{"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2945","title":"Protect master branch","comments":"Cool, I think we can do both :)","body":"After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:\r\n- 00cc036fea7c7745cfe722360036ed306796a3f2\r\n- 13ae8c98602bbad8197de3b9b425f4c78f582af1\r\n- ...\r\n\r\nI propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future:\r\n- [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch\r\n - Currently, simple merge commits are already disabled\r\n - I propose to disable rebase merging as well\r\n- ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~\r\n - ~~This protection would reject direct pushes to master branch~~\r\n - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~\r\n- [x] Protect the master branch only from direct pushing of **merge commits**\r\n - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch).\r\n - No need to disable\/re-enable this protection on each release \r\n\r\nThis purpose of this Issue is to open a discussion about this problem and to agree in a solution.","comment_length":8,"text":"Protect master branch\nAfter accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:\r\n- 00cc036fea7c7745cfe722360036ed306796a3f2\r\n- 13ae8c98602bbad8197de3b9b425f4c78f582af1\r\n- ...\r\n\r\nI propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future:\r\n- [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch\r\n - Currently, simple merge commits are already disabled\r\n - I propose to disable rebase merging as well\r\n- ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~\r\n - ~~This protection would reject direct pushes to master branch~~\r\n - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~\r\n- [x] Protect the master branch only from direct pushing of **merge commits**\r\n - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch).\r\n - No need to disable\/re-enable this protection on each release \r\n\r\nThis purpose of this Issue is to open a discussion about this problem and to agree in a solution.\nCool, I think we can do both :)","embeddings":[-0.2390099317,-0.0966292247,-0.0759872422,-0.0984264985,-0.0993319154,-0.114640519,0.0078532966,0.239554435,-0.029255297,-0.0760675296,0.3045841157,-0.0697917044,-0.1556909084,0.2472542524,-0.0918809772,0.2528043091,0.1680651754,0.000099914,-0.099873215,0.0868200585,-0.0079817604,-0.1016465425,0.1242587939,0.0464673564,-0.2174060941,0.0417824537,0.0499481186,0.1391150504,-0.4184132516,-0.4594882727,0.1383522302,0.1707512289,0.0033366256,0.5652734041,-0.0001042576,0.0216173306,0.2390211523,0.0108919349,-0.1405786723,-0.1769673228,-0.5225075483,-0.4049667418,-0.1415199637,0.0287014097,0.0748746395,0.06865637,-0.0910198018,0.002452912,0.3400871456,0.0976495519,0.2408002317,-0.1941754073,0.1793913692,-0.0078934981,0.3311534822,0.5539640784,-0.0518133156,0.22529535,0.0045611048,-0.0102421539,0.1125852838,0.2086250186,-0.0837501734,-0.2613006532,0.2912853956,-0.1761568338,0.5298508406,-0.3999355435,-0.1809146106,-0.1078854129,0.0704470351,-0.201073736,-0.3955834508,-0.2669799924,-0.1179080084,-0.0097809853,0.0528227128,0.2314448655,-0.0157607514,0.0743468106,-0.3007844985,-0.3514812589,0.0444366112,0.0477003977,0.10768985,-0.1619907767,0.0425621495,-0.0867641419,0.3245029449,0.0032727672,-0.1678859442,-0.2417692542,-0.3024554849,-0.0731806457,-0.1582976431,-0.166840598,-0.2660009265,-0.3856663704,0.4048146307,0.3533258736,-0.360022366,-0.049508784,0.002726417,-0.0647690669,0.2959358394,0.0719453767,0.011164953,0.0792105719,0.5598432422,0.3058459461,0.0217627417,0.3349214792,0.3275439441,-0.0626471415,0.0689635947,0.5172939301,0.5336612463,-0.2621914148,0.0377011895,0.1680787653,-0.3323733509,-0.3190573454,0.0054372335,-0.0610207208,0.0830994621,0.0703226849,-0.1091590971,0.0525613427,0.1415696889,-0.0016033769,-0.3210552037,-0.1457319856,-0.3556134701,-0.0552582555,0.1311259568,-0.4680353403,0.0908330008,0.4519638121,-0.1013908461,-0.2686688304,0.0358462371,0.1716575772,0.0330468714,0.2908923328,-0.1433112174,-0.3825237751,0.2031245977,0.0965870917,-0.0456247479,0.2545187175,-0.1935967952,-0.2513652146,-0.1866476983,0.3089985549,-0.1121701747,0.3287965655,-0.548268497,0.0025912365,0.0102211721,0.3561609685,0.4972600937,0.385391444,0.2942132354,-0.1614470184,0.0424310081,0.1232627109,0.3384101689,-0.0511925742,-0.0100153778,-0.2738169432,0.1200275049,0.2083038092,-0.2946696877,-0.0103482995,-0.0507854298,-0.1232980564,0.0196767803,-0.048742909,-0.204722017,-0.0717563778,-0.4525829852,-0.0241770819,-0.0117112603,0.0191363525,0.1543500572,-0.3124671876,-0.0787971914,0.1779266298,-0.182664752,0.1171256378,-0.3127051294,-0.6241436601,-0.1159733236,-0.0647230893,0.1302614063,0.1962026656,0.3260959387,0.0762788281,0.0484594442,-0.1529644579,-0.0238246322,0.1797253937,0.3943619728,-0.1591625661,-0.2575499117,0.0904227868,-0.3276287019,0.0429003425,-0.0230592508,-0.0326709487,0.0328225382,-0.4201252162,0.2604811192,-0.0395855084,-0.089378275,-0.1092364043,0.2325994372,0.0994815007,-0.2429064363,-0.2273789346,-0.1815333217,0.2851662338,-0.242647782,0.2751372755,0.099704735,0.0815574303,-0.1571651399,-0.0062341182,-0.0228314847,0.3285116851,0.1736403704,-0.0701381266,-0.04674327,0.1801669151,-0.0945321396,0.1870126128,0.1769949794,0.4175725579,0.3424830437,-0.1347359419,-0.0105623221,-0.1067967415,-0.1460241973,0.022494385,-0.2684330642,0.2932992876,0.0991126224,-0.2107091695,-0.249318257,-0.0687692985,-0.1965132803,-0.265663594,-0.1541765183,0.0057348073,-0.1855473518,0.1418430507,-0.2901535332,0.2540666461,-0.2189239413,0.1275674105,0.0980846509,-0.201123789,-0.0243991297,0.0869311392,-0.0290957168,-0.057731878,0.100935325,0.3220874071,0.0902120993,0.183042407,-0.0160376783,0.0336617008,0.2433368862,-0.0592605695,0.362724781,0.1478046626,-0.0405387431,0.0476890579,0.3128943741,0.1199872941,0.0844289884,0.3880383074,-0.0403469466,0.1266814917,-0.2439862639,-0.0744038522,-0.2096193433,-0.5341181755,0.1785611957,-0.1322654784,-0.311395973,-0.0278486367,-0.053532742,0.1170226783,-0.4517551064,0.1408915371,0.2340205312,0.2509391308,-0.3069441617,-0.0698819757,0.1471458375,-0.0289604329,0.0071960338,0.1242603362,0.4291484952,-0.2824301422,0.4524761438,0.2389160544,-0.1856772155,-0.4897386432,-0.4018841088,0.0115547851,-0.1668282598,-0.1153612956,0.0861761943,-0.0717044175,0.1023176908,-0.401612252,-0.3951958418,0.0295078922,-0.3599407971,-0.1393553913,0.0470686667,0.1184691191,-0.2150190026,-0.0969456658,-0.2935731709,-0.2871156633,0.5013284087,-0.0580234416,0.0385896452,0.0112076271,-0.2732456625,-0.1094526052,0.1394193172,0.0409913287,-0.0648903027,-0.3088234365,-0.0299696531,0.0454311818,-0.0143893873,0.368963629,0.2126246095,-0.3675787747,0.0550749078,-0.350450635,-0.1397126317,-0.0122180469,0.4691388905,0.256077528,-0.1587453038,0.0662926659,0.0626281053,-0.1768465936,0.1262467802,-0.1262247413,0.1179754212,0.3565621078,0.3317129016,-0.0136891427,0.1207261309,0.0874625519,0.6525163054,0.3680907488,-0.0682478175,0.1437117904,0.2640963793,0.0372981504,0.0415912867,-0.2894337177,0.0679146126,-0.0695572942,-0.0514505655,0.0428000949,-0.1358909458,-0.1156823337,-0.0192327965,-0.4392817616,0.0292560495,-0.4280707836,-0.212833792,-0.3319002986,0.1305010617,-0.0154377855,-0.4608305693,0.1475661546,-0.0283782165,0.2772078216,0.2418593913,0.1750047207,0.1717233062,0.0286009591,-0.0267298147,-0.143258512,-0.0784367099,-0.0277465973,0.2193704545,0.0662960187,-0.120667249,0.0654256046,0.0472218953,0.401168257,-0.0122205056,0.1574847549,-0.2568204701,-0.3890761137,0.6179253459,-0.2597557604,0.1685838103,0.3741835654,0.5948965549,0.4432282746,-0.3229280412,-0.1968980581,0.2175525427,-0.0058352049,0.1632378548,0.1841427237,0.0639553592,0.2376962006,-0.4858786464,0.2991431355,0.2369734645,-0.2511768043,-0.0432059951,-0.0785255134,-0.1137046441,0.0831007659,-0.0332846567,-0.2697017193,0.0042523337,-0.0428405702,0.1484801471,0.103806518,0.1191856787,0.3641275465,0.5810796022,0.2270472646,-0.4135166109,-0.0292455852,-0.1834749579,-0.0771063045,0.3103395998,0.1818852574,0.2265654951,0.0643974394,0.2080049366,-0.1813084334,-0.2263067961,-0.2212891281,0.0309540853,-0.2094345242,-0.0052669551,0.1052183583,0.0515394434,-0.0904067382,0.0824889988,0.0987536013,-0.3212070465,0.1318824291,0.5714728832,0.955177784,0.3502957225,0.1090738475,-0.2922923863,-0.2720266581,0.3813621998,-0.357526511,-0.1286929101,-0.3113082349,-0.1811920255,-0.1677663922,-0.3074550629,-0.0780642331,-0.4934645891,-0.3175655603,0.1523682475,-0.1961845458,0.5375774503,0.0177650489,0.1167705655,0.1282786429,-0.038145341,0.2970950305,0.1892168671,0.0111535219,0.1593013853,-0.0028083171,-0.4545554519,-0.0279463418,0.0025342032,-0.4306589961,0.0386116318,-0.3405902982,0.0399823673,-0.2339083552,-0.2824401855,0.1117530316,-0.1104251817,0.6192416549,0.2143154591,-0.0984719917,-0.0485647954,0.2703552842,0.3980438113,-0.197127685,-0.196609512,0.5138126016,-0.2046020627,-0.1562071592,-0.0987637788,-0.0285022315,-0.3348375857,-0.2822057903,-0.0767591447,-0.3614266813,0.0732668266,-0.0306378119,0.2732182145,-0.0691128522,-0.1619696319,0.1780715436,0.111415118,-0.1410685033,-0.1422604918,-0.1091866344,-0.2943277955,-0.0380600765,0.1902545691,0.2271509767,-0.1096624061,-0.0836395845,0.0573430769,-0.1975595355,-0.2153514922,-0.2259288728,0.1200701594,-0.1281341463,0.0691441,-0.4934980571,-0.3398421407,0.1174373701,0.1943581998,0.2834068835,0.3471649885,0.07340087,-0.165871352,-0.2640971243,0.2091121227,-0.3319115043,0.1677211374,0.0679885969,0.2423418313,-0.0730114356,0.0273042507,-0.3893203735,0.1249633506,-0.1246866807,-0.0371800624,-0.1938935816,-0.076455161,0.002154053,-0.4198532403,0.0430044867,0.0236390047,-0.1842414141,-0.2793491185,0.0818327814,0.0688198656,0.0645905882,0.0382995978,-0.0577248856,-0.1069227234,0.1288665682,0.3231302798,0.2298199683,-0.0998132899,0.1544985622,-0.0540686958,0.1755025834,0.0916104764,-0.0348069631,0.1178810969,-0.012650379,-0.2803532481,-0.1100896522,0.3647060692,-0.1755682975,-0.1734805107,-0.3497366011,0.1640637666,0.0974830911,0.6140381098,0.110363245,0.1082694158,-0.4196391404,0.2307030708,-0.0214151721,-0.0301906541,-0.1783225387,-0.1215486825,-0.0753639936,0.2950595319,-0.0994521976,-0.3054563999,0.2349793315,0.1782335341,0.0241322219,0.0836851299,0.5049234629,0.1231228486,-0.2381349504,0.4243248105,0.2228861451,-0.1286406517,0.455909878,0.5233042836,-0.1415461004,-0.0785631537,-0.0360233933,0.0871152431,0.1805791706,-0.1823303849,0.0883068964,0.3238530457,0.6459390521,-0.1694083512,0.3705488145,0.0255201813,0.3059286177,0.2132881433,0.0270452425,-0.2294175178,0.4330947399,0.2575822473,0.3352282941,-0.5901472569,0.1695657074,0.5056748986,-0.023568213,0.2167177498,-0.3594747484,-0.0223750621,-0.0552352853,-0.0159900282,-0.0719330311,0.1218386665,0.1490119845,0.2901292443,-0.0673369542,-0.0273434464,-0.0467063077,0.2230517119,0.2663088441,-0.1428412646,-0.2008356899,0.2622175515,-0.2484770417,0.0786395818,0.2177159339,0.1616025418,-0.0978315324,-0.0521366596,0.1798922271,0.4497513771,-0.1746967882,-0.039889738,0.0317630544,-0.1530031413,0.1404958814,0.0627516881,0.2746909261,-0.0635648817,0.1233454049,0.0005846465,-0.383130908,-0.0437247008,0.4243560433,0.4585390389,-0.3600539863,0.41664958,0.174680531,0.2586427629,-0.0644906461,0.4355578721,-0.0931281298,0.1283278465,-0.0560730807,0.1473985761,0.0403031036,0.217051357,0.0914419889,-0.0287602842,-0.0969899297,0.0412180498,-0.2160513252,0.2098333389,0.3064318895,0.1730033755,0.4223076701,-0.3031860292,0.1374098659,0.1526105702,-0.0731561556,-0.2435802221,-0.047290422,0.1329396963,-0.3592749536,0.0106856637,0.2144466639,-0.0846365243,0.01132154,0.064479664,0.4179263711,0.262011826,0.1000918522,0.0653353333,-0.0138661312,0.2816164494,-0.3647170663,0.2353287637,0.1579896212,-0.1405763328,-0.2575822473,-0.0880536586,-0.0056584454,-0.0124874227,-0.4783657789,-0.0724288374,-0.1397930533,0.4400902689,-0.1443308294,0.0493659228,-0.3884942234,0.0546305478,-0.0210841242,0.2141258866,-0.3707188964,0.1099852324,0.0246811155,0.0773969591,0.0523463786,0.1700744331,0.0661983937,0.2172149122,-0.2984472513,-0.1805216223,0.3678478599,0.0187848061,0.0980769023,-0.15001674,0.4350650012,0.5103737712,-0.2477454543,-0.3580597639,-0.2498808056,-0.0220319424,-0.2263912112,-0.4824006557,0.2833116651,0.0979687423,0.2627729475,-0.0858350545,0.0814130977,-0.069908455,-0.2276804298,0.164998427,-0.2625057101]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2945","title":"Protect master branch","comments":"@lhoestq now the 2 are implemented.\r\n\r\nPlease note that for the the second protection, finally I have chosen to protect the master branch only from **merge commits** (see update comment above), so no need to disable\/re-enable the protection on each release (direct commits, different from merge commits, can be pushed to the remote master branch; and eventually reverted without messing up the repo history).","body":"After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:\r\n- 00cc036fea7c7745cfe722360036ed306796a3f2\r\n- 13ae8c98602bbad8197de3b9b425f4c78f582af1\r\n- ...\r\n\r\nI propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future:\r\n- [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch\r\n - Currently, simple merge commits are already disabled\r\n - I propose to disable rebase merging as well\r\n- ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~\r\n - ~~This protection would reject direct pushes to master branch~~\r\n - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~\r\n- [x] Protect the master branch only from direct pushing of **merge commits**\r\n - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch).\r\n - No need to disable\/re-enable this protection on each release \r\n\r\nThis purpose of this Issue is to open a discussion about this problem and to agree in a solution.","comment_length":64,"text":"Protect master branch\nAfter accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:\r\n- 00cc036fea7c7745cfe722360036ed306796a3f2\r\n- 13ae8c98602bbad8197de3b9b425f4c78f582af1\r\n- ...\r\n\r\nI propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future:\r\n- [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch\r\n - Currently, simple merge commits are already disabled\r\n - I propose to disable rebase merging as well\r\n- ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~\r\n - ~~This protection would reject direct pushes to master branch~~\r\n - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~\r\n- [x] Protect the master branch only from direct pushing of **merge commits**\r\n - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch).\r\n - No need to disable\/re-enable this protection on each release \r\n\r\nThis purpose of this Issue is to open a discussion about this problem and to agree in a solution.\n@lhoestq now the 2 are implemented.\r\n\r\nPlease note that for the the second protection, finally I have chosen to protect the master branch only from **merge commits** (see update comment above), so no need to disable\/re-enable the protection on each release (direct commits, different from merge commits, can be pushed to the remote master branch; and eventually reverted without messing up the repo history).","embeddings":[-0.1553194672,-0.1002306268,-0.070321016,-0.0798171908,-0.1042534709,-0.1879874468,0.0104035577,0.2728635967,-0.0098489737,-0.0841399208,0.2926148772,-0.0778751224,-0.1324572563,0.2158903182,-0.0562292449,0.1705456823,0.2003232092,-0.0407981202,-0.0923330337,0.0279233586,0.0226084664,-0.0687029436,0.0832000151,0.0436301306,-0.1837819666,0.0054013738,-0.0148117775,0.1564962417,-0.4397666156,-0.4695702493,0.1452465504,0.2322402596,0.0251065325,0.570535183,-0.0001043637,-0.0003454557,0.1694866419,-0.0406010039,-0.1401022524,-0.2154636681,-0.4901072085,-0.4447958767,-0.07006336,-0.0083542913,0.1288067251,0.0528678969,-0.132499963,0.0238457564,0.3848451674,0.1055523977,0.2386109382,-0.1028517336,0.1812405735,0.0137694413,0.392008841,0.5474528074,-0.066634573,0.3031997979,0.0540179275,0.0006757497,0.0611058585,0.1741086096,-0.0823729113,-0.2866700888,0.2270194292,-0.1779912561,0.5181742311,-0.4243531227,-0.2415838391,-0.0831176639,0.0523793846,-0.1781465262,-0.4089041948,-0.2419367433,-0.1499476582,0.0076339091,0.0677802935,0.2468644679,0.039739579,0.0036496122,-0.3773649931,-0.2718090415,-0.004407926,0.0492040403,0.141940549,-0.1626293063,0.0664263666,-0.1270385236,0.2942230105,-0.0473279878,-0.1469833106,-0.3676346838,-0.2413333207,-0.0125874653,-0.1026133746,-0.1957468837,-0.2462065965,-0.3248277903,0.4251304567,0.3889683783,-0.3832727969,0.0137396995,-0.0304892361,-0.0810169131,0.2071851492,0.1178578511,-0.0104900496,0.0584886484,0.530713141,0.3144408166,0.0051584225,0.3255857527,0.3015701473,0.0030471815,0.0344459116,0.574118197,0.4289444685,-0.26292032,0.1276476383,0.1959735751,-0.3412007391,-0.3845081031,0.0711311549,-0.0861726329,0.1103260964,0.0271337908,-0.1147114784,0.0668015406,0.0451908335,-0.0380863659,-0.3079027534,-0.141485557,-0.3744001687,-0.0685305148,0.1860401928,-0.4571051598,0.1491677761,0.4325242639,-0.0680372268,-0.1992845535,0.0445522442,0.1292257011,0.0555874184,0.2961196899,-0.1537979692,-0.4526284337,0.2235967368,0.1313333064,-0.0668994114,0.1940043122,-0.1598993689,-0.2616221011,-0.1049757376,0.2977656722,-0.1482721567,0.3071561158,-0.5237665772,0.0363732092,-0.0044837268,0.3506707847,0.4910375476,0.3025480509,0.3280785084,-0.1685939282,0.0617566295,0.1903455257,0.3235116601,-0.0516864397,0.0623045377,-0.2877363563,0.1095039099,0.1450733244,-0.3121602237,0.0175557248,-0.0052360594,-0.1312822402,0.0107761566,-0.0714070052,-0.1119878516,-0.0989294425,-0.3666853309,0.0694030076,-0.0407375731,-0.1007770002,0.2079568803,-0.3201756477,-0.0947941095,0.2050814927,-0.1481231451,0.0208985135,-0.2919872105,-0.6061385274,-0.1636558473,-0.06343887,0.1249144152,0.2275642306,0.3607693315,-0.0050310493,-0.0283617247,-0.1714023799,-0.0093972106,0.1125069708,0.3421364129,-0.1496780068,-0.2415423691,0.1322465539,-0.3545884192,0.0911409035,-0.0785121024,-0.0097747799,0.0344485678,-0.3157884181,0.3412798941,-0.0440529883,-0.1208016798,-0.1141509414,0.2415172756,0.095951654,-0.1747335494,-0.2039210498,-0.2074069828,0.2601909339,-0.2408306301,0.3018657863,0.1300136894,0.1026374176,-0.1411807239,-0.0364533328,-0.0250960924,0.36193645,0.2363044769,-0.0899174288,-0.0591584407,0.1687609404,-0.0490626656,0.2035702467,0.2246947736,0.4252423346,0.3736152053,-0.1026815102,0.0626009405,-0.0917925984,-0.0787246898,0.0245947056,-0.3291273713,0.3280371726,0.073694557,-0.1394644678,-0.1740321219,-0.1033326834,-0.1740740985,-0.2202692628,-0.1707722992,-0.012951361,-0.1690982729,0.1943652332,-0.3365918398,0.2223692834,-0.2772080004,0.148579374,0.0150526064,-0.2516064942,-0.1024010181,0.0682980418,-0.0059674503,-0.1327149123,0.1241709664,0.3773829341,0.1278743744,0.1701000035,0.0042741634,0.0201513059,0.2572831213,-0.1164359227,0.353325963,0.1443069726,-0.1054452285,0.054064367,0.3364559114,0.1812186539,0.0643318817,0.346919328,-0.0033243245,0.0352581739,-0.2084068358,-0.0459848903,-0.2602106631,-0.5283415914,0.2192376256,-0.110637784,-0.3552163243,-0.0085041393,-0.0453661792,0.047418233,-0.5248722434,0.136029914,0.1835808307,0.2905260623,-0.3178029358,-0.1199780703,0.1263155937,-0.0225951802,0.0123441136,0.1285599768,0.3631602824,-0.2266809344,0.45945099,0.2389809936,-0.1695563644,-0.4726397991,-0.3434132934,0.0191301201,-0.2039363831,-0.1179748252,0.0098189935,-0.0989309773,0.0921538174,-0.3585401177,-0.3257124722,-0.018773742,-0.3730046451,-0.1600959152,0.0695955083,0.0628860295,-0.2346176803,-0.114040032,-0.386590004,-0.2821527123,0.4793556035,-0.0164330266,0.1156589612,-0.0431727543,-0.2923772335,-0.0459178463,0.0874827728,0.1366390288,-0.0000448525,-0.3609322906,0.0325667374,-0.0406603813,-0.0293471292,0.352609843,0.1854910851,-0.4302407205,0.0712517202,-0.3217604458,-0.1558190286,-0.0520843901,0.4864443243,0.3141199946,-0.0973737389,0.0710799471,0.1035510823,-0.1919001341,0.0550895296,-0.1816118211,0.1472238153,0.3454884887,0.2857186198,0.0378993787,0.1458674669,0.0254973993,0.7140879035,0.3778497577,-0.059657298,0.210337624,0.3106974065,-0.0277180839,0.0039902786,-0.2601569593,0.065046832,-0.1012431234,-0.0000209612,0.0867039785,-0.1295782924,-0.1229653433,-0.0112106856,-0.4777139425,0.0091558872,-0.4146761596,-0.2396630496,-0.3181839585,0.1842502356,-0.0225256812,-0.4854888022,0.1313801706,0.0537709594,0.2305762768,0.2159709483,0.2142065167,0.1812160015,-0.0202590171,-0.0510845743,-0.1200738922,-0.067726247,-0.0043309759,0.145250544,0.0905801877,-0.1670407504,0.0318057276,0.0695502236,0.3299908936,-0.0148684094,0.1402913779,-0.1708071232,-0.5094846487,0.6086307168,-0.3185944259,0.1394028515,0.2990989089,0.5112247467,0.4397634268,-0.2954540849,-0.1953165382,0.2071224451,0.0325558819,0.1978364438,0.1773502678,0.1449773014,0.2783072591,-0.4903811812,0.2887739539,0.1548331976,-0.2504818439,-0.1551757604,-0.0039056637,-0.0270501971,0.1264622211,-0.0263037495,-0.2453983575,0.0091275191,-0.1094373837,0.1734458208,0.1329204887,0.0698278099,0.4152193964,0.594658792,0.1349339038,-0.3891987503,-0.0482011773,-0.1750733256,-0.0742834732,0.3271439075,0.1983511001,0.24822779,0.0203486141,0.1776862442,-0.1234726384,-0.2662384808,-0.2033664733,-0.038491372,-0.0767425969,0.0075890222,0.0873292238,-0.0458977483,-0.0909918919,0.0130487047,0.1399983913,-0.3661002219,0.1637848169,0.6180499792,0.9825440049,0.3975676596,0.0890336856,-0.2546566427,-0.3273829818,0.3259905577,-0.340107888,-0.1664507836,-0.2470169514,-0.1235878542,-0.1842461079,-0.3403116167,-0.0386944152,-0.4898471832,-0.3528554738,0.1351124644,-0.2180551142,0.5279191732,0.008177679,0.1233861297,0.1910188198,-0.12231455,0.2551510632,0.1982705742,0.061811205,0.104095228,-0.0040276642,-0.4687906206,-0.0022327867,0.0205386523,-0.3772695959,0.0891471654,-0.3437781334,0.0388470143,-0.2063531876,-0.3160520196,0.115827091,-0.0870479271,0.6059409976,0.1666302532,-0.0794522315,-0.0446208119,0.2913893759,0.4645246863,-0.2288517058,-0.219530493,0.5223725438,-0.2133638412,-0.1302172095,-0.2047614008,-0.0803344026,-0.2370928973,-0.2609480619,-0.1280429661,-0.3282325566,0.0795017928,0.041714251,0.2636537552,-0.0910693035,-0.1676060557,0.1859930903,0.1154857874,-0.1075729132,-0.118584998,-0.1695205718,-0.2610394955,-0.0678159222,0.2016011029,0.161605835,-0.1206071749,-0.0767639652,0.099657312,-0.2443365902,-0.2207894772,-0.2072923779,0.105987981,-0.1394685954,0.0929813311,-0.5486782193,-0.32993415,0.0517466478,0.2461933494,0.279415518,0.3304983079,0.1179999709,-0.0997705981,-0.2244495153,0.2186434418,-0.3279267251,0.1686490178,0.0235842206,0.1588117927,-0.0260097533,0.0404301658,-0.3905956745,0.160315454,-0.1254802048,-0.0211261567,-0.2255206406,-0.1017537564,-0.000630246,-0.3081870973,0.0324874744,0.0234331191,-0.1957566142,-0.2830150425,0.0926690474,0.0656956136,0.0647893623,-0.0570135079,-0.0574608184,-0.1081577986,0.0827530622,0.2720321715,0.2396921068,-0.1584913284,0.1097129956,-0.0322248675,0.1710882336,0.1105851158,-0.0247724894,0.0892416015,0.0146322586,-0.2658016682,-0.0353747047,0.267159909,-0.1817460209,-0.2171436995,-0.2685393691,0.1888355613,0.1056337059,0.5840907693,0.2028801143,0.1442169249,-0.405082494,0.1754089892,-0.0531977862,-0.0036356284,-0.1692355126,-0.0969402716,-0.0849622712,0.2793601751,-0.0867193192,-0.3035972714,0.2713837624,0.1881819069,0.0007486099,0.1049711928,0.5280760527,0.1296208501,-0.1896519363,0.4221355915,0.2712435424,-0.0876758322,0.494900018,0.470233053,-0.1131819189,-0.0203250889,-0.082736142,0.116614446,0.2306870967,-0.2540961802,0.1284718663,0.2962196171,0.6462965012,-0.1688840091,0.3935549557,-0.0330962837,0.3437395692,0.1979931891,0.0587407239,-0.2460495532,0.3927094042,0.2117075175,0.3301511109,-0.6189893484,0.1363731921,0.4914032817,0.0492829382,0.1995816529,-0.400105536,-0.0492327139,-0.0584757105,-0.0483128205,-0.0781826749,0.1634954214,0.1237376407,0.2829664946,0.0065842085,-0.088377893,-0.043440666,0.1850482225,0.2415616512,-0.1169977486,-0.1811731756,0.357754916,-0.2761978507,0.0664936379,0.215460062,0.1350683123,-0.074716486,-0.0686341003,0.1802342385,0.4503947198,-0.1699341387,0.0225111321,0.0557496101,-0.1625677198,0.1577177346,0.0801470727,0.2580322921,-0.0533408038,0.1128025502,0.0354268774,-0.4152979553,-0.025322238,0.4330811501,0.5564644337,-0.3377302289,0.4601393938,0.1192681268,0.2847007811,-0.0326848105,0.4374352098,-0.0471260622,0.1163921505,-0.0051536481,0.143078804,0.02808984,0.2575880587,0.1414998174,-0.0333454311,-0.1592887938,0.0301065128,-0.1215411872,0.127435267,0.2968399227,0.17307055,0.419888705,-0.2947140038,0.1275687069,0.1282345057,-0.0875694603,-0.3413968384,0.0095217237,0.0940847918,-0.3605292141,0.0042476766,0.2224816233,-0.0694397017,-0.0241943654,0.1189339831,0.4094779193,0.1927902848,0.1120927632,0.1051306576,-0.0493477993,0.3209429979,-0.3918459415,0.2249616981,0.1892230064,-0.1029298455,-0.2218884081,-0.0510987751,0.0073658284,0.0212806091,-0.4355088174,-0.1401801854,-0.162226066,0.4209837019,-0.1355972886,0.0682982057,-0.3722180128,0.0687672645,-0.0371344872,0.2683559954,-0.3000992239,0.0864084363,0.0525472984,0.017337583,0.088370651,0.1305798739,0.068434529,0.2528506517,-0.3523156941,-0.1248895153,0.3586619198,-0.0194706563,0.0910360888,-0.2172687799,0.4182040691,0.4654460251,-0.1480811983,-0.3321150839,-0.2635183036,-0.0111563709,-0.2798339725,-0.542089045,0.2986377478,0.1378680468,0.1974387169,-0.1058711782,0.0366066247,-0.0743264556,-0.1527637988,0.0611719489,-0.1987861097]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"Hi ! I guess the caching mechanism should have considered the new `filter` to be different from the old one, and don't use cached results from the old `filter`.\r\nTo avoid other users from having this issue we could make the caching differentiate the two, what do you think ?","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":50,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nHi ! I guess the caching mechanism should have considered the new `filter` to be different from the old one, and don't use cached results from the old `filter`.\r\nTo avoid other users from having this issue we could make the caching differentiate the two, what do you think ?","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"If it's easy enough to implement, then yes please \ud83d\ude04 But this issue can be low-priority, since I've only encountered it in a couple of `transformers` CI tests.","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":28,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nIf it's easy enough to implement, then yes please \ud83d\ude04 But this issue can be low-priority, since I've only encountered it in a couple of `transformers` CI tests.","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"Well it can cause issue with anyone that updates `datasets` and re-run some code that uses filter, so I'm creating a PR","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":22,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nWell it can cause issue with anyone that updates `datasets` and re-run some code that uses filter, so I'm creating a PR","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"I just merged a fix, let me know if you're still having this kind of issues :)\r\n\r\nWe'll do a release soon to make this fix available","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":27,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nI just merged a fix, let me know if you're still having this kind of issues :)\r\n\r\nWe'll do a release soon to make this fix available","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"Definitely works on several manual cases with our dummy datasets, thank you @lhoestq !","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":14,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nDefinitely works on several manual cases with our dummy datasets, thank you @lhoestq !","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"Fixed by #2947.","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":3,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nFixed by #2947.","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2941","title":"OSCAR unshuffled_original_ko: NonMatchingSplitsSizesError","comments":"I tried `unshuffled_original_da` and it is also not working","body":"## Describe the bug\r\n\r\nCannot download OSCAR `unshuffled_original_ko` due to `NonMatchingSplitsSizesError`.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> dataset = datasets.load_dataset('oscar', 'unshuffled_original_ko')\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=25292102197, num_examples=7345075, dataset_name='oscar'), 'recorded': SplitInfo(name='train', num_bytes=25284578514, num_examples=7344907, dataset_name='oscar')}]\r\n```\r\n\r\n## Expected results\r\n\r\nLoading is successful.\r\n\r\n## Actual results\r\n\r\nLoading throws above error.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.4.0-81-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":9,"text":"OSCAR unshuffled_original_ko: NonMatchingSplitsSizesError\n## Describe the bug\r\n\r\nCannot download OSCAR `unshuffled_original_ko` due to `NonMatchingSplitsSizesError`.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> dataset = datasets.load_dataset('oscar', 'unshuffled_original_ko')\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=25292102197, num_examples=7345075, dataset_name='oscar'), 'recorded': SplitInfo(name='train', num_bytes=25284578514, num_examples=7344907, dataset_name='oscar')}]\r\n```\r\n\r\n## Expected results\r\n\r\nLoading is successful.\r\n\r\n## Actual results\r\n\r\nLoading throws above error.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.4.0-81-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nI tried `unshuffled_original_da` and it is also not working","embeddings":[-0.2700701952,-0.3562059104,0.0761256963,0.3269021809,0.2630259991,-0.0228649452,-0.0282609034,0.4201454818,-0.0057781218,0.2675718367,-0.2937448621,0.3795824349,-0.175184533,0.0020705916,-0.1521454006,-0.3282049894,0.0055401283,0.0813263729,0.0597292781,0.0331693552,-0.2423605472,0.2973343432,-0.3931003809,0.0419516861,-0.2768164873,-0.0585272759,-0.0746447295,0.1445952952,-0.1675010622,-0.2716379166,0.3857498765,-0.0588463284,0.232816115,0.5122596025,-0.0001237657,0.0590177849,0.4652545154,-0.3251428604,-0.0443532206,-0.3920538425,-0.0886344537,-0.0983479768,-0.0657844767,-0.3845169842,-0.0129822195,0.146921888,0.0608098917,-0.0645650402,0.5164781213,0.3238581419,0.120624736,0.0500490926,0.0403883979,-0.2224902213,0.3106618822,-0.2548942566,-0.3315367401,0.0346963368,-0.0179840401,-0.1685461998,0.1427149326,0.1541062146,-0.0700180754,0.1261722445,-0.1210525259,-0.0195195898,-0.0040011993,-0.0904255956,0.2117808163,0.374551028,0.3375654519,-0.3247810006,-0.3132323921,-0.3029401302,0.0337986127,-0.3283481598,0.3009405434,0.1460143775,-0.2429262847,0.2017060965,-0.0670581907,0.0023002515,-0.1089840308,0.1734156609,0.0992210284,0.2540248334,0.068739228,0.1693916321,0.1453325152,0.1248955354,0.3683691025,-0.164793402,0.1000258476,0.0660203919,-0.2375979722,-0.1042992249,-0.1702269167,0.0275418647,0.1742199361,0.299749881,0.3533665538,-0.050712835,0.23539868,0.0687230006,0.1073275581,0.1424306631,-0.1735071242,0.4756015539,-0.038373746,0.4154306054,-0.1177193224,-0.0899807885,0.0127301393,-0.0728332028,0.0770530924,0.0638824552,0.3465489149,-0.1475606412,-0.5405621529,-0.0646340847,-0.4408827126,-0.2136350721,-0.1520271748,0.3651462495,-0.2901136577,0.1182295531,-0.1973749697,0.2700160146,-0.1282483786,0.043910075,-0.2218047082,0.1398160458,-0.1717366427,0.1828228682,0.1118824482,-0.204669714,0.495223105,-0.0624024421,0.2177726328,-0.1745524108,0.2678282857,-0.0713035017,-0.2555179,0.0733668804,0.0693256706,0.2275202274,-0.0512226708,-0.1955409944,-0.1835986227,0.1897325665,-0.1677214205,-0.5662183762,0.1641817391,0.1281939894,-0.3114531338,-0.0909331217,0.2412121892,-0.06287238,0.0155445421,-0.1351996511,-0.0021220825,-0.0958602279,0.1657218486,-0.2155779004,0.2271082997,0.3483718336,-0.1850482076,-0.1841768026,0.006626484,-0.2098627985,0.5653936863,0.3381557167,0.0346748829,-0.0691285431,-0.3662846684,-0.035242945,0.2375092804,-0.1608842909,-0.6755478978,0.0946134478,0.0492238738,0.5803261995,0.346126765,-0.0058211032,0.2652187347,0.1256498843,0.2001138926,0.2200994045,-0.2519147992,-0.3296817243,-0.4716974497,-0.4451117814,0.1734127849,0.1898487657,0.3207941949,0.1099775285,-0.058488369,0.2988939583,0.3775659204,0.0346811637,0.0107769752,-0.0222215783,0.1482665092,0.3807443678,-0.076375179,-0.3940159082,-0.2746164203,0.2904383838,-0.1236760765,-0.156783849,-0.2122336924,-0.527250886,-0.5870789289,-0.165048629,-0.1906133443,-0.162727192,0.0038345398,0.0500118583,0.4496047795,0.1891618371,-0.1783928275,0.2574459016,0.0525147803,0.1223757267,-0.1573136151,0.4813087285,-0.2699065506,-0.4087940454,0.1640718877,0.1615529507,0.2917622626,-0.2825148404,-0.1289041787,0.5438861251,-0.0126622617,0.1553732008,-0.0856645703,-0.2544069886,0.1181099713,-0.167449981,-0.1110598743,0.3721723258,0.0395096019,0.0105408933,-0.0509868823,0.5377941132,-0.0667465031,-0.0382976942,0.066784896,0.0064676078,0.2481441945,-0.1786244661,0.0006179169,-0.1696993709,0.3078433871,-0.0635774508,-0.0126545299,-0.1987084448,-0.5619574189,-0.1079929173,0.7206522226,-0.0952707604,0.0389066115,0.0032789083,-0.1174628735,0.1458038688,-0.0419397578,0.1237870455,0.2527246773,0.3046537042,-0.0554859377,0.1650494337,-0.0884823874,-0.2226391435,0.1139543355,0.0752876177,0.3927039802,0.5039468408,-0.1236994937,0.0763900355,-0.4040017426,-0.1757810861,-0.0385836139,0.3962652385,-0.0995420963,-0.251657933,-0.3585939407,-0.0895016566,-0.3763104379,0.0013211808,-0.1984645426,-0.1646792293,0.0933002308,0.4677838087,0.03062311,0.456158936,-0.2225827277,0.0516182445,-0.0948610157,0.0804057792,-0.2349837869,-0.2070280761,-0.2165757865,-0.0062309718,0.1760603487,-0.1895347685,0.4101438522,-0.1628828794,-0.1106793508,0.0016370228,-0.2100382894,-0.0284292269,0.0802166983,-0.1457641721,0.1499073058,0.1545280069,0.0492399521,-0.2327251136,0.3539457917,-0.253189683,-0.12509799,-0.0579356998,0.3628317118,0.1966164857,-0.0150078526,-0.6001400352,-0.3433532715,-0.2735929787,0.1192297861,0.171149075,0.2071451992,-0.4059786499,0.2124399543,0.0293269083,0.0146261109,-0.108423464,-0.1782411635,-0.2187789083,0.4179231524,-0.0541067719,-0.2413789183,-0.0135591552,-0.0757790357,-0.207618624,0.1081376076,-0.2254442275,-0.2374272496,-0.1528240591,0.0388922058,0.0000410417,-0.1447942406,0.4625220001,-0.0406659842,0.0200901981,-0.087360315,-0.2536268234,0.0319263563,-0.0870750993,0.4643537402,-0.0956633613,0.5923442245,0.1564027518,1.0513212681,0.3848242462,0.4277364314,0.2050092816,-0.2445764095,-0.0273402222,-0.1820255518,-0.1257591844,-0.0314858034,-0.1415696144,0.1885153949,0.1321771443,-0.0758097246,-0.1143291369,-0.2025897503,0.0363576189,-0.4454913139,-0.1845194995,-0.050208386,0.0200488307,-0.0743915215,0.2265842706,0.1607146412,0.0683756843,-0.0109747304,0.1947402656,0.2533482909,-0.095236063,0.0733086839,-0.1778615266,0.0034823706,-0.4660815597,0.4068314433,0.1445844173,0.155201152,0.105390504,-0.2482506037,0.2122097313,-0.2690775096,0.8391341567,-0.3196516037,0.021446595,-0.0050391215,-0.1162589341,-0.2233094126,-0.3100708723,-0.2493042648,0.0977460891,0.5004122257,0.401702255,-0.3936049342,0.1048643887,0.1261661202,-0.0172707625,-0.1737206727,-0.1427190602,-0.3067505658,-0.416105777,-0.521835506,0.0087642455,-0.10390421,0.4622172713,-0.103750132,-0.1932160407,-0.2216338515,-0.0615733713,0.1528482884,0.0589452758,0.1914952844,0.0529781282,0.2729120851,0.0143474173,0.5659828782,0.3270242512,0.433042109,-0.0510436259,-0.2842541039,-0.2222711891,-0.2774918079,-0.0777949691,0.3124807179,0.0741247982,0.0557200424,-0.2258193046,0.0286223143,-0.0232267417,-0.0609401241,0.2382368445,-0.074377045,-0.4884041846,-0.4397383332,0.2568946183,0.2011566609,-0.084970735,0.2157151252,0.1666444838,-0.1530286074,0.3322291672,-0.311853826,1.0656917095,-0.0349132605,-0.0590303577,0.1567310691,-0.4327569604,0.4600604773,0.2230525613,0.2934070528,-0.3985666931,-0.5422777534,-0.0626290441,-0.0495161489,0.2620194554,-0.0702634975,-0.2358459234,0.3924998343,-0.1383437216,0.059872508,-0.0403107777,0.0872835815,-0.0655479878,0.2615909278,-0.2137582153,-0.0282094814,-0.0062475386,0.6459135413,0.0490520187,-0.1735088825,-0.3161623478,-0.1341487914,-0.2399571091,0.2692095339,-0.0620518588,0.1432400197,-0.1940255165,-0.3515720665,-0.0938458592,-0.0665402114,0.1554252207,-0.3038245738,-0.1423941851,0.2567335367,-0.0502586178,-0.3049413562,0.0468185209,0.0734311789,0.1184430718,-0.2098308355,-0.1948067993,0.2171829045,0.081996724,-0.2234715521,0.2245999873,-0.344912529,0.4029507935,-0.2193651795,-0.3630694151,-0.1238867566,0.053775616,-0.4139072001,0.0482330695,-0.3476900756,-0.1793532968,0.0908351913,-0.0968465805,-0.1877360195,-0.2056991607,0.6468623877,0.2441388518,0.0889284089,0.4455735385,-0.1392742097,-0.1209909022,-0.0694017634,-0.0522745661,-0.0961838588,-0.6258867979,0.2478279322,-0.2725172639,0.2561436892,0.0560390577,0.1203305498,0.5102044344,0.1091660559,-0.0226831455,-0.315593183,-0.1827117205,-0.006129405,0.3605924249,0.2034677416,-0.2084128559,0.3939769268,0.0435558408,0.1410049051,-0.2295722514,0.1756058782,-0.0111234821,0.0903083682,-0.0150688291,-0.0624070913,0.0482578315,0.1309499592,0.0656671524,0.2593142688,-0.107485652,-0.1093422994,-0.1164527461,0.217103675,-0.0287401099,-0.0509174541,-0.0877763107,0.0597584993,-0.014149596,0.0015855627,-0.0073475796,0.3024415374,0.0025629182,0.1959821284,0.2877029181,0.0203361772,-0.3707616925,-0.0546788871,-0.1704271585,0.4047979116,-0.0256696772,0.2317973226,-0.0267457049,-0.0181804169,-0.0715063438,-0.3175280988,-0.0306168441,0.2115057111,0.3874916732,-0.347990185,-0.0214854032,0.3729046881,0.2182453871,0.5003054142,-0.1019885391,-0.1947451532,0.2794112563,0.132047534,-0.5035033822,0.1693591028,0.6532273293,-0.0279648732,0.0474482439,0.2889039218,-0.0722825229,-0.0049079503,-0.2585618496,0.0728518143,0.311209619,0.0896963254,0.3967244029,0.6500381827,0.2969564795,0.0609108508,0.1803033948,0.0377423726,0.3314584792,0.4642525315,-0.348226279,0.2353744805,-0.3763629198,0.2772921026,-0.0461748801,-0.2185561657,-0.2172773033,0.3019171655,-0.263828367,0.2700886428,0.147561118,0.2971718013,-0.0907789618,-0.0566396452,-0.2395830452,0.0068022935,0.0361580849,-0.1487154663,-0.23768875,-0.3082600832,-0.3154895604,0.1420746595,0.0483601578,0.2234884501,0.4846854806,-0.0899223089,-0.0839001238,-0.261703372,-0.1129587889,0.1009543613,0.4807308614,-0.359243691,0.002184781,0.1473571062,-0.2205985636,0.0831326023,0.3578595817,0.2714527547,0.2029879987,-0.224526599,0.0407990441,0.2606113255,0.0910466984,0.1569049805,0.3494497836,0.3068242073,-0.0663663,0.2790664434,0.0528981723,-0.1823626012,0.1601461917,0.0584523901,-0.1115810052,-0.2788190246,0.3917230964,-0.2243009657,0.102521576,0.057680171,0.0891512558,-0.4057340622,-0.2379486412,-0.2201809883,0.0126067307,0.231259793,0.0904681012,0.0248190816,-0.3940022886,0.1682776958,0.2681089044,-0.1561311036,-0.1225692406,-0.2420620918,-0.5433617234,0.2975563407,-0.2807193995,0.0935659185,0.0746108741,0.3382504284,-0.3834771514,-0.0064511839,0.1178494468,0.5332386494,0.0087339627,0.4412532747,-0.2814124823,-0.1415888518,-0.226051569,-0.0913680568,0.1125744134,-0.3669516146,0.0487546846,-0.2756848037,-0.0543771461,-0.0466812104,-0.121877186,-0.0234336779,-0.1646656841,0.2823671103,0.2395053953,0.2693460882,-0.1665234864,-0.0291301347,0.1338399649,-0.1567696631,-0.065012753,0.3058659434,0.3186143041,0.2530109882,-0.0928714797,0.2424036115,-0.1774176359,0.5707028508,0.1469569206,-0.1631754786,-0.10322088,-0.1555235088,-0.0427126624,0.0198512953,0.0677086264,0.4105406106,-0.1185427532,0.1515361816,-0.2512541115,-0.2376576364,0.6050724983,-0.1785651296,-0.1379883289,-0.1678245664,0.2456205785,0.3388739526,0.0138582885,-0.3756624758,0.1801710874,0.4207406342,0.0298809856,-0.0401440747,0.1496393681,-0.3060053885,-0.1257539541,0.0442536511,0.0996872783,0.149877876,-0.3360651135,0.4139497876,-0.1505003273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2937","title":"load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied","comments":"Hi @daqieq, thanks for reporting.\r\n\r\nUnfortunately, I was not able to reproduce this bug:\r\n```ipython\r\nIn [1]: from datasets import load_dataset\r\n ...: ds = load_dataset('wiki_bio')\r\nDownloading: 7.58kB [00:00, 26.3kB\/s]\r\nDownloading: 2.71kB [00:00, ?B\/s]\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\\r\n1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nDownloading: 334MB [01:17, 4.32MB\/s]\r\nDataset wiki_bio downloaded and prepared to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9. Subsequent calls will reuse thi\r\ns data.\r\n```\r\n\r\nThis kind of error messages usually happen because:\r\n- Your running Python script hasn't write access to that directory\r\n- You have another program (the File Explorer?) already browsing inside that directory","body":"## Describe the bug\r\nStandard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wiki_bio')\r\n```\r\n\r\n## Expected results\r\nIt is expected that the dataset downloads without any errors.\r\n\r\n## Actual results\r\nPermissionError see trace below:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\contextlib.py\", line 120, in __exit__\r\n next(self.gen)\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 598, in incomplete_dir\r\n os.rename(tmp_dir, dirname)\r\nPermissionError: [WinError 5] Access is denied: 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9.incomplete' -> 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9'\r\n```\r\nBy commenting out the os.rename() [L604](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L604) and the shutil.rmtree() [L607](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L607) lines, in my virtual environment, I was able to get the load process to complete, rename the directory manually and then rerun the `load_dataset('wiki_bio')` to get what I needed.\r\n\r\nIt seems that os.rename() in the `incomplete_dir` content manager is the culprit. Here's another project [Conan](https:\/\/github.com\/conan-io\/conan\/issues\/6560) with similar issue with os.rename() if it helps debug this issue.\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Windows-10-10.0.22449-SP0\r\n- Python version: 3.8.12\r\n- PyArrow version: 5.0.0\r\n","comment_length":109,"text":"load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied\n## Describe the bug\r\nStandard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wiki_bio')\r\n```\r\n\r\n## Expected results\r\nIt is expected that the dataset downloads without any errors.\r\n\r\n## Actual results\r\nPermissionError see trace below:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\contextlib.py\", line 120, in __exit__\r\n next(self.gen)\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 598, in incomplete_dir\r\n os.rename(tmp_dir, dirname)\r\nPermissionError: [WinError 5] Access is denied: 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9.incomplete' -> 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9'\r\n```\r\nBy commenting out the os.rename() [L604](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L604) and the shutil.rmtree() [L607](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L607) lines, in my virtual environment, I was able to get the load process to complete, rename the directory manually and then rerun the `load_dataset('wiki_bio')` to get what I needed.\r\n\r\nIt seems that os.rename() in the `incomplete_dir` content manager is the culprit. Here's another project [Conan](https:\/\/github.com\/conan-io\/conan\/issues\/6560) with similar issue with os.rename() if it helps debug this issue.\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Windows-10-10.0.22449-SP0\r\n- Python version: 3.8.12\r\n- PyArrow version: 5.0.0\r\n\nHi @daqieq, thanks for reporting.\r\n\r\nUnfortunately, I was not able to reproduce this bug:\r\n```ipython\r\nIn [1]: from datasets import load_dataset\r\n ...: ds = load_dataset('wiki_bio')\r\nDownloading: 7.58kB [00:00, 26.3kB\/s]\r\nDownloading: 2.71kB [00:00, ?B\/s]\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\\r\n1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nDownloading: 334MB [01:17, 4.32MB\/s]\r\nDataset wiki_bio downloaded and prepared to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9. Subsequent calls will reuse thi\r\ns data.\r\n```\r\n\r\nThis kind of error messages usually happen because:\r\n- Your running Python script hasn't write access to that directory\r\n- You have another program (the File Explorer?) already browsing inside that directory","embeddings":[-0.2289305478,0.3834535182,0.040345341,0.2550398409,-0.0121113108,0.2622572184,0.5036683083,0.1329532564,0.3442817032,0.1357175708,-0.1089743897,0.0772839189,0.073150374,-0.0056987684,-0.1256225109,0.0789283216,0.0364452153,-0.0053882273,0.214949578,0.1039548293,-0.4238833189,0.1072954684,-0.172855407,0.1861151755,0.0260949377,-0.1363423616,0.1212603897,0.3607298434,-0.0307886619,-0.3922062814,0.3530231118,0.0634532124,0.4639182985,0.2582460642,-0.000119246,-0.0137069086,0.2306841165,0.1244994402,-0.0737274736,0.0870998204,-0.1839853674,-0.1409676671,0.1652300805,-0.0976862535,0.1767440438,0.3634920418,-0.1021372527,-0.2009297311,-0.0130413789,0.379034102,0.1573001444,-0.0381855965,-0.1186785623,-0.0752038136,0.4933359921,-0.1592982262,-0.133179754,0.2798380852,0.0790250301,-0.11457026,-0.0989914164,0.1995727867,-0.1621495485,0.1471577436,0.0186581202,0.0311292373,0.2162310332,-0.4581103027,0.3521417677,0.1512579024,0.9656507373,-0.1955735534,-0.1914045215,-0.0887490958,0.0133037586,-0.0330028497,0.4326035678,0.4286232889,-0.3245685697,0.2523709536,0.0659693852,0.0835653022,-0.0825510696,-0.0019090191,0.0712853372,0.1486521512,-0.0504703857,0.1195566207,-0.1147312596,-0.0834005401,0.3863353431,-0.2138270736,-0.254601866,0.191666767,-0.3394370973,0.2898705602,0.0534532256,0.7778053284,0.157877773,0.3155301809,0.1250241548,-0.0540518314,0.2136244029,0.0651903003,0.0085929586,0.3214896023,-0.1122384295,-0.0169230793,0.4271060824,0.1174419895,-0.0816169605,-0.0089906445,-0.0200457647,-0.4294741154,0.3538792133,0.1054070443,0.4081188738,-0.2856419981,-0.1217012107,-0.0002951851,0.1581317037,-0.0239709932,0.0588830411,0.5011533499,-0.2717713416,0.0426839963,0.0682517141,0.1794451177,0.0186107904,0.0977605507,-0.0877998695,-0.0890028104,-0.2667180896,-0.0514384918,0.408646673,-0.3225747645,0.0955644995,0.2566769719,0.0270482041,-0.1792946011,0.1918320805,-0.0414977111,-0.0328572541,0.3576870263,0.397295326,0.3376339078,0.174932614,-0.2122963965,-0.0349953882,0.1863114685,-0.4872902334,-0.4025040567,-0.0336632058,0.1213277355,-0.0833240524,0.0632052422,-0.1980368942,-0.0482573472,0.3823129535,-0.3008547723,-0.0618359111,-0.0137255481,-0.2206203043,-0.5042767525,0.040707428,0.4012890756,-0.3596991003,0.1303672194,-0.0915537477,-0.2640866041,0.1504325122,0.0056375656,0.0056341127,0.3616814017,-0.2812198102,0.0156073133,0.2774584889,-0.3938943148,-1.0015972853,-0.0210030079,0.0730995089,0.1660424769,0.2444063723,0.1191371009,0.3659378588,-0.0303458292,0.1704769284,0.3187128901,0.1300752312,0.1119742319,-0.0629962236,-0.2038549483,0.2796608508,0.1339266002,0.0883391201,0.3009034991,0.312600106,-0.0569339506,0.251095593,-0.1872612089,0.116804406,0.1925157905,0.0114956526,-0.0351362228,0.1375146806,-0.1277379841,-0.420409292,0.3514482081,-0.00811728,0.0851668194,0.0759790465,-0.2027431279,-0.3184642494,0.0913407728,-0.5447552204,-0.2889970243,0.0851063654,0.2938987911,0.2124249339,0.0236818604,0.1891641766,0.3374149799,-0.2082268894,0.1271959245,-0.1811406165,0.0386061855,-0.1883375496,-0.1364492327,-0.2122476399,0.0469153933,0.2001280636,0.2527483404,-0.2134325951,0.1961012632,0.151213333,0.0374986753,-0.3785572946,-0.0595560744,-0.0223776083,0.1217297763,0.1568485051,0.1381781846,0.1483232975,-0.0443943143,-0.1922187656,0.1377540231,-0.1596345007,0.14373146,-0.3266511559,0.2430362701,0.2447061241,0.0900503173,0.0856966749,0.0344529897,0.3054705858,0.0241174567,0.0745961443,-0.2858042717,0.1012615561,-0.0778078139,0.0593021847,0.0225473158,0.0812815726,0.2497533113,-0.1746012717,0.0168862902,0.2869935334,0.349679023,0.5377136469,-0.0409865156,0.0515500903,0.0667970031,0.0479120463,0.0896691009,0.1192515045,0.0857714713,-0.0702273324,0.2634492815,0.1284338385,0.2241659462,-0.0647975802,-0.1138611957,-0.107541278,0.4194297791,-0.4875138104,-0.084387593,0.1000552922,0.0748399198,0.1812013388,-0.2428474277,-0.0887820199,-0.1460988075,-0.2430461347,0.3116364777,0.0940705463,0.0835373625,-0.511957705,0.164441213,0.0075377976,-0.2018063366,-0.1137571484,0.1238797531,-0.1658339351,-0.0141071854,0.1862449348,-0.1799398512,0.1873875707,-0.2951588631,0.2437517196,-0.3901760876,-0.1954325885,0.0151099013,0.0578858256,0.4098414481,0.1254311204,0.1551068723,0.1864105314,-0.3417646587,0.0566025972,-0.1075194404,-0.1510609835,0.2333712131,-0.0550265461,0.0938554928,0.0341966748,-0.2493757308,-0.5408244133,-0.2501789927,-0.0095754005,0.3858899772,0.0983009189,-0.0028510524,-0.1126530915,0.0033437847,0.267029494,0.3937678933,-0.2108280957,-0.2492020875,0.2102492452,-0.2092680335,-0.3346993625,0.0819493756,-0.1042304859,0.0445711389,0.2523835003,-0.4788489342,-0.0906342939,-0.3600659072,0.4096082449,-0.232615903,0.0938282236,0.3463047147,0.0892930999,0.1463989317,-0.0417488404,-0.2757158875,-0.0651353374,-0.2946413159,-0.068232581,0.3381980062,0.4274490774,-0.0575702377,0.3770318925,0.3002279699,0.3604496717,0.3344713151,0.04584153,0.3711133003,0.0542609356,-0.2594737113,0.1087757051,-0.1943459958,-0.140160799,0.3007430136,0.0889823362,-0.1952640116,-0.052531682,-0.3804523051,-0.2807892859,-0.2852858603,-0.0992422476,-0.2463323325,0.2537561357,0.1191406399,0.1100426763,-0.4462445676,-0.3691142201,0.0288443286,0.5546328425,-0.3098391891,0.1115756556,-0.3508035243,0.1366987079,-0.6549720168,0.2522459924,0.0340248011,0.4958057404,-0.1741242707,0.1266969591,-0.0015080259,-0.1074288562,0.5570681095,-0.6522978544,0.4618299305,0.0581227317,0.1157312021,0.0101971151,-0.2243451625,-0.0922563747,0.2416802943,0.1369419247,0.2145550996,0.1639221013,0.0440419056,0.0762091056,0.1665931493,-0.2001322806,-0.279057771,-0.2271348834,-0.3418877125,-0.3903853595,0.128006041,0.0421996191,0.1854700744,-0.1517433375,0.1171616763,-0.4046885669,0.2369778156,0.1301799268,0.1602822691,0.187791869,0.0645751134,0.2181593776,-0.0480453596,0.0753483474,-0.1291563809,0.5240867138,-0.2804769576,-0.328379035,-0.0524905957,-0.199985683,-0.1171171516,0.1122419834,-0.3265575469,-0.0500914566,-0.1384819001,-0.1327372044,-0.3143305779,0.117074959,0.0426035039,-0.1354341507,-0.3776838183,-0.3201939762,0.3641437292,0.2117004246,-0.4467277825,0.3404740095,0.1198474169,-0.4280630946,0.1824693233,0.0517599583,0.7466567159,0.0798895434,0.2780212462,0.0972766355,-0.1122953147,0.2001173049,0.1161781773,0.0565570518,-0.3728869557,-0.4541019201,-0.0988719016,-0.2205425799,-0.134537518,-0.0249406099,-0.2337937057,0.2817453742,-0.1129783094,0.3188050985,0.274310261,0.048037339,-0.3741956651,-0.3764992654,-0.1209942549,0.1251865476,-0.1938239634,0.1358688474,-0.2274900377,0.0911153406,0.0849043131,-0.1673949957,-0.439121902,0.1967898756,-0.4114460945,0.1381170154,-0.2974178493,-0.3836832643,0.0877998993,0.5960645676,0.2849669755,0.3604593873,-0.1543026865,0.0600666888,0.0793892592,0.1179443896,0.024958102,0.0074552288,0.1962639689,-0.0806936249,-0.2981881797,-0.0151690254,-0.2781275809,-0.3215303123,-0.1055536345,0.2862682045,0.4039052129,-0.3201119602,-0.2076887786,-0.3106458187,0.0849739909,-0.2798955739,0.0937029347,0.0470425822,-0.0756008551,0.3795401752,-0.3368881643,-0.425712347,-0.1475177705,0.5242761374,-0.079077132,-0.1337566078,0.5711378455,-0.0485029556,-0.1785390973,-0.0130321439,0.1454919577,-0.1009839401,-0.462458998,-0.1003440917,-0.1225365251,0.1078282446,-0.029106807,0.1698505282,0.2787975669,-0.0963102505,0.0534037836,-0.7581394315,-0.2621001899,-0.1129847094,-0.2311439961,0.1716352552,0.1831110716,-0.4405415952,-0.1353047937,-0.2840806246,-0.203002885,0.0441007577,-0.2147331834,-0.1294415742,0.2779014409,0.154557094,-0.001340596,0.1245135888,-0.0026244703,0.0926584527,0.1227548569,-0.1713698655,-0.1237100959,0.1475240737,0.1037818864,-0.0917996764,0.0267537702,-0.3284253776,-0.0735971555,-0.0984259099,0.3940449953,0.2593082786,-0.0281576384,0.0679169372,-0.1276832074,0.2894711792,-0.2827050388,0.4151190817,-0.3472835422,-0.0251300745,-0.0133727901,0.3347693682,-0.1413164884,0.0541265719,-0.11745359,-0.0824945197,0.1532541662,0.0678972974,0.0646063909,-0.3769121468,0.085026145,0.1787576079,0.0593337528,0.0623646304,-0.23318322,0.0148412948,0.2450501472,0.1589923799,-0.0922699794,-0.1294877678,-0.0120077264,-0.3234773874,-0.1790443361,0.3004153967,0.1639576256,-0.3836256266,0.1359916031,-0.0324052311,0.5581638813,-0.2246947289,-0.0577207766,0.4813490808,-0.2174006104,0.2738089263,0.218585968,0.2596439123,-0.0843797997,0.5514788032,0.1820428967,0.1367382705,-0.0058679264,-0.1720480621,-0.1030526981,-0.6628072858,0.28002882,0.0652534962,-0.0736904144,0.0509437881,-0.0296167657,-0.0086359354,-0.3041151464,-0.0021537442,-0.4382816255,-0.2168946117,-0.1808834672,-0.283659488,0.0684353933,-0.2338435054,0.1725392044,0.0927761719,-0.1707766503,-0.14519687,-0.069977276,-0.0427515879,-0.3035758734,-0.268468231,0.1116327941,0.4133677781,-0.2662928104,-0.1926916391,0.0890493989,0.0064425515,0.2395810932,0.0886423066,0.0961362123,0.7200041413,0.3696607947,-0.186706394,-0.0827123448,0.2043794394,0.08498016,-0.1013351902,0.3095628321,0.1992664337,0.1772322208,0.3453566134,0.1196848005,-0.124466449,0.0374585725,0.0380601026,-0.008947121,-0.0834797546,0.1873595268,-0.1764346659,-0.168423444,0.1405667216,0.1980234534,-0.0727778077,0.1700269282,0.2323683053,0.0627374277,0.1945085675,-0.1900534481,0.0200616512,0.1487799883,0.3155461252,0.3503623307,0.0980699509,-0.2335534841,-0.068358846,-0.5775797367,0.2545953691,-0.0182250105,-0.3800154626,-0.0322349034,-0.1324881017,-0.3368008435,0.1061356142,0.1619229615,-0.1090964749,0.0555578545,0.2668960989,-0.1567609012,-0.2936111093,0.3894377947,-0.0377689339,-0.0781006962,-0.6786634922,0.0366012938,-0.0887601152,0.0320076123,-0.1506992131,-0.1666385084,0.0341500044,-0.0260742661,0.5437771082,0.2644129097,0.4555599689,0.0078377938,0.0918127969,-0.562205255,-0.0687982887,-0.1452899277,-0.0336945578,-0.077826038,0.271348238,-0.2275042981,0.0414364263,-0.1241922826,0.1670939773,0.2662992477,0.3795417845,-0.0054561738,0.2279049158,0.1809432358,0.1571899056,0.3478443921,0.0699722916,-0.0961795077,0.4460362792,-0.2409338355,-0.342048198,0.2109487653,-0.3941510618,-0.3340475857,-0.0192074794,0.3251572549,-0.2549189031,-0.1531336755,-0.2379637361,0.2061088979,0.2478395253,-0.1434804201,-0.3868976235,0.1413453221,-0.1570352912,0.140291512,-0.0227520186,0.0342236757,0.0101351459,-0.225390628,0.1708674133,-0.2374369949]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2937","title":"load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied","comments":"Thanks @albertvillanova for looking at it! I tried on my personal Windows machine and it downloaded just fine.\r\n\r\nRunning on my work machine and on a colleague's machine it is consistently hitting this error. It's not a write access issue because the `.incomplete` directory is written just fine. It just won't rename and then it deletes the directory in the `finally` step. Also the zip file is written and extracted fine in the downloads directory.\r\n\r\nThat leaves another program that might be interfering, and there are plenty of those in my work machine ... (full antivirus, data loss prevention, etc.). So the question remains, why not extend the `try` block to allow catching the error and circle back to the rename after the unknown program is finished doing its 'stuff'. This is the approach that I read about in the linked repo (see my comments above).\r\n\r\nIf it's not high priority, that's fine. However, if someone were to write an PR that solved this issue in our environment in an `except` clause, would it be reviewed for inclusion in a future release? Just wondering whether I should spend any more time on this issue.","body":"## Describe the bug\r\nStandard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wiki_bio')\r\n```\r\n\r\n## Expected results\r\nIt is expected that the dataset downloads without any errors.\r\n\r\n## Actual results\r\nPermissionError see trace below:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\contextlib.py\", line 120, in __exit__\r\n next(self.gen)\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 598, in incomplete_dir\r\n os.rename(tmp_dir, dirname)\r\nPermissionError: [WinError 5] Access is denied: 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9.incomplete' -> 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9'\r\n```\r\nBy commenting out the os.rename() [L604](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L604) and the shutil.rmtree() [L607](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L607) lines, in my virtual environment, I was able to get the load process to complete, rename the directory manually and then rerun the `load_dataset('wiki_bio')` to get what I needed.\r\n\r\nIt seems that os.rename() in the `incomplete_dir` content manager is the culprit. Here's another project [Conan](https:\/\/github.com\/conan-io\/conan\/issues\/6560) with similar issue with os.rename() if it helps debug this issue.\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Windows-10-10.0.22449-SP0\r\n- Python version: 3.8.12\r\n- PyArrow version: 5.0.0\r\n","comment_length":194,"text":"load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied\n## Describe the bug\r\nStandard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wiki_bio')\r\n```\r\n\r\n## Expected results\r\nIt is expected that the dataset downloads without any errors.\r\n\r\n## Actual results\r\nPermissionError see trace below:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\contextlib.py\", line 120, in __exit__\r\n next(self.gen)\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 598, in incomplete_dir\r\n os.rename(tmp_dir, dirname)\r\nPermissionError: [WinError 5] Access is denied: 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9.incomplete' -> 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9'\r\n```\r\nBy commenting out the os.rename() [L604](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L604) and the shutil.rmtree() [L607](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L607) lines, in my virtual environment, I was able to get the load process to complete, rename the directory manually and then rerun the `load_dataset('wiki_bio')` to get what I needed.\r\n\r\nIt seems that os.rename() in the `incomplete_dir` content manager is the culprit. Here's another project [Conan](https:\/\/github.com\/conan-io\/conan\/issues\/6560) with similar issue with os.rename() if it helps debug this issue.\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Windows-10-10.0.22449-SP0\r\n- Python version: 3.8.12\r\n- PyArrow version: 5.0.0\r\n\nThanks @albertvillanova for looking at it! I tried on my personal Windows machine and it downloaded just fine.\r\n\r\nRunning on my work machine and on a colleague's machine it is consistently hitting this error. It's not a write access issue because the `.incomplete` directory is written just fine. It just won't rename and then it deletes the directory in the `finally` step. Also the zip file is written and extracted fine in the downloads directory.\r\n\r\nThat leaves another program that might be interfering, and there are plenty of those in my work machine ... (full antivirus, data loss prevention, etc.). So the question remains, why not extend the `try` block to allow catching the error and circle back to the rename after the unknown program is finished doing its 'stuff'. This is the approach that I read about in the linked repo (see my comments above).\r\n\r\nIf it's not high priority, that's fine. However, if someone were to write an PR that solved this issue in our environment in an `except` clause, would it be reviewed for inclusion in a future release? Just wondering whether I should spend any more time on this issue.","embeddings":[-0.2289305478,0.3834535182,0.040345341,0.2550398409,-0.0121113108,0.2622572184,0.5036683083,0.1329532564,0.3442817032,0.1357175708,-0.1089743897,0.0772839189,0.073150374,-0.0056987684,-0.1256225109,0.0789283216,0.0364452153,-0.0053882273,0.214949578,0.1039548293,-0.4238833189,0.1072954684,-0.172855407,0.1861151755,0.0260949377,-0.1363423616,0.1212603897,0.3607298434,-0.0307886619,-0.3922062814,0.3530231118,0.0634532124,0.4639182985,0.2582460642,-0.000119246,-0.0137069086,0.2306841165,0.1244994402,-0.0737274736,0.0870998204,-0.1839853674,-0.1409676671,0.1652300805,-0.0976862535,0.1767440438,0.3634920418,-0.1021372527,-0.2009297311,-0.0130413789,0.379034102,0.1573001444,-0.0381855965,-0.1186785623,-0.0752038136,0.4933359921,-0.1592982262,-0.133179754,0.2798380852,0.0790250301,-0.11457026,-0.0989914164,0.1995727867,-0.1621495485,0.1471577436,0.0186581202,0.0311292373,0.2162310332,-0.4581103027,0.3521417677,0.1512579024,0.9656507373,-0.1955735534,-0.1914045215,-0.0887490958,0.0133037586,-0.0330028497,0.4326035678,0.4286232889,-0.3245685697,0.2523709536,0.0659693852,0.0835653022,-0.0825510696,-0.0019090191,0.0712853372,0.1486521512,-0.0504703857,0.1195566207,-0.1147312596,-0.0834005401,0.3863353431,-0.2138270736,-0.254601866,0.191666767,-0.3394370973,0.2898705602,0.0534532256,0.7778053284,0.157877773,0.3155301809,0.1250241548,-0.0540518314,0.2136244029,0.0651903003,0.0085929586,0.3214896023,-0.1122384295,-0.0169230793,0.4271060824,0.1174419895,-0.0816169605,-0.0089906445,-0.0200457647,-0.4294741154,0.3538792133,0.1054070443,0.4081188738,-0.2856419981,-0.1217012107,-0.0002951851,0.1581317037,-0.0239709932,0.0588830411,0.5011533499,-0.2717713416,0.0426839963,0.0682517141,0.1794451177,0.0186107904,0.0977605507,-0.0877998695,-0.0890028104,-0.2667180896,-0.0514384918,0.408646673,-0.3225747645,0.0955644995,0.2566769719,0.0270482041,-0.1792946011,0.1918320805,-0.0414977111,-0.0328572541,0.3576870263,0.397295326,0.3376339078,0.174932614,-0.2122963965,-0.0349953882,0.1863114685,-0.4872902334,-0.4025040567,-0.0336632058,0.1213277355,-0.0833240524,0.0632052422,-0.1980368942,-0.0482573472,0.3823129535,-0.3008547723,-0.0618359111,-0.0137255481,-0.2206203043,-0.5042767525,0.040707428,0.4012890756,-0.3596991003,0.1303672194,-0.0915537477,-0.2640866041,0.1504325122,0.0056375656,0.0056341127,0.3616814017,-0.2812198102,0.0156073133,0.2774584889,-0.3938943148,-1.0015972853,-0.0210030079,0.0730995089,0.1660424769,0.2444063723,0.1191371009,0.3659378588,-0.0303458292,0.1704769284,0.3187128901,0.1300752312,0.1119742319,-0.0629962236,-0.2038549483,0.2796608508,0.1339266002,0.0883391201,0.3009034991,0.312600106,-0.0569339506,0.251095593,-0.1872612089,0.116804406,0.1925157905,0.0114956526,-0.0351362228,0.1375146806,-0.1277379841,-0.420409292,0.3514482081,-0.00811728,0.0851668194,0.0759790465,-0.2027431279,-0.3184642494,0.0913407728,-0.5447552204,-0.2889970243,0.0851063654,0.2938987911,0.2124249339,0.0236818604,0.1891641766,0.3374149799,-0.2082268894,0.1271959245,-0.1811406165,0.0386061855,-0.1883375496,-0.1364492327,-0.2122476399,0.0469153933,0.2001280636,0.2527483404,-0.2134325951,0.1961012632,0.151213333,0.0374986753,-0.3785572946,-0.0595560744,-0.0223776083,0.1217297763,0.1568485051,0.1381781846,0.1483232975,-0.0443943143,-0.1922187656,0.1377540231,-0.1596345007,0.14373146,-0.3266511559,0.2430362701,0.2447061241,0.0900503173,0.0856966749,0.0344529897,0.3054705858,0.0241174567,0.0745961443,-0.2858042717,0.1012615561,-0.0778078139,0.0593021847,0.0225473158,0.0812815726,0.2497533113,-0.1746012717,0.0168862902,0.2869935334,0.349679023,0.5377136469,-0.0409865156,0.0515500903,0.0667970031,0.0479120463,0.0896691009,0.1192515045,0.0857714713,-0.0702273324,0.2634492815,0.1284338385,0.2241659462,-0.0647975802,-0.1138611957,-0.107541278,0.4194297791,-0.4875138104,-0.084387593,0.1000552922,0.0748399198,0.1812013388,-0.2428474277,-0.0887820199,-0.1460988075,-0.2430461347,0.3116364777,0.0940705463,0.0835373625,-0.511957705,0.164441213,0.0075377976,-0.2018063366,-0.1137571484,0.1238797531,-0.1658339351,-0.0141071854,0.1862449348,-0.1799398512,0.1873875707,-0.2951588631,0.2437517196,-0.3901760876,-0.1954325885,0.0151099013,0.0578858256,0.4098414481,0.1254311204,0.1551068723,0.1864105314,-0.3417646587,0.0566025972,-0.1075194404,-0.1510609835,0.2333712131,-0.0550265461,0.0938554928,0.0341966748,-0.2493757308,-0.5408244133,-0.2501789927,-0.0095754005,0.3858899772,0.0983009189,-0.0028510524,-0.1126530915,0.0033437847,0.267029494,0.3937678933,-0.2108280957,-0.2492020875,0.2102492452,-0.2092680335,-0.3346993625,0.0819493756,-0.1042304859,0.0445711389,0.2523835003,-0.4788489342,-0.0906342939,-0.3600659072,0.4096082449,-0.232615903,0.0938282236,0.3463047147,0.0892930999,0.1463989317,-0.0417488404,-0.2757158875,-0.0651353374,-0.2946413159,-0.068232581,0.3381980062,0.4274490774,-0.0575702377,0.3770318925,0.3002279699,0.3604496717,0.3344713151,0.04584153,0.3711133003,0.0542609356,-0.2594737113,0.1087757051,-0.1943459958,-0.140160799,0.3007430136,0.0889823362,-0.1952640116,-0.052531682,-0.3804523051,-0.2807892859,-0.2852858603,-0.0992422476,-0.2463323325,0.2537561357,0.1191406399,0.1100426763,-0.4462445676,-0.3691142201,0.0288443286,0.5546328425,-0.3098391891,0.1115756556,-0.3508035243,0.1366987079,-0.6549720168,0.2522459924,0.0340248011,0.4958057404,-0.1741242707,0.1266969591,-0.0015080259,-0.1074288562,0.5570681095,-0.6522978544,0.4618299305,0.0581227317,0.1157312021,0.0101971151,-0.2243451625,-0.0922563747,0.2416802943,0.1369419247,0.2145550996,0.1639221013,0.0440419056,0.0762091056,0.1665931493,-0.2001322806,-0.279057771,-0.2271348834,-0.3418877125,-0.3903853595,0.128006041,0.0421996191,0.1854700744,-0.1517433375,0.1171616763,-0.4046885669,0.2369778156,0.1301799268,0.1602822691,0.187791869,0.0645751134,0.2181593776,-0.0480453596,0.0753483474,-0.1291563809,0.5240867138,-0.2804769576,-0.328379035,-0.0524905957,-0.199985683,-0.1171171516,0.1122419834,-0.3265575469,-0.0500914566,-0.1384819001,-0.1327372044,-0.3143305779,0.117074959,0.0426035039,-0.1354341507,-0.3776838183,-0.3201939762,0.3641437292,0.2117004246,-0.4467277825,0.3404740095,0.1198474169,-0.4280630946,0.1824693233,0.0517599583,0.7466567159,0.0798895434,0.2780212462,0.0972766355,-0.1122953147,0.2001173049,0.1161781773,0.0565570518,-0.3728869557,-0.4541019201,-0.0988719016,-0.2205425799,-0.134537518,-0.0249406099,-0.2337937057,0.2817453742,-0.1129783094,0.3188050985,0.274310261,0.048037339,-0.3741956651,-0.3764992654,-0.1209942549,0.1251865476,-0.1938239634,0.1358688474,-0.2274900377,0.0911153406,0.0849043131,-0.1673949957,-0.439121902,0.1967898756,-0.4114460945,0.1381170154,-0.2974178493,-0.3836832643,0.0877998993,0.5960645676,0.2849669755,0.3604593873,-0.1543026865,0.0600666888,0.0793892592,0.1179443896,0.024958102,0.0074552288,0.1962639689,-0.0806936249,-0.2981881797,-0.0151690254,-0.2781275809,-0.3215303123,-0.1055536345,0.2862682045,0.4039052129,-0.3201119602,-0.2076887786,-0.3106458187,0.0849739909,-0.2798955739,0.0937029347,0.0470425822,-0.0756008551,0.3795401752,-0.3368881643,-0.425712347,-0.1475177705,0.5242761374,-0.079077132,-0.1337566078,0.5711378455,-0.0485029556,-0.1785390973,-0.0130321439,0.1454919577,-0.1009839401,-0.462458998,-0.1003440917,-0.1225365251,0.1078282446,-0.029106807,0.1698505282,0.2787975669,-0.0963102505,0.0534037836,-0.7581394315,-0.2621001899,-0.1129847094,-0.2311439961,0.1716352552,0.1831110716,-0.4405415952,-0.1353047937,-0.2840806246,-0.203002885,0.0441007577,-0.2147331834,-0.1294415742,0.2779014409,0.154557094,-0.001340596,0.1245135888,-0.0026244703,0.0926584527,0.1227548569,-0.1713698655,-0.1237100959,0.1475240737,0.1037818864,-0.0917996764,0.0267537702,-0.3284253776,-0.0735971555,-0.0984259099,0.3940449953,0.2593082786,-0.0281576384,0.0679169372,-0.1276832074,0.2894711792,-0.2827050388,0.4151190817,-0.3472835422,-0.0251300745,-0.0133727901,0.3347693682,-0.1413164884,0.0541265719,-0.11745359,-0.0824945197,0.1532541662,0.0678972974,0.0646063909,-0.3769121468,0.085026145,0.1787576079,0.0593337528,0.0623646304,-0.23318322,0.0148412948,0.2450501472,0.1589923799,-0.0922699794,-0.1294877678,-0.0120077264,-0.3234773874,-0.1790443361,0.3004153967,0.1639576256,-0.3836256266,0.1359916031,-0.0324052311,0.5581638813,-0.2246947289,-0.0577207766,0.4813490808,-0.2174006104,0.2738089263,0.218585968,0.2596439123,-0.0843797997,0.5514788032,0.1820428967,0.1367382705,-0.0058679264,-0.1720480621,-0.1030526981,-0.6628072858,0.28002882,0.0652534962,-0.0736904144,0.0509437881,-0.0296167657,-0.0086359354,-0.3041151464,-0.0021537442,-0.4382816255,-0.2168946117,-0.1808834672,-0.283659488,0.0684353933,-0.2338435054,0.1725392044,0.0927761719,-0.1707766503,-0.14519687,-0.069977276,-0.0427515879,-0.3035758734,-0.268468231,0.1116327941,0.4133677781,-0.2662928104,-0.1926916391,0.0890493989,0.0064425515,0.2395810932,0.0886423066,0.0961362123,0.7200041413,0.3696607947,-0.186706394,-0.0827123448,0.2043794394,0.08498016,-0.1013351902,0.3095628321,0.1992664337,0.1772322208,0.3453566134,0.1196848005,-0.124466449,0.0374585725,0.0380601026,-0.008947121,-0.0834797546,0.1873595268,-0.1764346659,-0.168423444,0.1405667216,0.1980234534,-0.0727778077,0.1700269282,0.2323683053,0.0627374277,0.1945085675,-0.1900534481,0.0200616512,0.1487799883,0.3155461252,0.3503623307,0.0980699509,-0.2335534841,-0.068358846,-0.5775797367,0.2545953691,-0.0182250105,-0.3800154626,-0.0322349034,-0.1324881017,-0.3368008435,0.1061356142,0.1619229615,-0.1090964749,0.0555578545,0.2668960989,-0.1567609012,-0.2936111093,0.3894377947,-0.0377689339,-0.0781006962,-0.6786634922,0.0366012938,-0.0887601152,0.0320076123,-0.1506992131,-0.1666385084,0.0341500044,-0.0260742661,0.5437771082,0.2644129097,0.4555599689,0.0078377938,0.0918127969,-0.562205255,-0.0687982887,-0.1452899277,-0.0336945578,-0.077826038,0.271348238,-0.2275042981,0.0414364263,-0.1241922826,0.1670939773,0.2662992477,0.3795417845,-0.0054561738,0.2279049158,0.1809432358,0.1571899056,0.3478443921,0.0699722916,-0.0961795077,0.4460362792,-0.2409338355,-0.342048198,0.2109487653,-0.3941510618,-0.3340475857,-0.0192074794,0.3251572549,-0.2549189031,-0.1531336755,-0.2379637361,0.2061088979,0.2478395253,-0.1434804201,-0.3868976235,0.1413453221,-0.1570352912,0.140291512,-0.0227520186,0.0342236757,0.0101351459,-0.225390628,0.1708674133,-0.2374369949]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2934","title":"to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows","comments":"I did some investigation and, as it seems, the bug stems from [this line](https:\/\/github.com\/huggingface\/datasets\/blob\/8004d7c3e1d74b29c3e5b0d1660331cd26758363\/src\/datasets\/arrow_dataset.py#L325). The lifecycle of the dataset from the linked line is bound to one of the returned `tf.data.Dataset`. So my (hacky) solution involves wrapping the linked dataset with `weakref.proxy` and adding a custom `__del__` to `tf.python.data.ops.dataset_ops.TensorSliceDataset` (this is the type of a dataset that is returned by `tf.data.Dataset.from_tensor_slices`; this works for TF 2.x, but I'm not sure `tf.python.data.ops.dataset_ops` is a valid path for TF 1.x) that deletes the linked dataset, which is assigned to the dataset object as a property. Will open a draft PR soon!","body":"To reproduce:\r\n```python\r\nimport datasets as ds\r\nimport weakref\r\nimport gc\r\n\r\nd = ds.load_dataset(\"mnist\", split=\"train\")\r\nref = weakref.ref(d._data.table)\r\ntfd = d.to_tf_dataset(\"image\", batch_size=1, shuffle=False, label_cols=\"label\")\r\ndel tfd, d\r\ngc.collect()\r\nassert ref() is None, \"Error: there is at least one reference left\"\r\n```\r\n\r\nThis causes issues because the table holds a reference to an open arrow file that should be closed. So on windows it's not possible to delete or move the arrow file afterwards.\r\n\r\nMoreover the CI test of the `to_tf_dataset` method isn't able to clean up the temporary arrow files because of this.\r\n\r\ncc @Rocketknight1 ","comment_length":99,"text":"to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows\nTo reproduce:\r\n```python\r\nimport datasets as ds\r\nimport weakref\r\nimport gc\r\n\r\nd = ds.load_dataset(\"mnist\", split=\"train\")\r\nref = weakref.ref(d._data.table)\r\ntfd = d.to_tf_dataset(\"image\", batch_size=1, shuffle=False, label_cols=\"label\")\r\ndel tfd, d\r\ngc.collect()\r\nassert ref() is None, \"Error: there is at least one reference left\"\r\n```\r\n\r\nThis causes issues because the table holds a reference to an open arrow file that should be closed. So on windows it's not possible to delete or move the arrow file afterwards.\r\n\r\nMoreover the CI test of the `to_tf_dataset` method isn't able to clean up the temporary arrow files because of this.\r\n\r\ncc @Rocketknight1 \nI did some investigation and, as it seems, the bug stems from [this line](https:\/\/github.com\/huggingface\/datasets\/blob\/8004d7c3e1d74b29c3e5b0d1660331cd26758363\/src\/datasets\/arrow_dataset.py#L325). The lifecycle of the dataset from the linked line is bound to one of the returned `tf.data.Dataset`. So my (hacky) solution involves wrapping the linked dataset with `weakref.proxy` and adding a custom `__del__` to `tf.python.data.ops.dataset_ops.TensorSliceDataset` (this is the type of a dataset that is returned by `tf.data.Dataset.from_tensor_slices`; this works for TF 2.x, but I'm not sure `tf.python.data.ops.dataset_ops` is a valid path for TF 1.x) that deletes the linked dataset, which is assigned to the dataset object as a property. Will open a draft PR soon!","embeddings":[0.0456905663,0.334508419,0.1148890257,0.080451034,0.2304305583,0.118887417,0.3988847136,0.2548755109,-0.1099170819,0.2674386799,-0.3353850842,0.4040720165,-0.1644811481,-0.1074629501,-0.0342661999,-0.1066431478,0.0511111543,0.1451645941,-0.1666638404,-0.1226122081,-0.2103365809,0.0351907536,-0.1274755448,0.0721159056,-0.1387630254,-0.2811448276,0.0105889933,0.1006439403,0.2441080213,-0.0183550231,0.0682790056,0.0211760364,-0.177796185,0.4797623456,-0.0001166652,0.4142073989,-0.0061280271,0.1301537901,-0.235580802,0.0565112084,-0.0690313205,0.1760492623,0.1245584637,-0.0599729307,0.1349684894,-0.2177767009,-0.0255724229,0.0832738504,0.2824953496,0.4290578365,0.1738771498,0.6678111553,0.049049627,-0.0070738695,0.3827846944,0.0638971701,-0.2742590308,0.1161701158,-0.2370098084,-0.1935149431,0.0655605793,0.5781883597,0.0173822027,-0.0747985765,-0.1346866488,0.0793781951,-0.2474848479,-0.2132097334,0.099155657,0.3864619732,0.2942140996,-0.3177287579,-0.0704426542,-0.1286842972,-0.0647842139,-0.1454346925,0.1913480461,-0.046207156,0.017386917,0.1348499954,0.0582152456,-0.1622668952,-0.2615066469,-0.0061726281,-0.2810887992,-0.1072200239,0.0857438222,0.0801384374,0.414106369,-0.0271513741,-0.068637602,0.0416288003,0.0201638471,0.1307915747,-0.2023018003,-0.1820205152,-0.039258711,-0.1770301908,-0.0281791855,-0.0708624199,0.2392520756,-0.1570445448,-0.2443445474,0.0677398071,0.1644194871,-0.0705889463,-0.6655332446,0.1451997906,0.2680508792,-0.1933338344,-0.1119194329,-0.0286712628,0.0027096812,-0.4981434345,0.2861165106,-0.2107922286,0.4638061523,0.0174133722,-0.5834477544,0.1703355759,-0.1943601817,0.4077934325,-0.0392662771,0.3421227634,-0.0097245919,0.2793687582,0.3236465752,-0.0036117642,-0.4970700741,-0.0162012,-0.0546676517,-0.0396631137,-0.2995604873,-0.3319226205,0.1675927192,0.0439531542,-0.0478496067,0.1728075743,0.0610827021,0.1025774032,0.2539456487,-0.2808155119,0.5246316791,0.4812721908,-0.0750983432,0.2450949699,0.094958134,-0.2507022023,-0.1292222887,0.3534173071,-0.271070987,-0.1585650742,-0.2603646219,0.1245645955,-0.1202328429,-0.1552468389,-0.2238021493,0.0450044461,0.0343124531,-0.0595466755,0.0862919539,-0.1370920688,-0.5523484945,-0.2646558583,0.0109041464,0.3650912344,-0.1991525888,0.1795514375,-0.0254384503,-0.3931910396,-0.1594114155,0.2656008899,0.0646049082,0.315382421,-0.4097916186,-0.1304864734,0.4838194549,-0.2320678681,-0.4512842894,0.2350057364,-0.2089946121,0.2142525762,-0.3518173099,0.1303172708,0.2125170231,-0.2066658884,0.0123629132,-0.0070450343,-0.0623103678,-0.009127249,-0.3236092329,-0.0108151864,-0.0680190474,-0.1952706128,-0.0157284178,-0.0124156829,0.1318649352,-0.0493954197,0.3902634382,0.2196914554,0.1933261156,0.1417482793,0.3756607473,-0.1729133576,-0.0790883675,0.0432887785,-0.3775175512,0.1395243853,0.4056796134,0.0777031407,-0.1249045953,-0.1479362994,-0.0149693638,0.1985371262,-0.1919496357,0.0200802218,0.0784440786,-0.1194245666,-0.3427836597,0.2097568363,-0.0087737013,-0.1780350357,-0.2768910527,0.0820086896,0.1028026193,0.4687242806,0.2687642872,-0.0055732708,-0.2448134571,0.1943388134,-0.1494650543,-0.0300622191,-0.2719019651,0.4269395471,0.1747722477,0.3178130984,-0.1541973948,0.5814441442,0.1066848487,-0.5981359482,-0.2948667407,0.345680505,0.1319477111,-0.028623648,-0.0179552659,0.1456983984,-0.2007223219,0.171325475,0.0537690409,0.2392641008,0.0408273004,0.0214642044,-0.0363970362,-0.3410162628,-0.048381418,0.3614164293,-0.3146298528,0.2986148596,-0.6084029675,0.1965304911,0.3866648376,0.0311069563,-0.1689249873,0.2358304709,-0.1694948971,0.0617295206,0.0715380758,0.3181353509,0.476054281,0.0223621521,0.1727905273,0.1367983967,-0.193739742,-0.0902521908,0.2373642176,0.0887134001,0.3124321103,0.2039208263,0.25704512,0.193103686,-0.3024033606,0.0330183432,0.130203411,0.1728958189,-0.4525787234,0.1747216135,-0.1807385981,-0.0388464183,-0.307489574,0.0117461048,0.0377215482,-0.134800002,0.3154597282,0.3967652917,-0.4561233222,0.153922379,-0.3289419115,0.1290333718,0.0227688793,-0.2959315479,0.1803024411,-0.169179216,-0.2714836895,0.0344429053,0.3643865883,-0.24427782,0.4180335999,-0.0013580687,-0.0747494251,-0.1528537571,-0.2631629407,0.0897795483,-0.0949533954,0.1932279766,0.2937165499,0.3368107378,-0.1232672036,-0.1301765293,0.1462340504,-0.4851278663,-0.1786133051,0.1372669786,-0.1227175817,0.209010601,0.0203978065,-0.1520884037,0.0741765723,-0.3432299793,0.0670822263,-0.191379711,0.127728045,0.6203324199,0.1520412415,-0.0800016448,0.3789710402,0.10767968,-0.1736004055,-0.1127346307,0.200914368,-0.2037230134,-0.4296337366,0.2909044921,0.0338889062,0.1325789392,0.295558393,-0.6434158087,-0.1689423323,-0.0437356085,0.1776636243,-0.1088872477,-0.0470069051,0.1665738225,0.1164368093,-0.1197147071,-0.3432469964,0.0827863663,0.3821257651,-0.227080822,-0.0485802852,0.0422705784,0.0882449597,0.313570559,0.6243577003,0.0508327,-0.3429434299,0.0208036136,-0.2982521057,0.5073607564,-0.159090668,-0.2795748413,0.0158007275,0.0420313291,-0.13273637,0.2462733835,-0.0795774087,-0.2103789449,0.0578622855,0.2167329341,-0.0375917666,-0.5290263295,0.1369023621,-0.3965076804,0.1983652562,0.1509983689,-0.0425542817,-0.1581777185,-0.4451107979,0.2289596349,0.2040211558,0.0930258259,-0.1366162896,-0.1987896413,-0.2823874056,-0.3425536454,0.0680471882,0.1306476295,0.5035336614,0.0921356902,-0.2987198234,0.1360505223,0.2798964679,0.7769178748,-0.0301158391,-0.0992753878,0.2773328424,0.3224091232,-0.4041952789,0.0555643141,-0.1050164551,-0.0046288599,0.028162742,0.2416950762,0.0772358328,-0.3887937963,0.2129498273,0.1766911447,-0.2477184087,0.2110395133,-0.3918560147,-0.398602128,0.1013680622,0.0878460184,0.260494113,0.0497563928,-0.1087201983,-0.1115121841,-0.3012365699,-0.2482331544,0.0881446972,-0.0156271141,0.3283712864,0.3708717823,0.3975336552,0.0533410273,0.5702848434,0.3938702047,0.4075015187,-0.5431131124,0.021427609,0.0755058676,0.1911333948,0.125191927,0.4420394599,-0.1866856217,0.0531414859,-0.4151880443,0.0002785593,-0.3923731148,0.5302588344,0.1395115256,-0.2031341493,-0.0013096452,-0.0922305658,0.1333688647,-0.0103618503,-0.1489023864,0.6897454262,0.2165830582,-0.1877485067,0.2651528418,0.1582136005,1.1080405712,-0.0935677141,0.2334691584,0.5235840082,-0.2643699646,0.2126546949,-0.0553808175,0.1353131831,-0.0632255748,-0.5360183716,0.0201393031,-0.0125073828,-0.3608244359,-0.2005607784,0.1893274188,0.0023706069,0.1572063267,0.410684675,0.2291459441,0.0776234716,0.2959440947,-0.2186191976,0.0396565311,0.0224721488,0.4466592371,-0.1266208887,-0.2026067823,-0.1223222986,0.1659467816,0.0271231998,-0.0344647914,-0.0096408138,-0.7225638628,0.2756160796,0.0422483645,-0.0904388428,-0.0240593981,0.4012312889,0.3712239861,-0.0988768116,-0.1523476243,0.2883424163,0.3773627877,0.2930121124,0.0938072205,-0.0167857539,0.4750951231,0.0383024365,-0.0803633481,0.1676332504,-0.0544717051,-0.0525227487,-0.0370946191,-0.1225825995,0.0382938497,-0.1205560043,0.044584427,-0.0995850563,-0.2459054291,-0.0579577163,0.1035773382,-0.180487588,-0.1070264578,0.2344724834,-0.1927336901,-0.4316695929,0.2392567247,0.3922551274,-0.1205129474,0.1710344702,0.3036819696,-0.0761046782,-0.0731279776,-0.2562387288,0.3445347846,0.0378310904,-0.5670926571,0.2645981014,-0.1810758114,0.3202282488,-0.1079885513,0.5147185326,-0.1970301867,0.0209408961,0.0221601725,-0.1397224218,-0.3977710903,0.2361740619,0.0018207144,0.4124109447,-0.2537195086,0.4747990668,-0.1287534237,-0.2527932227,-0.2997979224,0.2343676686,-0.0281137675,0.0787903517,-0.1058863029,0.0576506145,0.3181125224,-0.2779511809,0.057696119,0.1073206663,-0.1381114721,-0.1384907216,-0.1637144834,0.1632773131,0.0280179083,-0.3234009743,-0.0129801221,-0.1738720238,0.076291658,-0.1113803163,0.3281935751,0.0700529516,-0.3296965957,0.0270981658,-0.213646844,0.1309412271,0.1392472386,0.0077438001,-0.0305922497,0.3758689463,0.0896954685,0.0981485397,-0.2366094887,-0.2325259447,-0.0303854793,0.2842668295,-0.0356378406,-0.3600642979,0.0377127007,-0.0558477119,-0.2938879132,-0.0599053092,0.3766296208,0.1799759567,-0.4634622633,-0.1362240314,0.2592977285,0.2242168635,-0.240421176,0.16245085,-0.121767506,0.0198641177,0.0385113843,0.166296646,0.1840274334,-0.1465545297,-0.1428779364,0.177231282,-0.2008037418,-0.2036248147,0.0310160294,0.5156292915,-0.0325355567,0.2373232841,-0.1476540118,0.1036643386,0.1264172643,0.2481572181,-0.0781786367,0.4826430678,0.1981102079,-0.0654385388,-0.0354681835,-0.3134524226,-0.0555218123,0.1641895622,-0.123816818,0.2421867996,-0.0361499302,0.2217094153,0.1549313664,-0.4245827198,-0.0117485141,0.2324923277,-0.0859795734,-0.3671880066,-0.3900308013,-0.208055228,-0.2723349631,0.0311247166,-0.0055279625,-0.0865737349,0.0094038406,0.1060144007,-0.0662822947,-0.2825777233,-0.1198218912,0.3330772221,0.2778065503,-0.2325664908,0.0656126365,-0.0248095151,-0.065677695,-0.0021082226,0.2082015872,0.0835288912,0.0284219794,-0.1759397835,0.0114991637,-0.0346538313,0.0922640339,-0.040092878,0.2144379616,-0.0489174612,-0.1405197233,0.2157975733,-0.0491833091,-0.070024997,0.1828447282,0.0519715101,0.0976671353,-0.4182255566,0.1498892158,-0.0033793463,-0.101680629,-0.2287625968,-0.0051180227,0.050867077,-0.108679451,0.4808011055,-0.4602059424,0.3644775152,-0.0873917863,0.0682187453,-0.0095315883,0.2219795287,0.1271091551,0.3292393088,-0.2891252935,-0.3842819929,-0.5309660435,0.2625992,-0.0495068468,-0.0506452397,0.2118176818,0.0317716971,-0.0888985768,-0.0664601475,0.0274049416,-0.1865465939,0.113322854,0.267254889,-0.2658275366,-0.0360959657,-0.4830030203,0.0984326154,0.0199032538,-0.5478616953,-0.1800811589,-0.1762706935,0.0375101008,-0.0818457901,-0.0215519853,-0.0166112091,0.5460247397,0.5247312188,0.1254571527,0.0754962265,0.1195667684,-0.102561526,-0.2674234807,-0.179499656,-0.3766336739,-0.1837470978,0.0070535499,-0.0237968545,-0.0529440977,0.197137922,0.0137047442,-0.2977714539,0.1801951528,0.3653975129,-0.4400824606,0.1956054866,-0.2059031129,0.3725327849,0.2628458738,0.4234926105,-0.1296609044,0.2166781127,-0.3998849392,-0.3820687532,0.4928815663,-0.5328620076,-0.093006812,-0.024597371,0.3842858076,-0.0938029662,0.0362871028,-0.5819953084,-0.176555559,0.3898159564,-0.083492741,-0.5030674338,0.1706117392,0.130146578,0.1384345293,0.1181671172,0.5913186073,0.1401758492,-0.1933630854,0.0364979543,-0.2862138152]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2934","title":"to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows","comments":"Thanks a lot for investigating !","body":"To reproduce:\r\n```python\r\nimport datasets as ds\r\nimport weakref\r\nimport gc\r\n\r\nd = ds.load_dataset(\"mnist\", split=\"train\")\r\nref = weakref.ref(d._data.table)\r\ntfd = d.to_tf_dataset(\"image\", batch_size=1, shuffle=False, label_cols=\"label\")\r\ndel tfd, d\r\ngc.collect()\r\nassert ref() is None, \"Error: there is at least one reference left\"\r\n```\r\n\r\nThis causes issues because the table holds a reference to an open arrow file that should be closed. So on windows it's not possible to delete or move the arrow file afterwards.\r\n\r\nMoreover the CI test of the `to_tf_dataset` method isn't able to clean up the temporary arrow files because of this.\r\n\r\ncc @Rocketknight1 ","comment_length":6,"text":"to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows\nTo reproduce:\r\n```python\r\nimport datasets as ds\r\nimport weakref\r\nimport gc\r\n\r\nd = ds.load_dataset(\"mnist\", split=\"train\")\r\nref = weakref.ref(d._data.table)\r\ntfd = d.to_tf_dataset(\"image\", batch_size=1, shuffle=False, label_cols=\"label\")\r\ndel tfd, d\r\ngc.collect()\r\nassert ref() is None, \"Error: there is at least one reference left\"\r\n```\r\n\r\nThis causes issues because the table holds a reference to an open arrow file that should be closed. So on windows it's not possible to delete or move the arrow file afterwards.\r\n\r\nMoreover the CI test of the `to_tf_dataset` method isn't able to clean up the temporary arrow files because of this.\r\n\r\ncc @Rocketknight1 \nThanks a lot for investigating !","embeddings":[-0.017654635,0.1541891694,0.0769107938,0.0744181946,0.1725985259,0.1645845324,0.3400904834,0.1861168444,-0.1919683069,0.3822576702,-0.2824491262,0.2848426998,-0.1505767554,-0.1510132104,-0.1580371708,-0.0992337465,0.1257585436,0.167851299,-0.1901147068,-0.1841688305,-0.2898720801,-0.0215148497,-0.327370286,0.0795662105,-0.2852760255,-0.3267048597,-0.0916857198,0.1191536784,0.1170001402,-0.07841409,-0.1573167741,-0.0711032003,-0.1773468852,0.4747584462,-0.0001205474,0.3584452868,0.0146335,0.1292270571,-0.1646926254,0.1001513228,-0.102570191,0.1868531406,0.1098253652,-0.1309705675,0.1541632265,-0.2203837484,0.047294464,0.1913519353,0.2669393718,0.3942332268,0.1664138734,0.5976775289,-0.0148456199,-0.0244104229,0.4810651839,0.0666921884,-0.3110128641,0.228202045,-0.183252871,-0.2691649497,0.1474390477,0.4888600409,0.0831765756,-0.0415764973,-0.1837423146,0.0297772828,-0.3578903675,-0.2360781282,0.1937207282,0.2784611583,0.3239397705,-0.2224481106,0.0123386877,0.0420895405,-0.12151099,-0.0689414591,0.1815636605,0.0662881956,0.1170332655,0.1451949328,-0.1794295907,-0.0714711398,-0.2178188562,-0.0269813742,-0.2697104514,-0.1216941476,0.0063582025,0.1643116474,0.3995058239,-0.0270410925,-0.2186173201,-0.0294112433,-0.0166256763,0.0714359283,-0.1651377082,-0.1348735392,-0.1345437765,0.0025094899,-0.1078164205,0.0273076948,0.1821828932,-0.2230697572,-0.1323369145,0.1000298783,0.1016594917,-0.1187836677,-0.5958840847,0.3131462336,0.295281142,-0.2285301536,-0.2712132633,-0.1387895048,-0.1088327542,-0.6633168459,0.4933824241,-0.2781928778,0.498590976,-0.1741571277,-0.5532178283,0.1169041544,-0.1342375726,0.3653871417,-0.115196079,0.4193829894,0.1590043753,0.0912155584,0.2282477468,0.1190212443,-0.4121915698,0.0310258958,0.0330229588,-0.1256711781,-0.360565722,-0.3782528937,0.1396561265,0.0635450929,-0.0230039731,0.2391600609,0.0726201385,0.1766205281,0.2837675512,-0.431430608,0.6493878365,0.4464761317,-0.0134130185,0.2591511011,0.1623892933,-0.1892025024,-0.1497623026,0.4186564386,-0.2162784487,-0.0137974694,-0.157305494,0.115259409,0.0606884286,-0.1436565369,-0.203731969,-0.1010478437,-0.0786174759,-0.1635623276,0.2012656033,-0.1976070255,-0.5147459507,-0.2963284552,-0.1082999781,0.3792219162,-0.1645184755,0.3078974187,0.1431201845,-0.3972130418,0.0059953704,0.363016516,0.1134514436,0.2382273823,-0.4455956221,-0.1024718434,0.4214687347,-0.2397156507,-0.4494386911,0.2381397039,-0.1294758916,0.0941175446,-0.2015377432,0.1184352636,0.143118009,-0.2037379295,0.0316181332,0.0183395352,-0.0693732202,-0.076740399,-0.3000811338,0.1583014727,0.0204451066,-0.0509014651,-0.154321,0.05273645,0.3393533826,-0.2361169606,0.3411521316,0.1847975552,0.0689592734,0.1607869565,0.321493119,-0.3075342774,-0.0852688402,0.0619392544,-0.1281856,0.1172516495,0.364707619,0.0094336923,-0.0719543397,-0.113814719,-0.113396585,0.1810445189,-0.2156182081,0.037566293,0.0035135944,-0.0378339328,-0.2137085199,0.2067743689,-0.0004930327,-0.1253623217,-0.3070148826,0.026798455,0.296400249,0.4681534469,0.1054550856,-0.0451783091,-0.1720358729,0.1316904277,-0.1075824574,-0.066053614,-0.2887434065,0.469099164,0.2476580292,0.1296931356,0.0165344253,0.4050133228,0.0425857641,-0.691002965,-0.1249036565,0.4314128757,0.1971428692,-0.1063851342,0.0224665478,0.0856082737,-0.2597537935,0.2148093581,0.100669235,0.2228789479,0.0058220923,0.0665586069,0.0792132393,-0.3537418544,0.096439071,0.4207526743,-0.3075872362,0.2093724161,-0.6096150279,0.0245354343,0.4899153709,-0.0103561152,-0.0894310772,0.2794166505,-0.1514907628,0.1749186516,0.0958402604,0.1407725215,0.5122002363,0.035527043,0.0858851746,0.1443796307,-0.2817312479,-0.066528663,0.2520436347,0.0302203242,0.2120308727,0.1943588257,0.1666267216,0.2204724401,-0.3269425929,-0.0779463053,0.100687407,0.0708008334,-0.3873489797,0.2037564516,-0.141062662,-0.1401050091,-0.2412522137,0.0849737749,-0.0051670396,-0.1253742427,0.3121213317,0.2641050816,-0.3669916689,0.1250952035,-0.4197683334,0.1320843995,0.0480908081,-0.2092866898,0.1106635705,-0.1319660395,-0.3402947783,0.0305394866,0.4360516667,-0.1923869103,0.4651909173,-0.2223875225,0.1703501493,-0.1473429501,-0.1536527276,0.0298600588,-0.0900716111,0.2436603159,0.2737259269,0.3140886426,-0.2141556144,-0.0708004832,0.2795857489,-0.5427033305,-0.2861479223,0.223249197,-0.0327365473,-0.1293388009,0.083870694,-0.0071252678,0.0840421841,-0.3345767558,0.0517720655,-0.102933906,0.0879867971,0.5746050477,0.191638723,-0.0430085361,0.2481711358,0.0756601617,-0.1675255299,-0.0089375759,0.2714706361,-0.207121104,-0.5179976225,0.3182075024,0.1386426985,-0.0428776555,0.3085761368,-0.6279010177,-0.1342575699,-0.2336850762,0.100689888,-0.1068340912,0.1550725847,0.0871676654,0.0960529,-0.1449069381,-0.3435560763,0.0048925113,0.2848476768,-0.2362652719,0.0006516685,-0.0349547826,0.0629395843,0.1345201433,0.4822039604,0.0318293944,-0.1654725671,0.200495705,-0.3054215014,0.6046196222,-0.2599270642,-0.2400037944,0.0599017553,0.1142046899,-0.1535027176,0.234093219,-0.0606404692,-0.1787326336,-0.0133119877,0.1014661565,-0.1196279526,-0.3492877185,0.0903233215,-0.3583359718,0.1568625569,0.1205483004,-0.0469608344,-0.0249231085,-0.3814910352,0.2835078239,0.1921522617,0.0423277169,-0.2172423154,-0.3888531327,-0.454675436,-0.10856013,0.1654713601,0.1470241696,0.5874216557,0.0064070574,-0.2722571194,0.1946412474,0.2674406469,0.7342179418,-0.1391730607,-0.0503010638,0.3903177679,0.3925023973,-0.3472799659,-0.0092262616,-0.2599988878,0.0213550497,-0.1138023883,0.2361126691,0.1701416969,-0.280518651,0.2421051562,0.2854965031,-0.2466304898,0.1876034737,-0.5327395797,-0.343834877,0.0030114246,0.1400165558,0.2260756344,-0.0294362605,-0.1500397772,-0.0529747792,-0.3793028295,-0.2613240778,0.0598471276,0.0349562019,0.4408480227,0.1943806857,0.3049543202,-0.0261976104,0.6141986251,0.3659821153,0.4922839701,-0.5450040698,0.1024127901,0.2489347011,0.0703405738,0.2837764621,0.4125652015,-0.1229415163,0.0962915942,-0.4376490116,0.1911679357,-0.3466957211,0.4170636237,0.20371674,-0.1481145918,-0.0529025272,0.0637534186,0.0556594357,-0.0155519368,-0.1059591547,0.6313262582,0.269338429,-0.0313745104,0.2944715321,0.1011382863,1.2075374126,-0.1255977303,0.3258848488,0.3455070555,-0.4239888489,-0.0855606198,0.0722757354,0.1262742877,-0.0671244711,-0.5808190703,0.0595162548,-0.0741003975,-0.2176912725,-0.1429721564,0.1091328785,0.0651676208,0.1057809964,0.5431858897,0.235861212,-0.0209707171,0.1053520069,-0.0252951458,0.016069822,-0.0120059047,0.5678728819,-0.1512120366,-0.3326204121,-0.087358214,0.0550680682,-0.0100521231,0.0195954721,-0.0133461254,-0.6196104288,0.2293964177,0.0329568237,-0.0710296631,0.0058566988,0.4161221981,0.3843311965,0.0319344103,-0.1466644853,0.3212939501,0.4691763222,0.2076121867,0.1192942485,0.0424969308,0.4516296089,-0.0465639569,-0.1062377393,0.2462542951,-0.1292869449,-0.0807303339,-0.1673709899,-0.0603705309,0.038864404,-0.1431066394,-0.048150599,-0.2568143904,-0.270617038,-0.059519697,0.0889462605,-0.2223586291,0.0245296545,0.3424888253,-0.1079980955,-0.5061223507,0.2096646726,0.276037395,-0.2299588323,0.1969762743,0.3436801732,-0.1816192269,-0.1202588603,-0.256157279,0.3886810839,0.0609171651,-0.5470731258,0.2829530239,-0.1245821118,0.429043591,-0.0138290934,0.5695757866,-0.0682384297,-0.0612271167,-0.0057996651,-0.2351913005,-0.3249398172,0.2648646533,0.030601548,0.3474400342,-0.1750466079,0.3146694005,-0.0190683883,-0.3286039531,-0.2675201893,0.2184214294,-0.195070371,0.1273489296,-0.1537469476,0.0576180927,0.2064946145,-0.2634270787,0.085400559,0.1924275458,-0.0160661768,-0.1238691434,-0.0520596728,0.1746418625,0.0810231417,-0.3536424935,-0.0041074217,-0.2538175881,0.0943080038,-0.047871232,0.1810651571,0.0045773825,-0.3380604386,-0.1056242958,-0.1124090403,0.0229120702,0.0872407481,0.058114782,-0.150458619,0.3196382821,0.0846866742,-0.0637011901,-0.2290772349,-0.2141297311,0.019038396,0.1848994642,0.0154668232,-0.3869718611,0.0789402574,-0.070252046,-0.127346918,-0.1185506508,0.4420285225,0.250708729,-0.4983851016,-0.1735827476,0.4078104496,0.2135398388,-0.3365501165,0.1308618188,-0.2609257698,0.003469344,-0.088597402,0.1060303822,0.1725834906,-0.2121965587,-0.1787487715,0.1510977447,-0.4579036236,-0.0566668771,-0.2281567156,0.6297215819,0.0878878087,0.2766057849,-0.0458956733,0.0726211742,0.0864037275,0.2339900881,-0.0945891514,0.5667414069,0.1749604195,-0.0305551048,-0.0447542705,-0.319888413,0.0433935151,0.2390377373,0.0055899844,0.2459430695,-0.0508207157,0.0674172193,0.1743658781,-0.4217628837,-0.0918770209,0.1651830971,-0.0224139057,-0.2720594108,-0.3387703598,-0.2180982232,-0.2728937566,0.12181665,-0.0307897348,0.0357925035,-0.0137483692,0.132101655,-0.0327218994,-0.2596278489,-0.0538723655,0.2903204858,0.344977051,-0.2803207636,0.036957033,-0.0194630641,0.0688491538,0.1412667036,0.3474190235,0.2181869447,0.0993962884,-0.2041796148,-0.0227078684,-0.0136043811,0.0684709847,0.0822915956,0.3425371349,0.0605486557,-0.1241256222,0.3070169687,-0.0458528548,-0.097131595,0.1166850254,-0.0437982343,-0.0444129519,-0.334315598,-0.0237693973,-0.0223184638,-0.2086401284,-0.1853766292,-0.0822281688,-0.1447233111,-0.0550839342,0.3361219466,-0.5006502271,0.2306670398,-0.1364663988,0.046309419,0.0424813442,0.2000268549,0.2953695655,0.5002297163,-0.3583451509,-0.5849112868,-0.5454111695,0.2477872968,-0.0550521836,-0.1414011866,0.1582048833,0.0288414918,-0.0622714907,-0.0681782961,0.274038583,-0.0469748341,0.0142406924,0.4672357142,-0.2215175331,0.0888676867,-0.3798103631,0.1021500453,-0.1448572278,-0.6390478611,-0.0195399132,-0.2257535011,0.0618604831,-0.0238049068,0.0974493399,-0.0718916506,0.6155150533,0.5243155956,0.0908330306,0.1478902549,0.1301814914,0.0306081325,-0.1639400721,-0.1677481234,-0.1326243728,-0.1898752004,0.0500835441,-0.0861456469,0.0079657948,0.2537888587,0.1178609505,-0.2608484626,0.1832638979,0.4560765028,-0.2930698693,0.0799366981,-0.2245865464,0.2090032101,0.3347372711,0.3146527112,-0.0712094083,0.1140887961,-0.4180307984,-0.3265278637,0.3847385943,-0.441431284,-0.1340732574,0.0491738394,0.3551390767,-0.1623359621,-0.0116730267,-0.5308157206,-0.1808762699,0.3966633379,-0.0071508149,-0.3978721499,0.2556317747,0.0600792915,0.1819682717,0.1403826475,0.7158242464,-0.03503526,-0.2116898149,-0.0162410177,-0.1845760942]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2932","title":"Conda build fails","comments":"Why 1.9 ?\r\n\r\nhttps:\/\/anaconda.org\/HuggingFace\/datasets currently says 1.11","body":"## Describe the bug\r\nCurrent `datasets` version in conda is 1.9 instead of 1.12.\r\n\r\nThe build of the conda package fails.\r\n","comment_length":7,"text":"Conda build fails\n## Describe the bug\r\nCurrent `datasets` version in conda is 1.9 instead of 1.12.\r\n\r\nThe build of the conda package fails.\r\n\nWhy 1.9 ?\r\n\r\nhttps:\/\/anaconda.org\/HuggingFace\/datasets currently says 1.11","embeddings":[-0.4641540647,0.0640663132,-0.1675162613,0.0911422074,0.0242198519,-0.1167648509,0.0046573328,0.4584467709,-0.1488352567,0.0218314379,-0.0215561539,0.1672185659,0.1742765456,0.4530319273,-0.0687930807,-0.0687004402,0.2214742899,0.1521472484,-0.3632488847,-0.0315286219,-0.2346586734,0.3556938469,-0.1418616623,0.042954348,-0.1667399108,-0.1111968309,-0.0725726113,-0.0722472295,-0.3525906503,-0.309075892,0.5409511924,-0.1080390662,0.0791708231,0.5754619241,-0.0001009286,-0.0032789898,0.5302111506,0.0530045889,-0.2712946236,-0.1803977042,-0.3502198458,-0.0796987191,-0.0399142914,0.1341502666,-0.025258461,-0.3136871755,-0.0039637419,-0.0812839568,0.2737195194,0.1549416184,0.3308033943,0.0851677209,0.4155243039,-0.4425582886,-0.4505780041,0.2404605895,-0.3100655675,0.1933912933,-0.0950218216,-0.0654576346,0.3125593662,0.2748890221,-0.0108567653,-0.2175896168,0.0571583509,-0.1458919346,-0.1065396816,-0.1608880013,0.2085220516,0.082794033,0.5673777461,-0.3365124464,-0.4459105134,0.116349265,0.0547229312,-0.1276331395,0.2863733172,-0.0994853154,-0.1852693409,0.2289404869,-0.3072545826,-0.1659559906,-0.0998794213,0.0542916507,0.0351006612,0.0843344927,-0.2238170356,0.0395075008,0.0850799978,-0.0795847178,0.0889546126,-0.017170459,-0.0044747209,-0.0164357647,-0.2207264602,-0.1931111813,-0.1908893287,0.0040363134,0.290076822,0.0416318104,-0.2814685106,-0.2031846493,0.0022204889,0.0698052868,-0.057144139,0.0626951158,0.4606170654,0.077450268,0.145462364,0.2514044344,0.1704059988,0.0070430655,0.0021299806,-0.3041796386,-0.2461816221,-0.0392220579,0.2495234162,-0.2043797076,-0.1059830338,0.1044391021,0.0609567054,-0.0279124603,0.0041456302,0.0737251118,-0.098237142,-0.1265262663,0.0580965318,0.1755417287,0.0175039303,0.0253259484,-0.2480798662,-0.19085069,-0.0307969321,0.012003894,0.2311127931,-0.3263165355,0.4701302052,-0.1246728003,0.1668923795,0.0062446366,-0.1614912301,0.078064017,-0.3607374132,0.4940434396,-0.2841105163,0.1163817868,-0.0540333688,-0.0690908656,-0.0722390711,0.009847221,-0.1026801541,-0.0875664055,-0.4648688734,0.2391197681,0.1965070069,0.1023772582,-0.1715472639,-0.0221420992,0.0536693893,-0.0455940999,-0.0935841873,0.0454386994,-0.0532059669,-0.1457587034,0.2406151742,0.2125408351,-0.1583735943,0.1107756644,0.1150244549,-0.2624725103,-0.2327995002,0.0774188638,-0.0671860948,-0.1484007537,-0.0090418523,-0.2777799666,0.1654792577,-0.4407516718,-0.5214886069,0.0708692968,0.1568902731,-0.1486879587,-0.0275132805,-0.1143989637,0.1878380924,-0.1143638939,0.1809624135,0.0786408931,0.005723747,-0.2770836055,-0.1697548479,-0.275437206,-0.241848886,0.0517572574,-0.0461218953,0.1780284196,0.042331446,-0.2044620216,0.1104750931,0.143998459,-0.2036080956,0.2029421628,0.4378283918,-0.1347674131,0.0343065746,-0.1481755525,-0.5067104101,0.0793106779,0.1355768591,-0.0216997135,-0.0685060993,-0.1599982828,-0.3334034681,0.2762608528,0.0679381713,-0.217081219,0.2146058977,-0.1150966808,0.1284432113,0.0793034956,-0.1678137481,0.7080689073,-0.137064442,0.2539079785,-0.1866996437,0.1552656889,-0.2871536613,-0.1260931939,0.2259406298,0.1750397533,0.2101011425,-0.1859097183,-0.1064059809,0.3100512624,-0.1132724211,-0.1732333302,0.1977748871,-0.2093673199,0.0850393698,-0.0069135549,0.1005597562,-0.0600860864,-0.1986969262,0.2078127116,0.0507737435,-0.0467366427,0.1806784272,-0.0241424423,-0.0142447082,0.2336301059,0.4198890924,0.0989498571,0.0189749841,-0.1595772505,0.1046580002,0.1180455238,0.2821315825,0.0236277115,-0.2964583933,0.1159171835,0.3984741867,-0.1097792536,-0.0726787671,0.2535006702,0.0919811726,0.1495138854,0.3302159607,0.2453039587,0.1037769392,0.0531402938,-0.095606029,0.2007383853,-0.3480142057,-0.0298374817,0.1060575619,0.0322836116,0.2812074125,-0.0845356435,-0.1015645266,0.0695140362,-0.0760767683,-0.0694011524,-0.1414503753,0.0929866508,0.0086003188,0.0421742573,-0.1739253849,-0.0793259069,-0.255325973,-0.2271770239,-0.2650867403,-0.020353796,-0.0007564036,0.1378406584,0.210903734,0.4877940714,0.0403436422,0.0395924002,-0.0504935272,0.0288728029,-0.1351604462,0.1316828281,-0.1524738818,0.1581934094,0.0931431428,-0.0417627729,0.1490940601,-0.4683484435,0.1915226579,-0.0488868766,-0.5337531567,0.162495628,-0.2750543058,0.4730536938,0.2408772856,0.0991982818,0.0325651355,-0.2097351402,0.1801730543,-0.1365850121,-0.1675789505,-0.219999522,-0.0309822541,-0.3386186361,-0.0958381668,-0.1521292478,-0.135665834,-0.1753454208,0.3863707185,-0.0221327562,0.0562425889,0.3409312069,0.4604199231,0.0470611565,-0.1918748617,0.0891415849,-0.1435628682,-0.4498470724,0.056229163,-0.2765717506,-0.2497362643,-0.0989951342,0.2478436381,0.4750163257,0.0358756483,-0.5661168694,-0.1986954808,-0.0168789178,0.1650635004,0.0965101942,-0.013050952,0.2417072803,0.0429633595,-0.1247905418,-0.2604943216,-0.3049151897,0.1147871763,-0.2061828226,0.3588501811,-0.0101826554,0.4184118509,-0.1961857677,0.2067244649,0.570433557,-0.193848446,0.1289104074,0.1958947629,0.646984756,-0.2325790077,-0.2919476926,0.1788204014,0.2425242811,0.2087279856,-0.0302555487,0.2250890583,-0.0930705443,-0.2837232053,-0.2128251344,-0.1717529297,-0.1405042112,-0.2474866956,0.1902405322,0.1072169468,0.1035116091,0.250531435,-0.1334473193,0.0827420875,0.1420256048,0.2684525549,-0.1062390804,-0.0453671105,-0.2940100133,-0.0090564089,-0.5889351964,0.3405347168,-0.115719378,0.5059463382,0.0934728608,0.2877984345,-0.1646546274,0.1292834878,0.7751023173,-0.1143795177,-0.1694623381,0.0409080386,0.1196236387,-0.386511147,0.1661832929,0.269202739,0.4197353423,0.2771253586,0.570371449,-0.2825520635,-0.2901853323,0.4174397886,0.3029113114,0.064699769,-0.1520482302,-0.2908979654,-0.3592730463,-0.0160685554,-0.270186305,0.0235399101,0.0294448528,-0.0739184543,-0.3009604812,-0.2542800009,-0.1923899204,-0.0933020413,0.1007940471,0.1852276027,-0.1531287879,0.4446290135,0.1214143708,0.143140167,0.3937529624,0.3233439922,-0.3388970494,-0.4432545602,-0.1255219132,-0.0592223518,0.2138877064,0.0718777403,-0.2636139393,-0.4265669286,-0.0393774845,0.0544238426,0.0062244446,-0.1491227001,0.4722716808,0.4071490765,-0.4898706079,-0.3163699806,0.3107614219,0.1555151641,0.0137672732,0.6786463261,0.1637614667,0.0148167852,0.0191572513,0.0330405645,0.6511470079,-0.0443928279,-0.0499101281,0.0446797237,-0.2227658033,0.3669653833,0.0488393344,-0.0652379617,-0.1855852902,-0.3706175685,0.0375149027,0.0188984033,0.3522101343,-0.0635798946,-0.1897932589,0.2009338439,0.0050997678,-0.0510391779,0.1347304136,0.0466747731,0.0570177846,-0.0017096179,-0.2023081779,0.2671092749,0.1218866855,0.0911121443,-0.0820011273,0.0522740521,0.0302037653,-0.2407122105,-0.3367681503,-0.0089284312,-0.052127514,0.0197862741,0.5075984001,-0.3320956826,-0.006780732,0.1831906736,0.532916069,0.268796891,-0.2907969654,0.0985455588,-0.1932935715,-0.1898310483,-0.0396836028,0.08748281,0.1513931453,-0.1723753661,-0.2542519271,-0.081403546,-0.0534898341,-0.0884619206,0.1103957668,-0.2348884344,-0.2580360174,-0.2222904563,0.0855715796,-0.0798651874,0.1912613213,-0.3580775261,0.2370942533,-0.0410242192,0.1347132623,0.1535924971,0.0453106724,-0.1981552094,-0.0836486146,0.2549034357,0.0502661504,-0.06083728,0.1065000668,-0.0302821025,-0.283223778,-0.199485153,-0.2953700125,-0.1355858594,-0.2921869159,-0.0426430628,-0.1656433195,-0.1487552971,0.030801909,0.2645925581,0.1511131227,-0.1485049576,0.1459294707,-0.217853874,-0.4190172255,0.3893738687,-0.2418382168,0.081914641,0.1848844141,0.3065513074,0.2461230457,-0.1169423535,-0.3740441203,-0.0096954424,-0.132256791,-0.0716296583,0.1841001213,0.0253096037,0.1819106936,-0.1267617196,0.1651816815,0.166282177,-0.3246473372,-0.2387083918,-0.1031778455,0.0037646173,-0.0249224957,-0.0605644509,0.1770996302,-0.1955480129,-0.3712851703,0.0236387346,-0.0425952226,0.1294345856,0.0975574777,-0.047995396,-0.0325053371,0.1657387912,0.1664978564,0.1940087974,0.2027938217,0.2586369216,-0.0825744346,-0.0052247988,-0.1060191244,-0.0838461593,-0.1263361126,0.0679822192,-0.3276896775,0.2343242615,0.3120971024,-0.153106302,0.0876108259,0.3479898274,0.4070643485,0.143855691,-0.0349122398,0.0139444303,0.4124389291,0.2923622727,-0.398332715,-0.037430156,0.2405215502,0.2383926958,-0.129919067,0.1489732713,0.0590852872,-0.1693943292,0.4098250866,0.1403006017,0.2205741554,0.0602487288,0.3973987997,0.3222332597,0.2497254461,0.0663221255,0.4399477243,0.2902185321,0.2228898108,0.4001783133,-0.2888497114,0.0846696123,-0.0285459496,0.1135114208,0.1290114373,-0.104021512,0.2614093721,-0.0871152207,-0.0681195483,-0.1083463579,-0.1178631559,0.2789145708,-0.36383003,-0.0831985399,-0.3642092049,0.0756527483,-0.1014542431,-0.0847678632,-0.0112250736,-0.0103246085,0.3061100245,-0.1312603951,-0.1614249945,-0.1096174344,0.5349951982,-0.1963897645,-0.0826527104,-0.4234129786,-0.0658413172,-0.0159965903,-0.0684605166,0.1299069077,0.5619087219,0.0393128321,-0.0317944586,-0.0642656311,0.5191246271,0.3900793493,0.045922596,0.1561653614,-0.1683365554,-0.0400273055,-0.0259133242,-0.136891067,0.3559701443,0.1754811555,0.0503647886,0.1609129608,0.269453615,-0.1784174144,0.0201286301,0.1036785692,0.2914521694,-0.2207628787,0.122631833,-0.3412408829,-0.1359685957,-0.1008708104,0.0563278571,-0.2205191106,0.1479242891,0.2504945099,-0.1503033936,-0.0945490748,-0.1575134844,0.1470509917,-0.1119528338,0.4386317134,0.2454130054,0.2079600096,0.0663516819,-0.411449343,-0.267690748,0.2406549454,-0.0822616667,0.1564208269,-0.0161260664,-0.1036028862,0.0204328652,0.1059478,0.3543419242,-0.0168153346,-0.0535218529,-0.2916656733,-0.325381577,-0.1110316217,-0.1225745976,-0.0263410211,0.162237078,-0.1866420358,0.18117024,-0.1827753931,0.0607612543,-0.1764393598,0.1855309606,-0.0066312873,-0.2995755672,0.3030668199,-0.0313812457,0.0356941372,-0.011500272,-0.4884319603,-0.2003731579,-0.3177161813,-0.059405148,0.3417550623,0.0319521949,0.131634593,-0.3392523825,-0.0623484664,-0.1945575625,0.3203814328,0.2225255221,0.0371220186,-0.1918862164,0.1454050988,-0.08730115,-0.2772326469,0.2167028487,0.3893757761,-0.1664120555,0.2395646125,-0.16181238,-0.3641056418,0.4493977427,-0.1083183736,-0.1574521214,-0.0138091575,-0.0823283941,0.0318232365,-0.2974265218,-0.5770468116,0.3174734712,0.1961872876,0.1672439873,-0.1092537194,0.1885131449,-0.1230864972,0.0585776381,0.0498388484,0.7527417541,-0.0445063934,-0.200931415,0.3679227531,0.028568225]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2932","title":"Conda build fails","comments":"Alright I added 1.12.0 and 1.12.1 and fixed the conda build #2952 ","body":"## Describe the bug\r\nCurrent `datasets` version in conda is 1.9 instead of 1.12.\r\n\r\nThe build of the conda package fails.\r\n","comment_length":12,"text":"Conda build fails\n## Describe the bug\r\nCurrent `datasets` version in conda is 1.9 instead of 1.12.\r\n\r\nThe build of the conda package fails.\r\n\nAlright I added 1.12.0 and 1.12.1 and fixed the conda build #2952 ","embeddings":[-0.5863975883,0.2035934627,-0.232974872,0.1968558878,0.2075668573,-0.0917660967,0.0932786614,0.5642169118,-0.3096557856,0.0915806666,0.0279443692,0.2152974308,0.0675806627,0.4218905568,-0.1395219564,0.0038826414,0.0308199879,0.2241077572,-0.4182020426,-0.0123287281,-0.1885185838,0.2447249144,-0.1071585938,0.0497225188,-0.245254755,-0.1354384422,0.0322817974,0.0517660826,-0.3269813061,-0.3223083019,0.2895521522,-0.0699542463,0.029348569,0.4992179871,-0.0000913996,0.0741490647,0.3882876933,-0.0257503297,-0.2825210094,-0.0777058378,-0.3032828271,-0.1225462258,-0.0787501559,-0.0245858394,-0.1226112768,-0.2728967667,-0.1174333841,-0.1095196307,0.1959647387,0.2553437352,0.4446544051,-0.0068132705,0.2757855952,-0.4212940931,-0.1860865951,-0.0436917767,-0.2034923285,0.0971226022,-0.0531155542,-0.0714639053,0.0847257301,0.2414616197,-0.0279956367,-0.0224732459,-0.0340468511,-0.164331913,0.1127335355,-0.135215342,0.2132426351,-0.0154786697,0.6562217474,-0.3307790458,-0.2999447286,0.1405354589,0.0567250699,-0.1853858232,0.2922054827,0.0008688812,-0.0837874711,0.2183299512,-0.2021050006,-0.1321849227,-0.1009486616,-0.0226571206,-0.0519170277,0.1495386511,-0.1100725085,0.0149995303,0.0504197516,-0.2023219615,0.362195164,-0.0173723605,-0.0971301347,-0.15897879,-0.2414802462,-0.1807620078,-0.0219156146,-0.1191595793,0.1366502941,-0.0261811931,-0.1909336001,-0.0144785335,0.0841854289,0.1068868414,0.0498245768,-0.0046015838,0.2775596678,0.2492183447,0.2414739579,-0.0120759141,-0.0095909256,0.0504127778,-0.0319391079,-0.4023247957,-0.196750626,0.0584249012,0.0760341808,-0.2049568146,-0.2722491622,0.1437727064,0.0510823801,0.0147776939,0.0015100588,0.2345990539,-0.016674662,-0.1202628314,0.0807097182,0.1372875571,-0.0713210627,-0.0524630882,-0.3114102781,-0.1119067445,-0.121408619,-0.0771802515,0.3006842136,-0.1038021892,0.3707338572,-0.0830180496,0.0617507473,-0.032412041,0.0544019453,0.027759565,-0.2395564318,0.3888090551,-0.0499059856,-0.0003740962,0.0252180807,-0.069863148,-0.027930323,0.0828347281,-0.0366372727,-0.2251679599,-0.4237506986,0.3854449987,0.1938515007,0.0662305206,0.0716949105,-0.0636549145,0.1177643538,-0.1366869509,-0.2729662657,0.0184946395,0.0636605322,-0.0871889219,0.2279902995,0.2070731521,-0.2756123841,0.1650138944,0.1333748549,-0.3019257784,-0.0072152917,-0.0268017929,-0.1443531364,-0.0073989532,-0.058277104,-0.1680708975,0.3321492374,-0.295216471,-0.4482250512,0.1187651679,0.0744725466,-0.328696847,-0.0668687671,-0.0782488808,0.196417436,-0.0056965332,0.1062466353,0.0992733911,-0.0286835041,-0.0740173683,-0.2667303383,-0.2071607858,-0.1118903831,0.1681469083,0.057727553,0.1447627693,0.1092381552,-0.2082567811,0.0788387135,0.0829256773,-0.0572477691,0.180361703,0.3837525547,-0.0759066194,0.0168578904,-0.1095519215,-0.3079583943,0.1055386439,0.0256883148,0.0204669945,-0.064672716,-0.0481392927,-0.3070678413,0.114816241,-0.0371600352,-0.0541843511,0.38391608,-0.0289845057,-0.0508067682,0.0003028249,-0.0711915568,0.3884094656,-0.150223583,0.0347469635,0.113964498,0.1847821027,-0.3705210686,-0.2101033777,0.2375807166,0.0689672604,0.2542420626,-0.1483214498,-0.1656304598,0.3099377453,-0.0822563544,-0.1912937462,0.3436691761,-0.1129948199,0.0022904254,-0.0102867652,0.0606409721,-0.0951388702,-0.1239355877,0.0992038921,0.0730591044,0.0231340732,0.1158568263,-0.1084979028,0.0092544667,0.189959541,0.5791200399,0.0050684898,0.1632045656,-0.2162691951,0.0145309176,0.055825673,0.1821321249,0.037359044,-0.2978574932,0.0825711638,0.3980308175,0.0076435688,0.0726711303,0.0926921591,-0.0076113031,0.1124799475,0.2614999413,0.2583378851,0.2037272304,0.1349072009,0.0081493929,0.1047509387,-0.2374267727,-0.0164957829,0.2083378881,-0.0552917421,0.2838928998,0.1663821787,-0.0411133058,-0.0695776567,-0.2452527285,-0.1380629987,0.0534085073,0.2105703056,0.037322551,-0.1095859855,-0.2824200094,-0.0017022508,-0.1233533695,-0.1432923526,-0.0640132949,-0.1281411797,0.1245221123,0.0895596594,0.00977821,0.5170032978,0.0143528832,0.0652859882,0.0990787596,0.009667472,-0.0518003739,-0.0790770501,-0.127101317,0.2408400774,0.2044294775,0.0509005859,0.2018431723,-0.3337310851,0.1491089165,-0.2748525441,-0.366984427,0.1274038702,-0.094208695,0.3852852583,0.2254270911,0.0919218436,0.0747521743,-0.2698901594,0.110820815,-0.0580758601,-0.0898231268,0.0070468765,-0.0787847564,-0.3627088368,-0.2367329299,-0.3228822947,-0.1648404151,-0.3588816226,0.1358786523,-0.060873311,0.0982486829,0.0736362413,0.3923217952,0.0536427498,-0.0562113784,0.0783555135,-0.222058624,-0.2932739556,0.1643790603,-0.3600728214,-0.3115933836,-0.0019090399,0.1679629087,0.5059683919,0.0179702938,-0.5560306907,0.0464412272,0.0780357271,0.1711068302,-0.0487899296,-0.0328627974,0.2273453474,0.2445125133,-0.2243379951,-0.3177289069,-0.2355287969,0.1948930621,-0.153514728,0.213289544,-0.2728606462,0.3523001075,-0.1319694966,0.166004017,0.2222482264,-0.2386436909,0.219208017,0.1830454022,0.5038248897,-0.1305069923,-0.1817170829,0.2021608651,0.1581531167,0.119042255,0.1395495832,-0.0279069915,-0.3791646361,-0.2756888568,-0.1051902995,-0.2455618232,-0.1936589777,-0.2074110061,0.0622599088,0.1791433245,0.1262440085,0.1772680581,-0.1568305194,0.1184369698,0.0211998075,0.0384883247,-0.0041033318,-0.1069697961,-0.1779532284,0.0358848907,-0.2792441845,0.2821165323,-0.2120809704,0.3916369379,-0.0623135567,0.2048662305,-0.0723558962,0.0009504542,0.6512207389,-0.2112644315,-0.0737034976,0.1125226542,0.1181229427,-0.2302242666,0.0650186986,0.083207123,0.3839910328,0.3091752231,0.3659373522,-0.3039242029,-0.1437313408,0.2834871113,0.1504956931,-0.0098037198,-0.2056625634,-0.0874546692,-0.2531458437,-0.1671919972,-0.1062273532,-0.0554941334,0.1245535091,-0.1537984908,-0.4580590427,-0.2820667922,-0.1694457382,-0.01758972,0.1693876237,0.2554765642,-0.1612110734,0.5345661044,0.0045534945,0.1155918241,0.4735500813,0.2798543572,-0.189198032,-0.2257909626,0.0917203054,-0.2555652261,0.0836775675,0.0862136558,-0.2363713384,-0.341285795,-0.2195706218,0.1084240749,0.076264888,-0.026122503,0.3726834059,0.2485917658,-0.4293208718,-0.2518617511,0.242318511,0.0162319206,-0.0728726611,0.5016472936,-0.1048982367,-0.0741726682,0.0676231533,0.0824791938,0.66672194,0.0610501505,-0.1266774684,0.006856665,-0.1592148095,0.1626461744,-0.1145113334,0.0037972457,-0.2875882685,-0.4622149169,0.0564569943,-0.0223851558,0.2537106872,0.2087925076,-0.3033083379,0.0743170977,-0.0972160399,-0.2132728994,0.0625187531,0.0630918145,0.0538320169,-0.0261505786,-0.0835556015,0.361900717,0.0005899728,-0.0727578327,-0.1062728912,-0.067716673,0.0350575037,-0.1675877124,-0.2483343184,0.0498329103,0.0010737955,0.2386963516,0.1627294123,-0.2036982179,0.0009662647,0.1117749289,0.3256909251,0.338988781,-0.2267630547,0.2255694866,0.0059769298,-0.1343333125,-0.0487091094,0.1586330235,0.0849824697,-0.1907640249,-0.2566301227,0.0247879233,-0.0578859113,-0.0663252398,0.1320600957,-0.1353257447,-0.0472149029,-0.2453776896,0.0064868447,0.0142924003,0.0114533948,-0.3371916115,0.3463756442,-0.0527707152,-0.0052369772,0.3470881283,0.012192864,-0.1826528311,-0.1017469913,0.2007384747,0.1102424562,0.0318651125,0.1323222965,-0.0796627328,-0.3194353282,-0.3690176308,-0.1674910337,-0.2219544947,-0.2326253057,0.0004745419,-0.0364109874,-0.2213421166,0.1392650902,0.2272441387,0.103966862,0.1268080324,0.12728405,-0.1871058792,-0.2903102636,0.1435471773,-0.2305998057,0.224498108,-0.0108646248,0.3226473331,0.1365102828,-0.0325094052,-0.5056607127,-0.0034738155,-0.376393348,0.0241533406,0.0283821672,-0.0198156852,0.0323873013,-0.1437138319,0.2947608829,0.2062872499,-0.2226619273,-0.3661328256,-0.0312338322,-0.0175669845,0.0497309305,0.0167891141,0.1219146401,-0.1331655234,-0.4057909846,0.0281089693,-0.018162312,0.0284162387,0.0422412306,-0.0115903253,0.1159084886,0.0983861461,0.1967673749,0.1101816744,0.1371367872,0.2352458239,-0.086849317,0.008582918,-0.0457893237,-0.0563464426,-0.0349864177,-0.0246918034,-0.2511456311,0.2289187759,0.3598465621,-0.2082441896,0.0332717411,0.3909105957,0.334284842,0.207790032,-0.1314865947,-0.0822631642,0.3971404433,0.4947049916,-0.4555192888,-0.1380333006,0.0083031477,0.0364327542,0.0368086807,0.0361693576,0.0356303714,-0.2292745262,0.308706969,0.2625923753,0.2341880202,-0.2097451538,0.2820364535,0.3012544215,0.0510902666,-0.0560819656,0.3553821146,0.1626077741,0.0981256738,0.6015193462,-0.2439955473,0.1260926723,0.0874224454,0.0610313267,0.0384354778,-0.1500719041,0.2947190106,-0.0056637153,-0.0532938018,-0.1579737663,-0.2138314545,0.0789796114,-0.327932924,-0.0442125946,-0.3130210936,0.047839202,-0.1836959124,0.0594023429,-0.0732315704,-0.0382933915,0.2386280447,-0.0940453112,-0.0338380709,0.0506904498,0.4240409732,-0.0262736827,-0.0643444732,-0.3497653008,0.0606286004,0.028568374,-0.1260315925,0.0108015751,0.4940970242,0.143856734,-0.0359248295,-0.0589989088,0.3479895294,0.348203361,0.277752161,0.0534951985,-0.1082824618,-0.0092249922,-0.0953803509,-0.2411097437,0.3029560149,0.2273861617,0.0720100328,0.2493595928,0.3347663879,-0.2555821836,-0.0169894919,0.1393467486,0.2789635658,-0.0335982628,0.0914347172,0.0385823473,-0.1384752989,-0.0504327938,-0.0994241536,-0.2475196272,0.0530853234,0.3749343157,-0.1021480113,-0.0129441777,-0.20851098,0.2110289633,-0.1338755935,0.4848892093,-0.0640231967,0.2046496123,-0.0398341753,-0.3171172142,-0.3321355581,0.2472712398,-0.025997106,0.1577498317,-0.0925640464,-0.0027004264,0.0405297764,0.0034110083,0.2962839603,-0.0514624342,-0.0067159436,-0.1673950255,-0.4477352202,-0.031719882,-0.0999102145,-0.0099834604,0.1832084507,-0.2746391296,0.1216634437,-0.1069333702,0.2742959857,-0.1479437649,-0.01545012,0.1173128188,-0.004334955,0.3953119218,0.0229871981,0.3068906069,-0.2120409459,-0.2795613408,-0.139135778,-0.4778122306,-0.1043489054,0.1476300657,0.2280675769,0.1021462083,-0.2248517722,0.0063110883,-0.1715018302,0.265201062,0.1091065183,0.2118775547,-0.1095602289,0.157030344,0.054823719,-0.0997829214,0.0586517639,0.253433466,-0.067046538,0.2125879377,-0.2155797929,-0.3528474867,0.4893417656,0.0142854201,-0.245224759,0.0129644563,-0.0260217786,-0.0059275106,-0.1479931176,-0.4293839633,0.229219079,0.2826619446,0.1066514403,-0.1077293307,0.1915610731,-0.1371898651,0.1114795953,0.0018529911,0.573530376,0.0091546699,-0.2639088035,0.1011295021,-0.0598786213]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2930","title":"Mutable columns argument breaks set_format","comments":"Pushed a fix to my branch #2731 ","body":"## Describe the bug\r\nIf you pass a mutable list to the `columns` argument of `set_format` and then change the list afterwards, the returned columns also change.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"glue\", \"cola\")\r\n\r\ncolumn_list = [\"idx\", \"label\"]\r\ndataset.set_format(\"python\", columns=column_list)\r\ncolumn_list[1] = \"foo\" # Change the list after we call `set_format`\r\ndataset['train'][:4].keys()\r\n```\r\n\r\n## Expected results\r\n```python\r\ndict_keys(['idx', 'label'])\r\n```\r\n\r\n## Actual results\r\n```python\r\ndict_keys(['idx'])\r\n```","comment_length":7,"text":"Mutable columns argument breaks set_format\n## Describe the bug\r\nIf you pass a mutable list to the `columns` argument of `set_format` and then change the list afterwards, the returned columns also change.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"glue\", \"cola\")\r\n\r\ncolumn_list = [\"idx\", \"label\"]\r\ndataset.set_format(\"python\", columns=column_list)\r\ncolumn_list[1] = \"foo\" # Change the list after we call `set_format`\r\ndataset['train'][:4].keys()\r\n```\r\n\r\n## Expected results\r\n```python\r\ndict_keys(['idx', 'label'])\r\n```\r\n\r\n## Actual results\r\n```python\r\ndict_keys(['idx'])\r\n```\nPushed a fix to my branch #2731 ","embeddings":[-0.0634390414,-0.1930613369,-0.0847119913,0.0437259898,0.5914995074,0.0606024899,0.7192493677,0.2848295569,0.1028594598,0.1568417102,-0.116555728,0.7383283377,0.0102749048,0.2990810573,-0.2102174014,-0.2156233937,0.1305874735,0.1316983253,-0.131610781,0.0363861509,-0.2623939514,0.0679177791,-0.375447154,-0.1826451868,-0.1687054485,-0.144836247,-0.2592204511,-0.0473862961,-0.0947361514,-0.271669805,0.1168658212,0.1954511404,-0.2182203233,0.4592967629,-0.0001044443,0.0152202081,0.0200444236,-0.1021577045,-0.2109158784,-0.0972076952,-0.2637723386,-0.0789669305,0.1564198732,-0.3649504483,-0.2155194879,-0.0359100625,-0.4558363259,-0.2674218416,0.2459866256,0.3095501363,0.337433964,0.0961965173,-0.0191967264,-0.0778871626,0.350938499,0.024526272,-0.2062668651,0.1694167107,0.0360042304,-0.0574511774,0.2977561057,0.3134259582,-0.353706181,0.0584034547,0.0068529551,0.0332615748,-0.1524679512,-0.0831187963,0.2396808267,0.182026431,0.2787374258,-0.4583109617,-0.2539326847,-0.2258167565,0.0679665506,-0.4645034373,0.0320331641,0.0488984473,0.1365063339,0.1395174265,0.1109698266,0.0990392044,-0.0189186633,0.1066981032,-0.1106197387,0.4013152719,-0.164026618,0.0244787335,-0.0885713398,-0.0331682153,-0.0251186453,0.00836109,-0.2377885878,-0.0240884684,-0.2194893956,-0.1676763147,0.0029882274,-0.0322995409,0.0095377434,-0.1139194816,-0.046752058,-0.030285297,0.1399154961,0.0920782611,0.3197984397,0.278681159,-0.0635075048,0.4963833392,0.0732233971,-0.0988840014,-0.0371107832,-0.1538478732,0.308842361,-0.1950437427,0.3155840933,0.1642861962,0.238843292,0.0682739392,-0.2911200523,0.0995498076,-0.1539213657,-0.0634099916,-0.0316827819,0.3574388027,-0.0768988431,0.4067205787,-0.0575630032,0.3268230855,-0.0287995134,0.2033368051,-0.2828402519,0.130673483,-0.16348508,-0.2244134843,0.084832713,0.0561836064,0.0570182092,0.3165373206,0.0141386166,-0.2182422727,-0.1813179404,-0.2205047458,0.3185226619,0.1042528227,-0.0451851599,0.2784454525,0.2753438056,-0.391156137,-0.1288427263,0.0011100398,-0.2692237198,-0.0980003327,-0.2712520361,0.301181078,-0.0521826521,0.006486427,-0.0049253348,0.0343260728,0.3726559877,0.1105766967,0.2605312169,-0.263920784,-0.1242048368,-0.3503037989,0.0914837793,0.2431913614,-0.3241707087,0.1112797782,0.3844877183,0.2423685193,0.1121881604,0.2034005076,-0.0473191589,0.0279094465,-0.0148240486,-0.1700301617,0.2520844638,-0.3341756165,-0.2848739028,0.0553748496,0.0027873039,0.0685167909,0.2248959839,-0.1666803211,0.218622759,0.0578945205,0.4702925682,0.1355687827,0.1329183131,0.2321659923,-0.2272477895,-0.0266459659,0.1384196877,-0.0202689543,-0.2603830695,0.1265508235,-0.0973207727,0.0009012923,0.2695514858,-0.1020145565,-0.1591219455,0.1050361097,0.3889386952,0.1192682534,0.0584359989,-0.1191093028,-0.3901537359,-0.1211959422,0.5693612695,0.0484296009,-0.1748030633,-0.420571506,-0.4216828644,0.1798273921,-0.2152584344,-0.0824195817,0.3075227141,-0.003220771,0.0642843321,0.0407320149,-0.2096620351,0.1370342225,0.0039426237,-0.0304533746,-0.0208814349,0.0401547514,0.0119094858,-0.1834635288,-0.0599086955,0.4251461327,0.3132659197,-0.0375363454,-0.167879343,0.4503110647,0.3329874873,-0.1549506932,-0.3464752436,0.2249492705,0.0076201176,0.1943212301,-0.0460130088,0.0911375135,0.0759302452,0.0359970815,-0.3030140102,0.4775167704,-0.3167341053,0.0824443698,-0.234490633,0.1565556824,0.0981085524,0.0481777042,-0.2115463167,-0.5392458439,-0.062703222,0.0825194418,-0.0306131057,-0.0419437215,-0.470782876,0.1967414916,0.467872113,0.083343856,0.2176243067,-0.2235009074,-0.0992184877,-0.0115429945,0.0731081441,0.2889002562,0.3600578606,0.2259647995,-0.1365745217,0.0370337591,-0.0854712948,-0.2084178478,0.2183267325,0.0393683836,-0.0967671201,0.1753202826,0.3845235407,-0.0736434236,-0.1551766545,-0.0109753758,-0.0839595348,0.1736816615,-0.383297354,0.1738462895,-0.357047528,-0.0811517164,-0.1417519152,-0.2927251756,0.1747904569,-0.4272394776,-0.1525552273,0.1162470058,-0.0045933891,0.1375183463,-0.0211186819,-0.0319030136,0.2776056528,-0.1820731759,0.0417787805,-0.4216853678,-0.0939931944,0.1455974728,0.0303103533,-0.2044923753,0.143839553,-0.0300900023,-0.2565354407,0.0638367087,-0.1259914637,0.1559157372,-0.2809145153,0.1430458874,0.4829552472,0.3382579684,-0.0894901603,0.0548964813,0.3780276179,0.2132316977,0.2059809715,0.138030231,0.0068890937,0.0550270453,-0.2424753159,-0.3511870801,-0.0555490851,-0.1354611665,0.1351916194,-0.1916234791,0.0870930701,0.2025233358,0.2921889722,-0.1789580435,0.0213473085,0.2048195601,-0.3525743186,-0.0658937842,0.1719931066,-0.1936250478,-0.2129869014,-0.0455462337,-0.0391085483,0.0343543291,0.003133951,-0.1914339364,-0.5200380683,-0.1579808742,0.0683984458,-0.1270274371,0.2740614712,0.054213278,0.1716973037,-0.0757222623,-0.124540098,-0.1934963614,-0.0346360952,0.2626588643,-0.1748870909,-0.0929444581,0.2631197572,0.0199016724,0.1118289381,0.0432161056,-0.0100050457,0.2867324948,-0.2556657493,0.3692888916,-0.3020857573,-0.3446291685,0.0101491911,-0.0587351248,-0.0385484286,0.0579442456,0.0086890599,-0.0651061609,-0.03837841,0.2229370773,-0.0282931291,-0.2501943707,0.1912529469,-0.1141608059,0.0595322475,0.0302135851,0.3910180032,-0.2894249856,-0.0952822343,0.0404491015,0.0114112077,-0.2488003522,-0.2327436656,-0.3056177199,-0.4484499991,-0.2699826956,0.1899857521,0.1403895319,0.4892662764,0.1334533989,-0.2362642586,0.0847799107,0.0963971168,0.6822872162,-0.4301690459,-0.3574624956,0.4292676151,0.0859615058,-0.3903828263,-0.2652634382,-0.3802113831,0.0143659599,0.0227175374,0.3876759708,-0.2589642107,-0.244254753,0.2214453816,0.1897400767,-0.0913613886,-0.1592662036,-0.2549227178,-0.0928166062,-0.2348954529,0.093459256,0.0586991273,0.2659591138,-0.1441631168,0.080433853,-0.120664537,-0.1190443635,-0.073037453,0.2516750395,0.3982740939,0.0575780161,-0.0467027016,-0.1222939566,0.2817413211,-0.073820971,0.1757423133,-0.3656084239,0.0336791202,-0.043546617,-0.3389855623,0.1086297184,0.1856503338,-0.0393812321,0.2526772916,-0.4333719611,0.0946321487,-0.1003584713,-0.0472176298,0.4953793287,0.0461364612,-0.2542179227,-0.685565114,0.1015549749,0.0903888866,-0.0064569241,0.3456730247,0.0725662932,-0.1828046888,0.7794381976,0.0889574587,0.7490982413,-0.0867960528,0.2408003211,0.2939781249,-0.2088262588,0.3112678528,0.0492751785,0.1543674469,-0.2872996628,-0.1380244195,-0.0044919541,-0.058305636,0.0247331839,-0.0898002461,-0.4097694159,0.2892744839,-0.5145075917,0.2875412405,-0.1983894557,-0.0730571076,-0.0874249563,-0.2267119884,-0.4677990079,0.2220413983,0.0646434128,-0.2087830454,0.0112568969,0.1907317191,-0.065343827,-0.1698995829,-0.1873892993,-0.1199399978,-0.1559698284,-0.060999155,0.0678645521,-0.4865007401,0.0141786886,0.2893814445,0.3962036669,0.3095730245,0.0078290785,0.0623751841,0.3810586631,0.2517932355,0.0658424273,-0.0208446104,0.3145283163,0.2877014875,-0.0982795581,0.133476764,0.1272155195,-0.0942069888,-0.208242327,0.0529583395,0.1855789423,-0.5744428039,-0.1757421643,-0.1527145803,0.2630667686,-0.3218447566,0.2347415537,-0.0754718408,0.0486424267,0.5379700661,-0.1782665998,-0.2557305992,-0.1464184821,0.1999298036,0.1669089943,0.0693257004,0.4183680713,0.209888503,-0.2242714167,-0.3183072805,-0.0735745728,-0.0647625104,-0.0756294131,0.3513303101,-0.0830074921,0.0055152066,-0.0013726874,0.3915545642,-0.2344167382,0.3513912261,-0.0538246259,-0.205533132,-0.1155246943,0.0326848663,0.2926522493,0.1945950687,0.2316054702,0.2053347826,-0.3007096052,-0.0630117357,-0.4568855166,-0.1410906166,-0.0862020776,0.0651070401,0.2362469286,0.0590859801,0.0111563224,-0.0795943588,0.3580130339,0.1183542982,0.1108416393,-0.4295145273,-0.325049907,0.0110547999,-0.0234215762,0.0853728354,-0.1708765328,-0.1580568254,-0.1967953146,-0.3962603509,0.2666503191,0.2017664611,-0.0845334008,0.1594115496,-0.1476027668,0.2420492321,0.069533959,-0.0918390304,-0.1183197945,-0.2192181051,0.1840180457,0.1944650263,-0.0407677777,-0.0157293361,-0.1871426105,0.105258435,-0.0939975604,0.1689095646,0.1012573317,-0.1896534264,-0.0661151484,0.0987796709,0.4865126908,0.1566231549,-0.2018299252,-0.0108075626,0.2645915449,0.3579041362,-0.1296218932,-0.1321053356,0.2255125493,0.2314445674,0.3111750484,0.450210005,0.1397311538,0.0756184161,-0.2331319749,-0.0009882856,0.0487412177,-0.1216368228,0.1389746517,0.5687338114,0.1403339207,0.0290854629,0.1549241543,0.0297774058,0.1120429337,0.3498232961,0.3880805969,0.2931537628,0.4028396904,-0.1947244108,0.0580541939,-0.1981632113,0.3190847337,-0.1623570174,0.0311275516,0.1651246846,0.0851635262,0.2890863717,-0.2256009579,-0.1322600543,-0.0517679639,0.0788591281,-0.2149592489,-0.1735067964,-0.0896615535,-0.0712928176,-0.0761795044,-0.0599798895,-0.2564581037,0.1198756173,0.3142152429,-0.0564944707,0.1837552637,-0.3391599059,-0.0654969811,0.000147028,0.4076169133,-0.1922232509,0.0069478415,-0.0030230694,0.0994503275,0.0413250551,0.0327654257,0.5266547799,0.1835280061,-0.1379824132,-0.2399745435,-0.0410318933,-0.1680832505,0.1132266372,0.0703418106,-0.0068472442,0.1387183964,0.1747312844,0.3146359026,-0.2387895435,0.0065306732,0.2093479186,0.1968953162,-0.176381886,-0.0169236436,0.0693721697,-0.2573129833,-0.1397525519,-0.0770672634,-0.1836999059,-0.0613855198,0.1926912814,0.338663131,0.3150190115,-0.144917652,0.1909878105,0.3556967378,0.2456722111,-0.0497522913,0.0769164115,-0.1025438681,-0.0973583683,-0.2256001085,0.0825739801,-0.1283246875,-0.2780284286,0.0039949431,0.056390807,0.1623618454,0.0148027819,0.0437278971,-0.0130933793,0.0254860874,0.4582059085,-0.2675381601,-0.1802484095,-0.0817581043,-0.1243195757,0.0804289132,-0.2435839325,0.0038390581,0.1951596886,0.2180403769,-0.3459209502,-0.0041370536,-0.1667109132,0.0345008858,0.0528789237,0.0232742596,0.109929055,-0.0425669812,-0.0815995336,-0.2123945504,-0.0985484049,-0.1917688251,0.0674044117,-0.1040592343,0.2106078565,-0.007643329,-0.1253238022,-0.132859543,0.2341602147,0.1957191378,0.0204240587,-0.1200679019,0.257250309,0.0979948491,-0.1329333782,-0.0794446543,0.3134340048,-0.0020583791,0.3305855095,-0.0650895014,-0.4050293565,0.382350266,-0.3645904958,-0.5083729029,-0.0518184118,0.0601342954,-0.1210649461,-0.0529172905,-0.3612802029,0.0644469261,0.1545197666,-0.1112097576,-0.3463765383,0.1828248799,0.0455008373,0.1964670569,-0.1405544728,0.1704312861,-0.0222192314,-0.0930835903,-0.1773957312,-0.1250364035]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2927","title":"Datasets 1.12 dataset.filter TypeError: get_indices_from_mask_function() got an unexpected keyword argument","comments":"Thanks for reporting, I'm looking into it :)","body":"## Describe the bug\r\nUpgrading to 1.12 caused `dataset.filter` call to fail with \r\n\r\n> get_indices_from_mask_function() got an unexpected keyword argument valid_rel_labels\r\n\r\n\r\n## Steps to reproduce the bug\r\n```pythondef \r\n\r\nfilter_good_rows(\r\n ex: Dict,\r\n valid_rel_labels: Set[str],\r\n valid_ner_labels: Set[str],\r\n tokenizer: PreTrainedTokenizerFast,\r\n) -> bool:\r\n \"\"\"Get the good rows\"\"\"\r\n encoding = get_encoding_for_text(text=ex[\"text\"], tokenizer=tokenizer)\r\n ex[\"encoding\"] = encoding\r\n for relation in ex[\"relations\"]:\r\n if not is_valid_relation(relation, valid_rel_labels):\r\n return False\r\n for span in ex[\"spans\"]:\r\n if not is_valid_span(span, valid_ner_labels, encoding):\r\n return False\r\n return True\r\n \r\ndef get_dataset(): \r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path,\r\n name=\"prodigy-dataset\",\r\n data_files=sorted(file_paths),\r\n cache_dir=cache_dir,\r\n )[\"train\"]\r\n\r\n valid_ner_labels = set(vocab.ner_category)\r\n valid_relations = set(vocab.relation_types.keys())\r\n ds = ds.filter(\r\n filter_good_rows,\r\n fn_kwargs=dict(\r\n valid_rel_labels=valid_relations,\r\n valid_ner_labels=valid_ner_labels,\r\n tokenizer=vocab.tokenizer,\r\n ),\r\n keep_in_memory=True,\r\n num_proc=num_proc,\r\n )\r\n\r\n```\r\n\r\n`ds` is a `DatasetDict` produced by a jsonl dataset.\r\nThis runs fine on 1.11 but fails on 1.12\r\n\r\n**Stack Trace**\r\n\r\n\r\n\r\n## Expected results\r\n\r\nI expect 1.12 datasets filter to filter the dataset without raising as it does on 1.11\r\n\r\n## Actual results\r\n```\r\ntf_ner_rel_lib\/dataset.py:695: in load_prodigy_arrow_datasets_from_jsonl\r\n ds = ds.filter(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2169: in filter\r\n indices = self.map(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1686: in map\r\n return self._map_single(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2048: in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\ninputs = {'_input_hash': [2108817714, 1477695082, -1021597032, 2130671338, -1260483858, -1203431639, ...], '_task_hash': [18070...ons', 'relations', 'relations', ...], 'answer': ['accept', 'accept', 'accept', 'accept', 'accept', 'accept', ...], ...}\r\nindices = [0, 1, 2, 3, 4, 5, ...], check_same_num_examples = False, offset = 0\r\n\r\n def apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples=False, offset=0):\r\n \"\"\"Utility to apply the function on a selection of columns.\"\"\"\r\n nonlocal update_data\r\n fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n if offset == 0:\r\n effective_indices = indices\r\n else:\r\n effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n processed_inputs = (\r\n> function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n )\r\nE TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'valid_rel_labels'\r\n\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1939: TypeError\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Mac\r\n- Python version: 3.8.9\r\n- PyArrow version: pyarrow==5.0.0\r\n\r\n","comment_length":8,"text":"Datasets 1.12 dataset.filter TypeError: get_indices_from_mask_function() got an unexpected keyword argument\n## Describe the bug\r\nUpgrading to 1.12 caused `dataset.filter` call to fail with \r\n\r\n> get_indices_from_mask_function() got an unexpected keyword argument valid_rel_labels\r\n\r\n\r\n## Steps to reproduce the bug\r\n```pythondef \r\n\r\nfilter_good_rows(\r\n ex: Dict,\r\n valid_rel_labels: Set[str],\r\n valid_ner_labels: Set[str],\r\n tokenizer: PreTrainedTokenizerFast,\r\n) -> bool:\r\n \"\"\"Get the good rows\"\"\"\r\n encoding = get_encoding_for_text(text=ex[\"text\"], tokenizer=tokenizer)\r\n ex[\"encoding\"] = encoding\r\n for relation in ex[\"relations\"]:\r\n if not is_valid_relation(relation, valid_rel_labels):\r\n return False\r\n for span in ex[\"spans\"]:\r\n if not is_valid_span(span, valid_ner_labels, encoding):\r\n return False\r\n return True\r\n \r\ndef get_dataset(): \r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path,\r\n name=\"prodigy-dataset\",\r\n data_files=sorted(file_paths),\r\n cache_dir=cache_dir,\r\n )[\"train\"]\r\n\r\n valid_ner_labels = set(vocab.ner_category)\r\n valid_relations = set(vocab.relation_types.keys())\r\n ds = ds.filter(\r\n filter_good_rows,\r\n fn_kwargs=dict(\r\n valid_rel_labels=valid_relations,\r\n valid_ner_labels=valid_ner_labels,\r\n tokenizer=vocab.tokenizer,\r\n ),\r\n keep_in_memory=True,\r\n num_proc=num_proc,\r\n )\r\n\r\n```\r\n\r\n`ds` is a `DatasetDict` produced by a jsonl dataset.\r\nThis runs fine on 1.11 but fails on 1.12\r\n\r\n**Stack Trace**\r\n\r\n\r\n\r\n## Expected results\r\n\r\nI expect 1.12 datasets filter to filter the dataset without raising as it does on 1.11\r\n\r\n## Actual results\r\n```\r\ntf_ner_rel_lib\/dataset.py:695: in load_prodigy_arrow_datasets_from_jsonl\r\n ds = ds.filter(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2169: in filter\r\n indices = self.map(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1686: in map\r\n return self._map_single(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2048: in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\ninputs = {'_input_hash': [2108817714, 1477695082, -1021597032, 2130671338, -1260483858, -1203431639, ...], '_task_hash': [18070...ons', 'relations', 'relations', ...], 'answer': ['accept', 'accept', 'accept', 'accept', 'accept', 'accept', ...], ...}\r\nindices = [0, 1, 2, 3, 4, 5, ...], check_same_num_examples = False, offset = 0\r\n\r\n def apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples=False, offset=0):\r\n \"\"\"Utility to apply the function on a selection of columns.\"\"\"\r\n nonlocal update_data\r\n fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n if offset == 0:\r\n effective_indices = indices\r\n else:\r\n effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n processed_inputs = (\r\n> function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n )\r\nE TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'valid_rel_labels'\r\n\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1939: TypeError\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Mac\r\n- Python version: 3.8.9\r\n- PyArrow version: pyarrow==5.0.0\r\n\r\n\nThanks for reporting, I'm looking into it :)","embeddings":[-0.2829699218,0.2777142525,-0.0049711736,0.1934766024,0.0521399193,-0.0259464737,0.2624210715,0.4762106538,-0.1806556284,0.0836368799,-0.1311300099,0.3645940423,-0.1310276091,0.1083575189,-0.1563561261,-0.2259478569,0.0624334253,0.0486550108,-0.0484171957,0.0325096361,-0.3871050477,0.4132161736,-0.4211299717,0.1235269159,0.008685492,-0.1654407978,0.1102141142,-0.0451107584,0.0099631641,-0.4235654175,0.4564154148,0.0699703991,0.0381553695,0.2566260993,-0.0001171321,0.2396819293,0.366684258,-0.0842551515,-0.2250449806,-0.405336529,-0.1809279323,-0.0339020006,0.1180416197,-0.1935501844,-0.0818986967,0.0174917243,-0.249624908,-0.2356579155,0.2611079514,0.4325256944,0.1972838491,0.0717553645,-0.0651654452,0.0339874998,0.0209346432,0.2666781545,0.0185224004,-0.0007021982,0.0587287322,-0.1879914552,0.2919843793,0.5048494339,-0.2368306518,-0.2743239999,0.2988181114,-0.2294282466,0.223627314,-0.3586401939,0.2479036301,-0.0648386478,0.0247386377,-0.0833451524,-0.3223271072,-0.1697276235,-0.2576664686,-0.1120616794,0.3498052657,-0.1584596336,-0.244265154,-0.0208345316,-0.07258939,-0.0044453437,-0.0193625409,0.0009967602,-0.0110017173,0.5335709453,0.051298894,0.0513096936,0.0644063577,-0.1729889512,0.3066953421,-0.1076792255,-0.1007701308,0.0617668256,-0.170351848,0.0725231543,0.2387878448,-0.100716427,0.0953230187,-0.0929184034,-0.0909817293,-0.1606885344,0.0432230234,-0.0053427084,-0.0694495961,0.0904969722,0.3082866073,0.6318762302,0.0672923923,0.058587458,-0.0616019517,0.0373482928,-0.0557346605,0.0807316527,0.0906984136,0.2471883744,0.3795975745,-0.3135053217,-0.6206545234,0.1616326869,-0.5770969987,-0.1607390344,0.237689808,0.0602699555,0.2683047056,0.0857219398,-0.0559860393,0.0184919778,-0.0884351358,0.0967272371,-0.1329723448,0.0300575532,-0.0661212951,-0.288934499,0.1902687103,-0.5589625239,0.0788640752,-0.1231013834,-0.090868406,-0.0088757947,-0.0592832975,0.0524173155,0.1024440303,0.3941213489,-0.4122775495,0.1066873968,0.2004962116,-0.5756821632,-0.1944885254,0.2767592072,-0.4454355538,-0.0839394405,0.0214587208,0.2041888237,0.1366723627,-0.1789838076,-0.3426181376,0.5803837776,0.0123843485,-0.2307372093,-0.0704242811,-0.2239028215,-0.3683023453,-0.0239756647,0.1340631545,0.1689580381,-0.8365051746,-0.2499466091,0.2335013896,-0.0030520104,-0.0626956224,-0.2084405422,-0.1818583906,-0.0219790861,-0.1172454953,-0.0013353637,0.3576940596,-0.485848546,-0.6574547887,0.1390666068,0.1225496158,0.1945742369,0.1314316094,-0.1025109664,0.3647884429,-0.0999429971,0.1756555587,0.2742819488,0.0124631217,-0.0347300246,-0.2012303621,0.0749125555,0.2941287458,-0.0680078194,0.1666537374,0.27637887,-0.1620120853,-0.0194072127,0.2861098051,0.1138249561,-0.0791773424,0.0638790876,0.2018899024,-0.1175088063,0.232898742,-0.4075967968,-0.1830836833,-0.0563953593,0.0861253291,0.130115062,-0.3378736973,-0.214805305,-0.3373113871,0.0745304599,-0.2046329081,-0.1652506888,0.1793425977,0.1503546089,-0.0387527496,-0.0507269315,-0.0374752656,0.4382365942,0.2428901494,0.1432452947,-0.2216559798,0.2244707495,-0.0795594603,-0.080613561,0.0983143821,-0.0522183664,0.36705634,-0.0016741207,-0.3518117964,0.2039128393,0.0591471232,-0.3466408551,-0.0529597551,0.0107605178,0.0352687202,0.0651153848,0.1794721931,0.3757503927,-0.0078226523,0.1407691687,-0.1973981559,0.4055749774,-0.0408483408,0.4687744379,0.0199777223,0.0771241337,0.3994879127,0.2407884896,-0.1524210721,-0.1703110337,0.0525777675,0.3843008578,0.2716696858,-0.1308127642,-0.0575984567,0.2184957862,0.3087431192,-0.0334121324,0.0245152265,0.3014428616,0.015626844,0.0211441629,0.1812755615,0.4510476291,0.150378406,0.1575035751,-0.1162214428,0.0736183971,-0.2908762395,-0.0003076858,0.2586247027,0.1770678163,-0.2355194986,-0.0977184847,0.2307223231,0.0618963651,-0.3269087672,-0.4945808649,-0.1177324802,0.1571106911,-0.2101406008,0.4223466814,-0.2563110888,-0.360643059,0.2112966329,-0.1757659763,-0.1615487486,-0.4771515429,0.0838324726,0.1105318815,0.0211028233,0.4142694175,-0.1944871247,0.2770675421,0.1981321126,-0.2057120055,-0.2324998677,-0.3686217666,-0.3903269172,0.0790402368,0.1200828031,0.0579170287,0.1671684682,-0.0545275025,0.1354142576,0.083329387,-0.7349098921,0.1015762314,-0.3447894454,0.4745761752,0.1016463861,0.038837377,-0.3328078389,-0.2596282661,0.1520894766,0.304394573,-0.2402678132,0.1635410488,0.0431459434,0.2904618084,-0.2289169878,-0.4294990301,-0.2123535424,-0.1180267707,-0.1469037235,-0.1846597344,0.3282931149,0.5408180952,0.0719795302,-0.0907747671,0.1236420944,0.184421882,-0.3507416248,-0.4034056962,0.144136861,-0.0788861513,-0.1249593124,-0.1968769431,-0.2286561579,0.0848402604,0.2183059156,-0.5605329275,-0.4192049503,0.0349870473,0.0410459861,0.1570181549,-0.1139656156,0.1058698297,0.2774389684,0.030448081,-0.2724855542,-0.3693442345,0.0357308015,-0.2114411592,0.0473164991,0.0135325789,0.9601014853,0.0164168794,0.0682780072,0.1694822013,0.2156193256,0.3308912516,0.0938933268,0.4396736324,-0.2999125719,-0.2751029134,-0.1022948325,0.0323724411,-0.0618773438,-0.2904603481,0.0270541478,-0.0472293831,-0.1075963899,-0.1457635909,-0.2428792566,-0.307492286,-0.1465963423,-0.1706662029,0.0713224858,0.0963586718,0.3636464477,-0.308096081,-0.0618637502,0.0288702194,0.2998217344,-0.0606535003,0.0504504703,-0.4331361651,-0.0584681481,-0.3299600482,0.3512754738,0.0836806521,0.5775329471,0.0337941647,0.1314383298,-0.1877772957,-0.0064897859,0.6513140798,-0.37467134,0.0990580469,0.3981027007,-0.0517896451,-0.105050236,-0.0820436478,-0.3614759147,0.1523979008,-0.130425632,0.3490183353,-0.2240297794,-0.0515188314,0.3174758554,0.0232320037,-0.1624078602,0.1051052287,-0.3738526404,-0.5830719471,-0.0338545628,0.2991501689,0.0823923871,0.4687190056,0.0837701112,0.0468495339,-0.0865380093,-0.4693993032,0.1605705619,0.2456996441,0.1313659102,0.1721284539,0.1322498471,0.1905105859,0.1852574944,0.2601747811,0.3494502306,-0.1209676489,-0.429931134,-0.1555662006,-0.003165609,0.2964791059,0.0886024907,-0.2174897939,-0.0622401759,-0.1156746596,-0.1378526092,-0.0112768495,0.1836106479,0.2737488151,0.2099698633,-0.4123848975,-0.5313845873,0.1796308607,0.2255197465,-0.1604507267,0.3734816909,0.2652886808,-0.4143254757,0.4304826856,-0.1537723541,0.7024047375,0.2602522671,-0.1502133161,0.3903979361,-0.1639682204,0.3863707483,0.1703226268,0.1472315341,-0.0537542589,-0.3721880913,-0.0091732787,0.0341162309,-0.1155766919,0.0897843763,-0.2464871407,0.3867236376,-0.2271939069,-0.1297411919,0.1523538828,0.0656017885,0.072543852,0.1573487818,-0.436416477,0.1157073081,-0.1356859058,-0.1174689382,-0.0174163952,-0.157321766,0.2357757241,-0.4989627004,-0.2169850469,0.0593276396,-0.2186802477,0.1991744936,0.1808251143,-0.5248704553,-0.0648283362,-0.2381993383,0.1592858881,0.1625673175,0.0071186177,0.1369998753,-0.1568153501,0.4103268683,0.1870173216,-0.0307340231,0.4770958126,0.1361288428,-0.0680285394,0.1569158882,0.016583018,-0.282654196,-0.3184965551,-0.1825746,-0.0308478139,-0.188172102,0.1618311256,-0.1791258454,-0.2661722004,-0.3849771619,0.1085400954,-0.1376332939,0.0371233523,0.1283319592,0.1249253303,-0.1639446914,0.0098909764,0.2198198438,0.0911266208,-0.3888247311,0.3381742537,-0.1077874154,-0.1485274583,-0.1235791966,0.1685680747,0.2814444602,0.0629519448,0.3650933504,-0.0872263014,-0.2079828233,0.1700563133,0.2281175256,-0.1518964171,0.2485146672,-0.1161972806,-0.3701640368,-0.2486976683,0.4615723789,-0.1306284964,0.0186716523,-0.0521034189,0.0991068631,0.003739581,-0.0162424091,-0.2511475086,0.0053935279,0.0735629424,0.1900277734,-0.4759955704,0.1512910128,0.2287879288,0.110932909,0.0573695824,0.2607388794,0.0007869502,-0.0999411941,-0.2160836458,0.1331126839,0.1301591247,-0.1403605193,-0.1228807792,0.0188462827,-0.0731661841,-0.0717758238,0.216513291,0.0835128203,-0.0267807152,0.2172831297,0.1311217993,0.1191414148,0.3056656718,0.0038306452,-0.0522037186,0.124015674,0.1425275654,0.1671767533,-0.0698422417,-0.0610502064,-0.5958340168,0.0865877271,-0.1721179634,0.334825635,0.3047714829,-0.2367942035,0.0305576939,0.0753731653,0.3694211245,0.314019382,0.0147067439,-0.0130339265,0.0858267918,0.1819656938,-0.3525706828,-0.1376184076,0.3225528002,0.0614038594,0.1956574023,0.1185259968,-0.3290822804,0.2033440918,0.2233221978,-0.120464392,0.4577661753,0.0393730588,0.2156214267,0.5496623516,0.0822913945,0.1059112772,0.4552771747,-0.1346557885,-0.0528844185,0.7670977712,0.0000362798,0.4767209888,0.5375198126,0.1209944114,0.1651364565,-0.116949372,-0.061264392,0.0139843347,-0.0742452666,0.2528257966,-0.1285436898,-0.2894105911,-0.2081596702,-0.0167670101,-0.1109244302,-0.0039245086,-0.1681597084,-0.2484375685,-0.1625422835,-0.2450120449,-0.0422880724,-0.0708364993,-0.160419032,-0.0501407646,0.1269110143,-0.3143103123,-0.0143398643,-0.1467229724,-0.2234831899,-0.0081214029,0.2016516626,-0.0929878354,0.3207439482,0.2998452187,0.159827143,0.3399278522,0.2196566463,0.3316202462,0.443985045,-0.0927284956,-0.1078115478,-0.1613139808,-0.0841439664,-0.1108931452,0.2211236358,-0.0047986256,0.1254571676,0.1096822023,0.1607336253,-0.0612043366,0.3804636002,0.3565916419,0.0261916313,-0.3428668082,0.0200863183,-0.0340408646,0.0160514023,-0.0020940849,-0.0784246698,-0.0881817192,-0.2545608282,0.4558415115,0.3104175329,0.1873828918,-0.1252919883,0.0795623884,-0.0890269503,0.3357637227,0.2054750472,0.0817027614,-0.1674567759,0.0917910859,-0.4356390238,-0.0776815191,0.0969927013,-0.0606095158,0.2765798271,-0.0421167389,0.0987063646,0.2266128808,0.0339960195,0.1146242544,-0.1884645075,-0.1508150697,-0.5087311268,-0.089195773,0.3614899218,0.1643898338,0.071918197,-0.2774338424,0.0469553731,-0.0643579736,0.0770169273,-0.192634359,-0.1640274078,0.1316321939,0.5628841519,0.121633105,0.3136529624,0.4868503809,0.0407279916,-0.0813758746,-0.3092046678,-0.3655738533,-0.2650110424,0.2840080559,-0.0509732217,0.5357193947,-0.1638861448,-0.3490848541,-0.0664884001,0.264821738,-0.157160446,-0.188663289,-0.4559234977,0.2295293361,0.0384258516,0.0980730504,-0.4108909369,0.3706010282,0.111948207,0.4879986346,-0.2702848613,-0.6187790036,0.4536240399,-0.2140198201,-0.4248342216,-0.2510511577,0.067547515,0.1526652873,-0.1547737271,-0.4326290786,-0.0599381812,0.4278156459,-0.0611086115,-0.2852191627,0.0017881198,0.2702943087,0.3620921075,-0.1158800274,-0.1328708529,0.1866863668,-0.0891441852,0.1879659444,-0.1533495486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2927","title":"Datasets 1.12 dataset.filter TypeError: get_indices_from_mask_function() got an unexpected keyword argument","comments":"Fixed by #2950.","body":"## Describe the bug\r\nUpgrading to 1.12 caused `dataset.filter` call to fail with \r\n\r\n> get_indices_from_mask_function() got an unexpected keyword argument valid_rel_labels\r\n\r\n\r\n## Steps to reproduce the bug\r\n```pythondef \r\n\r\nfilter_good_rows(\r\n ex: Dict,\r\n valid_rel_labels: Set[str],\r\n valid_ner_labels: Set[str],\r\n tokenizer: PreTrainedTokenizerFast,\r\n) -> bool:\r\n \"\"\"Get the good rows\"\"\"\r\n encoding = get_encoding_for_text(text=ex[\"text\"], tokenizer=tokenizer)\r\n ex[\"encoding\"] = encoding\r\n for relation in ex[\"relations\"]:\r\n if not is_valid_relation(relation, valid_rel_labels):\r\n return False\r\n for span in ex[\"spans\"]:\r\n if not is_valid_span(span, valid_ner_labels, encoding):\r\n return False\r\n return True\r\n \r\ndef get_dataset(): \r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path,\r\n name=\"prodigy-dataset\",\r\n data_files=sorted(file_paths),\r\n cache_dir=cache_dir,\r\n )[\"train\"]\r\n\r\n valid_ner_labels = set(vocab.ner_category)\r\n valid_relations = set(vocab.relation_types.keys())\r\n ds = ds.filter(\r\n filter_good_rows,\r\n fn_kwargs=dict(\r\n valid_rel_labels=valid_relations,\r\n valid_ner_labels=valid_ner_labels,\r\n tokenizer=vocab.tokenizer,\r\n ),\r\n keep_in_memory=True,\r\n num_proc=num_proc,\r\n )\r\n\r\n```\r\n\r\n`ds` is a `DatasetDict` produced by a jsonl dataset.\r\nThis runs fine on 1.11 but fails on 1.12\r\n\r\n**Stack Trace**\r\n\r\n\r\n\r\n## Expected results\r\n\r\nI expect 1.12 datasets filter to filter the dataset without raising as it does on 1.11\r\n\r\n## Actual results\r\n```\r\ntf_ner_rel_lib\/dataset.py:695: in load_prodigy_arrow_datasets_from_jsonl\r\n ds = ds.filter(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2169: in filter\r\n indices = self.map(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1686: in map\r\n return self._map_single(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2048: in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\ninputs = {'_input_hash': [2108817714, 1477695082, -1021597032, 2130671338, -1260483858, -1203431639, ...], '_task_hash': [18070...ons', 'relations', 'relations', ...], 'answer': ['accept', 'accept', 'accept', 'accept', 'accept', 'accept', ...], ...}\r\nindices = [0, 1, 2, 3, 4, 5, ...], check_same_num_examples = False, offset = 0\r\n\r\n def apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples=False, offset=0):\r\n \"\"\"Utility to apply the function on a selection of columns.\"\"\"\r\n nonlocal update_data\r\n fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n if offset == 0:\r\n effective_indices = indices\r\n else:\r\n effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n processed_inputs = (\r\n> function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n )\r\nE TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'valid_rel_labels'\r\n\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1939: TypeError\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Mac\r\n- Python version: 3.8.9\r\n- PyArrow version: pyarrow==5.0.0\r\n\r\n","comment_length":3,"text":"Datasets 1.12 dataset.filter TypeError: get_indices_from_mask_function() got an unexpected keyword argument\n## Describe the bug\r\nUpgrading to 1.12 caused `dataset.filter` call to fail with \r\n\r\n> get_indices_from_mask_function() got an unexpected keyword argument valid_rel_labels\r\n\r\n\r\n## Steps to reproduce the bug\r\n```pythondef \r\n\r\nfilter_good_rows(\r\n ex: Dict,\r\n valid_rel_labels: Set[str],\r\n valid_ner_labels: Set[str],\r\n tokenizer: PreTrainedTokenizerFast,\r\n) -> bool:\r\n \"\"\"Get the good rows\"\"\"\r\n encoding = get_encoding_for_text(text=ex[\"text\"], tokenizer=tokenizer)\r\n ex[\"encoding\"] = encoding\r\n for relation in ex[\"relations\"]:\r\n if not is_valid_relation(relation, valid_rel_labels):\r\n return False\r\n for span in ex[\"spans\"]:\r\n if not is_valid_span(span, valid_ner_labels, encoding):\r\n return False\r\n return True\r\n \r\ndef get_dataset(): \r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path,\r\n name=\"prodigy-dataset\",\r\n data_files=sorted(file_paths),\r\n cache_dir=cache_dir,\r\n )[\"train\"]\r\n\r\n valid_ner_labels = set(vocab.ner_category)\r\n valid_relations = set(vocab.relation_types.keys())\r\n ds = ds.filter(\r\n filter_good_rows,\r\n fn_kwargs=dict(\r\n valid_rel_labels=valid_relations,\r\n valid_ner_labels=valid_ner_labels,\r\n tokenizer=vocab.tokenizer,\r\n ),\r\n keep_in_memory=True,\r\n num_proc=num_proc,\r\n )\r\n\r\n```\r\n\r\n`ds` is a `DatasetDict` produced by a jsonl dataset.\r\nThis runs fine on 1.11 but fails on 1.12\r\n\r\n**Stack Trace**\r\n\r\n\r\n\r\n## Expected results\r\n\r\nI expect 1.12 datasets filter to filter the dataset without raising as it does on 1.11\r\n\r\n## Actual results\r\n```\r\ntf_ner_rel_lib\/dataset.py:695: in load_prodigy_arrow_datasets_from_jsonl\r\n ds = ds.filter(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2169: in filter\r\n indices = self.map(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1686: in map\r\n return self._map_single(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2048: in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\ninputs = {'_input_hash': [2108817714, 1477695082, -1021597032, 2130671338, -1260483858, -1203431639, ...], '_task_hash': [18070...ons', 'relations', 'relations', ...], 'answer': ['accept', 'accept', 'accept', 'accept', 'accept', 'accept', ...], ...}\r\nindices = [0, 1, 2, 3, 4, 5, ...], check_same_num_examples = False, offset = 0\r\n\r\n def apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples=False, offset=0):\r\n \"\"\"Utility to apply the function on a selection of columns.\"\"\"\r\n nonlocal update_data\r\n fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n if offset == 0:\r\n effective_indices = indices\r\n else:\r\n effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n processed_inputs = (\r\n> function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n )\r\nE TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'valid_rel_labels'\r\n\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1939: TypeError\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Mac\r\n- Python version: 3.8.9\r\n- PyArrow version: pyarrow==5.0.0\r\n\r\n\nFixed by #2950.","embeddings":[-0.2829699218,0.2777142525,-0.0049711736,0.1934766024,0.0521399193,-0.0259464737,0.2624210715,0.4762106538,-0.1806556284,0.0836368799,-0.1311300099,0.3645940423,-0.1310276091,0.1083575189,-0.1563561261,-0.2259478569,0.0624334253,0.0486550108,-0.0484171957,0.0325096361,-0.3871050477,0.4132161736,-0.4211299717,0.1235269159,0.008685492,-0.1654407978,0.1102141142,-0.0451107584,0.0099631641,-0.4235654175,0.4564154148,0.0699703991,0.0381553695,0.2566260993,-0.0001171321,0.2396819293,0.366684258,-0.0842551515,-0.2250449806,-0.405336529,-0.1809279323,-0.0339020006,0.1180416197,-0.1935501844,-0.0818986967,0.0174917243,-0.249624908,-0.2356579155,0.2611079514,0.4325256944,0.1972838491,0.0717553645,-0.0651654452,0.0339874998,0.0209346432,0.2666781545,0.0185224004,-0.0007021982,0.0587287322,-0.1879914552,0.2919843793,0.5048494339,-0.2368306518,-0.2743239999,0.2988181114,-0.2294282466,0.223627314,-0.3586401939,0.2479036301,-0.0648386478,0.0247386377,-0.0833451524,-0.3223271072,-0.1697276235,-0.2576664686,-0.1120616794,0.3498052657,-0.1584596336,-0.244265154,-0.0208345316,-0.07258939,-0.0044453437,-0.0193625409,0.0009967602,-0.0110017173,0.5335709453,0.051298894,0.0513096936,0.0644063577,-0.1729889512,0.3066953421,-0.1076792255,-0.1007701308,0.0617668256,-0.170351848,0.0725231543,0.2387878448,-0.100716427,0.0953230187,-0.0929184034,-0.0909817293,-0.1606885344,0.0432230234,-0.0053427084,-0.0694495961,0.0904969722,0.3082866073,0.6318762302,0.0672923923,0.058587458,-0.0616019517,0.0373482928,-0.0557346605,0.0807316527,0.0906984136,0.2471883744,0.3795975745,-0.3135053217,-0.6206545234,0.1616326869,-0.5770969987,-0.1607390344,0.237689808,0.0602699555,0.2683047056,0.0857219398,-0.0559860393,0.0184919778,-0.0884351358,0.0967272371,-0.1329723448,0.0300575532,-0.0661212951,-0.288934499,0.1902687103,-0.5589625239,0.0788640752,-0.1231013834,-0.090868406,-0.0088757947,-0.0592832975,0.0524173155,0.1024440303,0.3941213489,-0.4122775495,0.1066873968,0.2004962116,-0.5756821632,-0.1944885254,0.2767592072,-0.4454355538,-0.0839394405,0.0214587208,0.2041888237,0.1366723627,-0.1789838076,-0.3426181376,0.5803837776,0.0123843485,-0.2307372093,-0.0704242811,-0.2239028215,-0.3683023453,-0.0239756647,0.1340631545,0.1689580381,-0.8365051746,-0.2499466091,0.2335013896,-0.0030520104,-0.0626956224,-0.2084405422,-0.1818583906,-0.0219790861,-0.1172454953,-0.0013353637,0.3576940596,-0.485848546,-0.6574547887,0.1390666068,0.1225496158,0.1945742369,0.1314316094,-0.1025109664,0.3647884429,-0.0999429971,0.1756555587,0.2742819488,0.0124631217,-0.0347300246,-0.2012303621,0.0749125555,0.2941287458,-0.0680078194,0.1666537374,0.27637887,-0.1620120853,-0.0194072127,0.2861098051,0.1138249561,-0.0791773424,0.0638790876,0.2018899024,-0.1175088063,0.232898742,-0.4075967968,-0.1830836833,-0.0563953593,0.0861253291,0.130115062,-0.3378736973,-0.214805305,-0.3373113871,0.0745304599,-0.2046329081,-0.1652506888,0.1793425977,0.1503546089,-0.0387527496,-0.0507269315,-0.0374752656,0.4382365942,0.2428901494,0.1432452947,-0.2216559798,0.2244707495,-0.0795594603,-0.080613561,0.0983143821,-0.0522183664,0.36705634,-0.0016741207,-0.3518117964,0.2039128393,0.0591471232,-0.3466408551,-0.0529597551,0.0107605178,0.0352687202,0.0651153848,0.1794721931,0.3757503927,-0.0078226523,0.1407691687,-0.1973981559,0.4055749774,-0.0408483408,0.4687744379,0.0199777223,0.0771241337,0.3994879127,0.2407884896,-0.1524210721,-0.1703110337,0.0525777675,0.3843008578,0.2716696858,-0.1308127642,-0.0575984567,0.2184957862,0.3087431192,-0.0334121324,0.0245152265,0.3014428616,0.015626844,0.0211441629,0.1812755615,0.4510476291,0.150378406,0.1575035751,-0.1162214428,0.0736183971,-0.2908762395,-0.0003076858,0.2586247027,0.1770678163,-0.2355194986,-0.0977184847,0.2307223231,0.0618963651,-0.3269087672,-0.4945808649,-0.1177324802,0.1571106911,-0.2101406008,0.4223466814,-0.2563110888,-0.360643059,0.2112966329,-0.1757659763,-0.1615487486,-0.4771515429,0.0838324726,0.1105318815,0.0211028233,0.4142694175,-0.1944871247,0.2770675421,0.1981321126,-0.2057120055,-0.2324998677,-0.3686217666,-0.3903269172,0.0790402368,0.1200828031,0.0579170287,0.1671684682,-0.0545275025,0.1354142576,0.083329387,-0.7349098921,0.1015762314,-0.3447894454,0.4745761752,0.1016463861,0.038837377,-0.3328078389,-0.2596282661,0.1520894766,0.304394573,-0.2402678132,0.1635410488,0.0431459434,0.2904618084,-0.2289169878,-0.4294990301,-0.2123535424,-0.1180267707,-0.1469037235,-0.1846597344,0.3282931149,0.5408180952,0.0719795302,-0.0907747671,0.1236420944,0.184421882,-0.3507416248,-0.4034056962,0.144136861,-0.0788861513,-0.1249593124,-0.1968769431,-0.2286561579,0.0848402604,0.2183059156,-0.5605329275,-0.4192049503,0.0349870473,0.0410459861,0.1570181549,-0.1139656156,0.1058698297,0.2774389684,0.030448081,-0.2724855542,-0.3693442345,0.0357308015,-0.2114411592,0.0473164991,0.0135325789,0.9601014853,0.0164168794,0.0682780072,0.1694822013,0.2156193256,0.3308912516,0.0938933268,0.4396736324,-0.2999125719,-0.2751029134,-0.1022948325,0.0323724411,-0.0618773438,-0.2904603481,0.0270541478,-0.0472293831,-0.1075963899,-0.1457635909,-0.2428792566,-0.307492286,-0.1465963423,-0.1706662029,0.0713224858,0.0963586718,0.3636464477,-0.308096081,-0.0618637502,0.0288702194,0.2998217344,-0.0606535003,0.0504504703,-0.4331361651,-0.0584681481,-0.3299600482,0.3512754738,0.0836806521,0.5775329471,0.0337941647,0.1314383298,-0.1877772957,-0.0064897859,0.6513140798,-0.37467134,0.0990580469,0.3981027007,-0.0517896451,-0.105050236,-0.0820436478,-0.3614759147,0.1523979008,-0.130425632,0.3490183353,-0.2240297794,-0.0515188314,0.3174758554,0.0232320037,-0.1624078602,0.1051052287,-0.3738526404,-0.5830719471,-0.0338545628,0.2991501689,0.0823923871,0.4687190056,0.0837701112,0.0468495339,-0.0865380093,-0.4693993032,0.1605705619,0.2456996441,0.1313659102,0.1721284539,0.1322498471,0.1905105859,0.1852574944,0.2601747811,0.3494502306,-0.1209676489,-0.429931134,-0.1555662006,-0.003165609,0.2964791059,0.0886024907,-0.2174897939,-0.0622401759,-0.1156746596,-0.1378526092,-0.0112768495,0.1836106479,0.2737488151,0.2099698633,-0.4123848975,-0.5313845873,0.1796308607,0.2255197465,-0.1604507267,0.3734816909,0.2652886808,-0.4143254757,0.4304826856,-0.1537723541,0.7024047375,0.2602522671,-0.1502133161,0.3903979361,-0.1639682204,0.3863707483,0.1703226268,0.1472315341,-0.0537542589,-0.3721880913,-0.0091732787,0.0341162309,-0.1155766919,0.0897843763,-0.2464871407,0.3867236376,-0.2271939069,-0.1297411919,0.1523538828,0.0656017885,0.072543852,0.1573487818,-0.436416477,0.1157073081,-0.1356859058,-0.1174689382,-0.0174163952,-0.157321766,0.2357757241,-0.4989627004,-0.2169850469,0.0593276396,-0.2186802477,0.1991744936,0.1808251143,-0.5248704553,-0.0648283362,-0.2381993383,0.1592858881,0.1625673175,0.0071186177,0.1369998753,-0.1568153501,0.4103268683,0.1870173216,-0.0307340231,0.4770958126,0.1361288428,-0.0680285394,0.1569158882,0.016583018,-0.282654196,-0.3184965551,-0.1825746,-0.0308478139,-0.188172102,0.1618311256,-0.1791258454,-0.2661722004,-0.3849771619,0.1085400954,-0.1376332939,0.0371233523,0.1283319592,0.1249253303,-0.1639446914,0.0098909764,0.2198198438,0.0911266208,-0.3888247311,0.3381742537,-0.1077874154,-0.1485274583,-0.1235791966,0.1685680747,0.2814444602,0.0629519448,0.3650933504,-0.0872263014,-0.2079828233,0.1700563133,0.2281175256,-0.1518964171,0.2485146672,-0.1161972806,-0.3701640368,-0.2486976683,0.4615723789,-0.1306284964,0.0186716523,-0.0521034189,0.0991068631,0.003739581,-0.0162424091,-0.2511475086,0.0053935279,0.0735629424,0.1900277734,-0.4759955704,0.1512910128,0.2287879288,0.110932909,0.0573695824,0.2607388794,0.0007869502,-0.0999411941,-0.2160836458,0.1331126839,0.1301591247,-0.1403605193,-0.1228807792,0.0188462827,-0.0731661841,-0.0717758238,0.216513291,0.0835128203,-0.0267807152,0.2172831297,0.1311217993,0.1191414148,0.3056656718,0.0038306452,-0.0522037186,0.124015674,0.1425275654,0.1671767533,-0.0698422417,-0.0610502064,-0.5958340168,0.0865877271,-0.1721179634,0.334825635,0.3047714829,-0.2367942035,0.0305576939,0.0753731653,0.3694211245,0.314019382,0.0147067439,-0.0130339265,0.0858267918,0.1819656938,-0.3525706828,-0.1376184076,0.3225528002,0.0614038594,0.1956574023,0.1185259968,-0.3290822804,0.2033440918,0.2233221978,-0.120464392,0.4577661753,0.0393730588,0.2156214267,0.5496623516,0.0822913945,0.1059112772,0.4552771747,-0.1346557885,-0.0528844185,0.7670977712,0.0000362798,0.4767209888,0.5375198126,0.1209944114,0.1651364565,-0.116949372,-0.061264392,0.0139843347,-0.0742452666,0.2528257966,-0.1285436898,-0.2894105911,-0.2081596702,-0.0167670101,-0.1109244302,-0.0039245086,-0.1681597084,-0.2484375685,-0.1625422835,-0.2450120449,-0.0422880724,-0.0708364993,-0.160419032,-0.0501407646,0.1269110143,-0.3143103123,-0.0143398643,-0.1467229724,-0.2234831899,-0.0081214029,0.2016516626,-0.0929878354,0.3207439482,0.2998452187,0.159827143,0.3399278522,0.2196566463,0.3316202462,0.443985045,-0.0927284956,-0.1078115478,-0.1613139808,-0.0841439664,-0.1108931452,0.2211236358,-0.0047986256,0.1254571676,0.1096822023,0.1607336253,-0.0612043366,0.3804636002,0.3565916419,0.0261916313,-0.3428668082,0.0200863183,-0.0340408646,0.0160514023,-0.0020940849,-0.0784246698,-0.0881817192,-0.2545608282,0.4558415115,0.3104175329,0.1873828918,-0.1252919883,0.0795623884,-0.0890269503,0.3357637227,0.2054750472,0.0817027614,-0.1674567759,0.0917910859,-0.4356390238,-0.0776815191,0.0969927013,-0.0606095158,0.2765798271,-0.0421167389,0.0987063646,0.2266128808,0.0339960195,0.1146242544,-0.1884645075,-0.1508150697,-0.5087311268,-0.089195773,0.3614899218,0.1643898338,0.071918197,-0.2774338424,0.0469553731,-0.0643579736,0.0770169273,-0.192634359,-0.1640274078,0.1316321939,0.5628841519,0.121633105,0.3136529624,0.4868503809,0.0407279916,-0.0813758746,-0.3092046678,-0.3655738533,-0.2650110424,0.2840080559,-0.0509732217,0.5357193947,-0.1638861448,-0.3490848541,-0.0664884001,0.264821738,-0.157160446,-0.188663289,-0.4559234977,0.2295293361,0.0384258516,0.0980730504,-0.4108909369,0.3706010282,0.111948207,0.4879986346,-0.2702848613,-0.6187790036,0.4536240399,-0.2140198201,-0.4248342216,-0.2510511577,0.067547515,0.1526652873,-0.1547737271,-0.4326290786,-0.0599381812,0.4278156459,-0.0611086115,-0.2852191627,0.0017881198,0.2702943087,0.3620921075,-0.1158800274,-0.1328708529,0.1866863668,-0.0891441852,0.1879659444,-0.1533495486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2924","title":"\"File name too long\" error for file locks","comments":"Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock\")\r\n154\r\n```\r\nso not sure why it's considered too long for your filesystem.\r\n(also note that the lock files we use always have smaller filenames than 255)\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/5d1a9f1e3c6c495dc0610b459e39d2eb8893f152\/src\/datasets\/utils\/filelock.py#L135-L135","body":"## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t\/test\")`:\r\n\r\n```\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nWhere the user cache dir (e.g. `~\/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"gar1t\/test\")\r\n```\r\n\r\n## Expected results\r\n\r\nExpect the function to return without an error.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 765, in _save_info\r\n with FileLock(lock_path):\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 403, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.7\r\n- PyArrow version: 5.0.0\r\n","comment_length":39,"text":"\"File name too long\" error for file locks\n## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t\/test\")`:\r\n\r\n```\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nWhere the user cache dir (e.g. `~\/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"gar1t\/test\")\r\n```\r\n\r\n## Expected results\r\n\r\nExpect the function to return without an error.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 765, in _save_info\r\n with FileLock(lock_path):\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 403, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.7\r\n- PyArrow version: 5.0.0\r\n\nHi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock\")\r\n154\r\n```\r\nso not sure why it's considered too long for your filesystem.\r\n(also note that the lock files we use always have smaller filenames than 255)\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/5d1a9f1e3c6c495dc0610b459e39d2eb8893f152\/src\/datasets\/utils\/filelock.py#L135-L135","embeddings":[0.0493393503,0.0920015797,-0.0714939609,0.401848346,0.4151671231,0.2358950824,0.636662364,0.2435030937,0.2364584506,0.2313297093,0.0185690634,0.0077435691,-0.1442685276,-0.3100371361,-0.2015356272,-0.2443143427,-0.1581087261,-0.0416213386,-0.177078411,0.2177188098,-0.1549794227,0.3787363768,0.0246380791,0.1570838541,-0.536006391,0.0900789052,-0.1295958906,0.3886749148,-0.0215065964,-0.3729913831,0.051118616,-0.0467165932,0.068814449,0.811357677,-0.0001220686,-0.3334413767,0.3542338908,-0.0562992319,-0.3271892667,-0.2243083715,-0.1685816944,-0.5117114782,-0.0345845483,-0.4988024831,0.1872130185,-0.0951189175,-0.0331902057,-0.8021566272,0.0410523824,0.3733062148,0.1268745214,-0.0783471763,0.0910248384,-0.2579201758,0.3323119283,-0.2779891491,0.0040849657,0.3564026058,0.3556975722,-0.0988676101,-0.1546510309,0.2032182515,-0.0368982404,0.027324792,0.2471582741,0.079537563,-0.0874831453,-0.3261650205,0.2958056033,0.4273941517,0.4714379907,-0.0930620804,-0.2793129385,-0.4676968753,0.1424980164,-0.1251056492,0.4885306358,-0.0762403011,-0.222074613,0.1514885277,-0.1704189926,-0.0196209587,-0.1235974729,-0.1102384105,-0.101975739,0.0650619194,0.0405084826,0.0362790599,0.3111130893,-0.3483147621,0.2395959646,0.0048368936,0.0742740706,0.2694604397,-0.6542128921,0.1469123662,0.0842684656,0.3854928017,0.2534896433,0.0576435812,-0.2659727335,-0.1483078748,0.1947487742,-0.0104660569,-0.0649549589,0.374206692,0.1944440156,0.1612814218,0.3012429774,0.0779469088,-0.3367469013,-0.1069349945,-0.0764871612,-0.5534629822,0.445138216,0.133591041,0.0741629824,-0.3437117934,0.1504577994,0.5705774426,0.1057047993,-0.0104902675,0.2287023664,0.2733797133,0.005434711,0.1107665151,0.0830746666,-0.0899362862,-0.0889612734,0.1008345857,-0.1473124325,-0.1107303202,-0.2270203382,0.1124099791,0.0892055854,-0.3083957136,0.2792796195,-0.227465108,0.3152402341,-0.1249576584,-0.0829687566,-0.2377426773,-0.0667323172,0.1559667885,-0.0591069087,0.0739138052,0.2452137023,-0.2607334852,-0.1920612901,-0.0771116391,-0.3397490382,-0.2851879001,-0.0617215782,0.0947027877,-0.165494591,0.2159024328,0.2415551245,-0.2671947181,0.6555529237,-0.0798197463,0.0127971917,-0.2747654617,-0.1418095827,0.0071923966,0.0408245735,0.5534662008,0.1325238049,-0.0175623372,-0.0687896535,0.1282823086,0.0086658485,0.3722593188,0.1806448698,-0.0535966232,-0.4219526947,0.3124936521,0.2268715054,-0.2402810752,-0.486985594,0.3215382397,-0.394307375,0.0994933695,0.3229117393,0.0228835605,0.0517107099,-0.024582373,0.2853384614,0.1822132915,-0.0260959864,0.0234954469,-0.202524215,-0.1073302031,-0.0406307913,0.1691952497,0.0109546073,0.0670013204,0.0796193257,-0.1751918048,0.2715687454,0.0314637274,-0.2040443867,0.3736798167,0.1382424235,0.3267191648,0.1860822737,-0.106327638,-0.5937504768,0.2637207508,-0.1218474954,-0.1787318438,-0.2175928354,-0.1868047863,-0.1749890149,-0.0072898441,0.0034599197,0.1351371109,-0.0204084311,0.2832494974,0.1417792886,-0.1317639649,0.0934399366,0.5883799791,-0.1702439636,0.0175973084,-0.4781179428,-0.17112647,-0.1178534552,-0.1125399023,0.0315071233,0.0016395918,0.3693964779,-0.1089679822,-0.3443510532,0.4480753541,0.3332768381,0.0975331739,-0.1789251119,-0.0044306694,-0.0568642691,0.220775336,0.0726701096,0.0398121551,0.0077854944,0.0142915584,-0.2135758698,0.3000163734,-0.14081572,0.27127707,0.0286138058,0.1682408899,0.2461784184,-0.1318140626,0.0643453151,-0.1761228889,0.6458842158,-0.2163528502,0.3048405647,0.0547799319,0.0452316478,-0.1448624283,0.6458194852,0.045392707,-0.0001772655,0.1977425367,0.1562013328,-0.0408157147,-0.0193730351,0.3064890206,0.5111442208,0.1920584738,-0.0301231258,-0.2065057307,0.3941576481,-0.2238344699,0.256319046,0.0159084182,-0.0027928618,0.3657026589,0.0496179983,-0.1290875971,0.0514840819,-0.8095089197,-0.050442148,0.223697722,-0.3198090196,0.1480822265,-0.1393256634,-0.0432045013,-0.0237995591,0.0004903108,-0.0803984031,-0.4162820876,-0.1888936311,0.2785063386,-0.1418365985,0.2145332694,-0.1188357845,-0.0032407197,0.2712830305,-0.0145387761,-0.2950273752,-0.0016123495,0.0803438574,-0.0837603286,0.2720724344,-0.5375033021,0.0838043541,-0.1256639212,-0.0631072521,-0.5083525181,-0.2498805523,0.1116677895,-0.1684809178,0.2935642302,0.3413662016,0.2078575194,0.1666625887,0.0588969849,0.0049610781,-0.2622792721,0.0236695651,0.2326634228,0.1779004186,-0.0228179917,-0.1917994022,-0.023455549,-0.3463737965,-0.4160333872,0.3207841218,0.0303027015,0.2279285491,0.3411892653,-0.0708731115,0.2978130579,-0.0179723371,0.2472920567,-0.0879067779,-0.446370095,0.4170611799,-0.1525222659,-0.2990590334,-0.4318080544,0.2980134785,-0.0355094597,0.1055189967,-0.4223643541,-0.1790713519,-0.401260376,0.3389849961,-0.3357198238,0.0054093283,0.1970204413,-0.0852006599,-0.0563488565,-0.1069383249,0.0849020928,0.2325119227,0.0110361269,-0.009520852,0.0929816738,0.2951667607,-0.1160778701,0.4532136321,0.3616654873,0.0753480569,0.1953798383,-0.1376693696,0.1843574643,-0.2590326071,-0.1269457787,0.0239379536,0.1744713485,-0.221584484,0.1342053711,-0.0853912458,0.0140818916,-0.1333202869,-0.0577918552,0.1078078598,-0.2960295677,0.0073167868,0.0485861041,0.0524898916,0.031080164,0.1456944346,-0.1948870718,-0.1309104413,-0.0240893289,0.3084784746,0.0540789627,-0.0233560894,-0.2468729615,0.0915106237,-0.3371418118,0.4440097511,0.1830917299,0.2137979567,-0.069359839,-0.1601748466,-0.0800488889,0.0460065827,0.6831169128,0.0991293713,0.3160929978,0.1554959416,0.2547036707,-0.1230872124,0.0391722023,0.0795527846,0.3701882362,0.1390409023,0.4079330266,-0.2308098525,0.037608888,-0.183316946,0.1855425686,-0.1166330427,-0.4492169619,-0.0262856334,0.0087962123,-0.2636858821,0.1616199315,0.0336052962,0.2442290634,0.0616719015,0.0049132649,-0.0601768605,-0.0747785419,0.1601556242,0.1748355776,0.2812633216,-0.4468779564,0.264365375,-0.027559543,-0.1595295519,0.2304127961,0.5639728904,-0.4439823329,-0.2478427738,0.4144729376,-0.356559068,0.0764389262,0.4287089705,-0.3107663095,-0.1584243178,0.3582065701,0.6321594715,0.1236561164,0.0257820133,0.0325685404,0.1308744252,-0.1530217081,-0.3346109092,0.2163310051,0.243095085,-0.0178354532,0.5360283852,0.0242514107,-0.3025455177,0.2613933384,-0.2059544474,1.02094841,-0.1854486763,0.1871506423,-0.0190800447,0.1287675947,0.4992987216,0.0092902463,-0.0842248872,-0.4268136919,-0.4827860296,0.1402025223,-0.149804309,0.2397724092,-0.0366778336,-0.1301997006,0.3139286041,-0.2595103979,0.0177618489,-0.1745316088,0.4527935386,-0.6466593742,-0.2542138696,-0.2503230572,0.0479833744,-0.0439814702,0.1101506725,0.0617118813,0.168685928,-0.329262048,-0.2781674564,-0.2272064239,0.0370319821,-0.4157636762,0.0835297033,-0.2439022213,-0.282207489,-0.0902421251,0.2852126062,-0.0320501514,0.4013364017,-0.2035952657,0.0923504159,0.0828223079,0.0968723148,0.073000282,-0.0632440969,0.3176957965,0.0168975573,0.0743537098,-0.2857038081,0.0053972583,-0.1422331929,0.032482598,-0.02342714,0.1504140943,-0.2308297455,-0.3237563372,-0.3392715752,-0.2038627863,-0.1504354775,0.0521603599,0.3335699141,-0.2387240082,-0.0088765109,-0.2857566476,-0.4090466499,-0.0096647162,0.3055093884,-0.148109138,-0.047204867,0.4327270687,0.3000046313,-0.2102095187,-0.1779171973,-0.0190174319,-0.1799966097,-0.6621257663,0.0823084489,0.0360400341,-0.0427111723,-0.2416065782,-0.2668305337,0.2053605467,0.1265298724,0.1113983393,-0.46979931,-0.4082916677,-0.2329967916,-0.2976679206,0.2016152292,-0.0566246063,-0.3551495075,-0.258842051,0.0327734798,-0.2122098804,0.0599939972,-0.1423732638,-0.0618365444,0.4633510113,-0.1418498307,0.2244604528,-0.1397035271,0.046355471,0.4012090564,0.0694132596,-0.1816515177,-0.3011113703,0.1451775283,0.0411616899,0.0585848205,-0.0672907755,-0.3786494434,-0.052751489,-0.2798325121,0.1326039732,0.3364621699,0.2424450219,0.0756183788,0.09433043,0.098653309,-0.0325292088,0.3306849301,-0.1649502367,-0.1383991838,0.0095238341,0.2725932598,-0.2107255012,0.1175259277,-0.3212481141,0.1536602676,-0.0095391795,-0.0361817405,0.0752809569,-0.0265370756,-0.1096814424,0.5564427376,0.2327184081,0.2920384705,-0.2722925246,-0.2302523255,0.2756311297,0.1339128017,0.0165831745,-0.2051161677,0.1073662043,-0.1286123842,-0.0995853841,0.3734355271,-0.0706336424,-0.1355883777,-0.1899372786,-0.1248223633,0.5824737549,0.0935555845,-0.0696668848,0.2455470562,-0.0136265624,0.0776752979,0.4613701999,0.2502685785,0.0169347022,0.5474853516,-0.3130877316,0.0298885182,-0.1627934128,-0.0062184911,-0.2004130483,-0.4473776221,0.1316621155,0.3107807636,-0.0724566802,0.187366575,-0.1468187869,0.3045750558,-0.2501968741,0.1819011271,0.0543667413,0.049220752,-0.1095726714,-0.0142696071,0.3495133519,-0.0250102021,-0.2757626772,-0.0046619414,-0.11504668,0.1466555595,0.3796023428,0.2733509541,0.1335933954,-0.1753616929,0.0041428227,0.2133577764,0.1663140357,-0.482229948,0.1435226351,0.3909594119,-0.088678211,0.2622956634,-0.0017311651,0.5537607074,0.1397955269,0.0983077362,-0.0007230406,-0.1064622477,-0.011388314,-0.1354410052,0.4446231723,-0.0308796242,-0.1025131792,0.3387673199,0.0280349217,-0.1576936543,0.1146624386,0.1928842962,0.1237160265,-0.1751527786,0.2012574673,0.0158905946,-0.2077661604,-0.1786319464,0.1449805051,-0.3786619902,-0.1633839011,0.2516909838,0.0403841957,0.2624582648,-0.1355939955,0.0489800386,-0.1194112673,0.1483440697,0.3839921951,0.2433493137,-0.2853369117,-0.0405620374,-0.6024463177,0.2246279418,-0.0487730019,-0.1710739583,0.0451600216,-0.1129757613,-0.0471117832,-0.1518046707,0.3878982961,-0.0159281138,0.111534372,0.2670121491,-0.4752603173,-0.2240053713,0.1703025699,-0.0663013086,0.2272349,-0.3988763988,0.1303192973,-0.266743809,-0.0137350131,-0.1426671296,-0.1962016076,0.1286680251,0.3134683967,0.3546344638,0.1124142408,0.2980175614,0.0608979315,-0.249863565,-0.3230270147,0.1182483435,-0.1321166456,0.0156514794,0.2199660689,0.2973281145,-0.4038825631,-0.4329209328,-0.3770737052,0.5200780034,-0.1345874667,0.0714471936,-0.1244051009,-0.0493410081,-0.0306261051,0.2739892602,0.2633121014,0.2337562293,0.070637241,0.4157318175,-0.1906892061,-0.299829632,0.659506321,-0.1065887734,-0.4276301563,0.0981976986,0.4086528718,0.0720963702,-0.1614484638,-0.4712712765,0.1547385156,0.324437052,0.0784328952,-0.2918297052,0.2315367162,-0.2830388844,0.1289331019,-0.104005374,0.2516618371,0.2692824006,-0.2087137401,-0.0591599233,-0.1967329979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2924","title":"\"File name too long\" error for file locks","comments":"Yes, you're right! I need to get you more info here. Either there's something going with the name itself that the file system doesn't like (an encoding that blows up the name length??) or perhaps there's something with the path that's causing the entire string to be used as a name. I haven't seen this on any system before and the Internet's not forthcoming with any info.","body":"## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t\/test\")`:\r\n\r\n```\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nWhere the user cache dir (e.g. `~\/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"gar1t\/test\")\r\n```\r\n\r\n## Expected results\r\n\r\nExpect the function to return without an error.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 765, in _save_info\r\n with FileLock(lock_path):\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 403, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.7\r\n- PyArrow version: 5.0.0\r\n","comment_length":67,"text":"\"File name too long\" error for file locks\n## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t\/test\")`:\r\n\r\n```\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nWhere the user cache dir (e.g. `~\/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"gar1t\/test\")\r\n```\r\n\r\n## Expected results\r\n\r\nExpect the function to return without an error.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 765, in _save_info\r\n with FileLock(lock_path):\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 403, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.7\r\n- PyArrow version: 5.0.0\r\n\nYes, you're right! I need to get you more info here. Either there's something going with the name itself that the file system doesn't like (an encoding that blows up the name length??) or perhaps there's something with the path that's causing the entire string to be used as a name. I haven't seen this on any system before and the Internet's not forthcoming with any info.","embeddings":[0.0493393503,0.0920015797,-0.0714939609,0.401848346,0.4151671231,0.2358950824,0.636662364,0.2435030937,0.2364584506,0.2313297093,0.0185690634,0.0077435691,-0.1442685276,-0.3100371361,-0.2015356272,-0.2443143427,-0.1581087261,-0.0416213386,-0.177078411,0.2177188098,-0.1549794227,0.3787363768,0.0246380791,0.1570838541,-0.536006391,0.0900789052,-0.1295958906,0.3886749148,-0.0215065964,-0.3729913831,0.051118616,-0.0467165932,0.068814449,0.811357677,-0.0001220686,-0.3334413767,0.3542338908,-0.0562992319,-0.3271892667,-0.2243083715,-0.1685816944,-0.5117114782,-0.0345845483,-0.4988024831,0.1872130185,-0.0951189175,-0.0331902057,-0.8021566272,0.0410523824,0.3733062148,0.1268745214,-0.0783471763,0.0910248384,-0.2579201758,0.3323119283,-0.2779891491,0.0040849657,0.3564026058,0.3556975722,-0.0988676101,-0.1546510309,0.2032182515,-0.0368982404,0.027324792,0.2471582741,0.079537563,-0.0874831453,-0.3261650205,0.2958056033,0.4273941517,0.4714379907,-0.0930620804,-0.2793129385,-0.4676968753,0.1424980164,-0.1251056492,0.4885306358,-0.0762403011,-0.222074613,0.1514885277,-0.1704189926,-0.0196209587,-0.1235974729,-0.1102384105,-0.101975739,0.0650619194,0.0405084826,0.0362790599,0.3111130893,-0.3483147621,0.2395959646,0.0048368936,0.0742740706,0.2694604397,-0.6542128921,0.1469123662,0.0842684656,0.3854928017,0.2534896433,0.0576435812,-0.2659727335,-0.1483078748,0.1947487742,-0.0104660569,-0.0649549589,0.374206692,0.1944440156,0.1612814218,0.3012429774,0.0779469088,-0.3367469013,-0.1069349945,-0.0764871612,-0.5534629822,0.445138216,0.133591041,0.0741629824,-0.3437117934,0.1504577994,0.5705774426,0.1057047993,-0.0104902675,0.2287023664,0.2733797133,0.005434711,0.1107665151,0.0830746666,-0.0899362862,-0.0889612734,0.1008345857,-0.1473124325,-0.1107303202,-0.2270203382,0.1124099791,0.0892055854,-0.3083957136,0.2792796195,-0.227465108,0.3152402341,-0.1249576584,-0.0829687566,-0.2377426773,-0.0667323172,0.1559667885,-0.0591069087,0.0739138052,0.2452137023,-0.2607334852,-0.1920612901,-0.0771116391,-0.3397490382,-0.2851879001,-0.0617215782,0.0947027877,-0.165494591,0.2159024328,0.2415551245,-0.2671947181,0.6555529237,-0.0798197463,0.0127971917,-0.2747654617,-0.1418095827,0.0071923966,0.0408245735,0.5534662008,0.1325238049,-0.0175623372,-0.0687896535,0.1282823086,0.0086658485,0.3722593188,0.1806448698,-0.0535966232,-0.4219526947,0.3124936521,0.2268715054,-0.2402810752,-0.486985594,0.3215382397,-0.394307375,0.0994933695,0.3229117393,0.0228835605,0.0517107099,-0.024582373,0.2853384614,0.1822132915,-0.0260959864,0.0234954469,-0.202524215,-0.1073302031,-0.0406307913,0.1691952497,0.0109546073,0.0670013204,0.0796193257,-0.1751918048,0.2715687454,0.0314637274,-0.2040443867,0.3736798167,0.1382424235,0.3267191648,0.1860822737,-0.106327638,-0.5937504768,0.2637207508,-0.1218474954,-0.1787318438,-0.2175928354,-0.1868047863,-0.1749890149,-0.0072898441,0.0034599197,0.1351371109,-0.0204084311,0.2832494974,0.1417792886,-0.1317639649,0.0934399366,0.5883799791,-0.1702439636,0.0175973084,-0.4781179428,-0.17112647,-0.1178534552,-0.1125399023,0.0315071233,0.0016395918,0.3693964779,-0.1089679822,-0.3443510532,0.4480753541,0.3332768381,0.0975331739,-0.1789251119,-0.0044306694,-0.0568642691,0.220775336,0.0726701096,0.0398121551,0.0077854944,0.0142915584,-0.2135758698,0.3000163734,-0.14081572,0.27127707,0.0286138058,0.1682408899,0.2461784184,-0.1318140626,0.0643453151,-0.1761228889,0.6458842158,-0.2163528502,0.3048405647,0.0547799319,0.0452316478,-0.1448624283,0.6458194852,0.045392707,-0.0001772655,0.1977425367,0.1562013328,-0.0408157147,-0.0193730351,0.3064890206,0.5111442208,0.1920584738,-0.0301231258,-0.2065057307,0.3941576481,-0.2238344699,0.256319046,0.0159084182,-0.0027928618,0.3657026589,0.0496179983,-0.1290875971,0.0514840819,-0.8095089197,-0.050442148,0.223697722,-0.3198090196,0.1480822265,-0.1393256634,-0.0432045013,-0.0237995591,0.0004903108,-0.0803984031,-0.4162820876,-0.1888936311,0.2785063386,-0.1418365985,0.2145332694,-0.1188357845,-0.0032407197,0.2712830305,-0.0145387761,-0.2950273752,-0.0016123495,0.0803438574,-0.0837603286,0.2720724344,-0.5375033021,0.0838043541,-0.1256639212,-0.0631072521,-0.5083525181,-0.2498805523,0.1116677895,-0.1684809178,0.2935642302,0.3413662016,0.2078575194,0.1666625887,0.0588969849,0.0049610781,-0.2622792721,0.0236695651,0.2326634228,0.1779004186,-0.0228179917,-0.1917994022,-0.023455549,-0.3463737965,-0.4160333872,0.3207841218,0.0303027015,0.2279285491,0.3411892653,-0.0708731115,0.2978130579,-0.0179723371,0.2472920567,-0.0879067779,-0.446370095,0.4170611799,-0.1525222659,-0.2990590334,-0.4318080544,0.2980134785,-0.0355094597,0.1055189967,-0.4223643541,-0.1790713519,-0.401260376,0.3389849961,-0.3357198238,0.0054093283,0.1970204413,-0.0852006599,-0.0563488565,-0.1069383249,0.0849020928,0.2325119227,0.0110361269,-0.009520852,0.0929816738,0.2951667607,-0.1160778701,0.4532136321,0.3616654873,0.0753480569,0.1953798383,-0.1376693696,0.1843574643,-0.2590326071,-0.1269457787,0.0239379536,0.1744713485,-0.221584484,0.1342053711,-0.0853912458,0.0140818916,-0.1333202869,-0.0577918552,0.1078078598,-0.2960295677,0.0073167868,0.0485861041,0.0524898916,0.031080164,0.1456944346,-0.1948870718,-0.1309104413,-0.0240893289,0.3084784746,0.0540789627,-0.0233560894,-0.2468729615,0.0915106237,-0.3371418118,0.4440097511,0.1830917299,0.2137979567,-0.069359839,-0.1601748466,-0.0800488889,0.0460065827,0.6831169128,0.0991293713,0.3160929978,0.1554959416,0.2547036707,-0.1230872124,0.0391722023,0.0795527846,0.3701882362,0.1390409023,0.4079330266,-0.2308098525,0.037608888,-0.183316946,0.1855425686,-0.1166330427,-0.4492169619,-0.0262856334,0.0087962123,-0.2636858821,0.1616199315,0.0336052962,0.2442290634,0.0616719015,0.0049132649,-0.0601768605,-0.0747785419,0.1601556242,0.1748355776,0.2812633216,-0.4468779564,0.264365375,-0.027559543,-0.1595295519,0.2304127961,0.5639728904,-0.4439823329,-0.2478427738,0.4144729376,-0.356559068,0.0764389262,0.4287089705,-0.3107663095,-0.1584243178,0.3582065701,0.6321594715,0.1236561164,0.0257820133,0.0325685404,0.1308744252,-0.1530217081,-0.3346109092,0.2163310051,0.243095085,-0.0178354532,0.5360283852,0.0242514107,-0.3025455177,0.2613933384,-0.2059544474,1.02094841,-0.1854486763,0.1871506423,-0.0190800447,0.1287675947,0.4992987216,0.0092902463,-0.0842248872,-0.4268136919,-0.4827860296,0.1402025223,-0.149804309,0.2397724092,-0.0366778336,-0.1301997006,0.3139286041,-0.2595103979,0.0177618489,-0.1745316088,0.4527935386,-0.6466593742,-0.2542138696,-0.2503230572,0.0479833744,-0.0439814702,0.1101506725,0.0617118813,0.168685928,-0.329262048,-0.2781674564,-0.2272064239,0.0370319821,-0.4157636762,0.0835297033,-0.2439022213,-0.282207489,-0.0902421251,0.2852126062,-0.0320501514,0.4013364017,-0.2035952657,0.0923504159,0.0828223079,0.0968723148,0.073000282,-0.0632440969,0.3176957965,0.0168975573,0.0743537098,-0.2857038081,0.0053972583,-0.1422331929,0.032482598,-0.02342714,0.1504140943,-0.2308297455,-0.3237563372,-0.3392715752,-0.2038627863,-0.1504354775,0.0521603599,0.3335699141,-0.2387240082,-0.0088765109,-0.2857566476,-0.4090466499,-0.0096647162,0.3055093884,-0.148109138,-0.047204867,0.4327270687,0.3000046313,-0.2102095187,-0.1779171973,-0.0190174319,-0.1799966097,-0.6621257663,0.0823084489,0.0360400341,-0.0427111723,-0.2416065782,-0.2668305337,0.2053605467,0.1265298724,0.1113983393,-0.46979931,-0.4082916677,-0.2329967916,-0.2976679206,0.2016152292,-0.0566246063,-0.3551495075,-0.258842051,0.0327734798,-0.2122098804,0.0599939972,-0.1423732638,-0.0618365444,0.4633510113,-0.1418498307,0.2244604528,-0.1397035271,0.046355471,0.4012090564,0.0694132596,-0.1816515177,-0.3011113703,0.1451775283,0.0411616899,0.0585848205,-0.0672907755,-0.3786494434,-0.052751489,-0.2798325121,0.1326039732,0.3364621699,0.2424450219,0.0756183788,0.09433043,0.098653309,-0.0325292088,0.3306849301,-0.1649502367,-0.1383991838,0.0095238341,0.2725932598,-0.2107255012,0.1175259277,-0.3212481141,0.1536602676,-0.0095391795,-0.0361817405,0.0752809569,-0.0265370756,-0.1096814424,0.5564427376,0.2327184081,0.2920384705,-0.2722925246,-0.2302523255,0.2756311297,0.1339128017,0.0165831745,-0.2051161677,0.1073662043,-0.1286123842,-0.0995853841,0.3734355271,-0.0706336424,-0.1355883777,-0.1899372786,-0.1248223633,0.5824737549,0.0935555845,-0.0696668848,0.2455470562,-0.0136265624,0.0776752979,0.4613701999,0.2502685785,0.0169347022,0.5474853516,-0.3130877316,0.0298885182,-0.1627934128,-0.0062184911,-0.2004130483,-0.4473776221,0.1316621155,0.3107807636,-0.0724566802,0.187366575,-0.1468187869,0.3045750558,-0.2501968741,0.1819011271,0.0543667413,0.049220752,-0.1095726714,-0.0142696071,0.3495133519,-0.0250102021,-0.2757626772,-0.0046619414,-0.11504668,0.1466555595,0.3796023428,0.2733509541,0.1335933954,-0.1753616929,0.0041428227,0.2133577764,0.1663140357,-0.482229948,0.1435226351,0.3909594119,-0.088678211,0.2622956634,-0.0017311651,0.5537607074,0.1397955269,0.0983077362,-0.0007230406,-0.1064622477,-0.011388314,-0.1354410052,0.4446231723,-0.0308796242,-0.1025131792,0.3387673199,0.0280349217,-0.1576936543,0.1146624386,0.1928842962,0.1237160265,-0.1751527786,0.2012574673,0.0158905946,-0.2077661604,-0.1786319464,0.1449805051,-0.3786619902,-0.1633839011,0.2516909838,0.0403841957,0.2624582648,-0.1355939955,0.0489800386,-0.1194112673,0.1483440697,0.3839921951,0.2433493137,-0.2853369117,-0.0405620374,-0.6024463177,0.2246279418,-0.0487730019,-0.1710739583,0.0451600216,-0.1129757613,-0.0471117832,-0.1518046707,0.3878982961,-0.0159281138,0.111534372,0.2670121491,-0.4752603173,-0.2240053713,0.1703025699,-0.0663013086,0.2272349,-0.3988763988,0.1303192973,-0.266743809,-0.0137350131,-0.1426671296,-0.1962016076,0.1286680251,0.3134683967,0.3546344638,0.1124142408,0.2980175614,0.0608979315,-0.249863565,-0.3230270147,0.1182483435,-0.1321166456,0.0156514794,0.2199660689,0.2973281145,-0.4038825631,-0.4329209328,-0.3770737052,0.5200780034,-0.1345874667,0.0714471936,-0.1244051009,-0.0493410081,-0.0306261051,0.2739892602,0.2633121014,0.2337562293,0.070637241,0.4157318175,-0.1906892061,-0.299829632,0.659506321,-0.1065887734,-0.4276301563,0.0981976986,0.4086528718,0.0720963702,-0.1614484638,-0.4712712765,0.1547385156,0.324437052,0.0784328952,-0.2918297052,0.2315367162,-0.2830388844,0.1289331019,-0.104005374,0.2516618371,0.2692824006,-0.2087137401,-0.0591599233,-0.1967329979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2919","title":"Unwanted progress bars when accessing examples","comments":"doing a patch release now :)","body":"When accessing examples from a dataset formatted for pytorch, some progress bars appear when accessing examples:\r\n```python\r\nIn [1]: import datasets as ds \r\n\r\nIn [2]: d = ds.Dataset.from_dict({\"a\": [0, 1, 2]}).with_format(\"torch\") \r\n\r\nIn [3]: d[0] \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 3172.70it\/s]\r\nOut[3]: {'a': tensor(0)}\r\n```\r\n\r\nThis is because the pytorch formatter calls `map_nested` that uses progress bars\r\n\r\ncc @sgugger ","comment_length":6,"text":"Unwanted progress bars when accessing examples\nWhen accessing examples from a dataset formatted for pytorch, some progress bars appear when accessing examples:\r\n```python\r\nIn [1]: import datasets as ds \r\n\r\nIn [2]: d = ds.Dataset.from_dict({\"a\": [0, 1, 2]}).with_format(\"torch\") \r\n\r\nIn [3]: d[0] \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 3172.70it\/s]\r\nOut[3]: {'a': tensor(0)}\r\n```\r\n\r\nThis is because the pytorch formatter calls `map_nested` that uses progress bars\r\n\r\ncc @sgugger \ndoing a patch release now :)","embeddings":[-0.0175480675,-0.2450868934,0.0048732767,0.10296496,0.1993564218,0.0186564457,0.5456594825,0.2686175108,-0.3862209618,0.233160913,0.1730663329,0.4301009178,-0.0968903452,0.163059175,0.0540391617,-0.2791727781,-0.0362301841,0.0565355495,0.0279643871,0.077860862,-0.0710608438,-0.1534580141,-0.1631929129,0.2491238117,-0.4254311323,-0.1769718528,0.0204986315,-0.2497630417,0.2259828746,-0.6468563676,0.3624524474,0.017228175,0.0925377831,0.4540927112,-0.0001153858,0.1453293562,0.4285333753,0.2160392404,-0.2830248773,-0.1382309198,0.3537838757,-0.5126107335,0.3308672309,-0.3774378598,-0.0873389021,-0.5281800628,-0.2761972547,-0.4125932157,0.2876147032,0.0940865576,0.219826296,0.4712334275,-0.2902378142,-0.0055018864,0.5578870773,0.1376104206,-0.3271023929,0.1375557929,0.5464651585,0.0474077016,-0.2612787783,0.6541268229,-0.1014311612,0.3388422132,0.1663392037,-0.0622609183,0.0632258281,-0.1512287408,-0.1227654591,0.3897278607,0.1704966575,-0.0895446837,-0.2203477472,-0.810939312,-0.0621742755,-0.0982443467,-0.1076176837,0.1732621491,0.0438141599,0.0303935967,-0.5963996649,0.0290020015,-0.27146402,0.0937305689,0.0468122512,-0.1223285422,-0.0942170694,0.0973090604,-0.0284371544,0.083435297,0.1243030503,-0.1994698942,0.0932707414,0.0010196987,-0.046441935,0.1807018369,0.1487292498,-0.170697704,-0.1335336268,0.1875990182,-0.0521841459,0.1841346174,-0.0532620661,0.1660489142,-0.0838754028,-0.0406403653,0.1388794035,-0.2265987098,0.3347287774,0.0163198523,0.2717235684,0.1755359918,-0.037750639,-0.2225538939,0.1156542897,0.1093147323,0.1090649813,0.0211095344,-0.1003607139,0.2777105868,-0.3311299682,0.0219744816,0.2182542384,0.2214021385,-0.1495082378,-0.2717446387,0.1687204242,0.2794264853,-0.0600529797,0.2132000178,-0.0743473321,-0.1063092723,-0.4007383883,-0.1235286221,0.2540350258,0.1500249505,0.2033672929,0.1358764023,0.1028876156,-0.0182450209,0.3721589446,0.0875214264,0.4961630404,0.1029135436,-0.2043189853,0.2071680725,0.075304538,0.4455814064,-0.0290748477,0.271618098,-0.0666681826,-0.3141361773,-0.5112269521,0.1377267092,-0.4437701702,0.2145982981,-0.0623717569,-0.0707329661,0.3211912811,-0.0310045853,-0.0121499458,-0.3121537268,-0.247789517,0.0666654259,0.2291434705,0.2016990036,-0.1900742948,-0.2416289151,0.2099226117,0.0013152339,0.0640754029,0.1185466424,-0.23551175,0.3596096337,-0.2419328094,-0.026544407,0.1872638762,-0.1814072579,-0.3709934652,-0.1421762258,0.0234307237,0.3529036641,0.0315213762,-0.1182669997,0.3322122693,0.0806723833,-0.0189346746,0.1771131754,0.2529223263,-0.0826835856,-0.2729922533,0.0518658161,0.3036836088,0.2500388026,-0.0991399065,-0.1353184134,-0.2295820117,-0.2512821555,0.1488023102,-0.2186713666,0.1137110293,-0.0783538818,-0.0333985202,-0.1114408895,0.128908366,-0.200934723,-0.1400860697,0.334146589,-0.1916212291,-0.0585839003,-0.1843481064,-0.1486791074,-0.2163711041,-0.0423979238,-0.2948601842,-0.0875450671,0.0705130398,-0.1551547796,-0.0718744695,-0.0183241945,0.0944948941,-0.1325680614,-0.3775212765,-0.0472109951,0.0076890243,0.3263398409,0.1543031484,-0.3600919843,-0.1122498885,0.4872063696,0.0123988334,-0.0924425796,-0.1616445929,0.0323477574,0.1806732118,0.020979777,-0.418775171,0.1870616972,0.3267300427,0.1015354991,-0.0053531798,0.2033281922,-0.110525474,-0.0142406756,-0.2096013576,0.275554955,0.3913537562,0.3871296048,-0.0188226514,0.1576325446,0.2299357653,0.2287838459,-0.19063811,-0.3708530664,0.3794209957,0.1164492592,0.2275259644,-0.1128709093,-0.056149736,0.2054813057,0.2419088632,-0.0129112294,0.0710802451,0.1942463815,-0.3472250402,-0.0823546052,0.2821888328,-0.0113203805,0.0383519419,0.0980123356,0.1815937907,0.1735030264,-0.0646654963,-0.2774474919,-0.002367727,0.252974838,0.1037782133,0.0188018717,0.1014887169,0.0363334157,-0.0996529385,-0.2078756839,-0.2531711757,0.0459414124,-0.4298397005,0.2782195807,-0.3938655257,-0.2349227965,-0.1121626124,-0.1845910549,0.0518490709,-0.2728966475,0.1209051535,0.5843741298,-0.1538914591,0.3002850115,0.3276317716,0.1584009379,0.0250879861,0.1818053126,-0.0510158017,-0.510720849,-0.2618277967,0.0802433044,-0.0849289596,-0.1672920287,0.5056151152,-0.1877110004,0.4327579141,-0.3314349949,-0.1271976531,0.1429183781,-0.0935863927,0.0404994525,-0.0486703366,0.1458623558,-0.0911693275,0.233792603,0.160188511,-0.1380708367,-0.0214611273,0.16621387,-0.2902114093,-0.0336513706,-0.2353742123,0.0583931804,-0.3945191801,-0.1011367291,0.0673801824,-0.1514690369,0.1948434263,-0.0160474218,-0.3517791629,0.4189587235,0.1875047535,-0.1340169311,-0.1357982308,-0.3630971313,0.2848750055,-0.41330266,-0.2727486789,-0.0870164037,-0.2269277126,-0.2414059043,0.2249186188,-0.4970188141,-0.4690299034,-0.2819567025,0.2401717752,0.0068959016,-0.0885310322,0.2275055051,0.2388396263,-0.0490171686,-0.1502595395,-0.1286485642,0.2684911489,-0.2023165226,-0.0526690036,-0.2280971706,-0.1113249883,0.1931582093,0.4944924414,0.1924008131,-0.2912912369,0.1187348738,0.0306330826,0.1123298332,-0.1802462488,-0.1826853007,-0.1176944822,-0.4110670388,0.0464733355,-0.0148358615,0.0150283165,-0.0615319833,0.2057303935,-0.0157856271,-0.1149998009,-0.292308867,-0.1367049962,-0.5465351939,0.2146653235,-0.0885391533,0.3431994319,-0.1929283291,-0.1536388397,-0.1729557216,-0.1027353704,0.3973326683,0.0111482572,-0.1668602377,-0.256231606,-0.0367813185,0.3502275944,0.4951102138,0.0122325225,-0.0494005047,-0.0513418317,0.4644909203,0.4761361778,0.2832567096,0.1001850367,-0.1279858649,0.2342715114,-0.2401283532,-0.7125119567,-0.41146034,-0.1997616738,0.4119268656,0.2129512131,0.4722346961,0.0457137078,-0.086379692,0.0880065113,0.4283269346,-0.1754733473,-0.1151453927,-0.164950341,-0.3428955376,-0.3676280975,0.2961567342,0.1100679711,-0.0125856651,-0.2604800761,-0.2776361108,-0.0639647469,0.0514944792,-0.0319752358,0.3141577542,0.229007259,0.0247712489,0.1257308871,0.4538356066,0.4744880199,0.2547186613,0.0707441568,0.1482186466,-0.048177246,0.2233390659,-0.4182990193,-0.0396085083,0.2151462138,-0.0113113131,0.300827384,-0.2532792091,0.0221573282,-0.2611340284,0.0794604346,0.2527319491,-0.0078818072,-0.1881323904,-0.316670537,0.2333788425,-0.0685120448,-0.178301841,0.2720800936,-0.3529309928,-0.1420414448,0.1504247189,0.1549986601,0.6575912833,-0.0286711324,0.1168500483,0.511208117,-0.2467370182,0.2707988918,0.2357586771,0.1782183349,-0.2843172848,-0.2260917276,0.077235803,-0.446272552,0.0647938997,-0.0009366414,-0.0220275056,-0.131005004,-0.5273835063,0.2063754499,0.0190239213,-0.0590235889,0.3956880569,-0.3413526118,-0.2414980084,0.1582078487,-0.1195397153,0.3477330506,-0.153694585,0.105606243,0.0029755994,0.0559373796,-0.2887510955,0.0984202698,0.1579602063,0.3901692331,0.0510993376,-0.7411983609,-0.0375015289,0.1320708096,0.3103530407,-0.0464583524,-0.120931454,0.1860230863,0.0712058544,0.1173128709,0.2080661058,-0.0445344783,0.3048805892,-0.2177551091,-0.2133199573,0.289378047,0.0879581645,-0.2650895715,-0.0903161094,0.1210575998,0.0623679422,-0.0285778753,0.146940738,-0.1023302898,0.1881256551,-0.179275915,0.1194247305,0.098932758,-0.2934565246,0.3730723858,-0.1679872125,-0.2851409614,0.023308998,0.0684312284,0.0924657211,-0.0535454601,0.5884126425,0.2965933979,-0.3969217241,-0.0453655347,0.2278076261,0.2701043487,-0.2893387973,0.1174756289,0.2835693359,0.1774771065,-0.413019985,-0.0458238162,-0.0454670712,-0.026506193,-0.2763679922,-0.3110716641,-0.3819492757,-0.1838888377,-0.037768133,0.3098086715,0.0986221954,0.2046764046,-0.1496843994,-0.1149211973,-0.2811182737,0.2810724676,-0.1052022502,0.2823551893,-0.0097785192,-0.2017995566,-0.0716521591,-0.3540430963,-0.0920337066,0.0676924735,-0.2509935796,-0.2414507866,-0.2945031524,0.0720853955,-0.1283766329,-0.5431334972,0.066728361,0.1387889385,0.1103348732,-0.0959893838,0.4568045437,0.3058486581,0.0772642195,0.3269970715,-0.2034392655,0.2739566267,-0.0409143828,-0.020709984,-0.1116553992,0.3173391223,0.1663935035,-0.0878654346,0.2291099429,0.1036329418,0.1654630154,-0.0290552266,0.0285506975,0.3309161067,0.1152022108,-0.1805236489,0.2272562385,-0.1822111905,0.3503676355,0.0519738756,-0.2619753778,-0.2643000782,0.2569488585,0.1530782729,-0.0500352234,-0.2258733511,0.4774631262,0.0756995752,0.0340617336,0.1253063083,0.2247715294,0.5043908358,-0.0475612022,0.0738995597,-0.081570439,0.0508443005,0.3201390505,0.2587703466,-0.0798705667,-0.0055938778,0.0351336226,0.0449205413,0.0840099305,-0.0845154375,0.3153921962,0.6276469231,0.5006055236,-0.0922196507,0.0945133418,-0.263574332,-0.4391740263,-0.3639207184,-0.1227432266,0.0407780483,-0.0924314335,0.2919994593,-0.2905572057,-0.4293654561,-0.3388067782,-0.0357804783,-0.1445909292,-0.070831351,-0.034298256,0.0904835537,0.0901739672,-0.0436631069,-0.1581738442,0.1526149809,0.0199174788,-0.1052327827,-0.0309135877,-0.0641777217,0.0459759124,-0.0262360517,0.3488769233,0.1154232547,-0.0487071984,0.1308671981,0.0627249703,-0.1025522426,0.1025527045,0.1545532346,0.2601504624,-0.3949838877,-0.1687651724,-0.1504912376,0.1469617635,-0.0068152202,0.3295758367,-0.0582019277,-0.4498440027,0.2931940854,0.0710852221,-0.1421114355,0.6278657317,0.0279215667,0.1311353743,-0.2680278718,0.3336524963,-0.2577918768,-0.1453121305,0.0927531719,-0.3478544056,0.0340452418,-0.1008718088,0.8285409808,-0.0055257455,0.0478106216,-0.1501996219,0.0873420835,-0.1911148876,0.6613516212,0.1349851936,0.5720160007,-0.0729165375,0.0926222354,-0.3475141227,0.1489682943,-0.1697594225,0.0487481356,0.2350684851,-0.1679376215,-0.3383406103,0.4007802308,0.5639624,-0.2407636493,-0.249806881,-0.3299007118,-0.3072220385,0.0551428385,0.0805551186,-0.0869708359,-0.0475652367,-0.1825208068,0.0548046343,-0.0901909247,0.1151650399,0.0051526413,-0.1455165148,-0.0173772313,-0.0279609356,0.4075471163,0.2661443949,0.1931025535,0.2716137767,0.0248533059,0.0354542583,-0.0741970539,0.0967927873,0.4841645956,0.0150432829,0.329911083,-0.0158772171,0.0391109847,-0.1803999096,0.4807221889,0.3238476217,0.0178111568,-0.1421049386,0.0984386057,-0.2610158622,0.2574439645,-0.016893791,0.0306532588,-0.0154680144,0.4048753083,-0.0077654449,-0.1700899154,0.5261824131,-0.4042277336,-0.3670017421,-0.0829792991,0.2184514552,0.1859294027,-0.0921535343,-0.3823102415,-0.2459062785,0.2568610907,-0.1912846565,-0.1312142015,0.1051476598,0.0230771396,-0.1514546424,-0.0044824672,0.5724635124,0.0191494972,0.0986652225,-0.0518430099,-0.3697446585]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2918","title":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming","comments":"Hi @SBrandeis, thanks for reporting! ^^\r\n\r\nI think this is an issue with `fsspec`: https:\/\/github.com\/intake\/filesystem_spec\/issues\/389\r\n\r\nI will ask them if they are planning to fix it...","body":"## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n","comment_length":26,"text":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming\n## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n\nHi @SBrandeis, thanks for reporting! ^^\r\n\r\nI think this is an issue with `fsspec`: https:\/\/github.com\/intake\/filesystem_spec\/issues\/389\r\n\r\nI will ask them if they are planning to fix it...","embeddings":[-0.3864652812,-0.2204957455,0.0894813165,0.4387415349,0.2157086134,0.1229229122,-0.0369101539,0.3093882799,0.2422775328,0.0955026373,-0.2264311761,0.4261536002,-0.0641756877,0.3035327792,-0.0094654057,-0.2584786713,-0.0045060921,0.2337655127,-0.1049250886,0.1268760264,-0.0005556306,0.1941259056,-0.2469931841,-0.0304233823,0.0522285365,-0.1081046611,0.0124065261,0.2805253863,-0.059583243,-0.4462363422,0.1259049624,-0.0615482517,0.4121141732,0.3375741839,-0.0001145032,0.2734483778,0.4721427262,-0.1227841005,-0.4853440225,-0.2121149004,-0.218665719,0.0961271524,0.0428345352,-0.1566341817,-0.0091345869,0.0711133331,-0.0400046185,-0.7192984819,0.3178500533,0.3456359804,0.1778500229,0.0426006429,-0.0489487238,0.0268223863,0.0688386783,-0.2986646891,-0.0920066237,0.1218173951,0.2367273569,0.2287404835,-0.359573245,0.3845903277,-0.0658884048,0.0601840951,0.0944148228,0.0967238322,-0.1749501675,-0.4655051231,-0.0073944977,0.1884141415,0.6776787043,-0.2555166781,-0.4161978662,-0.2319881767,-0.0021713183,-0.6144573689,0.1465310603,0.1489273459,-0.4673651159,0.1280597448,0.2459630817,-0.0087530883,-0.327994585,-0.0458728224,-0.246799171,0.3882223368,-0.0767857283,-0.0237270929,0.0272943433,-0.0866289064,0.4627741277,-0.2110346556,-0.28459391,-0.0639790371,-0.4397552311,-0.0420655161,-0.0765123144,-0.094109349,0.1681466401,0.215389967,0.2953203917,0.0882523879,0.1543645263,0.0926442817,0.3555265665,0.3608669043,0.0598316565,0.1853774786,0.1311810017,0.0994313955,0.1366237104,-0.1818924248,-0.146488592,-0.0372809209,0.0281632524,-0.0131140957,0.262858659,-0.1912374645,-0.4138476253,0.0297920927,-0.3057210743,-0.0794672072,0.0055905147,0.2817168832,0.0958337784,0.4187803268,0.0194585584,0.1178673506,-0.2396837324,-0.4573938251,-0.0878112167,-0.1247182265,-0.0010082311,0.1233467832,0.0153353186,-0.1886832416,0.1515737325,-0.0465371571,0.1621010602,-0.0012185768,-0.0208561029,-0.1100598425,0.0216292907,0.2049841285,0.2180556655,0.1640353203,0.1795151085,-0.1997850239,-0.130682379,0.0886226073,-0.1416330338,-0.1671275645,-0.1788374335,0.2166469097,-0.0094024939,-0.3506264389,-0.0937892571,0.2814688385,0.1409590542,-0.2037903368,-0.1410701424,-0.0391465798,-0.1241319329,-0.0865156949,0.1792775393,0.5796180964,-0.1855915189,-0.1380216628,-0.1521381587,0.0008939481,0.5833057165,0.3344025612,-0.0607914478,0.0897968188,-0.170722425,0.1476544142,0.5017140508,-0.1436584145,-0.5796036124,0.5436661243,-0.0396082066,0.5637281537,0.2707032561,0.0225366186,0.1742344499,0.0431196988,0.313529104,0.3299024701,-0.1484592706,0.0074822325,-0.3334669471,-0.0399836674,0.4416950643,0.1912285686,0.0674290732,0.2502861321,-0.0288742278,-0.0626677349,0.3689779639,-0.0752563924,0.0763451308,0.0410080664,0.1468882412,0.0120370993,0.1379382461,-0.2090770155,-0.310682714,0.2972311378,0.0749307722,0.0127100674,-0.3174338341,0.2031723857,-0.1906455308,0.0017469627,-0.2628925443,-0.2590405643,0.1243173257,0.3170801699,0.0756207108,0.0505895875,-0.3084766865,0.7747215033,-0.078255251,-0.0131349247,-0.4549141228,0.2024929523,-0.0707768798,-0.2014573216,0.0164402872,-0.1265656799,0.0672556311,-0.0161919687,-0.2579914331,0.3313200772,-0.2366138101,0.2640713155,-0.1954375803,-0.1731976271,0.2897448838,-0.3644788265,-0.0866843313,0.2706028223,0.1075036898,0.125209704,-0.0176983867,0.1480839401,0.2070915401,0.2361335903,0.0434485786,0.1143487841,0.1961675137,-0.0816748515,-0.0686837137,-0.0850090533,0.3240485787,-0.3548789024,0.0467500761,-0.1989230514,-0.2722995877,0.0823631957,0.3586791456,-0.1403829306,-0.0492737889,0.3303245604,-0.2885322273,-0.1639078707,0.3145605624,0.1940099597,0.4723360538,0.0815219656,0.2573353648,-0.0455882996,-0.1886483878,-0.0735060051,0.3051697612,0.1374839693,0.0314638466,0.2464114279,0.0300063528,0.1635747999,-0.3229160309,-0.3497166634,-0.0566318817,0.1123416647,-0.3201874793,0.306302011,-0.38826859,-0.4651744664,0.0463509336,0.1560085416,0.0689049438,-0.3235587776,-0.1817550063,0.2325055152,-0.0038823313,0.1636838913,-0.2682063282,0.0895485729,0.2671770155,-0.0134687945,-0.3418698907,-0.1808474958,-0.2729629874,0.0591977872,0.1749951094,-0.4228037596,0.1769195795,0.0074260007,-0.1160532683,-0.1401735991,-0.076423429,0.3197818398,-0.1554995328,0.1568958163,-0.2774865627,0.1493911892,0.2213633955,-0.2503491044,0.1219482496,0.0908335969,-0.1187844425,0.2801248431,0.0940897167,-0.0241610892,-0.0305832736,-0.3010807931,-0.3024666011,-0.6177494526,0.0549683981,0.0454257689,0.2135515809,0.1643050015,0.0170347784,0.0968246013,0.3292069137,0.0991601422,-0.114212282,-0.4199434817,0.3747448325,-0.1670146883,-0.2868409157,-0.3248367012,-0.0681597292,0.1972901672,0.2589270473,-0.5295243263,0.2950162292,-0.4081035852,0.0848925635,-0.3834516108,-0.0974560827,-0.0261515174,0.0762885734,-0.0349893123,-0.1878576428,0.0707459524,-0.1169513538,0.0990076959,0.2066107839,0.1086307243,0.3698712885,0.1008493528,0.3327504098,0.4812112153,-0.0195496976,0.3407551944,0.082891129,0.1538007557,-0.1004334986,-0.0627885908,-0.0955621898,-0.2280800343,-0.0261197276,0.0981990993,-0.0962127745,-0.5762300491,-0.2840516269,0.0549149849,-0.0692367032,-0.2383105308,0.0528838076,-0.511775136,0.0420482606,-0.1872570366,0.1583397985,0.0807970166,0.0520012863,0.0860457644,0.1772524714,0.1261645257,0.1721196771,-0.185021311,-0.0639052019,-0.4881900847,0.1825860441,0.1051262617,0.4215749502,-0.1678105742,-0.1327165961,0.0791147202,-0.0652071014,0.3487038612,-0.2754168212,0.003567247,0.2673943937,0.1421353668,-0.1415521204,0.0917675421,-0.3209457397,0.1368288398,0.235954389,0.0974201635,-0.1718239039,0.0333400816,-0.0884195939,-0.0724961758,-0.1253028065,-0.0163991563,-0.1704963744,-0.2492375523,-0.152009964,-0.134242788,0.2032724619,0.1256832927,-0.0124605224,-0.0567925759,-0.0460816547,-0.0396244489,0.2418593764,0.2213254869,0.1399096996,-0.072280176,0.2290114909,0.0671992972,0.2844450474,0.2292162329,0.836432457,0.1181946918,-0.5624260902,0.2350820601,0.2339563966,0.1660992503,0.1305397898,-0.148232609,-0.0085166721,-0.0403849594,0.2698210478,-0.3024733663,0.2146548331,0.2112327665,-0.0507208817,-0.5149303675,-0.4357535839,0.5883390903,-0.039368242,-0.0761907846,0.7872929573,-0.1665049642,-0.3171558082,0.4775415361,0.0234355517,0.7585879564,0.0943952724,0.2168687731,0.2407481223,-0.3207278252,0.5003329515,-0.2320687622,0.3346687257,-0.1206100509,-0.3350296021,-0.0831092745,-0.1041097715,0.1709026545,-0.147603929,-0.6302542686,0.1709609181,-0.0596441738,-0.2528051734,0.3353525698,0.2742269933,-0.6082689762,-0.3062928915,-0.0675497502,0.1388825476,-0.2213602215,0.3239610493,0.1382443607,-0.2202336341,-0.0515731983,-0.0843103677,-0.4702361524,0.4758630693,-0.2765566409,0.1624139398,-0.2390307039,-0.4760982394,0.1513123959,0.2664022446,-0.1421327889,-0.2257857472,-0.1582512558,0.2350440472,0.1860539466,0.128023684,0.0993071273,0.1514580995,0.3883590102,0.0532772318,-0.0168129094,0.1884423494,-0.0089801019,-0.1606287062,-0.0637812093,0.0554574467,0.3183139861,-0.2681806087,0.2368578017,-0.1058510691,0.0429251082,-0.1842529923,0.0869722143,-0.0188314319,-0.1740541756,-0.0738878325,-0.08129704,-0.2196760625,-0.0847442299,0.3771200478,0.1454578936,0.3179713786,0.4994189143,0.0281169843,-0.1778806895,-0.1036716029,-0.0551506244,-0.2931399345,-0.252612263,0.0693599209,-0.2414764762,-0.1610280573,-0.2706694007,0.0038265749,0.0448373631,-0.103554666,-0.1137386858,-0.2993660271,-0.1906369925,-0.0266146194,-0.1242436618,0.2199304104,-0.1722201109,0.0222039819,0.0031197246,0.1290859282,-0.270683825,-0.1228595003,-0.0753692091,0.0231177211,0.2672106326,0.0292674527,0.2592337132,0.1096690521,0.1476223618,0.3139047623,-0.0276747048,-0.1609539241,-0.1787950993,0.138489455,0.0165312476,-0.348051548,-0.1548943967,-0.05662065,0.0126910359,-0.125136137,0.3503144085,0.181230098,0.0380968302,0.2706013322,0.1240136698,0.0563360825,-0.2326476574,0.1542367488,-0.2642730176,0.2401308864,0.0811118633,0.0928405821,-0.0755398721,-0.0118692657,-0.3646134436,0.0622693747,0.1017946154,-0.0280170571,0.5909268856,-0.4241805971,0.1720450372,-0.0497748405,0.0660299286,0.4045228958,-0.2296188772,-0.2089618146,0.0850852206,0.2000514269,-0.0432033539,-0.2126210332,-0.4010480046,-0.1016407683,0.0112383589,0.1459593624,-0.1068896204,-0.2721833587,0.0999634638,-0.2103084773,0.4859834313,0.3358429968,0.2670664489,0.4836392403,-0.1811283231,0.0228473805,-0.2260247022,0.0204593837,0.1999064386,0.2970430553,0.0035512666,-0.0078696832,-0.1613658667,0.2178194076,-0.1859746873,-0.482195437,-0.0797479451,0.1024470255,-0.1647549272,0.3901939094,0.1489096284,0.1521606892,-0.2779354751,-0.0519221425,-0.0458415113,-0.0661262944,-0.0142799746,-0.096080415,0.1068764329,-0.0128562739,-0.3279225528,0.0976088271,0.0123996465,0.0633685365,-0.0198641457,0.2808055282,-0.2884631157,-0.1185014918,-0.0822276548,0.1043173745,0.2999251485,-0.3544151187,0.4440693855,0.3661904037,-0.1568607092,0.0282512195,0.0700021908,0.4816074371,0.2853379846,-0.1766507179,0.0946500823,0.3374164999,-0.0436370596,-0.0402214266,0.0442541391,0.2830949128,0.1793196946,0.4566897452,0.0887393281,-0.1485419422,0.1045718119,0.0814351737,0.2538890839,-0.2799384296,0.4623198807,-0.1195058972,0.0366923138,0.0272421893,-0.3365718424,-0.5624158978,-0.0282046329,0.0348961949,-0.2285001576,0.0745235756,-0.0081122555,0.0628783256,0.1139546707,0.4926685393,0.3248410225,-0.0188453235,-0.2245095521,-0.1508785933,-0.235583514,0.2332020253,-0.0674795657,-0.1231484041,0.1752853394,0.2219228297,-0.0048415069,0.1680491567,-0.1584448814,0.0204550214,-0.2672839463,0.1405724436,-0.3995800018,-0.3515672684,0.1919635981,0.1735794991,-0.0483116023,-0.576480031,0.1881390065,-0.1502387971,0.1608176678,0.0926132202,-0.241848141,0.101145193,-0.2756578326,0.5297338963,0.2331105173,0.3792004883,-0.184295252,-0.4119567871,-0.0992468074,-0.3762306273,-0.2749142647,-0.0155847622,0.1083334535,0.5696322322,-0.1493341029,-0.2373841256,-0.3574708402,0.3262623847,0.0608894229,-0.2656728923,-0.2735387981,0.1733704358,-0.1194598079,0.2131261081,0.3050509691,0.380345583,-0.0287131406,0.1639950573,-0.2235154659,-0.4333209693,0.5773110986,-0.3461063206,-0.2659671307,0.2418412119,0.0638139024,0.3264979422,0.2410199344,-0.4596662819,0.0893656164,0.3556547463,-0.2354129702,-0.1511856616,0.1421462744,0.2259705812,-0.0607201755,-0.0850398242,0.45531106,0.2423961759,-0.249204129,0.1750813872,-0.1786384881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2918","title":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming","comments":"Code to reproduce the bug: `ClientPayloadError: 400, message='Can not decode content-encoding: gzip'`\r\n```python\r\nIn [1]: import fsspec\r\n\r\nIn [2]: import json\r\n\r\nIn [3]: with fsspec.open('https:\/\/raw.githubusercontent.com\/allenai\/scitldr\/master\/SciTLDR-Data\/SciTLDR-FullText\/test.jsonl', encoding=\"utf-8\") as f:\r\n ...: for row in f:\r\n ...: data = json.loads(row)\r\n ...:\r\n---------------------------------------------------------------------------\r\nClientPayloadError Traceback (most recent call last)\r\n```","body":"## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n","comment_length":46,"text":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming\n## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n\nCode to reproduce the bug: `ClientPayloadError: 400, message='Can not decode content-encoding: gzip'`\r\n```python\r\nIn [1]: import fsspec\r\n\r\nIn [2]: import json\r\n\r\nIn [3]: with fsspec.open('https:\/\/raw.githubusercontent.com\/allenai\/scitldr\/master\/SciTLDR-Data\/SciTLDR-FullText\/test.jsonl', encoding=\"utf-8\") as f:\r\n ...: for row in f:\r\n ...: data = json.loads(row)\r\n ...:\r\n---------------------------------------------------------------------------\r\nClientPayloadError Traceback (most recent call last)\r\n```","embeddings":[-0.3864652812,-0.2204957455,0.0894813165,0.4387415349,0.2157086134,0.1229229122,-0.0369101539,0.3093882799,0.2422775328,0.0955026373,-0.2264311761,0.4261536002,-0.0641756877,0.3035327792,-0.0094654057,-0.2584786713,-0.0045060921,0.2337655127,-0.1049250886,0.1268760264,-0.0005556306,0.1941259056,-0.2469931841,-0.0304233823,0.0522285365,-0.1081046611,0.0124065261,0.2805253863,-0.059583243,-0.4462363422,0.1259049624,-0.0615482517,0.4121141732,0.3375741839,-0.0001145032,0.2734483778,0.4721427262,-0.1227841005,-0.4853440225,-0.2121149004,-0.218665719,0.0961271524,0.0428345352,-0.1566341817,-0.0091345869,0.0711133331,-0.0400046185,-0.7192984819,0.3178500533,0.3456359804,0.1778500229,0.0426006429,-0.0489487238,0.0268223863,0.0688386783,-0.2986646891,-0.0920066237,0.1218173951,0.2367273569,0.2287404835,-0.359573245,0.3845903277,-0.0658884048,0.0601840951,0.0944148228,0.0967238322,-0.1749501675,-0.4655051231,-0.0073944977,0.1884141415,0.6776787043,-0.2555166781,-0.4161978662,-0.2319881767,-0.0021713183,-0.6144573689,0.1465310603,0.1489273459,-0.4673651159,0.1280597448,0.2459630817,-0.0087530883,-0.327994585,-0.0458728224,-0.246799171,0.3882223368,-0.0767857283,-0.0237270929,0.0272943433,-0.0866289064,0.4627741277,-0.2110346556,-0.28459391,-0.0639790371,-0.4397552311,-0.0420655161,-0.0765123144,-0.094109349,0.1681466401,0.215389967,0.2953203917,0.0882523879,0.1543645263,0.0926442817,0.3555265665,0.3608669043,0.0598316565,0.1853774786,0.1311810017,0.0994313955,0.1366237104,-0.1818924248,-0.146488592,-0.0372809209,0.0281632524,-0.0131140957,0.262858659,-0.1912374645,-0.4138476253,0.0297920927,-0.3057210743,-0.0794672072,0.0055905147,0.2817168832,0.0958337784,0.4187803268,0.0194585584,0.1178673506,-0.2396837324,-0.4573938251,-0.0878112167,-0.1247182265,-0.0010082311,0.1233467832,0.0153353186,-0.1886832416,0.1515737325,-0.0465371571,0.1621010602,-0.0012185768,-0.0208561029,-0.1100598425,0.0216292907,0.2049841285,0.2180556655,0.1640353203,0.1795151085,-0.1997850239,-0.130682379,0.0886226073,-0.1416330338,-0.1671275645,-0.1788374335,0.2166469097,-0.0094024939,-0.3506264389,-0.0937892571,0.2814688385,0.1409590542,-0.2037903368,-0.1410701424,-0.0391465798,-0.1241319329,-0.0865156949,0.1792775393,0.5796180964,-0.1855915189,-0.1380216628,-0.1521381587,0.0008939481,0.5833057165,0.3344025612,-0.0607914478,0.0897968188,-0.170722425,0.1476544142,0.5017140508,-0.1436584145,-0.5796036124,0.5436661243,-0.0396082066,0.5637281537,0.2707032561,0.0225366186,0.1742344499,0.0431196988,0.313529104,0.3299024701,-0.1484592706,0.0074822325,-0.3334669471,-0.0399836674,0.4416950643,0.1912285686,0.0674290732,0.2502861321,-0.0288742278,-0.0626677349,0.3689779639,-0.0752563924,0.0763451308,0.0410080664,0.1468882412,0.0120370993,0.1379382461,-0.2090770155,-0.310682714,0.2972311378,0.0749307722,0.0127100674,-0.3174338341,0.2031723857,-0.1906455308,0.0017469627,-0.2628925443,-0.2590405643,0.1243173257,0.3170801699,0.0756207108,0.0505895875,-0.3084766865,0.7747215033,-0.078255251,-0.0131349247,-0.4549141228,0.2024929523,-0.0707768798,-0.2014573216,0.0164402872,-0.1265656799,0.0672556311,-0.0161919687,-0.2579914331,0.3313200772,-0.2366138101,0.2640713155,-0.1954375803,-0.1731976271,0.2897448838,-0.3644788265,-0.0866843313,0.2706028223,0.1075036898,0.125209704,-0.0176983867,0.1480839401,0.2070915401,0.2361335903,0.0434485786,0.1143487841,0.1961675137,-0.0816748515,-0.0686837137,-0.0850090533,0.3240485787,-0.3548789024,0.0467500761,-0.1989230514,-0.2722995877,0.0823631957,0.3586791456,-0.1403829306,-0.0492737889,0.3303245604,-0.2885322273,-0.1639078707,0.3145605624,0.1940099597,0.4723360538,0.0815219656,0.2573353648,-0.0455882996,-0.1886483878,-0.0735060051,0.3051697612,0.1374839693,0.0314638466,0.2464114279,0.0300063528,0.1635747999,-0.3229160309,-0.3497166634,-0.0566318817,0.1123416647,-0.3201874793,0.306302011,-0.38826859,-0.4651744664,0.0463509336,0.1560085416,0.0689049438,-0.3235587776,-0.1817550063,0.2325055152,-0.0038823313,0.1636838913,-0.2682063282,0.0895485729,0.2671770155,-0.0134687945,-0.3418698907,-0.1808474958,-0.2729629874,0.0591977872,0.1749951094,-0.4228037596,0.1769195795,0.0074260007,-0.1160532683,-0.1401735991,-0.076423429,0.3197818398,-0.1554995328,0.1568958163,-0.2774865627,0.1493911892,0.2213633955,-0.2503491044,0.1219482496,0.0908335969,-0.1187844425,0.2801248431,0.0940897167,-0.0241610892,-0.0305832736,-0.3010807931,-0.3024666011,-0.6177494526,0.0549683981,0.0454257689,0.2135515809,0.1643050015,0.0170347784,0.0968246013,0.3292069137,0.0991601422,-0.114212282,-0.4199434817,0.3747448325,-0.1670146883,-0.2868409157,-0.3248367012,-0.0681597292,0.1972901672,0.2589270473,-0.5295243263,0.2950162292,-0.4081035852,0.0848925635,-0.3834516108,-0.0974560827,-0.0261515174,0.0762885734,-0.0349893123,-0.1878576428,0.0707459524,-0.1169513538,0.0990076959,0.2066107839,0.1086307243,0.3698712885,0.1008493528,0.3327504098,0.4812112153,-0.0195496976,0.3407551944,0.082891129,0.1538007557,-0.1004334986,-0.0627885908,-0.0955621898,-0.2280800343,-0.0261197276,0.0981990993,-0.0962127745,-0.5762300491,-0.2840516269,0.0549149849,-0.0692367032,-0.2383105308,0.0528838076,-0.511775136,0.0420482606,-0.1872570366,0.1583397985,0.0807970166,0.0520012863,0.0860457644,0.1772524714,0.1261645257,0.1721196771,-0.185021311,-0.0639052019,-0.4881900847,0.1825860441,0.1051262617,0.4215749502,-0.1678105742,-0.1327165961,0.0791147202,-0.0652071014,0.3487038612,-0.2754168212,0.003567247,0.2673943937,0.1421353668,-0.1415521204,0.0917675421,-0.3209457397,0.1368288398,0.235954389,0.0974201635,-0.1718239039,0.0333400816,-0.0884195939,-0.0724961758,-0.1253028065,-0.0163991563,-0.1704963744,-0.2492375523,-0.152009964,-0.134242788,0.2032724619,0.1256832927,-0.0124605224,-0.0567925759,-0.0460816547,-0.0396244489,0.2418593764,0.2213254869,0.1399096996,-0.072280176,0.2290114909,0.0671992972,0.2844450474,0.2292162329,0.836432457,0.1181946918,-0.5624260902,0.2350820601,0.2339563966,0.1660992503,0.1305397898,-0.148232609,-0.0085166721,-0.0403849594,0.2698210478,-0.3024733663,0.2146548331,0.2112327665,-0.0507208817,-0.5149303675,-0.4357535839,0.5883390903,-0.039368242,-0.0761907846,0.7872929573,-0.1665049642,-0.3171558082,0.4775415361,0.0234355517,0.7585879564,0.0943952724,0.2168687731,0.2407481223,-0.3207278252,0.5003329515,-0.2320687622,0.3346687257,-0.1206100509,-0.3350296021,-0.0831092745,-0.1041097715,0.1709026545,-0.147603929,-0.6302542686,0.1709609181,-0.0596441738,-0.2528051734,0.3353525698,0.2742269933,-0.6082689762,-0.3062928915,-0.0675497502,0.1388825476,-0.2213602215,0.3239610493,0.1382443607,-0.2202336341,-0.0515731983,-0.0843103677,-0.4702361524,0.4758630693,-0.2765566409,0.1624139398,-0.2390307039,-0.4760982394,0.1513123959,0.2664022446,-0.1421327889,-0.2257857472,-0.1582512558,0.2350440472,0.1860539466,0.128023684,0.0993071273,0.1514580995,0.3883590102,0.0532772318,-0.0168129094,0.1884423494,-0.0089801019,-0.1606287062,-0.0637812093,0.0554574467,0.3183139861,-0.2681806087,0.2368578017,-0.1058510691,0.0429251082,-0.1842529923,0.0869722143,-0.0188314319,-0.1740541756,-0.0738878325,-0.08129704,-0.2196760625,-0.0847442299,0.3771200478,0.1454578936,0.3179713786,0.4994189143,0.0281169843,-0.1778806895,-0.1036716029,-0.0551506244,-0.2931399345,-0.252612263,0.0693599209,-0.2414764762,-0.1610280573,-0.2706694007,0.0038265749,0.0448373631,-0.103554666,-0.1137386858,-0.2993660271,-0.1906369925,-0.0266146194,-0.1242436618,0.2199304104,-0.1722201109,0.0222039819,0.0031197246,0.1290859282,-0.270683825,-0.1228595003,-0.0753692091,0.0231177211,0.2672106326,0.0292674527,0.2592337132,0.1096690521,0.1476223618,0.3139047623,-0.0276747048,-0.1609539241,-0.1787950993,0.138489455,0.0165312476,-0.348051548,-0.1548943967,-0.05662065,0.0126910359,-0.125136137,0.3503144085,0.181230098,0.0380968302,0.2706013322,0.1240136698,0.0563360825,-0.2326476574,0.1542367488,-0.2642730176,0.2401308864,0.0811118633,0.0928405821,-0.0755398721,-0.0118692657,-0.3646134436,0.0622693747,0.1017946154,-0.0280170571,0.5909268856,-0.4241805971,0.1720450372,-0.0497748405,0.0660299286,0.4045228958,-0.2296188772,-0.2089618146,0.0850852206,0.2000514269,-0.0432033539,-0.2126210332,-0.4010480046,-0.1016407683,0.0112383589,0.1459593624,-0.1068896204,-0.2721833587,0.0999634638,-0.2103084773,0.4859834313,0.3358429968,0.2670664489,0.4836392403,-0.1811283231,0.0228473805,-0.2260247022,0.0204593837,0.1999064386,0.2970430553,0.0035512666,-0.0078696832,-0.1613658667,0.2178194076,-0.1859746873,-0.482195437,-0.0797479451,0.1024470255,-0.1647549272,0.3901939094,0.1489096284,0.1521606892,-0.2779354751,-0.0519221425,-0.0458415113,-0.0661262944,-0.0142799746,-0.096080415,0.1068764329,-0.0128562739,-0.3279225528,0.0976088271,0.0123996465,0.0633685365,-0.0198641457,0.2808055282,-0.2884631157,-0.1185014918,-0.0822276548,0.1043173745,0.2999251485,-0.3544151187,0.4440693855,0.3661904037,-0.1568607092,0.0282512195,0.0700021908,0.4816074371,0.2853379846,-0.1766507179,0.0946500823,0.3374164999,-0.0436370596,-0.0402214266,0.0442541391,0.2830949128,0.1793196946,0.4566897452,0.0887393281,-0.1485419422,0.1045718119,0.0814351737,0.2538890839,-0.2799384296,0.4623198807,-0.1195058972,0.0366923138,0.0272421893,-0.3365718424,-0.5624158978,-0.0282046329,0.0348961949,-0.2285001576,0.0745235756,-0.0081122555,0.0628783256,0.1139546707,0.4926685393,0.3248410225,-0.0188453235,-0.2245095521,-0.1508785933,-0.235583514,0.2332020253,-0.0674795657,-0.1231484041,0.1752853394,0.2219228297,-0.0048415069,0.1680491567,-0.1584448814,0.0204550214,-0.2672839463,0.1405724436,-0.3995800018,-0.3515672684,0.1919635981,0.1735794991,-0.0483116023,-0.576480031,0.1881390065,-0.1502387971,0.1608176678,0.0926132202,-0.241848141,0.101145193,-0.2756578326,0.5297338963,0.2331105173,0.3792004883,-0.184295252,-0.4119567871,-0.0992468074,-0.3762306273,-0.2749142647,-0.0155847622,0.1083334535,0.5696322322,-0.1493341029,-0.2373841256,-0.3574708402,0.3262623847,0.0608894229,-0.2656728923,-0.2735387981,0.1733704358,-0.1194598079,0.2131261081,0.3050509691,0.380345583,-0.0287131406,0.1639950573,-0.2235154659,-0.4333209693,0.5773110986,-0.3461063206,-0.2659671307,0.2418412119,0.0638139024,0.3264979422,0.2410199344,-0.4596662819,0.0893656164,0.3556547463,-0.2354129702,-0.1511856616,0.1421462744,0.2259705812,-0.0607201755,-0.0850398242,0.45531106,0.2423961759,-0.249204129,0.1750813872,-0.1786384881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2918","title":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming","comments":"Thanks for investigating @albertvillanova ! \ud83e\udd17 ","body":"## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n","comment_length":6,"text":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming\n## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n\nThanks for investigating @albertvillanova ! \ud83e\udd17 ","embeddings":[-0.3864652812,-0.2204957455,0.0894813165,0.4387415349,0.2157086134,0.1229229122,-0.0369101539,0.3093882799,0.2422775328,0.0955026373,-0.2264311761,0.4261536002,-0.0641756877,0.3035327792,-0.0094654057,-0.2584786713,-0.0045060921,0.2337655127,-0.1049250886,0.1268760264,-0.0005556306,0.1941259056,-0.2469931841,-0.0304233823,0.0522285365,-0.1081046611,0.0124065261,0.2805253863,-0.059583243,-0.4462363422,0.1259049624,-0.0615482517,0.4121141732,0.3375741839,-0.0001145032,0.2734483778,0.4721427262,-0.1227841005,-0.4853440225,-0.2121149004,-0.218665719,0.0961271524,0.0428345352,-0.1566341817,-0.0091345869,0.0711133331,-0.0400046185,-0.7192984819,0.3178500533,0.3456359804,0.1778500229,0.0426006429,-0.0489487238,0.0268223863,0.0688386783,-0.2986646891,-0.0920066237,0.1218173951,0.2367273569,0.2287404835,-0.359573245,0.3845903277,-0.0658884048,0.0601840951,0.0944148228,0.0967238322,-0.1749501675,-0.4655051231,-0.0073944977,0.1884141415,0.6776787043,-0.2555166781,-0.4161978662,-0.2319881767,-0.0021713183,-0.6144573689,0.1465310603,0.1489273459,-0.4673651159,0.1280597448,0.2459630817,-0.0087530883,-0.327994585,-0.0458728224,-0.246799171,0.3882223368,-0.0767857283,-0.0237270929,0.0272943433,-0.0866289064,0.4627741277,-0.2110346556,-0.28459391,-0.0639790371,-0.4397552311,-0.0420655161,-0.0765123144,-0.094109349,0.1681466401,0.215389967,0.2953203917,0.0882523879,0.1543645263,0.0926442817,0.3555265665,0.3608669043,0.0598316565,0.1853774786,0.1311810017,0.0994313955,0.1366237104,-0.1818924248,-0.146488592,-0.0372809209,0.0281632524,-0.0131140957,0.262858659,-0.1912374645,-0.4138476253,0.0297920927,-0.3057210743,-0.0794672072,0.0055905147,0.2817168832,0.0958337784,0.4187803268,0.0194585584,0.1178673506,-0.2396837324,-0.4573938251,-0.0878112167,-0.1247182265,-0.0010082311,0.1233467832,0.0153353186,-0.1886832416,0.1515737325,-0.0465371571,0.1621010602,-0.0012185768,-0.0208561029,-0.1100598425,0.0216292907,0.2049841285,0.2180556655,0.1640353203,0.1795151085,-0.1997850239,-0.130682379,0.0886226073,-0.1416330338,-0.1671275645,-0.1788374335,0.2166469097,-0.0094024939,-0.3506264389,-0.0937892571,0.2814688385,0.1409590542,-0.2037903368,-0.1410701424,-0.0391465798,-0.1241319329,-0.0865156949,0.1792775393,0.5796180964,-0.1855915189,-0.1380216628,-0.1521381587,0.0008939481,0.5833057165,0.3344025612,-0.0607914478,0.0897968188,-0.170722425,0.1476544142,0.5017140508,-0.1436584145,-0.5796036124,0.5436661243,-0.0396082066,0.5637281537,0.2707032561,0.0225366186,0.1742344499,0.0431196988,0.313529104,0.3299024701,-0.1484592706,0.0074822325,-0.3334669471,-0.0399836674,0.4416950643,0.1912285686,0.0674290732,0.2502861321,-0.0288742278,-0.0626677349,0.3689779639,-0.0752563924,0.0763451308,0.0410080664,0.1468882412,0.0120370993,0.1379382461,-0.2090770155,-0.310682714,0.2972311378,0.0749307722,0.0127100674,-0.3174338341,0.2031723857,-0.1906455308,0.0017469627,-0.2628925443,-0.2590405643,0.1243173257,0.3170801699,0.0756207108,0.0505895875,-0.3084766865,0.7747215033,-0.078255251,-0.0131349247,-0.4549141228,0.2024929523,-0.0707768798,-0.2014573216,0.0164402872,-0.1265656799,0.0672556311,-0.0161919687,-0.2579914331,0.3313200772,-0.2366138101,0.2640713155,-0.1954375803,-0.1731976271,0.2897448838,-0.3644788265,-0.0866843313,0.2706028223,0.1075036898,0.125209704,-0.0176983867,0.1480839401,0.2070915401,0.2361335903,0.0434485786,0.1143487841,0.1961675137,-0.0816748515,-0.0686837137,-0.0850090533,0.3240485787,-0.3548789024,0.0467500761,-0.1989230514,-0.2722995877,0.0823631957,0.3586791456,-0.1403829306,-0.0492737889,0.3303245604,-0.2885322273,-0.1639078707,0.3145605624,0.1940099597,0.4723360538,0.0815219656,0.2573353648,-0.0455882996,-0.1886483878,-0.0735060051,0.3051697612,0.1374839693,0.0314638466,0.2464114279,0.0300063528,0.1635747999,-0.3229160309,-0.3497166634,-0.0566318817,0.1123416647,-0.3201874793,0.306302011,-0.38826859,-0.4651744664,0.0463509336,0.1560085416,0.0689049438,-0.3235587776,-0.1817550063,0.2325055152,-0.0038823313,0.1636838913,-0.2682063282,0.0895485729,0.2671770155,-0.0134687945,-0.3418698907,-0.1808474958,-0.2729629874,0.0591977872,0.1749951094,-0.4228037596,0.1769195795,0.0074260007,-0.1160532683,-0.1401735991,-0.076423429,0.3197818398,-0.1554995328,0.1568958163,-0.2774865627,0.1493911892,0.2213633955,-0.2503491044,0.1219482496,0.0908335969,-0.1187844425,0.2801248431,0.0940897167,-0.0241610892,-0.0305832736,-0.3010807931,-0.3024666011,-0.6177494526,0.0549683981,0.0454257689,0.2135515809,0.1643050015,0.0170347784,0.0968246013,0.3292069137,0.0991601422,-0.114212282,-0.4199434817,0.3747448325,-0.1670146883,-0.2868409157,-0.3248367012,-0.0681597292,0.1972901672,0.2589270473,-0.5295243263,0.2950162292,-0.4081035852,0.0848925635,-0.3834516108,-0.0974560827,-0.0261515174,0.0762885734,-0.0349893123,-0.1878576428,0.0707459524,-0.1169513538,0.0990076959,0.2066107839,0.1086307243,0.3698712885,0.1008493528,0.3327504098,0.4812112153,-0.0195496976,0.3407551944,0.082891129,0.1538007557,-0.1004334986,-0.0627885908,-0.0955621898,-0.2280800343,-0.0261197276,0.0981990993,-0.0962127745,-0.5762300491,-0.2840516269,0.0549149849,-0.0692367032,-0.2383105308,0.0528838076,-0.511775136,0.0420482606,-0.1872570366,0.1583397985,0.0807970166,0.0520012863,0.0860457644,0.1772524714,0.1261645257,0.1721196771,-0.185021311,-0.0639052019,-0.4881900847,0.1825860441,0.1051262617,0.4215749502,-0.1678105742,-0.1327165961,0.0791147202,-0.0652071014,0.3487038612,-0.2754168212,0.003567247,0.2673943937,0.1421353668,-0.1415521204,0.0917675421,-0.3209457397,0.1368288398,0.235954389,0.0974201635,-0.1718239039,0.0333400816,-0.0884195939,-0.0724961758,-0.1253028065,-0.0163991563,-0.1704963744,-0.2492375523,-0.152009964,-0.134242788,0.2032724619,0.1256832927,-0.0124605224,-0.0567925759,-0.0460816547,-0.0396244489,0.2418593764,0.2213254869,0.1399096996,-0.072280176,0.2290114909,0.0671992972,0.2844450474,0.2292162329,0.836432457,0.1181946918,-0.5624260902,0.2350820601,0.2339563966,0.1660992503,0.1305397898,-0.148232609,-0.0085166721,-0.0403849594,0.2698210478,-0.3024733663,0.2146548331,0.2112327665,-0.0507208817,-0.5149303675,-0.4357535839,0.5883390903,-0.039368242,-0.0761907846,0.7872929573,-0.1665049642,-0.3171558082,0.4775415361,0.0234355517,0.7585879564,0.0943952724,0.2168687731,0.2407481223,-0.3207278252,0.5003329515,-0.2320687622,0.3346687257,-0.1206100509,-0.3350296021,-0.0831092745,-0.1041097715,0.1709026545,-0.147603929,-0.6302542686,0.1709609181,-0.0596441738,-0.2528051734,0.3353525698,0.2742269933,-0.6082689762,-0.3062928915,-0.0675497502,0.1388825476,-0.2213602215,0.3239610493,0.1382443607,-0.2202336341,-0.0515731983,-0.0843103677,-0.4702361524,0.4758630693,-0.2765566409,0.1624139398,-0.2390307039,-0.4760982394,0.1513123959,0.2664022446,-0.1421327889,-0.2257857472,-0.1582512558,0.2350440472,0.1860539466,0.128023684,0.0993071273,0.1514580995,0.3883590102,0.0532772318,-0.0168129094,0.1884423494,-0.0089801019,-0.1606287062,-0.0637812093,0.0554574467,0.3183139861,-0.2681806087,0.2368578017,-0.1058510691,0.0429251082,-0.1842529923,0.0869722143,-0.0188314319,-0.1740541756,-0.0738878325,-0.08129704,-0.2196760625,-0.0847442299,0.3771200478,0.1454578936,0.3179713786,0.4994189143,0.0281169843,-0.1778806895,-0.1036716029,-0.0551506244,-0.2931399345,-0.252612263,0.0693599209,-0.2414764762,-0.1610280573,-0.2706694007,0.0038265749,0.0448373631,-0.103554666,-0.1137386858,-0.2993660271,-0.1906369925,-0.0266146194,-0.1242436618,0.2199304104,-0.1722201109,0.0222039819,0.0031197246,0.1290859282,-0.270683825,-0.1228595003,-0.0753692091,0.0231177211,0.2672106326,0.0292674527,0.2592337132,0.1096690521,0.1476223618,0.3139047623,-0.0276747048,-0.1609539241,-0.1787950993,0.138489455,0.0165312476,-0.348051548,-0.1548943967,-0.05662065,0.0126910359,-0.125136137,0.3503144085,0.181230098,0.0380968302,0.2706013322,0.1240136698,0.0563360825,-0.2326476574,0.1542367488,-0.2642730176,0.2401308864,0.0811118633,0.0928405821,-0.0755398721,-0.0118692657,-0.3646134436,0.0622693747,0.1017946154,-0.0280170571,0.5909268856,-0.4241805971,0.1720450372,-0.0497748405,0.0660299286,0.4045228958,-0.2296188772,-0.2089618146,0.0850852206,0.2000514269,-0.0432033539,-0.2126210332,-0.4010480046,-0.1016407683,0.0112383589,0.1459593624,-0.1068896204,-0.2721833587,0.0999634638,-0.2103084773,0.4859834313,0.3358429968,0.2670664489,0.4836392403,-0.1811283231,0.0228473805,-0.2260247022,0.0204593837,0.1999064386,0.2970430553,0.0035512666,-0.0078696832,-0.1613658667,0.2178194076,-0.1859746873,-0.482195437,-0.0797479451,0.1024470255,-0.1647549272,0.3901939094,0.1489096284,0.1521606892,-0.2779354751,-0.0519221425,-0.0458415113,-0.0661262944,-0.0142799746,-0.096080415,0.1068764329,-0.0128562739,-0.3279225528,0.0976088271,0.0123996465,0.0633685365,-0.0198641457,0.2808055282,-0.2884631157,-0.1185014918,-0.0822276548,0.1043173745,0.2999251485,-0.3544151187,0.4440693855,0.3661904037,-0.1568607092,0.0282512195,0.0700021908,0.4816074371,0.2853379846,-0.1766507179,0.0946500823,0.3374164999,-0.0436370596,-0.0402214266,0.0442541391,0.2830949128,0.1793196946,0.4566897452,0.0887393281,-0.1485419422,0.1045718119,0.0814351737,0.2538890839,-0.2799384296,0.4623198807,-0.1195058972,0.0366923138,0.0272421893,-0.3365718424,-0.5624158978,-0.0282046329,0.0348961949,-0.2285001576,0.0745235756,-0.0081122555,0.0628783256,0.1139546707,0.4926685393,0.3248410225,-0.0188453235,-0.2245095521,-0.1508785933,-0.235583514,0.2332020253,-0.0674795657,-0.1231484041,0.1752853394,0.2219228297,-0.0048415069,0.1680491567,-0.1584448814,0.0204550214,-0.2672839463,0.1405724436,-0.3995800018,-0.3515672684,0.1919635981,0.1735794991,-0.0483116023,-0.576480031,0.1881390065,-0.1502387971,0.1608176678,0.0926132202,-0.241848141,0.101145193,-0.2756578326,0.5297338963,0.2331105173,0.3792004883,-0.184295252,-0.4119567871,-0.0992468074,-0.3762306273,-0.2749142647,-0.0155847622,0.1083334535,0.5696322322,-0.1493341029,-0.2373841256,-0.3574708402,0.3262623847,0.0608894229,-0.2656728923,-0.2735387981,0.1733704358,-0.1194598079,0.2131261081,0.3050509691,0.380345583,-0.0287131406,0.1639950573,-0.2235154659,-0.4333209693,0.5773110986,-0.3461063206,-0.2659671307,0.2418412119,0.0638139024,0.3264979422,0.2410199344,-0.4596662819,0.0893656164,0.3556547463,-0.2354129702,-0.1511856616,0.1421462744,0.2259705812,-0.0607201755,-0.0850398242,0.45531106,0.2423961759,-0.249204129,0.1750813872,-0.1786384881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2917","title":"windows download abnormal","comments":"Hi ! Is there some kind of proxy that is configured in your browser that gives you access to internet ? If it's the case it could explain why it doesn't work in the code, since the proxy wouldn't be used","body":"## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n","comment_length":41,"text":"windows download abnormal\n## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n\nHi ! Is there some kind of proxy that is configured in your browser that gives you access to internet ? If it's the case it could explain why it doesn't work in the code, since the proxy wouldn't be used","embeddings":[-0.1240484565,-0.2018248588,-0.0703396499,0.1894593984,0.2147738934,-0.1120505109,0.0276671592,0.0718739256,0.2805814743,0.0850560516,0.1648736149,0.1174071059,0.1554815173,0.1022035182,0.1588474363,-0.2615045309,0.0301855393,0.0393100753,-0.1691515595,0.0088643078,-0.3215381503,0.0859283432,-0.3066691756,-0.1127134934,0.1635590941,0.0324402116,-0.2886959612,0.3004715145,-0.2545996904,-0.1694043726,0.099369511,-0.0709087476,0.1850661635,0.3519495726,-0.0001087431,0.0813257769,0.3922842741,0.0235310104,0.0174283162,-0.0581749007,-0.2938911617,-0.1413686872,-0.1911946684,-0.3151716888,0.0997065529,0.2347301394,0.0454377979,-0.1112284884,0.244362548,0.4491443336,0.3060822189,0.4809369445,-0.1150025278,0.0517268442,0.4554628432,0.0505648442,-0.1816764176,0.2491989583,0.3063491285,-0.1419377476,0.2570019662,-0.059411183,-0.1210118234,0.0115583753,0.0310037415,0.1206353381,-0.0494647175,-0.5093158484,0.1234319583,0.1743744612,0.31266132,-0.1531091034,-0.1585402191,0.1158476546,-0.1493998766,-0.185696885,0.2733875215,0.2762599587,-0.1393169761,0.1871818155,-0.2805486917,0.1743876189,-0.2056800723,0.2596932352,-0.1191073507,0.3299077153,-0.0759498328,0.1894465089,-0.1710143685,0.3167176545,0.03771016,-0.3272551894,-0.0156770907,-0.0039893929,0.0856893435,0.105707854,0.0493557453,0.5118294954,0.05730417,0.2643461227,0.0334886052,-0.0560189262,0.2410898954,0.0700188726,0.090272449,0.0949797034,-0.2958666682,0.066043593,0.3817819655,0.313916266,-0.0065793474,-0.026200084,-0.0330495574,-0.3155067265,0.2965119183,-0.0053916001,0.486035049,-0.04742698,-0.556910634,0.0026351295,0.0146284271,0.0908918232,-0.0023530021,0.2526860833,-0.32183218,-0.022580795,0.1233704016,0.3000146747,-0.1240452528,0.1943567246,0.004063094,0.1806814522,-0.2520693839,-0.2260728776,0.3475835323,-0.0671294406,0.12012472,0.1935619116,-0.0914587379,-0.042873878,0.0132011967,-0.0805152804,0.0271826498,0.4316020012,0.1952559203,0.3702535331,-0.0094595077,0.0780785382,-0.0707269013,0.2062932998,-0.1463157386,-0.0249700788,0.021134628,0.2608017921,-0.3265893757,-0.1885598153,0.0397941098,-0.2664661705,-0.0409113131,-0.0293878373,0.2495985031,-0.1964796036,-0.2891480029,-0.3841315806,0.0929920301,0.5247913599,-0.3482265472,0.2068513781,0.0614641719,-0.271443069,0.4014218748,0.0586097874,0.1309562773,0.1605100483,-0.4302997291,0.0415051356,0.117131643,-0.3501324654,-0.4970429242,0.3584975302,-0.1899531484,0.3206911981,0.1783815473,0.0919841528,0.293910265,-0.0833817273,-0.012038324,0.3098945022,-0.022483252,0.0582434274,-0.1245782077,-0.1297618896,0.1300003082,0.1085215658,0.1208592206,0.1432953179,0.293230623,-0.1360394806,0.3164023459,0.1788195372,0.0964889973,0.2375457287,0.2550174594,0.2680895925,0.1236548349,-0.2269855589,0.0015702655,0.2861069441,-0.0804182291,-0.1590327024,-0.2410328686,-0.2218061388,-0.5582574606,0.0224495344,-0.1886620373,-0.2219815105,0.1653012484,0.1372602582,0.249552846,0.0426937975,0.0754473731,0.179336831,-0.0071864342,-0.068459399,0.0751905069,0.3602052331,-0.3439503014,-0.050523717,-0.096596159,-0.0721758753,0.2299172729,-0.0566853061,-0.0152284885,0.197268486,0.0202615354,-0.0426289514,0.0362087451,0.0486600362,0.0167178959,-0.1805609912,0.2716660798,0.5724846125,0.2685571015,-0.0967838913,-0.0322295241,0.0053620976,-0.075025402,0.0177561473,-0.0222576819,0.1486974061,0.1936407834,-0.188959837,-0.1011113226,0.0021014542,0.3638005555,0.0613047183,-0.0991422012,-0.1685334742,-0.1679939777,0.167080909,0.5450258255,-0.2752330005,0.3014795184,0.0706219375,-0.0966639817,0.2094732076,-0.1800906807,0.1059181914,0.6080446839,0.033296559,0.112960726,0.3711541593,0.0088629806,-0.1436058879,0.2149284184,0.1461233199,-0.2989706993,0.1730046868,-0.0520784967,0.1769192368,-0.2095150501,-0.2376021594,-0.0571574084,0.0526774004,-0.3452227414,0.2429896444,-0.1614445746,-0.1740004569,-0.3291527927,-0.086957626,-0.0766705945,-0.0632937625,0.0246917997,0.2807940841,-0.1450743377,-0.1818347871,-0.4370901287,0.092590116,0.2119114399,-0.3590200245,-0.0015145239,0.1992023885,-0.2863779962,0.1333586127,0.1517552882,-0.0984132662,0.5485373139,-0.5067276955,0.0579211973,-0.3667593896,-0.1009359881,-0.0654943958,0.0553083122,0.3579990566,0.182639122,0.1857500374,0.0548134558,0.1031560078,0.2081247717,-0.2389806062,-0.179515481,0.1597154588,0.0564894304,-0.3985366821,-0.2559876144,-0.2491457164,-0.2066613436,-0.3509724736,0.0161345936,0.1100101173,0.1269879043,-0.1030690521,0.0080765542,0.0372453369,0.0731868744,0.2246696055,-0.083026208,0.0263501108,0.3867479861,-0.0423500165,-0.6439450383,0.2405601293,-0.0189747661,-0.3578401804,0.3052423,-0.5030599833,-0.132392019,-0.2873800397,0.2506045699,0.0892270058,0.2778801024,0.3353272378,-0.2777280211,-0.0245965142,-0.3039755225,-0.0659073442,-0.2901861966,-0.1096735597,0.2428965271,0.1826853007,0.3031284213,-0.0289380215,0.2855456471,0.2206116915,0.1294395328,0.4378712475,-0.3864388764,0.4741620421,0.001300477,-0.1765860617,0.035116639,-0.2722176611,0.0180080682,0.2376574576,0.022387892,-0.004943266,-0.2312289178,-0.1656051874,-0.231309548,-0.2478713095,0.1438776106,0.1106841117,0.0407827273,0.1241698265,0.2241891772,0.1028406844,-0.1557345092,0.1141644493,0.5970383286,-0.2615623474,0.1439668685,0.0227814205,-0.1640785187,-0.4693872631,0.0716748014,0.1027646363,0.4488425553,-0.0301668625,0.0443425849,0.2156197429,-0.1619966328,0.7495212555,-0.2151187062,0.3322592676,0.2032127231,-0.0067156502,-0.3780357242,-0.2583532333,-0.5120751262,0.123098731,0.0817756206,-0.0529246628,-0.0555245318,0.1852327734,0.1150577143,0.2468607128,-0.1428902447,-0.3871606886,-0.3885157406,-0.6134721041,-0.2445401102,0.085090965,0.4038743079,-0.0435088724,-0.3796861768,-0.1664035767,-0.0642804652,0.1533765495,-0.0725113228,-0.0277298614,0.0795960501,-0.017775014,0.1780821234,-0.2506742775,0.2862488925,-0.1739084572,0.6596170068,0.1054916233,-0.1496224999,0.1620587111,-0.2683568299,-0.2014200389,0.2051974386,-0.1649345309,-0.220199585,-0.0797214359,0.0492504165,-0.3298475146,-0.0799625888,0.5729669333,-0.101683788,-0.0054186624,-0.2596149147,0.1787786633,0.0138970232,-0.1558488905,0.2674739361,0.0348851718,-0.0924990326,-0.108341001,-0.2890718579,0.651114285,0.2145095021,0.2358662635,0.2590305209,-0.5253791809,0.0419862084,0.2136851996,0.0738660768,-0.2070637792,-0.2838785946,-0.0373834483,-0.1332621872,0.0825860873,-0.0341505371,-0.2779364288,0.2207877785,-0.0865820795,0.1747573763,0.2744155824,-0.1581595838,-0.313105762,0.1084561571,-0.0328159146,0.195117116,0.2385889739,0.6629055142,-0.2662183642,0.0184965637,-0.1715994924,-0.2779659629,-0.3965134323,0.255030632,-0.1340843886,0.0074540563,-0.1599158049,0.1396337897,0.0225496311,-0.0147371292,0.0021659129,0.1571852863,-0.2298111022,0.0857386962,0.1341610104,-0.0822363645,0.263004005,-0.0994025618,0.1180735081,-0.2447074354,-0.4229418635,-0.0727604851,0.0214396194,-0.0027245332,-0.0277121738,0.2639417648,0.2187709659,-0.1882333159,-0.1784814149,-0.3920820057,0.083862029,-0.0708945021,0.2175155431,0.0490465052,0.0885434523,0.5029727817,-0.2861326635,-0.2359030694,-0.1554895788,0.3932054639,-0.1673704088,-0.0755681172,0.1102359146,0.0948679075,-0.3069735765,-0.2523458302,0.0131361112,-0.3942269385,-0.2205286175,0.0540067367,-0.0543695949,0.2869075835,-0.1471764296,0.0090060364,0.1335751414,-0.0578686707,-0.0151845738,-0.6815165877,0.1296804845,-0.2197964042,0.412201792,0.0466443598,0.0262459274,-0.157206744,0.4631694555,-0.5489163995,-0.3702111244,-0.0008696999,0.0495049134,-0.1290583014,0.1779977232,0.1302772611,0.1998716146,0.0118945697,0.1212280467,0.0002453557,-0.0661850348,-0.2484522611,-0.1650175601,0.1045997515,-0.0540886931,-0.2632085681,0.1900731623,-0.4511168599,0.0091411378,0.18899028,0.14730452,0.3043759167,-0.0957731083,0.0366581641,0.0956626311,0.1146368384,-0.2398875803,0.2471019179,-0.4936737418,0.5263616443,0.0056513948,0.1713121086,0.2247191817,0.0065553351,-0.054802306,-0.281349808,0.062704742,0.1826201379,0.3916451633,-0.1167972237,0.2606552839,-0.0006228012,0.283251822,0.5525352955,-0.352715075,-0.0893501639,0.1433627754,0.275811404,-0.2590067089,-0.0346851833,0.0842292532,0.1091726124,0.0576287955,0.1179928631,0.0555240661,-0.2826647758,-0.0912482589,-0.0500898212,0.4335025847,0.3670791388,-0.1808381081,0.3947139084,0.0065175421,0.1476498097,-0.0576697998,-0.0217515342,-0.0486023426,0.4234354198,0.0409806445,0.0692659318,-0.1091255769,0.0251742806,-0.1315766871,-0.2052455992,0.0766841099,-0.2012568414,0.0254510138,0.097710222,0.0770515501,-0.1540478617,-0.2748750746,0.2177974731,-0.3454121649,-0.0678378046,0.0694723278,-0.0125074452,-0.3036753833,-0.0721507445,-0.1196656898,0.0252670608,0.1473329514,-0.0054127262,0.259942174,0.0745272934,-0.2369666696,-0.3160700798,0.0126224859,-0.2048031986,0.1134118363,-0.2885008752,0.0778895468,0.3658398986,0.1674333662,0.1078387499,0.1148513034,0.5545535088,0.3852181435,-0.1072471216,-0.061104849,0.0784613341,-0.1331625879,-0.3258143961,0.43635571,0.071632877,0.193093881,0.2363510728,0.2184062153,-0.2301051915,0.2584095299,-0.030818494,-0.0795634016,-0.0589464083,0.1791744381,-0.2505965829,0.0937965512,-0.3897891343,-0.1013927385,-0.4083535373,-0.098580189,0.1772197485,-0.0879954323,0.1667040288,-0.3462209702,0.1035932824,0.0375302359,0.2181000561,0.3795768917,0.3196724355,-0.088166751,-0.0573232137,-0.5226137042,0.3419146836,-0.2258433551,-0.5296083689,-0.0877915993,0.0447274148,-0.6105858684,0.0722363666,0.3415539265,0.2679618895,-0.0636082962,0.3222130537,-0.1352423877,-0.3121239245,0.0335871987,0.1525713652,0.0759184957,-0.5940473676,0.2918206453,-0.4170546532,0.1697357893,-0.2700899541,-0.1863706708,-0.0113285938,-0.0088556763,0.5608647466,0.1507781595,0.5199138522,-0.1912190616,-0.1256828904,-0.4629922211,0.0781015754,-0.1126953214,0.2469523996,0.116675429,-0.1811442822,0.0191525444,-0.2678036988,-0.0600166209,0.3138544858,-0.0368412472,0.0905802175,-0.0571234785,-0.0728074685,-0.183093071,0.1488180161,0.0811512843,0.1466140598,-0.3787932098,-0.1018520966,-0.0337751284,-0.5434590578,0.2741713226,-0.2698807716,-0.3332621753,-0.2687217593,0.2311914861,-0.4032007754,0.0730406642,-0.0221859496,0.1500967294,0.1658335179,-0.0674838051,0.0399886183,0.2936095595,-0.3748459816,0.0406060629,0.25768736,0.4179147184,-0.0696628317,-0.1712902337,0.3069999814,-0.2148606628]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2917","title":"windows download abnormal","comments":"It is indeed an agency problem, thank you very, very much","body":"## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n","comment_length":11,"text":"windows download abnormal\n## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n\nIt is indeed an agency problem, thank you very, very much","embeddings":[-0.1819480658,-0.0709162951,-0.0238682684,0.1505493969,0.2515722811,-0.0938239619,0.0032474413,0.11024113,0.362229526,0.0661070794,0.1379210949,0.2134204656,0.0683521777,0.2074184567,-0.0183983669,-0.2120418251,0.0408425257,0.169603616,-0.109351553,0.0084896954,-0.3778957427,0.1222561225,-0.244990021,-0.0322376676,0.118513003,-0.1276566386,-0.3018625379,0.1873607785,-0.2236432433,-0.0762688667,0.1565723866,-0.1330198646,0.1389445513,0.3082361221,-0.0001110555,0.0270548705,0.4110395014,0.0012154208,0.0296383202,-0.0051854542,-0.1887717694,-0.1067669615,-0.2910613418,-0.3018800914,0.163738668,0.3125549555,0.0271366872,-0.1251074672,0.2193036079,0.4728519619,0.3004208505,0.4755435586,-0.1050411314,-0.052202031,0.2329810262,0.0768294707,-0.1763174385,0.1497375667,0.254214406,-0.2113004178,0.2170041651,-0.0690852255,-0.0610753894,-0.0063095409,-0.0902022049,0.0914523378,0.0073744375,-0.5927760601,0.1896012127,0.1313620359,0.401514262,-0.1367759109,-0.1460928917,0.1193385571,0.0144195985,-0.0813066885,0.2506349087,0.3949985504,-0.1587389559,0.0788909793,-0.2028708309,0.2833296955,-0.2204892486,0.1841510683,-0.1222558171,0.1910018027,-0.0857133567,0.2090443075,-0.0596853197,0.2083201408,0.0718123764,-0.2013075501,-0.0888202116,-0.0001829596,-0.0580069907,0.0415586531,-0.0736014992,0.6245207787,0.131835565,0.1753447503,0.0869020745,-0.1592784822,0.183907643,0.0403068922,0.224851951,0.1441524029,-0.2167216688,0.1089039966,0.2697381079,0.2915846109,-0.0308599882,0.1096391231,-0.107437022,-0.41546157,0.2698237002,0.0131871272,0.4077474773,-0.1028263122,-0.62843436,0.0436515138,0.1092530116,0.1674703658,-0.1276319921,0.1536726505,-0.2733654976,0.013872534,0.0132438513,0.32657969,-0.1094577163,0.206694752,0.0036723863,0.0926430076,-0.2967072725,-0.2458346933,0.3576645553,-0.1119512916,0.1324940771,0.2963950634,-0.1462853998,-0.01702445,0.0105879866,-0.1995972842,-0.1029506847,0.439473182,0.2255644649,0.3950291276,-0.1827936321,0.0152565576,-0.0298757032,0.1933602542,-0.1561779529,-0.0408560708,-0.0541052632,0.2324374616,-0.3855369985,-0.1968738139,-0.039049834,-0.2038401216,-0.0327578783,0.1615654677,0.2465070188,-0.2326217592,-0.2305738032,-0.3819427192,-0.0289984588,0.5887625813,-0.2635556459,0.2373927385,-0.0847060829,-0.328499347,0.4486765862,0.1127441376,0.1913455874,0.1516014636,-0.4607596099,0.0760051906,0.0519300029,-0.4398525655,-0.5598500967,0.4142220914,-0.0977686942,0.1617045999,0.1845384538,0.1501137912,0.3036822379,-0.0325751416,-0.0825505555,0.4031662345,-0.0779337958,0.1080535129,-0.1352510601,-0.134622559,0.201709345,0.168760702,0.1124669537,0.0058825724,0.1826209873,-0.1374899149,0.298524797,0.2201125473,0.0792058855,0.2166628391,0.2818375528,0.1327269375,0.1090106666,-0.1992979348,-0.0353883244,0.2307479382,-0.068367891,-0.1274635792,-0.1719899923,-0.2560539246,-0.4767713547,0.0034119347,-0.2548181415,-0.193141818,0.1301574707,0.0973193944,0.1700637639,0.1048899293,0.1259745061,0.1654153466,0.0545962192,0.0097603537,0.1804260761,0.302990526,-0.2577700615,-0.0457421467,0.0831682235,-0.0594857298,0.2411195934,0.0036581284,-0.0234827735,0.1238250807,0.0947378799,-0.0369792469,0.0112583535,-0.0961492285,-0.0187985655,-0.0046684532,0.208516553,0.5677586794,0.1324958503,-0.0747223124,0.0045230803,0.1056594327,-0.1307276189,-0.0183902904,-0.0695005655,0.2344380766,0.2568475604,-0.3220861256,-0.0040925732,-0.0434314571,0.2829574049,0.0912019163,-0.0603718944,-0.0726193562,-0.1558505297,0.1877027154,0.700084269,-0.1830890179,0.3046873808,0.1612100303,-0.1223507747,0.1930639148,-0.137361154,0.2834716737,0.7158129215,0.0091603678,0.0177161265,0.2950727344,0.002406968,-0.1490684301,0.2080704123,0.0997634754,-0.2241471559,0.3055584729,-0.0021933594,0.2467545569,-0.3240906596,-0.2490481883,0.0142765054,0.1652551144,-0.2981254756,0.098265931,-0.1337950528,-0.2373199612,-0.4030925632,0.0705316141,-0.0321190171,-0.0354837701,0.0339757949,0.1372851282,-0.0919834822,-0.1605060548,-0.3157018423,0.1221830696,0.1355545223,-0.1588221341,-0.0273725614,0.2578795254,-0.2540774345,0.1120541915,0.2056393325,-0.0287933052,0.5603420734,-0.496091783,0.2043798715,-0.4564227164,-0.2349586785,0.0054968395,0.0737605169,0.429425329,0.1740064919,0.074083142,0.0736172348,0.0474765077,0.1184441224,-0.2257839143,-0.2208357453,0.1726488918,-0.0539861992,-0.4455712736,-0.3471133113,-0.2787243128,-0.2446427345,-0.3425118625,0.228518784,0.1010006294,0.1010874212,-0.1100757569,0.0160217341,-0.098175019,-0.0168962702,0.1484609544,-0.0863984153,0.1665725112,0.4026210904,-0.0914862081,-0.6598010659,0.226908803,-0.0615841225,-0.2246567756,0.2210018188,-0.4095596969,-0.1262203753,-0.342110157,0.129096806,0.0070878975,0.2367419153,0.2802664936,-0.2545440495,0.0095087094,-0.2566065192,-0.113554813,-0.3147451878,-0.0000580463,0.1881667525,0.0922107473,0.2084324509,0.0313312039,0.1634161323,0.3067245483,-0.0213810727,0.3113523424,-0.4172437489,0.4244993925,0.0466330908,-0.2214517891,0.0983962864,-0.2498928308,0.1606142521,0.3106133938,-0.0131996665,-0.0498652756,-0.0531375594,-0.1525879204,-0.2962233722,-0.1872186512,0.1845183671,0.1975913942,0.0270271003,0.1604973674,0.2584456503,0.0593817346,-0.1810683012,0.1808519512,0.5908366442,-0.2561921775,0.1567899436,0.0854340792,-0.1405245364,-0.3897050917,0.0586795174,-0.048919417,0.3696776927,0.0397022516,0.092400454,0.183471337,-0.1580636948,0.7567191124,-0.2351808399,0.335219115,0.0931395069,0.0759490356,-0.3002094626,-0.2571650147,-0.5032290816,0.1778117418,0.1313455105,0.1499896944,-0.0213212427,0.1512990743,0.0676208735,0.2166710496,-0.1317122281,-0.4037706852,-0.2964296639,-0.6127716899,-0.2049366981,0.117291458,0.3299233913,-0.0507941172,-0.3962244689,-0.2173570246,-0.0589839593,0.1337932795,0.0097574433,-0.0469903946,0.1613981575,-0.0620574243,0.354998976,-0.2106681168,0.2135154009,0.0068420973,0.6776823997,0.0515572429,-0.20278126,0.1849692166,-0.4547348022,-0.075928323,0.1978503317,-0.2219791412,-0.181578353,-0.2288752943,0.0577406138,-0.1304596514,-0.1154361218,0.6724388599,0.0116147408,-0.0880290642,-0.2290315926,0.0269637574,0.0930446237,-0.1775280386,0.1311566979,0.0083262734,0.0038521371,-0.0504921637,-0.3116265535,0.5648178458,0.1825739145,0.2294044197,0.25097996,-0.6363233924,0.132541433,0.1731054485,-0.0348364413,-0.3449183702,-0.1180073246,-0.060735397,-0.0839627385,0.1357378662,0.0965342596,-0.2864951193,0.2465907186,-0.0923556387,0.1755324602,0.2961422801,-0.053293705,-0.4048813879,0.0279879328,0.0016065218,0.1267795116,0.2139146924,0.5632892847,-0.1866210252,-0.0936785042,-0.0864489675,-0.2423260212,-0.4809702337,0.2019794881,-0.1480094939,-0.0914589316,-0.0404894277,0.2275454849,0.2673154175,0.1050150692,0.0534894355,0.1554015577,-0.2471711636,0.09378694,0.2429651618,-0.163590923,0.3407642841,-0.1084667817,0.1350386739,-0.2384069264,-0.2937361002,-0.2282719314,0.0748440102,0.0276120156,-0.116148755,0.2478477508,0.0971163362,-0.2641325593,-0.1859139502,-0.4118072689,0.0480490588,-0.0216537248,0.1968362927,0.0426948182,0.2514934838,0.5543159842,-0.2680101395,-0.2960169911,-0.1482276917,0.316008538,-0.0751970038,-0.0281304382,0.0849173367,-0.0338713862,-0.2844862938,-0.2949290574,-0.1094078198,-0.3924708068,-0.4937892854,0.0646348968,0.0028452519,0.2016657442,0.0639257208,-0.0053070709,0.0866598338,-0.1805602014,-0.0810055435,-0.6708059907,0.1208183914,-0.2188697308,0.3121037781,-0.0310322922,-0.080994457,-0.0924195573,0.462344259,-0.56297791,-0.3270189762,0.0689965785,0.0888255239,-0.0834360719,0.109193787,0.139086917,0.1852610856,-0.0086610345,0.0958006084,0.0629086271,-0.1129151657,-0.2175378799,-0.239239648,0.123343803,-0.0251545962,-0.1054501235,0.3249388933,-0.3739489019,-0.1129374206,0.2286068797,0.0371736549,0.2270472944,-0.1074041575,0.0664063394,0.1346448511,0.1107195541,-0.1786526144,0.2178373933,-0.4177097082,0.502301991,-0.0636654794,0.2077580541,0.0905166268,0.0062949709,-0.042831663,-0.2727173567,0.0531402491,0.3122295439,0.3746222556,-0.1874159575,0.2840179205,-0.0366584845,0.2948142886,0.4220358133,-0.3256885707,-0.1145882607,0.2006989568,0.2581072748,-0.3778952956,-0.1126070246,0.0249982718,0.0281744618,-0.0173770655,0.0426898114,0.1704592407,-0.3231846988,-0.0715658963,-0.0891949311,0.510776937,0.3488326073,-0.3181062639,0.4961268604,-0.0357793383,0.2045413852,0.1308365911,0.0022999949,-0.123532027,0.466999203,0.1138676703,-0.0657802895,-0.07365264,-0.0520526879,-0.0127656963,-0.3265745342,0.0969003737,-0.2036779374,0.0883152783,0.1655903906,0.2275578827,0.1800905913,-0.1424449831,0.1382020563,-0.3902211487,-0.1678872854,0.0352386832,0.1181734353,-0.3001885116,0.0077317753,-0.1403809637,0.1666193604,0.03115898,0.0294490382,0.3884918988,0.1017772108,-0.2867660522,-0.2919543087,-0.065970242,-0.1406376809,0.1262633651,-0.3228229582,0.0812824517,0.1998783797,0.0639883801,0.1520850211,0.1378840059,0.6043738127,0.4262299836,-0.1217401624,-0.052427981,0.0836070701,-0.035255909,-0.2587223053,0.4822923541,0.1293968409,0.2001360506,0.219263345,0.1742515415,-0.2169545889,0.1601615399,-0.0443608537,-0.0834314376,0.0095301662,0.1602160633,-0.226329416,0.0906851217,-0.4063223898,-0.0822347403,-0.356235534,-0.2838271558,0.1814249009,-0.0034272557,0.1390607953,-0.5132735968,0.0922186598,0.0056273444,0.2130190283,0.3403390646,0.3399863839,-0.1434624493,-0.0310119316,-0.5878449082,0.4222592115,-0.1601106822,-0.469481945,-0.0580243543,-0.074073568,-0.4996267855,-0.0051813284,0.3259088099,0.3747652471,-0.034929134,0.2777367532,-0.2828331888,-0.2070655376,-0.0261246599,0.1599916518,0.1422616839,-0.6948549747,0.2628783882,-0.3964836597,0.1354710013,-0.3572916687,-0.1515267193,0.0148491617,-0.0411220454,0.4819484055,0.0784450695,0.5681082606,-0.2400528491,-0.1901470125,-0.3536834717,0.150347814,-0.1731881648,0.2326658219,0.2059119046,-0.2228829563,0.0862541646,-0.1185080707,-0.0339981914,0.2086851001,-0.0462474152,0.2048280835,-0.1679049134,0.1487743109,-0.1486237049,0.0804283768,-0.0019246727,0.2268226743,-0.4055153131,-0.2166316658,-0.0701380074,-0.4929824471,0.2949214578,-0.4310215712,-0.2514631152,-0.1839419752,0.2517825961,-0.3235113621,0.0435998961,-0.0431705825,0.1134158745,0.1523936093,-0.0991116017,-0.0711204708,0.2593789101,-0.3562481999,-0.0355296209,0.2653440535,0.4769489169,-0.0163087025,-0.180393517,0.3521855772,-0.1993495673]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2917","title":"windows download abnormal","comments":"Let me know if you have other questions :)\r\n\r\nClosing this issue now","body":"## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n","comment_length":13,"text":"windows download abnormal\n## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n\nLet me know if you have other questions :)\r\n\r\nClosing this issue now","embeddings":[-0.2207755297,-0.1332957745,-0.0691334903,0.1743752807,0.2265743613,-0.0526336357,-0.0071485089,0.1346075088,0.2471247762,0.0947366357,0.1111562997,0.2011167854,0.0616793036,0.1772477627,0.0522657,-0.2514466345,0.0436070114,0.1884659082,-0.1824369282,-0.0201365389,-0.3421658576,0.1442255229,-0.2688935101,-0.0980346352,0.1254004985,0.0039493642,-0.3086517155,0.2426276207,-0.2599076033,-0.1714349836,0.0742157772,-0.0879949704,0.163236618,0.2892299891,-0.0001064452,0.0440166481,0.3960578144,0.0089306133,0.0559709035,-0.0490174629,-0.2207491994,-0.1396053284,-0.2216101736,-0.293787092,0.1231971383,0.110559687,0.0049971924,-0.0325602069,0.0923154652,0.3783936501,0.3404253125,0.4842950702,-0.0722764954,-0.0313038714,0.4031114578,0.1105747074,-0.1911314577,0.1367606372,0.2970518172,-0.2148809433,0.1355500519,-0.0241315197,-0.0474884994,0.0302238148,-0.0182570498,0.0709910616,-0.0401855186,-0.4947353899,0.1354618669,0.1454054117,0.3116519451,-0.1623973399,-0.1045670062,0.1049401239,-0.0483258218,-0.2304380387,0.1805960536,0.3206201792,-0.1367184669,0.1314445287,-0.2873217762,0.2481869012,-0.2228381038,0.1704220772,-0.0978163704,0.2842999995,-0.0906764865,0.1684194952,-0.0802360848,0.2447992563,0.0083303163,-0.2286545932,-0.100851655,0.0021302979,-0.0380191021,0.0297235865,0.0100703146,0.5421230197,0.129743889,0.2278441787,-0.0172666032,-0.1089707017,0.2379163802,0.0550119728,0.1292164624,0.1321160793,-0.1616736948,0.0282277036,0.3516741991,0.2831096351,0.0294332672,0.1096406654,-0.0448804945,-0.409791708,0.3063461781,0.067747876,0.3790406585,-0.1142017618,-0.6207893491,0.0279846601,0.070389539,0.1223948523,-0.0729451403,0.2053755671,-0.2115510255,-0.038713377,0.0287676323,0.3363829255,-0.1473443359,0.1672639549,-0.0085347267,0.1934087873,-0.2336297184,-0.2789251506,0.4062646925,-0.1174081489,0.1876767725,0.2447303534,-0.1050559208,0.0019086151,-0.0402509943,-0.1226891279,-0.0408820026,0.4309408367,0.2251872271,0.4080506861,-0.1081939563,0.1009383127,-0.0526482053,0.2426031977,-0.1097498462,-0.0552099682,-0.0306426436,0.2899795175,-0.3140995502,-0.2293543965,-0.0028886429,-0.2081403881,-0.0288612023,0.0473718569,0.2172876298,-0.2606675923,-0.2903537452,-0.4342307448,0.0709750429,0.4895909131,-0.4144755304,0.227134794,0.0270461105,-0.3316147923,0.4150057137,0.1907803267,0.1251585186,0.099305965,-0.4734765291,0.1356181353,0.0067809615,-0.3681674302,-0.4561111033,0.3584577441,-0.1014415994,0.1938920319,0.1865333468,0.0966966599,0.288595438,-0.0928407386,-0.0392433889,0.2651661932,-0.0274843816,0.0421565957,-0.1506972611,-0.1795333475,0.1650287658,0.1755206734,0.0803700909,0.0484602489,0.2923360765,-0.1043607593,0.3381632268,0.1145961881,0.0599303469,0.2569989264,0.3408920169,0.1021471545,0.1144763753,-0.243753314,-0.0720155612,0.2930027843,-0.0744742453,-0.1432997733,-0.2104561776,-0.2281522602,-0.4664123952,0.0491136797,-0.2203699201,-0.2093744129,0.1927045584,0.105987832,0.2294083834,0.0463595949,0.0988513902,0.1179435104,-0.0049782838,-0.0852696747,0.1478122175,0.2976120412,-0.358138293,-0.0983111262,0.0428683013,-0.0618663542,0.2359926403,-0.0797711536,0.0072130114,0.1748583913,0.0909606516,0.0120516652,-0.0225844402,-0.0551137365,-0.0018050423,-0.0328586437,0.2095954716,0.4951538742,0.1691873521,-0.1198024824,-0.0891863182,0.0866177082,-0.056391336,-0.0603130832,-0.0616283976,0.135127157,0.2473331243,-0.2254518121,-0.0318140052,-0.0709816664,0.4197178185,0.1235188469,-0.1139750704,-0.1175652519,-0.1888785362,0.154271245,0.6313266158,-0.2423981279,0.2632025778,0.0781768337,-0.0857005417,0.1904331446,-0.1035305858,0.2065784037,0.6523746848,0.0705253407,0.0944053531,0.3197440505,-0.0311157349,-0.1489898115,0.2229828387,0.0814782307,-0.1967646033,0.2652468383,-0.0275604334,0.1729865074,-0.3210701644,-0.1806147695,0.0191281699,0.1348359287,-0.3165897727,0.1074841619,-0.1814183295,-0.0536290482,-0.3422429264,-0.0320985392,-0.0447294153,-0.053749226,0.0935111567,0.2653478086,-0.1439779252,-0.1548583955,-0.2668774128,0.0174200926,0.2345918864,-0.2694985569,-0.021260269,0.218075648,-0.2366104424,0.1633965224,0.1720452756,-0.0932388231,0.5667923689,-0.4775111079,0.1574508995,-0.4538096488,-0.1942285001,-0.0239324756,-0.0515272506,0.3197554946,0.2340763509,0.1066003665,0.0461247899,0.0562638193,0.1573544741,-0.2489885539,-0.2195565104,0.2129605412,0.0073923734,-0.4503124058,-0.3328312635,-0.2961428463,-0.202512458,-0.3447829485,0.1225349307,0.1772155315,0.1006463096,-0.0672200993,0.0358825438,-0.0186522044,0.047959242,0.0959775299,-0.1114475951,0.1012131572,0.4370254278,-0.108312875,-0.668559134,0.2404889315,0.0451720208,-0.2690183222,0.241299063,-0.5037426949,-0.088098526,-0.3273606896,0.2437198013,0.0548889041,0.2400947064,0.2727158964,-0.2193379253,-0.0504980423,-0.2639805079,-0.0492048077,-0.2931634486,-0.0319044366,0.1824257225,0.0572677068,0.2499418408,0.0213780347,0.2403487414,0.2202055752,0.0174942203,0.3821138442,-0.3713877201,0.4587187171,-0.0153283775,-0.1699820757,0.0924867392,-0.2480327785,0.0673405975,0.2961802781,-0.0331741124,-0.0549514927,-0.1332375258,-0.1691046953,-0.1654629558,-0.1967912763,0.1489178687,0.1905439347,-0.0007379284,0.1263190657,0.2819885314,0.126558423,-0.1203787774,0.1470764428,0.529684782,-0.2528937459,0.1417438388,-0.0430495441,-0.1875661463,-0.4491520822,0.0247008894,0.028167218,0.381575644,0.0108286748,0.0817865953,0.1806859076,-0.1728951335,0.7310193181,-0.1994722933,0.3743469417,0.131025508,-0.0181296635,-0.3187329769,-0.2922997177,-0.5790176392,0.1523871869,0.0467407405,0.0785343796,-0.0655063391,0.1080228686,0.1322008371,0.3294463456,-0.1184682176,-0.4172126651,-0.2886999547,-0.5888552666,-0.3001114428,0.0594724566,0.4062659144,0.000087206,-0.4491736591,-0.1950608939,-0.0533062816,0.0927784741,-0.0568115413,-0.0113462741,0.1471142918,-0.0048097325,0.2928668261,-0.2623037398,0.2470901459,-0.028746685,0.6279499531,0.070920743,-0.1808263958,0.1777650118,-0.3867741823,-0.1477529556,0.143608436,-0.2414682657,-0.2455293983,-0.1764716357,0.0498459563,-0.2587479949,-0.0836134553,0.6548749208,-0.0413557142,-0.0423992425,-0.1491366029,0.1900945753,0.0426516347,-0.1285643429,0.2071285695,0.0451181121,-0.0856232122,-0.0590202399,-0.2113350332,0.5505878925,0.2270111889,0.2372392416,0.2324950695,-0.585196197,0.0767249316,0.2548376322,0.0133874621,-0.3171902597,-0.2484219521,-0.0332701951,-0.0792204291,0.1030970961,0.081428133,-0.2686470747,0.2291331589,-0.0590135045,0.1270036101,0.2279467285,-0.1004103422,-0.2982287109,0.0521233603,0.0186875034,0.1960546076,0.2562512159,0.5732926726,-0.2369992733,-0.0070824078,-0.094170548,-0.2494924664,-0.4075240493,0.2002867758,-0.1548224837,0.0800937116,-0.1750008166,0.0427773744,0.0968228355,0.0378658399,-0.0155429896,0.1895344555,-0.2424435318,0.0958389044,0.2168214917,-0.1496645361,0.2837164998,-0.0939572006,0.1009400859,-0.221022293,-0.3328936398,-0.1587910354,0.0103745395,0.0131876329,-0.0436179899,0.2026299983,0.0993676856,-0.1678388119,-0.2867918909,-0.3931019604,0.0494828448,-0.059844099,0.2332053334,0.0822550058,0.1473239809,0.5253173113,-0.2927761972,-0.2554986179,-0.177690357,0.3423021436,-0.1086063087,-0.081868045,0.123623386,0.1187664419,-0.275470823,-0.3161422312,-0.0015436411,-0.3686634302,-0.3794449866,0.0258730557,0.0202567708,0.286070317,-0.0100809755,0.0859160051,0.1481716931,-0.1539181918,-0.0308849066,-0.6326345801,0.0793285519,-0.1875317991,0.3171585798,0.0826699287,0.0632378832,-0.1279541105,0.4226350784,-0.5515289903,-0.3977765143,0.0291253291,0.1112820432,-0.1331367344,0.1548172086,0.1501194984,0.2122796327,0.0203980245,0.1442107111,0.0092959199,-0.0687797889,-0.2692624331,-0.1742177755,0.0721313804,-0.0414409004,-0.1086429581,0.275123924,-0.4210032523,-0.0978210792,0.1876570284,0.17270872,0.2825715542,-0.0950304568,0.1360922009,0.1394856572,0.0868963525,-0.1750506014,0.1824970245,-0.4027463198,0.442858845,-0.0126204034,0.1944123656,0.1665093303,0.0109840715,-0.0307105407,-0.2862520814,-0.0017864812,0.2384109199,0.3138039112,-0.1555044502,0.2605932653,-0.0349944904,0.3879834116,0.4773438573,-0.3409121931,-0.1270853281,0.1898004264,0.3115430474,-0.3598316312,-0.1136988103,-0.0361874923,0.1267793924,0.0516493917,0.1107208356,0.0998693407,-0.2879479527,-0.0161045827,-0.0412315056,0.4778434932,0.3516988456,-0.2632001042,0.5142486095,-0.0232837703,0.1377200931,0.0572274812,0.0515077263,-0.0481213368,0.4958332777,0.0732803345,0.0498245247,-0.0417858437,-0.0422224067,-0.0472286642,-0.1962688118,0.098315455,-0.1797437668,0.075355649,0.1101114526,0.1492832899,0.0244431049,-0.221565485,0.1172366291,-0.3538033366,-0.0532904789,-0.0065229121,0.0325645544,-0.2743485868,-0.0475490838,-0.0886465758,0.1344543397,0.0166579895,0.0213944465,0.2973858118,0.1095635965,-0.2179983556,-0.3157918155,0.0369854607,-0.1580412835,0.0982746109,-0.2871119678,0.040060088,0.3295312226,0.0932035521,0.1183491424,0.134796381,0.5907421708,0.4009403586,-0.1221763343,-0.0771894902,0.0635202974,-0.0590710416,-0.2384877056,0.4153927267,0.039325472,0.1890621036,0.3091205657,0.2307442129,-0.2469318658,0.1552436352,0.080356203,-0.1620993912,0.020642532,0.1388530433,-0.2244401574,0.0271097682,-0.3510095477,-0.1198782027,-0.3236604631,-0.2240209132,0.1686892658,-0.0143957799,0.1356601864,-0.3788411021,0.1349952817,0.002670008,0.2053229809,0.3138542473,0.272441268,-0.1191345304,-0.0695681348,-0.5928861499,0.3832478225,-0.1678831279,-0.4602458775,-0.0931928083,0.0156863183,-0.4653169513,-0.0051278067,0.3414212167,0.2859823406,-0.0933431908,0.2965569794,-0.2014313191,-0.192971006,0.0049811103,0.0661877319,0.1054957882,-0.5753059983,0.318833977,-0.3736201823,0.1851435602,-0.3877812922,-0.09884388,0.0152286161,-0.0082453201,0.5097563863,0.0851208419,0.50721246,-0.254583776,-0.1529332697,-0.3806297183,0.0885771438,-0.1390194148,0.2699479163,0.1687265784,-0.13552773,0.075280495,-0.2353149801,-0.0574993305,0.2708546817,-0.0405838639,0.1234099939,-0.1145117506,0.1258966476,-0.1428504884,0.0675647482,0.0728959367,0.1774835438,-0.3727541566,-0.1398797184,-0.0954014957,-0.5806180835,0.2051665485,-0.2872015834,-0.2997911274,-0.1383069307,0.1946542859,-0.3773559034,0.0320957825,-0.111780636,0.1643281281,0.1397599727,-0.1000155509,-0.0343191288,0.2832554877,-0.3728188276,-0.0389520638,0.262173295,0.4476927817,0.0322523899,-0.2558436394,0.2842577696,-0.2356747538]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2914","title":"Having a dependency defining fsspec entrypoint raises an AttributeError when importing datasets","comments":"Closed by #2915.","body":"## Describe the bug\r\nIn one of my project, I defined a custom fsspec filesystem with an entrypoint.\r\nMy guess is that by doing so, a variable named `spec` is created in the module `fsspec` (created by entering a for loop as there are entrypoints defined, see the loop in question [here](https:\/\/github.com\/intake\/filesystem_spec\/blob\/0589358d8a029ed6b60d031018f52be2eb721291\/fsspec\/__init__.py#L55)).\r\nSo that `fsspec.spec`, that was previously referring to the `spec` submodule, is now referring to that `spec` variable.\r\nThis make the import of datasets failing as it is using that `fsspec.spec`.\r\n\r\n## Steps to reproduce the bug\r\nI could reproduce the bug with a dummy poetry project.\r\n\r\nHere is the pyproject.toml:\r\n```toml\r\n[tool.poetry]\r\nname = \"debug-datasets\"\r\nversion = \"0.1.0\"\r\ndescription = \"\"\r\nauthors = [\"Pierre Godard\"]\r\n\r\n[tool.poetry.dependencies]\r\npython = \"^3.8\"\r\ndatasets = \"^1.11.0\"\r\n\r\n[tool.poetry.dev-dependencies]\r\n\r\n[build-system]\r\nrequires = [\"poetry-core>=1.0.0\"]\r\nbuild-backend = \"poetry.core.masonry.api\"\r\n\r\n[tool.poetry.plugins.\"fsspec.specs\"]\r\n\"file2\" = \"fsspec.implementations.local.LocalFileSystem\"\r\n```\r\n\r\nThe only other file being a `debug_datasets\/__init__.py` empty file.\r\n\r\nThe overall structure of the project is as follows:\r\n```\r\n.\r\n\u251c\u2500\u2500 pyproject.toml\r\n\u2514\u2500\u2500 debug_datasets\r\n \u2514\u2500\u2500 __init__.py\r\n```\r\n\r\nThen, within the project folder run:\r\n\r\n```\r\npoetry install\r\npoetry run python\r\n```\r\n\r\nAnd in the python interpreter, try to import `datasets`:\r\n\r\n```\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nThe import should run successfully.\r\n\r\n## Actual results\r\n\r\nHere is the trace of the error I get:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/__init__.py\", line 33, in \r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 48, in \r\n from .filesystems import extract_path_from_uri, is_remote_filesystem\r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/filesystems\/__init__.py\", line 30, in \r\n def is_remote_filesystem(fs: fsspec.spec.AbstractFileSystem) -> bool:\r\nAttributeError: 'EntryPoint' object has no attribute 'AbstractFileSystem'\r\n```\r\n\r\n## Suggested fix\r\n\r\n`datasets\/filesystems\/__init__.py`, line 30, replace:\r\n```\r\n def is_remote_filesystem(fs: fsspec.spec.AbstractFileSystem) -> bool:\r\n```\r\nby:\r\n```\r\n def is_remote_filesystem(fs: fsspec.AbstractFileSystem) -> bool:\r\n```\r\n\r\nI will come up with a PR soon if this effectively solves the issue.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: WSL2 (Ubuntu 20.04.1 LTS)\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n- `fsspec` version: 2021.8.1\r\n","comment_length":3,"text":"Having a dependency defining fsspec entrypoint raises an AttributeError when importing datasets\n## Describe the bug\r\nIn one of my project, I defined a custom fsspec filesystem with an entrypoint.\r\nMy guess is that by doing so, a variable named `spec` is created in the module `fsspec` (created by entering a for loop as there are entrypoints defined, see the loop in question [here](https:\/\/github.com\/intake\/filesystem_spec\/blob\/0589358d8a029ed6b60d031018f52be2eb721291\/fsspec\/__init__.py#L55)).\r\nSo that `fsspec.spec`, that was previously referring to the `spec` submodule, is now referring to that `spec` variable.\r\nThis make the import of datasets failing as it is using that `fsspec.spec`.\r\n\r\n## Steps to reproduce the bug\r\nI could reproduce the bug with a dummy poetry project.\r\n\r\nHere is the pyproject.toml:\r\n```toml\r\n[tool.poetry]\r\nname = \"debug-datasets\"\r\nversion = \"0.1.0\"\r\ndescription = \"\"\r\nauthors = [\"Pierre Godard\"]\r\n\r\n[tool.poetry.dependencies]\r\npython = \"^3.8\"\r\ndatasets = \"^1.11.0\"\r\n\r\n[tool.poetry.dev-dependencies]\r\n\r\n[build-system]\r\nrequires = [\"poetry-core>=1.0.0\"]\r\nbuild-backend = \"poetry.core.masonry.api\"\r\n\r\n[tool.poetry.plugins.\"fsspec.specs\"]\r\n\"file2\" = \"fsspec.implementations.local.LocalFileSystem\"\r\n```\r\n\r\nThe only other file being a `debug_datasets\/__init__.py` empty file.\r\n\r\nThe overall structure of the project is as follows:\r\n```\r\n.\r\n\u251c\u2500\u2500 pyproject.toml\r\n\u2514\u2500\u2500 debug_datasets\r\n \u2514\u2500\u2500 __init__.py\r\n```\r\n\r\nThen, within the project folder run:\r\n\r\n```\r\npoetry install\r\npoetry run python\r\n```\r\n\r\nAnd in the python interpreter, try to import `datasets`:\r\n\r\n```\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nThe import should run successfully.\r\n\r\n## Actual results\r\n\r\nHere is the trace of the error I get:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/__init__.py\", line 33, in \r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 48, in \r\n from .filesystems import extract_path_from_uri, is_remote_filesystem\r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/filesystems\/__init__.py\", line 30, in \r\n def is_remote_filesystem(fs: fsspec.spec.AbstractFileSystem) -> bool:\r\nAttributeError: 'EntryPoint' object has no attribute 'AbstractFileSystem'\r\n```\r\n\r\n## Suggested fix\r\n\r\n`datasets\/filesystems\/__init__.py`, line 30, replace:\r\n```\r\n def is_remote_filesystem(fs: fsspec.spec.AbstractFileSystem) -> bool:\r\n```\r\nby:\r\n```\r\n def is_remote_filesystem(fs: fsspec.AbstractFileSystem) -> bool:\r\n```\r\n\r\nI will come up with a PR soon if this effectively solves the issue.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: WSL2 (Ubuntu 20.04.1 LTS)\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n- `fsspec` version: 2021.8.1\r\n\nClosed by #2915.","embeddings":[-0.0430343486,0.1729615629,-0.0092811715,0.3249996006,0.3824198246,-0.027578136,0.3696733415,-0.0281330235,0.0717082471,0.1155877039,0.2619345784,0.4130678475,-0.0622646138,0.0250749756,-0.1028326377,0.2277256399,0.0461320765,0.1964916587,-0.1118648723,0.0814422593,-0.2052113712,0.284779042,-0.20766671,0.1609047055,-0.3610854745,0.1328371316,0.0609395057,0.3386956155,-0.1826843172,-0.6150017977,0.3787585497,-0.1575102955,-0.0890944004,0.1836814284,-0.0001192106,0.0413611904,0.1138052121,-0.081654422,-0.2293013632,0.0555289909,-0.5713313818,-0.1993799657,0.1814039797,-0.2640390098,0.4023723602,-0.2299601138,-0.0492568053,-0.720643878,-0.0350703783,0.4295676649,0.151214689,-0.0737293884,0.1342175305,-0.078581512,0.1368629038,0.0073615429,-0.2163085043,0.0665525422,-0.1590744555,-0.1137120426,-0.0247463994,0.028592892,-0.073725529,0.2388720959,0.7355163693,-0.1537969112,0.4934940934,-0.2031433433,-0.0418940075,0.1256616861,0.4556665123,-0.4343681931,-0.3134965301,-0.1376075,0.0694526061,-0.099931106,0.2053408474,-0.0052408343,0.0365756415,0.1147769988,0.4034270048,-0.0220960025,-0.1103871539,0.0099158697,-0.1579840034,-0.1949465275,-0.2566277087,0.1149913594,-0.0176931862,-0.1435348243,-0.293458432,-0.1421632171,-0.1392901242,0.0056826379,-0.3321452439,0.0178288762,0.2203202993,-0.1728139669,-0.2198076099,0.2929090858,-0.0524715595,0.0291536804,0.1376225799,0.3627481163,0.1954966486,0.1123554483,0.3818732202,0.0941143036,0.2671805918,0.0105761532,-0.1745803803,-0.0067355181,-0.0322308168,-0.4421686828,0.196050778,-0.0149847884,0.7509050965,-0.3278894126,-0.3549762964,0.2934402525,-0.2196975499,0.0935777649,-0.0379396789,0.2039052695,0.2131644487,0.5074154139,0.110865429,0.1131861806,-0.2682110369,0.0728461221,-0.0765823871,0.0951074734,-0.1301978678,0.1238338947,-0.0917826369,-0.0535966307,0.3636643291,-0.0692519695,0.087492682,-0.2760293186,0.0895308852,-0.1705299467,-0.2756180465,0.268853128,-0.4858374298,0.0634655356,0.0485958643,-0.1812145114,-0.1407059431,-0.0891787559,-0.223374173,-0.2241652012,-0.2036291659,0.1713668406,-0.227433145,0.0364288799,-0.2167210281,-0.2242526561,0.2060965598,-0.0521678776,0.1591060758,-0.2531750202,-0.255607456,-0.1615860462,-0.0356961787,0.3934701979,-0.2941275835,-0.1991277188,-0.1428540945,-0.2064636052,-0.1786636561,-0.1581739336,-0.1267938614,0.4374836683,-0.2819532752,0.1414128244,0.2326729596,-0.4089822769,-0.1433600485,0.2878492177,0.0954126939,-0.1979203522,0.1566754282,-0.1284318268,0.1135182604,-0.1574372202,0.1001495048,0.0776153803,0.1633862257,0.007081199,-0.0188464597,0.036375124,-0.1544718444,-0.0118767954,-0.2443443686,0.0282867458,0.1455506384,0.1295337081,-0.1325169504,-0.1437395513,0.102992475,0.446297735,0.5053649545,0.2926414013,0.0018090364,-0.3757137656,-0.1410629153,0.2792656124,0.0050844387,-0.0705011711,-0.0173224658,-0.0369233266,-0.2272954136,0.4188956916,-0.0087850615,-0.3483618796,-0.0021392696,0.1587059945,-0.2076747417,-0.1167936027,-0.2682749331,0.0590893626,-0.1198496372,0.0904613882,-0.2748877108,0.0908167139,-0.0701375529,-0.0042610434,-0.0235086512,0.3459734023,0.1507644355,0.122721754,-0.1112476885,0.3404074907,0.5937909484,0.2090201527,-0.0751314461,0.3819218576,0.1279350668,-0.2491389513,-0.0412581488,0.1454391629,0.2145802826,-0.0491112433,0.4379881918,0.2754839659,0.2406787723,0.3727573156,-0.0556240045,0.1264895946,0.242703557,-0.1039064452,-0.1578007489,-0.2713132203,0.1646561325,-0.0192193184,0.4520537555,0.2100941837,0.0325967483,-0.0543766543,-0.0089729214,0.0672252774,0.0164770391,-0.0364091173,-0.2857769132,0.0901333243,0.2991757691,0.056403894,0.7109383345,0.0651810989,-0.3750181198,0.1294146478,-0.020994015,-0.0924102291,0.232258141,0.1908759773,0.1215177178,0.1294934303,0.2365701497,-0.1311048567,0.0034811324,-0.2910775244,-0.0515734069,0.1591396928,-0.7218433022,0.3073238134,-0.4488064051,0.0365511924,-0.3674746156,-0.1099792644,0.039912153,-0.3449760377,-0.1204983741,0.0795523301,-0.1346967071,0.1161788553,-0.3224466741,0.0531995259,-0.0138493739,-0.4038999677,0.0998631865,-0.1038146392,-0.0749545097,-0.0629806891,0.2017032653,0.4358921051,0.1518811285,0.2062558383,-0.0725508034,-0.2794296741,-0.1373124123,0.0227615349,-0.0418572798,0.7079858184,0.4984821379,-0.0236271117,0.334713608,-0.052583456,-0.022778647,-0.0508446507,-0.1497538984,0.1021313444,-0.0401481725,-0.159014374,-0.1050804704,-0.2614705265,0.0915697217,-0.2867671847,-0.0876169056,-0.0462019704,0.0119425636,0.4248128235,0.1451478451,0.4045563936,0.2662217617,-0.0025793067,-0.0106500452,0.1442047656,0.1359658688,-0.1538913995,-0.2444033176,0.2343038619,-0.2083691955,-0.0294864997,-0.0952343643,-0.4535905719,-0.25282076,-0.1953865737,0.3757030666,-0.0514751524,-0.1112330109,0.4003503621,0.165106073,-0.0839413479,-0.3391038477,-0.2688634694,0.0547733977,0.0875356421,-0.0184010845,0.1420247257,-0.1215976626,-0.1658104211,0.374792397,0.2599062324,-0.1457856596,0.2739960253,-0.1751101613,0.2564560473,0.1007364541,-0.3438697755,-0.0130711058,-0.085844934,-0.2883581519,0.1482560337,0.0481070541,0.2359811664,-0.0565045625,0.2738620937,0.261869669,-0.2000104636,-0.2162149996,-0.2536721528,0.3697989285,-0.0000891886,0.1281913221,0.1720805168,0.223276034,0.0398766212,0.3417465091,-0.0693826824,-0.1782999188,-0.2556450963,0.2239852697,-0.149227649,0.1481966078,0.118940182,0.2389065027,0.0841452926,0.0368388407,0.0327930972,-0.0729915872,0.107063055,-0.0424628891,0.0199763998,0.2870316803,0.0729298294,-0.6140675545,0.013722131,-0.0617609844,0.2754461467,0.1362575889,0.1719761193,-0.2309075296,-0.0642381907,0.0313319452,0.1695097834,-0.0079485681,0.0463472158,-0.3716210723,-0.3436375558,-0.3559315503,-0.3648687899,-0.1220872253,0.1460855603,0.608284533,0.4415367246,-0.2861590385,-0.1434668005,0.1396352649,-0.0774980411,-0.0251135901,-0.3337855041,0.2506456077,0.0822203383,-0.2227955461,0.438881278,0.7413416505,-0.0107491156,-0.6289779544,0.0093631083,0.1382879615,0.2119753063,0.3244391382,-0.2618112862,0.1188502163,-0.363063544,-0.1881043762,-0.3619200289,0.1122573316,0.260430187,0.2203682214,0.0700321048,-0.4762194157,0.1530032158,-0.1998498738,-0.0410767794,0.3159738183,0.1726385653,-0.2884833515,0.2569683492,0.0741914511,0.7808514237,0.569355309,-0.1249206737,0.3599708676,-0.2424308062,-0.0542152859,-0.1202839613,-0.1542701572,-0.2387532294,0.1913201809,-0.0544839613,-0.18723239,0.705799818,-0.1415283531,-0.1244052649,0.1290933788,0.1822909117,-0.1393857598,0.1165291592,0.1594590992,0.0814604759,-0.4633861482,-0.370397687,0.1029222086,-0.0320705138,0.0683007315,-0.0650951117,-0.0247276314,-0.0777822435,-0.2294555604,-0.32167539,0.0439463444,-0.4455278814,0.3251912594,0.1124332696,0.1910561025,0.2862206399,0.3550289273,-0.0675555989,0.1610818058,-0.1015330851,0.2352689207,0.1239409521,-0.0098448126,0.0602400079,-0.227035746,0.1665501446,-0.0625454262,-0.087164484,-0.1301974505,-0.1741476953,-0.1828138232,0.1021620706,0.0863435939,-0.1087821126,-0.3041732013,-0.1606344432,0.0950354114,0.1120003834,-0.221758917,0.0907387882,0.0084382929,-0.1957605779,0.1542771906,-0.1038391814,-0.3994420469,0.0711361542,0.2759751081,-0.1620726883,0.1611830741,0.381315589,-0.1241321787,-0.1759859473,-0.1711550504,0.4069006741,-0.2283793688,-0.2342585176,0.094457373,0.0810037404,-0.2219625711,-0.120175615,0.5494086742,-0.0067346175,-0.0117426496,0.0497230366,-0.4125763774,-0.0088035138,0.0351061262,-0.2378330827,0.0009349291,-0.0294312052,0.0923814923,-0.3538710773,0.3067423701,-0.2256021351,0.2494763881,-0.0625165701,0.1239562109,0.4566327929,0.0589139163,0.1883669645,-0.272287935,0.0895297304,0.2248979509,-0.2632036507,-0.0840326101,-0.3775589168,0.1614944786,0.0968005434,-0.0412362032,0.1767305434,-0.0410813652,-0.0104668569,-0.1951070875,-0.2089901417,0.4560409784,-0.3352125585,0.2502237558,-0.1950406581,0.1619997472,-0.1527923942,0.1199247018,0.0177820474,0.3442685306,-0.0704655424,0.2242517322,-0.1402976066,0.0524093583,-0.1809961349,0.286863625,-0.2309779823,0.001652966,0.0643716156,-0.2714308202,0.0075101843,0.1741512418,0.051411096,0.3920331597,-0.3538575768,0.2180334181,0.1009309292,0.0863336474,-0.2097055018,0.0577151366,0.2425736338,-0.2602337599,0.0286733247,0.32045421,0.3694347143,-0.1328757405,-0.1426737905,0.1771979332,-0.1642621458,0.1670656502,0.2125457823,0.0809638053,-0.1030210778,-0.0222702622,0.3884674013,0.1377125829,0.2725866735,0.6721334457,-0.0822553411,0.4087929726,0.3187718987,0.0578281432,-0.0882372782,-0.3026813567,0.6395866871,0.0971084386,0.0903770179,0.0212999135,0.1801173836,0.3370145261,-0.2536561191,-0.0584542826,0.1592061371,-0.0090570366,-0.0695601627,-0.1169230044,0.2838150859,0.0470363572,0.0753637478,-0.1204745993,0.099752076,-0.2900804579,0.4320870042,-0.0070109242,-0.2397236079,-0.4791888297,0.2327571511,0.0517540872,0.2455652803,-0.3497766256,0.0935187489,-0.2294338495,0.009274709,0.2234476656,0.2447924465,0.6747469306,0.2802542746,0.0613406897,-0.0361753739,-0.1880470365,-0.0251632184,0.1080025136,0.0423466004,-0.4021561146,0.4986065328,0.3744322956,0.0134312417,-0.1475075185,-0.3044832051,0.2202391624,0.3330422342,-0.5749848485,0.4656461477,-0.1772824079,0.2455034703,-0.0323761702,0.301497072,-0.2611684501,-0.3874022961,0.4785009027,-0.0177927166,0.0497761108,-0.1358003616,0.0600086786,-0.1562406123,0.4533146918,0.2446237355,0.0289993435,-0.0898260772,0.0194680337,-0.7921037078,0.09198457,-0.0585266203,-0.1397465765,0.0542413443,0.132478714,0.0207924377,0.3848239481,-0.055609718,0.0733580142,0.3911984265,0.2168242484,-0.3189651668,-0.1849366575,0.1098671705,-0.0723528266,0.1301362216,-0.5238112211,-0.1766480356,0.0221980996,0.0278460272,0.0507279634,-0.2994850278,0.2058833688,-0.1167204082,0.4986041486,-0.0291798674,0.5323891044,-0.144431591,-0.0973311812,-0.1218846962,-0.4071162939,-0.2981791794,0.0227600001,-0.027476877,0.1542173773,-0.233344689,-0.0661291555,-0.3824612796,0.4765124917,0.0400805809,-0.0109937387,-0.3236443698,0.3815689981,-0.4507586658,0.1908286214,-0.4913522601,0.1234193966,-0.0205241553,0.3339468837,-0.2606340647,-0.5488712192,0.4859133363,-0.534476161,-0.1120796576,0.1190102845,0.4382752478,0.0190124456,-0.2618913352,-0.7699956894,0.0116949342,0.3630550206,-0.0525717586,-0.2893211842,0.0585939176,-0.398167491,0.0147053022,-0.050671231,0.4721556902,-0.1411389709,0.0548643433,0.3597792387,-0.139134869]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2913","title":"timit_asr dataset only includes one text phrase","comments":"Hi @margotwagner, \r\nThis bug was fixed in #1995. Upgrading the datasets should work (min v1.8.0 ideally)","body":"## Describe the bug\r\nThe dataset 'timit_asr' only includes one text phrase. It only includes the transcription \"Would such an act of refusal be useful?\" multiple times rather than different phrases.\r\n\r\n## Steps to reproduce the bug\r\nNote: I am following the tutorial https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\n1. Install the dataset and other packages\r\n```python\r\n!pip install datasets>=1.5.0\r\n!pip install transformers==4.4.0\r\n!pip install soundfile\r\n!pip install jiwer\r\n```\r\n2. Load the dataset\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n```\r\n3. Remove columns that we don't want\r\n```python\r\ntimit = timit.remove_columns([\"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \"sentence_type\", \"speaker_id\"])\r\n```\r\n4. Write a short function to display some random samples of the dataset.\r\n```python\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n \r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\nshow_random_elements(timit[\"train\"].remove_columns([\"file\"]))\r\n```\r\n\r\n## Expected results\r\n10 random different transcription phrases.\r\n\r\n## Actual results\r\n10 of the same transcription phrase \"Would such an act of refusal be useful?\"\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.4.1\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: not listed\r\n","comment_length":16,"text":"timit_asr dataset only includes one text phrase\n## Describe the bug\r\nThe dataset 'timit_asr' only includes one text phrase. It only includes the transcription \"Would such an act of refusal be useful?\" multiple times rather than different phrases.\r\n\r\n## Steps to reproduce the bug\r\nNote: I am following the tutorial https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\n1. Install the dataset and other packages\r\n```python\r\n!pip install datasets>=1.5.0\r\n!pip install transformers==4.4.0\r\n!pip install soundfile\r\n!pip install jiwer\r\n```\r\n2. Load the dataset\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n```\r\n3. Remove columns that we don't want\r\n```python\r\ntimit = timit.remove_columns([\"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \"sentence_type\", \"speaker_id\"])\r\n```\r\n4. Write a short function to display some random samples of the dataset.\r\n```python\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n \r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\nshow_random_elements(timit[\"train\"].remove_columns([\"file\"]))\r\n```\r\n\r\n## Expected results\r\n10 random different transcription phrases.\r\n\r\n## Actual results\r\n10 of the same transcription phrase \"Would such an act of refusal be useful?\"\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.4.1\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: not listed\r\n\nHi @margotwagner, \r\nThis bug was fixed in #1995. Upgrading the datasets should work (min v1.8.0 ideally)","embeddings":[0.161290288,-0.138053894,-0.0205535926,0.2273859978,0.1176241562,-0.1575979143,0.2670057416,0.2965668142,-0.5154975653,0.2804078162,0.1454417259,0.4508609474,-0.0935282409,-0.1810210198,0.0642997622,-0.0575755686,0.1221427992,0.2484459728,-0.0715344921,-0.2342915833,0.1333340406,0.3007956445,-0.257455945,-0.064661555,-0.2737528682,0.0847610086,-0.0766698644,-0.2693378031,-0.0532864109,-0.4953400791,0.1539394706,0.011122671,0.1049450785,0.3338848054,-0.0001155144,-0.0851756856,0.0369373672,0.0938122272,-0.2382418513,-0.157761991,-0.0969283953,0.0886226818,-0.0698391646,-0.0270568896,-0.0834104121,0.1937476546,0.0529345647,-0.0981618613,0.5147064328,0.2915839553,0.1400677115,-0.0102435909,-0.252120316,-0.0659528822,0.3198694587,0.0313076191,0.0850752518,-0.1780210286,0.2137241364,0.0314790495,-0.0516679287,0.6180685163,-0.2813329399,0.3980180621,-0.2359145284,0.1242306605,-0.3711774647,-0.5308332443,0.266836673,0.2161651254,0.6465566754,-0.2296243757,-0.2044666409,-0.2378615439,0.3199083507,-0.1560648084,-0.0421698391,-0.0161785055,-0.2728021145,0.2350625694,-0.0094950655,0.047647696,-0.2204018831,0.1288805455,0.055811543,-0.1074196771,-0.0731363669,0.1385569125,-0.177100122,0.0994098634,-0.085863933,0.2703583241,-0.1442609429,0.0341251902,-0.149762243,-0.0674901083,0.1925161034,-0.3049418628,0.2917183042,-0.0602199472,0.3855580986,0.0374377258,-0.1856715977,-0.0881002173,0.3239292204,0.019808244,-0.0233677495,0.1496766508,0.2298973352,-0.2340689301,-0.2531856596,0.0737158284,-0.0144180572,0.2162737995,0.331879437,-0.265029937,0.2829249203,-0.367644608,-0.5658828616,0.0642497912,-0.4400376976,-0.1030385345,-0.2528013289,0.1816738844,0.1693309546,0.3009272516,0.1449253261,0.2226763219,0.0693393499,-0.5628325343,-0.1818621904,0.0267980024,0.2098604441,0.0308408234,0.211329326,-0.3237065375,0.3398516774,0.3417337835,0.195452705,-0.4315991104,-0.2874112725,-0.1301766485,0.0255610328,0.0091858534,-0.0264903046,0.3916681111,0.0738202855,-0.0182900447,0.0072356486,0.0610072762,-0.106556423,0.0068646008,0.2134016603,0.1364102364,-0.0805880576,-0.167114675,0.1232486665,0.3605877757,0.0779450312,-0.3008698225,0.0030109661,-0.1827422976,-0.3919641674,-0.0386591218,0.0773314983,0.3024609089,-0.3355484903,0.0954273939,0.178534463,0.4213981032,0.3944750428,0.2484540939,-0.0465098992,0.3734385073,-0.1168166548,0.3001444042,0.2041984349,-0.4621309042,-0.3920545578,0.2112352401,-0.049603641,0.2464142591,0.1548911184,-0.1627833694,0.3027338982,0.1427075267,0.3759355247,0.153478384,0.1457359195,0.0334793814,-0.2446887195,-0.057892438,0.0868306831,-0.0131406998,0.0021537091,0.0955836996,0.0370380655,0.232970655,0.5271517038,0.0745838359,0.0924516246,0.1029673368,0.1659648567,-0.1012314931,0.2239779383,-0.2631005943,0.2564861178,-0.0488976687,0.2068272382,0.2837985456,0.1540315896,0.0793671757,-0.3015450537,-0.2188883722,-0.2991884053,-0.3250726461,0.1145813912,0.3257373571,-0.0144317094,0.0420466512,-0.1898579001,0.3409897685,-0.2203264982,-0.0741857439,0.0233059432,0.0991652608,0.0638567135,-0.2447074801,0.5248855948,0.2528992593,0.18514961,0.2016844004,-0.1669406295,0.22355178,-0.0147676235,0.0219675135,-0.2637168467,-0.3288924992,0.1099477187,-0.5017575622,-0.1113948897,0.2344709635,0.3181717396,-0.0044999369,-0.3036757708,-0.1834709942,0.1390077174,0.2569387853,-0.0554442666,-0.0651190057,0.3011586368,0.0662527978,-0.3537996113,-0.3246975541,0.2526897192,-0.234980613,0.2662216723,0.0202848148,-0.3417770565,0.1585986316,0.3928949833,0.2245113999,0.085208714,0.2943886817,-0.2468326539,-0.027017273,-0.1227289811,0.0949118286,0.2907651365,0.2881050408,0.0543022901,-0.3764547706,-0.1287488788,-0.1132321432,0.1393357515,-0.1307561845,-0.3556297719,0.4351170063,0.3585450351,0.0863876268,-0.4493839145,-0.1341284662,-0.1424040794,0.4703677297,-0.3349600732,-0.2316320539,-0.4356057346,-0.1205524206,-0.109005712,0.0662912875,0.0797044486,0.225550577,0.2190566808,-0.0127121434,0.0054567931,0.1367991865,0.1719908118,0.2340619862,0.0712051392,0.1117983311,0.086305052,0.3191404045,-0.306956917,-0.0242702272,0.0259727612,0.271612525,0.1074295118,-0.1422471404,-0.1639021039,-0.4067965448,0.0590180866,0.2675643563,-0.3301686049,0.5113489628,0.1133633777,0.1602485478,-0.2251659632,-0.3130813539,0.0026724054,-0.0268870983,-0.1686861217,0.3555683196,0.0829433203,-0.0056142341,-0.0482558794,-0.698974669,0.099392809,-0.2500089407,0.2359433323,-0.0328674652,0.2103785425,0.3641358316,0.0606071204,0.0812700838,-0.0587117635,0.2053699046,-0.410209775,-0.2545935214,0.3561107218,0.2606044412,-0.3348694146,-0.3405695558,0.0509579033,0.2652952075,-0.0322290994,-0.4001087248,0.1003541723,-0.2674658895,0.1453480273,-0.2972603142,0.2826198339,0.3854675591,0.0129939336,-0.1495047212,-0.1692215204,-0.1078334898,0.2658356428,0.3553734422,0.0058843573,0.1011472642,0.0699137822,-0.0593266487,0.3506978154,0.6459594965,0.2025162727,0.2600066662,-0.1765539348,0.0942354053,-0.0324934721,-0.3995798528,0.0270797946,-0.1173247173,-0.062567167,0.4981202781,-0.0899957195,0.0536311306,-0.1695882976,-0.0572774038,-0.3290627599,-0.1686403602,0.0674927011,-0.2359398454,0.1839738041,0.1747753173,0.0528291985,0.1619518846,0.0347351544,-0.2861667871,-0.2988380492,0.2851738334,-0.1499728858,-0.0820301399,-0.2836113274,-0.377288729,-0.1241017282,0.1385834366,0.5095545053,-0.2117005289,-0.037520647,0.3600184023,0.291478008,0.3810289502,-0.3069941998,0.0908769667,0.0103903497,0.2287908047,0.0055077169,0.0054613417,-0.5205008984,-0.1967586279,-0.0475005209,-0.043032378,-0.1668928564,-0.0080181081,0.108959198,0.3927445412,-0.0097480956,-0.2408112288,-0.2557696104,-0.0757930055,-0.3301421404,0.1615659893,0.3422310054,0.0443936288,-0.1651335955,-0.1203253195,0.1603984088,-0.0627515167,0.5332366228,0.0798594728,0.375176698,0.0739652514,0.3011747897,0.2748407722,-0.3577797413,-0.2591435611,0.3608151674,0.100969635,-0.1420252174,-0.00367438,-0.3468615413,0.3289579153,0.4931568503,-0.223191008,0.2419292778,-0.2970614135,0.2568941116,-0.1529492587,0.0957464501,0.3195284307,-0.0749837682,-0.0751821995,-0.410286963,0.3822805583,-0.0364946276,-0.1979487538,0.3445512056,0.0909191668,-0.3103821576,0.4712514579,-0.2578842938,0.8118340373,0.2264072746,0.3666303754,0.3155016303,0.1381517202,0.3261640966,-0.1960099638,0.0902330428,-0.0220177118,-0.0018170826,-0.0215603653,0.0583430231,-0.0778571218,-0.0417594425,-0.5769283772,0.1775290817,0.0647163838,0.0288711693,0.1868025362,-0.0118521098,-0.1294962317,-0.1064071804,0.1252996773,0.0325710885,-0.3839387298,0.3074729145,0.3099324405,-0.0242960695,0.0314041749,-0.5837308764,-0.8224813938,-0.0301681384,-0.0825177506,-0.2259411067,-0.1996800154,-0.3336216509,0.3681984842,0.2585997581,0.1915538907,0.3256590664,-0.1379685849,-0.0852642953,0.3682431877,0.2776513696,-0.2679565549,-0.0255918875,0.1445183456,-0.0860315636,-0.3883327842,0.1766039878,0.0551586859,0.0989629775,-0.3880658448,0.2538315058,-0.1146805435,-0.2331590801,0.0857373253,0.4289047122,0.1415170133,-0.5119704008,0.10320732,0.1241723523,-0.4588638544,0.3286851645,0.1725837886,-0.4755195975,-0.0637625083,0.259340167,0.1283604354,0.0002506027,0.6646648645,-0.0414828882,0.0483881421,-0.37047562,0.0406970792,-0.0209288169,-0.2641952336,0.1791121364,0.0586630926,0.1739625186,-0.2265277505,0.2535494268,0.0265137423,0.204802826,-0.2814756632,-0.3617306948,-0.0619779229,-0.0314538814,0.1234360188,0.0552398637,-0.2911967039,0.1181857958,0.0688160807,0.3186088502,-0.2981917858,0.0059745917,-0.0517939329,0.2173203677,-0.0840429664,0.0813204795,-0.0378096215,0.0851734653,0.2000400573,0.3185155094,0.0672922209,-0.1598699093,-0.3062146604,0.1838445365,0.0417443737,0.1562646031,-0.040768452,0.100105688,0.1492390484,-0.1337077022,-0.1010111123,0.334582895,0.1253624856,0.2183783352,-0.0296323262,0.0370717831,0.1851567924,-0.1598648131,-0.1839772761,-0.172459051,0.3399853706,0.1806136072,0.0487441346,-0.0481960177,0.1186248809,0.0399239436,-0.1340517849,0.1186952814,-0.0778344646,-0.2655762732,0.1333703995,0.2876859903,-0.1194795594,0.3786883652,-0.3459939361,-0.1118993759,0.091195181,0.235143885,-0.2654986084,-0.1237113848,-0.3232501447,-0.1065506563,-0.0583089553,0.2784046829,-0.0598313436,-0.4723803997,0.121467337,-0.0649383292,0.336020112,-0.5360029936,-0.0734213144,0.3985144794,0.0149740381,-0.3611198664,-0.1933412552,-0.0151967062,0.0208013449,0.5450986028,-0.3675437868,0.1373579353,-0.3386060596,0.0263877269,0.191548273,0.0256050285,-0.0382200703,0.4056419134,-0.3271881938,-0.0489630923,-0.0716247186,0.3050477505,-0.1176727563,0.0334460959,-0.1280483007,-0.0514146574,-0.1239193901,-0.1248619258,-0.3380397856,-0.2734287083,-0.2303761989,0.0061822413,0.0975418091,-0.0125555322,0.1405136287,0.1916218102,-0.2668848932,-0.5332419276,-0.1980937272,-0.0377839021,0.1720763147,-0.3900322318,0.3403875828,0.3084276021,-0.221769765,0.2584382296,0.2433644086,0.519982934,0.0715812668,0.0334475301,0.1065192446,-0.5706885457,-0.0919026062,-0.0517226532,0.3297682703,-0.1242420226,0.0587560534,0.0218634959,0.0158726722,-0.095639661,0.0293477289,0.2285464853,-0.0756259263,-0.563256681,0.2190777361,-0.1411624253,0.1117327288,0.0406589434,0.0749845207,-0.4736695588,0.4417389929,0.3421675265,0.0575963221,-0.0106266811,0.0021905343,0.0813886747,-0.1263748854,-0.1126130521,0.2673044801,0.1579869688,-0.2770343423,-0.1591356099,-0.6735790372,-0.071482718,-0.0488048792,-0.0049527902,0.0006292545,0.1189806312,0.1586837918,0.1362603009,0.0692648441,-0.0260408036,-0.5322327018,0.3291485906,-0.215462625,-0.0053817439,-0.0877875164,0.4366607666,0.0414892845,-0.1064981595,0.0108353356,0.094691515,0.0340188518,0.0655476525,0.0059462963,0.017562082,-0.1206619516,0.1582386047,0.1423452646,0.1708153635,-0.234853968,0.1895660013,-0.2832585871,0.0292929802,-0.1982437521,0.1561561525,0.2698151469,0.5158427358,-0.183451429,-0.0686847195,-0.0636009946,0.165361613,0.1112443656,-0.0649597496,-0.1914123297,0.089769423,-0.0087000672,0.2714968026,-0.2622262836,0.1059425324,0.0038290906,0.0026387305,-0.4920392334,-0.7109149694,0.4522462487,-0.4033147991,-0.5287759304,-0.0796565562,0.590672791,0.2403544337,-0.032153476,-0.7086960673,-0.0359898657,0.3959985971,0.2443489283,-0.3021709323,0.2129863203,0.1219399646,0.0085307239,0.0144249219,0.6226139069,0.3105249107,-0.2327545434,0.2679713964,-0.4604581892]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2913","title":"timit_asr dataset only includes one text phrase","comments":"Hi @margotwagner,\r\n\r\nYes, as @bhavitvyamalik has commented, this bug was fixed in `datasets` version 1.5.0. You need to update it, as your current version is 1.4.1:\r\n> Environment info\r\n> - `datasets` version: 1.4.1","body":"## Describe the bug\r\nThe dataset 'timit_asr' only includes one text phrase. It only includes the transcription \"Would such an act of refusal be useful?\" multiple times rather than different phrases.\r\n\r\n## Steps to reproduce the bug\r\nNote: I am following the tutorial https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\n1. Install the dataset and other packages\r\n```python\r\n!pip install datasets>=1.5.0\r\n!pip install transformers==4.4.0\r\n!pip install soundfile\r\n!pip install jiwer\r\n```\r\n2. Load the dataset\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n```\r\n3. Remove columns that we don't want\r\n```python\r\ntimit = timit.remove_columns([\"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \"sentence_type\", \"speaker_id\"])\r\n```\r\n4. Write a short function to display some random samples of the dataset.\r\n```python\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n \r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\nshow_random_elements(timit[\"train\"].remove_columns([\"file\"]))\r\n```\r\n\r\n## Expected results\r\n10 random different transcription phrases.\r\n\r\n## Actual results\r\n10 of the same transcription phrase \"Would such an act of refusal be useful?\"\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.4.1\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: not listed\r\n","comment_length":34,"text":"timit_asr dataset only includes one text phrase\n## Describe the bug\r\nThe dataset 'timit_asr' only includes one text phrase. It only includes the transcription \"Would such an act of refusal be useful?\" multiple times rather than different phrases.\r\n\r\n## Steps to reproduce the bug\r\nNote: I am following the tutorial https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\n1. Install the dataset and other packages\r\n```python\r\n!pip install datasets>=1.5.0\r\n!pip install transformers==4.4.0\r\n!pip install soundfile\r\n!pip install jiwer\r\n```\r\n2. Load the dataset\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n```\r\n3. Remove columns that we don't want\r\n```python\r\ntimit = timit.remove_columns([\"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \"sentence_type\", \"speaker_id\"])\r\n```\r\n4. Write a short function to display some random samples of the dataset.\r\n```python\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n \r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\nshow_random_elements(timit[\"train\"].remove_columns([\"file\"]))\r\n```\r\n\r\n## Expected results\r\n10 random different transcription phrases.\r\n\r\n## Actual results\r\n10 of the same transcription phrase \"Would such an act of refusal be useful?\"\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.4.1\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: not listed\r\n\nHi @margotwagner,\r\n\r\nYes, as @bhavitvyamalik has commented, this bug was fixed in `datasets` version 1.5.0. You need to update it, as your current version is 1.4.1:\r\n> Environment info\r\n> - `datasets` version: 1.4.1","embeddings":[0.161290288,-0.138053894,-0.0205535926,0.2273859978,0.1176241562,-0.1575979143,0.2670057416,0.2965668142,-0.5154975653,0.2804078162,0.1454417259,0.4508609474,-0.0935282409,-0.1810210198,0.0642997622,-0.0575755686,0.1221427992,0.2484459728,-0.0715344921,-0.2342915833,0.1333340406,0.3007956445,-0.257455945,-0.064661555,-0.2737528682,0.0847610086,-0.0766698644,-0.2693378031,-0.0532864109,-0.4953400791,0.1539394706,0.011122671,0.1049450785,0.3338848054,-0.0001155144,-0.0851756856,0.0369373672,0.0938122272,-0.2382418513,-0.157761991,-0.0969283953,0.0886226818,-0.0698391646,-0.0270568896,-0.0834104121,0.1937476546,0.0529345647,-0.0981618613,0.5147064328,0.2915839553,0.1400677115,-0.0102435909,-0.252120316,-0.0659528822,0.3198694587,0.0313076191,0.0850752518,-0.1780210286,0.2137241364,0.0314790495,-0.0516679287,0.6180685163,-0.2813329399,0.3980180621,-0.2359145284,0.1242306605,-0.3711774647,-0.5308332443,0.266836673,0.2161651254,0.6465566754,-0.2296243757,-0.2044666409,-0.2378615439,0.3199083507,-0.1560648084,-0.0421698391,-0.0161785055,-0.2728021145,0.2350625694,-0.0094950655,0.047647696,-0.2204018831,0.1288805455,0.055811543,-0.1074196771,-0.0731363669,0.1385569125,-0.177100122,0.0994098634,-0.085863933,0.2703583241,-0.1442609429,0.0341251902,-0.149762243,-0.0674901083,0.1925161034,-0.3049418628,0.2917183042,-0.0602199472,0.3855580986,0.0374377258,-0.1856715977,-0.0881002173,0.3239292204,0.019808244,-0.0233677495,0.1496766508,0.2298973352,-0.2340689301,-0.2531856596,0.0737158284,-0.0144180572,0.2162737995,0.331879437,-0.265029937,0.2829249203,-0.367644608,-0.5658828616,0.0642497912,-0.4400376976,-0.1030385345,-0.2528013289,0.1816738844,0.1693309546,0.3009272516,0.1449253261,0.2226763219,0.0693393499,-0.5628325343,-0.1818621904,0.0267980024,0.2098604441,0.0308408234,0.211329326,-0.3237065375,0.3398516774,0.3417337835,0.195452705,-0.4315991104,-0.2874112725,-0.1301766485,0.0255610328,0.0091858534,-0.0264903046,0.3916681111,0.0738202855,-0.0182900447,0.0072356486,0.0610072762,-0.106556423,0.0068646008,0.2134016603,0.1364102364,-0.0805880576,-0.167114675,0.1232486665,0.3605877757,0.0779450312,-0.3008698225,0.0030109661,-0.1827422976,-0.3919641674,-0.0386591218,0.0773314983,0.3024609089,-0.3355484903,0.0954273939,0.178534463,0.4213981032,0.3944750428,0.2484540939,-0.0465098992,0.3734385073,-0.1168166548,0.3001444042,0.2041984349,-0.4621309042,-0.3920545578,0.2112352401,-0.049603641,0.2464142591,0.1548911184,-0.1627833694,0.3027338982,0.1427075267,0.3759355247,0.153478384,0.1457359195,0.0334793814,-0.2446887195,-0.057892438,0.0868306831,-0.0131406998,0.0021537091,0.0955836996,0.0370380655,0.232970655,0.5271517038,0.0745838359,0.0924516246,0.1029673368,0.1659648567,-0.1012314931,0.2239779383,-0.2631005943,0.2564861178,-0.0488976687,0.2068272382,0.2837985456,0.1540315896,0.0793671757,-0.3015450537,-0.2188883722,-0.2991884053,-0.3250726461,0.1145813912,0.3257373571,-0.0144317094,0.0420466512,-0.1898579001,0.3409897685,-0.2203264982,-0.0741857439,0.0233059432,0.0991652608,0.0638567135,-0.2447074801,0.5248855948,0.2528992593,0.18514961,0.2016844004,-0.1669406295,0.22355178,-0.0147676235,0.0219675135,-0.2637168467,-0.3288924992,0.1099477187,-0.5017575622,-0.1113948897,0.2344709635,0.3181717396,-0.0044999369,-0.3036757708,-0.1834709942,0.1390077174,0.2569387853,-0.0554442666,-0.0651190057,0.3011586368,0.0662527978,-0.3537996113,-0.3246975541,0.2526897192,-0.234980613,0.2662216723,0.0202848148,-0.3417770565,0.1585986316,0.3928949833,0.2245113999,0.085208714,0.2943886817,-0.2468326539,-0.027017273,-0.1227289811,0.0949118286,0.2907651365,0.2881050408,0.0543022901,-0.3764547706,-0.1287488788,-0.1132321432,0.1393357515,-0.1307561845,-0.3556297719,0.4351170063,0.3585450351,0.0863876268,-0.4493839145,-0.1341284662,-0.1424040794,0.4703677297,-0.3349600732,-0.2316320539,-0.4356057346,-0.1205524206,-0.109005712,0.0662912875,0.0797044486,0.225550577,0.2190566808,-0.0127121434,0.0054567931,0.1367991865,0.1719908118,0.2340619862,0.0712051392,0.1117983311,0.086305052,0.3191404045,-0.306956917,-0.0242702272,0.0259727612,0.271612525,0.1074295118,-0.1422471404,-0.1639021039,-0.4067965448,0.0590180866,0.2675643563,-0.3301686049,0.5113489628,0.1133633777,0.1602485478,-0.2251659632,-0.3130813539,0.0026724054,-0.0268870983,-0.1686861217,0.3555683196,0.0829433203,-0.0056142341,-0.0482558794,-0.698974669,0.099392809,-0.2500089407,0.2359433323,-0.0328674652,0.2103785425,0.3641358316,0.0606071204,0.0812700838,-0.0587117635,0.2053699046,-0.410209775,-0.2545935214,0.3561107218,0.2606044412,-0.3348694146,-0.3405695558,0.0509579033,0.2652952075,-0.0322290994,-0.4001087248,0.1003541723,-0.2674658895,0.1453480273,-0.2972603142,0.2826198339,0.3854675591,0.0129939336,-0.1495047212,-0.1692215204,-0.1078334898,0.2658356428,0.3553734422,0.0058843573,0.1011472642,0.0699137822,-0.0593266487,0.3506978154,0.6459594965,0.2025162727,0.2600066662,-0.1765539348,0.0942354053,-0.0324934721,-0.3995798528,0.0270797946,-0.1173247173,-0.062567167,0.4981202781,-0.0899957195,0.0536311306,-0.1695882976,-0.0572774038,-0.3290627599,-0.1686403602,0.0674927011,-0.2359398454,0.1839738041,0.1747753173,0.0528291985,0.1619518846,0.0347351544,-0.2861667871,-0.2988380492,0.2851738334,-0.1499728858,-0.0820301399,-0.2836113274,-0.377288729,-0.1241017282,0.1385834366,0.5095545053,-0.2117005289,-0.037520647,0.3600184023,0.291478008,0.3810289502,-0.3069941998,0.0908769667,0.0103903497,0.2287908047,0.0055077169,0.0054613417,-0.5205008984,-0.1967586279,-0.0475005209,-0.043032378,-0.1668928564,-0.0080181081,0.108959198,0.3927445412,-0.0097480956,-0.2408112288,-0.2557696104,-0.0757930055,-0.3301421404,0.1615659893,0.3422310054,0.0443936288,-0.1651335955,-0.1203253195,0.1603984088,-0.0627515167,0.5332366228,0.0798594728,0.375176698,0.0739652514,0.3011747897,0.2748407722,-0.3577797413,-0.2591435611,0.3608151674,0.100969635,-0.1420252174,-0.00367438,-0.3468615413,0.3289579153,0.4931568503,-0.223191008,0.2419292778,-0.2970614135,0.2568941116,-0.1529492587,0.0957464501,0.3195284307,-0.0749837682,-0.0751821995,-0.410286963,0.3822805583,-0.0364946276,-0.1979487538,0.3445512056,0.0909191668,-0.3103821576,0.4712514579,-0.2578842938,0.8118340373,0.2264072746,0.3666303754,0.3155016303,0.1381517202,0.3261640966,-0.1960099638,0.0902330428,-0.0220177118,-0.0018170826,-0.0215603653,0.0583430231,-0.0778571218,-0.0417594425,-0.5769283772,0.1775290817,0.0647163838,0.0288711693,0.1868025362,-0.0118521098,-0.1294962317,-0.1064071804,0.1252996773,0.0325710885,-0.3839387298,0.3074729145,0.3099324405,-0.0242960695,0.0314041749,-0.5837308764,-0.8224813938,-0.0301681384,-0.0825177506,-0.2259411067,-0.1996800154,-0.3336216509,0.3681984842,0.2585997581,0.1915538907,0.3256590664,-0.1379685849,-0.0852642953,0.3682431877,0.2776513696,-0.2679565549,-0.0255918875,0.1445183456,-0.0860315636,-0.3883327842,0.1766039878,0.0551586859,0.0989629775,-0.3880658448,0.2538315058,-0.1146805435,-0.2331590801,0.0857373253,0.4289047122,0.1415170133,-0.5119704008,0.10320732,0.1241723523,-0.4588638544,0.3286851645,0.1725837886,-0.4755195975,-0.0637625083,0.259340167,0.1283604354,0.0002506027,0.6646648645,-0.0414828882,0.0483881421,-0.37047562,0.0406970792,-0.0209288169,-0.2641952336,0.1791121364,0.0586630926,0.1739625186,-0.2265277505,0.2535494268,0.0265137423,0.204802826,-0.2814756632,-0.3617306948,-0.0619779229,-0.0314538814,0.1234360188,0.0552398637,-0.2911967039,0.1181857958,0.0688160807,0.3186088502,-0.2981917858,0.0059745917,-0.0517939329,0.2173203677,-0.0840429664,0.0813204795,-0.0378096215,0.0851734653,0.2000400573,0.3185155094,0.0672922209,-0.1598699093,-0.3062146604,0.1838445365,0.0417443737,0.1562646031,-0.040768452,0.100105688,0.1492390484,-0.1337077022,-0.1010111123,0.334582895,0.1253624856,0.2183783352,-0.0296323262,0.0370717831,0.1851567924,-0.1598648131,-0.1839772761,-0.172459051,0.3399853706,0.1806136072,0.0487441346,-0.0481960177,0.1186248809,0.0399239436,-0.1340517849,0.1186952814,-0.0778344646,-0.2655762732,0.1333703995,0.2876859903,-0.1194795594,0.3786883652,-0.3459939361,-0.1118993759,0.091195181,0.235143885,-0.2654986084,-0.1237113848,-0.3232501447,-0.1065506563,-0.0583089553,0.2784046829,-0.0598313436,-0.4723803997,0.121467337,-0.0649383292,0.336020112,-0.5360029936,-0.0734213144,0.3985144794,0.0149740381,-0.3611198664,-0.1933412552,-0.0151967062,0.0208013449,0.5450986028,-0.3675437868,0.1373579353,-0.3386060596,0.0263877269,0.191548273,0.0256050285,-0.0382200703,0.4056419134,-0.3271881938,-0.0489630923,-0.0716247186,0.3050477505,-0.1176727563,0.0334460959,-0.1280483007,-0.0514146574,-0.1239193901,-0.1248619258,-0.3380397856,-0.2734287083,-0.2303761989,0.0061822413,0.0975418091,-0.0125555322,0.1405136287,0.1916218102,-0.2668848932,-0.5332419276,-0.1980937272,-0.0377839021,0.1720763147,-0.3900322318,0.3403875828,0.3084276021,-0.221769765,0.2584382296,0.2433644086,0.519982934,0.0715812668,0.0334475301,0.1065192446,-0.5706885457,-0.0919026062,-0.0517226532,0.3297682703,-0.1242420226,0.0587560534,0.0218634959,0.0158726722,-0.095639661,0.0293477289,0.2285464853,-0.0756259263,-0.563256681,0.2190777361,-0.1411624253,0.1117327288,0.0406589434,0.0749845207,-0.4736695588,0.4417389929,0.3421675265,0.0575963221,-0.0106266811,0.0021905343,0.0813886747,-0.1263748854,-0.1126130521,0.2673044801,0.1579869688,-0.2770343423,-0.1591356099,-0.6735790372,-0.071482718,-0.0488048792,-0.0049527902,0.0006292545,0.1189806312,0.1586837918,0.1362603009,0.0692648441,-0.0260408036,-0.5322327018,0.3291485906,-0.215462625,-0.0053817439,-0.0877875164,0.4366607666,0.0414892845,-0.1064981595,0.0108353356,0.094691515,0.0340188518,0.0655476525,0.0059462963,0.017562082,-0.1206619516,0.1582386047,0.1423452646,0.1708153635,-0.234853968,0.1895660013,-0.2832585871,0.0292929802,-0.1982437521,0.1561561525,0.2698151469,0.5158427358,-0.183451429,-0.0686847195,-0.0636009946,0.165361613,0.1112443656,-0.0649597496,-0.1914123297,0.089769423,-0.0087000672,0.2714968026,-0.2622262836,0.1059425324,0.0038290906,0.0026387305,-0.4920392334,-0.7109149694,0.4522462487,-0.4033147991,-0.5287759304,-0.0796565562,0.590672791,0.2403544337,-0.032153476,-0.7086960673,-0.0359898657,0.3959985971,0.2443489283,-0.3021709323,0.2129863203,0.1219399646,0.0085307239,0.0144249219,0.6226139069,0.3105249107,-0.2327545434,0.2679713964,-0.4604581892]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2904","title":"FORCE_REDOWNLOAD does not work","comments":"Hi ! Thanks for reporting. The error seems to happen only if you use compressed files.\r\n\r\nThe second dataset is prepared in another dataset cache directory than the first - which is normal, since the source file is different. However, it doesn't uncompress the new data file because it finds the old uncompressed data in the extraction cache directory.\r\n\r\nIf we fix the extraction cache mechanism to uncompress a local file if it changed then it should fix the issue.\r\nCurrently the extraction cache mechanism only takes into account the path of the compressed file, which is an issue.","body":"## Describe the bug\r\nWith GenerateMode.FORCE_REDOWNLOAD, the documentation says \r\n +------------------------------------+-----------+---------+\r\n | | Downloads | Dataset |\r\n +====================================+===========+=========+\r\n | `REUSE_DATASET_IF_EXISTS` (default)| Reuse | Reuse |\r\n +------------------------------------+-----------+---------+\r\n | `REUSE_CACHE_IF_EXISTS` | Reuse | Fresh |\r\n +------------------------------------+-----------+---------+\r\n | `FORCE_REDOWNLOAD` | Fresh | Fresh |\r\n +------------------------------------+-----------+---------+\r\n\r\nHowever, the old dataset is loaded even when FORCE_REDOWNLOAD is chosen.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\nimport pandas as pd\r\nfrom datasets import load_dataset, GenerateMode\r\npd.DataFrame(range(5), columns=['numbers']).to_csv('\/tmp\/test.tsv.gz', index=False)\r\nee = load_dataset('csv', data_files=['\/tmp\/test.tsv.gz'], delimiter='\\t', split='train', download_mode=GenerateMode.FORCE_REDOWNLOAD)\r\nprint(ee)\r\npd.DataFrame(range(10), columns=['numerals']).to_csv('\/tmp\/test.tsv.gz', index=False)\r\nee = load_dataset('csv', data_files=['\/tmp\/test.tsv.gz'], delimiter='\\t', split='train', download_mode=GenerateMode.FORCE_REDOWNLOAD)\r\nprint(ee)\r\n\r\n```\r\n\r\n## Expected results\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\nDataset({\r\n features: ['numerals'],\r\n num_rows: 10\r\n})\r\n\r\n## Actual results\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-4.14.181-108.257.amzn1.x86_64-x86_64-with-glibc2.10\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":99,"text":"FORCE_REDOWNLOAD does not work\n## Describe the bug\r\nWith GenerateMode.FORCE_REDOWNLOAD, the documentation says \r\n +------------------------------------+-----------+---------+\r\n | | Downloads | Dataset |\r\n +====================================+===========+=========+\r\n | `REUSE_DATASET_IF_EXISTS` (default)| Reuse | Reuse |\r\n +------------------------------------+-----------+---------+\r\n | `REUSE_CACHE_IF_EXISTS` | Reuse | Fresh |\r\n +------------------------------------+-----------+---------+\r\n | `FORCE_REDOWNLOAD` | Fresh | Fresh |\r\n +------------------------------------+-----------+---------+\r\n\r\nHowever, the old dataset is loaded even when FORCE_REDOWNLOAD is chosen.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\nimport pandas as pd\r\nfrom datasets import load_dataset, GenerateMode\r\npd.DataFrame(range(5), columns=['numbers']).to_csv('\/tmp\/test.tsv.gz', index=False)\r\nee = load_dataset('csv', data_files=['\/tmp\/test.tsv.gz'], delimiter='\\t', split='train', download_mode=GenerateMode.FORCE_REDOWNLOAD)\r\nprint(ee)\r\npd.DataFrame(range(10), columns=['numerals']).to_csv('\/tmp\/test.tsv.gz', index=False)\r\nee = load_dataset('csv', data_files=['\/tmp\/test.tsv.gz'], delimiter='\\t', split='train', download_mode=GenerateMode.FORCE_REDOWNLOAD)\r\nprint(ee)\r\n\r\n```\r\n\r\n## Expected results\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\nDataset({\r\n features: ['numerals'],\r\n num_rows: 10\r\n})\r\n\r\n## Actual results\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-4.14.181-108.257.amzn1.x86_64-x86_64-with-glibc2.10\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\nHi ! Thanks for reporting. The error seems to happen only if you use compressed files.\r\n\r\nThe second dataset is prepared in another dataset cache directory than the first - which is normal, since the source file is different. However, it doesn't uncompress the new data file because it finds the old uncompressed data in the extraction cache directory.\r\n\r\nIf we fix the extraction cache mechanism to uncompress a local file if it changed then it should fix the issue.\r\nCurrently the extraction cache mechanism only takes into account the path of the compressed file, which is an issue.","embeddings":[-0.1253399998,0.0207806267,0.0185128842,0.0257142521,0.1821040809,0.0251934566,0.4788660109,0.2438331693,0.145796001,-0.2338699996,-0.1162827983,0.2687395215,0.0850747079,0.1121127084,-0.0132031068,0.3137487769,0.1101020277,0.2144826502,-0.0722493976,-0.0051923986,-0.2910993099,0.0634416714,-0.1833185554,-0.2195639163,-0.2065767497,0.2937990129,-0.0426936708,0.3026081324,0.0084062209,-0.3731202781,0.1836812794,0.2550069988,0.1820975393,0.6611076593,-0.000109716,0.0617883578,0.1586239785,-0.1793150753,-0.1704429984,-0.1858083606,0.0419964381,-0.0455286205,-0.3461463749,-0.0842089579,-0.0005326021,-0.2207529545,-0.1748757958,-0.346149832,0.5304936767,0.4028010368,0.2320071906,-0.1104161516,0.1202409416,-0.0721178129,0.1787120998,0.1414143443,-0.0699524358,0.3557813466,0.1045252755,-0.2072234303,-0.090631254,0.071447365,-0.1301329136,0.0515931137,-0.100776881,-0.0266386643,0.3280929923,-0.1360899806,0.0856568739,0.1706219614,0.60730654,-0.3377125561,-0.4172843397,0.0072846375,0.0183145087,-0.2442766875,0.1650239527,0.1792488247,-0.0074186656,0.368666321,-0.0126197841,-0.0254521314,0.0721380934,-0.1948064566,-0.0308580399,-0.117819868,-0.1004166082,-0.0362135768,0.0516206622,0.0602480508,0.4213585556,-0.021616118,-0.0566806123,0.0342854224,-0.029152602,-0.1342204958,0.2153061479,-0.3018356264,0.0747370049,-0.0281923842,-0.0764476284,-0.1077359319,-0.0592024475,0.04897983,0.0738615543,0.2261448056,-0.1387252212,-0.090120174,0.3910107613,0.1767412573,-0.251378119,-0.0726185068,0.0294129029,-0.2027270049,0.5261009336,-0.0647600591,0.2614033818,-0.1472218335,-0.4044618905,-0.0181814078,-0.1070784777,0.0032655229,-0.2742642164,0.1767624319,0.0207748469,0.3530659974,0.1341454834,0.1856252104,0.0581398085,-0.206356883,-0.2226337641,-0.1226871684,-0.3091842234,0.0251519568,0.2947835922,-0.1708823442,0.3440055251,0.5470414758,-0.3297725916,-0.5027301908,-0.0621353462,-0.0314323269,0.145252198,0.2964736819,-0.0633033141,-0.0863981768,0.2666994929,-0.0901367441,-0.0168399569,0.3552974463,-0.0914471522,-0.3069287539,0.1436540186,0.1975634098,0.0511063114,0.1958488077,-0.127950877,0.0155278603,0.3042720556,-0.0705190748,-0.0457962081,-0.0500780307,-0.0136534544,-0.2484514862,0.0476663969,0.7337824702,-0.4681162834,0.0081418594,-0.2922013104,-0.0890840143,0.2404547036,0.0785243809,0.0047035967,0.0423925631,-0.4686902165,-0.3603310585,0.1667583287,-0.1669529229,-0.4369600415,0.0058205375,-0.2142421305,0.2575328946,0.1665146351,-0.0449789762,-0.1501356214,-0.0172337554,0.0970478505,0.2490073442,0.0923939124,0.044823505,-0.3246170282,-0.288821131,0.0455681123,-0.1329662651,0.2855156362,0.405680865,0.3521436155,0.166620627,0.3699389398,0.0625928491,0.0155102359,0.0143866315,0.3367364407,-0.0180097409,0.0330760218,0.1765212268,-0.5221686959,0.3075172901,-0.0336038433,-0.2515624166,0.0728959292,-0.0701956674,-0.5719491243,-0.228338778,-0.0720977858,-0.3877546787,0.1725475937,0.4713421166,-0.033435449,0.0184990596,-0.1434893459,0.4725981653,-0.3845011294,-0.0480710976,-0.0413006209,0.2855195999,-0.1932275444,-0.1814612597,-0.2195678353,-0.1187965199,0.2447945774,-0.1625743508,-0.1215589717,0.4299698472,0.186048165,0.2720868587,-0.3619140387,-0.1950368881,0.1652197391,-0.016063001,-0.1465488821,0.1185459942,0.2459682077,-0.0776265264,-0.0173951481,0.2789047956,-0.4031001925,-0.0885871574,0.071374476,-0.0462191142,0.2666130662,-0.2367841005,0.2828178704,-0.340397954,-0.3403313756,0.2947020233,0.0018241567,0.1279272139,0.0989251807,0.1121156812,0.2867794335,0.0066712452,0.047518149,-0.0890123621,-0.1765656024,-0.2945634723,0.0131209595,0.3500014544,0.5929611921,0.1657259762,-0.0166612044,0.1451728493,-0.0583074652,-0.253809005,0.2043861002,-0.2613338828,-0.0064706737,0.3311824501,0.1878525168,0.0763362646,-0.6370830536,0.186015591,0.5140926838,0.2118025422,-0.2195149958,-0.1909853667,-0.1673647612,0.4125154912,0.0630170777,0.1435509324,-0.0300561506,-0.1065669209,-0.138272956,0.5955287814,-0.0589990281,0.073292084,-0.1850811243,-0.0749260634,-0.0461661331,-0.1426966637,-0.1472802907,-0.1764150858,-0.1414824873,0.0910957381,0.2313582748,-0.3419226408,0.1477311403,-0.1828764379,-0.1267566234,-0.3267619908,-0.0833430737,-0.1293941885,0.0340040401,0.2224018276,-0.2250489891,0.1471374333,0.169757694,0.0325683653,-0.004715248,-0.3554546833,-0.3341683745,-0.1029282659,0.1511999369,0.2633759975,-0.2600107789,-0.4554549158,0.1090232432,-0.3711392879,-0.0406573713,0.1015611589,-0.0919760466,0.3833361566,0.2238112539,-0.0310825258,0.4020083547,0.2698833942,-0.4376188517,-0.2790603936,0.0805035159,-0.1014774665,-0.2633699775,0.0147913452,0.0609156676,-0.0121200867,0.2737061381,-0.4307680726,-0.0845259503,-0.2824058533,0.0837785453,-0.0585749894,0.2336959094,0.2081224918,0.1207609028,-0.2136411071,-0.1040799022,-0.0956789032,-0.041495204,0.233491689,0.126982823,0.0212046467,0.2292778045,0.0935394913,0.6404239535,0.0503102541,-0.1254811585,0.5086991787,0.125423491,0.7294853926,-0.074036397,-0.339060694,-0.0056349542,-0.1845843047,0.319257617,0.0420288928,-0.1061969697,0.1309799999,-0.1564230025,0.0242274236,-0.1678121686,-0.2236509025,-0.1179092973,-0.2493901998,-0.0220435485,0.314229995,-0.1047639698,-0.1341655105,0.0813808367,-0.0185252242,0.0876276046,0.3220412135,0.0884925127,0.0994393826,0.5284665227,-0.1080815345,-0.0133388601,-0.0877977833,0.3645614684,0.2040727586,-0.202382952,0.0585922338,-0.023118237,0.7181988358,-0.2194465101,0.313290745,0.2286951095,0.0114228521,-0.0870219991,-0.196765542,0.1032617912,-0.1487740576,0.1887097061,0.3649292588,0.0375474505,0.0390015617,-0.1003576666,0.2032171041,0.0621288046,-0.1629727781,-0.2788121998,-0.1599720269,-0.4588589966,0.3117338121,0.1915880591,0.0890043527,-0.320887208,0.0846341178,-0.2478237301,-0.4116822779,-0.0395846739,0.1951411813,0.2008189261,0.1139606908,0.0489872061,-0.0161832664,-0.1171152368,0.2674762607,0.4194466472,-0.1649921089,0.0667295754,-0.1280568391,0.1093956754,-0.067581296,0.319385618,-0.0657960176,-0.1120607778,-0.3978951871,-0.1618822664,0.0620966889,0.0673459694,0.002971882,-0.3354582787,-0.0954507068,-0.4370097518,0.2871852815,-0.0682167336,-0.0515976958,0.4572842717,0.2364326417,-0.2039115429,0.1625162512,0.0528943725,0.8192458153,-0.1221480891,0.165192157,0.0327902324,-0.0593867525,0.4034763277,-0.2886822224,0.194058761,-0.1767736971,-0.3061233759,0.1044934094,-0.0408082418,0.2328797132,0.2079223692,-0.1264044195,0.1385401487,-0.0984730124,0.4167669713,0.2764787674,-0.0030264198,-0.2224701941,-0.1413925886,-0.2291935533,0.1468190849,0.2811729312,0.1926800758,-0.1454819441,-0.0086173862,0.1220361069,-0.0229478516,-0.2685617507,0.2866361737,-0.1642533094,0.3783459961,-0.4507656693,-0.1098404899,0.0326456428,0.2600927949,0.2549611032,0.2578200102,0.0146770021,0.0830926001,0.4122848809,0.1323083341,-0.1551139206,0.0008572826,0.4158841372,0.1006288007,-0.3379714787,0.1988293082,-0.0907847136,-0.1476554275,0.0037588191,0.2748386562,-0.0433281027,-0.1125033721,-0.092763938,0.0935379267,-0.074974604,-0.4186442494,0.1443356574,-0.0522535145,-0.1282395571,0.121255748,-0.4481029809,-0.4539896846,-0.2699021995,0.5651443005,0.1653032899,0.0896883756,0.3432316482,-0.1958156079,0.009230705,-0.1956038326,0.0348324627,-0.1255518347,-0.3250435591,0.5315566063,-0.2140324712,-0.1011318862,0.1889260113,-0.173754245,0.1105415151,0.3808698356,0.1637355983,-0.2844932377,-0.5291988254,-0.3904183209,-0.0524902157,0.2282985598,0.050847441,-0.0072952523,-0.0107132103,-0.0045601041,-0.3350776434,0.2403877079,0.0919069722,0.0729334131,0.1242790818,-0.1043476984,-0.2560699284,-0.2592730224,0.1775953174,0.3209490776,-0.0254106764,-0.2779419422,-0.3296145797,0.0958329737,0.3220802546,-0.1090267152,-0.1992628872,-0.0858467817,-0.2128356099,-0.1136852056,0.2098957598,0.4146845043,-0.0505065322,-0.0152493529,0.3320708871,0.0022462029,-0.1997736692,-0.095850639,-0.2755773664,-0.1480358839,0.0696581528,0.2725877762,-0.0361606181,0.1273908317,-0.4385435283,-0.0649156943,0.1195174456,0.0885495842,0.1913103312,-0.2284960747,-0.1663822532,-0.0860028714,0.2218851149,0.4013639688,0.0489362255,-0.1580483168,0.2325664312,0.2389936596,-0.5628596544,-0.0106937625,0.1943591684,-0.003649774,0.098564215,0.1986977458,0.2104046792,-0.1681738794,-0.0613068901,0.1347883046,0.2547552884,-0.0625275373,0.3317419291,0.6826942563,0.1762798131,0.248175472,0.2562503219,0.1975747496,0.164253071,0.329222858,0.0229687337,-0.0241935551,-0.2815636098,0.0184473433,-0.379707098,-0.3709614277,0.3160880208,0.4231722653,-0.2143074274,0.236592263,0.0353621766,-0.3099908233,-0.0264861919,0.0299492758,-0.2805517018,0.2032455802,-0.1236360446,-0.3274005949,0.3878690004,-0.2026110739,-0.1777496487,0.1964985281,-0.0410761125,0.0262600426,-0.1356980056,-0.0048754462,-0.2721968591,-0.4071184695,-0.0643913075,0.1155254096,0.1304227114,-0.2057668418,-0.0984744281,0.3018858433,-0.1364281178,-0.0018364664,0.2399721444,0.3788969219,0.2111078948,0.1534903944,-0.0022816143,0.1738818735,-0.0088148732,-0.3157631755,0.4200020134,0.0662739351,-0.0240946896,-0.1370903552,0.203938961,-0.2506957352,-0.0812061727,0.1522063017,0.0141364988,-0.1547079533,0.4992974699,0.0719960332,-0.3092764318,-0.0027599297,0.2752628028,0.0251966976,-0.079818204,0.3267326951,-0.1332303882,0.2819019258,-0.2300074697,0.1109809577,0.1148270667,-0.034752652,0.4555131197,-0.4898700416,-0.2097949684,0.0988615081,-0.816400528,0.0120940423,0.0328483358,0.2008297592,-0.1417099983,0.0833700895,-0.0083174789,0.0212943275,0.1145128533,0.3138949871,0.0724570304,0.3102653921,-0.0539459884,-0.1346898079,-0.0958660915,-0.1477009207,-0.1163652837,-0.5594716668,0.0759588927,-0.1853275597,0.2095306814,-0.037059404,0.0065100822,0.2180145383,0.2444203049,0.4945931435,0.0287282132,-0.0332200117,-0.0520022735,-0.1335050911,-0.5665888786,-0.039800819,-0.2466407716,0.0500838161,0.1520656943,0.1390793025,-0.3625163436,-0.3033446372,-0.2144049257,0.3089033067,0.2004514635,-0.0365050435,-0.3225580454,-0.0226420909,-0.0864753276,0.0237421617,0.2618903816,0.2817649543,-0.3526063561,0.3924575746,-0.3913657367,-0.3119346499,0.2078100592,-0.3488782942,-0.3597286046,-0.1109973863,0.2195085585,-0.059010148,-0.0838459656,0.080401659,0.3842255771,0.2200256139,-0.1292812973,-0.2282675058,0.1182789207,-0.403898716,0.1522310525,-0.0926432759,0.1583367139,0.1514638215,-0.2153838128,0.1477641314,-0.0796994865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2902","title":"Add WIT Dataset","comments":"@hassiahk is working on it #2810 ","body":"## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":6,"text":"Add WIT Dataset\n## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n@hassiahk is working on it #2810 ","embeddings":[-0.1296591461,-0.1891998202,-0.2115951777,-0.0000080944,0.0460982546,0.0357038304,0.4529203176,0.1215223595,0.2780340016,0.1513489336,-0.029059099,0.1144830734,-0.1435055733,0.2445029169,0.1256047338,-0.1864110827,0.0068955352,-0.0093742143,-0.0735348091,-0.1070073843,-0.0773350447,0.1636914909,-0.1567373872,-0.2530801892,-0.4015705884,-0.1087995097,-0.146388337,-0.0190194007,-0.1731389165,0.0142502328,-0.2311714441,0.0920530111,-0.1416087747,0.4282389283,-0.0000961792,-0.0348447226,0.1924369335,-0.1517420858,-0.0877345204,-0.1121518463,0.0036920926,-0.0255717319,-0.2709282935,-0.3336799145,-0.1762119234,-0.1874346882,0.2792802751,-0.038891498,0.2992183864,0.1237806827,0.3313299716,-0.1193304285,0.167231366,-0.0636842549,0.1344261765,0.1325438321,-0.2437879592,-0.1032884791,0.0572029762,-0.1843628138,-0.2473023236,0.6803191304,0.1149657592,0.0394390561,-0.0199051965,0.0259543676,-0.0537805147,-0.2517703772,0.1902595907,0.4509062171,0.5951299071,-0.1380434632,-0.1741087288,0.0608816259,-0.0706331432,0.0042669997,0.0225002747,0.3247715533,-0.1227583364,0.053050328,-0.1813238263,-0.1304020733,-0.1617643535,0.2954109609,-0.1025401354,0.3974758983,-0.0176089592,-0.0563631766,-0.0496043675,-0.1633326858,-0.2125443816,0.02245792,0.240905866,0.0170114562,0.0630257428,-0.129550159,0.054668501,-0.1277145147,0.3031286895,-0.2512640655,-0.039158728,0.1198865026,-0.3077050745,0.2759459615,0.1043220907,-0.2439876199,-0.1338686198,-0.1969237626,0.2190607339,0.0052139987,-0.2566974163,0.1069463864,0.1158736348,-0.0193225145,-0.1939061284,-0.1762420088,0.0073827389,0.1123326346,0.0784640163,-0.0886906609,-0.124503158,-0.055896312,-0.3169686496,0.2779684961,0.1109271422,0.0749763697,0.0985479653,0.2271314412,-0.1937326193,-0.1465904415,-0.115335606,0.1956622303,-0.1542165428,0.1144959927,0.1543435156,0.5087482929,0.2681051195,-0.0842734575,0.0915060565,0.0306925531,0.1029897705,-0.0939951912,0.1980237216,-0.0445276089,0.1020599678,-0.1297828108,-0.0630924329,-0.288815707,-0.1791048646,0.048106607,-0.0815851614,0.0402405933,-0.4307381809,0.2774644792,0.1309566647,-0.1127759367,0.1063605249,0.7211184502,0.0591425896,-0.0758869126,0.1704614013,0.2683287561,-0.3786454499,-0.0175030548,0.1434833258,0.3921846449,-0.1378667355,-0.0694735646,0.1149419025,0.2635501623,-0.0323280878,0.2083413899,-0.0748511776,0.0976687744,0.0117110927,0.2452786714,0.1070185974,-0.1362077594,-0.0952911451,-0.1215822399,-0.0758547634,0.0205433052,0.1006369516,0.3590261936,-0.0074512311,0.0786975026,0.0008037383,0.6429982781,-0.101937674,0.2204965353,-0.2547368109,-0.412805438,-0.031592954,0.3136671782,0.1404488534,-0.3794309199,0.3183832169,0.2673033774,0.1137722209,-0.2929435074,0.1085533649,0.0377520844,-0.0135402372,0.1493370086,0.0558074787,-0.2918334305,-0.0843841732,-0.1060758308,0.1630552709,0.4458846748,-0.1963102072,-0.2546880245,0.0034701391,-0.1770533621,-0.0542313494,-0.2643735409,0.3854094744,0.2371143699,-0.1219256148,0.2192856669,0.0515208244,-0.0828982666,-0.2751990557,-0.0682251155,-0.0554965064,0.1411690116,-0.2138494253,-0.0074874591,0.0453312024,0.162361756,-0.2836136818,0.0503010452,0.1026869044,0.2066954374,0.0853403509,0.3449212611,0.3921612203,0.1152452454,0.2651148438,-0.5831927061,0.0691323876,0.0750517696,0.0019045175,-0.0033749938,-0.3692139387,0.0319884233,0.2239721268,-0.0128715681,0.1168907061,0.0606278479,0.2472451031,-0.0976875871,0.001401014,-0.3726579845,-0.0102998558,0.162639454,-0.1726347953,-0.1651332676,-0.2103906274,0.0991069302,0.5462095737,0.2400736511,0.5376595855,0.0971913859,-0.1212614775,-0.2618568838,0.073506847,-0.2561023533,-0.0930368453,0.3516949713,0.0182377622,-0.1866730601,-0.0738957152,-0.1027568281,0.1412557811,-0.0225975774,0.1650807112,0.0877017826,0.297613889,-0.1069796085,-0.5398007035,-0.0681409612,-0.0508023016,-0.2064673901,0.0649895668,0.0703736097,-0.0205571149,-0.2685602605,-0.0878075063,0.1285372823,-0.2698201537,-0.0652996004,0.3480983078,-0.058700081,-0.144668743,0.2157517225,0.2154152393,0.2665415406,0.1056036651,-0.0490454584,-0.0699434429,-0.4251724482,-0.0013740142,0.2867534161,0.2798976302,0.0790984556,0.468409121,-0.1529595554,0.019352071,-0.2396499366,-0.6391475201,0.1912251115,-0.1763069183,0.2191690207,0.0712262616,0.0936918706,-0.1991403848,-0.2139495015,0.1469445527,-0.1367393434,-0.1214991733,-0.0671442822,0.0764834508,0.0814846978,-0.1663844138,-0.2625015378,-0.1379565299,-0.3415474296,0.2776284814,0.2449719161,0.1046371013,-0.0269246474,0.3842811882,0.0516841039,-0.0223082379,0.2923726141,-0.3290747404,-0.029114319,0.1770710647,-0.3507783413,-0.3409646749,-0.241635114,-0.2149017453,0.2913351059,0.0963894725,-0.452812463,-0.3489452302,0.0134124234,0.2051300406,0.1298554689,0.076393947,0.282122463,0.1753068119,-0.3087322414,-0.1432863772,-0.0113858152,0.0432145521,-0.1399844736,0.0589655936,0.1224881411,-0.0000328349,0.2479563504,0.4598668814,-0.2399621606,0.0897038355,0.4043138623,-0.015272229,0.2053738832,-0.2565110326,-0.1264286041,0.4514034986,0.0070677064,0.0613207631,0.30798769,0.1180268228,-0.4592045248,-0.0890442282,0.174862057,-0.5091278553,-0.1667373627,0.1673555821,0.1669268012,0.1635840684,-0.1377123594,-0.2422106862,-0.1948408335,-0.1597946584,0.1669388264,0.2362881452,-0.0687819272,-0.1252325177,-0.2005453855,0.0608569533,-0.1316801757,0.0990372747,0.2825544477,-0.123838447,-0.1642559022,-0.161138624,-0.0587822609,-0.1694849581,0.2437805086,-0.3938077092,-0.1649195999,0.1007589772,-0.0390315913,-0.1142330319,0.2840626538,-0.3339577615,-0.2328336537,0.0622188188,0.2880499065,-0.3703322113,0.0837903321,0.4168727994,0.1480499059,-0.0266231615,-0.0325266942,-0.1320562661,-0.2437737137,-0.4280227423,0.0575218908,-0.0885785222,0.1218918115,0.3068802357,-0.0476196334,-0.0697411075,-0.0784421787,-0.0828517675,0.2587498724,-0.0900649056,0.3041183949,-0.1872166842,-0.138946712,-0.0720472485,-0.1653235257,0.1967177689,-0.0868958831,-0.0660888404,-0.1588893533,-0.2913936675,0.0738343373,0.0961803868,0.2457379699,0.0848319456,-0.0663812011,-0.0204767045,-0.3289154172,0.2266960442,0.2494198382,0.1433276385,-0.0007672762,-0.2630187869,0.3372961879,-0.0651260763,0.0367082134,0.3666113317,-0.0947311148,-0.4259827435,0.3040920198,0.3670836389,0.9130336642,-0.1412870288,0.2655439973,0.0825656652,0.0368013196,0.4112624526,-0.2702554166,0.0926808119,0.063380003,0.0450684801,-0.042350892,0.2032682449,-0.0341339484,0.0403719842,-0.2530496418,-0.1733001024,0.3581714332,0.1081538126,0.0419844128,0.3453404605,-0.0444174036,-0.4325706661,-0.3199236393,0.2508710325,0.060147658,-0.1543311328,-0.0326714627,-0.0739602372,0.0480866395,-0.0830819085,-0.3174287677,-0.0862657726,-0.2957362831,0.1991055459,-0.3531002998,-0.3360367119,0.1190520227,0.1976566911,0.1793552339,0.1488007456,-0.2352505326,0.1980030835,-0.1115192845,-0.136059463,-0.0225254111,-0.1173774973,0.3119098544,-0.1344185174,-0.1809410751,0.0406746306,0.2653738558,-0.2495407313,-0.4706448913,-0.1169789657,0.3314366341,0.1948352754,0.0485579409,0.2501215041,-0.413610667,-0.2901029289,0.2998668849,-0.1534142941,0.1727076322,0.0644296631,0.234110564,-0.0873228759,-0.0742139593,0.2355686873,0.2749872208,0.0819436833,0.2199673355,0.2496194392,-0.1277595311,-0.4075970948,-0.2303928286,0.0248409584,-0.3103953898,-0.08661329,0.0937269107,-0.3579361737,-0.1972269416,0.2842388749,0.0858474672,-0.0124888159,-0.0472396538,-0.0732503757,-0.0512179583,0.4948905706,-0.1107646599,0.2192007452,0.0759131759,0.1513412446,0.1365086883,0.2520558536,-0.5044214129,0.0618388243,-0.1622692049,0.1555999666,0.1093536094,-0.1887407005,0.0446312688,0.1271794736,0.297516793,-0.0032108144,-0.1694126576,-0.3926710784,-0.1479100138,0.0459250957,0.00995667,-0.0352526307,0.1959041208,-0.0861305743,-0.0785156563,-0.2286321074,0.1431579441,-0.0590985827,-0.221162349,0.2085623145,-0.1347504705,0.1107792184,0.1136089787,-0.1726914197,-0.0558690876,0.2453998476,0.1603109837,-0.1366208494,0.2261314541,-0.1058626622,0.074613668,0.22358042,-0.1432922632,0.1838464141,0.210678488,0.1371192336,0.0155197922,0.1411829144,0.2470877767,-0.1947020739,-0.338175416,-0.0111634769,0.168082267,0.4273661077,-0.3924217224,-0.0953474045,0.2009430379,-0.4014649987,0.4477969408,0.2921489179,0.1483340859,-0.002726747,-0.0974504873,0.067161195,-0.1852512658,-0.3622380793,-0.0507756211,0.0865137428,0.0177711882,0.0699414983,-0.1085267439,0.0994353145,0.2459332198,0.3117552102,-0.0358871408,0.4007888138,-0.0078987358,0.0618646182,0.3345318437,0.2172142565,0.4865316749,0.1046640351,0.0416013673,0.3430802226,-0.0083274813,0.2223719209,0.0999372602,-0.0087033538,-0.1931295693,0.1871576011,-0.0184282884,-0.1056677848,-0.275636971,-0.1552461386,-0.1231860667,0.0284091607,-0.0035716088,-0.0306761824,0.1527671069,0.0474604145,-0.1027932689,-0.1923794597,0.0851862952,-0.0813491642,0.4018681645,-0.2289473712,0.2052509189,0.2295409143,-0.0062884679,0.0375786684,0.4565461576,0.1032586545,-0.0071680066,-0.0938500986,0.0924604014,-0.1135674641,-0.0638433546,-0.2580824494,0.0472246632,0.3923101425,-0.0717268735,0.2726536989,0.267921716,-0.3474231064,-0.1523036957,0.4223014116,-0.2082197219,-0.0883307829,-0.5123988986,0.0371011719,0.0079047736,-0.2581389248,0.0969417989,-0.5972567797,-0.0656792223,0.0051606284,-0.0808611438,0.0543405078,-0.102023758,0.1447420567,0.1186306775,0.4803176522,0.2922181487,0.1430308968,0.0741488263,-0.3569711447,-0.473400563,0.1692283154,-0.1185174435,-0.0800530314,-0.1962443441,0.3564409912,0.296964705,0.2779829502,-0.0123385582,-0.1476554275,-0.0388901271,0.0978951603,-0.3775681555,0.1366155446,0.1053182557,0.1353593618,-0.0266050361,-0.1613134742,0.0928840712,-0.0417491868,0.2913910151,-0.1669008136,0.0616998039,-0.0978731364,-0.0602027848,0.2010447979,-0.0838134512,0.1094563082,-0.4794981182,-0.3175835311,-0.1471890509,-0.1113790944,-0.1586659849,0.2541112304,0.2663548291,0.4172587097,-0.0143395821,-0.0704477504,0.2176796794,0.4207631946,0.0117856711,0.1037914306,0.0464718007,-0.0413710028,0.2674077749,-0.1472931802,-0.1270537078,0.4860852957,-0.1144139916,-0.1948803961,-0.0594344959,-0.5313600898,0.0803394318,-0.3131714165,-0.2420081943,-0.0785194188,-0.0384221189,0.0641538873,0.2597292364,-0.3967080116,0.1920229495,0.2368176877,0.2900294065,-0.0751894638,0.3233103752,0.1524246633,-0.1856960952,-0.1501988024,-0.0521546677,0.0246847868,-0.1107614785,-0.2156294882,-0.4013026655]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2902","title":"Add WIT Dataset","comments":"WikiMedia is now hosting the pixel values directly which should make it a lot easier!\r\nThe files can be found here:\r\nhttps:\/\/techblog.wikimedia.org\/2021\/09\/09\/the-wikipedia-image-caption-matching-challenge-and-a-huge-release-of-image-data-for-research\/\r\nhttps:\/\/analytics.wikimedia.org\/published\/datasets\/one-off\/caption_competition\/training\/image_pixels\/","body":"## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":23,"text":"Add WIT Dataset\n## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nWikiMedia is now hosting the pixel values directly which should make it a lot easier!\r\nThe files can be found here:\r\nhttps:\/\/techblog.wikimedia.org\/2021\/09\/09\/the-wikipedia-image-caption-matching-challenge-and-a-huge-release-of-image-data-for-research\/\r\nhttps:\/\/analytics.wikimedia.org\/published\/datasets\/one-off\/caption_competition\/training\/image_pixels\/","embeddings":[-0.0497860387,-0.0547444522,-0.1280666143,0.0003004618,-0.039354112,-0.0071313963,0.3546046019,0.1659145951,0.1936855763,0.2055810541,0.0639799759,0.1460067928,-0.0242924541,0.1997424215,-0.01487609,-0.2061530054,-0.0474179126,0.007556899,-0.095439963,-0.0953547359,-0.1237470731,0.1472279876,-0.1375917941,-0.2112537324,-0.4580724537,-0.0875768587,-0.1730553955,-0.0253455658,-0.1721293628,-0.0776184052,-0.1209003329,0.1993556619,-0.1435845345,0.3342963159,-0.0000973471,-0.1190001816,0.2309069782,-0.1802841574,-0.0871033221,0.0576613285,0.1233595088,-0.0183559302,-0.3197696805,-0.4497412145,-0.1917415559,-0.0919262245,0.250259757,0.0872264057,0.220723629,-0.0171945672,0.3280394673,-0.1210444793,0.10394793,0.0242980495,0.0408663712,0.1121114418,-0.1330121011,-0.0007004917,0.0150091052,-0.3246178627,-0.1604316086,0.6847470999,0.0171727017,0.0272535756,0.0614412166,0.0563659444,-0.1721417308,-0.3297246695,0.1790350825,0.4349500835,0.5310005546,0.0022286586,-0.1660662293,-0.0531980433,-0.0875593573,-0.0650093853,0.1432961375,0.2999428809,-0.1368227452,0.0239426252,-0.2666973472,0.0028924709,-0.2545390427,0.1593592912,-0.1959471852,0.4131114483,0.021566527,0.0326524712,-0.000182682,-0.2101522833,-0.4369372725,-0.0127320085,0.2125476599,0.0099950507,0.0381090157,-0.2509273589,-0.0503862202,-0.1546090245,0.3692227602,-0.197378248,-0.0403584838,0.0733242184,-0.1825137436,0.3001660109,0.1151571646,-0.273580879,-0.1728782058,0.0215191487,0.1827842295,0.0874059945,-0.2466638684,0.1344642043,0.2253025919,-0.1029728055,-0.1772990078,-0.087800771,0.1070196927,-0.0466616042,-0.1075610965,-0.008583298,-0.146704793,-0.0710129738,-0.4360583127,0.2165165991,-0.0335759073,0.1098506823,0.1278970838,0.1473859996,-0.2188078016,-0.1533710063,-0.1057867333,0.1009146869,0.0557398014,0.2790454626,0.1211024597,0.3311381638,0.256734401,-0.0824029595,0.1746457219,0.0947210044,0.0416478626,-0.0953107029,0.2555246055,-0.008912893,0.0164618678,-0.1227277219,-0.051241383,-0.2793621719,-0.19902426,0.0270652734,-0.1084727868,0.2036321461,-0.3212899566,0.2668367922,0.1446974277,-0.0583522059,-0.0045630517,0.6982100606,0.0959759504,0.0308421906,0.1365310401,0.1445246041,-0.3958842158,-0.0636466146,0.2453516275,0.4490688741,-0.0886282399,0.0255282167,0.3061005771,0.3564617038,0.0193696469,0.1932205856,0.0245506912,-0.0181516893,0.0488758124,0.2630175054,-0.1056210548,-0.2358684838,-0.1195387468,-0.1759907305,-0.0965119898,0.1020767167,0.0326437689,0.2860510647,0.270050317,-0.102775909,0.093534559,0.6013006568,-0.1892949641,0.2137656063,-0.2130067647,-0.3863318861,-0.0069002835,0.3273807466,0.1285638511,-0.4685770869,0.2250231653,0.1909397244,0.0319985636,-0.3845872879,0.0987228006,0.0075334026,-0.0310356412,0.0995845795,0.091828078,-0.2398949265,-0.0079521695,0.0546789393,0.2049163729,0.4832212925,-0.1773444712,-0.2481343299,-0.038995903,-0.1513748169,-0.1065134704,-0.3698706031,0.3537430465,0.1205859855,-0.0045493012,0.2210574299,0.0784497336,-0.0594155379,-0.3163442016,-0.0456017256,0.1199062467,0.1867598891,-0.2623721361,-0.0185750853,0.1218658164,0.0860536918,-0.0835613608,0.1807654053,0.0911751837,0.1913248599,0.0461769477,0.3455418348,0.2677955925,0.0530583225,0.3559015393,-0.6900479198,-0.0690085068,0.19123438,0.0030610359,0.0122600449,-0.2659728527,0.1819274873,0.2717046738,0.0452171303,-0.0804450437,-0.0083080204,0.1473466903,-0.0244393963,0.0583699085,-0.3646619022,-0.0810375884,0.2129996717,-0.2515609264,-0.2326057106,-0.1890244186,0.1808756441,0.5132712126,0.2433172613,0.3687907755,0.1684178561,-0.1199678779,-0.1401048452,-0.0068703438,-0.0960154831,-0.1688125134,0.3742109239,-0.053772375,-0.1260635257,-0.2724704742,-0.0922951698,0.1780956537,-0.0158920027,0.2231312841,0.0486644618,0.2492239773,-0.0249909423,-0.4577436149,-0.1271143258,-0.1574495137,-0.1973033249,0.056282267,0.0784645602,-0.071331881,-0.1991609186,-0.066964902,0.2833610475,-0.3087125421,-0.1143085882,0.3522223532,0.0152183259,-0.1103117839,0.183313325,0.0873240456,0.1733572483,0.1862344891,0.0245582648,-0.1337936968,-0.5455780029,-0.0611162968,0.2686133087,0.1743589044,-0.0161512867,0.5377413034,-0.1781012118,0.0367738679,-0.2884244919,-0.705337286,0.1020467132,-0.1840780228,0.1969555616,0.0155186988,0.2760653794,-0.2234164178,-0.1541128606,0.1546294391,-0.1017461941,-0.1935486495,-0.2209588885,0.0480773337,0.0420998633,-0.2291868478,-0.2148961276,-0.1442073584,-0.2447485179,0.1066020951,0.3489588201,0.1721870601,0.0647365227,0.3838174343,0.0280661602,-0.0238699075,0.2764138281,-0.3571163714,-0.032202661,0.3682462573,-0.3039095998,-0.403814584,-0.1645279825,-0.182767272,0.1699132323,0.0591959544,-0.3852038085,-0.3932056427,0.0160926636,0.1964574009,0.2606021762,0.0776553899,0.1688962281,0.1696974039,-0.2888213396,-0.120130457,-0.0905002728,0.1087626144,0.0804397017,0.0327175856,0.2156327516,0.0916362107,0.2265037298,0.6132438183,-0.1613393426,-0.0281569362,0.4119825959,-0.0383804776,0.224104315,-0.0002331532,-0.0705114752,0.497284919,0.0091163954,0.1581450105,0.2489970326,0.1093951836,-0.4575267434,-0.1425068378,0.0741071776,-0.4309840202,-0.1831286401,0.1126235873,0.282223165,0.1359028369,-0.1255722344,-0.1118768379,-0.1950071603,-0.2140990347,0.1501378864,0.1365459114,-0.033066839,-0.0854402483,-0.1478022784,0.0440795906,-0.2099582404,0.1326003671,0.3808392882,-0.1713135391,-0.1235297993,-0.064189747,-0.1482826024,-0.2265735269,0.2456169724,-0.3066806197,-0.0396218561,-0.0001861092,-0.0057738582,0.0171943922,0.2211831212,-0.2088258266,-0.3443025649,-0.0452317037,0.247772336,-0.4293949008,0.0881709754,0.3906086981,0.1875751317,-0.0982594416,0.0725138634,-0.1535127908,-0.4466047287,-0.4639799893,0.0759280697,-0.0179230105,0.234052375,0.1692747027,-0.0461039618,-0.0495606475,-0.0345803201,-0.0313565917,0.3599346876,0.0232606847,0.2451539487,-0.1866826117,0.0159258991,-0.0362759829,-0.0392687432,0.2646231949,0.0655591339,0.0620991588,-0.1904884726,-0.1798860431,-0.0081319464,0.1325878799,0.2491253167,-0.0325292051,0.0099018356,0.0323803313,-0.3869562447,0.2138520479,0.1015860438,0.1107611284,-0.0148324464,-0.2682086527,0.3575321436,0.0139691914,0.0896827802,0.389649421,0.0827483088,-0.4077329636,0.13139759,0.606361568,0.8766955137,-0.1784271747,0.1968854815,0.1156533733,-0.122769326,0.490136683,-0.0125109544,0.0893897936,-0.0903045982,0.0462060049,-0.0482571162,0.1810766608,0.0233861692,-0.0071799215,-0.2383960932,-0.0591488257,0.4950783849,0.0509889983,0.0223183818,0.2454401106,-0.0180004686,-0.3712224662,-0.1638911664,0.2396210581,-0.1347939372,-0.0386185609,0.0205545817,-0.1282645911,-0.0563069582,-0.2408725768,-0.2661218941,-0.0375264436,-0.3923751414,0.2052637786,-0.2520484924,-0.232131511,0.0540006123,0.1925714016,0.1538775563,-0.0118546505,-0.1931021065,0.2699621618,-0.2166959643,-0.1297280639,0.0994486734,-0.0778556168,0.3134611547,-0.1621257216,-0.2310339063,0.0682638064,0.3530699909,-0.2003534883,-0.3784057498,-0.1596205533,0.2148566991,0.3098063767,-0.0714116693,0.0368100926,-0.4080921412,-0.2550562024,0.2662607431,-0.1720845252,0.1280960441,0.0953225568,0.338157177,-0.0661718771,0.0120415185,0.279966861,0.1513243467,0.0692156926,0.252928853,0.1970508397,-0.1593777239,-0.262548238,-0.1428267211,-0.2143792063,-0.3051334023,-0.1352979839,0.0671666935,-0.2043531686,-0.1338025331,0.3641974926,0.1625054628,-0.0255595706,0.0916185379,-0.0791770592,-0.0435318984,0.3351586759,-0.2645081282,0.2620086372,0.1362607479,0.0919233859,0.0621007904,0.2466659248,-0.4697705805,0.0008081495,-0.276823312,0.220438391,0.2841959,-0.1862068176,-0.0280307755,0.125368759,0.3021201789,0.0017990991,-0.1685612202,-0.3627264202,-0.2431438267,0.0422772169,0.0334107876,-0.1346720457,0.0254573394,-0.0790322646,-0.0836148933,-0.1622535139,0.0931934118,-0.2240975499,-0.2472412735,0.2228255719,-0.1823137701,0.0700710714,0.094846718,-0.1861076504,-0.0950848833,0.4156052172,0.0535985343,-0.1083418354,0.2208644897,-0.1395664066,-0.0260134172,0.2909347117,-0.2630017996,0.0133082671,0.2766338587,0.1419247389,-0.0343248211,0.2441204339,0.2616045177,-0.21285294,-0.2781764269,-0.070929803,0.2034056783,0.3885403872,-0.3480496109,-0.0002685353,0.3647504449,-0.4026501775,0.4058004618,0.3052506745,0.1680756807,-0.0470950566,-0.258131206,0.0765385777,-0.0169942752,-0.4340025783,-0.020303281,0.1415560097,0.1105765998,0.0487362295,-0.0601711161,0.159247458,0.2790450156,0.2702627778,-0.0576715805,0.4003246725,0.0700719953,0.0536593162,0.2820592821,0.2163243592,0.5008888245,0.1520393342,0.0615330748,0.3375333846,0.0667119846,-0.0489498749,0.0231666267,-0.1415083408,-0.1176150665,0.3134403527,-0.0147500196,-0.0715856478,-0.3975217044,-0.2155854702,-0.0502537452,0.0992040187,0.0855665654,-0.1450577676,0.2354638726,0.0524963774,-0.1196645424,-0.2443861812,-0.0430718996,-0.0664650947,0.5451594591,-0.1991377473,0.0696593449,0.2658621371,-0.0356095135,0.0963545516,0.402122885,0.1985036433,0.0789157301,-0.2381747067,0.1088349149,-0.0445786901,-0.1440981328,-0.1425260752,-0.0305382255,0.147759378,-0.2748593986,0.349375397,0.2631461918,-0.3019365072,-0.3211148977,0.3872309923,-0.1345878541,-0.0100665605,-0.6193216443,-0.002628427,0.0443207249,-0.1521770805,0.1048784405,-0.5363394022,-0.0952254608,-0.0885432363,-0.0321343392,0.0277196635,-0.1648291796,0.1338630617,0.0752426833,0.4413732886,0.2891284525,0.0606763028,0.0619646944,-0.2279345095,-0.5278967023,0.116753839,-0.0799406394,-0.4166087508,-0.1129243895,0.3505662084,0.2454901189,0.2385322601,0.1003209874,-0.1506654024,0.2423653901,0.0709742904,-0.2875417769,0.0808582902,0.0561882034,0.0541797392,-0.0941887349,-0.1719164848,0.1251953542,-0.2471843213,0.3233294487,-0.241175577,0.1120429486,-0.1701765209,0.0729581118,0.0646916777,-0.0272890981,0.1814373881,-0.4356124699,-0.3198751509,-0.0362593234,0.0115888026,-0.1519787014,0.1135638803,0.156106621,0.5797201991,0.0692493692,-0.0713019744,0.1066633314,0.4916709661,0.0201849677,0.0295590572,0.0948422104,0.031026803,0.2517523468,-0.1997822076,-0.078859888,0.396351546,-0.0907165185,-0.1283352971,-0.055131916,-0.620397687,0.1190891713,-0.439666748,-0.1874820441,-0.0373212919,-0.1397859603,0.1088624224,0.3996539116,-0.3055769801,0.2087804824,0.2238700241,0.1930085868,-0.2440024018,0.3604778051,0.1155558676,-0.1429820657,-0.0835569948,-0.1805303544,-0.0429826528,-0.0756496042,-0.2174596637,-0.3298113346]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2902","title":"Add WIT Dataset","comments":"> @hassiahk is working on it #2810\r\n\r\nThank you @bhavitvyamalik! Added this issue so we could track progress \ud83d\ude04 . Just linked the PR as well for visibility. ","body":"## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":28,"text":"Add WIT Dataset\n## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n> @hassiahk is working on it #2810\r\n\r\nThank you @bhavitvyamalik! Added this issue so we could track progress \ud83d\ude04 . Just linked the PR as well for visibility. ","embeddings":[-0.0809517056,-0.2214551866,-0.1989731491,0.0133092012,0.0548566766,0.0012881586,0.4317961633,0.0905278698,0.2972510159,0.1470830888,-0.0140026174,0.117820628,-0.1801292151,0.2750623822,0.1759230644,-0.2347300649,0.0528856255,-0.0525587946,-0.111166127,-0.0794195235,-0.0522569418,0.1980129927,-0.193837896,-0.2477924973,-0.398671031,-0.1057200357,-0.1739888638,-0.0169518292,-0.1846639961,0.0327739008,-0.2608135045,0.0733176619,-0.1156978458,0.4770053923,-0.0000964032,-0.064579986,0.2143842131,-0.1423904598,-0.0545602031,-0.1777072847,0.041535154,-0.0049116011,-0.2549009025,-0.2813225985,-0.1839788407,-0.1538777053,0.2700455189,-0.0029812963,0.3053423464,0.1210975945,0.3088106811,-0.1474685222,0.2071530968,-0.0514375716,0.1326429099,0.1544086039,-0.248274833,-0.0964364931,0.0966394767,-0.2267570943,-0.2421554029,0.6854541302,0.1160565093,0.0250154659,-0.0252820924,0.0070958599,-0.0926502421,-0.2954958975,0.188639313,0.4700401127,0.5937217474,-0.1630618572,-0.1721827239,0.0919661224,-0.0487032495,0.0397691727,0.0408558957,0.3571046293,-0.1618400067,0.0649592951,-0.1783177257,-0.15133214,-0.1726877838,0.2914958894,-0.0949109718,0.4078658223,-0.0315455608,-0.0555812605,-0.0373772085,-0.1777147949,-0.2164238542,0.0523082763,0.2419596165,0.0662366971,0.0335867517,-0.1256911457,0.0385495387,-0.1135115251,0.3417511284,-0.2774721086,-0.0565746911,0.132901594,-0.323112458,0.3092396557,0.1385564804,-0.2174253464,-0.1450333446,-0.1854815036,0.1951016188,0.0325002782,-0.2619886994,0.1505577266,0.099721089,-0.0214241892,-0.2273956388,-0.160768792,-0.0516468845,0.0911392644,0.1193964332,-0.0707594156,-0.1421585381,-0.0661800429,-0.3177126944,0.2407010645,0.1010043994,0.0815588385,0.0965300798,0.2475896478,-0.1470988691,-0.1765169948,-0.0881534964,0.2248412669,-0.1903703958,0.1265021414,0.1592800766,0.5113365054,0.2537501454,-0.0733771846,0.1129089221,0.0018325694,0.0856605247,-0.1001399159,0.2069840133,-0.0557915792,0.0809679776,-0.1524138749,-0.073373422,-0.3102889955,-0.1555388868,0.0773518682,-0.0889911428,0.049377501,-0.4161147475,0.2530652583,0.1284849197,-0.0809181705,0.0908475667,0.7699179053,0.0582486838,-0.0342353247,0.1798475385,0.2833342552,-0.3797134161,-0.0025154876,0.1757738888,0.3905650079,-0.092420809,-0.0642552599,0.1328477561,0.2665217221,-0.0583941862,0.2173905224,-0.0567309335,0.087856859,0.0246495996,0.2322876602,0.0694603547,-0.1395289451,-0.0435003601,-0.1011100262,-0.0813610405,0.0106059145,0.0836546645,0.298145026,-0.0127935633,0.096528627,0.0080791311,0.6753434539,-0.1098164916,0.2286004871,-0.2737777531,-0.4278278649,-0.0113982931,0.328744024,0.145377025,-0.4055054188,0.3278430998,0.2322438657,0.0949867368,-0.3136895597,0.1085856259,0.0223203544,-0.0005051232,0.1467118561,0.1092848554,-0.2964729071,-0.0836762711,-0.130402267,0.1318242103,0.483746618,-0.248289302,-0.2480282038,0.0042855954,-0.2053590268,-0.0621923134,-0.2704664171,0.3617103696,0.17163001,-0.1306215972,0.2724723816,0.0430581756,-0.0919279084,-0.2533151209,-0.0622685514,-0.0406291224,0.1581876874,-0.2046279311,-0.0177770704,0.0472491309,0.1987008452,-0.2961709499,0.0533348955,0.1038889065,0.1865651459,0.0792258084,0.3698846102,0.3873916864,0.1120804548,0.2921239138,-0.6291121244,0.1154151708,0.0760285705,-0.006126096,-0.0176064018,-0.3841995299,0.0472670086,0.19283472,-0.0016580714,0.1144336909,0.073393099,0.2081537247,-0.0956783444,-0.0049937302,-0.366854459,0.0016061328,0.1734439284,-0.1884854436,-0.1824615896,-0.2030260712,0.0807694122,0.5237088799,0.2258421183,0.5505995154,0.1254282445,-0.1420795768,-0.28885746,0.0549856015,-0.2941262424,-0.1655779034,0.3116272688,-0.0251247585,-0.1782427877,-0.0482349657,-0.085527949,0.1338160634,-0.0276928395,0.1904964596,0.0914436951,0.2914770544,-0.1149811968,-0.5630301237,-0.0705270022,-0.0223842766,-0.2100635022,0.0955632254,0.0532289781,0.0085637094,-0.2813546956,-0.0914483368,0.1090581045,-0.2741340697,-0.0460894257,0.3448168337,-0.0612961911,-0.1088433415,0.2226289511,0.2646664679,0.2659244537,0.0535228476,-0.0327079035,-0.0569329187,-0.4480122924,-0.0012578141,0.2744467258,0.3094526529,0.0774371698,0.4492533505,-0.1706188321,-0.0032700298,-0.1970399022,-0.6828343272,0.1931857914,-0.1799332798,0.222297132,0.0465350263,0.0598002523,-0.2162089497,-0.1784383208,0.1267048568,-0.1550560743,-0.1278322637,-0.0514384434,0.0643266812,0.0791968629,-0.171182245,-0.2651809454,-0.1608145237,-0.3226848841,0.3241012394,0.2365675569,0.11344897,-0.0567815825,0.3845591247,0.0518014841,-0.0231321547,0.2783186138,-0.3127382994,-0.0589283146,0.1515590996,-0.3646040261,-0.3476807475,-0.2444108576,-0.1988812536,0.2525679469,0.0808433667,-0.4292116463,-0.3522478342,0.0311818309,0.2061463892,0.1419826597,0.0842591301,0.2448365092,0.190408811,-0.2793112099,-0.1548068672,-0.0041617351,0.0255124755,-0.1157210916,0.0760340467,0.1783276647,0.0139507679,0.2965382338,0.4465199113,-0.2176057845,0.1476778239,0.3842280805,0.0036174529,0.1714651883,-0.2575018108,-0.1082883179,0.5114590526,-0.0155883245,0.0532231331,0.3055923581,0.1387230456,-0.493773073,-0.1047745943,0.1593842357,-0.5112806559,-0.1331510544,0.1892561316,0.1403896511,0.1600528061,-0.1513719559,-0.3017922044,-0.175591737,-0.1879560053,0.1834862083,0.255494982,-0.0663223863,-0.0857872739,-0.1768642962,0.0369856954,-0.121113494,0.0471084453,0.3025898635,-0.1103605628,-0.1657053381,-0.1256040633,-0.0691812932,-0.1423523277,0.2711536586,-0.3945429325,-0.1802496612,0.1024449319,-0.0627269968,-0.0776354223,0.2761702836,-0.330293119,-0.2617784739,0.0633344948,0.3113533854,-0.3719662428,0.0762865916,0.4057539105,0.1531731486,0.0045682145,-0.0382465012,-0.0839486346,-0.2303083241,-0.4416763484,0.0756094828,-0.0475242175,0.1286154091,0.2946779728,-0.0733427703,-0.0314382799,-0.0484548584,-0.0901324525,0.2570098937,-0.0899744332,0.2848593593,-0.1690306813,-0.1259722412,-0.1103044376,-0.1840956509,0.1707034856,-0.0802198648,-0.0567748919,-0.2014106512,-0.2966367304,0.0954907387,0.0731393769,0.2481430173,0.0807809383,-0.0638768226,-0.0167128611,-0.3351784945,0.1609588414,0.2288006246,0.2169885933,0.0056775264,-0.2819871008,0.3445268273,-0.073021248,0.039783787,0.3688854277,-0.1323992908,-0.4168664217,0.3294692636,0.3667802513,0.9066294432,-0.1412285864,0.2516909838,0.0521108657,0.0443725064,0.379629761,-0.2828923464,0.0604830906,0.0479294434,0.0563858896,-0.0548547022,0.1923683733,-0.0117860539,0.0325935893,-0.275164932,-0.1745013148,0.3504127562,0.1096888706,0.0581242368,0.3463208973,-0.0501323305,-0.4585259557,-0.2810158432,0.2333678752,0.0736545548,-0.1327342242,-0.0186843965,-0.0988537818,0.0398562253,-0.0720824152,-0.3393507898,-0.0706250221,-0.3228843808,0.2183895111,-0.308116734,-0.3472005725,0.1668397635,0.2316637933,0.2130430043,0.1262043267,-0.2604316473,0.2059072405,-0.0819388479,-0.1436516792,0.0000350554,-0.0936641917,0.364066422,-0.1577276438,-0.1965638399,0.0312893242,0.2719175816,-0.1914389282,-0.4962573647,-0.1763629317,0.3231757283,0.2469760776,0.0594019704,0.2734151781,-0.420081526,-0.279900074,0.2923044562,-0.1240278035,0.139734447,0.090626359,0.2457036376,-0.0870910808,-0.0481114164,0.2409695685,0.2811488807,0.0817728937,0.2042908669,0.3132542372,-0.1231459677,-0.4118822515,-0.2268985659,0.0112832189,-0.3043795824,-0.0997028574,0.075029701,-0.3576362729,-0.1853453815,0.2977905869,0.0945064574,-0.0108859651,-0.0280812215,-0.104707174,-0.0504857972,0.4871656299,-0.1233899519,0.1704572886,0.0498322472,0.1368637681,0.1661450267,0.2778989673,-0.4913753569,0.0999070555,-0.1608650386,0.1617065072,0.1140983626,-0.2118062377,0.0391213,0.1311940104,0.2852498591,-0.0028897552,-0.1713041216,-0.3745980859,-0.1606842279,0.0489701405,0.0406194441,-0.0218042042,0.2007368803,-0.1021691784,-0.075192146,-0.2093645781,0.170140624,-0.0332107544,-0.235269323,0.1865300387,-0.1608070433,0.1187050864,0.1141011491,-0.1843488514,-0.0202733669,0.2034232765,0.1748362631,-0.1693175435,0.2350271195,-0.090575017,0.0716959387,0.2352731675,-0.1609387249,0.1508632898,0.22368294,0.1156171039,0.046740178,0.0889115855,0.2442272305,-0.2031773925,-0.3575416207,0.0098930569,0.2128424644,0.3950456977,-0.3788528442,-0.1032667682,0.1637059599,-0.3952121437,0.4144688547,0.2580401301,0.1857469529,-0.0102851801,-0.1041494682,0.0706411228,-0.1911442429,-0.3604948521,-0.0429542065,0.0436507836,0.0383596979,0.0862421021,-0.1259579062,0.1100242585,0.2090563923,0.2760358453,-0.0548093542,0.3306506574,-0.0164462365,0.0951900482,0.3387245536,0.2413451821,0.4937801957,0.0987096876,0.0264338162,0.3570222855,-0.0418578535,0.2647001147,0.1094167084,-0.034785334,-0.1574991196,0.187232703,0.0186632369,-0.1193159148,-0.2749849856,-0.1580919325,-0.1344575584,0.0357709043,-0.0121378247,-0.0608970672,0.1425048411,0.0659468845,-0.1128230691,-0.201325193,0.0615972467,-0.0594317168,0.4502527118,-0.2346452177,0.2587008178,0.1951896548,-0.0046939291,0.0203922391,0.4529597461,0.112104319,0.019977089,-0.1175721884,0.0875313953,-0.114308767,-0.061771743,-0.1920992732,0.0724602044,0.3947676718,-0.0750949383,0.2475585938,0.236138463,-0.3471648097,-0.1523065567,0.4345367253,-0.1978103966,-0.0468036309,-0.5462716818,0.0078052361,0.0408355705,-0.2892584205,0.0690495074,-0.637196064,-0.0430860594,-0.0008575172,-0.0671569929,0.0473668277,-0.0951450542,0.1405508965,0.1236739159,0.4421503842,0.3087599874,0.1302765012,0.0961523876,-0.3838952184,-0.446438849,0.165471822,-0.1000231653,-0.0665667132,-0.1452491432,0.321154058,0.2846552432,0.2771971524,-0.0085254181,-0.1527163386,-0.0146443518,0.1127069667,-0.3975604475,0.1767693162,0.0961188823,0.1993730366,-0.0299789477,-0.1019918174,0.1034432575,-0.056199532,0.2615951002,-0.2036397755,0.102962628,-0.1265001148,-0.0658517629,0.1697728336,-0.1092779413,0.0920383409,-0.4555152953,-0.3111367822,-0.1391869038,-0.1233393773,-0.1475510448,0.2659368515,0.2873288691,0.4175904393,-0.0261233281,-0.1063955426,0.2146098614,0.4109917879,-0.0069877491,0.081328705,0.0300202575,-0.0731000528,0.3030554354,-0.1591516882,-0.098505877,0.4811171591,-0.1448982358,-0.2175184041,-0.0372310281,-0.5459349155,0.0817107856,-0.3308123648,-0.2261419147,-0.0765321776,-0.0824959278,0.0465272143,0.2357187569,-0.4232021868,0.1949178725,0.2078907192,0.3215321004,-0.0627836511,0.3240291178,0.1668283939,-0.1704506576,-0.1509878784,-0.0996923894,0.043273434,-0.0744077042,-0.1939665377,-0.382758081]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2901","title":"Incompatibility with pytest","comments":"Sorry, my bad... When implementing `xpathopen`, I just considered the use case in the COUNTER dataset... I'm fixing it!","body":"## Describe the bug\r\n\r\npytest complains about xpathopen \/ path.open(\"w\")\r\n\r\n## Steps to reproduce the bug\r\n\r\nCreate a test file, `test.py`:\r\n\r\n```python\r\nimport datasets as ds\r\ndef load_dataset():\r\n ds.load_dataset(\"counter\", split=\"train\", streaming=True)\r\n```\r\n\r\nAnd launch it with pytest:\r\n\r\n```bash\r\npython -m pytest test.py\r\n```\r\n\r\n## Expected results\r\n\r\nIt should give something like:\r\n\r\n```\r\ncollected 1 item\r\n\r\ntest.py . [100%]\r\n\r\n======= 1 passed in 3.15s =======\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\n============================================================================================================================= test session starts ==============================================================================================================================\r\nplatform linux -- Python 3.8.11, pytest-6.2.5, py-1.10.0, pluggy-1.0.0\r\nrootdir: \/home\/slesage\/hf\/datasets-preview-backend, configfile: pyproject.toml\r\nplugins: anyio-3.3.1\r\ncollected 1 item\r\n\r\ntests\/queries\/test_rows.py . [100%]Traceback (most recent call last):\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/runpy.py\", line 194, in _run_module_as_main\r\n return _run_code(code, main_globals, None,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/runpy.py\", line 87, in _run_code\r\n exec(code, run_globals)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pytest\/__main__.py\", line 5, in \r\n raise SystemExit(pytest.console_main())\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/config\/__init__.py\", line 185, in console_main\r\n code = main()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/config\/__init__.py\", line 162, in main\r\n ret: Union[ExitCode, int] = config.hook.pytest_cmdline_main(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_hooks.py\", line 265, in __call__\r\n return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_manager.py\", line 80, in _hookexec\r\n return self._inner_hookexec(hook_name, methods, kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 60, in _multicall\r\n return outcome.get_result()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_result.py\", line 60, in get_result\r\n raise ex[1].with_traceback(ex[2])\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 39, in _multicall\r\n res = hook_impl.function(*args)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/main.py\", line 316, in pytest_cmdline_main\r\n return wrap_session(config, _main)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/main.py\", line 304, in wrap_session\r\n config.hook.pytest_sessionfinish(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_hooks.py\", line 265, in __call__\r\n return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_manager.py\", line 80, in _hookexec\r\n return self._inner_hookexec(hook_name, methods, kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 55, in _multicall\r\n gen.send(outcome)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/terminal.py\", line 803, in pytest_sessionfinish\r\n outcome.get_result()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_result.py\", line 60, in get_result\r\n raise ex[1].with_traceback(ex[2])\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 39, in _multicall\r\n res = hook_impl.function(*args)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/cacheprovider.py\", line 428, in pytest_sessionfinish\r\n config.cache.set(\"cache\/nodeids\", sorted(self.cached_nodeids))\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/cacheprovider.py\", line 188, in set\r\n f = path.open(\"w\")\r\nTypeError: xpathopen() takes 1 positional argument but 2 were given\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Linux-5.11.0-1017-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":19,"text":"Incompatibility with pytest\n## Describe the bug\r\n\r\npytest complains about xpathopen \/ path.open(\"w\")\r\n\r\n## Steps to reproduce the bug\r\n\r\nCreate a test file, `test.py`:\r\n\r\n```python\r\nimport datasets as ds\r\ndef load_dataset():\r\n ds.load_dataset(\"counter\", split=\"train\", streaming=True)\r\n```\r\n\r\nAnd launch it with pytest:\r\n\r\n```bash\r\npython -m pytest test.py\r\n```\r\n\r\n## Expected results\r\n\r\nIt should give something like:\r\n\r\n```\r\ncollected 1 item\r\n\r\ntest.py . [100%]\r\n\r\n======= 1 passed in 3.15s =======\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\n============================================================================================================================= test session starts ==============================================================================================================================\r\nplatform linux -- Python 3.8.11, pytest-6.2.5, py-1.10.0, pluggy-1.0.0\r\nrootdir: \/home\/slesage\/hf\/datasets-preview-backend, configfile: pyproject.toml\r\nplugins: anyio-3.3.1\r\ncollected 1 item\r\n\r\ntests\/queries\/test_rows.py . [100%]Traceback (most recent call last):\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/runpy.py\", line 194, in _run_module_as_main\r\n return _run_code(code, main_globals, None,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/runpy.py\", line 87, in _run_code\r\n exec(code, run_globals)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pytest\/__main__.py\", line 5, in \r\n raise SystemExit(pytest.console_main())\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/config\/__init__.py\", line 185, in console_main\r\n code = main()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/config\/__init__.py\", line 162, in main\r\n ret: Union[ExitCode, int] = config.hook.pytest_cmdline_main(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_hooks.py\", line 265, in __call__\r\n return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_manager.py\", line 80, in _hookexec\r\n return self._inner_hookexec(hook_name, methods, kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 60, in _multicall\r\n return outcome.get_result()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_result.py\", line 60, in get_result\r\n raise ex[1].with_traceback(ex[2])\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 39, in _multicall\r\n res = hook_impl.function(*args)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/main.py\", line 316, in pytest_cmdline_main\r\n return wrap_session(config, _main)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/main.py\", line 304, in wrap_session\r\n config.hook.pytest_sessionfinish(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_hooks.py\", line 265, in __call__\r\n return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_manager.py\", line 80, in _hookexec\r\n return self._inner_hookexec(hook_name, methods, kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 55, in _multicall\r\n gen.send(outcome)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/terminal.py\", line 803, in pytest_sessionfinish\r\n outcome.get_result()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_result.py\", line 60, in get_result\r\n raise ex[1].with_traceback(ex[2])\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 39, in _multicall\r\n res = hook_impl.function(*args)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/cacheprovider.py\", line 428, in pytest_sessionfinish\r\n config.cache.set(\"cache\/nodeids\", sorted(self.cached_nodeids))\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/cacheprovider.py\", line 188, in set\r\n f = path.open(\"w\")\r\nTypeError: xpathopen() takes 1 positional argument but 2 were given\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Linux-5.11.0-1017-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nSorry, my bad... When implementing `xpathopen`, I just considered the use case in the COUNTER dataset... I'm fixing it!","embeddings":[-0.3016421199,-0.1903872788,-0.0241492949,0.0990293771,0.3130396903,-0.1301406622,0.3114988804,0.2646481693,-0.1026777923,0.2515563369,-0.0329412185,0.3579744101,-0.0499672778,0.0878717452,-0.1619878709,0.1268500686,-0.0505254902,0.0685187206,0.0408935249,0.1136002094,-0.5712682605,-0.1570851803,-0.2168759555,0.0572490133,0.0126794586,0.0686679184,0.0916296914,0.2422822416,-0.0774931237,-0.469889611,0.1960368752,-0.0263328627,-0.0422579646,0.4722329378,-0.0001045304,-0.0628838241,0.4584931433,0.2026191652,-0.4859764278,-0.2095862329,-0.1490363926,0.2063343376,0.2531500459,-0.254386425,-0.1224498227,-0.2215161026,0.0800599158,-0.5224384069,0.0328101404,0.2298583686,0.2646005154,0.5858446956,0.1764907539,-0.1220323741,0.3920012712,-0.1865620166,-0.0988033265,0.1512855589,0.2791418135,0.0342658088,-0.1552184969,0.2738155425,-0.251668334,0.2581939995,0.0830629617,0.0403149687,0.1071799397,0.1236744151,0.0054618004,0.1477268934,-0.0244684517,-0.2447356284,-0.187163204,0.0132809207,-0.1560323685,-0.4709234536,0.0660294145,0.2990931869,-0.1927743256,0.0707391798,-0.1533978879,0.285312593,-0.2178352475,0.2086598277,-0.0414643139,0.474942416,0.0106467595,0.0582296215,0.0282520335,0.0535728522,0.3718392253,-0.029089909,0.0258877985,-0.2166382521,-0.0353300683,0.0447735675,0.3098430037,0.1299089044,0.1545516998,0.1684173644,-0.0050124396,-0.0647071525,0.1099802256,0.1257055998,-0.0156705379,-0.0932240486,-0.0586434491,0.1788802147,0.1195266545,0.1648669541,-0.1360280216,0.0084639816,-0.0668128058,-0.3070941567,0.0900237709,0.1294367164,0.5126132965,-0.0637784451,-0.3585409522,-0.1637833863,-0.3168658912,0.1315581501,-0.0887979344,0.1646843106,-0.3344549835,0.0525158457,0.2567434609,0.1665857583,-0.3985979259,0.1734841466,-0.1729390174,-0.0025560693,-0.2071772516,-0.1981868595,0.0244949311,-0.0061054123,0.0369944014,0.0258494634,-0.0169168096,-0.0553664789,0.1729714125,-0.0304495171,0.0978282169,0.0834876001,0.1038077101,0.0230855085,0.0646957457,0.2365556359,0.0004667089,0.4194874167,-0.1334887296,-0.2440327555,0.1164916679,0.35601753,0.0041353889,-0.1565836817,0.1820517033,-0.3767896891,-0.1220365986,-0.0800681487,0.0026528577,-0.2985853255,-0.1478585154,-0.2639125884,0.2254217416,0.2425370067,0.2605780959,0.0505344905,-0.0282764584,0.1627342254,0.3797311783,0.1669939458,-0.1939092427,-0.0896540061,-0.1379573345,0.1596079618,-0.0486291014,-0.4239594936,-0.3278156817,-0.1607752889,-0.1731502265,-0.07503663,0.049631156,0.0323371701,0.3179516196,0.0479337424,-0.0318585075,0.412953496,-0.0740917623,0.0748518556,-0.4666415751,-0.3798174262,0.5641200542,0.151705578,0.0570301823,-0.3497416377,0.2054489851,0.0608398058,0.5148956776,-0.1897210926,-0.1518411934,-0.2993411422,0.752466023,-0.0474503674,0.0110379141,-0.0960698351,0.0665274039,0.0204419531,0.2316295356,-0.2802456915,-0.1499119997,-0.0891542956,-0.1524517834,0.1319255084,-0.3466832638,-0.382157594,0.2474724054,0.2782579362,-0.1456098258,-0.1309039593,-0.0118953492,-0.0933069214,-0.3045151532,0.100494273,0.0386145003,0.1053284183,-0.0396609642,-0.094124943,-0.1193556041,0.298748225,0.162939176,-0.0600543693,-0.1179709211,0.3312662542,0.2449479997,0.1955460757,-0.0468190089,0.0072500408,0.0820179358,-0.3156889975,-0.2024822384,0.3221007586,0.0975408554,-0.002581625,0.0227513239,0.1893991232,-0.2354014963,0.1002926081,0.2986403406,0.0012304113,-0.0390900932,-0.1708397567,0.0159268808,0.3352193236,0.253539145,0.4091952145,0.4384313226,0.0471034944,-0.1185913086,-0.0291307271,0.1720729768,0.0163347404,0.2327632755,0.1700844765,-0.400013119,-0.0777249038,-0.0843988508,0.3033127487,0.3952056468,0.275578618,0.0969723538,0.0887242258,-0.1324126124,-0.2870429754,0.2992976308,0.1152055711,0.0208725408,0.2415834963,0.2297437042,0.1021427438,-0.2554797828,-0.0732267201,-0.1389924884,0.0566167571,-0.1460091919,-0.0478304811,-0.1640127003,0.2011270672,-0.1140680909,-0.5005605817,0.1816411167,-0.2008605003,0.0143492073,0.3810865879,-0.2022752911,0.1200969592,-0.2770578265,0.0942122191,0.0885685533,-0.0884958655,0.1415922791,-0.2295171022,-0.1760160774,0.1794952601,0.0489556827,-0.1769523323,0.3923769593,-0.0935229883,-0.0265750755,-0.1780978739,-0.1293669194,0.0228035133,0.2426382601,0.0753103569,0.111499086,0.05481565,0.108430557,-0.2810976505,0.3347675204,-0.2257572263,-0.0527905673,0.1222594678,0.0974359363,-0.0859619454,-0.1513888091,-0.6303583384,-0.1750367284,-0.3355152905,0.0044383388,-0.107812658,0.1289740354,0.4354717731,0.0387976281,0.1618680805,-0.1536655724,0.121643737,-0.0872593373,-0.0211115107,0.2169368863,-0.1349027008,-0.4010963142,-0.198347047,-0.1213339865,-0.1045445204,0.433926791,-0.2609218061,-0.2592196167,-0.2313274145,0.0872525126,-0.1294534057,-0.0232153423,0.3986389339,0.2484317124,-0.2569616735,-0.2864581048,0.0643375963,-0.0433718897,-0.0686287582,0.0381864421,-0.087544404,0.1517151892,0.1127801165,0.2772788107,-0.0392005183,-0.0387176163,0.1129376143,-0.2495963275,0.1978779882,-0.0433227047,-0.256916225,0.1621052921,-0.2219971567,-0.3214866221,0.0484665148,0.2195314169,0.1148819104,0.0141728139,0.0208034664,-0.3543073237,-0.2621160448,-0.2463587523,0.3780513704,0.0837339088,-0.0672197193,0.1626053154,-0.1316152364,-0.1256516725,0.1881820112,0.3759931922,0.1093446314,-0.0119339237,0.2482949048,0.0123445923,-0.0862110257,0.0998784602,0.0090176854,0.1281881332,0.0566877872,-0.1912219673,0.0785223171,-0.0200078636,0.164327085,0.3458769023,0.387481302,0.3010594249,-0.1472473443,-0.2551456392,0.0189636461,-0.5182095766,-0.3018001318,0.0700561032,0.1623141319,-0.2750703692,0.0511011817,0.3745272458,0.1937805861,0.0627759919,0.0518822372,0.0086369924,-0.4112387002,-0.0766517594,0.0733634308,-0.0133935297,0.4511975646,-0.036897406,-0.2186553776,-0.3560897112,-0.333645165,-0.0098668383,-0.0695359781,0.410104692,-0.161490202,0.1817578077,0.0064769732,0.3672036827,0.3949998319,0.21570158,0.1406856179,0.1515156925,-0.007653513,-0.1830333024,0.0305453707,0.1035782769,-0.352021575,0.2259793282,-0.3184482455,0.2777011096,-0.2909503579,0.1956375539,0.3502264917,0.1843060404,0.2732339501,0.0076183779,0.1056670472,-0.1933567673,-0.0941651911,0.5531533957,-0.1504649967,-0.2331054062,-0.1227031499,0.1427025497,0.798979938,0.2282719463,0.1575280577,0.8578522205,-0.379422754,-0.1998089552,0.3360408843,0.1782059669,-0.3571628928,0.0383414142,0.0812166259,-0.228775233,0.0273007564,-0.2691111565,-0.0185519699,-0.1856172085,-0.0207111835,0.0446326248,0.1084755659,0.0523586944,0.040253371,0.0424530245,-0.2928382456,0.1929394305,-0.0179150142,0.1270106137,-0.0508781746,-0.0839509666,0.2480623722,-0.1010967717,-0.2043585628,0.3419599831,-0.5799904466,0.1721819341,-0.2692230642,-0.4562358558,-0.1075189933,-0.0367897227,-0.2285494804,0.1678056121,-0.034031637,0.1088864431,-0.025018584,0.1675112695,0.2350722551,0.0307645444,0.2065542489,-0.2243107855,0.0097956462,-0.1305310428,0.0516777299,-0.4160594046,-0.0600328259,0.0467677265,0.494930625,-0.2594941854,-0.2402201593,-0.2332240641,-0.1814632267,-0.3135359883,0.2200746089,-0.0131482948,0.1351002008,0.1709565818,0.0870712101,-0.3995954692,-0.1094917506,-0.0123062516,0.0119483415,0.1383818239,0.4005638361,0.1243489757,-0.3258515,-0.2226490825,0.1260189861,0.1640535593,-0.0271104369,0.0868695006,-0.0412157513,0.1951783746,-0.1118135676,0.0058581447,0.2094234228,0.0991211757,-0.1979446858,-0.4055492878,-0.1660560668,0.1417739689,0.2028817385,0.1181820855,-0.0264829546,0.2930473983,-0.0213118494,-0.250485301,-0.3906231225,0.1107914075,0.2336137593,0.1192180067,-0.1454875469,-0.2076455355,0.3068408668,-0.0104081798,0.2542708516,-0.051687777,-0.1347726136,-0.3061599433,-0.3753732443,0.0859213248,0.0782093406,-0.3529872596,0.1614413857,-0.1700136662,0.0037516817,-0.1876564473,0.1137397587,0.3402270079,-0.2084238827,-0.0727987811,0.3792553246,0.1400492191,-0.2362306863,0.0073809871,-0.2387416065,0.2812373936,-0.1352637857,-0.0331278555,0.1245579347,-0.0565308221,-0.1441801339,-0.2388965189,0.2134499103,0.1605943888,0.2762319148,-0.1227833033,-0.11529129,-0.1517725885,0.5579532981,0.198146686,-0.1018952206,-0.0254687592,0.0657821447,0.377563864,-0.2160446793,-0.0246563014,-0.1456676871,-0.1745386422,0.1609793305,-0.06970983,-0.4467020631,0.0724190399,-0.1336331218,0.0843940005,-0.0011720849,-0.1661704034,-0.0122293616,0.1776824147,-0.1440191418,-0.0499875061,0.0935866535,-0.0745285526,0.1813107729,0.2331058681,-0.0133844428,0.6520751119,0.5496276617,-0.2443293333,0.0571718849,-0.0549315177,0.0069860495,-0.1236407459,-0.0362573415,0.0326864682,0.0409415178,0.2288151681,-0.2829382718,0.1047509387,-0.0511482842,-0.0435396358,-0.0990097001,-0.2635801733,0.0496004447,-0.0047050342,-0.0728645399,-0.0293682516,-0.1020355001,0.0224539246,0.0303376559,0.0476331823,-0.1048749685,-0.3494413197,0.0213044882,0.2079137266,0.1573521048,-0.3372322619,0.0876498446,0.218774423,0.143193692,-0.1703726351,0.4733753502,0.330448091,0.4719161391,-0.2293340415,-0.1304816902,-0.1675468981,-0.0063679451,-0.0394412205,-0.0647637993,0.013415969,0.1269380897,0.2298047692,0.3690746725,-0.3679998517,0.2118322849,-0.0242782552,-0.2714368403,-0.1849635541,0.2795254588,0.1443216503,-0.2951877415,0.0125996275,-0.0279961135,-0.2422025651,-0.0668567345,0.2900865078,0.2645173967,0.0876090899,-0.0397080444,0.1404092163,0.1440564841,0.1232805997,0.1019615456,-0.007067997,-0.4234194756,0.0328579731,-0.4711897075,0.4904586673,-0.1441804022,-0.446585238,0.0725421757,-0.0549177974,-0.2503592968,0.2345560342,0.3334582746,0.3571134806,0.1524520814,-0.0025301664,-0.255433023,0.1322989017,0.2028841376,-0.1032729968,-0.1260176003,-0.3461897969,0.2244306803,-0.2395664901,0.2582560182,0.0934694856,0.0410272479,0.2902907431,0.0138179325,0.017829489,-0.0170892756,0.395065397,-0.320538342,0.0409307852,-0.1458014548,-0.2275678366,0.0252156015,0.0532084107,0.0172631126,0.0858793035,0.1734698564,-0.5986049175,-0.434243381,0.4738562405,-0.3005673587,-0.1423755139,0.1085444987,0.3310374022,0.0011340193,0.1555804163,-0.037431892,0.1749423742,-0.1264825761,0.0589323528,-0.2458039671,-0.7254006267,0.2943550944,-0.7327613235,-0.2969195843,-0.2304133028,0.460382551,0.0212493073,0.238044709,-0.2867960334,0.1030114964,0.2128991187,0.0478451774,-0.1016833112,0.1259424686,0.017026538,-0.023835374,0.1139790639,0.6136496067,0.1038713083,-0.0976273566,-0.0312423818,-0.282785058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2892","title":"Error when encoding a dataset with None objects with a Sequence feature","comments":"This has been fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/2900\r\nWe're doing a new release 1.12 today to make the fix available :)","body":"There is an error when encoding a dataset with None objects with a Sequence feature\r\n\r\nTo reproduce:\r\n```python\r\nfrom datasets import Dataset, Features, Value, Sequence\r\ndata = {\"a\": [[0], None]}\r\nfeatures = Features({\"a\": Sequence(Value(\"int32\"))})\r\ndataset = Dataset.from_dict(data, features=features)\r\n```\r\nraises\r\n\r\n```python\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 2 data = {\"a\": [[0], None]}\r\n 3 features = Features({\"a\": Sequence(Value(\"int32\"))})\r\n----> 4 dataset = Dataset.from_dict(data, features=features)\r\n[...]\r\n~\/datasets\/features.py in encode_nested_example(schema, obj)\r\n 888 if isinstance(obj, str): # don't interpret a string as a list\r\n 889 raise ValueError(\"Got a string but expected a list instead: '{}'\".format(obj))\r\n--> 890 return [encode_nested_example(schema.feature, o) for o in obj]\r\n 891 # Object with special encoding:\r\n 892 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n\r\nTypeError: 'NoneType' object is not iterable\r\n```\r\n\r\nInstead, if should run without error, as if the `features` were not passed","comment_length":19,"text":"Error when encoding a dataset with None objects with a Sequence feature\nThere is an error when encoding a dataset with None objects with a Sequence feature\r\n\r\nTo reproduce:\r\n```python\r\nfrom datasets import Dataset, Features, Value, Sequence\r\ndata = {\"a\": [[0], None]}\r\nfeatures = Features({\"a\": Sequence(Value(\"int32\"))})\r\ndataset = Dataset.from_dict(data, features=features)\r\n```\r\nraises\r\n\r\n```python\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 2 data = {\"a\": [[0], None]}\r\n 3 features = Features({\"a\": Sequence(Value(\"int32\"))})\r\n----> 4 dataset = Dataset.from_dict(data, features=features)\r\n[...]\r\n~\/datasets\/features.py in encode_nested_example(schema, obj)\r\n 888 if isinstance(obj, str): # don't interpret a string as a list\r\n 889 raise ValueError(\"Got a string but expected a list instead: '{}'\".format(obj))\r\n--> 890 return [encode_nested_example(schema.feature, o) for o in obj]\r\n 891 # Object with special encoding:\r\n 892 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n\r\nTypeError: 'NoneType' object is not iterable\r\n```\r\n\r\nInstead, if should run without error, as if the `features` were not passed\nThis has been fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/2900\r\nWe're doing a new release 1.12 today to make the fix available :)","embeddings":[-0.0830345005,-0.2502439916,-0.0386255719,0.3015390933,0.2076890171,0.0788367391,0.5001555681,0.1494158357,0.1255292594,0.1421060264,0.3285635114,0.0508400798,-0.2967574298,0.3420606554,-0.2623355389,-0.2604982257,0.034535449,0.2224871665,0.0154916933,-0.0224474147,-0.3130667508,0.1709361374,-0.364877075,-0.2221626192,-0.1550774723,-0.2108083814,-0.2160241902,-0.2960246801,0.0542513058,-0.5987132192,0.2819482088,-0.0695153251,0.0921094567,0.2383219302,-0.0001144155,0.105492115,0.6381921768,-0.1453159899,-0.2628047466,-0.3394742906,-0.5618312359,-0.4227804542,0.1241655424,-0.3578473926,-0.0965121984,-0.0749715865,0.094718039,-0.3724971116,0.2042126656,0.2839667201,0.2130593061,0.3830007613,-0.0845509171,-0.0744024739,0.206820786,0.4162615538,-0.2607547045,-0.1798537225,0.035758879,0.2699487805,0.0946282744,0.3492396772,-0.111404337,-0.1305068284,0.2094347328,0.0849353001,0.0885619074,-0.4521939754,-0.2671251595,0.2014953941,0.2924095094,0.0013618435,-0.4084641933,-0.2180884778,0.0188714378,-0.6638943553,0.2666681111,0.055101525,-0.0386278071,-0.0269636735,-0.168392688,0.145001635,-0.2138954103,0.1660171747,-0.1209752634,0.0807909966,-0.1394540668,0.0341406651,-0.2402791679,-0.2700186968,0.0587494336,-0.3706248403,-0.0257517211,0.2111456394,-0.2515126467,-0.1936304271,0.167350471,0.0101849167,0.1697200388,0.1379934251,-0.0658706501,-0.1037984863,-0.0693587288,-0.0169596635,0.2529113591,0.2136024684,0.4112964869,-0.0680698305,-0.0581751354,0.1961243749,0.0996622145,-0.017598724,0.11928422,0.0517393723,0.0169261098,0.4072842002,0.6467193961,-0.0525628068,-0.393481791,0.2361950129,-0.3853594661,-0.0304983035,-0.0211958643,0.0061908085,0.0841203332,0.2916485071,-0.1465796083,0.2161677331,-0.2736037076,-0.250882864,-0.2348464429,0.0721954107,0.1018402427,0.0885985568,-0.0547403768,-0.0332995802,0.0658167601,0.1560105532,0.0796802491,0.1009850726,-0.0555230267,-0.365808934,0.0925061703,0.1523720324,-0.0070467223,0.2921590805,0.3762358427,-0.2581280768,-0.1091617644,0.1182429716,-0.2788110375,-0.217513755,-0.005827894,0.1725294441,-0.3795679808,0.0441264361,-0.0734036565,0.2263460904,0.2108575404,-0.0775901377,-0.0965437591,-0.0869289562,-0.2196164131,-0.1535144448,-0.0919654146,0.6465163827,-0.1091333926,-0.012587267,0.1191491932,-0.0898958445,0.3521232605,0.2327210456,0.0481479168,-0.057248231,-0.3404105306,0.1900297105,0.0831347257,-0.1115483418,-0.208607167,0.5657577515,0.0239907876,0.3598697484,0.1134639457,0.0959460512,0.1585700363,-0.3072297871,0.0130883651,0.1975988597,-0.1874651164,-0.2206795961,-0.2418985963,-0.1231627762,0.5206927657,-0.0743467882,-0.0640905127,0.2067834437,-0.3010617197,-0.0231112521,-0.1406468004,-0.3186881542,0.1288141012,0.5053259134,0.21923922,0.0375964157,-0.0967492238,-0.4028446674,-0.362685889,0.1312113106,-0.179603979,0.2839995623,-0.4932679534,-0.1799385846,0.0735202953,0.019041175,-0.3947464824,0.0015028021,0.1547810137,0.1214925349,-0.1535111219,0.1999227107,-0.0773415267,0.2772004306,0.1561118215,-0.1248554364,-0.567322731,0.3699874282,0.0695009977,-0.2262255251,0.0071669328,0.3033728004,0.3317838013,-0.0899984464,0.018318845,0.1505862623,0.0259702709,-0.3309904039,-0.3067805767,-0.1571382731,0.3190588653,-0.4185928106,-0.2214330137,0.3747457862,0.1825433373,-0.1073410809,-0.1209880039,0.6575422287,0.0157377589,0.4647912979,0.0042069959,0.1760141253,0.0537278354,-0.0018615762,-0.327142626,-0.2869090438,0.1755582988,0.2026411891,-0.1650790572,-0.0399117619,-0.3015944362,0.0058686477,0.4108926356,0.088436991,0.0354627408,0.0570802502,-0.0770579055,-0.0368382186,0.1092406586,-0.159593001,0.299883306,0.0505586714,-0.0732196867,0.0735782608,0.0075443899,0.0745387301,0.1758682579,-0.0350369215,0.2430823594,0.0285679307,0.1076480895,0.1176982298,-0.1779721677,-0.5438489318,-0.0398581289,0.1430890262,-0.2491215467,0.2492818087,-0.4795846939,-0.2056158483,0.130416885,-0.2173653096,-0.386295557,-0.4440228939,-0.1639413238,0.0275585894,-0.14083983,0.2236978859,0.0148140779,-0.26245597,0.2985749543,-0.0243482068,-0.0831376389,-0.1392291635,-0.0242327861,0.0463199466,0.1656642258,-0.0481024235,0.2467772514,0.0332959294,-0.3499837816,-0.2126868665,-0.2085345984,0.0565498397,-0.3503640592,0.0777079239,0.3079494238,0.1277272999,-0.04940686,-0.1998873949,0.4227285385,0.4243725836,-0.2120994478,0.3217695653,-0.102062583,-0.0354735814,-0.2110650837,-0.0959021896,-0.1572214514,-0.2278930247,0.1076047793,0.2109689265,0.2101610452,0.4471023381,0.0470193811,0.120879285,0.1355805248,-0.175628379,0.0008855638,-0.0150361629,0.4824748039,-0.1827880889,-0.2712480724,0.1487995982,-0.2987255454,0.0875612497,0.2041173279,-0.3575526774,-0.1763042659,-0.1413789541,0.1611307114,0.0745360553,-0.1046581939,0.3775742948,0.2816057205,-0.1083455756,-0.0367697217,-0.0508110113,0.2569004595,-0.0705505088,0.0287486054,0.229687199,0.5177567601,0.0398100056,0.0972354859,0.4376426637,0.1970631927,0.4261662662,-0.0823322833,0.0689360127,-0.1690145433,0.0005423338,-0.1711448282,0.107939221,-0.0178557411,0.0836275071,-0.2327573746,-0.1191914901,-0.0793068409,-0.0630706996,0.1038516238,-0.2357092947,0.129370749,-0.4676784277,0.442966491,-0.0864085481,0.0182908159,-0.3036096394,-0.0047005508,-0.1684736162,-0.0083024697,0.322756052,-0.1478672475,-0.64961797,-0.172231406,-0.2026044726,0.1215542853,0.3011380136,0.5467792153,-0.0337624177,-0.0624163561,-0.1075646654,-0.0191717129,0.4020332694,-0.1643890142,-0.1837625653,0.2454210073,-0.2638014555,-0.2387931347,-0.0567740202,-0.3692710102,0.3344596624,-0.2252197117,0.3618203402,-0.3405365348,-0.2225623876,0.2660458982,0.0742505565,-0.0913682804,-0.1132186204,-0.0910779387,-0.223561734,-0.1456012279,0.0129503747,0.2087670118,0.262157619,-0.1340866685,-0.0023913924,-0.5279824138,0.0307506509,0.3058232069,-0.2428039312,0.3555576503,-0.0859422684,0.1051599681,-0.1363119334,0.1940722018,0.2076593935,0.3552796543,0.377152741,-0.648413837,0.0111574912,-0.2274465859,0.1626977772,0.2772224844,-0.0934667066,0.0936513171,0.0833641216,0.1915255636,-0.2102062851,-0.3317123055,0.4189606011,-0.0805432349,-0.6726543307,-0.2872206569,0.2649118006,0.0967709273,0.2421812862,-0.1336611062,0.2232288569,-0.2113029659,0.3653744459,-0.0063009313,0.6850805283,0.4925385714,0.0419002809,0.5505322814,0.447142154,0.6043536663,0.4047942162,0.2582224309,-0.1625066251,-0.6721657515,-0.1485058814,-0.0585912243,0.4232234955,0.1191022769,-0.4531514645,0.489102751,-0.1281993687,-0.0638250262,-0.0981585309,-0.2965756059,-0.1317236423,-0.0439694487,-0.2515955269,0.1245702505,0.0693780556,0.1620190293,0.0573866554,-0.3246811926,0.0522958711,-0.1667520255,-0.1947969943,0.169654116,-0.1433667392,0.0941162854,0.0364244618,-0.3463888764,0.2818099856,0.2232395709,0.1879022419,-0.1312100291,-0.1560928375,-0.1969094723,0.0400298387,0.1034141481,-0.0801590979,-0.0777151361,0.3071270287,-0.0268944614,-0.1485314518,0.1475492716,-0.2142958194,0.1054213047,0.3429113925,0.0590512119,-0.1748204976,-0.1611669809,-0.453250885,0.0302399192,0.1710843444,-0.0500850677,0.1164650768,0.0069036391,-0.2920267582,-0.0495645739,-0.1978931725,-0.0086773541,-0.0171592813,0.3072853088,0.5229854584,0.0768130943,0.6932663918,0.191862449,-0.2426931262,-0.1660664529,0.1627267897,0.1830895394,-0.5128666759,0.3100555539,-0.1137086079,0.1459202468,-0.2419808209,0.4570830464,-0.2819413543,0.2090833783,-0.0393038131,-0.1175033078,-0.1571840644,0.2325696945,-0.1332800984,0.1057097316,-0.2226439267,0.1152285784,-0.0790469125,-0.2166758925,-0.2946244478,-0.0973041132,-0.4076433778,0.3669984043,0.1148904189,0.0858326256,0.2190838307,0.1607996225,0.1681335419,0.0856592953,-0.1367595494,-0.196850881,-0.0751663223,0.1268492639,0.1073024794,-0.1864513606,0.1410098374,0.0269025378,-0.0628488362,-0.1389683783,0.0139900362,0.1784923375,-0.0842979401,0.4305182099,0.3046940267,0.3934035599,-0.1181101874,0.160622254,-0.0320248976,0.2289986759,-0.0731261298,0.0898118019,-0.0715093166,0.0787385702,-0.4437326789,0.1356508732,-0.129166916,0.1195798442,0.2263306379,-0.3602280319,0.0355114751,0.3236742318,-0.0311139822,0.0755379125,-0.198323831,-0.0149546945,0.090865083,0.1674839258,-0.161023438,0.0044072028,0.1927999556,0.098377116,0.055364117,0.2497331053,0.0238154791,0.2197780013,0.1523765475,0.0410668887,0.0576352365,-0.1304038465,0.4767722189,0.2502485216,0.106006749,-0.005961901,-0.0451643243,-0.1361158639,0.2374778241,0.4095085859,-0.0225137156,0.1133541986,0.4354737401,-0.0183572639,-0.0299489778,-0.2667696178,-0.183229357,0.0874819532,0.3381926715,0.2028533518,0.1870117784,-0.3521421552,-0.3258129656,-0.2701639831,-0.069582805,0.3099346757,-0.0583300889,-0.0029459991,0.0356674194,0.0114391381,-0.3740221262,-0.2415987253,-0.0964923352,-0.1092930734,0.2701234221,0.2186554223,0.1762788445,-0.3771931529,0.27475968,0.2808687687,0.2680353522,-0.1144816577,0.2841907442,0.3918447495,0.0941931978,0.1932379305,0.5161973834,0.3695507646,0.2945841849,-0.020104045,0.1129500493,-0.0390780009,-0.3450121284,0.0877992213,-0.059845686,-0.2485655695,0.2879635692,0.1558267176,0.1228244901,-0.1254130453,0.3168381751,0.0544719622,-0.1103277802,-0.2857102156,0.7373996973,-0.2703433633,0.0276814178,-0.0651036352,-0.2521202862,-0.4581924081,-0.0652412251,0.381400764,0.1698071212,0.149515003,0.2177317739,0.0697703585,-0.4082437456,0.330963999,0.3403661847,0.0769058913,-0.2651322186,0.0617089309,-0.2750614285,0.1116446406,-0.1015280634,-0.0691227019,0.2495637238,0.1278142929,-0.1932554841,0.1148602813,-0.2204782218,-0.1393817365,-0.104323633,0.25245592,-0.3717963696,-0.1946512759,0.2883968651,-0.0892380476,-0.1251148432,-0.2652546763,0.0353319235,0.0039455947,0.0739230812,0.0550829805,-0.0448910445,-0.3227532506,-0.0278719831,0.057251703,0.6041781306,0.1074118167,-0.0593114123,-0.3922547102,-0.2225103676,-0.272254765,-0.1133324504,0.5871050954,0.2825134397,0.6090587378,-0.1371040046,0.2531865239,-0.0552069545,0.3776589036,-0.0609906539,0.0438415594,-0.4244903922,0.5184668899,-0.3719528913,0.1762697846,0.0068774358,0.1811163276,0.1390513331,0.1774709225,-0.0504962131,-0.2920152545,0.3989830613,-0.3084869087,-0.3456190228,0.2019387633,-0.0813750997,0.1927640736,-0.0473877303,-0.3078513145,-0.1343808621,0.4044968188,-0.1600380987,-0.2725828886,0.0046120416,0.1497466266,-0.0676551685,-0.120562166,0.1607695073,-0.0707525089,-0.2598537803,0.3602844477,-0.3102980554]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2888","title":"v1.11.1 release date","comments":"Hi ! Probably 1.12 on monday :)\r\n","body":"Hello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.\r\n\r\nWhen do you plan to publush v1.11.1 release?","comment_length":7,"text":"v1.11.1 release date\nHello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.\r\n\r\nWhen do you plan to publush v1.11.1 release?\nHi ! Probably 1.12 on monday :)\r\n","embeddings":[-0.3028725386,-0.1802771538,-0.2127547413,-0.0963228866,0.0652879179,-0.1868623346,0.0738429576,0.3128882051,-0.2114172131,0.4218455255,0.2351345271,0.0822070688,-0.0895844847,0.4837630689,-0.1932857186,-0.2654596269,0.0769366696,0.0144002829,-0.1551253647,0.0024792501,-0.1483270228,0.119360745,-0.3253111243,0.0831760392,0.0989715233,-0.064572975,-0.3526091278,-0.1743175536,-0.5363405347,-0.3933886588,0.461257726,0.1939074099,0.2547554672,0.2986641824,-0.000103304,-0.4014018774,0.5556403995,0.2845861912,-0.2422691882,-0.2193116248,-0.3339813054,-0.5906572938,-0.1154627725,0.0800851658,-0.2921097875,-0.1461395621,0.082330212,-0.0849997699,-0.13677302,-0.1171071827,0.2824233174,0.0947450623,0.1915313601,-0.3989170194,0.284394294,-0.0203845855,-0.2789727747,-0.3155975044,0.5697530508,0.1857773513,0.1898058355,0.326002121,0.2075845897,-0.1623990536,0.1387228072,0.1931113005,0.1122244373,-0.32729882,-0.0393120684,0.1010244489,0.9985591769,-0.0018671668,-0.2434326708,0.2223907411,-0.0925616995,-0.3847976029,0.1220049784,-0.0718759969,0.1220379025,0.1305280179,0.0270749014,-0.3837861121,-0.1746051461,0.2439880967,0.0338811204,0.3361899853,-0.1041846797,-0.0660860538,0.1154744327,-0.2463802546,0.0967559218,-0.0215033758,-0.1613555253,0.2227307111,0.0464089848,-0.4094306231,0.1858401746,-0.0288592987,0.1305842102,0.3231448531,-0.1524199545,-0.1115325615,-0.1302562058,-0.0906753838,0.5744841099,0.0935288742,0.377846837,0.2240085304,0.3023318052,0.1646759659,0.210433796,-0.0148354508,0.131557256,0.2080590278,-0.1345603913,0.092921719,0.2413697988,-0.5841081142,0.3566501141,0.0036136017,0.0791958421,-0.3253816366,-0.3280638158,-0.1975245178,-0.0646178052,0.1883967668,-0.1564619541,-0.1827139109,0.1210180596,-0.2046548128,-0.3577520847,0.0363553949,-0.2325457633,-0.0549016744,0.1167488173,-0.2526122332,-0.1254964024,0.0305748545,-0.04656725,0.0622468628,0.0610297844,0.1928608567,-0.2869565487,0.2402708977,-0.1486318558,0.2325121164,-0.3784801364,0.3326986134,-0.1420154572,0.2017378062,-0.0192655772,-0.2063460052,-0.235487327,0.3137744367,-0.1372713745,-0.3029890358,0.19962731,0.3279361129,-0.3900982738,-0.1182468757,-0.1860611141,0.2779344022,-0.1112236157,-0.2089126557,0.0612550527,-0.0017352636,-0.2293154597,0.0255627967,-0.3211435378,-0.2947054803,0.1805807501,-0.0630620569,0.0061964095,-0.2280418724,0.1311280131,-0.185826987,0.4691230059,-0.3665311038,-0.5793189406,0.4294295013,-0.1866938472,-0.4467723966,0.4338716865,0.2241641581,0.3097352386,-0.1575234383,-0.2617045343,0.2637113929,-0.1526574194,-0.320348084,-0.2554035187,-0.3677665293,-0.0720299929,0.1507190168,0.1602411419,0.2366000116,0.2579503655,0.2858259678,0.1898807436,0.1925684959,0.1466663629,0.0294717029,0.5004534125,-0.10350997,0.2177486867,-0.249899298,-0.2581174374,-0.0229802001,-0.0547287352,0.2486825585,0.5317671299,-0.2098047286,-0.168516919,-0.1740678996,0.2359928489,-0.0324108824,0.1631854326,0.1089351401,-0.0164421871,0.1266786158,-0.3025694489,0.0803501308,0.1472560316,0.0244728141,-0.075657174,0.1248692945,-0.2761505842,0.0051354663,0.2405391037,-0.0050578136,0.0540098287,-0.2141433656,0.1412835717,0.0534515046,-0.0641530901,0.3371422291,0.6113381386,0.1324200183,0.3750956953,0.1302689165,0.219974041,0.0903351158,-0.0883826166,0.2412555814,-0.269962877,0.312500447,0.238511622,-0.2063528448,-0.1451906115,0.118214719,0.1887120157,-0.0891980454,-0.0327405259,-0.3398402631,0.2254811525,-0.1658332348,-0.2750526965,-0.0365721621,-0.2551523447,0.3512178659,0.3392684758,-0.1608422101,-0.1638178527,0.0186253563,0.0652871504,-0.0031810594,0.2478724569,0.0328481123,-0.1552314162,0.3118087649,0.1120094359,-0.0968455672,0.0139505379,0.0497668646,0.0382369012,0.0392541103,-0.3878073692,-0.221236378,0.0975770727,0.1155316979,0.0994631052,0.1521302015,0.1214670911,0.2327489555,0.2580001354,-0.344601959,-0.3168298304,-0.2266455889,0.1146346107,-0.1705584228,0.1445552707,-0.1704138368,0.0943085477,0.0777231157,0.2056838423,0.1746924818,-0.0169805512,0.7022944689,-0.2754337788,0.4286308587,-0.22941567,-0.0400346629,-0.287625581,0.3418615162,-0.1031061858,-0.2282226235,0.195717901,0.0163822714,0.3134722412,-0.4504329264,-0.4006796479,0.0965960771,-0.190621376,-0.0871734247,0.1440135986,-0.3507573307,-0.0002354266,0.2681168616,0.020206226,-0.1884005666,-0.1797890365,-0.1004347429,-0.1091894209,0.0140760578,-0.2644398212,-0.5476667881,-0.23607862,-0.032877408,0.0195357129,0.1396642774,0.0098538799,-0.2097719312,-0.0314573459,-0.0043667466,-0.025507886,-0.3325499892,-0.13492091,-0.1378961504,0.1646966338,-0.1210952327,-0.3690144718,0.0233245771,0.2133821547,-0.1906823069,0.0343830734,-0.608182013,-0.2255241424,-0.07357122,0.0902312696,0.271065414,0.0482119732,0.3331973553,0.0043859896,-0.2407487333,-0.0966798887,-0.289496839,-0.0617358088,0.2790717185,0.3425323367,0.1480440795,0.2153837532,0.0382699557,0.4080422521,0.0405399092,-0.0831079558,0.1420852691,0.0724038258,0.1340251565,-0.0351542011,-0.1449554712,0.1592445076,0.0636743233,0.2118257731,0.0847029686,0.2609232962,-0.1647253633,-0.4551307261,-0.0843807682,0.0422558002,0.1397813708,-0.2756437659,0.2367210686,0.3786954284,0.094233036,0.0404279269,-0.1744975001,-0.2230151296,0.1618852317,0.0282293838,0.155143857,-0.037190333,-0.0875179842,0.2474852353,-0.4120295048,0.3248883486,-0.1162716597,-0.0254962426,-0.1575712413,0.1632208079,0.0907445326,-0.0197371934,-0.1022816896,-0.4347622991,0.1979690641,0.3334328234,-0.3282330632,0.209832564,0.0612220913,-0.5530356169,-0.0694036111,-0.1418507844,0.490211755,0.0862114429,-0.2527614832,0.5576377511,0.0898270458,-0.2430394441,0.089087598,-0.0483200476,-0.0593439825,0.0106714554,-0.0859256014,-0.3630439341,0.028471671,-0.0130609032,-0.2682625651,0.0257274024,-0.0579449162,0.0380475894,-0.0671024472,0.098899141,-0.0086529767,0.1888565868,0.2343615294,-0.0548734739,0.2192328572,0.2210982591,0.1266529709,-0.219884634,0.0629318357,-0.3096889853,0.2338983119,0.2519245148,-0.0197273009,0.0526784435,-0.174160555,0.0401716121,-0.1222358719,-0.1953085959,0.4756614864,-0.0125389127,-0.1958249062,-0.2044544071,0.2814204693,0.2040803432,-0.323269546,0.6049668789,0.2233847231,0.0581449308,0.0844930708,-0.0875702798,0.8768109083,0.0091802431,-0.0530386232,-0.103736043,-0.258508563,0.3260923326,0.1500751227,0.1819668561,-0.2611414194,-0.1667617857,-0.1415490061,0.1881760061,0.2261874974,0.0333929323,-0.4414962232,0.281454742,0.0162123907,0.318469286,0.0218495298,0.4039916396,0.2129945159,0.1521504074,0.0103357509,0.2614460289,0.007627523,0.2971936762,-0.027748853,-0.0637189299,0.0742425397,0.0526656769,-0.2426996827,-0.083447732,0.2240836024,0.1241263077,0.0071352581,-0.4132660627,-0.1007736176,-0.1466791779,0.2693107724,0.1400369257,-0.2019508928,0.0580885448,-0.0842173621,-0.1220868751,0.0292737298,0.2089927346,0.2616169453,-0.3624403477,-0.2537929416,0.1820687056,0.0940711349,-0.1514322162,-0.3724226952,-0.1742647737,0.0492790043,-0.117994085,0.2095771134,0.3122289181,-0.1254249513,-0.2380960584,0.2550587356,0.233158201,0.0837606713,0.0246355869,0.2637301683,0.0132632833,-0.2026557177,0.2164986283,0.2368977517,-0.2317263782,-0.1114870906,0.032164149,-0.0729064792,-0.3280212283,0.149072662,0.2185632586,-0.032606367,0.0898795649,-0.0617410839,-0.0693863109,0.1985311955,0.7674746513,0.1134763584,-0.1324970871,-0.3815554082,-0.1036905646,-0.1598554105,0.4016879797,-0.3281000853,0.0328431837,-0.3015990853,-0.1019492298,0.1833198965,-0.1321831942,-0.4311744273,0.0374722108,0.2304802239,-0.1759789288,0.3346586227,0.0276554301,-0.0175769292,-0.0240464881,0.1310301274,-0.3019618094,-0.3366426528,-0.2488733232,0.0150217954,0.0964988172,-0.100690268,0.0258981884,-0.2341493517,-0.1691185385,-0.2735385597,0.3086502552,0.06740731,0.0497181714,0.2783055902,0.1942646354,-0.0693481043,-0.1438840032,0.0461286157,0.1568335444,0.3562457561,-0.0707835779,0.1494660378,0.2881281674,0.3032661378,-0.0937201083,-0.1116363481,-0.0131287752,-0.0468256325,0.2652699351,0.1904975176,0.0371222459,0.1332924515,0.2392027676,0.299526006,0.122852698,0.0362787955,-0.1393833905,0.0654303581,0.3110370636,-0.2561894059,-0.3114360273,0.0423956625,0.2429854125,-0.2813114524,-0.0419045128,0.2741584182,0.3505083323,0.22629264,0.1385433227,0.2401577383,-0.3199465871,0.1219916716,0.2591587603,0.2279812098,-0.0197042357,0.2900158763,0.1830374002,-0.0682588518,0.093497701,0.0996332839,0.2804675102,0.5949040055,0.3702272177,0.2393652797,-0.1198549196,0.0408343188,0.0919148624,-0.0366935171,-0.0142002311,0.0637082756,0.2750998139,-0.29993397,0.0468942784,-0.3736778796,-0.2152929455,0.0454041921,-0.1816400886,-0.0064777839,-0.1451922357,-0.1916889995,-0.053099826,0.0388124809,0.3415044248,0.4628058672,-0.2635577023,0.0336105339,-0.1391046792,-0.0992713794,0.0667115003,0.2791768909,-0.0249665789,0.2348811775,-0.396173954,0.1176120788,0.1963460892,0.3659482896,0.0244652964,-0.109329991,-0.2823936045,-0.0752786174,-0.1506735981,-0.2029793561,0.1160526648,0.3965930045,0.1634795666,0.0897685811,0.1343604624,0.145347029,-0.1465705633,0.4180464149,0.0394908339,-0.1240979508,-0.451328218,0.2151078582,0.1355981827,-0.1942361742,-0.0209452938,0.0969316363,-0.1266513616,-0.3759785891,0.180424124,0.1019164398,-0.1711923629,-0.1600877494,0.1484104097,-0.0683724433,0.124356918,0.1947842687,0.1296981573,0.0286450516,-0.3705144227,-0.3561777174,0.3234818578,0.2191070914,-0.2648040056,-0.1161450446,0.246768415,0.1100329086,0.3032544553,0.1898021251,-0.1368157864,-0.244817704,-0.1841135025,-0.4536907673,0.1890252829,0.2329204381,0.2602323592,-0.12943694,0.0105283875,0.1758193523,-0.3087698519,0.061227791,0.0883698463,-0.2749614716,0.2900063097,0.1081978455,0.4732691348,-0.0953155756,0.1714043915,-0.0163458809,-0.175800845,-0.3269743919,-0.1704187542,-0.0873152837,-0.1886828095,-0.0751526654,0.0581922047,-0.237008065,0.1586024314,-0.1818601787,0.1955085397,-0.0558605976,0.3494592011,0.0367623195,0.0512245521,0.0115491403,-0.4316511452,0.1641193479,0.0917870998,-0.1045387313,-0.1162620783,-0.3184365332,-0.2898872793,0.266521275,0.1041490585,-0.119506143,-0.3795951605,-0.1625541449,0.3523990214,-0.1871525049,-0.4342991412,0.0063330201,0.1459901631,0.1065577641,-0.1277641356,0.1525983065,0.193662852,-0.0439628996,-0.0201323126,0.1976662725,0.0432184488,-0.044095356,-0.2465994656,0.1530583352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2888","title":"v1.11.1 release date","comments":"@albertvillanova i think this issue is still valid and should not be closed till `>1.11.0` is published :)","body":"Hello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.\r\n\r\nWhen do you plan to publush v1.11.1 release?","comment_length":18,"text":"v1.11.1 release date\nHello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.\r\n\r\nWhen do you plan to publush v1.11.1 release?\n@albertvillanova i think this issue is still valid and should not be closed till `>1.11.0` is published :)","embeddings":[-0.2534325421,-0.1435918212,-0.1170087606,0.1250979602,-0.0803622156,-0.2069433331,0.3182261586,0.2595030665,-0.2662116587,0.2432924211,0.4498234391,-0.0775993839,-0.0092685595,0.3775973022,-0.3972175717,-0.0890159234,0.0787209198,0.1089016348,-0.0526827313,0.0449188799,-0.2720680833,0.068657726,-0.3964475691,0.1608792841,0.0663995072,-0.0716037601,-0.1215934381,-0.254347086,-0.6671610475,-0.5313281417,0.3133550286,0.2034503967,0.0760271847,0.271275878,-0.0001110468,-0.423866421,0.5205672979,0.1486699134,-0.2614662647,-0.1513033807,-0.416443944,-0.5353802443,-0.1008592322,-0.0137669845,-0.3072520196,-0.1077144369,0.0117483158,-0.1943197548,-0.0628163069,-0.0156020885,0.2480115443,0.1673386395,0.4019794762,-0.3542853892,0.1863051802,-0.1159243658,-0.3703640699,-0.2499655932,0.5717850924,0.2568038106,0.2268878967,0.384983182,0.1912631989,-0.1922257096,0.1507626772,-0.0097723147,0.2396764457,-0.266032964,0.0415400378,0.0777787417,0.9955607653,0.0203635413,-0.3793464899,0.2034740597,-0.0927332118,-0.2712348402,0.2412474453,0.027283011,-0.0292447396,0.149242878,0.0873354375,-0.4584193528,-0.1516828388,0.1602145284,-0.0272049885,0.459104538,-0.0887170136,0.1067651212,-0.0215255283,-0.3015371263,0.2346984148,0.0200975835,0.0850308686,0.156445384,0.0514521152,-0.4402125776,0.1060789451,-0.0172465947,0.1498136967,0.1924144477,-0.2606184483,-0.249166131,-0.0464439057,0.0133378366,0.6594331861,0.0745627508,0.3405662179,0.188009277,0.4329450428,0.2551412284,0.0644607022,0.0997430384,0.0618424639,-0.0808662996,-0.1455136389,0.2052437365,0.4584421515,-0.5718432069,0.3661362231,0.1454990655,0.0115315011,-0.073997356,-0.3209313154,-0.1508467942,-0.0373871252,0.3257052004,0.0893606171,-0.1983124912,0.1348611414,-0.1805576384,-0.1951919943,-0.1088341549,-0.200981006,-0.0183628146,-0.031561859,-0.3116750121,-0.2775349915,0.1134495363,-0.1254057586,-0.0434127748,0.0564601496,0.1302085966,-0.2246665657,0.4118489623,-0.2466320097,0.082844384,-0.3096274436,0.2402789593,-0.0819456652,0.3500746489,-0.0424564332,-0.1780534983,-0.3391320705,0.2919041216,-0.1666521877,-0.0776967406,0.2701286674,0.1953231692,-0.2030190676,-0.2382987589,-0.0494303294,0.1722414643,-0.2394903451,-0.256931901,-0.0695379525,0.0432331599,-0.1193417981,0.1540283114,-0.2632963657,-0.2000599653,0.1197276041,0.0051401095,-0.1842305511,-0.3186813593,0.1262588203,-0.2455536276,0.3016958833,-0.3433002532,-0.639559865,0.3815340102,-0.284142822,-0.3894364536,0.4114672244,0.1741337925,0.2954373658,-0.3212958872,-0.1409738809,-0.0816188827,-0.1616334766,-0.3559683263,-0.4418994188,-0.4019602537,0.0429010168,0.0181122757,0.2204887122,0.1503366083,0.0952749476,0.4235877395,0.0749824941,0.2797828615,0.2078803927,0.0765660107,0.482803762,-0.1286068708,0.0529898442,-0.1006552279,-0.2767008245,0.0938937142,-0.1435723305,0.1591186672,0.4384330511,-0.3189628124,-0.2085864544,-0.0561050922,0.3260715008,-0.0709523261,0.0817816257,0.0915738642,-0.2478206754,0.1673493236,-0.4060981572,0.1368636489,-0.0098665319,-0.07066378,-0.088441737,0.0452465005,-0.2908729911,0.0994255021,0.1788440347,0.0838302895,0.1694515795,-0.2390569299,0.1745637655,-0.0815123394,-0.0723995641,0.3089856207,0.3561557233,0.1862898916,0.2350969911,0.2040317804,0.0746104494,0.1710554659,-0.0586112514,0.2303554267,-0.2253153324,0.3460877836,0.3549923301,-0.2632007301,-0.0564746819,0.144726038,0.1307961047,-0.2599101663,-0.0981654152,-0.2556478679,0.0817662179,-0.221117571,-0.2914198935,0.2252827138,-0.2763938606,0.2965317667,0.5968988538,0.0184195228,-0.047829736,0.0809677541,0.2573701739,-0.0809906572,0.1679659784,-0.0462063886,0.0884929076,0.2539899945,-0.021018533,0.0049677342,0.1221921593,0.0425285064,0.1027600989,0.0455321558,-0.4042115211,-0.0299377795,0.0530678742,0.0633863062,0.115278326,0.1767398119,0.1918154806,0.2266006619,0.0650810748,-0.3978857994,-0.2835028172,-0.0675668642,0.0291238204,-0.0194902401,0.0619625486,0.0534454733,0.1382157803,0.1468727291,0.0316265896,0.1826277375,-0.1548491269,0.7969358563,-0.3889087141,0.4114349484,-0.156639576,-0.1199053377,-0.3110970855,0.2345053405,-0.0124809062,-0.2776047289,0.2567169964,0.0537328571,0.2090214789,-0.6258595586,-0.4577414989,0.0528508835,-0.144126296,0.0452978648,0.0343037061,-0.2054453045,-0.107023187,0.2207032591,0.0384894237,-0.0587673448,-0.2087223232,-0.1666585952,-0.0068196445,0.0721502379,-0.2570260763,-0.6848427057,-0.1318627,-0.3363646269,-0.0185509361,-0.1078772321,0.1569403857,-0.1232450008,-0.051425416,-0.1049095914,0.0217202213,-0.3342834115,-0.0948758051,-0.048731003,0.0945054814,-0.0082828654,-0.4569739103,0.1993883997,0.1774878055,-0.2025116682,0.1531127393,-0.5314667225,-0.2725237608,0.206520468,-0.043862544,0.1573558748,-0.1776329577,0.5536837578,-0.1086382791,-0.2538419664,-0.1345501691,-0.2866043448,0.0720614642,0.1712462157,0.3246344626,0.409180969,0.2624562085,0.0242744964,0.3892090321,0.1903090924,0.0267370865,0.2341599315,0.1795129925,0.1924058646,-0.0937660858,-0.2098539323,0.0676550344,0.1260128468,0.1618110985,0.1551786363,0.212505579,-0.0854245946,-0.3764601052,-0.0002547773,-0.0050693392,0.1066496,-0.2045838088,0.19392021,0.6248509884,0.1559808254,-0.0690406263,-0.1899852306,-0.050280299,-0.1242645904,0.2394972146,0.218652606,-0.0803116336,0.0049764337,0.3057085574,-0.2312407643,0.3685397804,-0.0162633043,0.3051949441,-0.1745799035,0.183699578,0.167574808,-0.0165351611,-0.002247988,-0.4258732796,0.2501198947,0.5538979769,-0.2576999366,0.1758041531,0.0319327712,-0.1597495675,-0.1175884306,-0.1852885038,0.5539381504,-0.0616523623,-0.3503075838,0.5332707167,0.0958007053,-0.1159934327,0.0116964756,-0.148977384,-0.0391741842,0.0361851342,-0.0692064166,-0.1455716044,0.0450960174,0.0707545504,-0.2735245228,-0.2130563408,-0.2302908301,-0.0789414719,-0.1162233725,0.2302821875,-0.0054950924,0.2637892067,0.1018653214,-0.006938694,0.2944147289,0.310501188,0.110301502,-0.2645044625,0.1126402467,-0.0766577423,0.1321163625,0.3621128798,-0.1440612823,0.1353558898,-0.0324910507,0.0674986169,-0.0920338705,-0.2673429549,0.4327611923,-0.1655806452,-0.2975933552,-0.0365021341,0.2060316652,0.1686381698,-0.3554573059,0.5559965968,0.2967342436,0.0080543449,-0.0236563794,-0.2770451903,1.0293949842,-0.0287329629,-0.0709773898,-0.2229948789,-0.2106604427,0.3116071224,0.1518327743,0.1275588423,-0.1479117125,-0.1200410053,-0.156358555,0.025576435,0.3566545844,0.0137892039,-0.4522585273,0.1800192744,-0.1066877097,0.3458714783,0.1815039814,0.4181370139,0.2039348483,0.2149993479,-0.1357867271,0.1812182963,0.0415112488,0.3444061577,-0.1115806252,-0.1123415828,0.0506095029,-0.1684783846,-0.3772748411,-0.162755996,0.0946233049,0.1284955442,-0.0418820232,-0.3295052052,0.1501638293,-0.2067037821,0.2394614667,0.0587115698,-0.1543533653,0.1066337228,-0.1002984643,-0.2956922352,0.0422360674,0.2025135756,0.2493470311,-0.365799576,-0.1296645254,0.207518965,-0.0665956438,-0.1957606375,-0.3805607557,0.0144149913,0.1905001253,-0.0858497769,0.0902350396,0.0793108046,0.0000393472,-0.2687571943,0.2166418582,0.2363907099,0.173072502,-0.070472464,0.0909126326,-0.0943970904,-0.0462283231,0.331574887,0.4488438964,-0.0869515613,-0.0577264875,-0.104647994,-0.0253153257,-0.2967375219,0.1651073247,0.1483710557,-0.1926550269,0.3166780174,0.0451815948,-0.177145049,0.0966312066,0.8037606478,0.1132905334,-0.0867026523,-0.3089384437,-0.2269143313,-0.201273486,0.4007463753,-0.3507455587,-0.0632593036,-0.1728218198,0.003514756,0.1783655137,-0.2588581443,-0.3582352698,0.1094941497,-0.0229328293,-0.1558413357,0.3805615604,-0.1397653073,-0.1268882602,-0.1695237756,0.0683612898,-0.1914259791,-0.4219688773,-0.1825899184,-0.0068806806,0.1795591116,-0.099498719,0.1105903983,-0.2188424021,-0.3331711888,-0.3294802904,0.5655476451,-0.0574626066,0.2105582207,0.265817821,-0.0013697128,-0.1911033839,0.083206743,-0.0473593064,0.1444205344,0.2916863263,0.0616737306,0.0525817275,0.2192509323,0.3352403939,-0.051245898,-0.3188316226,-0.0857311264,0.028379539,0.1082434431,0.2533852756,0.0845074058,0.0778547525,0.3177062273,0.261341095,0.1477471739,0.0235879123,-0.10451556,0.1499727219,0.2290463895,-0.4348066449,-0.2144445181,-0.1024437621,0.1954161376,-0.3519690037,-0.0586642176,0.4161963165,0.3745106161,0.3306231499,0.2353342474,0.3252157569,-0.2875549793,0.1753366739,0.298651427,0.2821616828,0.0819243565,0.2702084482,0.3563085496,0.1198417321,-0.1174164787,0.1162222177,0.2579935193,0.6770775318,0.4611133635,0.2945824265,-0.1674147099,0.1528614014,0.145076111,0.0248568561,-0.0097037228,-0.028063925,0.2341251522,-0.2223802507,0.000882468,-0.252620697,-0.1989824623,-0.1395764947,-0.2066176832,0.0804793462,-0.2013021708,-0.1639318615,-0.0490338057,-0.0818435028,0.3881368041,0.4651462138,-0.172336027,0.0645914972,-0.1019973978,-0.126423344,0.1757827997,0.2655941546,-0.1845968068,0.3018839657,-0.4171665907,0.1055940539,0.2046003491,0.5415052176,0.0668870881,-0.0055139861,-0.0966519117,-0.0346165001,-0.2049902081,-0.0964519158,0.119966276,0.4839955568,0.2810302973,0.3264023364,0.0233848933,0.0755379423,-0.1421592236,0.4168718159,-0.1079073548,-0.1283363253,-0.6276838779,0.2174899578,-0.1051726788,-0.1579537541,-0.0612815022,0.1260119081,-0.0126639176,-0.3415772617,0.2715965509,0.2166879475,-0.0643055439,-0.248646915,0.1112187207,-0.2308488041,0.0420173332,0.0135724219,-0.0086574517,-0.0197401382,-0.3096870184,-0.4489578605,0.3907983005,0.1115981862,-0.2408151627,0.1052808017,0.2133009285,0.0889786929,0.3081979156,0.0924179628,-0.1790391803,-0.3171945512,-0.1446876526,-0.402273953,0.1941866428,0.3912087083,0.0847943872,-0.0422423035,-0.2774840295,0.2742191255,-0.2932401896,0.0322426744,0.1044757292,-0.3337162137,0.386005193,0.1406664401,0.5287458301,0.032131739,0.1304536164,0.2330349982,-0.1344192475,-0.4677241743,-0.1328878999,-0.0336711966,-0.1085676104,0.0418233424,0.0438944586,-0.2677304447,0.1162145585,-0.1990454048,0.0510428324,-0.3705941141,0.4114429653,0.0662547424,0.109508045,0.0662498847,-0.2569505572,0.2648650408,0.0552677512,-0.1163000241,-0.2598813772,-0.4447279274,-0.319673717,0.2948858142,0.0099982917,-0.1469426304,-0.1186853945,0.075142175,0.2601584792,-0.263087213,-0.4047777951,-0.1358789206,0.3026438653,0.2290502042,-0.1205541492,0.0037204858,0.0116371764,-0.1853573173,0.0748225152,0.3017756641,-0.147349447,-0.1105487272,-0.1910789162,0.0743446872]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2885","title":"Adding an Elastic Search index to a Dataset","comments":"Hi, is this bug deterministic in your poetry env ? I mean, does it always stop at 90% or is it random ?\r\n\r\nAlso, can you try using another version of Elasticsearch ? Maybe there's an issue with the one of you poetry env","body":"## Describe the bug\r\nWhen trying to index documents from the squad dataset, the connection to ElasticSearch seems to break:\r\n\r\nReusing dataset squad (\/Users\/andreasmotz\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/d6ec3ceb99ca480ce37cdd35555d6cb2511d223b9150cce08a837ef62ffea453)\r\n 90%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 9501\/10570 [00:01<00:00, 6335.61docs\/s]\r\n\r\nNo error is thrown, but the indexing breaks ~90%.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nfrom datasets import load_dataset\r\nfrom elasticsearch import Elasticsearch\r\nes = Elasticsearch()\r\nsquad = load_dataset('squad', split='validation')\r\nindex_name = \"corpus\"\r\nes_config = {\r\n \"settings\": {\r\n \"number_of_shards\": 1,\r\n \"analysis\": {\"analyzer\": {\"stop_standard\": {\"type\": \"standard\", \" stopwords\": \"_english_\"}}},\r\n },\r\n \"mappings\": {\r\n \"properties\": {\r\n \"idx\" : {\"type\" : \"keyword\"},\r\n \"title\" : {\"type\" : \"keyword\"},\r\n \"text\": {\r\n \"type\": \"text\",\r\n \"analyzer\": \"standard\",\r\n \"similarity\": \"BM25\"\r\n },\r\n }\r\n },\r\n}\r\nclass IndexBuilder:\r\n \"\"\"\r\n Elastic search indexing of a corpus\r\n \"\"\"\r\n def __init__(\r\n self,\r\n *args,\r\n #corpus : None,\r\n dataset : squad,\r\n index_name = str,\r\n query = str,\r\n config = dict,\r\n **kwargs,\r\n ):\r\n #instantiate HuggingFace dataset\r\n self.dataset = dataset\r\n #instantiate ElasticSearch config\r\n self.config = config\r\n self.es = Elasticsearch()\r\n self.index_name = index_name\r\n self.query = query\r\n def elastic_index(self):\r\n print(self.es.info)\r\n self.es.indices.delete(index=self.index_name, ignore=[400, 404])\r\n search_index = self.dataset.add_elasticsearch_index(column='context', host='localhost', port='9200', es_index_name=self.index_name, es_index_config=self.config)\r\n return search_index\r\n def exact_match_method(self, index):\r\n scores, retrieved_examples = index.get_nearest_examples('context', query=self.query, k=1)\r\n return scores, retrieved_examples\r\nif __name__ == \"__main__\":\r\n print(type(squad))\r\n Index = IndexBuilder(dataset=squad, index_name='corpus_index', query='Where was Chopin born?', config=es_config)\r\n search_index = Index.elastic_index()\r\n scores, examples = Index.exact_match_method(search_index)\r\n print(scores, examples)\r\n for name in squad.column_names:\r\n print(type(squad[name]))\r\n```\r\n\r\n## Environment info\r\nWe run the code in Poetry. This might be the issue, since the script runs successfully in our local environment.\r\n\r\nPoetry:\r\n- Python version: 3.8\r\n- PyArrow: 4.0.1\r\n- Elasticsearch: 7.13.4\r\n- datasets: 1.10.2\r\n\r\nLocal:\r\n- Python version: 3.8\r\n- PyArrow: 3.0.0\r\n- Elasticsearch: 7.7.1\r\n- datasets: 1.7.0\r\n","comment_length":44,"text":"Adding an Elastic Search index to a Dataset\n## Describe the bug\r\nWhen trying to index documents from the squad dataset, the connection to ElasticSearch seems to break:\r\n\r\nReusing dataset squad (\/Users\/andreasmotz\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/d6ec3ceb99ca480ce37cdd35555d6cb2511d223b9150cce08a837ef62ffea453)\r\n 90%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 9501\/10570 [00:01<00:00, 6335.61docs\/s]\r\n\r\nNo error is thrown, but the indexing breaks ~90%.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nfrom datasets import load_dataset\r\nfrom elasticsearch import Elasticsearch\r\nes = Elasticsearch()\r\nsquad = load_dataset('squad', split='validation')\r\nindex_name = \"corpus\"\r\nes_config = {\r\n \"settings\": {\r\n \"number_of_shards\": 1,\r\n \"analysis\": {\"analyzer\": {\"stop_standard\": {\"type\": \"standard\", \" stopwords\": \"_english_\"}}},\r\n },\r\n \"mappings\": {\r\n \"properties\": {\r\n \"idx\" : {\"type\" : \"keyword\"},\r\n \"title\" : {\"type\" : \"keyword\"},\r\n \"text\": {\r\n \"type\": \"text\",\r\n \"analyzer\": \"standard\",\r\n \"similarity\": \"BM25\"\r\n },\r\n }\r\n },\r\n}\r\nclass IndexBuilder:\r\n \"\"\"\r\n Elastic search indexing of a corpus\r\n \"\"\"\r\n def __init__(\r\n self,\r\n *args,\r\n #corpus : None,\r\n dataset : squad,\r\n index_name = str,\r\n query = str,\r\n config = dict,\r\n **kwargs,\r\n ):\r\n #instantiate HuggingFace dataset\r\n self.dataset = dataset\r\n #instantiate ElasticSearch config\r\n self.config = config\r\n self.es = Elasticsearch()\r\n self.index_name = index_name\r\n self.query = query\r\n def elastic_index(self):\r\n print(self.es.info)\r\n self.es.indices.delete(index=self.index_name, ignore=[400, 404])\r\n search_index = self.dataset.add_elasticsearch_index(column='context', host='localhost', port='9200', es_index_name=self.index_name, es_index_config=self.config)\r\n return search_index\r\n def exact_match_method(self, index):\r\n scores, retrieved_examples = index.get_nearest_examples('context', query=self.query, k=1)\r\n return scores, retrieved_examples\r\nif __name__ == \"__main__\":\r\n print(type(squad))\r\n Index = IndexBuilder(dataset=squad, index_name='corpus_index', query='Where was Chopin born?', config=es_config)\r\n search_index = Index.elastic_index()\r\n scores, examples = Index.exact_match_method(search_index)\r\n print(scores, examples)\r\n for name in squad.column_names:\r\n print(type(squad[name]))\r\n```\r\n\r\n## Environment info\r\nWe run the code in Poetry. This might be the issue, since the script runs successfully in our local environment.\r\n\r\nPoetry:\r\n- Python version: 3.8\r\n- PyArrow: 4.0.1\r\n- Elasticsearch: 7.13.4\r\n- datasets: 1.10.2\r\n\r\nLocal:\r\n- Python version: 3.8\r\n- PyArrow: 3.0.0\r\n- Elasticsearch: 7.7.1\r\n- datasets: 1.7.0\r\n\nHi, is this bug deterministic in your poetry env ? I mean, does it always stop at 90% or is it random ?\r\n\r\nAlso, can you try using another version of Elasticsearch ? Maybe there's an issue with the one of you poetry env","embeddings":[-0.1423357278,0.172727868,-0.0044709793,0.0400959924,0.0997370332,0.0273672696,-0.013204325,0.0684226677,0.0138257872,0.0531148016,-0.0395590216,0.3000072539,0.3499440253,-0.3781624734,-0.1049256846,0.1306153685,0.0909727961,0.0941495076,0.2756751478,0.1979281008,-0.0359147601,0.095908314,-0.1247543246,-0.0892915428,-0.3346315026,0.0212135501,-0.0316708237,-0.1235372126,-0.0287176576,-0.4167428911,0.3019476235,-0.1501852423,-0.1225858629,0.3576337695,-0.0001185079,0.0780737773,0.4348543286,-0.0329714008,-0.0183227472,-0.2907201052,-0.0723804981,-0.1354442686,0.279198885,-0.1591316015,-0.0953472778,0.1617633402,0.0987489,-0.6114605665,0.4415228069,0.3961337507,0.221011281,0.3180061877,0.3437509239,-0.1243836433,0.4198286533,-0.1281127334,-0.0489515252,-0.2564239502,0.2530343533,-0.0389002413,-0.0290674474,0.1452289224,-0.0944889709,-0.2707443833,0.108504355,0.1481345296,-0.4625434577,-0.2957313359,0.3149895072,0.1134225652,0.1910260916,-0.0712343082,-0.2519367039,0.0451840535,-0.0008995398,0.1622108668,0.1994201243,0.1894568503,0.0606099963,0.1190423146,0.0653904825,-0.1248976737,-0.2391608059,0.4135331213,-0.0537097417,0.1734440774,-0.097541891,0.1064755395,0.1365941018,-0.146080032,0.0782291442,0.2522954047,-0.2729330957,-0.0555161387,-0.3764431775,0.1024475098,-0.1168943718,-0.2002719194,0.4185051024,-0.0634525344,-0.0926844999,0.1597323418,0.6514115334,0.082370162,0.2860307097,-0.0631037876,0.0613816716,0.3085791171,0.0735873207,-0.3172648847,-0.3174724579,-0.2593558729,0.285479039,-0.2149694413,0.4180778563,-0.4818145633,0.2134571224,-0.0300275907,-0.159065932,0.2193003744,-0.5879529715,-0.0061579379,-0.0753503293,0.1353493631,0.1364613473,-0.4953055084,0.0383693911,0.2725166976,-0.2289299369,0.1452205926,-0.2356664091,-0.2659835815,0.0408693254,0.0179617349,0.1160360128,-0.6583971977,0.2295712978,0.0021879284,0.0842600539,-0.3945707381,0.2630119622,-0.0829003006,0.1773642153,0.218855992,0.0615348555,0.023488922,0.0232272167,-0.4385558963,-0.0351555534,-0.1221938878,-0.3814532757,-0.2521406412,-0.2764842808,0.2102274299,-0.1385770142,0.0171780493,0.1357520074,0.2181541324,0.1103124693,-0.0082744509,0.0775755048,-0.0403560661,-0.2897684872,0.0814163461,0.2997894585,0.2074805498,0.1160847694,-0.1290065497,0.034548115,0.4276873767,-0.2609300017,0.2141182125,-0.2313275784,0.0364292637,-0.1731410325,0.4179225564,0.3350299299,-0.4630022347,-0.613461256,0.1387205869,-0.2306159735,-0.0950086787,-0.272356391,0.0607838072,0.6483929753,0.0107715735,0.1798678637,0.2476815581,-0.1762580425,-0.1678980738,-0.1523010135,-0.0007278286,-0.1430873871,-0.177202031,0.0628991276,0.0903355628,0.2576126158,0.2004109323,-0.0055740727,0.1895459443,0.0115436343,0.3066953421,0.3564867079,0.1077488884,0.2526801229,-0.1288016289,-0.2565320432,0.2232705653,0.0895305648,0.0836865902,-0.1827768236,-0.1198632121,-0.1379051954,0.0071618739,-0.1569918245,0.0309961122,0.107497111,-0.2377432734,-0.2198884338,0.1922539026,0.0002298418,-0.1667473465,0.0728107169,0.2382162064,-0.2376390845,0.3120537996,-0.0830118656,-0.0867530704,-0.1883609593,-0.0127894552,0.2368167192,-0.2060383856,-0.1446202993,0.332541883,-0.0994061977,0.3171947896,0.1542841047,0.025194684,0.1261184365,-0.6889179945,-0.1038779616,0.1420570463,0.0571958348,-0.2761912048,-0.0003128937,0.0152491853,0.0072259768,0.3724019229,-0.2171028852,0.0868740231,-0.0073972992,-0.1734815091,-0.4597905576,-0.0337329842,-0.0198460873,-0.3641361892,-0.0424907617,-0.1129411384,-0.118290931,0.0319062024,0.0161155853,0.1542524546,0.104662396,0.3709858358,0.0202301163,-0.10019207,-0.3800418377,-0.0208591763,0.2398631126,0.1784040332,0.1413077563,-0.0003945187,0.016000025,-0.0571260825,0.3174203336,0.3449330926,0.0851912573,0.3284900784,0.4267836213,0.080735378,-0.2349987477,-0.0556238145,-0.2187708616,-0.0349604785,-0.2483057678,0.4837775528,-0.0381795317,0.1216896772,-0.142181322,0.036698062,-0.2608760893,-0.440820545,-0.1000710577,0.1918876469,-0.0298058689,0.010196479,-0.389636904,-0.0280155968,0.2536537051,-0.2719688416,0.1400227845,-0.13825804,-0.1423628628,-0.0342508368,0.1852953732,0.0559359342,0.239090845,0.2277609557,-0.1308041066,-0.3857408166,-0.4401707351,0.1353846639,0.0270662643,0.3381630778,0.1061841398,-0.2070374787,-0.4189092517,-0.139235571,0.2391747087,-0.1569141895,0.0960531235,0.0989907682,0.0812273249,0.165456146,0.252183646,-0.2988062501,-0.1856297553,-0.1773062944,0.0319295116,-0.1300621033,-0.0701901093,0.3265482187,-0.0100668529,0.1334317029,-0.117993705,-0.0943661481,-0.1205612123,-0.1552785039,0.0512902848,-0.1617365777,-0.3146482408,0.1664025784,-0.0358661972,-0.2476519793,0.2399175614,-0.3951127529,0.2129151374,0.1262060553,-0.2579472363,-0.3092864454,0.0559353307,0.0856449679,-0.2632970214,-0.0152955577,-0.0331743136,0.0292060673,0.1754637212,0.2231558114,0.0925877765,-0.0978393182,0.2361078113,-0.0709363371,0.6817715764,0.1603981704,-0.0134615982,0.1587851346,-0.1708619744,0.3219510317,-0.2369931191,-0.5023804903,-0.1412394941,-0.1059440747,-0.1054208204,-0.0283705853,0.109574452,-0.2827525437,-0.115599893,-0.2426892221,-0.0605009161,-0.54604882,0.1893932819,0.1710993499,0.0693428218,0.0568653457,0.1564288139,0.1233408228,-0.4360743463,-0.4252508581,-0.1272193491,-0.1278421879,0.0157799479,-0.2675440609,0.1987197399,-0.6068624258,0.1646698713,-0.0542106107,0.5287730098,0.0551700518,0.0009338917,0.1420914084,0.1686774641,0.7477430105,-0.2251750976,-0.1185975447,-0.0158345401,0.0547930151,-0.5928254128,0.0987460539,-0.1476437747,0.6286976933,0.2092178911,0.3737720251,0.0288933534,0.1745180041,0.0755364895,-0.1005474627,0.079070352,-0.077841498,-0.2256833166,-0.2200263143,0.3044427037,0.2742739916,0.3445194364,0.0810018852,0.3223793507,-0.2284775674,-0.2969676554,-0.0309453886,0.0437185802,-0.2290928662,0.2575103939,-0.0781481341,0.2960500419,0.4340247512,0.2177850455,0.2762292624,0.408185482,-0.2440405637,0.0885933414,-0.0196736883,-0.0375479199,0.1764096618,0.3452102542,0.2333936393,0.1530003399,0.1788771152,0.0665417537,-0.2147426754,-0.1878597289,0.5031189919,0.1382859349,-0.4760869145,0.0670427978,0.3620651662,-0.0627520233,-0.0308657922,0.0745157823,0.5442926288,-0.3988694847,0.6471534371,0.1135982275,0.6712856889,0.089194715,-0.0725566596,-0.2630755305,0.1109268218,0.7352765799,-0.3347020149,0.25493747,-0.2097499371,-0.5855283737,-0.0690590367,-0.2858499587,0.0086796787,-0.2234174311,-0.3603481948,0.0166503042,-0.3822994828,0.2821260095,-0.0841552094,0.3374670744,-0.1510460228,-0.1274801344,0.0352684073,0.1547400504,0.2227140963,0.2692329884,0.1916231364,0.0424441211,-0.0174229462,-0.2775602937,-0.494461298,-0.0191288274,-0.1274602115,0.1551729739,0.1827656329,-0.0453223921,0.3815469444,0.0988958329,0.0481477119,0.107557185,-0.1812240034,0.2406510413,-0.0080882208,-0.0569829531,0.4446687996,-0.220583424,0.0711796582,0.1791232228,-0.4024159312,-0.0184437446,0.2618075609,-0.1858665645,-0.0822697282,0.3471039236,-0.2522739768,-0.3732704222,-0.2347560972,-0.1285910159,-0.067549251,-0.1432647109,0.0892322212,0.0084903808,-0.1242847145,0.2830715179,0.3689793646,-0.2495323718,0.0721010566,0.8503792286,-0.2160516232,-0.1386739761,0.2074775249,0.060395591,-0.1718024909,-0.1100918576,0.3264315724,0.285936892,-0.1506772935,-0.2592438161,-0.4627506733,0.2283157706,0.0369627327,0.037715055,0.2122460753,-0.0092372168,-0.184262231,-0.1904779226,0.1433349401,0.1566762626,0.4442587197,0.1021856293,-0.3198876679,-0.405808568,-0.0958278179,0.1375002265,-0.243810907,0.3268664777,-0.2520278394,0.0064505953,-0.129592225,-0.0516157895,0.5856395364,-0.3752815127,0.0591960773,0.1744844466,-0.0512193516,-0.1685403883,-0.0637019202,0.1191376373,-0.0450762287,-0.1070081294,-0.1666346788,-0.2316609472,-0.0473485813,-0.2600593269,0.0268881936,0.4245568514,0.129269436,-0.1663024724,-0.1522525251,-0.1260861605,0.3549259007,-0.0860512331,0.2210680395,0.2574340999,0.0169345103,0.2850543857,0.1095146686,-0.0313541964,-0.1166432649,0.2705664635,-0.0437093563,0.0783668011,0.2038312554,-0.2371039093,-0.0961813927,0.5086316466,0.4194711745,0.394295156,-0.3104347289,-0.1650985628,-0.0005253511,0.2413312495,0.0492484048,-0.2417104393,0.6126328111,0.090026781,0.1283886582,0.2012028694,0.0890386701,-0.3130941987,-0.3525202274,0.0291854721,0.7695252299,0.1349657774,0.106252335,-0.1847231537,-0.0298995078,-0.0307875723,0.0969654322,0.1540451199,0.3628753126,0.2108799964,0.0093515022,0.4457625151,0.0946389288,0.2002345026,0.195820272,0.1463286877,0.2819838822,0.2778770626,0.1983456612,0.0648833215,-0.0310216248,0.815058291,-0.0081768036,-0.3089000583,-0.1074742749,0.0123540768,-0.0718853697,-0.1381998509,0.2133601159,-0.1653792709,-0.0008473348,-0.2416391075,0.0849005878,-0.0269591268,-0.3901277184,0.1575505435,0.1124823689,-0.043095123,0.0508896858,-0.141497016,0.0767197385,-0.1664367169,0.0036225964,-0.1519070715,0.1252125502,0.2351682633,0.3312087953,0.4035592377,-0.074723579,-0.1400466859,0.4058650434,-0.0350114331,-0.1445527524,0.1079212949,0.0318367966,-0.0002727603,-0.5440949202,0.147816807,0.0487125851,-0.0635885522,0.0140420524,0.2640837133,0.4321984351,-0.1540914923,-0.3470183909,-0.2106317282,0.0744327679,-0.1783645749,-0.1506300569,-0.1600381285,0.0988606885,0.2971235514,-0.2123093009,0.1440627426,0.1429017782,0.0628881827,-0.086921744,0.1966791004,0.1160484254,-0.0550464578,-0.033396706,0.074752003,-0.2059546411,0.3186418712,-0.4508095384,0.169267714,0.5591514707,0.2960522175,-0.2047137916,0.1067671329,0.1807710826,-0.2350661755,-0.2993133366,-0.0772450045,-0.0160411503,-0.2238667756,-0.1015606448,0.0261780359,-0.0101414658,-0.0251561124,0.1090446934,-0.5017803311,0.0234621968,-0.2353968024,-0.1540967077,-0.1983961016,-0.1749744862,0.1152511016,0.0764285624,0.2546803653,-0.0108258994,0.0705437288,-0.1942732483,-0.4698742628,-0.2817729712,0.2653594911,0.0957904756,0.0309691727,-0.2558007538,-0.4981395304,-0.2089369297,0.1482592225,0.0951690748,-0.1051607206,-0.4935901165,0.0211842135,-0.0664745718,0.1254032254,-0.2824284732,0.3217877746,-0.0584607683,0.4555459917,0.0576231778,-0.4273312688,0.472407788,-0.3319600224,0.005578591,-0.0409944504,0.2926785946,0.6038848758,0.1409938782,-0.2386948913,-0.1673590839,0.2756532133,-0.1943458468,-0.2440593839,-0.098173894,-0.0150665436,0.1926191449,-0.0259733181,-0.1619805992,0.2742896378,-0.0292484723,0.0151276793,-0.0911699384]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2882","title":"`load_dataset('docred')` results in a `NonMatchingChecksumError` ","comments":"Hi @tmpr, thanks for reporting.\r\n\r\nTwo weeks ago (23th Aug), the host of the source `docred` dataset updated one of the files (`dev.json`): you can see it [here](https:\/\/drive.google.com\/drive\/folders\/1c5-0YwnoJx8NS6CV2f-NoTHR__BdkNqw).\r\n\r\nTherefore, the checksum needs to be updated.\r\n\r\nNormally, in the meantime, you could avoid the error by passing `ignore_verifications=True` to `load_dataset`. However, as the old link points to a non-existing file, the link must be updated too.\r\n\r\nI'm fixing all this.\r\n\r\n","body":"## Describe the bug\r\nI get consistent `NonMatchingChecksumError: Checksums didn't match for dataset source files` errors when trying to execute `datasets.load_dataset('docred')`.\r\n\r\n## Steps to reproduce the bug\r\nIt is quasi only this code:\r\n```python\r\nimport datasets\r\ndata = datasets.load_dataset('docred')\r\n```\r\n\r\n## Expected results\r\nThe DocRED dataset should be loaded without any problems.\r\n\r\n## Actual results\r\n```\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in \r\n----> 1 d = datasets.load_dataset('docred')\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 845 \r\n 846 # Download and prepare data\r\n--> 847 builder_instance.download_and_prepare(\r\n 848 download_config=download_config,\r\n 849 download_mode=download_mode,\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 613 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 614 if not downloaded_from_gcs:\r\n--> 615 self._download_and_prepare(\r\n 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 617 )\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 673 # Checksums verification\r\n 674 if verify_infos:\r\n--> 675 verify_checksums(\r\n 676 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 677 )\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 38 if len(bad_urls) > 0:\r\n 39 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 40 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 41 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 42 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/uc?export=download&id=1fDmfUUo5G7gfaoqWWvK81u08m71TK2g7']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.11.0-7633-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n\r\nThis error also happened on my Windows-partition, after freshly installing python 3.9 and `datasets`.\r\n\r\n## Remarks\r\n\r\n- I have already called `rm -rf \/home\/\/.cache\/huggingface`, i.e., I have tried clearing the cache.\r\n- The problem does not exist for other datasets, i.e., it seems to be DocRED-specific.","comment_length":69,"text":"`load_dataset('docred')` results in a `NonMatchingChecksumError` \n## Describe the bug\r\nI get consistent `NonMatchingChecksumError: Checksums didn't match for dataset source files` errors when trying to execute `datasets.load_dataset('docred')`.\r\n\r\n## Steps to reproduce the bug\r\nIt is quasi only this code:\r\n```python\r\nimport datasets\r\ndata = datasets.load_dataset('docred')\r\n```\r\n\r\n## Expected results\r\nThe DocRED dataset should be loaded without any problems.\r\n\r\n## Actual results\r\n```\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in \r\n----> 1 d = datasets.load_dataset('docred')\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 845 \r\n 846 # Download and prepare data\r\n--> 847 builder_instance.download_and_prepare(\r\n 848 download_config=download_config,\r\n 849 download_mode=download_mode,\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 613 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 614 if not downloaded_from_gcs:\r\n--> 615 self._download_and_prepare(\r\n 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 617 )\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 673 # Checksums verification\r\n 674 if verify_infos:\r\n--> 675 verify_checksums(\r\n 676 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 677 )\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 38 if len(bad_urls) > 0:\r\n 39 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 40 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 41 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 42 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/uc?export=download&id=1fDmfUUo5G7gfaoqWWvK81u08m71TK2g7']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.11.0-7633-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n\r\nThis error also happened on my Windows-partition, after freshly installing python 3.9 and `datasets`.\r\n\r\n## Remarks\r\n\r\n- I have already called `rm -rf \/home\/\/.cache\/huggingface`, i.e., I have tried clearing the cache.\r\n- The problem does not exist for other datasets, i.e., it seems to be DocRED-specific.\nHi @tmpr, thanks for reporting.\r\n\r\nTwo weeks ago (23th Aug), the host of the source `docred` dataset updated one of the files (`dev.json`): you can see it [here](https:\/\/drive.google.com\/drive\/folders\/1c5-0YwnoJx8NS6CV2f-NoTHR__BdkNqw).\r\n\r\nTherefore, the checksum needs to be updated.\r\n\r\nNormally, in the meantime, you could avoid the error by passing `ignore_verifications=True` to `load_dataset`. However, as the old link points to a non-existing file, the link must be updated too.\r\n\r\nI'm fixing all this.\r\n\r\n","embeddings":[-0.2659113109,0.3422692716,0.0355651341,0.3098572195,0.214197889,0.0404327884,0.3646671176,0.4201550186,0.3040324748,0.0715031847,-0.2768665254,0.190118432,0.1306955665,-0.1951108724,-0.1612918228,0.3339123726,0.1246936545,0.1053829789,-0.2242237628,-0.1252798289,-0.3106808662,0.1985796541,-0.2184446454,-0.2843059599,-0.0529530197,0.1835617423,0.1532600224,0.2324340791,-0.120700039,-0.3587095439,0.2788212895,0.2785292566,0.173633635,0.4296127558,-0.0001249282,0.1207987443,0.2817707658,-0.0242785923,-0.501280129,-0.2013867348,-0.563354671,-0.3742280006,-0.0070420792,-0.1583375037,-0.0344875716,0.2363334298,-0.0177811794,-0.2750721872,0.0001792605,0.3736942112,0.1168587953,0.3918669224,0.1610017419,0.0871690363,0.3812574446,0.0504982956,-0.0901036188,0.4119758606,0.219390288,-0.0385965593,-0.2582117915,0.1683268249,-0.4234545529,0.3069981039,0.232055366,0.0618822426,0.0648707673,-0.1373666972,0.253241241,0.2994743288,0.4981358647,-0.4146616459,-0.3278160989,-0.2285789698,-0.298303932,-0.3188637793,0.3791346848,0.1398549974,-0.1967773736,0.0336276442,-0.3680784702,0.1499671191,0.0037718869,0.1940130293,0.2422606051,0.0545322224,0.0316536799,-0.0187701844,0.0221102182,0.0298728757,0.3937070668,-0.5399644375,-0.2046732903,0.2048074454,-0.5590269566,0.0730827823,0.0080572087,0.2777058184,0.352638036,0.4925115705,0.08901117,0.2042775005,-0.0375440791,0.188513726,0.1643131375,0.2119750381,-0.0647413805,0.4100411534,0.2708570957,0.3656848669,-0.1347787827,0.0432133712,-0.0054317582,-0.1764253378,0.6712611914,-0.1259947717,0.3084286749,-0.4890398979,-0.4388926029,0.271324724,0.0082045635,-0.0947970897,0.0869470164,0.322021246,-0.226584658,0.1516590118,0.1470181793,0.1476904005,-0.2125054002,-0.0541327372,-0.187337622,-0.0764364675,-0.0596771426,0.1976072788,0.2942281067,-0.1821978688,0.2553656697,0.1083121598,0.2190237492,-0.1960808337,0.2999760807,-0.1955657154,-0.0533633046,0.3180696368,-0.0322253928,0.0961275995,0.2476602346,-0.1812210828,-0.1439391971,0.1828757226,-0.2557730675,-0.2137421519,-0.1269117147,0.1511545181,-0.3514274061,0.0082077142,-0.2378200293,-0.400021553,0.4063311517,-0.3223438859,-0.0154792899,-0.3542110622,-0.2637994885,-0.353944093,0.2389694899,0.4025940299,0.0419260897,0.0426536016,-0.0905655473,-0.0114815729,0.294190079,0.183512345,0.0093251551,0.1056279317,-0.4124949574,-0.0635623261,0.2606512606,-0.3542023897,-0.6170941591,0.1036195084,0.1277518421,0.6295824647,0.0476492308,0.0750021562,0.1831660569,-0.0502420217,0.2216965407,0.1532353014,-0.038785696,0.1244795322,-0.252807796,-0.0325587615,-0.0434744433,0.1044865996,0.0201719031,0.2498004735,0.2889224589,-0.1661153138,0.2637803555,-0.0953656435,-0.1257363111,0.2321082801,0.2991938293,0.046871759,0.0639734343,-0.0167484935,-0.7693794966,0.3423186839,-0.1505941451,-0.0280080661,-0.0733518675,-0.0444171354,-0.1803081781,-0.1141534746,-0.3073681295,-0.0912047699,-0.0125983637,0.3906361461,0.2074201256,0.0201071296,0.0379508324,0.4137744009,-0.3047952354,0.1248284727,-0.3346244991,0.4118489027,0.0743400902,-0.1029386371,0.1318240315,-0.0030252386,0.1328218132,-0.0562920161,-0.2745426893,0.4387036562,0.3811519444,0.1363846362,-0.1962373555,0.3414201438,-0.0699199662,-0.1395473629,-0.012621915,0.4393106699,0.1203998923,-0.1690554321,-0.1533752978,0.3473964334,-0.0006330401,0.1946744621,0.0062628747,0.0575816073,0.2450569719,-0.1167023629,-0.1461344063,-0.261870563,0.3650653958,0.1296010762,0.2266256809,0.196940884,-0.1572066098,-0.0258856118,0.4515280128,-0.0735068321,-0.0205244031,0.1288514733,0.0474654101,-0.0436953641,0.0753767639,0.2370779216,0.5027197599,0.0668875426,-0.04627854,0.1425257325,-0.1718104333,-0.0678285286,0.0968001559,-0.0407392457,0.1989537925,0.6168555617,0.1380470097,-0.1526708454,-0.2818833888,-0.0774433017,-0.1217622161,0.3029148579,-0.4903008938,0.0769678801,-0.1072904691,0.0858299136,-0.3731725514,-0.1268986762,-0.1855977178,-0.2465110421,-0.2533238232,0.6255096793,0.0291794557,0.1770886779,-0.543838203,0.0261698551,-0.0482735336,-0.3658800423,0.0178647786,0.2034313083,-0.2111107856,-0.154736951,0.5788052082,-0.0910895392,0.3742569685,-0.4110652506,-0.0268509947,-0.5500833392,-0.2233636677,0.0945177004,0.0162294693,0.2236746848,0.171183303,0.0406548828,-0.0192004945,-0.2501042485,0.2409217954,-0.0442294776,-0.2682688534,0.3236261308,-0.1510273069,-0.0955459252,0.1709357947,-0.1985139698,-0.0594486669,-0.3050559163,-0.1297381818,0.2319205552,0.2133797705,0.1936335266,-0.0064717545,0.0152918147,0.1389674693,0.337371707,-0.1682054251,-0.6134564281,0.4448585808,-0.0378274098,-0.220205918,-0.1443342119,-0.1069903895,0.1174569875,0.3974053264,-0.5095450878,0.0167492125,-0.288091749,0.0945630446,0.0950631127,-0.0240606703,0.2160272449,0.145353362,0.0829630345,-0.2499599308,-0.2719066143,0.1908720434,-0.1372274011,0.1996578574,0.0226447172,0.3025070429,-0.2673497498,0.5462415814,0.4274444282,0.0061363615,0.2280853242,0.0467436984,0.526879847,-0.1603343338,-0.3179928064,-0.0785081387,-0.2098642141,-0.1336312145,0.1367129833,-0.025604425,-0.1290741563,-0.1602381468,0.0565649085,-0.2643253803,-0.2664655745,0.0557018742,-0.2999257445,0.341845274,0.1807065904,0.1787686795,-0.1449635029,-0.3675031662,-0.0111274412,0.2551630437,-0.0621705726,-0.064657025,-0.2754209936,0.0361303054,-0.0963153616,0.3009771109,0.2166886777,0.4385888875,0.1286434382,-0.0846547931,-0.0186727438,0.023536874,0.2270695567,-0.2749712467,0.2956960201,0.1485588551,-0.0079607638,-0.1783613861,-0.2544582188,-0.0432508513,0.0618162937,0.0433870666,0.55647856,-0.3165850043,-0.0163535215,0.2413590103,0.4100426435,-0.1461860985,-0.2883686125,-0.440358609,-0.1881668419,-0.2969538569,-0.1807302386,-0.1434826255,0.1207043454,-0.2336365879,-0.1006112695,-0.017981587,0.0570383631,0.297167033,-0.0569480471,0.4486523867,-0.062713787,0.4389171004,0.0435416959,0.0607733615,0.4035254121,0.8385460377,-0.3677003384,-0.4420888424,-0.0795324668,-0.1300328821,-0.1770050377,0.0727124736,-0.0297103859,-0.1339790672,0.2369088978,-0.1334208101,-0.0876609385,-0.0725696161,0.2104186416,0.1998604238,-0.4859289825,-0.4705499709,0.1424980611,0.0854976997,-0.0564965196,0.579628408,-0.0678586066,-0.17265971,0.0975924656,0.1562774181,0.7910794616,-0.0374635085,0.2097775638,0.2671902478,-0.321650207,-0.0153308865,-0.2884845138,0.1176057309,-0.5410556197,-0.4565611482,0.0023745324,-0.410885036,0.0694896355,0.0244534891,-0.0058735418,0.3644321561,-0.1365393251,0.3639833331,0.1626484692,0.150323987,-0.1333116144,-0.019252805,-0.0246920493,0.0563706756,-0.1063247249,0.2817863524,-0.1043262407,0.0744758397,-0.0837384537,-0.2491396219,-0.3359660208,0.26861462,-0.0484248325,0.3063894212,0.3144675493,-0.1562400758,-0.1932050735,0.3079376817,0.1931915879,0.1256464273,-0.202425912,-0.096611917,0.4446141422,0.1859632283,0.1080899015,-0.1174749732,0.4294384122,-0.0164770056,-0.1706129313,-0.0195093155,-0.1567083597,-0.5109879971,-0.0638424307,-0.1865808964,-0.0139560504,-0.3327461779,-0.1821816862,-0.1491386443,-0.0026479869,-0.179938361,0.1034976915,0.1509203017,-0.1339936107,0.3799248934,-0.0015962229,-0.460085392,-0.1274677515,0.5131280422,0.1256911457,-0.0666733161,0.52543962,-0.0354989879,-0.2741117477,-0.1303777695,-0.0239341632,0.0429769866,-0.2795833051,0.1564933807,-0.2923184931,0.3266904354,-0.1656529903,0.230919674,0.1746156663,-0.02821506,0.0803735107,-0.5436577201,-0.1149597615,-0.0545217469,0.0714057386,0.2256109267,-0.096170336,-0.1318736374,-0.1686248928,-0.1643115431,-0.219647333,-0.0509599522,-0.2850660086,-0.0503922813,0.0666950047,0.1091088355,0.095470123,-0.0847984403,0.1070708409,-0.1395269483,-0.1586124599,-0.0982483998,-0.2473136783,0.1913026422,0.007808188,-0.2195226997,0.1011908725,-0.2507101297,-0.1231054664,-0.2926606834,0.041283261,0.333221972,-0.1663248092,0.0240588076,0.0479134135,-0.0059108702,-0.0702901557,0.059866488,-0.3567721844,-0.0949055552,0.0353378989,0.0851883814,0.0412023254,-0.0427586734,0.0382646769,0.2007402033,0.0974865928,-0.0341187306,0.4755352736,-0.4363497496,0.0402533934,0.1563017517,0.2356877029,0.4222279489,-0.3856227994,-0.1175228581,0.13966766,0.1710722893,-0.2858359516,-0.1420300454,0.1406693459,-0.2054573745,0.0249944143,0.2416557819,0.1442941874,-0.5312721729,0.3154070675,0.0365457907,0.1829687357,-0.1733940989,0.0952961743,0.4206505418,0.1013368517,0.1403630525,0.1325903833,-0.0310446788,-0.0105605861,0.0478046462,-0.0750081539,0.3036599159,-0.3571105599,0.0880224854,-0.2201276869,-0.5279682279,0.3481909633,0.2502381504,-0.0698255226,0.1256925017,0.0152221387,0.0834440887,-0.0615339875,0.2207084745,-0.3515161574,-0.11219991,-0.1160394847,0.0300300941,0.0056752041,-0.3382467031,0.0237839688,0.2282619774,-0.0487359576,0.1440568417,0.1589454263,0.1219190583,-0.1803323925,-0.2696266174,0.2587604523,0.0842613652,-0.0409872234,-0.1821427047,0.2920826972,0.0527241006,0.0603767931,0.0833991989,0.2604502141,0.4520942867,0.5969523787,0.1655570716,-0.2217404097,-0.0189450718,-0.0231258385,-0.0942030549,0.7059627175,-0.1801467538,-0.2599339485,0.3139159977,0.0530928262,-0.1471292824,0.0253863428,-0.0972277522,-0.2940345109,-0.0907661393,0.5079358816,-0.0236264933,0.0174965095,-0.0995528772,0.2220041156,-0.2275825441,-0.249173522,0.4326468706,0.2270053178,0.1376857758,0.0776116177,0.0341995284,-0.0186288971,0.1766093224,0.2722485065,0.0902166814,-0.378665477,-0.342133522,-0.7264214158,0.3454352617,-0.1576506495,-0.0619914234,0.1172379106,0.1326255202,0.0116173243,0.1041900218,0.314648211,-0.0641261786,-0.2050143182,-0.0098502068,-0.1759048551,-0.3004679978,-0.1592682153,-0.1377318949,-0.1920399964,-0.5703731179,0.2017878592,0.1151962057,0.0482627079,-0.1229476482,0.2097163796,0.0996090695,0.2080218643,0.1852984279,0.1540618092,0.6352700591,-0.0647114143,-0.2254853845,-0.0971970707,-0.1141671911,-0.1586394906,0.3163270652,-0.1148585752,0.3463056982,-0.1001497358,-0.3019786775,-0.3574885726,0.3911955953,0.2497394234,-0.3221425116,-0.0838530287,0.0728310719,-0.2557837963,0.2184863389,0.0259722173,0.1138151437,0.1335904598,0.4593373537,-0.1675980389,-0.2450248748,0.464666605,-0.2160876095,-0.159683153,-0.1233223602,0.257687211,-0.2743937969,-0.2150910795,-0.6132447124,0.1759458482,0.4812847674,-0.1356473714,-0.1669290215,0.2582945228,-0.2036875784,0.3876015544,0.016391566,0.3485515118,0.0122529911,-0.222143665,-0.0536836199,-0.1185232848]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2879","title":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"","comments":"Hi @rcgale, thanks for reporting.\r\n\r\nPlease note that this bug was fixed on `datasets` version 1.5.0: https:\/\/github.com\/huggingface\/datasets\/commit\/a23c73e526e1c30263834164f16f1fdf76722c8c#diff-f12a7a42d4673bb6c2ca5a40c92c29eb4fe3475908c84fd4ce4fad5dc2514878\r\n\r\nIf you update `datasets` version, that should work.\r\n\r\nOn the other hand, would it be possible for @patrickvonplaten to update the [blog post](https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english) with the correct version of `datasets`?","body":"## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n","comment_length":46,"text":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"\n## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n\nHi @rcgale, thanks for reporting.\r\n\r\nPlease note that this bug was fixed on `datasets` version 1.5.0: https:\/\/github.com\/huggingface\/datasets\/commit\/a23c73e526e1c30263834164f16f1fdf76722c8c#diff-f12a7a42d4673bb6c2ca5a40c92c29eb4fe3475908c84fd4ce4fad5dc2514878\r\n\r\nIf you update `datasets` version, that should work.\r\n\r\nOn the other hand, would it be possible for @patrickvonplaten to update the [blog post](https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english) with the correct version of `datasets`?","embeddings":[0.124852851,-0.0535938032,0.0293307602,0.0739796385,0.2047281116,-0.0245766733,0.2489377707,0.4213005006,-0.4746908545,0.1751622707,0.1361838877,0.5513864756,-0.3347690701,-0.0850955397,0.0344839208,0.1865644157,-0.0921123177,0.1384560466,-0.2678906322,-0.2130741179,0.0398925841,0.3963992298,-0.1676281989,-0.0685962737,-0.3741768599,0.244877398,0.2273771465,-0.2314912826,0.1054832488,-0.4600336254,0.2924744487,0.1486980319,0.0194595307,0.3702304661,-0.0001191299,0.0063226023,0.2757813334,0.0352541432,-0.2775231004,-0.4479430318,-0.0650265291,0.1085864976,0.0530338921,0.1475536525,-0.4656309485,-0.1793126911,0.2337317616,-0.3409471512,0.0292004794,0.4485591054,0.0685843602,-0.0634173304,-0.2496642619,-0.0824669749,0.3689354062,0.0125433672,-0.0625187159,0.0592703931,-0.0683355555,0.2411289513,0.2196747363,0.3755480051,-0.0313242152,0.0129956314,-0.1954175979,-0.0935640186,-0.2540271878,-0.3487864733,0.0747668445,0.2519178092,0.7106900215,-0.2958623171,-0.3215118349,-0.2683330774,0.3369950652,-0.0542203225,0.2911206484,0.0254176874,-0.2944933772,0.1627251506,-0.5629917383,0.0742394626,-0.3041745424,0.1429014206,-0.0974289998,-0.359634608,-0.143771261,0.1845409572,-0.1975932419,0.1163858026,0.114155367,0.1884384304,-0.0396124236,-0.0923857763,-0.3206233084,-0.0985560194,0.0559260733,-0.0815377459,0.0717144236,0.3172878623,0.2842969596,0.0675945058,-0.4174665213,-0.135505408,0.0322321728,0.1481688768,0.1823079735,-0.1778396666,0.101101391,-0.0720301494,-0.1533745825,0.0139923282,0.0189836212,-0.0720990822,-0.1062536985,-0.0153032569,0.3571920395,-0.1175772175,-0.5631766319,0.1525668353,-0.5025030375,0.2135901749,-0.0520560816,0.1947724819,-0.2154174149,0.1350183487,0.2971945703,0.0870628357,0.02487945,-0.3115249574,-0.0998023972,-0.1971572787,0.1828780323,0.0968628898,0.2440094203,-0.2103739977,0.2123047858,0.257135421,0.2615037858,-0.1538840383,-0.0926653668,-0.0326226279,0.2529811561,0.2668441832,-0.1248196512,0.3939130902,-0.0565123409,0.1157927513,0.106670782,0.2571235001,-0.0651218817,-0.0061710193,0.5623431206,0.060588941,-0.1836533099,-0.3006524444,-0.1071216986,0.3333047032,0.0442795344,-0.0668319687,0.2504209578,-0.2751781344,-0.2271512002,0.0158824958,0.1358926445,0.1566538513,-0.5853290558,0.0243286006,0.1945606619,0.0684714913,0.3222911954,0.3032742143,0.1651955247,0.1811709255,-0.0816664621,0.1170219034,-0.219933033,-0.3883649707,-0.2060429454,0.2092838287,-0.2708128095,0.264259249,-0.0832697079,-0.2192167342,0.0802669078,0.0991331413,0.3640693426,-0.1891600937,0.24668172,-0.1056700721,-0.3831144273,-0.1445571929,0.1893382818,0.1755265892,-0.0177115817,0.0532647818,-0.2982362509,-0.0164561383,0.4027053714,0.0394103304,-0.0291296821,-0.025731707,0.3644500375,-0.1775300503,0.263861388,-0.1554233581,0.2872300744,0.0272912737,0.1604733467,0.2198717296,0.5499910116,-0.1693307608,-0.1779452264,-0.0762331262,-0.0728189349,-0.0683157369,0.0582602769,0.1486589164,-0.1489184052,0.3247836828,-0.0285170637,0.4712553024,-0.4881338477,-0.0153098824,-0.0879812539,-0.1088415384,0.3033908308,-0.0987990201,0.1165293604,0.2411418706,0.0172993075,0.0872968137,-0.110142529,0.3636822999,0.2269500643,0.1219172627,-0.3682011068,-0.4027226269,0.2316687405,-0.5078880787,-0.4552857876,0.5677568913,-0.0276648402,0.2008169293,-0.1534452736,0.060441643,0.0807491466,0.0479999073,0.0348503962,0.0360783227,-0.019409189,-0.1079213247,-0.470849663,-0.1607666463,0.2607947588,-0.2098269463,0.1590214819,-0.1161290258,-0.320378989,0.1368902475,0.315872401,-0.0502400734,0.0395470038,0.2744301558,-0.1326723844,0.209983319,0.0375697017,0.1719760001,0.2373924702,0.2009201497,0.0807271153,-0.0535479188,-0.0781926885,-0.2605938315,0.2923918068,0.1945406348,-0.3651409447,0.3840377629,0.2203695476,0.0489813201,-0.4372256994,-0.0049824319,-0.1904428899,0.2824560404,-0.449403286,-0.1749291867,-0.3977611363,0.1078346595,-0.1361607611,-0.0218015406,0.0710958019,0.0436779261,0.3717918992,-0.0027077466,-0.0859973133,0.2743452489,0.1501078755,0.1915652007,-0.110422954,0.1713414043,-0.0096689994,0.1939062178,-0.4710183442,0.0452458076,-0.1606466472,-0.2644656301,0.1504383832,-0.1925743371,-0.1099141166,-0.4143466055,-0.071546413,0.1478731036,-0.1916618943,0.4232137799,0.2326952517,-0.0426136926,-0.1860425919,-0.194430843,-0.1887598336,-0.0310719088,-0.0671303943,0.0714476034,-0.1346394867,0.0415853858,-0.256131798,-0.8888216615,0.1003023162,-0.2204675525,0.7185165882,0.0599351153,-0.0460496917,0.5412366986,0.2822042704,0.2012919486,0.0792539567,0.1863065064,-0.3713881075,-0.1230798736,0.0792584717,0.0293458365,-0.4573093057,-0.298314631,0.1936385185,0.1548020095,0.2074109167,-0.4953384995,0.3497347236,-0.1089945436,-0.1306295842,0.0102862855,-0.1989106387,0.2906790078,-0.2038048357,-0.1159159765,-0.2554523647,0.1175914109,0.3896949589,0.1017955169,0.112711817,0.3398635685,0.2160488367,0.2003065348,0.3506759107,0.7158643007,0.0828187764,0.039017871,-0.1280344874,0.4244537055,-0.0728549138,-0.052832868,0.1600767821,0.0590854175,-0.1087116003,0.4583721757,-0.0842429325,-0.0255117994,-0.0293365568,-0.0642682612,-0.0904889032,-0.3277724385,-0.022520544,-0.4580339193,0.0512960665,0.1535509676,0.2963804603,0.1280051917,-0.1144675538,-0.0452858321,0.1441749334,0.3194569945,-0.1257756054,-0.3150956631,-0.3613077104,-0.415389359,0.0612773709,0.1869184822,0.5611051917,-0.2880935073,-0.201971963,0.2231436074,0.1043023616,0.4215117395,-0.3380554318,-0.042292852,0.1251643896,0.2728692293,-0.2976222634,-0.0735439733,-0.3262295723,-0.289985925,0.279160589,0.0972543359,-0.1177073866,-0.1351196617,0.0686608776,0.058421474,0.0167316217,-0.099004969,-0.0332190059,0.0306317639,-0.0272785462,0.1637768447,0.1725981981,-0.1248045266,-0.491910845,-0.3849042654,-0.0554664992,0.1896433085,0.5516821146,0.0221093055,0.2840608358,-0.1308684945,0.2439791858,0.235736981,-0.2960318625,0.2580799758,0.1936600059,0.1173487008,-0.0414018072,0.1370378286,-0.1260076612,0.3830217123,0.3855090439,-0.4349811971,-0.0189787298,-0.3385869563,0.1840335429,-0.2829484046,0.3459416926,0.4015236199,-0.1441545635,-0.1260796487,-0.4360538721,0.418184042,0.0229241867,-0.1640126258,0.4557494521,0.161981672,-0.2460984588,0.1927152127,-0.1567512751,0.8095886111,0.1871284097,0.2667512596,0.3340376318,0.1109344587,0.1609196216,-0.4287750423,0.0356836617,-0.0055845086,0.0698633865,-0.1306088716,0.089701362,-0.1681430936,0.3797655702,-0.2919414937,-0.0783907324,0.2101620287,0.197283715,0.2774983346,-0.0788463205,-0.2581412792,0.0186328627,0.0897454992,0.0458310619,0.058744546,0.3057942092,0.2888186574,-0.1282046884,0.0097754784,-0.265276581,-0.6595223546,0.0203248579,-0.599691987,-0.09471035,0.1489200592,-0.3863419294,0.183830902,0.3397494555,0.1764554679,0.4337405264,-0.4633069336,-0.0267995987,0.2511091232,0.1131987423,-0.0246270187,-0.0830141678,0.3481529355,-0.0840320587,-0.1434060186,-0.0710047334,0.0281203054,0.16605708,-0.1545241773,-0.0496859439,-0.1185674593,-0.3426322937,0.0497712791,0.0989603326,0.1527892351,-0.2369833142,0.0680043399,-0.1038809866,-0.0651224554,0.2742635012,0.2779468596,-0.5032378435,-0.1089579538,0.5379329324,-0.0835737288,0.1077651456,0.6408995986,-0.1893859357,-0.0512990244,-0.2775192261,0.0168702956,0.2590861022,-0.3368996084,0.1719175428,0.0321254991,0.30066517,0.0966014713,-0.0444742106,0.0463051498,-0.0589915104,-0.2513128221,-0.2389338315,-0.081248261,0.1200233996,0.0252261981,0.2911800742,0.0278385542,0.3665986061,0.0767617822,-0.347037375,-0.2017673552,-0.0237252936,0.0557583757,0.0804858208,-0.0478359275,0.0278626494,0.1107278466,0.0545184948,0.0149586732,0.1027084291,-0.1335515976,-0.0833661035,-0.2990173697,0.1511185914,0.2011461705,0.1130731627,-0.0153080691,0.0650832951,-0.0542607009,-0.0600513965,0.0761152506,0.3808921576,0.0174185708,0.3620303869,0.1188446134,0.3976139128,-0.0936357677,-0.2623862922,-0.0354406275,-0.0775718689,0.1644935161,-0.0624371767,-0.1850817502,-0.0580990687,-0.1324884593,0.0220973957,0.0893222392,0.1518605053,-0.0546601489,-0.2896298468,0.2135311365,0.1549112052,0.0506647453,0.3179009259,-0.1531378031,-0.2708627284,-0.1102764606,0.0789906234,-0.0246088952,0.0176900197,-0.3351352513,-0.0152753768,-0.0255123302,0.0031253609,0.0554810911,-0.3541377783,0.3623180091,-0.0125009436,0.1121347845,-0.5960302353,0.1129645109,0.3746998906,0.0613390915,-0.0079180803,-0.2586423457,-0.2529646456,-0.0035204799,0.4860927463,-0.2427177429,0.2749043107,-0.1475844979,-0.0262315013,0.2557944953,-0.2615301609,-0.2477716953,0.2055236101,-0.4631261826,0.160956189,0.101355657,0.4035341442,-0.0463260636,0.1006667241,-0.2105019838,-0.0210334621,-0.1480747759,-0.1586381793,-0.3575797379,-0.0887535512,-0.1383645236,-0.1716290116,-0.0802430212,-0.0285868235,0.1238230765,0.2392329723,-0.2778191268,-0.5319434404,-0.2491969317,-0.1834071875,0.251419425,-0.1140987724,0.3520128727,0.103128694,-0.0186601989,0.1629291028,0.3187715411,0.5043587089,0.1790820807,-0.1014207229,0.1203379259,-0.299844265,-0.0056224484,0.020882478,0.62394315,0.0326780751,0.1173429191,0.0509871952,0.0518923216,0.0013063183,0.0413613394,-0.0084985485,0.0400381945,-0.6968811154,0.4176638126,0.0176094957,0.2393384725,-0.2014484107,-0.0412869416,-0.2195517421,0.3418336213,0.6031695008,0.1864216924,0.0186088011,-0.0934841856,0.0555081293,-0.1810229719,0.1212917343,0.0766027942,-0.0048448374,-0.1068457067,-0.0195507258,-0.2139687836,0.0899255127,0.265216738,0.2187431008,0.0150080938,-0.1638361514,0.0671154484,0.0615044869,-0.0098666847,-0.1182038561,-0.2539238632,0.2244302034,-0.1969873458,0.0151608726,-0.0069401781,0.1406158358,0.1495863199,-0.0737226456,-0.1326639354,0.1910309643,-0.1454409212,-0.057077501,-0.0759500638,0.1773371249,0.3487201631,0.0686021373,0.0964073241,0.3599496484,-0.0915555507,0.07731089,-0.4108827412,-0.1420516521,-0.2299602181,0.2994242013,0.049812071,0.3389099836,0.0066421395,-0.3442056477,0.1354150027,0.2597314715,0.3330140114,-0.162827,-0.1400983036,0.0477316156,-0.323091656,0.2139669657,0.0000565254,0.2621964514,0.0430665314,0.1607147604,-0.5121689439,-0.9930148125,0.6250440478,-0.7181706429,-0.2236547172,-0.1427765787,0.2892617285,0.3226871789,-0.1015825123,-0.5634548068,-0.0728689656,0.4693164527,-0.0401696749,-0.0920308381,0.112321049,0.1479113549,0.3100144565,-0.164433822,0.9622838497,0.3237745166,-0.1628977358,0.1872403026,-0.4706269205]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2879","title":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"","comments":"I just proposed a change in the blog post.\r\n\r\nI had assumed there was a data format change that broke a previous version of the code, since presumably @patrickvonplaten tested the tutorial with the version they explicitly referenced. But that fix you linked suggests a problem in the code, which surprised me.\r\n\r\nI still wonder, though, is there a way for downloads to be invalidated server-side? If the client can announce its version during a download request, perhaps the server could reject known incompatibilities? It would save much valuable time if `datasets` raised an informative error on a known problem (\"Error: the requested data set requires `datasets>=1.5.0`.\"). This kind of API versioning is a prudent move anyhow, as there will surely come a time when you'll need to make a breaking change to data.","body":"## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n","comment_length":134,"text":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"\n## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n\nI just proposed a change in the blog post.\r\n\r\nI had assumed there was a data format change that broke a previous version of the code, since presumably @patrickvonplaten tested the tutorial with the version they explicitly referenced. But that fix you linked suggests a problem in the code, which surprised me.\r\n\r\nI still wonder, though, is there a way for downloads to be invalidated server-side? If the client can announce its version during a download request, perhaps the server could reject known incompatibilities? It would save much valuable time if `datasets` raised an informative error on a known problem (\"Error: the requested data set requires `datasets>=1.5.0`.\"). This kind of API versioning is a prudent move anyhow, as there will surely come a time when you'll need to make a breaking change to data.","embeddings":[-0.088485904,0.037396457,-0.0213857722,0.1496120393,-0.0189264342,-0.0777354389,0.2245492935,0.3621777296,-0.429489553,0.1633367538,0.1525295675,0.2624569833,-0.432490617,-0.0201306585,-0.071119301,0.2396375537,-0.1268624365,0.25981462,-0.3329852521,-0.0231152549,0.0825956166,0.2425760627,-0.0296556633,0.0283614192,-0.0865963995,-0.0143052982,0.2008777112,-0.093485117,-0.3122866154,-0.6014047265,0.2304865569,0.2764073014,0.0768745914,0.1653757393,-0.0001274555,0.1065067053,0.1652288288,-0.0230602305,-0.4907235503,-0.1459860653,-0.180824548,0.1132927239,0.011158362,-0.0032401655,0.1294730902,-0.2341334075,0.2061720639,-0.1966678202,0.3206728995,0.5292428732,0.0049742479,0.2393674999,-0.1227483228,0.1194096953,0.3548440039,-0.1067724153,-0.0561466739,0.0766564906,0.0378676206,0.2733804584,0.2134608924,0.3169699907,0.086405769,-0.1138325781,-0.290795207,-0.193882212,-0.1545206457,-0.2383064628,-0.2170854211,0.3288349211,0.753831327,-0.3150245845,-0.6087892652,-0.3401043713,0.1912582219,-0.1430546939,0.3923214972,-0.1902457625,-0.2573828101,0.3908389211,-0.6390121579,-0.2049098164,-0.3079322577,0.1295725256,0.015353418,-0.1147583276,0.0225799382,0.1235876381,-0.1734608263,0.0642666593,0.1816413701,0.1108352542,0.038729921,-0.0109756803,-0.1475488394,-0.3412714601,0.0423946492,-0.1861126125,0.0421029404,0.4260554612,0.5204828382,0.0642907992,-0.5445080996,-0.0775154456,0.1737653315,0.2407868356,-0.0252583344,-0.2616614103,0.15692994,0.0426035337,0.1903326213,0.1748280823,0.0237038396,-0.3405395448,-0.1293011457,0.1194925904,0.3347736299,-0.1888237745,-0.5020070672,0.0776049346,-0.454580307,0.1854378134,-0.0716408938,0.120723933,-0.2241946012,0.419606179,0.1554582864,0.0640978739,-0.1348512471,-0.4040712416,0.1883217096,-0.1559716016,0.1908436567,-0.0499088056,0.2960176468,-0.3722983003,0.0881588981,0.1427764744,0.1810636073,-0.1214631051,-0.1279874742,0.1010570377,0.1780343056,0.2132166326,-0.1896715909,0.1723005623,-0.0805888027,0.0884558037,-0.0172109585,-0.0199682843,-0.1741253138,-0.1103576198,0.3484823704,-0.0193801634,-0.2260963023,-0.2067113668,-0.0028787269,0.1341750324,-0.0354007855,-0.2685537934,0.1478761584,-0.115805082,-0.3188738525,-0.0152796609,0.0157302301,0.2144396305,-0.6616707444,-0.1349116415,0.1824918687,-0.0098238224,-0.0802729949,0.2591881156,-0.018249182,0.1335975379,-0.0214253534,0.061384391,-0.0431077182,-0.4288671315,-0.3187329769,0.1890932769,-0.2718712687,0.2365475744,0.1732592136,-0.2317657024,0.3257893324,-0.1745207906,0.1762593389,-0.2554750144,0.1786212027,-0.1710287333,-0.4389976859,-0.4511151314,0.031018842,-0.0533370189,0.068120487,0.0648189187,0.0564532466,0.0423984602,0.3488164246,0.0829287097,0.0118160006,-0.0338022485,0.200045526,-0.0843634456,0.0186142046,-0.224367395,0.0837669671,0.2081040293,0.0765817836,0.1705437154,0.424785018,-0.1788806766,-0.2476136684,0.0383234285,0.1872706562,0.126285255,-0.0395877548,0.2273307592,-0.0825186446,0.3504412174,-0.2704238892,0.4691356719,-0.4019192457,0.0372846611,-0.0536281206,-0.1635846496,0.2474771291,0.0825913176,0.1019182429,-0.0394122601,-0.1032757536,-0.0926558524,-0.024524264,0.518794775,0.1530638039,0.2767786682,-0.1384766251,-0.2536575794,0.2143708169,-0.2618277371,-0.3397199214,0.5486403704,-0.0972192958,0.4484133124,-0.0255129375,0.1951500177,0.1404719353,-0.040309187,0.0705885664,-0.0731970817,-0.0213092566,-0.0093516512,-0.5304045081,-0.1736079603,0.1465597302,-0.2134280056,0.1417781264,-0.1460386813,-0.2046640664,-0.0846562311,0.2301444709,-0.0212985929,0.0592088923,0.1565668881,0.0832527056,0.211786449,0.0553550608,0.3183932006,0.3474945128,0.0831101686,0.0246619973,0.0214087591,-0.2565384805,-0.1302022934,0.2768206596,0.2188893259,-0.3133857846,0.3226563931,0.2063322961,0.0317539014,-0.4395200908,0.0319717452,-0.0409543738,0.0367235951,-0.6262536049,-0.0090599591,-0.5513135791,0.1690156907,-0.3306770027,-0.08182659,-0.1623638719,0.0047671883,0.5723306537,0.1694382876,-0.0812915936,0.2194667459,-0.087634325,0.3680031896,-0.1986395717,-0.0618027821,0.0382376686,0.235950008,-0.4149255753,0.0028021303,0.0273967143,-0.3462299407,0.2414893061,-0.4148816168,-0.3009605408,-0.2322787941,-0.0679201856,0.1891996413,0.0653937161,0.2402111292,0.3457305431,0.1270100325,-0.0146292243,-0.0418712981,-0.2779378891,0.0484673493,-0.1159968078,-0.1051494107,0.1344976276,0.0982847586,-0.2827249169,-0.8775423765,0.0841022357,-0.1689092368,0.4604318142,0.0795201138,0.1856422573,0.3558037281,0.346203953,0.1059153229,-0.0441331379,0.1544438154,-0.0882074237,-0.2302579731,0.0935372785,0.0836211443,-0.2186214924,-0.0057524103,0.2806986868,0.0527223162,0.3915783763,-0.4168854058,0.2677829564,0.2194800079,-0.1391707957,0.001931228,-0.368018806,0.5466492772,-0.0834411085,0.0156952888,-0.2377285361,0.205624491,0.1374920607,0.122576125,0.2549115419,0.3231088817,0.2804227471,0.2665038407,0.5263994336,0.3263044357,-0.2103032321,0.2247066796,-0.1228169724,0.4726446271,0.0411540791,0.0922420695,-0.0651079491,-0.0960020721,-0.1824163198,0.3834352195,-0.0364108309,-0.0770952627,-0.1888981909,0.2259602696,-0.2749902606,-0.272918582,0.0425743796,-0.3952446878,0.1360569596,0.2876459062,0.3582558036,0.10156288,0.0322360359,-0.0319893025,0.2270751446,0.3397321999,-0.0098600397,-0.205221802,-0.0876502991,-0.2094879746,0.2032209635,0.4069237113,0.6541618705,-0.1864350736,-0.0886587799,0.1288477182,-0.0191108957,0.2862899005,-0.4888174534,-0.0456894152,-0.0792832151,0.3533954322,-0.1327290684,-0.1294834763,-0.1357855201,-0.3984862864,0.1170825735,0.0794346184,-0.1247220188,-0.0330578312,-0.0315417647,0.1264813244,-0.1087495461,0.1763421148,-0.1473098397,-0.0187249426,-0.1351140738,-0.008912567,0.0840163231,-0.278080076,-0.4168049693,-0.0725171417,-0.0888437629,0.2681666613,0.3895969093,0.1621502638,0.2874218822,-0.1167992502,0.0239205156,0.0930395275,0.0980092809,0.3213100731,0.2371076494,0.1738164723,-0.1843378693,0.0630599037,0.0820485875,0.101782836,0.5416073799,-0.336302489,-0.1183999628,-0.1456222981,0.0961889178,-0.2815686762,0.3941451907,0.539711833,-0.2854302824,-0.3250682056,-0.4289255738,0.4246477187,-0.0904475898,-0.2281326652,0.3700994253,0.1016527489,-0.3230112493,0.1511365473,0.0276337042,0.9220039845,0.2088724077,0.2067637593,0.1859174073,-0.1246660948,0.4331418574,-0.5090640783,0.1522121876,0.1198083013,0.0101093641,-0.3668872118,0.1092012301,0.0915936753,0.439971447,-0.1749594063,0.0485190079,0.1938182414,0.3199470937,0.243715927,-0.0224065706,-0.4007433057,0.095558621,-0.1448121965,-0.0000623971,0.1024669558,0.2897033691,0.0265425239,-0.0909916908,-0.0498299003,-0.4163598716,-0.7078104615,-0.2266657054,-0.4140581489,-0.1478801817,-0.0445111394,-0.119061254,0.1190106422,0.2762684822,0.2014925331,0.2418511808,-0.3773293197,0.0200512931,0.0144654009,0.0742946267,0.1023879573,-0.0653457791,0.3629058599,-0.0745208785,0.0413267799,0.0539514199,-0.0798187256,-0.0671611503,-0.102445133,-0.0455864854,-0.0551971756,-0.3396161795,0.1325672865,-0.1026446521,0.0545611791,-0.1985493749,-0.0092557035,-0.2127401829,0.1121369526,0.1885818541,0.0892388821,-0.4363735914,-0.0978496596,0.6311433911,-0.2886928022,0.1289472431,0.4841496348,-0.3160094321,-0.0333173946,-0.1941363215,-0.0213660095,-0.1992967427,-0.5194515586,0.0147852106,-0.0446038842,0.0800760463,-0.127022177,0.073364988,0.1802175939,0.1770586073,-0.1799434125,-0.3772439957,0.0420021489,0.2579897642,0.2447490245,0.3469192088,0.0996372178,0.2187812775,-0.0042199041,-0.5193978548,-0.1086388677,-0.1001783013,0.1144042015,0.0074447608,0.0580743849,0.3495641649,0.0714690611,-0.2546701133,-0.1440987289,0.1539246291,-0.1497370154,0.0224602874,-0.2373713255,0.2285162508,0.0468573086,0.1454444677,-0.0425269417,-0.0164453555,-0.0305593424,0.0448148474,-0.0886746049,0.3584471941,0.1018417403,0.191381067,0.1421771497,0.4053522646,-0.2433033288,-0.0650703013,0.068246454,0.1981734782,0.1225374714,0.2254868448,-0.2010090202,-0.0713312626,-0.1664502025,0.2583247721,0.2247251868,0.0503340773,0.0594959557,-0.1397160292,0.2466144562,0.0786284432,0.3503730297,0.6881967783,-0.1185354069,-0.4578858316,0.1832003444,-0.0270989817,-0.1317240745,-0.025419686,-0.3005431294,0.0465860665,-0.1627551615,0.1273007691,0.0149540277,-0.3517835438,0.460877955,0.1213573515,0.1400361806,-0.4294632971,0.1858078986,0.6682862639,0.0034412649,0.2189691067,-0.4196910262,-0.3038707972,0.1169409975,0.3169998527,-0.3769645095,0.2690323591,-0.1761023998,-0.0776479989,0.0045769424,-0.1181805208,-0.3301261067,0.2845059335,-0.0586046278,0.4738274813,0.5105355978,0.5142284632,-0.0172043759,-0.0909561291,-0.4596327543,0.0934664384,0.0878878832,-0.1986487806,-0.5646128058,-0.2639584839,-0.1481105834,-0.0202088729,-0.1672478616,0.1114047915,0.2192398906,0.28841874,-0.2779063284,-0.5320429802,-0.2158830315,-0.2374021709,0.3170849979,-0.2049438506,0.2701926231,0.2036055326,0.077412039,0.2706266344,0.3161221743,0.405833602,0.091557838,0.0079941442,0.0263181683,-0.2494978458,0.2277772725,0.0264974926,0.4185414016,-0.0022926908,0.1665539294,0.3360890448,-0.0296551175,0.0563846491,0.1897279322,-0.3109255135,0.0160298627,-0.8132576346,0.3678290546,-0.0193114169,0.2532857358,-0.205313772,0.0359720923,-0.0377472825,0.3639247119,0.374653399,0.2076763511,0.0715543926,-0.001427614,-0.0013188893,0.042796094,0.1832635552,0.3172470033,-0.0597401001,0.0115216132,-0.0177085325,-0.5737120509,0.0423169583,0.0510372594,0.0625441298,-0.0109656798,-0.2795841694,0.2245611697,0.3119564056,0.1439362615,-0.0390034355,-0.2761303186,0.3555093706,-0.1490615904,0.0674827844,-0.0336016417,0.2049396783,0.0936795622,-0.4758703113,0.1259448081,0.0696859658,-0.2321201712,-0.0411871411,-0.023927616,0.1523931175,0.1186016798,0.2039049417,0.0417625122,0.3559782207,-0.0677288547,-0.1148146093,-0.5655712485,-0.0974949151,-0.245974496,0.312805146,-0.1374589652,0.3027483225,-0.1450625956,0.0440846756,-0.0077923709,0.3880821764,0.3565621674,0.0990418866,-0.1623943746,-0.2619998753,-0.3227412105,0.1392250359,0.265111953,0.2450653315,-0.0310609378,0.1129158437,-0.4395873845,-0.7549797893,0.4463395476,-0.5544971228,-0.0952443779,0.0162800103,0.3003892601,0.3546909094,-0.0889625624,-0.5038310885,-0.1051778495,0.3698131144,0.0651458949,-0.0476119258,0.0583503135,0.1136684343,0.0546242669,-0.02786153,1.0447940826,0.216182217,-0.1762980819,0.1037134603,-0.4345042109]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2879","title":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"","comments":"Also, thank you for a quick and helpful reply!","body":"## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n","comment_length":9,"text":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"\n## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n\nAlso, thank you for a quick and helpful reply!","embeddings":[0.1843589097,-0.0476369262,0.0212153625,0.1307302862,0.1825789064,-0.0150703816,0.222696349,0.4102399647,-0.4767715335,0.1482010931,0.202155903,0.528619945,-0.3603378236,-0.1392592937,0.0697514489,0.1514802277,-0.0296482705,0.1163090765,-0.2856072187,-0.2430297881,0.0460388102,0.3836102486,-0.181789279,-0.0110180005,-0.3692496121,0.2394319922,0.2274626493,-0.1810913682,0.1650811285,-0.439029336,0.2945018709,0.0983762294,0.0515014045,0.3712564111,-0.0001205389,-0.0167328734,0.2610318065,0.0087951152,-0.2081606984,-0.3689297736,-0.1382844448,0.1195933074,0.0133538358,0.12241216,-0.4130392373,-0.2065744102,0.229093805,-0.3684004843,0.0639212877,0.4605449438,0.0608402193,-0.139359951,-0.2579361498,-0.0793896466,0.3681408465,-0.0151606631,-0.0316133946,0.0299798381,-0.0864327401,0.2322303355,0.2330624908,0.3767850101,-0.0501611307,0.0072294213,-0.2207338065,-0.0934685469,-0.2900455594,-0.3867571652,0.0662252977,0.191508621,0.7059831023,-0.3136530221,-0.2619046569,-0.2177537978,0.3014097512,-0.0254760459,0.2628924251,0.0440048538,-0.2687057555,0.1821450889,-0.5459147692,0.1098416448,-0.3146458268,0.1415527016,-0.0640842691,-0.3725824058,-0.0981010497,0.2094152719,-0.2525474727,0.1218328848,0.1371880621,0.2098523825,-0.0407567285,-0.088758558,-0.3333061934,-0.0827903822,0.0905025452,-0.055269625,0.0512395315,0.2592660189,0.323356241,0.0425352305,-0.3752723634,-0.1163501963,0.0656884834,0.1570367366,0.0961320028,-0.1948441863,0.118978247,-0.1403936744,-0.1913475394,0.040307723,-0.0055863252,-0.091643773,-0.0389522314,-0.0102081764,0.3480204046,-0.1361118853,-0.5908363461,0.1955148876,-0.5599816442,0.2329349965,-0.0647064596,0.2073190659,-0.187874347,0.1397468895,0.3171749115,0.0926715434,-0.0059219012,-0.3403543532,-0.0963281542,-0.1902716309,0.1971511841,0.0616736077,0.2055280805,-0.1423812211,0.2137727439,0.2923474014,0.2695505917,-0.1473722309,-0.0583484769,-0.0180806592,0.219842881,0.2483473718,-0.1362519711,0.4016667604,-0.0248719845,0.1940863729,0.1011622474,0.2526570261,-0.1005024016,-0.0175146516,0.5515517592,0.0515450686,-0.179110989,-0.2624935806,-0.0348146893,0.3113076985,0.0846941844,-0.1433292776,0.2118196636,-0.2840275764,-0.1747703999,-0.003525354,0.0937831625,0.1310754865,-0.5885374546,0.0558972545,0.228044942,0.1303474754,0.3747636676,0.2952876389,0.1747010648,0.2124052495,-0.0759727508,0.1145521328,-0.1961544603,-0.4299621284,-0.214312613,0.1697297394,-0.2615546882,0.2817715108,-0.0361457802,-0.2367807031,0.0825164616,0.075459443,0.409607172,-0.2213379294,0.1860471368,-0.0808987841,-0.3983337283,-0.1363315433,0.2186726034,0.1355755478,-0.0341216251,0.0313469507,-0.2578755319,-0.0243165027,0.3640209436,0.0280066933,-0.004362273,-0.0479702353,0.3482239544,-0.2454301864,0.3254101574,-0.0976360664,0.3093805313,0.0151669383,0.1946799606,0.1750653833,0.6044173241,-0.1565310061,-0.170942992,-0.0799789056,-0.0831656903,-0.0937429518,0.0424369089,0.1493053585,-0.1577068269,0.3208678961,-0.0529016294,0.496171236,-0.4533744752,-0.0638343394,-0.0522209033,-0.1226176918,0.2798168063,-0.0861343965,0.0933275223,0.2543967366,0.0215519834,0.0870476365,-0.0931508169,0.3177514076,0.2538267672,0.1660675555,-0.3391086459,-0.3895223737,0.2347876579,-0.5418210626,-0.418391794,0.6494663954,0.0136631234,0.2044710368,-0.1246106178,0.0206717309,0.0992714986,0.0414252579,0.0030764376,0.0376889668,-0.0036091113,-0.0781878829,-0.4513742328,-0.0986297056,0.2685841024,-0.2420790792,0.1837171465,-0.1044279113,-0.3128755689,0.1960173547,0.3068779707,-0.0273673404,0.0484327525,0.2494100779,-0.1171611398,0.1943250597,0.0596161298,0.2010252476,0.235914126,0.1829334795,0.0925755948,-0.0770207793,-0.0987522975,-0.221516192,0.2697022557,0.1580257267,-0.3681939244,0.4219260216,0.2311407924,0.0666007549,-0.3795827329,0.0026456879,-0.2107334882,0.3031838238,-0.4800261259,-0.1368284374,-0.3495121598,0.1713373214,-0.1380009055,-0.0323438756,0.1311494857,0.0800623074,0.370326221,-0.0135218948,-0.1168963611,0.2293738723,0.1091258749,0.2434615046,-0.0670005679,0.2318317592,0.0222459622,0.2099670172,-0.4693998694,0.015322485,-0.1299001575,-0.2572244108,0.1167961061,-0.1661652029,-0.0994341001,-0.3659780025,-0.0571099631,0.151061967,-0.1844133586,0.4728702605,0.2280860841,-0.0554880984,-0.1941481084,-0.2376766503,-0.2153888792,0.0371617377,-0.0857695267,0.097694777,-0.1236358881,-0.0152868517,-0.2216340005,-0.9135431647,0.0924015716,-0.2011582255,0.6722809076,0.0548606068,-0.0086461287,0.4621789753,0.2635646164,0.1870574206,0.138909936,0.2119944692,-0.3575545549,-0.106063135,0.0823722705,0.0353392884,-0.4551185369,-0.3046335578,0.1577614248,0.1924328506,0.2105116844,-0.4597027302,0.4100583196,-0.0954082683,-0.1507000029,-0.0038921644,-0.2178034633,0.3120282888,-0.1800539792,-0.0859067366,-0.2840114534,0.1318731755,0.3946647942,0.1762969643,0.1079193801,0.3068368733,0.1843467206,0.1532683223,0.3919531703,0.7317343354,0.1164943278,0.0385542214,-0.1042457595,0.3726763427,-0.0699758604,-0.0704685152,0.1748739481,0.0528061725,-0.2046083659,0.4510253072,-0.0907783583,-0.1019510776,-0.0252721105,-0.0426658839,-0.1361756176,-0.2871514559,-0.0318492763,-0.4308846891,0.0729355887,0.1623313129,0.2992659807,0.1566134095,-0.0742487907,-0.0494469069,0.1143832132,0.3183398247,-0.1257064044,-0.3194031715,-0.405916661,-0.4125874341,0.0458473675,0.1671956629,0.555062592,-0.2708597481,-0.2594034672,0.2183424979,0.0556486286,0.4326915443,-0.3497639894,-0.0037947437,0.1397782266,0.3352205157,-0.2780133188,-0.0910181105,-0.3201158047,-0.2889475822,0.2534392178,-0.0221099649,-0.0810841843,-0.0848103017,0.0755050629,0.0781073496,-0.0023477995,-0.0481448509,-0.0511059426,0.0388418064,0.0123623246,0.1542546749,0.1453919858,-0.1345046908,-0.5091332197,-0.3764441311,-0.0203373488,0.197188735,0.5483554006,0.0524414107,0.3182862401,-0.1757279932,0.25930655,0.2130944431,-0.3146297634,0.2446758151,0.1653646082,0.1780990511,-0.0280466471,0.1700089127,-0.1276296973,0.4133140445,0.3716487288,-0.4411073625,-0.0122789433,-0.3978454173,0.1657234579,-0.2691042423,0.3032995164,0.4508009553,-0.1429353505,-0.1288248301,-0.4156200588,0.3972937763,0.0255686529,-0.2025948465,0.4813393652,0.1077151895,-0.2843468487,0.2128332257,-0.1285715252,0.8244360089,0.2676287889,0.3113027811,0.3640472591,0.1112722233,0.1022783965,-0.4576124549,0.0514674373,-0.0159365274,0.1021314114,-0.1459210366,0.0862596482,-0.1860912293,0.3048148155,-0.3261877596,-0.073388651,0.1394101828,0.2140429318,0.2349338233,-0.2076794058,-0.219132632,0.0628909096,0.1287455559,0.0310560446,0.0454352461,0.2764178514,0.2411041856,-0.141391173,-0.0235028788,-0.3043451309,-0.660400033,0.0256115347,-0.5773504972,-0.0898341462,0.1136530787,-0.3812579215,0.1459791064,0.3813396096,0.1606675088,0.3877428174,-0.4796431065,-0.0438009612,0.3158276379,0.1177165657,-0.0222175177,-0.1181874424,0.3234867454,-0.074841246,-0.1256530583,-0.0707839131,0.0170849375,0.1853922606,-0.1924300343,0.010592225,-0.1102366969,-0.325163871,-0.0179609898,0.1750960201,0.1333635449,-0.2425092608,0.0644174665,-0.1092584059,-0.1126320586,0.3161510825,0.2585337162,-0.5235105157,-0.0768600926,0.5095415115,-0.05945234,0.1247373223,0.6273759007,-0.2404686511,-0.0542512424,-0.275106281,0.0932660177,0.2636379004,-0.3139445484,0.1428882778,-0.0181948114,0.2839416862,0.0443141162,-0.0612466894,0.0597400367,-0.033192832,-0.2500084341,-0.2497020364,-0.0879926533,0.1590440422,0.0418185443,0.2457344234,-0.0120161958,0.283083111,0.0409386642,-0.3694509268,-0.1864431202,-0.0472353101,0.0774314255,0.0784023553,-0.0327920094,0.0524554141,0.0643761083,0.0200634375,0.0109805102,0.0574669391,-0.1392928064,-0.0746839494,-0.3100247085,0.1609307975,0.1798962355,0.1112993583,-0.0083414037,0.0556965023,-0.0509433709,-0.0491366871,0.0704786181,0.3807947636,0.0208634771,0.3144367337,0.0940055251,0.3792684674,-0.0854802057,-0.2379428893,-0.0430187918,-0.1058757231,0.1939326227,-0.0524486452,-0.188233614,-0.0853269547,-0.0319134593,0.0517282598,0.0332106687,0.1841089427,-0.0188174192,-0.2987866402,0.2861653864,0.1175718531,0.0132432804,0.3378902376,-0.2234210521,-0.248046279,-0.0910891816,0.0822845176,-0.0294995066,0.0094374726,-0.3606777191,-0.0174245946,-0.0946175754,-0.0077177826,0.0422165394,-0.4409456551,0.3382523358,0.0060723345,0.1259869188,-0.617092371,0.1211182028,0.4069335461,0.0471916907,0.0044105747,-0.3285924792,-0.2809558213,-0.0062797265,0.5289809704,-0.2464639395,0.2679083049,-0.0634808689,-0.0275064819,0.2425554693,-0.2426433712,-0.2594917715,0.2379479855,-0.4665975571,0.1766881943,0.1410185397,0.3673169315,-0.0485259555,0.0456962809,-0.240078032,-0.0296354685,-0.1623637378,-0.1589698941,-0.3425230086,-0.1122073233,-0.1174796894,-0.1609565914,-0.0854225755,0.0221237019,0.0852115899,0.2617628276,-0.2954109311,-0.5197121501,-0.2400555909,-0.1835346073,0.2313261777,-0.1279245317,0.3786039948,0.10281308,-0.0337539278,0.1412706077,0.3286019266,0.5195795298,0.2080286592,-0.1701619327,0.1073838472,-0.2793742716,-0.0176598448,0.015043973,0.6274505258,0.0008645584,0.1270019561,0.0226037018,0.0276397411,0.0152008086,0.0441954024,0.0171362832,-0.0371454172,-0.7086550593,0.4125433564,-0.0367944427,0.2317141145,-0.1660160571,-0.0049939775,-0.2587536573,0.3277559578,0.621121943,0.1853351891,0.0533866957,-0.0270174704,0.0449458137,-0.1548006386,0.1108004078,0.0185913462,0.049389597,-0.1143924668,-0.0499974191,-0.1892929673,0.0780865923,0.2131735981,0.2548357248,-0.026068788,-0.1599316001,0.0669078305,0.0486767702,0.002179692,-0.0890494362,-0.2402725965,0.2845950425,-0.227225706,0.0184556376,0.014658561,0.145174697,0.1081263795,-0.154785648,-0.1130904257,0.2015322894,-0.1334021538,-0.0073894253,-0.1033231467,0.1589587033,0.3646920919,0.1079781055,0.0701356903,0.4246194661,-0.0901377499,0.1119568273,-0.4082079232,-0.1465588659,-0.196915254,0.2526400387,0.0476315953,0.3335848749,-0.022573825,-0.3501974046,0.1585386097,0.2298382223,0.3724767268,-0.0926154405,-0.1263129711,0.0814921409,-0.3730500937,0.2268996835,-0.0261833817,0.2177128643,0.0186660364,0.1487790346,-0.4728176296,-0.960010469,0.6462187767,-0.7156583071,-0.2400015891,-0.1577966362,0.3278788328,0.2929249704,-0.1062858477,-0.5133001208,-0.118125312,0.4647008777,-0.0780134201,-0.1065440103,0.1552323401,0.1537450105,0.2477950007,-0.1198170334,1.0183166265,0.3149241507,-0.161011979,0.2086753994,-0.4965496659]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"I have changed line 288 to `if int(datasets.config.PYARROW_VERSION.split(\".\")[0]) < 3:` just to get around it.","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":15,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nI have changed line 288 to `if int(datasets.config.PYARROW_VERSION.split(\".\")[0]) < 3:` just to get around it.","embeddings":[-0.4137373269,0.0904145688,0.0879315063,0.069035016,0.284868449,0.1465740502,0.1234253496,0.3872671723,-0.1835217625,0.1793182641,0.3690841198,0.3502765,-0.2362233698,0.1078671366,-0.2026903927,0.0031209812,0.0010433139,0.2549073994,0.1760860234,-0.0239745975,-0.1098649725,0.1690122485,-0.2097337544,0.2731696069,-0.2981304824,0.0923148543,0.0032627089,-0.0122110816,-0.1285185814,-0.5752780437,0.2998543382,-0.1617958844,0.2332223952,0.391263485,-0.000127086,0.0779013038,0.3378002942,-0.0440747142,-0.2238939106,-0.4914156199,-0.085651435,0.0122515438,0.3829289079,-0.0937147513,0.0937459767,-0.6206139922,0.0335906781,0.0384339988,-0.1429862529,0.3645019233,0.127537936,0.0344468243,0.1980867237,-0.0754810423,0.2223289162,0.083245784,-0.0282249879,0.5229272842,0.1158733368,-0.1320393234,0.233240068,-0.102022633,0.0575134605,-0.0043903464,0.0711091682,0.1914277673,0.2576049566,-0.1857423782,0.2388329953,0.0887792706,0.5562196374,-0.6069078445,-0.4294553101,0.0136653185,0.174752906,-0.2191064954,0.2185028791,0.2610872686,-0.1244944483,0.1687700599,-0.0923660845,-0.1329844445,-0.0807051584,-0.0779537112,-0.433144331,0.3678042889,0.1932168156,0.1645069569,-0.0400829166,-0.04866606,-0.0215911791,-0.0890382826,0.1259018928,-0.0294876229,-0.2407489419,-0.0118983239,-0.0643955991,0.1943915635,0.2201173902,0.1590977758,0.0649064705,-0.234773919,0.2536158264,0.0735231191,0.0475472286,0.0977316201,-0.0197430961,0.146631524,0.0936663523,-0.0873535872,0.0256926231,-0.0208997894,-0.0806714743,-0.0734729469,0.3739596009,0.0080980845,0.7578920722,-0.3194511235,-0.2312780768,0.1659606099,-0.3343805671,-0.184103936,-0.2247074097,0.2836925983,0.0676930696,0.3056817055,-0.1982494593,0.3802600801,-0.1360655129,-0.1857322007,-0.0950188115,0.1552126408,0.0275490843,-0.1461474597,0.0764700472,-0.2719792128,-0.0986580551,0.3349583149,0.0094830692,0.1213422492,-0.0362757482,-0.0120166596,-0.1315119714,0.4732261598,-0.2890482843,0.2727305889,0.2008104473,-0.4936734736,-0.3582565486,0.2217560709,-0.1869192868,-0.431489408,-0.2999134958,0.1262726486,-0.1483470798,-0.2103191763,-0.0671084151,-0.0523799025,0.1624187529,-0.1384892911,0.0297997743,-0.651668191,0.1586196274,-0.3059960306,0.2577728629,0.161972031,-0.6419220567,0.0579073988,0.1789008826,-0.1889813095,0.1506783217,-0.1405111104,0.0306624472,-0.1734109372,0.074796848,0.0583436862,0.3201945126,-0.2379271239,-0.3039623797,0.1230635121,0.0718067214,0.0206128471,0.2803038359,-0.3171919584,-0.0282199513,0.1652483642,-0.1928093135,0.1018786877,0.0079200668,0.0062634307,-0.1814045608,-0.0036330107,0.4458725452,0.2917095721,0.1499163508,-0.2299220115,-0.0262244828,-0.1855148673,0.1473240703,-0.0819691345,0.0311352387,-0.0489151254,0.3686681986,-0.2390524894,-0.0064461622,-0.2394002527,-0.2902263105,0.2341018915,0.0907700509,0.0002655792,-0.3193365932,-0.219354555,-0.2210743129,0.0846685618,-0.128561303,-0.011503078,-0.0059905695,0.3605172932,0.0959836021,0.0846642852,-0.3152723312,0.1812611818,0.0422432944,0.1910732836,-0.0209711008,0.4087104499,-0.4075936377,-0.3384674788,0.0850214958,0.0640254617,0.1028376147,0.002861667,0.0299142897,0.3328282237,0.0242639706,0.0389278978,0.0299662594,0.2096638978,-0.0284052063,-0.2111189663,-0.1741658747,0.3066881597,-0.0031621177,0.3704423904,0.0565479696,0.4024959803,0.5626875758,0.3306960762,-0.0120106991,0.0881720185,-0.1884381175,-0.0055437423,0.1211171523,-0.1257602423,0.0624467172,-0.1393158436,0.1693660468,-0.0199615061,-0.5397575498,0.0718814433,0.3745397031,-0.0722794607,0.1365482211,0.1192966476,-0.22845155,0.0563648492,0.0482595488,0.0757280812,0.249298811,0.1221992299,-0.1398308128,0.1419418901,-0.3319807649,0.1508725435,0.1715394557,0.053618446,0.1923853457,0.1373240501,0.2172163874,0.0542263351,-0.0518643856,-0.4222477078,0.1137933508,0.4095291495,-0.4994865358,0.2577461302,-0.2975929677,0.2889801562,-0.0827534124,-0.1969304681,-0.4268323183,-0.2678608894,0.0104828775,0.0935454518,0.1283202618,0.1767116934,-0.3569593728,-0.1166164875,0.0463055074,-0.5190931559,-0.252766639,-0.2070385814,-0.4546803832,0.0009253586,0.2450136989,-0.2954372466,0.0536534749,-0.2570302188,-0.2196664661,-0.2246728688,-0.1574143022,0.0125030046,-0.269426465,0.3128561378,0.2743914127,0.0396343656,-0.0582421124,-0.0491061211,0.3296924531,-0.3633875251,-0.2326972187,0.1670558602,-0.0441559106,-0.018140588,-0.0896702781,-0.5271107554,-0.2856226861,-0.174629271,-0.0225634612,0.0724860951,0.1668928266,-0.0621382482,0.2801332772,-0.0787437782,-0.1362159103,0.0422251076,0.0579393916,0.0014274698,0.2768362463,0.0047421837,-0.2914088666,0.010116118,-0.0066247168,0.1385644227,0.0028628155,-0.2945348024,-0.1095552146,0.0360653736,0.5227450132,-0.2845128775,0.0628193095,0.3160390556,0.2878689766,-0.0280942582,0.0421851873,-0.1316174716,-0.1662545651,0.1143491268,0.1350112706,0.1534068286,0.1728003323,-0.0685666278,0.6282868385,-0.0858015046,-0.2661091983,0.19313097,-0.3469312489,0.3226404786,0.0537074953,-0.1815225929,-0.0795435756,0.02120829,0.000880775,0.265568316,-0.202450186,-0.1276522726,-0.1136756316,-0.21760571,-0.2410297394,-0.0568175204,0.0081763715,0.0084404759,0.1998902708,-0.1092223823,0.0509707965,-0.0197257306,-0.0720054805,0.1099524423,0.3429470062,-0.2275398523,0.0822050944,-0.1880298704,-0.2194179147,-0.1872704327,0.4171535075,0.0073181456,0.4000034034,0.0758238211,-0.203572914,0.0676551983,-0.2018064111,0.5138761997,-0.1867123246,0.1210603341,0.2132676095,-0.1637834311,-0.4432633817,-0.4106668234,-0.5143594146,0.1034789085,-0.0334381461,0.163722381,-0.2714032829,-0.1977230757,0.2114056796,0.195720017,0.081494987,-0.0392780565,-0.461995393,-0.0728343725,-0.2166758329,0.2309352309,-0.1200748757,0.3576822281,-0.2043959945,0.5000504851,-0.1328083277,-0.1695317477,0.2581602931,0.2766701877,0.3142804205,-0.3184945583,0.0112793241,-0.2040047497,0.1823281199,0.537861824,0.5164781213,0.3591929376,-0.2701449394,0.1650438756,0.2548321784,0.5881131291,0.0124148075,-0.4793679416,-0.1435153484,-0.2246030867,0.0869210511,-0.055749163,0.0208998341,0.5394140482,-0.1436982155,-0.5579455495,-0.0006600026,0.583478868,0.1682926714,-0.0880463868,0.0389920734,0.4312092662,-0.1566647142,0.3882396221,-0.1652097702,0.7541747689,-0.0415149927,0.2032355219,0.1119088903,-0.3435962796,0.1856690794,0.0110415602,0.0853404105,-0.5221113563,-0.3559496403,-0.1561163366,-0.2469584793,0.1788385957,0.0282494351,-0.239993453,0.3036089838,0.0095087113,0.3756101727,0.3913848698,0.3188383281,-0.1210615188,0.3092973828,-0.2204645425,-0.0153665738,-0.0465542451,0.0767170414,0.1382596344,0.0722189397,-0.2499359548,-0.4431240261,-0.3027442992,0.0162141509,-0.2184962183,-0.1166363657,0.4740677774,-0.5513356328,0.1450444162,0.0259818938,-0.1213886365,0.0205297451,-0.1986920089,-0.0232671183,-0.132011652,0.1351816356,0.1822996736,0.0830752775,0.4477312863,0.0146222953,-0.3590239882,0.276563555,-0.0613866225,0.0766119584,0.0154431676,-0.0289588962,0.2499645054,-0.3149304092,0.1435465068,-0.5809526443,0.1648852527,-0.1438597441,0.0154867964,-0.1039261594,0.0030873038,0.0458917692,0.1371153742,-0.3003840148,-0.2312101871,0.2688654065,-0.011931234,0.2826273143,0.4671837091,-0.0964071751,-0.0290105585,-0.082691431,0.0665704608,0.3482677639,-0.3652615249,-0.0533109345,-0.0104720378,-0.0259197932,0.1340213865,0.2309290469,0.2441840917,0.2051969171,0.0225443952,-0.5867090225,-0.0156736895,0.1940641701,-0.1296601743,0.1184900925,0.1128129438,0.2511689067,0.2188224196,-0.1819416583,-0.1868157536,0.0008195406,-0.0286003239,0.1301870495,-0.2333950996,0.183599174,0.1691423208,-0.0048594819,0.1116361246,-0.2133685946,-0.1062259227,-0.0898168311,-0.1932307929,0.2013582289,0.0601094924,0.2008146048,0.1493998915,0.0889126062,0.0152651286,-0.0439368822,0.0690269545,0.1395592093,-0.2084579766,0.3365767598,0.4670518041,-0.1786255389,-0.3895530701,-0.0425577238,-0.2097871155,-0.0377730764,0.1553899646,0.3876922727,-0.1798083931,0.1443675905,-0.3008017838,0.0501429252,-0.2175431103,-0.0649992079,0.1967027634,-0.382879734,-0.2250436544,-0.0442183875,0.4638454616,0.2430250347,-0.2055921704,-0.0171396378,0.1072365046,0.0048530609,-0.4478969276,0.1075395718,0.2346013784,0.1743809432,0.0245360527,0.5318636894,0.4067471027,-0.3246034682,-0.0742826611,0.164279595,0.4959196746,0.2708314657,0.1485649049,0.4434695542,-0.0889714733,0.1561041474,0.4896700978,0.1625347584,-0.0874976143,0.3993495703,-0.1631687433,0.1351977885,0.106490612,-0.1803329736,-0.0160063393,-0.1632703543,0.2863210738,0.017498076,0.3956526518,-0.2600688338,0.2041577548,0.3499060869,0.1165681928,-0.1896251142,-0.092125386,0.0162337255,-0.0093539674,-0.0449644402,0.135077551,-0.2281519026,-0.2638019025,-0.0954566151,-0.1304732263,-0.2548211217,0.7299135923,0.2903822362,-0.1669585854,-0.2881604731,0.1720102429,0.1545021385,0.0507362522,-0.2968220413,0.3778815866,0.1143395901,-0.074858211,0.2609856129,0.3107405007,0.8127143383,0.2363016456,-0.1687476635,-0.2336084694,-0.2092974782,0.0317804553,0.291803956,-0.0576246791,-0.028149046,0.0891762078,0.2830816209,0.0896404311,0.0992728993,-0.0181849543,0.0861287341,-0.0633634105,-0.3945927024,0.320821017,0.1960968226,-0.1580650061,-0.0013216215,0.2794041336,-0.2914029956,-0.0390996598,0.1896317601,0.5572887659,-0.1122135371,-0.1232099086,-0.0135795595,0.2728503048,0.3097299635,0.5396063328,0.2115000337,-0.2938640416,-0.156917423,-0.4964832366,0.3750751913,-0.197897613,-0.1750097573,0.2619361579,0.1028776541,0.0685790107,0.170511499,0.3704152703,0.3715750575,-0.3486966193,0.1317123473,-0.3753184974,-0.2713752091,-0.0102243423,-0.1323414594,-0.0787532032,-0.0523299165,0.1964601576,-0.6846225858,0.0174367577,-0.1400099099,-0.2763794065,0.1993915886,-0.0236322135,0.4791323841,-0.0159135181,0.6772804856,0.0820816085,-0.15606305,-0.21632348,-0.0316805914,-0.4409960806,0.5421468019,-0.2116658688,0.2731279433,-0.1207861826,-0.3032947481,-0.2011571079,0.2629315555,0.0276308414,0.0753050745,0.06213497,0.1530801654,0.0903730989,-0.0284187868,-0.0078368885,-0.0506124459,-0.032961823,0.2801569998,-0.5794116259,-0.7179991007,0.5333272219,-0.2753734291,-0.3065237999,-0.3928979933,0.2762829959,-0.0685962513,-0.0433019735,-0.5575699806,0.0828975588,0.4288054407,0.0711633116,-0.2072430998,-0.138777554,-0.0423490182,0.3275108933,-0.1119783968,0.5136018395,-0.3148345053,-0.1485731006,-0.225925386,-0.2605141401]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"Hi @bwang482,\r\n\r\nI'm sorry but I'm not able to reproduce your bug.\r\n\r\nPlease note that in our current master branch, we made a commit (d03223d4d64b89e76b48b00602aba5aa2f817f1e) that simultaneously modified:\r\n- test_dataset_common.py: https:\/\/github.com\/huggingface\/datasets\/commit\/d03223d4d64b89e76b48b00602aba5aa2f817f1e#diff-a1bc225bd9a5bade373d1f140e24d09cbbdc97971c2f73bb627daaa803ada002L289 that introduces the usage of `datasets.config.PYARROW_VERSION.major`\r\n- but also changed config.py: https:\/\/github.com\/huggingface\/datasets\/commit\/d03223d4d64b89e76b48b00602aba5aa2f817f1e#diff-e021fcfc41811fb970fab889b8d245e68382bca8208e63eaafc9a396a336f8f2L40, so that `datasets.config.PYARROW_VERSION.major` exists\r\n","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":47,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nHi @bwang482,\r\n\r\nI'm sorry but I'm not able to reproduce your bug.\r\n\r\nPlease note that in our current master branch, we made a commit (d03223d4d64b89e76b48b00602aba5aa2f817f1e) that simultaneously modified:\r\n- test_dataset_common.py: https:\/\/github.com\/huggingface\/datasets\/commit\/d03223d4d64b89e76b48b00602aba5aa2f817f1e#diff-a1bc225bd9a5bade373d1f140e24d09cbbdc97971c2f73bb627daaa803ada002L289 that introduces the usage of `datasets.config.PYARROW_VERSION.major`\r\n- but also changed config.py: https:\/\/github.com\/huggingface\/datasets\/commit\/d03223d4d64b89e76b48b00602aba5aa2f817f1e#diff-e021fcfc41811fb970fab889b8d245e68382bca8208e63eaafc9a396a336f8f2L40, so that `datasets.config.PYARROW_VERSION.major` exists\r\n","embeddings":[-0.3742081225,0.1478579193,0.0884070471,0.0458860211,0.2804539502,0.1281492561,0.1531104892,0.3776514232,-0.2427559048,0.2011216879,0.3052013814,0.4177031815,-0.1865011901,0.1149607599,-0.1479020417,0.0099561084,0.0530921482,0.2520797253,0.1291816086,0.0060467068,-0.1062645391,0.1722441018,-0.1942494661,0.2815794349,-0.2942166626,0.1177279726,-0.0419812873,-0.0479392149,-0.1290827245,-0.5930629969,0.3376209438,-0.1589786112,0.2437154949,0.3399909437,-0.0001267467,0.1114126891,0.4063085616,-0.0081883511,-0.2549678385,-0.5021974444,-0.1217937693,-0.0536484718,0.3664952815,-0.0817655995,0.102711834,-0.7000570297,0.013414585,0.049935773,-0.1659477353,0.329718858,0.1360016465,0.0637335405,0.2616915703,-0.097520113,0.2251864374,0.0621270798,-0.0079679107,0.4649414122,0.0723097548,-0.1101893187,0.2356003225,-0.0848251805,0.0687128976,-0.0513685308,0.0992346108,0.1825542748,0.2629990876,-0.1627095938,0.1989027411,0.0318586826,0.5882601738,-0.626342833,-0.4298146665,-0.0176614802,0.220206961,-0.2607834935,0.2262852788,0.2429365963,-0.10094136,0.1628725231,-0.0570024997,-0.1206212416,-0.063535206,-0.0164811891,-0.4222541451,0.3365793526,0.1881286055,0.140308246,-0.1132394746,-0.009329834,0.0025072652,-0.1097331718,0.10199292,-0.078015849,-0.2746736109,-0.0144315083,-0.0775343105,0.1633089334,0.1757595986,0.208408609,0.0030351472,-0.2792110145,0.2704052031,0.0860595629,0.1035962552,0.1269080788,-0.0204388909,0.1562927812,0.1005842909,-0.1070765704,0.0690039098,-0.0031423518,-0.0581314154,-0.0568243526,0.3815536797,0.0283413101,0.742669642,-0.2355119437,-0.2066710442,0.1739302725,-0.3724194169,-0.179325074,-0.2660050988,0.2187235057,0.0305577051,0.3607175052,-0.1849215776,0.3320734203,-0.0956049785,-0.2356070876,-0.113595061,0.1457725167,0.0022741908,-0.1549615115,0.0268715788,-0.3101926744,-0.0878965631,0.3905238211,0.0014736976,0.1260318905,-0.1415989846,-0.0541735664,-0.1536239386,0.4754760563,-0.2805734873,0.2540762722,0.1305656582,-0.4681425095,-0.3429700136,0.2116095126,-0.2033770382,-0.4039486051,-0.4072811902,0.1281087101,-0.1272283196,-0.1809758395,-0.1497177035,-0.0052619977,0.2306341529,-0.1509694159,-0.0146371135,-0.6668522954,0.1032236964,-0.319904536,0.2467015982,0.1543766707,-0.6335021853,0.0087504741,0.1704790741,-0.1900213808,0.165815562,-0.1197979152,0.005515072,-0.1398759037,0.0290386174,0.0476634167,0.3468930423,-0.1955857724,-0.3685946465,0.0728684291,0.058083415,-0.0002380217,0.2594134212,-0.3139027059,0.0474226773,0.1808189899,-0.1679071486,0.0811104253,0.001909597,0.0160693731,-0.1740115881,-0.0421568789,0.3828534484,0.2783960104,0.1499212235,-0.1601757854,0.0293607879,-0.2005846798,0.1286726147,-0.0868610367,0.0851052552,0.0264915284,0.4046365619,-0.1430395395,0.033946842,-0.2175856382,-0.3197008371,0.1867123246,0.0698746145,0.0265328661,-0.3560784161,-0.1848592609,-0.1795900613,0.1088649258,-0.1084898785,-0.0283278916,0.0058021918,0.3832255006,0.0711459517,0.1115474254,-0.2974824309,0.2475375384,0.0242408346,0.1781353056,-0.0212112162,0.4727578163,-0.4082576334,-0.3153122663,0.0902254656,0.0457065813,0.102861464,0.0391666442,-0.0002053427,0.2879419625,-0.0028909601,0.0584653802,-0.0276767816,0.2250142992,0.0138650322,-0.1725012511,-0.192661047,0.2565892935,0.0011452289,0.4009828568,0.1184246242,0.3715967536,0.5518648624,0.3002988398,-0.0234602094,0.09365049,-0.1508531719,-0.0168528631,0.1085570678,-0.0839205682,0.0393525958,-0.1069295108,0.2100666016,-0.0077000619,-0.5035149455,0.0840201378,0.3833167255,-0.0595607199,0.1452025324,0.1436303854,-0.2609054446,0.0484301373,0.0339332297,0.1345786303,0.2741984427,0.1518546194,-0.1246563643,0.1309271455,-0.3188162744,0.164799884,0.1828395575,0.1089477018,0.1759979427,0.1831200421,0.2829288542,0.086618349,-0.0567072183,-0.3240150213,0.1387683749,0.3523797393,-0.4344475269,0.2530074716,-0.3444136083,0.250282228,-0.1669087112,-0.1862104386,-0.4147616029,-0.2429396659,-0.0304380916,0.0963686109,0.1005030423,0.2397005558,-0.3122081459,-0.1296312362,0.0462817326,-0.3988128304,-0.2186687887,-0.1758570671,-0.4486293495,-0.019090917,0.2013365775,-0.3951259255,0.039651487,-0.2199624628,-0.2137857676,-0.3060598373,-0.1851968169,0.0833885372,-0.2393283546,0.2992102802,0.2597303987,-0.038242083,-0.0473477803,-0.0988906175,0.3235451281,-0.3321583569,-0.2847580612,0.1006459966,-0.0942858085,0.0356610715,-0.0603935421,-0.4970048368,-0.3575113714,-0.1462482959,0.0124236261,0.041099906,0.1555687338,-0.0624720603,0.2309518009,-0.1249248758,-0.1489116102,-0.0105736004,-0.0276867412,0.0039672754,0.2373251915,0.0062804506,-0.2969486415,-0.0511556789,0.0064248098,0.0896267369,0.0367034785,-0.3144406974,-0.1439506412,0.0547291487,0.5389251113,-0.3207369447,0.0192000363,0.2970771492,0.2694778144,-0.0367961302,0.0630886257,-0.1787241399,-0.1892798841,0.1446789801,0.1415685862,0.1744132191,0.2410916984,-0.0849480182,0.5790010095,-0.0703439489,-0.298630327,0.2510004938,-0.29209131,0.3411943913,0.0630541816,-0.1798293442,-0.0886980519,-0.0483298227,0.0156402048,0.2517925501,-0.2184791267,-0.1496745795,-0.1239425465,-0.2539528012,-0.221225664,-0.042609781,-0.0612108223,0.0359983332,0.2440867573,-0.0551080666,0.0298614856,-0.0403753035,-0.054597225,0.1456445307,0.3016491532,-0.2157077044,0.1017485484,-0.1830856204,-0.152707696,-0.1803331375,0.3837957382,0.0439700186,0.4377391338,0.1552539915,-0.1999727935,0.0833069757,-0.1641715169,0.5928218365,-0.1187062114,0.1239395142,0.1585399061,-0.1371153593,-0.3955423534,-0.4143469036,-0.4618233144,0.083942309,0.00496139,0.1973275989,-0.2952902317,-0.1378474683,0.2305309922,0.1844815016,0.0506920479,-0.0268239118,-0.4164669812,-0.0658165142,-0.16919218,0.2111219019,-0.126962468,0.3692860305,-0.1657081842,0.4781220555,-0.1405241489,-0.1776957959,0.1654678583,0.2207294255,0.338403821,-0.3253705204,-0.0032508457,-0.1957048923,0.2242892385,0.6268402338,0.5107086897,0.3266793787,-0.3425282538,0.2157906592,0.2769287527,0.5334735513,-0.019866595,-0.482755363,-0.0958601087,-0.2471357435,0.0602886491,-0.0659033507,-0.032102935,0.5593261719,-0.1096102819,-0.5021070242,-0.0202025585,0.5698218346,0.1894673705,-0.0793709159,0.0957337916,0.3778422773,-0.1557371318,0.3602989912,-0.1334694028,0.7824115753,0.0175452381,0.2209131867,0.1155825779,-0.4412602186,0.2862614989,-0.0613791905,0.0911196545,-0.5176571608,-0.3366225064,-0.1464529783,-0.2395018488,0.1303436756,0.0145639181,-0.206828028,0.2747061551,0.047589004,0.4116238654,0.3609587848,0.3171049058,-0.0719520301,0.2720511854,-0.252182126,-0.0217562318,-0.0593513064,0.0917054042,0.1281195432,0.0586952567,-0.2239385247,-0.4248760045,-0.3383872807,0.0584696122,-0.2342780083,-0.0421386547,0.4508948922,-0.5700247288,0.118259497,0.0426155031,-0.1173641682,-0.0329631642,-0.1581275463,-0.0164402071,-0.0627333745,0.1840555221,0.1415057629,0.0766142309,0.4222027957,-0.011285712,-0.3214787543,0.2916197777,-0.0392378792,0.0190975238,-0.0512054153,-0.0198285989,0.1973825842,-0.2740911245,0.1948555112,-0.5745176673,0.2753389776,-0.2236594856,0.018647559,-0.1024588272,0.0007495828,-0.0163126327,0.1178036332,-0.2898898125,-0.2125737667,0.3666992486,0.0528494865,0.282312274,0.5269957185,-0.0676542446,-0.0817728192,-0.0454060286,0.0528813414,0.2541744709,-0.3957654536,-0.0471700318,-0.0543740988,0.03274839,0.0790852234,0.1984802634,0.2237325758,0.2555364966,0.0201187208,-0.5458897352,-0.0552603416,0.174301967,-0.1844223738,0.0464428887,0.202157706,0.2829450071,0.1664931476,-0.200484857,-0.1850072443,-0.0044213291,-0.0431108475,0.139864549,-0.1864715964,0.1530421674,0.1342078894,-0.0325627141,0.1182299331,-0.21702604,-0.0706173331,-0.074098587,-0.2204393148,0.1982030272,0.0624699034,0.239382863,0.1128435358,0.0884267464,-0.0120732607,-0.0264758989,0.1066364795,0.1785605848,-0.2272362262,0.3060315549,0.4929539561,-0.2622906864,-0.3427831531,-0.0386489853,-0.1529734135,-0.1157705635,0.1555685252,0.4499003589,-0.1631075442,0.1811010689,-0.2985671461,0.0400748439,-0.238818258,-0.0395774134,0.180485189,-0.4256383181,-0.2396808267,-0.0433830358,0.4485315979,0.2221607566,-0.2303750366,-0.098058641,0.1173264757,0.0090721501,-0.449053973,0.0356157497,0.264077127,0.1550159156,-0.0410299823,0.4426931143,0.4489092529,-0.2873223126,-0.0974351838,0.1774400622,0.4901801348,0.2559277713,0.1621134728,0.5159202218,-0.0438941792,0.1582222879,0.4991028309,0.2005461901,-0.0998899266,0.4372412562,-0.112453796,0.1261387914,0.1164664105,-0.2075036913,0.0423413739,-0.1860411316,0.2623975575,0.0324441902,0.3752452433,-0.2654273808,0.1894052476,0.3312999606,0.205977574,-0.2006644458,-0.1015894637,0.0663827136,-0.0794084296,-0.0449131653,0.1113160029,-0.2534873486,-0.229363516,-0.0513659008,-0.1425132304,-0.2100434154,0.6946986914,0.2750623524,-0.1903173476,-0.3078861237,0.1221392453,0.1174288988,0.0489128195,-0.2738500237,0.3728533387,0.080597274,-0.0812714919,0.2231171429,0.3539702296,0.8215025067,0.2467615157,-0.0757995397,-0.3166638911,-0.1711066663,0.0428912602,0.2550635338,-0.01291807,0.0328659192,0.0312773809,0.2523137927,0.0757340938,0.0942193642,-0.0254333951,0.1040955782,-0.0186157599,-0.4002721012,0.3342917562,0.1377340406,-0.232922852,0.0334776565,0.3048773706,-0.2463754267,-0.0460598394,0.2661556005,0.5599740148,-0.1200682446,-0.1416714638,-0.0092102,0.2585321069,0.3113091886,0.5630508661,0.1585423648,-0.2242979407,-0.1859928071,-0.4756968617,0.4308848083,-0.2150819898,-0.1064559594,0.2760234177,0.0639475137,0.0704118907,0.1569856107,0.3471615314,0.4209796786,-0.3910065293,0.1344710886,-0.3857069016,-0.2570206225,0.0194006171,-0.1237015724,-0.0705663562,-0.0510134622,0.1715069413,-0.5911445022,0.0238804184,-0.145424813,-0.2321884781,0.172145009,-0.0550273545,0.5031514764,0.007037668,0.6460561752,0.0470305867,-0.1947334409,-0.2284395993,-0.1045869589,-0.4692241848,0.6017167568,-0.2683476508,0.2230907828,-0.1235534325,-0.3353490829,-0.2232298255,0.2529105246,0.0031415622,0.0884141028,0.0043125926,0.1209901646,0.0713862255,-0.0406132862,-0.0173885003,-0.0586990155,0.015012959,0.3013275564,-0.5924042463,-0.6977592707,0.4954080284,-0.3196092844,-0.300034225,-0.4309999943,0.2480504811,0.0159253236,-0.0336424895,-0.5722135901,0.1170051396,0.4677253067,0.0741850808,-0.2211979777,-0.1495643854,-0.077095449,0.3797605634,-0.1158954501,0.4920510352,-0.2867052257,-0.1916608512,-0.2343590558,-0.2670706809]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"Sorted. Thanks!","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":2,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nSorted. Thanks!","embeddings":[-0.3823499978,0.0788534731,0.0835613832,0.0859657899,0.254578352,0.1588205546,0.1258136332,0.378461957,-0.1942313612,0.1517066956,0.3442723751,0.3460409343,-0.2396473587,0.1160570979,-0.1717098057,-0.0130585758,0.0182114933,0.2685298622,0.1562249959,-0.0352682173,-0.0831688866,0.1723991781,-0.221207574,0.2769206166,-0.290784806,0.1055777818,-0.034875419,-0.0441389605,-0.1249132976,-0.5678355694,0.3373684287,-0.1683481783,0.259375453,0.3960514963,-0.0001278901,0.0832418501,0.355887711,-0.0235939845,-0.2130245715,-0.4989036024,-0.1226171777,-0.0021165852,0.3659736514,-0.0908794627,0.0975263193,-0.645819068,0.0478101447,0.0595673434,-0.1573934406,0.3680754006,0.1251091063,0.040394716,0.2063285261,-0.0739030614,0.2163523734,0.092064105,-0.0217345729,0.534501493,0.1071522757,-0.1533436328,0.2649391592,-0.1108957157,0.0430092327,-0.0005163433,0.0668540001,0.1680730581,0.2481788546,-0.173890233,0.2404453605,0.0941150188,0.6074262857,-0.5957273245,-0.4052781761,0.0457946397,0.190304175,-0.2164417058,0.2071692944,0.2712188065,-0.1031831801,0.1470091492,-0.0970233083,-0.1404293329,-0.0962809622,-0.0484454148,-0.4191841185,0.3295250833,0.1847844124,0.1627088934,-0.0797763839,-0.0509254895,0.0048299716,-0.1046580449,0.1460140944,-0.0359727815,-0.2669477761,-0.0234076288,-0.0792864859,0.217488572,0.217696771,0.1644843668,0.0548282899,-0.2486527711,0.2624913454,0.0772697702,0.0593933463,0.1027095169,0.0099531254,0.1515246481,0.0868951976,-0.0923113823,0.0064102691,0.0057760188,-0.0834435672,-0.0731900558,0.4025044441,-0.0256594121,0.7480243444,-0.2999703586,-0.2222675681,0.1814114749,-0.3430295587,-0.2135788947,-0.2431992441,0.2692058682,0.0672401413,0.310472101,-0.2242441922,0.3650549948,-0.1089137867,-0.199377805,-0.0976025686,0.1639591455,0.0397986546,-0.1415777802,0.0776998177,-0.2863925397,-0.0622021332,0.3482494056,-0.0095234104,0.1233666539,-0.0444290265,-0.0483978391,-0.1564430743,0.4665051401,-0.2961569726,0.2662153244,0.1866068244,-0.4694723487,-0.3469584286,0.228558898,-0.2082118392,-0.4199297726,-0.3010490537,0.1158056408,-0.1309493929,-0.2246068269,-0.1018703878,-0.0314013213,0.1599574685,-0.1390951127,0.0193729699,-0.6764339805,0.1813001037,-0.3093072474,0.2377792299,0.1493155062,-0.64209795,0.0522496849,0.1860803813,-0.1793422401,0.1692923605,-0.1044029519,0.0251485948,-0.1446892172,0.074836567,0.0642969012,0.3613587618,-0.2439746708,-0.3121292293,0.0909591764,0.1042780653,0.0162541959,0.2491032332,-0.3086020947,-0.0181004889,0.1892558783,-0.1845408976,0.0676994994,0.013640088,0.0026977183,-0.1708292961,0.0054476419,0.438311249,0.2977380157,0.1350409687,-0.2180688083,-0.0289862547,-0.2037833929,0.1285439283,-0.0818165392,0.0136220716,-0.0355041698,0.407566458,-0.211303249,-0.0053179162,-0.21974428,-0.2898447216,0.2070418745,0.0864594579,0.017564917,-0.310914129,-0.1828339547,-0.2235690951,0.092757836,-0.1389970034,-0.0187330078,-0.008984318,0.3708506823,0.0991881639,0.0983951762,-0.3148404062,0.1957128495,0.0238641016,0.1941994578,0.0052944184,0.4397611022,-0.4184099436,-0.3409918547,0.0828048363,0.0324428193,0.1147675738,-0.0007691359,0.0294004232,0.3283750117,0.0075974376,0.0259973463,0.0376454331,0.2178497612,-0.0152834198,-0.2490369678,-0.1754437536,0.2634848952,-0.0124529032,0.3771674037,0.0714703649,0.367174536,0.5579252839,0.311139673,-0.0011530713,0.081929028,-0.1928504854,0.0089487033,0.1333372742,-0.0902685896,0.0926108137,-0.1520457119,0.1808014959,-0.0404996648,-0.5255686641,0.0326813981,0.378573209,-0.0664351285,0.1505247355,0.1364178061,-0.2508790195,0.0396844186,0.0591744334,0.0680330917,0.2792570293,0.1244287118,-0.1432555616,0.1382844001,-0.3376169503,0.1645191908,0.1703767031,0.0325545408,0.1882993728,0.1368130594,0.2204149663,0.053874325,-0.0650231689,-0.3855646849,0.1114446223,0.4056575298,-0.4613749981,0.2288228422,-0.2899155617,0.3022912443,-0.0999656469,-0.1894143373,-0.4212338924,-0.2496228665,-0.0001924169,0.090336822,0.1711301655,0.1835670471,-0.3250343204,-0.1124665737,0.0609607361,-0.4758637249,-0.2720769644,-0.1949386746,-0.4537057579,-0.0113761248,0.2226081789,-0.3075243831,0.0475775488,-0.2794205546,-0.222134456,-0.2541949451,-0.156630829,0.0200997163,-0.2725733817,0.3184693456,0.2527645528,0.0669448599,-0.0692736134,-0.0613622926,0.3347893357,-0.3718841076,-0.2399427444,0.1490340978,-0.0621652342,-0.0054474128,-0.0345259421,-0.5232025981,-0.3153493404,-0.1216146052,0.0009835535,0.0737264007,0.1543922126,-0.0844632015,0.2646680176,-0.0991962701,-0.1470929235,0.0729259551,0.0268436,-0.0291636419,0.2536785007,0.0024928947,-0.2990637422,-0.015819408,0.006221428,0.1566198468,0.0125369066,-0.2939115465,-0.1314835399,0.0421486981,0.5094447136,-0.2864129841,0.0493873321,0.3148921132,0.2830879986,-0.0151688168,0.0654160902,-0.1438310146,-0.1625919193,0.1416301429,0.1366788298,0.1592234224,0.1884554476,-0.0858525932,0.6183635592,-0.066043742,-0.27083534,0.2227285802,-0.349478066,0.3271769881,0.0535552427,-0.1810140163,-0.0944271237,-0.0012391821,0.0283627696,0.2522964776,-0.2121816278,-0.1313127875,-0.1222149208,-0.2319981158,-0.2465988696,-0.0176805239,-0.0523007289,0.0142141161,0.2401451319,-0.0935465693,0.034567073,-0.0206153709,-0.0605601259,0.1290478557,0.3503609002,-0.2477122843,0.0702728033,-0.2205438316,-0.2247415632,-0.173276186,0.3985607326,-0.0039343834,0.4307931662,0.0841581747,-0.1833612025,0.0811553076,-0.1943857968,0.5388733149,-0.1934172809,0.1095087007,0.1825060248,-0.1371613145,-0.4131250083,-0.4019485116,-0.5120395422,0.0875898749,-0.0420206934,0.1591536552,-0.2654743791,-0.1794775724,0.2284601927,0.1847786903,0.0931383222,-0.0519520231,-0.4445238411,-0.057030078,-0.1848578751,0.2046760619,-0.1663069725,0.3586477637,-0.1895323992,0.4917454422,-0.1368259937,-0.1512624323,0.2483527064,0.287710011,0.3000683784,-0.3307991326,-0.0130946636,-0.1949541271,0.1695131212,0.5847437978,0.4820328057,0.3315340579,-0.294988364,0.1528007686,0.2649532855,0.5828864574,-0.0272528008,-0.483433634,-0.1476513296,-0.2480645627,0.0608517006,-0.0455113314,0.0324646272,0.5878044367,-0.1079147011,-0.5646884441,-0.0115431584,0.5832077861,0.2118635923,-0.0714421794,0.0664260015,0.4306585193,-0.1481029093,0.3573056757,-0.1868436038,0.7492056489,-0.0570857488,0.2151276916,0.1019691601,-0.3711522222,0.1966659725,-0.0153630348,0.0793213546,-0.5118100643,-0.3569800556,-0.1609403491,-0.2538842559,0.1990831494,0.0451388434,-0.221278891,0.3177910745,0.0354344957,0.3949761987,0.3592294157,0.3050839007,-0.0906670541,0.3530260026,-0.2177924514,-0.0141585823,-0.0614609346,0.0901510194,0.1194941476,0.0920847505,-0.2436117381,-0.4295100272,-0.305162847,0.0007013087,-0.2089609057,-0.0932991058,0.4825862348,-0.564635396,0.1532411426,0.0374579653,-0.1119042113,0.0024960313,-0.2098170668,-0.0293328371,-0.1158581823,0.1289699525,0.1722138822,0.0898154378,0.4245472848,-0.0003973609,-0.362603873,0.304105103,-0.0489400961,0.0594808012,-0.003578915,-0.0241917279,0.2139300406,-0.3155358732,0.132243067,-0.5739024878,0.1775269806,-0.1474528909,0.0127399089,-0.1176820323,-0.0013163961,0.0483027846,0.1547810137,-0.283978492,-0.2273022979,0.2939545214,-0.0005660221,0.2803093195,0.4978891015,-0.115971446,-0.0311297104,-0.0739877224,0.0684594959,0.3330630958,-0.3416400552,-0.0628984123,-0.0199738983,0.0104450407,0.1303909272,0.2286181152,0.2601722181,0.2260191292,0.0391300023,-0.5462092161,-0.0374827012,0.1696475744,-0.1299935579,0.1054470316,0.1329135299,0.2669435143,0.2379645258,-0.1843246967,-0.1761202514,0.0102633117,-0.0304712392,0.1358926743,-0.2163457274,0.158505246,0.138336271,0.0027873158,0.1136001423,-0.2447290123,-0.1037489846,-0.0822046772,-0.1622231752,0.2020552754,0.0761264116,0.1932974756,0.1445910037,0.0521461889,0.0121984174,-0.031037217,0.0573093072,0.117708832,-0.2250270993,0.3234240115,0.4698260427,-0.2181081176,-0.3864577413,-0.048631914,-0.2083722055,-0.0919483081,0.1622691602,0.3923228979,-0.1639600098,0.1349995434,-0.3085337281,0.0645157397,-0.2191734761,-0.084187828,0.1953693926,-0.4075606465,-0.2250798941,-0.05221726,0.4547858238,0.2435606569,-0.2076168656,-0.0345069468,0.1239121035,-0.0041082399,-0.4457924366,0.080068253,0.2681975961,0.166590631,0.0092392126,0.4907566011,0.4225642681,-0.3052988052,-0.0838325247,0.1534824818,0.5127706528,0.2449548542,0.1527925283,0.4591118395,-0.0687162206,0.1594157666,0.5113036633,0.1964520961,-0.1306294948,0.3847328126,-0.1721257567,0.1426737756,0.0957871377,-0.1894542873,0.0161806419,-0.1555673629,0.2666930258,0.0196503624,0.3913426101,-0.2815415561,0.1803604662,0.3507416546,0.1539365053,-0.2007782906,-0.1109496057,0.0122991148,-0.0259589273,-0.0553431511,0.127328977,-0.2358577847,-0.2330396473,-0.0735607669,-0.1582722217,-0.2579681575,0.7183961868,0.2728726566,-0.1813760847,-0.3079471588,0.1664167345,0.1576441526,0.0288805664,-0.2608737051,0.3783064485,0.0894762874,-0.0856343582,0.232608512,0.3378421664,0.8134999275,0.2410013527,-0.1378462315,-0.2350551188,-0.1959068179,0.036293149,0.3027204871,-0.0511013158,-0.0022525107,0.0649085566,0.2457795441,0.0829157159,0.1131313965,-0.0042747431,0.080337815,-0.0697649494,-0.4125570953,0.2890695632,0.1622858495,-0.1878182888,0.0006666749,0.294282347,-0.2899018824,-0.0241305344,0.2330348194,0.567232132,-0.1211466789,-0.1184538007,-0.0200917888,0.2552170753,0.2999455333,0.5425080657,0.2283957005,-0.2810824215,-0.1846067607,-0.491732955,0.3696128726,-0.188915655,-0.1681735516,0.2723663449,0.0870510712,0.0668255761,0.1543128341,0.358499825,0.3908868134,-0.3710202575,0.133395642,-0.3752774596,-0.2423381209,-0.0293359756,-0.1051227599,-0.0941129327,-0.0155493794,0.2134482265,-0.6555412412,-0.0000092105,-0.1405493617,-0.242702499,0.1772046834,-0.0279113445,0.4945253432,-0.039074365,0.6881918907,0.0928875506,-0.1293089986,-0.1993017346,-0.0419701189,-0.4117202461,0.5732808709,-0.2209887207,0.2203930467,-0.1318157911,-0.3068338335,-0.1680376679,0.2684181929,0.0365713872,0.0693127438,0.074137643,0.1095281318,0.1249653995,-0.0653188229,-0.0048772795,-0.0606577806,-0.0253131129,0.2815610468,-0.5856456757,-0.6785210967,0.5218515396,-0.2936740816,-0.2915141881,-0.4276632667,0.2554471791,-0.0313607417,-0.0421098396,-0.5466675758,0.0804826915,0.4648813307,0.0448707342,-0.2022099197,-0.1116632447,-0.022803925,0.3597259521,-0.1047497168,0.4848444164,-0.3160455525,-0.1668283343,-0.2473701835,-0.2702198327]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"Reopening this. Although the `test_dataset_common.py` script works fine now.\r\n\r\nHas this got something to do with my pull request not passing `ci\/circleci: run_dataset_script_tests_pyarrow` tests?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/2873","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":25,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nReopening this. Although the `test_dataset_common.py` script works fine now.\r\n\r\nHas this got something to do with my pull request not passing `ci\/circleci: run_dataset_script_tests_pyarrow` tests?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/2873","embeddings":[-0.3413378894,0.2006237954,0.049372945,0.0571891814,0.1107941419,0.0854446366,0.2353403866,0.2816726863,-0.1760852039,0.1386214346,0.438138634,0.3304321766,-0.1294845939,0.297735095,-0.1947026998,0.0704638883,0.0133854281,0.1427540034,0.3433145285,0.0562438518,-0.1254506111,0.239258036,-0.1946723908,0.1283952147,-0.1329981238,0.0372122005,-0.1708889008,0.0408576243,-0.1387818605,-0.4680930376,0.4385558963,0.0902260914,0.1819554418,0.6217026114,-0.0001299516,0.1380508989,0.4102708101,-0.0885429606,-0.2731285393,-0.5470474362,0.0774918571,-0.0126978597,0.3988084793,-0.1040790603,0.0965566337,-0.3847824633,-0.0448813178,0.1092322022,-0.067902863,0.3774314225,0.0892633647,0.2561304271,0.0892846957,0.0130179506,0.2664367259,0.3202229738,-0.1790551394,0.4346646965,0.163472265,-0.1368527114,0.18223162,-0.0818933249,-0.0010008782,0.0381129086,0.058766406,0.0650137067,-0.0054221773,-0.2756882608,0.2095608413,-0.0046765376,0.4738269448,-0.5655617118,-0.4999672771,0.0325702988,0.1310693324,-0.2759967446,0.1989559233,0.1092876121,-0.2475142032,0.1120673195,-0.1768936068,-0.1210495085,-0.0797204599,-0.0597053356,-0.452073127,0.1955433339,0.1415925175,0.1933697611,0.0005672796,-0.0080453893,0.0825758576,-0.1077340245,0.1832358539,-0.0202006698,-0.2383793592,-0.0761251375,0.0723483413,0.3277744949,0.2713033557,0.4195830226,0.0403551459,-0.1306673586,0.174266085,0.0946188346,0.0676531792,0.2015321553,-0.0325329341,0.1100755706,0.26883623,0.0022651241,0.0251369309,-0.0340531245,-0.0820883214,-0.118339479,0.2201674134,0.0307946801,0.8329048157,-0.3100616932,-0.264413327,0.0647196248,-0.4131411612,-0.1456345022,-0.1949632913,0.2120336592,0.1063020974,0.3991495371,-0.054112941,0.3935264349,-0.0928063095,-0.0335153863,-0.1254985034,0.0960216671,0.0741094276,-0.0745831132,0.2243506163,-0.3418166041,-0.0214521512,0.2785852551,0.2317495495,0.1688358784,0.0847869664,0.1707445085,-0.0158238746,0.5882755518,-0.1170846,0.3030574322,0.1808865517,-0.4662673175,-0.3240087628,0.2942477167,-0.3417190313,-0.3119837642,-0.1544470787,0.0395626612,-0.3309824169,-0.2279408276,-0.2230107188,-0.199565202,0.230595544,-0.09554369,0.0377825759,-0.6431552768,0.1101010367,-0.2451968342,0.2660878897,0.2522025704,-0.4532644153,0.0551288687,-0.0058161258,0.0077627655,0.1686531156,0.108184278,-0.0699362084,-0.1368115842,-0.0259974599,0.0800719559,0.4520925283,-0.4171760976,-0.3087326586,0.2504969537,-0.0502103381,0.0525274239,0.2544584572,-0.4099239409,0.0670625865,0.1091564149,-0.1023410782,-0.0559723116,-0.013213655,0.1008469164,-0.1170869917,-0.0518476814,0.4205104113,0.2592461705,0.0945970491,-0.1817267984,-0.0615676045,-0.3520296216,0.2080374062,-0.058173541,-0.00003425,-0.0472278409,0.4154104888,-0.3017407358,0.0376632102,-0.0054553566,-0.2810015082,0.2353983223,0.1397245526,0.1832517534,-0.4007819295,-0.2460963428,-0.2957423627,0.1120926812,-0.1815855503,-0.0250835977,-0.0734229907,0.3676602542,0.0943571031,0.0908512399,-0.2459866107,0.1231615022,-0.0549701601,0.121638678,-0.0157055464,0.36100775,-0.39767465,-0.3114204407,0.1486184299,0.0152983777,0.1552105695,-0.0953354239,0.0492075197,0.1977923065,0.018292509,-0.0615071394,0.0249138251,0.2371177673,0.0974489674,-0.1439389586,-0.1320542544,0.2738496065,-0.0527514368,0.2288922668,0.0089537343,0.4483506978,0.433740288,0.4135373831,-0.0052913735,0.133137241,-0.0789800286,0.0464649163,-0.1026584506,0.0235418081,0.0623203032,-0.1579686403,0.2385576516,-0.0083107622,-0.4083735943,0.1789080948,0.4755972028,0.0045280191,0.0512079373,0.1263181567,-0.0464047939,0.0135155227,0.1427060068,0.1196604073,0.3690716922,0.1315706223,-0.1269986928,0.1280188709,-0.3089667261,-0.000240603,0.0881366134,0.0855403543,-0.0177239105,0.1066989228,0.2934764922,0.0094751427,-0.1714618802,-0.5069288611,0.1386284977,0.3523923159,-0.5544535518,0.2957420945,-0.2599960566,0.2412185669,-0.0563951358,-0.099517554,-0.3512249887,-0.3400330842,-0.0426338091,0.1322477013,0.1855418086,0.1987326741,-0.3156895339,0.085814856,0.0278856121,-0.5474075079,-0.3988513947,-0.2250475734,-0.3432812393,-0.0082413657,0.335031271,-0.3440192044,0.1043744832,-0.3536006212,-0.0615599714,-0.3151578307,-0.3276818395,0.0971032903,-0.3790067136,0.3400279582,0.264110893,0.1390078962,-0.0936909392,-0.0775850788,0.2825623751,-0.3451506793,-0.3498209715,0.0558034815,-0.0393706858,-0.081658788,-0.2297518551,-0.4395770431,-0.1756117046,-0.1256152838,0.1797543317,0.1345760971,0.0894499645,0.1602129638,0.2617307603,-0.1119602174,0.0015942445,-0.080045335,0.0416262671,-0.1853195578,0.311385572,-0.1229889318,-0.2905154526,0.0768143982,0.0901876315,0.2567841113,0.0905639157,-0.3195661008,-0.2088564485,0.0071465177,0.5721033216,-0.1528172642,0.0589576438,0.3950414658,0.1801723242,0.033580374,0.0303174295,-0.2007779479,-0.0994261503,0.1576837599,0.0561942831,0.1956194937,0.2270442694,-0.0645815507,0.7777855992,-0.0912658498,-0.247752741,0.3013986051,-0.3482937813,0.4125493765,0.0924599543,-0.3143632114,-0.1148337424,-0.0043492815,-0.0983640924,0.1719359308,-0.1672710329,-0.196849227,-0.131608218,-0.2043840885,-0.1171986982,-0.0901859552,-0.0458139852,-0.0148275411,0.2799755633,-0.0396879055,0.1678755432,-0.1324225217,-0.0496874787,0.1742268801,0.442769289,-0.0196975451,0.0237190947,-0.2624483407,-0.2940678298,-0.2959463596,0.3724193871,0.061157085,0.4008819461,-0.026681086,-0.1610317528,0.1031048149,-0.3025604188,0.6163403392,-0.2401427329,0.0806937441,0.3421912193,-0.2250520885,-0.4068379104,-0.320012629,-0.4515895844,0.1839895546,-0.0279769711,0.2594494224,-0.2801221311,-0.3214555085,0.2671273351,0.1050608456,0.0664297938,0.0288712382,-0.5246195793,-0.4429920614,-0.2986671627,0.2140108347,0.0544981584,0.2545920312,-0.3563589156,0.4444724619,-0.0712299645,-0.2317398638,0.2484552115,0.2951732576,0.2508275807,-0.1330598593,-0.0640557408,-0.1815859675,0.0097758155,0.477694422,0.5681917667,0.3777179122,-0.2568677068,0.3769102693,0.2031817287,0.5758686662,0.2257894129,-0.4059256017,-0.0527210794,-0.1149656773,-0.0379138701,-0.080989413,0.1220759079,0.5143463016,-0.2152787,-0.3515267074,-0.0673798025,0.3762578964,0.290491879,-0.0491693653,0.1210093647,0.4069837034,-0.1331271678,0.3132099211,-0.1312802881,0.8649080396,0.057380721,0.256983012,0.1712691486,-0.322670877,0.0820893273,-0.003961972,0.0175484009,-0.4571155906,-0.1686368138,-0.1796827912,-0.2954652309,0.2304840833,-0.0226178914,-0.1865593791,0.229722783,0.0601244867,0.4995066226,0.4300904572,0.3966604471,-0.2316233218,0.3145558834,-0.2194425762,-0.0172984209,0.0293021239,0.2286307663,0.124477379,-0.002792815,-0.2746600807,-0.4961460829,-0.37046206,-0.0256939139,-0.2890318036,-0.1104236618,0.3295895755,-0.7482529283,0.2486482412,0.0339220054,0.1098260209,0.0075352951,-0.3014063835,0.0366480127,-0.1004758477,0.1164056063,0.2305586487,-0.0126885399,0.3798055649,0.0022485489,-0.369571954,0.248747468,-0.077589035,-0.1284065992,-0.10458038,0.0652091727,-0.0416565202,-0.3378791511,0.1743742973,-0.5917771459,0.1822234839,-0.0884274393,-0.0206442904,-0.1191925257,0.0958006158,0.0254339762,0.2053674906,-0.2256229073,-0.1250716597,0.2743127644,-0.0698661953,0.1688301414,0.5413680077,-0.0138974981,0.0417850837,-0.0379354209,0.033793889,0.2881394327,-0.4225570858,0.0193725899,-0.0664407313,0.0207807571,-0.0079836641,0.509545207,0.0708176792,0.2001803219,-0.0391862728,-0.4795256853,-0.0093072541,0.2478470504,-0.0300485119,0.1569505185,0.1399680674,0.2862483859,0.2056425959,-0.405495137,-0.1714913249,-0.0400294513,-0.000537709,0.0970597789,-0.1244992316,0.0023443503,0.171981141,-0.0211473666,0.024566818,-0.1681083292,-0.0766649842,-0.0485599563,-0.1229086444,0.2095736861,0.0366380587,0.1580324769,0.1997240484,0.0126829194,0.0098167099,-0.0131300623,0.1601453722,0.0829429477,-0.1392778456,0.093243286,0.1914658844,-0.1461140364,-0.3684512675,0.0260951091,-0.1601081192,0.0838330016,0.1734333187,0.3831002414,-0.1243301556,0.1700462848,-0.1644287854,0.136464268,-0.1596181542,0.013914519,0.2238865793,-0.3504797816,-0.2246349007,-0.0711626559,0.5016793013,0.2746834755,-0.1880669147,0.0494724512,0.0735972151,-0.0657561943,-0.3324206769,0.0811484829,0.3131485581,0.1644701362,-0.0142199947,0.5696216226,0.3624671996,-0.2893332541,0.1590238959,0.1196841523,0.7548341155,0.2659273148,0.0646335185,0.1735058725,-0.1727526039,0.1701314747,0.4858205616,0.0595215857,-0.0590773523,0.1139845327,-0.1660731435,0.1806773692,0.0633857921,-0.1396620274,0.001239076,-0.1661031544,0.3387806416,-0.264516741,0.3434379101,-0.2066052258,0.1852727979,0.4347059727,0.0754559636,-0.1739108562,-0.2398148179,0.1066153198,-0.0229048841,-0.0884501562,-0.1149521023,-0.2485224754,-0.1854368746,-0.0515635498,-0.1260222495,-0.1242799088,0.6650737524,0.1825416833,-0.1802328378,-0.2559813559,-0.0092405602,0.0285580959,0.0777159259,-0.263761878,0.4144214392,0.0703170374,-0.1295516491,0.0619428381,0.3388366401,0.9245889783,0.1384816468,-0.119341433,-0.3575937152,-0.2355310768,0.0099260984,0.138694793,0.0076980577,-0.0125988983,0.1495130807,0.1344967633,0.0862469524,0.071466282,-0.0586302131,0.1186977923,-0.1172768399,-0.4760383666,0.3648820519,-0.0168354753,-0.1769959629,-0.0490625612,0.1928962469,-0.2198527902,0.0186762493,0.1510175467,0.4744874835,-0.1413980275,-0.2653834522,-0.0160904955,0.1495954692,0.3990335464,0.4798596203,0.2936071754,-0.2884002924,-0.1620029062,-0.630964756,0.2499451637,-0.1089177728,-0.0733902678,0.1761762798,0.1923172176,0.1261360198,0.1565837115,0.2528316379,0.1572821587,-0.2526715398,0.0740477443,-0.2817492485,-0.1842305511,-0.1125999242,-0.1680740267,-0.0837305263,-0.0515798405,0.1233399138,-0.6046460867,-0.0450350232,-0.2213482559,-0.3686109781,0.1300189048,-0.2846067846,0.4359494448,-0.0453370288,0.7685061693,0.1256030947,-0.1501159221,-0.3288882077,-0.1370660514,-0.285241276,0.5235645771,-0.3340819478,0.2942072451,-0.0719057769,-0.380186826,-0.2367647588,0.2125569731,0.1473571807,0.015084289,0.139054656,0.18910411,-0.0995963514,-0.0554136485,-0.0435973816,-0.030801123,-0.0259375852,0.2249903977,-0.5183079839,-0.6421110034,0.5877329111,-0.3876187503,-0.2811570168,-0.3539839387,0.2209096998,-0.1894461215,-0.1547939181,-0.4908481538,0.021292137,0.3034403026,0.1360800862,-0.2562732697,0.0768880621,0.0237822514,0.290207088,-0.0882469937,0.499348551,-0.3625539243,-0.0745926052,-0.1757297665,-0.1907047182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"Hi @bwang482,\r\n\r\nIf you click on `Details` (on the right of your non passing CI test names: `ci\/circleci: run_dataset_script_tests_pyarrow`), you can have more information about the non-passing tests.\r\n\r\nFor example, for [\"ci\/circleci: run_dataset_script_tests_pyarrow_1\" details](https:\/\/circleci.com\/gh\/huggingface\/datasets\/46324?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link), you can see that the only non-passing test has to do with the dataset card (missing information in the `README.md` file): `test_changed_dataset_card`\r\n```\r\n=========================== short test summary info ============================\r\nFAILED tests\/test_dataset_cards.py::test_changed_dataset_card[swedish_medical_ner]\r\n= 1 failed, 3214 passed, 2874 skipped, 2 xfailed, 1 xpassed, 15 warnings in 175.59s (0:02:55) =\r\n```\r\n\r\nTherefore, your PR non-passing test has nothing to do with this issue.","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":95,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nHi @bwang482,\r\n\r\nIf you click on `Details` (on the right of your non passing CI test names: `ci\/circleci: run_dataset_script_tests_pyarrow`), you can have more information about the non-passing tests.\r\n\r\nFor example, for [\"ci\/circleci: run_dataset_script_tests_pyarrow_1\" details](https:\/\/circleci.com\/gh\/huggingface\/datasets\/46324?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link), you can see that the only non-passing test has to do with the dataset card (missing information in the `README.md` file): `test_changed_dataset_card`\r\n```\r\n=========================== short test summary info ============================\r\nFAILED tests\/test_dataset_cards.py::test_changed_dataset_card[swedish_medical_ner]\r\n= 1 failed, 3214 passed, 2874 skipped, 2 xfailed, 1 xpassed, 15 warnings in 175.59s (0:02:55) =\r\n```\r\n\r\nTherefore, your PR non-passing test has nothing to do with this issue.","embeddings":[-0.4502635896,0.1362894326,0.0727608278,-0.0077378419,0.210440293,0.1455685943,0.1347785145,0.3891866505,-0.2612731457,0.1966487318,0.2423591018,0.4409922659,-0.174534291,0.1955467612,-0.2465572804,0.0201375429,-0.0359608233,0.1991022974,0.1349838227,0.0418360345,-0.1486246139,0.1625906974,-0.2162213922,0.2647956908,-0.1537034959,-0.0562001877,-0.0889388993,-0.0443796068,-0.1304095089,-0.543780148,0.374103874,-0.0596004017,0.1746742874,0.5324847698,-0.0001234938,0.1340290904,0.3579029739,-0.099743627,-0.3201577663,-0.4985058904,0.0303555876,0.0250716098,0.3309290409,-0.1497163624,0.0457028411,-0.5905317068,-0.0560617782,0.011204035,-0.0264583249,0.4921305478,0.1557275653,0.1648605168,0.0755554661,-0.137879923,0.210254997,0.089855887,-0.0760812536,0.4706763327,0.1216390133,-0.0717137381,0.1600072682,-0.1023208573,-0.0217084177,-0.0166188478,-0.0197366606,0.0890951231,0.2701399326,-0.2540642619,0.2758241594,0.0213556439,0.6139847636,-0.6099302769,-0.402813077,0.0584969297,0.1510484666,-0.1977946013,0.2043727338,0.1740199775,-0.2345986068,0.1370747983,-0.1940619051,-0.0381158032,-0.0790201724,-0.1012421995,-0.3796566129,0.2510077655,0.1984764487,0.1712024063,-0.0156713389,-0.1160794795,0.0763115138,-0.1649231166,0.170402199,0.0014240057,-0.2107874453,-0.129683286,-0.0683172196,0.3605410159,0.2446825057,0.2437425703,0.0395206213,-0.2193709165,0.2899466455,0.1493432522,-0.0002065366,0.060830459,0.0282580554,0.1525740027,0.1156841591,-0.0743039697,0.0116414698,0.0125757102,-0.0356712267,-0.1347546428,0.2806890607,-0.0175586436,0.747718215,-0.386092782,-0.3294264078,0.1127082333,-0.3191033304,-0.1121121868,-0.1496730894,0.3698425591,0.0801952556,0.3064732254,-0.2025549859,0.3027328849,-0.148769632,-0.1783850491,-0.0953058898,0.0878898203,-0.0219249018,-0.1331822574,0.1687428951,-0.2979690433,-0.0465056896,0.3469100296,0.1025219187,0.1879935563,0.0652616769,-0.0026953628,-0.0544139817,0.4611659646,-0.2218509614,0.2516638935,0.1780836433,-0.5087434649,-0.3634715378,0.2743989229,-0.2164251655,-0.3534741104,-0.2432028204,0.151281327,-0.0920187533,-0.2586004138,-0.0489280932,-0.0613507144,0.1802086085,-0.1575635076,-0.0042943428,-0.6692387462,0.1291497648,-0.3172443509,0.2979747355,0.2355631441,-0.5403274298,0.0511988066,0.1171005592,-0.1538690478,0.2208563685,-0.0480547622,0.0285186488,-0.1684476435,0.008382176,0.0004034062,0.3619309366,-0.3157839477,-0.2808323205,0.2037823498,0.1373576522,-0.0054161474,0.258110404,-0.3559542,0.0676776916,0.1332530528,-0.2449017912,0.0707520768,-0.0634788871,0.0394995026,-0.1724347621,0.0111094378,0.4310903251,0.2942523658,0.0776945651,-0.1869189143,0.0541623756,-0.2536583543,0.1724235415,-0.153491497,0.0145558426,-0.095210731,0.4228506982,-0.3071483076,0.0091015995,-0.0820851251,-0.3720258772,0.214207381,0.1426869631,0.0427836291,-0.3331008852,-0.2545279264,-0.3418053687,0.0929807276,-0.1178567037,-0.0309027005,0.0290229879,0.3834606111,0.0836455449,0.0666082948,-0.1785346568,0.1259467155,0.0109810112,0.1843898743,-0.021914199,0.4052859545,-0.3848957419,-0.3287788928,0.1164525375,0.0147691462,0.2110700905,-0.0747004673,-0.0244301055,0.3569452167,0.0668140128,0.0324969031,0.0375453681,0.2812748551,0.0095077008,-0.2744845748,-0.2255578786,0.2845185697,-0.0563144833,0.335890919,-0.0477804728,0.4245945215,0.4690216482,0.2440936267,0.0434890911,0.1207157895,-0.0876624584,0.0307110269,0.0175956078,-0.1027696207,0.1239546984,-0.0975145176,0.2429556996,0.0073814266,-0.4534864426,0.0181120038,0.3326230049,-0.1036967561,0.079722181,0.1088910624,-0.1348168701,0.0713322908,0.1310128421,0.1286019683,0.2671003044,0.1545218527,-0.0664283633,0.1459353566,-0.3957768083,0.0493688993,0.1474568546,0.046951104,0.0843830109,0.1178493723,0.2030754983,0.0692258701,-0.1964526772,-0.4386540353,0.1690077484,0.3975126445,-0.4556099176,0.2525030673,-0.2593023777,0.2879627645,-0.1762989759,-0.1013129577,-0.2812756002,-0.2871729434,0.0298087746,0.1613865793,0.1280528903,0.2844814956,-0.3949874341,0.1308537275,0.0604371689,-0.5389617682,-0.3587738574,-0.1788505465,-0.4358465075,0.0564847142,0.2930062115,-0.2936736047,0.1870015562,-0.3385720551,-0.1041571423,-0.2780030966,-0.2546269596,0.1000234187,-0.2732539177,0.4224722683,0.2331346571,0.0653007999,-0.0509411618,-0.1394465566,0.2985119224,-0.5006918907,-0.2382101864,0.1075783893,-0.0609142967,-0.0281601492,-0.0992578343,-0.6026321054,-0.2700941563,-0.1767314076,-0.0738294199,0.1012948826,0.1171792299,0.0153186703,0.1794993281,-0.0414384827,-0.1064003855,0.040731471,0.0227799695,-0.0935673937,0.2638745904,-0.1412262917,-0.247062102,0.0030329677,0.0593990013,0.3075082004,0.0263136849,-0.3157670796,-0.1785085499,0.0870808735,0.5713326931,-0.213754639,-0.0088340752,0.2017015219,0.3253875971,-0.0461313352,0.0438692309,-0.1561382264,-0.2257906049,0.0831051841,0.1174168065,0.1173686087,0.2478049397,-0.0001471604,0.7000100017,0.0901702568,-0.2647122443,0.2456341684,-0.290442735,0.3646113276,0.0215920843,-0.2104302496,-0.0293033496,-0.0032851116,-0.0899685696,0.2567013502,-0.1834515184,-0.2540484667,-0.0747877732,-0.1884061694,-0.2296490818,-0.0697398409,-0.0385251641,-0.0507598296,0.2692644894,-0.0709148198,0.1493453532,-0.0700604767,-0.1060432568,0.2032533139,0.3706054389,-0.2543409765,0.0308764242,-0.2697114348,-0.200737372,-0.2348984778,0.4246786833,-0.0129908873,0.4772621691,-0.016963236,-0.1294088811,0.0974079669,-0.2080836296,0.4789678752,-0.2268110663,0.0811510012,0.2893922925,-0.170969218,-0.44634372,-0.3620918989,-0.4987749159,0.0781412125,-0.0376700051,0.2260660231,-0.2343872935,-0.2319649607,0.2269615382,0.1185379326,0.0725107491,0.0037124723,-0.4310705662,-0.2614602745,-0.2430775613,0.3034367263,-0.1024958342,0.3525213897,-0.3325912952,0.4075814784,-0.0961001888,-0.1727211326,0.2449412048,0.3544303477,0.2249419391,-0.2119112909,0.0260769092,-0.1250227988,0.2323424816,0.5753104091,0.5339670777,0.2321098298,-0.2221550047,0.1890183538,0.2378938645,0.6089592576,0.0379683226,-0.4350992441,-0.1311749667,-0.2159736902,0.0436225496,-0.0090597766,0.1054687575,0.5369826555,-0.1258685887,-0.5352920294,0.0339511372,0.47798118,0.2792463005,-0.136687398,0.0759350955,0.3156251609,-0.1674428582,0.4376081526,-0.1025658101,0.8173331022,-0.0009261135,0.1719650477,0.138971284,-0.3020308912,0.1291724443,0.0200279076,0.0107886679,-0.4951123297,-0.3523802161,-0.1284809411,-0.2606924474,0.2495348901,0.0580628514,-0.343375206,0.3005758524,0.0194640867,0.4572975338,0.4209087789,0.2916842699,-0.1678933203,0.2989626825,-0.2507620752,0.0380312651,0.0244168658,0.0722335279,0.1154952571,0.0764666349,-0.141377911,-0.4948067069,-0.2790382206,0.0151465517,-0.2930012643,-0.1004594862,0.3964217007,-0.5733570457,0.2118080556,0.0720924959,0.04818381,0.0640392303,-0.2333733588,0.0129593583,-0.1457787901,0.1276945472,0.183948487,0.0475329347,0.4533499479,-0.0087215574,-0.3378638029,0.3581652641,-0.0813947842,-0.0119384425,0.0148284594,-0.1256122291,0.1773061901,-0.3368036747,0.2464153767,-0.628660202,0.1099734455,-0.1781975776,0.045037251,-0.1984385252,0.0036137204,0.1091617271,0.2272614092,-0.3318321109,-0.2118319571,0.2083907127,0.0059665497,0.2285706252,0.5235672593,-0.0887126327,-0.04807809,-0.1036171764,0.0581538342,0.3272997141,-0.3375204802,-0.0491202101,-0.0297179222,0.065609552,0.1523433775,0.3768545389,0.2776260078,0.2131869495,-0.0744468719,-0.5255508423,-0.0729655996,0.2375106514,-0.0712140799,0.1975880712,0.1174603403,0.3237075806,0.2238592505,-0.1134935394,-0.2397029251,-0.03295701,-0.1323354989,0.1919335723,-0.0868279263,0.0964449942,0.051700104,0.0634484664,0.0887569487,-0.192830056,-0.1422792673,-0.1109476462,-0.1475551575,0.1835369468,0.1018511131,0.1381859332,0.1264923364,0.0560475253,0.0490148067,-0.0467185937,0.1385231018,0.0393861048,-0.1792175919,0.2147742212,0.34734115,-0.1672769934,-0.3585488498,-0.0370983481,-0.2041278034,-0.0396423712,0.1470280886,0.3860643506,-0.2622255683,0.0606436282,-0.183980599,0.0891807824,-0.0386753567,-0.058118239,0.1145231798,-0.3145363331,-0.2465545535,0.0138950851,0.4490870833,0.2860864401,-0.2015555203,-0.030049989,0.235149473,0.0334136561,-0.4120021164,0.0085355006,0.2943555713,0.1643490195,0.0969960317,0.4992057681,0.3679380715,-0.337335825,0.0763207898,0.054108277,0.4890707135,0.2271460444,0.1624105126,0.4281882644,-0.120604977,0.148973614,0.500107348,0.0717378333,-0.068675071,0.3365833759,-0.226596266,0.2901752591,0.0518489964,-0.0947274566,0.0120786726,-0.1402914524,0.33037889,-0.0836483538,0.3031802773,-0.1854189783,0.2107686251,0.3404168189,0.0991525203,-0.3103975058,-0.264610827,0.0073891794,-0.0769663379,-0.1239804476,-0.0777561739,-0.248784706,-0.3113349378,-0.0424013659,-0.1833255887,-0.2136457413,0.6925194263,0.2094757855,-0.1458517164,-0.2887672782,0.099605754,0.1255612969,0.0661409795,-0.2412374616,0.4364032447,-0.0067151501,-0.0461819135,0.1553471386,0.2981884778,0.8132764697,0.2181598842,-0.1735399514,-0.3258236647,-0.2322149277,0.0645576119,0.2646583021,0.0630190447,-0.0373362675,0.0212638974,0.2920491993,0.1168117151,0.0277587902,-0.0784622729,0.0650030896,-0.1590969265,-0.4842327237,0.2990456522,0.1847798973,-0.1018589437,0.0164223909,0.1767463088,-0.2542063892,0.0268722139,0.2197711021,0.4025389552,-0.0997011513,-0.1280664951,0.0130578224,0.2641711831,0.3753708899,0.4660414457,0.3042026758,-0.3263893723,-0.2322641909,-0.5930103064,0.3081550002,-0.1118557751,-0.0641729385,0.2311535031,0.0703425854,0.1258249283,0.2151816934,0.30558604,0.3428937793,-0.2636965513,0.1026008427,-0.4001891017,-0.1769956946,-0.0904512405,-0.1819176972,-0.0943459645,-0.1717551947,0.1907263249,-0.5734676123,0.0147785563,-0.1284413487,-0.3459657729,0.200353846,-0.19253923,0.4588808715,-0.0203613099,0.7599465251,0.057064373,-0.1314642429,-0.1995707601,-0.15740785,-0.3382564485,0.508472383,-0.2153967023,0.2439021915,-0.0355879702,-0.2223462462,-0.1448097527,0.2334987372,0.1409814954,0.0157278813,0.0833625644,0.0926648304,0.0641425624,-0.0942074135,-0.0230031405,-0.0197342616,-0.0309021324,0.2972713113,-0.4680343568,-0.672316432,0.540030539,-0.338793844,-0.235279426,-0.4275918007,0.2575130761,-0.0969465226,-0.0870121717,-0.4807568491,0.1060673445,0.3818357885,0.0325497799,-0.2270982862,-0.0324551389,0.0326505639,0.3201504648,-0.1376316845,0.5239678621,-0.2877165973,-0.1514039785,-0.1682212055,-0.2710402012]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"Hi, @Chenfei-Kang.\r\n\r\nI'm sorry, but I'm not able to reproduce your bug:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"glue\", 'cola')\r\nds\r\n```\r\n```\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 8551\r\n })\r\n validation: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 1043\r\n })\r\n test: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 1063\r\n })\r\n})\r\n```\r\n\r\nCould you please give more details and environment info (platform, PyArrow version)?","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":66,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\nHi, @Chenfei-Kang.\r\n\r\nI'm sorry, but I'm not able to reproduce your bug:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"glue\", 'cola')\r\nds\r\n```\r\n```\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 8551\r\n })\r\n validation: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 1043\r\n })\r\n test: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 1063\r\n })\r\n})\r\n```\r\n\r\nCould you please give more details and environment info (platform, PyArrow version)?","embeddings":[-0.0594903305,-0.1199711189,0.0148641206,0.2438831329,0.4732634127,-0.0013416682,0.4414617419,0.1253976971,-0.0142953452,0.2675631642,-0.1799774915,0.4658890367,-0.1191558838,0.1054266095,0.0933514312,-0.1634460539,-0.1493485421,0.1789052486,-0.0950517431,0.0041558379,-0.2951413095,-0.1227453202,-0.2647020519,0.1601062417,-0.4545831382,-0.2473363578,-0.1237274557,-0.0289071202,-0.2947974801,-0.3072052002,0.3905930817,0.045531936,0.0802016407,0.573389709,-0.0001046333,0.085742943,0.5220624804,0.0795245767,-0.1019385606,-0.4897305667,-0.1364545226,-0.1668332815,0.2139172852,-0.393879205,-0.174881503,-0.0434178077,0.0156303495,0.1047198623,0.3135638833,0.4849516451,0.3089385629,0.4509330392,0.0464522541,-0.2673178315,0.2966395915,0.1515992731,-0.0532518923,0.3601016104,-0.0297630094,-0.1026376411,0.1846834719,0.1567065567,-0.2669480145,-0.0967591628,0.1903866827,0.0687865466,-0.0415391438,-0.290898174,-0.0751495287,0.4143870771,0.3387809098,-0.4593112171,-0.2788108885,0.1180645153,0.2122831494,-0.2495187223,0.2602464557,0.0856177285,-0.0128446911,0.1975728273,-0.0867947638,0.2796069086,-0.2793028951,0.3049078584,-0.164794296,0.1444952488,-0.0630867183,0.0345241651,0.2060508877,-0.1919636279,0.059791442,-0.1310273856,0.0356741138,0.1007023752,-0.243607074,-0.1369945854,0.1091515124,-0.0550249554,0.1305607557,-0.2993285656,0.1565279961,-0.0825746059,-0.0047007087,0.4082542658,0.183615461,0.3742778003,0.2353915423,0.146733135,0.2031041682,0.1250560284,-0.2914271057,0.0851671919,-0.1916304976,-0.0795564353,0.3350311518,0.141778037,0.4543803036,0.004213843,-0.4538616538,-0.0484698229,-0.4231696129,0.0795800835,0.2269208282,0.4578568935,-0.1490905881,0.111972034,0.0448685922,0.1539013535,-0.1707451195,-0.2831352949,-0.2869074643,0.4119275212,-0.3327414989,-0.2005894333,0.1725695282,0.0631759688,-0.0440938734,-0.0142761068,-0.124532178,0.0147927273,0.0312171597,-0.4864667356,-0.1069177911,0.1802367717,0.0366909429,-0.1434192806,0.2177432328,-0.4385626912,0.0038601621,0.2104463279,-0.1390264481,-0.2205248028,-0.0644018948,0.2921482325,-0.3297555149,-0.1617497504,-0.1220805645,0.1581521481,0.158894524,0.0152816763,0.0386043973,-0.0845265612,-0.0218608137,-0.3067058027,0.0268780459,0.3902682364,-0.3743338883,-0.2277735025,-0.1163482741,-0.0573342443,0.3309760094,-0.0552941114,-0.0139303515,0.2842520177,-0.0162361618,0.1270191222,0.6166725755,-0.3811469376,-0.3189837933,-0.0871518478,-0.0525565632,-0.0323508754,-0.0000088416,-0.1415092349,0.0952094793,0.1851929128,0.4305967093,0.1825711876,0.0108494731,-0.0068763741,-0.2708131671,-0.2111279964,0.1911941022,0.0895309374,0.1389753819,-0.0566433519,-0.0780088603,-0.24225308,0.0179314129,-0.1191949025,-0.1323612928,-0.0616726913,0.5806082487,-0.0895553157,-0.1220896021,-0.5331310034,-0.2384607494,0.1451272964,-0.0384060368,0.4400658906,-0.1647625715,-0.2132972479,-0.4087546468,0.0467493199,-0.0534322485,-0.0422442332,0.2887170017,0.0327202827,-0.162372008,0.1476718634,-0.093931675,0.0690900609,-0.1891066879,0.1402491182,0.1150690317,0.3248268962,-0.052656617,-0.4338489175,0.0160555448,0.2450485379,0.4278887212,0.0615985915,-0.0518304221,0.3103211224,0.0966809914,-0.2166442722,-0.332929194,-0.0375642255,0.1061588004,-0.3284879327,-0.1916384846,-0.02861006,0.1814015508,-0.0039089932,0.0155871715,0.5066564679,0.073046267,0.0751549006,-0.0834585428,0.1962543428,-0.0521300845,-0.1202241629,-0.1189233959,-0.0935665593,0.0853727981,-0.0516477935,0.1604620069,0.0839664191,-0.1486921012,-0.0939611942,0.6276284456,0.066133365,0.4066179395,-0.1278576404,-0.253084898,0.1469340771,0.0753541663,0.0935161412,0.3560036421,0.309184581,-0.1741830856,-0.012002632,-0.1699138284,0.0200339202,0.0936975554,-0.0300577246,0.3375006616,-0.0191667695,0.326565057,0.0025228276,-0.3129945993,0.1154984683,0.0383162536,0.1894647479,-0.3931485116,0.1806956828,-0.3155397177,-0.1539050788,0.0592735447,-0.1398235261,-0.0421994813,-0.2707495987,-0.085526906,0.1815255284,0.0589705072,0.264362663,0.1597475857,-0.0327072591,0.0847381577,-0.0655731633,-0.102235347,-0.2201068252,-0.1743910462,0.1428499222,0.1274092495,0.2078600228,0.2995347381,-0.1425538957,0.0480940603,0.0138055841,-0.3602250516,0.0857338384,-0.1215342283,0.4401930273,0.3425597847,0.0728836656,-0.2368693203,0.0173933227,0.3619660437,-0.3473314047,-0.108051084,0.3983776271,-0.1979978234,0.0502504855,-0.2161714584,-0.3248876333,-0.510535121,-0.318955183,0.1311067194,-0.0294710808,0.0872791484,0.3062438071,0.2281634808,0.450894624,-0.052571062,0.0395837203,-0.1739951223,-0.0077805598,0.3531797528,-0.0148417056,-0.3720351756,0.0079913763,-0.1651002616,0.1313911527,-0.0117581366,-0.1790904701,-0.5396494269,-0.1400610059,0.288399756,0.0229437444,-0.0957489237,0.3742546439,0.0539732315,-0.1510512084,-0.1974764913,-0.082263872,0.0317742862,0.3172571063,0.0219645612,-0.069488965,0.2908799648,0.0552209876,0.2811395824,-0.113323316,-0.2513748407,0.4198065102,-0.2915056348,0.2461425811,-0.0716580451,-0.4766027927,-0.0776230395,-0.0555009358,-0.0022872451,-0.0360607766,-0.2989491224,-0.2200709581,-0.297008574,-0.1219761595,-0.3595552146,-0.1541765779,0.0609147735,0.089118734,0.0558054894,0.1404543817,0.155065909,-0.3061159551,0.0088768518,0.0000187792,0.1075965613,0.0372078121,-0.2144530863,-0.2214943469,-0.2509330511,-0.2910080254,0.3232398033,0.0477950834,-0.0251004752,-0.1071845666,-0.1048598811,0.084771201,-0.054349944,0.6767696738,-0.1964523345,0.0710406229,0.4911931157,0.2105484456,-0.432089746,0.02492279,-0.4132964909,0.1125910804,0.1601796597,0.3240831196,-0.3951091766,-0.0542449914,0.2625240684,-0.0359851383,-0.0594216958,-0.269669354,-0.2357648313,-0.2557882369,-0.1077657789,0.1962948442,0.0353272893,0.4339828193,-0.2649331689,-0.1105708033,-0.2390617579,0.0428951569,0.1868531257,0.2545810938,0.2625078559,-0.1614429355,-0.1597974896,-0.1762635708,0.058802031,0.2780003548,0.2882107198,0.0445406772,-0.4137676656,0.0157187078,0.0581392795,0.2611902654,0.0464258008,-0.1333332211,-0.0307984203,-0.0403293222,0.0974305719,0.1435642242,0.2177045345,0.386084348,0.059595976,-0.3346404433,-0.2437654287,0.3455174267,0.1484420449,0.0505361073,0.0170092452,0.2781257927,-0.2117078453,0.2881139517,-0.0201545823,1.0111832619,0.0690524206,-0.0597921535,0.5816109776,-0.0952846482,0.5560985208,-0.0648017898,0.1861322969,-0.2962427735,-0.3566812873,-0.0365724601,-0.1678650081,0.2394370884,0.0646672547,-0.6077862382,0.2306772918,-0.2865700722,0.1413257867,0.037199948,-0.0313408002,-0.0069267428,-0.0241819527,-0.3856900334,0.2118214965,-0.1228692979,0.1014267579,-0.0845644027,-0.10637898,-0.1233802363,-0.1591433734,-0.1048097834,0.0431123562,-0.5134391785,-0.1531040519,0.0520422086,-0.1772312224,-0.0729138181,0.1872905791,0.0402178206,0.0775228292,-0.1871171445,0.1700199097,0.1999508739,0.0999664664,-0.0692627802,-0.0332883447,0.4308818877,-0.0916927382,-0.3986715674,0.1275028139,-0.0470685437,0.1355867386,0.3548524976,0.0182911213,-0.0207855981,-0.4515144229,-0.0984330326,-0.1435841471,0.2059473395,-0.3068613708,0.2612115443,-0.0072328709,-0.1372915208,0.0108396858,-0.0357730947,-0.018511625,-0.0398774929,0.3423734009,0.3180928528,0.4484286904,0.2465310693,0.2419533879,-0.1212224513,-0.3200915158,0.1532203853,0.230802089,-0.4224052429,0.2418849766,0.2260043472,-0.1214663088,-0.053419821,0.4723202586,0.0875411034,0.1689212471,-0.1257396638,-0.148903504,-0.2344330549,0.3241607547,0.1022827998,0.0040802173,0.1147460192,0.4286702871,0.2379003763,0.017730521,-0.4440496266,-0.0286853164,-0.4677805901,0.070526503,-0.0260700732,-0.2602921128,0.2004087865,-0.0038485706,0.2606965303,0.1246233732,-0.0757152289,-0.3621689975,-0.1629702449,0.0902142152,-0.002169311,0.1705976725,0.1876295209,0.0336485803,-0.230336979,-0.1281399429,0.0921760574,-0.055329401,-0.1467017829,0.21817936,0.2054448128,0.3228447139,0.0179912299,-0.0136706028,0.0152384248,-0.0924621522,-0.0348306708,-0.0050635072,0.1245716438,-0.0292172786,-0.0914055929,-0.0829483122,0.1921166331,0.1356510073,0.0203486923,-0.2578234971,-0.2741311491,-0.1145555899,0.2158735543,0.1211166307,-0.1575217694,0.1765080243,-0.0109323282,0.3513850868,-0.3763158917,-0.0102533428,0.2301227152,-0.0349603593,0.2269831598,0.0463015363,-0.2856045365,0.3148671389,-0.2031256407,0.090680182,0.2280460447,-0.0443477929,0.1081102192,0.2768344283,-0.1694982648,-0.0334678218,0.4123375118,0.070386678,0.1482765526,0.02117116,0.0128191616,0.1463374197,-0.0221673734,0.0858672485,0.3090984523,-0.3254109621,0.0323023833,0.2096980363,0.0278553236,-0.0797003284,-0.2609014511,0.516067028,-0.2635023892,-0.1445791423,-0.3344384432,0.1385084242,-0.200316295,0.0369850062,-0.2015442103,-0.1111779511,-0.1869933307,-0.3066197634,0.0259554833,-0.0909686759,0.2416601479,-0.0493581742,-0.0289359875,-0.2844852209,-0.0523675457,0.3596524894,0.2271602452,-0.20385921,0.1030545309,0.0247038733,0.1235098615,-0.1864320636,0.5136396885,0.4270206392,0.2857337296,-0.1543841809,-0.0936766863,0.0345495343,-0.3090969026,0.0706597194,0.0969653726,0.2919444442,0.2387932241,0.0644041747,0.2788921297,-0.283341378,0.1762279719,-0.0601931773,0.0752444565,-0.3707785904,0.3305317461,-0.1943280846,-0.1633215249,-0.1663211137,-0.0086910268,-0.4805711806,0.1700413376,0.3345673084,0.1505533755,0.18274571,-0.1898440421,0.1426260471,-0.2063407302,0.2242961377,-0.0433819778,0.3339892626,-0.279900521,0.1178943589,-0.4362550676,0.3936928809,-0.0448485799,-0.0489642657,0.0756321251,-0.0840160921,-0.1616720408,0.0349483863,-0.0744248405,0.2408670038,0.1278519928,0.0993816927,-0.6071122885,-0.199648127,-0.1119104177,-0.2307693958,0.1361706406,-0.2414505482,-0.047243461,-0.1864420474,0.1147017404,0.0427970104,-0.3028406501,0.1948673278,0.2954314947,0.6067336798,0.1576177031,0.4007611871,-0.013851488,0.0997695774,0.0633205548,-0.2700609863,-0.1527519971,0.1926423162,0.1350933015,0.3505522013,0.0757688209,0.232017532,-0.0688422769,0.0777027383,0.0588588454,-0.124382481,-0.1221504211,0.163699314,-0.1173157468,-0.2261742949,-0.255618602,0.0282795932,0.0752818286,0.0470450297,-0.0877219141,-0.2791985869,0.670065105,-0.3341488242,-0.2696869671,-0.0509191118,0.2001962513,0.2284435928,-0.2409944087,-0.6293930411,-0.0089962306,0.4204948246,0.0220112894,-0.2528333664,0.2136524618,-0.0641363934,0.1553049386,-0.0634708256,0.0211865623,-0.1120157316,-0.0563756153,0.047504697,-0.2333326787]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"> Hi, @Chenfei-Kang.\r\n> \r\n> I'm sorry, but I'm not able to reproduce your bug:\r\n> \r\n> ```python\r\n> from datasets import load_dataset\r\n> \r\n> ds = load_dataset(\"glue\", 'cola')\r\n> ds\r\n> ```\r\n> \r\n> ```\r\n> DatasetDict({\r\n> train: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 8551\r\n> })\r\n> validation: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 1043\r\n> })\r\n> test: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 1063\r\n> })\r\n> })\r\n> ```\r\n> \r\n> Could you please give more details and environment info (platform, PyArrow version)?\r\n\r\nSorry to reply you so late.\r\nplatform: pycharm 2021 + anaconda with python 3.7\r\nPyArrow version: 5.0.0\r\nhuggingface-hub: 0.0.16\r\ndatasets: 1.9.0\r\n","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":116,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\n> Hi, @Chenfei-Kang.\r\n> \r\n> I'm sorry, but I'm not able to reproduce your bug:\r\n> \r\n> ```python\r\n> from datasets import load_dataset\r\n> \r\n> ds = load_dataset(\"glue\", 'cola')\r\n> ds\r\n> ```\r\n> \r\n> ```\r\n> DatasetDict({\r\n> train: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 8551\r\n> })\r\n> validation: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 1043\r\n> })\r\n> test: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 1063\r\n> })\r\n> })\r\n> ```\r\n> \r\n> Could you please give more details and environment info (platform, PyArrow version)?\r\n\r\nSorry to reply you so late.\r\nplatform: pycharm 2021 + anaconda with python 3.7\r\nPyArrow version: 5.0.0\r\nhuggingface-hub: 0.0.16\r\ndatasets: 1.9.0\r\n","embeddings":[-0.0170860495,-0.3240526319,0.062121436,0.2103672326,0.3840116858,-0.0155181745,0.415794313,0.1785911471,0.1879396886,0.3116176724,-0.2562711835,0.2617366314,-0.0900785625,0.1555979401,0.134646982,-0.1783233583,-0.1780708879,0.2362426817,-0.1429384202,-0.0610416606,-0.307266444,0.016021125,-0.2765933871,0.1563270539,-0.3710355163,-0.2232295871,-0.0727495775,-0.0337860063,-0.3516702652,-0.3736093938,0.3502697647,-0.0149404071,0.1077515855,0.3936695755,-0.000109273,0.039748542,0.5531897545,0.0916725695,-0.1215192825,-0.486445725,0.0033005883,-0.153592959,0.2640893161,-0.2971876264,-0.2491431981,-0.1678692102,-0.0046053282,0.1546724737,0.3977606893,0.5068286061,0.2735918462,0.5179857016,0.1557451636,-0.3057215512,0.2476199716,0.1295629144,-0.0830335468,0.3645498157,0.0868025422,0.0324823782,0.2476591766,0.22567074,-0.2615945935,-0.1020329148,0.2481998503,0.0714492425,0.0565817356,-0.3359338641,-0.0874417275,0.3458363712,0.2176503092,-0.4307984114,-0.2740110457,-0.0156725049,0.1589335203,-0.2778740823,0.2152642608,0.1117749512,-0.0614302717,0.2610106766,-0.0477596298,0.2916436493,-0.2724869251,0.2264740765,0.0332289934,0.0998286605,-0.0974352807,0.0864828005,0.3184731901,-0.1978602111,-0.0712477863,-0.0744847059,0.0765407234,0.1997876465,-0.2625720799,-0.1569988281,0.0412150547,-0.0527565517,0.2478982806,-0.0623121895,0.010698121,-0.0784372389,0.0866332129,0.3945228755,0.1457946301,0.3695580363,0.2746926844,0.0596430153,0.147126019,0.2683903873,-0.1179920584,0.0550732911,-0.1812906712,-0.0505757257,0.2737024724,0.1010781154,0.4157111049,-0.0123857539,-0.3835292757,-0.0340619497,-0.249388665,0.2141660452,0.2170223594,0.4983271658,-0.1098527387,-0.006581971,0.1920767128,0.1782709658,-0.1211596727,-0.1304862946,-0.2357181609,0.34223032,-0.3215116858,-0.0617322512,0.2265961319,0.0308971405,0.0039272024,-0.0047287471,0.0682608932,-0.0418036729,-0.0186945554,-0.3479027152,-0.0821475908,0.2535481453,0.032342352,-0.0685248151,0.2264001071,-0.410043478,-0.0981175676,0.1309780478,-0.0746355355,-0.1795981675,-0.2304216176,0.2343917638,-0.2560875714,-0.0213031564,-0.0934348702,0.0212540068,0.1447509378,0.1013658345,0.1272418797,-0.0190788247,0.0178890768,-0.2180711776,0.1380720735,0.3654916883,-0.100190863,-0.3205661774,0.0321522094,-0.0025296691,0.1171867996,-0.0173707716,0.0019335701,0.2421703041,-0.0305478908,0.1668857634,0.4985628724,-0.62414819,-0.3442799449,-0.2747382522,-0.0806370825,-0.0161460973,0.0393288732,-0.2286882102,0.1914647818,0.2106938213,0.4453007579,0.1527808011,0.1522087157,0.0167638846,-0.2598265111,-0.2841210067,-0.1492304802,0.1216640398,0.0572228469,-0.0611823201,-0.1246782616,-0.3315826058,0.0439961776,-0.1464243233,-0.0947787762,-0.0285352468,0.4802513719,-0.0267939679,-0.0334615521,-0.5103312731,-0.3110117316,0.1444107145,-0.0827431381,0.4963111877,-0.2935872972,-0.2401018441,-0.3797984123,0.0216788147,-0.0260307416,-0.1073646471,0.2608529925,-0.0292999409,-0.0793902054,0.2039854974,-0.148156628,0.3061811626,-0.0629678369,0.3361934423,-0.0899120644,0.2520775199,-0.0519085974,-0.4535894096,0.0178125389,0.3152451813,0.407485038,0.0184305087,0.0400744975,0.2059644312,0.120073542,-0.2516606152,-0.3879702389,-0.2059795707,0.0908953026,-0.2052687109,-0.1253317297,-0.1193971634,0.1875582188,0.043237146,0.0816024616,0.4845341146,0.0002889726,0.1447360814,-0.0010887757,0.2192268968,0.0464254729,-0.2223508209,-0.1943704933,-0.1116545573,0.1098374575,0.0302241202,0.2480703294,0.0179758705,-0.1454162598,0.0168016292,0.6618235111,0.0420250148,0.258257091,-0.1182069257,-0.3660911322,0.2513479888,0.1739427298,0.087606214,0.3652841747,0.2524144948,-0.2956422269,0.0845419914,-0.1635932028,0.0505567454,0.0976488069,-0.0679601654,0.3684989512,-0.0912532359,0.3075386584,0.0185329355,-0.3090053499,0.0744575188,-0.0793073252,0.1062876061,-0.3757323921,0.1150494143,-0.2569516301,-0.1478515416,-0.0088819563,-0.133373037,0.0266572647,-0.1974048913,-0.0598685481,0.3584897518,0.0893847048,0.282640487,0.2069196403,0.0192367453,-0.0590613931,-0.1030487418,-0.1551313996,-0.108617492,-0.11700131,0.0823388919,0.0588655174,0.1915499568,0.3331587911,-0.3143273294,0.0909607261,0.1370236278,-0.3026359379,0.0858223662,-0.1740436107,0.2888210118,0.2684489191,0.1589440256,-0.1759042591,-0.0624604672,0.4340486825,-0.3824962378,-0.192428574,0.2827731073,-0.1414361447,-0.0550831221,-0.2623379529,-0.2640984952,-0.5197219253,-0.2946316004,0.2550328076,0.0227000788,0.0717564151,0.3472696841,0.234048292,0.452318579,-0.3834219277,-0.0229388513,-0.1441199481,0.0412894003,0.2728778422,-0.0820257813,-0.4259680212,-0.0077680945,-0.0855097622,0.2539603114,-0.1286979169,-0.0568948649,-0.4479478002,-0.1266304255,0.3238252997,0.0086144321,-0.1531503648,0.6029988527,0.0156400762,-0.1078105569,-0.1583341658,-0.1767390072,0.1045932546,0.1382757425,0.0056557627,-0.0700706616,0.2453286648,-0.0265153348,0.351564467,-0.0049929833,-0.2051898241,0.3471206427,-0.304443121,0.3146957457,-0.1218410507,-0.5708717108,-0.0727783665,-0.0265240464,0.0520668849,-0.1330883354,-0.2462659478,-0.0897768289,-0.3285742104,-0.1738543063,-0.4415084422,-0.2324540317,0.0581631921,0.1306092143,-0.0827693418,0.0958863199,0.0774856657,-0.3568073511,0.0563106313,-0.0132170403,0.212240696,0.0356143676,-0.1878067106,-0.4234542251,-0.2692607045,-0.3553898036,0.3830406666,0.0468409471,0.0110894823,-0.0506009758,-0.1024787277,0.1410767138,0.0076352716,0.6806127429,-0.0301307887,-0.0055903825,0.416000694,0.1417528987,-0.6230165362,0.0706069246,-0.3021711707,0.0531921424,0.1227650046,0.3985318244,-0.4662339985,-0.1559689939,0.3363058567,-0.0372493379,-0.0328230187,-0.2179400921,-0.2882821262,-0.3978614211,-0.213686794,0.1251843274,0.1448903382,0.3859930634,-0.225021705,-0.1091162786,-0.124067463,0.0948003605,0.0340228826,0.2724271715,0.2362679839,-0.228084147,-0.095412448,0.0113189863,0.0838170201,0.2514652312,0.2685821652,-0.0433206074,-0.5514891148,-0.1514342278,0.1056525484,0.2161636204,0.0786006078,-0.1357617527,0.0079719489,-0.0911873356,0.1339932978,0.2157751322,0.1181629151,0.4902128577,0.169891879,-0.2991678119,-0.191720441,0.452149719,0.1290634274,0.0447196178,0.1472747773,0.4835731387,-0.169077903,0.2779710591,-0.1196076274,1.0848121643,-0.0030835283,-0.005263906,0.5609253645,-0.1063345298,0.6737447977,0.0900183097,0.2337673157,-0.36399436,-0.1940107197,-0.0178164672,-0.1334923059,0.2924979925,-0.0650964603,-0.5087834001,0.1753665209,-0.2362163216,0.1639979631,0.0437814966,0.0470250808,-0.1116012633,-0.1634422541,-0.319111824,0.1553267986,-0.0189958215,0.2381698191,-0.0293426774,-0.1032173112,-0.2884888649,-0.2500365078,-0.1799219549,0.0486413091,-0.4893500209,-0.2644637525,0.2028944343,-0.1735015661,-0.2620542645,0.035055887,0.1736058295,0.0271786191,-0.2295172513,0.085042268,-0.0116313575,-0.0057277782,-0.0612283163,-0.1201168895,0.4516772926,-0.1806232482,-0.5559980273,-0.0021970393,0.0283423346,0.1189502105,0.3397734463,0.0533939674,-0.0720781907,-0.3765566647,-0.1956294924,-0.0863566697,0.2171403468,-0.2849394381,0.215872243,-0.0756182298,-0.1748709828,-0.0020164615,-0.0216978751,-0.0546836369,0.0181161389,0.3995351791,0.2500368953,0.3563725352,0.2536879778,0.1453519017,-0.1277043223,-0.2788094282,0.0454512127,0.2644447982,-0.5634872317,0.1480279118,0.2177876085,-0.1283232868,-0.0758905932,0.4255404174,0.0340973884,0.0199873578,-0.1081608087,-0.1711105853,-0.2474561185,0.4519791007,-0.0532878414,-0.053389363,0.2934162915,0.5443694592,0.2228721976,0.0077437623,-0.4028413296,0.025847666,-0.3943938017,0.0516046882,0.0506236702,-0.2735677063,0.2741547823,-0.0607873648,0.2235942632,0.2008981258,-0.0648372918,-0.3292416632,-0.1574597955,0.1020702645,0.0132466042,0.1558352858,0.1404897571,0.0429231934,-0.1865489632,-0.1436392814,0.1646613628,-0.0484871604,-0.0519672707,0.2371664941,0.1691443026,0.3118648231,0.0495464392,0.0909304991,-0.0028736435,-0.1294036061,-0.0817828923,0.0178820342,0.0095177591,0.0264174677,-0.0565491654,-0.1073949337,0.1240915209,0.1748039126,0.0907406211,-0.2451831847,-0.3529208899,-0.2803744972,0.1817104667,0.1533816606,-0.1543813348,0.1704663485,-0.0670958683,0.2948552966,-0.4210546017,0.0239637289,0.2194127142,-0.0067794458,0.179174453,0.1512840837,-0.2732408047,0.2063111812,-0.0471630208,0.0283583831,0.4666559994,0.0181072932,0.0889500454,0.2631120682,-0.1317112595,-0.0344259217,0.4894234538,0.1561266631,0.2242719233,-0.0069709676,-0.0884023011,0.2236964554,-0.1240096763,0.074921973,0.2926060557,-0.216833517,-0.0012596183,0.1987000406,0.0942938253,-0.127527684,-0.3349516094,0.5498442054,-0.2395892292,-0.2107713521,-0.1777208298,0.1950450093,-0.2097447366,-0.0012295916,-0.1400179267,-0.0810490996,-0.1268614382,-0.4116225243,0.0561529286,-0.1514625102,0.4862962961,-0.1211124882,-0.0871192738,-0.2842113078,-0.2175373286,0.408749193,0.3050468564,-0.154292956,0.1880830228,0.0706383362,0.1759784222,-0.2439061552,0.5857952833,0.4205303788,0.2032050192,-0.0877042264,-0.0394125432,-0.0987181142,-0.3073500991,0.0758499652,0.1601038575,0.3437224329,0.2210115194,0.0020970211,0.2533309758,-0.2443259656,0.2270522863,-0.1037411615,0.0440704562,-0.4042601585,0.3363026083,-0.3260605931,-0.2499150783,-0.3251251876,0.0070015988,-0.4086143374,0.1709913462,0.3403113782,0.0040550446,0.1531253904,-0.3124785423,0.1025997922,-0.1237484664,0.287279129,-0.0421427861,0.2992600203,-0.1812504232,0.0731786937,-0.4249604642,0.395984292,-0.1252929121,-0.112262547,0.0873513669,-0.1618833095,-0.1488880217,0.0399846993,0.0557344407,0.4386814237,0.056703303,-0.059051577,-0.5389719605,-0.1532924324,-0.1344299763,-0.2273552269,0.2009892464,-0.2520205975,0.0800778195,-0.1442407817,0.0637667999,0.0295061786,-0.0963596702,0.1442123055,0.0670828372,0.6210793257,0.1881039739,0.3392416537,-0.0430545397,-0.0689965189,0.0249925554,-0.174067378,-0.2406356484,0.2803048193,0.2160280198,0.2756659687,0.0625708327,0.0870543271,-0.0475448668,0.1085865125,0.0825032964,-0.2264692187,-0.0616678037,0.119794324,-0.0202589016,-0.2373049855,-0.1996346414,0.0610714257,0.005293489,0.0769687742,-0.2243449986,-0.2481006831,0.7152537704,-0.2563150823,-0.2637996078,0.055393666,0.2930933833,0.1165040582,-0.4328460097,-0.7834493518,0.0166447517,0.2939421535,0.083926104,-0.2240283042,0.2948251665,0.0100844931,0.0482179038,-0.0746177435,0.1352374554,0.0407708175,-0.0168586764,0.0712476522,-0.2007396221]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"- For the platform, we need to know the operating system of your machine. Could you please run the command `datasets-cli env` and copy-and-paste its output below?\r\n- In relation with the error, you just gave us the error type and message (`TypeError: 'NoneType' object is not callable`). Could you please copy-paste the complete stack trace, so that we know exactly which part of the code threw the error?","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":69,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\n- For the platform, we need to know the operating system of your machine. Could you please run the command `datasets-cli env` and copy-and-paste its output below?\r\n- In relation with the error, you just gave us the error type and message (`TypeError: 'NoneType' object is not callable`). Could you please copy-paste the complete stack trace, so that we know exactly which part of the code threw the error?","embeddings":[-0.2277715802,-0.3193501234,0.0008596373,0.3918522596,0.3677237034,0.0557456017,0.3176339269,0.1545283645,0.1733444631,0.2960280478,-0.1426855326,0.4552958608,-0.0776570067,0.2298437357,0.0493332893,-0.1651720107,-0.2105034888,0.2611335516,-0.1191705689,0.0262278002,-0.5092200041,-0.0741478875,-0.1640852094,0.1349925548,-0.237754494,-0.2719812989,-0.1240942478,0.0408258066,-0.1921427697,-0.2838498354,0.3645710647,-0.0801881999,0.2122620046,0.5988925099,-0.0001008489,-0.1514540613,0.435205698,0.0475999489,-0.1254921108,-0.3523042202,-0.2946639061,-0.3847207725,0.18985717,-0.3827019334,-0.0743031427,-0.0082017854,0.0286093634,-0.2426373065,0.3481251895,0.3484524488,0.3506994545,0.415902108,0.0833576843,-0.2823812366,0.1884763539,0.0775724575,-0.1048533171,0.2966803014,0.0623726882,-0.0164799206,0.3145338893,0.1602044255,-0.2830132842,-0.0542992242,0.2551215291,0.0645872205,-0.0053246845,-0.3559243679,0.0310364459,0.3066703379,0.4195721149,-0.3946249187,-0.1872214228,0.0607764684,0.1299087554,-0.2234973758,0.1539727449,-0.0115599288,-0.0064773369,0.0658862591,-0.0040367073,0.1557925344,-0.236804232,0.2183835655,-0.1664240062,0.0901961327,-0.2121003419,0.107063584,0.0670970306,-0.1992394775,0.0452290624,-0.1839859337,0.0351355337,0.1096044704,-0.221126616,-0.1452792883,0.1299011409,0.1111679673,0.1270107031,-0.2455778271,0.1168725938,-0.0483121127,-0.0236970503,0.4071713388,0.239894405,0.2160654962,0.197072491,0.1496332735,0.2160288244,0.0441346094,-0.1853791922,-0.0519746169,-0.1255763024,-0.1014563665,0.2814617455,0.1962061971,0.5696896911,-0.0633770376,-0.4451116025,-0.0369891338,-0.207089901,0.0923674777,0.2541407347,0.3783914447,-0.1840863526,0.1531827301,0.2149179727,0.0695832223,-0.1426405609,-0.2327828705,-0.2402686328,0.3336758912,-0.2726967335,-0.232448265,0.0732233152,0.0800087824,-0.1016519293,0.0515346751,-0.0466053039,0.0736266375,0.1052564383,-0.382973969,-0.1715993732,0.1613502353,0.1342369616,-0.0109220734,0.3575491905,-0.413230598,-0.1455421895,0.1080859974,-0.3179209232,-0.0927535221,-0.0903914943,0.3212557733,-0.2362898439,-0.079209581,-0.1951911002,-0.0047342735,0.0772167519,-0.2021209151,0.0243879221,-0.1612926275,0.0099704852,-0.4124752581,0.0396714061,0.5434519649,-0.3702253997,-0.131931603,-0.3560011387,-0.0526704639,0.3650825024,-0.1617242545,0.0869184732,0.2049263865,-0.1619556099,-0.0783798695,0.5953560472,-0.3626013696,-0.1942619234,0.0556299053,-0.079251498,-0.0580402203,-0.067448847,-0.0140706375,0.0950176865,0.1139309034,0.4095413983,0.2211544514,-0.0633310452,-0.027870506,-0.1950395554,-0.1733900309,0.1649025828,0.0918372497,0.1436722875,0.1626258194,0.0630830303,-0.2037843466,0.0184938218,-0.0397915058,-0.09291742,-0.0652414113,0.6565243602,-0.0872236714,-0.0788739324,-0.5945641994,-0.3253455162,0.1684785038,-0.0677354112,0.2807234824,-0.1447329372,-0.1707682014,-0.3664011657,0.0677765012,-0.0498286709,0.0480884425,0.34265396,0.1271567047,-0.1998722106,0.0451182313,-0.1443472803,0.1779200137,-0.1498600394,0.1019958928,-0.0057357312,0.3032817245,-0.0105569726,-0.3550938368,0.0437213145,0.2375442684,0.5392077565,0.0301111341,-0.0714498162,0.2961714268,0.0839738399,-0.1378158778,-0.1490661353,-0.0828541517,0.1191909611,-0.2615092099,0.0761332735,0.113510862,0.341347903,-0.0806762725,0.1379154474,0.550291419,0.070542492,0.218250677,-0.0425794087,0.2359925807,-0.0458750986,-0.0404749922,-0.09132009,-0.1391232908,0.1310136467,0.0969333351,0.1783950478,-0.0640903935,-0.1166954562,0.0628789961,0.5675632358,0.0985855833,0.3846487105,-0.0710379705,-0.3002262712,0.1548450738,0.0890177488,0.2822793424,0.4458229542,0.3082370758,-0.1348168105,0.0028449628,-0.105473116,0.0909257606,0.069577001,-0.008273948,0.2112074643,-0.0183985941,0.1782516539,0.0799446777,-0.1778929234,-0.0452206172,-0.2235442847,0.2644327581,-0.3588179946,0.1916420162,-0.2675382495,-0.1693612486,0.0416815057,-0.0662412271,-0.1583892852,-0.343996793,-0.2322412729,0.0256404057,0.1133051515,0.1315740049,0.114424713,0.1559722424,0.0080657853,0.0855939835,-0.1656484157,-0.1026751921,-0.1289932728,0.1639403254,0.1211261749,0.2124070823,0.4851669073,-0.2281131297,0.2450820208,-0.0264441986,-0.227994293,0.0638851374,-0.0960038304,0.4924280643,0.4533854723,0.170710057,-0.0828904957,-0.0167101845,0.3197818995,-0.3656996191,0.0297314506,0.2110722065,-0.2212308943,0.1077203006,-0.1586056501,-0.1741414219,-0.5684420466,-0.4111053944,0.1035611331,0.1680119336,0.1657065153,0.1624201983,0.1247417033,0.5334134102,0.1375749111,0.1933974028,-0.195846498,-0.0487356074,0.2836108506,-0.122860238,-0.2586844563,0.1470841169,-0.1050434485,0.2065625638,0.0155344047,-0.1752419621,-0.5393431783,-0.1731456071,0.1546034515,-0.0829334036,0.0919005796,0.2602021098,0.1155401021,-0.1390545517,-0.2339776456,-0.1428654939,-0.1661490351,0.2557976246,0.0672933385,0.0401755236,0.3397756815,-0.178367734,0.1294477135,0.0065674935,-0.2346490473,0.3460306227,-0.2629454136,0.3489436209,-0.1248321906,-0.5024338961,-0.1444402039,0.0058216783,-0.0900051892,-0.0492299162,-0.1613157392,-0.1666391939,-0.3757922053,-0.12356738,-0.4473450482,-0.173208639,-0.02683492,0.213281557,0.0196586438,-0.0107808877,0.1226413622,-0.4682358205,0.0489344634,-0.0516467467,0.2382136434,0.0197485723,-0.2432868481,-0.2479406893,-0.3090171814,-0.1854937226,0.4490612149,-0.1629894227,0.0337656327,-0.1722394228,-0.0209151395,0.0463810414,-0.0175552871,0.5029234886,-0.3419142962,0.0797719881,0.4985005856,0.1387616396,-0.3840284348,0.1739851683,-0.4070257545,0.3183746338,0.1517318189,0.2837773263,-0.3160907328,-0.0206106473,0.2402315289,0.1300393194,-0.1309812963,-0.2083940059,-0.1448206306,-0.2342955619,-0.1278313398,0.0245009754,-0.0037171121,0.3006121814,-0.1316144913,-0.1478490233,-0.3589732349,0.0226067789,0.2575157881,0.2095248997,0.2955670655,-0.0968761295,-0.0735841319,-0.068001546,0.0821264386,0.1955122501,0.274494797,0.0189455654,-0.391548723,0.0630288497,0.0565042421,0.2511690259,0.0781225115,-0.1441668868,-0.0535798632,-0.0800635144,0.1943779588,-0.0523580499,0.1001058295,0.2666032314,0.0928936079,-0.4500440657,-0.3444149792,0.388646096,0.191103518,-0.0468111858,0.2515259981,0.2554996312,-0.2587035596,0.2411348969,0.0585117638,1.0566794872,0.0451303609,-0.1790634096,0.4914720356,-0.1253181845,0.5688884258,0.0427971072,0.0154062081,-0.3378975689,-0.2656220496,-0.0510648116,-0.1546158791,0.3252924979,-0.0990185961,-0.6059663892,0.2722361684,-0.4187105596,-0.0113039594,-0.0516176261,-0.1450726092,-0.1615752131,-0.130570069,-0.4006145895,0.2759853601,-0.1767890304,0.1705224812,-0.1539548337,-0.0489460118,0.0174538828,-0.1243231222,-0.1045431644,-0.00681563,-0.4354237616,-0.2015477568,-0.1179327592,-0.1617850065,0.017373981,0.1457440257,0.1715212911,0.1336825788,-0.2375279218,0.0901856199,0.0600658134,0.0691440105,-0.0843933225,0.0353280008,0.2427478284,-0.0197919644,-0.3155231774,0.1725399643,0.0347914137,0.1483977884,0.3645579815,0.0996133909,-0.0831791535,-0.3884962499,-0.1447437406,-0.0601997264,0.2035056949,-0.2569366693,0.2733273804,0.1267134398,-0.075365968,0.1074780077,-0.0244918708,-0.0938684121,-0.0563318618,0.2750467658,0.3504133224,0.2657757103,0.0927115083,0.201552093,-0.2149729133,-0.3738940656,0.2654359341,0.0770252869,-0.22658512,0.2399508804,0.1424390525,-0.0477566309,-0.0740084425,0.328137219,0.1060338765,0.0496939942,-0.1529465616,-0.2888395488,-0.2707046568,0.2582178116,0.1262251735,0.0776324049,-0.0284722429,0.2088868767,0.1913177222,-0.0492134839,-0.4822713137,0.0605647266,-0.3836083114,-0.0870803818,0.0551759526,-0.0908131376,0.1894112378,-0.0005163328,0.2610564232,0.0888261497,-0.0810422897,-0.3424244523,-0.1005696654,0.0964426696,0.0030663002,0.0352182798,0.1485673785,-0.1207061037,-0.1707189232,-0.1411678195,0.0266073775,0.0593882687,-0.1587965488,0.1150495261,0.1243259236,0.2831470072,0.0353000462,0.2056675106,0.0769665316,-0.0229453649,-0.0435421765,0.0187972002,0.0036229359,-0.0640793815,-0.0301972888,0.0279702153,0.1451300681,0.1402733177,0.1255751401,-0.1008501798,-0.2102200091,0.0878476053,0.1658077836,0.1559677124,-0.141983971,0.2770460844,-0.0868307948,0.3664443195,-0.274410367,0.0608054735,0.1910197437,-0.1125192195,0.2319393456,0.1514575779,-0.3416772187,0.2574727237,-0.160598442,-0.024174612,0.3763528764,-0.1595057696,0.154295519,0.2322441638,-0.1036700085,-0.0269345958,0.3910489976,0.2250012904,0.089335233,0.0697669387,-0.1052329168,0.1659094244,-0.0354886912,0.0229820106,0.23917602,-0.4080982804,0.2459843904,0.2516362369,-0.0291607659,-0.0656304657,-0.1686448157,0.2762493193,-0.5681785941,-0.1020831168,-0.3088049889,0.0113962851,-0.1732568741,0.1290312707,-0.128262803,-0.1437105089,-0.2699719965,-0.3319649398,0.0556728579,-0.1790091693,0.195461601,-0.1371625513,0.1048784405,-0.1952640265,-0.0611734428,0.2457263917,0.1937719285,-0.2082418054,-0.0402503274,-0.0125757577,0.2360838503,-0.1725241244,0.5394890308,0.3954853415,0.3322067857,-0.1617974192,-0.0274389703,0.0731182173,-0.2820368111,-0.0603159703,0.0485493392,0.0340395607,0.2409244031,0.0659549609,0.3405122161,-0.3197934926,0.0967679396,-0.0659998953,0.0167914163,-0.3043793738,0.4304428101,-0.3063565493,-0.1328262538,-0.1199343801,-0.1177720949,-0.5592741966,0.0602604449,0.4006091654,0.0640409663,0.1086076871,-0.1182630286,0.1509963721,-0.1716412604,0.1180160418,-0.0006215716,0.3370815516,-0.2615213096,0.0773840994,-0.4892735481,0.3647934198,0.0276360251,-0.1982382685,0.0315376259,-0.119094789,-0.1470418572,0.009699991,0.0401231274,0.238081336,0.2386036366,0.0062592286,-0.5759982467,-0.3410032392,-0.1011373773,-0.2159598768,0.1086912006,-0.2300252169,-0.0447582565,-0.1500625759,0.1826710105,-0.0428104773,-0.257914722,0.0840296447,0.1191138402,0.6550236344,0.1934750229,0.3989144266,-0.12089663,-0.0226046145,0.1570947468,-0.3202399313,-0.0817679614,0.2735521197,0.1210139245,0.2002992332,0.0415458567,0.1631483734,-0.2030333132,0.0024250993,0.0274581667,0.0585398674,-0.1464440376,0.3077630401,-0.1645611525,-0.2361537367,-0.2291445434,0.1095362157,0.0604569279,0.1458997875,-0.0198298264,-0.1106984243,0.624981761,-0.1692866236,-0.2598852217,-0.0570382588,0.3724466562,0.0044079828,-0.1203285828,-0.4922212958,0.003138799,0.4646417499,-0.0488136932,-0.1663854867,0.2039655447,-0.1955647022,-0.0195632055,-0.0393957794,0.1110214889,-0.0188861452,0.0038260752,0.2579205334,-0.1187548116]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"> * For the platform, we need to know the operating system of your machine. Could you please run the command `datasets-cli env` and copy-and-paste its output below?\r\n> * In relation with the error, you just gave us the error type and message (`TypeError: 'NoneType' object is not callable`). Could you please copy-paste the complete stack trace, so that we know exactly which part of the code threw the error?\r\n\r\n1. For the platform, here are the output:\r\n - datasets` version: 1.11.0\r\n - Platform: Windows-10-10.0.19041-SP0\r\n - Python version: 3.7.10\r\n - PyArrow version: 5.0.0\r\n2. For the code and error\uff1a\r\n ```python\r\n from datasets import load_dataset, load_metric\r\n dataset = load_dataset(\"glue\", \"cola\")\r\n ```\r\n ```python\r\n Traceback (most recent call last):\r\n ....\r\n ....\r\n File \"my_file.py\", line 2, in \r\n dataset = load_dataset(\"glue\", \"cola\")\r\n File \"My environments\\lib\\site-packages\\datasets\\load.py\", line 830, in load_dataset\r\n **config_kwargs,\r\n File \"My environments\\lib\\site-packages\\datasets\\load.py\", line 710, in load_dataset_builder\r\n **config_kwargs,\r\n TypeError: 'NoneType' object is not callable\r\n ```\r\n Thank you!","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":154,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\n> * For the platform, we need to know the operating system of your machine. Could you please run the command `datasets-cli env` and copy-and-paste its output below?\r\n> * In relation with the error, you just gave us the error type and message (`TypeError: 'NoneType' object is not callable`). Could you please copy-paste the complete stack trace, so that we know exactly which part of the code threw the error?\r\n\r\n1. For the platform, here are the output:\r\n - datasets` version: 1.11.0\r\n - Platform: Windows-10-10.0.19041-SP0\r\n - Python version: 3.7.10\r\n - PyArrow version: 5.0.0\r\n2. For the code and error\uff1a\r\n ```python\r\n from datasets import load_dataset, load_metric\r\n dataset = load_dataset(\"glue\", \"cola\")\r\n ```\r\n ```python\r\n Traceback (most recent call last):\r\n ....\r\n ....\r\n File \"my_file.py\", line 2, in \r\n dataset = load_dataset(\"glue\", \"cola\")\r\n File \"My environments\\lib\\site-packages\\datasets\\load.py\", line 830, in load_dataset\r\n **config_kwargs,\r\n File \"My environments\\lib\\site-packages\\datasets\\load.py\", line 710, in load_dataset_builder\r\n **config_kwargs,\r\n TypeError: 'NoneType' object is not callable\r\n ```\r\n Thank you!","embeddings":[-0.2333926857,-0.2121910155,0.0151841119,0.3540363908,0.431312561,0.0565467253,0.4275258183,0.1668837517,0.1619586647,0.2543876767,-0.1491201818,0.4838864207,-0.032627441,0.2008449435,0.1351143867,-0.1875769347,-0.1780486703,0.1836340725,-0.2075739056,0.0554907173,-0.530339241,-0.0694576576,-0.131100297,0.1142207906,-0.165304631,-0.2441832423,-0.1580201536,0.0987920165,-0.1795856655,-0.3501777351,0.4064702392,-0.1164285839,0.2196373641,0.562065959,-0.0001015193,-0.0459203348,0.4622377753,0.0741885006,-0.1826767772,-0.3177576065,-0.2952965498,-0.3818572462,0.1700716317,-0.3912402391,-0.0481698103,-0.0371986367,-0.0039750608,-0.2519207001,0.2980602384,0.3795799315,0.328802675,0.4538740516,0.0911629125,-0.2488764822,0.1904570758,0.0910028517,-0.0621268973,0.3023785055,0.0538949445,-0.0567670763,0.2113882601,0.1328090131,-0.3628790081,-0.0556369275,0.2719312906,0.0631388128,0.0336346067,-0.3873294294,0.0748671964,0.3000453711,0.4482415617,-0.3896304667,-0.2469754517,0.0235931426,0.1223488674,-0.1994509548,0.2123644501,0.0509505086,0.0008063291,0.0616286434,0.0196442436,0.1935778111,-0.1930177808,0.2497536093,-0.1554912627,0.1598448157,-0.1672741175,0.1182282194,0.0708542541,-0.2033922374,0.049795296,-0.2369977683,0.007765641,0.1136214361,-0.2622630298,-0.0578037463,0.1595869213,0.1328142434,0.0924718305,-0.2031244338,0.0840387568,-0.0385783762,0.0005321783,0.3734622598,0.2657110691,0.2719011307,0.2630656064,0.1053493395,0.3113079071,0.0968508199,-0.1035405621,0.0171599817,-0.1741403788,-0.1962881684,0.2919623256,0.27147156,0.5614536405,-0.0071394164,-0.4003361762,-0.112911351,-0.1762718856,0.0414780118,0.1914937347,0.3488394022,-0.1659963131,0.1930406094,0.2003636211,0.1248331815,-0.2139800191,-0.2602629662,-0.2342291027,0.277436316,-0.2562621832,-0.1992109269,0.1157746315,-0.0126312999,-0.0741725713,0.0112992479,-0.1771455258,0.1546337306,0.1403729916,-0.3297455609,-0.1945308298,0.1805401146,0.0513517223,0.0468542874,0.3396641314,-0.3767827153,-0.1449093074,0.0857277736,-0.2972788513,-0.1406729817,-0.1555437893,0.3137725294,-0.203810975,-0.0596022606,-0.2211033255,-0.0271769408,0.1290350109,-0.1592079401,0.0224947892,-0.1568120122,-0.0343936272,-0.4068643153,0.0919899419,0.6147862673,-0.4071787894,-0.1555482447,-0.3654867709,-0.1499285698,0.3206854463,-0.2192046493,0.0420476161,0.205342114,-0.1884394288,-0.0670418441,0.5587269068,-0.4109594524,-0.2658071816,0.071330905,-0.1287924349,-0.0868593082,0.0406506956,-0.0299981683,0.0722305849,0.1137080416,0.4100970328,0.275131911,0.0011142003,0.0079012448,-0.1864448488,-0.2357116938,0.1991765499,0.0649753883,0.1384467632,0.1023558527,0.1406214237,-0.1971790642,0.0246631335,-0.0487639643,-0.0327245519,-0.057789661,0.5558087826,-0.0468545221,-0.044843737,-0.5518350005,-0.4227520227,0.1897544414,-0.0799524039,0.3541664779,-0.1335036159,-0.1872806996,-0.3264653981,0.0583028831,-0.1389628798,-0.0200660415,0.3176531494,0.1545748264,-0.1688733399,0.0288684927,-0.0564694032,0.2119377255,-0.0756313056,0.0736123845,-0.14293015,0.3080175817,-0.057636708,-0.3339879513,0.0377683677,0.2538016438,0.5117610097,-0.0004488377,-0.0633197874,0.3244804144,0.0866748691,-0.0496543907,-0.1650345176,-0.0999800116,0.1509415507,-0.2289367616,0.0971660241,0.0872650221,0.3168590069,-0.0662048236,0.0954951942,0.4775231183,0.0889250115,0.2869647741,-0.0390479192,0.2417568564,-0.0411586203,-0.0287830383,-0.0849919841,-0.1743532121,0.0761678517,0.1894836426,0.2709217072,-0.0320920832,-0.0377041772,0.0921959504,0.5580118299,0.0443215631,0.3538882732,-0.0804426596,-0.3818507195,0.0884140804,0.1868627965,0.2703929543,0.4774660766,0.3163056076,-0.1177275926,-0.0135746626,-0.0687375143,0.068971917,0.0899053961,0.0000686368,0.2182746232,0.0326346494,0.2071536779,0.0608280376,-0.2113714963,-0.0529055037,-0.1733764559,0.2231471837,-0.4077748656,0.1184415296,-0.3216187358,-0.1545066088,-0.0203860439,0.0009128465,-0.1665667146,-0.3700627089,-0.2923040688,0.0090758018,0.0586140528,0.1355309337,0.0729432553,0.0848084092,0.0303801391,0.0038109233,-0.1432915479,-0.1506637037,-0.197236672,0.1309271753,0.1724003851,0.2035219669,0.4579099715,-0.1723938286,0.1496831626,-0.0668234155,-0.2592738867,0.0691069737,-0.0358921215,0.5192118287,0.4751577079,0.1759489179,0.0373363346,-0.0723712072,0.3917131424,-0.337565273,-0.0363603607,0.2813171744,-0.2262491435,0.1912381798,-0.1627744734,-0.2709549367,-0.5850613117,-0.4314647317,0.0900498033,0.1581933945,0.1574169397,0.2290733755,0.1376304924,0.4989827573,0.0554460995,0.173627466,-0.1574210674,-0.0517017953,0.2869262695,-0.128268525,-0.3188622594,0.0627037063,-0.1610585004,0.2123719454,0.0095064696,-0.2091803849,-0.5421624184,-0.1937738359,0.248898387,-0.0088543072,0.1406644732,0.2820861042,0.1652371287,-0.171896413,-0.2008978277,-0.1451961696,-0.0958748311,0.2647386193,0.003278581,0.1268489063,0.3415829241,-0.1613709927,0.1591750383,0.0212520342,-0.1993087977,0.414999783,-0.2640288472,0.3553538322,-0.1553532481,-0.501110673,-0.1315999925,-0.034155868,0.0119620711,-0.0257774498,-0.1498845816,-0.1616336107,-0.3645355999,-0.1962905228,-0.4402103126,-0.1943258345,-0.0451902971,0.199670434,0.0204199702,-0.0293068476,0.1270928979,-0.4808782339,0.0326096974,0.0426807031,0.2218273133,-0.0436264426,-0.1767756939,-0.2460242808,-0.2401109934,-0.2647053599,0.4546621442,-0.158620581,0.0367192812,-0.1481581628,-0.0604273044,0.0030126467,-0.0713322535,0.5519664884,-0.3454786837,0.167946592,0.5127314329,0.1982301772,-0.3972994685,0.1448308527,-0.3467212319,0.197871536,0.0658447966,0.3129544258,-0.3585754633,-0.0375085883,0.2515778542,0.2019121051,-0.129298076,-0.1740512401,-0.1447782964,-0.2503760755,-0.1316992491,-0.058486212,-0.1246138215,0.3594864011,-0.1265377104,-0.1583637297,-0.3093515933,-0.0301203411,0.1789499223,0.1840766668,0.2802935541,-0.0759279355,-0.0367612876,-0.0464058407,0.071699515,0.2319861799,0.3442593217,-0.0239473321,-0.360599637,0.0667763501,0.0140341548,0.2356529385,0.0585258938,-0.1843719929,-0.1035960093,-0.0533856153,0.200205341,-0.1060454622,0.0749847591,0.3068714738,0.0640616715,-0.3982105851,-0.3787680268,0.4075770378,0.1482772082,-0.055876337,0.2646415234,0.2493728548,-0.3079307079,0.2636658549,0.0765032694,0.9972296357,0.0634360686,-0.1427229047,0.5291199684,-0.0992499292,0.5708351135,0.0216645449,0.0737809911,-0.3225640059,-0.2403181046,-0.0621667355,-0.1458812952,0.3637820482,0.0177567303,-0.554625988,0.2570112944,-0.4040381014,0.0338811316,-0.0993947461,-0.0540791601,-0.168008253,-0.1555937231,-0.4221657813,0.2615644336,-0.1391600221,0.1646663249,-0.0964661688,-0.0559898913,0.0007044189,-0.1505103707,-0.0544889867,0.0260486845,-0.470529139,-0.092832163,-0.1959897131,-0.1402023137,-0.0150172981,0.2117722034,0.1081975773,0.0798802003,-0.2208486944,0.074763,0.1003000066,0.1113611236,-0.0745936707,0.0186240617,0.2860429585,-0.0365927033,-0.3421170413,0.145158276,-0.0346250087,0.1750632823,0.3781581819,0.0789964721,-0.0583959855,-0.3603264987,-0.1006426886,-0.0581629425,0.2370610982,-0.2568027377,0.2513565421,0.0545558929,-0.0355337635,0.0711976513,-0.1054593101,-0.0828102529,-0.0973028988,0.3941631317,0.2912923396,0.2772480845,0.2169934809,0.2351005822,-0.2244954705,-0.3636664152,0.2889499366,0.0990770832,-0.3086126447,0.198417455,0.2202891111,-0.0483231917,-0.0357774086,0.3472393751,0.0114345681,0.049836304,-0.1350132227,-0.2807761729,-0.22253865,0.2674214244,0.0906429961,0.0844058841,0.028326951,0.1621989608,0.1774459928,-0.0547243915,-0.451469332,-0.011736189,-0.3375882506,-0.0379747562,0.1141254455,-0.0688912645,0.2117233276,0.0469736308,0.266826123,-0.027095424,-0.095401898,-0.3234717548,-0.1007948369,0.1007452458,-0.0319782905,-0.0041054767,0.2243584841,-0.1309217066,-0.1626257896,-0.1651719809,0.1755297929,0.0577325188,-0.1902176887,0.1038779467,0.1569740027,0.2834070921,0.0430249199,0.2171774507,-0.0266571995,-0.0581222437,-0.0226649251,0.0391197726,0.0156680252,-0.0498561636,-0.0160412788,0.0438439623,0.0941356197,0.1154540554,0.1023316532,-0.1328705847,-0.2397717088,0.1010522768,0.1707957089,0.1672703177,-0.1169895381,0.2498272508,-0.0852398649,0.343233645,-0.3024182916,0.0496240407,0.1867102683,-0.1545248628,0.2260235697,0.1156060994,-0.3048253357,0.2245938927,-0.14696531,-0.0125416005,0.4131203294,-0.1344489008,0.2140961289,0.3211934268,-0.0835602805,-0.0166897029,0.3776884973,0.1835822165,0.1086514145,0.0316770524,-0.0579561144,0.1470866799,-0.0196429528,0.0855417624,0.2414101213,-0.5050227046,0.2710326314,0.2405994534,-0.0243101902,-0.1121373996,-0.163989529,0.3712483943,-0.5745438933,-0.0636712015,-0.3064317703,0.0594137311,-0.1685567945,0.1159525663,-0.1114942878,-0.1400388479,-0.256551981,-0.2784867585,0.06332881,-0.2414887398,0.2439766079,-0.1118730009,0.0211273301,-0.2714125514,-0.0401527472,0.2217802703,0.1276277751,-0.1941972971,0.0097096395,0.0506009683,0.2041798085,-0.1806842834,0.552272737,0.3988933563,0.3800224066,-0.1354636401,-0.0664787814,0.1149364188,-0.2638256848,-0.0904527828,0.0105616841,-0.0362640731,0.2590774,0.1043943912,0.3203140199,-0.3022005558,-0.0068912487,-0.0190034341,0.0057310155,-0.2694809437,0.3833438158,-0.2941297293,-0.1339461505,-0.1130259633,-0.0764845833,-0.5723281503,0.0453014746,0.4337003529,0.0553493984,0.1160499007,-0.1790060252,0.1447362602,-0.1381019503,0.1920721382,-0.0137894452,0.2897506058,-0.266952455,0.0805813819,-0.5247184038,0.3150629103,0.0391136967,-0.1924300641,-0.0284713041,-0.1205905452,-0.1418034434,0.0033041995,0.0252633821,0.2740843296,0.2506788373,0.0156387016,-0.544901669,-0.293649137,-0.0416166857,-0.1840641648,0.0557728298,-0.3411940038,-0.002358932,-0.1740486026,0.1497212201,0.0006315968,-0.249928534,0.1440658867,0.0977682769,0.6802558899,0.1865135878,0.4030945301,-0.0869980007,-0.0436691344,0.0363891684,-0.2932219505,-0.1100367904,0.2340059429,0.0753232017,0.2027742267,0.0346169695,0.1456870288,-0.2451830357,0.0854274631,0.0174441207,0.1328043193,-0.1475849748,0.3109081388,-0.1509093344,-0.2978287041,-0.1912623942,0.0007492236,0.1050974503,0.1029956341,-0.069436349,-0.2385252118,0.5177173018,-0.2855589986,-0.1767610461,-0.0903273001,0.3520115018,-0.0403811969,-0.1172361895,-0.5067312717,0.054644987,0.4887304604,-0.002228745,-0.1500933021,0.1638828069,-0.2067309022,0.0215018056,-0.0706165358,0.0114977323,-0.0409061126,0.000829379,0.1980810612,-0.1320041418]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"For that environment, I am sorry but I can't reproduce the bug: I can load the dataset without any problem.","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":20,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\nFor that environment, I am sorry but I can't reproduce the bug: I can load the dataset without any problem.","embeddings":[-0.2608940005,-0.2007066011,0.0660000592,0.400195688,0.4645994604,0.0115246382,0.3240968585,0.0941318348,0.1896000355,0.3001818657,-0.2849094868,0.4667589068,-0.0366437919,0.1668581367,0.0777649358,-0.1253727525,-0.2606472373,0.2636680901,-0.1777470112,0.0613667145,-0.5233634114,0.0231986102,-0.2373635471,0.0858903304,-0.2768001556,-0.1533993483,-0.0281501114,0.079713583,-0.2462176234,-0.2805505991,0.4611611068,-0.0159628261,0.1640476584,0.5104134083,-0.0001023479,-0.0184572823,0.5836995244,0.0530032516,-0.0761517882,-0.4211900532,-0.2255626768,-0.3977999389,0.2870038748,-0.3548148572,-0.1275422126,0.0055780676,0.0405133776,-0.1819047779,0.262139827,0.3204336464,0.3242153823,0.5296897888,0.0114422301,-0.3599714339,0.1965120882,0.0808303058,-0.0636900738,0.3710546196,0.0265845973,-0.0474929884,0.2439645529,0.0870305225,-0.3353538513,0.0527608916,0.2847363651,0.0382548533,0.0095302081,-0.270170182,0.0964757726,0.3694062233,0.3855828047,-0.382620275,-0.1514433473,0.1750092059,0.2239649594,-0.2032708824,0.211649254,0.0468775555,0.0163049828,0.0907445624,-0.0130718844,0.1712200791,-0.2445577234,0.2749375999,-0.2020063251,0.1065498963,-0.1935495585,0.1447966695,0.0849972516,-0.1188601851,0.0763338953,-0.1345976889,-0.0619969517,0.0890901908,-0.2083802223,0.0349296518,0.1152055487,0.1535798907,0.1636114269,-0.2843680978,0.1018852517,-0.0238335412,0.0397984199,0.4501174688,0.2414609641,0.1897560805,0.2724782526,0.088692598,0.2660125494,0.066787757,-0.2417135537,-0.0120890485,-0.1543584764,-0.0610723533,0.2427943051,0.1411847323,0.5124645233,0.0189152192,-0.4501633942,-0.1742377281,-0.172133401,0.0913005546,0.2716411948,0.4277633727,-0.1526402384,0.0820783153,0.1332417876,0.1743281931,-0.1884514242,-0.2918041646,-0.3183253706,0.3539678454,-0.2528480887,-0.2349549532,0.1911386847,0.0579657778,-0.0449788868,0.0134671535,-0.0958473086,0.1027007028,0.0886686146,-0.2908604145,-0.2310647517,0.3132710755,0.2443055063,-0.0628327206,0.2800607383,-0.4668283463,-0.091596745,0.2292028666,-0.2898909748,-0.1671735048,-0.0732437372,0.2667024732,-0.3390546739,-0.0697349459,-0.2441662848,0.0466724299,0.0489577651,-0.1566343755,-0.1024087444,-0.0962615982,-0.112656869,-0.4234251976,0.0446183719,0.5165780783,-0.3751866817,-0.1992407143,-0.3078093231,-0.1091819853,0.352284044,-0.147301212,0.0145670017,0.0629001632,-0.1706566364,-0.0006857364,0.6008648276,-0.3847208321,-0.2815137208,0.0563009419,-0.0371869579,-0.1165539101,-0.0517222472,-0.0470438264,0.0718874335,0.135642916,0.4128215313,0.2446535677,0.0493588299,-0.0324675739,-0.2109587342,-0.2381093949,0.1280453652,0.151658386,0.1272704005,0.1630408317,0.0712505579,-0.2506299913,-0.0557164401,-0.0061886362,-0.1022448465,-0.0301638599,0.5365030169,-0.1599618644,-0.0399641804,-0.532881856,-0.3950176537,0.2330813259,-0.0779283196,0.2541694939,-0.105489023,-0.1700357646,-0.352982074,0.1019510254,-0.0170240067,0.0607618615,0.2853280008,0.0909614265,-0.2011494935,0.1454881579,-0.0809026584,0.1637727916,-0.0873494446,0.0512318462,0.0335052349,0.291667223,0.0747782141,-0.3356175423,0.0352951288,0.1897852123,0.4574481249,-0.0104598496,-0.0672703311,0.2640901804,0.070892185,-0.0788056329,-0.2070860565,-0.05381817,0.0879925042,-0.313064307,-0.0303661674,0.0379851758,0.3187099099,-0.0943738967,0.0625385866,0.6038931608,0.0668431222,0.1744620055,-0.0730603263,0.2714183331,0.0653238744,-0.0385558642,-0.149077177,-0.1286280155,0.0459849499,0.1176183671,0.2486371994,0.0061038891,-0.0880215764,-0.0231885035,0.5769147873,0.0913832113,0.3607913256,-0.1375178844,-0.4204133153,0.1137951836,0.0370061696,0.2525543571,0.5032669902,0.332315892,-0.1068636402,0.023629915,-0.0571720079,0.0404726267,0.1032921076,0.0494471341,0.2328974009,-0.0000240203,0.3057212234,-0.0209420472,-0.3462260365,0.1613915414,-0.053847719,0.1842062026,-0.3760870993,0.1845387518,-0.2747030258,-0.076757893,0.1425513774,-0.1027247906,-0.125005424,-0.3286290169,-0.2664072216,0.1081375405,0.1210905612,0.1938206255,0.1422476321,0.0763199255,0.0087453714,-0.0255289134,-0.1826611906,-0.1326549202,-0.1483921856,0.0880739987,0.1653355807,0.181941092,0.4119828641,-0.1496375948,0.2284882814,-0.0511268564,-0.245598793,0.100177817,-0.0573291741,0.5299291015,0.4086084068,0.0602920428,-0.1410644501,0.1004512459,0.4188910425,-0.4658019245,0.0264790654,0.363583684,-0.3070645034,0.0837264657,-0.2518776655,-0.1060993373,-0.5478299856,-0.4280380607,0.0683310926,0.0958201066,0.0449768677,0.172417596,0.1191176772,0.444096446,0.112392433,0.0157600436,-0.2330423892,-0.0854077637,0.3311132789,-0.1693256646,-0.3608907759,0.2042488754,-0.1323478222,0.2548744977,0.0285864808,-0.1609783471,-0.5243912339,-0.1556174606,0.2053533643,-0.005188724,-0.0110916914,0.2912253737,0.0402853899,-0.1367866844,-0.2726752162,-0.1413018703,-0.0488153882,0.2352696806,-0.0683882236,0.0471838973,0.2825388908,-0.1223018318,0.1436379254,-0.1094394699,-0.1853011698,0.4059675932,-0.3138792515,0.3742715716,-0.1514657587,-0.4665456712,-0.2108906806,0.0278709959,0.0422425531,0.0484344922,-0.1828998625,-0.2429656833,-0.3873915374,-0.184057802,-0.3091237545,-0.1176473424,-0.0448231548,0.1869143099,-0.0344660766,0.1282412708,0.1506693661,-0.4099556506,-0.0440214574,-0.0145162204,0.1907593459,0.0003334501,-0.1874935776,-0.1343231499,-0.2394137084,-0.3349730968,0.466514647,-0.130417794,-0.0214519892,-0.1048111022,-0.0228138175,0.1187153012,-0.0458018035,0.5801782012,-0.2944788039,0.1719288677,0.5542984009,0.0904794037,-0.5516597629,0.1077317894,-0.331844449,0.3005692363,0.1331286132,0.2810209692,-0.3470429182,-0.0193871576,0.2158617675,0.0685858876,-0.1151087955,-0.2407927513,-0.2334081084,-0.2933899462,-0.0492701679,-0.0665732101,-0.0586115457,0.4225768149,-0.1354665607,-0.1273059845,-0.4123404026,-0.0367351063,0.1656020731,0.1939739734,0.3422520459,-0.1359105408,-0.0259811543,-0.0587919764,0.0830501467,0.1330539882,0.3568017781,0.0888106078,-0.4469739497,0.1390765905,0.0736007243,0.2392614037,0.0024299552,-0.2127716392,-0.1564223021,-0.0973184034,0.1256743819,0.0694909394,0.1411263794,0.2321473658,0.0137663297,-0.3646833301,-0.23113361,0.4325918257,0.1745893359,0.0250268523,0.2719003558,0.2346458733,-0.2254342437,0.2237933278,-0.0649300292,0.9164009094,-0.0016310633,-0.1102356687,0.6165456772,-0.0979906321,0.5674658418,0.0650797188,0.0679337382,-0.3704264462,-0.2879863977,-0.0052313902,-0.1453758329,0.2548450828,-0.0493531264,-0.5216739774,0.2098066807,-0.3250356913,0.0563870929,-0.0338756666,-0.0862727463,-0.0688716248,-0.1625674218,-0.3169096708,0.2203887552,-0.1160019636,0.1652479023,-0.0905514881,-0.0729376525,0.0201872438,-0.0793530717,-0.0859890431,0.0518678725,-0.4105747044,-0.1025814638,0.0088864788,-0.1489424556,-0.049235668,0.1097768545,0.0439088494,0.0238108411,-0.3483451903,0.2360750288,0.0525084697,0.0174742565,-0.0742553324,0.0752038509,0.2092088163,-0.0593716875,-0.2678528428,0.1072407961,-0.0471316501,0.167549476,0.3512031436,0.1351451129,-0.1057788879,-0.3018784225,-0.2150235772,-0.1387380958,0.2945331931,-0.2950992286,0.2529356182,0.0883854032,-0.1128398553,0.0237117633,0.0370561406,-0.1807857901,-0.0391988792,0.4054492116,0.3315434754,0.3266708255,0.1742117852,0.2139461488,-0.1330504417,-0.3562078476,0.2356424928,0.07449314,-0.3519974351,0.3284549117,0.2739292681,-0.0456543639,0.0271229651,0.4190958738,0.0177649334,-0.0406256802,-0.1278355718,-0.2977449894,-0.2052850723,0.2201710194,0.0076375799,-0.0133153005,0.1036312953,0.2353747338,0.1969152242,-0.0753166899,-0.4256239235,0.1184609309,-0.3420971334,-0.1064495519,0.0358140729,-0.0839432031,0.1991972327,0.094303377,0.2200210541,0.04124403,-0.1173671484,-0.3198773861,-0.1708855778,0.11573001,-0.0603523888,0.0984599441,0.1912232041,-0.0978017226,-0.3399778008,-0.1138171703,0.0863472894,0.0186832566,-0.2485669553,0.0389387496,0.1841407865,0.2960723639,0.0243147407,0.2214778364,0.0648288131,-0.0516980924,-0.1667803824,0.0070171691,0.0738608539,-0.0377257206,0.0453988388,-0.0242696367,0.146423474,0.1258281916,0.1472238451,-0.1821626425,-0.1376380622,-0.0106543619,0.1622682214,0.1054019704,-0.1678972542,0.3169001639,-0.151139766,0.3191595078,-0.3713175952,-0.0181264002,0.2591815293,-0.0037934883,0.2337247133,0.0487347431,-0.2177034914,0.2784507871,-0.100174889,-0.0086973989,0.4537391067,-0.1244415864,0.0648771003,0.238508001,-0.1500530988,0.0674678758,0.4041028619,0.2606096864,0.0971257016,0.1352504194,-0.0299772583,0.1296595037,-0.0773413926,0.0511166528,0.1697779894,-0.5580494404,0.186106354,0.1685796976,0.0023652951,-0.0791588277,-0.2323796898,0.4102463126,-0.4937512875,-0.0605658107,-0.3794310689,0.0643697307,-0.163028717,-0.0013760449,-0.1433874965,-0.1149196997,-0.1717617661,-0.2528249323,0.0416352265,-0.0990594178,0.2412425876,-0.1137952805,0.1046868265,-0.1930136085,-0.0399430357,0.278688699,0.0477989689,-0.1555451304,0.0398130342,-0.0601673685,0.25269261,-0.1166727543,0.4984675646,0.4282827377,0.265019238,-0.2017355114,-0.1184142306,0.0438968129,-0.294929415,-0.1482043564,0.1448875815,0.1621525288,0.1861361414,0.0748439059,0.2893256247,-0.284999311,0.1308235079,-0.0036748131,-0.0078192856,-0.2805196047,0.3590799868,-0.3625504673,-0.1178260595,-0.1554860473,-0.1072709784,-0.4793089926,0.098795183,0.3647942543,0.1218607798,0.1348394006,-0.1550246477,0.1431170255,-0.2600436211,0.2424018681,-0.0232327823,0.3653592467,-0.2398388982,0.0257782973,-0.5430261493,0.4481772184,0.0129255764,-0.0782645717,0.0848244727,-0.121702157,-0.1154602692,-0.0344456807,-0.0048418576,0.3630883992,0.2708312571,-0.0477743857,-0.5591887236,-0.3568104208,-0.0946177766,-0.2772197127,0.0461378619,-0.2907800674,-0.0501833335,-0.1604976356,0.1326508671,-0.0683541149,-0.2736573815,0.1152127311,0.2708913684,0.6669099331,0.2045687437,0.407979399,-0.0874733254,0.0932615772,0.07337749,-0.2984471619,-0.0754200667,0.1905672997,0.1010607332,0.136260882,0.0150285382,0.1734221131,-0.1472680867,-0.0376885645,-0.0404813625,0.0578768551,-0.2393028438,0.313811034,-0.2897048891,-0.1945980787,-0.2668882906,0.0952340364,0.084310323,0.1096617356,0.0218840633,-0.0939410552,0.6466474533,-0.2551367581,-0.3148619533,0.002964878,0.2552378476,0.0340935066,-0.1752844602,-0.4145071208,0.0628261641,0.443211019,-0.0549863242,-0.218359828,0.2427147627,-0.1774560511,0.0503365546,-0.1022099629,0.0111536505,0.0516408011,-0.0675834417,0.1296829134,-0.0830133334]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"One naive question: do you have internet access from the machine where you execute the code?","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":16,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\nOne naive question: do you have internet access from the machine where you execute the code?","embeddings":[-0.2466751933,-0.2166070193,-0.0038837458,0.3586860299,0.4009475708,0.035536956,0.302520901,0.1521895081,0.1650357544,0.3566101193,-0.1779196858,0.4452134073,-0.0382566825,0.1990447938,0.0816193596,-0.0899617374,-0.2033715993,0.2447062582,-0.1842225194,0.010521763,-0.472876668,-0.0185202286,-0.1821441352,0.0199327096,-0.2480738759,-0.2002862543,-0.1163984835,0.0645010024,-0.2148991227,-0.2530295849,0.3632104993,0.0450640544,0.112367928,0.5467213392,-0.0001019842,-0.0862105861,0.4975403249,0.0762401372,-0.1133384928,-0.2227149755,-0.2704135478,-0.3431518078,0.1823247373,-0.4260804653,-0.0916544795,0.028890904,0.0123812426,-0.2481914461,0.3053150177,0.3072139323,0.339708358,0.4426216781,-0.0190175083,-0.3322893977,0.2162288725,0.0421040244,-0.0323108882,0.362215519,0.0633602291,-0.0537314415,0.1767412126,0.1824111044,-0.2867747545,-0.0029550304,0.2504888773,0.0876018703,-0.0245356578,-0.2595610619,0.1022056341,0.2668823004,0.2856886089,-0.4520843923,-0.1245521009,0.1114801839,0.1043578833,-0.2367105186,0.1430763751,-0.0036112065,-0.0197007451,0.1060361415,-0.0351593755,0.133691445,-0.231664747,0.2396647781,-0.209117204,0.1492936164,-0.1894275993,0.0967105776,0.0752299801,-0.0852814317,0.0532211438,-0.1713415682,0.021075597,0.0105691114,-0.1507948637,-0.0472187474,0.1830180287,0.1402942985,0.1534052491,-0.254579097,0.1078871489,-0.0058753905,0.1186242029,0.4159373343,0.154383108,0.2415641695,0.2150634825,0.0880800635,0.2187133878,0.1302518398,-0.2568061352,-0.0115337186,-0.0923354477,-0.0363530554,0.3084871173,0.1605716944,0.4978731573,0.0148101076,-0.43729496,-0.0482422002,-0.2314789593,0.1063998938,0.2851623595,0.3828612864,-0.2197893709,0.0280763004,0.1954128742,0.1463474184,-0.167907998,-0.2896850407,-0.3228367269,0.3451814651,-0.3253240883,-0.1996293813,0.1255358905,0.0956025869,-0.078586556,0.0480372384,-0.035712216,0.0553010367,0.0763480887,-0.2473653257,-0.1706801951,0.1856051236,0.2096802443,-0.1042110175,0.3234429955,-0.4836687446,-0.1027234495,0.1058211848,-0.2070778757,-0.0896052122,-0.0359199308,0.3180175722,-0.3513986468,-0.1008676514,-0.151536569,0.1008075848,0.0612458289,-0.1549359858,0.0268286895,-0.0838796198,-0.0460206904,-0.443877846,0.0987593383,0.5773766041,-0.3923660815,-0.1583535075,-0.3149743676,-0.0410091579,0.3569568694,-0.2316169292,0.1328972578,0.1612990648,-0.1505951583,0.0307376627,0.5494579673,-0.3594247699,-0.298289597,-0.0194503367,-0.1195071116,-0.0845604688,-0.0080690281,-0.1045471951,0.0758646503,0.1502961367,0.4117963314,0.2450776398,-0.0244919565,-0.0403468348,-0.2356397361,-0.2162138522,0.0301526804,0.107242316,0.1855029613,0.160957098,0.1086669937,-0.2254877985,0.0150828492,-0.0165604558,-0.0678272471,-0.0664632171,0.5562682748,-0.1198019162,-0.1030108109,-0.548601687,-0.3364979625,0.2137777209,-0.0644096881,0.2583818138,-0.1622745097,-0.1771243662,-0.4169964492,0.1173746884,-0.0591691211,0.0415956117,0.3479426801,0.1584892571,-0.1316381693,0.1201645657,-0.038923569,0.1554773301,-0.0975586995,0.0615419187,0.002867274,0.3624258637,0.0288160499,-0.3402500749,0.0125437025,0.2390864491,0.492028594,0.0251288041,-0.0366603509,0.3425714076,0.0153975571,-0.0589240976,-0.0713611618,0.0129818767,0.0355378203,-0.3102554679,0.0903075784,0.0740944669,0.274332881,-0.0501489863,0.0455586836,0.5526733398,0.0639084801,0.1631180197,-0.0774783418,0.2747522891,-0.1043333039,-0.0652063265,-0.1591193974,-0.1482455432,0.1165027544,0.0529977977,0.1941261292,-0.0445721783,-0.102411747,0.0711867437,0.6148303747,0.07447882,0.410061866,-0.1350679696,-0.3626484573,0.1363148987,0.0632819161,0.2627360523,0.4373947978,0.3606951535,-0.0552522913,0.0725060552,-0.0956002772,0.0463584624,0.0818468109,0.0780013651,0.1710066646,-0.0235593766,0.2648818195,0.0000485086,-0.302197367,0.0393880606,-0.2115725577,0.1999685317,-0.3442276716,0.2703234255,-0.2761572301,-0.1083705425,0.0665704906,-0.1341651529,-0.1119984984,-0.3972260058,-0.189691484,0.0495734848,0.0846979097,0.1596637964,0.1331482828,0.1780963689,0.0435850658,-0.0506396033,-0.2169364691,-0.0929441899,-0.1491832733,0.124519974,0.1127889976,0.2195498496,0.4563970566,-0.2264528573,0.2653894424,0.0405263342,-0.2521411479,0.1091040745,-0.1113009229,0.5382031798,0.4217015207,0.0849712491,-0.1375489533,0.0479233935,0.4025655687,-0.3987693489,0.0345607921,0.2760668993,-0.1733168364,0.0993671566,-0.2214981169,-0.1091302112,-0.4937289059,-0.430516243,0.0535355024,0.0940166637,0.0907848328,0.1661356688,0.1647697985,0.5161489248,0.0903599858,0.0853421465,-0.2309448719,-0.0928328782,0.2652721107,-0.1356292665,-0.3170882165,0.1164372712,-0.1327296942,0.1882433891,0.0346632898,-0.2173748314,-0.5696399808,-0.1761069298,0.2179042101,-0.0343912281,0.0933210403,0.26999861,0.0949118435,-0.1569222659,-0.2845886648,-0.1625180691,-0.09149611,0.1541405916,-0.0318299644,-0.0249708015,0.2786458135,-0.1161085069,0.1320557296,-0.0665605962,-0.2075443119,0.3825691044,-0.3480911255,0.3984667063,-0.0996405408,-0.5200996399,-0.1736067683,0.0353287831,-0.0609325208,0.0489658341,-0.1714958549,-0.216622591,-0.3766874969,-0.0562461317,-0.3627006412,-0.1873274595,-0.0391595475,0.1560379118,-0.0368834846,0.0749153569,0.1888145953,-0.4491058588,0.0570053346,-0.0169513598,0.1537907422,-0.0894060507,-0.1648553312,-0.1185744405,-0.2468067259,-0.2861296535,0.4680056274,-0.1499364823,-0.056393221,-0.2046160102,-0.0660794526,0.0905419141,-0.0358795337,0.5689522028,-0.3503737152,0.1897085607,0.4768269658,0.1251972616,-0.4568893611,0.1350766271,-0.4355596006,0.2517589331,0.2185436338,0.2802282274,-0.3060492873,0.0191356,0.2399570048,0.1026442349,-0.1106582806,-0.2221298069,-0.221788317,-0.3183932304,-0.0803543776,0.0420141369,-0.0252923761,0.4193621278,-0.0983455256,-0.1282833368,-0.3264181316,-0.0256801695,0.2025925368,0.2596056759,0.2894299626,-0.1231492087,-0.1225850731,-0.0823648274,0.0792324618,0.1359493136,0.3235682249,0.0793575421,-0.3446296751,0.0907276347,0.0888884291,0.2781876624,0.0573618561,-0.1762590408,-0.1440508962,-0.0869545788,0.2175625265,0.0061369422,0.1769185215,0.2548579276,0.0897866562,-0.3811888397,-0.2307589352,0.4337727427,0.1766114533,-0.0381243378,0.2948857546,0.2082063407,-0.2366477698,0.226096943,-0.0221611597,1.0133308172,0.0526240282,-0.1635152251,0.5793563128,-0.1488889903,0.5672776699,-0.0199497938,0.0393971093,-0.3318083882,-0.2921088338,-0.0050110044,-0.1491662711,0.2573121488,-0.0939406306,-0.6075484157,0.1944481134,-0.3985604346,0.0609820001,0.0078523858,-0.1274123043,-0.1191684529,-0.1094275266,-0.3458572328,0.2644385099,-0.1378703862,0.1916236281,-0.1486469507,-0.0662537143,-0.081767492,-0.109817706,-0.1274234951,0.0437017307,-0.3669551611,-0.1849412918,-0.0509839356,-0.1615005583,-0.1080818698,0.0942901149,0.0813174099,0.1470959932,-0.2779953182,0.0880830884,0.0845176354,0.024237819,-0.0069385697,0.0085471412,0.1958828121,-0.0370822027,-0.2943578064,0.1068727002,0.0457872525,0.11517708,0.3087917566,0.1035760045,-0.0675774515,-0.3244170845,-0.1034835204,-0.0974463224,0.2481093854,-0.3131102622,0.2714842558,0.0891997069,-0.1201467961,0.0999452099,-0.0129307564,-0.1413194686,-0.0845809951,0.3065213561,0.2475549877,0.3110371828,0.0651139915,0.2337538004,-0.1451797485,-0.3778772056,0.1946107447,0.1266462058,-0.3094281554,0.2451846749,0.1132033244,-0.1002142429,-0.018966902,0.3583429158,0.0872674361,0.0121289548,-0.1319913417,-0.3177200258,-0.2275741249,0.1974657923,0.1262835264,0.0467932597,0.0409467742,0.2409867495,0.2160649598,-0.0149477664,-0.4777274728,0.0233572591,-0.35480088,-0.1210959181,0.0553099439,-0.131202504,0.2106182128,0.0307738371,0.2626608312,0.0611371771,-0.1230163798,-0.3603873551,-0.1196228266,0.0878794044,-0.0960554257,0.1212469712,0.1928565055,-0.0655948073,-0.2737081051,-0.1821471155,0.0655785874,-0.004897011,-0.1561484635,0.0742097273,0.1955157518,0.3057770133,0.0542224348,0.0768087059,0.1181700379,-0.0472166464,-0.0940910056,0.0504168384,0.0114915501,-0.091620788,0.0615547486,0.0473166667,0.1579477042,0.120423086,0.1462675333,-0.1426482201,-0.1761249602,0.0342858694,0.2458388805,0.1684488058,-0.1826538593,0.2451196015,-0.2031995654,0.368602097,-0.3018474579,0.0448667556,0.233446002,-0.0057722093,0.2976733148,0.1492141932,-0.2892270386,0.2251113206,-0.1543214321,0.0022290335,0.4208866656,-0.0848289132,0.0941332951,0.255379945,-0.1617472917,0.0726773366,0.4207955301,0.215765819,0.0808967128,0.1147266775,-0.0120389042,0.2092016637,-0.0595478676,-0.0088517275,0.1725978404,-0.4266685545,0.1862843037,0.1980996728,0.0223718807,-0.0381644145,-0.2197990566,0.415300101,-0.508476913,-0.0565226451,-0.3237141967,0.0487760864,-0.1741153598,0.0605982803,-0.1682130247,-0.1289482117,-0.2265689969,-0.2804172635,0.1312571615,-0.1609176397,0.2799320817,-0.1695612073,0.1022945419,-0.1938088834,-0.0666625947,0.2105489671,0.1612253487,-0.2031766772,0.0029694394,-0.0198347457,0.2243938595,-0.1444881409,0.4721126258,0.4282221496,0.2557366192,-0.1851379275,-0.0929748416,-0.0065720282,-0.2934385538,-0.1710650176,0.0933051482,0.175207302,0.1753773838,0.107457526,0.3313249648,-0.3341614604,0.0941621438,-0.023793228,-0.0138929961,-0.2267360091,0.3832151592,-0.2947542667,-0.0855249688,-0.1702571511,-0.1365145743,-0.5748503804,0.0405889116,0.3419461846,0.0930171311,0.1262921244,-0.1605507731,0.1524770707,-0.1748819649,0.1722018272,0.0148292333,0.4054068625,-0.258834511,-0.0027993866,-0.5410189033,0.4275760949,0.0155638382,-0.1588082761,0.0315136537,-0.0331475437,-0.1146663353,-0.017911721,0.022063138,0.3360385299,0.2262089252,-0.0256580152,-0.5814420581,-0.2901699543,-0.1248456612,-0.2796956301,0.1322259605,-0.2572001219,-0.0358443893,-0.2061264217,0.2160844207,-0.037930768,-0.2939603329,0.0869979188,0.1420842558,0.6847292781,0.177642867,0.4054076374,-0.1641926318,0.0316925645,0.0117772399,-0.2423882335,-0.1033019796,0.2090906501,0.0755653828,0.1736918837,0.0127599053,0.1335804164,-0.1275078207,0.0320263021,0.1104811877,0.0701001137,-0.1159140542,0.2584398985,-0.2325378805,-0.1188334078,-0.1874582469,0.0807324499,0.088373743,0.2076720297,-0.0338687859,-0.1781326681,0.6508178115,-0.1992694587,-0.3754538,-0.0056762015,0.3330244422,0.0251546893,-0.1825997382,-0.4330867231,0.0186706223,0.4222237468,-0.0488338508,-0.2132585198,0.2750219703,-0.2542455494,0.0392030962,-0.023348771,0.0766803846,0.0071151704,-0.0488673002,0.1620789319,-0.1435601711]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"> For that environment, I am sorry but I can't reproduce the bug: I can load the dataset without any problem.\r\n\r\nBut I can download other task dataset such as `dataset = load_dataset('squad')`. I don't know what went wrong. Thank you so much!","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":43,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\n> For that environment, I am sorry but I can't reproduce the bug: I can load the dataset without any problem.\r\n\r\nBut I can download other task dataset such as `dataset = load_dataset('squad')`. I don't know what went wrong. Thank you so much!","embeddings":[-0.2479815185,-0.3217719793,0.0763500705,0.4701036215,0.4953816831,0.0185086466,0.3540442288,0.0614947863,0.2402586043,0.3128617108,-0.276409179,0.4540319443,-0.0626192093,0.1033334509,0.1556583196,-0.1557852626,-0.2255954146,0.1679677218,-0.0779853985,0.1020552889,-0.4996086359,-0.0293463878,-0.2632772624,0.1951634139,-0.1559653133,-0.1863708049,-0.0737683624,0.0807150304,-0.1478742063,-0.2537325919,0.5681656003,-0.1083382666,0.2122284323,0.5568339229,-0.0001079977,0.0223296881,0.5475671887,0.0222912077,-0.181262657,-0.5533769727,-0.1755931079,-0.4152663052,0.3201848865,-0.3743572831,-0.1145967916,0.07909403,0.1459887475,-0.1117727757,0.3590354919,0.4540637732,0.2855804861,0.7018626928,-0.0197485741,-0.342117399,0.1333463937,0.0541447364,-0.1285513192,0.2871590555,0.1371662915,-0.0082335789,0.4334596992,-0.0373312868,-0.3705081344,0.090445593,0.3338054717,0.0173116531,0.0456819162,-0.4084408581,0.0989959762,0.3526202738,0.429132849,-0.2789865732,-0.1445523649,0.1062167063,0.2393290848,-0.061995618,0.1765795946,0.0757426322,-0.0298103616,0.0413528346,0.0463871136,0.1718294621,-0.2779862285,0.2125990689,-0.1284300834,-0.0353670754,-0.1932722479,0.2176218927,0.038036257,-0.0971409827,0.0310842935,-0.1765830368,0.0018496352,0.1451958716,-0.3074558675,-0.0370389931,0.1003599092,0.0206406731,0.2188021392,-0.2397266477,0.2141465396,-0.0814069584,0.0014896236,0.3541437387,0.346270442,0.2386882752,0.2158100009,0.1408458501,0.1517539918,0.022729395,-0.1789033264,0.0621397234,-0.2835336924,-0.0460709594,0.2485666424,0.2519290149,0.62191993,0.0226177946,-0.5257906914,-0.1946161389,-0.0697027519,0.0744651854,0.1489258707,0.4044239819,-0.2522181571,0.0734108984,0.0253312159,0.2898434997,-0.1949325502,-0.3826381862,-0.250428021,0.1526836008,-0.268599242,-0.270571053,0.24026227,0.0925443023,0.0312479846,0.0430606492,-0.2270833105,0.0859438777,0.1672883034,-0.3710010648,-0.3183163404,0.203129068,0.2981905341,0.013275166,0.2797108889,-0.3238164485,-0.0456261188,0.2975198925,-0.3877721429,-0.1205919012,-0.1263929754,0.1935602129,-0.3629306555,-0.1565003395,-0.5364289284,0.0805210099,0.0401180312,-0.0892851651,-0.0573687814,-0.1425406635,-0.0697560161,-0.3583665192,-0.0551820286,0.4761920869,-0.2147590965,-0.2288669497,-0.3011833727,-0.2443191111,0.3596111238,-0.1455561668,-0.0476030186,0.2456554323,-0.1385620087,0.0195395816,0.642037034,-0.5051142573,-0.2810824811,0.1532612592,-0.1181739345,-0.076440759,-0.0949228108,-0.0446148701,0.2863956094,0.0758701265,0.3716195226,0.3349784911,0.0008509188,-0.0305648036,-0.131107524,-0.2105619609,0.2909449637,0.1949024647,0.2361101955,0.2211598307,0.0545335151,-0.1998473704,-0.0358049236,0.0527693219,-0.0813176557,-0.0201197267,0.4355514646,-0.0436968356,-0.0257663354,-0.5290393233,-0.5686805844,0.1899500042,-0.0841612145,0.1826199889,-0.115729481,-0.0673960596,-0.3040428758,0.042624291,-0.078021042,-0.0217457879,0.182378903,0.0794131309,-0.2668655217,0.0558810793,-0.0770036429,0.1909017116,-0.1526100785,0.095781371,0.1212609336,0.2611952722,0.1016982198,-0.3987404406,-0.0017724648,0.1362133324,0.4781362414,0.0183626059,-0.0826574564,0.1729604006,0.0458101444,-0.1020453051,-0.2171081603,-0.1762046367,0.0575558543,-0.3366027176,-0.0447619446,0.1347156912,0.3460848331,-0.160481751,0.0836437568,0.4571962357,-0.0167928878,0.3421861231,-0.1437793821,0.2406296581,0.0525164865,-0.0606726483,-0.1331075728,0.1804715544,0.1290700138,0.0913747177,0.2827326059,-0.0355854295,-0.2706822753,0.0046682684,0.6162089109,0.0299597997,0.3194147646,-0.0831516907,-0.33788234,0.1462022662,0.0000325413,0.0418264307,0.5353927612,0.2632282078,-0.1103599295,-0.0164734125,0.0007065669,0.0515690558,0.0333776362,0.069090277,0.2856582701,0.0461304933,0.2759691477,0.0386635847,-0.2005325556,0.1444482505,-0.0084361024,0.1951348335,-0.3969622254,0.1878786534,-0.1423569769,-0.1165724248,-0.0304603409,0.1370230019,-0.1042248011,-0.1990451217,-0.3134959638,0.1589778066,0.2137130797,0.1265260726,0.1616361737,0.0088405497,-0.0113126803,-0.0408857763,-0.1634947658,-0.1458583921,-0.1861838251,0.0462163053,0.1946974099,0.0735260546,0.3495370448,-0.1900552809,0.1245678738,-0.1000485346,-0.1240719408,0.0189164113,-0.0774424672,0.5409654975,0.2683835626,0.2594893575,-0.162481755,0.1194115579,0.3791869879,-0.4782688916,-0.0512078144,0.3330359757,-0.334943682,0.0967640132,-0.2121215761,-0.3186377287,-0.5942174196,-0.3938197494,0.2054920793,-0.0129524982,0.069686912,0.3293658495,0.1026407331,0.4716491699,0.2177550197,0.0916105509,-0.1844632626,-0.0342024155,0.2794252634,-0.1346675754,-0.3884647787,0.1575419456,-0.0971030593,0.2852425575,0.0799094364,-0.1245648041,-0.2630615234,-0.0669115633,0.197574988,-0.1246708706,-0.1207072735,0.3231537342,-0.0845761523,-0.0469946526,-0.2115352452,-0.1213957146,-0.0482246727,0.3319038451,-0.0071725766,0.0945167914,0.3415936828,-0.1624031961,0.3309932351,-0.0868292972,-0.0393967107,0.4093084633,-0.3032317758,0.2776997089,-0.0965931192,-0.4811325371,0.0073162122,0.0563486181,0.0693825185,0.0789806917,-0.2715215087,-0.2479568273,-0.3175480962,-0.2529810071,-0.4196513891,-0.0761848167,-0.0940631106,0.1780696958,0.0090043806,0.1709620655,0.1889690459,-0.4353352487,-0.1018504202,0.0006975317,0.23761563,0.0146928616,-0.1129617095,-0.2612105012,-0.2340477854,-0.3139867485,0.345592469,-0.2557941675,0.083301276,-0.140028879,0.0895220935,0.1889620423,0.0027529181,0.7596032023,-0.1823073924,0.2290213704,0.38683483,0.2017732263,-0.6543713808,-0.0279390458,-0.2924177051,0.3361033797,0.0906111449,0.3959363401,-0.4269514382,-0.0443884917,0.1525252908,0.0314957127,-0.0939231589,-0.2425147295,-0.2687611878,-0.3051568568,-0.0087272041,0.0218200311,0.0090193832,0.4349669814,-0.1700991541,-0.1275983006,-0.4085874557,0.1602924466,0.1126146093,-0.0253704134,0.2526399791,-0.056815505,-0.0093024755,0.0635729954,-0.0077941241,0.1148457006,0.3582159579,0.0272814278,-0.4528546631,0.2615330219,-0.0211641658,0.2443993241,0.0709576383,-0.1279031187,-0.0851995796,0.0087838434,0.0780045241,-0.0141761946,0.0905466229,0.3425087929,0.0383048356,-0.3234171569,-0.4045231044,0.2545499206,0.3031949997,0.0401418433,0.3135276437,0.0083070444,-0.1673143357,0.0610586889,-0.1346745938,0.8996839523,-0.0605399869,-0.1017853916,0.6653998494,-0.2444549948,0.4785053432,0.0302947909,0.0903261378,-0.2928251624,-0.2700549364,-0.0155689307,-0.2568275332,0.2637600899,0.0815494508,-0.4304995835,0.3003613055,-0.291116178,0.2078086585,0.0213771164,-0.0544004999,-0.060865432,-0.1126743257,-0.3436706066,0.1629805863,-0.2664074898,0.2688083351,-0.1147507727,-0.14413324,-0.0331183821,-0.0753438398,-0.1297870278,0.1565764546,-0.3821193576,-0.1631519198,0.1158264056,-0.2006135285,0.0768847913,0.1819679886,0.0739785209,-0.10281232,-0.3268237114,0.1788987815,-0.0758843198,0.1364337355,-0.087842159,0.0000826565,0.2469836175,-0.1007041708,-0.33282125,0.1236411333,-0.0640807524,0.21982342,0.2899284065,0.180278793,-0.072996825,-0.2621715069,-0.2122358978,-0.1799507886,0.3643169701,-0.2759153545,0.1959189624,0.009252294,-0.1024515927,0.0800813809,-0.0507112592,-0.1253606826,0.0088543976,0.4873361588,0.4263882637,0.2872929573,0.2660657465,0.2743073404,-0.0806822106,-0.3562266827,0.2828985155,0.1164565831,-0.33517012,0.3059953153,0.1482208073,0.0116094733,0.0161372833,0.4952471256,0.0256405957,-0.0226025004,-0.0756149366,-0.2714717686,-0.294061482,0.211046204,0.0070540542,-0.0486660786,0.1137508452,0.1883585602,0.2880287468,-0.153834179,-0.3571674824,0.1202098504,-0.3974452615,-0.026375806,0.0220237225,-0.0883372873,0.2679443061,0.0743281767,0.1367656738,-0.0063327239,-0.1159371808,-0.2485037595,-0.2217644304,0.1363088489,-0.0278415494,0.1512162387,0.1515443176,-0.2163798362,-0.2809672058,-0.0129056145,0.0469002761,0.0201656688,-0.1744152457,0.0505833849,0.1910213977,0.3978153467,-0.0217826068,0.2827041447,-0.0067478046,0.0345397852,-0.1433786154,0.02376814,0.2030914575,0.0252375361,-0.2087370902,-0.0644891113,0.2508503795,0.1676190645,0.0940858424,-0.3053337336,-0.1231551021,-0.0976448059,0.0022103242,0.1308603883,-0.1723631322,0.2990866005,-0.1103164032,0.2701034248,-0.2779805362,0.0524252616,0.326957196,-0.1212088093,0.0678809881,0.0283454955,-0.3222197592,0.2012974918,-0.1126514226,0.1511472762,0.4406094253,-0.1247906014,0.0929894373,0.2174029052,-0.1162052229,-0.015790185,0.460701853,0.2020016313,0.0393340774,0.1420467645,0.0226332583,0.0366369858,-0.0482860841,0.1063207835,0.2827980518,-0.4781712294,0.0035832161,0.1094033644,-0.0017196483,-0.0950419977,-0.3151477277,0.5316855907,-0.4729452133,-0.1242428645,-0.2933817804,-0.0078632785,-0.188767314,0.1411719173,-0.1920025945,-0.1300992668,-0.0960251093,-0.1964515448,0.0039726589,-0.1452821344,0.2983595133,-0.1298789233,-0.0621965863,-0.1563254148,-0.0753151849,0.4235470593,-0.0483771041,-0.169693917,0.0714745149,-0.1513211429,0.2156936377,-0.0976124927,0.5807609558,0.332120806,0.3286684453,-0.2131840289,-0.1695679575,0.2545269728,-0.3181824982,-0.1070876792,0.1548376232,0.1585799307,0.1606116891,0.0785211027,0.2066705674,-0.2241433114,0.142018795,0.0436597429,-0.1447922885,-0.3188250661,0.4858053029,-0.3899853826,-0.1255370826,-0.127706483,-0.0744743422,-0.5568339229,0.1190805957,0.2711359262,0.2195585668,0.0988916159,-0.1955020726,0.0958702639,-0.3225300908,0.2215004414,0.020388402,0.3648219109,-0.3144544363,0.1156628281,-0.5239733458,0.4592312872,-0.0107843764,0.0105118509,0.1541946232,-0.0931218565,-0.2654405832,-0.0039149406,0.030693708,0.2047602385,0.399890393,0.1453907788,-0.5745072961,-0.3909095824,-0.0477377437,-0.0727452859,0.0916486382,-0.2621004581,0.0129266009,-0.1517878324,0.023295382,-0.1117513552,-0.3203566968,0.1511974633,0.1406897157,0.7023329139,0.0446438454,0.4411260486,0.028585989,-0.011712499,0.1216360629,-0.3079591691,-0.105045788,0.176432088,-0.004793454,0.0528662838,-0.020993961,0.2726189494,-0.2175622582,-0.1812321395,-0.1439599544,0.1281732172,-0.3053524792,0.311968416,-0.3086757362,-0.1393731236,-0.3422217667,-0.0057940949,0.1176716611,0.0452723876,-0.1081117392,0.0493608788,0.6277851462,-0.1653215438,-0.3440380692,-0.1038250551,0.3128980398,-0.0359871238,-0.112662144,-0.5198081136,0.0568199977,0.4852424562,-0.0689152628,-0.1922020912,0.2341876179,-0.1415186673,0.1324385554,-0.0214217473,-0.0941569433,-0.0095509402,-0.0314085595,0.2175465375,-0.1435788721]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Hi @severo, thanks for reporting.\r\n\r\nJust note that currently not all canonical datasets support streaming mode: this is one case!\r\n\r\nAll datasets that use `pathlib` joins (using `\/`) instead of `os.path.join` (as in this dataset) do not support streaming mode yet.","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":41,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nHi @severo, thanks for reporting.\r\n\r\nJust note that currently not all canonical datasets support streaming mode: this is one case!\r\n\r\nAll datasets that use `pathlib` joins (using `\/`) instead of `os.path.join` (as in this dataset) do not support streaming mode yet.","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"OK. Do you think it's possible to detect this, and raise an exception (maybe `NotImplementedError`, or a specific `StreamingError`)?","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":19,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nOK. Do you think it's possible to detect this, and raise an exception (maybe `NotImplementedError`, or a specific `StreamingError`)?","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"We should definitely support datasets using `pathlib` in streaming mode...\r\n\r\nFor non-supported datasets in streaming mode, we have already a request of raising an error\/warning: see #2654.","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":27,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nWe should definitely support datasets using `pathlib` in streaming mode...\r\n\r\nFor non-supported datasets in streaming mode, we have already a request of raising an error\/warning: see #2654.","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Hi @severo, please note that \"counter\" dataset will be streamable (at least until it arrives at the missing file, error already in normal mode) once these PRs are merged:\r\n- #2874\r\n- #2876\r\n- #2880\r\n\r\nI have tested it. \ud83d\ude09 ","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":40,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nHi @severo, please note that \"counter\" dataset will be streamable (at least until it arrives at the missing file, error already in normal mode) once these PRs are merged:\r\n- #2874\r\n- #2876\r\n- #2880\r\n\r\nI have tested it. \ud83d\ude09 ","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Now (on master), we get:\r\n\r\n```\r\nimport datasets as ds\r\nds.load_dataset('counter', split=\"train\", streaming=False)\r\n```\r\n\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\nThe error is now the same with or without streaming. I close the issue, thanks @albertvillanova and @lhoestq!\r\n","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":191,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nNow (on master), we get:\r\n\r\n```\r\nimport datasets as ds\r\nds.load_dataset('counter', split=\"train\", streaming=False)\r\n```\r\n\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\nThe error is now the same with or without streaming. I close the issue, thanks @albertvillanova and @lhoestq!\r\n","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Note that we might want to open an issue to fix the \"counter\" dataset by itself, but I let it up to you.","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":23,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nNote that we might want to open an issue to fix the \"counter\" dataset by itself, but I let it up to you.","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Fixed here: https:\/\/github.com\/huggingface\/datasets\/pull\/2894. Thanks @albertvillanova ","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":5,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nFixed here: https:\/\/github.com\/huggingface\/datasets\/pull\/2894. Thanks @albertvillanova ","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2860","title":"Cannot download TOTTO dataset","comments":"Hola @mrm8488, thanks for reporting.\r\n\r\nApparently, the data source host changed their URL one week ago: https:\/\/github.com\/google-research-datasets\/ToTTo\/commit\/cebeb430ec2a97747e704d16a9354f7d9073ff8f\r\n\r\nI'm fixing it.","body":"Error: Couldn't find file at https:\/\/storage.googleapis.com\/totto\/totto_data.zip\r\n\r\n`datasets version: 1.11.0`\r\n# How to reproduce:\r\n\r\n```py\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('totto')\r\n```\r\n\r\n\r\n","comment_length":20,"text":"Cannot download TOTTO dataset\nError: Couldn't find file at https:\/\/storage.googleapis.com\/totto\/totto_data.zip\r\n\r\n`datasets version: 1.11.0`\r\n# How to reproduce:\r\n\r\n```py\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('totto')\r\n```\r\n\r\n\r\n\nHola @mrm8488, thanks for reporting.\r\n\r\nApparently, the data source host changed their URL one week ago: https:\/\/github.com\/google-research-datasets\/ToTTo\/commit\/cebeb430ec2a97747e704d16a9354f7d9073ff8f\r\n\r\nI'm fixing it.","embeddings":[-0.294410497,0.3900933862,-0.1317222118,0.0641968921,0.4353683293,0.1393284649,0.1679639369,0.5879850388,-0.0261453092,0.234341681,-0.1947766393,-0.0473091528,0.0594171956,0.318708539,0.0643599778,-0.3084107637,0.0652796775,-0.0130061815,-0.1395519972,0.0308540575,-0.0759798363,0.2782622278,0.0100256903,0.0654641986,0.0181091316,0.0936704874,-0.0290926713,-0.0979447067,-0.2686378956,-0.4347373247,0.415338546,0.0613755472,0.1697377264,0.3808855712,-0.0001073767,0.1472330391,0.3291011751,0.0086168032,-0.2500196695,-0.3531953394,-0.3288857937,-0.1141764298,-0.1965035796,-0.1480810791,-0.0830533355,0.2127362341,0.2038826346,-0.2607125342,0.0417160429,0.5573630333,0.2718278766,0.0907747969,0.3079364896,-0.3647713363,0.2657265365,0.1475394666,-0.0365158468,0.0705632791,0.0785895288,0.0108668972,0.3562817872,0.0460000634,-0.133997038,0.0743308142,-0.1553672999,-0.0840224102,-0.0924824327,-0.3424230218,0.2141403407,0.3073774278,0.5096561313,-0.2044652998,-0.3611649573,0.2267780751,-0.0598624535,-0.279548645,0.2698038518,0.1802222282,0.0900912583,0.1698183566,0.1732570976,-0.4855846167,-0.1890135705,0.0528651476,-0.4141792953,0.107604593,-0.1177637801,0.0048586847,-0.105921559,-0.1203932837,0.3016455472,-0.0547915176,-0.0375608727,0.2590126991,-0.2745456994,-0.1841468364,0.1035571247,0.0311484244,0.1920053959,0.0490685366,0.1362317801,-0.1017651632,-0.5062013865,-0.0666135922,0.2418693751,-0.1095379815,-0.000267334,0.3804118037,0.2949267328,0.2416359335,0.1484333426,0.0473140553,-0.1586337984,-0.358691752,0.0269331522,-0.113104403,0.4928361177,-0.2085420638,-0.3899927437,0.1616978347,-0.090081729,-0.1214746758,-0.0176121797,0.1517945826,-0.0244206693,0.0753581598,0.0918698311,0.1404550672,-0.008870828,-0.1377051771,-0.2661000192,0.2041446567,-0.0125199547,0.0713024735,0.2309996784,-0.3448672891,0.0952998549,-0.0258917771,-0.4624833763,0.0828349143,-0.1361531913,-0.0750162899,-0.2458771169,0.3632289171,0.2265905589,0.1045650393,0.0013822372,0.1134399772,-0.2949780524,0.2792558968,-0.3678761125,-0.0678893328,-0.1227772906,0.2941953242,-0.3227276504,-0.2466399968,-0.5071802735,-0.0339648984,-0.0525038354,-0.3056934178,-0.1234595478,-0.1745043397,-0.1805351675,-0.2515176237,0.1124921739,0.6013523936,-0.2063911855,0.020885611,-0.4196625054,-0.4201267958,0.1899637729,0.269320339,-0.0341406576,0.4477233887,-0.3025847673,0.0885180831,0.4392078221,-0.1254837364,-0.7618820667,-0.0033206183,-0.2596687675,-0.1898869723,0.1790016443,0.0790314004,0.2497510761,-0.1000057384,0.2460572869,0.1370749325,0.012917717,-0.0459471196,-0.2387001067,-0.0105643263,-0.0045891618,0.1839413047,0.1048776805,0.2050982267,0.1501675397,-0.0725682601,0.2357553095,0.1607331038,0.0994152948,0.3408914506,0.3313904405,-0.0265133232,-0.0199446026,-0.1835746318,-0.1992596537,0.1232289374,-0.2162778378,0.1084155291,-0.2086808383,-0.1304647177,-0.4467419088,0.0594626628,-0.0653845966,-0.093042925,0.1499165893,-0.0002756922,0.310736388,0.3321071863,-0.02035648,0.320682168,0.2677029073,0.1030120254,-0.1375443637,0.5665908456,-0.1339783072,0.0568048991,0.4536594152,-0.2412828058,0.236403808,-0.0816816241,-0.0623007454,0.1545941979,0.0604691543,0.3422275782,0.4771338403,0.2396415472,0.1721659005,-0.4354346693,0.2116640955,0.343268007,0.0783024356,0.1411449164,-0.1909071058,0.0213618409,0.1539851874,0.0350163281,-0.0176072661,0.2682003379,0.544712007,0.0763235763,0.0633015558,-0.0642063022,0.2679164112,0.5545850396,0.2224782109,-0.0677626878,-0.2419174314,0.1457102746,0.538705945,-0.0372180901,0.0527861379,0.2355953008,-0.1476681978,-0.0631909966,-0.1953097135,0.2097867578,0.1718096882,0.0596293472,0.1565221548,0.0106348377,-0.0777703449,-0.2667880654,0.0551331863,0.0725212246,0.1452946216,0.1418423057,0.1219759658,0.2061336339,-0.08200939,-0.3087290227,0.0309936032,0.3056665659,-0.1311020702,-0.0795388594,-0.2592677772,0.0049411356,-0.2732787728,-0.3159630299,-0.2113306969,-0.3952906728,0.0605420209,0.2657611072,0.0656001344,-0.0102982745,-0.0560371764,0.0340000875,0.2436152548,-0.3098428249,-0.1341092587,-0.150049597,-0.1041894853,0.170614779,0.2377597988,-0.3077156246,0.3982344866,-0.6368309259,0.1199382693,-0.5480054617,-0.234621942,-0.043117255,-0.1899068207,0.0942285731,0.175614506,0.4022015929,0.3427299261,0.2505168021,0.1937542409,-0.0247938484,-0.0510064363,0.0891098529,-0.1387198567,-0.0441237539,0.1827340573,-0.4613678753,-0.3021485507,-0.2280495763,0.0444356278,0.0629076734,0.2366307676,-0.2858436108,0.1709893048,0.0659947768,0.1328113973,-0.3080635071,-0.1222961098,-0.1796017885,0.4985628128,-0.2371805012,-0.2333060801,0.1059011444,0.130261451,0.4757525921,0.1674677283,-0.4497697651,0.179243952,-0.3459710777,0.1534799039,-0.0468465574,0.0758880824,0.3201242983,0.1453730166,-0.1137198061,-0.0569933727,-0.1610036939,-0.050038822,-0.3855092824,0.2330114543,-0.0367576629,0.2503889203,-0.1210072637,0.398199141,0.2551098466,0.07454741,0.3040669858,-0.1943264455,0.1306042373,-0.1279257536,-0.1889470071,-0.0984253138,0.1577033997,-0.1163241193,-0.1967870146,-0.0184634868,0.0741213709,-0.1694130898,-0.1042783111,-0.2042790651,-0.2836247385,-0.1303623915,0.0177816842,0.4172818065,0.185910061,0.0156836808,-0.1674310863,-0.0935420543,0.2443633378,0.1396629214,-0.0014964864,-0.018045662,-0.1251368523,-0.1843106151,-0.6704334021,0.3441170454,-0.1174282879,0.2353997678,-0.0927243307,0.185250476,0.1104813442,-0.0463375226,0.5471497774,0.062537156,0.3733004928,0.3563941717,0.2766824961,-0.146045804,-0.0525100045,-0.1304521114,0.2965848744,-0.0263616275,0.3771854639,-0.2650294006,-0.0554963201,0.2222447246,0.1888717711,-0.1031564474,0.1805047095,-0.0639477074,-0.4556207061,-0.2093116641,-0.1772014797,-0.0870981961,0.2738859355,-0.1530112326,-0.0872611031,0.0672675371,-0.147369504,-0.0175553579,-0.1686567217,0.381292969,-0.1637212038,0.177162528,-0.0096636331,-0.2981503308,0.5925562382,0.9116903543,-0.0094228163,-0.4505523443,-0.0699971095,0.1041783616,-0.1924929768,0.0370545797,-0.1077148691,-0.155616954,-0.1347660273,0.121474579,-0.1697997302,0.2179342657,0.2668156624,0.0785118788,-0.4189471304,-0.3319608569,0.407774061,0.1675906181,-0.0422660559,0.4756697118,0.1231786013,0.0052154721,-0.1512746066,-0.146587804,0.7076441646,-0.1708522737,0.121878989,0.3979227841,-0.1670890003,0.2951293588,0.0539895035,0.054950837,-0.356028378,0.2241553664,-0.0507987142,-0.0956090316,0.0422886536,0.0331598967,-0.2261527926,0.3760743439,-0.1912707835,0.0057616672,0.2342994213,0.2299364954,-0.2325203121,0.1433078796,-0.2494061738,0.2325792462,-0.1700909585,0.5963363051,0.012605018,-0.1578868926,-0.2352524251,-0.0887922347,-0.4586822689,0.2226306796,-0.3531464338,0.0472171605,-0.1291957051,-0.2200131863,0.0645607188,0.1586482823,0.3044909239,0.2362559736,-0.5237365365,0.2670378089,-0.3025308251,-0.3513713479,0.172851041,0.2117986232,0.2136156857,-0.0355904885,-0.2754929662,0.1930139661,0.0165787246,-0.1336362511,-0.1597443968,-0.3023325503,0.0960842669,-0.2871341109,0.0308847092,-0.1784550101,-0.0639164969,-0.2592556775,0.210878104,0.2444291562,0.1790868342,-0.0770321116,-0.2008534521,-0.2742711306,-0.1253176183,0.4576685429,-0.0899077132,0.0362664238,0.3003350198,0.1418553889,-0.0421103351,-0.2370263934,-0.0758946836,0.0870876238,-0.0705426484,0.2010704726,-0.246197015,0.4602616429,-0.1983534098,0.0310583059,0.0072811441,-0.0602627471,-0.0174485762,-0.4184100926,-0.0034990332,0.1399504542,0.068998836,0.0080933636,0.0801748037,0.33794415,0.0753722563,-0.2415237427,-0.3141222596,0.1533349752,0.010092549,-0.2602766156,0.3766644895,0.1265031993,0.2459721863,-0.1016764343,0.1848554462,-0.2197219133,-0.2210969478,-0.2645950317,0.0000300764,0.0581651703,-0.0433961898,-0.0884144753,0.1235888079,-0.3815771639,-0.2894198596,0.2447361052,0.0846430287,0.4267973006,-0.0242630001,-0.062661536,0.1742569208,0.0256446283,-0.0715736002,0.0522936583,-0.0506931208,0.0010746485,-0.0242586527,0.3078612983,0.0098361075,-0.1488315761,-0.3080678284,-0.012468027,0.069576934,-0.0835743919,0.4794929922,-0.238189742,0.0586935356,0.3324653208,0.1315368712,0.3096744716,-0.5152159929,0.1130948365,0.0408594236,0.286698699,-0.3299861252,-0.1299768537,-0.2388359457,0.0133060208,-0.1226402,-0.023742402,-0.0654749423,0.1589624137,0.2537431121,0.1377555579,0.3771550059,-0.1948017776,0.2289052606,0.309329778,0.1026867107,0.1862534583,0.2217633277,0.0799650326,-0.096968554,0.4350672066,-0.2293919921,0.1500754207,-0.0677117333,0.0790822133,0.1858036816,-0.2016682029,-0.3423802555,0.2099467218,-0.1070173159,-0.0416853353,0.0382248834,0.2646930814,-0.3808471262,0.2599398196,-0.2471588105,0.2180426419,0.0396916568,-0.107037656,-0.0710211098,-0.277212888,-0.1615739018,0.1111863405,0.0544033013,-0.0574874617,-0.1172462255,-0.0004238633,-0.4112986624,-0.4172866046,0.0453437679,0.1922140568,0.1519670635,-0.1251610518,0.4158020914,-0.0666866601,-0.2817391157,0.1539526582,0.4888014793,0.4909828007,0.1652413905,0.2514764071,-0.1447680593,-0.2571274042,-0.2132361531,0.0702499524,0.1148378253,-0.0334707685,0.0071211583,0.2830833197,0.2091301531,-0.1540863514,0.1244486645,0.0029568451,0.34194839,0.0563798025,0.3701932728,-0.3759686649,-0.0931663662,-0.2818811238,-0.0891278237,-0.3936681449,-0.1934796274,0.1249653772,0.0378298834,0.0091357706,-0.2682196498,0.1284017563,0.0321747214,0.1845884472,0.2696240544,0.2058680654,-0.2469167262,-0.3324057162,-0.5042065382,0.1490001231,-0.1049715057,-0.2146742195,-0.1411429644,0.0215014555,-0.2108839601,0.0625294596,0.2979361415,-0.0173048489,-0.1892402917,-0.1613342911,-0.1859367937,-0.0138215581,-0.175954029,0.0314951167,0.0316190124,-0.3335469961,0.2791323066,-0.348980993,0.0721447468,-0.202991724,0.058277525,0.0845616013,-0.0209053755,0.2773117125,0.1782321632,0.1633214355,-0.0716613159,-0.1242836416,-0.4604788125,-0.0752218738,-0.1951590627,0.0604053661,-0.1034028381,0.0961736962,-0.1812672317,-0.4262296855,-0.1299846321,0.0591744781,-0.151055932,0.0678856224,-0.1810762137,-0.0299206339,-0.0501213819,0.1820138097,0.1748444289,0.0471048988,-0.0015658198,0.0969647095,-0.2998532355,-0.2840430439,0.1638270468,0.0188097768,-0.0074548773,-0.0921064466,0.048402749,0.2315627784,0.064388819,-0.1891824305,0.2362127751,0.4156240821,0.009564341,-0.1696297377,0.1937181801,-0.111114867,0.1457280368,0.1789949685,0.4593870044,-0.0406309515,-0.1174679548,0.0221763458,-0.0482680798]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2859","title":"Loading allenai\/c4 in streaming mode does too many HEAD requests","comments":"https:\/\/github.com\/huggingface\/datasets\/blob\/6c766f9115d686182d76b1b937cb27e099c45d68\/src\/datasets\/builder.py#L179-L186","body":"This does 60,000+ HEAD requests to get all the ETags of all the data files:\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"allenai\/c4\", streaming=True)\r\n```\r\nIt makes loading the dataset completely impractical.\r\n\r\nThe ETags are used to compute the config id (it must depend on the data files being used).\r\nInstead of using the ETags, we could simply use the commit hash of the dataset repository on the hub, as well and the glob pattern used to resolve the files (here it's `*` by default, to load all the files of the repository)","comment_length":1,"text":"Loading allenai\/c4 in streaming mode does too many HEAD requests\nThis does 60,000+ HEAD requests to get all the ETags of all the data files:\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"allenai\/c4\", streaming=True)\r\n```\r\nIt makes loading the dataset completely impractical.\r\n\r\nThe ETags are used to compute the config id (it must depend on the data files being used).\r\nInstead of using the ETags, we could simply use the commit hash of the dataset repository on the hub, as well and the glob pattern used to resolve the files (here it's `*` by default, to load all the files of the repository)\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/6c766f9115d686182d76b1b937cb27e099c45d68\/src\/datasets\/builder.py#L179-L186","embeddings":[-0.5428535938,-0.0302166119,-0.1265247166,0.3286388516,0.256604135,-0.0104140276,0.197608754,0.3729439676,0.1844934821,0.2685716748,-0.1927886903,0.3850267828,0.0771837533,0.2340062857,-0.143449083,0.136008963,-0.1367563605,0.3763958514,0.1254129112,0.1078248546,-0.0619285703,0.0006759175,-0.0417456739,-0.0773263052,-0.2133178711,0.2759331763,0.1665377468,0.2257279307,-0.0134929046,-0.4966956973,0.240168795,0.2341412455,-0.0487481132,0.2447642982,-0.0001018645,0.1439338028,0.4033434391,-0.0089729438,-0.085589841,0.4268731475,-0.4157178402,-0.0739421472,0.0224096645,-0.3959259391,-0.2034465969,-0.2251017988,0.0632516742,-0.0980081558,0.2041195333,0.2827344537,0.260243088,0.4249944985,0.0377698317,-0.1970410794,0.2039705664,-0.316323489,0.0617868043,-0.056926813,0.0797685385,0.1593059301,-0.0834435821,0.2483358234,-0.0834733918,0.0802756175,0.2913605869,-0.1320183277,-0.0032377529,-0.2171800733,0.3939404786,0.2030942291,0.3443941772,-0.3286153674,-0.1332122386,-0.2533044219,0.0619671606,-0.1955298185,0.0476448163,0.1538827866,-0.1587007791,0.1192615852,-0.0326113962,-0.0544691943,-0.1602670252,-0.0494049191,-0.0776140019,0.089980714,-0.0840734094,0.1207940727,0.136279732,0.058017496,0.0522075854,-0.2175264806,-0.0846055672,-0.004862661,-0.3694218993,-0.0516511574,0.3653528988,0.4273298979,0.1478993595,0.1687032282,-0.1197926477,0.2871201038,0.0435164198,0.0685389638,0.0575072095,0.0231996793,-0.0176825654,-0.389597863,0.5508090258,0.0899150446,-0.1071902066,-0.2005494833,-0.0709187463,-0.1060602814,0.0840810388,-0.1210096851,0.1117649302,-0.2036800086,-0.0995911956,-0.1678549349,-0.0444756113,0.0720648617,0.1929066479,0.4125182927,-0.1070380881,0.2573559582,-0.2619225979,0.0917120874,-0.1934460849,-0.0925472528,-0.3160288334,-0.2928213477,-0.1694728285,0.0569894873,0.1707771719,-0.34016487,0.4697290957,-0.0191245824,0.3383053243,0.0339444056,0.0842757225,0.1081046611,-0.0809869394,0.4188500047,0.2263165563,-0.0787535161,-0.0641879365,-0.1923369914,-0.291814357,-0.1569433808,0.0312824622,-0.4784304202,0.1928419471,0.2877151072,-0.0427191779,0.1639586836,-0.3409174085,0.0171625558,0.0027538103,0.1268018633,-0.1082377136,-0.0361211002,0.1082094535,-0.2483724952,0.1618676782,0.4884290099,-0.0835868865,-0.0830648839,-0.039552927,-0.1286983341,-0.0772875845,0.3747945726,-0.3466023803,0.120490931,-0.0308591686,0.0409644805,0.032458812,-0.425350368,-0.6205683351,0.1140533686,0.0639335662,-0.0276992563,0.2386657745,0.3460126519,0.1039455682,-0.107216388,0.0222086739,0.1709899902,0.073792845,0.2256556302,-0.1190583929,-0.2345744073,-0.2999114394,0.3787722886,-0.1165996566,-0.1365634799,0.1347563714,-0.1876992285,0.2526454926,-0.3332684338,0.0380112678,-0.1584161669,0.0938546658,-0.0603501089,0.0127826966,0.2236943841,-0.4421748519,0.3059828579,0.2439787388,0.1666173935,-0.1472782493,-0.3505738974,0.0228462406,-0.0090298792,-0.3510281146,-0.0372455232,0.1835951656,0.0712855086,0.0677272379,-0.1504935324,-0.2270190865,0.2682373822,-0.3091340661,-0.0179539192,-0.4381204545,0.1312066019,0.0892122611,0.0468895361,0.2672764957,-0.1050158292,0.0074687083,0.0492864773,0.0621649623,0.1636259705,-0.0330553278,0.4530902505,0.0818722695,0.7110303044,0.1389819533,-0.2826481164,0.3203724325,0.2067766637,0.1654448509,-0.2394565195,-0.1613846868,0.4084748626,0.3778287172,0.3252526224,0.1585008949,-0.0329761878,0.410119772,-0.0721899271,-0.1317404956,-0.0330927819,-0.00823195,-0.0209802482,0.3309631944,-0.023005046,-0.2523918152,0.3234723508,0.3031021953,0.1111327931,-0.1492699087,-0.0652219951,-0.298992157,-0.1263385117,0.2080144435,-0.0855184197,0.2448243797,0.4307595193,-0.0929894522,0.1461498737,0.0120118773,-0.256077677,0.3098646998,0.0617690906,0.0520799011,0.0422569439,0.1629692465,-0.2966534495,-0.4100134969,-0.3106255233,-0.1572241336,0.1993097663,-0.2276444584,-0.0174489729,-0.0920694247,-0.3194893003,0.0963194221,-0.0115835099,0.0477257594,-0.2335455716,0.0685858503,0.2594055235,-0.2168204486,0.1381093711,-0.1156482473,0.2903750539,0.0892408416,-0.1132856086,-0.2681921422,-0.2419697493,-0.1095879748,0.1148864031,0.4646658599,0.0141319409,0.3488751948,-0.3224317729,0.1775819212,-0.3888134956,-0.1769560575,0.1015825272,-0.1227035671,0.0946206078,0.0673972368,0.1070112586,0.2800557911,-0.125384286,0.0769697279,-0.2286968827,-0.0885543525,0.1452503055,-0.154563725,-0.2497059852,0.0414588526,-0.3924611509,-0.0118674375,-0.3728982508,0.342957437,0.2925284505,0.2317622006,-0.0143444026,0.0886501297,0.003775317,0.0996953025,-0.221465528,-0.2427807301,-0.3542366624,0.1676867455,-0.1924644858,-0.1902693361,0.0533196107,0.1380496174,0.4981618226,-0.0522706732,-0.4013659656,-0.0050968095,-0.4799565375,0.4729455709,0.0018389857,0.1385525167,0.2309796959,0.1050632596,-0.2131042331,-0.003395986,0.0059769182,0.1519519687,-0.0271408688,-0.2601141036,-0.0332135223,0.1607500166,-0.1440254897,0.287691772,0.401422888,-0.1580083966,0.2020960897,-0.1712821573,0.3033681214,-0.4020245373,-0.0107392669,0.1449472159,0.1542857438,-0.1868642867,0.3402228355,0.1952575743,-0.5144734979,-0.2171178758,-0.1385502368,-0.047275681,-0.3612928391,0.0551773384,0.3035261631,0.0228444245,0.0812270045,0.2058289647,-0.0380034633,-0.1516912431,0.2849894762,0.0916408524,0.0076755234,-0.2801522613,-0.5164752007,-0.0290395562,-0.7274724245,0.4155051708,-0.1272853166,-0.1209579483,0.1984863281,-0.1490302235,0.1140256822,-0.1446226537,0.3893289268,-0.360753417,0.0653221086,0.1890885085,0.2773555815,-0.2066984922,0.0153291468,-0.1949689388,0.2554828525,0.0409896933,0.4853951335,-0.3089270294,-0.2090707272,-0.1388294697,0.1643549055,-0.1718056649,-0.1440299302,-0.2748757303,-0.2858329415,-0.2627595961,0.1193227097,-0.0391619354,0.003167425,-0.0440098308,-0.0841071829,-0.1187053919,0.1033410579,0.1640647948,0.4710362256,0.1158270538,0.0934364945,0.1659183055,0.0374369659,-0.1580857188,0.2071107477,0.6684245467,0.1755816042,-0.2218962312,-0.1182288826,0.1617062241,0.1396004111,0.3151801229,0.1193937883,0.0207385924,-0.1292494237,0.3238544762,-0.3029144704,0.193803668,0.1370225847,-0.0925901309,-0.2631414235,-0.3615525663,0.4471847117,-0.1883913875,-0.1312835515,0.4696905613,0.3428766429,-0.1417624652,0.2846679986,0.2487651408,0.8667517304,-0.0928715691,0.1275847405,0.2757603824,-0.3555955291,0.1609345227,-0.3366043568,0.0796381533,-0.351541996,0.1143317819,0.1026704684,0.0138060609,-0.17082645,0.1024274305,-0.1321756542,0.2881008685,-0.1443882436,0.3143622875,-0.1412631124,0.435343951,-0.3882615864,-0.4119451344,-0.6139753461,0.2683770061,0.3304165006,-0.1537439227,-0.0472401492,-0.0246126465,0.195830375,0.0197973438,-0.0137850698,0.2106341124,-0.1912306547,0.092871733,-0.275851965,0.081352964,-0.1425484568,0.0267527327,0.0469446369,0.1992893517,-0.2087278515,0.3614895642,0.0190596003,0.1589624435,0.1899458021,-0.2031857073,-0.1272608638,-0.1174134687,-0.2530823648,0.1769057661,0.0884269029,-0.3375614583,0.022240879,-0.0522600412,0.0008730987,-0.1094607189,0.0189550538,0.2294987887,-0.2908079028,-0.1482636034,0.2226371914,-0.0878083035,-0.0815093443,0.3045401871,0.0436069034,-0.3001805246,-0.1483344734,0.2632359266,0.0427140892,0.0369828083,0.121416837,0.0237175189,-0.2068173587,-0.302369684,-0.1465962678,0.2409197986,-0.4375190437,-0.0331906378,-0.0494346097,0.0033178083,0.0506808609,-0.0272150077,-0.0058739791,-0.0840052441,-0.0371301249,-0.1992988437,-0.2851664126,0.3043626845,0.1010983586,0.3375176191,-0.0635118112,-0.1917614192,-0.2429069132,-0.0476204529,-0.4567089677,0.0419154353,-0.2124951035,0.05605492,-0.2779247761,0.0023317703,0.4613576531,-0.0997968242,0.2634966969,-0.0443392321,-0.4361954033,-0.3544987738,0.0609365441,0.1033121794,-0.0436229631,0.0142967226,0.0523092151,-0.076503098,-0.2957774997,-0.313400954,0.6046929359,0.1704049408,-0.1896798611,-0.2239180356,0.3594099283,-0.1177664101,-0.2642332911,0.215486899,0.1292701811,-0.0318126939,-0.068104282,0.0756035671,-0.341378659,-0.2242005467,0.4636442661,0.4278444946,0.1715426743,0.1539797187,0.1951588243,-0.0621873923,-0.0180493202,0.2234244496,0.1387220323,0.402947098,-0.0099452743,0.001162872,0.1709284782,0.3553200066,-0.2987711132,-0.3309018016,0.0836583748,-0.0448817126,0.1490396261,0.1131789684,0.2794712782,-0.1548952758,0.0715526789,0.0051325476,0.2009383142,-0.1418718398,-0.168976441,0.2532060742,-0.1036994383,-0.1869090796,0.1869180202,0.155501321,0.2906024754,0.5382152796,-0.1976806223,0.205840528,0.0884427577,0.1080901399,0.0100663658,-0.2402389348,-0.4833901823,0.0265654586,-0.1290313751,0.1863304228,0.202718392,0.3596491516,0.1560903788,-0.1159326583,-0.257940203,0.0854742229,-0.1045068949,-0.0051630386,-0.1370507926,-0.2379113585,-0.1614997536,0.2349171042,-0.1079137623,-0.0510151051,0.063251324,0.1068846062,-0.045994781,-0.0559367277,0.0670065805,0.0569071136,0.4399585724,-0.2259953767,0.0312765054,0.2127195597,0.1286762804,-0.0329716057,0.1760226488,0.3772985041,0.0412282757,-0.08742892,0.2876946926,0.1841208041,-0.0804393515,-0.2539972365,0.1091934517,0.0874144956,0.2268330157,0.1731670797,0.2240009159,-0.222832948,-0.2325480133,-0.2485449314,0.0837869942,-0.0438246764,0.1640124619,0.0560980551,-0.24888964,-0.0865307972,0.1873739958,-0.3417313993,0.0528506376,0.329139322,-0.1795737147,0.0524501763,-0.111730136,0.141595602,-0.3015827537,0.6976209283,0.0254463758,0.2618788183,-0.1180256903,-0.3995813727,-0.4921222329,0.0522904173,-0.0073115826,0.0935412943,0.0256276205,0.2950900793,0.1332703829,0.2454700172,0.0224462841,-0.368126899,0.255061239,0.2647507787,-0.2174696624,0.1709866524,0.0450598449,-0.1318834722,0.0978959724,-0.4364307523,0.1563719362,-0.0252211951,0.2079412937,0.0352058187,0.2020336241,0.1742803603,-0.2212349325,0.6054376364,-0.0029269236,0.1959469169,-0.1875263155,-0.1295257062,-0.4434603453,-0.0303481836,-0.2917251289,-0.215718776,-0.0807414427,0.1567137688,0.0566738807,0.100940749,-0.219702065,0.1784289032,0.2410254925,0.0916849822,-0.1354959309,-0.1076691151,-0.2465330511,0.3003370166,-0.2350136191,0.2327117324,0.0816560239,-0.25500682,-0.125608623,-0.3737340271,0.6916484833,-0.2861724496,-0.1407487988,-0.0290631168,0.4134774208,0.2459308952,-0.1011010483,-0.2815151513,0.2617774606,0.1991793066,-0.1879443675,-0.2414782196,0.3538321257,-0.0719252452,-0.1372228563,-0.1907454431,0.0154669201,0.0112963067,-0.2225657403,-0.25459674,-0.4584097266]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2945","title":"Protect master branch","comments":"Cool, I think we can do both :)","body":"After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:\r\n- 00cc036fea7c7745cfe722360036ed306796a3f2\r\n- 13ae8c98602bbad8197de3b9b425f4c78f582af1\r\n- ...\r\n\r\nI propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future:\r\n- [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch\r\n - Currently, simple merge commits are already disabled\r\n - I propose to disable rebase merging as well\r\n- ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~\r\n - ~~This protection would reject direct pushes to master branch~~\r\n - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~\r\n- [x] Protect the master branch only from direct pushing of **merge commits**\r\n - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch).\r\n - No need to disable\/re-enable this protection on each release \r\n\r\nThis purpose of this Issue is to open a discussion about this problem and to agree in a solution.","comment_length":8,"text":"Protect master branch\nAfter accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:\r\n- 00cc036fea7c7745cfe722360036ed306796a3f2\r\n- 13ae8c98602bbad8197de3b9b425f4c78f582af1\r\n- ...\r\n\r\nI propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future:\r\n- [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch\r\n - Currently, simple merge commits are already disabled\r\n - I propose to disable rebase merging as well\r\n- ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~\r\n - ~~This protection would reject direct pushes to master branch~~\r\n - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~\r\n- [x] Protect the master branch only from direct pushing of **merge commits**\r\n - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch).\r\n - No need to disable\/re-enable this protection on each release \r\n\r\nThis purpose of this Issue is to open a discussion about this problem and to agree in a solution.\nCool, I think we can do both :)","embeddings":[-0.2390099317,-0.0966292247,-0.0759872422,-0.0984264985,-0.0993319154,-0.114640519,0.0078532966,0.239554435,-0.029255297,-0.0760675296,0.3045841157,-0.0697917044,-0.1556909084,0.2472542524,-0.0918809772,0.2528043091,0.1680651754,0.000099914,-0.099873215,0.0868200585,-0.0079817604,-0.1016465425,0.1242587939,0.0464673564,-0.2174060941,0.0417824537,0.0499481186,0.1391150504,-0.4184132516,-0.4594882727,0.1383522302,0.1707512289,0.0033366256,0.5652734041,-0.0001042576,0.0216173306,0.2390211523,0.0108919349,-0.1405786723,-0.1769673228,-0.5225075483,-0.4049667418,-0.1415199637,0.0287014097,0.0748746395,0.06865637,-0.0910198018,0.002452912,0.3400871456,0.0976495519,0.2408002317,-0.1941754073,0.1793913692,-0.0078934981,0.3311534822,0.5539640784,-0.0518133156,0.22529535,0.0045611048,-0.0102421539,0.1125852838,0.2086250186,-0.0837501734,-0.2613006532,0.2912853956,-0.1761568338,0.5298508406,-0.3999355435,-0.1809146106,-0.1078854129,0.0704470351,-0.201073736,-0.3955834508,-0.2669799924,-0.1179080084,-0.0097809853,0.0528227128,0.2314448655,-0.0157607514,0.0743468106,-0.3007844985,-0.3514812589,0.0444366112,0.0477003977,0.10768985,-0.1619907767,0.0425621495,-0.0867641419,0.3245029449,0.0032727672,-0.1678859442,-0.2417692542,-0.3024554849,-0.0731806457,-0.1582976431,-0.166840598,-0.2660009265,-0.3856663704,0.4048146307,0.3533258736,-0.360022366,-0.049508784,0.002726417,-0.0647690669,0.2959358394,0.0719453767,0.011164953,0.0792105719,0.5598432422,0.3058459461,0.0217627417,0.3349214792,0.3275439441,-0.0626471415,0.0689635947,0.5172939301,0.5336612463,-0.2621914148,0.0377011895,0.1680787653,-0.3323733509,-0.3190573454,0.0054372335,-0.0610207208,0.0830994621,0.0703226849,-0.1091590971,0.0525613427,0.1415696889,-0.0016033769,-0.3210552037,-0.1457319856,-0.3556134701,-0.0552582555,0.1311259568,-0.4680353403,0.0908330008,0.4519638121,-0.1013908461,-0.2686688304,0.0358462371,0.1716575772,0.0330468714,0.2908923328,-0.1433112174,-0.3825237751,0.2031245977,0.0965870917,-0.0456247479,0.2545187175,-0.1935967952,-0.2513652146,-0.1866476983,0.3089985549,-0.1121701747,0.3287965655,-0.548268497,0.0025912365,0.0102211721,0.3561609685,0.4972600937,0.385391444,0.2942132354,-0.1614470184,0.0424310081,0.1232627109,0.3384101689,-0.0511925742,-0.0100153778,-0.2738169432,0.1200275049,0.2083038092,-0.2946696877,-0.0103482995,-0.0507854298,-0.1232980564,0.0196767803,-0.048742909,-0.204722017,-0.0717563778,-0.4525829852,-0.0241770819,-0.0117112603,0.0191363525,0.1543500572,-0.3124671876,-0.0787971914,0.1779266298,-0.182664752,0.1171256378,-0.3127051294,-0.6241436601,-0.1159733236,-0.0647230893,0.1302614063,0.1962026656,0.3260959387,0.0762788281,0.0484594442,-0.1529644579,-0.0238246322,0.1797253937,0.3943619728,-0.1591625661,-0.2575499117,0.0904227868,-0.3276287019,0.0429003425,-0.0230592508,-0.0326709487,0.0328225382,-0.4201252162,0.2604811192,-0.0395855084,-0.089378275,-0.1092364043,0.2325994372,0.0994815007,-0.2429064363,-0.2273789346,-0.1815333217,0.2851662338,-0.242647782,0.2751372755,0.099704735,0.0815574303,-0.1571651399,-0.0062341182,-0.0228314847,0.3285116851,0.1736403704,-0.0701381266,-0.04674327,0.1801669151,-0.0945321396,0.1870126128,0.1769949794,0.4175725579,0.3424830437,-0.1347359419,-0.0105623221,-0.1067967415,-0.1460241973,0.022494385,-0.2684330642,0.2932992876,0.0991126224,-0.2107091695,-0.249318257,-0.0687692985,-0.1965132803,-0.265663594,-0.1541765183,0.0057348073,-0.1855473518,0.1418430507,-0.2901535332,0.2540666461,-0.2189239413,0.1275674105,0.0980846509,-0.201123789,-0.0243991297,0.0869311392,-0.0290957168,-0.057731878,0.100935325,0.3220874071,0.0902120993,0.183042407,-0.0160376783,0.0336617008,0.2433368862,-0.0592605695,0.362724781,0.1478046626,-0.0405387431,0.0476890579,0.3128943741,0.1199872941,0.0844289884,0.3880383074,-0.0403469466,0.1266814917,-0.2439862639,-0.0744038522,-0.2096193433,-0.5341181755,0.1785611957,-0.1322654784,-0.311395973,-0.0278486367,-0.053532742,0.1170226783,-0.4517551064,0.1408915371,0.2340205312,0.2509391308,-0.3069441617,-0.0698819757,0.1471458375,-0.0289604329,0.0071960338,0.1242603362,0.4291484952,-0.2824301422,0.4524761438,0.2389160544,-0.1856772155,-0.4897386432,-0.4018841088,0.0115547851,-0.1668282598,-0.1153612956,0.0861761943,-0.0717044175,0.1023176908,-0.401612252,-0.3951958418,0.0295078922,-0.3599407971,-0.1393553913,0.0470686667,0.1184691191,-0.2150190026,-0.0969456658,-0.2935731709,-0.2871156633,0.5013284087,-0.0580234416,0.0385896452,0.0112076271,-0.2732456625,-0.1094526052,0.1394193172,0.0409913287,-0.0648903027,-0.3088234365,-0.0299696531,0.0454311818,-0.0143893873,0.368963629,0.2126246095,-0.3675787747,0.0550749078,-0.350450635,-0.1397126317,-0.0122180469,0.4691388905,0.256077528,-0.1587453038,0.0662926659,0.0626281053,-0.1768465936,0.1262467802,-0.1262247413,0.1179754212,0.3565621078,0.3317129016,-0.0136891427,0.1207261309,0.0874625519,0.6525163054,0.3680907488,-0.0682478175,0.1437117904,0.2640963793,0.0372981504,0.0415912867,-0.2894337177,0.0679146126,-0.0695572942,-0.0514505655,0.0428000949,-0.1358909458,-0.1156823337,-0.0192327965,-0.4392817616,0.0292560495,-0.4280707836,-0.212833792,-0.3319002986,0.1305010617,-0.0154377855,-0.4608305693,0.1475661546,-0.0283782165,0.2772078216,0.2418593913,0.1750047207,0.1717233062,0.0286009591,-0.0267298147,-0.143258512,-0.0784367099,-0.0277465973,0.2193704545,0.0662960187,-0.120667249,0.0654256046,0.0472218953,0.401168257,-0.0122205056,0.1574847549,-0.2568204701,-0.3890761137,0.6179253459,-0.2597557604,0.1685838103,0.3741835654,0.5948965549,0.4432282746,-0.3229280412,-0.1968980581,0.2175525427,-0.0058352049,0.1632378548,0.1841427237,0.0639553592,0.2376962006,-0.4858786464,0.2991431355,0.2369734645,-0.2511768043,-0.0432059951,-0.0785255134,-0.1137046441,0.0831007659,-0.0332846567,-0.2697017193,0.0042523337,-0.0428405702,0.1484801471,0.103806518,0.1191856787,0.3641275465,0.5810796022,0.2270472646,-0.4135166109,-0.0292455852,-0.1834749579,-0.0771063045,0.3103395998,0.1818852574,0.2265654951,0.0643974394,0.2080049366,-0.1813084334,-0.2263067961,-0.2212891281,0.0309540853,-0.2094345242,-0.0052669551,0.1052183583,0.0515394434,-0.0904067382,0.0824889988,0.0987536013,-0.3212070465,0.1318824291,0.5714728832,0.955177784,0.3502957225,0.1090738475,-0.2922923863,-0.2720266581,0.3813621998,-0.357526511,-0.1286929101,-0.3113082349,-0.1811920255,-0.1677663922,-0.3074550629,-0.0780642331,-0.4934645891,-0.3175655603,0.1523682475,-0.1961845458,0.5375774503,0.0177650489,0.1167705655,0.1282786429,-0.038145341,0.2970950305,0.1892168671,0.0111535219,0.1593013853,-0.0028083171,-0.4545554519,-0.0279463418,0.0025342032,-0.4306589961,0.0386116318,-0.3405902982,0.0399823673,-0.2339083552,-0.2824401855,0.1117530316,-0.1104251817,0.6192416549,0.2143154591,-0.0984719917,-0.0485647954,0.2703552842,0.3980438113,-0.197127685,-0.196609512,0.5138126016,-0.2046020627,-0.1562071592,-0.0987637788,-0.0285022315,-0.3348375857,-0.2822057903,-0.0767591447,-0.3614266813,0.0732668266,-0.0306378119,0.2732182145,-0.0691128522,-0.1619696319,0.1780715436,0.111415118,-0.1410685033,-0.1422604918,-0.1091866344,-0.2943277955,-0.0380600765,0.1902545691,0.2271509767,-0.1096624061,-0.0836395845,0.0573430769,-0.1975595355,-0.2153514922,-0.2259288728,0.1200701594,-0.1281341463,0.0691441,-0.4934980571,-0.3398421407,0.1174373701,0.1943581998,0.2834068835,0.3471649885,0.07340087,-0.165871352,-0.2640971243,0.2091121227,-0.3319115043,0.1677211374,0.0679885969,0.2423418313,-0.0730114356,0.0273042507,-0.3893203735,0.1249633506,-0.1246866807,-0.0371800624,-0.1938935816,-0.076455161,0.002154053,-0.4198532403,0.0430044867,0.0236390047,-0.1842414141,-0.2793491185,0.0818327814,0.0688198656,0.0645905882,0.0382995978,-0.0577248856,-0.1069227234,0.1288665682,0.3231302798,0.2298199683,-0.0998132899,0.1544985622,-0.0540686958,0.1755025834,0.0916104764,-0.0348069631,0.1178810969,-0.012650379,-0.2803532481,-0.1100896522,0.3647060692,-0.1755682975,-0.1734805107,-0.3497366011,0.1640637666,0.0974830911,0.6140381098,0.110363245,0.1082694158,-0.4196391404,0.2307030708,-0.0214151721,-0.0301906541,-0.1783225387,-0.1215486825,-0.0753639936,0.2950595319,-0.0994521976,-0.3054563999,0.2349793315,0.1782335341,0.0241322219,0.0836851299,0.5049234629,0.1231228486,-0.2381349504,0.4243248105,0.2228861451,-0.1286406517,0.455909878,0.5233042836,-0.1415461004,-0.0785631537,-0.0360233933,0.0871152431,0.1805791706,-0.1823303849,0.0883068964,0.3238530457,0.6459390521,-0.1694083512,0.3705488145,0.0255201813,0.3059286177,0.2132881433,0.0270452425,-0.2294175178,0.4330947399,0.2575822473,0.3352282941,-0.5901472569,0.1695657074,0.5056748986,-0.023568213,0.2167177498,-0.3594747484,-0.0223750621,-0.0552352853,-0.0159900282,-0.0719330311,0.1218386665,0.1490119845,0.2901292443,-0.0673369542,-0.0273434464,-0.0467063077,0.2230517119,0.2663088441,-0.1428412646,-0.2008356899,0.2622175515,-0.2484770417,0.0786395818,0.2177159339,0.1616025418,-0.0978315324,-0.0521366596,0.1798922271,0.4497513771,-0.1746967882,-0.039889738,0.0317630544,-0.1530031413,0.1404958814,0.0627516881,0.2746909261,-0.0635648817,0.1233454049,0.0005846465,-0.383130908,-0.0437247008,0.4243560433,0.4585390389,-0.3600539863,0.41664958,0.174680531,0.2586427629,-0.0644906461,0.4355578721,-0.0931281298,0.1283278465,-0.0560730807,0.1473985761,0.0403031036,0.217051357,0.0914419889,-0.0287602842,-0.0969899297,0.0412180498,-0.2160513252,0.2098333389,0.3064318895,0.1730033755,0.4223076701,-0.3031860292,0.1374098659,0.1526105702,-0.0731561556,-0.2435802221,-0.047290422,0.1329396963,-0.3592749536,0.0106856637,0.2144466639,-0.0846365243,0.01132154,0.064479664,0.4179263711,0.262011826,0.1000918522,0.0653353333,-0.0138661312,0.2816164494,-0.3647170663,0.2353287637,0.1579896212,-0.1405763328,-0.2575822473,-0.0880536586,-0.0056584454,-0.0124874227,-0.4783657789,-0.0724288374,-0.1397930533,0.4400902689,-0.1443308294,0.0493659228,-0.3884942234,0.0546305478,-0.0210841242,0.2141258866,-0.3707188964,0.1099852324,0.0246811155,0.0773969591,0.0523463786,0.1700744331,0.0661983937,0.2172149122,-0.2984472513,-0.1805216223,0.3678478599,0.0187848061,0.0980769023,-0.15001674,0.4350650012,0.5103737712,-0.2477454543,-0.3580597639,-0.2498808056,-0.0220319424,-0.2263912112,-0.4824006557,0.2833116651,0.0979687423,0.2627729475,-0.0858350545,0.0814130977,-0.069908455,-0.2276804298,0.164998427,-0.2625057101]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2945","title":"Protect master branch","comments":"@lhoestq now the 2 are implemented.\r\n\r\nPlease note that for the the second protection, finally I have chosen to protect the master branch only from **merge commits** (see update comment above), so no need to disable\/re-enable the protection on each release (direct commits, different from merge commits, can be pushed to the remote master branch; and eventually reverted without messing up the repo history).","body":"After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:\r\n- 00cc036fea7c7745cfe722360036ed306796a3f2\r\n- 13ae8c98602bbad8197de3b9b425f4c78f582af1\r\n- ...\r\n\r\nI propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future:\r\n- [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch\r\n - Currently, simple merge commits are already disabled\r\n - I propose to disable rebase merging as well\r\n- ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~\r\n - ~~This protection would reject direct pushes to master branch~~\r\n - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~\r\n- [x] Protect the master branch only from direct pushing of **merge commits**\r\n - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch).\r\n - No need to disable\/re-enable this protection on each release \r\n\r\nThis purpose of this Issue is to open a discussion about this problem and to agree in a solution.","comment_length":64,"text":"Protect master branch\nAfter accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:\r\n- 00cc036fea7c7745cfe722360036ed306796a3f2\r\n- 13ae8c98602bbad8197de3b9b425f4c78f582af1\r\n- ...\r\n\r\nI propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future:\r\n- [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch\r\n - Currently, simple merge commits are already disabled\r\n - I propose to disable rebase merging as well\r\n- ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~\r\n - ~~This protection would reject direct pushes to master branch~~\r\n - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~\r\n- [x] Protect the master branch only from direct pushing of **merge commits**\r\n - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch).\r\n - No need to disable\/re-enable this protection on each release \r\n\r\nThis purpose of this Issue is to open a discussion about this problem and to agree in a solution.\n@lhoestq now the 2 are implemented.\r\n\r\nPlease note that for the the second protection, finally I have chosen to protect the master branch only from **merge commits** (see update comment above), so no need to disable\/re-enable the protection on each release (direct commits, different from merge commits, can be pushed to the remote master branch; and eventually reverted without messing up the repo history).","embeddings":[-0.1553194672,-0.1002306268,-0.070321016,-0.0798171908,-0.1042534709,-0.1879874468,0.0104035577,0.2728635967,-0.0098489737,-0.0841399208,0.2926148772,-0.0778751224,-0.1324572563,0.2158903182,-0.0562292449,0.1705456823,0.2003232092,-0.0407981202,-0.0923330337,0.0279233586,0.0226084664,-0.0687029436,0.0832000151,0.0436301306,-0.1837819666,0.0054013738,-0.0148117775,0.1564962417,-0.4397666156,-0.4695702493,0.1452465504,0.2322402596,0.0251065325,0.570535183,-0.0001043637,-0.0003454557,0.1694866419,-0.0406010039,-0.1401022524,-0.2154636681,-0.4901072085,-0.4447958767,-0.07006336,-0.0083542913,0.1288067251,0.0528678969,-0.132499963,0.0238457564,0.3848451674,0.1055523977,0.2386109382,-0.1028517336,0.1812405735,0.0137694413,0.392008841,0.5474528074,-0.066634573,0.3031997979,0.0540179275,0.0006757497,0.0611058585,0.1741086096,-0.0823729113,-0.2866700888,0.2270194292,-0.1779912561,0.5181742311,-0.4243531227,-0.2415838391,-0.0831176639,0.0523793846,-0.1781465262,-0.4089041948,-0.2419367433,-0.1499476582,0.0076339091,0.0677802935,0.2468644679,0.039739579,0.0036496122,-0.3773649931,-0.2718090415,-0.004407926,0.0492040403,0.141940549,-0.1626293063,0.0664263666,-0.1270385236,0.2942230105,-0.0473279878,-0.1469833106,-0.3676346838,-0.2413333207,-0.0125874653,-0.1026133746,-0.1957468837,-0.2462065965,-0.3248277903,0.4251304567,0.3889683783,-0.3832727969,0.0137396995,-0.0304892361,-0.0810169131,0.2071851492,0.1178578511,-0.0104900496,0.0584886484,0.530713141,0.3144408166,0.0051584225,0.3255857527,0.3015701473,0.0030471815,0.0344459116,0.574118197,0.4289444685,-0.26292032,0.1276476383,0.1959735751,-0.3412007391,-0.3845081031,0.0711311549,-0.0861726329,0.1103260964,0.0271337908,-0.1147114784,0.0668015406,0.0451908335,-0.0380863659,-0.3079027534,-0.141485557,-0.3744001687,-0.0685305148,0.1860401928,-0.4571051598,0.1491677761,0.4325242639,-0.0680372268,-0.1992845535,0.0445522442,0.1292257011,0.0555874184,0.2961196899,-0.1537979692,-0.4526284337,0.2235967368,0.1313333064,-0.0668994114,0.1940043122,-0.1598993689,-0.2616221011,-0.1049757376,0.2977656722,-0.1482721567,0.3071561158,-0.5237665772,0.0363732092,-0.0044837268,0.3506707847,0.4910375476,0.3025480509,0.3280785084,-0.1685939282,0.0617566295,0.1903455257,0.3235116601,-0.0516864397,0.0623045377,-0.2877363563,0.1095039099,0.1450733244,-0.3121602237,0.0175557248,-0.0052360594,-0.1312822402,0.0107761566,-0.0714070052,-0.1119878516,-0.0989294425,-0.3666853309,0.0694030076,-0.0407375731,-0.1007770002,0.2079568803,-0.3201756477,-0.0947941095,0.2050814927,-0.1481231451,0.0208985135,-0.2919872105,-0.6061385274,-0.1636558473,-0.06343887,0.1249144152,0.2275642306,0.3607693315,-0.0050310493,-0.0283617247,-0.1714023799,-0.0093972106,0.1125069708,0.3421364129,-0.1496780068,-0.2415423691,0.1322465539,-0.3545884192,0.0911409035,-0.0785121024,-0.0097747799,0.0344485678,-0.3157884181,0.3412798941,-0.0440529883,-0.1208016798,-0.1141509414,0.2415172756,0.095951654,-0.1747335494,-0.2039210498,-0.2074069828,0.2601909339,-0.2408306301,0.3018657863,0.1300136894,0.1026374176,-0.1411807239,-0.0364533328,-0.0250960924,0.36193645,0.2363044769,-0.0899174288,-0.0591584407,0.1687609404,-0.0490626656,0.2035702467,0.2246947736,0.4252423346,0.3736152053,-0.1026815102,0.0626009405,-0.0917925984,-0.0787246898,0.0245947056,-0.3291273713,0.3280371726,0.073694557,-0.1394644678,-0.1740321219,-0.1033326834,-0.1740740985,-0.2202692628,-0.1707722992,-0.012951361,-0.1690982729,0.1943652332,-0.3365918398,0.2223692834,-0.2772080004,0.148579374,0.0150526064,-0.2516064942,-0.1024010181,0.0682980418,-0.0059674503,-0.1327149123,0.1241709664,0.3773829341,0.1278743744,0.1701000035,0.0042741634,0.0201513059,0.2572831213,-0.1164359227,0.353325963,0.1443069726,-0.1054452285,0.054064367,0.3364559114,0.1812186539,0.0643318817,0.346919328,-0.0033243245,0.0352581739,-0.2084068358,-0.0459848903,-0.2602106631,-0.5283415914,0.2192376256,-0.110637784,-0.3552163243,-0.0085041393,-0.0453661792,0.047418233,-0.5248722434,0.136029914,0.1835808307,0.2905260623,-0.3178029358,-0.1199780703,0.1263155937,-0.0225951802,0.0123441136,0.1285599768,0.3631602824,-0.2266809344,0.45945099,0.2389809936,-0.1695563644,-0.4726397991,-0.3434132934,0.0191301201,-0.2039363831,-0.1179748252,0.0098189935,-0.0989309773,0.0921538174,-0.3585401177,-0.3257124722,-0.018773742,-0.3730046451,-0.1600959152,0.0695955083,0.0628860295,-0.2346176803,-0.114040032,-0.386590004,-0.2821527123,0.4793556035,-0.0164330266,0.1156589612,-0.0431727543,-0.2923772335,-0.0459178463,0.0874827728,0.1366390288,-0.0000448525,-0.3609322906,0.0325667374,-0.0406603813,-0.0293471292,0.352609843,0.1854910851,-0.4302407205,0.0712517202,-0.3217604458,-0.1558190286,-0.0520843901,0.4864443243,0.3141199946,-0.0973737389,0.0710799471,0.1035510823,-0.1919001341,0.0550895296,-0.1816118211,0.1472238153,0.3454884887,0.2857186198,0.0378993787,0.1458674669,0.0254973993,0.7140879035,0.3778497577,-0.059657298,0.210337624,0.3106974065,-0.0277180839,0.0039902786,-0.2601569593,0.065046832,-0.1012431234,-0.0000209612,0.0867039785,-0.1295782924,-0.1229653433,-0.0112106856,-0.4777139425,0.0091558872,-0.4146761596,-0.2396630496,-0.3181839585,0.1842502356,-0.0225256812,-0.4854888022,0.1313801706,0.0537709594,0.2305762768,0.2159709483,0.2142065167,0.1812160015,-0.0202590171,-0.0510845743,-0.1200738922,-0.067726247,-0.0043309759,0.145250544,0.0905801877,-0.1670407504,0.0318057276,0.0695502236,0.3299908936,-0.0148684094,0.1402913779,-0.1708071232,-0.5094846487,0.6086307168,-0.3185944259,0.1394028515,0.2990989089,0.5112247467,0.4397634268,-0.2954540849,-0.1953165382,0.2071224451,0.0325558819,0.1978364438,0.1773502678,0.1449773014,0.2783072591,-0.4903811812,0.2887739539,0.1548331976,-0.2504818439,-0.1551757604,-0.0039056637,-0.0270501971,0.1264622211,-0.0263037495,-0.2453983575,0.0091275191,-0.1094373837,0.1734458208,0.1329204887,0.0698278099,0.4152193964,0.594658792,0.1349339038,-0.3891987503,-0.0482011773,-0.1750733256,-0.0742834732,0.3271439075,0.1983511001,0.24822779,0.0203486141,0.1776862442,-0.1234726384,-0.2662384808,-0.2033664733,-0.038491372,-0.0767425969,0.0075890222,0.0873292238,-0.0458977483,-0.0909918919,0.0130487047,0.1399983913,-0.3661002219,0.1637848169,0.6180499792,0.9825440049,0.3975676596,0.0890336856,-0.2546566427,-0.3273829818,0.3259905577,-0.340107888,-0.1664507836,-0.2470169514,-0.1235878542,-0.1842461079,-0.3403116167,-0.0386944152,-0.4898471832,-0.3528554738,0.1351124644,-0.2180551142,0.5279191732,0.008177679,0.1233861297,0.1910188198,-0.12231455,0.2551510632,0.1982705742,0.061811205,0.104095228,-0.0040276642,-0.4687906206,-0.0022327867,0.0205386523,-0.3772695959,0.0891471654,-0.3437781334,0.0388470143,-0.2063531876,-0.3160520196,0.115827091,-0.0870479271,0.6059409976,0.1666302532,-0.0794522315,-0.0446208119,0.2913893759,0.4645246863,-0.2288517058,-0.219530493,0.5223725438,-0.2133638412,-0.1302172095,-0.2047614008,-0.0803344026,-0.2370928973,-0.2609480619,-0.1280429661,-0.3282325566,0.0795017928,0.041714251,0.2636537552,-0.0910693035,-0.1676060557,0.1859930903,0.1154857874,-0.1075729132,-0.118584998,-0.1695205718,-0.2610394955,-0.0678159222,0.2016011029,0.161605835,-0.1206071749,-0.0767639652,0.099657312,-0.2443365902,-0.2207894772,-0.2072923779,0.105987981,-0.1394685954,0.0929813311,-0.5486782193,-0.32993415,0.0517466478,0.2461933494,0.279415518,0.3304983079,0.1179999709,-0.0997705981,-0.2244495153,0.2186434418,-0.3279267251,0.1686490178,0.0235842206,0.1588117927,-0.0260097533,0.0404301658,-0.3905956745,0.160315454,-0.1254802048,-0.0211261567,-0.2255206406,-0.1017537564,-0.000630246,-0.3081870973,0.0324874744,0.0234331191,-0.1957566142,-0.2830150425,0.0926690474,0.0656956136,0.0647893623,-0.0570135079,-0.0574608184,-0.1081577986,0.0827530622,0.2720321715,0.2396921068,-0.1584913284,0.1097129956,-0.0322248675,0.1710882336,0.1105851158,-0.0247724894,0.0892416015,0.0146322586,-0.2658016682,-0.0353747047,0.267159909,-0.1817460209,-0.2171436995,-0.2685393691,0.1888355613,0.1056337059,0.5840907693,0.2028801143,0.1442169249,-0.405082494,0.1754089892,-0.0531977862,-0.0036356284,-0.1692355126,-0.0969402716,-0.0849622712,0.2793601751,-0.0867193192,-0.3035972714,0.2713837624,0.1881819069,0.0007486099,0.1049711928,0.5280760527,0.1296208501,-0.1896519363,0.4221355915,0.2712435424,-0.0876758322,0.494900018,0.470233053,-0.1131819189,-0.0203250889,-0.082736142,0.116614446,0.2306870967,-0.2540961802,0.1284718663,0.2962196171,0.6462965012,-0.1688840091,0.3935549557,-0.0330962837,0.3437395692,0.1979931891,0.0587407239,-0.2460495532,0.3927094042,0.2117075175,0.3301511109,-0.6189893484,0.1363731921,0.4914032817,0.0492829382,0.1995816529,-0.400105536,-0.0492327139,-0.0584757105,-0.0483128205,-0.0781826749,0.1634954214,0.1237376407,0.2829664946,0.0065842085,-0.088377893,-0.043440666,0.1850482225,0.2415616512,-0.1169977486,-0.1811731756,0.357754916,-0.2761978507,0.0664936379,0.215460062,0.1350683123,-0.074716486,-0.0686341003,0.1802342385,0.4503947198,-0.1699341387,0.0225111321,0.0557496101,-0.1625677198,0.1577177346,0.0801470727,0.2580322921,-0.0533408038,0.1128025502,0.0354268774,-0.4152979553,-0.025322238,0.4330811501,0.5564644337,-0.3377302289,0.4601393938,0.1192681268,0.2847007811,-0.0326848105,0.4374352098,-0.0471260622,0.1163921505,-0.0051536481,0.143078804,0.02808984,0.2575880587,0.1414998174,-0.0333454311,-0.1592887938,0.0301065128,-0.1215411872,0.127435267,0.2968399227,0.17307055,0.419888705,-0.2947140038,0.1275687069,0.1282345057,-0.0875694603,-0.3413968384,0.0095217237,0.0940847918,-0.3605292141,0.0042476766,0.2224816233,-0.0694397017,-0.0241943654,0.1189339831,0.4094779193,0.1927902848,0.1120927632,0.1051306576,-0.0493477993,0.3209429979,-0.3918459415,0.2249616981,0.1892230064,-0.1029298455,-0.2218884081,-0.0510987751,0.0073658284,0.0212806091,-0.4355088174,-0.1401801854,-0.162226066,0.4209837019,-0.1355972886,0.0682982057,-0.3722180128,0.0687672645,-0.0371344872,0.2683559954,-0.3000992239,0.0864084363,0.0525472984,0.017337583,0.088370651,0.1305798739,0.068434529,0.2528506517,-0.3523156941,-0.1248895153,0.3586619198,-0.0194706563,0.0910360888,-0.2172687799,0.4182040691,0.4654460251,-0.1480811983,-0.3321150839,-0.2635183036,-0.0111563709,-0.2798339725,-0.542089045,0.2986377478,0.1378680468,0.1974387169,-0.1058711782,0.0366066247,-0.0743264556,-0.1527637988,0.0611719489,-0.1987861097]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"Hi ! I guess the caching mechanism should have considered the new `filter` to be different from the old one, and don't use cached results from the old `filter`.\r\nTo avoid other users from having this issue we could make the caching differentiate the two, what do you think ?","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":50,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nHi ! I guess the caching mechanism should have considered the new `filter` to be different from the old one, and don't use cached results from the old `filter`.\r\nTo avoid other users from having this issue we could make the caching differentiate the two, what do you think ?","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"If it's easy enough to implement, then yes please \ud83d\ude04 But this issue can be low-priority, since I've only encountered it in a couple of `transformers` CI tests.","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":28,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nIf it's easy enough to implement, then yes please \ud83d\ude04 But this issue can be low-priority, since I've only encountered it in a couple of `transformers` CI tests.","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"Well it can cause issue with anyone that updates `datasets` and re-run some code that uses filter, so I'm creating a PR","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":22,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nWell it can cause issue with anyone that updates `datasets` and re-run some code that uses filter, so I'm creating a PR","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"I just merged a fix, let me know if you're still having this kind of issues :)\r\n\r\nWe'll do a release soon to make this fix available","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":27,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nI just merged a fix, let me know if you're still having this kind of issues :)\r\n\r\nWe'll do a release soon to make this fix available","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"Definitely works on several manual cases with our dummy datasets, thank you @lhoestq !","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":14,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nDefinitely works on several manual cases with our dummy datasets, thank you @lhoestq !","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2943","title":"Backwards compatibility broken for cached datasets that use `.filter()`","comments":"Fixed by #2947.","body":"## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":3,"text":"Backwards compatibility broken for cached datasets that use `.filter()`\n## Describe the bug\r\nAfter upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with \r\n`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}`\r\n\r\nRelated feature: https:\/\/github.com\/huggingface\/datasets\/pull\/2836\r\n\r\n:question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) \r\n\r\n## Workaround\r\nRemove the cache for the given dataset, e.g. `rm -rf ~\/.cache\/huggingface\/datasets\/librispeech_asr`.\r\n\r\n## Steps to reproduce the bug\r\n1. Delete `~\/.cache\/huggingface\/datasets\/librispeech_asr` if it exists.\r\n\r\n2. `pip install datasets==1.11.0` and run the following snippet:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nids = [\"1272-141231-0000\"]\r\nds = load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\", \"clean\", split=\"validation\")\r\nds = ds.filter(lambda x: x[\"id\"] in ids)\r\n```\r\n3. `pip install datasets==1.12.1` and re-run the code again\r\n\r\n## Expected results\r\nSame result as with the previous `datasets` version.\r\n\r\n## Actual results\r\n```bash\r\nReusing dataset librispeech_asr (.\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1)\r\nLoading cached processed dataset at .\/.cache\/huggingface\/datasets\/librispeech_asr\/clean\/2.1.0\/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1\/cache-cd1c29844fdbc87a.arrow\r\nTraceback (most recent call last):\r\n File \".\/repos\/transformers\/src\/transformers\/models\/wav2vec2\/try_dataset.py\", line 5, in \r\n ds = ds.filter(lambda x: x[\"id\"] in ids)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 2169, in filter\r\n indices = self.map(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1686, in map\r\n return self._map_single(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 185, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py\", line 398, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1896, in _map_single\r\n return Dataset.from_file(cache_file_name, info=info, split=self.split)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 343, in from_file\r\n return cls(\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 282, in __init__\r\n self.info.features = self.info.features.reorder_fields_as(inferred_features)\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1151, in reorder_fields_as\r\n return Features(recursive_reorder(self, other))\r\n File \".\/envs\/transformers\/lib\/python3.8\/site-packages\/datasets\/features.py\", line 1140, in recursive_reorder\r\n raise ValueError(f\"Keys mismatch: between {source} and {target}\" + stack_position)\r\nValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nFixed by #2947.","embeddings":[-0.3049958348,0.1243880838,-0.0465583764,0.2368971109,0.1517290324,-0.0666598156,-0.0069413707,0.3285393715,0.1737383604,0.0293814819,-0.2313861847,0.3581927419,-0.1503304988,0.2527370155,-0.2977862358,-0.1378120482,0.0444693454,0.0321247876,-0.0614953265,0.166412577,-0.1239175051,0.4988580644,-0.1957765073,0.0070795678,0.2221652865,0.0513210297,0.1091236398,0.138566196,0.0798620358,-0.3392693698,0.5981172919,0.1064069718,-0.0348691642,0.498096019,-0.0001213098,0.099867776,0.4398066998,-0.1073976383,-0.4579497874,-0.3686879575,-0.192854017,0.1319934726,0.2131137848,0.0811629742,-0.0715122968,0.1143416911,-0.303493619,-0.346046567,0.2488765419,0.1989328116,0.2267285585,0.2136144787,0.2657418847,0.066587165,0.170179069,0.0826525018,0.0259718653,-0.0490864292,0.2940285802,0.014261202,0.043685697,0.4806104004,-0.3283523321,-0.2146463245,0.3046185374,-0.2982533276,0.1030709222,-0.3957037926,0.2999309897,0.0781370997,0.1462642252,-0.255518198,-0.5773367882,-0.3483263254,-0.5121472478,-0.2868291736,0.4164735079,-0.216358915,-0.2009968162,0.2200830877,-0.3342647851,-0.0270038545,0.0934983194,0.0535101071,-0.0584819205,0.2941842079,-0.1318320185,-0.0048626242,0.0149830757,-0.132900551,0.1924084276,-0.2750048339,-0.1968749017,0.1725800484,-0.0155595224,-0.0546131916,0.0954222009,0.3323281705,0.140570581,0.0356287211,0.0192098524,0.0845401734,-0.0409641638,0.0536773168,0.1373275816,0.475378722,0.4901195765,0.2467906624,0.2727875113,0.2854887545,-0.0078853024,-0.0013800253,0.5281852484,-0.1666412055,0.0747071281,0.1517307162,0.196695298,-0.4048683643,-0.3602132797,0.1689626127,-0.0769995078,-0.1250401586,0.2207264006,0.0534940958,0.1005648673,0.2142349631,0.1320811808,0.0338568501,-0.1964725256,0.0612344779,-0.1725096703,-0.1739457697,0.0011616903,-0.1112991124,-0.0473052748,-0.89977777,0.1747701168,0.0482602492,0.0063488153,0.1058098748,-0.0697732121,0.207434997,0.1730166823,0.5229284167,-0.2213587761,0.0839003325,0.1671907902,-0.548327744,-0.1386865377,0.0125284484,-0.2415378094,-0.1112066656,0.2085879147,0.1735862345,-0.3296715915,0.0126515068,-0.17029953,0.4515431225,0.2592140436,-0.5466870666,-0.0779574588,-0.213380143,-0.4382343888,-0.1997140199,0.0971902013,0.4131370485,-0.5942543149,-0.3216674924,0.0216870401,0.0319067985,0.0399160981,-0.1229341999,-0.063439101,-0.2444722205,-0.1077689156,-0.1384182572,0.3606542647,-0.3195546865,-0.733892858,0.1537256837,0.2360213548,0.5342661738,-0.0709591061,-0.1374839246,0.2813950479,-0.2578290701,0.1459920853,0.0848445743,-0.1649033576,-0.0079940017,-0.3409692049,-0.2346946597,0.2238697708,-0.1502910852,0.0904174298,0.3348585963,0.0790057704,-0.2542206049,0.0502917804,0.0350500681,0.1049243882,0.1990925223,0.2453232259,0.2237652242,0.2863062918,-0.4365186989,-0.2801407278,0.2506181598,0.2141302824,-0.0600011088,-0.1524066925,-0.1982273012,-0.2678262889,-0.0884182155,-0.1589775681,-0.3044096231,0.0788738504,0.0225765873,0.3843636513,-0.0385034606,-0.1145259663,0.6218649745,0.3084535897,0.0944804996,-0.5049411654,0.229641512,-0.0548514128,0.0051691518,-0.0591857955,0.0377827175,0.2939595878,-0.0685256049,-0.2895757854,0.407875061,0.0392592028,0.084168233,-0.0276337974,0.1937460154,0.1773459613,0.2105839252,0.0639338642,0.2741822898,-0.0148837138,0.2087214142,-0.195456177,0.557248354,0.0293753464,0.3240505159,0.1208347678,0.148756206,0.3702490628,0.1305633932,-0.2189950645,-0.1370786875,0.3204495311,-0.4587228,0.0626674443,0.0123394234,-0.1297932267,0.3776263595,0.0716826022,0.0755119473,-0.0334524736,0.3119212389,-0.1593657285,-0.042190969,0.1813624054,0.4024861157,0.4014509916,0.2088322341,0.0507172868,0.0758817792,-0.1615392268,-0.0248774402,0.2082498521,0.1293462515,-0.1696399748,0.1535750628,0.3034210205,0.2732039392,-0.3061974645,0.0933934823,-0.2356239706,0.1195331365,-0.270581305,0.3616366684,-0.2810101509,-0.2280866951,0.076141715,-0.1678324938,-0.2615140676,-0.3875222206,0.089899376,0.3934392929,-0.086219281,0.2567750216,0.0147161586,0.2445412278,0.136167407,-0.1996820718,-0.4005717337,-0.2078815252,-0.3476482034,0.0053976052,0.2051323503,-0.1568491161,0.2970358729,-0.0763099566,-0.2066247612,-0.390294075,-0.9079098701,-0.0486286655,-0.123964861,0.6180298328,0.3176228106,-0.0916265026,-0.1352719367,-0.2327560335,0.1665794849,-0.0413150303,-0.4310828745,0.1568385363,0.0666227341,0.1682221591,-0.2003851682,-0.2270279974,-0.1135412008,-0.1458190233,0.0079398574,-0.1276904047,0.0849654675,0.3094039559,0.1123534068,-0.3165157139,-0.0872011557,0.1330857426,-0.3335401416,-0.0773374513,0.0739419162,-0.054162316,-0.2302434891,-0.034139622,-0.0319806486,0.0831893757,0.3736632466,-0.4069587886,-0.3791774213,-0.2713731527,-0.0048744264,0.2311773747,0.068360962,0.2319667786,0.1899949759,0.0391035639,-0.2794680893,-0.4342857301,-0.0886206329,-0.2365958691,0.1095489264,0.1482250243,0.6309939027,0.0345636867,0.4764270186,0.2947264612,0.0489611253,0.1398259848,0.1815548688,0.4972691238,-0.3258462548,-0.4095653892,-0.0467088632,-0.0461938493,0.0505124666,-0.1531914324,0.1795819253,-0.1120608747,-0.3487883806,0.000088976,-0.2515632212,-0.2498103678,-0.2386271805,-0.0136443963,0.2690052986,0.3906219602,0.1407706589,-0.3225954771,-0.1763848662,-0.0081372689,0.4007481039,0.1035773978,0.2502198517,-0.1815419495,-0.0471423045,-0.4509257674,0.492559284,0.2151459008,0.2648919821,-0.1177540421,-0.1629662067,-0.3294481337,-0.0492410697,0.5928120017,-0.3631767035,0.0443488397,0.1019544229,-0.2081836015,-0.382881999,0.0511292815,0.0076627661,0.3649113178,0.0543546379,0.3508075774,-0.0821149051,-0.0730158612,-0.0869157314,0.2223753035,-0.1245945767,0.0196017139,-0.2743616104,-0.2049434334,-0.143308565,0.0314200111,-0.1528896391,0.1650674194,-0.1800080687,-0.0685375854,-0.1131430715,-0.0563930348,0.1153809577,0.1347035617,0.2945029438,-0.1000063792,0.1187398955,0.1441811323,0.2405287623,0.04236985,0.458391726,-0.2755981982,-0.2780961096,-0.0670714304,0.1208044589,-0.114640221,0.3437377512,-0.2040164918,-0.0173406545,-0.0444572084,-0.1020425931,-0.0212400388,-0.1411033124,-0.001939329,0.3703025281,-0.204925105,-0.2889373302,0.2329640687,0.2390516698,-0.2043491751,0.436775893,0.3422557414,-0.2809350789,0.1687894613,-0.1999594271,0.7136852741,0.1521688104,-0.0298533011,0.0877994671,-0.1695175022,0.4524547458,0.3125461042,0.1559776068,-0.0667682663,-0.2370487005,-0.0655061752,-0.0347471349,0.1580221951,-0.1110610664,-0.2036339194,0.5631492734,-0.0592185557,-0.0353867561,0.1755868793,-0.1173290387,-0.2023077309,0.0190884974,-0.4372626245,0.099125281,-0.0213232394,0.2521654963,0.0010581244,-0.0767222121,0.0552208163,-0.3087654114,-0.2637814581,-0.0523921177,-0.3992073536,-0.0659277439,0.2081235945,-0.433280319,0.0089732604,0.0526266396,0.2566778362,-0.0081807002,-0.0373007692,-0.0075372811,0.0702684447,0.4977722168,0.2345933914,0.0250330679,0.2863307297,0.0890943483,0.1867841035,0.0195984095,-0.1420115829,-0.3608390987,-0.0186531637,0.0038103643,0.1180328727,-0.1588664055,0.0288226064,-0.2428553104,0.0213073175,-0.364861697,0.0650937036,-0.0898337811,-0.0098389015,0.0943212733,-0.0165627208,-0.226817742,-0.0053528175,0.6357886791,-0.1438153237,-0.4367529452,0.127720207,-0.0055662361,-0.2235843986,0.0353587158,-0.2044414431,0.0274193734,-0.2750842869,0.2978782654,-0.2511248887,-0.1071279123,0.2691295445,0.3227922022,0.1886078864,0.2146122009,-0.1516674459,-0.3945676982,-0.1370823681,0.2854580283,0.0442744903,0.0549807698,-0.0311486106,-0.3236714602,-0.1444216073,-0.2066645175,-0.2257885635,0.0087526664,0.3313151598,0.0712086409,0.1904872656,0.2205638587,0.1548970491,-0.1995647997,-0.0134116681,0.0725863725,-0.203820318,-0.102092959,-0.093345888,0.0935418084,-0.0911601558,-0.2648990452,-0.1546195447,-0.0057543651,0.0614032373,-0.1646934003,0.4182681739,0.3214158118,-0.1182498485,-0.0536907464,-0.0327779986,0.0472252741,0.3035388887,0.1383125186,0.0185656752,0.0326717421,0.0427048579,0.1994483471,-0.1875816137,0.014096451,0.082016468,0.0423960686,-0.2510475516,0.0270903334,0.3569348454,-0.295296222,-0.0216397811,0.5987740755,0.3562681079,0.2437597811,-0.1398067921,0.1138302758,-0.0777392983,0.1533572823,-0.2739046812,-0.3326784968,0.205778718,0.2732258439,0.0609572381,0.1860822588,-0.0021066871,0.0291597173,0.6171163917,0.2176484168,0.4478169084,-0.1253162473,0.3508082032,0.5271652937,0.2326287627,0.1843764484,0.241720289,0.0737254471,0.0160929821,0.5080396533,-0.0660088882,0.2586891651,0.2197473645,0.1287705451,-0.0355183855,-0.4662743509,-0.1656748652,0.0283220075,-0.0797490031,0.2506761551,0.1661577374,0.0856602341,-0.2232845426,-0.3345984221,0.0114881797,0.024897946,-0.193626374,-0.2161095738,0.1446396858,-0.3639173806,-0.0571490377,0.013486567,-0.1208442077,0.0703997388,0.4529311359,-0.2469489723,-0.128362909,-0.4731904268,-0.3793772757,0.0500456095,0.1747638434,-0.0140105868,0.4047631919,0.258780688,-0.2298548669,0.2219883502,0.2503442466,0.4856500924,0.2468240112,0.027692223,0.0133466832,-0.1420494765,-0.0540064909,-0.0907230228,0.3356420696,-0.0870835707,0.0305686034,0.1879348457,0.1134100109,0.0007339246,0.2053903639,-0.0408540145,0.3989400268,-0.4661819339,0.0895344913,-0.2369527221,0.2848792076,0.0150147714,0.1000138745,-0.2490254641,-0.1181152612,0.171215117,0.2318110913,0.2039028108,-0.1064854339,0.0748107806,0.1348544508,0.0739088506,0.4687882364,-0.0930132717,-0.0892460048,-0.1464888304,-0.653509438,-0.0558623448,0.1423008889,0.1280672401,-0.0635151789,0.0375300944,0.0490150824,0.1846857667,0.0434734374,-0.1836016178,0.1054587811,-0.2515150607,-0.2000013739,-0.1759469211,0.2301196903,-0.2421481013,0.0548487157,-0.2659107745,0.0627719611,-0.0416926146,0.0269530844,-0.005927782,-0.0628840253,0.0675144792,0.3473338187,-0.1370735615,0.3493901491,0.2663169801,0.03349353,-0.0034476253,-0.2782979906,-0.1931354105,-0.2683142722,0.0700851008,-0.2921240628,0.4918605983,-0.4018596709,-0.3019634485,-0.1804217845,0.2724736333,0.1233783066,-0.1268284917,-0.4767788649,0.1658581197,-0.0587571412,0.0838124454,0.0044377646,0.508733511,0.059986867,0.3379509747,-0.2266310304,-0.3412219286,0.5358169079,-0.4137625992,-0.2617653012,-0.1319275051,-0.0639740229,0.6073155403,-0.4836450815,-0.3043731153,-0.2037671804,0.2719914615,-0.1200053543,-0.1490247399,0.1555725485,-0.2652980089,0.1745556146,-0.1892018616,-0.0870604515,0.4393822849,-0.0882679075,0.2773536742,-0.1299419552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2941","title":"OSCAR unshuffled_original_ko: NonMatchingSplitsSizesError","comments":"I tried `unshuffled_original_da` and it is also not working","body":"## Describe the bug\r\n\r\nCannot download OSCAR `unshuffled_original_ko` due to `NonMatchingSplitsSizesError`.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> dataset = datasets.load_dataset('oscar', 'unshuffled_original_ko')\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=25292102197, num_examples=7345075, dataset_name='oscar'), 'recorded': SplitInfo(name='train', num_bytes=25284578514, num_examples=7344907, dataset_name='oscar')}]\r\n```\r\n\r\n## Expected results\r\n\r\nLoading is successful.\r\n\r\n## Actual results\r\n\r\nLoading throws above error.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.4.0-81-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n","comment_length":9,"text":"OSCAR unshuffled_original_ko: NonMatchingSplitsSizesError\n## Describe the bug\r\n\r\nCannot download OSCAR `unshuffled_original_ko` due to `NonMatchingSplitsSizesError`.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> dataset = datasets.load_dataset('oscar', 'unshuffled_original_ko')\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=25292102197, num_examples=7345075, dataset_name='oscar'), 'recorded': SplitInfo(name='train', num_bytes=25284578514, num_examples=7344907, dataset_name='oscar')}]\r\n```\r\n\r\n## Expected results\r\n\r\nLoading is successful.\r\n\r\n## Actual results\r\n\r\nLoading throws above error.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.4.0-81-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 5.0.0\r\n\nI tried `unshuffled_original_da` and it is also not working","embeddings":[-0.2700701952,-0.3562059104,0.0761256963,0.3269021809,0.2630259991,-0.0228649452,-0.0282609034,0.4201454818,-0.0057781218,0.2675718367,-0.2937448621,0.3795824349,-0.175184533,0.0020705916,-0.1521454006,-0.3282049894,0.0055401283,0.0813263729,0.0597292781,0.0331693552,-0.2423605472,0.2973343432,-0.3931003809,0.0419516861,-0.2768164873,-0.0585272759,-0.0746447295,0.1445952952,-0.1675010622,-0.2716379166,0.3857498765,-0.0588463284,0.232816115,0.5122596025,-0.0001237657,0.0590177849,0.4652545154,-0.3251428604,-0.0443532206,-0.3920538425,-0.0886344537,-0.0983479768,-0.0657844767,-0.3845169842,-0.0129822195,0.146921888,0.0608098917,-0.0645650402,0.5164781213,0.3238581419,0.120624736,0.0500490926,0.0403883979,-0.2224902213,0.3106618822,-0.2548942566,-0.3315367401,0.0346963368,-0.0179840401,-0.1685461998,0.1427149326,0.1541062146,-0.0700180754,0.1261722445,-0.1210525259,-0.0195195898,-0.0040011993,-0.0904255956,0.2117808163,0.374551028,0.3375654519,-0.3247810006,-0.3132323921,-0.3029401302,0.0337986127,-0.3283481598,0.3009405434,0.1460143775,-0.2429262847,0.2017060965,-0.0670581907,0.0023002515,-0.1089840308,0.1734156609,0.0992210284,0.2540248334,0.068739228,0.1693916321,0.1453325152,0.1248955354,0.3683691025,-0.164793402,0.1000258476,0.0660203919,-0.2375979722,-0.1042992249,-0.1702269167,0.0275418647,0.1742199361,0.299749881,0.3533665538,-0.050712835,0.23539868,0.0687230006,0.1073275581,0.1424306631,-0.1735071242,0.4756015539,-0.038373746,0.4154306054,-0.1177193224,-0.0899807885,0.0127301393,-0.0728332028,0.0770530924,0.0638824552,0.3465489149,-0.1475606412,-0.5405621529,-0.0646340847,-0.4408827126,-0.2136350721,-0.1520271748,0.3651462495,-0.2901136577,0.1182295531,-0.1973749697,0.2700160146,-0.1282483786,0.043910075,-0.2218047082,0.1398160458,-0.1717366427,0.1828228682,0.1118824482,-0.204669714,0.495223105,-0.0624024421,0.2177726328,-0.1745524108,0.2678282857,-0.0713035017,-0.2555179,0.0733668804,0.0693256706,0.2275202274,-0.0512226708,-0.1955409944,-0.1835986227,0.1897325665,-0.1677214205,-0.5662183762,0.1641817391,0.1281939894,-0.3114531338,-0.0909331217,0.2412121892,-0.06287238,0.0155445421,-0.1351996511,-0.0021220825,-0.0958602279,0.1657218486,-0.2155779004,0.2271082997,0.3483718336,-0.1850482076,-0.1841768026,0.006626484,-0.2098627985,0.5653936863,0.3381557167,0.0346748829,-0.0691285431,-0.3662846684,-0.035242945,0.2375092804,-0.1608842909,-0.6755478978,0.0946134478,0.0492238738,0.5803261995,0.346126765,-0.0058211032,0.2652187347,0.1256498843,0.2001138926,0.2200994045,-0.2519147992,-0.3296817243,-0.4716974497,-0.4451117814,0.1734127849,0.1898487657,0.3207941949,0.1099775285,-0.058488369,0.2988939583,0.3775659204,0.0346811637,0.0107769752,-0.0222215783,0.1482665092,0.3807443678,-0.076375179,-0.3940159082,-0.2746164203,0.2904383838,-0.1236760765,-0.156783849,-0.2122336924,-0.527250886,-0.5870789289,-0.165048629,-0.1906133443,-0.162727192,0.0038345398,0.0500118583,0.4496047795,0.1891618371,-0.1783928275,0.2574459016,0.0525147803,0.1223757267,-0.1573136151,0.4813087285,-0.2699065506,-0.4087940454,0.1640718877,0.1615529507,0.2917622626,-0.2825148404,-0.1289041787,0.5438861251,-0.0126622617,0.1553732008,-0.0856645703,-0.2544069886,0.1181099713,-0.167449981,-0.1110598743,0.3721723258,0.0395096019,0.0105408933,-0.0509868823,0.5377941132,-0.0667465031,-0.0382976942,0.066784896,0.0064676078,0.2481441945,-0.1786244661,0.0006179169,-0.1696993709,0.3078433871,-0.0635774508,-0.0126545299,-0.1987084448,-0.5619574189,-0.1079929173,0.7206522226,-0.0952707604,0.0389066115,0.0032789083,-0.1174628735,0.1458038688,-0.0419397578,0.1237870455,0.2527246773,0.3046537042,-0.0554859377,0.1650494337,-0.0884823874,-0.2226391435,0.1139543355,0.0752876177,0.3927039802,0.5039468408,-0.1236994937,0.0763900355,-0.4040017426,-0.1757810861,-0.0385836139,0.3962652385,-0.0995420963,-0.251657933,-0.3585939407,-0.0895016566,-0.3763104379,0.0013211808,-0.1984645426,-0.1646792293,0.0933002308,0.4677838087,0.03062311,0.456158936,-0.2225827277,0.0516182445,-0.0948610157,0.0804057792,-0.2349837869,-0.2070280761,-0.2165757865,-0.0062309718,0.1760603487,-0.1895347685,0.4101438522,-0.1628828794,-0.1106793508,0.0016370228,-0.2100382894,-0.0284292269,0.0802166983,-0.1457641721,0.1499073058,0.1545280069,0.0492399521,-0.2327251136,0.3539457917,-0.253189683,-0.12509799,-0.0579356998,0.3628317118,0.1966164857,-0.0150078526,-0.6001400352,-0.3433532715,-0.2735929787,0.1192297861,0.171149075,0.2071451992,-0.4059786499,0.2124399543,0.0293269083,0.0146261109,-0.108423464,-0.1782411635,-0.2187789083,0.4179231524,-0.0541067719,-0.2413789183,-0.0135591552,-0.0757790357,-0.207618624,0.1081376076,-0.2254442275,-0.2374272496,-0.1528240591,0.0388922058,0.0000410417,-0.1447942406,0.4625220001,-0.0406659842,0.0200901981,-0.087360315,-0.2536268234,0.0319263563,-0.0870750993,0.4643537402,-0.0956633613,0.5923442245,0.1564027518,1.0513212681,0.3848242462,0.4277364314,0.2050092816,-0.2445764095,-0.0273402222,-0.1820255518,-0.1257591844,-0.0314858034,-0.1415696144,0.1885153949,0.1321771443,-0.0758097246,-0.1143291369,-0.2025897503,0.0363576189,-0.4454913139,-0.1845194995,-0.050208386,0.0200488307,-0.0743915215,0.2265842706,0.1607146412,0.0683756843,-0.0109747304,0.1947402656,0.2533482909,-0.095236063,0.0733086839,-0.1778615266,0.0034823706,-0.4660815597,0.4068314433,0.1445844173,0.155201152,0.105390504,-0.2482506037,0.2122097313,-0.2690775096,0.8391341567,-0.3196516037,0.021446595,-0.0050391215,-0.1162589341,-0.2233094126,-0.3100708723,-0.2493042648,0.0977460891,0.5004122257,0.401702255,-0.3936049342,0.1048643887,0.1261661202,-0.0172707625,-0.1737206727,-0.1427190602,-0.3067505658,-0.416105777,-0.521835506,0.0087642455,-0.10390421,0.4622172713,-0.103750132,-0.1932160407,-0.2216338515,-0.0615733713,0.1528482884,0.0589452758,0.1914952844,0.0529781282,0.2729120851,0.0143474173,0.5659828782,0.3270242512,0.433042109,-0.0510436259,-0.2842541039,-0.2222711891,-0.2774918079,-0.0777949691,0.3124807179,0.0741247982,0.0557200424,-0.2258193046,0.0286223143,-0.0232267417,-0.0609401241,0.2382368445,-0.074377045,-0.4884041846,-0.4397383332,0.2568946183,0.2011566609,-0.084970735,0.2157151252,0.1666444838,-0.1530286074,0.3322291672,-0.311853826,1.0656917095,-0.0349132605,-0.0590303577,0.1567310691,-0.4327569604,0.4600604773,0.2230525613,0.2934070528,-0.3985666931,-0.5422777534,-0.0626290441,-0.0495161489,0.2620194554,-0.0702634975,-0.2358459234,0.3924998343,-0.1383437216,0.059872508,-0.0403107777,0.0872835815,-0.0655479878,0.2615909278,-0.2137582153,-0.0282094814,-0.0062475386,0.6459135413,0.0490520187,-0.1735088825,-0.3161623478,-0.1341487914,-0.2399571091,0.2692095339,-0.0620518588,0.1432400197,-0.1940255165,-0.3515720665,-0.0938458592,-0.0665402114,0.1554252207,-0.3038245738,-0.1423941851,0.2567335367,-0.0502586178,-0.3049413562,0.0468185209,0.0734311789,0.1184430718,-0.2098308355,-0.1948067993,0.2171829045,0.081996724,-0.2234715521,0.2245999873,-0.344912529,0.4029507935,-0.2193651795,-0.3630694151,-0.1238867566,0.053775616,-0.4139072001,0.0482330695,-0.3476900756,-0.1793532968,0.0908351913,-0.0968465805,-0.1877360195,-0.2056991607,0.6468623877,0.2441388518,0.0889284089,0.4455735385,-0.1392742097,-0.1209909022,-0.0694017634,-0.0522745661,-0.0961838588,-0.6258867979,0.2478279322,-0.2725172639,0.2561436892,0.0560390577,0.1203305498,0.5102044344,0.1091660559,-0.0226831455,-0.315593183,-0.1827117205,-0.006129405,0.3605924249,0.2034677416,-0.2084128559,0.3939769268,0.0435558408,0.1410049051,-0.2295722514,0.1756058782,-0.0111234821,0.0903083682,-0.0150688291,-0.0624070913,0.0482578315,0.1309499592,0.0656671524,0.2593142688,-0.107485652,-0.1093422994,-0.1164527461,0.217103675,-0.0287401099,-0.0509174541,-0.0877763107,0.0597584993,-0.014149596,0.0015855627,-0.0073475796,0.3024415374,0.0025629182,0.1959821284,0.2877029181,0.0203361772,-0.3707616925,-0.0546788871,-0.1704271585,0.4047979116,-0.0256696772,0.2317973226,-0.0267457049,-0.0181804169,-0.0715063438,-0.3175280988,-0.0306168441,0.2115057111,0.3874916732,-0.347990185,-0.0214854032,0.3729046881,0.2182453871,0.5003054142,-0.1019885391,-0.1947451532,0.2794112563,0.132047534,-0.5035033822,0.1693591028,0.6532273293,-0.0279648732,0.0474482439,0.2889039218,-0.0722825229,-0.0049079503,-0.2585618496,0.0728518143,0.311209619,0.0896963254,0.3967244029,0.6500381827,0.2969564795,0.0609108508,0.1803033948,0.0377423726,0.3314584792,0.4642525315,-0.348226279,0.2353744805,-0.3763629198,0.2772921026,-0.0461748801,-0.2185561657,-0.2172773033,0.3019171655,-0.263828367,0.2700886428,0.147561118,0.2971718013,-0.0907789618,-0.0566396452,-0.2395830452,0.0068022935,0.0361580849,-0.1487154663,-0.23768875,-0.3082600832,-0.3154895604,0.1420746595,0.0483601578,0.2234884501,0.4846854806,-0.0899223089,-0.0839001238,-0.261703372,-0.1129587889,0.1009543613,0.4807308614,-0.359243691,0.002184781,0.1473571062,-0.2205985636,0.0831326023,0.3578595817,0.2714527547,0.2029879987,-0.224526599,0.0407990441,0.2606113255,0.0910466984,0.1569049805,0.3494497836,0.3068242073,-0.0663663,0.2790664434,0.0528981723,-0.1823626012,0.1601461917,0.0584523901,-0.1115810052,-0.2788190246,0.3917230964,-0.2243009657,0.102521576,0.057680171,0.0891512558,-0.4057340622,-0.2379486412,-0.2201809883,0.0126067307,0.231259793,0.0904681012,0.0248190816,-0.3940022886,0.1682776958,0.2681089044,-0.1561311036,-0.1225692406,-0.2420620918,-0.5433617234,0.2975563407,-0.2807193995,0.0935659185,0.0746108741,0.3382504284,-0.3834771514,-0.0064511839,0.1178494468,0.5332386494,0.0087339627,0.4412532747,-0.2814124823,-0.1415888518,-0.226051569,-0.0913680568,0.1125744134,-0.3669516146,0.0487546846,-0.2756848037,-0.0543771461,-0.0466812104,-0.121877186,-0.0234336779,-0.1646656841,0.2823671103,0.2395053953,0.2693460882,-0.1665234864,-0.0291301347,0.1338399649,-0.1567696631,-0.065012753,0.3058659434,0.3186143041,0.2530109882,-0.0928714797,0.2424036115,-0.1774176359,0.5707028508,0.1469569206,-0.1631754786,-0.10322088,-0.1555235088,-0.0427126624,0.0198512953,0.0677086264,0.4105406106,-0.1185427532,0.1515361816,-0.2512541115,-0.2376576364,0.6050724983,-0.1785651296,-0.1379883289,-0.1678245664,0.2456205785,0.3388739526,0.0138582885,-0.3756624758,0.1801710874,0.4207406342,0.0298809856,-0.0401440747,0.1496393681,-0.3060053885,-0.1257539541,0.0442536511,0.0996872783,0.149877876,-0.3360651135,0.4139497876,-0.1505003273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2937","title":"load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied","comments":"Hi @daqieq, thanks for reporting.\r\n\r\nUnfortunately, I was not able to reproduce this bug:\r\n```ipython\r\nIn [1]: from datasets import load_dataset\r\n ...: ds = load_dataset('wiki_bio')\r\nDownloading: 7.58kB [00:00, 26.3kB\/s]\r\nDownloading: 2.71kB [00:00, ?B\/s]\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\\r\n1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nDownloading: 334MB [01:17, 4.32MB\/s]\r\nDataset wiki_bio downloaded and prepared to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9. Subsequent calls will reuse thi\r\ns data.\r\n```\r\n\r\nThis kind of error messages usually happen because:\r\n- Your running Python script hasn't write access to that directory\r\n- You have another program (the File Explorer?) already browsing inside that directory","body":"## Describe the bug\r\nStandard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wiki_bio')\r\n```\r\n\r\n## Expected results\r\nIt is expected that the dataset downloads without any errors.\r\n\r\n## Actual results\r\nPermissionError see trace below:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\contextlib.py\", line 120, in __exit__\r\n next(self.gen)\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 598, in incomplete_dir\r\n os.rename(tmp_dir, dirname)\r\nPermissionError: [WinError 5] Access is denied: 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9.incomplete' -> 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9'\r\n```\r\nBy commenting out the os.rename() [L604](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L604) and the shutil.rmtree() [L607](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L607) lines, in my virtual environment, I was able to get the load process to complete, rename the directory manually and then rerun the `load_dataset('wiki_bio')` to get what I needed.\r\n\r\nIt seems that os.rename() in the `incomplete_dir` content manager is the culprit. Here's another project [Conan](https:\/\/github.com\/conan-io\/conan\/issues\/6560) with similar issue with os.rename() if it helps debug this issue.\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Windows-10-10.0.22449-SP0\r\n- Python version: 3.8.12\r\n- PyArrow version: 5.0.0\r\n","comment_length":109,"text":"load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied\n## Describe the bug\r\nStandard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wiki_bio')\r\n```\r\n\r\n## Expected results\r\nIt is expected that the dataset downloads without any errors.\r\n\r\n## Actual results\r\nPermissionError see trace below:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\contextlib.py\", line 120, in __exit__\r\n next(self.gen)\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 598, in incomplete_dir\r\n os.rename(tmp_dir, dirname)\r\nPermissionError: [WinError 5] Access is denied: 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9.incomplete' -> 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9'\r\n```\r\nBy commenting out the os.rename() [L604](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L604) and the shutil.rmtree() [L607](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L607) lines, in my virtual environment, I was able to get the load process to complete, rename the directory manually and then rerun the `load_dataset('wiki_bio')` to get what I needed.\r\n\r\nIt seems that os.rename() in the `incomplete_dir` content manager is the culprit. Here's another project [Conan](https:\/\/github.com\/conan-io\/conan\/issues\/6560) with similar issue with os.rename() if it helps debug this issue.\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Windows-10-10.0.22449-SP0\r\n- Python version: 3.8.12\r\n- PyArrow version: 5.0.0\r\n\nHi @daqieq, thanks for reporting.\r\n\r\nUnfortunately, I was not able to reproduce this bug:\r\n```ipython\r\nIn [1]: from datasets import load_dataset\r\n ...: ds = load_dataset('wiki_bio')\r\nDownloading: 7.58kB [00:00, 26.3kB\/s]\r\nDownloading: 2.71kB [00:00, ?B\/s]\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\\r\n1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nDownloading: 334MB [01:17, 4.32MB\/s]\r\nDataset wiki_bio downloaded and prepared to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9. Subsequent calls will reuse thi\r\ns data.\r\n```\r\n\r\nThis kind of error messages usually happen because:\r\n- Your running Python script hasn't write access to that directory\r\n- You have another program (the File Explorer?) already browsing inside that directory","embeddings":[-0.2289305478,0.3834535182,0.040345341,0.2550398409,-0.0121113108,0.2622572184,0.5036683083,0.1329532564,0.3442817032,0.1357175708,-0.1089743897,0.0772839189,0.073150374,-0.0056987684,-0.1256225109,0.0789283216,0.0364452153,-0.0053882273,0.214949578,0.1039548293,-0.4238833189,0.1072954684,-0.172855407,0.1861151755,0.0260949377,-0.1363423616,0.1212603897,0.3607298434,-0.0307886619,-0.3922062814,0.3530231118,0.0634532124,0.4639182985,0.2582460642,-0.000119246,-0.0137069086,0.2306841165,0.1244994402,-0.0737274736,0.0870998204,-0.1839853674,-0.1409676671,0.1652300805,-0.0976862535,0.1767440438,0.3634920418,-0.1021372527,-0.2009297311,-0.0130413789,0.379034102,0.1573001444,-0.0381855965,-0.1186785623,-0.0752038136,0.4933359921,-0.1592982262,-0.133179754,0.2798380852,0.0790250301,-0.11457026,-0.0989914164,0.1995727867,-0.1621495485,0.1471577436,0.0186581202,0.0311292373,0.2162310332,-0.4581103027,0.3521417677,0.1512579024,0.9656507373,-0.1955735534,-0.1914045215,-0.0887490958,0.0133037586,-0.0330028497,0.4326035678,0.4286232889,-0.3245685697,0.2523709536,0.0659693852,0.0835653022,-0.0825510696,-0.0019090191,0.0712853372,0.1486521512,-0.0504703857,0.1195566207,-0.1147312596,-0.0834005401,0.3863353431,-0.2138270736,-0.254601866,0.191666767,-0.3394370973,0.2898705602,0.0534532256,0.7778053284,0.157877773,0.3155301809,0.1250241548,-0.0540518314,0.2136244029,0.0651903003,0.0085929586,0.3214896023,-0.1122384295,-0.0169230793,0.4271060824,0.1174419895,-0.0816169605,-0.0089906445,-0.0200457647,-0.4294741154,0.3538792133,0.1054070443,0.4081188738,-0.2856419981,-0.1217012107,-0.0002951851,0.1581317037,-0.0239709932,0.0588830411,0.5011533499,-0.2717713416,0.0426839963,0.0682517141,0.1794451177,0.0186107904,0.0977605507,-0.0877998695,-0.0890028104,-0.2667180896,-0.0514384918,0.408646673,-0.3225747645,0.0955644995,0.2566769719,0.0270482041,-0.1792946011,0.1918320805,-0.0414977111,-0.0328572541,0.3576870263,0.397295326,0.3376339078,0.174932614,-0.2122963965,-0.0349953882,0.1863114685,-0.4872902334,-0.4025040567,-0.0336632058,0.1213277355,-0.0833240524,0.0632052422,-0.1980368942,-0.0482573472,0.3823129535,-0.3008547723,-0.0618359111,-0.0137255481,-0.2206203043,-0.5042767525,0.040707428,0.4012890756,-0.3596991003,0.1303672194,-0.0915537477,-0.2640866041,0.1504325122,0.0056375656,0.0056341127,0.3616814017,-0.2812198102,0.0156073133,0.2774584889,-0.3938943148,-1.0015972853,-0.0210030079,0.0730995089,0.1660424769,0.2444063723,0.1191371009,0.3659378588,-0.0303458292,0.1704769284,0.3187128901,0.1300752312,0.1119742319,-0.0629962236,-0.2038549483,0.2796608508,0.1339266002,0.0883391201,0.3009034991,0.312600106,-0.0569339506,0.251095593,-0.1872612089,0.116804406,0.1925157905,0.0114956526,-0.0351362228,0.1375146806,-0.1277379841,-0.420409292,0.3514482081,-0.00811728,0.0851668194,0.0759790465,-0.2027431279,-0.3184642494,0.0913407728,-0.5447552204,-0.2889970243,0.0851063654,0.2938987911,0.2124249339,0.0236818604,0.1891641766,0.3374149799,-0.2082268894,0.1271959245,-0.1811406165,0.0386061855,-0.1883375496,-0.1364492327,-0.2122476399,0.0469153933,0.2001280636,0.2527483404,-0.2134325951,0.1961012632,0.151213333,0.0374986753,-0.3785572946,-0.0595560744,-0.0223776083,0.1217297763,0.1568485051,0.1381781846,0.1483232975,-0.0443943143,-0.1922187656,0.1377540231,-0.1596345007,0.14373146,-0.3266511559,0.2430362701,0.2447061241,0.0900503173,0.0856966749,0.0344529897,0.3054705858,0.0241174567,0.0745961443,-0.2858042717,0.1012615561,-0.0778078139,0.0593021847,0.0225473158,0.0812815726,0.2497533113,-0.1746012717,0.0168862902,0.2869935334,0.349679023,0.5377136469,-0.0409865156,0.0515500903,0.0667970031,0.0479120463,0.0896691009,0.1192515045,0.0857714713,-0.0702273324,0.2634492815,0.1284338385,0.2241659462,-0.0647975802,-0.1138611957,-0.107541278,0.4194297791,-0.4875138104,-0.084387593,0.1000552922,0.0748399198,0.1812013388,-0.2428474277,-0.0887820199,-0.1460988075,-0.2430461347,0.3116364777,0.0940705463,0.0835373625,-0.511957705,0.164441213,0.0075377976,-0.2018063366,-0.1137571484,0.1238797531,-0.1658339351,-0.0141071854,0.1862449348,-0.1799398512,0.1873875707,-0.2951588631,0.2437517196,-0.3901760876,-0.1954325885,0.0151099013,0.0578858256,0.4098414481,0.1254311204,0.1551068723,0.1864105314,-0.3417646587,0.0566025972,-0.1075194404,-0.1510609835,0.2333712131,-0.0550265461,0.0938554928,0.0341966748,-0.2493757308,-0.5408244133,-0.2501789927,-0.0095754005,0.3858899772,0.0983009189,-0.0028510524,-0.1126530915,0.0033437847,0.267029494,0.3937678933,-0.2108280957,-0.2492020875,0.2102492452,-0.2092680335,-0.3346993625,0.0819493756,-0.1042304859,0.0445711389,0.2523835003,-0.4788489342,-0.0906342939,-0.3600659072,0.4096082449,-0.232615903,0.0938282236,0.3463047147,0.0892930999,0.1463989317,-0.0417488404,-0.2757158875,-0.0651353374,-0.2946413159,-0.068232581,0.3381980062,0.4274490774,-0.0575702377,0.3770318925,0.3002279699,0.3604496717,0.3344713151,0.04584153,0.3711133003,0.0542609356,-0.2594737113,0.1087757051,-0.1943459958,-0.140160799,0.3007430136,0.0889823362,-0.1952640116,-0.052531682,-0.3804523051,-0.2807892859,-0.2852858603,-0.0992422476,-0.2463323325,0.2537561357,0.1191406399,0.1100426763,-0.4462445676,-0.3691142201,0.0288443286,0.5546328425,-0.3098391891,0.1115756556,-0.3508035243,0.1366987079,-0.6549720168,0.2522459924,0.0340248011,0.4958057404,-0.1741242707,0.1266969591,-0.0015080259,-0.1074288562,0.5570681095,-0.6522978544,0.4618299305,0.0581227317,0.1157312021,0.0101971151,-0.2243451625,-0.0922563747,0.2416802943,0.1369419247,0.2145550996,0.1639221013,0.0440419056,0.0762091056,0.1665931493,-0.2001322806,-0.279057771,-0.2271348834,-0.3418877125,-0.3903853595,0.128006041,0.0421996191,0.1854700744,-0.1517433375,0.1171616763,-0.4046885669,0.2369778156,0.1301799268,0.1602822691,0.187791869,0.0645751134,0.2181593776,-0.0480453596,0.0753483474,-0.1291563809,0.5240867138,-0.2804769576,-0.328379035,-0.0524905957,-0.199985683,-0.1171171516,0.1122419834,-0.3265575469,-0.0500914566,-0.1384819001,-0.1327372044,-0.3143305779,0.117074959,0.0426035039,-0.1354341507,-0.3776838183,-0.3201939762,0.3641437292,0.2117004246,-0.4467277825,0.3404740095,0.1198474169,-0.4280630946,0.1824693233,0.0517599583,0.7466567159,0.0798895434,0.2780212462,0.0972766355,-0.1122953147,0.2001173049,0.1161781773,0.0565570518,-0.3728869557,-0.4541019201,-0.0988719016,-0.2205425799,-0.134537518,-0.0249406099,-0.2337937057,0.2817453742,-0.1129783094,0.3188050985,0.274310261,0.048037339,-0.3741956651,-0.3764992654,-0.1209942549,0.1251865476,-0.1938239634,0.1358688474,-0.2274900377,0.0911153406,0.0849043131,-0.1673949957,-0.439121902,0.1967898756,-0.4114460945,0.1381170154,-0.2974178493,-0.3836832643,0.0877998993,0.5960645676,0.2849669755,0.3604593873,-0.1543026865,0.0600666888,0.0793892592,0.1179443896,0.024958102,0.0074552288,0.1962639689,-0.0806936249,-0.2981881797,-0.0151690254,-0.2781275809,-0.3215303123,-0.1055536345,0.2862682045,0.4039052129,-0.3201119602,-0.2076887786,-0.3106458187,0.0849739909,-0.2798955739,0.0937029347,0.0470425822,-0.0756008551,0.3795401752,-0.3368881643,-0.425712347,-0.1475177705,0.5242761374,-0.079077132,-0.1337566078,0.5711378455,-0.0485029556,-0.1785390973,-0.0130321439,0.1454919577,-0.1009839401,-0.462458998,-0.1003440917,-0.1225365251,0.1078282446,-0.029106807,0.1698505282,0.2787975669,-0.0963102505,0.0534037836,-0.7581394315,-0.2621001899,-0.1129847094,-0.2311439961,0.1716352552,0.1831110716,-0.4405415952,-0.1353047937,-0.2840806246,-0.203002885,0.0441007577,-0.2147331834,-0.1294415742,0.2779014409,0.154557094,-0.001340596,0.1245135888,-0.0026244703,0.0926584527,0.1227548569,-0.1713698655,-0.1237100959,0.1475240737,0.1037818864,-0.0917996764,0.0267537702,-0.3284253776,-0.0735971555,-0.0984259099,0.3940449953,0.2593082786,-0.0281576384,0.0679169372,-0.1276832074,0.2894711792,-0.2827050388,0.4151190817,-0.3472835422,-0.0251300745,-0.0133727901,0.3347693682,-0.1413164884,0.0541265719,-0.11745359,-0.0824945197,0.1532541662,0.0678972974,0.0646063909,-0.3769121468,0.085026145,0.1787576079,0.0593337528,0.0623646304,-0.23318322,0.0148412948,0.2450501472,0.1589923799,-0.0922699794,-0.1294877678,-0.0120077264,-0.3234773874,-0.1790443361,0.3004153967,0.1639576256,-0.3836256266,0.1359916031,-0.0324052311,0.5581638813,-0.2246947289,-0.0577207766,0.4813490808,-0.2174006104,0.2738089263,0.218585968,0.2596439123,-0.0843797997,0.5514788032,0.1820428967,0.1367382705,-0.0058679264,-0.1720480621,-0.1030526981,-0.6628072858,0.28002882,0.0652534962,-0.0736904144,0.0509437881,-0.0296167657,-0.0086359354,-0.3041151464,-0.0021537442,-0.4382816255,-0.2168946117,-0.1808834672,-0.283659488,0.0684353933,-0.2338435054,0.1725392044,0.0927761719,-0.1707766503,-0.14519687,-0.069977276,-0.0427515879,-0.3035758734,-0.268468231,0.1116327941,0.4133677781,-0.2662928104,-0.1926916391,0.0890493989,0.0064425515,0.2395810932,0.0886423066,0.0961362123,0.7200041413,0.3696607947,-0.186706394,-0.0827123448,0.2043794394,0.08498016,-0.1013351902,0.3095628321,0.1992664337,0.1772322208,0.3453566134,0.1196848005,-0.124466449,0.0374585725,0.0380601026,-0.008947121,-0.0834797546,0.1873595268,-0.1764346659,-0.168423444,0.1405667216,0.1980234534,-0.0727778077,0.1700269282,0.2323683053,0.0627374277,0.1945085675,-0.1900534481,0.0200616512,0.1487799883,0.3155461252,0.3503623307,0.0980699509,-0.2335534841,-0.068358846,-0.5775797367,0.2545953691,-0.0182250105,-0.3800154626,-0.0322349034,-0.1324881017,-0.3368008435,0.1061356142,0.1619229615,-0.1090964749,0.0555578545,0.2668960989,-0.1567609012,-0.2936111093,0.3894377947,-0.0377689339,-0.0781006962,-0.6786634922,0.0366012938,-0.0887601152,0.0320076123,-0.1506992131,-0.1666385084,0.0341500044,-0.0260742661,0.5437771082,0.2644129097,0.4555599689,0.0078377938,0.0918127969,-0.562205255,-0.0687982887,-0.1452899277,-0.0336945578,-0.077826038,0.271348238,-0.2275042981,0.0414364263,-0.1241922826,0.1670939773,0.2662992477,0.3795417845,-0.0054561738,0.2279049158,0.1809432358,0.1571899056,0.3478443921,0.0699722916,-0.0961795077,0.4460362792,-0.2409338355,-0.342048198,0.2109487653,-0.3941510618,-0.3340475857,-0.0192074794,0.3251572549,-0.2549189031,-0.1531336755,-0.2379637361,0.2061088979,0.2478395253,-0.1434804201,-0.3868976235,0.1413453221,-0.1570352912,0.140291512,-0.0227520186,0.0342236757,0.0101351459,-0.225390628,0.1708674133,-0.2374369949]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2937","title":"load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied","comments":"Thanks @albertvillanova for looking at it! I tried on my personal Windows machine and it downloaded just fine.\r\n\r\nRunning on my work machine and on a colleague's machine it is consistently hitting this error. It's not a write access issue because the `.incomplete` directory is written just fine. It just won't rename and then it deletes the directory in the `finally` step. Also the zip file is written and extracted fine in the downloads directory.\r\n\r\nThat leaves another program that might be interfering, and there are plenty of those in my work machine ... (full antivirus, data loss prevention, etc.). So the question remains, why not extend the `try` block to allow catching the error and circle back to the rename after the unknown program is finished doing its 'stuff'. This is the approach that I read about in the linked repo (see my comments above).\r\n\r\nIf it's not high priority, that's fine. However, if someone were to write an PR that solved this issue in our environment in an `except` clause, would it be reviewed for inclusion in a future release? Just wondering whether I should spend any more time on this issue.","body":"## Describe the bug\r\nStandard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wiki_bio')\r\n```\r\n\r\n## Expected results\r\nIt is expected that the dataset downloads without any errors.\r\n\r\n## Actual results\r\nPermissionError see trace below:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\contextlib.py\", line 120, in __exit__\r\n next(self.gen)\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 598, in incomplete_dir\r\n os.rename(tmp_dir, dirname)\r\nPermissionError: [WinError 5] Access is denied: 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9.incomplete' -> 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9'\r\n```\r\nBy commenting out the os.rename() [L604](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L604) and the shutil.rmtree() [L607](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L607) lines, in my virtual environment, I was able to get the load process to complete, rename the directory manually and then rerun the `load_dataset('wiki_bio')` to get what I needed.\r\n\r\nIt seems that os.rename() in the `incomplete_dir` content manager is the culprit. Here's another project [Conan](https:\/\/github.com\/conan-io\/conan\/issues\/6560) with similar issue with os.rename() if it helps debug this issue.\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Windows-10-10.0.22449-SP0\r\n- Python version: 3.8.12\r\n- PyArrow version: 5.0.0\r\n","comment_length":194,"text":"load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied\n## Describe the bug\r\nStandard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wiki_bio')\r\n```\r\n\r\n## Expected results\r\nIt is expected that the dataset downloads without any errors.\r\n\r\n## Actual results\r\nPermissionError see trace below:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset wiki_bio\/default (download: 318.53 MiB, generated: 736.94 MiB, post-processed: Unknown size, total: 1.03 GiB) to C:\\Users\\username\\.cache\\huggingface\\datasets\\wiki_bio\\default\\1.1.0\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\contextlib.py\", line 120, in __exit__\r\n next(self.gen)\r\n File \"C:\\Users\\username\\.conda\\envs\\hf\\lib\\site-packages\\datasets\\builder.py\", line 598, in incomplete_dir\r\n os.rename(tmp_dir, dirname)\r\nPermissionError: [WinError 5] Access is denied: 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9.incomplete' -> 'C:\\\\Users\\\\username\\\\.cache\\\\huggingface\\\\datasets\\\\wiki_bio\\\\default\\\\1.1.0\\\\5293ce565954ba965dada626f1e79684e98172d950371d266bf3caaf87e911c9'\r\n```\r\nBy commenting out the os.rename() [L604](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L604) and the shutil.rmtree() [L607](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L607) lines, in my virtual environment, I was able to get the load process to complete, rename the directory manually and then rerun the `load_dataset('wiki_bio')` to get what I needed.\r\n\r\nIt seems that os.rename() in the `incomplete_dir` content manager is the culprit. Here's another project [Conan](https:\/\/github.com\/conan-io\/conan\/issues\/6560) with similar issue with os.rename() if it helps debug this issue.\r\n\r\n## Environment info\r\n- `datasets` version: 1.12.1\r\n- Platform: Windows-10-10.0.22449-SP0\r\n- Python version: 3.8.12\r\n- PyArrow version: 5.0.0\r\n\nThanks @albertvillanova for looking at it! I tried on my personal Windows machine and it downloaded just fine.\r\n\r\nRunning on my work machine and on a colleague's machine it is consistently hitting this error. It's not a write access issue because the `.incomplete` directory is written just fine. It just won't rename and then it deletes the directory in the `finally` step. Also the zip file is written and extracted fine in the downloads directory.\r\n\r\nThat leaves another program that might be interfering, and there are plenty of those in my work machine ... (full antivirus, data loss prevention, etc.). So the question remains, why not extend the `try` block to allow catching the error and circle back to the rename after the unknown program is finished doing its 'stuff'. This is the approach that I read about in the linked repo (see my comments above).\r\n\r\nIf it's not high priority, that's fine. However, if someone were to write an PR that solved this issue in our environment in an `except` clause, would it be reviewed for inclusion in a future release? Just wondering whether I should spend any more time on this issue.","embeddings":[-0.2289305478,0.3834535182,0.040345341,0.2550398409,-0.0121113108,0.2622572184,0.5036683083,0.1329532564,0.3442817032,0.1357175708,-0.1089743897,0.0772839189,0.073150374,-0.0056987684,-0.1256225109,0.0789283216,0.0364452153,-0.0053882273,0.214949578,0.1039548293,-0.4238833189,0.1072954684,-0.172855407,0.1861151755,0.0260949377,-0.1363423616,0.1212603897,0.3607298434,-0.0307886619,-0.3922062814,0.3530231118,0.0634532124,0.4639182985,0.2582460642,-0.000119246,-0.0137069086,0.2306841165,0.1244994402,-0.0737274736,0.0870998204,-0.1839853674,-0.1409676671,0.1652300805,-0.0976862535,0.1767440438,0.3634920418,-0.1021372527,-0.2009297311,-0.0130413789,0.379034102,0.1573001444,-0.0381855965,-0.1186785623,-0.0752038136,0.4933359921,-0.1592982262,-0.133179754,0.2798380852,0.0790250301,-0.11457026,-0.0989914164,0.1995727867,-0.1621495485,0.1471577436,0.0186581202,0.0311292373,0.2162310332,-0.4581103027,0.3521417677,0.1512579024,0.9656507373,-0.1955735534,-0.1914045215,-0.0887490958,0.0133037586,-0.0330028497,0.4326035678,0.4286232889,-0.3245685697,0.2523709536,0.0659693852,0.0835653022,-0.0825510696,-0.0019090191,0.0712853372,0.1486521512,-0.0504703857,0.1195566207,-0.1147312596,-0.0834005401,0.3863353431,-0.2138270736,-0.254601866,0.191666767,-0.3394370973,0.2898705602,0.0534532256,0.7778053284,0.157877773,0.3155301809,0.1250241548,-0.0540518314,0.2136244029,0.0651903003,0.0085929586,0.3214896023,-0.1122384295,-0.0169230793,0.4271060824,0.1174419895,-0.0816169605,-0.0089906445,-0.0200457647,-0.4294741154,0.3538792133,0.1054070443,0.4081188738,-0.2856419981,-0.1217012107,-0.0002951851,0.1581317037,-0.0239709932,0.0588830411,0.5011533499,-0.2717713416,0.0426839963,0.0682517141,0.1794451177,0.0186107904,0.0977605507,-0.0877998695,-0.0890028104,-0.2667180896,-0.0514384918,0.408646673,-0.3225747645,0.0955644995,0.2566769719,0.0270482041,-0.1792946011,0.1918320805,-0.0414977111,-0.0328572541,0.3576870263,0.397295326,0.3376339078,0.174932614,-0.2122963965,-0.0349953882,0.1863114685,-0.4872902334,-0.4025040567,-0.0336632058,0.1213277355,-0.0833240524,0.0632052422,-0.1980368942,-0.0482573472,0.3823129535,-0.3008547723,-0.0618359111,-0.0137255481,-0.2206203043,-0.5042767525,0.040707428,0.4012890756,-0.3596991003,0.1303672194,-0.0915537477,-0.2640866041,0.1504325122,0.0056375656,0.0056341127,0.3616814017,-0.2812198102,0.0156073133,0.2774584889,-0.3938943148,-1.0015972853,-0.0210030079,0.0730995089,0.1660424769,0.2444063723,0.1191371009,0.3659378588,-0.0303458292,0.1704769284,0.3187128901,0.1300752312,0.1119742319,-0.0629962236,-0.2038549483,0.2796608508,0.1339266002,0.0883391201,0.3009034991,0.312600106,-0.0569339506,0.251095593,-0.1872612089,0.116804406,0.1925157905,0.0114956526,-0.0351362228,0.1375146806,-0.1277379841,-0.420409292,0.3514482081,-0.00811728,0.0851668194,0.0759790465,-0.2027431279,-0.3184642494,0.0913407728,-0.5447552204,-0.2889970243,0.0851063654,0.2938987911,0.2124249339,0.0236818604,0.1891641766,0.3374149799,-0.2082268894,0.1271959245,-0.1811406165,0.0386061855,-0.1883375496,-0.1364492327,-0.2122476399,0.0469153933,0.2001280636,0.2527483404,-0.2134325951,0.1961012632,0.151213333,0.0374986753,-0.3785572946,-0.0595560744,-0.0223776083,0.1217297763,0.1568485051,0.1381781846,0.1483232975,-0.0443943143,-0.1922187656,0.1377540231,-0.1596345007,0.14373146,-0.3266511559,0.2430362701,0.2447061241,0.0900503173,0.0856966749,0.0344529897,0.3054705858,0.0241174567,0.0745961443,-0.2858042717,0.1012615561,-0.0778078139,0.0593021847,0.0225473158,0.0812815726,0.2497533113,-0.1746012717,0.0168862902,0.2869935334,0.349679023,0.5377136469,-0.0409865156,0.0515500903,0.0667970031,0.0479120463,0.0896691009,0.1192515045,0.0857714713,-0.0702273324,0.2634492815,0.1284338385,0.2241659462,-0.0647975802,-0.1138611957,-0.107541278,0.4194297791,-0.4875138104,-0.084387593,0.1000552922,0.0748399198,0.1812013388,-0.2428474277,-0.0887820199,-0.1460988075,-0.2430461347,0.3116364777,0.0940705463,0.0835373625,-0.511957705,0.164441213,0.0075377976,-0.2018063366,-0.1137571484,0.1238797531,-0.1658339351,-0.0141071854,0.1862449348,-0.1799398512,0.1873875707,-0.2951588631,0.2437517196,-0.3901760876,-0.1954325885,0.0151099013,0.0578858256,0.4098414481,0.1254311204,0.1551068723,0.1864105314,-0.3417646587,0.0566025972,-0.1075194404,-0.1510609835,0.2333712131,-0.0550265461,0.0938554928,0.0341966748,-0.2493757308,-0.5408244133,-0.2501789927,-0.0095754005,0.3858899772,0.0983009189,-0.0028510524,-0.1126530915,0.0033437847,0.267029494,0.3937678933,-0.2108280957,-0.2492020875,0.2102492452,-0.2092680335,-0.3346993625,0.0819493756,-0.1042304859,0.0445711389,0.2523835003,-0.4788489342,-0.0906342939,-0.3600659072,0.4096082449,-0.232615903,0.0938282236,0.3463047147,0.0892930999,0.1463989317,-0.0417488404,-0.2757158875,-0.0651353374,-0.2946413159,-0.068232581,0.3381980062,0.4274490774,-0.0575702377,0.3770318925,0.3002279699,0.3604496717,0.3344713151,0.04584153,0.3711133003,0.0542609356,-0.2594737113,0.1087757051,-0.1943459958,-0.140160799,0.3007430136,0.0889823362,-0.1952640116,-0.052531682,-0.3804523051,-0.2807892859,-0.2852858603,-0.0992422476,-0.2463323325,0.2537561357,0.1191406399,0.1100426763,-0.4462445676,-0.3691142201,0.0288443286,0.5546328425,-0.3098391891,0.1115756556,-0.3508035243,0.1366987079,-0.6549720168,0.2522459924,0.0340248011,0.4958057404,-0.1741242707,0.1266969591,-0.0015080259,-0.1074288562,0.5570681095,-0.6522978544,0.4618299305,0.0581227317,0.1157312021,0.0101971151,-0.2243451625,-0.0922563747,0.2416802943,0.1369419247,0.2145550996,0.1639221013,0.0440419056,0.0762091056,0.1665931493,-0.2001322806,-0.279057771,-0.2271348834,-0.3418877125,-0.3903853595,0.128006041,0.0421996191,0.1854700744,-0.1517433375,0.1171616763,-0.4046885669,0.2369778156,0.1301799268,0.1602822691,0.187791869,0.0645751134,0.2181593776,-0.0480453596,0.0753483474,-0.1291563809,0.5240867138,-0.2804769576,-0.328379035,-0.0524905957,-0.199985683,-0.1171171516,0.1122419834,-0.3265575469,-0.0500914566,-0.1384819001,-0.1327372044,-0.3143305779,0.117074959,0.0426035039,-0.1354341507,-0.3776838183,-0.3201939762,0.3641437292,0.2117004246,-0.4467277825,0.3404740095,0.1198474169,-0.4280630946,0.1824693233,0.0517599583,0.7466567159,0.0798895434,0.2780212462,0.0972766355,-0.1122953147,0.2001173049,0.1161781773,0.0565570518,-0.3728869557,-0.4541019201,-0.0988719016,-0.2205425799,-0.134537518,-0.0249406099,-0.2337937057,0.2817453742,-0.1129783094,0.3188050985,0.274310261,0.048037339,-0.3741956651,-0.3764992654,-0.1209942549,0.1251865476,-0.1938239634,0.1358688474,-0.2274900377,0.0911153406,0.0849043131,-0.1673949957,-0.439121902,0.1967898756,-0.4114460945,0.1381170154,-0.2974178493,-0.3836832643,0.0877998993,0.5960645676,0.2849669755,0.3604593873,-0.1543026865,0.0600666888,0.0793892592,0.1179443896,0.024958102,0.0074552288,0.1962639689,-0.0806936249,-0.2981881797,-0.0151690254,-0.2781275809,-0.3215303123,-0.1055536345,0.2862682045,0.4039052129,-0.3201119602,-0.2076887786,-0.3106458187,0.0849739909,-0.2798955739,0.0937029347,0.0470425822,-0.0756008551,0.3795401752,-0.3368881643,-0.425712347,-0.1475177705,0.5242761374,-0.079077132,-0.1337566078,0.5711378455,-0.0485029556,-0.1785390973,-0.0130321439,0.1454919577,-0.1009839401,-0.462458998,-0.1003440917,-0.1225365251,0.1078282446,-0.029106807,0.1698505282,0.2787975669,-0.0963102505,0.0534037836,-0.7581394315,-0.2621001899,-0.1129847094,-0.2311439961,0.1716352552,0.1831110716,-0.4405415952,-0.1353047937,-0.2840806246,-0.203002885,0.0441007577,-0.2147331834,-0.1294415742,0.2779014409,0.154557094,-0.001340596,0.1245135888,-0.0026244703,0.0926584527,0.1227548569,-0.1713698655,-0.1237100959,0.1475240737,0.1037818864,-0.0917996764,0.0267537702,-0.3284253776,-0.0735971555,-0.0984259099,0.3940449953,0.2593082786,-0.0281576384,0.0679169372,-0.1276832074,0.2894711792,-0.2827050388,0.4151190817,-0.3472835422,-0.0251300745,-0.0133727901,0.3347693682,-0.1413164884,0.0541265719,-0.11745359,-0.0824945197,0.1532541662,0.0678972974,0.0646063909,-0.3769121468,0.085026145,0.1787576079,0.0593337528,0.0623646304,-0.23318322,0.0148412948,0.2450501472,0.1589923799,-0.0922699794,-0.1294877678,-0.0120077264,-0.3234773874,-0.1790443361,0.3004153967,0.1639576256,-0.3836256266,0.1359916031,-0.0324052311,0.5581638813,-0.2246947289,-0.0577207766,0.4813490808,-0.2174006104,0.2738089263,0.218585968,0.2596439123,-0.0843797997,0.5514788032,0.1820428967,0.1367382705,-0.0058679264,-0.1720480621,-0.1030526981,-0.6628072858,0.28002882,0.0652534962,-0.0736904144,0.0509437881,-0.0296167657,-0.0086359354,-0.3041151464,-0.0021537442,-0.4382816255,-0.2168946117,-0.1808834672,-0.283659488,0.0684353933,-0.2338435054,0.1725392044,0.0927761719,-0.1707766503,-0.14519687,-0.069977276,-0.0427515879,-0.3035758734,-0.268468231,0.1116327941,0.4133677781,-0.2662928104,-0.1926916391,0.0890493989,0.0064425515,0.2395810932,0.0886423066,0.0961362123,0.7200041413,0.3696607947,-0.186706394,-0.0827123448,0.2043794394,0.08498016,-0.1013351902,0.3095628321,0.1992664337,0.1772322208,0.3453566134,0.1196848005,-0.124466449,0.0374585725,0.0380601026,-0.008947121,-0.0834797546,0.1873595268,-0.1764346659,-0.168423444,0.1405667216,0.1980234534,-0.0727778077,0.1700269282,0.2323683053,0.0627374277,0.1945085675,-0.1900534481,0.0200616512,0.1487799883,0.3155461252,0.3503623307,0.0980699509,-0.2335534841,-0.068358846,-0.5775797367,0.2545953691,-0.0182250105,-0.3800154626,-0.0322349034,-0.1324881017,-0.3368008435,0.1061356142,0.1619229615,-0.1090964749,0.0555578545,0.2668960989,-0.1567609012,-0.2936111093,0.3894377947,-0.0377689339,-0.0781006962,-0.6786634922,0.0366012938,-0.0887601152,0.0320076123,-0.1506992131,-0.1666385084,0.0341500044,-0.0260742661,0.5437771082,0.2644129097,0.4555599689,0.0078377938,0.0918127969,-0.562205255,-0.0687982887,-0.1452899277,-0.0336945578,-0.077826038,0.271348238,-0.2275042981,0.0414364263,-0.1241922826,0.1670939773,0.2662992477,0.3795417845,-0.0054561738,0.2279049158,0.1809432358,0.1571899056,0.3478443921,0.0699722916,-0.0961795077,0.4460362792,-0.2409338355,-0.342048198,0.2109487653,-0.3941510618,-0.3340475857,-0.0192074794,0.3251572549,-0.2549189031,-0.1531336755,-0.2379637361,0.2061088979,0.2478395253,-0.1434804201,-0.3868976235,0.1413453221,-0.1570352912,0.140291512,-0.0227520186,0.0342236757,0.0101351459,-0.225390628,0.1708674133,-0.2374369949]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2934","title":"to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows","comments":"I did some investigation and, as it seems, the bug stems from [this line](https:\/\/github.com\/huggingface\/datasets\/blob\/8004d7c3e1d74b29c3e5b0d1660331cd26758363\/src\/datasets\/arrow_dataset.py#L325). The lifecycle of the dataset from the linked line is bound to one of the returned `tf.data.Dataset`. So my (hacky) solution involves wrapping the linked dataset with `weakref.proxy` and adding a custom `__del__` to `tf.python.data.ops.dataset_ops.TensorSliceDataset` (this is the type of a dataset that is returned by `tf.data.Dataset.from_tensor_slices`; this works for TF 2.x, but I'm not sure `tf.python.data.ops.dataset_ops` is a valid path for TF 1.x) that deletes the linked dataset, which is assigned to the dataset object as a property. Will open a draft PR soon!","body":"To reproduce:\r\n```python\r\nimport datasets as ds\r\nimport weakref\r\nimport gc\r\n\r\nd = ds.load_dataset(\"mnist\", split=\"train\")\r\nref = weakref.ref(d._data.table)\r\ntfd = d.to_tf_dataset(\"image\", batch_size=1, shuffle=False, label_cols=\"label\")\r\ndel tfd, d\r\ngc.collect()\r\nassert ref() is None, \"Error: there is at least one reference left\"\r\n```\r\n\r\nThis causes issues because the table holds a reference to an open arrow file that should be closed. So on windows it's not possible to delete or move the arrow file afterwards.\r\n\r\nMoreover the CI test of the `to_tf_dataset` method isn't able to clean up the temporary arrow files because of this.\r\n\r\ncc @Rocketknight1 ","comment_length":99,"text":"to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows\nTo reproduce:\r\n```python\r\nimport datasets as ds\r\nimport weakref\r\nimport gc\r\n\r\nd = ds.load_dataset(\"mnist\", split=\"train\")\r\nref = weakref.ref(d._data.table)\r\ntfd = d.to_tf_dataset(\"image\", batch_size=1, shuffle=False, label_cols=\"label\")\r\ndel tfd, d\r\ngc.collect()\r\nassert ref() is None, \"Error: there is at least one reference left\"\r\n```\r\n\r\nThis causes issues because the table holds a reference to an open arrow file that should be closed. So on windows it's not possible to delete or move the arrow file afterwards.\r\n\r\nMoreover the CI test of the `to_tf_dataset` method isn't able to clean up the temporary arrow files because of this.\r\n\r\ncc @Rocketknight1 \nI did some investigation and, as it seems, the bug stems from [this line](https:\/\/github.com\/huggingface\/datasets\/blob\/8004d7c3e1d74b29c3e5b0d1660331cd26758363\/src\/datasets\/arrow_dataset.py#L325). The lifecycle of the dataset from the linked line is bound to one of the returned `tf.data.Dataset`. So my (hacky) solution involves wrapping the linked dataset with `weakref.proxy` and adding a custom `__del__` to `tf.python.data.ops.dataset_ops.TensorSliceDataset` (this is the type of a dataset that is returned by `tf.data.Dataset.from_tensor_slices`; this works for TF 2.x, but I'm not sure `tf.python.data.ops.dataset_ops` is a valid path for TF 1.x) that deletes the linked dataset, which is assigned to the dataset object as a property. Will open a draft PR soon!","embeddings":[0.0456905663,0.334508419,0.1148890257,0.080451034,0.2304305583,0.118887417,0.3988847136,0.2548755109,-0.1099170819,0.2674386799,-0.3353850842,0.4040720165,-0.1644811481,-0.1074629501,-0.0342661999,-0.1066431478,0.0511111543,0.1451645941,-0.1666638404,-0.1226122081,-0.2103365809,0.0351907536,-0.1274755448,0.0721159056,-0.1387630254,-0.2811448276,0.0105889933,0.1006439403,0.2441080213,-0.0183550231,0.0682790056,0.0211760364,-0.177796185,0.4797623456,-0.0001166652,0.4142073989,-0.0061280271,0.1301537901,-0.235580802,0.0565112084,-0.0690313205,0.1760492623,0.1245584637,-0.0599729307,0.1349684894,-0.2177767009,-0.0255724229,0.0832738504,0.2824953496,0.4290578365,0.1738771498,0.6678111553,0.049049627,-0.0070738695,0.3827846944,0.0638971701,-0.2742590308,0.1161701158,-0.2370098084,-0.1935149431,0.0655605793,0.5781883597,0.0173822027,-0.0747985765,-0.1346866488,0.0793781951,-0.2474848479,-0.2132097334,0.099155657,0.3864619732,0.2942140996,-0.3177287579,-0.0704426542,-0.1286842972,-0.0647842139,-0.1454346925,0.1913480461,-0.046207156,0.017386917,0.1348499954,0.0582152456,-0.1622668952,-0.2615066469,-0.0061726281,-0.2810887992,-0.1072200239,0.0857438222,0.0801384374,0.414106369,-0.0271513741,-0.068637602,0.0416288003,0.0201638471,0.1307915747,-0.2023018003,-0.1820205152,-0.039258711,-0.1770301908,-0.0281791855,-0.0708624199,0.2392520756,-0.1570445448,-0.2443445474,0.0677398071,0.1644194871,-0.0705889463,-0.6655332446,0.1451997906,0.2680508792,-0.1933338344,-0.1119194329,-0.0286712628,0.0027096812,-0.4981434345,0.2861165106,-0.2107922286,0.4638061523,0.0174133722,-0.5834477544,0.1703355759,-0.1943601817,0.4077934325,-0.0392662771,0.3421227634,-0.0097245919,0.2793687582,0.3236465752,-0.0036117642,-0.4970700741,-0.0162012,-0.0546676517,-0.0396631137,-0.2995604873,-0.3319226205,0.1675927192,0.0439531542,-0.0478496067,0.1728075743,0.0610827021,0.1025774032,0.2539456487,-0.2808155119,0.5246316791,0.4812721908,-0.0750983432,0.2450949699,0.094958134,-0.2507022023,-0.1292222887,0.3534173071,-0.271070987,-0.1585650742,-0.2603646219,0.1245645955,-0.1202328429,-0.1552468389,-0.2238021493,0.0450044461,0.0343124531,-0.0595466755,0.0862919539,-0.1370920688,-0.5523484945,-0.2646558583,0.0109041464,0.3650912344,-0.1991525888,0.1795514375,-0.0254384503,-0.3931910396,-0.1594114155,0.2656008899,0.0646049082,0.315382421,-0.4097916186,-0.1304864734,0.4838194549,-0.2320678681,-0.4512842894,0.2350057364,-0.2089946121,0.2142525762,-0.3518173099,0.1303172708,0.2125170231,-0.2066658884,0.0123629132,-0.0070450343,-0.0623103678,-0.009127249,-0.3236092329,-0.0108151864,-0.0680190474,-0.1952706128,-0.0157284178,-0.0124156829,0.1318649352,-0.0493954197,0.3902634382,0.2196914554,0.1933261156,0.1417482793,0.3756607473,-0.1729133576,-0.0790883675,0.0432887785,-0.3775175512,0.1395243853,0.4056796134,0.0777031407,-0.1249045953,-0.1479362994,-0.0149693638,0.1985371262,-0.1919496357,0.0200802218,0.0784440786,-0.1194245666,-0.3427836597,0.2097568363,-0.0087737013,-0.1780350357,-0.2768910527,0.0820086896,0.1028026193,0.4687242806,0.2687642872,-0.0055732708,-0.2448134571,0.1943388134,-0.1494650543,-0.0300622191,-0.2719019651,0.4269395471,0.1747722477,0.3178130984,-0.1541973948,0.5814441442,0.1066848487,-0.5981359482,-0.2948667407,0.345680505,0.1319477111,-0.028623648,-0.0179552659,0.1456983984,-0.2007223219,0.171325475,0.0537690409,0.2392641008,0.0408273004,0.0214642044,-0.0363970362,-0.3410162628,-0.048381418,0.3614164293,-0.3146298528,0.2986148596,-0.6084029675,0.1965304911,0.3866648376,0.0311069563,-0.1689249873,0.2358304709,-0.1694948971,0.0617295206,0.0715380758,0.3181353509,0.476054281,0.0223621521,0.1727905273,0.1367983967,-0.193739742,-0.0902521908,0.2373642176,0.0887134001,0.3124321103,0.2039208263,0.25704512,0.193103686,-0.3024033606,0.0330183432,0.130203411,0.1728958189,-0.4525787234,0.1747216135,-0.1807385981,-0.0388464183,-0.307489574,0.0117461048,0.0377215482,-0.134800002,0.3154597282,0.3967652917,-0.4561233222,0.153922379,-0.3289419115,0.1290333718,0.0227688793,-0.2959315479,0.1803024411,-0.169179216,-0.2714836895,0.0344429053,0.3643865883,-0.24427782,0.4180335999,-0.0013580687,-0.0747494251,-0.1528537571,-0.2631629407,0.0897795483,-0.0949533954,0.1932279766,0.2937165499,0.3368107378,-0.1232672036,-0.1301765293,0.1462340504,-0.4851278663,-0.1786133051,0.1372669786,-0.1227175817,0.209010601,0.0203978065,-0.1520884037,0.0741765723,-0.3432299793,0.0670822263,-0.191379711,0.127728045,0.6203324199,0.1520412415,-0.0800016448,0.3789710402,0.10767968,-0.1736004055,-0.1127346307,0.200914368,-0.2037230134,-0.4296337366,0.2909044921,0.0338889062,0.1325789392,0.295558393,-0.6434158087,-0.1689423323,-0.0437356085,0.1776636243,-0.1088872477,-0.0470069051,0.1665738225,0.1164368093,-0.1197147071,-0.3432469964,0.0827863663,0.3821257651,-0.227080822,-0.0485802852,0.0422705784,0.0882449597,0.313570559,0.6243577003,0.0508327,-0.3429434299,0.0208036136,-0.2982521057,0.5073607564,-0.159090668,-0.2795748413,0.0158007275,0.0420313291,-0.13273637,0.2462733835,-0.0795774087,-0.2103789449,0.0578622855,0.2167329341,-0.0375917666,-0.5290263295,0.1369023621,-0.3965076804,0.1983652562,0.1509983689,-0.0425542817,-0.1581777185,-0.4451107979,0.2289596349,0.2040211558,0.0930258259,-0.1366162896,-0.1987896413,-0.2823874056,-0.3425536454,0.0680471882,0.1306476295,0.5035336614,0.0921356902,-0.2987198234,0.1360505223,0.2798964679,0.7769178748,-0.0301158391,-0.0992753878,0.2773328424,0.3224091232,-0.4041952789,0.0555643141,-0.1050164551,-0.0046288599,0.028162742,0.2416950762,0.0772358328,-0.3887937963,0.2129498273,0.1766911447,-0.2477184087,0.2110395133,-0.3918560147,-0.398602128,0.1013680622,0.0878460184,0.260494113,0.0497563928,-0.1087201983,-0.1115121841,-0.3012365699,-0.2482331544,0.0881446972,-0.0156271141,0.3283712864,0.3708717823,0.3975336552,0.0533410273,0.5702848434,0.3938702047,0.4075015187,-0.5431131124,0.021427609,0.0755058676,0.1911333948,0.125191927,0.4420394599,-0.1866856217,0.0531414859,-0.4151880443,0.0002785593,-0.3923731148,0.5302588344,0.1395115256,-0.2031341493,-0.0013096452,-0.0922305658,0.1333688647,-0.0103618503,-0.1489023864,0.6897454262,0.2165830582,-0.1877485067,0.2651528418,0.1582136005,1.1080405712,-0.0935677141,0.2334691584,0.5235840082,-0.2643699646,0.2126546949,-0.0553808175,0.1353131831,-0.0632255748,-0.5360183716,0.0201393031,-0.0125073828,-0.3608244359,-0.2005607784,0.1893274188,0.0023706069,0.1572063267,0.410684675,0.2291459441,0.0776234716,0.2959440947,-0.2186191976,0.0396565311,0.0224721488,0.4466592371,-0.1266208887,-0.2026067823,-0.1223222986,0.1659467816,0.0271231998,-0.0344647914,-0.0096408138,-0.7225638628,0.2756160796,0.0422483645,-0.0904388428,-0.0240593981,0.4012312889,0.3712239861,-0.0988768116,-0.1523476243,0.2883424163,0.3773627877,0.2930121124,0.0938072205,-0.0167857539,0.4750951231,0.0383024365,-0.0803633481,0.1676332504,-0.0544717051,-0.0525227487,-0.0370946191,-0.1225825995,0.0382938497,-0.1205560043,0.044584427,-0.0995850563,-0.2459054291,-0.0579577163,0.1035773382,-0.180487588,-0.1070264578,0.2344724834,-0.1927336901,-0.4316695929,0.2392567247,0.3922551274,-0.1205129474,0.1710344702,0.3036819696,-0.0761046782,-0.0731279776,-0.2562387288,0.3445347846,0.0378310904,-0.5670926571,0.2645981014,-0.1810758114,0.3202282488,-0.1079885513,0.5147185326,-0.1970301867,0.0209408961,0.0221601725,-0.1397224218,-0.3977710903,0.2361740619,0.0018207144,0.4124109447,-0.2537195086,0.4747990668,-0.1287534237,-0.2527932227,-0.2997979224,0.2343676686,-0.0281137675,0.0787903517,-0.1058863029,0.0576506145,0.3181125224,-0.2779511809,0.057696119,0.1073206663,-0.1381114721,-0.1384907216,-0.1637144834,0.1632773131,0.0280179083,-0.3234009743,-0.0129801221,-0.1738720238,0.076291658,-0.1113803163,0.3281935751,0.0700529516,-0.3296965957,0.0270981658,-0.213646844,0.1309412271,0.1392472386,0.0077438001,-0.0305922497,0.3758689463,0.0896954685,0.0981485397,-0.2366094887,-0.2325259447,-0.0303854793,0.2842668295,-0.0356378406,-0.3600642979,0.0377127007,-0.0558477119,-0.2938879132,-0.0599053092,0.3766296208,0.1799759567,-0.4634622633,-0.1362240314,0.2592977285,0.2242168635,-0.240421176,0.16245085,-0.121767506,0.0198641177,0.0385113843,0.166296646,0.1840274334,-0.1465545297,-0.1428779364,0.177231282,-0.2008037418,-0.2036248147,0.0310160294,0.5156292915,-0.0325355567,0.2373232841,-0.1476540118,0.1036643386,0.1264172643,0.2481572181,-0.0781786367,0.4826430678,0.1981102079,-0.0654385388,-0.0354681835,-0.3134524226,-0.0555218123,0.1641895622,-0.123816818,0.2421867996,-0.0361499302,0.2217094153,0.1549313664,-0.4245827198,-0.0117485141,0.2324923277,-0.0859795734,-0.3671880066,-0.3900308013,-0.208055228,-0.2723349631,0.0311247166,-0.0055279625,-0.0865737349,0.0094038406,0.1060144007,-0.0662822947,-0.2825777233,-0.1198218912,0.3330772221,0.2778065503,-0.2325664908,0.0656126365,-0.0248095151,-0.065677695,-0.0021082226,0.2082015872,0.0835288912,0.0284219794,-0.1759397835,0.0114991637,-0.0346538313,0.0922640339,-0.040092878,0.2144379616,-0.0489174612,-0.1405197233,0.2157975733,-0.0491833091,-0.070024997,0.1828447282,0.0519715101,0.0976671353,-0.4182255566,0.1498892158,-0.0033793463,-0.101680629,-0.2287625968,-0.0051180227,0.050867077,-0.108679451,0.4808011055,-0.4602059424,0.3644775152,-0.0873917863,0.0682187453,-0.0095315883,0.2219795287,0.1271091551,0.3292393088,-0.2891252935,-0.3842819929,-0.5309660435,0.2625992,-0.0495068468,-0.0506452397,0.2118176818,0.0317716971,-0.0888985768,-0.0664601475,0.0274049416,-0.1865465939,0.113322854,0.267254889,-0.2658275366,-0.0360959657,-0.4830030203,0.0984326154,0.0199032538,-0.5478616953,-0.1800811589,-0.1762706935,0.0375101008,-0.0818457901,-0.0215519853,-0.0166112091,0.5460247397,0.5247312188,0.1254571527,0.0754962265,0.1195667684,-0.102561526,-0.2674234807,-0.179499656,-0.3766336739,-0.1837470978,0.0070535499,-0.0237968545,-0.0529440977,0.197137922,0.0137047442,-0.2977714539,0.1801951528,0.3653975129,-0.4400824606,0.1956054866,-0.2059031129,0.3725327849,0.2628458738,0.4234926105,-0.1296609044,0.2166781127,-0.3998849392,-0.3820687532,0.4928815663,-0.5328620076,-0.093006812,-0.024597371,0.3842858076,-0.0938029662,0.0362871028,-0.5819953084,-0.176555559,0.3898159564,-0.083492741,-0.5030674338,0.1706117392,0.130146578,0.1384345293,0.1181671172,0.5913186073,0.1401758492,-0.1933630854,0.0364979543,-0.2862138152]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2934","title":"to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows","comments":"Thanks a lot for investigating !","body":"To reproduce:\r\n```python\r\nimport datasets as ds\r\nimport weakref\r\nimport gc\r\n\r\nd = ds.load_dataset(\"mnist\", split=\"train\")\r\nref = weakref.ref(d._data.table)\r\ntfd = d.to_tf_dataset(\"image\", batch_size=1, shuffle=False, label_cols=\"label\")\r\ndel tfd, d\r\ngc.collect()\r\nassert ref() is None, \"Error: there is at least one reference left\"\r\n```\r\n\r\nThis causes issues because the table holds a reference to an open arrow file that should be closed. So on windows it's not possible to delete or move the arrow file afterwards.\r\n\r\nMoreover the CI test of the `to_tf_dataset` method isn't able to clean up the temporary arrow files because of this.\r\n\r\ncc @Rocketknight1 ","comment_length":6,"text":"to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows\nTo reproduce:\r\n```python\r\nimport datasets as ds\r\nimport weakref\r\nimport gc\r\n\r\nd = ds.load_dataset(\"mnist\", split=\"train\")\r\nref = weakref.ref(d._data.table)\r\ntfd = d.to_tf_dataset(\"image\", batch_size=1, shuffle=False, label_cols=\"label\")\r\ndel tfd, d\r\ngc.collect()\r\nassert ref() is None, \"Error: there is at least one reference left\"\r\n```\r\n\r\nThis causes issues because the table holds a reference to an open arrow file that should be closed. So on windows it's not possible to delete or move the arrow file afterwards.\r\n\r\nMoreover the CI test of the `to_tf_dataset` method isn't able to clean up the temporary arrow files because of this.\r\n\r\ncc @Rocketknight1 \nThanks a lot for investigating !","embeddings":[-0.017654635,0.1541891694,0.0769107938,0.0744181946,0.1725985259,0.1645845324,0.3400904834,0.1861168444,-0.1919683069,0.3822576702,-0.2824491262,0.2848426998,-0.1505767554,-0.1510132104,-0.1580371708,-0.0992337465,0.1257585436,0.167851299,-0.1901147068,-0.1841688305,-0.2898720801,-0.0215148497,-0.327370286,0.0795662105,-0.2852760255,-0.3267048597,-0.0916857198,0.1191536784,0.1170001402,-0.07841409,-0.1573167741,-0.0711032003,-0.1773468852,0.4747584462,-0.0001205474,0.3584452868,0.0146335,0.1292270571,-0.1646926254,0.1001513228,-0.102570191,0.1868531406,0.1098253652,-0.1309705675,0.1541632265,-0.2203837484,0.047294464,0.1913519353,0.2669393718,0.3942332268,0.1664138734,0.5976775289,-0.0148456199,-0.0244104229,0.4810651839,0.0666921884,-0.3110128641,0.228202045,-0.183252871,-0.2691649497,0.1474390477,0.4888600409,0.0831765756,-0.0415764973,-0.1837423146,0.0297772828,-0.3578903675,-0.2360781282,0.1937207282,0.2784611583,0.3239397705,-0.2224481106,0.0123386877,0.0420895405,-0.12151099,-0.0689414591,0.1815636605,0.0662881956,0.1170332655,0.1451949328,-0.1794295907,-0.0714711398,-0.2178188562,-0.0269813742,-0.2697104514,-0.1216941476,0.0063582025,0.1643116474,0.3995058239,-0.0270410925,-0.2186173201,-0.0294112433,-0.0166256763,0.0714359283,-0.1651377082,-0.1348735392,-0.1345437765,0.0025094899,-0.1078164205,0.0273076948,0.1821828932,-0.2230697572,-0.1323369145,0.1000298783,0.1016594917,-0.1187836677,-0.5958840847,0.3131462336,0.295281142,-0.2285301536,-0.2712132633,-0.1387895048,-0.1088327542,-0.6633168459,0.4933824241,-0.2781928778,0.498590976,-0.1741571277,-0.5532178283,0.1169041544,-0.1342375726,0.3653871417,-0.115196079,0.4193829894,0.1590043753,0.0912155584,0.2282477468,0.1190212443,-0.4121915698,0.0310258958,0.0330229588,-0.1256711781,-0.360565722,-0.3782528937,0.1396561265,0.0635450929,-0.0230039731,0.2391600609,0.0726201385,0.1766205281,0.2837675512,-0.431430608,0.6493878365,0.4464761317,-0.0134130185,0.2591511011,0.1623892933,-0.1892025024,-0.1497623026,0.4186564386,-0.2162784487,-0.0137974694,-0.157305494,0.115259409,0.0606884286,-0.1436565369,-0.203731969,-0.1010478437,-0.0786174759,-0.1635623276,0.2012656033,-0.1976070255,-0.5147459507,-0.2963284552,-0.1082999781,0.3792219162,-0.1645184755,0.3078974187,0.1431201845,-0.3972130418,0.0059953704,0.363016516,0.1134514436,0.2382273823,-0.4455956221,-0.1024718434,0.4214687347,-0.2397156507,-0.4494386911,0.2381397039,-0.1294758916,0.0941175446,-0.2015377432,0.1184352636,0.143118009,-0.2037379295,0.0316181332,0.0183395352,-0.0693732202,-0.076740399,-0.3000811338,0.1583014727,0.0204451066,-0.0509014651,-0.154321,0.05273645,0.3393533826,-0.2361169606,0.3411521316,0.1847975552,0.0689592734,0.1607869565,0.321493119,-0.3075342774,-0.0852688402,0.0619392544,-0.1281856,0.1172516495,0.364707619,0.0094336923,-0.0719543397,-0.113814719,-0.113396585,0.1810445189,-0.2156182081,0.037566293,0.0035135944,-0.0378339328,-0.2137085199,0.2067743689,-0.0004930327,-0.1253623217,-0.3070148826,0.026798455,0.296400249,0.4681534469,0.1054550856,-0.0451783091,-0.1720358729,0.1316904277,-0.1075824574,-0.066053614,-0.2887434065,0.469099164,0.2476580292,0.1296931356,0.0165344253,0.4050133228,0.0425857641,-0.691002965,-0.1249036565,0.4314128757,0.1971428692,-0.1063851342,0.0224665478,0.0856082737,-0.2597537935,0.2148093581,0.100669235,0.2228789479,0.0058220923,0.0665586069,0.0792132393,-0.3537418544,0.096439071,0.4207526743,-0.3075872362,0.2093724161,-0.6096150279,0.0245354343,0.4899153709,-0.0103561152,-0.0894310772,0.2794166505,-0.1514907628,0.1749186516,0.0958402604,0.1407725215,0.5122002363,0.035527043,0.0858851746,0.1443796307,-0.2817312479,-0.066528663,0.2520436347,0.0302203242,0.2120308727,0.1943588257,0.1666267216,0.2204724401,-0.3269425929,-0.0779463053,0.100687407,0.0708008334,-0.3873489797,0.2037564516,-0.141062662,-0.1401050091,-0.2412522137,0.0849737749,-0.0051670396,-0.1253742427,0.3121213317,0.2641050816,-0.3669916689,0.1250952035,-0.4197683334,0.1320843995,0.0480908081,-0.2092866898,0.1106635705,-0.1319660395,-0.3402947783,0.0305394866,0.4360516667,-0.1923869103,0.4651909173,-0.2223875225,0.1703501493,-0.1473429501,-0.1536527276,0.0298600588,-0.0900716111,0.2436603159,0.2737259269,0.3140886426,-0.2141556144,-0.0708004832,0.2795857489,-0.5427033305,-0.2861479223,0.223249197,-0.0327365473,-0.1293388009,0.083870694,-0.0071252678,0.0840421841,-0.3345767558,0.0517720655,-0.102933906,0.0879867971,0.5746050477,0.191638723,-0.0430085361,0.2481711358,0.0756601617,-0.1675255299,-0.0089375759,0.2714706361,-0.207121104,-0.5179976225,0.3182075024,0.1386426985,-0.0428776555,0.3085761368,-0.6279010177,-0.1342575699,-0.2336850762,0.100689888,-0.1068340912,0.1550725847,0.0871676654,0.0960529,-0.1449069381,-0.3435560763,0.0048925113,0.2848476768,-0.2362652719,0.0006516685,-0.0349547826,0.0629395843,0.1345201433,0.4822039604,0.0318293944,-0.1654725671,0.200495705,-0.3054215014,0.6046196222,-0.2599270642,-0.2400037944,0.0599017553,0.1142046899,-0.1535027176,0.234093219,-0.0606404692,-0.1787326336,-0.0133119877,0.1014661565,-0.1196279526,-0.3492877185,0.0903233215,-0.3583359718,0.1568625569,0.1205483004,-0.0469608344,-0.0249231085,-0.3814910352,0.2835078239,0.1921522617,0.0423277169,-0.2172423154,-0.3888531327,-0.454675436,-0.10856013,0.1654713601,0.1470241696,0.5874216557,0.0064070574,-0.2722571194,0.1946412474,0.2674406469,0.7342179418,-0.1391730607,-0.0503010638,0.3903177679,0.3925023973,-0.3472799659,-0.0092262616,-0.2599988878,0.0213550497,-0.1138023883,0.2361126691,0.1701416969,-0.280518651,0.2421051562,0.2854965031,-0.2466304898,0.1876034737,-0.5327395797,-0.343834877,0.0030114246,0.1400165558,0.2260756344,-0.0294362605,-0.1500397772,-0.0529747792,-0.3793028295,-0.2613240778,0.0598471276,0.0349562019,0.4408480227,0.1943806857,0.3049543202,-0.0261976104,0.6141986251,0.3659821153,0.4922839701,-0.5450040698,0.1024127901,0.2489347011,0.0703405738,0.2837764621,0.4125652015,-0.1229415163,0.0962915942,-0.4376490116,0.1911679357,-0.3466957211,0.4170636237,0.20371674,-0.1481145918,-0.0529025272,0.0637534186,0.0556594357,-0.0155519368,-0.1059591547,0.6313262582,0.269338429,-0.0313745104,0.2944715321,0.1011382863,1.2075374126,-0.1255977303,0.3258848488,0.3455070555,-0.4239888489,-0.0855606198,0.0722757354,0.1262742877,-0.0671244711,-0.5808190703,0.0595162548,-0.0741003975,-0.2176912725,-0.1429721564,0.1091328785,0.0651676208,0.1057809964,0.5431858897,0.235861212,-0.0209707171,0.1053520069,-0.0252951458,0.016069822,-0.0120059047,0.5678728819,-0.1512120366,-0.3326204121,-0.087358214,0.0550680682,-0.0100521231,0.0195954721,-0.0133461254,-0.6196104288,0.2293964177,0.0329568237,-0.0710296631,0.0058566988,0.4161221981,0.3843311965,0.0319344103,-0.1466644853,0.3212939501,0.4691763222,0.2076121867,0.1192942485,0.0424969308,0.4516296089,-0.0465639569,-0.1062377393,0.2462542951,-0.1292869449,-0.0807303339,-0.1673709899,-0.0603705309,0.038864404,-0.1431066394,-0.048150599,-0.2568143904,-0.270617038,-0.059519697,0.0889462605,-0.2223586291,0.0245296545,0.3424888253,-0.1079980955,-0.5061223507,0.2096646726,0.276037395,-0.2299588323,0.1969762743,0.3436801732,-0.1816192269,-0.1202588603,-0.256157279,0.3886810839,0.0609171651,-0.5470731258,0.2829530239,-0.1245821118,0.429043591,-0.0138290934,0.5695757866,-0.0682384297,-0.0612271167,-0.0057996651,-0.2351913005,-0.3249398172,0.2648646533,0.030601548,0.3474400342,-0.1750466079,0.3146694005,-0.0190683883,-0.3286039531,-0.2675201893,0.2184214294,-0.195070371,0.1273489296,-0.1537469476,0.0576180927,0.2064946145,-0.2634270787,0.085400559,0.1924275458,-0.0160661768,-0.1238691434,-0.0520596728,0.1746418625,0.0810231417,-0.3536424935,-0.0041074217,-0.2538175881,0.0943080038,-0.047871232,0.1810651571,0.0045773825,-0.3380604386,-0.1056242958,-0.1124090403,0.0229120702,0.0872407481,0.058114782,-0.150458619,0.3196382821,0.0846866742,-0.0637011901,-0.2290772349,-0.2141297311,0.019038396,0.1848994642,0.0154668232,-0.3869718611,0.0789402574,-0.070252046,-0.127346918,-0.1185506508,0.4420285225,0.250708729,-0.4983851016,-0.1735827476,0.4078104496,0.2135398388,-0.3365501165,0.1308618188,-0.2609257698,0.003469344,-0.088597402,0.1060303822,0.1725834906,-0.2121965587,-0.1787487715,0.1510977447,-0.4579036236,-0.0566668771,-0.2281567156,0.6297215819,0.0878878087,0.2766057849,-0.0458956733,0.0726211742,0.0864037275,0.2339900881,-0.0945891514,0.5667414069,0.1749604195,-0.0305551048,-0.0447542705,-0.319888413,0.0433935151,0.2390377373,0.0055899844,0.2459430695,-0.0508207157,0.0674172193,0.1743658781,-0.4217628837,-0.0918770209,0.1651830971,-0.0224139057,-0.2720594108,-0.3387703598,-0.2180982232,-0.2728937566,0.12181665,-0.0307897348,0.0357925035,-0.0137483692,0.132101655,-0.0327218994,-0.2596278489,-0.0538723655,0.2903204858,0.344977051,-0.2803207636,0.036957033,-0.0194630641,0.0688491538,0.1412667036,0.3474190235,0.2181869447,0.0993962884,-0.2041796148,-0.0227078684,-0.0136043811,0.0684709847,0.0822915956,0.3425371349,0.0605486557,-0.1241256222,0.3070169687,-0.0458528548,-0.097131595,0.1166850254,-0.0437982343,-0.0444129519,-0.334315598,-0.0237693973,-0.0223184638,-0.2086401284,-0.1853766292,-0.0822281688,-0.1447233111,-0.0550839342,0.3361219466,-0.5006502271,0.2306670398,-0.1364663988,0.046309419,0.0424813442,0.2000268549,0.2953695655,0.5002297163,-0.3583451509,-0.5849112868,-0.5454111695,0.2477872968,-0.0550521836,-0.1414011866,0.1582048833,0.0288414918,-0.0622714907,-0.0681782961,0.274038583,-0.0469748341,0.0142406924,0.4672357142,-0.2215175331,0.0888676867,-0.3798103631,0.1021500453,-0.1448572278,-0.6390478611,-0.0195399132,-0.2257535011,0.0618604831,-0.0238049068,0.0974493399,-0.0718916506,0.6155150533,0.5243155956,0.0908330306,0.1478902549,0.1301814914,0.0306081325,-0.1639400721,-0.1677481234,-0.1326243728,-0.1898752004,0.0500835441,-0.0861456469,0.0079657948,0.2537888587,0.1178609505,-0.2608484626,0.1832638979,0.4560765028,-0.2930698693,0.0799366981,-0.2245865464,0.2090032101,0.3347372711,0.3146527112,-0.0712094083,0.1140887961,-0.4180307984,-0.3265278637,0.3847385943,-0.441431284,-0.1340732574,0.0491738394,0.3551390767,-0.1623359621,-0.0116730267,-0.5308157206,-0.1808762699,0.3966633379,-0.0071508149,-0.3978721499,0.2556317747,0.0600792915,0.1819682717,0.1403826475,0.7158242464,-0.03503526,-0.2116898149,-0.0162410177,-0.1845760942]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2932","title":"Conda build fails","comments":"Why 1.9 ?\r\n\r\nhttps:\/\/anaconda.org\/HuggingFace\/datasets currently says 1.11","body":"## Describe the bug\r\nCurrent `datasets` version in conda is 1.9 instead of 1.12.\r\n\r\nThe build of the conda package fails.\r\n","comment_length":7,"text":"Conda build fails\n## Describe the bug\r\nCurrent `datasets` version in conda is 1.9 instead of 1.12.\r\n\r\nThe build of the conda package fails.\r\n\nWhy 1.9 ?\r\n\r\nhttps:\/\/anaconda.org\/HuggingFace\/datasets currently says 1.11","embeddings":[-0.4641540647,0.0640663132,-0.1675162613,0.0911422074,0.0242198519,-0.1167648509,0.0046573328,0.4584467709,-0.1488352567,0.0218314379,-0.0215561539,0.1672185659,0.1742765456,0.4530319273,-0.0687930807,-0.0687004402,0.2214742899,0.1521472484,-0.3632488847,-0.0315286219,-0.2346586734,0.3556938469,-0.1418616623,0.042954348,-0.1667399108,-0.1111968309,-0.0725726113,-0.0722472295,-0.3525906503,-0.309075892,0.5409511924,-0.1080390662,0.0791708231,0.5754619241,-0.0001009286,-0.0032789898,0.5302111506,0.0530045889,-0.2712946236,-0.1803977042,-0.3502198458,-0.0796987191,-0.0399142914,0.1341502666,-0.025258461,-0.3136871755,-0.0039637419,-0.0812839568,0.2737195194,0.1549416184,0.3308033943,0.0851677209,0.4155243039,-0.4425582886,-0.4505780041,0.2404605895,-0.3100655675,0.1933912933,-0.0950218216,-0.0654576346,0.3125593662,0.2748890221,-0.0108567653,-0.2175896168,0.0571583509,-0.1458919346,-0.1065396816,-0.1608880013,0.2085220516,0.082794033,0.5673777461,-0.3365124464,-0.4459105134,0.116349265,0.0547229312,-0.1276331395,0.2863733172,-0.0994853154,-0.1852693409,0.2289404869,-0.3072545826,-0.1659559906,-0.0998794213,0.0542916507,0.0351006612,0.0843344927,-0.2238170356,0.0395075008,0.0850799978,-0.0795847178,0.0889546126,-0.017170459,-0.0044747209,-0.0164357647,-0.2207264602,-0.1931111813,-0.1908893287,0.0040363134,0.290076822,0.0416318104,-0.2814685106,-0.2031846493,0.0022204889,0.0698052868,-0.057144139,0.0626951158,0.4606170654,0.077450268,0.145462364,0.2514044344,0.1704059988,0.0070430655,0.0021299806,-0.3041796386,-0.2461816221,-0.0392220579,0.2495234162,-0.2043797076,-0.1059830338,0.1044391021,0.0609567054,-0.0279124603,0.0041456302,0.0737251118,-0.098237142,-0.1265262663,0.0580965318,0.1755417287,0.0175039303,0.0253259484,-0.2480798662,-0.19085069,-0.0307969321,0.012003894,0.2311127931,-0.3263165355,0.4701302052,-0.1246728003,0.1668923795,0.0062446366,-0.1614912301,0.078064017,-0.3607374132,0.4940434396,-0.2841105163,0.1163817868,-0.0540333688,-0.0690908656,-0.0722390711,0.009847221,-0.1026801541,-0.0875664055,-0.4648688734,0.2391197681,0.1965070069,0.1023772582,-0.1715472639,-0.0221420992,0.0536693893,-0.0455940999,-0.0935841873,0.0454386994,-0.0532059669,-0.1457587034,0.2406151742,0.2125408351,-0.1583735943,0.1107756644,0.1150244549,-0.2624725103,-0.2327995002,0.0774188638,-0.0671860948,-0.1484007537,-0.0090418523,-0.2777799666,0.1654792577,-0.4407516718,-0.5214886069,0.0708692968,0.1568902731,-0.1486879587,-0.0275132805,-0.1143989637,0.1878380924,-0.1143638939,0.1809624135,0.0786408931,0.005723747,-0.2770836055,-0.1697548479,-0.275437206,-0.241848886,0.0517572574,-0.0461218953,0.1780284196,0.042331446,-0.2044620216,0.1104750931,0.143998459,-0.2036080956,0.2029421628,0.4378283918,-0.1347674131,0.0343065746,-0.1481755525,-0.5067104101,0.0793106779,0.1355768591,-0.0216997135,-0.0685060993,-0.1599982828,-0.3334034681,0.2762608528,0.0679381713,-0.217081219,0.2146058977,-0.1150966808,0.1284432113,0.0793034956,-0.1678137481,0.7080689073,-0.137064442,0.2539079785,-0.1866996437,0.1552656889,-0.2871536613,-0.1260931939,0.2259406298,0.1750397533,0.2101011425,-0.1859097183,-0.1064059809,0.3100512624,-0.1132724211,-0.1732333302,0.1977748871,-0.2093673199,0.0850393698,-0.0069135549,0.1005597562,-0.0600860864,-0.1986969262,0.2078127116,0.0507737435,-0.0467366427,0.1806784272,-0.0241424423,-0.0142447082,0.2336301059,0.4198890924,0.0989498571,0.0189749841,-0.1595772505,0.1046580002,0.1180455238,0.2821315825,0.0236277115,-0.2964583933,0.1159171835,0.3984741867,-0.1097792536,-0.0726787671,0.2535006702,0.0919811726,0.1495138854,0.3302159607,0.2453039587,0.1037769392,0.0531402938,-0.095606029,0.2007383853,-0.3480142057,-0.0298374817,0.1060575619,0.0322836116,0.2812074125,-0.0845356435,-0.1015645266,0.0695140362,-0.0760767683,-0.0694011524,-0.1414503753,0.0929866508,0.0086003188,0.0421742573,-0.1739253849,-0.0793259069,-0.255325973,-0.2271770239,-0.2650867403,-0.020353796,-0.0007564036,0.1378406584,0.210903734,0.4877940714,0.0403436422,0.0395924002,-0.0504935272,0.0288728029,-0.1351604462,0.1316828281,-0.1524738818,0.1581934094,0.0931431428,-0.0417627729,0.1490940601,-0.4683484435,0.1915226579,-0.0488868766,-0.5337531567,0.162495628,-0.2750543058,0.4730536938,0.2408772856,0.0991982818,0.0325651355,-0.2097351402,0.1801730543,-0.1365850121,-0.1675789505,-0.219999522,-0.0309822541,-0.3386186361,-0.0958381668,-0.1521292478,-0.135665834,-0.1753454208,0.3863707185,-0.0221327562,0.0562425889,0.3409312069,0.4604199231,0.0470611565,-0.1918748617,0.0891415849,-0.1435628682,-0.4498470724,0.056229163,-0.2765717506,-0.2497362643,-0.0989951342,0.2478436381,0.4750163257,0.0358756483,-0.5661168694,-0.1986954808,-0.0168789178,0.1650635004,0.0965101942,-0.013050952,0.2417072803,0.0429633595,-0.1247905418,-0.2604943216,-0.3049151897,0.1147871763,-0.2061828226,0.3588501811,-0.0101826554,0.4184118509,-0.1961857677,0.2067244649,0.570433557,-0.193848446,0.1289104074,0.1958947629,0.646984756,-0.2325790077,-0.2919476926,0.1788204014,0.2425242811,0.2087279856,-0.0302555487,0.2250890583,-0.0930705443,-0.2837232053,-0.2128251344,-0.1717529297,-0.1405042112,-0.2474866956,0.1902405322,0.1072169468,0.1035116091,0.250531435,-0.1334473193,0.0827420875,0.1420256048,0.2684525549,-0.1062390804,-0.0453671105,-0.2940100133,-0.0090564089,-0.5889351964,0.3405347168,-0.115719378,0.5059463382,0.0934728608,0.2877984345,-0.1646546274,0.1292834878,0.7751023173,-0.1143795177,-0.1694623381,0.0409080386,0.1196236387,-0.386511147,0.1661832929,0.269202739,0.4197353423,0.2771253586,0.570371449,-0.2825520635,-0.2901853323,0.4174397886,0.3029113114,0.064699769,-0.1520482302,-0.2908979654,-0.3592730463,-0.0160685554,-0.270186305,0.0235399101,0.0294448528,-0.0739184543,-0.3009604812,-0.2542800009,-0.1923899204,-0.0933020413,0.1007940471,0.1852276027,-0.1531287879,0.4446290135,0.1214143708,0.143140167,0.3937529624,0.3233439922,-0.3388970494,-0.4432545602,-0.1255219132,-0.0592223518,0.2138877064,0.0718777403,-0.2636139393,-0.4265669286,-0.0393774845,0.0544238426,0.0062244446,-0.1491227001,0.4722716808,0.4071490765,-0.4898706079,-0.3163699806,0.3107614219,0.1555151641,0.0137672732,0.6786463261,0.1637614667,0.0148167852,0.0191572513,0.0330405645,0.6511470079,-0.0443928279,-0.0499101281,0.0446797237,-0.2227658033,0.3669653833,0.0488393344,-0.0652379617,-0.1855852902,-0.3706175685,0.0375149027,0.0188984033,0.3522101343,-0.0635798946,-0.1897932589,0.2009338439,0.0050997678,-0.0510391779,0.1347304136,0.0466747731,0.0570177846,-0.0017096179,-0.2023081779,0.2671092749,0.1218866855,0.0911121443,-0.0820011273,0.0522740521,0.0302037653,-0.2407122105,-0.3367681503,-0.0089284312,-0.052127514,0.0197862741,0.5075984001,-0.3320956826,-0.006780732,0.1831906736,0.532916069,0.268796891,-0.2907969654,0.0985455588,-0.1932935715,-0.1898310483,-0.0396836028,0.08748281,0.1513931453,-0.1723753661,-0.2542519271,-0.081403546,-0.0534898341,-0.0884619206,0.1103957668,-0.2348884344,-0.2580360174,-0.2222904563,0.0855715796,-0.0798651874,0.1912613213,-0.3580775261,0.2370942533,-0.0410242192,0.1347132623,0.1535924971,0.0453106724,-0.1981552094,-0.0836486146,0.2549034357,0.0502661504,-0.06083728,0.1065000668,-0.0302821025,-0.283223778,-0.199485153,-0.2953700125,-0.1355858594,-0.2921869159,-0.0426430628,-0.1656433195,-0.1487552971,0.030801909,0.2645925581,0.1511131227,-0.1485049576,0.1459294707,-0.217853874,-0.4190172255,0.3893738687,-0.2418382168,0.081914641,0.1848844141,0.3065513074,0.2461230457,-0.1169423535,-0.3740441203,-0.0096954424,-0.132256791,-0.0716296583,0.1841001213,0.0253096037,0.1819106936,-0.1267617196,0.1651816815,0.166282177,-0.3246473372,-0.2387083918,-0.1031778455,0.0037646173,-0.0249224957,-0.0605644509,0.1770996302,-0.1955480129,-0.3712851703,0.0236387346,-0.0425952226,0.1294345856,0.0975574777,-0.047995396,-0.0325053371,0.1657387912,0.1664978564,0.1940087974,0.2027938217,0.2586369216,-0.0825744346,-0.0052247988,-0.1060191244,-0.0838461593,-0.1263361126,0.0679822192,-0.3276896775,0.2343242615,0.3120971024,-0.153106302,0.0876108259,0.3479898274,0.4070643485,0.143855691,-0.0349122398,0.0139444303,0.4124389291,0.2923622727,-0.398332715,-0.037430156,0.2405215502,0.2383926958,-0.129919067,0.1489732713,0.0590852872,-0.1693943292,0.4098250866,0.1403006017,0.2205741554,0.0602487288,0.3973987997,0.3222332597,0.2497254461,0.0663221255,0.4399477243,0.2902185321,0.2228898108,0.4001783133,-0.2888497114,0.0846696123,-0.0285459496,0.1135114208,0.1290114373,-0.104021512,0.2614093721,-0.0871152207,-0.0681195483,-0.1083463579,-0.1178631559,0.2789145708,-0.36383003,-0.0831985399,-0.3642092049,0.0756527483,-0.1014542431,-0.0847678632,-0.0112250736,-0.0103246085,0.3061100245,-0.1312603951,-0.1614249945,-0.1096174344,0.5349951982,-0.1963897645,-0.0826527104,-0.4234129786,-0.0658413172,-0.0159965903,-0.0684605166,0.1299069077,0.5619087219,0.0393128321,-0.0317944586,-0.0642656311,0.5191246271,0.3900793493,0.045922596,0.1561653614,-0.1683365554,-0.0400273055,-0.0259133242,-0.136891067,0.3559701443,0.1754811555,0.0503647886,0.1609129608,0.269453615,-0.1784174144,0.0201286301,0.1036785692,0.2914521694,-0.2207628787,0.122631833,-0.3412408829,-0.1359685957,-0.1008708104,0.0563278571,-0.2205191106,0.1479242891,0.2504945099,-0.1503033936,-0.0945490748,-0.1575134844,0.1470509917,-0.1119528338,0.4386317134,0.2454130054,0.2079600096,0.0663516819,-0.411449343,-0.267690748,0.2406549454,-0.0822616667,0.1564208269,-0.0161260664,-0.1036028862,0.0204328652,0.1059478,0.3543419242,-0.0168153346,-0.0535218529,-0.2916656733,-0.325381577,-0.1110316217,-0.1225745976,-0.0263410211,0.162237078,-0.1866420358,0.18117024,-0.1827753931,0.0607612543,-0.1764393598,0.1855309606,-0.0066312873,-0.2995755672,0.3030668199,-0.0313812457,0.0356941372,-0.011500272,-0.4884319603,-0.2003731579,-0.3177161813,-0.059405148,0.3417550623,0.0319521949,0.131634593,-0.3392523825,-0.0623484664,-0.1945575625,0.3203814328,0.2225255221,0.0371220186,-0.1918862164,0.1454050988,-0.08730115,-0.2772326469,0.2167028487,0.3893757761,-0.1664120555,0.2395646125,-0.16181238,-0.3641056418,0.4493977427,-0.1083183736,-0.1574521214,-0.0138091575,-0.0823283941,0.0318232365,-0.2974265218,-0.5770468116,0.3174734712,0.1961872876,0.1672439873,-0.1092537194,0.1885131449,-0.1230864972,0.0585776381,0.0498388484,0.7527417541,-0.0445063934,-0.200931415,0.3679227531,0.028568225]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2932","title":"Conda build fails","comments":"Alright I added 1.12.0 and 1.12.1 and fixed the conda build #2952 ","body":"## Describe the bug\r\nCurrent `datasets` version in conda is 1.9 instead of 1.12.\r\n\r\nThe build of the conda package fails.\r\n","comment_length":12,"text":"Conda build fails\n## Describe the bug\r\nCurrent `datasets` version in conda is 1.9 instead of 1.12.\r\n\r\nThe build of the conda package fails.\r\n\nAlright I added 1.12.0 and 1.12.1 and fixed the conda build #2952 ","embeddings":[-0.5863975883,0.2035934627,-0.232974872,0.1968558878,0.2075668573,-0.0917660967,0.0932786614,0.5642169118,-0.3096557856,0.0915806666,0.0279443692,0.2152974308,0.0675806627,0.4218905568,-0.1395219564,0.0038826414,0.0308199879,0.2241077572,-0.4182020426,-0.0123287281,-0.1885185838,0.2447249144,-0.1071585938,0.0497225188,-0.245254755,-0.1354384422,0.0322817974,0.0517660826,-0.3269813061,-0.3223083019,0.2895521522,-0.0699542463,0.029348569,0.4992179871,-0.0000913996,0.0741490647,0.3882876933,-0.0257503297,-0.2825210094,-0.0777058378,-0.3032828271,-0.1225462258,-0.0787501559,-0.0245858394,-0.1226112768,-0.2728967667,-0.1174333841,-0.1095196307,0.1959647387,0.2553437352,0.4446544051,-0.0068132705,0.2757855952,-0.4212940931,-0.1860865951,-0.0436917767,-0.2034923285,0.0971226022,-0.0531155542,-0.0714639053,0.0847257301,0.2414616197,-0.0279956367,-0.0224732459,-0.0340468511,-0.164331913,0.1127335355,-0.135215342,0.2132426351,-0.0154786697,0.6562217474,-0.3307790458,-0.2999447286,0.1405354589,0.0567250699,-0.1853858232,0.2922054827,0.0008688812,-0.0837874711,0.2183299512,-0.2021050006,-0.1321849227,-0.1009486616,-0.0226571206,-0.0519170277,0.1495386511,-0.1100725085,0.0149995303,0.0504197516,-0.2023219615,0.362195164,-0.0173723605,-0.0971301347,-0.15897879,-0.2414802462,-0.1807620078,-0.0219156146,-0.1191595793,0.1366502941,-0.0261811931,-0.1909336001,-0.0144785335,0.0841854289,0.1068868414,0.0498245768,-0.0046015838,0.2775596678,0.2492183447,0.2414739579,-0.0120759141,-0.0095909256,0.0504127778,-0.0319391079,-0.4023247957,-0.196750626,0.0584249012,0.0760341808,-0.2049568146,-0.2722491622,0.1437727064,0.0510823801,0.0147776939,0.0015100588,0.2345990539,-0.016674662,-0.1202628314,0.0807097182,0.1372875571,-0.0713210627,-0.0524630882,-0.3114102781,-0.1119067445,-0.121408619,-0.0771802515,0.3006842136,-0.1038021892,0.3707338572,-0.0830180496,0.0617507473,-0.032412041,0.0544019453,0.027759565,-0.2395564318,0.3888090551,-0.0499059856,-0.0003740962,0.0252180807,-0.069863148,-0.027930323,0.0828347281,-0.0366372727,-0.2251679599,-0.4237506986,0.3854449987,0.1938515007,0.0662305206,0.0716949105,-0.0636549145,0.1177643538,-0.1366869509,-0.2729662657,0.0184946395,0.0636605322,-0.0871889219,0.2279902995,0.2070731521,-0.2756123841,0.1650138944,0.1333748549,-0.3019257784,-0.0072152917,-0.0268017929,-0.1443531364,-0.0073989532,-0.058277104,-0.1680708975,0.3321492374,-0.295216471,-0.4482250512,0.1187651679,0.0744725466,-0.328696847,-0.0668687671,-0.0782488808,0.196417436,-0.0056965332,0.1062466353,0.0992733911,-0.0286835041,-0.0740173683,-0.2667303383,-0.2071607858,-0.1118903831,0.1681469083,0.057727553,0.1447627693,0.1092381552,-0.2082567811,0.0788387135,0.0829256773,-0.0572477691,0.180361703,0.3837525547,-0.0759066194,0.0168578904,-0.1095519215,-0.3079583943,0.1055386439,0.0256883148,0.0204669945,-0.064672716,-0.0481392927,-0.3070678413,0.114816241,-0.0371600352,-0.0541843511,0.38391608,-0.0289845057,-0.0508067682,0.0003028249,-0.0711915568,0.3884094656,-0.150223583,0.0347469635,0.113964498,0.1847821027,-0.3705210686,-0.2101033777,0.2375807166,0.0689672604,0.2542420626,-0.1483214498,-0.1656304598,0.3099377453,-0.0822563544,-0.1912937462,0.3436691761,-0.1129948199,0.0022904254,-0.0102867652,0.0606409721,-0.0951388702,-0.1239355877,0.0992038921,0.0730591044,0.0231340732,0.1158568263,-0.1084979028,0.0092544667,0.189959541,0.5791200399,0.0050684898,0.1632045656,-0.2162691951,0.0145309176,0.055825673,0.1821321249,0.037359044,-0.2978574932,0.0825711638,0.3980308175,0.0076435688,0.0726711303,0.0926921591,-0.0076113031,0.1124799475,0.2614999413,0.2583378851,0.2037272304,0.1349072009,0.0081493929,0.1047509387,-0.2374267727,-0.0164957829,0.2083378881,-0.0552917421,0.2838928998,0.1663821787,-0.0411133058,-0.0695776567,-0.2452527285,-0.1380629987,0.0534085073,0.2105703056,0.037322551,-0.1095859855,-0.2824200094,-0.0017022508,-0.1233533695,-0.1432923526,-0.0640132949,-0.1281411797,0.1245221123,0.0895596594,0.00977821,0.5170032978,0.0143528832,0.0652859882,0.0990787596,0.009667472,-0.0518003739,-0.0790770501,-0.127101317,0.2408400774,0.2044294775,0.0509005859,0.2018431723,-0.3337310851,0.1491089165,-0.2748525441,-0.366984427,0.1274038702,-0.094208695,0.3852852583,0.2254270911,0.0919218436,0.0747521743,-0.2698901594,0.110820815,-0.0580758601,-0.0898231268,0.0070468765,-0.0787847564,-0.3627088368,-0.2367329299,-0.3228822947,-0.1648404151,-0.3588816226,0.1358786523,-0.060873311,0.0982486829,0.0736362413,0.3923217952,0.0536427498,-0.0562113784,0.0783555135,-0.222058624,-0.2932739556,0.1643790603,-0.3600728214,-0.3115933836,-0.0019090399,0.1679629087,0.5059683919,0.0179702938,-0.5560306907,0.0464412272,0.0780357271,0.1711068302,-0.0487899296,-0.0328627974,0.2273453474,0.2445125133,-0.2243379951,-0.3177289069,-0.2355287969,0.1948930621,-0.153514728,0.213289544,-0.2728606462,0.3523001075,-0.1319694966,0.166004017,0.2222482264,-0.2386436909,0.219208017,0.1830454022,0.5038248897,-0.1305069923,-0.1817170829,0.2021608651,0.1581531167,0.119042255,0.1395495832,-0.0279069915,-0.3791646361,-0.2756888568,-0.1051902995,-0.2455618232,-0.1936589777,-0.2074110061,0.0622599088,0.1791433245,0.1262440085,0.1772680581,-0.1568305194,0.1184369698,0.0211998075,0.0384883247,-0.0041033318,-0.1069697961,-0.1779532284,0.0358848907,-0.2792441845,0.2821165323,-0.2120809704,0.3916369379,-0.0623135567,0.2048662305,-0.0723558962,0.0009504542,0.6512207389,-0.2112644315,-0.0737034976,0.1125226542,0.1181229427,-0.2302242666,0.0650186986,0.083207123,0.3839910328,0.3091752231,0.3659373522,-0.3039242029,-0.1437313408,0.2834871113,0.1504956931,-0.0098037198,-0.2056625634,-0.0874546692,-0.2531458437,-0.1671919972,-0.1062273532,-0.0554941334,0.1245535091,-0.1537984908,-0.4580590427,-0.2820667922,-0.1694457382,-0.01758972,0.1693876237,0.2554765642,-0.1612110734,0.5345661044,0.0045534945,0.1155918241,0.4735500813,0.2798543572,-0.189198032,-0.2257909626,0.0917203054,-0.2555652261,0.0836775675,0.0862136558,-0.2363713384,-0.341285795,-0.2195706218,0.1084240749,0.076264888,-0.026122503,0.3726834059,0.2485917658,-0.4293208718,-0.2518617511,0.242318511,0.0162319206,-0.0728726611,0.5016472936,-0.1048982367,-0.0741726682,0.0676231533,0.0824791938,0.66672194,0.0610501505,-0.1266774684,0.006856665,-0.1592148095,0.1626461744,-0.1145113334,0.0037972457,-0.2875882685,-0.4622149169,0.0564569943,-0.0223851558,0.2537106872,0.2087925076,-0.3033083379,0.0743170977,-0.0972160399,-0.2132728994,0.0625187531,0.0630918145,0.0538320169,-0.0261505786,-0.0835556015,0.361900717,0.0005899728,-0.0727578327,-0.1062728912,-0.067716673,0.0350575037,-0.1675877124,-0.2483343184,0.0498329103,0.0010737955,0.2386963516,0.1627294123,-0.2036982179,0.0009662647,0.1117749289,0.3256909251,0.338988781,-0.2267630547,0.2255694866,0.0059769298,-0.1343333125,-0.0487091094,0.1586330235,0.0849824697,-0.1907640249,-0.2566301227,0.0247879233,-0.0578859113,-0.0663252398,0.1320600957,-0.1353257447,-0.0472149029,-0.2453776896,0.0064868447,0.0142924003,0.0114533948,-0.3371916115,0.3463756442,-0.0527707152,-0.0052369772,0.3470881283,0.012192864,-0.1826528311,-0.1017469913,0.2007384747,0.1102424562,0.0318651125,0.1323222965,-0.0796627328,-0.3194353282,-0.3690176308,-0.1674910337,-0.2219544947,-0.2326253057,0.0004745419,-0.0364109874,-0.2213421166,0.1392650902,0.2272441387,0.103966862,0.1268080324,0.12728405,-0.1871058792,-0.2903102636,0.1435471773,-0.2305998057,0.224498108,-0.0108646248,0.3226473331,0.1365102828,-0.0325094052,-0.5056607127,-0.0034738155,-0.376393348,0.0241533406,0.0283821672,-0.0198156852,0.0323873013,-0.1437138319,0.2947608829,0.2062872499,-0.2226619273,-0.3661328256,-0.0312338322,-0.0175669845,0.0497309305,0.0167891141,0.1219146401,-0.1331655234,-0.4057909846,0.0281089693,-0.018162312,0.0284162387,0.0422412306,-0.0115903253,0.1159084886,0.0983861461,0.1967673749,0.1101816744,0.1371367872,0.2352458239,-0.086849317,0.008582918,-0.0457893237,-0.0563464426,-0.0349864177,-0.0246918034,-0.2511456311,0.2289187759,0.3598465621,-0.2082441896,0.0332717411,0.3909105957,0.334284842,0.207790032,-0.1314865947,-0.0822631642,0.3971404433,0.4947049916,-0.4555192888,-0.1380333006,0.0083031477,0.0364327542,0.0368086807,0.0361693576,0.0356303714,-0.2292745262,0.308706969,0.2625923753,0.2341880202,-0.2097451538,0.2820364535,0.3012544215,0.0510902666,-0.0560819656,0.3553821146,0.1626077741,0.0981256738,0.6015193462,-0.2439955473,0.1260926723,0.0874224454,0.0610313267,0.0384354778,-0.1500719041,0.2947190106,-0.0056637153,-0.0532938018,-0.1579737663,-0.2138314545,0.0789796114,-0.327932924,-0.0442125946,-0.3130210936,0.047839202,-0.1836959124,0.0594023429,-0.0732315704,-0.0382933915,0.2386280447,-0.0940453112,-0.0338380709,0.0506904498,0.4240409732,-0.0262736827,-0.0643444732,-0.3497653008,0.0606286004,0.028568374,-0.1260315925,0.0108015751,0.4940970242,0.143856734,-0.0359248295,-0.0589989088,0.3479895294,0.348203361,0.277752161,0.0534951985,-0.1082824618,-0.0092249922,-0.0953803509,-0.2411097437,0.3029560149,0.2273861617,0.0720100328,0.2493595928,0.3347663879,-0.2555821836,-0.0169894919,0.1393467486,0.2789635658,-0.0335982628,0.0914347172,0.0385823473,-0.1384752989,-0.0504327938,-0.0994241536,-0.2475196272,0.0530853234,0.3749343157,-0.1021480113,-0.0129441777,-0.20851098,0.2110289633,-0.1338755935,0.4848892093,-0.0640231967,0.2046496123,-0.0398341753,-0.3171172142,-0.3321355581,0.2472712398,-0.025997106,0.1577498317,-0.0925640464,-0.0027004264,0.0405297764,0.0034110083,0.2962839603,-0.0514624342,-0.0067159436,-0.1673950255,-0.4477352202,-0.031719882,-0.0999102145,-0.0099834604,0.1832084507,-0.2746391296,0.1216634437,-0.1069333702,0.2742959857,-0.1479437649,-0.01545012,0.1173128188,-0.004334955,0.3953119218,0.0229871981,0.3068906069,-0.2120409459,-0.2795613408,-0.139135778,-0.4778122306,-0.1043489054,0.1476300657,0.2280675769,0.1021462083,-0.2248517722,0.0063110883,-0.1715018302,0.265201062,0.1091065183,0.2118775547,-0.1095602289,0.157030344,0.054823719,-0.0997829214,0.0586517639,0.253433466,-0.067046538,0.2125879377,-0.2155797929,-0.3528474867,0.4893417656,0.0142854201,-0.245224759,0.0129644563,-0.0260217786,-0.0059275106,-0.1479931176,-0.4293839633,0.229219079,0.2826619446,0.1066514403,-0.1077293307,0.1915610731,-0.1371898651,0.1114795953,0.0018529911,0.573530376,0.0091546699,-0.2639088035,0.1011295021,-0.0598786213]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2930","title":"Mutable columns argument breaks set_format","comments":"Pushed a fix to my branch #2731 ","body":"## Describe the bug\r\nIf you pass a mutable list to the `columns` argument of `set_format` and then change the list afterwards, the returned columns also change.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"glue\", \"cola\")\r\n\r\ncolumn_list = [\"idx\", \"label\"]\r\ndataset.set_format(\"python\", columns=column_list)\r\ncolumn_list[1] = \"foo\" # Change the list after we call `set_format`\r\ndataset['train'][:4].keys()\r\n```\r\n\r\n## Expected results\r\n```python\r\ndict_keys(['idx', 'label'])\r\n```\r\n\r\n## Actual results\r\n```python\r\ndict_keys(['idx'])\r\n```","comment_length":7,"text":"Mutable columns argument breaks set_format\n## Describe the bug\r\nIf you pass a mutable list to the `columns` argument of `set_format` and then change the list afterwards, the returned columns also change.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"glue\", \"cola\")\r\n\r\ncolumn_list = [\"idx\", \"label\"]\r\ndataset.set_format(\"python\", columns=column_list)\r\ncolumn_list[1] = \"foo\" # Change the list after we call `set_format`\r\ndataset['train'][:4].keys()\r\n```\r\n\r\n## Expected results\r\n```python\r\ndict_keys(['idx', 'label'])\r\n```\r\n\r\n## Actual results\r\n```python\r\ndict_keys(['idx'])\r\n```\nPushed a fix to my branch #2731 ","embeddings":[-0.0634390414,-0.1930613369,-0.0847119913,0.0437259898,0.5914995074,0.0606024899,0.7192493677,0.2848295569,0.1028594598,0.1568417102,-0.116555728,0.7383283377,0.0102749048,0.2990810573,-0.2102174014,-0.2156233937,0.1305874735,0.1316983253,-0.131610781,0.0363861509,-0.2623939514,0.0679177791,-0.375447154,-0.1826451868,-0.1687054485,-0.144836247,-0.2592204511,-0.0473862961,-0.0947361514,-0.271669805,0.1168658212,0.1954511404,-0.2182203233,0.4592967629,-0.0001044443,0.0152202081,0.0200444236,-0.1021577045,-0.2109158784,-0.0972076952,-0.2637723386,-0.0789669305,0.1564198732,-0.3649504483,-0.2155194879,-0.0359100625,-0.4558363259,-0.2674218416,0.2459866256,0.3095501363,0.337433964,0.0961965173,-0.0191967264,-0.0778871626,0.350938499,0.024526272,-0.2062668651,0.1694167107,0.0360042304,-0.0574511774,0.2977561057,0.3134259582,-0.353706181,0.0584034547,0.0068529551,0.0332615748,-0.1524679512,-0.0831187963,0.2396808267,0.182026431,0.2787374258,-0.4583109617,-0.2539326847,-0.2258167565,0.0679665506,-0.4645034373,0.0320331641,0.0488984473,0.1365063339,0.1395174265,0.1109698266,0.0990392044,-0.0189186633,0.1066981032,-0.1106197387,0.4013152719,-0.164026618,0.0244787335,-0.0885713398,-0.0331682153,-0.0251186453,0.00836109,-0.2377885878,-0.0240884684,-0.2194893956,-0.1676763147,0.0029882274,-0.0322995409,0.0095377434,-0.1139194816,-0.046752058,-0.030285297,0.1399154961,0.0920782611,0.3197984397,0.278681159,-0.0635075048,0.4963833392,0.0732233971,-0.0988840014,-0.0371107832,-0.1538478732,0.308842361,-0.1950437427,0.3155840933,0.1642861962,0.238843292,0.0682739392,-0.2911200523,0.0995498076,-0.1539213657,-0.0634099916,-0.0316827819,0.3574388027,-0.0768988431,0.4067205787,-0.0575630032,0.3268230855,-0.0287995134,0.2033368051,-0.2828402519,0.130673483,-0.16348508,-0.2244134843,0.084832713,0.0561836064,0.0570182092,0.3165373206,0.0141386166,-0.2182422727,-0.1813179404,-0.2205047458,0.3185226619,0.1042528227,-0.0451851599,0.2784454525,0.2753438056,-0.391156137,-0.1288427263,0.0011100398,-0.2692237198,-0.0980003327,-0.2712520361,0.301181078,-0.0521826521,0.006486427,-0.0049253348,0.0343260728,0.3726559877,0.1105766967,0.2605312169,-0.263920784,-0.1242048368,-0.3503037989,0.0914837793,0.2431913614,-0.3241707087,0.1112797782,0.3844877183,0.2423685193,0.1121881604,0.2034005076,-0.0473191589,0.0279094465,-0.0148240486,-0.1700301617,0.2520844638,-0.3341756165,-0.2848739028,0.0553748496,0.0027873039,0.0685167909,0.2248959839,-0.1666803211,0.218622759,0.0578945205,0.4702925682,0.1355687827,0.1329183131,0.2321659923,-0.2272477895,-0.0266459659,0.1384196877,-0.0202689543,-0.2603830695,0.1265508235,-0.0973207727,0.0009012923,0.2695514858,-0.1020145565,-0.1591219455,0.1050361097,0.3889386952,0.1192682534,0.0584359989,-0.1191093028,-0.3901537359,-0.1211959422,0.5693612695,0.0484296009,-0.1748030633,-0.420571506,-0.4216828644,0.1798273921,-0.2152584344,-0.0824195817,0.3075227141,-0.003220771,0.0642843321,0.0407320149,-0.2096620351,0.1370342225,0.0039426237,-0.0304533746,-0.0208814349,0.0401547514,0.0119094858,-0.1834635288,-0.0599086955,0.4251461327,0.3132659197,-0.0375363454,-0.167879343,0.4503110647,0.3329874873,-0.1549506932,-0.3464752436,0.2249492705,0.0076201176,0.1943212301,-0.0460130088,0.0911375135,0.0759302452,0.0359970815,-0.3030140102,0.4775167704,-0.3167341053,0.0824443698,-0.234490633,0.1565556824,0.0981085524,0.0481777042,-0.2115463167,-0.5392458439,-0.062703222,0.0825194418,-0.0306131057,-0.0419437215,-0.470782876,0.1967414916,0.467872113,0.083343856,0.2176243067,-0.2235009074,-0.0992184877,-0.0115429945,0.0731081441,0.2889002562,0.3600578606,0.2259647995,-0.1365745217,0.0370337591,-0.0854712948,-0.2084178478,0.2183267325,0.0393683836,-0.0967671201,0.1753202826,0.3845235407,-0.0736434236,-0.1551766545,-0.0109753758,-0.0839595348,0.1736816615,-0.383297354,0.1738462895,-0.357047528,-0.0811517164,-0.1417519152,-0.2927251756,0.1747904569,-0.4272394776,-0.1525552273,0.1162470058,-0.0045933891,0.1375183463,-0.0211186819,-0.0319030136,0.2776056528,-0.1820731759,0.0417787805,-0.4216853678,-0.0939931944,0.1455974728,0.0303103533,-0.2044923753,0.143839553,-0.0300900023,-0.2565354407,0.0638367087,-0.1259914637,0.1559157372,-0.2809145153,0.1430458874,0.4829552472,0.3382579684,-0.0894901603,0.0548964813,0.3780276179,0.2132316977,0.2059809715,0.138030231,0.0068890937,0.0550270453,-0.2424753159,-0.3511870801,-0.0555490851,-0.1354611665,0.1351916194,-0.1916234791,0.0870930701,0.2025233358,0.2921889722,-0.1789580435,0.0213473085,0.2048195601,-0.3525743186,-0.0658937842,0.1719931066,-0.1936250478,-0.2129869014,-0.0455462337,-0.0391085483,0.0343543291,0.003133951,-0.1914339364,-0.5200380683,-0.1579808742,0.0683984458,-0.1270274371,0.2740614712,0.054213278,0.1716973037,-0.0757222623,-0.124540098,-0.1934963614,-0.0346360952,0.2626588643,-0.1748870909,-0.0929444581,0.2631197572,0.0199016724,0.1118289381,0.0432161056,-0.0100050457,0.2867324948,-0.2556657493,0.3692888916,-0.3020857573,-0.3446291685,0.0101491911,-0.0587351248,-0.0385484286,0.0579442456,0.0086890599,-0.0651061609,-0.03837841,0.2229370773,-0.0282931291,-0.2501943707,0.1912529469,-0.1141608059,0.0595322475,0.0302135851,0.3910180032,-0.2894249856,-0.0952822343,0.0404491015,0.0114112077,-0.2488003522,-0.2327436656,-0.3056177199,-0.4484499991,-0.2699826956,0.1899857521,0.1403895319,0.4892662764,0.1334533989,-0.2362642586,0.0847799107,0.0963971168,0.6822872162,-0.4301690459,-0.3574624956,0.4292676151,0.0859615058,-0.3903828263,-0.2652634382,-0.3802113831,0.0143659599,0.0227175374,0.3876759708,-0.2589642107,-0.244254753,0.2214453816,0.1897400767,-0.0913613886,-0.1592662036,-0.2549227178,-0.0928166062,-0.2348954529,0.093459256,0.0586991273,0.2659591138,-0.1441631168,0.080433853,-0.120664537,-0.1190443635,-0.073037453,0.2516750395,0.3982740939,0.0575780161,-0.0467027016,-0.1222939566,0.2817413211,-0.073820971,0.1757423133,-0.3656084239,0.0336791202,-0.043546617,-0.3389855623,0.1086297184,0.1856503338,-0.0393812321,0.2526772916,-0.4333719611,0.0946321487,-0.1003584713,-0.0472176298,0.4953793287,0.0461364612,-0.2542179227,-0.685565114,0.1015549749,0.0903888866,-0.0064569241,0.3456730247,0.0725662932,-0.1828046888,0.7794381976,0.0889574587,0.7490982413,-0.0867960528,0.2408003211,0.2939781249,-0.2088262588,0.3112678528,0.0492751785,0.1543674469,-0.2872996628,-0.1380244195,-0.0044919541,-0.058305636,0.0247331839,-0.0898002461,-0.4097694159,0.2892744839,-0.5145075917,0.2875412405,-0.1983894557,-0.0730571076,-0.0874249563,-0.2267119884,-0.4677990079,0.2220413983,0.0646434128,-0.2087830454,0.0112568969,0.1907317191,-0.065343827,-0.1698995829,-0.1873892993,-0.1199399978,-0.1559698284,-0.060999155,0.0678645521,-0.4865007401,0.0141786886,0.2893814445,0.3962036669,0.3095730245,0.0078290785,0.0623751841,0.3810586631,0.2517932355,0.0658424273,-0.0208446104,0.3145283163,0.2877014875,-0.0982795581,0.133476764,0.1272155195,-0.0942069888,-0.208242327,0.0529583395,0.1855789423,-0.5744428039,-0.1757421643,-0.1527145803,0.2630667686,-0.3218447566,0.2347415537,-0.0754718408,0.0486424267,0.5379700661,-0.1782665998,-0.2557305992,-0.1464184821,0.1999298036,0.1669089943,0.0693257004,0.4183680713,0.209888503,-0.2242714167,-0.3183072805,-0.0735745728,-0.0647625104,-0.0756294131,0.3513303101,-0.0830074921,0.0055152066,-0.0013726874,0.3915545642,-0.2344167382,0.3513912261,-0.0538246259,-0.205533132,-0.1155246943,0.0326848663,0.2926522493,0.1945950687,0.2316054702,0.2053347826,-0.3007096052,-0.0630117357,-0.4568855166,-0.1410906166,-0.0862020776,0.0651070401,0.2362469286,0.0590859801,0.0111563224,-0.0795943588,0.3580130339,0.1183542982,0.1108416393,-0.4295145273,-0.325049907,0.0110547999,-0.0234215762,0.0853728354,-0.1708765328,-0.1580568254,-0.1967953146,-0.3962603509,0.2666503191,0.2017664611,-0.0845334008,0.1594115496,-0.1476027668,0.2420492321,0.069533959,-0.0918390304,-0.1183197945,-0.2192181051,0.1840180457,0.1944650263,-0.0407677777,-0.0157293361,-0.1871426105,0.105258435,-0.0939975604,0.1689095646,0.1012573317,-0.1896534264,-0.0661151484,0.0987796709,0.4865126908,0.1566231549,-0.2018299252,-0.0108075626,0.2645915449,0.3579041362,-0.1296218932,-0.1321053356,0.2255125493,0.2314445674,0.3111750484,0.450210005,0.1397311538,0.0756184161,-0.2331319749,-0.0009882856,0.0487412177,-0.1216368228,0.1389746517,0.5687338114,0.1403339207,0.0290854629,0.1549241543,0.0297774058,0.1120429337,0.3498232961,0.3880805969,0.2931537628,0.4028396904,-0.1947244108,0.0580541939,-0.1981632113,0.3190847337,-0.1623570174,0.0311275516,0.1651246846,0.0851635262,0.2890863717,-0.2256009579,-0.1322600543,-0.0517679639,0.0788591281,-0.2149592489,-0.1735067964,-0.0896615535,-0.0712928176,-0.0761795044,-0.0599798895,-0.2564581037,0.1198756173,0.3142152429,-0.0564944707,0.1837552637,-0.3391599059,-0.0654969811,0.000147028,0.4076169133,-0.1922232509,0.0069478415,-0.0030230694,0.0994503275,0.0413250551,0.0327654257,0.5266547799,0.1835280061,-0.1379824132,-0.2399745435,-0.0410318933,-0.1680832505,0.1132266372,0.0703418106,-0.0068472442,0.1387183964,0.1747312844,0.3146359026,-0.2387895435,0.0065306732,0.2093479186,0.1968953162,-0.176381886,-0.0169236436,0.0693721697,-0.2573129833,-0.1397525519,-0.0770672634,-0.1836999059,-0.0613855198,0.1926912814,0.338663131,0.3150190115,-0.144917652,0.1909878105,0.3556967378,0.2456722111,-0.0497522913,0.0769164115,-0.1025438681,-0.0973583683,-0.2256001085,0.0825739801,-0.1283246875,-0.2780284286,0.0039949431,0.056390807,0.1623618454,0.0148027819,0.0437278971,-0.0130933793,0.0254860874,0.4582059085,-0.2675381601,-0.1802484095,-0.0817581043,-0.1243195757,0.0804289132,-0.2435839325,0.0038390581,0.1951596886,0.2180403769,-0.3459209502,-0.0041370536,-0.1667109132,0.0345008858,0.0528789237,0.0232742596,0.109929055,-0.0425669812,-0.0815995336,-0.2123945504,-0.0985484049,-0.1917688251,0.0674044117,-0.1040592343,0.2106078565,-0.007643329,-0.1253238022,-0.132859543,0.2341602147,0.1957191378,0.0204240587,-0.1200679019,0.257250309,0.0979948491,-0.1329333782,-0.0794446543,0.3134340048,-0.0020583791,0.3305855095,-0.0650895014,-0.4050293565,0.382350266,-0.3645904958,-0.5083729029,-0.0518184118,0.0601342954,-0.1210649461,-0.0529172905,-0.3612802029,0.0644469261,0.1545197666,-0.1112097576,-0.3463765383,0.1828248799,0.0455008373,0.1964670569,-0.1405544728,0.1704312861,-0.0222192314,-0.0930835903,-0.1773957312,-0.1250364035]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2927","title":"Datasets 1.12 dataset.filter TypeError: get_indices_from_mask_function() got an unexpected keyword argument","comments":"Thanks for reporting, I'm looking into it :)","body":"## Describe the bug\r\nUpgrading to 1.12 caused `dataset.filter` call to fail with \r\n\r\n> get_indices_from_mask_function() got an unexpected keyword argument valid_rel_labels\r\n\r\n\r\n## Steps to reproduce the bug\r\n```pythondef \r\n\r\nfilter_good_rows(\r\n ex: Dict,\r\n valid_rel_labels: Set[str],\r\n valid_ner_labels: Set[str],\r\n tokenizer: PreTrainedTokenizerFast,\r\n) -> bool:\r\n \"\"\"Get the good rows\"\"\"\r\n encoding = get_encoding_for_text(text=ex[\"text\"], tokenizer=tokenizer)\r\n ex[\"encoding\"] = encoding\r\n for relation in ex[\"relations\"]:\r\n if not is_valid_relation(relation, valid_rel_labels):\r\n return False\r\n for span in ex[\"spans\"]:\r\n if not is_valid_span(span, valid_ner_labels, encoding):\r\n return False\r\n return True\r\n \r\ndef get_dataset(): \r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path,\r\n name=\"prodigy-dataset\",\r\n data_files=sorted(file_paths),\r\n cache_dir=cache_dir,\r\n )[\"train\"]\r\n\r\n valid_ner_labels = set(vocab.ner_category)\r\n valid_relations = set(vocab.relation_types.keys())\r\n ds = ds.filter(\r\n filter_good_rows,\r\n fn_kwargs=dict(\r\n valid_rel_labels=valid_relations,\r\n valid_ner_labels=valid_ner_labels,\r\n tokenizer=vocab.tokenizer,\r\n ),\r\n keep_in_memory=True,\r\n num_proc=num_proc,\r\n )\r\n\r\n```\r\n\r\n`ds` is a `DatasetDict` produced by a jsonl dataset.\r\nThis runs fine on 1.11 but fails on 1.12\r\n\r\n**Stack Trace**\r\n\r\n\r\n\r\n## Expected results\r\n\r\nI expect 1.12 datasets filter to filter the dataset without raising as it does on 1.11\r\n\r\n## Actual results\r\n```\r\ntf_ner_rel_lib\/dataset.py:695: in load_prodigy_arrow_datasets_from_jsonl\r\n ds = ds.filter(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2169: in filter\r\n indices = self.map(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1686: in map\r\n return self._map_single(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2048: in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\ninputs = {'_input_hash': [2108817714, 1477695082, -1021597032, 2130671338, -1260483858, -1203431639, ...], '_task_hash': [18070...ons', 'relations', 'relations', ...], 'answer': ['accept', 'accept', 'accept', 'accept', 'accept', 'accept', ...], ...}\r\nindices = [0, 1, 2, 3, 4, 5, ...], check_same_num_examples = False, offset = 0\r\n\r\n def apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples=False, offset=0):\r\n \"\"\"Utility to apply the function on a selection of columns.\"\"\"\r\n nonlocal update_data\r\n fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n if offset == 0:\r\n effective_indices = indices\r\n else:\r\n effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n processed_inputs = (\r\n> function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n )\r\nE TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'valid_rel_labels'\r\n\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1939: TypeError\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Mac\r\n- Python version: 3.8.9\r\n- PyArrow version: pyarrow==5.0.0\r\n\r\n","comment_length":8,"text":"Datasets 1.12 dataset.filter TypeError: get_indices_from_mask_function() got an unexpected keyword argument\n## Describe the bug\r\nUpgrading to 1.12 caused `dataset.filter` call to fail with \r\n\r\n> get_indices_from_mask_function() got an unexpected keyword argument valid_rel_labels\r\n\r\n\r\n## Steps to reproduce the bug\r\n```pythondef \r\n\r\nfilter_good_rows(\r\n ex: Dict,\r\n valid_rel_labels: Set[str],\r\n valid_ner_labels: Set[str],\r\n tokenizer: PreTrainedTokenizerFast,\r\n) -> bool:\r\n \"\"\"Get the good rows\"\"\"\r\n encoding = get_encoding_for_text(text=ex[\"text\"], tokenizer=tokenizer)\r\n ex[\"encoding\"] = encoding\r\n for relation in ex[\"relations\"]:\r\n if not is_valid_relation(relation, valid_rel_labels):\r\n return False\r\n for span in ex[\"spans\"]:\r\n if not is_valid_span(span, valid_ner_labels, encoding):\r\n return False\r\n return True\r\n \r\ndef get_dataset(): \r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path,\r\n name=\"prodigy-dataset\",\r\n data_files=sorted(file_paths),\r\n cache_dir=cache_dir,\r\n )[\"train\"]\r\n\r\n valid_ner_labels = set(vocab.ner_category)\r\n valid_relations = set(vocab.relation_types.keys())\r\n ds = ds.filter(\r\n filter_good_rows,\r\n fn_kwargs=dict(\r\n valid_rel_labels=valid_relations,\r\n valid_ner_labels=valid_ner_labels,\r\n tokenizer=vocab.tokenizer,\r\n ),\r\n keep_in_memory=True,\r\n num_proc=num_proc,\r\n )\r\n\r\n```\r\n\r\n`ds` is a `DatasetDict` produced by a jsonl dataset.\r\nThis runs fine on 1.11 but fails on 1.12\r\n\r\n**Stack Trace**\r\n\r\n\r\n\r\n## Expected results\r\n\r\nI expect 1.12 datasets filter to filter the dataset without raising as it does on 1.11\r\n\r\n## Actual results\r\n```\r\ntf_ner_rel_lib\/dataset.py:695: in load_prodigy_arrow_datasets_from_jsonl\r\n ds = ds.filter(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2169: in filter\r\n indices = self.map(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1686: in map\r\n return self._map_single(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2048: in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\ninputs = {'_input_hash': [2108817714, 1477695082, -1021597032, 2130671338, -1260483858, -1203431639, ...], '_task_hash': [18070...ons', 'relations', 'relations', ...], 'answer': ['accept', 'accept', 'accept', 'accept', 'accept', 'accept', ...], ...}\r\nindices = [0, 1, 2, 3, 4, 5, ...], check_same_num_examples = False, offset = 0\r\n\r\n def apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples=False, offset=0):\r\n \"\"\"Utility to apply the function on a selection of columns.\"\"\"\r\n nonlocal update_data\r\n fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n if offset == 0:\r\n effective_indices = indices\r\n else:\r\n effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n processed_inputs = (\r\n> function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n )\r\nE TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'valid_rel_labels'\r\n\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1939: TypeError\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Mac\r\n- Python version: 3.8.9\r\n- PyArrow version: pyarrow==5.0.0\r\n\r\n\nThanks for reporting, I'm looking into it :)","embeddings":[-0.2829699218,0.2777142525,-0.0049711736,0.1934766024,0.0521399193,-0.0259464737,0.2624210715,0.4762106538,-0.1806556284,0.0836368799,-0.1311300099,0.3645940423,-0.1310276091,0.1083575189,-0.1563561261,-0.2259478569,0.0624334253,0.0486550108,-0.0484171957,0.0325096361,-0.3871050477,0.4132161736,-0.4211299717,0.1235269159,0.008685492,-0.1654407978,0.1102141142,-0.0451107584,0.0099631641,-0.4235654175,0.4564154148,0.0699703991,0.0381553695,0.2566260993,-0.0001171321,0.2396819293,0.366684258,-0.0842551515,-0.2250449806,-0.405336529,-0.1809279323,-0.0339020006,0.1180416197,-0.1935501844,-0.0818986967,0.0174917243,-0.249624908,-0.2356579155,0.2611079514,0.4325256944,0.1972838491,0.0717553645,-0.0651654452,0.0339874998,0.0209346432,0.2666781545,0.0185224004,-0.0007021982,0.0587287322,-0.1879914552,0.2919843793,0.5048494339,-0.2368306518,-0.2743239999,0.2988181114,-0.2294282466,0.223627314,-0.3586401939,0.2479036301,-0.0648386478,0.0247386377,-0.0833451524,-0.3223271072,-0.1697276235,-0.2576664686,-0.1120616794,0.3498052657,-0.1584596336,-0.244265154,-0.0208345316,-0.07258939,-0.0044453437,-0.0193625409,0.0009967602,-0.0110017173,0.5335709453,0.051298894,0.0513096936,0.0644063577,-0.1729889512,0.3066953421,-0.1076792255,-0.1007701308,0.0617668256,-0.170351848,0.0725231543,0.2387878448,-0.100716427,0.0953230187,-0.0929184034,-0.0909817293,-0.1606885344,0.0432230234,-0.0053427084,-0.0694495961,0.0904969722,0.3082866073,0.6318762302,0.0672923923,0.058587458,-0.0616019517,0.0373482928,-0.0557346605,0.0807316527,0.0906984136,0.2471883744,0.3795975745,-0.3135053217,-0.6206545234,0.1616326869,-0.5770969987,-0.1607390344,0.237689808,0.0602699555,0.2683047056,0.0857219398,-0.0559860393,0.0184919778,-0.0884351358,0.0967272371,-0.1329723448,0.0300575532,-0.0661212951,-0.288934499,0.1902687103,-0.5589625239,0.0788640752,-0.1231013834,-0.090868406,-0.0088757947,-0.0592832975,0.0524173155,0.1024440303,0.3941213489,-0.4122775495,0.1066873968,0.2004962116,-0.5756821632,-0.1944885254,0.2767592072,-0.4454355538,-0.0839394405,0.0214587208,0.2041888237,0.1366723627,-0.1789838076,-0.3426181376,0.5803837776,0.0123843485,-0.2307372093,-0.0704242811,-0.2239028215,-0.3683023453,-0.0239756647,0.1340631545,0.1689580381,-0.8365051746,-0.2499466091,0.2335013896,-0.0030520104,-0.0626956224,-0.2084405422,-0.1818583906,-0.0219790861,-0.1172454953,-0.0013353637,0.3576940596,-0.485848546,-0.6574547887,0.1390666068,0.1225496158,0.1945742369,0.1314316094,-0.1025109664,0.3647884429,-0.0999429971,0.1756555587,0.2742819488,0.0124631217,-0.0347300246,-0.2012303621,0.0749125555,0.2941287458,-0.0680078194,0.1666537374,0.27637887,-0.1620120853,-0.0194072127,0.2861098051,0.1138249561,-0.0791773424,0.0638790876,0.2018899024,-0.1175088063,0.232898742,-0.4075967968,-0.1830836833,-0.0563953593,0.0861253291,0.130115062,-0.3378736973,-0.214805305,-0.3373113871,0.0745304599,-0.2046329081,-0.1652506888,0.1793425977,0.1503546089,-0.0387527496,-0.0507269315,-0.0374752656,0.4382365942,0.2428901494,0.1432452947,-0.2216559798,0.2244707495,-0.0795594603,-0.080613561,0.0983143821,-0.0522183664,0.36705634,-0.0016741207,-0.3518117964,0.2039128393,0.0591471232,-0.3466408551,-0.0529597551,0.0107605178,0.0352687202,0.0651153848,0.1794721931,0.3757503927,-0.0078226523,0.1407691687,-0.1973981559,0.4055749774,-0.0408483408,0.4687744379,0.0199777223,0.0771241337,0.3994879127,0.2407884896,-0.1524210721,-0.1703110337,0.0525777675,0.3843008578,0.2716696858,-0.1308127642,-0.0575984567,0.2184957862,0.3087431192,-0.0334121324,0.0245152265,0.3014428616,0.015626844,0.0211441629,0.1812755615,0.4510476291,0.150378406,0.1575035751,-0.1162214428,0.0736183971,-0.2908762395,-0.0003076858,0.2586247027,0.1770678163,-0.2355194986,-0.0977184847,0.2307223231,0.0618963651,-0.3269087672,-0.4945808649,-0.1177324802,0.1571106911,-0.2101406008,0.4223466814,-0.2563110888,-0.360643059,0.2112966329,-0.1757659763,-0.1615487486,-0.4771515429,0.0838324726,0.1105318815,0.0211028233,0.4142694175,-0.1944871247,0.2770675421,0.1981321126,-0.2057120055,-0.2324998677,-0.3686217666,-0.3903269172,0.0790402368,0.1200828031,0.0579170287,0.1671684682,-0.0545275025,0.1354142576,0.083329387,-0.7349098921,0.1015762314,-0.3447894454,0.4745761752,0.1016463861,0.038837377,-0.3328078389,-0.2596282661,0.1520894766,0.304394573,-0.2402678132,0.1635410488,0.0431459434,0.2904618084,-0.2289169878,-0.4294990301,-0.2123535424,-0.1180267707,-0.1469037235,-0.1846597344,0.3282931149,0.5408180952,0.0719795302,-0.0907747671,0.1236420944,0.184421882,-0.3507416248,-0.4034056962,0.144136861,-0.0788861513,-0.1249593124,-0.1968769431,-0.2286561579,0.0848402604,0.2183059156,-0.5605329275,-0.4192049503,0.0349870473,0.0410459861,0.1570181549,-0.1139656156,0.1058698297,0.2774389684,0.030448081,-0.2724855542,-0.3693442345,0.0357308015,-0.2114411592,0.0473164991,0.0135325789,0.9601014853,0.0164168794,0.0682780072,0.1694822013,0.2156193256,0.3308912516,0.0938933268,0.4396736324,-0.2999125719,-0.2751029134,-0.1022948325,0.0323724411,-0.0618773438,-0.2904603481,0.0270541478,-0.0472293831,-0.1075963899,-0.1457635909,-0.2428792566,-0.307492286,-0.1465963423,-0.1706662029,0.0713224858,0.0963586718,0.3636464477,-0.308096081,-0.0618637502,0.0288702194,0.2998217344,-0.0606535003,0.0504504703,-0.4331361651,-0.0584681481,-0.3299600482,0.3512754738,0.0836806521,0.5775329471,0.0337941647,0.1314383298,-0.1877772957,-0.0064897859,0.6513140798,-0.37467134,0.0990580469,0.3981027007,-0.0517896451,-0.105050236,-0.0820436478,-0.3614759147,0.1523979008,-0.130425632,0.3490183353,-0.2240297794,-0.0515188314,0.3174758554,0.0232320037,-0.1624078602,0.1051052287,-0.3738526404,-0.5830719471,-0.0338545628,0.2991501689,0.0823923871,0.4687190056,0.0837701112,0.0468495339,-0.0865380093,-0.4693993032,0.1605705619,0.2456996441,0.1313659102,0.1721284539,0.1322498471,0.1905105859,0.1852574944,0.2601747811,0.3494502306,-0.1209676489,-0.429931134,-0.1555662006,-0.003165609,0.2964791059,0.0886024907,-0.2174897939,-0.0622401759,-0.1156746596,-0.1378526092,-0.0112768495,0.1836106479,0.2737488151,0.2099698633,-0.4123848975,-0.5313845873,0.1796308607,0.2255197465,-0.1604507267,0.3734816909,0.2652886808,-0.4143254757,0.4304826856,-0.1537723541,0.7024047375,0.2602522671,-0.1502133161,0.3903979361,-0.1639682204,0.3863707483,0.1703226268,0.1472315341,-0.0537542589,-0.3721880913,-0.0091732787,0.0341162309,-0.1155766919,0.0897843763,-0.2464871407,0.3867236376,-0.2271939069,-0.1297411919,0.1523538828,0.0656017885,0.072543852,0.1573487818,-0.436416477,0.1157073081,-0.1356859058,-0.1174689382,-0.0174163952,-0.157321766,0.2357757241,-0.4989627004,-0.2169850469,0.0593276396,-0.2186802477,0.1991744936,0.1808251143,-0.5248704553,-0.0648283362,-0.2381993383,0.1592858881,0.1625673175,0.0071186177,0.1369998753,-0.1568153501,0.4103268683,0.1870173216,-0.0307340231,0.4770958126,0.1361288428,-0.0680285394,0.1569158882,0.016583018,-0.282654196,-0.3184965551,-0.1825746,-0.0308478139,-0.188172102,0.1618311256,-0.1791258454,-0.2661722004,-0.3849771619,0.1085400954,-0.1376332939,0.0371233523,0.1283319592,0.1249253303,-0.1639446914,0.0098909764,0.2198198438,0.0911266208,-0.3888247311,0.3381742537,-0.1077874154,-0.1485274583,-0.1235791966,0.1685680747,0.2814444602,0.0629519448,0.3650933504,-0.0872263014,-0.2079828233,0.1700563133,0.2281175256,-0.1518964171,0.2485146672,-0.1161972806,-0.3701640368,-0.2486976683,0.4615723789,-0.1306284964,0.0186716523,-0.0521034189,0.0991068631,0.003739581,-0.0162424091,-0.2511475086,0.0053935279,0.0735629424,0.1900277734,-0.4759955704,0.1512910128,0.2287879288,0.110932909,0.0573695824,0.2607388794,0.0007869502,-0.0999411941,-0.2160836458,0.1331126839,0.1301591247,-0.1403605193,-0.1228807792,0.0188462827,-0.0731661841,-0.0717758238,0.216513291,0.0835128203,-0.0267807152,0.2172831297,0.1311217993,0.1191414148,0.3056656718,0.0038306452,-0.0522037186,0.124015674,0.1425275654,0.1671767533,-0.0698422417,-0.0610502064,-0.5958340168,0.0865877271,-0.1721179634,0.334825635,0.3047714829,-0.2367942035,0.0305576939,0.0753731653,0.3694211245,0.314019382,0.0147067439,-0.0130339265,0.0858267918,0.1819656938,-0.3525706828,-0.1376184076,0.3225528002,0.0614038594,0.1956574023,0.1185259968,-0.3290822804,0.2033440918,0.2233221978,-0.120464392,0.4577661753,0.0393730588,0.2156214267,0.5496623516,0.0822913945,0.1059112772,0.4552771747,-0.1346557885,-0.0528844185,0.7670977712,0.0000362798,0.4767209888,0.5375198126,0.1209944114,0.1651364565,-0.116949372,-0.061264392,0.0139843347,-0.0742452666,0.2528257966,-0.1285436898,-0.2894105911,-0.2081596702,-0.0167670101,-0.1109244302,-0.0039245086,-0.1681597084,-0.2484375685,-0.1625422835,-0.2450120449,-0.0422880724,-0.0708364993,-0.160419032,-0.0501407646,0.1269110143,-0.3143103123,-0.0143398643,-0.1467229724,-0.2234831899,-0.0081214029,0.2016516626,-0.0929878354,0.3207439482,0.2998452187,0.159827143,0.3399278522,0.2196566463,0.3316202462,0.443985045,-0.0927284956,-0.1078115478,-0.1613139808,-0.0841439664,-0.1108931452,0.2211236358,-0.0047986256,0.1254571676,0.1096822023,0.1607336253,-0.0612043366,0.3804636002,0.3565916419,0.0261916313,-0.3428668082,0.0200863183,-0.0340408646,0.0160514023,-0.0020940849,-0.0784246698,-0.0881817192,-0.2545608282,0.4558415115,0.3104175329,0.1873828918,-0.1252919883,0.0795623884,-0.0890269503,0.3357637227,0.2054750472,0.0817027614,-0.1674567759,0.0917910859,-0.4356390238,-0.0776815191,0.0969927013,-0.0606095158,0.2765798271,-0.0421167389,0.0987063646,0.2266128808,0.0339960195,0.1146242544,-0.1884645075,-0.1508150697,-0.5087311268,-0.089195773,0.3614899218,0.1643898338,0.071918197,-0.2774338424,0.0469553731,-0.0643579736,0.0770169273,-0.192634359,-0.1640274078,0.1316321939,0.5628841519,0.121633105,0.3136529624,0.4868503809,0.0407279916,-0.0813758746,-0.3092046678,-0.3655738533,-0.2650110424,0.2840080559,-0.0509732217,0.5357193947,-0.1638861448,-0.3490848541,-0.0664884001,0.264821738,-0.157160446,-0.188663289,-0.4559234977,0.2295293361,0.0384258516,0.0980730504,-0.4108909369,0.3706010282,0.111948207,0.4879986346,-0.2702848613,-0.6187790036,0.4536240399,-0.2140198201,-0.4248342216,-0.2510511577,0.067547515,0.1526652873,-0.1547737271,-0.4326290786,-0.0599381812,0.4278156459,-0.0611086115,-0.2852191627,0.0017881198,0.2702943087,0.3620921075,-0.1158800274,-0.1328708529,0.1866863668,-0.0891441852,0.1879659444,-0.1533495486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2927","title":"Datasets 1.12 dataset.filter TypeError: get_indices_from_mask_function() got an unexpected keyword argument","comments":"Fixed by #2950.","body":"## Describe the bug\r\nUpgrading to 1.12 caused `dataset.filter` call to fail with \r\n\r\n> get_indices_from_mask_function() got an unexpected keyword argument valid_rel_labels\r\n\r\n\r\n## Steps to reproduce the bug\r\n```pythondef \r\n\r\nfilter_good_rows(\r\n ex: Dict,\r\n valid_rel_labels: Set[str],\r\n valid_ner_labels: Set[str],\r\n tokenizer: PreTrainedTokenizerFast,\r\n) -> bool:\r\n \"\"\"Get the good rows\"\"\"\r\n encoding = get_encoding_for_text(text=ex[\"text\"], tokenizer=tokenizer)\r\n ex[\"encoding\"] = encoding\r\n for relation in ex[\"relations\"]:\r\n if not is_valid_relation(relation, valid_rel_labels):\r\n return False\r\n for span in ex[\"spans\"]:\r\n if not is_valid_span(span, valid_ner_labels, encoding):\r\n return False\r\n return True\r\n \r\ndef get_dataset(): \r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path,\r\n name=\"prodigy-dataset\",\r\n data_files=sorted(file_paths),\r\n cache_dir=cache_dir,\r\n )[\"train\"]\r\n\r\n valid_ner_labels = set(vocab.ner_category)\r\n valid_relations = set(vocab.relation_types.keys())\r\n ds = ds.filter(\r\n filter_good_rows,\r\n fn_kwargs=dict(\r\n valid_rel_labels=valid_relations,\r\n valid_ner_labels=valid_ner_labels,\r\n tokenizer=vocab.tokenizer,\r\n ),\r\n keep_in_memory=True,\r\n num_proc=num_proc,\r\n )\r\n\r\n```\r\n\r\n`ds` is a `DatasetDict` produced by a jsonl dataset.\r\nThis runs fine on 1.11 but fails on 1.12\r\n\r\n**Stack Trace**\r\n\r\n\r\n\r\n## Expected results\r\n\r\nI expect 1.12 datasets filter to filter the dataset without raising as it does on 1.11\r\n\r\n## Actual results\r\n```\r\ntf_ner_rel_lib\/dataset.py:695: in load_prodigy_arrow_datasets_from_jsonl\r\n ds = ds.filter(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2169: in filter\r\n indices = self.map(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1686: in map\r\n return self._map_single(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2048: in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\ninputs = {'_input_hash': [2108817714, 1477695082, -1021597032, 2130671338, -1260483858, -1203431639, ...], '_task_hash': [18070...ons', 'relations', 'relations', ...], 'answer': ['accept', 'accept', 'accept', 'accept', 'accept', 'accept', ...], ...}\r\nindices = [0, 1, 2, 3, 4, 5, ...], check_same_num_examples = False, offset = 0\r\n\r\n def apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples=False, offset=0):\r\n \"\"\"Utility to apply the function on a selection of columns.\"\"\"\r\n nonlocal update_data\r\n fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n if offset == 0:\r\n effective_indices = indices\r\n else:\r\n effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n processed_inputs = (\r\n> function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n )\r\nE TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'valid_rel_labels'\r\n\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1939: TypeError\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Mac\r\n- Python version: 3.8.9\r\n- PyArrow version: pyarrow==5.0.0\r\n\r\n","comment_length":3,"text":"Datasets 1.12 dataset.filter TypeError: get_indices_from_mask_function() got an unexpected keyword argument\n## Describe the bug\r\nUpgrading to 1.12 caused `dataset.filter` call to fail with \r\n\r\n> get_indices_from_mask_function() got an unexpected keyword argument valid_rel_labels\r\n\r\n\r\n## Steps to reproduce the bug\r\n```pythondef \r\n\r\nfilter_good_rows(\r\n ex: Dict,\r\n valid_rel_labels: Set[str],\r\n valid_ner_labels: Set[str],\r\n tokenizer: PreTrainedTokenizerFast,\r\n) -> bool:\r\n \"\"\"Get the good rows\"\"\"\r\n encoding = get_encoding_for_text(text=ex[\"text\"], tokenizer=tokenizer)\r\n ex[\"encoding\"] = encoding\r\n for relation in ex[\"relations\"]:\r\n if not is_valid_relation(relation, valid_rel_labels):\r\n return False\r\n for span in ex[\"spans\"]:\r\n if not is_valid_span(span, valid_ner_labels, encoding):\r\n return False\r\n return True\r\n \r\ndef get_dataset(): \r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path,\r\n name=\"prodigy-dataset\",\r\n data_files=sorted(file_paths),\r\n cache_dir=cache_dir,\r\n )[\"train\"]\r\n\r\n valid_ner_labels = set(vocab.ner_category)\r\n valid_relations = set(vocab.relation_types.keys())\r\n ds = ds.filter(\r\n filter_good_rows,\r\n fn_kwargs=dict(\r\n valid_rel_labels=valid_relations,\r\n valid_ner_labels=valid_ner_labels,\r\n tokenizer=vocab.tokenizer,\r\n ),\r\n keep_in_memory=True,\r\n num_proc=num_proc,\r\n )\r\n\r\n```\r\n\r\n`ds` is a `DatasetDict` produced by a jsonl dataset.\r\nThis runs fine on 1.11 but fails on 1.12\r\n\r\n**Stack Trace**\r\n\r\n\r\n\r\n## Expected results\r\n\r\nI expect 1.12 datasets filter to filter the dataset without raising as it does on 1.11\r\n\r\n## Actual results\r\n```\r\ntf_ner_rel_lib\/dataset.py:695: in load_prodigy_arrow_datasets_from_jsonl\r\n ds = ds.filter(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2169: in filter\r\n indices = self.map(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1686: in map\r\n return self._map_single(\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:185: in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py:398: in wrapper\r\n out = func(self, *args, **kwargs)\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:2048: in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\ninputs = {'_input_hash': [2108817714, 1477695082, -1021597032, 2130671338, -1260483858, -1203431639, ...], '_task_hash': [18070...ons', 'relations', 'relations', ...], 'answer': ['accept', 'accept', 'accept', 'accept', 'accept', 'accept', ...], ...}\r\nindices = [0, 1, 2, 3, 4, 5, ...], check_same_num_examples = False, offset = 0\r\n\r\n def apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples=False, offset=0):\r\n \"\"\"Utility to apply the function on a selection of columns.\"\"\"\r\n nonlocal update_data\r\n fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n if offset == 0:\r\n effective_indices = indices\r\n else:\r\n effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n processed_inputs = (\r\n> function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n )\r\nE TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'valid_rel_labels'\r\n\r\n..\/..\/..\/..\/.pyenv\/versions\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:1939: TypeError\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Mac\r\n- Python version: 3.8.9\r\n- PyArrow version: pyarrow==5.0.0\r\n\r\n\nFixed by #2950.","embeddings":[-0.2829699218,0.2777142525,-0.0049711736,0.1934766024,0.0521399193,-0.0259464737,0.2624210715,0.4762106538,-0.1806556284,0.0836368799,-0.1311300099,0.3645940423,-0.1310276091,0.1083575189,-0.1563561261,-0.2259478569,0.0624334253,0.0486550108,-0.0484171957,0.0325096361,-0.3871050477,0.4132161736,-0.4211299717,0.1235269159,0.008685492,-0.1654407978,0.1102141142,-0.0451107584,0.0099631641,-0.4235654175,0.4564154148,0.0699703991,0.0381553695,0.2566260993,-0.0001171321,0.2396819293,0.366684258,-0.0842551515,-0.2250449806,-0.405336529,-0.1809279323,-0.0339020006,0.1180416197,-0.1935501844,-0.0818986967,0.0174917243,-0.249624908,-0.2356579155,0.2611079514,0.4325256944,0.1972838491,0.0717553645,-0.0651654452,0.0339874998,0.0209346432,0.2666781545,0.0185224004,-0.0007021982,0.0587287322,-0.1879914552,0.2919843793,0.5048494339,-0.2368306518,-0.2743239999,0.2988181114,-0.2294282466,0.223627314,-0.3586401939,0.2479036301,-0.0648386478,0.0247386377,-0.0833451524,-0.3223271072,-0.1697276235,-0.2576664686,-0.1120616794,0.3498052657,-0.1584596336,-0.244265154,-0.0208345316,-0.07258939,-0.0044453437,-0.0193625409,0.0009967602,-0.0110017173,0.5335709453,0.051298894,0.0513096936,0.0644063577,-0.1729889512,0.3066953421,-0.1076792255,-0.1007701308,0.0617668256,-0.170351848,0.0725231543,0.2387878448,-0.100716427,0.0953230187,-0.0929184034,-0.0909817293,-0.1606885344,0.0432230234,-0.0053427084,-0.0694495961,0.0904969722,0.3082866073,0.6318762302,0.0672923923,0.058587458,-0.0616019517,0.0373482928,-0.0557346605,0.0807316527,0.0906984136,0.2471883744,0.3795975745,-0.3135053217,-0.6206545234,0.1616326869,-0.5770969987,-0.1607390344,0.237689808,0.0602699555,0.2683047056,0.0857219398,-0.0559860393,0.0184919778,-0.0884351358,0.0967272371,-0.1329723448,0.0300575532,-0.0661212951,-0.288934499,0.1902687103,-0.5589625239,0.0788640752,-0.1231013834,-0.090868406,-0.0088757947,-0.0592832975,0.0524173155,0.1024440303,0.3941213489,-0.4122775495,0.1066873968,0.2004962116,-0.5756821632,-0.1944885254,0.2767592072,-0.4454355538,-0.0839394405,0.0214587208,0.2041888237,0.1366723627,-0.1789838076,-0.3426181376,0.5803837776,0.0123843485,-0.2307372093,-0.0704242811,-0.2239028215,-0.3683023453,-0.0239756647,0.1340631545,0.1689580381,-0.8365051746,-0.2499466091,0.2335013896,-0.0030520104,-0.0626956224,-0.2084405422,-0.1818583906,-0.0219790861,-0.1172454953,-0.0013353637,0.3576940596,-0.485848546,-0.6574547887,0.1390666068,0.1225496158,0.1945742369,0.1314316094,-0.1025109664,0.3647884429,-0.0999429971,0.1756555587,0.2742819488,0.0124631217,-0.0347300246,-0.2012303621,0.0749125555,0.2941287458,-0.0680078194,0.1666537374,0.27637887,-0.1620120853,-0.0194072127,0.2861098051,0.1138249561,-0.0791773424,0.0638790876,0.2018899024,-0.1175088063,0.232898742,-0.4075967968,-0.1830836833,-0.0563953593,0.0861253291,0.130115062,-0.3378736973,-0.214805305,-0.3373113871,0.0745304599,-0.2046329081,-0.1652506888,0.1793425977,0.1503546089,-0.0387527496,-0.0507269315,-0.0374752656,0.4382365942,0.2428901494,0.1432452947,-0.2216559798,0.2244707495,-0.0795594603,-0.080613561,0.0983143821,-0.0522183664,0.36705634,-0.0016741207,-0.3518117964,0.2039128393,0.0591471232,-0.3466408551,-0.0529597551,0.0107605178,0.0352687202,0.0651153848,0.1794721931,0.3757503927,-0.0078226523,0.1407691687,-0.1973981559,0.4055749774,-0.0408483408,0.4687744379,0.0199777223,0.0771241337,0.3994879127,0.2407884896,-0.1524210721,-0.1703110337,0.0525777675,0.3843008578,0.2716696858,-0.1308127642,-0.0575984567,0.2184957862,0.3087431192,-0.0334121324,0.0245152265,0.3014428616,0.015626844,0.0211441629,0.1812755615,0.4510476291,0.150378406,0.1575035751,-0.1162214428,0.0736183971,-0.2908762395,-0.0003076858,0.2586247027,0.1770678163,-0.2355194986,-0.0977184847,0.2307223231,0.0618963651,-0.3269087672,-0.4945808649,-0.1177324802,0.1571106911,-0.2101406008,0.4223466814,-0.2563110888,-0.360643059,0.2112966329,-0.1757659763,-0.1615487486,-0.4771515429,0.0838324726,0.1105318815,0.0211028233,0.4142694175,-0.1944871247,0.2770675421,0.1981321126,-0.2057120055,-0.2324998677,-0.3686217666,-0.3903269172,0.0790402368,0.1200828031,0.0579170287,0.1671684682,-0.0545275025,0.1354142576,0.083329387,-0.7349098921,0.1015762314,-0.3447894454,0.4745761752,0.1016463861,0.038837377,-0.3328078389,-0.2596282661,0.1520894766,0.304394573,-0.2402678132,0.1635410488,0.0431459434,0.2904618084,-0.2289169878,-0.4294990301,-0.2123535424,-0.1180267707,-0.1469037235,-0.1846597344,0.3282931149,0.5408180952,0.0719795302,-0.0907747671,0.1236420944,0.184421882,-0.3507416248,-0.4034056962,0.144136861,-0.0788861513,-0.1249593124,-0.1968769431,-0.2286561579,0.0848402604,0.2183059156,-0.5605329275,-0.4192049503,0.0349870473,0.0410459861,0.1570181549,-0.1139656156,0.1058698297,0.2774389684,0.030448081,-0.2724855542,-0.3693442345,0.0357308015,-0.2114411592,0.0473164991,0.0135325789,0.9601014853,0.0164168794,0.0682780072,0.1694822013,0.2156193256,0.3308912516,0.0938933268,0.4396736324,-0.2999125719,-0.2751029134,-0.1022948325,0.0323724411,-0.0618773438,-0.2904603481,0.0270541478,-0.0472293831,-0.1075963899,-0.1457635909,-0.2428792566,-0.307492286,-0.1465963423,-0.1706662029,0.0713224858,0.0963586718,0.3636464477,-0.308096081,-0.0618637502,0.0288702194,0.2998217344,-0.0606535003,0.0504504703,-0.4331361651,-0.0584681481,-0.3299600482,0.3512754738,0.0836806521,0.5775329471,0.0337941647,0.1314383298,-0.1877772957,-0.0064897859,0.6513140798,-0.37467134,0.0990580469,0.3981027007,-0.0517896451,-0.105050236,-0.0820436478,-0.3614759147,0.1523979008,-0.130425632,0.3490183353,-0.2240297794,-0.0515188314,0.3174758554,0.0232320037,-0.1624078602,0.1051052287,-0.3738526404,-0.5830719471,-0.0338545628,0.2991501689,0.0823923871,0.4687190056,0.0837701112,0.0468495339,-0.0865380093,-0.4693993032,0.1605705619,0.2456996441,0.1313659102,0.1721284539,0.1322498471,0.1905105859,0.1852574944,0.2601747811,0.3494502306,-0.1209676489,-0.429931134,-0.1555662006,-0.003165609,0.2964791059,0.0886024907,-0.2174897939,-0.0622401759,-0.1156746596,-0.1378526092,-0.0112768495,0.1836106479,0.2737488151,0.2099698633,-0.4123848975,-0.5313845873,0.1796308607,0.2255197465,-0.1604507267,0.3734816909,0.2652886808,-0.4143254757,0.4304826856,-0.1537723541,0.7024047375,0.2602522671,-0.1502133161,0.3903979361,-0.1639682204,0.3863707483,0.1703226268,0.1472315341,-0.0537542589,-0.3721880913,-0.0091732787,0.0341162309,-0.1155766919,0.0897843763,-0.2464871407,0.3867236376,-0.2271939069,-0.1297411919,0.1523538828,0.0656017885,0.072543852,0.1573487818,-0.436416477,0.1157073081,-0.1356859058,-0.1174689382,-0.0174163952,-0.157321766,0.2357757241,-0.4989627004,-0.2169850469,0.0593276396,-0.2186802477,0.1991744936,0.1808251143,-0.5248704553,-0.0648283362,-0.2381993383,0.1592858881,0.1625673175,0.0071186177,0.1369998753,-0.1568153501,0.4103268683,0.1870173216,-0.0307340231,0.4770958126,0.1361288428,-0.0680285394,0.1569158882,0.016583018,-0.282654196,-0.3184965551,-0.1825746,-0.0308478139,-0.188172102,0.1618311256,-0.1791258454,-0.2661722004,-0.3849771619,0.1085400954,-0.1376332939,0.0371233523,0.1283319592,0.1249253303,-0.1639446914,0.0098909764,0.2198198438,0.0911266208,-0.3888247311,0.3381742537,-0.1077874154,-0.1485274583,-0.1235791966,0.1685680747,0.2814444602,0.0629519448,0.3650933504,-0.0872263014,-0.2079828233,0.1700563133,0.2281175256,-0.1518964171,0.2485146672,-0.1161972806,-0.3701640368,-0.2486976683,0.4615723789,-0.1306284964,0.0186716523,-0.0521034189,0.0991068631,0.003739581,-0.0162424091,-0.2511475086,0.0053935279,0.0735629424,0.1900277734,-0.4759955704,0.1512910128,0.2287879288,0.110932909,0.0573695824,0.2607388794,0.0007869502,-0.0999411941,-0.2160836458,0.1331126839,0.1301591247,-0.1403605193,-0.1228807792,0.0188462827,-0.0731661841,-0.0717758238,0.216513291,0.0835128203,-0.0267807152,0.2172831297,0.1311217993,0.1191414148,0.3056656718,0.0038306452,-0.0522037186,0.124015674,0.1425275654,0.1671767533,-0.0698422417,-0.0610502064,-0.5958340168,0.0865877271,-0.1721179634,0.334825635,0.3047714829,-0.2367942035,0.0305576939,0.0753731653,0.3694211245,0.314019382,0.0147067439,-0.0130339265,0.0858267918,0.1819656938,-0.3525706828,-0.1376184076,0.3225528002,0.0614038594,0.1956574023,0.1185259968,-0.3290822804,0.2033440918,0.2233221978,-0.120464392,0.4577661753,0.0393730588,0.2156214267,0.5496623516,0.0822913945,0.1059112772,0.4552771747,-0.1346557885,-0.0528844185,0.7670977712,0.0000362798,0.4767209888,0.5375198126,0.1209944114,0.1651364565,-0.116949372,-0.061264392,0.0139843347,-0.0742452666,0.2528257966,-0.1285436898,-0.2894105911,-0.2081596702,-0.0167670101,-0.1109244302,-0.0039245086,-0.1681597084,-0.2484375685,-0.1625422835,-0.2450120449,-0.0422880724,-0.0708364993,-0.160419032,-0.0501407646,0.1269110143,-0.3143103123,-0.0143398643,-0.1467229724,-0.2234831899,-0.0081214029,0.2016516626,-0.0929878354,0.3207439482,0.2998452187,0.159827143,0.3399278522,0.2196566463,0.3316202462,0.443985045,-0.0927284956,-0.1078115478,-0.1613139808,-0.0841439664,-0.1108931452,0.2211236358,-0.0047986256,0.1254571676,0.1096822023,0.1607336253,-0.0612043366,0.3804636002,0.3565916419,0.0261916313,-0.3428668082,0.0200863183,-0.0340408646,0.0160514023,-0.0020940849,-0.0784246698,-0.0881817192,-0.2545608282,0.4558415115,0.3104175329,0.1873828918,-0.1252919883,0.0795623884,-0.0890269503,0.3357637227,0.2054750472,0.0817027614,-0.1674567759,0.0917910859,-0.4356390238,-0.0776815191,0.0969927013,-0.0606095158,0.2765798271,-0.0421167389,0.0987063646,0.2266128808,0.0339960195,0.1146242544,-0.1884645075,-0.1508150697,-0.5087311268,-0.089195773,0.3614899218,0.1643898338,0.071918197,-0.2774338424,0.0469553731,-0.0643579736,0.0770169273,-0.192634359,-0.1640274078,0.1316321939,0.5628841519,0.121633105,0.3136529624,0.4868503809,0.0407279916,-0.0813758746,-0.3092046678,-0.3655738533,-0.2650110424,0.2840080559,-0.0509732217,0.5357193947,-0.1638861448,-0.3490848541,-0.0664884001,0.264821738,-0.157160446,-0.188663289,-0.4559234977,0.2295293361,0.0384258516,0.0980730504,-0.4108909369,0.3706010282,0.111948207,0.4879986346,-0.2702848613,-0.6187790036,0.4536240399,-0.2140198201,-0.4248342216,-0.2510511577,0.067547515,0.1526652873,-0.1547737271,-0.4326290786,-0.0599381812,0.4278156459,-0.0611086115,-0.2852191627,0.0017881198,0.2702943087,0.3620921075,-0.1158800274,-0.1328708529,0.1866863668,-0.0891441852,0.1879659444,-0.1533495486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2924","title":"\"File name too long\" error for file locks","comments":"Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock\")\r\n154\r\n```\r\nso not sure why it's considered too long for your filesystem.\r\n(also note that the lock files we use always have smaller filenames than 255)\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/5d1a9f1e3c6c495dc0610b459e39d2eb8893f152\/src\/datasets\/utils\/filelock.py#L135-L135","body":"## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t\/test\")`:\r\n\r\n```\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nWhere the user cache dir (e.g. `~\/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"gar1t\/test\")\r\n```\r\n\r\n## Expected results\r\n\r\nExpect the function to return without an error.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 765, in _save_info\r\n with FileLock(lock_path):\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 403, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.7\r\n- PyArrow version: 5.0.0\r\n","comment_length":39,"text":"\"File name too long\" error for file locks\n## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t\/test\")`:\r\n\r\n```\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nWhere the user cache dir (e.g. `~\/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"gar1t\/test\")\r\n```\r\n\r\n## Expected results\r\n\r\nExpect the function to return without an error.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 765, in _save_info\r\n with FileLock(lock_path):\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 403, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.7\r\n- PyArrow version: 5.0.0\r\n\nHi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock\")\r\n154\r\n```\r\nso not sure why it's considered too long for your filesystem.\r\n(also note that the lock files we use always have smaller filenames than 255)\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/5d1a9f1e3c6c495dc0610b459e39d2eb8893f152\/src\/datasets\/utils\/filelock.py#L135-L135","embeddings":[0.0493393503,0.0920015797,-0.0714939609,0.401848346,0.4151671231,0.2358950824,0.636662364,0.2435030937,0.2364584506,0.2313297093,0.0185690634,0.0077435691,-0.1442685276,-0.3100371361,-0.2015356272,-0.2443143427,-0.1581087261,-0.0416213386,-0.177078411,0.2177188098,-0.1549794227,0.3787363768,0.0246380791,0.1570838541,-0.536006391,0.0900789052,-0.1295958906,0.3886749148,-0.0215065964,-0.3729913831,0.051118616,-0.0467165932,0.068814449,0.811357677,-0.0001220686,-0.3334413767,0.3542338908,-0.0562992319,-0.3271892667,-0.2243083715,-0.1685816944,-0.5117114782,-0.0345845483,-0.4988024831,0.1872130185,-0.0951189175,-0.0331902057,-0.8021566272,0.0410523824,0.3733062148,0.1268745214,-0.0783471763,0.0910248384,-0.2579201758,0.3323119283,-0.2779891491,0.0040849657,0.3564026058,0.3556975722,-0.0988676101,-0.1546510309,0.2032182515,-0.0368982404,0.027324792,0.2471582741,0.079537563,-0.0874831453,-0.3261650205,0.2958056033,0.4273941517,0.4714379907,-0.0930620804,-0.2793129385,-0.4676968753,0.1424980164,-0.1251056492,0.4885306358,-0.0762403011,-0.222074613,0.1514885277,-0.1704189926,-0.0196209587,-0.1235974729,-0.1102384105,-0.101975739,0.0650619194,0.0405084826,0.0362790599,0.3111130893,-0.3483147621,0.2395959646,0.0048368936,0.0742740706,0.2694604397,-0.6542128921,0.1469123662,0.0842684656,0.3854928017,0.2534896433,0.0576435812,-0.2659727335,-0.1483078748,0.1947487742,-0.0104660569,-0.0649549589,0.374206692,0.1944440156,0.1612814218,0.3012429774,0.0779469088,-0.3367469013,-0.1069349945,-0.0764871612,-0.5534629822,0.445138216,0.133591041,0.0741629824,-0.3437117934,0.1504577994,0.5705774426,0.1057047993,-0.0104902675,0.2287023664,0.2733797133,0.005434711,0.1107665151,0.0830746666,-0.0899362862,-0.0889612734,0.1008345857,-0.1473124325,-0.1107303202,-0.2270203382,0.1124099791,0.0892055854,-0.3083957136,0.2792796195,-0.227465108,0.3152402341,-0.1249576584,-0.0829687566,-0.2377426773,-0.0667323172,0.1559667885,-0.0591069087,0.0739138052,0.2452137023,-0.2607334852,-0.1920612901,-0.0771116391,-0.3397490382,-0.2851879001,-0.0617215782,0.0947027877,-0.165494591,0.2159024328,0.2415551245,-0.2671947181,0.6555529237,-0.0798197463,0.0127971917,-0.2747654617,-0.1418095827,0.0071923966,0.0408245735,0.5534662008,0.1325238049,-0.0175623372,-0.0687896535,0.1282823086,0.0086658485,0.3722593188,0.1806448698,-0.0535966232,-0.4219526947,0.3124936521,0.2268715054,-0.2402810752,-0.486985594,0.3215382397,-0.394307375,0.0994933695,0.3229117393,0.0228835605,0.0517107099,-0.024582373,0.2853384614,0.1822132915,-0.0260959864,0.0234954469,-0.202524215,-0.1073302031,-0.0406307913,0.1691952497,0.0109546073,0.0670013204,0.0796193257,-0.1751918048,0.2715687454,0.0314637274,-0.2040443867,0.3736798167,0.1382424235,0.3267191648,0.1860822737,-0.106327638,-0.5937504768,0.2637207508,-0.1218474954,-0.1787318438,-0.2175928354,-0.1868047863,-0.1749890149,-0.0072898441,0.0034599197,0.1351371109,-0.0204084311,0.2832494974,0.1417792886,-0.1317639649,0.0934399366,0.5883799791,-0.1702439636,0.0175973084,-0.4781179428,-0.17112647,-0.1178534552,-0.1125399023,0.0315071233,0.0016395918,0.3693964779,-0.1089679822,-0.3443510532,0.4480753541,0.3332768381,0.0975331739,-0.1789251119,-0.0044306694,-0.0568642691,0.220775336,0.0726701096,0.0398121551,0.0077854944,0.0142915584,-0.2135758698,0.3000163734,-0.14081572,0.27127707,0.0286138058,0.1682408899,0.2461784184,-0.1318140626,0.0643453151,-0.1761228889,0.6458842158,-0.2163528502,0.3048405647,0.0547799319,0.0452316478,-0.1448624283,0.6458194852,0.045392707,-0.0001772655,0.1977425367,0.1562013328,-0.0408157147,-0.0193730351,0.3064890206,0.5111442208,0.1920584738,-0.0301231258,-0.2065057307,0.3941576481,-0.2238344699,0.256319046,0.0159084182,-0.0027928618,0.3657026589,0.0496179983,-0.1290875971,0.0514840819,-0.8095089197,-0.050442148,0.223697722,-0.3198090196,0.1480822265,-0.1393256634,-0.0432045013,-0.0237995591,0.0004903108,-0.0803984031,-0.4162820876,-0.1888936311,0.2785063386,-0.1418365985,0.2145332694,-0.1188357845,-0.0032407197,0.2712830305,-0.0145387761,-0.2950273752,-0.0016123495,0.0803438574,-0.0837603286,0.2720724344,-0.5375033021,0.0838043541,-0.1256639212,-0.0631072521,-0.5083525181,-0.2498805523,0.1116677895,-0.1684809178,0.2935642302,0.3413662016,0.2078575194,0.1666625887,0.0588969849,0.0049610781,-0.2622792721,0.0236695651,0.2326634228,0.1779004186,-0.0228179917,-0.1917994022,-0.023455549,-0.3463737965,-0.4160333872,0.3207841218,0.0303027015,0.2279285491,0.3411892653,-0.0708731115,0.2978130579,-0.0179723371,0.2472920567,-0.0879067779,-0.446370095,0.4170611799,-0.1525222659,-0.2990590334,-0.4318080544,0.2980134785,-0.0355094597,0.1055189967,-0.4223643541,-0.1790713519,-0.401260376,0.3389849961,-0.3357198238,0.0054093283,0.1970204413,-0.0852006599,-0.0563488565,-0.1069383249,0.0849020928,0.2325119227,0.0110361269,-0.009520852,0.0929816738,0.2951667607,-0.1160778701,0.4532136321,0.3616654873,0.0753480569,0.1953798383,-0.1376693696,0.1843574643,-0.2590326071,-0.1269457787,0.0239379536,0.1744713485,-0.221584484,0.1342053711,-0.0853912458,0.0140818916,-0.1333202869,-0.0577918552,0.1078078598,-0.2960295677,0.0073167868,0.0485861041,0.0524898916,0.031080164,0.1456944346,-0.1948870718,-0.1309104413,-0.0240893289,0.3084784746,0.0540789627,-0.0233560894,-0.2468729615,0.0915106237,-0.3371418118,0.4440097511,0.1830917299,0.2137979567,-0.069359839,-0.1601748466,-0.0800488889,0.0460065827,0.6831169128,0.0991293713,0.3160929978,0.1554959416,0.2547036707,-0.1230872124,0.0391722023,0.0795527846,0.3701882362,0.1390409023,0.4079330266,-0.2308098525,0.037608888,-0.183316946,0.1855425686,-0.1166330427,-0.4492169619,-0.0262856334,0.0087962123,-0.2636858821,0.1616199315,0.0336052962,0.2442290634,0.0616719015,0.0049132649,-0.0601768605,-0.0747785419,0.1601556242,0.1748355776,0.2812633216,-0.4468779564,0.264365375,-0.027559543,-0.1595295519,0.2304127961,0.5639728904,-0.4439823329,-0.2478427738,0.4144729376,-0.356559068,0.0764389262,0.4287089705,-0.3107663095,-0.1584243178,0.3582065701,0.6321594715,0.1236561164,0.0257820133,0.0325685404,0.1308744252,-0.1530217081,-0.3346109092,0.2163310051,0.243095085,-0.0178354532,0.5360283852,0.0242514107,-0.3025455177,0.2613933384,-0.2059544474,1.02094841,-0.1854486763,0.1871506423,-0.0190800447,0.1287675947,0.4992987216,0.0092902463,-0.0842248872,-0.4268136919,-0.4827860296,0.1402025223,-0.149804309,0.2397724092,-0.0366778336,-0.1301997006,0.3139286041,-0.2595103979,0.0177618489,-0.1745316088,0.4527935386,-0.6466593742,-0.2542138696,-0.2503230572,0.0479833744,-0.0439814702,0.1101506725,0.0617118813,0.168685928,-0.329262048,-0.2781674564,-0.2272064239,0.0370319821,-0.4157636762,0.0835297033,-0.2439022213,-0.282207489,-0.0902421251,0.2852126062,-0.0320501514,0.4013364017,-0.2035952657,0.0923504159,0.0828223079,0.0968723148,0.073000282,-0.0632440969,0.3176957965,0.0168975573,0.0743537098,-0.2857038081,0.0053972583,-0.1422331929,0.032482598,-0.02342714,0.1504140943,-0.2308297455,-0.3237563372,-0.3392715752,-0.2038627863,-0.1504354775,0.0521603599,0.3335699141,-0.2387240082,-0.0088765109,-0.2857566476,-0.4090466499,-0.0096647162,0.3055093884,-0.148109138,-0.047204867,0.4327270687,0.3000046313,-0.2102095187,-0.1779171973,-0.0190174319,-0.1799966097,-0.6621257663,0.0823084489,0.0360400341,-0.0427111723,-0.2416065782,-0.2668305337,0.2053605467,0.1265298724,0.1113983393,-0.46979931,-0.4082916677,-0.2329967916,-0.2976679206,0.2016152292,-0.0566246063,-0.3551495075,-0.258842051,0.0327734798,-0.2122098804,0.0599939972,-0.1423732638,-0.0618365444,0.4633510113,-0.1418498307,0.2244604528,-0.1397035271,0.046355471,0.4012090564,0.0694132596,-0.1816515177,-0.3011113703,0.1451775283,0.0411616899,0.0585848205,-0.0672907755,-0.3786494434,-0.052751489,-0.2798325121,0.1326039732,0.3364621699,0.2424450219,0.0756183788,0.09433043,0.098653309,-0.0325292088,0.3306849301,-0.1649502367,-0.1383991838,0.0095238341,0.2725932598,-0.2107255012,0.1175259277,-0.3212481141,0.1536602676,-0.0095391795,-0.0361817405,0.0752809569,-0.0265370756,-0.1096814424,0.5564427376,0.2327184081,0.2920384705,-0.2722925246,-0.2302523255,0.2756311297,0.1339128017,0.0165831745,-0.2051161677,0.1073662043,-0.1286123842,-0.0995853841,0.3734355271,-0.0706336424,-0.1355883777,-0.1899372786,-0.1248223633,0.5824737549,0.0935555845,-0.0696668848,0.2455470562,-0.0136265624,0.0776752979,0.4613701999,0.2502685785,0.0169347022,0.5474853516,-0.3130877316,0.0298885182,-0.1627934128,-0.0062184911,-0.2004130483,-0.4473776221,0.1316621155,0.3107807636,-0.0724566802,0.187366575,-0.1468187869,0.3045750558,-0.2501968741,0.1819011271,0.0543667413,0.049220752,-0.1095726714,-0.0142696071,0.3495133519,-0.0250102021,-0.2757626772,-0.0046619414,-0.11504668,0.1466555595,0.3796023428,0.2733509541,0.1335933954,-0.1753616929,0.0041428227,0.2133577764,0.1663140357,-0.482229948,0.1435226351,0.3909594119,-0.088678211,0.2622956634,-0.0017311651,0.5537607074,0.1397955269,0.0983077362,-0.0007230406,-0.1064622477,-0.011388314,-0.1354410052,0.4446231723,-0.0308796242,-0.1025131792,0.3387673199,0.0280349217,-0.1576936543,0.1146624386,0.1928842962,0.1237160265,-0.1751527786,0.2012574673,0.0158905946,-0.2077661604,-0.1786319464,0.1449805051,-0.3786619902,-0.1633839011,0.2516909838,0.0403841957,0.2624582648,-0.1355939955,0.0489800386,-0.1194112673,0.1483440697,0.3839921951,0.2433493137,-0.2853369117,-0.0405620374,-0.6024463177,0.2246279418,-0.0487730019,-0.1710739583,0.0451600216,-0.1129757613,-0.0471117832,-0.1518046707,0.3878982961,-0.0159281138,0.111534372,0.2670121491,-0.4752603173,-0.2240053713,0.1703025699,-0.0663013086,0.2272349,-0.3988763988,0.1303192973,-0.266743809,-0.0137350131,-0.1426671296,-0.1962016076,0.1286680251,0.3134683967,0.3546344638,0.1124142408,0.2980175614,0.0608979315,-0.249863565,-0.3230270147,0.1182483435,-0.1321166456,0.0156514794,0.2199660689,0.2973281145,-0.4038825631,-0.4329209328,-0.3770737052,0.5200780034,-0.1345874667,0.0714471936,-0.1244051009,-0.0493410081,-0.0306261051,0.2739892602,0.2633121014,0.2337562293,0.070637241,0.4157318175,-0.1906892061,-0.299829632,0.659506321,-0.1065887734,-0.4276301563,0.0981976986,0.4086528718,0.0720963702,-0.1614484638,-0.4712712765,0.1547385156,0.324437052,0.0784328952,-0.2918297052,0.2315367162,-0.2830388844,0.1289331019,-0.104005374,0.2516618371,0.2692824006,-0.2087137401,-0.0591599233,-0.1967329979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2924","title":"\"File name too long\" error for file locks","comments":"Yes, you're right! I need to get you more info here. Either there's something going with the name itself that the file system doesn't like (an encoding that blows up the name length??) or perhaps there's something with the path that's causing the entire string to be used as a name. I haven't seen this on any system before and the Internet's not forthcoming with any info.","body":"## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t\/test\")`:\r\n\r\n```\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nWhere the user cache dir (e.g. `~\/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"gar1t\/test\")\r\n```\r\n\r\n## Expected results\r\n\r\nExpect the function to return without an error.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 765, in _save_info\r\n with FileLock(lock_path):\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 403, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.7\r\n- PyArrow version: 5.0.0\r\n","comment_length":67,"text":"\"File name too long\" error for file locks\n## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t\/test\")`:\r\n\r\n```\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nWhere the user cache dir (e.g. `~\/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"gar1t\/test\")\r\n```\r\n\r\n## Expected results\r\n\r\nExpect the function to return without an error.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 644, in download_and_prepare\r\n self._save_info()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 765, in _save_info\r\n with FileLock(lock_path):\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/lib\/python3.9\/site-packages\/datasets\/utils\/filelock.py\", line 403, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 36] File name too long: '\/.cache\/huggingface\/datasets\/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.1\r\n- Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.7\r\n- PyArrow version: 5.0.0\r\n\nYes, you're right! I need to get you more info here. Either there's something going with the name itself that the file system doesn't like (an encoding that blows up the name length??) or perhaps there's something with the path that's causing the entire string to be used as a name. I haven't seen this on any system before and the Internet's not forthcoming with any info.","embeddings":[0.0493393503,0.0920015797,-0.0714939609,0.401848346,0.4151671231,0.2358950824,0.636662364,0.2435030937,0.2364584506,0.2313297093,0.0185690634,0.0077435691,-0.1442685276,-0.3100371361,-0.2015356272,-0.2443143427,-0.1581087261,-0.0416213386,-0.177078411,0.2177188098,-0.1549794227,0.3787363768,0.0246380791,0.1570838541,-0.536006391,0.0900789052,-0.1295958906,0.3886749148,-0.0215065964,-0.3729913831,0.051118616,-0.0467165932,0.068814449,0.811357677,-0.0001220686,-0.3334413767,0.3542338908,-0.0562992319,-0.3271892667,-0.2243083715,-0.1685816944,-0.5117114782,-0.0345845483,-0.4988024831,0.1872130185,-0.0951189175,-0.0331902057,-0.8021566272,0.0410523824,0.3733062148,0.1268745214,-0.0783471763,0.0910248384,-0.2579201758,0.3323119283,-0.2779891491,0.0040849657,0.3564026058,0.3556975722,-0.0988676101,-0.1546510309,0.2032182515,-0.0368982404,0.027324792,0.2471582741,0.079537563,-0.0874831453,-0.3261650205,0.2958056033,0.4273941517,0.4714379907,-0.0930620804,-0.2793129385,-0.4676968753,0.1424980164,-0.1251056492,0.4885306358,-0.0762403011,-0.222074613,0.1514885277,-0.1704189926,-0.0196209587,-0.1235974729,-0.1102384105,-0.101975739,0.0650619194,0.0405084826,0.0362790599,0.3111130893,-0.3483147621,0.2395959646,0.0048368936,0.0742740706,0.2694604397,-0.6542128921,0.1469123662,0.0842684656,0.3854928017,0.2534896433,0.0576435812,-0.2659727335,-0.1483078748,0.1947487742,-0.0104660569,-0.0649549589,0.374206692,0.1944440156,0.1612814218,0.3012429774,0.0779469088,-0.3367469013,-0.1069349945,-0.0764871612,-0.5534629822,0.445138216,0.133591041,0.0741629824,-0.3437117934,0.1504577994,0.5705774426,0.1057047993,-0.0104902675,0.2287023664,0.2733797133,0.005434711,0.1107665151,0.0830746666,-0.0899362862,-0.0889612734,0.1008345857,-0.1473124325,-0.1107303202,-0.2270203382,0.1124099791,0.0892055854,-0.3083957136,0.2792796195,-0.227465108,0.3152402341,-0.1249576584,-0.0829687566,-0.2377426773,-0.0667323172,0.1559667885,-0.0591069087,0.0739138052,0.2452137023,-0.2607334852,-0.1920612901,-0.0771116391,-0.3397490382,-0.2851879001,-0.0617215782,0.0947027877,-0.165494591,0.2159024328,0.2415551245,-0.2671947181,0.6555529237,-0.0798197463,0.0127971917,-0.2747654617,-0.1418095827,0.0071923966,0.0408245735,0.5534662008,0.1325238049,-0.0175623372,-0.0687896535,0.1282823086,0.0086658485,0.3722593188,0.1806448698,-0.0535966232,-0.4219526947,0.3124936521,0.2268715054,-0.2402810752,-0.486985594,0.3215382397,-0.394307375,0.0994933695,0.3229117393,0.0228835605,0.0517107099,-0.024582373,0.2853384614,0.1822132915,-0.0260959864,0.0234954469,-0.202524215,-0.1073302031,-0.0406307913,0.1691952497,0.0109546073,0.0670013204,0.0796193257,-0.1751918048,0.2715687454,0.0314637274,-0.2040443867,0.3736798167,0.1382424235,0.3267191648,0.1860822737,-0.106327638,-0.5937504768,0.2637207508,-0.1218474954,-0.1787318438,-0.2175928354,-0.1868047863,-0.1749890149,-0.0072898441,0.0034599197,0.1351371109,-0.0204084311,0.2832494974,0.1417792886,-0.1317639649,0.0934399366,0.5883799791,-0.1702439636,0.0175973084,-0.4781179428,-0.17112647,-0.1178534552,-0.1125399023,0.0315071233,0.0016395918,0.3693964779,-0.1089679822,-0.3443510532,0.4480753541,0.3332768381,0.0975331739,-0.1789251119,-0.0044306694,-0.0568642691,0.220775336,0.0726701096,0.0398121551,0.0077854944,0.0142915584,-0.2135758698,0.3000163734,-0.14081572,0.27127707,0.0286138058,0.1682408899,0.2461784184,-0.1318140626,0.0643453151,-0.1761228889,0.6458842158,-0.2163528502,0.3048405647,0.0547799319,0.0452316478,-0.1448624283,0.6458194852,0.045392707,-0.0001772655,0.1977425367,0.1562013328,-0.0408157147,-0.0193730351,0.3064890206,0.5111442208,0.1920584738,-0.0301231258,-0.2065057307,0.3941576481,-0.2238344699,0.256319046,0.0159084182,-0.0027928618,0.3657026589,0.0496179983,-0.1290875971,0.0514840819,-0.8095089197,-0.050442148,0.223697722,-0.3198090196,0.1480822265,-0.1393256634,-0.0432045013,-0.0237995591,0.0004903108,-0.0803984031,-0.4162820876,-0.1888936311,0.2785063386,-0.1418365985,0.2145332694,-0.1188357845,-0.0032407197,0.2712830305,-0.0145387761,-0.2950273752,-0.0016123495,0.0803438574,-0.0837603286,0.2720724344,-0.5375033021,0.0838043541,-0.1256639212,-0.0631072521,-0.5083525181,-0.2498805523,0.1116677895,-0.1684809178,0.2935642302,0.3413662016,0.2078575194,0.1666625887,0.0588969849,0.0049610781,-0.2622792721,0.0236695651,0.2326634228,0.1779004186,-0.0228179917,-0.1917994022,-0.023455549,-0.3463737965,-0.4160333872,0.3207841218,0.0303027015,0.2279285491,0.3411892653,-0.0708731115,0.2978130579,-0.0179723371,0.2472920567,-0.0879067779,-0.446370095,0.4170611799,-0.1525222659,-0.2990590334,-0.4318080544,0.2980134785,-0.0355094597,0.1055189967,-0.4223643541,-0.1790713519,-0.401260376,0.3389849961,-0.3357198238,0.0054093283,0.1970204413,-0.0852006599,-0.0563488565,-0.1069383249,0.0849020928,0.2325119227,0.0110361269,-0.009520852,0.0929816738,0.2951667607,-0.1160778701,0.4532136321,0.3616654873,0.0753480569,0.1953798383,-0.1376693696,0.1843574643,-0.2590326071,-0.1269457787,0.0239379536,0.1744713485,-0.221584484,0.1342053711,-0.0853912458,0.0140818916,-0.1333202869,-0.0577918552,0.1078078598,-0.2960295677,0.0073167868,0.0485861041,0.0524898916,0.031080164,0.1456944346,-0.1948870718,-0.1309104413,-0.0240893289,0.3084784746,0.0540789627,-0.0233560894,-0.2468729615,0.0915106237,-0.3371418118,0.4440097511,0.1830917299,0.2137979567,-0.069359839,-0.1601748466,-0.0800488889,0.0460065827,0.6831169128,0.0991293713,0.3160929978,0.1554959416,0.2547036707,-0.1230872124,0.0391722023,0.0795527846,0.3701882362,0.1390409023,0.4079330266,-0.2308098525,0.037608888,-0.183316946,0.1855425686,-0.1166330427,-0.4492169619,-0.0262856334,0.0087962123,-0.2636858821,0.1616199315,0.0336052962,0.2442290634,0.0616719015,0.0049132649,-0.0601768605,-0.0747785419,0.1601556242,0.1748355776,0.2812633216,-0.4468779564,0.264365375,-0.027559543,-0.1595295519,0.2304127961,0.5639728904,-0.4439823329,-0.2478427738,0.4144729376,-0.356559068,0.0764389262,0.4287089705,-0.3107663095,-0.1584243178,0.3582065701,0.6321594715,0.1236561164,0.0257820133,0.0325685404,0.1308744252,-0.1530217081,-0.3346109092,0.2163310051,0.243095085,-0.0178354532,0.5360283852,0.0242514107,-0.3025455177,0.2613933384,-0.2059544474,1.02094841,-0.1854486763,0.1871506423,-0.0190800447,0.1287675947,0.4992987216,0.0092902463,-0.0842248872,-0.4268136919,-0.4827860296,0.1402025223,-0.149804309,0.2397724092,-0.0366778336,-0.1301997006,0.3139286041,-0.2595103979,0.0177618489,-0.1745316088,0.4527935386,-0.6466593742,-0.2542138696,-0.2503230572,0.0479833744,-0.0439814702,0.1101506725,0.0617118813,0.168685928,-0.329262048,-0.2781674564,-0.2272064239,0.0370319821,-0.4157636762,0.0835297033,-0.2439022213,-0.282207489,-0.0902421251,0.2852126062,-0.0320501514,0.4013364017,-0.2035952657,0.0923504159,0.0828223079,0.0968723148,0.073000282,-0.0632440969,0.3176957965,0.0168975573,0.0743537098,-0.2857038081,0.0053972583,-0.1422331929,0.032482598,-0.02342714,0.1504140943,-0.2308297455,-0.3237563372,-0.3392715752,-0.2038627863,-0.1504354775,0.0521603599,0.3335699141,-0.2387240082,-0.0088765109,-0.2857566476,-0.4090466499,-0.0096647162,0.3055093884,-0.148109138,-0.047204867,0.4327270687,0.3000046313,-0.2102095187,-0.1779171973,-0.0190174319,-0.1799966097,-0.6621257663,0.0823084489,0.0360400341,-0.0427111723,-0.2416065782,-0.2668305337,0.2053605467,0.1265298724,0.1113983393,-0.46979931,-0.4082916677,-0.2329967916,-0.2976679206,0.2016152292,-0.0566246063,-0.3551495075,-0.258842051,0.0327734798,-0.2122098804,0.0599939972,-0.1423732638,-0.0618365444,0.4633510113,-0.1418498307,0.2244604528,-0.1397035271,0.046355471,0.4012090564,0.0694132596,-0.1816515177,-0.3011113703,0.1451775283,0.0411616899,0.0585848205,-0.0672907755,-0.3786494434,-0.052751489,-0.2798325121,0.1326039732,0.3364621699,0.2424450219,0.0756183788,0.09433043,0.098653309,-0.0325292088,0.3306849301,-0.1649502367,-0.1383991838,0.0095238341,0.2725932598,-0.2107255012,0.1175259277,-0.3212481141,0.1536602676,-0.0095391795,-0.0361817405,0.0752809569,-0.0265370756,-0.1096814424,0.5564427376,0.2327184081,0.2920384705,-0.2722925246,-0.2302523255,0.2756311297,0.1339128017,0.0165831745,-0.2051161677,0.1073662043,-0.1286123842,-0.0995853841,0.3734355271,-0.0706336424,-0.1355883777,-0.1899372786,-0.1248223633,0.5824737549,0.0935555845,-0.0696668848,0.2455470562,-0.0136265624,0.0776752979,0.4613701999,0.2502685785,0.0169347022,0.5474853516,-0.3130877316,0.0298885182,-0.1627934128,-0.0062184911,-0.2004130483,-0.4473776221,0.1316621155,0.3107807636,-0.0724566802,0.187366575,-0.1468187869,0.3045750558,-0.2501968741,0.1819011271,0.0543667413,0.049220752,-0.1095726714,-0.0142696071,0.3495133519,-0.0250102021,-0.2757626772,-0.0046619414,-0.11504668,0.1466555595,0.3796023428,0.2733509541,0.1335933954,-0.1753616929,0.0041428227,0.2133577764,0.1663140357,-0.482229948,0.1435226351,0.3909594119,-0.088678211,0.2622956634,-0.0017311651,0.5537607074,0.1397955269,0.0983077362,-0.0007230406,-0.1064622477,-0.011388314,-0.1354410052,0.4446231723,-0.0308796242,-0.1025131792,0.3387673199,0.0280349217,-0.1576936543,0.1146624386,0.1928842962,0.1237160265,-0.1751527786,0.2012574673,0.0158905946,-0.2077661604,-0.1786319464,0.1449805051,-0.3786619902,-0.1633839011,0.2516909838,0.0403841957,0.2624582648,-0.1355939955,0.0489800386,-0.1194112673,0.1483440697,0.3839921951,0.2433493137,-0.2853369117,-0.0405620374,-0.6024463177,0.2246279418,-0.0487730019,-0.1710739583,0.0451600216,-0.1129757613,-0.0471117832,-0.1518046707,0.3878982961,-0.0159281138,0.111534372,0.2670121491,-0.4752603173,-0.2240053713,0.1703025699,-0.0663013086,0.2272349,-0.3988763988,0.1303192973,-0.266743809,-0.0137350131,-0.1426671296,-0.1962016076,0.1286680251,0.3134683967,0.3546344638,0.1124142408,0.2980175614,0.0608979315,-0.249863565,-0.3230270147,0.1182483435,-0.1321166456,0.0156514794,0.2199660689,0.2973281145,-0.4038825631,-0.4329209328,-0.3770737052,0.5200780034,-0.1345874667,0.0714471936,-0.1244051009,-0.0493410081,-0.0306261051,0.2739892602,0.2633121014,0.2337562293,0.070637241,0.4157318175,-0.1906892061,-0.299829632,0.659506321,-0.1065887734,-0.4276301563,0.0981976986,0.4086528718,0.0720963702,-0.1614484638,-0.4712712765,0.1547385156,0.324437052,0.0784328952,-0.2918297052,0.2315367162,-0.2830388844,0.1289331019,-0.104005374,0.2516618371,0.2692824006,-0.2087137401,-0.0591599233,-0.1967329979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2919","title":"Unwanted progress bars when accessing examples","comments":"doing a patch release now :)","body":"When accessing examples from a dataset formatted for pytorch, some progress bars appear when accessing examples:\r\n```python\r\nIn [1]: import datasets as ds \r\n\r\nIn [2]: d = ds.Dataset.from_dict({\"a\": [0, 1, 2]}).with_format(\"torch\") \r\n\r\nIn [3]: d[0] \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 3172.70it\/s]\r\nOut[3]: {'a': tensor(0)}\r\n```\r\n\r\nThis is because the pytorch formatter calls `map_nested` that uses progress bars\r\n\r\ncc @sgugger ","comment_length":6,"text":"Unwanted progress bars when accessing examples\nWhen accessing examples from a dataset formatted for pytorch, some progress bars appear when accessing examples:\r\n```python\r\nIn [1]: import datasets as ds \r\n\r\nIn [2]: d = ds.Dataset.from_dict({\"a\": [0, 1, 2]}).with_format(\"torch\") \r\n\r\nIn [3]: d[0] \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 3172.70it\/s]\r\nOut[3]: {'a': tensor(0)}\r\n```\r\n\r\nThis is because the pytorch formatter calls `map_nested` that uses progress bars\r\n\r\ncc @sgugger \ndoing a patch release now :)","embeddings":[-0.0175480675,-0.2450868934,0.0048732767,0.10296496,0.1993564218,0.0186564457,0.5456594825,0.2686175108,-0.3862209618,0.233160913,0.1730663329,0.4301009178,-0.0968903452,0.163059175,0.0540391617,-0.2791727781,-0.0362301841,0.0565355495,0.0279643871,0.077860862,-0.0710608438,-0.1534580141,-0.1631929129,0.2491238117,-0.4254311323,-0.1769718528,0.0204986315,-0.2497630417,0.2259828746,-0.6468563676,0.3624524474,0.017228175,0.0925377831,0.4540927112,-0.0001153858,0.1453293562,0.4285333753,0.2160392404,-0.2830248773,-0.1382309198,0.3537838757,-0.5126107335,0.3308672309,-0.3774378598,-0.0873389021,-0.5281800628,-0.2761972547,-0.4125932157,0.2876147032,0.0940865576,0.219826296,0.4712334275,-0.2902378142,-0.0055018864,0.5578870773,0.1376104206,-0.3271023929,0.1375557929,0.5464651585,0.0474077016,-0.2612787783,0.6541268229,-0.1014311612,0.3388422132,0.1663392037,-0.0622609183,0.0632258281,-0.1512287408,-0.1227654591,0.3897278607,0.1704966575,-0.0895446837,-0.2203477472,-0.810939312,-0.0621742755,-0.0982443467,-0.1076176837,0.1732621491,0.0438141599,0.0303935967,-0.5963996649,0.0290020015,-0.27146402,0.0937305689,0.0468122512,-0.1223285422,-0.0942170694,0.0973090604,-0.0284371544,0.083435297,0.1243030503,-0.1994698942,0.0932707414,0.0010196987,-0.046441935,0.1807018369,0.1487292498,-0.170697704,-0.1335336268,0.1875990182,-0.0521841459,0.1841346174,-0.0532620661,0.1660489142,-0.0838754028,-0.0406403653,0.1388794035,-0.2265987098,0.3347287774,0.0163198523,0.2717235684,0.1755359918,-0.037750639,-0.2225538939,0.1156542897,0.1093147323,0.1090649813,0.0211095344,-0.1003607139,0.2777105868,-0.3311299682,0.0219744816,0.2182542384,0.2214021385,-0.1495082378,-0.2717446387,0.1687204242,0.2794264853,-0.0600529797,0.2132000178,-0.0743473321,-0.1063092723,-0.4007383883,-0.1235286221,0.2540350258,0.1500249505,0.2033672929,0.1358764023,0.1028876156,-0.0182450209,0.3721589446,0.0875214264,0.4961630404,0.1029135436,-0.2043189853,0.2071680725,0.075304538,0.4455814064,-0.0290748477,0.271618098,-0.0666681826,-0.3141361773,-0.5112269521,0.1377267092,-0.4437701702,0.2145982981,-0.0623717569,-0.0707329661,0.3211912811,-0.0310045853,-0.0121499458,-0.3121537268,-0.247789517,0.0666654259,0.2291434705,0.2016990036,-0.1900742948,-0.2416289151,0.2099226117,0.0013152339,0.0640754029,0.1185466424,-0.23551175,0.3596096337,-0.2419328094,-0.026544407,0.1872638762,-0.1814072579,-0.3709934652,-0.1421762258,0.0234307237,0.3529036641,0.0315213762,-0.1182669997,0.3322122693,0.0806723833,-0.0189346746,0.1771131754,0.2529223263,-0.0826835856,-0.2729922533,0.0518658161,0.3036836088,0.2500388026,-0.0991399065,-0.1353184134,-0.2295820117,-0.2512821555,0.1488023102,-0.2186713666,0.1137110293,-0.0783538818,-0.0333985202,-0.1114408895,0.128908366,-0.200934723,-0.1400860697,0.334146589,-0.1916212291,-0.0585839003,-0.1843481064,-0.1486791074,-0.2163711041,-0.0423979238,-0.2948601842,-0.0875450671,0.0705130398,-0.1551547796,-0.0718744695,-0.0183241945,0.0944948941,-0.1325680614,-0.3775212765,-0.0472109951,0.0076890243,0.3263398409,0.1543031484,-0.3600919843,-0.1122498885,0.4872063696,0.0123988334,-0.0924425796,-0.1616445929,0.0323477574,0.1806732118,0.020979777,-0.418775171,0.1870616972,0.3267300427,0.1015354991,-0.0053531798,0.2033281922,-0.110525474,-0.0142406756,-0.2096013576,0.275554955,0.3913537562,0.3871296048,-0.0188226514,0.1576325446,0.2299357653,0.2287838459,-0.19063811,-0.3708530664,0.3794209957,0.1164492592,0.2275259644,-0.1128709093,-0.056149736,0.2054813057,0.2419088632,-0.0129112294,0.0710802451,0.1942463815,-0.3472250402,-0.0823546052,0.2821888328,-0.0113203805,0.0383519419,0.0980123356,0.1815937907,0.1735030264,-0.0646654963,-0.2774474919,-0.002367727,0.252974838,0.1037782133,0.0188018717,0.1014887169,0.0363334157,-0.0996529385,-0.2078756839,-0.2531711757,0.0459414124,-0.4298397005,0.2782195807,-0.3938655257,-0.2349227965,-0.1121626124,-0.1845910549,0.0518490709,-0.2728966475,0.1209051535,0.5843741298,-0.1538914591,0.3002850115,0.3276317716,0.1584009379,0.0250879861,0.1818053126,-0.0510158017,-0.510720849,-0.2618277967,0.0802433044,-0.0849289596,-0.1672920287,0.5056151152,-0.1877110004,0.4327579141,-0.3314349949,-0.1271976531,0.1429183781,-0.0935863927,0.0404994525,-0.0486703366,0.1458623558,-0.0911693275,0.233792603,0.160188511,-0.1380708367,-0.0214611273,0.16621387,-0.2902114093,-0.0336513706,-0.2353742123,0.0583931804,-0.3945191801,-0.1011367291,0.0673801824,-0.1514690369,0.1948434263,-0.0160474218,-0.3517791629,0.4189587235,0.1875047535,-0.1340169311,-0.1357982308,-0.3630971313,0.2848750055,-0.41330266,-0.2727486789,-0.0870164037,-0.2269277126,-0.2414059043,0.2249186188,-0.4970188141,-0.4690299034,-0.2819567025,0.2401717752,0.0068959016,-0.0885310322,0.2275055051,0.2388396263,-0.0490171686,-0.1502595395,-0.1286485642,0.2684911489,-0.2023165226,-0.0526690036,-0.2280971706,-0.1113249883,0.1931582093,0.4944924414,0.1924008131,-0.2912912369,0.1187348738,0.0306330826,0.1123298332,-0.1802462488,-0.1826853007,-0.1176944822,-0.4110670388,0.0464733355,-0.0148358615,0.0150283165,-0.0615319833,0.2057303935,-0.0157856271,-0.1149998009,-0.292308867,-0.1367049962,-0.5465351939,0.2146653235,-0.0885391533,0.3431994319,-0.1929283291,-0.1536388397,-0.1729557216,-0.1027353704,0.3973326683,0.0111482572,-0.1668602377,-0.256231606,-0.0367813185,0.3502275944,0.4951102138,0.0122325225,-0.0494005047,-0.0513418317,0.4644909203,0.4761361778,0.2832567096,0.1001850367,-0.1279858649,0.2342715114,-0.2401283532,-0.7125119567,-0.41146034,-0.1997616738,0.4119268656,0.2129512131,0.4722346961,0.0457137078,-0.086379692,0.0880065113,0.4283269346,-0.1754733473,-0.1151453927,-0.164950341,-0.3428955376,-0.3676280975,0.2961567342,0.1100679711,-0.0125856651,-0.2604800761,-0.2776361108,-0.0639647469,0.0514944792,-0.0319752358,0.3141577542,0.229007259,0.0247712489,0.1257308871,0.4538356066,0.4744880199,0.2547186613,0.0707441568,0.1482186466,-0.048177246,0.2233390659,-0.4182990193,-0.0396085083,0.2151462138,-0.0113113131,0.300827384,-0.2532792091,0.0221573282,-0.2611340284,0.0794604346,0.2527319491,-0.0078818072,-0.1881323904,-0.316670537,0.2333788425,-0.0685120448,-0.178301841,0.2720800936,-0.3529309928,-0.1420414448,0.1504247189,0.1549986601,0.6575912833,-0.0286711324,0.1168500483,0.511208117,-0.2467370182,0.2707988918,0.2357586771,0.1782183349,-0.2843172848,-0.2260917276,0.077235803,-0.446272552,0.0647938997,-0.0009366414,-0.0220275056,-0.131005004,-0.5273835063,0.2063754499,0.0190239213,-0.0590235889,0.3956880569,-0.3413526118,-0.2414980084,0.1582078487,-0.1195397153,0.3477330506,-0.153694585,0.105606243,0.0029755994,0.0559373796,-0.2887510955,0.0984202698,0.1579602063,0.3901692331,0.0510993376,-0.7411983609,-0.0375015289,0.1320708096,0.3103530407,-0.0464583524,-0.120931454,0.1860230863,0.0712058544,0.1173128709,0.2080661058,-0.0445344783,0.3048805892,-0.2177551091,-0.2133199573,0.289378047,0.0879581645,-0.2650895715,-0.0903161094,0.1210575998,0.0623679422,-0.0285778753,0.146940738,-0.1023302898,0.1881256551,-0.179275915,0.1194247305,0.098932758,-0.2934565246,0.3730723858,-0.1679872125,-0.2851409614,0.023308998,0.0684312284,0.0924657211,-0.0535454601,0.5884126425,0.2965933979,-0.3969217241,-0.0453655347,0.2278076261,0.2701043487,-0.2893387973,0.1174756289,0.2835693359,0.1774771065,-0.413019985,-0.0458238162,-0.0454670712,-0.026506193,-0.2763679922,-0.3110716641,-0.3819492757,-0.1838888377,-0.037768133,0.3098086715,0.0986221954,0.2046764046,-0.1496843994,-0.1149211973,-0.2811182737,0.2810724676,-0.1052022502,0.2823551893,-0.0097785192,-0.2017995566,-0.0716521591,-0.3540430963,-0.0920337066,0.0676924735,-0.2509935796,-0.2414507866,-0.2945031524,0.0720853955,-0.1283766329,-0.5431334972,0.066728361,0.1387889385,0.1103348732,-0.0959893838,0.4568045437,0.3058486581,0.0772642195,0.3269970715,-0.2034392655,0.2739566267,-0.0409143828,-0.020709984,-0.1116553992,0.3173391223,0.1663935035,-0.0878654346,0.2291099429,0.1036329418,0.1654630154,-0.0290552266,0.0285506975,0.3309161067,0.1152022108,-0.1805236489,0.2272562385,-0.1822111905,0.3503676355,0.0519738756,-0.2619753778,-0.2643000782,0.2569488585,0.1530782729,-0.0500352234,-0.2258733511,0.4774631262,0.0756995752,0.0340617336,0.1253063083,0.2247715294,0.5043908358,-0.0475612022,0.0738995597,-0.081570439,0.0508443005,0.3201390505,0.2587703466,-0.0798705667,-0.0055938778,0.0351336226,0.0449205413,0.0840099305,-0.0845154375,0.3153921962,0.6276469231,0.5006055236,-0.0922196507,0.0945133418,-0.263574332,-0.4391740263,-0.3639207184,-0.1227432266,0.0407780483,-0.0924314335,0.2919994593,-0.2905572057,-0.4293654561,-0.3388067782,-0.0357804783,-0.1445909292,-0.070831351,-0.034298256,0.0904835537,0.0901739672,-0.0436631069,-0.1581738442,0.1526149809,0.0199174788,-0.1052327827,-0.0309135877,-0.0641777217,0.0459759124,-0.0262360517,0.3488769233,0.1154232547,-0.0487071984,0.1308671981,0.0627249703,-0.1025522426,0.1025527045,0.1545532346,0.2601504624,-0.3949838877,-0.1687651724,-0.1504912376,0.1469617635,-0.0068152202,0.3295758367,-0.0582019277,-0.4498440027,0.2931940854,0.0710852221,-0.1421114355,0.6278657317,0.0279215667,0.1311353743,-0.2680278718,0.3336524963,-0.2577918768,-0.1453121305,0.0927531719,-0.3478544056,0.0340452418,-0.1008718088,0.8285409808,-0.0055257455,0.0478106216,-0.1501996219,0.0873420835,-0.1911148876,0.6613516212,0.1349851936,0.5720160007,-0.0729165375,0.0926222354,-0.3475141227,0.1489682943,-0.1697594225,0.0487481356,0.2350684851,-0.1679376215,-0.3383406103,0.4007802308,0.5639624,-0.2407636493,-0.249806881,-0.3299007118,-0.3072220385,0.0551428385,0.0805551186,-0.0869708359,-0.0475652367,-0.1825208068,0.0548046343,-0.0901909247,0.1151650399,0.0051526413,-0.1455165148,-0.0173772313,-0.0279609356,0.4075471163,0.2661443949,0.1931025535,0.2716137767,0.0248533059,0.0354542583,-0.0741970539,0.0967927873,0.4841645956,0.0150432829,0.329911083,-0.0158772171,0.0391109847,-0.1803999096,0.4807221889,0.3238476217,0.0178111568,-0.1421049386,0.0984386057,-0.2610158622,0.2574439645,-0.016893791,0.0306532588,-0.0154680144,0.4048753083,-0.0077654449,-0.1700899154,0.5261824131,-0.4042277336,-0.3670017421,-0.0829792991,0.2184514552,0.1859294027,-0.0921535343,-0.3823102415,-0.2459062785,0.2568610907,-0.1912846565,-0.1312142015,0.1051476598,0.0230771396,-0.1514546424,-0.0044824672,0.5724635124,0.0191494972,0.0986652225,-0.0518430099,-0.3697446585]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2918","title":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming","comments":"Hi @SBrandeis, thanks for reporting! ^^\r\n\r\nI think this is an issue with `fsspec`: https:\/\/github.com\/intake\/filesystem_spec\/issues\/389\r\n\r\nI will ask them if they are planning to fix it...","body":"## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n","comment_length":26,"text":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming\n## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n\nHi @SBrandeis, thanks for reporting! ^^\r\n\r\nI think this is an issue with `fsspec`: https:\/\/github.com\/intake\/filesystem_spec\/issues\/389\r\n\r\nI will ask them if they are planning to fix it...","embeddings":[-0.3864652812,-0.2204957455,0.0894813165,0.4387415349,0.2157086134,0.1229229122,-0.0369101539,0.3093882799,0.2422775328,0.0955026373,-0.2264311761,0.4261536002,-0.0641756877,0.3035327792,-0.0094654057,-0.2584786713,-0.0045060921,0.2337655127,-0.1049250886,0.1268760264,-0.0005556306,0.1941259056,-0.2469931841,-0.0304233823,0.0522285365,-0.1081046611,0.0124065261,0.2805253863,-0.059583243,-0.4462363422,0.1259049624,-0.0615482517,0.4121141732,0.3375741839,-0.0001145032,0.2734483778,0.4721427262,-0.1227841005,-0.4853440225,-0.2121149004,-0.218665719,0.0961271524,0.0428345352,-0.1566341817,-0.0091345869,0.0711133331,-0.0400046185,-0.7192984819,0.3178500533,0.3456359804,0.1778500229,0.0426006429,-0.0489487238,0.0268223863,0.0688386783,-0.2986646891,-0.0920066237,0.1218173951,0.2367273569,0.2287404835,-0.359573245,0.3845903277,-0.0658884048,0.0601840951,0.0944148228,0.0967238322,-0.1749501675,-0.4655051231,-0.0073944977,0.1884141415,0.6776787043,-0.2555166781,-0.4161978662,-0.2319881767,-0.0021713183,-0.6144573689,0.1465310603,0.1489273459,-0.4673651159,0.1280597448,0.2459630817,-0.0087530883,-0.327994585,-0.0458728224,-0.246799171,0.3882223368,-0.0767857283,-0.0237270929,0.0272943433,-0.0866289064,0.4627741277,-0.2110346556,-0.28459391,-0.0639790371,-0.4397552311,-0.0420655161,-0.0765123144,-0.094109349,0.1681466401,0.215389967,0.2953203917,0.0882523879,0.1543645263,0.0926442817,0.3555265665,0.3608669043,0.0598316565,0.1853774786,0.1311810017,0.0994313955,0.1366237104,-0.1818924248,-0.146488592,-0.0372809209,0.0281632524,-0.0131140957,0.262858659,-0.1912374645,-0.4138476253,0.0297920927,-0.3057210743,-0.0794672072,0.0055905147,0.2817168832,0.0958337784,0.4187803268,0.0194585584,0.1178673506,-0.2396837324,-0.4573938251,-0.0878112167,-0.1247182265,-0.0010082311,0.1233467832,0.0153353186,-0.1886832416,0.1515737325,-0.0465371571,0.1621010602,-0.0012185768,-0.0208561029,-0.1100598425,0.0216292907,0.2049841285,0.2180556655,0.1640353203,0.1795151085,-0.1997850239,-0.130682379,0.0886226073,-0.1416330338,-0.1671275645,-0.1788374335,0.2166469097,-0.0094024939,-0.3506264389,-0.0937892571,0.2814688385,0.1409590542,-0.2037903368,-0.1410701424,-0.0391465798,-0.1241319329,-0.0865156949,0.1792775393,0.5796180964,-0.1855915189,-0.1380216628,-0.1521381587,0.0008939481,0.5833057165,0.3344025612,-0.0607914478,0.0897968188,-0.170722425,0.1476544142,0.5017140508,-0.1436584145,-0.5796036124,0.5436661243,-0.0396082066,0.5637281537,0.2707032561,0.0225366186,0.1742344499,0.0431196988,0.313529104,0.3299024701,-0.1484592706,0.0074822325,-0.3334669471,-0.0399836674,0.4416950643,0.1912285686,0.0674290732,0.2502861321,-0.0288742278,-0.0626677349,0.3689779639,-0.0752563924,0.0763451308,0.0410080664,0.1468882412,0.0120370993,0.1379382461,-0.2090770155,-0.310682714,0.2972311378,0.0749307722,0.0127100674,-0.3174338341,0.2031723857,-0.1906455308,0.0017469627,-0.2628925443,-0.2590405643,0.1243173257,0.3170801699,0.0756207108,0.0505895875,-0.3084766865,0.7747215033,-0.078255251,-0.0131349247,-0.4549141228,0.2024929523,-0.0707768798,-0.2014573216,0.0164402872,-0.1265656799,0.0672556311,-0.0161919687,-0.2579914331,0.3313200772,-0.2366138101,0.2640713155,-0.1954375803,-0.1731976271,0.2897448838,-0.3644788265,-0.0866843313,0.2706028223,0.1075036898,0.125209704,-0.0176983867,0.1480839401,0.2070915401,0.2361335903,0.0434485786,0.1143487841,0.1961675137,-0.0816748515,-0.0686837137,-0.0850090533,0.3240485787,-0.3548789024,0.0467500761,-0.1989230514,-0.2722995877,0.0823631957,0.3586791456,-0.1403829306,-0.0492737889,0.3303245604,-0.2885322273,-0.1639078707,0.3145605624,0.1940099597,0.4723360538,0.0815219656,0.2573353648,-0.0455882996,-0.1886483878,-0.0735060051,0.3051697612,0.1374839693,0.0314638466,0.2464114279,0.0300063528,0.1635747999,-0.3229160309,-0.3497166634,-0.0566318817,0.1123416647,-0.3201874793,0.306302011,-0.38826859,-0.4651744664,0.0463509336,0.1560085416,0.0689049438,-0.3235587776,-0.1817550063,0.2325055152,-0.0038823313,0.1636838913,-0.2682063282,0.0895485729,0.2671770155,-0.0134687945,-0.3418698907,-0.1808474958,-0.2729629874,0.0591977872,0.1749951094,-0.4228037596,0.1769195795,0.0074260007,-0.1160532683,-0.1401735991,-0.076423429,0.3197818398,-0.1554995328,0.1568958163,-0.2774865627,0.1493911892,0.2213633955,-0.2503491044,0.1219482496,0.0908335969,-0.1187844425,0.2801248431,0.0940897167,-0.0241610892,-0.0305832736,-0.3010807931,-0.3024666011,-0.6177494526,0.0549683981,0.0454257689,0.2135515809,0.1643050015,0.0170347784,0.0968246013,0.3292069137,0.0991601422,-0.114212282,-0.4199434817,0.3747448325,-0.1670146883,-0.2868409157,-0.3248367012,-0.0681597292,0.1972901672,0.2589270473,-0.5295243263,0.2950162292,-0.4081035852,0.0848925635,-0.3834516108,-0.0974560827,-0.0261515174,0.0762885734,-0.0349893123,-0.1878576428,0.0707459524,-0.1169513538,0.0990076959,0.2066107839,0.1086307243,0.3698712885,0.1008493528,0.3327504098,0.4812112153,-0.0195496976,0.3407551944,0.082891129,0.1538007557,-0.1004334986,-0.0627885908,-0.0955621898,-0.2280800343,-0.0261197276,0.0981990993,-0.0962127745,-0.5762300491,-0.2840516269,0.0549149849,-0.0692367032,-0.2383105308,0.0528838076,-0.511775136,0.0420482606,-0.1872570366,0.1583397985,0.0807970166,0.0520012863,0.0860457644,0.1772524714,0.1261645257,0.1721196771,-0.185021311,-0.0639052019,-0.4881900847,0.1825860441,0.1051262617,0.4215749502,-0.1678105742,-0.1327165961,0.0791147202,-0.0652071014,0.3487038612,-0.2754168212,0.003567247,0.2673943937,0.1421353668,-0.1415521204,0.0917675421,-0.3209457397,0.1368288398,0.235954389,0.0974201635,-0.1718239039,0.0333400816,-0.0884195939,-0.0724961758,-0.1253028065,-0.0163991563,-0.1704963744,-0.2492375523,-0.152009964,-0.134242788,0.2032724619,0.1256832927,-0.0124605224,-0.0567925759,-0.0460816547,-0.0396244489,0.2418593764,0.2213254869,0.1399096996,-0.072280176,0.2290114909,0.0671992972,0.2844450474,0.2292162329,0.836432457,0.1181946918,-0.5624260902,0.2350820601,0.2339563966,0.1660992503,0.1305397898,-0.148232609,-0.0085166721,-0.0403849594,0.2698210478,-0.3024733663,0.2146548331,0.2112327665,-0.0507208817,-0.5149303675,-0.4357535839,0.5883390903,-0.039368242,-0.0761907846,0.7872929573,-0.1665049642,-0.3171558082,0.4775415361,0.0234355517,0.7585879564,0.0943952724,0.2168687731,0.2407481223,-0.3207278252,0.5003329515,-0.2320687622,0.3346687257,-0.1206100509,-0.3350296021,-0.0831092745,-0.1041097715,0.1709026545,-0.147603929,-0.6302542686,0.1709609181,-0.0596441738,-0.2528051734,0.3353525698,0.2742269933,-0.6082689762,-0.3062928915,-0.0675497502,0.1388825476,-0.2213602215,0.3239610493,0.1382443607,-0.2202336341,-0.0515731983,-0.0843103677,-0.4702361524,0.4758630693,-0.2765566409,0.1624139398,-0.2390307039,-0.4760982394,0.1513123959,0.2664022446,-0.1421327889,-0.2257857472,-0.1582512558,0.2350440472,0.1860539466,0.128023684,0.0993071273,0.1514580995,0.3883590102,0.0532772318,-0.0168129094,0.1884423494,-0.0089801019,-0.1606287062,-0.0637812093,0.0554574467,0.3183139861,-0.2681806087,0.2368578017,-0.1058510691,0.0429251082,-0.1842529923,0.0869722143,-0.0188314319,-0.1740541756,-0.0738878325,-0.08129704,-0.2196760625,-0.0847442299,0.3771200478,0.1454578936,0.3179713786,0.4994189143,0.0281169843,-0.1778806895,-0.1036716029,-0.0551506244,-0.2931399345,-0.252612263,0.0693599209,-0.2414764762,-0.1610280573,-0.2706694007,0.0038265749,0.0448373631,-0.103554666,-0.1137386858,-0.2993660271,-0.1906369925,-0.0266146194,-0.1242436618,0.2199304104,-0.1722201109,0.0222039819,0.0031197246,0.1290859282,-0.270683825,-0.1228595003,-0.0753692091,0.0231177211,0.2672106326,0.0292674527,0.2592337132,0.1096690521,0.1476223618,0.3139047623,-0.0276747048,-0.1609539241,-0.1787950993,0.138489455,0.0165312476,-0.348051548,-0.1548943967,-0.05662065,0.0126910359,-0.125136137,0.3503144085,0.181230098,0.0380968302,0.2706013322,0.1240136698,0.0563360825,-0.2326476574,0.1542367488,-0.2642730176,0.2401308864,0.0811118633,0.0928405821,-0.0755398721,-0.0118692657,-0.3646134436,0.0622693747,0.1017946154,-0.0280170571,0.5909268856,-0.4241805971,0.1720450372,-0.0497748405,0.0660299286,0.4045228958,-0.2296188772,-0.2089618146,0.0850852206,0.2000514269,-0.0432033539,-0.2126210332,-0.4010480046,-0.1016407683,0.0112383589,0.1459593624,-0.1068896204,-0.2721833587,0.0999634638,-0.2103084773,0.4859834313,0.3358429968,0.2670664489,0.4836392403,-0.1811283231,0.0228473805,-0.2260247022,0.0204593837,0.1999064386,0.2970430553,0.0035512666,-0.0078696832,-0.1613658667,0.2178194076,-0.1859746873,-0.482195437,-0.0797479451,0.1024470255,-0.1647549272,0.3901939094,0.1489096284,0.1521606892,-0.2779354751,-0.0519221425,-0.0458415113,-0.0661262944,-0.0142799746,-0.096080415,0.1068764329,-0.0128562739,-0.3279225528,0.0976088271,0.0123996465,0.0633685365,-0.0198641457,0.2808055282,-0.2884631157,-0.1185014918,-0.0822276548,0.1043173745,0.2999251485,-0.3544151187,0.4440693855,0.3661904037,-0.1568607092,0.0282512195,0.0700021908,0.4816074371,0.2853379846,-0.1766507179,0.0946500823,0.3374164999,-0.0436370596,-0.0402214266,0.0442541391,0.2830949128,0.1793196946,0.4566897452,0.0887393281,-0.1485419422,0.1045718119,0.0814351737,0.2538890839,-0.2799384296,0.4623198807,-0.1195058972,0.0366923138,0.0272421893,-0.3365718424,-0.5624158978,-0.0282046329,0.0348961949,-0.2285001576,0.0745235756,-0.0081122555,0.0628783256,0.1139546707,0.4926685393,0.3248410225,-0.0188453235,-0.2245095521,-0.1508785933,-0.235583514,0.2332020253,-0.0674795657,-0.1231484041,0.1752853394,0.2219228297,-0.0048415069,0.1680491567,-0.1584448814,0.0204550214,-0.2672839463,0.1405724436,-0.3995800018,-0.3515672684,0.1919635981,0.1735794991,-0.0483116023,-0.576480031,0.1881390065,-0.1502387971,0.1608176678,0.0926132202,-0.241848141,0.101145193,-0.2756578326,0.5297338963,0.2331105173,0.3792004883,-0.184295252,-0.4119567871,-0.0992468074,-0.3762306273,-0.2749142647,-0.0155847622,0.1083334535,0.5696322322,-0.1493341029,-0.2373841256,-0.3574708402,0.3262623847,0.0608894229,-0.2656728923,-0.2735387981,0.1733704358,-0.1194598079,0.2131261081,0.3050509691,0.380345583,-0.0287131406,0.1639950573,-0.2235154659,-0.4333209693,0.5773110986,-0.3461063206,-0.2659671307,0.2418412119,0.0638139024,0.3264979422,0.2410199344,-0.4596662819,0.0893656164,0.3556547463,-0.2354129702,-0.1511856616,0.1421462744,0.2259705812,-0.0607201755,-0.0850398242,0.45531106,0.2423961759,-0.249204129,0.1750813872,-0.1786384881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2918","title":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming","comments":"Code to reproduce the bug: `ClientPayloadError: 400, message='Can not decode content-encoding: gzip'`\r\n```python\r\nIn [1]: import fsspec\r\n\r\nIn [2]: import json\r\n\r\nIn [3]: with fsspec.open('https:\/\/raw.githubusercontent.com\/allenai\/scitldr\/master\/SciTLDR-Data\/SciTLDR-FullText\/test.jsonl', encoding=\"utf-8\") as f:\r\n ...: for row in f:\r\n ...: data = json.loads(row)\r\n ...:\r\n---------------------------------------------------------------------------\r\nClientPayloadError Traceback (most recent call last)\r\n```","body":"## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n","comment_length":46,"text":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming\n## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n\nCode to reproduce the bug: `ClientPayloadError: 400, message='Can not decode content-encoding: gzip'`\r\n```python\r\nIn [1]: import fsspec\r\n\r\nIn [2]: import json\r\n\r\nIn [3]: with fsspec.open('https:\/\/raw.githubusercontent.com\/allenai\/scitldr\/master\/SciTLDR-Data\/SciTLDR-FullText\/test.jsonl', encoding=\"utf-8\") as f:\r\n ...: for row in f:\r\n ...: data = json.loads(row)\r\n ...:\r\n---------------------------------------------------------------------------\r\nClientPayloadError Traceback (most recent call last)\r\n```","embeddings":[-0.3864652812,-0.2204957455,0.0894813165,0.4387415349,0.2157086134,0.1229229122,-0.0369101539,0.3093882799,0.2422775328,0.0955026373,-0.2264311761,0.4261536002,-0.0641756877,0.3035327792,-0.0094654057,-0.2584786713,-0.0045060921,0.2337655127,-0.1049250886,0.1268760264,-0.0005556306,0.1941259056,-0.2469931841,-0.0304233823,0.0522285365,-0.1081046611,0.0124065261,0.2805253863,-0.059583243,-0.4462363422,0.1259049624,-0.0615482517,0.4121141732,0.3375741839,-0.0001145032,0.2734483778,0.4721427262,-0.1227841005,-0.4853440225,-0.2121149004,-0.218665719,0.0961271524,0.0428345352,-0.1566341817,-0.0091345869,0.0711133331,-0.0400046185,-0.7192984819,0.3178500533,0.3456359804,0.1778500229,0.0426006429,-0.0489487238,0.0268223863,0.0688386783,-0.2986646891,-0.0920066237,0.1218173951,0.2367273569,0.2287404835,-0.359573245,0.3845903277,-0.0658884048,0.0601840951,0.0944148228,0.0967238322,-0.1749501675,-0.4655051231,-0.0073944977,0.1884141415,0.6776787043,-0.2555166781,-0.4161978662,-0.2319881767,-0.0021713183,-0.6144573689,0.1465310603,0.1489273459,-0.4673651159,0.1280597448,0.2459630817,-0.0087530883,-0.327994585,-0.0458728224,-0.246799171,0.3882223368,-0.0767857283,-0.0237270929,0.0272943433,-0.0866289064,0.4627741277,-0.2110346556,-0.28459391,-0.0639790371,-0.4397552311,-0.0420655161,-0.0765123144,-0.094109349,0.1681466401,0.215389967,0.2953203917,0.0882523879,0.1543645263,0.0926442817,0.3555265665,0.3608669043,0.0598316565,0.1853774786,0.1311810017,0.0994313955,0.1366237104,-0.1818924248,-0.146488592,-0.0372809209,0.0281632524,-0.0131140957,0.262858659,-0.1912374645,-0.4138476253,0.0297920927,-0.3057210743,-0.0794672072,0.0055905147,0.2817168832,0.0958337784,0.4187803268,0.0194585584,0.1178673506,-0.2396837324,-0.4573938251,-0.0878112167,-0.1247182265,-0.0010082311,0.1233467832,0.0153353186,-0.1886832416,0.1515737325,-0.0465371571,0.1621010602,-0.0012185768,-0.0208561029,-0.1100598425,0.0216292907,0.2049841285,0.2180556655,0.1640353203,0.1795151085,-0.1997850239,-0.130682379,0.0886226073,-0.1416330338,-0.1671275645,-0.1788374335,0.2166469097,-0.0094024939,-0.3506264389,-0.0937892571,0.2814688385,0.1409590542,-0.2037903368,-0.1410701424,-0.0391465798,-0.1241319329,-0.0865156949,0.1792775393,0.5796180964,-0.1855915189,-0.1380216628,-0.1521381587,0.0008939481,0.5833057165,0.3344025612,-0.0607914478,0.0897968188,-0.170722425,0.1476544142,0.5017140508,-0.1436584145,-0.5796036124,0.5436661243,-0.0396082066,0.5637281537,0.2707032561,0.0225366186,0.1742344499,0.0431196988,0.313529104,0.3299024701,-0.1484592706,0.0074822325,-0.3334669471,-0.0399836674,0.4416950643,0.1912285686,0.0674290732,0.2502861321,-0.0288742278,-0.0626677349,0.3689779639,-0.0752563924,0.0763451308,0.0410080664,0.1468882412,0.0120370993,0.1379382461,-0.2090770155,-0.310682714,0.2972311378,0.0749307722,0.0127100674,-0.3174338341,0.2031723857,-0.1906455308,0.0017469627,-0.2628925443,-0.2590405643,0.1243173257,0.3170801699,0.0756207108,0.0505895875,-0.3084766865,0.7747215033,-0.078255251,-0.0131349247,-0.4549141228,0.2024929523,-0.0707768798,-0.2014573216,0.0164402872,-0.1265656799,0.0672556311,-0.0161919687,-0.2579914331,0.3313200772,-0.2366138101,0.2640713155,-0.1954375803,-0.1731976271,0.2897448838,-0.3644788265,-0.0866843313,0.2706028223,0.1075036898,0.125209704,-0.0176983867,0.1480839401,0.2070915401,0.2361335903,0.0434485786,0.1143487841,0.1961675137,-0.0816748515,-0.0686837137,-0.0850090533,0.3240485787,-0.3548789024,0.0467500761,-0.1989230514,-0.2722995877,0.0823631957,0.3586791456,-0.1403829306,-0.0492737889,0.3303245604,-0.2885322273,-0.1639078707,0.3145605624,0.1940099597,0.4723360538,0.0815219656,0.2573353648,-0.0455882996,-0.1886483878,-0.0735060051,0.3051697612,0.1374839693,0.0314638466,0.2464114279,0.0300063528,0.1635747999,-0.3229160309,-0.3497166634,-0.0566318817,0.1123416647,-0.3201874793,0.306302011,-0.38826859,-0.4651744664,0.0463509336,0.1560085416,0.0689049438,-0.3235587776,-0.1817550063,0.2325055152,-0.0038823313,0.1636838913,-0.2682063282,0.0895485729,0.2671770155,-0.0134687945,-0.3418698907,-0.1808474958,-0.2729629874,0.0591977872,0.1749951094,-0.4228037596,0.1769195795,0.0074260007,-0.1160532683,-0.1401735991,-0.076423429,0.3197818398,-0.1554995328,0.1568958163,-0.2774865627,0.1493911892,0.2213633955,-0.2503491044,0.1219482496,0.0908335969,-0.1187844425,0.2801248431,0.0940897167,-0.0241610892,-0.0305832736,-0.3010807931,-0.3024666011,-0.6177494526,0.0549683981,0.0454257689,0.2135515809,0.1643050015,0.0170347784,0.0968246013,0.3292069137,0.0991601422,-0.114212282,-0.4199434817,0.3747448325,-0.1670146883,-0.2868409157,-0.3248367012,-0.0681597292,0.1972901672,0.2589270473,-0.5295243263,0.2950162292,-0.4081035852,0.0848925635,-0.3834516108,-0.0974560827,-0.0261515174,0.0762885734,-0.0349893123,-0.1878576428,0.0707459524,-0.1169513538,0.0990076959,0.2066107839,0.1086307243,0.3698712885,0.1008493528,0.3327504098,0.4812112153,-0.0195496976,0.3407551944,0.082891129,0.1538007557,-0.1004334986,-0.0627885908,-0.0955621898,-0.2280800343,-0.0261197276,0.0981990993,-0.0962127745,-0.5762300491,-0.2840516269,0.0549149849,-0.0692367032,-0.2383105308,0.0528838076,-0.511775136,0.0420482606,-0.1872570366,0.1583397985,0.0807970166,0.0520012863,0.0860457644,0.1772524714,0.1261645257,0.1721196771,-0.185021311,-0.0639052019,-0.4881900847,0.1825860441,0.1051262617,0.4215749502,-0.1678105742,-0.1327165961,0.0791147202,-0.0652071014,0.3487038612,-0.2754168212,0.003567247,0.2673943937,0.1421353668,-0.1415521204,0.0917675421,-0.3209457397,0.1368288398,0.235954389,0.0974201635,-0.1718239039,0.0333400816,-0.0884195939,-0.0724961758,-0.1253028065,-0.0163991563,-0.1704963744,-0.2492375523,-0.152009964,-0.134242788,0.2032724619,0.1256832927,-0.0124605224,-0.0567925759,-0.0460816547,-0.0396244489,0.2418593764,0.2213254869,0.1399096996,-0.072280176,0.2290114909,0.0671992972,0.2844450474,0.2292162329,0.836432457,0.1181946918,-0.5624260902,0.2350820601,0.2339563966,0.1660992503,0.1305397898,-0.148232609,-0.0085166721,-0.0403849594,0.2698210478,-0.3024733663,0.2146548331,0.2112327665,-0.0507208817,-0.5149303675,-0.4357535839,0.5883390903,-0.039368242,-0.0761907846,0.7872929573,-0.1665049642,-0.3171558082,0.4775415361,0.0234355517,0.7585879564,0.0943952724,0.2168687731,0.2407481223,-0.3207278252,0.5003329515,-0.2320687622,0.3346687257,-0.1206100509,-0.3350296021,-0.0831092745,-0.1041097715,0.1709026545,-0.147603929,-0.6302542686,0.1709609181,-0.0596441738,-0.2528051734,0.3353525698,0.2742269933,-0.6082689762,-0.3062928915,-0.0675497502,0.1388825476,-0.2213602215,0.3239610493,0.1382443607,-0.2202336341,-0.0515731983,-0.0843103677,-0.4702361524,0.4758630693,-0.2765566409,0.1624139398,-0.2390307039,-0.4760982394,0.1513123959,0.2664022446,-0.1421327889,-0.2257857472,-0.1582512558,0.2350440472,0.1860539466,0.128023684,0.0993071273,0.1514580995,0.3883590102,0.0532772318,-0.0168129094,0.1884423494,-0.0089801019,-0.1606287062,-0.0637812093,0.0554574467,0.3183139861,-0.2681806087,0.2368578017,-0.1058510691,0.0429251082,-0.1842529923,0.0869722143,-0.0188314319,-0.1740541756,-0.0738878325,-0.08129704,-0.2196760625,-0.0847442299,0.3771200478,0.1454578936,0.3179713786,0.4994189143,0.0281169843,-0.1778806895,-0.1036716029,-0.0551506244,-0.2931399345,-0.252612263,0.0693599209,-0.2414764762,-0.1610280573,-0.2706694007,0.0038265749,0.0448373631,-0.103554666,-0.1137386858,-0.2993660271,-0.1906369925,-0.0266146194,-0.1242436618,0.2199304104,-0.1722201109,0.0222039819,0.0031197246,0.1290859282,-0.270683825,-0.1228595003,-0.0753692091,0.0231177211,0.2672106326,0.0292674527,0.2592337132,0.1096690521,0.1476223618,0.3139047623,-0.0276747048,-0.1609539241,-0.1787950993,0.138489455,0.0165312476,-0.348051548,-0.1548943967,-0.05662065,0.0126910359,-0.125136137,0.3503144085,0.181230098,0.0380968302,0.2706013322,0.1240136698,0.0563360825,-0.2326476574,0.1542367488,-0.2642730176,0.2401308864,0.0811118633,0.0928405821,-0.0755398721,-0.0118692657,-0.3646134436,0.0622693747,0.1017946154,-0.0280170571,0.5909268856,-0.4241805971,0.1720450372,-0.0497748405,0.0660299286,0.4045228958,-0.2296188772,-0.2089618146,0.0850852206,0.2000514269,-0.0432033539,-0.2126210332,-0.4010480046,-0.1016407683,0.0112383589,0.1459593624,-0.1068896204,-0.2721833587,0.0999634638,-0.2103084773,0.4859834313,0.3358429968,0.2670664489,0.4836392403,-0.1811283231,0.0228473805,-0.2260247022,0.0204593837,0.1999064386,0.2970430553,0.0035512666,-0.0078696832,-0.1613658667,0.2178194076,-0.1859746873,-0.482195437,-0.0797479451,0.1024470255,-0.1647549272,0.3901939094,0.1489096284,0.1521606892,-0.2779354751,-0.0519221425,-0.0458415113,-0.0661262944,-0.0142799746,-0.096080415,0.1068764329,-0.0128562739,-0.3279225528,0.0976088271,0.0123996465,0.0633685365,-0.0198641457,0.2808055282,-0.2884631157,-0.1185014918,-0.0822276548,0.1043173745,0.2999251485,-0.3544151187,0.4440693855,0.3661904037,-0.1568607092,0.0282512195,0.0700021908,0.4816074371,0.2853379846,-0.1766507179,0.0946500823,0.3374164999,-0.0436370596,-0.0402214266,0.0442541391,0.2830949128,0.1793196946,0.4566897452,0.0887393281,-0.1485419422,0.1045718119,0.0814351737,0.2538890839,-0.2799384296,0.4623198807,-0.1195058972,0.0366923138,0.0272421893,-0.3365718424,-0.5624158978,-0.0282046329,0.0348961949,-0.2285001576,0.0745235756,-0.0081122555,0.0628783256,0.1139546707,0.4926685393,0.3248410225,-0.0188453235,-0.2245095521,-0.1508785933,-0.235583514,0.2332020253,-0.0674795657,-0.1231484041,0.1752853394,0.2219228297,-0.0048415069,0.1680491567,-0.1584448814,0.0204550214,-0.2672839463,0.1405724436,-0.3995800018,-0.3515672684,0.1919635981,0.1735794991,-0.0483116023,-0.576480031,0.1881390065,-0.1502387971,0.1608176678,0.0926132202,-0.241848141,0.101145193,-0.2756578326,0.5297338963,0.2331105173,0.3792004883,-0.184295252,-0.4119567871,-0.0992468074,-0.3762306273,-0.2749142647,-0.0155847622,0.1083334535,0.5696322322,-0.1493341029,-0.2373841256,-0.3574708402,0.3262623847,0.0608894229,-0.2656728923,-0.2735387981,0.1733704358,-0.1194598079,0.2131261081,0.3050509691,0.380345583,-0.0287131406,0.1639950573,-0.2235154659,-0.4333209693,0.5773110986,-0.3461063206,-0.2659671307,0.2418412119,0.0638139024,0.3264979422,0.2410199344,-0.4596662819,0.0893656164,0.3556547463,-0.2354129702,-0.1511856616,0.1421462744,0.2259705812,-0.0607201755,-0.0850398242,0.45531106,0.2423961759,-0.249204129,0.1750813872,-0.1786384881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2918","title":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming","comments":"Thanks for investigating @albertvillanova ! \ud83e\udd17 ","body":"## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n","comment_length":6,"text":"`Can not decode content-encoding: gzip` when loading `scitldr` dataset with streaming\n## Describe the bug\r\n\r\nTrying to load the `\"FullText\"` config of the `\"scitldr\"` dataset with `streaming=True` raises an error from `aiohttp`:\r\n```python\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\ncc @lhoestq \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\niter_dset = iter(\r\n load_dataset(\"scitldr\", name=\"FullText\", split=\"test\", streaming=True)\r\n)\r\n\r\nnext(iter_dset)\r\n```\r\n\r\n## Expected results\r\nReturns the first sample of the dataset\r\n\r\n## Actual results\r\nCalling `__next__` crashes with the following Traceback:\r\n\r\n```python\r\n----> 1 next(dset_iter)\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 339\r\n 340 def __iter__(self):\r\n--> 341 for key, example in self._iter():\r\n 342 if self.features:\r\n 343 # we encode the example for ClassLabel feature types for example\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in _iter(self)\r\n 336 else:\r\n 337 ex_iterable = self._ex_iterable\r\n--> 338 yield from ex_iterable\r\n 339\r\n 340 def __iter__(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\datasets\\iterable_dataset.py in __iter__(self)\r\n 76\r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80\r\n\r\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\scitldr\\72d6e2195786c57e1d343066fb2cc4f93ea39c5e381e53e6ae7c44bbfd1f05ef\\scitldr.py in _generate_examples(self, filepath, split)\r\n 162\r\n 163 with open(filepath, encoding=\"utf-8\") as f:\r\n--> 164 for id_, row in enumerate(f):\r\n 165 data = json.loads(row)\r\n 166 if self.config.name == \"AIC\":\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in read(self, length)\r\n 496 else:\r\n 497 length = min(self.size - self.loc, length)\r\n--> 498 return super().read(length)\r\n 499\r\n 500 async def async_fetch_all(self):\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\spec.py in read(self, length)\r\n 1481 # don't even bother calling fetch\r\n 1482 return b\"\"\r\n-> 1483 out = self.cache._fetch(self.loc, self.loc + length)\r\n 1484 self.loc += len(out)\r\n 1485 return out\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\caching.py in _fetch(self, start, end)\r\n 378 elif start < self.start:\r\n 379 if self.end - end > self.blocksize:\r\n--> 380 self.cache = self.fetcher(start, bend)\r\n 381 self.start = start\r\n 382 else:\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in wrapper(*args, **kwargs)\r\n 86 def wrapper(*args, **kwargs):\r\n 87 self = obj or args[0]\r\n---> 88 return sync(self.loop, func, *args, **kwargs)\r\n 89\r\n 90 return wrapper\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in sync(loop, func, timeout, *args, **kwargs)\r\n 67 raise FSTimeoutError\r\n 68 if isinstance(result[0], BaseException):\r\n---> 69 raise result[0]\r\n 70 return result[0]\r\n 71\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\asyn.py in _runner(event, coro, result, timeout)\r\n 23 coro = asyncio.wait_for(coro, timeout=timeout)\r\n 24 try:\r\n---> 25 result[0] = await coro\r\n 26 except Exception as ex:\r\n 27 result[0] = ex\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\fsspec\\implementations\\http.py in async_fetch_range(self, start, end)\r\n 538 if r.status == 206:\r\n 539 # partial content, as expected\r\n--> 540 out = await r.read()\r\n 541 elif \"Content-Length\" in r.headers:\r\n 542 cl = int(r.headers[\"Content-Length\"])\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\client_reqrep.py in read(self)\r\n 1030 if self._body is None:\r\n 1031 try:\r\n-> 1032 self._body = await self.content.read()\r\n 1033 for trace in self._traces:\r\n 1034 await trace.send_response_chunk_received(\r\n\r\n~\\miniconda3\\envs\\datasets\\lib\\site-packages\\aiohttp\\streams.py in read(self, n)\r\n 342 async def read(self, n: int = -1) -> bytes:\r\n 343 if self._exception is not None:\r\n--> 344 raise self._exception\r\n 345\r\n 346 # migration problem; with DataQueue you have to catch\r\n\r\nClientPayloadError: 400, message='Can not decode content-encoding: gzip'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n- aiohttp version: 3.7.4.post0\r\n\nThanks for investigating @albertvillanova ! \ud83e\udd17 ","embeddings":[-0.3864652812,-0.2204957455,0.0894813165,0.4387415349,0.2157086134,0.1229229122,-0.0369101539,0.3093882799,0.2422775328,0.0955026373,-0.2264311761,0.4261536002,-0.0641756877,0.3035327792,-0.0094654057,-0.2584786713,-0.0045060921,0.2337655127,-0.1049250886,0.1268760264,-0.0005556306,0.1941259056,-0.2469931841,-0.0304233823,0.0522285365,-0.1081046611,0.0124065261,0.2805253863,-0.059583243,-0.4462363422,0.1259049624,-0.0615482517,0.4121141732,0.3375741839,-0.0001145032,0.2734483778,0.4721427262,-0.1227841005,-0.4853440225,-0.2121149004,-0.218665719,0.0961271524,0.0428345352,-0.1566341817,-0.0091345869,0.0711133331,-0.0400046185,-0.7192984819,0.3178500533,0.3456359804,0.1778500229,0.0426006429,-0.0489487238,0.0268223863,0.0688386783,-0.2986646891,-0.0920066237,0.1218173951,0.2367273569,0.2287404835,-0.359573245,0.3845903277,-0.0658884048,0.0601840951,0.0944148228,0.0967238322,-0.1749501675,-0.4655051231,-0.0073944977,0.1884141415,0.6776787043,-0.2555166781,-0.4161978662,-0.2319881767,-0.0021713183,-0.6144573689,0.1465310603,0.1489273459,-0.4673651159,0.1280597448,0.2459630817,-0.0087530883,-0.327994585,-0.0458728224,-0.246799171,0.3882223368,-0.0767857283,-0.0237270929,0.0272943433,-0.0866289064,0.4627741277,-0.2110346556,-0.28459391,-0.0639790371,-0.4397552311,-0.0420655161,-0.0765123144,-0.094109349,0.1681466401,0.215389967,0.2953203917,0.0882523879,0.1543645263,0.0926442817,0.3555265665,0.3608669043,0.0598316565,0.1853774786,0.1311810017,0.0994313955,0.1366237104,-0.1818924248,-0.146488592,-0.0372809209,0.0281632524,-0.0131140957,0.262858659,-0.1912374645,-0.4138476253,0.0297920927,-0.3057210743,-0.0794672072,0.0055905147,0.2817168832,0.0958337784,0.4187803268,0.0194585584,0.1178673506,-0.2396837324,-0.4573938251,-0.0878112167,-0.1247182265,-0.0010082311,0.1233467832,0.0153353186,-0.1886832416,0.1515737325,-0.0465371571,0.1621010602,-0.0012185768,-0.0208561029,-0.1100598425,0.0216292907,0.2049841285,0.2180556655,0.1640353203,0.1795151085,-0.1997850239,-0.130682379,0.0886226073,-0.1416330338,-0.1671275645,-0.1788374335,0.2166469097,-0.0094024939,-0.3506264389,-0.0937892571,0.2814688385,0.1409590542,-0.2037903368,-0.1410701424,-0.0391465798,-0.1241319329,-0.0865156949,0.1792775393,0.5796180964,-0.1855915189,-0.1380216628,-0.1521381587,0.0008939481,0.5833057165,0.3344025612,-0.0607914478,0.0897968188,-0.170722425,0.1476544142,0.5017140508,-0.1436584145,-0.5796036124,0.5436661243,-0.0396082066,0.5637281537,0.2707032561,0.0225366186,0.1742344499,0.0431196988,0.313529104,0.3299024701,-0.1484592706,0.0074822325,-0.3334669471,-0.0399836674,0.4416950643,0.1912285686,0.0674290732,0.2502861321,-0.0288742278,-0.0626677349,0.3689779639,-0.0752563924,0.0763451308,0.0410080664,0.1468882412,0.0120370993,0.1379382461,-0.2090770155,-0.310682714,0.2972311378,0.0749307722,0.0127100674,-0.3174338341,0.2031723857,-0.1906455308,0.0017469627,-0.2628925443,-0.2590405643,0.1243173257,0.3170801699,0.0756207108,0.0505895875,-0.3084766865,0.7747215033,-0.078255251,-0.0131349247,-0.4549141228,0.2024929523,-0.0707768798,-0.2014573216,0.0164402872,-0.1265656799,0.0672556311,-0.0161919687,-0.2579914331,0.3313200772,-0.2366138101,0.2640713155,-0.1954375803,-0.1731976271,0.2897448838,-0.3644788265,-0.0866843313,0.2706028223,0.1075036898,0.125209704,-0.0176983867,0.1480839401,0.2070915401,0.2361335903,0.0434485786,0.1143487841,0.1961675137,-0.0816748515,-0.0686837137,-0.0850090533,0.3240485787,-0.3548789024,0.0467500761,-0.1989230514,-0.2722995877,0.0823631957,0.3586791456,-0.1403829306,-0.0492737889,0.3303245604,-0.2885322273,-0.1639078707,0.3145605624,0.1940099597,0.4723360538,0.0815219656,0.2573353648,-0.0455882996,-0.1886483878,-0.0735060051,0.3051697612,0.1374839693,0.0314638466,0.2464114279,0.0300063528,0.1635747999,-0.3229160309,-0.3497166634,-0.0566318817,0.1123416647,-0.3201874793,0.306302011,-0.38826859,-0.4651744664,0.0463509336,0.1560085416,0.0689049438,-0.3235587776,-0.1817550063,0.2325055152,-0.0038823313,0.1636838913,-0.2682063282,0.0895485729,0.2671770155,-0.0134687945,-0.3418698907,-0.1808474958,-0.2729629874,0.0591977872,0.1749951094,-0.4228037596,0.1769195795,0.0074260007,-0.1160532683,-0.1401735991,-0.076423429,0.3197818398,-0.1554995328,0.1568958163,-0.2774865627,0.1493911892,0.2213633955,-0.2503491044,0.1219482496,0.0908335969,-0.1187844425,0.2801248431,0.0940897167,-0.0241610892,-0.0305832736,-0.3010807931,-0.3024666011,-0.6177494526,0.0549683981,0.0454257689,0.2135515809,0.1643050015,0.0170347784,0.0968246013,0.3292069137,0.0991601422,-0.114212282,-0.4199434817,0.3747448325,-0.1670146883,-0.2868409157,-0.3248367012,-0.0681597292,0.1972901672,0.2589270473,-0.5295243263,0.2950162292,-0.4081035852,0.0848925635,-0.3834516108,-0.0974560827,-0.0261515174,0.0762885734,-0.0349893123,-0.1878576428,0.0707459524,-0.1169513538,0.0990076959,0.2066107839,0.1086307243,0.3698712885,0.1008493528,0.3327504098,0.4812112153,-0.0195496976,0.3407551944,0.082891129,0.1538007557,-0.1004334986,-0.0627885908,-0.0955621898,-0.2280800343,-0.0261197276,0.0981990993,-0.0962127745,-0.5762300491,-0.2840516269,0.0549149849,-0.0692367032,-0.2383105308,0.0528838076,-0.511775136,0.0420482606,-0.1872570366,0.1583397985,0.0807970166,0.0520012863,0.0860457644,0.1772524714,0.1261645257,0.1721196771,-0.185021311,-0.0639052019,-0.4881900847,0.1825860441,0.1051262617,0.4215749502,-0.1678105742,-0.1327165961,0.0791147202,-0.0652071014,0.3487038612,-0.2754168212,0.003567247,0.2673943937,0.1421353668,-0.1415521204,0.0917675421,-0.3209457397,0.1368288398,0.235954389,0.0974201635,-0.1718239039,0.0333400816,-0.0884195939,-0.0724961758,-0.1253028065,-0.0163991563,-0.1704963744,-0.2492375523,-0.152009964,-0.134242788,0.2032724619,0.1256832927,-0.0124605224,-0.0567925759,-0.0460816547,-0.0396244489,0.2418593764,0.2213254869,0.1399096996,-0.072280176,0.2290114909,0.0671992972,0.2844450474,0.2292162329,0.836432457,0.1181946918,-0.5624260902,0.2350820601,0.2339563966,0.1660992503,0.1305397898,-0.148232609,-0.0085166721,-0.0403849594,0.2698210478,-0.3024733663,0.2146548331,0.2112327665,-0.0507208817,-0.5149303675,-0.4357535839,0.5883390903,-0.039368242,-0.0761907846,0.7872929573,-0.1665049642,-0.3171558082,0.4775415361,0.0234355517,0.7585879564,0.0943952724,0.2168687731,0.2407481223,-0.3207278252,0.5003329515,-0.2320687622,0.3346687257,-0.1206100509,-0.3350296021,-0.0831092745,-0.1041097715,0.1709026545,-0.147603929,-0.6302542686,0.1709609181,-0.0596441738,-0.2528051734,0.3353525698,0.2742269933,-0.6082689762,-0.3062928915,-0.0675497502,0.1388825476,-0.2213602215,0.3239610493,0.1382443607,-0.2202336341,-0.0515731983,-0.0843103677,-0.4702361524,0.4758630693,-0.2765566409,0.1624139398,-0.2390307039,-0.4760982394,0.1513123959,0.2664022446,-0.1421327889,-0.2257857472,-0.1582512558,0.2350440472,0.1860539466,0.128023684,0.0993071273,0.1514580995,0.3883590102,0.0532772318,-0.0168129094,0.1884423494,-0.0089801019,-0.1606287062,-0.0637812093,0.0554574467,0.3183139861,-0.2681806087,0.2368578017,-0.1058510691,0.0429251082,-0.1842529923,0.0869722143,-0.0188314319,-0.1740541756,-0.0738878325,-0.08129704,-0.2196760625,-0.0847442299,0.3771200478,0.1454578936,0.3179713786,0.4994189143,0.0281169843,-0.1778806895,-0.1036716029,-0.0551506244,-0.2931399345,-0.252612263,0.0693599209,-0.2414764762,-0.1610280573,-0.2706694007,0.0038265749,0.0448373631,-0.103554666,-0.1137386858,-0.2993660271,-0.1906369925,-0.0266146194,-0.1242436618,0.2199304104,-0.1722201109,0.0222039819,0.0031197246,0.1290859282,-0.270683825,-0.1228595003,-0.0753692091,0.0231177211,0.2672106326,0.0292674527,0.2592337132,0.1096690521,0.1476223618,0.3139047623,-0.0276747048,-0.1609539241,-0.1787950993,0.138489455,0.0165312476,-0.348051548,-0.1548943967,-0.05662065,0.0126910359,-0.125136137,0.3503144085,0.181230098,0.0380968302,0.2706013322,0.1240136698,0.0563360825,-0.2326476574,0.1542367488,-0.2642730176,0.2401308864,0.0811118633,0.0928405821,-0.0755398721,-0.0118692657,-0.3646134436,0.0622693747,0.1017946154,-0.0280170571,0.5909268856,-0.4241805971,0.1720450372,-0.0497748405,0.0660299286,0.4045228958,-0.2296188772,-0.2089618146,0.0850852206,0.2000514269,-0.0432033539,-0.2126210332,-0.4010480046,-0.1016407683,0.0112383589,0.1459593624,-0.1068896204,-0.2721833587,0.0999634638,-0.2103084773,0.4859834313,0.3358429968,0.2670664489,0.4836392403,-0.1811283231,0.0228473805,-0.2260247022,0.0204593837,0.1999064386,0.2970430553,0.0035512666,-0.0078696832,-0.1613658667,0.2178194076,-0.1859746873,-0.482195437,-0.0797479451,0.1024470255,-0.1647549272,0.3901939094,0.1489096284,0.1521606892,-0.2779354751,-0.0519221425,-0.0458415113,-0.0661262944,-0.0142799746,-0.096080415,0.1068764329,-0.0128562739,-0.3279225528,0.0976088271,0.0123996465,0.0633685365,-0.0198641457,0.2808055282,-0.2884631157,-0.1185014918,-0.0822276548,0.1043173745,0.2999251485,-0.3544151187,0.4440693855,0.3661904037,-0.1568607092,0.0282512195,0.0700021908,0.4816074371,0.2853379846,-0.1766507179,0.0946500823,0.3374164999,-0.0436370596,-0.0402214266,0.0442541391,0.2830949128,0.1793196946,0.4566897452,0.0887393281,-0.1485419422,0.1045718119,0.0814351737,0.2538890839,-0.2799384296,0.4623198807,-0.1195058972,0.0366923138,0.0272421893,-0.3365718424,-0.5624158978,-0.0282046329,0.0348961949,-0.2285001576,0.0745235756,-0.0081122555,0.0628783256,0.1139546707,0.4926685393,0.3248410225,-0.0188453235,-0.2245095521,-0.1508785933,-0.235583514,0.2332020253,-0.0674795657,-0.1231484041,0.1752853394,0.2219228297,-0.0048415069,0.1680491567,-0.1584448814,0.0204550214,-0.2672839463,0.1405724436,-0.3995800018,-0.3515672684,0.1919635981,0.1735794991,-0.0483116023,-0.576480031,0.1881390065,-0.1502387971,0.1608176678,0.0926132202,-0.241848141,0.101145193,-0.2756578326,0.5297338963,0.2331105173,0.3792004883,-0.184295252,-0.4119567871,-0.0992468074,-0.3762306273,-0.2749142647,-0.0155847622,0.1083334535,0.5696322322,-0.1493341029,-0.2373841256,-0.3574708402,0.3262623847,0.0608894229,-0.2656728923,-0.2735387981,0.1733704358,-0.1194598079,0.2131261081,0.3050509691,0.380345583,-0.0287131406,0.1639950573,-0.2235154659,-0.4333209693,0.5773110986,-0.3461063206,-0.2659671307,0.2418412119,0.0638139024,0.3264979422,0.2410199344,-0.4596662819,0.0893656164,0.3556547463,-0.2354129702,-0.1511856616,0.1421462744,0.2259705812,-0.0607201755,-0.0850398242,0.45531106,0.2423961759,-0.249204129,0.1750813872,-0.1786384881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2917","title":"windows download abnormal","comments":"Hi ! Is there some kind of proxy that is configured in your browser that gives you access to internet ? If it's the case it could explain why it doesn't work in the code, since the proxy wouldn't be used","body":"## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n","comment_length":41,"text":"windows download abnormal\n## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n\nHi ! Is there some kind of proxy that is configured in your browser that gives you access to internet ? If it's the case it could explain why it doesn't work in the code, since the proxy wouldn't be used","embeddings":[-0.1240484565,-0.2018248588,-0.0703396499,0.1894593984,0.2147738934,-0.1120505109,0.0276671592,0.0718739256,0.2805814743,0.0850560516,0.1648736149,0.1174071059,0.1554815173,0.1022035182,0.1588474363,-0.2615045309,0.0301855393,0.0393100753,-0.1691515595,0.0088643078,-0.3215381503,0.0859283432,-0.3066691756,-0.1127134934,0.1635590941,0.0324402116,-0.2886959612,0.3004715145,-0.2545996904,-0.1694043726,0.099369511,-0.0709087476,0.1850661635,0.3519495726,-0.0001087431,0.0813257769,0.3922842741,0.0235310104,0.0174283162,-0.0581749007,-0.2938911617,-0.1413686872,-0.1911946684,-0.3151716888,0.0997065529,0.2347301394,0.0454377979,-0.1112284884,0.244362548,0.4491443336,0.3060822189,0.4809369445,-0.1150025278,0.0517268442,0.4554628432,0.0505648442,-0.1816764176,0.2491989583,0.3063491285,-0.1419377476,0.2570019662,-0.059411183,-0.1210118234,0.0115583753,0.0310037415,0.1206353381,-0.0494647175,-0.5093158484,0.1234319583,0.1743744612,0.31266132,-0.1531091034,-0.1585402191,0.1158476546,-0.1493998766,-0.185696885,0.2733875215,0.2762599587,-0.1393169761,0.1871818155,-0.2805486917,0.1743876189,-0.2056800723,0.2596932352,-0.1191073507,0.3299077153,-0.0759498328,0.1894465089,-0.1710143685,0.3167176545,0.03771016,-0.3272551894,-0.0156770907,-0.0039893929,0.0856893435,0.105707854,0.0493557453,0.5118294954,0.05730417,0.2643461227,0.0334886052,-0.0560189262,0.2410898954,0.0700188726,0.090272449,0.0949797034,-0.2958666682,0.066043593,0.3817819655,0.313916266,-0.0065793474,-0.026200084,-0.0330495574,-0.3155067265,0.2965119183,-0.0053916001,0.486035049,-0.04742698,-0.556910634,0.0026351295,0.0146284271,0.0908918232,-0.0023530021,0.2526860833,-0.32183218,-0.022580795,0.1233704016,0.3000146747,-0.1240452528,0.1943567246,0.004063094,0.1806814522,-0.2520693839,-0.2260728776,0.3475835323,-0.0671294406,0.12012472,0.1935619116,-0.0914587379,-0.042873878,0.0132011967,-0.0805152804,0.0271826498,0.4316020012,0.1952559203,0.3702535331,-0.0094595077,0.0780785382,-0.0707269013,0.2062932998,-0.1463157386,-0.0249700788,0.021134628,0.2608017921,-0.3265893757,-0.1885598153,0.0397941098,-0.2664661705,-0.0409113131,-0.0293878373,0.2495985031,-0.1964796036,-0.2891480029,-0.3841315806,0.0929920301,0.5247913599,-0.3482265472,0.2068513781,0.0614641719,-0.271443069,0.4014218748,0.0586097874,0.1309562773,0.1605100483,-0.4302997291,0.0415051356,0.117131643,-0.3501324654,-0.4970429242,0.3584975302,-0.1899531484,0.3206911981,0.1783815473,0.0919841528,0.293910265,-0.0833817273,-0.012038324,0.3098945022,-0.022483252,0.0582434274,-0.1245782077,-0.1297618896,0.1300003082,0.1085215658,0.1208592206,0.1432953179,0.293230623,-0.1360394806,0.3164023459,0.1788195372,0.0964889973,0.2375457287,0.2550174594,0.2680895925,0.1236548349,-0.2269855589,0.0015702655,0.2861069441,-0.0804182291,-0.1590327024,-0.2410328686,-0.2218061388,-0.5582574606,0.0224495344,-0.1886620373,-0.2219815105,0.1653012484,0.1372602582,0.249552846,0.0426937975,0.0754473731,0.179336831,-0.0071864342,-0.068459399,0.0751905069,0.3602052331,-0.3439503014,-0.050523717,-0.096596159,-0.0721758753,0.2299172729,-0.0566853061,-0.0152284885,0.197268486,0.0202615354,-0.0426289514,0.0362087451,0.0486600362,0.0167178959,-0.1805609912,0.2716660798,0.5724846125,0.2685571015,-0.0967838913,-0.0322295241,0.0053620976,-0.075025402,0.0177561473,-0.0222576819,0.1486974061,0.1936407834,-0.188959837,-0.1011113226,0.0021014542,0.3638005555,0.0613047183,-0.0991422012,-0.1685334742,-0.1679939777,0.167080909,0.5450258255,-0.2752330005,0.3014795184,0.0706219375,-0.0966639817,0.2094732076,-0.1800906807,0.1059181914,0.6080446839,0.033296559,0.112960726,0.3711541593,0.0088629806,-0.1436058879,0.2149284184,0.1461233199,-0.2989706993,0.1730046868,-0.0520784967,0.1769192368,-0.2095150501,-0.2376021594,-0.0571574084,0.0526774004,-0.3452227414,0.2429896444,-0.1614445746,-0.1740004569,-0.3291527927,-0.086957626,-0.0766705945,-0.0632937625,0.0246917997,0.2807940841,-0.1450743377,-0.1818347871,-0.4370901287,0.092590116,0.2119114399,-0.3590200245,-0.0015145239,0.1992023885,-0.2863779962,0.1333586127,0.1517552882,-0.0984132662,0.5485373139,-0.5067276955,0.0579211973,-0.3667593896,-0.1009359881,-0.0654943958,0.0553083122,0.3579990566,0.182639122,0.1857500374,0.0548134558,0.1031560078,0.2081247717,-0.2389806062,-0.179515481,0.1597154588,0.0564894304,-0.3985366821,-0.2559876144,-0.2491457164,-0.2066613436,-0.3509724736,0.0161345936,0.1100101173,0.1269879043,-0.1030690521,0.0080765542,0.0372453369,0.0731868744,0.2246696055,-0.083026208,0.0263501108,0.3867479861,-0.0423500165,-0.6439450383,0.2405601293,-0.0189747661,-0.3578401804,0.3052423,-0.5030599833,-0.132392019,-0.2873800397,0.2506045699,0.0892270058,0.2778801024,0.3353272378,-0.2777280211,-0.0245965142,-0.3039755225,-0.0659073442,-0.2901861966,-0.1096735597,0.2428965271,0.1826853007,0.3031284213,-0.0289380215,0.2855456471,0.2206116915,0.1294395328,0.4378712475,-0.3864388764,0.4741620421,0.001300477,-0.1765860617,0.035116639,-0.2722176611,0.0180080682,0.2376574576,0.022387892,-0.004943266,-0.2312289178,-0.1656051874,-0.231309548,-0.2478713095,0.1438776106,0.1106841117,0.0407827273,0.1241698265,0.2241891772,0.1028406844,-0.1557345092,0.1141644493,0.5970383286,-0.2615623474,0.1439668685,0.0227814205,-0.1640785187,-0.4693872631,0.0716748014,0.1027646363,0.4488425553,-0.0301668625,0.0443425849,0.2156197429,-0.1619966328,0.7495212555,-0.2151187062,0.3322592676,0.2032127231,-0.0067156502,-0.3780357242,-0.2583532333,-0.5120751262,0.123098731,0.0817756206,-0.0529246628,-0.0555245318,0.1852327734,0.1150577143,0.2468607128,-0.1428902447,-0.3871606886,-0.3885157406,-0.6134721041,-0.2445401102,0.085090965,0.4038743079,-0.0435088724,-0.3796861768,-0.1664035767,-0.0642804652,0.1533765495,-0.0725113228,-0.0277298614,0.0795960501,-0.017775014,0.1780821234,-0.2506742775,0.2862488925,-0.1739084572,0.6596170068,0.1054916233,-0.1496224999,0.1620587111,-0.2683568299,-0.2014200389,0.2051974386,-0.1649345309,-0.220199585,-0.0797214359,0.0492504165,-0.3298475146,-0.0799625888,0.5729669333,-0.101683788,-0.0054186624,-0.2596149147,0.1787786633,0.0138970232,-0.1558488905,0.2674739361,0.0348851718,-0.0924990326,-0.108341001,-0.2890718579,0.651114285,0.2145095021,0.2358662635,0.2590305209,-0.5253791809,0.0419862084,0.2136851996,0.0738660768,-0.2070637792,-0.2838785946,-0.0373834483,-0.1332621872,0.0825860873,-0.0341505371,-0.2779364288,0.2207877785,-0.0865820795,0.1747573763,0.2744155824,-0.1581595838,-0.313105762,0.1084561571,-0.0328159146,0.195117116,0.2385889739,0.6629055142,-0.2662183642,0.0184965637,-0.1715994924,-0.2779659629,-0.3965134323,0.255030632,-0.1340843886,0.0074540563,-0.1599158049,0.1396337897,0.0225496311,-0.0147371292,0.0021659129,0.1571852863,-0.2298111022,0.0857386962,0.1341610104,-0.0822363645,0.263004005,-0.0994025618,0.1180735081,-0.2447074354,-0.4229418635,-0.0727604851,0.0214396194,-0.0027245332,-0.0277121738,0.2639417648,0.2187709659,-0.1882333159,-0.1784814149,-0.3920820057,0.083862029,-0.0708945021,0.2175155431,0.0490465052,0.0885434523,0.5029727817,-0.2861326635,-0.2359030694,-0.1554895788,0.3932054639,-0.1673704088,-0.0755681172,0.1102359146,0.0948679075,-0.3069735765,-0.2523458302,0.0131361112,-0.3942269385,-0.2205286175,0.0540067367,-0.0543695949,0.2869075835,-0.1471764296,0.0090060364,0.1335751414,-0.0578686707,-0.0151845738,-0.6815165877,0.1296804845,-0.2197964042,0.412201792,0.0466443598,0.0262459274,-0.157206744,0.4631694555,-0.5489163995,-0.3702111244,-0.0008696999,0.0495049134,-0.1290583014,0.1779977232,0.1302772611,0.1998716146,0.0118945697,0.1212280467,0.0002453557,-0.0661850348,-0.2484522611,-0.1650175601,0.1045997515,-0.0540886931,-0.2632085681,0.1900731623,-0.4511168599,0.0091411378,0.18899028,0.14730452,0.3043759167,-0.0957731083,0.0366581641,0.0956626311,0.1146368384,-0.2398875803,0.2471019179,-0.4936737418,0.5263616443,0.0056513948,0.1713121086,0.2247191817,0.0065553351,-0.054802306,-0.281349808,0.062704742,0.1826201379,0.3916451633,-0.1167972237,0.2606552839,-0.0006228012,0.283251822,0.5525352955,-0.352715075,-0.0893501639,0.1433627754,0.275811404,-0.2590067089,-0.0346851833,0.0842292532,0.1091726124,0.0576287955,0.1179928631,0.0555240661,-0.2826647758,-0.0912482589,-0.0500898212,0.4335025847,0.3670791388,-0.1808381081,0.3947139084,0.0065175421,0.1476498097,-0.0576697998,-0.0217515342,-0.0486023426,0.4234354198,0.0409806445,0.0692659318,-0.1091255769,0.0251742806,-0.1315766871,-0.2052455992,0.0766841099,-0.2012568414,0.0254510138,0.097710222,0.0770515501,-0.1540478617,-0.2748750746,0.2177974731,-0.3454121649,-0.0678378046,0.0694723278,-0.0125074452,-0.3036753833,-0.0721507445,-0.1196656898,0.0252670608,0.1473329514,-0.0054127262,0.259942174,0.0745272934,-0.2369666696,-0.3160700798,0.0126224859,-0.2048031986,0.1134118363,-0.2885008752,0.0778895468,0.3658398986,0.1674333662,0.1078387499,0.1148513034,0.5545535088,0.3852181435,-0.1072471216,-0.061104849,0.0784613341,-0.1331625879,-0.3258143961,0.43635571,0.071632877,0.193093881,0.2363510728,0.2184062153,-0.2301051915,0.2584095299,-0.030818494,-0.0795634016,-0.0589464083,0.1791744381,-0.2505965829,0.0937965512,-0.3897891343,-0.1013927385,-0.4083535373,-0.098580189,0.1772197485,-0.0879954323,0.1667040288,-0.3462209702,0.1035932824,0.0375302359,0.2181000561,0.3795768917,0.3196724355,-0.088166751,-0.0573232137,-0.5226137042,0.3419146836,-0.2258433551,-0.5296083689,-0.0877915993,0.0447274148,-0.6105858684,0.0722363666,0.3415539265,0.2679618895,-0.0636082962,0.3222130537,-0.1352423877,-0.3121239245,0.0335871987,0.1525713652,0.0759184957,-0.5940473676,0.2918206453,-0.4170546532,0.1697357893,-0.2700899541,-0.1863706708,-0.0113285938,-0.0088556763,0.5608647466,0.1507781595,0.5199138522,-0.1912190616,-0.1256828904,-0.4629922211,0.0781015754,-0.1126953214,0.2469523996,0.116675429,-0.1811442822,0.0191525444,-0.2678036988,-0.0600166209,0.3138544858,-0.0368412472,0.0905802175,-0.0571234785,-0.0728074685,-0.183093071,0.1488180161,0.0811512843,0.1466140598,-0.3787932098,-0.1018520966,-0.0337751284,-0.5434590578,0.2741713226,-0.2698807716,-0.3332621753,-0.2687217593,0.2311914861,-0.4032007754,0.0730406642,-0.0221859496,0.1500967294,0.1658335179,-0.0674838051,0.0399886183,0.2936095595,-0.3748459816,0.0406060629,0.25768736,0.4179147184,-0.0696628317,-0.1712902337,0.3069999814,-0.2148606628]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2917","title":"windows download abnormal","comments":"It is indeed an agency problem, thank you very, very much","body":"## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n","comment_length":11,"text":"windows download abnormal\n## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n\nIt is indeed an agency problem, thank you very, very much","embeddings":[-0.1819480658,-0.0709162951,-0.0238682684,0.1505493969,0.2515722811,-0.0938239619,0.0032474413,0.11024113,0.362229526,0.0661070794,0.1379210949,0.2134204656,0.0683521777,0.2074184567,-0.0183983669,-0.2120418251,0.0408425257,0.169603616,-0.109351553,0.0084896954,-0.3778957427,0.1222561225,-0.244990021,-0.0322376676,0.118513003,-0.1276566386,-0.3018625379,0.1873607785,-0.2236432433,-0.0762688667,0.1565723866,-0.1330198646,0.1389445513,0.3082361221,-0.0001110555,0.0270548705,0.4110395014,0.0012154208,0.0296383202,-0.0051854542,-0.1887717694,-0.1067669615,-0.2910613418,-0.3018800914,0.163738668,0.3125549555,0.0271366872,-0.1251074672,0.2193036079,0.4728519619,0.3004208505,0.4755435586,-0.1050411314,-0.052202031,0.2329810262,0.0768294707,-0.1763174385,0.1497375667,0.254214406,-0.2113004178,0.2170041651,-0.0690852255,-0.0610753894,-0.0063095409,-0.0902022049,0.0914523378,0.0073744375,-0.5927760601,0.1896012127,0.1313620359,0.401514262,-0.1367759109,-0.1460928917,0.1193385571,0.0144195985,-0.0813066885,0.2506349087,0.3949985504,-0.1587389559,0.0788909793,-0.2028708309,0.2833296955,-0.2204892486,0.1841510683,-0.1222558171,0.1910018027,-0.0857133567,0.2090443075,-0.0596853197,0.2083201408,0.0718123764,-0.2013075501,-0.0888202116,-0.0001829596,-0.0580069907,0.0415586531,-0.0736014992,0.6245207787,0.131835565,0.1753447503,0.0869020745,-0.1592784822,0.183907643,0.0403068922,0.224851951,0.1441524029,-0.2167216688,0.1089039966,0.2697381079,0.2915846109,-0.0308599882,0.1096391231,-0.107437022,-0.41546157,0.2698237002,0.0131871272,0.4077474773,-0.1028263122,-0.62843436,0.0436515138,0.1092530116,0.1674703658,-0.1276319921,0.1536726505,-0.2733654976,0.013872534,0.0132438513,0.32657969,-0.1094577163,0.206694752,0.0036723863,0.0926430076,-0.2967072725,-0.2458346933,0.3576645553,-0.1119512916,0.1324940771,0.2963950634,-0.1462853998,-0.01702445,0.0105879866,-0.1995972842,-0.1029506847,0.439473182,0.2255644649,0.3950291276,-0.1827936321,0.0152565576,-0.0298757032,0.1933602542,-0.1561779529,-0.0408560708,-0.0541052632,0.2324374616,-0.3855369985,-0.1968738139,-0.039049834,-0.2038401216,-0.0327578783,0.1615654677,0.2465070188,-0.2326217592,-0.2305738032,-0.3819427192,-0.0289984588,0.5887625813,-0.2635556459,0.2373927385,-0.0847060829,-0.328499347,0.4486765862,0.1127441376,0.1913455874,0.1516014636,-0.4607596099,0.0760051906,0.0519300029,-0.4398525655,-0.5598500967,0.4142220914,-0.0977686942,0.1617045999,0.1845384538,0.1501137912,0.3036822379,-0.0325751416,-0.0825505555,0.4031662345,-0.0779337958,0.1080535129,-0.1352510601,-0.134622559,0.201709345,0.168760702,0.1124669537,0.0058825724,0.1826209873,-0.1374899149,0.298524797,0.2201125473,0.0792058855,0.2166628391,0.2818375528,0.1327269375,0.1090106666,-0.1992979348,-0.0353883244,0.2307479382,-0.068367891,-0.1274635792,-0.1719899923,-0.2560539246,-0.4767713547,0.0034119347,-0.2548181415,-0.193141818,0.1301574707,0.0973193944,0.1700637639,0.1048899293,0.1259745061,0.1654153466,0.0545962192,0.0097603537,0.1804260761,0.302990526,-0.2577700615,-0.0457421467,0.0831682235,-0.0594857298,0.2411195934,0.0036581284,-0.0234827735,0.1238250807,0.0947378799,-0.0369792469,0.0112583535,-0.0961492285,-0.0187985655,-0.0046684532,0.208516553,0.5677586794,0.1324958503,-0.0747223124,0.0045230803,0.1056594327,-0.1307276189,-0.0183902904,-0.0695005655,0.2344380766,0.2568475604,-0.3220861256,-0.0040925732,-0.0434314571,0.2829574049,0.0912019163,-0.0603718944,-0.0726193562,-0.1558505297,0.1877027154,0.700084269,-0.1830890179,0.3046873808,0.1612100303,-0.1223507747,0.1930639148,-0.137361154,0.2834716737,0.7158129215,0.0091603678,0.0177161265,0.2950727344,0.002406968,-0.1490684301,0.2080704123,0.0997634754,-0.2241471559,0.3055584729,-0.0021933594,0.2467545569,-0.3240906596,-0.2490481883,0.0142765054,0.1652551144,-0.2981254756,0.098265931,-0.1337950528,-0.2373199612,-0.4030925632,0.0705316141,-0.0321190171,-0.0354837701,0.0339757949,0.1372851282,-0.0919834822,-0.1605060548,-0.3157018423,0.1221830696,0.1355545223,-0.1588221341,-0.0273725614,0.2578795254,-0.2540774345,0.1120541915,0.2056393325,-0.0287933052,0.5603420734,-0.496091783,0.2043798715,-0.4564227164,-0.2349586785,0.0054968395,0.0737605169,0.429425329,0.1740064919,0.074083142,0.0736172348,0.0474765077,0.1184441224,-0.2257839143,-0.2208357453,0.1726488918,-0.0539861992,-0.4455712736,-0.3471133113,-0.2787243128,-0.2446427345,-0.3425118625,0.228518784,0.1010006294,0.1010874212,-0.1100757569,0.0160217341,-0.098175019,-0.0168962702,0.1484609544,-0.0863984153,0.1665725112,0.4026210904,-0.0914862081,-0.6598010659,0.226908803,-0.0615841225,-0.2246567756,0.2210018188,-0.4095596969,-0.1262203753,-0.342110157,0.129096806,0.0070878975,0.2367419153,0.2802664936,-0.2545440495,0.0095087094,-0.2566065192,-0.113554813,-0.3147451878,-0.0000580463,0.1881667525,0.0922107473,0.2084324509,0.0313312039,0.1634161323,0.3067245483,-0.0213810727,0.3113523424,-0.4172437489,0.4244993925,0.0466330908,-0.2214517891,0.0983962864,-0.2498928308,0.1606142521,0.3106133938,-0.0131996665,-0.0498652756,-0.0531375594,-0.1525879204,-0.2962233722,-0.1872186512,0.1845183671,0.1975913942,0.0270271003,0.1604973674,0.2584456503,0.0593817346,-0.1810683012,0.1808519512,0.5908366442,-0.2561921775,0.1567899436,0.0854340792,-0.1405245364,-0.3897050917,0.0586795174,-0.048919417,0.3696776927,0.0397022516,0.092400454,0.183471337,-0.1580636948,0.7567191124,-0.2351808399,0.335219115,0.0931395069,0.0759490356,-0.3002094626,-0.2571650147,-0.5032290816,0.1778117418,0.1313455105,0.1499896944,-0.0213212427,0.1512990743,0.0676208735,0.2166710496,-0.1317122281,-0.4037706852,-0.2964296639,-0.6127716899,-0.2049366981,0.117291458,0.3299233913,-0.0507941172,-0.3962244689,-0.2173570246,-0.0589839593,0.1337932795,0.0097574433,-0.0469903946,0.1613981575,-0.0620574243,0.354998976,-0.2106681168,0.2135154009,0.0068420973,0.6776823997,0.0515572429,-0.20278126,0.1849692166,-0.4547348022,-0.075928323,0.1978503317,-0.2219791412,-0.181578353,-0.2288752943,0.0577406138,-0.1304596514,-0.1154361218,0.6724388599,0.0116147408,-0.0880290642,-0.2290315926,0.0269637574,0.0930446237,-0.1775280386,0.1311566979,0.0083262734,0.0038521371,-0.0504921637,-0.3116265535,0.5648178458,0.1825739145,0.2294044197,0.25097996,-0.6363233924,0.132541433,0.1731054485,-0.0348364413,-0.3449183702,-0.1180073246,-0.060735397,-0.0839627385,0.1357378662,0.0965342596,-0.2864951193,0.2465907186,-0.0923556387,0.1755324602,0.2961422801,-0.053293705,-0.4048813879,0.0279879328,0.0016065218,0.1267795116,0.2139146924,0.5632892847,-0.1866210252,-0.0936785042,-0.0864489675,-0.2423260212,-0.4809702337,0.2019794881,-0.1480094939,-0.0914589316,-0.0404894277,0.2275454849,0.2673154175,0.1050150692,0.0534894355,0.1554015577,-0.2471711636,0.09378694,0.2429651618,-0.163590923,0.3407642841,-0.1084667817,0.1350386739,-0.2384069264,-0.2937361002,-0.2282719314,0.0748440102,0.0276120156,-0.116148755,0.2478477508,0.0971163362,-0.2641325593,-0.1859139502,-0.4118072689,0.0480490588,-0.0216537248,0.1968362927,0.0426948182,0.2514934838,0.5543159842,-0.2680101395,-0.2960169911,-0.1482276917,0.316008538,-0.0751970038,-0.0281304382,0.0849173367,-0.0338713862,-0.2844862938,-0.2949290574,-0.1094078198,-0.3924708068,-0.4937892854,0.0646348968,0.0028452519,0.2016657442,0.0639257208,-0.0053070709,0.0866598338,-0.1805602014,-0.0810055435,-0.6708059907,0.1208183914,-0.2188697308,0.3121037781,-0.0310322922,-0.080994457,-0.0924195573,0.462344259,-0.56297791,-0.3270189762,0.0689965785,0.0888255239,-0.0834360719,0.109193787,0.139086917,0.1852610856,-0.0086610345,0.0958006084,0.0629086271,-0.1129151657,-0.2175378799,-0.239239648,0.123343803,-0.0251545962,-0.1054501235,0.3249388933,-0.3739489019,-0.1129374206,0.2286068797,0.0371736549,0.2270472944,-0.1074041575,0.0664063394,0.1346448511,0.1107195541,-0.1786526144,0.2178373933,-0.4177097082,0.502301991,-0.0636654794,0.2077580541,0.0905166268,0.0062949709,-0.042831663,-0.2727173567,0.0531402491,0.3122295439,0.3746222556,-0.1874159575,0.2840179205,-0.0366584845,0.2948142886,0.4220358133,-0.3256885707,-0.1145882607,0.2006989568,0.2581072748,-0.3778952956,-0.1126070246,0.0249982718,0.0281744618,-0.0173770655,0.0426898114,0.1704592407,-0.3231846988,-0.0715658963,-0.0891949311,0.510776937,0.3488326073,-0.3181062639,0.4961268604,-0.0357793383,0.2045413852,0.1308365911,0.0022999949,-0.123532027,0.466999203,0.1138676703,-0.0657802895,-0.07365264,-0.0520526879,-0.0127656963,-0.3265745342,0.0969003737,-0.2036779374,0.0883152783,0.1655903906,0.2275578827,0.1800905913,-0.1424449831,0.1382020563,-0.3902211487,-0.1678872854,0.0352386832,0.1181734353,-0.3001885116,0.0077317753,-0.1403809637,0.1666193604,0.03115898,0.0294490382,0.3884918988,0.1017772108,-0.2867660522,-0.2919543087,-0.065970242,-0.1406376809,0.1262633651,-0.3228229582,0.0812824517,0.1998783797,0.0639883801,0.1520850211,0.1378840059,0.6043738127,0.4262299836,-0.1217401624,-0.052427981,0.0836070701,-0.035255909,-0.2587223053,0.4822923541,0.1293968409,0.2001360506,0.219263345,0.1742515415,-0.2169545889,0.1601615399,-0.0443608537,-0.0834314376,0.0095301662,0.1602160633,-0.226329416,0.0906851217,-0.4063223898,-0.0822347403,-0.356235534,-0.2838271558,0.1814249009,-0.0034272557,0.1390607953,-0.5132735968,0.0922186598,0.0056273444,0.2130190283,0.3403390646,0.3399863839,-0.1434624493,-0.0310119316,-0.5878449082,0.4222592115,-0.1601106822,-0.469481945,-0.0580243543,-0.074073568,-0.4996267855,-0.0051813284,0.3259088099,0.3747652471,-0.034929134,0.2777367532,-0.2828331888,-0.2070655376,-0.0261246599,0.1599916518,0.1422616839,-0.6948549747,0.2628783882,-0.3964836597,0.1354710013,-0.3572916687,-0.1515267193,0.0148491617,-0.0411220454,0.4819484055,0.0784450695,0.5681082606,-0.2400528491,-0.1901470125,-0.3536834717,0.150347814,-0.1731881648,0.2326658219,0.2059119046,-0.2228829563,0.0862541646,-0.1185080707,-0.0339981914,0.2086851001,-0.0462474152,0.2048280835,-0.1679049134,0.1487743109,-0.1486237049,0.0804283768,-0.0019246727,0.2268226743,-0.4055153131,-0.2166316658,-0.0701380074,-0.4929824471,0.2949214578,-0.4310215712,-0.2514631152,-0.1839419752,0.2517825961,-0.3235113621,0.0435998961,-0.0431705825,0.1134158745,0.1523936093,-0.0991116017,-0.0711204708,0.2593789101,-0.3562481999,-0.0355296209,0.2653440535,0.4769489169,-0.0163087025,-0.180393517,0.3521855772,-0.1993495673]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2917","title":"windows download abnormal","comments":"Let me know if you have other questions :)\r\n\r\nClosing this issue now","body":"## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n","comment_length":13,"text":"windows download abnormal\n## Describe the bug\r\nThe script clearly exists (accessible from the browser), but the script download fails on windows. Then I tried it again and it can be downloaded normally on linux. why??\r\n## Steps to reproduce the bug\r\n```python3.7 + windows\r\n![image](https:\/\/user-images.githubusercontent.com\/52347799\/133436174-4303f847-55d5-434f-a749-08da3bb9b654.png)\r\n\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\nIt can be downloaded normally.\r\n\r\n## Actual results\r\nit cann't\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.11.0\r\n- Platform:windows\r\n- Python version:3.7\r\n- PyArrow version:\r\n\nLet me know if you have other questions :)\r\n\r\nClosing this issue now","embeddings":[-0.2207755297,-0.1332957745,-0.0691334903,0.1743752807,0.2265743613,-0.0526336357,-0.0071485089,0.1346075088,0.2471247762,0.0947366357,0.1111562997,0.2011167854,0.0616793036,0.1772477627,0.0522657,-0.2514466345,0.0436070114,0.1884659082,-0.1824369282,-0.0201365389,-0.3421658576,0.1442255229,-0.2688935101,-0.0980346352,0.1254004985,0.0039493642,-0.3086517155,0.2426276207,-0.2599076033,-0.1714349836,0.0742157772,-0.0879949704,0.163236618,0.2892299891,-0.0001064452,0.0440166481,0.3960578144,0.0089306133,0.0559709035,-0.0490174629,-0.2207491994,-0.1396053284,-0.2216101736,-0.293787092,0.1231971383,0.110559687,0.0049971924,-0.0325602069,0.0923154652,0.3783936501,0.3404253125,0.4842950702,-0.0722764954,-0.0313038714,0.4031114578,0.1105747074,-0.1911314577,0.1367606372,0.2970518172,-0.2148809433,0.1355500519,-0.0241315197,-0.0474884994,0.0302238148,-0.0182570498,0.0709910616,-0.0401855186,-0.4947353899,0.1354618669,0.1454054117,0.3116519451,-0.1623973399,-0.1045670062,0.1049401239,-0.0483258218,-0.2304380387,0.1805960536,0.3206201792,-0.1367184669,0.1314445287,-0.2873217762,0.2481869012,-0.2228381038,0.1704220772,-0.0978163704,0.2842999995,-0.0906764865,0.1684194952,-0.0802360848,0.2447992563,0.0083303163,-0.2286545932,-0.100851655,0.0021302979,-0.0380191021,0.0297235865,0.0100703146,0.5421230197,0.129743889,0.2278441787,-0.0172666032,-0.1089707017,0.2379163802,0.0550119728,0.1292164624,0.1321160793,-0.1616736948,0.0282277036,0.3516741991,0.2831096351,0.0294332672,0.1096406654,-0.0448804945,-0.409791708,0.3063461781,0.067747876,0.3790406585,-0.1142017618,-0.6207893491,0.0279846601,0.070389539,0.1223948523,-0.0729451403,0.2053755671,-0.2115510255,-0.038713377,0.0287676323,0.3363829255,-0.1473443359,0.1672639549,-0.0085347267,0.1934087873,-0.2336297184,-0.2789251506,0.4062646925,-0.1174081489,0.1876767725,0.2447303534,-0.1050559208,0.0019086151,-0.0402509943,-0.1226891279,-0.0408820026,0.4309408367,0.2251872271,0.4080506861,-0.1081939563,0.1009383127,-0.0526482053,0.2426031977,-0.1097498462,-0.0552099682,-0.0306426436,0.2899795175,-0.3140995502,-0.2293543965,-0.0028886429,-0.2081403881,-0.0288612023,0.0473718569,0.2172876298,-0.2606675923,-0.2903537452,-0.4342307448,0.0709750429,0.4895909131,-0.4144755304,0.227134794,0.0270461105,-0.3316147923,0.4150057137,0.1907803267,0.1251585186,0.099305965,-0.4734765291,0.1356181353,0.0067809615,-0.3681674302,-0.4561111033,0.3584577441,-0.1014415994,0.1938920319,0.1865333468,0.0966966599,0.288595438,-0.0928407386,-0.0392433889,0.2651661932,-0.0274843816,0.0421565957,-0.1506972611,-0.1795333475,0.1650287658,0.1755206734,0.0803700909,0.0484602489,0.2923360765,-0.1043607593,0.3381632268,0.1145961881,0.0599303469,0.2569989264,0.3408920169,0.1021471545,0.1144763753,-0.243753314,-0.0720155612,0.2930027843,-0.0744742453,-0.1432997733,-0.2104561776,-0.2281522602,-0.4664123952,0.0491136797,-0.2203699201,-0.2093744129,0.1927045584,0.105987832,0.2294083834,0.0463595949,0.0988513902,0.1179435104,-0.0049782838,-0.0852696747,0.1478122175,0.2976120412,-0.358138293,-0.0983111262,0.0428683013,-0.0618663542,0.2359926403,-0.0797711536,0.0072130114,0.1748583913,0.0909606516,0.0120516652,-0.0225844402,-0.0551137365,-0.0018050423,-0.0328586437,0.2095954716,0.4951538742,0.1691873521,-0.1198024824,-0.0891863182,0.0866177082,-0.056391336,-0.0603130832,-0.0616283976,0.135127157,0.2473331243,-0.2254518121,-0.0318140052,-0.0709816664,0.4197178185,0.1235188469,-0.1139750704,-0.1175652519,-0.1888785362,0.154271245,0.6313266158,-0.2423981279,0.2632025778,0.0781768337,-0.0857005417,0.1904331446,-0.1035305858,0.2065784037,0.6523746848,0.0705253407,0.0944053531,0.3197440505,-0.0311157349,-0.1489898115,0.2229828387,0.0814782307,-0.1967646033,0.2652468383,-0.0275604334,0.1729865074,-0.3210701644,-0.1806147695,0.0191281699,0.1348359287,-0.3165897727,0.1074841619,-0.1814183295,-0.0536290482,-0.3422429264,-0.0320985392,-0.0447294153,-0.053749226,0.0935111567,0.2653478086,-0.1439779252,-0.1548583955,-0.2668774128,0.0174200926,0.2345918864,-0.2694985569,-0.021260269,0.218075648,-0.2366104424,0.1633965224,0.1720452756,-0.0932388231,0.5667923689,-0.4775111079,0.1574508995,-0.4538096488,-0.1942285001,-0.0239324756,-0.0515272506,0.3197554946,0.2340763509,0.1066003665,0.0461247899,0.0562638193,0.1573544741,-0.2489885539,-0.2195565104,0.2129605412,0.0073923734,-0.4503124058,-0.3328312635,-0.2961428463,-0.202512458,-0.3447829485,0.1225349307,0.1772155315,0.1006463096,-0.0672200993,0.0358825438,-0.0186522044,0.047959242,0.0959775299,-0.1114475951,0.1012131572,0.4370254278,-0.108312875,-0.668559134,0.2404889315,0.0451720208,-0.2690183222,0.241299063,-0.5037426949,-0.088098526,-0.3273606896,0.2437198013,0.0548889041,0.2400947064,0.2727158964,-0.2193379253,-0.0504980423,-0.2639805079,-0.0492048077,-0.2931634486,-0.0319044366,0.1824257225,0.0572677068,0.2499418408,0.0213780347,0.2403487414,0.2202055752,0.0174942203,0.3821138442,-0.3713877201,0.4587187171,-0.0153283775,-0.1699820757,0.0924867392,-0.2480327785,0.0673405975,0.2961802781,-0.0331741124,-0.0549514927,-0.1332375258,-0.1691046953,-0.1654629558,-0.1967912763,0.1489178687,0.1905439347,-0.0007379284,0.1263190657,0.2819885314,0.126558423,-0.1203787774,0.1470764428,0.529684782,-0.2528937459,0.1417438388,-0.0430495441,-0.1875661463,-0.4491520822,0.0247008894,0.028167218,0.381575644,0.0108286748,0.0817865953,0.1806859076,-0.1728951335,0.7310193181,-0.1994722933,0.3743469417,0.131025508,-0.0181296635,-0.3187329769,-0.2922997177,-0.5790176392,0.1523871869,0.0467407405,0.0785343796,-0.0655063391,0.1080228686,0.1322008371,0.3294463456,-0.1184682176,-0.4172126651,-0.2886999547,-0.5888552666,-0.3001114428,0.0594724566,0.4062659144,0.000087206,-0.4491736591,-0.1950608939,-0.0533062816,0.0927784741,-0.0568115413,-0.0113462741,0.1471142918,-0.0048097325,0.2928668261,-0.2623037398,0.2470901459,-0.028746685,0.6279499531,0.070920743,-0.1808263958,0.1777650118,-0.3867741823,-0.1477529556,0.143608436,-0.2414682657,-0.2455293983,-0.1764716357,0.0498459563,-0.2587479949,-0.0836134553,0.6548749208,-0.0413557142,-0.0423992425,-0.1491366029,0.1900945753,0.0426516347,-0.1285643429,0.2071285695,0.0451181121,-0.0856232122,-0.0590202399,-0.2113350332,0.5505878925,0.2270111889,0.2372392416,0.2324950695,-0.585196197,0.0767249316,0.2548376322,0.0133874621,-0.3171902597,-0.2484219521,-0.0332701951,-0.0792204291,0.1030970961,0.081428133,-0.2686470747,0.2291331589,-0.0590135045,0.1270036101,0.2279467285,-0.1004103422,-0.2982287109,0.0521233603,0.0186875034,0.1960546076,0.2562512159,0.5732926726,-0.2369992733,-0.0070824078,-0.094170548,-0.2494924664,-0.4075240493,0.2002867758,-0.1548224837,0.0800937116,-0.1750008166,0.0427773744,0.0968228355,0.0378658399,-0.0155429896,0.1895344555,-0.2424435318,0.0958389044,0.2168214917,-0.1496645361,0.2837164998,-0.0939572006,0.1009400859,-0.221022293,-0.3328936398,-0.1587910354,0.0103745395,0.0131876329,-0.0436179899,0.2026299983,0.0993676856,-0.1678388119,-0.2867918909,-0.3931019604,0.0494828448,-0.059844099,0.2332053334,0.0822550058,0.1473239809,0.5253173113,-0.2927761972,-0.2554986179,-0.177690357,0.3423021436,-0.1086063087,-0.081868045,0.123623386,0.1187664419,-0.275470823,-0.3161422312,-0.0015436411,-0.3686634302,-0.3794449866,0.0258730557,0.0202567708,0.286070317,-0.0100809755,0.0859160051,0.1481716931,-0.1539181918,-0.0308849066,-0.6326345801,0.0793285519,-0.1875317991,0.3171585798,0.0826699287,0.0632378832,-0.1279541105,0.4226350784,-0.5515289903,-0.3977765143,0.0291253291,0.1112820432,-0.1331367344,0.1548172086,0.1501194984,0.2122796327,0.0203980245,0.1442107111,0.0092959199,-0.0687797889,-0.2692624331,-0.1742177755,0.0721313804,-0.0414409004,-0.1086429581,0.275123924,-0.4210032523,-0.0978210792,0.1876570284,0.17270872,0.2825715542,-0.0950304568,0.1360922009,0.1394856572,0.0868963525,-0.1750506014,0.1824970245,-0.4027463198,0.442858845,-0.0126204034,0.1944123656,0.1665093303,0.0109840715,-0.0307105407,-0.2862520814,-0.0017864812,0.2384109199,0.3138039112,-0.1555044502,0.2605932653,-0.0349944904,0.3879834116,0.4773438573,-0.3409121931,-0.1270853281,0.1898004264,0.3115430474,-0.3598316312,-0.1136988103,-0.0361874923,0.1267793924,0.0516493917,0.1107208356,0.0998693407,-0.2879479527,-0.0161045827,-0.0412315056,0.4778434932,0.3516988456,-0.2632001042,0.5142486095,-0.0232837703,0.1377200931,0.0572274812,0.0515077263,-0.0481213368,0.4958332777,0.0732803345,0.0498245247,-0.0417858437,-0.0422224067,-0.0472286642,-0.1962688118,0.098315455,-0.1797437668,0.075355649,0.1101114526,0.1492832899,0.0244431049,-0.221565485,0.1172366291,-0.3538033366,-0.0532904789,-0.0065229121,0.0325645544,-0.2743485868,-0.0475490838,-0.0886465758,0.1344543397,0.0166579895,0.0213944465,0.2973858118,0.1095635965,-0.2179983556,-0.3157918155,0.0369854607,-0.1580412835,0.0982746109,-0.2871119678,0.040060088,0.3295312226,0.0932035521,0.1183491424,0.134796381,0.5907421708,0.4009403586,-0.1221763343,-0.0771894902,0.0635202974,-0.0590710416,-0.2384877056,0.4153927267,0.039325472,0.1890621036,0.3091205657,0.2307442129,-0.2469318658,0.1552436352,0.080356203,-0.1620993912,0.020642532,0.1388530433,-0.2244401574,0.0271097682,-0.3510095477,-0.1198782027,-0.3236604631,-0.2240209132,0.1686892658,-0.0143957799,0.1356601864,-0.3788411021,0.1349952817,0.002670008,0.2053229809,0.3138542473,0.272441268,-0.1191345304,-0.0695681348,-0.5928861499,0.3832478225,-0.1678831279,-0.4602458775,-0.0931928083,0.0156863183,-0.4653169513,-0.0051278067,0.3414212167,0.2859823406,-0.0933431908,0.2965569794,-0.2014313191,-0.192971006,0.0049811103,0.0661877319,0.1054957882,-0.5753059983,0.318833977,-0.3736201823,0.1851435602,-0.3877812922,-0.09884388,0.0152286161,-0.0082453201,0.5097563863,0.0851208419,0.50721246,-0.254583776,-0.1529332697,-0.3806297183,0.0885771438,-0.1390194148,0.2699479163,0.1687265784,-0.13552773,0.075280495,-0.2353149801,-0.0574993305,0.2708546817,-0.0405838639,0.1234099939,-0.1145117506,0.1258966476,-0.1428504884,0.0675647482,0.0728959367,0.1774835438,-0.3727541566,-0.1398797184,-0.0954014957,-0.5806180835,0.2051665485,-0.2872015834,-0.2997911274,-0.1383069307,0.1946542859,-0.3773559034,0.0320957825,-0.111780636,0.1643281281,0.1397599727,-0.1000155509,-0.0343191288,0.2832554877,-0.3728188276,-0.0389520638,0.262173295,0.4476927817,0.0322523899,-0.2558436394,0.2842577696,-0.2356747538]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2914","title":"Having a dependency defining fsspec entrypoint raises an AttributeError when importing datasets","comments":"Closed by #2915.","body":"## Describe the bug\r\nIn one of my project, I defined a custom fsspec filesystem with an entrypoint.\r\nMy guess is that by doing so, a variable named `spec` is created in the module `fsspec` (created by entering a for loop as there are entrypoints defined, see the loop in question [here](https:\/\/github.com\/intake\/filesystem_spec\/blob\/0589358d8a029ed6b60d031018f52be2eb721291\/fsspec\/__init__.py#L55)).\r\nSo that `fsspec.spec`, that was previously referring to the `spec` submodule, is now referring to that `spec` variable.\r\nThis make the import of datasets failing as it is using that `fsspec.spec`.\r\n\r\n## Steps to reproduce the bug\r\nI could reproduce the bug with a dummy poetry project.\r\n\r\nHere is the pyproject.toml:\r\n```toml\r\n[tool.poetry]\r\nname = \"debug-datasets\"\r\nversion = \"0.1.0\"\r\ndescription = \"\"\r\nauthors = [\"Pierre Godard\"]\r\n\r\n[tool.poetry.dependencies]\r\npython = \"^3.8\"\r\ndatasets = \"^1.11.0\"\r\n\r\n[tool.poetry.dev-dependencies]\r\n\r\n[build-system]\r\nrequires = [\"poetry-core>=1.0.0\"]\r\nbuild-backend = \"poetry.core.masonry.api\"\r\n\r\n[tool.poetry.plugins.\"fsspec.specs\"]\r\n\"file2\" = \"fsspec.implementations.local.LocalFileSystem\"\r\n```\r\n\r\nThe only other file being a `debug_datasets\/__init__.py` empty file.\r\n\r\nThe overall structure of the project is as follows:\r\n```\r\n.\r\n\u251c\u2500\u2500 pyproject.toml\r\n\u2514\u2500\u2500 debug_datasets\r\n \u2514\u2500\u2500 __init__.py\r\n```\r\n\r\nThen, within the project folder run:\r\n\r\n```\r\npoetry install\r\npoetry run python\r\n```\r\n\r\nAnd in the python interpreter, try to import `datasets`:\r\n\r\n```\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nThe import should run successfully.\r\n\r\n## Actual results\r\n\r\nHere is the trace of the error I get:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/__init__.py\", line 33, in \r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 48, in \r\n from .filesystems import extract_path_from_uri, is_remote_filesystem\r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/filesystems\/__init__.py\", line 30, in \r\n def is_remote_filesystem(fs: fsspec.spec.AbstractFileSystem) -> bool:\r\nAttributeError: 'EntryPoint' object has no attribute 'AbstractFileSystem'\r\n```\r\n\r\n## Suggested fix\r\n\r\n`datasets\/filesystems\/__init__.py`, line 30, replace:\r\n```\r\n def is_remote_filesystem(fs: fsspec.spec.AbstractFileSystem) -> bool:\r\n```\r\nby:\r\n```\r\n def is_remote_filesystem(fs: fsspec.AbstractFileSystem) -> bool:\r\n```\r\n\r\nI will come up with a PR soon if this effectively solves the issue.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: WSL2 (Ubuntu 20.04.1 LTS)\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n- `fsspec` version: 2021.8.1\r\n","comment_length":3,"text":"Having a dependency defining fsspec entrypoint raises an AttributeError when importing datasets\n## Describe the bug\r\nIn one of my project, I defined a custom fsspec filesystem with an entrypoint.\r\nMy guess is that by doing so, a variable named `spec` is created in the module `fsspec` (created by entering a for loop as there are entrypoints defined, see the loop in question [here](https:\/\/github.com\/intake\/filesystem_spec\/blob\/0589358d8a029ed6b60d031018f52be2eb721291\/fsspec\/__init__.py#L55)).\r\nSo that `fsspec.spec`, that was previously referring to the `spec` submodule, is now referring to that `spec` variable.\r\nThis make the import of datasets failing as it is using that `fsspec.spec`.\r\n\r\n## Steps to reproduce the bug\r\nI could reproduce the bug with a dummy poetry project.\r\n\r\nHere is the pyproject.toml:\r\n```toml\r\n[tool.poetry]\r\nname = \"debug-datasets\"\r\nversion = \"0.1.0\"\r\ndescription = \"\"\r\nauthors = [\"Pierre Godard\"]\r\n\r\n[tool.poetry.dependencies]\r\npython = \"^3.8\"\r\ndatasets = \"^1.11.0\"\r\n\r\n[tool.poetry.dev-dependencies]\r\n\r\n[build-system]\r\nrequires = [\"poetry-core>=1.0.0\"]\r\nbuild-backend = \"poetry.core.masonry.api\"\r\n\r\n[tool.poetry.plugins.\"fsspec.specs\"]\r\n\"file2\" = \"fsspec.implementations.local.LocalFileSystem\"\r\n```\r\n\r\nThe only other file being a `debug_datasets\/__init__.py` empty file.\r\n\r\nThe overall structure of the project is as follows:\r\n```\r\n.\r\n\u251c\u2500\u2500 pyproject.toml\r\n\u2514\u2500\u2500 debug_datasets\r\n \u2514\u2500\u2500 __init__.py\r\n```\r\n\r\nThen, within the project folder run:\r\n\r\n```\r\npoetry install\r\npoetry run python\r\n```\r\n\r\nAnd in the python interpreter, try to import `datasets`:\r\n\r\n```\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nThe import should run successfully.\r\n\r\n## Actual results\r\n\r\nHere is the trace of the error I get:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/__init__.py\", line 33, in \r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 48, in \r\n from .filesystems import extract_path_from_uri, is_remote_filesystem\r\n File \"\/home\/godarpi\/.cache\/pypoetry\/virtualenvs\/debug-datasets-JuFzTKL--py3.8\/lib\/python3.8\/site-packages\/datasets\/filesystems\/__init__.py\", line 30, in \r\n def is_remote_filesystem(fs: fsspec.spec.AbstractFileSystem) -> bool:\r\nAttributeError: 'EntryPoint' object has no attribute 'AbstractFileSystem'\r\n```\r\n\r\n## Suggested fix\r\n\r\n`datasets\/filesystems\/__init__.py`, line 30, replace:\r\n```\r\n def is_remote_filesystem(fs: fsspec.spec.AbstractFileSystem) -> bool:\r\n```\r\nby:\r\n```\r\n def is_remote_filesystem(fs: fsspec.AbstractFileSystem) -> bool:\r\n```\r\n\r\nI will come up with a PR soon if this effectively solves the issue.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: WSL2 (Ubuntu 20.04.1 LTS)\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n- `fsspec` version: 2021.8.1\r\n\nClosed by #2915.","embeddings":[-0.0430343486,0.1729615629,-0.0092811715,0.3249996006,0.3824198246,-0.027578136,0.3696733415,-0.0281330235,0.0717082471,0.1155877039,0.2619345784,0.4130678475,-0.0622646138,0.0250749756,-0.1028326377,0.2277256399,0.0461320765,0.1964916587,-0.1118648723,0.0814422593,-0.2052113712,0.284779042,-0.20766671,0.1609047055,-0.3610854745,0.1328371316,0.0609395057,0.3386956155,-0.1826843172,-0.6150017977,0.3787585497,-0.1575102955,-0.0890944004,0.1836814284,-0.0001192106,0.0413611904,0.1138052121,-0.081654422,-0.2293013632,0.0555289909,-0.5713313818,-0.1993799657,0.1814039797,-0.2640390098,0.4023723602,-0.2299601138,-0.0492568053,-0.720643878,-0.0350703783,0.4295676649,0.151214689,-0.0737293884,0.1342175305,-0.078581512,0.1368629038,0.0073615429,-0.2163085043,0.0665525422,-0.1590744555,-0.1137120426,-0.0247463994,0.028592892,-0.073725529,0.2388720959,0.7355163693,-0.1537969112,0.4934940934,-0.2031433433,-0.0418940075,0.1256616861,0.4556665123,-0.4343681931,-0.3134965301,-0.1376075,0.0694526061,-0.099931106,0.2053408474,-0.0052408343,0.0365756415,0.1147769988,0.4034270048,-0.0220960025,-0.1103871539,0.0099158697,-0.1579840034,-0.1949465275,-0.2566277087,0.1149913594,-0.0176931862,-0.1435348243,-0.293458432,-0.1421632171,-0.1392901242,0.0056826379,-0.3321452439,0.0178288762,0.2203202993,-0.1728139669,-0.2198076099,0.2929090858,-0.0524715595,0.0291536804,0.1376225799,0.3627481163,0.1954966486,0.1123554483,0.3818732202,0.0941143036,0.2671805918,0.0105761532,-0.1745803803,-0.0067355181,-0.0322308168,-0.4421686828,0.196050778,-0.0149847884,0.7509050965,-0.3278894126,-0.3549762964,0.2934402525,-0.2196975499,0.0935777649,-0.0379396789,0.2039052695,0.2131644487,0.5074154139,0.110865429,0.1131861806,-0.2682110369,0.0728461221,-0.0765823871,0.0951074734,-0.1301978678,0.1238338947,-0.0917826369,-0.0535966307,0.3636643291,-0.0692519695,0.087492682,-0.2760293186,0.0895308852,-0.1705299467,-0.2756180465,0.268853128,-0.4858374298,0.0634655356,0.0485958643,-0.1812145114,-0.1407059431,-0.0891787559,-0.223374173,-0.2241652012,-0.2036291659,0.1713668406,-0.227433145,0.0364288799,-0.2167210281,-0.2242526561,0.2060965598,-0.0521678776,0.1591060758,-0.2531750202,-0.255607456,-0.1615860462,-0.0356961787,0.3934701979,-0.2941275835,-0.1991277188,-0.1428540945,-0.2064636052,-0.1786636561,-0.1581739336,-0.1267938614,0.4374836683,-0.2819532752,0.1414128244,0.2326729596,-0.4089822769,-0.1433600485,0.2878492177,0.0954126939,-0.1979203522,0.1566754282,-0.1284318268,0.1135182604,-0.1574372202,0.1001495048,0.0776153803,0.1633862257,0.007081199,-0.0188464597,0.036375124,-0.1544718444,-0.0118767954,-0.2443443686,0.0282867458,0.1455506384,0.1295337081,-0.1325169504,-0.1437395513,0.102992475,0.446297735,0.5053649545,0.2926414013,0.0018090364,-0.3757137656,-0.1410629153,0.2792656124,0.0050844387,-0.0705011711,-0.0173224658,-0.0369233266,-0.2272954136,0.4188956916,-0.0087850615,-0.3483618796,-0.0021392696,0.1587059945,-0.2076747417,-0.1167936027,-0.2682749331,0.0590893626,-0.1198496372,0.0904613882,-0.2748877108,0.0908167139,-0.0701375529,-0.0042610434,-0.0235086512,0.3459734023,0.1507644355,0.122721754,-0.1112476885,0.3404074907,0.5937909484,0.2090201527,-0.0751314461,0.3819218576,0.1279350668,-0.2491389513,-0.0412581488,0.1454391629,0.2145802826,-0.0491112433,0.4379881918,0.2754839659,0.2406787723,0.3727573156,-0.0556240045,0.1264895946,0.242703557,-0.1039064452,-0.1578007489,-0.2713132203,0.1646561325,-0.0192193184,0.4520537555,0.2100941837,0.0325967483,-0.0543766543,-0.0089729214,0.0672252774,0.0164770391,-0.0364091173,-0.2857769132,0.0901333243,0.2991757691,0.056403894,0.7109383345,0.0651810989,-0.3750181198,0.1294146478,-0.020994015,-0.0924102291,0.232258141,0.1908759773,0.1215177178,0.1294934303,0.2365701497,-0.1311048567,0.0034811324,-0.2910775244,-0.0515734069,0.1591396928,-0.7218433022,0.3073238134,-0.4488064051,0.0365511924,-0.3674746156,-0.1099792644,0.039912153,-0.3449760377,-0.1204983741,0.0795523301,-0.1346967071,0.1161788553,-0.3224466741,0.0531995259,-0.0138493739,-0.4038999677,0.0998631865,-0.1038146392,-0.0749545097,-0.0629806891,0.2017032653,0.4358921051,0.1518811285,0.2062558383,-0.0725508034,-0.2794296741,-0.1373124123,0.0227615349,-0.0418572798,0.7079858184,0.4984821379,-0.0236271117,0.334713608,-0.052583456,-0.022778647,-0.0508446507,-0.1497538984,0.1021313444,-0.0401481725,-0.159014374,-0.1050804704,-0.2614705265,0.0915697217,-0.2867671847,-0.0876169056,-0.0462019704,0.0119425636,0.4248128235,0.1451478451,0.4045563936,0.2662217617,-0.0025793067,-0.0106500452,0.1442047656,0.1359658688,-0.1538913995,-0.2444033176,0.2343038619,-0.2083691955,-0.0294864997,-0.0952343643,-0.4535905719,-0.25282076,-0.1953865737,0.3757030666,-0.0514751524,-0.1112330109,0.4003503621,0.165106073,-0.0839413479,-0.3391038477,-0.2688634694,0.0547733977,0.0875356421,-0.0184010845,0.1420247257,-0.1215976626,-0.1658104211,0.374792397,0.2599062324,-0.1457856596,0.2739960253,-0.1751101613,0.2564560473,0.1007364541,-0.3438697755,-0.0130711058,-0.085844934,-0.2883581519,0.1482560337,0.0481070541,0.2359811664,-0.0565045625,0.2738620937,0.261869669,-0.2000104636,-0.2162149996,-0.2536721528,0.3697989285,-0.0000891886,0.1281913221,0.1720805168,0.223276034,0.0398766212,0.3417465091,-0.0693826824,-0.1782999188,-0.2556450963,0.2239852697,-0.149227649,0.1481966078,0.118940182,0.2389065027,0.0841452926,0.0368388407,0.0327930972,-0.0729915872,0.107063055,-0.0424628891,0.0199763998,0.2870316803,0.0729298294,-0.6140675545,0.013722131,-0.0617609844,0.2754461467,0.1362575889,0.1719761193,-0.2309075296,-0.0642381907,0.0313319452,0.1695097834,-0.0079485681,0.0463472158,-0.3716210723,-0.3436375558,-0.3559315503,-0.3648687899,-0.1220872253,0.1460855603,0.608284533,0.4415367246,-0.2861590385,-0.1434668005,0.1396352649,-0.0774980411,-0.0251135901,-0.3337855041,0.2506456077,0.0822203383,-0.2227955461,0.438881278,0.7413416505,-0.0107491156,-0.6289779544,0.0093631083,0.1382879615,0.2119753063,0.3244391382,-0.2618112862,0.1188502163,-0.363063544,-0.1881043762,-0.3619200289,0.1122573316,0.260430187,0.2203682214,0.0700321048,-0.4762194157,0.1530032158,-0.1998498738,-0.0410767794,0.3159738183,0.1726385653,-0.2884833515,0.2569683492,0.0741914511,0.7808514237,0.569355309,-0.1249206737,0.3599708676,-0.2424308062,-0.0542152859,-0.1202839613,-0.1542701572,-0.2387532294,0.1913201809,-0.0544839613,-0.18723239,0.705799818,-0.1415283531,-0.1244052649,0.1290933788,0.1822909117,-0.1393857598,0.1165291592,0.1594590992,0.0814604759,-0.4633861482,-0.370397687,0.1029222086,-0.0320705138,0.0683007315,-0.0650951117,-0.0247276314,-0.0777822435,-0.2294555604,-0.32167539,0.0439463444,-0.4455278814,0.3251912594,0.1124332696,0.1910561025,0.2862206399,0.3550289273,-0.0675555989,0.1610818058,-0.1015330851,0.2352689207,0.1239409521,-0.0098448126,0.0602400079,-0.227035746,0.1665501446,-0.0625454262,-0.087164484,-0.1301974505,-0.1741476953,-0.1828138232,0.1021620706,0.0863435939,-0.1087821126,-0.3041732013,-0.1606344432,0.0950354114,0.1120003834,-0.221758917,0.0907387882,0.0084382929,-0.1957605779,0.1542771906,-0.1038391814,-0.3994420469,0.0711361542,0.2759751081,-0.1620726883,0.1611830741,0.381315589,-0.1241321787,-0.1759859473,-0.1711550504,0.4069006741,-0.2283793688,-0.2342585176,0.094457373,0.0810037404,-0.2219625711,-0.120175615,0.5494086742,-0.0067346175,-0.0117426496,0.0497230366,-0.4125763774,-0.0088035138,0.0351061262,-0.2378330827,0.0009349291,-0.0294312052,0.0923814923,-0.3538710773,0.3067423701,-0.2256021351,0.2494763881,-0.0625165701,0.1239562109,0.4566327929,0.0589139163,0.1883669645,-0.272287935,0.0895297304,0.2248979509,-0.2632036507,-0.0840326101,-0.3775589168,0.1614944786,0.0968005434,-0.0412362032,0.1767305434,-0.0410813652,-0.0104668569,-0.1951070875,-0.2089901417,0.4560409784,-0.3352125585,0.2502237558,-0.1950406581,0.1619997472,-0.1527923942,0.1199247018,0.0177820474,0.3442685306,-0.0704655424,0.2242517322,-0.1402976066,0.0524093583,-0.1809961349,0.286863625,-0.2309779823,0.001652966,0.0643716156,-0.2714308202,0.0075101843,0.1741512418,0.051411096,0.3920331597,-0.3538575768,0.2180334181,0.1009309292,0.0863336474,-0.2097055018,0.0577151366,0.2425736338,-0.2602337599,0.0286733247,0.32045421,0.3694347143,-0.1328757405,-0.1426737905,0.1771979332,-0.1642621458,0.1670656502,0.2125457823,0.0809638053,-0.1030210778,-0.0222702622,0.3884674013,0.1377125829,0.2725866735,0.6721334457,-0.0822553411,0.4087929726,0.3187718987,0.0578281432,-0.0882372782,-0.3026813567,0.6395866871,0.0971084386,0.0903770179,0.0212999135,0.1801173836,0.3370145261,-0.2536561191,-0.0584542826,0.1592061371,-0.0090570366,-0.0695601627,-0.1169230044,0.2838150859,0.0470363572,0.0753637478,-0.1204745993,0.099752076,-0.2900804579,0.4320870042,-0.0070109242,-0.2397236079,-0.4791888297,0.2327571511,0.0517540872,0.2455652803,-0.3497766256,0.0935187489,-0.2294338495,0.009274709,0.2234476656,0.2447924465,0.6747469306,0.2802542746,0.0613406897,-0.0361753739,-0.1880470365,-0.0251632184,0.1080025136,0.0423466004,-0.4021561146,0.4986065328,0.3744322956,0.0134312417,-0.1475075185,-0.3044832051,0.2202391624,0.3330422342,-0.5749848485,0.4656461477,-0.1772824079,0.2455034703,-0.0323761702,0.301497072,-0.2611684501,-0.3874022961,0.4785009027,-0.0177927166,0.0497761108,-0.1358003616,0.0600086786,-0.1562406123,0.4533146918,0.2446237355,0.0289993435,-0.0898260772,0.0194680337,-0.7921037078,0.09198457,-0.0585266203,-0.1397465765,0.0542413443,0.132478714,0.0207924377,0.3848239481,-0.055609718,0.0733580142,0.3911984265,0.2168242484,-0.3189651668,-0.1849366575,0.1098671705,-0.0723528266,0.1301362216,-0.5238112211,-0.1766480356,0.0221980996,0.0278460272,0.0507279634,-0.2994850278,0.2058833688,-0.1167204082,0.4986041486,-0.0291798674,0.5323891044,-0.144431591,-0.0973311812,-0.1218846962,-0.4071162939,-0.2981791794,0.0227600001,-0.027476877,0.1542173773,-0.233344689,-0.0661291555,-0.3824612796,0.4765124917,0.0400805809,-0.0109937387,-0.3236443698,0.3815689981,-0.4507586658,0.1908286214,-0.4913522601,0.1234193966,-0.0205241553,0.3339468837,-0.2606340647,-0.5488712192,0.4859133363,-0.534476161,-0.1120796576,0.1190102845,0.4382752478,0.0190124456,-0.2618913352,-0.7699956894,0.0116949342,0.3630550206,-0.0525717586,-0.2893211842,0.0585939176,-0.398167491,0.0147053022,-0.050671231,0.4721556902,-0.1411389709,0.0548643433,0.3597792387,-0.139134869]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2913","title":"timit_asr dataset only includes one text phrase","comments":"Hi @margotwagner, \r\nThis bug was fixed in #1995. Upgrading the datasets should work (min v1.8.0 ideally)","body":"## Describe the bug\r\nThe dataset 'timit_asr' only includes one text phrase. It only includes the transcription \"Would such an act of refusal be useful?\" multiple times rather than different phrases.\r\n\r\n## Steps to reproduce the bug\r\nNote: I am following the tutorial https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\n1. Install the dataset and other packages\r\n```python\r\n!pip install datasets>=1.5.0\r\n!pip install transformers==4.4.0\r\n!pip install soundfile\r\n!pip install jiwer\r\n```\r\n2. Load the dataset\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n```\r\n3. Remove columns that we don't want\r\n```python\r\ntimit = timit.remove_columns([\"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \"sentence_type\", \"speaker_id\"])\r\n```\r\n4. Write a short function to display some random samples of the dataset.\r\n```python\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n \r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\nshow_random_elements(timit[\"train\"].remove_columns([\"file\"]))\r\n```\r\n\r\n## Expected results\r\n10 random different transcription phrases.\r\n\r\n## Actual results\r\n10 of the same transcription phrase \"Would such an act of refusal be useful?\"\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.4.1\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: not listed\r\n","comment_length":16,"text":"timit_asr dataset only includes one text phrase\n## Describe the bug\r\nThe dataset 'timit_asr' only includes one text phrase. It only includes the transcription \"Would such an act of refusal be useful?\" multiple times rather than different phrases.\r\n\r\n## Steps to reproduce the bug\r\nNote: I am following the tutorial https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\n1. Install the dataset and other packages\r\n```python\r\n!pip install datasets>=1.5.0\r\n!pip install transformers==4.4.0\r\n!pip install soundfile\r\n!pip install jiwer\r\n```\r\n2. Load the dataset\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n```\r\n3. Remove columns that we don't want\r\n```python\r\ntimit = timit.remove_columns([\"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \"sentence_type\", \"speaker_id\"])\r\n```\r\n4. Write a short function to display some random samples of the dataset.\r\n```python\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n \r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\nshow_random_elements(timit[\"train\"].remove_columns([\"file\"]))\r\n```\r\n\r\n## Expected results\r\n10 random different transcription phrases.\r\n\r\n## Actual results\r\n10 of the same transcription phrase \"Would such an act of refusal be useful?\"\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.4.1\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: not listed\r\n\nHi @margotwagner, \r\nThis bug was fixed in #1995. Upgrading the datasets should work (min v1.8.0 ideally)","embeddings":[0.161290288,-0.138053894,-0.0205535926,0.2273859978,0.1176241562,-0.1575979143,0.2670057416,0.2965668142,-0.5154975653,0.2804078162,0.1454417259,0.4508609474,-0.0935282409,-0.1810210198,0.0642997622,-0.0575755686,0.1221427992,0.2484459728,-0.0715344921,-0.2342915833,0.1333340406,0.3007956445,-0.257455945,-0.064661555,-0.2737528682,0.0847610086,-0.0766698644,-0.2693378031,-0.0532864109,-0.4953400791,0.1539394706,0.011122671,0.1049450785,0.3338848054,-0.0001155144,-0.0851756856,0.0369373672,0.0938122272,-0.2382418513,-0.157761991,-0.0969283953,0.0886226818,-0.0698391646,-0.0270568896,-0.0834104121,0.1937476546,0.0529345647,-0.0981618613,0.5147064328,0.2915839553,0.1400677115,-0.0102435909,-0.252120316,-0.0659528822,0.3198694587,0.0313076191,0.0850752518,-0.1780210286,0.2137241364,0.0314790495,-0.0516679287,0.6180685163,-0.2813329399,0.3980180621,-0.2359145284,0.1242306605,-0.3711774647,-0.5308332443,0.266836673,0.2161651254,0.6465566754,-0.2296243757,-0.2044666409,-0.2378615439,0.3199083507,-0.1560648084,-0.0421698391,-0.0161785055,-0.2728021145,0.2350625694,-0.0094950655,0.047647696,-0.2204018831,0.1288805455,0.055811543,-0.1074196771,-0.0731363669,0.1385569125,-0.177100122,0.0994098634,-0.085863933,0.2703583241,-0.1442609429,0.0341251902,-0.149762243,-0.0674901083,0.1925161034,-0.3049418628,0.2917183042,-0.0602199472,0.3855580986,0.0374377258,-0.1856715977,-0.0881002173,0.3239292204,0.019808244,-0.0233677495,0.1496766508,0.2298973352,-0.2340689301,-0.2531856596,0.0737158284,-0.0144180572,0.2162737995,0.331879437,-0.265029937,0.2829249203,-0.367644608,-0.5658828616,0.0642497912,-0.4400376976,-0.1030385345,-0.2528013289,0.1816738844,0.1693309546,0.3009272516,0.1449253261,0.2226763219,0.0693393499,-0.5628325343,-0.1818621904,0.0267980024,0.2098604441,0.0308408234,0.211329326,-0.3237065375,0.3398516774,0.3417337835,0.195452705,-0.4315991104,-0.2874112725,-0.1301766485,0.0255610328,0.0091858534,-0.0264903046,0.3916681111,0.0738202855,-0.0182900447,0.0072356486,0.0610072762,-0.106556423,0.0068646008,0.2134016603,0.1364102364,-0.0805880576,-0.167114675,0.1232486665,0.3605877757,0.0779450312,-0.3008698225,0.0030109661,-0.1827422976,-0.3919641674,-0.0386591218,0.0773314983,0.3024609089,-0.3355484903,0.0954273939,0.178534463,0.4213981032,0.3944750428,0.2484540939,-0.0465098992,0.3734385073,-0.1168166548,0.3001444042,0.2041984349,-0.4621309042,-0.3920545578,0.2112352401,-0.049603641,0.2464142591,0.1548911184,-0.1627833694,0.3027338982,0.1427075267,0.3759355247,0.153478384,0.1457359195,0.0334793814,-0.2446887195,-0.057892438,0.0868306831,-0.0131406998,0.0021537091,0.0955836996,0.0370380655,0.232970655,0.5271517038,0.0745838359,0.0924516246,0.1029673368,0.1659648567,-0.1012314931,0.2239779383,-0.2631005943,0.2564861178,-0.0488976687,0.2068272382,0.2837985456,0.1540315896,0.0793671757,-0.3015450537,-0.2188883722,-0.2991884053,-0.3250726461,0.1145813912,0.3257373571,-0.0144317094,0.0420466512,-0.1898579001,0.3409897685,-0.2203264982,-0.0741857439,0.0233059432,0.0991652608,0.0638567135,-0.2447074801,0.5248855948,0.2528992593,0.18514961,0.2016844004,-0.1669406295,0.22355178,-0.0147676235,0.0219675135,-0.2637168467,-0.3288924992,0.1099477187,-0.5017575622,-0.1113948897,0.2344709635,0.3181717396,-0.0044999369,-0.3036757708,-0.1834709942,0.1390077174,0.2569387853,-0.0554442666,-0.0651190057,0.3011586368,0.0662527978,-0.3537996113,-0.3246975541,0.2526897192,-0.234980613,0.2662216723,0.0202848148,-0.3417770565,0.1585986316,0.3928949833,0.2245113999,0.085208714,0.2943886817,-0.2468326539,-0.027017273,-0.1227289811,0.0949118286,0.2907651365,0.2881050408,0.0543022901,-0.3764547706,-0.1287488788,-0.1132321432,0.1393357515,-0.1307561845,-0.3556297719,0.4351170063,0.3585450351,0.0863876268,-0.4493839145,-0.1341284662,-0.1424040794,0.4703677297,-0.3349600732,-0.2316320539,-0.4356057346,-0.1205524206,-0.109005712,0.0662912875,0.0797044486,0.225550577,0.2190566808,-0.0127121434,0.0054567931,0.1367991865,0.1719908118,0.2340619862,0.0712051392,0.1117983311,0.086305052,0.3191404045,-0.306956917,-0.0242702272,0.0259727612,0.271612525,0.1074295118,-0.1422471404,-0.1639021039,-0.4067965448,0.0590180866,0.2675643563,-0.3301686049,0.5113489628,0.1133633777,0.1602485478,-0.2251659632,-0.3130813539,0.0026724054,-0.0268870983,-0.1686861217,0.3555683196,0.0829433203,-0.0056142341,-0.0482558794,-0.698974669,0.099392809,-0.2500089407,0.2359433323,-0.0328674652,0.2103785425,0.3641358316,0.0606071204,0.0812700838,-0.0587117635,0.2053699046,-0.410209775,-0.2545935214,0.3561107218,0.2606044412,-0.3348694146,-0.3405695558,0.0509579033,0.2652952075,-0.0322290994,-0.4001087248,0.1003541723,-0.2674658895,0.1453480273,-0.2972603142,0.2826198339,0.3854675591,0.0129939336,-0.1495047212,-0.1692215204,-0.1078334898,0.2658356428,0.3553734422,0.0058843573,0.1011472642,0.0699137822,-0.0593266487,0.3506978154,0.6459594965,0.2025162727,0.2600066662,-0.1765539348,0.0942354053,-0.0324934721,-0.3995798528,0.0270797946,-0.1173247173,-0.062567167,0.4981202781,-0.0899957195,0.0536311306,-0.1695882976,-0.0572774038,-0.3290627599,-0.1686403602,0.0674927011,-0.2359398454,0.1839738041,0.1747753173,0.0528291985,0.1619518846,0.0347351544,-0.2861667871,-0.2988380492,0.2851738334,-0.1499728858,-0.0820301399,-0.2836113274,-0.377288729,-0.1241017282,0.1385834366,0.5095545053,-0.2117005289,-0.037520647,0.3600184023,0.291478008,0.3810289502,-0.3069941998,0.0908769667,0.0103903497,0.2287908047,0.0055077169,0.0054613417,-0.5205008984,-0.1967586279,-0.0475005209,-0.043032378,-0.1668928564,-0.0080181081,0.108959198,0.3927445412,-0.0097480956,-0.2408112288,-0.2557696104,-0.0757930055,-0.3301421404,0.1615659893,0.3422310054,0.0443936288,-0.1651335955,-0.1203253195,0.1603984088,-0.0627515167,0.5332366228,0.0798594728,0.375176698,0.0739652514,0.3011747897,0.2748407722,-0.3577797413,-0.2591435611,0.3608151674,0.100969635,-0.1420252174,-0.00367438,-0.3468615413,0.3289579153,0.4931568503,-0.223191008,0.2419292778,-0.2970614135,0.2568941116,-0.1529492587,0.0957464501,0.3195284307,-0.0749837682,-0.0751821995,-0.410286963,0.3822805583,-0.0364946276,-0.1979487538,0.3445512056,0.0909191668,-0.3103821576,0.4712514579,-0.2578842938,0.8118340373,0.2264072746,0.3666303754,0.3155016303,0.1381517202,0.3261640966,-0.1960099638,0.0902330428,-0.0220177118,-0.0018170826,-0.0215603653,0.0583430231,-0.0778571218,-0.0417594425,-0.5769283772,0.1775290817,0.0647163838,0.0288711693,0.1868025362,-0.0118521098,-0.1294962317,-0.1064071804,0.1252996773,0.0325710885,-0.3839387298,0.3074729145,0.3099324405,-0.0242960695,0.0314041749,-0.5837308764,-0.8224813938,-0.0301681384,-0.0825177506,-0.2259411067,-0.1996800154,-0.3336216509,0.3681984842,0.2585997581,0.1915538907,0.3256590664,-0.1379685849,-0.0852642953,0.3682431877,0.2776513696,-0.2679565549,-0.0255918875,0.1445183456,-0.0860315636,-0.3883327842,0.1766039878,0.0551586859,0.0989629775,-0.3880658448,0.2538315058,-0.1146805435,-0.2331590801,0.0857373253,0.4289047122,0.1415170133,-0.5119704008,0.10320732,0.1241723523,-0.4588638544,0.3286851645,0.1725837886,-0.4755195975,-0.0637625083,0.259340167,0.1283604354,0.0002506027,0.6646648645,-0.0414828882,0.0483881421,-0.37047562,0.0406970792,-0.0209288169,-0.2641952336,0.1791121364,0.0586630926,0.1739625186,-0.2265277505,0.2535494268,0.0265137423,0.204802826,-0.2814756632,-0.3617306948,-0.0619779229,-0.0314538814,0.1234360188,0.0552398637,-0.2911967039,0.1181857958,0.0688160807,0.3186088502,-0.2981917858,0.0059745917,-0.0517939329,0.2173203677,-0.0840429664,0.0813204795,-0.0378096215,0.0851734653,0.2000400573,0.3185155094,0.0672922209,-0.1598699093,-0.3062146604,0.1838445365,0.0417443737,0.1562646031,-0.040768452,0.100105688,0.1492390484,-0.1337077022,-0.1010111123,0.334582895,0.1253624856,0.2183783352,-0.0296323262,0.0370717831,0.1851567924,-0.1598648131,-0.1839772761,-0.172459051,0.3399853706,0.1806136072,0.0487441346,-0.0481960177,0.1186248809,0.0399239436,-0.1340517849,0.1186952814,-0.0778344646,-0.2655762732,0.1333703995,0.2876859903,-0.1194795594,0.3786883652,-0.3459939361,-0.1118993759,0.091195181,0.235143885,-0.2654986084,-0.1237113848,-0.3232501447,-0.1065506563,-0.0583089553,0.2784046829,-0.0598313436,-0.4723803997,0.121467337,-0.0649383292,0.336020112,-0.5360029936,-0.0734213144,0.3985144794,0.0149740381,-0.3611198664,-0.1933412552,-0.0151967062,0.0208013449,0.5450986028,-0.3675437868,0.1373579353,-0.3386060596,0.0263877269,0.191548273,0.0256050285,-0.0382200703,0.4056419134,-0.3271881938,-0.0489630923,-0.0716247186,0.3050477505,-0.1176727563,0.0334460959,-0.1280483007,-0.0514146574,-0.1239193901,-0.1248619258,-0.3380397856,-0.2734287083,-0.2303761989,0.0061822413,0.0975418091,-0.0125555322,0.1405136287,0.1916218102,-0.2668848932,-0.5332419276,-0.1980937272,-0.0377839021,0.1720763147,-0.3900322318,0.3403875828,0.3084276021,-0.221769765,0.2584382296,0.2433644086,0.519982934,0.0715812668,0.0334475301,0.1065192446,-0.5706885457,-0.0919026062,-0.0517226532,0.3297682703,-0.1242420226,0.0587560534,0.0218634959,0.0158726722,-0.095639661,0.0293477289,0.2285464853,-0.0756259263,-0.563256681,0.2190777361,-0.1411624253,0.1117327288,0.0406589434,0.0749845207,-0.4736695588,0.4417389929,0.3421675265,0.0575963221,-0.0106266811,0.0021905343,0.0813886747,-0.1263748854,-0.1126130521,0.2673044801,0.1579869688,-0.2770343423,-0.1591356099,-0.6735790372,-0.071482718,-0.0488048792,-0.0049527902,0.0006292545,0.1189806312,0.1586837918,0.1362603009,0.0692648441,-0.0260408036,-0.5322327018,0.3291485906,-0.215462625,-0.0053817439,-0.0877875164,0.4366607666,0.0414892845,-0.1064981595,0.0108353356,0.094691515,0.0340188518,0.0655476525,0.0059462963,0.017562082,-0.1206619516,0.1582386047,0.1423452646,0.1708153635,-0.234853968,0.1895660013,-0.2832585871,0.0292929802,-0.1982437521,0.1561561525,0.2698151469,0.5158427358,-0.183451429,-0.0686847195,-0.0636009946,0.165361613,0.1112443656,-0.0649597496,-0.1914123297,0.089769423,-0.0087000672,0.2714968026,-0.2622262836,0.1059425324,0.0038290906,0.0026387305,-0.4920392334,-0.7109149694,0.4522462487,-0.4033147991,-0.5287759304,-0.0796565562,0.590672791,0.2403544337,-0.032153476,-0.7086960673,-0.0359898657,0.3959985971,0.2443489283,-0.3021709323,0.2129863203,0.1219399646,0.0085307239,0.0144249219,0.6226139069,0.3105249107,-0.2327545434,0.2679713964,-0.4604581892]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2913","title":"timit_asr dataset only includes one text phrase","comments":"Hi @margotwagner,\r\n\r\nYes, as @bhavitvyamalik has commented, this bug was fixed in `datasets` version 1.5.0. You need to update it, as your current version is 1.4.1:\r\n> Environment info\r\n> - `datasets` version: 1.4.1","body":"## Describe the bug\r\nThe dataset 'timit_asr' only includes one text phrase. It only includes the transcription \"Would such an act of refusal be useful?\" multiple times rather than different phrases.\r\n\r\n## Steps to reproduce the bug\r\nNote: I am following the tutorial https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\n1. Install the dataset and other packages\r\n```python\r\n!pip install datasets>=1.5.0\r\n!pip install transformers==4.4.0\r\n!pip install soundfile\r\n!pip install jiwer\r\n```\r\n2. Load the dataset\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n```\r\n3. Remove columns that we don't want\r\n```python\r\ntimit = timit.remove_columns([\"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \"sentence_type\", \"speaker_id\"])\r\n```\r\n4. Write a short function to display some random samples of the dataset.\r\n```python\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n \r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\nshow_random_elements(timit[\"train\"].remove_columns([\"file\"]))\r\n```\r\n\r\n## Expected results\r\n10 random different transcription phrases.\r\n\r\n## Actual results\r\n10 of the same transcription phrase \"Would such an act of refusal be useful?\"\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.4.1\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: not listed\r\n","comment_length":34,"text":"timit_asr dataset only includes one text phrase\n## Describe the bug\r\nThe dataset 'timit_asr' only includes one text phrase. It only includes the transcription \"Would such an act of refusal be useful?\" multiple times rather than different phrases.\r\n\r\n## Steps to reproduce the bug\r\nNote: I am following the tutorial https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\n1. Install the dataset and other packages\r\n```python\r\n!pip install datasets>=1.5.0\r\n!pip install transformers==4.4.0\r\n!pip install soundfile\r\n!pip install jiwer\r\n```\r\n2. Load the dataset\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n```\r\n3. Remove columns that we don't want\r\n```python\r\ntimit = timit.remove_columns([\"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \"sentence_type\", \"speaker_id\"])\r\n```\r\n4. Write a short function to display some random samples of the dataset.\r\n```python\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n \r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\nshow_random_elements(timit[\"train\"].remove_columns([\"file\"]))\r\n```\r\n\r\n## Expected results\r\n10 random different transcription phrases.\r\n\r\n## Actual results\r\n10 of the same transcription phrase \"Would such an act of refusal be useful?\"\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.4.1\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: not listed\r\n\nHi @margotwagner,\r\n\r\nYes, as @bhavitvyamalik has commented, this bug was fixed in `datasets` version 1.5.0. You need to update it, as your current version is 1.4.1:\r\n> Environment info\r\n> - `datasets` version: 1.4.1","embeddings":[0.161290288,-0.138053894,-0.0205535926,0.2273859978,0.1176241562,-0.1575979143,0.2670057416,0.2965668142,-0.5154975653,0.2804078162,0.1454417259,0.4508609474,-0.0935282409,-0.1810210198,0.0642997622,-0.0575755686,0.1221427992,0.2484459728,-0.0715344921,-0.2342915833,0.1333340406,0.3007956445,-0.257455945,-0.064661555,-0.2737528682,0.0847610086,-0.0766698644,-0.2693378031,-0.0532864109,-0.4953400791,0.1539394706,0.011122671,0.1049450785,0.3338848054,-0.0001155144,-0.0851756856,0.0369373672,0.0938122272,-0.2382418513,-0.157761991,-0.0969283953,0.0886226818,-0.0698391646,-0.0270568896,-0.0834104121,0.1937476546,0.0529345647,-0.0981618613,0.5147064328,0.2915839553,0.1400677115,-0.0102435909,-0.252120316,-0.0659528822,0.3198694587,0.0313076191,0.0850752518,-0.1780210286,0.2137241364,0.0314790495,-0.0516679287,0.6180685163,-0.2813329399,0.3980180621,-0.2359145284,0.1242306605,-0.3711774647,-0.5308332443,0.266836673,0.2161651254,0.6465566754,-0.2296243757,-0.2044666409,-0.2378615439,0.3199083507,-0.1560648084,-0.0421698391,-0.0161785055,-0.2728021145,0.2350625694,-0.0094950655,0.047647696,-0.2204018831,0.1288805455,0.055811543,-0.1074196771,-0.0731363669,0.1385569125,-0.177100122,0.0994098634,-0.085863933,0.2703583241,-0.1442609429,0.0341251902,-0.149762243,-0.0674901083,0.1925161034,-0.3049418628,0.2917183042,-0.0602199472,0.3855580986,0.0374377258,-0.1856715977,-0.0881002173,0.3239292204,0.019808244,-0.0233677495,0.1496766508,0.2298973352,-0.2340689301,-0.2531856596,0.0737158284,-0.0144180572,0.2162737995,0.331879437,-0.265029937,0.2829249203,-0.367644608,-0.5658828616,0.0642497912,-0.4400376976,-0.1030385345,-0.2528013289,0.1816738844,0.1693309546,0.3009272516,0.1449253261,0.2226763219,0.0693393499,-0.5628325343,-0.1818621904,0.0267980024,0.2098604441,0.0308408234,0.211329326,-0.3237065375,0.3398516774,0.3417337835,0.195452705,-0.4315991104,-0.2874112725,-0.1301766485,0.0255610328,0.0091858534,-0.0264903046,0.3916681111,0.0738202855,-0.0182900447,0.0072356486,0.0610072762,-0.106556423,0.0068646008,0.2134016603,0.1364102364,-0.0805880576,-0.167114675,0.1232486665,0.3605877757,0.0779450312,-0.3008698225,0.0030109661,-0.1827422976,-0.3919641674,-0.0386591218,0.0773314983,0.3024609089,-0.3355484903,0.0954273939,0.178534463,0.4213981032,0.3944750428,0.2484540939,-0.0465098992,0.3734385073,-0.1168166548,0.3001444042,0.2041984349,-0.4621309042,-0.3920545578,0.2112352401,-0.049603641,0.2464142591,0.1548911184,-0.1627833694,0.3027338982,0.1427075267,0.3759355247,0.153478384,0.1457359195,0.0334793814,-0.2446887195,-0.057892438,0.0868306831,-0.0131406998,0.0021537091,0.0955836996,0.0370380655,0.232970655,0.5271517038,0.0745838359,0.0924516246,0.1029673368,0.1659648567,-0.1012314931,0.2239779383,-0.2631005943,0.2564861178,-0.0488976687,0.2068272382,0.2837985456,0.1540315896,0.0793671757,-0.3015450537,-0.2188883722,-0.2991884053,-0.3250726461,0.1145813912,0.3257373571,-0.0144317094,0.0420466512,-0.1898579001,0.3409897685,-0.2203264982,-0.0741857439,0.0233059432,0.0991652608,0.0638567135,-0.2447074801,0.5248855948,0.2528992593,0.18514961,0.2016844004,-0.1669406295,0.22355178,-0.0147676235,0.0219675135,-0.2637168467,-0.3288924992,0.1099477187,-0.5017575622,-0.1113948897,0.2344709635,0.3181717396,-0.0044999369,-0.3036757708,-0.1834709942,0.1390077174,0.2569387853,-0.0554442666,-0.0651190057,0.3011586368,0.0662527978,-0.3537996113,-0.3246975541,0.2526897192,-0.234980613,0.2662216723,0.0202848148,-0.3417770565,0.1585986316,0.3928949833,0.2245113999,0.085208714,0.2943886817,-0.2468326539,-0.027017273,-0.1227289811,0.0949118286,0.2907651365,0.2881050408,0.0543022901,-0.3764547706,-0.1287488788,-0.1132321432,0.1393357515,-0.1307561845,-0.3556297719,0.4351170063,0.3585450351,0.0863876268,-0.4493839145,-0.1341284662,-0.1424040794,0.4703677297,-0.3349600732,-0.2316320539,-0.4356057346,-0.1205524206,-0.109005712,0.0662912875,0.0797044486,0.225550577,0.2190566808,-0.0127121434,0.0054567931,0.1367991865,0.1719908118,0.2340619862,0.0712051392,0.1117983311,0.086305052,0.3191404045,-0.306956917,-0.0242702272,0.0259727612,0.271612525,0.1074295118,-0.1422471404,-0.1639021039,-0.4067965448,0.0590180866,0.2675643563,-0.3301686049,0.5113489628,0.1133633777,0.1602485478,-0.2251659632,-0.3130813539,0.0026724054,-0.0268870983,-0.1686861217,0.3555683196,0.0829433203,-0.0056142341,-0.0482558794,-0.698974669,0.099392809,-0.2500089407,0.2359433323,-0.0328674652,0.2103785425,0.3641358316,0.0606071204,0.0812700838,-0.0587117635,0.2053699046,-0.410209775,-0.2545935214,0.3561107218,0.2606044412,-0.3348694146,-0.3405695558,0.0509579033,0.2652952075,-0.0322290994,-0.4001087248,0.1003541723,-0.2674658895,0.1453480273,-0.2972603142,0.2826198339,0.3854675591,0.0129939336,-0.1495047212,-0.1692215204,-0.1078334898,0.2658356428,0.3553734422,0.0058843573,0.1011472642,0.0699137822,-0.0593266487,0.3506978154,0.6459594965,0.2025162727,0.2600066662,-0.1765539348,0.0942354053,-0.0324934721,-0.3995798528,0.0270797946,-0.1173247173,-0.062567167,0.4981202781,-0.0899957195,0.0536311306,-0.1695882976,-0.0572774038,-0.3290627599,-0.1686403602,0.0674927011,-0.2359398454,0.1839738041,0.1747753173,0.0528291985,0.1619518846,0.0347351544,-0.2861667871,-0.2988380492,0.2851738334,-0.1499728858,-0.0820301399,-0.2836113274,-0.377288729,-0.1241017282,0.1385834366,0.5095545053,-0.2117005289,-0.037520647,0.3600184023,0.291478008,0.3810289502,-0.3069941998,0.0908769667,0.0103903497,0.2287908047,0.0055077169,0.0054613417,-0.5205008984,-0.1967586279,-0.0475005209,-0.043032378,-0.1668928564,-0.0080181081,0.108959198,0.3927445412,-0.0097480956,-0.2408112288,-0.2557696104,-0.0757930055,-0.3301421404,0.1615659893,0.3422310054,0.0443936288,-0.1651335955,-0.1203253195,0.1603984088,-0.0627515167,0.5332366228,0.0798594728,0.375176698,0.0739652514,0.3011747897,0.2748407722,-0.3577797413,-0.2591435611,0.3608151674,0.100969635,-0.1420252174,-0.00367438,-0.3468615413,0.3289579153,0.4931568503,-0.223191008,0.2419292778,-0.2970614135,0.2568941116,-0.1529492587,0.0957464501,0.3195284307,-0.0749837682,-0.0751821995,-0.410286963,0.3822805583,-0.0364946276,-0.1979487538,0.3445512056,0.0909191668,-0.3103821576,0.4712514579,-0.2578842938,0.8118340373,0.2264072746,0.3666303754,0.3155016303,0.1381517202,0.3261640966,-0.1960099638,0.0902330428,-0.0220177118,-0.0018170826,-0.0215603653,0.0583430231,-0.0778571218,-0.0417594425,-0.5769283772,0.1775290817,0.0647163838,0.0288711693,0.1868025362,-0.0118521098,-0.1294962317,-0.1064071804,0.1252996773,0.0325710885,-0.3839387298,0.3074729145,0.3099324405,-0.0242960695,0.0314041749,-0.5837308764,-0.8224813938,-0.0301681384,-0.0825177506,-0.2259411067,-0.1996800154,-0.3336216509,0.3681984842,0.2585997581,0.1915538907,0.3256590664,-0.1379685849,-0.0852642953,0.3682431877,0.2776513696,-0.2679565549,-0.0255918875,0.1445183456,-0.0860315636,-0.3883327842,0.1766039878,0.0551586859,0.0989629775,-0.3880658448,0.2538315058,-0.1146805435,-0.2331590801,0.0857373253,0.4289047122,0.1415170133,-0.5119704008,0.10320732,0.1241723523,-0.4588638544,0.3286851645,0.1725837886,-0.4755195975,-0.0637625083,0.259340167,0.1283604354,0.0002506027,0.6646648645,-0.0414828882,0.0483881421,-0.37047562,0.0406970792,-0.0209288169,-0.2641952336,0.1791121364,0.0586630926,0.1739625186,-0.2265277505,0.2535494268,0.0265137423,0.204802826,-0.2814756632,-0.3617306948,-0.0619779229,-0.0314538814,0.1234360188,0.0552398637,-0.2911967039,0.1181857958,0.0688160807,0.3186088502,-0.2981917858,0.0059745917,-0.0517939329,0.2173203677,-0.0840429664,0.0813204795,-0.0378096215,0.0851734653,0.2000400573,0.3185155094,0.0672922209,-0.1598699093,-0.3062146604,0.1838445365,0.0417443737,0.1562646031,-0.040768452,0.100105688,0.1492390484,-0.1337077022,-0.1010111123,0.334582895,0.1253624856,0.2183783352,-0.0296323262,0.0370717831,0.1851567924,-0.1598648131,-0.1839772761,-0.172459051,0.3399853706,0.1806136072,0.0487441346,-0.0481960177,0.1186248809,0.0399239436,-0.1340517849,0.1186952814,-0.0778344646,-0.2655762732,0.1333703995,0.2876859903,-0.1194795594,0.3786883652,-0.3459939361,-0.1118993759,0.091195181,0.235143885,-0.2654986084,-0.1237113848,-0.3232501447,-0.1065506563,-0.0583089553,0.2784046829,-0.0598313436,-0.4723803997,0.121467337,-0.0649383292,0.336020112,-0.5360029936,-0.0734213144,0.3985144794,0.0149740381,-0.3611198664,-0.1933412552,-0.0151967062,0.0208013449,0.5450986028,-0.3675437868,0.1373579353,-0.3386060596,0.0263877269,0.191548273,0.0256050285,-0.0382200703,0.4056419134,-0.3271881938,-0.0489630923,-0.0716247186,0.3050477505,-0.1176727563,0.0334460959,-0.1280483007,-0.0514146574,-0.1239193901,-0.1248619258,-0.3380397856,-0.2734287083,-0.2303761989,0.0061822413,0.0975418091,-0.0125555322,0.1405136287,0.1916218102,-0.2668848932,-0.5332419276,-0.1980937272,-0.0377839021,0.1720763147,-0.3900322318,0.3403875828,0.3084276021,-0.221769765,0.2584382296,0.2433644086,0.519982934,0.0715812668,0.0334475301,0.1065192446,-0.5706885457,-0.0919026062,-0.0517226532,0.3297682703,-0.1242420226,0.0587560534,0.0218634959,0.0158726722,-0.095639661,0.0293477289,0.2285464853,-0.0756259263,-0.563256681,0.2190777361,-0.1411624253,0.1117327288,0.0406589434,0.0749845207,-0.4736695588,0.4417389929,0.3421675265,0.0575963221,-0.0106266811,0.0021905343,0.0813886747,-0.1263748854,-0.1126130521,0.2673044801,0.1579869688,-0.2770343423,-0.1591356099,-0.6735790372,-0.071482718,-0.0488048792,-0.0049527902,0.0006292545,0.1189806312,0.1586837918,0.1362603009,0.0692648441,-0.0260408036,-0.5322327018,0.3291485906,-0.215462625,-0.0053817439,-0.0877875164,0.4366607666,0.0414892845,-0.1064981595,0.0108353356,0.094691515,0.0340188518,0.0655476525,0.0059462963,0.017562082,-0.1206619516,0.1582386047,0.1423452646,0.1708153635,-0.234853968,0.1895660013,-0.2832585871,0.0292929802,-0.1982437521,0.1561561525,0.2698151469,0.5158427358,-0.183451429,-0.0686847195,-0.0636009946,0.165361613,0.1112443656,-0.0649597496,-0.1914123297,0.089769423,-0.0087000672,0.2714968026,-0.2622262836,0.1059425324,0.0038290906,0.0026387305,-0.4920392334,-0.7109149694,0.4522462487,-0.4033147991,-0.5287759304,-0.0796565562,0.590672791,0.2403544337,-0.032153476,-0.7086960673,-0.0359898657,0.3959985971,0.2443489283,-0.3021709323,0.2129863203,0.1219399646,0.0085307239,0.0144249219,0.6226139069,0.3105249107,-0.2327545434,0.2679713964,-0.4604581892]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2904","title":"FORCE_REDOWNLOAD does not work","comments":"Hi ! Thanks for reporting. The error seems to happen only if you use compressed files.\r\n\r\nThe second dataset is prepared in another dataset cache directory than the first - which is normal, since the source file is different. However, it doesn't uncompress the new data file because it finds the old uncompressed data in the extraction cache directory.\r\n\r\nIf we fix the extraction cache mechanism to uncompress a local file if it changed then it should fix the issue.\r\nCurrently the extraction cache mechanism only takes into account the path of the compressed file, which is an issue.","body":"## Describe the bug\r\nWith GenerateMode.FORCE_REDOWNLOAD, the documentation says \r\n +------------------------------------+-----------+---------+\r\n | | Downloads | Dataset |\r\n +====================================+===========+=========+\r\n | `REUSE_DATASET_IF_EXISTS` (default)| Reuse | Reuse |\r\n +------------------------------------+-----------+---------+\r\n | `REUSE_CACHE_IF_EXISTS` | Reuse | Fresh |\r\n +------------------------------------+-----------+---------+\r\n | `FORCE_REDOWNLOAD` | Fresh | Fresh |\r\n +------------------------------------+-----------+---------+\r\n\r\nHowever, the old dataset is loaded even when FORCE_REDOWNLOAD is chosen.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\nimport pandas as pd\r\nfrom datasets import load_dataset, GenerateMode\r\npd.DataFrame(range(5), columns=['numbers']).to_csv('\/tmp\/test.tsv.gz', index=False)\r\nee = load_dataset('csv', data_files=['\/tmp\/test.tsv.gz'], delimiter='\\t', split='train', download_mode=GenerateMode.FORCE_REDOWNLOAD)\r\nprint(ee)\r\npd.DataFrame(range(10), columns=['numerals']).to_csv('\/tmp\/test.tsv.gz', index=False)\r\nee = load_dataset('csv', data_files=['\/tmp\/test.tsv.gz'], delimiter='\\t', split='train', download_mode=GenerateMode.FORCE_REDOWNLOAD)\r\nprint(ee)\r\n\r\n```\r\n\r\n## Expected results\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\nDataset({\r\n features: ['numerals'],\r\n num_rows: 10\r\n})\r\n\r\n## Actual results\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-4.14.181-108.257.amzn1.x86_64-x86_64-with-glibc2.10\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":99,"text":"FORCE_REDOWNLOAD does not work\n## Describe the bug\r\nWith GenerateMode.FORCE_REDOWNLOAD, the documentation says \r\n +------------------------------------+-----------+---------+\r\n | | Downloads | Dataset |\r\n +====================================+===========+=========+\r\n | `REUSE_DATASET_IF_EXISTS` (default)| Reuse | Reuse |\r\n +------------------------------------+-----------+---------+\r\n | `REUSE_CACHE_IF_EXISTS` | Reuse | Fresh |\r\n +------------------------------------+-----------+---------+\r\n | `FORCE_REDOWNLOAD` | Fresh | Fresh |\r\n +------------------------------------+-----------+---------+\r\n\r\nHowever, the old dataset is loaded even when FORCE_REDOWNLOAD is chosen.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\nimport pandas as pd\r\nfrom datasets import load_dataset, GenerateMode\r\npd.DataFrame(range(5), columns=['numbers']).to_csv('\/tmp\/test.tsv.gz', index=False)\r\nee = load_dataset('csv', data_files=['\/tmp\/test.tsv.gz'], delimiter='\\t', split='train', download_mode=GenerateMode.FORCE_REDOWNLOAD)\r\nprint(ee)\r\npd.DataFrame(range(10), columns=['numerals']).to_csv('\/tmp\/test.tsv.gz', index=False)\r\nee = load_dataset('csv', data_files=['\/tmp\/test.tsv.gz'], delimiter='\\t', split='train', download_mode=GenerateMode.FORCE_REDOWNLOAD)\r\nprint(ee)\r\n\r\n```\r\n\r\n## Expected results\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\nDataset({\r\n features: ['numerals'],\r\n num_rows: 10\r\n})\r\n\r\n## Actual results\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\nDataset({\r\n features: ['numbers'],\r\n num_rows: 5\r\n})\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-4.14.181-108.257.amzn1.x86_64-x86_64-with-glibc2.10\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\nHi ! Thanks for reporting. The error seems to happen only if you use compressed files.\r\n\r\nThe second dataset is prepared in another dataset cache directory than the first - which is normal, since the source file is different. However, it doesn't uncompress the new data file because it finds the old uncompressed data in the extraction cache directory.\r\n\r\nIf we fix the extraction cache mechanism to uncompress a local file if it changed then it should fix the issue.\r\nCurrently the extraction cache mechanism only takes into account the path of the compressed file, which is an issue.","embeddings":[-0.1253399998,0.0207806267,0.0185128842,0.0257142521,0.1821040809,0.0251934566,0.4788660109,0.2438331693,0.145796001,-0.2338699996,-0.1162827983,0.2687395215,0.0850747079,0.1121127084,-0.0132031068,0.3137487769,0.1101020277,0.2144826502,-0.0722493976,-0.0051923986,-0.2910993099,0.0634416714,-0.1833185554,-0.2195639163,-0.2065767497,0.2937990129,-0.0426936708,0.3026081324,0.0084062209,-0.3731202781,0.1836812794,0.2550069988,0.1820975393,0.6611076593,-0.000109716,0.0617883578,0.1586239785,-0.1793150753,-0.1704429984,-0.1858083606,0.0419964381,-0.0455286205,-0.3461463749,-0.0842089579,-0.0005326021,-0.2207529545,-0.1748757958,-0.346149832,0.5304936767,0.4028010368,0.2320071906,-0.1104161516,0.1202409416,-0.0721178129,0.1787120998,0.1414143443,-0.0699524358,0.3557813466,0.1045252755,-0.2072234303,-0.090631254,0.071447365,-0.1301329136,0.0515931137,-0.100776881,-0.0266386643,0.3280929923,-0.1360899806,0.0856568739,0.1706219614,0.60730654,-0.3377125561,-0.4172843397,0.0072846375,0.0183145087,-0.2442766875,0.1650239527,0.1792488247,-0.0074186656,0.368666321,-0.0126197841,-0.0254521314,0.0721380934,-0.1948064566,-0.0308580399,-0.117819868,-0.1004166082,-0.0362135768,0.0516206622,0.0602480508,0.4213585556,-0.021616118,-0.0566806123,0.0342854224,-0.029152602,-0.1342204958,0.2153061479,-0.3018356264,0.0747370049,-0.0281923842,-0.0764476284,-0.1077359319,-0.0592024475,0.04897983,0.0738615543,0.2261448056,-0.1387252212,-0.090120174,0.3910107613,0.1767412573,-0.251378119,-0.0726185068,0.0294129029,-0.2027270049,0.5261009336,-0.0647600591,0.2614033818,-0.1472218335,-0.4044618905,-0.0181814078,-0.1070784777,0.0032655229,-0.2742642164,0.1767624319,0.0207748469,0.3530659974,0.1341454834,0.1856252104,0.0581398085,-0.206356883,-0.2226337641,-0.1226871684,-0.3091842234,0.0251519568,0.2947835922,-0.1708823442,0.3440055251,0.5470414758,-0.3297725916,-0.5027301908,-0.0621353462,-0.0314323269,0.145252198,0.2964736819,-0.0633033141,-0.0863981768,0.2666994929,-0.0901367441,-0.0168399569,0.3552974463,-0.0914471522,-0.3069287539,0.1436540186,0.1975634098,0.0511063114,0.1958488077,-0.127950877,0.0155278603,0.3042720556,-0.0705190748,-0.0457962081,-0.0500780307,-0.0136534544,-0.2484514862,0.0476663969,0.7337824702,-0.4681162834,0.0081418594,-0.2922013104,-0.0890840143,0.2404547036,0.0785243809,0.0047035967,0.0423925631,-0.4686902165,-0.3603310585,0.1667583287,-0.1669529229,-0.4369600415,0.0058205375,-0.2142421305,0.2575328946,0.1665146351,-0.0449789762,-0.1501356214,-0.0172337554,0.0970478505,0.2490073442,0.0923939124,0.044823505,-0.3246170282,-0.288821131,0.0455681123,-0.1329662651,0.2855156362,0.405680865,0.3521436155,0.166620627,0.3699389398,0.0625928491,0.0155102359,0.0143866315,0.3367364407,-0.0180097409,0.0330760218,0.1765212268,-0.5221686959,0.3075172901,-0.0336038433,-0.2515624166,0.0728959292,-0.0701956674,-0.5719491243,-0.228338778,-0.0720977858,-0.3877546787,0.1725475937,0.4713421166,-0.033435449,0.0184990596,-0.1434893459,0.4725981653,-0.3845011294,-0.0480710976,-0.0413006209,0.2855195999,-0.1932275444,-0.1814612597,-0.2195678353,-0.1187965199,0.2447945774,-0.1625743508,-0.1215589717,0.4299698472,0.186048165,0.2720868587,-0.3619140387,-0.1950368881,0.1652197391,-0.016063001,-0.1465488821,0.1185459942,0.2459682077,-0.0776265264,-0.0173951481,0.2789047956,-0.4031001925,-0.0885871574,0.071374476,-0.0462191142,0.2666130662,-0.2367841005,0.2828178704,-0.340397954,-0.3403313756,0.2947020233,0.0018241567,0.1279272139,0.0989251807,0.1121156812,0.2867794335,0.0066712452,0.047518149,-0.0890123621,-0.1765656024,-0.2945634723,0.0131209595,0.3500014544,0.5929611921,0.1657259762,-0.0166612044,0.1451728493,-0.0583074652,-0.253809005,0.2043861002,-0.2613338828,-0.0064706737,0.3311824501,0.1878525168,0.0763362646,-0.6370830536,0.186015591,0.5140926838,0.2118025422,-0.2195149958,-0.1909853667,-0.1673647612,0.4125154912,0.0630170777,0.1435509324,-0.0300561506,-0.1065669209,-0.138272956,0.5955287814,-0.0589990281,0.073292084,-0.1850811243,-0.0749260634,-0.0461661331,-0.1426966637,-0.1472802907,-0.1764150858,-0.1414824873,0.0910957381,0.2313582748,-0.3419226408,0.1477311403,-0.1828764379,-0.1267566234,-0.3267619908,-0.0833430737,-0.1293941885,0.0340040401,0.2224018276,-0.2250489891,0.1471374333,0.169757694,0.0325683653,-0.004715248,-0.3554546833,-0.3341683745,-0.1029282659,0.1511999369,0.2633759975,-0.2600107789,-0.4554549158,0.1090232432,-0.3711392879,-0.0406573713,0.1015611589,-0.0919760466,0.3833361566,0.2238112539,-0.0310825258,0.4020083547,0.2698833942,-0.4376188517,-0.2790603936,0.0805035159,-0.1014774665,-0.2633699775,0.0147913452,0.0609156676,-0.0121200867,0.2737061381,-0.4307680726,-0.0845259503,-0.2824058533,0.0837785453,-0.0585749894,0.2336959094,0.2081224918,0.1207609028,-0.2136411071,-0.1040799022,-0.0956789032,-0.041495204,0.233491689,0.126982823,0.0212046467,0.2292778045,0.0935394913,0.6404239535,0.0503102541,-0.1254811585,0.5086991787,0.125423491,0.7294853926,-0.074036397,-0.339060694,-0.0056349542,-0.1845843047,0.319257617,0.0420288928,-0.1061969697,0.1309799999,-0.1564230025,0.0242274236,-0.1678121686,-0.2236509025,-0.1179092973,-0.2493901998,-0.0220435485,0.314229995,-0.1047639698,-0.1341655105,0.0813808367,-0.0185252242,0.0876276046,0.3220412135,0.0884925127,0.0994393826,0.5284665227,-0.1080815345,-0.0133388601,-0.0877977833,0.3645614684,0.2040727586,-0.202382952,0.0585922338,-0.023118237,0.7181988358,-0.2194465101,0.313290745,0.2286951095,0.0114228521,-0.0870219991,-0.196765542,0.1032617912,-0.1487740576,0.1887097061,0.3649292588,0.0375474505,0.0390015617,-0.1003576666,0.2032171041,0.0621288046,-0.1629727781,-0.2788121998,-0.1599720269,-0.4588589966,0.3117338121,0.1915880591,0.0890043527,-0.320887208,0.0846341178,-0.2478237301,-0.4116822779,-0.0395846739,0.1951411813,0.2008189261,0.1139606908,0.0489872061,-0.0161832664,-0.1171152368,0.2674762607,0.4194466472,-0.1649921089,0.0667295754,-0.1280568391,0.1093956754,-0.067581296,0.319385618,-0.0657960176,-0.1120607778,-0.3978951871,-0.1618822664,0.0620966889,0.0673459694,0.002971882,-0.3354582787,-0.0954507068,-0.4370097518,0.2871852815,-0.0682167336,-0.0515976958,0.4572842717,0.2364326417,-0.2039115429,0.1625162512,0.0528943725,0.8192458153,-0.1221480891,0.165192157,0.0327902324,-0.0593867525,0.4034763277,-0.2886822224,0.194058761,-0.1767736971,-0.3061233759,0.1044934094,-0.0408082418,0.2328797132,0.2079223692,-0.1264044195,0.1385401487,-0.0984730124,0.4167669713,0.2764787674,-0.0030264198,-0.2224701941,-0.1413925886,-0.2291935533,0.1468190849,0.2811729312,0.1926800758,-0.1454819441,-0.0086173862,0.1220361069,-0.0229478516,-0.2685617507,0.2866361737,-0.1642533094,0.3783459961,-0.4507656693,-0.1098404899,0.0326456428,0.2600927949,0.2549611032,0.2578200102,0.0146770021,0.0830926001,0.4122848809,0.1323083341,-0.1551139206,0.0008572826,0.4158841372,0.1006288007,-0.3379714787,0.1988293082,-0.0907847136,-0.1476554275,0.0037588191,0.2748386562,-0.0433281027,-0.1125033721,-0.092763938,0.0935379267,-0.074974604,-0.4186442494,0.1443356574,-0.0522535145,-0.1282395571,0.121255748,-0.4481029809,-0.4539896846,-0.2699021995,0.5651443005,0.1653032899,0.0896883756,0.3432316482,-0.1958156079,0.009230705,-0.1956038326,0.0348324627,-0.1255518347,-0.3250435591,0.5315566063,-0.2140324712,-0.1011318862,0.1889260113,-0.173754245,0.1105415151,0.3808698356,0.1637355983,-0.2844932377,-0.5291988254,-0.3904183209,-0.0524902157,0.2282985598,0.050847441,-0.0072952523,-0.0107132103,-0.0045601041,-0.3350776434,0.2403877079,0.0919069722,0.0729334131,0.1242790818,-0.1043476984,-0.2560699284,-0.2592730224,0.1775953174,0.3209490776,-0.0254106764,-0.2779419422,-0.3296145797,0.0958329737,0.3220802546,-0.1090267152,-0.1992628872,-0.0858467817,-0.2128356099,-0.1136852056,0.2098957598,0.4146845043,-0.0505065322,-0.0152493529,0.3320708871,0.0022462029,-0.1997736692,-0.095850639,-0.2755773664,-0.1480358839,0.0696581528,0.2725877762,-0.0361606181,0.1273908317,-0.4385435283,-0.0649156943,0.1195174456,0.0885495842,0.1913103312,-0.2284960747,-0.1663822532,-0.0860028714,0.2218851149,0.4013639688,0.0489362255,-0.1580483168,0.2325664312,0.2389936596,-0.5628596544,-0.0106937625,0.1943591684,-0.003649774,0.098564215,0.1986977458,0.2104046792,-0.1681738794,-0.0613068901,0.1347883046,0.2547552884,-0.0625275373,0.3317419291,0.6826942563,0.1762798131,0.248175472,0.2562503219,0.1975747496,0.164253071,0.329222858,0.0229687337,-0.0241935551,-0.2815636098,0.0184473433,-0.379707098,-0.3709614277,0.3160880208,0.4231722653,-0.2143074274,0.236592263,0.0353621766,-0.3099908233,-0.0264861919,0.0299492758,-0.2805517018,0.2032455802,-0.1236360446,-0.3274005949,0.3878690004,-0.2026110739,-0.1777496487,0.1964985281,-0.0410761125,0.0262600426,-0.1356980056,-0.0048754462,-0.2721968591,-0.4071184695,-0.0643913075,0.1155254096,0.1304227114,-0.2057668418,-0.0984744281,0.3018858433,-0.1364281178,-0.0018364664,0.2399721444,0.3788969219,0.2111078948,0.1534903944,-0.0022816143,0.1738818735,-0.0088148732,-0.3157631755,0.4200020134,0.0662739351,-0.0240946896,-0.1370903552,0.203938961,-0.2506957352,-0.0812061727,0.1522063017,0.0141364988,-0.1547079533,0.4992974699,0.0719960332,-0.3092764318,-0.0027599297,0.2752628028,0.0251966976,-0.079818204,0.3267326951,-0.1332303882,0.2819019258,-0.2300074697,0.1109809577,0.1148270667,-0.034752652,0.4555131197,-0.4898700416,-0.2097949684,0.0988615081,-0.816400528,0.0120940423,0.0328483358,0.2008297592,-0.1417099983,0.0833700895,-0.0083174789,0.0212943275,0.1145128533,0.3138949871,0.0724570304,0.3102653921,-0.0539459884,-0.1346898079,-0.0958660915,-0.1477009207,-0.1163652837,-0.5594716668,0.0759588927,-0.1853275597,0.2095306814,-0.037059404,0.0065100822,0.2180145383,0.2444203049,0.4945931435,0.0287282132,-0.0332200117,-0.0520022735,-0.1335050911,-0.5665888786,-0.039800819,-0.2466407716,0.0500838161,0.1520656943,0.1390793025,-0.3625163436,-0.3033446372,-0.2144049257,0.3089033067,0.2004514635,-0.0365050435,-0.3225580454,-0.0226420909,-0.0864753276,0.0237421617,0.2618903816,0.2817649543,-0.3526063561,0.3924575746,-0.3913657367,-0.3119346499,0.2078100592,-0.3488782942,-0.3597286046,-0.1109973863,0.2195085585,-0.059010148,-0.0838459656,0.080401659,0.3842255771,0.2200256139,-0.1292812973,-0.2282675058,0.1182789207,-0.403898716,0.1522310525,-0.0926432759,0.1583367139,0.1514638215,-0.2153838128,0.1477641314,-0.0796994865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2902","title":"Add WIT Dataset","comments":"@hassiahk is working on it #2810 ","body":"## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":6,"text":"Add WIT Dataset\n## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n@hassiahk is working on it #2810 ","embeddings":[-0.1296591461,-0.1891998202,-0.2115951777,-0.0000080944,0.0460982546,0.0357038304,0.4529203176,0.1215223595,0.2780340016,0.1513489336,-0.029059099,0.1144830734,-0.1435055733,0.2445029169,0.1256047338,-0.1864110827,0.0068955352,-0.0093742143,-0.0735348091,-0.1070073843,-0.0773350447,0.1636914909,-0.1567373872,-0.2530801892,-0.4015705884,-0.1087995097,-0.146388337,-0.0190194007,-0.1731389165,0.0142502328,-0.2311714441,0.0920530111,-0.1416087747,0.4282389283,-0.0000961792,-0.0348447226,0.1924369335,-0.1517420858,-0.0877345204,-0.1121518463,0.0036920926,-0.0255717319,-0.2709282935,-0.3336799145,-0.1762119234,-0.1874346882,0.2792802751,-0.038891498,0.2992183864,0.1237806827,0.3313299716,-0.1193304285,0.167231366,-0.0636842549,0.1344261765,0.1325438321,-0.2437879592,-0.1032884791,0.0572029762,-0.1843628138,-0.2473023236,0.6803191304,0.1149657592,0.0394390561,-0.0199051965,0.0259543676,-0.0537805147,-0.2517703772,0.1902595907,0.4509062171,0.5951299071,-0.1380434632,-0.1741087288,0.0608816259,-0.0706331432,0.0042669997,0.0225002747,0.3247715533,-0.1227583364,0.053050328,-0.1813238263,-0.1304020733,-0.1617643535,0.2954109609,-0.1025401354,0.3974758983,-0.0176089592,-0.0563631766,-0.0496043675,-0.1633326858,-0.2125443816,0.02245792,0.240905866,0.0170114562,0.0630257428,-0.129550159,0.054668501,-0.1277145147,0.3031286895,-0.2512640655,-0.039158728,0.1198865026,-0.3077050745,0.2759459615,0.1043220907,-0.2439876199,-0.1338686198,-0.1969237626,0.2190607339,0.0052139987,-0.2566974163,0.1069463864,0.1158736348,-0.0193225145,-0.1939061284,-0.1762420088,0.0073827389,0.1123326346,0.0784640163,-0.0886906609,-0.124503158,-0.055896312,-0.3169686496,0.2779684961,0.1109271422,0.0749763697,0.0985479653,0.2271314412,-0.1937326193,-0.1465904415,-0.115335606,0.1956622303,-0.1542165428,0.1144959927,0.1543435156,0.5087482929,0.2681051195,-0.0842734575,0.0915060565,0.0306925531,0.1029897705,-0.0939951912,0.1980237216,-0.0445276089,0.1020599678,-0.1297828108,-0.0630924329,-0.288815707,-0.1791048646,0.048106607,-0.0815851614,0.0402405933,-0.4307381809,0.2774644792,0.1309566647,-0.1127759367,0.1063605249,0.7211184502,0.0591425896,-0.0758869126,0.1704614013,0.2683287561,-0.3786454499,-0.0175030548,0.1434833258,0.3921846449,-0.1378667355,-0.0694735646,0.1149419025,0.2635501623,-0.0323280878,0.2083413899,-0.0748511776,0.0976687744,0.0117110927,0.2452786714,0.1070185974,-0.1362077594,-0.0952911451,-0.1215822399,-0.0758547634,0.0205433052,0.1006369516,0.3590261936,-0.0074512311,0.0786975026,0.0008037383,0.6429982781,-0.101937674,0.2204965353,-0.2547368109,-0.412805438,-0.031592954,0.3136671782,0.1404488534,-0.3794309199,0.3183832169,0.2673033774,0.1137722209,-0.2929435074,0.1085533649,0.0377520844,-0.0135402372,0.1493370086,0.0558074787,-0.2918334305,-0.0843841732,-0.1060758308,0.1630552709,0.4458846748,-0.1963102072,-0.2546880245,0.0034701391,-0.1770533621,-0.0542313494,-0.2643735409,0.3854094744,0.2371143699,-0.1219256148,0.2192856669,0.0515208244,-0.0828982666,-0.2751990557,-0.0682251155,-0.0554965064,0.1411690116,-0.2138494253,-0.0074874591,0.0453312024,0.162361756,-0.2836136818,0.0503010452,0.1026869044,0.2066954374,0.0853403509,0.3449212611,0.3921612203,0.1152452454,0.2651148438,-0.5831927061,0.0691323876,0.0750517696,0.0019045175,-0.0033749938,-0.3692139387,0.0319884233,0.2239721268,-0.0128715681,0.1168907061,0.0606278479,0.2472451031,-0.0976875871,0.001401014,-0.3726579845,-0.0102998558,0.162639454,-0.1726347953,-0.1651332676,-0.2103906274,0.0991069302,0.5462095737,0.2400736511,0.5376595855,0.0971913859,-0.1212614775,-0.2618568838,0.073506847,-0.2561023533,-0.0930368453,0.3516949713,0.0182377622,-0.1866730601,-0.0738957152,-0.1027568281,0.1412557811,-0.0225975774,0.1650807112,0.0877017826,0.297613889,-0.1069796085,-0.5398007035,-0.0681409612,-0.0508023016,-0.2064673901,0.0649895668,0.0703736097,-0.0205571149,-0.2685602605,-0.0878075063,0.1285372823,-0.2698201537,-0.0652996004,0.3480983078,-0.058700081,-0.144668743,0.2157517225,0.2154152393,0.2665415406,0.1056036651,-0.0490454584,-0.0699434429,-0.4251724482,-0.0013740142,0.2867534161,0.2798976302,0.0790984556,0.468409121,-0.1529595554,0.019352071,-0.2396499366,-0.6391475201,0.1912251115,-0.1763069183,0.2191690207,0.0712262616,0.0936918706,-0.1991403848,-0.2139495015,0.1469445527,-0.1367393434,-0.1214991733,-0.0671442822,0.0764834508,0.0814846978,-0.1663844138,-0.2625015378,-0.1379565299,-0.3415474296,0.2776284814,0.2449719161,0.1046371013,-0.0269246474,0.3842811882,0.0516841039,-0.0223082379,0.2923726141,-0.3290747404,-0.029114319,0.1770710647,-0.3507783413,-0.3409646749,-0.241635114,-0.2149017453,0.2913351059,0.0963894725,-0.452812463,-0.3489452302,0.0134124234,0.2051300406,0.1298554689,0.076393947,0.282122463,0.1753068119,-0.3087322414,-0.1432863772,-0.0113858152,0.0432145521,-0.1399844736,0.0589655936,0.1224881411,-0.0000328349,0.2479563504,0.4598668814,-0.2399621606,0.0897038355,0.4043138623,-0.015272229,0.2053738832,-0.2565110326,-0.1264286041,0.4514034986,0.0070677064,0.0613207631,0.30798769,0.1180268228,-0.4592045248,-0.0890442282,0.174862057,-0.5091278553,-0.1667373627,0.1673555821,0.1669268012,0.1635840684,-0.1377123594,-0.2422106862,-0.1948408335,-0.1597946584,0.1669388264,0.2362881452,-0.0687819272,-0.1252325177,-0.2005453855,0.0608569533,-0.1316801757,0.0990372747,0.2825544477,-0.123838447,-0.1642559022,-0.161138624,-0.0587822609,-0.1694849581,0.2437805086,-0.3938077092,-0.1649195999,0.1007589772,-0.0390315913,-0.1142330319,0.2840626538,-0.3339577615,-0.2328336537,0.0622188188,0.2880499065,-0.3703322113,0.0837903321,0.4168727994,0.1480499059,-0.0266231615,-0.0325266942,-0.1320562661,-0.2437737137,-0.4280227423,0.0575218908,-0.0885785222,0.1218918115,0.3068802357,-0.0476196334,-0.0697411075,-0.0784421787,-0.0828517675,0.2587498724,-0.0900649056,0.3041183949,-0.1872166842,-0.138946712,-0.0720472485,-0.1653235257,0.1967177689,-0.0868958831,-0.0660888404,-0.1588893533,-0.2913936675,0.0738343373,0.0961803868,0.2457379699,0.0848319456,-0.0663812011,-0.0204767045,-0.3289154172,0.2266960442,0.2494198382,0.1433276385,-0.0007672762,-0.2630187869,0.3372961879,-0.0651260763,0.0367082134,0.3666113317,-0.0947311148,-0.4259827435,0.3040920198,0.3670836389,0.9130336642,-0.1412870288,0.2655439973,0.0825656652,0.0368013196,0.4112624526,-0.2702554166,0.0926808119,0.063380003,0.0450684801,-0.042350892,0.2032682449,-0.0341339484,0.0403719842,-0.2530496418,-0.1733001024,0.3581714332,0.1081538126,0.0419844128,0.3453404605,-0.0444174036,-0.4325706661,-0.3199236393,0.2508710325,0.060147658,-0.1543311328,-0.0326714627,-0.0739602372,0.0480866395,-0.0830819085,-0.3174287677,-0.0862657726,-0.2957362831,0.1991055459,-0.3531002998,-0.3360367119,0.1190520227,0.1976566911,0.1793552339,0.1488007456,-0.2352505326,0.1980030835,-0.1115192845,-0.136059463,-0.0225254111,-0.1173774973,0.3119098544,-0.1344185174,-0.1809410751,0.0406746306,0.2653738558,-0.2495407313,-0.4706448913,-0.1169789657,0.3314366341,0.1948352754,0.0485579409,0.2501215041,-0.413610667,-0.2901029289,0.2998668849,-0.1534142941,0.1727076322,0.0644296631,0.234110564,-0.0873228759,-0.0742139593,0.2355686873,0.2749872208,0.0819436833,0.2199673355,0.2496194392,-0.1277595311,-0.4075970948,-0.2303928286,0.0248409584,-0.3103953898,-0.08661329,0.0937269107,-0.3579361737,-0.1972269416,0.2842388749,0.0858474672,-0.0124888159,-0.0472396538,-0.0732503757,-0.0512179583,0.4948905706,-0.1107646599,0.2192007452,0.0759131759,0.1513412446,0.1365086883,0.2520558536,-0.5044214129,0.0618388243,-0.1622692049,0.1555999666,0.1093536094,-0.1887407005,0.0446312688,0.1271794736,0.297516793,-0.0032108144,-0.1694126576,-0.3926710784,-0.1479100138,0.0459250957,0.00995667,-0.0352526307,0.1959041208,-0.0861305743,-0.0785156563,-0.2286321074,0.1431579441,-0.0590985827,-0.221162349,0.2085623145,-0.1347504705,0.1107792184,0.1136089787,-0.1726914197,-0.0558690876,0.2453998476,0.1603109837,-0.1366208494,0.2261314541,-0.1058626622,0.074613668,0.22358042,-0.1432922632,0.1838464141,0.210678488,0.1371192336,0.0155197922,0.1411829144,0.2470877767,-0.1947020739,-0.338175416,-0.0111634769,0.168082267,0.4273661077,-0.3924217224,-0.0953474045,0.2009430379,-0.4014649987,0.4477969408,0.2921489179,0.1483340859,-0.002726747,-0.0974504873,0.067161195,-0.1852512658,-0.3622380793,-0.0507756211,0.0865137428,0.0177711882,0.0699414983,-0.1085267439,0.0994353145,0.2459332198,0.3117552102,-0.0358871408,0.4007888138,-0.0078987358,0.0618646182,0.3345318437,0.2172142565,0.4865316749,0.1046640351,0.0416013673,0.3430802226,-0.0083274813,0.2223719209,0.0999372602,-0.0087033538,-0.1931295693,0.1871576011,-0.0184282884,-0.1056677848,-0.275636971,-0.1552461386,-0.1231860667,0.0284091607,-0.0035716088,-0.0306761824,0.1527671069,0.0474604145,-0.1027932689,-0.1923794597,0.0851862952,-0.0813491642,0.4018681645,-0.2289473712,0.2052509189,0.2295409143,-0.0062884679,0.0375786684,0.4565461576,0.1032586545,-0.0071680066,-0.0938500986,0.0924604014,-0.1135674641,-0.0638433546,-0.2580824494,0.0472246632,0.3923101425,-0.0717268735,0.2726536989,0.267921716,-0.3474231064,-0.1523036957,0.4223014116,-0.2082197219,-0.0883307829,-0.5123988986,0.0371011719,0.0079047736,-0.2581389248,0.0969417989,-0.5972567797,-0.0656792223,0.0051606284,-0.0808611438,0.0543405078,-0.102023758,0.1447420567,0.1186306775,0.4803176522,0.2922181487,0.1430308968,0.0741488263,-0.3569711447,-0.473400563,0.1692283154,-0.1185174435,-0.0800530314,-0.1962443441,0.3564409912,0.296964705,0.2779829502,-0.0123385582,-0.1476554275,-0.0388901271,0.0978951603,-0.3775681555,0.1366155446,0.1053182557,0.1353593618,-0.0266050361,-0.1613134742,0.0928840712,-0.0417491868,0.2913910151,-0.1669008136,0.0616998039,-0.0978731364,-0.0602027848,0.2010447979,-0.0838134512,0.1094563082,-0.4794981182,-0.3175835311,-0.1471890509,-0.1113790944,-0.1586659849,0.2541112304,0.2663548291,0.4172587097,-0.0143395821,-0.0704477504,0.2176796794,0.4207631946,0.0117856711,0.1037914306,0.0464718007,-0.0413710028,0.2674077749,-0.1472931802,-0.1270537078,0.4860852957,-0.1144139916,-0.1948803961,-0.0594344959,-0.5313600898,0.0803394318,-0.3131714165,-0.2420081943,-0.0785194188,-0.0384221189,0.0641538873,0.2597292364,-0.3967080116,0.1920229495,0.2368176877,0.2900294065,-0.0751894638,0.3233103752,0.1524246633,-0.1856960952,-0.1501988024,-0.0521546677,0.0246847868,-0.1107614785,-0.2156294882,-0.4013026655]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2902","title":"Add WIT Dataset","comments":"WikiMedia is now hosting the pixel values directly which should make it a lot easier!\r\nThe files can be found here:\r\nhttps:\/\/techblog.wikimedia.org\/2021\/09\/09\/the-wikipedia-image-caption-matching-challenge-and-a-huge-release-of-image-data-for-research\/\r\nhttps:\/\/analytics.wikimedia.org\/published\/datasets\/one-off\/caption_competition\/training\/image_pixels\/","body":"## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":23,"text":"Add WIT Dataset\n## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nWikiMedia is now hosting the pixel values directly which should make it a lot easier!\r\nThe files can be found here:\r\nhttps:\/\/techblog.wikimedia.org\/2021\/09\/09\/the-wikipedia-image-caption-matching-challenge-and-a-huge-release-of-image-data-for-research\/\r\nhttps:\/\/analytics.wikimedia.org\/published\/datasets\/one-off\/caption_competition\/training\/image_pixels\/","embeddings":[-0.0497860387,-0.0547444522,-0.1280666143,0.0003004618,-0.039354112,-0.0071313963,0.3546046019,0.1659145951,0.1936855763,0.2055810541,0.0639799759,0.1460067928,-0.0242924541,0.1997424215,-0.01487609,-0.2061530054,-0.0474179126,0.007556899,-0.095439963,-0.0953547359,-0.1237470731,0.1472279876,-0.1375917941,-0.2112537324,-0.4580724537,-0.0875768587,-0.1730553955,-0.0253455658,-0.1721293628,-0.0776184052,-0.1209003329,0.1993556619,-0.1435845345,0.3342963159,-0.0000973471,-0.1190001816,0.2309069782,-0.1802841574,-0.0871033221,0.0576613285,0.1233595088,-0.0183559302,-0.3197696805,-0.4497412145,-0.1917415559,-0.0919262245,0.250259757,0.0872264057,0.220723629,-0.0171945672,0.3280394673,-0.1210444793,0.10394793,0.0242980495,0.0408663712,0.1121114418,-0.1330121011,-0.0007004917,0.0150091052,-0.3246178627,-0.1604316086,0.6847470999,0.0171727017,0.0272535756,0.0614412166,0.0563659444,-0.1721417308,-0.3297246695,0.1790350825,0.4349500835,0.5310005546,0.0022286586,-0.1660662293,-0.0531980433,-0.0875593573,-0.0650093853,0.1432961375,0.2999428809,-0.1368227452,0.0239426252,-0.2666973472,0.0028924709,-0.2545390427,0.1593592912,-0.1959471852,0.4131114483,0.021566527,0.0326524712,-0.000182682,-0.2101522833,-0.4369372725,-0.0127320085,0.2125476599,0.0099950507,0.0381090157,-0.2509273589,-0.0503862202,-0.1546090245,0.3692227602,-0.197378248,-0.0403584838,0.0733242184,-0.1825137436,0.3001660109,0.1151571646,-0.273580879,-0.1728782058,0.0215191487,0.1827842295,0.0874059945,-0.2466638684,0.1344642043,0.2253025919,-0.1029728055,-0.1772990078,-0.087800771,0.1070196927,-0.0466616042,-0.1075610965,-0.008583298,-0.146704793,-0.0710129738,-0.4360583127,0.2165165991,-0.0335759073,0.1098506823,0.1278970838,0.1473859996,-0.2188078016,-0.1533710063,-0.1057867333,0.1009146869,0.0557398014,0.2790454626,0.1211024597,0.3311381638,0.256734401,-0.0824029595,0.1746457219,0.0947210044,0.0416478626,-0.0953107029,0.2555246055,-0.008912893,0.0164618678,-0.1227277219,-0.051241383,-0.2793621719,-0.19902426,0.0270652734,-0.1084727868,0.2036321461,-0.3212899566,0.2668367922,0.1446974277,-0.0583522059,-0.0045630517,0.6982100606,0.0959759504,0.0308421906,0.1365310401,0.1445246041,-0.3958842158,-0.0636466146,0.2453516275,0.4490688741,-0.0886282399,0.0255282167,0.3061005771,0.3564617038,0.0193696469,0.1932205856,0.0245506912,-0.0181516893,0.0488758124,0.2630175054,-0.1056210548,-0.2358684838,-0.1195387468,-0.1759907305,-0.0965119898,0.1020767167,0.0326437689,0.2860510647,0.270050317,-0.102775909,0.093534559,0.6013006568,-0.1892949641,0.2137656063,-0.2130067647,-0.3863318861,-0.0069002835,0.3273807466,0.1285638511,-0.4685770869,0.2250231653,0.1909397244,0.0319985636,-0.3845872879,0.0987228006,0.0075334026,-0.0310356412,0.0995845795,0.091828078,-0.2398949265,-0.0079521695,0.0546789393,0.2049163729,0.4832212925,-0.1773444712,-0.2481343299,-0.038995903,-0.1513748169,-0.1065134704,-0.3698706031,0.3537430465,0.1205859855,-0.0045493012,0.2210574299,0.0784497336,-0.0594155379,-0.3163442016,-0.0456017256,0.1199062467,0.1867598891,-0.2623721361,-0.0185750853,0.1218658164,0.0860536918,-0.0835613608,0.1807654053,0.0911751837,0.1913248599,0.0461769477,0.3455418348,0.2677955925,0.0530583225,0.3559015393,-0.6900479198,-0.0690085068,0.19123438,0.0030610359,0.0122600449,-0.2659728527,0.1819274873,0.2717046738,0.0452171303,-0.0804450437,-0.0083080204,0.1473466903,-0.0244393963,0.0583699085,-0.3646619022,-0.0810375884,0.2129996717,-0.2515609264,-0.2326057106,-0.1890244186,0.1808756441,0.5132712126,0.2433172613,0.3687907755,0.1684178561,-0.1199678779,-0.1401048452,-0.0068703438,-0.0960154831,-0.1688125134,0.3742109239,-0.053772375,-0.1260635257,-0.2724704742,-0.0922951698,0.1780956537,-0.0158920027,0.2231312841,0.0486644618,0.2492239773,-0.0249909423,-0.4577436149,-0.1271143258,-0.1574495137,-0.1973033249,0.056282267,0.0784645602,-0.071331881,-0.1991609186,-0.066964902,0.2833610475,-0.3087125421,-0.1143085882,0.3522223532,0.0152183259,-0.1103117839,0.183313325,0.0873240456,0.1733572483,0.1862344891,0.0245582648,-0.1337936968,-0.5455780029,-0.0611162968,0.2686133087,0.1743589044,-0.0161512867,0.5377413034,-0.1781012118,0.0367738679,-0.2884244919,-0.705337286,0.1020467132,-0.1840780228,0.1969555616,0.0155186988,0.2760653794,-0.2234164178,-0.1541128606,0.1546294391,-0.1017461941,-0.1935486495,-0.2209588885,0.0480773337,0.0420998633,-0.2291868478,-0.2148961276,-0.1442073584,-0.2447485179,0.1066020951,0.3489588201,0.1721870601,0.0647365227,0.3838174343,0.0280661602,-0.0238699075,0.2764138281,-0.3571163714,-0.032202661,0.3682462573,-0.3039095998,-0.403814584,-0.1645279825,-0.182767272,0.1699132323,0.0591959544,-0.3852038085,-0.3932056427,0.0160926636,0.1964574009,0.2606021762,0.0776553899,0.1688962281,0.1696974039,-0.2888213396,-0.120130457,-0.0905002728,0.1087626144,0.0804397017,0.0327175856,0.2156327516,0.0916362107,0.2265037298,0.6132438183,-0.1613393426,-0.0281569362,0.4119825959,-0.0383804776,0.224104315,-0.0002331532,-0.0705114752,0.497284919,0.0091163954,0.1581450105,0.2489970326,0.1093951836,-0.4575267434,-0.1425068378,0.0741071776,-0.4309840202,-0.1831286401,0.1126235873,0.282223165,0.1359028369,-0.1255722344,-0.1118768379,-0.1950071603,-0.2140990347,0.1501378864,0.1365459114,-0.033066839,-0.0854402483,-0.1478022784,0.0440795906,-0.2099582404,0.1326003671,0.3808392882,-0.1713135391,-0.1235297993,-0.064189747,-0.1482826024,-0.2265735269,0.2456169724,-0.3066806197,-0.0396218561,-0.0001861092,-0.0057738582,0.0171943922,0.2211831212,-0.2088258266,-0.3443025649,-0.0452317037,0.247772336,-0.4293949008,0.0881709754,0.3906086981,0.1875751317,-0.0982594416,0.0725138634,-0.1535127908,-0.4466047287,-0.4639799893,0.0759280697,-0.0179230105,0.234052375,0.1692747027,-0.0461039618,-0.0495606475,-0.0345803201,-0.0313565917,0.3599346876,0.0232606847,0.2451539487,-0.1866826117,0.0159258991,-0.0362759829,-0.0392687432,0.2646231949,0.0655591339,0.0620991588,-0.1904884726,-0.1798860431,-0.0081319464,0.1325878799,0.2491253167,-0.0325292051,0.0099018356,0.0323803313,-0.3869562447,0.2138520479,0.1015860438,0.1107611284,-0.0148324464,-0.2682086527,0.3575321436,0.0139691914,0.0896827802,0.389649421,0.0827483088,-0.4077329636,0.13139759,0.606361568,0.8766955137,-0.1784271747,0.1968854815,0.1156533733,-0.122769326,0.490136683,-0.0125109544,0.0893897936,-0.0903045982,0.0462060049,-0.0482571162,0.1810766608,0.0233861692,-0.0071799215,-0.2383960932,-0.0591488257,0.4950783849,0.0509889983,0.0223183818,0.2454401106,-0.0180004686,-0.3712224662,-0.1638911664,0.2396210581,-0.1347939372,-0.0386185609,0.0205545817,-0.1282645911,-0.0563069582,-0.2408725768,-0.2661218941,-0.0375264436,-0.3923751414,0.2052637786,-0.2520484924,-0.232131511,0.0540006123,0.1925714016,0.1538775563,-0.0118546505,-0.1931021065,0.2699621618,-0.2166959643,-0.1297280639,0.0994486734,-0.0778556168,0.3134611547,-0.1621257216,-0.2310339063,0.0682638064,0.3530699909,-0.2003534883,-0.3784057498,-0.1596205533,0.2148566991,0.3098063767,-0.0714116693,0.0368100926,-0.4080921412,-0.2550562024,0.2662607431,-0.1720845252,0.1280960441,0.0953225568,0.338157177,-0.0661718771,0.0120415185,0.279966861,0.1513243467,0.0692156926,0.252928853,0.1970508397,-0.1593777239,-0.262548238,-0.1428267211,-0.2143792063,-0.3051334023,-0.1352979839,0.0671666935,-0.2043531686,-0.1338025331,0.3641974926,0.1625054628,-0.0255595706,0.0916185379,-0.0791770592,-0.0435318984,0.3351586759,-0.2645081282,0.2620086372,0.1362607479,0.0919233859,0.0621007904,0.2466659248,-0.4697705805,0.0008081495,-0.276823312,0.220438391,0.2841959,-0.1862068176,-0.0280307755,0.125368759,0.3021201789,0.0017990991,-0.1685612202,-0.3627264202,-0.2431438267,0.0422772169,0.0334107876,-0.1346720457,0.0254573394,-0.0790322646,-0.0836148933,-0.1622535139,0.0931934118,-0.2240975499,-0.2472412735,0.2228255719,-0.1823137701,0.0700710714,0.094846718,-0.1861076504,-0.0950848833,0.4156052172,0.0535985343,-0.1083418354,0.2208644897,-0.1395664066,-0.0260134172,0.2909347117,-0.2630017996,0.0133082671,0.2766338587,0.1419247389,-0.0343248211,0.2441204339,0.2616045177,-0.21285294,-0.2781764269,-0.070929803,0.2034056783,0.3885403872,-0.3480496109,-0.0002685353,0.3647504449,-0.4026501775,0.4058004618,0.3052506745,0.1680756807,-0.0470950566,-0.258131206,0.0765385777,-0.0169942752,-0.4340025783,-0.020303281,0.1415560097,0.1105765998,0.0487362295,-0.0601711161,0.159247458,0.2790450156,0.2702627778,-0.0576715805,0.4003246725,0.0700719953,0.0536593162,0.2820592821,0.2163243592,0.5008888245,0.1520393342,0.0615330748,0.3375333846,0.0667119846,-0.0489498749,0.0231666267,-0.1415083408,-0.1176150665,0.3134403527,-0.0147500196,-0.0715856478,-0.3975217044,-0.2155854702,-0.0502537452,0.0992040187,0.0855665654,-0.1450577676,0.2354638726,0.0524963774,-0.1196645424,-0.2443861812,-0.0430718996,-0.0664650947,0.5451594591,-0.1991377473,0.0696593449,0.2658621371,-0.0356095135,0.0963545516,0.402122885,0.1985036433,0.0789157301,-0.2381747067,0.1088349149,-0.0445786901,-0.1440981328,-0.1425260752,-0.0305382255,0.147759378,-0.2748593986,0.349375397,0.2631461918,-0.3019365072,-0.3211148977,0.3872309923,-0.1345878541,-0.0100665605,-0.6193216443,-0.002628427,0.0443207249,-0.1521770805,0.1048784405,-0.5363394022,-0.0952254608,-0.0885432363,-0.0321343392,0.0277196635,-0.1648291796,0.1338630617,0.0752426833,0.4413732886,0.2891284525,0.0606763028,0.0619646944,-0.2279345095,-0.5278967023,0.116753839,-0.0799406394,-0.4166087508,-0.1129243895,0.3505662084,0.2454901189,0.2385322601,0.1003209874,-0.1506654024,0.2423653901,0.0709742904,-0.2875417769,0.0808582902,0.0561882034,0.0541797392,-0.0941887349,-0.1719164848,0.1251953542,-0.2471843213,0.3233294487,-0.241175577,0.1120429486,-0.1701765209,0.0729581118,0.0646916777,-0.0272890981,0.1814373881,-0.4356124699,-0.3198751509,-0.0362593234,0.0115888026,-0.1519787014,0.1135638803,0.156106621,0.5797201991,0.0692493692,-0.0713019744,0.1066633314,0.4916709661,0.0201849677,0.0295590572,0.0948422104,0.031026803,0.2517523468,-0.1997822076,-0.078859888,0.396351546,-0.0907165185,-0.1283352971,-0.055131916,-0.620397687,0.1190891713,-0.439666748,-0.1874820441,-0.0373212919,-0.1397859603,0.1088624224,0.3996539116,-0.3055769801,0.2087804824,0.2238700241,0.1930085868,-0.2440024018,0.3604778051,0.1155558676,-0.1429820657,-0.0835569948,-0.1805303544,-0.0429826528,-0.0756496042,-0.2174596637,-0.3298113346]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2902","title":"Add WIT Dataset","comments":"> @hassiahk is working on it #2810\r\n\r\nThank you @bhavitvyamalik! Added this issue so we could track progress \ud83d\ude04 . Just linked the PR as well for visibility. ","body":"## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":28,"text":"Add WIT Dataset\n## Adding a Dataset\r\n- **Name:** *WIT*\r\n- **Description:** *Wikipedia-based Image Text Dataset*\r\n- **Paper:** *[WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning\r\n](https:\/\/arxiv.org\/abs\/2103.01913)*\r\n- **Data:** *https:\/\/github.com\/google-research-datasets\/wit*\r\n- **Motivation:** (excerpt from their Github README.md)\r\n\r\n> - The largest multimodal dataset (publicly available at the time of this writing) by the number of image-text examples.\r\n> - A massively multilingual dataset (first of its kind) with coverage for over 100+ languages.\r\n> - A collection of diverse set of concepts and real world entities.\r\n> - Brings forth challenging real-world test sets.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n> @hassiahk is working on it #2810\r\n\r\nThank you @bhavitvyamalik! Added this issue so we could track progress \ud83d\ude04 . Just linked the PR as well for visibility. ","embeddings":[-0.0809517056,-0.2214551866,-0.1989731491,0.0133092012,0.0548566766,0.0012881586,0.4317961633,0.0905278698,0.2972510159,0.1470830888,-0.0140026174,0.117820628,-0.1801292151,0.2750623822,0.1759230644,-0.2347300649,0.0528856255,-0.0525587946,-0.111166127,-0.0794195235,-0.0522569418,0.1980129927,-0.193837896,-0.2477924973,-0.398671031,-0.1057200357,-0.1739888638,-0.0169518292,-0.1846639961,0.0327739008,-0.2608135045,0.0733176619,-0.1156978458,0.4770053923,-0.0000964032,-0.064579986,0.2143842131,-0.1423904598,-0.0545602031,-0.1777072847,0.041535154,-0.0049116011,-0.2549009025,-0.2813225985,-0.1839788407,-0.1538777053,0.2700455189,-0.0029812963,0.3053423464,0.1210975945,0.3088106811,-0.1474685222,0.2071530968,-0.0514375716,0.1326429099,0.1544086039,-0.248274833,-0.0964364931,0.0966394767,-0.2267570943,-0.2421554029,0.6854541302,0.1160565093,0.0250154659,-0.0252820924,0.0070958599,-0.0926502421,-0.2954958975,0.188639313,0.4700401127,0.5937217474,-0.1630618572,-0.1721827239,0.0919661224,-0.0487032495,0.0397691727,0.0408558957,0.3571046293,-0.1618400067,0.0649592951,-0.1783177257,-0.15133214,-0.1726877838,0.2914958894,-0.0949109718,0.4078658223,-0.0315455608,-0.0555812605,-0.0373772085,-0.1777147949,-0.2164238542,0.0523082763,0.2419596165,0.0662366971,0.0335867517,-0.1256911457,0.0385495387,-0.1135115251,0.3417511284,-0.2774721086,-0.0565746911,0.132901594,-0.323112458,0.3092396557,0.1385564804,-0.2174253464,-0.1450333446,-0.1854815036,0.1951016188,0.0325002782,-0.2619886994,0.1505577266,0.099721089,-0.0214241892,-0.2273956388,-0.160768792,-0.0516468845,0.0911392644,0.1193964332,-0.0707594156,-0.1421585381,-0.0661800429,-0.3177126944,0.2407010645,0.1010043994,0.0815588385,0.0965300798,0.2475896478,-0.1470988691,-0.1765169948,-0.0881534964,0.2248412669,-0.1903703958,0.1265021414,0.1592800766,0.5113365054,0.2537501454,-0.0733771846,0.1129089221,0.0018325694,0.0856605247,-0.1001399159,0.2069840133,-0.0557915792,0.0809679776,-0.1524138749,-0.073373422,-0.3102889955,-0.1555388868,0.0773518682,-0.0889911428,0.049377501,-0.4161147475,0.2530652583,0.1284849197,-0.0809181705,0.0908475667,0.7699179053,0.0582486838,-0.0342353247,0.1798475385,0.2833342552,-0.3797134161,-0.0025154876,0.1757738888,0.3905650079,-0.092420809,-0.0642552599,0.1328477561,0.2665217221,-0.0583941862,0.2173905224,-0.0567309335,0.087856859,0.0246495996,0.2322876602,0.0694603547,-0.1395289451,-0.0435003601,-0.1011100262,-0.0813610405,0.0106059145,0.0836546645,0.298145026,-0.0127935633,0.096528627,0.0080791311,0.6753434539,-0.1098164916,0.2286004871,-0.2737777531,-0.4278278649,-0.0113982931,0.328744024,0.145377025,-0.4055054188,0.3278430998,0.2322438657,0.0949867368,-0.3136895597,0.1085856259,0.0223203544,-0.0005051232,0.1467118561,0.1092848554,-0.2964729071,-0.0836762711,-0.130402267,0.1318242103,0.483746618,-0.248289302,-0.2480282038,0.0042855954,-0.2053590268,-0.0621923134,-0.2704664171,0.3617103696,0.17163001,-0.1306215972,0.2724723816,0.0430581756,-0.0919279084,-0.2533151209,-0.0622685514,-0.0406291224,0.1581876874,-0.2046279311,-0.0177770704,0.0472491309,0.1987008452,-0.2961709499,0.0533348955,0.1038889065,0.1865651459,0.0792258084,0.3698846102,0.3873916864,0.1120804548,0.2921239138,-0.6291121244,0.1154151708,0.0760285705,-0.006126096,-0.0176064018,-0.3841995299,0.0472670086,0.19283472,-0.0016580714,0.1144336909,0.073393099,0.2081537247,-0.0956783444,-0.0049937302,-0.366854459,0.0016061328,0.1734439284,-0.1884854436,-0.1824615896,-0.2030260712,0.0807694122,0.5237088799,0.2258421183,0.5505995154,0.1254282445,-0.1420795768,-0.28885746,0.0549856015,-0.2941262424,-0.1655779034,0.3116272688,-0.0251247585,-0.1782427877,-0.0482349657,-0.085527949,0.1338160634,-0.0276928395,0.1904964596,0.0914436951,0.2914770544,-0.1149811968,-0.5630301237,-0.0705270022,-0.0223842766,-0.2100635022,0.0955632254,0.0532289781,0.0085637094,-0.2813546956,-0.0914483368,0.1090581045,-0.2741340697,-0.0460894257,0.3448168337,-0.0612961911,-0.1088433415,0.2226289511,0.2646664679,0.2659244537,0.0535228476,-0.0327079035,-0.0569329187,-0.4480122924,-0.0012578141,0.2744467258,0.3094526529,0.0774371698,0.4492533505,-0.1706188321,-0.0032700298,-0.1970399022,-0.6828343272,0.1931857914,-0.1799332798,0.222297132,0.0465350263,0.0598002523,-0.2162089497,-0.1784383208,0.1267048568,-0.1550560743,-0.1278322637,-0.0514384434,0.0643266812,0.0791968629,-0.171182245,-0.2651809454,-0.1608145237,-0.3226848841,0.3241012394,0.2365675569,0.11344897,-0.0567815825,0.3845591247,0.0518014841,-0.0231321547,0.2783186138,-0.3127382994,-0.0589283146,0.1515590996,-0.3646040261,-0.3476807475,-0.2444108576,-0.1988812536,0.2525679469,0.0808433667,-0.4292116463,-0.3522478342,0.0311818309,0.2061463892,0.1419826597,0.0842591301,0.2448365092,0.190408811,-0.2793112099,-0.1548068672,-0.0041617351,0.0255124755,-0.1157210916,0.0760340467,0.1783276647,0.0139507679,0.2965382338,0.4465199113,-0.2176057845,0.1476778239,0.3842280805,0.0036174529,0.1714651883,-0.2575018108,-0.1082883179,0.5114590526,-0.0155883245,0.0532231331,0.3055923581,0.1387230456,-0.493773073,-0.1047745943,0.1593842357,-0.5112806559,-0.1331510544,0.1892561316,0.1403896511,0.1600528061,-0.1513719559,-0.3017922044,-0.175591737,-0.1879560053,0.1834862083,0.255494982,-0.0663223863,-0.0857872739,-0.1768642962,0.0369856954,-0.121113494,0.0471084453,0.3025898635,-0.1103605628,-0.1657053381,-0.1256040633,-0.0691812932,-0.1423523277,0.2711536586,-0.3945429325,-0.1802496612,0.1024449319,-0.0627269968,-0.0776354223,0.2761702836,-0.330293119,-0.2617784739,0.0633344948,0.3113533854,-0.3719662428,0.0762865916,0.4057539105,0.1531731486,0.0045682145,-0.0382465012,-0.0839486346,-0.2303083241,-0.4416763484,0.0756094828,-0.0475242175,0.1286154091,0.2946779728,-0.0733427703,-0.0314382799,-0.0484548584,-0.0901324525,0.2570098937,-0.0899744332,0.2848593593,-0.1690306813,-0.1259722412,-0.1103044376,-0.1840956509,0.1707034856,-0.0802198648,-0.0567748919,-0.2014106512,-0.2966367304,0.0954907387,0.0731393769,0.2481430173,0.0807809383,-0.0638768226,-0.0167128611,-0.3351784945,0.1609588414,0.2288006246,0.2169885933,0.0056775264,-0.2819871008,0.3445268273,-0.073021248,0.039783787,0.3688854277,-0.1323992908,-0.4168664217,0.3294692636,0.3667802513,0.9066294432,-0.1412285864,0.2516909838,0.0521108657,0.0443725064,0.379629761,-0.2828923464,0.0604830906,0.0479294434,0.0563858896,-0.0548547022,0.1923683733,-0.0117860539,0.0325935893,-0.275164932,-0.1745013148,0.3504127562,0.1096888706,0.0581242368,0.3463208973,-0.0501323305,-0.4585259557,-0.2810158432,0.2333678752,0.0736545548,-0.1327342242,-0.0186843965,-0.0988537818,0.0398562253,-0.0720824152,-0.3393507898,-0.0706250221,-0.3228843808,0.2183895111,-0.308116734,-0.3472005725,0.1668397635,0.2316637933,0.2130430043,0.1262043267,-0.2604316473,0.2059072405,-0.0819388479,-0.1436516792,0.0000350554,-0.0936641917,0.364066422,-0.1577276438,-0.1965638399,0.0312893242,0.2719175816,-0.1914389282,-0.4962573647,-0.1763629317,0.3231757283,0.2469760776,0.0594019704,0.2734151781,-0.420081526,-0.279900074,0.2923044562,-0.1240278035,0.139734447,0.090626359,0.2457036376,-0.0870910808,-0.0481114164,0.2409695685,0.2811488807,0.0817728937,0.2042908669,0.3132542372,-0.1231459677,-0.4118822515,-0.2268985659,0.0112832189,-0.3043795824,-0.0997028574,0.075029701,-0.3576362729,-0.1853453815,0.2977905869,0.0945064574,-0.0108859651,-0.0280812215,-0.104707174,-0.0504857972,0.4871656299,-0.1233899519,0.1704572886,0.0498322472,0.1368637681,0.1661450267,0.2778989673,-0.4913753569,0.0999070555,-0.1608650386,0.1617065072,0.1140983626,-0.2118062377,0.0391213,0.1311940104,0.2852498591,-0.0028897552,-0.1713041216,-0.3745980859,-0.1606842279,0.0489701405,0.0406194441,-0.0218042042,0.2007368803,-0.1021691784,-0.075192146,-0.2093645781,0.170140624,-0.0332107544,-0.235269323,0.1865300387,-0.1608070433,0.1187050864,0.1141011491,-0.1843488514,-0.0202733669,0.2034232765,0.1748362631,-0.1693175435,0.2350271195,-0.090575017,0.0716959387,0.2352731675,-0.1609387249,0.1508632898,0.22368294,0.1156171039,0.046740178,0.0889115855,0.2442272305,-0.2031773925,-0.3575416207,0.0098930569,0.2128424644,0.3950456977,-0.3788528442,-0.1032667682,0.1637059599,-0.3952121437,0.4144688547,0.2580401301,0.1857469529,-0.0102851801,-0.1041494682,0.0706411228,-0.1911442429,-0.3604948521,-0.0429542065,0.0436507836,0.0383596979,0.0862421021,-0.1259579062,0.1100242585,0.2090563923,0.2760358453,-0.0548093542,0.3306506574,-0.0164462365,0.0951900482,0.3387245536,0.2413451821,0.4937801957,0.0987096876,0.0264338162,0.3570222855,-0.0418578535,0.2647001147,0.1094167084,-0.034785334,-0.1574991196,0.187232703,0.0186632369,-0.1193159148,-0.2749849856,-0.1580919325,-0.1344575584,0.0357709043,-0.0121378247,-0.0608970672,0.1425048411,0.0659468845,-0.1128230691,-0.201325193,0.0615972467,-0.0594317168,0.4502527118,-0.2346452177,0.2587008178,0.1951896548,-0.0046939291,0.0203922391,0.4529597461,0.112104319,0.019977089,-0.1175721884,0.0875313953,-0.114308767,-0.061771743,-0.1920992732,0.0724602044,0.3947676718,-0.0750949383,0.2475585938,0.236138463,-0.3471648097,-0.1523065567,0.4345367253,-0.1978103966,-0.0468036309,-0.5462716818,0.0078052361,0.0408355705,-0.2892584205,0.0690495074,-0.637196064,-0.0430860594,-0.0008575172,-0.0671569929,0.0473668277,-0.0951450542,0.1405508965,0.1236739159,0.4421503842,0.3087599874,0.1302765012,0.0961523876,-0.3838952184,-0.446438849,0.165471822,-0.1000231653,-0.0665667132,-0.1452491432,0.321154058,0.2846552432,0.2771971524,-0.0085254181,-0.1527163386,-0.0146443518,0.1127069667,-0.3975604475,0.1767693162,0.0961188823,0.1993730366,-0.0299789477,-0.1019918174,0.1034432575,-0.056199532,0.2615951002,-0.2036397755,0.102962628,-0.1265001148,-0.0658517629,0.1697728336,-0.1092779413,0.0920383409,-0.4555152953,-0.3111367822,-0.1391869038,-0.1233393773,-0.1475510448,0.2659368515,0.2873288691,0.4175904393,-0.0261233281,-0.1063955426,0.2146098614,0.4109917879,-0.0069877491,0.081328705,0.0300202575,-0.0731000528,0.3030554354,-0.1591516882,-0.098505877,0.4811171591,-0.1448982358,-0.2175184041,-0.0372310281,-0.5459349155,0.0817107856,-0.3308123648,-0.2261419147,-0.0765321776,-0.0824959278,0.0465272143,0.2357187569,-0.4232021868,0.1949178725,0.2078907192,0.3215321004,-0.0627836511,0.3240291178,0.1668283939,-0.1704506576,-0.1509878784,-0.0996923894,0.043273434,-0.0744077042,-0.1939665377,-0.382758081]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2901","title":"Incompatibility with pytest","comments":"Sorry, my bad... When implementing `xpathopen`, I just considered the use case in the COUNTER dataset... I'm fixing it!","body":"## Describe the bug\r\n\r\npytest complains about xpathopen \/ path.open(\"w\")\r\n\r\n## Steps to reproduce the bug\r\n\r\nCreate a test file, `test.py`:\r\n\r\n```python\r\nimport datasets as ds\r\ndef load_dataset():\r\n ds.load_dataset(\"counter\", split=\"train\", streaming=True)\r\n```\r\n\r\nAnd launch it with pytest:\r\n\r\n```bash\r\npython -m pytest test.py\r\n```\r\n\r\n## Expected results\r\n\r\nIt should give something like:\r\n\r\n```\r\ncollected 1 item\r\n\r\ntest.py . [100%]\r\n\r\n======= 1 passed in 3.15s =======\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\n============================================================================================================================= test session starts ==============================================================================================================================\r\nplatform linux -- Python 3.8.11, pytest-6.2.5, py-1.10.0, pluggy-1.0.0\r\nrootdir: \/home\/slesage\/hf\/datasets-preview-backend, configfile: pyproject.toml\r\nplugins: anyio-3.3.1\r\ncollected 1 item\r\n\r\ntests\/queries\/test_rows.py . [100%]Traceback (most recent call last):\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/runpy.py\", line 194, in _run_module_as_main\r\n return _run_code(code, main_globals, None,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/runpy.py\", line 87, in _run_code\r\n exec(code, run_globals)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pytest\/__main__.py\", line 5, in \r\n raise SystemExit(pytest.console_main())\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/config\/__init__.py\", line 185, in console_main\r\n code = main()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/config\/__init__.py\", line 162, in main\r\n ret: Union[ExitCode, int] = config.hook.pytest_cmdline_main(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_hooks.py\", line 265, in __call__\r\n return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_manager.py\", line 80, in _hookexec\r\n return self._inner_hookexec(hook_name, methods, kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 60, in _multicall\r\n return outcome.get_result()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_result.py\", line 60, in get_result\r\n raise ex[1].with_traceback(ex[2])\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 39, in _multicall\r\n res = hook_impl.function(*args)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/main.py\", line 316, in pytest_cmdline_main\r\n return wrap_session(config, _main)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/main.py\", line 304, in wrap_session\r\n config.hook.pytest_sessionfinish(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_hooks.py\", line 265, in __call__\r\n return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_manager.py\", line 80, in _hookexec\r\n return self._inner_hookexec(hook_name, methods, kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 55, in _multicall\r\n gen.send(outcome)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/terminal.py\", line 803, in pytest_sessionfinish\r\n outcome.get_result()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_result.py\", line 60, in get_result\r\n raise ex[1].with_traceback(ex[2])\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 39, in _multicall\r\n res = hook_impl.function(*args)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/cacheprovider.py\", line 428, in pytest_sessionfinish\r\n config.cache.set(\"cache\/nodeids\", sorted(self.cached_nodeids))\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/cacheprovider.py\", line 188, in set\r\n f = path.open(\"w\")\r\nTypeError: xpathopen() takes 1 positional argument but 2 were given\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Linux-5.11.0-1017-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":19,"text":"Incompatibility with pytest\n## Describe the bug\r\n\r\npytest complains about xpathopen \/ path.open(\"w\")\r\n\r\n## Steps to reproduce the bug\r\n\r\nCreate a test file, `test.py`:\r\n\r\n```python\r\nimport datasets as ds\r\ndef load_dataset():\r\n ds.load_dataset(\"counter\", split=\"train\", streaming=True)\r\n```\r\n\r\nAnd launch it with pytest:\r\n\r\n```bash\r\npython -m pytest test.py\r\n```\r\n\r\n## Expected results\r\n\r\nIt should give something like:\r\n\r\n```\r\ncollected 1 item\r\n\r\ntest.py . [100%]\r\n\r\n======= 1 passed in 3.15s =======\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\n============================================================================================================================= test session starts ==============================================================================================================================\r\nplatform linux -- Python 3.8.11, pytest-6.2.5, py-1.10.0, pluggy-1.0.0\r\nrootdir: \/home\/slesage\/hf\/datasets-preview-backend, configfile: pyproject.toml\r\nplugins: anyio-3.3.1\r\ncollected 1 item\r\n\r\ntests\/queries\/test_rows.py . [100%]Traceback (most recent call last):\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/runpy.py\", line 194, in _run_module_as_main\r\n return _run_code(code, main_globals, None,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/runpy.py\", line 87, in _run_code\r\n exec(code, run_globals)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pytest\/__main__.py\", line 5, in \r\n raise SystemExit(pytest.console_main())\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/config\/__init__.py\", line 185, in console_main\r\n code = main()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/config\/__init__.py\", line 162, in main\r\n ret: Union[ExitCode, int] = config.hook.pytest_cmdline_main(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_hooks.py\", line 265, in __call__\r\n return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_manager.py\", line 80, in _hookexec\r\n return self._inner_hookexec(hook_name, methods, kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 60, in _multicall\r\n return outcome.get_result()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_result.py\", line 60, in get_result\r\n raise ex[1].with_traceback(ex[2])\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 39, in _multicall\r\n res = hook_impl.function(*args)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/main.py\", line 316, in pytest_cmdline_main\r\n return wrap_session(config, _main)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/main.py\", line 304, in wrap_session\r\n config.hook.pytest_sessionfinish(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_hooks.py\", line 265, in __call__\r\n return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_manager.py\", line 80, in _hookexec\r\n return self._inner_hookexec(hook_name, methods, kwargs, firstresult)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 55, in _multicall\r\n gen.send(outcome)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/terminal.py\", line 803, in pytest_sessionfinish\r\n outcome.get_result()\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_result.py\", line 60, in get_result\r\n raise ex[1].with_traceback(ex[2])\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/pluggy\/_callers.py\", line 39, in _multicall\r\n res = hook_impl.function(*args)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/cacheprovider.py\", line 428, in pytest_sessionfinish\r\n config.cache.set(\"cache\/nodeids\", sorted(self.cached_nodeids))\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/_pytest\/cacheprovider.py\", line 188, in set\r\n f = path.open(\"w\")\r\nTypeError: xpathopen() takes 1 positional argument but 2 were given\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.12.0\r\n- Platform: Linux-5.11.0-1017-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nSorry, my bad... When implementing `xpathopen`, I just considered the use case in the COUNTER dataset... I'm fixing it!","embeddings":[-0.3016421199,-0.1903872788,-0.0241492949,0.0990293771,0.3130396903,-0.1301406622,0.3114988804,0.2646481693,-0.1026777923,0.2515563369,-0.0329412185,0.3579744101,-0.0499672778,0.0878717452,-0.1619878709,0.1268500686,-0.0505254902,0.0685187206,0.0408935249,0.1136002094,-0.5712682605,-0.1570851803,-0.2168759555,0.0572490133,0.0126794586,0.0686679184,0.0916296914,0.2422822416,-0.0774931237,-0.469889611,0.1960368752,-0.0263328627,-0.0422579646,0.4722329378,-0.0001045304,-0.0628838241,0.4584931433,0.2026191652,-0.4859764278,-0.2095862329,-0.1490363926,0.2063343376,0.2531500459,-0.254386425,-0.1224498227,-0.2215161026,0.0800599158,-0.5224384069,0.0328101404,0.2298583686,0.2646005154,0.5858446956,0.1764907539,-0.1220323741,0.3920012712,-0.1865620166,-0.0988033265,0.1512855589,0.2791418135,0.0342658088,-0.1552184969,0.2738155425,-0.251668334,0.2581939995,0.0830629617,0.0403149687,0.1071799397,0.1236744151,0.0054618004,0.1477268934,-0.0244684517,-0.2447356284,-0.187163204,0.0132809207,-0.1560323685,-0.4709234536,0.0660294145,0.2990931869,-0.1927743256,0.0707391798,-0.1533978879,0.285312593,-0.2178352475,0.2086598277,-0.0414643139,0.474942416,0.0106467595,0.0582296215,0.0282520335,0.0535728522,0.3718392253,-0.029089909,0.0258877985,-0.2166382521,-0.0353300683,0.0447735675,0.3098430037,0.1299089044,0.1545516998,0.1684173644,-0.0050124396,-0.0647071525,0.1099802256,0.1257055998,-0.0156705379,-0.0932240486,-0.0586434491,0.1788802147,0.1195266545,0.1648669541,-0.1360280216,0.0084639816,-0.0668128058,-0.3070941567,0.0900237709,0.1294367164,0.5126132965,-0.0637784451,-0.3585409522,-0.1637833863,-0.3168658912,0.1315581501,-0.0887979344,0.1646843106,-0.3344549835,0.0525158457,0.2567434609,0.1665857583,-0.3985979259,0.1734841466,-0.1729390174,-0.0025560693,-0.2071772516,-0.1981868595,0.0244949311,-0.0061054123,0.0369944014,0.0258494634,-0.0169168096,-0.0553664789,0.1729714125,-0.0304495171,0.0978282169,0.0834876001,0.1038077101,0.0230855085,0.0646957457,0.2365556359,0.0004667089,0.4194874167,-0.1334887296,-0.2440327555,0.1164916679,0.35601753,0.0041353889,-0.1565836817,0.1820517033,-0.3767896891,-0.1220365986,-0.0800681487,0.0026528577,-0.2985853255,-0.1478585154,-0.2639125884,0.2254217416,0.2425370067,0.2605780959,0.0505344905,-0.0282764584,0.1627342254,0.3797311783,0.1669939458,-0.1939092427,-0.0896540061,-0.1379573345,0.1596079618,-0.0486291014,-0.4239594936,-0.3278156817,-0.1607752889,-0.1731502265,-0.07503663,0.049631156,0.0323371701,0.3179516196,0.0479337424,-0.0318585075,0.412953496,-0.0740917623,0.0748518556,-0.4666415751,-0.3798174262,0.5641200542,0.151705578,0.0570301823,-0.3497416377,0.2054489851,0.0608398058,0.5148956776,-0.1897210926,-0.1518411934,-0.2993411422,0.752466023,-0.0474503674,0.0110379141,-0.0960698351,0.0665274039,0.0204419531,0.2316295356,-0.2802456915,-0.1499119997,-0.0891542956,-0.1524517834,0.1319255084,-0.3466832638,-0.382157594,0.2474724054,0.2782579362,-0.1456098258,-0.1309039593,-0.0118953492,-0.0933069214,-0.3045151532,0.100494273,0.0386145003,0.1053284183,-0.0396609642,-0.094124943,-0.1193556041,0.298748225,0.162939176,-0.0600543693,-0.1179709211,0.3312662542,0.2449479997,0.1955460757,-0.0468190089,0.0072500408,0.0820179358,-0.3156889975,-0.2024822384,0.3221007586,0.0975408554,-0.002581625,0.0227513239,0.1893991232,-0.2354014963,0.1002926081,0.2986403406,0.0012304113,-0.0390900932,-0.1708397567,0.0159268808,0.3352193236,0.253539145,0.4091952145,0.4384313226,0.0471034944,-0.1185913086,-0.0291307271,0.1720729768,0.0163347404,0.2327632755,0.1700844765,-0.400013119,-0.0777249038,-0.0843988508,0.3033127487,0.3952056468,0.275578618,0.0969723538,0.0887242258,-0.1324126124,-0.2870429754,0.2992976308,0.1152055711,0.0208725408,0.2415834963,0.2297437042,0.1021427438,-0.2554797828,-0.0732267201,-0.1389924884,0.0566167571,-0.1460091919,-0.0478304811,-0.1640127003,0.2011270672,-0.1140680909,-0.5005605817,0.1816411167,-0.2008605003,0.0143492073,0.3810865879,-0.2022752911,0.1200969592,-0.2770578265,0.0942122191,0.0885685533,-0.0884958655,0.1415922791,-0.2295171022,-0.1760160774,0.1794952601,0.0489556827,-0.1769523323,0.3923769593,-0.0935229883,-0.0265750755,-0.1780978739,-0.1293669194,0.0228035133,0.2426382601,0.0753103569,0.111499086,0.05481565,0.108430557,-0.2810976505,0.3347675204,-0.2257572263,-0.0527905673,0.1222594678,0.0974359363,-0.0859619454,-0.1513888091,-0.6303583384,-0.1750367284,-0.3355152905,0.0044383388,-0.107812658,0.1289740354,0.4354717731,0.0387976281,0.1618680805,-0.1536655724,0.121643737,-0.0872593373,-0.0211115107,0.2169368863,-0.1349027008,-0.4010963142,-0.198347047,-0.1213339865,-0.1045445204,0.433926791,-0.2609218061,-0.2592196167,-0.2313274145,0.0872525126,-0.1294534057,-0.0232153423,0.3986389339,0.2484317124,-0.2569616735,-0.2864581048,0.0643375963,-0.0433718897,-0.0686287582,0.0381864421,-0.087544404,0.1517151892,0.1127801165,0.2772788107,-0.0392005183,-0.0387176163,0.1129376143,-0.2495963275,0.1978779882,-0.0433227047,-0.256916225,0.1621052921,-0.2219971567,-0.3214866221,0.0484665148,0.2195314169,0.1148819104,0.0141728139,0.0208034664,-0.3543073237,-0.2621160448,-0.2463587523,0.3780513704,0.0837339088,-0.0672197193,0.1626053154,-0.1316152364,-0.1256516725,0.1881820112,0.3759931922,0.1093446314,-0.0119339237,0.2482949048,0.0123445923,-0.0862110257,0.0998784602,0.0090176854,0.1281881332,0.0566877872,-0.1912219673,0.0785223171,-0.0200078636,0.164327085,0.3458769023,0.387481302,0.3010594249,-0.1472473443,-0.2551456392,0.0189636461,-0.5182095766,-0.3018001318,0.0700561032,0.1623141319,-0.2750703692,0.0511011817,0.3745272458,0.1937805861,0.0627759919,0.0518822372,0.0086369924,-0.4112387002,-0.0766517594,0.0733634308,-0.0133935297,0.4511975646,-0.036897406,-0.2186553776,-0.3560897112,-0.333645165,-0.0098668383,-0.0695359781,0.410104692,-0.161490202,0.1817578077,0.0064769732,0.3672036827,0.3949998319,0.21570158,0.1406856179,0.1515156925,-0.007653513,-0.1830333024,0.0305453707,0.1035782769,-0.352021575,0.2259793282,-0.3184482455,0.2777011096,-0.2909503579,0.1956375539,0.3502264917,0.1843060404,0.2732339501,0.0076183779,0.1056670472,-0.1933567673,-0.0941651911,0.5531533957,-0.1504649967,-0.2331054062,-0.1227031499,0.1427025497,0.798979938,0.2282719463,0.1575280577,0.8578522205,-0.379422754,-0.1998089552,0.3360408843,0.1782059669,-0.3571628928,0.0383414142,0.0812166259,-0.228775233,0.0273007564,-0.2691111565,-0.0185519699,-0.1856172085,-0.0207111835,0.0446326248,0.1084755659,0.0523586944,0.040253371,0.0424530245,-0.2928382456,0.1929394305,-0.0179150142,0.1270106137,-0.0508781746,-0.0839509666,0.2480623722,-0.1010967717,-0.2043585628,0.3419599831,-0.5799904466,0.1721819341,-0.2692230642,-0.4562358558,-0.1075189933,-0.0367897227,-0.2285494804,0.1678056121,-0.034031637,0.1088864431,-0.025018584,0.1675112695,0.2350722551,0.0307645444,0.2065542489,-0.2243107855,0.0097956462,-0.1305310428,0.0516777299,-0.4160594046,-0.0600328259,0.0467677265,0.494930625,-0.2594941854,-0.2402201593,-0.2332240641,-0.1814632267,-0.3135359883,0.2200746089,-0.0131482948,0.1351002008,0.1709565818,0.0870712101,-0.3995954692,-0.1094917506,-0.0123062516,0.0119483415,0.1383818239,0.4005638361,0.1243489757,-0.3258515,-0.2226490825,0.1260189861,0.1640535593,-0.0271104369,0.0868695006,-0.0412157513,0.1951783746,-0.1118135676,0.0058581447,0.2094234228,0.0991211757,-0.1979446858,-0.4055492878,-0.1660560668,0.1417739689,0.2028817385,0.1181820855,-0.0264829546,0.2930473983,-0.0213118494,-0.250485301,-0.3906231225,0.1107914075,0.2336137593,0.1192180067,-0.1454875469,-0.2076455355,0.3068408668,-0.0104081798,0.2542708516,-0.051687777,-0.1347726136,-0.3061599433,-0.3753732443,0.0859213248,0.0782093406,-0.3529872596,0.1614413857,-0.1700136662,0.0037516817,-0.1876564473,0.1137397587,0.3402270079,-0.2084238827,-0.0727987811,0.3792553246,0.1400492191,-0.2362306863,0.0073809871,-0.2387416065,0.2812373936,-0.1352637857,-0.0331278555,0.1245579347,-0.0565308221,-0.1441801339,-0.2388965189,0.2134499103,0.1605943888,0.2762319148,-0.1227833033,-0.11529129,-0.1517725885,0.5579532981,0.198146686,-0.1018952206,-0.0254687592,0.0657821447,0.377563864,-0.2160446793,-0.0246563014,-0.1456676871,-0.1745386422,0.1609793305,-0.06970983,-0.4467020631,0.0724190399,-0.1336331218,0.0843940005,-0.0011720849,-0.1661704034,-0.0122293616,0.1776824147,-0.1440191418,-0.0499875061,0.0935866535,-0.0745285526,0.1813107729,0.2331058681,-0.0133844428,0.6520751119,0.5496276617,-0.2443293333,0.0571718849,-0.0549315177,0.0069860495,-0.1236407459,-0.0362573415,0.0326864682,0.0409415178,0.2288151681,-0.2829382718,0.1047509387,-0.0511482842,-0.0435396358,-0.0990097001,-0.2635801733,0.0496004447,-0.0047050342,-0.0728645399,-0.0293682516,-0.1020355001,0.0224539246,0.0303376559,0.0476331823,-0.1048749685,-0.3494413197,0.0213044882,0.2079137266,0.1573521048,-0.3372322619,0.0876498446,0.218774423,0.143193692,-0.1703726351,0.4733753502,0.330448091,0.4719161391,-0.2293340415,-0.1304816902,-0.1675468981,-0.0063679451,-0.0394412205,-0.0647637993,0.013415969,0.1269380897,0.2298047692,0.3690746725,-0.3679998517,0.2118322849,-0.0242782552,-0.2714368403,-0.1849635541,0.2795254588,0.1443216503,-0.2951877415,0.0125996275,-0.0279961135,-0.2422025651,-0.0668567345,0.2900865078,0.2645173967,0.0876090899,-0.0397080444,0.1404092163,0.1440564841,0.1232805997,0.1019615456,-0.007067997,-0.4234194756,0.0328579731,-0.4711897075,0.4904586673,-0.1441804022,-0.446585238,0.0725421757,-0.0549177974,-0.2503592968,0.2345560342,0.3334582746,0.3571134806,0.1524520814,-0.0025301664,-0.255433023,0.1322989017,0.2028841376,-0.1032729968,-0.1260176003,-0.3461897969,0.2244306803,-0.2395664901,0.2582560182,0.0934694856,0.0410272479,0.2902907431,0.0138179325,0.017829489,-0.0170892756,0.395065397,-0.320538342,0.0409307852,-0.1458014548,-0.2275678366,0.0252156015,0.0532084107,0.0172631126,0.0858793035,0.1734698564,-0.5986049175,-0.434243381,0.4738562405,-0.3005673587,-0.1423755139,0.1085444987,0.3310374022,0.0011340193,0.1555804163,-0.037431892,0.1749423742,-0.1264825761,0.0589323528,-0.2458039671,-0.7254006267,0.2943550944,-0.7327613235,-0.2969195843,-0.2304133028,0.460382551,0.0212493073,0.238044709,-0.2867960334,0.1030114964,0.2128991187,0.0478451774,-0.1016833112,0.1259424686,0.017026538,-0.023835374,0.1139790639,0.6136496067,0.1038713083,-0.0976273566,-0.0312423818,-0.282785058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2892","title":"Error when encoding a dataset with None objects with a Sequence feature","comments":"This has been fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/2900\r\nWe're doing a new release 1.12 today to make the fix available :)","body":"There is an error when encoding a dataset with None objects with a Sequence feature\r\n\r\nTo reproduce:\r\n```python\r\nfrom datasets import Dataset, Features, Value, Sequence\r\ndata = {\"a\": [[0], None]}\r\nfeatures = Features({\"a\": Sequence(Value(\"int32\"))})\r\ndataset = Dataset.from_dict(data, features=features)\r\n```\r\nraises\r\n\r\n```python\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 2 data = {\"a\": [[0], None]}\r\n 3 features = Features({\"a\": Sequence(Value(\"int32\"))})\r\n----> 4 dataset = Dataset.from_dict(data, features=features)\r\n[...]\r\n~\/datasets\/features.py in encode_nested_example(schema, obj)\r\n 888 if isinstance(obj, str): # don't interpret a string as a list\r\n 889 raise ValueError(\"Got a string but expected a list instead: '{}'\".format(obj))\r\n--> 890 return [encode_nested_example(schema.feature, o) for o in obj]\r\n 891 # Object with special encoding:\r\n 892 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n\r\nTypeError: 'NoneType' object is not iterable\r\n```\r\n\r\nInstead, if should run without error, as if the `features` were not passed","comment_length":19,"text":"Error when encoding a dataset with None objects with a Sequence feature\nThere is an error when encoding a dataset with None objects with a Sequence feature\r\n\r\nTo reproduce:\r\n```python\r\nfrom datasets import Dataset, Features, Value, Sequence\r\ndata = {\"a\": [[0], None]}\r\nfeatures = Features({\"a\": Sequence(Value(\"int32\"))})\r\ndataset = Dataset.from_dict(data, features=features)\r\n```\r\nraises\r\n\r\n```python\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 2 data = {\"a\": [[0], None]}\r\n 3 features = Features({\"a\": Sequence(Value(\"int32\"))})\r\n----> 4 dataset = Dataset.from_dict(data, features=features)\r\n[...]\r\n~\/datasets\/features.py in encode_nested_example(schema, obj)\r\n 888 if isinstance(obj, str): # don't interpret a string as a list\r\n 889 raise ValueError(\"Got a string but expected a list instead: '{}'\".format(obj))\r\n--> 890 return [encode_nested_example(schema.feature, o) for o in obj]\r\n 891 # Object with special encoding:\r\n 892 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n\r\nTypeError: 'NoneType' object is not iterable\r\n```\r\n\r\nInstead, if should run without error, as if the `features` were not passed\nThis has been fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/2900\r\nWe're doing a new release 1.12 today to make the fix available :)","embeddings":[-0.0830345005,-0.2502439916,-0.0386255719,0.3015390933,0.2076890171,0.0788367391,0.5001555681,0.1494158357,0.1255292594,0.1421060264,0.3285635114,0.0508400798,-0.2967574298,0.3420606554,-0.2623355389,-0.2604982257,0.034535449,0.2224871665,0.0154916933,-0.0224474147,-0.3130667508,0.1709361374,-0.364877075,-0.2221626192,-0.1550774723,-0.2108083814,-0.2160241902,-0.2960246801,0.0542513058,-0.5987132192,0.2819482088,-0.0695153251,0.0921094567,0.2383219302,-0.0001144155,0.105492115,0.6381921768,-0.1453159899,-0.2628047466,-0.3394742906,-0.5618312359,-0.4227804542,0.1241655424,-0.3578473926,-0.0965121984,-0.0749715865,0.094718039,-0.3724971116,0.2042126656,0.2839667201,0.2130593061,0.3830007613,-0.0845509171,-0.0744024739,0.206820786,0.4162615538,-0.2607547045,-0.1798537225,0.035758879,0.2699487805,0.0946282744,0.3492396772,-0.111404337,-0.1305068284,0.2094347328,0.0849353001,0.0885619074,-0.4521939754,-0.2671251595,0.2014953941,0.2924095094,0.0013618435,-0.4084641933,-0.2180884778,0.0188714378,-0.6638943553,0.2666681111,0.055101525,-0.0386278071,-0.0269636735,-0.168392688,0.145001635,-0.2138954103,0.1660171747,-0.1209752634,0.0807909966,-0.1394540668,0.0341406651,-0.2402791679,-0.2700186968,0.0587494336,-0.3706248403,-0.0257517211,0.2111456394,-0.2515126467,-0.1936304271,0.167350471,0.0101849167,0.1697200388,0.1379934251,-0.0658706501,-0.1037984863,-0.0693587288,-0.0169596635,0.2529113591,0.2136024684,0.4112964869,-0.0680698305,-0.0581751354,0.1961243749,0.0996622145,-0.017598724,0.11928422,0.0517393723,0.0169261098,0.4072842002,0.6467193961,-0.0525628068,-0.393481791,0.2361950129,-0.3853594661,-0.0304983035,-0.0211958643,0.0061908085,0.0841203332,0.2916485071,-0.1465796083,0.2161677331,-0.2736037076,-0.250882864,-0.2348464429,0.0721954107,0.1018402427,0.0885985568,-0.0547403768,-0.0332995802,0.0658167601,0.1560105532,0.0796802491,0.1009850726,-0.0555230267,-0.365808934,0.0925061703,0.1523720324,-0.0070467223,0.2921590805,0.3762358427,-0.2581280768,-0.1091617644,0.1182429716,-0.2788110375,-0.217513755,-0.005827894,0.1725294441,-0.3795679808,0.0441264361,-0.0734036565,0.2263460904,0.2108575404,-0.0775901377,-0.0965437591,-0.0869289562,-0.2196164131,-0.1535144448,-0.0919654146,0.6465163827,-0.1091333926,-0.012587267,0.1191491932,-0.0898958445,0.3521232605,0.2327210456,0.0481479168,-0.057248231,-0.3404105306,0.1900297105,0.0831347257,-0.1115483418,-0.208607167,0.5657577515,0.0239907876,0.3598697484,0.1134639457,0.0959460512,0.1585700363,-0.3072297871,0.0130883651,0.1975988597,-0.1874651164,-0.2206795961,-0.2418985963,-0.1231627762,0.5206927657,-0.0743467882,-0.0640905127,0.2067834437,-0.3010617197,-0.0231112521,-0.1406468004,-0.3186881542,0.1288141012,0.5053259134,0.21923922,0.0375964157,-0.0967492238,-0.4028446674,-0.362685889,0.1312113106,-0.179603979,0.2839995623,-0.4932679534,-0.1799385846,0.0735202953,0.019041175,-0.3947464824,0.0015028021,0.1547810137,0.1214925349,-0.1535111219,0.1999227107,-0.0773415267,0.2772004306,0.1561118215,-0.1248554364,-0.567322731,0.3699874282,0.0695009977,-0.2262255251,0.0071669328,0.3033728004,0.3317838013,-0.0899984464,0.018318845,0.1505862623,0.0259702709,-0.3309904039,-0.3067805767,-0.1571382731,0.3190588653,-0.4185928106,-0.2214330137,0.3747457862,0.1825433373,-0.1073410809,-0.1209880039,0.6575422287,0.0157377589,0.4647912979,0.0042069959,0.1760141253,0.0537278354,-0.0018615762,-0.327142626,-0.2869090438,0.1755582988,0.2026411891,-0.1650790572,-0.0399117619,-0.3015944362,0.0058686477,0.4108926356,0.088436991,0.0354627408,0.0570802502,-0.0770579055,-0.0368382186,0.1092406586,-0.159593001,0.299883306,0.0505586714,-0.0732196867,0.0735782608,0.0075443899,0.0745387301,0.1758682579,-0.0350369215,0.2430823594,0.0285679307,0.1076480895,0.1176982298,-0.1779721677,-0.5438489318,-0.0398581289,0.1430890262,-0.2491215467,0.2492818087,-0.4795846939,-0.2056158483,0.130416885,-0.2173653096,-0.386295557,-0.4440228939,-0.1639413238,0.0275585894,-0.14083983,0.2236978859,0.0148140779,-0.26245597,0.2985749543,-0.0243482068,-0.0831376389,-0.1392291635,-0.0242327861,0.0463199466,0.1656642258,-0.0481024235,0.2467772514,0.0332959294,-0.3499837816,-0.2126868665,-0.2085345984,0.0565498397,-0.3503640592,0.0777079239,0.3079494238,0.1277272999,-0.04940686,-0.1998873949,0.4227285385,0.4243725836,-0.2120994478,0.3217695653,-0.102062583,-0.0354735814,-0.2110650837,-0.0959021896,-0.1572214514,-0.2278930247,0.1076047793,0.2109689265,0.2101610452,0.4471023381,0.0470193811,0.120879285,0.1355805248,-0.175628379,0.0008855638,-0.0150361629,0.4824748039,-0.1827880889,-0.2712480724,0.1487995982,-0.2987255454,0.0875612497,0.2041173279,-0.3575526774,-0.1763042659,-0.1413789541,0.1611307114,0.0745360553,-0.1046581939,0.3775742948,0.2816057205,-0.1083455756,-0.0367697217,-0.0508110113,0.2569004595,-0.0705505088,0.0287486054,0.229687199,0.5177567601,0.0398100056,0.0972354859,0.4376426637,0.1970631927,0.4261662662,-0.0823322833,0.0689360127,-0.1690145433,0.0005423338,-0.1711448282,0.107939221,-0.0178557411,0.0836275071,-0.2327573746,-0.1191914901,-0.0793068409,-0.0630706996,0.1038516238,-0.2357092947,0.129370749,-0.4676784277,0.442966491,-0.0864085481,0.0182908159,-0.3036096394,-0.0047005508,-0.1684736162,-0.0083024697,0.322756052,-0.1478672475,-0.64961797,-0.172231406,-0.2026044726,0.1215542853,0.3011380136,0.5467792153,-0.0337624177,-0.0624163561,-0.1075646654,-0.0191717129,0.4020332694,-0.1643890142,-0.1837625653,0.2454210073,-0.2638014555,-0.2387931347,-0.0567740202,-0.3692710102,0.3344596624,-0.2252197117,0.3618203402,-0.3405365348,-0.2225623876,0.2660458982,0.0742505565,-0.0913682804,-0.1132186204,-0.0910779387,-0.223561734,-0.1456012279,0.0129503747,0.2087670118,0.262157619,-0.1340866685,-0.0023913924,-0.5279824138,0.0307506509,0.3058232069,-0.2428039312,0.3555576503,-0.0859422684,0.1051599681,-0.1363119334,0.1940722018,0.2076593935,0.3552796543,0.377152741,-0.648413837,0.0111574912,-0.2274465859,0.1626977772,0.2772224844,-0.0934667066,0.0936513171,0.0833641216,0.1915255636,-0.2102062851,-0.3317123055,0.4189606011,-0.0805432349,-0.6726543307,-0.2872206569,0.2649118006,0.0967709273,0.2421812862,-0.1336611062,0.2232288569,-0.2113029659,0.3653744459,-0.0063009313,0.6850805283,0.4925385714,0.0419002809,0.5505322814,0.447142154,0.6043536663,0.4047942162,0.2582224309,-0.1625066251,-0.6721657515,-0.1485058814,-0.0585912243,0.4232234955,0.1191022769,-0.4531514645,0.489102751,-0.1281993687,-0.0638250262,-0.0981585309,-0.2965756059,-0.1317236423,-0.0439694487,-0.2515955269,0.1245702505,0.0693780556,0.1620190293,0.0573866554,-0.3246811926,0.0522958711,-0.1667520255,-0.1947969943,0.169654116,-0.1433667392,0.0941162854,0.0364244618,-0.3463888764,0.2818099856,0.2232395709,0.1879022419,-0.1312100291,-0.1560928375,-0.1969094723,0.0400298387,0.1034141481,-0.0801590979,-0.0777151361,0.3071270287,-0.0268944614,-0.1485314518,0.1475492716,-0.2142958194,0.1054213047,0.3429113925,0.0590512119,-0.1748204976,-0.1611669809,-0.453250885,0.0302399192,0.1710843444,-0.0500850677,0.1164650768,0.0069036391,-0.2920267582,-0.0495645739,-0.1978931725,-0.0086773541,-0.0171592813,0.3072853088,0.5229854584,0.0768130943,0.6932663918,0.191862449,-0.2426931262,-0.1660664529,0.1627267897,0.1830895394,-0.5128666759,0.3100555539,-0.1137086079,0.1459202468,-0.2419808209,0.4570830464,-0.2819413543,0.2090833783,-0.0393038131,-0.1175033078,-0.1571840644,0.2325696945,-0.1332800984,0.1057097316,-0.2226439267,0.1152285784,-0.0790469125,-0.2166758925,-0.2946244478,-0.0973041132,-0.4076433778,0.3669984043,0.1148904189,0.0858326256,0.2190838307,0.1607996225,0.1681335419,0.0856592953,-0.1367595494,-0.196850881,-0.0751663223,0.1268492639,0.1073024794,-0.1864513606,0.1410098374,0.0269025378,-0.0628488362,-0.1389683783,0.0139900362,0.1784923375,-0.0842979401,0.4305182099,0.3046940267,0.3934035599,-0.1181101874,0.160622254,-0.0320248976,0.2289986759,-0.0731261298,0.0898118019,-0.0715093166,0.0787385702,-0.4437326789,0.1356508732,-0.129166916,0.1195798442,0.2263306379,-0.3602280319,0.0355114751,0.3236742318,-0.0311139822,0.0755379125,-0.198323831,-0.0149546945,0.090865083,0.1674839258,-0.161023438,0.0044072028,0.1927999556,0.098377116,0.055364117,0.2497331053,0.0238154791,0.2197780013,0.1523765475,0.0410668887,0.0576352365,-0.1304038465,0.4767722189,0.2502485216,0.106006749,-0.005961901,-0.0451643243,-0.1361158639,0.2374778241,0.4095085859,-0.0225137156,0.1133541986,0.4354737401,-0.0183572639,-0.0299489778,-0.2667696178,-0.183229357,0.0874819532,0.3381926715,0.2028533518,0.1870117784,-0.3521421552,-0.3258129656,-0.2701639831,-0.069582805,0.3099346757,-0.0583300889,-0.0029459991,0.0356674194,0.0114391381,-0.3740221262,-0.2415987253,-0.0964923352,-0.1092930734,0.2701234221,0.2186554223,0.1762788445,-0.3771931529,0.27475968,0.2808687687,0.2680353522,-0.1144816577,0.2841907442,0.3918447495,0.0941931978,0.1932379305,0.5161973834,0.3695507646,0.2945841849,-0.020104045,0.1129500493,-0.0390780009,-0.3450121284,0.0877992213,-0.059845686,-0.2485655695,0.2879635692,0.1558267176,0.1228244901,-0.1254130453,0.3168381751,0.0544719622,-0.1103277802,-0.2857102156,0.7373996973,-0.2703433633,0.0276814178,-0.0651036352,-0.2521202862,-0.4581924081,-0.0652412251,0.381400764,0.1698071212,0.149515003,0.2177317739,0.0697703585,-0.4082437456,0.330963999,0.3403661847,0.0769058913,-0.2651322186,0.0617089309,-0.2750614285,0.1116446406,-0.1015280634,-0.0691227019,0.2495637238,0.1278142929,-0.1932554841,0.1148602813,-0.2204782218,-0.1393817365,-0.104323633,0.25245592,-0.3717963696,-0.1946512759,0.2883968651,-0.0892380476,-0.1251148432,-0.2652546763,0.0353319235,0.0039455947,0.0739230812,0.0550829805,-0.0448910445,-0.3227532506,-0.0278719831,0.057251703,0.6041781306,0.1074118167,-0.0593114123,-0.3922547102,-0.2225103676,-0.272254765,-0.1133324504,0.5871050954,0.2825134397,0.6090587378,-0.1371040046,0.2531865239,-0.0552069545,0.3776589036,-0.0609906539,0.0438415594,-0.4244903922,0.5184668899,-0.3719528913,0.1762697846,0.0068774358,0.1811163276,0.1390513331,0.1774709225,-0.0504962131,-0.2920152545,0.3989830613,-0.3084869087,-0.3456190228,0.2019387633,-0.0813750997,0.1927640736,-0.0473877303,-0.3078513145,-0.1343808621,0.4044968188,-0.1600380987,-0.2725828886,0.0046120416,0.1497466266,-0.0676551685,-0.120562166,0.1607695073,-0.0707525089,-0.2598537803,0.3602844477,-0.3102980554]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2888","title":"v1.11.1 release date","comments":"Hi ! Probably 1.12 on monday :)\r\n","body":"Hello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.\r\n\r\nWhen do you plan to publush v1.11.1 release?","comment_length":7,"text":"v1.11.1 release date\nHello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.\r\n\r\nWhen do you plan to publush v1.11.1 release?\nHi ! Probably 1.12 on monday :)\r\n","embeddings":[-0.3028725386,-0.1802771538,-0.2127547413,-0.0963228866,0.0652879179,-0.1868623346,0.0738429576,0.3128882051,-0.2114172131,0.4218455255,0.2351345271,0.0822070688,-0.0895844847,0.4837630689,-0.1932857186,-0.2654596269,0.0769366696,0.0144002829,-0.1551253647,0.0024792501,-0.1483270228,0.119360745,-0.3253111243,0.0831760392,0.0989715233,-0.064572975,-0.3526091278,-0.1743175536,-0.5363405347,-0.3933886588,0.461257726,0.1939074099,0.2547554672,0.2986641824,-0.000103304,-0.4014018774,0.5556403995,0.2845861912,-0.2422691882,-0.2193116248,-0.3339813054,-0.5906572938,-0.1154627725,0.0800851658,-0.2921097875,-0.1461395621,0.082330212,-0.0849997699,-0.13677302,-0.1171071827,0.2824233174,0.0947450623,0.1915313601,-0.3989170194,0.284394294,-0.0203845855,-0.2789727747,-0.3155975044,0.5697530508,0.1857773513,0.1898058355,0.326002121,0.2075845897,-0.1623990536,0.1387228072,0.1931113005,0.1122244373,-0.32729882,-0.0393120684,0.1010244489,0.9985591769,-0.0018671668,-0.2434326708,0.2223907411,-0.0925616995,-0.3847976029,0.1220049784,-0.0718759969,0.1220379025,0.1305280179,0.0270749014,-0.3837861121,-0.1746051461,0.2439880967,0.0338811204,0.3361899853,-0.1041846797,-0.0660860538,0.1154744327,-0.2463802546,0.0967559218,-0.0215033758,-0.1613555253,0.2227307111,0.0464089848,-0.4094306231,0.1858401746,-0.0288592987,0.1305842102,0.3231448531,-0.1524199545,-0.1115325615,-0.1302562058,-0.0906753838,0.5744841099,0.0935288742,0.377846837,0.2240085304,0.3023318052,0.1646759659,0.210433796,-0.0148354508,0.131557256,0.2080590278,-0.1345603913,0.092921719,0.2413697988,-0.5841081142,0.3566501141,0.0036136017,0.0791958421,-0.3253816366,-0.3280638158,-0.1975245178,-0.0646178052,0.1883967668,-0.1564619541,-0.1827139109,0.1210180596,-0.2046548128,-0.3577520847,0.0363553949,-0.2325457633,-0.0549016744,0.1167488173,-0.2526122332,-0.1254964024,0.0305748545,-0.04656725,0.0622468628,0.0610297844,0.1928608567,-0.2869565487,0.2402708977,-0.1486318558,0.2325121164,-0.3784801364,0.3326986134,-0.1420154572,0.2017378062,-0.0192655772,-0.2063460052,-0.235487327,0.3137744367,-0.1372713745,-0.3029890358,0.19962731,0.3279361129,-0.3900982738,-0.1182468757,-0.1860611141,0.2779344022,-0.1112236157,-0.2089126557,0.0612550527,-0.0017352636,-0.2293154597,0.0255627967,-0.3211435378,-0.2947054803,0.1805807501,-0.0630620569,0.0061964095,-0.2280418724,0.1311280131,-0.185826987,0.4691230059,-0.3665311038,-0.5793189406,0.4294295013,-0.1866938472,-0.4467723966,0.4338716865,0.2241641581,0.3097352386,-0.1575234383,-0.2617045343,0.2637113929,-0.1526574194,-0.320348084,-0.2554035187,-0.3677665293,-0.0720299929,0.1507190168,0.1602411419,0.2366000116,0.2579503655,0.2858259678,0.1898807436,0.1925684959,0.1466663629,0.0294717029,0.5004534125,-0.10350997,0.2177486867,-0.249899298,-0.2581174374,-0.0229802001,-0.0547287352,0.2486825585,0.5317671299,-0.2098047286,-0.168516919,-0.1740678996,0.2359928489,-0.0324108824,0.1631854326,0.1089351401,-0.0164421871,0.1266786158,-0.3025694489,0.0803501308,0.1472560316,0.0244728141,-0.075657174,0.1248692945,-0.2761505842,0.0051354663,0.2405391037,-0.0050578136,0.0540098287,-0.2141433656,0.1412835717,0.0534515046,-0.0641530901,0.3371422291,0.6113381386,0.1324200183,0.3750956953,0.1302689165,0.219974041,0.0903351158,-0.0883826166,0.2412555814,-0.269962877,0.312500447,0.238511622,-0.2063528448,-0.1451906115,0.118214719,0.1887120157,-0.0891980454,-0.0327405259,-0.3398402631,0.2254811525,-0.1658332348,-0.2750526965,-0.0365721621,-0.2551523447,0.3512178659,0.3392684758,-0.1608422101,-0.1638178527,0.0186253563,0.0652871504,-0.0031810594,0.2478724569,0.0328481123,-0.1552314162,0.3118087649,0.1120094359,-0.0968455672,0.0139505379,0.0497668646,0.0382369012,0.0392541103,-0.3878073692,-0.221236378,0.0975770727,0.1155316979,0.0994631052,0.1521302015,0.1214670911,0.2327489555,0.2580001354,-0.344601959,-0.3168298304,-0.2266455889,0.1146346107,-0.1705584228,0.1445552707,-0.1704138368,0.0943085477,0.0777231157,0.2056838423,0.1746924818,-0.0169805512,0.7022944689,-0.2754337788,0.4286308587,-0.22941567,-0.0400346629,-0.287625581,0.3418615162,-0.1031061858,-0.2282226235,0.195717901,0.0163822714,0.3134722412,-0.4504329264,-0.4006796479,0.0965960771,-0.190621376,-0.0871734247,0.1440135986,-0.3507573307,-0.0002354266,0.2681168616,0.020206226,-0.1884005666,-0.1797890365,-0.1004347429,-0.1091894209,0.0140760578,-0.2644398212,-0.5476667881,-0.23607862,-0.032877408,0.0195357129,0.1396642774,0.0098538799,-0.2097719312,-0.0314573459,-0.0043667466,-0.025507886,-0.3325499892,-0.13492091,-0.1378961504,0.1646966338,-0.1210952327,-0.3690144718,0.0233245771,0.2133821547,-0.1906823069,0.0343830734,-0.608182013,-0.2255241424,-0.07357122,0.0902312696,0.271065414,0.0482119732,0.3331973553,0.0043859896,-0.2407487333,-0.0966798887,-0.289496839,-0.0617358088,0.2790717185,0.3425323367,0.1480440795,0.2153837532,0.0382699557,0.4080422521,0.0405399092,-0.0831079558,0.1420852691,0.0724038258,0.1340251565,-0.0351542011,-0.1449554712,0.1592445076,0.0636743233,0.2118257731,0.0847029686,0.2609232962,-0.1647253633,-0.4551307261,-0.0843807682,0.0422558002,0.1397813708,-0.2756437659,0.2367210686,0.3786954284,0.094233036,0.0404279269,-0.1744975001,-0.2230151296,0.1618852317,0.0282293838,0.155143857,-0.037190333,-0.0875179842,0.2474852353,-0.4120295048,0.3248883486,-0.1162716597,-0.0254962426,-0.1575712413,0.1632208079,0.0907445326,-0.0197371934,-0.1022816896,-0.4347622991,0.1979690641,0.3334328234,-0.3282330632,0.209832564,0.0612220913,-0.5530356169,-0.0694036111,-0.1418507844,0.490211755,0.0862114429,-0.2527614832,0.5576377511,0.0898270458,-0.2430394441,0.089087598,-0.0483200476,-0.0593439825,0.0106714554,-0.0859256014,-0.3630439341,0.028471671,-0.0130609032,-0.2682625651,0.0257274024,-0.0579449162,0.0380475894,-0.0671024472,0.098899141,-0.0086529767,0.1888565868,0.2343615294,-0.0548734739,0.2192328572,0.2210982591,0.1266529709,-0.219884634,0.0629318357,-0.3096889853,0.2338983119,0.2519245148,-0.0197273009,0.0526784435,-0.174160555,0.0401716121,-0.1222358719,-0.1953085959,0.4756614864,-0.0125389127,-0.1958249062,-0.2044544071,0.2814204693,0.2040803432,-0.323269546,0.6049668789,0.2233847231,0.0581449308,0.0844930708,-0.0875702798,0.8768109083,0.0091802431,-0.0530386232,-0.103736043,-0.258508563,0.3260923326,0.1500751227,0.1819668561,-0.2611414194,-0.1667617857,-0.1415490061,0.1881760061,0.2261874974,0.0333929323,-0.4414962232,0.281454742,0.0162123907,0.318469286,0.0218495298,0.4039916396,0.2129945159,0.1521504074,0.0103357509,0.2614460289,0.007627523,0.2971936762,-0.027748853,-0.0637189299,0.0742425397,0.0526656769,-0.2426996827,-0.083447732,0.2240836024,0.1241263077,0.0071352581,-0.4132660627,-0.1007736176,-0.1466791779,0.2693107724,0.1400369257,-0.2019508928,0.0580885448,-0.0842173621,-0.1220868751,0.0292737298,0.2089927346,0.2616169453,-0.3624403477,-0.2537929416,0.1820687056,0.0940711349,-0.1514322162,-0.3724226952,-0.1742647737,0.0492790043,-0.117994085,0.2095771134,0.3122289181,-0.1254249513,-0.2380960584,0.2550587356,0.233158201,0.0837606713,0.0246355869,0.2637301683,0.0132632833,-0.2026557177,0.2164986283,0.2368977517,-0.2317263782,-0.1114870906,0.032164149,-0.0729064792,-0.3280212283,0.149072662,0.2185632586,-0.032606367,0.0898795649,-0.0617410839,-0.0693863109,0.1985311955,0.7674746513,0.1134763584,-0.1324970871,-0.3815554082,-0.1036905646,-0.1598554105,0.4016879797,-0.3281000853,0.0328431837,-0.3015990853,-0.1019492298,0.1833198965,-0.1321831942,-0.4311744273,0.0374722108,0.2304802239,-0.1759789288,0.3346586227,0.0276554301,-0.0175769292,-0.0240464881,0.1310301274,-0.3019618094,-0.3366426528,-0.2488733232,0.0150217954,0.0964988172,-0.100690268,0.0258981884,-0.2341493517,-0.1691185385,-0.2735385597,0.3086502552,0.06740731,0.0497181714,0.2783055902,0.1942646354,-0.0693481043,-0.1438840032,0.0461286157,0.1568335444,0.3562457561,-0.0707835779,0.1494660378,0.2881281674,0.3032661378,-0.0937201083,-0.1116363481,-0.0131287752,-0.0468256325,0.2652699351,0.1904975176,0.0371222459,0.1332924515,0.2392027676,0.299526006,0.122852698,0.0362787955,-0.1393833905,0.0654303581,0.3110370636,-0.2561894059,-0.3114360273,0.0423956625,0.2429854125,-0.2813114524,-0.0419045128,0.2741584182,0.3505083323,0.22629264,0.1385433227,0.2401577383,-0.3199465871,0.1219916716,0.2591587603,0.2279812098,-0.0197042357,0.2900158763,0.1830374002,-0.0682588518,0.093497701,0.0996332839,0.2804675102,0.5949040055,0.3702272177,0.2393652797,-0.1198549196,0.0408343188,0.0919148624,-0.0366935171,-0.0142002311,0.0637082756,0.2750998139,-0.29993397,0.0468942784,-0.3736778796,-0.2152929455,0.0454041921,-0.1816400886,-0.0064777839,-0.1451922357,-0.1916889995,-0.053099826,0.0388124809,0.3415044248,0.4628058672,-0.2635577023,0.0336105339,-0.1391046792,-0.0992713794,0.0667115003,0.2791768909,-0.0249665789,0.2348811775,-0.396173954,0.1176120788,0.1963460892,0.3659482896,0.0244652964,-0.109329991,-0.2823936045,-0.0752786174,-0.1506735981,-0.2029793561,0.1160526648,0.3965930045,0.1634795666,0.0897685811,0.1343604624,0.145347029,-0.1465705633,0.4180464149,0.0394908339,-0.1240979508,-0.451328218,0.2151078582,0.1355981827,-0.1942361742,-0.0209452938,0.0969316363,-0.1266513616,-0.3759785891,0.180424124,0.1019164398,-0.1711923629,-0.1600877494,0.1484104097,-0.0683724433,0.124356918,0.1947842687,0.1296981573,0.0286450516,-0.3705144227,-0.3561777174,0.3234818578,0.2191070914,-0.2648040056,-0.1161450446,0.246768415,0.1100329086,0.3032544553,0.1898021251,-0.1368157864,-0.244817704,-0.1841135025,-0.4536907673,0.1890252829,0.2329204381,0.2602323592,-0.12943694,0.0105283875,0.1758193523,-0.3087698519,0.061227791,0.0883698463,-0.2749614716,0.2900063097,0.1081978455,0.4732691348,-0.0953155756,0.1714043915,-0.0163458809,-0.175800845,-0.3269743919,-0.1704187542,-0.0873152837,-0.1886828095,-0.0751526654,0.0581922047,-0.237008065,0.1586024314,-0.1818601787,0.1955085397,-0.0558605976,0.3494592011,0.0367623195,0.0512245521,0.0115491403,-0.4316511452,0.1641193479,0.0917870998,-0.1045387313,-0.1162620783,-0.3184365332,-0.2898872793,0.266521275,0.1041490585,-0.119506143,-0.3795951605,-0.1625541449,0.3523990214,-0.1871525049,-0.4342991412,0.0063330201,0.1459901631,0.1065577641,-0.1277641356,0.1525983065,0.193662852,-0.0439628996,-0.0201323126,0.1976662725,0.0432184488,-0.044095356,-0.2465994656,0.1530583352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2888","title":"v1.11.1 release date","comments":"@albertvillanova i think this issue is still valid and should not be closed till `>1.11.0` is published :)","body":"Hello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.\r\n\r\nWhen do you plan to publush v1.11.1 release?","comment_length":18,"text":"v1.11.1 release date\nHello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.\r\n\r\nWhen do you plan to publush v1.11.1 release?\n@albertvillanova i think this issue is still valid and should not be closed till `>1.11.0` is published :)","embeddings":[-0.2534325421,-0.1435918212,-0.1170087606,0.1250979602,-0.0803622156,-0.2069433331,0.3182261586,0.2595030665,-0.2662116587,0.2432924211,0.4498234391,-0.0775993839,-0.0092685595,0.3775973022,-0.3972175717,-0.0890159234,0.0787209198,0.1089016348,-0.0526827313,0.0449188799,-0.2720680833,0.068657726,-0.3964475691,0.1608792841,0.0663995072,-0.0716037601,-0.1215934381,-0.254347086,-0.6671610475,-0.5313281417,0.3133550286,0.2034503967,0.0760271847,0.271275878,-0.0001110468,-0.423866421,0.5205672979,0.1486699134,-0.2614662647,-0.1513033807,-0.416443944,-0.5353802443,-0.1008592322,-0.0137669845,-0.3072520196,-0.1077144369,0.0117483158,-0.1943197548,-0.0628163069,-0.0156020885,0.2480115443,0.1673386395,0.4019794762,-0.3542853892,0.1863051802,-0.1159243658,-0.3703640699,-0.2499655932,0.5717850924,0.2568038106,0.2268878967,0.384983182,0.1912631989,-0.1922257096,0.1507626772,-0.0097723147,0.2396764457,-0.266032964,0.0415400378,0.0777787417,0.9955607653,0.0203635413,-0.3793464899,0.2034740597,-0.0927332118,-0.2712348402,0.2412474453,0.027283011,-0.0292447396,0.149242878,0.0873354375,-0.4584193528,-0.1516828388,0.1602145284,-0.0272049885,0.459104538,-0.0887170136,0.1067651212,-0.0215255283,-0.3015371263,0.2346984148,0.0200975835,0.0850308686,0.156445384,0.0514521152,-0.4402125776,0.1060789451,-0.0172465947,0.1498136967,0.1924144477,-0.2606184483,-0.249166131,-0.0464439057,0.0133378366,0.6594331861,0.0745627508,0.3405662179,0.188009277,0.4329450428,0.2551412284,0.0644607022,0.0997430384,0.0618424639,-0.0808662996,-0.1455136389,0.2052437365,0.4584421515,-0.5718432069,0.3661362231,0.1454990655,0.0115315011,-0.073997356,-0.3209313154,-0.1508467942,-0.0373871252,0.3257052004,0.0893606171,-0.1983124912,0.1348611414,-0.1805576384,-0.1951919943,-0.1088341549,-0.200981006,-0.0183628146,-0.031561859,-0.3116750121,-0.2775349915,0.1134495363,-0.1254057586,-0.0434127748,0.0564601496,0.1302085966,-0.2246665657,0.4118489623,-0.2466320097,0.082844384,-0.3096274436,0.2402789593,-0.0819456652,0.3500746489,-0.0424564332,-0.1780534983,-0.3391320705,0.2919041216,-0.1666521877,-0.0776967406,0.2701286674,0.1953231692,-0.2030190676,-0.2382987589,-0.0494303294,0.1722414643,-0.2394903451,-0.256931901,-0.0695379525,0.0432331599,-0.1193417981,0.1540283114,-0.2632963657,-0.2000599653,0.1197276041,0.0051401095,-0.1842305511,-0.3186813593,0.1262588203,-0.2455536276,0.3016958833,-0.3433002532,-0.639559865,0.3815340102,-0.284142822,-0.3894364536,0.4114672244,0.1741337925,0.2954373658,-0.3212958872,-0.1409738809,-0.0816188827,-0.1616334766,-0.3559683263,-0.4418994188,-0.4019602537,0.0429010168,0.0181122757,0.2204887122,0.1503366083,0.0952749476,0.4235877395,0.0749824941,0.2797828615,0.2078803927,0.0765660107,0.482803762,-0.1286068708,0.0529898442,-0.1006552279,-0.2767008245,0.0938937142,-0.1435723305,0.1591186672,0.4384330511,-0.3189628124,-0.2085864544,-0.0561050922,0.3260715008,-0.0709523261,0.0817816257,0.0915738642,-0.2478206754,0.1673493236,-0.4060981572,0.1368636489,-0.0098665319,-0.07066378,-0.088441737,0.0452465005,-0.2908729911,0.0994255021,0.1788440347,0.0838302895,0.1694515795,-0.2390569299,0.1745637655,-0.0815123394,-0.0723995641,0.3089856207,0.3561557233,0.1862898916,0.2350969911,0.2040317804,0.0746104494,0.1710554659,-0.0586112514,0.2303554267,-0.2253153324,0.3460877836,0.3549923301,-0.2632007301,-0.0564746819,0.144726038,0.1307961047,-0.2599101663,-0.0981654152,-0.2556478679,0.0817662179,-0.221117571,-0.2914198935,0.2252827138,-0.2763938606,0.2965317667,0.5968988538,0.0184195228,-0.047829736,0.0809677541,0.2573701739,-0.0809906572,0.1679659784,-0.0462063886,0.0884929076,0.2539899945,-0.021018533,0.0049677342,0.1221921593,0.0425285064,0.1027600989,0.0455321558,-0.4042115211,-0.0299377795,0.0530678742,0.0633863062,0.115278326,0.1767398119,0.1918154806,0.2266006619,0.0650810748,-0.3978857994,-0.2835028172,-0.0675668642,0.0291238204,-0.0194902401,0.0619625486,0.0534454733,0.1382157803,0.1468727291,0.0316265896,0.1826277375,-0.1548491269,0.7969358563,-0.3889087141,0.4114349484,-0.156639576,-0.1199053377,-0.3110970855,0.2345053405,-0.0124809062,-0.2776047289,0.2567169964,0.0537328571,0.2090214789,-0.6258595586,-0.4577414989,0.0528508835,-0.144126296,0.0452978648,0.0343037061,-0.2054453045,-0.107023187,0.2207032591,0.0384894237,-0.0587673448,-0.2087223232,-0.1666585952,-0.0068196445,0.0721502379,-0.2570260763,-0.6848427057,-0.1318627,-0.3363646269,-0.0185509361,-0.1078772321,0.1569403857,-0.1232450008,-0.051425416,-0.1049095914,0.0217202213,-0.3342834115,-0.0948758051,-0.048731003,0.0945054814,-0.0082828654,-0.4569739103,0.1993883997,0.1774878055,-0.2025116682,0.1531127393,-0.5314667225,-0.2725237608,0.206520468,-0.043862544,0.1573558748,-0.1776329577,0.5536837578,-0.1086382791,-0.2538419664,-0.1345501691,-0.2866043448,0.0720614642,0.1712462157,0.3246344626,0.409180969,0.2624562085,0.0242744964,0.3892090321,0.1903090924,0.0267370865,0.2341599315,0.1795129925,0.1924058646,-0.0937660858,-0.2098539323,0.0676550344,0.1260128468,0.1618110985,0.1551786363,0.212505579,-0.0854245946,-0.3764601052,-0.0002547773,-0.0050693392,0.1066496,-0.2045838088,0.19392021,0.6248509884,0.1559808254,-0.0690406263,-0.1899852306,-0.050280299,-0.1242645904,0.2394972146,0.218652606,-0.0803116336,0.0049764337,0.3057085574,-0.2312407643,0.3685397804,-0.0162633043,0.3051949441,-0.1745799035,0.183699578,0.167574808,-0.0165351611,-0.002247988,-0.4258732796,0.2501198947,0.5538979769,-0.2576999366,0.1758041531,0.0319327712,-0.1597495675,-0.1175884306,-0.1852885038,0.5539381504,-0.0616523623,-0.3503075838,0.5332707167,0.0958007053,-0.1159934327,0.0116964756,-0.148977384,-0.0391741842,0.0361851342,-0.0692064166,-0.1455716044,0.0450960174,0.0707545504,-0.2735245228,-0.2130563408,-0.2302908301,-0.0789414719,-0.1162233725,0.2302821875,-0.0054950924,0.2637892067,0.1018653214,-0.006938694,0.2944147289,0.310501188,0.110301502,-0.2645044625,0.1126402467,-0.0766577423,0.1321163625,0.3621128798,-0.1440612823,0.1353558898,-0.0324910507,0.0674986169,-0.0920338705,-0.2673429549,0.4327611923,-0.1655806452,-0.2975933552,-0.0365021341,0.2060316652,0.1686381698,-0.3554573059,0.5559965968,0.2967342436,0.0080543449,-0.0236563794,-0.2770451903,1.0293949842,-0.0287329629,-0.0709773898,-0.2229948789,-0.2106604427,0.3116071224,0.1518327743,0.1275588423,-0.1479117125,-0.1200410053,-0.156358555,0.025576435,0.3566545844,0.0137892039,-0.4522585273,0.1800192744,-0.1066877097,0.3458714783,0.1815039814,0.4181370139,0.2039348483,0.2149993479,-0.1357867271,0.1812182963,0.0415112488,0.3444061577,-0.1115806252,-0.1123415828,0.0506095029,-0.1684783846,-0.3772748411,-0.162755996,0.0946233049,0.1284955442,-0.0418820232,-0.3295052052,0.1501638293,-0.2067037821,0.2394614667,0.0587115698,-0.1543533653,0.1066337228,-0.1002984643,-0.2956922352,0.0422360674,0.2025135756,0.2493470311,-0.365799576,-0.1296645254,0.207518965,-0.0665956438,-0.1957606375,-0.3805607557,0.0144149913,0.1905001253,-0.0858497769,0.0902350396,0.0793108046,0.0000393472,-0.2687571943,0.2166418582,0.2363907099,0.173072502,-0.070472464,0.0909126326,-0.0943970904,-0.0462283231,0.331574887,0.4488438964,-0.0869515613,-0.0577264875,-0.104647994,-0.0253153257,-0.2967375219,0.1651073247,0.1483710557,-0.1926550269,0.3166780174,0.0451815948,-0.177145049,0.0966312066,0.8037606478,0.1132905334,-0.0867026523,-0.3089384437,-0.2269143313,-0.201273486,0.4007463753,-0.3507455587,-0.0632593036,-0.1728218198,0.003514756,0.1783655137,-0.2588581443,-0.3582352698,0.1094941497,-0.0229328293,-0.1558413357,0.3805615604,-0.1397653073,-0.1268882602,-0.1695237756,0.0683612898,-0.1914259791,-0.4219688773,-0.1825899184,-0.0068806806,0.1795591116,-0.099498719,0.1105903983,-0.2188424021,-0.3331711888,-0.3294802904,0.5655476451,-0.0574626066,0.2105582207,0.265817821,-0.0013697128,-0.1911033839,0.083206743,-0.0473593064,0.1444205344,0.2916863263,0.0616737306,0.0525817275,0.2192509323,0.3352403939,-0.051245898,-0.3188316226,-0.0857311264,0.028379539,0.1082434431,0.2533852756,0.0845074058,0.0778547525,0.3177062273,0.261341095,0.1477471739,0.0235879123,-0.10451556,0.1499727219,0.2290463895,-0.4348066449,-0.2144445181,-0.1024437621,0.1954161376,-0.3519690037,-0.0586642176,0.4161963165,0.3745106161,0.3306231499,0.2353342474,0.3252157569,-0.2875549793,0.1753366739,0.298651427,0.2821616828,0.0819243565,0.2702084482,0.3563085496,0.1198417321,-0.1174164787,0.1162222177,0.2579935193,0.6770775318,0.4611133635,0.2945824265,-0.1674147099,0.1528614014,0.145076111,0.0248568561,-0.0097037228,-0.028063925,0.2341251522,-0.2223802507,0.000882468,-0.252620697,-0.1989824623,-0.1395764947,-0.2066176832,0.0804793462,-0.2013021708,-0.1639318615,-0.0490338057,-0.0818435028,0.3881368041,0.4651462138,-0.172336027,0.0645914972,-0.1019973978,-0.126423344,0.1757827997,0.2655941546,-0.1845968068,0.3018839657,-0.4171665907,0.1055940539,0.2046003491,0.5415052176,0.0668870881,-0.0055139861,-0.0966519117,-0.0346165001,-0.2049902081,-0.0964519158,0.119966276,0.4839955568,0.2810302973,0.3264023364,0.0233848933,0.0755379423,-0.1421592236,0.4168718159,-0.1079073548,-0.1283363253,-0.6276838779,0.2174899578,-0.1051726788,-0.1579537541,-0.0612815022,0.1260119081,-0.0126639176,-0.3415772617,0.2715965509,0.2166879475,-0.0643055439,-0.248646915,0.1112187207,-0.2308488041,0.0420173332,0.0135724219,-0.0086574517,-0.0197401382,-0.3096870184,-0.4489578605,0.3907983005,0.1115981862,-0.2408151627,0.1052808017,0.2133009285,0.0889786929,0.3081979156,0.0924179628,-0.1790391803,-0.3171945512,-0.1446876526,-0.402273953,0.1941866428,0.3912087083,0.0847943872,-0.0422423035,-0.2774840295,0.2742191255,-0.2932401896,0.0322426744,0.1044757292,-0.3337162137,0.386005193,0.1406664401,0.5287458301,0.032131739,0.1304536164,0.2330349982,-0.1344192475,-0.4677241743,-0.1328878999,-0.0336711966,-0.1085676104,0.0418233424,0.0438944586,-0.2677304447,0.1162145585,-0.1990454048,0.0510428324,-0.3705941141,0.4114429653,0.0662547424,0.109508045,0.0662498847,-0.2569505572,0.2648650408,0.0552677512,-0.1163000241,-0.2598813772,-0.4447279274,-0.319673717,0.2948858142,0.0099982917,-0.1469426304,-0.1186853945,0.075142175,0.2601584792,-0.263087213,-0.4047777951,-0.1358789206,0.3026438653,0.2290502042,-0.1205541492,0.0037204858,0.0116371764,-0.1853573173,0.0748225152,0.3017756641,-0.147349447,-0.1105487272,-0.1910789162,0.0743446872]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2885","title":"Adding an Elastic Search index to a Dataset","comments":"Hi, is this bug deterministic in your poetry env ? I mean, does it always stop at 90% or is it random ?\r\n\r\nAlso, can you try using another version of Elasticsearch ? Maybe there's an issue with the one of you poetry env","body":"## Describe the bug\r\nWhen trying to index documents from the squad dataset, the connection to ElasticSearch seems to break:\r\n\r\nReusing dataset squad (\/Users\/andreasmotz\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/d6ec3ceb99ca480ce37cdd35555d6cb2511d223b9150cce08a837ef62ffea453)\r\n 90%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 9501\/10570 [00:01<00:00, 6335.61docs\/s]\r\n\r\nNo error is thrown, but the indexing breaks ~90%.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nfrom datasets import load_dataset\r\nfrom elasticsearch import Elasticsearch\r\nes = Elasticsearch()\r\nsquad = load_dataset('squad', split='validation')\r\nindex_name = \"corpus\"\r\nes_config = {\r\n \"settings\": {\r\n \"number_of_shards\": 1,\r\n \"analysis\": {\"analyzer\": {\"stop_standard\": {\"type\": \"standard\", \" stopwords\": \"_english_\"}}},\r\n },\r\n \"mappings\": {\r\n \"properties\": {\r\n \"idx\" : {\"type\" : \"keyword\"},\r\n \"title\" : {\"type\" : \"keyword\"},\r\n \"text\": {\r\n \"type\": \"text\",\r\n \"analyzer\": \"standard\",\r\n \"similarity\": \"BM25\"\r\n },\r\n }\r\n },\r\n}\r\nclass IndexBuilder:\r\n \"\"\"\r\n Elastic search indexing of a corpus\r\n \"\"\"\r\n def __init__(\r\n self,\r\n *args,\r\n #corpus : None,\r\n dataset : squad,\r\n index_name = str,\r\n query = str,\r\n config = dict,\r\n **kwargs,\r\n ):\r\n #instantiate HuggingFace dataset\r\n self.dataset = dataset\r\n #instantiate ElasticSearch config\r\n self.config = config\r\n self.es = Elasticsearch()\r\n self.index_name = index_name\r\n self.query = query\r\n def elastic_index(self):\r\n print(self.es.info)\r\n self.es.indices.delete(index=self.index_name, ignore=[400, 404])\r\n search_index = self.dataset.add_elasticsearch_index(column='context', host='localhost', port='9200', es_index_name=self.index_name, es_index_config=self.config)\r\n return search_index\r\n def exact_match_method(self, index):\r\n scores, retrieved_examples = index.get_nearest_examples('context', query=self.query, k=1)\r\n return scores, retrieved_examples\r\nif __name__ == \"__main__\":\r\n print(type(squad))\r\n Index = IndexBuilder(dataset=squad, index_name='corpus_index', query='Where was Chopin born?', config=es_config)\r\n search_index = Index.elastic_index()\r\n scores, examples = Index.exact_match_method(search_index)\r\n print(scores, examples)\r\n for name in squad.column_names:\r\n print(type(squad[name]))\r\n```\r\n\r\n## Environment info\r\nWe run the code in Poetry. This might be the issue, since the script runs successfully in our local environment.\r\n\r\nPoetry:\r\n- Python version: 3.8\r\n- PyArrow: 4.0.1\r\n- Elasticsearch: 7.13.4\r\n- datasets: 1.10.2\r\n\r\nLocal:\r\n- Python version: 3.8\r\n- PyArrow: 3.0.0\r\n- Elasticsearch: 7.7.1\r\n- datasets: 1.7.0\r\n","comment_length":44,"text":"Adding an Elastic Search index to a Dataset\n## Describe the bug\r\nWhen trying to index documents from the squad dataset, the connection to ElasticSearch seems to break:\r\n\r\nReusing dataset squad (\/Users\/andreasmotz\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/d6ec3ceb99ca480ce37cdd35555d6cb2511d223b9150cce08a837ef62ffea453)\r\n 90%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 9501\/10570 [00:01<00:00, 6335.61docs\/s]\r\n\r\nNo error is thrown, but the indexing breaks ~90%.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nfrom datasets import load_dataset\r\nfrom elasticsearch import Elasticsearch\r\nes = Elasticsearch()\r\nsquad = load_dataset('squad', split='validation')\r\nindex_name = \"corpus\"\r\nes_config = {\r\n \"settings\": {\r\n \"number_of_shards\": 1,\r\n \"analysis\": {\"analyzer\": {\"stop_standard\": {\"type\": \"standard\", \" stopwords\": \"_english_\"}}},\r\n },\r\n \"mappings\": {\r\n \"properties\": {\r\n \"idx\" : {\"type\" : \"keyword\"},\r\n \"title\" : {\"type\" : \"keyword\"},\r\n \"text\": {\r\n \"type\": \"text\",\r\n \"analyzer\": \"standard\",\r\n \"similarity\": \"BM25\"\r\n },\r\n }\r\n },\r\n}\r\nclass IndexBuilder:\r\n \"\"\"\r\n Elastic search indexing of a corpus\r\n \"\"\"\r\n def __init__(\r\n self,\r\n *args,\r\n #corpus : None,\r\n dataset : squad,\r\n index_name = str,\r\n query = str,\r\n config = dict,\r\n **kwargs,\r\n ):\r\n #instantiate HuggingFace dataset\r\n self.dataset = dataset\r\n #instantiate ElasticSearch config\r\n self.config = config\r\n self.es = Elasticsearch()\r\n self.index_name = index_name\r\n self.query = query\r\n def elastic_index(self):\r\n print(self.es.info)\r\n self.es.indices.delete(index=self.index_name, ignore=[400, 404])\r\n search_index = self.dataset.add_elasticsearch_index(column='context', host='localhost', port='9200', es_index_name=self.index_name, es_index_config=self.config)\r\n return search_index\r\n def exact_match_method(self, index):\r\n scores, retrieved_examples = index.get_nearest_examples('context', query=self.query, k=1)\r\n return scores, retrieved_examples\r\nif __name__ == \"__main__\":\r\n print(type(squad))\r\n Index = IndexBuilder(dataset=squad, index_name='corpus_index', query='Where was Chopin born?', config=es_config)\r\n search_index = Index.elastic_index()\r\n scores, examples = Index.exact_match_method(search_index)\r\n print(scores, examples)\r\n for name in squad.column_names:\r\n print(type(squad[name]))\r\n```\r\n\r\n## Environment info\r\nWe run the code in Poetry. This might be the issue, since the script runs successfully in our local environment.\r\n\r\nPoetry:\r\n- Python version: 3.8\r\n- PyArrow: 4.0.1\r\n- Elasticsearch: 7.13.4\r\n- datasets: 1.10.2\r\n\r\nLocal:\r\n- Python version: 3.8\r\n- PyArrow: 3.0.0\r\n- Elasticsearch: 7.7.1\r\n- datasets: 1.7.0\r\n\nHi, is this bug deterministic in your poetry env ? I mean, does it always stop at 90% or is it random ?\r\n\r\nAlso, can you try using another version of Elasticsearch ? Maybe there's an issue with the one of you poetry env","embeddings":[-0.1423357278,0.172727868,-0.0044709793,0.0400959924,0.0997370332,0.0273672696,-0.013204325,0.0684226677,0.0138257872,0.0531148016,-0.0395590216,0.3000072539,0.3499440253,-0.3781624734,-0.1049256846,0.1306153685,0.0909727961,0.0941495076,0.2756751478,0.1979281008,-0.0359147601,0.095908314,-0.1247543246,-0.0892915428,-0.3346315026,0.0212135501,-0.0316708237,-0.1235372126,-0.0287176576,-0.4167428911,0.3019476235,-0.1501852423,-0.1225858629,0.3576337695,-0.0001185079,0.0780737773,0.4348543286,-0.0329714008,-0.0183227472,-0.2907201052,-0.0723804981,-0.1354442686,0.279198885,-0.1591316015,-0.0953472778,0.1617633402,0.0987489,-0.6114605665,0.4415228069,0.3961337507,0.221011281,0.3180061877,0.3437509239,-0.1243836433,0.4198286533,-0.1281127334,-0.0489515252,-0.2564239502,0.2530343533,-0.0389002413,-0.0290674474,0.1452289224,-0.0944889709,-0.2707443833,0.108504355,0.1481345296,-0.4625434577,-0.2957313359,0.3149895072,0.1134225652,0.1910260916,-0.0712343082,-0.2519367039,0.0451840535,-0.0008995398,0.1622108668,0.1994201243,0.1894568503,0.0606099963,0.1190423146,0.0653904825,-0.1248976737,-0.2391608059,0.4135331213,-0.0537097417,0.1734440774,-0.097541891,0.1064755395,0.1365941018,-0.146080032,0.0782291442,0.2522954047,-0.2729330957,-0.0555161387,-0.3764431775,0.1024475098,-0.1168943718,-0.2002719194,0.4185051024,-0.0634525344,-0.0926844999,0.1597323418,0.6514115334,0.082370162,0.2860307097,-0.0631037876,0.0613816716,0.3085791171,0.0735873207,-0.3172648847,-0.3174724579,-0.2593558729,0.285479039,-0.2149694413,0.4180778563,-0.4818145633,0.2134571224,-0.0300275907,-0.159065932,0.2193003744,-0.5879529715,-0.0061579379,-0.0753503293,0.1353493631,0.1364613473,-0.4953055084,0.0383693911,0.2725166976,-0.2289299369,0.1452205926,-0.2356664091,-0.2659835815,0.0408693254,0.0179617349,0.1160360128,-0.6583971977,0.2295712978,0.0021879284,0.0842600539,-0.3945707381,0.2630119622,-0.0829003006,0.1773642153,0.218855992,0.0615348555,0.023488922,0.0232272167,-0.4385558963,-0.0351555534,-0.1221938878,-0.3814532757,-0.2521406412,-0.2764842808,0.2102274299,-0.1385770142,0.0171780493,0.1357520074,0.2181541324,0.1103124693,-0.0082744509,0.0775755048,-0.0403560661,-0.2897684872,0.0814163461,0.2997894585,0.2074805498,0.1160847694,-0.1290065497,0.034548115,0.4276873767,-0.2609300017,0.2141182125,-0.2313275784,0.0364292637,-0.1731410325,0.4179225564,0.3350299299,-0.4630022347,-0.613461256,0.1387205869,-0.2306159735,-0.0950086787,-0.272356391,0.0607838072,0.6483929753,0.0107715735,0.1798678637,0.2476815581,-0.1762580425,-0.1678980738,-0.1523010135,-0.0007278286,-0.1430873871,-0.177202031,0.0628991276,0.0903355628,0.2576126158,0.2004109323,-0.0055740727,0.1895459443,0.0115436343,0.3066953421,0.3564867079,0.1077488884,0.2526801229,-0.1288016289,-0.2565320432,0.2232705653,0.0895305648,0.0836865902,-0.1827768236,-0.1198632121,-0.1379051954,0.0071618739,-0.1569918245,0.0309961122,0.107497111,-0.2377432734,-0.2198884338,0.1922539026,0.0002298418,-0.1667473465,0.0728107169,0.2382162064,-0.2376390845,0.3120537996,-0.0830118656,-0.0867530704,-0.1883609593,-0.0127894552,0.2368167192,-0.2060383856,-0.1446202993,0.332541883,-0.0994061977,0.3171947896,0.1542841047,0.025194684,0.1261184365,-0.6889179945,-0.1038779616,0.1420570463,0.0571958348,-0.2761912048,-0.0003128937,0.0152491853,0.0072259768,0.3724019229,-0.2171028852,0.0868740231,-0.0073972992,-0.1734815091,-0.4597905576,-0.0337329842,-0.0198460873,-0.3641361892,-0.0424907617,-0.1129411384,-0.118290931,0.0319062024,0.0161155853,0.1542524546,0.104662396,0.3709858358,0.0202301163,-0.10019207,-0.3800418377,-0.0208591763,0.2398631126,0.1784040332,0.1413077563,-0.0003945187,0.016000025,-0.0571260825,0.3174203336,0.3449330926,0.0851912573,0.3284900784,0.4267836213,0.080735378,-0.2349987477,-0.0556238145,-0.2187708616,-0.0349604785,-0.2483057678,0.4837775528,-0.0381795317,0.1216896772,-0.142181322,0.036698062,-0.2608760893,-0.440820545,-0.1000710577,0.1918876469,-0.0298058689,0.010196479,-0.389636904,-0.0280155968,0.2536537051,-0.2719688416,0.1400227845,-0.13825804,-0.1423628628,-0.0342508368,0.1852953732,0.0559359342,0.239090845,0.2277609557,-0.1308041066,-0.3857408166,-0.4401707351,0.1353846639,0.0270662643,0.3381630778,0.1061841398,-0.2070374787,-0.4189092517,-0.139235571,0.2391747087,-0.1569141895,0.0960531235,0.0989907682,0.0812273249,0.165456146,0.252183646,-0.2988062501,-0.1856297553,-0.1773062944,0.0319295116,-0.1300621033,-0.0701901093,0.3265482187,-0.0100668529,0.1334317029,-0.117993705,-0.0943661481,-0.1205612123,-0.1552785039,0.0512902848,-0.1617365777,-0.3146482408,0.1664025784,-0.0358661972,-0.2476519793,0.2399175614,-0.3951127529,0.2129151374,0.1262060553,-0.2579472363,-0.3092864454,0.0559353307,0.0856449679,-0.2632970214,-0.0152955577,-0.0331743136,0.0292060673,0.1754637212,0.2231558114,0.0925877765,-0.0978393182,0.2361078113,-0.0709363371,0.6817715764,0.1603981704,-0.0134615982,0.1587851346,-0.1708619744,0.3219510317,-0.2369931191,-0.5023804903,-0.1412394941,-0.1059440747,-0.1054208204,-0.0283705853,0.109574452,-0.2827525437,-0.115599893,-0.2426892221,-0.0605009161,-0.54604882,0.1893932819,0.1710993499,0.0693428218,0.0568653457,0.1564288139,0.1233408228,-0.4360743463,-0.4252508581,-0.1272193491,-0.1278421879,0.0157799479,-0.2675440609,0.1987197399,-0.6068624258,0.1646698713,-0.0542106107,0.5287730098,0.0551700518,0.0009338917,0.1420914084,0.1686774641,0.7477430105,-0.2251750976,-0.1185975447,-0.0158345401,0.0547930151,-0.5928254128,0.0987460539,-0.1476437747,0.6286976933,0.2092178911,0.3737720251,0.0288933534,0.1745180041,0.0755364895,-0.1005474627,0.079070352,-0.077841498,-0.2256833166,-0.2200263143,0.3044427037,0.2742739916,0.3445194364,0.0810018852,0.3223793507,-0.2284775674,-0.2969676554,-0.0309453886,0.0437185802,-0.2290928662,0.2575103939,-0.0781481341,0.2960500419,0.4340247512,0.2177850455,0.2762292624,0.408185482,-0.2440405637,0.0885933414,-0.0196736883,-0.0375479199,0.1764096618,0.3452102542,0.2333936393,0.1530003399,0.1788771152,0.0665417537,-0.2147426754,-0.1878597289,0.5031189919,0.1382859349,-0.4760869145,0.0670427978,0.3620651662,-0.0627520233,-0.0308657922,0.0745157823,0.5442926288,-0.3988694847,0.6471534371,0.1135982275,0.6712856889,0.089194715,-0.0725566596,-0.2630755305,0.1109268218,0.7352765799,-0.3347020149,0.25493747,-0.2097499371,-0.5855283737,-0.0690590367,-0.2858499587,0.0086796787,-0.2234174311,-0.3603481948,0.0166503042,-0.3822994828,0.2821260095,-0.0841552094,0.3374670744,-0.1510460228,-0.1274801344,0.0352684073,0.1547400504,0.2227140963,0.2692329884,0.1916231364,0.0424441211,-0.0174229462,-0.2775602937,-0.494461298,-0.0191288274,-0.1274602115,0.1551729739,0.1827656329,-0.0453223921,0.3815469444,0.0988958329,0.0481477119,0.107557185,-0.1812240034,0.2406510413,-0.0080882208,-0.0569829531,0.4446687996,-0.220583424,0.0711796582,0.1791232228,-0.4024159312,-0.0184437446,0.2618075609,-0.1858665645,-0.0822697282,0.3471039236,-0.2522739768,-0.3732704222,-0.2347560972,-0.1285910159,-0.067549251,-0.1432647109,0.0892322212,0.0084903808,-0.1242847145,0.2830715179,0.3689793646,-0.2495323718,0.0721010566,0.8503792286,-0.2160516232,-0.1386739761,0.2074775249,0.060395591,-0.1718024909,-0.1100918576,0.3264315724,0.285936892,-0.1506772935,-0.2592438161,-0.4627506733,0.2283157706,0.0369627327,0.037715055,0.2122460753,-0.0092372168,-0.184262231,-0.1904779226,0.1433349401,0.1566762626,0.4442587197,0.1021856293,-0.3198876679,-0.405808568,-0.0958278179,0.1375002265,-0.243810907,0.3268664777,-0.2520278394,0.0064505953,-0.129592225,-0.0516157895,0.5856395364,-0.3752815127,0.0591960773,0.1744844466,-0.0512193516,-0.1685403883,-0.0637019202,0.1191376373,-0.0450762287,-0.1070081294,-0.1666346788,-0.2316609472,-0.0473485813,-0.2600593269,0.0268881936,0.4245568514,0.129269436,-0.1663024724,-0.1522525251,-0.1260861605,0.3549259007,-0.0860512331,0.2210680395,0.2574340999,0.0169345103,0.2850543857,0.1095146686,-0.0313541964,-0.1166432649,0.2705664635,-0.0437093563,0.0783668011,0.2038312554,-0.2371039093,-0.0961813927,0.5086316466,0.4194711745,0.394295156,-0.3104347289,-0.1650985628,-0.0005253511,0.2413312495,0.0492484048,-0.2417104393,0.6126328111,0.090026781,0.1283886582,0.2012028694,0.0890386701,-0.3130941987,-0.3525202274,0.0291854721,0.7695252299,0.1349657774,0.106252335,-0.1847231537,-0.0298995078,-0.0307875723,0.0969654322,0.1540451199,0.3628753126,0.2108799964,0.0093515022,0.4457625151,0.0946389288,0.2002345026,0.195820272,0.1463286877,0.2819838822,0.2778770626,0.1983456612,0.0648833215,-0.0310216248,0.815058291,-0.0081768036,-0.3089000583,-0.1074742749,0.0123540768,-0.0718853697,-0.1381998509,0.2133601159,-0.1653792709,-0.0008473348,-0.2416391075,0.0849005878,-0.0269591268,-0.3901277184,0.1575505435,0.1124823689,-0.043095123,0.0508896858,-0.141497016,0.0767197385,-0.1664367169,0.0036225964,-0.1519070715,0.1252125502,0.2351682633,0.3312087953,0.4035592377,-0.074723579,-0.1400466859,0.4058650434,-0.0350114331,-0.1445527524,0.1079212949,0.0318367966,-0.0002727603,-0.5440949202,0.147816807,0.0487125851,-0.0635885522,0.0140420524,0.2640837133,0.4321984351,-0.1540914923,-0.3470183909,-0.2106317282,0.0744327679,-0.1783645749,-0.1506300569,-0.1600381285,0.0988606885,0.2971235514,-0.2123093009,0.1440627426,0.1429017782,0.0628881827,-0.086921744,0.1966791004,0.1160484254,-0.0550464578,-0.033396706,0.074752003,-0.2059546411,0.3186418712,-0.4508095384,0.169267714,0.5591514707,0.2960522175,-0.2047137916,0.1067671329,0.1807710826,-0.2350661755,-0.2993133366,-0.0772450045,-0.0160411503,-0.2238667756,-0.1015606448,0.0261780359,-0.0101414658,-0.0251561124,0.1090446934,-0.5017803311,0.0234621968,-0.2353968024,-0.1540967077,-0.1983961016,-0.1749744862,0.1152511016,0.0764285624,0.2546803653,-0.0108258994,0.0705437288,-0.1942732483,-0.4698742628,-0.2817729712,0.2653594911,0.0957904756,0.0309691727,-0.2558007538,-0.4981395304,-0.2089369297,0.1482592225,0.0951690748,-0.1051607206,-0.4935901165,0.0211842135,-0.0664745718,0.1254032254,-0.2824284732,0.3217877746,-0.0584607683,0.4555459917,0.0576231778,-0.4273312688,0.472407788,-0.3319600224,0.005578591,-0.0409944504,0.2926785946,0.6038848758,0.1409938782,-0.2386948913,-0.1673590839,0.2756532133,-0.1943458468,-0.2440593839,-0.098173894,-0.0150665436,0.1926191449,-0.0259733181,-0.1619805992,0.2742896378,-0.0292484723,0.0151276793,-0.0911699384]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2882","title":"`load_dataset('docred')` results in a `NonMatchingChecksumError` ","comments":"Hi @tmpr, thanks for reporting.\r\n\r\nTwo weeks ago (23th Aug), the host of the source `docred` dataset updated one of the files (`dev.json`): you can see it [here](https:\/\/drive.google.com\/drive\/folders\/1c5-0YwnoJx8NS6CV2f-NoTHR__BdkNqw).\r\n\r\nTherefore, the checksum needs to be updated.\r\n\r\nNormally, in the meantime, you could avoid the error by passing `ignore_verifications=True` to `load_dataset`. However, as the old link points to a non-existing file, the link must be updated too.\r\n\r\nI'm fixing all this.\r\n\r\n","body":"## Describe the bug\r\nI get consistent `NonMatchingChecksumError: Checksums didn't match for dataset source files` errors when trying to execute `datasets.load_dataset('docred')`.\r\n\r\n## Steps to reproduce the bug\r\nIt is quasi only this code:\r\n```python\r\nimport datasets\r\ndata = datasets.load_dataset('docred')\r\n```\r\n\r\n## Expected results\r\nThe DocRED dataset should be loaded without any problems.\r\n\r\n## Actual results\r\n```\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in \r\n----> 1 d = datasets.load_dataset('docred')\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 845 \r\n 846 # Download and prepare data\r\n--> 847 builder_instance.download_and_prepare(\r\n 848 download_config=download_config,\r\n 849 download_mode=download_mode,\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 613 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 614 if not downloaded_from_gcs:\r\n--> 615 self._download_and_prepare(\r\n 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 617 )\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 673 # Checksums verification\r\n 674 if verify_infos:\r\n--> 675 verify_checksums(\r\n 676 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 677 )\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 38 if len(bad_urls) > 0:\r\n 39 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 40 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 41 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 42 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/uc?export=download&id=1fDmfUUo5G7gfaoqWWvK81u08m71TK2g7']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.11.0-7633-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n\r\nThis error also happened on my Windows-partition, after freshly installing python 3.9 and `datasets`.\r\n\r\n## Remarks\r\n\r\n- I have already called `rm -rf \/home\/\/.cache\/huggingface`, i.e., I have tried clearing the cache.\r\n- The problem does not exist for other datasets, i.e., it seems to be DocRED-specific.","comment_length":69,"text":"`load_dataset('docred')` results in a `NonMatchingChecksumError` \n## Describe the bug\r\nI get consistent `NonMatchingChecksumError: Checksums didn't match for dataset source files` errors when trying to execute `datasets.load_dataset('docred')`.\r\n\r\n## Steps to reproduce the bug\r\nIt is quasi only this code:\r\n```python\r\nimport datasets\r\ndata = datasets.load_dataset('docred')\r\n```\r\n\r\n## Expected results\r\nThe DocRED dataset should be loaded without any problems.\r\n\r\n## Actual results\r\n```\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in \r\n----> 1 d = datasets.load_dataset('docred')\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 845 \r\n 846 # Download and prepare data\r\n--> 847 builder_instance.download_and_prepare(\r\n 848 download_config=download_config,\r\n 849 download_mode=download_mode,\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 613 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 614 if not downloaded_from_gcs:\r\n--> 615 self._download_and_prepare(\r\n 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 617 )\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 673 # Checksums verification\r\n 674 if verify_infos:\r\n--> 675 verify_checksums(\r\n 676 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 677 )\r\n\r\n~\/anaconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 38 if len(bad_urls) > 0:\r\n 39 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 40 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 41 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 42 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/uc?export=download&id=1fDmfUUo5G7gfaoqWWvK81u08m71TK2g7']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.11.0-7633-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n\r\nThis error also happened on my Windows-partition, after freshly installing python 3.9 and `datasets`.\r\n\r\n## Remarks\r\n\r\n- I have already called `rm -rf \/home\/\/.cache\/huggingface`, i.e., I have tried clearing the cache.\r\n- The problem does not exist for other datasets, i.e., it seems to be DocRED-specific.\nHi @tmpr, thanks for reporting.\r\n\r\nTwo weeks ago (23th Aug), the host of the source `docred` dataset updated one of the files (`dev.json`): you can see it [here](https:\/\/drive.google.com\/drive\/folders\/1c5-0YwnoJx8NS6CV2f-NoTHR__BdkNqw).\r\n\r\nTherefore, the checksum needs to be updated.\r\n\r\nNormally, in the meantime, you could avoid the error by passing `ignore_verifications=True` to `load_dataset`. However, as the old link points to a non-existing file, the link must be updated too.\r\n\r\nI'm fixing all this.\r\n\r\n","embeddings":[-0.2659113109,0.3422692716,0.0355651341,0.3098572195,0.214197889,0.0404327884,0.3646671176,0.4201550186,0.3040324748,0.0715031847,-0.2768665254,0.190118432,0.1306955665,-0.1951108724,-0.1612918228,0.3339123726,0.1246936545,0.1053829789,-0.2242237628,-0.1252798289,-0.3106808662,0.1985796541,-0.2184446454,-0.2843059599,-0.0529530197,0.1835617423,0.1532600224,0.2324340791,-0.120700039,-0.3587095439,0.2788212895,0.2785292566,0.173633635,0.4296127558,-0.0001249282,0.1207987443,0.2817707658,-0.0242785923,-0.501280129,-0.2013867348,-0.563354671,-0.3742280006,-0.0070420792,-0.1583375037,-0.0344875716,0.2363334298,-0.0177811794,-0.2750721872,0.0001792605,0.3736942112,0.1168587953,0.3918669224,0.1610017419,0.0871690363,0.3812574446,0.0504982956,-0.0901036188,0.4119758606,0.219390288,-0.0385965593,-0.2582117915,0.1683268249,-0.4234545529,0.3069981039,0.232055366,0.0618822426,0.0648707673,-0.1373666972,0.253241241,0.2994743288,0.4981358647,-0.4146616459,-0.3278160989,-0.2285789698,-0.298303932,-0.3188637793,0.3791346848,0.1398549974,-0.1967773736,0.0336276442,-0.3680784702,0.1499671191,0.0037718869,0.1940130293,0.2422606051,0.0545322224,0.0316536799,-0.0187701844,0.0221102182,0.0298728757,0.3937070668,-0.5399644375,-0.2046732903,0.2048074454,-0.5590269566,0.0730827823,0.0080572087,0.2777058184,0.352638036,0.4925115705,0.08901117,0.2042775005,-0.0375440791,0.188513726,0.1643131375,0.2119750381,-0.0647413805,0.4100411534,0.2708570957,0.3656848669,-0.1347787827,0.0432133712,-0.0054317582,-0.1764253378,0.6712611914,-0.1259947717,0.3084286749,-0.4890398979,-0.4388926029,0.271324724,0.0082045635,-0.0947970897,0.0869470164,0.322021246,-0.226584658,0.1516590118,0.1470181793,0.1476904005,-0.2125054002,-0.0541327372,-0.187337622,-0.0764364675,-0.0596771426,0.1976072788,0.2942281067,-0.1821978688,0.2553656697,0.1083121598,0.2190237492,-0.1960808337,0.2999760807,-0.1955657154,-0.0533633046,0.3180696368,-0.0322253928,0.0961275995,0.2476602346,-0.1812210828,-0.1439391971,0.1828757226,-0.2557730675,-0.2137421519,-0.1269117147,0.1511545181,-0.3514274061,0.0082077142,-0.2378200293,-0.400021553,0.4063311517,-0.3223438859,-0.0154792899,-0.3542110622,-0.2637994885,-0.353944093,0.2389694899,0.4025940299,0.0419260897,0.0426536016,-0.0905655473,-0.0114815729,0.294190079,0.183512345,0.0093251551,0.1056279317,-0.4124949574,-0.0635623261,0.2606512606,-0.3542023897,-0.6170941591,0.1036195084,0.1277518421,0.6295824647,0.0476492308,0.0750021562,0.1831660569,-0.0502420217,0.2216965407,0.1532353014,-0.038785696,0.1244795322,-0.252807796,-0.0325587615,-0.0434744433,0.1044865996,0.0201719031,0.2498004735,0.2889224589,-0.1661153138,0.2637803555,-0.0953656435,-0.1257363111,0.2321082801,0.2991938293,0.046871759,0.0639734343,-0.0167484935,-0.7693794966,0.3423186839,-0.1505941451,-0.0280080661,-0.0733518675,-0.0444171354,-0.1803081781,-0.1141534746,-0.3073681295,-0.0912047699,-0.0125983637,0.3906361461,0.2074201256,0.0201071296,0.0379508324,0.4137744009,-0.3047952354,0.1248284727,-0.3346244991,0.4118489027,0.0743400902,-0.1029386371,0.1318240315,-0.0030252386,0.1328218132,-0.0562920161,-0.2745426893,0.4387036562,0.3811519444,0.1363846362,-0.1962373555,0.3414201438,-0.0699199662,-0.1395473629,-0.012621915,0.4393106699,0.1203998923,-0.1690554321,-0.1533752978,0.3473964334,-0.0006330401,0.1946744621,0.0062628747,0.0575816073,0.2450569719,-0.1167023629,-0.1461344063,-0.261870563,0.3650653958,0.1296010762,0.2266256809,0.196940884,-0.1572066098,-0.0258856118,0.4515280128,-0.0735068321,-0.0205244031,0.1288514733,0.0474654101,-0.0436953641,0.0753767639,0.2370779216,0.5027197599,0.0668875426,-0.04627854,0.1425257325,-0.1718104333,-0.0678285286,0.0968001559,-0.0407392457,0.1989537925,0.6168555617,0.1380470097,-0.1526708454,-0.2818833888,-0.0774433017,-0.1217622161,0.3029148579,-0.4903008938,0.0769678801,-0.1072904691,0.0858299136,-0.3731725514,-0.1268986762,-0.1855977178,-0.2465110421,-0.2533238232,0.6255096793,0.0291794557,0.1770886779,-0.543838203,0.0261698551,-0.0482735336,-0.3658800423,0.0178647786,0.2034313083,-0.2111107856,-0.154736951,0.5788052082,-0.0910895392,0.3742569685,-0.4110652506,-0.0268509947,-0.5500833392,-0.2233636677,0.0945177004,0.0162294693,0.2236746848,0.171183303,0.0406548828,-0.0192004945,-0.2501042485,0.2409217954,-0.0442294776,-0.2682688534,0.3236261308,-0.1510273069,-0.0955459252,0.1709357947,-0.1985139698,-0.0594486669,-0.3050559163,-0.1297381818,0.2319205552,0.2133797705,0.1936335266,-0.0064717545,0.0152918147,0.1389674693,0.337371707,-0.1682054251,-0.6134564281,0.4448585808,-0.0378274098,-0.220205918,-0.1443342119,-0.1069903895,0.1174569875,0.3974053264,-0.5095450878,0.0167492125,-0.288091749,0.0945630446,0.0950631127,-0.0240606703,0.2160272449,0.145353362,0.0829630345,-0.2499599308,-0.2719066143,0.1908720434,-0.1372274011,0.1996578574,0.0226447172,0.3025070429,-0.2673497498,0.5462415814,0.4274444282,0.0061363615,0.2280853242,0.0467436984,0.526879847,-0.1603343338,-0.3179928064,-0.0785081387,-0.2098642141,-0.1336312145,0.1367129833,-0.025604425,-0.1290741563,-0.1602381468,0.0565649085,-0.2643253803,-0.2664655745,0.0557018742,-0.2999257445,0.341845274,0.1807065904,0.1787686795,-0.1449635029,-0.3675031662,-0.0111274412,0.2551630437,-0.0621705726,-0.064657025,-0.2754209936,0.0361303054,-0.0963153616,0.3009771109,0.2166886777,0.4385888875,0.1286434382,-0.0846547931,-0.0186727438,0.023536874,0.2270695567,-0.2749712467,0.2956960201,0.1485588551,-0.0079607638,-0.1783613861,-0.2544582188,-0.0432508513,0.0618162937,0.0433870666,0.55647856,-0.3165850043,-0.0163535215,0.2413590103,0.4100426435,-0.1461860985,-0.2883686125,-0.440358609,-0.1881668419,-0.2969538569,-0.1807302386,-0.1434826255,0.1207043454,-0.2336365879,-0.1006112695,-0.017981587,0.0570383631,0.297167033,-0.0569480471,0.4486523867,-0.062713787,0.4389171004,0.0435416959,0.0607733615,0.4035254121,0.8385460377,-0.3677003384,-0.4420888424,-0.0795324668,-0.1300328821,-0.1770050377,0.0727124736,-0.0297103859,-0.1339790672,0.2369088978,-0.1334208101,-0.0876609385,-0.0725696161,0.2104186416,0.1998604238,-0.4859289825,-0.4705499709,0.1424980611,0.0854976997,-0.0564965196,0.579628408,-0.0678586066,-0.17265971,0.0975924656,0.1562774181,0.7910794616,-0.0374635085,0.2097775638,0.2671902478,-0.321650207,-0.0153308865,-0.2884845138,0.1176057309,-0.5410556197,-0.4565611482,0.0023745324,-0.410885036,0.0694896355,0.0244534891,-0.0058735418,0.3644321561,-0.1365393251,0.3639833331,0.1626484692,0.150323987,-0.1333116144,-0.019252805,-0.0246920493,0.0563706756,-0.1063247249,0.2817863524,-0.1043262407,0.0744758397,-0.0837384537,-0.2491396219,-0.3359660208,0.26861462,-0.0484248325,0.3063894212,0.3144675493,-0.1562400758,-0.1932050735,0.3079376817,0.1931915879,0.1256464273,-0.202425912,-0.096611917,0.4446141422,0.1859632283,0.1080899015,-0.1174749732,0.4294384122,-0.0164770056,-0.1706129313,-0.0195093155,-0.1567083597,-0.5109879971,-0.0638424307,-0.1865808964,-0.0139560504,-0.3327461779,-0.1821816862,-0.1491386443,-0.0026479869,-0.179938361,0.1034976915,0.1509203017,-0.1339936107,0.3799248934,-0.0015962229,-0.460085392,-0.1274677515,0.5131280422,0.1256911457,-0.0666733161,0.52543962,-0.0354989879,-0.2741117477,-0.1303777695,-0.0239341632,0.0429769866,-0.2795833051,0.1564933807,-0.2923184931,0.3266904354,-0.1656529903,0.230919674,0.1746156663,-0.02821506,0.0803735107,-0.5436577201,-0.1149597615,-0.0545217469,0.0714057386,0.2256109267,-0.096170336,-0.1318736374,-0.1686248928,-0.1643115431,-0.219647333,-0.0509599522,-0.2850660086,-0.0503922813,0.0666950047,0.1091088355,0.095470123,-0.0847984403,0.1070708409,-0.1395269483,-0.1586124599,-0.0982483998,-0.2473136783,0.1913026422,0.007808188,-0.2195226997,0.1011908725,-0.2507101297,-0.1231054664,-0.2926606834,0.041283261,0.333221972,-0.1663248092,0.0240588076,0.0479134135,-0.0059108702,-0.0702901557,0.059866488,-0.3567721844,-0.0949055552,0.0353378989,0.0851883814,0.0412023254,-0.0427586734,0.0382646769,0.2007402033,0.0974865928,-0.0341187306,0.4755352736,-0.4363497496,0.0402533934,0.1563017517,0.2356877029,0.4222279489,-0.3856227994,-0.1175228581,0.13966766,0.1710722893,-0.2858359516,-0.1420300454,0.1406693459,-0.2054573745,0.0249944143,0.2416557819,0.1442941874,-0.5312721729,0.3154070675,0.0365457907,0.1829687357,-0.1733940989,0.0952961743,0.4206505418,0.1013368517,0.1403630525,0.1325903833,-0.0310446788,-0.0105605861,0.0478046462,-0.0750081539,0.3036599159,-0.3571105599,0.0880224854,-0.2201276869,-0.5279682279,0.3481909633,0.2502381504,-0.0698255226,0.1256925017,0.0152221387,0.0834440887,-0.0615339875,0.2207084745,-0.3515161574,-0.11219991,-0.1160394847,0.0300300941,0.0056752041,-0.3382467031,0.0237839688,0.2282619774,-0.0487359576,0.1440568417,0.1589454263,0.1219190583,-0.1803323925,-0.2696266174,0.2587604523,0.0842613652,-0.0409872234,-0.1821427047,0.2920826972,0.0527241006,0.0603767931,0.0833991989,0.2604502141,0.4520942867,0.5969523787,0.1655570716,-0.2217404097,-0.0189450718,-0.0231258385,-0.0942030549,0.7059627175,-0.1801467538,-0.2599339485,0.3139159977,0.0530928262,-0.1471292824,0.0253863428,-0.0972277522,-0.2940345109,-0.0907661393,0.5079358816,-0.0236264933,0.0174965095,-0.0995528772,0.2220041156,-0.2275825441,-0.249173522,0.4326468706,0.2270053178,0.1376857758,0.0776116177,0.0341995284,-0.0186288971,0.1766093224,0.2722485065,0.0902166814,-0.378665477,-0.342133522,-0.7264214158,0.3454352617,-0.1576506495,-0.0619914234,0.1172379106,0.1326255202,0.0116173243,0.1041900218,0.314648211,-0.0641261786,-0.2050143182,-0.0098502068,-0.1759048551,-0.3004679978,-0.1592682153,-0.1377318949,-0.1920399964,-0.5703731179,0.2017878592,0.1151962057,0.0482627079,-0.1229476482,0.2097163796,0.0996090695,0.2080218643,0.1852984279,0.1540618092,0.6352700591,-0.0647114143,-0.2254853845,-0.0971970707,-0.1141671911,-0.1586394906,0.3163270652,-0.1148585752,0.3463056982,-0.1001497358,-0.3019786775,-0.3574885726,0.3911955953,0.2497394234,-0.3221425116,-0.0838530287,0.0728310719,-0.2557837963,0.2184863389,0.0259722173,0.1138151437,0.1335904598,0.4593373537,-0.1675980389,-0.2450248748,0.464666605,-0.2160876095,-0.159683153,-0.1233223602,0.257687211,-0.2743937969,-0.2150910795,-0.6132447124,0.1759458482,0.4812847674,-0.1356473714,-0.1669290215,0.2582945228,-0.2036875784,0.3876015544,0.016391566,0.3485515118,0.0122529911,-0.222143665,-0.0536836199,-0.1185232848]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2879","title":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"","comments":"Hi @rcgale, thanks for reporting.\r\n\r\nPlease note that this bug was fixed on `datasets` version 1.5.0: https:\/\/github.com\/huggingface\/datasets\/commit\/a23c73e526e1c30263834164f16f1fdf76722c8c#diff-f12a7a42d4673bb6c2ca5a40c92c29eb4fe3475908c84fd4ce4fad5dc2514878\r\n\r\nIf you update `datasets` version, that should work.\r\n\r\nOn the other hand, would it be possible for @patrickvonplaten to update the [blog post](https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english) with the correct version of `datasets`?","body":"## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n","comment_length":46,"text":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"\n## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n\nHi @rcgale, thanks for reporting.\r\n\r\nPlease note that this bug was fixed on `datasets` version 1.5.0: https:\/\/github.com\/huggingface\/datasets\/commit\/a23c73e526e1c30263834164f16f1fdf76722c8c#diff-f12a7a42d4673bb6c2ca5a40c92c29eb4fe3475908c84fd4ce4fad5dc2514878\r\n\r\nIf you update `datasets` version, that should work.\r\n\r\nOn the other hand, would it be possible for @patrickvonplaten to update the [blog post](https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english) with the correct version of `datasets`?","embeddings":[0.124852851,-0.0535938032,0.0293307602,0.0739796385,0.2047281116,-0.0245766733,0.2489377707,0.4213005006,-0.4746908545,0.1751622707,0.1361838877,0.5513864756,-0.3347690701,-0.0850955397,0.0344839208,0.1865644157,-0.0921123177,0.1384560466,-0.2678906322,-0.2130741179,0.0398925841,0.3963992298,-0.1676281989,-0.0685962737,-0.3741768599,0.244877398,0.2273771465,-0.2314912826,0.1054832488,-0.4600336254,0.2924744487,0.1486980319,0.0194595307,0.3702304661,-0.0001191299,0.0063226023,0.2757813334,0.0352541432,-0.2775231004,-0.4479430318,-0.0650265291,0.1085864976,0.0530338921,0.1475536525,-0.4656309485,-0.1793126911,0.2337317616,-0.3409471512,0.0292004794,0.4485591054,0.0685843602,-0.0634173304,-0.2496642619,-0.0824669749,0.3689354062,0.0125433672,-0.0625187159,0.0592703931,-0.0683355555,0.2411289513,0.2196747363,0.3755480051,-0.0313242152,0.0129956314,-0.1954175979,-0.0935640186,-0.2540271878,-0.3487864733,0.0747668445,0.2519178092,0.7106900215,-0.2958623171,-0.3215118349,-0.2683330774,0.3369950652,-0.0542203225,0.2911206484,0.0254176874,-0.2944933772,0.1627251506,-0.5629917383,0.0742394626,-0.3041745424,0.1429014206,-0.0974289998,-0.359634608,-0.143771261,0.1845409572,-0.1975932419,0.1163858026,0.114155367,0.1884384304,-0.0396124236,-0.0923857763,-0.3206233084,-0.0985560194,0.0559260733,-0.0815377459,0.0717144236,0.3172878623,0.2842969596,0.0675945058,-0.4174665213,-0.135505408,0.0322321728,0.1481688768,0.1823079735,-0.1778396666,0.101101391,-0.0720301494,-0.1533745825,0.0139923282,0.0189836212,-0.0720990822,-0.1062536985,-0.0153032569,0.3571920395,-0.1175772175,-0.5631766319,0.1525668353,-0.5025030375,0.2135901749,-0.0520560816,0.1947724819,-0.2154174149,0.1350183487,0.2971945703,0.0870628357,0.02487945,-0.3115249574,-0.0998023972,-0.1971572787,0.1828780323,0.0968628898,0.2440094203,-0.2103739977,0.2123047858,0.257135421,0.2615037858,-0.1538840383,-0.0926653668,-0.0326226279,0.2529811561,0.2668441832,-0.1248196512,0.3939130902,-0.0565123409,0.1157927513,0.106670782,0.2571235001,-0.0651218817,-0.0061710193,0.5623431206,0.060588941,-0.1836533099,-0.3006524444,-0.1071216986,0.3333047032,0.0442795344,-0.0668319687,0.2504209578,-0.2751781344,-0.2271512002,0.0158824958,0.1358926445,0.1566538513,-0.5853290558,0.0243286006,0.1945606619,0.0684714913,0.3222911954,0.3032742143,0.1651955247,0.1811709255,-0.0816664621,0.1170219034,-0.219933033,-0.3883649707,-0.2060429454,0.2092838287,-0.2708128095,0.264259249,-0.0832697079,-0.2192167342,0.0802669078,0.0991331413,0.3640693426,-0.1891600937,0.24668172,-0.1056700721,-0.3831144273,-0.1445571929,0.1893382818,0.1755265892,-0.0177115817,0.0532647818,-0.2982362509,-0.0164561383,0.4027053714,0.0394103304,-0.0291296821,-0.025731707,0.3644500375,-0.1775300503,0.263861388,-0.1554233581,0.2872300744,0.0272912737,0.1604733467,0.2198717296,0.5499910116,-0.1693307608,-0.1779452264,-0.0762331262,-0.0728189349,-0.0683157369,0.0582602769,0.1486589164,-0.1489184052,0.3247836828,-0.0285170637,0.4712553024,-0.4881338477,-0.0153098824,-0.0879812539,-0.1088415384,0.3033908308,-0.0987990201,0.1165293604,0.2411418706,0.0172993075,0.0872968137,-0.110142529,0.3636822999,0.2269500643,0.1219172627,-0.3682011068,-0.4027226269,0.2316687405,-0.5078880787,-0.4552857876,0.5677568913,-0.0276648402,0.2008169293,-0.1534452736,0.060441643,0.0807491466,0.0479999073,0.0348503962,0.0360783227,-0.019409189,-0.1079213247,-0.470849663,-0.1607666463,0.2607947588,-0.2098269463,0.1590214819,-0.1161290258,-0.320378989,0.1368902475,0.315872401,-0.0502400734,0.0395470038,0.2744301558,-0.1326723844,0.209983319,0.0375697017,0.1719760001,0.2373924702,0.2009201497,0.0807271153,-0.0535479188,-0.0781926885,-0.2605938315,0.2923918068,0.1945406348,-0.3651409447,0.3840377629,0.2203695476,0.0489813201,-0.4372256994,-0.0049824319,-0.1904428899,0.2824560404,-0.449403286,-0.1749291867,-0.3977611363,0.1078346595,-0.1361607611,-0.0218015406,0.0710958019,0.0436779261,0.3717918992,-0.0027077466,-0.0859973133,0.2743452489,0.1501078755,0.1915652007,-0.110422954,0.1713414043,-0.0096689994,0.1939062178,-0.4710183442,0.0452458076,-0.1606466472,-0.2644656301,0.1504383832,-0.1925743371,-0.1099141166,-0.4143466055,-0.071546413,0.1478731036,-0.1916618943,0.4232137799,0.2326952517,-0.0426136926,-0.1860425919,-0.194430843,-0.1887598336,-0.0310719088,-0.0671303943,0.0714476034,-0.1346394867,0.0415853858,-0.256131798,-0.8888216615,0.1003023162,-0.2204675525,0.7185165882,0.0599351153,-0.0460496917,0.5412366986,0.2822042704,0.2012919486,0.0792539567,0.1863065064,-0.3713881075,-0.1230798736,0.0792584717,0.0293458365,-0.4573093057,-0.298314631,0.1936385185,0.1548020095,0.2074109167,-0.4953384995,0.3497347236,-0.1089945436,-0.1306295842,0.0102862855,-0.1989106387,0.2906790078,-0.2038048357,-0.1159159765,-0.2554523647,0.1175914109,0.3896949589,0.1017955169,0.112711817,0.3398635685,0.2160488367,0.2003065348,0.3506759107,0.7158643007,0.0828187764,0.039017871,-0.1280344874,0.4244537055,-0.0728549138,-0.052832868,0.1600767821,0.0590854175,-0.1087116003,0.4583721757,-0.0842429325,-0.0255117994,-0.0293365568,-0.0642682612,-0.0904889032,-0.3277724385,-0.022520544,-0.4580339193,0.0512960665,0.1535509676,0.2963804603,0.1280051917,-0.1144675538,-0.0452858321,0.1441749334,0.3194569945,-0.1257756054,-0.3150956631,-0.3613077104,-0.415389359,0.0612773709,0.1869184822,0.5611051917,-0.2880935073,-0.201971963,0.2231436074,0.1043023616,0.4215117395,-0.3380554318,-0.042292852,0.1251643896,0.2728692293,-0.2976222634,-0.0735439733,-0.3262295723,-0.289985925,0.279160589,0.0972543359,-0.1177073866,-0.1351196617,0.0686608776,0.058421474,0.0167316217,-0.099004969,-0.0332190059,0.0306317639,-0.0272785462,0.1637768447,0.1725981981,-0.1248045266,-0.491910845,-0.3849042654,-0.0554664992,0.1896433085,0.5516821146,0.0221093055,0.2840608358,-0.1308684945,0.2439791858,0.235736981,-0.2960318625,0.2580799758,0.1936600059,0.1173487008,-0.0414018072,0.1370378286,-0.1260076612,0.3830217123,0.3855090439,-0.4349811971,-0.0189787298,-0.3385869563,0.1840335429,-0.2829484046,0.3459416926,0.4015236199,-0.1441545635,-0.1260796487,-0.4360538721,0.418184042,0.0229241867,-0.1640126258,0.4557494521,0.161981672,-0.2460984588,0.1927152127,-0.1567512751,0.8095886111,0.1871284097,0.2667512596,0.3340376318,0.1109344587,0.1609196216,-0.4287750423,0.0356836617,-0.0055845086,0.0698633865,-0.1306088716,0.089701362,-0.1681430936,0.3797655702,-0.2919414937,-0.0783907324,0.2101620287,0.197283715,0.2774983346,-0.0788463205,-0.2581412792,0.0186328627,0.0897454992,0.0458310619,0.058744546,0.3057942092,0.2888186574,-0.1282046884,0.0097754784,-0.265276581,-0.6595223546,0.0203248579,-0.599691987,-0.09471035,0.1489200592,-0.3863419294,0.183830902,0.3397494555,0.1764554679,0.4337405264,-0.4633069336,-0.0267995987,0.2511091232,0.1131987423,-0.0246270187,-0.0830141678,0.3481529355,-0.0840320587,-0.1434060186,-0.0710047334,0.0281203054,0.16605708,-0.1545241773,-0.0496859439,-0.1185674593,-0.3426322937,0.0497712791,0.0989603326,0.1527892351,-0.2369833142,0.0680043399,-0.1038809866,-0.0651224554,0.2742635012,0.2779468596,-0.5032378435,-0.1089579538,0.5379329324,-0.0835737288,0.1077651456,0.6408995986,-0.1893859357,-0.0512990244,-0.2775192261,0.0168702956,0.2590861022,-0.3368996084,0.1719175428,0.0321254991,0.30066517,0.0966014713,-0.0444742106,0.0463051498,-0.0589915104,-0.2513128221,-0.2389338315,-0.081248261,0.1200233996,0.0252261981,0.2911800742,0.0278385542,0.3665986061,0.0767617822,-0.347037375,-0.2017673552,-0.0237252936,0.0557583757,0.0804858208,-0.0478359275,0.0278626494,0.1107278466,0.0545184948,0.0149586732,0.1027084291,-0.1335515976,-0.0833661035,-0.2990173697,0.1511185914,0.2011461705,0.1130731627,-0.0153080691,0.0650832951,-0.0542607009,-0.0600513965,0.0761152506,0.3808921576,0.0174185708,0.3620303869,0.1188446134,0.3976139128,-0.0936357677,-0.2623862922,-0.0354406275,-0.0775718689,0.1644935161,-0.0624371767,-0.1850817502,-0.0580990687,-0.1324884593,0.0220973957,0.0893222392,0.1518605053,-0.0546601489,-0.2896298468,0.2135311365,0.1549112052,0.0506647453,0.3179009259,-0.1531378031,-0.2708627284,-0.1102764606,0.0789906234,-0.0246088952,0.0176900197,-0.3351352513,-0.0152753768,-0.0255123302,0.0031253609,0.0554810911,-0.3541377783,0.3623180091,-0.0125009436,0.1121347845,-0.5960302353,0.1129645109,0.3746998906,0.0613390915,-0.0079180803,-0.2586423457,-0.2529646456,-0.0035204799,0.4860927463,-0.2427177429,0.2749043107,-0.1475844979,-0.0262315013,0.2557944953,-0.2615301609,-0.2477716953,0.2055236101,-0.4631261826,0.160956189,0.101355657,0.4035341442,-0.0463260636,0.1006667241,-0.2105019838,-0.0210334621,-0.1480747759,-0.1586381793,-0.3575797379,-0.0887535512,-0.1383645236,-0.1716290116,-0.0802430212,-0.0285868235,0.1238230765,0.2392329723,-0.2778191268,-0.5319434404,-0.2491969317,-0.1834071875,0.251419425,-0.1140987724,0.3520128727,0.103128694,-0.0186601989,0.1629291028,0.3187715411,0.5043587089,0.1790820807,-0.1014207229,0.1203379259,-0.299844265,-0.0056224484,0.020882478,0.62394315,0.0326780751,0.1173429191,0.0509871952,0.0518923216,0.0013063183,0.0413613394,-0.0084985485,0.0400381945,-0.6968811154,0.4176638126,0.0176094957,0.2393384725,-0.2014484107,-0.0412869416,-0.2195517421,0.3418336213,0.6031695008,0.1864216924,0.0186088011,-0.0934841856,0.0555081293,-0.1810229719,0.1212917343,0.0766027942,-0.0048448374,-0.1068457067,-0.0195507258,-0.2139687836,0.0899255127,0.265216738,0.2187431008,0.0150080938,-0.1638361514,0.0671154484,0.0615044869,-0.0098666847,-0.1182038561,-0.2539238632,0.2244302034,-0.1969873458,0.0151608726,-0.0069401781,0.1406158358,0.1495863199,-0.0737226456,-0.1326639354,0.1910309643,-0.1454409212,-0.057077501,-0.0759500638,0.1773371249,0.3487201631,0.0686021373,0.0964073241,0.3599496484,-0.0915555507,0.07731089,-0.4108827412,-0.1420516521,-0.2299602181,0.2994242013,0.049812071,0.3389099836,0.0066421395,-0.3442056477,0.1354150027,0.2597314715,0.3330140114,-0.162827,-0.1400983036,0.0477316156,-0.323091656,0.2139669657,0.0000565254,0.2621964514,0.0430665314,0.1607147604,-0.5121689439,-0.9930148125,0.6250440478,-0.7181706429,-0.2236547172,-0.1427765787,0.2892617285,0.3226871789,-0.1015825123,-0.5634548068,-0.0728689656,0.4693164527,-0.0401696749,-0.0920308381,0.112321049,0.1479113549,0.3100144565,-0.164433822,0.9622838497,0.3237745166,-0.1628977358,0.1872403026,-0.4706269205]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2879","title":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"","comments":"I just proposed a change in the blog post.\r\n\r\nI had assumed there was a data format change that broke a previous version of the code, since presumably @patrickvonplaten tested the tutorial with the version they explicitly referenced. But that fix you linked suggests a problem in the code, which surprised me.\r\n\r\nI still wonder, though, is there a way for downloads to be invalidated server-side? If the client can announce its version during a download request, perhaps the server could reject known incompatibilities? It would save much valuable time if `datasets` raised an informative error on a known problem (\"Error: the requested data set requires `datasets>=1.5.0`.\"). This kind of API versioning is a prudent move anyhow, as there will surely come a time when you'll need to make a breaking change to data.","body":"## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n","comment_length":134,"text":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"\n## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n\nI just proposed a change in the blog post.\r\n\r\nI had assumed there was a data format change that broke a previous version of the code, since presumably @patrickvonplaten tested the tutorial with the version they explicitly referenced. But that fix you linked suggests a problem in the code, which surprised me.\r\n\r\nI still wonder, though, is there a way for downloads to be invalidated server-side? If the client can announce its version during a download request, perhaps the server could reject known incompatibilities? It would save much valuable time if `datasets` raised an informative error on a known problem (\"Error: the requested data set requires `datasets>=1.5.0`.\"). This kind of API versioning is a prudent move anyhow, as there will surely come a time when you'll need to make a breaking change to data.","embeddings":[-0.088485904,0.037396457,-0.0213857722,0.1496120393,-0.0189264342,-0.0777354389,0.2245492935,0.3621777296,-0.429489553,0.1633367538,0.1525295675,0.2624569833,-0.432490617,-0.0201306585,-0.071119301,0.2396375537,-0.1268624365,0.25981462,-0.3329852521,-0.0231152549,0.0825956166,0.2425760627,-0.0296556633,0.0283614192,-0.0865963995,-0.0143052982,0.2008777112,-0.093485117,-0.3122866154,-0.6014047265,0.2304865569,0.2764073014,0.0768745914,0.1653757393,-0.0001274555,0.1065067053,0.1652288288,-0.0230602305,-0.4907235503,-0.1459860653,-0.180824548,0.1132927239,0.011158362,-0.0032401655,0.1294730902,-0.2341334075,0.2061720639,-0.1966678202,0.3206728995,0.5292428732,0.0049742479,0.2393674999,-0.1227483228,0.1194096953,0.3548440039,-0.1067724153,-0.0561466739,0.0766564906,0.0378676206,0.2733804584,0.2134608924,0.3169699907,0.086405769,-0.1138325781,-0.290795207,-0.193882212,-0.1545206457,-0.2383064628,-0.2170854211,0.3288349211,0.753831327,-0.3150245845,-0.6087892652,-0.3401043713,0.1912582219,-0.1430546939,0.3923214972,-0.1902457625,-0.2573828101,0.3908389211,-0.6390121579,-0.2049098164,-0.3079322577,0.1295725256,0.015353418,-0.1147583276,0.0225799382,0.1235876381,-0.1734608263,0.0642666593,0.1816413701,0.1108352542,0.038729921,-0.0109756803,-0.1475488394,-0.3412714601,0.0423946492,-0.1861126125,0.0421029404,0.4260554612,0.5204828382,0.0642907992,-0.5445080996,-0.0775154456,0.1737653315,0.2407868356,-0.0252583344,-0.2616614103,0.15692994,0.0426035337,0.1903326213,0.1748280823,0.0237038396,-0.3405395448,-0.1293011457,0.1194925904,0.3347736299,-0.1888237745,-0.5020070672,0.0776049346,-0.454580307,0.1854378134,-0.0716408938,0.120723933,-0.2241946012,0.419606179,0.1554582864,0.0640978739,-0.1348512471,-0.4040712416,0.1883217096,-0.1559716016,0.1908436567,-0.0499088056,0.2960176468,-0.3722983003,0.0881588981,0.1427764744,0.1810636073,-0.1214631051,-0.1279874742,0.1010570377,0.1780343056,0.2132166326,-0.1896715909,0.1723005623,-0.0805888027,0.0884558037,-0.0172109585,-0.0199682843,-0.1741253138,-0.1103576198,0.3484823704,-0.0193801634,-0.2260963023,-0.2067113668,-0.0028787269,0.1341750324,-0.0354007855,-0.2685537934,0.1478761584,-0.115805082,-0.3188738525,-0.0152796609,0.0157302301,0.2144396305,-0.6616707444,-0.1349116415,0.1824918687,-0.0098238224,-0.0802729949,0.2591881156,-0.018249182,0.1335975379,-0.0214253534,0.061384391,-0.0431077182,-0.4288671315,-0.3187329769,0.1890932769,-0.2718712687,0.2365475744,0.1732592136,-0.2317657024,0.3257893324,-0.1745207906,0.1762593389,-0.2554750144,0.1786212027,-0.1710287333,-0.4389976859,-0.4511151314,0.031018842,-0.0533370189,0.068120487,0.0648189187,0.0564532466,0.0423984602,0.3488164246,0.0829287097,0.0118160006,-0.0338022485,0.200045526,-0.0843634456,0.0186142046,-0.224367395,0.0837669671,0.2081040293,0.0765817836,0.1705437154,0.424785018,-0.1788806766,-0.2476136684,0.0383234285,0.1872706562,0.126285255,-0.0395877548,0.2273307592,-0.0825186446,0.3504412174,-0.2704238892,0.4691356719,-0.4019192457,0.0372846611,-0.0536281206,-0.1635846496,0.2474771291,0.0825913176,0.1019182429,-0.0394122601,-0.1032757536,-0.0926558524,-0.024524264,0.518794775,0.1530638039,0.2767786682,-0.1384766251,-0.2536575794,0.2143708169,-0.2618277371,-0.3397199214,0.5486403704,-0.0972192958,0.4484133124,-0.0255129375,0.1951500177,0.1404719353,-0.040309187,0.0705885664,-0.0731970817,-0.0213092566,-0.0093516512,-0.5304045081,-0.1736079603,0.1465597302,-0.2134280056,0.1417781264,-0.1460386813,-0.2046640664,-0.0846562311,0.2301444709,-0.0212985929,0.0592088923,0.1565668881,0.0832527056,0.211786449,0.0553550608,0.3183932006,0.3474945128,0.0831101686,0.0246619973,0.0214087591,-0.2565384805,-0.1302022934,0.2768206596,0.2188893259,-0.3133857846,0.3226563931,0.2063322961,0.0317539014,-0.4395200908,0.0319717452,-0.0409543738,0.0367235951,-0.6262536049,-0.0090599591,-0.5513135791,0.1690156907,-0.3306770027,-0.08182659,-0.1623638719,0.0047671883,0.5723306537,0.1694382876,-0.0812915936,0.2194667459,-0.087634325,0.3680031896,-0.1986395717,-0.0618027821,0.0382376686,0.235950008,-0.4149255753,0.0028021303,0.0273967143,-0.3462299407,0.2414893061,-0.4148816168,-0.3009605408,-0.2322787941,-0.0679201856,0.1891996413,0.0653937161,0.2402111292,0.3457305431,0.1270100325,-0.0146292243,-0.0418712981,-0.2779378891,0.0484673493,-0.1159968078,-0.1051494107,0.1344976276,0.0982847586,-0.2827249169,-0.8775423765,0.0841022357,-0.1689092368,0.4604318142,0.0795201138,0.1856422573,0.3558037281,0.346203953,0.1059153229,-0.0441331379,0.1544438154,-0.0882074237,-0.2302579731,0.0935372785,0.0836211443,-0.2186214924,-0.0057524103,0.2806986868,0.0527223162,0.3915783763,-0.4168854058,0.2677829564,0.2194800079,-0.1391707957,0.001931228,-0.368018806,0.5466492772,-0.0834411085,0.0156952888,-0.2377285361,0.205624491,0.1374920607,0.122576125,0.2549115419,0.3231088817,0.2804227471,0.2665038407,0.5263994336,0.3263044357,-0.2103032321,0.2247066796,-0.1228169724,0.4726446271,0.0411540791,0.0922420695,-0.0651079491,-0.0960020721,-0.1824163198,0.3834352195,-0.0364108309,-0.0770952627,-0.1888981909,0.2259602696,-0.2749902606,-0.272918582,0.0425743796,-0.3952446878,0.1360569596,0.2876459062,0.3582558036,0.10156288,0.0322360359,-0.0319893025,0.2270751446,0.3397321999,-0.0098600397,-0.205221802,-0.0876502991,-0.2094879746,0.2032209635,0.4069237113,0.6541618705,-0.1864350736,-0.0886587799,0.1288477182,-0.0191108957,0.2862899005,-0.4888174534,-0.0456894152,-0.0792832151,0.3533954322,-0.1327290684,-0.1294834763,-0.1357855201,-0.3984862864,0.1170825735,0.0794346184,-0.1247220188,-0.0330578312,-0.0315417647,0.1264813244,-0.1087495461,0.1763421148,-0.1473098397,-0.0187249426,-0.1351140738,-0.008912567,0.0840163231,-0.278080076,-0.4168049693,-0.0725171417,-0.0888437629,0.2681666613,0.3895969093,0.1621502638,0.2874218822,-0.1167992502,0.0239205156,0.0930395275,0.0980092809,0.3213100731,0.2371076494,0.1738164723,-0.1843378693,0.0630599037,0.0820485875,0.101782836,0.5416073799,-0.336302489,-0.1183999628,-0.1456222981,0.0961889178,-0.2815686762,0.3941451907,0.539711833,-0.2854302824,-0.3250682056,-0.4289255738,0.4246477187,-0.0904475898,-0.2281326652,0.3700994253,0.1016527489,-0.3230112493,0.1511365473,0.0276337042,0.9220039845,0.2088724077,0.2067637593,0.1859174073,-0.1246660948,0.4331418574,-0.5090640783,0.1522121876,0.1198083013,0.0101093641,-0.3668872118,0.1092012301,0.0915936753,0.439971447,-0.1749594063,0.0485190079,0.1938182414,0.3199470937,0.243715927,-0.0224065706,-0.4007433057,0.095558621,-0.1448121965,-0.0000623971,0.1024669558,0.2897033691,0.0265425239,-0.0909916908,-0.0498299003,-0.4163598716,-0.7078104615,-0.2266657054,-0.4140581489,-0.1478801817,-0.0445111394,-0.119061254,0.1190106422,0.2762684822,0.2014925331,0.2418511808,-0.3773293197,0.0200512931,0.0144654009,0.0742946267,0.1023879573,-0.0653457791,0.3629058599,-0.0745208785,0.0413267799,0.0539514199,-0.0798187256,-0.0671611503,-0.102445133,-0.0455864854,-0.0551971756,-0.3396161795,0.1325672865,-0.1026446521,0.0545611791,-0.1985493749,-0.0092557035,-0.2127401829,0.1121369526,0.1885818541,0.0892388821,-0.4363735914,-0.0978496596,0.6311433911,-0.2886928022,0.1289472431,0.4841496348,-0.3160094321,-0.0333173946,-0.1941363215,-0.0213660095,-0.1992967427,-0.5194515586,0.0147852106,-0.0446038842,0.0800760463,-0.127022177,0.073364988,0.1802175939,0.1770586073,-0.1799434125,-0.3772439957,0.0420021489,0.2579897642,0.2447490245,0.3469192088,0.0996372178,0.2187812775,-0.0042199041,-0.5193978548,-0.1086388677,-0.1001783013,0.1144042015,0.0074447608,0.0580743849,0.3495641649,0.0714690611,-0.2546701133,-0.1440987289,0.1539246291,-0.1497370154,0.0224602874,-0.2373713255,0.2285162508,0.0468573086,0.1454444677,-0.0425269417,-0.0164453555,-0.0305593424,0.0448148474,-0.0886746049,0.3584471941,0.1018417403,0.191381067,0.1421771497,0.4053522646,-0.2433033288,-0.0650703013,0.068246454,0.1981734782,0.1225374714,0.2254868448,-0.2010090202,-0.0713312626,-0.1664502025,0.2583247721,0.2247251868,0.0503340773,0.0594959557,-0.1397160292,0.2466144562,0.0786284432,0.3503730297,0.6881967783,-0.1185354069,-0.4578858316,0.1832003444,-0.0270989817,-0.1317240745,-0.025419686,-0.3005431294,0.0465860665,-0.1627551615,0.1273007691,0.0149540277,-0.3517835438,0.460877955,0.1213573515,0.1400361806,-0.4294632971,0.1858078986,0.6682862639,0.0034412649,0.2189691067,-0.4196910262,-0.3038707972,0.1169409975,0.3169998527,-0.3769645095,0.2690323591,-0.1761023998,-0.0776479989,0.0045769424,-0.1181805208,-0.3301261067,0.2845059335,-0.0586046278,0.4738274813,0.5105355978,0.5142284632,-0.0172043759,-0.0909561291,-0.4596327543,0.0934664384,0.0878878832,-0.1986487806,-0.5646128058,-0.2639584839,-0.1481105834,-0.0202088729,-0.1672478616,0.1114047915,0.2192398906,0.28841874,-0.2779063284,-0.5320429802,-0.2158830315,-0.2374021709,0.3170849979,-0.2049438506,0.2701926231,0.2036055326,0.077412039,0.2706266344,0.3161221743,0.405833602,0.091557838,0.0079941442,0.0263181683,-0.2494978458,0.2277772725,0.0264974926,0.4185414016,-0.0022926908,0.1665539294,0.3360890448,-0.0296551175,0.0563846491,0.1897279322,-0.3109255135,0.0160298627,-0.8132576346,0.3678290546,-0.0193114169,0.2532857358,-0.205313772,0.0359720923,-0.0377472825,0.3639247119,0.374653399,0.2076763511,0.0715543926,-0.001427614,-0.0013188893,0.042796094,0.1832635552,0.3172470033,-0.0597401001,0.0115216132,-0.0177085325,-0.5737120509,0.0423169583,0.0510372594,0.0625441298,-0.0109656798,-0.2795841694,0.2245611697,0.3119564056,0.1439362615,-0.0390034355,-0.2761303186,0.3555093706,-0.1490615904,0.0674827844,-0.0336016417,0.2049396783,0.0936795622,-0.4758703113,0.1259448081,0.0696859658,-0.2321201712,-0.0411871411,-0.023927616,0.1523931175,0.1186016798,0.2039049417,0.0417625122,0.3559782207,-0.0677288547,-0.1148146093,-0.5655712485,-0.0974949151,-0.245974496,0.312805146,-0.1374589652,0.3027483225,-0.1450625956,0.0440846756,-0.0077923709,0.3880821764,0.3565621674,0.0990418866,-0.1623943746,-0.2619998753,-0.3227412105,0.1392250359,0.265111953,0.2450653315,-0.0310609378,0.1129158437,-0.4395873845,-0.7549797893,0.4463395476,-0.5544971228,-0.0952443779,0.0162800103,0.3003892601,0.3546909094,-0.0889625624,-0.5038310885,-0.1051778495,0.3698131144,0.0651458949,-0.0476119258,0.0583503135,0.1136684343,0.0546242669,-0.02786153,1.0447940826,0.216182217,-0.1762980819,0.1037134603,-0.4345042109]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2879","title":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"","comments":"Also, thank you for a quick and helpful reply!","body":"## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n","comment_length":9,"text":"In v1.4.1, all TIMIT train transcripts are \"Would such an act of refusal be useful?\"\n## Describe the bug\r\nUsing version 1.4.1 of `datasets`, TIMIT transcripts are all the same.\r\n\r\n## Steps to reproduce the bug\r\nI was following this tutorial\r\n- https:\/\/huggingface.co\/blog\/fine-tune-wav2vec2-english\r\n\r\nBut here's a distilled repro:\r\n```python\r\n!pip install datasets==1.4.1\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\", cache_dir=\".\/temp\")\r\nunique_transcripts = set(timit[\"train\"][\"text\"])\r\nprint(unique_transcripts)\r\nassert len(unique_transcripts) > 1\r\n```\r\n## Expected results\r\nExpected the correct TIMIT data. Or an error saying that this version of `datasets` can't produce it.\r\n\r\n## Actual results\r\nEvery train transcript was \"Would such an act of refusal be useful?\" Every test transcript was \"The bungalow was pleasantly situated near the shore.\"\r\n\r\n## Environment info\r\n- `datasets` version: 1.4.1\r\n- Platform: Darwin-18.7.0-x86_64-i386-64bit\r\n- Python version: 3.7.9\r\n- PyTorch version (GPU?): 1.9.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: tried both\r\n- Using distributed or parallel set-up in script?: no\r\n- \r\n\r\n\nAlso, thank you for a quick and helpful reply!","embeddings":[0.1843589097,-0.0476369262,0.0212153625,0.1307302862,0.1825789064,-0.0150703816,0.222696349,0.4102399647,-0.4767715335,0.1482010931,0.202155903,0.528619945,-0.3603378236,-0.1392592937,0.0697514489,0.1514802277,-0.0296482705,0.1163090765,-0.2856072187,-0.2430297881,0.0460388102,0.3836102486,-0.181789279,-0.0110180005,-0.3692496121,0.2394319922,0.2274626493,-0.1810913682,0.1650811285,-0.439029336,0.2945018709,0.0983762294,0.0515014045,0.3712564111,-0.0001205389,-0.0167328734,0.2610318065,0.0087951152,-0.2081606984,-0.3689297736,-0.1382844448,0.1195933074,0.0133538358,0.12241216,-0.4130392373,-0.2065744102,0.229093805,-0.3684004843,0.0639212877,0.4605449438,0.0608402193,-0.139359951,-0.2579361498,-0.0793896466,0.3681408465,-0.0151606631,-0.0316133946,0.0299798381,-0.0864327401,0.2322303355,0.2330624908,0.3767850101,-0.0501611307,0.0072294213,-0.2207338065,-0.0934685469,-0.2900455594,-0.3867571652,0.0662252977,0.191508621,0.7059831023,-0.3136530221,-0.2619046569,-0.2177537978,0.3014097512,-0.0254760459,0.2628924251,0.0440048538,-0.2687057555,0.1821450889,-0.5459147692,0.1098416448,-0.3146458268,0.1415527016,-0.0640842691,-0.3725824058,-0.0981010497,0.2094152719,-0.2525474727,0.1218328848,0.1371880621,0.2098523825,-0.0407567285,-0.088758558,-0.3333061934,-0.0827903822,0.0905025452,-0.055269625,0.0512395315,0.2592660189,0.323356241,0.0425352305,-0.3752723634,-0.1163501963,0.0656884834,0.1570367366,0.0961320028,-0.1948441863,0.118978247,-0.1403936744,-0.1913475394,0.040307723,-0.0055863252,-0.091643773,-0.0389522314,-0.0102081764,0.3480204046,-0.1361118853,-0.5908363461,0.1955148876,-0.5599816442,0.2329349965,-0.0647064596,0.2073190659,-0.187874347,0.1397468895,0.3171749115,0.0926715434,-0.0059219012,-0.3403543532,-0.0963281542,-0.1902716309,0.1971511841,0.0616736077,0.2055280805,-0.1423812211,0.2137727439,0.2923474014,0.2695505917,-0.1473722309,-0.0583484769,-0.0180806592,0.219842881,0.2483473718,-0.1362519711,0.4016667604,-0.0248719845,0.1940863729,0.1011622474,0.2526570261,-0.1005024016,-0.0175146516,0.5515517592,0.0515450686,-0.179110989,-0.2624935806,-0.0348146893,0.3113076985,0.0846941844,-0.1433292776,0.2118196636,-0.2840275764,-0.1747703999,-0.003525354,0.0937831625,0.1310754865,-0.5885374546,0.0558972545,0.228044942,0.1303474754,0.3747636676,0.2952876389,0.1747010648,0.2124052495,-0.0759727508,0.1145521328,-0.1961544603,-0.4299621284,-0.214312613,0.1697297394,-0.2615546882,0.2817715108,-0.0361457802,-0.2367807031,0.0825164616,0.075459443,0.409607172,-0.2213379294,0.1860471368,-0.0808987841,-0.3983337283,-0.1363315433,0.2186726034,0.1355755478,-0.0341216251,0.0313469507,-0.2578755319,-0.0243165027,0.3640209436,0.0280066933,-0.004362273,-0.0479702353,0.3482239544,-0.2454301864,0.3254101574,-0.0976360664,0.3093805313,0.0151669383,0.1946799606,0.1750653833,0.6044173241,-0.1565310061,-0.170942992,-0.0799789056,-0.0831656903,-0.0937429518,0.0424369089,0.1493053585,-0.1577068269,0.3208678961,-0.0529016294,0.496171236,-0.4533744752,-0.0638343394,-0.0522209033,-0.1226176918,0.2798168063,-0.0861343965,0.0933275223,0.2543967366,0.0215519834,0.0870476365,-0.0931508169,0.3177514076,0.2538267672,0.1660675555,-0.3391086459,-0.3895223737,0.2347876579,-0.5418210626,-0.418391794,0.6494663954,0.0136631234,0.2044710368,-0.1246106178,0.0206717309,0.0992714986,0.0414252579,0.0030764376,0.0376889668,-0.0036091113,-0.0781878829,-0.4513742328,-0.0986297056,0.2685841024,-0.2420790792,0.1837171465,-0.1044279113,-0.3128755689,0.1960173547,0.3068779707,-0.0273673404,0.0484327525,0.2494100779,-0.1171611398,0.1943250597,0.0596161298,0.2010252476,0.235914126,0.1829334795,0.0925755948,-0.0770207793,-0.0987522975,-0.221516192,0.2697022557,0.1580257267,-0.3681939244,0.4219260216,0.2311407924,0.0666007549,-0.3795827329,0.0026456879,-0.2107334882,0.3031838238,-0.4800261259,-0.1368284374,-0.3495121598,0.1713373214,-0.1380009055,-0.0323438756,0.1311494857,0.0800623074,0.370326221,-0.0135218948,-0.1168963611,0.2293738723,0.1091258749,0.2434615046,-0.0670005679,0.2318317592,0.0222459622,0.2099670172,-0.4693998694,0.015322485,-0.1299001575,-0.2572244108,0.1167961061,-0.1661652029,-0.0994341001,-0.3659780025,-0.0571099631,0.151061967,-0.1844133586,0.4728702605,0.2280860841,-0.0554880984,-0.1941481084,-0.2376766503,-0.2153888792,0.0371617377,-0.0857695267,0.097694777,-0.1236358881,-0.0152868517,-0.2216340005,-0.9135431647,0.0924015716,-0.2011582255,0.6722809076,0.0548606068,-0.0086461287,0.4621789753,0.2635646164,0.1870574206,0.138909936,0.2119944692,-0.3575545549,-0.106063135,0.0823722705,0.0353392884,-0.4551185369,-0.3046335578,0.1577614248,0.1924328506,0.2105116844,-0.4597027302,0.4100583196,-0.0954082683,-0.1507000029,-0.0038921644,-0.2178034633,0.3120282888,-0.1800539792,-0.0859067366,-0.2840114534,0.1318731755,0.3946647942,0.1762969643,0.1079193801,0.3068368733,0.1843467206,0.1532683223,0.3919531703,0.7317343354,0.1164943278,0.0385542214,-0.1042457595,0.3726763427,-0.0699758604,-0.0704685152,0.1748739481,0.0528061725,-0.2046083659,0.4510253072,-0.0907783583,-0.1019510776,-0.0252721105,-0.0426658839,-0.1361756176,-0.2871514559,-0.0318492763,-0.4308846891,0.0729355887,0.1623313129,0.2992659807,0.1566134095,-0.0742487907,-0.0494469069,0.1143832132,0.3183398247,-0.1257064044,-0.3194031715,-0.405916661,-0.4125874341,0.0458473675,0.1671956629,0.555062592,-0.2708597481,-0.2594034672,0.2183424979,0.0556486286,0.4326915443,-0.3497639894,-0.0037947437,0.1397782266,0.3352205157,-0.2780133188,-0.0910181105,-0.3201158047,-0.2889475822,0.2534392178,-0.0221099649,-0.0810841843,-0.0848103017,0.0755050629,0.0781073496,-0.0023477995,-0.0481448509,-0.0511059426,0.0388418064,0.0123623246,0.1542546749,0.1453919858,-0.1345046908,-0.5091332197,-0.3764441311,-0.0203373488,0.197188735,0.5483554006,0.0524414107,0.3182862401,-0.1757279932,0.25930655,0.2130944431,-0.3146297634,0.2446758151,0.1653646082,0.1780990511,-0.0280466471,0.1700089127,-0.1276296973,0.4133140445,0.3716487288,-0.4411073625,-0.0122789433,-0.3978454173,0.1657234579,-0.2691042423,0.3032995164,0.4508009553,-0.1429353505,-0.1288248301,-0.4156200588,0.3972937763,0.0255686529,-0.2025948465,0.4813393652,0.1077151895,-0.2843468487,0.2128332257,-0.1285715252,0.8244360089,0.2676287889,0.3113027811,0.3640472591,0.1112722233,0.1022783965,-0.4576124549,0.0514674373,-0.0159365274,0.1021314114,-0.1459210366,0.0862596482,-0.1860912293,0.3048148155,-0.3261877596,-0.073388651,0.1394101828,0.2140429318,0.2349338233,-0.2076794058,-0.219132632,0.0628909096,0.1287455559,0.0310560446,0.0454352461,0.2764178514,0.2411041856,-0.141391173,-0.0235028788,-0.3043451309,-0.660400033,0.0256115347,-0.5773504972,-0.0898341462,0.1136530787,-0.3812579215,0.1459791064,0.3813396096,0.1606675088,0.3877428174,-0.4796431065,-0.0438009612,0.3158276379,0.1177165657,-0.0222175177,-0.1181874424,0.3234867454,-0.074841246,-0.1256530583,-0.0707839131,0.0170849375,0.1853922606,-0.1924300343,0.010592225,-0.1102366969,-0.325163871,-0.0179609898,0.1750960201,0.1333635449,-0.2425092608,0.0644174665,-0.1092584059,-0.1126320586,0.3161510825,0.2585337162,-0.5235105157,-0.0768600926,0.5095415115,-0.05945234,0.1247373223,0.6273759007,-0.2404686511,-0.0542512424,-0.275106281,0.0932660177,0.2636379004,-0.3139445484,0.1428882778,-0.0181948114,0.2839416862,0.0443141162,-0.0612466894,0.0597400367,-0.033192832,-0.2500084341,-0.2497020364,-0.0879926533,0.1590440422,0.0418185443,0.2457344234,-0.0120161958,0.283083111,0.0409386642,-0.3694509268,-0.1864431202,-0.0472353101,0.0774314255,0.0784023553,-0.0327920094,0.0524554141,0.0643761083,0.0200634375,0.0109805102,0.0574669391,-0.1392928064,-0.0746839494,-0.3100247085,0.1609307975,0.1798962355,0.1112993583,-0.0083414037,0.0556965023,-0.0509433709,-0.0491366871,0.0704786181,0.3807947636,0.0208634771,0.3144367337,0.0940055251,0.3792684674,-0.0854802057,-0.2379428893,-0.0430187918,-0.1058757231,0.1939326227,-0.0524486452,-0.188233614,-0.0853269547,-0.0319134593,0.0517282598,0.0332106687,0.1841089427,-0.0188174192,-0.2987866402,0.2861653864,0.1175718531,0.0132432804,0.3378902376,-0.2234210521,-0.248046279,-0.0910891816,0.0822845176,-0.0294995066,0.0094374726,-0.3606777191,-0.0174245946,-0.0946175754,-0.0077177826,0.0422165394,-0.4409456551,0.3382523358,0.0060723345,0.1259869188,-0.617092371,0.1211182028,0.4069335461,0.0471916907,0.0044105747,-0.3285924792,-0.2809558213,-0.0062797265,0.5289809704,-0.2464639395,0.2679083049,-0.0634808689,-0.0275064819,0.2425554693,-0.2426433712,-0.2594917715,0.2379479855,-0.4665975571,0.1766881943,0.1410185397,0.3673169315,-0.0485259555,0.0456962809,-0.240078032,-0.0296354685,-0.1623637378,-0.1589698941,-0.3425230086,-0.1122073233,-0.1174796894,-0.1609565914,-0.0854225755,0.0221237019,0.0852115899,0.2617628276,-0.2954109311,-0.5197121501,-0.2400555909,-0.1835346073,0.2313261777,-0.1279245317,0.3786039948,0.10281308,-0.0337539278,0.1412706077,0.3286019266,0.5195795298,0.2080286592,-0.1701619327,0.1073838472,-0.2793742716,-0.0176598448,0.015043973,0.6274505258,0.0008645584,0.1270019561,0.0226037018,0.0276397411,0.0152008086,0.0441954024,0.0171362832,-0.0371454172,-0.7086550593,0.4125433564,-0.0367944427,0.2317141145,-0.1660160571,-0.0049939775,-0.2587536573,0.3277559578,0.621121943,0.1853351891,0.0533866957,-0.0270174704,0.0449458137,-0.1548006386,0.1108004078,0.0185913462,0.049389597,-0.1143924668,-0.0499974191,-0.1892929673,0.0780865923,0.2131735981,0.2548357248,-0.026068788,-0.1599316001,0.0669078305,0.0486767702,0.002179692,-0.0890494362,-0.2402725965,0.2845950425,-0.227225706,0.0184556376,0.014658561,0.145174697,0.1081263795,-0.154785648,-0.1130904257,0.2015322894,-0.1334021538,-0.0073894253,-0.1033231467,0.1589587033,0.3646920919,0.1079781055,0.0701356903,0.4246194661,-0.0901377499,0.1119568273,-0.4082079232,-0.1465588659,-0.196915254,0.2526400387,0.0476315953,0.3335848749,-0.022573825,-0.3501974046,0.1585386097,0.2298382223,0.3724767268,-0.0926154405,-0.1263129711,0.0814921409,-0.3730500937,0.2268996835,-0.0261833817,0.2177128643,0.0186660364,0.1487790346,-0.4728176296,-0.960010469,0.6462187767,-0.7156583071,-0.2400015891,-0.1577966362,0.3278788328,0.2929249704,-0.1062858477,-0.5133001208,-0.118125312,0.4647008777,-0.0780134201,-0.1065440103,0.1552323401,0.1537450105,0.2477950007,-0.1198170334,1.0183166265,0.3149241507,-0.161011979,0.2086753994,-0.4965496659]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"I have changed line 288 to `if int(datasets.config.PYARROW_VERSION.split(\".\")[0]) < 3:` just to get around it.","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":15,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nI have changed line 288 to `if int(datasets.config.PYARROW_VERSION.split(\".\")[0]) < 3:` just to get around it.","embeddings":[-0.4137373269,0.0904145688,0.0879315063,0.069035016,0.284868449,0.1465740502,0.1234253496,0.3872671723,-0.1835217625,0.1793182641,0.3690841198,0.3502765,-0.2362233698,0.1078671366,-0.2026903927,0.0031209812,0.0010433139,0.2549073994,0.1760860234,-0.0239745975,-0.1098649725,0.1690122485,-0.2097337544,0.2731696069,-0.2981304824,0.0923148543,0.0032627089,-0.0122110816,-0.1285185814,-0.5752780437,0.2998543382,-0.1617958844,0.2332223952,0.391263485,-0.000127086,0.0779013038,0.3378002942,-0.0440747142,-0.2238939106,-0.4914156199,-0.085651435,0.0122515438,0.3829289079,-0.0937147513,0.0937459767,-0.6206139922,0.0335906781,0.0384339988,-0.1429862529,0.3645019233,0.127537936,0.0344468243,0.1980867237,-0.0754810423,0.2223289162,0.083245784,-0.0282249879,0.5229272842,0.1158733368,-0.1320393234,0.233240068,-0.102022633,0.0575134605,-0.0043903464,0.0711091682,0.1914277673,0.2576049566,-0.1857423782,0.2388329953,0.0887792706,0.5562196374,-0.6069078445,-0.4294553101,0.0136653185,0.174752906,-0.2191064954,0.2185028791,0.2610872686,-0.1244944483,0.1687700599,-0.0923660845,-0.1329844445,-0.0807051584,-0.0779537112,-0.433144331,0.3678042889,0.1932168156,0.1645069569,-0.0400829166,-0.04866606,-0.0215911791,-0.0890382826,0.1259018928,-0.0294876229,-0.2407489419,-0.0118983239,-0.0643955991,0.1943915635,0.2201173902,0.1590977758,0.0649064705,-0.234773919,0.2536158264,0.0735231191,0.0475472286,0.0977316201,-0.0197430961,0.146631524,0.0936663523,-0.0873535872,0.0256926231,-0.0208997894,-0.0806714743,-0.0734729469,0.3739596009,0.0080980845,0.7578920722,-0.3194511235,-0.2312780768,0.1659606099,-0.3343805671,-0.184103936,-0.2247074097,0.2836925983,0.0676930696,0.3056817055,-0.1982494593,0.3802600801,-0.1360655129,-0.1857322007,-0.0950188115,0.1552126408,0.0275490843,-0.1461474597,0.0764700472,-0.2719792128,-0.0986580551,0.3349583149,0.0094830692,0.1213422492,-0.0362757482,-0.0120166596,-0.1315119714,0.4732261598,-0.2890482843,0.2727305889,0.2008104473,-0.4936734736,-0.3582565486,0.2217560709,-0.1869192868,-0.431489408,-0.2999134958,0.1262726486,-0.1483470798,-0.2103191763,-0.0671084151,-0.0523799025,0.1624187529,-0.1384892911,0.0297997743,-0.651668191,0.1586196274,-0.3059960306,0.2577728629,0.161972031,-0.6419220567,0.0579073988,0.1789008826,-0.1889813095,0.1506783217,-0.1405111104,0.0306624472,-0.1734109372,0.074796848,0.0583436862,0.3201945126,-0.2379271239,-0.3039623797,0.1230635121,0.0718067214,0.0206128471,0.2803038359,-0.3171919584,-0.0282199513,0.1652483642,-0.1928093135,0.1018786877,0.0079200668,0.0062634307,-0.1814045608,-0.0036330107,0.4458725452,0.2917095721,0.1499163508,-0.2299220115,-0.0262244828,-0.1855148673,0.1473240703,-0.0819691345,0.0311352387,-0.0489151254,0.3686681986,-0.2390524894,-0.0064461622,-0.2394002527,-0.2902263105,0.2341018915,0.0907700509,0.0002655792,-0.3193365932,-0.219354555,-0.2210743129,0.0846685618,-0.128561303,-0.011503078,-0.0059905695,0.3605172932,0.0959836021,0.0846642852,-0.3152723312,0.1812611818,0.0422432944,0.1910732836,-0.0209711008,0.4087104499,-0.4075936377,-0.3384674788,0.0850214958,0.0640254617,0.1028376147,0.002861667,0.0299142897,0.3328282237,0.0242639706,0.0389278978,0.0299662594,0.2096638978,-0.0284052063,-0.2111189663,-0.1741658747,0.3066881597,-0.0031621177,0.3704423904,0.0565479696,0.4024959803,0.5626875758,0.3306960762,-0.0120106991,0.0881720185,-0.1884381175,-0.0055437423,0.1211171523,-0.1257602423,0.0624467172,-0.1393158436,0.1693660468,-0.0199615061,-0.5397575498,0.0718814433,0.3745397031,-0.0722794607,0.1365482211,0.1192966476,-0.22845155,0.0563648492,0.0482595488,0.0757280812,0.249298811,0.1221992299,-0.1398308128,0.1419418901,-0.3319807649,0.1508725435,0.1715394557,0.053618446,0.1923853457,0.1373240501,0.2172163874,0.0542263351,-0.0518643856,-0.4222477078,0.1137933508,0.4095291495,-0.4994865358,0.2577461302,-0.2975929677,0.2889801562,-0.0827534124,-0.1969304681,-0.4268323183,-0.2678608894,0.0104828775,0.0935454518,0.1283202618,0.1767116934,-0.3569593728,-0.1166164875,0.0463055074,-0.5190931559,-0.252766639,-0.2070385814,-0.4546803832,0.0009253586,0.2450136989,-0.2954372466,0.0536534749,-0.2570302188,-0.2196664661,-0.2246728688,-0.1574143022,0.0125030046,-0.269426465,0.3128561378,0.2743914127,0.0396343656,-0.0582421124,-0.0491061211,0.3296924531,-0.3633875251,-0.2326972187,0.1670558602,-0.0441559106,-0.018140588,-0.0896702781,-0.5271107554,-0.2856226861,-0.174629271,-0.0225634612,0.0724860951,0.1668928266,-0.0621382482,0.2801332772,-0.0787437782,-0.1362159103,0.0422251076,0.0579393916,0.0014274698,0.2768362463,0.0047421837,-0.2914088666,0.010116118,-0.0066247168,0.1385644227,0.0028628155,-0.2945348024,-0.1095552146,0.0360653736,0.5227450132,-0.2845128775,0.0628193095,0.3160390556,0.2878689766,-0.0280942582,0.0421851873,-0.1316174716,-0.1662545651,0.1143491268,0.1350112706,0.1534068286,0.1728003323,-0.0685666278,0.6282868385,-0.0858015046,-0.2661091983,0.19313097,-0.3469312489,0.3226404786,0.0537074953,-0.1815225929,-0.0795435756,0.02120829,0.000880775,0.265568316,-0.202450186,-0.1276522726,-0.1136756316,-0.21760571,-0.2410297394,-0.0568175204,0.0081763715,0.0084404759,0.1998902708,-0.1092223823,0.0509707965,-0.0197257306,-0.0720054805,0.1099524423,0.3429470062,-0.2275398523,0.0822050944,-0.1880298704,-0.2194179147,-0.1872704327,0.4171535075,0.0073181456,0.4000034034,0.0758238211,-0.203572914,0.0676551983,-0.2018064111,0.5138761997,-0.1867123246,0.1210603341,0.2132676095,-0.1637834311,-0.4432633817,-0.4106668234,-0.5143594146,0.1034789085,-0.0334381461,0.163722381,-0.2714032829,-0.1977230757,0.2114056796,0.195720017,0.081494987,-0.0392780565,-0.461995393,-0.0728343725,-0.2166758329,0.2309352309,-0.1200748757,0.3576822281,-0.2043959945,0.5000504851,-0.1328083277,-0.1695317477,0.2581602931,0.2766701877,0.3142804205,-0.3184945583,0.0112793241,-0.2040047497,0.1823281199,0.537861824,0.5164781213,0.3591929376,-0.2701449394,0.1650438756,0.2548321784,0.5881131291,0.0124148075,-0.4793679416,-0.1435153484,-0.2246030867,0.0869210511,-0.055749163,0.0208998341,0.5394140482,-0.1436982155,-0.5579455495,-0.0006600026,0.583478868,0.1682926714,-0.0880463868,0.0389920734,0.4312092662,-0.1566647142,0.3882396221,-0.1652097702,0.7541747689,-0.0415149927,0.2032355219,0.1119088903,-0.3435962796,0.1856690794,0.0110415602,0.0853404105,-0.5221113563,-0.3559496403,-0.1561163366,-0.2469584793,0.1788385957,0.0282494351,-0.239993453,0.3036089838,0.0095087113,0.3756101727,0.3913848698,0.3188383281,-0.1210615188,0.3092973828,-0.2204645425,-0.0153665738,-0.0465542451,0.0767170414,0.1382596344,0.0722189397,-0.2499359548,-0.4431240261,-0.3027442992,0.0162141509,-0.2184962183,-0.1166363657,0.4740677774,-0.5513356328,0.1450444162,0.0259818938,-0.1213886365,0.0205297451,-0.1986920089,-0.0232671183,-0.132011652,0.1351816356,0.1822996736,0.0830752775,0.4477312863,0.0146222953,-0.3590239882,0.276563555,-0.0613866225,0.0766119584,0.0154431676,-0.0289588962,0.2499645054,-0.3149304092,0.1435465068,-0.5809526443,0.1648852527,-0.1438597441,0.0154867964,-0.1039261594,0.0030873038,0.0458917692,0.1371153742,-0.3003840148,-0.2312101871,0.2688654065,-0.011931234,0.2826273143,0.4671837091,-0.0964071751,-0.0290105585,-0.082691431,0.0665704608,0.3482677639,-0.3652615249,-0.0533109345,-0.0104720378,-0.0259197932,0.1340213865,0.2309290469,0.2441840917,0.2051969171,0.0225443952,-0.5867090225,-0.0156736895,0.1940641701,-0.1296601743,0.1184900925,0.1128129438,0.2511689067,0.2188224196,-0.1819416583,-0.1868157536,0.0008195406,-0.0286003239,0.1301870495,-0.2333950996,0.183599174,0.1691423208,-0.0048594819,0.1116361246,-0.2133685946,-0.1062259227,-0.0898168311,-0.1932307929,0.2013582289,0.0601094924,0.2008146048,0.1493998915,0.0889126062,0.0152651286,-0.0439368822,0.0690269545,0.1395592093,-0.2084579766,0.3365767598,0.4670518041,-0.1786255389,-0.3895530701,-0.0425577238,-0.2097871155,-0.0377730764,0.1553899646,0.3876922727,-0.1798083931,0.1443675905,-0.3008017838,0.0501429252,-0.2175431103,-0.0649992079,0.1967027634,-0.382879734,-0.2250436544,-0.0442183875,0.4638454616,0.2430250347,-0.2055921704,-0.0171396378,0.1072365046,0.0048530609,-0.4478969276,0.1075395718,0.2346013784,0.1743809432,0.0245360527,0.5318636894,0.4067471027,-0.3246034682,-0.0742826611,0.164279595,0.4959196746,0.2708314657,0.1485649049,0.4434695542,-0.0889714733,0.1561041474,0.4896700978,0.1625347584,-0.0874976143,0.3993495703,-0.1631687433,0.1351977885,0.106490612,-0.1803329736,-0.0160063393,-0.1632703543,0.2863210738,0.017498076,0.3956526518,-0.2600688338,0.2041577548,0.3499060869,0.1165681928,-0.1896251142,-0.092125386,0.0162337255,-0.0093539674,-0.0449644402,0.135077551,-0.2281519026,-0.2638019025,-0.0954566151,-0.1304732263,-0.2548211217,0.7299135923,0.2903822362,-0.1669585854,-0.2881604731,0.1720102429,0.1545021385,0.0507362522,-0.2968220413,0.3778815866,0.1143395901,-0.074858211,0.2609856129,0.3107405007,0.8127143383,0.2363016456,-0.1687476635,-0.2336084694,-0.2092974782,0.0317804553,0.291803956,-0.0576246791,-0.028149046,0.0891762078,0.2830816209,0.0896404311,0.0992728993,-0.0181849543,0.0861287341,-0.0633634105,-0.3945927024,0.320821017,0.1960968226,-0.1580650061,-0.0013216215,0.2794041336,-0.2914029956,-0.0390996598,0.1896317601,0.5572887659,-0.1122135371,-0.1232099086,-0.0135795595,0.2728503048,0.3097299635,0.5396063328,0.2115000337,-0.2938640416,-0.156917423,-0.4964832366,0.3750751913,-0.197897613,-0.1750097573,0.2619361579,0.1028776541,0.0685790107,0.170511499,0.3704152703,0.3715750575,-0.3486966193,0.1317123473,-0.3753184974,-0.2713752091,-0.0102243423,-0.1323414594,-0.0787532032,-0.0523299165,0.1964601576,-0.6846225858,0.0174367577,-0.1400099099,-0.2763794065,0.1993915886,-0.0236322135,0.4791323841,-0.0159135181,0.6772804856,0.0820816085,-0.15606305,-0.21632348,-0.0316805914,-0.4409960806,0.5421468019,-0.2116658688,0.2731279433,-0.1207861826,-0.3032947481,-0.2011571079,0.2629315555,0.0276308414,0.0753050745,0.06213497,0.1530801654,0.0903730989,-0.0284187868,-0.0078368885,-0.0506124459,-0.032961823,0.2801569998,-0.5794116259,-0.7179991007,0.5333272219,-0.2753734291,-0.3065237999,-0.3928979933,0.2762829959,-0.0685962513,-0.0433019735,-0.5575699806,0.0828975588,0.4288054407,0.0711633116,-0.2072430998,-0.138777554,-0.0423490182,0.3275108933,-0.1119783968,0.5136018395,-0.3148345053,-0.1485731006,-0.225925386,-0.2605141401]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"Hi @bwang482,\r\n\r\nI'm sorry but I'm not able to reproduce your bug.\r\n\r\nPlease note that in our current master branch, we made a commit (d03223d4d64b89e76b48b00602aba5aa2f817f1e) that simultaneously modified:\r\n- test_dataset_common.py: https:\/\/github.com\/huggingface\/datasets\/commit\/d03223d4d64b89e76b48b00602aba5aa2f817f1e#diff-a1bc225bd9a5bade373d1f140e24d09cbbdc97971c2f73bb627daaa803ada002L289 that introduces the usage of `datasets.config.PYARROW_VERSION.major`\r\n- but also changed config.py: https:\/\/github.com\/huggingface\/datasets\/commit\/d03223d4d64b89e76b48b00602aba5aa2f817f1e#diff-e021fcfc41811fb970fab889b8d245e68382bca8208e63eaafc9a396a336f8f2L40, so that `datasets.config.PYARROW_VERSION.major` exists\r\n","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":47,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nHi @bwang482,\r\n\r\nI'm sorry but I'm not able to reproduce your bug.\r\n\r\nPlease note that in our current master branch, we made a commit (d03223d4d64b89e76b48b00602aba5aa2f817f1e) that simultaneously modified:\r\n- test_dataset_common.py: https:\/\/github.com\/huggingface\/datasets\/commit\/d03223d4d64b89e76b48b00602aba5aa2f817f1e#diff-a1bc225bd9a5bade373d1f140e24d09cbbdc97971c2f73bb627daaa803ada002L289 that introduces the usage of `datasets.config.PYARROW_VERSION.major`\r\n- but also changed config.py: https:\/\/github.com\/huggingface\/datasets\/commit\/d03223d4d64b89e76b48b00602aba5aa2f817f1e#diff-e021fcfc41811fb970fab889b8d245e68382bca8208e63eaafc9a396a336f8f2L40, so that `datasets.config.PYARROW_VERSION.major` exists\r\n","embeddings":[-0.3742081225,0.1478579193,0.0884070471,0.0458860211,0.2804539502,0.1281492561,0.1531104892,0.3776514232,-0.2427559048,0.2011216879,0.3052013814,0.4177031815,-0.1865011901,0.1149607599,-0.1479020417,0.0099561084,0.0530921482,0.2520797253,0.1291816086,0.0060467068,-0.1062645391,0.1722441018,-0.1942494661,0.2815794349,-0.2942166626,0.1177279726,-0.0419812873,-0.0479392149,-0.1290827245,-0.5930629969,0.3376209438,-0.1589786112,0.2437154949,0.3399909437,-0.0001267467,0.1114126891,0.4063085616,-0.0081883511,-0.2549678385,-0.5021974444,-0.1217937693,-0.0536484718,0.3664952815,-0.0817655995,0.102711834,-0.7000570297,0.013414585,0.049935773,-0.1659477353,0.329718858,0.1360016465,0.0637335405,0.2616915703,-0.097520113,0.2251864374,0.0621270798,-0.0079679107,0.4649414122,0.0723097548,-0.1101893187,0.2356003225,-0.0848251805,0.0687128976,-0.0513685308,0.0992346108,0.1825542748,0.2629990876,-0.1627095938,0.1989027411,0.0318586826,0.5882601738,-0.626342833,-0.4298146665,-0.0176614802,0.220206961,-0.2607834935,0.2262852788,0.2429365963,-0.10094136,0.1628725231,-0.0570024997,-0.1206212416,-0.063535206,-0.0164811891,-0.4222541451,0.3365793526,0.1881286055,0.140308246,-0.1132394746,-0.009329834,0.0025072652,-0.1097331718,0.10199292,-0.078015849,-0.2746736109,-0.0144315083,-0.0775343105,0.1633089334,0.1757595986,0.208408609,0.0030351472,-0.2792110145,0.2704052031,0.0860595629,0.1035962552,0.1269080788,-0.0204388909,0.1562927812,0.1005842909,-0.1070765704,0.0690039098,-0.0031423518,-0.0581314154,-0.0568243526,0.3815536797,0.0283413101,0.742669642,-0.2355119437,-0.2066710442,0.1739302725,-0.3724194169,-0.179325074,-0.2660050988,0.2187235057,0.0305577051,0.3607175052,-0.1849215776,0.3320734203,-0.0956049785,-0.2356070876,-0.113595061,0.1457725167,0.0022741908,-0.1549615115,0.0268715788,-0.3101926744,-0.0878965631,0.3905238211,0.0014736976,0.1260318905,-0.1415989846,-0.0541735664,-0.1536239386,0.4754760563,-0.2805734873,0.2540762722,0.1305656582,-0.4681425095,-0.3429700136,0.2116095126,-0.2033770382,-0.4039486051,-0.4072811902,0.1281087101,-0.1272283196,-0.1809758395,-0.1497177035,-0.0052619977,0.2306341529,-0.1509694159,-0.0146371135,-0.6668522954,0.1032236964,-0.319904536,0.2467015982,0.1543766707,-0.6335021853,0.0087504741,0.1704790741,-0.1900213808,0.165815562,-0.1197979152,0.005515072,-0.1398759037,0.0290386174,0.0476634167,0.3468930423,-0.1955857724,-0.3685946465,0.0728684291,0.058083415,-0.0002380217,0.2594134212,-0.3139027059,0.0474226773,0.1808189899,-0.1679071486,0.0811104253,0.001909597,0.0160693731,-0.1740115881,-0.0421568789,0.3828534484,0.2783960104,0.1499212235,-0.1601757854,0.0293607879,-0.2005846798,0.1286726147,-0.0868610367,0.0851052552,0.0264915284,0.4046365619,-0.1430395395,0.033946842,-0.2175856382,-0.3197008371,0.1867123246,0.0698746145,0.0265328661,-0.3560784161,-0.1848592609,-0.1795900613,0.1088649258,-0.1084898785,-0.0283278916,0.0058021918,0.3832255006,0.0711459517,0.1115474254,-0.2974824309,0.2475375384,0.0242408346,0.1781353056,-0.0212112162,0.4727578163,-0.4082576334,-0.3153122663,0.0902254656,0.0457065813,0.102861464,0.0391666442,-0.0002053427,0.2879419625,-0.0028909601,0.0584653802,-0.0276767816,0.2250142992,0.0138650322,-0.1725012511,-0.192661047,0.2565892935,0.0011452289,0.4009828568,0.1184246242,0.3715967536,0.5518648624,0.3002988398,-0.0234602094,0.09365049,-0.1508531719,-0.0168528631,0.1085570678,-0.0839205682,0.0393525958,-0.1069295108,0.2100666016,-0.0077000619,-0.5035149455,0.0840201378,0.3833167255,-0.0595607199,0.1452025324,0.1436303854,-0.2609054446,0.0484301373,0.0339332297,0.1345786303,0.2741984427,0.1518546194,-0.1246563643,0.1309271455,-0.3188162744,0.164799884,0.1828395575,0.1089477018,0.1759979427,0.1831200421,0.2829288542,0.086618349,-0.0567072183,-0.3240150213,0.1387683749,0.3523797393,-0.4344475269,0.2530074716,-0.3444136083,0.250282228,-0.1669087112,-0.1862104386,-0.4147616029,-0.2429396659,-0.0304380916,0.0963686109,0.1005030423,0.2397005558,-0.3122081459,-0.1296312362,0.0462817326,-0.3988128304,-0.2186687887,-0.1758570671,-0.4486293495,-0.019090917,0.2013365775,-0.3951259255,0.039651487,-0.2199624628,-0.2137857676,-0.3060598373,-0.1851968169,0.0833885372,-0.2393283546,0.2992102802,0.2597303987,-0.038242083,-0.0473477803,-0.0988906175,0.3235451281,-0.3321583569,-0.2847580612,0.1006459966,-0.0942858085,0.0356610715,-0.0603935421,-0.4970048368,-0.3575113714,-0.1462482959,0.0124236261,0.041099906,0.1555687338,-0.0624720603,0.2309518009,-0.1249248758,-0.1489116102,-0.0105736004,-0.0276867412,0.0039672754,0.2373251915,0.0062804506,-0.2969486415,-0.0511556789,0.0064248098,0.0896267369,0.0367034785,-0.3144406974,-0.1439506412,0.0547291487,0.5389251113,-0.3207369447,0.0192000363,0.2970771492,0.2694778144,-0.0367961302,0.0630886257,-0.1787241399,-0.1892798841,0.1446789801,0.1415685862,0.1744132191,0.2410916984,-0.0849480182,0.5790010095,-0.0703439489,-0.298630327,0.2510004938,-0.29209131,0.3411943913,0.0630541816,-0.1798293442,-0.0886980519,-0.0483298227,0.0156402048,0.2517925501,-0.2184791267,-0.1496745795,-0.1239425465,-0.2539528012,-0.221225664,-0.042609781,-0.0612108223,0.0359983332,0.2440867573,-0.0551080666,0.0298614856,-0.0403753035,-0.054597225,0.1456445307,0.3016491532,-0.2157077044,0.1017485484,-0.1830856204,-0.152707696,-0.1803331375,0.3837957382,0.0439700186,0.4377391338,0.1552539915,-0.1999727935,0.0833069757,-0.1641715169,0.5928218365,-0.1187062114,0.1239395142,0.1585399061,-0.1371153593,-0.3955423534,-0.4143469036,-0.4618233144,0.083942309,0.00496139,0.1973275989,-0.2952902317,-0.1378474683,0.2305309922,0.1844815016,0.0506920479,-0.0268239118,-0.4164669812,-0.0658165142,-0.16919218,0.2111219019,-0.126962468,0.3692860305,-0.1657081842,0.4781220555,-0.1405241489,-0.1776957959,0.1654678583,0.2207294255,0.338403821,-0.3253705204,-0.0032508457,-0.1957048923,0.2242892385,0.6268402338,0.5107086897,0.3266793787,-0.3425282538,0.2157906592,0.2769287527,0.5334735513,-0.019866595,-0.482755363,-0.0958601087,-0.2471357435,0.0602886491,-0.0659033507,-0.032102935,0.5593261719,-0.1096102819,-0.5021070242,-0.0202025585,0.5698218346,0.1894673705,-0.0793709159,0.0957337916,0.3778422773,-0.1557371318,0.3602989912,-0.1334694028,0.7824115753,0.0175452381,0.2209131867,0.1155825779,-0.4412602186,0.2862614989,-0.0613791905,0.0911196545,-0.5176571608,-0.3366225064,-0.1464529783,-0.2395018488,0.1303436756,0.0145639181,-0.206828028,0.2747061551,0.047589004,0.4116238654,0.3609587848,0.3171049058,-0.0719520301,0.2720511854,-0.252182126,-0.0217562318,-0.0593513064,0.0917054042,0.1281195432,0.0586952567,-0.2239385247,-0.4248760045,-0.3383872807,0.0584696122,-0.2342780083,-0.0421386547,0.4508948922,-0.5700247288,0.118259497,0.0426155031,-0.1173641682,-0.0329631642,-0.1581275463,-0.0164402071,-0.0627333745,0.1840555221,0.1415057629,0.0766142309,0.4222027957,-0.011285712,-0.3214787543,0.2916197777,-0.0392378792,0.0190975238,-0.0512054153,-0.0198285989,0.1973825842,-0.2740911245,0.1948555112,-0.5745176673,0.2753389776,-0.2236594856,0.018647559,-0.1024588272,0.0007495828,-0.0163126327,0.1178036332,-0.2898898125,-0.2125737667,0.3666992486,0.0528494865,0.282312274,0.5269957185,-0.0676542446,-0.0817728192,-0.0454060286,0.0528813414,0.2541744709,-0.3957654536,-0.0471700318,-0.0543740988,0.03274839,0.0790852234,0.1984802634,0.2237325758,0.2555364966,0.0201187208,-0.5458897352,-0.0552603416,0.174301967,-0.1844223738,0.0464428887,0.202157706,0.2829450071,0.1664931476,-0.200484857,-0.1850072443,-0.0044213291,-0.0431108475,0.139864549,-0.1864715964,0.1530421674,0.1342078894,-0.0325627141,0.1182299331,-0.21702604,-0.0706173331,-0.074098587,-0.2204393148,0.1982030272,0.0624699034,0.239382863,0.1128435358,0.0884267464,-0.0120732607,-0.0264758989,0.1066364795,0.1785605848,-0.2272362262,0.3060315549,0.4929539561,-0.2622906864,-0.3427831531,-0.0386489853,-0.1529734135,-0.1157705635,0.1555685252,0.4499003589,-0.1631075442,0.1811010689,-0.2985671461,0.0400748439,-0.238818258,-0.0395774134,0.180485189,-0.4256383181,-0.2396808267,-0.0433830358,0.4485315979,0.2221607566,-0.2303750366,-0.098058641,0.1173264757,0.0090721501,-0.449053973,0.0356157497,0.264077127,0.1550159156,-0.0410299823,0.4426931143,0.4489092529,-0.2873223126,-0.0974351838,0.1774400622,0.4901801348,0.2559277713,0.1621134728,0.5159202218,-0.0438941792,0.1582222879,0.4991028309,0.2005461901,-0.0998899266,0.4372412562,-0.112453796,0.1261387914,0.1164664105,-0.2075036913,0.0423413739,-0.1860411316,0.2623975575,0.0324441902,0.3752452433,-0.2654273808,0.1894052476,0.3312999606,0.205977574,-0.2006644458,-0.1015894637,0.0663827136,-0.0794084296,-0.0449131653,0.1113160029,-0.2534873486,-0.229363516,-0.0513659008,-0.1425132304,-0.2100434154,0.6946986914,0.2750623524,-0.1903173476,-0.3078861237,0.1221392453,0.1174288988,0.0489128195,-0.2738500237,0.3728533387,0.080597274,-0.0812714919,0.2231171429,0.3539702296,0.8215025067,0.2467615157,-0.0757995397,-0.3166638911,-0.1711066663,0.0428912602,0.2550635338,-0.01291807,0.0328659192,0.0312773809,0.2523137927,0.0757340938,0.0942193642,-0.0254333951,0.1040955782,-0.0186157599,-0.4002721012,0.3342917562,0.1377340406,-0.232922852,0.0334776565,0.3048773706,-0.2463754267,-0.0460598394,0.2661556005,0.5599740148,-0.1200682446,-0.1416714638,-0.0092102,0.2585321069,0.3113091886,0.5630508661,0.1585423648,-0.2242979407,-0.1859928071,-0.4756968617,0.4308848083,-0.2150819898,-0.1064559594,0.2760234177,0.0639475137,0.0704118907,0.1569856107,0.3471615314,0.4209796786,-0.3910065293,0.1344710886,-0.3857069016,-0.2570206225,0.0194006171,-0.1237015724,-0.0705663562,-0.0510134622,0.1715069413,-0.5911445022,0.0238804184,-0.145424813,-0.2321884781,0.172145009,-0.0550273545,0.5031514764,0.007037668,0.6460561752,0.0470305867,-0.1947334409,-0.2284395993,-0.1045869589,-0.4692241848,0.6017167568,-0.2683476508,0.2230907828,-0.1235534325,-0.3353490829,-0.2232298255,0.2529105246,0.0031415622,0.0884141028,0.0043125926,0.1209901646,0.0713862255,-0.0406132862,-0.0173885003,-0.0586990155,0.015012959,0.3013275564,-0.5924042463,-0.6977592707,0.4954080284,-0.3196092844,-0.300034225,-0.4309999943,0.2480504811,0.0159253236,-0.0336424895,-0.5722135901,0.1170051396,0.4677253067,0.0741850808,-0.2211979777,-0.1495643854,-0.077095449,0.3797605634,-0.1158954501,0.4920510352,-0.2867052257,-0.1916608512,-0.2343590558,-0.2670706809]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"Sorted. Thanks!","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":2,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nSorted. Thanks!","embeddings":[-0.3823499978,0.0788534731,0.0835613832,0.0859657899,0.254578352,0.1588205546,0.1258136332,0.378461957,-0.1942313612,0.1517066956,0.3442723751,0.3460409343,-0.2396473587,0.1160570979,-0.1717098057,-0.0130585758,0.0182114933,0.2685298622,0.1562249959,-0.0352682173,-0.0831688866,0.1723991781,-0.221207574,0.2769206166,-0.290784806,0.1055777818,-0.034875419,-0.0441389605,-0.1249132976,-0.5678355694,0.3373684287,-0.1683481783,0.259375453,0.3960514963,-0.0001278901,0.0832418501,0.355887711,-0.0235939845,-0.2130245715,-0.4989036024,-0.1226171777,-0.0021165852,0.3659736514,-0.0908794627,0.0975263193,-0.645819068,0.0478101447,0.0595673434,-0.1573934406,0.3680754006,0.1251091063,0.040394716,0.2063285261,-0.0739030614,0.2163523734,0.092064105,-0.0217345729,0.534501493,0.1071522757,-0.1533436328,0.2649391592,-0.1108957157,0.0430092327,-0.0005163433,0.0668540001,0.1680730581,0.2481788546,-0.173890233,0.2404453605,0.0941150188,0.6074262857,-0.5957273245,-0.4052781761,0.0457946397,0.190304175,-0.2164417058,0.2071692944,0.2712188065,-0.1031831801,0.1470091492,-0.0970233083,-0.1404293329,-0.0962809622,-0.0484454148,-0.4191841185,0.3295250833,0.1847844124,0.1627088934,-0.0797763839,-0.0509254895,0.0048299716,-0.1046580449,0.1460140944,-0.0359727815,-0.2669477761,-0.0234076288,-0.0792864859,0.217488572,0.217696771,0.1644843668,0.0548282899,-0.2486527711,0.2624913454,0.0772697702,0.0593933463,0.1027095169,0.0099531254,0.1515246481,0.0868951976,-0.0923113823,0.0064102691,0.0057760188,-0.0834435672,-0.0731900558,0.4025044441,-0.0256594121,0.7480243444,-0.2999703586,-0.2222675681,0.1814114749,-0.3430295587,-0.2135788947,-0.2431992441,0.2692058682,0.0672401413,0.310472101,-0.2242441922,0.3650549948,-0.1089137867,-0.199377805,-0.0976025686,0.1639591455,0.0397986546,-0.1415777802,0.0776998177,-0.2863925397,-0.0622021332,0.3482494056,-0.0095234104,0.1233666539,-0.0444290265,-0.0483978391,-0.1564430743,0.4665051401,-0.2961569726,0.2662153244,0.1866068244,-0.4694723487,-0.3469584286,0.228558898,-0.2082118392,-0.4199297726,-0.3010490537,0.1158056408,-0.1309493929,-0.2246068269,-0.1018703878,-0.0314013213,0.1599574685,-0.1390951127,0.0193729699,-0.6764339805,0.1813001037,-0.3093072474,0.2377792299,0.1493155062,-0.64209795,0.0522496849,0.1860803813,-0.1793422401,0.1692923605,-0.1044029519,0.0251485948,-0.1446892172,0.074836567,0.0642969012,0.3613587618,-0.2439746708,-0.3121292293,0.0909591764,0.1042780653,0.0162541959,0.2491032332,-0.3086020947,-0.0181004889,0.1892558783,-0.1845408976,0.0676994994,0.013640088,0.0026977183,-0.1708292961,0.0054476419,0.438311249,0.2977380157,0.1350409687,-0.2180688083,-0.0289862547,-0.2037833929,0.1285439283,-0.0818165392,0.0136220716,-0.0355041698,0.407566458,-0.211303249,-0.0053179162,-0.21974428,-0.2898447216,0.2070418745,0.0864594579,0.017564917,-0.310914129,-0.1828339547,-0.2235690951,0.092757836,-0.1389970034,-0.0187330078,-0.008984318,0.3708506823,0.0991881639,0.0983951762,-0.3148404062,0.1957128495,0.0238641016,0.1941994578,0.0052944184,0.4397611022,-0.4184099436,-0.3409918547,0.0828048363,0.0324428193,0.1147675738,-0.0007691359,0.0294004232,0.3283750117,0.0075974376,0.0259973463,0.0376454331,0.2178497612,-0.0152834198,-0.2490369678,-0.1754437536,0.2634848952,-0.0124529032,0.3771674037,0.0714703649,0.367174536,0.5579252839,0.311139673,-0.0011530713,0.081929028,-0.1928504854,0.0089487033,0.1333372742,-0.0902685896,0.0926108137,-0.1520457119,0.1808014959,-0.0404996648,-0.5255686641,0.0326813981,0.378573209,-0.0664351285,0.1505247355,0.1364178061,-0.2508790195,0.0396844186,0.0591744334,0.0680330917,0.2792570293,0.1244287118,-0.1432555616,0.1382844001,-0.3376169503,0.1645191908,0.1703767031,0.0325545408,0.1882993728,0.1368130594,0.2204149663,0.053874325,-0.0650231689,-0.3855646849,0.1114446223,0.4056575298,-0.4613749981,0.2288228422,-0.2899155617,0.3022912443,-0.0999656469,-0.1894143373,-0.4212338924,-0.2496228665,-0.0001924169,0.090336822,0.1711301655,0.1835670471,-0.3250343204,-0.1124665737,0.0609607361,-0.4758637249,-0.2720769644,-0.1949386746,-0.4537057579,-0.0113761248,0.2226081789,-0.3075243831,0.0475775488,-0.2794205546,-0.222134456,-0.2541949451,-0.156630829,0.0200997163,-0.2725733817,0.3184693456,0.2527645528,0.0669448599,-0.0692736134,-0.0613622926,0.3347893357,-0.3718841076,-0.2399427444,0.1490340978,-0.0621652342,-0.0054474128,-0.0345259421,-0.5232025981,-0.3153493404,-0.1216146052,0.0009835535,0.0737264007,0.1543922126,-0.0844632015,0.2646680176,-0.0991962701,-0.1470929235,0.0729259551,0.0268436,-0.0291636419,0.2536785007,0.0024928947,-0.2990637422,-0.015819408,0.006221428,0.1566198468,0.0125369066,-0.2939115465,-0.1314835399,0.0421486981,0.5094447136,-0.2864129841,0.0493873321,0.3148921132,0.2830879986,-0.0151688168,0.0654160902,-0.1438310146,-0.1625919193,0.1416301429,0.1366788298,0.1592234224,0.1884554476,-0.0858525932,0.6183635592,-0.066043742,-0.27083534,0.2227285802,-0.349478066,0.3271769881,0.0535552427,-0.1810140163,-0.0944271237,-0.0012391821,0.0283627696,0.2522964776,-0.2121816278,-0.1313127875,-0.1222149208,-0.2319981158,-0.2465988696,-0.0176805239,-0.0523007289,0.0142141161,0.2401451319,-0.0935465693,0.034567073,-0.0206153709,-0.0605601259,0.1290478557,0.3503609002,-0.2477122843,0.0702728033,-0.2205438316,-0.2247415632,-0.173276186,0.3985607326,-0.0039343834,0.4307931662,0.0841581747,-0.1833612025,0.0811553076,-0.1943857968,0.5388733149,-0.1934172809,0.1095087007,0.1825060248,-0.1371613145,-0.4131250083,-0.4019485116,-0.5120395422,0.0875898749,-0.0420206934,0.1591536552,-0.2654743791,-0.1794775724,0.2284601927,0.1847786903,0.0931383222,-0.0519520231,-0.4445238411,-0.057030078,-0.1848578751,0.2046760619,-0.1663069725,0.3586477637,-0.1895323992,0.4917454422,-0.1368259937,-0.1512624323,0.2483527064,0.287710011,0.3000683784,-0.3307991326,-0.0130946636,-0.1949541271,0.1695131212,0.5847437978,0.4820328057,0.3315340579,-0.294988364,0.1528007686,0.2649532855,0.5828864574,-0.0272528008,-0.483433634,-0.1476513296,-0.2480645627,0.0608517006,-0.0455113314,0.0324646272,0.5878044367,-0.1079147011,-0.5646884441,-0.0115431584,0.5832077861,0.2118635923,-0.0714421794,0.0664260015,0.4306585193,-0.1481029093,0.3573056757,-0.1868436038,0.7492056489,-0.0570857488,0.2151276916,0.1019691601,-0.3711522222,0.1966659725,-0.0153630348,0.0793213546,-0.5118100643,-0.3569800556,-0.1609403491,-0.2538842559,0.1990831494,0.0451388434,-0.221278891,0.3177910745,0.0354344957,0.3949761987,0.3592294157,0.3050839007,-0.0906670541,0.3530260026,-0.2177924514,-0.0141585823,-0.0614609346,0.0901510194,0.1194941476,0.0920847505,-0.2436117381,-0.4295100272,-0.305162847,0.0007013087,-0.2089609057,-0.0932991058,0.4825862348,-0.564635396,0.1532411426,0.0374579653,-0.1119042113,0.0024960313,-0.2098170668,-0.0293328371,-0.1158581823,0.1289699525,0.1722138822,0.0898154378,0.4245472848,-0.0003973609,-0.362603873,0.304105103,-0.0489400961,0.0594808012,-0.003578915,-0.0241917279,0.2139300406,-0.3155358732,0.132243067,-0.5739024878,0.1775269806,-0.1474528909,0.0127399089,-0.1176820323,-0.0013163961,0.0483027846,0.1547810137,-0.283978492,-0.2273022979,0.2939545214,-0.0005660221,0.2803093195,0.4978891015,-0.115971446,-0.0311297104,-0.0739877224,0.0684594959,0.3330630958,-0.3416400552,-0.0628984123,-0.0199738983,0.0104450407,0.1303909272,0.2286181152,0.2601722181,0.2260191292,0.0391300023,-0.5462092161,-0.0374827012,0.1696475744,-0.1299935579,0.1054470316,0.1329135299,0.2669435143,0.2379645258,-0.1843246967,-0.1761202514,0.0102633117,-0.0304712392,0.1358926743,-0.2163457274,0.158505246,0.138336271,0.0027873158,0.1136001423,-0.2447290123,-0.1037489846,-0.0822046772,-0.1622231752,0.2020552754,0.0761264116,0.1932974756,0.1445910037,0.0521461889,0.0121984174,-0.031037217,0.0573093072,0.117708832,-0.2250270993,0.3234240115,0.4698260427,-0.2181081176,-0.3864577413,-0.048631914,-0.2083722055,-0.0919483081,0.1622691602,0.3923228979,-0.1639600098,0.1349995434,-0.3085337281,0.0645157397,-0.2191734761,-0.084187828,0.1953693926,-0.4075606465,-0.2250798941,-0.05221726,0.4547858238,0.2435606569,-0.2076168656,-0.0345069468,0.1239121035,-0.0041082399,-0.4457924366,0.080068253,0.2681975961,0.166590631,0.0092392126,0.4907566011,0.4225642681,-0.3052988052,-0.0838325247,0.1534824818,0.5127706528,0.2449548542,0.1527925283,0.4591118395,-0.0687162206,0.1594157666,0.5113036633,0.1964520961,-0.1306294948,0.3847328126,-0.1721257567,0.1426737756,0.0957871377,-0.1894542873,0.0161806419,-0.1555673629,0.2666930258,0.0196503624,0.3913426101,-0.2815415561,0.1803604662,0.3507416546,0.1539365053,-0.2007782906,-0.1109496057,0.0122991148,-0.0259589273,-0.0553431511,0.127328977,-0.2358577847,-0.2330396473,-0.0735607669,-0.1582722217,-0.2579681575,0.7183961868,0.2728726566,-0.1813760847,-0.3079471588,0.1664167345,0.1576441526,0.0288805664,-0.2608737051,0.3783064485,0.0894762874,-0.0856343582,0.232608512,0.3378421664,0.8134999275,0.2410013527,-0.1378462315,-0.2350551188,-0.1959068179,0.036293149,0.3027204871,-0.0511013158,-0.0022525107,0.0649085566,0.2457795441,0.0829157159,0.1131313965,-0.0042747431,0.080337815,-0.0697649494,-0.4125570953,0.2890695632,0.1622858495,-0.1878182888,0.0006666749,0.294282347,-0.2899018824,-0.0241305344,0.2330348194,0.567232132,-0.1211466789,-0.1184538007,-0.0200917888,0.2552170753,0.2999455333,0.5425080657,0.2283957005,-0.2810824215,-0.1846067607,-0.491732955,0.3696128726,-0.188915655,-0.1681735516,0.2723663449,0.0870510712,0.0668255761,0.1543128341,0.358499825,0.3908868134,-0.3710202575,0.133395642,-0.3752774596,-0.2423381209,-0.0293359756,-0.1051227599,-0.0941129327,-0.0155493794,0.2134482265,-0.6555412412,-0.0000092105,-0.1405493617,-0.242702499,0.1772046834,-0.0279113445,0.4945253432,-0.039074365,0.6881918907,0.0928875506,-0.1293089986,-0.1993017346,-0.0419701189,-0.4117202461,0.5732808709,-0.2209887207,0.2203930467,-0.1318157911,-0.3068338335,-0.1680376679,0.2684181929,0.0365713872,0.0693127438,0.074137643,0.1095281318,0.1249653995,-0.0653188229,-0.0048772795,-0.0606577806,-0.0253131129,0.2815610468,-0.5856456757,-0.6785210967,0.5218515396,-0.2936740816,-0.2915141881,-0.4276632667,0.2554471791,-0.0313607417,-0.0421098396,-0.5466675758,0.0804826915,0.4648813307,0.0448707342,-0.2022099197,-0.1116632447,-0.022803925,0.3597259521,-0.1047497168,0.4848444164,-0.3160455525,-0.1668283343,-0.2473701835,-0.2702198327]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"Reopening this. Although the `test_dataset_common.py` script works fine now.\r\n\r\nHas this got something to do with my pull request not passing `ci\/circleci: run_dataset_script_tests_pyarrow` tests?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/2873","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":25,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nReopening this. Although the `test_dataset_common.py` script works fine now.\r\n\r\nHas this got something to do with my pull request not passing `ci\/circleci: run_dataset_script_tests_pyarrow` tests?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/2873","embeddings":[-0.3413378894,0.2006237954,0.049372945,0.0571891814,0.1107941419,0.0854446366,0.2353403866,0.2816726863,-0.1760852039,0.1386214346,0.438138634,0.3304321766,-0.1294845939,0.297735095,-0.1947026998,0.0704638883,0.0133854281,0.1427540034,0.3433145285,0.0562438518,-0.1254506111,0.239258036,-0.1946723908,0.1283952147,-0.1329981238,0.0372122005,-0.1708889008,0.0408576243,-0.1387818605,-0.4680930376,0.4385558963,0.0902260914,0.1819554418,0.6217026114,-0.0001299516,0.1380508989,0.4102708101,-0.0885429606,-0.2731285393,-0.5470474362,0.0774918571,-0.0126978597,0.3988084793,-0.1040790603,0.0965566337,-0.3847824633,-0.0448813178,0.1092322022,-0.067902863,0.3774314225,0.0892633647,0.2561304271,0.0892846957,0.0130179506,0.2664367259,0.3202229738,-0.1790551394,0.4346646965,0.163472265,-0.1368527114,0.18223162,-0.0818933249,-0.0010008782,0.0381129086,0.058766406,0.0650137067,-0.0054221773,-0.2756882608,0.2095608413,-0.0046765376,0.4738269448,-0.5655617118,-0.4999672771,0.0325702988,0.1310693324,-0.2759967446,0.1989559233,0.1092876121,-0.2475142032,0.1120673195,-0.1768936068,-0.1210495085,-0.0797204599,-0.0597053356,-0.452073127,0.1955433339,0.1415925175,0.1933697611,0.0005672796,-0.0080453893,0.0825758576,-0.1077340245,0.1832358539,-0.0202006698,-0.2383793592,-0.0761251375,0.0723483413,0.3277744949,0.2713033557,0.4195830226,0.0403551459,-0.1306673586,0.174266085,0.0946188346,0.0676531792,0.2015321553,-0.0325329341,0.1100755706,0.26883623,0.0022651241,0.0251369309,-0.0340531245,-0.0820883214,-0.118339479,0.2201674134,0.0307946801,0.8329048157,-0.3100616932,-0.264413327,0.0647196248,-0.4131411612,-0.1456345022,-0.1949632913,0.2120336592,0.1063020974,0.3991495371,-0.054112941,0.3935264349,-0.0928063095,-0.0335153863,-0.1254985034,0.0960216671,0.0741094276,-0.0745831132,0.2243506163,-0.3418166041,-0.0214521512,0.2785852551,0.2317495495,0.1688358784,0.0847869664,0.1707445085,-0.0158238746,0.5882755518,-0.1170846,0.3030574322,0.1808865517,-0.4662673175,-0.3240087628,0.2942477167,-0.3417190313,-0.3119837642,-0.1544470787,0.0395626612,-0.3309824169,-0.2279408276,-0.2230107188,-0.199565202,0.230595544,-0.09554369,0.0377825759,-0.6431552768,0.1101010367,-0.2451968342,0.2660878897,0.2522025704,-0.4532644153,0.0551288687,-0.0058161258,0.0077627655,0.1686531156,0.108184278,-0.0699362084,-0.1368115842,-0.0259974599,0.0800719559,0.4520925283,-0.4171760976,-0.3087326586,0.2504969537,-0.0502103381,0.0525274239,0.2544584572,-0.4099239409,0.0670625865,0.1091564149,-0.1023410782,-0.0559723116,-0.013213655,0.1008469164,-0.1170869917,-0.0518476814,0.4205104113,0.2592461705,0.0945970491,-0.1817267984,-0.0615676045,-0.3520296216,0.2080374062,-0.058173541,-0.00003425,-0.0472278409,0.4154104888,-0.3017407358,0.0376632102,-0.0054553566,-0.2810015082,0.2353983223,0.1397245526,0.1832517534,-0.4007819295,-0.2460963428,-0.2957423627,0.1120926812,-0.1815855503,-0.0250835977,-0.0734229907,0.3676602542,0.0943571031,0.0908512399,-0.2459866107,0.1231615022,-0.0549701601,0.121638678,-0.0157055464,0.36100775,-0.39767465,-0.3114204407,0.1486184299,0.0152983777,0.1552105695,-0.0953354239,0.0492075197,0.1977923065,0.018292509,-0.0615071394,0.0249138251,0.2371177673,0.0974489674,-0.1439389586,-0.1320542544,0.2738496065,-0.0527514368,0.2288922668,0.0089537343,0.4483506978,0.433740288,0.4135373831,-0.0052913735,0.133137241,-0.0789800286,0.0464649163,-0.1026584506,0.0235418081,0.0623203032,-0.1579686403,0.2385576516,-0.0083107622,-0.4083735943,0.1789080948,0.4755972028,0.0045280191,0.0512079373,0.1263181567,-0.0464047939,0.0135155227,0.1427060068,0.1196604073,0.3690716922,0.1315706223,-0.1269986928,0.1280188709,-0.3089667261,-0.000240603,0.0881366134,0.0855403543,-0.0177239105,0.1066989228,0.2934764922,0.0094751427,-0.1714618802,-0.5069288611,0.1386284977,0.3523923159,-0.5544535518,0.2957420945,-0.2599960566,0.2412185669,-0.0563951358,-0.099517554,-0.3512249887,-0.3400330842,-0.0426338091,0.1322477013,0.1855418086,0.1987326741,-0.3156895339,0.085814856,0.0278856121,-0.5474075079,-0.3988513947,-0.2250475734,-0.3432812393,-0.0082413657,0.335031271,-0.3440192044,0.1043744832,-0.3536006212,-0.0615599714,-0.3151578307,-0.3276818395,0.0971032903,-0.3790067136,0.3400279582,0.264110893,0.1390078962,-0.0936909392,-0.0775850788,0.2825623751,-0.3451506793,-0.3498209715,0.0558034815,-0.0393706858,-0.081658788,-0.2297518551,-0.4395770431,-0.1756117046,-0.1256152838,0.1797543317,0.1345760971,0.0894499645,0.1602129638,0.2617307603,-0.1119602174,0.0015942445,-0.080045335,0.0416262671,-0.1853195578,0.311385572,-0.1229889318,-0.2905154526,0.0768143982,0.0901876315,0.2567841113,0.0905639157,-0.3195661008,-0.2088564485,0.0071465177,0.5721033216,-0.1528172642,0.0589576438,0.3950414658,0.1801723242,0.033580374,0.0303174295,-0.2007779479,-0.0994261503,0.1576837599,0.0561942831,0.1956194937,0.2270442694,-0.0645815507,0.7777855992,-0.0912658498,-0.247752741,0.3013986051,-0.3482937813,0.4125493765,0.0924599543,-0.3143632114,-0.1148337424,-0.0043492815,-0.0983640924,0.1719359308,-0.1672710329,-0.196849227,-0.131608218,-0.2043840885,-0.1171986982,-0.0901859552,-0.0458139852,-0.0148275411,0.2799755633,-0.0396879055,0.1678755432,-0.1324225217,-0.0496874787,0.1742268801,0.442769289,-0.0196975451,0.0237190947,-0.2624483407,-0.2940678298,-0.2959463596,0.3724193871,0.061157085,0.4008819461,-0.026681086,-0.1610317528,0.1031048149,-0.3025604188,0.6163403392,-0.2401427329,0.0806937441,0.3421912193,-0.2250520885,-0.4068379104,-0.320012629,-0.4515895844,0.1839895546,-0.0279769711,0.2594494224,-0.2801221311,-0.3214555085,0.2671273351,0.1050608456,0.0664297938,0.0288712382,-0.5246195793,-0.4429920614,-0.2986671627,0.2140108347,0.0544981584,0.2545920312,-0.3563589156,0.4444724619,-0.0712299645,-0.2317398638,0.2484552115,0.2951732576,0.2508275807,-0.1330598593,-0.0640557408,-0.1815859675,0.0097758155,0.477694422,0.5681917667,0.3777179122,-0.2568677068,0.3769102693,0.2031817287,0.5758686662,0.2257894129,-0.4059256017,-0.0527210794,-0.1149656773,-0.0379138701,-0.080989413,0.1220759079,0.5143463016,-0.2152787,-0.3515267074,-0.0673798025,0.3762578964,0.290491879,-0.0491693653,0.1210093647,0.4069837034,-0.1331271678,0.3132099211,-0.1312802881,0.8649080396,0.057380721,0.256983012,0.1712691486,-0.322670877,0.0820893273,-0.003961972,0.0175484009,-0.4571155906,-0.1686368138,-0.1796827912,-0.2954652309,0.2304840833,-0.0226178914,-0.1865593791,0.229722783,0.0601244867,0.4995066226,0.4300904572,0.3966604471,-0.2316233218,0.3145558834,-0.2194425762,-0.0172984209,0.0293021239,0.2286307663,0.124477379,-0.002792815,-0.2746600807,-0.4961460829,-0.37046206,-0.0256939139,-0.2890318036,-0.1104236618,0.3295895755,-0.7482529283,0.2486482412,0.0339220054,0.1098260209,0.0075352951,-0.3014063835,0.0366480127,-0.1004758477,0.1164056063,0.2305586487,-0.0126885399,0.3798055649,0.0022485489,-0.369571954,0.248747468,-0.077589035,-0.1284065992,-0.10458038,0.0652091727,-0.0416565202,-0.3378791511,0.1743742973,-0.5917771459,0.1822234839,-0.0884274393,-0.0206442904,-0.1191925257,0.0958006158,0.0254339762,0.2053674906,-0.2256229073,-0.1250716597,0.2743127644,-0.0698661953,0.1688301414,0.5413680077,-0.0138974981,0.0417850837,-0.0379354209,0.033793889,0.2881394327,-0.4225570858,0.0193725899,-0.0664407313,0.0207807571,-0.0079836641,0.509545207,0.0708176792,0.2001803219,-0.0391862728,-0.4795256853,-0.0093072541,0.2478470504,-0.0300485119,0.1569505185,0.1399680674,0.2862483859,0.2056425959,-0.405495137,-0.1714913249,-0.0400294513,-0.000537709,0.0970597789,-0.1244992316,0.0023443503,0.171981141,-0.0211473666,0.024566818,-0.1681083292,-0.0766649842,-0.0485599563,-0.1229086444,0.2095736861,0.0366380587,0.1580324769,0.1997240484,0.0126829194,0.0098167099,-0.0131300623,0.1601453722,0.0829429477,-0.1392778456,0.093243286,0.1914658844,-0.1461140364,-0.3684512675,0.0260951091,-0.1601081192,0.0838330016,0.1734333187,0.3831002414,-0.1243301556,0.1700462848,-0.1644287854,0.136464268,-0.1596181542,0.013914519,0.2238865793,-0.3504797816,-0.2246349007,-0.0711626559,0.5016793013,0.2746834755,-0.1880669147,0.0494724512,0.0735972151,-0.0657561943,-0.3324206769,0.0811484829,0.3131485581,0.1644701362,-0.0142199947,0.5696216226,0.3624671996,-0.2893332541,0.1590238959,0.1196841523,0.7548341155,0.2659273148,0.0646335185,0.1735058725,-0.1727526039,0.1701314747,0.4858205616,0.0595215857,-0.0590773523,0.1139845327,-0.1660731435,0.1806773692,0.0633857921,-0.1396620274,0.001239076,-0.1661031544,0.3387806416,-0.264516741,0.3434379101,-0.2066052258,0.1852727979,0.4347059727,0.0754559636,-0.1739108562,-0.2398148179,0.1066153198,-0.0229048841,-0.0884501562,-0.1149521023,-0.2485224754,-0.1854368746,-0.0515635498,-0.1260222495,-0.1242799088,0.6650737524,0.1825416833,-0.1802328378,-0.2559813559,-0.0092405602,0.0285580959,0.0777159259,-0.263761878,0.4144214392,0.0703170374,-0.1295516491,0.0619428381,0.3388366401,0.9245889783,0.1384816468,-0.119341433,-0.3575937152,-0.2355310768,0.0099260984,0.138694793,0.0076980577,-0.0125988983,0.1495130807,0.1344967633,0.0862469524,0.071466282,-0.0586302131,0.1186977923,-0.1172768399,-0.4760383666,0.3648820519,-0.0168354753,-0.1769959629,-0.0490625612,0.1928962469,-0.2198527902,0.0186762493,0.1510175467,0.4744874835,-0.1413980275,-0.2653834522,-0.0160904955,0.1495954692,0.3990335464,0.4798596203,0.2936071754,-0.2884002924,-0.1620029062,-0.630964756,0.2499451637,-0.1089177728,-0.0733902678,0.1761762798,0.1923172176,0.1261360198,0.1565837115,0.2528316379,0.1572821587,-0.2526715398,0.0740477443,-0.2817492485,-0.1842305511,-0.1125999242,-0.1680740267,-0.0837305263,-0.0515798405,0.1233399138,-0.6046460867,-0.0450350232,-0.2213482559,-0.3686109781,0.1300189048,-0.2846067846,0.4359494448,-0.0453370288,0.7685061693,0.1256030947,-0.1501159221,-0.3288882077,-0.1370660514,-0.285241276,0.5235645771,-0.3340819478,0.2942072451,-0.0719057769,-0.380186826,-0.2367647588,0.2125569731,0.1473571807,0.015084289,0.139054656,0.18910411,-0.0995963514,-0.0554136485,-0.0435973816,-0.030801123,-0.0259375852,0.2249903977,-0.5183079839,-0.6421110034,0.5877329111,-0.3876187503,-0.2811570168,-0.3539839387,0.2209096998,-0.1894461215,-0.1547939181,-0.4908481538,0.021292137,0.3034403026,0.1360800862,-0.2562732697,0.0768880621,0.0237822514,0.290207088,-0.0882469937,0.499348551,-0.3625539243,-0.0745926052,-0.1757297665,-0.1907047182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2871","title":"datasets.config.PYARROW_VERSION has no attribute 'major'","comments":"Hi @bwang482,\r\n\r\nIf you click on `Details` (on the right of your non passing CI test names: `ci\/circleci: run_dataset_script_tests_pyarrow`), you can have more information about the non-passing tests.\r\n\r\nFor example, for [\"ci\/circleci: run_dataset_script_tests_pyarrow_1\" details](https:\/\/circleci.com\/gh\/huggingface\/datasets\/46324?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link), you can see that the only non-passing test has to do with the dataset card (missing information in the `README.md` file): `test_changed_dataset_card`\r\n```\r\n=========================== short test summary info ============================\r\nFAILED tests\/test_dataset_cards.py::test_changed_dataset_card[swedish_medical_ner]\r\n= 1 failed, 3214 passed, 2874 skipped, 2 xfailed, 1 xpassed, 15 warnings in 175.59s (0:02:55) =\r\n```\r\n\r\nTherefore, your PR non-passing test has nothing to do with this issue.","body":"In the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":95,"text":"datasets.config.PYARROW_VERSION has no attribute 'major'\nIn the test_dataset_common.py script, line 288-289\r\n\r\n```\r\nif datasets.config.PYARROW_VERSION.major < 3:\r\n packaged_datasets = [pd for pd in packaged_datasets if pd[\"dataset_name\"] != \"parquet\"]\r\n```\r\n\r\nwhich throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested this on both datasets.__version_=='1.11.0' and '1.9.0'. I am using Mac OS.\r\n\r\n```\r\nimport datasets\r\ndatasets.config.PYARROW_VERSION.major\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n\/var\/folders\/1f\/0wqmlgp90qjd5mpj53fnjq440000gn\/T\/ipykernel_73361\/2547517336.py in \r\n 1 import datasets\r\n----> 2 datasets.config.PYARROW_VERSION.major\r\n\r\nAttributeError: 'str' object has no attribute 'major'\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: Darwin-20.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.11\r\n- PyArrow version: 4.0.1\r\n\nHi @bwang482,\r\n\r\nIf you click on `Details` (on the right of your non passing CI test names: `ci\/circleci: run_dataset_script_tests_pyarrow`), you can have more information about the non-passing tests.\r\n\r\nFor example, for [\"ci\/circleci: run_dataset_script_tests_pyarrow_1\" details](https:\/\/circleci.com\/gh\/huggingface\/datasets\/46324?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link), you can see that the only non-passing test has to do with the dataset card (missing information in the `README.md` file): `test_changed_dataset_card`\r\n```\r\n=========================== short test summary info ============================\r\nFAILED tests\/test_dataset_cards.py::test_changed_dataset_card[swedish_medical_ner]\r\n= 1 failed, 3214 passed, 2874 skipped, 2 xfailed, 1 xpassed, 15 warnings in 175.59s (0:02:55) =\r\n```\r\n\r\nTherefore, your PR non-passing test has nothing to do with this issue.","embeddings":[-0.4502635896,0.1362894326,0.0727608278,-0.0077378419,0.210440293,0.1455685943,0.1347785145,0.3891866505,-0.2612731457,0.1966487318,0.2423591018,0.4409922659,-0.174534291,0.1955467612,-0.2465572804,0.0201375429,-0.0359608233,0.1991022974,0.1349838227,0.0418360345,-0.1486246139,0.1625906974,-0.2162213922,0.2647956908,-0.1537034959,-0.0562001877,-0.0889388993,-0.0443796068,-0.1304095089,-0.543780148,0.374103874,-0.0596004017,0.1746742874,0.5324847698,-0.0001234938,0.1340290904,0.3579029739,-0.099743627,-0.3201577663,-0.4985058904,0.0303555876,0.0250716098,0.3309290409,-0.1497163624,0.0457028411,-0.5905317068,-0.0560617782,0.011204035,-0.0264583249,0.4921305478,0.1557275653,0.1648605168,0.0755554661,-0.137879923,0.210254997,0.089855887,-0.0760812536,0.4706763327,0.1216390133,-0.0717137381,0.1600072682,-0.1023208573,-0.0217084177,-0.0166188478,-0.0197366606,0.0890951231,0.2701399326,-0.2540642619,0.2758241594,0.0213556439,0.6139847636,-0.6099302769,-0.402813077,0.0584969297,0.1510484666,-0.1977946013,0.2043727338,0.1740199775,-0.2345986068,0.1370747983,-0.1940619051,-0.0381158032,-0.0790201724,-0.1012421995,-0.3796566129,0.2510077655,0.1984764487,0.1712024063,-0.0156713389,-0.1160794795,0.0763115138,-0.1649231166,0.170402199,0.0014240057,-0.2107874453,-0.129683286,-0.0683172196,0.3605410159,0.2446825057,0.2437425703,0.0395206213,-0.2193709165,0.2899466455,0.1493432522,-0.0002065366,0.060830459,0.0282580554,0.1525740027,0.1156841591,-0.0743039697,0.0116414698,0.0125757102,-0.0356712267,-0.1347546428,0.2806890607,-0.0175586436,0.747718215,-0.386092782,-0.3294264078,0.1127082333,-0.3191033304,-0.1121121868,-0.1496730894,0.3698425591,0.0801952556,0.3064732254,-0.2025549859,0.3027328849,-0.148769632,-0.1783850491,-0.0953058898,0.0878898203,-0.0219249018,-0.1331822574,0.1687428951,-0.2979690433,-0.0465056896,0.3469100296,0.1025219187,0.1879935563,0.0652616769,-0.0026953628,-0.0544139817,0.4611659646,-0.2218509614,0.2516638935,0.1780836433,-0.5087434649,-0.3634715378,0.2743989229,-0.2164251655,-0.3534741104,-0.2432028204,0.151281327,-0.0920187533,-0.2586004138,-0.0489280932,-0.0613507144,0.1802086085,-0.1575635076,-0.0042943428,-0.6692387462,0.1291497648,-0.3172443509,0.2979747355,0.2355631441,-0.5403274298,0.0511988066,0.1171005592,-0.1538690478,0.2208563685,-0.0480547622,0.0285186488,-0.1684476435,0.008382176,0.0004034062,0.3619309366,-0.3157839477,-0.2808323205,0.2037823498,0.1373576522,-0.0054161474,0.258110404,-0.3559542,0.0676776916,0.1332530528,-0.2449017912,0.0707520768,-0.0634788871,0.0394995026,-0.1724347621,0.0111094378,0.4310903251,0.2942523658,0.0776945651,-0.1869189143,0.0541623756,-0.2536583543,0.1724235415,-0.153491497,0.0145558426,-0.095210731,0.4228506982,-0.3071483076,0.0091015995,-0.0820851251,-0.3720258772,0.214207381,0.1426869631,0.0427836291,-0.3331008852,-0.2545279264,-0.3418053687,0.0929807276,-0.1178567037,-0.0309027005,0.0290229879,0.3834606111,0.0836455449,0.0666082948,-0.1785346568,0.1259467155,0.0109810112,0.1843898743,-0.021914199,0.4052859545,-0.3848957419,-0.3287788928,0.1164525375,0.0147691462,0.2110700905,-0.0747004673,-0.0244301055,0.3569452167,0.0668140128,0.0324969031,0.0375453681,0.2812748551,0.0095077008,-0.2744845748,-0.2255578786,0.2845185697,-0.0563144833,0.335890919,-0.0477804728,0.4245945215,0.4690216482,0.2440936267,0.0434890911,0.1207157895,-0.0876624584,0.0307110269,0.0175956078,-0.1027696207,0.1239546984,-0.0975145176,0.2429556996,0.0073814266,-0.4534864426,0.0181120038,0.3326230049,-0.1036967561,0.079722181,0.1088910624,-0.1348168701,0.0713322908,0.1310128421,0.1286019683,0.2671003044,0.1545218527,-0.0664283633,0.1459353566,-0.3957768083,0.0493688993,0.1474568546,0.046951104,0.0843830109,0.1178493723,0.2030754983,0.0692258701,-0.1964526772,-0.4386540353,0.1690077484,0.3975126445,-0.4556099176,0.2525030673,-0.2593023777,0.2879627645,-0.1762989759,-0.1013129577,-0.2812756002,-0.2871729434,0.0298087746,0.1613865793,0.1280528903,0.2844814956,-0.3949874341,0.1308537275,0.0604371689,-0.5389617682,-0.3587738574,-0.1788505465,-0.4358465075,0.0564847142,0.2930062115,-0.2936736047,0.1870015562,-0.3385720551,-0.1041571423,-0.2780030966,-0.2546269596,0.1000234187,-0.2732539177,0.4224722683,0.2331346571,0.0653007999,-0.0509411618,-0.1394465566,0.2985119224,-0.5006918907,-0.2382101864,0.1075783893,-0.0609142967,-0.0281601492,-0.0992578343,-0.6026321054,-0.2700941563,-0.1767314076,-0.0738294199,0.1012948826,0.1171792299,0.0153186703,0.1794993281,-0.0414384827,-0.1064003855,0.040731471,0.0227799695,-0.0935673937,0.2638745904,-0.1412262917,-0.247062102,0.0030329677,0.0593990013,0.3075082004,0.0263136849,-0.3157670796,-0.1785085499,0.0870808735,0.5713326931,-0.213754639,-0.0088340752,0.2017015219,0.3253875971,-0.0461313352,0.0438692309,-0.1561382264,-0.2257906049,0.0831051841,0.1174168065,0.1173686087,0.2478049397,-0.0001471604,0.7000100017,0.0901702568,-0.2647122443,0.2456341684,-0.290442735,0.3646113276,0.0215920843,-0.2104302496,-0.0293033496,-0.0032851116,-0.0899685696,0.2567013502,-0.1834515184,-0.2540484667,-0.0747877732,-0.1884061694,-0.2296490818,-0.0697398409,-0.0385251641,-0.0507598296,0.2692644894,-0.0709148198,0.1493453532,-0.0700604767,-0.1060432568,0.2032533139,0.3706054389,-0.2543409765,0.0308764242,-0.2697114348,-0.200737372,-0.2348984778,0.4246786833,-0.0129908873,0.4772621691,-0.016963236,-0.1294088811,0.0974079669,-0.2080836296,0.4789678752,-0.2268110663,0.0811510012,0.2893922925,-0.170969218,-0.44634372,-0.3620918989,-0.4987749159,0.0781412125,-0.0376700051,0.2260660231,-0.2343872935,-0.2319649607,0.2269615382,0.1185379326,0.0725107491,0.0037124723,-0.4310705662,-0.2614602745,-0.2430775613,0.3034367263,-0.1024958342,0.3525213897,-0.3325912952,0.4075814784,-0.0961001888,-0.1727211326,0.2449412048,0.3544303477,0.2249419391,-0.2119112909,0.0260769092,-0.1250227988,0.2323424816,0.5753104091,0.5339670777,0.2321098298,-0.2221550047,0.1890183538,0.2378938645,0.6089592576,0.0379683226,-0.4350992441,-0.1311749667,-0.2159736902,0.0436225496,-0.0090597766,0.1054687575,0.5369826555,-0.1258685887,-0.5352920294,0.0339511372,0.47798118,0.2792463005,-0.136687398,0.0759350955,0.3156251609,-0.1674428582,0.4376081526,-0.1025658101,0.8173331022,-0.0009261135,0.1719650477,0.138971284,-0.3020308912,0.1291724443,0.0200279076,0.0107886679,-0.4951123297,-0.3523802161,-0.1284809411,-0.2606924474,0.2495348901,0.0580628514,-0.343375206,0.3005758524,0.0194640867,0.4572975338,0.4209087789,0.2916842699,-0.1678933203,0.2989626825,-0.2507620752,0.0380312651,0.0244168658,0.0722335279,0.1154952571,0.0764666349,-0.141377911,-0.4948067069,-0.2790382206,0.0151465517,-0.2930012643,-0.1004594862,0.3964217007,-0.5733570457,0.2118080556,0.0720924959,0.04818381,0.0640392303,-0.2333733588,0.0129593583,-0.1457787901,0.1276945472,0.183948487,0.0475329347,0.4533499479,-0.0087215574,-0.3378638029,0.3581652641,-0.0813947842,-0.0119384425,0.0148284594,-0.1256122291,0.1773061901,-0.3368036747,0.2464153767,-0.628660202,0.1099734455,-0.1781975776,0.045037251,-0.1984385252,0.0036137204,0.1091617271,0.2272614092,-0.3318321109,-0.2118319571,0.2083907127,0.0059665497,0.2285706252,0.5235672593,-0.0887126327,-0.04807809,-0.1036171764,0.0581538342,0.3272997141,-0.3375204802,-0.0491202101,-0.0297179222,0.065609552,0.1523433775,0.3768545389,0.2776260078,0.2131869495,-0.0744468719,-0.5255508423,-0.0729655996,0.2375106514,-0.0712140799,0.1975880712,0.1174603403,0.3237075806,0.2238592505,-0.1134935394,-0.2397029251,-0.03295701,-0.1323354989,0.1919335723,-0.0868279263,0.0964449942,0.051700104,0.0634484664,0.0887569487,-0.192830056,-0.1422792673,-0.1109476462,-0.1475551575,0.1835369468,0.1018511131,0.1381859332,0.1264923364,0.0560475253,0.0490148067,-0.0467185937,0.1385231018,0.0393861048,-0.1792175919,0.2147742212,0.34734115,-0.1672769934,-0.3585488498,-0.0370983481,-0.2041278034,-0.0396423712,0.1470280886,0.3860643506,-0.2622255683,0.0606436282,-0.183980599,0.0891807824,-0.0386753567,-0.058118239,0.1145231798,-0.3145363331,-0.2465545535,0.0138950851,0.4490870833,0.2860864401,-0.2015555203,-0.030049989,0.235149473,0.0334136561,-0.4120021164,0.0085355006,0.2943555713,0.1643490195,0.0969960317,0.4992057681,0.3679380715,-0.337335825,0.0763207898,0.054108277,0.4890707135,0.2271460444,0.1624105126,0.4281882644,-0.120604977,0.148973614,0.500107348,0.0717378333,-0.068675071,0.3365833759,-0.226596266,0.2901752591,0.0518489964,-0.0947274566,0.0120786726,-0.1402914524,0.33037889,-0.0836483538,0.3031802773,-0.1854189783,0.2107686251,0.3404168189,0.0991525203,-0.3103975058,-0.264610827,0.0073891794,-0.0769663379,-0.1239804476,-0.0777561739,-0.248784706,-0.3113349378,-0.0424013659,-0.1833255887,-0.2136457413,0.6925194263,0.2094757855,-0.1458517164,-0.2887672782,0.099605754,0.1255612969,0.0661409795,-0.2412374616,0.4364032447,-0.0067151501,-0.0461819135,0.1553471386,0.2981884778,0.8132764697,0.2181598842,-0.1735399514,-0.3258236647,-0.2322149277,0.0645576119,0.2646583021,0.0630190447,-0.0373362675,0.0212638974,0.2920491993,0.1168117151,0.0277587902,-0.0784622729,0.0650030896,-0.1590969265,-0.4842327237,0.2990456522,0.1847798973,-0.1018589437,0.0164223909,0.1767463088,-0.2542063892,0.0268722139,0.2197711021,0.4025389552,-0.0997011513,-0.1280664951,0.0130578224,0.2641711831,0.3753708899,0.4660414457,0.3042026758,-0.3263893723,-0.2322641909,-0.5930103064,0.3081550002,-0.1118557751,-0.0641729385,0.2311535031,0.0703425854,0.1258249283,0.2151816934,0.30558604,0.3428937793,-0.2636965513,0.1026008427,-0.4001891017,-0.1769956946,-0.0904512405,-0.1819176972,-0.0943459645,-0.1717551947,0.1907263249,-0.5734676123,0.0147785563,-0.1284413487,-0.3459657729,0.200353846,-0.19253923,0.4588808715,-0.0203613099,0.7599465251,0.057064373,-0.1314642429,-0.1995707601,-0.15740785,-0.3382564485,0.508472383,-0.2153967023,0.2439021915,-0.0355879702,-0.2223462462,-0.1448097527,0.2334987372,0.1409814954,0.0157278813,0.0833625644,0.0926648304,0.0641425624,-0.0942074135,-0.0230031405,-0.0197342616,-0.0309021324,0.2972713113,-0.4680343568,-0.672316432,0.540030539,-0.338793844,-0.235279426,-0.4275918007,0.2575130761,-0.0969465226,-0.0870121717,-0.4807568491,0.1060673445,0.3818357885,0.0325497799,-0.2270982862,-0.0324551389,0.0326505639,0.3201504648,-0.1376316845,0.5239678621,-0.2877165973,-0.1514039785,-0.1682212055,-0.2710402012]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"Hi, @Chenfei-Kang.\r\n\r\nI'm sorry, but I'm not able to reproduce your bug:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"glue\", 'cola')\r\nds\r\n```\r\n```\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 8551\r\n })\r\n validation: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 1043\r\n })\r\n test: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 1063\r\n })\r\n})\r\n```\r\n\r\nCould you please give more details and environment info (platform, PyArrow version)?","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":66,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\nHi, @Chenfei-Kang.\r\n\r\nI'm sorry, but I'm not able to reproduce your bug:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"glue\", 'cola')\r\nds\r\n```\r\n```\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 8551\r\n })\r\n validation: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 1043\r\n })\r\n test: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 1063\r\n })\r\n})\r\n```\r\n\r\nCould you please give more details and environment info (platform, PyArrow version)?","embeddings":[-0.0594903305,-0.1199711189,0.0148641206,0.2438831329,0.4732634127,-0.0013416682,0.4414617419,0.1253976971,-0.0142953452,0.2675631642,-0.1799774915,0.4658890367,-0.1191558838,0.1054266095,0.0933514312,-0.1634460539,-0.1493485421,0.1789052486,-0.0950517431,0.0041558379,-0.2951413095,-0.1227453202,-0.2647020519,0.1601062417,-0.4545831382,-0.2473363578,-0.1237274557,-0.0289071202,-0.2947974801,-0.3072052002,0.3905930817,0.045531936,0.0802016407,0.573389709,-0.0001046333,0.085742943,0.5220624804,0.0795245767,-0.1019385606,-0.4897305667,-0.1364545226,-0.1668332815,0.2139172852,-0.393879205,-0.174881503,-0.0434178077,0.0156303495,0.1047198623,0.3135638833,0.4849516451,0.3089385629,0.4509330392,0.0464522541,-0.2673178315,0.2966395915,0.1515992731,-0.0532518923,0.3601016104,-0.0297630094,-0.1026376411,0.1846834719,0.1567065567,-0.2669480145,-0.0967591628,0.1903866827,0.0687865466,-0.0415391438,-0.290898174,-0.0751495287,0.4143870771,0.3387809098,-0.4593112171,-0.2788108885,0.1180645153,0.2122831494,-0.2495187223,0.2602464557,0.0856177285,-0.0128446911,0.1975728273,-0.0867947638,0.2796069086,-0.2793028951,0.3049078584,-0.164794296,0.1444952488,-0.0630867183,0.0345241651,0.2060508877,-0.1919636279,0.059791442,-0.1310273856,0.0356741138,0.1007023752,-0.243607074,-0.1369945854,0.1091515124,-0.0550249554,0.1305607557,-0.2993285656,0.1565279961,-0.0825746059,-0.0047007087,0.4082542658,0.183615461,0.3742778003,0.2353915423,0.146733135,0.2031041682,0.1250560284,-0.2914271057,0.0851671919,-0.1916304976,-0.0795564353,0.3350311518,0.141778037,0.4543803036,0.004213843,-0.4538616538,-0.0484698229,-0.4231696129,0.0795800835,0.2269208282,0.4578568935,-0.1490905881,0.111972034,0.0448685922,0.1539013535,-0.1707451195,-0.2831352949,-0.2869074643,0.4119275212,-0.3327414989,-0.2005894333,0.1725695282,0.0631759688,-0.0440938734,-0.0142761068,-0.124532178,0.0147927273,0.0312171597,-0.4864667356,-0.1069177911,0.1802367717,0.0366909429,-0.1434192806,0.2177432328,-0.4385626912,0.0038601621,0.2104463279,-0.1390264481,-0.2205248028,-0.0644018948,0.2921482325,-0.3297555149,-0.1617497504,-0.1220805645,0.1581521481,0.158894524,0.0152816763,0.0386043973,-0.0845265612,-0.0218608137,-0.3067058027,0.0268780459,0.3902682364,-0.3743338883,-0.2277735025,-0.1163482741,-0.0573342443,0.3309760094,-0.0552941114,-0.0139303515,0.2842520177,-0.0162361618,0.1270191222,0.6166725755,-0.3811469376,-0.3189837933,-0.0871518478,-0.0525565632,-0.0323508754,-0.0000088416,-0.1415092349,0.0952094793,0.1851929128,0.4305967093,0.1825711876,0.0108494731,-0.0068763741,-0.2708131671,-0.2111279964,0.1911941022,0.0895309374,0.1389753819,-0.0566433519,-0.0780088603,-0.24225308,0.0179314129,-0.1191949025,-0.1323612928,-0.0616726913,0.5806082487,-0.0895553157,-0.1220896021,-0.5331310034,-0.2384607494,0.1451272964,-0.0384060368,0.4400658906,-0.1647625715,-0.2132972479,-0.4087546468,0.0467493199,-0.0534322485,-0.0422442332,0.2887170017,0.0327202827,-0.162372008,0.1476718634,-0.093931675,0.0690900609,-0.1891066879,0.1402491182,0.1150690317,0.3248268962,-0.052656617,-0.4338489175,0.0160555448,0.2450485379,0.4278887212,0.0615985915,-0.0518304221,0.3103211224,0.0966809914,-0.2166442722,-0.332929194,-0.0375642255,0.1061588004,-0.3284879327,-0.1916384846,-0.02861006,0.1814015508,-0.0039089932,0.0155871715,0.5066564679,0.073046267,0.0751549006,-0.0834585428,0.1962543428,-0.0521300845,-0.1202241629,-0.1189233959,-0.0935665593,0.0853727981,-0.0516477935,0.1604620069,0.0839664191,-0.1486921012,-0.0939611942,0.6276284456,0.066133365,0.4066179395,-0.1278576404,-0.253084898,0.1469340771,0.0753541663,0.0935161412,0.3560036421,0.309184581,-0.1741830856,-0.012002632,-0.1699138284,0.0200339202,0.0936975554,-0.0300577246,0.3375006616,-0.0191667695,0.326565057,0.0025228276,-0.3129945993,0.1154984683,0.0383162536,0.1894647479,-0.3931485116,0.1806956828,-0.3155397177,-0.1539050788,0.0592735447,-0.1398235261,-0.0421994813,-0.2707495987,-0.085526906,0.1815255284,0.0589705072,0.264362663,0.1597475857,-0.0327072591,0.0847381577,-0.0655731633,-0.102235347,-0.2201068252,-0.1743910462,0.1428499222,0.1274092495,0.2078600228,0.2995347381,-0.1425538957,0.0480940603,0.0138055841,-0.3602250516,0.0857338384,-0.1215342283,0.4401930273,0.3425597847,0.0728836656,-0.2368693203,0.0173933227,0.3619660437,-0.3473314047,-0.108051084,0.3983776271,-0.1979978234,0.0502504855,-0.2161714584,-0.3248876333,-0.510535121,-0.318955183,0.1311067194,-0.0294710808,0.0872791484,0.3062438071,0.2281634808,0.450894624,-0.052571062,0.0395837203,-0.1739951223,-0.0077805598,0.3531797528,-0.0148417056,-0.3720351756,0.0079913763,-0.1651002616,0.1313911527,-0.0117581366,-0.1790904701,-0.5396494269,-0.1400610059,0.288399756,0.0229437444,-0.0957489237,0.3742546439,0.0539732315,-0.1510512084,-0.1974764913,-0.082263872,0.0317742862,0.3172571063,0.0219645612,-0.069488965,0.2908799648,0.0552209876,0.2811395824,-0.113323316,-0.2513748407,0.4198065102,-0.2915056348,0.2461425811,-0.0716580451,-0.4766027927,-0.0776230395,-0.0555009358,-0.0022872451,-0.0360607766,-0.2989491224,-0.2200709581,-0.297008574,-0.1219761595,-0.3595552146,-0.1541765779,0.0609147735,0.089118734,0.0558054894,0.1404543817,0.155065909,-0.3061159551,0.0088768518,0.0000187792,0.1075965613,0.0372078121,-0.2144530863,-0.2214943469,-0.2509330511,-0.2910080254,0.3232398033,0.0477950834,-0.0251004752,-0.1071845666,-0.1048598811,0.084771201,-0.054349944,0.6767696738,-0.1964523345,0.0710406229,0.4911931157,0.2105484456,-0.432089746,0.02492279,-0.4132964909,0.1125910804,0.1601796597,0.3240831196,-0.3951091766,-0.0542449914,0.2625240684,-0.0359851383,-0.0594216958,-0.269669354,-0.2357648313,-0.2557882369,-0.1077657789,0.1962948442,0.0353272893,0.4339828193,-0.2649331689,-0.1105708033,-0.2390617579,0.0428951569,0.1868531257,0.2545810938,0.2625078559,-0.1614429355,-0.1597974896,-0.1762635708,0.058802031,0.2780003548,0.2882107198,0.0445406772,-0.4137676656,0.0157187078,0.0581392795,0.2611902654,0.0464258008,-0.1333332211,-0.0307984203,-0.0403293222,0.0974305719,0.1435642242,0.2177045345,0.386084348,0.059595976,-0.3346404433,-0.2437654287,0.3455174267,0.1484420449,0.0505361073,0.0170092452,0.2781257927,-0.2117078453,0.2881139517,-0.0201545823,1.0111832619,0.0690524206,-0.0597921535,0.5816109776,-0.0952846482,0.5560985208,-0.0648017898,0.1861322969,-0.2962427735,-0.3566812873,-0.0365724601,-0.1678650081,0.2394370884,0.0646672547,-0.6077862382,0.2306772918,-0.2865700722,0.1413257867,0.037199948,-0.0313408002,-0.0069267428,-0.0241819527,-0.3856900334,0.2118214965,-0.1228692979,0.1014267579,-0.0845644027,-0.10637898,-0.1233802363,-0.1591433734,-0.1048097834,0.0431123562,-0.5134391785,-0.1531040519,0.0520422086,-0.1772312224,-0.0729138181,0.1872905791,0.0402178206,0.0775228292,-0.1871171445,0.1700199097,0.1999508739,0.0999664664,-0.0692627802,-0.0332883447,0.4308818877,-0.0916927382,-0.3986715674,0.1275028139,-0.0470685437,0.1355867386,0.3548524976,0.0182911213,-0.0207855981,-0.4515144229,-0.0984330326,-0.1435841471,0.2059473395,-0.3068613708,0.2612115443,-0.0072328709,-0.1372915208,0.0108396858,-0.0357730947,-0.018511625,-0.0398774929,0.3423734009,0.3180928528,0.4484286904,0.2465310693,0.2419533879,-0.1212224513,-0.3200915158,0.1532203853,0.230802089,-0.4224052429,0.2418849766,0.2260043472,-0.1214663088,-0.053419821,0.4723202586,0.0875411034,0.1689212471,-0.1257396638,-0.148903504,-0.2344330549,0.3241607547,0.1022827998,0.0040802173,0.1147460192,0.4286702871,0.2379003763,0.017730521,-0.4440496266,-0.0286853164,-0.4677805901,0.070526503,-0.0260700732,-0.2602921128,0.2004087865,-0.0038485706,0.2606965303,0.1246233732,-0.0757152289,-0.3621689975,-0.1629702449,0.0902142152,-0.002169311,0.1705976725,0.1876295209,0.0336485803,-0.230336979,-0.1281399429,0.0921760574,-0.055329401,-0.1467017829,0.21817936,0.2054448128,0.3228447139,0.0179912299,-0.0136706028,0.0152384248,-0.0924621522,-0.0348306708,-0.0050635072,0.1245716438,-0.0292172786,-0.0914055929,-0.0829483122,0.1921166331,0.1356510073,0.0203486923,-0.2578234971,-0.2741311491,-0.1145555899,0.2158735543,0.1211166307,-0.1575217694,0.1765080243,-0.0109323282,0.3513850868,-0.3763158917,-0.0102533428,0.2301227152,-0.0349603593,0.2269831598,0.0463015363,-0.2856045365,0.3148671389,-0.2031256407,0.090680182,0.2280460447,-0.0443477929,0.1081102192,0.2768344283,-0.1694982648,-0.0334678218,0.4123375118,0.070386678,0.1482765526,0.02117116,0.0128191616,0.1463374197,-0.0221673734,0.0858672485,0.3090984523,-0.3254109621,0.0323023833,0.2096980363,0.0278553236,-0.0797003284,-0.2609014511,0.516067028,-0.2635023892,-0.1445791423,-0.3344384432,0.1385084242,-0.200316295,0.0369850062,-0.2015442103,-0.1111779511,-0.1869933307,-0.3066197634,0.0259554833,-0.0909686759,0.2416601479,-0.0493581742,-0.0289359875,-0.2844852209,-0.0523675457,0.3596524894,0.2271602452,-0.20385921,0.1030545309,0.0247038733,0.1235098615,-0.1864320636,0.5136396885,0.4270206392,0.2857337296,-0.1543841809,-0.0936766863,0.0345495343,-0.3090969026,0.0706597194,0.0969653726,0.2919444442,0.2387932241,0.0644041747,0.2788921297,-0.283341378,0.1762279719,-0.0601931773,0.0752444565,-0.3707785904,0.3305317461,-0.1943280846,-0.1633215249,-0.1663211137,-0.0086910268,-0.4805711806,0.1700413376,0.3345673084,0.1505533755,0.18274571,-0.1898440421,0.1426260471,-0.2063407302,0.2242961377,-0.0433819778,0.3339892626,-0.279900521,0.1178943589,-0.4362550676,0.3936928809,-0.0448485799,-0.0489642657,0.0756321251,-0.0840160921,-0.1616720408,0.0349483863,-0.0744248405,0.2408670038,0.1278519928,0.0993816927,-0.6071122885,-0.199648127,-0.1119104177,-0.2307693958,0.1361706406,-0.2414505482,-0.047243461,-0.1864420474,0.1147017404,0.0427970104,-0.3028406501,0.1948673278,0.2954314947,0.6067336798,0.1576177031,0.4007611871,-0.013851488,0.0997695774,0.0633205548,-0.2700609863,-0.1527519971,0.1926423162,0.1350933015,0.3505522013,0.0757688209,0.232017532,-0.0688422769,0.0777027383,0.0588588454,-0.124382481,-0.1221504211,0.163699314,-0.1173157468,-0.2261742949,-0.255618602,0.0282795932,0.0752818286,0.0470450297,-0.0877219141,-0.2791985869,0.670065105,-0.3341488242,-0.2696869671,-0.0509191118,0.2001962513,0.2284435928,-0.2409944087,-0.6293930411,-0.0089962306,0.4204948246,0.0220112894,-0.2528333664,0.2136524618,-0.0641363934,0.1553049386,-0.0634708256,0.0211865623,-0.1120157316,-0.0563756153,0.047504697,-0.2333326787]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"> Hi, @Chenfei-Kang.\r\n> \r\n> I'm sorry, but I'm not able to reproduce your bug:\r\n> \r\n> ```python\r\n> from datasets import load_dataset\r\n> \r\n> ds = load_dataset(\"glue\", 'cola')\r\n> ds\r\n> ```\r\n> \r\n> ```\r\n> DatasetDict({\r\n> train: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 8551\r\n> })\r\n> validation: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 1043\r\n> })\r\n> test: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 1063\r\n> })\r\n> })\r\n> ```\r\n> \r\n> Could you please give more details and environment info (platform, PyArrow version)?\r\n\r\nSorry to reply you so late.\r\nplatform: pycharm 2021 + anaconda with python 3.7\r\nPyArrow version: 5.0.0\r\nhuggingface-hub: 0.0.16\r\ndatasets: 1.9.0\r\n","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":116,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\n> Hi, @Chenfei-Kang.\r\n> \r\n> I'm sorry, but I'm not able to reproduce your bug:\r\n> \r\n> ```python\r\n> from datasets import load_dataset\r\n> \r\n> ds = load_dataset(\"glue\", 'cola')\r\n> ds\r\n> ```\r\n> \r\n> ```\r\n> DatasetDict({\r\n> train: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 8551\r\n> })\r\n> validation: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 1043\r\n> })\r\n> test: Dataset({\r\n> features: ['sentence', 'label', 'idx'],\r\n> num_rows: 1063\r\n> })\r\n> })\r\n> ```\r\n> \r\n> Could you please give more details and environment info (platform, PyArrow version)?\r\n\r\nSorry to reply you so late.\r\nplatform: pycharm 2021 + anaconda with python 3.7\r\nPyArrow version: 5.0.0\r\nhuggingface-hub: 0.0.16\r\ndatasets: 1.9.0\r\n","embeddings":[-0.0170860495,-0.3240526319,0.062121436,0.2103672326,0.3840116858,-0.0155181745,0.415794313,0.1785911471,0.1879396886,0.3116176724,-0.2562711835,0.2617366314,-0.0900785625,0.1555979401,0.134646982,-0.1783233583,-0.1780708879,0.2362426817,-0.1429384202,-0.0610416606,-0.307266444,0.016021125,-0.2765933871,0.1563270539,-0.3710355163,-0.2232295871,-0.0727495775,-0.0337860063,-0.3516702652,-0.3736093938,0.3502697647,-0.0149404071,0.1077515855,0.3936695755,-0.000109273,0.039748542,0.5531897545,0.0916725695,-0.1215192825,-0.486445725,0.0033005883,-0.153592959,0.2640893161,-0.2971876264,-0.2491431981,-0.1678692102,-0.0046053282,0.1546724737,0.3977606893,0.5068286061,0.2735918462,0.5179857016,0.1557451636,-0.3057215512,0.2476199716,0.1295629144,-0.0830335468,0.3645498157,0.0868025422,0.0324823782,0.2476591766,0.22567074,-0.2615945935,-0.1020329148,0.2481998503,0.0714492425,0.0565817356,-0.3359338641,-0.0874417275,0.3458363712,0.2176503092,-0.4307984114,-0.2740110457,-0.0156725049,0.1589335203,-0.2778740823,0.2152642608,0.1117749512,-0.0614302717,0.2610106766,-0.0477596298,0.2916436493,-0.2724869251,0.2264740765,0.0332289934,0.0998286605,-0.0974352807,0.0864828005,0.3184731901,-0.1978602111,-0.0712477863,-0.0744847059,0.0765407234,0.1997876465,-0.2625720799,-0.1569988281,0.0412150547,-0.0527565517,0.2478982806,-0.0623121895,0.010698121,-0.0784372389,0.0866332129,0.3945228755,0.1457946301,0.3695580363,0.2746926844,0.0596430153,0.147126019,0.2683903873,-0.1179920584,0.0550732911,-0.1812906712,-0.0505757257,0.2737024724,0.1010781154,0.4157111049,-0.0123857539,-0.3835292757,-0.0340619497,-0.249388665,0.2141660452,0.2170223594,0.4983271658,-0.1098527387,-0.006581971,0.1920767128,0.1782709658,-0.1211596727,-0.1304862946,-0.2357181609,0.34223032,-0.3215116858,-0.0617322512,0.2265961319,0.0308971405,0.0039272024,-0.0047287471,0.0682608932,-0.0418036729,-0.0186945554,-0.3479027152,-0.0821475908,0.2535481453,0.032342352,-0.0685248151,0.2264001071,-0.410043478,-0.0981175676,0.1309780478,-0.0746355355,-0.1795981675,-0.2304216176,0.2343917638,-0.2560875714,-0.0213031564,-0.0934348702,0.0212540068,0.1447509378,0.1013658345,0.1272418797,-0.0190788247,0.0178890768,-0.2180711776,0.1380720735,0.3654916883,-0.100190863,-0.3205661774,0.0321522094,-0.0025296691,0.1171867996,-0.0173707716,0.0019335701,0.2421703041,-0.0305478908,0.1668857634,0.4985628724,-0.62414819,-0.3442799449,-0.2747382522,-0.0806370825,-0.0161460973,0.0393288732,-0.2286882102,0.1914647818,0.2106938213,0.4453007579,0.1527808011,0.1522087157,0.0167638846,-0.2598265111,-0.2841210067,-0.1492304802,0.1216640398,0.0572228469,-0.0611823201,-0.1246782616,-0.3315826058,0.0439961776,-0.1464243233,-0.0947787762,-0.0285352468,0.4802513719,-0.0267939679,-0.0334615521,-0.5103312731,-0.3110117316,0.1444107145,-0.0827431381,0.4963111877,-0.2935872972,-0.2401018441,-0.3797984123,0.0216788147,-0.0260307416,-0.1073646471,0.2608529925,-0.0292999409,-0.0793902054,0.2039854974,-0.148156628,0.3061811626,-0.0629678369,0.3361934423,-0.0899120644,0.2520775199,-0.0519085974,-0.4535894096,0.0178125389,0.3152451813,0.407485038,0.0184305087,0.0400744975,0.2059644312,0.120073542,-0.2516606152,-0.3879702389,-0.2059795707,0.0908953026,-0.2052687109,-0.1253317297,-0.1193971634,0.1875582188,0.043237146,0.0816024616,0.4845341146,0.0002889726,0.1447360814,-0.0010887757,0.2192268968,0.0464254729,-0.2223508209,-0.1943704933,-0.1116545573,0.1098374575,0.0302241202,0.2480703294,0.0179758705,-0.1454162598,0.0168016292,0.6618235111,0.0420250148,0.258257091,-0.1182069257,-0.3660911322,0.2513479888,0.1739427298,0.087606214,0.3652841747,0.2524144948,-0.2956422269,0.0845419914,-0.1635932028,0.0505567454,0.0976488069,-0.0679601654,0.3684989512,-0.0912532359,0.3075386584,0.0185329355,-0.3090053499,0.0744575188,-0.0793073252,0.1062876061,-0.3757323921,0.1150494143,-0.2569516301,-0.1478515416,-0.0088819563,-0.133373037,0.0266572647,-0.1974048913,-0.0598685481,0.3584897518,0.0893847048,0.282640487,0.2069196403,0.0192367453,-0.0590613931,-0.1030487418,-0.1551313996,-0.108617492,-0.11700131,0.0823388919,0.0588655174,0.1915499568,0.3331587911,-0.3143273294,0.0909607261,0.1370236278,-0.3026359379,0.0858223662,-0.1740436107,0.2888210118,0.2684489191,0.1589440256,-0.1759042591,-0.0624604672,0.4340486825,-0.3824962378,-0.192428574,0.2827731073,-0.1414361447,-0.0550831221,-0.2623379529,-0.2640984952,-0.5197219253,-0.2946316004,0.2550328076,0.0227000788,0.0717564151,0.3472696841,0.234048292,0.452318579,-0.3834219277,-0.0229388513,-0.1441199481,0.0412894003,0.2728778422,-0.0820257813,-0.4259680212,-0.0077680945,-0.0855097622,0.2539603114,-0.1286979169,-0.0568948649,-0.4479478002,-0.1266304255,0.3238252997,0.0086144321,-0.1531503648,0.6029988527,0.0156400762,-0.1078105569,-0.1583341658,-0.1767390072,0.1045932546,0.1382757425,0.0056557627,-0.0700706616,0.2453286648,-0.0265153348,0.351564467,-0.0049929833,-0.2051898241,0.3471206427,-0.304443121,0.3146957457,-0.1218410507,-0.5708717108,-0.0727783665,-0.0265240464,0.0520668849,-0.1330883354,-0.2462659478,-0.0897768289,-0.3285742104,-0.1738543063,-0.4415084422,-0.2324540317,0.0581631921,0.1306092143,-0.0827693418,0.0958863199,0.0774856657,-0.3568073511,0.0563106313,-0.0132170403,0.212240696,0.0356143676,-0.1878067106,-0.4234542251,-0.2692607045,-0.3553898036,0.3830406666,0.0468409471,0.0110894823,-0.0506009758,-0.1024787277,0.1410767138,0.0076352716,0.6806127429,-0.0301307887,-0.0055903825,0.416000694,0.1417528987,-0.6230165362,0.0706069246,-0.3021711707,0.0531921424,0.1227650046,0.3985318244,-0.4662339985,-0.1559689939,0.3363058567,-0.0372493379,-0.0328230187,-0.2179400921,-0.2882821262,-0.3978614211,-0.213686794,0.1251843274,0.1448903382,0.3859930634,-0.225021705,-0.1091162786,-0.124067463,0.0948003605,0.0340228826,0.2724271715,0.2362679839,-0.228084147,-0.095412448,0.0113189863,0.0838170201,0.2514652312,0.2685821652,-0.0433206074,-0.5514891148,-0.1514342278,0.1056525484,0.2161636204,0.0786006078,-0.1357617527,0.0079719489,-0.0911873356,0.1339932978,0.2157751322,0.1181629151,0.4902128577,0.169891879,-0.2991678119,-0.191720441,0.452149719,0.1290634274,0.0447196178,0.1472747773,0.4835731387,-0.169077903,0.2779710591,-0.1196076274,1.0848121643,-0.0030835283,-0.005263906,0.5609253645,-0.1063345298,0.6737447977,0.0900183097,0.2337673157,-0.36399436,-0.1940107197,-0.0178164672,-0.1334923059,0.2924979925,-0.0650964603,-0.5087834001,0.1753665209,-0.2362163216,0.1639979631,0.0437814966,0.0470250808,-0.1116012633,-0.1634422541,-0.319111824,0.1553267986,-0.0189958215,0.2381698191,-0.0293426774,-0.1032173112,-0.2884888649,-0.2500365078,-0.1799219549,0.0486413091,-0.4893500209,-0.2644637525,0.2028944343,-0.1735015661,-0.2620542645,0.035055887,0.1736058295,0.0271786191,-0.2295172513,0.085042268,-0.0116313575,-0.0057277782,-0.0612283163,-0.1201168895,0.4516772926,-0.1806232482,-0.5559980273,-0.0021970393,0.0283423346,0.1189502105,0.3397734463,0.0533939674,-0.0720781907,-0.3765566647,-0.1956294924,-0.0863566697,0.2171403468,-0.2849394381,0.215872243,-0.0756182298,-0.1748709828,-0.0020164615,-0.0216978751,-0.0546836369,0.0181161389,0.3995351791,0.2500368953,0.3563725352,0.2536879778,0.1453519017,-0.1277043223,-0.2788094282,0.0454512127,0.2644447982,-0.5634872317,0.1480279118,0.2177876085,-0.1283232868,-0.0758905932,0.4255404174,0.0340973884,0.0199873578,-0.1081608087,-0.1711105853,-0.2474561185,0.4519791007,-0.0532878414,-0.053389363,0.2934162915,0.5443694592,0.2228721976,0.0077437623,-0.4028413296,0.025847666,-0.3943938017,0.0516046882,0.0506236702,-0.2735677063,0.2741547823,-0.0607873648,0.2235942632,0.2008981258,-0.0648372918,-0.3292416632,-0.1574597955,0.1020702645,0.0132466042,0.1558352858,0.1404897571,0.0429231934,-0.1865489632,-0.1436392814,0.1646613628,-0.0484871604,-0.0519672707,0.2371664941,0.1691443026,0.3118648231,0.0495464392,0.0909304991,-0.0028736435,-0.1294036061,-0.0817828923,0.0178820342,0.0095177591,0.0264174677,-0.0565491654,-0.1073949337,0.1240915209,0.1748039126,0.0907406211,-0.2451831847,-0.3529208899,-0.2803744972,0.1817104667,0.1533816606,-0.1543813348,0.1704663485,-0.0670958683,0.2948552966,-0.4210546017,0.0239637289,0.2194127142,-0.0067794458,0.179174453,0.1512840837,-0.2732408047,0.2063111812,-0.0471630208,0.0283583831,0.4666559994,0.0181072932,0.0889500454,0.2631120682,-0.1317112595,-0.0344259217,0.4894234538,0.1561266631,0.2242719233,-0.0069709676,-0.0884023011,0.2236964554,-0.1240096763,0.074921973,0.2926060557,-0.216833517,-0.0012596183,0.1987000406,0.0942938253,-0.127527684,-0.3349516094,0.5498442054,-0.2395892292,-0.2107713521,-0.1777208298,0.1950450093,-0.2097447366,-0.0012295916,-0.1400179267,-0.0810490996,-0.1268614382,-0.4116225243,0.0561529286,-0.1514625102,0.4862962961,-0.1211124882,-0.0871192738,-0.2842113078,-0.2175373286,0.408749193,0.3050468564,-0.154292956,0.1880830228,0.0706383362,0.1759784222,-0.2439061552,0.5857952833,0.4205303788,0.2032050192,-0.0877042264,-0.0394125432,-0.0987181142,-0.3073500991,0.0758499652,0.1601038575,0.3437224329,0.2210115194,0.0020970211,0.2533309758,-0.2443259656,0.2270522863,-0.1037411615,0.0440704562,-0.4042601585,0.3363026083,-0.3260605931,-0.2499150783,-0.3251251876,0.0070015988,-0.4086143374,0.1709913462,0.3403113782,0.0040550446,0.1531253904,-0.3124785423,0.1025997922,-0.1237484664,0.287279129,-0.0421427861,0.2992600203,-0.1812504232,0.0731786937,-0.4249604642,0.395984292,-0.1252929121,-0.112262547,0.0873513669,-0.1618833095,-0.1488880217,0.0399846993,0.0557344407,0.4386814237,0.056703303,-0.059051577,-0.5389719605,-0.1532924324,-0.1344299763,-0.2273552269,0.2009892464,-0.2520205975,0.0800778195,-0.1442407817,0.0637667999,0.0295061786,-0.0963596702,0.1442123055,0.0670828372,0.6210793257,0.1881039739,0.3392416537,-0.0430545397,-0.0689965189,0.0249925554,-0.174067378,-0.2406356484,0.2803048193,0.2160280198,0.2756659687,0.0625708327,0.0870543271,-0.0475448668,0.1085865125,0.0825032964,-0.2264692187,-0.0616678037,0.119794324,-0.0202589016,-0.2373049855,-0.1996346414,0.0610714257,0.005293489,0.0769687742,-0.2243449986,-0.2481006831,0.7152537704,-0.2563150823,-0.2637996078,0.055393666,0.2930933833,0.1165040582,-0.4328460097,-0.7834493518,0.0166447517,0.2939421535,0.083926104,-0.2240283042,0.2948251665,0.0100844931,0.0482179038,-0.0746177435,0.1352374554,0.0407708175,-0.0168586764,0.0712476522,-0.2007396221]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"- For the platform, we need to know the operating system of your machine. Could you please run the command `datasets-cli env` and copy-and-paste its output below?\r\n- In relation with the error, you just gave us the error type and message (`TypeError: 'NoneType' object is not callable`). Could you please copy-paste the complete stack trace, so that we know exactly which part of the code threw the error?","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":69,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\n- For the platform, we need to know the operating system of your machine. Could you please run the command `datasets-cli env` and copy-and-paste its output below?\r\n- In relation with the error, you just gave us the error type and message (`TypeError: 'NoneType' object is not callable`). Could you please copy-paste the complete stack trace, so that we know exactly which part of the code threw the error?","embeddings":[-0.2277715802,-0.3193501234,0.0008596373,0.3918522596,0.3677237034,0.0557456017,0.3176339269,0.1545283645,0.1733444631,0.2960280478,-0.1426855326,0.4552958608,-0.0776570067,0.2298437357,0.0493332893,-0.1651720107,-0.2105034888,0.2611335516,-0.1191705689,0.0262278002,-0.5092200041,-0.0741478875,-0.1640852094,0.1349925548,-0.237754494,-0.2719812989,-0.1240942478,0.0408258066,-0.1921427697,-0.2838498354,0.3645710647,-0.0801881999,0.2122620046,0.5988925099,-0.0001008489,-0.1514540613,0.435205698,0.0475999489,-0.1254921108,-0.3523042202,-0.2946639061,-0.3847207725,0.18985717,-0.3827019334,-0.0743031427,-0.0082017854,0.0286093634,-0.2426373065,0.3481251895,0.3484524488,0.3506994545,0.415902108,0.0833576843,-0.2823812366,0.1884763539,0.0775724575,-0.1048533171,0.2966803014,0.0623726882,-0.0164799206,0.3145338893,0.1602044255,-0.2830132842,-0.0542992242,0.2551215291,0.0645872205,-0.0053246845,-0.3559243679,0.0310364459,0.3066703379,0.4195721149,-0.3946249187,-0.1872214228,0.0607764684,0.1299087554,-0.2234973758,0.1539727449,-0.0115599288,-0.0064773369,0.0658862591,-0.0040367073,0.1557925344,-0.236804232,0.2183835655,-0.1664240062,0.0901961327,-0.2121003419,0.107063584,0.0670970306,-0.1992394775,0.0452290624,-0.1839859337,0.0351355337,0.1096044704,-0.221126616,-0.1452792883,0.1299011409,0.1111679673,0.1270107031,-0.2455778271,0.1168725938,-0.0483121127,-0.0236970503,0.4071713388,0.239894405,0.2160654962,0.197072491,0.1496332735,0.2160288244,0.0441346094,-0.1853791922,-0.0519746169,-0.1255763024,-0.1014563665,0.2814617455,0.1962061971,0.5696896911,-0.0633770376,-0.4451116025,-0.0369891338,-0.207089901,0.0923674777,0.2541407347,0.3783914447,-0.1840863526,0.1531827301,0.2149179727,0.0695832223,-0.1426405609,-0.2327828705,-0.2402686328,0.3336758912,-0.2726967335,-0.232448265,0.0732233152,0.0800087824,-0.1016519293,0.0515346751,-0.0466053039,0.0736266375,0.1052564383,-0.382973969,-0.1715993732,0.1613502353,0.1342369616,-0.0109220734,0.3575491905,-0.413230598,-0.1455421895,0.1080859974,-0.3179209232,-0.0927535221,-0.0903914943,0.3212557733,-0.2362898439,-0.079209581,-0.1951911002,-0.0047342735,0.0772167519,-0.2021209151,0.0243879221,-0.1612926275,0.0099704852,-0.4124752581,0.0396714061,0.5434519649,-0.3702253997,-0.131931603,-0.3560011387,-0.0526704639,0.3650825024,-0.1617242545,0.0869184732,0.2049263865,-0.1619556099,-0.0783798695,0.5953560472,-0.3626013696,-0.1942619234,0.0556299053,-0.079251498,-0.0580402203,-0.067448847,-0.0140706375,0.0950176865,0.1139309034,0.4095413983,0.2211544514,-0.0633310452,-0.027870506,-0.1950395554,-0.1733900309,0.1649025828,0.0918372497,0.1436722875,0.1626258194,0.0630830303,-0.2037843466,0.0184938218,-0.0397915058,-0.09291742,-0.0652414113,0.6565243602,-0.0872236714,-0.0788739324,-0.5945641994,-0.3253455162,0.1684785038,-0.0677354112,0.2807234824,-0.1447329372,-0.1707682014,-0.3664011657,0.0677765012,-0.0498286709,0.0480884425,0.34265396,0.1271567047,-0.1998722106,0.0451182313,-0.1443472803,0.1779200137,-0.1498600394,0.1019958928,-0.0057357312,0.3032817245,-0.0105569726,-0.3550938368,0.0437213145,0.2375442684,0.5392077565,0.0301111341,-0.0714498162,0.2961714268,0.0839738399,-0.1378158778,-0.1490661353,-0.0828541517,0.1191909611,-0.2615092099,0.0761332735,0.113510862,0.341347903,-0.0806762725,0.1379154474,0.550291419,0.070542492,0.218250677,-0.0425794087,0.2359925807,-0.0458750986,-0.0404749922,-0.09132009,-0.1391232908,0.1310136467,0.0969333351,0.1783950478,-0.0640903935,-0.1166954562,0.0628789961,0.5675632358,0.0985855833,0.3846487105,-0.0710379705,-0.3002262712,0.1548450738,0.0890177488,0.2822793424,0.4458229542,0.3082370758,-0.1348168105,0.0028449628,-0.105473116,0.0909257606,0.069577001,-0.008273948,0.2112074643,-0.0183985941,0.1782516539,0.0799446777,-0.1778929234,-0.0452206172,-0.2235442847,0.2644327581,-0.3588179946,0.1916420162,-0.2675382495,-0.1693612486,0.0416815057,-0.0662412271,-0.1583892852,-0.343996793,-0.2322412729,0.0256404057,0.1133051515,0.1315740049,0.114424713,0.1559722424,0.0080657853,0.0855939835,-0.1656484157,-0.1026751921,-0.1289932728,0.1639403254,0.1211261749,0.2124070823,0.4851669073,-0.2281131297,0.2450820208,-0.0264441986,-0.227994293,0.0638851374,-0.0960038304,0.4924280643,0.4533854723,0.170710057,-0.0828904957,-0.0167101845,0.3197818995,-0.3656996191,0.0297314506,0.2110722065,-0.2212308943,0.1077203006,-0.1586056501,-0.1741414219,-0.5684420466,-0.4111053944,0.1035611331,0.1680119336,0.1657065153,0.1624201983,0.1247417033,0.5334134102,0.1375749111,0.1933974028,-0.195846498,-0.0487356074,0.2836108506,-0.122860238,-0.2586844563,0.1470841169,-0.1050434485,0.2065625638,0.0155344047,-0.1752419621,-0.5393431783,-0.1731456071,0.1546034515,-0.0829334036,0.0919005796,0.2602021098,0.1155401021,-0.1390545517,-0.2339776456,-0.1428654939,-0.1661490351,0.2557976246,0.0672933385,0.0401755236,0.3397756815,-0.178367734,0.1294477135,0.0065674935,-0.2346490473,0.3460306227,-0.2629454136,0.3489436209,-0.1248321906,-0.5024338961,-0.1444402039,0.0058216783,-0.0900051892,-0.0492299162,-0.1613157392,-0.1666391939,-0.3757922053,-0.12356738,-0.4473450482,-0.173208639,-0.02683492,0.213281557,0.0196586438,-0.0107808877,0.1226413622,-0.4682358205,0.0489344634,-0.0516467467,0.2382136434,0.0197485723,-0.2432868481,-0.2479406893,-0.3090171814,-0.1854937226,0.4490612149,-0.1629894227,0.0337656327,-0.1722394228,-0.0209151395,0.0463810414,-0.0175552871,0.5029234886,-0.3419142962,0.0797719881,0.4985005856,0.1387616396,-0.3840284348,0.1739851683,-0.4070257545,0.3183746338,0.1517318189,0.2837773263,-0.3160907328,-0.0206106473,0.2402315289,0.1300393194,-0.1309812963,-0.2083940059,-0.1448206306,-0.2342955619,-0.1278313398,0.0245009754,-0.0037171121,0.3006121814,-0.1316144913,-0.1478490233,-0.3589732349,0.0226067789,0.2575157881,0.2095248997,0.2955670655,-0.0968761295,-0.0735841319,-0.068001546,0.0821264386,0.1955122501,0.274494797,0.0189455654,-0.391548723,0.0630288497,0.0565042421,0.2511690259,0.0781225115,-0.1441668868,-0.0535798632,-0.0800635144,0.1943779588,-0.0523580499,0.1001058295,0.2666032314,0.0928936079,-0.4500440657,-0.3444149792,0.388646096,0.191103518,-0.0468111858,0.2515259981,0.2554996312,-0.2587035596,0.2411348969,0.0585117638,1.0566794872,0.0451303609,-0.1790634096,0.4914720356,-0.1253181845,0.5688884258,0.0427971072,0.0154062081,-0.3378975689,-0.2656220496,-0.0510648116,-0.1546158791,0.3252924979,-0.0990185961,-0.6059663892,0.2722361684,-0.4187105596,-0.0113039594,-0.0516176261,-0.1450726092,-0.1615752131,-0.130570069,-0.4006145895,0.2759853601,-0.1767890304,0.1705224812,-0.1539548337,-0.0489460118,0.0174538828,-0.1243231222,-0.1045431644,-0.00681563,-0.4354237616,-0.2015477568,-0.1179327592,-0.1617850065,0.017373981,0.1457440257,0.1715212911,0.1336825788,-0.2375279218,0.0901856199,0.0600658134,0.0691440105,-0.0843933225,0.0353280008,0.2427478284,-0.0197919644,-0.3155231774,0.1725399643,0.0347914137,0.1483977884,0.3645579815,0.0996133909,-0.0831791535,-0.3884962499,-0.1447437406,-0.0601997264,0.2035056949,-0.2569366693,0.2733273804,0.1267134398,-0.075365968,0.1074780077,-0.0244918708,-0.0938684121,-0.0563318618,0.2750467658,0.3504133224,0.2657757103,0.0927115083,0.201552093,-0.2149729133,-0.3738940656,0.2654359341,0.0770252869,-0.22658512,0.2399508804,0.1424390525,-0.0477566309,-0.0740084425,0.328137219,0.1060338765,0.0496939942,-0.1529465616,-0.2888395488,-0.2707046568,0.2582178116,0.1262251735,0.0776324049,-0.0284722429,0.2088868767,0.1913177222,-0.0492134839,-0.4822713137,0.0605647266,-0.3836083114,-0.0870803818,0.0551759526,-0.0908131376,0.1894112378,-0.0005163328,0.2610564232,0.0888261497,-0.0810422897,-0.3424244523,-0.1005696654,0.0964426696,0.0030663002,0.0352182798,0.1485673785,-0.1207061037,-0.1707189232,-0.1411678195,0.0266073775,0.0593882687,-0.1587965488,0.1150495261,0.1243259236,0.2831470072,0.0353000462,0.2056675106,0.0769665316,-0.0229453649,-0.0435421765,0.0187972002,0.0036229359,-0.0640793815,-0.0301972888,0.0279702153,0.1451300681,0.1402733177,0.1255751401,-0.1008501798,-0.2102200091,0.0878476053,0.1658077836,0.1559677124,-0.141983971,0.2770460844,-0.0868307948,0.3664443195,-0.274410367,0.0608054735,0.1910197437,-0.1125192195,0.2319393456,0.1514575779,-0.3416772187,0.2574727237,-0.160598442,-0.024174612,0.3763528764,-0.1595057696,0.154295519,0.2322441638,-0.1036700085,-0.0269345958,0.3910489976,0.2250012904,0.089335233,0.0697669387,-0.1052329168,0.1659094244,-0.0354886912,0.0229820106,0.23917602,-0.4080982804,0.2459843904,0.2516362369,-0.0291607659,-0.0656304657,-0.1686448157,0.2762493193,-0.5681785941,-0.1020831168,-0.3088049889,0.0113962851,-0.1732568741,0.1290312707,-0.128262803,-0.1437105089,-0.2699719965,-0.3319649398,0.0556728579,-0.1790091693,0.195461601,-0.1371625513,0.1048784405,-0.1952640265,-0.0611734428,0.2457263917,0.1937719285,-0.2082418054,-0.0402503274,-0.0125757577,0.2360838503,-0.1725241244,0.5394890308,0.3954853415,0.3322067857,-0.1617974192,-0.0274389703,0.0731182173,-0.2820368111,-0.0603159703,0.0485493392,0.0340395607,0.2409244031,0.0659549609,0.3405122161,-0.3197934926,0.0967679396,-0.0659998953,0.0167914163,-0.3043793738,0.4304428101,-0.3063565493,-0.1328262538,-0.1199343801,-0.1177720949,-0.5592741966,0.0602604449,0.4006091654,0.0640409663,0.1086076871,-0.1182630286,0.1509963721,-0.1716412604,0.1180160418,-0.0006215716,0.3370815516,-0.2615213096,0.0773840994,-0.4892735481,0.3647934198,0.0276360251,-0.1982382685,0.0315376259,-0.119094789,-0.1470418572,0.009699991,0.0401231274,0.238081336,0.2386036366,0.0062592286,-0.5759982467,-0.3410032392,-0.1011373773,-0.2159598768,0.1086912006,-0.2300252169,-0.0447582565,-0.1500625759,0.1826710105,-0.0428104773,-0.257914722,0.0840296447,0.1191138402,0.6550236344,0.1934750229,0.3989144266,-0.12089663,-0.0226046145,0.1570947468,-0.3202399313,-0.0817679614,0.2735521197,0.1210139245,0.2002992332,0.0415458567,0.1631483734,-0.2030333132,0.0024250993,0.0274581667,0.0585398674,-0.1464440376,0.3077630401,-0.1645611525,-0.2361537367,-0.2291445434,0.1095362157,0.0604569279,0.1458997875,-0.0198298264,-0.1106984243,0.624981761,-0.1692866236,-0.2598852217,-0.0570382588,0.3724466562,0.0044079828,-0.1203285828,-0.4922212958,0.003138799,0.4646417499,-0.0488136932,-0.1663854867,0.2039655447,-0.1955647022,-0.0195632055,-0.0393957794,0.1110214889,-0.0188861452,0.0038260752,0.2579205334,-0.1187548116]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"> * For the platform, we need to know the operating system of your machine. Could you please run the command `datasets-cli env` and copy-and-paste its output below?\r\n> * In relation with the error, you just gave us the error type and message (`TypeError: 'NoneType' object is not callable`). Could you please copy-paste the complete stack trace, so that we know exactly which part of the code threw the error?\r\n\r\n1. For the platform, here are the output:\r\n - datasets` version: 1.11.0\r\n - Platform: Windows-10-10.0.19041-SP0\r\n - Python version: 3.7.10\r\n - PyArrow version: 5.0.0\r\n2. For the code and error\uff1a\r\n ```python\r\n from datasets import load_dataset, load_metric\r\n dataset = load_dataset(\"glue\", \"cola\")\r\n ```\r\n ```python\r\n Traceback (most recent call last):\r\n ....\r\n ....\r\n File \"my_file.py\", line 2, in \r\n dataset = load_dataset(\"glue\", \"cola\")\r\n File \"My environments\\lib\\site-packages\\datasets\\load.py\", line 830, in load_dataset\r\n **config_kwargs,\r\n File \"My environments\\lib\\site-packages\\datasets\\load.py\", line 710, in load_dataset_builder\r\n **config_kwargs,\r\n TypeError: 'NoneType' object is not callable\r\n ```\r\n Thank you!","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":154,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\n> * For the platform, we need to know the operating system of your machine. Could you please run the command `datasets-cli env` and copy-and-paste its output below?\r\n> * In relation with the error, you just gave us the error type and message (`TypeError: 'NoneType' object is not callable`). Could you please copy-paste the complete stack trace, so that we know exactly which part of the code threw the error?\r\n\r\n1. For the platform, here are the output:\r\n - datasets` version: 1.11.0\r\n - Platform: Windows-10-10.0.19041-SP0\r\n - Python version: 3.7.10\r\n - PyArrow version: 5.0.0\r\n2. For the code and error\uff1a\r\n ```python\r\n from datasets import load_dataset, load_metric\r\n dataset = load_dataset(\"glue\", \"cola\")\r\n ```\r\n ```python\r\n Traceback (most recent call last):\r\n ....\r\n ....\r\n File \"my_file.py\", line 2, in \r\n dataset = load_dataset(\"glue\", \"cola\")\r\n File \"My environments\\lib\\site-packages\\datasets\\load.py\", line 830, in load_dataset\r\n **config_kwargs,\r\n File \"My environments\\lib\\site-packages\\datasets\\load.py\", line 710, in load_dataset_builder\r\n **config_kwargs,\r\n TypeError: 'NoneType' object is not callable\r\n ```\r\n Thank you!","embeddings":[-0.2333926857,-0.2121910155,0.0151841119,0.3540363908,0.431312561,0.0565467253,0.4275258183,0.1668837517,0.1619586647,0.2543876767,-0.1491201818,0.4838864207,-0.032627441,0.2008449435,0.1351143867,-0.1875769347,-0.1780486703,0.1836340725,-0.2075739056,0.0554907173,-0.530339241,-0.0694576576,-0.131100297,0.1142207906,-0.165304631,-0.2441832423,-0.1580201536,0.0987920165,-0.1795856655,-0.3501777351,0.4064702392,-0.1164285839,0.2196373641,0.562065959,-0.0001015193,-0.0459203348,0.4622377753,0.0741885006,-0.1826767772,-0.3177576065,-0.2952965498,-0.3818572462,0.1700716317,-0.3912402391,-0.0481698103,-0.0371986367,-0.0039750608,-0.2519207001,0.2980602384,0.3795799315,0.328802675,0.4538740516,0.0911629125,-0.2488764822,0.1904570758,0.0910028517,-0.0621268973,0.3023785055,0.0538949445,-0.0567670763,0.2113882601,0.1328090131,-0.3628790081,-0.0556369275,0.2719312906,0.0631388128,0.0336346067,-0.3873294294,0.0748671964,0.3000453711,0.4482415617,-0.3896304667,-0.2469754517,0.0235931426,0.1223488674,-0.1994509548,0.2123644501,0.0509505086,0.0008063291,0.0616286434,0.0196442436,0.1935778111,-0.1930177808,0.2497536093,-0.1554912627,0.1598448157,-0.1672741175,0.1182282194,0.0708542541,-0.2033922374,0.049795296,-0.2369977683,0.007765641,0.1136214361,-0.2622630298,-0.0578037463,0.1595869213,0.1328142434,0.0924718305,-0.2031244338,0.0840387568,-0.0385783762,0.0005321783,0.3734622598,0.2657110691,0.2719011307,0.2630656064,0.1053493395,0.3113079071,0.0968508199,-0.1035405621,0.0171599817,-0.1741403788,-0.1962881684,0.2919623256,0.27147156,0.5614536405,-0.0071394164,-0.4003361762,-0.112911351,-0.1762718856,0.0414780118,0.1914937347,0.3488394022,-0.1659963131,0.1930406094,0.2003636211,0.1248331815,-0.2139800191,-0.2602629662,-0.2342291027,0.277436316,-0.2562621832,-0.1992109269,0.1157746315,-0.0126312999,-0.0741725713,0.0112992479,-0.1771455258,0.1546337306,0.1403729916,-0.3297455609,-0.1945308298,0.1805401146,0.0513517223,0.0468542874,0.3396641314,-0.3767827153,-0.1449093074,0.0857277736,-0.2972788513,-0.1406729817,-0.1555437893,0.3137725294,-0.203810975,-0.0596022606,-0.2211033255,-0.0271769408,0.1290350109,-0.1592079401,0.0224947892,-0.1568120122,-0.0343936272,-0.4068643153,0.0919899419,0.6147862673,-0.4071787894,-0.1555482447,-0.3654867709,-0.1499285698,0.3206854463,-0.2192046493,0.0420476161,0.205342114,-0.1884394288,-0.0670418441,0.5587269068,-0.4109594524,-0.2658071816,0.071330905,-0.1287924349,-0.0868593082,0.0406506956,-0.0299981683,0.0722305849,0.1137080416,0.4100970328,0.275131911,0.0011142003,0.0079012448,-0.1864448488,-0.2357116938,0.1991765499,0.0649753883,0.1384467632,0.1023558527,0.1406214237,-0.1971790642,0.0246631335,-0.0487639643,-0.0327245519,-0.057789661,0.5558087826,-0.0468545221,-0.044843737,-0.5518350005,-0.4227520227,0.1897544414,-0.0799524039,0.3541664779,-0.1335036159,-0.1872806996,-0.3264653981,0.0583028831,-0.1389628798,-0.0200660415,0.3176531494,0.1545748264,-0.1688733399,0.0288684927,-0.0564694032,0.2119377255,-0.0756313056,0.0736123845,-0.14293015,0.3080175817,-0.057636708,-0.3339879513,0.0377683677,0.2538016438,0.5117610097,-0.0004488377,-0.0633197874,0.3244804144,0.0866748691,-0.0496543907,-0.1650345176,-0.0999800116,0.1509415507,-0.2289367616,0.0971660241,0.0872650221,0.3168590069,-0.0662048236,0.0954951942,0.4775231183,0.0889250115,0.2869647741,-0.0390479192,0.2417568564,-0.0411586203,-0.0287830383,-0.0849919841,-0.1743532121,0.0761678517,0.1894836426,0.2709217072,-0.0320920832,-0.0377041772,0.0921959504,0.5580118299,0.0443215631,0.3538882732,-0.0804426596,-0.3818507195,0.0884140804,0.1868627965,0.2703929543,0.4774660766,0.3163056076,-0.1177275926,-0.0135746626,-0.0687375143,0.068971917,0.0899053961,0.0000686368,0.2182746232,0.0326346494,0.2071536779,0.0608280376,-0.2113714963,-0.0529055037,-0.1733764559,0.2231471837,-0.4077748656,0.1184415296,-0.3216187358,-0.1545066088,-0.0203860439,0.0009128465,-0.1665667146,-0.3700627089,-0.2923040688,0.0090758018,0.0586140528,0.1355309337,0.0729432553,0.0848084092,0.0303801391,0.0038109233,-0.1432915479,-0.1506637037,-0.197236672,0.1309271753,0.1724003851,0.2035219669,0.4579099715,-0.1723938286,0.1496831626,-0.0668234155,-0.2592738867,0.0691069737,-0.0358921215,0.5192118287,0.4751577079,0.1759489179,0.0373363346,-0.0723712072,0.3917131424,-0.337565273,-0.0363603607,0.2813171744,-0.2262491435,0.1912381798,-0.1627744734,-0.2709549367,-0.5850613117,-0.4314647317,0.0900498033,0.1581933945,0.1574169397,0.2290733755,0.1376304924,0.4989827573,0.0554460995,0.173627466,-0.1574210674,-0.0517017953,0.2869262695,-0.128268525,-0.3188622594,0.0627037063,-0.1610585004,0.2123719454,0.0095064696,-0.2091803849,-0.5421624184,-0.1937738359,0.248898387,-0.0088543072,0.1406644732,0.2820861042,0.1652371287,-0.171896413,-0.2008978277,-0.1451961696,-0.0958748311,0.2647386193,0.003278581,0.1268489063,0.3415829241,-0.1613709927,0.1591750383,0.0212520342,-0.1993087977,0.414999783,-0.2640288472,0.3553538322,-0.1553532481,-0.501110673,-0.1315999925,-0.034155868,0.0119620711,-0.0257774498,-0.1498845816,-0.1616336107,-0.3645355999,-0.1962905228,-0.4402103126,-0.1943258345,-0.0451902971,0.199670434,0.0204199702,-0.0293068476,0.1270928979,-0.4808782339,0.0326096974,0.0426807031,0.2218273133,-0.0436264426,-0.1767756939,-0.2460242808,-0.2401109934,-0.2647053599,0.4546621442,-0.158620581,0.0367192812,-0.1481581628,-0.0604273044,0.0030126467,-0.0713322535,0.5519664884,-0.3454786837,0.167946592,0.5127314329,0.1982301772,-0.3972994685,0.1448308527,-0.3467212319,0.197871536,0.0658447966,0.3129544258,-0.3585754633,-0.0375085883,0.2515778542,0.2019121051,-0.129298076,-0.1740512401,-0.1447782964,-0.2503760755,-0.1316992491,-0.058486212,-0.1246138215,0.3594864011,-0.1265377104,-0.1583637297,-0.3093515933,-0.0301203411,0.1789499223,0.1840766668,0.2802935541,-0.0759279355,-0.0367612876,-0.0464058407,0.071699515,0.2319861799,0.3442593217,-0.0239473321,-0.360599637,0.0667763501,0.0140341548,0.2356529385,0.0585258938,-0.1843719929,-0.1035960093,-0.0533856153,0.200205341,-0.1060454622,0.0749847591,0.3068714738,0.0640616715,-0.3982105851,-0.3787680268,0.4075770378,0.1482772082,-0.055876337,0.2646415234,0.2493728548,-0.3079307079,0.2636658549,0.0765032694,0.9972296357,0.0634360686,-0.1427229047,0.5291199684,-0.0992499292,0.5708351135,0.0216645449,0.0737809911,-0.3225640059,-0.2403181046,-0.0621667355,-0.1458812952,0.3637820482,0.0177567303,-0.554625988,0.2570112944,-0.4040381014,0.0338811316,-0.0993947461,-0.0540791601,-0.168008253,-0.1555937231,-0.4221657813,0.2615644336,-0.1391600221,0.1646663249,-0.0964661688,-0.0559898913,0.0007044189,-0.1505103707,-0.0544889867,0.0260486845,-0.470529139,-0.092832163,-0.1959897131,-0.1402023137,-0.0150172981,0.2117722034,0.1081975773,0.0798802003,-0.2208486944,0.074763,0.1003000066,0.1113611236,-0.0745936707,0.0186240617,0.2860429585,-0.0365927033,-0.3421170413,0.145158276,-0.0346250087,0.1750632823,0.3781581819,0.0789964721,-0.0583959855,-0.3603264987,-0.1006426886,-0.0581629425,0.2370610982,-0.2568027377,0.2513565421,0.0545558929,-0.0355337635,0.0711976513,-0.1054593101,-0.0828102529,-0.0973028988,0.3941631317,0.2912923396,0.2772480845,0.2169934809,0.2351005822,-0.2244954705,-0.3636664152,0.2889499366,0.0990770832,-0.3086126447,0.198417455,0.2202891111,-0.0483231917,-0.0357774086,0.3472393751,0.0114345681,0.049836304,-0.1350132227,-0.2807761729,-0.22253865,0.2674214244,0.0906429961,0.0844058841,0.028326951,0.1621989608,0.1774459928,-0.0547243915,-0.451469332,-0.011736189,-0.3375882506,-0.0379747562,0.1141254455,-0.0688912645,0.2117233276,0.0469736308,0.266826123,-0.027095424,-0.095401898,-0.3234717548,-0.1007948369,0.1007452458,-0.0319782905,-0.0041054767,0.2243584841,-0.1309217066,-0.1626257896,-0.1651719809,0.1755297929,0.0577325188,-0.1902176887,0.1038779467,0.1569740027,0.2834070921,0.0430249199,0.2171774507,-0.0266571995,-0.0581222437,-0.0226649251,0.0391197726,0.0156680252,-0.0498561636,-0.0160412788,0.0438439623,0.0941356197,0.1154540554,0.1023316532,-0.1328705847,-0.2397717088,0.1010522768,0.1707957089,0.1672703177,-0.1169895381,0.2498272508,-0.0852398649,0.343233645,-0.3024182916,0.0496240407,0.1867102683,-0.1545248628,0.2260235697,0.1156060994,-0.3048253357,0.2245938927,-0.14696531,-0.0125416005,0.4131203294,-0.1344489008,0.2140961289,0.3211934268,-0.0835602805,-0.0166897029,0.3776884973,0.1835822165,0.1086514145,0.0316770524,-0.0579561144,0.1470866799,-0.0196429528,0.0855417624,0.2414101213,-0.5050227046,0.2710326314,0.2405994534,-0.0243101902,-0.1121373996,-0.163989529,0.3712483943,-0.5745438933,-0.0636712015,-0.3064317703,0.0594137311,-0.1685567945,0.1159525663,-0.1114942878,-0.1400388479,-0.256551981,-0.2784867585,0.06332881,-0.2414887398,0.2439766079,-0.1118730009,0.0211273301,-0.2714125514,-0.0401527472,0.2217802703,0.1276277751,-0.1941972971,0.0097096395,0.0506009683,0.2041798085,-0.1806842834,0.552272737,0.3988933563,0.3800224066,-0.1354636401,-0.0664787814,0.1149364188,-0.2638256848,-0.0904527828,0.0105616841,-0.0362640731,0.2590774,0.1043943912,0.3203140199,-0.3022005558,-0.0068912487,-0.0190034341,0.0057310155,-0.2694809437,0.3833438158,-0.2941297293,-0.1339461505,-0.1130259633,-0.0764845833,-0.5723281503,0.0453014746,0.4337003529,0.0553493984,0.1160499007,-0.1790060252,0.1447362602,-0.1381019503,0.1920721382,-0.0137894452,0.2897506058,-0.266952455,0.0805813819,-0.5247184038,0.3150629103,0.0391136967,-0.1924300641,-0.0284713041,-0.1205905452,-0.1418034434,0.0033041995,0.0252633821,0.2740843296,0.2506788373,0.0156387016,-0.544901669,-0.293649137,-0.0416166857,-0.1840641648,0.0557728298,-0.3411940038,-0.002358932,-0.1740486026,0.1497212201,0.0006315968,-0.249928534,0.1440658867,0.0977682769,0.6802558899,0.1865135878,0.4030945301,-0.0869980007,-0.0436691344,0.0363891684,-0.2932219505,-0.1100367904,0.2340059429,0.0753232017,0.2027742267,0.0346169695,0.1456870288,-0.2451830357,0.0854274631,0.0174441207,0.1328043193,-0.1475849748,0.3109081388,-0.1509093344,-0.2978287041,-0.1912623942,0.0007492236,0.1050974503,0.1029956341,-0.069436349,-0.2385252118,0.5177173018,-0.2855589986,-0.1767610461,-0.0903273001,0.3520115018,-0.0403811969,-0.1172361895,-0.5067312717,0.054644987,0.4887304604,-0.002228745,-0.1500933021,0.1638828069,-0.2067309022,0.0215018056,-0.0706165358,0.0114977323,-0.0409061126,0.000829379,0.1980810612,-0.1320041418]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"For that environment, I am sorry but I can't reproduce the bug: I can load the dataset without any problem.","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":20,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\nFor that environment, I am sorry but I can't reproduce the bug: I can load the dataset without any problem.","embeddings":[-0.2608940005,-0.2007066011,0.0660000592,0.400195688,0.4645994604,0.0115246382,0.3240968585,0.0941318348,0.1896000355,0.3001818657,-0.2849094868,0.4667589068,-0.0366437919,0.1668581367,0.0777649358,-0.1253727525,-0.2606472373,0.2636680901,-0.1777470112,0.0613667145,-0.5233634114,0.0231986102,-0.2373635471,0.0858903304,-0.2768001556,-0.1533993483,-0.0281501114,0.079713583,-0.2462176234,-0.2805505991,0.4611611068,-0.0159628261,0.1640476584,0.5104134083,-0.0001023479,-0.0184572823,0.5836995244,0.0530032516,-0.0761517882,-0.4211900532,-0.2255626768,-0.3977999389,0.2870038748,-0.3548148572,-0.1275422126,0.0055780676,0.0405133776,-0.1819047779,0.262139827,0.3204336464,0.3242153823,0.5296897888,0.0114422301,-0.3599714339,0.1965120882,0.0808303058,-0.0636900738,0.3710546196,0.0265845973,-0.0474929884,0.2439645529,0.0870305225,-0.3353538513,0.0527608916,0.2847363651,0.0382548533,0.0095302081,-0.270170182,0.0964757726,0.3694062233,0.3855828047,-0.382620275,-0.1514433473,0.1750092059,0.2239649594,-0.2032708824,0.211649254,0.0468775555,0.0163049828,0.0907445624,-0.0130718844,0.1712200791,-0.2445577234,0.2749375999,-0.2020063251,0.1065498963,-0.1935495585,0.1447966695,0.0849972516,-0.1188601851,0.0763338953,-0.1345976889,-0.0619969517,0.0890901908,-0.2083802223,0.0349296518,0.1152055487,0.1535798907,0.1636114269,-0.2843680978,0.1018852517,-0.0238335412,0.0397984199,0.4501174688,0.2414609641,0.1897560805,0.2724782526,0.088692598,0.2660125494,0.066787757,-0.2417135537,-0.0120890485,-0.1543584764,-0.0610723533,0.2427943051,0.1411847323,0.5124645233,0.0189152192,-0.4501633942,-0.1742377281,-0.172133401,0.0913005546,0.2716411948,0.4277633727,-0.1526402384,0.0820783153,0.1332417876,0.1743281931,-0.1884514242,-0.2918041646,-0.3183253706,0.3539678454,-0.2528480887,-0.2349549532,0.1911386847,0.0579657778,-0.0449788868,0.0134671535,-0.0958473086,0.1027007028,0.0886686146,-0.2908604145,-0.2310647517,0.3132710755,0.2443055063,-0.0628327206,0.2800607383,-0.4668283463,-0.091596745,0.2292028666,-0.2898909748,-0.1671735048,-0.0732437372,0.2667024732,-0.3390546739,-0.0697349459,-0.2441662848,0.0466724299,0.0489577651,-0.1566343755,-0.1024087444,-0.0962615982,-0.112656869,-0.4234251976,0.0446183719,0.5165780783,-0.3751866817,-0.1992407143,-0.3078093231,-0.1091819853,0.352284044,-0.147301212,0.0145670017,0.0629001632,-0.1706566364,-0.0006857364,0.6008648276,-0.3847208321,-0.2815137208,0.0563009419,-0.0371869579,-0.1165539101,-0.0517222472,-0.0470438264,0.0718874335,0.135642916,0.4128215313,0.2446535677,0.0493588299,-0.0324675739,-0.2109587342,-0.2381093949,0.1280453652,0.151658386,0.1272704005,0.1630408317,0.0712505579,-0.2506299913,-0.0557164401,-0.0061886362,-0.1022448465,-0.0301638599,0.5365030169,-0.1599618644,-0.0399641804,-0.532881856,-0.3950176537,0.2330813259,-0.0779283196,0.2541694939,-0.105489023,-0.1700357646,-0.352982074,0.1019510254,-0.0170240067,0.0607618615,0.2853280008,0.0909614265,-0.2011494935,0.1454881579,-0.0809026584,0.1637727916,-0.0873494446,0.0512318462,0.0335052349,0.291667223,0.0747782141,-0.3356175423,0.0352951288,0.1897852123,0.4574481249,-0.0104598496,-0.0672703311,0.2640901804,0.070892185,-0.0788056329,-0.2070860565,-0.05381817,0.0879925042,-0.313064307,-0.0303661674,0.0379851758,0.3187099099,-0.0943738967,0.0625385866,0.6038931608,0.0668431222,0.1744620055,-0.0730603263,0.2714183331,0.0653238744,-0.0385558642,-0.149077177,-0.1286280155,0.0459849499,0.1176183671,0.2486371994,0.0061038891,-0.0880215764,-0.0231885035,0.5769147873,0.0913832113,0.3607913256,-0.1375178844,-0.4204133153,0.1137951836,0.0370061696,0.2525543571,0.5032669902,0.332315892,-0.1068636402,0.023629915,-0.0571720079,0.0404726267,0.1032921076,0.0494471341,0.2328974009,-0.0000240203,0.3057212234,-0.0209420472,-0.3462260365,0.1613915414,-0.053847719,0.1842062026,-0.3760870993,0.1845387518,-0.2747030258,-0.076757893,0.1425513774,-0.1027247906,-0.125005424,-0.3286290169,-0.2664072216,0.1081375405,0.1210905612,0.1938206255,0.1422476321,0.0763199255,0.0087453714,-0.0255289134,-0.1826611906,-0.1326549202,-0.1483921856,0.0880739987,0.1653355807,0.181941092,0.4119828641,-0.1496375948,0.2284882814,-0.0511268564,-0.245598793,0.100177817,-0.0573291741,0.5299291015,0.4086084068,0.0602920428,-0.1410644501,0.1004512459,0.4188910425,-0.4658019245,0.0264790654,0.363583684,-0.3070645034,0.0837264657,-0.2518776655,-0.1060993373,-0.5478299856,-0.4280380607,0.0683310926,0.0958201066,0.0449768677,0.172417596,0.1191176772,0.444096446,0.112392433,0.0157600436,-0.2330423892,-0.0854077637,0.3311132789,-0.1693256646,-0.3608907759,0.2042488754,-0.1323478222,0.2548744977,0.0285864808,-0.1609783471,-0.5243912339,-0.1556174606,0.2053533643,-0.005188724,-0.0110916914,0.2912253737,0.0402853899,-0.1367866844,-0.2726752162,-0.1413018703,-0.0488153882,0.2352696806,-0.0683882236,0.0471838973,0.2825388908,-0.1223018318,0.1436379254,-0.1094394699,-0.1853011698,0.4059675932,-0.3138792515,0.3742715716,-0.1514657587,-0.4665456712,-0.2108906806,0.0278709959,0.0422425531,0.0484344922,-0.1828998625,-0.2429656833,-0.3873915374,-0.184057802,-0.3091237545,-0.1176473424,-0.0448231548,0.1869143099,-0.0344660766,0.1282412708,0.1506693661,-0.4099556506,-0.0440214574,-0.0145162204,0.1907593459,0.0003334501,-0.1874935776,-0.1343231499,-0.2394137084,-0.3349730968,0.466514647,-0.130417794,-0.0214519892,-0.1048111022,-0.0228138175,0.1187153012,-0.0458018035,0.5801782012,-0.2944788039,0.1719288677,0.5542984009,0.0904794037,-0.5516597629,0.1077317894,-0.331844449,0.3005692363,0.1331286132,0.2810209692,-0.3470429182,-0.0193871576,0.2158617675,0.0685858876,-0.1151087955,-0.2407927513,-0.2334081084,-0.2933899462,-0.0492701679,-0.0665732101,-0.0586115457,0.4225768149,-0.1354665607,-0.1273059845,-0.4123404026,-0.0367351063,0.1656020731,0.1939739734,0.3422520459,-0.1359105408,-0.0259811543,-0.0587919764,0.0830501467,0.1330539882,0.3568017781,0.0888106078,-0.4469739497,0.1390765905,0.0736007243,0.2392614037,0.0024299552,-0.2127716392,-0.1564223021,-0.0973184034,0.1256743819,0.0694909394,0.1411263794,0.2321473658,0.0137663297,-0.3646833301,-0.23113361,0.4325918257,0.1745893359,0.0250268523,0.2719003558,0.2346458733,-0.2254342437,0.2237933278,-0.0649300292,0.9164009094,-0.0016310633,-0.1102356687,0.6165456772,-0.0979906321,0.5674658418,0.0650797188,0.0679337382,-0.3704264462,-0.2879863977,-0.0052313902,-0.1453758329,0.2548450828,-0.0493531264,-0.5216739774,0.2098066807,-0.3250356913,0.0563870929,-0.0338756666,-0.0862727463,-0.0688716248,-0.1625674218,-0.3169096708,0.2203887552,-0.1160019636,0.1652479023,-0.0905514881,-0.0729376525,0.0201872438,-0.0793530717,-0.0859890431,0.0518678725,-0.4105747044,-0.1025814638,0.0088864788,-0.1489424556,-0.049235668,0.1097768545,0.0439088494,0.0238108411,-0.3483451903,0.2360750288,0.0525084697,0.0174742565,-0.0742553324,0.0752038509,0.2092088163,-0.0593716875,-0.2678528428,0.1072407961,-0.0471316501,0.167549476,0.3512031436,0.1351451129,-0.1057788879,-0.3018784225,-0.2150235772,-0.1387380958,0.2945331931,-0.2950992286,0.2529356182,0.0883854032,-0.1128398553,0.0237117633,0.0370561406,-0.1807857901,-0.0391988792,0.4054492116,0.3315434754,0.3266708255,0.1742117852,0.2139461488,-0.1330504417,-0.3562078476,0.2356424928,0.07449314,-0.3519974351,0.3284549117,0.2739292681,-0.0456543639,0.0271229651,0.4190958738,0.0177649334,-0.0406256802,-0.1278355718,-0.2977449894,-0.2052850723,0.2201710194,0.0076375799,-0.0133153005,0.1036312953,0.2353747338,0.1969152242,-0.0753166899,-0.4256239235,0.1184609309,-0.3420971334,-0.1064495519,0.0358140729,-0.0839432031,0.1991972327,0.094303377,0.2200210541,0.04124403,-0.1173671484,-0.3198773861,-0.1708855778,0.11573001,-0.0603523888,0.0984599441,0.1912232041,-0.0978017226,-0.3399778008,-0.1138171703,0.0863472894,0.0186832566,-0.2485669553,0.0389387496,0.1841407865,0.2960723639,0.0243147407,0.2214778364,0.0648288131,-0.0516980924,-0.1667803824,0.0070171691,0.0738608539,-0.0377257206,0.0453988388,-0.0242696367,0.146423474,0.1258281916,0.1472238451,-0.1821626425,-0.1376380622,-0.0106543619,0.1622682214,0.1054019704,-0.1678972542,0.3169001639,-0.151139766,0.3191595078,-0.3713175952,-0.0181264002,0.2591815293,-0.0037934883,0.2337247133,0.0487347431,-0.2177034914,0.2784507871,-0.100174889,-0.0086973989,0.4537391067,-0.1244415864,0.0648771003,0.238508001,-0.1500530988,0.0674678758,0.4041028619,0.2606096864,0.0971257016,0.1352504194,-0.0299772583,0.1296595037,-0.0773413926,0.0511166528,0.1697779894,-0.5580494404,0.186106354,0.1685796976,0.0023652951,-0.0791588277,-0.2323796898,0.4102463126,-0.4937512875,-0.0605658107,-0.3794310689,0.0643697307,-0.163028717,-0.0013760449,-0.1433874965,-0.1149196997,-0.1717617661,-0.2528249323,0.0416352265,-0.0990594178,0.2412425876,-0.1137952805,0.1046868265,-0.1930136085,-0.0399430357,0.278688699,0.0477989689,-0.1555451304,0.0398130342,-0.0601673685,0.25269261,-0.1166727543,0.4984675646,0.4282827377,0.265019238,-0.2017355114,-0.1184142306,0.0438968129,-0.294929415,-0.1482043564,0.1448875815,0.1621525288,0.1861361414,0.0748439059,0.2893256247,-0.284999311,0.1308235079,-0.0036748131,-0.0078192856,-0.2805196047,0.3590799868,-0.3625504673,-0.1178260595,-0.1554860473,-0.1072709784,-0.4793089926,0.098795183,0.3647942543,0.1218607798,0.1348394006,-0.1550246477,0.1431170255,-0.2600436211,0.2424018681,-0.0232327823,0.3653592467,-0.2398388982,0.0257782973,-0.5430261493,0.4481772184,0.0129255764,-0.0782645717,0.0848244727,-0.121702157,-0.1154602692,-0.0344456807,-0.0048418576,0.3630883992,0.2708312571,-0.0477743857,-0.5591887236,-0.3568104208,-0.0946177766,-0.2772197127,0.0461378619,-0.2907800674,-0.0501833335,-0.1604976356,0.1326508671,-0.0683541149,-0.2736573815,0.1152127311,0.2708913684,0.6669099331,0.2045687437,0.407979399,-0.0874733254,0.0932615772,0.07337749,-0.2984471619,-0.0754200667,0.1905672997,0.1010607332,0.136260882,0.0150285382,0.1734221131,-0.1472680867,-0.0376885645,-0.0404813625,0.0578768551,-0.2393028438,0.313811034,-0.2897048891,-0.1945980787,-0.2668882906,0.0952340364,0.084310323,0.1096617356,0.0218840633,-0.0939410552,0.6466474533,-0.2551367581,-0.3148619533,0.002964878,0.2552378476,0.0340935066,-0.1752844602,-0.4145071208,0.0628261641,0.443211019,-0.0549863242,-0.218359828,0.2427147627,-0.1774560511,0.0503365546,-0.1022099629,0.0111536505,0.0516408011,-0.0675834417,0.1296829134,-0.0830133334]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"One naive question: do you have internet access from the machine where you execute the code?","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":16,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\nOne naive question: do you have internet access from the machine where you execute the code?","embeddings":[-0.2466751933,-0.2166070193,-0.0038837458,0.3586860299,0.4009475708,0.035536956,0.302520901,0.1521895081,0.1650357544,0.3566101193,-0.1779196858,0.4452134073,-0.0382566825,0.1990447938,0.0816193596,-0.0899617374,-0.2033715993,0.2447062582,-0.1842225194,0.010521763,-0.472876668,-0.0185202286,-0.1821441352,0.0199327096,-0.2480738759,-0.2002862543,-0.1163984835,0.0645010024,-0.2148991227,-0.2530295849,0.3632104993,0.0450640544,0.112367928,0.5467213392,-0.0001019842,-0.0862105861,0.4975403249,0.0762401372,-0.1133384928,-0.2227149755,-0.2704135478,-0.3431518078,0.1823247373,-0.4260804653,-0.0916544795,0.028890904,0.0123812426,-0.2481914461,0.3053150177,0.3072139323,0.339708358,0.4426216781,-0.0190175083,-0.3322893977,0.2162288725,0.0421040244,-0.0323108882,0.362215519,0.0633602291,-0.0537314415,0.1767412126,0.1824111044,-0.2867747545,-0.0029550304,0.2504888773,0.0876018703,-0.0245356578,-0.2595610619,0.1022056341,0.2668823004,0.2856886089,-0.4520843923,-0.1245521009,0.1114801839,0.1043578833,-0.2367105186,0.1430763751,-0.0036112065,-0.0197007451,0.1060361415,-0.0351593755,0.133691445,-0.231664747,0.2396647781,-0.209117204,0.1492936164,-0.1894275993,0.0967105776,0.0752299801,-0.0852814317,0.0532211438,-0.1713415682,0.021075597,0.0105691114,-0.1507948637,-0.0472187474,0.1830180287,0.1402942985,0.1534052491,-0.254579097,0.1078871489,-0.0058753905,0.1186242029,0.4159373343,0.154383108,0.2415641695,0.2150634825,0.0880800635,0.2187133878,0.1302518398,-0.2568061352,-0.0115337186,-0.0923354477,-0.0363530554,0.3084871173,0.1605716944,0.4978731573,0.0148101076,-0.43729496,-0.0482422002,-0.2314789593,0.1063998938,0.2851623595,0.3828612864,-0.2197893709,0.0280763004,0.1954128742,0.1463474184,-0.167907998,-0.2896850407,-0.3228367269,0.3451814651,-0.3253240883,-0.1996293813,0.1255358905,0.0956025869,-0.078586556,0.0480372384,-0.035712216,0.0553010367,0.0763480887,-0.2473653257,-0.1706801951,0.1856051236,0.2096802443,-0.1042110175,0.3234429955,-0.4836687446,-0.1027234495,0.1058211848,-0.2070778757,-0.0896052122,-0.0359199308,0.3180175722,-0.3513986468,-0.1008676514,-0.151536569,0.1008075848,0.0612458289,-0.1549359858,0.0268286895,-0.0838796198,-0.0460206904,-0.443877846,0.0987593383,0.5773766041,-0.3923660815,-0.1583535075,-0.3149743676,-0.0410091579,0.3569568694,-0.2316169292,0.1328972578,0.1612990648,-0.1505951583,0.0307376627,0.5494579673,-0.3594247699,-0.298289597,-0.0194503367,-0.1195071116,-0.0845604688,-0.0080690281,-0.1045471951,0.0758646503,0.1502961367,0.4117963314,0.2450776398,-0.0244919565,-0.0403468348,-0.2356397361,-0.2162138522,0.0301526804,0.107242316,0.1855029613,0.160957098,0.1086669937,-0.2254877985,0.0150828492,-0.0165604558,-0.0678272471,-0.0664632171,0.5562682748,-0.1198019162,-0.1030108109,-0.548601687,-0.3364979625,0.2137777209,-0.0644096881,0.2583818138,-0.1622745097,-0.1771243662,-0.4169964492,0.1173746884,-0.0591691211,0.0415956117,0.3479426801,0.1584892571,-0.1316381693,0.1201645657,-0.038923569,0.1554773301,-0.0975586995,0.0615419187,0.002867274,0.3624258637,0.0288160499,-0.3402500749,0.0125437025,0.2390864491,0.492028594,0.0251288041,-0.0366603509,0.3425714076,0.0153975571,-0.0589240976,-0.0713611618,0.0129818767,0.0355378203,-0.3102554679,0.0903075784,0.0740944669,0.274332881,-0.0501489863,0.0455586836,0.5526733398,0.0639084801,0.1631180197,-0.0774783418,0.2747522891,-0.1043333039,-0.0652063265,-0.1591193974,-0.1482455432,0.1165027544,0.0529977977,0.1941261292,-0.0445721783,-0.102411747,0.0711867437,0.6148303747,0.07447882,0.410061866,-0.1350679696,-0.3626484573,0.1363148987,0.0632819161,0.2627360523,0.4373947978,0.3606951535,-0.0552522913,0.0725060552,-0.0956002772,0.0463584624,0.0818468109,0.0780013651,0.1710066646,-0.0235593766,0.2648818195,0.0000485086,-0.302197367,0.0393880606,-0.2115725577,0.1999685317,-0.3442276716,0.2703234255,-0.2761572301,-0.1083705425,0.0665704906,-0.1341651529,-0.1119984984,-0.3972260058,-0.189691484,0.0495734848,0.0846979097,0.1596637964,0.1331482828,0.1780963689,0.0435850658,-0.0506396033,-0.2169364691,-0.0929441899,-0.1491832733,0.124519974,0.1127889976,0.2195498496,0.4563970566,-0.2264528573,0.2653894424,0.0405263342,-0.2521411479,0.1091040745,-0.1113009229,0.5382031798,0.4217015207,0.0849712491,-0.1375489533,0.0479233935,0.4025655687,-0.3987693489,0.0345607921,0.2760668993,-0.1733168364,0.0993671566,-0.2214981169,-0.1091302112,-0.4937289059,-0.430516243,0.0535355024,0.0940166637,0.0907848328,0.1661356688,0.1647697985,0.5161489248,0.0903599858,0.0853421465,-0.2309448719,-0.0928328782,0.2652721107,-0.1356292665,-0.3170882165,0.1164372712,-0.1327296942,0.1882433891,0.0346632898,-0.2173748314,-0.5696399808,-0.1761069298,0.2179042101,-0.0343912281,0.0933210403,0.26999861,0.0949118435,-0.1569222659,-0.2845886648,-0.1625180691,-0.09149611,0.1541405916,-0.0318299644,-0.0249708015,0.2786458135,-0.1161085069,0.1320557296,-0.0665605962,-0.2075443119,0.3825691044,-0.3480911255,0.3984667063,-0.0996405408,-0.5200996399,-0.1736067683,0.0353287831,-0.0609325208,0.0489658341,-0.1714958549,-0.216622591,-0.3766874969,-0.0562461317,-0.3627006412,-0.1873274595,-0.0391595475,0.1560379118,-0.0368834846,0.0749153569,0.1888145953,-0.4491058588,0.0570053346,-0.0169513598,0.1537907422,-0.0894060507,-0.1648553312,-0.1185744405,-0.2468067259,-0.2861296535,0.4680056274,-0.1499364823,-0.056393221,-0.2046160102,-0.0660794526,0.0905419141,-0.0358795337,0.5689522028,-0.3503737152,0.1897085607,0.4768269658,0.1251972616,-0.4568893611,0.1350766271,-0.4355596006,0.2517589331,0.2185436338,0.2802282274,-0.3060492873,0.0191356,0.2399570048,0.1026442349,-0.1106582806,-0.2221298069,-0.221788317,-0.3183932304,-0.0803543776,0.0420141369,-0.0252923761,0.4193621278,-0.0983455256,-0.1282833368,-0.3264181316,-0.0256801695,0.2025925368,0.2596056759,0.2894299626,-0.1231492087,-0.1225850731,-0.0823648274,0.0792324618,0.1359493136,0.3235682249,0.0793575421,-0.3446296751,0.0907276347,0.0888884291,0.2781876624,0.0573618561,-0.1762590408,-0.1440508962,-0.0869545788,0.2175625265,0.0061369422,0.1769185215,0.2548579276,0.0897866562,-0.3811888397,-0.2307589352,0.4337727427,0.1766114533,-0.0381243378,0.2948857546,0.2082063407,-0.2366477698,0.226096943,-0.0221611597,1.0133308172,0.0526240282,-0.1635152251,0.5793563128,-0.1488889903,0.5672776699,-0.0199497938,0.0393971093,-0.3318083882,-0.2921088338,-0.0050110044,-0.1491662711,0.2573121488,-0.0939406306,-0.6075484157,0.1944481134,-0.3985604346,0.0609820001,0.0078523858,-0.1274123043,-0.1191684529,-0.1094275266,-0.3458572328,0.2644385099,-0.1378703862,0.1916236281,-0.1486469507,-0.0662537143,-0.081767492,-0.109817706,-0.1274234951,0.0437017307,-0.3669551611,-0.1849412918,-0.0509839356,-0.1615005583,-0.1080818698,0.0942901149,0.0813174099,0.1470959932,-0.2779953182,0.0880830884,0.0845176354,0.024237819,-0.0069385697,0.0085471412,0.1958828121,-0.0370822027,-0.2943578064,0.1068727002,0.0457872525,0.11517708,0.3087917566,0.1035760045,-0.0675774515,-0.3244170845,-0.1034835204,-0.0974463224,0.2481093854,-0.3131102622,0.2714842558,0.0891997069,-0.1201467961,0.0999452099,-0.0129307564,-0.1413194686,-0.0845809951,0.3065213561,0.2475549877,0.3110371828,0.0651139915,0.2337538004,-0.1451797485,-0.3778772056,0.1946107447,0.1266462058,-0.3094281554,0.2451846749,0.1132033244,-0.1002142429,-0.018966902,0.3583429158,0.0872674361,0.0121289548,-0.1319913417,-0.3177200258,-0.2275741249,0.1974657923,0.1262835264,0.0467932597,0.0409467742,0.2409867495,0.2160649598,-0.0149477664,-0.4777274728,0.0233572591,-0.35480088,-0.1210959181,0.0553099439,-0.131202504,0.2106182128,0.0307738371,0.2626608312,0.0611371771,-0.1230163798,-0.3603873551,-0.1196228266,0.0878794044,-0.0960554257,0.1212469712,0.1928565055,-0.0655948073,-0.2737081051,-0.1821471155,0.0655785874,-0.004897011,-0.1561484635,0.0742097273,0.1955157518,0.3057770133,0.0542224348,0.0768087059,0.1181700379,-0.0472166464,-0.0940910056,0.0504168384,0.0114915501,-0.091620788,0.0615547486,0.0473166667,0.1579477042,0.120423086,0.1462675333,-0.1426482201,-0.1761249602,0.0342858694,0.2458388805,0.1684488058,-0.1826538593,0.2451196015,-0.2031995654,0.368602097,-0.3018474579,0.0448667556,0.233446002,-0.0057722093,0.2976733148,0.1492141932,-0.2892270386,0.2251113206,-0.1543214321,0.0022290335,0.4208866656,-0.0848289132,0.0941332951,0.255379945,-0.1617472917,0.0726773366,0.4207955301,0.215765819,0.0808967128,0.1147266775,-0.0120389042,0.2092016637,-0.0595478676,-0.0088517275,0.1725978404,-0.4266685545,0.1862843037,0.1980996728,0.0223718807,-0.0381644145,-0.2197990566,0.415300101,-0.508476913,-0.0565226451,-0.3237141967,0.0487760864,-0.1741153598,0.0605982803,-0.1682130247,-0.1289482117,-0.2265689969,-0.2804172635,0.1312571615,-0.1609176397,0.2799320817,-0.1695612073,0.1022945419,-0.1938088834,-0.0666625947,0.2105489671,0.1612253487,-0.2031766772,0.0029694394,-0.0198347457,0.2243938595,-0.1444881409,0.4721126258,0.4282221496,0.2557366192,-0.1851379275,-0.0929748416,-0.0065720282,-0.2934385538,-0.1710650176,0.0933051482,0.175207302,0.1753773838,0.107457526,0.3313249648,-0.3341614604,0.0941621438,-0.023793228,-0.0138929961,-0.2267360091,0.3832151592,-0.2947542667,-0.0855249688,-0.1702571511,-0.1365145743,-0.5748503804,0.0405889116,0.3419461846,0.0930171311,0.1262921244,-0.1605507731,0.1524770707,-0.1748819649,0.1722018272,0.0148292333,0.4054068625,-0.258834511,-0.0027993866,-0.5410189033,0.4275760949,0.0155638382,-0.1588082761,0.0315136537,-0.0331475437,-0.1146663353,-0.017911721,0.022063138,0.3360385299,0.2262089252,-0.0256580152,-0.5814420581,-0.2901699543,-0.1248456612,-0.2796956301,0.1322259605,-0.2572001219,-0.0358443893,-0.2061264217,0.2160844207,-0.037930768,-0.2939603329,0.0869979188,0.1420842558,0.6847292781,0.177642867,0.4054076374,-0.1641926318,0.0316925645,0.0117772399,-0.2423882335,-0.1033019796,0.2090906501,0.0755653828,0.1736918837,0.0127599053,0.1335804164,-0.1275078207,0.0320263021,0.1104811877,0.0701001137,-0.1159140542,0.2584398985,-0.2325378805,-0.1188334078,-0.1874582469,0.0807324499,0.088373743,0.2076720297,-0.0338687859,-0.1781326681,0.6508178115,-0.1992694587,-0.3754538,-0.0056762015,0.3330244422,0.0251546893,-0.1825997382,-0.4330867231,0.0186706223,0.4222237468,-0.0488338508,-0.2132585198,0.2750219703,-0.2542455494,0.0392030962,-0.023348771,0.0766803846,0.0071151704,-0.0488673002,0.1620789319,-0.1435601711]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2869","title":"TypeError: 'NoneType' object is not callable","comments":"> For that environment, I am sorry but I can't reproduce the bug: I can load the dataset without any problem.\r\n\r\nBut I can download other task dataset such as `dataset = load_dataset('squad')`. I don't know what went wrong. Thank you so much!","body":"## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":43,"text":"TypeError: 'NoneType' object is not callable\n## Describe the bug\r\n\r\nTypeError: 'NoneType' object is not callable\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\ndataset = datasets.load_dataset(\"glue\", 'cola')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform:\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\n> For that environment, I am sorry but I can't reproduce the bug: I can load the dataset without any problem.\r\n\r\nBut I can download other task dataset such as `dataset = load_dataset('squad')`. I don't know what went wrong. Thank you so much!","embeddings":[-0.2479815185,-0.3217719793,0.0763500705,0.4701036215,0.4953816831,0.0185086466,0.3540442288,0.0614947863,0.2402586043,0.3128617108,-0.276409179,0.4540319443,-0.0626192093,0.1033334509,0.1556583196,-0.1557852626,-0.2255954146,0.1679677218,-0.0779853985,0.1020552889,-0.4996086359,-0.0293463878,-0.2632772624,0.1951634139,-0.1559653133,-0.1863708049,-0.0737683624,0.0807150304,-0.1478742063,-0.2537325919,0.5681656003,-0.1083382666,0.2122284323,0.5568339229,-0.0001079977,0.0223296881,0.5475671887,0.0222912077,-0.181262657,-0.5533769727,-0.1755931079,-0.4152663052,0.3201848865,-0.3743572831,-0.1145967916,0.07909403,0.1459887475,-0.1117727757,0.3590354919,0.4540637732,0.2855804861,0.7018626928,-0.0197485741,-0.342117399,0.1333463937,0.0541447364,-0.1285513192,0.2871590555,0.1371662915,-0.0082335789,0.4334596992,-0.0373312868,-0.3705081344,0.090445593,0.3338054717,0.0173116531,0.0456819162,-0.4084408581,0.0989959762,0.3526202738,0.429132849,-0.2789865732,-0.1445523649,0.1062167063,0.2393290848,-0.061995618,0.1765795946,0.0757426322,-0.0298103616,0.0413528346,0.0463871136,0.1718294621,-0.2779862285,0.2125990689,-0.1284300834,-0.0353670754,-0.1932722479,0.2176218927,0.038036257,-0.0971409827,0.0310842935,-0.1765830368,0.0018496352,0.1451958716,-0.3074558675,-0.0370389931,0.1003599092,0.0206406731,0.2188021392,-0.2397266477,0.2141465396,-0.0814069584,0.0014896236,0.3541437387,0.346270442,0.2386882752,0.2158100009,0.1408458501,0.1517539918,0.022729395,-0.1789033264,0.0621397234,-0.2835336924,-0.0460709594,0.2485666424,0.2519290149,0.62191993,0.0226177946,-0.5257906914,-0.1946161389,-0.0697027519,0.0744651854,0.1489258707,0.4044239819,-0.2522181571,0.0734108984,0.0253312159,0.2898434997,-0.1949325502,-0.3826381862,-0.250428021,0.1526836008,-0.268599242,-0.270571053,0.24026227,0.0925443023,0.0312479846,0.0430606492,-0.2270833105,0.0859438777,0.1672883034,-0.3710010648,-0.3183163404,0.203129068,0.2981905341,0.013275166,0.2797108889,-0.3238164485,-0.0456261188,0.2975198925,-0.3877721429,-0.1205919012,-0.1263929754,0.1935602129,-0.3629306555,-0.1565003395,-0.5364289284,0.0805210099,0.0401180312,-0.0892851651,-0.0573687814,-0.1425406635,-0.0697560161,-0.3583665192,-0.0551820286,0.4761920869,-0.2147590965,-0.2288669497,-0.3011833727,-0.2443191111,0.3596111238,-0.1455561668,-0.0476030186,0.2456554323,-0.1385620087,0.0195395816,0.642037034,-0.5051142573,-0.2810824811,0.1532612592,-0.1181739345,-0.076440759,-0.0949228108,-0.0446148701,0.2863956094,0.0758701265,0.3716195226,0.3349784911,0.0008509188,-0.0305648036,-0.131107524,-0.2105619609,0.2909449637,0.1949024647,0.2361101955,0.2211598307,0.0545335151,-0.1998473704,-0.0358049236,0.0527693219,-0.0813176557,-0.0201197267,0.4355514646,-0.0436968356,-0.0257663354,-0.5290393233,-0.5686805844,0.1899500042,-0.0841612145,0.1826199889,-0.115729481,-0.0673960596,-0.3040428758,0.042624291,-0.078021042,-0.0217457879,0.182378903,0.0794131309,-0.2668655217,0.0558810793,-0.0770036429,0.1909017116,-0.1526100785,0.095781371,0.1212609336,0.2611952722,0.1016982198,-0.3987404406,-0.0017724648,0.1362133324,0.4781362414,0.0183626059,-0.0826574564,0.1729604006,0.0458101444,-0.1020453051,-0.2171081603,-0.1762046367,0.0575558543,-0.3366027176,-0.0447619446,0.1347156912,0.3460848331,-0.160481751,0.0836437568,0.4571962357,-0.0167928878,0.3421861231,-0.1437793821,0.2406296581,0.0525164865,-0.0606726483,-0.1331075728,0.1804715544,0.1290700138,0.0913747177,0.2827326059,-0.0355854295,-0.2706822753,0.0046682684,0.6162089109,0.0299597997,0.3194147646,-0.0831516907,-0.33788234,0.1462022662,0.0000325413,0.0418264307,0.5353927612,0.2632282078,-0.1103599295,-0.0164734125,0.0007065669,0.0515690558,0.0333776362,0.069090277,0.2856582701,0.0461304933,0.2759691477,0.0386635847,-0.2005325556,0.1444482505,-0.0084361024,0.1951348335,-0.3969622254,0.1878786534,-0.1423569769,-0.1165724248,-0.0304603409,0.1370230019,-0.1042248011,-0.1990451217,-0.3134959638,0.1589778066,0.2137130797,0.1265260726,0.1616361737,0.0088405497,-0.0113126803,-0.0408857763,-0.1634947658,-0.1458583921,-0.1861838251,0.0462163053,0.1946974099,0.0735260546,0.3495370448,-0.1900552809,0.1245678738,-0.1000485346,-0.1240719408,0.0189164113,-0.0774424672,0.5409654975,0.2683835626,0.2594893575,-0.162481755,0.1194115579,0.3791869879,-0.4782688916,-0.0512078144,0.3330359757,-0.334943682,0.0967640132,-0.2121215761,-0.3186377287,-0.5942174196,-0.3938197494,0.2054920793,-0.0129524982,0.069686912,0.3293658495,0.1026407331,0.4716491699,0.2177550197,0.0916105509,-0.1844632626,-0.0342024155,0.2794252634,-0.1346675754,-0.3884647787,0.1575419456,-0.0971030593,0.2852425575,0.0799094364,-0.1245648041,-0.2630615234,-0.0669115633,0.197574988,-0.1246708706,-0.1207072735,0.3231537342,-0.0845761523,-0.0469946526,-0.2115352452,-0.1213957146,-0.0482246727,0.3319038451,-0.0071725766,0.0945167914,0.3415936828,-0.1624031961,0.3309932351,-0.0868292972,-0.0393967107,0.4093084633,-0.3032317758,0.2776997089,-0.0965931192,-0.4811325371,0.0073162122,0.0563486181,0.0693825185,0.0789806917,-0.2715215087,-0.2479568273,-0.3175480962,-0.2529810071,-0.4196513891,-0.0761848167,-0.0940631106,0.1780696958,0.0090043806,0.1709620655,0.1889690459,-0.4353352487,-0.1018504202,0.0006975317,0.23761563,0.0146928616,-0.1129617095,-0.2612105012,-0.2340477854,-0.3139867485,0.345592469,-0.2557941675,0.083301276,-0.140028879,0.0895220935,0.1889620423,0.0027529181,0.7596032023,-0.1823073924,0.2290213704,0.38683483,0.2017732263,-0.6543713808,-0.0279390458,-0.2924177051,0.3361033797,0.0906111449,0.3959363401,-0.4269514382,-0.0443884917,0.1525252908,0.0314957127,-0.0939231589,-0.2425147295,-0.2687611878,-0.3051568568,-0.0087272041,0.0218200311,0.0090193832,0.4349669814,-0.1700991541,-0.1275983006,-0.4085874557,0.1602924466,0.1126146093,-0.0253704134,0.2526399791,-0.056815505,-0.0093024755,0.0635729954,-0.0077941241,0.1148457006,0.3582159579,0.0272814278,-0.4528546631,0.2615330219,-0.0211641658,0.2443993241,0.0709576383,-0.1279031187,-0.0851995796,0.0087838434,0.0780045241,-0.0141761946,0.0905466229,0.3425087929,0.0383048356,-0.3234171569,-0.4045231044,0.2545499206,0.3031949997,0.0401418433,0.3135276437,0.0083070444,-0.1673143357,0.0610586889,-0.1346745938,0.8996839523,-0.0605399869,-0.1017853916,0.6653998494,-0.2444549948,0.4785053432,0.0302947909,0.0903261378,-0.2928251624,-0.2700549364,-0.0155689307,-0.2568275332,0.2637600899,0.0815494508,-0.4304995835,0.3003613055,-0.291116178,0.2078086585,0.0213771164,-0.0544004999,-0.060865432,-0.1126743257,-0.3436706066,0.1629805863,-0.2664074898,0.2688083351,-0.1147507727,-0.14413324,-0.0331183821,-0.0753438398,-0.1297870278,0.1565764546,-0.3821193576,-0.1631519198,0.1158264056,-0.2006135285,0.0768847913,0.1819679886,0.0739785209,-0.10281232,-0.3268237114,0.1788987815,-0.0758843198,0.1364337355,-0.087842159,0.0000826565,0.2469836175,-0.1007041708,-0.33282125,0.1236411333,-0.0640807524,0.21982342,0.2899284065,0.180278793,-0.072996825,-0.2621715069,-0.2122358978,-0.1799507886,0.3643169701,-0.2759153545,0.1959189624,0.009252294,-0.1024515927,0.0800813809,-0.0507112592,-0.1253606826,0.0088543976,0.4873361588,0.4263882637,0.2872929573,0.2660657465,0.2743073404,-0.0806822106,-0.3562266827,0.2828985155,0.1164565831,-0.33517012,0.3059953153,0.1482208073,0.0116094733,0.0161372833,0.4952471256,0.0256405957,-0.0226025004,-0.0756149366,-0.2714717686,-0.294061482,0.211046204,0.0070540542,-0.0486660786,0.1137508452,0.1883585602,0.2880287468,-0.153834179,-0.3571674824,0.1202098504,-0.3974452615,-0.026375806,0.0220237225,-0.0883372873,0.2679443061,0.0743281767,0.1367656738,-0.0063327239,-0.1159371808,-0.2485037595,-0.2217644304,0.1363088489,-0.0278415494,0.1512162387,0.1515443176,-0.2163798362,-0.2809672058,-0.0129056145,0.0469002761,0.0201656688,-0.1744152457,0.0505833849,0.1910213977,0.3978153467,-0.0217826068,0.2827041447,-0.0067478046,0.0345397852,-0.1433786154,0.02376814,0.2030914575,0.0252375361,-0.2087370902,-0.0644891113,0.2508503795,0.1676190645,0.0940858424,-0.3053337336,-0.1231551021,-0.0976448059,0.0022103242,0.1308603883,-0.1723631322,0.2990866005,-0.1103164032,0.2701034248,-0.2779805362,0.0524252616,0.326957196,-0.1212088093,0.0678809881,0.0283454955,-0.3222197592,0.2012974918,-0.1126514226,0.1511472762,0.4406094253,-0.1247906014,0.0929894373,0.2174029052,-0.1162052229,-0.015790185,0.460701853,0.2020016313,0.0393340774,0.1420467645,0.0226332583,0.0366369858,-0.0482860841,0.1063207835,0.2827980518,-0.4781712294,0.0035832161,0.1094033644,-0.0017196483,-0.0950419977,-0.3151477277,0.5316855907,-0.4729452133,-0.1242428645,-0.2933817804,-0.0078632785,-0.188767314,0.1411719173,-0.1920025945,-0.1300992668,-0.0960251093,-0.1964515448,0.0039726589,-0.1452821344,0.2983595133,-0.1298789233,-0.0621965863,-0.1563254148,-0.0753151849,0.4235470593,-0.0483771041,-0.169693917,0.0714745149,-0.1513211429,0.2156936377,-0.0976124927,0.5807609558,0.332120806,0.3286684453,-0.2131840289,-0.1695679575,0.2545269728,-0.3181824982,-0.1070876792,0.1548376232,0.1585799307,0.1606116891,0.0785211027,0.2066705674,-0.2241433114,0.142018795,0.0436597429,-0.1447922885,-0.3188250661,0.4858053029,-0.3899853826,-0.1255370826,-0.127706483,-0.0744743422,-0.5568339229,0.1190805957,0.2711359262,0.2195585668,0.0988916159,-0.1955020726,0.0958702639,-0.3225300908,0.2215004414,0.020388402,0.3648219109,-0.3144544363,0.1156628281,-0.5239733458,0.4592312872,-0.0107843764,0.0105118509,0.1541946232,-0.0931218565,-0.2654405832,-0.0039149406,0.030693708,0.2047602385,0.399890393,0.1453907788,-0.5745072961,-0.3909095824,-0.0477377437,-0.0727452859,0.0916486382,-0.2621004581,0.0129266009,-0.1517878324,0.023295382,-0.1117513552,-0.3203566968,0.1511974633,0.1406897157,0.7023329139,0.0446438454,0.4411260486,0.028585989,-0.011712499,0.1216360629,-0.3079591691,-0.105045788,0.176432088,-0.004793454,0.0528662838,-0.020993961,0.2726189494,-0.2175622582,-0.1812321395,-0.1439599544,0.1281732172,-0.3053524792,0.311968416,-0.3086757362,-0.1393731236,-0.3422217667,-0.0057940949,0.1176716611,0.0452723876,-0.1081117392,0.0493608788,0.6277851462,-0.1653215438,-0.3440380692,-0.1038250551,0.3128980398,-0.0359871238,-0.112662144,-0.5198081136,0.0568199977,0.4852424562,-0.0689152628,-0.1922020912,0.2341876179,-0.1415186673,0.1324385554,-0.0214217473,-0.0941569433,-0.0095509402,-0.0314085595,0.2175465375,-0.1435788721]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Hi @severo, thanks for reporting.\r\n\r\nJust note that currently not all canonical datasets support streaming mode: this is one case!\r\n\r\nAll datasets that use `pathlib` joins (using `\/`) instead of `os.path.join` (as in this dataset) do not support streaming mode yet.","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":41,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nHi @severo, thanks for reporting.\r\n\r\nJust note that currently not all canonical datasets support streaming mode: this is one case!\r\n\r\nAll datasets that use `pathlib` joins (using `\/`) instead of `os.path.join` (as in this dataset) do not support streaming mode yet.","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"OK. Do you think it's possible to detect this, and raise an exception (maybe `NotImplementedError`, or a specific `StreamingError`)?","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":19,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nOK. Do you think it's possible to detect this, and raise an exception (maybe `NotImplementedError`, or a specific `StreamingError`)?","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"We should definitely support datasets using `pathlib` in streaming mode...\r\n\r\nFor non-supported datasets in streaming mode, we have already a request of raising an error\/warning: see #2654.","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":27,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nWe should definitely support datasets using `pathlib` in streaming mode...\r\n\r\nFor non-supported datasets in streaming mode, we have already a request of raising an error\/warning: see #2654.","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Hi @severo, please note that \"counter\" dataset will be streamable (at least until it arrives at the missing file, error already in normal mode) once these PRs are merged:\r\n- #2874\r\n- #2876\r\n- #2880\r\n\r\nI have tested it. \ud83d\ude09 ","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":40,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nHi @severo, please note that \"counter\" dataset will be streamable (at least until it arrives at the missing file, error already in normal mode) once these PRs are merged:\r\n- #2874\r\n- #2876\r\n- #2880\r\n\r\nI have tested it. \ud83d\ude09 ","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Now (on master), we get:\r\n\r\n```\r\nimport datasets as ds\r\nds.load_dataset('counter', split=\"train\", streaming=False)\r\n```\r\n\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\nThe error is now the same with or without streaming. I close the issue, thanks @albertvillanova and @lhoestq!\r\n","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":191,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nNow (on master), we get:\r\n\r\n```\r\nimport datasets as ds\r\nds.load_dataset('counter', split=\"train\", streaming=False)\r\n```\r\n\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets\/src\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\nThe error is now the same with or without streaming. I close the issue, thanks @albertvillanova and @lhoestq!\r\n","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Note that we might want to open an issue to fix the \"counter\" dataset by itself, but I let it up to you.","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":23,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nNote that we might want to open an issue to fix the \"counter\" dataset by itself, but I let it up to you.","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2866","title":"\"counter\" dataset raises an error in normal mode, but not in streaming mode","comments":"Fixed here: https:\/\/github.com\/huggingface\/datasets\/pull\/2894. Thanks @albertvillanova ","body":"## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":5,"text":"\"counter\" dataset raises an error in normal mode, but not in streaming mode\n## Describe the bug\r\n\r\n`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> a = ds.load_dataset('counter', split=\"train\", streaming=False)\r\nUsing custom data configuration default\r\nDownloading and preparing dataset counter\/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to \/home\/slesage\/.cache\/huggingface\/datasets\/counter\/default\/1.0.0\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9...\r\nTraceback (most recent call last):\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 726, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1124, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1185, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/slesage\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/counter\/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9\/counter.py\", line 161, in _generate_examples\r\n with derived_file.open(encoding=\"utf-8\") as f:\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1222, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n File \"\/home\/slesage\/.pyenv\/versions\/3.8.11\/lib\/python3.8\/pathlib.py\", line 1078, in _opener\r\n return self._accessor.open(self, flags, mode)\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 1112, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 636, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/slesage\/hf\/datasets-preview-backend\/.venv\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 728, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] No such file or directory: '\/home\/slesage\/.cache\/huggingface\/datasets\/downloads\/extracted\/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211\/COUNTER\/0032p.xml'\r\n```\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> b = ds.load_dataset('counter', split=\"train\", streaming=True)\r\nUsing custom data configuration default\r\n>>> list(b)\r\n[]\r\n```\r\n\r\n## Expected results\r\n\r\nAn exception should be raised in streaming mode\r\n\r\n## Actual results\r\n\r\nNo exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.1.dev0\r\n- Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nFixed here: https:\/\/github.com\/huggingface\/datasets\/pull\/2894. Thanks @albertvillanova ","embeddings":[-0.4171527922,-0.2417904586,0.0004065955,0.1936673522,0.187378943,0.0572046041,0.5056346059,0.1059950739,0.1942841411,0.1353349984,0.0698306859,0.1297213137,-0.2097095996,0.3238820434,0.0893105865,0.0787798092,0.0210811216,0.1588886231,0.0335764661,0.0774051622,-0.2157031149,0.2028147876,-0.3865993321,-0.3339452446,-0.1323988438,0.1463262439,0.1100258455,0.1485300958,0.1753474176,-0.6307143569,0.2604337335,0.0775746107,0.2229160666,0.7022691965,-0.0001109225,0.1807845384,0.4209491313,-0.138756454,-0.5919362307,-0.2953981757,-0.3683612645,-0.3184562027,0.1747525334,-0.1460207105,-0.1155984327,0.0416297689,-0.2645150125,-0.4016982019,0.2196300626,0.44546175,0.2117786556,0.4712955952,-0.1509929299,-0.1280123889,0.1274099499,0.0086588459,-0.1527633816,0.2987458408,0.0759947449,0.2464787662,-0.030185163,0.400646776,-0.0477484316,0.1930941492,0.2172417343,-0.0403118059,-0.0638400242,-0.2730312347,0.077968359,0.2267186195,0.3892460763,-0.3428210616,-0.1372255683,-0.4340849519,-0.0550197884,-0.4057786763,0.2461039424,-0.0120554632,-0.2402932793,0.0848756358,-0.3027442396,0.1529166549,0.0393640399,-0.1008400992,-0.0583141893,0.1315543056,-0.0259356182,0.1768467724,-0.0182867814,-0.0394550189,0.1186181977,-0.2750140131,-0.0770448223,0.1979830861,-0.5796865821,0.0617473088,0.3563539386,-0.483120203,0.2248359621,0.0766406581,0.3723810911,0.0808451474,0.1294206977,-0.0054175756,0.3990353644,0.2017412037,0.1839399487,-0.2446381003,0.1665526032,0.3514225185,-0.0046332767,-0.3092014194,0.0455646105,-0.0928959474,0.2623339891,0.0135595212,0.3132420182,0.0522785522,-0.3759042025,0.1984404474,-0.2598961294,0.0621640086,0.3526681066,0.2470078766,0.31240955,0.331681639,-0.1649695039,0.2566638589,-0.170162797,-0.4501440227,-0.2399167717,-0.3115689754,0.0298984032,0.0585324802,0.2430290878,-0.4115727246,0.2127870023,0.1143840179,0.2314704955,-0.1508325785,0.2350654602,-0.3559949696,0.2714752853,0.4170607924,0.2132242918,0.30559659,0.2130860984,-0.078121908,0.0865608603,0.1292979866,-0.0135739716,-0.3794960678,0.1188295856,0.241816327,-0.1854702979,0.0091848904,-0.1626182795,0.3421088755,0.0492594019,-0.490398556,0.1006153971,-0.1393199116,-0.1788205355,-0.1388723552,0.3767809272,0.4191721678,-0.3163481355,-0.0631755143,-0.2223777026,-0.1933458447,0.3068187237,0.0627034158,-0.1867482066,0.0545980632,-0.198907733,-0.0324022733,0.1652611792,-0.133823514,-0.4505629539,0.3839800358,-0.1003188565,0.4216839075,0.003368129,0.0348957814,0.0462126397,-0.1179660112,0.2082362026,-0.0482203104,-0.3429863751,0.1301704794,-0.2063286006,0.0962786973,0.1745322049,0.0263372958,0.1080550551,0.145677641,0.1164149791,-0.0307607576,0.1964359283,-0.0440106578,0.0053364923,0.2909404337,0.0424242578,-0.0588180684,0.000367176,-0.1810717881,-0.4702015221,0.3554164469,0.2276427895,0.0265536476,0.0678977966,0.0377307907,-0.3046529889,0.141935274,-0.2711749077,-0.4771591425,0.1949529499,0.3907288015,-0.1118490696,0.0888262317,-0.277343154,0.3020512462,-0.1696492136,-0.1002230644,-0.1528523117,0.1263132989,0.1340726018,-0.3565647304,-0.0337362103,0.0897880867,0.2519123256,-0.1344926953,-0.234999463,0.4969182014,0.1069902405,0.0699909925,-0.0781737417,-0.173212111,0.1972211748,-0.1314566433,-0.2321656495,0.4724216461,0.3397763968,-0.2434855402,-0.1393338144,0.186793372,-0.2051788419,0.2564637661,-0.0694206432,0.0912937298,0.328330636,0.3197959363,-0.0928376839,-0.0212596133,0.2077185959,-0.1777403057,0.1031239405,-0.1610922515,-0.2532593608,-0.0428625979,0.1946246028,0.1955534518,0.0406559408,0.0529748946,-0.3236100376,-0.2086987048,0.3706597984,0.0880206153,0.4839422405,0.0791207775,0.2721221745,0.0922638103,-0.0106658787,-0.1263325214,0.2971056998,0.0752388388,0.0984970778,0.3603755832,0.2497458905,-0.1204286814,-0.4016779959,-0.2251661271,0.1358991563,0.133663699,-0.3052351177,0.0297692902,-0.1282226741,-0.2748440802,0.1133073941,-0.4021544755,-0.2225447446,-0.3337677121,-0.2058949023,0.3451238573,-0.1469643116,0.1498992443,-0.180475831,-0.0882678479,0.308257699,-0.0031163315,-0.2322359681,-0.1135618612,-0.2594828606,0.03214081,0.0412808843,-0.3234791756,0.434748143,-0.1234986335,-0.1801847816,-0.1512944698,-0.0745449811,0.1740905792,0.0435857289,0.2001399845,0.2743394077,0.2197588384,0.039221812,0.0045702313,0.4241863489,-0.2919014394,0.0146440258,0.3762074709,0.2012005895,0.2666827142,-0.2297265977,-0.2490313053,-0.0779714212,-0.4310797453,-0.174230665,-0.0227724351,0.0273636784,0.4507275522,0.262483865,0.3891989887,0.3127643466,-0.0566316359,-0.3348741829,-0.3475183249,0.1322237253,-0.1974763721,-0.4097188413,0.1144938767,0.0011452077,0.1664847434,0.2781507671,-0.3012993634,0.0803638548,-0.0408363193,0.1188741624,-0.05930195,-0.1120122075,0.1432749033,-0.06137814,-0.1019272059,-0.3338415623,-0.1850921214,-0.1276152432,-0.1529603153,0.2324113101,0.0381841213,0.5950309038,0.207781747,0.5280631185,0.4836455286,0.0581972674,0.6157116294,-0.2377375066,0.2161288112,-0.3673551977,-0.3503070772,0.2225523591,-0.0085880933,-0.1320838332,0.208188355,-0.1470538974,-0.2681727409,-0.2567493916,0.035103634,-0.0647703409,-0.1755146086,0.0512214527,-0.2642833889,0.302411586,0.0359326154,0.2061444074,-0.2464565039,-0.0574943684,-0.2373392433,0.0760812387,-0.0045370539,0.006409138,-0.2967653871,0.0214666966,-0.5252018571,0.3021186888,0.2227880508,0.6154462695,-0.0391720943,-0.208869651,0.0450495034,0.0117582334,0.5304381847,-0.3789144456,-0.0240581483,0.2964757681,-0.0950774848,-0.4073814154,-0.1102449,-0.2325181812,0.3465290964,0.232893005,0.1428349912,-0.2224261612,-0.0886221677,0.1366074681,0.1360030472,-0.0751254857,-0.1846507341,-0.2278954089,-0.2202225477,-0.1945456862,-0.126737833,0.0843008831,0.2432254553,-0.1949771643,-0.0850351676,-0.1033551395,0.0288981795,0.0844489187,0.1044177935,0.3788572848,0.0903362259,0.2310871333,0.3695783913,0.1015721634,0.3100064099,0.5074752569,0.406558156,-0.3184788823,0.2045599073,0.0775870979,0.0856532231,0.288239032,-0.1116834059,-0.0276315771,0.0777177364,0.093457222,-0.2894927263,0.1943801939,0.2449228019,-0.0539539419,-0.1495429426,-0.5760563612,0.3336381912,0.030438222,0.0183755308,0.3064616621,-0.2464926243,-0.2605732381,-0.0352292471,-0.0546513498,0.4981874526,0.1343333125,0.0661568791,0.4267181158,-0.2779954374,0.2107367963,-0.1438204199,0.2207974344,-0.187934801,-0.3919576406,-0.0941024274,-0.10854128,0.0540997535,-0.0021506485,-0.2287910581,0.2563299835,0.0245750789,0.2516512573,0.0321382955,0.0513761006,-0.4211883843,-0.104438141,-0.2310300469,0.2266884446,0.0391741507,0.0595658608,-0.1965502948,-0.1511451453,0.2058582604,-0.3431107402,-0.2650854588,-0.0005393665,-0.2571261823,0.0304906834,-0.0789500475,-0.558648169,-0.0092750648,0.1646382511,-0.1165745929,-0.2574972808,-0.1603616029,0.0731903464,0.1823066473,0.4032897949,-0.0304601323,0.1647265106,0.3220483065,-0.0156900398,-0.1203032583,0.1713949293,-0.2056288868,-0.1386834234,0.0792042762,0.0525303334,0.1477532983,-0.1950283945,0.0720984861,-0.1244531721,0.0610796064,-0.2714118063,0.1379991472,-0.0522926897,-0.2443103045,-0.0718860403,-0.1764226109,-0.1069568917,-0.1404853761,0.3965620995,-0.1222574115,0.2639794946,0.5979731679,0.2628350556,-0.069922179,-0.2239042521,0.2011469901,0.066502966,-0.2591678798,0.3576523364,-0.0661282092,-0.0084151132,-0.0689552277,0.3735864162,-0.0137275895,-0.0653011873,-0.1425481886,-0.2276573032,-0.339094311,-0.0766485855,-0.0648032054,0.1809473783,-0.0614876933,0.0205044076,0.3440939784,-0.2483291179,-0.3373638988,-0.2206485569,-0.2257323116,0.1367567629,-0.1156357825,0.0903353691,0.2451532632,-0.0245363768,0.1401757002,0.1833451092,-0.0654687658,-0.2514533401,-0.072165139,0.0910971165,-0.0126871252,-0.2120142579,-0.1004446447,0.0691824108,-0.3066586256,-0.3149644732,0.2542549074,0.0261867605,-0.0648500547,0.3339463472,0.120893836,0.2269126773,-0.2458010316,-0.0012788579,-0.0492911786,-0.0714226738,-0.2654004991,0.1434953511,0.0741449967,0.0186536331,-0.0137111247,-0.030239949,-0.0359067172,-0.0958979055,0.3457251489,-0.3809930384,0.1985213012,0.2853876054,0.4326710999,0.3073008955,-0.2745672166,-0.0601298138,-0.0283836294,0.2137221843,-0.2702535689,0.1047894284,-0.1389295012,0.3404973149,0.181291014,0.123454757,0.042961996,-0.2800160646,0.0771737248,0.1745934039,0.0768419132,-0.1775951684,0.2129438967,0.4377642274,-0.1620197445,0.1636501998,-0.0975020081,0.0571440384,0.0120126558,0.4437315762,-0.0766298324,0.1632130891,0.2037422657,-0.1365545392,-0.0510564111,-0.3139661849,-0.2231898904,0.2932848632,-0.147838369,0.2211655676,0.0607555956,0.1148490384,-0.2439828813,0.124097243,-0.3192916512,-0.080800429,-0.1555945724,0.0469758473,-0.1781293303,-0.124873817,-0.202840507,-0.1335257292,0.1504873037,0.1290383786,0.3712652624,0.043258965,-0.0071963728,-0.6050085425,-0.0662222132,0.3256896436,0.0365514085,-0.2266713977,0.2377025038,0.4254244268,0.0902423635,0.3842978477,0.3726780713,0.3797657192,0.3374568522,-0.0666713566,0.1129238158,0.1543490291,-0.1648841947,-0.2762878537,0.1785657257,-0.0458157174,0.273303926,0.3034488857,0.1270707846,-0.1446415633,0.1170547977,0.0449929349,0.0575731881,-0.6023664474,0.426597923,-0.4701407552,0.0185302999,-0.1208022311,-0.1203802675,-0.2376078218,0.0821256414,0.2865520716,0.0400108956,0.2274750918,-0.0990611687,0.1067298427,-0.125540629,0.2446664572,-0.0027377591,-0.255812645,-0.2028058618,-0.1672923118,-0.8009281754,0.0148777906,0.0856726393,-0.1091405004,0.1444782764,0.1989774257,-0.1189473942,0.2464016676,0.3879629374,-0.2339880317,-0.1309391558,0.3268940747,-0.1953607798,-0.4079427123,-0.4459073246,-0.0250745453,-0.0120659471,-0.2429813445,0.2650612295,-0.2226771861,0.1452894956,0.1209749728,-0.0283921864,0.1159845218,-0.0041816803,0.3461862206,0.1672824472,0.314198494,-0.1446937025,0.1787152439,-0.329939276,-0.1000356674,-0.3050260544,0.3060719669,-0.1126585081,0.4945887029,-0.1233776212,-0.0124988612,-0.1637256145,0.2108146697,0.2584315836,-0.2667381167,-0.3577256799,0.2748013735,-0.2798830569,0.2512665987,0.0318002701,0.3652265668,-0.0365184397,0.135316059,-0.2697343826,-0.5320450664,0.4457719028,-0.4308793247,-0.3953136802,-0.0651005581,0.1863383502,0.2452222258,-0.2555918097,-0.3518375456,0.1235720217,0.3564775884,-0.1301043332,-0.2069166154,-0.0070771347,-0.1155157685,0.0543674044,-0.0779213235,0.145641759,0.1664129049,-0.157684952,0.2604365945,-0.0645067096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2860","title":"Cannot download TOTTO dataset","comments":"Hola @mrm8488, thanks for reporting.\r\n\r\nApparently, the data source host changed their URL one week ago: https:\/\/github.com\/google-research-datasets\/ToTTo\/commit\/cebeb430ec2a97747e704d16a9354f7d9073ff8f\r\n\r\nI'm fixing it.","body":"Error: Couldn't find file at https:\/\/storage.googleapis.com\/totto\/totto_data.zip\r\n\r\n`datasets version: 1.11.0`\r\n# How to reproduce:\r\n\r\n```py\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('totto')\r\n```\r\n\r\n\r\n","comment_length":20,"text":"Cannot download TOTTO dataset\nError: Couldn't find file at https:\/\/storage.googleapis.com\/totto\/totto_data.zip\r\n\r\n`datasets version: 1.11.0`\r\n# How to reproduce:\r\n\r\n```py\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('totto')\r\n```\r\n\r\n\r\n\nHola @mrm8488, thanks for reporting.\r\n\r\nApparently, the data source host changed their URL one week ago: https:\/\/github.com\/google-research-datasets\/ToTTo\/commit\/cebeb430ec2a97747e704d16a9354f7d9073ff8f\r\n\r\nI'm fixing it.","embeddings":[-0.294410497,0.3900933862,-0.1317222118,0.0641968921,0.4353683293,0.1393284649,0.1679639369,0.5879850388,-0.0261453092,0.234341681,-0.1947766393,-0.0473091528,0.0594171956,0.318708539,0.0643599778,-0.3084107637,0.0652796775,-0.0130061815,-0.1395519972,0.0308540575,-0.0759798363,0.2782622278,0.0100256903,0.0654641986,0.0181091316,0.0936704874,-0.0290926713,-0.0979447067,-0.2686378956,-0.4347373247,0.415338546,0.0613755472,0.1697377264,0.3808855712,-0.0001073767,0.1472330391,0.3291011751,0.0086168032,-0.2500196695,-0.3531953394,-0.3288857937,-0.1141764298,-0.1965035796,-0.1480810791,-0.0830533355,0.2127362341,0.2038826346,-0.2607125342,0.0417160429,0.5573630333,0.2718278766,0.0907747969,0.3079364896,-0.3647713363,0.2657265365,0.1475394666,-0.0365158468,0.0705632791,0.0785895288,0.0108668972,0.3562817872,0.0460000634,-0.133997038,0.0743308142,-0.1553672999,-0.0840224102,-0.0924824327,-0.3424230218,0.2141403407,0.3073774278,0.5096561313,-0.2044652998,-0.3611649573,0.2267780751,-0.0598624535,-0.279548645,0.2698038518,0.1802222282,0.0900912583,0.1698183566,0.1732570976,-0.4855846167,-0.1890135705,0.0528651476,-0.4141792953,0.107604593,-0.1177637801,0.0048586847,-0.105921559,-0.1203932837,0.3016455472,-0.0547915176,-0.0375608727,0.2590126991,-0.2745456994,-0.1841468364,0.1035571247,0.0311484244,0.1920053959,0.0490685366,0.1362317801,-0.1017651632,-0.5062013865,-0.0666135922,0.2418693751,-0.1095379815,-0.000267334,0.3804118037,0.2949267328,0.2416359335,0.1484333426,0.0473140553,-0.1586337984,-0.358691752,0.0269331522,-0.113104403,0.4928361177,-0.2085420638,-0.3899927437,0.1616978347,-0.090081729,-0.1214746758,-0.0176121797,0.1517945826,-0.0244206693,0.0753581598,0.0918698311,0.1404550672,-0.008870828,-0.1377051771,-0.2661000192,0.2041446567,-0.0125199547,0.0713024735,0.2309996784,-0.3448672891,0.0952998549,-0.0258917771,-0.4624833763,0.0828349143,-0.1361531913,-0.0750162899,-0.2458771169,0.3632289171,0.2265905589,0.1045650393,0.0013822372,0.1134399772,-0.2949780524,0.2792558968,-0.3678761125,-0.0678893328,-0.1227772906,0.2941953242,-0.3227276504,-0.2466399968,-0.5071802735,-0.0339648984,-0.0525038354,-0.3056934178,-0.1234595478,-0.1745043397,-0.1805351675,-0.2515176237,0.1124921739,0.6013523936,-0.2063911855,0.020885611,-0.4196625054,-0.4201267958,0.1899637729,0.269320339,-0.0341406576,0.4477233887,-0.3025847673,0.0885180831,0.4392078221,-0.1254837364,-0.7618820667,-0.0033206183,-0.2596687675,-0.1898869723,0.1790016443,0.0790314004,0.2497510761,-0.1000057384,0.2460572869,0.1370749325,0.012917717,-0.0459471196,-0.2387001067,-0.0105643263,-0.0045891618,0.1839413047,0.1048776805,0.2050982267,0.1501675397,-0.0725682601,0.2357553095,0.1607331038,0.0994152948,0.3408914506,0.3313904405,-0.0265133232,-0.0199446026,-0.1835746318,-0.1992596537,0.1232289374,-0.2162778378,0.1084155291,-0.2086808383,-0.1304647177,-0.4467419088,0.0594626628,-0.0653845966,-0.093042925,0.1499165893,-0.0002756922,0.310736388,0.3321071863,-0.02035648,0.320682168,0.2677029073,0.1030120254,-0.1375443637,0.5665908456,-0.1339783072,0.0568048991,0.4536594152,-0.2412828058,0.236403808,-0.0816816241,-0.0623007454,0.1545941979,0.0604691543,0.3422275782,0.4771338403,0.2396415472,0.1721659005,-0.4354346693,0.2116640955,0.343268007,0.0783024356,0.1411449164,-0.1909071058,0.0213618409,0.1539851874,0.0350163281,-0.0176072661,0.2682003379,0.544712007,0.0763235763,0.0633015558,-0.0642063022,0.2679164112,0.5545850396,0.2224782109,-0.0677626878,-0.2419174314,0.1457102746,0.538705945,-0.0372180901,0.0527861379,0.2355953008,-0.1476681978,-0.0631909966,-0.1953097135,0.2097867578,0.1718096882,0.0596293472,0.1565221548,0.0106348377,-0.0777703449,-0.2667880654,0.0551331863,0.0725212246,0.1452946216,0.1418423057,0.1219759658,0.2061336339,-0.08200939,-0.3087290227,0.0309936032,0.3056665659,-0.1311020702,-0.0795388594,-0.2592677772,0.0049411356,-0.2732787728,-0.3159630299,-0.2113306969,-0.3952906728,0.0605420209,0.2657611072,0.0656001344,-0.0102982745,-0.0560371764,0.0340000875,0.2436152548,-0.3098428249,-0.1341092587,-0.150049597,-0.1041894853,0.170614779,0.2377597988,-0.3077156246,0.3982344866,-0.6368309259,0.1199382693,-0.5480054617,-0.234621942,-0.043117255,-0.1899068207,0.0942285731,0.175614506,0.4022015929,0.3427299261,0.2505168021,0.1937542409,-0.0247938484,-0.0510064363,0.0891098529,-0.1387198567,-0.0441237539,0.1827340573,-0.4613678753,-0.3021485507,-0.2280495763,0.0444356278,0.0629076734,0.2366307676,-0.2858436108,0.1709893048,0.0659947768,0.1328113973,-0.3080635071,-0.1222961098,-0.1796017885,0.4985628128,-0.2371805012,-0.2333060801,0.1059011444,0.130261451,0.4757525921,0.1674677283,-0.4497697651,0.179243952,-0.3459710777,0.1534799039,-0.0468465574,0.0758880824,0.3201242983,0.1453730166,-0.1137198061,-0.0569933727,-0.1610036939,-0.050038822,-0.3855092824,0.2330114543,-0.0367576629,0.2503889203,-0.1210072637,0.398199141,0.2551098466,0.07454741,0.3040669858,-0.1943264455,0.1306042373,-0.1279257536,-0.1889470071,-0.0984253138,0.1577033997,-0.1163241193,-0.1967870146,-0.0184634868,0.0741213709,-0.1694130898,-0.1042783111,-0.2042790651,-0.2836247385,-0.1303623915,0.0177816842,0.4172818065,0.185910061,0.0156836808,-0.1674310863,-0.0935420543,0.2443633378,0.1396629214,-0.0014964864,-0.018045662,-0.1251368523,-0.1843106151,-0.6704334021,0.3441170454,-0.1174282879,0.2353997678,-0.0927243307,0.185250476,0.1104813442,-0.0463375226,0.5471497774,0.062537156,0.3733004928,0.3563941717,0.2766824961,-0.146045804,-0.0525100045,-0.1304521114,0.2965848744,-0.0263616275,0.3771854639,-0.2650294006,-0.0554963201,0.2222447246,0.1888717711,-0.1031564474,0.1805047095,-0.0639477074,-0.4556207061,-0.2093116641,-0.1772014797,-0.0870981961,0.2738859355,-0.1530112326,-0.0872611031,0.0672675371,-0.147369504,-0.0175553579,-0.1686567217,0.381292969,-0.1637212038,0.177162528,-0.0096636331,-0.2981503308,0.5925562382,0.9116903543,-0.0094228163,-0.4505523443,-0.0699971095,0.1041783616,-0.1924929768,0.0370545797,-0.1077148691,-0.155616954,-0.1347660273,0.121474579,-0.1697997302,0.2179342657,0.2668156624,0.0785118788,-0.4189471304,-0.3319608569,0.407774061,0.1675906181,-0.0422660559,0.4756697118,0.1231786013,0.0052154721,-0.1512746066,-0.146587804,0.7076441646,-0.1708522737,0.121878989,0.3979227841,-0.1670890003,0.2951293588,0.0539895035,0.054950837,-0.356028378,0.2241553664,-0.0507987142,-0.0956090316,0.0422886536,0.0331598967,-0.2261527926,0.3760743439,-0.1912707835,0.0057616672,0.2342994213,0.2299364954,-0.2325203121,0.1433078796,-0.2494061738,0.2325792462,-0.1700909585,0.5963363051,0.012605018,-0.1578868926,-0.2352524251,-0.0887922347,-0.4586822689,0.2226306796,-0.3531464338,0.0472171605,-0.1291957051,-0.2200131863,0.0645607188,0.1586482823,0.3044909239,0.2362559736,-0.5237365365,0.2670378089,-0.3025308251,-0.3513713479,0.172851041,0.2117986232,0.2136156857,-0.0355904885,-0.2754929662,0.1930139661,0.0165787246,-0.1336362511,-0.1597443968,-0.3023325503,0.0960842669,-0.2871341109,0.0308847092,-0.1784550101,-0.0639164969,-0.2592556775,0.210878104,0.2444291562,0.1790868342,-0.0770321116,-0.2008534521,-0.2742711306,-0.1253176183,0.4576685429,-0.0899077132,0.0362664238,0.3003350198,0.1418553889,-0.0421103351,-0.2370263934,-0.0758946836,0.0870876238,-0.0705426484,0.2010704726,-0.246197015,0.4602616429,-0.1983534098,0.0310583059,0.0072811441,-0.0602627471,-0.0174485762,-0.4184100926,-0.0034990332,0.1399504542,0.068998836,0.0080933636,0.0801748037,0.33794415,0.0753722563,-0.2415237427,-0.3141222596,0.1533349752,0.010092549,-0.2602766156,0.3766644895,0.1265031993,0.2459721863,-0.1016764343,0.1848554462,-0.2197219133,-0.2210969478,-0.2645950317,0.0000300764,0.0581651703,-0.0433961898,-0.0884144753,0.1235888079,-0.3815771639,-0.2894198596,0.2447361052,0.0846430287,0.4267973006,-0.0242630001,-0.062661536,0.1742569208,0.0256446283,-0.0715736002,0.0522936583,-0.0506931208,0.0010746485,-0.0242586527,0.3078612983,0.0098361075,-0.1488315761,-0.3080678284,-0.012468027,0.069576934,-0.0835743919,0.4794929922,-0.238189742,0.0586935356,0.3324653208,0.1315368712,0.3096744716,-0.5152159929,0.1130948365,0.0408594236,0.286698699,-0.3299861252,-0.1299768537,-0.2388359457,0.0133060208,-0.1226402,-0.023742402,-0.0654749423,0.1589624137,0.2537431121,0.1377555579,0.3771550059,-0.1948017776,0.2289052606,0.309329778,0.1026867107,0.1862534583,0.2217633277,0.0799650326,-0.096968554,0.4350672066,-0.2293919921,0.1500754207,-0.0677117333,0.0790822133,0.1858036816,-0.2016682029,-0.3423802555,0.2099467218,-0.1070173159,-0.0416853353,0.0382248834,0.2646930814,-0.3808471262,0.2599398196,-0.2471588105,0.2180426419,0.0396916568,-0.107037656,-0.0710211098,-0.277212888,-0.1615739018,0.1111863405,0.0544033013,-0.0574874617,-0.1172462255,-0.0004238633,-0.4112986624,-0.4172866046,0.0453437679,0.1922140568,0.1519670635,-0.1251610518,0.4158020914,-0.0666866601,-0.2817391157,0.1539526582,0.4888014793,0.4909828007,0.1652413905,0.2514764071,-0.1447680593,-0.2571274042,-0.2132361531,0.0702499524,0.1148378253,-0.0334707685,0.0071211583,0.2830833197,0.2091301531,-0.1540863514,0.1244486645,0.0029568451,0.34194839,0.0563798025,0.3701932728,-0.3759686649,-0.0931663662,-0.2818811238,-0.0891278237,-0.3936681449,-0.1934796274,0.1249653772,0.0378298834,0.0091357706,-0.2682196498,0.1284017563,0.0321747214,0.1845884472,0.2696240544,0.2058680654,-0.2469167262,-0.3324057162,-0.5042065382,0.1490001231,-0.1049715057,-0.2146742195,-0.1411429644,0.0215014555,-0.2108839601,0.0625294596,0.2979361415,-0.0173048489,-0.1892402917,-0.1613342911,-0.1859367937,-0.0138215581,-0.175954029,0.0314951167,0.0316190124,-0.3335469961,0.2791323066,-0.348980993,0.0721447468,-0.202991724,0.058277525,0.0845616013,-0.0209053755,0.2773117125,0.1782321632,0.1633214355,-0.0716613159,-0.1242836416,-0.4604788125,-0.0752218738,-0.1951590627,0.0604053661,-0.1034028381,0.0961736962,-0.1812672317,-0.4262296855,-0.1299846321,0.0591744781,-0.151055932,0.0678856224,-0.1810762137,-0.0299206339,-0.0501213819,0.1820138097,0.1748444289,0.0471048988,-0.0015658198,0.0969647095,-0.2998532355,-0.2840430439,0.1638270468,0.0188097768,-0.0074548773,-0.0921064466,0.048402749,0.2315627784,0.064388819,-0.1891824305,0.2362127751,0.4156240821,0.009564341,-0.1696297377,0.1937181801,-0.111114867,0.1457280368,0.1789949685,0.4593870044,-0.0406309515,-0.1174679548,0.0221763458,-0.0482680798]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2859","title":"Loading allenai\/c4 in streaming mode does too many HEAD requests","comments":"https:\/\/github.com\/huggingface\/datasets\/blob\/6c766f9115d686182d76b1b937cb27e099c45d68\/src\/datasets\/builder.py#L179-L186","body":"This does 60,000+ HEAD requests to get all the ETags of all the data files:\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"allenai\/c4\", streaming=True)\r\n```\r\nIt makes loading the dataset completely impractical.\r\n\r\nThe ETags are used to compute the config id (it must depend on the data files being used).\r\nInstead of using the ETags, we could simply use the commit hash of the dataset repository on the hub, as well and the glob pattern used to resolve the files (here it's `*` by default, to load all the files of the repository)","comment_length":1,"text":"Loading allenai\/c4 in streaming mode does too many HEAD requests\nThis does 60,000+ HEAD requests to get all the ETags of all the data files:\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"allenai\/c4\", streaming=True)\r\n```\r\nIt makes loading the dataset completely impractical.\r\n\r\nThe ETags are used to compute the config id (it must depend on the data files being used).\r\nInstead of using the ETags, we could simply use the commit hash of the dataset repository on the hub, as well and the glob pattern used to resolve the files (here it's `*` by default, to load all the files of the repository)\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/6c766f9115d686182d76b1b937cb27e099c45d68\/src\/datasets\/builder.py#L179-L186","embeddings":[-0.5428535938,-0.0302166119,-0.1265247166,0.3286388516,0.256604135,-0.0104140276,0.197608754,0.3729439676,0.1844934821,0.2685716748,-0.1927886903,0.3850267828,0.0771837533,0.2340062857,-0.143449083,0.136008963,-0.1367563605,0.3763958514,0.1254129112,0.1078248546,-0.0619285703,0.0006759175,-0.0417456739,-0.0773263052,-0.2133178711,0.2759331763,0.1665377468,0.2257279307,-0.0134929046,-0.4966956973,0.240168795,0.2341412455,-0.0487481132,0.2447642982,-0.0001018645,0.1439338028,0.4033434391,-0.0089729438,-0.085589841,0.4268731475,-0.4157178402,-0.0739421472,0.0224096645,-0.3959259391,-0.2034465969,-0.2251017988,0.0632516742,-0.0980081558,0.2041195333,0.2827344537,0.260243088,0.4249944985,0.0377698317,-0.1970410794,0.2039705664,-0.316323489,0.0617868043,-0.056926813,0.0797685385,0.1593059301,-0.0834435821,0.2483358234,-0.0834733918,0.0802756175,0.2913605869,-0.1320183277,-0.0032377529,-0.2171800733,0.3939404786,0.2030942291,0.3443941772,-0.3286153674,-0.1332122386,-0.2533044219,0.0619671606,-0.1955298185,0.0476448163,0.1538827866,-0.1587007791,0.1192615852,-0.0326113962,-0.0544691943,-0.1602670252,-0.0494049191,-0.0776140019,0.089980714,-0.0840734094,0.1207940727,0.136279732,0.058017496,0.0522075854,-0.2175264806,-0.0846055672,-0.004862661,-0.3694218993,-0.0516511574,0.3653528988,0.4273298979,0.1478993595,0.1687032282,-0.1197926477,0.2871201038,0.0435164198,0.0685389638,0.0575072095,0.0231996793,-0.0176825654,-0.389597863,0.5508090258,0.0899150446,-0.1071902066,-0.2005494833,-0.0709187463,-0.1060602814,0.0840810388,-0.1210096851,0.1117649302,-0.2036800086,-0.0995911956,-0.1678549349,-0.0444756113,0.0720648617,0.1929066479,0.4125182927,-0.1070380881,0.2573559582,-0.2619225979,0.0917120874,-0.1934460849,-0.0925472528,-0.3160288334,-0.2928213477,-0.1694728285,0.0569894873,0.1707771719,-0.34016487,0.4697290957,-0.0191245824,0.3383053243,0.0339444056,0.0842757225,0.1081046611,-0.0809869394,0.4188500047,0.2263165563,-0.0787535161,-0.0641879365,-0.1923369914,-0.291814357,-0.1569433808,0.0312824622,-0.4784304202,0.1928419471,0.2877151072,-0.0427191779,0.1639586836,-0.3409174085,0.0171625558,0.0027538103,0.1268018633,-0.1082377136,-0.0361211002,0.1082094535,-0.2483724952,0.1618676782,0.4884290099,-0.0835868865,-0.0830648839,-0.039552927,-0.1286983341,-0.0772875845,0.3747945726,-0.3466023803,0.120490931,-0.0308591686,0.0409644805,0.032458812,-0.425350368,-0.6205683351,0.1140533686,0.0639335662,-0.0276992563,0.2386657745,0.3460126519,0.1039455682,-0.107216388,0.0222086739,0.1709899902,0.073792845,0.2256556302,-0.1190583929,-0.2345744073,-0.2999114394,0.3787722886,-0.1165996566,-0.1365634799,0.1347563714,-0.1876992285,0.2526454926,-0.3332684338,0.0380112678,-0.1584161669,0.0938546658,-0.0603501089,0.0127826966,0.2236943841,-0.4421748519,0.3059828579,0.2439787388,0.1666173935,-0.1472782493,-0.3505738974,0.0228462406,-0.0090298792,-0.3510281146,-0.0372455232,0.1835951656,0.0712855086,0.0677272379,-0.1504935324,-0.2270190865,0.2682373822,-0.3091340661,-0.0179539192,-0.4381204545,0.1312066019,0.0892122611,0.0468895361,0.2672764957,-0.1050158292,0.0074687083,0.0492864773,0.0621649623,0.1636259705,-0.0330553278,0.4530902505,0.0818722695,0.7110303044,0.1389819533,-0.2826481164,0.3203724325,0.2067766637,0.1654448509,-0.2394565195,-0.1613846868,0.4084748626,0.3778287172,0.3252526224,0.1585008949,-0.0329761878,0.410119772,-0.0721899271,-0.1317404956,-0.0330927819,-0.00823195,-0.0209802482,0.3309631944,-0.023005046,-0.2523918152,0.3234723508,0.3031021953,0.1111327931,-0.1492699087,-0.0652219951,-0.298992157,-0.1263385117,0.2080144435,-0.0855184197,0.2448243797,0.4307595193,-0.0929894522,0.1461498737,0.0120118773,-0.256077677,0.3098646998,0.0617690906,0.0520799011,0.0422569439,0.1629692465,-0.2966534495,-0.4100134969,-0.3106255233,-0.1572241336,0.1993097663,-0.2276444584,-0.0174489729,-0.0920694247,-0.3194893003,0.0963194221,-0.0115835099,0.0477257594,-0.2335455716,0.0685858503,0.2594055235,-0.2168204486,0.1381093711,-0.1156482473,0.2903750539,0.0892408416,-0.1132856086,-0.2681921422,-0.2419697493,-0.1095879748,0.1148864031,0.4646658599,0.0141319409,0.3488751948,-0.3224317729,0.1775819212,-0.3888134956,-0.1769560575,0.1015825272,-0.1227035671,0.0946206078,0.0673972368,0.1070112586,0.2800557911,-0.125384286,0.0769697279,-0.2286968827,-0.0885543525,0.1452503055,-0.154563725,-0.2497059852,0.0414588526,-0.3924611509,-0.0118674375,-0.3728982508,0.342957437,0.2925284505,0.2317622006,-0.0143444026,0.0886501297,0.003775317,0.0996953025,-0.221465528,-0.2427807301,-0.3542366624,0.1676867455,-0.1924644858,-0.1902693361,0.0533196107,0.1380496174,0.4981618226,-0.0522706732,-0.4013659656,-0.0050968095,-0.4799565375,0.4729455709,0.0018389857,0.1385525167,0.2309796959,0.1050632596,-0.2131042331,-0.003395986,0.0059769182,0.1519519687,-0.0271408688,-0.2601141036,-0.0332135223,0.1607500166,-0.1440254897,0.287691772,0.401422888,-0.1580083966,0.2020960897,-0.1712821573,0.3033681214,-0.4020245373,-0.0107392669,0.1449472159,0.1542857438,-0.1868642867,0.3402228355,0.1952575743,-0.5144734979,-0.2171178758,-0.1385502368,-0.047275681,-0.3612928391,0.0551773384,0.3035261631,0.0228444245,0.0812270045,0.2058289647,-0.0380034633,-0.1516912431,0.2849894762,0.0916408524,0.0076755234,-0.2801522613,-0.5164752007,-0.0290395562,-0.7274724245,0.4155051708,-0.1272853166,-0.1209579483,0.1984863281,-0.1490302235,0.1140256822,-0.1446226537,0.3893289268,-0.360753417,0.0653221086,0.1890885085,0.2773555815,-0.2066984922,0.0153291468,-0.1949689388,0.2554828525,0.0409896933,0.4853951335,-0.3089270294,-0.2090707272,-0.1388294697,0.1643549055,-0.1718056649,-0.1440299302,-0.2748757303,-0.2858329415,-0.2627595961,0.1193227097,-0.0391619354,0.003167425,-0.0440098308,-0.0841071829,-0.1187053919,0.1033410579,0.1640647948,0.4710362256,0.1158270538,0.0934364945,0.1659183055,0.0374369659,-0.1580857188,0.2071107477,0.6684245467,0.1755816042,-0.2218962312,-0.1182288826,0.1617062241,0.1396004111,0.3151801229,0.1193937883,0.0207385924,-0.1292494237,0.3238544762,-0.3029144704,0.193803668,0.1370225847,-0.0925901309,-0.2631414235,-0.3615525663,0.4471847117,-0.1883913875,-0.1312835515,0.4696905613,0.3428766429,-0.1417624652,0.2846679986,0.2487651408,0.8667517304,-0.0928715691,0.1275847405,0.2757603824,-0.3555955291,0.1609345227,-0.3366043568,0.0796381533,-0.351541996,0.1143317819,0.1026704684,0.0138060609,-0.17082645,0.1024274305,-0.1321756542,0.2881008685,-0.1443882436,0.3143622875,-0.1412631124,0.435343951,-0.3882615864,-0.4119451344,-0.6139753461,0.2683770061,0.3304165006,-0.1537439227,-0.0472401492,-0.0246126465,0.195830375,0.0197973438,-0.0137850698,0.2106341124,-0.1912306547,0.092871733,-0.275851965,0.081352964,-0.1425484568,0.0267527327,0.0469446369,0.1992893517,-0.2087278515,0.3614895642,0.0190596003,0.1589624435,0.1899458021,-0.2031857073,-0.1272608638,-0.1174134687,-0.2530823648,0.1769057661,0.0884269029,-0.3375614583,0.022240879,-0.0522600412,0.0008730987,-0.1094607189,0.0189550538,0.2294987887,-0.2908079028,-0.1482636034,0.2226371914,-0.0878083035,-0.0815093443,0.3045401871,0.0436069034,-0.3001805246,-0.1483344734,0.2632359266,0.0427140892,0.0369828083,0.121416837,0.0237175189,-0.2068173587,-0.302369684,-0.1465962678,0.2409197986,-0.4375190437,-0.0331906378,-0.0494346097,0.0033178083,0.0506808609,-0.0272150077,-0.0058739791,-0.0840052441,-0.0371301249,-0.1992988437,-0.2851664126,0.3043626845,0.1010983586,0.3375176191,-0.0635118112,-0.1917614192,-0.2429069132,-0.0476204529,-0.4567089677,0.0419154353,-0.2124951035,0.05605492,-0.2779247761,0.0023317703,0.4613576531,-0.0997968242,0.2634966969,-0.0443392321,-0.4361954033,-0.3544987738,0.0609365441,0.1033121794,-0.0436229631,0.0142967226,0.0523092151,-0.076503098,-0.2957774997,-0.313400954,0.6046929359,0.1704049408,-0.1896798611,-0.2239180356,0.3594099283,-0.1177664101,-0.2642332911,0.215486899,0.1292701811,-0.0318126939,-0.068104282,0.0756035671,-0.341378659,-0.2242005467,0.4636442661,0.4278444946,0.1715426743,0.1539797187,0.1951588243,-0.0621873923,-0.0180493202,0.2234244496,0.1387220323,0.402947098,-0.0099452743,0.001162872,0.1709284782,0.3553200066,-0.2987711132,-0.3309018016,0.0836583748,-0.0448817126,0.1490396261,0.1131789684,0.2794712782,-0.1548952758,0.0715526789,0.0051325476,0.2009383142,-0.1418718398,-0.168976441,0.2532060742,-0.1036994383,-0.1869090796,0.1869180202,0.155501321,0.2906024754,0.5382152796,-0.1976806223,0.205840528,0.0884427577,0.1080901399,0.0100663658,-0.2402389348,-0.4833901823,0.0265654586,-0.1290313751,0.1863304228,0.202718392,0.3596491516,0.1560903788,-0.1159326583,-0.257940203,0.0854742229,-0.1045068949,-0.0051630386,-0.1370507926,-0.2379113585,-0.1614997536,0.2349171042,-0.1079137623,-0.0510151051,0.063251324,0.1068846062,-0.045994781,-0.0559367277,0.0670065805,0.0569071136,0.4399585724,-0.2259953767,0.0312765054,0.2127195597,0.1286762804,-0.0329716057,0.1760226488,0.3772985041,0.0412282757,-0.08742892,0.2876946926,0.1841208041,-0.0804393515,-0.2539972365,0.1091934517,0.0874144956,0.2268330157,0.1731670797,0.2240009159,-0.222832948,-0.2325480133,-0.2485449314,0.0837869942,-0.0438246764,0.1640124619,0.0560980551,-0.24888964,-0.0865307972,0.1873739958,-0.3417313993,0.0528506376,0.329139322,-0.1795737147,0.0524501763,-0.111730136,0.141595602,-0.3015827537,0.6976209283,0.0254463758,0.2618788183,-0.1180256903,-0.3995813727,-0.4921222329,0.0522904173,-0.0073115826,0.0935412943,0.0256276205,0.2950900793,0.1332703829,0.2454700172,0.0224462841,-0.368126899,0.255061239,0.2647507787,-0.2174696624,0.1709866524,0.0450598449,-0.1318834722,0.0978959724,-0.4364307523,0.1563719362,-0.0252211951,0.2079412937,0.0352058187,0.2020336241,0.1742803603,-0.2212349325,0.6054376364,-0.0029269236,0.1959469169,-0.1875263155,-0.1295257062,-0.4434603453,-0.0303481836,-0.2917251289,-0.215718776,-0.0807414427,0.1567137688,0.0566738807,0.100940749,-0.219702065,0.1784289032,0.2410254925,0.0916849822,-0.1354959309,-0.1076691151,-0.2465330511,0.3003370166,-0.2350136191,0.2327117324,0.0816560239,-0.25500682,-0.125608623,-0.3737340271,0.6916484833,-0.2861724496,-0.1407487988,-0.0290631168,0.4134774208,0.2459308952,-0.1011010483,-0.2815151513,0.2617774606,0.1991793066,-0.1879443675,-0.2414782196,0.3538321257,-0.0719252452,-0.1372228563,-0.1907454431,0.0154669201,0.0112963067,-0.2225657403,-0.25459674,-0.4584097266]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2846","title":"Negative timezone","comments":"Fixed by #2847.","body":"## Describe the bug\r\nThe load_dataset method do not accept a parquet file with a negative timezone, as it has the following regex:\r\n```\r\n\"^(s|ms|us|ns),\\s*tz=([a-zA-Z0-9\/_+:]*)$\"\r\n```\r\nSo a valid timestap ```timestamp[us, tz=-03:00]``` returns an error when loading parquet files.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Where the timestamp column has a tz of -03:00\r\ndatasets = load_dataset('parquet', data_files={'train': train_files, 'validation': validation_files,\r\n 'test': test_files}, cache_dir=\".\/cache_teste\/\")\r\n```\r\n\r\n## Expected results\r\nThe -03:00 is a valid tz so the regex should accept this without raising an error.\r\n\r\n## Actual results\r\nAs this regex disaproves a valid tz it raises the following error:\r\n```python\r\nraise ValueError(\r\n f\"{datasets_dtype} is not a validly formatted string representation of a pyarrow timestamp.\"\r\n f\"Examples include timestamp[us] or timestamp[us, tz=America\/New_York]\"\r\n f\"See: https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\"\r\n )\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Ubuntu 20.04\r\n- Python version: 3.8\r\n- PyArrow version: 5.0.0\r\n","comment_length":3,"text":"Negative timezone\n## Describe the bug\r\nThe load_dataset method do not accept a parquet file with a negative timezone, as it has the following regex:\r\n```\r\n\"^(s|ms|us|ns),\\s*tz=([a-zA-Z0-9\/_+:]*)$\"\r\n```\r\nSo a valid timestap ```timestamp[us, tz=-03:00]``` returns an error when loading parquet files.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Where the timestamp column has a tz of -03:00\r\ndatasets = load_dataset('parquet', data_files={'train': train_files, 'validation': validation_files,\r\n 'test': test_files}, cache_dir=\".\/cache_teste\/\")\r\n```\r\n\r\n## Expected results\r\nThe -03:00 is a valid tz so the regex should accept this without raising an error.\r\n\r\n## Actual results\r\nAs this regex disaproves a valid tz it raises the following error:\r\n```python\r\nraise ValueError(\r\n f\"{datasets_dtype} is not a validly formatted string representation of a pyarrow timestamp.\"\r\n f\"Examples include timestamp[us] or timestamp[us, tz=America\/New_York]\"\r\n f\"See: https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\"\r\n )\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Ubuntu 20.04\r\n- Python version: 3.8\r\n- PyArrow version: 5.0.0\r\n\nFixed by #2847.","embeddings":[-0.401596576,0.1829235852,0.07779257,0.0459456369,0.0025876896,-0.1361358166,0.2783327103,0.1863403022,-0.2822602391,-0.1084523499,0.0049508214,0.4612381458,-0.0764548704,0.011837787,0.0862466618,-0.0383578874,0.011814937,0.2273899317,-0.0372114666,0.0899889767,-0.2281404585,0.1430076659,-0.1861973256,0.0171979833,-0.0101191457,0.0847581327,-0.135529995,0.2957262397,-0.0179393198,-0.3312743306,0.0218966678,-0.2963922024,-0.0028354791,0.4744252563,-0.0001110772,0.1515261084,0.2427365482,-0.0586084239,-0.1286713034,-0.0980418921,0.0726354867,-0.1844875515,0.1201160625,-0.1410159767,-0.2818562984,-0.4092908502,-0.1490833312,-0.2255629301,0.4855223894,0.7533420324,0.1939110011,0.5161628723,-0.1995947659,0.1261487156,0.5708992481,-0.1920128316,-0.0927384943,0.0550048351,0.2838503122,-0.0549646839,0.0217351187,-0.0753623918,-0.0667717829,-0.1474190205,-0.0274879858,0.1315036863,0.1412029564,0.1112895012,-0.1078112423,0.0587866716,0.5331471562,-0.3926923573,-0.3223041892,-0.0117041012,-0.1060191542,-0.4288388491,0.241237238,0.3066093624,0.124086462,0.4001655281,0.2249842137,0.2806190848,-0.3247011304,0.2575951517,-0.0456544571,0.4889696538,0.2003501654,0.0241223089,-0.4663511813,0.0521042608,0.1108724698,0.1738105118,0.1420491487,0.3028146625,-0.0712178871,0.1471797675,-0.144489795,0.0858040303,0.1398317665,0.1655455232,0.0636238679,0.053757105,0.1454739124,0.1479489505,0.1892784685,0.0139465919,-0.1422267556,0.0613142997,-0.013570535,-0.2705182135,0.1799209118,-0.1905653626,0.3304877281,-0.3920660615,0.2817029655,0.1169816926,0.3362499475,-0.2698073685,-0.2274588794,0.1621998847,-0.1237005219,0.0080517251,0.0145802246,0.0851220116,-0.2806506157,0.0991542861,-0.0242676027,0.2271328568,-0.2589269876,-0.3209973276,-0.302365303,0.0338018872,-0.3716346025,0.1070889086,0.0459969416,-0.0455009155,0.100954026,0.5322930217,-0.1443692893,0.1839753389,-0.003500768,-0.0746635497,0.1978344023,0.2002501637,-0.3000686467,0.0292879231,0.0530873127,-0.3285584748,-0.1837910563,0.0785339326,-0.3674685955,-0.3055340052,-0.2438785285,0.2281315029,-0.1431456208,-0.0010055227,-0.3128293753,-0.0223961119,0.2348175347,-0.333059907,-0.004378831,-0.0979190767,-0.0376100279,-0.1344065964,0.2565334141,0.1289270669,-0.5483288765,0.0963393524,-0.2633677721,-0.1224089786,0.4269274771,0.0700964183,-0.1817486882,-0.1433549374,-0.0769442394,0.2165022492,0.1412177533,-0.2336217314,-0.1102606952,0.2337683588,-0.3173281252,-0.1192938015,-0.0990622118,-0.231791079,0.1512234509,0.020114962,-0.0212348569,0.3360035121,0.1083586439,0.1227589697,-0.5359722376,0.0807048976,0.3973245323,0.2341749817,0.0942998827,-0.1980790943,0.0284822267,-0.1203887314,0.1557871401,0.1266277283,0.0885832459,0.3448267877,0.0967899412,0.1287433058,0.1226942837,-0.1904807985,-0.4723727703,0.1862957329,-0.1095412374,0.1620388627,-0.1045084372,-0.0073415767,-0.4141269624,0.1481197923,-0.096865952,-0.0436174348,0.1869332194,0.0512472242,0.2622552812,0.249815464,-0.0642117113,0.3733778894,0.0246363953,0.0647729114,-0.0456543528,0.5021957159,-0.1471415311,-0.3258623183,0.1392658502,0.0646996722,0.3012734354,-0.037828967,0.0434902348,0.3261065483,0.0395886116,0.2853697836,-0.3048215508,-0.0612492934,0.1293693781,-0.3858837187,-0.207326442,0.4110406637,0.1310723275,0.0106387213,-0.1130714864,0.4067870975,0.1465439796,-0.0204076506,-0.2243354172,-0.2361697406,0.1568973809,-0.0847803056,-0.229190886,-0.2032954246,0.0412322916,0.1294880956,0.1418613344,0.1586416364,-0.2724761665,0.0539743006,0.433634758,0.0812699646,-0.0223389436,0.1786754131,-0.0640674978,-0.3206694722,0.0971370861,0.0145480074,0.3622031212,0.3019005358,-0.0451056398,0.1357600987,-0.3348754644,-0.0767820105,0.0611638278,0.064834334,0.1726130247,0.2490399033,0.3392724097,0.1360774487,-0.3790141344,-0.2310668826,0.1009448841,0.3705965579,-0.4107787907,0.2328082919,-0.5654392242,0.0682446212,-0.1355815381,-0.0026652142,-0.2121079564,-0.2839313447,0.1905758828,-0.0255046487,-0.1566009223,0.1916968375,-0.0622645728,0.1792233735,0.042999357,-0.2946195304,-0.4987807274,-0.3976953924,-0.361844182,0.0726953521,0.0359174013,0.1968455762,0.1091454104,-0.3954747915,-0.267641753,-0.1785188019,-0.2480862737,0.0385909453,-0.0182739533,0.3933659196,-0.1042994037,0.3079154491,0.01678426,-0.1031174734,0.3731988072,-0.1594765782,-0.1894491464,0.1921914071,-0.0201293398,0.2859966755,-0.1708589047,-0.3231950104,-0.1922650784,-0.3542550206,-0.0851473436,0.1331684142,-0.0649012774,0.1285061538,0.0572508425,-0.0619811751,-0.3311877549,0.2320606112,-0.141649574,-0.2909179926,0.3566639721,-0.2580299377,-0.4099886417,-0.0560200587,0.0336855054,0.0038646574,0.1677842438,-0.1878620535,-0.1797359437,0.0439801328,0.0636100173,-0.1022412255,-0.3338511884,0.2544527352,-0.0195600595,-0.1497008204,-0.0398604199,-0.1063863933,-0.0719658583,0.2644098401,-0.0550280772,0.2287327051,0.2309377044,0.0191080179,0.5462126136,-0.1312200427,-0.1270393878,0.6752419472,-0.3085885644,0.2448028922,-0.1913854629,0.1500792205,-0.1392898709,0.163982302,-0.0640363321,0.4088273048,-0.0085132401,-0.2498831898,-0.1730188578,-0.2420633286,-0.5097332001,-0.2107598931,0.0797342211,0.3322928846,0.1328580976,0.0374162197,0.093588315,-0.1892618537,-0.0967996344,0.24613823,0.1137009785,-0.0994739607,0.101370886,-0.1160054579,0.3933271468,-0.19981426,0.0003155158,-0.183611095,0.2889014482,-0.1266645193,-0.2722315788,0.1253264695,-0.2978845835,0.4570964575,-0.1393439025,0.4313631952,0.1596315056,0.2084527314,-0.7821110487,-0.2712784111,-0.116531767,-0.1613186449,-0.1825954616,0.5363669395,-0.4510548711,-0.2448557764,-0.0667103156,0.1571066231,-0.0744473487,-0.1041083932,-0.2140115201,-0.1238103285,-0.2010342628,0.2292434573,0.2167951167,0.5153964758,0.1347245574,0.1013604775,0.0044893981,0.2450048029,-0.2120156437,0.0690344125,0.4042893946,-0.3083541691,0.2448140681,-0.1298057139,0.1602814645,0.2333973497,0.3590833545,0.1447647363,-0.3983040154,0.0918503851,-0.0644038022,0.1998587698,0.1134951562,0.0331702121,-0.2388593405,0.2070671171,0.1435539126,-0.0894528925,0.2865354717,0.1903067082,-0.3289343417,-0.1299386173,-0.0705431998,0.4153870046,0.1418311,0.0720458478,0.0884489566,0.1744685471,-0.1770568937,0.654321909,0.3505808711,0.8752934337,0.1466898769,0.1826487184,0.6168476343,-0.4747037292,0.0518460944,-0.3435702324,-0.016636258,-0.2856911421,-0.226630196,0.0760395452,-0.1182271019,0.0768920034,0.5537091494,-0.0774042085,-0.0146157099,-0.1878613681,0.3469762206,0.1196667925,0.1804137081,0.1201629266,-0.0118263671,-0.4504956901,0.1670679748,-0.1673784703,-0.291362524,-0.1048002094,-0.2231215537,-0.2026486397,-0.1706275493,-0.2006442994,0.0875374004,-0.0596438237,-0.0937154889,0.0707079768,-0.4285850823,-0.2822828889,0.1662148833,-0.044940751,-0.2581043243,-0.0653549135,0.2501985133,-0.0336803868,-0.0489100963,-0.1062085629,-0.1479503661,0.2996114492,-0.1795483083,-0.1232356653,-0.1307693869,-0.3396779299,-0.1974482685,0.417812109,0.1991950423,0.2284066528,-0.0201991834,-0.0582342185,-0.1484427899,0.3907301724,-0.1985546201,0.1828110516,-0.020065872,-0.0223799739,-0.000300522,-0.1273802221,-0.4222848713,-0.3272140324,0.4109066725,-0.2652904093,0.4898923337,0.3383811116,0.3565918207,-0.3823345602,-0.1648204327,-0.0386129059,0.4703432322,-0.7984442711,0.0922832415,0.0204677582,0.2143622488,0.1994909495,0.268073529,0.1414806396,-0.241323933,0.0730182603,-0.4873651862,0.086264126,0.0245634913,-0.0697451606,-0.000755875,0.5695537925,-0.0068625091,0.0584739931,-0.3465655446,-0.3637188673,-0.1675873399,-0.0854966342,0.3382076323,0.1061993763,-0.1070086062,0.1598103046,0.1119166389,0.2124246359,0.0454903096,-0.073748894,-0.243686378,-0.1241589636,0.097991392,0.1985567808,0.0239714868,-0.0402043387,0.0892224982,-0.07815855,-0.0522341169,0.1997665912,0.2309246361,-0.0903074294,0.2646777034,0.6699783802,-0.1792663336,-0.4867533445,-0.1989911497,0.0251502711,0.1268214881,-0.0812057331,0.2192850411,-0.1263088137,-0.025605211,-0.2016952485,0.093462877,0.0198981892,0.0587094463,0.1900307834,-0.2137476951,-0.0726272687,0.2158380151,0.4578569531,0.1843313575,-0.0724706724,-0.3975322247,0.5031764507,0.1826037914,-0.2685511708,-0.3165690899,0.2468528897,-0.1336969286,-0.0477196239,0.0266932957,-0.0381873623,0.0752735883,-0.1198404655,0.2607041895,0.504401207,-0.1926892102,-0.0878861472,0.2724527121,-0.3107531071,0.1685589701,0.2133750618,0.1483507901,0.1025044098,0.1489689797,0.20622316,0.1051956341,0.3992113471,0.1431295574,-0.473456949,-0.3598192036,0.0699967518,0.3366046846,0.2457112074,0.0535413139,0.3172211349,0.266387403,0.4350030124,-0.1518821418,-0.2244282216,0.2220335752,0.016202312,0.1183996499,-0.0340054631,-0.0982813165,-0.1239158064,0.0900802687,-0.21148628,0.0601062886,0.3543044627,0.1013027951,0.114544794,-0.1070043296,0.035315115,-0.0315501802,-0.0463947058,-0.1057288721,0.3186493814,-0.0791990161,-0.2017089576,0.026172813,-0.0824860111,0.4072450697,0.068078652,0.0239264611,-0.0457544699,-0.0028744272,-0.0268376693,0.0753068998,0.2928326428,0.0645724535,-0.128736347,0.4553866088,0.2345588505,-0.2379465401,-0.0359845385,0.1916193664,0.069648467,-0.1054758728,0.0718202889,0.3355532289,-0.0562219433,-0.0824100152,0.0122020571,0.0747931153,-0.1825690717,-0.005066907,0.2672432661,0.1582802981,-0.1860693246,0.073324278,0.2545598149,0.4876082838,0.3626698852,0.0654503107,-0.2099902481,-0.1265234649,-0.6451534629,0.1798093915,-0.3399054706,0.2545686662,0.2821106911,0.0264912806,-0.0073886472,0.36225003,0.3437770605,0.2307071686,-0.1907526553,0.261398524,-0.2055655122,0.1758830696,0.30039379,0.0609397031,0.0205494203,-0.1969158947,0.159442842,-0.2053804547,0.227088213,-0.1554300934,-0.4755373299,-0.0215598438,-0.1579973847,0.5580276251,-0.0796601698,-0.0754231662,-0.0927572399,0.0953967571,-0.3974444568,-0.2756851017,-0.2345208079,0.2786346972,-0.3894477785,0.0357938036,-0.1875647604,-0.1576785594,-0.1896101683,0.1011595353,-0.1426366717,-0.1834889054,-0.113284409,0.2018488348,0.2061236054,-0.0246549379,-0.1752969921,-0.1790168434,0.1375781894,0.0488893948,-0.2294315696,-0.5924068093,0.3508055806,-0.3560836017,-0.1072248295,0.2015565932,0.0038441429,0.1043514758,0.012900807,-0.0146530587,0.1217737943,0.3825568855,0.0349973664,-0.251714021,-0.2496683151,0.1123021021,-0.1267711222,-0.0876492113,-0.0726160333,0.0274497513,-0.106626831,0.0420327261,-0.1960916668]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2842","title":"always requiring the username in the dataset name when there is one","comments":"From what I can understand, you want the saved arrow file directory to have username as well instead of just dataset name if it was downloaded with the user prefix?","body":"Me and now another person have been bitten by the `datasets`'s non-strictness on requiring a dataset creator's username when it's due.\r\n\r\nSo both of us started with `stas\/openwebtext-10k`, somewhere along the lines lost `stas\/` and continued using `openwebtext-10k` and it all was good until we published the software and things broke, since there is no `openwebtext-10k`\r\n\r\nSo this feature request is asking to tighten the checking and not allow dataset loading if it was downloaded with the user prefix, but then attempted to be used w\/o it.\r\n\r\nThe same in code:\r\n\r\n```\r\n# first run\r\npython -c \"from datasets import load_dataset; load_dataset('stas\/openwebtext-10k')\"\r\n# now run immediately\r\npython -c \"from datasets import load_dataset; load_dataset('openwebtext-10k')\"\r\n# the second command should fail, but it doesn't fail now.\r\n```\r\n\r\nPlease let me know if I explained myself clearly.\r\n\r\nThank you!","comment_length":30,"text":"always requiring the username in the dataset name when there is one\nMe and now another person have been bitten by the `datasets`'s non-strictness on requiring a dataset creator's username when it's due.\r\n\r\nSo both of us started with `stas\/openwebtext-10k`, somewhere along the lines lost `stas\/` and continued using `openwebtext-10k` and it all was good until we published the software and things broke, since there is no `openwebtext-10k`\r\n\r\nSo this feature request is asking to tighten the checking and not allow dataset loading if it was downloaded with the user prefix, but then attempted to be used w\/o it.\r\n\r\nThe same in code:\r\n\r\n```\r\n# first run\r\npython -c \"from datasets import load_dataset; load_dataset('stas\/openwebtext-10k')\"\r\n# now run immediately\r\npython -c \"from datasets import load_dataset; load_dataset('openwebtext-10k')\"\r\n# the second command should fail, but it doesn't fail now.\r\n```\r\n\r\nPlease let me know if I explained myself clearly.\r\n\r\nThank you!\nFrom what I can understand, you want the saved arrow file directory to have username as well instead of just dataset name if it was downloaded with the user prefix?","embeddings":[-0.1190379634,0.1679278612,-0.0471789353,0.083183147,0.0581347011,-0.1882434636,0.2695405185,0.0901198387,0.0316710472,0.300450474,0.1571788788,-0.0406884216,-0.0339300595,0.0137734013,-0.2211274058,0.0345355049,-0.037572436,0.4607703388,-0.1611818522,-0.0188344046,-0.2140433043,0.0586200505,0.0266007949,0.1466935724,0.092917338,0.058842849,0.3130096197,0.2868224382,-0.346788317,-0.4643648267,0.3015707433,0.2356016636,0.0029155812,0.2852761447,-0.0001282645,-0.0333672389,0.2038314193,-0.0626532808,-0.4011749923,-0.4367166162,-0.5333263278,-0.2659046352,0.3054177761,-0.3457644284,0.0780374855,-0.2222444117,0.0719516128,-0.7639085054,0.5935953856,0.173073858,0.0422012918,0.2400033772,-0.058429502,0.0676861182,0.0910241902,0.1828541756,-0.1983405799,0.0576104783,0.2272553295,0.1533893198,0.2952234745,0.2544378638,-0.1423670352,0.1849683672,0.4132423401,-0.2334423661,0.1443344504,-0.1818229556,-0.0621791855,0.4150875509,1.119764924,0.1163945422,-0.2934359312,-0.1209268048,-0.1289778799,0.0856915265,0.4294357896,-0.028681282,-0.2189789265,0.0680714622,0.049464494,-0.2302126884,-0.2342119068,0.0908740088,-0.2012763619,0.4058734477,-0.029753875,0.2471694201,0.0076992027,-0.1546998769,0.1335837245,0.0433808118,-0.0445635058,0.0337282456,-0.3212507665,-0.1692341864,0.090835236,-0.4048357606,-0.0476178452,0.5174331069,0.4020608962,0.0327300504,-0.3441611528,-0.0132745011,0.1745274514,0.190897122,0.6876799464,-0.0796756595,0.4477432966,-0.2132449448,0.1390418559,-0.1412651986,-0.2152579576,-0.2961265147,0.1313990057,0.0904852152,0.6935943961,-0.2948879004,-0.1523353606,0.0342134871,-0.1434054077,-0.2593048513,0.0486348569,0.0749922767,-0.0735887289,0.1086224318,0.1906049252,-0.0469755083,0.2274105549,0.0317912996,0.0382744707,-0.4006378651,-0.1337124705,-0.0184564274,0.2508587837,-0.3766240478,0.4235996902,-0.069896698,-0.0486346111,-0.0030262368,0.2811372578,0.0505682491,0.1693538427,0.1912305951,-0.0317290314,0.1745623946,0.0232970379,-0.2015249431,-0.2798311412,0.4101695418,-0.3928644657,-0.2087896764,-0.458591193,-0.0268326569,-0.2381677628,0.0136597864,-0.1689317673,-0.3450929523,-0.061842192,-0.0526142754,0.2095700502,0.2401569933,-0.061588455,-0.1526175439,-0.0715633631,0.3001872301,-0.2423455864,-0.1067422256,-0.2043185681,-0.0189731102,-0.158886537,0.1915175617,-0.4447416663,0.0448998325,-0.2232249677,0.1948380321,0.5952612162,-0.495640099,-0.0760816261,0.1809930801,-0.1531474739,0.0020545234,0.5405682325,-0.1667470187,-0.1783379763,-0.273727715,-0.5374604464,0.2701226771,0.2313288748,-0.0610755235,0.0770300031,-0.1622279584,0.0180178434,-0.2001359463,-0.237470299,0.0979825258,0.1585738361,-0.1249815747,0.3110917807,-0.2299071401,-0.0763425902,0.0260678306,0.280485779,0.1328603923,-0.0561477169,0.1408209652,-0.3887729943,0.209692806,0.0229794681,-0.0062835971,-0.1900233626,-0.3237221539,-0.2919830382,-0.0033373197,-0.1393821537,0.4665467143,-0.1532784253,0.2935825884,-0.2240790874,-0.1394454837,-0.1043931916,0.3928515017,-0.061280597,0.1351870298,-0.2509830892,-0.0169803556,0.136933744,0.2311661541,-0.049291268,-0.168064788,0.3726317585,-0.0220886804,-0.1133118421,0.5330518484,-0.0282319766,0.0319849886,-0.1048781797,0.1954684258,0.1114103943,-0.2514074445,0.1943351328,0.1735413671,0.1979680359,-0.110180378,0.0263108853,0.1806216091,-0.0178217087,0.0831885189,0.0754430369,-0.1566095501,-0.0618395284,-0.2291377485,-0.3702287674,-0.1156575456,0.0781684741,0.028383391,0.1068357155,-0.0578241237,-0.2347726524,-0.2743072808,0.3148424625,0.0879814997,0.1735295206,-0.0414846316,0.0935693234,-0.0817902759,0.1874943078,0.0105516203,0.414855361,0.1526489854,0.1386745274,-0.2018697709,0.3261217773,-0.1958365887,0.4617404044,0.0994198173,-0.1915585101,0.0323105305,-0.2889240384,-0.000456579,-0.2048773319,-0.1207685098,0.1294811517,-0.1390759945,-0.5578092933,-0.1775350273,-0.1947994381,-0.177322939,-0.0711520389,-0.389251858,-0.1408149898,-0.0222053286,-0.0295884032,0.3878869712,-0.0663037002,0.3042517006,-0.2504909933,0.2322497219,-0.2513068914,-0.4720901251,-0.2935732305,0.1879668981,0.1781593114,-0.0206833128,0.1983250529,-0.0375114866,0.3403944969,0.0168807711,0.3036705256,-0.4464102685,-0.2624958158,-0.1112394556,-0.0941314921,0.0098525006,0.1129646525,0.2278973162,0.1429516822,-0.3494535685,-0.0963938758,-0.130720526,-0.140706256,0.0009915548,0.1941831559,-0.0720808953,0.0421999432,-0.0923184529,-0.0686048046,-0.1738803238,0.1846844405,0.0589118898,0.3326393366,0.4182104766,0.024567036,0.0229594819,0.0603554808,0.0939202309,-0.1221084967,-0.4152963758,0.2814902365,-0.221095562,-0.0365174599,0.1737717241,0.232513532,0.0525959171,0.0665095448,-0.2961342633,-0.0420534499,-0.2180652767,0.4774214327,-0.2487527877,0.2532512844,0.567563653,0.3703916371,-0.1215839982,-0.1645440161,-0.0928736031,0.1102988049,0.0533315688,0.1783058792,0.1550331414,-0.0761385411,-0.3044638634,0.702835381,-0.0307355858,0.3346610367,0.5488693714,0.1978102028,0.3732857704,0.0515313633,-0.3388848603,-0.3989714384,0.1529436857,0.2582511008,0.1570234448,0.0615501143,0.1797024906,-0.3803019524,-0.1841921508,-0.1308833957,-0.2201764733,-0.0360447168,-0.3498950601,0.0543982945,0.0439988412,-0.1236758083,-0.1354579479,-0.0775798485,0.0839839429,0.4269081652,0.156512171,-0.0896431953,-0.4259313047,0.3228061199,-0.1297939271,0.3699968159,-0.0988408849,0.2729541957,0.1579927802,-0.0058774385,0.1153095588,0.2723063529,0.3054022789,-0.2086634785,0.3312802315,-0.1204936653,0.4135695994,-0.1751645505,0.031526763,0.2564969361,-0.0998786241,-0.2643826008,0.3027749956,-0.1913681328,-0.4059071839,0.1005333439,-0.2781323195,-0.0048160353,-0.0812905505,-0.2858083546,-0.2675290108,-0.0237786993,-0.1971609592,0.1056349501,0.076451458,0.2602632046,0.1318935901,-0.5443438292,-0.1031771302,0.3608002961,-0.1166566163,0.0351510048,0.1044920981,0.1588217318,0.0822712034,0.1861937642,0.2327169925,0.6167235374,-0.2998134792,-0.3326890171,-0.2465224266,-0.0142621538,0.3358976245,0.5567632914,-0.3770553172,0.0642206445,0.0524342544,0.0801536962,-0.1524664015,0.2843641341,0.0053211995,0.0515178256,-0.4108067155,-0.3268089294,0.542955935,0.1056067795,-0.300545007,0.7662752271,0.3692923188,-0.4027783573,0.1413379461,-0.0247143041,1.2238194942,0.2662164271,0.0439252891,0.0607517995,-0.4458107054,0.2821810544,0.0547335669,-0.1874690056,-0.1474926323,-0.303496927,-0.3116855323,-0.1509777904,0.568595767,0.1349815577,-0.1568014324,0.0573802292,-0.1468241811,0.0361378826,0.1009627208,0.7443876266,-0.1851266325,0.04694115,0.1732179075,-0.0364801101,-0.0447541773,0.2468228787,-0.0116908019,-0.1113996282,-0.1487347335,-0.5499303937,-0.0749664903,-0.0680466294,-0.7534858584,-0.0093349172,-0.379172951,-0.3681637645,0.5626470447,0.1523161083,0.0576119088,0.2965051234,-0.3175480068,0.255210191,0.0901815891,0.0149815697,-0.1803237647,0.1127088219,0.260605514,-0.0502237342,-0.0617956854,-0.0783602297,-0.283724755,-0.3321819901,-0.0255018845,0.0529483482,0.3253266513,-0.574577868,-0.0056035267,0.012261223,0.0434932373,-0.0465569086,-0.0350842215,0.0479125679,0.0039653662,0.1144410595,-0.2168416977,-0.3006615341,0.1790729016,0.2455641478,0.0463701449,0.0019493644,0.5263770819,-0.5754947066,0.0289999861,-0.1034607664,0.1901353896,0.2181659043,-0.4234993458,0.1825539023,0.3277737498,-0.182595998,-0.1071861684,0.3787519038,0.2603489161,0.3952935338,0.162694782,-0.1714950502,0.0690461472,-0.0812488645,-0.0692831352,0.3977308869,-0.2079311609,-0.2025918663,0.0163096469,0.0052000084,-0.161601454,0.138897121,0.2977437675,0.1350883842,0.1676586568,0.1225025132,0.2940799892,-0.2935943604,-0.09258084,-0.1401331574,-0.2513011098,-0.0639976412,-0.3148651421,0.3228757977,0.1089398041,-0.3098429739,0.1659444571,-0.442240566,0.0485299453,0.0441705473,-0.0342233293,0.1821005493,-0.0535595268,0.0724416599,0.3745588064,0.0742093921,-0.329554528,0.3735235929,-0.0287459753,0.19487679,-0.1757368892,0.2849723101,-0.2727879286,-0.0840634704,-0.0693758503,0.1563986391,0.239205569,0.1116321161,-0.1627604663,0.0437552892,-0.3610872924,-0.136868,0.2759779692,0.4564221203,-0.0846002474,-0.34220469,0.3734611273,0.0350232683,-0.1578491777,-0.1395681798,-0.0506657399,0.028796684,0.0680078715,0.2245224267,0.0949353799,-0.1718696654,0.0033878472,-0.0594214611,0.1257727742,0.2444269955,0.4302337468,0.5527543426,-0.1809055507,-0.2369037569,0.5231280923,0.0897443891,-0.0738714635,0.0243310891,-0.3417432308,0.2599059343,0.315195322,0.0353340469,-0.2184716314,-0.1708521694,0.1869697273,0.0286614317,-0.2586018741,0.2146742791,0.2728356421,0.2644332051,-0.1238499805,0.0831329077,-0.0688410774,0.1084795892,0.1066413075,-0.1596774608,-0.2735556364,-0.1660283357,-0.1107291281,-0.2326799929,-0.1443534493,-0.2382266223,0.0144023178,0.1292829812,-0.0178129598,-0.1071006209,-0.0757050067,-0.0576787591,0.1540258378,-0.3662449121,0.0070559145,0.1383771002,-0.0378998555,0.3139425218,0.5676707625,0.289681077,0.1228693351,0.2602412403,-0.0290090386,-0.0165426545,-0.0719606504,0.0499388836,0.3027605414,0.0829568058,0.0857042223,0.1471986175,-0.0053595686,-0.0027512922,0.0145407096,-0.0830346793,-0.4029019177,-0.3799490035,0.3665400445,-0.0053553027,-0.3656518161,0.0106148226,0.0804206654,0.1012569815,0.1627473533,0.5316871405,0.1874831468,-0.0080828816,-0.1555690914,-0.0273117982,-0.011067301,0.3597064316,0.2601287663,0.0083899647,-0.3220818043,-0.141349718,-0.6859127879,0.0416158698,0.1635604352,-0.1681409627,0.2027796209,-0.113917917,0.1069934443,0.1008727327,0.1230739281,0.0877577141,-0.0730656534,0.2201381624,-0.3137085736,0.1525007635,0.451246649,0.1160837561,0.0508197434,-0.6453543305,0.2798843086,-0.0242338665,-0.1862506419,-0.1720719188,-0.1087183952,0.4785988033,0.0000968219,0.7948836684,0.0299061965,0.3624134958,0.1474067122,0.1536778808,0.0480755605,0.0920405239,-0.4106507599,0.2657963634,0.6018586755,0.1037251875,-0.3340429068,0.1935338974,-0.3498045802,0.2718960643,-0.1413111091,0.2120958865,-0.1843793839,-0.2185983509,-0.0732575357,-0.0793107077,-0.0453754924,0.0925312266,-0.1731414348,0.1865433007,-0.6228410006,-0.5094375014,0.6428922415,-0.0273556225,-0.1840174496,-0.0626870766,0.7091805935,0.1909339726,-0.0494986698,-0.7029519677,0.1149335876,0.4967623949,0.0326750912,0.0374151506,0.3120481968,-0.1207026169,0.2600691319,-0.0263782069,0.4791584909,0.0988166481,-0.2374482751,-0.1279355139,-0.3826912344]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2842","title":"always requiring the username in the dataset name when there is one","comments":"I don't think the user cares of how this is done, but the 2nd command should fail, IMHO, as its dataset name is invalid:\r\n```\r\n# first run\r\npython -c \"from datasets import load_dataset; load_dataset('stas\/openwebtext-10k')\"\r\n# now run immediately\r\npython -c \"from datasets import load_dataset; load_dataset('openwebtext-10k')\"\r\n# the second command should fail, but it doesn't fail now.\r\n```\r\n\r\nMoreover, if someone were to create `openwebtext-10k` w\/o the prefix, they will now get the wrong dataset, if they previously downloaded `stas\/openwebtext-10k`.\r\n\r\nAnd if there are 2 users with the same dataset name `foo\/ds` and `bar\/ds` - currently this won't work to get the correct dataset.\r\n\r\nSo really there 3 unrelated issues hiding in the current behavior.","body":"Me and now another person have been bitten by the `datasets`'s non-strictness on requiring a dataset creator's username when it's due.\r\n\r\nSo both of us started with `stas\/openwebtext-10k`, somewhere along the lines lost `stas\/` and continued using `openwebtext-10k` and it all was good until we published the software and things broke, since there is no `openwebtext-10k`\r\n\r\nSo this feature request is asking to tighten the checking and not allow dataset loading if it was downloaded with the user prefix, but then attempted to be used w\/o it.\r\n\r\nThe same in code:\r\n\r\n```\r\n# first run\r\npython -c \"from datasets import load_dataset; load_dataset('stas\/openwebtext-10k')\"\r\n# now run immediately\r\npython -c \"from datasets import load_dataset; load_dataset('openwebtext-10k')\"\r\n# the second command should fail, but it doesn't fail now.\r\n```\r\n\r\nPlease let me know if I explained myself clearly.\r\n\r\nThank you!","comment_length":115,"text":"always requiring the username in the dataset name when there is one\nMe and now another person have been bitten by the `datasets`'s non-strictness on requiring a dataset creator's username when it's due.\r\n\r\nSo both of us started with `stas\/openwebtext-10k`, somewhere along the lines lost `stas\/` and continued using `openwebtext-10k` and it all was good until we published the software and things broke, since there is no `openwebtext-10k`\r\n\r\nSo this feature request is asking to tighten the checking and not allow dataset loading if it was downloaded with the user prefix, but then attempted to be used w\/o it.\r\n\r\nThe same in code:\r\n\r\n```\r\n# first run\r\npython -c \"from datasets import load_dataset; load_dataset('stas\/openwebtext-10k')\"\r\n# now run immediately\r\npython -c \"from datasets import load_dataset; load_dataset('openwebtext-10k')\"\r\n# the second command should fail, but it doesn't fail now.\r\n```\r\n\r\nPlease let me know if I explained myself clearly.\r\n\r\nThank you!\nI don't think the user cares of how this is done, but the 2nd command should fail, IMHO, as its dataset name is invalid:\r\n```\r\n# first run\r\npython -c \"from datasets import load_dataset; load_dataset('stas\/openwebtext-10k')\"\r\n# now run immediately\r\npython -c \"from datasets import load_dataset; load_dataset('openwebtext-10k')\"\r\n# the second command should fail, but it doesn't fail now.\r\n```\r\n\r\nMoreover, if someone were to create `openwebtext-10k` w\/o the prefix, they will now get the wrong dataset, if they previously downloaded `stas\/openwebtext-10k`.\r\n\r\nAnd if there are 2 users with the same dataset name `foo\/ds` and `bar\/ds` - currently this won't work to get the correct dataset.\r\n\r\nSo really there 3 unrelated issues hiding in the current behavior.","embeddings":[-0.0112160547,0.1447476298,-0.039360974,0.0958193168,0.0011399824,-0.21814017,0.4295825958,0.1111663133,0.1069055125,0.3369613886,0.1158719733,-0.0281225555,0.012058734,0.0387605466,-0.1626236439,0.15405114,-0.0621639118,0.3452987075,-0.1390585154,0.0082569737,-0.1663803458,-0.0051505985,0.0509563424,0.0512073487,0.1581596732,0.0974032432,0.3119137287,0.1957424432,-0.2975040972,-0.4824235439,0.3003055453,0.3615881801,-0.0639261305,0.4107335508,-0.0001179238,-0.0590646155,0.2076957226,0.0588892549,-0.5383250713,-0.4455489516,-0.4917058647,-0.2482866049,0.2102007866,-0.3056302369,0.0239957049,-0.0770888329,0.0180540141,-0.7141695619,0.4935949147,0.1425914317,0.1235402897,0.4247736037,-0.0561008267,-0.0546838194,0.0977879092,0.0445507616,-0.1430555433,0.0673177168,0.2824338078,0.1683505923,0.3565243781,0.2082258761,-0.2156513482,0.1110855341,0.3078671992,-0.297311604,0.1378888935,-0.1850421876,-0.0196566731,0.4707064927,1.0010164976,0.1690532565,-0.4670833349,-0.1757358611,-0.1048424318,0.0698701069,0.4081428647,0.1350550354,-0.1117157638,0.1679628044,-0.0835038573,-0.0350223184,-0.0327410065,-0.0183726028,-0.2701371312,0.463075012,-0.0382801704,0.2346923351,-0.126982227,-0.1310454905,0.1787928343,-0.0570791326,0.0163661037,-0.0085559338,-0.4005134404,-0.1691485345,0.1516158283,-0.4226424098,0.0428753421,0.3494897187,0.2651739419,0.0839743465,-0.3256967366,-0.0293966886,0.2081443816,0.2185648829,0.5727408528,-0.1588097215,0.4727654457,-0.2071726769,0.2513425052,-0.0986485332,-0.0985091105,-0.1068097875,0.0423010625,0.0020768261,0.5621961951,-0.2326089889,-0.2164977491,0.0521594696,-0.0958086252,-0.2232666016,0.0548773855,0.0210059099,-0.1488406956,0.1968661994,0.084558256,-0.0511213467,0.1819668114,-0.1001203507,-0.0620180219,-0.3828084171,-0.1790466756,0.0472363196,0.2418147027,-0.4406692982,0.3566652834,0.0988026187,-0.114215441,0.0436313599,0.1698496491,0.0908757001,-0.0172396172,0.190460071,-0.0580173172,0.1688790023,-0.0506810285,-0.2279134095,-0.2390823513,0.3085813522,-0.390411824,-0.2045666426,-0.4336534142,0.1097150519,-0.2935425639,0.066132538,-0.0812161043,-0.2224237472,-0.0917920545,-0.0883756503,0.1027795821,0.1798978746,-0.1870242059,-0.1584708095,-0.0523947552,0.3964365125,-0.1756321788,-0.0502207018,-0.1365360767,0.029464392,-0.1163142025,0.1394523829,-0.4678532481,0.1224646345,-0.1508981884,-0.0489705689,0.3005219698,-0.5388397574,-0.0999098346,0.2233630568,-0.1950829178,0.2370555699,0.515113771,-0.1638976336,-0.1690343618,-0.3403670788,-0.38592875,0.2341131121,0.1532597393,0.0108526954,-0.0428429842,-0.2005463839,0.1890724003,-0.0827106163,-0.1046393141,0.1461027265,0.1451874524,0.0288746953,0.2422480285,-0.2445032746,-0.1366341412,-0.1447138786,0.2217039317,0.0762651041,-0.1103753522,0.0332951918,-0.4837906957,0.3472396135,0.1079088002,0.2498538941,-0.0062445723,-0.3379675746,-0.2230136245,-0.0397204049,-0.0805989653,0.3676057756,-0.0520953275,0.2502617836,-0.1500829607,-0.1581352502,-0.0297868438,0.5793233514,-0.1334762275,0.1298864037,-0.3625038862,-0.1167224422,0.0630655214,0.3549443483,-0.0472435132,-0.0926375985,0.3039370775,-0.0752965957,-0.032413844,0.4737949967,0.1176315397,-0.163392067,-0.2493177205,0.2496862262,0.2226758897,-0.0240225773,0.1172254756,0.0922882482,0.3392164409,-0.1424901634,-0.0014673754,0.2404722124,-0.0266190972,0.0524699837,0.0816142038,-0.2937871218,0.0200445335,-0.2223831415,-0.3907833099,-0.1125158742,0.0381230712,0.0196339674,0.0680680349,0.0361104049,-0.1875891834,-0.2777249813,0.2804121077,0.0181260668,0.0682562515,-0.0759634972,0.2254895121,-0.1054001525,0.2398348451,-0.136148423,0.3855210245,0.2088996172,0.059797354,-0.2161650807,0.2650007308,-0.2814117074,0.3381402791,0.0773393586,-0.3269681633,-0.0710419416,-0.1665609181,0.1119682863,-0.1857122034,-0.1120791584,0.2062527388,-0.144352138,-0.5838578939,-0.1948499233,-0.1586288065,-0.0823749378,-0.0362053961,-0.3682667911,-0.2206246853,-0.1217518374,0.0648033395,0.3796051443,-0.1369002163,0.1926307976,-0.2595559359,0.2505998909,-0.387212038,-0.3063415289,-0.2005381584,0.1017693058,0.0251546223,0.0505025312,0.2794833779,-0.0904417783,0.4055159688,-0.1425036788,0.0767796338,-0.284624368,-0.4424414039,-0.0563289598,-0.1785662323,0.1260644644,0.0438687876,-0.0142092248,0.2531159818,-0.3800748587,-0.1869846284,-0.1745903641,-0.0908754691,0.0464824699,0.1490806341,-0.2406558543,-0.0902432725,-0.1752282232,-0.0152530055,-0.2159507722,0.0440757349,-0.0034682797,0.2913244665,0.5777443051,-0.1085720286,-0.092646189,0.0440641046,0.2822085023,-0.241314888,-0.4220570326,0.1361677647,-0.1282571852,-0.1011924967,0.1170816571,0.2195591033,0.1298719347,0.1245224699,-0.3237330914,-0.0901662037,-0.2222660482,0.4953493774,-0.1317195296,0.2276004553,0.4823291898,0.3495170772,-0.1312914342,-0.1352271438,-0.110282965,0.0061191386,-0.0955961198,0.1160098016,0.1497413665,-0.1784745306,-0.2178835422,0.5322826505,-0.0081038838,0.1795237362,0.4474458992,0.2515381873,0.4841234088,-0.0553931706,-0.3686906993,-0.2375951111,0.1669606715,0.3007070422,0.3614526987,-0.0288800951,0.176593259,-0.4512162507,-0.0420400053,-0.1621522605,-0.2278546095,-0.1509097517,-0.2837168872,0.0334209949,0.1741854548,0.0232751034,-0.1574230939,-0.1108555496,-0.0661252439,0.3186627328,0.1897315383,-0.1305613518,-0.5399690866,0.3867037296,-0.1150231808,0.3552759588,-0.0691340864,0.4952971637,0.1704865545,0.0651916191,0.0256934334,0.2471552193,0.3881770372,-0.213127628,0.2539444566,-0.10538771,0.2757460177,-0.2027953267,-0.021842584,0.116800338,-0.1649671048,-0.1291738451,0.2650550008,-0.4018349349,-0.3482812345,0.1696344763,-0.2113065422,0.032765612,-0.2741826177,-0.2963881195,-0.3259498477,-0.0105901444,-0.2670188248,0.0302371178,0.1780674607,0.1067327112,0.0705287009,-0.4568783045,0.0152228409,0.4961725473,0.0117481602,-0.1752828211,0.231042102,0.3731124699,0.077202715,0.2146691531,0.2025267482,0.6153803468,-0.3113231957,-0.3651433587,-0.2468257546,0.1075051278,0.203677386,0.62523669,-0.1257189214,0.0555853657,0.0182659421,0.0174131803,-0.1480576396,0.2311881781,0.0695190728,0.0075260377,-0.5689130425,-0.5124384165,0.4097725153,0.1962919682,-0.3164054751,0.5638622046,0.2210734338,-0.3777572811,0.2363743186,-0.0685734302,1.1841133833,0.1511303782,-0.0222883169,0.1086927578,-0.2743471861,0.4086295664,-0.0190224443,-0.1535325199,-0.1012921482,-0.3914802372,-0.2432575375,-0.0974706039,0.4923288524,0.1395298094,-0.0809419006,0.1936919689,-0.1467929333,0.310796082,0.1649026573,0.8204007149,-0.1968409717,-0.0432446785,0.1305499524,0.0713707358,-0.0698817968,0.0980889425,-0.0604219474,-0.0893918797,0.0003169497,-0.5520802736,-0.0432141349,-0.0595422648,-0.521519959,-0.0043296497,-0.3476559222,-0.1348288059,0.6052544713,-0.0314831883,0.2230160832,0.4870773554,-0.3037205637,0.1740720421,0.026037287,0.107130222,-0.1010565981,0.0885669738,0.2871973813,0.0180479717,-0.1029232144,-0.0717549324,-0.2084172666,-0.0926948488,-0.1091305017,0.0516612567,0.2960455418,-0.6112048626,0.1321746111,0.1655411422,0.1464852393,-0.1314795315,0.0616511032,-0.0469370857,-0.1123209149,0.2089915872,-0.2543317974,-0.3418778479,-0.0353129581,0.3156788051,0.1183122322,-0.0468297787,0.324785322,-0.4340351224,-0.0146591244,-0.1239699572,0.044387918,0.0820216313,-0.3374632299,0.1304941773,0.3694277108,-0.1517460495,-0.1111420095,0.4006417096,0.2646324635,0.2343738824,0.2556479275,-0.0740642846,0.1905229986,-0.0857848376,-0.0271288007,0.366823256,-0.0700080246,-0.0851100907,-0.2316470295,0.1306413561,-0.271759212,-0.0667752326,0.2494487613,0.1456221342,0.1512948722,0.0332742296,0.2019618154,-0.3044810295,0.0085814279,-0.1379687935,-0.2515730262,-0.1664135158,-0.2767710388,0.2686150968,0.1358598471,-0.2527415454,0.132998243,-0.3781419992,-0.0350330472,-0.0503023304,0.2388755977,0.3519418836,-0.0465129316,0.0376931243,0.3187004328,0.2507348359,-0.2451188564,0.474457562,-0.0569196828,0.1913256645,-0.1051184237,0.3228104413,-0.2103770524,-0.0249673203,-0.0326478817,0.2123124301,0.2147752196,0.1101787686,-0.1119844392,-0.005165522,-0.2259212136,0.0477607399,0.2360938936,0.4404348135,-0.0166109838,-0.397975415,0.3906178176,0.1331166774,-0.2027976215,-0.1477350891,-0.131935209,0.0933493972,0.1392949522,0.3862106204,0.2862531245,-0.1766945869,-0.0275046695,-0.1304129213,0.3384794593,0.2866654694,0.4526553452,0.559271574,-0.1542936862,-0.1854736209,0.5126243234,0.0068702386,-0.1419386864,0.0054718694,-0.2569516599,0.3793473542,0.274857074,0.120821327,-0.1378637701,-0.1511700749,0.3385250866,0.1294661313,-0.2375845462,0.1839798987,0.159532994,0.0375085473,-0.2583964467,0.1217844337,-0.191684112,0.0509817526,0.05328428,-0.053939417,-0.3588397801,-0.1637411118,-0.1485710144,-0.2200826854,-0.1351292431,-0.3176749051,0.0408213772,0.0658945143,-0.1486790329,-0.2505571544,-0.176974386,-0.0058159158,0.1997225434,-0.387365222,0.0502134189,0.170920372,-0.0937477499,0.2696216702,0.6034386158,0.3446468413,0.2489085644,0.2829256058,-0.0589632206,-0.0136465458,0.007323496,-0.0637469292,0.4808591306,-0.0753983334,0.1981627494,0.216433987,0.0723053068,-0.0627668574,-0.0630355924,-0.0711401254,-0.4051997662,-0.5685397387,0.5356151462,0.0218324475,-0.2458160818,0.0212234408,0.1753309816,0.0452518947,0.2339673936,0.5319804549,0.2475272864,0.0989459231,-0.220238179,0.0302972402,0.0421724506,0.320638448,0.2086677402,-0.0390397236,-0.3553540111,0.0335700065,-0.6624810696,0.1259515435,0.0534823611,-0.0521552861,0.1798840463,-0.2214946896,0.1530091166,0.1559294313,-0.1151130423,-0.0013093128,-0.0822173208,0.1474504471,-0.2808152139,0.1885608137,0.3135068715,-0.0348808281,0.0735729113,-0.6237818003,0.2831662893,-0.0909055471,-0.1089206561,-0.1977215707,0.0677645952,0.5854008794,-0.1355735809,0.6277521253,0.181301266,0.4258820117,0.1772142798,0.0965598673,-0.111102879,-0.0380102471,-0.413700521,0.244944185,0.4392441809,-0.1091395468,-0.1369019002,0.0600056089,-0.2991527319,0.3612939715,-0.0526058488,0.2648939788,-0.2576389909,-0.2502010763,-0.183387816,-0.0939578414,-0.0225651022,0.1579710096,-0.146263063,0.1357736737,-0.4803952277,-0.5864162445,0.5395393372,-0.107848525,-0.167233929,-0.2472512871,0.6490894556,0.0394797958,-0.2027115822,-0.527410686,0.1477640867,0.5112928152,-0.0624120533,0.0429566689,0.380864799,-0.1212850064,0.1125438362,-0.0627647713,0.1469632685,0.01512504,-0.371404022,0.017637834,-0.3359433115]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2839","title":"OpenWebText: NonMatchingSplitsSizesError","comments":"Thanks for reporting, I'm updating the verifications metadata","body":"## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n","comment_length":8,"text":"OpenWebText: NonMatchingSplitsSizesError\n## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n\nThanks for reporting, I'm updating the verifications metadata","embeddings":[-0.32564345,-0.0787488744,0.0416311324,0.4369296134,0.0632659569,0.1131168976,-0.0893040746,0.56578058,-0.0477535538,0.1396527141,-0.1303432435,0.0160794053,-0.0701626986,0.0707900152,-0.1303530782,0.0958568305,-0.1383086741,0.178688094,-0.2445953488,0.0833462924,-0.2331949174,0.238662079,-0.3176048398,-0.0529136099,-0.178170979,0.0412582904,-0.2001938969,0.1461547613,-0.2893433273,-0.4155524969,0.2356107086,-0.3650061786,0.0161703695,0.5185359716,-0.0001193553,0.0201675352,0.3312910795,-0.2299240083,-0.1270173341,-0.1205399781,-0.5110303164,-0.3880583048,-0.1229868382,-0.0879189298,0.2323465794,-0.1369426548,0.1524888724,-0.306979537,0.3225004375,0.3693405092,0.1601023823,0.4129099548,0.2254389822,0.021552898,0.3561557233,-0.3026716113,-0.1578760892,-0.0497807376,0.0644101724,0.0467742868,0.0078994706,0.0679835007,0.0253639258,0.2295645922,0.1037727892,-0.0813513994,0.3687720001,-0.2027950585,0.1101161689,0.4001135826,0.7251262665,-0.254214555,-0.3288719952,-0.2818359435,-0.4098591506,0.0485786758,0.2816064954,0.4939975441,0.0064983065,0.2092936784,-0.3392004669,0.0480482429,0.0248999894,0.1024883464,-0.2312752157,0.2969220281,0.0151579259,0.1188931093,-0.0856369436,0.0909081846,0.2815484703,-0.1452156156,-0.0107437195,0.164334178,-0.2652312815,-0.0475808568,0.0181642398,-0.2479332834,0.3492283225,0.3527331948,0.3432183862,-0.3055211902,-0.192064777,-0.0125056403,0.1979494095,0.444098413,-0.0279367957,0.2162253261,-0.186285302,0.0570947379,0.1257350445,-0.0588351376,-0.0491400436,-0.3709675372,0.0928653181,0.0482338257,0.6038799882,-0.2576237321,-0.6360960007,0.1065674201,-0.3070550263,-0.2868615985,-0.3296884,-0.1251932234,-0.3409199715,0.2569748759,0.1459303051,-0.1809077561,-0.2902568281,-0.3806399107,-0.2934208214,-0.0562996268,-0.1521889716,0.0781796873,-0.0896609351,-0.082058847,0.3928046227,-0.1357023567,-0.0144651448,-0.2292023301,0.1582430452,0.0301775411,-0.0029213957,0.1892298162,-0.1762201339,0.1275197715,-0.0034348734,0.0962300599,-0.3501027822,0.3865600526,-0.2331618369,-0.3938328028,-0.1949294209,0.1335442364,-0.3694763482,-0.2194156796,0.3746355176,-0.0974125564,0.2680396736,-0.1616348922,-0.1111938655,-0.1827877909,-0.1751283109,-0.1388487071,0.0836787596,0.3914381862,-0.2883639038,-0.2171717137,0.2311590761,-0.1092894375,0.3596263826,0.4289075434,0.0885053501,0.0088397749,-0.2161827683,0.2334822714,0.0998117998,-0.3768259883,-0.6785972714,0.3799767196,-0.1182072386,0.1913342476,0.128611818,-0.1917271912,0.047622133,-0.1238994002,-0.0667467564,0.1936255097,-0.2272973061,0.060548339,-0.5992887616,-0.7900931835,0.3947702646,-0.063123785,0.2197069973,-0.2049342543,0.0091069285,0.2623768151,0.486287266,0.1057340577,-0.009524175,0.1010838598,0.1586307287,-0.033941865,-0.0599811152,-0.2892177701,-0.1544214785,0.2258619666,-0.1643334925,-0.0389609374,0.2473222315,-0.1482165754,-0.54909271,-0.2039607763,-0.0497773364,-0.1486789435,0.0058486974,0.042424757,0.2637369931,0.0901236534,0.1350685954,0.3264502585,0.0346996002,0.1312265396,-0.2871116698,0.1862237304,-0.0674351677,-0.082457073,0.1553918421,-0.0730462819,0.1596544385,-0.0909905136,-0.3767090142,0.6851299405,-0.0080521759,0.126973927,-0.0209540874,0.1521532387,0.2290636748,-0.1977326274,0.131191656,0.3649396598,-0.0161853209,-0.0296166316,0.1177503616,-0.1503991187,0.1035164073,-0.0140225999,-0.0629861727,-0.031142151,-0.090478994,-0.1592348218,-0.0880051255,-0.146123305,0.0805791169,0.0339468122,-0.1628018022,0.0696686581,0.166589275,-0.1903180927,0.7757811546,-0.0441245958,-0.1039870605,0.2745919526,-0.3110389411,-0.1077228114,-0.1961886883,0.177483961,0.5276733041,0.2409575582,0.2021829635,0.2293022275,0.0225705542,-0.1215977743,0.1433231235,0.3013388216,0.2312081009,0.3488216102,0.0267643631,0.0699671805,-0.1673664153,0.0998206288,0.1198485568,0.3532116413,-0.4279280603,-0.2392227501,-0.4788658619,-0.0029847587,-0.3143445551,0.1545405537,-0.1642452776,-0.274020195,-0.0560249723,0.0417954139,-0.0652234107,0.2124938518,-0.239224419,0.1416530162,0.1828521639,0.1791202426,0.0933842212,-0.0148642175,-0.4930908382,0.0081185205,0.3259840906,-0.325419277,0.4200646281,-0.5714495182,-0.0197503902,-0.1275273114,-0.2239328325,-0.0891354233,0.0967677161,0.0591952428,0.2511111498,0.1434148997,0.1152847856,0.0002924891,0.1184536442,-0.2296147346,-0.255079627,-0.1175127402,0.0957346857,0.239151597,-0.0410392061,-0.5830168128,-0.2802815735,-0.3175727427,0.267768532,0.1152561978,0.2072443068,0.0615796708,0.0548707023,0.0056975153,-0.0092328396,0.1080072597,-0.2043620497,-0.4664912522,0.3166824579,0.0714167953,-0.2712298036,0.0126738995,0.256416589,-0.0446778126,0.0811583921,-0.5617858768,-0.2073500901,-0.3542568088,-0.0665474758,-0.1130505279,0.1958581656,0.4577782452,0.027986018,-0.0470941663,-0.1315357238,-0.3364978135,-0.0469963774,0.2581958175,0.5358541012,-0.1085429341,0.2180033773,0.3973363638,0.4600125849,0.5372806787,0.041154068,-0.0606082827,-0.0378829651,0.0961130187,-0.0800278112,0.0433650389,0.2252321988,-0.201586917,0.0214810502,0.3422173262,0.1082652658,0.122533761,-0.0004730427,-0.0970617086,-0.4285309315,-0.1527620852,-0.0802733153,-0.1112878099,0.2171887755,0.2329321802,0.0715923086,-0.2347638011,-0.2464827448,0.1354565471,-0.0045890836,-0.1431981474,0.0457561836,0.1936069876,0.3566380739,-0.3216092587,0.3874337077,0.1129570752,0.2624562979,0.1981515139,-0.2502702177,0.0953327939,0.0124490773,0.4080364108,-0.0591535382,-0.0573152713,0.074357979,0.1439663172,-0.4227099419,-0.3065178394,-0.1077827141,0.0364384092,0.5374099612,-0.0388277434,-0.4084883034,0.0206725355,0.3111313283,0.4094210267,-0.2320624292,-0.0502811149,-0.1600877196,-0.3571175039,-0.3017560542,-0.1640065163,0.0633711368,0.2974315286,-0.0089250868,-0.0683187321,-0.081477128,-0.1740711033,0.2115716189,-0.1440046579,0.2809153497,-0.0072218222,0.3390001953,0.2225872278,0.4083380401,0.2933750451,0.5753409863,0.0399561226,-0.2198426872,-0.0824764445,-0.0778567865,-0.2219630629,0.3040519059,0.0985011533,-0.3474415243,-0.009826568,0.170565486,-0.2748628557,0.2841237783,0.4026911855,-0.0993392542,-0.4422556758,-0.3618905246,0.448036015,0.2284169644,-0.0536170974,0.4399095774,-0.0560965948,-0.1993983537,-0.0351740383,-0.0688770562,1.0394178629,0.1332676262,-0.1174464971,0.2066734135,-0.5770884156,0.1889451444,-0.260543406,0.1740755737,-0.2990444601,-0.4450036883,0.0023927616,-0.2591135502,0.2143733054,-0.0442197658,-0.2002891898,0.2211339623,-0.0077696722,0.0683801845,0.1744779497,0.3442565501,-0.2637011409,0.1808253676,0.2966559529,0.0750432163,-0.1750821173,0.5030097961,-0.096618399,-0.1475435793,-0.0894345567,-0.1398896724,-0.5841290355,0.0426218025,-0.3402764201,0.2456290573,-0.1438892782,0.0775524005,0.0649253577,0.0523464084,0.1110498235,-0.0271709971,-0.3606568873,0.266343683,-0.0546128899,0.1516703963,0.2797600627,0.0689811707,0.1413345337,-0.1141878739,-0.2196468562,0.0418826453,-0.1090296358,-0.1005278528,-0.0551900342,-0.1821748763,0.3954920769,-0.6181647778,-0.154661268,-0.0731591061,0.1096793488,-0.1788410097,0.0799090713,0.1311027706,-0.072433874,0.4409589767,0.006746124,-0.4591042697,-0.1808200181,0.4458133578,0.114635393,0.0725739747,0.4087303877,-0.1945676953,-0.0564856976,-0.0052378997,0.4248681366,-0.1451554596,-0.4997896552,0.2524960041,-0.377926141,0.3413412571,0.0476019457,0.3166755736,0.5543801188,0.4230823815,-0.0041913255,-0.1374966353,0.0702374652,0.02195484,0.0486794449,0.3826151192,-0.0934591815,0.4492403269,0.0123212393,-0.0824739709,-0.2386768907,0.2522660792,-0.006313662,0.0311044939,0.1263194531,0.0359355137,0.1116643623,-0.040479403,-0.0491947792,0.2194381505,-0.2064635456,-0.1023788303,-0.1599382609,0.2150251865,0.1122192666,-0.109412387,0.0972078592,-0.0959301665,0.0959118679,-0.0661048293,-0.1156716868,0.1497949064,-0.0630305856,0.323399812,0.0677217618,-0.1879870296,-0.0506002083,0.2329144776,-0.2129188776,0.3926366866,0.0437697358,0.4021113515,0.1804009527,0.0005429484,-0.1165501997,-0.0354224928,0.0201880056,0.137938723,0.26992926,-0.211934045,0.1039823443,0.3449422121,0.1045872942,0.4217939079,-0.1827317476,-0.2936154008,0.1470703185,0.1735403538,-0.333981961,0.0166978203,0.4356484711,0.0674994737,-0.0713218972,0.0402785987,0.0992582589,-0.0365034863,-0.2805308104,0.1777226925,0.3709855974,0.1994870603,0.2782967687,0.8514111638,-0.0568386093,-0.0106518129,0.0311711188,-0.2115876824,0.2864341736,0.2603233159,-0.1862044036,0.3435190916,-0.0515650846,0.3375907242,-0.3570045829,-0.3758834004,-0.0913514197,0.2305946648,-0.0134398816,0.1718674004,0.2250275016,0.2972023487,-0.1338582635,0.0202030353,-0.4106066525,0.1846010685,0.0804802403,-0.0465590693,0.0563089475,-0.2348933071,-0.1628661901,0.0959311575,-0.1277887374,0.0229803808,0.290574193,-0.1822422296,-0.2544303536,-0.3419475257,-0.1527126431,-0.071968466,0.175948754,-0.4940128922,0.0769663826,0.1308668405,-0.2884328067,0.0894287378,0.6541632414,0.4437442124,0.2774838805,-0.0937441885,-0.1094948575,0.1787839532,0.1244596317,0.0756153241,0.2275884598,0.0269678812,-0.1599591374,0.2955514193,0.0459985174,-0.0748245791,0.0624247752,-0.268196702,-0.2347323447,-0.403103292,0.2820436358,-0.1490770876,0.0683924928,0.0182964262,0.0651223585,-0.3186810315,-0.1009630263,0.1417585462,0.000812163,0.2375723422,0.1097738594,0.0177985523,-0.2429865301,0.2181180269,0.4704924822,-0.3095637262,-0.160249427,-0.2318778336,-0.5288663507,0.2426986247,-0.1575481445,-0.2160964906,0.0095794424,0.3213546276,-0.2441986948,-0.0278440919,0.3870459795,0.2888368368,-0.2123795599,0.172443822,-0.3027431667,-0.1520148516,-0.0920263156,0.0277132206,0.1658377051,-0.2794111073,0.2063078284,-0.129235819,-0.0309759397,-0.2186823934,-0.0200429484,0.1742995083,-0.0219135154,0.4452887774,0.3286918998,0.3399773538,-0.037208505,-0.1565983444,-0.2059414387,-0.2183606476,-0.2168839872,0.4339432716,-0.0403193608,0.2017148435,-0.1970680505,0.0437811799,-0.1256293952,0.4511986971,0.0257218834,-0.0067428658,-0.212751776,-0.3006520271,-0.0633832961,-0.1204399168,0.3458029032,0.2411932796,-0.1523052305,0.1324471831,-0.0131171765,-0.4516012073,0.4643374979,0.0422768444,-0.0440979935,-0.142718032,0.1752106845,0.3098734915,-0.115505442,-0.4275672436,0.1146273464,0.6186627746,0.0115267765,-0.0637426376,0.1322178245,-0.0562384427,-0.0835311711,0.0753997713,0.1356141567,-0.0528067797,-0.4168241322,0.3720106184,-0.1234578863]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2839","title":"OpenWebText: NonMatchingSplitsSizesError","comments":"I just regenerated the verifications metadata and noticed that nothing changed: the data file is fine (the checksum didn't change), and the number of examples is still 8013769. Not sure how you managed to get 7982430 examples.\r\n\r\nCan you try to delete your cache ( by default at `~\/.cache\/huggingface\/datasets`) and try again please ?\r\nAlso, on which platform are you (linux\/macos\/windows) ?","body":"## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n","comment_length":62,"text":"OpenWebText: NonMatchingSplitsSizesError\n## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n\nI just regenerated the verifications metadata and noticed that nothing changed: the data file is fine (the checksum didn't change), and the number of examples is still 8013769. Not sure how you managed to get 7982430 examples.\r\n\r\nCan you try to delete your cache ( by default at `~\/.cache\/huggingface\/datasets`) and try again please ?\r\nAlso, on which platform are you (linux\/macos\/windows) ?","embeddings":[-0.2325653881,-0.1025962085,0.0594487116,0.4710067809,0.099145472,0.0983288512,0.001799254,0.5494816303,-0.0201674923,0.1316695958,-0.1176461726,0.0032898269,-0.1183154136,0.091308251,-0.0851507634,0.1036110669,-0.1839791536,0.1432445645,-0.2236059755,0.0837038904,-0.1660715044,0.2510773838,-0.2940531671,-0.136175856,-0.2839494944,-0.0273162425,-0.2651298642,0.2377939373,-0.2557264864,-0.4160161018,0.1758898795,-0.3987876177,-0.0372202508,0.5550357699,-0.0001198019,0.053337954,0.2946435809,-0.2280974835,-0.130750522,-0.1020404696,-0.4944339693,-0.3940509856,-0.1452365369,-0.1462632567,0.1926995069,-0.128324911,0.0508886799,-0.2793247402,0.45934093,0.4411795735,0.1552283317,0.397815913,0.2411608696,0.0371606573,0.3650759459,-0.2219923884,-0.131810233,-0.0362073407,0.0326931886,0.0405232348,0.0806393847,0.1172676533,0.010848294,0.208711639,-0.0005465942,-0.04086693,0.2286907285,-0.2362536341,0.1534994245,0.4934612215,0.6223924756,-0.136980623,-0.3544664681,-0.3586990833,-0.4166816473,0.0730490908,0.2820883095,0.4564271271,0.0207831394,0.2088509351,-0.3920858502,0.0627977401,0.0887393206,0.1010190621,-0.219721511,0.1737966686,-0.0613732822,0.0936699957,0.0337460227,0.1205220819,0.2228565961,-0.1298710406,-0.0552684665,0.130376339,-0.2470706999,-0.0399980396,0.0465838946,-0.1719465107,0.3725028634,0.3157573938,0.3307099342,-0.2253412157,-0.2792927325,-0.0260545649,0.1021305472,0.486409694,-0.0030222856,0.181135416,-0.0994259194,0.153441906,0.0841908902,-0.038324248,-0.0627070367,-0.3573563993,0.0782192796,0.0519293919,0.5506390929,-0.313762635,-0.5638862252,0.1355626136,-0.271795541,-0.2499685287,-0.2926042974,-0.0598699637,-0.3987960219,0.21013017,0.1625957936,-0.1578373164,-0.3149763942,-0.3888362348,-0.274458766,-0.0227039084,-0.1876686811,0.1219247207,0.0013910971,-0.1314144135,0.4226850271,-0.0792298242,-0.0405353419,-0.2938423455,0.16571486,0.011818096,0.0518370979,0.1675295681,-0.2069491893,0.1224142388,0.0686967969,0.0914774612,-0.2862041891,0.3105949759,-0.2906442583,-0.4093714058,-0.1381831467,0.1050514579,-0.3705904186,-0.160875082,0.3073712587,-0.0599066541,0.3331327736,-0.1210963205,-0.0874515325,-0.1715030372,-0.1489326507,-0.1385395974,0.136148721,0.4009007215,-0.2815036774,-0.1870612502,0.2009663135,-0.1067375243,0.2657059729,0.4419420362,0.0383116044,0.1077955142,-0.2697066367,0.1994151175,0.1035528705,-0.3222500086,-0.6370256543,0.3166928291,-0.0614389703,0.1890935451,0.1131784543,-0.2998300493,0.0997247845,-0.1133560538,-0.1074843481,0.1425459236,-0.1855406165,0.0653884783,-0.6042557955,-0.7404735684,0.4050538242,-0.0152202202,0.2199621052,-0.2009830177,0.0258024931,0.2023736984,0.4867910445,-0.0068813604,0.0082399799,0.1832945794,0.1684482396,0.0200096425,-0.0262848958,-0.3034965396,-0.1270769089,0.2578174174,-0.1189626977,-0.0252107475,0.2395293266,-0.1284899116,-0.5201492906,-0.2797634602,-0.027436683,-0.2890127301,-0.0011977525,0.0537340343,0.3314978778,0.070443891,0.1006710827,0.3630115986,0.0587042272,0.1622793376,-0.3886450529,0.2116378248,-0.0419298112,-0.1590762138,0.147472024,-0.0411890857,0.157839641,-0.1788159907,-0.3645519018,0.6570310593,0.1383420527,0.02780932,-0.0206335951,0.1761494279,0.2389790714,-0.232609421,0.1361853629,0.3155079186,-0.0367029049,-0.0358329862,0.0357469767,-0.0957837403,0.1101791039,-0.027026739,-0.0629577935,0.0031620259,-0.0677025169,-0.2016023397,-0.0996729061,-0.1194978952,0.0788498223,-0.0529226586,-0.1825649738,0.0738681331,0.1595273167,-0.1473862082,0.7876149416,-0.0265092328,-0.0810908601,0.2676542401,-0.2830883563,-0.1285564452,-0.183779493,0.1333068758,0.5281811357,0.2116181999,0.1979825199,0.2271813452,0.0595819056,-0.0967165157,0.1486262977,0.2788825929,0.2119769752,0.301492095,0.0253650583,0.0782746673,-0.1882289052,0.1017744318,0.116242066,0.3736377954,-0.4455088973,-0.1506477296,-0.499330163,0.0167021099,-0.2890394628,0.1519827545,-0.1953706294,-0.2340987176,-0.0569427386,0.0578246228,-0.0190754719,0.2200787216,-0.1844465733,0.1744143516,0.177943483,0.2221967876,0.0834514573,-0.0005307973,-0.5165869594,0.0194541551,0.3351318538,-0.249814257,0.4144765437,-0.5614904165,-0.0921603888,-0.201660797,-0.2970202863,-0.0893634483,0.0089835012,0.1129095405,0.292773813,0.2110989392,0.0249922592,-0.0226391368,0.1041431502,-0.320573926,-0.2713530958,-0.0989822373,0.0973132402,0.2661294639,-0.0641359314,-0.5627799034,-0.216370061,-0.2757206857,0.326700151,0.1139719784,0.202980265,0.0925925523,0.0694469884,0.1050399691,-0.0523555838,0.1160848886,-0.2542546391,-0.5217090845,0.3188512623,-0.0026786192,-0.3333176374,-0.0070148506,0.1511377543,-0.0392282307,0.0255860798,-0.6288368106,-0.2420555353,-0.4020483494,-0.0416228026,-0.1522629708,0.2767257392,0.4846800268,-0.030819742,-0.0632560626,-0.1522991359,-0.3440429568,-0.0237116944,0.1624833941,0.6210068464,-0.1053256467,0.1852699518,0.4056099951,0.4629566371,0.5385028124,0.0481480174,-0.0136958975,-0.0321367905,0.1083640978,-0.1104782,-0.0401165858,0.2920781076,-0.1999634057,0.0356370732,0.2973732054,0.1355915368,0.1561637968,-0.0030652257,-0.0703070462,-0.4779709876,-0.2007860243,-0.0245935228,-0.1111411899,0.1895927191,0.1881502569,0.0732369721,-0.2434804589,-0.1945626438,0.1598565429,-0.0046626697,-0.1435011923,0.020765001,0.0671190917,0.3567481041,-0.2626105249,0.3809084594,0.1341329366,0.2630775273,0.1486925185,-0.2384483814,0.1142474264,0.005393642,0.5108354092,-0.0777720734,-0.0712624267,0.0950103551,0.0725387558,-0.403919071,-0.3115305305,-0.1836721748,0.0334490202,0.5596083403,0.045381356,-0.4185118973,0.0558074079,0.3097251654,0.4309919775,-0.1897706985,-0.0285510439,-0.1574657857,-0.4026028514,-0.3548040688,-0.1288215369,0.0678973868,0.2978068292,-0.0918349847,-0.0646309704,-0.1403725445,-0.1978233904,0.196538955,-0.1222217157,0.2837555408,-0.0024830059,0.314786464,0.2665962279,0.352255255,0.2715790272,0.5941634178,0.0000016587,-0.2287432402,-0.050626453,-0.0613596663,-0.2377571166,0.3324658871,0.071513772,-0.3414058387,0.0568000898,0.247422725,-0.1786895394,0.3185909688,0.4256512523,-0.0635496825,-0.4400918484,-0.3524298072,0.3958845139,0.2119295299,-0.0502379648,0.5040233731,-0.0495395437,-0.1787898839,-0.0557667576,-0.0423582681,1.1023490429,0.1564745158,-0.1077495366,0.2527418733,-0.4704375565,0.2593771815,-0.2291093171,0.1341121495,-0.2876505554,-0.5424720049,0.0213857889,-0.2812007964,0.2467852831,-0.0431677885,-0.2238034606,0.1670550406,-0.0107187983,0.0595708787,0.1506897509,0.3264892995,-0.2793125808,0.1247564852,0.1668467671,0.0791604966,-0.1236406341,0.4728925824,-0.0779321864,-0.1369110644,-0.1266555786,-0.1621517539,-0.6035356522,0.054857716,-0.3796156347,0.2448875308,0.0042179055,0.050631199,0.0325433724,0.0662536696,0.2591130137,0.0856591314,-0.3663952351,0.236521557,-0.0726003945,0.1324077845,0.2462103367,0.011354005,0.1980232298,-0.1248903424,-0.2756637633,0.0476575159,-0.1618162543,-0.103570722,0.0045223096,-0.1750111282,0.3070130944,-0.6238247156,-0.0981967375,-0.0558684058,0.1288654506,-0.2135449052,0.0828685015,0.1996137351,-0.1079261824,0.4750551581,0.0140790306,-0.4118836224,-0.1699378192,0.4765078127,0.0673943311,0.0415952392,0.4440050721,-0.0966847464,-0.0302062351,-0.0609109029,0.4412086904,-0.0831031129,-0.474270314,0.227412805,-0.3301911652,0.3447002769,0.0210015997,0.3565444648,0.5420540571,0.4391899109,-0.0135716889,-0.2046358436,0.0159000345,0.0229217652,0.050128907,0.423733592,-0.0923482701,0.4085063636,0.0041386271,-0.0276931729,-0.2374144495,0.3453592956,-0.0479806401,0.0406343043,0.1409527957,-0.0342337973,0.144965902,-0.0253790226,-0.0578118153,0.2546966076,-0.1896177828,-0.1297222078,-0.1877345145,0.218500942,0.111317575,-0.1485711634,0.1090276912,-0.0480880216,0.0964983627,-0.065180175,0.0203978792,0.0826219395,-0.0614790805,0.2990712523,0.046455007,-0.1582014561,-0.0322582908,0.1964221299,-0.2496592551,0.3323284984,0.131318897,0.3523203433,0.1779555529,0.0039510238,-0.0866929516,0.005066534,0.0330318101,0.0843183324,0.2701558173,-0.2523284554,0.0954454169,0.3207378685,0.1293789744,0.3699232936,-0.1670776755,-0.1981525272,0.2500639558,0.1845028996,-0.2806071937,0.0274397749,0.4826545417,0.079204753,-0.061439041,0.1219393238,0.0515225008,-0.0403979048,-0.1860175133,0.1256128252,0.3057819903,0.1706677526,0.3275693357,0.7761434913,-0.1907459348,-0.0441235155,0.0329700969,-0.2791087627,0.2220818996,0.1788907945,-0.1657695472,0.3593076766,-0.1619860679,0.3304784894,-0.290073663,-0.4505563378,-0.080542706,0.143808797,-0.0844437107,0.1313305348,0.0786818638,0.433490485,-0.2061226368,-0.0349456258,-0.4442700148,0.2482859939,0.0669215247,0.0064120521,0.0686031654,-0.2048148364,-0.1650122106,0.0923437849,-0.0643353909,-0.016690718,0.3275226355,-0.1424655169,-0.3697812855,-0.4538449049,-0.2162763029,-0.0822580755,0.2116422951,-0.4417620599,0.1198693514,0.1042456627,-0.2983011603,0.0144472541,0.6690344214,0.3910433054,0.2221409976,-0.0082954373,-0.0840206519,0.1398146898,0.1586985588,0.0923110023,0.301276505,0.0906743333,-0.1768054068,0.3088213205,0.0299497005,-0.0884334967,0.0700699911,-0.3639446199,-0.2119321674,-0.4602619708,0.2309454978,-0.1386169493,0.0885575712,-0.0725027323,0.0463712141,-0.3338874578,-0.0766064972,0.1268920898,-0.0344366953,0.2461351901,0.0962803215,0.0243322775,-0.3029870689,0.2154331952,0.4572522938,-0.2380200773,-0.1272331923,-0.2411009371,-0.5297257304,0.1988669485,-0.1453111619,-0.2223797739,-0.0128744366,0.2722304761,-0.2468560934,0.0310955066,0.4103512168,0.2545855939,-0.1320250481,0.1670610309,-0.3741107285,-0.1049029902,-0.1244053394,0.0631966516,0.1704512686,-0.2594733536,0.2376660407,-0.0988555923,-0.0403691195,-0.1549960822,-0.035338778,0.0598153509,-0.0658439919,0.4095597267,0.3740644157,0.2854384184,0.0370702557,-0.1465518475,-0.2708669305,-0.1762216985,-0.1887483746,0.4648070335,0.0033473216,0.2654676139,-0.1592422873,0.0305482578,-0.1501217484,0.5124750733,0.1208250895,-0.0202244129,-0.2466941923,-0.2782547474,-0.0599183328,-0.0759468973,0.4266418517,0.2766796947,-0.1649906486,0.1061381027,-0.0258041844,-0.4128347039,0.4644300342,0.00816123,-0.0649360195,-0.1816580743,0.2398122549,0.317119807,-0.2281712592,-0.4576901197,0.0420844778,0.6015406847,-0.0326097496,-0.1080123931,0.1667660028,-0.0720524564,-0.0610164069,0.0857727081,0.1562547535,-0.0138093578,-0.3180502653,0.4304473996,-0.0633088499]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2839","title":"OpenWebText: NonMatchingSplitsSizesError","comments":"I'll try without deleting the whole cache (we have large datasets already stored). I was under the impression that `download_mode=\"force_redownload\"` would bypass cache.\r\nSorry plateform should be linux (Redhat version 8.1)","body":"## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n","comment_length":31,"text":"OpenWebText: NonMatchingSplitsSizesError\n## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n\nI'll try without deleting the whole cache (we have large datasets already stored). I was under the impression that `download_mode=\"force_redownload\"` would bypass cache.\r\nSorry plateform should be linux (Redhat version 8.1)","embeddings":[-0.3607826829,-0.239854753,0.0516973063,0.4607081115,-0.0456267372,0.1327427775,0.066052109,0.6086549163,0.1439708173,0.1374783516,-0.2283655256,0.1216922924,-0.0511874668,0.24087888,0.0173501056,0.0554530993,-0.1288683265,0.209086746,0.0327513143,0.0964000374,-0.2407321781,0.165215224,-0.3307679594,-0.0877942294,-0.0919432417,0.0468051881,-0.0785730928,0.2525186837,-0.1423963159,-0.3720365465,0.3101969957,-0.3209818304,0.0526632667,0.4554589093,-0.0001236256,0.058411181,0.3002471924,-0.2492422163,-0.16436176,-0.0202188827,-0.4952323437,-0.4582964778,-0.1188494414,-0.1141500697,0.184963122,-0.0521582067,0.0923798084,-0.2381899059,0.3523260951,0.1760241985,0.1076085791,0.2688066065,0.1909793615,-0.0141789233,0.3994997442,-0.288500309,-0.2018007636,-0.026038751,0.2025960684,0.2345232815,-0.0004501377,0.2051186413,0.0155433789,0.3176065981,0.1230182052,-0.0852457508,0.1734517515,-0.223595798,0.1765745282,0.4632994831,0.7116635442,-0.2908660173,-0.3333330154,-0.4549492896,-0.4818343818,-0.0767983198,0.3402119577,0.3721658885,0.0110199759,0.2835704088,-0.2383954376,-0.0144351302,0.0867683887,0.0248056687,-0.2652613223,0.1660380363,-0.0655873343,0.0525521077,-0.0053646825,0.1384350061,0.2647070587,-0.0735872909,0.0548304617,0.1759223789,-0.2408612818,-0.0740365535,0.0165432449,-0.2352584451,0.3727194965,0.3354320526,0.3635959923,-0.1954646856,-0.3061286211,-0.0127849197,0.202025786,0.4631952047,-0.1391366273,0.0402779281,-0.1740580797,0.0575832874,0.111650236,-0.1003556475,-0.0169872101,-0.3424649537,0.1643748879,0.0634849742,0.6322448254,-0.3608763218,-0.5638146996,0.0620397069,-0.2243329287,-0.2449309826,-0.2871406674,-0.0788733289,-0.3087580204,0.3446002007,0.1083343402,-0.1410493702,-0.250838846,-0.4097034633,-0.2513109446,-0.1282017827,-0.1443449259,0.0415491126,0.0319383219,-0.138546139,0.2920025587,0.0044018342,-0.0483897887,-0.1885268092,0.1772954911,-0.0495248884,0.0395917334,0.2417313159,-0.0637856051,0.2164584994,0.0311726723,0.1261219084,-0.3001105189,0.3434175551,-0.3207673132,-0.3199581802,-0.0984371156,0.0781468749,-0.2246572673,-0.1677706391,0.1053349003,-0.047128059,0.3671295941,-0.2251332849,-0.0991620123,-0.1212022156,-0.1245307252,-0.1096512154,0.0449683294,0.5349205136,-0.4528928995,-0.1801821291,0.185575366,-0.1425025314,0.3732614815,0.3822270036,0.0877135545,0.0240434296,-0.2359959036,0.0825527161,0.0203862526,-0.4213356078,-0.8024846911,0.4292296767,-0.063970007,0.1961514801,0.1568332464,-0.1446538568,0.1170874834,-0.1508550644,0.057605356,0.1717762351,-0.2016098052,0.0442328788,-0.629065454,-0.8196583986,0.3851351142,-0.0254817307,0.2239643931,-0.1056236476,0.0249994807,0.2732915282,0.368870616,0.1013237163,0.0660170689,0.1658230275,0.0039065285,0.0954902247,-0.0445364453,-0.2933856547,-0.3191492856,0.2778995037,-0.2501968443,-0.1287551522,0.3011008203,-0.1411423236,-0.5185483098,-0.2888207138,-0.0862790644,-0.0926862955,-0.0315534584,0.0389909446,0.2252077758,0.0351627357,0.0298353322,0.3739862442,0.116735056,0.1384649873,-0.2917839289,0.1245396957,-0.0396486931,-0.0166225843,0.0037614766,-0.1315949559,0.189512223,-0.1118277237,-0.3834511042,0.6283049583,-0.0011398828,0.1150420979,-0.090775691,-0.0132446112,0.2655419707,-0.1831507087,0.1516956985,0.3305782974,-0.0388676003,-0.1138824895,0.0377740338,-0.1808829308,0.1214482114,0.0156273618,-0.1123446822,0.0120707881,-0.0592523217,-0.1683395058,0.0520322248,-0.1470990628,0.041033186,-0.132320255,0.1531793624,-0.058693409,0.089498423,-0.0775124133,0.6427761316,-0.0461898148,-0.0024128065,0.2773858607,-0.31353724,-0.1165984571,-0.149243772,0.3093588948,0.4315313101,0.2351769954,0.1816487759,0.214854151,-0.0480404198,-0.0663257837,0.0623969994,0.2788154185,0.1103492528,0.278703928,0.0314646326,0.0604010746,-0.2338596284,0.1105607077,0.1959756166,0.3361184299,-0.4241134524,-0.1764782369,-0.4222451448,-0.0984081179,-0.2253969163,0.1779851317,-0.1684412062,-0.228486225,-0.0869306996,0.0673669577,-0.0112149464,0.18163158,-0.2745671272,0.1598093063,0.139281556,0.0402397811,0.0125676868,0.0235119388,-0.4680196047,-0.0232420117,0.3328770697,-0.314013958,0.3491358757,-0.535438776,-0.0107724341,-0.0256563537,-0.1533284187,0.0009073421,0.082536146,-0.0696029216,0.2342223376,0.2336276174,0.0563837104,0.1101676375,0.1208643317,-0.270549655,-0.2377162576,-0.1075411141,0.1214817017,0.3437823653,-0.0233938284,-0.5561491847,-0.2882370651,-0.3905978799,0.2452228218,0.1488028616,0.2224647105,0.3056887388,0.081562236,0.1264165938,0.1255260706,0.086612016,-0.2552498579,-0.3846787512,0.27545017,0.0955306813,-0.2659647465,-0.0228641741,0.2200033963,-0.0600050166,0.1335960776,-0.5093490481,-0.1830071062,-0.4022891521,-0.2084587067,-0.1236945465,0.2260227352,0.4829974174,0.0146279922,-0.0300645493,-0.1376815438,-0.330088079,0.0368260704,0.1950068027,0.4865730703,-0.1092981622,0.238016054,0.3572044671,0.4601337314,0.468192935,0.1195461452,0.0832175612,0.0417757891,0.2068868279,-0.1063739881,0.0597097725,0.1245609075,-0.2355274707,-0.0282634329,0.2690125108,0.1604710221,-0.0877640173,-0.0937861949,-0.1780083925,-0.2976430357,-0.1028991342,-0.056094151,-0.0675401241,0.1703018546,0.2686711848,0.1051749662,-0.3179878891,-0.2011032701,0.1288355589,0.0375514328,-0.0185770057,0.0662474185,0.0950690508,0.4359333217,-0.3008748293,0.408465147,0.1136155277,0.2396367192,0.1571036577,-0.2272114605,0.1036046445,0.0330567695,0.4885368645,-0.2571599185,0.0951224789,0.0809625834,0.150860101,-0.3015558124,-0.2861626744,-0.1340235919,0.0312607549,0.5175762773,0.0374496356,-0.4427628517,0.0397897884,0.2095132023,0.4867495,-0.2145173252,-0.0063501117,-0.1836827546,-0.3508098423,-0.4405160546,0.0100522069,0.1441052556,0.1535165757,0.0876208469,-0.1092209294,-0.1667594165,-0.0914118215,0.1370179802,-0.1000970826,0.3531214297,0.0272447821,0.2810154259,0.2407773435,0.3583734035,0.2556761801,0.5383527279,-0.021165546,-0.0488367788,0.0128531642,-0.0926643014,-0.2564074397,0.4061823487,0.0760536641,-0.3070344329,0.1012065113,0.1428366005,-0.2415323555,0.2489228249,0.2346798331,-0.1833533943,-0.5043689013,-0.6911815405,0.4553481638,0.1715047657,-0.0941056758,0.4634655714,-0.0732957423,-0.1497593373,0.0628624707,-0.1080331802,1.1349259615,-0.0019084004,-0.0780185983,0.0873545408,-0.5329984426,0.1726213545,-0.2816605866,0.1941205561,-0.2007126808,-0.4522471428,-0.0641660839,-0.2595258057,0.0983032063,0.0390519239,-0.1713236123,0.2763011754,0.0749046281,0.1446793526,0.1882211268,0.2370787114,-0.3019885421,0.0262208786,0.1655744165,0.0374213457,-0.2098885775,0.512563765,-0.1052570567,-0.2427012324,-0.1034007668,-0.1232609674,-0.5089899898,0.1167298257,-0.3765400648,0.2764251828,-0.2374932915,-0.0862786472,0.0357935578,-0.0525439978,0.0248796251,-0.0449727252,-0.3938284516,0.1956785172,0.0594527237,0.2147149146,0.2943235338,0.0335743017,0.2583121061,0.0008409771,-0.1066407859,0.140414089,-0.0813289285,-0.2316496968,-0.0270814281,-0.0436839983,0.3094876111,-0.5829216242,-0.216356352,-0.1027067527,0.1244824603,-0.2333381176,0.034378849,0.0879767686,-0.0746369436,0.3408400714,-0.0635321289,-0.471717298,-0.1994462013,0.5206711292,0.1819891036,0.0293352511,0.3584669232,-0.1413273364,-0.0639839992,0.0470415317,0.3087407053,-0.2237901986,-0.7066394687,0.3328604102,-0.500685215,0.2621082962,0.0529283509,0.1818521917,0.4850798845,0.4176092148,-0.0780598596,-0.1445489377,-0.0488503315,-0.0138431787,0.0381969698,0.3852779269,0.0485504679,0.4218764901,0.1068172455,-0.1626463085,-0.2120172083,0.2745403945,0.0531282164,-0.0390793979,0.0639520735,0.0605904683,0.1390803456,-0.1253376156,-0.1304654181,0.2742481828,-0.2284352481,-0.0878518745,-0.1836317629,0.2317643613,0.0993773937,-0.1137636006,-0.0607136041,-0.1676470786,0.107526198,-0.1030435935,-0.0204722751,0.0792441368,0.004505835,0.1212018579,0.0752922595,-0.1436760724,-0.1643702239,0.2273164988,-0.127459228,0.2589283586,0.0485452674,0.3721717894,0.1187281236,0.0677689314,-0.2113079429,-0.051788453,0.1161625832,0.1264194101,0.2635839283,-0.3530114591,0.122291714,0.4328971803,0.1321386844,0.333717227,-0.1584495008,-0.2856967151,0.0900920331,0.1263714731,-0.311929673,0.0799823329,0.5263721347,0.0300190262,-0.0502244681,0.16433236,0.2634984851,-0.0278823003,-0.2550130785,0.2430293709,0.5129175782,0.160427317,0.2251694947,1.0427386761,0.0106978891,0.1018011421,0.1038975194,0.0040842388,0.3406729698,0.412668705,0.0849268958,0.2741696537,-0.0247918442,0.3909768462,-0.413439393,-0.4724245965,-0.1134139895,0.3169428706,-0.013725725,0.278760612,0.1678983569,0.218736425,-0.2551123798,0.0223544799,-0.3514285684,0.1848168969,0.0168443173,-0.032872919,0.324711591,-0.2265097946,-0.1403006762,0.0862738118,-0.160579592,0.1383538097,0.2371824533,-0.185989216,-0.1939421594,-0.3010249138,-0.1874037087,0.010577946,0.1624984443,-0.4342554808,0.1783868223,0.1304658353,-0.2807535231,0.0324490704,0.6316363811,0.4081062376,0.3006579578,-0.0689980611,-0.1026685834,0.3005135357,0.1360225081,-0.0912715718,0.1052781269,0.0703729913,-0.067407608,0.3016679883,0.0214814637,-0.1085116491,0.0153886583,-0.2776354551,-0.1697630137,-0.3523252308,0.3387854695,-0.2293528318,0.069656916,-0.032013718,0.040482156,-0.374260664,0.0021905468,0.1292652786,-0.055863563,0.2296980768,0.1696502119,0.0025295273,-0.148401022,0.2151013166,0.4300907552,-0.3279687464,-0.1806129813,-0.055387985,-0.6753011346,0.1929343939,-0.2446181476,-0.1638778597,0.0704149455,0.3578231037,-0.1726098359,-0.0341279916,0.437810719,0.3769516051,-0.111942701,0.2010066062,-0.1759392619,-0.2322308719,0.0149960201,0.0364370532,0.1134869978,-0.427033782,0.21207726,0.1316030324,-0.0673153698,-0.2084578574,-0.0971877351,0.1151321456,-0.1476906687,0.5114874244,0.2450016737,0.3539223671,0.0070622731,-0.291864723,-0.2748839855,-0.1809360236,-0.2635334432,0.3029268682,0.0074895052,0.1913788617,-0.2344797552,-0.0533136092,-0.1607613415,0.560459137,0.1186750978,0.0577345639,-0.2615610957,-0.2985357344,-0.2029914409,-0.0543149337,0.428090781,0.2124760151,-0.195509851,0.119935818,-0.0338993222,-0.5013495684,0.3801289201,0.04241715,-0.0864379257,-0.0998112932,0.2353791893,0.4515444338,-0.1079054475,-0.3443450928,0.1179368272,0.4826676548,0.0082232049,-0.1074975803,0.1426823884,-0.1259907484,-0.1082381979,0.0097631421,0.0963457599,0.0574400462,-0.416595459,0.3547329605,-0.1080687344]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2839","title":"OpenWebText: NonMatchingSplitsSizesError","comments":"Hi @thomasw21 , are you still having this issue after clearing your cache ?","body":"## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n","comment_length":14,"text":"OpenWebText: NonMatchingSplitsSizesError\n## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n\nHi @thomasw21 , are you still having this issue after clearing your cache ?","embeddings":[-0.3699499965,-0.1261989325,0.0325992927,0.4002228081,0.0890990272,0.1182889119,-0.0533089116,0.6034545302,-0.0335419439,0.1091241091,-0.1687633097,0.0282330066,-0.0962741897,0.0463814177,-0.1560756266,0.0372467712,-0.1961586177,0.2053611428,-0.2641817033,0.1242605746,-0.2632367015,0.2561306357,-0.2909330726,-0.0491796099,-0.1904989481,0.0199674014,-0.1706748307,0.1613520384,-0.221933037,-0.4056100547,0.2906131744,-0.3588837385,0.0469210073,0.5125917792,-0.000114122,0.0556628332,0.3204028606,-0.1646400541,-0.1723579913,-0.1874836534,-0.4532294571,-0.3731660545,-0.0388759747,-0.112672627,0.1566630751,-0.1037599817,0.1046383306,-0.3784665465,0.3734263182,0.330273509,0.2179907411,0.3707486689,0.2069821954,0.0058436338,0.4057067931,-0.3193127215,-0.2311393917,-0.1044008136,0.1066185907,0.0624603406,-0.0348753482,0.0600125566,0.0215322468,0.3120785952,0.077384375,-0.0503110252,0.3094576597,-0.2281142473,0.1407149136,0.3834085763,0.7641160488,-0.2195935696,-0.283985287,-0.3051543236,-0.3677956462,-0.0343249999,0.2471122444,0.483710438,-0.020346934,0.2292669863,-0.3121861517,0.0388080738,0.0425118431,0.1139018014,-0.2064270824,0.2632624805,-0.0242877677,0.1047253832,-0.0661162212,0.0650689006,0.3052019179,-0.1170110404,-0.0257206298,0.1538783908,-0.2919646204,-0.0112022916,0.0328578874,-0.2262672037,0.2527976036,0.3594455421,0.3436971903,-0.2234082371,-0.1035509557,-0.0320001915,0.2006347477,0.4552832246,-0.0448409393,0.1968113631,-0.0917967036,-0.0303141531,0.0936576501,-0.0837954581,-0.0554752834,-0.3193314373,0.0959796682,0.0045037735,0.5925169587,-0.2778443098,-0.6033614874,0.0504921898,-0.2114330232,-0.2764321566,-0.3233760893,-0.1093845218,-0.3041671515,0.2534267902,0.1797761768,-0.1391205937,-0.3311912417,-0.3503164649,-0.3089872301,-0.1209257022,-0.1772671044,0.032847248,-0.0027473106,-0.0529400334,0.3967550993,-0.1295852065,-0.0087198969,-0.2121687531,0.2079534382,-0.0091037238,-0.0446107127,0.1839971989,-0.1260671616,0.1843945831,0.0385377072,0.1092341244,-0.3352973461,0.3740764558,-0.2801018655,-0.4088790119,-0.1939189583,0.189780131,-0.3035056293,-0.1956293732,0.3018482625,-0.0626553893,0.2803269625,-0.1122081727,-0.1693761051,-0.1711202711,-0.1893253773,-0.1533607095,0.0882912874,0.3502240479,-0.257031709,-0.1984224617,0.2106315792,-0.0680614114,0.3246780038,0.4196591675,-0.0076634637,0.0039387178,-0.2138770074,0.1999334693,0.1432218552,-0.2937815785,-0.6871965528,0.3707432449,-0.114377968,0.1523077488,0.0933001712,-0.1211867109,0.1059551015,-0.1198136285,-0.0463212542,0.2644116879,-0.2024022788,0.0870940834,-0.5790618062,-0.7767994404,0.3755401075,-0.0242120661,0.2111732364,-0.1639950573,0.006461286,0.2177522928,0.4441138506,0.0989785343,0.0500627458,0.1430724412,0.0939549729,0.0429766849,-0.061436139,-0.279608041,-0.221799925,0.2446716726,-0.1376405358,-0.0303829275,0.2486105412,-0.1314045936,-0.5314114094,-0.2417081743,-0.1061991081,-0.1340866238,0.0522408374,-0.0163130574,0.2626656592,0.0801130608,0.080397673,0.3879548609,0.0932108238,0.1257617027,-0.2942942381,0.2135344148,-0.0467306525,-0.1446321309,0.178412959,-0.0720042661,0.1614710093,-0.0529614687,-0.4088436663,0.656724751,-0.0448579974,0.1105701253,-0.0014170395,0.0634532049,0.2236272097,-0.2137767524,0.1476328522,0.3671287,-0.0286879614,-0.0925415456,0.1367339641,-0.1318332106,0.1761523336,0.0378681459,-0.1045079455,-0.0272626355,-0.0866388455,-0.1919618994,0.0127834119,-0.1355392635,0.121220693,-0.0005551229,-0.1029190198,0.0522346124,0.1520875394,-0.1716966927,0.7244181037,-0.0214816853,-0.0341246277,0.2815185785,-0.334707588,-0.0916437805,-0.1870049685,0.1764208823,0.4675450623,0.2792059779,0.1903891414,0.1980343014,0.0257366169,-0.1207682714,0.1171555221,0.2820116878,0.2443300039,0.2476788908,0.0440932438,0.0549047031,-0.2191032767,0.127850458,0.0708637387,0.3302167356,-0.3441953063,-0.2616869509,-0.4838453829,-0.0104162619,-0.2235880345,0.1671076715,-0.156897679,-0.3010331094,-0.0594916195,0.0940519497,-0.0327388905,0.2392505854,-0.2775541842,0.1586039364,0.1916082054,0.1540816426,0.0908739641,-0.0944102034,-0.5167679787,0.0587307364,0.2828665674,-0.3290631771,0.4073050916,-0.5408995748,0.0238137655,-0.0917047858,-0.1395531148,-0.1037957743,0.0741046965,0.0165922679,0.1612576097,0.1679470241,0.0378816687,-0.0143355317,0.1781898737,-0.3044912815,-0.2063006163,-0.0532360338,0.1252015829,0.2432821542,-0.0463878065,-0.6165706515,-0.291761905,-0.3603462577,0.2076623291,0.0761469081,0.1756553352,0.0283724535,0.0443828218,-0.0005369574,0.0527454726,0.0789406672,-0.2449579984,-0.4761647284,0.3822352588,-0.0113914954,-0.2909473777,0.0234442204,0.2967124879,-0.0753872171,0.0756534189,-0.6022984385,-0.1599006504,-0.4078228176,-0.0391741544,-0.108949624,0.2167443335,0.4038220942,0.0629792586,-0.0997803211,-0.1050412282,-0.3229443431,-0.0342823453,0.2531094849,0.4984205365,-0.075255096,0.2646891475,0.3646192551,0.4557607174,0.4932022393,0.0393025428,0.0251377858,-0.0237086453,0.1077448577,-0.1484603286,0.0608350486,0.1915815771,-0.2333793342,0.0665346235,0.3000492752,0.2092967182,-0.0172890276,-0.0782203227,-0.0475675128,-0.3858370483,-0.1967263967,-0.0444011129,-0.0784836709,0.1743656099,0.234790504,0.0342591368,-0.2259376496,-0.2762153447,0.1112004444,-0.0675346926,-0.1429969519,0.0296185333,0.1646965295,0.4048630297,-0.4343224466,0.3805494606,0.0658688918,0.190784514,0.1779663116,-0.2261306047,0.1277927756,0.0619105622,0.4063279927,-0.0373143628,-0.0581962466,0.0986392945,0.1192357019,-0.4282928705,-0.2882971466,-0.0731879696,0.1474189758,0.4618077874,-0.0963401347,-0.3969036639,0.0115620997,0.284748435,0.3944990337,-0.2492479831,-0.0591817088,-0.1590930372,-0.3731490374,-0.3331924677,-0.163165465,0.1015041023,0.2927089334,-0.025393337,-0.110274598,-0.1241560876,-0.2215584219,0.1548478305,-0.1730535477,0.3194341362,-0.0157437511,0.3513678312,0.2734590471,0.3738481104,0.216216743,0.6072022319,-0.0133419409,-0.1398139596,-0.0264540762,-0.0848596394,-0.2349039167,0.2699095309,0.0967658311,-0.3419226706,-0.0765482485,0.242423147,-0.2905293703,0.2411602437,0.3828800321,-0.1577095836,-0.4289746881,-0.4100594521,0.4611559212,0.1847285926,-0.0395510942,0.4683814943,-0.0681901574,-0.1618482023,0.0223922133,-0.1416904777,1.0075243711,0.0473518558,-0.070832409,0.1603209227,-0.4676517546,0.2012103051,-0.2283052355,0.2453926504,-0.2333680987,-0.3698821962,0.0185682103,-0.2234342098,0.1259595305,-0.0248139221,-0.1877573729,0.2538684607,0.0037012172,0.1132142916,0.1456156224,0.2981158793,-0.2324315012,0.1620375812,0.2557009757,0.1349364668,-0.2352220565,0.4734431505,-0.0882345811,-0.1289013922,-0.059737619,-0.1225551814,-0.5021666884,0.07698109,-0.3497204185,0.3533706069,-0.2537010014,-0.0112943575,0.0301967245,0.0693132058,0.0641643181,-0.0456334837,-0.3567199111,0.288267076,-0.0491707437,0.1585377902,0.2078833282,0.1017516553,0.0963010117,-0.0898899734,-0.2434322983,0.0874571428,-0.1024627984,-0.1643781662,-0.0210669991,-0.1699323058,0.3804636598,-0.5987073183,-0.1298674494,-0.0560918897,0.1143864244,-0.1916318238,0.125151962,0.1493874937,-0.0565422587,0.4149746299,0.0331454426,-0.4573384225,-0.2126535028,0.4191862047,0.1492619514,0.0418783054,0.4388936162,-0.1515882313,-0.0777815506,-0.0204084069,0.3754278719,-0.1178459674,-0.525010705,0.2747764587,-0.3725500405,0.3430575132,0.1198816821,0.2394055426,0.4780520201,0.4091352224,-0.0040369891,-0.1364207119,0.053976614,0.0129884854,0.0321094207,0.405670315,-0.0683615729,0.3495720029,0.0551644228,-0.0754781663,-0.2944043577,0.2543922961,0.0035267975,0.0515492447,0.0972042978,-0.0195180774,0.1351340264,-0.0175564829,-0.0069645499,0.2041044682,-0.2563955188,-0.1647906005,-0.133107841,0.1834753454,0.0535559617,-0.1581362039,-0.0237283576,-0.1615855843,0.1350013465,-0.1121332198,-0.0332108177,0.1630435586,-0.0651337057,0.218713671,0.1464077979,-0.2388555706,-0.047596395,0.242839843,-0.1983966976,0.3675199449,0.0656004921,0.3472638428,0.1828948259,0.029625861,-0.0774962828,-0.0980018973,0.0194602907,0.0936442539,0.2962960005,-0.2937561572,0.0647013038,0.3576254249,0.15895468,0.3955867589,-0.1450508535,-0.2880072594,0.1373763233,0.2351198792,-0.2910705507,-0.00102274,0.4893831313,0.0906786248,-0.0360482633,0.061286971,0.1359153241,0.00004355,-0.2684246004,0.1829939187,0.4793517888,0.1680790186,0.3154784441,0.7748761773,-0.0139485104,-0.0920302644,0.050207749,-0.1478105783,0.3275920153,0.2920737863,-0.1787822098,0.3930224478,-0.0138913244,0.3724783957,-0.4011479914,-0.4172951877,-0.0690851808,0.2302530259,-0.0381208509,0.1952132881,0.1545183361,0.2349814624,-0.205132544,-0.0329739116,-0.3573937416,0.1953632534,0.0132825412,-0.0334880836,0.1450294852,-0.2101859599,-0.1938864887,0.0848090574,-0.1631283015,0.0196652226,0.2517290413,-0.1770727187,-0.2166511267,-0.3352458477,-0.150279209,-0.0993154943,0.206591785,-0.4288110137,0.0856537372,0.1212227046,-0.2970786095,0.0337097608,0.6495663524,0.39911291,0.277544409,-0.1779762655,-0.0753611326,0.2537455559,0.0985254347,-0.0123101119,0.1917613894,0.0324676782,-0.168019563,0.2745341957,0.1051601693,-0.1410874724,0.0826658383,-0.2783308625,-0.1329126209,-0.3866247535,0.1847351789,-0.1417805105,0.0418033265,0.0239878725,0.0048595183,-0.3268780708,-0.0127508193,0.2200702578,-0.083038725,0.2020141035,0.0824789926,0.0501657724,-0.1901606172,0.2122752368,0.4538312852,-0.2883225381,-0.1841077805,-0.2728342414,-0.5499471426,0.2420024425,-0.2210352123,-0.1362965405,0.0553482734,0.3255945742,-0.2462527305,0.0084377937,0.4023242593,0.3190976083,-0.1830889136,0.1065202579,-0.3580295444,-0.1319987029,-0.0476974994,0.0569486618,0.1757245213,-0.3006260097,0.1647506654,-0.0643829256,0.014002122,-0.2621038258,-0.031362839,0.1674591452,-0.0671422929,0.4410232604,0.3077579141,0.247195214,0.0354929566,-0.1491001844,-0.2292175144,-0.3135262132,-0.1659494936,0.3160116076,-0.0073687932,0.201423347,-0.2448646724,0.0274212118,-0.1734057218,0.4470086694,0.0180720352,0.0514634289,-0.1807844788,-0.2943457067,-0.0613124147,-0.1014600024,0.3610028028,0.3020106852,-0.1695098728,0.1225596592,-0.028407665,-0.4301031232,0.424536556,0.0289269555,-0.068247743,-0.1417369843,0.1950233132,0.3503942192,-0.0881268233,-0.3723471463,0.1681749076,0.5794219971,0.0252883397,-0.0621412545,0.1718127131,-0.057613682,-0.111045301,0.0953206122,0.1655536294,-0.0319843888,-0.4057445824,0.3418639004,-0.1213817671]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2839","title":"OpenWebText: NonMatchingSplitsSizesError","comments":"Sorry I haven't had time to work on this. I'll close and re-open if I can't figure out why I'm having this issue. Thanks for taking a look !","body":"## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n","comment_length":29,"text":"OpenWebText: NonMatchingSplitsSizesError\n## Describe the bug\r\n\r\nWhen downloading `openwebtext`, I'm getting:\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430, dataset_name='openwebtext')}]\r\n```\r\n\r\nI suspect that the file we download from has changed since the size doesn't look like to match with documentation\r\n\r\n`Downloading: 0%| | 0.00\/12.9G [00:00\r\n- `datasets` version: 1.10.2\r\n- Platform: linux (Redhat version 8.1)\r\n- Python version: 3.8\r\n- PyArrow version: 4.0.1\r\n\nSorry I haven't had time to work on this. I'll close and re-open if I can't figure out why I'm having this issue. Thanks for taking a look !","embeddings":[-0.3353740871,-0.18841286,0.0487625636,0.4328111112,0.0591351986,0.1115414873,-0.0852611586,0.5720363855,-0.0169446841,0.1263842881,-0.1240672097,0.0030228929,-0.0780212209,0.0841974169,-0.1284149289,0.028532682,-0.1261273474,0.1875724196,-0.2567581236,0.1037800536,-0.289475292,0.2635649145,-0.3147125542,0.0043351529,-0.1761120707,0.0403093137,-0.1997045577,0.1530286819,-0.2597185671,-0.3653438091,0.2717806399,-0.3823173642,0.0590334535,0.4936454296,-0.0001182356,0.0295847729,0.3093135059,-0.2089018524,-0.0846336931,-0.1598861217,-0.4955505431,-0.4291180968,-0.0661034137,-0.1210819632,0.1902426779,-0.1171414033,0.126828596,-0.3275127709,0.3349781334,0.3231547475,0.178308785,0.4126829505,0.2072996944,0.0013006629,0.3759169579,-0.2919977009,-0.2213473916,-0.0704078153,0.0860572755,0.0268128999,0.0031038474,0.0268938132,0.0232798997,0.2867479622,0.1630622447,-0.0368349515,0.3667397797,-0.2062833607,0.1011864692,0.421010673,0.7274805903,-0.215399608,-0.2575476766,-0.3073272705,-0.3769135177,0.0297992546,0.2773856521,0.4866474867,0.0031598224,0.1984095871,-0.3036155999,0.057540372,0.038236998,0.1616750956,-0.2182648033,0.2967064977,-0.0009074002,0.1527033448,-0.0929155275,0.0851415694,0.2732588649,-0.1273201704,0.0084036961,0.1611264944,-0.2615137696,0.0117760235,-0.028219644,-0.2475543171,0.2653850913,0.3092905581,0.3690367043,-0.3131037354,-0.1379642636,-0.0037725242,0.239037782,0.379905194,-0.0437345132,0.1846986711,-0.1888067573,0.0326588899,0.1132931337,-0.0755623877,-0.0282561257,-0.3621478677,0.0659071729,-0.0113287931,0.6134747267,-0.2487007678,-0.6051997542,0.0658672377,-0.2793935835,-0.2565443218,-0.3295621574,-0.1635749787,-0.3112907708,0.2582394481,0.1950361133,-0.0819413513,-0.3276638687,-0.3721604943,-0.2869725823,-0.1174980029,-0.1620406508,0.0461693704,-0.104130201,-0.0131914904,0.3832283616,-0.0587805249,-0.0423808992,-0.2699390352,0.1448607892,0.0324001946,-0.0699524432,0.150408566,-0.163368687,0.1828367412,0.0472206064,0.084708415,-0.356433779,0.3897977173,-0.2615618706,-0.3636491001,-0.2217823118,0.1456974,-0.3320489228,-0.2151723057,0.3394041359,-0.0417384505,0.2415718585,-0.2057961076,-0.1408115029,-0.2203736901,-0.2054502219,-0.1400124133,0.0715175718,0.3740119636,-0.2969316244,-0.218428269,0.2159541398,-0.0582996793,0.4005611539,0.4265969098,0.0306359641,0.0068852236,-0.1978981346,0.251529634,0.1011661589,-0.2532138526,-0.6780224442,0.4099708498,-0.1634244025,0.1814756691,0.1188286543,-0.0990723297,0.0737057328,-0.131755352,-0.0850457847,0.2120765597,-0.2106408179,0.0707255006,-0.5871102214,-0.7869003415,0.3780463934,-0.0584511533,0.1627453417,-0.1797488481,-0.0278539229,0.2599582672,0.4960872531,0.1067426801,0.0147278421,0.1068839282,0.0614952855,0.0681387484,-0.0840681568,-0.2501348555,-0.1588561684,0.2199242413,-0.1815293133,-0.0562988445,0.3016198874,-0.1083739847,-0.5671508908,-0.205630973,-0.0594865642,-0.0970369652,0.0005349878,0.0013908687,0.1822182983,0.0878181085,0.1549400836,0.3509557545,0.0925608948,0.1153081357,-0.3258306086,0.2187996805,-0.042067606,-0.1144090295,0.1862325072,-0.0807211772,0.1894461066,-0.0506584011,-0.4008100629,0.6621257663,-0.0508649424,0.1304217279,-0.0485616773,0.0759109631,0.208716765,-0.1647550017,0.1396922916,0.4040056467,-0.0078118043,-0.0893657655,0.1285756379,-0.1466649324,0.1681570411,0.0385813452,-0.1399838179,0.0219172258,-0.1082341596,-0.1906882823,-0.0218825191,-0.1363818944,0.1214792728,-0.0143770063,-0.1486689448,0.0490543544,0.1307161152,-0.1663719416,0.735114038,-0.0395391472,-0.0493242703,0.3084146678,-0.3009499013,-0.0696082339,-0.1869142652,0.2009025812,0.5170977712,0.2373419106,0.1705388427,0.2232394367,0.0059838467,-0.1573754251,0.1209169701,0.298640728,0.2252731025,0.2332677841,-0.0040379744,0.0442569368,-0.1431460232,0.1358934343,0.054894004,0.3265087605,-0.4214719534,-0.2221026719,-0.4799452126,-0.0070513934,-0.332985729,0.185629949,-0.1814439297,-0.3045488298,-0.0691658854,0.0729711354,-0.0868554711,0.1772733927,-0.2617514133,0.1166194379,0.1726839989,0.1684265584,0.1131397113,-0.0296697319,-0.5200146437,0.0163424835,0.2846112549,-0.2646366358,0.3774171472,-0.553154707,0.0195056908,-0.0720698833,-0.1443551183,-0.1200665608,0.0596865043,0.0407049321,0.1904513389,0.1615996957,0.0788518563,-0.0324806385,0.1823148876,-0.2791775167,-0.1931467801,-0.0665636063,0.1417395473,0.2153919935,-0.0159327202,-0.5946922898,-0.2640362382,-0.3566413522,0.2469536364,0.0686296374,0.1870593131,0.0002924684,0.082385689,0.0191314239,0.042840708,0.0874992684,-0.2150725871,-0.4413594306,0.3366400301,0.0674884766,-0.3040826321,-0.0139753176,0.2830919027,-0.0750178322,0.0852657259,-0.6037398577,-0.1121485978,-0.3842428923,-0.1020442545,-0.1089007631,0.1909778714,0.4259518683,0.0320949741,-0.0522993766,-0.1426744163,-0.3267737031,-0.0091122296,0.2860001922,0.5351224542,-0.0486740395,0.2700761557,0.3071492612,0.437469542,0.536436379,0.0791866481,0.0161778517,-0.038535431,0.0828423277,-0.1028325781,0.053590294,0.191081807,-0.2176369429,0.0492427573,0.356510669,0.1696504056,0.0127552291,-0.0368495211,-0.0876021162,-0.3898725808,-0.1862393767,-0.0227672923,-0.0313762613,0.1551938504,0.2314361185,0.0815003365,-0.2151930481,-0.2685636282,0.1059815288,-0.0300254636,-0.1459790915,0.042329181,0.1752462387,0.3730886281,-0.3741367757,0.396332413,0.0969317332,0.237070322,0.1760120243,-0.2623289227,0.1341694742,0.0665083379,0.3930379748,-0.0159508865,-0.0892399326,0.0761585608,0.1335611939,-0.4909276068,-0.3171027899,-0.1052994579,0.1085404456,0.5230340362,-0.0738658905,-0.3801212013,0.0209024921,0.3068942428,0.4333490133,-0.2359808534,-0.0729079694,-0.125231415,-0.3763367236,-0.285029918,-0.1912844479,0.0464635305,0.2928822637,0.00435649,-0.0980404466,-0.0733076483,-0.2081430852,0.174622193,-0.1836490929,0.2925070226,-0.000494563,0.3188732862,0.2693992853,0.386737138,0.2052209973,0.5904605389,0.0436040871,-0.1867178828,-0.0387384258,-0.1282729656,-0.2344355732,0.3033230901,0.1187586039,-0.3259145319,-0.0401388668,0.195691213,-0.3326937258,0.2428995073,0.4663271904,-0.1162643284,-0.4240007997,-0.4418559968,0.4975842535,0.2040061206,-0.0451389961,0.4790868461,-0.079894267,-0.1677797139,0.0333918668,-0.1459572464,1.0368591547,0.0999154299,-0.0907199085,0.2094315439,-0.5453346372,0.171004191,-0.19995597,0.2204816341,-0.2778279483,-0.4051022232,0.0069353119,-0.2201943398,0.1650538594,-0.0130685773,-0.2186738849,0.2340096682,-0.0390748158,0.0920773447,0.1622831225,0.2636517882,-0.2369296104,0.1853297502,0.2661128938,0.0833000615,-0.2128532529,0.5148091912,-0.0760418549,-0.160902977,-0.1491775662,-0.1373740286,-0.5546498895,0.0599502251,-0.3144536614,0.2885302305,-0.2171665728,0.0433867201,0.0880814642,0.0923925862,0.0142036108,-0.059497349,-0.3775972724,0.2553164065,-0.0525227785,0.1412575394,0.26208812,0.0832052827,0.1101575568,-0.0912031159,-0.2372803092,0.0405232869,-0.0837490484,-0.1229647547,-0.0970741287,-0.1680072099,0.3867132366,-0.6676743031,-0.1573059112,-0.103407383,0.1436843872,-0.1959362477,0.0874958634,0.102919735,-0.0474428125,0.4278289676,0.0163506046,-0.4587227404,-0.1943021268,0.4307106733,0.1279550046,0.0916890055,0.444259882,-0.1370531917,-0.0376776531,0.0117737707,0.4629659355,-0.1131052375,-0.5498889089,0.2595935464,-0.4044882059,0.3425936699,0.0981684327,0.2810602784,0.5407380462,0.3864488602,-0.0264150258,-0.1146284193,0.0669916049,0.0011304211,0.07464616,0.3588258028,-0.0805719569,0.4045130908,0.0672632754,-0.0704931542,-0.2422812879,0.25442788,0.0156776998,0.0428695045,0.1455546021,0.0286411587,0.1355085522,-0.0117970947,-0.0390093587,0.1854173392,-0.2432226539,-0.0989609435,-0.1615243405,0.2079800814,0.0546912886,-0.1530820429,0.0456557199,-0.151010558,0.1255456358,-0.1162612811,-0.1137405038,0.1645741761,-0.0513619855,0.2183627039,0.1268445253,-0.2066084892,-0.0665399879,0.2602390945,-0.2344353348,0.4273049533,0.0245208964,0.3384079039,0.2047903687,-0.0108745098,-0.0850313157,-0.0598098226,0.0003589948,0.1567872465,0.301349014,-0.250828743,0.1260694265,0.3643406928,0.1374808848,0.4328534901,-0.131325081,-0.2906668186,0.1045469418,0.1784314662,-0.299235791,0.0161652621,0.4957776964,0.1222691759,-0.0567720011,0.0761141255,0.0944880024,-0.0311806705,-0.332786411,0.1645711809,0.4324413836,0.2256922126,0.3186547458,0.8666628599,-0.0204310995,-0.0256544836,0.0237761121,-0.1841366738,0.3195383251,0.2806243598,-0.1578601599,0.3437280059,-0.0056774742,0.3505115509,-0.3397983313,-0.367202431,-0.1076571792,0.2417443395,-0.0121823633,0.2198258638,0.2063263953,0.3108752072,-0.2241532803,-0.0063813683,-0.372656554,0.1676117182,0.0706150681,-0.0313569121,0.0524179302,-0.225984484,-0.2060041875,0.0640666187,-0.1661808938,0.0202165302,0.2706856728,-0.1795413345,-0.2028777748,-0.3367297351,-0.1908313483,-0.1135760844,0.1885900944,-0.4949007034,0.04654615,0.1343567818,-0.2942207754,0.1290152371,0.6200796962,0.4436339438,0.2979873419,-0.1408046186,-0.1054059044,0.2517394125,0.102725774,0.0467174351,0.1857618392,0.0489167087,-0.1277914792,0.3198865354,0.0569142103,-0.0990992934,0.0934144482,-0.2364149839,-0.1541102082,-0.3987446427,0.2347800583,-0.208598569,0.0547558181,0.0016943603,-0.0284705535,-0.3427087963,-0.0356006362,0.1651673317,-0.064795047,0.2333302945,0.1454223245,0.0216183551,-0.2116189897,0.2247880846,0.4161574543,-0.3166663051,-0.1842965931,-0.2350704372,-0.5166273117,0.2552207708,-0.2080881894,-0.1786124557,0.0614984892,0.318290025,-0.225325197,-0.0437424481,0.4135928452,0.2970720828,-0.2436300516,0.1413473636,-0.3079405427,-0.2109190077,-0.0562669411,0.0403397158,0.1673977524,-0.2650242448,0.2208800167,-0.1054110229,-0.0257281661,-0.2519944012,-0.033062879,0.134844467,-0.0843252838,0.4137382507,0.3345900476,0.2951545417,-0.0177687351,-0.1573621333,-0.2531405687,-0.2753590345,-0.2024574131,0.4000014365,-0.0340067893,0.2347235531,-0.2256833315,0.0578506254,-0.1563293189,0.4413224757,0.0167557131,0.0550507046,-0.2007739246,-0.2971840203,-0.0977731198,-0.1419035941,0.3616728187,0.3011946976,-0.1548299044,0.1341492832,0.0166279189,-0.4505012631,0.4645195901,0.0743985549,-0.0788485929,-0.1006376445,0.150979802,0.3283902109,-0.0582994074,-0.3652930856,0.1082647368,0.5874559879,0.0201444477,-0.0254815463,0.1324418783,-0.0440023355,-0.143429637,0.1158751696,0.1667719781,-0.0006540735,-0.4427075982,0.4158499539,-0.1364068687]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2837","title":"prepare_module issue when loading from read-only fs","comments":"Hello, I opened #2887 to fix this.","body":"## Describe the bug\r\n\r\nWhen we use prepare_module from a readonly file system, we create a FileLock using the `local_path`.\r\nThis path is not necessarily writable.\r\n\r\n`lock_path = local_path + \".lock\"`\r\n\r\n\r\n## Steps to reproduce the bug\r\n\r\nRun `load_dataset` on a readonly python loader file.\r\n```python\r\nds = load_dataset(\r\n python_loader, data_files={\"train\": train_path, \"test\": test_path}\r\n )\r\n```\r\n\r\nwhere `python_loader` is a path to a file located in a readonly folder.\r\n\r\n## Expected results\r\nThis should work I think?\r\n\r\n## Actual results\r\n\r\n```python\r\n return load_dataset(\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/load.py\", line 711, in load_dataset\r\n module_path, hash, resolved_file_path = prepare_module(\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/load.py\", line 465, in prepare_module\r\n with FileLock(lock_path):\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/utils\/filelock.py\", line 314, in __enter__\r\n self.acquire()\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/utils\/filelock.py\", line 263, in acquire\r\n self._acquire()\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/utils\/filelock.py\", line 378, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 30] Read-only file system: 'YOUR_FILE.py.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.8\r\n- PyArrow version: 3.0.0\r\n","comment_length":7,"text":"prepare_module issue when loading from read-only fs\n## Describe the bug\r\n\r\nWhen we use prepare_module from a readonly file system, we create a FileLock using the `local_path`.\r\nThis path is not necessarily writable.\r\n\r\n`lock_path = local_path + \".lock\"`\r\n\r\n\r\n## Steps to reproduce the bug\r\n\r\nRun `load_dataset` on a readonly python loader file.\r\n```python\r\nds = load_dataset(\r\n python_loader, data_files={\"train\": train_path, \"test\": test_path}\r\n )\r\n```\r\n\r\nwhere `python_loader` is a path to a file located in a readonly folder.\r\n\r\n## Expected results\r\nThis should work I think?\r\n\r\n## Actual results\r\n\r\n```python\r\n return load_dataset(\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/load.py\", line 711, in load_dataset\r\n module_path, hash, resolved_file_path = prepare_module(\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/load.py\", line 465, in prepare_module\r\n with FileLock(lock_path):\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/utils\/filelock.py\", line 314, in __enter__\r\n self.acquire()\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/utils\/filelock.py\", line 263, in acquire\r\n self._acquire()\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/datasets\/utils\/filelock.py\", line 378, in _acquire\r\n fd = os.open(self._lock_file, open_mode)\r\nOSError: [Errno 30] Read-only file system: 'YOUR_FILE.py.lock'\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.8\r\n- PyArrow version: 3.0.0\r\n\nHello, I opened #2887 to fix this.","embeddings":[-0.3615451157,0.0755590349,0.0289529301,0.1704782993,0.4606169164,0.0472189672,0.5346610546,0.2162520736,0.1541241556,-0.0172812417,-0.1880602241,0.5080477595,-0.2517548203,-0.1020716801,-0.0777731612,0.2706075609,0.0200665444,0.0112962145,-0.2530965805,-0.1416644603,-0.3549022675,0.083093904,-0.4137744606,-0.1727414429,-0.3624045849,-0.0010009212,0.0150155798,0.5845605135,-0.2485121489,-0.5039462447,0.1725386381,0.1473281085,0.1436156631,0.4981012642,-0.0001130025,0.1002123281,0.2270591855,-0.1892979294,-0.4881866574,-0.3046151698,-0.105461657,-0.1274337918,0.1908809245,-0.1867613792,-0.1393137425,-0.5304034948,-0.2640208304,-0.4016046822,0.3502630293,0.4838982224,0.2176769376,0.2739233971,0.0604529195,-0.0791170597,0.2170758843,0.1471763104,-0.1317860484,0.3534235954,0.0887826458,-0.2009129822,-0.1079101413,0.0262013674,-0.3538956642,-0.1007885262,0.3614571393,0.0793452784,0.2608650327,-0.0544152707,-0.1264526695,0.0608647391,0.1374147534,-0.1659426987,-0.3974211812,-0.6213773489,-0.0089646839,0.1135625988,0.2981308997,0.1206043288,-0.028576104,0.2407972813,0.0269516315,-0.1052403376,-0.0331734978,0.2052266598,0.154052645,0.3671686947,-0.180500254,0.1311748326,0.1440232545,0.1402127147,0.4073286653,-0.0684278086,0.0509521961,0.4604339898,-0.2682097852,0.233228758,0.0367487818,-0.0787241757,-0.0215088688,0.0229562968,-0.3697723448,-0.013175237,-0.0801893175,0.1265997291,0.2813589871,0.1346696764,-0.0728763118,0.4764606655,0.0243565273,-0.0190150123,-0.0353303105,-0.085020341,0.0485505611,-0.4161615372,0.0899737999,-0.0126990555,0.1692366898,-0.2055101991,-0.1103947237,0.1301618814,0.2248141915,-0.0499092154,0.0725574344,0.2323269993,0.138743788,0.1530821919,0.3242757618,0.0326625034,-0.3230699301,0.1935004294,-0.1256914288,0.0694168359,-0.313726604,0.1015437543,-0.0295194108,-0.053814251,0.2219032049,0.1839742959,-0.1543634087,-0.3048380017,0.3266907036,-0.0741040707,0.465683639,0.4492875636,0.0261696745,0.1335900873,0.2800299823,-0.4442369938,0.0123438919,0.3613704741,-0.2363121659,-0.4464117885,0.0718439817,0.1992449015,-0.1278736293,0.3288271129,0.1265934855,-0.3136154711,0.2732582092,-0.0790606961,0.170081228,-0.1526500136,-0.1688032597,-0.0858845711,0.1860392839,0.5844793916,0.134581238,0.0104342708,0.0764218792,-0.2388762832,0.1630235165,0.1288185865,0.0057586203,0.1372639537,-0.514036715,-0.1008746326,0.5935893655,-0.3952892721,-0.3501677215,0.6340183616,-0.3967165053,0.0048385118,-0.0934316888,0.0626902506,-0.166485697,-0.0275396761,0.2753303647,0.2470519692,0.1967827529,-0.0344532467,-0.4271615744,-0.1154866144,0.0498031043,0.0919567347,0.214132905,0.1385164559,0.1722756624,0.0567654446,0.576087594,0.0357400812,-0.0619192496,0.3428143859,0.1387649626,0.1469044387,0.0154709714,-0.2648651302,-0.2714800239,0.2259671688,-0.2695763111,-0.0706239566,-0.3946177959,0.0283291154,-0.1896777302,0.0181036517,-0.3803439736,-0.0279352795,0.1331651956,0.1340157241,-0.2499096096,0.1763952225,-0.2169623375,0.3940416873,-0.2785477638,0.0401713327,-0.148022145,0.0975229666,0.0217893887,-0.3167933226,-0.0070293364,0.1229512542,0.2782101631,-0.193747133,-0.2061221004,0.5448225737,0.2801107466,0.2208112627,-0.4886832535,0.1083517522,0.0526437797,0.1362427771,-0.2638491094,-0.1477313936,0.2089962065,-0.1264919341,0.0085178427,0.3692706227,-0.3772579134,0.2040560395,-0.1571952105,0.1213087067,-0.1043784693,0.0197371058,-0.0840966851,-0.4397590756,0.1749591827,0.2276804894,0.3190254867,0.239686653,0.274345845,-0.2504338622,0.3213247061,-0.0315745883,0.147301048,-0.2104073316,-0.0419110656,-0.0665487573,0.1402689368,0.2549570501,0.3706813455,0.0545233265,0.1089363694,0.5284523368,-0.0026596836,-0.2469557673,0.2032437325,-0.0277540442,0.17541641,0.4235455692,-0.0954023823,0.0573463179,-0.303596884,0.091017276,0.0315084085,0.3276870549,-0.4230449498,0.2000810355,-0.3785784245,0.6610643268,-0.0804011077,-0.0024226895,-0.0371042341,-0.2586273849,-0.289814949,0.415081799,-0.0393697545,0.0511742458,-0.4830292165,-0.1440440565,0.0967916921,-0.1691366881,-0.0456406921,-0.0908025131,-0.1695038676,-0.0301344581,0.3343069553,0.0038354748,0.2205538899,-0.1680856645,-0.3577037752,0.0026093912,-0.156212911,0.1012139767,-0.0958915204,0.2751527131,0.1804755926,-0.2171112895,0.2612597942,0.1766691655,0.2082221806,-0.1173339263,-0.0566928573,0.0787561536,-0.0823798701,-0.0695103854,-0.4069253504,-0.3410567343,0.0462749861,-0.3884555995,0.2614655197,0.1027328745,-0.0414920747,0.29346174,0.071922712,0.1918885708,-0.2724782526,0.2787229419,-0.2098328173,-0.3823747337,0.3349430263,-0.2632080019,-0.191938594,0.0326675735,0.091447182,-0.0494222417,0.1870202273,-0.3827704191,-0.1279921085,-0.1663241982,-0.0052514705,-0.4764178395,0.0121695977,0.1992472559,0.2870612741,-0.0297039747,-0.2782786787,0.1061678231,0.2256313711,-0.2836254835,-0.1796613038,-0.0542474277,0.1620976925,0.2196423113,0.2789154649,0.0494113378,0.2672840655,0.2495362312,-0.0872432664,0.601487875,-0.2262182087,-0.2318103164,-0.1176411659,0.1415583491,-0.0918707475,0.0974385291,-0.0824679062,0.1802934408,-0.1910606772,-0.117612496,-0.0696774498,-0.4342952073,0.066268377,0.0176196024,0.3826934099,-0.0425229818,0.4123843312,0.0386898965,0.0394405685,0.3091135323,0.222222358,0.3552841246,-0.0775424168,-0.1985890269,0.3166635036,-0.1220260561,-0.0746105239,-0.0493275486,0.6063399315,0.0643918961,-0.1706492305,0.1342133284,-0.1887139082,0.463975668,0.2841457427,0.1854385883,0.4695688784,0.0061455532,-0.4709876478,-0.0953393951,0.0593526214,0.1957202703,-0.1744700223,0.2657916248,0.0012027553,-0.3747076094,-0.1961618662,0.2092413902,-0.284113884,-0.57484442,-0.0602412671,-0.1175609604,-0.5115852952,-0.0298701953,0.0779625773,-0.001417728,-0.1903055757,-0.0648643449,-0.2446544617,-0.0683475211,-0.0242092591,-0.1280129403,0.141795814,-0.1169269085,0.2914534509,0.4778870642,-0.2165145427,0.0474382117,0.572752893,-0.1095179021,0.0142181171,0.0492609665,-0.3631398976,-0.0699389428,0.3055059016,-0.2395135462,-0.0459696166,0.2824029624,-0.1126741469,-0.3837956786,0.0030722667,0.1914007515,-0.2450546175,-0.0560250692,-0.5997166634,0.2412534207,0.118539378,-0.2235057354,0.1170910448,0.2133340091,-0.415946424,0.5759008527,0.1936140805,0.7597817183,-0.1088913307,0.264213711,0.3265396655,0.0003088665,-0.1690466553,-0.1940424144,-0.1271480471,-0.4136402011,-0.2599995434,-0.0107920887,-0.3368870914,0.159552902,0.0309137683,0.2243182063,-0.0030712553,-0.2117538899,-0.4166780412,-0.0148398438,0.0591158159,-0.0451406203,-0.1436314881,-0.3729967177,0.1872561723,0.2511239648,0.0437246859,0.036838863,0.0647738725,0.0696307048,-0.0890980884,-0.050260324,0.0937275514,-0.1036123931,0.211707145,-0.1442716867,0.0830783695,0.1310213953,0.1724354327,0.1655641347,0.1476754844,0.0153289586,-0.1186896339,-0.0003178229,0.1287066638,-0.0654998645,0.0644868612,0.4023944438,0.0262893755,-0.1039805487,-0.0024635077,-0.3104089797,0.03907381,0.1498880237,0.164171949,0.3338457942,-0.2537143826,-0.3359835148,-0.1192316115,0.1811166108,-0.1282703429,0.1330957562,0.3400221765,-0.154437393,0.204138428,-0.1948488802,-0.4045990109,-0.0032015988,0.6593111753,-0.0247135144,0.0077341939,0.529576242,0.0431572124,-0.2265129536,-0.1641213298,0.1040750891,-0.2617243528,-0.1685379893,0.0791290626,0.0155376717,0.1345810443,-0.0760851204,0.2573824525,0.1497277468,-0.1516906917,-0.0288136844,-0.290453434,-0.6129739881,-0.3375506997,-0.034060698,0.1864067018,0.2397548407,-0.0965103135,-0.298448801,-0.1132525131,-0.3038001359,-0.0175061598,-0.1531082392,0.0584409386,0.1029860601,-0.0057780514,0.0974783674,-0.0904091969,0.1868555546,0.126792863,-0.0437172838,-0.1864301413,-0.3665138483,0.1899230629,0.3445058763,-0.0511882864,0.0245632976,-0.4087621272,-0.2329588979,-0.172446236,0.2015915811,0.6735688448,0.0654939264,0.1407678276,-0.0267195385,0.1106121019,-0.3436069787,0.1412900239,-0.4008731842,-0.2820969224,-0.0469953716,-0.0145156244,-0.1941575706,0.0676129684,-0.0438858829,-0.1702922136,0.1523228586,0.2536627054,0.4651986361,-0.2271796316,0.1018027589,0.0810985565,0.1945170462,0.2208149731,-0.2026748061,-0.1200169697,0.3134765923,0.192284137,-0.0411984213,-0.0604379363,-0.1193829328,0.048549708,0.1092850789,0.1839449704,0.0794543102,-0.1112131551,-0.0539432988,0.2337377667,-0.1029600427,-0.0738672838,0.0407083742,0.0679225102,-0.0439740457,0.0436786525,0.1204121038,-0.0653446689,0.0583736971,0.4206710458,-0.2134313732,0.2311044782,-0.2272040844,0.0674328059,-0.2625122368,-0.5102348924,0.3012460768,0.010563679,0.0249814782,0.1316508353,-0.2630089223,-0.0643992946,-0.1580971479,0.132748276,0.3357828259,0.1966502219,-0.2287924439,-0.1627687365,0.2611636519,-0.1991122812,-0.351210624,-0.0863004029,0.0362357944,-0.1596467942,0.1395542026,0.3181827366,0.1312958747,-0.1749210656,-0.0021455688,0.4520097077,0.202346772,-0.3160943389,0.00314878,0.0840832591,0.2191612571,0.1363385618,-0.2629844546,0.3859756589,0.37821123,-0.0800785944,-0.1186521277,0.0766458511,-0.0718717873,-0.1554455757,0.3632100224,-0.3234453201,0.272999227,0.32299909,0.1604667008,-0.2178542912,0.0162076764,0.0957674682,-0.0174971446,-0.2103169262,0.5438808799,0.2923335731,-0.2348906994,-0.1776558459,0.2322452515,-0.1330354363,-0.0784996301,0.4625235498,0.1248167232,0.3864832819,-0.1769706458,0.0619905144,0.2888757288,0.3146129251,0.0897049382,0.0788578019,-0.1823353767,0.0708802119,-0.490283668,0.1510576457,0.105628565,-0.0458901338,-0.1274360865,-0.1867986768,0.070763737,0.0293396432,0.117317684,0.0960991681,0.2131087482,0.2911879718,0.0624628626,-0.1925524324,-0.1473698765,-0.1330533177,-0.1236855611,-0.3090548217,0.0245395862,-0.2560054958,0.1193127707,-0.0983079597,-0.0034548808,-0.210264191,0.1687549949,0.5240650177,0.2034036964,0.2290897071,-0.2489489913,-0.33784163,-0.5055366158,0.0874785706,-0.0207588375,-0.0255117211,0.0106428638,0.3047583997,-0.3538104892,-0.1242372096,-0.4722252488,0.2078437358,-0.0853549168,0.001077548,-0.5824391842,0.407544136,-0.1239091232,0.2060529143,0.1441422254,0.091065459,0.198119387,0.3234096169,-0.2430427819,-0.2224521041,0.5419303179,-0.2000518441,-0.1375153363,0.2815092206,0.0338524282,0.0139344456,-0.1124353111,-0.5952610373,0.0466245078,0.4774067402,-0.1012789011,0.17818892,-0.1132893562,-0.1231916323,0.2070800066,-0.1508630663,0.2771522105,-0.0426432155,-0.147565484,0.1236010343,-0.2099152803]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2831","title":"ArrowInvalid when mapping dataset with missing values","comments":"Hi ! It fails because of the feature type inference.\r\n\r\nBecause the first 1000 examples all have null values in the \"match\" field, then it infers that the type for this field is `null` type before writing the data on disk. But as soon as it tries to map an example with a non-null \"match\" field, then it fails.\r\n\r\nTo fix that you can either:\r\n- increase the writer_batch_size to >2000 (default is 1000) so that some non-null values will be in the first batch written to disk\r\n```python\r\ndatasets = datasets.map(lambda e: {'labels': e['match']}, remove_columns=['id'], writer_batch_size=2000)\r\n```\r\n- OR force the feature type with:\r\n```python\r\nfrom datasets import Features, Value\r\n\r\nfeatures = Features({\r\n 'conflict': Value('int64'),\r\n 'date': Value('string'),\r\n 'headline': Value('string'),\r\n 'match': Value('float64'),\r\n 'label': Value('float64')\r\n})\r\ndatasets = datasets.map(lambda e: {'labels': e['match']}, remove_columns=['id'], features=features)\r\n```","body":"## Describe the bug\r\nI encountered an `ArrowInvalid` when mapping dataset with missing values. \r\nHere are the files for a minimal example. The exception is only thrown when the first line in the csv has a missing value (if you move the last line to the top it isn't thrown).\r\n[data_small.csv](https:\/\/github.com\/huggingface\/datasets\/files\/7037838\/data_small.csv)\r\n[data.csv](https:\/\/github.com\/huggingface\/datasets\/files\/7037842\/data.csv)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndatasets = load_dataset(\"csv\", data_files=['data_small.csv'])\r\n\r\ndatasets = datasets.map(lambda e: {'labels': e['match']},\r\n remove_columns=['id'])\r\n```\r\n\r\n## Expected results\r\nNo error\r\n\r\n## Actual results\r\n```\r\nFile \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Invalid null value\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.5.0\r\n- Platform: Linux-5.11.0-25-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyTorch version (GPU?): 1.7.1+cpu (False)\r\n- Tensorflow version (GPU?): 2.4.1 (False)\r\n- Using GPU in script?: no\r\n- Using distributed or parallel set-up in script?: no\r\n","comment_length":134,"text":"ArrowInvalid when mapping dataset with missing values\n## Describe the bug\r\nI encountered an `ArrowInvalid` when mapping dataset with missing values. \r\nHere are the files for a minimal example. The exception is only thrown when the first line in the csv has a missing value (if you move the last line to the top it isn't thrown).\r\n[data_small.csv](https:\/\/github.com\/huggingface\/datasets\/files\/7037838\/data_small.csv)\r\n[data.csv](https:\/\/github.com\/huggingface\/datasets\/files\/7037842\/data.csv)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndatasets = load_dataset(\"csv\", data_files=['data_small.csv'])\r\n\r\ndatasets = datasets.map(lambda e: {'labels': e['match']},\r\n remove_columns=['id'])\r\n```\r\n\r\n## Expected results\r\nNo error\r\n\r\n## Actual results\r\n```\r\nFile \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Invalid null value\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.5.0\r\n- Platform: Linux-5.11.0-25-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyTorch version (GPU?): 1.7.1+cpu (False)\r\n- Tensorflow version (GPU?): 2.4.1 (False)\r\n- Using GPU in script?: no\r\n- Using distributed or parallel set-up in script?: no\r\n\nHi ! It fails because of the feature type inference.\r\n\r\nBecause the first 1000 examples all have null values in the \"match\" field, then it infers that the type for this field is `null` type before writing the data on disk. But as soon as it tries to map an example with a non-null \"match\" field, then it fails.\r\n\r\nTo fix that you can either:\r\n- increase the writer_batch_size to >2000 (default is 1000) so that some non-null values will be in the first batch written to disk\r\n```python\r\ndatasets = datasets.map(lambda e: {'labels': e['match']}, remove_columns=['id'], writer_batch_size=2000)\r\n```\r\n- OR force the feature type with:\r\n```python\r\nfrom datasets import Features, Value\r\n\r\nfeatures = Features({\r\n 'conflict': Value('int64'),\r\n 'date': Value('string'),\r\n 'headline': Value('string'),\r\n 'match': Value('float64'),\r\n 'label': Value('float64')\r\n})\r\ndatasets = datasets.map(lambda e: {'labels': e['match']}, remove_columns=['id'], features=features)\r\n```","embeddings":[-0.0327268355,-0.1531262994,0.1040895432,0.2286237329,0.0726004243,0.1366074681,0.3329794109,0.4905757904,-0.0207695644,0.2300736755,0.2338506132,0.4390177727,0.0271520261,-0.1494164914,-0.1128671244,-0.0434916839,0.036237888,0.1175362915,-0.0013278106,-0.0447311327,-0.378742218,0.0844402388,-0.3389022946,0.0575111024,-0.3012346625,-0.1077298597,-0.0090291733,0.0034838135,0.1298887879,-0.4910511672,0.2923193276,-0.5293874741,-0.1505012959,0.1626194119,-0.0001158711,-0.0339868702,0.2518496215,0.0519128256,-0.0784546211,-0.2758250237,-0.1568179429,-0.2185934782,-0.0141735272,-0.1653571576,0.1762025803,-0.0375164859,0.0760230497,-0.1961084604,-0.0201382134,0.2709065378,0.1460493654,0.4126615822,0.217869252,0.0892388672,0.4545611143,0.0507913604,-0.2142468095,0.3872686923,0.150176838,-0.2261426896,0.1055756882,0.3145453632,-0.0312069766,0.0160816535,0.3101096451,0.176736027,-0.0633132234,-0.2293644398,-0.0956780314,0.1412264705,0.121313028,-0.4014458358,-0.430261761,-0.2851817012,0.0836299956,-0.3390862942,0.3403961062,0.168567881,-0.1564158648,0.1623672247,0.1155143604,0.2014273405,-0.192848891,0.1499879211,-0.1042830124,0.0833818689,0.0503504686,0.2609374523,-0.0560738258,-0.2023003697,-0.2976311743,-0.1810262501,0.0063066739,0.2630151212,-0.3672583699,-0.2198308855,-0.1504853517,-0.1560584158,0.1592742801,-0.2276551127,0.2990320325,-0.1067543626,0.0777331293,0.4128786623,0.0608040318,-0.1506104469,0.1527083069,0.108326219,0.1780028194,-0.4061113,0.2636467516,-0.0342918187,0.1340156496,-0.5859510303,0.5141164064,0.2606024444,0.6728469133,-0.1126297936,-0.4689024091,0.1276107281,-0.7892329097,0.3337589502,-0.135780707,0.1627140641,0.0130745769,0.0863919556,0.0484194085,0.2774145603,-0.1620678008,-0.2743251026,0.0147257512,-0.0746785253,-0.0540683046,-0.0446719974,0.3431840837,-0.189890027,0.2458141446,0.1759653986,-0.1122789308,-0.0045988979,0.2626313269,-0.3046376705,0.1480023861,0.1737484187,-0.4271009266,0.2794370651,0.0387199447,-0.146589011,-0.1892317235,0.2775795162,-0.1205456331,-0.1588583738,-0.2165751904,0.1482956111,0.0022299229,-0.2128342688,-0.0540005676,-0.0019999465,0.242374599,-0.173490569,0.1846102625,-0.219979018,0.0187497139,-0.3760899603,-0.0113685085,0.3736341894,-0.6385915279,0.070836775,0.1267937571,-0.0185744036,-0.0337321162,0.3634842634,-0.1870690137,-0.013671454,-0.0591617823,0.5252671838,-0.0086465375,-0.1326466799,-0.4373976886,0.131386131,-0.074395977,-0.2444548905,0.0401039533,0.2435707748,0.1417698264,-0.1496624351,-0.01733483,-0.1213846132,-0.1300859749,0.2243183255,-0.3260702193,-0.0411789082,0.1720778644,0.0506165735,-0.1469437182,-0.2109326124,0.1945364773,-0.6835855842,0.0329079628,-0.2068118751,0.2666673362,0.3962259293,0.4044350684,-0.0465761051,0.1552023888,-0.1333547831,-0.8110236526,0.1985868216,0.100847736,-0.1821604669,-0.4811460078,-0.4098103642,-0.1061270237,0.3230985403,0.169046849,0.1539756507,0.0751769021,-0.0067508742,-0.122685194,-0.0598925687,-0.2849860191,-0.298654139,-0.2172655761,0.1865486056,-0.2249355167,0.2575946748,-0.013920987,-0.2167794704,-0.0960751399,0.0344255976,0.1761991829,-0.1273743063,0.013591324,0.4820096791,0.1067622378,0.1463912725,-0.1182710975,0.0272354595,0.3414061964,-0.4875248373,-0.2777242064,0.3717956841,0.3092466295,0.0654949769,-0.1880118698,0.367030859,-0.0085762525,0.142661944,-0.243231982,-0.0240107384,0.0476458743,-0.0050884029,-0.1706681848,-0.0740773231,-0.0837703869,-0.0132249454,0.2308350503,0.1707009226,-0.648462832,0.0064496514,0.1777651906,-0.1262826473,0.1471400559,0.0463455319,-0.2821579576,0.3659059703,0.3918358088,0.0322905146,0.3437526524,0.197664246,-0.0775276646,0.0050951457,-0.1521711648,-0.0824590325,0.3486306369,-0.0050706733,0.0830795988,0.1302965432,0.1771517992,0.2778594494,-0.1754120141,0.0187847074,0.2271973938,0.1906450242,-0.5900598764,-0.0668672845,-0.1082044542,0.0791586116,-0.3532581627,-0.1602440476,-0.0837709084,-0.2679263949,-0.0238328073,0.1787119508,-0.4540548921,0.2962757349,-0.0503031909,0.0645017549,0.2097569555,-0.2873474061,-0.1986872554,-0.1805445403,-0.206972599,0.0005408119,0.1689912528,-0.0692994669,0.2951590121,0.2671799064,-0.105649069,-0.5031932592,-0.4341720939,0.0272096228,-0.228622973,0.1287030727,0.1987026632,0.2116244137,-0.0771777779,-0.3438918293,0.1970792711,0.0353334136,-0.2919718623,0.1465553045,-0.0171925724,-0.0445718765,-0.1751176119,-0.4850569963,-0.038820982,-0.228159681,0.2234225869,-0.3103996813,0.0502251722,0.163063243,0.3292430937,0.1380271018,-0.1170724183,-0.1576803476,0.1402166784,0.2776943147,0.5713930726,-0.0197698306,-0.3324659467,0.2355109453,-0.1852093339,-0.0800568014,0.4506347477,-0.2126406878,-0.0239198282,-0.1131706908,0.6397444606,-0.0702653602,-0.0265891757,0.3226291835,0.2941908538,0.0133324713,-0.1408387423,-0.3169084787,0.062488731,-0.0236808807,0.2940809429,0.0561678745,0.3970873654,0.0143699041,0.7554779053,0.1163189709,-0.2453815788,0.1562790424,-0.2184980959,-0.0008914333,0.1268464029,-0.1204357743,-0.1177736074,-0.1749807596,-0.0615210682,-0.1283867359,-0.1914948374,-0.5315887928,0.0941079631,-0.0040597455,-0.4546507299,-0.3176428974,0.2134097368,-0.0659967661,0.3006450236,0.0850822702,0.0237074234,-0.2731817663,-0.0983999297,0.2626481056,-0.0619853921,0.2583997548,-0.1927739233,0.0116458945,-0.2957198918,-0.2309028059,0.4249314368,0.2806850672,0.1446990371,0.1786822975,-0.2188019603,-0.0082682148,0.1248107031,0.6407935023,-0.1843993962,-0.0035300697,-0.0733780116,0.0795039535,-0.4673798978,0.1887799054,-0.2618241608,0.1212380379,0.2620203197,0.3791087568,-0.4262292981,0.0409371145,0.2719637454,0.0301522668,0.0009182586,-0.1788763106,-0.1175876558,0.0538972989,-0.1162515879,-0.0002532091,0.2568055391,-0.0368428715,0.1809669733,-0.1021706387,-0.3370203078,-0.0148665449,0.007732376,-0.0051575401,0.3421080709,-0.2208801657,0.2672509849,0.0290986337,0.3201949894,0.5259217024,0.3236470222,0.230843693,-0.591368556,0.0479080267,-0.0493211299,0.3719320893,0.2339028269,-0.0153828561,0.0391952023,-0.0169088654,0.0757478252,-0.1975085735,-0.0415865444,0.320069164,0.1617128402,-0.0220546499,0.0735812411,0.545021534,-0.2178991288,-0.0845392942,0.5981555581,0.4256738424,-0.1784007847,0.3142663538,0.265111655,0.8266257644,0.242348671,-0.0428322181,0.5020028949,-0.2361523062,0.375300765,0.4359337986,0.1408500075,-0.4725413322,0.0318128802,-0.0056982269,-0.0902223662,0.0566545278,0.156603232,-0.0879654139,0.1055393666,0.1479057968,0.3768041134,0.0117138168,-0.2954432964,-0.1822736114,-0.2319571376,-0.048127763,0.0287756603,0.0588858016,0.0464371666,-0.0251412205,-0.2543234229,-0.3001084924,-0.1459304094,-0.0274330061,-0.0137783093,-0.6243045926,0.1249296293,0.3038370013,-0.1498272121,0.1048352942,0.0283966642,-0.0029960247,0.045398768,-0.1749745607,-0.0308621209,0.1676077694,0.0359348468,0.1077979282,-0.1268626601,0.1436228752,0.0088677462,-0.3379888535,0.1709951311,-0.3489298522,-0.0887772962,0.2501184046,-0.0525682755,-0.2960265875,-0.0757637918,-0.1102009714,-0.1591259986,-0.1553000361,-0.0104976213,0.0702018738,-0.1015283167,0.0871764198,-0.0822184831,-0.1431222707,-0.2805766165,0.0563413911,0.3491201997,-0.1585402638,0.2250405103,0.746440649,-0.1259889603,-0.0978427157,-0.1068751514,0.275321126,0.2847056687,-0.7513645887,-0.033911176,-0.2546555102,0.1422490925,0.0434250832,0.3635244071,0.4133291841,-0.0414074101,-0.1273898929,-0.2120484859,-0.3010553122,0.2500478029,0.1703124046,0.1966131479,-0.1225019917,0.1032615528,0.0622730441,-0.2535413504,-0.2832318246,0.1140101403,-0.082818985,0.2677064836,0.1502032876,0.284286201,-0.0032054107,-0.2481000274,0.0826087147,0.2167167962,-0.0332854278,-0.122351639,-0.1673523188,0.1068496183,-0.0166369546,0.0131727001,0.0193716213,-0.1955142915,-0.0152198179,-0.066260092,0.0393886007,0.0908396393,-0.1936000437,0.3321723938,0.0795067772,0.1329206973,-0.3743384778,-0.0004383257,-0.0297724344,0.0948932096,-0.1321002543,0.2827996314,-0.2116814405,0.1072418839,0.0784711912,0.099616304,-0.3790988028,-0.1276966631,0.3835837841,-0.3355983794,-0.2621270418,0.2332983464,0.1428565234,0.0325820073,-0.2208074629,0.0428407565,0.103923969,0.20104298,-0.3584828973,0.2307639271,0.0241831783,0.1520148665,0.3160282671,0.1300290376,0.1271678805,-0.1918479353,0.0712350607,-0.0569212548,0.101578407,-0.215393424,-0.2106611878,0.5099640489,-0.0421366692,0.1484289914,0.3295753002,0.3196731806,0.503620863,0.0196692962,0.2491696477,0.2726070881,0.1587906629,0.0403381251,-0.0816662535,-0.2444848716,0.1164995059,0.1713621467,-0.100070931,-0.020089997,0.4890686572,0.4165976644,-0.0280174501,-0.3324064314,0.161188826,-0.0731745884,-0.2269110233,0.0226643905,-0.6775984764,-0.0025767274,-0.193576768,-0.1397781968,-0.1691484302,-0.1219594181,0.8168914914,0.2213880569,-0.0209993217,-0.3886559308,0.1076374725,0.3059972227,0.487688601,-0.2539097965,0.1352225989,0.1234420314,-0.2828816175,-0.1400599182,0.4913206995,0.4069623053,0.2249293029,-0.1212654039,0.0403519124,-0.0347090177,-0.0543514192,0.1517041326,0.1644178778,0.0987065881,0.1409580708,0.36448282,0.0870506987,-0.0920832455,0.0718838647,0.013167874,0.2570971251,-0.2345177084,-0.1019177288,-0.3865156174,-0.438046515,-0.0280458182,-0.2071051896,-0.2199875414,0.0009460147,0.4115730226,0.088200137,0.0986720473,0.1032278687,0.0817171261,0.0474650115,0.2402690351,0.5085468292,-0.1268954724,-0.2856565118,0.0516216941,-0.715718329,0.2013088912,-0.0577675365,0.2043696493,0.2648020089,0.3000234962,0.1720285267,0.2032229155,0.1652730405,-0.0015439541,0.0217056293,0.2742359638,-0.1224540249,-0.1925009936,-0.0138758328,-0.280415982,0.0767689422,-0.4350309372,0.232589215,-0.0922950506,-0.0119196782,-0.0938092545,-0.0700891167,0.1152580082,0.0317848884,0.4940911233,0.0795048028,0.1511560529,0.038688466,-0.1448906809,-0.1115941107,-0.2258879393,-0.2643378079,0.2339619845,0.0837763399,0.3281964958,0.0952427238,-0.2597880065,-0.0061515686,0.0315794311,-0.034317892,-0.2390241325,-0.2667894363,0.3094367087,-0.2738668621,-0.0049880832,-0.1516632885,0.0633130521,0.0026383053,0.1307356358,-0.319498688,-0.4689128399,0.3250766098,-0.3774596155,-0.1080545411,0.0280196536,0.3486168683,0.485396862,0.1548854113,-0.5569854975,-0.1189387292,0.4156110287,-0.1695117801,-0.3143323064,-0.1939266771,0.0120867556,0.0381229669,0.0517548621,0.5593414903,0.1047479436,0.0732036233,0.1368073076,-0.4698798954]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2826","title":"Add a Text Classification dataset: KanHope","comments":"Hi ! In your script it looks like you're trying to load the dataset `bn_hate_speech,`, not KanHope.\r\n\r\nMoreover the error `KeyError: ' '` means that you have a feature of type ClassLabel, but for a certain example of the dataset, it looks like the label is empty (it's just a string with a space). Can you make sure that the data don't have missing labels, and that your dataset script parses the labels correctly ?","body":"## Adding a Dataset\r\n- **Name:** *KanHope*\r\n- **Description:** *A code-mixed English-Kannada dataset for Hope speech detection*\r\n- **Paper:** *https:\/\/arxiv.org\/abs\/2108.04616* (I am the author of the paper}\r\n- **Author:** *[AdeepH](https:\/\/github.com\/adeepH)*\r\n- **Data:** *https:\/\/github.com\/adeepH\/KanHope\/tree\/main\/dataset*\r\n- **Motivation:** *The dataset is amongst the very few resources available for code-mixed Dravidian languages*\r\n\r\n- I tried following the steps as per the instructions. However, could not resolve an error. Any help would be appreciated.\r\n\r\n- The dataset card and the scripts for the dataset *https:\/\/github.com\/adeepH\/datasets\/tree\/multilingual-hope-speech\/datasets\/mhs_eval*\r\n\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset bn_hate_speech\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/bn_hate_speech\/default\/0.0.0\/5f417ddc89777278abd29988f909f39495f0ec802090f7d8fa63b5bffb121762...\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 data = load_dataset('\/content\/bn')\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 850 ignore_verifications=ignore_verifications,\r\n 851 try_from_hf_gcs=try_from_hf_gcs,\r\n--> 852 use_auth_token=use_auth_token,\r\n 853 )\r\n 854 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 614 if not downloaded_from_gcs:\r\n 615 self._download_and_prepare(\r\n--> 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 617 )\r\n 618 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 691 try:\r\n 692 # Prepare split will record examples associated to the split\r\n--> 693 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 694 except OSError as e:\r\n 695 raise OSError(\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 1107 disable=bool(logging.get_verbosity() == logging.NOTSET),\r\n 1108 ):\r\n-> 1109 example = self.info.features.encode_example(record)\r\n 1110 writer.write(example, key)\r\n 1111 finally:\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in encode_example(self, example)\r\n 1015 \"\"\"\r\n 1016 example = cast_to_python_objects(example)\r\n-> 1017 return encode_nested_example(self, example)\r\n 1018 \r\n 1019 def encode_batch(self, batch):\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in encode_nested_example(schema, obj)\r\n 863 if isinstance(schema, dict):\r\n 864 return {\r\n--> 865 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 866 }\r\n 867 elif isinstance(schema, (list, tuple)):\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in (.0)\r\n 863 if isinstance(schema, dict):\r\n 864 return {\r\n--> 865 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 866 }\r\n 867 elif isinstance(schema, (list, tuple)):\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in encode_nested_example(schema, obj)\r\n 890 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 891 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 892 return schema.encode_example(obj)\r\n 893 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 894 return obj\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in encode_example(self, example_data)\r\n 665 # If a string is given, convert to associated integer\r\n 666 if isinstance(example_data, str):\r\n--> 667 example_data = self.str2int(example_data)\r\n 668 \r\n 669 # Allowing -1 to mean no label.\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in str2int(self, values)\r\n 623 if value not in self._str2int:\r\n 624 value = str(value).strip()\r\n--> 625 output.append(self._str2int[str(value)])\r\n 626 else:\r\n 627 # No names provided, try to integerize\r\n\r\nKeyError: ' '\r\n```","comment_length":75,"text":"Add a Text Classification dataset: KanHope\n## Adding a Dataset\r\n- **Name:** *KanHope*\r\n- **Description:** *A code-mixed English-Kannada dataset for Hope speech detection*\r\n- **Paper:** *https:\/\/arxiv.org\/abs\/2108.04616* (I am the author of the paper}\r\n- **Author:** *[AdeepH](https:\/\/github.com\/adeepH)*\r\n- **Data:** *https:\/\/github.com\/adeepH\/KanHope\/tree\/main\/dataset*\r\n- **Motivation:** *The dataset is amongst the very few resources available for code-mixed Dravidian languages*\r\n\r\n- I tried following the steps as per the instructions. However, could not resolve an error. Any help would be appreciated.\r\n\r\n- The dataset card and the scripts for the dataset *https:\/\/github.com\/adeepH\/datasets\/tree\/multilingual-hope-speech\/datasets\/mhs_eval*\r\n\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset bn_hate_speech\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/bn_hate_speech\/default\/0.0.0\/5f417ddc89777278abd29988f909f39495f0ec802090f7d8fa63b5bffb121762...\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 data = load_dataset('\/content\/bn')\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 850 ignore_verifications=ignore_verifications,\r\n 851 try_from_hf_gcs=try_from_hf_gcs,\r\n--> 852 use_auth_token=use_auth_token,\r\n 853 )\r\n 854 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 614 if not downloaded_from_gcs:\r\n 615 self._download_and_prepare(\r\n--> 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 617 )\r\n 618 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 691 try:\r\n 692 # Prepare split will record examples associated to the split\r\n--> 693 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 694 except OSError as e:\r\n 695 raise OSError(\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 1107 disable=bool(logging.get_verbosity() == logging.NOTSET),\r\n 1108 ):\r\n-> 1109 example = self.info.features.encode_example(record)\r\n 1110 writer.write(example, key)\r\n 1111 finally:\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in encode_example(self, example)\r\n 1015 \"\"\"\r\n 1016 example = cast_to_python_objects(example)\r\n-> 1017 return encode_nested_example(self, example)\r\n 1018 \r\n 1019 def encode_batch(self, batch):\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in encode_nested_example(schema, obj)\r\n 863 if isinstance(schema, dict):\r\n 864 return {\r\n--> 865 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 866 }\r\n 867 elif isinstance(schema, (list, tuple)):\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in (.0)\r\n 863 if isinstance(schema, dict):\r\n 864 return {\r\n--> 865 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 866 }\r\n 867 elif isinstance(schema, (list, tuple)):\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in encode_nested_example(schema, obj)\r\n 890 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 891 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 892 return schema.encode_example(obj)\r\n 893 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 894 return obj\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in encode_example(self, example_data)\r\n 665 # If a string is given, convert to associated integer\r\n 666 if isinstance(example_data, str):\r\n--> 667 example_data = self.str2int(example_data)\r\n 668 \r\n 669 # Allowing -1 to mean no label.\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/features.py in str2int(self, values)\r\n 623 if value not in self._str2int:\r\n 624 value = str(value).strip()\r\n--> 625 output.append(self._str2int[str(value)])\r\n 626 else:\r\n 627 # No names provided, try to integerize\r\n\r\nKeyError: ' '\r\n```\nHi ! In your script it looks like you're trying to load the dataset `bn_hate_speech,`, not KanHope.\r\n\r\nMoreover the error `KeyError: ' '` means that you have a feature of type ClassLabel, but for a certain example of the dataset, it looks like the label is empty (it's just a string with a space). Can you make sure that the data don't have missing labels, and that your dataset script parses the labels correctly ?","embeddings":[-0.2438095212,-0.0837450847,-0.1076410264,0.1681506783,0.2711259723,0.0154425018,0.1767380387,0.4127865434,0.1465682834,0.086858876,-0.1648917794,-0.0779896751,-0.2590732574,0.2801799476,-0.2317627817,-0.2915211916,-0.0884709209,0.1522613019,-0.1177391112,-0.0643954799,-0.0891423598,0.3205003142,-0.1597214788,0.1476871818,-0.2826568484,-0.3294512033,-0.1136720181,0.0935218558,-0.3491749763,-0.4404600263,0.265022397,-0.052807495,0.1387756914,0.3857922852,-0.0001095746,-0.0700783134,0.31012097,-0.1751648784,-0.3034772277,-0.2981868386,-0.0789678693,-0.0469495505,-0.023372205,-0.3721512854,-0.2534294724,0.1050642058,-0.0610323809,-0.4590233564,0.5346906781,0.650282979,0.229791224,0.0913035348,0.3871519566,-0.2322711349,0.0315533839,0.2178065032,-0.2045287341,0.3952790499,0.2633341849,-0.1257786155,0.0324165039,0.3559170365,0.0472595692,0.1438392997,0.2195971012,-0.015439433,0.1111363098,-0.2860589921,0.2297817022,0.1622222811,0.320484072,-0.4211920202,-0.1435187608,-0.1698827595,0.0571510829,-0.3703622818,0.3384325504,0.0452004038,-0.3210283518,0.219812423,0.2442415953,-0.0520663597,-0.1897224188,0.1275731176,0.2060876638,0.3765225112,-0.102869004,0.0157828555,-0.0386532173,-0.1966774017,0.0999093354,-0.2033937275,0.0466274358,0.2891245186,-0.3035188317,-0.0678539053,0.0391921364,-0.1798283607,0.1154339984,-0.2513130009,-0.0251798052,0.0710902959,0.002385335,0.009314036,0.0932091624,0.1263341308,0.2573832273,0.2938033938,0.3097373545,0.2206825316,-0.0685122982,0.0444711931,0.0038334976,-0.0998946726,0.0476477519,0.1706878841,0.2578557134,-0.1094403118,-0.3618175387,0.0034739668,-0.0324972756,0.0244196821,-0.00028303,0.3031305075,0.1252112687,-0.1238038316,0.1699262112,0.2675572634,-0.1435555369,-0.324788481,-0.1181280389,0.3308382332,-0.1890414208,-0.1475578994,0.2492350042,-0.1676839888,-0.0145412395,-0.0785941333,0.1480413526,-0.0763944089,-0.1346729398,-0.1982849091,-0.1027288511,0.1339935511,-0.0330300145,0.0655056462,0.3730125427,-0.528439343,-0.1443895847,0.054698877,-0.0276944656,-0.2500066161,-0.0474639162,0.2256289274,0.0221192595,0.0573181361,-0.3491019309,0.4549744427,0.2764905989,-0.0304751769,-0.0142187122,-0.2228385508,-0.0456286781,-0.1861785799,0.1986172944,0.5018196702,-0.3949016333,-0.0255267676,-0.295882225,-0.0158854276,0.1709475815,0.3286664486,-0.1296211034,0.3348804116,-0.3377097845,0.2680733502,0.3865487278,-0.3286767006,-0.265244931,0.0421010889,-0.0843798369,-0.016684752,0.1469514221,-0.0085748956,0.1543220282,0.0568281338,0.212637499,0.199009344,0.0098339142,0.0903028995,-0.2157630771,0.0875527486,0.2370745391,0.1217467189,0.2510007024,-0.0429093987,0.1854168773,0.1080474779,0.2132850289,-0.3157053888,0.3144257963,0.0235330593,0.2331555039,-0.064878948,-0.0976154953,-0.3182422221,-0.034065105,0.1612092853,-0.0904948041,0.3825849593,-0.5090478659,-0.1393896788,-0.4510042071,0.0097899446,-0.4143909812,-0.2540181875,0.2000551075,0.0524621159,0.1434838027,0.1402629018,-0.1512691677,0.5673833489,-0.3430083096,0.0166805387,-0.0704813972,0.3489124179,-0.3280907273,-0.2981008291,0.2780768275,0.2281981111,0.0782784671,-0.1092452332,-0.138141498,0.286190182,-0.253243506,0.002743836,0.1318633556,-0.2189347744,-0.0235489141,-0.4651159942,0.1460891962,0.1815542579,0.2387862951,-0.1993078142,0.0126365162,0.4493567646,0.323233515,0.1039566621,-0.0041023707,0.0095627103,0.4009858966,0.0525740013,-0.0183920767,-0.198618561,0.2568436861,-0.0751996413,0.2822484672,-0.2066947073,-0.2829215825,0.0135723483,0.3775028884,0.0491406284,0.4471918046,0.1533788741,-0.1032971516,-0.0195986778,-0.1601633728,0.1207781211,0.3303133547,0.0281339604,-0.0462334566,0.2144807875,-0.1015012562,-0.184566617,0.2178310454,-0.0088697011,0.0528022572,0.2028618306,0.1282225102,-0.0992851406,-0.3807548583,-0.2188048363,-0.0102537293,0.3387620449,-0.2657980621,0.0051065776,-0.127150625,-0.3741125762,-0.1996125132,0.2105915397,-0.1274860203,-0.2177482992,0.0389078073,-0.0407144018,0.0290992465,0.1327520162,0.0913390741,0.0010963383,0.1120714322,-0.1972098649,-0.0206984226,-0.2795613706,-0.1944819093,0.1158320531,0.3588798046,0.1520616114,0.4850715101,-0.3050047457,0.0137912193,-0.1548298895,-0.4496572316,-0.0058443835,-0.3447466791,0.0721900761,0.0923592448,0.2036583275,-0.0428540371,-0.459580332,0.3144107461,-0.0724989846,-0.3400882185,0.1471819133,-0.0350828357,-0.0772237629,-0.0065320386,-0.7115244865,-0.302619487,-0.375590235,-0.0545475073,-0.1178930029,0.1017842889,0.0756666735,0.2219262719,0.233664915,0.0522240996,0.1142273992,-0.2015258968,-0.2000739127,0.3361239731,-0.3108819127,-0.3760727942,0.0944502652,-0.1340978891,0.3507673144,-0.2822908759,-0.3692141473,0.20770441,-0.2674362361,0.4491765797,0.0926628783,-0.1399540901,0.3198357522,0.0110754361,-0.0922598988,0.0202658288,-0.221143201,0.1299603432,-0.0406076089,0.3319708109,0.0096705416,0.47392717,0.0118622035,0.623893857,0.2154695541,0.0228207409,0.4156720936,0.1736146361,0.26692608,-0.3062385917,-0.4208293557,0.2276843935,-0.0628337711,0.119622983,0.0323038623,-0.1091859862,-0.2895950675,-0.350995928,-0.3293344378,-0.2467863113,-0.3040595651,0.1780607104,-0.1339372247,0.3296235204,-0.1510362029,-0.1977496296,-0.1477027237,0.0920325071,-0.0135462917,0.3433564901,-0.3183005452,0.088877663,-0.3328643143,-0.2641271651,-0.4069662988,0.2210445553,0.1843544841,0.0799063742,-0.3019348383,0.1988075078,-0.0035238955,-0.0613079406,0.1738230288,-0.1815279722,0.207480669,0.2146141827,0.0727858245,-0.0819840953,0.0115695866,-0.3653075397,0.1526092142,0.2256942093,0.1723292321,-0.4129521251,0.0829494223,0.0229442474,0.1335794032,-0.1013845727,-0.0803679898,-0.108610779,-0.3618460596,-0.53320086,0.2236565799,0.3853721619,0.4265888929,-0.0034150661,0.2268608212,0.1188860461,-0.0876852274,0.1869895756,0.1162846982,0.0482704081,0.1831129342,0.1843984425,0.0821003169,0.0087144561,-0.0682033077,0.6501699686,0.2422444224,-0.5056651235,-0.1315882057,-0.2369926721,0.1060508266,0.1945034713,-0.0631324947,0.0218656808,-0.0195111893,-0.0468131267,0.071032621,0.2546421587,0.2530525327,0.3129629493,-0.424028933,-0.3952422738,0.0929747894,-0.2013558447,-0.0196746048,-0.0248848554,0.252009958,-0.425501585,0.2545580566,-0.0045936257,0.9560571909,0.0347711518,0.2339397669,0.1494144201,-0.1839836389,0.1841397583,0.0286606662,-0.059349604,-0.1531006694,0.1560572982,-0.0298918523,-0.0247323941,0.2963774502,-0.1079888791,-0.3020828664,0.0964457616,-0.0278940573,-0.1122333705,-0.0903785229,0.2975540459,-0.141435802,-0.265757978,-0.5213631988,0.1996451467,-0.0886174217,0.1661653817,0.0547117069,-0.1375193149,-0.1689305604,-0.3871940672,-0.5968745947,-0.0085987905,-0.087220706,0.2162306309,-0.163077116,-0.4250732064,0.5653030276,0.3926413953,0.6571986079,0.1344224215,-0.117561616,0.1307078898,-0.18848598,-0.2416857928,0.1248124912,0.0581636094,0.3777551651,-0.2391423881,-0.3237068951,0.2999587953,-0.0002486202,-0.1739885956,-0.304099977,-0.1643059552,-0.094383046,-0.2417110056,-0.1273347288,-0.0950589105,-0.2003800571,-0.3413677812,0.1872369647,0.1308596432,-0.2187298238,0.1325452626,-0.0359440967,-0.2003850341,-0.0723344162,0.3325362206,0.2005321085,0.0265499782,0.7404000759,0.2742212713,-0.2460490167,-0.268645674,0.0441007763,0.2170727849,-0.1088809669,0.1044075266,0.0407160185,-0.1658088416,-0.1185244694,0.5871521831,0.1434551924,-0.0834691599,0.0892258957,-0.3736430705,-0.5136302114,0.1403672695,0.0344948918,0.0711072758,0.160532698,0.5175585747,0.1939789206,0.259758383,-0.3545671999,-0.0199799687,-0.1265084743,0.2201581299,0.256428957,-0.1030636728,0.1695173532,0.0986917764,0.1692528874,0.2568180263,-0.0744998753,-0.2562862933,-0.0297071077,0.0684918836,0.060977336,-0.0182958432,0.0627055243,-0.4180742204,0.0465571806,-0.0198126025,0.0902304277,0.1790376902,-0.1040823013,0.2523134947,-0.0954293832,0.1027104855,0.2190435678,-0.0049355277,-0.222917825,-0.031602677,0.1160967052,-0.0118865352,-0.0080475342,-0.1322987974,-0.0837075412,0.0394298024,-0.1864338517,0.1457431465,0.4332692623,-0.3386834264,0.0652096644,0.0485561118,0.2000084072,0.0899415761,-0.3356072307,0.2155048102,0.1496501267,0.1620500386,-0.5012565851,-0.0260765236,-0.1048647538,-0.1312339455,0.1752729416,0.0897104964,-0.0530102439,0.1718425751,-0.0669194758,0.0415269621,0.2185550034,-0.1688288152,-0.018157078,0.1704832166,-0.0814407617,-0.0229545627,0.396853596,0.1706307679,0.0817142576,0.754280448,-0.4638420343,0.2728507221,0.0283806231,0.0798992813,0.0751684383,-0.045380421,0.2522485852,0.0665887371,-0.157911703,0.1691103131,0.0004675018,0.2341300398,-0.1701087952,0.10330531,0.0690735728,0.030340312,-0.1861152053,0.0451406166,-0.2474006712,-0.1647726893,-0.0838667229,-0.1383020133,-0.1634885669,-0.0046158647,0.31667611,0.0317494161,-0.164608717,-0.3379205763,0.1691758186,0.0094465986,0.203090325,-0.2248623073,0.2859092653,0.4393926561,0.0068217875,0.2072544843,0.272285223,0.584860146,0.2557890415,-0.3220208883,0.2581300437,-0.1916150004,-0.3510195613,0.0153380474,0.208209008,0.3283077776,0.4378299415,0.2944050133,0.2015208602,-0.1613500714,-0.0089410357,0.2797061205,0.0970080495,-0.0775780678,0.1749489009,-0.2106502205,-0.1400338113,-0.276357621,0.0291909594,-0.4062030315,-0.0574478693,0.2399069965,0.0581412055,0.1419747919,-0.255405277,0.1037548557,-0.0013099245,0.4438290298,0.2324422449,0.0214441083,-0.427441299,-0.2046918124,-0.7065240145,-0.0420197882,-0.0805960074,-0.2868054807,-0.2410641909,0.0310068224,0.389321804,0.104071185,0.182817325,-0.0337896831,-0.2051883489,-0.0601090416,-0.2338097095,-0.2392845601,0.0388932489,0.1976132393,0.0517265648,-0.0728837699,0.1848481596,0.4200256765,0.0988643393,-0.1860603541,0.0358880647,0.074406974,-0.2081200182,0.0546483062,0.1556433737,0.6450904608,-0.1854212284,0.0329485387,-0.0414810479,-0.0796917751,-0.1384912729,0.4796761572,0.2791798711,0.6704034805,-0.2028396279,-0.031105075,-0.0715384558,0.4367156923,-0.0289897583,-0.3920423985,-0.2773582935,0.2260175943,0.1188498661,0.155267939,-0.1216093674,0.3387446105,-0.0839007869,0.0555880181,-0.3392180502,-0.5602796674,0.3288574815,-0.1307483763,-0.4304898083,0.1566852182,0.0662876815,0.3002919555,0.13716425,-0.7134615183,0.2555774152,0.2097670287,-0.0324450843,-0.1433790028,0.1805782616,0.0106321322,0.1486517489,-0.048737064,0.4346379042,-0.035497725,-0.076419875,-0.0083866827,-0.3005461991]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2825","title":"The datasets.map function does not load cached dataset after moving python script","comments":"This also happened to me on COLAB.\r\nDetails:\r\nI ran the `run_mlm.py` in two different notebooks. \r\nIn the first notebook, I do tokenization since I can get 4 CPU cores without any GPUs, and save the cache into a folder which I copy to drive.\r\nIn the second notebook, I copy the cache folder from drive and re-run the run_mlm.py script (this time I uncomment the trainer code which happens after the tokenization)\r\n\r\nNote: I didn't change anything in the arguments, not even the preprocessing_num_workers\r\n ","body":"## Describe the bug\r\nThe datasets.map function caches the processed data to a certain directory. When the map function is called another time with totally the same parameters, the cached data are supposed to be reloaded instead of re-processing. However, it doesn't reuse cached data sometimes. I use the common data processing in different tasks, the datasets are processed again, the only difference is that I run them in different files.\r\n\r\n## Steps to reproduce the bug\r\nJust run the following codes in different .py files.\r\n```python\r\nif __name__ == '__main__':\r\n from datasets import load_dataset\r\n from transformers import AutoTokenizer\r\n raw_datasets = load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\n\r\n def tokenize_function(examples):\r\n return tokenizer(examples[\"text\"], padding=\"max_length\", truncation=True)\r\n\r\n\r\n tokenized_datasets = raw_datasets.map(tokenize_function, batched=True)\r\n```\r\n\r\n## Expected results\r\nThe map function should reload data in the second or any later runs.\r\n\r\n## Actual results\r\nThe processing happens in each run.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: linux\r\n- Python version: 3.7.6\r\n- PyArrow version: 3.0.0\r\n\r\nThis is the first time I report a bug. If there is any problem or confusing description, please let me know \ud83d\ude04.\r\n","comment_length":85,"text":"The datasets.map function does not load cached dataset after moving python script\n## Describe the bug\r\nThe datasets.map function caches the processed data to a certain directory. When the map function is called another time with totally the same parameters, the cached data are supposed to be reloaded instead of re-processing. However, it doesn't reuse cached data sometimes. I use the common data processing in different tasks, the datasets are processed again, the only difference is that I run them in different files.\r\n\r\n## Steps to reproduce the bug\r\nJust run the following codes in different .py files.\r\n```python\r\nif __name__ == '__main__':\r\n from datasets import load_dataset\r\n from transformers import AutoTokenizer\r\n raw_datasets = load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\n\r\n def tokenize_function(examples):\r\n return tokenizer(examples[\"text\"], padding=\"max_length\", truncation=True)\r\n\r\n\r\n tokenized_datasets = raw_datasets.map(tokenize_function, batched=True)\r\n```\r\n\r\n## Expected results\r\nThe map function should reload data in the second or any later runs.\r\n\r\n## Actual results\r\nThe processing happens in each run.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: linux\r\n- Python version: 3.7.6\r\n- PyArrow version: 3.0.0\r\n\r\nThis is the first time I report a bug. If there is any problem or confusing description, please let me know \ud83d\ude04.\r\n\nThis also happened to me on COLAB.\r\nDetails:\r\nI ran the `run_mlm.py` in two different notebooks. \r\nIn the first notebook, I do tokenization since I can get 4 CPU cores without any GPUs, and save the cache into a folder which I copy to drive.\r\nIn the second notebook, I copy the cache folder from drive and re-run the run_mlm.py script (this time I uncomment the trainer code which happens after the tokenization)\r\n\r\nNote: I didn't change anything in the arguments, not even the preprocessing_num_workers\r\n ","embeddings":[-0.1612097174,0.0666727647,0.1235988811,0.1421179473,0.121106185,-0.0925723389,0.4181308448,0.1090156436,0.1731479019,-0.0780501887,0.1154462621,0.6156936288,-0.0211400185,-0.1331947744,0.1611653715,0.2321514785,0.1759635508,0.1501921713,0.0869101956,-0.1811600775,-0.1934700161,0.0917448997,-0.4173083007,-0.1069656685,-0.4597243071,0.0247058906,-0.1255613714,-0.0259201769,0.230769515,-0.3739830256,0.3416953087,0.0286043081,0.3293195367,0.4340651929,-0.0001240348,-0.0629330575,0.1899118572,-0.119130753,-0.2047812641,-0.2416981906,-0.0412899964,-0.063502565,0.1722025722,-0.1777267605,0.0049677868,0.1066701785,-0.0636675432,-0.5965551734,0.3043339252,0.2677555978,0.0943721831,0.0372929089,-0.2589368224,0.2234524488,-0.0972795263,-0.1032885909,-0.0847678334,0.035640616,0.2718014717,-0.4042436481,-0.1610348374,0.3430397511,-0.2559062541,0.1986912936,0.3800358176,0.0639560893,-0.1315136701,-0.3624012768,0.3555093408,-0.1061243713,0.282738626,-0.4287838936,-0.1180434525,-0.337542206,-0.1085550115,-0.1223061308,0.2084060758,-0.1288083941,0.1678153425,0.1231883615,-0.3631304502,-0.0782385394,0.1985265017,0.2072054893,-0.1494708955,0.4666341841,-0.1864189059,0.2081397772,-0.1478581727,0.0597921051,0.2998939157,-0.1723808944,0.0578929037,0.4553029537,-0.1857805848,0.00498839,0.1042604446,0.0164342467,-0.0691987127,-0.0429935753,0.099240154,0.0713901743,-0.0783548951,0.0563075989,0.1188397035,0.4971541762,0.0955592021,0.3987930715,0.1923970729,-0.0712575316,-0.7449443936,-0.1728950292,-0.1321693659,0.0351381004,0.4841654003,0.1077986285,0.1735487729,-0.0119950669,0.0163378939,0.0511335172,-0.2896785736,-0.2111107856,0.1056540087,0.4156179726,-0.0801942125,0.153526172,-0.0949861333,0.143733874,-0.1945284456,0.0410147682,-0.1212418005,-0.1458109915,-0.2652318776,0.3765378296,0.2610456347,0.1239273027,0.4008917511,0.1099497601,-0.0333948508,-0.2456994355,0.3013131022,-0.0583071299,0.5958386064,0.162357986,-0.0118661886,0.5882652402,0.2430468202,0.0027757902,-0.1312852502,0.2340513617,-0.6722649932,-0.0946235657,0.4809675515,-0.0038467753,-0.1870749593,-0.0904574916,-0.2192808092,0.058336135,0.5520724058,-0.2391305864,0.0510410778,-0.337028563,-0.370714277,-0.4002167881,0.1697919369,0.528999269,-0.090806447,-0.1419551224,-0.2303371578,0.4682983458,0.2079978734,0.1423209459,-0.3177858889,0.2858117521,-0.3038775325,0.3247823119,0.2061853856,-0.400685817,-0.6898708344,0.0444125161,-0.018911235,0.2681943774,-0.2062691599,-0.0692022368,0.2436883897,-0.1089235768,0.0306177288,0.2620004714,0.1978547126,0.3179399371,-0.2945934534,-0.0697201416,0.2854961157,-0.239878878,0.1903080344,0.0873956457,0.0487780645,0.0684907213,0.1300728172,0.0050785085,0.1377951652,0.0802836195,0.1884408891,-0.0057985964,0.2942882478,-0.0042526033,-0.261926651,0.2479481548,-0.2068691701,-0.4193128049,-0.1266342551,-0.0922151729,-0.0158471409,-0.1833045632,-0.4220011532,-0.3649426699,0.0538757145,0.2663186193,0.1915195435,0.1526272893,-0.1701246202,0.2396815121,0.2925300002,-0.1287483722,-0.0264476351,0.2079433352,0.172969386,-0.1440100074,-0.2889925539,0.1064643487,0.4621390104,-0.13235946,-0.2746540904,0.1695510745,0.288733989,0.1612867117,-0.2219236195,0.0296782088,0.1039761752,0.0752551183,0.2001729906,0.2023634762,0.1230779141,-0.0977190658,0.1575126499,0.3327405155,0.1918524504,0.1902764589,-0.0139093883,-0.0960238799,0.0941475928,-0.1803358346,0.1143564135,-0.2381325364,-0.1593662947,-0.0464133397,0.4003399611,0.2147738487,0.2689982355,0.3867591918,0.6588147879,0.2824723125,0.1987578273,-0.0951084644,-0.705016911,-0.2481067032,0.0504059009,0.2260113806,0.361297518,-0.0615942776,-0.0229833983,0.0431727059,0.1944842339,-0.0213191751,0.0175201781,-0.0430983566,0.2224767506,0.2194302827,0.2184744775,0.2063397467,-0.1053415611,0.1863428652,0.2211709321,0.2342161089,-0.1675949842,0.1337196678,-0.4766334891,0.2224489599,-0.1026736647,-0.1299252659,-0.0875684917,-0.3107097149,-0.1644132733,0.4240451753,0.199869737,0.2177815288,0.2266703546,0.054220736,-0.0503890812,-0.286282748,-0.2200453281,-0.3234049678,-0.2892321944,-0.1396640241,0.2144162953,-0.0832653791,-0.001452321,0.0057454226,0.0393406712,-0.3005688488,-0.311855644,0.0178975463,0.1531227678,0.0891461894,-0.3017293513,-0.0930188298,-0.3721902966,-0.0375695117,0.1113606691,-0.4386264682,-0.048365064,0.0744518787,0.0122522805,-0.0959481075,0.1013624594,-0.223274827,-0.257391423,0.0301529076,0.077672258,-0.0911214426,-0.0248405728,-0.0852264613,-0.0929128602,-0.094437249,-0.178216517,-0.171350956,-0.3487738669,-0.2494967878,0.3564105034,-0.2292416245,-0.1970822513,-0.0286220517,-0.0213072728,0.2215861082,0.2303259373,-0.4229950011,-0.1399636716,-0.0464505367,0.1276432127,0.040781498,0.1231173649,0.4344370961,0.082470797,0.1826771349,-0.2157482803,-0.2161741704,0.0058982498,0.2345543951,0.092919603,0.1258677691,0.1468118429,-0.0391458198,0.8137372732,0.3329548538,-0.2891389728,0.3945438266,-0.0771217495,0.199116677,-0.2179757506,-0.3984623551,-0.0565709285,-0.1827714294,-0.190834552,0.1693387181,-0.0378273651,-0.2157123536,-0.0502852462,-0.1973913163,-0.3199198544,-0.2977833748,0.1104878336,-0.1729243994,0.3215156496,0.3823325634,0.323205471,-0.5896300077,-0.1851330847,0.1468659788,0.0718662813,0.377453059,0.0667560101,-0.2743141055,-0.2742203176,-0.2854088247,0.2252298445,0.0543754138,0.456194967,0.0394051336,-0.1448571831,0.0027040169,0.0972270295,0.2747096717,-0.4058230221,0.0144489277,0.2318762243,-0.169030115,-0.0841191635,-0.1644562036,0.0765198544,0.4333055317,0.196318686,0.4632546008,0.2221888453,0.0458111949,-0.3195922375,0.3016912341,-0.2415540814,-0.0776091293,-0.3578511775,0.035837926,-0.4233498871,-0.053378325,-0.0722151697,-0.1255103201,0.0304971002,-0.1190648898,-0.2477416843,-0.1286759675,0.0159275848,0.2879603207,0.4373827875,0.1520212591,0.2777386904,0.2438932806,-0.0925457403,0.2242038101,0.2301967591,-0.1936238706,-0.1622547805,-0.0704884082,0.0264500119,0.258575052,0.0949298218,-0.3744323552,-0.0717406645,-0.1470507532,-0.1333134174,-0.1991483122,0.0390074551,0.1198321804,0.2040016651,-0.2184669226,-0.4964649975,0.1315729022,0.114811644,-0.2241777331,0.478428632,-0.526822865,-0.2337674201,0.4640763104,-0.1589085609,0.6939270496,-0.2030059248,-0.0131727364,-0.1569578201,0.2673536241,0.317594707,-0.2601320446,0.1933846027,-0.4079014659,-0.0705298185,0.0176699478,-0.2399065495,-0.0603284501,0.0622503869,-0.3646001816,0.2890155017,0.0456750728,0.3356089294,-0.0115410462,0.2104549557,0.2857830822,-0.1372208595,-0.0568747781,0.0819549635,0.0685290098,0.4716668427,-0.0036627227,0.0043436391,0.0710884109,-0.0449547209,-0.0913441032,0.1123438329,-0.7211812139,0.3068857491,-0.1747056991,0.0336415544,-0.0516440906,0.2431949675,0.3510490656,0.1937649846,-0.128208071,0.1868287325,0.1930041909,0.2486364394,-0.0477928445,-0.3190559447,0.1430826783,0.1476634443,-0.1788543612,-0.0234667622,-0.0624391437,-0.3910655677,-0.253793627,0.2718144655,0.4700975716,-0.1907345951,-0.1381174773,0.1859202832,-0.02691615,-0.1034908593,0.027351575,0.1545143276,0.0978820845,0.5066225529,-0.3040743768,-0.3529683053,0.0917914584,0.540384233,0.1394846886,-0.0734277144,0.3789063692,-0.2389588505,-0.0740443394,-0.1299583018,0.0430366248,-0.0469727926,-0.2355395705,0.2335426658,-0.0914036706,0.1868425012,0.0900250375,-0.0355315544,-0.2137664855,0.0162534006,-0.3044367433,-0.3737491667,-0.3308830261,-0.0863465518,0.0511150062,-0.0865011737,0.4272451699,0.2910868227,0.0601286106,0.2189546973,-0.189519003,0.2773419917,0.3314101696,0.1650745422,-0.1432035267,-0.0637516454,0.0176947471,-0.1543926746,-0.0480545722,0.1709011048,-0.0499137975,-0.1203969643,-0.1994163394,0.2219584435,0.1711622924,-0.0217314493,-0.2919609547,-0.3067833781,0.1630872488,-0.5828201771,0.1274298429,0.2063224465,-0.0932626054,0.0095130429,0.3784586191,-0.198389411,-0.069982864,0.0737883598,-0.3343176246,-0.1535220146,0.0062807426,0.2818330228,0.1830986142,0.190015465,-0.0367168598,-0.2239575982,0.1002553105,0.0655635223,0.3466318846,-0.2847490311,-0.1031518281,0.0366776139,-0.2593297362,0.2991609871,-0.1855588406,-0.2341835201,0.287458092,0.0721718073,-0.2727124691,-0.1777084321,0.4632629454,0.1624135226,-0.0538191088,0.2088624835,0.3971565366,-0.1730994284,0.2225059867,0.0720846951,0.6409052014,-0.2101064175,0.3045118749,0.0442039669,0.0877536312,0.2990761101,0.2292415798,0.2153177857,-0.1304768324,0.2860995829,0.0502045415,0.2895384133,0.2598907053,-0.0994988829,-0.0054142154,-0.4730812907,0.0157533865,0.1380017549,-0.3188564479,0.398581028,-0.0782034919,0.0821984038,-0.0805381685,-0.1319543421,-0.02026961,0.2977826297,-0.360398531,-0.1637579501,0.1918500811,-0.0986475274,0.0516189039,0.0529705361,0.0217677001,0.09859211,0.4778100848,0.082221359,-0.3910200596,-0.0628982931,-0.3392602205,0.1961570084,0.4229792953,-0.1809888631,0.2635442615,-0.3162927628,0.0390433483,-0.1846109927,0.109095268,0.3859155178,0.4559960067,-0.2092548907,0.2142004073,0.2236176729,0.0662406757,-0.1414040178,0.3060624897,-0.1343111098,-0.082782425,-0.0785425678,0.020350147,-0.0050411136,-0.0961627066,-0.0192351025,0.3018614948,-0.0153999981,0.0497265682,0.0164831262,-0.0332654119,-0.1139935106,0.2352301776,-0.2441951185,0.2029223144,0.5567872524,-0.0086038653,0.1496214271,-0.3217227757,0.034883704,0.1219478175,0.2171915621,0.3149913251,-0.2996428907,-0.4097650945,-0.2676977217,-0.7540673018,0.3354357779,-0.1838145107,0.2210133523,-0.0587712973,0.08054021,-0.0903601348,-0.1343165785,0.165517807,-0.0526569895,0.0231831577,0.1846774369,-0.2335088551,-0.1410393119,-0.3344892561,0.08516711,0.1468009502,-0.3403093219,0.2742243111,-0.1483900845,0.0004372018,-0.2321799695,-0.1087325513,0.0131375305,0.0082705924,0.8180852532,0.1553950757,0.1773556471,-0.0341654532,0.1784492582,-0.2499229163,0.0768158585,-0.0369193144,0.0287617687,-0.0001644013,0.0974197388,-0.1688402146,-0.1767754704,-0.504983604,-0.1605619043,0.2636039853,-0.1114461645,-0.0918082595,0.0327226706,-0.0583393164,0.3533480167,-0.0920662284,0.4632125497,-0.1938390583,0.3721648753,-0.2800157368,-0.1676999778,0.3528719842,-0.6539241076,-0.445569247,-0.2375842035,0.3042317927,-0.2936923802,-0.2092626095,-0.5068480968,-0.0203976296,0.444563359,-0.0114233419,-0.2154309452,0.0537237488,-0.0901708454,0.2256145328,-0.2515746653,0.2279302627,0.2003179044,-0.2147218585,0.0753315911,-0.1473741829]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2825","title":"The datasets.map function does not load cached dataset after moving python script","comments":"Thanks for reporting ! This is indeed a bug, I'm looking into it","body":"## Describe the bug\r\nThe datasets.map function caches the processed data to a certain directory. When the map function is called another time with totally the same parameters, the cached data are supposed to be reloaded instead of re-processing. However, it doesn't reuse cached data sometimes. I use the common data processing in different tasks, the datasets are processed again, the only difference is that I run them in different files.\r\n\r\n## Steps to reproduce the bug\r\nJust run the following codes in different .py files.\r\n```python\r\nif __name__ == '__main__':\r\n from datasets import load_dataset\r\n from transformers import AutoTokenizer\r\n raw_datasets = load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\n\r\n def tokenize_function(examples):\r\n return tokenizer(examples[\"text\"], padding=\"max_length\", truncation=True)\r\n\r\n\r\n tokenized_datasets = raw_datasets.map(tokenize_function, batched=True)\r\n```\r\n\r\n## Expected results\r\nThe map function should reload data in the second or any later runs.\r\n\r\n## Actual results\r\nThe processing happens in each run.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: linux\r\n- Python version: 3.7.6\r\n- PyArrow version: 3.0.0\r\n\r\nThis is the first time I report a bug. If there is any problem or confusing description, please let me know \ud83d\ude04.\r\n","comment_length":13,"text":"The datasets.map function does not load cached dataset after moving python script\n## Describe the bug\r\nThe datasets.map function caches the processed data to a certain directory. When the map function is called another time with totally the same parameters, the cached data are supposed to be reloaded instead of re-processing. However, it doesn't reuse cached data sometimes. I use the common data processing in different tasks, the datasets are processed again, the only difference is that I run them in different files.\r\n\r\n## Steps to reproduce the bug\r\nJust run the following codes in different .py files.\r\n```python\r\nif __name__ == '__main__':\r\n from datasets import load_dataset\r\n from transformers import AutoTokenizer\r\n raw_datasets = load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\n\r\n def tokenize_function(examples):\r\n return tokenizer(examples[\"text\"], padding=\"max_length\", truncation=True)\r\n\r\n\r\n tokenized_datasets = raw_datasets.map(tokenize_function, batched=True)\r\n```\r\n\r\n## Expected results\r\nThe map function should reload data in the second or any later runs.\r\n\r\n## Actual results\r\nThe processing happens in each run.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: linux\r\n- Python version: 3.7.6\r\n- PyArrow version: 3.0.0\r\n\r\nThis is the first time I report a bug. If there is any problem or confusing description, please let me know \ud83d\ude04.\r\n\nThanks for reporting ! This is indeed a bug, I'm looking into it","embeddings":[-0.0928246528,0.0353672095,0.0930575877,0.0993780345,0.1389549971,-0.0688583553,0.4122239053,0.1713828295,0.2286131829,-0.1237098053,0.1240525097,0.5175410509,-0.0142520489,-0.1841217875,0.1500011832,0.1871795505,0.2208417654,0.1862142682,0.0343401097,-0.2101508528,-0.1906058341,0.0074796495,-0.3999150991,-0.0556141362,-0.3895706534,0.1418250501,-0.144321546,-0.0767576918,0.2297321707,-0.4079437852,0.3964934945,0.0969622955,0.2728181481,0.4235104024,-0.0001219823,-0.0338355266,0.2292539924,-0.1005551964,-0.1314365715,-0.2216355652,-0.0886393413,-0.0659393072,0.1086358428,-0.1759864837,0.0406188443,-0.0302871969,-0.0742248744,-0.5787385702,0.3091609776,0.2556922138,0.115243569,0.0504508354,-0.2082477063,0.1805128753,-0.1063698232,-0.095673725,-0.0220883805,-0.1044369861,0.2900579572,-0.3806003332,-0.1846182644,0.2567872107,-0.1888342202,0.105700098,0.3405888379,0.016221758,-0.1603529602,-0.3042694032,0.2797772586,-0.0931538567,0.3588237464,-0.4208958745,-0.2226949483,-0.2972559631,-0.1442857087,-0.0071968995,0.1576484293,-0.1826785803,0.2009661943,0.1336780041,-0.4482117295,-0.1309143901,0.1029749811,0.2108209878,-0.0913619697,0.3717657328,-0.1069849804,0.22270827,-0.1753474623,0.034918651,0.1962803602,-0.1709039807,0.0531557202,0.5050052404,-0.2086565346,-0.0022408273,0.133885324,-0.0162170604,0.057870362,0.0831476524,0.0329411179,0.1465694159,-0.1785293818,-0.0176028628,0.2223400772,0.4795157015,0.1250340939,0.4396397173,0.1034809649,-0.1098446921,-0.7153984308,-0.1525016129,-0.1062359512,-0.0545984544,0.6554469466,0.0222750884,0.2374232262,-0.0149531895,0.1160370484,0.1094723716,-0.1231237873,-0.2185614854,0.0279546063,0.4383969605,-0.0606766269,0.2895187736,-0.0927915499,0.1993542016,-0.1540943086,0.0469528325,-0.1616937369,-0.2420498431,-0.2045531124,0.2962532341,0.1838106513,0.0810131282,0.3403462768,0.1439111829,-0.0440091006,-0.2238905132,0.2629828453,0.0289249718,0.6173426509,0.1202462316,-0.0526015833,0.5951809287,0.2461516112,0.0445202515,-0.1578405946,0.2223032713,-0.7051262856,-0.1446986645,0.4036145508,0.0342218243,-0.1923199892,-0.1066038385,-0.1792516261,0.0036804196,0.5243843794,-0.255482465,0.0571590662,-0.2893354595,-0.3743081689,-0.4061015844,0.1506380886,0.4997980595,-0.004401274,-0.1656034142,-0.1477142572,0.3303782642,0.1386244595,0.1156267375,-0.3449488878,0.3154698312,-0.2998971343,0.3331964016,0.3271439373,-0.4190366268,-0.7177855968,0.0751584247,-0.0045209173,0.2787913978,-0.2330899686,-0.0558174178,0.222856611,-0.1392142922,-0.0994028077,0.2826619148,0.2052944899,0.3346381783,-0.2289548218,-0.1065657213,0.34298262,-0.1810037345,0.1750920266,0.0812563822,-0.0000550816,0.0480283722,0.1802407056,0.0236316677,0.1945720464,0.1927071214,0.1390551031,0.0986294448,0.3337622285,-0.0595292971,-0.2820928693,0.2654383779,-0.1981913596,-0.5161175132,-0.0475431494,-0.0765885189,-0.0904142484,-0.1833398938,-0.5219665766,-0.2661530972,0.0725814924,0.2676863968,0.2641972005,0.1398716271,-0.1747345477,0.1988597512,0.3232674003,-0.1634768546,-0.08758834,0.1838061363,0.2016696483,-0.0720206052,-0.3083678484,0.0649294779,0.4533135295,-0.1110672653,-0.2810949087,0.2153314203,0.2526636124,0.2492101192,-0.1198031604,-0.0196989737,0.0365078375,-0.023098018,0.1569166481,0.2410008609,0.1003436819,-0.1328220069,0.1561343968,0.202255398,0.175908193,0.2006759048,-0.0164723005,-0.0874907598,0.0911512971,-0.1904923022,0.094818674,-0.2022498399,-0.1780011952,-0.1641681343,0.4223390222,0.1883176118,0.3436411023,0.2869781256,0.6627221704,0.1788884997,0.1711745858,0.018564973,-0.6508919597,-0.2396672517,-0.1064246148,0.1712076366,0.3237538934,-0.0569468886,-0.0800732747,-0.0462446921,0.209755376,0.0706920475,0.0392953083,-0.0379815251,0.175695613,0.2264041007,0.2048121095,0.177065134,-0.0670218915,0.2043900788,0.2082093656,0.1890830249,-0.1625415087,0.0767255127,-0.3909190297,0.3666087687,-0.068560034,-0.1921299845,-0.0622236952,-0.2952159643,-0.1365936697,0.3060539365,0.0991413817,0.2357823402,0.0930634439,-0.0500938073,0.0003513621,-0.4086705148,-0.2552194595,-0.370323807,-0.2227658629,-0.1258556694,0.1664728522,-0.1475484967,0.0871744007,-0.0013492005,0.0629387572,-0.3593187034,-0.240680024,-0.0698246136,0.1407874376,-0.0132178161,-0.3100545108,-0.0072165597,-0.2632973492,-0.0663152039,0.1533186883,-0.5259952545,-0.0533175804,0.188634634,0.0279808939,-0.0374453999,0.1455306262,-0.3022510707,-0.2096186131,0.0403193571,0.1105431318,-0.0543356463,-0.039356567,-0.0802846774,-0.0696940273,-0.1704852581,-0.2920866609,-0.1377395093,-0.3719044924,-0.3158812821,0.3483986259,-0.193798542,-0.2463772446,-0.1309868097,-0.0301985908,0.1985356957,0.2263653427,-0.3835794032,-0.0527390763,-0.127749294,0.1799549609,-0.0229097921,0.158635512,0.4902326763,0.0729219094,0.1569582522,-0.2417710125,-0.0768942535,0.0610387623,0.2066959739,0.0173275005,0.124157533,0.0612797588,-0.0731715262,0.8447428346,0.3750460744,-0.2395498455,0.4339180589,-0.1242032573,0.1559933573,-0.2364548296,-0.3758387566,-0.1434130967,-0.2178285122,-0.1108889133,0.1275508404,0.0304627232,-0.2130716294,-0.018781634,-0.2146094143,-0.4178936779,-0.2764394283,0.0530615374,-0.1290223897,0.2872948945,0.3051960766,0.2294481695,-0.6220389605,-0.2146760374,0.2077193111,-0.0329644345,0.4447860122,0.0009736089,-0.2783771753,-0.1507144272,-0.2242325246,0.1996689737,0.0721212775,0.4041115344,0.0837482214,-0.189480111,-0.0134108756,0.1144384742,0.2869142294,-0.3712111115,0.0903801918,0.2023197263,-0.1258850694,-0.1593343168,-0.1525787413,0.1120038629,0.2984237075,0.1595720649,0.4297607839,0.1938183755,0.0325590894,-0.3667456508,0.3707756996,-0.2597859502,-0.114257209,-0.3271948993,0.0884747803,-0.4602726996,-0.0919382945,-0.0913776383,-0.1079575643,-0.0109050162,-0.1133172736,-0.2615408003,-0.0545394234,-0.041678831,0.194476366,0.4526271522,0.0286847409,0.3426935077,0.2780859172,-0.1762031019,0.1364350766,0.299529165,-0.1375357509,-0.2169118226,-0.1742755175,0.0947208256,0.2804147899,0.0028124982,-0.3161248565,-0.069879584,-0.1523030251,-0.0805635378,-0.3296914101,-0.0116260825,0.2260961533,0.1105515733,-0.2684786022,-0.4908463657,0.2091236413,0.0981090516,-0.2013097703,0.5697401762,-0.410910964,-0.1549281329,0.4900784791,-0.2227364779,0.6698690653,-0.2544949055,0.0256498773,-0.1605151296,0.30040434,0.2402479053,-0.2775893807,0.0432773344,-0.4373539984,-0.0793169439,0.0313826874,-0.2288014293,-0.0601102747,0.1101601049,-0.3473621905,0.2662377357,-0.0410998054,0.4165115952,-0.0033621988,0.2364912927,0.2673930824,-0.0929264352,0.0670116097,0.0969400555,0.1242618784,0.464329958,0.0484998301,-0.0473349839,0.039439816,-0.0779561996,-0.1329875439,0.1328337193,-0.5107427835,0.3210960031,-0.1541863382,0.1124994531,-0.0773027539,0.3645000756,0.3381867409,0.1547324657,-0.2069272101,0.2277360857,0.1569675505,0.2691049874,-0.0063426448,-0.226877898,0.0644017383,0.1912046522,-0.2837256789,0.0375011005,-0.0803442597,-0.4013679624,-0.3720651865,0.27790308,0.4389127195,-0.3181234598,-0.1861631125,0.1888598949,-0.0183269102,-0.0971026048,0.0633447617,0.1931180656,0.0781806484,0.4670720696,-0.285960108,-0.3394742012,0.0985542536,0.5539428592,0.1574303359,-0.0891848654,0.4062616229,-0.3077758253,-0.0314525366,-0.1451089829,0.0546298996,0.044570826,-0.3717840016,0.1382370293,-0.0392307825,0.2710923553,0.0749620572,-0.0957323238,-0.3017767668,0.0246333685,-0.3069173694,-0.2918444574,-0.3193426728,-0.0943848565,0.0496077761,-0.0205662996,0.3800570071,0.1509476751,0.0756855309,0.2781947255,-0.210169822,0.2410122007,0.389896661,0.1799956113,-0.181224972,-0.0508018546,0.0156558454,-0.2081274539,-0.004152196,0.096401453,-0.0449937694,-0.1464496553,-0.1618010104,0.2024939656,0.1696569473,0.0389281996,-0.2793242633,-0.2493662238,0.175191015,-0.5360283256,0.1491341889,0.2523556352,-0.0624468848,0.1519960165,0.4219889045,-0.2565125525,-0.0557903387,0.0638465732,-0.3750688732,-0.0628033057,-0.0880325362,0.2123445868,0.1389136761,0.1438453645,-0.1469783783,-0.1768975556,0.2108755112,0.0691963434,0.3749499023,-0.2957571745,-0.0487527251,0.1230907962,-0.1653466076,0.2767085433,-0.3283170462,-0.1693475991,0.2505047023,0.1382929534,-0.2606954575,-0.2727978528,0.4128356576,0.1433226466,-0.1123846918,0.056402985,0.2925511301,-0.3135686815,0.3032242954,0.0464540236,0.5587585568,-0.2876917124,0.338470459,0.2365060449,0.1580397338,0.1666840613,0.136705339,0.2150016278,-0.096154511,0.3128519058,0.0177459661,0.2945986688,0.1948968172,-0.0888017267,-0.0092838211,-0.4697328508,-0.0337138288,0.1261710972,-0.3483223915,0.3593920767,-0.0375496261,0.1433276236,0.0720328838,-0.1121677235,-0.110886395,0.3026529253,-0.3052933514,-0.1673251539,0.2243474126,-0.0767309293,0.0525531359,0.0025169235,0.0143131306,0.0683228448,0.4486830533,0.1085495576,-0.3328796327,-0.0616294779,-0.224594146,0.1390073448,0.4831044078,-0.1581938416,0.1845668107,-0.2319359779,0.0140547482,-0.0975998268,0.1668659002,0.3799286485,0.4430533946,-0.1481638104,0.351616025,0.1520383358,0.0271630622,-0.0965811908,0.390106827,-0.2279262245,-0.1441182643,-0.0544350222,0.0654741153,-0.0352086127,-0.1429272741,-0.0151414266,0.3786447048,-0.0796127543,0.1650294811,0.0397062227,-0.0452669524,-0.1049192697,0.1887014955,-0.267098099,0.1108142585,0.5441742539,0.0738949552,0.1876515001,-0.3705578446,0.0462144129,0.143771857,0.174605608,0.3369737267,-0.2889632583,-0.4057234824,-0.2916530669,-0.7126657963,0.3394765556,-0.1694641858,0.2732893527,0.0644730404,0.043443013,-0.1857524961,-0.1183680221,0.2397449315,-0.0535907075,-0.0421033166,0.2503176928,-0.1993057877,-0.0917779654,-0.2811604142,0.1799590737,0.1436006576,-0.3986141682,0.2988144457,-0.1648105681,0.0395180732,-0.191458568,-0.031722378,0.0602946617,0.0012448059,0.8092890978,0.1159820557,0.1138470024,-0.0567313246,0.2298415154,-0.2968367934,0.0546786189,-0.0002243936,0.0132693211,-0.0415909141,0.0151802506,-0.2264107466,-0.2175167799,-0.4425337017,-0.2809734643,0.2237388045,-0.0639511347,-0.1287063658,-0.0596592389,0.0081469119,0.3285406828,-0.0946541801,0.4309675395,-0.1998093128,0.3871890903,-0.2032527179,-0.1928033829,0.3517588079,-0.5707944632,-0.4315943122,-0.1778105646,0.3120351434,-0.3041443527,-0.2027008384,-0.5087888241,0.1133152023,0.514434576,0.1337886453,-0.1194233969,0.0146736875,-0.0940273851,0.2663268745,-0.1888667047,0.3684408963,0.1690643281,-0.1008474827,0.1587291956,-0.1010639444]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2825","title":"The datasets.map function does not load cached dataset after moving python script","comments":"#2854 fixed the issue :)\r\n\r\nWe'll do a new release of `datasets` soon to make the fix available.\r\nIn the meantime, feel free to try it out by installing `datasets` from source\r\n\r\nIf you have other issues or any question, feel free to re-open the issue :)","body":"## Describe the bug\r\nThe datasets.map function caches the processed data to a certain directory. When the map function is called another time with totally the same parameters, the cached data are supposed to be reloaded instead of re-processing. However, it doesn't reuse cached data sometimes. I use the common data processing in different tasks, the datasets are processed again, the only difference is that I run them in different files.\r\n\r\n## Steps to reproduce the bug\r\nJust run the following codes in different .py files.\r\n```python\r\nif __name__ == '__main__':\r\n from datasets import load_dataset\r\n from transformers import AutoTokenizer\r\n raw_datasets = load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\n\r\n def tokenize_function(examples):\r\n return tokenizer(examples[\"text\"], padding=\"max_length\", truncation=True)\r\n\r\n\r\n tokenized_datasets = raw_datasets.map(tokenize_function, batched=True)\r\n```\r\n\r\n## Expected results\r\nThe map function should reload data in the second or any later runs.\r\n\r\n## Actual results\r\nThe processing happens in each run.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: linux\r\n- Python version: 3.7.6\r\n- PyArrow version: 3.0.0\r\n\r\nThis is the first time I report a bug. If there is any problem or confusing description, please let me know \ud83d\ude04.\r\n","comment_length":47,"text":"The datasets.map function does not load cached dataset after moving python script\n## Describe the bug\r\nThe datasets.map function caches the processed data to a certain directory. When the map function is called another time with totally the same parameters, the cached data are supposed to be reloaded instead of re-processing. However, it doesn't reuse cached data sometimes. I use the common data processing in different tasks, the datasets are processed again, the only difference is that I run them in different files.\r\n\r\n## Steps to reproduce the bug\r\nJust run the following codes in different .py files.\r\n```python\r\nif __name__ == '__main__':\r\n from datasets import load_dataset\r\n from transformers import AutoTokenizer\r\n raw_datasets = load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\n\r\n def tokenize_function(examples):\r\n return tokenizer(examples[\"text\"], padding=\"max_length\", truncation=True)\r\n\r\n\r\n tokenized_datasets = raw_datasets.map(tokenize_function, batched=True)\r\n```\r\n\r\n## Expected results\r\nThe map function should reload data in the second or any later runs.\r\n\r\n## Actual results\r\nThe processing happens in each run.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: linux\r\n- Python version: 3.7.6\r\n- PyArrow version: 3.0.0\r\n\r\nThis is the first time I report a bug. If there is any problem or confusing description, please let me know \ud83d\ude04.\r\n\n#2854 fixed the issue :)\r\n\r\nWe'll do a new release of `datasets` soon to make the fix available.\r\nIn the meantime, feel free to try it out by installing `datasets` from source\r\n\r\nIf you have other issues or any question, feel free to re-open the issue :)","embeddings":[-0.1273975223,0.0724143684,0.0914586037,0.0095804129,0.1599638164,-0.0660084039,0.4047463834,0.246048376,0.2259855419,-0.0873799026,0.1144409701,0.5168568492,-0.0583694465,-0.1686522663,0.1454256177,0.1743037403,0.2147076875,0.1613187045,0.0151957124,-0.1748613268,-0.2260828018,-0.0004946861,-0.3888833523,-0.0530203693,-0.4167590141,0.1189529076,-0.1496831626,-0.0641775802,0.2103359103,-0.4711386859,0.3424901962,0.1252883971,0.2924120724,0.3714997768,-0.0001175742,-0.0325822458,0.2381018996,-0.0887213871,-0.1687702239,-0.1984855384,-0.0911653638,-0.0784300789,0.1140753031,-0.1740205139,0.0148621257,-0.0670339316,-0.0580180101,-0.5517778993,0.3449013233,0.2660760283,0.1569478512,0.091664575,-0.1324367374,0.1409147382,-0.0919476375,-0.1474609375,-0.0402100831,-0.1767199039,0.288394928,-0.4073186815,-0.1861928105,0.2407978177,-0.1470916569,0.0880867541,0.3459538817,0.0264889579,-0.1978945434,-0.3031310439,0.3096983433,-0.0609247796,0.3651449382,-0.4083526433,-0.2633101344,-0.2716084719,-0.1292909235,-0.0306372046,0.1167385951,-0.2174400091,0.2154372633,0.1586684734,-0.4689263701,-0.1160352752,0.1113872156,0.1911925524,-0.126292944,0.4278686345,-0.0862307549,0.1805843711,-0.1728989333,-0.0524413548,0.1630068719,-0.1279247999,0.0592711233,0.5038397312,-0.1977031082,-0.0271315686,0.1539091021,-0.0486140214,0.0574085824,0.1103376374,-0.009577659,0.2165228575,-0.1987567395,-0.0295949746,0.1525702477,0.5008949041,0.1595460474,0.4105488658,0.1037071869,-0.0912700221,-0.6493957043,-0.1686208397,-0.1285147518,-0.0576798096,0.6006772518,0.0352326371,0.203829214,-0.0009124889,0.1098072156,0.0979165137,-0.071602203,-0.1900148988,0.0337545089,0.449154079,-0.0397840627,0.3059886694,-0.0941899493,0.2121893764,-0.1355014145,0.0562306903,-0.1956659108,-0.2108473182,-0.1978409588,0.2894202173,0.2324400395,0.1111553609,0.3557316959,0.0897305757,-0.0325618498,-0.120944269,0.277336508,0.0440268628,0.6040584445,0.1215034574,-0.0231417157,0.5818362236,0.2637910247,0.0014681084,-0.151149407,0.1716978252,-0.6476752162,-0.1697181016,0.3885863125,0.0899603292,-0.2056822926,-0.1571826786,-0.1573634148,-0.0107510462,0.5134230852,-0.2149523944,-0.0122531233,-0.3363991678,-0.3412780762,-0.3847541511,0.1996460408,0.4401876926,-0.051230289,-0.1549954563,-0.1365014464,0.3355631232,0.1325760931,0.1070944518,-0.3555155993,0.3302602172,-0.2539390326,0.3772550225,0.2882091701,-0.448125571,-0.7045115232,0.0289333984,0.014852711,0.2682565451,-0.2539271712,-0.1451850682,0.2909582853,-0.1724201739,-0.0794617459,0.2683820128,0.2199956477,0.3350924253,-0.2560369372,-0.1363706589,0.3257187307,-0.1690098792,0.2152385265,0.0236729011,-0.0220000949,0.0955117047,0.1455851644,-0.0097120069,0.2444512248,0.1607978344,0.1761437058,0.0880905166,0.3228500485,-0.0743065029,-0.2499892563,0.2898178399,-0.1547434926,-0.4490094483,-0.097189784,-0.106667392,-0.1024616584,-0.1602702439,-0.5025526285,-0.2756501436,0.1304307282,0.2702072859,0.2430558652,0.1169801876,-0.2288282216,0.1204118058,0.3192523122,-0.1263379157,-0.0870229453,0.1495259702,0.1631319225,-0.0988768414,-0.2654795647,0.0836536586,0.4113591611,-0.091093339,-0.2791558206,0.2201692164,0.2130369991,0.2288496047,-0.06588386,0.0021975208,0.023006035,-0.0422747508,0.1455105543,0.2342197299,0.0880005285,-0.1172928885,0.1115299612,0.2227391154,0.2169865966,0.2037428319,-0.0055456567,-0.0320434906,0.1114534512,-0.1695540398,0.0991046652,-0.2482886761,-0.2217267752,-0.1447315514,0.3873695731,0.1267309636,0.3542512357,0.334225744,0.6654902101,0.1740915775,0.1755018085,0.0477037206,-0.6600905657,-0.2747760713,-0.0900236368,0.1625278145,0.2718919814,-0.003267417,-0.0523946099,-0.0592408516,0.1783802509,0.0688536093,0.0634294674,-0.0239987858,0.1010288224,0.1833835542,0.1937384009,0.1529010087,-0.1032355726,0.174034372,0.1970168203,0.2634407282,-0.1638142318,0.0178135596,-0.4033915699,0.4058876038,-0.011080293,-0.2489236444,-0.0473926626,-0.3511256576,-0.0560280718,0.2632542551,0.0609169528,0.2647619247,0.0576650612,-0.0182358995,0.006251486,-0.3572273254,-0.2716269493,-0.4146781266,-0.2249661982,-0.0644470006,0.159751907,-0.1184373647,0.1242043227,-0.0222660135,0.0828909203,-0.3509310484,-0.2879324257,-0.0605935492,0.0784289464,-0.0372277647,-0.2462186068,0.0106649986,-0.2674310803,-0.1137792692,0.1633513719,-0.5391669869,-0.0414720327,0.1899370402,0.0173565503,-0.0609043166,0.1396540403,-0.334335655,-0.1855303198,0.0146545516,0.0790277645,-0.0025065762,-0.0581584163,-0.0783223137,-0.0806789175,-0.1290310621,-0.2679095566,-0.1495465487,-0.3555653691,-0.3247364461,0.3774479628,-0.2381370068,-0.2658275664,-0.0918252617,-0.0233523957,0.2156021148,0.218480289,-0.3069497049,-0.0684558377,-0.183655411,0.1931986213,0.0127494996,0.1517698914,0.4884290695,0.1360828131,0.0904133543,-0.215189144,-0.0438366532,0.0594989993,0.1414827704,0.0310218632,0.0702757984,0.0630455017,-0.0158293452,0.8712025881,0.3394305706,-0.2434746325,0.3906833827,-0.1166394204,0.1712316871,-0.2813171446,-0.368688643,-0.0921083391,-0.1965304464,-0.0361909419,0.1398809999,0.0031318127,-0.1993825585,0.0029483326,-0.1436542124,-0.421038568,-0.2925756872,0.0749979168,-0.095634304,0.2742871344,0.324069798,0.2811284661,-0.6588519216,-0.2407615036,0.208889395,-0.0415099226,0.3959528208,-0.0165566411,-0.3597285748,-0.0942782909,-0.331176728,0.2504333258,0.0488030389,0.4142693877,0.0938733891,-0.2136970162,-0.0453537293,0.0769306943,0.2219329327,-0.3888384402,0.0473807938,0.2300531268,-0.1469590813,-0.1887865067,-0.1479176581,0.0483716764,0.2653465569,0.1881247908,0.4231220186,0.1420800388,0.0004303179,-0.3593809903,0.3494275808,-0.2431687266,-0.0385589041,-0.3027107418,0.083901085,-0.5145257711,-0.0996667966,-0.0840012208,-0.0430639014,-0.0338879228,-0.1100077704,-0.2868148983,-0.1015879735,-0.0205446184,0.2367705554,0.4534938633,0.0386607423,0.3291157186,0.2824527025,-0.1834538877,0.1781745702,0.2456832826,-0.1328651905,-0.2484772056,-0.1681522578,0.0905909985,0.2635990381,0.0388325565,-0.2994982898,-0.0738540739,-0.1791797578,-0.0201591309,-0.3251717687,0.0845126733,0.2234952003,0.1066389307,-0.2621914744,-0.4880998135,0.1662352532,0.0820382535,-0.1978458911,0.5077859759,-0.3288021088,-0.1782568693,0.5263878107,-0.167229712,0.7030587792,-0.2614830434,0.0227349121,-0.1484288424,0.3079768419,0.2823582888,-0.2782203555,0.0377055146,-0.4485217929,-0.0738487169,0.0370742343,-0.1756739914,-0.0501448512,0.0857838467,-0.3708004355,0.2630684972,-0.0307351425,0.4242716134,0.0008204111,0.2950986028,0.2175413519,-0.161154449,0.0709258392,0.1475217342,0.0955183953,0.4126780331,0.0385906659,-0.0242801625,0.0524465889,-0.0692033842,-0.1368081868,0.1307314932,-0.4697889686,0.3341296911,-0.1529870331,0.1190001294,-0.0603015199,0.3621901572,0.3968267441,0.1983506531,-0.256606549,0.2645443678,0.0978826508,0.2188669145,0.0133805219,-0.207746774,0.0840460137,0.1926829964,-0.2933793366,0.0694650561,-0.0555588156,-0.3760706782,-0.3659450114,0.1993739605,0.4698028564,-0.3571555018,-0.1596153677,0.1972794086,-0.0259875115,-0.1202722192,0.1127587929,0.1794176251,0.0735144988,0.4738390148,-0.2469421029,-0.3006403446,0.0969439596,0.5473600626,0.086856626,-0.0703573152,0.4687364995,-0.2822931707,-0.0759916306,-0.1681775451,0.062826477,0.1285603493,-0.3534423411,0.0983731821,-0.0091100838,0.2103241235,0.1147621945,-0.0235994626,-0.2950790226,-0.0067663845,-0.2794577181,-0.2637412846,-0.3657056689,-0.0859150887,0.0335573852,0.043395739,0.3683927059,0.1731268167,0.0272705927,0.2843603492,-0.2767242789,0.2214020342,0.3566819429,0.1540384591,-0.2106957734,-0.1227830574,0.0563885234,-0.2020298392,0.0518809259,0.1056120396,-0.0767099038,-0.2032084316,-0.149872005,0.1702451557,0.1818999946,-0.0067304159,-0.2689258754,-0.1823142916,0.1438468546,-0.5760858655,0.1551049948,0.2457036674,-0.0718843564,0.1519443691,0.4375988245,-0.2316621393,-0.0271024555,0.017064603,-0.3173016906,-0.0495208055,-0.1127286255,0.2077107728,0.1210681722,0.1175579205,-0.0437983572,-0.1282811016,0.2395606786,0.0934648514,0.3994583189,-0.2951737046,-0.0739271864,0.1158022583,-0.1481214315,0.2610611618,-0.2970363498,-0.2015527189,0.3140524626,0.1927070767,-0.2493180633,-0.2417922467,0.3572395146,0.137401998,-0.0605151393,0.0118390005,0.2675012648,-0.3133398294,0.2877171338,0.0523609631,0.5716959834,-0.320587039,0.2812640071,0.206658259,0.1167723015,0.1284619272,0.1284874678,0.1781931818,-0.0683909506,0.3175272644,-0.0119767888,0.3628058136,0.1818665713,-0.0623005144,-0.0099807987,-0.4547534585,-0.0277174935,0.1175221875,-0.2924536467,0.3628990352,-0.0239516106,0.1671731323,0.0781309903,-0.1214324385,-0.1191205233,0.3189847469,-0.3081088364,-0.1804355383,0.1479912996,-0.0687662065,0.0302227773,0.0211466402,-0.0196929816,0.0529189035,0.3540977538,0.0921824947,-0.3180074394,-0.0996504426,-0.1993022412,0.0919285864,0.4701643884,-0.146915406,0.1937372088,-0.1774657071,-0.0135230236,-0.0918060765,0.2231097072,0.4018775821,0.45461303,-0.1482905447,0.3653244078,0.1168749332,0.0095231291,-0.1005837545,0.3791520298,-0.2120788693,-0.1514657438,-0.0042177239,0.1118993983,-0.0705499873,-0.2121526003,-0.0576564893,0.3330321014,-0.0077678049,0.146998927,0.1022792161,-0.0578393489,-0.1023925692,0.180457592,-0.288815856,0.1078632399,0.5013055801,0.0990960747,0.1918560117,-0.3919435441,0.0726611242,0.1490288526,0.1968065649,0.3455561996,-0.3320749402,-0.4023283124,-0.3072879612,-0.6909540296,0.3282343149,-0.1399887949,0.2348546535,0.0173072442,0.0463602841,-0.1554466039,-0.1246749461,0.2110840529,-0.0270854849,-0.068589516,0.2278555632,-0.2249159217,0.0068836836,-0.2312878668,0.1672779918,0.137252003,-0.3825232685,0.2694316208,-0.1653431356,0.0654588193,-0.1703134477,-0.0070523522,0.0670723468,-0.0568998232,0.7861573696,0.1311318129,0.1361974329,-0.0896755606,0.2551853657,-0.3458391428,0.0296220463,-0.0395722426,0.0076721921,-0.0578936674,0.0625895932,-0.1507781148,-0.2193167359,-0.4397763312,-0.2626307309,0.2159584612,-0.0473887809,-0.1473796666,-0.0183445662,0.0134375496,0.3102091551,-0.0609849356,0.4418364465,-0.1894945651,0.3719442487,-0.2240031213,-0.2603968382,0.3844135702,-0.5583379269,-0.469915241,-0.2035344094,0.2741136253,-0.2419668734,-0.2197262198,-0.4971866906,0.1145918518,0.4538178146,0.1593986899,-0.1826810688,0.0075181648,-0.0619167425,0.2377015054,-0.2190838903,0.3804850578,0.1765696704,-0.0737811327,0.0922997296,-0.0781693682]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2823","title":"HF_DATASETS_CACHE variable in Windows","comments":"Agh - I'm a muppet. No quote marks are needed.\r\nset HF_DATASETS_CACHE = C:\\Datasets\r\nworks as intended.","body":"I can't seem to use a custom Cache directory in Windows. I have tried:\r\n\r\nset HF_DATASETS_CACHE = \"C:\\Datasets\"\r\nset HF_DATASETS_CACHE = \"C:\/Datasets\"\r\nset HF_DATASETS_CACHE = \"C:\\\\Datasets\"\r\nset HF_DATASETS_CACHE = \"r'C:\\Datasets'\"\r\nset HF_DATASETS_CACHE = \"\\Datasets\"\r\nset HF_DATASETS_CACHE = \"\/Datasets\"\r\n\r\nIn each instance I get the \"[WinError 123] The filename, directory name, or volume label syntax is incorrect\" error when attempting to load a dataset","comment_length":17,"text":"HF_DATASETS_CACHE variable in Windows\nI can't seem to use a custom Cache directory in Windows. I have tried:\r\n\r\nset HF_DATASETS_CACHE = \"C:\\Datasets\"\r\nset HF_DATASETS_CACHE = \"C:\/Datasets\"\r\nset HF_DATASETS_CACHE = \"C:\\\\Datasets\"\r\nset HF_DATASETS_CACHE = \"r'C:\\Datasets'\"\r\nset HF_DATASETS_CACHE = \"\\Datasets\"\r\nset HF_DATASETS_CACHE = \"\/Datasets\"\r\n\r\nIn each instance I get the \"[WinError 123] The filename, directory name, or volume label syntax is incorrect\" error when attempting to load a dataset\nAgh - I'm a muppet. No quote marks are needed.\r\nset HF_DATASETS_CACHE = C:\\Datasets\r\nworks as intended.","embeddings":[-0.2557515204,0.4681499004,-0.0086679254,0.0938130617,-0.0932577625,0.2809450626,0.3591798544,0.0130967433,0.5492925644,0.1723245382,-0.1086298451,-0.3207066953,0.0269634761,-0.1362952143,-0.0884396136,0.1600556374,-0.0124639487,0.2051330358,0.3308285177,0.1052643135,-0.3471739888,0.194610998,-0.1163545176,0.1732594371,-0.2201039195,-0.0273021515,-0.2286344022,0.3452448845,0.1202253625,-0.0732737556,0.2985122204,0.1823600531,0.4611557126,0.4373255372,-0.0001192506,-0.0092273718,0.1367600113,-0.1833856255,-0.1629035026,0.1200159863,-0.2397752106,0.2823110223,-0.2644101381,0.0286023542,-0.002813136,0.2012717724,0.008684691,-0.633574605,-0.2559023499,0.3518752158,0.144456163,-0.2560602129,-0.4754669964,-0.1310121119,0.3233233988,0.4254634976,-0.1895282269,0.1743329912,0.0022010743,-0.2547373772,-0.1301951408,0.1984852701,0.0741537735,0.3853290975,0.3594753742,0.2023977488,-0.0893624648,0.1987679154,0.2981185019,-0.028904222,0.911711812,-0.3926505446,-0.067581974,0.1137470528,-0.1882463098,-0.3810411096,0.3569802046,0.1096177548,-0.2152440995,0.0923623815,0.0322469994,-0.209403649,-0.2581728101,0.1735568196,0.1536321193,-0.0576701201,0.0166171566,-0.0853283554,0.209992066,-0.1379805952,0.304087162,-0.242267251,-0.4997873604,0.1586566865,-0.2232942432,0.1928556412,-0.2934031487,0.5975059867,-0.0478701368,0.034979865,0.0655475706,0.1633851677,-0.1220712364,-0.0252837315,-0.0709528327,0.4531860054,0.0865988582,0.0687403828,0.0591183044,-0.1701724678,-0.1758125871,-0.088145569,-0.2163306922,-0.449667871,0.218222633,0.060151238,-0.1500684768,-0.5533458591,0.0617661439,-0.2378155887,0.1814159453,-0.1227409691,0.282826215,0.3287512064,0.181951344,-0.155424729,-0.2453553677,0.1222565249,-0.1248539314,-0.0093623875,-0.044213336,0.115095675,-0.3545196056,0.4502945244,0.4435072243,0.1504650712,-0.0859052166,-0.1223051623,0.2869002819,0.1731900722,-0.0152334422,-0.2569104731,-0.0175518487,0.1623936146,-0.0230819732,0.3286900818,0.0982614011,-0.1849636436,-0.2361326665,0.2819387317,-0.2798592448,-0.4200795293,-0.1352794021,-0.004328853,-0.073356919,-0.3321877122,-0.2920281291,0.0190325361,0.0367763527,-0.0160885863,0.0514025465,0.0777065381,-0.292544961,-0.1867620945,-0.1598431617,0.4258094728,-0.6456723809,0.0549008623,-0.1299296767,0.0349943377,0.1590850204,0.3617969155,-0.1314047426,0.144155547,-0.3559282422,0.0094692772,0.291824162,-0.520745635,-0.457478404,0.0855572671,0.3869653642,-0.2977321446,0.284693718,0.0809179693,0.1606728584,-0.1620028168,0.0116999075,0.216101557,0.0686244294,-0.2469922751,-0.0120328413,-0.1828505844,0.0438937731,0.2415065765,-0.0034101778,0.4013135135,0.1921704412,0.0047577857,0.0524615049,-0.0943140462,0.1270525903,0.30668208,0.3205009103,0.4424470663,0.1727309972,0.1762866676,-0.2260034233,0.291208148,-0.1170075089,-0.2172814012,0.2878508568,0.0115911169,0.0161375888,-0.2953485548,-0.220785588,-0.1811322868,-0.0076177837,0.2608892024,0.2349206507,-0.2933544219,-0.02679554,0.6589478254,0.2265408784,-0.1941644102,0.2957310677,-0.0269145686,-0.1871454716,-0.0424646065,-0.1201222688,-0.0311330575,0.2132748365,0.1775698513,-0.2325893193,0.3005902171,0.0022563951,-0.0776350871,0.3121986389,0.3358964622,-0.0462791063,-0.0129095605,0.5341506004,0.0875013471,-0.0112856096,-0.0671750233,-0.2536233366,-0.028579792,-0.1555690765,0.3571642339,-0.2178194523,-0.0085013164,0.1321295351,-0.0853915066,0.3848129809,-0.2251411974,0.3857040703,-0.2674026787,0.1608100682,-0.204515174,-0.0223741475,0.0500294492,0.2000289559,0.0464820378,0.0900041834,0.3841385543,-0.2304574251,-0.061300002,0.1181385964,0.1175849214,0.6406487823,0.1460811198,-0.0062330784,0.0381106734,-0.1333918869,-0.0382689051,0.1126071587,-0.1810551733,-0.0053234501,-0.040674448,-0.0634999946,-0.1481620371,0.1249394044,-0.353921324,-0.0089608701,0.0603265539,-0.402530998,0.125486508,-0.2663137317,-0.2450944632,-0.0418880805,-0.3795005977,-0.1935695559,-0.1571077108,-0.0439107344,0.0742222071,0.3458555937,-0.0554470196,-0.6163679361,0.2907798588,-0.1938139796,-0.5244791508,-0.2568773627,-0.0308192391,-0.3730294704,-0.0545444898,0.0315489955,0.0334092863,0.0565818474,0.0013542117,0.402936697,-0.433188349,0.2553689778,0.172015205,0.3753192127,0.2596966922,0.0151372654,-0.0903709978,0.0453084819,-0.1888089627,0.2970713973,-0.0839478523,0.1273833066,-0.0312315002,-0.1942108572,-0.0280021261,0.0193983261,0.2274570316,-0.3716921508,-0.2120231241,-0.2424019873,0.143338412,0.0797449723,0.3831820488,-0.2004886121,0.0953654572,0.2932655513,-0.3508400619,-0.3323981464,-0.3654141128,0.5393200517,-0.4161731005,0.0118386988,0.1936699301,0.2055070102,-0.2015186995,0.2151324451,-0.4242047369,0.0964951068,-0.1795964092,0.0807430893,0.0342938155,-0.088356927,0.1362947226,-0.2433892041,-0.0392160565,-0.2442471087,-0.5972727537,0.2365330309,-0.0724125803,0.4560070038,0.3650609851,0.1059790179,-0.0916182622,0.4457309246,0.0964218825,0.1595916003,0.4605713785,-0.0735017136,0.3184109926,-0.0136250146,-0.0809273124,0.1316006333,-0.0595001951,-0.2526791394,0.2065048218,0.2332798243,-0.1624370068,0.025440922,-0.0352298506,0.151282534,-0.1032656729,0.1092287526,-0.3504682183,0.243460685,-0.0305181406,-0.1385994554,-0.34907493,-0.4369841516,0.470767349,0.5245847702,-0.2454287112,0.4193021953,-0.1807518005,0.111305207,-0.2360460013,0.4586577415,-0.119205825,0.3818719983,-0.4501280785,0.3092700541,-0.1123982519,-0.0147014949,0.5110964179,-0.4883116782,0.1709703803,0.4405175447,0.2215117812,0.1607983112,-0.2506406009,0.0454852134,0.3383985758,-0.4176425338,0.090141952,0.2738480568,-0.2527697384,0.053219907,-0.0272600707,-0.4673559666,-0.3163568676,-0.0918906704,-0.0569565557,0.064101018,-0.2776578069,-0.1184407249,0.0401036739,0.3600781262,0.3120995164,-0.1930749714,0.1677638292,0.3788413703,-0.143060863,0.3332851827,0.2072020471,-0.0836262926,0.0844900906,0.1856536567,0.1926142126,0.2615593374,-0.1616356969,-0.1108481511,0.1743608713,-0.2599956691,0.1231755093,0.0483884849,-0.1970623732,-0.1120580658,0.0306535121,-0.0253591854,-0.3931269646,0.209609881,-0.0403304249,-0.1396387368,-0.2823480666,-0.3919416964,0.4164451063,0.0744006187,-0.1704442501,0.4063184559,0.3550950885,-0.3551276624,0.2467372417,0.0781498849,0.6511047482,-0.2317091674,0.0413221121,-0.2924341559,-0.459600538,-0.1842951775,-0.1948485225,0.0385683216,-0.12402758,-0.2318082899,-0.150369212,-0.0903323218,0.3203728795,0.4188167453,-0.1880275905,0.5310227275,0.0854854882,0.0732927173,0.1065138802,-0.1120631695,0.0741085932,-0.3800901473,0.2787130475,0.1939138621,-0.0752650425,0.2040399015,-0.0995933264,0.1389640868,-0.0254558157,0.1124390215,-0.4690798223,-0.0862769336,-0.5284095407,0.0550687462,-0.1582950801,-0.1572606713,0.1016687676,0.1178226769,0.4569378793,-0.0447974652,-0.2122693062,0.1578858346,0.1468140036,0.0920069292,-0.213471368,-0.0168731213,0.0020503178,-0.1468966156,-0.233698681,-0.1124359667,0.0008228201,-0.5835405588,-0.2684060037,-0.2306132168,0.2315963358,-0.002110668,-0.0922531486,-0.1822307557,0.230701074,0.2200394422,0.0676135048,-0.0096413596,0.209912315,0.1866279542,-0.4496470392,-0.0763368979,-0.1141119376,-0.0535052717,-0.4296340942,-0.2064580321,0.2676545084,0.0008903425,-0.0931643173,-0.0913444757,-0.0895103067,0.031917084,-0.1400099844,-0.2013543397,-0.0653561279,-0.1735588163,0.156319648,-0.0680225715,0.3295845985,-0.1584560871,-0.0559698716,-0.4187110066,-0.1215190813,-0.1466019452,-0.3597600162,0.1639639735,-0.2259476483,-0.3933383822,0.2311780304,-0.1383239925,-0.1911327541,0.0001045569,0.1545092314,-0.0920527205,-0.1364142746,0.1935989261,0.2963394225,0.1862527579,-0.0671018586,-0.2437205315,-0.0876429677,-0.0772349015,-0.2528389394,0.1778157502,0.019922616,0.1567851156,0.0384651385,-0.2315372527,0.0238007102,-0.1065338552,0.1360592097,0.4663167,-0.1788537949,0.3055639267,0.0394601189,-0.3041663468,-0.3280591369,0.0740611553,0.3465181291,0.3882786334,0.0188465286,0.1808875203,-0.072000131,-0.0405948199,-0.2922016084,0.061184559,-0.038822528,-0.0349996351,-0.2215377092,-0.2312077433,0.3110741377,0.2208743542,0.0948446169,-0.0335947089,-0.0226293914,0.0144147649,0.374923557,0.0413136259,0.0023533721,-0.266915977,-0.1157440096,-0.2078693658,-0.1973181814,0.4525607228,0.3078547418,0.0413310342,0.0699677244,-0.1190053597,0.5813835859,-0.1159400716,-0.3443711996,0.2724404931,-0.1668227762,0.2911586165,0.4228002131,0.48921296,-0.02943944,0.8494073153,0.2110210806,-0.02398693,0.3741270304,-0.0340198241,-0.2246545404,-0.584741354,-0.173425734,0.1263131201,0.2699767947,0.2626489699,-0.3228519261,0.1719605625,-0.1433338523,-0.1161764339,-0.1790141165,-0.4177320898,0.1715865731,-0.3374357224,0.3003840148,0.0978951156,-0.1745757163,0.092420727,0.0679262727,-0.0253723301,0.1279136986,-0.2228805125,0.0852088109,0.0766691491,0.3389742672,0.2513722777,-0.1098921672,-0.128888011,0.2159267515,-0.2958231568,0.1166778281,0.4527239203,0.0408511087,0.1794822216,0.2273029834,-0.2377319932,-0.0539041087,-0.2109176069,0.0057932292,-0.0493150577,0.0424005501,-0.036436744,-0.1378655136,0.1513020843,0.1018326283,0.033102978,0.253862679,-0.2496426851,0.316180706,0.0220187567,0.2991948128,0.2000438869,0.192247197,0.2069408596,0.2616453469,-0.2504758835,-0.1989166886,0.3196480572,-0.0838438794,0.2157267928,0.2454298586,0.0192266889,0.1832866967,0.423425585,0.1119875908,0.3532936871,-0.2295434028,-0.0899119377,-0.476319015,-0.0003468155,-0.0555821285,-0.2668922246,-0.0472123288,0.0526003689,-0.0892351121,-0.1452381015,0.4688036442,-0.0866777599,0.0472869799,0.2267997712,-0.3549312949,-0.0105335843,0.2644642293,0.4981248677,0.1589110941,-0.5530814528,0.0672092214,0.455622673,-0.0252314135,-0.0870580971,-0.0961905569,0.4168117046,0.0889464468,0.4691800773,-0.2186760008,0.2127042413,0.2576371133,0.1848213226,-0.4217242301,0.0459140837,0.0423344076,-0.3932206333,-0.0256986078,0.1231914535,-0.6304084063,-0.029116191,-0.2919798791,-0.0300271418,0.1483543962,0.2352732122,-0.0766762197,0.3016290665,-0.0349002667,0.1033950448,0.4050494432,0.2613582015,-0.1189399958,0.2810581923,-0.2620744407,0.1222636029,0.3842461109,-0.2537790239,-0.3603968024,0.0744023025,0.3058955967,0.0528524891,-0.1602582037,-0.208542183,0.0974018201,0.2296550572,0.2220070958,-0.2222031653,0.0691856667,-0.1327781975,0.1328295916,0.0688432753,0.041508913,-0.0183698833,-0.0479435734,-0.3267987072,0.0389851779]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2821","title":"Cannot load linnaeus dataset","comments":"Thanks for reporting ! #2852 fixed this error\r\n\r\nWe'll do a new release of `datasets` soon :)","body":"## Describe the bug\r\nThe [linnaeus](https:\/\/huggingface.co\/datasets\/linnaeus) dataset cannot be loaded. To reproduce:\r\n```\r\nfrom datasets import load_dataset\r\n\r\ndatasets = load_dataset(\"linnaeus\")\r\n```\r\nThis results in:\r\n```\r\nDownloading and preparing dataset linnaeus\/linnaeus (download: 17.36 MiB, generated: 8.74 MiB, post-processed: Unknown size, total: 26.10 MiB) to \/root\/.cache\/huggingface\/datasets\/linnaeus\/linnaeus\/1.0.0\/2ff05dbc256108233262f596e09e322dbc3db067202de14286913607cd9cb704...\r\n---------------------------------------------------------------------------\r\nConnectionError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 datasets = load_dataset(\"linnaeus\")\r\n\r\n11 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 603 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 604 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 605 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 606 \r\n 607 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/drive.google.com\/u\/0\/uc?id=1OletxmPYNkz2ltOr9pyT0b0iBtUWxslh&export=download\/\r\n```","comment_length":17,"text":"Cannot load linnaeus dataset\n## Describe the bug\r\nThe [linnaeus](https:\/\/huggingface.co\/datasets\/linnaeus) dataset cannot be loaded. To reproduce:\r\n```\r\nfrom datasets import load_dataset\r\n\r\ndatasets = load_dataset(\"linnaeus\")\r\n```\r\nThis results in:\r\n```\r\nDownloading and preparing dataset linnaeus\/linnaeus (download: 17.36 MiB, generated: 8.74 MiB, post-processed: Unknown size, total: 26.10 MiB) to \/root\/.cache\/huggingface\/datasets\/linnaeus\/linnaeus\/1.0.0\/2ff05dbc256108233262f596e09e322dbc3db067202de14286913607cd9cb704...\r\n---------------------------------------------------------------------------\r\nConnectionError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 datasets = load_dataset(\"linnaeus\")\r\n\r\n11 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 603 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 604 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 605 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 606 \r\n 607 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/drive.google.com\/u\/0\/uc?id=1OletxmPYNkz2ltOr9pyT0b0iBtUWxslh&export=download\/\r\n```\nThanks for reporting ! #2852 fixed this error\r\n\r\nWe'll do a new release of `datasets` soon :)","embeddings":[-0.1098577157,-0.143165946,-0.0036667234,0.5683761835,0.2294618189,-0.0716874748,0.0764331371,0.3127034903,0.1373578608,0.0487763658,-0.3296411037,0.0350638591,-0.0180285536,-0.1834695786,0.1603863835,-0.2352236807,-0.1053614989,-0.0622279905,-0.4106493592,0.1044703573,-0.1308865696,0.3101219535,-0.0382798947,-0.0052360063,-0.0595476739,-0.0635992363,0.0269933958,0.4038961232,-0.0817026794,-0.5142930746,0.3658116162,-0.025758775,0.2221749574,0.6335695386,-0.0001129523,0.1002365649,0.3394467235,-0.0037233825,-0.417912811,-0.3696373105,-0.1447288692,-0.1747751683,0.147974357,-0.1763244122,-0.1769499779,0.1925733835,-0.0365836434,-0.4481865764,0.2486859858,0.2116878182,0.2418952137,0.2672031522,0.3284202516,-0.1788068116,0.0594662465,-0.0254503135,-0.104390353,0.7655639052,0.1173543409,0.2063794136,0.0510814972,0.3760934174,0.0908738375,0.16046004,0.4749838114,0.0394841954,0.2304551601,-0.2688692212,0.1130298451,0.1952230185,0.4000908136,-0.1607928276,-0.3665402234,-0.1934892684,0.0951567963,-0.3805069327,0.3234522939,0.1300037503,-0.2127992958,0.1132354736,-0.123492457,-0.1773454696,-0.2792692482,0.2881274521,-0.0447788946,-0.1690954268,-0.1575406343,0.1463273019,0.1806666553,-0.1248240247,-0.0539279319,0.0094063003,-0.0262687579,0.2579780817,-0.2460853308,0.1773656458,-0.0310983844,0.1817291528,0.2558598816,0.0904838815,-0.1404848844,0.0120945303,-0.0953951403,0.1886289269,0.298963815,0.0496456474,-0.0871780366,0.0046279761,0.2070226222,0.5039040446,-0.2293083221,-0.1274598688,0.0119244009,-0.2724320292,0.1770056784,-0.1214359626,0.4966045618,-0.3770104945,-0.1082005426,0.0917365104,-0.0458589755,0.0481627844,0.151230976,0.4551626444,-0.25800246,0.1544597,0.095444411,0.0614505634,-0.1204147637,-0.1457547247,-0.2277960777,0.1879870743,-0.177434206,0.0885476619,0.3645627201,-0.5190119743,0.1639645547,-0.014977023,0.2522332966,-0.1167270392,-0.0186835006,-0.2165719122,-0.107789509,0.311350435,0.0228776522,0.0862024426,0.1767017841,-0.1567439437,-0.1832847595,-0.0628963485,-0.3582116961,-0.2106400877,-0.0830561221,0.1629568785,-0.2142596841,0.0662850142,-0.3244718015,-0.0464086831,-0.1090323478,-0.1943716407,-0.1086106449,-0.1355385631,-0.1367982626,-0.209141925,0.3754203022,0.6375408173,-0.0808179602,0.0584750362,-0.1234021932,-0.0431577228,0.1322187781,0.3780945539,-0.1198968068,-0.0201250315,-0.2446826398,-0.1078016236,0.2805714011,-0.3060962558,-0.3442858756,0.11498072,-0.0337683484,0.3318838775,-0.0532563291,-0.0374132134,-0.0318786912,0.1323215216,0.1931566149,0.2832884789,-0.0563489199,-0.1293168962,-0.2122825384,-0.259457767,0.1632746458,0.2174010128,0.1880996227,0.1501255631,0.0770919248,0.0468290523,0.130879879,-0.0778299496,0.0669316649,0.1227842644,0.1351918876,0.4310694635,-0.0718185753,-0.3716206551,-0.4436382651,0.3363806009,-0.03146106,0.1913928241,-0.146351248,0.0657709762,-0.4101037383,0.0308009479,-0.0787293762,-0.1457380056,0.1097632498,0.0998461172,0.0353457816,0.2634499669,-0.2242711782,0.4821530581,0.0530593172,0.1775317788,-0.4551772773,0.257725209,-0.034705095,-0.0915881023,0.03112779,0.1576486081,0.1089980304,-0.0902797282,-0.103441529,0.3063961267,0.0274115913,-0.0405845679,0.1088259593,-0.1070616171,0.0980541557,-0.1691553593,0.0670722127,0.0534188636,0.1921955347,-0.0061213728,-0.3353888392,0.1313283145,-0.1543354392,0.2098008394,0.0107588526,0.0329193026,0.3360375762,-0.0081185251,-0.0462266766,-0.0963451713,0.3444593251,-0.108463861,0.4994331598,-0.0924285054,-0.3686579168,-0.1824690849,-0.1115126386,0.0226241816,0.0863872617,0.2621725202,-0.1716367751,-0.0400449112,0.0067801168,0.1041773707,0.3363501132,0.0850468576,0.0939544365,0.0940393284,0.0719725564,-0.0245300326,0.1491383314,-0.0482987016,0.2200093418,0.0769660249,0.1324644238,0.064665772,-0.4159032702,-0.3394742906,-0.0652086064,0.3232426643,-0.3425343037,0.1666306853,-0.1244778037,-0.3035856783,-0.2150221616,-0.2546468079,-0.4462381005,-0.2368631661,-0.3029857278,0.5309480429,0.1654843688,0.0352378562,-0.3215266764,0.0583712086,-0.040452648,-0.1734482944,-0.1826367676,-0.0247984342,-0.1803610772,0.0409550183,0.2866387963,0.1651409119,0.4407707751,-0.0842784494,-0.0442237966,-0.2323412746,-0.3730412424,-0.0294400081,0.0091494331,0.4914520383,0.1717602611,0.4271701574,-0.1075664908,-0.2444170862,0.4784312844,-0.0908168703,0.0685489252,0.0768222138,-0.203672111,0.0741951391,0.0008809214,-0.5234006047,-0.3201754391,-0.296497196,0.3565408289,0.0888761356,0.0241043288,0.4940642715,-0.0086904792,0.1140922159,-0.0923546031,0.1628053635,-0.3461090624,-0.618103981,0.3059367537,-0.2781543732,-0.4102417231,0.0600088499,0.1253888458,0.3633446097,-0.0327594392,-0.6117784381,-0.0045396043,-0.1223770529,0.1743086725,-0.0899428129,-0.1310316175,0.0207807552,-0.1267365217,-0.0387925059,0.0129932296,-0.1978799403,-0.1798412353,-0.0088726301,0.1235184073,0.0039644828,0.3006074727,-0.0057001635,0.5055919886,0.1251589656,0.0853411406,0.3692272902,0.05112122,0.4881687164,-0.3019167483,-0.5070818067,-0.1335797608,-0.1893318295,0.0256695691,0.0054524126,0.2143648267,-0.1302414685,-0.3036271632,-0.2458688319,-0.2516880035,-0.2852415442,0.1365142763,-0.0992162749,0.1378863752,0.0290458277,-0.1467989385,-0.1984380037,-0.1798201799,-0.1898526847,0.5857322812,0.2828261554,0.2692273855,-0.4927919507,0.2591663003,-0.5437831283,0.1316825151,-0.0735209212,0.4005352259,-0.2426699698,0.1835493892,0.1185221151,0.0203854684,0.3855633736,0.1478426605,0.4297837615,0.2452249527,-0.3469170034,-0.4284437597,0.1479193121,0.1272664666,-0.00303634,0.4096187055,0.3735685647,-0.2575898767,-0.030873524,0.160746485,0.2486558855,-0.1679402739,-0.2994202673,-0.2562253773,-0.2295896709,-0.3067691922,-0.0736396387,-0.011630307,0.3966581523,0.3705374599,0.0559865683,-0.1036729142,0.0487801284,0.2401845753,0.2222827077,0.1429222524,0.0422944129,0.4125692546,0.4882103503,0.1184626594,0.0244307518,0.6174946427,-0.0276214257,-0.6739399433,-0.0755878687,0.2311675847,-0.0158604011,0.1919622719,-0.1942848861,-0.1250990629,0.1588992923,0.1396012604,-0.0812206715,0.2773296237,0.0574068241,0.1291486174,-0.4189870059,-0.4065654278,0.4614156485,0.0728990734,-0.0248796791,0.3173590899,0.1852024347,-0.1900639683,0.0608675033,-0.1444210112,0.9203289747,-0.0377891921,0.129033193,0.3462245166,-0.0819157138,0.5078811646,0.0777398795,0.0116894366,-0.2593929768,-0.21444197,-0.0198573191,-0.1125155911,0.1943310499,-0.1327902973,-0.0507676788,0.4270289242,0.044167161,0.0233233124,0.0325411893,0.1910652965,-0.0050378563,-0.273011744,-0.550560534,0.1895080507,-0.1917114109,0.471381247,-0.1486088485,-0.1439522505,-0.0907168984,-0.1318972707,-0.0942242444,0.1017732993,-0.1582664102,0.3013027012,-0.0296492074,-0.3180105388,-0.0180561673,0.1300046146,-0.059777569,0.1640034467,-0.3081128299,0.2088374496,-0.172481373,-0.200623855,0.1004102528,0.1756676733,0.1397547275,-0.1888620406,-0.1557838321,0.0061076637,-0.0746376589,-0.3447106183,0.0671191514,-0.0123197734,-0.0949990526,-0.1176988035,-0.2597034872,-0.1651997566,0.1919189394,-0.0706138164,0.1399460435,0.2263596505,0.0399863645,-0.1092118546,0.2789070308,-0.2633683383,-0.093457967,0.6049341559,-0.0435961299,-0.2974512279,0.3684125245,0.2711165249,-0.1591368467,-0.0773260817,0.0296810027,-0.052925095,-0.5901020169,0.1162419692,0.0221219528,0.3236812055,-0.1370934099,-0.1013464108,0.2540922463,-0.3785056174,0.1729757339,-0.6587608457,-0.2781509459,0.2384362221,-0.2570829988,0.1325252652,0.1375771165,0.0773725137,-0.1278800815,0.0461509079,-0.2828451097,0.2159305364,-0.0680316389,-0.0153448246,0.4675156772,-0.0501257405,0.1522693038,-0.1469821334,0.1267457157,0.0237754527,-0.1837632358,-0.1875150204,-0.1944703758,0.1109411493,-0.1449028552,-0.2054656297,-0.0715201199,-0.2063226998,-0.0248486735,-0.1483222842,0.0929172039,0.4028233886,-0.0033535645,-0.0944644883,0.0791764185,0.1351013631,-0.2862477005,0.3571726382,0.0124075832,-0.0937272534,0.0591622069,-0.0165073853,-0.1653703302,-0.1095689386,-0.3179785311,-0.0277944282,-0.0287066959,0.0027031854,0.5015749335,-0.0748102739,0.064654015,0.2290384471,0.1794671416,0.1800268292,-0.0876839161,0.1274280101,0.0532912388,0.1653984338,-0.3052541316,-0.0202535875,0.1616721302,0.0090713101,-0.0189885125,0.1019241884,-0.0128978314,0.0412564278,-0.1517837048,0.0929423571,0.4723512828,-0.0426666662,0.1329584718,0.1896082014,-0.1569470465,-0.123482883,0.2636792958,0.1779013723,0.0847951621,0.6215701699,-0.115148738,0.1475511193,-0.1662748605,0.0233472232,-0.006759115,-0.2583770454,-0.3491345942,0.3788992465,0.0265688337,0.052099254,-0.0912096277,0.2827593386,-0.2950371504,-0.2381262779,-0.1105131358,0.1157539263,-0.1052901894,0.0295345671,0.0085168285,-0.2364259064,-0.1409330815,0.0166417677,-0.1832159609,-0.2049672008,0.3693362772,0.2010635585,-0.2949988246,-0.4677318335,-0.3516994119,0.3162498474,-0.0246515088,-0.1387357861,0.2355430573,0.397221446,-0.1357555836,0.1825453341,0.3464144468,0.5268877745,0.1495817453,0.1987739801,0.1265126914,-0.0913282633,-0.0762350783,-0.1071182638,0.1564548314,0.0565242246,0.1400633603,0.3333317041,0.2052952796,-0.1565948129,0.2190282792,-0.1009882614,0.322303921,-0.4391763806,0.3071796596,-0.2974657714,0.0331949778,-0.1077923998,0.0944547802,-0.2446489632,0.0657378957,0.2872646451,0.0876968727,-0.0261811167,-0.0293238517,0.063077189,0.0814956799,0.5135881901,0.2220079005,0.0571157895,-0.4488470554,-0.2352756113,-0.8884506822,0.107644096,-0.2813794315,-0.1105363816,0.2011599839,-0.0130072813,-0.0319682062,0.2776533663,-0.0550291725,0.0415961146,-0.0766474456,-0.01457729,-0.2234962285,-0.242813915,-0.0475633405,-0.0204235055,-0.0823250785,-0.3094788194,0.1845414937,-0.2136323005,0.0460995547,-0.0357470661,0.0285498742,0.012782828,0.2756361663,0.3079813719,0.2152011991,0.6308562756,0.0293047745,-0.2888162732,-0.3502052724,-0.5530298948,0.1024943143,0.4333720207,0.2024979442,0.5229716301,-0.1057329923,-0.3106374741,-0.2398288101,0.6266832352,-0.1735196263,0.0534491092,-0.1775565594,0.0458979234,0.0100549022,0.3290853798,-0.0232772212,0.3484038711,-0.1739281863,0.1046599299,-0.1560463905,-0.3812662065,0.6492167115,-0.4117067158,-0.27528584,-0.2286066562,0.144055292,0.2265027463,-0.1033521369,-0.4917485416,0.1499257386,0.3424094021,0.1057307124,-0.3156130612,0.2155573815,-0.2473880053,-0.0007568364,-0.0607644804,0.5921410918,0.015126002,-0.3535507321,0.2004110068,-0.0902967453]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset reddit\/default (download: 2.93 GiB, generated: 17.64 GiB, post-processed: Unknown size, total: 20.57 GiB) to \/Volumes\/My Passport for Mac\/og-chat-data\/reddit\/default\/1.0.0\/98ba5abea674d3178f7588aa6518a5510dc0c6fa8176d9653a3546d5afcb3969...\r\nDownloading: 13%\r\n403M\/3.14G [44:39<2:27:09, 310kB\/s]\r\n---------------------------------------------------------------------------\r\ntimeout Traceback (most recent call last)\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in _error_catcher(self)\r\n 437 try:\r\n--> 438 yield\r\n 439 \r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in read(self, amt, decode_content, cache_content)\r\n 518 cache_content = False\r\n--> 519 data = self._fp.read(amt) if not fp_closed else b\"\"\r\n 520 if (\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/http\/client.py in read(self, amt)\r\n 458 b = bytearray(amt)\r\n--> 459 n = self.readinto(b)\r\n 460 return memoryview(b)[:n].tobytes()\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/http\/client.py in readinto(self, b)\r\n 502 # (for example, reading in 1k chunks)\r\n--> 503 n = self.fp.readinto(b)\r\n 504 if not n and b:\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/socket.py in readinto(self, b)\r\n 703 try:\r\n--> 704 return self._sock.recv_into(b)\r\n 705 except timeout:\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/ssl.py in recv_into(self, buffer, nbytes, flags)\r\n 1240 self.__class__)\r\n-> 1241 return self.read(nbytes, buffer)\r\n 1242 else:\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/ssl.py in read(self, len, buffer)\r\n 1098 if buffer is not None:\r\n-> 1099 return self._sslobj.read(len, buffer)\r\n 1100 else:\r\n\r\ntimeout: The read operation timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nReadTimeoutError Traceback (most recent call last)\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/requests\/models.py in generate()\r\n 757 try:\r\n--> 758 for chunk in self.raw.stream(chunk_size, decode_content=True):\r\n 759 yield chunk\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in stream(self, amt, decode_content)\r\n 575 while not is_fp_closed(self._fp):\r\n--> 576 data = self.read(amt=amt, decode_content=decode_content)\r\n 577 \r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in read(self, amt, decode_content, cache_content)\r\n 540 # Content-Length are caught.\r\n--> 541 raise IncompleteRead(self._fp_bytes_read, self.length_remaining)\r\n 542 \r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/contextlib.py in __exit__(self, type, value, traceback)\r\n 134 try:\r\n--> 135 self.gen.throw(type, value, traceback)\r\n 136 except StopIteration as exc:\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in _error_catcher(self)\r\n 442 # there is yet no clean way to get at it from this context.\r\n--> 443 raise ReadTimeoutError(self._pool, None, \"Read timed out.\")\r\n 444 \r\n\r\nReadTimeoutError: HTTPSConnectionPool(host='zenodo.org', port=443): Read timed out.\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n\/var\/folders\/3f\/md0t9sgj6rz8xy01fskttqdc0000gn\/T\/ipykernel_89016\/1133441872.py in \r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 845 \r\n 846 # Download and prepare data\r\n--> 847 builder_instance.download_and_prepare(\r\n 848 download_config=download_config,\r\n 849 download_mode=download_mode,\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 613 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 614 if not downloaded_from_gcs:\r\n--> 615 self._download_and_prepare(\r\n 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 617 )\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 669 split_dict = SplitDict(dataset_name=self.name)\r\n 670 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 671 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 672 \r\n 673 # Checksums verification\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/reddit\/98ba5abea674d3178f7588aa6518a5510dc0c6fa8176d9653a3546d5afcb3969\/reddit.py in _split_generators(self, dl_manager)\r\n 73 def _split_generators(self, dl_manager):\r\n 74 \"\"\"Returns SplitGenerators.\"\"\"\r\n---> 75 dl_path = dl_manager.download_and_extract(_URL)\r\n 76 return [\r\n 77 datasets.SplitGenerator(\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py in download_and_extract(self, url_or_urls)\r\n 287 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 288 \"\"\"\r\n--> 289 return self.extract(self.download(url_or_urls))\r\n 290 \r\n 291 def get_recorded_sizes_checksums(self):\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py in download(self, url_or_urls)\r\n 195 \r\n 196 start_time = datetime.now()\r\n--> 197 downloaded_path_or_paths = map_nested(\r\n 198 download_func,\r\n 199 url_or_urls,\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 194 # Singleton\r\n 195 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 196 return function(data_struct)\r\n 197 \r\n 198 disable_tqdm = bool(logger.getEffectiveLevel() > logging.INFO) or not utils.is_progress_bar_enabled()\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py in _download(self, url_or_filename, download_config)\r\n 218 # append the relative path to the base_path\r\n 219 url_or_filename = url_or_path_join(self._base_path, url_or_filename)\r\n--> 220 return cached_path(url_or_filename, download_config=download_config)\r\n 221 \r\n 222 def iter_archive(self, path):\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 286 if is_remote_url(url_or_filename):\r\n 287 # URL, so get it from the cache (downloading if necessary)\r\n--> 288 output_path = get_from_cache(\r\n 289 url_or_filename,\r\n 290 cache_dir=cache_dir,\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 643 ftp_get(url, temp_file)\r\n 644 else:\r\n--> 645 http_get(\r\n 646 url,\r\n 647 temp_file,\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py in http_get(url, temp_file, proxies, resume_size, headers, cookies, timeout, max_retries)\r\n 451 disable=bool(logging.get_verbosity() == logging.NOTSET),\r\n 452 )\r\n--> 453 for chunk in response.iter_content(chunk_size=1024):\r\n 454 if chunk: # filter out keep-alive new chunks\r\n 455 progress.update(len(chunk))\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/requests\/models.py in generate()\r\n 763 raise ContentDecodingError(e)\r\n 764 except ReadTimeoutError as e:\r\n--> 765 raise ConnectionError(e)\r\n 766 else:\r\n 767 # Standard file-like object.\r\n\r\nConnectionError: HTTPSConnectionPool(host='zenodo.org', port=443): Read timed out.\r\n```","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":646,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset reddit\/default (download: 2.93 GiB, generated: 17.64 GiB, post-processed: Unknown size, total: 20.57 GiB) to \/Volumes\/My Passport for Mac\/og-chat-data\/reddit\/default\/1.0.0\/98ba5abea674d3178f7588aa6518a5510dc0c6fa8176d9653a3546d5afcb3969...\r\nDownloading: 13%\r\n403M\/3.14G [44:39<2:27:09, 310kB\/s]\r\n---------------------------------------------------------------------------\r\ntimeout Traceback (most recent call last)\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in _error_catcher(self)\r\n 437 try:\r\n--> 438 yield\r\n 439 \r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in read(self, amt, decode_content, cache_content)\r\n 518 cache_content = False\r\n--> 519 data = self._fp.read(amt) if not fp_closed else b\"\"\r\n 520 if (\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/http\/client.py in read(self, amt)\r\n 458 b = bytearray(amt)\r\n--> 459 n = self.readinto(b)\r\n 460 return memoryview(b)[:n].tobytes()\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/http\/client.py in readinto(self, b)\r\n 502 # (for example, reading in 1k chunks)\r\n--> 503 n = self.fp.readinto(b)\r\n 504 if not n and b:\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/socket.py in readinto(self, b)\r\n 703 try:\r\n--> 704 return self._sock.recv_into(b)\r\n 705 except timeout:\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/ssl.py in recv_into(self, buffer, nbytes, flags)\r\n 1240 self.__class__)\r\n-> 1241 return self.read(nbytes, buffer)\r\n 1242 else:\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/ssl.py in read(self, len, buffer)\r\n 1098 if buffer is not None:\r\n-> 1099 return self._sslobj.read(len, buffer)\r\n 1100 else:\r\n\r\ntimeout: The read operation timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nReadTimeoutError Traceback (most recent call last)\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/requests\/models.py in generate()\r\n 757 try:\r\n--> 758 for chunk in self.raw.stream(chunk_size, decode_content=True):\r\n 759 yield chunk\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in stream(self, amt, decode_content)\r\n 575 while not is_fp_closed(self._fp):\r\n--> 576 data = self.read(amt=amt, decode_content=decode_content)\r\n 577 \r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in read(self, amt, decode_content, cache_content)\r\n 540 # Content-Length are caught.\r\n--> 541 raise IncompleteRead(self._fp_bytes_read, self.length_remaining)\r\n 542 \r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/contextlib.py in __exit__(self, type, value, traceback)\r\n 134 try:\r\n--> 135 self.gen.throw(type, value, traceback)\r\n 136 except StopIteration as exc:\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/urllib3\/response.py in _error_catcher(self)\r\n 442 # there is yet no clean way to get at it from this context.\r\n--> 443 raise ReadTimeoutError(self._pool, None, \"Read timed out.\")\r\n 444 \r\n\r\nReadTimeoutError: HTTPSConnectionPool(host='zenodo.org', port=443): Read timed out.\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n\/var\/folders\/3f\/md0t9sgj6rz8xy01fskttqdc0000gn\/T\/ipykernel_89016\/1133441872.py in \r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 845 \r\n 846 # Download and prepare data\r\n--> 847 builder_instance.download_and_prepare(\r\n 848 download_config=download_config,\r\n 849 download_mode=download_mode,\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 613 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 614 if not downloaded_from_gcs:\r\n--> 615 self._download_and_prepare(\r\n 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 617 )\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 669 split_dict = SplitDict(dataset_name=self.name)\r\n 670 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 671 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 672 \r\n 673 # Checksums verification\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/reddit\/98ba5abea674d3178f7588aa6518a5510dc0c6fa8176d9653a3546d5afcb3969\/reddit.py in _split_generators(self, dl_manager)\r\n 73 def _split_generators(self, dl_manager):\r\n 74 \"\"\"Returns SplitGenerators.\"\"\"\r\n---> 75 dl_path = dl_manager.download_and_extract(_URL)\r\n 76 return [\r\n 77 datasets.SplitGenerator(\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py in download_and_extract(self, url_or_urls)\r\n 287 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 288 \"\"\"\r\n--> 289 return self.extract(self.download(url_or_urls))\r\n 290 \r\n 291 def get_recorded_sizes_checksums(self):\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py in download(self, url_or_urls)\r\n 195 \r\n 196 start_time = datetime.now()\r\n--> 197 downloaded_path_or_paths = map_nested(\r\n 198 download_func,\r\n 199 url_or_urls,\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 194 # Singleton\r\n 195 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 196 return function(data_struct)\r\n 197 \r\n 198 disable_tqdm = bool(logger.getEffectiveLevel() > logging.INFO) or not utils.is_progress_bar_enabled()\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py in _download(self, url_or_filename, download_config)\r\n 218 # append the relative path to the base_path\r\n 219 url_or_filename = url_or_path_join(self._base_path, url_or_filename)\r\n--> 220 return cached_path(url_or_filename, download_config=download_config)\r\n 221 \r\n 222 def iter_archive(self, path):\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 286 if is_remote_url(url_or_filename):\r\n 287 # URL, so get it from the cache (downloading if necessary)\r\n--> 288 output_path = get_from_cache(\r\n 289 url_or_filename,\r\n 290 cache_dir=cache_dir,\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 643 ftp_get(url, temp_file)\r\n 644 else:\r\n--> 645 http_get(\r\n 646 url,\r\n 647 temp_file,\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py in http_get(url, temp_file, proxies, resume_size, headers, cookies, timeout, max_retries)\r\n 451 disable=bool(logging.get_verbosity() == logging.NOTSET),\r\n 452 )\r\n--> 453 for chunk in response.iter_content(chunk_size=1024):\r\n 454 if chunk: # filter out keep-alive new chunks\r\n 455 progress.update(len(chunk))\r\n\r\n\/usr\/local\/anaconda3\/envs\/og-data-env\/lib\/python3.9\/site-packages\/requests\/models.py in generate()\r\n 763 raise ContentDecodingError(e)\r\n 764 except ReadTimeoutError as e:\r\n--> 765 raise ConnectionError(e)\r\n 766 else:\r\n 767 # Standard file-like object.\r\n\r\nConnectionError: HTTPSConnectionPool(host='zenodo.org', port=443): Read timed out.\r\n```","embeddings":[-0.3638951778,-0.1421309114,-0.0797189623,0.1200471893,0.166841507,0.1872061193,0.0386046171,0.2438804507,-0.0571571589,-0.1331617236,-0.0627646521,0.1441400796,0.3488289118,-0.0019752642,-0.0935343206,0.2313060313,0.0492559485,-0.1813283712,-0.0657732934,0.0541189387,-0.0644872263,0.1003924385,-0.1485519111,-0.1538381428,-0.1942003816,-0.0592060834,-0.1093631983,-0.1202174351,-0.2129873037,-0.2641814351,0.31763798,0.1678793132,-0.0866455138,0.5624829531,-0.0001095642,-0.2061458379,0.3377981782,0.0788483992,-0.4498393536,-0.4648267329,-0.261334002,-0.1185770333,0.0888731331,0.049372673,0.0792772025,0.1067883223,0.0705106482,-0.4283182025,0.317329675,0.4327323437,0.2017596811,0.178360939,0.2336223423,-0.1382245421,0.1232367456,0.0802627355,0.0134522729,0.4224596918,0.1959228814,-0.1228379458,0.2677885592,0.0269400552,-0.0859171823,0.2206049412,-0.1385429054,-0.0203245785,0.1799729913,-0.4049907625,-0.0455894917,0.1644668132,0.8760608435,-0.0498092026,-0.412339747,-0.2699545324,0.1099989936,-0.0107127959,0.3056143522,0.3553996682,-0.4313032925,0.2249759734,-0.2838400006,0.0700306892,0.0303666573,0.3651785851,0.1223645285,-0.0159724392,0.0214315131,0.2920616865,0.0948145241,-0.0455701128,0.2709402442,-0.3481400013,0.0262930784,-0.0731001273,-0.4070403874,-0.1091760322,0.0345305689,0.1573893279,0.1790525764,0.1141359657,-0.0329113863,-0.1888143271,-0.0093015349,0.1332781017,0.3627251089,-0.0514031611,0.0227710307,0.1718333662,0.2054774463,-0.0747328475,-0.0656795129,-0.0005851863,0.0896900445,-0.0134078031,0.1178109944,-0.1248832345,0.333463639,-0.2927748561,-0.264020294,-0.0683694184,0.1210783646,-0.0408359207,-0.1997187883,0.1755993515,-0.4186320901,0.1737193018,-0.0235131476,0.12073607,-0.2860050499,-0.3834685087,-0.0888734832,-0.0542666838,-0.2270246595,-0.0607817583,-0.0255941823,-0.0167295821,0.0956359059,0.1742198765,0.0805225372,-0.3122357428,-0.0038488104,-0.2295095474,-0.1563984901,0.1975986063,-0.1175193116,0.4563015997,0.0885375962,0.1118027046,-0.0920328721,-0.0471643582,-0.1970358193,-0.2452467978,-0.1614522636,0.2359225899,-0.209383145,-0.2534177899,-0.1873574108,-0.10031268,0.2682973742,0.0549779795,-0.0631795004,-0.0574919246,0.0485177338,-0.10668277,-0.0858378708,0.3686818182,-0.4268350601,0.32156232,-0.2427403778,0.1214148998,0.2209982276,0.2091517597,-0.2188085467,0.0070210267,-0.1598643512,-0.1748522222,0.2636782527,-0.2803451121,-0.6896670461,0.0776350945,-0.3067837954,0.0360194631,-0.1616358459,0.2824025154,0.3321720064,-0.1226263493,0.0851028785,0.4505003691,-0.2175913453,0.0324916169,-0.407615304,-0.1520121694,-0.0775400996,0.1463225335,-0.0387063511,0.016049223,-0.1382055879,-0.0291029382,0.2990521491,0.4490671158,0.1406909823,0.2543386519,0.09554144,0.2424472421,0.0095893964,-0.1858817786,-0.3774220943,0.1745990217,0.5166456699,0.0398234054,-0.234704867,-0.0346935764,-0.4541685879,0.0166554451,-0.0438087694,0.0101372059,0.0289865471,0.1677503139,0.288700372,0.1997802109,0.0198443532,0.6505975723,-0.0681825802,0.3121290505,-0.0864717141,0.2073747516,-0.1682162732,-0.1355968118,0.1590008885,-0.3182855546,0.1997077614,-0.1743712276,-0.1728224605,0.3266712427,0.0589874312,0.101190798,-0.0256031081,-0.2437624186,0.2322071791,-0.140318647,-0.1642286181,0.3527719676,0.10086371,0.0236945394,-0.2600862682,-0.0127559723,0.2208568305,0.1852484792,-0.2386240661,0.1970193684,0.1934301555,-0.0050591128,0.1545086652,-0.0676915795,0.2783944309,-0.2629438937,0.1345344037,-0.2053741664,0.0025805165,0.3789856434,0.3293512464,0.1056887582,-0.2198491246,0.1804495305,0.1296850443,-0.1198631078,-0.0056320135,0.5065407753,0.3067332804,0.1704119295,0.2266139984,-0.0701207072,-0.2436689585,-0.1433323473,0.1890535206,0.0757508501,0.0555560812,-0.0969582424,0.1349174827,-0.0243189335,-0.0290592741,-0.3912774026,0.1723955423,0.3068226576,-0.0378288329,0.0120927421,-0.2496584654,-0.1520566344,0.1568963826,-0.1114778221,-0.2432442009,-0.5550534725,-0.1170367375,0.2648683786,-0.0485743247,0.0361475125,-0.1683561802,-0.1456118822,0.3770270646,-0.2226378322,-0.0041139335,-0.15879637,-0.0620049685,0.0850587562,0.2173152864,-0.1189311743,0.3502157032,-0.2401047051,-0.2311211377,-0.3148239851,-0.0313548632,0.0154778501,-0.1471475661,0.3772631288,-0.228700012,0.4140965343,0.0851948634,0.1193854585,0.1587168425,-0.1645901799,-0.0706984699,-0.1050101146,0.0962466225,0.1030078828,-0.0575042479,-0.2533045411,-0.2038371563,-0.3415128887,0.1356620789,-0.1374951899,-0.0539904311,0.1529115736,0.2279939353,0.0301787183,-0.1151634753,0.0764709264,-0.1927661747,-0.3900161684,0.4926212132,-0.0878467113,-0.1976853907,0.0133778267,0.212452352,0.0070858877,0.3136509359,-0.6629764438,0.0568315722,-0.3185949028,0.2716561258,-0.0546660349,-0.2824904025,0.2686043084,-0.4917546213,-0.1053478792,0.0398719497,0.0277596097,-0.0297617838,0.1054959446,0.4257929027,-0.1738022715,0.2901479602,-0.0058969217,0.1678647697,0.2327540368,-0.0230306946,0.0852697045,0.1211807057,0.2642677128,-0.0063393824,-0.3034160435,-0.0176089443,-0.0341337211,-0.136857003,-0.0107164923,-0.0085253892,-0.2189045548,-0.1899486035,0.0047977958,-0.2691439986,-0.2852286696,-0.1181231514,-0.2087269723,0.0319991447,0.3415756524,0.2162771821,0.0480177924,-0.1770515442,-0.1402965635,-0.0034896294,-0.1222753674,-0.2203560472,-0.5491005778,0.1339450777,-0.3536432087,0.2230889201,-0.0601286404,0.3213074803,-0.080739364,0.1125659198,0.1893660724,0.0379196741,0.6117430329,-0.1809263527,0.0362371579,0.0179805923,0.2915143669,-0.4782929122,-0.0769509822,0.1523824632,0.2845622897,0.5226516724,0.0370355844,-0.3182994425,0.0222990029,0.078677617,0.2848935127,0.1279173493,-0.4066620767,0.0861360207,-0.1845363528,-0.1715354323,0.1556864232,0.0990253389,0.1399312615,-0.0063083726,-0.2252762318,0.1577490568,0.1879136413,-0.0114681767,-0.1006641537,0.1955480576,-0.2800946534,0.5696733594,0.0582194738,-0.2168166041,0.3292991817,0.6071470976,0.2037016153,-0.4783625901,-0.0077772723,-0.0352760479,-0.1191034093,0.2941103578,0.0688014701,-0.0737716928,-0.2819933891,0.3526284695,0.0820840821,0.157606706,0.2579836845,0.0414124168,-0.4130149186,-0.4910191298,0.1987432241,0.0407051407,0.0077407714,0.7631573677,-0.0821264759,0.1172933951,0.2169814259,0.0036683986,1.0106487274,-0.1849299222,0.0555885024,0.0269521791,-0.1968937367,0.4470204413,-0.2444500625,0.0533534288,-0.0744405091,0.0985226557,0.0790510774,-0.0900908038,-0.1041536033,0.249537006,-0.2677375376,0.0934129357,-0.1866249442,0.1198794767,-0.106873922,0.1602323949,-0.3221085668,-0.0493133478,-0.073877573,0.2528643012,-0.0190036763,0.2423348874,0.0503329895,0.1021399796,-0.2165174931,-0.0511650369,-0.6515235305,0.060792096,-0.1059935763,-0.0249388851,-0.3161874413,-0.2694265842,-0.039735727,0.0647982806,0.1809837967,0.2945515513,-0.1085293517,0.2734043598,-0.161736235,-0.0828538686,0.0070874556,-0.25620386,0.1700795889,0.0257484168,-0.4043705463,0.1728888899,-0.0120191434,-0.1925395578,0.2453389317,0.245125249,0.1537724733,-0.1405610144,0.0705392212,0.0044918926,0.1839322299,-0.3315316439,0.1646397859,0.2733990252,0.1160796955,0.293476671,-0.1306786388,-0.3417874575,-0.312403053,0.420914948,0.0147224758,-0.0182725787,0.2777169347,0.1946513504,-0.3831721246,-0.3084495962,0.0483951569,-0.5737427473,-0.3759396374,0.27087605,-0.4809353948,0.3688181043,-0.0525627024,-0.1945745945,0.1614543945,0.0410522744,0.1765296459,-0.5769954324,-0.0723821744,-0.1342255026,-0.3124864399,0.0035801211,-0.1218181923,-0.0174361859,0.2209631354,-0.0283059459,-0.3059004247,0.1072414219,-0.2990958393,0.1666987836,0.3630222976,-0.1178333908,0.3600859046,-0.1476009041,0.0719487518,0.2135466933,-0.1259144396,-0.1867566407,0.008877093,0.1204207763,-0.0983191803,-0.1987834275,0.1021670327,-0.4653491676,-0.1139157265,0.0836558416,-0.1327847391,0.1149014458,0.0751330703,-0.2157939076,0.1564480662,0.0608537719,0.0319545716,0.1744048297,-0.081452854,0.434597373,0.2335431874,0.3212583363,0.1280277222,0.0476791412,-0.5013427734,-0.0021893682,-0.1555356383,0.0086164866,0.5000967979,-0.2849594355,0.0574563406,0.3465841115,0.358238101,0.5142099857,0.0217753146,-0.1660113037,0.3493447304,0.2776166797,-0.0072569977,-0.0931516886,0.31616503,0.1954575479,0.0127301645,0.1659449488,-0.0286612287,-0.1532746255,0.1752324998,0.2132846713,0.7456157207,0.059435267,0.2120887935,0.2359481901,0.1325783432,-0.0256505627,0.4962303042,0.2893052995,0.0648460463,0.1836969107,-0.1896064878,-0.0699173361,-0.0696899593,0.1187767163,-0.0925604329,-0.2562949061,0.0800467879,0.1729655415,-0.2709694207,-0.0536396429,0.1682775766,0.1517081857,-0.2412230521,-0.0526957028,-0.1684112102,0.2435312122,0.0784868076,0.2163597792,0.1559387296,-0.1166476682,-0.1253601015,0.0594988652,0.0061543956,-0.2609735727,0.4184811115,0.1232506558,0.0963147506,-0.5866312981,-0.0014923572,-0.0868233368,0.1881744117,-0.474390924,0.0474524312,0.0180923548,-0.2157314867,-0.0612208322,0.2576412857,0.5287145972,0.157876119,-0.0736208335,0.1887330711,-0.0066006649,-0.0833327174,-0.0630711988,0.4645082355,0.089026697,-0.0761165842,0.4138446152,0.1446330249,-0.2284878492,0.1453460157,0.1089742184,0.2862819731,-0.2727353573,-0.0084777623,-0.0594219491,-0.1286308616,-0.3655264676,-0.1453654021,-0.1309562474,-0.0148780029,0.4429959059,-0.3660323322,-0.0384186283,-0.265814513,0.1107900962,-0.1612299383,0.3122112453,0.4844064713,0.2282981426,-0.1740746796,-0.3672617972,-0.38946262,0.0742105469,-0.6396659017,-0.150228858,-0.1438748389,0.5153074861,0.1265914589,0.3481556773,0.3572769761,0.0897677764,-0.4011443853,0.4775125086,-0.2764579952,-0.376594305,-0.0509812124,0.0072242636,0.0850610882,-0.3032618165,0.1778670996,-0.2237764895,0.0236348938,-0.1377318203,-0.2066952288,0.2231236398,-0.1374203265,0.4727042019,0.0809809789,0.2350996286,0.2856599391,-0.5054975748,-0.4817369282,-0.2052441388,0.0052701994,0.0731386915,0.0634792745,0.3176403642,-0.3422477543,0.0149215916,-0.1203675345,0.1972674429,-0.1143393368,-0.2799225152,0.2240277082,-0.2387311459,-0.0295138694,0.3290416002,-0.0437298194,0.1548280567,-0.0523173921,0.1966982484,-0.2919937968,-0.344068259,0.2796042562,0.0334829576,-0.4112816751,0.0386853963,0.1673302054,0.1059004515,-0.0483771563,-0.2123416066,0.0996661484,0.4178960025,-0.0632095933,-0.2242977172,0.2084812075,0.0545513555,0.158358708,-0.0366299339,0.4831286371,0.0268595964,-0.2417451888,0.3183769584,-0.1463124752]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"Hey @lhoestq should I try to fix this issue ?","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":10,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\nHey @lhoestq should I try to fix this issue ?","embeddings":[-0.3229391873,-0.1452181786,-0.0921981707,0.1495675594,0.1223356798,0.2118237019,0.0725628659,0.1619715393,-0.0042136936,-0.179710865,-0.002579777,0.1080995128,0.2991375327,-0.0202094316,-0.0535588674,0.2257830948,0.0980871841,-0.2138081789,-0.0274107065,0.0541314632,-0.0396531634,0.0795041919,-0.1696907878,-0.1037186161,-0.1970556229,-0.0834876448,-0.0887065902,-0.0881419256,-0.1635095179,-0.22556822,0.2328706533,0.1906133443,-0.08636415,0.5323535204,-0.0001098788,-0.1932763457,0.3116383553,0.0794507042,-0.3774964809,-0.4368223846,-0.2054940462,-0.1277656853,0.0653153136,0.0371546894,0.1271589845,0.1339381188,0.0613027476,-0.3696908653,0.2714517117,0.3567239344,0.200300768,0.3396145403,0.2257603556,-0.0364479944,0.1874397248,0.0483288169,0.0738473982,0.4294619858,0.1973480731,-0.1066404283,0.2583864331,0.0646374077,-0.1005405933,0.227908805,-0.0676368922,0.0118788928,0.1615293175,-0.4194114208,-0.0912717655,0.12934196,0.8974468708,-0.006717965,-0.4444300532,-0.2724119425,0.1446119845,0.0279446673,0.3546218574,0.3928590417,-0.3586822152,0.174208343,-0.2515689135,-0.0596538298,0.0619719066,0.3606230915,0.0795979351,-0.0586893335,0.0543124266,0.3006879389,0.1167341247,-0.0253120344,0.2730750144,-0.3923239708,0.0315246359,-0.120367229,-0.4185859561,-0.0751451105,-0.008666913,0.256883353,0.1549490988,0.0281615146,-0.0092111798,-0.2176924199,-0.1296387762,0.1286874264,0.4609533846,-0.0846338123,-0.021766521,0.1847823262,0.1996580064,-0.1273577213,-0.0499224514,0.0622118078,0.0676101968,-0.0569905229,0.1174766719,-0.1309590936,0.2997891903,-0.2239587754,-0.2243892252,0.0199308563,0.1070002392,-0.116439648,-0.1968941987,0.15718095,-0.5001404285,0.282037735,-0.0591908023,0.0897831023,-0.3374088109,-0.46206972,-0.0781457648,0.0293507595,-0.2490952313,-0.0732373744,-0.1214543208,-0.0279665366,-0.0205290858,0.1493876725,0.0389321297,-0.3205437064,-0.0554194562,-0.2115112245,-0.1496171951,0.1463934332,-0.1507051736,0.4690842032,0.0205134004,0.1644591391,-0.0645269603,0.0455471426,-0.305152297,-0.2255772352,-0.0837342441,0.2294354439,-0.2565513849,-0.306913197,-0.227144137,-0.0992048904,0.2145469189,-0.0132150343,-0.090332374,-0.0714858845,0.1028874218,-0.093312256,-0.1204740852,0.3781053722,-0.4907187521,0.3337835073,-0.2151438892,0.1045017093,0.2559295595,0.2922604382,-0.2118861079,0.0425635427,-0.1975020468,-0.1300389767,0.3101243079,-0.2202607542,-0.6869509816,0.0718825459,-0.3223353028,0.006946607,-0.1906253397,0.3404941261,0.266295135,-0.1222445592,0.0441370681,0.356422931,-0.2246155441,0.0489984639,-0.4274553061,-0.1760314256,0.0863730535,0.2211072296,0.006440192,-0.0156055931,-0.2201594114,-0.0154925939,0.3019465506,0.4858202934,0.1000773534,0.2127553523,0.0066641881,0.2881125808,-0.0434355289,-0.1862495989,-0.3396910727,0.1394865811,0.5685458183,0.0362259485,-0.2001799494,-0.0432874374,-0.3849168122,0.0472230278,-0.0296893194,0.0944157541,0.0103548365,0.1204789206,0.2333741188,0.2488060594,-0.0068094316,0.5920253992,-0.036432527,0.2961692214,-0.1134766787,0.2655540705,-0.190109387,-0.0758355632,0.1585974842,-0.3358879983,0.1565392315,-0.1582795233,-0.1301783472,0.2735829949,0.0461659357,0.1550990939,-0.0088297436,-0.2064001113,0.3236560822,-0.1493874788,-0.2201239616,0.3200768232,0.1316653639,0.015926199,-0.2117775232,-0.0468312874,0.2223205268,0.1811868399,-0.2609057426,0.132086724,0.0904310122,0.0356036313,0.1487475634,-0.0297680236,0.2846351564,-0.2747428119,0.1233049855,-0.1726231426,-0.0406183302,0.3562107682,0.3532289863,0.0572855696,-0.2207048088,0.2090945393,0.1724737883,-0.1181943715,0.0209374651,0.3768256605,0.2726873457,0.1829893142,0.1817195117,-0.1001244262,-0.1576252729,-0.1328181773,0.1219856739,0.1303930283,-0.0202917829,-0.0710608289,0.0974666849,-0.0203403346,-0.0277200937,-0.3598120511,0.1830209792,0.3374426365,-0.0684072822,-0.0071426779,-0.219995141,-0.125729382,0.2173401713,-0.0961502865,-0.3208750486,-0.5690428019,-0.1160383895,0.2196944803,-0.087720044,0.0540348291,-0.1596696228,-0.1464293897,0.3939836621,-0.2068844587,0.0193288978,-0.1703544855,-0.0465683155,0.0981427133,0.2302041203,-0.1257002652,0.3081567883,-0.2789928913,-0.251839608,-0.4589740932,0.0036427723,0.0580042452,-0.1739594638,0.3086137176,-0.2075053453,0.4488535225,0.019191036,0.149183616,0.1537816525,-0.2346425205,-0.0090575581,-0.1197700351,0.0833665654,0.0977742076,0.0082746847,-0.2852861583,-0.2616789639,-0.3184665442,0.1394189298,-0.152561754,-0.0456795469,0.0721958652,0.1570874304,-0.0095705763,-0.0869150683,0.064859353,-0.2259548903,-0.4070048034,0.4899906218,-0.062106099,-0.2405166924,0.0764892027,0.179158181,-0.0469040312,0.2807112038,-0.7032278776,0.0868571252,-0.2939401269,0.2561586499,-0.1028992981,-0.2088770419,0.2067396194,-0.5187262893,-0.1010919139,0.0839680731,0.0477390513,-0.0127560738,0.1357416064,0.4230009019,-0.0845860466,0.2549438775,-0.088434495,0.1675830483,0.1658005863,0.0040437453,0.065840371,0.0913862064,0.246755451,0.0516305156,-0.2601689696,-0.00323467,-0.0192011502,-0.1915653795,0.0363776498,-0.0518689379,-0.1829180568,-0.2068374455,-0.0555759184,-0.3045925796,-0.3400904238,-0.1228976175,-0.1587321013,0.0118439207,0.2988443971,0.2148422152,0.0509686619,-0.2290703952,-0.1046523228,0.0603385456,-0.021730233,-0.2436377406,-0.5250362158,0.1484563649,-0.3766491711,0.2041376382,-0.0775991604,0.3665100038,-0.0565216877,0.1079761609,0.1723058671,-0.0322542787,0.6097729802,-0.1568298191,0.0541811474,-0.0678962842,0.3280013502,-0.4298905134,-0.0914605707,0.1365546137,0.2551421225,0.4804638624,-0.1013379097,-0.3312020004,0.0573515892,0.0944498926,0.2967836559,0.1135168746,-0.481150806,0.0424285308,-0.1137489304,-0.1290588826,0.1443316787,0.1037494838,0.0814013109,0.0248103682,-0.1911548376,0.192555666,0.1372222304,0.0948434621,-0.1461683065,0.1899271756,-0.2857371569,0.5257727504,0.0884392038,-0.2245810628,0.3064742386,0.6800438762,0.2292501926,-0.4589339495,0.036740467,-0.0574066974,-0.1117961332,0.3125545979,0.052880805,-0.0370640457,-0.1722669452,0.3471660614,0.0377560034,0.1671719551,0.2752718329,0.0793982893,-0.4532320499,-0.5117076039,0.1645373255,0.074265644,0.0170164183,0.7101881504,-0.1010909602,0.1426479518,0.1844957173,-0.0353070945,1.0351309776,-0.1938214451,0.1246707439,0.0243207552,-0.0986221209,0.4625009894,-0.3301754892,0.0114303054,-0.0694777668,0.0947039053,0.0423637666,-0.0789257586,-0.1305075139,0.3125604689,-0.2249970287,0.1553881764,-0.1651517749,0.1678152531,-0.1734104007,0.2039464414,-0.3205725551,0.0056592813,-0.1161248982,0.2229984552,-0.0354014486,0.2604043782,0.0299671963,0.1112598255,-0.1971564889,-0.031650845,-0.6402619481,-0.00259282,-0.1099351346,-0.0782934725,-0.3622608781,-0.2781798244,0.0035796701,0.1153822467,0.1141586676,0.3081608713,-0.0788476318,0.2729528844,-0.1686133891,-0.0603758544,-0.0299991965,-0.2016790956,0.1367286593,0.018560363,-0.3911022246,0.2405423522,-0.0481391177,-0.16942361,0.1429902017,0.2913776934,0.2220751345,-0.1549342722,0.0112299975,-0.028955847,0.1595211178,-0.2494760454,0.1533636153,0.3502978086,0.1950357854,0.2519264519,-0.1441348791,-0.2753319442,-0.3875776231,0.4191140532,0.0290134195,-0.0197356604,0.2464025319,0.2482145578,-0.4117497206,-0.2951807976,0.0821789652,-0.5475595593,-0.368789345,0.2303786129,-0.4228589535,0.4171248376,-0.0713559464,-0.1298573762,0.2080071718,0.0426398106,0.2178039253,-0.5665858984,-0.0661822036,-0.0884595737,-0.2654107809,0.0044813505,-0.0545442179,-0.0669877306,0.2125707865,-0.1089632362,-0.2965841889,0.055455137,-0.2796338797,0.0943910107,0.3182209432,-0.0895007625,0.3497310877,-0.1654298753,0.0933734253,0.1705220044,-0.1360297948,-0.1659141183,0.0456219055,0.1349406391,-0.0579168871,-0.1680571437,0.1392920464,-0.4407623112,-0.1025438681,0.1346337497,-0.0803185105,0.1397664845,0.0939316973,-0.1928618997,0.2093704343,0.0010754223,-0.0575286187,0.2206061333,-0.0847280696,0.5286200047,0.2150290161,0.2474509776,0.2180312872,0.0778279155,-0.5206773281,-0.0415005535,-0.1290927678,0.0376551449,0.4728071392,-0.3343703747,0.0954352245,0.3918638527,0.4231960475,0.5649775863,0.0012741482,-0.1750257462,0.3973003328,0.2857113183,-0.0150063839,-0.1167183593,0.3022102416,0.1539904922,0.0074652908,0.1613985896,-0.056965135,-0.0663776025,0.1434796602,0.2579844594,0.7116305232,0.0273758192,0.1998497397,0.1504392028,0.0969894379,0.025205316,0.4835990667,0.2878908217,-0.0670275092,0.102397576,-0.1739576757,-0.1342239529,0.0182164498,0.193296954,-0.0830386803,-0.3136624098,0.0416437984,0.1705549955,-0.2826194465,-0.0389012434,0.1557422578,0.1761849076,-0.2286928892,-0.1233062521,-0.1451182663,0.2653869987,0.0769547075,0.2881632447,0.2690483034,-0.1302958131,-0.1052541882,0.0734287202,0.0191805456,-0.2896296382,0.3290356398,0.232155323,0.1606858969,-0.5691657066,0.0163570754,-0.1572817415,0.2107225657,-0.4413031042,0.0410056598,-0.0222074501,-0.2414121926,0.0235276502,0.1943901181,0.4988883138,0.1826110333,-0.0524717979,0.087746948,-0.0110522183,-0.0255389456,-0.0114667704,0.4648129642,0.0792378485,-0.0999821275,0.4754454494,0.1392788589,-0.2335128486,0.1607190669,0.0703763887,0.2371118069,-0.3442747295,0.0006955045,-0.0665494725,-0.1551597565,-0.40171206,-0.1928277165,-0.1895328462,-0.0067093889,0.4390450418,-0.3304584026,-0.0334252529,-0.2754209936,0.1056854203,-0.1803042889,0.3488104045,0.5027753115,0.3127340078,-0.1888481975,-0.2689041495,-0.4547498822,0.1023759246,-0.6168171763,-0.1967224926,-0.0962870717,0.4458437562,0.1235105693,0.4042377472,0.3505080342,0.0551721789,-0.4405975938,0.4939446151,-0.2874500155,-0.349255532,-0.0491150729,0.0317249484,0.049667675,-0.2563816011,0.2067451924,-0.2299333811,-0.0044142115,-0.1775527447,-0.2214927375,0.1781205386,-0.0859958157,0.5094978213,0.0619645789,0.3038068712,0.2616734207,-0.5011414289,-0.4645992815,-0.2597398162,0.0028752238,0.0794362575,0.0338151827,0.3148884475,-0.328353107,0.0462092161,-0.178184092,0.1587014198,-0.1035756469,-0.2572789192,0.1626215428,-0.2424263954,0.0564552881,0.341750592,0.0046893149,0.1724648029,-0.015336114,0.1752430946,-0.2748341262,-0.3486074507,0.2796503007,0.0278691053,-0.3510866761,0.049683854,0.1401320994,0.1677316725,-0.0278902538,-0.2746706009,0.0792034268,0.4314221442,-0.0492156558,-0.1817262322,0.190356791,0.0525109842,0.1493711323,-0.0255442616,0.4684515595,-0.0271999557,-0.3019453287,0.3573344648,-0.1362569481]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"It also doesn't seem to be \"smart caching\" and I received an error about a file not being found...","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":19,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\nIt also doesn't seem to be \"smart caching\" and I received an error about a file not being found...","embeddings":[-0.3514554501,-0.1771149635,-0.0796713531,0.1682872921,0.1475948989,0.223283723,0.0226032361,0.2211398929,0.0123542016,-0.1897378862,-0.0006052735,0.0798814148,0.2927660644,-0.0234991312,0.004596815,0.2605564594,0.1746075749,-0.1837051511,0.0295606758,0.0212488603,0.0039186464,0.1042710021,-0.2045951933,-0.189121902,-0.2868826687,-0.0632349402,-0.1026237905,-0.0651948974,-0.1613499969,-0.2562071085,0.2678031325,0.1747539937,-0.1125701368,0.5722392797,-0.0001108744,-0.200259611,0.3267357647,0.0606783293,-0.3902634084,-0.4879054725,-0.2514712811,-0.1367273331,0.0648576394,0.0613183603,0.1100816876,0.0650017411,0.0765786842,-0.3915479481,0.2778304815,0.4178677201,0.1937408149,0.3087404072,0.1997336745,-0.1075381562,0.2146615088,-0.0059832106,0.0433712266,0.3852337301,0.1850037426,-0.2150467038,0.2256681919,0.0387121551,-0.0873250589,0.2124781311,-0.0270646233,0.0023178114,0.0624300987,-0.4295829535,-0.0266268235,0.1627349108,0.9098489881,0.0155012272,-0.4210056663,-0.2765198648,0.0643732026,0.0191605408,0.4107702971,0.3963220119,-0.3528384864,0.2195534706,-0.2865401804,0.0169617962,0.0585765578,0.3455071747,0.1300351769,-0.0365858451,0.0514852069,0.2121029496,0.077693738,-0.0127777942,0.318225354,-0.3606897593,0.0638849661,-0.0454120524,-0.4223773479,-0.059610758,0.041024752,0.2965488732,0.2180076838,0.0376025513,0.0090417573,-0.2331798077,-0.1563536525,0.1810906082,0.414400667,0.0328045376,-0.0863438174,0.1566242725,0.2490015477,-0.1563360095,-0.0637261122,0.0281155314,0.0535564683,-0.0046373708,0.2060956359,-0.160881415,0.2642780542,-0.2941417098,-0.2341362834,-0.0317058675,0.1319109648,-0.1088596955,-0.2119629532,0.2084040791,-0.5190799236,0.2473203689,-0.1339633018,0.1147477552,-0.3465745449,-0.3769268692,-0.0995051712,0.047918506,-0.2433881313,-0.0284864716,-0.0462480299,-0.0000035711,0.0765833184,0.2044056356,0.0176847167,-0.3638130426,-0.0099054035,-0.2387177795,-0.1129431501,0.1991788596,-0.151095584,0.4225296676,0.0376464613,0.1037934422,-0.0805965811,-0.0162504371,-0.3319402337,-0.2365739495,-0.1203527078,0.2128539681,-0.3113930821,-0.3375158906,-0.1700793654,-0.0349231586,0.2402767092,-0.0020042018,-0.0957966819,-0.0359608307,0.0914469063,-0.1080953255,-0.0803297758,0.417519629,-0.4671492875,0.2881300449,-0.1859707385,0.1363562047,0.1721723527,0.2570499778,-0.1467755288,0.0454530977,-0.2080725729,-0.1775383949,0.2739038169,-0.1860808581,-0.7467337847,0.057802964,-0.2843503654,0.0424199812,-0.1178643554,0.369484961,0.2407179475,-0.1426741034,0.0391038395,0.4280474484,-0.2491465658,0.0049372939,-0.393066287,-0.2006654292,0.023305757,0.1867526621,0.0173271932,0.0746781379,-0.1123708561,-0.0463864729,0.2413773984,0.42074278,0.1031232849,0.2581168711,0.027649628,0.3117121458,0.034169782,-0.2208822668,-0.4634932578,0.1929716021,0.4974279106,-0.0402292684,-0.2256632,-0.0038947146,-0.3863794506,-0.0058577508,-0.0664516315,0.0201453753,0.0067058112,0.1483602524,0.2730265558,0.2620780468,-0.0428324305,0.6063830853,0.0634972528,0.2556881607,-0.1081701815,0.2015016973,-0.1358214915,-0.1573535502,0.1642462015,-0.331902653,0.2064648718,-0.1426869333,-0.1712403446,0.3065044582,0.0427156873,0.1618183255,0.0114986664,-0.1000133455,0.2866928577,-0.143895641,-0.1582967639,0.3258193433,0.1136549339,0.0031757897,-0.2952041924,0.017541334,0.160854429,0.1463085264,-0.2422055751,0.128293097,0.118898876,0.0218801927,0.2346511185,-0.0179281887,0.3345374763,-0.281360507,0.1407999247,-0.1899896413,-0.0198666658,0.3656054139,0.3421819508,0.1151358187,-0.1574933529,0.1737980098,0.1267798096,-0.1751209199,-0.0009873186,0.4331418276,0.2764455974,0.1672484577,0.2333202958,-0.0859269425,-0.1909573674,-0.163375169,0.1053178534,0.0348649882,-0.0363411382,-0.0639247745,0.1348118186,-0.0042759427,-0.0754822642,-0.3622912169,0.1658114791,0.2567993999,-0.0312261395,0.0473792218,-0.2350285798,-0.0895154327,0.1797646433,-0.0626701862,-0.3132677972,-0.534617722,-0.0570352897,0.2572369874,-0.0239659008,0.0314369202,-0.2107655704,-0.0662339777,0.4184938669,-0.2246605158,-0.0019356724,-0.1966605633,-0.0031246259,0.0666251704,0.2288979292,-0.1572305411,0.2978439629,-0.2575805485,-0.2320070267,-0.4382397234,0.010009652,0.0240634885,-0.1659139246,0.3262791634,-0.2745301723,0.4014875293,0.0467613973,0.184700042,0.1293170452,-0.215952307,-0.0568516403,-0.0901853219,0.1098197922,0.0916875973,0.0130148195,-0.2832642794,-0.2611865997,-0.3255770802,0.1303075105,-0.1285695136,-0.05958331,0.0994416028,0.2131880373,-0.0286169928,-0.1400176436,0.1327975541,-0.2126912624,-0.4661632776,0.574690938,-0.1025391966,-0.2443284094,0.0728578046,0.2274207771,-0.0849176496,0.3090695441,-0.730183363,0.0651278421,-0.3316624165,0.2532573938,-0.0661315992,-0.1861431897,0.2206130922,-0.4908768535,-0.0971385464,0.0012669493,0.036649555,-0.0588362068,0.0865296572,0.430654943,-0.0872969702,0.2449732572,-0.0117747048,0.2368469834,0.1969712526,0.0393538512,0.1359301209,0.1315881014,0.2887610793,0.038164176,-0.1975024194,0.005636089,-0.0781105012,-0.1713756174,0.0641343966,0.0207211263,-0.2424149513,-0.1966176182,-0.0034133128,-0.2802681029,-0.2028690279,-0.1154053956,-0.2398322076,-0.0162449833,0.3200686872,0.2180338204,0.0189306457,-0.2698965669,-0.0365622826,0.0151882265,-0.0470498092,-0.1482366025,-0.5069867969,0.1340525299,-0.3120477796,0.2034245431,-0.0954854637,0.3548944294,-0.1077156886,0.0429648906,0.155305475,0.0076523894,0.5914546847,-0.1041169018,0.0983438641,-0.071587421,0.2899121642,-0.3268386722,-0.10911268,0.1869495064,0.2797989845,0.3968087435,-0.0841598511,-0.2665337622,0.1139556617,0.0210893434,0.3152824044,0.111203298,-0.4647839367,0.0356425457,-0.0712441728,-0.1073167473,0.1279919893,0.0826845989,0.0197342448,-0.0182539765,-0.1706736982,0.1882760078,0.1699028313,0.0500942841,-0.1184466928,0.2467597872,-0.3339700401,0.479537338,0.1493544579,-0.2253588289,0.2193413675,0.7089198232,0.1852729917,-0.4553846121,0.0336596631,0.0208239816,-0.0945588127,0.2738794088,0.0477418639,-0.0804545134,-0.1090874597,0.3035274446,-0.007923591,0.147912696,0.243076086,0.0518566072,-0.4522884786,-0.5552822351,0.2607029378,0.1119510457,0.0036001226,0.7280551195,-0.1057209596,0.1578547955,0.2373596877,-0.0994788632,0.9929692745,-0.1621007472,0.1493358016,-0.0333844759,-0.0478535444,0.4863727987,-0.3597359061,0.0773661286,-0.0537776947,-0.003917024,0.0284776054,-0.1073969379,-0.2344082594,0.2651615143,-0.2373554111,0.2292892337,-0.2026773542,0.206604749,-0.129842788,0.1543676257,-0.3203789592,-0.0110400608,-0.088974297,0.2219662815,0.0260919053,0.2512620986,0.0174465142,0.1118969247,-0.2052426338,0.0056720665,-0.6313534379,0.0718435049,-0.1243227422,-0.0102214729,-0.3196527958,-0.3208350539,-0.0335325301,0.1296310425,0.1616758704,0.2259854078,-0.1288221478,0.3198728561,-0.0963644385,-0.0537404232,-0.0080448994,-0.2188044488,0.1929916888,0.0218503047,-0.3007855713,0.1726766676,-0.0692984238,-0.16043441,0.1041513458,0.2808120549,0.1768756509,-0.0942789987,0.051310651,-0.0405613743,0.1964176893,-0.2893080115,0.1639479846,0.4116243422,0.1906171143,0.2454124987,-0.1340900064,-0.2700133026,-0.4032472968,0.4275887311,0.0022110525,-0.1101713255,0.2125894576,0.2454711944,-0.4511732757,-0.2619058788,0.0425997861,-0.5402831435,-0.396068573,0.2499278784,-0.5254061818,0.3360442817,-0.083750993,-0.1635059565,0.1754437238,0.041530963,0.1881674826,-0.5935431719,-0.0810241774,-0.1003884971,-0.2837699652,0.0128185665,-0.1215308905,-0.1770421267,0.2575616539,-0.0340102091,-0.2953881323,0.0378854945,-0.2645798326,0.0987284631,0.2678186595,-0.0715711713,0.3697848916,-0.1025820673,0.0607985109,0.1411417872,-0.1148399338,-0.1495312154,0.0062993551,0.1379148215,-0.0633908659,-0.1479607522,0.0588900521,-0.4454343915,-0.0665020645,0.0915175751,-0.0156409107,0.137248829,0.0614575222,-0.2270420641,0.1766365319,-0.0205348283,-0.0309892613,0.2033306062,-0.065585494,0.4361163378,0.1910603791,0.21707201,0.153766036,0.0498855002,-0.4696024358,-0.0289687794,-0.1080414206,-0.0012210421,0.5069380403,-0.3495698273,0.0856208503,0.3764311969,0.3894277215,0.5139318705,0.0112119503,-0.1490623355,0.4427869618,0.2743194997,-0.0154407136,-0.1695469916,0.3015776277,0.1912882179,0.0245828386,0.1108362377,-0.0200664364,-0.146534428,0.1181681976,0.2566966414,0.7322998047,-0.0011011017,0.1976030916,0.0902481526,0.1754857302,0.0045162989,0.4779079258,0.3105092049,-0.0350732766,0.2082247585,-0.0762824565,-0.1176383123,0.0369771533,0.1323200017,-0.1063781008,-0.2636118829,0.0359183103,0.1957786232,-0.2781755626,-0.0142179737,0.102069132,0.1815608144,-0.2402112484,-0.0688338131,-0.1920789629,0.2047709227,0.0795007423,0.2561769783,0.3552342653,-0.1377823204,-0.0822461769,0.0817601383,0.0413426384,-0.2312030643,0.3926378787,0.1953658909,0.1165902764,-0.5178903937,0.005887107,-0.0900005028,0.2304942012,-0.4653824568,0.0823567808,-0.0151105756,-0.295053184,-0.0217591617,0.2152808011,0.529571414,0.1805810183,-0.0744957626,0.1209974959,-0.0446575768,-0.0673837513,-0.0638553873,0.505407095,0.0441162996,-0.1488218755,0.3898946345,0.114591822,-0.2024145275,0.2241458446,0.056145221,0.2569638789,-0.3091994524,-0.0556072108,-0.09218622,-0.1618699729,-0.3807839155,-0.1619996428,-0.2222255468,-0.0357230492,0.4328232408,-0.3476443887,-0.0325892903,-0.2878080606,0.1002570391,-0.131763801,0.3615846932,0.5655294657,0.3356921077,-0.1526848078,-0.2620870769,-0.4610744715,0.0843296722,-0.5949679017,-0.0758613572,-0.1460902989,0.4706688821,0.07032419,0.3799139261,0.3234637082,0.0588407703,-0.410841763,0.4678776264,-0.3202511966,-0.3522126973,-0.0815750659,0.0398679115,0.0524407141,-0.249667719,0.1971653104,-0.3352738917,0.0077865152,-0.1428929418,-0.2132218778,0.1658730656,-0.051189594,0.4950648546,0.0748207197,0.3029790819,0.2944583595,-0.4896107018,-0.5438292027,-0.2805481553,0.0079492573,-0.0352593884,0.0365844704,0.2936923504,-0.3994548619,0.0398656726,-0.1119329333,0.2174253315,-0.1450024396,-0.2861656249,0.1758307219,-0.2617860734,-0.0467304327,0.3729984462,0.0740540847,0.180900678,-0.0931034237,0.1903899312,-0.2506165802,-0.3158220649,0.235268265,-0.0230837893,-0.3548648953,0.0035197225,0.1605774611,0.1762246192,-0.0140078403,-0.1961223632,0.0897569507,0.3986312449,0.0048919539,-0.1920953244,0.1823785156,0.0494246557,0.18333368,-0.011886091,0.4769107997,0.0252300557,-0.2749190927,0.3731456697,-0.1237805635]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"To be clear, the error I get when I try to \"re-instantiate\" the download after failure is: \r\n```\r\nOSError: Cannot find data file. \r\nOriginal error:\r\n[Errno 20] Not a directory: \/.cache\/huggingface\/datasets\/downloads\/1ec12301abba4daa60eb3a90e53529b5b173296b22dc3bef3186e205c75e594c\/corpus-webis-tldr-17.json'\r\n```","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":32,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\nTo be clear, the error I get when I try to \"re-instantiate\" the download after failure is: \r\n```\r\nOSError: Cannot find data file. \r\nOriginal error:\r\n[Errno 20] Not a directory: \/.cache\/huggingface\/datasets\/downloads\/1ec12301abba4daa60eb3a90e53529b5b173296b22dc3bef3186e205c75e594c\/corpus-webis-tldr-17.json'\r\n```","embeddings":[-0.3457060754,-0.1134913117,-0.0889911205,0.1294443458,0.1474923193,0.2143732458,0.0480176024,0.1998876333,0.0338550657,-0.0955315158,-0.030202182,0.1168750599,0.2437810749,-0.0771484822,-0.0491200723,0.231423229,0.0577855818,-0.1622307599,-0.0559198335,0.0735022649,0.011059083,0.1274359077,-0.0714068413,-0.1832770556,-0.2978641093,-0.0655245036,-0.1329081357,0.0713861063,-0.2183753848,-0.3113506734,0.2569487691,0.1230735406,0.0103951171,0.6663289666,-0.0001116103,-0.1592568904,0.2437488288,0.0426406115,-0.4877470732,-0.5513868928,-0.1342119128,-0.1945220381,0.0699981824,-0.0238278359,0.1194951162,0.0872305483,0.1862618923,-0.4260832667,0.2967655361,0.4254150689,0.1753057986,0.2387185842,0.1722792983,-0.0251178443,0.1737603545,0.1074786335,0.0559896566,0.4737626314,0.1752824336,-0.1304802895,0.2195627391,-0.0209688228,-0.1029600576,0.2084075212,-0.0964718014,0.1018205136,0.0989019349,-0.4662339687,-0.0699396804,0.1454073787,0.931850791,-0.0477511846,-0.4556200504,-0.2536899745,0.0674694479,-0.0081540942,0.3597680926,0.3981687725,-0.3702106476,0.2384305149,-0.2310329676,-0.0379162356,0.0456236973,0.3663991988,0.1222335845,0.0011730166,0.0717665702,0.1935005784,0.0912001356,-0.0108822351,0.1897068471,-0.2875150144,0.0827833563,-0.0230276976,-0.3163494468,-0.0040634186,0.0655484796,0.1514580846,0.1957388222,0.0782986283,0.0556203872,-0.2048885673,-0.1919862777,0.1541327089,0.3801209629,0.0180899296,-0.031185098,0.0855460837,0.3153648973,-0.0658992603,-0.0671005026,0.0107101351,0.0720502138,0.0049376562,0.2472783923,-0.1595617533,0.3136942387,-0.1568962187,-0.2649344802,-0.0032858315,-0.0083135916,-0.1022393033,-0.1726380885,0.1798466295,-0.4610124528,0.2657701671,0.0795594454,0.0874037296,-0.373341918,-0.3142513633,-0.0771974847,0.0610377826,-0.2870959342,-0.0180023052,0.0141784763,-0.1022643745,0.1352740228,0.1678624302,0.0020395012,-0.3097982407,-0.017695101,-0.1359445453,-0.1976197511,0.2188748419,-0.1506328285,0.4216026962,0.0583742112,0.1022207737,-0.0403342545,-0.0383255184,-0.2669510841,-0.2788278162,-0.1351445317,0.2164530307,-0.2373245806,-0.2125957608,-0.1456176937,-0.1163150445,0.2005085945,-0.012750349,-0.0898415819,-0.0423631743,0.1261980236,-0.1159448773,-0.0254689995,0.4587135613,-0.5205498934,0.2254990339,-0.210218206,0.105047226,0.1674224436,0.2220856547,-0.1680978537,0.125601545,-0.3032050431,-0.1276328117,0.326361984,-0.2088168263,-0.6565337181,-0.0064632804,-0.4225059748,0.0436369553,-0.0398843214,0.3199849427,0.2116360217,-0.0621474087,0.0237390995,0.3340667784,-0.192074582,0.0146882283,-0.3758515716,-0.1495868415,-0.0532919168,0.1565191895,0.0131836329,0.0386526622,-0.0940399021,0.0106654167,0.291351825,0.3134852052,0.1198113337,0.280466944,0.0698081106,0.3056702912,0.0819984227,-0.2470279485,-0.4421099126,0.2378842086,0.505831778,0.0975984633,-0.2918375731,-0.0104449233,-0.4422037899,0.0312912799,-0.0256959517,0.0083430177,0.0235647857,0.1981825382,0.2984959185,0.2305762917,-0.0688344389,0.5827645063,-0.0630007908,0.2926049829,-0.2162380219,0.3001052439,-0.185246408,-0.1520486623,0.073565647,-0.3078400791,0.0781649947,-0.2067961991,-0.1422501504,0.342199862,-0.0093691116,0.1589464247,0.0476749353,-0.065210022,0.2794476449,-0.1224916577,-0.1327505857,0.3763543665,0.0903322697,0.0723876134,-0.2474758774,-0.0002933883,0.1602976024,0.1797787845,-0.1811097562,0.1444840431,0.1153942198,-0.0085444888,0.1329468489,-0.044731155,0.4311594963,-0.3700483143,0.1274306178,-0.2081591934,-0.0152934119,0.3702557981,0.2613644898,0.0569560155,-0.1585609466,0.170138821,0.124994725,-0.1409881264,-0.0852991566,0.3960923254,0.3165245056,0.1875201315,0.2048451602,-0.0945082903,-0.1836341172,-0.1254289597,0.1274700314,0.0232612975,-0.0187184308,0.0660415143,0.1705086082,0.0169539042,-0.1588126868,-0.4864618182,0.1523246616,0.34003371,-0.1289244741,0.0425567515,-0.258289516,-0.0811866969,0.1528228819,-0.0556136668,-0.3009979725,-0.5474972725,-0.0488037132,0.3517224491,-0.0317387059,0.0333206505,-0.168702364,-0.1326798201,0.3515864015,-0.1978571713,-0.0380522572,-0.1242396235,-0.0250989161,0.0519403741,0.2081225663,-0.1554661691,0.3499416113,-0.3303993344,-0.215082854,-0.5077780485,-0.0573956966,0.0420378633,-0.071146816,0.345436275,-0.1364248991,0.4982603192,0.051556401,0.0235980954,0.1339967996,-0.2258448005,-0.109969832,-0.0887669101,0.0489946045,0.1446504891,-0.0279510859,-0.4025306702,-0.198005408,-0.3813938498,0.1214804798,-0.1050283685,-0.0400379263,0.1701100171,0.1836489588,0.0213121362,0.0021810639,0.1341613382,-0.1613623798,-0.4538007975,0.4539409876,-0.1210563257,-0.1833670288,0.0870497376,0.1361739635,-0.0487819202,0.3200146556,-0.7256786227,0.0329235345,-0.2690056562,0.2915369272,-0.1593283266,-0.1423478276,0.2895046473,-0.403314203,-0.1086893454,0.0458066911,0.0284443609,-0.0452256687,0.1700766385,0.4130652845,-0.1201204509,0.2722564936,-0.023795791,0.3074049652,0.0999026597,-0.0174084418,0.1361951232,0.124123089,0.2862432301,-0.0135767134,-0.2499906719,-0.0356642194,-0.0534126982,-0.2116122693,0.0808124095,0.069836691,-0.1821412146,-0.1863361746,0.0072596869,-0.3295259476,-0.234416768,-0.0848799646,-0.2164970189,-0.0028232927,0.292429626,0.0991397351,-0.0508784316,-0.179363057,-0.1007219404,0.0946593881,0.0371528678,-0.1377045661,-0.5816249847,0.0662401021,-0.3708145916,0.1462101191,-0.014009634,0.2827035487,-0.0599452779,0.1073074192,0.1634768099,-0.0294181611,0.6919802427,-0.2152400017,0.0371403322,-0.0299268346,0.3537363708,-0.4249563813,-0.0459860116,0.186232239,0.3114411533,0.4815077484,-0.0239964873,-0.3255311847,0.0689761713,0.1253195852,0.3304878771,0.0509944446,-0.4549201131,0.0984178632,-0.1383912712,-0.2370644659,0.0883581638,0.0409537554,0.067175597,0.0519362427,-0.1399580687,0.2137626708,0.1497508883,0.0670159981,-0.1214296147,0.2189557403,-0.2838029563,0.5442034006,0.1434605122,-0.1760974824,0.3162301779,0.7088836432,0.1405872554,-0.4247154891,0.0590028018,0.0815856308,-0.1278009415,0.2983317077,0.1206627339,-0.1569913328,-0.1940547377,0.3669216633,-0.0035607074,0.1559975296,0.2746644914,0.0425252765,-0.4269053638,-0.4922953546,0.2004536539,0.1067945212,0.0310725905,0.6868847609,-0.0483279563,0.1477255225,0.3077605069,-0.0741992593,1.0407176018,-0.2179802358,0.1197610497,0.0413824096,-0.0214084443,0.5579091907,-0.4630913436,0.0748168528,-0.1090738401,0.0262700357,-0.0109224757,-0.15298751,-0.1878227293,0.3844901025,-0.1779046208,0.1537816972,-0.2142794281,0.1738723516,-0.1212547049,0.2222172767,-0.4063543379,-0.0682746917,-0.1278559864,0.2052882761,0.0070128446,0.2733254135,-0.0355601311,0.2405059338,-0.2309954166,-0.0035609459,-0.6009572744,0.0482944362,-0.1448885053,0.028513927,-0.4012339711,-0.3080592453,-0.1364906281,0.2661265731,0.1745536923,0.2025553733,-0.0600729436,0.2379786372,-0.1382526159,-0.1495207995,-0.021051798,-0.2287067324,0.2524911463,0.0927499384,-0.3896206319,0.1058072075,-0.1077684984,-0.1877213567,0.1917234957,0.2199915946,0.1235309169,-0.0989543051,0.0367771573,0.0433514006,0.1285727322,-0.2819080353,0.1095124707,0.3493512869,0.1423767209,0.2095102817,-0.1349360645,-0.2640582621,-0.3781038821,0.4134252071,0.0218759384,-0.0857010037,0.3029447198,0.2900459766,-0.4362707734,-0.245623529,0.037844345,-0.5051119924,-0.4053316712,0.2660213709,-0.3573107123,0.3312881589,-0.0398349985,-0.1801972836,0.1913964301,0.0492339507,0.1025515795,-0.6354929805,-0.0688726977,-0.0282220077,-0.2394969761,0.0633844286,-0.0798383579,-0.1808781177,0.130954206,-0.1610224396,-0.2876026928,0.0951898545,-0.2553077936,0.0912962779,0.3782756925,-0.104643181,0.3756402433,-0.1644013375,0.0805489421,0.1534279585,-0.1319138855,-0.1438295543,-0.0169358291,0.1424365342,-0.056654457,-0.1216360554,0.1616321206,-0.4505066574,-0.0557559431,0.0261626095,-0.0478166007,0.239534542,0.0799981281,-0.2451112568,0.1829366088,0.0039198347,-0.0365811586,0.1725434959,-0.0338566005,0.4763280153,0.2203836888,0.2932705879,0.0837580636,0.0673233271,-0.5774546862,0.0289658252,-0.1202875599,0.0637435541,0.4671874046,-0.3326795995,0.0473774225,0.3258732557,0.4628502727,0.544955194,-0.0101526445,-0.0936515406,0.3874695599,0.2539565563,0.0218380913,-0.1331838816,0.3642087281,0.1027419046,0.0447580405,0.1804490685,-0.0285880174,-0.125051558,0.1498920768,0.2051858157,0.7734168768,-0.0462942682,0.1690049171,0.1242434382,0.1203421727,0.0064934744,0.4481608272,0.2540308535,0.1060197577,0.0917703509,-0.1066582054,-0.0925568789,0.0123034744,0.1724871546,-0.1829885095,-0.3379840553,0.0398431346,0.1847477108,-0.1709378958,0.0186976455,0.1239756271,0.2378612906,-0.2728780806,-0.0608684197,-0.1533220857,0.2601199448,0.0617554151,0.2621082962,0.2569058239,-0.1947939694,-0.1001044884,0.0892075524,0.0280524176,-0.2530421615,0.2741194665,0.1652854085,0.0852826461,-0.6464281678,0.0074516474,-0.1105929241,0.1793162823,-0.4398787618,0.0459658802,0.0450883657,-0.2883310318,-0.0789096206,0.2779345512,0.5579769015,0.164406836,-0.0585424714,0.1780638993,-0.1288062185,-0.0518230386,-0.1080694795,0.3691693544,0.0417232849,-0.1889354587,0.3852721453,0.1355163455,-0.2180576622,0.1424547583,0.1456933916,0.3020739853,-0.2055168152,-0.079981409,-0.0809611231,-0.1015383527,-0.3095261455,-0.1737587303,-0.1946857572,-0.0438487716,0.3546403944,-0.336546123,-0.0382005982,-0.2726593912,0.1017242372,-0.1623690128,0.201533258,0.5906613469,0.2629623413,-0.1339448541,-0.2415398806,-0.4791153371,0.0867289677,-0.5807116628,-0.1464292407,-0.2388314456,0.4592356384,0.0935254022,0.3251010776,0.4186297953,0.0528813452,-0.3567662239,0.5118842721,-0.2942032516,-0.4408628047,-0.0702454075,0.0721712038,0.0406721458,-0.2368917763,0.0800139233,-0.3003018498,-0.0044025877,-0.1498901397,-0.1932371855,0.1880463958,-0.1040893421,0.5126104951,0.067648463,0.3033854365,0.1854617149,-0.5745764971,-0.5213245749,-0.3028706908,-0.0419492945,0.0739932433,0.0750670433,0.5065416694,-0.4804158807,-0.0385461301,-0.1881070733,0.1270098984,-0.0792131573,-0.2347914875,0.1491236687,-0.2023097277,-0.0865536779,0.3757200241,0.0314426646,0.2095029801,-0.0320796147,0.2295545042,-0.2249717116,-0.3684900403,0.2885981798,0.078787379,-0.2721015811,0.0808417723,0.1974207014,0.1026080549,-0.0161032565,-0.2055115253,0.0939313546,0.4026018083,-0.0389592089,-0.290909797,0.097097002,-0.0008468208,0.1037741527,-0.0382439606,0.4988758564,0.0005766893,-0.2682355642,0.373096168,-0.1041219458]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"Here is a new error:\r\n```\r\nConnectionError: Couldn't reach https:\/\/zenodo.org\/record\/1043504\/files\/corpus-webis-tldr-17.zip?download=1\r\n```","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":11,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\nHere is a new error:\r\n```\r\nConnectionError: Couldn't reach https:\/\/zenodo.org\/record\/1043504\/files\/corpus-webis-tldr-17.zip?download=1\r\n```","embeddings":[-0.4081909955,-0.1089026034,-0.1094463468,0.1515832692,0.1130267084,0.2142616063,0.0840727761,0.3068364561,-0.0660675541,-0.1119623184,-0.1068131104,0.0975017324,0.4113184214,-0.1342690736,-0.020653069,0.1525490731,0.0492633507,-0.1653548777,-0.0979540125,0.0594536178,-0.0448087826,0.0783080235,-0.1501569748,-0.0578017421,-0.1563995779,-0.1251861304,-0.1238110811,-0.0363061428,-0.1859293431,-0.2525510788,0.2547755241,0.2108639032,-0.0711481422,0.5528725982,-0.0001104491,-0.1708034724,0.303276062,0.0838445053,-0.3630517721,-0.5232587457,-0.3281781673,-0.2090961784,-0.0013488565,0.0142952995,0.1560145319,0.0485828333,0.1005177498,-0.359201014,0.2358051986,0.4755813777,0.1897135824,0.3804078996,0.2906827331,-0.1239369214,0.2243192345,-0.1041233465,0.0385189094,0.4120749831,0.1672575176,-0.1618065387,0.1880881339,0.0503783897,-0.1184409633,0.2123211175,-0.1259353161,0.0221845284,0.1949920803,-0.4404372275,-0.0263109915,0.2092333734,0.8917365074,0.0031863831,-0.4279877245,-0.1679777801,0.0923425108,0.0559783094,0.328299582,0.3929519355,-0.4279763997,0.2088214159,-0.2413973659,-0.0293730926,-0.0733857974,0.4148551822,0.1484912932,0.0643221661,0.1130663753,0.2836157382,0.1244029775,-0.0530418828,0.3478829563,-0.3741147518,0.0406066775,-0.0005886079,-0.3751925528,-0.1098412201,0.0257266741,0.1771164834,0.2515984774,0.0700600743,0.0581130758,-0.2245886475,-0.0209847521,0.1040591225,0.4548796117,-0.0743478984,-0.097479701,0.0804721341,0.2500208318,-0.0718758851,-0.046056062,0.0413865373,0.0376243442,-0.0067251669,0.2050542086,-0.1374206543,0.3924167752,-0.244096145,-0.1673046649,-0.0051755584,0.0620819405,-0.1462557614,-0.2499830872,0.1582598835,-0.5128635764,0.2408862263,-0.1027885452,0.1135737449,-0.3741767406,-0.3846269846,-0.0790497288,-0.0016438161,-0.1783504188,-0.077796571,-0.0413664281,-0.1577402651,0.0549152121,0.1443082839,0.0027681997,-0.3104592562,-0.0111484239,-0.2226958424,-0.1405958533,0.1680752635,-0.0305002145,0.3910940289,0.0812286064,0.1923751831,-0.1069694981,0.0074424045,-0.3014906645,-0.2636810541,-0.1507117152,0.2227704674,-0.2637268007,-0.3494066,0.0191754438,-0.0580090769,0.2642919719,-0.0498161539,-0.1509559751,-0.1041602418,0.0521344505,-0.0531885996,-0.0871585384,0.3474592566,-0.3874460757,0.2904416621,-0.17200616,0.0873992369,0.1776934862,0.2009917647,-0.2368229181,0.0680518374,-0.1696184129,-0.1373031437,0.3532176614,-0.3156419396,-0.7326756716,0.0926484093,-0.3593492806,0.0416964889,-0.1123183295,0.3382844031,0.3295727074,-0.059866555,0.0855931938,0.4176938832,-0.2770448923,0.0440625213,-0.3489669263,-0.1351130307,-0.0280524492,0.1519370526,0.0762357041,0.0203800593,-0.114074491,0.0544904247,0.3380609453,0.4503048062,0.1472552121,0.2226033211,-0.0057480838,0.2412635535,0.0233918969,-0.2662078738,-0.3647723198,0.122192122,0.5901842713,0.1320095807,-0.2266763896,0.0344343074,-0.4239556193,0.0387280546,-0.0627057031,0.1105188429,0.0376650952,0.1450965255,0.2888616025,0.2484779954,0.0210071467,0.506314218,0.0463684499,0.3348506391,-0.1375990957,0.2692663372,-0.2054700702,-0.1332794875,0.1299682111,-0.2975554764,0.1656808704,-0.1223552972,-0.179339096,0.2660220861,-0.0314551704,0.1338777095,0.1013865769,-0.0733303875,0.2483551651,-0.2689513862,-0.1393833011,0.4149530232,0.1353232563,0.0645292848,-0.1538651735,0.0028647273,0.1649567038,0.2260372043,-0.1712017804,0.1451054513,0.1743328124,0.0181002133,0.128447935,0.0369021967,0.3375707567,-0.3655296266,0.1061580777,-0.2130673081,-0.0513913147,0.3360942602,0.4006820023,0.0583098419,-0.1840437055,0.154171735,0.1229731739,-0.1374609321,-0.0140439039,0.3862383068,0.2657440305,0.1666996181,0.1802147329,-0.0937796682,-0.2758416831,-0.1425926387,0.1736548096,0.1664084792,0.0503152944,0.0324002355,0.1071993858,0.0266018566,-0.097502932,-0.461751163,0.0951284915,0.3407601714,-0.0253633615,-0.0152014382,-0.2249924392,-0.144770816,0.1276366115,-0.0736897886,-0.3270601034,-0.6042021513,-0.1006063297,0.3908727169,-0.0338944942,0.0052045002,-0.298720628,-0.1191317141,0.433755815,-0.143675223,-0.0291539337,-0.234858945,-0.0579949915,0.1087324917,0.2161482871,-0.1993200332,0.3114432096,-0.2422725707,-0.1581113786,-0.5059799552,-0.0470055006,0.0666292906,-0.1818380058,0.3266986907,-0.2115068883,0.5471208096,0.1117614955,0.1207652763,0.1914097518,-0.2455101162,-0.0643480495,-0.0501725748,-0.0135854213,0.0862809196,0.094470866,-0.3664765358,-0.2132635862,-0.3699297905,0.0576130524,-0.1652804613,0.0080882944,0.210745275,0.1907398552,0.0111657064,-0.0713911802,0.1404019594,-0.1365866512,-0.4253164232,0.5006203055,-0.0875897855,-0.28320086,0.1112621427,0.2111269981,-0.0185153522,0.2835907042,-0.6413965225,0.1639854908,-0.272982657,0.298369199,-0.1239206195,-0.1983539611,0.2312436849,-0.4095801413,-0.1077380851,0.0451735742,0.1224518418,-0.087958999,0.090821825,0.4324681461,-0.0395724475,0.274877727,0.0064265579,0.2511159182,0.1461914033,0.0493600294,0.1355623752,0.0485610031,0.2548144162,0.0600316375,-0.3394828141,-0.1025329903,-0.0665283948,-0.1578672379,0.0581114665,-0.0007291579,-0.1952277422,-0.2681167722,-0.0694631636,-0.2433260828,-0.27864182,-0.1021451503,-0.1771101952,-0.009252023,0.3756552935,0.1636719108,-0.0176344942,-0.2872249186,-0.0895479247,0.0202990118,-0.063529633,-0.1342819035,-0.5123952627,0.170778662,-0.4307094514,0.18241404,0.0116764344,0.3388959467,-0.1305282563,0.0752177015,0.1505348384,-0.0229311679,0.6054639816,-0.2180060744,0.1173634827,-0.0114462394,0.3227596283,-0.4498859346,-0.0376003422,0.0344039351,0.2764469385,0.4018400311,0.0064568436,-0.2792449296,0.012938505,0.1770111918,0.2211467624,0.1156264544,-0.4411495924,0.0534354039,-0.1756308824,-0.2104160786,0.1455452144,0.0472462699,0.1587452292,-0.0139926504,-0.2431384325,0.262219131,0.1586796343,0.0278884582,-0.0930484757,0.0996662378,-0.2298464775,0.5381723642,0.0904124528,-0.2871627808,0.3406682909,0.638374567,0.2017240822,-0.5629165769,0.0188947879,-0.0538018495,-0.13731426,0.2534309626,0.1121188924,-0.0825561807,-0.1706787348,0.3123084009,0.0455223806,0.1507003903,0.3447977901,0.1160266325,-0.4762827456,-0.434435457,0.124029167,0.0121071711,0.0236054026,0.7168967128,-0.1324732304,0.1403861195,0.1727220863,-0.0568324775,1.0651465654,-0.2242844254,0.1498897672,0.0143463993,-0.1043693796,0.4616220295,-0.2956160903,0.0695652738,-0.0709175467,-0.0536904819,0.0408597142,-0.0911600962,-0.1024529412,0.2723587453,-0.2048674673,0.1106119975,-0.1986512393,0.1972106844,-0.0402260982,0.1656456143,-0.3229209483,-0.0154002849,-0.1394459307,0.2260106802,-0.0443267412,0.2786281109,-0.0320583433,0.1081972793,-0.2420550883,-0.0644151792,-0.5881773829,0.1696118116,0.0116148647,-0.0160254315,-0.4168026447,-0.1784943491,-0.1844672412,0.1712024808,0.0568092093,0.2147735655,-0.0961229354,0.2443031967,-0.1398154944,-0.1220993325,0.0423766226,-0.1694138646,0.173820883,-0.0008121931,-0.4612194896,0.2522578239,-0.0474844947,-0.2020090967,0.092975989,0.2054017186,0.1348574013,-0.1012586728,0.1022443101,-0.0710169747,0.1302691251,-0.2776732147,0.1415476203,0.2547678351,0.1400682777,0.2160898149,-0.0827108994,-0.2813751101,-0.2953568399,0.4644492269,0.0422753878,0.0006649707,0.232094124,0.2622019649,-0.4003342986,-0.2796896994,0.1233879328,-0.5116579533,-0.3536352217,0.1877388656,-0.5340178013,0.4027259946,-0.1124906018,-0.0663741827,0.1882253587,0.0270991381,0.1535607427,-0.5865340233,-0.0377112217,-0.074785687,-0.1805466115,-0.0085416446,-0.0678743273,-0.1240791455,0.1911657304,-0.0964972824,-0.3055936098,-0.0051101046,-0.2788057327,0.1170172691,0.3360910714,-0.0872067139,0.3655140698,-0.0797903687,0.0820945278,0.1091827601,-0.1335918158,-0.1511659175,-0.0178726856,0.1296094656,-0.0388030373,-0.2033949792,0.1259355247,-0.4660389423,-0.0801742524,0.1037873849,-0.0891907588,0.1262204945,0.0682752803,-0.200402841,0.2321671247,0.0608962774,-0.0568326637,0.2117624879,-0.0741795525,0.5051373839,0.1987506896,0.2952888906,0.1450666338,0.0259091631,-0.575330019,-0.0162743255,-0.035353262,-0.0215700585,0.5457751751,-0.2486973256,0.1209726334,0.3923819363,0.4351459444,0.5112317204,0.0446756557,-0.1397817284,0.2479624152,0.2740482688,-0.001346303,-0.1739205718,0.3098240793,0.0871524885,-0.0083130943,0.1016525701,-0.1646696925,-0.1097521558,0.1875065267,0.1612596661,0.828334868,-0.0255359076,0.1997552067,0.2295868844,0.0300468188,-0.0111820158,0.4545129538,0.2147212923,-0.0037078939,0.1024573371,-0.1875474453,-0.0974105299,0.0357835442,0.1867590249,-0.0821633264,-0.2863271832,0.0892688856,0.1414832324,-0.1785375625,-0.0587740503,0.1598269343,0.1505255103,-0.15081653,-0.0377522297,-0.2225744724,0.2729139328,0.0949387252,0.222790122,0.0902482718,-0.1482682377,-0.0515150055,0.0850149095,0.0555325076,-0.258664906,0.3376890123,0.2006182373,0.122237131,-0.6588637829,-0.0510838255,-0.0956509709,0.1528795063,-0.4626125395,0.0733065456,0.0137156071,-0.263197273,-0.0489087887,0.2547800839,0.5092695355,0.1565107852,0.0200365782,0.1766693592,-0.0764170066,-0.0950213969,0.0052554831,0.4508048892,0.1305164248,-0.0291121751,0.4160615802,0.1279015541,-0.2223623544,0.2160473168,0.1577023566,0.2383845747,-0.2103034109,-0.1339183301,-0.0226398911,-0.1382410973,-0.2803114057,-0.1560059488,-0.1664187908,-0.0107981069,0.4410131276,-0.3874879181,-0.0404740721,-0.2920849621,0.091421783,-0.1681881547,0.3196053505,0.504993856,0.2238189131,-0.138535738,-0.2212182134,-0.4413924515,0.2086920887,-0.5922015309,-0.2127123922,-0.1758978516,0.4406628013,0.0937333554,0.3859078586,0.3946072459,-0.0928434655,-0.3016571701,0.4948923886,-0.3109632134,-0.368730247,0.0030861248,0.0873818621,0.0380316451,-0.2598726451,0.2002878785,-0.2337639034,0.0098616593,-0.1697882563,-0.277161926,0.2045567632,-0.1853707731,0.4458235204,0.0681075901,0.4357322752,0.2343308032,-0.4613886178,-0.5193771124,-0.2658767104,-0.0694313422,0.1045595855,0.0557896383,0.3026757836,-0.323957473,-0.0067665344,-0.1195061877,0.1565442085,-0.1857493222,-0.1968384832,0.1651036143,-0.2531904578,0.0407338142,0.3137263954,-0.0751488432,0.0663720816,-0.0122715253,0.1597040743,-0.2519939244,-0.3326855004,0.2609618902,0.008945846,-0.3601915538,0.0265543144,0.1463392377,-0.0099131018,-0.013080108,-0.1742476821,0.0925789401,0.4298184514,-0.027578,-0.2111691684,0.1747496873,0.0376659259,0.1385248899,-0.0340417512,0.3491087258,0.000829451,-0.3108901381,0.3493559957,-0.1170323789]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"Hi ! Since https:\/\/github.com\/huggingface\/datasets\/pull\/2803 we've changed the time out from 10sec to 100sec.\r\nThis should prevent the `ReadTimeoutError`. Feel free to try it out by installing `datasets` from source\r\n```\r\npip install git+https:\/\/github.com\/huggingface\/datasets.git\r\n```\r\n\r\nWhen re-running your code you said you get a `OSError`, could you try deleting the file at the path returned by the error ? (the one after `[Errno 20] Not a directory:`). Ideally when a download fails you should be able to re-run it without error; there might be an issue here.\r\n\r\nFinally not sure what we can do about `ConnectionError`, this must be an issue from zenodo. If it happens you simply need to try again\r\n","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":111,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\nHi ! Since https:\/\/github.com\/huggingface\/datasets\/pull\/2803 we've changed the time out from 10sec to 100sec.\r\nThis should prevent the `ReadTimeoutError`. Feel free to try it out by installing `datasets` from source\r\n```\r\npip install git+https:\/\/github.com\/huggingface\/datasets.git\r\n```\r\n\r\nWhen re-running your code you said you get a `OSError`, could you try deleting the file at the path returned by the error ? (the one after `[Errno 20] Not a directory:`). Ideally when a download fails you should be able to re-run it without error; there might be an issue here.\r\n\r\nFinally not sure what we can do about `ConnectionError`, this must be an issue from zenodo. If it happens you simply need to try again\r\n","embeddings":[-0.4538246691,-0.0359525084,-0.1072578579,0.2148068547,0.225401029,0.0983382016,0.1591716409,0.2268794477,-0.0526379123,0.0000226978,-0.162777245,0.1547378004,0.1955923587,-0.1334099025,-0.097253561,0.0610446818,-0.0565706789,-0.0677602589,-0.1615533531,0.0819472596,0.0188205875,0.1436862499,-0.1121324673,-0.0026506882,-0.0766938329,-0.1644875854,-0.1467983723,0.050724186,-0.2956551611,-0.3328927457,0.1694848686,0.2418556511,-0.0690903589,0.6850526929,-0.0001101031,-0.1397206187,0.2245097011,0.0490238145,-0.3112405539,-0.5094980597,-0.1957669705,-0.1742007434,-0.0006458978,-0.0291792974,0.185485512,0.0554979406,0.1089929864,-0.2749736309,0.3389205039,0.5729048848,0.1995919943,0.3697789013,0.3693856597,-0.0535559133,0.2664019465,-0.012151652,0.031192204,0.4169182479,0.138078779,-0.0596361831,0.1828976423,0.0227427837,-0.0756145343,0.1545686871,-0.1193917766,0.0333615504,0.2555910051,-0.3887176514,-0.1383722574,0.2219341695,0.7953239083,-0.0027446952,-0.4068219364,-0.2871252596,0.1696534157,-0.0129158003,0.3349570334,0.2615113556,-0.4169999659,0.2174756229,-0.2613837123,-0.0808658972,-0.1266333461,0.4406132996,0.1170140058,0.0012068402,-0.0256973505,0.3007272482,0.2633582652,-0.1087947264,0.2221509814,-0.3168979883,-0.0292462241,0.0137489038,-0.3383377492,-0.1888522357,0.0329177901,0.1638691276,0.3037260771,0.0318557173,0.0471077301,-0.1961562186,-0.0839972943,0.1398137212,0.3832516074,0.0297968816,0.0367336459,0.0558998697,0.2139374614,0.0805045292,-0.0963447616,0.0382728539,-0.0520347729,0.0375237539,0.10670311,-0.1745736599,0.4148418903,-0.2983430624,-0.1657698452,0.0255850274,0.0757377893,-0.1702812314,-0.1688547879,0.2133611888,-0.4682379067,0.2446539104,-0.0922744349,0.0652114749,-0.3638776839,-0.3287384808,-0.0513934009,0.0137415705,-0.1487400681,-0.0431218706,-0.0900757164,-0.2335726917,0.1790361106,0.1448258609,0.0673104748,-0.2530994117,0.0061690058,-0.2436126471,-0.1551749706,0.1279544085,-0.0695514008,0.2915445566,0.1237974241,0.1339074969,-0.0637521893,0.0186464153,-0.2753818631,-0.3336540759,-0.200213477,0.2280961424,-0.335934788,-0.2744778097,0.025252834,-0.133858785,0.3139727712,-0.1475003511,-0.1339813918,-0.1570851356,0.0600346029,-0.0604759827,-0.0414292365,0.3157823086,-0.3593474329,0.1914790273,-0.2504243255,-0.0099842027,0.1681512147,0.2309772223,-0.2553583384,0.1200761124,-0.2136332244,-0.2048165798,0.3608438373,-0.3011219203,-0.600212872,0.003167416,-0.3776432574,0.1416860372,-0.1392541677,0.237651974,0.3605267107,-0.0069612009,-0.0734062865,0.3193845451,-0.2023209482,-0.0070757899,-0.3626311421,-0.1656437218,-0.1414755285,0.2422949374,0.0788861513,-0.0042997641,-0.0377065167,0.0240184031,0.2975076735,0.3169288635,0.0668257326,0.2782160342,0.1488048136,0.3036114872,-0.0502669141,-0.2682843506,-0.2213944495,0.1641264856,0.5927707553,0.1514855921,-0.2055570185,0.055662524,-0.3700690269,0.0064080376,-0.0968770906,-0.0029626812,0.0381577797,0.2506614029,0.2869864702,0.2637932003,-0.0415531583,0.4684087932,-0.0279490668,0.3550293744,-0.1794379801,0.2723223269,-0.2326211184,-0.1905997396,0.0346551053,-0.2605832219,0.1190185398,-0.1672925353,-0.1346308589,0.3733447194,0.0169509817,0.0580253042,0.1264173687,-0.1198541522,0.2207273692,-0.1083708256,-0.1503271461,0.248680532,0.1149940789,0.1063638106,-0.1766043752,0.1053012237,0.143037051,0.2057059705,-0.1959834993,0.1009418294,0.1564978808,0.0406317189,0.0823936462,0.0261111017,0.4374154508,-0.3641561568,0.0406906791,-0.1805986017,-0.0071351598,0.3008781075,0.410961926,0.0310156327,-0.1827630848,0.1770051122,0.1970965564,-0.089612104,-0.0317267142,0.4135085344,0.4253745079,0.2163395286,0.0385433398,0.0027145692,-0.17714113,-0.0891942158,0.166731149,0.0989230424,0.1816506088,0.0949445963,-0.0189945735,-0.0067830626,-0.0840521604,-0.5713555217,0.0794092417,0.4571004212,-0.0892287493,-0.137603879,-0.1908154786,-0.0958798304,0.1683706343,0.016017437,-0.475083977,-0.5786948204,-0.0676092505,0.452991873,-0.0524929017,0.0596475191,-0.2232982218,-0.1037176475,0.3926707804,-0.0736510977,0.0445313528,-0.2548263669,0.0125565547,0.0856887326,0.2230580002,-0.1910485476,0.2809529901,-0.3305025995,-0.2099800855,-0.4389460683,-0.043910604,0.1156763509,-0.2164092809,0.3292762339,-0.1231699139,0.5877863169,0.0931101665,0.0436491631,0.2046536505,-0.3436298072,-0.0776778907,-0.056204807,0.0713258833,0.1451002359,0.0598632395,-0.2994942069,-0.2301851958,-0.4278173745,0.1740367413,-0.0479382537,0.0908307359,0.2794734538,0.2059393525,0.0834495574,-0.2537406683,0.1692144722,-0.1258783191,-0.5617018342,0.4327825308,-0.1177528277,-0.263663888,0.0996503979,0.305850327,-0.020622395,0.1772894859,-0.6788051724,-0.0035098523,-0.2368997186,0.3396564424,-0.2022636831,-0.1831388772,0.3164763153,-0.4094335437,-0.1001709104,0.0228217896,0.109788537,-0.0111628193,0.1557815075,0.4228840172,-0.067544274,0.3687635362,-0.0110914931,0.3627903163,0.0647312254,0.0253924541,0.0362944342,0.0257080924,0.3557808995,0.0485108197,-0.3596302271,-0.0280463509,-0.0110422065,-0.1352111846,0.1352456212,0.0259224158,-0.0934613571,-0.2718134522,-0.0847134814,-0.2287119031,-0.316334635,-0.0509185828,-0.1603989899,-0.0765900984,0.2381764054,0.1666227579,-0.0253077913,-0.2131285518,-0.0903376415,0.106533356,-0.0434229225,-0.1533289999,-0.6302165985,0.2468212396,-0.4078631103,0.1728399694,-0.048500035,0.2384988368,-0.177538529,0.0593122952,0.1668082327,-0.014791213,0.5320880413,-0.1183647066,-0.0065427362,-0.0412879363,0.2439120412,-0.468411833,-0.0319941863,-0.0179598294,0.2684278786,0.3909966946,0.1017176658,-0.3571455479,-0.0560006164,0.1374456435,0.2031075507,0.1029755026,-0.3991025686,0.0702966973,-0.2252368033,-0.3088773489,0.1659785062,-0.0910201445,0.1600352675,-0.04411624,-0.0862021372,0.3067347109,0.2041906118,0.0026236123,-0.1327371746,0.1401609331,-0.2742887139,0.5147088766,0.1473813802,-0.2570005059,0.5399026871,0.6204403043,0.1855611056,-0.5246990323,-0.0027769879,-0.1413169503,-0.1575628966,0.2658459544,0.1414180845,-0.0564617962,-0.1830666214,0.449201405,0.1471313387,0.224517718,0.3479778469,0.112206161,-0.3846345842,-0.3544855714,-0.0518103093,0.0050303196,0.0098074041,0.6732231379,0.0225052554,0.0605005026,0.2474566996,-0.049816329,1.1006953716,-0.1692030281,0.1691567749,0.1133905277,-0.1307371557,0.5093159676,-0.2213900089,0.0868217498,-0.1367001832,-0.1483263522,0.0517303944,-0.0947281942,0.0730001926,0.1920678169,-0.2430835217,0.0707540065,-0.1546960175,0.1884455085,-0.0883382112,0.1380520016,-0.3857750595,-0.0779629722,-0.1396879852,0.2257314026,-0.0752516761,0.269074291,0.0206742529,0.0868844017,-0.2398618907,-0.0202986859,-0.5047807693,0.0938810632,-0.0931527093,-0.0274194721,-0.4438469708,-0.155970633,-0.1467036903,0.2542875707,0.0000526351,0.1928377002,-0.0549912862,0.2169499546,-0.1297851056,-0.0963955447,0.0818099231,-0.2373087853,0.1440313011,-0.0193442367,-0.496923238,0.2323336452,-0.0716103092,-0.3198162019,0.1333940327,0.1087034121,0.0935107544,-0.1285556257,0.0663377494,-0.1190061644,0.0405748896,-0.3209662437,0.1466360986,0.2985940576,-0.0046109455,0.2261171788,-0.0861766785,-0.3036797047,-0.3014829755,0.3936604559,-0.0082371915,-0.0050014216,0.2596047819,0.2101564556,-0.3978389204,-0.2830137312,0.3068806231,-0.4383847117,-0.3919934928,0.0434712321,-0.5033024549,0.3654732704,-0.0944298431,-0.1962878257,0.2207839787,0.050057102,0.1057844609,-0.6435926557,-0.112725161,-0.0441772342,-0.087554574,0.0152264861,-0.0437841825,-0.2011600733,0.0685492754,-0.0929156765,-0.3221847713,0.038775444,-0.2956762314,0.1002294123,0.4034521878,-0.1686833501,0.4216539264,-0.2036422491,0.0921621248,0.107712768,-0.1632175446,-0.1457540393,0.0300687142,0.1397892535,-0.0278808344,-0.2190429121,0.1264558136,-0.4023299217,-0.1240626797,0.0671283454,-0.0584938414,0.1805595756,0.0082785152,-0.2347159088,0.188729465,0.0936195254,-0.0575819127,0.1896219105,-0.0769715086,0.4825540185,0.1868838072,0.3720387518,0.0232267752,-0.0362637416,-0.4318806231,0.0704969913,-0.0795288682,0.0478319526,0.5599256158,-0.286220789,0.059289176,0.3281361461,0.5346974134,0.5222051144,-0.0060021468,-0.1228737012,0.3413521945,0.2643199563,0.028864935,-0.1323685199,0.4328159988,0.0564259887,0.0538241304,0.1736360788,-0.2305140644,-0.0726387054,0.1665361524,0.1727385372,0.8266090751,0.1331849098,0.1718324572,0.2565765381,-0.113578096,-0.0658832267,0.4332050979,0.2657815218,-0.0021295596,0.1296424121,-0.289693743,-0.0243610162,-0.0274886396,0.1395929009,-0.0272971652,-0.312753886,0.0777793378,0.140043959,-0.1352997273,-0.1218784377,0.0924050808,0.1922758967,-0.2082992792,-0.1651860476,-0.1020177677,0.3825664222,0.0385682918,0.3544970155,0.0017309929,-0.1745239496,-0.1743298769,0.0220151804,0.0964681655,-0.2946985364,0.3090139329,0.250826478,0.0474553704,-0.7139669061,-0.1418219358,-0.0674925819,0.1310342848,-0.490365684,0.1432151049,0.0534580015,-0.162340641,-0.0222107507,0.307228297,0.5712034106,0.1255346984,0.1071017906,0.2486964017,-0.0851117447,-0.0366279483,0.043594256,0.3503609002,0.06011457,0.0243838765,0.3931989968,0.1480340809,-0.2403560281,0.1648120135,0.1535256654,0.2381442189,-0.2155444175,-0.1856283396,0.0254312642,-0.1125086248,-0.2143724412,-0.1487719417,-0.1726246774,-0.0003801415,0.3418622911,-0.288359791,0.0096085807,-0.2853538692,0.0952519923,-0.1689324379,0.2129429877,0.5421634912,0.1865495741,-0.0928706005,-0.2022402883,-0.4226510823,0.184549734,-0.5492637157,-0.3295221031,-0.1850746125,0.3340445161,0.174022153,0.4482896328,0.4680608809,-0.0904588029,-0.2368764579,0.3994177282,-0.3604365885,-0.3117973506,0.0756682381,0.05952923,0.0811509565,-0.2063947171,0.2447949201,-0.2420951575,0.0437555611,-0.061561048,-0.2038389742,0.1301213205,-0.2715986073,0.3323514163,0.0370632634,0.3929243088,0.199374944,-0.4113180935,-0.3911923468,-0.2847712934,-0.0875257179,0.117393285,0.1652342826,0.4006875455,-0.3302830756,-0.1996988654,-0.212048173,0.3072327673,-0.1407973915,-0.2316658795,0.1535109878,-0.2409273386,0.1307951212,0.268887341,-0.036486946,0.0119676935,-0.008522762,0.1623505056,-0.2913174033,-0.2029954046,0.3976862431,0.0385884345,-0.2475548834,-0.0036511149,0.1197994277,0.0990649313,-0.1700856984,-0.2421632707,0.088280201,0.369887799,-0.07661888,-0.2562713325,0.0985955074,-0.0127007272,0.1415995806,0.0295892842,0.3484928012,-0.0263655577,-0.3384765387,0.3663103282,-0.0425285622]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"@lhoestq thanks for the update. The directory specified by the OSError ie. \r\n```\r\n1ec12301abba4daa60eb3a90e53529b5b173296b22dc3bef3186e205c75e594c\/corpus-webis-tldr-17.json \r\n```\r\n was not actually in that directory so I can't delete it. ","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":26,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\n@lhoestq thanks for the update. The directory specified by the OSError ie. \r\n```\r\n1ec12301abba4daa60eb3a90e53529b5b173296b22dc3bef3186e205c75e594c\/corpus-webis-tldr-17.json \r\n```\r\n was not actually in that directory so I can't delete it. ","embeddings":[-0.3095022142,-0.1504017413,-0.1248522475,0.1671922207,0.0626938269,0.2322175056,0.0452464111,0.2481846213,0.0208498016,-0.0671563223,-0.0702519044,0.163301155,0.2610349953,-0.1121932715,-0.133146435,0.2049571574,0.0665125772,-0.2218913734,-0.0956013799,0.1206625998,-0.0917953029,0.1593911499,-0.1283442378,-0.1198260337,-0.141182825,-0.1201372966,-0.1345248669,-0.0305773001,-0.2029904276,-0.3059111536,0.2204692513,0.1168280691,-0.029211536,0.6436533928,-0.0001098637,-0.1421745867,0.2654624581,0.0524743497,-0.4409989417,-0.5108650327,-0.2619158626,-0.12985304,0.0590727143,-0.031244399,0.2131990045,0.0836230367,0.1228473783,-0.3912442327,0.2117971182,0.4887261689,0.1886624247,0.3496263623,0.2179190367,-0.0172980651,0.1792479008,0.0728332773,0.0780197382,0.3980514407,0.1822239906,-0.141372472,0.1812858433,0.0218105223,-0.1129294187,0.2673594356,-0.1171443388,-0.0004770057,0.1453633755,-0.4685249627,-0.0216492452,0.1269262284,0.9099088311,-0.0185254905,-0.4328919947,-0.1754165143,0.0527640246,0.0815893486,0.4559943676,0.4742854834,-0.3605511487,0.2056500167,-0.2597310841,-0.0197784621,-0.0729322284,0.3834029138,0.0901441798,0.0476595201,0.0735123977,0.2317583114,0.0998640582,-0.0492647849,0.2535369396,-0.3695167303,0.0089942543,-0.0261943266,-0.3241217434,-0.1076200381,0.0020491134,0.1523838639,0.1226414815,0.0144249778,-0.0172259416,-0.2915516496,-0.2147482336,0.1797130406,0.4288505018,0.0015061132,-0.0348932892,0.1349713206,0.2598968446,-0.0260889512,-0.0193673577,0.081011802,0.1183842346,0.0089719594,0.0835823938,-0.1330717355,0.3993260264,-0.2017674446,-0.2572386563,-0.0288927276,0.0433466434,-0.1179031804,-0.2280004919,0.1583617777,-0.4906696975,0.2088637054,0.0793649256,0.1370234489,-0.362824887,-0.3598408997,-0.0434101857,0.1102728173,-0.1638957858,-0.1707212925,-0.0593388528,-0.1105477959,0.0511401817,0.1225289926,-0.0155340414,-0.3069616854,-0.0715052262,-0.1762696803,-0.1245331988,0.23968032,-0.1146276444,0.4079996347,0.1003963202,0.116927132,-0.0820333585,0.0389702171,-0.330312103,-0.3403403163,-0.1272415668,0.2374211401,-0.2133954018,-0.3039638698,-0.125053376,-0.1104238927,0.2541174293,0.0337237269,-0.0793847069,-0.0963231623,0.0166097991,-0.0846725777,-0.1574709415,0.3429022431,-0.5084065199,0.3213284612,-0.196583122,0.1041990742,0.1816943586,0.3182139695,-0.1496501714,0.0935464799,-0.2771325409,-0.1204956993,0.2729804218,-0.2532721758,-0.6182508469,0.0820885748,-0.3610656559,0.0075530135,-0.0865056813,0.2850093246,0.279789716,-0.1380776763,-0.0132090636,0.3752258122,-0.2172137499,0.0725085139,-0.3490177691,-0.1847133785,-0.021370193,0.1756252348,-0.0187508408,-0.0033899227,-0.1430781782,-0.0335804336,0.3587705791,0.3870671391,0.1483341455,0.3216244578,0.0243828204,0.2783790231,0.0534567907,-0.2024547607,-0.3310602605,0.1612716019,0.5941175818,0.1395786703,-0.2706468105,0.0003419686,-0.4284191728,0.0866755694,-0.0213273335,0.0571298748,0.0183077063,0.1815392971,0.2960186005,0.2478752285,-0.0115101915,0.4557389915,-0.0579033233,0.2721843719,-0.2114622444,0.2506326139,-0.2167890966,-0.1202930138,0.1560701877,-0.3509273529,0.0836673751,-0.2268949896,-0.2199625522,0.3265954256,0.0438918583,0.1196844354,0.0249896757,-0.0596186854,0.2224897742,-0.2003027201,-0.134572506,0.4064633548,0.1330251992,0.0998228788,-0.2701017261,-0.0196894165,0.2064270675,0.2266620547,-0.1862728298,0.1116251722,0.090895474,0.0398536697,0.204728812,-0.0470396541,0.4167723656,-0.3639966249,0.0535619073,-0.2369760275,0.0256813709,0.3321652412,0.3510746062,0.0703334957,-0.1828645021,0.2540558577,0.1528062075,-0.1309075952,-0.0735404864,0.3477202654,0.263831526,0.1764098108,0.1616751999,-0.1241317987,-0.1138275862,-0.1220456958,0.1897414774,0.126205489,-0.0436989442,-0.0182874035,0.1533727646,-0.0168347899,-0.1188560873,-0.4593299329,0.0727738142,0.4072105885,-0.0815584809,0.0731035173,-0.2246208042,-0.1491611749,0.1998507828,-0.0577890761,-0.3491258323,-0.5447545648,-0.0516445972,0.365608573,-0.0903268829,0.0406891108,-0.0846578106,-0.117372863,0.4572752118,-0.158053726,-0.010764692,-0.133191973,-0.0478320904,0.0734729022,0.242655322,-0.1946613342,0.3340427577,-0.2936217785,-0.2266913205,-0.5305721164,-0.018986728,0.0168310758,-0.1706739366,0.2737117112,-0.1221019253,0.5325040817,0.0582203455,0.0858037025,0.0869639963,-0.256645292,-0.1234242693,-0.0716627911,0.0591876544,0.1748991907,-0.0101410402,-0.3868838847,-0.1918511838,-0.404558748,0.106361039,-0.1477372348,0.0216791723,0.1028052643,0.1835914254,0.0243516024,0.0331155285,0.1411416382,-0.1265153438,-0.4522252679,0.4309332371,-0.0664703473,-0.2162594199,0.104318358,0.2256389409,-0.0498273633,0.215276137,-0.7514492273,-0.0145282028,-0.278327018,0.2697307765,-0.1753168106,-0.189715907,0.2162373364,-0.4299247563,-0.0869578943,0.0475652181,0.0349085368,-0.0592548326,0.1991394758,0.3676381409,-0.0787483528,0.3557338417,-0.0129544698,0.199818179,0.1184440255,0.0496494584,0.1700463295,0.1690751761,0.2795865238,0.0121602742,-0.2339679748,-0.0585169196,-0.0014460189,-0.2347747833,0.1252498478,0.0456063636,-0.1891673505,-0.1527427137,-0.0685337633,-0.2882629931,-0.2439082563,-0.060317494,-0.1026765034,0.0170278233,0.2532618642,0.0990905166,0.0114231948,-0.277900815,-0.0780699775,0.1083372235,-0.039546717,-0.1868322045,-0.4770647883,0.0852853358,-0.3877303004,0.2447301596,-0.0003017061,0.2891821265,-0.0396191142,0.118132405,0.1726474166,-0.0341566354,0.6463543773,-0.1588416696,0.1628613174,-0.0375560522,0.2864968181,-0.4513315856,-0.0555176809,0.0547593981,0.3185351491,0.5014467835,-0.0028705033,-0.2623036206,0.0188497007,0.1009352133,0.2959567904,0.0538651645,-0.4968960285,0.0795541257,-0.2046012729,-0.2623451054,0.1263023913,0.1091593131,0.0686356425,0.0350880027,-0.1038482264,0.2448898256,0.0855247676,0.0520429313,-0.1704456061,0.1298332959,-0.2464815378,0.4957830012,0.0454470031,-0.2059773058,0.369571954,0.684181273,0.1612985283,-0.4628184736,0.0990600362,-0.0322146751,-0.1157461032,0.2643632591,0.1172473058,-0.0979505256,-0.1994203925,0.368955195,0.0208758786,0.1572159827,0.2718421221,0.0810242444,-0.4527803659,-0.4301779866,0.113894105,0.0375651084,0.0289464127,0.6303082705,-0.0606518313,0.115968734,0.3418049514,-0.0649461225,1.006020546,-0.2283347249,0.1064811051,0.0492464192,-0.0904068276,0.4886732697,-0.411364764,0.0716899708,-0.1177021787,0.0135577703,0.031148959,-0.1090933979,-0.1918100715,0.3354752958,-0.2478168607,0.0934099853,-0.1912205815,0.1450520307,-0.1241278946,0.2234558016,-0.2989013791,-0.0206946321,-0.1991699636,0.2247374803,-0.0198315158,0.2506230175,-0.0100728571,0.2096994966,-0.1463121176,-0.0804262012,-0.6208152175,0.0453265384,-0.0195442941,0.0316737182,-0.3655813336,-0.1974221915,-0.0437818058,0.2366869748,0.1416718513,0.2241814584,-0.0163395386,0.2341423333,-0.1062711254,-0.1597452164,-0.0008495637,-0.1496988982,0.1444966495,0.0135768149,-0.2735590935,0.2143666893,-0.1091816798,-0.2239749134,0.2120580226,0.2389528602,0.0794445053,-0.136567235,0.0419685021,-0.1080686823,0.1240836978,-0.2395192981,0.131440714,0.3121679127,0.1640829444,0.2492801994,-0.1205941066,-0.3549574018,-0.3437150121,0.4136879742,0.0535841659,-0.0073905475,0.3306728005,0.3128746152,-0.4030729532,-0.309256196,0.0924070925,-0.5343806744,-0.3205618262,0.3328488767,-0.4157615602,0.3471693397,-0.062788561,-0.1378178,0.203016609,0.1054211184,0.1233612001,-0.6414012909,-0.0647877008,-0.0836532116,-0.2053898126,0.0240563769,-0.1460573226,-0.1997723281,0.1127578914,-0.1422039866,-0.295127064,0.0484703258,-0.2745977938,0.0958570838,0.3683098853,-0.0902536511,0.4197207689,-0.0623872951,0.0887815133,0.1003364697,-0.1311419457,-0.1470998079,0.0164527185,0.1390110552,-0.0766675621,-0.191554606,0.1097426638,-0.5065022707,-0.0802384764,0.0977504477,-0.0953500867,0.1345021576,0.0178678334,-0.2570755184,0.177574113,-0.0285100583,0.0442681648,0.2221822143,-0.1041632965,0.5471839905,0.196433723,0.2797546685,0.1488572359,0.0148905022,-0.5414075851,-0.0015548152,-0.0919464007,0.0543605089,0.3810917735,-0.3114630282,0.072883606,0.3426959515,0.417276144,0.503536284,-0.0343137421,-0.1695257127,0.2993564904,0.2728626132,0.0618403852,-0.1616173089,0.3489340544,0.0864299759,-0.0092736557,0.1848194748,-0.1044386402,-0.058701802,0.1856523305,0.2389995307,0.7759051919,0.0514618903,0.1726847142,0.0957849845,0.0030998725,-0.0193010122,0.3979817033,0.1976183057,0.0659400225,0.0808458552,-0.1315656304,-0.0837668926,0.1171829179,0.1486099064,-0.0626389384,-0.3323945403,0.034631364,0.1082113907,-0.1654981673,-0.008060243,0.2052625418,0.0791040212,-0.2886404991,-0.0663575009,-0.2051645815,0.2875978053,0.046893511,0.2713527679,0.1986185759,-0.146088779,-0.0298795588,0.1195137128,-0.0047301957,-0.2354220152,0.2322648019,0.2363520116,0.1051327437,-0.5865030289,-0.0447985083,-0.1485786736,0.1900198609,-0.4724008143,0.0266745593,0.0291519966,-0.2448444963,0.0099696508,0.2574675977,0.5680678487,0.2143834978,-0.0539006367,0.1223606989,-0.1421367675,-0.0592411794,-0.0138143562,0.4059752226,0.0000049295,-0.0787821338,0.4521322846,0.1227863133,-0.2082473636,0.1654384434,0.1053877324,0.2668370008,-0.1839603931,-0.1674202383,-0.0302672703,-0.0702591389,-0.3013981581,-0.2018900663,-0.1886167526,-0.0504566021,0.3965955377,-0.3332307935,-0.0050685559,-0.1903362125,0.0933342427,-0.1705901325,0.3120704293,0.6364371777,0.2870883644,-0.1428942978,-0.2970386147,-0.388443917,0.1540689319,-0.5896973014,-0.1935309023,-0.2076192796,0.4632063508,0.0898938775,0.3273732066,0.4216398597,0.0515935719,-0.409479022,0.4779759347,-0.3611396551,-0.3932264149,0.0418681987,0.0766325071,0.0460634008,-0.2648559511,0.1191573292,-0.2247226387,-0.0011504382,-0.1998833418,-0.1851944476,0.1524484307,-0.0451054536,0.4760362804,0.0896722227,0.3005132377,0.197111845,-0.5851706266,-0.4740627706,-0.280194819,-0.0704731345,0.0913138092,0.088868849,0.4465154409,-0.4278231859,-0.0006203317,-0.1412664354,0.1998655498,-0.1642399281,-0.2047735453,0.1525726467,-0.189626202,-0.0411875509,0.4074382484,-0.0385515131,0.0942021385,0.0715055093,0.1940276623,-0.2873764038,-0.3336482644,0.3049746752,0.0193077903,-0.3906154633,0.0559618808,0.1058753431,0.0637906641,-0.0104585364,-0.2375225276,0.1077349633,0.4397536218,-0.0333892703,-0.2493492365,0.103724584,-0.007916797,0.1705269516,-0.0169549529,0.5145740509,0.0177307185,-0.3262771368,0.3846450448,-0.1136140972]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"Oh, then could you try deleting the parent directory `1ec12301abba4daa60eb3a90e53529b5b173296b22dc3bef3186e205c75e594c` instead ?\r\nThis way the download manager will know that it has to uncompress the data again","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":27,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\nOh, then could you try deleting the parent directory `1ec12301abba4daa60eb3a90e53529b5b173296b22dc3bef3186e205c75e594c` instead ?\r\nThis way the download manager will know that it has to uncompress the data again","embeddings":[-0.3229547739,-0.1548745632,-0.0958515629,0.2122767121,0.1727839857,0.2089084834,0.0311661009,0.2200947851,-0.0430057198,-0.0564695038,-0.0722842738,0.1025691852,0.2630372643,-0.0284359343,-0.1015666798,0.2591447234,0.0751857385,-0.1532897949,-0.134002611,0.093762137,-0.0453595072,0.1273859441,-0.1522027403,-0.1518076509,-0.2515835166,-0.0813376531,-0.1123948395,-0.0133465808,-0.2417642772,-0.3136616349,0.2217783332,0.1977145821,-0.0670335516,0.6311795712,-0.0001082193,-0.1661073864,0.2970055044,0.0607004948,-0.4434760809,-0.5188804865,-0.215862155,-0.1876016706,0.0243747514,0.0184114669,0.1335865259,0.0764069483,0.0671559721,-0.3609519005,0.3123517632,0.3766123056,0.2216305137,0.2938492298,0.1855626553,-0.1055657491,0.1551821679,0.0236803293,0.0808866918,0.4364547729,0.1442379951,-0.1347500831,0.1696618199,0.0660864338,-0.0995426252,0.2206776589,-0.1114257351,-0.0089217285,0.1844033003,-0.4553858042,-0.0359131321,0.1294727474,0.8906873465,-0.0397688448,-0.3760148585,-0.2968247235,0.1337402761,0.0270687733,0.4262497425,0.4283311367,-0.3859825432,0.2389271259,-0.266323179,-0.0228596497,0.0520490929,0.336849153,0.0947214887,-0.0229382161,-0.0076306579,0.1902257353,0.192469582,-0.0476929285,0.3261296749,-0.3512678742,0.0251167547,-0.0928472653,-0.3812167645,-0.0857717767,0.0138084851,0.17820324,0.1665260047,0.0568927824,-0.0361496769,-0.2665970623,-0.1561307907,0.1570051759,0.4282910824,0.0166722089,-0.0358770266,0.147023052,0.2235934436,-0.0597804934,-0.0454057567,0.0735234693,0.0501849167,-0.0342417248,0.1098434329,-0.0982877016,0.2976776063,-0.1693594307,-0.2574085295,-0.0296510197,0.1553851068,-0.1137762293,-0.253431648,0.1721235663,-0.456833899,0.3014909625,-0.0733489394,0.103349559,-0.3224068582,-0.3715544641,-0.0770322829,0.0178418625,-0.2535534203,-0.114999868,-0.0478267111,-0.0865907222,0.0556586869,0.2549179494,-0.0034713037,-0.3169311881,-0.0021104482,-0.2342968434,-0.0961762667,0.2281832099,-0.0899430886,0.4282738268,0.0563638732,0.1315538138,-0.0567201562,-0.0037883378,-0.3410360515,-0.2845475376,-0.1375982761,0.248136878,-0.2255693376,-0.2428179085,-0.152455315,-0.0870867968,0.1295226216,0.0438936315,-0.095159106,-0.0379463248,0.1215736866,-0.1129857972,-0.0805240273,0.3380515575,-0.4335297644,0.2870418429,-0.2611810863,0.0653512627,0.2386940569,0.2952560186,-0.2074873894,0.0207775664,-0.2533313632,-0.1445062757,0.2831653357,-0.166037336,-0.7092910409,0.0670329779,-0.3865574896,0.0544933341,-0.0951566547,0.2693455815,0.2380082756,-0.1328064352,0.0359675288,0.4281702936,-0.240654856,0.0688112453,-0.3896734715,-0.2089087516,0.0580077805,0.1221133545,-0.0680812821,0.0384233408,-0.1475014538,-0.0306050535,0.3082167804,0.3910114169,0.0956725478,0.3066321015,-0.0085600819,0.267819345,0.0175748114,-0.1669080406,-0.3697960079,0.181688875,0.4722077549,0.0637292862,-0.2675577104,-0.0389684476,-0.4558921158,0.0080367196,-0.0186161622,0.0451203398,0.0606043823,0.2065854967,0.2361144871,0.2035789937,-0.0189470612,0.54888767,0.0479960591,0.2936903834,-0.1253661215,0.1977761686,-0.1798651814,-0.1205080673,0.1851098835,-0.363132596,0.1816332191,-0.2229929417,-0.1726033986,0.3127637804,0.0540517494,0.1943844557,0.0477315299,-0.2152786702,0.2617830634,-0.1447607279,-0.1480248272,0.4077906013,0.1012227014,0.059215676,-0.2254851013,0.005725048,0.2113429308,0.1735051274,-0.2585220337,0.1331920773,0.1047107801,-0.0047231638,0.2456095964,-0.0934913978,0.3387973905,-0.248947531,0.1024932265,-0.1653571427,0.0717944503,0.3274765313,0.3101591766,0.054316476,-0.2026435286,0.2211813182,0.0865456536,-0.1284258813,-0.0187284015,0.4358161092,0.2632165253,0.1839127392,0.2763888538,-0.0652680472,-0.1404545903,-0.1486685276,0.1811475903,0.0201219767,-0.0072506336,-0.0460743718,0.1301448345,-0.0028621759,-0.1005294994,-0.3409176171,0.1631841809,0.3060806096,-0.0270071886,-0.0144398874,-0.3174676895,-0.1755051315,0.2243029475,-0.0507291108,-0.2910987437,-0.5441323519,-0.0666458905,0.2732475698,-0.0621724129,0.0719594136,-0.1098337919,-0.1164225638,0.3750080466,-0.1574857235,-0.0236032493,-0.1277323514,0.0175013412,0.10863024,0.2345093787,-0.2008160204,0.3059839606,-0.2675752044,-0.2694882452,-0.4199394584,-0.0039956979,0.0205786712,-0.1766990721,0.2704044878,-0.1738457233,0.469563365,0.1040984094,0.1371683776,0.1168253645,-0.1985445321,-0.0714371502,-0.0763074234,0.088183172,0.1546370238,-0.0620138794,-0.3198411167,-0.2525842488,-0.3489462435,0.1429210901,-0.1798455417,-0.010337851,0.0810252801,0.195770517,-0.0042211888,-0.0697445124,0.1001497284,-0.2080837488,-0.4533269703,0.4969059229,-0.0752501264,-0.2108899504,0.0671744496,0.2153131217,-0.0347476453,0.2523428798,-0.6985370517,0.1177348047,-0.3417280614,0.2311582863,-0.0959614962,-0.0915403962,0.1849167198,-0.4445741773,-0.1377573609,0.0578001961,0.0158680771,-0.119964458,0.1828929335,0.4253915846,-0.1589241475,0.2361423671,0.0096748201,0.1478535831,0.1611280739,0.0944745913,0.1329361498,0.0986260995,0.3116336167,0.0382882319,-0.236718595,-0.0082221748,-0.0521577522,-0.1688452512,0.0556841232,0.0042465311,-0.1978373826,-0.1728802323,-0.0491411276,-0.303101629,-0.282676518,-0.0783567354,-0.2005131096,-0.0245784316,0.2768312097,0.0961313173,0.0270894021,-0.2244980931,-0.0682405382,0.0223049074,-0.0316494815,-0.1880517006,-0.4782491922,0.1640078574,-0.3637948632,0.2138873786,-0.0782149583,0.2673681974,-0.0601456426,0.0691483319,0.1800273657,0.0097442539,0.5972951651,-0.1228708401,0.0711162314,-0.0421208367,0.2663243711,-0.3959361613,-0.0773223788,0.130768165,0.2371252328,0.4672703445,-0.0770895779,-0.2989766598,0.0393826775,0.0109477593,0.3277234435,0.0852342024,-0.4379781485,0.0409490801,-0.1564339101,-0.2209914327,0.103896156,0.0966446996,0.0743274242,-0.0187686961,-0.1925458014,0.2158980221,0.1383886039,0.1050995812,-0.1186629757,0.2249954641,-0.2883522809,0.4566147327,0.1170315593,-0.1751348227,0.3143304884,0.7368856668,0.2488532066,-0.4469808936,0.0710989982,-0.0204800628,-0.1305081248,0.3000690639,0.0614250936,-0.0396419838,-0.2185840458,0.4220646322,-0.0212411899,0.1574662775,0.2910261452,0.0361128859,-0.4399998486,-0.4556054175,0.1672197133,0.0991086513,0.0398140065,0.7080286741,-0.1060628518,0.1349106431,0.2072743922,-0.0953416973,1.0428043604,-0.2065405548,0.1418238878,0.0392075256,-0.094096981,0.4687246084,-0.3534321487,0.0853755698,-0.0433649644,0.0304953903,0.0344041474,-0.084532842,-0.1817820221,0.3168224692,-0.2818124592,0.1816873699,-0.2366980314,0.1214529872,-0.157609567,0.1928666383,-0.3748887777,-0.0243160594,-0.1379537135,0.2495529056,0.0385167785,0.2364375889,0.0259888303,0.1212934405,-0.1247036085,-0.0537204556,-0.6156312227,0.0517454483,-0.1698077768,-0.0385154895,-0.3800042868,-0.2348268777,-0.0605930015,0.1536155939,0.1128549129,0.2116150558,-0.0490338914,0.2427803576,-0.144032836,-0.0439157225,-0.0471792631,-0.1807094067,0.1273679137,0.0231983345,-0.3375990391,0.2221047282,-0.0565145239,-0.1635820121,0.2309992015,0.2578212023,0.0968183056,-0.1043442488,0.0605711527,-0.0869510174,0.154744938,-0.2690963149,0.174286142,0.2979972661,0.1570421606,0.2275492698,-0.1475709677,-0.3382980525,-0.3909550607,0.3864549398,0.0356899574,-0.086984098,0.2533664405,0.2230577171,-0.3960119784,-0.3125069439,0.1102764532,-0.5467652082,-0.3165750206,0.2736478448,-0.4207128882,0.2876586318,-0.0433814339,-0.1575329304,0.2123346925,0.0839530528,0.1333031505,-0.5559974909,-0.1051083952,-0.1152151674,-0.2786164284,0.031419117,-0.1133975685,-0.1451276094,0.2314849794,-0.0455208011,-0.3330334425,0.1012120098,-0.2488736361,0.1280422509,0.2624683678,-0.0927807838,0.3442625701,-0.1348825991,0.1149704978,0.1428751349,-0.1397125423,-0.1804019064,-0.0137787368,0.1184450537,-0.1072441339,-0.1202157065,0.096537374,-0.4541833699,-0.0819671527,0.0703591779,-0.0191849936,0.0994389877,0.0988778323,-0.2201792449,0.1730370969,-0.0524032079,-0.0701597258,0.1650610864,-0.077280499,0.5235565901,0.2017830163,0.2585534751,0.1099747494,0.0459432714,-0.5424665213,-0.0166705772,-0.0623557046,0.0409433357,0.4354705215,-0.256242156,0.0311667752,0.304849714,0.4342504442,0.5167433619,-0.0207804851,-0.1835751832,0.3832824826,0.2985503376,-0.0159600973,-0.1510686874,0.2647130489,0.171720773,-0.0071416483,0.1411260217,-0.0328542106,-0.0482021496,0.1285698265,0.2093429714,0.6953504086,0.0386541374,0.2273381799,0.1782874912,0.1388129294,-0.0286272932,0.468357414,0.2474291772,0.0253023282,0.1073262542,-0.1378973573,-0.0369270034,0.0003921087,0.1407735646,-0.1127588823,-0.3116643727,0.0422544256,0.2148376256,-0.2325130701,-0.0299793277,0.1322474778,0.1798060834,-0.2827203274,-0.1676199287,-0.164477855,0.2919849753,0.064123258,0.2669749856,0.2639375329,-0.0923782289,-0.1687733382,0.0813644677,0.0371085331,-0.2284736037,0.3985338807,0.2368303984,0.1304698884,-0.5860459208,-0.054081127,-0.0530214459,0.195715338,-0.4779833257,0.0264094789,0.0305640008,-0.2674971819,-0.0185983106,0.2658746839,0.5785292983,0.1724967659,-0.0661332086,0.0794493482,-0.0307389572,-0.0551208891,-0.0800942555,0.4403755665,0.0343762152,-0.0934068188,0.3708055615,0.1527278274,-0.260795325,0.1164208204,0.0944370553,0.2666628361,-0.2454841584,-0.0307158846,-0.0360297561,-0.1037130803,-0.3233611584,-0.2196986377,-0.2076780051,0.0101851858,0.5007896423,-0.3325722516,-0.0063664569,-0.2588770986,0.1201225817,-0.1117340028,0.3275038898,0.5890837312,0.232092157,-0.1620317549,-0.3375647068,-0.5112208128,0.1102174744,-0.5844736695,-0.1985817403,-0.1481408328,0.4136403799,0.1261549145,0.4157708585,0.3685455024,0.1126013696,-0.3510943055,0.4230380952,-0.3570303321,-0.3375181258,-0.0206216946,0.0368579477,0.0541546345,-0.303840965,0.1973661929,-0.2808120251,0.0178733207,-0.1091532484,-0.1492456198,0.1750957072,-0.0382720493,0.5005349517,0.0738065317,0.2318065464,0.1809499413,-0.5287168026,-0.490619421,-0.2379537076,-0.0188400894,-0.0321709588,0.1138350517,0.3460378647,-0.3993895352,0.0892246738,-0.1867887825,0.202719301,-0.082118921,-0.2360038161,0.1945278496,-0.2277840078,0.0029831314,0.3618310094,0.0315354057,0.2227500677,-0.0178697873,0.2073406875,-0.247781381,-0.3570301533,0.2816219628,0.1005692557,-0.4137580991,0.0535922423,0.170096457,0.1150686592,-0.0047739833,-0.239549011,0.165006578,0.3653581142,-0.0137979509,-0.2470384687,0.1575022936,0.0031551281,0.1486157924,-0.0199631806,0.515188694,0.0958430693,-0.2933823168,0.3243213892,-0.0978853479]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"It seems to have worked. It only took like 20min! I think the extra timeout length did the trick! One thing is that it downloaded a total of 41gb instead of 20gb but at least it finished. ","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":37,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\nIt seems to have worked. It only took like 20min! I think the extra timeout length did the trick! One thing is that it downloaded a total of 41gb instead of 20gb but at least it finished. ","embeddings":[-0.3443332016,-0.1409978122,-0.0883192793,0.1077195704,0.1344687641,0.141922012,0.0290933382,0.2305127382,-0.0500460416,-0.148842454,-0.0651002824,0.1319421828,0.2865752578,0.0448663943,-0.0551885068,0.258646518,0.0915640593,-0.211843878,-0.0709677488,0.0641470179,-0.0160000175,0.116680488,-0.1866284907,-0.1631224155,-0.2236128598,-0.0258419942,-0.1102018952,-0.0971679464,-0.2112138122,-0.2538017631,0.2425089926,0.1964347512,-0.1058989242,0.5974957943,-0.0001101712,-0.1901295632,0.3014584184,0.0749201775,-0.3920445442,-0.5111631155,-0.1665061861,-0.1499351114,0.0677512959,0.0589529686,0.1219026893,0.1088145897,0.0431350544,-0.34569031,0.2524756789,0.4062392712,0.1974427402,0.3186733127,0.2356890738,-0.0650089309,0.1889505833,0.0090737212,0.0612702146,0.4360184073,0.2413877398,-0.148238197,0.2863081694,0.0281024091,-0.1203852221,0.2467680275,-0.1053779125,0.0088083874,0.1443118006,-0.4500547051,-0.0667435303,0.1656840891,0.853022337,0.0380398817,-0.4033251107,-0.2840456665,0.1371211112,0.0184613839,0.3736425936,0.4260642231,-0.411041677,0.1774050742,-0.2858318686,-0.0086823329,0.0160078201,0.3706652522,0.1045642868,-0.0335832164,0.0563669764,0.264218986,0.1433646083,-0.0444081984,0.250808537,-0.3367406428,0.0569291227,-0.0836199448,-0.4266654253,-0.0696586743,0.0054502673,0.2371156067,0.1993168294,0.010983482,-0.0365413129,-0.2065038532,-0.1257815361,0.148792088,0.3903265595,-0.0173849408,-0.0382208414,0.1559549421,0.2482758462,-0.152148962,-0.072568655,0.0817225948,0.0498883277,-0.0067599411,0.0552344956,-0.1608839929,0.2735024989,-0.2745932341,-0.1909499168,-0.0340160131,0.1356941015,-0.0971092135,-0.2648789585,0.1821255982,-0.478856504,0.2605337203,-0.0654438287,0.1063378602,-0.3326458335,-0.3925067484,-0.061796315,0.0165220648,-0.2331841141,-0.0992078036,-0.0376826786,-0.0509471409,0.0369332843,0.1624528915,0.0429841168,-0.2928953469,0.0006367006,-0.1750113368,-0.1017823368,0.1838416308,-0.1402421892,0.4254395664,0.026733594,0.1841025501,-0.0658647045,0.0560505725,-0.28247118,-0.2472639978,-0.1145937964,0.2255188972,-0.2630476654,-0.3427596688,-0.1906621456,-0.0305896848,0.1723275483,0.0454448573,-0.1256667525,-0.0434742197,0.1271689236,-0.0709539503,-0.0858591422,0.3016451299,-0.5062198043,0.2912123799,-0.1940139383,0.1040057391,0.2185676098,0.2560478151,-0.1935175061,-0.0209170692,-0.2023592144,-0.1125776619,0.2611092329,-0.1710769832,-0.684820354,0.0519659072,-0.3113511801,0.0178578943,-0.1778743118,0.2989582121,0.2475437522,-0.0631885007,0.0114180548,0.4559701085,-0.2054668367,0.0613629557,-0.3999524415,-0.2040450722,0.0406922288,0.2002241611,0.0118118962,-0.0178570766,-0.1679245234,-0.0560864769,0.3344512284,0.4108452201,0.0995041579,0.2582981884,0.0189514179,0.2807394862,-0.0070325038,-0.1741079837,-0.3393336236,0.1816743016,0.5479783416,0.0701307803,-0.2165244222,-0.0037395069,-0.3784163296,0.0311311688,0.0122752711,0.0492905341,0.0164882336,0.1553180516,0.2597070932,0.2316702157,-0.03329641,0.5538898706,0.0001849726,0.304079175,-0.087167792,0.2223829329,-0.1737217754,-0.1235968173,0.1765867621,-0.3449999392,0.1806219965,-0.2068094462,-0.1502500921,0.2995541692,0.0429318473,0.0923922881,0.0319355465,-0.1928210557,0.277308166,-0.1518629789,-0.1848557144,0.3243631721,0.1011921838,0.0295746531,-0.2526453733,-0.0034100425,0.2128176242,0.2231773883,-0.2287004888,0.1309569925,0.0696672201,0.0324921906,0.1943875551,0.0144695174,0.3520648479,-0.2859787345,0.0815066993,-0.1778687537,-0.0075532594,0.3557195663,0.3431658745,0.0843384042,-0.2139345407,0.210994333,0.1186634749,-0.1527645886,0.00645896,0.3508174121,0.2607823014,0.187229529,0.2359099537,-0.1029999852,-0.1805161834,-0.1531548351,0.1440916806,0.0476365313,-0.0420634039,-0.0848126337,0.127474308,-0.0298430398,-0.0700510293,-0.3875254393,0.155090034,0.3091116846,0.0100950506,-0.0407449864,-0.238657102,-0.1384452283,0.2339590937,-0.1119870991,-0.3209646046,-0.5685407519,-0.0882725269,0.2417791188,-0.0517280474,0.074503012,-0.1412697732,-0.1699120402,0.3960913122,-0.2323114276,0.0296001863,-0.2021946758,-0.012911275,0.099621594,0.226380676,-0.1435131431,0.2954112589,-0.2167696059,-0.284511894,-0.4060569704,-0.0627470836,0.0189122111,-0.1931961328,0.3072670102,-0.2622005045,0.4708539546,0.029715132,0.1916576326,0.1111353263,-0.2169248015,-0.0593548417,-0.0859669521,0.0615779348,0.1041705459,-0.0240694899,-0.2971173227,-0.2785845399,-0.3193431795,0.1920940727,-0.1284935921,-0.0572024547,0.0882264897,0.23023507,0.0005931304,-0.0859977752,0.0964079723,-0.1858760118,-0.4200664461,0.5475694537,-0.0646037608,-0.2359232455,0.0955689922,0.238853097,-0.0694149956,0.2651896775,-0.6985403895,0.1034328267,-0.3445623517,0.2580441236,-0.0691322535,-0.1756858975,0.1853321493,-0.5137167573,-0.1090017334,0.0757054985,0.0311742164,-0.0786784291,0.1408202946,0.3928355277,-0.1135045812,0.2219333947,0.0011653552,0.1633586138,0.1411470175,0.0361656956,0.0860657394,0.1088595986,0.269469589,0.0133499419,-0.2448961288,0.0432928205,-0.0245824251,-0.1301951259,0.0362731107,-0.0228019692,-0.2440935969,-0.1782369912,-0.0317853354,-0.2656563222,-0.2509200573,-0.0718886405,-0.1711310446,-0.0590667129,0.3505452871,0.1751366109,0.0286385883,-0.2856182158,-0.0759189501,-0.0141120842,-0.0857097656,-0.1887924969,-0.530837357,0.1519774348,-0.3958620727,0.2477414608,-0.1128497794,0.2813424468,-0.0573084839,0.0412857048,0.1840789616,-0.0091780527,0.5779333711,-0.0978500396,0.0265222862,-0.0464659072,0.2701937258,-0.4184784293,-0.1023058444,0.1767010093,0.2083654851,0.52390486,-0.0358796492,-0.3270113766,0.1232310012,0.0759908333,0.3333395123,0.1616407931,-0.4344725013,0.0406917557,-0.1140852496,-0.1964607835,0.1450041085,0.0693398863,0.0846049264,-0.0492773019,-0.1960893869,0.2146620303,0.136298269,0.0859561116,-0.1118162125,0.1846216172,-0.3388823867,0.4967478216,0.0981099606,-0.1820325255,0.2965050638,0.6978909969,0.2273445278,-0.4742147326,0.0299002007,-0.056542404,-0.0924849436,0.3000592291,0.0675783157,-0.026383331,-0.2150105089,0.4110398889,0.0790702179,0.198398456,0.2396914512,0.0864008218,-0.4709188044,-0.480022788,0.1404571533,0.0652398095,0.0255754236,0.7254694104,-0.1011429653,0.1864499003,0.2429662943,-0.1138596237,1.0272656679,-0.2083415538,0.129852891,0.0084400047,-0.0949073806,0.4633690715,-0.3317197859,0.0691339672,-0.0628282353,0.0812856033,0.0680284575,-0.0684603974,-0.1782367527,0.3203916848,-0.2527095079,0.1336237043,-0.165655002,0.1418566555,-0.1555574983,0.2258991897,-0.296071589,-0.0161666162,-0.0694218948,0.2329600006,-0.0125219645,0.209631592,0.0571886674,0.1308651865,-0.2241944522,-0.0392896198,-0.6747110486,0.0534944423,-0.1191190854,-0.0541615449,-0.34417665,-0.2679920495,-0.0499538444,0.1242875308,0.1085304469,0.2949840426,-0.0722758844,0.317217499,-0.1573856622,-0.0684300661,0.0089837164,-0.238450706,0.1120078936,0.0059924466,-0.3947493136,0.2043996453,-0.0380242728,-0.2180021554,0.1471639872,0.2849779129,0.2082932442,-0.0738233104,0.061935287,-0.0328315645,0.1700588912,-0.2620545626,0.1565243602,0.3397988975,0.1525166184,0.2784980536,-0.1270455569,-0.3394216597,-0.4096662402,0.4108077586,0.0111025684,-0.0399703346,0.2374463677,0.2874158323,-0.4115638137,-0.2906267345,0.0608270802,-0.533208847,-0.3351154029,0.2490385324,-0.4125131965,0.3651679158,-0.0505498126,-0.14699395,0.1469285339,0.0774997622,0.1968628764,-0.5906020403,-0.0626296327,-0.1093577296,-0.2917641997,0.025508089,-0.1097740233,-0.0930524617,0.2142011225,-0.0552366972,-0.2938351333,0.1041359454,-0.2343114614,0.1285980791,0.3187293112,-0.1083915755,0.3753854632,-0.1432560682,0.0887853354,0.1932702214,-0.1047508419,-0.1710686833,0.0111001609,0.1229714677,-0.0484923162,-0.1617569923,0.1111613065,-0.4427928925,-0.0904797316,0.1024796367,-0.0471708551,0.1314586997,0.0653812736,-0.2333442569,0.2394034117,-0.0071447901,0.0015383622,0.20369789,-0.053377185,0.5051440597,0.2251237631,0.2564190626,0.1778503805,0.0664253011,-0.4852145314,-0.0160634406,-0.1081020609,0.0566375516,0.4679362476,-0.2862653136,0.049256999,0.3104961216,0.4150467515,0.5625929236,0.0313829631,-0.1846255958,0.4165688753,0.2802244723,-0.0111239953,-0.1266764104,0.3446883559,0.1587620974,0.0120021934,0.1286051124,-0.0494400673,-0.0430684984,0.1444677562,0.2352799773,0.6867423654,0.0569095649,0.1836857796,0.1002985612,0.1104222089,-0.040523693,0.4586593211,0.2718220055,-0.0080098454,0.1491598487,-0.1766095161,-0.1374293566,-0.0225102622,0.1113672182,-0.0861218646,-0.277199775,0.0229270235,0.1849718392,-0.2480676174,-0.0364360996,0.1623393595,0.1757764071,-0.2252263576,-0.1415656209,-0.1562655568,0.2823522389,0.1131406575,0.2638893127,0.2492496371,-0.1177722886,-0.1352325827,0.1094321534,0.053133212,-0.2901604474,0.3783887923,0.2352379858,0.1187347472,-0.6347622275,-0.013239488,-0.1434112191,0.1692614555,-0.4513990879,0.008547253,-0.0373290069,-0.2354823351,0.0193488859,0.1819727719,0.5284031034,0.1656909287,-0.0857023746,0.1386563778,-0.0580241419,-0.0462211035,-0.0590845123,0.4760381281,0.0833448991,-0.0854204148,0.427846998,0.126039207,-0.2391672134,0.1376172453,0.0726826414,0.2613367736,-0.2798641324,-0.0890470818,-0.023789661,-0.1442390084,-0.434555918,-0.1916846484,-0.1944942921,-0.0115776993,0.4266687036,-0.3257023096,-0.0417138599,-0.3187475801,0.1037225798,-0.1676358581,0.322306931,0.5741569996,0.2755864859,-0.1699838489,-0.2984001935,-0.4349109232,0.0711828247,-0.5885080099,-0.1753151119,-0.1917393655,0.4649917781,0.1373919845,0.3879269063,0.3642197847,0.0924223065,-0.3997996449,0.4755358994,-0.3340258896,-0.2970176637,-0.0493943393,0.0344098508,0.0781392902,-0.2549127042,0.1945787221,-0.2697484493,0.0031861621,-0.1564327627,-0.2445120364,0.1896850616,-0.0470439568,0.5041050315,0.0649292916,0.3030158877,0.2709483206,-0.4769492447,-0.4949392378,-0.2478876412,-0.0227418765,0.0370464884,0.0322431549,0.3441215754,-0.3469092846,0.0326548889,-0.1335644424,0.2137101889,-0.1422483027,-0.3069442809,0.2198387831,-0.2569900751,0.0220552348,0.4016231298,-0.0177027006,0.1567055732,-0.0345554613,0.1839788258,-0.2789247632,-0.3133023083,0.2667181194,0.0398919284,-0.3934292793,0.0392880775,0.0985214412,0.171653226,-0.0038625647,-0.2305061221,0.1069827527,0.3953539133,-0.0248194281,-0.213444978,0.1907691807,0.0378333777,0.1685906947,-0.0826662332,0.4905624092,0.0174225848,-0.2901221514,0.2927956879,-0.118358016]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2820","title":"Downloading \u201creddit\u201d dataset keeps timing out.","comments":"Great ! The timeout change will be available in the next release of `datasets` :)","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n","comment_length":15,"text":"Downloading \u201creddit\u201d dataset keeps timing out.\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\nEverytime I try and download the reddit dataset it times out before finishing and I have to try again.\r\n\r\nThere is some timeout error that I will post once it happens again.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"reddit\", ignore_verifications=True, cache_dir=\"\/Volumes\/My Passport for Mac\/og-chat-data\")\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\nI would expect the download to finish, or at least provide a parameter to extend the read timeout window.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nShown below in error message.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS \r\n- Python version: 3.9.6 (conda env)\r\n- PyArrow version: N\/A\r\n\nGreat ! The timeout change will be available in the next release of `datasets` :)","embeddings":[-0.332565397,-0.1843312979,-0.0898885429,0.1235421896,0.1154088974,0.20342125,0.0692033097,0.1948271543,0.0080888374,-0.1127361357,-0.0388174579,0.1267369688,0.3198180497,-0.0085039539,-0.0534012467,0.2132210582,0.1126272678,-0.2199580222,-0.0561524816,0.0619684793,-0.0259977113,0.0980479941,-0.1872974634,-0.1489140242,-0.220628351,-0.0725399777,-0.1046981961,-0.0864836276,-0.1910952181,-0.2625173628,0.2348134369,0.1820387691,-0.1121262386,0.5685609579,-0.0001091252,-0.1931512803,0.3014320135,0.0898453146,-0.3972260356,-0.5302575827,-0.2087555379,-0.1177487597,0.0930002406,0.061220292,0.1349250674,0.1349701434,0.0797204524,-0.3765141666,0.2988954782,0.3885107934,0.2085336894,0.3388108909,0.2051558644,-0.1082680672,0.1981246471,0.0113136321,0.0747744441,0.4239095449,0.2181670666,-0.1494691819,0.2451155335,0.0394018441,-0.0965771899,0.2266110033,-0.0973844752,0.0136142848,0.1451986432,-0.466991365,-0.0684821606,0.1813105047,0.8911020756,-0.021712631,-0.4001736641,-0.258687079,0.1316109747,0.0451037139,0.3587510884,0.4323689342,-0.3790051937,0.1980970055,-0.2649651766,-0.0067034294,0.0245510247,0.3593546152,0.1257101297,-0.0500247814,0.0599562377,0.2658752501,0.1105549186,-0.0422214232,0.2587308884,-0.3523654938,0.0610199235,-0.0886870623,-0.4141874611,-0.0804832205,0.0218804479,0.2407666445,0.1934000254,0.0341239087,-0.0387467556,-0.2400173843,-0.1413883269,0.1893355399,0.4078042209,-0.0674613193,-0.0250381436,0.1478048414,0.2107926309,-0.1134847179,-0.063833952,0.0535912849,0.0507275313,-0.001687436,0.1109579876,-0.1491280347,0.3092374504,-0.2474888116,-0.2105293274,-0.0269542653,0.125160262,-0.0809645131,-0.2457515597,0.1964538395,-0.5022117496,0.2473296523,-0.0735310167,0.1055911705,-0.3395315409,-0.4422063231,-0.0916342214,0.0280213058,-0.2442216426,-0.1150388494,-0.0712599307,-0.047228042,0.0468768962,0.2181249708,0.0374335498,-0.3078325391,-0.0163889453,-0.206464842,-0.1239155382,0.1609276831,-0.0937518403,0.4133200347,0.0374888629,0.1654024422,-0.0652775839,0.0276503228,-0.3055583239,-0.2328856885,-0.1218143851,0.2369157076,-0.2371754795,-0.3388174772,-0.1863103807,-0.0377295539,0.1709715873,0.0210310817,-0.0845663473,-0.0661638007,0.1166251376,-0.1068172753,-0.0984747112,0.3593381345,-0.4697035253,0.2985408604,-0.2076730877,0.0728382245,0.2052048892,0.2603732049,-0.2101275176,-0.0254207626,-0.1851094365,-0.1022650674,0.2913559973,-0.2133266628,-0.681019485,0.056532532,-0.3472000062,0.0144900512,-0.1609448344,0.3321326971,0.2602753937,-0.0993376374,0.0142122479,0.4430049062,-0.2201122791,0.0456481539,-0.3771333992,-0.1766481698,0.0516937077,0.1702148467,-0.0078179399,0.0026247192,-0.1507890224,-0.0129004139,0.3397394419,0.4552467167,0.0842002109,0.2524072826,0.0205734707,0.3088770807,-0.0064780796,-0.1913855225,-0.3553208411,0.138116613,0.5240982175,0.0408933237,-0.2050648481,-0.0284127519,-0.4121263325,0.0505748093,-0.0000021077,0.0729467496,0.0257431362,0.1518879235,0.2323637158,0.2388239652,-0.0077819144,0.5655297637,-0.0059307143,0.322208941,-0.1272541732,0.2284726799,-0.1741695553,-0.1189317331,0.1881214976,-0.3326369226,0.2035652101,-0.1496482641,-0.146946311,0.3235112727,0.057451345,0.1518115997,0.0318307318,-0.183159098,0.2925489843,-0.1807374805,-0.1830328107,0.340198338,0.1121049151,0.0572339483,-0.2511285543,-0.0207545441,0.2344272286,0.1870603561,-0.2349492908,0.1554428041,0.0685226843,0.0231037196,0.2082873434,-0.0297265444,0.3219424188,-0.3080088198,0.09263625,-0.2048376501,-0.0397098921,0.3594606221,0.3424139917,0.0979071632,-0.1939586997,0.1800067425,0.1248006448,-0.1532145888,0.0064187064,0.3499200046,0.2413647175,0.1969719231,0.2253946811,-0.095875442,-0.20352377,-0.1658248752,0.1265263557,0.0747643486,-0.0526282378,-0.1120940149,0.1223186627,-0.0058328584,-0.0606317855,-0.3991521895,0.1475276053,0.2961424887,-0.0177233238,-0.0119953807,-0.2549981475,-0.1382015795,0.2136799246,-0.0808982626,-0.3331087232,-0.5795000792,-0.0937828124,0.2571898997,-0.0573051646,0.044193089,-0.1736050397,-0.1520492882,0.4003954828,-0.1704133898,0.0116825104,-0.189627707,-0.0267506223,0.1072913557,0.1927867383,-0.122613132,0.3229691386,-0.2326078266,-0.2575902045,-0.4179015458,-0.0182764512,0.0373634733,-0.1815707982,0.3370355368,-0.2264543176,0.4573836029,0.0296916999,0.1729751229,0.1142643914,-0.2258428633,-0.0619651787,-0.0934238657,0.070507206,0.1020833775,0.0106945494,-0.3248880208,-0.235027805,-0.316229552,0.1535179764,-0.1548067331,-0.0545148477,0.0896725506,0.1977946907,0.009981053,-0.1039225012,0.1252058446,-0.1936142743,-0.4263325632,0.500207603,-0.079557322,-0.231386289,0.1009609327,0.2388248295,-0.0827391967,0.2645408511,-0.7040193081,0.0736210197,-0.3148973286,0.2641815245,-0.0818821043,-0.1966903955,0.18640095,-0.5066794753,-0.1164627522,0.0416130349,0.0295686778,-0.0704500601,0.1185321361,0.4156368971,-0.1115132794,0.2220181972,-0.0169511884,0.1548875421,0.1686028838,0.0414596945,0.0808101222,0.0879838765,0.2456268966,0.0407300964,-0.2441373467,-0.0002098388,-0.0093373945,-0.1418362409,0.044388257,-0.0214639772,-0.2164234668,-0.1938350201,-0.0503177531,-0.2807759345,-0.2775853872,-0.1303253174,-0.2086200565,-0.0134997973,0.3138259053,0.1774721891,0.0545960404,-0.2412600815,-0.0623092838,0.0242560003,-0.054005675,-0.196218133,-0.518239677,0.1494181901,-0.3419925869,0.2126323283,-0.082898967,0.3195330203,-0.1019363776,0.0668947771,0.1847009659,0.0213289317,0.5837029815,-0.1438643932,0.0639943033,-0.041873876,0.3054300845,-0.4174593389,-0.0732556731,0.1212947667,0.2215480953,0.4734686911,-0.0575774834,-0.2972850502,0.0651210323,0.0918284133,0.3058170676,0.1363924593,-0.4373358488,0.0346037149,-0.0944503397,-0.1677962393,0.148077637,0.077172935,0.0923233852,-0.0270763747,-0.1864672303,0.2143795341,0.1640803218,0.0745122582,-0.114336811,0.1905725151,-0.3220095038,0.4795773923,0.1085000038,-0.2212834656,0.2767581642,0.6749264598,0.2207655609,-0.4847328663,0.0185487214,-0.0146206599,-0.1072434932,0.284611553,0.0658755302,-0.004433251,-0.1812390834,0.3611374199,0.0625729486,0.1960870475,0.2665808797,0.0496378876,-0.4517481327,-0.5033305287,0.1771738231,0.065895848,0.0117840841,0.719861269,-0.1256470531,0.181616962,0.2188906968,-0.0778709948,1.0315356255,-0.1797211915,0.1018662304,0.0197582562,-0.1266185194,0.4483973682,-0.316336453,0.0452278443,-0.098411195,0.0641556159,0.0440440215,-0.0876465067,-0.1701059639,0.2907742262,-0.2612811923,0.143528223,-0.1905480921,0.1509622782,-0.1378015727,0.2003513277,-0.3368636966,-0.004560254,-0.0806891173,0.2376323193,-0.0004357065,0.2267304361,0.0255865306,0.1086976752,-0.1769999713,-0.0378316119,-0.6513311863,0.0642652959,-0.1341704428,-0.0497850068,-0.3370926678,-0.2558038235,-0.0276734103,0.1454526782,0.1298295408,0.2976088524,-0.071592249,0.2866640985,-0.1478486359,-0.1080525517,-0.022995254,-0.1987811327,0.150427714,0.0240461454,-0.417656213,0.2283020616,-0.0286010168,-0.1800690442,0.1448517442,0.2823989093,0.2047573924,-0.106223762,0.0670920685,-0.0494172573,0.1892652512,-0.2825333774,0.1651626974,0.3497476578,0.1672386378,0.228847608,-0.1107202247,-0.3023985624,-0.3804409206,0.4252966642,0.01397599,-0.0529895984,0.2292639464,0.2907762229,-0.4126146436,-0.3104163706,0.0755002722,-0.5295309424,-0.3556112349,0.2654338777,-0.4488745332,0.3835633695,-0.0795330852,-0.1437752396,0.1785692871,0.0673664212,0.2089008391,-0.6014701128,-0.0501641296,-0.0748803467,-0.2990021706,0.0008020907,-0.0611567609,-0.1183697358,0.2324630767,-0.0524831861,-0.3133786321,0.0791647062,-0.2534817457,0.131735757,0.322996676,-0.1072486416,0.3681426346,-0.1318911612,0.0953141972,0.1483918279,-0.124208279,-0.1744576693,-0.0144083397,0.117457673,-0.074741818,-0.1516759396,0.0916461274,-0.4514706731,-0.0853850618,0.1140954867,-0.0319341943,0.1217533574,0.0702395812,-0.1928872019,0.2283814251,0.0119228642,-0.0264302902,0.1748304963,-0.0681027099,0.4618448317,0.2131423205,0.2228418291,0.1763655692,0.0428800397,-0.4920851588,-0.0153298629,-0.0792124867,0.0177037288,0.4668798149,-0.2738249004,0.0817041993,0.3484230936,0.4055718184,0.5130241513,0.0192066487,-0.1529607177,0.3950096667,0.2907663882,0.0115008401,-0.1424344182,0.32559219,0.189084515,0.0098868804,0.1148243323,-0.0518186428,-0.0406497829,0.1015927717,0.246859625,0.6968792081,0.0084672086,0.2089839131,0.1545238793,0.0882261097,-0.0255789496,0.4657320082,0.2926680744,-0.0189954769,0.1164047793,-0.1435752958,-0.1191978157,-0.0107773589,0.1063177809,-0.072024785,-0.2935923338,0.0366912261,0.1817636639,-0.2606488168,-0.0289472304,0.1459549814,0.1911061555,-0.2283805609,-0.0879242644,-0.1606379449,0.2806740403,0.0840837806,0.2524560392,0.2495897114,-0.1197518706,-0.1521423906,0.0840588585,0.0603456683,-0.2768293023,0.3771804273,0.2152775228,0.1361166686,-0.6255585551,-0.0017479411,-0.0882864594,0.1864312887,-0.4565320909,0.0005294789,-0.0254332572,-0.2339548618,0.0152835669,0.1944603473,0.5240771174,0.1339980811,-0.0916119963,0.105875507,-0.0330090895,-0.068114996,-0.0270495266,0.4576022327,0.0753960907,-0.075382039,0.4270864725,0.1439127773,-0.2385238707,0.1826465428,0.0872111842,0.2724862993,-0.3126870394,-0.029425431,-0.0577726141,-0.1541365087,-0.4035849571,-0.1798866987,-0.2024888992,-0.0191920772,0.4477645457,-0.3383000493,-0.0512937717,-0.3140487373,0.1078596264,-0.1230980828,0.3184839785,0.5826185346,0.2792212367,-0.154325217,-0.2865278125,-0.4265450835,0.0884305462,-0.5802479982,-0.1833609939,-0.1429676414,0.4548461139,0.101003401,0.4136739075,0.3601632714,0.0577728115,-0.4000045955,0.4943615496,-0.2935335338,-0.317959249,-0.0293882843,0.0138507895,0.0600167736,-0.2512103021,0.1815416068,-0.2636130452,0.0049526356,-0.1743647158,-0.2442423552,0.2019723356,-0.0929164812,0.5008657575,0.0723370463,0.2897492647,0.2557418942,-0.4755592346,-0.4844291508,-0.267275393,-0.0113760903,0.0238975435,0.0229856651,0.3448058665,-0.337767303,0.0154124191,-0.1105157733,0.1843410432,-0.1265572309,-0.2969739437,0.1717035323,-0.2617947757,0.0158320814,0.3815042675,-0.010380582,0.1345930248,-0.0313098319,0.1885896325,-0.2590397596,-0.3457368314,0.2579984963,0.0385637023,-0.3825317919,0.0210036598,0.1547961533,0.1468198001,-0.0303863529,-0.2383127213,0.1209158078,0.4083229005,-0.0261226688,-0.2012388706,0.2060900182,0.0548751578,0.1820363849,-0.0315946043,0.4965365231,0.0253865924,-0.2866041064,0.3604632616,-0.1326176673]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2818","title":"cannot load data from my loacal path","comments":"Hi ! The `data_files` parameter must be a string, a list\/tuple or a python dict.\r\n\r\nCan you check the type of your `config.train_path` please ? Or use `data_files=str(config.train_path)` ?","body":"## Describe the bug\r\nI just want to directly load data from my local path,but find a bug.And I compare it with pandas to provide my local path is real.\r\n\r\nhere is my code\r\n```python3\r\n# print my local path\r\nprint(config.train_path)\r\n# read data and print data length\r\ntarin=pd.read_csv(config.train_path)\r\nprint(len(tarin))\r\n\r\n# loading data by load_dataset \r\ndata = load_dataset('csv',data_files=config.train_path)\r\n\r\nprint(len(data))\r\n```\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nC:\\Users\\wie\\Documents\\\u9879\u76ee\\\u6587\u672c\u5206\u7c7b\\data\\train.csv\r\n7613\r\nTraceback (most recent call last):\r\n File \"c:\/Users\/wie\/Documents\/\u9879\u76ee\/\u6587\u672c\u5206\u7c7b\/lib\/DataPrecess.py\", line 17, in \r\n data = load_dataset('csv',data_files=config.train_path)\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\load.py\", line 830, in load_dataset\r\n **config_kwargs,\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\load.py\", line 710, in load_dataset_builder\r\n **config_kwargs,\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\builder.py\", line 271, in __init__\r\n **config_kwargs,\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\builder.py\", line 386, in _create_builder_config\r\n config_kwargs, custom_features=custom_features, use_auth_token=self.use_auth_token\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\builder.py\", line 156, in create_config_id\r\n raise ValueError(\"Please provide a valid `data_files` in `DatasetBuilder`\")\r\nValueError: Please provide a valid `data_files` in `DatasetBuilder`\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: win10\r\n- Python version: 3.7.9\r\n- PyArrow version: 5.0.0\r\n","comment_length":29,"text":"cannot load data from my loacal path\n## Describe the bug\r\nI just want to directly load data from my local path,but find a bug.And I compare it with pandas to provide my local path is real.\r\n\r\nhere is my code\r\n```python3\r\n# print my local path\r\nprint(config.train_path)\r\n# read data and print data length\r\ntarin=pd.read_csv(config.train_path)\r\nprint(len(tarin))\r\n\r\n# loading data by load_dataset \r\ndata = load_dataset('csv',data_files=config.train_path)\r\n\r\nprint(len(data))\r\n```\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nC:\\Users\\wie\\Documents\\\u9879\u76ee\\\u6587\u672c\u5206\u7c7b\\data\\train.csv\r\n7613\r\nTraceback (most recent call last):\r\n File \"c:\/Users\/wie\/Documents\/\u9879\u76ee\/\u6587\u672c\u5206\u7c7b\/lib\/DataPrecess.py\", line 17, in \r\n data = load_dataset('csv',data_files=config.train_path)\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\load.py\", line 830, in load_dataset\r\n **config_kwargs,\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\load.py\", line 710, in load_dataset_builder\r\n **config_kwargs,\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\builder.py\", line 271, in __init__\r\n **config_kwargs,\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\builder.py\", line 386, in _create_builder_config\r\n config_kwargs, custom_features=custom_features, use_auth_token=self.use_auth_token\r\n File \"C:\\Users\\wie\\Miniconda3\\lib\\site-packages\\datasets\\builder.py\", line 156, in create_config_id\r\n raise ValueError(\"Please provide a valid `data_files` in `DatasetBuilder`\")\r\nValueError: Please provide a valid `data_files` in `DatasetBuilder`\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: win10\r\n- Python version: 3.7.9\r\n- PyArrow version: 5.0.0\r\n\nHi ! The `data_files` parameter must be a string, a list\/tuple or a python dict.\r\n\r\nCan you check the type of your `config.train_path` please ? Or use `data_files=str(config.train_path)` ?","embeddings":[-0.0574387945,-0.0158846732,0.0742202699,0.5506054163,0.3027433455,-0.1935664266,0.4147061408,0.0769136027,0.0825464055,0.0638927221,-0.0089269225,0.586360693,0.0122840842,-0.0284527745,0.1138430834,-0.0146789085,0.1873939633,0.085528031,-0.12411616,-0.2560271919,-0.3155732453,0.1688574553,-0.1460751295,0.1901938915,-0.0699701011,0.2109431922,-0.2822049856,0.4675122499,0.0227660947,-0.2501043975,0.4729166329,-0.280921042,0.4408823252,0.4645849764,-0.0001151561,-0.0097315684,0.475538075,0.0375795662,-0.2390377969,-0.607242763,-0.1662139148,-0.2815756202,0.4258596599,-0.1936817765,-0.066541642,-0.1174363792,-0.223506853,-0.1684545875,0.4106475413,0.4155715406,0.1782934666,0.1498636007,-0.2141826451,-0.1690315753,0.3525300026,0.1742020547,0.0461488776,0.4088488817,0.0037086802,-0.251622349,0.1068805382,-0.0742010474,-0.1378999054,0.0324060842,0.2563998401,0.2106460184,-0.0583171584,0.002279914,0.1797527224,0.3070548475,0.5116797686,-0.0626182184,-0.3027721345,-0.0975569487,-0.0389354527,-0.1113292128,0.2599177659,0.0741580278,0.0159020461,0.1967173517,0.1175490245,0.2056593299,-0.0650330633,0.4197382033,-0.1461278051,0.1639842838,-0.1312114894,0.0889243931,0.0116657214,0.1690495759,0.6599962711,-0.0752312168,0.0472111329,0.3930088878,-0.1994341314,0.127945751,0.3355509639,0.1851821095,-0.0519484505,-0.0531266108,0.3470786512,-0.1044920757,-0.1862101108,0.0079265162,0.3611758351,0.1985619366,-0.3046008646,0.2134357691,0.189971745,0.2259806991,-0.0859122425,-0.1602514088,-0.2420356125,-0.2013465911,-0.1269379407,0.1843207031,0.3140102029,-0.3728873134,-0.2214430571,-0.0581827797,0.0346098766,0.1836173087,0.2119463384,0.5136289597,-0.2088429928,0.5151565671,0.0225076247,0.2508969009,-0.0922638774,0.1784021258,-0.2614133954,0.0694378614,-0.1460306197,-0.1126512215,-0.0430093072,0.021556098,0.2521870434,-0.0096220551,-0.1722253859,-0.3173284531,-0.3492340446,-0.4206415713,0.0861275718,0.0746736377,-0.0392393172,0.0950153396,0.2757533491,-0.0933145955,-0.2179812938,0.1728485972,-0.2247143984,-0.4391539097,-0.3228242695,0.1312871128,-0.2310711443,0.0401465781,0.088622123,-0.2725886703,0.0588152781,0.3045935929,-0.1377795637,0.1218146011,-0.3652853966,-0.2556673288,0.4419429898,0.6512079239,-0.4164176881,0.2217454165,-0.2636246383,0.0769634545,0.410280019,0.1699589044,-0.3566013277,0.381391257,-0.1161803231,-0.151099667,0.3857566714,-0.5012984872,-0.3714636862,0.599517107,-0.0560738929,-0.2375229001,-0.0514174961,-0.0200658031,-0.1240407601,0.0844895691,0.3112476468,0.4983115494,0.1864834577,0.1523343176,-0.4068825841,-0.4395462871,-0.126450792,0.2480695993,-0.118607074,0.05363556,0.241335094,-0.3943933249,0.3047114909,-0.0594210476,-0.2168198824,0.4881590903,0.3690141439,0.0848213509,-0.0759779662,0.1972519308,-0.2346253097,0.2298525572,-0.0617966019,-0.1506754607,-0.1290706098,-0.003085624,-0.4466368556,0.0966235995,-0.2612614632,-0.1387730539,0.0527373739,0.1784128845,-0.0079229819,0.0783594474,-0.1965729594,0.1575737298,-0.1202824935,-0.1758037955,-0.1213267222,0.2304284424,-0.0853113458,0.0754689351,0.0699482262,0.4098211527,0.216741994,-0.0333156399,-0.0657609478,0.3856574595,0.1693173498,0.2863016129,-0.1166194975,-0.3804411292,0.2069358528,-0.0697807446,0.0948281735,0.0371947475,0.2641209066,0.1456101686,-0.2078539133,0.1532488018,-0.1123906225,0.0651696399,0.045180466,0.0005378057,0.5214226842,-0.2430495471,0.0346963443,-0.3120870292,0.2649819553,0.4043836594,0.5120335221,0.1754516065,-0.3331454694,-0.0802893192,0.1967051625,0.1612119824,0.0592381768,0.0705425516,-0.0783512518,-0.0168797802,-0.0027853975,0.4694555998,0.4198943377,0.1177514121,0.146663323,0.1681153625,0.0641562715,-0.1689205617,0.206251353,-0.0474292114,0.2692136168,0.0539128967,-0.0148290703,0.3034391105,-0.1236819029,0.0268181115,0.0028711224,0.299459219,-0.3273457587,0.0614307411,-0.3752581179,-0.1776741594,-0.145605877,-0.331272155,0.2078628689,-0.1352641135,-0.3229803741,0.032826785,-0.1912081838,-0.0376699418,-0.1218343377,0.0566690713,-0.0594361313,-0.2732097805,-0.2022083998,-0.1779984683,-0.525094986,-0.0782819316,0.3557379246,0.157008186,-0.087319918,-0.177217871,-0.0133176837,-0.099757798,0.1238210499,0.02284877,0.178985104,0.6002935767,0.1714735478,0.0395605043,0.1790844202,-0.1626320332,0.2645680308,-0.2557523549,-0.1662978083,0.006235044,0.0991331264,-0.1168443635,-0.1966537833,-0.1383365691,-0.3782124221,-0.2874901891,-0.0389488526,0.0985015407,0.1825945973,0.1845143437,0.216906026,0.2336980253,0.1198514402,-0.077794157,-0.1604256183,-0.4417062402,0.2158462107,-0.028256245,-0.5365150571,-0.3180679977,-0.1649757177,0.3878697157,-0.152408421,-0.5176630616,-0.0888612121,-0.2314541489,0.1201902553,-0.0813609213,-0.1304138452,0.346337527,-0.1021287441,-0.0298768412,-0.0978019238,-0.0786362141,0.0038152558,0.183000952,0.0856306627,0.1478830129,0.3921531439,-0.1356151104,0.6314616799,0.0274883565,-0.0541715883,0.5767683983,-0.1661159545,0.4468920827,-0.1700372249,-0.3950130641,-0.2945566177,0.0130058555,-0.2421169132,0.3109371364,0.0554624498,-0.347533226,-0.3655242324,0.0218573827,-0.4551153183,-0.1747784019,-0.1583245695,-0.2738893032,0.0217914227,-0.1425581872,0.1417834759,-0.0741286799,-0.0378821865,0.0210161135,0.0313278809,0.03297057,-0.0628153831,0.1278610677,-0.0329137221,-0.4352096915,0.1097882614,-0.0478189327,0.4363663495,0.006618523,-0.2862240672,0.0781521127,-0.0584604293,0.6754683852,0.0223823413,0.1359011531,0.1814872026,-0.0211425666,-0.6554043293,-0.0295493454,0.1211227551,0.0229199473,0.0767608136,0.796644032,0.1024092734,-0.1237041876,0.0252632182,0.0859775096,0.0182036962,-0.4049771726,-0.3497110009,-0.3659102619,0.0984825492,-0.1340129077,-0.0836043432,0.3783411384,-0.0212986916,-0.1552290022,-0.086149931,-0.2001954764,-0.2223150283,0.1005428582,0.3425101638,-0.0169256367,0.2756721079,-0.0321035348,0.3114451766,-0.1748637855,0.6477417946,-0.1171410605,-0.3216336668,0.0788393393,-0.3234080076,0.0986616537,0.2470776439,-0.1465616673,0.0478248075,0.0174015816,-0.2901143432,-0.0325319953,0.1599928886,-0.0270323791,0.0252429135,0.0956066251,-0.2298050821,0.517645061,0.1183605492,-0.1103813797,0.3470835686,-0.5550871491,-0.2694754303,0.2309324294,0.239632234,0.8387879729,-0.1660595536,0.1618093699,0.320600003,-0.2457161695,0.2124182731,0.1730218828,-0.1323700994,-0.4516242743,-0.1949019432,0.069234632,-0.1457591653,0.1558912843,-0.0642124265,0.0811752826,0.0778916329,-0.1818905026,0.3544170856,0.3318727612,0.0835025236,0.054602012,-0.1341724694,-0.3800011873,0.0537459552,0.2156202495,0.1067383438,-0.0813261792,0.0350171961,-0.2643424869,-0.1704931408,0.0468566269,0.3553574979,-0.3430225253,0.0182530973,-0.107673876,-0.4711862504,-0.3164750934,0.6356455684,0.2303337604,0.0492921919,-0.2496906221,0.0883892402,0.035733968,0.1797053814,-0.0138684912,-0.1586066037,-0.0513027012,-0.1195732951,-0.1252864301,-0.1987668574,-0.2385016084,-0.1723289341,0.1757486016,0.3630549312,0.1982508898,-0.0026166497,-0.3738718927,0.0690092966,0.3177512884,-0.2005014867,0.1159919724,-0.0185712557,-0.0291237906,-0.1364469528,-0.1668360829,-0.4678330719,0.0192847811,0.5623624921,0.00611198,0.5865507722,0.490696907,-0.0581142716,0.0934346914,-0.1508752704,0.2538242638,0.0914068371,-0.1445089728,-0.1839631945,-0.1803078949,0.0121869957,-0.1138370112,0.0961937606,0.3451667428,-0.1079843044,-0.1119632646,-0.6838929057,-0.448482573,0.0404831022,-0.1864709854,0.0638710037,0.0737024248,0.1408946365,-0.0244399644,-0.2554089427,-0.2226366401,0.2571768463,-0.4180083871,0.0198031422,-0.1412283778,-0.0413050689,0.24452281,-0.0154003156,0.1508336216,-0.1639546603,0.0708342493,-0.1837785244,-0.1056208014,0.1013326868,0.085903883,-0.1847527474,-0.1455666125,-0.2703591883,-0.0890598744,-0.0855412558,-0.2252631932,0.2039053142,-0.0406455807,0.0950419009,-0.1309648007,0.3166359067,-0.2054282129,0.3459956646,-0.3354070783,0.010914214,0.149894461,0.1973810792,0.0708735287,-0.0026603967,0.1882465184,-0.2834925354,-0.2006029338,0.1394177526,0.2105656564,-0.092192255,0.0599513575,0.1898698211,0.0745634884,0.2924563587,-0.1836612225,0.099225536,0.5471073985,0.1454543918,-0.2937477827,-0.0604266897,-0.0628634617,-0.1081198007,-0.0155177945,0.18011792,0.1219177693,-0.2686055005,0.0843513682,-0.0607593544,0.4588413537,0.0993796885,0.1458263099,0.3395939171,-0.2543954253,0.1626715958,-0.0428986587,0.2295009196,0.0336795188,0.5187291503,0.0486087054,-0.0806713253,-0.2998912632,0.404823035,-0.0787528455,-0.6417547464,0.3375591338,0.0485763289,-0.3438985646,-0.2575294375,-0.2667192221,0.2372455001,-0.0933442935,0.1150254458,-0.1276447773,0.1020797864,-0.0705744997,-0.2714782953,-0.4754407108,-0.1534811556,0.1385405958,-0.0363482945,-0.009251547,-0.1910673529,-0.084154062,0.1059214026,-0.2127449811,-0.2071151435,0.1263237894,0.3150521219,-0.1272409409,-0.316398412,0.4282380342,-0.2456131577,0.2148877382,-0.0139596118,-0.0502042845,0.2021694928,0.6500501633,-0.0500279628,-0.1874515265,0.1273712516,-0.195182845,-0.1654112935,0.1244853139,0.0986938849,0.4059478045,0.5674875975,0.1458260268,-0.1257262528,0.382663548,0.0502285734,0.4492844939,-0.2906240225,0.3194684386,-0.0897679776,-0.0078343432,-0.0442407504,0.2483790815,-0.416639477,0.073987402,0.230932191,-0.0469475649,0.368147552,0.0715056956,0.0608653016,0.0146221211,0.4513391256,0.4840821922,0.3376534283,-0.2196336687,-0.2611141205,-0.6287825108,0.12541686,0.0330737717,0.1006624103,-0.0362791307,0.1746265143,-0.1181007624,0.0348033048,-0.0451579317,0.3480482101,0.0746761262,-0.0032194327,-0.1826473325,-0.3485901952,-0.3953167796,0.0473146178,-0.0942979828,-0.4356854856,-0.0121801393,0.1142916009,0.0336442627,0.1058668792,-0.2598726451,-0.1744865924,0.103974469,0.3186103404,0.1113508493,0.1091753095,0.0312557705,0.1002476215,-0.379299283,-0.1152853072,-0.073720254,0.2165313363,0.0278788358,-0.3506174386,-0.1900012493,0.1097136959,-0.3974589109,0.1685300022,0.0821398497,0.119525224,0.0046526068,-0.169753328,-0.0053001377,-0.0939360783,-0.0222134106,0.0084341569,-0.2914646268,0.041347377,-0.0685116723,-0.0535092391,0.3320354521,-0.4292975068,-0.0542798042,-0.0027846175,0.1854250282,-0.0521517396,0.1116239205,-0.0562810972,-0.1143700555,0.4793233573,-0.0975359008,-0.005254426,-0.0017493165,-0.1760842651,0.2696327865,-0.1027432382,0.172288239,0.0208203886,0.0066081607,0.0037639646,-0.3554125726]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2816","title":"Add Mostly Basic Python Problems Dataset","comments":"I started working on that.","body":"## Adding a Dataset\r\n- **Name:** Mostly Basic Python Problems Dataset\r\n- **Description:** The benchmark consists of around 1,000 crowd-sourced Python programming problems, designed to be solvable by entry level programmers, covering programming fundamentals, standard library functionality, and so on. Each problem consists of a task description, code solution and 3 automated test cases.\r\n- **Paper:** *link to the dataset paper if available*\r\n- **Data:** https:\/\/github.com\/google-research\/google-research\/tree\/master\/mbpp\r\n- **Motivation:** Simple, small dataset related to coding problems.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":5,"text":"Add Mostly Basic Python Problems Dataset\n## Adding a Dataset\r\n- **Name:** Mostly Basic Python Problems Dataset\r\n- **Description:** The benchmark consists of around 1,000 crowd-sourced Python programming problems, designed to be solvable by entry level programmers, covering programming fundamentals, standard library functionality, and so on. Each problem consists of a task description, code solution and 3 automated test cases.\r\n- **Paper:** *link to the dataset paper if available*\r\n- **Data:** https:\/\/github.com\/google-research\/google-research\/tree\/master\/mbpp\r\n- **Motivation:** Simple, small dataset related to coding problems.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nI started working on that.","embeddings":[-0.2849132717,-0.133457914,-0.2878871858,-0.1627224237,0.0817829296,0.1678224653,0.1912839413,0.4287959337,-0.0074734227,0.210512802,0.2690508068,0.1734536737,-0.2547970116,0.573654294,0.2218417376,-0.1999996901,-0.1182349026,0.2328725308,0.2143574059,0.0789715797,-0.0444036089,-0.1006354392,-0.1761155576,-0.0717737973,-0.0602941476,-0.2451325059,-0.1391011029,0.1347881705,-0.2574696243,-0.4787394404,0.0592083596,-0.0328190811,-0.0435738936,0.2327648848,-0.0000939827,-0.1238563731,0.2540135086,-0.0613972284,-0.3312395513,0.0047494345,0.0648178086,-0.2565566897,-0.0716578811,-0.1465823054,-0.1690939963,-0.172266975,0.0171775371,-0.2113060057,0.4378979504,0.204744041,0.3965813816,0.3091777861,-0.006565854,-0.1522907019,0.216149956,0.024506947,-0.3803349137,0.1085891798,0.4649888277,-0.1904764622,-0.0253517609,0.0958340913,-0.0297297817,0.166222617,0.1383808106,-0.1287764162,-0.0920542553,-0.3865365684,0.0112141734,0.2208591253,0.2777346075,-0.2721240819,-0.199903965,-0.4757342637,-0.0428696312,-0.3654705584,-0.3429593742,0.2389941216,-0.1624166071,0.2200182825,0.0626452565,-0.1403585076,-0.1508650333,0.0799098536,0.1195256338,0.186825335,-0.1486431509,-0.2035667002,0.0157113913,0.0141214374,-0.0299392939,-0.043952506,0.0844788328,0.1030525267,-0.0776382312,-0.326043278,0.4767831266,0.0591339916,0.3409461379,0.1629697233,0.0790242255,0.1004548818,0.1058621556,-0.0145056518,0.0812987313,-0.0355491675,0.0155223291,0.0251007862,0.4136144519,0.0323383398,0.0539993495,0.2117567509,-0.0601265579,0.0352759212,0.2056613415,0.0244944692,0.1068023965,0.08631479,-0.2509212494,0.174749285,-0.1515491754,-0.0128302816,-0.0539118201,0.2836790979,-0.0774912685,0.0061077382,-0.1606379747,0.2821772695,-0.1541848183,-0.0615469925,-0.194296211,0.0142778065,-0.0936883837,-0.1387249231,0.3490798771,-0.0221535675,0.2100770026,-0.0154607417,-0.0234164577,-0.0740897432,0.527869463,0.1111602262,0.0096492833,-0.0003645252,0.010580943,-0.1216617376,0.0536291227,-0.1226709709,-0.218266204,0.0287689678,0.0097443331,-0.1943154335,-0.2205077559,0.3808838427,-0.0785726458,-0.03023188,-0.0619468652,0.2399370819,-0.0009556578,-0.0700846314,-0.0413206704,0.107826665,0.0277574304,-0.2108882815,0.2414938956,0.3934024274,0.1215469688,0.029138796,-0.0930505171,0.1230977476,-0.1034983024,0.0949857607,-0.0594481155,0.4386395812,-0.1583972126,0.0002152469,0.1732854694,-0.1555610001,-0.2258829772,0.2075191289,-0.1442385614,-0.0984919816,0.0532773063,0.2819795907,0.0427116863,-0.0072683352,0.1027461216,0.3658614755,-0.277700454,0.2478543371,-0.2792246938,-0.1979110539,-0.0562232845,0.2866658568,0.2445278317,-0.1036788374,-0.0271261968,-0.141008243,0.104257524,-0.2878671885,-0.1209466979,0.1444403976,0.3846429288,-0.206984058,-0.0294839405,-0.4035709202,-0.1740601361,-0.1135928109,0.1878543347,0.0001660053,0.1051434875,-0.1259521991,-0.3165053129,-0.0199470874,-0.1153551638,-0.1701505333,0.3375154436,-0.0397514366,0.0888382196,-0.1658346653,-0.2843676805,0.2442146987,-0.1412384808,0.0214176681,-0.1593115628,0.2416948825,-0.1673507988,-0.159970969,0.3613834977,0.1923901439,0.0006691717,-0.0876550302,0.1539141536,0.1522364467,0.0688945875,-0.068937093,0.3547674716,0.2883714736,0.0814196095,-0.2527677417,-0.0614659935,0.4031417072,-0.0114574805,0.0475117564,-0.112842761,0.4404624403,0.1782226264,0.1843335032,0.1788933128,0.0583372042,0.080048278,-0.0155238835,-0.10869921,0.0456027873,0.3763151169,0.0773899853,0.2836897671,-0.1078732833,-0.5110245943,0.1460869461,0.2634853721,0.0052963854,0.4311172366,0.1204944625,-0.1064375415,-0.0698128119,0.0869167075,0.0788849518,0.1781291515,0.36779809,-0.0641684383,0.1296677142,-0.0156422555,-0.1245661154,0.0670226514,0.0587196611,-0.119154036,0.0994987264,0.1193532199,0.0447520949,-0.1776721478,-0.1595429033,-0.356238693,0.0591747127,-0.201352492,0.0800890699,0.0072302818,-0.0371136218,0.1928287596,0.1426100731,-0.0705291778,-0.309636116,0.3106985092,0.0039120247,-0.3180580437,0.1192706749,-0.1378332078,0.3182047606,-0.1154433787,0.059579879,-0.0891058072,-0.4489895403,-0.0686900169,0.3469485641,0.1656770557,0.0099303592,0.7775645852,-0.4257620871,0.0841796324,-0.1114786789,-0.5076244473,0.0903270319,-0.10588222,0.4892371297,0.2175707072,0.2885817587,-0.1303798854,-0.1785420328,0.0121866399,-0.3277238011,-0.0488845818,-0.0984865427,0.0274520759,-0.107781373,-0.1009424031,-0.3282198012,-0.213459447,-0.5073335171,0.2270195782,0.1714576632,0.1278980374,0.0494404919,0.160942167,0.0711264536,0.2024968266,0.3002057374,-0.0295888744,-0.331191659,0.0964476764,-0.3714588284,-0.2795067728,-0.2746471763,-0.3080466688,0.0787712857,-0.106021367,-0.2787248194,-0.0600926355,-0.1415893883,0.2842505276,-0.0094558261,0.2864073217,0.0421348065,0.1579408795,-0.2639485002,-0.0872828513,-0.2491760701,-0.1346005648,-0.0176452603,0.0338529721,-0.0915968344,0.1364734024,-0.029201759,0.6323778629,-0.0437675975,0.0396507345,0.274980545,-0.0951089412,0.3145446777,-0.2266472131,-0.2177451551,0.1323497742,0.1845987737,-0.1663608253,0.3419016004,0.2598410249,-0.0545279011,-0.1045615301,-0.1451695859,-0.3454625309,-0.0577728003,0.0807973966,0.0073522013,0.1809813082,-0.0353948623,0.1664307714,-0.3587761819,0.0026473503,-0.0290876031,0.1582148671,0.0133782281,-0.1972995251,-0.3953419924,0.0645690337,-0.3348983824,0.3358369768,0.041489657,-0.1409664154,-0.1448580474,-0.2054610848,-0.1375602335,-0.1178281382,-0.1000991389,-0.1408055574,-0.0400669165,0.0846497864,0.0220727958,-0.2038372308,0.0480490327,-0.2899057269,0.3177028,0.3884712458,0.2171136886,-0.3953396976,0.0446692929,0.2301255018,0.1758732796,0.0652521551,0.0362986624,-0.1971364319,-0.5246984959,-0.1820788234,0.2671554685,0.0400625467,0.0144246863,0.100121811,-0.0905779824,-0.1432419866,-0.1190516353,0.2236286998,0.1664918214,0.0686051697,-0.1021133438,0.2474269271,-0.2636699378,0.0543497987,0.2084483057,0.2128447294,-0.1321889013,-0.017673023,0.1463393718,-0.1298807561,0.4288018048,0.2115851939,-0.0310485028,-0.0317209437,-0.009645585,0.2370041609,-0.3951739371,0.1484133005,0.2273263335,0.0173148867,-0.3671790063,-0.3667575717,0.2812284529,0.0609747842,-0.1191611961,0.2862706482,-0.1520745754,-0.0427081399,0.2569087744,0.0041393614,0.7928612828,-0.0999678522,0.1300785542,0.1561889946,-0.4552157223,0.2975474298,-0.0260259863,0.0869164616,-0.3499881327,-0.1780227423,0.1301014572,-0.1910932213,0.0204733573,-0.1164493561,-0.1875609756,0.2096138299,-0.1517485827,0.0211159699,0.0806960389,0.4276665747,0.0035557202,-0.1122982949,-0.6275213957,0.3527128696,-0.0592700541,0.055084005,-0.0957840011,-0.094110474,-0.1216754392,-0.1753355712,-0.3040816784,0.0688271523,-0.0819580257,-0.0140285194,0.1852906495,-0.289991349,0.0614970028,-0.0560386032,0.2199302316,0.411085546,-0.1494042277,0.0876062661,-0.0459685549,-0.0399256237,0.1574093699,0.1865114123,0.1310010403,-0.1573210806,-0.077538088,0.1257893145,-0.0165198147,-0.3455993235,-0.4156636,0.0761799514,0.1302779615,-0.235797599,0.0676300004,-0.0486941151,-0.0925663784,-0.3562155366,0.319152832,0.2949682474,0.0340279639,0.1288042068,0.1629132926,-0.0830563828,-0.1699594557,-0.1456392109,0.2088894844,-0.0103840297,0.097329244,0.0497086309,-0.1631789953,-0.5027208924,0.2805196047,0.274531424,0.1604718417,0.2389351726,0.0471353605,-0.2242142707,-0.0470979102,0.0503411666,0.1080031022,-0.0275026374,-0.0657624975,0.0156294145,-0.3691862226,0.1414543986,0.1368465126,0.3040755689,-0.0650479719,0.1141097024,0.0034399545,0.0501704849,-0.5499732494,-0.0523175299,-0.383456409,-0.0199775826,0.0414214283,-0.1249154359,-0.0245237965,0.0321942084,0.3503103256,-0.0804084241,-0.3943235576,-0.4513368309,-0.1011980847,0.033290863,-0.2454203218,0.0175139047,0.3055770993,0.0559042506,-0.1150428206,-0.2049759179,0.1094421893,-0.0628470033,0.1027886942,-0.0570826009,-0.0378283747,0.1265892237,0.1321827918,0.0052135549,-0.1836564839,0.2465239316,0.1092364639,0.0875936896,0.1570901424,-0.048100505,0.2043523043,0.1361744851,0.1567183584,0.0757217184,0.2024168521,-0.1890129745,-0.0768097863,0.3814417422,0.2264214605,0.2491586655,-0.138778463,0.2964679003,0.0682555437,0.4484824538,-0.2751523554,0.0099071655,0.13540712,-0.0499501899,0.3374973834,0.2310186923,0.1051962748,-0.0294341668,-0.0045157555,-0.0359240361,0.3266936541,-0.0052267071,0.1200972795,0.0908199325,-0.0949622244,-0.163230747,-0.1218854561,-0.0158446003,0.1250858158,0.0669253096,0.0373379514,0.3269079924,0.0918026641,-0.0028089383,0.0019441035,0.0687978044,0.3089620769,0.1360459626,0.0662463307,0.1082996354,0.119298853,0.3474561274,-0.1851810366,0.0759227201,-0.1562440991,-0.1139819026,-0.0642299429,-0.0785170272,0.0372920707,-0.1200308502,-0.2012816817,-0.1325361431,0.0439911485,-0.1894602925,0.1587169021,-0.1031046361,0.1600947678,-0.2180506438,0.0393019281,0.0442594588,0.3786991239,-0.2109910548,0.1054970622,-0.1609446257,0.0328658931,-0.1398211122,0.3307124078,0.5601598024,0.1211885512,-0.3152107,-0.0080715362,-0.1592412144,-0.0768523067,-0.2886967063,0.033032842,0.1179808974,0.3127059639,0.1622148603,0.3244030774,-0.3064067066,-0.0574292168,0.0431822613,-0.3236804307,-0.3411235809,0.0474533364,-0.1020902544,-0.2670582831,-0.0983202085,-0.027120363,-0.3235022128,-0.1106706262,0.2747683227,0.0064977249,0.0108977519,-0.2007917613,0.1924992651,-0.130554378,0.4602719545,0.2361584008,0.3637224138,-0.1539571434,-0.1167379916,-0.38258636,0.1371040344,-0.1922585964,-0.0478795879,0.0565933585,0.1103364453,0.0645984709,0.307295382,0.0687862858,0.0319577865,-0.1541932225,-0.0271717403,-0.4460570514,0.2466765344,0.1374053508,0.022925457,0.1934008896,-0.2839234769,-0.0025260057,-0.3172266483,0.3206676841,-0.1528250426,-0.0116221141,0.3197624683,-0.1758912504,0.5544427633,-0.0219294205,0.2818313241,-0.2144765109,-0.4851514101,-0.088024877,-0.3162638545,-0.003983466,0.2112577856,0.2206900716,0.2752889991,-0.0699977726,-0.329949826,-0.0114827501,0.1840766668,0.2173782289,0.0440488271,0.0415664986,0.0234162565,0.1511238962,0.0875175595,-0.1170225367,0.1542575657,-0.2034400702,-0.0452892259,-0.109629117,-0.3547621071,0.2678392529,0.0264546853,-0.2194169015,-0.2771323621,0.259618789,0.177227512,0.1006659046,-0.3608134687,0.1104133874,0.3004675508,-0.0152471839,-0.1078148335,0.3440016508,-0.0157194529,-0.0998590738,0.0057671033,0.1892267466,-0.2367936224,-0.1179514676,0.1228453368,-0.3366601169]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2813","title":"Remove compression from xopen","comments":"After discussing with @lhoestq, a reasonable alternative:\r\n- `download_manager.extract(urlpath)` adds prefixes to `urlpath` in the same way as `fsspec` does for protocols, but we implement custom prefixes for all compression formats: \r\n `bz2::http:\/\/domain.org\/filename.bz2`\r\n- `xopen` parses the `urlpath` and extracts the `compression` parameter and passes it to `fsspec.open`:\r\n `fsspec.open(\"http:\/\/domain.org\/filename.bz2\", compression=\"bz2\")`\r\n\r\nPros:\r\n- clean solution that continues giving support to all compression formats\r\n- no breaking change when opening non-decompressed files: if no compression-protocol-like is passed, fsspec.open does not uncompress (passes compression=None)\r\n\r\nCons:\r\n- we create a \"private\" convention for the format of `urlpath`: although similar to `fsspec` protocols, we add custom prefixes for the `compression` argument","body":"We implemented support for streaming with 2 requirements:\r\n- transparent use for the end user: just needs to pass the parameter `streaming=True`\r\n- no additional work for the contributors: previous loading scripts should also work in streaming mode with no (or minor) changes; and new loading scripts should not involve additional code to support streaming\r\n\r\nIn order to fulfill these requirements, streaming implementation patched some Python functions:\r\n- the `open(urlpath)` function was patched with `fsspec.open(urlpath)`\r\n- the `os.path.join(urlpath, *others)` function was patched in order to add to `urlpath` hops (`::`) and extractor protocols (`zip:\/\/`), which are required by `fsspec.open`\r\n\r\nRecently, we implemented support for streaming all archive+compression formats: zip, tar, gz, bz2, lz4, xz, zst; tar.gz, tar.bz2,...\r\nUnder the hood, the implementation:\r\n- passes an additional parameter `compression` to `fsspec.open`, so that it performs the decompression on the fly: `fsspec.open(urlpath, compression=...)`\r\n\r\nSome concerns have been raised about passing the parameter `compression` to `fsspec.open`:\r\n- https:\/\/github.com\/huggingface\/datasets\/pull\/2786#discussion_r689550254\r\n- #2811 \r\n\r\nThe main argument is that if `open` decompresses the file and afterwards we call `gzip.open` on it, that will raise an error in `oscar` dataset:\r\n```python\r\ngzip.open(open(urlpath\r\n```\r\nWhile this is true:\r\n- it is not natural\/usual to call `open` inside `gzip.open` (never seen this before)\r\n- indeed, this was recently (2 months ago) coded that way in `datasets` in order to allow streaming support (with previous implementation of streaming)\r\n\r\nIn this particular case, there is a natural fix solution: #2811:\r\n- Revert the `open` inside the `gzip.open` (change done 2 months ago): `gzip.open(open(urlpath` => `gzip.open(urlpath`\r\n- Patch `gzip.open(urlpath` with `fsspec.open(urlpath, compression=\"gzip\"` \r\n\r\nAre there other issues apart from this?\r\n\r\nNote that there is an issue just because the open inside of the gzip.open. There is no issue in the other cases where datasets loading scripts use just\r\n- `gzip.open` \r\n- `open` (after having called dl_manager.download_and_extract)\r\n\r\nTODO:\r\n- [ ] Is this really an issue? Please enumerate the `datasets` loading scripts where this is problematic.\r\n - For the moment, there are only 3 datasets where we have an `open` inside a `gzip.open`:\r\n - oscar (since 23 June), mc4 (since 2 July) and c4 (since 2 July)\r\n - In the 3 datasets, the only reason to put an open inside a gzip.open was indeed to force supporting streaming\r\n- [ ] If this is indeed an issue, which are the possible alternatives? Pros\/cons?","comment_length":105,"text":"Remove compression from xopen\nWe implemented support for streaming with 2 requirements:\r\n- transparent use for the end user: just needs to pass the parameter `streaming=True`\r\n- no additional work for the contributors: previous loading scripts should also work in streaming mode with no (or minor) changes; and new loading scripts should not involve additional code to support streaming\r\n\r\nIn order to fulfill these requirements, streaming implementation patched some Python functions:\r\n- the `open(urlpath)` function was patched with `fsspec.open(urlpath)`\r\n- the `os.path.join(urlpath, *others)` function was patched in order to add to `urlpath` hops (`::`) and extractor protocols (`zip:\/\/`), which are required by `fsspec.open`\r\n\r\nRecently, we implemented support for streaming all archive+compression formats: zip, tar, gz, bz2, lz4, xz, zst; tar.gz, tar.bz2,...\r\nUnder the hood, the implementation:\r\n- passes an additional parameter `compression` to `fsspec.open`, so that it performs the decompression on the fly: `fsspec.open(urlpath, compression=...)`\r\n\r\nSome concerns have been raised about passing the parameter `compression` to `fsspec.open`:\r\n- https:\/\/github.com\/huggingface\/datasets\/pull\/2786#discussion_r689550254\r\n- #2811 \r\n\r\nThe main argument is that if `open` decompresses the file and afterwards we call `gzip.open` on it, that will raise an error in `oscar` dataset:\r\n```python\r\ngzip.open(open(urlpath\r\n```\r\nWhile this is true:\r\n- it is not natural\/usual to call `open` inside `gzip.open` (never seen this before)\r\n- indeed, this was recently (2 months ago) coded that way in `datasets` in order to allow streaming support (with previous implementation of streaming)\r\n\r\nIn this particular case, there is a natural fix solution: #2811:\r\n- Revert the `open` inside the `gzip.open` (change done 2 months ago): `gzip.open(open(urlpath` => `gzip.open(urlpath`\r\n- Patch `gzip.open(urlpath` with `fsspec.open(urlpath, compression=\"gzip\"` \r\n\r\nAre there other issues apart from this?\r\n\r\nNote that there is an issue just because the open inside of the gzip.open. There is no issue in the other cases where datasets loading scripts use just\r\n- `gzip.open` \r\n- `open` (after having called dl_manager.download_and_extract)\r\n\r\nTODO:\r\n- [ ] Is this really an issue? Please enumerate the `datasets` loading scripts where this is problematic.\r\n - For the moment, there are only 3 datasets where we have an `open` inside a `gzip.open`:\r\n - oscar (since 23 June), mc4 (since 2 July) and c4 (since 2 July)\r\n - In the 3 datasets, the only reason to put an open inside a gzip.open was indeed to force supporting streaming\r\n- [ ] If this is indeed an issue, which are the possible alternatives? Pros\/cons?\nAfter discussing with @lhoestq, a reasonable alternative:\r\n- `download_manager.extract(urlpath)` adds prefixes to `urlpath` in the same way as `fsspec` does for protocols, but we implement custom prefixes for all compression formats: \r\n `bz2::http:\/\/domain.org\/filename.bz2`\r\n- `xopen` parses the `urlpath` and extracts the `compression` parameter and passes it to `fsspec.open`:\r\n `fsspec.open(\"http:\/\/domain.org\/filename.bz2\", compression=\"bz2\")`\r\n\r\nPros:\r\n- clean solution that continues giving support to all compression formats\r\n- no breaking change when opening non-decompressed files: if no compression-protocol-like is passed, fsspec.open does not uncompress (passes compression=None)\r\n\r\nCons:\r\n- we create a \"private\" convention for the format of `urlpath`: although similar to `fsspec` protocols, we add custom prefixes for the `compression` argument","embeddings":[-0.4191690683,-0.0069032027,0.00591273,0.0238578357,0.1248178035,-0.3256881237,-0.2292778045,0.4713353813,0.088664867,0.2409460843,-0.2471538186,0.4001294672,0.0819666311,0.1371937096,-0.0226641782,-0.1370156556,-0.0278308373,0.1331306547,0.0882370248,-0.0054111537,-0.4790536165,0.1909059584,-0.134004131,-0.0106720226,0.2445950508,0.0356198996,0.2556262016,0.21884045,-0.3985928297,-0.1497644484,-0.1308254302,0.1495487243,0.0225962121,0.0869437903,-0.0001130888,0.0266869236,0.2988415658,-0.1954084933,-0.2381751686,-0.2629078329,-0.1178576946,-0.0889441147,0.0825306773,-0.0007839249,0.1002834663,-0.7044277787,0.0207543392,-0.8813657761,0.5043359399,0.2567861676,0.1220125854,0.1561299562,0.0596263148,0.0020174929,0.275047034,-0.0995139927,-0.1907603443,-0.1044160053,0.2203274369,0.332131207,-0.1137185469,0.423512727,-0.1898332685,-0.1889518499,0.3292146027,-0.2509939373,0.0750826746,-0.2273344249,-0.1807321608,0.1074701399,0.1452671587,-0.1285867393,-0.5220016241,-0.1854573786,0.068858102,-0.5120434165,0.1385540515,0.3591740727,0.04782594,0.1616456658,-0.0850937963,-0.2431392819,0.0070688804,0.0468943752,0.197272405,0.2863021791,-0.0803830177,0.0017110787,0.3852153718,0.1243219376,0.1140473187,-0.2616316974,-0.0275921207,-0.0973567069,-0.2123515308,0.0272930954,0.1381485462,0.0571020357,0.0816577971,0.4363040924,0.2570689917,0.2995754182,-0.1212359965,0.0256679691,0.0512153357,0.1565018743,-0.0218391381,-0.1820706427,0.0376392566,0.1823161095,0.2186764479,-0.2348976135,0.2453594506,-0.1651804149,-0.3647239506,0.322506845,0.6144118309,0.007971433,-0.1655450612,-0.0997920781,-0.1505320221,-0.210909456,0.0533471107,0.0897925794,-0.0646989644,0.335244,0.0172747541,0.3113516271,-0.2142479122,-0.0174160339,-0.1873303652,-0.0547256842,-0.1313954145,0.2216723412,0.0550615899,-0.7223923802,0.0392942131,0.2271921635,0.5497029424,-0.0975679234,0.1636150926,-0.0467937253,0.2022970468,0.244686082,-0.0780420974,-0.0580973811,0.2379582673,0.0552627258,-0.2282734215,0.1757036,-0.5017311573,-0.2378914654,-0.106171757,0.158489123,-0.02071709,0.1946989447,0.0252138488,0.0411648452,-0.4355071485,-0.4308710992,0.0343736745,-0.0243316777,-0.2750070095,-0.1244571283,0.3272199333,0.0292744879,0.0415952876,-0.2197496891,-0.1632013023,0.1786553413,0.5669157505,0.2861439884,-0.3046621382,-0.3236417174,-0.2175180316,0.0354337394,0.5478261709,-0.228663072,-0.4149609506,0.3734907508,0.0073397402,0.3961355388,0.2859599888,-0.3961561024,0.4834241569,-0.1646109968,0.1445796639,0.220534265,0.1359208971,0.1277269721,-0.2463243306,-0.0944504738,0.0922186598,-0.0883124396,-0.1049429476,0.1127613038,-0.0099599687,-0.0584183298,0.1922488362,-0.0892307684,0.0422890447,-0.2404192537,0.1896105409,-0.0265081748,-0.0270191841,0.0127173373,0.1164058223,-0.0961750224,-0.159837842,-0.1996869147,-0.1320573688,-0.2483507842,-0.0809583366,0.0238576662,0.0341969728,-0.2237561792,0.0648615435,0.1565391421,-0.2515935302,0.0391189046,-0.2730352283,0.2372072935,0.0206574015,0.022600593,-0.2585276067,-0.0280286372,0.1524491757,-0.0354438387,0.0742466897,-0.0298156347,0.2051520944,-0.1311266571,-0.3151764572,0.4424102306,-0.0684272274,0.3711241782,-0.2005501091,0.0440545827,0.290427357,-0.3242277205,0.1412657499,0.5436943173,0.1446817219,0.1829971224,-0.1114250496,0.3421912491,0.0299910307,-0.0450014248,0.1369954199,-0.0521548279,-0.0120326634,0.0070299641,-0.5576578379,0.0399456583,-0.1061250046,0.1598307788,0.3208061159,0.1401841789,-0.1437506378,-0.082701534,0.5054670572,-0.1650294513,-0.1705068797,0.4521291554,0.024848735,-0.2451156825,0.1435942054,0.1345321834,0.6052958369,0.2483297735,0.169898048,0.1075602025,0.1086897478,-0.2464023232,0.5393251777,0.1573459655,0.1208355129,0.2046387345,0.0883006975,0.3496350348,-0.4763229489,-0.1577930897,-0.0188913811,-0.3481453061,-0.2709912658,0.1504784226,-0.5044305325,-0.1620633155,0.1140392795,0.1126582697,0.0804836974,-0.1830219477,0.1574531943,0.2160220593,-0.2817361057,0.3589549661,-0.2321249843,0.3441582918,-0.0611000583,-0.2766939104,-0.1615319401,0.1091090292,0.1053586677,0.0525800064,-0.0900605768,-0.4445866942,0.1202621311,0.222739622,0.0490903743,-0.2063014507,-0.197283864,0.0614701025,-0.0634209588,-0.1260740757,-0.0545571148,-0.1704532355,-0.0619699284,-0.4755946696,0.0886214152,-0.2605989277,0.0925254077,-0.2269986868,-0.0709561333,0.2812229097,-0.4269396663,-0.1482831389,0.0539467521,-0.5750749707,0.3633087873,-0.3330125809,0.1032284722,0.355273813,-0.1464030147,-0.264279753,0.1123070046,0.1083967686,-0.3649926782,-0.6582285166,0.3468936682,-0.1203555539,-0.3256308436,-0.1368480623,0.0312408563,-0.0274973791,0.5976182222,-0.0717615634,0.0724809617,-0.1209234446,-0.1464298666,-0.1731392592,-0.0612983964,0.3393907547,0.1769471616,-0.22264117,-0.3091018498,-0.1057067811,-0.1121419072,0.1643157601,0.2035526633,0.0604088716,0.2159347534,-0.0364295542,0.2296808362,0.1189593077,0.3798584342,0.6151396036,0.2907935679,0.4680997431,-0.1208091229,-0.0692044199,0.2056080848,-0.1757374555,-0.0238053165,0.1030081213,0.1163585857,-0.1173081622,-0.1242893711,0.1389081478,-0.3419250846,-0.3283257484,0.0274633393,-0.0866463184,0.1399342716,0.0537560433,0.2511183321,-0.1893440783,0.1527810842,0.0916983187,0.2383137196,0.6755878925,-0.15241234,0.107014589,0.3319439292,-0.4048944116,0.3504059017,-0.1224255115,0.0183258634,-0.0397147201,-0.2039787769,0.0501860343,0.1413901448,0.3463095129,-0.0763344616,-0.2647818327,0.2832125425,-0.040618889,-0.5618639588,0.0392826907,-0.0522551313,-0.1368769556,-0.0117746349,0.3609420061,-0.2811939418,-0.3290457428,-0.0402591228,-0.2512136698,0.1511800885,-0.0644916147,-0.2048472017,-0.3617077768,-0.157032758,-0.2812048197,0.1525585651,0.2550778091,-0.2527911365,-0.0108838575,-0.1123664826,-0.0053259502,0.3447751701,-0.29168576,0.4291557074,0.2251055986,-0.1502024382,-0.0146231242,0.4382825494,-0.1271054149,0.398761034,0.1898753047,-0.6032568216,0.1646266431,0.310446918,0.1171529293,0.3649796247,-0.325535953,-0.04942118,-0.228967756,0.3229519427,-0.3455232084,0.3161297739,0.3063803017,-0.2636548877,-0.0443640761,-0.4398933053,0.2175514996,0.1642788202,-0.1890220344,0.8254839778,0.0352724269,-0.1773246378,0.122579813,0.1263153255,1.0271830559,0.0779224932,-0.0172915496,0.525633812,-0.1789595783,0.2232170254,-0.252496779,0.1508210152,0.4147315919,-0.6663112044,0.0116243744,0.0635504648,0.2472855747,-0.2237026244,-0.31313923,0.1947860569,0.1261041015,-0.3749637008,0.2340984941,0.1640256047,-0.4982544482,-0.5518414378,-0.6275808215,0.1332900077,-0.1251038462,0.1028622761,0.1514483541,-0.4420361817,0.349096626,-0.2442368269,0.0693465471,0.3628324866,-0.2472995967,-0.0080772815,-0.3074160218,-0.2275934517,0.1119299084,0.1407986879,-0.195042789,-0.1791714579,-0.003805822,0.218735218,-0.2408133894,0.2785665095,-0.0977738127,-0.0204975288,0.0606458113,-0.0285229497,-0.0298124757,0.1814388633,-0.1000713333,-0.2371849865,0.2475835532,-0.0636254102,0.1865832061,-0.0623586252,-0.0895967335,-0.2573811114,-0.1082181931,-0.0805354714,0.0791487619,-0.1657522321,0.0583991781,0.0762596279,-0.1058047861,-0.4336339533,-0.137647599,0.24259381,-0.0383617878,0.0819688141,0.241481021,-0.4261102676,-0.2348876745,-0.010146019,-0.0361519679,-0.3330216408,-0.1468376219,0.045588959,-0.0184945576,-0.1030839384,-0.1638296247,0.1580751389,-0.0849083215,0.3173320591,0.0215174872,-0.2094064355,-0.0381982699,0.1747953296,-0.320145458,0.061814934,0.2344076335,0.1284126937,-0.0325051434,-0.0868131444,-0.2891893089,0.1344661415,0.4174242914,-0.0773126855,-0.1533911824,-0.0050636814,0.4084664583,-0.1125337109,0.0429866761,0.2514637709,-0.2160812616,-0.1384045929,-0.2821722031,0.2124687433,-0.009444044,-0.3162380159,-0.1214062348,-0.0168030038,-0.0527550057,-0.315230757,0.3412930667,0.0054235808,0.0268558692,-0.0939990953,0.2126724422,-0.0124069257,0.0356110968,0.0362169482,0.0572795644,0.6462957263,-0.0696859583,-0.0376833379,-0.1834296882,-0.1723701358,-0.3989319205,-0.1285023838,0.3483402729,0.3660281003,0.5126979947,0.1758350283,-0.1910285503,-0.124236241,0.3427618742,0.1125371084,-0.0456462763,-0.1857048869,-0.0685047284,0.152340129,-0.1664695442,0.0045213206,-0.02806684,0.4545364082,0.0036067872,-0.0321892984,-0.0679436252,0.0796379074,0.0212428775,-0.0878722146,0.1140840501,0.3200615048,0.5746148825,0.2170225382,0.1167597994,0.2304069847,-0.0640204921,-0.026594732,0.0575587302,-0.1954071373,-0.116558753,0.4438080788,0.2932678759,0.2682131827,-0.1862258613,-0.4940246344,-0.0866636112,0.1054815501,-0.1801442206,0.4586992562,0.259776175,0.13445732,-0.205948621,-0.1786029041,0.0593893863,0.4335688353,0.2376706004,-0.2742589116,0.0818656683,0.087352477,0.0576240793,0.1591817886,0.0930175558,0.3212279975,0.2338700294,0.120174706,0.0564838536,0.1418886632,-0.3399507999,0.034546461,0.523691833,-0.1745196134,0.1115561128,0.307135433,0.0080914106,0.2147071362,0.4515011609,0.2882674336,-0.0392870829,-0.0057261591,0.1468675733,0.1493417919,0.0900042504,0.025803579,0.0969763994,0.5032496452,0.1215637773,0.2795298696,0.1764881611,-0.0703050941,0.2152403146,-0.1154240668,0.3489600122,-0.2589488029,0.2532716095,0.1675393581,0.1809652448,0.0671691969,-0.2221703976,-0.2454078496,0.1245906502,0.1826526821,0.1845455617,0.0149036711,-0.1429315209,0.0868986025,0.249523133,0.1764581501,-0.1061422825,-0.1412615478,-0.2266934514,-0.3937927485,-0.3562112153,0.2376065552,0.1957727224,-0.0760525614,0.202798143,-0.0151425721,0.3195686936,0.2645491362,-0.2284376472,-0.0823647305,0.2203802615,-0.0818236172,-0.1618982702,0.1388604641,0.3744253814,0.1084784195,-0.0013724865,-0.1735455096,0.2423675805,-0.1927788556,0.081259042,0.0070030959,0.0112384735,-0.3363492489,-0.1374450028,0.3763563633,0.2580069304,-0.0482212938,-0.4752644897,-0.1529146284,0.0744278505,0.0543088578,-0.0074467245,-0.1139924452,-0.0779763982,0.2417218834,-0.0826463923,-0.0084010959,-0.4220407009,0.6201142669,-0.1255545765,-0.4073031545,-0.4189613461,0.0136868926,-0.0503152721,-0.1109454632,-0.1100521535,0.3306161463,-0.1201392338,0.2010824978,-0.1084934026,-0.4842882156,0.7212756872,-0.3065904379,-0.0180206485,0.0600795522,0.3401736319,0.0934052989,0.0044341399,-0.3561251462,0.1665582955,0.1113158315,-0.0360377356,-0.0797713101,0.1227187291,-0.1704528183,-0.1375095099,-0.0629704893,0.234298408,0.1840940863,-0.2812791765,0.0636586398,-0.1641805917]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2799","title":"Loading JSON throws ArrowNotImplementedError","comments":"Hi @lewtun, thanks for reporting.\r\n\r\nApparently, `pyarrow.json` tries to cast timestamp-like fields in your JSON file to pyarrow timestamp type, and it fails with `ArrowNotImplementedError`.\r\n\r\nI will investigate if there is a way to tell pyarrow not to try that timestamp casting.","body":"## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":42,"text":"Loading JSON throws ArrowNotImplementedError\n## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nHi @lewtun, thanks for reporting.\r\n\r\nApparently, `pyarrow.json` tries to cast timestamp-like fields in your JSON file to pyarrow timestamp type, and it fails with `ArrowNotImplementedError`.\r\n\r\nI will investigate if there is a way to tell pyarrow not to try that timestamp casting.","embeddings":[-0.0681871027,0.1841620058,0.0364155807,0.3780699968,0.2355371863,-0.0062702368,0.4846684635,0.3670629561,0.4180777371,-0.0028386964,0.0605546869,0.5898874402,0.0276210085,-0.148857981,-0.2172110528,-0.1838730276,0.0313557796,0.1905159801,0.1274458319,0.0285853427,-0.3122327328,0.2942055166,-0.214816317,0.1017739549,-0.0391265377,-0.0952268541,0.1041623503,0.3933393061,-0.126418829,-0.3826465905,0.4134063721,-0.0130348513,0.2161162943,0.5308527946,-0.0001203599,0.1964653581,0.4152528644,-0.092444241,-0.2868113518,-0.4780595005,-0.2790301442,-0.3253917992,0.537475884,-0.1200379506,0.1052053347,-0.2437469363,-0.0497698858,-0.2207819521,0.6015406847,0.2307226658,0.1048027351,0.429939568,0.0260289963,0.0192190502,0.0637280419,0.2108367532,-0.1217756793,0.4542036653,0.008248603,0.1270525903,-0.1715968549,-0.0036361166,-0.102641046,0.1538670212,0.2191231996,0.0476008654,-0.1704021692,0.0818345174,0.0492499694,0.2946319878,0.2845771015,-0.0508809648,-0.4122377634,-0.3489953876,-0.1207850203,0.1302726865,0.2388218939,0.2473598123,-0.1509169936,0.1259337515,0.0328827612,-0.0184712112,-0.274197191,0.2383805364,0.0036435723,0.0111869918,-0.1582194567,0.10729184,-0.1295856088,-0.020824587,-0.1499335319,-0.2244829088,-0.3057584167,0.1283780038,0.0374987274,0.11600402,-0.0355978459,-0.4100777805,0.2355834544,0.3558476567,0.3206709325,0.2080192715,-0.2673651874,0.2078464776,0.5903261304,0.2016133815,-0.0549013875,0.0114986571,0.2564998567,0.330514431,0.0607299209,-0.0435533673,-0.2637991011,-0.1108330786,0.0103772897,-0.197450906,0.3086808324,-0.1726259738,-0.2238530517,0.0755632967,-0.6134242415,-0.0573209412,-0.0002183132,0.3672920167,-0.1971875429,0.4987778664,0.2744282484,0.4868384302,0.0442172214,0.0317274667,-0.1022381261,-0.0353484266,-0.1634835154,-0.2962282002,0.2371660173,0.0525447652,0.0861053467,0.1265179962,-0.0630014017,-0.4126393497,-0.0742599964,0.0477150232,-0.0974051356,0.16696693,0.1920519024,0.132871151,0.1559260786,-0.4045022726,-0.1376460344,0.0323025733,-0.376919657,0.0055578146,-0.3176852167,0.0546303503,-0.2702908218,-0.0120817386,-0.7623512149,-0.0521201082,-0.1122782081,0.0660267845,0.0638756305,-0.0244168527,0.2227595448,-0.2470848262,0.0749816447,0.28331393,-0.3005584776,-0.2378236353,-0.0629288927,-0.037491519,0.1896335036,0.2650878727,-0.2005772442,0.2845451832,-0.1637352109,0.2106698602,0.3121518195,-0.3327314556,-0.2219264358,0.605520308,0.0193049405,0.5248504877,0.0198842697,-0.2308956832,-0.0092007685,0.1884041429,0.3035949767,0.4013707638,0.1422874182,0.0679152384,-0.2457050234,-0.2022635043,-0.1346425563,0.3127692342,0.0487670116,0.1889399141,0.0995356888,0.0244781338,0.2506872714,-0.1951391548,-0.0266671889,0.1624232382,0.0911331177,0.0282151625,-0.0057259151,-0.0044113393,-0.6142941117,0.1499558836,-0.0511226989,0.109201923,-0.408085674,0.0093207462,-0.2547832429,0.1260678917,-0.1280752718,0.1847681105,0.0374528877,0.1086534187,-0.1821872592,0.2108015567,-0.1838937998,0.3152327836,0.1626450419,0.2816524506,-0.4834640026,0.4377061725,0.210255906,-0.1212381497,0.0271184649,0.1853716373,-0.097817637,-0.1626708806,-0.2391357571,0.1727546602,0.1394181401,0.1797138602,-0.0562768616,-0.2375711501,-0.0822836161,0.0838592127,-0.1951901019,0.3068773448,0.2217241377,-0.0245843939,-0.0280968156,0.8052088618,-0.0741070136,0.4431970716,-0.015723357,-0.0205509756,0.4561716616,0.0615659803,-0.364153266,-0.3053402305,0.2388348728,0.3186245859,0.0603672415,0.0040045432,-0.3794290721,-0.1885213256,0.3982873261,0.005179801,0.2424625754,0.1532346159,-0.0953567475,-0.0996729881,0.0158600174,0.1401785463,0.2382987291,0.0758770034,-0.2070478052,-0.0434467606,-0.0409476832,-0.0247451495,0.1555987,0.306535244,0.2641779482,0.0933471769,0.1558886766,-0.020962242,-0.2815460265,-0.238003552,-0.2863601744,0.1189386174,-0.6182570457,0.2991375625,-0.4743469954,0.1188910827,0.035333585,-0.3410826325,-0.126299873,-0.322971493,-0.3007420599,0.2734274864,-0.1147493422,-0.1341666132,-0.2667239308,0.3151707649,-0.0232129041,-0.3463473022,-0.2113419473,-0.0117184091,-0.184648335,-0.0234055538,0.3875870109,0.1092963666,0.1544730067,-0.1289402992,0.047313679,0.1418192536,-0.0863964856,0.1482000202,-0.1994989067,0.2889936268,0.2154709399,0.1969099492,-0.0978527516,-0.1516460776,0.2886652946,0.0957521275,-0.3427708745,0.3241927326,-0.1354984939,0.0278997961,-0.1492165327,-0.0291694198,0.1219549254,-0.4392543733,0.5247157216,0.2458687872,0.0581384897,-0.0170072187,0.0242614429,0.1799215674,0.1687064171,0.0159113072,-0.1107467487,-0.1546859592,0.2326568216,-0.160832569,-0.2501766384,0.0764547884,-0.1965518147,0.1685908139,-0.0865133032,-0.390976578,0.0259424336,0.0707196668,0.1136565655,-0.0444040187,-0.295239538,0.0588247478,-0.0127509227,0.0851902887,-0.1855715215,-0.2022252828,0.1206316799,-0.0438129231,0.0216035321,0.0565150753,0.7252899408,-0.0942959189,0.1807754487,0.3117997646,-0.0382717438,0.394974649,-0.0264437217,0.1030089855,-0.4057941437,-0.1191259399,-0.3808179796,-0.0971131027,0.0786372274,0.2193027288,0.1096423194,-0.2879481912,-0.2405597121,-0.0622841977,-0.4587052464,-0.3489450216,0.1333795786,-0.3543255031,-0.3382495046,0.0320333503,0.2205641121,-0.0587302446,-0.0581021123,-0.0303620175,0.4014143944,0.2062794119,-0.0455005579,-0.2057112902,0.0343681015,-0.3544535339,0.3606956899,0.0746525601,0.3657507002,0.1974582672,-0.028097624,0.2447486371,-0.2093995959,0.4490728974,-0.120011054,0.0227281768,0.0280953031,-0.1236653179,-0.6418688893,0.1011512503,-0.1271315068,0.2907666266,0.2359819263,0.2095480561,-0.4572590888,-0.2961414456,0.1018688306,-0.1504392922,-0.1058704555,-0.1064765677,-0.3640385568,-0.2923769653,-0.2448295504,-0.1525726318,0.0274700802,0.2387111634,0.0874371156,-0.0987067744,-0.2664907575,-0.0690853894,0.1225850284,0.2164159268,0.1880879253,-0.2173919529,0.0409795344,0.0061119525,0.3643347025,0.3959253132,0.8496598601,0.2561551034,-0.5346414447,0.2092527151,-0.0245482698,0.0420461036,0.1865970343,0.0230351947,0.0835103542,0.2850650549,0.1448794007,-0.2175282389,-0.1765887737,0.4804262519,-0.1738846451,-0.4621180296,-0.346393168,0.4906897545,-0.1168812439,0.0213299748,0.2263398916,0.2678977251,-0.1467050314,0.758928895,0.0028833551,0.7930403352,0.1487369686,0.1980527639,0.7221454978,-0.5317468047,0.400482446,-0.0322305784,-0.0708860084,-0.2383532822,0.5987789035,-0.1820171326,-0.151582703,-0.0095944684,0.192508921,-0.1030690968,0.0225751493,-0.3046782017,-0.0162525531,0.0787486061,0.128597334,-0.1141062975,-0.2810204625,-0.7139284611,-0.0139217712,-0.1206339747,0.0340203084,-0.0035200173,-0.3064055443,-0.2558885217,-0.188852787,-0.3464414477,0.2579855025,0.1630945057,-0.0184752066,0.2230367512,-0.0496139601,0.0946815684,-0.0524797849,0.2925044,-0.0933387354,-0.1754464954,0.187010482,-0.1364040524,-0.0991375446,-0.0924463496,-0.016065022,0.5005759001,-0.0768829286,-0.1589543819,-0.2050989121,-0.2627161145,-0.5380346775,0.2646922767,0.4163651466,-0.3466082215,-0.240976274,-0.393114835,-0.0475282073,0.2723071873,-0.1570670009,0.0498477258,0.2624671161,0.1839359552,-0.0674586892,0.2908512056,-0.3609032929,-0.1087285951,0.3222659528,-0.0353102535,0.0564952418,0.5033448339,0.2862301469,-0.2179954052,-0.1904615164,0.2036407739,0.8105605841,-0.6269248128,0.3065258563,-0.1730780452,-0.0535799339,-0.0472417548,0.3028697371,0.0699545369,-0.0232218802,0.0566213802,-0.6213235855,-0.1062008291,0.4122470915,-0.1543131769,-0.0214453992,0.1351440996,-0.0169346631,-0.2405516505,-0.1720611304,-0.17293562,0.2192620784,-0.1576303691,-0.2045778781,-0.3086437285,0.0632046908,0.5788154602,0.0068058115,0.0865861177,0.0817479491,-0.1436324716,-0.0885987878,-0.3009876311,0.2072960436,0.1043563411,-0.0927105248,-0.0784604624,0.0101217115,-0.1790237874,-0.135748744,0.2507112324,0.2667553127,-0.0477175564,-0.1331558526,0.0813497081,0.1524973363,-0.2447343022,0.1382133961,-0.159083724,0.5401737094,0.0858471692,-0.0476722531,-0.2291520685,0.1443036646,-0.2354615629,0.2881273329,0.2030762136,0.2164718509,0.4846534729,-0.427249223,0.0382829905,-0.2070930153,0.0976076871,0.3163167238,-0.2574578226,0.0504572801,-0.16220209,0.0547889322,0.0131520722,0.0525103472,0.1396277249,-0.0389830656,0.0427073203,0.333689034,-0.1388189495,0.042830985,-0.2334407568,0.0458435602,0.4255111814,0.1589772552,0.051013343,0.2905644774,-0.2684130967,0.0809304491,0.0158537552,0.1407580674,0.2980785966,-0.0071150423,0.0680872425,0.0217248164,0.1491240412,0.2423139662,0.0005278526,-0.4159061313,-0.0420816727,0.3158295155,-0.0500427932,0.3092298806,0.1688649356,0.3333734274,-0.1892764717,0.0969858468,-0.4036626816,-0.1670233607,-0.0538140684,-0.2211986184,-0.5511418581,-0.2069526762,-0.3715268075,-0.1109172553,-0.1350990832,-0.0615067855,-0.2741713226,0.1886223704,0.0503452383,-0.330363065,0.0814077035,-0.1955392063,0.2041691542,-0.1523468047,0.2263703197,-0.0093282107,0.0362457894,0.3188800812,0.0760150179,0.5093174577,0.2283807099,-0.0963346064,-0.2487950623,-0.0241237804,-0.142178461,0.040377859,0.3468794227,0.1857326627,0.0395644121,0.2420821488,0.1516417563,-0.1133231595,0.3565611541,0.002253989,0.263456285,-0.0992489904,0.5178786516,-0.467658788,0.1000935584,-0.1672317088,-0.3525443077,-0.1460466087,-0.3188263476,-0.1780048162,-0.1945900023,-0.0235637128,-0.058697395,0.0232757386,-0.0869168863,0.4179295003,0.394969821,0.1819574684,0.1531835794,-0.033079084,-0.2972991467,0.2351235002,-0.1318951845,0.0307580009,0.2157548368,0.2165423781,0.0767683834,0.1726402193,0.0412504598,0.0387746431,0.0111108012,0.2165732384,-0.2433100045,-0.060975045,0.2753140926,0.0173532814,-0.1357738823,-0.2809374332,-0.0888741985,-0.0585748106,-0.0598330982,0.1518500149,-0.0595391653,-0.0654157251,-0.1698684841,0.4425504208,0.1324360222,0.2011087984,-0.0575092882,-0.4760095477,-0.2689436674,-0.4144521952,-0.1308504641,0.3707846403,-0.1804432422,0.6111345291,0.0342233665,0.1484044045,-0.0185195096,-0.0448632054,-0.1875070333,-0.0982181877,-0.5277866125,0.2939006686,-0.3957848847,0.1447350532,-0.1315573007,0.1057318822,0.1239854619,0.1936321557,-0.0772358403,-0.2317139655,0.4982314706,-0.2602384984,-0.3133595288,0.2727927864,0.0889144912,-0.1325899363,-0.2973820269,-0.6104027033,0.0109965969,0.2189054042,0.0305512454,-0.2178259343,-0.1124497503,-0.124261193,-0.05467695,-0.1363347769,0.2223986536,0.0136693334,-0.1138071492,0.1441099644,-0.1869696826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2799","title":"Loading JSON throws ArrowNotImplementedError","comments":"I think the issue is more complex than that...\r\n\r\nI just took one of your JSON lines and pyarrow.json read it without problem.","body":"## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":23,"text":"Loading JSON throws ArrowNotImplementedError\n## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nI think the issue is more complex than that...\r\n\r\nI just took one of your JSON lines and pyarrow.json read it without problem.","embeddings":[-0.0681871027,0.1841620058,0.0364155807,0.3780699968,0.2355371863,-0.0062702368,0.4846684635,0.3670629561,0.4180777371,-0.0028386964,0.0605546869,0.5898874402,0.0276210085,-0.148857981,-0.2172110528,-0.1838730276,0.0313557796,0.1905159801,0.1274458319,0.0285853427,-0.3122327328,0.2942055166,-0.214816317,0.1017739549,-0.0391265377,-0.0952268541,0.1041623503,0.3933393061,-0.126418829,-0.3826465905,0.4134063721,-0.0130348513,0.2161162943,0.5308527946,-0.0001203599,0.1964653581,0.4152528644,-0.092444241,-0.2868113518,-0.4780595005,-0.2790301442,-0.3253917992,0.537475884,-0.1200379506,0.1052053347,-0.2437469363,-0.0497698858,-0.2207819521,0.6015406847,0.2307226658,0.1048027351,0.429939568,0.0260289963,0.0192190502,0.0637280419,0.2108367532,-0.1217756793,0.4542036653,0.008248603,0.1270525903,-0.1715968549,-0.0036361166,-0.102641046,0.1538670212,0.2191231996,0.0476008654,-0.1704021692,0.0818345174,0.0492499694,0.2946319878,0.2845771015,-0.0508809648,-0.4122377634,-0.3489953876,-0.1207850203,0.1302726865,0.2388218939,0.2473598123,-0.1509169936,0.1259337515,0.0328827612,-0.0184712112,-0.274197191,0.2383805364,0.0036435723,0.0111869918,-0.1582194567,0.10729184,-0.1295856088,-0.020824587,-0.1499335319,-0.2244829088,-0.3057584167,0.1283780038,0.0374987274,0.11600402,-0.0355978459,-0.4100777805,0.2355834544,0.3558476567,0.3206709325,0.2080192715,-0.2673651874,0.2078464776,0.5903261304,0.2016133815,-0.0549013875,0.0114986571,0.2564998567,0.330514431,0.0607299209,-0.0435533673,-0.2637991011,-0.1108330786,0.0103772897,-0.197450906,0.3086808324,-0.1726259738,-0.2238530517,0.0755632967,-0.6134242415,-0.0573209412,-0.0002183132,0.3672920167,-0.1971875429,0.4987778664,0.2744282484,0.4868384302,0.0442172214,0.0317274667,-0.1022381261,-0.0353484266,-0.1634835154,-0.2962282002,0.2371660173,0.0525447652,0.0861053467,0.1265179962,-0.0630014017,-0.4126393497,-0.0742599964,0.0477150232,-0.0974051356,0.16696693,0.1920519024,0.132871151,0.1559260786,-0.4045022726,-0.1376460344,0.0323025733,-0.376919657,0.0055578146,-0.3176852167,0.0546303503,-0.2702908218,-0.0120817386,-0.7623512149,-0.0521201082,-0.1122782081,0.0660267845,0.0638756305,-0.0244168527,0.2227595448,-0.2470848262,0.0749816447,0.28331393,-0.3005584776,-0.2378236353,-0.0629288927,-0.037491519,0.1896335036,0.2650878727,-0.2005772442,0.2845451832,-0.1637352109,0.2106698602,0.3121518195,-0.3327314556,-0.2219264358,0.605520308,0.0193049405,0.5248504877,0.0198842697,-0.2308956832,-0.0092007685,0.1884041429,0.3035949767,0.4013707638,0.1422874182,0.0679152384,-0.2457050234,-0.2022635043,-0.1346425563,0.3127692342,0.0487670116,0.1889399141,0.0995356888,0.0244781338,0.2506872714,-0.1951391548,-0.0266671889,0.1624232382,0.0911331177,0.0282151625,-0.0057259151,-0.0044113393,-0.6142941117,0.1499558836,-0.0511226989,0.109201923,-0.408085674,0.0093207462,-0.2547832429,0.1260678917,-0.1280752718,0.1847681105,0.0374528877,0.1086534187,-0.1821872592,0.2108015567,-0.1838937998,0.3152327836,0.1626450419,0.2816524506,-0.4834640026,0.4377061725,0.210255906,-0.1212381497,0.0271184649,0.1853716373,-0.097817637,-0.1626708806,-0.2391357571,0.1727546602,0.1394181401,0.1797138602,-0.0562768616,-0.2375711501,-0.0822836161,0.0838592127,-0.1951901019,0.3068773448,0.2217241377,-0.0245843939,-0.0280968156,0.8052088618,-0.0741070136,0.4431970716,-0.015723357,-0.0205509756,0.4561716616,0.0615659803,-0.364153266,-0.3053402305,0.2388348728,0.3186245859,0.0603672415,0.0040045432,-0.3794290721,-0.1885213256,0.3982873261,0.005179801,0.2424625754,0.1532346159,-0.0953567475,-0.0996729881,0.0158600174,0.1401785463,0.2382987291,0.0758770034,-0.2070478052,-0.0434467606,-0.0409476832,-0.0247451495,0.1555987,0.306535244,0.2641779482,0.0933471769,0.1558886766,-0.020962242,-0.2815460265,-0.238003552,-0.2863601744,0.1189386174,-0.6182570457,0.2991375625,-0.4743469954,0.1188910827,0.035333585,-0.3410826325,-0.126299873,-0.322971493,-0.3007420599,0.2734274864,-0.1147493422,-0.1341666132,-0.2667239308,0.3151707649,-0.0232129041,-0.3463473022,-0.2113419473,-0.0117184091,-0.184648335,-0.0234055538,0.3875870109,0.1092963666,0.1544730067,-0.1289402992,0.047313679,0.1418192536,-0.0863964856,0.1482000202,-0.1994989067,0.2889936268,0.2154709399,0.1969099492,-0.0978527516,-0.1516460776,0.2886652946,0.0957521275,-0.3427708745,0.3241927326,-0.1354984939,0.0278997961,-0.1492165327,-0.0291694198,0.1219549254,-0.4392543733,0.5247157216,0.2458687872,0.0581384897,-0.0170072187,0.0242614429,0.1799215674,0.1687064171,0.0159113072,-0.1107467487,-0.1546859592,0.2326568216,-0.160832569,-0.2501766384,0.0764547884,-0.1965518147,0.1685908139,-0.0865133032,-0.390976578,0.0259424336,0.0707196668,0.1136565655,-0.0444040187,-0.295239538,0.0588247478,-0.0127509227,0.0851902887,-0.1855715215,-0.2022252828,0.1206316799,-0.0438129231,0.0216035321,0.0565150753,0.7252899408,-0.0942959189,0.1807754487,0.3117997646,-0.0382717438,0.394974649,-0.0264437217,0.1030089855,-0.4057941437,-0.1191259399,-0.3808179796,-0.0971131027,0.0786372274,0.2193027288,0.1096423194,-0.2879481912,-0.2405597121,-0.0622841977,-0.4587052464,-0.3489450216,0.1333795786,-0.3543255031,-0.3382495046,0.0320333503,0.2205641121,-0.0587302446,-0.0581021123,-0.0303620175,0.4014143944,0.2062794119,-0.0455005579,-0.2057112902,0.0343681015,-0.3544535339,0.3606956899,0.0746525601,0.3657507002,0.1974582672,-0.028097624,0.2447486371,-0.2093995959,0.4490728974,-0.120011054,0.0227281768,0.0280953031,-0.1236653179,-0.6418688893,0.1011512503,-0.1271315068,0.2907666266,0.2359819263,0.2095480561,-0.4572590888,-0.2961414456,0.1018688306,-0.1504392922,-0.1058704555,-0.1064765677,-0.3640385568,-0.2923769653,-0.2448295504,-0.1525726318,0.0274700802,0.2387111634,0.0874371156,-0.0987067744,-0.2664907575,-0.0690853894,0.1225850284,0.2164159268,0.1880879253,-0.2173919529,0.0409795344,0.0061119525,0.3643347025,0.3959253132,0.8496598601,0.2561551034,-0.5346414447,0.2092527151,-0.0245482698,0.0420461036,0.1865970343,0.0230351947,0.0835103542,0.2850650549,0.1448794007,-0.2175282389,-0.1765887737,0.4804262519,-0.1738846451,-0.4621180296,-0.346393168,0.4906897545,-0.1168812439,0.0213299748,0.2263398916,0.2678977251,-0.1467050314,0.758928895,0.0028833551,0.7930403352,0.1487369686,0.1980527639,0.7221454978,-0.5317468047,0.400482446,-0.0322305784,-0.0708860084,-0.2383532822,0.5987789035,-0.1820171326,-0.151582703,-0.0095944684,0.192508921,-0.1030690968,0.0225751493,-0.3046782017,-0.0162525531,0.0787486061,0.128597334,-0.1141062975,-0.2810204625,-0.7139284611,-0.0139217712,-0.1206339747,0.0340203084,-0.0035200173,-0.3064055443,-0.2558885217,-0.188852787,-0.3464414477,0.2579855025,0.1630945057,-0.0184752066,0.2230367512,-0.0496139601,0.0946815684,-0.0524797849,0.2925044,-0.0933387354,-0.1754464954,0.187010482,-0.1364040524,-0.0991375446,-0.0924463496,-0.016065022,0.5005759001,-0.0768829286,-0.1589543819,-0.2050989121,-0.2627161145,-0.5380346775,0.2646922767,0.4163651466,-0.3466082215,-0.240976274,-0.393114835,-0.0475282073,0.2723071873,-0.1570670009,0.0498477258,0.2624671161,0.1839359552,-0.0674586892,0.2908512056,-0.3609032929,-0.1087285951,0.3222659528,-0.0353102535,0.0564952418,0.5033448339,0.2862301469,-0.2179954052,-0.1904615164,0.2036407739,0.8105605841,-0.6269248128,0.3065258563,-0.1730780452,-0.0535799339,-0.0472417548,0.3028697371,0.0699545369,-0.0232218802,0.0566213802,-0.6213235855,-0.1062008291,0.4122470915,-0.1543131769,-0.0214453992,0.1351440996,-0.0169346631,-0.2405516505,-0.1720611304,-0.17293562,0.2192620784,-0.1576303691,-0.2045778781,-0.3086437285,0.0632046908,0.5788154602,0.0068058115,0.0865861177,0.0817479491,-0.1436324716,-0.0885987878,-0.3009876311,0.2072960436,0.1043563411,-0.0927105248,-0.0784604624,0.0101217115,-0.1790237874,-0.135748744,0.2507112324,0.2667553127,-0.0477175564,-0.1331558526,0.0813497081,0.1524973363,-0.2447343022,0.1382133961,-0.159083724,0.5401737094,0.0858471692,-0.0476722531,-0.2291520685,0.1443036646,-0.2354615629,0.2881273329,0.2030762136,0.2164718509,0.4846534729,-0.427249223,0.0382829905,-0.2070930153,0.0976076871,0.3163167238,-0.2574578226,0.0504572801,-0.16220209,0.0547889322,0.0131520722,0.0525103472,0.1396277249,-0.0389830656,0.0427073203,0.333689034,-0.1388189495,0.042830985,-0.2334407568,0.0458435602,0.4255111814,0.1589772552,0.051013343,0.2905644774,-0.2684130967,0.0809304491,0.0158537552,0.1407580674,0.2980785966,-0.0071150423,0.0680872425,0.0217248164,0.1491240412,0.2423139662,0.0005278526,-0.4159061313,-0.0420816727,0.3158295155,-0.0500427932,0.3092298806,0.1688649356,0.3333734274,-0.1892764717,0.0969858468,-0.4036626816,-0.1670233607,-0.0538140684,-0.2211986184,-0.5511418581,-0.2069526762,-0.3715268075,-0.1109172553,-0.1350990832,-0.0615067855,-0.2741713226,0.1886223704,0.0503452383,-0.330363065,0.0814077035,-0.1955392063,0.2041691542,-0.1523468047,0.2263703197,-0.0093282107,0.0362457894,0.3188800812,0.0760150179,0.5093174577,0.2283807099,-0.0963346064,-0.2487950623,-0.0241237804,-0.142178461,0.040377859,0.3468794227,0.1857326627,0.0395644121,0.2420821488,0.1516417563,-0.1133231595,0.3565611541,0.002253989,0.263456285,-0.0992489904,0.5178786516,-0.467658788,0.1000935584,-0.1672317088,-0.3525443077,-0.1460466087,-0.3188263476,-0.1780048162,-0.1945900023,-0.0235637128,-0.058697395,0.0232757386,-0.0869168863,0.4179295003,0.394969821,0.1819574684,0.1531835794,-0.033079084,-0.2972991467,0.2351235002,-0.1318951845,0.0307580009,0.2157548368,0.2165423781,0.0767683834,0.1726402193,0.0412504598,0.0387746431,0.0111108012,0.2165732384,-0.2433100045,-0.060975045,0.2753140926,0.0173532814,-0.1357738823,-0.2809374332,-0.0888741985,-0.0585748106,-0.0598330982,0.1518500149,-0.0595391653,-0.0654157251,-0.1698684841,0.4425504208,0.1324360222,0.2011087984,-0.0575092882,-0.4760095477,-0.2689436674,-0.4144521952,-0.1308504641,0.3707846403,-0.1804432422,0.6111345291,0.0342233665,0.1484044045,-0.0185195096,-0.0448632054,-0.1875070333,-0.0982181877,-0.5277866125,0.2939006686,-0.3957848847,0.1447350532,-0.1315573007,0.1057318822,0.1239854619,0.1936321557,-0.0772358403,-0.2317139655,0.4982314706,-0.2602384984,-0.3133595288,0.2727927864,0.0889144912,-0.1325899363,-0.2973820269,-0.6104027033,0.0109965969,0.2189054042,0.0305512454,-0.2178259343,-0.1124497503,-0.124261193,-0.05467695,-0.1363347769,0.2223986536,0.0136693334,-0.1138071492,0.1441099644,-0.1869696826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2799","title":"Loading JSON throws ArrowNotImplementedError","comments":"> I just took one of your JSON lines an pyarrow.json read it without problem.\r\n\r\nyes, and for some peculiar reason the error is non-deterministic (i was eventually able to load the whole dataset by just re-running the `load_dataset` cell multiple times \ud83e\udd14)\r\n\r\nthanks for looking into this \ud83d\ude4f !","body":"## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":50,"text":"Loading JSON throws ArrowNotImplementedError\n## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\n> I just took one of your JSON lines an pyarrow.json read it without problem.\r\n\r\nyes, and for some peculiar reason the error is non-deterministic (i was eventually able to load the whole dataset by just re-running the `load_dataset` cell multiple times \ud83e\udd14)\r\n\r\nthanks for looking into this \ud83d\ude4f !","embeddings":[-0.0681871027,0.1841620058,0.0364155807,0.3780699968,0.2355371863,-0.0062702368,0.4846684635,0.3670629561,0.4180777371,-0.0028386964,0.0605546869,0.5898874402,0.0276210085,-0.148857981,-0.2172110528,-0.1838730276,0.0313557796,0.1905159801,0.1274458319,0.0285853427,-0.3122327328,0.2942055166,-0.214816317,0.1017739549,-0.0391265377,-0.0952268541,0.1041623503,0.3933393061,-0.126418829,-0.3826465905,0.4134063721,-0.0130348513,0.2161162943,0.5308527946,-0.0001203599,0.1964653581,0.4152528644,-0.092444241,-0.2868113518,-0.4780595005,-0.2790301442,-0.3253917992,0.537475884,-0.1200379506,0.1052053347,-0.2437469363,-0.0497698858,-0.2207819521,0.6015406847,0.2307226658,0.1048027351,0.429939568,0.0260289963,0.0192190502,0.0637280419,0.2108367532,-0.1217756793,0.4542036653,0.008248603,0.1270525903,-0.1715968549,-0.0036361166,-0.102641046,0.1538670212,0.2191231996,0.0476008654,-0.1704021692,0.0818345174,0.0492499694,0.2946319878,0.2845771015,-0.0508809648,-0.4122377634,-0.3489953876,-0.1207850203,0.1302726865,0.2388218939,0.2473598123,-0.1509169936,0.1259337515,0.0328827612,-0.0184712112,-0.274197191,0.2383805364,0.0036435723,0.0111869918,-0.1582194567,0.10729184,-0.1295856088,-0.020824587,-0.1499335319,-0.2244829088,-0.3057584167,0.1283780038,0.0374987274,0.11600402,-0.0355978459,-0.4100777805,0.2355834544,0.3558476567,0.3206709325,0.2080192715,-0.2673651874,0.2078464776,0.5903261304,0.2016133815,-0.0549013875,0.0114986571,0.2564998567,0.330514431,0.0607299209,-0.0435533673,-0.2637991011,-0.1108330786,0.0103772897,-0.197450906,0.3086808324,-0.1726259738,-0.2238530517,0.0755632967,-0.6134242415,-0.0573209412,-0.0002183132,0.3672920167,-0.1971875429,0.4987778664,0.2744282484,0.4868384302,0.0442172214,0.0317274667,-0.1022381261,-0.0353484266,-0.1634835154,-0.2962282002,0.2371660173,0.0525447652,0.0861053467,0.1265179962,-0.0630014017,-0.4126393497,-0.0742599964,0.0477150232,-0.0974051356,0.16696693,0.1920519024,0.132871151,0.1559260786,-0.4045022726,-0.1376460344,0.0323025733,-0.376919657,0.0055578146,-0.3176852167,0.0546303503,-0.2702908218,-0.0120817386,-0.7623512149,-0.0521201082,-0.1122782081,0.0660267845,0.0638756305,-0.0244168527,0.2227595448,-0.2470848262,0.0749816447,0.28331393,-0.3005584776,-0.2378236353,-0.0629288927,-0.037491519,0.1896335036,0.2650878727,-0.2005772442,0.2845451832,-0.1637352109,0.2106698602,0.3121518195,-0.3327314556,-0.2219264358,0.605520308,0.0193049405,0.5248504877,0.0198842697,-0.2308956832,-0.0092007685,0.1884041429,0.3035949767,0.4013707638,0.1422874182,0.0679152384,-0.2457050234,-0.2022635043,-0.1346425563,0.3127692342,0.0487670116,0.1889399141,0.0995356888,0.0244781338,0.2506872714,-0.1951391548,-0.0266671889,0.1624232382,0.0911331177,0.0282151625,-0.0057259151,-0.0044113393,-0.6142941117,0.1499558836,-0.0511226989,0.109201923,-0.408085674,0.0093207462,-0.2547832429,0.1260678917,-0.1280752718,0.1847681105,0.0374528877,0.1086534187,-0.1821872592,0.2108015567,-0.1838937998,0.3152327836,0.1626450419,0.2816524506,-0.4834640026,0.4377061725,0.210255906,-0.1212381497,0.0271184649,0.1853716373,-0.097817637,-0.1626708806,-0.2391357571,0.1727546602,0.1394181401,0.1797138602,-0.0562768616,-0.2375711501,-0.0822836161,0.0838592127,-0.1951901019,0.3068773448,0.2217241377,-0.0245843939,-0.0280968156,0.8052088618,-0.0741070136,0.4431970716,-0.015723357,-0.0205509756,0.4561716616,0.0615659803,-0.364153266,-0.3053402305,0.2388348728,0.3186245859,0.0603672415,0.0040045432,-0.3794290721,-0.1885213256,0.3982873261,0.005179801,0.2424625754,0.1532346159,-0.0953567475,-0.0996729881,0.0158600174,0.1401785463,0.2382987291,0.0758770034,-0.2070478052,-0.0434467606,-0.0409476832,-0.0247451495,0.1555987,0.306535244,0.2641779482,0.0933471769,0.1558886766,-0.020962242,-0.2815460265,-0.238003552,-0.2863601744,0.1189386174,-0.6182570457,0.2991375625,-0.4743469954,0.1188910827,0.035333585,-0.3410826325,-0.126299873,-0.322971493,-0.3007420599,0.2734274864,-0.1147493422,-0.1341666132,-0.2667239308,0.3151707649,-0.0232129041,-0.3463473022,-0.2113419473,-0.0117184091,-0.184648335,-0.0234055538,0.3875870109,0.1092963666,0.1544730067,-0.1289402992,0.047313679,0.1418192536,-0.0863964856,0.1482000202,-0.1994989067,0.2889936268,0.2154709399,0.1969099492,-0.0978527516,-0.1516460776,0.2886652946,0.0957521275,-0.3427708745,0.3241927326,-0.1354984939,0.0278997961,-0.1492165327,-0.0291694198,0.1219549254,-0.4392543733,0.5247157216,0.2458687872,0.0581384897,-0.0170072187,0.0242614429,0.1799215674,0.1687064171,0.0159113072,-0.1107467487,-0.1546859592,0.2326568216,-0.160832569,-0.2501766384,0.0764547884,-0.1965518147,0.1685908139,-0.0865133032,-0.390976578,0.0259424336,0.0707196668,0.1136565655,-0.0444040187,-0.295239538,0.0588247478,-0.0127509227,0.0851902887,-0.1855715215,-0.2022252828,0.1206316799,-0.0438129231,0.0216035321,0.0565150753,0.7252899408,-0.0942959189,0.1807754487,0.3117997646,-0.0382717438,0.394974649,-0.0264437217,0.1030089855,-0.4057941437,-0.1191259399,-0.3808179796,-0.0971131027,0.0786372274,0.2193027288,0.1096423194,-0.2879481912,-0.2405597121,-0.0622841977,-0.4587052464,-0.3489450216,0.1333795786,-0.3543255031,-0.3382495046,0.0320333503,0.2205641121,-0.0587302446,-0.0581021123,-0.0303620175,0.4014143944,0.2062794119,-0.0455005579,-0.2057112902,0.0343681015,-0.3544535339,0.3606956899,0.0746525601,0.3657507002,0.1974582672,-0.028097624,0.2447486371,-0.2093995959,0.4490728974,-0.120011054,0.0227281768,0.0280953031,-0.1236653179,-0.6418688893,0.1011512503,-0.1271315068,0.2907666266,0.2359819263,0.2095480561,-0.4572590888,-0.2961414456,0.1018688306,-0.1504392922,-0.1058704555,-0.1064765677,-0.3640385568,-0.2923769653,-0.2448295504,-0.1525726318,0.0274700802,0.2387111634,0.0874371156,-0.0987067744,-0.2664907575,-0.0690853894,0.1225850284,0.2164159268,0.1880879253,-0.2173919529,0.0409795344,0.0061119525,0.3643347025,0.3959253132,0.8496598601,0.2561551034,-0.5346414447,0.2092527151,-0.0245482698,0.0420461036,0.1865970343,0.0230351947,0.0835103542,0.2850650549,0.1448794007,-0.2175282389,-0.1765887737,0.4804262519,-0.1738846451,-0.4621180296,-0.346393168,0.4906897545,-0.1168812439,0.0213299748,0.2263398916,0.2678977251,-0.1467050314,0.758928895,0.0028833551,0.7930403352,0.1487369686,0.1980527639,0.7221454978,-0.5317468047,0.400482446,-0.0322305784,-0.0708860084,-0.2383532822,0.5987789035,-0.1820171326,-0.151582703,-0.0095944684,0.192508921,-0.1030690968,0.0225751493,-0.3046782017,-0.0162525531,0.0787486061,0.128597334,-0.1141062975,-0.2810204625,-0.7139284611,-0.0139217712,-0.1206339747,0.0340203084,-0.0035200173,-0.3064055443,-0.2558885217,-0.188852787,-0.3464414477,0.2579855025,0.1630945057,-0.0184752066,0.2230367512,-0.0496139601,0.0946815684,-0.0524797849,0.2925044,-0.0933387354,-0.1754464954,0.187010482,-0.1364040524,-0.0991375446,-0.0924463496,-0.016065022,0.5005759001,-0.0768829286,-0.1589543819,-0.2050989121,-0.2627161145,-0.5380346775,0.2646922767,0.4163651466,-0.3466082215,-0.240976274,-0.393114835,-0.0475282073,0.2723071873,-0.1570670009,0.0498477258,0.2624671161,0.1839359552,-0.0674586892,0.2908512056,-0.3609032929,-0.1087285951,0.3222659528,-0.0353102535,0.0564952418,0.5033448339,0.2862301469,-0.2179954052,-0.1904615164,0.2036407739,0.8105605841,-0.6269248128,0.3065258563,-0.1730780452,-0.0535799339,-0.0472417548,0.3028697371,0.0699545369,-0.0232218802,0.0566213802,-0.6213235855,-0.1062008291,0.4122470915,-0.1543131769,-0.0214453992,0.1351440996,-0.0169346631,-0.2405516505,-0.1720611304,-0.17293562,0.2192620784,-0.1576303691,-0.2045778781,-0.3086437285,0.0632046908,0.5788154602,0.0068058115,0.0865861177,0.0817479491,-0.1436324716,-0.0885987878,-0.3009876311,0.2072960436,0.1043563411,-0.0927105248,-0.0784604624,0.0101217115,-0.1790237874,-0.135748744,0.2507112324,0.2667553127,-0.0477175564,-0.1331558526,0.0813497081,0.1524973363,-0.2447343022,0.1382133961,-0.159083724,0.5401737094,0.0858471692,-0.0476722531,-0.2291520685,0.1443036646,-0.2354615629,0.2881273329,0.2030762136,0.2164718509,0.4846534729,-0.427249223,0.0382829905,-0.2070930153,0.0976076871,0.3163167238,-0.2574578226,0.0504572801,-0.16220209,0.0547889322,0.0131520722,0.0525103472,0.1396277249,-0.0389830656,0.0427073203,0.333689034,-0.1388189495,0.042830985,-0.2334407568,0.0458435602,0.4255111814,0.1589772552,0.051013343,0.2905644774,-0.2684130967,0.0809304491,0.0158537552,0.1407580674,0.2980785966,-0.0071150423,0.0680872425,0.0217248164,0.1491240412,0.2423139662,0.0005278526,-0.4159061313,-0.0420816727,0.3158295155,-0.0500427932,0.3092298806,0.1688649356,0.3333734274,-0.1892764717,0.0969858468,-0.4036626816,-0.1670233607,-0.0538140684,-0.2211986184,-0.5511418581,-0.2069526762,-0.3715268075,-0.1109172553,-0.1350990832,-0.0615067855,-0.2741713226,0.1886223704,0.0503452383,-0.330363065,0.0814077035,-0.1955392063,0.2041691542,-0.1523468047,0.2263703197,-0.0093282107,0.0362457894,0.3188800812,0.0760150179,0.5093174577,0.2283807099,-0.0963346064,-0.2487950623,-0.0241237804,-0.142178461,0.040377859,0.3468794227,0.1857326627,0.0395644121,0.2420821488,0.1516417563,-0.1133231595,0.3565611541,0.002253989,0.263456285,-0.0992489904,0.5178786516,-0.467658788,0.1000935584,-0.1672317088,-0.3525443077,-0.1460466087,-0.3188263476,-0.1780048162,-0.1945900023,-0.0235637128,-0.058697395,0.0232757386,-0.0869168863,0.4179295003,0.394969821,0.1819574684,0.1531835794,-0.033079084,-0.2972991467,0.2351235002,-0.1318951845,0.0307580009,0.2157548368,0.2165423781,0.0767683834,0.1726402193,0.0412504598,0.0387746431,0.0111108012,0.2165732384,-0.2433100045,-0.060975045,0.2753140926,0.0173532814,-0.1357738823,-0.2809374332,-0.0888741985,-0.0585748106,-0.0598330982,0.1518500149,-0.0595391653,-0.0654157251,-0.1698684841,0.4425504208,0.1324360222,0.2011087984,-0.0575092882,-0.4760095477,-0.2689436674,-0.4144521952,-0.1308504641,0.3707846403,-0.1804432422,0.6111345291,0.0342233665,0.1484044045,-0.0185195096,-0.0448632054,-0.1875070333,-0.0982181877,-0.5277866125,0.2939006686,-0.3957848847,0.1447350532,-0.1315573007,0.1057318822,0.1239854619,0.1936321557,-0.0772358403,-0.2317139655,0.4982314706,-0.2602384984,-0.3133595288,0.2727927864,0.0889144912,-0.1325899363,-0.2973820269,-0.6104027033,0.0109965969,0.2189054042,0.0305512454,-0.2178259343,-0.1124497503,-0.124261193,-0.05467695,-0.1363347769,0.2223986536,0.0136693334,-0.1138071492,0.1441099644,-0.1869696826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2799","title":"Loading JSON throws ArrowNotImplementedError","comments":"I think the error is generated by the `pyarrow.json.read()` option: `read_options=paj.ReadOptions(block_size=block_size)`...\r\ncc: @lhoestq ","body":"## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":13,"text":"Loading JSON throws ArrowNotImplementedError\n## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nI think the error is generated by the `pyarrow.json.read()` option: `read_options=paj.ReadOptions(block_size=block_size)`...\r\ncc: @lhoestq ","embeddings":[-0.0681871027,0.1841620058,0.0364155807,0.3780699968,0.2355371863,-0.0062702368,0.4846684635,0.3670629561,0.4180777371,-0.0028386964,0.0605546869,0.5898874402,0.0276210085,-0.148857981,-0.2172110528,-0.1838730276,0.0313557796,0.1905159801,0.1274458319,0.0285853427,-0.3122327328,0.2942055166,-0.214816317,0.1017739549,-0.0391265377,-0.0952268541,0.1041623503,0.3933393061,-0.126418829,-0.3826465905,0.4134063721,-0.0130348513,0.2161162943,0.5308527946,-0.0001203599,0.1964653581,0.4152528644,-0.092444241,-0.2868113518,-0.4780595005,-0.2790301442,-0.3253917992,0.537475884,-0.1200379506,0.1052053347,-0.2437469363,-0.0497698858,-0.2207819521,0.6015406847,0.2307226658,0.1048027351,0.429939568,0.0260289963,0.0192190502,0.0637280419,0.2108367532,-0.1217756793,0.4542036653,0.008248603,0.1270525903,-0.1715968549,-0.0036361166,-0.102641046,0.1538670212,0.2191231996,0.0476008654,-0.1704021692,0.0818345174,0.0492499694,0.2946319878,0.2845771015,-0.0508809648,-0.4122377634,-0.3489953876,-0.1207850203,0.1302726865,0.2388218939,0.2473598123,-0.1509169936,0.1259337515,0.0328827612,-0.0184712112,-0.274197191,0.2383805364,0.0036435723,0.0111869918,-0.1582194567,0.10729184,-0.1295856088,-0.020824587,-0.1499335319,-0.2244829088,-0.3057584167,0.1283780038,0.0374987274,0.11600402,-0.0355978459,-0.4100777805,0.2355834544,0.3558476567,0.3206709325,0.2080192715,-0.2673651874,0.2078464776,0.5903261304,0.2016133815,-0.0549013875,0.0114986571,0.2564998567,0.330514431,0.0607299209,-0.0435533673,-0.2637991011,-0.1108330786,0.0103772897,-0.197450906,0.3086808324,-0.1726259738,-0.2238530517,0.0755632967,-0.6134242415,-0.0573209412,-0.0002183132,0.3672920167,-0.1971875429,0.4987778664,0.2744282484,0.4868384302,0.0442172214,0.0317274667,-0.1022381261,-0.0353484266,-0.1634835154,-0.2962282002,0.2371660173,0.0525447652,0.0861053467,0.1265179962,-0.0630014017,-0.4126393497,-0.0742599964,0.0477150232,-0.0974051356,0.16696693,0.1920519024,0.132871151,0.1559260786,-0.4045022726,-0.1376460344,0.0323025733,-0.376919657,0.0055578146,-0.3176852167,0.0546303503,-0.2702908218,-0.0120817386,-0.7623512149,-0.0521201082,-0.1122782081,0.0660267845,0.0638756305,-0.0244168527,0.2227595448,-0.2470848262,0.0749816447,0.28331393,-0.3005584776,-0.2378236353,-0.0629288927,-0.037491519,0.1896335036,0.2650878727,-0.2005772442,0.2845451832,-0.1637352109,0.2106698602,0.3121518195,-0.3327314556,-0.2219264358,0.605520308,0.0193049405,0.5248504877,0.0198842697,-0.2308956832,-0.0092007685,0.1884041429,0.3035949767,0.4013707638,0.1422874182,0.0679152384,-0.2457050234,-0.2022635043,-0.1346425563,0.3127692342,0.0487670116,0.1889399141,0.0995356888,0.0244781338,0.2506872714,-0.1951391548,-0.0266671889,0.1624232382,0.0911331177,0.0282151625,-0.0057259151,-0.0044113393,-0.6142941117,0.1499558836,-0.0511226989,0.109201923,-0.408085674,0.0093207462,-0.2547832429,0.1260678917,-0.1280752718,0.1847681105,0.0374528877,0.1086534187,-0.1821872592,0.2108015567,-0.1838937998,0.3152327836,0.1626450419,0.2816524506,-0.4834640026,0.4377061725,0.210255906,-0.1212381497,0.0271184649,0.1853716373,-0.097817637,-0.1626708806,-0.2391357571,0.1727546602,0.1394181401,0.1797138602,-0.0562768616,-0.2375711501,-0.0822836161,0.0838592127,-0.1951901019,0.3068773448,0.2217241377,-0.0245843939,-0.0280968156,0.8052088618,-0.0741070136,0.4431970716,-0.015723357,-0.0205509756,0.4561716616,0.0615659803,-0.364153266,-0.3053402305,0.2388348728,0.3186245859,0.0603672415,0.0040045432,-0.3794290721,-0.1885213256,0.3982873261,0.005179801,0.2424625754,0.1532346159,-0.0953567475,-0.0996729881,0.0158600174,0.1401785463,0.2382987291,0.0758770034,-0.2070478052,-0.0434467606,-0.0409476832,-0.0247451495,0.1555987,0.306535244,0.2641779482,0.0933471769,0.1558886766,-0.020962242,-0.2815460265,-0.238003552,-0.2863601744,0.1189386174,-0.6182570457,0.2991375625,-0.4743469954,0.1188910827,0.035333585,-0.3410826325,-0.126299873,-0.322971493,-0.3007420599,0.2734274864,-0.1147493422,-0.1341666132,-0.2667239308,0.3151707649,-0.0232129041,-0.3463473022,-0.2113419473,-0.0117184091,-0.184648335,-0.0234055538,0.3875870109,0.1092963666,0.1544730067,-0.1289402992,0.047313679,0.1418192536,-0.0863964856,0.1482000202,-0.1994989067,0.2889936268,0.2154709399,0.1969099492,-0.0978527516,-0.1516460776,0.2886652946,0.0957521275,-0.3427708745,0.3241927326,-0.1354984939,0.0278997961,-0.1492165327,-0.0291694198,0.1219549254,-0.4392543733,0.5247157216,0.2458687872,0.0581384897,-0.0170072187,0.0242614429,0.1799215674,0.1687064171,0.0159113072,-0.1107467487,-0.1546859592,0.2326568216,-0.160832569,-0.2501766384,0.0764547884,-0.1965518147,0.1685908139,-0.0865133032,-0.390976578,0.0259424336,0.0707196668,0.1136565655,-0.0444040187,-0.295239538,0.0588247478,-0.0127509227,0.0851902887,-0.1855715215,-0.2022252828,0.1206316799,-0.0438129231,0.0216035321,0.0565150753,0.7252899408,-0.0942959189,0.1807754487,0.3117997646,-0.0382717438,0.394974649,-0.0264437217,0.1030089855,-0.4057941437,-0.1191259399,-0.3808179796,-0.0971131027,0.0786372274,0.2193027288,0.1096423194,-0.2879481912,-0.2405597121,-0.0622841977,-0.4587052464,-0.3489450216,0.1333795786,-0.3543255031,-0.3382495046,0.0320333503,0.2205641121,-0.0587302446,-0.0581021123,-0.0303620175,0.4014143944,0.2062794119,-0.0455005579,-0.2057112902,0.0343681015,-0.3544535339,0.3606956899,0.0746525601,0.3657507002,0.1974582672,-0.028097624,0.2447486371,-0.2093995959,0.4490728974,-0.120011054,0.0227281768,0.0280953031,-0.1236653179,-0.6418688893,0.1011512503,-0.1271315068,0.2907666266,0.2359819263,0.2095480561,-0.4572590888,-0.2961414456,0.1018688306,-0.1504392922,-0.1058704555,-0.1064765677,-0.3640385568,-0.2923769653,-0.2448295504,-0.1525726318,0.0274700802,0.2387111634,0.0874371156,-0.0987067744,-0.2664907575,-0.0690853894,0.1225850284,0.2164159268,0.1880879253,-0.2173919529,0.0409795344,0.0061119525,0.3643347025,0.3959253132,0.8496598601,0.2561551034,-0.5346414447,0.2092527151,-0.0245482698,0.0420461036,0.1865970343,0.0230351947,0.0835103542,0.2850650549,0.1448794007,-0.2175282389,-0.1765887737,0.4804262519,-0.1738846451,-0.4621180296,-0.346393168,0.4906897545,-0.1168812439,0.0213299748,0.2263398916,0.2678977251,-0.1467050314,0.758928895,0.0028833551,0.7930403352,0.1487369686,0.1980527639,0.7221454978,-0.5317468047,0.400482446,-0.0322305784,-0.0708860084,-0.2383532822,0.5987789035,-0.1820171326,-0.151582703,-0.0095944684,0.192508921,-0.1030690968,0.0225751493,-0.3046782017,-0.0162525531,0.0787486061,0.128597334,-0.1141062975,-0.2810204625,-0.7139284611,-0.0139217712,-0.1206339747,0.0340203084,-0.0035200173,-0.3064055443,-0.2558885217,-0.188852787,-0.3464414477,0.2579855025,0.1630945057,-0.0184752066,0.2230367512,-0.0496139601,0.0946815684,-0.0524797849,0.2925044,-0.0933387354,-0.1754464954,0.187010482,-0.1364040524,-0.0991375446,-0.0924463496,-0.016065022,0.5005759001,-0.0768829286,-0.1589543819,-0.2050989121,-0.2627161145,-0.5380346775,0.2646922767,0.4163651466,-0.3466082215,-0.240976274,-0.393114835,-0.0475282073,0.2723071873,-0.1570670009,0.0498477258,0.2624671161,0.1839359552,-0.0674586892,0.2908512056,-0.3609032929,-0.1087285951,0.3222659528,-0.0353102535,0.0564952418,0.5033448339,0.2862301469,-0.2179954052,-0.1904615164,0.2036407739,0.8105605841,-0.6269248128,0.3065258563,-0.1730780452,-0.0535799339,-0.0472417548,0.3028697371,0.0699545369,-0.0232218802,0.0566213802,-0.6213235855,-0.1062008291,0.4122470915,-0.1543131769,-0.0214453992,0.1351440996,-0.0169346631,-0.2405516505,-0.1720611304,-0.17293562,0.2192620784,-0.1576303691,-0.2045778781,-0.3086437285,0.0632046908,0.5788154602,0.0068058115,0.0865861177,0.0817479491,-0.1436324716,-0.0885987878,-0.3009876311,0.2072960436,0.1043563411,-0.0927105248,-0.0784604624,0.0101217115,-0.1790237874,-0.135748744,0.2507112324,0.2667553127,-0.0477175564,-0.1331558526,0.0813497081,0.1524973363,-0.2447343022,0.1382133961,-0.159083724,0.5401737094,0.0858471692,-0.0476722531,-0.2291520685,0.1443036646,-0.2354615629,0.2881273329,0.2030762136,0.2164718509,0.4846534729,-0.427249223,0.0382829905,-0.2070930153,0.0976076871,0.3163167238,-0.2574578226,0.0504572801,-0.16220209,0.0547889322,0.0131520722,0.0525103472,0.1396277249,-0.0389830656,0.0427073203,0.333689034,-0.1388189495,0.042830985,-0.2334407568,0.0458435602,0.4255111814,0.1589772552,0.051013343,0.2905644774,-0.2684130967,0.0809304491,0.0158537552,0.1407580674,0.2980785966,-0.0071150423,0.0680872425,0.0217248164,0.1491240412,0.2423139662,0.0005278526,-0.4159061313,-0.0420816727,0.3158295155,-0.0500427932,0.3092298806,0.1688649356,0.3333734274,-0.1892764717,0.0969858468,-0.4036626816,-0.1670233607,-0.0538140684,-0.2211986184,-0.5511418581,-0.2069526762,-0.3715268075,-0.1109172553,-0.1350990832,-0.0615067855,-0.2741713226,0.1886223704,0.0503452383,-0.330363065,0.0814077035,-0.1955392063,0.2041691542,-0.1523468047,0.2263703197,-0.0093282107,0.0362457894,0.3188800812,0.0760150179,0.5093174577,0.2283807099,-0.0963346064,-0.2487950623,-0.0241237804,-0.142178461,0.040377859,0.3468794227,0.1857326627,0.0395644121,0.2420821488,0.1516417563,-0.1133231595,0.3565611541,0.002253989,0.263456285,-0.0992489904,0.5178786516,-0.467658788,0.1000935584,-0.1672317088,-0.3525443077,-0.1460466087,-0.3188263476,-0.1780048162,-0.1945900023,-0.0235637128,-0.058697395,0.0232757386,-0.0869168863,0.4179295003,0.394969821,0.1819574684,0.1531835794,-0.033079084,-0.2972991467,0.2351235002,-0.1318951845,0.0307580009,0.2157548368,0.2165423781,0.0767683834,0.1726402193,0.0412504598,0.0387746431,0.0111108012,0.2165732384,-0.2433100045,-0.060975045,0.2753140926,0.0173532814,-0.1357738823,-0.2809374332,-0.0888741985,-0.0585748106,-0.0598330982,0.1518500149,-0.0595391653,-0.0654157251,-0.1698684841,0.4425504208,0.1324360222,0.2011087984,-0.0575092882,-0.4760095477,-0.2689436674,-0.4144521952,-0.1308504641,0.3707846403,-0.1804432422,0.6111345291,0.0342233665,0.1484044045,-0.0185195096,-0.0448632054,-0.1875070333,-0.0982181877,-0.5277866125,0.2939006686,-0.3957848847,0.1447350532,-0.1315573007,0.1057318822,0.1239854619,0.1936321557,-0.0772358403,-0.2317139655,0.4982314706,-0.2602384984,-0.3133595288,0.2727927864,0.0889144912,-0.1325899363,-0.2973820269,-0.6104027033,0.0109965969,0.2189054042,0.0305512454,-0.2178259343,-0.1124497503,-0.124261193,-0.05467695,-0.1363347769,0.2223986536,0.0136693334,-0.1138071492,0.1441099644,-0.1869696826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2799","title":"Loading JSON throws ArrowNotImplementedError","comments":"The code works fine on my side.\r\nNot sure what's going on here :\/\r\n\r\nI remember we did a few changes in the JSON loader in #2638 , did you do an update `datasets` when debugging this ?\r\n","body":"## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":38,"text":"Loading JSON throws ArrowNotImplementedError\n## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nThe code works fine on my side.\r\nNot sure what's going on here :\/\r\n\r\nI remember we did a few changes in the JSON loader in #2638 , did you do an update `datasets` when debugging this ?\r\n","embeddings":[-0.0681871027,0.1841620058,0.0364155807,0.3780699968,0.2355371863,-0.0062702368,0.4846684635,0.3670629561,0.4180777371,-0.0028386964,0.0605546869,0.5898874402,0.0276210085,-0.148857981,-0.2172110528,-0.1838730276,0.0313557796,0.1905159801,0.1274458319,0.0285853427,-0.3122327328,0.2942055166,-0.214816317,0.1017739549,-0.0391265377,-0.0952268541,0.1041623503,0.3933393061,-0.126418829,-0.3826465905,0.4134063721,-0.0130348513,0.2161162943,0.5308527946,-0.0001203599,0.1964653581,0.4152528644,-0.092444241,-0.2868113518,-0.4780595005,-0.2790301442,-0.3253917992,0.537475884,-0.1200379506,0.1052053347,-0.2437469363,-0.0497698858,-0.2207819521,0.6015406847,0.2307226658,0.1048027351,0.429939568,0.0260289963,0.0192190502,0.0637280419,0.2108367532,-0.1217756793,0.4542036653,0.008248603,0.1270525903,-0.1715968549,-0.0036361166,-0.102641046,0.1538670212,0.2191231996,0.0476008654,-0.1704021692,0.0818345174,0.0492499694,0.2946319878,0.2845771015,-0.0508809648,-0.4122377634,-0.3489953876,-0.1207850203,0.1302726865,0.2388218939,0.2473598123,-0.1509169936,0.1259337515,0.0328827612,-0.0184712112,-0.274197191,0.2383805364,0.0036435723,0.0111869918,-0.1582194567,0.10729184,-0.1295856088,-0.020824587,-0.1499335319,-0.2244829088,-0.3057584167,0.1283780038,0.0374987274,0.11600402,-0.0355978459,-0.4100777805,0.2355834544,0.3558476567,0.3206709325,0.2080192715,-0.2673651874,0.2078464776,0.5903261304,0.2016133815,-0.0549013875,0.0114986571,0.2564998567,0.330514431,0.0607299209,-0.0435533673,-0.2637991011,-0.1108330786,0.0103772897,-0.197450906,0.3086808324,-0.1726259738,-0.2238530517,0.0755632967,-0.6134242415,-0.0573209412,-0.0002183132,0.3672920167,-0.1971875429,0.4987778664,0.2744282484,0.4868384302,0.0442172214,0.0317274667,-0.1022381261,-0.0353484266,-0.1634835154,-0.2962282002,0.2371660173,0.0525447652,0.0861053467,0.1265179962,-0.0630014017,-0.4126393497,-0.0742599964,0.0477150232,-0.0974051356,0.16696693,0.1920519024,0.132871151,0.1559260786,-0.4045022726,-0.1376460344,0.0323025733,-0.376919657,0.0055578146,-0.3176852167,0.0546303503,-0.2702908218,-0.0120817386,-0.7623512149,-0.0521201082,-0.1122782081,0.0660267845,0.0638756305,-0.0244168527,0.2227595448,-0.2470848262,0.0749816447,0.28331393,-0.3005584776,-0.2378236353,-0.0629288927,-0.037491519,0.1896335036,0.2650878727,-0.2005772442,0.2845451832,-0.1637352109,0.2106698602,0.3121518195,-0.3327314556,-0.2219264358,0.605520308,0.0193049405,0.5248504877,0.0198842697,-0.2308956832,-0.0092007685,0.1884041429,0.3035949767,0.4013707638,0.1422874182,0.0679152384,-0.2457050234,-0.2022635043,-0.1346425563,0.3127692342,0.0487670116,0.1889399141,0.0995356888,0.0244781338,0.2506872714,-0.1951391548,-0.0266671889,0.1624232382,0.0911331177,0.0282151625,-0.0057259151,-0.0044113393,-0.6142941117,0.1499558836,-0.0511226989,0.109201923,-0.408085674,0.0093207462,-0.2547832429,0.1260678917,-0.1280752718,0.1847681105,0.0374528877,0.1086534187,-0.1821872592,0.2108015567,-0.1838937998,0.3152327836,0.1626450419,0.2816524506,-0.4834640026,0.4377061725,0.210255906,-0.1212381497,0.0271184649,0.1853716373,-0.097817637,-0.1626708806,-0.2391357571,0.1727546602,0.1394181401,0.1797138602,-0.0562768616,-0.2375711501,-0.0822836161,0.0838592127,-0.1951901019,0.3068773448,0.2217241377,-0.0245843939,-0.0280968156,0.8052088618,-0.0741070136,0.4431970716,-0.015723357,-0.0205509756,0.4561716616,0.0615659803,-0.364153266,-0.3053402305,0.2388348728,0.3186245859,0.0603672415,0.0040045432,-0.3794290721,-0.1885213256,0.3982873261,0.005179801,0.2424625754,0.1532346159,-0.0953567475,-0.0996729881,0.0158600174,0.1401785463,0.2382987291,0.0758770034,-0.2070478052,-0.0434467606,-0.0409476832,-0.0247451495,0.1555987,0.306535244,0.2641779482,0.0933471769,0.1558886766,-0.020962242,-0.2815460265,-0.238003552,-0.2863601744,0.1189386174,-0.6182570457,0.2991375625,-0.4743469954,0.1188910827,0.035333585,-0.3410826325,-0.126299873,-0.322971493,-0.3007420599,0.2734274864,-0.1147493422,-0.1341666132,-0.2667239308,0.3151707649,-0.0232129041,-0.3463473022,-0.2113419473,-0.0117184091,-0.184648335,-0.0234055538,0.3875870109,0.1092963666,0.1544730067,-0.1289402992,0.047313679,0.1418192536,-0.0863964856,0.1482000202,-0.1994989067,0.2889936268,0.2154709399,0.1969099492,-0.0978527516,-0.1516460776,0.2886652946,0.0957521275,-0.3427708745,0.3241927326,-0.1354984939,0.0278997961,-0.1492165327,-0.0291694198,0.1219549254,-0.4392543733,0.5247157216,0.2458687872,0.0581384897,-0.0170072187,0.0242614429,0.1799215674,0.1687064171,0.0159113072,-0.1107467487,-0.1546859592,0.2326568216,-0.160832569,-0.2501766384,0.0764547884,-0.1965518147,0.1685908139,-0.0865133032,-0.390976578,0.0259424336,0.0707196668,0.1136565655,-0.0444040187,-0.295239538,0.0588247478,-0.0127509227,0.0851902887,-0.1855715215,-0.2022252828,0.1206316799,-0.0438129231,0.0216035321,0.0565150753,0.7252899408,-0.0942959189,0.1807754487,0.3117997646,-0.0382717438,0.394974649,-0.0264437217,0.1030089855,-0.4057941437,-0.1191259399,-0.3808179796,-0.0971131027,0.0786372274,0.2193027288,0.1096423194,-0.2879481912,-0.2405597121,-0.0622841977,-0.4587052464,-0.3489450216,0.1333795786,-0.3543255031,-0.3382495046,0.0320333503,0.2205641121,-0.0587302446,-0.0581021123,-0.0303620175,0.4014143944,0.2062794119,-0.0455005579,-0.2057112902,0.0343681015,-0.3544535339,0.3606956899,0.0746525601,0.3657507002,0.1974582672,-0.028097624,0.2447486371,-0.2093995959,0.4490728974,-0.120011054,0.0227281768,0.0280953031,-0.1236653179,-0.6418688893,0.1011512503,-0.1271315068,0.2907666266,0.2359819263,0.2095480561,-0.4572590888,-0.2961414456,0.1018688306,-0.1504392922,-0.1058704555,-0.1064765677,-0.3640385568,-0.2923769653,-0.2448295504,-0.1525726318,0.0274700802,0.2387111634,0.0874371156,-0.0987067744,-0.2664907575,-0.0690853894,0.1225850284,0.2164159268,0.1880879253,-0.2173919529,0.0409795344,0.0061119525,0.3643347025,0.3959253132,0.8496598601,0.2561551034,-0.5346414447,0.2092527151,-0.0245482698,0.0420461036,0.1865970343,0.0230351947,0.0835103542,0.2850650549,0.1448794007,-0.2175282389,-0.1765887737,0.4804262519,-0.1738846451,-0.4621180296,-0.346393168,0.4906897545,-0.1168812439,0.0213299748,0.2263398916,0.2678977251,-0.1467050314,0.758928895,0.0028833551,0.7930403352,0.1487369686,0.1980527639,0.7221454978,-0.5317468047,0.400482446,-0.0322305784,-0.0708860084,-0.2383532822,0.5987789035,-0.1820171326,-0.151582703,-0.0095944684,0.192508921,-0.1030690968,0.0225751493,-0.3046782017,-0.0162525531,0.0787486061,0.128597334,-0.1141062975,-0.2810204625,-0.7139284611,-0.0139217712,-0.1206339747,0.0340203084,-0.0035200173,-0.3064055443,-0.2558885217,-0.188852787,-0.3464414477,0.2579855025,0.1630945057,-0.0184752066,0.2230367512,-0.0496139601,0.0946815684,-0.0524797849,0.2925044,-0.0933387354,-0.1754464954,0.187010482,-0.1364040524,-0.0991375446,-0.0924463496,-0.016065022,0.5005759001,-0.0768829286,-0.1589543819,-0.2050989121,-0.2627161145,-0.5380346775,0.2646922767,0.4163651466,-0.3466082215,-0.240976274,-0.393114835,-0.0475282073,0.2723071873,-0.1570670009,0.0498477258,0.2624671161,0.1839359552,-0.0674586892,0.2908512056,-0.3609032929,-0.1087285951,0.3222659528,-0.0353102535,0.0564952418,0.5033448339,0.2862301469,-0.2179954052,-0.1904615164,0.2036407739,0.8105605841,-0.6269248128,0.3065258563,-0.1730780452,-0.0535799339,-0.0472417548,0.3028697371,0.0699545369,-0.0232218802,0.0566213802,-0.6213235855,-0.1062008291,0.4122470915,-0.1543131769,-0.0214453992,0.1351440996,-0.0169346631,-0.2405516505,-0.1720611304,-0.17293562,0.2192620784,-0.1576303691,-0.2045778781,-0.3086437285,0.0632046908,0.5788154602,0.0068058115,0.0865861177,0.0817479491,-0.1436324716,-0.0885987878,-0.3009876311,0.2072960436,0.1043563411,-0.0927105248,-0.0784604624,0.0101217115,-0.1790237874,-0.135748744,0.2507112324,0.2667553127,-0.0477175564,-0.1331558526,0.0813497081,0.1524973363,-0.2447343022,0.1382133961,-0.159083724,0.5401737094,0.0858471692,-0.0476722531,-0.2291520685,0.1443036646,-0.2354615629,0.2881273329,0.2030762136,0.2164718509,0.4846534729,-0.427249223,0.0382829905,-0.2070930153,0.0976076871,0.3163167238,-0.2574578226,0.0504572801,-0.16220209,0.0547889322,0.0131520722,0.0525103472,0.1396277249,-0.0389830656,0.0427073203,0.333689034,-0.1388189495,0.042830985,-0.2334407568,0.0458435602,0.4255111814,0.1589772552,0.051013343,0.2905644774,-0.2684130967,0.0809304491,0.0158537552,0.1407580674,0.2980785966,-0.0071150423,0.0680872425,0.0217248164,0.1491240412,0.2423139662,0.0005278526,-0.4159061313,-0.0420816727,0.3158295155,-0.0500427932,0.3092298806,0.1688649356,0.3333734274,-0.1892764717,0.0969858468,-0.4036626816,-0.1670233607,-0.0538140684,-0.2211986184,-0.5511418581,-0.2069526762,-0.3715268075,-0.1109172553,-0.1350990832,-0.0615067855,-0.2741713226,0.1886223704,0.0503452383,-0.330363065,0.0814077035,-0.1955392063,0.2041691542,-0.1523468047,0.2263703197,-0.0093282107,0.0362457894,0.3188800812,0.0760150179,0.5093174577,0.2283807099,-0.0963346064,-0.2487950623,-0.0241237804,-0.142178461,0.040377859,0.3468794227,0.1857326627,0.0395644121,0.2420821488,0.1516417563,-0.1133231595,0.3565611541,0.002253989,0.263456285,-0.0992489904,0.5178786516,-0.467658788,0.1000935584,-0.1672317088,-0.3525443077,-0.1460466087,-0.3188263476,-0.1780048162,-0.1945900023,-0.0235637128,-0.058697395,0.0232757386,-0.0869168863,0.4179295003,0.394969821,0.1819574684,0.1531835794,-0.033079084,-0.2972991467,0.2351235002,-0.1318951845,0.0307580009,0.2157548368,0.2165423781,0.0767683834,0.1726402193,0.0412504598,0.0387746431,0.0111108012,0.2165732384,-0.2433100045,-0.060975045,0.2753140926,0.0173532814,-0.1357738823,-0.2809374332,-0.0888741985,-0.0585748106,-0.0598330982,0.1518500149,-0.0595391653,-0.0654157251,-0.1698684841,0.4425504208,0.1324360222,0.2011087984,-0.0575092882,-0.4760095477,-0.2689436674,-0.4144521952,-0.1308504641,0.3707846403,-0.1804432422,0.6111345291,0.0342233665,0.1484044045,-0.0185195096,-0.0448632054,-0.1875070333,-0.0982181877,-0.5277866125,0.2939006686,-0.3957848847,0.1447350532,-0.1315573007,0.1057318822,0.1239854619,0.1936321557,-0.0772358403,-0.2317139655,0.4982314706,-0.2602384984,-0.3133595288,0.2727927864,0.0889144912,-0.1325899363,-0.2973820269,-0.6104027033,0.0109965969,0.2189054042,0.0305512454,-0.2178259343,-0.1124497503,-0.124261193,-0.05467695,-0.1363347769,0.2223986536,0.0136693334,-0.1138071492,0.1441099644,-0.1869696826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2799","title":"Loading JSON throws ArrowNotImplementedError","comments":"OK after upgrading `datasets` to v1.12.1 the issue seems to have gone away. Closing this now :)","body":"## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":17,"text":"Loading JSON throws ArrowNotImplementedError\n## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nOK after upgrading `datasets` to v1.12.1 the issue seems to have gone away. Closing this now :)","embeddings":[-0.0681871027,0.1841620058,0.0364155807,0.3780699968,0.2355371863,-0.0062702368,0.4846684635,0.3670629561,0.4180777371,-0.0028386964,0.0605546869,0.5898874402,0.0276210085,-0.148857981,-0.2172110528,-0.1838730276,0.0313557796,0.1905159801,0.1274458319,0.0285853427,-0.3122327328,0.2942055166,-0.214816317,0.1017739549,-0.0391265377,-0.0952268541,0.1041623503,0.3933393061,-0.126418829,-0.3826465905,0.4134063721,-0.0130348513,0.2161162943,0.5308527946,-0.0001203599,0.1964653581,0.4152528644,-0.092444241,-0.2868113518,-0.4780595005,-0.2790301442,-0.3253917992,0.537475884,-0.1200379506,0.1052053347,-0.2437469363,-0.0497698858,-0.2207819521,0.6015406847,0.2307226658,0.1048027351,0.429939568,0.0260289963,0.0192190502,0.0637280419,0.2108367532,-0.1217756793,0.4542036653,0.008248603,0.1270525903,-0.1715968549,-0.0036361166,-0.102641046,0.1538670212,0.2191231996,0.0476008654,-0.1704021692,0.0818345174,0.0492499694,0.2946319878,0.2845771015,-0.0508809648,-0.4122377634,-0.3489953876,-0.1207850203,0.1302726865,0.2388218939,0.2473598123,-0.1509169936,0.1259337515,0.0328827612,-0.0184712112,-0.274197191,0.2383805364,0.0036435723,0.0111869918,-0.1582194567,0.10729184,-0.1295856088,-0.020824587,-0.1499335319,-0.2244829088,-0.3057584167,0.1283780038,0.0374987274,0.11600402,-0.0355978459,-0.4100777805,0.2355834544,0.3558476567,0.3206709325,0.2080192715,-0.2673651874,0.2078464776,0.5903261304,0.2016133815,-0.0549013875,0.0114986571,0.2564998567,0.330514431,0.0607299209,-0.0435533673,-0.2637991011,-0.1108330786,0.0103772897,-0.197450906,0.3086808324,-0.1726259738,-0.2238530517,0.0755632967,-0.6134242415,-0.0573209412,-0.0002183132,0.3672920167,-0.1971875429,0.4987778664,0.2744282484,0.4868384302,0.0442172214,0.0317274667,-0.1022381261,-0.0353484266,-0.1634835154,-0.2962282002,0.2371660173,0.0525447652,0.0861053467,0.1265179962,-0.0630014017,-0.4126393497,-0.0742599964,0.0477150232,-0.0974051356,0.16696693,0.1920519024,0.132871151,0.1559260786,-0.4045022726,-0.1376460344,0.0323025733,-0.376919657,0.0055578146,-0.3176852167,0.0546303503,-0.2702908218,-0.0120817386,-0.7623512149,-0.0521201082,-0.1122782081,0.0660267845,0.0638756305,-0.0244168527,0.2227595448,-0.2470848262,0.0749816447,0.28331393,-0.3005584776,-0.2378236353,-0.0629288927,-0.037491519,0.1896335036,0.2650878727,-0.2005772442,0.2845451832,-0.1637352109,0.2106698602,0.3121518195,-0.3327314556,-0.2219264358,0.605520308,0.0193049405,0.5248504877,0.0198842697,-0.2308956832,-0.0092007685,0.1884041429,0.3035949767,0.4013707638,0.1422874182,0.0679152384,-0.2457050234,-0.2022635043,-0.1346425563,0.3127692342,0.0487670116,0.1889399141,0.0995356888,0.0244781338,0.2506872714,-0.1951391548,-0.0266671889,0.1624232382,0.0911331177,0.0282151625,-0.0057259151,-0.0044113393,-0.6142941117,0.1499558836,-0.0511226989,0.109201923,-0.408085674,0.0093207462,-0.2547832429,0.1260678917,-0.1280752718,0.1847681105,0.0374528877,0.1086534187,-0.1821872592,0.2108015567,-0.1838937998,0.3152327836,0.1626450419,0.2816524506,-0.4834640026,0.4377061725,0.210255906,-0.1212381497,0.0271184649,0.1853716373,-0.097817637,-0.1626708806,-0.2391357571,0.1727546602,0.1394181401,0.1797138602,-0.0562768616,-0.2375711501,-0.0822836161,0.0838592127,-0.1951901019,0.3068773448,0.2217241377,-0.0245843939,-0.0280968156,0.8052088618,-0.0741070136,0.4431970716,-0.015723357,-0.0205509756,0.4561716616,0.0615659803,-0.364153266,-0.3053402305,0.2388348728,0.3186245859,0.0603672415,0.0040045432,-0.3794290721,-0.1885213256,0.3982873261,0.005179801,0.2424625754,0.1532346159,-0.0953567475,-0.0996729881,0.0158600174,0.1401785463,0.2382987291,0.0758770034,-0.2070478052,-0.0434467606,-0.0409476832,-0.0247451495,0.1555987,0.306535244,0.2641779482,0.0933471769,0.1558886766,-0.020962242,-0.2815460265,-0.238003552,-0.2863601744,0.1189386174,-0.6182570457,0.2991375625,-0.4743469954,0.1188910827,0.035333585,-0.3410826325,-0.126299873,-0.322971493,-0.3007420599,0.2734274864,-0.1147493422,-0.1341666132,-0.2667239308,0.3151707649,-0.0232129041,-0.3463473022,-0.2113419473,-0.0117184091,-0.184648335,-0.0234055538,0.3875870109,0.1092963666,0.1544730067,-0.1289402992,0.047313679,0.1418192536,-0.0863964856,0.1482000202,-0.1994989067,0.2889936268,0.2154709399,0.1969099492,-0.0978527516,-0.1516460776,0.2886652946,0.0957521275,-0.3427708745,0.3241927326,-0.1354984939,0.0278997961,-0.1492165327,-0.0291694198,0.1219549254,-0.4392543733,0.5247157216,0.2458687872,0.0581384897,-0.0170072187,0.0242614429,0.1799215674,0.1687064171,0.0159113072,-0.1107467487,-0.1546859592,0.2326568216,-0.160832569,-0.2501766384,0.0764547884,-0.1965518147,0.1685908139,-0.0865133032,-0.390976578,0.0259424336,0.0707196668,0.1136565655,-0.0444040187,-0.295239538,0.0588247478,-0.0127509227,0.0851902887,-0.1855715215,-0.2022252828,0.1206316799,-0.0438129231,0.0216035321,0.0565150753,0.7252899408,-0.0942959189,0.1807754487,0.3117997646,-0.0382717438,0.394974649,-0.0264437217,0.1030089855,-0.4057941437,-0.1191259399,-0.3808179796,-0.0971131027,0.0786372274,0.2193027288,0.1096423194,-0.2879481912,-0.2405597121,-0.0622841977,-0.4587052464,-0.3489450216,0.1333795786,-0.3543255031,-0.3382495046,0.0320333503,0.2205641121,-0.0587302446,-0.0581021123,-0.0303620175,0.4014143944,0.2062794119,-0.0455005579,-0.2057112902,0.0343681015,-0.3544535339,0.3606956899,0.0746525601,0.3657507002,0.1974582672,-0.028097624,0.2447486371,-0.2093995959,0.4490728974,-0.120011054,0.0227281768,0.0280953031,-0.1236653179,-0.6418688893,0.1011512503,-0.1271315068,0.2907666266,0.2359819263,0.2095480561,-0.4572590888,-0.2961414456,0.1018688306,-0.1504392922,-0.1058704555,-0.1064765677,-0.3640385568,-0.2923769653,-0.2448295504,-0.1525726318,0.0274700802,0.2387111634,0.0874371156,-0.0987067744,-0.2664907575,-0.0690853894,0.1225850284,0.2164159268,0.1880879253,-0.2173919529,0.0409795344,0.0061119525,0.3643347025,0.3959253132,0.8496598601,0.2561551034,-0.5346414447,0.2092527151,-0.0245482698,0.0420461036,0.1865970343,0.0230351947,0.0835103542,0.2850650549,0.1448794007,-0.2175282389,-0.1765887737,0.4804262519,-0.1738846451,-0.4621180296,-0.346393168,0.4906897545,-0.1168812439,0.0213299748,0.2263398916,0.2678977251,-0.1467050314,0.758928895,0.0028833551,0.7930403352,0.1487369686,0.1980527639,0.7221454978,-0.5317468047,0.400482446,-0.0322305784,-0.0708860084,-0.2383532822,0.5987789035,-0.1820171326,-0.151582703,-0.0095944684,0.192508921,-0.1030690968,0.0225751493,-0.3046782017,-0.0162525531,0.0787486061,0.128597334,-0.1141062975,-0.2810204625,-0.7139284611,-0.0139217712,-0.1206339747,0.0340203084,-0.0035200173,-0.3064055443,-0.2558885217,-0.188852787,-0.3464414477,0.2579855025,0.1630945057,-0.0184752066,0.2230367512,-0.0496139601,0.0946815684,-0.0524797849,0.2925044,-0.0933387354,-0.1754464954,0.187010482,-0.1364040524,-0.0991375446,-0.0924463496,-0.016065022,0.5005759001,-0.0768829286,-0.1589543819,-0.2050989121,-0.2627161145,-0.5380346775,0.2646922767,0.4163651466,-0.3466082215,-0.240976274,-0.393114835,-0.0475282073,0.2723071873,-0.1570670009,0.0498477258,0.2624671161,0.1839359552,-0.0674586892,0.2908512056,-0.3609032929,-0.1087285951,0.3222659528,-0.0353102535,0.0564952418,0.5033448339,0.2862301469,-0.2179954052,-0.1904615164,0.2036407739,0.8105605841,-0.6269248128,0.3065258563,-0.1730780452,-0.0535799339,-0.0472417548,0.3028697371,0.0699545369,-0.0232218802,0.0566213802,-0.6213235855,-0.1062008291,0.4122470915,-0.1543131769,-0.0214453992,0.1351440996,-0.0169346631,-0.2405516505,-0.1720611304,-0.17293562,0.2192620784,-0.1576303691,-0.2045778781,-0.3086437285,0.0632046908,0.5788154602,0.0068058115,0.0865861177,0.0817479491,-0.1436324716,-0.0885987878,-0.3009876311,0.2072960436,0.1043563411,-0.0927105248,-0.0784604624,0.0101217115,-0.1790237874,-0.135748744,0.2507112324,0.2667553127,-0.0477175564,-0.1331558526,0.0813497081,0.1524973363,-0.2447343022,0.1382133961,-0.159083724,0.5401737094,0.0858471692,-0.0476722531,-0.2291520685,0.1443036646,-0.2354615629,0.2881273329,0.2030762136,0.2164718509,0.4846534729,-0.427249223,0.0382829905,-0.2070930153,0.0976076871,0.3163167238,-0.2574578226,0.0504572801,-0.16220209,0.0547889322,0.0131520722,0.0525103472,0.1396277249,-0.0389830656,0.0427073203,0.333689034,-0.1388189495,0.042830985,-0.2334407568,0.0458435602,0.4255111814,0.1589772552,0.051013343,0.2905644774,-0.2684130967,0.0809304491,0.0158537552,0.1407580674,0.2980785966,-0.0071150423,0.0680872425,0.0217248164,0.1491240412,0.2423139662,0.0005278526,-0.4159061313,-0.0420816727,0.3158295155,-0.0500427932,0.3092298806,0.1688649356,0.3333734274,-0.1892764717,0.0969858468,-0.4036626816,-0.1670233607,-0.0538140684,-0.2211986184,-0.5511418581,-0.2069526762,-0.3715268075,-0.1109172553,-0.1350990832,-0.0615067855,-0.2741713226,0.1886223704,0.0503452383,-0.330363065,0.0814077035,-0.1955392063,0.2041691542,-0.1523468047,0.2263703197,-0.0093282107,0.0362457894,0.3188800812,0.0760150179,0.5093174577,0.2283807099,-0.0963346064,-0.2487950623,-0.0241237804,-0.142178461,0.040377859,0.3468794227,0.1857326627,0.0395644121,0.2420821488,0.1516417563,-0.1133231595,0.3565611541,0.002253989,0.263456285,-0.0992489904,0.5178786516,-0.467658788,0.1000935584,-0.1672317088,-0.3525443077,-0.1460466087,-0.3188263476,-0.1780048162,-0.1945900023,-0.0235637128,-0.058697395,0.0232757386,-0.0869168863,0.4179295003,0.394969821,0.1819574684,0.1531835794,-0.033079084,-0.2972991467,0.2351235002,-0.1318951845,0.0307580009,0.2157548368,0.2165423781,0.0767683834,0.1726402193,0.0412504598,0.0387746431,0.0111108012,0.2165732384,-0.2433100045,-0.060975045,0.2753140926,0.0173532814,-0.1357738823,-0.2809374332,-0.0888741985,-0.0585748106,-0.0598330982,0.1518500149,-0.0595391653,-0.0654157251,-0.1698684841,0.4425504208,0.1324360222,0.2011087984,-0.0575092882,-0.4760095477,-0.2689436674,-0.4144521952,-0.1308504641,0.3707846403,-0.1804432422,0.6111345291,0.0342233665,0.1484044045,-0.0185195096,-0.0448632054,-0.1875070333,-0.0982181877,-0.5277866125,0.2939006686,-0.3957848847,0.1447350532,-0.1315573007,0.1057318822,0.1239854619,0.1936321557,-0.0772358403,-0.2317139655,0.4982314706,-0.2602384984,-0.3133595288,0.2727927864,0.0889144912,-0.1325899363,-0.2973820269,-0.6104027033,0.0109965969,0.2189054042,0.0305512454,-0.2178259343,-0.1124497503,-0.124261193,-0.05467695,-0.1363347769,0.2223986536,0.0136693334,-0.1138071492,0.1441099644,-0.1869696826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2799","title":"Loading JSON throws ArrowNotImplementedError","comments":"Oops, I spoke too soon \ud83d\ude13 \r\n\r\nAfter deleting the cache and trying the above code snippet again I am hitting the same error. You can also reproduce it in the Colab notebook I linked to in the issue description. ","body":"## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":39,"text":"Loading JSON throws ArrowNotImplementedError\n## Describe the bug\r\nI have created a [dataset](https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below).\r\n\r\nCuriously, there is no problem loading the dataset with `pandas` which suggests some incorrect type inference is being made on the `datasets` side. For example, the stack trace indicates that some URL fields are being parsed as timestamps.\r\n\r\nYou can find a Colab notebook which reproduces the error [here](https:\/\/colab.research.google.com\/drive\/1YUCM0j1vx5ZrouQbYSzal6RwB4-Aoh4o?usp=sharing).\r\n\r\n**Edit:** If one repeatedly tries to load the dataset, it _eventually_ works but I think it would still be good to understand why it fails in the first place :)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom huggingface_hub import hf_hub_url\r\nimport pandas as pd\r\n\r\n# returns https:\/\/huggingface.co\/datasets\/lewtun\/github-issues-test\/resolve\/main\/issues-datasets.jsonl\r\ndata_files = hf_hub_url(repo_id=\"lewtun\/github-issues-test\", filename=\"issues-datasets.jsonl\", repo_type=\"dataset\")\r\n# throws ArrowNotImplementedError\r\ndset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n# no problem with pandas ...\r\ndf = pd.read_json(data_files, orient=\"records\", lines=True)\r\ndf.head()\r\n```\r\n\r\n## Expected results\r\nI can load any line-separated JSON file, similar to `pandas`.\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n----> 1 dset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: JSON conversion to struct, open_issues: int64, closed_issues: int64, state: timestamp[s], created_at: timestamp[s], updated_at: timestamp[s], due_on: timestamp[s], closed_at: timestamp[s]> is not supported\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nOops, I spoke too soon \ud83d\ude13 \r\n\r\nAfter deleting the cache and trying the above code snippet again I am hitting the same error. You can also reproduce it in the Colab notebook I linked to in the issue description. ","embeddings":[-0.0681871027,0.1841620058,0.0364155807,0.3780699968,0.2355371863,-0.0062702368,0.4846684635,0.3670629561,0.4180777371,-0.0028386964,0.0605546869,0.5898874402,0.0276210085,-0.148857981,-0.2172110528,-0.1838730276,0.0313557796,0.1905159801,0.1274458319,0.0285853427,-0.3122327328,0.2942055166,-0.214816317,0.1017739549,-0.0391265377,-0.0952268541,0.1041623503,0.3933393061,-0.126418829,-0.3826465905,0.4134063721,-0.0130348513,0.2161162943,0.5308527946,-0.0001203599,0.1964653581,0.4152528644,-0.092444241,-0.2868113518,-0.4780595005,-0.2790301442,-0.3253917992,0.537475884,-0.1200379506,0.1052053347,-0.2437469363,-0.0497698858,-0.2207819521,0.6015406847,0.2307226658,0.1048027351,0.429939568,0.0260289963,0.0192190502,0.0637280419,0.2108367532,-0.1217756793,0.4542036653,0.008248603,0.1270525903,-0.1715968549,-0.0036361166,-0.102641046,0.1538670212,0.2191231996,0.0476008654,-0.1704021692,0.0818345174,0.0492499694,0.2946319878,0.2845771015,-0.0508809648,-0.4122377634,-0.3489953876,-0.1207850203,0.1302726865,0.2388218939,0.2473598123,-0.1509169936,0.1259337515,0.0328827612,-0.0184712112,-0.274197191,0.2383805364,0.0036435723,0.0111869918,-0.1582194567,0.10729184,-0.1295856088,-0.020824587,-0.1499335319,-0.2244829088,-0.3057584167,0.1283780038,0.0374987274,0.11600402,-0.0355978459,-0.4100777805,0.2355834544,0.3558476567,0.3206709325,0.2080192715,-0.2673651874,0.2078464776,0.5903261304,0.2016133815,-0.0549013875,0.0114986571,0.2564998567,0.330514431,0.0607299209,-0.0435533673,-0.2637991011,-0.1108330786,0.0103772897,-0.197450906,0.3086808324,-0.1726259738,-0.2238530517,0.0755632967,-0.6134242415,-0.0573209412,-0.0002183132,0.3672920167,-0.1971875429,0.4987778664,0.2744282484,0.4868384302,0.0442172214,0.0317274667,-0.1022381261,-0.0353484266,-0.1634835154,-0.2962282002,0.2371660173,0.0525447652,0.0861053467,0.1265179962,-0.0630014017,-0.4126393497,-0.0742599964,0.0477150232,-0.0974051356,0.16696693,0.1920519024,0.132871151,0.1559260786,-0.4045022726,-0.1376460344,0.0323025733,-0.376919657,0.0055578146,-0.3176852167,0.0546303503,-0.2702908218,-0.0120817386,-0.7623512149,-0.0521201082,-0.1122782081,0.0660267845,0.0638756305,-0.0244168527,0.2227595448,-0.2470848262,0.0749816447,0.28331393,-0.3005584776,-0.2378236353,-0.0629288927,-0.037491519,0.1896335036,0.2650878727,-0.2005772442,0.2845451832,-0.1637352109,0.2106698602,0.3121518195,-0.3327314556,-0.2219264358,0.605520308,0.0193049405,0.5248504877,0.0198842697,-0.2308956832,-0.0092007685,0.1884041429,0.3035949767,0.4013707638,0.1422874182,0.0679152384,-0.2457050234,-0.2022635043,-0.1346425563,0.3127692342,0.0487670116,0.1889399141,0.0995356888,0.0244781338,0.2506872714,-0.1951391548,-0.0266671889,0.1624232382,0.0911331177,0.0282151625,-0.0057259151,-0.0044113393,-0.6142941117,0.1499558836,-0.0511226989,0.109201923,-0.408085674,0.0093207462,-0.2547832429,0.1260678917,-0.1280752718,0.1847681105,0.0374528877,0.1086534187,-0.1821872592,0.2108015567,-0.1838937998,0.3152327836,0.1626450419,0.2816524506,-0.4834640026,0.4377061725,0.210255906,-0.1212381497,0.0271184649,0.1853716373,-0.097817637,-0.1626708806,-0.2391357571,0.1727546602,0.1394181401,0.1797138602,-0.0562768616,-0.2375711501,-0.0822836161,0.0838592127,-0.1951901019,0.3068773448,0.2217241377,-0.0245843939,-0.0280968156,0.8052088618,-0.0741070136,0.4431970716,-0.015723357,-0.0205509756,0.4561716616,0.0615659803,-0.364153266,-0.3053402305,0.2388348728,0.3186245859,0.0603672415,0.0040045432,-0.3794290721,-0.1885213256,0.3982873261,0.005179801,0.2424625754,0.1532346159,-0.0953567475,-0.0996729881,0.0158600174,0.1401785463,0.2382987291,0.0758770034,-0.2070478052,-0.0434467606,-0.0409476832,-0.0247451495,0.1555987,0.306535244,0.2641779482,0.0933471769,0.1558886766,-0.020962242,-0.2815460265,-0.238003552,-0.2863601744,0.1189386174,-0.6182570457,0.2991375625,-0.4743469954,0.1188910827,0.035333585,-0.3410826325,-0.126299873,-0.322971493,-0.3007420599,0.2734274864,-0.1147493422,-0.1341666132,-0.2667239308,0.3151707649,-0.0232129041,-0.3463473022,-0.2113419473,-0.0117184091,-0.184648335,-0.0234055538,0.3875870109,0.1092963666,0.1544730067,-0.1289402992,0.047313679,0.1418192536,-0.0863964856,0.1482000202,-0.1994989067,0.2889936268,0.2154709399,0.1969099492,-0.0978527516,-0.1516460776,0.2886652946,0.0957521275,-0.3427708745,0.3241927326,-0.1354984939,0.0278997961,-0.1492165327,-0.0291694198,0.1219549254,-0.4392543733,0.5247157216,0.2458687872,0.0581384897,-0.0170072187,0.0242614429,0.1799215674,0.1687064171,0.0159113072,-0.1107467487,-0.1546859592,0.2326568216,-0.160832569,-0.2501766384,0.0764547884,-0.1965518147,0.1685908139,-0.0865133032,-0.390976578,0.0259424336,0.0707196668,0.1136565655,-0.0444040187,-0.295239538,0.0588247478,-0.0127509227,0.0851902887,-0.1855715215,-0.2022252828,0.1206316799,-0.0438129231,0.0216035321,0.0565150753,0.7252899408,-0.0942959189,0.1807754487,0.3117997646,-0.0382717438,0.394974649,-0.0264437217,0.1030089855,-0.4057941437,-0.1191259399,-0.3808179796,-0.0971131027,0.0786372274,0.2193027288,0.1096423194,-0.2879481912,-0.2405597121,-0.0622841977,-0.4587052464,-0.3489450216,0.1333795786,-0.3543255031,-0.3382495046,0.0320333503,0.2205641121,-0.0587302446,-0.0581021123,-0.0303620175,0.4014143944,0.2062794119,-0.0455005579,-0.2057112902,0.0343681015,-0.3544535339,0.3606956899,0.0746525601,0.3657507002,0.1974582672,-0.028097624,0.2447486371,-0.2093995959,0.4490728974,-0.120011054,0.0227281768,0.0280953031,-0.1236653179,-0.6418688893,0.1011512503,-0.1271315068,0.2907666266,0.2359819263,0.2095480561,-0.4572590888,-0.2961414456,0.1018688306,-0.1504392922,-0.1058704555,-0.1064765677,-0.3640385568,-0.2923769653,-0.2448295504,-0.1525726318,0.0274700802,0.2387111634,0.0874371156,-0.0987067744,-0.2664907575,-0.0690853894,0.1225850284,0.2164159268,0.1880879253,-0.2173919529,0.0409795344,0.0061119525,0.3643347025,0.3959253132,0.8496598601,0.2561551034,-0.5346414447,0.2092527151,-0.0245482698,0.0420461036,0.1865970343,0.0230351947,0.0835103542,0.2850650549,0.1448794007,-0.2175282389,-0.1765887737,0.4804262519,-0.1738846451,-0.4621180296,-0.346393168,0.4906897545,-0.1168812439,0.0213299748,0.2263398916,0.2678977251,-0.1467050314,0.758928895,0.0028833551,0.7930403352,0.1487369686,0.1980527639,0.7221454978,-0.5317468047,0.400482446,-0.0322305784,-0.0708860084,-0.2383532822,0.5987789035,-0.1820171326,-0.151582703,-0.0095944684,0.192508921,-0.1030690968,0.0225751493,-0.3046782017,-0.0162525531,0.0787486061,0.128597334,-0.1141062975,-0.2810204625,-0.7139284611,-0.0139217712,-0.1206339747,0.0340203084,-0.0035200173,-0.3064055443,-0.2558885217,-0.188852787,-0.3464414477,0.2579855025,0.1630945057,-0.0184752066,0.2230367512,-0.0496139601,0.0946815684,-0.0524797849,0.2925044,-0.0933387354,-0.1754464954,0.187010482,-0.1364040524,-0.0991375446,-0.0924463496,-0.016065022,0.5005759001,-0.0768829286,-0.1589543819,-0.2050989121,-0.2627161145,-0.5380346775,0.2646922767,0.4163651466,-0.3466082215,-0.240976274,-0.393114835,-0.0475282073,0.2723071873,-0.1570670009,0.0498477258,0.2624671161,0.1839359552,-0.0674586892,0.2908512056,-0.3609032929,-0.1087285951,0.3222659528,-0.0353102535,0.0564952418,0.5033448339,0.2862301469,-0.2179954052,-0.1904615164,0.2036407739,0.8105605841,-0.6269248128,0.3065258563,-0.1730780452,-0.0535799339,-0.0472417548,0.3028697371,0.0699545369,-0.0232218802,0.0566213802,-0.6213235855,-0.1062008291,0.4122470915,-0.1543131769,-0.0214453992,0.1351440996,-0.0169346631,-0.2405516505,-0.1720611304,-0.17293562,0.2192620784,-0.1576303691,-0.2045778781,-0.3086437285,0.0632046908,0.5788154602,0.0068058115,0.0865861177,0.0817479491,-0.1436324716,-0.0885987878,-0.3009876311,0.2072960436,0.1043563411,-0.0927105248,-0.0784604624,0.0101217115,-0.1790237874,-0.135748744,0.2507112324,0.2667553127,-0.0477175564,-0.1331558526,0.0813497081,0.1524973363,-0.2447343022,0.1382133961,-0.159083724,0.5401737094,0.0858471692,-0.0476722531,-0.2291520685,0.1443036646,-0.2354615629,0.2881273329,0.2030762136,0.2164718509,0.4846534729,-0.427249223,0.0382829905,-0.2070930153,0.0976076871,0.3163167238,-0.2574578226,0.0504572801,-0.16220209,0.0547889322,0.0131520722,0.0525103472,0.1396277249,-0.0389830656,0.0427073203,0.333689034,-0.1388189495,0.042830985,-0.2334407568,0.0458435602,0.4255111814,0.1589772552,0.051013343,0.2905644774,-0.2684130967,0.0809304491,0.0158537552,0.1407580674,0.2980785966,-0.0071150423,0.0680872425,0.0217248164,0.1491240412,0.2423139662,0.0005278526,-0.4159061313,-0.0420816727,0.3158295155,-0.0500427932,0.3092298806,0.1688649356,0.3333734274,-0.1892764717,0.0969858468,-0.4036626816,-0.1670233607,-0.0538140684,-0.2211986184,-0.5511418581,-0.2069526762,-0.3715268075,-0.1109172553,-0.1350990832,-0.0615067855,-0.2741713226,0.1886223704,0.0503452383,-0.330363065,0.0814077035,-0.1955392063,0.2041691542,-0.1523468047,0.2263703197,-0.0093282107,0.0362457894,0.3188800812,0.0760150179,0.5093174577,0.2283807099,-0.0963346064,-0.2487950623,-0.0241237804,-0.142178461,0.040377859,0.3468794227,0.1857326627,0.0395644121,0.2420821488,0.1516417563,-0.1133231595,0.3565611541,0.002253989,0.263456285,-0.0992489904,0.5178786516,-0.467658788,0.1000935584,-0.1672317088,-0.3525443077,-0.1460466087,-0.3188263476,-0.1780048162,-0.1945900023,-0.0235637128,-0.058697395,0.0232757386,-0.0869168863,0.4179295003,0.394969821,0.1819574684,0.1531835794,-0.033079084,-0.2972991467,0.2351235002,-0.1318951845,0.0307580009,0.2157548368,0.2165423781,0.0767683834,0.1726402193,0.0412504598,0.0387746431,0.0111108012,0.2165732384,-0.2433100045,-0.060975045,0.2753140926,0.0173532814,-0.1357738823,-0.2809374332,-0.0888741985,-0.0585748106,-0.0598330982,0.1518500149,-0.0595391653,-0.0654157251,-0.1698684841,0.4425504208,0.1324360222,0.2011087984,-0.0575092882,-0.4760095477,-0.2689436674,-0.4144521952,-0.1308504641,0.3707846403,-0.1804432422,0.6111345291,0.0342233665,0.1484044045,-0.0185195096,-0.0448632054,-0.1875070333,-0.0982181877,-0.5277866125,0.2939006686,-0.3957848847,0.1447350532,-0.1315573007,0.1057318822,0.1239854619,0.1936321557,-0.0772358403,-0.2317139655,0.4982314706,-0.2602384984,-0.3133595288,0.2727927864,0.0889144912,-0.1325899363,-0.2973820269,-0.6104027033,0.0109965969,0.2189054042,0.0305512454,-0.2178259343,-0.1124497503,-0.124261193,-0.05467695,-0.1363347769,0.2223986536,0.0136693334,-0.1138071492,0.1441099644,-0.1869696826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2788","title":"How to sample every file in a list of files making up a split in a dataset when loading?","comments":"Hi ! This is not possible just with `load_dataset`.\r\n\r\nYou can do something like this instead:\r\n```python\r\nseed=42\r\ndata_files_dict = {\r\n \"train\": [train_file1, train_file2],\r\n \"test\": [test_file1, test_file2],\r\n \"val\": [val_file1, val_file2]\r\n}\r\ndataset = datasets.load_dataset(\r\n \"csv\",\r\n data_files=data_files_dict,\r\n).shuffle(seed=seed)\r\n\r\nsample_dataset = {splitname: split.select(range(8)) for splitname, split in dataset.items()}\r\n```\r\n\r\nAnother alternative is loading each file separately with `split=\"train[:8]\"` and then use `concatenate_datasets` to merge the sample of each file.","body":"I am loading a dataset with multiple train, test, and validation files like this:\r\n\r\n```\r\ndata_files_dict = {\r\n \"train\": [train_file1, train_file2],\r\n \"test\": [test_file1, test_file2],\r\n \"val\": [val_file1, val_file2]\r\n}\r\ndataset = datasets.load_dataset(\r\n \"csv\",\r\n data_files=data_files_dict,\r\n split=['train[:8]', 'test[:8]', 'val[:8]']\r\n)\r\n\r\n```\r\n\r\nHowever, this only selects the first 8 rows from train_file1, test_file1, val_file1, since they are the first files in the lists.\r\n\r\nI'm trying to formulate a split argument that can sample from each file specified in my list of files that make up each split.\r\n\r\nIs this type of splitting supported? If so, how can I do it?","comment_length":67,"text":"How to sample every file in a list of files making up a split in a dataset when loading?\nI am loading a dataset with multiple train, test, and validation files like this:\r\n\r\n```\r\ndata_files_dict = {\r\n \"train\": [train_file1, train_file2],\r\n \"test\": [test_file1, test_file2],\r\n \"val\": [val_file1, val_file2]\r\n}\r\ndataset = datasets.load_dataset(\r\n \"csv\",\r\n data_files=data_files_dict,\r\n split=['train[:8]', 'test[:8]', 'val[:8]']\r\n)\r\n\r\n```\r\n\r\nHowever, this only selects the first 8 rows from train_file1, test_file1, val_file1, since they are the first files in the lists.\r\n\r\nI'm trying to formulate a split argument that can sample from each file specified in my list of files that make up each split.\r\n\r\nIs this type of splitting supported? If so, how can I do it?\nHi ! This is not possible just with `load_dataset`.\r\n\r\nYou can do something like this instead:\r\n```python\r\nseed=42\r\ndata_files_dict = {\r\n \"train\": [train_file1, train_file2],\r\n \"test\": [test_file1, test_file2],\r\n \"val\": [val_file1, val_file2]\r\n}\r\ndataset = datasets.load_dataset(\r\n \"csv\",\r\n data_files=data_files_dict,\r\n).shuffle(seed=seed)\r\n\r\nsample_dataset = {splitname: split.select(range(8)) for splitname, split in dataset.items()}\r\n```\r\n\r\nAnother alternative is loading each file separately with `split=\"train[:8]\"` and then use `concatenate_datasets` to merge the sample of each file.","embeddings":[-0.3138794303,-0.1796855778,-0.1142874882,0.145461306,-0.0157872085,0.3574570119,0.444955796,0.4948764741,0.556786716,0.0195083302,-0.0745141283,0.198577702,0.0281437356,0.1434578001,0.1276390553,-0.3237335086,-0.0534550883,0.1187504977,0.2719002068,0.0101069016,-0.1883502007,0.091860868,-0.4363408387,-0.3013267517,-0.1955500692,0.0335028172,-0.3832268119,0.0586887076,0.0298341662,-0.2322053164,0.0476163216,0.2474765778,0.0852342919,0.011206409,-0.0001181749,0.0020298953,0.1929554939,-0.3049511909,-0.2270892859,-0.3292902708,-0.0461659841,0.0632799491,-0.0771353766,-0.3171573579,-0.1542430818,-0.1601599008,-0.0316538103,-0.138571009,0.4643646777,0.1826563776,0.0203301627,0.1105126143,-0.3269145489,-0.1088883802,0.3746608496,0.3522197604,0.2187461704,0.0444287285,0.4358339012,-0.0470805466,-0.1141447574,-0.193022579,-0.0340736024,0.2639250457,0.1277291179,0.1246679351,-0.3276381791,-0.3322709799,-0.1635397971,0.4533859789,0.2487736046,0.0091191055,-0.4796769917,-0.6428300142,0.0095056836,-0.3940193355,-0.1848472953,0.4530909657,-0.3995965123,0.0952166766,0.1818254739,-0.0572032072,0.0219684429,0.2663127482,0.0569325797,0.1254368424,-0.067655161,0.0810493603,0.7294518948,0.1416414231,-0.0421174206,-0.2076791376,-0.079792425,0.1188273728,-0.4636723995,-0.1048142835,-0.2771571279,-0.0089037791,0.1892953068,0.3294673264,0.0680096671,-0.2414481789,0.1494103819,0.170034349,0.6460336447,-0.2381347865,-0.0347361565,0.1712915152,-0.1695314646,0.0176247302,0.0331937373,-0.0454387814,-0.0381043367,-0.1814244986,-0.3282948732,0.1058130413,-0.1274303645,0.0860395506,-0.1948265135,-0.4890435636,-0.2440893203,-0.5034528971,0.0762869567,0.3599081933,0.2342460155,-0.0838767216,-0.1735524237,0.4428135753,-0.0889260322,-0.274707526,-0.1148847938,0.2223401219,-0.1089815348,0.0057191653,0.1465914994,-0.188432321,0.2893144488,-0.0865546241,0.2269657701,-0.2306074351,0.3035942018,-0.138725698,0.1588564962,0.134075433,0.2950143814,-0.0119690951,0.2660682797,-0.0398498178,-0.1329244077,0.2388462722,-0.2571510375,-0.191285789,0.2006834,-0.0225589946,0.0105375061,0.1674260944,-0.362051636,0.2915106118,0.1189149395,0.1359364688,-0.1135267764,-0.1590327024,0.0445085838,-0.0197431128,0.1557989419,0.3986886144,-0.2346587479,-0.123491399,-0.2077334225,-0.1717362553,0.3626834452,0.0350856408,-0.3634830415,0.241454497,-0.3660298884,0.7993331552,0.3345078826,0.1338744164,-0.0137383081,0.4271205068,0.0220962726,0.0380513445,0.2731009722,0.0632520467,0.5865426064,-0.0518562123,0.0598010197,0.505310595,-0.2618050277,0.0240476914,0.1377211064,-0.0513437763,0.1542183459,0.3551742136,-0.0846145749,0.1678067297,-0.0063042361,0.2048270851,0.5283377767,-0.2432685345,-0.036706645,0.0304862726,-0.0026645798,0.2762807608,0.0030848575,-0.209514752,-0.1535419524,0.1135422587,0.278737247,-0.5721077919,0.3249298036,-0.3869333565,-0.122447297,-0.0005192388,-0.1241058335,0.1367531568,0.0101979896,0.1256721169,-0.1288661659,-0.5308008194,-0.3429537714,-0.0541972294,-0.2381863445,0.1940638423,-0.2353833616,0.3334453106,0.2488534898,0.0130883576,0.1835725605,-0.0925863832,-0.0169581082,-0.3653472364,0.0474871881,0.2036670148,0.8958287835,-0.0555032603,0.1376837194,-0.2449238896,0.0282209273,0.0709381327,-0.1368812919,-0.1341382563,-0.0789420307,-0.1361997575,-0.2074255496,0.5919697881,-0.5605755448,0.4632727206,-0.1452792734,0.2544754744,0.1349827349,-0.0727637112,-0.2149717957,-0.0677865893,0.0911614895,-0.1049719304,0.3877750039,0.062067505,-0.2700069547,-0.1236140803,0.226440832,-0.2351535559,-0.021593364,0.1420031786,0.0250722636,0.0769358501,0.1343010515,0.2210197002,0.2190335542,0.1367043406,0.2197372913,-0.2373443842,0.1761601716,-0.1678719819,0.090482682,0.2804896235,-0.1848636568,0.2034763992,-0.5162175894,-0.2349596769,-0.1231027246,0.0242495108,0.2009834945,-0.0500389449,-0.1504891515,0.0563311279,-0.1548143774,0.0014892309,-0.1733236015,-0.2208731472,0.1084049791,-0.0190758575,-0.2101282328,0.1752094328,-0.014967124,-0.046751678,-0.0501355603,0.3053672314,-0.0454382449,-0.172108233,0.0894083381,-0.1400024593,0.1235911474,0.0462605245,0.5049254298,0.5502218604,0.1374114454,0.0413787812,-0.216308251,0.2009606212,0.3659947217,0.0553424805,-0.1022311673,0.1487470418,0.0951430202,0.393389374,-0.0603435375,-0.3966614306,-0.2915034294,0.4472600818,-0.05621567,-0.1207681373,0.1564860195,0.0250614211,-0.1724932492,-0.2683087289,-0.3343343437,-0.1703346223,0.1064798757,0.1056603864,0.0908488706,-0.0613080114,0.0446206331,-0.1298354268,-0.2531605363,-0.3052235842,-0.0670007989,-0.1719997525,0.028853571,-0.0586904548,-0.1425167173,0.0911052972,-0.1706534475,-0.1698983461,0.3151335418,0.0565422364,-0.119005613,0.0536041893,-0.0231910441,-0.1919582039,-0.0447354652,0.0237636454,-0.2283338606,0.1782976985,-0.1228153259,0.076862894,0.3081637621,-0.1329180747,0.2364634871,0.2309126407,0.0722262487,0.0287414677,0.9346376061,0.4259983301,0.133753702,0.1054857075,-0.0050867302,-0.1113102809,0.0277696364,-0.2091869265,0.0696029216,-0.0040837573,-0.2223213613,0.2278610617,-0.053333424,-0.0510844141,-0.0543309748,0.4830429852,-0.3878196478,-0.4113050103,0.25141868,-0.1339634657,0.1430372745,-0.063463524,0.135205552,0.0034175126,0.0518518314,-0.323864311,0.3723500967,0.6840326786,0.0787975788,-0.3335604966,-0.2034336179,0.1091626957,-0.1188940927,0.2429923266,0.236023739,0.0331168249,-0.1723501384,0.1852572113,0.2139886171,1.0035449266,0.0698846206,-0.0713159665,0.1751419008,-0.1078397557,0.223814711,-0.0718893707,-0.1749577373,0.1892586946,0.0739918724,0.1319837421,-0.4061217904,-0.2338041663,-0.0051983781,0.4461130798,-0.1055282876,-0.0611943975,0.0108930236,0.0180290118,-0.1926851273,0.1312042326,0.1845132858,0.0162767153,-0.1188838109,-0.1092984229,0.0061728503,-0.2024412155,0.1696754843,0.1074824855,0.4021864831,0.119180128,0.2199012786,0.265286088,-0.2765387893,0.2222146094,0.3064244688,-0.3874230385,-0.3323573768,-0.1303471476,-0.5022568107,0.3288090825,0.1313901693,-0.1304757297,0.0695616528,-0.196965009,-0.1734095812,-0.1204662025,0.1301077753,0.28518188,-0.1736651212,-0.1530525833,-0.6710407138,0.3475161195,0.0427141786,-0.0904696956,0.5216256976,-0.340926677,-0.152146548,0.4326778948,-0.0926360115,0.5991636515,-0.170889765,0.0016946049,0.0064230105,0.537494719,0.1572350562,-0.0105162533,0.085755989,-0.2693865597,-0.4592579603,-0.2316503972,-0.2514381111,0.0591211393,0.5377877355,-0.0869729146,0.0531826504,-0.2273864895,-0.1006668061,-0.1539772302,-0.0106278928,-0.0455473699,-0.1159325168,-0.0164215509,-0.0479992293,-0.1417924464,0.1763977408,0.1815107614,-0.1568814218,0.0286266003,-0.0045379074,-0.3593708873,-0.0207994971,-0.0406891517,-0.165133059,0.0170693509,-0.0868792608,0.0052189683,0.2841290832,-0.2209650129,-0.0513923354,0.1355526,0.1300029904,-0.1193240955,-0.0126097109,-0.3487888575,-0.0778232366,0.3172205091,0.1575842202,-0.130166173,0.1060090438,-0.0572708361,-0.4175531268,-0.1477218121,0.2041468918,0.2804667056,-0.2656110525,-0.5376275182,0.1008565947,-0.0632007644,0.0431189463,0.0061470293,0.1896751225,-0.222538799,0.3468295634,-0.0925849602,-0.0056073032,0.0320769362,-0.1889261454,0.0692635104,-0.1685343087,0.3529903591,0.0854190513,-0.0536854751,-0.2866612673,0.0814151391,-0.1341737658,-0.058207836,0.2703179717,-0.0604624599,0.3530196846,0.036713779,0.0682967752,0.215334788,0.1412826329,-0.1821472794,-0.6045979261,0.24487634,0.1488193721,0.0904414505,0.3770522773,-0.0822673291,0.266807735,-0.1099334732,0.2634677887,-0.112606369,0.0930376798,0.2155744284,0.1863121986,0.134630397,0.1324447393,-0.0327640735,0.0740718096,-0.0101237055,0.104290612,0.0286130793,-0.1058967337,-0.0680630133,0.1173505411,0.1163081601,0.1006411538,0.1008267179,0.0226469953,-0.2740134895,-0.3027728796,0.1793282181,-0.2539671659,0.0554066226,0.1510189623,0.768152535,0.0637191832,-0.3797898293,0.0452203304,-0.1893551797,0.1555936784,0.0956037715,0.2181667835,0.1093070656,-0.0047505689,-0.4768045545,-0.0098437723,0.2439012825,-0.1089685336,-0.0905510709,0.0835218579,0.2070000768,0.0270670112,0.4364597201,0.7404636145,0.1301483512,0.4848807752,0.2463205755,0.1492878795,-0.1434832811,0.0682018176,0.1512396783,0.1214119866,0.0020375762,0.2861407399,-0.0152082061,0.0409551524,-0.2007864565,-0.2533937991,0.2422301918,0.2562247813,0.4884402156,0.9639608264,0.072473444,0.0361188389,-0.1169641688,0.1322025359,0.3788408935,0.2384470105,-0.4292243123,0.538477242,-0.004234883,-0.1628546268,-0.0874175876,-0.7342131734,0.2144223452,0.3714231253,0.1754190326,0.1611477882,-0.0401279815,0.7600802779,-0.3678719699,-0.175085336,-0.259591341,0.1456206143,0.1086178571,-0.1780230552,-0.2276837081,-0.0662628934,-0.3687976003,-0.0719542205,0.2731823325,-0.3108788133,-0.0655130595,0.0593590438,0.146046415,-0.1520741582,-0.0059195315,-0.0339517407,-0.2421080023,-0.400865227,-0.2064206749,-0.1082673296,0.2427560091,0.6257146597,-0.3303347826,0.0083575435,0.0491568074,0.0817204043,-0.2312256694,-0.3645603359,0.0110467579,0.0062826145,-0.0571063906,-0.116769284,0.095698379,0.5006061792,0.0493136086,-0.0362400934,0.1601584554,0.104337059,-0.1322453022,-0.4573705494,0.3525461853,0.0718609095,0.016215371,-0.2391441166,-0.29176265,-0.1211856976,-0.0244758744,-0.0647133067,-0.1321182996,0.0504421741,0.4550064802,0.0464036688,-0.1049969718,-0.087212339,-0.1087279767,0.0174958576,-0.1154665574,-0.1533858478,-0.4683508873,-0.4039641619,-0.2181062996,-0.1178192124,0.1383768171,0.2690055668,0.0030818151,0.236174494,-0.1081134602,-0.0170291979,0.189727962,0.4029572308,-0.1841488183,-0.3247696161,-0.147824198,0.318146795,-0.1806169897,-0.2607774138,-0.0513554364,0.0818233266,-0.1259544045,-0.0365122668,-0.1497748792,0.384450376,0.0456246957,-0.1752358526,-0.2018456459,0.4326042533,0.0438907631,0.2329794168,-0.108843796,-0.0967771411,-0.2256961316,-0.0982257947,-0.0252844654,0.5491028428,-0.1545085162,0.0189270228,-0.2223613709,0.0939050242,0.0328006707,-0.2733963728,-0.3613688648,0.0466827713,-0.3458651304,0.1328794956,0.3630694151,0.1326952279,-0.1692213565,-0.0482826158,-0.0207226053,-0.0508955941,0.2685925066,-0.5031941533,-0.0445315354,-0.1722718775,0.0902007148,-0.1271914095,0.0717598349,0.0037254195,-0.0348261446,0.4160357416,-0.1965292245,0.221266076,0.0803876445,0.3423400223,-0.0423950814,-0.1226456165,0.000727763,0.2510319054,-0.3887094855,0.0162322745,-0.2482425869]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2787","title":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com","comments":"the bug code locate in \uff1a\r\n if data_args.task_name is not None:\r\n # Downloading and loading a dataset from the hub.\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)","body":"Hello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n","comment_length":25,"text":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\nHello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n\nthe bug code locate in \uff1a\r\n if data_args.task_name is not None:\r\n # Downloading and loading a dataset from the hub.\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)","embeddings":[-0.2038715184,-0.1601331234,-0.0819259807,0.046902068,0.2665438056,-0.0723191276,0.1546953321,0.3012448549,0.1090682521,-0.0955984294,-0.1769392192,-0.1569310576,0.0424281619,0.0098089986,0.0353741758,-0.1757960469,-0.0955913514,-0.0462971069,-0.2514628172,0.187785551,-0.1379356235,0.1785493493,-0.0379736349,0.2863889635,-0.1777602732,-0.2044938058,-0.0835378692,0.2092444748,-0.1527038217,-0.2665781081,0.4261870086,-0.1104120836,0.1023879945,0.4276525974,-0.0001111302,0.2171017975,0.4489411414,0.0764487162,-0.1543845832,-0.3074429333,-0.0379915237,0.0689299181,0.2134657204,-0.2233344316,-0.1960697025,0.3461775184,0.0491150022,-0.1808769554,0.4765046239,0.2705875933,0.2488420606,0.3377011716,0.4446749985,-0.2152077854,0.0332203731,-0.2912712395,0.0024016653,0.6216793656,0.1213354543,-0.2661054432,0.2426152825,0.2062644213,-0.1433541477,-0.0120363627,0.0456216745,-0.1558198184,0.0829060301,-0.2861295044,0.1528282017,0.0686278045,-0.141816169,-0.2192167342,-0.2591913342,-0.0737397969,-0.0262958892,-0.1296845824,0.331905216,0.1610924304,-0.2213149071,0.2277653813,-0.0923094675,-0.1391131133,-0.2099060416,0.2466940284,0.0454544127,0.2128365934,-0.285688132,0.2644905746,0.4337483943,-0.0378364623,-0.2852620184,0.0696709752,0.1732451916,0.1920874119,-0.2251850963,0.0389768854,-0.3047998548,0.0818184689,0.0858520195,0.0211909339,-0.2492952347,-0.1187235191,0.2668730021,0.1873599738,-0.0882738084,0.3309412599,0.0141430199,0.1095975041,0.3459482193,0.3594096899,-0.1148068979,0.0365825891,-0.2656762898,-0.46562168,-0.3594941199,0.0650140643,0.2611832321,0.0185204595,-0.1696608961,0.0698689297,-0.1784015596,0.0330274105,0.1020842791,0.5010195374,-0.264885366,-0.3661019206,0.0398671776,0.2935795486,0.0180174913,0.0218921565,-0.1782663316,0.157259956,-0.2640621364,0.0905434713,0.0618727505,-0.254320085,0.2353595346,-0.2026710063,0.0230699349,-0.1711889654,0.1205860972,-0.2005208731,-0.158144623,0.3388409019,0.0797365904,0.180475682,0.4291989803,-0.3894393742,-0.0596620627,-0.0774279088,-0.2149029821,-0.4968355596,0.0238060597,0.1501094401,-0.0620120503,0.1604385972,-0.0989380255,-0.248925671,0.2021318972,0.1912706196,-0.0140952608,-0.0100923842,0.1434962451,-0.2804526091,0.5142765641,0.4544944763,0.1708977968,-0.1090477556,-0.0004218963,0.0640206262,-0.0517981537,0.2329597026,-0.0699741095,0.1279715449,-0.3756048977,0.1551721245,0.3715333343,-0.653388679,-0.2841059566,0.3186047971,-0.274140209,-0.0540166125,0.1668800265,-0.2321895212,-0.0450625047,0.0239412375,0.3900935948,0.0740385205,0.0700897276,-0.2052421421,-0.0241824016,-0.3783530891,0.1280156225,0.0518320054,0.1179565564,0.0941041037,0.0289575048,-0.2993083894,0.2291134,0.1804360002,0.0232359786,0.22256051,0.0875867456,0.1844524741,-0.0589135289,-0.1022363752,-0.0982042775,0.1569275558,-0.13933146,0.3022058308,-0.4215773046,0.0652287006,-0.3841099143,0.0516632423,-0.2309429497,-0.3040854037,0.1636356562,-0.0034321281,0.4423214495,0.1496356577,-0.0936227217,0.4874028862,-0.1470745057,0.2835756838,-0.3597308099,0.2300837487,-0.3057549596,-0.1062674224,-0.156371057,0.166188091,0.351362288,-0.2343126982,-0.0265929922,0.4231664836,-0.2641439438,0.0973605886,0.2657681406,0.0963225588,0.031523183,-0.2190815657,0.0253635254,-0.1022461355,-0.0166736674,0.0638581291,0.6379241347,0.0479976051,0.1513416469,0.7245124578,0.1906802058,0.2567310035,0.1417647898,-0.0172172338,-0.2446911037,0.0072813323,0.2052772939,0.0509162173,0.4524137974,-0.3265386522,0.1677902639,-0.2114031017,0.1441133022,-0.035479892,0.2344904095,0.1491603702,-0.3617250323,0.2157289535,-0.1119084656,0.3587546647,0.3501404822,0.0897394866,-0.0167159215,0.2718507648,-0.1022959575,-0.0745456368,0.3431275487,0.0640020669,0.0710157901,-0.1801293045,-0.0368262008,-0.1345124245,-0.1243019998,-0.2394788861,-0.1977582723,0.1469757408,-0.2880876958,0.1693797261,-0.1144179553,-0.3013064861,-0.1543430984,-0.3460415006,-0.4186742306,-0.2409940213,-0.3461467028,0.0472732708,0.0670877323,0.2756136656,0.1089546755,0.0105085308,-0.0052667689,-0.1842108667,-0.034712214,-0.1745305657,-0.1391935796,0.0206357259,-0.0188447759,-0.2574251294,0.2771224976,-0.1947549433,-0.1127352193,0.0472366884,-0.2349187285,0.0479723103,-0.0904428661,0.1711955369,0.328553766,0.3805794716,0.1841480136,-0.3997780979,0.5727685094,0.0603625476,0.1144243404,-0.1596369445,0.1083576158,0.0168828517,0.1767283529,0.0041359356,-0.6324492693,-0.3507916033,0.3485687375,-0.0266018584,0.1511572897,0.1762987822,0.0596724488,0.2824338078,-0.3841683567,0.0531042032,-0.2631133497,-0.4323770106,0.1425144523,-0.2498616874,-0.220136717,-0.0182346441,0.1905058622,0.2702789605,-0.2122463733,-0.3237496614,-0.3732507825,-0.2246032059,-0.0373766795,0.1037429646,0.1593064964,0.2925694287,-0.1132495999,-0.0387336724,0.1269527972,-0.2187675238,0.2200421691,-0.197969988,0.1281968504,0.2286254913,0.3870925903,0.0413929857,0.5444860458,0.2436498553,-0.0796830058,0.3192813396,-0.1745523363,0.2922652662,-0.037890166,-0.5508058667,-0.1344224811,-0.1521711648,0.0345692784,0.0439348668,0.1621201336,0.1705301851,-0.5553136468,-0.2198973298,-0.191320166,-0.2798695266,-0.0953921154,0.0088962717,0.175069645,0.0263881776,-0.1763007343,-0.2223936319,0.1937036365,0.2320328504,0.5205129981,-0.0382330604,0.0609751306,-0.1520732641,-0.3924021125,-0.4625766575,0.4786426425,0.0343587212,0.1926789582,-0.2925172746,-0.0163767822,-0.0434862077,0.1060830727,0.7086897492,-0.0712133721,0.2219744027,0.0485518277,-0.0863394961,-0.3645547032,-0.0629987493,0.0806383789,0.2895561159,0.5007997751,0.4849519432,-0.3278111815,0.096823208,0.2301485389,-0.0766402856,-0.0308216345,-0.0739789009,-0.2963504195,-0.3656420708,-0.267806977,0.0053085051,0.2438388169,0.2123910487,0.3308703899,-0.1466614455,0.1623217314,-0.2427598834,-0.059580896,-0.0181475859,0.0382346138,-0.201649636,-0.0326519087,0.1781357527,0.3460509479,-0.3148242235,0.4281020164,-0.2402162999,-0.2235836387,0.1684294194,0.3209204674,0.4058360159,0.3009501994,-0.1991278529,0.1601253003,0.2677814364,0.3612563014,-0.1262014061,0.1613375992,0.4501324892,0.3550080657,-0.1951705515,-0.2329108715,0.2528257668,-0.0729128048,0.0030149017,0.2449313849,0.4155495167,-0.2275766283,0.165534988,-0.2975753248,1.1979163885,0.0573303141,-0.0391714573,0.1336840391,-0.237947315,0.5808716416,-0.156599924,0.1131992862,-0.1766391993,-0.0917727649,0.1218969896,-0.2235735059,0.3505120277,-0.1980931312,-0.0589326397,0.2558603883,-0.1141535416,0.3006235659,0.2014454603,0.2694348395,-0.0904738307,0.0046210554,-0.257040292,0.1390346289,-0.0044752457,0.4249934554,-0.0580967255,-0.1384331137,0.0146737667,-0.3370604515,-0.3443261981,0.1544762552,-0.7144252062,0.1664410383,0.0121583911,-0.082237266,-0.1178105026,-0.0211141706,0.3912243843,0.1099306643,-0.219323352,0.1842373759,-0.0037481382,-0.3129947782,0.1039159521,0.1227223426,-0.0625953376,-0.269305557,-0.2032667845,0.141741544,-0.1730962843,-0.422695756,0.2038718164,-0.1461987346,0.1515293568,-0.1836434305,-0.0226265304,-0.1612545848,0.0242267586,-0.1469756216,0.112170592,-0.0693805888,-0.2479232848,-0.1085977331,0.0223695394,-0.3606722057,0.0174629949,0.4281620383,-0.144581303,0.0723203868,0.4555252492,0.1247401312,-0.131731987,-0.1284465343,-0.2199524939,0.0430559218,-0.4091939628,-0.0014768161,-0.0063629537,-0.1436571628,-0.230215624,-0.0912814885,0.3861811757,0.2269382328,-0.0249439478,-0.5732638836,-0.1236653402,0.1283735037,-0.1113202199,-0.0436672159,0.1352208406,0.2921498716,0.0813923702,0.0772792622,-0.2949266732,-0.0582201816,-0.1980380267,-0.0909823924,0.4604614675,-0.4231759012,0.5564394593,0.2472602725,0.18336308,0.0185599104,-0.1767931283,-0.1913488358,-0.0861806348,0.123901844,-0.104550086,-0.2903296351,0.0198584348,-0.3376125097,0.0041401763,-0.0067034117,0.0077222949,0.0418777503,0.0665356293,-0.2817328274,-0.0320656784,-0.0895531774,-0.1053852886,0.1841530949,-0.1359493136,0.178451702,0.0287575107,-0.0061410554,-0.2754328847,-0.1757724285,-0.2360019982,-0.0797953382,0.0297373626,0.1081585288,0.3297698498,-0.1942861974,-0.2923170328,-0.1087816954,0.4173107445,-0.0037687309,-0.1419619322,0.059019573,0.1690811962,0.1616690904,-0.1311531216,0.0745707229,0.1629071534,-0.1565767974,0.1184273064,0.2068682462,-0.0237029567,0.1638732702,0.0392775722,0.0193122122,0.8280630112,0.2137528658,0.084507376,0.1391806453,0.0415813588,-0.0628211871,0.1529559493,0.0035156396,0.2465283573,-0.0891733319,-0.3549365401,-0.0614763871,-0.0593232363,-0.0681368411,0.2297047675,-0.0699740872,0.0251145214,0.1838400215,-0.0988485664,-0.1319564432,-0.4080098569,0.6134052873,-0.4409067929,-0.0357085466,-0.1996202171,0.3630471528,0.0331471078,-0.1725615412,0.1452637911,-0.1420483887,-0.2319661677,-0.3037571013,0.0896623954,-0.3973191082,0.3674946129,-0.0679671764,0.0307095051,-0.0729464665,-0.3076351881,0.1328382194,0.2419953793,-0.2333100587,0.3818818331,0.2886329293,-0.0376283564,-0.1605471075,0.4000950456,0.6762163043,0.2699618042,-0.2351969033,0.2674933672,0.0332841128,-0.052856721,-0.0259344038,-0.005995648,0.2812909782,-0.0034649612,0.2339333296,0.233676374,-0.2260972857,0.0528747365,-0.1916484237,0.3153737783,-0.0148682706,0.4126878381,-0.3227872252,-0.1426360309,-0.3223921061,0.1364282221,-0.8104321361,0.1720464826,0.2672681212,-0.4068028927,0.0215156246,-0.1731269211,0.080315344,-0.0752660111,0.7511890531,0.2013044804,0.3521663547,-0.4966924787,-0.3168994188,-0.2904440463,0.2075130939,-0.1204550713,-0.1377062201,-0.0594685376,-0.0786989182,-0.1366242021,0.1496257037,0.1224855855,0.1239469945,-0.2708899975,-0.1634471714,-0.0168661717,0.0420122892,-0.0634030178,0.1023459509,0.27331236,-0.2260944992,0.2341933399,-0.0447947569,0.0383970998,0.2165072858,-0.0195575412,0.1787251234,-0.3686765432,0.3712461293,0.1257679909,0.347453624,0.1640767455,-0.0576720797,-0.1508484334,-0.2002634406,-0.0525047928,0.1181119606,-0.0094588045,0.0054064239,-0.2345675826,-0.2379542887,-0.5355015397,0.1884404123,0.0324563906,-0.0615439638,0.0410017781,-0.1977436841,0.1453060806,0.1729564071,0.2607688904,0.0437686816,-0.0029306645,0.1317536086,-0.3068507314,-0.266634196,0.6317306161,-0.4022055566,-0.2421782613,0.0718518123,0.36091277,0.1383718401,0.1080266684,-0.4578807056,0.0367780551,0.3818689883,0.3051508665,-0.2807103693,0.2636131942,-0.2206344306,0.0750978142,0.1570173502,0.1133459508,-0.1142306253,-0.3479038477,-0.0901176855,-0.1272863001]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2787","title":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com","comments":"Hi @jinec,\r\n\r\nFrom time to time we get this kind of `ConnectionError` coming from the github.com website: https:\/\/raw.githubusercontent.com\r\n\r\nNormally, it should work if you wait a little and then retry.\r\n\r\nCould you please confirm if the problem persists?","body":"Hello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n","comment_length":38,"text":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\nHello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n\nHi @jinec,\r\n\r\nFrom time to time we get this kind of `ConnectionError` coming from the github.com website: https:\/\/raw.githubusercontent.com\r\n\r\nNormally, it should work if you wait a little and then retry.\r\n\r\nCould you please confirm if the problem persists?","embeddings":[-0.2038715184,-0.1601331234,-0.0819259807,0.046902068,0.2665438056,-0.0723191276,0.1546953321,0.3012448549,0.1090682521,-0.0955984294,-0.1769392192,-0.1569310576,0.0424281619,0.0098089986,0.0353741758,-0.1757960469,-0.0955913514,-0.0462971069,-0.2514628172,0.187785551,-0.1379356235,0.1785493493,-0.0379736349,0.2863889635,-0.1777602732,-0.2044938058,-0.0835378692,0.2092444748,-0.1527038217,-0.2665781081,0.4261870086,-0.1104120836,0.1023879945,0.4276525974,-0.0001111302,0.2171017975,0.4489411414,0.0764487162,-0.1543845832,-0.3074429333,-0.0379915237,0.0689299181,0.2134657204,-0.2233344316,-0.1960697025,0.3461775184,0.0491150022,-0.1808769554,0.4765046239,0.2705875933,0.2488420606,0.3377011716,0.4446749985,-0.2152077854,0.0332203731,-0.2912712395,0.0024016653,0.6216793656,0.1213354543,-0.2661054432,0.2426152825,0.2062644213,-0.1433541477,-0.0120363627,0.0456216745,-0.1558198184,0.0829060301,-0.2861295044,0.1528282017,0.0686278045,-0.141816169,-0.2192167342,-0.2591913342,-0.0737397969,-0.0262958892,-0.1296845824,0.331905216,0.1610924304,-0.2213149071,0.2277653813,-0.0923094675,-0.1391131133,-0.2099060416,0.2466940284,0.0454544127,0.2128365934,-0.285688132,0.2644905746,0.4337483943,-0.0378364623,-0.2852620184,0.0696709752,0.1732451916,0.1920874119,-0.2251850963,0.0389768854,-0.3047998548,0.0818184689,0.0858520195,0.0211909339,-0.2492952347,-0.1187235191,0.2668730021,0.1873599738,-0.0882738084,0.3309412599,0.0141430199,0.1095975041,0.3459482193,0.3594096899,-0.1148068979,0.0365825891,-0.2656762898,-0.46562168,-0.3594941199,0.0650140643,0.2611832321,0.0185204595,-0.1696608961,0.0698689297,-0.1784015596,0.0330274105,0.1020842791,0.5010195374,-0.264885366,-0.3661019206,0.0398671776,0.2935795486,0.0180174913,0.0218921565,-0.1782663316,0.157259956,-0.2640621364,0.0905434713,0.0618727505,-0.254320085,0.2353595346,-0.2026710063,0.0230699349,-0.1711889654,0.1205860972,-0.2005208731,-0.158144623,0.3388409019,0.0797365904,0.180475682,0.4291989803,-0.3894393742,-0.0596620627,-0.0774279088,-0.2149029821,-0.4968355596,0.0238060597,0.1501094401,-0.0620120503,0.1604385972,-0.0989380255,-0.248925671,0.2021318972,0.1912706196,-0.0140952608,-0.0100923842,0.1434962451,-0.2804526091,0.5142765641,0.4544944763,0.1708977968,-0.1090477556,-0.0004218963,0.0640206262,-0.0517981537,0.2329597026,-0.0699741095,0.1279715449,-0.3756048977,0.1551721245,0.3715333343,-0.653388679,-0.2841059566,0.3186047971,-0.274140209,-0.0540166125,0.1668800265,-0.2321895212,-0.0450625047,0.0239412375,0.3900935948,0.0740385205,0.0700897276,-0.2052421421,-0.0241824016,-0.3783530891,0.1280156225,0.0518320054,0.1179565564,0.0941041037,0.0289575048,-0.2993083894,0.2291134,0.1804360002,0.0232359786,0.22256051,0.0875867456,0.1844524741,-0.0589135289,-0.1022363752,-0.0982042775,0.1569275558,-0.13933146,0.3022058308,-0.4215773046,0.0652287006,-0.3841099143,0.0516632423,-0.2309429497,-0.3040854037,0.1636356562,-0.0034321281,0.4423214495,0.1496356577,-0.0936227217,0.4874028862,-0.1470745057,0.2835756838,-0.3597308099,0.2300837487,-0.3057549596,-0.1062674224,-0.156371057,0.166188091,0.351362288,-0.2343126982,-0.0265929922,0.4231664836,-0.2641439438,0.0973605886,0.2657681406,0.0963225588,0.031523183,-0.2190815657,0.0253635254,-0.1022461355,-0.0166736674,0.0638581291,0.6379241347,0.0479976051,0.1513416469,0.7245124578,0.1906802058,0.2567310035,0.1417647898,-0.0172172338,-0.2446911037,0.0072813323,0.2052772939,0.0509162173,0.4524137974,-0.3265386522,0.1677902639,-0.2114031017,0.1441133022,-0.035479892,0.2344904095,0.1491603702,-0.3617250323,0.2157289535,-0.1119084656,0.3587546647,0.3501404822,0.0897394866,-0.0167159215,0.2718507648,-0.1022959575,-0.0745456368,0.3431275487,0.0640020669,0.0710157901,-0.1801293045,-0.0368262008,-0.1345124245,-0.1243019998,-0.2394788861,-0.1977582723,0.1469757408,-0.2880876958,0.1693797261,-0.1144179553,-0.3013064861,-0.1543430984,-0.3460415006,-0.4186742306,-0.2409940213,-0.3461467028,0.0472732708,0.0670877323,0.2756136656,0.1089546755,0.0105085308,-0.0052667689,-0.1842108667,-0.034712214,-0.1745305657,-0.1391935796,0.0206357259,-0.0188447759,-0.2574251294,0.2771224976,-0.1947549433,-0.1127352193,0.0472366884,-0.2349187285,0.0479723103,-0.0904428661,0.1711955369,0.328553766,0.3805794716,0.1841480136,-0.3997780979,0.5727685094,0.0603625476,0.1144243404,-0.1596369445,0.1083576158,0.0168828517,0.1767283529,0.0041359356,-0.6324492693,-0.3507916033,0.3485687375,-0.0266018584,0.1511572897,0.1762987822,0.0596724488,0.2824338078,-0.3841683567,0.0531042032,-0.2631133497,-0.4323770106,0.1425144523,-0.2498616874,-0.220136717,-0.0182346441,0.1905058622,0.2702789605,-0.2122463733,-0.3237496614,-0.3732507825,-0.2246032059,-0.0373766795,0.1037429646,0.1593064964,0.2925694287,-0.1132495999,-0.0387336724,0.1269527972,-0.2187675238,0.2200421691,-0.197969988,0.1281968504,0.2286254913,0.3870925903,0.0413929857,0.5444860458,0.2436498553,-0.0796830058,0.3192813396,-0.1745523363,0.2922652662,-0.037890166,-0.5508058667,-0.1344224811,-0.1521711648,0.0345692784,0.0439348668,0.1621201336,0.1705301851,-0.5553136468,-0.2198973298,-0.191320166,-0.2798695266,-0.0953921154,0.0088962717,0.175069645,0.0263881776,-0.1763007343,-0.2223936319,0.1937036365,0.2320328504,0.5205129981,-0.0382330604,0.0609751306,-0.1520732641,-0.3924021125,-0.4625766575,0.4786426425,0.0343587212,0.1926789582,-0.2925172746,-0.0163767822,-0.0434862077,0.1060830727,0.7086897492,-0.0712133721,0.2219744027,0.0485518277,-0.0863394961,-0.3645547032,-0.0629987493,0.0806383789,0.2895561159,0.5007997751,0.4849519432,-0.3278111815,0.096823208,0.2301485389,-0.0766402856,-0.0308216345,-0.0739789009,-0.2963504195,-0.3656420708,-0.267806977,0.0053085051,0.2438388169,0.2123910487,0.3308703899,-0.1466614455,0.1623217314,-0.2427598834,-0.059580896,-0.0181475859,0.0382346138,-0.201649636,-0.0326519087,0.1781357527,0.3460509479,-0.3148242235,0.4281020164,-0.2402162999,-0.2235836387,0.1684294194,0.3209204674,0.4058360159,0.3009501994,-0.1991278529,0.1601253003,0.2677814364,0.3612563014,-0.1262014061,0.1613375992,0.4501324892,0.3550080657,-0.1951705515,-0.2329108715,0.2528257668,-0.0729128048,0.0030149017,0.2449313849,0.4155495167,-0.2275766283,0.165534988,-0.2975753248,1.1979163885,0.0573303141,-0.0391714573,0.1336840391,-0.237947315,0.5808716416,-0.156599924,0.1131992862,-0.1766391993,-0.0917727649,0.1218969896,-0.2235735059,0.3505120277,-0.1980931312,-0.0589326397,0.2558603883,-0.1141535416,0.3006235659,0.2014454603,0.2694348395,-0.0904738307,0.0046210554,-0.257040292,0.1390346289,-0.0044752457,0.4249934554,-0.0580967255,-0.1384331137,0.0146737667,-0.3370604515,-0.3443261981,0.1544762552,-0.7144252062,0.1664410383,0.0121583911,-0.082237266,-0.1178105026,-0.0211141706,0.3912243843,0.1099306643,-0.219323352,0.1842373759,-0.0037481382,-0.3129947782,0.1039159521,0.1227223426,-0.0625953376,-0.269305557,-0.2032667845,0.141741544,-0.1730962843,-0.422695756,0.2038718164,-0.1461987346,0.1515293568,-0.1836434305,-0.0226265304,-0.1612545848,0.0242267586,-0.1469756216,0.112170592,-0.0693805888,-0.2479232848,-0.1085977331,0.0223695394,-0.3606722057,0.0174629949,0.4281620383,-0.144581303,0.0723203868,0.4555252492,0.1247401312,-0.131731987,-0.1284465343,-0.2199524939,0.0430559218,-0.4091939628,-0.0014768161,-0.0063629537,-0.1436571628,-0.230215624,-0.0912814885,0.3861811757,0.2269382328,-0.0249439478,-0.5732638836,-0.1236653402,0.1283735037,-0.1113202199,-0.0436672159,0.1352208406,0.2921498716,0.0813923702,0.0772792622,-0.2949266732,-0.0582201816,-0.1980380267,-0.0909823924,0.4604614675,-0.4231759012,0.5564394593,0.2472602725,0.18336308,0.0185599104,-0.1767931283,-0.1913488358,-0.0861806348,0.123901844,-0.104550086,-0.2903296351,0.0198584348,-0.3376125097,0.0041401763,-0.0067034117,0.0077222949,0.0418777503,0.0665356293,-0.2817328274,-0.0320656784,-0.0895531774,-0.1053852886,0.1841530949,-0.1359493136,0.178451702,0.0287575107,-0.0061410554,-0.2754328847,-0.1757724285,-0.2360019982,-0.0797953382,0.0297373626,0.1081585288,0.3297698498,-0.1942861974,-0.2923170328,-0.1087816954,0.4173107445,-0.0037687309,-0.1419619322,0.059019573,0.1690811962,0.1616690904,-0.1311531216,0.0745707229,0.1629071534,-0.1565767974,0.1184273064,0.2068682462,-0.0237029567,0.1638732702,0.0392775722,0.0193122122,0.8280630112,0.2137528658,0.084507376,0.1391806453,0.0415813588,-0.0628211871,0.1529559493,0.0035156396,0.2465283573,-0.0891733319,-0.3549365401,-0.0614763871,-0.0593232363,-0.0681368411,0.2297047675,-0.0699740872,0.0251145214,0.1838400215,-0.0988485664,-0.1319564432,-0.4080098569,0.6134052873,-0.4409067929,-0.0357085466,-0.1996202171,0.3630471528,0.0331471078,-0.1725615412,0.1452637911,-0.1420483887,-0.2319661677,-0.3037571013,0.0896623954,-0.3973191082,0.3674946129,-0.0679671764,0.0307095051,-0.0729464665,-0.3076351881,0.1328382194,0.2419953793,-0.2333100587,0.3818818331,0.2886329293,-0.0376283564,-0.1605471075,0.4000950456,0.6762163043,0.2699618042,-0.2351969033,0.2674933672,0.0332841128,-0.052856721,-0.0259344038,-0.005995648,0.2812909782,-0.0034649612,0.2339333296,0.233676374,-0.2260972857,0.0528747365,-0.1916484237,0.3153737783,-0.0148682706,0.4126878381,-0.3227872252,-0.1426360309,-0.3223921061,0.1364282221,-0.8104321361,0.1720464826,0.2672681212,-0.4068028927,0.0215156246,-0.1731269211,0.080315344,-0.0752660111,0.7511890531,0.2013044804,0.3521663547,-0.4966924787,-0.3168994188,-0.2904440463,0.2075130939,-0.1204550713,-0.1377062201,-0.0594685376,-0.0786989182,-0.1366242021,0.1496257037,0.1224855855,0.1239469945,-0.2708899975,-0.1634471714,-0.0168661717,0.0420122892,-0.0634030178,0.1023459509,0.27331236,-0.2260944992,0.2341933399,-0.0447947569,0.0383970998,0.2165072858,-0.0195575412,0.1787251234,-0.3686765432,0.3712461293,0.1257679909,0.347453624,0.1640767455,-0.0576720797,-0.1508484334,-0.2002634406,-0.0525047928,0.1181119606,-0.0094588045,0.0054064239,-0.2345675826,-0.2379542887,-0.5355015397,0.1884404123,0.0324563906,-0.0615439638,0.0410017781,-0.1977436841,0.1453060806,0.1729564071,0.2607688904,0.0437686816,-0.0029306645,0.1317536086,-0.3068507314,-0.266634196,0.6317306161,-0.4022055566,-0.2421782613,0.0718518123,0.36091277,0.1383718401,0.1080266684,-0.4578807056,0.0367780551,0.3818689883,0.3051508665,-0.2807103693,0.2636131942,-0.2206344306,0.0750978142,0.1570173502,0.1133459508,-0.1142306253,-0.3479038477,-0.0901176855,-0.1272863001]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2787","title":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com","comments":"cannot connect\uff0ceven by Web browser\uff0cplease check that there is some problems\u3002","body":"Hello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n","comment_length":11,"text":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\nHello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n\ncannot connect\uff0ceven by Web browser\uff0cplease check that there is some problems\u3002","embeddings":[-0.2038715184,-0.1601331234,-0.0819259807,0.046902068,0.2665438056,-0.0723191276,0.1546953321,0.3012448549,0.1090682521,-0.0955984294,-0.1769392192,-0.1569310576,0.0424281619,0.0098089986,0.0353741758,-0.1757960469,-0.0955913514,-0.0462971069,-0.2514628172,0.187785551,-0.1379356235,0.1785493493,-0.0379736349,0.2863889635,-0.1777602732,-0.2044938058,-0.0835378692,0.2092444748,-0.1527038217,-0.2665781081,0.4261870086,-0.1104120836,0.1023879945,0.4276525974,-0.0001111302,0.2171017975,0.4489411414,0.0764487162,-0.1543845832,-0.3074429333,-0.0379915237,0.0689299181,0.2134657204,-0.2233344316,-0.1960697025,0.3461775184,0.0491150022,-0.1808769554,0.4765046239,0.2705875933,0.2488420606,0.3377011716,0.4446749985,-0.2152077854,0.0332203731,-0.2912712395,0.0024016653,0.6216793656,0.1213354543,-0.2661054432,0.2426152825,0.2062644213,-0.1433541477,-0.0120363627,0.0456216745,-0.1558198184,0.0829060301,-0.2861295044,0.1528282017,0.0686278045,-0.141816169,-0.2192167342,-0.2591913342,-0.0737397969,-0.0262958892,-0.1296845824,0.331905216,0.1610924304,-0.2213149071,0.2277653813,-0.0923094675,-0.1391131133,-0.2099060416,0.2466940284,0.0454544127,0.2128365934,-0.285688132,0.2644905746,0.4337483943,-0.0378364623,-0.2852620184,0.0696709752,0.1732451916,0.1920874119,-0.2251850963,0.0389768854,-0.3047998548,0.0818184689,0.0858520195,0.0211909339,-0.2492952347,-0.1187235191,0.2668730021,0.1873599738,-0.0882738084,0.3309412599,0.0141430199,0.1095975041,0.3459482193,0.3594096899,-0.1148068979,0.0365825891,-0.2656762898,-0.46562168,-0.3594941199,0.0650140643,0.2611832321,0.0185204595,-0.1696608961,0.0698689297,-0.1784015596,0.0330274105,0.1020842791,0.5010195374,-0.264885366,-0.3661019206,0.0398671776,0.2935795486,0.0180174913,0.0218921565,-0.1782663316,0.157259956,-0.2640621364,0.0905434713,0.0618727505,-0.254320085,0.2353595346,-0.2026710063,0.0230699349,-0.1711889654,0.1205860972,-0.2005208731,-0.158144623,0.3388409019,0.0797365904,0.180475682,0.4291989803,-0.3894393742,-0.0596620627,-0.0774279088,-0.2149029821,-0.4968355596,0.0238060597,0.1501094401,-0.0620120503,0.1604385972,-0.0989380255,-0.248925671,0.2021318972,0.1912706196,-0.0140952608,-0.0100923842,0.1434962451,-0.2804526091,0.5142765641,0.4544944763,0.1708977968,-0.1090477556,-0.0004218963,0.0640206262,-0.0517981537,0.2329597026,-0.0699741095,0.1279715449,-0.3756048977,0.1551721245,0.3715333343,-0.653388679,-0.2841059566,0.3186047971,-0.274140209,-0.0540166125,0.1668800265,-0.2321895212,-0.0450625047,0.0239412375,0.3900935948,0.0740385205,0.0700897276,-0.2052421421,-0.0241824016,-0.3783530891,0.1280156225,0.0518320054,0.1179565564,0.0941041037,0.0289575048,-0.2993083894,0.2291134,0.1804360002,0.0232359786,0.22256051,0.0875867456,0.1844524741,-0.0589135289,-0.1022363752,-0.0982042775,0.1569275558,-0.13933146,0.3022058308,-0.4215773046,0.0652287006,-0.3841099143,0.0516632423,-0.2309429497,-0.3040854037,0.1636356562,-0.0034321281,0.4423214495,0.1496356577,-0.0936227217,0.4874028862,-0.1470745057,0.2835756838,-0.3597308099,0.2300837487,-0.3057549596,-0.1062674224,-0.156371057,0.166188091,0.351362288,-0.2343126982,-0.0265929922,0.4231664836,-0.2641439438,0.0973605886,0.2657681406,0.0963225588,0.031523183,-0.2190815657,0.0253635254,-0.1022461355,-0.0166736674,0.0638581291,0.6379241347,0.0479976051,0.1513416469,0.7245124578,0.1906802058,0.2567310035,0.1417647898,-0.0172172338,-0.2446911037,0.0072813323,0.2052772939,0.0509162173,0.4524137974,-0.3265386522,0.1677902639,-0.2114031017,0.1441133022,-0.035479892,0.2344904095,0.1491603702,-0.3617250323,0.2157289535,-0.1119084656,0.3587546647,0.3501404822,0.0897394866,-0.0167159215,0.2718507648,-0.1022959575,-0.0745456368,0.3431275487,0.0640020669,0.0710157901,-0.1801293045,-0.0368262008,-0.1345124245,-0.1243019998,-0.2394788861,-0.1977582723,0.1469757408,-0.2880876958,0.1693797261,-0.1144179553,-0.3013064861,-0.1543430984,-0.3460415006,-0.4186742306,-0.2409940213,-0.3461467028,0.0472732708,0.0670877323,0.2756136656,0.1089546755,0.0105085308,-0.0052667689,-0.1842108667,-0.034712214,-0.1745305657,-0.1391935796,0.0206357259,-0.0188447759,-0.2574251294,0.2771224976,-0.1947549433,-0.1127352193,0.0472366884,-0.2349187285,0.0479723103,-0.0904428661,0.1711955369,0.328553766,0.3805794716,0.1841480136,-0.3997780979,0.5727685094,0.0603625476,0.1144243404,-0.1596369445,0.1083576158,0.0168828517,0.1767283529,0.0041359356,-0.6324492693,-0.3507916033,0.3485687375,-0.0266018584,0.1511572897,0.1762987822,0.0596724488,0.2824338078,-0.3841683567,0.0531042032,-0.2631133497,-0.4323770106,0.1425144523,-0.2498616874,-0.220136717,-0.0182346441,0.1905058622,0.2702789605,-0.2122463733,-0.3237496614,-0.3732507825,-0.2246032059,-0.0373766795,0.1037429646,0.1593064964,0.2925694287,-0.1132495999,-0.0387336724,0.1269527972,-0.2187675238,0.2200421691,-0.197969988,0.1281968504,0.2286254913,0.3870925903,0.0413929857,0.5444860458,0.2436498553,-0.0796830058,0.3192813396,-0.1745523363,0.2922652662,-0.037890166,-0.5508058667,-0.1344224811,-0.1521711648,0.0345692784,0.0439348668,0.1621201336,0.1705301851,-0.5553136468,-0.2198973298,-0.191320166,-0.2798695266,-0.0953921154,0.0088962717,0.175069645,0.0263881776,-0.1763007343,-0.2223936319,0.1937036365,0.2320328504,0.5205129981,-0.0382330604,0.0609751306,-0.1520732641,-0.3924021125,-0.4625766575,0.4786426425,0.0343587212,0.1926789582,-0.2925172746,-0.0163767822,-0.0434862077,0.1060830727,0.7086897492,-0.0712133721,0.2219744027,0.0485518277,-0.0863394961,-0.3645547032,-0.0629987493,0.0806383789,0.2895561159,0.5007997751,0.4849519432,-0.3278111815,0.096823208,0.2301485389,-0.0766402856,-0.0308216345,-0.0739789009,-0.2963504195,-0.3656420708,-0.267806977,0.0053085051,0.2438388169,0.2123910487,0.3308703899,-0.1466614455,0.1623217314,-0.2427598834,-0.059580896,-0.0181475859,0.0382346138,-0.201649636,-0.0326519087,0.1781357527,0.3460509479,-0.3148242235,0.4281020164,-0.2402162999,-0.2235836387,0.1684294194,0.3209204674,0.4058360159,0.3009501994,-0.1991278529,0.1601253003,0.2677814364,0.3612563014,-0.1262014061,0.1613375992,0.4501324892,0.3550080657,-0.1951705515,-0.2329108715,0.2528257668,-0.0729128048,0.0030149017,0.2449313849,0.4155495167,-0.2275766283,0.165534988,-0.2975753248,1.1979163885,0.0573303141,-0.0391714573,0.1336840391,-0.237947315,0.5808716416,-0.156599924,0.1131992862,-0.1766391993,-0.0917727649,0.1218969896,-0.2235735059,0.3505120277,-0.1980931312,-0.0589326397,0.2558603883,-0.1141535416,0.3006235659,0.2014454603,0.2694348395,-0.0904738307,0.0046210554,-0.257040292,0.1390346289,-0.0044752457,0.4249934554,-0.0580967255,-0.1384331137,0.0146737667,-0.3370604515,-0.3443261981,0.1544762552,-0.7144252062,0.1664410383,0.0121583911,-0.082237266,-0.1178105026,-0.0211141706,0.3912243843,0.1099306643,-0.219323352,0.1842373759,-0.0037481382,-0.3129947782,0.1039159521,0.1227223426,-0.0625953376,-0.269305557,-0.2032667845,0.141741544,-0.1730962843,-0.422695756,0.2038718164,-0.1461987346,0.1515293568,-0.1836434305,-0.0226265304,-0.1612545848,0.0242267586,-0.1469756216,0.112170592,-0.0693805888,-0.2479232848,-0.1085977331,0.0223695394,-0.3606722057,0.0174629949,0.4281620383,-0.144581303,0.0723203868,0.4555252492,0.1247401312,-0.131731987,-0.1284465343,-0.2199524939,0.0430559218,-0.4091939628,-0.0014768161,-0.0063629537,-0.1436571628,-0.230215624,-0.0912814885,0.3861811757,0.2269382328,-0.0249439478,-0.5732638836,-0.1236653402,0.1283735037,-0.1113202199,-0.0436672159,0.1352208406,0.2921498716,0.0813923702,0.0772792622,-0.2949266732,-0.0582201816,-0.1980380267,-0.0909823924,0.4604614675,-0.4231759012,0.5564394593,0.2472602725,0.18336308,0.0185599104,-0.1767931283,-0.1913488358,-0.0861806348,0.123901844,-0.104550086,-0.2903296351,0.0198584348,-0.3376125097,0.0041401763,-0.0067034117,0.0077222949,0.0418777503,0.0665356293,-0.2817328274,-0.0320656784,-0.0895531774,-0.1053852886,0.1841530949,-0.1359493136,0.178451702,0.0287575107,-0.0061410554,-0.2754328847,-0.1757724285,-0.2360019982,-0.0797953382,0.0297373626,0.1081585288,0.3297698498,-0.1942861974,-0.2923170328,-0.1087816954,0.4173107445,-0.0037687309,-0.1419619322,0.059019573,0.1690811962,0.1616690904,-0.1311531216,0.0745707229,0.1629071534,-0.1565767974,0.1184273064,0.2068682462,-0.0237029567,0.1638732702,0.0392775722,0.0193122122,0.8280630112,0.2137528658,0.084507376,0.1391806453,0.0415813588,-0.0628211871,0.1529559493,0.0035156396,0.2465283573,-0.0891733319,-0.3549365401,-0.0614763871,-0.0593232363,-0.0681368411,0.2297047675,-0.0699740872,0.0251145214,0.1838400215,-0.0988485664,-0.1319564432,-0.4080098569,0.6134052873,-0.4409067929,-0.0357085466,-0.1996202171,0.3630471528,0.0331471078,-0.1725615412,0.1452637911,-0.1420483887,-0.2319661677,-0.3037571013,0.0896623954,-0.3973191082,0.3674946129,-0.0679671764,0.0307095051,-0.0729464665,-0.3076351881,0.1328382194,0.2419953793,-0.2333100587,0.3818818331,0.2886329293,-0.0376283564,-0.1605471075,0.4000950456,0.6762163043,0.2699618042,-0.2351969033,0.2674933672,0.0332841128,-0.052856721,-0.0259344038,-0.005995648,0.2812909782,-0.0034649612,0.2339333296,0.233676374,-0.2260972857,0.0528747365,-0.1916484237,0.3153737783,-0.0148682706,0.4126878381,-0.3227872252,-0.1426360309,-0.3223921061,0.1364282221,-0.8104321361,0.1720464826,0.2672681212,-0.4068028927,0.0215156246,-0.1731269211,0.080315344,-0.0752660111,0.7511890531,0.2013044804,0.3521663547,-0.4966924787,-0.3168994188,-0.2904440463,0.2075130939,-0.1204550713,-0.1377062201,-0.0594685376,-0.0786989182,-0.1366242021,0.1496257037,0.1224855855,0.1239469945,-0.2708899975,-0.1634471714,-0.0168661717,0.0420122892,-0.0634030178,0.1023459509,0.27331236,-0.2260944992,0.2341933399,-0.0447947569,0.0383970998,0.2165072858,-0.0195575412,0.1787251234,-0.3686765432,0.3712461293,0.1257679909,0.347453624,0.1640767455,-0.0576720797,-0.1508484334,-0.2002634406,-0.0525047928,0.1181119606,-0.0094588045,0.0054064239,-0.2345675826,-0.2379542887,-0.5355015397,0.1884404123,0.0324563906,-0.0615439638,0.0410017781,-0.1977436841,0.1453060806,0.1729564071,0.2607688904,0.0437686816,-0.0029306645,0.1317536086,-0.3068507314,-0.266634196,0.6317306161,-0.4022055566,-0.2421782613,0.0718518123,0.36091277,0.1383718401,0.1080266684,-0.4578807056,0.0367780551,0.3818689883,0.3051508665,-0.2807103693,0.2636131942,-0.2206344306,0.0750978142,0.1570173502,0.1133459508,-0.1142306253,-0.3479038477,-0.0901176855,-0.1272863001]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2787","title":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com","comments":"I can access https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py without problem...","body":"Hello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n","comment_length":6,"text":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\nHello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n\nI can access https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py without problem...","embeddings":[-0.2038715184,-0.1601331234,-0.0819259807,0.046902068,0.2665438056,-0.0723191276,0.1546953321,0.3012448549,0.1090682521,-0.0955984294,-0.1769392192,-0.1569310576,0.0424281619,0.0098089986,0.0353741758,-0.1757960469,-0.0955913514,-0.0462971069,-0.2514628172,0.187785551,-0.1379356235,0.1785493493,-0.0379736349,0.2863889635,-0.1777602732,-0.2044938058,-0.0835378692,0.2092444748,-0.1527038217,-0.2665781081,0.4261870086,-0.1104120836,0.1023879945,0.4276525974,-0.0001111302,0.2171017975,0.4489411414,0.0764487162,-0.1543845832,-0.3074429333,-0.0379915237,0.0689299181,0.2134657204,-0.2233344316,-0.1960697025,0.3461775184,0.0491150022,-0.1808769554,0.4765046239,0.2705875933,0.2488420606,0.3377011716,0.4446749985,-0.2152077854,0.0332203731,-0.2912712395,0.0024016653,0.6216793656,0.1213354543,-0.2661054432,0.2426152825,0.2062644213,-0.1433541477,-0.0120363627,0.0456216745,-0.1558198184,0.0829060301,-0.2861295044,0.1528282017,0.0686278045,-0.141816169,-0.2192167342,-0.2591913342,-0.0737397969,-0.0262958892,-0.1296845824,0.331905216,0.1610924304,-0.2213149071,0.2277653813,-0.0923094675,-0.1391131133,-0.2099060416,0.2466940284,0.0454544127,0.2128365934,-0.285688132,0.2644905746,0.4337483943,-0.0378364623,-0.2852620184,0.0696709752,0.1732451916,0.1920874119,-0.2251850963,0.0389768854,-0.3047998548,0.0818184689,0.0858520195,0.0211909339,-0.2492952347,-0.1187235191,0.2668730021,0.1873599738,-0.0882738084,0.3309412599,0.0141430199,0.1095975041,0.3459482193,0.3594096899,-0.1148068979,0.0365825891,-0.2656762898,-0.46562168,-0.3594941199,0.0650140643,0.2611832321,0.0185204595,-0.1696608961,0.0698689297,-0.1784015596,0.0330274105,0.1020842791,0.5010195374,-0.264885366,-0.3661019206,0.0398671776,0.2935795486,0.0180174913,0.0218921565,-0.1782663316,0.157259956,-0.2640621364,0.0905434713,0.0618727505,-0.254320085,0.2353595346,-0.2026710063,0.0230699349,-0.1711889654,0.1205860972,-0.2005208731,-0.158144623,0.3388409019,0.0797365904,0.180475682,0.4291989803,-0.3894393742,-0.0596620627,-0.0774279088,-0.2149029821,-0.4968355596,0.0238060597,0.1501094401,-0.0620120503,0.1604385972,-0.0989380255,-0.248925671,0.2021318972,0.1912706196,-0.0140952608,-0.0100923842,0.1434962451,-0.2804526091,0.5142765641,0.4544944763,0.1708977968,-0.1090477556,-0.0004218963,0.0640206262,-0.0517981537,0.2329597026,-0.0699741095,0.1279715449,-0.3756048977,0.1551721245,0.3715333343,-0.653388679,-0.2841059566,0.3186047971,-0.274140209,-0.0540166125,0.1668800265,-0.2321895212,-0.0450625047,0.0239412375,0.3900935948,0.0740385205,0.0700897276,-0.2052421421,-0.0241824016,-0.3783530891,0.1280156225,0.0518320054,0.1179565564,0.0941041037,0.0289575048,-0.2993083894,0.2291134,0.1804360002,0.0232359786,0.22256051,0.0875867456,0.1844524741,-0.0589135289,-0.1022363752,-0.0982042775,0.1569275558,-0.13933146,0.3022058308,-0.4215773046,0.0652287006,-0.3841099143,0.0516632423,-0.2309429497,-0.3040854037,0.1636356562,-0.0034321281,0.4423214495,0.1496356577,-0.0936227217,0.4874028862,-0.1470745057,0.2835756838,-0.3597308099,0.2300837487,-0.3057549596,-0.1062674224,-0.156371057,0.166188091,0.351362288,-0.2343126982,-0.0265929922,0.4231664836,-0.2641439438,0.0973605886,0.2657681406,0.0963225588,0.031523183,-0.2190815657,0.0253635254,-0.1022461355,-0.0166736674,0.0638581291,0.6379241347,0.0479976051,0.1513416469,0.7245124578,0.1906802058,0.2567310035,0.1417647898,-0.0172172338,-0.2446911037,0.0072813323,0.2052772939,0.0509162173,0.4524137974,-0.3265386522,0.1677902639,-0.2114031017,0.1441133022,-0.035479892,0.2344904095,0.1491603702,-0.3617250323,0.2157289535,-0.1119084656,0.3587546647,0.3501404822,0.0897394866,-0.0167159215,0.2718507648,-0.1022959575,-0.0745456368,0.3431275487,0.0640020669,0.0710157901,-0.1801293045,-0.0368262008,-0.1345124245,-0.1243019998,-0.2394788861,-0.1977582723,0.1469757408,-0.2880876958,0.1693797261,-0.1144179553,-0.3013064861,-0.1543430984,-0.3460415006,-0.4186742306,-0.2409940213,-0.3461467028,0.0472732708,0.0670877323,0.2756136656,0.1089546755,0.0105085308,-0.0052667689,-0.1842108667,-0.034712214,-0.1745305657,-0.1391935796,0.0206357259,-0.0188447759,-0.2574251294,0.2771224976,-0.1947549433,-0.1127352193,0.0472366884,-0.2349187285,0.0479723103,-0.0904428661,0.1711955369,0.328553766,0.3805794716,0.1841480136,-0.3997780979,0.5727685094,0.0603625476,0.1144243404,-0.1596369445,0.1083576158,0.0168828517,0.1767283529,0.0041359356,-0.6324492693,-0.3507916033,0.3485687375,-0.0266018584,0.1511572897,0.1762987822,0.0596724488,0.2824338078,-0.3841683567,0.0531042032,-0.2631133497,-0.4323770106,0.1425144523,-0.2498616874,-0.220136717,-0.0182346441,0.1905058622,0.2702789605,-0.2122463733,-0.3237496614,-0.3732507825,-0.2246032059,-0.0373766795,0.1037429646,0.1593064964,0.2925694287,-0.1132495999,-0.0387336724,0.1269527972,-0.2187675238,0.2200421691,-0.197969988,0.1281968504,0.2286254913,0.3870925903,0.0413929857,0.5444860458,0.2436498553,-0.0796830058,0.3192813396,-0.1745523363,0.2922652662,-0.037890166,-0.5508058667,-0.1344224811,-0.1521711648,0.0345692784,0.0439348668,0.1621201336,0.1705301851,-0.5553136468,-0.2198973298,-0.191320166,-0.2798695266,-0.0953921154,0.0088962717,0.175069645,0.0263881776,-0.1763007343,-0.2223936319,0.1937036365,0.2320328504,0.5205129981,-0.0382330604,0.0609751306,-0.1520732641,-0.3924021125,-0.4625766575,0.4786426425,0.0343587212,0.1926789582,-0.2925172746,-0.0163767822,-0.0434862077,0.1060830727,0.7086897492,-0.0712133721,0.2219744027,0.0485518277,-0.0863394961,-0.3645547032,-0.0629987493,0.0806383789,0.2895561159,0.5007997751,0.4849519432,-0.3278111815,0.096823208,0.2301485389,-0.0766402856,-0.0308216345,-0.0739789009,-0.2963504195,-0.3656420708,-0.267806977,0.0053085051,0.2438388169,0.2123910487,0.3308703899,-0.1466614455,0.1623217314,-0.2427598834,-0.059580896,-0.0181475859,0.0382346138,-0.201649636,-0.0326519087,0.1781357527,0.3460509479,-0.3148242235,0.4281020164,-0.2402162999,-0.2235836387,0.1684294194,0.3209204674,0.4058360159,0.3009501994,-0.1991278529,0.1601253003,0.2677814364,0.3612563014,-0.1262014061,0.1613375992,0.4501324892,0.3550080657,-0.1951705515,-0.2329108715,0.2528257668,-0.0729128048,0.0030149017,0.2449313849,0.4155495167,-0.2275766283,0.165534988,-0.2975753248,1.1979163885,0.0573303141,-0.0391714573,0.1336840391,-0.237947315,0.5808716416,-0.156599924,0.1131992862,-0.1766391993,-0.0917727649,0.1218969896,-0.2235735059,0.3505120277,-0.1980931312,-0.0589326397,0.2558603883,-0.1141535416,0.3006235659,0.2014454603,0.2694348395,-0.0904738307,0.0046210554,-0.257040292,0.1390346289,-0.0044752457,0.4249934554,-0.0580967255,-0.1384331137,0.0146737667,-0.3370604515,-0.3443261981,0.1544762552,-0.7144252062,0.1664410383,0.0121583911,-0.082237266,-0.1178105026,-0.0211141706,0.3912243843,0.1099306643,-0.219323352,0.1842373759,-0.0037481382,-0.3129947782,0.1039159521,0.1227223426,-0.0625953376,-0.269305557,-0.2032667845,0.141741544,-0.1730962843,-0.422695756,0.2038718164,-0.1461987346,0.1515293568,-0.1836434305,-0.0226265304,-0.1612545848,0.0242267586,-0.1469756216,0.112170592,-0.0693805888,-0.2479232848,-0.1085977331,0.0223695394,-0.3606722057,0.0174629949,0.4281620383,-0.144581303,0.0723203868,0.4555252492,0.1247401312,-0.131731987,-0.1284465343,-0.2199524939,0.0430559218,-0.4091939628,-0.0014768161,-0.0063629537,-0.1436571628,-0.230215624,-0.0912814885,0.3861811757,0.2269382328,-0.0249439478,-0.5732638836,-0.1236653402,0.1283735037,-0.1113202199,-0.0436672159,0.1352208406,0.2921498716,0.0813923702,0.0772792622,-0.2949266732,-0.0582201816,-0.1980380267,-0.0909823924,0.4604614675,-0.4231759012,0.5564394593,0.2472602725,0.18336308,0.0185599104,-0.1767931283,-0.1913488358,-0.0861806348,0.123901844,-0.104550086,-0.2903296351,0.0198584348,-0.3376125097,0.0041401763,-0.0067034117,0.0077222949,0.0418777503,0.0665356293,-0.2817328274,-0.0320656784,-0.0895531774,-0.1053852886,0.1841530949,-0.1359493136,0.178451702,0.0287575107,-0.0061410554,-0.2754328847,-0.1757724285,-0.2360019982,-0.0797953382,0.0297373626,0.1081585288,0.3297698498,-0.1942861974,-0.2923170328,-0.1087816954,0.4173107445,-0.0037687309,-0.1419619322,0.059019573,0.1690811962,0.1616690904,-0.1311531216,0.0745707229,0.1629071534,-0.1565767974,0.1184273064,0.2068682462,-0.0237029567,0.1638732702,0.0392775722,0.0193122122,0.8280630112,0.2137528658,0.084507376,0.1391806453,0.0415813588,-0.0628211871,0.1529559493,0.0035156396,0.2465283573,-0.0891733319,-0.3549365401,-0.0614763871,-0.0593232363,-0.0681368411,0.2297047675,-0.0699740872,0.0251145214,0.1838400215,-0.0988485664,-0.1319564432,-0.4080098569,0.6134052873,-0.4409067929,-0.0357085466,-0.1996202171,0.3630471528,0.0331471078,-0.1725615412,0.1452637911,-0.1420483887,-0.2319661677,-0.3037571013,0.0896623954,-0.3973191082,0.3674946129,-0.0679671764,0.0307095051,-0.0729464665,-0.3076351881,0.1328382194,0.2419953793,-0.2333100587,0.3818818331,0.2886329293,-0.0376283564,-0.1605471075,0.4000950456,0.6762163043,0.2699618042,-0.2351969033,0.2674933672,0.0332841128,-0.052856721,-0.0259344038,-0.005995648,0.2812909782,-0.0034649612,0.2339333296,0.233676374,-0.2260972857,0.0528747365,-0.1916484237,0.3153737783,-0.0148682706,0.4126878381,-0.3227872252,-0.1426360309,-0.3223921061,0.1364282221,-0.8104321361,0.1720464826,0.2672681212,-0.4068028927,0.0215156246,-0.1731269211,0.080315344,-0.0752660111,0.7511890531,0.2013044804,0.3521663547,-0.4966924787,-0.3168994188,-0.2904440463,0.2075130939,-0.1204550713,-0.1377062201,-0.0594685376,-0.0786989182,-0.1366242021,0.1496257037,0.1224855855,0.1239469945,-0.2708899975,-0.1634471714,-0.0168661717,0.0420122892,-0.0634030178,0.1023459509,0.27331236,-0.2260944992,0.2341933399,-0.0447947569,0.0383970998,0.2165072858,-0.0195575412,0.1787251234,-0.3686765432,0.3712461293,0.1257679909,0.347453624,0.1640767455,-0.0576720797,-0.1508484334,-0.2002634406,-0.0525047928,0.1181119606,-0.0094588045,0.0054064239,-0.2345675826,-0.2379542887,-0.5355015397,0.1884404123,0.0324563906,-0.0615439638,0.0410017781,-0.1977436841,0.1453060806,0.1729564071,0.2607688904,0.0437686816,-0.0029306645,0.1317536086,-0.3068507314,-0.266634196,0.6317306161,-0.4022055566,-0.2421782613,0.0718518123,0.36091277,0.1383718401,0.1080266684,-0.4578807056,0.0367780551,0.3818689883,0.3051508665,-0.2807103693,0.2636131942,-0.2206344306,0.0750978142,0.1570173502,0.1133459508,-0.1142306253,-0.3479038477,-0.0901176855,-0.1272863001]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2787","title":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com","comments":"> I can access https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py without problem...\r\n\r\nI can not access https:\/\/raw.githubusercontent.com\/huggingface\/datasets either, I am in China","body":"Hello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n","comment_length":17,"text":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\nHello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n\n> I can access https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py without problem...\r\n\r\nI can not access https:\/\/raw.githubusercontent.com\/huggingface\/datasets either, I am in China","embeddings":[-0.2038715184,-0.1601331234,-0.0819259807,0.046902068,0.2665438056,-0.0723191276,0.1546953321,0.3012448549,0.1090682521,-0.0955984294,-0.1769392192,-0.1569310576,0.0424281619,0.0098089986,0.0353741758,-0.1757960469,-0.0955913514,-0.0462971069,-0.2514628172,0.187785551,-0.1379356235,0.1785493493,-0.0379736349,0.2863889635,-0.1777602732,-0.2044938058,-0.0835378692,0.2092444748,-0.1527038217,-0.2665781081,0.4261870086,-0.1104120836,0.1023879945,0.4276525974,-0.0001111302,0.2171017975,0.4489411414,0.0764487162,-0.1543845832,-0.3074429333,-0.0379915237,0.0689299181,0.2134657204,-0.2233344316,-0.1960697025,0.3461775184,0.0491150022,-0.1808769554,0.4765046239,0.2705875933,0.2488420606,0.3377011716,0.4446749985,-0.2152077854,0.0332203731,-0.2912712395,0.0024016653,0.6216793656,0.1213354543,-0.2661054432,0.2426152825,0.2062644213,-0.1433541477,-0.0120363627,0.0456216745,-0.1558198184,0.0829060301,-0.2861295044,0.1528282017,0.0686278045,-0.141816169,-0.2192167342,-0.2591913342,-0.0737397969,-0.0262958892,-0.1296845824,0.331905216,0.1610924304,-0.2213149071,0.2277653813,-0.0923094675,-0.1391131133,-0.2099060416,0.2466940284,0.0454544127,0.2128365934,-0.285688132,0.2644905746,0.4337483943,-0.0378364623,-0.2852620184,0.0696709752,0.1732451916,0.1920874119,-0.2251850963,0.0389768854,-0.3047998548,0.0818184689,0.0858520195,0.0211909339,-0.2492952347,-0.1187235191,0.2668730021,0.1873599738,-0.0882738084,0.3309412599,0.0141430199,0.1095975041,0.3459482193,0.3594096899,-0.1148068979,0.0365825891,-0.2656762898,-0.46562168,-0.3594941199,0.0650140643,0.2611832321,0.0185204595,-0.1696608961,0.0698689297,-0.1784015596,0.0330274105,0.1020842791,0.5010195374,-0.264885366,-0.3661019206,0.0398671776,0.2935795486,0.0180174913,0.0218921565,-0.1782663316,0.157259956,-0.2640621364,0.0905434713,0.0618727505,-0.254320085,0.2353595346,-0.2026710063,0.0230699349,-0.1711889654,0.1205860972,-0.2005208731,-0.158144623,0.3388409019,0.0797365904,0.180475682,0.4291989803,-0.3894393742,-0.0596620627,-0.0774279088,-0.2149029821,-0.4968355596,0.0238060597,0.1501094401,-0.0620120503,0.1604385972,-0.0989380255,-0.248925671,0.2021318972,0.1912706196,-0.0140952608,-0.0100923842,0.1434962451,-0.2804526091,0.5142765641,0.4544944763,0.1708977968,-0.1090477556,-0.0004218963,0.0640206262,-0.0517981537,0.2329597026,-0.0699741095,0.1279715449,-0.3756048977,0.1551721245,0.3715333343,-0.653388679,-0.2841059566,0.3186047971,-0.274140209,-0.0540166125,0.1668800265,-0.2321895212,-0.0450625047,0.0239412375,0.3900935948,0.0740385205,0.0700897276,-0.2052421421,-0.0241824016,-0.3783530891,0.1280156225,0.0518320054,0.1179565564,0.0941041037,0.0289575048,-0.2993083894,0.2291134,0.1804360002,0.0232359786,0.22256051,0.0875867456,0.1844524741,-0.0589135289,-0.1022363752,-0.0982042775,0.1569275558,-0.13933146,0.3022058308,-0.4215773046,0.0652287006,-0.3841099143,0.0516632423,-0.2309429497,-0.3040854037,0.1636356562,-0.0034321281,0.4423214495,0.1496356577,-0.0936227217,0.4874028862,-0.1470745057,0.2835756838,-0.3597308099,0.2300837487,-0.3057549596,-0.1062674224,-0.156371057,0.166188091,0.351362288,-0.2343126982,-0.0265929922,0.4231664836,-0.2641439438,0.0973605886,0.2657681406,0.0963225588,0.031523183,-0.2190815657,0.0253635254,-0.1022461355,-0.0166736674,0.0638581291,0.6379241347,0.0479976051,0.1513416469,0.7245124578,0.1906802058,0.2567310035,0.1417647898,-0.0172172338,-0.2446911037,0.0072813323,0.2052772939,0.0509162173,0.4524137974,-0.3265386522,0.1677902639,-0.2114031017,0.1441133022,-0.035479892,0.2344904095,0.1491603702,-0.3617250323,0.2157289535,-0.1119084656,0.3587546647,0.3501404822,0.0897394866,-0.0167159215,0.2718507648,-0.1022959575,-0.0745456368,0.3431275487,0.0640020669,0.0710157901,-0.1801293045,-0.0368262008,-0.1345124245,-0.1243019998,-0.2394788861,-0.1977582723,0.1469757408,-0.2880876958,0.1693797261,-0.1144179553,-0.3013064861,-0.1543430984,-0.3460415006,-0.4186742306,-0.2409940213,-0.3461467028,0.0472732708,0.0670877323,0.2756136656,0.1089546755,0.0105085308,-0.0052667689,-0.1842108667,-0.034712214,-0.1745305657,-0.1391935796,0.0206357259,-0.0188447759,-0.2574251294,0.2771224976,-0.1947549433,-0.1127352193,0.0472366884,-0.2349187285,0.0479723103,-0.0904428661,0.1711955369,0.328553766,0.3805794716,0.1841480136,-0.3997780979,0.5727685094,0.0603625476,0.1144243404,-0.1596369445,0.1083576158,0.0168828517,0.1767283529,0.0041359356,-0.6324492693,-0.3507916033,0.3485687375,-0.0266018584,0.1511572897,0.1762987822,0.0596724488,0.2824338078,-0.3841683567,0.0531042032,-0.2631133497,-0.4323770106,0.1425144523,-0.2498616874,-0.220136717,-0.0182346441,0.1905058622,0.2702789605,-0.2122463733,-0.3237496614,-0.3732507825,-0.2246032059,-0.0373766795,0.1037429646,0.1593064964,0.2925694287,-0.1132495999,-0.0387336724,0.1269527972,-0.2187675238,0.2200421691,-0.197969988,0.1281968504,0.2286254913,0.3870925903,0.0413929857,0.5444860458,0.2436498553,-0.0796830058,0.3192813396,-0.1745523363,0.2922652662,-0.037890166,-0.5508058667,-0.1344224811,-0.1521711648,0.0345692784,0.0439348668,0.1621201336,0.1705301851,-0.5553136468,-0.2198973298,-0.191320166,-0.2798695266,-0.0953921154,0.0088962717,0.175069645,0.0263881776,-0.1763007343,-0.2223936319,0.1937036365,0.2320328504,0.5205129981,-0.0382330604,0.0609751306,-0.1520732641,-0.3924021125,-0.4625766575,0.4786426425,0.0343587212,0.1926789582,-0.2925172746,-0.0163767822,-0.0434862077,0.1060830727,0.7086897492,-0.0712133721,0.2219744027,0.0485518277,-0.0863394961,-0.3645547032,-0.0629987493,0.0806383789,0.2895561159,0.5007997751,0.4849519432,-0.3278111815,0.096823208,0.2301485389,-0.0766402856,-0.0308216345,-0.0739789009,-0.2963504195,-0.3656420708,-0.267806977,0.0053085051,0.2438388169,0.2123910487,0.3308703899,-0.1466614455,0.1623217314,-0.2427598834,-0.059580896,-0.0181475859,0.0382346138,-0.201649636,-0.0326519087,0.1781357527,0.3460509479,-0.3148242235,0.4281020164,-0.2402162999,-0.2235836387,0.1684294194,0.3209204674,0.4058360159,0.3009501994,-0.1991278529,0.1601253003,0.2677814364,0.3612563014,-0.1262014061,0.1613375992,0.4501324892,0.3550080657,-0.1951705515,-0.2329108715,0.2528257668,-0.0729128048,0.0030149017,0.2449313849,0.4155495167,-0.2275766283,0.165534988,-0.2975753248,1.1979163885,0.0573303141,-0.0391714573,0.1336840391,-0.237947315,0.5808716416,-0.156599924,0.1131992862,-0.1766391993,-0.0917727649,0.1218969896,-0.2235735059,0.3505120277,-0.1980931312,-0.0589326397,0.2558603883,-0.1141535416,0.3006235659,0.2014454603,0.2694348395,-0.0904738307,0.0046210554,-0.257040292,0.1390346289,-0.0044752457,0.4249934554,-0.0580967255,-0.1384331137,0.0146737667,-0.3370604515,-0.3443261981,0.1544762552,-0.7144252062,0.1664410383,0.0121583911,-0.082237266,-0.1178105026,-0.0211141706,0.3912243843,0.1099306643,-0.219323352,0.1842373759,-0.0037481382,-0.3129947782,0.1039159521,0.1227223426,-0.0625953376,-0.269305557,-0.2032667845,0.141741544,-0.1730962843,-0.422695756,0.2038718164,-0.1461987346,0.1515293568,-0.1836434305,-0.0226265304,-0.1612545848,0.0242267586,-0.1469756216,0.112170592,-0.0693805888,-0.2479232848,-0.1085977331,0.0223695394,-0.3606722057,0.0174629949,0.4281620383,-0.144581303,0.0723203868,0.4555252492,0.1247401312,-0.131731987,-0.1284465343,-0.2199524939,0.0430559218,-0.4091939628,-0.0014768161,-0.0063629537,-0.1436571628,-0.230215624,-0.0912814885,0.3861811757,0.2269382328,-0.0249439478,-0.5732638836,-0.1236653402,0.1283735037,-0.1113202199,-0.0436672159,0.1352208406,0.2921498716,0.0813923702,0.0772792622,-0.2949266732,-0.0582201816,-0.1980380267,-0.0909823924,0.4604614675,-0.4231759012,0.5564394593,0.2472602725,0.18336308,0.0185599104,-0.1767931283,-0.1913488358,-0.0861806348,0.123901844,-0.104550086,-0.2903296351,0.0198584348,-0.3376125097,0.0041401763,-0.0067034117,0.0077222949,0.0418777503,0.0665356293,-0.2817328274,-0.0320656784,-0.0895531774,-0.1053852886,0.1841530949,-0.1359493136,0.178451702,0.0287575107,-0.0061410554,-0.2754328847,-0.1757724285,-0.2360019982,-0.0797953382,0.0297373626,0.1081585288,0.3297698498,-0.1942861974,-0.2923170328,-0.1087816954,0.4173107445,-0.0037687309,-0.1419619322,0.059019573,0.1690811962,0.1616690904,-0.1311531216,0.0745707229,0.1629071534,-0.1565767974,0.1184273064,0.2068682462,-0.0237029567,0.1638732702,0.0392775722,0.0193122122,0.8280630112,0.2137528658,0.084507376,0.1391806453,0.0415813588,-0.0628211871,0.1529559493,0.0035156396,0.2465283573,-0.0891733319,-0.3549365401,-0.0614763871,-0.0593232363,-0.0681368411,0.2297047675,-0.0699740872,0.0251145214,0.1838400215,-0.0988485664,-0.1319564432,-0.4080098569,0.6134052873,-0.4409067929,-0.0357085466,-0.1996202171,0.3630471528,0.0331471078,-0.1725615412,0.1452637911,-0.1420483887,-0.2319661677,-0.3037571013,0.0896623954,-0.3973191082,0.3674946129,-0.0679671764,0.0307095051,-0.0729464665,-0.3076351881,0.1328382194,0.2419953793,-0.2333100587,0.3818818331,0.2886329293,-0.0376283564,-0.1605471075,0.4000950456,0.6762163043,0.2699618042,-0.2351969033,0.2674933672,0.0332841128,-0.052856721,-0.0259344038,-0.005995648,0.2812909782,-0.0034649612,0.2339333296,0.233676374,-0.2260972857,0.0528747365,-0.1916484237,0.3153737783,-0.0148682706,0.4126878381,-0.3227872252,-0.1426360309,-0.3223921061,0.1364282221,-0.8104321361,0.1720464826,0.2672681212,-0.4068028927,0.0215156246,-0.1731269211,0.080315344,-0.0752660111,0.7511890531,0.2013044804,0.3521663547,-0.4966924787,-0.3168994188,-0.2904440463,0.2075130939,-0.1204550713,-0.1377062201,-0.0594685376,-0.0786989182,-0.1366242021,0.1496257037,0.1224855855,0.1239469945,-0.2708899975,-0.1634471714,-0.0168661717,0.0420122892,-0.0634030178,0.1023459509,0.27331236,-0.2260944992,0.2341933399,-0.0447947569,0.0383970998,0.2165072858,-0.0195575412,0.1787251234,-0.3686765432,0.3712461293,0.1257679909,0.347453624,0.1640767455,-0.0576720797,-0.1508484334,-0.2002634406,-0.0525047928,0.1181119606,-0.0094588045,0.0054064239,-0.2345675826,-0.2379542887,-0.5355015397,0.1884404123,0.0324563906,-0.0615439638,0.0410017781,-0.1977436841,0.1453060806,0.1729564071,0.2607688904,0.0437686816,-0.0029306645,0.1317536086,-0.3068507314,-0.266634196,0.6317306161,-0.4022055566,-0.2421782613,0.0718518123,0.36091277,0.1383718401,0.1080266684,-0.4578807056,0.0367780551,0.3818689883,0.3051508665,-0.2807103693,0.2636131942,-0.2206344306,0.0750978142,0.1570173502,0.1133459508,-0.1142306253,-0.3479038477,-0.0901176855,-0.1272863001]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2787","title":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com","comments":"Finally i can access it, by the superfast software. Thanks","body":"Hello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n","comment_length":10,"text":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\nHello,\r\nI am trying to run run_glue.py and it gives me this error -\r\n\r\nTraceback (most recent call last):\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 546, in \r\n main()\r\n File \"E:\/BERT\/pytorch_hugging\/transformers\/examples\/pytorch\/text-classification\/run_glue.py\", line 250, in main\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 718, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"C:\\install\\Anaconda3\\envs\\huggingface\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/datasets\/glue\/glue.py\r\n\r\nTrying to do python run_glue.py --model_name_or_path\r\nbert-base-cased\r\n--task_name\r\nmrpc\r\n--do_train\r\n--do_eval\r\n--max_seq_length\r\n128\r\n--per_device_train_batch_size\r\n32\r\n--learning_rate\r\n2e-5\r\n--num_train_epochs\r\n3\r\n--output_dir\r\n.\/tmp\/mrpc\/\r\n\r\nIs this something on my end? From what I can tell, this was re-fixeded by @fullyz a few months ago.\r\nThank you!\r\n\nFinally i can access it, by the superfast software. Thanks","embeddings":[-0.2038715184,-0.1601331234,-0.0819259807,0.046902068,0.2665438056,-0.0723191276,0.1546953321,0.3012448549,0.1090682521,-0.0955984294,-0.1769392192,-0.1569310576,0.0424281619,0.0098089986,0.0353741758,-0.1757960469,-0.0955913514,-0.0462971069,-0.2514628172,0.187785551,-0.1379356235,0.1785493493,-0.0379736349,0.2863889635,-0.1777602732,-0.2044938058,-0.0835378692,0.2092444748,-0.1527038217,-0.2665781081,0.4261870086,-0.1104120836,0.1023879945,0.4276525974,-0.0001111302,0.2171017975,0.4489411414,0.0764487162,-0.1543845832,-0.3074429333,-0.0379915237,0.0689299181,0.2134657204,-0.2233344316,-0.1960697025,0.3461775184,0.0491150022,-0.1808769554,0.4765046239,0.2705875933,0.2488420606,0.3377011716,0.4446749985,-0.2152077854,0.0332203731,-0.2912712395,0.0024016653,0.6216793656,0.1213354543,-0.2661054432,0.2426152825,0.2062644213,-0.1433541477,-0.0120363627,0.0456216745,-0.1558198184,0.0829060301,-0.2861295044,0.1528282017,0.0686278045,-0.141816169,-0.2192167342,-0.2591913342,-0.0737397969,-0.0262958892,-0.1296845824,0.331905216,0.1610924304,-0.2213149071,0.2277653813,-0.0923094675,-0.1391131133,-0.2099060416,0.2466940284,0.0454544127,0.2128365934,-0.285688132,0.2644905746,0.4337483943,-0.0378364623,-0.2852620184,0.0696709752,0.1732451916,0.1920874119,-0.2251850963,0.0389768854,-0.3047998548,0.0818184689,0.0858520195,0.0211909339,-0.2492952347,-0.1187235191,0.2668730021,0.1873599738,-0.0882738084,0.3309412599,0.0141430199,0.1095975041,0.3459482193,0.3594096899,-0.1148068979,0.0365825891,-0.2656762898,-0.46562168,-0.3594941199,0.0650140643,0.2611832321,0.0185204595,-0.1696608961,0.0698689297,-0.1784015596,0.0330274105,0.1020842791,0.5010195374,-0.264885366,-0.3661019206,0.0398671776,0.2935795486,0.0180174913,0.0218921565,-0.1782663316,0.157259956,-0.2640621364,0.0905434713,0.0618727505,-0.254320085,0.2353595346,-0.2026710063,0.0230699349,-0.1711889654,0.1205860972,-0.2005208731,-0.158144623,0.3388409019,0.0797365904,0.180475682,0.4291989803,-0.3894393742,-0.0596620627,-0.0774279088,-0.2149029821,-0.4968355596,0.0238060597,0.1501094401,-0.0620120503,0.1604385972,-0.0989380255,-0.248925671,0.2021318972,0.1912706196,-0.0140952608,-0.0100923842,0.1434962451,-0.2804526091,0.5142765641,0.4544944763,0.1708977968,-0.1090477556,-0.0004218963,0.0640206262,-0.0517981537,0.2329597026,-0.0699741095,0.1279715449,-0.3756048977,0.1551721245,0.3715333343,-0.653388679,-0.2841059566,0.3186047971,-0.274140209,-0.0540166125,0.1668800265,-0.2321895212,-0.0450625047,0.0239412375,0.3900935948,0.0740385205,0.0700897276,-0.2052421421,-0.0241824016,-0.3783530891,0.1280156225,0.0518320054,0.1179565564,0.0941041037,0.0289575048,-0.2993083894,0.2291134,0.1804360002,0.0232359786,0.22256051,0.0875867456,0.1844524741,-0.0589135289,-0.1022363752,-0.0982042775,0.1569275558,-0.13933146,0.3022058308,-0.4215773046,0.0652287006,-0.3841099143,0.0516632423,-0.2309429497,-0.3040854037,0.1636356562,-0.0034321281,0.4423214495,0.1496356577,-0.0936227217,0.4874028862,-0.1470745057,0.2835756838,-0.3597308099,0.2300837487,-0.3057549596,-0.1062674224,-0.156371057,0.166188091,0.351362288,-0.2343126982,-0.0265929922,0.4231664836,-0.2641439438,0.0973605886,0.2657681406,0.0963225588,0.031523183,-0.2190815657,0.0253635254,-0.1022461355,-0.0166736674,0.0638581291,0.6379241347,0.0479976051,0.1513416469,0.7245124578,0.1906802058,0.2567310035,0.1417647898,-0.0172172338,-0.2446911037,0.0072813323,0.2052772939,0.0509162173,0.4524137974,-0.3265386522,0.1677902639,-0.2114031017,0.1441133022,-0.035479892,0.2344904095,0.1491603702,-0.3617250323,0.2157289535,-0.1119084656,0.3587546647,0.3501404822,0.0897394866,-0.0167159215,0.2718507648,-0.1022959575,-0.0745456368,0.3431275487,0.0640020669,0.0710157901,-0.1801293045,-0.0368262008,-0.1345124245,-0.1243019998,-0.2394788861,-0.1977582723,0.1469757408,-0.2880876958,0.1693797261,-0.1144179553,-0.3013064861,-0.1543430984,-0.3460415006,-0.4186742306,-0.2409940213,-0.3461467028,0.0472732708,0.0670877323,0.2756136656,0.1089546755,0.0105085308,-0.0052667689,-0.1842108667,-0.034712214,-0.1745305657,-0.1391935796,0.0206357259,-0.0188447759,-0.2574251294,0.2771224976,-0.1947549433,-0.1127352193,0.0472366884,-0.2349187285,0.0479723103,-0.0904428661,0.1711955369,0.328553766,0.3805794716,0.1841480136,-0.3997780979,0.5727685094,0.0603625476,0.1144243404,-0.1596369445,0.1083576158,0.0168828517,0.1767283529,0.0041359356,-0.6324492693,-0.3507916033,0.3485687375,-0.0266018584,0.1511572897,0.1762987822,0.0596724488,0.2824338078,-0.3841683567,0.0531042032,-0.2631133497,-0.4323770106,0.1425144523,-0.2498616874,-0.220136717,-0.0182346441,0.1905058622,0.2702789605,-0.2122463733,-0.3237496614,-0.3732507825,-0.2246032059,-0.0373766795,0.1037429646,0.1593064964,0.2925694287,-0.1132495999,-0.0387336724,0.1269527972,-0.2187675238,0.2200421691,-0.197969988,0.1281968504,0.2286254913,0.3870925903,0.0413929857,0.5444860458,0.2436498553,-0.0796830058,0.3192813396,-0.1745523363,0.2922652662,-0.037890166,-0.5508058667,-0.1344224811,-0.1521711648,0.0345692784,0.0439348668,0.1621201336,0.1705301851,-0.5553136468,-0.2198973298,-0.191320166,-0.2798695266,-0.0953921154,0.0088962717,0.175069645,0.0263881776,-0.1763007343,-0.2223936319,0.1937036365,0.2320328504,0.5205129981,-0.0382330604,0.0609751306,-0.1520732641,-0.3924021125,-0.4625766575,0.4786426425,0.0343587212,0.1926789582,-0.2925172746,-0.0163767822,-0.0434862077,0.1060830727,0.7086897492,-0.0712133721,0.2219744027,0.0485518277,-0.0863394961,-0.3645547032,-0.0629987493,0.0806383789,0.2895561159,0.5007997751,0.4849519432,-0.3278111815,0.096823208,0.2301485389,-0.0766402856,-0.0308216345,-0.0739789009,-0.2963504195,-0.3656420708,-0.267806977,0.0053085051,0.2438388169,0.2123910487,0.3308703899,-0.1466614455,0.1623217314,-0.2427598834,-0.059580896,-0.0181475859,0.0382346138,-0.201649636,-0.0326519087,0.1781357527,0.3460509479,-0.3148242235,0.4281020164,-0.2402162999,-0.2235836387,0.1684294194,0.3209204674,0.4058360159,0.3009501994,-0.1991278529,0.1601253003,0.2677814364,0.3612563014,-0.1262014061,0.1613375992,0.4501324892,0.3550080657,-0.1951705515,-0.2329108715,0.2528257668,-0.0729128048,0.0030149017,0.2449313849,0.4155495167,-0.2275766283,0.165534988,-0.2975753248,1.1979163885,0.0573303141,-0.0391714573,0.1336840391,-0.237947315,0.5808716416,-0.156599924,0.1131992862,-0.1766391993,-0.0917727649,0.1218969896,-0.2235735059,0.3505120277,-0.1980931312,-0.0589326397,0.2558603883,-0.1141535416,0.3006235659,0.2014454603,0.2694348395,-0.0904738307,0.0046210554,-0.257040292,0.1390346289,-0.0044752457,0.4249934554,-0.0580967255,-0.1384331137,0.0146737667,-0.3370604515,-0.3443261981,0.1544762552,-0.7144252062,0.1664410383,0.0121583911,-0.082237266,-0.1178105026,-0.0211141706,0.3912243843,0.1099306643,-0.219323352,0.1842373759,-0.0037481382,-0.3129947782,0.1039159521,0.1227223426,-0.0625953376,-0.269305557,-0.2032667845,0.141741544,-0.1730962843,-0.422695756,0.2038718164,-0.1461987346,0.1515293568,-0.1836434305,-0.0226265304,-0.1612545848,0.0242267586,-0.1469756216,0.112170592,-0.0693805888,-0.2479232848,-0.1085977331,0.0223695394,-0.3606722057,0.0174629949,0.4281620383,-0.144581303,0.0723203868,0.4555252492,0.1247401312,-0.131731987,-0.1284465343,-0.2199524939,0.0430559218,-0.4091939628,-0.0014768161,-0.0063629537,-0.1436571628,-0.230215624,-0.0912814885,0.3861811757,0.2269382328,-0.0249439478,-0.5732638836,-0.1236653402,0.1283735037,-0.1113202199,-0.0436672159,0.1352208406,0.2921498716,0.0813923702,0.0772792622,-0.2949266732,-0.0582201816,-0.1980380267,-0.0909823924,0.4604614675,-0.4231759012,0.5564394593,0.2472602725,0.18336308,0.0185599104,-0.1767931283,-0.1913488358,-0.0861806348,0.123901844,-0.104550086,-0.2903296351,0.0198584348,-0.3376125097,0.0041401763,-0.0067034117,0.0077222949,0.0418777503,0.0665356293,-0.2817328274,-0.0320656784,-0.0895531774,-0.1053852886,0.1841530949,-0.1359493136,0.178451702,0.0287575107,-0.0061410554,-0.2754328847,-0.1757724285,-0.2360019982,-0.0797953382,0.0297373626,0.1081585288,0.3297698498,-0.1942861974,-0.2923170328,-0.1087816954,0.4173107445,-0.0037687309,-0.1419619322,0.059019573,0.1690811962,0.1616690904,-0.1311531216,0.0745707229,0.1629071534,-0.1565767974,0.1184273064,0.2068682462,-0.0237029567,0.1638732702,0.0392775722,0.0193122122,0.8280630112,0.2137528658,0.084507376,0.1391806453,0.0415813588,-0.0628211871,0.1529559493,0.0035156396,0.2465283573,-0.0891733319,-0.3549365401,-0.0614763871,-0.0593232363,-0.0681368411,0.2297047675,-0.0699740872,0.0251145214,0.1838400215,-0.0988485664,-0.1319564432,-0.4080098569,0.6134052873,-0.4409067929,-0.0357085466,-0.1996202171,0.3630471528,0.0331471078,-0.1725615412,0.1452637911,-0.1420483887,-0.2319661677,-0.3037571013,0.0896623954,-0.3973191082,0.3674946129,-0.0679671764,0.0307095051,-0.0729464665,-0.3076351881,0.1328382194,0.2419953793,-0.2333100587,0.3818818331,0.2886329293,-0.0376283564,-0.1605471075,0.4000950456,0.6762163043,0.2699618042,-0.2351969033,0.2674933672,0.0332841128,-0.052856721,-0.0259344038,-0.005995648,0.2812909782,-0.0034649612,0.2339333296,0.233676374,-0.2260972857,0.0528747365,-0.1916484237,0.3153737783,-0.0148682706,0.4126878381,-0.3227872252,-0.1426360309,-0.3223921061,0.1364282221,-0.8104321361,0.1720464826,0.2672681212,-0.4068028927,0.0215156246,-0.1731269211,0.080315344,-0.0752660111,0.7511890531,0.2013044804,0.3521663547,-0.4966924787,-0.3168994188,-0.2904440463,0.2075130939,-0.1204550713,-0.1377062201,-0.0594685376,-0.0786989182,-0.1366242021,0.1496257037,0.1224855855,0.1239469945,-0.2708899975,-0.1634471714,-0.0168661717,0.0420122892,-0.0634030178,0.1023459509,0.27331236,-0.2260944992,0.2341933399,-0.0447947569,0.0383970998,0.2165072858,-0.0195575412,0.1787251234,-0.3686765432,0.3712461293,0.1257679909,0.347453624,0.1640767455,-0.0576720797,-0.1508484334,-0.2002634406,-0.0525047928,0.1181119606,-0.0094588045,0.0054064239,-0.2345675826,-0.2379542887,-0.5355015397,0.1884404123,0.0324563906,-0.0615439638,0.0410017781,-0.1977436841,0.1453060806,0.1729564071,0.2607688904,0.0437686816,-0.0029306645,0.1317536086,-0.3068507314,-0.266634196,0.6317306161,-0.4022055566,-0.2421782613,0.0718518123,0.36091277,0.1383718401,0.1080266684,-0.4578807056,0.0367780551,0.3818689883,0.3051508665,-0.2807103693,0.2636131942,-0.2206344306,0.0750978142,0.1570173502,0.1133459508,-0.1142306253,-0.3479038477,-0.0901176855,-0.1272863001]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2775","title":"`generate_random_fingerprint()` deterministic with \ud83e\udd17Transformers' `set_seed()`","comments":"I dug into what I believe is the root of this issue and added a repro in my comment. If this is better addressed as a cross-team issue, let me know and I can open an issue in the Transformers repo","body":"## Describe the bug\r\n\r\n**Update:** I dug into this to try to reproduce the underlying issue, and I believe it's that `set_seed()` from the `transformers` library makes the \"random\" fingerprint identical each time. I believe this is still a bug, because `datasets` is used exactly this way in `transformers` after `set_seed()` has been called, and I think that using `set_seed()` is a standard procedure to aid reproducibility. I've added more details to reproduce this below.\r\n\r\nHi there! I'm using my own local dataset and custom preprocessing function. My preprocessing function seems to be unpickle-able, perhaps because it is from a closure (will debug this separately). I get this warning, which is expected:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L260-L265\r\n\r\nHowever, what's not expected is that the `datasets` actually _does_ seem to cache and reuse this dataset between runs! After that line, the next thing that's logged looks like:\r\n\r\n```text\r\n Loading cached processed dataset at \/home\/xxx\/.cache\/huggingface\/datasets\/csv\/default-xxx\/0.0.0\/xxx\/cache-xxx.arrow\r\n```\r\n\r\nThe path is exactly the same each run (e.g., last 26 runs).\r\n\r\nThis becomes a problem because I'll pass in the `--max_eval_samples` flag to the HuggingFace example script I'm running off of ([run_swag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/pytorch\/multiple-choice\/run_swag.py)). The fact that the cached dataset is reused means this flag gets ignored. I'll try to load 100 examples, and it will load the full cached 1,000,000.\r\n\r\nI think that\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L248\r\n\r\n... is actually consistent because randomness is being controlled in HuggingFace\/Transformers for reproducibility. I've added a demo of this below.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n# Contents of print_fingerprint.py\r\nfrom transformers import set_seed\r\nfrom datasets.fingerprint import generate_random_fingerprint\r\nset_seed(42)\r\nprint(generate_random_fingerprint())\r\n```\r\n\r\n```bash\r\nfor i in {0..10}; do\r\n python print_fingerprint.py\r\ndone\r\n\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n```\r\n\r\n## Expected results\r\nAfter the \"random hash\" warning is emitted, a random hash is generated, and no outdated cached datasets are reused.\r\n\r\n## Actual results\r\nAfter the \"random hash\" warning is emitted, an identical hash is generated each time, and an outdated cached dataset is reused each run.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.8.0-1038-gcp-x86_64-with-glibc2.31\r\n- Python version: 3.9.6\r\n- PyArrow version: 4.0.1","comment_length":41,"text":"`generate_random_fingerprint()` deterministic with \ud83e\udd17Transformers' `set_seed()`\n## Describe the bug\r\n\r\n**Update:** I dug into this to try to reproduce the underlying issue, and I believe it's that `set_seed()` from the `transformers` library makes the \"random\" fingerprint identical each time. I believe this is still a bug, because `datasets` is used exactly this way in `transformers` after `set_seed()` has been called, and I think that using `set_seed()` is a standard procedure to aid reproducibility. I've added more details to reproduce this below.\r\n\r\nHi there! I'm using my own local dataset and custom preprocessing function. My preprocessing function seems to be unpickle-able, perhaps because it is from a closure (will debug this separately). I get this warning, which is expected:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L260-L265\r\n\r\nHowever, what's not expected is that the `datasets` actually _does_ seem to cache and reuse this dataset between runs! After that line, the next thing that's logged looks like:\r\n\r\n```text\r\n Loading cached processed dataset at \/home\/xxx\/.cache\/huggingface\/datasets\/csv\/default-xxx\/0.0.0\/xxx\/cache-xxx.arrow\r\n```\r\n\r\nThe path is exactly the same each run (e.g., last 26 runs).\r\n\r\nThis becomes a problem because I'll pass in the `--max_eval_samples` flag to the HuggingFace example script I'm running off of ([run_swag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/pytorch\/multiple-choice\/run_swag.py)). The fact that the cached dataset is reused means this flag gets ignored. I'll try to load 100 examples, and it will load the full cached 1,000,000.\r\n\r\nI think that\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L248\r\n\r\n... is actually consistent because randomness is being controlled in HuggingFace\/Transformers for reproducibility. I've added a demo of this below.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n# Contents of print_fingerprint.py\r\nfrom transformers import set_seed\r\nfrom datasets.fingerprint import generate_random_fingerprint\r\nset_seed(42)\r\nprint(generate_random_fingerprint())\r\n```\r\n\r\n```bash\r\nfor i in {0..10}; do\r\n python print_fingerprint.py\r\ndone\r\n\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n```\r\n\r\n## Expected results\r\nAfter the \"random hash\" warning is emitted, a random hash is generated, and no outdated cached datasets are reused.\r\n\r\n## Actual results\r\nAfter the \"random hash\" warning is emitted, an identical hash is generated each time, and an outdated cached dataset is reused each run.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.8.0-1038-gcp-x86_64-with-glibc2.31\r\n- Python version: 3.9.6\r\n- PyArrow version: 4.0.1\nI dug into what I believe is the root of this issue and added a repro in my comment. If this is better addressed as a cross-team issue, let me know and I can open an issue in the Transformers repo","embeddings":[-0.1622430533,-0.0606823638,0.0938825384,0.0836025402,0.3144044876,-0.1429734379,0.6139574647,0.0132132526,-0.1196164414,0.0058159092,0.165034622,0.2017952353,-0.0787890479,0.0945226923,-0.0190071128,0.2130856961,0.0278785881,-0.1402870119,-0.2763957083,-0.1410640776,-0.2284770757,0.152345553,-0.0874382332,-0.1617389768,-0.6726455092,-0.102102913,-0.1329014301,0.1628411114,0.1795904189,-0.4169393778,0.1931263804,0.23521474,-0.1014345139,0.3875524104,-0.0001199422,-0.0937381238,0.3522936404,-0.0650267377,-0.2904088199,-0.1039032489,0.0416674875,0.1874276549,-0.0534762479,-0.1191118434,-0.0600779131,0.0070903306,0.1324010938,-0.4158923626,0.3937298059,-0.1208229214,0.1669470221,0.2793320715,-0.117078729,0.0917676166,0.1264911741,0.700026989,-0.2477301359,-0.0409514718,0.0038184212,0.0352796055,-0.2997260094,0.2550401092,0.1456892937,0.1156733781,0.5079758167,0.1923030913,-0.1786665469,-0.1532367319,0.1148629934,0.2767552435,0.2311255336,-0.3748752773,-0.3369910121,-0.6084536314,-0.2362681925,-0.2321816534,0.1527197361,0.0096535189,-0.1577541232,0.0793170258,-0.3549202085,0.1265327483,0.0707441419,-0.1952131391,0.2468887717,0.221877411,0.0072371704,0.1679401398,0.1587349176,0.0604980215,0.169935748,-0.1181586757,-0.0938741639,0.039496325,0.0180432964,0.1057436615,0.1586054713,0.1534633934,0.3189108968,0.5014269352,0.0610095821,0.1023851931,0.1677440405,-0.073842831,0.2467009127,0.2197239995,-0.058061745,0.3566723764,0.0670858175,0.0370436385,-0.0029727847,0.0142731853,0.2652676702,-0.0426026322,0.501455307,-0.0270120259,0.2397684753,-0.0510366186,0.0650820434,0.1498213857,0.0121367406,-0.0750536099,-0.0779314637,0.2501386106,-0.1982081831,-0.0242463928,-0.0956238508,-0.0449064262,-0.1348835528,-0.0826703608,-0.2837889791,-0.1389882714,-0.2515237331,0.2813723683,-0.0713063851,-0.080778636,0.3410976827,0.2519500256,0.1854823083,-0.1556297839,0.1785332561,-0.0622604117,0.4497914612,0.1312009692,-0.1929613948,0.1577107608,0.1532040387,0.0786590576,-0.1610734612,0.0881237611,0.086571835,-0.2364135236,0.1173688397,0.0525634661,-0.5501559377,0.1222918853,-0.0983982906,-0.0366495587,0.4920062125,0.1749353409,0.2651320398,-0.1899252236,-0.4315011203,-0.291819036,0.072933957,0.7805565596,0.0292768963,-0.2914335132,0.0271981601,0.1518498659,0.178817749,0.1570243537,-0.1241752878,0.2212250829,-0.460788548,0.2205772847,-0.2322400659,-0.1423754692,-0.2322938591,0.3027459085,-0.0936336219,0.383153975,0.038360171,0.0561094806,-0.111499615,-0.1254639775,-0.238865599,0.2245150059,-0.0068208943,0.1098389551,-0.4005638957,-0.3897634745,0.0655296817,-0.0483645014,0.0519645959,0.0094370665,0.1318699718,-0.1288129389,0.2877875865,-0.106851168,0.1581378281,0.1528533399,0.1587304026,0.2246325463,0.0351854488,0.168915838,-0.541993916,0.3820074201,-0.241048187,-0.1893585026,-0.1494206786,-0.3210414052,-0.0753021613,0.0960759893,-0.4060974121,-0.3152598143,-0.0154513624,0.5506615639,0.2708693445,-0.0051996605,0.0968249068,0.2403063625,0.1423283666,-0.0027180747,-0.9128639102,0.0567852408,0.0906823352,-0.1725624204,-0.2306703925,0.0778862163,0.0968683288,-0.2287880331,0.0384364836,0.3598063886,0.1580559313,0.262817502,-0.0984206945,0.1928106546,0.0074188476,-0.1086183786,-0.0394932814,0.1521769762,0.0554043837,-0.1021172255,-0.0983805135,0.4918055832,-0.2128903121,0.1159908548,-0.1526506692,-0.0862316415,-0.0628082529,-0.0525967814,-0.2161844373,-0.1971542835,0.0899825916,-0.0390489846,0.5391991735,0.2282948643,-0.3941819072,0.2312160581,-0.0016536118,-0.0491704121,0.1437010616,0.0852040425,0.0712452978,-0.3131351769,0.0560087487,0.3389393389,0.4275982082,0.0159966983,0.0583637543,0.0512276031,-0.1064069122,-0.166511476,-0.0071328133,0.000592747,-0.0298429485,0.2195874751,0.1412554532,0.0112569081,-0.4387521446,-0.0286888238,0.1356961429,-0.0273100324,-0.3191818893,0.2899666727,-0.1755524725,0.445558995,-0.3127735555,-0.1545071006,-0.2895255387,-0.2514543235,-0.1213659272,0.4109185636,-0.1166299582,0.1950098127,-0.1968613714,0.2971345484,-0.0853328109,-0.2759700119,-0.2541822493,-0.1093628407,-0.2183861136,-0.0602291338,0.120375514,-0.0034396118,0.2321113497,-0.1100810915,-0.0712264851,-0.3098180592,-0.1055286676,0.222975418,-0.1086345837,0.5486388803,0.2477721125,-0.2188455909,-0.1793799549,-0.2130807936,0.2121192366,-0.1938274652,-0.0862900466,-0.1702485383,-0.0922699794,-0.0537805334,0.0295689162,0.0690156519,-0.0948401466,0.0121203233,0.2557660639,-0.0645927638,0.0405180492,0.5897526145,0.0284133237,-0.0574366823,-0.1054536477,-0.0440537892,-0.5146172047,-0.4949686527,0.2611507177,-0.1634007245,-0.2163863778,-0.3258367181,-0.2972288728,0.0188390184,0.1888897568,-0.6068610549,-0.0516650379,-0.2685230672,0.4238205254,-0.2032543421,0.0177746955,0.3344092071,0.0985963792,0.1076801494,-0.0211104527,-0.0474801324,0.2567791939,-0.3898849487,0.3817112148,0.1048754081,0.060253825,0.0892471969,1.0046882629,0.2359312177,-0.3699997962,0.1978331804,-0.0644092187,0.4749623835,-0.297985971,-0.2079370916,0.1310502589,-0.0590401702,-0.2528951168,0.1606173813,-0.07951691,0.2031422853,0.1621236354,0.2374079078,-0.1579894274,-0.363694042,0.1179601401,-0.0183762703,0.0290566292,0.0220860969,0.3120568693,-0.2099297345,-0.1583347768,0.0411260091,0.0356698744,0.4221745431,0.0522913039,-0.3100935221,0.1876141131,-0.6171491146,0.0540949032,0.2187434435,0.2380439937,-0.1098272726,-0.3385640681,0.1094959974,-0.0808582827,0.7635926604,-0.0211593658,0.2726782262,0.0645891279,-0.4564518631,-0.177855581,-0.0194794573,0.0387323424,-0.2118596435,0.0743650422,0.1087059602,-0.1690746844,0.1645171344,-0.1442149431,0.1973522007,-0.1298978031,-0.1945911944,-0.2662622631,-0.3079314828,-0.1788805723,0.1033713222,0.0640035197,-0.0608189479,-0.082603313,0.0196878575,-0.3017717302,-0.2599675655,0.0435260981,0.070662871,0.3443909287,0.0018843182,0.4046282768,0.4547137916,-0.1982673109,0.0894361213,0.4374314845,-0.2148355991,0.008554481,0.220231086,-0.0396816097,-0.233437255,0.5466172099,-0.1877536774,-0.0894529521,0.0587288588,0.0108100008,-0.4439824224,0.0664634183,0.1466405839,0.1137157753,-0.0281976406,-0.2624799609,0.3312913179,0.0272615924,-0.275875181,0.439270556,-0.3099086881,-0.123923704,0.4247116446,0.2471437454,0.8804854751,-0.1272103786,0.3180259168,0.1907135099,0.2768242657,0.1486351341,0.4829757512,0.0115162758,-0.4281357229,-0.3656575084,0.0060816314,-0.2024412155,0.0555919074,0.0869948715,0.0665673837,0.1132726148,-0.0630428717,0.1032283306,-0.1567851007,0.1560461223,0.0246322453,-0.215686366,0.1735640764,0.0843450055,0.1718550324,0.5333346128,-0.069990024,0.0953961015,0.0607641228,-0.2466620356,-0.3635313511,0.208606571,-0.2790142596,0.202284053,0.3691955805,-0.0258358717,-0.2583649755,0.1351563483,0.5098417401,0.1552302986,-0.1020523608,0.1477964669,0.257832408,0.1927008927,-0.1315274686,-0.1301498562,0.2111311555,0.1236370206,-0.1206821948,-0.151934728,-0.0478053614,-0.2716751993,-0.4598906636,-0.0121336933,-0.0979532823,-0.3023136854,0.0790294111,-0.1466923654,0.0272654649,-0.1541841626,0.0286998507,0.2207656801,0.0228120275,0.1299972385,-0.1785639971,-0.4069522619,-0.2265017629,0.4325009286,-0.061210122,0.0727250427,0.2602235377,0.0922238529,-0.0315433852,-0.1085575074,0.2351074517,0.1377259791,-0.5122736096,0.2434004992,-0.2169811875,-0.0195956621,-0.0813552439,0.2316167206,0.1115598679,0.0348231681,-0.0617025979,-0.1523831934,-0.3719450533,-0.1704631299,0.0821972638,0.3519157469,0.2928989828,0.0990412161,-0.105128184,0.312489599,-0.243337661,-0.0502341352,0.0076264087,0.0296813976,0.0670681223,-0.3035534024,0.286791712,-0.0681442842,-0.0063585672,0.2946644425,-0.368863374,-0.1184284538,-0.4529237151,0.1593662649,0.0088854134,-0.0048954943,-0.1571958214,0.177975595,0.0237191357,-0.3966794312,0.3917924166,0.5151258707,0.014996076,-0.0830858946,0.4291324615,0.0537452325,-0.0980246663,-0.1753560305,-0.1844471544,0.0748553947,0.2480771095,0.1079167798,-0.1506762952,0.0071840128,-0.2314375341,0.028811736,0.2302110642,-0.0676556826,0.0180185717,-0.1343324035,-0.268111676,-0.0723344311,0.1703024507,0.161169529,-0.163512826,-0.3935705125,0.1354323775,0.1497947723,-0.0725999027,-0.2825774848,-0.3876026571,0.2493042201,0.0868987143,0.5509284735,0.314143002,-0.1225685328,0.057235755,0.0346190669,0.5128440857,-0.3987120092,0.2396622002,0.3090468943,-0.238540411,0.0252349377,0.4938820601,0.2923137546,0.2063568383,0.1732746661,-0.0384309776,0.5072751641,-0.3027773201,0.0972182304,-0.1392077655,-0.6456997991,0.3082389832,0.4405922592,-0.1516468227,0.4086826742,0.0774330646,0.4832053185,-0.1660363823,-0.0858012736,-0.3003623486,0.3024359941,0.014399589,-0.5353155136,0.0377790593,-0.0834386721,-0.3826372325,0.0900417939,-0.0035493243,-0.2081626803,0.6144252419,-0.0752098262,-0.2515937686,-0.3642039597,-0.0378359221,0.0142612047,0.1197622865,-0.3115850389,0.2505505383,0.1605973095,-0.2969707847,0.101381667,0.1886999458,0.5075894594,0.3388091326,0.1338160038,0.1075945273,-0.2951599956,0.1907003224,-0.2481291443,0.3066643178,-0.2367557883,-0.2121828943,0.2729892135,0.0783088952,-0.1230099425,0.0755241662,0.060490232,0.4505946636,-0.1969412714,0.6952243447,-0.1496911049,0.1883820593,-0.0011686402,0.3480207622,-0.3697022498,0.0585602708,0.262393862,-0.1507827938,0.199068889,-0.1322580129,0.0340269804,-0.048897896,0.3877412081,0.2560419142,-0.3483611047,-0.1792102009,0.1957639307,-0.5142297149,0.3533166647,-0.199963212,0.1005276814,0.0804019272,0.1455357522,-0.2276922911,-0.0096826553,-0.1714702696,0.2559454441,0.0081330761,0.3090419173,-0.156830743,0.0880401358,-0.4392120242,-0.0144750355,0.2130461782,-0.3841431737,0.2073270231,0.1999246478,-0.0297078062,0.0550276823,-0.1730278581,-0.0277811494,-0.0068210261,0.401745975,0.2652513385,0.0799823776,-0.1851325482,-0.1826791316,-0.5771654248,-0.1551558077,-0.1146446019,0.1160615385,0.1047956571,0.2232986093,-0.2068719119,-0.2676874399,-0.4154664278,0.2365451455,0.2599455416,-0.13951011,-0.1362609118,0.2388964891,-0.3561549187,0.328877151,0.0919604897,0.4796586633,-0.02437279,0.2805638015,-0.1349282712,-0.4061133265,0.7216504216,-0.5009364486,-0.2477237582,-0.3238893747,0.0440613255,0.0715888962,-0.4523603022,-0.4823102951,0.1665855348,0.1452373564,0.0978136137,-0.0011369397,-0.1485641301,-0.248302117,0.0275907088,-0.0975219384,0.5400007963,0.1959822625,-0.1516948193,0.3653633893,-0.1505270004]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2775","title":"`generate_random_fingerprint()` deterministic with \ud83e\udd17Transformers' `set_seed()`","comments":"Hi !\r\n\r\nIMO we shouldn't try to modify `set_seed` from transformers but maybe make `datasets` have its own RNG just to generate random fingerprints.\r\n\r\nAny opinion on this @LysandreJik ?","body":"## Describe the bug\r\n\r\n**Update:** I dug into this to try to reproduce the underlying issue, and I believe it's that `set_seed()` from the `transformers` library makes the \"random\" fingerprint identical each time. I believe this is still a bug, because `datasets` is used exactly this way in `transformers` after `set_seed()` has been called, and I think that using `set_seed()` is a standard procedure to aid reproducibility. I've added more details to reproduce this below.\r\n\r\nHi there! I'm using my own local dataset and custom preprocessing function. My preprocessing function seems to be unpickle-able, perhaps because it is from a closure (will debug this separately). I get this warning, which is expected:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L260-L265\r\n\r\nHowever, what's not expected is that the `datasets` actually _does_ seem to cache and reuse this dataset between runs! After that line, the next thing that's logged looks like:\r\n\r\n```text\r\n Loading cached processed dataset at \/home\/xxx\/.cache\/huggingface\/datasets\/csv\/default-xxx\/0.0.0\/xxx\/cache-xxx.arrow\r\n```\r\n\r\nThe path is exactly the same each run (e.g., last 26 runs).\r\n\r\nThis becomes a problem because I'll pass in the `--max_eval_samples` flag to the HuggingFace example script I'm running off of ([run_swag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/pytorch\/multiple-choice\/run_swag.py)). The fact that the cached dataset is reused means this flag gets ignored. I'll try to load 100 examples, and it will load the full cached 1,000,000.\r\n\r\nI think that\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L248\r\n\r\n... is actually consistent because randomness is being controlled in HuggingFace\/Transformers for reproducibility. I've added a demo of this below.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n# Contents of print_fingerprint.py\r\nfrom transformers import set_seed\r\nfrom datasets.fingerprint import generate_random_fingerprint\r\nset_seed(42)\r\nprint(generate_random_fingerprint())\r\n```\r\n\r\n```bash\r\nfor i in {0..10}; do\r\n python print_fingerprint.py\r\ndone\r\n\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n```\r\n\r\n## Expected results\r\nAfter the \"random hash\" warning is emitted, a random hash is generated, and no outdated cached datasets are reused.\r\n\r\n## Actual results\r\nAfter the \"random hash\" warning is emitted, an identical hash is generated each time, and an outdated cached dataset is reused each run.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.8.0-1038-gcp-x86_64-with-glibc2.31\r\n- Python version: 3.9.6\r\n- PyArrow version: 4.0.1","comment_length":30,"text":"`generate_random_fingerprint()` deterministic with \ud83e\udd17Transformers' `set_seed()`\n## Describe the bug\r\n\r\n**Update:** I dug into this to try to reproduce the underlying issue, and I believe it's that `set_seed()` from the `transformers` library makes the \"random\" fingerprint identical each time. I believe this is still a bug, because `datasets` is used exactly this way in `transformers` after `set_seed()` has been called, and I think that using `set_seed()` is a standard procedure to aid reproducibility. I've added more details to reproduce this below.\r\n\r\nHi there! I'm using my own local dataset and custom preprocessing function. My preprocessing function seems to be unpickle-able, perhaps because it is from a closure (will debug this separately). I get this warning, which is expected:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L260-L265\r\n\r\nHowever, what's not expected is that the `datasets` actually _does_ seem to cache and reuse this dataset between runs! After that line, the next thing that's logged looks like:\r\n\r\n```text\r\n Loading cached processed dataset at \/home\/xxx\/.cache\/huggingface\/datasets\/csv\/default-xxx\/0.0.0\/xxx\/cache-xxx.arrow\r\n```\r\n\r\nThe path is exactly the same each run (e.g., last 26 runs).\r\n\r\nThis becomes a problem because I'll pass in the `--max_eval_samples` flag to the HuggingFace example script I'm running off of ([run_swag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/pytorch\/multiple-choice\/run_swag.py)). The fact that the cached dataset is reused means this flag gets ignored. I'll try to load 100 examples, and it will load the full cached 1,000,000.\r\n\r\nI think that\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L248\r\n\r\n... is actually consistent because randomness is being controlled in HuggingFace\/Transformers for reproducibility. I've added a demo of this below.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n# Contents of print_fingerprint.py\r\nfrom transformers import set_seed\r\nfrom datasets.fingerprint import generate_random_fingerprint\r\nset_seed(42)\r\nprint(generate_random_fingerprint())\r\n```\r\n\r\n```bash\r\nfor i in {0..10}; do\r\n python print_fingerprint.py\r\ndone\r\n\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n```\r\n\r\n## Expected results\r\nAfter the \"random hash\" warning is emitted, a random hash is generated, and no outdated cached datasets are reused.\r\n\r\n## Actual results\r\nAfter the \"random hash\" warning is emitted, an identical hash is generated each time, and an outdated cached dataset is reused each run.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.8.0-1038-gcp-x86_64-with-glibc2.31\r\n- Python version: 3.9.6\r\n- PyArrow version: 4.0.1\nHi !\r\n\r\nIMO we shouldn't try to modify `set_seed` from transformers but maybe make `datasets` have its own RNG just to generate random fingerprints.\r\n\r\nAny opinion on this @LysandreJik ?","embeddings":[-0.1622430533,-0.0606823638,0.0938825384,0.0836025402,0.3144044876,-0.1429734379,0.6139574647,0.0132132526,-0.1196164414,0.0058159092,0.165034622,0.2017952353,-0.0787890479,0.0945226923,-0.0190071128,0.2130856961,0.0278785881,-0.1402870119,-0.2763957083,-0.1410640776,-0.2284770757,0.152345553,-0.0874382332,-0.1617389768,-0.6726455092,-0.102102913,-0.1329014301,0.1628411114,0.1795904189,-0.4169393778,0.1931263804,0.23521474,-0.1014345139,0.3875524104,-0.0001199422,-0.0937381238,0.3522936404,-0.0650267377,-0.2904088199,-0.1039032489,0.0416674875,0.1874276549,-0.0534762479,-0.1191118434,-0.0600779131,0.0070903306,0.1324010938,-0.4158923626,0.3937298059,-0.1208229214,0.1669470221,0.2793320715,-0.117078729,0.0917676166,0.1264911741,0.700026989,-0.2477301359,-0.0409514718,0.0038184212,0.0352796055,-0.2997260094,0.2550401092,0.1456892937,0.1156733781,0.5079758167,0.1923030913,-0.1786665469,-0.1532367319,0.1148629934,0.2767552435,0.2311255336,-0.3748752773,-0.3369910121,-0.6084536314,-0.2362681925,-0.2321816534,0.1527197361,0.0096535189,-0.1577541232,0.0793170258,-0.3549202085,0.1265327483,0.0707441419,-0.1952131391,0.2468887717,0.221877411,0.0072371704,0.1679401398,0.1587349176,0.0604980215,0.169935748,-0.1181586757,-0.0938741639,0.039496325,0.0180432964,0.1057436615,0.1586054713,0.1534633934,0.3189108968,0.5014269352,0.0610095821,0.1023851931,0.1677440405,-0.073842831,0.2467009127,0.2197239995,-0.058061745,0.3566723764,0.0670858175,0.0370436385,-0.0029727847,0.0142731853,0.2652676702,-0.0426026322,0.501455307,-0.0270120259,0.2397684753,-0.0510366186,0.0650820434,0.1498213857,0.0121367406,-0.0750536099,-0.0779314637,0.2501386106,-0.1982081831,-0.0242463928,-0.0956238508,-0.0449064262,-0.1348835528,-0.0826703608,-0.2837889791,-0.1389882714,-0.2515237331,0.2813723683,-0.0713063851,-0.080778636,0.3410976827,0.2519500256,0.1854823083,-0.1556297839,0.1785332561,-0.0622604117,0.4497914612,0.1312009692,-0.1929613948,0.1577107608,0.1532040387,0.0786590576,-0.1610734612,0.0881237611,0.086571835,-0.2364135236,0.1173688397,0.0525634661,-0.5501559377,0.1222918853,-0.0983982906,-0.0366495587,0.4920062125,0.1749353409,0.2651320398,-0.1899252236,-0.4315011203,-0.291819036,0.072933957,0.7805565596,0.0292768963,-0.2914335132,0.0271981601,0.1518498659,0.178817749,0.1570243537,-0.1241752878,0.2212250829,-0.460788548,0.2205772847,-0.2322400659,-0.1423754692,-0.2322938591,0.3027459085,-0.0936336219,0.383153975,0.038360171,0.0561094806,-0.111499615,-0.1254639775,-0.238865599,0.2245150059,-0.0068208943,0.1098389551,-0.4005638957,-0.3897634745,0.0655296817,-0.0483645014,0.0519645959,0.0094370665,0.1318699718,-0.1288129389,0.2877875865,-0.106851168,0.1581378281,0.1528533399,0.1587304026,0.2246325463,0.0351854488,0.168915838,-0.541993916,0.3820074201,-0.241048187,-0.1893585026,-0.1494206786,-0.3210414052,-0.0753021613,0.0960759893,-0.4060974121,-0.3152598143,-0.0154513624,0.5506615639,0.2708693445,-0.0051996605,0.0968249068,0.2403063625,0.1423283666,-0.0027180747,-0.9128639102,0.0567852408,0.0906823352,-0.1725624204,-0.2306703925,0.0778862163,0.0968683288,-0.2287880331,0.0384364836,0.3598063886,0.1580559313,0.262817502,-0.0984206945,0.1928106546,0.0074188476,-0.1086183786,-0.0394932814,0.1521769762,0.0554043837,-0.1021172255,-0.0983805135,0.4918055832,-0.2128903121,0.1159908548,-0.1526506692,-0.0862316415,-0.0628082529,-0.0525967814,-0.2161844373,-0.1971542835,0.0899825916,-0.0390489846,0.5391991735,0.2282948643,-0.3941819072,0.2312160581,-0.0016536118,-0.0491704121,0.1437010616,0.0852040425,0.0712452978,-0.3131351769,0.0560087487,0.3389393389,0.4275982082,0.0159966983,0.0583637543,0.0512276031,-0.1064069122,-0.166511476,-0.0071328133,0.000592747,-0.0298429485,0.2195874751,0.1412554532,0.0112569081,-0.4387521446,-0.0286888238,0.1356961429,-0.0273100324,-0.3191818893,0.2899666727,-0.1755524725,0.445558995,-0.3127735555,-0.1545071006,-0.2895255387,-0.2514543235,-0.1213659272,0.4109185636,-0.1166299582,0.1950098127,-0.1968613714,0.2971345484,-0.0853328109,-0.2759700119,-0.2541822493,-0.1093628407,-0.2183861136,-0.0602291338,0.120375514,-0.0034396118,0.2321113497,-0.1100810915,-0.0712264851,-0.3098180592,-0.1055286676,0.222975418,-0.1086345837,0.5486388803,0.2477721125,-0.2188455909,-0.1793799549,-0.2130807936,0.2121192366,-0.1938274652,-0.0862900466,-0.1702485383,-0.0922699794,-0.0537805334,0.0295689162,0.0690156519,-0.0948401466,0.0121203233,0.2557660639,-0.0645927638,0.0405180492,0.5897526145,0.0284133237,-0.0574366823,-0.1054536477,-0.0440537892,-0.5146172047,-0.4949686527,0.2611507177,-0.1634007245,-0.2163863778,-0.3258367181,-0.2972288728,0.0188390184,0.1888897568,-0.6068610549,-0.0516650379,-0.2685230672,0.4238205254,-0.2032543421,0.0177746955,0.3344092071,0.0985963792,0.1076801494,-0.0211104527,-0.0474801324,0.2567791939,-0.3898849487,0.3817112148,0.1048754081,0.060253825,0.0892471969,1.0046882629,0.2359312177,-0.3699997962,0.1978331804,-0.0644092187,0.4749623835,-0.297985971,-0.2079370916,0.1310502589,-0.0590401702,-0.2528951168,0.1606173813,-0.07951691,0.2031422853,0.1621236354,0.2374079078,-0.1579894274,-0.363694042,0.1179601401,-0.0183762703,0.0290566292,0.0220860969,0.3120568693,-0.2099297345,-0.1583347768,0.0411260091,0.0356698744,0.4221745431,0.0522913039,-0.3100935221,0.1876141131,-0.6171491146,0.0540949032,0.2187434435,0.2380439937,-0.1098272726,-0.3385640681,0.1094959974,-0.0808582827,0.7635926604,-0.0211593658,0.2726782262,0.0645891279,-0.4564518631,-0.177855581,-0.0194794573,0.0387323424,-0.2118596435,0.0743650422,0.1087059602,-0.1690746844,0.1645171344,-0.1442149431,0.1973522007,-0.1298978031,-0.1945911944,-0.2662622631,-0.3079314828,-0.1788805723,0.1033713222,0.0640035197,-0.0608189479,-0.082603313,0.0196878575,-0.3017717302,-0.2599675655,0.0435260981,0.070662871,0.3443909287,0.0018843182,0.4046282768,0.4547137916,-0.1982673109,0.0894361213,0.4374314845,-0.2148355991,0.008554481,0.220231086,-0.0396816097,-0.233437255,0.5466172099,-0.1877536774,-0.0894529521,0.0587288588,0.0108100008,-0.4439824224,0.0664634183,0.1466405839,0.1137157753,-0.0281976406,-0.2624799609,0.3312913179,0.0272615924,-0.275875181,0.439270556,-0.3099086881,-0.123923704,0.4247116446,0.2471437454,0.8804854751,-0.1272103786,0.3180259168,0.1907135099,0.2768242657,0.1486351341,0.4829757512,0.0115162758,-0.4281357229,-0.3656575084,0.0060816314,-0.2024412155,0.0555919074,0.0869948715,0.0665673837,0.1132726148,-0.0630428717,0.1032283306,-0.1567851007,0.1560461223,0.0246322453,-0.215686366,0.1735640764,0.0843450055,0.1718550324,0.5333346128,-0.069990024,0.0953961015,0.0607641228,-0.2466620356,-0.3635313511,0.208606571,-0.2790142596,0.202284053,0.3691955805,-0.0258358717,-0.2583649755,0.1351563483,0.5098417401,0.1552302986,-0.1020523608,0.1477964669,0.257832408,0.1927008927,-0.1315274686,-0.1301498562,0.2111311555,0.1236370206,-0.1206821948,-0.151934728,-0.0478053614,-0.2716751993,-0.4598906636,-0.0121336933,-0.0979532823,-0.3023136854,0.0790294111,-0.1466923654,0.0272654649,-0.1541841626,0.0286998507,0.2207656801,0.0228120275,0.1299972385,-0.1785639971,-0.4069522619,-0.2265017629,0.4325009286,-0.061210122,0.0727250427,0.2602235377,0.0922238529,-0.0315433852,-0.1085575074,0.2351074517,0.1377259791,-0.5122736096,0.2434004992,-0.2169811875,-0.0195956621,-0.0813552439,0.2316167206,0.1115598679,0.0348231681,-0.0617025979,-0.1523831934,-0.3719450533,-0.1704631299,0.0821972638,0.3519157469,0.2928989828,0.0990412161,-0.105128184,0.312489599,-0.243337661,-0.0502341352,0.0076264087,0.0296813976,0.0670681223,-0.3035534024,0.286791712,-0.0681442842,-0.0063585672,0.2946644425,-0.368863374,-0.1184284538,-0.4529237151,0.1593662649,0.0088854134,-0.0048954943,-0.1571958214,0.177975595,0.0237191357,-0.3966794312,0.3917924166,0.5151258707,0.014996076,-0.0830858946,0.4291324615,0.0537452325,-0.0980246663,-0.1753560305,-0.1844471544,0.0748553947,0.2480771095,0.1079167798,-0.1506762952,0.0071840128,-0.2314375341,0.028811736,0.2302110642,-0.0676556826,0.0180185717,-0.1343324035,-0.268111676,-0.0723344311,0.1703024507,0.161169529,-0.163512826,-0.3935705125,0.1354323775,0.1497947723,-0.0725999027,-0.2825774848,-0.3876026571,0.2493042201,0.0868987143,0.5509284735,0.314143002,-0.1225685328,0.057235755,0.0346190669,0.5128440857,-0.3987120092,0.2396622002,0.3090468943,-0.238540411,0.0252349377,0.4938820601,0.2923137546,0.2063568383,0.1732746661,-0.0384309776,0.5072751641,-0.3027773201,0.0972182304,-0.1392077655,-0.6456997991,0.3082389832,0.4405922592,-0.1516468227,0.4086826742,0.0774330646,0.4832053185,-0.1660363823,-0.0858012736,-0.3003623486,0.3024359941,0.014399589,-0.5353155136,0.0377790593,-0.0834386721,-0.3826372325,0.0900417939,-0.0035493243,-0.2081626803,0.6144252419,-0.0752098262,-0.2515937686,-0.3642039597,-0.0378359221,0.0142612047,0.1197622865,-0.3115850389,0.2505505383,0.1605973095,-0.2969707847,0.101381667,0.1886999458,0.5075894594,0.3388091326,0.1338160038,0.1075945273,-0.2951599956,0.1907003224,-0.2481291443,0.3066643178,-0.2367557883,-0.2121828943,0.2729892135,0.0783088952,-0.1230099425,0.0755241662,0.060490232,0.4505946636,-0.1969412714,0.6952243447,-0.1496911049,0.1883820593,-0.0011686402,0.3480207622,-0.3697022498,0.0585602708,0.262393862,-0.1507827938,0.199068889,-0.1322580129,0.0340269804,-0.048897896,0.3877412081,0.2560419142,-0.3483611047,-0.1792102009,0.1957639307,-0.5142297149,0.3533166647,-0.199963212,0.1005276814,0.0804019272,0.1455357522,-0.2276922911,-0.0096826553,-0.1714702696,0.2559454441,0.0081330761,0.3090419173,-0.156830743,0.0880401358,-0.4392120242,-0.0144750355,0.2130461782,-0.3841431737,0.2073270231,0.1999246478,-0.0297078062,0.0550276823,-0.1730278581,-0.0277811494,-0.0068210261,0.401745975,0.2652513385,0.0799823776,-0.1851325482,-0.1826791316,-0.5771654248,-0.1551558077,-0.1146446019,0.1160615385,0.1047956571,0.2232986093,-0.2068719119,-0.2676874399,-0.4154664278,0.2365451455,0.2599455416,-0.13951011,-0.1362609118,0.2388964891,-0.3561549187,0.328877151,0.0919604897,0.4796586633,-0.02437279,0.2805638015,-0.1349282712,-0.4061133265,0.7216504216,-0.5009364486,-0.2477237582,-0.3238893747,0.0440613255,0.0715888962,-0.4523603022,-0.4823102951,0.1665855348,0.1452373564,0.0978136137,-0.0011369397,-0.1485641301,-0.248302117,0.0275907088,-0.0975219384,0.5400007963,0.1959822625,-0.1516948193,0.3653633893,-0.1505270004]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2775","title":"`generate_random_fingerprint()` deterministic with \ud83e\udd17Transformers' `set_seed()`","comments":"Yes, this sounds good @lhoestq ","body":"## Describe the bug\r\n\r\n**Update:** I dug into this to try to reproduce the underlying issue, and I believe it's that `set_seed()` from the `transformers` library makes the \"random\" fingerprint identical each time. I believe this is still a bug, because `datasets` is used exactly this way in `transformers` after `set_seed()` has been called, and I think that using `set_seed()` is a standard procedure to aid reproducibility. I've added more details to reproduce this below.\r\n\r\nHi there! I'm using my own local dataset and custom preprocessing function. My preprocessing function seems to be unpickle-able, perhaps because it is from a closure (will debug this separately). I get this warning, which is expected:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L260-L265\r\n\r\nHowever, what's not expected is that the `datasets` actually _does_ seem to cache and reuse this dataset between runs! After that line, the next thing that's logged looks like:\r\n\r\n```text\r\n Loading cached processed dataset at \/home\/xxx\/.cache\/huggingface\/datasets\/csv\/default-xxx\/0.0.0\/xxx\/cache-xxx.arrow\r\n```\r\n\r\nThe path is exactly the same each run (e.g., last 26 runs).\r\n\r\nThis becomes a problem because I'll pass in the `--max_eval_samples` flag to the HuggingFace example script I'm running off of ([run_swag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/pytorch\/multiple-choice\/run_swag.py)). The fact that the cached dataset is reused means this flag gets ignored. I'll try to load 100 examples, and it will load the full cached 1,000,000.\r\n\r\nI think that\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L248\r\n\r\n... is actually consistent because randomness is being controlled in HuggingFace\/Transformers for reproducibility. I've added a demo of this below.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n# Contents of print_fingerprint.py\r\nfrom transformers import set_seed\r\nfrom datasets.fingerprint import generate_random_fingerprint\r\nset_seed(42)\r\nprint(generate_random_fingerprint())\r\n```\r\n\r\n```bash\r\nfor i in {0..10}; do\r\n python print_fingerprint.py\r\ndone\r\n\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n```\r\n\r\n## Expected results\r\nAfter the \"random hash\" warning is emitted, a random hash is generated, and no outdated cached datasets are reused.\r\n\r\n## Actual results\r\nAfter the \"random hash\" warning is emitted, an identical hash is generated each time, and an outdated cached dataset is reused each run.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.8.0-1038-gcp-x86_64-with-glibc2.31\r\n- Python version: 3.9.6\r\n- PyArrow version: 4.0.1","comment_length":5,"text":"`generate_random_fingerprint()` deterministic with \ud83e\udd17Transformers' `set_seed()`\n## Describe the bug\r\n\r\n**Update:** I dug into this to try to reproduce the underlying issue, and I believe it's that `set_seed()` from the `transformers` library makes the \"random\" fingerprint identical each time. I believe this is still a bug, because `datasets` is used exactly this way in `transformers` after `set_seed()` has been called, and I think that using `set_seed()` is a standard procedure to aid reproducibility. I've added more details to reproduce this below.\r\n\r\nHi there! I'm using my own local dataset and custom preprocessing function. My preprocessing function seems to be unpickle-able, perhaps because it is from a closure (will debug this separately). I get this warning, which is expected:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L260-L265\r\n\r\nHowever, what's not expected is that the `datasets` actually _does_ seem to cache and reuse this dataset between runs! After that line, the next thing that's logged looks like:\r\n\r\n```text\r\n Loading cached processed dataset at \/home\/xxx\/.cache\/huggingface\/datasets\/csv\/default-xxx\/0.0.0\/xxx\/cache-xxx.arrow\r\n```\r\n\r\nThe path is exactly the same each run (e.g., last 26 runs).\r\n\r\nThis becomes a problem because I'll pass in the `--max_eval_samples` flag to the HuggingFace example script I'm running off of ([run_swag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/pytorch\/multiple-choice\/run_swag.py)). The fact that the cached dataset is reused means this flag gets ignored. I'll try to load 100 examples, and it will load the full cached 1,000,000.\r\n\r\nI think that\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/450b9174765374111e5c6daab0ed294bc3d9b639\/src\/datasets\/fingerprint.py#L248\r\n\r\n... is actually consistent because randomness is being controlled in HuggingFace\/Transformers for reproducibility. I've added a demo of this below.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\n# Contents of print_fingerprint.py\r\nfrom transformers import set_seed\r\nfrom datasets.fingerprint import generate_random_fingerprint\r\nset_seed(42)\r\nprint(generate_random_fingerprint())\r\n```\r\n\r\n```bash\r\nfor i in {0..10}; do\r\n python print_fingerprint.py\r\ndone\r\n\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n1c80317fa3b1799d\r\n```\r\n\r\n## Expected results\r\nAfter the \"random hash\" warning is emitted, a random hash is generated, and no outdated cached datasets are reused.\r\n\r\n## Actual results\r\nAfter the \"random hash\" warning is emitted, an identical hash is generated each time, and an outdated cached dataset is reused each run.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.8.0-1038-gcp-x86_64-with-glibc2.31\r\n- Python version: 3.9.6\r\n- PyArrow version: 4.0.1\nYes, this sounds good @lhoestq ","embeddings":[-0.1622430533,-0.0606823638,0.0938825384,0.0836025402,0.3144044876,-0.1429734379,0.6139574647,0.0132132526,-0.1196164414,0.0058159092,0.165034622,0.2017952353,-0.0787890479,0.0945226923,-0.0190071128,0.2130856961,0.0278785881,-0.1402870119,-0.2763957083,-0.1410640776,-0.2284770757,0.152345553,-0.0874382332,-0.1617389768,-0.6726455092,-0.102102913,-0.1329014301,0.1628411114,0.1795904189,-0.4169393778,0.1931263804,0.23521474,-0.1014345139,0.3875524104,-0.0001199422,-0.0937381238,0.3522936404,-0.0650267377,-0.2904088199,-0.1039032489,0.0416674875,0.1874276549,-0.0534762479,-0.1191118434,-0.0600779131,0.0070903306,0.1324010938,-0.4158923626,0.3937298059,-0.1208229214,0.1669470221,0.2793320715,-0.117078729,0.0917676166,0.1264911741,0.700026989,-0.2477301359,-0.0409514718,0.0038184212,0.0352796055,-0.2997260094,0.2550401092,0.1456892937,0.1156733781,0.5079758167,0.1923030913,-0.1786665469,-0.1532367319,0.1148629934,0.2767552435,0.2311255336,-0.3748752773,-0.3369910121,-0.6084536314,-0.2362681925,-0.2321816534,0.1527197361,0.0096535189,-0.1577541232,0.0793170258,-0.3549202085,0.1265327483,0.0707441419,-0.1952131391,0.2468887717,0.221877411,0.0072371704,0.1679401398,0.1587349176,0.0604980215,0.169935748,-0.1181586757,-0.0938741639,0.039496325,0.0180432964,0.1057436615,0.1586054713,0.1534633934,0.3189108968,0.5014269352,0.0610095821,0.1023851931,0.1677440405,-0.073842831,0.2467009127,0.2197239995,-0.058061745,0.3566723764,0.0670858175,0.0370436385,-0.0029727847,0.0142731853,0.2652676702,-0.0426026322,0.501455307,-0.0270120259,0.2397684753,-0.0510366186,0.0650820434,0.1498213857,0.0121367406,-0.0750536099,-0.0779314637,0.2501386106,-0.1982081831,-0.0242463928,-0.0956238508,-0.0449064262,-0.1348835528,-0.0826703608,-0.2837889791,-0.1389882714,-0.2515237331,0.2813723683,-0.0713063851,-0.080778636,0.3410976827,0.2519500256,0.1854823083,-0.1556297839,0.1785332561,-0.0622604117,0.4497914612,0.1312009692,-0.1929613948,0.1577107608,0.1532040387,0.0786590576,-0.1610734612,0.0881237611,0.086571835,-0.2364135236,0.1173688397,0.0525634661,-0.5501559377,0.1222918853,-0.0983982906,-0.0366495587,0.4920062125,0.1749353409,0.2651320398,-0.1899252236,-0.4315011203,-0.291819036,0.072933957,0.7805565596,0.0292768963,-0.2914335132,0.0271981601,0.1518498659,0.178817749,0.1570243537,-0.1241752878,0.2212250829,-0.460788548,0.2205772847,-0.2322400659,-0.1423754692,-0.2322938591,0.3027459085,-0.0936336219,0.383153975,0.038360171,0.0561094806,-0.111499615,-0.1254639775,-0.238865599,0.2245150059,-0.0068208943,0.1098389551,-0.4005638957,-0.3897634745,0.0655296817,-0.0483645014,0.0519645959,0.0094370665,0.1318699718,-0.1288129389,0.2877875865,-0.106851168,0.1581378281,0.1528533399,0.1587304026,0.2246325463,0.0351854488,0.168915838,-0.541993916,0.3820074201,-0.241048187,-0.1893585026,-0.1494206786,-0.3210414052,-0.0753021613,0.0960759893,-0.4060974121,-0.3152598143,-0.0154513624,0.5506615639,0.2708693445,-0.0051996605,0.0968249068,0.2403063625,0.1423283666,-0.0027180747,-0.9128639102,0.0567852408,0.0906823352,-0.1725624204,-0.2306703925,0.0778862163,0.0968683288,-0.2287880331,0.0384364836,0.3598063886,0.1580559313,0.262817502,-0.0984206945,0.1928106546,0.0074188476,-0.1086183786,-0.0394932814,0.1521769762,0.0554043837,-0.1021172255,-0.0983805135,0.4918055832,-0.2128903121,0.1159908548,-0.1526506692,-0.0862316415,-0.0628082529,-0.0525967814,-0.2161844373,-0.1971542835,0.0899825916,-0.0390489846,0.5391991735,0.2282948643,-0.3941819072,0.2312160581,-0.0016536118,-0.0491704121,0.1437010616,0.0852040425,0.0712452978,-0.3131351769,0.0560087487,0.3389393389,0.4275982082,0.0159966983,0.0583637543,0.0512276031,-0.1064069122,-0.166511476,-0.0071328133,0.000592747,-0.0298429485,0.2195874751,0.1412554532,0.0112569081,-0.4387521446,-0.0286888238,0.1356961429,-0.0273100324,-0.3191818893,0.2899666727,-0.1755524725,0.445558995,-0.3127735555,-0.1545071006,-0.2895255387,-0.2514543235,-0.1213659272,0.4109185636,-0.1166299582,0.1950098127,-0.1968613714,0.2971345484,-0.0853328109,-0.2759700119,-0.2541822493,-0.1093628407,-0.2183861136,-0.0602291338,0.120375514,-0.0034396118,0.2321113497,-0.1100810915,-0.0712264851,-0.3098180592,-0.1055286676,0.222975418,-0.1086345837,0.5486388803,0.2477721125,-0.2188455909,-0.1793799549,-0.2130807936,0.2121192366,-0.1938274652,-0.0862900466,-0.1702485383,-0.0922699794,-0.0537805334,0.0295689162,0.0690156519,-0.0948401466,0.0121203233,0.2557660639,-0.0645927638,0.0405180492,0.5897526145,0.0284133237,-0.0574366823,-0.1054536477,-0.0440537892,-0.5146172047,-0.4949686527,0.2611507177,-0.1634007245,-0.2163863778,-0.3258367181,-0.2972288728,0.0188390184,0.1888897568,-0.6068610549,-0.0516650379,-0.2685230672,0.4238205254,-0.2032543421,0.0177746955,0.3344092071,0.0985963792,0.1076801494,-0.0211104527,-0.0474801324,0.2567791939,-0.3898849487,0.3817112148,0.1048754081,0.060253825,0.0892471969,1.0046882629,0.2359312177,-0.3699997962,0.1978331804,-0.0644092187,0.4749623835,-0.297985971,-0.2079370916,0.1310502589,-0.0590401702,-0.2528951168,0.1606173813,-0.07951691,0.2031422853,0.1621236354,0.2374079078,-0.1579894274,-0.363694042,0.1179601401,-0.0183762703,0.0290566292,0.0220860969,0.3120568693,-0.2099297345,-0.1583347768,0.0411260091,0.0356698744,0.4221745431,0.0522913039,-0.3100935221,0.1876141131,-0.6171491146,0.0540949032,0.2187434435,0.2380439937,-0.1098272726,-0.3385640681,0.1094959974,-0.0808582827,0.7635926604,-0.0211593658,0.2726782262,0.0645891279,-0.4564518631,-0.177855581,-0.0194794573,0.0387323424,-0.2118596435,0.0743650422,0.1087059602,-0.1690746844,0.1645171344,-0.1442149431,0.1973522007,-0.1298978031,-0.1945911944,-0.2662622631,-0.3079314828,-0.1788805723,0.1033713222,0.0640035197,-0.0608189479,-0.082603313,0.0196878575,-0.3017717302,-0.2599675655,0.0435260981,0.070662871,0.3443909287,0.0018843182,0.4046282768,0.4547137916,-0.1982673109,0.0894361213,0.4374314845,-0.2148355991,0.008554481,0.220231086,-0.0396816097,-0.233437255,0.5466172099,-0.1877536774,-0.0894529521,0.0587288588,0.0108100008,-0.4439824224,0.0664634183,0.1466405839,0.1137157753,-0.0281976406,-0.2624799609,0.3312913179,0.0272615924,-0.275875181,0.439270556,-0.3099086881,-0.123923704,0.4247116446,0.2471437454,0.8804854751,-0.1272103786,0.3180259168,0.1907135099,0.2768242657,0.1486351341,0.4829757512,0.0115162758,-0.4281357229,-0.3656575084,0.0060816314,-0.2024412155,0.0555919074,0.0869948715,0.0665673837,0.1132726148,-0.0630428717,0.1032283306,-0.1567851007,0.1560461223,0.0246322453,-0.215686366,0.1735640764,0.0843450055,0.1718550324,0.5333346128,-0.069990024,0.0953961015,0.0607641228,-0.2466620356,-0.3635313511,0.208606571,-0.2790142596,0.202284053,0.3691955805,-0.0258358717,-0.2583649755,0.1351563483,0.5098417401,0.1552302986,-0.1020523608,0.1477964669,0.257832408,0.1927008927,-0.1315274686,-0.1301498562,0.2111311555,0.1236370206,-0.1206821948,-0.151934728,-0.0478053614,-0.2716751993,-0.4598906636,-0.0121336933,-0.0979532823,-0.3023136854,0.0790294111,-0.1466923654,0.0272654649,-0.1541841626,0.0286998507,0.2207656801,0.0228120275,0.1299972385,-0.1785639971,-0.4069522619,-0.2265017629,0.4325009286,-0.061210122,0.0727250427,0.2602235377,0.0922238529,-0.0315433852,-0.1085575074,0.2351074517,0.1377259791,-0.5122736096,0.2434004992,-0.2169811875,-0.0195956621,-0.0813552439,0.2316167206,0.1115598679,0.0348231681,-0.0617025979,-0.1523831934,-0.3719450533,-0.1704631299,0.0821972638,0.3519157469,0.2928989828,0.0990412161,-0.105128184,0.312489599,-0.243337661,-0.0502341352,0.0076264087,0.0296813976,0.0670681223,-0.3035534024,0.286791712,-0.0681442842,-0.0063585672,0.2946644425,-0.368863374,-0.1184284538,-0.4529237151,0.1593662649,0.0088854134,-0.0048954943,-0.1571958214,0.177975595,0.0237191357,-0.3966794312,0.3917924166,0.5151258707,0.014996076,-0.0830858946,0.4291324615,0.0537452325,-0.0980246663,-0.1753560305,-0.1844471544,0.0748553947,0.2480771095,0.1079167798,-0.1506762952,0.0071840128,-0.2314375341,0.028811736,0.2302110642,-0.0676556826,0.0180185717,-0.1343324035,-0.268111676,-0.0723344311,0.1703024507,0.161169529,-0.163512826,-0.3935705125,0.1354323775,0.1497947723,-0.0725999027,-0.2825774848,-0.3876026571,0.2493042201,0.0868987143,0.5509284735,0.314143002,-0.1225685328,0.057235755,0.0346190669,0.5128440857,-0.3987120092,0.2396622002,0.3090468943,-0.238540411,0.0252349377,0.4938820601,0.2923137546,0.2063568383,0.1732746661,-0.0384309776,0.5072751641,-0.3027773201,0.0972182304,-0.1392077655,-0.6456997991,0.3082389832,0.4405922592,-0.1516468227,0.4086826742,0.0774330646,0.4832053185,-0.1660363823,-0.0858012736,-0.3003623486,0.3024359941,0.014399589,-0.5353155136,0.0377790593,-0.0834386721,-0.3826372325,0.0900417939,-0.0035493243,-0.2081626803,0.6144252419,-0.0752098262,-0.2515937686,-0.3642039597,-0.0378359221,0.0142612047,0.1197622865,-0.3115850389,0.2505505383,0.1605973095,-0.2969707847,0.101381667,0.1886999458,0.5075894594,0.3388091326,0.1338160038,0.1075945273,-0.2951599956,0.1907003224,-0.2481291443,0.3066643178,-0.2367557883,-0.2121828943,0.2729892135,0.0783088952,-0.1230099425,0.0755241662,0.060490232,0.4505946636,-0.1969412714,0.6952243447,-0.1496911049,0.1883820593,-0.0011686402,0.3480207622,-0.3697022498,0.0585602708,0.262393862,-0.1507827938,0.199068889,-0.1322580129,0.0340269804,-0.048897896,0.3877412081,0.2560419142,-0.3483611047,-0.1792102009,0.1957639307,-0.5142297149,0.3533166647,-0.199963212,0.1005276814,0.0804019272,0.1455357522,-0.2276922911,-0.0096826553,-0.1714702696,0.2559454441,0.0081330761,0.3090419173,-0.156830743,0.0880401358,-0.4392120242,-0.0144750355,0.2130461782,-0.3841431737,0.2073270231,0.1999246478,-0.0297078062,0.0550276823,-0.1730278581,-0.0277811494,-0.0068210261,0.401745975,0.2652513385,0.0799823776,-0.1851325482,-0.1826791316,-0.5771654248,-0.1551558077,-0.1146446019,0.1160615385,0.1047956571,0.2232986093,-0.2068719119,-0.2676874399,-0.4154664278,0.2365451455,0.2599455416,-0.13951011,-0.1362609118,0.2388964891,-0.3561549187,0.328877151,0.0919604897,0.4796586633,-0.02437279,0.2805638015,-0.1349282712,-0.4061133265,0.7216504216,-0.5009364486,-0.2477237582,-0.3238893747,0.0440613255,0.0715888962,-0.4523603022,-0.4823102951,0.1665855348,0.1452373564,0.0978136137,-0.0011369397,-0.1485641301,-0.248302117,0.0275907088,-0.0975219384,0.5400007963,0.1959822625,-0.1516948193,0.3653633893,-0.1505270004]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2768","title":"`ArrowInvalid: Added column's length must match table's length.` after using `select`","comments":"Hi,\r\n\r\nthe `select` method creates an indices mapping and doesn't modify the underlying PyArrow table by default for better performance. To modify the underlying table after the `select` call, call `flatten_indices` on the dataset object as follows:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\nds = ds.flatten_indices()\r\nds = ds.add_column('ones', [1]*128)\r\n```","body":"## Describe the bug\r\nI would like to add a column to a downsampled dataset. However I get an error message saying the length don't match with the length of the unsampled dataset indicated. I suspect that the dataset size is not updated when calling `select`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\nds = ds.add_column('ones', [1]*128)\r\n```\r\n\r\n## Expected results\r\nI would expect a new column named `ones` filled with `1`. When I check the length of `ds` it says `128`. Interestingly, it works when calling `ds = ds.map(lambda x: x)` before adding the column.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```python\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n\/var\/folders\/l4\/2905jygx4tx5jv8_kn03vxsw0000gn\/T\/ipykernel_6301\/868709636.py in \r\n 1 from datasets import load_dataset\r\n 2 ds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\n----> 3 ds = ds.add_column('ones', [0]*128)\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 183 }\r\n 184 # apply actual function\r\n--> 185 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 186 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 187 # re-apply format to the output\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in add_column(self, name, column, new_fingerprint)\r\n 2965 column_table = InMemoryTable.from_pydict({name: column})\r\n 2966 # Concatenate tables horizontally\r\n-> 2967 table = ConcatenationTable.from_tables([self._data, column_table], axis=1)\r\n 2968 # Update features\r\n 2969 info = self.info.copy()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in from_tables(cls, tables, axis)\r\n 715 table_blocks = to_blocks(table)\r\n 716 blocks = _extend_blocks(blocks, table_blocks, axis=axis)\r\n--> 717 return cls.from_blocks(blocks)\r\n 718 \r\n 719 @property\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in from_blocks(cls, blocks)\r\n 663 return cls(table, blocks)\r\n 664 else:\r\n--> 665 table = cls._concat_blocks_horizontally_and_vertically(blocks)\r\n 666 return cls(table, blocks)\r\n 667 \r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in _concat_blocks_horizontally_and_vertically(cls, blocks)\r\n 623 if not tables:\r\n 624 continue\r\n--> 625 pa_table_horizontally_concatenated = cls._concat_blocks(tables, axis=1)\r\n 626 pa_tables_to_concat_vertically.append(pa_table_horizontally_concatenated)\r\n 627 return cls._concat_blocks(pa_tables_to_concat_vertically, axis=0)\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in _concat_blocks(blocks, axis)\r\n 612 else:\r\n 613 for name, col in zip(table.column_names, table.columns):\r\n--> 614 pa_table = pa_table.append_column(name, col)\r\n 615 return pa_table\r\n 616 else:\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.append_column()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.add_column()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Added column's length must match table's length. Expected length 31962 but got length 128\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS-10.16-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n","comment_length":53,"text":"`ArrowInvalid: Added column's length must match table's length.` after using `select`\n## Describe the bug\r\nI would like to add a column to a downsampled dataset. However I get an error message saying the length don't match with the length of the unsampled dataset indicated. I suspect that the dataset size is not updated when calling `select`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\nds = ds.add_column('ones', [1]*128)\r\n```\r\n\r\n## Expected results\r\nI would expect a new column named `ones` filled with `1`. When I check the length of `ds` it says `128`. Interestingly, it works when calling `ds = ds.map(lambda x: x)` before adding the column.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```python\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n\/var\/folders\/l4\/2905jygx4tx5jv8_kn03vxsw0000gn\/T\/ipykernel_6301\/868709636.py in \r\n 1 from datasets import load_dataset\r\n 2 ds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\n----> 3 ds = ds.add_column('ones', [0]*128)\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 183 }\r\n 184 # apply actual function\r\n--> 185 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 186 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 187 # re-apply format to the output\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in add_column(self, name, column, new_fingerprint)\r\n 2965 column_table = InMemoryTable.from_pydict({name: column})\r\n 2966 # Concatenate tables horizontally\r\n-> 2967 table = ConcatenationTable.from_tables([self._data, column_table], axis=1)\r\n 2968 # Update features\r\n 2969 info = self.info.copy()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in from_tables(cls, tables, axis)\r\n 715 table_blocks = to_blocks(table)\r\n 716 blocks = _extend_blocks(blocks, table_blocks, axis=axis)\r\n--> 717 return cls.from_blocks(blocks)\r\n 718 \r\n 719 @property\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in from_blocks(cls, blocks)\r\n 663 return cls(table, blocks)\r\n 664 else:\r\n--> 665 table = cls._concat_blocks_horizontally_and_vertically(blocks)\r\n 666 return cls(table, blocks)\r\n 667 \r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in _concat_blocks_horizontally_and_vertically(cls, blocks)\r\n 623 if not tables:\r\n 624 continue\r\n--> 625 pa_table_horizontally_concatenated = cls._concat_blocks(tables, axis=1)\r\n 626 pa_tables_to_concat_vertically.append(pa_table_horizontally_concatenated)\r\n 627 return cls._concat_blocks(pa_tables_to_concat_vertically, axis=0)\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in _concat_blocks(blocks, axis)\r\n 612 else:\r\n 613 for name, col in zip(table.column_names, table.columns):\r\n--> 614 pa_table = pa_table.append_column(name, col)\r\n 615 return pa_table\r\n 616 else:\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.append_column()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.add_column()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Added column's length must match table's length. Expected length 31962 but got length 128\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS-10.16-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n\nHi,\r\n\r\nthe `select` method creates an indices mapping and doesn't modify the underlying PyArrow table by default for better performance. To modify the underlying table after the `select` call, call `flatten_indices` on the dataset object as follows:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\nds = ds.flatten_indices()\r\nds = ds.add_column('ones', [1]*128)\r\n```","embeddings":[-0.2274188101,-0.1898306608,-0.0089051146,0.0325093158,0.0982829332,0.0415741466,0.0728505924,0.1837980449,0.0604875386,0.164753437,0.2048401088,0.6761353016,0.0139961764,-0.3138830364,0.0691086799,-0.2314466685,0.0780239403,0.1225910485,-0.3308289647,-0.0652318522,-0.3497133255,0.0132752853,-0.3354176581,-0.3151179254,-0.35253039,-0.1137229875,-0.1477196962,-0.076647304,-0.3253406882,-0.3277154565,0.081593357,-0.3840053678,0.055132024,0.2379106879,-0.0001264378,-0.1121125668,0.2349809557,-0.1064973325,-0.2713165581,0.0904975608,-0.1300846487,-0.3753260374,0.1473688334,-0.1593483537,0.1864854842,-0.1655192822,-0.6239047647,-0.3172651231,-0.0759102702,0.1652894318,0.161454007,0.2517942488,0.0120389545,-0.2805069983,-0.0970271379,0.1960537583,-0.2620949745,0.3448806703,0.060268797,-0.0941637009,0.3812075257,0.3098258078,0.2030231357,-0.0177224353,0.1855849028,-0.0706433281,-0.0906358734,-0.0210477207,0.4785783589,-0.0847314149,0.7143198848,-0.2248954773,-0.2491259128,-0.1734179258,0.4096648395,-0.3938191831,-0.0271732341,-0.1862057149,0.2897660434,0.0206637345,-0.1405392736,0.0848093703,0.1604089737,0.1899269074,-0.0069546569,0.5650565028,0.2024693191,0.3704702258,-0.192640841,-0.1704318523,0.0081759496,0.0942436606,-0.2345474809,0.3636321723,-0.3872019649,-0.2049816102,-0.2336306423,-0.2155933231,0.1200502291,0.3310541511,-0.1377871931,-0.2782112956,0.3323386014,-0.2003704607,0.2972340584,0.3154058158,0.0448100008,0.606928587,-0.3473197818,-0.2117927074,0.3857125938,0.0074288007,0.1638017446,-0.2288597971,0.3544318676,-0.2207029164,0.1014282778,0.0487019904,-0.218203485,0.2133063972,-0.1640473157,-0.0807819366,-0.1688757539,0.1329697818,0.1167246401,0.090414986,0.1235700473,0.3109091222,-0.1185306236,-0.0169276446,-0.2082299143,-0.0679130778,-0.0348821245,-0.0294247828,-0.0645576492,0.3572279811,0.2836541533,-0.0721221864,-0.2275483906,-0.2286396921,-0.1471447945,-0.0540051572,0.1518873572,0.2009018809,-0.2485596836,0.0838380605,-0.0692494884,-0.3781675398,-0.1850917041,0.3647522926,-0.2289008796,-0.2965632975,-0.6092182994,0.0443059616,0.0293099396,-0.1413225532,0.1089118347,0.268869698,0.4819933474,-0.1184430197,0.4236471653,-0.3621317446,-0.3157013059,-0.1269381344,0.151450038,0.2686063945,-0.8413779736,-0.065250054,0.0082486253,0.2485842705,0.1492479891,0.2491782606,-0.0317912586,-0.1156833395,0.1023319066,0.2750084698,0.444047004,0.2298650444,-0.8025736809,-0.1254057884,-0.4141768813,0.1022443846,-0.119038485,0.1956773549,0.0839703605,0.2012964487,0.3524823189,0.095638141,-0.0832370371,-0.1414693147,-0.3857779205,-0.1018519029,0.2032636553,-0.1336515397,0.0435380042,-0.0106330048,0.3127953708,-0.0638056919,0.3429020643,0.3277736902,0.2038774341,0.3033575714,0.2194056958,-0.0338491276,0.2668210864,-0.2382794619,-0.2710856199,0.0977192894,0.1223185733,-0.1519711763,-0.1563115418,-0.240345329,-0.7025293708,0.2183455229,0.1364157349,0.005338741,-0.0692306608,-0.0590435602,0.218089819,-0.1341995001,-0.0326836184,-0.0548646972,-0.3558381498,0.1471114308,-0.1049330831,0.2896157801,0.0534480587,-0.3693044186,-0.2833218575,0.1731369942,0.1423828155,0.1439751387,-0.3318285346,0.5284386277,-0.02505772,-0.1870957315,-0.0710229203,-0.3655413687,0.1858620644,-0.0491627306,0.2788014412,-0.0748981237,0.1783838868,0.0045647947,0.011803302,-0.0678275377,-0.2379916459,0.2784359455,-0.3470553756,-0.1087519154,0.0590355732,0.1315242648,-0.1129913703,-0.3320047855,-0.2876385152,0.0122851627,-0.0364267863,-0.0151267387,-0.229577601,0.060114406,0.2146398723,-0.0264009442,0.1114841998,0.1653437465,-0.2672156692,0.1005395725,0.3662687838,0.2415162772,0.7434340119,0.1991255879,0.093924351,-0.0243315827,-0.1202091128,-0.0079243649,0.046591796,-0.0483894534,0.0148665728,0.2591744661,0.2337099463,-0.1083140671,-0.0253790636,0.3794060349,0.1265302747,0.1463618428,-0.3006027639,0.0304230396,-0.3964523673,0.1090054885,-0.1138181686,-0.1068870127,-0.04723268,-0.167769134,0.0350237526,0.0338079929,-0.2729739249,0.275567174,-0.4887412786,0.1287105381,0.2696689665,-0.11721275,0.1353976578,0.1072930694,-0.1601349413,-0.0256597102,-0.3478003442,0.0767987147,0.1265491843,0.3528582156,-0.0809799954,-0.1846956015,-0.6639268398,-0.0905888379,-0.1658451408,0.1347124577,0.1048660129,0.1017715037,-0.1720081121,-0.1064303219,0.1442970484,0.2994417846,0.0071018846,0.0552978106,0.0515188649,-0.1211503819,-0.1023785919,-0.4178765714,-0.2243611962,-0.2537752986,0.254355371,0.1677352786,-0.0982267484,-0.1927160919,0.4928369224,-0.3301513791,0.0152297141,-0.0172036998,-0.3184478879,0.36839661,0.4220629632,-0.0695865527,-0.1449093372,0.0337172411,-0.2432203293,-0.0239139777,0.0918513238,-0.1472073197,0.093353346,-0.1383074075,0.1208327562,-0.032450594,0.2197681218,0.5915466547,-0.0773879215,0.1023480073,-0.1245045438,-0.4013107717,0.2273926437,0.2349444479,0.5312868953,-0.1336196959,0.4529953897,-0.0000410322,0.2885741591,0.0917197391,-0.2183245718,-0.0316433199,-0.2098070681,0.1782779098,-0.3310292363,0.0301831886,-0.2201043218,-0.0530982092,-0.0377524719,-0.0310935639,0.0038577351,-0.2196581662,0.2738206685,0.006902128,0.0954334736,0.13206397,0.0047286232,-0.1869534701,-0.0142247137,0.2118297219,0.2793754339,-0.2753639519,0.033407405,-0.1057039648,-0.2404872328,-0.0680912137,-0.0263268817,-0.4333118498,-0.0451960526,-0.2838226259,0.3899294138,0.1940863729,0.54674685,0.1754097641,-0.5557872653,0.1311583966,0.2818832994,0.2410366982,-0.3274645209,-0.4251196086,0.0470249802,0.1985604763,-0.4720601141,-0.1163353547,0.0849930868,-0.0265570357,0.0530414805,0.1417142898,-0.1842363924,0.3784877658,0.5383037329,0.2780979276,-0.1354471892,-0.174876824,-0.2113422304,-0.1154458001,0.0383576415,0.2666169405,0.0506884083,0.1190377176,0.334777981,-0.0588509105,-0.217922166,-0.4106576443,-0.2320038676,0.1078915447,0.5795092583,0.1309526712,0.343416363,-0.1727490723,0.1780847311,0.1462423056,0.6778901815,-0.3160172701,-0.1287690848,0.1859256476,-0.2585476339,0.285241425,0.3564641774,0.1048220843,0.4273752868,-0.0976130515,0.1615415961,-0.1236966774,-0.3316584229,0.3483012319,-0.036491584,0.1592893302,-0.1429035068,0.5838422179,0.055739563,-0.0640686005,0.5355708003,0.3855295777,-0.0477329008,0.178987056,0.2290593386,1.0119824409,0.398732692,-0.2096713781,0.1493921429,-0.6083494425,0.1385125369,0.1710830629,0.0415348597,-0.1949031204,-0.2967480719,0.1062929258,-0.0384661742,0.0778827444,-0.0383502841,-0.0741887763,-0.022515092,-0.4111111462,-0.0608634241,-0.0994131342,0.0356370509,-0.2062990814,-0.1018423364,0.0768304914,0.014451935,0.1541109532,0.1061988473,-0.0259004571,0.2839572728,-0.3806120157,-0.5120562911,-0.0994393602,0.0156706218,0.1089029536,0.0575731285,0.0102587184,-0.4089047015,0.1461764127,-0.1549652517,-0.0874759331,0.133685708,0.1438664049,0.3423863351,0.2724767327,0.2239227146,0.1456787735,0.1136241034,0.0948991403,0.2152918577,-0.2369085848,0.0916554704,-0.0856527761,-0.1643978357,-0.2890014052,0.140919596,-0.0327984244,-0.5026828051,-0.3177187145,0.1072989404,0.1824521124,-0.1735684574,0.0370034799,-0.2570529282,0.012125059,0.2574302554,0.1259057671,-0.320546329,-0.1250503808,0.7137432098,0.2430962622,0.0370569788,0.5456565022,-0.0620369837,-0.2005740851,0.0578546934,-0.2416698039,0.1941676289,-0.4723762274,0.0387659967,-0.3199873567,0.1019138321,0.0100216297,0.0781955346,0.2685832977,0.2327012122,-0.3625876307,0.1678191125,-0.3429454863,0.3639068604,0.1460609585,0.3142026663,-0.1675263643,-0.2009896785,-0.0601713657,0.3776643574,-0.169047609,-0.2857521474,0.0747615173,0.2211058885,0.1200480834,0.261272192,-0.2349202335,-0.1687364578,0.0742569268,0.4132665396,0.1892871559,-0.0425816476,-0.0177383237,0.1136992425,0.1303786635,0.1159602925,-0.194168821,-0.3900858164,-0.0220029783,-0.073395282,-0.0131697385,0.043819204,-0.1892604828,-0.168611899,0.1748149097,-0.2541188896,0.0460724011,0.3447584808,-0.1099528894,0.1676507443,-0.0650224537,0.1859346628,0.2655059397,0.1638247669,-0.354803741,0.0101155164,-0.2249089777,-0.0252966601,0.2344941795,-0.2191301584,0.1044161245,-0.0908749998,0.2551119924,0.2529684007,-0.0010135361,-0.3300299942,0.0061342614,0.0920823365,-0.3814071715,-0.0281085372,-0.3035818636,0.2432120293,0.0974200964,0.3086265624,0.0931272879,0.1900091767,-0.326125294,0.1838027239,0.2700856328,-0.1610836536,-0.0497952513,0.8157376647,0.0274299849,-0.0618518181,0.3706049323,0.2224710733,0.2035601437,0.060965471,0.0033504188,0.4796197116,0.2929706573,0.0817171708,-0.1237499118,0.0599611104,0.1983148754,0.0828077421,0.0365374982,-0.0929830149,0.411836952,-0.0237150416,0.0725683495,-0.3503089845,-0.084490113,0.075655967,-0.1859807372,0.0007158036,-0.6281936765,-0.1837962121,-0.3110828698,-0.3594712019,-0.1047758311,-0.0567477718,0.7304555178,-0.2808421254,0.0407896116,-0.258308053,-0.2098976225,0.0031496177,0.0761209428,-0.312835753,0.2475763708,0.2272415012,-0.0591459647,0.332310617,0.0456520617,0.3254354,0.1328353137,0.2016155869,-0.0929412022,0.2371743768,-0.2348897606,0.2276574969,0.3966606855,0.1293011755,-0.1632917374,0.1752874851,0.0534346253,-0.0952450112,0.2036039382,0.1524476111,0.4200280011,-0.3563271463,0.1430589259,-0.1308436096,-0.33670488,-0.0797395185,-0.047032699,-0.0371607617,0.086217992,0.0980572253,-0.0910399333,0.0292178206,-0.1615636796,0.0033695088,-0.1287699789,0.1948776245,0.214407295,-0.2921981215,0.0550785251,-0.1844687313,-0.2614018023,-0.0671689883,-0.5196347833,-0.5024821758,0.0391684882,0.2446741015,0.201341182,-0.0925038978,0.4790179431,0.2301711291,0.0132905981,0.4627120793,-0.0188116543,-0.2972557247,-0.1463599205,0.0459298454,0.0431483723,-0.1846254766,0.04131395,0.2230617553,-0.1242952943,-0.0850944668,-0.1141512617,0.1872731745,0.2204159051,0.4074783623,0.0221222211,0.5489421487,0.1126407534,0.2246771753,-0.2629415989,-0.3424998224,-0.2419401556,0.290606916,0.0979023352,0.3872912228,0.0580057316,0.0127178384,0.1250769198,0.1995161027,-0.3370289505,0.1343454421,-0.2346699536,-0.0472320057,-0.1939323097,0.0765421987,0.1249756813,0.050142277,0.2796938717,0.3591469228,0.1647763401,-0.4475569129,0.4196082652,0.0277057178,-0.2497125417,-0.0874990299,0.0219654422,0.3229506612,-0.1594501436,-0.2185301036,-0.1609862149,0.304910332,-0.1044651344,0.3495863974,0.1398440003,-0.0841187909,0.1383666545,-0.2282846421,0.5186856389,-0.2141585946,0.1019174904,-0.203262791,-0.342454493]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2768","title":"`ArrowInvalid: Added column's length must match table's length.` after using `select`","comments":"Thanks for the question @lvwerra. And thanks for the answer @mariosasko. ^^","body":"## Describe the bug\r\nI would like to add a column to a downsampled dataset. However I get an error message saying the length don't match with the length of the unsampled dataset indicated. I suspect that the dataset size is not updated when calling `select`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\nds = ds.add_column('ones', [1]*128)\r\n```\r\n\r\n## Expected results\r\nI would expect a new column named `ones` filled with `1`. When I check the length of `ds` it says `128`. Interestingly, it works when calling `ds = ds.map(lambda x: x)` before adding the column.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```python\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n\/var\/folders\/l4\/2905jygx4tx5jv8_kn03vxsw0000gn\/T\/ipykernel_6301\/868709636.py in \r\n 1 from datasets import load_dataset\r\n 2 ds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\n----> 3 ds = ds.add_column('ones', [0]*128)\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 183 }\r\n 184 # apply actual function\r\n--> 185 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 186 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 187 # re-apply format to the output\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in add_column(self, name, column, new_fingerprint)\r\n 2965 column_table = InMemoryTable.from_pydict({name: column})\r\n 2966 # Concatenate tables horizontally\r\n-> 2967 table = ConcatenationTable.from_tables([self._data, column_table], axis=1)\r\n 2968 # Update features\r\n 2969 info = self.info.copy()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in from_tables(cls, tables, axis)\r\n 715 table_blocks = to_blocks(table)\r\n 716 blocks = _extend_blocks(blocks, table_blocks, axis=axis)\r\n--> 717 return cls.from_blocks(blocks)\r\n 718 \r\n 719 @property\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in from_blocks(cls, blocks)\r\n 663 return cls(table, blocks)\r\n 664 else:\r\n--> 665 table = cls._concat_blocks_horizontally_and_vertically(blocks)\r\n 666 return cls(table, blocks)\r\n 667 \r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in _concat_blocks_horizontally_and_vertically(cls, blocks)\r\n 623 if not tables:\r\n 624 continue\r\n--> 625 pa_table_horizontally_concatenated = cls._concat_blocks(tables, axis=1)\r\n 626 pa_tables_to_concat_vertically.append(pa_table_horizontally_concatenated)\r\n 627 return cls._concat_blocks(pa_tables_to_concat_vertically, axis=0)\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in _concat_blocks(blocks, axis)\r\n 612 else:\r\n 613 for name, col in zip(table.column_names, table.columns):\r\n--> 614 pa_table = pa_table.append_column(name, col)\r\n 615 return pa_table\r\n 616 else:\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.append_column()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.add_column()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Added column's length must match table's length. Expected length 31962 but got length 128\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS-10.16-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n","comment_length":12,"text":"`ArrowInvalid: Added column's length must match table's length.` after using `select`\n## Describe the bug\r\nI would like to add a column to a downsampled dataset. However I get an error message saying the length don't match with the length of the unsampled dataset indicated. I suspect that the dataset size is not updated when calling `select`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\nds = ds.add_column('ones', [1]*128)\r\n```\r\n\r\n## Expected results\r\nI would expect a new column named `ones` filled with `1`. When I check the length of `ds` it says `128`. Interestingly, it works when calling `ds = ds.map(lambda x: x)` before adding the column.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```python\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n\/var\/folders\/l4\/2905jygx4tx5jv8_kn03vxsw0000gn\/T\/ipykernel_6301\/868709636.py in \r\n 1 from datasets import load_dataset\r\n 2 ds = load_dataset(\"tweets_hate_speech_detection\")['train'].select(range(128))\r\n----> 3 ds = ds.add_column('ones', [0]*128)\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 183 }\r\n 184 # apply actual function\r\n--> 185 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 186 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 187 # re-apply format to the output\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in add_column(self, name, column, new_fingerprint)\r\n 2965 column_table = InMemoryTable.from_pydict({name: column})\r\n 2966 # Concatenate tables horizontally\r\n-> 2967 table = ConcatenationTable.from_tables([self._data, column_table], axis=1)\r\n 2968 # Update features\r\n 2969 info = self.info.copy()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in from_tables(cls, tables, axis)\r\n 715 table_blocks = to_blocks(table)\r\n 716 blocks = _extend_blocks(blocks, table_blocks, axis=axis)\r\n--> 717 return cls.from_blocks(blocks)\r\n 718 \r\n 719 @property\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in from_blocks(cls, blocks)\r\n 663 return cls(table, blocks)\r\n 664 else:\r\n--> 665 table = cls._concat_blocks_horizontally_and_vertically(blocks)\r\n 666 return cls(table, blocks)\r\n 667 \r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in _concat_blocks_horizontally_and_vertically(cls, blocks)\r\n 623 if not tables:\r\n 624 continue\r\n--> 625 pa_table_horizontally_concatenated = cls._concat_blocks(tables, axis=1)\r\n 626 pa_tables_to_concat_vertically.append(pa_table_horizontally_concatenated)\r\n 627 return cls._concat_blocks(pa_tables_to_concat_vertically, axis=0)\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/datasets\/table.py in _concat_blocks(blocks, axis)\r\n 612 else:\r\n 613 for name, col in zip(table.column_names, table.columns):\r\n--> 614 pa_table = pa_table.append_column(name, col)\r\n 615 return pa_table\r\n 616 else:\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.append_column()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.add_column()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/git\/semantic-clustering\/env\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Added column's length must match table's length. Expected length 31962 but got length 128\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS-10.16-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 5.0.0\r\n\nThanks for the question @lvwerra. And thanks for the answer @mariosasko. ^^","embeddings":[-0.2274188101,-0.1898306608,-0.0089051146,0.0325093158,0.0982829332,0.0415741466,0.0728505924,0.1837980449,0.0604875386,0.164753437,0.2048401088,0.6761353016,0.0139961764,-0.3138830364,0.0691086799,-0.2314466685,0.0780239403,0.1225910485,-0.3308289647,-0.0652318522,-0.3497133255,0.0132752853,-0.3354176581,-0.3151179254,-0.35253039,-0.1137229875,-0.1477196962,-0.076647304,-0.3253406882,-0.3277154565,0.081593357,-0.3840053678,0.055132024,0.2379106879,-0.0001264378,-0.1121125668,0.2349809557,-0.1064973325,-0.2713165581,0.0904975608,-0.1300846487,-0.3753260374,0.1473688334,-0.1593483537,0.1864854842,-0.1655192822,-0.6239047647,-0.3172651231,-0.0759102702,0.1652894318,0.161454007,0.2517942488,0.0120389545,-0.2805069983,-0.0970271379,0.1960537583,-0.2620949745,0.3448806703,0.060268797,-0.0941637009,0.3812075257,0.3098258078,0.2030231357,-0.0177224353,0.1855849028,-0.0706433281,-0.0906358734,-0.0210477207,0.4785783589,-0.0847314149,0.7143198848,-0.2248954773,-0.2491259128,-0.1734179258,0.4096648395,-0.3938191831,-0.0271732341,-0.1862057149,0.2897660434,0.0206637345,-0.1405392736,0.0848093703,0.1604089737,0.1899269074,-0.0069546569,0.5650565028,0.2024693191,0.3704702258,-0.192640841,-0.1704318523,0.0081759496,0.0942436606,-0.2345474809,0.3636321723,-0.3872019649,-0.2049816102,-0.2336306423,-0.2155933231,0.1200502291,0.3310541511,-0.1377871931,-0.2782112956,0.3323386014,-0.2003704607,0.2972340584,0.3154058158,0.0448100008,0.606928587,-0.3473197818,-0.2117927074,0.3857125938,0.0074288007,0.1638017446,-0.2288597971,0.3544318676,-0.2207029164,0.1014282778,0.0487019904,-0.218203485,0.2133063972,-0.1640473157,-0.0807819366,-0.1688757539,0.1329697818,0.1167246401,0.090414986,0.1235700473,0.3109091222,-0.1185306236,-0.0169276446,-0.2082299143,-0.0679130778,-0.0348821245,-0.0294247828,-0.0645576492,0.3572279811,0.2836541533,-0.0721221864,-0.2275483906,-0.2286396921,-0.1471447945,-0.0540051572,0.1518873572,0.2009018809,-0.2485596836,0.0838380605,-0.0692494884,-0.3781675398,-0.1850917041,0.3647522926,-0.2289008796,-0.2965632975,-0.6092182994,0.0443059616,0.0293099396,-0.1413225532,0.1089118347,0.268869698,0.4819933474,-0.1184430197,0.4236471653,-0.3621317446,-0.3157013059,-0.1269381344,0.151450038,0.2686063945,-0.8413779736,-0.065250054,0.0082486253,0.2485842705,0.1492479891,0.2491782606,-0.0317912586,-0.1156833395,0.1023319066,0.2750084698,0.444047004,0.2298650444,-0.8025736809,-0.1254057884,-0.4141768813,0.1022443846,-0.119038485,0.1956773549,0.0839703605,0.2012964487,0.3524823189,0.095638141,-0.0832370371,-0.1414693147,-0.3857779205,-0.1018519029,0.2032636553,-0.1336515397,0.0435380042,-0.0106330048,0.3127953708,-0.0638056919,0.3429020643,0.3277736902,0.2038774341,0.3033575714,0.2194056958,-0.0338491276,0.2668210864,-0.2382794619,-0.2710856199,0.0977192894,0.1223185733,-0.1519711763,-0.1563115418,-0.240345329,-0.7025293708,0.2183455229,0.1364157349,0.005338741,-0.0692306608,-0.0590435602,0.218089819,-0.1341995001,-0.0326836184,-0.0548646972,-0.3558381498,0.1471114308,-0.1049330831,0.2896157801,0.0534480587,-0.3693044186,-0.2833218575,0.1731369942,0.1423828155,0.1439751387,-0.3318285346,0.5284386277,-0.02505772,-0.1870957315,-0.0710229203,-0.3655413687,0.1858620644,-0.0491627306,0.2788014412,-0.0748981237,0.1783838868,0.0045647947,0.011803302,-0.0678275377,-0.2379916459,0.2784359455,-0.3470553756,-0.1087519154,0.0590355732,0.1315242648,-0.1129913703,-0.3320047855,-0.2876385152,0.0122851627,-0.0364267863,-0.0151267387,-0.229577601,0.060114406,0.2146398723,-0.0264009442,0.1114841998,0.1653437465,-0.2672156692,0.1005395725,0.3662687838,0.2415162772,0.7434340119,0.1991255879,0.093924351,-0.0243315827,-0.1202091128,-0.0079243649,0.046591796,-0.0483894534,0.0148665728,0.2591744661,0.2337099463,-0.1083140671,-0.0253790636,0.3794060349,0.1265302747,0.1463618428,-0.3006027639,0.0304230396,-0.3964523673,0.1090054885,-0.1138181686,-0.1068870127,-0.04723268,-0.167769134,0.0350237526,0.0338079929,-0.2729739249,0.275567174,-0.4887412786,0.1287105381,0.2696689665,-0.11721275,0.1353976578,0.1072930694,-0.1601349413,-0.0256597102,-0.3478003442,0.0767987147,0.1265491843,0.3528582156,-0.0809799954,-0.1846956015,-0.6639268398,-0.0905888379,-0.1658451408,0.1347124577,0.1048660129,0.1017715037,-0.1720081121,-0.1064303219,0.1442970484,0.2994417846,0.0071018846,0.0552978106,0.0515188649,-0.1211503819,-0.1023785919,-0.4178765714,-0.2243611962,-0.2537752986,0.254355371,0.1677352786,-0.0982267484,-0.1927160919,0.4928369224,-0.3301513791,0.0152297141,-0.0172036998,-0.3184478879,0.36839661,0.4220629632,-0.0695865527,-0.1449093372,0.0337172411,-0.2432203293,-0.0239139777,0.0918513238,-0.1472073197,0.093353346,-0.1383074075,0.1208327562,-0.032450594,0.2197681218,0.5915466547,-0.0773879215,0.1023480073,-0.1245045438,-0.4013107717,0.2273926437,0.2349444479,0.5312868953,-0.1336196959,0.4529953897,-0.0000410322,0.2885741591,0.0917197391,-0.2183245718,-0.0316433199,-0.2098070681,0.1782779098,-0.3310292363,0.0301831886,-0.2201043218,-0.0530982092,-0.0377524719,-0.0310935639,0.0038577351,-0.2196581662,0.2738206685,0.006902128,0.0954334736,0.13206397,0.0047286232,-0.1869534701,-0.0142247137,0.2118297219,0.2793754339,-0.2753639519,0.033407405,-0.1057039648,-0.2404872328,-0.0680912137,-0.0263268817,-0.4333118498,-0.0451960526,-0.2838226259,0.3899294138,0.1940863729,0.54674685,0.1754097641,-0.5557872653,0.1311583966,0.2818832994,0.2410366982,-0.3274645209,-0.4251196086,0.0470249802,0.1985604763,-0.4720601141,-0.1163353547,0.0849930868,-0.0265570357,0.0530414805,0.1417142898,-0.1842363924,0.3784877658,0.5383037329,0.2780979276,-0.1354471892,-0.174876824,-0.2113422304,-0.1154458001,0.0383576415,0.2666169405,0.0506884083,0.1190377176,0.334777981,-0.0588509105,-0.217922166,-0.4106576443,-0.2320038676,0.1078915447,0.5795092583,0.1309526712,0.343416363,-0.1727490723,0.1780847311,0.1462423056,0.6778901815,-0.3160172701,-0.1287690848,0.1859256476,-0.2585476339,0.285241425,0.3564641774,0.1048220843,0.4273752868,-0.0976130515,0.1615415961,-0.1236966774,-0.3316584229,0.3483012319,-0.036491584,0.1592893302,-0.1429035068,0.5838422179,0.055739563,-0.0640686005,0.5355708003,0.3855295777,-0.0477329008,0.178987056,0.2290593386,1.0119824409,0.398732692,-0.2096713781,0.1493921429,-0.6083494425,0.1385125369,0.1710830629,0.0415348597,-0.1949031204,-0.2967480719,0.1062929258,-0.0384661742,0.0778827444,-0.0383502841,-0.0741887763,-0.022515092,-0.4111111462,-0.0608634241,-0.0994131342,0.0356370509,-0.2062990814,-0.1018423364,0.0768304914,0.014451935,0.1541109532,0.1061988473,-0.0259004571,0.2839572728,-0.3806120157,-0.5120562911,-0.0994393602,0.0156706218,0.1089029536,0.0575731285,0.0102587184,-0.4089047015,0.1461764127,-0.1549652517,-0.0874759331,0.133685708,0.1438664049,0.3423863351,0.2724767327,0.2239227146,0.1456787735,0.1136241034,0.0948991403,0.2152918577,-0.2369085848,0.0916554704,-0.0856527761,-0.1643978357,-0.2890014052,0.140919596,-0.0327984244,-0.5026828051,-0.3177187145,0.1072989404,0.1824521124,-0.1735684574,0.0370034799,-0.2570529282,0.012125059,0.2574302554,0.1259057671,-0.320546329,-0.1250503808,0.7137432098,0.2430962622,0.0370569788,0.5456565022,-0.0620369837,-0.2005740851,0.0578546934,-0.2416698039,0.1941676289,-0.4723762274,0.0387659967,-0.3199873567,0.1019138321,0.0100216297,0.0781955346,0.2685832977,0.2327012122,-0.3625876307,0.1678191125,-0.3429454863,0.3639068604,0.1460609585,0.3142026663,-0.1675263643,-0.2009896785,-0.0601713657,0.3776643574,-0.169047609,-0.2857521474,0.0747615173,0.2211058885,0.1200480834,0.261272192,-0.2349202335,-0.1687364578,0.0742569268,0.4132665396,0.1892871559,-0.0425816476,-0.0177383237,0.1136992425,0.1303786635,0.1159602925,-0.194168821,-0.3900858164,-0.0220029783,-0.073395282,-0.0131697385,0.043819204,-0.1892604828,-0.168611899,0.1748149097,-0.2541188896,0.0460724011,0.3447584808,-0.1099528894,0.1676507443,-0.0650224537,0.1859346628,0.2655059397,0.1638247669,-0.354803741,0.0101155164,-0.2249089777,-0.0252966601,0.2344941795,-0.2191301584,0.1044161245,-0.0908749998,0.2551119924,0.2529684007,-0.0010135361,-0.3300299942,0.0061342614,0.0920823365,-0.3814071715,-0.0281085372,-0.3035818636,0.2432120293,0.0974200964,0.3086265624,0.0931272879,0.1900091767,-0.326125294,0.1838027239,0.2700856328,-0.1610836536,-0.0497952513,0.8157376647,0.0274299849,-0.0618518181,0.3706049323,0.2224710733,0.2035601437,0.060965471,0.0033504188,0.4796197116,0.2929706573,0.0817171708,-0.1237499118,0.0599611104,0.1983148754,0.0828077421,0.0365374982,-0.0929830149,0.411836952,-0.0237150416,0.0725683495,-0.3503089845,-0.084490113,0.075655967,-0.1859807372,0.0007158036,-0.6281936765,-0.1837962121,-0.3110828698,-0.3594712019,-0.1047758311,-0.0567477718,0.7304555178,-0.2808421254,0.0407896116,-0.258308053,-0.2098976225,0.0031496177,0.0761209428,-0.312835753,0.2475763708,0.2272415012,-0.0591459647,0.332310617,0.0456520617,0.3254354,0.1328353137,0.2016155869,-0.0929412022,0.2371743768,-0.2348897606,0.2276574969,0.3966606855,0.1293011755,-0.1632917374,0.1752874851,0.0534346253,-0.0952450112,0.2036039382,0.1524476111,0.4200280011,-0.3563271463,0.1430589259,-0.1308436096,-0.33670488,-0.0797395185,-0.047032699,-0.0371607617,0.086217992,0.0980572253,-0.0910399333,0.0292178206,-0.1615636796,0.0033695088,-0.1287699789,0.1948776245,0.214407295,-0.2921981215,0.0550785251,-0.1844687313,-0.2614018023,-0.0671689883,-0.5196347833,-0.5024821758,0.0391684882,0.2446741015,0.201341182,-0.0925038978,0.4790179431,0.2301711291,0.0132905981,0.4627120793,-0.0188116543,-0.2972557247,-0.1463599205,0.0459298454,0.0431483723,-0.1846254766,0.04131395,0.2230617553,-0.1242952943,-0.0850944668,-0.1141512617,0.1872731745,0.2204159051,0.4074783623,0.0221222211,0.5489421487,0.1126407534,0.2246771753,-0.2629415989,-0.3424998224,-0.2419401556,0.290606916,0.0979023352,0.3872912228,0.0580057316,0.0127178384,0.1250769198,0.1995161027,-0.3370289505,0.1343454421,-0.2346699536,-0.0472320057,-0.1939323097,0.0765421987,0.1249756813,0.050142277,0.2796938717,0.3591469228,0.1647763401,-0.4475569129,0.4196082652,0.0277057178,-0.2497125417,-0.0874990299,0.0219654422,0.3229506612,-0.1594501436,-0.2185301036,-0.1609862149,0.304910332,-0.1044651344,0.3495863974,0.1398440003,-0.0841187909,0.1383666545,-0.2282846421,0.5186856389,-0.2141585946,0.1019174904,-0.203262791,-0.342454493]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2767","title":"equal operation to perform unbatch for huggingface datasets ","comments":"Hi @lhoestq \r\nMaybe this is clearer to explain like this, currently map function, map one example to \"one\" modified one, lets assume we want to map one example to \"multiple\" examples, in which we do not know in advance how many examples they would be per each entry. I greatly appreciate telling me how I can handle this operation, thanks a lot","body":"Hi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n","comment_length":62,"text":"equal operation to perform unbatch for huggingface datasets \nHi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n\nHi @lhoestq \r\nMaybe this is clearer to explain like this, currently map function, map one example to \"one\" modified one, lets assume we want to map one example to \"multiple\" examples, in which we do not know in advance how many examples they would be per each entry. I greatly appreciate telling me how I can handle this operation, thanks a lot","embeddings":[-0.0286772158,-0.7930662632,0.0335285924,-0.0818505213,-0.0049160868,-0.1033486947,0.2304782718,-0.0015914276,0.4284729958,0.2395398617,-0.324056536,-0.045957718,0.0236226693,0.4699031711,0.1768155247,-0.234804824,0.0796521008,0.1120863855,-0.2576453686,-0.0738277659,-0.1641778648,-0.0214592591,-0.5562424064,-0.1872487962,0.3530071676,-0.3256388605,-0.1314136684,-0.2317972332,-0.1461296082,0.315525502,0.3914791048,0.3159993589,0.1450021267,0.4359943569,-0.000124996,0.2683015168,-0.2683853507,-0.160119608,0.124135524,-0.1457855105,-0.1318628341,-0.0308620147,-0.086130254,-0.2909473479,-0.1553762704,0.0247164331,0.0838053226,-0.3370779455,0.6689295173,0.1163539812,0.0033910642,0.2023021877,-0.0828164369,0.0687010959,-0.0968771428,0.0643510893,0.0033202854,0.0944168419,0.0819834173,-0.0196641665,0.0691599995,0.3062365949,-0.0207522139,-0.0487893857,0.0508214012,0.07096719,0.0294626951,-0.4244825244,0.0995178819,0.3601127863,0.1911817938,-0.3335200548,-0.5461017489,-0.0424865298,0.0593188368,-0.0293313079,-0.2361653447,0.157397002,-0.1050995514,0.2068406343,-0.4860775769,-0.1199070066,-0.0401413664,0.0794255883,0.1040407047,-0.0883648247,-0.3136317432,0.304833591,0.0857107341,0.0686647147,-0.1580198109,-0.1046706513,0.1547605395,0.2256595939,-0.6143533587,-0.2198613882,-0.2562040687,-0.018030664,0.3166288137,0.3536277115,0.5462899804,-0.0548854396,-0.0812086761,0.1295626163,0.0867581517,0.5366975069,-0.0814011171,-0.2418981045,-0.0120940544,0.1348869503,0.0322757177,-0.0191909596,0.3165329695,-0.0755865127,-0.364189893,0.1535307169,0.3737836182,-0.0768215284,-0.3276385069,-0.2054927647,-0.355368495,-0.1100299507,0.013368411,0.149778232,-0.1751033515,0.0324243531,0.1829235554,0.0582386367,-0.2670663595,-0.0648584366,-0.15654625,0.2346510738,-0.1623347849,0.0303167943,-0.0835497528,-0.1439877152,0.0143885342,0.4878042042,-0.0132796234,-0.0494600348,0.2268330753,-0.1175174713,0.3311557174,-0.012914151,0.1702439636,-0.1079322323,0.1320791095,0.048105862,-0.047323171,-0.020109592,-0.1124603525,-0.251969099,0.0791813955,-0.0342014916,-0.280818522,0.0415980518,-0.2271566242,0.3425572515,-0.3150840104,0.3320280612,0.4480588734,0.3958907723,-0.0926601663,-0.08121676,0.3662144542,0.0135225635,-0.0299308486,-0.2975441515,0.3306922317,-0.0206188466,0.2846172154,0.2693359554,-0.0772321746,0.2560566664,-0.2479436398,-0.1416988522,0.2170896083,-0.1550867856,-0.422665,0.2669496834,-0.1116820797,0.3041907847,-0.0293439664,0.0453690514,-0.0606513955,0.1691665053,-0.0211248212,0.2345538884,-0.0976088271,-0.0916917697,-0.0077355248,-0.1554212123,0.3678473532,-0.1237492189,-0.1123933941,0.3373726308,0.1141549349,-0.0311683882,0.2400601208,-0.4657922089,0.2133105695,0.2559942603,0.2300509512,0.1441592425,0.2475076914,-0.162230745,-0.3966977596,0.1127589867,-0.3874371946,0.1097859219,0.1357524693,-0.3187190294,-0.1147378087,-0.2083285451,-0.1158750355,-0.0275504105,-0.0913064778,-0.07691627,0.4252487719,0.0141551876,-0.4040580094,0.2884075344,-0.1211187318,0.1116678566,-0.4417361319,0.0506424047,0.2840945125,0.0014731303,-0.0312433951,0.169615671,-0.0143173551,-0.3851283193,0.0938997865,0.3167589009,-0.2716920376,-0.0050706319,-0.0730890334,0.0714602321,0.0649735183,-0.1072305143,-0.1467594951,0.1161518022,0.0738630369,-0.0635086298,-0.1146612242,0.755053401,-0.0631501153,0.1573880762,0.1018702015,-0.0422718599,-0.1553642303,-0.2558642626,-0.1365821511,0.2526302338,0.1801749468,-0.2043041736,0.2563439608,-0.1358107626,-0.227570802,-0.2907540798,0.2738241255,0.0077716927,0.3027178347,0.1412793249,0.0636070445,0.1619970351,0.0168047398,-0.1653858274,0.3542358279,-0.0300219264,0.2275405973,0.3308396637,0.2882860005,0.1496148556,0.0311066695,-0.2790497243,-0.040879406,-0.0289257523,0.0171885472,0.0865658298,-0.3360079825,-0.0112243909,0.0899158865,0.1401436925,-0.0238192286,-0.0793158188,-0.1206796318,-0.1589755714,-0.198860988,0.0308305621,-0.2242706716,-0.028090097,-0.0675436109,0.0957195312,0.0862422884,0.4852786958,0.3077232838,0.4453638792,-0.0495040603,0.3921653926,-0.0835309774,0.0408271775,0.0993405282,-0.0459782109,-0.1255243868,-0.2114326358,0.204008922,-0.0066473591,-0.0031043291,-0.514117837,-0.2244881839,0.2302787453,0.0412578695,-0.2499296367,0.3766584098,0.0238632113,-0.5575558543,-0.1051275283,-0.029765375,-0.0915054232,-0.5620676875,0.2115024328,-0.1757505834,0.1443349719,-0.2802774608,-0.3391255736,-0.0725459903,-0.2252959311,0.9792701006,-0.3612010479,0.0411205776,0.1631681174,-0.2286432534,0.0631909594,-0.0209391639,0.1144876555,-0.0962952077,-0.543505311,-0.171720624,0.0018520691,-0.1162723675,-0.1359869838,-0.0556084886,0.0778616071,0.5772361159,0.0053357044,-0.190570876,-0.3950435221,-0.2697645128,-0.0862358287,0.4126008451,0.2028821409,-0.3294523656,0.1059352309,-0.0606171004,-0.0498360544,0.233362332,0.0149459997,-0.0965534598,0.216025725,-0.0281821564,0.3622387648,0.3341759145,0.3676950336,-0.1023174226,0.0332149342,-0.2708882987,-0.0947812423,0.1031512022,-0.3197565675,0.0953792408,0.1244688332,0.2572000325,0.0873058438,-0.2040501833,0.2834171355,-0.1026234627,-0.0883007497,-0.3075081408,-0.3365748823,0.0534683838,-0.210079208,0.2320919186,-0.1446572244,-0.2437559813,-0.03635877,-0.1027688831,0.2560223341,0.2257698923,0.1445205063,0.1072359905,-0.4300121665,-0.2547133565,-0.4477873445,0.6707225442,-0.0858910829,0.234377712,0.0605876595,-0.1017306522,0.1097610369,0.0608188212,1.0629650354,-0.1269094646,-0.1150798798,-0.4581466615,-0.517885983,-0.3554793298,0.2233658582,-0.2517984807,0.5846662521,-0.2207652479,0.4985593855,-0.3260095119,-0.2020828575,0.3566174805,-0.021687692,-0.1113523543,-0.1126884148,-0.0517902076,-0.5832823515,-0.1669629663,-0.0096345497,0.1398490071,0.1057272702,0.073675327,0.0825316831,-0.345015347,0.2281898409,0.3295410573,-0.1223947406,-0.0531050675,0.2326716632,-0.1102341712,0.1479382813,0.3024218976,-0.1586880237,0.1321974844,-0.0909596533,-0.4877898991,-0.2094704211,0.1972485483,0.4393365383,-0.0067398567,0.1547267884,0.5029050708,0.3022942841,0.2137037516,-0.1864867806,0.0794921443,0.0667030066,0.4177776873,-0.2173391283,-0.3686799705,0.2620944381,0.2697609663,-0.1836909205,0.321077913,0.3597310185,-0.1764508486,0.0623170249,-0.0336873084,0.9954258204,-0.2392773777,0.032742627,0.0933552831,0.0993018448,0.5693767667,-0.3401606381,0.1846494526,-0.1260444671,-0.4899755418,-0.135716036,-0.079919152,0.0968293175,0.1936994195,-0.3488971889,0.3828977346,0.2853866816,-0.0825762823,-0.0641366765,-0.0280254558,-0.1656277478,-0.1071335822,-0.14671655,-0.1079679504,0.0220566317,0.5606478453,0.0056524188,-0.0185837746,0.2328845263,-0.0159533806,-0.0225528292,-0.1988405585,-0.5518344641,0.2577758431,0.2468711287,-0.0929161385,-0.0591979399,0.6065933704,0.5822288394,-0.1418762654,0.0419400744,-0.1534196436,0.069488734,0.0006188672,-0.2725731432,-0.0609657317,0.1631481647,0.0406530648,-0.2638309896,0.0023189462,0.0961168781,0.1285231262,-0.0629003942,0.0654477552,-0.0298105013,-0.3008282781,-0.0218396429,0.1121982411,-0.1306973249,0.1083184853,-0.0872032344,-0.0722611919,-0.5252497196,0.1937134266,-0.1698551774,-0.4661316276,0.0094296681,0.3949134946,0.2523920536,-0.1013392061,0.3541485369,0.0105433352,0.129339844,-0.0748833567,-0.2623842955,0.3379665017,-0.4243885875,0.4755460322,-0.3102811575,0.2922950089,0.074296169,0.1960959882,0.1080114469,0.2043850571,-0.2514828146,-0.2861175537,-0.1518701166,0.0957859606,0.2197881043,0.1917156875,0.0758672729,0.2850635052,0.1013225168,-0.1431701332,-0.1664751172,0.258251518,-0.2909674644,0.2511137724,0.1690468341,-0.0266354885,0.1901881695,0.2332659215,-0.0508921742,0.2125862688,-0.1234488487,-0.0858648717,-0.033523567,0.1654632837,-0.0731757209,0.0120541649,0.2206756324,0.2774683535,0.001632168,-0.1156456396,0.6620077491,0.3135236502,0.1242309287,0.4530085921,-0.0392362289,-0.0300653055,-0.4302762449,-0.2596962452,0.205622673,0.0129287913,-0.041948501,0.0790596232,-0.0580811575,-0.0484758131,-0.0968901217,-0.0315731056,0.5867382884,-0.2456670105,-0.0277612451,0.1022151783,-0.0098981168,-0.3602332771,0.3858983517,-0.1729997694,-0.2082863003,0.0613645613,-0.0469940193,0.014611749,-0.0997826457,0.2281923741,0.3033629656,-0.2837214768,0.0247397069,0.3674831986,-0.11263486,0.2506143451,-0.1545862705,-0.2312275469,-0.0144274021,-0.1954026222,0.1038140133,0.4507317543,-0.0486226827,0.0728783384,-0.0790012255,0.1152454913,0.3343017101,-0.0837697089,-0.1827017814,0.4700679481,-0.2861013114,0.2373531908,0.3596649468,-0.4437984824,0.1341635585,0.4912834466,0.0105598317,0.4207003713,-0.0713460594,0.5603382587,-0.0149856564,-0.2897993326,0.1662007421,0.4697264135,-0.1804602444,-0.1805506796,-0.1419833004,-0.1482178122,-0.1495567709,0.0911211446,-0.0643826574,-0.0144673623,0.3167575598,0.0442402326,0.1189552695,0.0375941619,-0.1458500922,0.3263544738,0.3060722351,-0.0280870218,-0.1908961236,0.3235077262,-0.2239843905,0.0667230859,0.6331109405,-0.0457763001,-0.0151909618,-0.3315422535,0.6930217147,0.2356721908,-0.1277495921,0.0481511913,-0.0293366425,0.0372452475,0.0154018644,-0.0318471566,-0.0016323165,-0.0284106713,0.1966878474,-0.1164305806,0.145705089,-0.4811047316,0.0732528418,-0.0500028133,-0.0742101371,-0.2154103816,0.2200533003,-0.2281733751,-0.3667743206,0.117242679,-0.1666602045,0.075022608,0.1193062142,-0.0270116944,-0.183455199,0.2962087393,0.3321763873,0.4024871886,-0.1092905551,-0.189754948,-0.0769579262,0.0380624048,0.2055287808,-0.2358456552,0.3291265965,0.0614926144,-0.0284663942,0.0228521321,-0.0437223464,-0.1772944182,-0.251696229,0.1158900335,-0.3434189856,-0.2081310451,-0.0887909234,0.2309343368,0.2883775234,-0.3365623951,0.0835811719,0.3303910792,-0.2150851339,-0.0639268383,0.4400002062,-0.006996654,0.0228792578,0.0210187528,0.1909422576,-0.1019744873,-0.1159372553,-0.0735417157,0.4731511772,0.1208456382,-0.3382180333,0.3409439325,-0.0889389142,0.1021969095,-0.2532986999,-0.3499638736,-0.0432649888,0.1976893693,0.5831398964,-0.4377908409,-0.2445881814,0.1392468661,0.0341706313,-0.0471587554,0.3726726174,0.8478620648,-0.114143014,-0.0920383036,-0.1728164554,-0.0630580634,0.5526186824,-0.2391415536,-0.1933091879,0.0768830702,0.2288538814,-0.1346839964,-0.4252274632,-0.6996849775,0.1381092519,0.2961443067,-0.0848561004,-0.1005780622,0.3815325201,-0.0538215898,-0.1655156612,0.0913752243,0.0765388831,0.3595644236,-0.3701280653,0.2556760907,-0.3200045824]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2767","title":"equal operation to perform unbatch for huggingface datasets ","comments":"Hi,\r\nthis is also my question on how to perform similar operation as \"unbatch\" in tensorflow in great huggingface dataset library. \r\nthanks.","body":"Hi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n","comment_length":22,"text":"equal operation to perform unbatch for huggingface datasets \nHi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n\nHi,\r\nthis is also my question on how to perform similar operation as \"unbatch\" in tensorflow in great huggingface dataset library. \r\nthanks.","embeddings":[-0.0424975231,-0.7478794456,0.0441974327,-0.0638020411,0.0602153614,-0.0741190612,0.2164928764,0.0234034453,0.4041276276,0.2328329533,-0.375790894,-0.0309473332,0.008743871,0.4603340924,0.2217881382,-0.2381002754,0.0493126996,0.0999169722,-0.2166887224,-0.0692890733,-0.07732407,-0.0117729409,-0.5677039623,-0.1649728417,0.3511196375,-0.2511515319,-0.1151801869,-0.2309311926,-0.1251270175,0.2847944796,0.4793148041,0.2426975667,0.1669390798,0.4491536617,-0.0001256292,0.2773347199,-0.2205358744,-0.1515981555,0.1575898677,-0.151048407,-0.1850503683,-0.0384183079,-0.0539824218,-0.3068354428,-0.1764235646,-0.0184586626,0.0631059781,-0.3345097601,0.6362299919,0.1719640344,0.0013318665,0.2099072486,-0.0798648074,0.0820844993,-0.1022054255,0.0367429368,-0.0256353393,0.1146042943,0.0846524984,0.0455240831,0.0681281611,0.2344957441,-0.0564624779,-0.0299290791,0.0353240147,0.1095685884,-0.032894548,-0.3611135781,0.1364402771,0.3937362134,0.2183602303,-0.373719424,-0.5333204269,-0.0268120263,0.0756629482,-0.0685920864,-0.205910027,0.2482982725,-0.0961421728,0.2391859442,-0.4491079748,-0.1502820253,-0.0860688835,0.0137132183,0.0574082434,-0.1004855558,-0.3021975756,0.2693094313,0.0576845296,0.0820068941,-0.0920690149,-0.0770096332,0.1063661948,0.158407554,-0.672842741,-0.2440767139,-0.2414087653,0.0239156168,0.2683830857,0.3667656779,0.5453345776,-0.0524723865,-0.0262592211,0.0717241764,0.050923761,0.5504601598,-0.0927488953,-0.2812029421,0.037321765,0.1110856161,-0.0138820438,-0.0510671809,0.3221217394,-0.0530241765,-0.344066143,0.14720577,0.3968828321,-0.1028705165,-0.355604291,-0.1796554625,-0.3000112772,-0.1298760325,0.0673620328,0.1227707714,-0.1877583712,0.016536355,0.1142916307,0.0416766889,-0.2851560414,-0.0781527162,-0.1623554677,0.2975337207,-0.1561620086,0.0250501819,-0.0284833647,-0.2422840297,0.0106540537,0.487816751,-0.0249357056,-0.0006227787,0.2213852406,-0.1916823983,0.2998886406,-0.0221225787,0.1274335682,-0.1672376692,0.1175889149,0.0563347861,-0.0360738076,0.0110055776,-0.0750050694,-0.2657646239,0.0381360538,-0.0416547023,-0.3246122003,0.0032485053,-0.1793220937,0.2697196901,-0.3806002438,0.4313870072,0.4621528983,0.3615436256,-0.0420184471,-0.0687471926,0.4144029617,-0.0749240369,-0.0679476485,-0.3054028749,0.3379549384,-0.0421895646,0.3259534538,0.2724405229,-0.0680568144,0.3023763001,-0.2311029583,-0.1621000916,0.235185951,-0.1755775064,-0.4458527267,0.1876093298,-0.0978450701,0.2507426143,-0.0481086299,0.0729169548,-0.0773575604,0.2221725732,-0.0441651233,0.2555157542,-0.0487267859,-0.1383284032,-0.015506614,-0.1830752939,0.3481126726,-0.0824953467,-0.0941848382,0.3812470138,0.1271146983,-0.0217743367,0.2218708396,-0.4553268254,0.1601547897,0.2863127887,0.2274770886,0.1333514005,0.2319566011,-0.1823991835,-0.3266887963,0.1416940689,-0.3598377407,0.0889272392,0.115725413,-0.3049005866,-0.1305134594,-0.2349703014,-0.1013237908,0.0059276456,-0.1083109602,-0.0970226675,0.4688789845,0.0513625555,-0.4691463411,0.353730619,-0.1237802282,0.1276918054,-0.4118320942,0.0413681194,0.2954889536,-0.0169274807,-0.0135378065,0.2284951657,-0.0155147873,-0.3583829999,0.0819730461,0.3086032569,-0.3341932893,0.0336165279,-0.0568824895,0.0462707952,0.0354421847,-0.1527609974,-0.1786395758,0.0535832122,0.0892216489,-0.011261384,-0.0550484695,0.7477804422,-0.0993619338,0.1371473223,0.1400215626,-0.0569110103,-0.1730741113,-0.3248525262,-0.1202724278,0.2766099274,0.2228631377,-0.2027987391,0.2167761475,-0.1463133544,-0.2560457289,-0.33674106,0.3105781972,0.0451067872,0.3367241919,0.1336293221,0.0417049378,0.2017525434,0.0126549862,-0.1468297839,0.4227997363,-0.0309492908,0.2567645311,0.379553616,0.3169096708,0.1312439889,0.0841790512,-0.2490986586,-0.0797817633,-0.0145397317,0.0557714365,0.0699486136,-0.3710449636,-0.0041458462,0.0700621605,0.1621102691,0.0241403021,-0.1014374867,-0.1798603833,-0.2002354413,-0.2028765678,0.0098963231,-0.1672875881,-0.0132869389,-0.0906698033,0.0028247344,0.1491720974,0.4050225914,0.2213249952,0.4259082973,-0.09626095,0.3810700178,-0.1173515394,0.0341566578,0.091918461,-0.0392951705,-0.1389828175,-0.1874711066,0.2197643965,0.0147329243,0.0398427658,-0.5104528069,-0.163260147,0.2542126179,0.0426331311,-0.2061795145,0.3554162085,0.056791164,-0.535736084,-0.0846047625,-0.028251566,-0.0885254443,-0.5490326285,0.1930105537,-0.1596266627,0.1691580564,-0.3059814274,-0.3406267464,-0.0838258937,-0.216371581,0.9462254643,-0.3164279759,0.0208242256,0.1576191187,-0.3140150905,0.1281804889,-0.0236970019,0.1545455158,-0.0536154322,-0.5723289847,-0.1633879691,0.0267136283,-0.1361408532,-0.1201191023,-0.0198404156,0.1375264078,0.5128785968,-0.0431780182,-0.1274823099,-0.4869631529,-0.2536957264,-0.0911310539,0.4112828374,0.1251698136,-0.3662730157,0.1088478789,-0.0184606872,-0.0162020139,0.213346824,0.004052876,-0.0966634229,0.2594733834,-0.0480677076,0.3650169373,0.3248707056,0.3047578335,-0.1051920429,0.0402547941,-0.297282666,-0.0865996405,0.0679808035,-0.3497266769,0.1237851828,0.1597072482,0.327542007,0.1114405021,-0.1926475465,0.3412344754,-0.1267317832,-0.0945030898,-0.2697210312,-0.3193699419,-0.0557211973,-0.1540178657,0.2477357388,-0.1236863285,-0.1904548705,0.0278493799,-0.1654296517,0.2789090276,0.2406210452,0.1410547048,0.0966712087,-0.3829194307,-0.3035737872,-0.4883563817,0.6454593539,-0.1398071945,0.2618300617,0.0665872544,-0.1270617098,0.1306856424,0.0341162197,1.0370573997,-0.0324867778,-0.1396744102,-0.4750311077,-0.4710195661,-0.3139974475,0.18416062,-0.2221711129,0.5769687891,-0.2369395047,0.3767886758,-0.2941345572,-0.222318843,0.366845876,-0.0364616141,-0.0791642293,-0.1018247157,-0.0539003946,-0.6249426007,-0.0915392339,-0.0180134848,0.0656028315,0.154986158,0.041196391,0.1325081736,-0.3746035993,0.1941810846,0.3676533997,-0.1300132424,-0.0035208396,0.149178803,-0.1001515388,0.1031873077,0.3238515556,-0.229727298,0.1300437897,-0.032159511,-0.4681832492,-0.2133583724,0.1570042968,0.4705612957,-0.0856447294,0.1072182879,0.6108522415,0.2382157147,0.1436378807,-0.1965652704,0.0670796335,0.0644117147,0.4300856888,-0.188441664,-0.3638094664,0.2624984384,0.3062618971,-0.1562134922,0.2497635037,0.3646601439,-0.2032862604,-0.0084194783,-0.0541925244,0.9401324987,-0.2121051848,0.0104363058,0.054609973,0.0788148791,0.6255307198,-0.2705157995,0.2517609298,-0.1583642662,-0.5409907699,-0.1582000256,-0.1358966976,0.1604440808,0.1070916802,-0.3282728493,0.3783901334,0.2835077941,-0.0956671759,-0.034036871,-0.0009841957,-0.1824061424,-0.0806929842,-0.1565550119,-0.1136664227,-0.0327011421,0.5507885218,0.0213549193,0.0167274065,0.2169836313,0.0469508097,-0.0409515128,-0.15431045,-0.4644551575,0.2687178552,0.3073612452,-0.1465978622,0.0131554436,0.6149058938,0.6690641642,-0.1112763211,-0.0006842926,-0.1205121353,0.0515998974,0.0170154609,-0.2183578461,-0.0521085449,0.1305121332,0.0501958393,-0.2650192678,0.0233010948,0.1552675813,0.1165151596,-0.0890503824,0.0902366787,0.0055774255,-0.274508059,-0.0404995382,0.1044759452,-0.117905736,0.1013720632,-0.0872135237,-0.0688399225,-0.4957576692,0.178932935,-0.1613684744,-0.4757240713,-0.0061027063,0.3508865237,0.2941737175,-0.064324528,0.4002082646,0.0366977863,0.1323281229,-0.0667798668,-0.2662048638,0.2626368105,-0.3881516755,0.4740816355,-0.2939937115,0.3173548877,0.1067781299,0.1131158397,0.0162121095,0.1288261861,-0.288934648,-0.2663587332,-0.1245181933,0.0764646828,0.1795401722,0.1779039204,0.0750110969,0.2766519189,0.0476269647,-0.2402656823,-0.1558551043,0.2842417061,-0.2833643258,0.2184417099,0.132666558,-0.0157710034,0.1630739421,0.2165210843,-0.0567008518,0.2261711359,-0.0547173433,-0.0868413076,-0.01728989,0.1616122276,-0.1124465838,0.0084790913,0.1674675941,0.2956520319,-0.0330698378,-0.0734099969,0.6825496554,0.2951779664,0.1258980483,0.4761397541,-0.036271926,-0.0462831594,-0.4582363665,-0.3013265729,0.1785136312,0.0314401649,-0.0072616506,0.1081374511,-0.0615027174,-0.0164691657,-0.142190516,-0.0376242734,0.5748140812,-0.2740880549,-0.0295528639,0.0608416609,-0.0396312661,-0.3905901909,0.4094214737,-0.1740113944,-0.2018861175,0.020180963,-0.020628171,-0.0053494461,-0.1028336734,0.2194721401,0.2927230597,-0.2715494633,0.025117347,0.3670117557,-0.1269934773,0.2906769216,-0.2144141793,-0.2452571392,0.0469378382,-0.1940744817,0.1204538196,0.4268241525,0.0451319404,0.0252438337,-0.0401616096,0.0988043994,0.2951332927,-0.0160153396,-0.1837643534,0.4372751713,-0.2256069779,0.2472945005,0.3573013246,-0.4244349301,0.1144540086,0.4516224563,-0.0041886684,0.4051045775,-0.0818894431,0.545558095,-0.0006798513,-0.250179112,0.146315828,0.4758297801,-0.1705471277,-0.1944057196,-0.0980265141,-0.1143437549,-0.093543008,0.115186803,-0.053310886,0.0250720847,0.1784051657,0.0636573061,0.108123824,0.0506287813,-0.0329837799,0.3312443197,0.2496910095,0.0193046499,-0.1996648163,0.3107480109,-0.1767770797,0.0455963947,0.6782322526,-0.0700295419,-0.0051777847,-0.3357172012,0.6726943851,0.2287703753,-0.1550967544,0.0435780771,0.0125395311,0.0421920866,0.066052638,-0.0302558243,-0.0014334109,0.0101446854,0.1819347292,-0.0780007392,0.1584955901,-0.538995564,0.1471691281,0.0061968779,-0.0930639133,-0.2471258938,0.2212389112,-0.275174439,-0.4047088921,0.1331683993,-0.140824616,0.0647846833,0.1152140126,-0.0306608416,-0.2047548443,0.297388494,0.356315583,0.4698463678,-0.0787715688,-0.1468904912,-0.0865190625,-0.0149453208,0.2726433277,-0.2390535176,0.3346012235,0.0375351049,-0.0850001201,-0.0304190386,-0.0827731267,-0.1391343027,-0.2084831595,0.0215337258,-0.3523412943,-0.2075141966,-0.0560427867,0.2039812207,0.1930284947,-0.2936784625,0.1306679398,0.304652065,-0.2205507457,-0.0346958302,0.4123051167,0.0229639858,0.0205563549,-0.0387004614,0.1113634408,-0.0904418677,-0.0774443597,-0.0414636843,0.4800552726,0.0969666541,-0.303117305,0.3314430118,-0.1199898049,0.109498933,-0.2138832361,-0.3864007592,-0.0466757901,0.227021277,0.5675137639,-0.447263062,-0.2105763406,0.1187098771,0.0214072932,-0.0336179882,0.356982708,0.8135994673,-0.126994282,-0.1103093475,-0.2217966169,-0.0645324886,0.5682218671,-0.2965456545,-0.1726727486,0.049536515,0.2286392152,-0.1154554784,-0.3957335651,-0.6892159581,0.1222510561,0.3059706688,-0.1044685543,-0.1229855791,0.4051800966,-0.0281280261,-0.0896987393,0.111638844,0.0332249366,0.3858937919,-0.3727833629,0.25028193,-0.2413191646]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2767","title":"equal operation to perform unbatch for huggingface datasets ","comments":"Hi,\r\n\r\n`Dataset.map` in the batched mode allows you to map a single row to multiple rows. So to perform \"unbatch\", you can do the following:\r\n```python\r\nimport collections\r\n\r\ndef unbatch(batch):\r\n new_batch = collections.defaultdict(list)\r\n keys = batch.keys()\r\n for values in zip(*batch.values()):\r\n ex = {k: v for k, v in zip(keys, values)}\r\n inputs = f\"record query: {ex['query']} entities: {', '.join(ex['entities'])} passage: {ex['passage']}\"\r\n new_batch[\"inputs\"].extend([inputs] * len(ex[\"answers\"]))\r\n new_batch[\"targets\"].extend(ex[\"answers\"])\r\n return new_batch\r\n\r\ndset = dset.map(unbatch, batched=True, remove_columns=dset.column_names)\r\n```","body":"Hi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n","comment_length":72,"text":"equal operation to perform unbatch for huggingface datasets \nHi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n\nHi,\r\n\r\n`Dataset.map` in the batched mode allows you to map a single row to multiple rows. So to perform \"unbatch\", you can do the following:\r\n```python\r\nimport collections\r\n\r\ndef unbatch(batch):\r\n new_batch = collections.defaultdict(list)\r\n keys = batch.keys()\r\n for values in zip(*batch.values()):\r\n ex = {k: v for k, v in zip(keys, values)}\r\n inputs = f\"record query: {ex['query']} entities: {', '.join(ex['entities'])} passage: {ex['passage']}\"\r\n new_batch[\"inputs\"].extend([inputs] * len(ex[\"answers\"]))\r\n new_batch[\"targets\"].extend(ex[\"answers\"])\r\n return new_batch\r\n\r\ndset = dset.map(unbatch, batched=True, remove_columns=dset.column_names)\r\n```","embeddings":[-0.0753810331,-0.7735245228,0.0261516012,-0.0693187863,0.0228945483,-0.0482676737,0.3034118414,0.040893551,0.397575289,0.2398023307,-0.3650856912,-0.0645896122,0.0264364704,0.411909461,0.179759115,-0.2043053061,0.0633047,0.0990086794,-0.2375567555,-0.0585383028,-0.1362456232,0.0407442078,-0.5339214802,-0.176741302,0.3353881836,-0.2973919511,-0.1161579415,-0.2635371983,-0.1034479588,0.3052071929,0.4921708405,0.309329927,0.2077566832,0.4856838882,-0.0001246521,0.3010813892,-0.2492474616,-0.1526912153,0.1145772412,-0.0873676911,-0.1308563501,-0.0882824585,-0.0693368986,-0.3121809363,-0.1677762419,-0.0135150421,0.086169593,-0.3338784277,0.6199630499,0.107661739,0.0035845411,0.3279936612,-0.0736484453,0.0471059866,-0.161874786,0.0315583572,-0.0162977967,0.1471144408,0.1239845455,0.0130380867,0.0811594427,0.3311883807,-0.1018417254,-0.034806475,0.0685208738,0.1417723894,0.0737149864,-0.4227611721,0.1653945744,0.3434442282,0.1286982298,-0.3457928002,-0.5522564054,-0.1110518277,0.0664463043,-0.0784133896,-0.2424910516,0.1549121588,-0.1231821254,0.1915919334,-0.4587093592,-0.085062556,-0.064205125,0.0513771512,0.0663282946,-0.0215883255,-0.2763303518,0.270816803,0.0563461557,0.0941273719,-0.1524153501,-0.1206801981,0.1649204046,0.2293738127,-0.6707261205,-0.2004776746,-0.2176058739,-0.0827744827,0.2486542463,0.3064738214,0.5127056837,-0.0695319474,-0.0014590117,0.0912642032,0.069061473,0.5457304716,-0.1010435522,-0.2234711945,0.010707383,0.1190956309,-0.0006037139,-0.0132912463,0.3079895079,-0.0960773826,-0.2861174643,0.186826244,0.3991137147,-0.0048522572,-0.3670807183,-0.1827206463,-0.3305942416,-0.1504132301,0.0168250147,0.1619478613,-0.195544973,0.038826786,0.1834592521,0.096247904,-0.2616742551,-0.0341715328,-0.1754765362,0.2005480081,-0.1355688274,-0.0149284443,0.0025351986,-0.1964241862,0.0182543136,0.405243963,0.045959089,-0.0078745354,0.1850574166,-0.1379251629,0.3300073445,0.0267928485,0.2054771781,-0.1043841615,0.1196802184,0.1109361276,-0.0676129758,0.0062329699,-0.0762393475,-0.2370306104,0.1029392928,-0.0261911377,-0.290127635,-0.0064336713,-0.2655008733,0.2805640399,-0.3139200509,0.4107523561,0.4642283022,0.3083906174,-0.0579389855,-0.0884879678,0.3909344673,-0.033580035,-0.1268098652,-0.270321548,0.3334641159,-0.0070434152,0.3165129125,0.3111931682,-0.0583803356,0.3475653827,-0.2451771051,-0.0472126268,0.2052667141,-0.1820423901,-0.4236242473,0.2309683859,-0.0761009604,0.2990359664,-0.035675738,0.0611703806,-0.0361164622,0.1729475856,0.0400792658,0.2885244787,-0.0879403278,-0.0819173232,0.0025714759,-0.1750308424,0.2911466062,-0.0937554762,-0.0195052046,0.2881815135,0.0503391959,-0.1503593475,0.2743276358,-0.5115571618,0.2057096809,0.2351751029,0.2331880778,0.1862762272,0.2376929522,-0.1802090406,-0.4620914459,0.1702177674,-0.375782311,0.1592279524,0.0510786287,-0.3364458084,-0.1325394809,-0.1899868399,-0.1449082196,0.0438793339,-0.0932907164,-0.133311078,0.462906003,0.0250245333,-0.4002581835,0.2373046726,-0.0836174414,0.1097686961,-0.4155571163,0.0613827072,0.2461878359,-0.0186510831,-0.0090569798,0.2117616385,0.0163055379,-0.3814625442,0.1080989093,0.2936584353,-0.3484901488,0.019219039,-0.0798182786,0.0817219391,0.0673079863,-0.0596716553,-0.129542917,0.0506830029,0.0736499205,-0.0641615167,-0.0339300036,0.8116818666,-0.0951392353,0.1268706173,0.067018576,-0.0672764555,-0.1310650855,-0.2365904897,-0.1849679053,0.2044692636,0.1865125,-0.23161982,0.2167523503,-0.1897063702,-0.2400353551,-0.2519877851,0.291700542,0.0077059921,0.3112775683,0.1127476171,0.042579215,0.2121856213,-0.0388295315,-0.1469406784,0.3771910965,0.0011109207,0.2564551532,0.3336131275,0.2784111798,0.1565333158,0.0501440316,-0.260941714,-0.0400376283,0.010296341,0.0104519352,0.1022086889,-0.3529451191,-0.0023584473,0.0579043962,0.1635893732,0.0091688875,-0.0685236752,-0.093329832,-0.1590349078,-0.2032993138,-0.0484064296,-0.199980557,-0.0636533871,-0.0596891753,0.0650865659,0.0827660188,0.4527523816,0.2323810607,0.4247006178,-0.0046510827,0.3034425378,-0.1403293908,0.0124633079,0.0758061558,-0.0436794795,-0.1188975722,-0.236073032,0.2823859453,-0.0057214219,0.0372352451,-0.5395433903,-0.1603367329,0.2113296688,-0.065205209,-0.2380534261,0.3718391657,0.1210858002,-0.5385232568,-0.1301630884,0.0426791012,-0.1647605151,-0.54490906,0.1823602468,-0.151724875,0.1190217435,-0.2554599643,-0.3128781915,-0.1122351065,-0.193388164,0.9071428776,-0.360363394,0.0387682952,0.1697901636,-0.2453997433,0.0909720957,0.0120815188,0.0438339747,-0.1046573743,-0.5249735713,-0.1223860905,0.0322275795,-0.137395978,-0.1270463914,-0.0300666932,0.1196217984,0.5564443469,-0.0042878212,-0.146527648,-0.4110369682,-0.1484779865,-0.0592323169,0.4029557705,0.1972219199,-0.3497493267,0.1031162441,-0.0321001783,-0.0572186373,0.1777200252,-0.0003169307,-0.1117025539,0.2010336071,0.014789477,0.3613646626,0.423307687,0.2689908445,-0.1124324426,0.0211761966,-0.2406370044,-0.0781438947,0.0687934458,-0.4165217578,0.0867044628,0.0967586637,0.2121867537,0.0890761316,-0.2410807461,0.2765483558,-0.0950957611,-0.0303139426,-0.3282959163,-0.3560399711,0.0593812838,-0.2432753146,0.2499993294,-0.1077308729,-0.1504729241,-0.0759818628,-0.1181452349,0.2302044183,0.1959138662,0.1247708499,0.0806263834,-0.3856316507,-0.29920578,-0.5825085044,0.6898991466,-0.0906803608,0.2828050256,0.1038244441,-0.0963239819,0.0918035209,0.0738144964,1.1266227961,-0.0902917236,-0.067985177,-0.4532372952,-0.4792626202,-0.3658041656,0.1747479439,-0.2462021858,0.4959308207,-0.1369210929,0.4861794114,-0.3012235761,-0.2158021927,0.3842225671,-0.0708498433,-0.1029943228,-0.1085480675,-0.1039466858,-0.614331007,-0.1452139765,0.0402121022,0.1736764461,0.1835606694,0.0963888317,0.1442807615,-0.3470229208,0.2196639627,0.3789550662,-0.0680103526,-0.0419562869,0.2102966458,-0.1067141891,0.114185214,0.2900209129,-0.1538269371,0.0499993153,-0.1119307205,-0.5091649294,-0.2036356926,0.1387948841,0.44538638,0.0021075138,0.0893900841,0.4498021007,0.3313249648,0.1878379136,-0.2095423639,0.0752900615,0.122947298,0.476790607,-0.2456014603,-0.3756901026,0.1890125275,0.2901027799,-0.1251985878,0.2583325803,0.3687479496,-0.2165210247,0.0750940368,-0.0771681815,0.9015494585,-0.2241622955,0.0572488122,0.1993955672,0.1786247045,0.5808358192,-0.285525769,0.257675916,-0.1110954359,-0.5086899996,-0.1233623922,-0.1007771716,0.143522501,0.1798946857,-0.2801915705,0.3612534404,0.3504683971,-0.0795020163,-0.0773687959,-0.0594200194,-0.134112075,-0.1367684603,-0.1793332994,-0.106296584,-0.0217228923,0.5700871944,-0.0459106192,-0.0014356059,0.1819747984,-0.004344637,-0.0726044476,-0.2204621583,-0.5403102636,0.2581214011,0.2453534007,-0.1466521174,-0.1070109308,0.5531075597,0.6465962529,-0.1431564689,0.0185717084,-0.1360114664,0.0638046265,0.0116694449,-0.2535888851,-0.1350872368,0.1461563259,0.0387848727,-0.259575367,0.0008656349,0.1302725524,0.042016115,-0.0150953932,0.0775441974,-0.1029819176,-0.3029366434,-0.0792436227,0.0985207185,-0.0989727452,0.0807273164,-0.0858889073,-0.0727722347,-0.5481367707,0.2476285398,-0.1678896993,-0.4914943278,-0.0106162066,0.4057464898,0.2606990933,-0.1742279232,0.4102480114,0.0836177319,0.093614988,-0.0570735969,-0.2861724794,0.2830045521,-0.4344593287,0.5060538054,-0.3074496388,0.2821793258,0.0642789677,0.1685796976,0.12941131,0.1295879632,-0.255628556,-0.2817218304,-0.1472710967,0.0785300583,0.2440841645,0.2441805154,0.1258409619,0.3767596781,0.0902140588,-0.1705210507,-0.1665884703,0.2332267612,-0.2345494777,0.1979041547,0.140725255,-0.0713166445,0.2908710837,0.2017937303,-0.0633118823,0.1915008575,-0.0993861705,-0.1101252362,-0.0246681683,0.1482065171,-0.0789695159,0.0089468732,0.2233039588,0.2798078656,-0.0272522606,-0.1051924303,0.6996047497,0.3694262803,0.1226846352,0.4460561872,0.000668635,0.0151981646,-0.3611882031,-0.2898232639,0.2136497796,0.0124402745,0.0064391419,0.1228917539,-0.0764205828,-0.0361728407,-0.1091239452,-0.0680708811,0.5871477127,-0.2230691314,0.0146059226,0.0528352074,-0.1149175167,-0.3639836311,0.4361190796,-0.1670497358,-0.1969697177,-0.0572258122,-0.0879222304,0.0174715333,-0.1485780776,0.2439451367,0.3767095208,-0.262863934,0.0559383705,0.3968823552,-0.1516262293,0.3247960508,-0.1985861808,-0.2218667418,0.0902816132,-0.1761721224,0.1324374825,0.4299080968,0.057751257,0.1098402962,-0.0126127396,0.1312818527,0.322417587,-0.0335141085,-0.1583712399,0.4137467146,-0.2783847153,0.2369359583,0.2978637815,-0.4279221594,0.0664772391,0.4004288018,-0.0287378132,0.4075155854,-0.0916058198,0.5527042747,-0.0889406651,-0.3047347069,0.1227997616,0.4881363511,-0.2441277057,-0.1703254133,-0.1985963583,-0.1303695887,-0.0382661261,0.0753605515,-0.0483205169,0.0260792896,0.1791818738,0.035792809,0.1077525765,0.0507127456,-0.150303334,0.2482164055,0.2431674153,0.0162209235,-0.2239834368,0.2543329895,-0.2170743793,0.0394469686,0.7196279764,0.0188232269,-0.0095197074,-0.2957319915,0.6617161036,0.1789482087,-0.1640623957,-0.0295566432,-0.0209624842,0.0190350693,0.0180690605,-0.0139458394,0.0262741987,-0.0292681474,0.1912162602,-0.1057024002,0.161187768,-0.4103523791,0.0813199133,-0.081707336,-0.1073371843,-0.219711408,0.1943629384,-0.2433102876,-0.3114367425,0.1205803305,-0.1593330652,0.1079164222,0.0862915814,-0.0176184643,-0.1551500559,0.3381443024,0.3547686636,0.3820835352,-0.1703349501,-0.0973507687,-0.1417099237,0.0107850516,0.1978826821,-0.2089080364,0.334354043,0.1071263477,-0.0026871511,-0.0359332785,0.021387428,-0.1784843355,-0.2717832327,0.121122703,-0.3433920145,-0.204581812,-0.1264019758,0.195732981,0.2733252943,-0.3446793258,0.075491406,0.35812217,-0.213390246,-0.0684988946,0.37402004,0.0175867807,0.0997598693,-0.0087883919,0.1813976318,-0.0638087764,-0.0679592937,-0.0506675281,0.3892513216,0.1280761957,-0.3385195732,0.3446708024,-0.0772373602,0.1299190819,-0.2593948841,-0.3769942522,-0.074344568,0.2412814796,0.5907958746,-0.4905029535,-0.2275039852,0.1521978825,-0.0045223483,-0.0331159681,0.3418536186,0.823520422,-0.1254417151,-0.0221549254,-0.2195950449,-0.0871911496,0.6015579104,-0.2815897167,-0.226382345,-0.0047984147,0.1956968158,-0.103582792,-0.3497138023,-0.6360821724,0.1418419331,0.2863026559,-0.075530231,-0.1086892188,0.4168684781,-0.083490707,-0.1523137242,0.0869465694,0.0916837305,0.35708794,-0.411241889,0.1835698485,-0.3044159412]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2767","title":"equal operation to perform unbatch for huggingface datasets ","comments":"Dear @mariosasko \r\nFirst, thank you very much for coming back to me on this, I appreciate it a lot. I tried this solution, I am getting errors, do you mind\r\ngiving me one test example to be able to run your code, to understand better the format of the inputs to your function?\r\nin this function https:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L952 they copy each example to the number of \"answers\", do you mean one should not do the copying part and use directly your function? \r\n\r\n\r\nthank you very much for your help and time.","body":"Hi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n","comment_length":90,"text":"equal operation to perform unbatch for huggingface datasets \nHi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n\nDear @mariosasko \r\nFirst, thank you very much for coming back to me on this, I appreciate it a lot. I tried this solution, I am getting errors, do you mind\r\ngiving me one test example to be able to run your code, to understand better the format of the inputs to your function?\r\nin this function https:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L952 they copy each example to the number of \"answers\", do you mean one should not do the copying part and use directly your function? \r\n\r\n\r\nthank you very much for your help and time.","embeddings":[-0.0108777815,-0.7745167613,0.0500379056,-0.0438205153,0.0585958362,-0.1315132678,0.2449042946,0.0355795249,0.403403908,0.207198292,-0.3647301793,-0.0576848835,0.0258308109,0.4528597593,0.1368795037,-0.2415413558,0.0354214869,0.130074814,-0.3146570027,-0.0971326157,-0.1279555559,0.0090236766,-0.5449327826,-0.1741686314,0.3397312164,-0.285574764,-0.1111921817,-0.2267976999,-0.2018394172,0.3307215273,0.421207577,0.276545018,0.1566703469,0.5160192847,-0.0001262681,0.2817387581,-0.2239812016,-0.1529459655,0.0926806554,-0.1573763192,-0.141733855,-0.0440882891,-0.0630615577,-0.2752128839,-0.1816470772,-0.0017648741,0.0454082973,-0.3386259675,0.6838319302,0.1817790866,-0.0048463992,0.206730932,-0.0597356409,0.092067562,-0.1350352913,0.0671231896,-0.0204418618,0.0735917613,0.0742384195,0.0391847678,0.0825988725,0.2869920731,-0.0342607386,-0.0616690628,0.0821961388,0.1065226719,0.0024306127,-0.3898434937,0.1111200899,0.3619070351,0.2063346505,-0.3542043865,-0.5544352531,-0.0521407835,0.0253042858,-0.047070466,-0.2078715265,0.1657986045,-0.1159716845,0.2144903541,-0.4811854959,-0.0987051576,-0.0723049492,0.0731838271,0.0504783355,-0.1077801734,-0.2960430086,0.2540514767,0.0699314401,0.0905945897,-0.1174799427,-0.1175418869,0.123034589,0.2084047049,-0.6481525302,-0.2360532731,-0.2577502429,-0.0207786467,0.263964802,0.3931050301,0.5452445745,-0.105997473,-0.0683631003,0.0775518268,0.0718001723,0.5941019654,-0.093124032,-0.2449739575,-0.0216667969,0.1149332821,0.033849474,-0.045613125,0.3210249543,-0.0735206679,-0.3377156556,0.1442278922,0.3928581178,-0.098125428,-0.3077066541,-0.1714103371,-0.3482732773,-0.1062351093,0.0052448702,0.1235989332,-0.1738000065,0.0956065878,0.2250313759,0.0844191462,-0.2863166034,-0.0946582407,-0.1676394641,0.2888308764,-0.1838473678,0.0078385184,-0.0555583052,-0.1803662032,-0.0213175565,0.5038354993,0.0103862332,-0.0349442065,0.2061427534,-0.1644645035,0.3446747363,-0.0276211277,0.1581510454,-0.1279955357,0.1393094361,-0.0182058513,-0.0253074467,0.0004724024,-0.0577381849,-0.2445571721,0.0647586286,-0.0385767519,-0.3195185959,0.0288104545,-0.193225801,0.3196071088,-0.3369585574,0.3980179131,0.454992801,0.3295466602,-0.0548327714,-0.0711227283,0.4004134536,0.0320708752,-0.0892611146,-0.2960408628,0.3439548612,0.0093569336,0.3034724593,0.2937402427,-0.0132582132,0.2759523392,-0.2429623306,-0.0962143987,0.2000613511,-0.1653339267,-0.3978582919,0.2438246906,-0.0936977118,0.2858174145,-0.0008192139,0.0327162445,-0.0864739269,0.1881908178,-0.0270804409,0.2525263131,-0.0661386475,-0.1053133681,-0.0651599318,-0.1617499441,0.3505885303,-0.0655955747,-0.0829995051,0.3464676738,0.0652354285,-0.0619511642,0.2206347138,-0.5067725778,0.1742602289,0.2860680223,0.2737710476,0.1422733814,0.2397804707,-0.1471372098,-0.3221267164,0.1042289734,-0.3572678864,0.1865509748,0.1231021062,-0.2822453678,-0.174836114,-0.2101640403,-0.1613721251,-0.0220703632,-0.092828609,-0.0894160867,0.4296045601,0.062740095,-0.3935889304,0.3531248271,-0.1103326455,0.11737369,-0.4565909505,0.0766896084,0.2712408602,-0.0212028157,-0.0294926483,0.202091679,0.0250198003,-0.3182093203,0.0931531414,0.2768290639,-0.3111835718,-0.0000497935,-0.0807028562,0.1129693165,0.0448838323,-0.0933870897,-0.1356566399,0.1158975288,0.0989421085,-0.0575681403,-0.0803855881,0.7735747695,-0.0630798414,0.1183306128,0.1151418835,-0.0541957915,-0.1563129872,-0.2935832739,-0.1384200007,0.2834094763,0.2511129677,-0.2120889574,0.2328744829,-0.1283389032,-0.2621282637,-0.2862565815,0.2700942159,0.0284279771,0.3281134963,0.1195155457,0.0314621255,0.1946714073,-0.0153375231,-0.1554953903,0.3815603852,-0.0071744379,0.2247863263,0.3685889542,0.3105094731,0.1431070119,0.0553678721,-0.2981645763,-0.0729001313,0.0393902995,0.0086060138,0.1030274034,-0.3797222078,0.0236477815,0.0302486364,0.1702512354,-0.081622079,-0.0730784461,-0.1584304422,-0.1832448989,-0.2011881322,0.0363762155,-0.1941183209,-0.0438365713,-0.0806341022,0.0661064312,0.1269943118,0.4456271529,0.2406216264,0.4628773928,-0.0280903801,0.3665049374,-0.1363861561,0.0575946607,0.0705956444,-0.0461736582,-0.1194234192,-0.2033166885,0.2038011402,-0.0031708898,0.0439757034,-0.4660874009,-0.2374643683,0.2677345872,0.0232883915,-0.2113489062,0.3508021235,0.0557054542,-0.5462496281,-0.0783203766,-0.0216935202,-0.0884826779,-0.5615617633,0.252897054,-0.1770302355,0.1233159378,-0.3099218905,-0.3661047518,-0.0872962326,-0.2064203322,0.9370263219,-0.3762081861,0.0057926476,0.1996505111,-0.2222321033,0.0677220523,0.0126034934,0.1556550413,-0.0976721346,-0.5041771531,-0.073125653,0.0281708445,-0.1376679242,-0.148702845,-0.0546869412,0.0802529082,0.5497124195,-0.0282657202,-0.1994812638,-0.4520470798,-0.2961638272,-0.0728063732,0.4209002852,0.1771828979,-0.3457151055,0.1205206513,-0.0375313312,-0.0695850253,0.2295768261,0.0187303331,-0.1270513982,0.2347080708,0.0014079961,0.3598975837,0.3538529873,0.3470194936,-0.0990255922,0.0370055921,-0.2861382365,-0.0852763057,0.0900597051,-0.3696490228,0.1021969095,0.1357011646,0.263286978,0.07791017,-0.1912818551,0.2992965579,-0.1509897858,-0.0565325208,-0.3372487128,-0.3283590376,0.0419884883,-0.1391087919,0.2262818068,-0.1436557472,-0.1614686102,-0.0105280802,-0.1394328028,0.2685930431,0.2348733693,0.1099194214,0.1222356781,-0.3618437648,-0.2483720332,-0.4749386609,0.6477642655,-0.1054954007,0.2582255006,0.0596568771,-0.1224758923,0.1019009203,0.0552801192,1.0442065001,-0.0928623453,-0.1418884248,-0.4542002976,-0.5523269773,-0.3331586421,0.2196693718,-0.2337745577,0.5441312194,-0.1980513185,0.4110106528,-0.3555338383,-0.1987074465,0.3440606296,-0.0428050794,-0.1176277995,-0.1123756245,-0.0697600096,-0.6193229556,-0.11438068,-0.0228117667,0.1505482495,0.1685724407,0.0634071305,0.1474557221,-0.3576940298,0.2287350148,0.3958051503,-0.1185991764,-0.0005935804,0.1395819336,-0.1285169572,0.1251049638,0.312931776,-0.2290610671,0.1828499436,-0.0389183797,-0.525521338,-0.2011207193,0.1282041371,0.4326058626,0.0043082265,0.1211856157,0.4952029288,0.2905820906,0.1692709029,-0.1751951128,0.0514556281,0.0547314472,0.4383256137,-0.2018345594,-0.3718878925,0.2034461945,0.2566279173,-0.1311087757,0.2950930893,0.4055140018,-0.170587033,0.0070119798,-0.0641239062,1.0264148712,-0.1779045761,0.0226665214,0.1446870267,0.0649365485,0.5472710133,-0.3608383536,0.2080309838,-0.1670423448,-0.5076099634,-0.1528960019,-0.1149318218,0.1049193591,0.1264861375,-0.3500165343,0.3597266674,0.2949793041,-0.1077529937,-0.0378293172,0.0068124472,-0.1752336174,-0.0723744109,-0.156309545,-0.1104513705,-0.0091842255,0.5706517696,0.0093581229,-0.0183522459,0.1865911186,-0.0138797946,-0.0601734407,-0.1625116765,-0.5331320167,0.2263108343,0.306632638,-0.1381725669,0.014043631,0.6243480444,0.6611810327,-0.1327867806,-0.0027744118,-0.1644930989,0.0180114824,0.0469275117,-0.2306063324,-0.0574955977,0.1657326967,0.04810578,-0.2576411366,-0.0092641786,0.1547838598,0.098936297,-0.0698966458,0.0188649725,-0.0771572068,-0.3292374313,-0.0292537492,0.1232707128,-0.1089475304,0.1232991293,-0.0864138678,-0.0403812714,-0.5168852806,0.1755291075,-0.1288746297,-0.4917961955,0.0165395774,0.4037632346,0.2822576165,-0.1160470992,0.4003125727,0.0940636545,0.0872102231,-0.0527308844,-0.2767204642,0.3507880569,-0.4622550011,0.4693782628,-0.3270950913,0.271769315,0.0570278764,0.1779133976,0.1025838703,0.1646062881,-0.2312485874,-0.2901343405,-0.1018912345,0.0940673798,0.2068044096,0.1520269215,0.0983555764,0.2725200355,0.1068393067,-0.2188543677,-0.1683713794,0.2467847019,-0.2557554245,0.1999899596,0.1347340196,-0.0919883773,0.1986120045,0.1977641433,-0.0685090199,0.2219430655,-0.1261403114,-0.0883877352,-0.0391468555,0.1671304852,-0.0806709304,0.0134148505,0.1874517947,0.2687897384,-0.001930296,-0.0806286037,0.6413139105,0.3312113583,0.1150878593,0.5001331568,-0.051440578,-0.0175249744,-0.4890014529,-0.2775426805,0.2126242667,0.0310199838,0.027646685,0.0917783976,-0.0246267356,-0.0275032893,-0.135947302,0.002769843,0.5822210312,-0.2275114357,-0.0127435261,0.0468721949,-0.02021355,-0.3827409446,0.3991596401,-0.1654493362,-0.2494139373,0.0360155739,-0.0512168407,0.0068901936,-0.0835429505,0.2543728948,0.3149311841,-0.2773489654,0.0299966112,0.3806999028,-0.1358973086,0.2611198425,-0.1441325247,-0.2044881582,0.0264261067,-0.2047907114,0.105783008,0.418795079,-0.0145583153,0.0733329803,-0.0609219447,0.1031606793,0.2356014252,-0.0655149519,-0.2108763009,0.4138633311,-0.2922664583,0.2327293903,0.2990344465,-0.3924899697,0.0951183811,0.4452514946,0.0023688325,0.4268141091,-0.0494985953,0.5840528607,0.0175101068,-0.2804224789,0.1194979697,0.4449959993,-0.2014257461,-0.1196044758,-0.1712664217,-0.1596518755,-0.1877388805,0.090019457,-0.0751736015,0.0028118589,0.2173173428,0.0717944652,0.0639649704,0.0240853317,-0.0795658976,0.2913348675,0.2723922729,-0.0090198955,-0.1765927523,0.3458016813,-0.1892184317,0.0963714942,0.6774992347,-0.0281568971,0.017948227,-0.3747056127,0.6705471873,0.2169339061,-0.1634797454,0.0480992757,-0.0109066209,0.0236371811,0.0606255494,-0.0125561543,0.0033276803,-0.0127487667,0.1845303774,-0.1218116954,0.1319627762,-0.4920049012,0.1160301492,-0.0095948102,-0.062035922,-0.2028695792,0.1745490432,-0.2787022889,-0.3275158405,0.1406188458,-0.109123677,0.0932399631,0.0925557986,-0.032507766,-0.1501616389,0.2595950365,0.3713560104,0.4355632663,-0.1231247187,-0.1570314169,-0.091663681,0.0360565111,0.2634945214,-0.2674753368,0.3377637863,0.0507329926,0.0305756424,0.032336276,-0.0640156642,-0.1622574478,-0.2586784661,0.0873731747,-0.3793010116,-0.2062282413,-0.0468783677,0.2062362283,0.227014482,-0.3029272854,0.0698186308,0.3544505537,-0.1974746287,-0.0493625961,0.4218857586,-0.0391922258,0.0747133642,-0.0641381815,0.1939850748,-0.0614002123,-0.0922132432,-0.0477710813,0.5073210001,0.1155316234,-0.3652403355,0.3364468515,-0.0941130444,0.1249055564,-0.2463935316,-0.3963310122,-0.0742866546,0.2382933646,0.5800231695,-0.4254644513,-0.2457300872,0.1356075853,0.0331823193,-0.0066491105,0.326990515,0.812299788,-0.1088394523,-0.0509543754,-0.1931656897,-0.0889547914,0.5671644211,-0.2700323761,-0.2003219724,0.0761687234,0.1930247992,-0.0961933509,-0.3720490336,-0.7107865214,0.1285331249,0.259074837,-0.1107787639,-0.0877092108,0.4188708961,-0.0624388754,-0.122663945,0.1063420847,0.0659079403,0.3857209682,-0.3582318127,0.2123966366,-0.2672464252]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2767","title":"equal operation to perform unbatch for huggingface datasets ","comments":"Hi @mariosasko \r\nI think finally I got this, I think you mean to do things in one step, here is the full example for completeness:\r\n\r\n```\r\ndef unbatch(batch):\r\n new_batch = collections.defaultdict(list)\r\n keys = batch.keys()\r\n for values in zip(*batch.values()):\r\n ex = {k: v for k, v in zip(keys, values)}\r\n # updates the passage.\r\n passage = ex['passage']\r\n passage = re.sub(r'(\\.|\\?|\\!|\\\"|\\')\\n@highlight\\n', r'\\1 ', passage)\r\n passage = re.sub(r'\\n@highlight\\n', '. ', passage)\r\n inputs = f\"record query: {ex['query']} entities: {', '.join(ex['entities'])} passage: {passage}\"\r\n # duplicates the samples based on number of answers.\r\n num_answers = len(ex[\"answers\"])\r\n num_duplicates = np.maximum(1, num_answers)\r\n new_batch[\"inputs\"].extend([inputs] * num_duplicates) #len(ex[\"answers\"]))\r\n new_batch[\"targets\"].extend(ex[\"answers\"] if num_answers > 0 else [\"\"])\r\n return new_batch\r\n\r\ndata = datasets.load_dataset('super_glue', 'record', split=\"train\", script_version=\"master\")\r\ndata = data.map(unbatch, batched=True, remove_columns=data.column_names)\r\n```\r\n\r\nThanks a lot again, this was a super great way to do it.","body":"Hi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n","comment_length":131,"text":"equal operation to perform unbatch for huggingface datasets \nHi\r\nI need to use \"unbatch\" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve:\r\n\r\nI am considering \"record\" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to make it similar to what T5 originally did:\r\n\r\nhttps:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/3c58859b8fe72c2dbca6a43bc775aa510ba7e706\/t5\/data\/preprocessors.py#L925\r\n\r\nHere please find an example:\r\n\r\n For example, a typical example from ReCoRD might look like\r\n {\r\n 'passsage': 'This is the passage.',\r\n 'query': 'A @placeholder is a bird.',\r\n 'entities': ['penguin', 'potato', 'pigeon'],\r\n 'answers': ['penguin', 'pigeon'],\r\n }\r\n and I need a prosessor which would turn this example into the following two examples:\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'penguin',\r\n }\r\n and\r\n {\r\n 'inputs': 'record query: A @placeholder is a bird. entities: penguin, '\r\n 'potato, pigeon passage: This is the passage.',\r\n 'targets': 'pigeon',\r\n }\r\n\r\n\r\nFor doing this, one need unbatch, as each entry can map to multiple samples depending on the number of answers, I am not sure how to perform this operation with huggingface datasets library and greatly appreciate your help\r\n\r\n@lhoestq \r\n\r\nThank you very much.\r\n\nHi @mariosasko \r\nI think finally I got this, I think you mean to do things in one step, here is the full example for completeness:\r\n\r\n```\r\ndef unbatch(batch):\r\n new_batch = collections.defaultdict(list)\r\n keys = batch.keys()\r\n for values in zip(*batch.values()):\r\n ex = {k: v for k, v in zip(keys, values)}\r\n # updates the passage.\r\n passage = ex['passage']\r\n passage = re.sub(r'(\\.|\\?|\\!|\\\"|\\')\\n@highlight\\n', r'\\1 ', passage)\r\n passage = re.sub(r'\\n@highlight\\n', '. ', passage)\r\n inputs = f\"record query: {ex['query']} entities: {', '.join(ex['entities'])} passage: {passage}\"\r\n # duplicates the samples based on number of answers.\r\n num_answers = len(ex[\"answers\"])\r\n num_duplicates = np.maximum(1, num_answers)\r\n new_batch[\"inputs\"].extend([inputs] * num_duplicates) #len(ex[\"answers\"]))\r\n new_batch[\"targets\"].extend(ex[\"answers\"] if num_answers > 0 else [\"\"])\r\n return new_batch\r\n\r\ndata = datasets.load_dataset('super_glue', 'record', split=\"train\", script_version=\"master\")\r\ndata = data.map(unbatch, batched=True, remove_columns=data.column_names)\r\n```\r\n\r\nThanks a lot again, this was a super great way to do it.","embeddings":[-0.0244649611,-0.8023660183,0.0363254026,-0.0556131192,0.0216097701,-0.0733311102,0.2536010146,0.0494495481,0.3815232515,0.241488114,-0.3534253836,-0.0654722452,0.0393712558,0.4445899725,0.1729435027,-0.2635638416,0.0600867756,0.1323877722,-0.2501762509,-0.0962025076,-0.1124165952,0.0328882895,-0.5142920017,-0.179090023,0.3367595375,-0.2938053906,-0.1501288116,-0.25004673,-0.1282937527,0.3018818796,0.4609180391,0.2909281254,0.2123069316,0.4509742856,-0.0001249902,0.292660445,-0.2429701388,-0.1337586045,0.1145015657,-0.0880547985,-0.0968523026,-0.094203338,-0.0344842076,-0.3119998872,-0.1772888601,-0.0073471898,0.0576628409,-0.3176602423,0.6801812053,0.1236015633,0.0111872517,0.2365360111,-0.0750250593,0.0690964758,-0.1374193281,0.0161666367,-0.0209393706,0.0664196014,0.0917303488,0.0396539941,0.1078623459,0.3016746044,-0.0618220083,-0.0389599092,0.0509128645,0.130355835,0.0570813902,-0.4106472731,0.1430849731,0.371230334,0.1794088334,-0.3381175995,-0.5565244555,-0.0625604987,0.0454883538,-0.1229719147,-0.2359695733,0.1706840247,-0.1305094957,0.2065273523,-0.410728395,-0.0659740493,-0.0432806909,0.0421258323,0.076096341,-0.0634093508,-0.2866249084,0.2641929984,0.0687270314,0.118497096,-0.1335151792,-0.1068736017,0.1343878061,0.2130965739,-0.6556074023,-0.2227413803,-0.243380785,-0.0494689941,0.278621912,0.3848898113,0.5002590418,-0.0848697722,-0.0207165629,0.0651059002,0.0739711523,0.5639448166,-0.0917669907,-0.2520310283,-0.0166104343,0.1402390152,0.0175611489,-0.0419357605,0.3055624664,-0.0990368202,-0.2601804733,0.2037353814,0.4017173946,-0.071072191,-0.3239939213,-0.1834992617,-0.3271969855,-0.1466225237,-0.0070036929,0.1262951344,-0.1984352022,0.0695280209,0.1843550801,0.0698361024,-0.2867665887,-0.0619153231,-0.1795290262,0.2524328232,-0.1391302496,-0.0016128101,-0.0439372435,-0.2109626979,0.0124201532,0.4680207968,0.0447812714,-0.0492153391,0.222948581,-0.1477324069,0.2782618105,-0.0027510433,0.1948138475,-0.1123818979,0.1089916527,0.0534727201,-0.0731872618,-0.0069980486,-0.0442639664,-0.2340401858,0.1050291061,-0.0208555199,-0.2917470336,0.0252854489,-0.2057438791,0.2881164253,-0.3421461284,0.4115400612,0.4393236637,0.3590464294,-0.0536828786,-0.0564322136,0.4131744206,-0.019270733,-0.0965282768,-0.26799196,0.3482652009,0.0231547076,0.3075779974,0.2882292271,-0.0513803102,0.3208956718,-0.2531568706,-0.0951119587,0.2231654227,-0.1747700125,-0.4179792106,0.2593773901,-0.0811822712,0.2943813801,0.0126556782,0.0581352301,-0.1151295975,0.1797288209,-0.0140454564,0.2801768184,-0.0816658288,-0.0973386616,-0.0200460423,-0.1788855791,0.3267515302,-0.0937372372,-0.0511453375,0.3327536583,0.041573327,-0.1167278588,0.2596202195,-0.4877478182,0.181680128,0.2622479498,0.2451246232,0.1432557255,0.2428885102,-0.152628094,-0.3867460489,0.1505290866,-0.3342051208,0.1509622037,0.0888974518,-0.3247162998,-0.1425441802,-0.2196986079,-0.1516406536,-0.0205253456,-0.0847825333,-0.113110289,0.4627571106,0.0237465501,-0.3968051374,0.276607275,-0.082532309,0.0979911983,-0.4222601652,0.0488031805,0.2384614348,0.0088137798,-0.0059175733,0.1739051789,0.0389285833,-0.3525582552,0.0861845389,0.2756893635,-0.3201452792,0.007060708,-0.0751751736,0.0669735745,0.0377308093,-0.0897159949,-0.1346691698,0.0368004777,0.0751844123,-0.0604840592,-0.0372871533,0.7736294866,-0.0813849419,0.1259882301,0.1066019908,-0.0489501767,-0.1498593986,-0.2755477428,-0.1673119962,0.2761720717,0.2356814444,-0.2201273739,0.2520665228,-0.1806640178,-0.2266646326,-0.2730321288,0.2895759344,0.0222219992,0.3233869076,0.1219824627,0.0554224737,0.2160241306,-0.0402212627,-0.1559975892,0.3685371578,-0.003990531,0.254512459,0.3580181301,0.2879855633,0.152347818,0.0509685427,-0.2573584616,-0.074441731,0.0054119737,-0.010024149,0.1193401515,-0.3581905961,0.0063078338,0.0412622169,0.1751889735,0.0069136517,-0.0505397953,-0.1088419259,-0.1534914821,-0.1993155479,-0.0013042735,-0.2027613223,-0.047777351,-0.0915739834,0.0144527415,0.0921797231,0.4318853021,0.2483216375,0.4400375187,-0.0472345874,0.3629509509,-0.1473098248,0.0066064852,0.0675346255,-0.0385228246,-0.1326177418,-0.2288424373,0.2457871139,-0.0202647522,0.0167320855,-0.5255942941,-0.1940316558,0.2381523252,-0.0047633923,-0.23789832,0.3722145557,0.0893638283,-0.5399628878,-0.111228779,0.018306395,-0.1314908564,-0.5532984734,0.2090689838,-0.1560288221,0.131594196,-0.2751115561,-0.3274399936,-0.1012824178,-0.1940861493,0.9283853173,-0.3411178291,0.0465672724,0.1961683631,-0.2678838968,0.073897168,0.0258137938,0.1224486679,-0.0997033864,-0.5181102157,-0.1364708543,0.0263615157,-0.1522428393,-0.1412647218,-0.063107878,0.1339441687,0.5387528539,-0.0023167669,-0.2072089016,-0.4411291778,-0.227887556,-0.1010209024,0.4474504292,0.1770002246,-0.3411288857,0.1188594401,-0.0365313701,-0.053765934,0.2113044113,-0.0068767234,-0.1427294016,0.1921971738,0.025802467,0.3620700836,0.3707354367,0.3190152943,-0.0835306942,0.0542934053,-0.2608816326,-0.1027698666,0.0708574727,-0.3895360231,0.0945675373,0.136618644,0.2134462148,0.064443633,-0.211458236,0.3141698539,-0.1198179498,-0.0176075287,-0.3152264953,-0.3304156065,0.0021315089,-0.1976332664,0.2163100541,-0.1318219304,-0.1573126167,-0.0392751396,-0.0884592235,0.2690512538,0.2098579705,0.1098339781,0.1071186438,-0.4077515006,-0.3052113652,-0.5409753323,0.6420145631,-0.0964473858,0.2413145304,0.0919438303,-0.0866520405,0.0848415345,0.066159308,1.0803837776,-0.0872084275,-0.1028820053,-0.4584878087,-0.5351157188,-0.352935195,0.1782926619,-0.2240314633,0.5216464996,-0.2015162557,0.4366714954,-0.3163002729,-0.2294137627,0.3831054866,-0.0391894169,-0.0868258774,-0.114705801,-0.1034919843,-0.6376012564,-0.10896907,0.0142843528,0.1750251651,0.1664241105,0.0678805187,0.1198185086,-0.3701828122,0.2059961855,0.353264004,-0.1239747629,-0.018102359,0.1862320006,-0.094244495,0.0939731672,0.3174675405,-0.2403257191,0.1054546833,-0.0841723755,-0.511026144,-0.1814254373,0.1448629498,0.4306387901,-0.0065280613,0.0908252075,0.4707623124,0.2709288895,0.1709093899,-0.2007093132,0.0886756629,0.0992484316,0.469899863,-0.213220939,-0.3708055019,0.2332923561,0.2992036939,-0.1240188703,0.3007999957,0.3552348912,-0.1722936034,0.0196966548,-0.0850256756,0.9305514097,-0.2057654113,0.0553850867,0.166240558,0.1524593234,0.5891494155,-0.311837703,0.242113471,-0.1372167021,-0.5443879962,-0.1400336623,-0.1119210497,0.1085160747,0.1334474385,-0.3048271239,0.3906707168,0.3042517602,-0.0892803743,-0.0534521714,-0.0337566845,-0.1932050735,-0.088241227,-0.1691993326,-0.1124444678,-0.0340409428,0.5939669013,-0.0418732166,0.0029528579,0.1844166815,-0.0214712452,-0.0345331058,-0.1751814038,-0.5350192189,0.2652752399,0.2661238611,-0.1540381163,-0.0759143904,0.5716874599,0.6944461465,-0.1355420947,0.0254093241,-0.1389091015,0.0475386307,0.0059843841,-0.2499772459,-0.079294689,0.1179979146,0.0279094744,-0.2371240854,-0.0089593446,0.1250245571,0.0945895314,-0.0359660909,0.0241513439,-0.0430498682,-0.2812277377,-0.038163662,0.1298306286,-0.1041664854,0.0669883564,-0.0777908415,-0.0373957939,-0.5551764369,0.1727187932,-0.1398658752,-0.4718235433,0.0163040496,0.3586140275,0.2728106081,-0.1444614977,0.4105927646,0.0793774128,0.1235169992,-0.0582536422,-0.3121371567,0.2812635601,-0.4466728866,0.4826536477,-0.3332633078,0.2762415111,0.0468734056,0.1860536933,0.099235408,0.1407223493,-0.2239490598,-0.259115696,-0.1163416281,0.0624319836,0.2013016343,0.2372535616,0.100003697,0.3003735542,0.0758328289,-0.1464378089,-0.1659802049,0.2176052928,-0.2373042554,0.1934839189,0.1431108266,-0.0761533529,0.251360476,0.2321752012,-0.0578263067,0.2195382118,-0.0989718288,-0.1025244892,-0.0458366647,0.1538626999,-0.0928682759,0.0233610477,0.2305043191,0.3027408719,-0.0158406328,-0.0984349623,0.6481759548,0.29995206,0.1284629554,0.5035308599,0.0074528884,-0.0021046081,-0.4534750283,-0.2945229113,0.2117395103,0.0077519231,0.0200660713,0.0980636701,-0.0531902388,-0.0284612756,-0.1261399835,-0.0733382031,0.5804975629,-0.2278455645,-0.0304169338,0.0826798528,-0.0677458271,-0.3896819353,0.4106399417,-0.168134287,-0.2066210061,-0.0145438826,-0.0696037039,0.0262061264,-0.1212480515,0.2267165184,0.3431794941,-0.2564218044,0.02456115,0.3773986697,-0.1493621469,0.3176618516,-0.1793254763,-0.2306487858,0.0359412991,-0.180405125,0.1404880434,0.4447540939,0.0361010879,0.0967202932,-0.0780335814,0.1191921979,0.2925365567,-0.0684206188,-0.2099957466,0.417432934,-0.3103857338,0.238117829,0.2955986559,-0.4263084531,0.0692508221,0.39716506,-0.0107809538,0.3950863183,-0.0795759261,0.57693398,-0.0424113832,-0.2923170924,0.1233978122,0.4711918235,-0.2232030332,-0.1391162872,-0.159606263,-0.1254362166,-0.0932332277,0.0786693841,-0.0378241204,0.023122387,0.1900473982,0.0685061216,0.1057491675,0.0426235944,-0.0970226526,0.2900155783,0.2817672491,-0.0030866901,-0.2167579234,0.3081443012,-0.2293579131,0.0245838761,0.6840304136,-0.0212529749,-0.0011890217,-0.33669433,0.6613734365,0.23841919,-0.1470415741,0.0022663674,-0.0645725504,0.027070811,0.0554500036,-0.0251231138,0.0127936238,-0.0312582254,0.1785208434,-0.120761618,0.1366900951,-0.4874284863,0.1445138305,-0.0462260507,-0.0735687688,-0.222711727,0.2186488807,-0.2597885132,-0.3497095108,0.1194812357,-0.108291164,0.0792769939,0.0722764209,-0.0218795743,-0.1936153024,0.2946291268,0.3574353456,0.4273125827,-0.1635020822,-0.1160747185,-0.107279636,0.0179323144,0.2351539433,-0.2670191824,0.3736535013,0.0878038704,-0.0091180606,-0.019941425,-0.0227677207,-0.1291804761,-0.2303566635,0.0730217248,-0.3877384067,-0.2342642397,-0.0709631592,0.1866690814,0.2684719265,-0.321487993,0.0706241727,0.3622841835,-0.2053527683,-0.0401140191,0.388110131,0.007868208,0.0918074474,-0.0195938218,0.1772874296,-0.0671827495,-0.0797691643,-0.0543443337,0.4619969428,0.1400754005,-0.349322319,0.3237389624,-0.1057210639,0.1021877825,-0.2573279142,-0.3548609018,-0.1049618199,0.2201499343,0.5885673165,-0.4777234197,-0.2191802412,0.1400903463,0.024754934,0.0037987174,0.3690429926,0.8464453816,-0.1257858276,-0.0567209311,-0.2206315249,-0.0846719444,0.5872298479,-0.2355547845,-0.1941432655,0.0436771102,0.2272451967,-0.0792713091,-0.3721283972,-0.6097920537,0.1408359706,0.274710536,-0.095180124,-0.1169880629,0.4490294158,-0.0427572541,-0.1332986355,0.1149808764,0.0669758096,0.3785043657,-0.3848003745,0.1986691058,-0.3128778338]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2765","title":"BERTScore Error","comments":"Hi,\r\n\r\nThe `use_fast_tokenizer` argument has been recently added to the bert-score lib. I've opened a PR with the fix. In the meantime, you can try to downgrade the version of bert-score with the following command to make the code work:\r\n```\r\npip uninstall bert-score\r\npip install \"bert-score<0.3.10\"\r\n```","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\npredictions = [\"hello there\", \"general kenobi\"]\r\nreferences = [\"hello there\", \"general kenobi\"]\r\nbert = load_metric('bertscore')\r\nbert.compute(predictions=predictions, references=references,lang='en')\r\n```\r\n\r\n# Bug\r\n`TypeError: get_hash() missing 1 required positional argument: 'use_fast_tokenizer'`\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version:\r\n- Platform: Colab \r\n- Python version:\r\n- PyArrow version:\r\n","comment_length":48,"text":"BERTScore Error\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\npredictions = [\"hello there\", \"general kenobi\"]\r\nreferences = [\"hello there\", \"general kenobi\"]\r\nbert = load_metric('bertscore')\r\nbert.compute(predictions=predictions, references=references,lang='en')\r\n```\r\n\r\n# Bug\r\n`TypeError: get_hash() missing 1 required positional argument: 'use_fast_tokenizer'`\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version:\r\n- Platform: Colab \r\n- Python version:\r\n- PyArrow version:\r\n\nHi,\r\n\r\nThe `use_fast_tokenizer` argument has been recently added to the bert-score lib. I've opened a PR with the fix. In the meantime, you can try to downgrade the version of bert-score with the following command to make the code work:\r\n```\r\npip uninstall bert-score\r\npip install \"bert-score<0.3.10\"\r\n```","embeddings":[-0.1357685477,0.1683022529,0.0354213715,0.2092981488,0.3473314941,0.0116616022,0.228991881,0.3530413806,-0.1199427098,0.3440085948,0.0769439936,0.4752254784,-0.1141471267,-0.116274558,-0.0818043575,-0.3451915979,0.0918747634,0.3889160156,0.0978735834,-0.1766297966,-0.2902543843,-0.0321730375,-0.1890511215,0.14117378,-0.2226971686,0.0321636721,-0.2295636237,-0.1066831648,-0.2944519818,-0.4702112675,0.244017005,-0.0262592826,-0.0497719832,0.4349934757,-0.00011068,-0.0363304466,0.2661401331,-0.042828802,0.0534035116,-0.3680902421,-0.5844059587,-0.2591186166,0.0363761783,-0.3373253345,0.0206000004,0.0272910316,-0.0091573633,-0.205401808,0.2047609687,0.3369059265,0.2373391241,0.1403486729,-0.1228036061,-0.3247485459,0.3542505503,-0.2880485058,0.0359722786,0.239534542,-0.2517361939,-0.1955946982,-0.1582666934,0.3248482943,-0.0513245836,0.0017433926,0.2958139181,0.244579941,-0.0856363773,0.0226623286,0.0879988819,0.1774827838,0.0153080029,-0.2262357622,-0.3905266821,0.2721064687,-0.0508809835,-0.2326862812,0.008509024,-0.2376612872,0.0329267941,-0.2067704797,-0.3441335857,-0.1278514266,-0.1945914477,0.0551801622,-0.0131272878,0.3923610151,0.0053676334,0.0391234159,0.3017946482,-0.163325876,-0.0880642459,-0.0763620809,0.104990907,0.2026166916,-0.0702825934,0.0108945398,0.2954882681,0.090971075,0.1854871511,-0.5502127409,0.0220071506,0.0181753691,0.0533534288,0.20646061,0.0191061515,0.3505941331,-0.1541594267,0.216454044,0.2325025052,-0.0649732351,-0.171160236,-0.0547930077,-0.031594649,-0.0788692012,0.489043355,0.1721761376,-0.2210557312,-0.2010472119,-0.2376409769,0.1853826642,-0.4741438329,-0.0452962145,0.210642457,0.4373003542,-0.1917456836,0.1406194717,-0.1900238246,0.0471833199,-0.1788125485,0.1952169985,-0.2828834355,0.2730635405,-0.3546588421,-0.0977147445,0.1049495712,-0.1991661042,0.37559551,0.0675270855,0.4369859695,0.1716331095,-0.1321011335,-0.2742588222,0.1550827473,-0.0190750062,0.0458559394,-0.0588260926,0.1923587322,-0.1388753951,-0.2338359207,0.1322214901,-0.2897829115,-0.3272272348,0.0407599024,0.2200025469,-0.1817957163,-0.0427128002,-0.2142801881,0.0517423637,0.1565991342,-0.2517822087,-0.196469456,-0.2444114089,-0.3549409509,-0.2286157608,0.3791269064,0.1625283957,-0.2698113024,-0.0479812771,-0.1046242043,0.2611281872,0.2512706518,0.193728596,-0.2317173481,0.3143136501,-0.1103289425,0.0838198364,0.3269455433,-0.5421458483,-0.4934960902,-0.2773634493,0.1178701892,-0.1706846952,0.0330090113,0.0672046989,0.0950100273,0.039079804,0.2723297775,-0.1352109462,0.0036767162,0.1628976017,-0.2928624153,-0.1766243279,0.1568981707,-0.0429335237,0.4267222881,-0.2014783621,-0.1615751088,0.2878130078,-0.0385640301,-0.1464606225,-0.0206569377,0.080988571,0.5700151324,-0.286860615,0.2906683087,-0.2317186296,0.0181312673,0.1943648458,-0.0077656615,0.2645762265,-0.2120677233,-0.1155014187,-0.4618649483,0.1162807196,-0.0927695483,0.1053282917,0.1548778564,-0.0557345711,0.2700780928,0.1727413982,-0.0629527047,-0.2985925674,0.0740361884,-0.1582509279,0.0757353231,0.2299352288,-0.2951509356,-0.2042697817,0.3250257969,0.2769492567,0.3096359074,0.0713861287,-0.1819415092,0.1881348491,-0.0460106172,-0.0693427995,0.0917252153,0.2145000547,0.0582688972,-0.4049297869,0.082875289,0.1304061264,0.3137616217,0.0475625284,-0.1744232923,0.4764544368,0.0648464933,0.1258464605,-0.0076349145,-0.0176294688,0.267881155,0.1416642368,0.2068474591,-0.3284746706,0.1557706892,-0.0036571291,0.1681103259,-0.021475194,0.1855958551,-0.1167835146,0.5808283687,0.3723367751,0.0887974873,0.2555126846,-0.227700904,-0.1590948552,-0.1334581673,-0.2853350341,0.3417616487,0.2336738855,-0.0786180198,-0.0336476155,-0.1537461579,-0.0775062963,0.1249396354,0.1540715545,-0.0738965794,-0.0380365551,0.3315090835,-0.0597750694,-0.190403372,-0.1201440543,-0.0203250069,0.4591867328,-0.3168023527,0.2947408557,-0.1316542774,0.1890826821,-0.0493894964,-0.6532367468,-0.059199784,-0.3793861568,0.2373259962,-0.0883880109,-0.0403929539,0.3508841395,0.057935819,0.1776039749,0.1711506397,-0.1074415892,-0.2738973796,-0.5320510864,-0.2634779811,0.0622547679,-0.1156594306,0.0657804832,0.2517986298,-0.2744704187,0.1120415255,-0.1333208084,-0.5113040209,0.0996796861,-0.3408310115,0.4703126848,0.3321333826,-0.1779076606,-0.5641243458,-0.2329175919,0.4654655755,-0.189506948,-0.0704404712,0.2736673951,-0.0602836832,-0.0971532837,0.0020504901,-0.3575296402,0.0341204032,0.0028462575,0.0291580129,0.0796931237,0.1652203798,-0.1709430367,0.2613537312,0.2987735868,-0.0680780932,-0.0711867362,-0.2196063846,-0.4031369686,0.5500207543,-0.1909077018,-0.2825049758,-0.1657274663,-0.169257924,0.198781997,0.0487793349,-0.1322614551,-0.59427917,-0.0986286029,0.1793272197,-0.0764526576,0.2097877115,0.1528716385,0.3122169971,-0.1486094594,-0.2454244941,-0.1619692296,0.1745098978,-0.1710664481,0.1097458526,-0.2303149998,0.3325130045,0.1021188423,0.92150563,0.1928438544,-0.2365649939,0.260007441,-0.1881907582,0.1139804795,-0.1771749705,-0.424674809,0.0952086225,-0.0468004905,-0.1012853161,0.3070327938,-0.0522174649,0.1064388976,-0.2177164406,-0.0757566616,-0.2872132957,-0.3465437889,0.1041225493,-0.055419188,0.2312720418,0.2452373803,0.2311151326,-0.4509416819,-0.2976679504,0.1437606812,-0.3137770295,0.0095746545,-0.1711256355,0.0973925442,-0.3218021393,-0.340377897,0.427082032,0.4881606102,0.2987392545,0.045781035,0.1343841702,-0.0436954387,-0.1290864795,0.4031524956,-0.0241340231,0.1742092371,0.0091173053,0.0497267395,-0.0384635553,0.122155197,-0.4014152884,0.1189657226,0.0262240805,0.4224999845,-0.294880271,0.1139015332,0.1677736491,0.0341985524,-0.1736882925,-0.0720651075,-0.4913318455,-0.106716916,-0.0734213144,0.3320941925,0.0136509165,0.304274708,0.0199649893,-0.0697354823,0.0623824559,-0.2804695964,0.2110203803,0.3583402634,0.195629105,-0.0093369316,0.2691898644,-0.0080667669,-0.23225528,0.5070555806,0.1846844107,-0.247304827,-0.6065497994,-0.1910001785,0.0006405315,0.5480855107,0.0741695613,-0.1902205199,-0.0997796431,-0.0731616318,0.0602185987,-0.0679093152,0.3978590965,0.4066798687,0.0434782095,-0.3172378242,-0.1784374714,0.5495812297,0.2827202678,-0.1120696142,0.0196733698,0.5535880923,-0.2378884107,0.4279626012,0.3329145908,0.9520226717,-0.1216431931,-0.0361436643,0.268777281,0.0591817424,0.5790296793,0.1037805453,0.2235223204,-0.4127732813,-0.2533380687,0.019740941,-0.1117411256,-0.0586737208,-0.1041252688,-0.3604677618,0.4359687269,-0.2815932333,0.2554154694,0.0881394073,-0.0858828574,0.1566115916,-0.193749994,-0.3508883119,0.1715563685,-0.064783141,0.0104419654,-0.1781091094,-0.0838004351,-0.129350841,-0.3548137844,-0.246096015,0.0518612675,-0.0493004397,0.1146284193,-0.1172227412,-0.3084894717,0.0602367409,0.0566899739,0.4155307114,0.4209897816,-0.2582201362,0.2679956257,-0.0648816973,0.1811761558,0.1085587814,0.0717811137,0.1041906998,-0.0552384891,-0.15788275,0.2358063757,-0.0819146931,-0.1071696877,-0.0834335163,0.0156912003,0.2633230388,-0.2350267321,0.0473706797,-0.1392345279,0.1451860815,-0.3295619786,0.1802753955,0.1452372968,-0.1570014954,0.3131348193,0.0018812018,-0.1545768678,0.1093607619,0.4137751758,0.1025107577,-0.0221784804,0.4374265373,-0.1429377645,-0.2596268058,-0.1375302821,-0.161458686,0.1663228571,-0.1670380235,0.1860702783,0.0946158022,0.2306497693,0.0522691682,0.5719162226,0.1177775487,-0.1696172059,0.1363872588,-0.3092880249,-0.1082825512,0.1192801669,-0.0864584893,-0.0076830038,0.0035059617,0.1689174324,-0.1505281776,0.3258641362,-0.3601907492,0.0126508325,-0.4700366855,0.2061266601,-0.2473837733,-0.0173085425,0.2452022731,-0.0230984781,0.1007350758,0.2630828619,-0.1777683049,-0.2205410451,-0.1287358701,0.0916389376,0.1104478836,0.0546092987,-0.0064813052,0.1228998378,-0.1365970671,-0.2155858278,-0.1219802871,0.2991619706,-0.0076051247,0.2453022748,0.1821097881,0.1080811024,0.1998056173,-0.2148442417,0.1180736795,0.1997876018,-0.0293727759,0.2218607664,-0.0079315035,-0.1382147223,0.2635356188,0.0294998083,0.4643994868,-0.0224980228,0.2424798608,-0.1946604103,-0.1003611311,0.1491227597,0.2866361439,0.3719915748,-0.1622954756,-0.1772243679,0.3476004601,0.197204262,-0.4130782783,-0.1833365262,-0.0150852064,0.2177569568,0.0468954928,0.1587257087,-0.0353802666,-0.0545021631,0.0274950396,-0.0682402924,0.1834892631,-0.2067254633,-0.0511266589,0.0133193266,-0.073046729,0.1287654042,0.3544396758,0.0895644203,0.3318196237,0.5564888716,-0.2681508362,0.4829030037,-0.2044484466,0.0878312662,-0.1325692683,0.0285283234,-0.107297197,0.1365872622,0.1400083601,0.0589682907,0.0158567447,0.8266776204,-0.0803471059,0.1014607698,-0.3887360692,0.1683852077,-0.0119952066,-0.1171091273,-0.0547749065,-0.2624056339,-0.2106729448,0.081116721,0.015891362,0.124582544,0.4215361476,0.2092754245,0.144351691,-0.2554765344,-0.1748764366,0.4411932528,0.2574707568,-0.2895908654,0.1098924056,0.159224391,-0.1117751077,-0.1280726194,0.3511741459,0.2745372057,0.3522098064,-0.17761603,-0.148036927,-0.2610130608,-0.0431445315,-0.0026830877,0.3742747307,0.1880339533,0.0464068763,0.1489734203,0.1528083235,-0.1990123391,0.2535201907,0.0206502602,0.4681150317,-0.0658477545,0.317407459,0.0398341939,0.0324196406,0.0036228662,-0.2002907991,-0.2993038297,0.1330287755,0.0517547503,0.2607646286,-0.0237016641,-0.2117721587,0.1049003229,0.1395422667,0.440942049,0.279127419,0.1469460428,-0.640408814,-0.0449286811,-0.4518197477,0.1629372835,-0.0208123401,0.3008379638,-0.1564245075,-0.0414830558,-0.1359417588,0.2649830282,-0.0204073358,-0.1017611548,-0.2103607208,0.073061645,-0.4150728285,0.0396690741,-0.3952053487,-0.4520669281,0.1354969144,-0.1934475899,-0.0277335308,-0.3400938511,0.1221163422,-0.2560794055,-0.2969569862,0.1214851886,0.1431155801,0.6734743714,0.2715201676,0.4250793159,-0.0546117947,0.05037608,0.082502611,-0.1654841006,-0.0256075133,-0.0099344309,0.0087098172,0.1472877711,-0.1428937018,0.0028557279,-0.3011906445,0.1760221273,0.0736622289,-0.2087876797,-0.3524345756,0.10069938,-0.0179183614,-0.0920893475,-0.2663926482,0.1667681336,-0.0187939499,0.3740313351,-0.4058200717,-0.2653809488,0.5419955254,-0.5197664499,-0.3598220646,-0.080749765,0.2593533695,0.0481357872,-0.3689701259,-0.7990337014,-0.0074765887,0.3657462597,0.2452615201,-0.2954538763,-0.0031833539,-0.0610721596,0.1361377984,-0.1206355691,0.2416770458,-0.0414120071,-0.1126293764,0.0680974275,-0.2633719742]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2763","title":"English wikipedia datasets is not clean","comments":"Hi ! Certain users might need these data (for training or simply to explore\/index the dataset).\r\n\r\nFeel free to implement a map function that gets rid of these paragraphs and process the wikipedia dataset with it before training","body":"## Describe the bug\r\nWikipedia english dumps contain many wikipedia paragraphs like \"References\", \"Category:\" and \"See Also\" that should not be used for training.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nfrom datasets import load_dataset\r\nw = load_dataset('wikipedia', '20200501.en')\r\nprint(w['train'][0]['text'])\r\n```\r\n\r\n> 'Yangliuqing () is a market town in Xiqing District, in the western suburbs of Tianjin, People\\'s Republic of China. Despite its relatively small size, it has been named since 2006 in the \"famous historical and cultural market towns in China\".\\n\\nIt is best known in China for creating nianhua or Yangliuqing nianhua. For more than 400 years, Yangliuqing has in effect specialised in the creation of these woodcuts for the New Year. wood block prints using vivid colourschemes to portray traditional scenes of children\\'s games often interwoven with auspiciouse objects.\\n\\n, it had 27 residential communities () and 25 villages under its administration.\\n\\nShi Family Grand Courtyard\\n\\nShi Family Grand Courtyard (Ti\u0101nj\u012bn Sh\u00ed Ji\u0101 D\u00e0 Yu\u00e0n, \u5929\u6d25\u77f3\u5bb6\u5927\u9662) is situated in Yangliuqing Town of Xiqing District, which is the former residence of wealthy merchant Shi Yuanshi - the 4th son of Shi Wancheng, one of the eight great masters in Tianjin. First built in 1875, it covers over 6,000 square meters, including large and small yards and over 200 folk houses, a theater and over 275 rooms that served as apartments and places of business and worship for this powerful family. Shifu Garden, which finished its expansion in October 2003, covers 1,200 square meters, incorporates the elegance of imperial garden and delicacy of south garden. Now the courtyard of Shi family covers about 10,000 square meters, which is called the first mansion in North China. Now it serves as the folk custom museum in Yangliuqing, which has a large collection of folk custom museum in Yanliuqing, which has a large collection of folk art pieces like Yanliuqing New Year pictures, brick sculpture.\\n\\nShi\\'s ancestor came from Dong\\'e County in Shandong Province, engaged in water transport of grain. As the wealth gradually accumulated, the Shi Family moved to Yangliuqing and bought large tracts of land and set up their residence. Shi Yuanshi came from the fourth generation of the family, who was a successful businessman and a good household manager, and the residence was thus enlarged for several times until it acquired the present scale. It is believed to be the first mansion in the west of Tianjin.\\n\\nThe residence is symmetric based on the axis formed by a passageway in the middle, on which there are four archways. On the east side of the courtyard, there are traditional single-story houses with rows of rooms around the four sides, which was once the living area for the Shi Family. The rooms on north side were the accountants\\' office. On the west are the major constructions including the family hall for worshipping Buddha, theater and the south reception room. On both sides of the residence are side yard rooms for maids and servants.\\n\\nToday, the Shi mansion, located in the township of Yangliuqing to the west of central Tianjin, stands as a surprisingly well-preserved monument to China\\'s pre-revolution mercantile spirit. It also serves as an on-location shoot for many of China\\'s popular historical dramas. Many of the rooms feature period furniture, paintings and calligraphy, and the extensive Shifu Garden.\\n\\nPart of the complex has been turned into the Yangliuqing Museum, which includes displays focused on symbolic aspects of the courtyards\\' construction, local folk art and customs, and traditional period furnishings and crafts.\\n\\n**See also \\n\\nList of township-level divisions of Tianjin\\n\\nReferences \\n\\n http:\/\/arts.cultural-china.com\/en\/65Arts4795.html\\n\\nCategory:Towns in Tianjin'**\r\n\r\n## Expected results\r\nI expect no junk in the data.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":38,"text":"English wikipedia datasets is not clean\n## Describe the bug\r\nWikipedia english dumps contain many wikipedia paragraphs like \"References\", \"Category:\" and \"See Also\" that should not be used for training.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nfrom datasets import load_dataset\r\nw = load_dataset('wikipedia', '20200501.en')\r\nprint(w['train'][0]['text'])\r\n```\r\n\r\n> 'Yangliuqing () is a market town in Xiqing District, in the western suburbs of Tianjin, People\\'s Republic of China. Despite its relatively small size, it has been named since 2006 in the \"famous historical and cultural market towns in China\".\\n\\nIt is best known in China for creating nianhua or Yangliuqing nianhua. For more than 400 years, Yangliuqing has in effect specialised in the creation of these woodcuts for the New Year. wood block prints using vivid colourschemes to portray traditional scenes of children\\'s games often interwoven with auspiciouse objects.\\n\\n, it had 27 residential communities () and 25 villages under its administration.\\n\\nShi Family Grand Courtyard\\n\\nShi Family Grand Courtyard (Ti\u0101nj\u012bn Sh\u00ed Ji\u0101 D\u00e0 Yu\u00e0n, \u5929\u6d25\u77f3\u5bb6\u5927\u9662) is situated in Yangliuqing Town of Xiqing District, which is the former residence of wealthy merchant Shi Yuanshi - the 4th son of Shi Wancheng, one of the eight great masters in Tianjin. First built in 1875, it covers over 6,000 square meters, including large and small yards and over 200 folk houses, a theater and over 275 rooms that served as apartments and places of business and worship for this powerful family. Shifu Garden, which finished its expansion in October 2003, covers 1,200 square meters, incorporates the elegance of imperial garden and delicacy of south garden. Now the courtyard of Shi family covers about 10,000 square meters, which is called the first mansion in North China. Now it serves as the folk custom museum in Yangliuqing, which has a large collection of folk custom museum in Yanliuqing, which has a large collection of folk art pieces like Yanliuqing New Year pictures, brick sculpture.\\n\\nShi\\'s ancestor came from Dong\\'e County in Shandong Province, engaged in water transport of grain. As the wealth gradually accumulated, the Shi Family moved to Yangliuqing and bought large tracts of land and set up their residence. Shi Yuanshi came from the fourth generation of the family, who was a successful businessman and a good household manager, and the residence was thus enlarged for several times until it acquired the present scale. It is believed to be the first mansion in the west of Tianjin.\\n\\nThe residence is symmetric based on the axis formed by a passageway in the middle, on which there are four archways. On the east side of the courtyard, there are traditional single-story houses with rows of rooms around the four sides, which was once the living area for the Shi Family. The rooms on north side were the accountants\\' office. On the west are the major constructions including the family hall for worshipping Buddha, theater and the south reception room. On both sides of the residence are side yard rooms for maids and servants.\\n\\nToday, the Shi mansion, located in the township of Yangliuqing to the west of central Tianjin, stands as a surprisingly well-preserved monument to China\\'s pre-revolution mercantile spirit. It also serves as an on-location shoot for many of China\\'s popular historical dramas. Many of the rooms feature period furniture, paintings and calligraphy, and the extensive Shifu Garden.\\n\\nPart of the complex has been turned into the Yangliuqing Museum, which includes displays focused on symbolic aspects of the courtyards\\' construction, local folk art and customs, and traditional period furnishings and crafts.\\n\\n**See also \\n\\nList of township-level divisions of Tianjin\\n\\nReferences \\n\\n http:\/\/arts.cultural-china.com\/en\/65Arts4795.html\\n\\nCategory:Towns in Tianjin'**\r\n\r\n## Expected results\r\nI expect no junk in the data.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n\nHi ! Certain users might need these data (for training or simply to explore\/index the dataset).\r\n\r\nFeel free to implement a map function that gets rid of these paragraphs and process the wikipedia dataset with it before training","embeddings":[-0.0217499249,0.218051672,-0.1063896716,0.4800746441,0.2548156381,0.1268990338,0.3188252151,0.3147780299,0.210529834,0.0805180967,-0.1847252399,0.2396585047,0.3409660161,-0.222655803,0.2059661597,-0.3597016037,0.2095890939,0.0540651046,-0.3535144031,-0.2804773152,-0.2011397183,-0.0445804894,-0.2302432656,0.25117594,-0.3913946152,-0.0602716692,-0.1513203532,0.0464148596,-0.0278708208,-0.2779792249,0.0563255362,-0.2114447802,-0.2633556426,0.373516947,-0.0001048581,-0.035167627,0.3287271857,0.0319341868,-0.181307897,0.161454156,-0.4102666676,-0.2440759987,-0.177390784,-0.429876864,0.0149512319,-0.1356499642,0.0609836355,0.1326296329,0.2397331744,0.2525061369,0.2651383877,-0.192565009,-0.0854851902,-0.1452975869,0.230780825,0.2006862164,-0.1339762956,0.1043765694,-0.1793815941,-0.0810348317,-0.3428086936,0.4742647111,-0.1878856868,-0.0846635252,-0.0016512723,0.0582641847,0.003401729,-0.1630994231,0.3464976549,0.4074309766,0.3778395355,-0.2557986677,-0.2104946524,0.0359945074,-0.1938414872,-0.0199063215,0.2206408828,0.4031757116,-0.1245583221,0.2314280868,-0.0001320894,-0.1658485681,-0.1803619564,0.3355080485,-0.0778018385,0.3513061702,0.0784134865,-0.0682719573,-0.1236614957,-0.2863451242,0.1012386754,-0.1975318342,0.0682558045,0.3470073342,0.0446781218,0.052592054,0.1136643887,0.2080913633,0.2858697772,-0.1333636343,-0.2652031183,-0.0882652476,-0.1367068142,0.0373129472,0.1167579442,-0.2515017986,-0.0322292931,0.1256176084,0.2220259458,0.156022504,-0.106958203,0.2413073927,0.1984235495,-0.2149846852,-0.0874885544,-0.1044268012,0.0813943073,-0.1835741401,-0.3046195209,0.5948886275,-0.3711348474,-0.1318762749,-0.2979770899,0.3457996249,-0.1281636208,-0.0458850078,0.1415418983,0.0513534211,-0.2917692959,-0.1837873012,-0.2216851264,0.2715581059,-0.1283211112,0.0329070985,-0.004677685,-0.1679580659,0.208050102,0.3736848533,-0.1623074114,-0.1798124909,-0.3341544867,-0.2787850797,0.1935303211,0.1823902726,-0.1924865395,-0.0175302457,0.1466522515,-0.466809839,-0.0528691001,0.21004498,-0.1489783525,0.0081559578,-0.3441658616,0.2914662659,-0.0529970787,-0.0388295949,0.2570791245,0.158239007,0.3505287468,0.1839469671,0.3142883182,0.3185166121,-0.1166176945,-0.0557745323,0.1568236202,0.3804237247,-0.1299568713,-0.2104954273,-0.0621755831,-0.0130195459,0.33330369,0.3461742103,0.0501938872,0.6008312106,-0.0544615202,0.1438074559,0.2068926543,-0.0154538471,-0.3052205145,0.2074411064,0.0376535654,0.0793067887,0.019768076,0.0175007787,0.1682174504,0.2418496162,0.3015216887,0.2445599139,0.3042181432,0.1639282256,-0.3787766695,-0.3602706492,0.332654953,-0.0616248921,0.1865497828,-0.0377615765,0.1737394184,0.4814153612,0.3857314885,-0.1446896344,0.1019091085,0.3164463043,-0.0144040678,0.1753053218,0.2112829834,-0.1105099916,0.0269022789,-0.0118263736,0.1372775584,0.2664980292,-0.2442848831,-0.3616583943,-0.2867643535,0.0922363847,-0.3069060445,-0.3633959889,0.2632481456,-0.0178854689,-0.1031325832,0.2370324135,0.1850254238,-0.0453620292,-0.4526405931,-0.016540071,-0.2000683248,0.2641223669,-0.3693933189,0.1141751856,-0.2537211776,0.2936862409,-0.0100539131,0.257689327,-0.1580757201,0.0685129091,0.0015483522,0.3471232653,0.1509388089,-0.026889421,0.2413432896,-0.5917641521,0.1313875765,0.3716413677,0.132237047,0.0131708914,-0.2830155194,0.1229898632,0.1055339724,-0.0603140444,-0.066121228,-0.0409795716,0.1868358105,-0.1494055092,0.1104725078,-0.3986352682,0.1634108126,0.2137004137,0.0442372411,0.2923955619,-0.5543634295,-0.0755150691,0.3965308666,-0.0039930712,0.4506634772,-0.0317551494,-0.0955358297,-0.1722249091,0.3157480359,-0.1668299884,-0.006180983,0.0669405386,0.1214140356,0.0968192741,-0.1436297745,-0.381041348,0.2992761135,0.0703402981,0.1209891886,0.0078817969,0.1264990121,0.1745315045,-0.4669395089,0.2897341847,-0.1302144378,0.2766365707,-0.5091982484,0.0974701717,-0.1547906697,-0.4750180244,-0.2219113111,-0.0872937739,-0.1623680741,-0.1715699434,0.0400904901,-0.2734844089,-0.3408176303,0.2923766971,0.0714333206,0.1096787453,0.1521874815,-0.0311822668,-0.1206420511,-0.3250731826,-0.1509714723,0.1725412905,0.1331773698,0.1109377518,0.299443543,-0.2933179736,-0.1828276962,-0.5233896971,-0.4014012516,0.2149630487,-0.0341928676,0.1937803179,-0.0156564992,0.3246045411,-0.1813620329,-0.2817063928,0.1334258765,0.1305948943,-0.0667662844,0.0104153864,-0.2001212835,0.1518593282,-0.1306820512,-0.4890944958,-0.0460658856,-0.223866567,0.096482031,-0.052053228,0.3584499061,0.4512475729,0.1462246478,0.1525146812,0.1314287633,0.2828705013,-0.3825257719,-0.0908682272,0.2340758145,-0.1905329227,-0.3137817085,-0.10740982,-0.1780990213,-0.0916708782,0.0535360165,-0.5115405917,0.0990553349,-0.0660326034,0.4441334307,0.2005378455,0.0849117339,-0.0107278209,0.0398279652,-0.0819234997,-0.1840941757,0.209616825,-0.0431088991,-0.1930231005,0.1679575592,-0.0162276011,0.5018720031,0.2536744177,0.5470895767,0.2564651668,0.2209027708,0.1681793779,-0.1750086099,0.2536382973,-0.0764030665,-0.2170144171,0.2947618067,-0.0201398134,-0.1961782277,0.3155837655,-0.190696761,-0.2418267876,-0.0645511597,0.08986222,-0.4327481389,-0.215068832,0.0688226372,0.0117369331,0.2894493937,0.0313613042,0.2015414387,0.165694654,-0.2568355501,0.0745147467,0.0153124565,-0.0868685842,0.0118714711,0.1580387056,-0.2125458121,-0.2947313786,-0.025816448,0.1346649528,0.0923661813,-0.2648640275,-0.1244309321,0.120708473,0.0238640178,0.2426898628,-0.2924233675,0.1627910882,0.1081888303,0.2368687987,-0.2457910329,0.099656783,-0.2826502323,0.3831590712,0.4021661878,0.1113402694,-0.2951384485,-0.0998468325,0.5115591884,0.2186970711,-0.1767619252,0.1225276068,-0.1420457959,-0.1168026999,-0.2688203156,0.3743785322,0.1339215338,0.1794767529,-0.0843837559,0.1417194158,0.1423677653,0.5251242518,0.0477726124,0.1895801872,0.2533740997,0.2911843657,-0.0996597111,-0.0260245055,-0.078183122,0.0705487058,0.4936422706,-0.1975689977,-0.2303668261,-0.2939320505,-0.2943986356,0.0649925023,0.3974117935,-0.0331266709,-0.0881756991,0.1749217063,-0.1704996228,-0.1304392666,-0.0950147733,0.1914545894,0.0488738678,-0.1440315098,-0.4761123061,0.6198274493,-0.1264691055,0.0273769647,0.0217139758,-0.0269501489,-0.4105686545,0.1700829715,0.2546803057,0.9464235902,0.2910658121,0.1372578293,0.3389448524,0.1367831677,0.1745024025,0.1369130164,-0.1498888582,-0.3785795867,0.1422254294,0.0048096394,0.0986764729,0.0799717605,0.0561742783,-0.2926945388,0.0920923576,-0.4051761329,0.148754403,0.2856565416,0.0871682763,0.110602513,-0.2564540505,-0.2089186907,0.1717503518,-0.067459248,0.050156489,-0.2172477096,-0.0479169525,-0.2214342654,-0.0593932942,-0.4298385382,0.1101737618,-0.014606528,0.4323617816,-0.047437381,-0.3589806855,-0.1632196605,0.3142880797,0.0068219728,0.2323222309,-0.2190514505,0.1638811082,0.055122979,0.104590632,0.1837894619,0.2056868821,0.4542143643,-0.0410167836,-0.3303290009,0.0718730465,-0.1509423107,-0.0843537524,-0.1562633514,-0.0905573294,-0.1486929953,-0.1267619431,-0.2030929327,0.0374683961,-0.144942522,-0.3290224671,0.2100215703,0.0266082883,-0.0055104503,-0.0627180487,0.1027225256,-0.4178873301,-0.0673539788,0.1203259081,0.201243937,0.2464583069,0.3070988059,0.338580966,-0.3948931992,-0.3434700966,0.0073800921,-0.2924907804,-0.2570524216,-0.1034270525,-0.2936595678,-0.0158264749,-0.2636457682,0.1282202154,0.015287742,0.0136042871,-0.1466554105,-0.4271979332,-0.0610139258,0.3576854467,0.1551736146,0.0271694176,0.1670845151,0.0822836533,0.206235379,-0.0465490036,-0.4155221581,0.0919839293,-0.3135485947,0.1315243393,-0.0567871779,0.0673396885,-0.2217293829,-0.2155216187,0.1608971953,-0.2391644865,-0.0959363952,-0.312594831,-0.1546696872,0.094618246,0.069693394,-0.098050952,0.0308927577,-0.0046277787,0.1005858183,-0.2756653428,0.0421901159,0.1222029403,-0.0092295622,0.3724422157,-0.0267859027,0.1925994158,0.0317435078,-0.1391220391,-0.0515317135,0.2592616677,-0.0716816708,0.2771883309,0.0490556583,-0.0567286238,0.0147448881,0.0316862911,-0.0732395872,0.0380925871,0.2449267358,-0.2375312001,-0.0414853171,0.3766607344,0.108349897,0.3124468923,-0.3910881877,-0.0867063552,-0.1380978078,0.3295797706,-0.1939186305,0.0695069581,0.0026596144,-0.2876421809,0.0547299348,0.1131540462,0.0939885452,-0.2414284199,-0.0041060247,0.1915750653,0.0746798217,-0.2529520094,0.2269826829,0.3803837001,0.0138364304,0.177862376,-0.0633262172,0.0702819377,-0.0616164319,0.3406343162,0.3166194558,-0.0080495244,0.2981328666,0.2172823995,0.0415306017,0.0035943035,0.3930797577,0.167706117,-0.4627998471,0.0103609664,0.1248842105,0.0746581405,0.3124029934,-0.1110320538,-0.0987591147,0.1237552911,-0.369142741,-0.1521781981,-0.1100705266,-0.1553021967,0.0493780635,0.0431636982,-0.0157786701,-0.2311102152,0.0295922942,-0.0264924616,-0.3002961278,-0.4345846176,0.5299290419,0.1440940797,0.117573902,-0.4787161946,0.1488221288,0.0081930645,-0.0973189026,-0.3266991675,0.6328185797,0.3283832371,0.2006080151,-0.1508840322,0.0393685512,-0.1578694433,-0.1300507039,0.0183743257,-0.0011689853,0.1858093888,-0.1209186986,0.4000377655,0.1490638256,-0.2466212809,-0.1811035275,0.5726721287,0.2986347377,0.0300126858,-0.3170942068,-0.063595742,-0.1525593102,0.0646835417,-0.1351116151,-0.4166483581,0.0175686851,0.5091844797,0.0315330513,0.2170378417,-0.1578041166,0.1260560751,0.1256030351,0.5439803004,0.4771564007,0.3215119541,0.0334349088,-0.2102819085,-0.5221916437,0.1139221638,-0.3307038844,0.0611843728,0.193330124,0.2705697417,-0.0403807126,0.247134164,-0.1648397893,-0.0854507312,-0.0238445532,0.1717416346,-0.2824478745,-0.0814173147,0.1719482243,0.1485818624,-0.2805481851,-0.3322792649,0.3170485795,-0.2754182518,0.1466152966,-0.0726718158,-0.180958882,0.1821578592,0.3728329241,0.3264461458,0.1073773578,0.212205261,-0.2185538709,-0.0122219864,-0.0883901045,-0.2975090444,0.0703276694,0.6526842117,0.0642661527,0.1192352846,0.1520586312,-0.2892920971,0.1741575152,0.5524716973,0.0602064542,-0.0746276975,-0.2985337079,-0.024961615,0.1073964387,-0.0485404916,-0.2636002898,0.1822526455,-0.1529707462,0.0791880041,-0.1050487235,-0.6924180984,0.179192856,-0.474367708,-0.3494185805,-0.2618203163,0.2053394616,0.1779121608,-0.047801245,-0.4104752243,-0.135132134,0.3146881461,-0.1238038763,-0.39163661,-0.0282040834,-0.0388242304,0.1189856157,-0.1689133495,0.21359016,0.0560979061,-0.3986985683,0.0426545031,-0.5303530097]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2762","title":"Add RVL-CDIP dataset","comments":"cc @nateraw ","body":"## Adding a Dataset\r\n- **Name:** RVL-CDIP\r\n- **Description:** The RVL-CDIP (Ryerson Vision Lab Complex Document Information Processing) dataset consists of 400,000 grayscale images in 16 classes, with 25,000 images per class. There are 320,000 training images, 40,000 validation images, and 40,000 test images. The images are sized so their largest dimension does not exceed 1000 pixels.\r\n- **Paper:** https:\/\/www.cs.cmu.edu\/~aharley\/icdar15\/\r\n- **Data:** https:\/\/www.cs.cmu.edu\/~aharley\/rvl-cdip\/\r\n- **Motivation:** I'm currently adding LayoutLMv2 and LayoutXLM to HuggingFace Transformers. LayoutLM (v1) already exists in the library. This dataset has a large value for document image classification (i.e. classifying scanned documents). LayoutLM models obtain SOTA on this dataset, so would be great to directly use it in notebooks.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":2,"text":"Add RVL-CDIP dataset\n## Adding a Dataset\r\n- **Name:** RVL-CDIP\r\n- **Description:** The RVL-CDIP (Ryerson Vision Lab Complex Document Information Processing) dataset consists of 400,000 grayscale images in 16 classes, with 25,000 images per class. There are 320,000 training images, 40,000 validation images, and 40,000 test images. The images are sized so their largest dimension does not exceed 1000 pixels.\r\n- **Paper:** https:\/\/www.cs.cmu.edu\/~aharley\/icdar15\/\r\n- **Data:** https:\/\/www.cs.cmu.edu\/~aharley\/rvl-cdip\/\r\n- **Motivation:** I'm currently adding LayoutLMv2 and LayoutXLM to HuggingFace Transformers. LayoutLM (v1) already exists in the library. This dataset has a large value for document image classification (i.e. classifying scanned documents). LayoutLM models obtain SOTA on this dataset, so would be great to directly use it in notebooks.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\ncc @nateraw ","embeddings":[-0.4235295653,-0.6050093174,-0.0136897219,0.1312966496,0.0744160488,0.0326845311,0.1773322076,0.1531055421,-0.083571814,0.018977087,-0.3533639014,0.2439338863,-0.0217019934,0.072023496,0.2045475096,-0.3612084687,-0.0637265816,0.0468908697,-0.2771911323,-0.1904319972,-0.1820402145,0.3107316494,0.0541405752,-0.1580822915,-0.4117602706,-0.2349641472,-0.1937828958,0.1136880443,-0.3480368257,0.0108187413,-0.0922602713,0.1320497394,0.1466498971,0.2647447586,-0.0001065592,0.0187948979,-0.01774415,-0.1133833528,0.2115951329,0.0164041277,0.2746759057,-0.3394529819,-0.2739129066,-0.3155261874,-0.3253223002,-0.258105129,-0.138216123,0.1185634732,0.4089779556,0.1404997259,0.2868819237,0.2546234131,0.1054849476,-0.0785947666,0.2497178316,0.272551626,-0.30292961,0.1786174923,0.2321746647,-0.1573553234,-0.2770443857,0.6161986589,0.1204697415,-0.2030275613,0.4005105793,0.2674842775,-0.0644168928,0.006494835,0.0080439979,0.2896797061,0.2110324353,-0.2218527496,-0.2832958996,-0.243441686,0.0270570256,-0.3127811551,-0.0312139876,0.2478997111,-0.1373451203,0.0465247966,-0.4051018059,-0.2457090765,-0.1211205274,-0.0998147726,-0.0462005436,0.3456833959,-0.1590660512,-0.1461544931,0.4216436446,0.0622891895,-0.1349771172,0.075460352,0.0543553382,-0.164660126,-0.2244494408,-0.1001714244,0.2504344583,0.1095959917,0.1943096519,-0.0497705527,0.4354341626,-0.1800407618,0.0954850987,0.091192916,-0.1943182945,0.4253619313,-0.2095894665,0.2584174275,0.1072037593,0.2046618015,0.3382173777,-0.012444023,-0.011810841,-0.0572707132,-0.2496788353,-0.224672243,0.0288942568,0.0406937599,-0.1288602948,0.0884686559,0.2069873512,0.1692823619,-0.0215231683,0.3205248117,0.2462232411,0.0059604123,0.362526536,0.0064875865,-0.2586929798,-0.0006943267,-0.1274559349,0.2836678922,-0.1692771018,0.1572210044,-0.019413678,0.0964952484,0.0319395065,-0.2029990554,-0.2654071748,0.1976125091,-0.0670025796,-0.0260688998,0.1944788843,0.1633542776,-0.2581086457,-0.2605816424,0.0050856099,0.2021829933,-0.285801053,-0.0847831666,0.1388389766,0.0938278213,-0.0575366132,0.2118556052,0.0101774288,-0.1059736013,-0.0671047494,0.4708725512,0.0006015004,0.1306479424,0.241032362,0.015112116,-0.3328617215,0.1309047639,-0.0080846287,0.3574265838,-0.1159758717,-0.2120061517,-0.011592431,-0.0486473292,-0.1178015172,0.2772051096,-0.0109583354,0.2765824497,-0.078292042,-0.0711971894,-0.00372109,-0.3285591006,-0.4364301562,0.0852158517,-0.2690641582,0.0037181759,0.1436513066,0.330445528,0.0768363699,0.1628699601,-0.1698916405,0.1437911838,-0.1331208199,0.1523771435,-0.2623205483,-0.2100396901,-0.0922902673,0.3021109402,0.0955509618,-0.1019924805,-0.2069029957,-0.1136794016,0.3012248576,-0.2424896806,0.1169848293,0.2169220299,0.378002733,-0.0966484323,-0.0097397566,-0.2184305787,-0.3532038331,-0.0661710799,0.0501192212,0.3007566929,0.0759040415,-0.1708235443,-0.132664144,0.0142906392,-0.0186755043,-0.3222814798,0.2013837248,0.1845606565,0.2076387703,-0.1080814824,-0.0531414561,0.2346434295,-0.0867497399,0.2332694829,-0.3378702998,0.1478333473,-0.0336149372,0.1141727865,0.1592749357,0.2590860724,-0.2835580707,-0.1468141675,0.2001098096,0.296508044,-0.1102533489,0.0119160051,0.0752441809,0.1703256071,0.2682084739,-0.4217996299,-0.0571809337,-0.0282006711,-0.0079170447,0.3321003318,-0.5096785426,0.3010706902,-0.0119156707,-0.0405236855,0.0724276006,0.0591028631,0.1208358407,-0.1341485828,-0.1750265062,-0.0552734546,0.3399146795,0.1842273921,0.4693416059,0.0572355352,-0.0902262926,-0.0016380326,0.0171153843,-0.0476030819,0.2080054134,0.0927868336,-0.2515598238,0.2461876124,0.1334871501,0.1896291226,0.0995224938,0.3576408923,-0.1562275589,0.1550513953,-0.2551834285,-0.0418113098,-0.0018165474,0.1475818008,-0.065992035,0.2184211314,0.1675091684,0.0092136012,-0.6620985866,-0.1729354411,-0.0809698552,0.0756950304,0.1448045373,0.0778824016,0.3050026,-0.2590825558,-0.2244625688,0.2188713998,-0.1494889408,-0.0388926938,0.1985915154,-0.0519460551,-0.2465087324,0.2526426613,-0.1475192457,0.5964338779,0.0144994659,-0.0690575987,-0.0590026826,-0.0650994778,-0.0336832553,0.2011669874,0.0682175905,0.0198121574,0.5593736768,0.0142233958,0.4105928242,-0.4432581365,-0.6206678152,0.2140417397,-0.2587843537,0.1348548084,0.172923699,-0.199988246,0.1719183475,-0.2285513282,0.2683303058,-0.1405115277,-0.2521427572,-0.1096378937,-0.1593298167,-0.2201211601,-0.2864072919,-0.2941485643,0.0575457215,-0.5107776523,0.3836238086,-0.0515894815,-0.0735889971,0.3048487604,0.1345030069,0.1134327874,-0.1707373261,0.037112873,0.0071838633,0.0759944394,0.2964988649,-0.1817099899,-0.3410192132,-0.4029166698,-0.3328656852,0.2227272093,0.2894533277,-0.5181148648,-0.2595885992,-0.4046509564,-0.0979363993,0.1363006979,0.0584054142,0.3081404567,0.0041737356,-0.1337128133,-0.0365733542,-0.0952399895,0.1437298059,-0.0825313926,0.2196144313,-0.234057039,0.1243824512,0.0220159329,0.070076853,-0.0510838144,-0.2446267307,0.4307413697,0.0699668825,0.5303213,-0.0776209235,-0.4336575568,0.0322930254,0.0079225581,0.0880720466,0.3208144009,0.0801837221,-0.1344010681,-0.1273029,-0.2054393142,-0.06468907,-0.267870307,0.2333110571,0.5966069698,0.3693214059,-0.1487521976,0.0286411848,-0.1904284656,-0.0883540288,0.2325053662,0.2859807014,0.0400497988,-0.1458453238,-0.6221173406,0.2632019222,-0.4855470061,0.1779460907,0.0886694044,-0.2147367299,-0.134166196,-0.2471738011,-0.0698918849,-0.1596143544,0.221302107,0.0605671778,-0.147787109,0.2133919299,-0.4921296537,-0.0993314385,0.3009005785,-0.2773652971,-0.3393535018,0.3262186646,0.5858231187,-0.4331932664,0.0577573963,0.2916110754,-0.1205636188,-0.1171429902,-0.1828918308,-0.2271765471,-0.1606378257,-0.0934675485,0.2259438932,-0.1581119895,0.1385804564,0.1457807869,0.0722235218,0.0204905067,-0.4377142489,0.0325986333,0.2167355865,0.1873887926,-0.0257801842,0.1537255645,0.3326520622,0.2842483521,0.2106839418,0.420516938,-0.1526357234,-0.4608754516,0.1158064678,-0.1341502964,0.1449112892,0.3655847609,0.1420652568,-0.0859637409,0.100782752,0.2388846427,-0.4095500112,0.2243489474,0.0982198641,0.1174607053,-0.1121091098,0.1422736198,0.2986398041,0.051308874,-0.0369525105,-0.1962186247,0.277323693,-0.2588630617,0.2504976094,0.2197674662,0.9291275144,-0.2442261875,0.3038822711,0.2210435718,0.0163758174,0.6039122343,-0.1222566292,0.2333583832,-0.2020043135,-0.3721788526,0.0813563317,0.0764164478,0.2022609562,-0.132720679,-0.1490048468,-0.1550290138,0.3045494258,0.0176026523,-0.1982481033,0.3044421375,-0.209123835,-0.4136846662,-0.2782652676,0.1819972098,0.1473396271,0.0696430728,-0.0468595475,-0.0423125476,-0.3854658306,-0.2039593905,-0.2295353264,-0.120658271,-0.3041383326,0.1166681573,0.0076938146,-0.2416686565,-0.0530752353,0.4568975866,0.0758012757,0.1170797944,-0.3447546065,0.2510144114,-0.2892121971,-0.217111811,0.0535829552,-0.0996107981,0.2616426647,-0.0096044606,0.0916918814,0.0713680163,0.2710697353,-0.1062991172,-0.2131081223,-0.0851838887,0.2132451981,-0.2778496444,0.2475293428,0.0482180081,-0.2118631303,-0.1774332523,0.1711888015,-0.0224345215,0.0955686346,-0.0951324329,0.171209529,-0.1721329242,-0.054595843,0.0505963117,0.1018185616,-0.1979719847,0.0347161442,0.1395695657,-0.1838904619,-0.3093277216,-0.0297083203,0.5112108588,-0.5155857801,0.0084694335,0.0345307402,-0.4674131274,0.060089048,0.573759973,0.4746878445,-0.3675758839,-0.0770998374,0.1228512749,-0.4582303464,0.5098072886,0.0080544902,0.4807537198,-0.0955823362,0.2961559594,0.0063684578,0.3245731294,-0.4196091294,-0.0990698263,-0.2043651789,0.0493960716,0.2852256298,-0.2626973093,0.2570278049,0.1101772413,0.1839331537,0.269246161,-0.288012445,-0.1954979897,-0.3289353848,0.0788490325,0.0755601972,-0.1724989563,0.1805173159,0.0915966108,-0.1308641136,-0.2461307496,0.1180942953,-0.104410708,0.0498757735,-0.2543011308,0.0481252186,-0.2161459476,0.1305120438,-0.0221583266,-0.0749795958,-0.0345221907,0.2869910896,-0.0438179076,-0.0231433157,-0.0311532523,0.0643880367,0.2193820775,0.4704363346,-0.1334253699,0.2724251151,0.2083205432,0.1099073589,0.3799269795,0.5595136881,-0.1656282991,-0.0849351361,0.1247793809,0.0891096741,0.3096292615,-0.4454111755,-0.1792194545,0.0806735232,-0.1995284408,0.3205584288,0.3574013114,0.0998850465,0.2633017302,-0.0891732872,-0.0912660584,0.4639575779,-0.1013369858,-0.1773624569,0.0547057204,0.0200309642,0.213889733,0.4926331937,0.1897459775,0.3135468662,0.120442532,-0.0394419841,0.0124295466,-0.3077331185,0.3297132254,0.3257528841,-0.2407817692,-0.0443943217,0.0750690177,0.2137322277,0.3525604606,-0.3358432949,0.1773631871,0.1907114238,0.0261046737,-0.3274530172,0.4097513258,-0.271535337,-0.1176285222,0.1428260356,-0.2613988817,-0.2194768935,0.2469081879,0.0694395006,-0.1821265966,0.3511440158,-0.2594938874,-0.0093243886,-0.155198589,-0.1002508327,0.0155708706,0.1759782434,-0.3552975059,0.250125736,0.1912691295,0.1334860921,0.1725219488,0.4948211312,0.0996941626,0.1043620035,0.2325454205,0.1452892274,-0.1586695164,0.0557166822,-0.3469654322,0.1568077058,0.2664892673,0.2995811105,0.3310653865,0.1824028939,-0.2453657985,-0.0723698065,-0.2397516668,-0.0670535266,-0.1526950747,-0.0382589512,-0.1892791241,-0.2117007226,-0.3813470304,0.003048708,-0.4070925117,0.1521706581,-0.2005288005,-0.0437946022,0.1016824618,-0.0986234918,0.0912683755,-0.0629428402,0.3323911726,0.4239867628,0.1304457784,-0.2630107701,-0.2959092855,-0.514713943,0.0946790874,-0.0405933522,-0.1162134185,-0.1372529864,0.1663169563,0.2671421468,0.0925728455,0.0215102918,-0.0796077773,-0.0498314351,-0.0858363658,-0.3778330982,0.1413139999,0.1686926335,-0.0299571138,0.245986104,-0.2205045521,0.0649857223,0.2508813441,0.1058385819,-0.1179234907,0.2528412342,-0.2770433128,0.1129975617,0.1304099858,0.0542902946,0.3391413093,-0.304271549,-0.3976641595,-0.3352033496,-0.1973171383,-0.2953970134,0.035466671,0.3923990428,0.6013519168,-0.323233217,-0.3374992907,0.0457635671,0.2210098207,0.2798245549,-0.0524528809,-0.2831208706,0.2350938022,-0.1711228639,-0.2178131044,0.1039375737,0.4913273752,0.0735716596,-0.0873401761,-0.131597206,-0.4974164069,0.3540226519,-0.2215640247,-0.018055629,0.1933280826,-0.0674820468,0.6120026112,0.0843113512,-0.5384672284,0.0197059587,0.0753989592,0.0960793644,-0.1180932075,0.3622705042,0.0808750018,-0.3450689614,-0.0752397329,0.5512399673,-0.024891803,-0.2480751276,0.0404570624,-0.2355949879]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2761","title":"Error loading C4 realnewslike dataset","comments":"Hi @danshirron, \r\n`c4` was updated few days back by @lhoestq. The new configs are `['en', 'en.noclean', 'en.realnewslike', 'en.webtextlike'].` You'll need to remove any older version of this dataset you previously downloaded and then run `load_dataset` again with new configuration.","body":"## Describe the bug\r\nError loading C4 realnewslike dataset. Validation part mismatch\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir)\r\n## Expected results\r\nsuccess on data loading\r\n## Actual results\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 15.3M\/15.3M [00:00<00:00, 28.1MB\/s]Traceback (most recent call last): \r\n File \"run_mlm_tf.py\", line 794, in \r\n main() \r\n File \"run_mlm_tf.py\", line 425, in main \r\n raw_datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 843, in load_dataset \r\n builder_instance.download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 608, in download_and_prepare \r\n self._download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 698, in _download_and_prepare verify_splits(self.info.splits, split_dict) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits \r\n raise NonMatchingSplitsSizesError(str(bad_splits)) \r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=38165657946, num_examples=13799838, dataset_name='c4'), 'recorded': SplitInfo(name='validation', num_bytes=37875873, num_examples=13863, dataset_name='c4')}] \r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1","comment_length":39,"text":"Error loading C4 realnewslike dataset\n## Describe the bug\r\nError loading C4 realnewslike dataset. Validation part mismatch\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir)\r\n## Expected results\r\nsuccess on data loading\r\n## Actual results\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 15.3M\/15.3M [00:00<00:00, 28.1MB\/s]Traceback (most recent call last): \r\n File \"run_mlm_tf.py\", line 794, in \r\n main() \r\n File \"run_mlm_tf.py\", line 425, in main \r\n raw_datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 843, in load_dataset \r\n builder_instance.download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 608, in download_and_prepare \r\n self._download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 698, in _download_and_prepare verify_splits(self.info.splits, split_dict) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits \r\n raise NonMatchingSplitsSizesError(str(bad_splits)) \r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=38165657946, num_examples=13799838, dataset_name='c4'), 'recorded': SplitInfo(name='validation', num_bytes=37875873, num_examples=13863, dataset_name='c4')}] \r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\nHi @danshirron, \r\n`c4` was updated few days back by @lhoestq. The new configs are `['en', 'en.noclean', 'en.realnewslike', 'en.webtextlike'].` You'll need to remove any older version of this dataset you previously downloaded and then run `load_dataset` again with new configuration.","embeddings":[-0.2315260619,0.0246403441,-0.0223660674,0.4574569166,0.1744244695,0.2295811772,-0.0073711821,0.5081834197,-0.0316528082,0.0191251654,-0.1842969209,-0.0640127659,-0.0091638044,0.3338081241,-0.043365553,-0.2194700241,-0.1107666269,0.1606068611,0.1407687962,0.0597507246,-0.1448137611,0.2009049803,-0.2154413909,-0.0725571662,-0.2707114816,-0.0130072096,0.1758372933,0.3205577433,0.0850609317,-0.5288669467,0.2025581449,-0.1739532501,0.0258219484,0.5030124187,-0.0001112284,0.0486640595,0.4246914983,-0.0443259999,-0.3031130135,0.0802712366,-0.309867084,-0.242859751,0.144365117,-0.0320097134,-0.3313594162,-0.1901502609,-0.2336005718,-0.337558955,0.2493240535,0.6167075038,0.238580361,0.3121445775,0.0608730987,0.0165075231,-0.1109841093,-0.1475731134,-0.022677483,0.3512011766,-0.0215199199,0.031706702,0.1784890443,0.2144178748,0.028754618,0.1094985083,0.1285811663,-0.1285620183,0.2033363879,-0.1063502431,0.1449157298,0.3846841753,0.3563730419,-0.2420736998,-0.1519511491,-0.3080828488,-0.1986611784,-0.2773949504,0.256582588,0.1999745071,-0.162488386,0.1661835164,-0.3945821524,0.2987675369,0.0589543805,0.0859668255,-0.1932324171,0.2412162274,0.151707381,-0.0265462417,0.0266118441,0.0962596685,0.1632707119,-0.1978420466,-0.0648603812,0.2136579305,-0.6099298596,-0.2218074054,-0.3099172413,0.4977063537,-0.0144231617,-0.051483281,-0.1140118986,-0.0171039887,0.1712173671,-0.186304301,0.1688324213,0.2142098099,-0.198009029,0.1207100525,0.1434043944,-0.1067880541,0.0161335897,-0.0791779235,0.162792176,-0.1318754554,-0.0637650043,0.2420955002,0.0574146733,-0.317915082,-0.6927220821,0.2458004653,-0.3038004041,-0.036881227,0.0007450588,0.3466166258,-0.3842547238,0.2641151547,0.1001107544,-0.1724211723,-0.2799533308,-0.2688709497,-0.2056975961,0.0522305109,-0.3685680926,0.0012723056,0.2524729669,-0.2094730288,0.4003704488,-0.0511978604,0.1751805246,-0.0478972159,0.209127903,-0.3762721717,-0.1081526875,0.2255997509,-0.3169129789,0.037633013,0.2689840496,-0.0927065387,0.0103949495,-0.0342599787,0.1593452543,-0.3435829878,-0.1637990624,0.287226826,-0.0145503711,0.0288387816,0.0684598163,-0.0540249497,0.3395628631,-0.1157185137,-0.1957776248,-0.2394090295,0.0330544971,-0.058229614,0.1889537871,0.4896891117,-0.4754011929,0.0806836635,-0.3405704498,0.1221416891,0.4467144907,0.2935122252,-0.0330286399,0.276563555,-0.2077404857,0.0237064268,-0.0174367968,-0.3686727285,-0.3444623053,0.5175626874,-0.0822370797,0.345344156,-0.0547711328,0.043166969,0.1374017298,0.0436606482,0.1547780186,0.2047380656,-0.1418510228,0.1543143243,-0.562957108,-0.1220002472,0.3635740578,0.0133565171,0.1890151352,-0.2069269419,0.1235904247,-0.0846622884,0.4912214279,-0.0084104314,-0.1206199005,-0.0321739204,0.2111113369,-0.0240932405,0.1006423309,-0.2743466496,-0.5518809557,0.2668113708,-0.0237271488,0.152853027,0.1286136508,-0.046923276,-0.589758873,-0.1467690319,-0.1562295854,-0.2404751182,0.1061195433,0.1226577386,0.2428137511,0.0796359479,-0.0546249487,0.5485526323,-0.0577965789,0.1413292736,-0.0799551159,0.341332972,-0.1525628865,-0.0305092093,-0.0302448831,0.1346586496,0.2463358045,-0.09839461,-0.3567022681,0.6420468092,0.1248638704,0.2066819072,-0.4426276088,0.0053086914,0.3827438354,-0.3125794828,0.0957020447,0.5969313979,0.2378956974,0.0181498863,0.0953133032,0.424510628,-0.0290682651,-0.0525034592,0.1060230434,-0.1878120452,0.1633400321,-0.1469224691,0.200978145,-0.3520150185,0.4401290417,-0.086315088,0.3760469556,-0.0243641734,-0.1455283016,0.0485069752,0.4009242058,-0.0737881437,-0.0185004827,-0.1634851992,-0.1819177121,-0.1184910983,0.3020989597,0.3241461515,0.6374415755,0.1745533198,-0.2192538083,0.2381288111,-0.000430839,-0.2599272132,0.4831356406,-0.1653624773,0.0127727846,0.4583308399,0.008977511,0.1091809645,-0.4487721622,0.1450789571,0.123402968,0.1818011999,-0.360792309,0.0788164586,-0.215221718,0.0968676656,-0.3518635631,-0.2619572878,0.029402364,-0.313379705,-0.0178715326,0.1368231773,-0.0503815673,0.2308113575,-0.4672618806,0.0846562535,0.1124338806,-0.123098582,-0.2004061192,0.0737325698,-0.3612599671,0.0988924429,0.253513217,0.0678496435,0.2908097208,-0.1956795603,-0.1511840522,0.0241016988,-0.4036478996,0.1278229356,0.1180695146,0.2275419384,0.0969556868,0.0744806901,0.3463863432,-0.4121953249,-0.071068041,-0.1308414191,-0.2183184475,0.17054151,0.0332118087,0.0247350875,-0.1156041622,-0.5987548232,-0.4374191463,-0.3892827332,-0.0498279557,-0.1645166576,-0.0728507191,0.1988957822,-0.1343782842,0.0514261425,0.1836190671,-0.0720977336,-0.4009529352,0.2122821659,0.1445584893,-0.0392751917,-0.2241797745,-0.1384652853,-0.0772775933,-0.0527735427,-0.0458171144,-0.3023410439,0.254326731,-0.4559999704,0.2435563952,-0.1985964179,-0.213123858,0.2872216702,0.0318476632,-0.0354612023,-0.1347290725,-0.0344382711,-0.0722915977,-0.1071643755,0.1636640579,-0.1976882517,0.394243598,0.1363981962,0.4813589752,0.2416753024,-0.5439124703,0.0680660531,0.1705723852,0.2048651129,-0.0990280136,-0.3353740275,0.3906064034,-0.0682675168,0.0239580069,0.2752487063,-0.3505270779,-0.2781621218,-0.0356105082,0.1964913607,-0.3761669397,-0.0256565753,-0.1338630915,-0.1636420041,0.2058549374,0.3254787922,0.2892023027,0.0354885459,-0.1637172848,0.2503463328,0.2493042946,0.0125087956,-0.1440640688,-0.1290749609,-0.1602590382,-0.2274690866,0.3832500875,0.3412743509,0.2640645802,0.0742882937,-0.1175951138,-0.0429789424,-0.0687436461,0.6774853468,-0.1602197737,0.1394208074,-0.0770248547,0.1332771629,-0.1524718404,-0.2240352929,-0.0229289755,0.0237797741,0.294780612,0.5989146829,-0.3209036589,-0.1446436495,0.3440994918,0.3235155046,-0.1614699215,-0.3134260774,-0.0346945636,-0.0641082749,-0.0452169776,0.1154499054,0.0462888815,0.2783490717,-0.1482951194,-0.153140083,0.0351074636,-0.0507183559,0.2441392988,0.3141935766,0.3286376297,-0.0386056788,0.3620902002,0.1574015468,0.302365303,0.372184962,0.2111066729,-0.345465064,-0.058314722,-0.1266359091,0.0932570174,-0.001311927,0.2818886042,-0.0490357913,-0.1346836835,0.2506635785,0.068419002,-0.152463302,0.3452251852,-0.0177854709,0.0991846919,-0.1798811108,-0.3432311416,0.2316904664,-0.1548494846,0.0145705231,0.1157020479,0.067900978,-0.3640975356,-0.0880728513,0.0034845076,0.7999356985,0.3829322457,-0.1509274393,0.2441919446,-0.3533546329,0.0227454789,-0.3782048523,0.1704315543,-0.3321321905,-0.1670113057,-0.0512643084,-0.1950098127,-0.0709553435,0.0745388493,-0.2098713219,0.2186295688,-0.1244219914,0.0774676353,0.2048077136,0.2370273769,0.0730773434,0.1332217753,-0.1807198673,0.1464785039,-0.0612236559,0.1584163606,-0.2090026289,-0.0780172944,0.0462242775,-0.1878198981,-0.2343229651,0.2004466653,-0.671544075,0.0274442788,0.0306347199,-0.0971903205,-0.0029017092,-0.0963324159,0.239186883,0.2715616524,-0.1229086146,0.0358495861,0.1376942098,0.153495729,0.0529686101,-0.0387600996,0.2003014684,-0.135901317,-0.1790282279,0.0824056566,-0.3000933826,-0.1808879673,0.2318426222,-0.0816883668,-0.0288241655,-0.3130398691,-0.1693854779,-0.0866780058,0.1064577252,-0.2034928203,0.1535087526,-0.2164582908,-0.3282732069,0.0480187908,0.0514248461,-0.3208444118,-0.3036953807,0.2600752115,0.1588468254,-0.0262651835,0.3788428903,0.0115331747,-0.3269037902,-0.2525985241,-0.1472427696,0.1546261013,-0.4197560251,0.0971366242,-0.3318922222,-0.2562456727,0.1576458812,0.2788511515,0.5067481399,0.0376338139,0.0551861934,-0.5445108414,-0.105076164,0.0463757105,-0.167886883,0.2210354954,-0.0551569201,0.6132634878,-0.3147180974,0.1814591736,-0.3386438787,0.0392292291,-0.1233162209,0.1893967837,0.000827592,-0.2967326045,0.2796502113,0.098179169,0.2074516565,0.3625247777,-0.1502892226,-0.1964601725,-0.1700348705,0.1034775451,0.0979663134,-0.2308011055,0.062586233,0.158974424,-0.0434559509,-0.008686211,0.1131301001,-0.1068085581,-0.0433675498,0.1655225754,0.3677061796,-0.0724938735,-0.2427455038,0.0912591964,-0.1443977952,0.0855922028,0.0484122522,0.1299035996,-0.148056075,0.0475814752,0.158301577,-0.0559338145,0.2772265077,0.2009176612,0.2070959061,-0.2499206066,-0.2945112884,0.2605382502,0.1768852025,0.4284108877,0.0114917271,-0.158310473,0.0698025599,0.2637248337,-0.3338602185,-0.0800075009,-0.0841673911,-0.2150553316,0.1919243038,0.0787677839,0.0997455642,-0.1859395355,-0.0469003171,0.2622558475,0.4332890213,-0.0860139802,0.109123379,0.398732245,-0.0716096759,0.0295816381,0.1499782354,-0.2502602935,0.0625841767,0.3447422087,-0.04796784,0.1358725876,0.0101845749,0.1770618111,-0.0911619663,-0.1617771238,0.2650981247,0.4197490513,-0.4040268958,0.2929238975,0.2387186885,-0.1023296788,-0.3085986972,-0.1845877022,-0.4084452093,-0.1939706504,-0.0719643459,0.1064393893,0.0173226986,-0.342833668,-0.2268976718,-0.0113695161,-0.0335704125,0.0478483923,0.5259116292,0.0245203599,-0.2330120802,-0.1281084269,-0.2837902904,0.2017405778,0.21901384,-0.3413544297,0.3417457938,0.2017515898,-0.1879087836,-0.2357747108,0.190875873,0.1929077357,0.4227327704,-0.2607415318,-0.1073196456,0.144697234,0.0094177518,0.123514995,0.4184183478,0.0590738058,-0.0722034201,0.226265043,0.1355629712,-0.2458421141,0.1419692785,-0.0804052204,-0.1136842966,-0.6303744316,0.3398411274,-0.0270323902,-0.0053561358,-0.106356062,0.2040739954,-0.3419698477,0.0560279228,0.3937089741,-0.1050297618,0.2485336363,0.1354949623,0.0990920588,-0.1431972235,0.3879002631,0.1573156416,0.0696986988,-0.5092074871,-0.1368087679,-0.4429211318,0.231569469,-0.1639965326,0.1835734993,-0.0463663936,0.0475714915,0.1086042672,0.18673639,0.1462301016,0.3793209493,0.1809259057,-0.0441591442,-0.2704292834,-0.1495058537,-0.1365631223,-0.2033466548,0.0330277979,-0.5399720669,0.0781174973,0.1010801494,0.1184633076,-0.0927306265,-0.2510053813,0.3491327763,0.2162149996,0.3482257128,0.3603941202,0.5763022304,-0.1550000161,-0.1805732548,0.002104782,-0.046220392,-0.1851202548,0.1420908719,0.1194762066,0.3269408047,0.0800936669,-0.1135649532,-0.2312363982,0.4827974141,0.3263706267,-0.2479363978,-0.255174309,-0.0562407076,-0.3647657037,-0.0320115685,0.2655400634,0.397505343,-0.0032958677,0.052043993,0.0077277757,-0.3785646856,0.4797550738,-0.5694298744,-0.0697973073,-0.1935780793,0.2503972948,-0.0434154011,-0.0811097994,-0.4199209809,0.0652301982,0.5131907463,-0.0961262435,-0.0506019965,0.1511285007,0.1545130163,-0.1854586005,0.0249983277,0.3290072381,0.3633755445,-0.2964656353,0.058436729,-0.2315704226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2761","title":"Error loading C4 realnewslike dataset","comments":"@bhavitvyamalik @lhoestq , just tried the above and got:\r\n>>> a=datasets.load_dataset('c4','en.realnewslike')\r\nDownloading: 3.29kB [00:00, 1.66MB\/s] \r\nDownloading: 2.40MB [00:00, 12.6MB\/s] \r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 819, in load_dataset\r\n builder_instance = load_dataset_builder(\r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 701, in load_dataset_builder\r\n builder_instance: DatasetBuilder = builder_cls(\r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1049, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 268, in __init__\r\n self.config, self.config_id = self._create_builder_config(\r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 360, in _create_builder_config\r\n raise ValueError(\r\nValueError: BuilderConfig en.realnewslike not found. Available: ['en', 'realnewslike', 'en.noblocklist', 'en.noclean']\r\n>>> \r\n\r\ndatasets version is 1.11.0\r\n","body":"## Describe the bug\r\nError loading C4 realnewslike dataset. Validation part mismatch\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir)\r\n## Expected results\r\nsuccess on data loading\r\n## Actual results\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 15.3M\/15.3M [00:00<00:00, 28.1MB\/s]Traceback (most recent call last): \r\n File \"run_mlm_tf.py\", line 794, in \r\n main() \r\n File \"run_mlm_tf.py\", line 425, in main \r\n raw_datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 843, in load_dataset \r\n builder_instance.download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 608, in download_and_prepare \r\n self._download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 698, in _download_and_prepare verify_splits(self.info.splits, split_dict) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits \r\n raise NonMatchingSplitsSizesError(str(bad_splits)) \r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=38165657946, num_examples=13799838, dataset_name='c4'), 'recorded': SplitInfo(name='validation', num_bytes=37875873, num_examples=13863, dataset_name='c4')}] \r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1","comment_length":91,"text":"Error loading C4 realnewslike dataset\n## Describe the bug\r\nError loading C4 realnewslike dataset. Validation part mismatch\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir)\r\n## Expected results\r\nsuccess on data loading\r\n## Actual results\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 15.3M\/15.3M [00:00<00:00, 28.1MB\/s]Traceback (most recent call last): \r\n File \"run_mlm_tf.py\", line 794, in \r\n main() \r\n File \"run_mlm_tf.py\", line 425, in main \r\n raw_datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 843, in load_dataset \r\n builder_instance.download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 608, in download_and_prepare \r\n self._download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 698, in _download_and_prepare verify_splits(self.info.splits, split_dict) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits \r\n raise NonMatchingSplitsSizesError(str(bad_splits)) \r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=38165657946, num_examples=13799838, dataset_name='c4'), 'recorded': SplitInfo(name='validation', num_bytes=37875873, num_examples=13863, dataset_name='c4')}] \r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\n@bhavitvyamalik @lhoestq , just tried the above and got:\r\n>>> a=datasets.load_dataset('c4','en.realnewslike')\r\nDownloading: 3.29kB [00:00, 1.66MB\/s] \r\nDownloading: 2.40MB [00:00, 12.6MB\/s] \r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 819, in load_dataset\r\n builder_instance = load_dataset_builder(\r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 701, in load_dataset_builder\r\n builder_instance: DatasetBuilder = builder_cls(\r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1049, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 268, in __init__\r\n self.config, self.config_id = self._create_builder_config(\r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 360, in _create_builder_config\r\n raise ValueError(\r\nValueError: BuilderConfig en.realnewslike not found. Available: ['en', 'realnewslike', 'en.noblocklist', 'en.noclean']\r\n>>> \r\n\r\ndatasets version is 1.11.0\r\n","embeddings":[-0.2315260619,0.0246403441,-0.0223660674,0.4574569166,0.1744244695,0.2295811772,-0.0073711821,0.5081834197,-0.0316528082,0.0191251654,-0.1842969209,-0.0640127659,-0.0091638044,0.3338081241,-0.043365553,-0.2194700241,-0.1107666269,0.1606068611,0.1407687962,0.0597507246,-0.1448137611,0.2009049803,-0.2154413909,-0.0725571662,-0.2707114816,-0.0130072096,0.1758372933,0.3205577433,0.0850609317,-0.5288669467,0.2025581449,-0.1739532501,0.0258219484,0.5030124187,-0.0001112284,0.0486640595,0.4246914983,-0.0443259999,-0.3031130135,0.0802712366,-0.309867084,-0.242859751,0.144365117,-0.0320097134,-0.3313594162,-0.1901502609,-0.2336005718,-0.337558955,0.2493240535,0.6167075038,0.238580361,0.3121445775,0.0608730987,0.0165075231,-0.1109841093,-0.1475731134,-0.022677483,0.3512011766,-0.0215199199,0.031706702,0.1784890443,0.2144178748,0.028754618,0.1094985083,0.1285811663,-0.1285620183,0.2033363879,-0.1063502431,0.1449157298,0.3846841753,0.3563730419,-0.2420736998,-0.1519511491,-0.3080828488,-0.1986611784,-0.2773949504,0.256582588,0.1999745071,-0.162488386,0.1661835164,-0.3945821524,0.2987675369,0.0589543805,0.0859668255,-0.1932324171,0.2412162274,0.151707381,-0.0265462417,0.0266118441,0.0962596685,0.1632707119,-0.1978420466,-0.0648603812,0.2136579305,-0.6099298596,-0.2218074054,-0.3099172413,0.4977063537,-0.0144231617,-0.051483281,-0.1140118986,-0.0171039887,0.1712173671,-0.186304301,0.1688324213,0.2142098099,-0.198009029,0.1207100525,0.1434043944,-0.1067880541,0.0161335897,-0.0791779235,0.162792176,-0.1318754554,-0.0637650043,0.2420955002,0.0574146733,-0.317915082,-0.6927220821,0.2458004653,-0.3038004041,-0.036881227,0.0007450588,0.3466166258,-0.3842547238,0.2641151547,0.1001107544,-0.1724211723,-0.2799533308,-0.2688709497,-0.2056975961,0.0522305109,-0.3685680926,0.0012723056,0.2524729669,-0.2094730288,0.4003704488,-0.0511978604,0.1751805246,-0.0478972159,0.209127903,-0.3762721717,-0.1081526875,0.2255997509,-0.3169129789,0.037633013,0.2689840496,-0.0927065387,0.0103949495,-0.0342599787,0.1593452543,-0.3435829878,-0.1637990624,0.287226826,-0.0145503711,0.0288387816,0.0684598163,-0.0540249497,0.3395628631,-0.1157185137,-0.1957776248,-0.2394090295,0.0330544971,-0.058229614,0.1889537871,0.4896891117,-0.4754011929,0.0806836635,-0.3405704498,0.1221416891,0.4467144907,0.2935122252,-0.0330286399,0.276563555,-0.2077404857,0.0237064268,-0.0174367968,-0.3686727285,-0.3444623053,0.5175626874,-0.0822370797,0.345344156,-0.0547711328,0.043166969,0.1374017298,0.0436606482,0.1547780186,0.2047380656,-0.1418510228,0.1543143243,-0.562957108,-0.1220002472,0.3635740578,0.0133565171,0.1890151352,-0.2069269419,0.1235904247,-0.0846622884,0.4912214279,-0.0084104314,-0.1206199005,-0.0321739204,0.2111113369,-0.0240932405,0.1006423309,-0.2743466496,-0.5518809557,0.2668113708,-0.0237271488,0.152853027,0.1286136508,-0.046923276,-0.589758873,-0.1467690319,-0.1562295854,-0.2404751182,0.1061195433,0.1226577386,0.2428137511,0.0796359479,-0.0546249487,0.5485526323,-0.0577965789,0.1413292736,-0.0799551159,0.341332972,-0.1525628865,-0.0305092093,-0.0302448831,0.1346586496,0.2463358045,-0.09839461,-0.3567022681,0.6420468092,0.1248638704,0.2066819072,-0.4426276088,0.0053086914,0.3827438354,-0.3125794828,0.0957020447,0.5969313979,0.2378956974,0.0181498863,0.0953133032,0.424510628,-0.0290682651,-0.0525034592,0.1060230434,-0.1878120452,0.1633400321,-0.1469224691,0.200978145,-0.3520150185,0.4401290417,-0.086315088,0.3760469556,-0.0243641734,-0.1455283016,0.0485069752,0.4009242058,-0.0737881437,-0.0185004827,-0.1634851992,-0.1819177121,-0.1184910983,0.3020989597,0.3241461515,0.6374415755,0.1745533198,-0.2192538083,0.2381288111,-0.000430839,-0.2599272132,0.4831356406,-0.1653624773,0.0127727846,0.4583308399,0.008977511,0.1091809645,-0.4487721622,0.1450789571,0.123402968,0.1818011999,-0.360792309,0.0788164586,-0.215221718,0.0968676656,-0.3518635631,-0.2619572878,0.029402364,-0.313379705,-0.0178715326,0.1368231773,-0.0503815673,0.2308113575,-0.4672618806,0.0846562535,0.1124338806,-0.123098582,-0.2004061192,0.0737325698,-0.3612599671,0.0988924429,0.253513217,0.0678496435,0.2908097208,-0.1956795603,-0.1511840522,0.0241016988,-0.4036478996,0.1278229356,0.1180695146,0.2275419384,0.0969556868,0.0744806901,0.3463863432,-0.4121953249,-0.071068041,-0.1308414191,-0.2183184475,0.17054151,0.0332118087,0.0247350875,-0.1156041622,-0.5987548232,-0.4374191463,-0.3892827332,-0.0498279557,-0.1645166576,-0.0728507191,0.1988957822,-0.1343782842,0.0514261425,0.1836190671,-0.0720977336,-0.4009529352,0.2122821659,0.1445584893,-0.0392751917,-0.2241797745,-0.1384652853,-0.0772775933,-0.0527735427,-0.0458171144,-0.3023410439,0.254326731,-0.4559999704,0.2435563952,-0.1985964179,-0.213123858,0.2872216702,0.0318476632,-0.0354612023,-0.1347290725,-0.0344382711,-0.0722915977,-0.1071643755,0.1636640579,-0.1976882517,0.394243598,0.1363981962,0.4813589752,0.2416753024,-0.5439124703,0.0680660531,0.1705723852,0.2048651129,-0.0990280136,-0.3353740275,0.3906064034,-0.0682675168,0.0239580069,0.2752487063,-0.3505270779,-0.2781621218,-0.0356105082,0.1964913607,-0.3761669397,-0.0256565753,-0.1338630915,-0.1636420041,0.2058549374,0.3254787922,0.2892023027,0.0354885459,-0.1637172848,0.2503463328,0.2493042946,0.0125087956,-0.1440640688,-0.1290749609,-0.1602590382,-0.2274690866,0.3832500875,0.3412743509,0.2640645802,0.0742882937,-0.1175951138,-0.0429789424,-0.0687436461,0.6774853468,-0.1602197737,0.1394208074,-0.0770248547,0.1332771629,-0.1524718404,-0.2240352929,-0.0229289755,0.0237797741,0.294780612,0.5989146829,-0.3209036589,-0.1446436495,0.3440994918,0.3235155046,-0.1614699215,-0.3134260774,-0.0346945636,-0.0641082749,-0.0452169776,0.1154499054,0.0462888815,0.2783490717,-0.1482951194,-0.153140083,0.0351074636,-0.0507183559,0.2441392988,0.3141935766,0.3286376297,-0.0386056788,0.3620902002,0.1574015468,0.302365303,0.372184962,0.2111066729,-0.345465064,-0.058314722,-0.1266359091,0.0932570174,-0.001311927,0.2818886042,-0.0490357913,-0.1346836835,0.2506635785,0.068419002,-0.152463302,0.3452251852,-0.0177854709,0.0991846919,-0.1798811108,-0.3432311416,0.2316904664,-0.1548494846,0.0145705231,0.1157020479,0.067900978,-0.3640975356,-0.0880728513,0.0034845076,0.7999356985,0.3829322457,-0.1509274393,0.2441919446,-0.3533546329,0.0227454789,-0.3782048523,0.1704315543,-0.3321321905,-0.1670113057,-0.0512643084,-0.1950098127,-0.0709553435,0.0745388493,-0.2098713219,0.2186295688,-0.1244219914,0.0774676353,0.2048077136,0.2370273769,0.0730773434,0.1332217753,-0.1807198673,0.1464785039,-0.0612236559,0.1584163606,-0.2090026289,-0.0780172944,0.0462242775,-0.1878198981,-0.2343229651,0.2004466653,-0.671544075,0.0274442788,0.0306347199,-0.0971903205,-0.0029017092,-0.0963324159,0.239186883,0.2715616524,-0.1229086146,0.0358495861,0.1376942098,0.153495729,0.0529686101,-0.0387600996,0.2003014684,-0.135901317,-0.1790282279,0.0824056566,-0.3000933826,-0.1808879673,0.2318426222,-0.0816883668,-0.0288241655,-0.3130398691,-0.1693854779,-0.0866780058,0.1064577252,-0.2034928203,0.1535087526,-0.2164582908,-0.3282732069,0.0480187908,0.0514248461,-0.3208444118,-0.3036953807,0.2600752115,0.1588468254,-0.0262651835,0.3788428903,0.0115331747,-0.3269037902,-0.2525985241,-0.1472427696,0.1546261013,-0.4197560251,0.0971366242,-0.3318922222,-0.2562456727,0.1576458812,0.2788511515,0.5067481399,0.0376338139,0.0551861934,-0.5445108414,-0.105076164,0.0463757105,-0.167886883,0.2210354954,-0.0551569201,0.6132634878,-0.3147180974,0.1814591736,-0.3386438787,0.0392292291,-0.1233162209,0.1893967837,0.000827592,-0.2967326045,0.2796502113,0.098179169,0.2074516565,0.3625247777,-0.1502892226,-0.1964601725,-0.1700348705,0.1034775451,0.0979663134,-0.2308011055,0.062586233,0.158974424,-0.0434559509,-0.008686211,0.1131301001,-0.1068085581,-0.0433675498,0.1655225754,0.3677061796,-0.0724938735,-0.2427455038,0.0912591964,-0.1443977952,0.0855922028,0.0484122522,0.1299035996,-0.148056075,0.0475814752,0.158301577,-0.0559338145,0.2772265077,0.2009176612,0.2070959061,-0.2499206066,-0.2945112884,0.2605382502,0.1768852025,0.4284108877,0.0114917271,-0.158310473,0.0698025599,0.2637248337,-0.3338602185,-0.0800075009,-0.0841673911,-0.2150553316,0.1919243038,0.0787677839,0.0997455642,-0.1859395355,-0.0469003171,0.2622558475,0.4332890213,-0.0860139802,0.109123379,0.398732245,-0.0716096759,0.0295816381,0.1499782354,-0.2502602935,0.0625841767,0.3447422087,-0.04796784,0.1358725876,0.0101845749,0.1770618111,-0.0911619663,-0.1617771238,0.2650981247,0.4197490513,-0.4040268958,0.2929238975,0.2387186885,-0.1023296788,-0.3085986972,-0.1845877022,-0.4084452093,-0.1939706504,-0.0719643459,0.1064393893,0.0173226986,-0.342833668,-0.2268976718,-0.0113695161,-0.0335704125,0.0478483923,0.5259116292,0.0245203599,-0.2330120802,-0.1281084269,-0.2837902904,0.2017405778,0.21901384,-0.3413544297,0.3417457938,0.2017515898,-0.1879087836,-0.2357747108,0.190875873,0.1929077357,0.4227327704,-0.2607415318,-0.1073196456,0.144697234,0.0094177518,0.123514995,0.4184183478,0.0590738058,-0.0722034201,0.226265043,0.1355629712,-0.2458421141,0.1419692785,-0.0804052204,-0.1136842966,-0.6303744316,0.3398411274,-0.0270323902,-0.0053561358,-0.106356062,0.2040739954,-0.3419698477,0.0560279228,0.3937089741,-0.1050297618,0.2485336363,0.1354949623,0.0990920588,-0.1431972235,0.3879002631,0.1573156416,0.0696986988,-0.5092074871,-0.1368087679,-0.4429211318,0.231569469,-0.1639965326,0.1835734993,-0.0463663936,0.0475714915,0.1086042672,0.18673639,0.1462301016,0.3793209493,0.1809259057,-0.0441591442,-0.2704292834,-0.1495058537,-0.1365631223,-0.2033466548,0.0330277979,-0.5399720669,0.0781174973,0.1010801494,0.1184633076,-0.0927306265,-0.2510053813,0.3491327763,0.2162149996,0.3482257128,0.3603941202,0.5763022304,-0.1550000161,-0.1805732548,0.002104782,-0.046220392,-0.1851202548,0.1420908719,0.1194762066,0.3269408047,0.0800936669,-0.1135649532,-0.2312363982,0.4827974141,0.3263706267,-0.2479363978,-0.255174309,-0.0562407076,-0.3647657037,-0.0320115685,0.2655400634,0.397505343,-0.0032958677,0.052043993,0.0077277757,-0.3785646856,0.4797550738,-0.5694298744,-0.0697973073,-0.1935780793,0.2503972948,-0.0434154011,-0.0811097994,-0.4199209809,0.0652301982,0.5131907463,-0.0961262435,-0.0506019965,0.1511285007,0.1545130163,-0.1854586005,0.0249983277,0.3290072381,0.3633755445,-0.2964656353,0.058436729,-0.2315704226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2761","title":"Error loading C4 realnewslike dataset","comments":"I think I had an older version of datasets installed and that's why I commented the old configurations in my last comment, my bad! I re-checked and updated it to latest version (`datasets==1.11.0`) and it's showing `available configs: ['en', 'realnewslike', 'en.noblocklist', 'en.noclean']`. \r\n\r\nI tried `raw_datasets = load_dataset('c4', 'realnewslike')` and the download started. Make sure you don't have any old copy of this dataset and you download it fresh using the latest version of datasets. Sorry for the mix up!","body":"## Describe the bug\r\nError loading C4 realnewslike dataset. Validation part mismatch\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir)\r\n## Expected results\r\nsuccess on data loading\r\n## Actual results\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 15.3M\/15.3M [00:00<00:00, 28.1MB\/s]Traceback (most recent call last): \r\n File \"run_mlm_tf.py\", line 794, in \r\n main() \r\n File \"run_mlm_tf.py\", line 425, in main \r\n raw_datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 843, in load_dataset \r\n builder_instance.download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 608, in download_and_prepare \r\n self._download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 698, in _download_and_prepare verify_splits(self.info.splits, split_dict) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits \r\n raise NonMatchingSplitsSizesError(str(bad_splits)) \r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=38165657946, num_examples=13799838, dataset_name='c4'), 'recorded': SplitInfo(name='validation', num_bytes=37875873, num_examples=13863, dataset_name='c4')}] \r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1","comment_length":79,"text":"Error loading C4 realnewslike dataset\n## Describe the bug\r\nError loading C4 realnewslike dataset. Validation part mismatch\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir)\r\n## Expected results\r\nsuccess on data loading\r\n## Actual results\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 15.3M\/15.3M [00:00<00:00, 28.1MB\/s]Traceback (most recent call last): \r\n File \"run_mlm_tf.py\", line 794, in \r\n main() \r\n File \"run_mlm_tf.py\", line 425, in main \r\n raw_datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 843, in load_dataset \r\n builder_instance.download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 608, in download_and_prepare \r\n self._download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 698, in _download_and_prepare verify_splits(self.info.splits, split_dict) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits \r\n raise NonMatchingSplitsSizesError(str(bad_splits)) \r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=38165657946, num_examples=13799838, dataset_name='c4'), 'recorded': SplitInfo(name='validation', num_bytes=37875873, num_examples=13863, dataset_name='c4')}] \r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\nI think I had an older version of datasets installed and that's why I commented the old configurations in my last comment, my bad! I re-checked and updated it to latest version (`datasets==1.11.0`) and it's showing `available configs: ['en', 'realnewslike', 'en.noblocklist', 'en.noclean']`. \r\n\r\nI tried `raw_datasets = load_dataset('c4', 'realnewslike')` and the download started. Make sure you don't have any old copy of this dataset and you download it fresh using the latest version of datasets. Sorry for the mix up!","embeddings":[-0.2315260619,0.0246403441,-0.0223660674,0.4574569166,0.1744244695,0.2295811772,-0.0073711821,0.5081834197,-0.0316528082,0.0191251654,-0.1842969209,-0.0640127659,-0.0091638044,0.3338081241,-0.043365553,-0.2194700241,-0.1107666269,0.1606068611,0.1407687962,0.0597507246,-0.1448137611,0.2009049803,-0.2154413909,-0.0725571662,-0.2707114816,-0.0130072096,0.1758372933,0.3205577433,0.0850609317,-0.5288669467,0.2025581449,-0.1739532501,0.0258219484,0.5030124187,-0.0001112284,0.0486640595,0.4246914983,-0.0443259999,-0.3031130135,0.0802712366,-0.309867084,-0.242859751,0.144365117,-0.0320097134,-0.3313594162,-0.1901502609,-0.2336005718,-0.337558955,0.2493240535,0.6167075038,0.238580361,0.3121445775,0.0608730987,0.0165075231,-0.1109841093,-0.1475731134,-0.022677483,0.3512011766,-0.0215199199,0.031706702,0.1784890443,0.2144178748,0.028754618,0.1094985083,0.1285811663,-0.1285620183,0.2033363879,-0.1063502431,0.1449157298,0.3846841753,0.3563730419,-0.2420736998,-0.1519511491,-0.3080828488,-0.1986611784,-0.2773949504,0.256582588,0.1999745071,-0.162488386,0.1661835164,-0.3945821524,0.2987675369,0.0589543805,0.0859668255,-0.1932324171,0.2412162274,0.151707381,-0.0265462417,0.0266118441,0.0962596685,0.1632707119,-0.1978420466,-0.0648603812,0.2136579305,-0.6099298596,-0.2218074054,-0.3099172413,0.4977063537,-0.0144231617,-0.051483281,-0.1140118986,-0.0171039887,0.1712173671,-0.186304301,0.1688324213,0.2142098099,-0.198009029,0.1207100525,0.1434043944,-0.1067880541,0.0161335897,-0.0791779235,0.162792176,-0.1318754554,-0.0637650043,0.2420955002,0.0574146733,-0.317915082,-0.6927220821,0.2458004653,-0.3038004041,-0.036881227,0.0007450588,0.3466166258,-0.3842547238,0.2641151547,0.1001107544,-0.1724211723,-0.2799533308,-0.2688709497,-0.2056975961,0.0522305109,-0.3685680926,0.0012723056,0.2524729669,-0.2094730288,0.4003704488,-0.0511978604,0.1751805246,-0.0478972159,0.209127903,-0.3762721717,-0.1081526875,0.2255997509,-0.3169129789,0.037633013,0.2689840496,-0.0927065387,0.0103949495,-0.0342599787,0.1593452543,-0.3435829878,-0.1637990624,0.287226826,-0.0145503711,0.0288387816,0.0684598163,-0.0540249497,0.3395628631,-0.1157185137,-0.1957776248,-0.2394090295,0.0330544971,-0.058229614,0.1889537871,0.4896891117,-0.4754011929,0.0806836635,-0.3405704498,0.1221416891,0.4467144907,0.2935122252,-0.0330286399,0.276563555,-0.2077404857,0.0237064268,-0.0174367968,-0.3686727285,-0.3444623053,0.5175626874,-0.0822370797,0.345344156,-0.0547711328,0.043166969,0.1374017298,0.0436606482,0.1547780186,0.2047380656,-0.1418510228,0.1543143243,-0.562957108,-0.1220002472,0.3635740578,0.0133565171,0.1890151352,-0.2069269419,0.1235904247,-0.0846622884,0.4912214279,-0.0084104314,-0.1206199005,-0.0321739204,0.2111113369,-0.0240932405,0.1006423309,-0.2743466496,-0.5518809557,0.2668113708,-0.0237271488,0.152853027,0.1286136508,-0.046923276,-0.589758873,-0.1467690319,-0.1562295854,-0.2404751182,0.1061195433,0.1226577386,0.2428137511,0.0796359479,-0.0546249487,0.5485526323,-0.0577965789,0.1413292736,-0.0799551159,0.341332972,-0.1525628865,-0.0305092093,-0.0302448831,0.1346586496,0.2463358045,-0.09839461,-0.3567022681,0.6420468092,0.1248638704,0.2066819072,-0.4426276088,0.0053086914,0.3827438354,-0.3125794828,0.0957020447,0.5969313979,0.2378956974,0.0181498863,0.0953133032,0.424510628,-0.0290682651,-0.0525034592,0.1060230434,-0.1878120452,0.1633400321,-0.1469224691,0.200978145,-0.3520150185,0.4401290417,-0.086315088,0.3760469556,-0.0243641734,-0.1455283016,0.0485069752,0.4009242058,-0.0737881437,-0.0185004827,-0.1634851992,-0.1819177121,-0.1184910983,0.3020989597,0.3241461515,0.6374415755,0.1745533198,-0.2192538083,0.2381288111,-0.000430839,-0.2599272132,0.4831356406,-0.1653624773,0.0127727846,0.4583308399,0.008977511,0.1091809645,-0.4487721622,0.1450789571,0.123402968,0.1818011999,-0.360792309,0.0788164586,-0.215221718,0.0968676656,-0.3518635631,-0.2619572878,0.029402364,-0.313379705,-0.0178715326,0.1368231773,-0.0503815673,0.2308113575,-0.4672618806,0.0846562535,0.1124338806,-0.123098582,-0.2004061192,0.0737325698,-0.3612599671,0.0988924429,0.253513217,0.0678496435,0.2908097208,-0.1956795603,-0.1511840522,0.0241016988,-0.4036478996,0.1278229356,0.1180695146,0.2275419384,0.0969556868,0.0744806901,0.3463863432,-0.4121953249,-0.071068041,-0.1308414191,-0.2183184475,0.17054151,0.0332118087,0.0247350875,-0.1156041622,-0.5987548232,-0.4374191463,-0.3892827332,-0.0498279557,-0.1645166576,-0.0728507191,0.1988957822,-0.1343782842,0.0514261425,0.1836190671,-0.0720977336,-0.4009529352,0.2122821659,0.1445584893,-0.0392751917,-0.2241797745,-0.1384652853,-0.0772775933,-0.0527735427,-0.0458171144,-0.3023410439,0.254326731,-0.4559999704,0.2435563952,-0.1985964179,-0.213123858,0.2872216702,0.0318476632,-0.0354612023,-0.1347290725,-0.0344382711,-0.0722915977,-0.1071643755,0.1636640579,-0.1976882517,0.394243598,0.1363981962,0.4813589752,0.2416753024,-0.5439124703,0.0680660531,0.1705723852,0.2048651129,-0.0990280136,-0.3353740275,0.3906064034,-0.0682675168,0.0239580069,0.2752487063,-0.3505270779,-0.2781621218,-0.0356105082,0.1964913607,-0.3761669397,-0.0256565753,-0.1338630915,-0.1636420041,0.2058549374,0.3254787922,0.2892023027,0.0354885459,-0.1637172848,0.2503463328,0.2493042946,0.0125087956,-0.1440640688,-0.1290749609,-0.1602590382,-0.2274690866,0.3832500875,0.3412743509,0.2640645802,0.0742882937,-0.1175951138,-0.0429789424,-0.0687436461,0.6774853468,-0.1602197737,0.1394208074,-0.0770248547,0.1332771629,-0.1524718404,-0.2240352929,-0.0229289755,0.0237797741,0.294780612,0.5989146829,-0.3209036589,-0.1446436495,0.3440994918,0.3235155046,-0.1614699215,-0.3134260774,-0.0346945636,-0.0641082749,-0.0452169776,0.1154499054,0.0462888815,0.2783490717,-0.1482951194,-0.153140083,0.0351074636,-0.0507183559,0.2441392988,0.3141935766,0.3286376297,-0.0386056788,0.3620902002,0.1574015468,0.302365303,0.372184962,0.2111066729,-0.345465064,-0.058314722,-0.1266359091,0.0932570174,-0.001311927,0.2818886042,-0.0490357913,-0.1346836835,0.2506635785,0.068419002,-0.152463302,0.3452251852,-0.0177854709,0.0991846919,-0.1798811108,-0.3432311416,0.2316904664,-0.1548494846,0.0145705231,0.1157020479,0.067900978,-0.3640975356,-0.0880728513,0.0034845076,0.7999356985,0.3829322457,-0.1509274393,0.2441919446,-0.3533546329,0.0227454789,-0.3782048523,0.1704315543,-0.3321321905,-0.1670113057,-0.0512643084,-0.1950098127,-0.0709553435,0.0745388493,-0.2098713219,0.2186295688,-0.1244219914,0.0774676353,0.2048077136,0.2370273769,0.0730773434,0.1332217753,-0.1807198673,0.1464785039,-0.0612236559,0.1584163606,-0.2090026289,-0.0780172944,0.0462242775,-0.1878198981,-0.2343229651,0.2004466653,-0.671544075,0.0274442788,0.0306347199,-0.0971903205,-0.0029017092,-0.0963324159,0.239186883,0.2715616524,-0.1229086146,0.0358495861,0.1376942098,0.153495729,0.0529686101,-0.0387600996,0.2003014684,-0.135901317,-0.1790282279,0.0824056566,-0.3000933826,-0.1808879673,0.2318426222,-0.0816883668,-0.0288241655,-0.3130398691,-0.1693854779,-0.0866780058,0.1064577252,-0.2034928203,0.1535087526,-0.2164582908,-0.3282732069,0.0480187908,0.0514248461,-0.3208444118,-0.3036953807,0.2600752115,0.1588468254,-0.0262651835,0.3788428903,0.0115331747,-0.3269037902,-0.2525985241,-0.1472427696,0.1546261013,-0.4197560251,0.0971366242,-0.3318922222,-0.2562456727,0.1576458812,0.2788511515,0.5067481399,0.0376338139,0.0551861934,-0.5445108414,-0.105076164,0.0463757105,-0.167886883,0.2210354954,-0.0551569201,0.6132634878,-0.3147180974,0.1814591736,-0.3386438787,0.0392292291,-0.1233162209,0.1893967837,0.000827592,-0.2967326045,0.2796502113,0.098179169,0.2074516565,0.3625247777,-0.1502892226,-0.1964601725,-0.1700348705,0.1034775451,0.0979663134,-0.2308011055,0.062586233,0.158974424,-0.0434559509,-0.008686211,0.1131301001,-0.1068085581,-0.0433675498,0.1655225754,0.3677061796,-0.0724938735,-0.2427455038,0.0912591964,-0.1443977952,0.0855922028,0.0484122522,0.1299035996,-0.148056075,0.0475814752,0.158301577,-0.0559338145,0.2772265077,0.2009176612,0.2070959061,-0.2499206066,-0.2945112884,0.2605382502,0.1768852025,0.4284108877,0.0114917271,-0.158310473,0.0698025599,0.2637248337,-0.3338602185,-0.0800075009,-0.0841673911,-0.2150553316,0.1919243038,0.0787677839,0.0997455642,-0.1859395355,-0.0469003171,0.2622558475,0.4332890213,-0.0860139802,0.109123379,0.398732245,-0.0716096759,0.0295816381,0.1499782354,-0.2502602935,0.0625841767,0.3447422087,-0.04796784,0.1358725876,0.0101845749,0.1770618111,-0.0911619663,-0.1617771238,0.2650981247,0.4197490513,-0.4040268958,0.2929238975,0.2387186885,-0.1023296788,-0.3085986972,-0.1845877022,-0.4084452093,-0.1939706504,-0.0719643459,0.1064393893,0.0173226986,-0.342833668,-0.2268976718,-0.0113695161,-0.0335704125,0.0478483923,0.5259116292,0.0245203599,-0.2330120802,-0.1281084269,-0.2837902904,0.2017405778,0.21901384,-0.3413544297,0.3417457938,0.2017515898,-0.1879087836,-0.2357747108,0.190875873,0.1929077357,0.4227327704,-0.2607415318,-0.1073196456,0.144697234,0.0094177518,0.123514995,0.4184183478,0.0590738058,-0.0722034201,0.226265043,0.1355629712,-0.2458421141,0.1419692785,-0.0804052204,-0.1136842966,-0.6303744316,0.3398411274,-0.0270323902,-0.0053561358,-0.106356062,0.2040739954,-0.3419698477,0.0560279228,0.3937089741,-0.1050297618,0.2485336363,0.1354949623,0.0990920588,-0.1431972235,0.3879002631,0.1573156416,0.0696986988,-0.5092074871,-0.1368087679,-0.4429211318,0.231569469,-0.1639965326,0.1835734993,-0.0463663936,0.0475714915,0.1086042672,0.18673639,0.1462301016,0.3793209493,0.1809259057,-0.0441591442,-0.2704292834,-0.1495058537,-0.1365631223,-0.2033466548,0.0330277979,-0.5399720669,0.0781174973,0.1010801494,0.1184633076,-0.0927306265,-0.2510053813,0.3491327763,0.2162149996,0.3482257128,0.3603941202,0.5763022304,-0.1550000161,-0.1805732548,0.002104782,-0.046220392,-0.1851202548,0.1420908719,0.1194762066,0.3269408047,0.0800936669,-0.1135649532,-0.2312363982,0.4827974141,0.3263706267,-0.2479363978,-0.255174309,-0.0562407076,-0.3647657037,-0.0320115685,0.2655400634,0.397505343,-0.0032958677,0.052043993,0.0077277757,-0.3785646856,0.4797550738,-0.5694298744,-0.0697973073,-0.1935780793,0.2503972948,-0.0434154011,-0.0811097994,-0.4199209809,0.0652301982,0.5131907463,-0.0961262435,-0.0506019965,0.1511285007,0.1545130163,-0.1854586005,0.0249983277,0.3290072381,0.3633755445,-0.2964656353,0.058436729,-0.2315704226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2761","title":"Error loading C4 realnewslike dataset","comments":"It works. I probably had some issue with the cache. after cleaning it im able to download the dataset. Thanks","body":"## Describe the bug\r\nError loading C4 realnewslike dataset. Validation part mismatch\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir)\r\n## Expected results\r\nsuccess on data loading\r\n## Actual results\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 15.3M\/15.3M [00:00<00:00, 28.1MB\/s]Traceback (most recent call last): \r\n File \"run_mlm_tf.py\", line 794, in \r\n main() \r\n File \"run_mlm_tf.py\", line 425, in main \r\n raw_datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 843, in load_dataset \r\n builder_instance.download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 608, in download_and_prepare \r\n self._download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 698, in _download_and_prepare verify_splits(self.info.splits, split_dict) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits \r\n raise NonMatchingSplitsSizesError(str(bad_splits)) \r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=38165657946, num_examples=13799838, dataset_name='c4'), 'recorded': SplitInfo(name='validation', num_bytes=37875873, num_examples=13863, dataset_name='c4')}] \r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1","comment_length":20,"text":"Error loading C4 realnewslike dataset\n## Describe the bug\r\nError loading C4 realnewslike dataset. Validation part mismatch\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir)\r\n## Expected results\r\nsuccess on data loading\r\n## Actual results\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 15.3M\/15.3M [00:00<00:00, 28.1MB\/s]Traceback (most recent call last): \r\n File \"run_mlm_tf.py\", line 794, in \r\n main() \r\n File \"run_mlm_tf.py\", line 425, in main \r\n raw_datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 843, in load_dataset \r\n builder_instance.download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 608, in download_and_prepare \r\n self._download_and_prepare( \r\n File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 698, in _download_and_prepare verify_splits(self.info.splits, split_dict) File \"\/home\/dshirron\/.local\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits \r\n raise NonMatchingSplitsSizesError(str(bad_splits)) \r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=38165657946, num_examples=13799838, dataset_name='c4'), 'recorded': SplitInfo(name='validation', num_bytes=37875873, num_examples=13863, dataset_name='c4')}] \r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\nIt works. I probably had some issue with the cache. after cleaning it im able to download the dataset. Thanks","embeddings":[-0.2315260619,0.0246403441,-0.0223660674,0.4574569166,0.1744244695,0.2295811772,-0.0073711821,0.5081834197,-0.0316528082,0.0191251654,-0.1842969209,-0.0640127659,-0.0091638044,0.3338081241,-0.043365553,-0.2194700241,-0.1107666269,0.1606068611,0.1407687962,0.0597507246,-0.1448137611,0.2009049803,-0.2154413909,-0.0725571662,-0.2707114816,-0.0130072096,0.1758372933,0.3205577433,0.0850609317,-0.5288669467,0.2025581449,-0.1739532501,0.0258219484,0.5030124187,-0.0001112284,0.0486640595,0.4246914983,-0.0443259999,-0.3031130135,0.0802712366,-0.309867084,-0.242859751,0.144365117,-0.0320097134,-0.3313594162,-0.1901502609,-0.2336005718,-0.337558955,0.2493240535,0.6167075038,0.238580361,0.3121445775,0.0608730987,0.0165075231,-0.1109841093,-0.1475731134,-0.022677483,0.3512011766,-0.0215199199,0.031706702,0.1784890443,0.2144178748,0.028754618,0.1094985083,0.1285811663,-0.1285620183,0.2033363879,-0.1063502431,0.1449157298,0.3846841753,0.3563730419,-0.2420736998,-0.1519511491,-0.3080828488,-0.1986611784,-0.2773949504,0.256582588,0.1999745071,-0.162488386,0.1661835164,-0.3945821524,0.2987675369,0.0589543805,0.0859668255,-0.1932324171,0.2412162274,0.151707381,-0.0265462417,0.0266118441,0.0962596685,0.1632707119,-0.1978420466,-0.0648603812,0.2136579305,-0.6099298596,-0.2218074054,-0.3099172413,0.4977063537,-0.0144231617,-0.051483281,-0.1140118986,-0.0171039887,0.1712173671,-0.186304301,0.1688324213,0.2142098099,-0.198009029,0.1207100525,0.1434043944,-0.1067880541,0.0161335897,-0.0791779235,0.162792176,-0.1318754554,-0.0637650043,0.2420955002,0.0574146733,-0.317915082,-0.6927220821,0.2458004653,-0.3038004041,-0.036881227,0.0007450588,0.3466166258,-0.3842547238,0.2641151547,0.1001107544,-0.1724211723,-0.2799533308,-0.2688709497,-0.2056975961,0.0522305109,-0.3685680926,0.0012723056,0.2524729669,-0.2094730288,0.4003704488,-0.0511978604,0.1751805246,-0.0478972159,0.209127903,-0.3762721717,-0.1081526875,0.2255997509,-0.3169129789,0.037633013,0.2689840496,-0.0927065387,0.0103949495,-0.0342599787,0.1593452543,-0.3435829878,-0.1637990624,0.287226826,-0.0145503711,0.0288387816,0.0684598163,-0.0540249497,0.3395628631,-0.1157185137,-0.1957776248,-0.2394090295,0.0330544971,-0.058229614,0.1889537871,0.4896891117,-0.4754011929,0.0806836635,-0.3405704498,0.1221416891,0.4467144907,0.2935122252,-0.0330286399,0.276563555,-0.2077404857,0.0237064268,-0.0174367968,-0.3686727285,-0.3444623053,0.5175626874,-0.0822370797,0.345344156,-0.0547711328,0.043166969,0.1374017298,0.0436606482,0.1547780186,0.2047380656,-0.1418510228,0.1543143243,-0.562957108,-0.1220002472,0.3635740578,0.0133565171,0.1890151352,-0.2069269419,0.1235904247,-0.0846622884,0.4912214279,-0.0084104314,-0.1206199005,-0.0321739204,0.2111113369,-0.0240932405,0.1006423309,-0.2743466496,-0.5518809557,0.2668113708,-0.0237271488,0.152853027,0.1286136508,-0.046923276,-0.589758873,-0.1467690319,-0.1562295854,-0.2404751182,0.1061195433,0.1226577386,0.2428137511,0.0796359479,-0.0546249487,0.5485526323,-0.0577965789,0.1413292736,-0.0799551159,0.341332972,-0.1525628865,-0.0305092093,-0.0302448831,0.1346586496,0.2463358045,-0.09839461,-0.3567022681,0.6420468092,0.1248638704,0.2066819072,-0.4426276088,0.0053086914,0.3827438354,-0.3125794828,0.0957020447,0.5969313979,0.2378956974,0.0181498863,0.0953133032,0.424510628,-0.0290682651,-0.0525034592,0.1060230434,-0.1878120452,0.1633400321,-0.1469224691,0.200978145,-0.3520150185,0.4401290417,-0.086315088,0.3760469556,-0.0243641734,-0.1455283016,0.0485069752,0.4009242058,-0.0737881437,-0.0185004827,-0.1634851992,-0.1819177121,-0.1184910983,0.3020989597,0.3241461515,0.6374415755,0.1745533198,-0.2192538083,0.2381288111,-0.000430839,-0.2599272132,0.4831356406,-0.1653624773,0.0127727846,0.4583308399,0.008977511,0.1091809645,-0.4487721622,0.1450789571,0.123402968,0.1818011999,-0.360792309,0.0788164586,-0.215221718,0.0968676656,-0.3518635631,-0.2619572878,0.029402364,-0.313379705,-0.0178715326,0.1368231773,-0.0503815673,0.2308113575,-0.4672618806,0.0846562535,0.1124338806,-0.123098582,-0.2004061192,0.0737325698,-0.3612599671,0.0988924429,0.253513217,0.0678496435,0.2908097208,-0.1956795603,-0.1511840522,0.0241016988,-0.4036478996,0.1278229356,0.1180695146,0.2275419384,0.0969556868,0.0744806901,0.3463863432,-0.4121953249,-0.071068041,-0.1308414191,-0.2183184475,0.17054151,0.0332118087,0.0247350875,-0.1156041622,-0.5987548232,-0.4374191463,-0.3892827332,-0.0498279557,-0.1645166576,-0.0728507191,0.1988957822,-0.1343782842,0.0514261425,0.1836190671,-0.0720977336,-0.4009529352,0.2122821659,0.1445584893,-0.0392751917,-0.2241797745,-0.1384652853,-0.0772775933,-0.0527735427,-0.0458171144,-0.3023410439,0.254326731,-0.4559999704,0.2435563952,-0.1985964179,-0.213123858,0.2872216702,0.0318476632,-0.0354612023,-0.1347290725,-0.0344382711,-0.0722915977,-0.1071643755,0.1636640579,-0.1976882517,0.394243598,0.1363981962,0.4813589752,0.2416753024,-0.5439124703,0.0680660531,0.1705723852,0.2048651129,-0.0990280136,-0.3353740275,0.3906064034,-0.0682675168,0.0239580069,0.2752487063,-0.3505270779,-0.2781621218,-0.0356105082,0.1964913607,-0.3761669397,-0.0256565753,-0.1338630915,-0.1636420041,0.2058549374,0.3254787922,0.2892023027,0.0354885459,-0.1637172848,0.2503463328,0.2493042946,0.0125087956,-0.1440640688,-0.1290749609,-0.1602590382,-0.2274690866,0.3832500875,0.3412743509,0.2640645802,0.0742882937,-0.1175951138,-0.0429789424,-0.0687436461,0.6774853468,-0.1602197737,0.1394208074,-0.0770248547,0.1332771629,-0.1524718404,-0.2240352929,-0.0229289755,0.0237797741,0.294780612,0.5989146829,-0.3209036589,-0.1446436495,0.3440994918,0.3235155046,-0.1614699215,-0.3134260774,-0.0346945636,-0.0641082749,-0.0452169776,0.1154499054,0.0462888815,0.2783490717,-0.1482951194,-0.153140083,0.0351074636,-0.0507183559,0.2441392988,0.3141935766,0.3286376297,-0.0386056788,0.3620902002,0.1574015468,0.302365303,0.372184962,0.2111066729,-0.345465064,-0.058314722,-0.1266359091,0.0932570174,-0.001311927,0.2818886042,-0.0490357913,-0.1346836835,0.2506635785,0.068419002,-0.152463302,0.3452251852,-0.0177854709,0.0991846919,-0.1798811108,-0.3432311416,0.2316904664,-0.1548494846,0.0145705231,0.1157020479,0.067900978,-0.3640975356,-0.0880728513,0.0034845076,0.7999356985,0.3829322457,-0.1509274393,0.2441919446,-0.3533546329,0.0227454789,-0.3782048523,0.1704315543,-0.3321321905,-0.1670113057,-0.0512643084,-0.1950098127,-0.0709553435,0.0745388493,-0.2098713219,0.2186295688,-0.1244219914,0.0774676353,0.2048077136,0.2370273769,0.0730773434,0.1332217753,-0.1807198673,0.1464785039,-0.0612236559,0.1584163606,-0.2090026289,-0.0780172944,0.0462242775,-0.1878198981,-0.2343229651,0.2004466653,-0.671544075,0.0274442788,0.0306347199,-0.0971903205,-0.0029017092,-0.0963324159,0.239186883,0.2715616524,-0.1229086146,0.0358495861,0.1376942098,0.153495729,0.0529686101,-0.0387600996,0.2003014684,-0.135901317,-0.1790282279,0.0824056566,-0.3000933826,-0.1808879673,0.2318426222,-0.0816883668,-0.0288241655,-0.3130398691,-0.1693854779,-0.0866780058,0.1064577252,-0.2034928203,0.1535087526,-0.2164582908,-0.3282732069,0.0480187908,0.0514248461,-0.3208444118,-0.3036953807,0.2600752115,0.1588468254,-0.0262651835,0.3788428903,0.0115331747,-0.3269037902,-0.2525985241,-0.1472427696,0.1546261013,-0.4197560251,0.0971366242,-0.3318922222,-0.2562456727,0.1576458812,0.2788511515,0.5067481399,0.0376338139,0.0551861934,-0.5445108414,-0.105076164,0.0463757105,-0.167886883,0.2210354954,-0.0551569201,0.6132634878,-0.3147180974,0.1814591736,-0.3386438787,0.0392292291,-0.1233162209,0.1893967837,0.000827592,-0.2967326045,0.2796502113,0.098179169,0.2074516565,0.3625247777,-0.1502892226,-0.1964601725,-0.1700348705,0.1034775451,0.0979663134,-0.2308011055,0.062586233,0.158974424,-0.0434559509,-0.008686211,0.1131301001,-0.1068085581,-0.0433675498,0.1655225754,0.3677061796,-0.0724938735,-0.2427455038,0.0912591964,-0.1443977952,0.0855922028,0.0484122522,0.1299035996,-0.148056075,0.0475814752,0.158301577,-0.0559338145,0.2772265077,0.2009176612,0.2070959061,-0.2499206066,-0.2945112884,0.2605382502,0.1768852025,0.4284108877,0.0114917271,-0.158310473,0.0698025599,0.2637248337,-0.3338602185,-0.0800075009,-0.0841673911,-0.2150553316,0.1919243038,0.0787677839,0.0997455642,-0.1859395355,-0.0469003171,0.2622558475,0.4332890213,-0.0860139802,0.109123379,0.398732245,-0.0716096759,0.0295816381,0.1499782354,-0.2502602935,0.0625841767,0.3447422087,-0.04796784,0.1358725876,0.0101845749,0.1770618111,-0.0911619663,-0.1617771238,0.2650981247,0.4197490513,-0.4040268958,0.2929238975,0.2387186885,-0.1023296788,-0.3085986972,-0.1845877022,-0.4084452093,-0.1939706504,-0.0719643459,0.1064393893,0.0173226986,-0.342833668,-0.2268976718,-0.0113695161,-0.0335704125,0.0478483923,0.5259116292,0.0245203599,-0.2330120802,-0.1281084269,-0.2837902904,0.2017405778,0.21901384,-0.3413544297,0.3417457938,0.2017515898,-0.1879087836,-0.2357747108,0.190875873,0.1929077357,0.4227327704,-0.2607415318,-0.1073196456,0.144697234,0.0094177518,0.123514995,0.4184183478,0.0590738058,-0.0722034201,0.226265043,0.1355629712,-0.2458421141,0.1419692785,-0.0804052204,-0.1136842966,-0.6303744316,0.3398411274,-0.0270323902,-0.0053561358,-0.106356062,0.2040739954,-0.3419698477,0.0560279228,0.3937089741,-0.1050297618,0.2485336363,0.1354949623,0.0990920588,-0.1431972235,0.3879002631,0.1573156416,0.0696986988,-0.5092074871,-0.1368087679,-0.4429211318,0.231569469,-0.1639965326,0.1835734993,-0.0463663936,0.0475714915,0.1086042672,0.18673639,0.1462301016,0.3793209493,0.1809259057,-0.0441591442,-0.2704292834,-0.1495058537,-0.1365631223,-0.2033466548,0.0330277979,-0.5399720669,0.0781174973,0.1010801494,0.1184633076,-0.0927306265,-0.2510053813,0.3491327763,0.2162149996,0.3482257128,0.3603941202,0.5763022304,-0.1550000161,-0.1805732548,0.002104782,-0.046220392,-0.1851202548,0.1420908719,0.1194762066,0.3269408047,0.0800936669,-0.1135649532,-0.2312363982,0.4827974141,0.3263706267,-0.2479363978,-0.255174309,-0.0562407076,-0.3647657037,-0.0320115685,0.2655400634,0.397505343,-0.0032958677,0.052043993,0.0077277757,-0.3785646856,0.4797550738,-0.5694298744,-0.0697973073,-0.1935780793,0.2503972948,-0.0434154011,-0.0811097994,-0.4199209809,0.0652301982,0.5131907463,-0.0961262435,-0.0506019965,0.1511285007,0.1545130163,-0.1854586005,0.0249983277,0.3290072381,0.3633755445,-0.2964656353,0.058436729,-0.2315704226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2759","title":"the meteor metric seems not consist with the official version","comments":"the issue is caused by the differences between varied meteor versions:\r\nmeteor1.0 is for https:\/\/aclanthology.org\/W07-0734.pdf\r\nmeteor1.5 is for https:\/\/aclanthology.org\/W14-3348.pdf\r\n\r\nhere is a very similar issue in NLTK\r\nhttps:\/\/github.com\/nltk\/nltk\/issues\/2655","body":"## Describe the bug\r\nThe computed meteor score seems strange because the value is very different from the scores computed by other tools. For example, I use the meteor score computed by [NLGeval](https:\/\/github.com\/Maluuba\/nlg-eval) as the reference (which reuses the official jar file for the computation)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_metric\r\nfrom nlgeval import NLGEval, compute_individual_metrics\r\n\r\nmeteor = load_metric('meteor')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = meteor.compute(predictions=predictions, references=references)\r\n# print the actual result\r\nprint(round(results[\"meteor\"], 4))\r\nmetrics_dict = compute_individual_metrics(references, predictions[0])\r\n# print the expected result\r\nprint(round(metrics_dict[\"METEOR\"], 4))\r\n```\r\nBy the way, you need to install the `nlg-eval` library first. Please check the installation guide [here](https:\/\/github.com\/Maluuba\/nlg-eval#setup), thanks!\r\n\r\n## Expected results\r\n`0.4474`\r\n\r\n## Actual results\r\n`0.7398`\r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: macOS-10.16-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 4.0.1\r\n","comment_length":28,"text":"the meteor metric seems not consist with the official version\n## Describe the bug\r\nThe computed meteor score seems strange because the value is very different from the scores computed by other tools. For example, I use the meteor score computed by [NLGeval](https:\/\/github.com\/Maluuba\/nlg-eval) as the reference (which reuses the official jar file for the computation)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_metric\r\nfrom nlgeval import NLGEval, compute_individual_metrics\r\n\r\nmeteor = load_metric('meteor')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = meteor.compute(predictions=predictions, references=references)\r\n# print the actual result\r\nprint(round(results[\"meteor\"], 4))\r\nmetrics_dict = compute_individual_metrics(references, predictions[0])\r\n# print the expected result\r\nprint(round(metrics_dict[\"METEOR\"], 4))\r\n```\r\nBy the way, you need to install the `nlg-eval` library first. Please check the installation guide [here](https:\/\/github.com\/Maluuba\/nlg-eval#setup), thanks!\r\n\r\n## Expected results\r\n`0.4474`\r\n\r\n## Actual results\r\n`0.7398`\r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: macOS-10.16-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 4.0.1\r\n\nthe issue is caused by the differences between varied meteor versions:\r\nmeteor1.0 is for https:\/\/aclanthology.org\/W07-0734.pdf\r\nmeteor1.5 is for https:\/\/aclanthology.org\/W14-3348.pdf\r\n\r\nhere is a very similar issue in NLTK\r\nhttps:\/\/github.com\/nltk\/nltk\/issues\/2655","embeddings":[-0.1358162314,-0.2349779308,-0.0059329974,0.4061324,0.1851489544,-0.2260374576,-0.1716093123,0.2357305288,-0.0438345596,0.0740853623,0.0668207631,0.3859927952,-0.0958541557,-0.4473044574,-0.0392022952,-0.0029859832,-0.0017699267,-0.2732459307,-0.1561685205,-0.1142284125,-0.0879566446,0.2557639182,-0.0586425364,0.0983389989,0.1512336284,0.2146469653,0.2339310795,-0.0853132606,-0.0735771358,-0.2372512668,0.4464174807,-0.2547119856,0.1061593369,0.5718839765,-0.0001092228,-0.1097544208,0.5471622348,-0.0057448437,-0.1640698314,-0.3211094439,-0.2009696811,-0.3649702668,0.3099010885,-0.2298337817,-0.1332615465,0.0013612013,-0.2734860778,-0.0977881551,0.5416399837,0.2425263971,0.2468080968,-0.0707297027,-0.032946784,-0.0735689253,0.0056162365,0.0035616672,-0.1855246723,0.4780148864,-0.2764458358,0.0538799241,-0.0193888284,0.1980319619,0.192500025,0.288261801,0.3047470748,-0.0374400355,0.205665499,0.1865397841,0.1254914254,0.2122563273,0.2753864527,-0.3759817779,-0.3653778434,0.1167431921,-0.0383625068,-0.4420903921,0.1010578945,0.0597524419,0.0617230311,-0.2433200181,-0.417296648,0.1959596425,-0.1231979579,0.1582983881,-0.0885476321,0.4110731184,-0.1813751906,0.1683813632,-0.1701655388,0.0795141235,0.1006190255,-0.1356270909,-0.0568454787,-0.0334441997,-0.1606684327,0.0442760922,-0.1364964545,-0.1545810848,0.0622734055,0.0040408052,0.2865042686,0.0556127653,0.4499130845,-0.0662461743,-0.083218731,0.1952988952,0.4034390152,-0.0914597362,0.2229097635,0.2841408253,0.071210146,0.0319353864,0.1968729347,-0.2307173014,0.5221300125,0.1731864214,-0.0789206848,-0.2163648307,-0.3225902021,0.0125966817,0.0460242592,0.0061025503,0.0650951937,-0.2966371477,-0.1263106912,-0.1950011104,0.0078040059,0.1437561363,-0.2404645085,-0.0113249114,-0.2011920661,0.1084214449,-0.1019063964,0.3199006915,-0.0643684641,0.1869410425,0.377910465,-0.0284285676,0.2165042758,-0.2593049705,-0.1498461068,-0.0044140611,-0.0484350361,0.0778284371,-0.1474400461,0.2168372571,0.2294588089,-0.2374290079,-0.1664121747,-0.0783710554,-0.0966554061,-0.2084736526,0.1554417312,0.2021888196,-0.3552140296,0.0330935195,-0.0279508755,0.1025886759,0.0691780522,-0.0243306719,0.0344141945,-0.2691453397,-0.2135478854,-0.0733912811,0.3713421822,-0.00008299,-0.2837857902,-0.0432695337,-0.3907486498,0.3113590181,0.109224923,-0.1440795809,-0.0074118418,0.0101120882,-0.1593013555,-0.4902365506,0.383169353,-0.2884459496,-0.3212443292,-0.1635175347,0.273311615,0.2212444246,-0.0917895138,0.0542549081,-0.2431739569,-0.0811820924,0.3172591031,-0.0166001562,0.3480672836,-0.1355618834,-0.5385566354,-0.0393849127,0.1130519435,0.0065353308,0.3192085624,-0.4252249002,0.1794968545,0.2294232994,0.0155098522,0.1722091585,0.0421790294,0.0602997728,0.0968743935,-0.4579890966,0.0290459003,-0.2238137126,-0.0304726809,0.2628813088,-0.204327181,-0.0593313463,0.1946983635,-0.0019937321,-0.5488238931,-0.058180809,-0.0294286348,-0.1989923418,0.1879660785,0.1149051413,-0.0626429021,-0.0436301716,0.0775676966,-0.1338417977,-0.3600866795,-0.0965371579,-0.0805342868,-0.0284555368,-0.1556739807,-0.1163060144,0.3951698244,0.2015009224,0.2209815085,-0.0824580118,-0.254796654,0.3142406344,0.2135628462,0.075458698,0.215813756,0.085991405,-0.1157252565,-0.2263933122,-0.1310671717,0.3775625229,0.1458840221,0.1624935418,-0.1309821159,0.3952777386,0.3242181242,0.093959786,-0.0954202637,-0.1156595871,-0.0092390869,-0.0147527177,-0.1969424337,-0.2568131983,0.1512787044,-0.0860581547,-0.021663608,-0.1682965606,0.2429612726,-0.1447513402,0.3565318286,0.0545579791,0.229637295,0.1990041733,-0.2496376783,-0.2357937098,-0.212897718,-0.162734434,0.3426095545,0.0317688026,0.0964511633,0.0694159269,-0.2506141961,-0.3128019273,0.1340085268,0.290913552,-0.0806699321,0.3633867204,0.3177281618,0.0116890101,-0.1174107641,0.2528041005,-0.0544438027,0.0673096851,-0.0671748221,0.2033380419,-0.2143352032,0.2841469049,-0.4203428626,-0.507707417,-0.6033649445,-0.4780286551,0.0246390644,-0.0339770317,0.2161393613,0.1809400916,0.3163756132,0.1196284816,0.0307187699,0.1664606631,-0.0013496839,-0.3262858391,-0.4050019383,0.1059098169,-0.0300928317,-0.0299215578,0.342389971,-0.1375697702,0.1093680933,-0.1515025944,-0.4955455661,0.02520556,-0.0457470231,0.7252534032,0.1718586832,-0.4527870119,0.0372483842,0.3458481431,0.1602809876,-0.3668874204,0.1735024303,0.2218369693,-0.2250281572,-0.2084215581,-0.0190014392,-0.22030209,0.1442066282,-0.1351414174,-0.2315960824,0.348626852,0.07076215,0.2468064278,-0.1577878743,0.1344600469,0.1748230159,0.3467097878,-0.2913876176,-0.1006377265,0.2914351225,-0.1825645417,-0.329429239,-0.0032992463,0.1110964864,0.4584417343,0.0616888627,-0.2635655701,-0.6254201531,-0.028324496,-0.0230475776,-0.0609051622,-0.1638613194,0.0816947371,-0.2560128272,-0.0483233407,-0.2598661482,-0.2592063546,-0.0988845974,0.1421607584,0.363828063,-0.1304811835,0.0940264612,0.1602395475,0.63237679,0.2696518302,-0.1276919395,-0.1066972688,0.1169157997,0.1045963392,-0.2340123951,-0.0122036031,0.1513127834,0.0846143886,0.335414201,0.3545050323,0.1112591177,0.3003501594,-0.2106336802,-0.0865532756,-0.3757596016,-0.2479578555,-0.16291821,0.0215300135,0.0458200909,0.2445279956,0.2455225885,-0.1743801236,-0.207165122,0.1453406364,0.4250761569,-0.1590637714,0.111364089,0.2787347138,0.0978732929,-0.2737770975,0.3948976696,0.1274559647,0.2267421484,0.1160897911,0.0500301905,0.0698362216,-0.1368580759,-0.04884905,-0.2523266971,0.3218267262,0.1352012604,-0.14239268,-0.6760870218,-0.2181343436,-0.2263398021,0.0616554059,0.2319736034,0.4251721799,-0.1331087947,-0.183606714,-0.1844701767,0.1671068221,-0.1929726154,-0.1721216142,-0.358448863,-0.0296944398,0.1986019462,-0.199478358,-0.2214978188,0.1997010559,-0.2016652375,0.0602731891,-0.1834780723,-0.2629290819,-0.0587269515,0.0503733456,0.2531851828,-0.2955349684,0.2033886611,-0.3119794726,-0.0680169091,-0.2577797174,0.0658239871,0.301826179,-0.1768380255,-0.0956353098,-0.1792342514,0.0770723745,0.1359236091,-0.1510064602,-0.1022670418,-0.4794323742,0.2801920176,0.0283420775,-0.089699097,0.1653415114,0.0140919434,-0.1515243798,0.0536121391,0.5751378536,0.3157077134,-0.1766789109,0.2900370657,-0.0357901603,-0.33055076,-0.1364216208,0.2808996439,0.6272659302,0.4066708982,0.0119327772,0.0492441654,-0.2753590643,0.3018888831,0.3433157802,0.1594590247,-0.4138987958,-0.2181342989,0.0412715152,0.1394530535,0.0428555608,-0.3080528975,0.1133087799,0.5284907818,-0.0817252174,0.1583693326,-0.0079945466,0.146564275,0.0654188618,0.1467654854,-0.0621022284,0.2710240483,-0.1619922519,0.2423302978,-0.001878597,0.0690987483,-0.2022750378,-0.0296115782,-0.1416733712,-0.0009688644,0.5955713391,0.0765041932,0.2002232373,-0.455109179,0.048600398,-0.117503956,0.5312615633,0.3216865659,-0.2299820185,0.26341784,-0.5472123027,0.2441687733,-0.1307070702,0.127983883,-0.069757238,-0.1185006127,0.0187034551,-0.1369009316,-0.0360005349,0.0825165063,-0.062099535,-0.1582513452,0.3788276017,-0.0411832295,0.4483365715,-0.3621811271,-0.0472270213,-0.2663274407,0.2131647766,-0.2188328803,0.0014460111,0.2188296169,0.094892852,-0.3635932505,0.0428369008,0.4329023063,-0.0534552149,0.4348675311,-0.0092122843,-0.178900063,0.0522656702,-0.1079750359,-0.2123864442,-0.0846213847,-0.1047981605,0.0343203582,-0.1124680787,0.2750742733,0.2135679573,0.4239366353,0.2719598711,-0.183675915,-0.155567944,-0.5252074599,0.2682492137,-0.032894481,-0.1070640609,0.0073942388,-0.1867925227,-0.1103931814,0.1266532838,0.2420267761,-0.3334038258,-0.0147571433,0.0707560033,0.0505085289,-0.0600869544,-0.2526158392,-0.0834790319,0.2036734223,0.0670769364,0.1873745024,-0.2901427746,-0.2291766852,-0.289362371,0.0908685401,0.1242933571,0.0077958126,0.0496295542,0.213177219,-0.1830178648,-0.2457816899,-0.1480923891,0.1141904965,0.1359606981,0.3024585247,-0.0604503006,0.5086019039,0.0653214604,-0.0861155838,0.0335970782,0.2548256516,-0.0884656087,0.1084715575,0.187520802,-0.0298928618,-0.1848643869,-0.0308391582,-0.1448867768,0.1882310361,0.2696504593,0.1290748268,0.0684604645,0.1303534657,0.2986498177,0.2236769944,-0.063790895,-0.444234252,-0.2394886166,0.1733602881,-0.2343720943,-0.0406975411,-0.2714272439,0.3211047947,-0.0014931345,0.0613797009,0.1544311941,-0.1894699186,-0.1897162944,-0.0988617092,0.7578188777,-0.4481823444,0.0421200916,-0.1400887519,0.2631900907,0.0463284627,0.429222554,-0.2257142961,0.2814006209,0.2795659304,-0.2038105279,0.3184913695,-0.1076937616,0.0860003307,0.0314964801,0.1829087436,0.1649773121,0.3285511732,0.2327073663,-0.0589885786,-0.0093524456,0.3833883703,-0.2099046856,0.0567064695,-0.4581983685,-0.1379719526,0.2848409712,-0.0877969638,-0.4621794224,-0.1289704442,-0.1818467081,-0.0258141421,-0.0887721628,0.0600941367,0.0153890736,-0.0791159868,-0.121839039,-0.3050372601,-0.2565439641,0.364189446,0.1647266746,-0.2404817194,0.29067716,0.432559967,0.1009295061,-0.1496160775,0.0716873407,0.161982283,0.1205393299,-0.2520706654,-0.0765159205,0.3310955465,-0.0832405239,0.1071996987,0.2756618857,-0.0569793694,-0.103544496,0.3694298863,0.2167431265,-0.2008022815,0.269274354,0.005107671,0.0936138481,-0.3605012596,0.5348919034,-0.1433045715,0.1428109109,0.0146245332,0.1432060748,-0.2738912404,-0.3752201498,-0.1137516871,0.3000074327,0.2451164573,-0.0801248029,0.1131700203,-0.2373938411,0.4121416211,-0.1286602467,-0.4160782993,-0.0516928881,-0.1181043461,-0.4005631506,0.2340830266,0.1683342606,0.2376131415,-0.1082196757,-0.0538191833,-0.3112965524,0.1834893525,-0.0009353615,0.352842927,-0.083482042,-0.1732328236,-0.5152145028,-0.0506459847,0.0662170798,-0.5063577294,0.0115721766,-0.0732128546,0.0404262394,-0.5476977825,0.1411287487,0.111779891,-0.3006656766,0.5468631387,0.364628911,0.4902895093,0.2103084028,0.2337276191,-0.1005822495,-0.1806226969,-0.0830208361,0.1566406041,-0.092849277,0.5090836287,-0.4025939107,0.269013077,0.0803936422,0.086972557,-0.0826480612,0.4936469495,0.09498097,-0.2323929965,-0.0688754991,-0.1761599332,-0.1346302181,-0.0426749066,-0.0072233402,0.2450737506,-0.0961180329,0.301433295,-0.1895337105,-0.3276013434,0.320858866,-0.1526229531,-0.2464816123,-0.131466791,0.2957193851,0.0376136713,-0.0539723337,-0.2076196671,0.0782645643,0.4740905762,-0.0011416521,0.2411159426,0.2820770144,-0.155720219,0.0282734074,-0.0181427374,0.2410161495,0.251188308,-0.1877973825,-0.0099866576,-0.0490341522]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2759","title":"the meteor metric seems not consist with the official version","comments":"Hi @jianguda, thanks for reporting.\r\n\r\nCurrently, at \ud83e\udd17 `datasets` we are using METEOR 1.0 (indeed using NLTK: `from nltk.translate import meteor_score`): See the [citation here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/meteor\/meteor.py#L23-L35).\r\n\r\nIf there is some open source implementation of METEOR 1.5, that could be an interesting contribution! \ud83d\ude09 ","body":"## Describe the bug\r\nThe computed meteor score seems strange because the value is very different from the scores computed by other tools. For example, I use the meteor score computed by [NLGeval](https:\/\/github.com\/Maluuba\/nlg-eval) as the reference (which reuses the official jar file for the computation)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_metric\r\nfrom nlgeval import NLGEval, compute_individual_metrics\r\n\r\nmeteor = load_metric('meteor')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = meteor.compute(predictions=predictions, references=references)\r\n# print the actual result\r\nprint(round(results[\"meteor\"], 4))\r\nmetrics_dict = compute_individual_metrics(references, predictions[0])\r\n# print the expected result\r\nprint(round(metrics_dict[\"METEOR\"], 4))\r\n```\r\nBy the way, you need to install the `nlg-eval` library first. Please check the installation guide [here](https:\/\/github.com\/Maluuba\/nlg-eval#setup), thanks!\r\n\r\n## Expected results\r\n`0.4474`\r\n\r\n## Actual results\r\n`0.7398`\r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: macOS-10.16-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 4.0.1\r\n","comment_length":42,"text":"the meteor metric seems not consist with the official version\n## Describe the bug\r\nThe computed meteor score seems strange because the value is very different from the scores computed by other tools. For example, I use the meteor score computed by [NLGeval](https:\/\/github.com\/Maluuba\/nlg-eval) as the reference (which reuses the official jar file for the computation)\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_metric\r\nfrom nlgeval import NLGEval, compute_individual_metrics\r\n\r\nmeteor = load_metric('meteor')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = meteor.compute(predictions=predictions, references=references)\r\n# print the actual result\r\nprint(round(results[\"meteor\"], 4))\r\nmetrics_dict = compute_individual_metrics(references, predictions[0])\r\n# print the expected result\r\nprint(round(metrics_dict[\"METEOR\"], 4))\r\n```\r\nBy the way, you need to install the `nlg-eval` library first. Please check the installation guide [here](https:\/\/github.com\/Maluuba\/nlg-eval#setup), thanks!\r\n\r\n## Expected results\r\n`0.4474`\r\n\r\n## Actual results\r\n`0.7398`\r\n\r\n## Environment info\r\n- `datasets` version: 1.10.2\r\n- Platform: macOS-10.16-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 4.0.1\r\n\nHi @jianguda, thanks for reporting.\r\n\r\nCurrently, at \ud83e\udd17 `datasets` we are using METEOR 1.0 (indeed using NLTK: `from nltk.translate import meteor_score`): See the [citation here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/meteor\/meteor.py#L23-L35).\r\n\r\nIf there is some open source implementation of METEOR 1.5, that could be an interesting contribution! \ud83d\ude09 ","embeddings":[-0.1478186399,-0.2521868348,-0.0246564019,0.3410737813,0.231181547,-0.2251470685,-0.1337345839,0.2375303209,-0.1029678956,0.1020672619,0.0819613785,0.3635589182,-0.1630420536,-0.3619390726,-0.049654711,0.0434393324,-0.0236538462,-0.3071354032,-0.1534652114,-0.0484819748,-0.0770225972,0.2145742178,-0.0249223225,0.0925159156,0.1617095172,0.2413862497,0.1707571447,-0.0400183573,-0.144275859,-0.2580814362,0.3748160005,-0.1964260042,0.10105367,0.5309472084,-0.0001044121,-0.1293847114,0.5862187147,0.0081376778,-0.1794004887,-0.3455775976,-0.2490601987,-0.3852468133,0.2880985439,-0.1919128895,-0.0959120914,0.0437685139,-0.2066351175,-0.1008442417,0.5769690275,0.2530090511,0.292075187,-0.1218493581,0.0126701789,-0.0997717828,0.0463839918,0.0661911368,-0.1800993085,0.4117848575,-0.2385418862,0.0982186273,-0.0425276309,0.2050881833,0.193847999,0.2316541821,0.2611716688,-0.0546889082,0.2156092823,0.1711399853,0.0779589489,0.191200003,0.3276813328,-0.3993017972,-0.3931284249,0.1359330714,-0.0082493555,-0.4920450151,0.0499257036,0.0782915428,0.0225335881,-0.21199435,-0.400894165,0.1289479434,-0.122258082,0.1270890534,-0.0585687608,0.3907246888,-0.2248826474,0.1203368232,-0.1347391307,0.0584191754,0.0831416845,-0.0383946113,-0.0343862474,-0.0310811587,-0.1347252578,0.0342390947,-0.1566369385,-0.1525467634,0.1690117717,-0.0029825966,0.2433558255,0.1008022577,0.4526496232,-0.0251307711,-0.0411129817,0.1939606369,0.4406630397,-0.1113480702,0.250462085,0.3186845779,0.1044172049,0.0632328242,0.1617822349,-0.1916848868,0.5070535541,0.1325170547,-0.0137895914,-0.1788939983,-0.3429356813,0.0918602422,0.0669184998,0.008175333,0.1029940397,-0.1799635738,-0.1395748407,-0.1464832872,-0.0156325214,0.0464368649,-0.1846251488,-0.0662353337,-0.2085932642,0.1549058557,-0.2039872557,0.2892301381,-0.0416440107,0.1677290797,0.3400670886,-0.0389067456,0.2023890913,-0.2363194972,-0.1112823412,0.0133613609,-0.0415274799,0.0646547079,-0.1101486385,0.1739303023,0.1842858344,-0.2790244222,-0.1020318046,-0.0391604044,-0.1144148707,-0.1948991865,0.0797411352,0.2515948415,-0.3690243363,0.0499073751,-0.0140078478,0.1826043427,0.0627038777,-0.0204932317,0.052719783,-0.1326204091,-0.1579742134,-0.0940806121,0.3570547402,0.11149133,-0.2813995779,-0.0093765743,-0.4308559597,0.3145482242,0.1247430742,-0.2065574676,0.0224438179,-0.0567487441,-0.161562711,-0.5349422097,0.3647163808,-0.2356968522,-0.3380417228,-0.2351226211,0.2033908218,0.1315128952,-0.0577308834,0.0421362966,-0.2469061017,-0.0696172565,0.2785610855,-0.0286377519,0.2881574929,-0.1320801675,-0.5351973176,-0.0517961942,0.1169945449,0.0778272301,0.301897198,-0.4567933679,0.2465071231,0.2115737647,-0.0054074037,0.2137972265,0.0483120047,0.0966469124,0.1147342548,-0.4557758868,0.1202820092,-0.2431789339,-0.0412268676,0.2640697956,-0.1666318029,-0.120225735,0.2189391404,0.0205227211,-0.53795892,0.0368409492,-0.0812163129,-0.182063356,0.2198257893,0.061349161,-0.1401595324,-0.0993563458,-0.0055019571,-0.1313671023,-0.3810173869,-0.0816131607,-0.1145739183,-0.0536090322,-0.1969023198,-0.1277427226,0.3600816727,0.1595160812,0.1943795532,-0.0530176349,-0.2322896719,0.3451606035,0.2800037861,0.0231145658,0.2230317891,0.1184838116,-0.087079905,-0.2262252718,-0.1516230553,0.2881505489,0.1753278822,0.173419565,-0.2755355239,0.3950206041,0.2253572941,0.0255402755,-0.1097524241,-0.0231604408,0.0361028165,0.0109119397,-0.2923180163,-0.2639077008,0.1407266557,-0.1132562086,0.0147086708,-0.1323924512,0.1913963258,-0.152247414,0.3871747851,0.0431432165,0.3020338416,0.2685679197,-0.2179817557,-0.2699264884,-0.2127495408,-0.1749303043,0.2869413197,0.1010900438,0.1103068516,0.0748123899,-0.1497457176,-0.2546248436,0.1458888501,0.2289117575,-0.1467743069,0.2909068763,0.312820226,0.0072740866,-0.1140569821,0.2003953904,-0.0769301653,0.0734622404,-0.0910370052,0.120904319,-0.1940790713,0.3012647033,-0.3821586072,-0.5017484426,-0.6062213182,-0.4944030643,0.0655424744,-0.0839283392,0.2129885256,0.1757662743,0.3269561827,0.1595528871,0.0667396188,0.2071737945,-0.0170070548,-0.3314951956,-0.2945849299,0.1425486505,-0.080276832,-0.0440404974,0.3598217964,-0.1186938062,0.1431424469,-0.12410786,-0.4709443152,0.0456022993,0.0082603088,0.6918196678,0.2361387759,-0.4747566581,-0.0206395537,0.4039651155,0.1259763986,-0.387466222,0.1384506524,0.2865402699,-0.2143689692,-0.2014358789,0.010843494,-0.1430825889,0.111751087,-0.134806022,-0.2115037143,0.390283227,0.0237564035,0.2410504073,-0.1284732223,0.2649903297,0.1560121626,0.3295111954,-0.3026740551,-0.2036079019,0.3007966578,-0.1732979566,-0.3139922917,0.0291742813,0.1207088381,0.3786696792,-0.0149592198,-0.1902527809,-0.6665781736,-0.0313941762,-0.0594150573,-0.0360873938,-0.1200876758,0.1167293563,-0.2164951265,-0.0864942819,-0.2213131785,-0.226139009,-0.1013961434,0.1149089932,0.350844413,-0.1825340539,0.0895560682,0.2299666256,0.7172644734,0.2367880195,-0.1143335998,-0.1029946432,0.0696392357,0.1311946213,-0.2151097208,-0.0539521091,0.1754362881,0.053729713,0.263405025,0.3654416203,0.1269430816,0.3131904304,-0.2360469699,-0.1155113205,-0.3265323639,-0.1952673346,-0.2356277108,-0.007068912,-0.050479982,0.2195651233,0.2327741683,-0.087667428,-0.2071623951,0.1731901616,0.4234602749,-0.158857584,0.1027152166,0.3724788725,0.1140843332,-0.3238576651,0.331469208,0.0909412876,0.1864419281,0.1126782075,0.0550086275,-0.003114427,-0.1909400076,-0.0209784582,-0.2531184852,0.310133338,0.0446419381,-0.1225215346,-0.5995672345,-0.1632290483,-0.2427881062,0.0216916818,0.1464215219,0.3101508915,-0.0808231831,-0.1328075081,-0.1513459533,0.1208471656,-0.1552976221,-0.1163096428,-0.3726927042,-0.0631828532,0.1353500485,-0.1255544871,-0.1635237336,0.1212916449,-0.1136774644,0.001340196,-0.174462229,-0.2919600308,-0.0350033902,0.0306291897,0.1390445977,-0.3460538685,0.1883102804,-0.2414290607,-0.0478549898,-0.2576870024,0.102518566,0.3154451251,-0.2469968349,-0.0589596741,-0.1768931448,0.1138710678,0.0974818543,-0.2113996893,-0.092966266,-0.5180155635,0.2825884223,0.0095768394,-0.090242222,0.1223739237,0.0630613193,-0.1648563296,0.0411987342,0.5064479113,0.2582514882,-0.1460028142,0.3888777494,-0.016643066,-0.2525936365,-0.1229887158,0.3109763563,0.6403035522,0.3419134021,-0.0344642662,0.0447286777,-0.2284908891,0.2664886415,0.286185205,0.1647363752,-0.4142992496,-0.2808918357,0.0430981629,0.1353817433,0.0765112564,-0.3224915862,0.0715386569,0.4974339604,-0.1406417936,0.0987119228,0.0170096662,0.1865603775,0.077574119,0.0481401645,-0.0448863991,0.2939742804,-0.1924897134,0.1315662414,-0.0096635111,0.1223281696,-0.0720281824,-0.0076266704,-0.1705517024,0.0580733493,0.5772729516,0.0531663522,0.1628954262,-0.3432329297,0.019839026,-0.0865589827,0.4654538035,0.3328102529,-0.2302966416,0.2989504933,-0.5688626766,0.2628832161,-0.1434999853,0.1299234927,0.0013416361,-0.094451651,0.0707464889,-0.0955878869,0.0549443401,0.1475526243,-0.0062728622,-0.1776949763,0.3491710424,-0.0172587149,0.4607383907,-0.3853222132,-0.1181579754,-0.2469914705,0.2481600493,-0.1255541742,-0.0896086767,0.1540971845,0.1558179408,-0.3281252682,0.0292922053,0.5054631829,-0.1028209776,0.4280501306,-0.0792577192,-0.1355879754,0.0373902917,-0.1363711655,-0.0528768487,-0.0011773357,-0.0515268184,0.047164809,-0.147827372,0.3012148738,0.2551988065,0.4241328835,0.2505405545,-0.2228532135,-0.132834509,-0.5114203691,0.2083126456,-0.045368351,-0.0960854217,-0.0417178795,-0.1661293209,-0.1720039696,0.1735002548,0.258081615,-0.385946393,-0.0752197653,0.0398499854,0.0059607588,-0.0258199181,-0.2938759029,-0.1433079243,0.1278073937,0.119267188,0.15937078,-0.2800629139,-0.2723571062,-0.2823260427,0.0656557456,0.0733794719,0.0689810216,0.0012033993,0.2080414295,-0.2676582336,-0.2233207226,-0.1572520137,0.1703042537,0.1034846827,0.2949127257,-0.0643940791,0.4062260389,0.0453626029,-0.1708671302,0.0585310794,0.2147109956,-0.0531996414,0.1193390563,0.2025176287,-0.06298922,-0.1279336065,0.0352382772,-0.0675161406,0.1731337011,0.3118720055,0.0562356971,0.0884849206,0.1232145503,0.3532485664,0.124633424,-0.1060488448,-0.4063402116,-0.217629537,0.2336735427,-0.2540642321,-0.1200785711,-0.2477079332,0.3661583066,0.003107371,0.0788147673,0.0793543831,-0.1804215908,-0.1613420546,-0.0794909522,0.7314050198,-0.4885142446,0.0216299221,-0.1493761092,0.2265654057,0.0350743458,0.440523088,-0.1372164488,0.3410734832,0.3740026951,-0.2450247407,0.4250045419,-0.0429956205,0.2151365727,0.0139032546,0.1537510902,0.2210005522,0.3062160611,0.203381449,-0.0450340658,-0.0749172643,0.3186627626,-0.1925454885,0.1538996696,-0.408044368,-0.1106438041,0.2830381393,-0.1046667546,-0.3735601008,-0.1484468579,-0.2269961536,-0.0403976589,-0.0507255234,0.0631420016,0.0036714289,-0.0868485048,-0.0783483982,-0.2781256735,-0.182688117,0.3552906811,0.2407272756,-0.3074507415,0.2518776655,0.3919894397,0.1222327426,-0.2010670006,-0.0159658249,0.1012648568,0.1369148046,-0.3087234497,-0.0339081697,0.2608788013,-0.1441622972,0.0951404721,0.2885707319,-0.0707431808,-0.0812909827,0.3715008795,0.262139082,-0.2381005287,0.2789750099,0.0364155397,0.0859423056,-0.3769672215,0.4557329714,-0.1082673892,0.0914920568,-0.0132459812,0.144167006,-0.2951063216,-0.4457588792,-0.0430305749,0.2353134304,0.1749767959,-0.1530628949,0.1383683383,-0.1613853127,0.3768999577,-0.1761392951,-0.378228724,-0.0125433449,-0.0520633422,-0.4072153568,0.1669828147,0.2578726411,0.2493457347,-0.0116366223,-0.0554786623,-0.3329471648,0.2094543725,-0.0443082564,0.4313583672,-0.1041297391,-0.1514206082,-0.4020168483,-0.0338057354,-0.0191725343,-0.433052063,-0.015140499,-0.0386177972,0.0190421827,-0.5741429329,0.1770795584,0.1041552052,-0.3532292247,0.5408803225,0.3004406989,0.4227412343,0.1769685447,0.1479733139,-0.1514249444,-0.0963379592,-0.1034426689,0.107019119,-0.1194014326,0.4396036267,-0.3794595599,0.1813824177,0.1088851839,0.0769578516,-0.1336252689,0.4373620749,0.0923435986,-0.1658055186,-0.0845742375,-0.1421273649,-0.0927540511,-0.0068176389,-0.0510414653,0.1910166889,-0.0871870518,0.2374994308,-0.1445524246,-0.2393935025,0.3237603605,-0.1295939982,-0.2335127294,-0.1668452621,0.2707190812,0.0491230451,-0.0625127777,-0.222731173,0.0414360724,0.4593601227,0.0213779677,0.1399921477,0.2271161079,-0.1391268373,0.05677744,-0.0739607066,0.2057380825,0.2255966961,-0.0973947719,0.0037193885,-0.0437505208]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2757","title":"Unexpected type after `concatenate_datasets`","comments":"Hi @JulesBelveze, thanks for your question.\r\n\r\nNote that \ud83e\udd17 `datasets` internally store their data in Apache Arrow format.\r\n\r\nHowever, when accessing dataset columns, by default they are returned as native Python objects (lists in this case).\r\n\r\nIf you would like their columns to be returned in a more suitable format for your use case (torch arrays), you can use the method `set_format()`:\r\n```python\r\nconcat_dataset.set_format(type=\"torch\")\r\n```\r\n\r\nYou have detailed information in our docs:\r\n- [Using a Dataset with PyTorch\/Tensorflow](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html)\r\n- [Dataset.set_format()](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html#datasets.Dataset.set_format)","body":"## Describe the bug\r\nI am trying to concatenate two `Dataset` using `concatenate_datasets` but it turns out that after concatenation the features are casted from `torch.Tensor` to `list`. \r\nIt then leads to a weird tensors when trying to convert it to a `DataLoader`. However, if I use each `Dataset` separately everything behave as expected.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> featurized_teacher\r\nDataset({\r\n features: ['t_labels', 't_input_ids', 't_token_type_ids', 't_attention_mask'],\r\n num_rows: 502\r\n})\r\n>>> for f in featurized_teacher.features:\r\n print(featurized_teacher[f].shape)\r\ntorch.Size([502])\r\ntorch.Size([502, 300])\r\ntorch.Size([502, 300])\r\ntorch.Size([502, 300])\r\n\r\n>>> featurized_student\r\nDataset({\r\n features: ['s_features', 's_labels'],\r\n num_rows: 502\r\n})\r\n>>> for f in featurized_student.features:\r\n print(featurized_student[f].shape)\r\ntorch.Size([502, 64])\r\ntorch.Size([502])\r\n```\r\nThe shapes seem alright to me. Then the results after concatenation are as follow:\r\n```python\r\n>>> concat_dataset = datasets.concatenate_datasets([featurized_student, featurized_teacher], axis=1)\r\n>>> type(concat_dataset[\"t_labels\"])\r\n\r\n```\r\nOne would expect to obtain the same type as the one before concatenation.\r\n\r\nAm I doing something wrong here? Any idea on how to fix this unexpected behavior?\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-10.14.6-x86_64-i386-64bit\r\n- Python version: 3.9.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":80,"text":"Unexpected type after `concatenate_datasets`\n## Describe the bug\r\nI am trying to concatenate two `Dataset` using `concatenate_datasets` but it turns out that after concatenation the features are casted from `torch.Tensor` to `list`. \r\nIt then leads to a weird tensors when trying to convert it to a `DataLoader`. However, if I use each `Dataset` separately everything behave as expected.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> featurized_teacher\r\nDataset({\r\n features: ['t_labels', 't_input_ids', 't_token_type_ids', 't_attention_mask'],\r\n num_rows: 502\r\n})\r\n>>> for f in featurized_teacher.features:\r\n print(featurized_teacher[f].shape)\r\ntorch.Size([502])\r\ntorch.Size([502, 300])\r\ntorch.Size([502, 300])\r\ntorch.Size([502, 300])\r\n\r\n>>> featurized_student\r\nDataset({\r\n features: ['s_features', 's_labels'],\r\n num_rows: 502\r\n})\r\n>>> for f in featurized_student.features:\r\n print(featurized_student[f].shape)\r\ntorch.Size([502, 64])\r\ntorch.Size([502])\r\n```\r\nThe shapes seem alright to me. Then the results after concatenation are as follow:\r\n```python\r\n>>> concat_dataset = datasets.concatenate_datasets([featurized_student, featurized_teacher], axis=1)\r\n>>> type(concat_dataset[\"t_labels\"])\r\n\r\n```\r\nOne would expect to obtain the same type as the one before concatenation.\r\n\r\nAm I doing something wrong here? Any idea on how to fix this unexpected behavior?\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-10.14.6-x86_64-i386-64bit\r\n- Python version: 3.9.5\r\n- PyArrow version: 3.0.0\r\n\nHi @JulesBelveze, thanks for your question.\r\n\r\nNote that \ud83e\udd17 `datasets` internally store their data in Apache Arrow format.\r\n\r\nHowever, when accessing dataset columns, by default they are returned as native Python objects (lists in this case).\r\n\r\nIf you would like their columns to be returned in a more suitable format for your use case (torch arrays), you can use the method `set_format()`:\r\n```python\r\nconcat_dataset.set_format(type=\"torch\")\r\n```\r\n\r\nYou have detailed information in our docs:\r\n- [Using a Dataset with PyTorch\/Tensorflow](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html)\r\n- [Dataset.set_format()](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html#datasets.Dataset.set_format)","embeddings":[0.0659025609,-0.2509687543,0.0296784658,0.507555902,0.4053797126,0.1719074398,0.4817784131,0.1346458942,-0.3654304147,-0.1315238774,-0.1543308496,0.395609051,-0.0383024774,-0.0193488616,-0.089732863,-0.2919967473,0.2896149158,-0.0016607346,-0.2081793547,-0.2273828089,-0.0381968655,0.1582772583,-0.2841176987,0.035756506,-0.2161132842,0.03157226,-0.0633054227,-0.1039584205,0.0063897204,0.0390482917,0.4256393015,0.0004014391,0.2480084449,0.5711277723,-0.0001217747,0.0700064898,0.1109226272,0.0230530966,-0.0054579936,-0.0247396491,-0.2727138102,-0.3866924644,0.1001271084,-0.2212410122,0.0955420285,-0.276202023,-0.2802899182,-0.456353426,-0.0949754938,0.1694531739,0.151410982,0.2291548699,0.0298358481,0.1134988666,0.1770875156,0.1690293849,-0.1362749636,-0.1136704534,-0.0572797284,0.2449533194,0.447096467,-0.0400952436,-0.1846583337,-0.1947271824,0.0227859411,0.2623610497,-0.314884007,-0.3963642418,-0.1733084768,0.1674891263,0.526042819,-0.1522032768,-0.1825706363,-0.1398912817,0.0140495589,0.1007185876,0.1467597932,0.1790630221,0.1674614847,-0.0629968941,0.001977049,0.2578364313,-0.0612737238,0.2807202637,-0.2607089579,-0.0113037117,-0.1205294207,0.2740452588,-0.0921653509,-0.1820497215,0.186672464,-0.2658762336,-0.0710545853,-0.0162656344,-0.5050219297,-0.0840443373,0.0590057038,-0.509517312,-0.1505635381,-0.142667532,0.1211892888,0.0527787209,-0.2076213062,0.1359153986,0.2071558088,0.3062790036,-0.0643244088,0.3882654905,-0.0904146731,-0.0109213414,-0.0627332702,0.1054688171,-0.030391179,-0.3052833378,0.0818851665,0.025435878,0.3865280449,-0.0147889825,-0.3727706671,0.0108852461,-0.3847967088,0.2573918104,0.0848566517,-0.1533523351,-0.3881433904,0.5925773978,0.2986513972,0.2811685205,0.0565008447,-0.0198840369,-0.0571242161,-0.2771323323,-0.0768730938,-0.2397636622,0.0317757875,-0.1434997618,-0.2257492691,0.4488455653,0.1257311851,0.0125266463,-0.0595633425,-0.0809916928,0.2891119421,-0.0434466861,-0.0761660263,0.2290890217,0.1811477989,-0.0985061228,-0.0362648219,0.2386940569,-0.4926404357,-0.0590336435,-0.381783247,0.1135702878,-0.016453132,0.0104641356,-0.073190771,0.183647275,0.5666251183,0.2480074018,0.0704077184,-0.2878100574,0.0038009512,-0.2995419502,-0.0008175711,0.0209132563,-0.2185069621,-0.0127195939,-0.1650024205,0.0993296802,0.4306170046,0.3159378171,-0.1027809456,0.3083853126,-0.3631752133,0.0680128336,0.5011175275,-0.0826230198,-0.0558355302,-0.0620212071,-0.007606979,0.4844593406,-0.0192016512,0.0540582873,-0.0028558637,0.0904539824,0.4210165143,0.2623525858,0.1196406558,-0.0385204628,-0.172674045,0.099473469,0.6978393793,-0.09616445,-0.1197120473,0.3986996412,-0.028801024,0.0267649945,0.1838664412,-0.20131962,0.0032166641,0.0487347618,-0.1022671163,0.1278575659,0.1434339583,0.0885536224,-0.477280885,0.0218084715,-0.1082340553,0.1001599357,-0.201389879,-0.1823511124,0.172307089,-0.0312149823,-0.0837009475,0.2704252601,0.0632626712,0.0763835013,-0.4588813782,-0.0922061875,0.1424329728,0.1000564471,0.1427499056,0.0255112257,-0.4670328796,0.2207724154,0.0516190268,-0.0008800761,-0.3370990753,0.1691160947,0.1627229452,-0.0121848844,-0.1386409104,0.2362546623,-0.240252912,0.0432870612,-0.1662945598,0.0633411631,0.0786366761,0.0362607315,-0.1191926226,0.1988856643,0.1499744952,-0.2119936496,0.296115756,0.469109118,0.2151079774,0.1262290925,-0.3541630805,0.1088888571,0.0366480425,0.0466697365,-0.1513889283,0.0986699015,0.0332320556,0.1303197294,-0.1636215597,0.1962623745,-0.5942215919,0.205019325,0.4457781315,-0.029396804,0.1831894219,-0.0318305567,-0.1152455062,0.1445436776,0.0400426909,-0.4003143907,0.598656714,-0.0762853101,0.160049051,-0.062399704,0.0319034867,-0.0884366781,0.1578813493,0.1634771675,-0.1648807526,0.2422760278,0.5488060713,-0.0258924793,0.0026322759,0.2411987334,-0.0258664507,-0.0695296004,-0.4299575984,0.40895316,-0.4493633211,-0.0130516933,-0.4337830544,-0.4268899262,-0.2325565964,-0.1914565712,-0.3151691854,0.0831534639,-0.0308005959,0.1409175694,0.2315021604,0.208858788,0.4176336229,-0.3352909088,-0.0116184205,0.2565828562,-0.2936905921,-0.0194220729,0.1011185199,-0.474614054,-0.0927021652,-0.0140165407,0.1141560152,-0.2371975183,-0.3689379394,-0.0497705899,-0.2383907735,0.0373716392,0.2184629291,-0.0252511688,-0.0695573762,-0.3585180342,0.0914961845,0.3910309672,0.0402582474,0.5428994298,-0.120610252,-0.1459313184,-0.2737415433,-0.414254427,-0.1589875221,-0.1897293776,0.3760147989,0.0556129292,0.0255450513,0.2554220557,0.1048957482,-0.1140453443,0.3653511405,0.1799469888,-0.2550261319,-0.3012900949,0.5462107658,-0.0352578685,-0.2991886735,-0.1919014752,-0.181500569,0.086480394,0.3125531077,-0.3383074999,0.2758895457,-0.2269130498,0.5231872797,-0.264123112,0.037150573,0.033702679,0.2198939174,0.0525674894,-0.1942944378,0.2089818865,0.2619978786,0.1683866382,0.4185818434,-0.0266035572,-0.0517275445,0.0500230528,0.2800492644,0.0553948171,-0.5754981637,0.0552767143,0.0369585305,0.2148905545,-0.1227348149,-0.5823745728,-0.0003764618,-0.2560166121,0.0677064732,-0.1807641685,-0.2100424021,-0.2295659631,-0.0138648478,0.4533790052,0.0392069519,-0.2866450846,0.1987097114,-0.2816284001,0.1523375362,-0.0028861118,-0.0913680419,-0.2553856373,-0.2479195148,0.1207726598,0.2488175631,-0.1429386437,-0.0366866924,-0.332705766,-0.3845030069,-0.0645513684,0.3166655302,0.2268413305,0.3978821039,0.2228398919,-0.1883240789,-0.1549575329,0.4459351003,0.8358085752,0.1393992305,-0.3315662444,0.3097257316,0.1563317329,-0.3661013544,-0.4271095395,-0.2192660421,0.2107050568,0.2837590575,0.2501065135,-0.2852648199,-0.0509405881,-0.0448640734,0.3663127124,-0.1162713468,0.0555991903,-0.1983711421,-0.0186834671,0.0620130785,-0.2956489027,0.0021542709,0.3204758763,-0.5425037146,-0.4796466529,-0.2594844997,-0.1547110677,0.1353881508,-0.1421689391,0.2919121981,-0.0179595929,0.085913226,-0.2203519046,0.226555407,0.2167867273,0.4099766612,0.2381253988,-0.6073964834,0.0864774585,-0.0816748217,0.1548353285,0.1625726968,-0.3722438812,0.1816199869,-0.2765264809,-0.0311316121,-0.0456786379,0.4009890258,0.5867688656,0.2809723616,-0.5694266558,-0.4078258872,0.2210954875,0.2447007,-0.0408971868,0.3986698091,-0.5758920908,-0.4458216131,0.0151686622,-0.1313024014,0.9348531365,0.227441147,0.3354800045,0.3603384197,0.2283131182,0.2585175335,-0.0876077637,0.2597103715,-0.2174419314,-0.2065098882,-0.1460086405,-0.3474820256,0.0850355774,-0.0712469071,-0.4455394149,-0.083542563,-0.6744431853,0.2634075284,0.0200508833,-0.1967615038,0.211319983,-0.3538209498,-0.0215711687,0.0079714665,-0.104398869,0.1108678281,-0.1389360726,-0.1853519529,-0.3907166123,-0.0414992683,-0.1276089549,0.264203012,-0.1600133628,0.4613226354,0.2610339224,-0.3412076831,0.1806777567,0.3065123558,0.2656790614,0.1745449156,-0.0411951058,-0.0961856693,0.1947005242,0.2346387357,0.2287783772,-0.3867944777,0.382373184,0.3281327486,-0.153571412,0.0813624263,0.1307567507,0.0036409386,-0.2941784561,0.1807542443,0.0310393367,-0.6678335071,0.0257427804,0.1144257635,0.2623029649,0.0078851208,0.0455635339,0.0326087773,-0.1880735308,0.1570516378,-0.0923690647,-0.152303189,-0.0486062318,0.2529429495,0.6211352348,0.1648054272,0.2510537207,0.1555795223,0.0163083728,-0.0108200936,0.3169968426,-0.2276737392,-0.3221340775,0.2020926178,-0.3763022423,-0.163161993,-0.367860049,0.3481138945,-0.2231416255,0.1864342093,-0.1074504107,-0.1279439032,-0.1636229008,-0.0874899924,0.0984858274,0.1391125023,-0.022963658,0.4937678576,0.0096237138,-0.2948608696,-0.1424332559,0.0263493862,-0.0642633513,0.2000518143,-0.1198728234,0.0365071781,0.1525546163,-0.1129384413,0.035251148,0.0947505534,0.1068756357,-0.1395224631,-0.2098544687,0.1725347042,0.1596056372,-0.0051732743,0.0608734861,-0.0223292187,0.0280726943,-0.1015719324,0.0722630173,0.0310427789,-0.0159951691,0.1429526359,-0.1008783579,0.1879900843,0.1577151716,0.0051406827,-0.0138960024,-0.0258001667,0.0863282084,0.2060129046,-0.2112694979,-0.0456977971,-0.4530457556,0.1093631461,0.0555424467,0.2954211831,0.2815393209,-0.4541044831,-0.1130690724,-0.247367546,0.1694397628,0.3495014608,-0.3330458999,-0.1357988864,-0.1574985981,0.0786973462,0.0047278283,-0.2634816766,-0.1379800141,0.1281900555,-0.1968494952,0.1794257611,-0.0343556441,-0.0298032034,-0.09308213,-0.0316716693,0.1060748771,0.0081215315,0.4864069223,0.2356226444,-0.0351790041,0.1831521839,0.2757528126,0.0298430324,-0.0677540079,0.2069119364,0.379934907,0.0728012174,0.475764513,-0.1776733547,0.4583591223,-0.1887035817,-0.0897480473,0.0986289382,-0.0208554938,0.3439779282,0.2496027052,0.1893983036,-0.3167268038,-0.4845971167,-0.0677669346,0.0744091123,-0.3010366559,0.1076430902,0.1744672209,0.0114028892,0.017471971,-0.2452272922,0.0484874994,0.044603277,0.347339958,0.0695022419,-0.1298007071,-0.332385838,0.2785465419,0.2137667537,0.1685858369,-0.0849741623,0.0482949317,0.2870333195,-0.2116377801,0.0310052577,0.328419596,0.4566881359,0.3523456752,-0.0556692705,-0.034788359,0.0796087384,-0.0909845829,0.0379796326,0.2267523557,0.0310630444,-0.2312520891,0.1849548221,0.0787102506,-0.0896127522,0.2247816771,0.0495844595,0.2855033576,-0.5732682943,0.4974719584,-0.3825766444,-0.1653485745,-0.2320639789,-0.0655600354,-0.3300944269,-0.0028859051,0.7613857985,0.1631412208,0.2303946018,-0.045602113,0.0524735153,0.1485673338,0.3819586039,-0.0265991706,0.1260825247,-0.0758734643,0.1160466596,0.1345690638,0.3428659439,0.3775585592,0.2901946008,0.0708572268,0.3413305283,0.0616459809,0.1583906859,-0.02624912,-0.2620299459,0.0305724684,0.1289074868,-0.3855929971,-0.3565680981,-0.0862102658,-0.103219308,0.0023236114,-0.3629136682,0.4041659236,-0.0406728163,0.0443604067,-0.1156619489,-0.2173946798,0.2016637772,0.2793516517,0.3577322662,0.3725616634,0.250300765,-0.1102252305,-0.1550386995,-0.0887542367,-0.5083963871,-0.2221994847,0.4509863555,-0.2360429317,0.1048239097,-0.2071297914,0.2530995607,-0.1397610009,0.1418414414,0.4245605469,-0.1707579643,-0.398073554,0.299046874,-0.3292228878,-0.0033926696,-0.0399658829,0.5031766295,-0.0017838628,0.2349109203,-0.0036315413,-0.4555346072,0.4450863898,-0.3086916506,-0.4081063569,0.0040454934,-0.0593627244,-0.0185005981,0.1491985172,-0.5609920621,-0.3100958169,0.4338613451,-0.3143926561,-0.0531018376,0.0299371406,-0.0753480196,0.0506678186,-0.2577311993,0.2257412672,-0.0080876872,-0.2879448831,0.4016746879,-0.2544682026]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2757","title":"Unexpected type after `concatenate_datasets`","comments":"Thanks @albertvillanova it indeed\u00a0did the job \ud83d\ude03 \r\nThanks for your answer!","body":"## Describe the bug\r\nI am trying to concatenate two `Dataset` using `concatenate_datasets` but it turns out that after concatenation the features are casted from `torch.Tensor` to `list`. \r\nIt then leads to a weird tensors when trying to convert it to a `DataLoader`. However, if I use each `Dataset` separately everything behave as expected.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> featurized_teacher\r\nDataset({\r\n features: ['t_labels', 't_input_ids', 't_token_type_ids', 't_attention_mask'],\r\n num_rows: 502\r\n})\r\n>>> for f in featurized_teacher.features:\r\n print(featurized_teacher[f].shape)\r\ntorch.Size([502])\r\ntorch.Size([502, 300])\r\ntorch.Size([502, 300])\r\ntorch.Size([502, 300])\r\n\r\n>>> featurized_student\r\nDataset({\r\n features: ['s_features', 's_labels'],\r\n num_rows: 502\r\n})\r\n>>> for f in featurized_student.features:\r\n print(featurized_student[f].shape)\r\ntorch.Size([502, 64])\r\ntorch.Size([502])\r\n```\r\nThe shapes seem alright to me. Then the results after concatenation are as follow:\r\n```python\r\n>>> concat_dataset = datasets.concatenate_datasets([featurized_student, featurized_teacher], axis=1)\r\n>>> type(concat_dataset[\"t_labels\"])\r\n\r\n```\r\nOne would expect to obtain the same type as the one before concatenation.\r\n\r\nAm I doing something wrong here? Any idea on how to fix this unexpected behavior?\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-10.14.6-x86_64-i386-64bit\r\n- Python version: 3.9.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":12,"text":"Unexpected type after `concatenate_datasets`\n## Describe the bug\r\nI am trying to concatenate two `Dataset` using `concatenate_datasets` but it turns out that after concatenation the features are casted from `torch.Tensor` to `list`. \r\nIt then leads to a weird tensors when trying to convert it to a `DataLoader`. However, if I use each `Dataset` separately everything behave as expected.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> featurized_teacher\r\nDataset({\r\n features: ['t_labels', 't_input_ids', 't_token_type_ids', 't_attention_mask'],\r\n num_rows: 502\r\n})\r\n>>> for f in featurized_teacher.features:\r\n print(featurized_teacher[f].shape)\r\ntorch.Size([502])\r\ntorch.Size([502, 300])\r\ntorch.Size([502, 300])\r\ntorch.Size([502, 300])\r\n\r\n>>> featurized_student\r\nDataset({\r\n features: ['s_features', 's_labels'],\r\n num_rows: 502\r\n})\r\n>>> for f in featurized_student.features:\r\n print(featurized_student[f].shape)\r\ntorch.Size([502, 64])\r\ntorch.Size([502])\r\n```\r\nThe shapes seem alright to me. Then the results after concatenation are as follow:\r\n```python\r\n>>> concat_dataset = datasets.concatenate_datasets([featurized_student, featurized_teacher], axis=1)\r\n>>> type(concat_dataset[\"t_labels\"])\r\n\r\n```\r\nOne would expect to obtain the same type as the one before concatenation.\r\n\r\nAm I doing something wrong here? Any idea on how to fix this unexpected behavior?\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-10.14.6-x86_64-i386-64bit\r\n- Python version: 3.9.5\r\n- PyArrow version: 3.0.0\r\n\nThanks @albertvillanova it indeed\u00a0did the job \ud83d\ude03 \r\nThanks for your answer!","embeddings":[0.0659190714,-0.2437318563,0.0309370346,0.4964453578,0.4149608314,0.1757521927,0.4837472141,0.1384313703,-0.3720063865,-0.134859845,-0.156314671,0.3944970071,-0.0366082974,-0.0296746287,-0.0745012835,-0.2937272787,0.294506073,-0.0014214056,-0.2037576437,-0.2272137105,-0.0386763476,0.1527753621,-0.2757367492,0.0347044952,-0.2138401121,0.0433406271,-0.0622530654,-0.1045721769,0.0157314241,0.0415251255,0.4239585698,-0.0021293538,0.2352160662,0.5650517344,-0.0001220819,0.0738590807,0.1022619978,0.0249297973,-0.0154521186,-0.0110048624,-0.2690013647,-0.3836075366,0.1033526734,-0.2186095864,0.0939235315,-0.2825351059,-0.2790849209,-0.4440074563,-0.097580947,0.1644082814,0.1471364498,0.2375485301,0.0298640952,0.1196675524,0.1865439117,0.165979296,-0.1456199437,-0.1074627936,-0.0598189496,0.2500637174,0.4331701398,-0.0468632802,-0.1849826723,-0.1953752786,0.0277906153,0.2591166496,-0.3254513741,-0.3925424814,-0.1725143939,0.1616290957,0.5209312439,-0.1575306058,-0.1835737377,-0.1490740031,0.0178554766,0.1022604108,0.1570604295,0.1806984991,0.1621251106,-0.0688207597,-0.0001436032,0.2619385123,-0.0541171767,0.2739671469,-0.2616983056,-0.0057369536,-0.1143608689,0.2706641257,-0.0917317346,-0.1695450544,0.1946017742,-0.2625325918,-0.0670424402,-0.0137376338,-0.5007714629,-0.0804302171,0.0671504065,-0.5047886968,-0.1523769647,-0.138415724,0.1303620636,0.053540308,-0.2088373154,0.125428617,0.206956163,0.3058359623,-0.0745112747,0.3823685348,-0.0955081955,-0.0104729887,-0.0577265583,0.1088388637,-0.0284929797,-0.2990461588,0.0802759677,0.0229053553,0.3922742009,-0.0095209852,-0.3741963506,0.021246478,-0.3829902709,0.2581012845,0.0750822723,-0.152745381,-0.3886459768,0.5850807428,0.2911901176,0.2835937738,0.0566592142,-0.0235448405,-0.0580969006,-0.2789310217,-0.081606172,-0.2414485067,0.0405072644,-0.1405511349,-0.2217605412,0.4507351518,0.1302915961,0.012507285,-0.0606594197,-0.0752940103,0.2919274271,-0.0306541231,-0.0690622926,0.2359744459,0.1845577061,-0.1006291434,-0.0358700901,0.239116326,-0.4935937524,-0.0602483973,-0.372064352,0.1110176072,-0.0262362845,0.0089457966,-0.0683113933,0.1796612144,0.5736274123,0.2451362312,0.0779191926,-0.2917569578,0.0116681112,-0.2986975014,0.0047487598,0.0214220695,-0.230518654,-0.0118534919,-0.1720362604,0.0998854116,0.4243899882,0.3167899847,-0.0998590663,0.3156577349,-0.3608244658,0.0701709688,0.5017547011,-0.0730383843,-0.0577745214,-0.0602548271,-0.0123138046,0.4981867075,-0.0248697195,0.0516324528,-0.0103875948,0.0878681317,0.4176945984,0.2619722784,0.1075595766,-0.0340657234,-0.1704995036,0.1092750058,0.7022425532,-0.0907333344,-0.1126430929,0.3934443891,-0.0338219106,0.0266338941,0.1877878457,-0.211167261,0.0082701612,0.0489922501,-0.1141024008,0.1250620037,0.1506378353,0.0856390446,-0.4757116437,0.0223593749,-0.1108546034,0.1066574231,-0.2000025809,-0.1850707978,0.1749994308,-0.0273593664,-0.0850962773,0.2619085312,0.0581697561,0.0630431026,-0.4643322825,-0.0952296406,0.1446460485,0.0798946843,0.1525879055,0.0302381311,-0.4583376348,0.2111876607,0.0490964241,-0.0031566585,-0.3407337964,0.1717142314,0.152107805,-0.0120640518,-0.1395008862,0.2329477519,-0.2427451611,0.0440465063,-0.180571124,0.0639604852,0.0783368349,0.0389712602,-0.1176954135,0.1948289275,0.1455775499,-0.216666624,0.2979933619,0.4656653106,0.2053004056,0.1310785711,-0.3571459651,0.1087006703,0.0358278602,0.0492625572,-0.1511350721,0.106050387,0.0382558815,0.1197566241,-0.1622099876,0.197761029,-0.6009560823,0.2125840634,0.4330871403,-0.0248388927,0.1795607209,-0.0372989774,-0.1181646138,0.1389309019,0.0444847085,-0.3980384171,0.6007566452,-0.079544507,0.1579300314,-0.0601493493,0.0253960099,-0.095920071,0.1579210907,0.1628264338,-0.1676989794,0.2511103451,0.5483640432,-0.0159940906,-0.0050752526,0.253772527,-0.0185723733,-0.0761334673,-0.4396650791,0.4208841622,-0.4437838197,-0.0174180251,-0.435288161,-0.4252430797,-0.2362214625,-0.1906720996,-0.3029824197,0.0834383741,-0.0448433384,0.1447667927,0.2258214206,0.2112877965,0.4156981409,-0.3432815671,-0.0162838902,0.2587583661,-0.2815701962,-0.0196906514,0.100178957,-0.4739817679,-0.090527378,-0.0094651924,0.1142466068,-0.2354081869,-0.3638991416,-0.0486782901,-0.2494111061,0.0372026302,0.2167593688,-0.0339495055,-0.0723435432,-0.3606536388,0.0904072523,0.3892674744,0.0397109874,0.545830667,-0.1192243472,-0.1480497271,-0.2790913284,-0.4045184255,-0.1491466165,-0.1835487038,0.3781381249,0.0575050823,0.0229560286,0.2577911317,0.1071214601,-0.1149613187,0.3671506643,0.1796973497,-0.2514235675,-0.294097513,0.5419860482,-0.0356419906,-0.3014376462,-0.1992533207,-0.1817018688,0.0905760899,0.3263100684,-0.3364426196,0.2817229629,-0.2243236005,0.5209254622,-0.2591611743,0.0303355046,0.0375356935,0.2177197784,0.0573520102,-0.1959762573,0.2094525397,0.2639806271,0.1689645648,0.4193356931,-0.0397721604,-0.0549184121,0.0481513701,0.3025141656,0.0352141261,-0.5685999393,0.0553254969,0.0350533091,0.2173440307,-0.1228222549,-0.5846559405,0.0082154451,-0.2600182295,0.0603483543,-0.1800439507,-0.2105576694,-0.2356047183,-0.0061458158,0.457937032,0.0466430709,-0.2883174419,0.2086817622,-0.2883477211,0.1490685642,0.0081363814,-0.0877592489,-0.2503910065,-0.2519666553,0.1200975403,0.2320447266,-0.1411392838,-0.0414165482,-0.3184079528,-0.3972735703,-0.059951406,0.305208534,0.2257097065,0.3916657865,0.2215428799,-0.1990561932,-0.1490066051,0.439681977,0.8412864804,0.1467217803,-0.3226118982,0.3043781221,0.156125173,-0.3746886849,-0.4323106408,-0.2151249051,0.2049172819,0.2872171104,0.248815015,-0.2804675698,-0.0433344431,-0.0354327895,0.3723092675,-0.1194225401,0.0645473748,-0.2004886717,-0.0256132241,0.0667794496,-0.2828763425,0.0018850242,0.312446028,-0.5491112471,-0.4713169932,-0.2564314008,-0.140782848,0.1309546381,-0.1431929916,0.2775426805,-0.0185364261,0.0927215889,-0.2249986678,0.2306561619,0.209965542,0.4140523076,0.2464790791,-0.6019338965,0.0827508196,-0.0783830211,0.1527935565,0.1694580615,-0.3691346347,0.173313424,-0.2745176256,-0.042359326,-0.0451426059,0.4040772617,0.5784541965,0.2799608707,-0.5684007406,-0.4052292407,0.2184126228,0.2392736673,-0.0410096645,0.3936131299,-0.5667901039,-0.4457779825,0.0235239118,-0.1242587268,0.9340797663,0.2308175862,0.3456716537,0.3732998371,0.2399658561,0.2560908794,-0.1033544689,0.2644013166,-0.2245924771,-0.221119687,-0.1466373801,-0.3485500813,0.0739446357,-0.0784315541,-0.4365439117,-0.0880262703,-0.6684350371,0.2682411075,0.0261553526,-0.2080345452,0.2103532553,-0.3531008661,-0.0106017282,0.0054257945,-0.1033933684,0.1108468845,-0.137214154,-0.1818611026,-0.3963865042,-0.0454925485,-0.1307304054,0.2725886106,-0.1515927613,0.4549675286,0.2611647546,-0.349208504,0.184138447,0.3059735596,0.2685470581,0.1701468825,-0.0446689464,-0.0843160823,0.2016795427,0.2381330878,0.2372993827,-0.3901323378,0.3904404044,0.3274179399,-0.1414780915,0.0788508132,0.1258701235,0.0087321913,-0.2953163683,0.1780688465,0.0203251522,-0.6751877666,0.0298371017,0.1152692437,0.2685189843,0.0004568308,0.0421645306,0.0207220949,-0.178845644,0.1551955044,-0.091522485,-0.1505638808,-0.0534625612,0.2534361482,0.6186258197,0.1669643372,0.2536993027,0.1569332778,0.0159194898,-0.0125537543,0.3054921329,-0.2259809524,-0.3427065611,0.2033070326,-0.3914577067,-0.1684547812,-0.3752190769,0.3356194496,-0.2183358967,0.1896243691,-0.1109470353,-0.1149859428,-0.1622962207,-0.08817783,0.1065446213,0.1316623539,-0.0279081333,0.5006034374,0.0165001918,-0.2925832868,-0.1371895671,0.0226956382,-0.0588526502,0.2001273036,-0.1181067154,0.0401985273,0.1519039422,-0.1143579856,0.0339846462,0.0885577872,0.1149579138,-0.1365960091,-0.2134457231,0.1733842343,0.1591496319,-0.0043457202,0.0649097562,-0.018019069,0.0279821642,-0.1068837568,0.0811678842,0.0237983055,-0.014589699,0.1377933919,-0.094016701,0.1933321357,0.1635741293,0.0025728878,-0.0201155655,-0.0203013811,0.084204793,0.2068062574,-0.2102957368,-0.040840745,-0.4518904984,0.1150537208,0.054140497,0.294146657,0.2826563418,-0.4562207162,-0.1160700172,-0.2541514635,0.1796671003,0.3579172194,-0.3368768692,-0.1357103735,-0.1595805734,0.0752875581,0.0078126816,-0.2595174611,-0.1360135525,0.1268317997,-0.2025453448,0.1767612696,-0.0388191342,-0.0413513035,-0.0977966785,-0.0277735088,0.1099702716,0.0083550438,0.4828452766,0.2385091335,-0.0351808853,0.1886693388,0.2742007971,0.021020487,-0.0748418719,0.2056278437,0.3777002692,0.0635857135,0.4782323837,-0.1842932701,0.4554717839,-0.1834283173,-0.0968687683,0.0944720358,-0.0329707153,0.337955296,0.2592892647,0.1908072233,-0.3133335412,-0.4887455702,-0.0720971823,0.08803913,-0.3046379089,0.1119706929,0.1788182557,0.0093932543,0.0222913325,-0.2360963374,0.0427698679,0.0485089533,0.3496249616,0.0633888766,-0.1314380914,-0.3350439966,0.2885661125,0.2040568143,0.1650207043,-0.0869811401,0.0438410714,0.2820618153,-0.2192261517,0.0340555198,0.3236372471,0.4572915435,0.351919651,-0.0521783791,-0.0359972306,0.0710648745,-0.0910080001,0.0353325494,0.2338298708,0.0319620259,-0.2241384238,0.1931258589,0.0757309422,-0.0881401002,0.2216559649,0.0551672019,0.2863171399,-0.5725302696,0.5041823983,-0.3828968108,-0.1731821448,-0.2240831703,-0.0637259111,-0.3255920112,0.0000336106,0.7593532205,0.1710291505,0.2244994938,-0.0497369915,0.0510604344,0.1556119472,0.3958733678,-0.0257883631,0.1295874864,-0.0745333508,0.1228386387,0.1331328303,0.3439053595,0.380109787,0.2921630442,0.0695113391,0.342217803,0.0604491532,0.1620337814,-0.0300504155,-0.2565834224,0.0331310295,0.128100872,-0.3727923334,-0.3519048393,-0.0976075679,-0.0946318358,-0.0083656078,-0.3612941504,0.4089610875,-0.0399104841,0.0433909707,-0.1115440279,-0.2311214656,0.2101143748,0.2832202911,0.3579321504,0.3602249026,0.2596922517,-0.1079410613,-0.1546716541,-0.1005014852,-0.5149802566,-0.23291035,0.4482857287,-0.2343309671,0.0961690694,-0.2016962618,0.253488481,-0.1363619864,0.1435457915,0.4287969172,-0.1598206609,-0.3926214278,0.3061795831,-0.3349909782,0.0059345146,-0.0366111696,0.4954218864,-0.0023916978,0.2370787114,-0.0047936337,-0.4628365934,0.4442895949,-0.3086186647,-0.4006411731,0.0021128498,-0.05872925,-0.0249924343,0.1366714686,-0.5604068041,-0.3176694214,0.4305161536,-0.3152371943,-0.0597375929,0.0323444419,-0.080996044,0.0461709425,-0.2604374886,0.2283602804,-0.0095951259,-0.2973435521,0.4006498456,-0.2546716332]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2750","title":"Second concatenation of datasets produces errors","comments":"@albertvillanova ","body":"Hi,\r\n\r\nI am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`.\r\n\r\n```\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\ndata = load_dataset('trec')['train']\r\nconcatenated = concatenate_datasets([data, data])\r\nconcatenated_2 = concatenate_datasets([concatenated, concatenated])\r\nprint('True features of features:', concatenated.features)\r\nprint('\\nProduced features of features:', concatenated_2.features)\r\n```\r\noutputs \r\n\r\n```\r\nTrue features of features: {'label-coarse': ClassLabel(num_classes=6, names=['DESC', 'ENTY', 'ABBR', 'HUM', 'NUM', 'LOC'], names_file=None, id=None), 'label-fine': ClassLabel(num_classes=47, names=['manner', 'cremat', 'animal', 'exp', 'ind', 'gr', 'title', 'def', 'date', 'reason', 'event', 'state', 'desc', 'count', 'other', 'letter', 'religion', 'food', 'country', 'color', 'termeq', 'city', 'body', 'dismed', 'mount', 'money', 'product', 'period', 'substance', 'sport', 'plant', 'techmeth', 'volsize', 'instru', 'abb', 'speed', 'word', 'lang', 'perc', 'code', 'dist', 'temp', 'symbol', 'ord', 'veh', 'weight', 'currency'], names_file=None, id=None), 'text': Value(dtype='string', id=None)}\r\n\r\nProduced features of features: {'label-coarse': Value(dtype='int64', id=None), 'label-fine': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}\r\n```\r\n\r\nI am using `datasets` v.1.11.0","comment_length":1,"text":"Second concatenation of datasets produces errors\nHi,\r\n\r\nI am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`.\r\n\r\n```\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\ndata = load_dataset('trec')['train']\r\nconcatenated = concatenate_datasets([data, data])\r\nconcatenated_2 = concatenate_datasets([concatenated, concatenated])\r\nprint('True features of features:', concatenated.features)\r\nprint('\\nProduced features of features:', concatenated_2.features)\r\n```\r\noutputs \r\n\r\n```\r\nTrue features of features: {'label-coarse': ClassLabel(num_classes=6, names=['DESC', 'ENTY', 'ABBR', 'HUM', 'NUM', 'LOC'], names_file=None, id=None), 'label-fine': ClassLabel(num_classes=47, names=['manner', 'cremat', 'animal', 'exp', 'ind', 'gr', 'title', 'def', 'date', 'reason', 'event', 'state', 'desc', 'count', 'other', 'letter', 'religion', 'food', 'country', 'color', 'termeq', 'city', 'body', 'dismed', 'mount', 'money', 'product', 'period', 'substance', 'sport', 'plant', 'techmeth', 'volsize', 'instru', 'abb', 'speed', 'word', 'lang', 'perc', 'code', 'dist', 'temp', 'symbol', 'ord', 'veh', 'weight', 'currency'], names_file=None, id=None), 'text': Value(dtype='string', id=None)}\r\n\r\nProduced features of features: {'label-coarse': Value(dtype='int64', id=None), 'label-fine': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}\r\n```\r\n\r\nI am using `datasets` v.1.11.0\n@albertvillanova ","embeddings":[-0.1504657269,-0.0591697134,-0.0577093922,0.2559515834,0.193053782,0.0901743174,0.0821338445,0.2622757852,-0.2609196305,-0.0551395863,0.0358312577,0.1265247911,0.2174618989,0.090199098,-0.2850414217,-0.1801928431,0.1788396835,0.0559669882,0.0485580564,-0.118900381,-0.1244952604,0.1265652031,-0.3151787817,0.0769735575,-0.4671195447,0.1625408679,-0.3227909505,0.007621164,-0.0032958225,-0.3098783493,-0.0556449778,0.3040577769,-0.0451283157,0.32534796,-0.0001139846,-0.1404050738,0.020385908,-0.0589029714,-0.1106781885,-0.1804829538,-0.1020738631,-0.5905038118,-0.0913802311,-0.2002821118,0.1462708861,-0.6070088148,-0.2705745995,-0.2687054873,0.2620135248,-0.1385006309,0.159783259,-0.0676152855,0.1747838706,-0.1806190014,-0.1066621318,0.0965181813,0.0347622149,-0.0004105456,-0.1437833905,-0.0325336047,0.4414526224,0.3706357181,-0.1782261729,-0.3017144203,-0.1555945128,0.1617710441,0.0063124946,-0.3817135096,-0.0536329299,0.1203863546,0.3290899098,-0.3443950415,-0.3714585006,-0.4346679449,0.1693651974,-0.2393826842,0.03127525,-0.0691222697,0.2867306471,0.2179713845,-0.0308873206,-0.189049691,0.0395146236,0.1386504024,-0.0608814657,0.2488592714,0.0057220296,0.1144914106,0.1244022697,-0.5368865728,0.0223361254,-0.4283556044,-0.0197522454,0.0175422635,-0.5440704823,-0.0473727919,0.0594298653,-0.6148066521,0.2270136178,0.1771862656,-0.0383978337,0.0237068944,-0.0668149516,-0.0024822024,0.2470452338,0.4188669622,0.2312257588,0.2866076529,-0.0470639057,-0.225259766,-0.2453390211,-0.0734846815,-0.1264434159,-0.2911265194,0.1000443324,0.0113737918,0.396143049,-0.0294776019,-0.0793904811,0.11111404,-0.3544274569,-0.1133779958,0.1028654203,0.0086840028,-0.0921811163,0.6188397408,-0.1172470599,0.2566657662,0.1120014265,-0.1708726287,-0.1573576778,-0.1931791455,0.0105560739,0.1908484846,-0.009235274,-0.2236082852,0.0366123281,0.2651329041,-0.0237969682,-0.1290379465,0.0272331573,-0.3101315796,0.0045604357,0.0218815319,0.1250576526,0.0520775132,0.0854914114,-0.0011943799,0.024049921,0.0454844385,-0.5595525503,-0.1502062678,-0.3701137304,0.2136218846,0.0852637514,0.2996404767,0.0801235214,0.302719593,0.2509583235,0.0506003797,-0.1885678768,-0.0870354995,0.0076499549,-0.1807177514,0.082131654,0.0898533612,0.0677425563,0.0127707971,0.1103849784,0.1843047738,0.2871360183,0.4206843376,-0.1703023314,0.3344914913,-0.2040976286,0.4971638322,0.5261321664,-0.0743064508,-0.1810768843,0.2042457908,-0.0661210865,0.3550088406,0.1406594813,-0.1781974137,0.0503419302,0.0754448324,0.4194709361,0.123826772,0.0369044282,-0.3195022941,-0.0728152394,0.0247223936,0.7286319733,-0.3416662514,-0.2831137478,0.2036883235,-0.0616782382,-0.0981448218,0.1693049967,-0.3445923924,-0.05869589,0.1913804859,-0.125977084,0.1830993742,0.0787284821,0.0552581102,-0.3375196755,0.0893760026,-0.0399330445,-0.2428482473,-0.167930007,-0.2065499425,0.0221018586,-0.0164528918,-0.1647251695,0.2443644553,0.1765640825,0.3324424326,-0.1891664118,-0.0546176396,-0.0573655665,0.4379684627,0.0619030967,0.0928421468,-0.1567480415,0.3573679924,0.0526526645,0.0333255939,-0.0547466129,0.1343888491,0.5898091793,-0.1381362081,-0.0987757519,0.4499067068,-0.1939736903,-0.1137822419,-0.0809097514,-0.0667266548,0.10392946,0.0040466455,0.0680115074,-0.1234541014,-0.0436430983,-0.1680187434,0.0705930367,0.2275555879,0.2849643826,0.1425679475,0.0359224901,-0.0949881747,0.1146106496,-0.1052120999,-0.1035783067,-0.4422333241,0.2089213729,0.0312789194,-0.2695364952,0.0007309509,-0.4664670825,0.2603234649,0.3221491277,0.1330774277,0.0780856013,0.0291613564,0.0404994898,0.0632191226,0.0192058738,-0.2066784352,0.4321274161,0.1638639122,-0.003215817,0.1260049194,-0.094315432,0.0408444293,0.182420522,-0.0683083534,0.0810810626,0.4733741879,0.2709379196,0.0947590172,-0.0473091118,-0.1575174481,0.0607751161,-0.1024566963,-0.1658180505,0.128069371,-0.6087813377,-0.1715297848,-0.2396755964,-0.5274429917,-0.2216403484,-0.4587311447,-0.1826675832,0.0086168805,-0.1529063731,0.172349304,0.1213834211,-0.1309144199,0.1670149118,0.0995271653,0.1163859367,0.1581566781,-0.0811959282,-0.0527282991,-0.1512535363,-0.0578469522,0.0354970507,-0.0889788717,-0.0157376137,-0.3222573996,-0.2392690182,0.1334006041,-0.2884916961,-0.1630610228,-0.0221155547,-0.0977419615,0.04354183,-0.4574145079,0.3913764656,0.5567986965,0.0928269625,0.2175679207,0.3165223002,-0.3759053051,-0.2535746992,-0.3681758046,-0.2974682748,-0.0178930797,0.1713598073,-0.256580472,0.1625857204,0.5247507095,0.0695422068,-0.2070610821,0.2132777125,0.2021140903,-0.1964456588,-0.1422774345,0.351760447,-0.2002482265,0.0002508448,-0.2620691359,-0.0276876092,-0.1174924597,0.2562609017,-0.2546110153,-0.0034958834,0.029826317,0.0769765601,-0.3703232408,0.0944289044,0.0625614673,0.3755552471,-0.1497161239,0.0357851461,-0.1428721547,0.0463107303,0.3056418896,0.4419403374,-0.182499513,-0.0159339756,-0.1022930294,-0.0357602537,0.2336682528,-0.0100663975,0.2705745697,-0.0379700698,0.2105386555,-0.0199496839,-0.6065669656,-0.2009226382,-0.2038234323,0.1288241297,-0.0388549417,-0.1556237787,0.2491723299,0.0688511431,0.1483340561,0.0078343702,-0.1901330203,-0.0095134685,-0.1038508192,0.4142340422,-0.0300220437,0.005481204,-0.2045525163,0.2662065029,0.0557391569,-0.0655134395,0.1078738794,-0.2636257708,-0.5134298801,-0.2311079204,-0.3840075135,0.3098147213,0.1988602579,0.370665431,0.1436106563,0.0540110432,-0.2669863403,0.1135599613,0.7118365169,-0.1150130257,-0.385193944,0.4801951647,0.0020542217,-0.076846607,-0.3024656773,-0.0549161136,0.2780228853,0.1245328113,0.5626254082,-0.1685641408,-0.1218298748,0.1117850393,0.6294497848,-0.1057713926,-0.1756478846,-0.112501964,-0.0003586744,-0.3241675198,-0.51160115,-0.2100287527,0.4020390213,-0.3005683124,-0.1428620964,-0.2321521044,-0.3157801628,0.3009279072,0.2026450038,0.466409564,0.1533275843,0.2751990557,-0.0724410638,0.2755639553,0.3012104332,0.6378158927,0.1912370026,-0.7376356125,-0.1446775496,-0.0231771059,0.3582291007,0.0496122055,-0.2540135682,0.4031372368,-0.4295672476,0.3281929791,-0.1531596333,0.1966497302,0.5844236016,0.1333208829,-0.5558376312,-0.1940577179,0.165182367,0.1863946617,-0.0991586596,0.2788518667,-0.4781163037,-0.384187907,0.0734137297,-0.1341848373,0.8999783993,0.2740628421,-0.1333556175,-0.0925272033,0.0792021602,0.2988036871,-0.2731703818,0.1949479282,-0.0499312393,-0.2394651622,-0.0514806062,0.0525455475,0.1468616128,0.0526653193,-0.7124878764,0.3619901538,-0.4484539628,0.0995696783,-0.2969274521,0.2878984809,0.3922804892,-0.3144977689,-0.2345916331,0.0546487682,0.046234075,-0.1462475955,-0.0337212905,-0.0758748576,0.2458808869,-0.2128926218,-0.1421933025,0.2426852882,-0.3409103751,0.4390045404,-0.0448306575,-0.2151910663,0.1573583931,0.0909858793,0.3253347576,0.2573347986,0.0987038314,-0.2113574743,0.0851860791,0.0636720881,-0.3257013559,-0.3507932425,0.1022851691,0.1928027868,-0.3504643738,0.1404039115,0.2587255538,0.0485343188,-0.2336755842,0.0428020135,0.2895253897,-0.5032480359,0.2250385284,0.3453092873,0.111779809,-0.1729832143,0.1053814292,0.1110715345,-0.0358319208,0.0005476316,-0.1623958945,-0.1433448941,-0.0809889063,0.2497196048,0.3761561513,-0.0039944556,0.2022247612,-0.0740246847,-0.0276717376,-0.1196447387,-0.060740862,-0.4044419825,0.0639665425,0.0911104605,-0.0903158113,0.0084972512,-0.2556840777,0.2764193416,0.0067527718,0.2241262347,-0.0271066073,-0.2180060446,-0.1631318778,0.0002046105,0.1853839755,0.1254906803,-0.1021481752,0.4127794206,-0.1707359403,0.338138938,-0.2539671659,-0.088140294,-0.1201729178,0.1959517151,-0.1684363633,-0.0022603103,-0.1586277634,-0.0449553803,0.1960983127,0.219684884,0.0685281083,-0.1981114149,-0.1344372034,0.1306734681,0.1136477888,-0.1090045646,-0.1214502305,-0.1655697376,0.2467256337,-0.3233593404,0.108999148,0.1586366594,0.1293053627,0.2555963993,0.0440423004,-0.03267904,0.0062531224,0.0737168118,0.0863091946,0.2640117407,0.0799779817,0.1482904255,-0.2073283792,-0.0573069565,-0.629260838,-0.1103538871,0.0366320759,-0.1719390899,0.1182309017,-0.1929187179,-0.1658097059,-0.059261661,0.2319018394,0.6020119786,-0.398537457,-0.0572810806,0.144917056,0.1897978336,-0.2437744141,-0.3710727394,0.0801532641,0.2015955448,-0.1089889631,0.257812351,0.2650195956,0.0011206222,0.0896895975,-0.0691619813,0.0189596098,0.2882750034,0.4360206127,0.2229431868,0.0661664829,0.0573681854,0.0454018898,0.1606493294,0.1576923132,0.1019946709,0.029053824,0.1478938162,0.3706709445,0.053790722,0.0660661161,-0.1295982599,0.0943690389,0.2220945805,-0.3501650095,0.1290225536,-0.1412718296,0.0181891341,-0.1675129086,-0.4559591711,-0.0761231035,0.5514985919,-0.0913946405,-0.2190891057,0.3085815609,-0.0410497412,-0.1610819846,-0.1548285931,0.2097725868,-0.0945969373,0.2522943914,-0.146456182,0.0588625632,-0.2706637979,0.084284611,0.3438919187,0.3819179833,-0.3340295255,0.2084652036,0.5395169258,0.0179144628,-0.0434292704,0.3910596371,0.6031447649,0.2915300727,0.0115458295,0.0487371832,-0.1092322096,-0.0658707321,0.1837105006,-0.0105245728,-0.1284661442,-0.3860157728,0.3023032248,0.1048435122,-0.1672857404,0.1264677048,0.2400977015,0.0715405345,-0.4240857661,0.2296592891,0.2455328405,-0.0862919539,-0.0205462538,-0.1232278645,-0.3437090218,-0.3410773873,0.7281009555,0.1575152576,0.1374147385,0.0866666958,0.1286887079,-0.0492461026,0.5275999308,0.2656449676,0.0637388453,-0.2873804569,-0.2236574143,-0.1651127785,0.0836650729,-0.1968185157,0.0170675218,0.2934058607,0.1322610676,-0.0237307828,0.2036402673,0.0310453307,-0.1009669825,-0.0647990257,-0.1896103621,-0.5950562358,-0.1486581415,0.0926938057,-0.0080743739,0.0550474077,-0.0918947682,0.4462240636,-0.1644046754,0.0500477813,-0.0800517946,0.1472428292,0.1145500168,0.1729750782,0.6292982101,0.2540768981,0.3092817664,-0.1138431132,0.1047936156,0.0906459317,-0.3408109248,0.0195527766,0.2614530325,0.3153954744,-0.0783053488,-0.2710303366,0.204288125,-0.1040892825,0.1056802645,0.4074398279,-0.0661337674,-0.1134404391,0.1953424811,0.211631611,0.0576640815,-0.0634947717,0.3801524341,-0.0141822305,0.5103456378,0.1568274051,-0.5181285143,0.4936011136,0.0660220161,-0.4652317762,-0.1163864583,0.134999454,0.1205678955,0.1267579198,-0.6108980775,-0.2417511493,0.5591214895,-0.0480036363,-0.3332654238,0.1799198836,0.0029888649,-0.1483451575,0.0958362445,0.2016952634,-0.0232245196,-0.3746020198,0.1889448166,-0.0530792288]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2750","title":"Second concatenation of datasets produces errors","comments":"Hi @Aktsvigun, thanks for reporting.\r\n\r\nI'm investigating this.","body":"Hi,\r\n\r\nI am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`.\r\n\r\n```\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\ndata = load_dataset('trec')['train']\r\nconcatenated = concatenate_datasets([data, data])\r\nconcatenated_2 = concatenate_datasets([concatenated, concatenated])\r\nprint('True features of features:', concatenated.features)\r\nprint('\\nProduced features of features:', concatenated_2.features)\r\n```\r\noutputs \r\n\r\n```\r\nTrue features of features: {'label-coarse': ClassLabel(num_classes=6, names=['DESC', 'ENTY', 'ABBR', 'HUM', 'NUM', 'LOC'], names_file=None, id=None), 'label-fine': ClassLabel(num_classes=47, names=['manner', 'cremat', 'animal', 'exp', 'ind', 'gr', 'title', 'def', 'date', 'reason', 'event', 'state', 'desc', 'count', 'other', 'letter', 'religion', 'food', 'country', 'color', 'termeq', 'city', 'body', 'dismed', 'mount', 'money', 'product', 'period', 'substance', 'sport', 'plant', 'techmeth', 'volsize', 'instru', 'abb', 'speed', 'word', 'lang', 'perc', 'code', 'dist', 'temp', 'symbol', 'ord', 'veh', 'weight', 'currency'], names_file=None, id=None), 'text': Value(dtype='string', id=None)}\r\n\r\nProduced features of features: {'label-coarse': Value(dtype='int64', id=None), 'label-fine': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}\r\n```\r\n\r\nI am using `datasets` v.1.11.0","comment_length":8,"text":"Second concatenation of datasets produces errors\nHi,\r\n\r\nI am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`.\r\n\r\n```\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\ndata = load_dataset('trec')['train']\r\nconcatenated = concatenate_datasets([data, data])\r\nconcatenated_2 = concatenate_datasets([concatenated, concatenated])\r\nprint('True features of features:', concatenated.features)\r\nprint('\\nProduced features of features:', concatenated_2.features)\r\n```\r\noutputs \r\n\r\n```\r\nTrue features of features: {'label-coarse': ClassLabel(num_classes=6, names=['DESC', 'ENTY', 'ABBR', 'HUM', 'NUM', 'LOC'], names_file=None, id=None), 'label-fine': ClassLabel(num_classes=47, names=['manner', 'cremat', 'animal', 'exp', 'ind', 'gr', 'title', 'def', 'date', 'reason', 'event', 'state', 'desc', 'count', 'other', 'letter', 'religion', 'food', 'country', 'color', 'termeq', 'city', 'body', 'dismed', 'mount', 'money', 'product', 'period', 'substance', 'sport', 'plant', 'techmeth', 'volsize', 'instru', 'abb', 'speed', 'word', 'lang', 'perc', 'code', 'dist', 'temp', 'symbol', 'ord', 'veh', 'weight', 'currency'], names_file=None, id=None), 'text': Value(dtype='string', id=None)}\r\n\r\nProduced features of features: {'label-coarse': Value(dtype='int64', id=None), 'label-fine': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}\r\n```\r\n\r\nI am using `datasets` v.1.11.0\nHi @Aktsvigun, thanks for reporting.\r\n\r\nI'm investigating this.","embeddings":[-0.1504657269,-0.0591697134,-0.0577093922,0.2559515834,0.193053782,0.0901743174,0.0821338445,0.2622757852,-0.2609196305,-0.0551395863,0.0358312577,0.1265247911,0.2174618989,0.090199098,-0.2850414217,-0.1801928431,0.1788396835,0.0559669882,0.0485580564,-0.118900381,-0.1244952604,0.1265652031,-0.3151787817,0.0769735575,-0.4671195447,0.1625408679,-0.3227909505,0.007621164,-0.0032958225,-0.3098783493,-0.0556449778,0.3040577769,-0.0451283157,0.32534796,-0.0001139846,-0.1404050738,0.020385908,-0.0589029714,-0.1106781885,-0.1804829538,-0.1020738631,-0.5905038118,-0.0913802311,-0.2002821118,0.1462708861,-0.6070088148,-0.2705745995,-0.2687054873,0.2620135248,-0.1385006309,0.159783259,-0.0676152855,0.1747838706,-0.1806190014,-0.1066621318,0.0965181813,0.0347622149,-0.0004105456,-0.1437833905,-0.0325336047,0.4414526224,0.3706357181,-0.1782261729,-0.3017144203,-0.1555945128,0.1617710441,0.0063124946,-0.3817135096,-0.0536329299,0.1203863546,0.3290899098,-0.3443950415,-0.3714585006,-0.4346679449,0.1693651974,-0.2393826842,0.03127525,-0.0691222697,0.2867306471,0.2179713845,-0.0308873206,-0.189049691,0.0395146236,0.1386504024,-0.0608814657,0.2488592714,0.0057220296,0.1144914106,0.1244022697,-0.5368865728,0.0223361254,-0.4283556044,-0.0197522454,0.0175422635,-0.5440704823,-0.0473727919,0.0594298653,-0.6148066521,0.2270136178,0.1771862656,-0.0383978337,0.0237068944,-0.0668149516,-0.0024822024,0.2470452338,0.4188669622,0.2312257588,0.2866076529,-0.0470639057,-0.225259766,-0.2453390211,-0.0734846815,-0.1264434159,-0.2911265194,0.1000443324,0.0113737918,0.396143049,-0.0294776019,-0.0793904811,0.11111404,-0.3544274569,-0.1133779958,0.1028654203,0.0086840028,-0.0921811163,0.6188397408,-0.1172470599,0.2566657662,0.1120014265,-0.1708726287,-0.1573576778,-0.1931791455,0.0105560739,0.1908484846,-0.009235274,-0.2236082852,0.0366123281,0.2651329041,-0.0237969682,-0.1290379465,0.0272331573,-0.3101315796,0.0045604357,0.0218815319,0.1250576526,0.0520775132,0.0854914114,-0.0011943799,0.024049921,0.0454844385,-0.5595525503,-0.1502062678,-0.3701137304,0.2136218846,0.0852637514,0.2996404767,0.0801235214,0.302719593,0.2509583235,0.0506003797,-0.1885678768,-0.0870354995,0.0076499549,-0.1807177514,0.082131654,0.0898533612,0.0677425563,0.0127707971,0.1103849784,0.1843047738,0.2871360183,0.4206843376,-0.1703023314,0.3344914913,-0.2040976286,0.4971638322,0.5261321664,-0.0743064508,-0.1810768843,0.2042457908,-0.0661210865,0.3550088406,0.1406594813,-0.1781974137,0.0503419302,0.0754448324,0.4194709361,0.123826772,0.0369044282,-0.3195022941,-0.0728152394,0.0247223936,0.7286319733,-0.3416662514,-0.2831137478,0.2036883235,-0.0616782382,-0.0981448218,0.1693049967,-0.3445923924,-0.05869589,0.1913804859,-0.125977084,0.1830993742,0.0787284821,0.0552581102,-0.3375196755,0.0893760026,-0.0399330445,-0.2428482473,-0.167930007,-0.2065499425,0.0221018586,-0.0164528918,-0.1647251695,0.2443644553,0.1765640825,0.3324424326,-0.1891664118,-0.0546176396,-0.0573655665,0.4379684627,0.0619030967,0.0928421468,-0.1567480415,0.3573679924,0.0526526645,0.0333255939,-0.0547466129,0.1343888491,0.5898091793,-0.1381362081,-0.0987757519,0.4499067068,-0.1939736903,-0.1137822419,-0.0809097514,-0.0667266548,0.10392946,0.0040466455,0.0680115074,-0.1234541014,-0.0436430983,-0.1680187434,0.0705930367,0.2275555879,0.2849643826,0.1425679475,0.0359224901,-0.0949881747,0.1146106496,-0.1052120999,-0.1035783067,-0.4422333241,0.2089213729,0.0312789194,-0.2695364952,0.0007309509,-0.4664670825,0.2603234649,0.3221491277,0.1330774277,0.0780856013,0.0291613564,0.0404994898,0.0632191226,0.0192058738,-0.2066784352,0.4321274161,0.1638639122,-0.003215817,0.1260049194,-0.094315432,0.0408444293,0.182420522,-0.0683083534,0.0810810626,0.4733741879,0.2709379196,0.0947590172,-0.0473091118,-0.1575174481,0.0607751161,-0.1024566963,-0.1658180505,0.128069371,-0.6087813377,-0.1715297848,-0.2396755964,-0.5274429917,-0.2216403484,-0.4587311447,-0.1826675832,0.0086168805,-0.1529063731,0.172349304,0.1213834211,-0.1309144199,0.1670149118,0.0995271653,0.1163859367,0.1581566781,-0.0811959282,-0.0527282991,-0.1512535363,-0.0578469522,0.0354970507,-0.0889788717,-0.0157376137,-0.3222573996,-0.2392690182,0.1334006041,-0.2884916961,-0.1630610228,-0.0221155547,-0.0977419615,0.04354183,-0.4574145079,0.3913764656,0.5567986965,0.0928269625,0.2175679207,0.3165223002,-0.3759053051,-0.2535746992,-0.3681758046,-0.2974682748,-0.0178930797,0.1713598073,-0.256580472,0.1625857204,0.5247507095,0.0695422068,-0.2070610821,0.2132777125,0.2021140903,-0.1964456588,-0.1422774345,0.351760447,-0.2002482265,0.0002508448,-0.2620691359,-0.0276876092,-0.1174924597,0.2562609017,-0.2546110153,-0.0034958834,0.029826317,0.0769765601,-0.3703232408,0.0944289044,0.0625614673,0.3755552471,-0.1497161239,0.0357851461,-0.1428721547,0.0463107303,0.3056418896,0.4419403374,-0.182499513,-0.0159339756,-0.1022930294,-0.0357602537,0.2336682528,-0.0100663975,0.2705745697,-0.0379700698,0.2105386555,-0.0199496839,-0.6065669656,-0.2009226382,-0.2038234323,0.1288241297,-0.0388549417,-0.1556237787,0.2491723299,0.0688511431,0.1483340561,0.0078343702,-0.1901330203,-0.0095134685,-0.1038508192,0.4142340422,-0.0300220437,0.005481204,-0.2045525163,0.2662065029,0.0557391569,-0.0655134395,0.1078738794,-0.2636257708,-0.5134298801,-0.2311079204,-0.3840075135,0.3098147213,0.1988602579,0.370665431,0.1436106563,0.0540110432,-0.2669863403,0.1135599613,0.7118365169,-0.1150130257,-0.385193944,0.4801951647,0.0020542217,-0.076846607,-0.3024656773,-0.0549161136,0.2780228853,0.1245328113,0.5626254082,-0.1685641408,-0.1218298748,0.1117850393,0.6294497848,-0.1057713926,-0.1756478846,-0.112501964,-0.0003586744,-0.3241675198,-0.51160115,-0.2100287527,0.4020390213,-0.3005683124,-0.1428620964,-0.2321521044,-0.3157801628,0.3009279072,0.2026450038,0.466409564,0.1533275843,0.2751990557,-0.0724410638,0.2755639553,0.3012104332,0.6378158927,0.1912370026,-0.7376356125,-0.1446775496,-0.0231771059,0.3582291007,0.0496122055,-0.2540135682,0.4031372368,-0.4295672476,0.3281929791,-0.1531596333,0.1966497302,0.5844236016,0.1333208829,-0.5558376312,-0.1940577179,0.165182367,0.1863946617,-0.0991586596,0.2788518667,-0.4781163037,-0.384187907,0.0734137297,-0.1341848373,0.8999783993,0.2740628421,-0.1333556175,-0.0925272033,0.0792021602,0.2988036871,-0.2731703818,0.1949479282,-0.0499312393,-0.2394651622,-0.0514806062,0.0525455475,0.1468616128,0.0526653193,-0.7124878764,0.3619901538,-0.4484539628,0.0995696783,-0.2969274521,0.2878984809,0.3922804892,-0.3144977689,-0.2345916331,0.0546487682,0.046234075,-0.1462475955,-0.0337212905,-0.0758748576,0.2458808869,-0.2128926218,-0.1421933025,0.2426852882,-0.3409103751,0.4390045404,-0.0448306575,-0.2151910663,0.1573583931,0.0909858793,0.3253347576,0.2573347986,0.0987038314,-0.2113574743,0.0851860791,0.0636720881,-0.3257013559,-0.3507932425,0.1022851691,0.1928027868,-0.3504643738,0.1404039115,0.2587255538,0.0485343188,-0.2336755842,0.0428020135,0.2895253897,-0.5032480359,0.2250385284,0.3453092873,0.111779809,-0.1729832143,0.1053814292,0.1110715345,-0.0358319208,0.0005476316,-0.1623958945,-0.1433448941,-0.0809889063,0.2497196048,0.3761561513,-0.0039944556,0.2022247612,-0.0740246847,-0.0276717376,-0.1196447387,-0.060740862,-0.4044419825,0.0639665425,0.0911104605,-0.0903158113,0.0084972512,-0.2556840777,0.2764193416,0.0067527718,0.2241262347,-0.0271066073,-0.2180060446,-0.1631318778,0.0002046105,0.1853839755,0.1254906803,-0.1021481752,0.4127794206,-0.1707359403,0.338138938,-0.2539671659,-0.088140294,-0.1201729178,0.1959517151,-0.1684363633,-0.0022603103,-0.1586277634,-0.0449553803,0.1960983127,0.219684884,0.0685281083,-0.1981114149,-0.1344372034,0.1306734681,0.1136477888,-0.1090045646,-0.1214502305,-0.1655697376,0.2467256337,-0.3233593404,0.108999148,0.1586366594,0.1293053627,0.2555963993,0.0440423004,-0.03267904,0.0062531224,0.0737168118,0.0863091946,0.2640117407,0.0799779817,0.1482904255,-0.2073283792,-0.0573069565,-0.629260838,-0.1103538871,0.0366320759,-0.1719390899,0.1182309017,-0.1929187179,-0.1658097059,-0.059261661,0.2319018394,0.6020119786,-0.398537457,-0.0572810806,0.144917056,0.1897978336,-0.2437744141,-0.3710727394,0.0801532641,0.2015955448,-0.1089889631,0.257812351,0.2650195956,0.0011206222,0.0896895975,-0.0691619813,0.0189596098,0.2882750034,0.4360206127,0.2229431868,0.0661664829,0.0573681854,0.0454018898,0.1606493294,0.1576923132,0.1019946709,0.029053824,0.1478938162,0.3706709445,0.053790722,0.0660661161,-0.1295982599,0.0943690389,0.2220945805,-0.3501650095,0.1290225536,-0.1412718296,0.0181891341,-0.1675129086,-0.4559591711,-0.0761231035,0.5514985919,-0.0913946405,-0.2190891057,0.3085815609,-0.0410497412,-0.1610819846,-0.1548285931,0.2097725868,-0.0945969373,0.2522943914,-0.146456182,0.0588625632,-0.2706637979,0.084284611,0.3438919187,0.3819179833,-0.3340295255,0.2084652036,0.5395169258,0.0179144628,-0.0434292704,0.3910596371,0.6031447649,0.2915300727,0.0115458295,0.0487371832,-0.1092322096,-0.0658707321,0.1837105006,-0.0105245728,-0.1284661442,-0.3860157728,0.3023032248,0.1048435122,-0.1672857404,0.1264677048,0.2400977015,0.0715405345,-0.4240857661,0.2296592891,0.2455328405,-0.0862919539,-0.0205462538,-0.1232278645,-0.3437090218,-0.3410773873,0.7281009555,0.1575152576,0.1374147385,0.0866666958,0.1286887079,-0.0492461026,0.5275999308,0.2656449676,0.0637388453,-0.2873804569,-0.2236574143,-0.1651127785,0.0836650729,-0.1968185157,0.0170675218,0.2934058607,0.1322610676,-0.0237307828,0.2036402673,0.0310453307,-0.1009669825,-0.0647990257,-0.1896103621,-0.5950562358,-0.1486581415,0.0926938057,-0.0080743739,0.0550474077,-0.0918947682,0.4462240636,-0.1644046754,0.0500477813,-0.0800517946,0.1472428292,0.1145500168,0.1729750782,0.6292982101,0.2540768981,0.3092817664,-0.1138431132,0.1047936156,0.0906459317,-0.3408109248,0.0195527766,0.2614530325,0.3153954744,-0.0783053488,-0.2710303366,0.204288125,-0.1040892825,0.1056802645,0.4074398279,-0.0661337674,-0.1134404391,0.1953424811,0.211631611,0.0576640815,-0.0634947717,0.3801524341,-0.0141822305,0.5103456378,0.1568274051,-0.5181285143,0.4936011136,0.0660220161,-0.4652317762,-0.1163864583,0.134999454,0.1205678955,0.1267579198,-0.6108980775,-0.2417511493,0.5591214895,-0.0480036363,-0.3332654238,0.1799198836,0.0029888649,-0.1483451575,0.0958362445,0.2016952634,-0.0232245196,-0.3746020198,0.1889448166,-0.0530792288]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2750","title":"Second concatenation of datasets produces errors","comments":"Hi @albertvillanova ,\r\nany update on this? Can I probably help in some way?","body":"Hi,\r\n\r\nI am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`.\r\n\r\n```\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\ndata = load_dataset('trec')['train']\r\nconcatenated = concatenate_datasets([data, data])\r\nconcatenated_2 = concatenate_datasets([concatenated, concatenated])\r\nprint('True features of features:', concatenated.features)\r\nprint('\\nProduced features of features:', concatenated_2.features)\r\n```\r\noutputs \r\n\r\n```\r\nTrue features of features: {'label-coarse': ClassLabel(num_classes=6, names=['DESC', 'ENTY', 'ABBR', 'HUM', 'NUM', 'LOC'], names_file=None, id=None), 'label-fine': ClassLabel(num_classes=47, names=['manner', 'cremat', 'animal', 'exp', 'ind', 'gr', 'title', 'def', 'date', 'reason', 'event', 'state', 'desc', 'count', 'other', 'letter', 'religion', 'food', 'country', 'color', 'termeq', 'city', 'body', 'dismed', 'mount', 'money', 'product', 'period', 'substance', 'sport', 'plant', 'techmeth', 'volsize', 'instru', 'abb', 'speed', 'word', 'lang', 'perc', 'code', 'dist', 'temp', 'symbol', 'ord', 'veh', 'weight', 'currency'], names_file=None, id=None), 'text': Value(dtype='string', id=None)}\r\n\r\nProduced features of features: {'label-coarse': Value(dtype='int64', id=None), 'label-fine': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}\r\n```\r\n\r\nI am using `datasets` v.1.11.0","comment_length":14,"text":"Second concatenation of datasets produces errors\nHi,\r\n\r\nI am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`.\r\n\r\n```\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\ndata = load_dataset('trec')['train']\r\nconcatenated = concatenate_datasets([data, data])\r\nconcatenated_2 = concatenate_datasets([concatenated, concatenated])\r\nprint('True features of features:', concatenated.features)\r\nprint('\\nProduced features of features:', concatenated_2.features)\r\n```\r\noutputs \r\n\r\n```\r\nTrue features of features: {'label-coarse': ClassLabel(num_classes=6, names=['DESC', 'ENTY', 'ABBR', 'HUM', 'NUM', 'LOC'], names_file=None, id=None), 'label-fine': ClassLabel(num_classes=47, names=['manner', 'cremat', 'animal', 'exp', 'ind', 'gr', 'title', 'def', 'date', 'reason', 'event', 'state', 'desc', 'count', 'other', 'letter', 'religion', 'food', 'country', 'color', 'termeq', 'city', 'body', 'dismed', 'mount', 'money', 'product', 'period', 'substance', 'sport', 'plant', 'techmeth', 'volsize', 'instru', 'abb', 'speed', 'word', 'lang', 'perc', 'code', 'dist', 'temp', 'symbol', 'ord', 'veh', 'weight', 'currency'], names_file=None, id=None), 'text': Value(dtype='string', id=None)}\r\n\r\nProduced features of features: {'label-coarse': Value(dtype='int64', id=None), 'label-fine': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}\r\n```\r\n\r\nI am using `datasets` v.1.11.0\nHi @albertvillanova ,\r\nany update on this? Can I probably help in some way?","embeddings":[-0.1504657269,-0.0591697134,-0.0577093922,0.2559515834,0.193053782,0.0901743174,0.0821338445,0.2622757852,-0.2609196305,-0.0551395863,0.0358312577,0.1265247911,0.2174618989,0.090199098,-0.2850414217,-0.1801928431,0.1788396835,0.0559669882,0.0485580564,-0.118900381,-0.1244952604,0.1265652031,-0.3151787817,0.0769735575,-0.4671195447,0.1625408679,-0.3227909505,0.007621164,-0.0032958225,-0.3098783493,-0.0556449778,0.3040577769,-0.0451283157,0.32534796,-0.0001139846,-0.1404050738,0.020385908,-0.0589029714,-0.1106781885,-0.1804829538,-0.1020738631,-0.5905038118,-0.0913802311,-0.2002821118,0.1462708861,-0.6070088148,-0.2705745995,-0.2687054873,0.2620135248,-0.1385006309,0.159783259,-0.0676152855,0.1747838706,-0.1806190014,-0.1066621318,0.0965181813,0.0347622149,-0.0004105456,-0.1437833905,-0.0325336047,0.4414526224,0.3706357181,-0.1782261729,-0.3017144203,-0.1555945128,0.1617710441,0.0063124946,-0.3817135096,-0.0536329299,0.1203863546,0.3290899098,-0.3443950415,-0.3714585006,-0.4346679449,0.1693651974,-0.2393826842,0.03127525,-0.0691222697,0.2867306471,0.2179713845,-0.0308873206,-0.189049691,0.0395146236,0.1386504024,-0.0608814657,0.2488592714,0.0057220296,0.1144914106,0.1244022697,-0.5368865728,0.0223361254,-0.4283556044,-0.0197522454,0.0175422635,-0.5440704823,-0.0473727919,0.0594298653,-0.6148066521,0.2270136178,0.1771862656,-0.0383978337,0.0237068944,-0.0668149516,-0.0024822024,0.2470452338,0.4188669622,0.2312257588,0.2866076529,-0.0470639057,-0.225259766,-0.2453390211,-0.0734846815,-0.1264434159,-0.2911265194,0.1000443324,0.0113737918,0.396143049,-0.0294776019,-0.0793904811,0.11111404,-0.3544274569,-0.1133779958,0.1028654203,0.0086840028,-0.0921811163,0.6188397408,-0.1172470599,0.2566657662,0.1120014265,-0.1708726287,-0.1573576778,-0.1931791455,0.0105560739,0.1908484846,-0.009235274,-0.2236082852,0.0366123281,0.2651329041,-0.0237969682,-0.1290379465,0.0272331573,-0.3101315796,0.0045604357,0.0218815319,0.1250576526,0.0520775132,0.0854914114,-0.0011943799,0.024049921,0.0454844385,-0.5595525503,-0.1502062678,-0.3701137304,0.2136218846,0.0852637514,0.2996404767,0.0801235214,0.302719593,0.2509583235,0.0506003797,-0.1885678768,-0.0870354995,0.0076499549,-0.1807177514,0.082131654,0.0898533612,0.0677425563,0.0127707971,0.1103849784,0.1843047738,0.2871360183,0.4206843376,-0.1703023314,0.3344914913,-0.2040976286,0.4971638322,0.5261321664,-0.0743064508,-0.1810768843,0.2042457908,-0.0661210865,0.3550088406,0.1406594813,-0.1781974137,0.0503419302,0.0754448324,0.4194709361,0.123826772,0.0369044282,-0.3195022941,-0.0728152394,0.0247223936,0.7286319733,-0.3416662514,-0.2831137478,0.2036883235,-0.0616782382,-0.0981448218,0.1693049967,-0.3445923924,-0.05869589,0.1913804859,-0.125977084,0.1830993742,0.0787284821,0.0552581102,-0.3375196755,0.0893760026,-0.0399330445,-0.2428482473,-0.167930007,-0.2065499425,0.0221018586,-0.0164528918,-0.1647251695,0.2443644553,0.1765640825,0.3324424326,-0.1891664118,-0.0546176396,-0.0573655665,0.4379684627,0.0619030967,0.0928421468,-0.1567480415,0.3573679924,0.0526526645,0.0333255939,-0.0547466129,0.1343888491,0.5898091793,-0.1381362081,-0.0987757519,0.4499067068,-0.1939736903,-0.1137822419,-0.0809097514,-0.0667266548,0.10392946,0.0040466455,0.0680115074,-0.1234541014,-0.0436430983,-0.1680187434,0.0705930367,0.2275555879,0.2849643826,0.1425679475,0.0359224901,-0.0949881747,0.1146106496,-0.1052120999,-0.1035783067,-0.4422333241,0.2089213729,0.0312789194,-0.2695364952,0.0007309509,-0.4664670825,0.2603234649,0.3221491277,0.1330774277,0.0780856013,0.0291613564,0.0404994898,0.0632191226,0.0192058738,-0.2066784352,0.4321274161,0.1638639122,-0.003215817,0.1260049194,-0.094315432,0.0408444293,0.182420522,-0.0683083534,0.0810810626,0.4733741879,0.2709379196,0.0947590172,-0.0473091118,-0.1575174481,0.0607751161,-0.1024566963,-0.1658180505,0.128069371,-0.6087813377,-0.1715297848,-0.2396755964,-0.5274429917,-0.2216403484,-0.4587311447,-0.1826675832,0.0086168805,-0.1529063731,0.172349304,0.1213834211,-0.1309144199,0.1670149118,0.0995271653,0.1163859367,0.1581566781,-0.0811959282,-0.0527282991,-0.1512535363,-0.0578469522,0.0354970507,-0.0889788717,-0.0157376137,-0.3222573996,-0.2392690182,0.1334006041,-0.2884916961,-0.1630610228,-0.0221155547,-0.0977419615,0.04354183,-0.4574145079,0.3913764656,0.5567986965,0.0928269625,0.2175679207,0.3165223002,-0.3759053051,-0.2535746992,-0.3681758046,-0.2974682748,-0.0178930797,0.1713598073,-0.256580472,0.1625857204,0.5247507095,0.0695422068,-0.2070610821,0.2132777125,0.2021140903,-0.1964456588,-0.1422774345,0.351760447,-0.2002482265,0.0002508448,-0.2620691359,-0.0276876092,-0.1174924597,0.2562609017,-0.2546110153,-0.0034958834,0.029826317,0.0769765601,-0.3703232408,0.0944289044,0.0625614673,0.3755552471,-0.1497161239,0.0357851461,-0.1428721547,0.0463107303,0.3056418896,0.4419403374,-0.182499513,-0.0159339756,-0.1022930294,-0.0357602537,0.2336682528,-0.0100663975,0.2705745697,-0.0379700698,0.2105386555,-0.0199496839,-0.6065669656,-0.2009226382,-0.2038234323,0.1288241297,-0.0388549417,-0.1556237787,0.2491723299,0.0688511431,0.1483340561,0.0078343702,-0.1901330203,-0.0095134685,-0.1038508192,0.4142340422,-0.0300220437,0.005481204,-0.2045525163,0.2662065029,0.0557391569,-0.0655134395,0.1078738794,-0.2636257708,-0.5134298801,-0.2311079204,-0.3840075135,0.3098147213,0.1988602579,0.370665431,0.1436106563,0.0540110432,-0.2669863403,0.1135599613,0.7118365169,-0.1150130257,-0.385193944,0.4801951647,0.0020542217,-0.076846607,-0.3024656773,-0.0549161136,0.2780228853,0.1245328113,0.5626254082,-0.1685641408,-0.1218298748,0.1117850393,0.6294497848,-0.1057713926,-0.1756478846,-0.112501964,-0.0003586744,-0.3241675198,-0.51160115,-0.2100287527,0.4020390213,-0.3005683124,-0.1428620964,-0.2321521044,-0.3157801628,0.3009279072,0.2026450038,0.466409564,0.1533275843,0.2751990557,-0.0724410638,0.2755639553,0.3012104332,0.6378158927,0.1912370026,-0.7376356125,-0.1446775496,-0.0231771059,0.3582291007,0.0496122055,-0.2540135682,0.4031372368,-0.4295672476,0.3281929791,-0.1531596333,0.1966497302,0.5844236016,0.1333208829,-0.5558376312,-0.1940577179,0.165182367,0.1863946617,-0.0991586596,0.2788518667,-0.4781163037,-0.384187907,0.0734137297,-0.1341848373,0.8999783993,0.2740628421,-0.1333556175,-0.0925272033,0.0792021602,0.2988036871,-0.2731703818,0.1949479282,-0.0499312393,-0.2394651622,-0.0514806062,0.0525455475,0.1468616128,0.0526653193,-0.7124878764,0.3619901538,-0.4484539628,0.0995696783,-0.2969274521,0.2878984809,0.3922804892,-0.3144977689,-0.2345916331,0.0546487682,0.046234075,-0.1462475955,-0.0337212905,-0.0758748576,0.2458808869,-0.2128926218,-0.1421933025,0.2426852882,-0.3409103751,0.4390045404,-0.0448306575,-0.2151910663,0.1573583931,0.0909858793,0.3253347576,0.2573347986,0.0987038314,-0.2113574743,0.0851860791,0.0636720881,-0.3257013559,-0.3507932425,0.1022851691,0.1928027868,-0.3504643738,0.1404039115,0.2587255538,0.0485343188,-0.2336755842,0.0428020135,0.2895253897,-0.5032480359,0.2250385284,0.3453092873,0.111779809,-0.1729832143,0.1053814292,0.1110715345,-0.0358319208,0.0005476316,-0.1623958945,-0.1433448941,-0.0809889063,0.2497196048,0.3761561513,-0.0039944556,0.2022247612,-0.0740246847,-0.0276717376,-0.1196447387,-0.060740862,-0.4044419825,0.0639665425,0.0911104605,-0.0903158113,0.0084972512,-0.2556840777,0.2764193416,0.0067527718,0.2241262347,-0.0271066073,-0.2180060446,-0.1631318778,0.0002046105,0.1853839755,0.1254906803,-0.1021481752,0.4127794206,-0.1707359403,0.338138938,-0.2539671659,-0.088140294,-0.1201729178,0.1959517151,-0.1684363633,-0.0022603103,-0.1586277634,-0.0449553803,0.1960983127,0.219684884,0.0685281083,-0.1981114149,-0.1344372034,0.1306734681,0.1136477888,-0.1090045646,-0.1214502305,-0.1655697376,0.2467256337,-0.3233593404,0.108999148,0.1586366594,0.1293053627,0.2555963993,0.0440423004,-0.03267904,0.0062531224,0.0737168118,0.0863091946,0.2640117407,0.0799779817,0.1482904255,-0.2073283792,-0.0573069565,-0.629260838,-0.1103538871,0.0366320759,-0.1719390899,0.1182309017,-0.1929187179,-0.1658097059,-0.059261661,0.2319018394,0.6020119786,-0.398537457,-0.0572810806,0.144917056,0.1897978336,-0.2437744141,-0.3710727394,0.0801532641,0.2015955448,-0.1089889631,0.257812351,0.2650195956,0.0011206222,0.0896895975,-0.0691619813,0.0189596098,0.2882750034,0.4360206127,0.2229431868,0.0661664829,0.0573681854,0.0454018898,0.1606493294,0.1576923132,0.1019946709,0.029053824,0.1478938162,0.3706709445,0.053790722,0.0660661161,-0.1295982599,0.0943690389,0.2220945805,-0.3501650095,0.1290225536,-0.1412718296,0.0181891341,-0.1675129086,-0.4559591711,-0.0761231035,0.5514985919,-0.0913946405,-0.2190891057,0.3085815609,-0.0410497412,-0.1610819846,-0.1548285931,0.2097725868,-0.0945969373,0.2522943914,-0.146456182,0.0588625632,-0.2706637979,0.084284611,0.3438919187,0.3819179833,-0.3340295255,0.2084652036,0.5395169258,0.0179144628,-0.0434292704,0.3910596371,0.6031447649,0.2915300727,0.0115458295,0.0487371832,-0.1092322096,-0.0658707321,0.1837105006,-0.0105245728,-0.1284661442,-0.3860157728,0.3023032248,0.1048435122,-0.1672857404,0.1264677048,0.2400977015,0.0715405345,-0.4240857661,0.2296592891,0.2455328405,-0.0862919539,-0.0205462538,-0.1232278645,-0.3437090218,-0.3410773873,0.7281009555,0.1575152576,0.1374147385,0.0866666958,0.1286887079,-0.0492461026,0.5275999308,0.2656449676,0.0637388453,-0.2873804569,-0.2236574143,-0.1651127785,0.0836650729,-0.1968185157,0.0170675218,0.2934058607,0.1322610676,-0.0237307828,0.2036402673,0.0310453307,-0.1009669825,-0.0647990257,-0.1896103621,-0.5950562358,-0.1486581415,0.0926938057,-0.0080743739,0.0550474077,-0.0918947682,0.4462240636,-0.1644046754,0.0500477813,-0.0800517946,0.1472428292,0.1145500168,0.1729750782,0.6292982101,0.2540768981,0.3092817664,-0.1138431132,0.1047936156,0.0906459317,-0.3408109248,0.0195527766,0.2614530325,0.3153954744,-0.0783053488,-0.2710303366,0.204288125,-0.1040892825,0.1056802645,0.4074398279,-0.0661337674,-0.1134404391,0.1953424811,0.211631611,0.0576640815,-0.0634947717,0.3801524341,-0.0141822305,0.5103456378,0.1568274051,-0.5181285143,0.4936011136,0.0660220161,-0.4652317762,-0.1163864583,0.134999454,0.1205678955,0.1267579198,-0.6108980775,-0.2417511493,0.5591214895,-0.0480036363,-0.3332654238,0.1799198836,0.0029888649,-0.1483451575,0.0958362445,0.2016952634,-0.0232245196,-0.3746020198,0.1889448166,-0.0530792288]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2750","title":"Second concatenation of datasets produces errors","comments":"Hi @Aktsvigun! We are planning to address this issue before our next release, in a couple of weeks at most. \ud83d\ude05 \r\n\r\nIn the meantime, if you would like to contribute, feel free to open a Pull Request. You are welcome. Here you can find more information: [How to contribute to Datasets?](CONTRIBUTING.md)","body":"Hi,\r\n\r\nI am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`.\r\n\r\n```\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\ndata = load_dataset('trec')['train']\r\nconcatenated = concatenate_datasets([data, data])\r\nconcatenated_2 = concatenate_datasets([concatenated, concatenated])\r\nprint('True features of features:', concatenated.features)\r\nprint('\\nProduced features of features:', concatenated_2.features)\r\n```\r\noutputs \r\n\r\n```\r\nTrue features of features: {'label-coarse': ClassLabel(num_classes=6, names=['DESC', 'ENTY', 'ABBR', 'HUM', 'NUM', 'LOC'], names_file=None, id=None), 'label-fine': ClassLabel(num_classes=47, names=['manner', 'cremat', 'animal', 'exp', 'ind', 'gr', 'title', 'def', 'date', 'reason', 'event', 'state', 'desc', 'count', 'other', 'letter', 'religion', 'food', 'country', 'color', 'termeq', 'city', 'body', 'dismed', 'mount', 'money', 'product', 'period', 'substance', 'sport', 'plant', 'techmeth', 'volsize', 'instru', 'abb', 'speed', 'word', 'lang', 'perc', 'code', 'dist', 'temp', 'symbol', 'ord', 'veh', 'weight', 'currency'], names_file=None, id=None), 'text': Value(dtype='string', id=None)}\r\n\r\nProduced features of features: {'label-coarse': Value(dtype='int64', id=None), 'label-fine': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}\r\n```\r\n\r\nI am using `datasets` v.1.11.0","comment_length":51,"text":"Second concatenation of datasets produces errors\nHi,\r\n\r\nI am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`.\r\n\r\n```\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\ndata = load_dataset('trec')['train']\r\nconcatenated = concatenate_datasets([data, data])\r\nconcatenated_2 = concatenate_datasets([concatenated, concatenated])\r\nprint('True features of features:', concatenated.features)\r\nprint('\\nProduced features of features:', concatenated_2.features)\r\n```\r\noutputs \r\n\r\n```\r\nTrue features of features: {'label-coarse': ClassLabel(num_classes=6, names=['DESC', 'ENTY', 'ABBR', 'HUM', 'NUM', 'LOC'], names_file=None, id=None), 'label-fine': ClassLabel(num_classes=47, names=['manner', 'cremat', 'animal', 'exp', 'ind', 'gr', 'title', 'def', 'date', 'reason', 'event', 'state', 'desc', 'count', 'other', 'letter', 'religion', 'food', 'country', 'color', 'termeq', 'city', 'body', 'dismed', 'mount', 'money', 'product', 'period', 'substance', 'sport', 'plant', 'techmeth', 'volsize', 'instru', 'abb', 'speed', 'word', 'lang', 'perc', 'code', 'dist', 'temp', 'symbol', 'ord', 'veh', 'weight', 'currency'], names_file=None, id=None), 'text': Value(dtype='string', id=None)}\r\n\r\nProduced features of features: {'label-coarse': Value(dtype='int64', id=None), 'label-fine': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}\r\n```\r\n\r\nI am using `datasets` v.1.11.0\nHi @Aktsvigun! We are planning to address this issue before our next release, in a couple of weeks at most. \ud83d\ude05 \r\n\r\nIn the meantime, if you would like to contribute, feel free to open a Pull Request. You are welcome. Here you can find more information: [How to contribute to Datasets?](CONTRIBUTING.md)","embeddings":[-0.1504657269,-0.0591697134,-0.0577093922,0.2559515834,0.193053782,0.0901743174,0.0821338445,0.2622757852,-0.2609196305,-0.0551395863,0.0358312577,0.1265247911,0.2174618989,0.090199098,-0.2850414217,-0.1801928431,0.1788396835,0.0559669882,0.0485580564,-0.118900381,-0.1244952604,0.1265652031,-0.3151787817,0.0769735575,-0.4671195447,0.1625408679,-0.3227909505,0.007621164,-0.0032958225,-0.3098783493,-0.0556449778,0.3040577769,-0.0451283157,0.32534796,-0.0001139846,-0.1404050738,0.020385908,-0.0589029714,-0.1106781885,-0.1804829538,-0.1020738631,-0.5905038118,-0.0913802311,-0.2002821118,0.1462708861,-0.6070088148,-0.2705745995,-0.2687054873,0.2620135248,-0.1385006309,0.159783259,-0.0676152855,0.1747838706,-0.1806190014,-0.1066621318,0.0965181813,0.0347622149,-0.0004105456,-0.1437833905,-0.0325336047,0.4414526224,0.3706357181,-0.1782261729,-0.3017144203,-0.1555945128,0.1617710441,0.0063124946,-0.3817135096,-0.0536329299,0.1203863546,0.3290899098,-0.3443950415,-0.3714585006,-0.4346679449,0.1693651974,-0.2393826842,0.03127525,-0.0691222697,0.2867306471,0.2179713845,-0.0308873206,-0.189049691,0.0395146236,0.1386504024,-0.0608814657,0.2488592714,0.0057220296,0.1144914106,0.1244022697,-0.5368865728,0.0223361254,-0.4283556044,-0.0197522454,0.0175422635,-0.5440704823,-0.0473727919,0.0594298653,-0.6148066521,0.2270136178,0.1771862656,-0.0383978337,0.0237068944,-0.0668149516,-0.0024822024,0.2470452338,0.4188669622,0.2312257588,0.2866076529,-0.0470639057,-0.225259766,-0.2453390211,-0.0734846815,-0.1264434159,-0.2911265194,0.1000443324,0.0113737918,0.396143049,-0.0294776019,-0.0793904811,0.11111404,-0.3544274569,-0.1133779958,0.1028654203,0.0086840028,-0.0921811163,0.6188397408,-0.1172470599,0.2566657662,0.1120014265,-0.1708726287,-0.1573576778,-0.1931791455,0.0105560739,0.1908484846,-0.009235274,-0.2236082852,0.0366123281,0.2651329041,-0.0237969682,-0.1290379465,0.0272331573,-0.3101315796,0.0045604357,0.0218815319,0.1250576526,0.0520775132,0.0854914114,-0.0011943799,0.024049921,0.0454844385,-0.5595525503,-0.1502062678,-0.3701137304,0.2136218846,0.0852637514,0.2996404767,0.0801235214,0.302719593,0.2509583235,0.0506003797,-0.1885678768,-0.0870354995,0.0076499549,-0.1807177514,0.082131654,0.0898533612,0.0677425563,0.0127707971,0.1103849784,0.1843047738,0.2871360183,0.4206843376,-0.1703023314,0.3344914913,-0.2040976286,0.4971638322,0.5261321664,-0.0743064508,-0.1810768843,0.2042457908,-0.0661210865,0.3550088406,0.1406594813,-0.1781974137,0.0503419302,0.0754448324,0.4194709361,0.123826772,0.0369044282,-0.3195022941,-0.0728152394,0.0247223936,0.7286319733,-0.3416662514,-0.2831137478,0.2036883235,-0.0616782382,-0.0981448218,0.1693049967,-0.3445923924,-0.05869589,0.1913804859,-0.125977084,0.1830993742,0.0787284821,0.0552581102,-0.3375196755,0.0893760026,-0.0399330445,-0.2428482473,-0.167930007,-0.2065499425,0.0221018586,-0.0164528918,-0.1647251695,0.2443644553,0.1765640825,0.3324424326,-0.1891664118,-0.0546176396,-0.0573655665,0.4379684627,0.0619030967,0.0928421468,-0.1567480415,0.3573679924,0.0526526645,0.0333255939,-0.0547466129,0.1343888491,0.5898091793,-0.1381362081,-0.0987757519,0.4499067068,-0.1939736903,-0.1137822419,-0.0809097514,-0.0667266548,0.10392946,0.0040466455,0.0680115074,-0.1234541014,-0.0436430983,-0.1680187434,0.0705930367,0.2275555879,0.2849643826,0.1425679475,0.0359224901,-0.0949881747,0.1146106496,-0.1052120999,-0.1035783067,-0.4422333241,0.2089213729,0.0312789194,-0.2695364952,0.0007309509,-0.4664670825,0.2603234649,0.3221491277,0.1330774277,0.0780856013,0.0291613564,0.0404994898,0.0632191226,0.0192058738,-0.2066784352,0.4321274161,0.1638639122,-0.003215817,0.1260049194,-0.094315432,0.0408444293,0.182420522,-0.0683083534,0.0810810626,0.4733741879,0.2709379196,0.0947590172,-0.0473091118,-0.1575174481,0.0607751161,-0.1024566963,-0.1658180505,0.128069371,-0.6087813377,-0.1715297848,-0.2396755964,-0.5274429917,-0.2216403484,-0.4587311447,-0.1826675832,0.0086168805,-0.1529063731,0.172349304,0.1213834211,-0.1309144199,0.1670149118,0.0995271653,0.1163859367,0.1581566781,-0.0811959282,-0.0527282991,-0.1512535363,-0.0578469522,0.0354970507,-0.0889788717,-0.0157376137,-0.3222573996,-0.2392690182,0.1334006041,-0.2884916961,-0.1630610228,-0.0221155547,-0.0977419615,0.04354183,-0.4574145079,0.3913764656,0.5567986965,0.0928269625,0.2175679207,0.3165223002,-0.3759053051,-0.2535746992,-0.3681758046,-0.2974682748,-0.0178930797,0.1713598073,-0.256580472,0.1625857204,0.5247507095,0.0695422068,-0.2070610821,0.2132777125,0.2021140903,-0.1964456588,-0.1422774345,0.351760447,-0.2002482265,0.0002508448,-0.2620691359,-0.0276876092,-0.1174924597,0.2562609017,-0.2546110153,-0.0034958834,0.029826317,0.0769765601,-0.3703232408,0.0944289044,0.0625614673,0.3755552471,-0.1497161239,0.0357851461,-0.1428721547,0.0463107303,0.3056418896,0.4419403374,-0.182499513,-0.0159339756,-0.1022930294,-0.0357602537,0.2336682528,-0.0100663975,0.2705745697,-0.0379700698,0.2105386555,-0.0199496839,-0.6065669656,-0.2009226382,-0.2038234323,0.1288241297,-0.0388549417,-0.1556237787,0.2491723299,0.0688511431,0.1483340561,0.0078343702,-0.1901330203,-0.0095134685,-0.1038508192,0.4142340422,-0.0300220437,0.005481204,-0.2045525163,0.2662065029,0.0557391569,-0.0655134395,0.1078738794,-0.2636257708,-0.5134298801,-0.2311079204,-0.3840075135,0.3098147213,0.1988602579,0.370665431,0.1436106563,0.0540110432,-0.2669863403,0.1135599613,0.7118365169,-0.1150130257,-0.385193944,0.4801951647,0.0020542217,-0.076846607,-0.3024656773,-0.0549161136,0.2780228853,0.1245328113,0.5626254082,-0.1685641408,-0.1218298748,0.1117850393,0.6294497848,-0.1057713926,-0.1756478846,-0.112501964,-0.0003586744,-0.3241675198,-0.51160115,-0.2100287527,0.4020390213,-0.3005683124,-0.1428620964,-0.2321521044,-0.3157801628,0.3009279072,0.2026450038,0.466409564,0.1533275843,0.2751990557,-0.0724410638,0.2755639553,0.3012104332,0.6378158927,0.1912370026,-0.7376356125,-0.1446775496,-0.0231771059,0.3582291007,0.0496122055,-0.2540135682,0.4031372368,-0.4295672476,0.3281929791,-0.1531596333,0.1966497302,0.5844236016,0.1333208829,-0.5558376312,-0.1940577179,0.165182367,0.1863946617,-0.0991586596,0.2788518667,-0.4781163037,-0.384187907,0.0734137297,-0.1341848373,0.8999783993,0.2740628421,-0.1333556175,-0.0925272033,0.0792021602,0.2988036871,-0.2731703818,0.1949479282,-0.0499312393,-0.2394651622,-0.0514806062,0.0525455475,0.1468616128,0.0526653193,-0.7124878764,0.3619901538,-0.4484539628,0.0995696783,-0.2969274521,0.2878984809,0.3922804892,-0.3144977689,-0.2345916331,0.0546487682,0.046234075,-0.1462475955,-0.0337212905,-0.0758748576,0.2458808869,-0.2128926218,-0.1421933025,0.2426852882,-0.3409103751,0.4390045404,-0.0448306575,-0.2151910663,0.1573583931,0.0909858793,0.3253347576,0.2573347986,0.0987038314,-0.2113574743,0.0851860791,0.0636720881,-0.3257013559,-0.3507932425,0.1022851691,0.1928027868,-0.3504643738,0.1404039115,0.2587255538,0.0485343188,-0.2336755842,0.0428020135,0.2895253897,-0.5032480359,0.2250385284,0.3453092873,0.111779809,-0.1729832143,0.1053814292,0.1110715345,-0.0358319208,0.0005476316,-0.1623958945,-0.1433448941,-0.0809889063,0.2497196048,0.3761561513,-0.0039944556,0.2022247612,-0.0740246847,-0.0276717376,-0.1196447387,-0.060740862,-0.4044419825,0.0639665425,0.0911104605,-0.0903158113,0.0084972512,-0.2556840777,0.2764193416,0.0067527718,0.2241262347,-0.0271066073,-0.2180060446,-0.1631318778,0.0002046105,0.1853839755,0.1254906803,-0.1021481752,0.4127794206,-0.1707359403,0.338138938,-0.2539671659,-0.088140294,-0.1201729178,0.1959517151,-0.1684363633,-0.0022603103,-0.1586277634,-0.0449553803,0.1960983127,0.219684884,0.0685281083,-0.1981114149,-0.1344372034,0.1306734681,0.1136477888,-0.1090045646,-0.1214502305,-0.1655697376,0.2467256337,-0.3233593404,0.108999148,0.1586366594,0.1293053627,0.2555963993,0.0440423004,-0.03267904,0.0062531224,0.0737168118,0.0863091946,0.2640117407,0.0799779817,0.1482904255,-0.2073283792,-0.0573069565,-0.629260838,-0.1103538871,0.0366320759,-0.1719390899,0.1182309017,-0.1929187179,-0.1658097059,-0.059261661,0.2319018394,0.6020119786,-0.398537457,-0.0572810806,0.144917056,0.1897978336,-0.2437744141,-0.3710727394,0.0801532641,0.2015955448,-0.1089889631,0.257812351,0.2650195956,0.0011206222,0.0896895975,-0.0691619813,0.0189596098,0.2882750034,0.4360206127,0.2229431868,0.0661664829,0.0573681854,0.0454018898,0.1606493294,0.1576923132,0.1019946709,0.029053824,0.1478938162,0.3706709445,0.053790722,0.0660661161,-0.1295982599,0.0943690389,0.2220945805,-0.3501650095,0.1290225536,-0.1412718296,0.0181891341,-0.1675129086,-0.4559591711,-0.0761231035,0.5514985919,-0.0913946405,-0.2190891057,0.3085815609,-0.0410497412,-0.1610819846,-0.1548285931,0.2097725868,-0.0945969373,0.2522943914,-0.146456182,0.0588625632,-0.2706637979,0.084284611,0.3438919187,0.3819179833,-0.3340295255,0.2084652036,0.5395169258,0.0179144628,-0.0434292704,0.3910596371,0.6031447649,0.2915300727,0.0115458295,0.0487371832,-0.1092322096,-0.0658707321,0.1837105006,-0.0105245728,-0.1284661442,-0.3860157728,0.3023032248,0.1048435122,-0.1672857404,0.1264677048,0.2400977015,0.0715405345,-0.4240857661,0.2296592891,0.2455328405,-0.0862919539,-0.0205462538,-0.1232278645,-0.3437090218,-0.3410773873,0.7281009555,0.1575152576,0.1374147385,0.0866666958,0.1286887079,-0.0492461026,0.5275999308,0.2656449676,0.0637388453,-0.2873804569,-0.2236574143,-0.1651127785,0.0836650729,-0.1968185157,0.0170675218,0.2934058607,0.1322610676,-0.0237307828,0.2036402673,0.0310453307,-0.1009669825,-0.0647990257,-0.1896103621,-0.5950562358,-0.1486581415,0.0926938057,-0.0080743739,0.0550474077,-0.0918947682,0.4462240636,-0.1644046754,0.0500477813,-0.0800517946,0.1472428292,0.1145500168,0.1729750782,0.6292982101,0.2540768981,0.3092817664,-0.1138431132,0.1047936156,0.0906459317,-0.3408109248,0.0195527766,0.2614530325,0.3153954744,-0.0783053488,-0.2710303366,0.204288125,-0.1040892825,0.1056802645,0.4074398279,-0.0661337674,-0.1134404391,0.1953424811,0.211631611,0.0576640815,-0.0634947717,0.3801524341,-0.0141822305,0.5103456378,0.1568274051,-0.5181285143,0.4936011136,0.0660220161,-0.4652317762,-0.1163864583,0.134999454,0.1205678955,0.1267579198,-0.6108980775,-0.2417511493,0.5591214895,-0.0480036363,-0.3332654238,0.1799198836,0.0029888649,-0.1483451575,0.0958362445,0.2016952634,-0.0232245196,-0.3746020198,0.1889448166,-0.0530792288]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2749","title":"Raise a proper exception when trying to stream a dataset that requires to manually download files","comments":"Hi @severo, thanks for reporting.\r\n\r\nAs discussed, datasets requiring manual download should be:\r\n- programmatically identifiable\r\n- properly handled with more clear error message when trying to load them with streaming\r\n\r\nIn relation with programmatically identifiability, note that for datasets requiring manual download, their builder have a property `manual_download_instructions` which is not None:\r\n```python\r\n# Dataset requiring manual download:\r\nbuilder.manual_download_instructions is not None\r\n```","body":"## Describe the bug\r\n\r\nAt least for 'reclor', 'telugu_books', 'turkish_movie_sentiment', 'ubuntu_dialogs_corpus', 'wikihow', trying to `load_dataset` in streaming mode raises a `TypeError` without any detail about why it fails.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"reclor\", streaming=True)\r\n```\r\n\r\n## Expected results\r\n\r\nIdeally: raise a specific exception, something like `ManualDownloadError`.\r\n\r\nOr at least give the reason in the message, as when we load in normal mode:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"reclor\")\r\n```\r\n\r\n```\r\nAssertionError: The dataset reclor with config default requires manual data.\r\n Please follow the manual download instructions: to use ReClor you need to download it manually. Please go to its homepage (http:\/\/whyu.me\/reclor\/) fill the google\r\n form and you will receive a download link and a password to extract it.Please extract all files in one folder and use the path folder in datasets.load_dataset('reclor', data_dir='path\/to\/folder\/folder_name')\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(reclor, data_dir='')\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS-11.5-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":64,"text":"Raise a proper exception when trying to stream a dataset that requires to manually download files\n## Describe the bug\r\n\r\nAt least for 'reclor', 'telugu_books', 'turkish_movie_sentiment', 'ubuntu_dialogs_corpus', 'wikihow', trying to `load_dataset` in streaming mode raises a `TypeError` without any detail about why it fails.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"reclor\", streaming=True)\r\n```\r\n\r\n## Expected results\r\n\r\nIdeally: raise a specific exception, something like `ManualDownloadError`.\r\n\r\nOr at least give the reason in the message, as when we load in normal mode:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"reclor\")\r\n```\r\n\r\n```\r\nAssertionError: The dataset reclor with config default requires manual data.\r\n Please follow the manual download instructions: to use ReClor you need to download it manually. Please go to its homepage (http:\/\/whyu.me\/reclor\/) fill the google\r\n form and you will receive a download link and a password to extract it.Please extract all files in one folder and use the path folder in datasets.load_dataset('reclor', data_dir='path\/to\/folder\/folder_name')\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(reclor, data_dir='')\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS-11.5-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nHi @severo, thanks for reporting.\r\n\r\nAs discussed, datasets requiring manual download should be:\r\n- programmatically identifiable\r\n- properly handled with more clear error message when trying to load them with streaming\r\n\r\nIn relation with programmatically identifiability, note that for datasets requiring manual download, their builder have a property `manual_download_instructions` which is not None:\r\n```python\r\n# Dataset requiring manual download:\r\nbuilder.manual_download_instructions is not None\r\n```","embeddings":[-0.2240490466,0.0735050663,0.068855688,0.1454883069,0.3467077017,0.0917349607,0.1795138866,0.0385224037,0.1048629507,0.0732556209,0.0714531541,0.2278849632,-0.306107223,0.168981418,-0.0743809417,-0.0853384212,-0.0969869718,0.1296140105,0.2995474637,0.1218056008,-0.1967432946,0.037302237,-0.2146420479,0.0178642627,-0.0018434762,-0.1568493396,0.2580464482,0.2388842106,-0.12881051,-0.3902961612,0.3372131884,0.0121841822,0.4855319858,0.5165213346,-0.0001199634,0.0070316265,0.3886614144,-0.1586243212,-0.5747519135,-0.1283361316,-0.4686380327,-0.0777676925,0.0901651233,-0.0698137358,0.1049859673,-0.1720953882,0.162341401,-0.3645288646,0.2047697306,0.4541603923,0.1066825688,0.2829991579,0.2462426722,0.163706407,0.1327375621,0.351267308,-0.0302184448,0.3713254631,0.0608687028,0.1510274708,0.1606027037,0.1628775001,-0.1228826195,0.0393839516,0.1880534738,-0.1661041975,-0.2295856774,-0.4277891219,-0.1103247255,0.4421196878,0.5148940086,-0.3407017291,-0.4609524012,-0.3798696399,0.0458856784,-0.1476799548,0.1311267167,0.4878360331,-0.4216932058,0.3501260877,-0.0883329585,-0.3158322573,-0.3065256178,0.2305598706,0.0059048333,-0.1519111991,0.055758439,0.0611203685,-0.0835495442,0.1057094559,0.2180747986,-0.309345305,-0.0396562181,0.0888883397,-0.204697147,-0.0041701873,-0.0657843724,-0.0414948389,0.1965882033,0.0238626655,0.3733995557,0.0567336902,-0.1529811174,0.2431636602,0.1901954114,0.0264550317,0.0991011336,0.1258841455,0.4811724126,0.3951644301,0.1332431287,-0.1177569702,-0.0253069643,0.0886065438,0.3152233064,0.2048721611,0.4681082666,-0.1049640924,-0.4039188921,0.0096668303,-0.6102744937,0.1056832075,-0.02290681,0.1641360372,-0.1616206318,0.2007362247,0.2162867635,0.3693430424,-0.0011653588,-0.2878369689,-0.0102987355,-0.0643678904,0.1261957884,0.0260482784,0.1795365959,-0.5311354995,0.0900712088,0.093942605,-0.3286801875,-0.0527605303,0.3502986729,-0.0545802228,-0.159295544,0.3757533729,0.1402717531,0.0273190979,0.1152361333,-0.1576324105,-0.016434364,0.2604483366,-0.0598954596,-0.531606853,0.1063527539,0.1647614092,-0.0167459399,-0.0777090043,-0.5121468306,0.0411837399,0.0013843826,-0.3260016441,-0.0188462473,-0.1534196436,0.3346794546,-0.2917099595,0.0453913435,0.5106771588,-0.3724917471,0.0134214601,-0.4481051266,-0.2220885307,0.7759043574,0.1512121409,-0.2369718254,0.2332099974,-0.3799983263,0.0408191308,0.4848953784,-0.0353461541,-0.3184098303,0.4427749515,0.0254455581,0.4210849106,0.1377783716,0.2182543427,-0.0400385782,-0.1267595738,-0.0261647534,0.5302296281,-0.2185987681,0.0431420021,-0.2089989781,-0.3620250821,0.2253880501,0.2703531682,0.0270372368,0.2280398458,0.3364054561,0.0141229415,0.2515027523,-0.1984639168,-0.1834649891,-0.0087441904,0.4122576416,-0.0866532698,-0.2348504364,-0.1183793917,-0.3734786808,0.2647877336,0.0439619459,0.0354769155,-0.2135419101,-0.3262125254,-0.1027309895,0.0139823742,-0.2305373102,0.0171953775,0.0360648334,0.3708552122,-0.3016510606,0.1690964848,-0.1053365096,0.4814932644,-0.5749636889,0.2593649924,-0.1490311772,0.1025912687,0.0344819985,-0.1005370393,-0.1384553462,-0.3296217024,0.0505902618,-0.0311502721,-0.2466446459,0.4579785466,0.2499203086,0.341388911,-0.2688518167,-0.0825616494,0.0779676735,-0.197544679,-0.227867946,0.5077203512,0.1895116568,-0.0363306254,-0.0994233862,0.2132187188,0.2448489964,0.1440677643,0.0862258375,0.213060692,0.1893789172,0.2060934901,-0.1933785826,0.0910151675,0.1567736268,-0.1239111423,-0.0441933163,-0.2678855062,-0.7136908174,-0.2057197094,0.0648716241,0.0464329794,0.4829679728,0.3096090853,-0.1382143945,0.0907322541,0.25253883,0.3169167042,0.4295144379,0.0733744875,0.11609786,0.0864411369,0.1807482988,-0.2610627711,0.254388839,0.2185361236,-0.2846085429,0.1634678841,0.1206299067,0.1041521356,-0.2988964915,-0.163327828,-0.0516937599,0.1894502938,-0.5312809348,0.0377002135,0.0376430117,-0.1813216358,-0.3296357691,-0.3708743155,-0.1727309972,-0.1337380558,-0.1167202145,0.3897580206,-0.1965971291,0.207007736,-0.0349070318,0.2335080653,-0.0607416183,-0.4499779046,-0.0715346262,0.21721977,-0.2908616662,-0.0184098668,0.3773886561,-0.1365685612,0.4272453785,-0.2955581248,0.0621297061,-0.3178853393,-0.1042977273,-0.0536005981,0.2045689821,0.2286831588,0.3041532934,0.4670447111,0.3198855221,-0.3050571978,0.0509364195,-0.0395637155,-0.0034178786,0.1977665573,-0.0418911129,0.3320870996,-0.1186343729,-0.2866623104,-0.4388065636,-0.4254140556,-0.0446366221,-0.0688903108,0.1023061424,0.3437128961,0.1085468233,0.1080609486,0.2089250386,0.1163057312,0.0747015178,-0.172325626,0.2136213034,0.0754034743,-0.1863212883,0.095716998,-0.0790749639,-0.1870462745,0.6288800836,-0.4854286611,-0.0773327276,-0.2160004526,0.0934364274,-0.1314488202,-0.163055405,0.1723690778,0.0485098325,0.0638696179,0.0521184541,-0.0092215231,-0.1064659432,0.1419488341,0.3870232403,0.1204741821,0.4440130293,-0.0056693312,0.4280276,0.1485658288,0.0279103853,0.5554195046,0.037950702,0.2381644249,0.1094087139,-0.1581960022,-0.0390722267,-0.0468433239,-0.2562383711,0.071328871,-0.0349691436,-0.3694889247,-0.2737471759,-0.1336821467,-0.3286108673,-0.3432630599,0.196297884,-0.0909453332,0.3309890628,0.0070275976,0.1219304875,0.004929495,-0.1368207335,-0.0896774083,0.596214056,0.2017941326,0.1043582782,0.1359161437,-0.0845663026,-0.282364279,0.3272709846,0.1492863297,0.6370370388,-0.0328637362,-0.1540554464,-0.056194216,0.158978194,0.1335155964,-0.4564893842,0.0867487937,-0.1212903932,0.3060760498,-0.3691339195,-0.0652158037,-0.2730306387,0.2272705734,0.2525944412,0.1201704666,-0.3754138947,0.1627915949,-0.1078445688,0.175225839,0.0209251717,-0.2673688531,-0.2003458291,-0.2474560589,-0.2632332444,0.0795574188,0.0657881126,0.0128566511,-0.1496168375,-0.103602469,-0.2790402174,0.2156237364,0.2718067467,-0.0223996621,0.2523845136,-0.0458869226,-0.010379089,0.0428165495,0.0614267439,0.4101091623,0.5562981367,-0.0217061248,-0.6266164184,-0.0508677959,0.0560513176,0.0753826052,0.3416007161,-0.2250460386,-0.0010806679,0.2516185939,-0.0133126276,-0.0711944476,0.430477351,0.2217739373,-0.1622162163,-0.5013062954,-0.7254967093,0.194761768,0.0292814542,-0.081960924,0.3846815228,-0.008730351,-0.4788636863,0.2372295856,-0.1070184931,0.9966676235,0.1952316761,0.094668813,0.2934130132,-0.3080472052,0.2921258211,-0.2087788433,0.0780226961,-0.185116142,-0.043530371,-0.204889074,-0.126767844,0.27618891,-0.1122849211,-0.3511961699,0.3305287957,-0.3107739985,0.2130016387,0.2617628276,0.3083787858,-0.5707214475,-0.2149680257,-0.3299856186,0.0228233673,-0.2422361374,0.224169597,-0.0468313433,-0.1632672399,0.1981519908,-0.2673217952,-0.3233048022,0.0179481097,-0.4679001868,-0.013950876,-0.3990938067,-0.370266825,0.3562408984,0.4006227553,0.0616326667,0.0464915819,-0.2373075485,0.0984614119,-0.0071981372,0.1846889108,0.2170572728,-0.0641302392,0.2954823375,0.0699849501,-0.0635706931,0.1846473515,-0.2823320329,-0.1665865034,-0.0062390687,-0.1703841537,0.1936393231,-0.1883354783,-0.186947912,-0.3215478957,-0.1350565553,-0.0895299092,0.0798922628,-0.0824367702,-0.0284026209,0.176592797,-0.1195740253,-0.1898990273,-0.2120892853,0.3894803226,-0.1033363491,0.2676964402,0.4218558371,-0.0047922353,-0.0913328007,-0.1607122719,0.2487591505,-0.0593642034,-0.2879115343,-0.0581788905,-0.1485290676,0.1245779991,-0.0560325831,0.1912638694,0.2498514056,-0.1431521475,-0.1989729255,-0.5332922935,-0.2473437488,-0.0458375625,0.16446127,0.2406936437,-0.102132231,0.0534521043,0.0385988429,-0.3022888601,-0.2069284469,0.0686588064,-0.0765221268,-0.0498192236,-0.029495623,0.3428830504,0.4114540219,-0.1261543334,0.0110747982,0.079566218,-0.0242402814,-0.052744247,-0.0257125273,0.2072196156,0.0961388424,-0.0974363014,-0.056055475,-0.4067094326,-0.0844239742,-0.0434160754,0.0589241274,-0.0078552756,-0.1694695354,0.1477720141,0.1635249257,0.3355441391,-0.4046207666,0.2251184136,-0.4388322532,0.1751460731,0.0661600605,0.1257627159,0.0047380072,-0.0282686036,0.0239321236,0.0112660592,0.3392134905,0.1978636235,0.2448913604,-0.4220344424,0.0937912613,0.0007048298,0.3157902956,0.4360058904,-0.3139345944,-0.067789048,0.1305146515,0.1659675688,-0.172185421,0.0850876272,0.0192722548,-0.0723217726,-0.0239331108,0.2342981845,-0.0105341244,-0.0923530981,-0.351210475,0.1030642986,0.4362009466,0.2195510417,0.1078603864,0.3562318087,-0.0963432342,0.1072605699,-0.104281947,0.1574417502,0.1261450946,0.0263399407,-0.291826874,0.091523692,0.071654506,-0.0933473185,-0.0582971461,-0.5852045417,0.032785859,0.4007576108,-0.142184332,0.0789683089,0.27359882,0.389793843,-0.3853033483,-0.044326432,-0.3138195872,-0.3281126022,-0.0007676274,0.260597378,-0.1108107045,-0.1875741184,0.0702332184,-0.211680308,0.0564196296,0.0578414686,-0.2524810135,0.0885447115,-0.0331500508,-0.5446571708,-0.1316628754,0.2515082061,0.2410064787,-0.4297248721,-0.0410517678,0.2998056412,-0.051803492,0.1709399372,0.1605734676,0.4806586802,0.2976831496,-0.0374794789,0.1335161477,-0.0354826078,0.0498824492,0.0439476147,0.0491816886,0.0891289189,0.192810908,0.2950775027,0.1464643031,-0.1442900151,0.1522221416,-0.0911778733,0.0842920914,-0.4881499112,0.329772979,-0.1628800929,0.037109483,-0.2197921276,-0.213310346,-0.1095037013,0.0350301191,0.4889464974,0.0746517032,0.0925180689,0.0753408968,0.0058635292,-0.0019570594,0.2779946625,0.4232470393,0.0432397723,-0.2326324135,-0.150240913,-0.6847051978,0.0838006139,-0.038199354,-0.1071620658,0.2841880322,0.0338020921,-0.0358228758,0.2476684749,-0.1277788132,0.3261429667,-0.2290099114,0.2403309643,-0.1857775748,-0.5727883577,0.0407370552,-0.1749348789,-0.2027567774,-0.4563086927,0.272790879,-0.2935861349,0.0265793651,-0.0100464253,0.0213187933,0.2098412514,0.2045656443,0.5075956583,0.1037400886,0.3434555531,-0.0579454415,-0.3387697041,0.1108119339,-0.052334778,-0.2122107595,0.0282862708,-0.1680826694,0.585057795,-0.1310547739,0.1079022437,-0.5024103522,0.4316817522,0.222895354,-0.174565196,-0.4910794497,0.1436585933,-0.2482076138,0.1482337713,-0.0705923364,0.0483024828,-0.0161168743,0.1445305198,-0.2441436052,-0.4462406635,0.5733362436,-0.1650540084,-0.0774964318,0.0621335879,0.2239474654,0.0654136688,0.1732180268,-0.4505608976,0.1712160558,0.3813616931,-0.2176904529,-0.0624277852,-0.063949503,-0.0404645987,0.3741620481,0.0852383673,0.0944983959,-0.0947523117,-0.1604730487,0.1903830618,-0.1142488122]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2749","title":"Raise a proper exception when trying to stream a dataset that requires to manually download files","comments":"Thanks @albertvillanova ","body":"## Describe the bug\r\n\r\nAt least for 'reclor', 'telugu_books', 'turkish_movie_sentiment', 'ubuntu_dialogs_corpus', 'wikihow', trying to `load_dataset` in streaming mode raises a `TypeError` without any detail about why it fails.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"reclor\", streaming=True)\r\n```\r\n\r\n## Expected results\r\n\r\nIdeally: raise a specific exception, something like `ManualDownloadError`.\r\n\r\nOr at least give the reason in the message, as when we load in normal mode:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"reclor\")\r\n```\r\n\r\n```\r\nAssertionError: The dataset reclor with config default requires manual data.\r\n Please follow the manual download instructions: to use ReClor you need to download it manually. Please go to its homepage (http:\/\/whyu.me\/reclor\/) fill the google\r\n form and you will receive a download link and a password to extract it.Please extract all files in one folder and use the path folder in datasets.load_dataset('reclor', data_dir='path\/to\/folder\/folder_name')\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(reclor, data_dir='')\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS-11.5-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":2,"text":"Raise a proper exception when trying to stream a dataset that requires to manually download files\n## Describe the bug\r\n\r\nAt least for 'reclor', 'telugu_books', 'turkish_movie_sentiment', 'ubuntu_dialogs_corpus', 'wikihow', trying to `load_dataset` in streaming mode raises a `TypeError` without any detail about why it fails.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"reclor\", streaming=True)\r\n```\r\n\r\n## Expected results\r\n\r\nIdeally: raise a specific exception, something like `ManualDownloadError`.\r\n\r\nOr at least give the reason in the message, as when we load in normal mode:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"reclor\")\r\n```\r\n\r\n```\r\nAssertionError: The dataset reclor with config default requires manual data.\r\n Please follow the manual download instructions: to use ReClor you need to download it manually. Please go to its homepage (http:\/\/whyu.me\/reclor\/) fill the google\r\n form and you will receive a download link and a password to extract it.Please extract all files in one folder and use the path folder in datasets.load_dataset('reclor', data_dir='path\/to\/folder\/folder_name')\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(reclor, data_dir='')\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: macOS-11.5-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nThanks @albertvillanova ","embeddings":[-0.1975738853,-0.0646141395,0.1009067744,0.1245716289,0.3010841608,0.1299112439,0.1154407859,0.0606036633,0.1342097819,0.068457,0.0895295516,0.1881507188,-0.2988877296,0.1655600816,-0.0889565647,-0.0553256758,-0.0640125349,0.170885399,0.246596545,0.0845115185,-0.1935782135,-0.0137948189,-0.2771826386,0.0294405986,-0.0049575884,-0.1352417469,0.2491986454,0.2622209191,-0.1448888332,-0.3700896204,0.4256049991,-0.0763254464,0.5232368112,0.5654091239,-0.0001238985,-0.0032435411,0.4001612067,-0.1125989854,-0.5048841238,-0.1028546095,-0.4910790026,-0.0852655917,0.1412660033,-0.0603791997,0.0756291598,-0.2200090587,0.1193614379,-0.3511401415,0.2703941166,0.4962480068,0.0812106207,0.2581282258,0.2068020552,0.093825087,0.1560520977,0.3605677485,-0.0017650049,0.4128413796,0.0865037292,0.0993387625,0.2270660698,0.1077716351,-0.1517692655,0.0596095175,0.1583005041,-0.1655384898,-0.2805836201,-0.4110898972,-0.0693731159,0.4448114932,0.5491130352,-0.3268358707,-0.4224393368,-0.3403680921,0.0262219794,-0.1067586169,0.1277688146,0.4465924501,-0.4286827147,0.3472199142,-0.0255359728,-0.4110637307,-0.3274691999,0.2742733061,-0.0299841706,-0.1672925055,0.0335006118,0.0478859991,-0.0302793626,0.1493332386,0.2098626196,-0.3144459426,-0.055731684,0.0635735169,-0.2561118603,-0.0493388958,-0.0902472064,-0.0203403588,0.1715281755,-0.0152351595,0.3008235395,0.0377094597,-0.1970220208,0.2465382218,0.293374151,-0.0485421121,0.0950214714,0.1062958986,0.4937499166,0.2837011814,0.1439351737,-0.1241644174,0.067215994,0.080959484,0.3093741238,0.1799200475,0.5027268529,-0.139037028,-0.2760890126,-0.0139248446,-0.6197627187,0.0951069072,-0.0586613007,0.1464385837,-0.1775619239,0.2529213727,0.2392013371,0.4172401726,0.031076042,-0.2827266753,0.0008300364,-0.0474554934,0.115205422,0.0264634155,0.1955853403,-0.5165011287,0.0752975196,0.1583190262,-0.311945051,-0.0737458691,0.3899900615,-0.0977003276,-0.1560954899,0.4156757891,0.1810763627,0.0909808353,0.1430080533,-0.1368178576,-0.0154574076,0.322642982,-0.1135138422,-0.5192280412,0.0929766372,0.1188297719,0.0077864658,-0.1409145296,-0.5632780194,-0.0087203244,-0.0137426993,-0.3171853125,0.0102818115,-0.1859032363,0.3533522189,-0.3048823178,0.0516930968,0.4448133707,-0.4284979999,0.0550041795,-0.5329200625,-0.1991171837,0.7489522099,0.1366569996,-0.2956022024,0.2800389826,-0.3661877215,0.096004039,0.5354598761,-0.0407016836,-0.3190988004,0.4313433468,0.0520697683,0.3557985425,0.1483365148,0.247750774,-0.049393747,-0.0733502507,0.0030936135,0.4752152562,-0.2463573664,0.0317447297,-0.1426499933,-0.2735365927,0.2456280142,0.28460899,0.0139940213,0.2285872996,0.330714643,0.0152606489,0.3275569081,-0.2200641632,-0.2210525572,0.0404808335,0.4100925624,-0.0617787465,-0.2257291377,-0.1455771327,-0.3618359268,0.2068763971,0.0802498013,-0.008888253,-0.1226311773,-0.2924204171,-0.1681331396,-0.0193331111,-0.1860383451,0.0476020426,-0.0084380982,0.3261317015,-0.3519518375,0.1760224253,-0.1662981957,0.4238096774,-0.5557566881,0.2522362471,-0.1254434437,0.1746751368,0.0287769847,-0.1139499247,-0.1279858649,-0.394228667,0.0574553534,0.0042252811,-0.2448651344,0.4434017539,0.2621207833,0.3120868504,-0.2662701607,-0.1522647291,0.1325532347,-0.2394411564,-0.2468323261,0.5195616484,0.1766687334,-0.0532146469,-0.1265839189,0.177454263,0.2841694057,0.1520512849,0.041496627,0.2358413637,0.1634016931,0.2074415088,-0.1891226172,0.1678085476,0.2119253725,-0.1213127673,-0.1044299752,-0.2750173509,-0.6770871878,-0.2584836781,0.0913474113,0.0504837334,0.5568054318,0.3636954129,-0.1637730747,0.1314199567,0.2556049526,0.3112256229,0.4888988733,0.0449741781,0.1558320522,0.1332707703,0.2024517655,-0.2412426472,0.1877892911,0.2043153197,-0.2582996488,0.1336990297,0.0786461011,0.0815460235,-0.2716888785,-0.0866297781,-0.10471531,0.2292313427,-0.4836162329,0.0501140393,-0.0187116973,-0.1449034661,-0.2933523357,-0.3515416086,-0.1468345374,-0.1351743191,-0.1552658528,0.3633770645,-0.1391572505,0.1584842056,-0.061872676,0.1563211977,0.0043041324,-0.4579693377,-0.0838544965,0.2486248463,-0.2506049871,-0.0462400243,0.3349947035,-0.0983723849,0.3765210211,-0.2060984522,0.0071432646,-0.2682625651,-0.0624975339,-0.0942271352,0.1734650284,0.2082577646,0.3031947315,0.5581041574,0.2829796374,-0.2389394194,0.0384702273,-0.0248730443,0.034069445,0.222478956,-0.0278462078,0.3938463628,-0.1008334905,-0.3428724408,-0.4036191106,-0.3848366737,-0.0526582189,-0.1102566198,0.0429576598,0.2735595405,0.1924037486,0.1256449521,0.1240624264,0.0649771243,0.0729717389,-0.1136818901,0.2325005531,0.0943287984,-0.2120794207,0.1239567399,-0.0524869524,-0.2451669127,0.6101920009,-0.5060183406,-0.1255127192,-0.2692707479,0.0556569137,-0.2091535926,-0.0958912969,0.1843014359,-0.0340403095,0.0756532922,0.0829082057,0.0445126779,-0.1205143556,0.1287115514,0.3492148817,0.1128054708,0.4649344385,-0.0487550721,0.3096621931,0.158772096,0.0882144198,0.5750720501,-0.0116795413,0.3106136024,0.1076426506,-0.196192503,-0.1102588996,-0.0517691858,-0.2554695308,0.0234511066,-0.0401870683,-0.3565119505,-0.3103529811,-0.1992713809,-0.3130309284,-0.3157689571,0.1794984043,-0.0446635559,0.3212704062,-0.0380387902,0.1096267998,0.0573679954,-0.136533305,-0.072530143,0.5885931849,0.1885402948,0.0708069429,0.0465513244,-0.1772979945,-0.2915691435,0.3154720664,0.130025208,0.6931040883,-0.0423886664,-0.0647268817,-0.0425950438,0.1911508441,0.1917791665,-0.4346577227,0.1175433919,-0.166692704,0.2778237462,-0.4344634712,0.0113199521,-0.3083179295,0.2520684302,0.2553509176,0.1739417464,-0.3064784706,0.1729821414,-0.0962667763,0.216299057,0.0400379747,-0.2745391428,-0.2082114816,-0.31862849,-0.2041926682,0.074252896,0.1139583364,0.0090665286,-0.0866307095,-0.1420052797,-0.2735147476,0.1324746758,0.2436266243,-0.0731947199,0.2247017622,-0.0622171722,-0.0580096617,-0.0207700972,-0.0200034808,0.4123843312,0.5402238965,0.0310669728,-0.6061320901,-0.0329909176,0.0440418683,0.1293848902,0.2849460244,-0.3089448214,0.0694979951,0.287248224,-0.0221481826,-0.0099568302,0.3599005044,0.26541996,-0.1120681092,-0.4758700728,-0.6815021038,0.2436866462,0.0668724775,-0.0497266091,0.3791180551,0.05474253,-0.399609834,0.2078880966,-0.1681960076,0.9599691629,0.184772566,0.138263464,0.2773056924,-0.2958838642,0.3553560972,-0.2245488018,0.0673221424,-0.1584539413,-0.1360498965,-0.2272124588,-0.1952988505,0.3011102974,-0.1383203715,-0.3542455435,0.3589805663,-0.3473806977,0.178876251,0.2988596261,0.2595911622,-0.5522857904,-0.1461444348,-0.3292716146,-0.0077075982,-0.2177015394,0.2300803214,-0.0562403947,-0.1348083168,0.1250351369,-0.2894686759,-0.386449039,0.0196945518,-0.430613637,-0.1055280045,-0.3684163392,-0.403755486,0.3970203698,0.4749249518,0.0696459338,0.1381573677,-0.2808059752,0.0643949062,-0.0939636007,0.1372372657,0.1907990128,-0.0151347909,0.1896858662,0.0665871873,-0.1556336731,0.3155349195,-0.2639328837,-0.2672538161,0.040877331,-0.1034855172,0.1411716938,-0.1781279445,-0.2466555238,-0.3122947514,-0.1148746684,-0.1106561944,0.0583116263,-0.0145728979,-0.0171425156,0.2512451112,-0.1010310873,-0.1946640611,-0.2186662406,0.4341676831,-0.0610055737,0.2536640167,0.4833369255,0.0664529502,-0.0914841741,-0.1458890438,0.2955719233,-0.0613909997,-0.1989216506,-0.0292381775,-0.1711985469,0.1561332643,-0.0700367391,0.1390265375,0.2602269053,-0.164945215,-0.2026525587,-0.5787573457,-0.2200440615,-0.05092071,0.2272560298,0.2307594121,-0.0636097193,0.0308820009,0.097299546,-0.3386579156,-0.1637792587,0.1258933246,-0.0517325625,-0.0203867331,0.0484208278,0.3266091943,0.445576936,-0.1534848958,-0.0103431745,0.0897080824,0.0331358053,-0.0426575281,0.0291422997,0.220466882,0.1112815961,-0.0757997856,-0.0340955108,-0.465795517,-0.1075451449,0.0073665599,-0.0244569797,0.0016123517,-0.1468917727,0.139575839,0.1722260863,0.2771118581,-0.4392724633,0.2668138742,-0.463552773,0.1393502653,0.0349519812,0.0941721648,0.0981972963,-0.0497636311,-0.064843066,-0.007925164,0.2784922123,0.1699183285,0.2232794017,-0.4156040251,0.0980095118,-0.0268737208,0.4075753987,0.4285072386,-0.2880628705,0.0269305613,0.1438801736,0.1208021343,-0.1737537831,0.0976292789,0.0979811475,0.0191614795,-0.0420136079,0.2614986598,-0.1373831779,-0.1163995788,-0.2966488302,0.0737142265,0.3611924946,0.2738353312,0.1366084665,0.3623420894,-0.1255868226,0.0227429271,-0.147298485,0.2138113081,0.1281973869,0.0636529252,-0.2184108049,0.0210818406,0.0671528876,-0.1513585299,-0.1319669187,-0.6089710593,0.0366578996,0.4200115502,-0.1309493184,0.0015291735,0.2211922556,0.3716903925,-0.3646056354,-0.0630227178,-0.2785800993,-0.3185749054,0.027285222,0.306891799,0.0007309947,-0.1652013659,0.0680670887,-0.2167413831,0.0783068538,0.0464689396,-0.2401611358,0.1138848141,-0.0394652188,-0.4711329937,-0.0448373519,0.2415728569,0.1173040271,-0.3655276299,-0.0543694831,0.2437286228,-0.0010026359,0.1844011545,0.1564211696,0.3769582808,0.343929559,-0.062083222,0.1961861253,-0.0238543451,0.0183315948,0.1003387719,0.0159933008,0.1206837818,0.2131553739,0.2582154572,0.1318858117,-0.1313931942,0.1855686605,-0.1462122947,0.0665581152,-0.550660789,0.3187341094,-0.1287657917,0.0098425569,-0.2270540446,-0.2429974079,-0.1124460623,0.0386722833,0.4892801046,0.0632962137,0.0388072655,0.0972640589,-0.0248156004,-0.0477523915,0.2757925391,0.4354126155,0.155869633,-0.2191243321,-0.1660970151,-0.6674462557,0.0405322388,-0.0948008299,-0.1305868477,0.3083191812,-0.0039925035,-0.044723127,0.2757860124,-0.0086027738,0.3423649371,-0.2617117763,0.2088597119,-0.1278426498,-0.6246593595,0.0556058325,-0.1312208325,-0.2174884528,-0.4897020161,0.3004279137,-0.2962175012,0.0164023377,0.0226922929,-0.0562987812,0.1410411447,0.1887453645,0.541005671,0.0537836067,0.3645676076,-0.0190624334,-0.3440586627,0.1698723733,-0.0528270528,-0.1945764571,0.0221238229,-0.1828952283,0.5842747092,-0.2429691851,0.0975375473,-0.4882458448,0.4655016363,0.211640358,-0.231221199,-0.4304596484,0.1715040356,-0.2243919522,0.1434067339,-0.0908920914,-0.0167918392,-0.0500158034,0.2074625492,-0.248605296,-0.3645605147,0.5968320966,-0.1657389402,-0.0485330001,0.0402221829,0.2384093255,0.0809071064,0.1799874157,-0.4141158164,0.13146016,0.4311361015,-0.1714917421,-0.0533395037,-0.0657250956,-0.0941065326,0.3344257772,0.1151045039,0.2415604293,-0.1166984737,-0.1827428043,0.2012242377,-0.1252136379]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2746","title":"Cannot load `few-nerd` dataset","comments":"Hi @Mehrad0711,\r\n\r\nI'm afraid there is no \"canonical\" Hugging Face dataset named \"few-nerd\".\r\n\r\nThere are 2 kinds of datasets hosted at the Hugging Face Hub:\r\n- canonical datasets (their identifier contains no slash \"\/\"): we, the Hugging Face team, supervise their implementation and we make sure they work correctly by means of our test suite\r\n- community datasets (their identifier contains a slash \"\/\", where before the slash it is the username or the organization name): those datasets are uploaded to the Hub by the community, and we, the Hugging Face team, do not supervise them; it is the responsibility of the user\/organization implementing them properly if they want them to be used by other users.\r\n\r\nIn this specific case, there is no \"canonical\" dataset named \"few-nerd\". On the other hand, there are two \"community\" datasets named \"few-nerd\":\r\n- [\"nbroad\/few-nerd\"](https:\/\/huggingface.co\/datasets\/nbroad\/few-nerd)\r\n- [\"dfki-nlp\/few-nerd\"](https:\/\/huggingface.co\/datasets\/dfki-nlp\/few-nerd)\r\n\r\nIf they were properly implemented, you should be able to load them this way:\r\n```python\r\n# \"nbroad\/few-nerd\" community dataset\r\nds = load_dataset(\"nbroad\/few-nerd\", \"supervised\")\r\n\r\n# \"dfki-nlp\/few-nerd\" community dataset\r\nds = load_dataset(\"dfki-nlp\/few-nerd\", \"supervised\")\r\n```\r\n\r\nHowever, they are not correctly implemented and both of them give errors:\r\n- \"nbroad\/few-nerd\":\r\n ```\r\n TypeError: expected str, bytes or os.PathLike object, not dict\r\n ```\r\n- \"dfki-nlp\/few-nerd\":\r\n ```\r\n ConnectionError: Couldn't reach https:\/\/cloud.tsinghua.edu.cn\/f\/09265750ae6340429827\/?dl=1\r\n ```\r\n\r\nYou could try to contact their users\/organizations to inform them about their bugs and ask them if they are planning to fix them. Alternatively you could try to implement your own script for this dataset.","body":"## Describe the bug\r\n\r\nCannot load `few-nerd` dataset.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset('few-nerd', 'supervised')\r\n```\r\n\r\n## Actual results\r\n\r\nExecuting above code will give the following error:\r\n\r\n```\r\nUsing the latest cached version of the module from \/Users\/Mehrad\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/few-nerd\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53 (last modified on Wed Jun 2 11:34:25 2021) since it couldn't be found locally at \/Users\/Mehrad\/Documents\/GitHub\/genienlp\/few-nerd\/few-nerd.py, or remotely (FileNotFoundError).\r\nDownloading and preparing dataset few_nerd\/supervised (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/Users\/Mehrad\/.cache\/huggingface\/datasets\/few_nerd\/supervised\/0.0.0\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53...\r\nTraceback (most recent call last):\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 693, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1107, in _prepare_split\r\n disable=bool(logging.get_verbosity() == logging.NOTSET),\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/Users\/Mehrad\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/few-nerd\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53\/few-nerd.py\", line 196, in _generate_examples\r\n with open(filepath, encoding=\"utf-8\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/Users\/Mehrad\/.cache\/huggingface\/datasets\/downloads\/supervised\/train.json'\r\n```\r\nThe bug is probably in identifying and downloading the dataset. If I download the json splits directly from [link](https:\/\/github.com\/nbroad1881\/few-nerd\/tree\/main\/uncompressed) and put them under the downloads directory, they will be processed into arrow format correctly. \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Python version: 3.8\r\n- PyArrow version: 1.0.1\r\n","comment_length":242,"text":"Cannot load `few-nerd` dataset\n## Describe the bug\r\n\r\nCannot load `few-nerd` dataset.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset('few-nerd', 'supervised')\r\n```\r\n\r\n## Actual results\r\n\r\nExecuting above code will give the following error:\r\n\r\n```\r\nUsing the latest cached version of the module from \/Users\/Mehrad\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/few-nerd\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53 (last modified on Wed Jun 2 11:34:25 2021) since it couldn't be found locally at \/Users\/Mehrad\/Documents\/GitHub\/genienlp\/few-nerd\/few-nerd.py, or remotely (FileNotFoundError).\r\nDownloading and preparing dataset few_nerd\/supervised (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/Users\/Mehrad\/.cache\/huggingface\/datasets\/few_nerd\/supervised\/0.0.0\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53...\r\nTraceback (most recent call last):\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 693, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1107, in _prepare_split\r\n disable=bool(logging.get_verbosity() == logging.NOTSET),\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/Users\/Mehrad\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/few-nerd\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53\/few-nerd.py\", line 196, in _generate_examples\r\n with open(filepath, encoding=\"utf-8\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/Users\/Mehrad\/.cache\/huggingface\/datasets\/downloads\/supervised\/train.json'\r\n```\r\nThe bug is probably in identifying and downloading the dataset. If I download the json splits directly from [link](https:\/\/github.com\/nbroad1881\/few-nerd\/tree\/main\/uncompressed) and put them under the downloads directory, they will be processed into arrow format correctly. \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Python version: 3.8\r\n- PyArrow version: 1.0.1\r\n\nHi @Mehrad0711,\r\n\r\nI'm afraid there is no \"canonical\" Hugging Face dataset named \"few-nerd\".\r\n\r\nThere are 2 kinds of datasets hosted at the Hugging Face Hub:\r\n- canonical datasets (their identifier contains no slash \"\/\"): we, the Hugging Face team, supervise their implementation and we make sure they work correctly by means of our test suite\r\n- community datasets (their identifier contains a slash \"\/\", where before the slash it is the username or the organization name): those datasets are uploaded to the Hub by the community, and we, the Hugging Face team, do not supervise them; it is the responsibility of the user\/organization implementing them properly if they want them to be used by other users.\r\n\r\nIn this specific case, there is no \"canonical\" dataset named \"few-nerd\". On the other hand, there are two \"community\" datasets named \"few-nerd\":\r\n- [\"nbroad\/few-nerd\"](https:\/\/huggingface.co\/datasets\/nbroad\/few-nerd)\r\n- [\"dfki-nlp\/few-nerd\"](https:\/\/huggingface.co\/datasets\/dfki-nlp\/few-nerd)\r\n\r\nIf they were properly implemented, you should be able to load them this way:\r\n```python\r\n# \"nbroad\/few-nerd\" community dataset\r\nds = load_dataset(\"nbroad\/few-nerd\", \"supervised\")\r\n\r\n# \"dfki-nlp\/few-nerd\" community dataset\r\nds = load_dataset(\"dfki-nlp\/few-nerd\", \"supervised\")\r\n```\r\n\r\nHowever, they are not correctly implemented and both of them give errors:\r\n- \"nbroad\/few-nerd\":\r\n ```\r\n TypeError: expected str, bytes or os.PathLike object, not dict\r\n ```\r\n- \"dfki-nlp\/few-nerd\":\r\n ```\r\n ConnectionError: Couldn't reach https:\/\/cloud.tsinghua.edu.cn\/f\/09265750ae6340429827\/?dl=1\r\n ```\r\n\r\nYou could try to contact their users\/organizations to inform them about their bugs and ask them if they are planning to fix them. Alternatively you could try to implement your own script for this dataset.","embeddings":[-0.3042272031,-0.0985626653,0.007841493,0.5131354332,0.328545779,-0.0576422513,0.5201516747,0.2553012073,0.1991576403,0.1681091636,-0.4604775012,-0.0910487175,-0.2877757847,-0.2970745862,0.4347799718,0.0320020169,-0.02518804,-0.0349554755,0.1119609401,0.0084387101,-0.2257288843,0.2362412065,-0.2453141659,0.0727761388,-0.419000715,-0.0549565963,0.0889210105,0.4892740846,-0.0719758198,-0.6296027303,0.460472703,0.014389392,0.2181515396,0.4235012233,-0.0001091994,0.144184947,0.4189922512,-0.019266123,-0.2884244621,-0.3601104915,0.0940838531,-0.5144337416,0.1233332306,-0.2608023882,-0.406347245,0.0236194301,0.1217370182,-0.2709337175,0.490170449,0.488686204,0.288490653,0.1309136897,-0.0160082523,-0.3164563179,0.4296740592,0.0675480366,-0.1612952203,0.2777442336,0.3432971835,-0.0043072193,-0.0363358036,0.0306214392,-0.0457585603,-0.1231167316,0.39013201,0.0024617843,-0.1256728023,-0.2126000524,0.1296324283,0.4381466508,0.2064512968,-0.2134825289,-0.2328190356,-0.1929797977,0.0738780946,-0.1819786429,0.1654340327,0.0319197588,-0.1912625879,0.0741785392,-0.0262068659,-0.0630680993,-0.1722616404,0.1423974484,-0.110440284,0.1547157764,-0.150649339,0.0314562917,-0.1483332068,-0.0989765152,-0.2203025371,0.1539326459,-0.0331301875,0.2842571437,-0.3529803753,0.1692639887,0.0968754515,0.3016742766,0.2464153916,0.34680143,-0.0859935731,-0.011230737,-0.031366203,0.1054866239,0.0539561436,0.1550135016,0.3106099367,0.1709685922,0.2803744376,0.4544360042,-0.0940064788,0.0249504186,-0.2067869008,-0.1653557122,0.0443974584,0.0359360576,0.3279303312,-0.233940199,-0.4098480344,0.1534145772,0.2209211886,0.088786453,0.2583982944,0.4830324352,-0.2361518294,-0.0013515315,-0.0064407205,0.0255305003,-0.1511354744,-0.1756568849,-0.3392073214,0.1940045655,-0.080814667,0.0127045149,0.3254746199,-0.1232190207,0.3092910647,-0.1528524309,-0.0015991911,-0.1232205778,-0.1105265766,-0.1869745404,-0.0999141708,0.2051265985,-0.0700348094,0.0878425166,0.1172278821,-0.3199294508,-0.2411322892,-0.1124539599,-0.1270280331,-0.3474444747,-0.0937899575,0.2281974107,-0.2979833782,0.1321427822,-0.0671058372,0.0436785258,0.0764682218,-0.2537538111,-0.1240771115,-0.1402088106,-0.4273744822,-0.3767170906,0.6006539464,0.6969179511,-0.2007522732,-0.1508898139,-0.5498247147,-0.3447906077,0.105987072,0.3750484884,-0.2740003765,0.1505955309,-0.2850112915,0.3159655631,0.3138710558,-0.4321829379,-0.4340073168,0.2631569505,-0.2292088717,0.4060312808,-0.0842331722,0.085828267,0.1212194487,-0.1092525572,0.3998244703,0.4777024686,0.0203810725,-0.1821041256,-0.2581339777,-0.1423353404,0.1062361747,0.1785901934,0.2633876503,0.078766875,0.1548193544,0.0659918264,0.2326668501,0.208025232,0.0624008998,0.1595210284,0.2724159062,0.2057394385,0.0333990641,-0.4820550382,-0.6707289219,0.4395003617,-0.1612329334,-0.0476558656,-0.1690658629,-0.1581422389,-0.419980526,0.0298832543,-0.0987409204,-0.2591835856,0.179764092,0.2151272893,0.2117394209,0.0778774843,-0.2501817942,0.5149942636,0.0565136932,0.0629970059,-0.3822929561,0.1305712163,-0.1241281852,-0.1103749722,-0.1185646728,0.3505547345,0.2312965542,-0.185125351,0.0676198527,0.5000377893,0.0044357101,-0.1740285009,-0.0906504989,-0.4309688807,0.0396165289,-0.1043563485,0.1348259002,0.2153398097,0.1075613499,-0.0511244535,0.1082193404,0.3727986515,0.0354981348,0.2296619713,0.0288257133,-0.001395671,0.1831604987,0.0764853358,0.1574743092,-0.2301817834,0.4892827272,0.2627514303,0.1726315022,0.207790941,-0.2014015764,-0.0117084198,0.085846737,-0.0487292819,0.0544365048,0.2945663631,-0.1510484666,-0.0472731255,0.0568785742,0.3697474301,0.3691950142,0.1709932387,-0.026174238,0.0965143219,-0.0314382352,-0.0691685751,0.1280409545,0.0516601652,0.0810704231,0.2116550356,0.069557637,0.1970311254,-0.2670096457,-0.2561357021,-0.0517689362,0.3250854015,-0.3932227194,0.1741223335,-0.2071428001,-0.1251394451,-0.0495271608,0.0053799627,-0.4605352283,-0.0769572705,0.0503376685,0.1106792986,0.226770401,0.1833500862,-0.1267197728,0.1852060109,-0.3189344406,0.0570205934,-0.2792580724,0.0003344137,-0.3292350471,0.0763938129,0.3980509639,-0.0389914773,0.3228530884,-0.2194560468,-0.2569383085,-0.1921415925,-0.0463258214,-0.1839065105,0.0817929432,0.6579142213,0.1881101578,0.248079896,-0.0574077852,-0.1280138791,0.3673250675,-0.302280426,-0.1845697612,0.0297332909,-0.1068385243,0.0871499255,-0.1282043159,-0.3432548344,-0.6057456136,-0.4530194402,0.309641391,0.1487502158,0.1859292686,0.3676010668,-0.1163644344,0.092038922,-0.048556868,0.2006089389,-0.1484298408,-0.2746390998,0.2829445004,-0.4200580418,-0.3737150729,-0.1102969795,0.0587631874,0.0354793705,0.0825509056,-0.45354864,0.3604623675,-0.1324108988,0.0797707886,-0.1327626556,-0.0603401028,0.0686467588,-0.1221249402,-0.0532959439,0.0417382829,-0.1169573069,0.0133847743,-0.2853544354,0.0561641678,-0.0636483952,0.3072242141,-0.0019497601,0.4964491725,0.0580832027,-0.1058206409,0.2891736031,-0.08743117,0.4604646862,-0.1876951754,-0.437345475,0.2259440571,0.0792094916,0.102100268,0.1729521453,0.0360884257,-0.0474607423,-0.1986845881,-0.1381447911,-0.1718092412,-0.1961095184,0.1420131624,-0.1423948854,-0.0142822741,0.081840232,0.0758626238,-0.1404197812,-0.2760429084,-0.1373603642,0.2724564373,0.1131136864,0.151174441,-0.3932623863,0.1020023748,-0.3729324937,0.3129017055,0.0480136499,0.2415397763,-0.1663651317,-0.0114003224,0.1493565589,-0.0688032657,0.5504934788,-0.1515315622,0.2541805208,0.3367012739,0.1286232769,-0.7504607439,-0.0662807971,0.0233678222,0.3294961751,0.24641864,0.4543984532,-0.2505264878,-0.1369054765,0.2337837964,0.2959230542,-0.1432938278,-0.2281201482,-0.2188386768,-0.1805829257,-0.2202906311,-0.1829197705,-0.0182928052,0.269656539,-0.0675438568,-0.0292059965,-0.1527411491,0.1380231827,0.1742379218,-0.1038874611,0.3798680902,-0.0627666339,0.3176010251,0.1641202271,0.1555960625,0.0896447152,0.3713906407,-0.1347903013,-0.1864742786,-0.0433625132,0.3078988791,-0.0182066802,0.2944717705,-0.1834383756,0.055939272,0.2339202911,0.0878412351,-0.0088506117,0.1489190608,0.1992045939,-0.1655572653,-0.1703934222,-0.3570542932,0.4624353647,0.1836014539,0.163052693,0.0863773376,0.1469563395,-0.1434808671,0.0609284192,-0.0672204494,0.8010356426,-0.1922806054,0.2192895114,0.0859913528,-0.0177428629,0.7876724601,-0.2983103693,0.0326753631,-0.4788469076,-0.1975042075,0.0700773224,-0.0545484759,0.0387883745,0.1448954791,-0.0947466865,0.1931800246,-0.2254080325,-0.1533672214,-0.0042766198,0.114478372,0.0007141341,-0.1973515749,-0.1570111811,0.1364552677,-0.1990411282,0.2564733028,-0.0531497002,0.1679503918,-0.1742462516,-0.1574965268,-0.3562295139,0.0992634818,-0.2688148618,0.4265545607,0.3584544063,-0.4234371185,0.0714093,0.264308244,0.0898208544,0.1571880877,-0.1224652976,0.2339867651,-0.03810459,-0.1398673356,0.1586802006,0.0800117627,0.4458076656,-0.0480822548,-0.2589701414,-0.0948810577,-0.1892416775,-0.162511915,0.0411392711,0.1250686049,-0.0068883188,-0.1418985277,-0.2220694423,-0.1977455169,0.2485524863,-0.271815002,0.1637997925,0.3282660246,-0.2938086092,-0.0524130277,-0.0341585241,-0.1917771846,0.0028740948,0.4908804595,0.0114549724,-0.0632133782,0.2723681629,0.2820203006,-0.1826473922,-0.1906582266,-0.0402478427,0.3351081014,-0.4386791587,0.0629253983,0.0017517607,0.1089850441,-0.0831965357,0.0866381824,0.1854896396,-0.3316171169,-0.0497146435,-0.5123913884,-0.4241811931,0.0617065169,-0.1854267418,0.092766732,0.1852404624,-0.090490602,-0.0750614107,0.1424363106,-0.3271410167,0.1679143757,-0.4357074797,0.0970702097,0.0867010206,-0.0913310274,0.0261145886,0.0056106979,0.2133780122,0.1586080641,-0.1716596335,-0.2343904823,-0.2015677094,0.1274099499,-0.1638601273,0.0023351691,0.01506483,0.0391616076,-0.1514984071,-0.2010219991,0.1095058471,0.1515682936,0.029777294,-0.0893435031,-0.1283659935,0.0794053972,-0.2303027958,0.2807925344,-0.0297160465,0.0850751102,0.0738382638,0.1805206984,-0.0865613297,-0.0737274513,-0.2952458858,-0.031328395,-0.0808858797,0.1275688112,0.2736900151,-0.2907218337,-0.1748833209,0.2719103694,0.2462235689,0.3950810432,-0.1202238202,0.1272371411,0.2753774226,0.1972280592,-0.3205420971,-0.1585118622,0.2482056916,-0.1605396867,0.0703758299,0.2588653266,-0.0053533805,-0.1616201401,0.0416752957,0.1095610037,0.3554962277,-0.1286274791,0.1150353029,-0.1055823714,-0.003897764,-0.1538096666,0.066234082,0.0418538824,0.2328319252,0.7362142801,-0.1257958263,0.1040981039,-0.4056333899,0.1632438153,-0.0487122089,-0.690713048,0.1636545956,0.375107348,-0.2004184276,0.0376651213,-0.2002519369,0.4359369576,-0.4588370025,-0.0304424949,-0.182907477,0.1219528094,-0.2052728683,-0.0144398743,0.1458822936,-0.1242274046,-0.2728482187,-0.0481203832,-0.0287610441,-0.1988630593,0.3789027035,0.1070108488,-0.2630307972,-0.2015572935,0.1357066333,0.1699098796,0.2180937231,-0.3574670553,0.2478876263,-0.0779876709,-0.0820785686,0.129903093,0.1735709608,0.5013978481,0.28092134,0.0707979724,0.2174236029,0.022981571,-0.1285044551,-0.2220837176,0.2500049174,0.0155393863,0.339538902,0.2071842849,0.1622638255,-0.1437848657,-0.0835736021,0.1611058265,0.2301980704,-0.22403723,0.502263844,-0.3898711503,-0.2849798203,-0.2181780487,0.1364949495,-0.5349044204,0.0649265274,0.0866794363,0.1992293149,0.2909302115,-0.2516123056,0.110848926,-0.0459295884,0.3562528789,0.165802598,0.0648055151,-0.2455940545,-0.2477184236,-0.7577752471,0.2465137541,-0.0787993744,-0.0816405714,0.1872541457,0.1978064179,-0.2379469126,-0.0519824661,-0.0008516288,0.0256338697,0.3602199852,0.1129859537,-0.3824447691,-0.1387330294,0.0782743394,-0.0492958985,0.0544910766,-0.3725679815,0.1263582855,-0.507992208,0.0351008214,0.0988070816,-0.1641854942,0.196913287,0.0705325976,0.4982855618,-0.0376918763,0.3045731783,0.0173745751,-0.1567618996,-0.6732808352,-0.0673637316,-0.1394059956,0.3545388281,0.1344243586,0.3404389918,-0.291760385,-0.0594922602,-0.2812422514,0.2554526925,0.1446000338,0.0735234767,-0.0078475745,0.0126822898,-0.1204615384,0.1674374044,0.2154098004,0.0834819153,-0.0127500156,-0.0437047109,-0.3620735109,-0.4809339941,0.7215983868,-0.0405092165,-0.1685136259,-0.2162558436,0.410028547,0.16133973,-0.2120658606,-0.3983775675,0.1647351533,0.5024667382,0.138381958,-0.2245300412,0.0906325281,-0.2353197932,-0.0083670923,-0.0951787382,0.039505396,-0.0266619623,-0.1184293926,0.3332899213,-0.1559804082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2746","title":"Cannot load `few-nerd` dataset","comments":"Thanks @albertvillanova for your detailed explanation! I will resort to my own scripts for now. ","body":"## Describe the bug\r\n\r\nCannot load `few-nerd` dataset.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset('few-nerd', 'supervised')\r\n```\r\n\r\n## Actual results\r\n\r\nExecuting above code will give the following error:\r\n\r\n```\r\nUsing the latest cached version of the module from \/Users\/Mehrad\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/few-nerd\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53 (last modified on Wed Jun 2 11:34:25 2021) since it couldn't be found locally at \/Users\/Mehrad\/Documents\/GitHub\/genienlp\/few-nerd\/few-nerd.py, or remotely (FileNotFoundError).\r\nDownloading and preparing dataset few_nerd\/supervised (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/Users\/Mehrad\/.cache\/huggingface\/datasets\/few_nerd\/supervised\/0.0.0\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53...\r\nTraceback (most recent call last):\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 693, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1107, in _prepare_split\r\n disable=bool(logging.get_verbosity() == logging.NOTSET),\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/Users\/Mehrad\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/few-nerd\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53\/few-nerd.py\", line 196, in _generate_examples\r\n with open(filepath, encoding=\"utf-8\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/Users\/Mehrad\/.cache\/huggingface\/datasets\/downloads\/supervised\/train.json'\r\n```\r\nThe bug is probably in identifying and downloading the dataset. If I download the json splits directly from [link](https:\/\/github.com\/nbroad1881\/few-nerd\/tree\/main\/uncompressed) and put them under the downloads directory, they will be processed into arrow format correctly. \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Python version: 3.8\r\n- PyArrow version: 1.0.1\r\n","comment_length":15,"text":"Cannot load `few-nerd` dataset\n## Describe the bug\r\n\r\nCannot load `few-nerd` dataset.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset('few-nerd', 'supervised')\r\n```\r\n\r\n## Actual results\r\n\r\nExecuting above code will give the following error:\r\n\r\n```\r\nUsing the latest cached version of the module from \/Users\/Mehrad\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/few-nerd\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53 (last modified on Wed Jun 2 11:34:25 2021) since it couldn't be found locally at \/Users\/Mehrad\/Documents\/GitHub\/genienlp\/few-nerd\/few-nerd.py, or remotely (FileNotFoundError).\r\nDownloading and preparing dataset few_nerd\/supervised (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/Users\/Mehrad\/.cache\/huggingface\/datasets\/few_nerd\/supervised\/0.0.0\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53...\r\nTraceback (most recent call last):\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 693, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1107, in _prepare_split\r\n disable=bool(logging.get_verbosity() == logging.NOTSET),\r\n File \"\/Users\/Mehrad\/opt\/anaconda3\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/Users\/Mehrad\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/few-nerd\/62464ace912a40a0f33a11a8310f9041c9dc3590ff2b3c77c14d83ca53cfec53\/few-nerd.py\", line 196, in _generate_examples\r\n with open(filepath, encoding=\"utf-8\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/Users\/Mehrad\/.cache\/huggingface\/datasets\/downloads\/supervised\/train.json'\r\n```\r\nThe bug is probably in identifying and downloading the dataset. If I download the json splits directly from [link](https:\/\/github.com\/nbroad1881\/few-nerd\/tree\/main\/uncompressed) and put them under the downloads directory, they will be processed into arrow format correctly. \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Python version: 3.8\r\n- PyArrow version: 1.0.1\r\n\nThanks @albertvillanova for your detailed explanation! I will resort to my own scripts for now. ","embeddings":[-0.3042272031,-0.0985626653,0.007841493,0.5131354332,0.328545779,-0.0576422513,0.5201516747,0.2553012073,0.1991576403,0.1681091636,-0.4604775012,-0.0910487175,-0.2877757847,-0.2970745862,0.4347799718,0.0320020169,-0.02518804,-0.0349554755,0.1119609401,0.0084387101,-0.2257288843,0.2362412065,-0.2453141659,0.0727761388,-0.419000715,-0.0549565963,0.0889210105,0.4892740846,-0.0719758198,-0.6296027303,0.460472703,0.014389392,0.2181515396,0.4235012233,-0.0001091994,0.144184947,0.4189922512,-0.019266123,-0.2884244621,-0.3601104915,0.0940838531,-0.5144337416,0.1233332306,-0.2608023882,-0.406347245,0.0236194301,0.1217370182,-0.2709337175,0.490170449,0.488686204,0.288490653,0.1309136897,-0.0160082523,-0.3164563179,0.4296740592,0.0675480366,-0.1612952203,0.2777442336,0.3432971835,-0.0043072193,-0.0363358036,0.0306214392,-0.0457585603,-0.1231167316,0.39013201,0.0024617843,-0.1256728023,-0.2126000524,0.1296324283,0.4381466508,0.2064512968,-0.2134825289,-0.2328190356,-0.1929797977,0.0738780946,-0.1819786429,0.1654340327,0.0319197588,-0.1912625879,0.0741785392,-0.0262068659,-0.0630680993,-0.1722616404,0.1423974484,-0.110440284,0.1547157764,-0.150649339,0.0314562917,-0.1483332068,-0.0989765152,-0.2203025371,0.1539326459,-0.0331301875,0.2842571437,-0.3529803753,0.1692639887,0.0968754515,0.3016742766,0.2464153916,0.34680143,-0.0859935731,-0.011230737,-0.031366203,0.1054866239,0.0539561436,0.1550135016,0.3106099367,0.1709685922,0.2803744376,0.4544360042,-0.0940064788,0.0249504186,-0.2067869008,-0.1653557122,0.0443974584,0.0359360576,0.3279303312,-0.233940199,-0.4098480344,0.1534145772,0.2209211886,0.088786453,0.2583982944,0.4830324352,-0.2361518294,-0.0013515315,-0.0064407205,0.0255305003,-0.1511354744,-0.1756568849,-0.3392073214,0.1940045655,-0.080814667,0.0127045149,0.3254746199,-0.1232190207,0.3092910647,-0.1528524309,-0.0015991911,-0.1232205778,-0.1105265766,-0.1869745404,-0.0999141708,0.2051265985,-0.0700348094,0.0878425166,0.1172278821,-0.3199294508,-0.2411322892,-0.1124539599,-0.1270280331,-0.3474444747,-0.0937899575,0.2281974107,-0.2979833782,0.1321427822,-0.0671058372,0.0436785258,0.0764682218,-0.2537538111,-0.1240771115,-0.1402088106,-0.4273744822,-0.3767170906,0.6006539464,0.6969179511,-0.2007522732,-0.1508898139,-0.5498247147,-0.3447906077,0.105987072,0.3750484884,-0.2740003765,0.1505955309,-0.2850112915,0.3159655631,0.3138710558,-0.4321829379,-0.4340073168,0.2631569505,-0.2292088717,0.4060312808,-0.0842331722,0.085828267,0.1212194487,-0.1092525572,0.3998244703,0.4777024686,0.0203810725,-0.1821041256,-0.2581339777,-0.1423353404,0.1062361747,0.1785901934,0.2633876503,0.078766875,0.1548193544,0.0659918264,0.2326668501,0.208025232,0.0624008998,0.1595210284,0.2724159062,0.2057394385,0.0333990641,-0.4820550382,-0.6707289219,0.4395003617,-0.1612329334,-0.0476558656,-0.1690658629,-0.1581422389,-0.419980526,0.0298832543,-0.0987409204,-0.2591835856,0.179764092,0.2151272893,0.2117394209,0.0778774843,-0.2501817942,0.5149942636,0.0565136932,0.0629970059,-0.3822929561,0.1305712163,-0.1241281852,-0.1103749722,-0.1185646728,0.3505547345,0.2312965542,-0.185125351,0.0676198527,0.5000377893,0.0044357101,-0.1740285009,-0.0906504989,-0.4309688807,0.0396165289,-0.1043563485,0.1348259002,0.2153398097,0.1075613499,-0.0511244535,0.1082193404,0.3727986515,0.0354981348,0.2296619713,0.0288257133,-0.001395671,0.1831604987,0.0764853358,0.1574743092,-0.2301817834,0.4892827272,0.2627514303,0.1726315022,0.207790941,-0.2014015764,-0.0117084198,0.085846737,-0.0487292819,0.0544365048,0.2945663631,-0.1510484666,-0.0472731255,0.0568785742,0.3697474301,0.3691950142,0.1709932387,-0.026174238,0.0965143219,-0.0314382352,-0.0691685751,0.1280409545,0.0516601652,0.0810704231,0.2116550356,0.069557637,0.1970311254,-0.2670096457,-0.2561357021,-0.0517689362,0.3250854015,-0.3932227194,0.1741223335,-0.2071428001,-0.1251394451,-0.0495271608,0.0053799627,-0.4605352283,-0.0769572705,0.0503376685,0.1106792986,0.226770401,0.1833500862,-0.1267197728,0.1852060109,-0.3189344406,0.0570205934,-0.2792580724,0.0003344137,-0.3292350471,0.0763938129,0.3980509639,-0.0389914773,0.3228530884,-0.2194560468,-0.2569383085,-0.1921415925,-0.0463258214,-0.1839065105,0.0817929432,0.6579142213,0.1881101578,0.248079896,-0.0574077852,-0.1280138791,0.3673250675,-0.302280426,-0.1845697612,0.0297332909,-0.1068385243,0.0871499255,-0.1282043159,-0.3432548344,-0.6057456136,-0.4530194402,0.309641391,0.1487502158,0.1859292686,0.3676010668,-0.1163644344,0.092038922,-0.048556868,0.2006089389,-0.1484298408,-0.2746390998,0.2829445004,-0.4200580418,-0.3737150729,-0.1102969795,0.0587631874,0.0354793705,0.0825509056,-0.45354864,0.3604623675,-0.1324108988,0.0797707886,-0.1327626556,-0.0603401028,0.0686467588,-0.1221249402,-0.0532959439,0.0417382829,-0.1169573069,0.0133847743,-0.2853544354,0.0561641678,-0.0636483952,0.3072242141,-0.0019497601,0.4964491725,0.0580832027,-0.1058206409,0.2891736031,-0.08743117,0.4604646862,-0.1876951754,-0.437345475,0.2259440571,0.0792094916,0.102100268,0.1729521453,0.0360884257,-0.0474607423,-0.1986845881,-0.1381447911,-0.1718092412,-0.1961095184,0.1420131624,-0.1423948854,-0.0142822741,0.081840232,0.0758626238,-0.1404197812,-0.2760429084,-0.1373603642,0.2724564373,0.1131136864,0.151174441,-0.3932623863,0.1020023748,-0.3729324937,0.3129017055,0.0480136499,0.2415397763,-0.1663651317,-0.0114003224,0.1493565589,-0.0688032657,0.5504934788,-0.1515315622,0.2541805208,0.3367012739,0.1286232769,-0.7504607439,-0.0662807971,0.0233678222,0.3294961751,0.24641864,0.4543984532,-0.2505264878,-0.1369054765,0.2337837964,0.2959230542,-0.1432938278,-0.2281201482,-0.2188386768,-0.1805829257,-0.2202906311,-0.1829197705,-0.0182928052,0.269656539,-0.0675438568,-0.0292059965,-0.1527411491,0.1380231827,0.1742379218,-0.1038874611,0.3798680902,-0.0627666339,0.3176010251,0.1641202271,0.1555960625,0.0896447152,0.3713906407,-0.1347903013,-0.1864742786,-0.0433625132,0.3078988791,-0.0182066802,0.2944717705,-0.1834383756,0.055939272,0.2339202911,0.0878412351,-0.0088506117,0.1489190608,0.1992045939,-0.1655572653,-0.1703934222,-0.3570542932,0.4624353647,0.1836014539,0.163052693,0.0863773376,0.1469563395,-0.1434808671,0.0609284192,-0.0672204494,0.8010356426,-0.1922806054,0.2192895114,0.0859913528,-0.0177428629,0.7876724601,-0.2983103693,0.0326753631,-0.4788469076,-0.1975042075,0.0700773224,-0.0545484759,0.0387883745,0.1448954791,-0.0947466865,0.1931800246,-0.2254080325,-0.1533672214,-0.0042766198,0.114478372,0.0007141341,-0.1973515749,-0.1570111811,0.1364552677,-0.1990411282,0.2564733028,-0.0531497002,0.1679503918,-0.1742462516,-0.1574965268,-0.3562295139,0.0992634818,-0.2688148618,0.4265545607,0.3584544063,-0.4234371185,0.0714093,0.264308244,0.0898208544,0.1571880877,-0.1224652976,0.2339867651,-0.03810459,-0.1398673356,0.1586802006,0.0800117627,0.4458076656,-0.0480822548,-0.2589701414,-0.0948810577,-0.1892416775,-0.162511915,0.0411392711,0.1250686049,-0.0068883188,-0.1418985277,-0.2220694423,-0.1977455169,0.2485524863,-0.271815002,0.1637997925,0.3282660246,-0.2938086092,-0.0524130277,-0.0341585241,-0.1917771846,0.0028740948,0.4908804595,0.0114549724,-0.0632133782,0.2723681629,0.2820203006,-0.1826473922,-0.1906582266,-0.0402478427,0.3351081014,-0.4386791587,0.0629253983,0.0017517607,0.1089850441,-0.0831965357,0.0866381824,0.1854896396,-0.3316171169,-0.0497146435,-0.5123913884,-0.4241811931,0.0617065169,-0.1854267418,0.092766732,0.1852404624,-0.090490602,-0.0750614107,0.1424363106,-0.3271410167,0.1679143757,-0.4357074797,0.0970702097,0.0867010206,-0.0913310274,0.0261145886,0.0056106979,0.2133780122,0.1586080641,-0.1716596335,-0.2343904823,-0.2015677094,0.1274099499,-0.1638601273,0.0023351691,0.01506483,0.0391616076,-0.1514984071,-0.2010219991,0.1095058471,0.1515682936,0.029777294,-0.0893435031,-0.1283659935,0.0794053972,-0.2303027958,0.2807925344,-0.0297160465,0.0850751102,0.0738382638,0.1805206984,-0.0865613297,-0.0737274513,-0.2952458858,-0.031328395,-0.0808858797,0.1275688112,0.2736900151,-0.2907218337,-0.1748833209,0.2719103694,0.2462235689,0.3950810432,-0.1202238202,0.1272371411,0.2753774226,0.1972280592,-0.3205420971,-0.1585118622,0.2482056916,-0.1605396867,0.0703758299,0.2588653266,-0.0053533805,-0.1616201401,0.0416752957,0.1095610037,0.3554962277,-0.1286274791,0.1150353029,-0.1055823714,-0.003897764,-0.1538096666,0.066234082,0.0418538824,0.2328319252,0.7362142801,-0.1257958263,0.1040981039,-0.4056333899,0.1632438153,-0.0487122089,-0.690713048,0.1636545956,0.375107348,-0.2004184276,0.0376651213,-0.2002519369,0.4359369576,-0.4588370025,-0.0304424949,-0.182907477,0.1219528094,-0.2052728683,-0.0144398743,0.1458822936,-0.1242274046,-0.2728482187,-0.0481203832,-0.0287610441,-0.1988630593,0.3789027035,0.1070108488,-0.2630307972,-0.2015572935,0.1357066333,0.1699098796,0.2180937231,-0.3574670553,0.2478876263,-0.0779876709,-0.0820785686,0.129903093,0.1735709608,0.5013978481,0.28092134,0.0707979724,0.2174236029,0.022981571,-0.1285044551,-0.2220837176,0.2500049174,0.0155393863,0.339538902,0.2071842849,0.1622638255,-0.1437848657,-0.0835736021,0.1611058265,0.2301980704,-0.22403723,0.502263844,-0.3898711503,-0.2849798203,-0.2181780487,0.1364949495,-0.5349044204,0.0649265274,0.0866794363,0.1992293149,0.2909302115,-0.2516123056,0.110848926,-0.0459295884,0.3562528789,0.165802598,0.0648055151,-0.2455940545,-0.2477184236,-0.7577752471,0.2465137541,-0.0787993744,-0.0816405714,0.1872541457,0.1978064179,-0.2379469126,-0.0519824661,-0.0008516288,0.0256338697,0.3602199852,0.1129859537,-0.3824447691,-0.1387330294,0.0782743394,-0.0492958985,0.0544910766,-0.3725679815,0.1263582855,-0.507992208,0.0351008214,0.0988070816,-0.1641854942,0.196913287,0.0705325976,0.4982855618,-0.0376918763,0.3045731783,0.0173745751,-0.1567618996,-0.6732808352,-0.0673637316,-0.1394059956,0.3545388281,0.1344243586,0.3404389918,-0.291760385,-0.0594922602,-0.2812422514,0.2554526925,0.1446000338,0.0735234767,-0.0078475745,0.0126822898,-0.1204615384,0.1674374044,0.2154098004,0.0834819153,-0.0127500156,-0.0437047109,-0.3620735109,-0.4809339941,0.7215983868,-0.0405092165,-0.1685136259,-0.2162558436,0.410028547,0.16133973,-0.2120658606,-0.3983775675,0.1647351533,0.5024667382,0.138381958,-0.2245300412,0.0906325281,-0.2353197932,-0.0083670923,-0.0951787382,0.039505396,-0.0266619623,-0.1184293926,0.3332899213,-0.1559804082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2743","title":"Dataset JSON is incorrect","comments":"As discussed, the metadata JSON files must be regenerated because the keys were nor properly generated and they will not be read by the builder:\r\n> Indeed there is some problem\/bug while reading the datasets_info.json file: there is a mismatch with the config.name keys in the file...\r\nIn the meanwhile, in order to be able to use the datasets_info.json file content, you can create the builder without passing the name :\r\n```\r\nIn [25]: builder = datasets.load_dataset_builder(\"journalists_questions\")\r\nIn [26]: builder.info.splits\r\nOut[26]: {'train': SplitInfo(name='train', num_bytes=342296, num_examples=10077, dataset_name='journalists_questions')}\r\n```\r\n\r\nAfter regenerating the metadata JSON file for this dataset, I get the right key:\r\n```\r\n{\"plain_text\": {\"description\": \"The journalists_questions corpus (\r\n```","body":"## Describe the bug\r\n\r\nThe JSON file generated for https:\/\/github.com\/huggingface\/datasets\/blob\/573f3d35081cee239d1b962878206e9abe6cde91\/datasets\/journalists_questions\/journalists_questions.py is https:\/\/github.com\/huggingface\/datasets\/blob\/573f3d35081cee239d1b962878206e9abe6cde91\/datasets\/journalists_questions\/dataset_infos.json.\r\n\r\nThe only config should be `plain_text`, but the first key in the JSON is `journalists_questions` (the dataset id) instead.\r\n\r\n```json\r\n{\r\n \"journalists_questions\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nLook at the files.\r\n\r\n## Expected results\r\n\r\nThe first key should be `plain_text`:\r\n\r\n```json\r\n{\r\n \"plain_text\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n## Actual results\r\n\r\n```json\r\n{\r\n \"journalists_questions\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n","comment_length":109,"text":"Dataset JSON is incorrect\n## Describe the bug\r\n\r\nThe JSON file generated for https:\/\/github.com\/huggingface\/datasets\/blob\/573f3d35081cee239d1b962878206e9abe6cde91\/datasets\/journalists_questions\/journalists_questions.py is https:\/\/github.com\/huggingface\/datasets\/blob\/573f3d35081cee239d1b962878206e9abe6cde91\/datasets\/journalists_questions\/dataset_infos.json.\r\n\r\nThe only config should be `plain_text`, but the first key in the JSON is `journalists_questions` (the dataset id) instead.\r\n\r\n```json\r\n{\r\n \"journalists_questions\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nLook at the files.\r\n\r\n## Expected results\r\n\r\nThe first key should be `plain_text`:\r\n\r\n```json\r\n{\r\n \"plain_text\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n## Actual results\r\n\r\n```json\r\n{\r\n \"journalists_questions\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n\nAs discussed, the metadata JSON files must be regenerated because the keys were nor properly generated and they will not be read by the builder:\r\n> Indeed there is some problem\/bug while reading the datasets_info.json file: there is a mismatch with the config.name keys in the file...\r\nIn the meanwhile, in order to be able to use the datasets_info.json file content, you can create the builder without passing the name :\r\n```\r\nIn [25]: builder = datasets.load_dataset_builder(\"journalists_questions\")\r\nIn [26]: builder.info.splits\r\nOut[26]: {'train': SplitInfo(name='train', num_bytes=342296, num_examples=10077, dataset_name='journalists_questions')}\r\n```\r\n\r\nAfter regenerating the metadata JSON file for this dataset, I get the right key:\r\n```\r\n{\"plain_text\": {\"description\": \"The journalists_questions corpus (\r\n```","embeddings":[0.1457878947,0.0401426293,-0.052989196,0.4768224955,0.106341809,0.2229720503,0.1550247669,0.3400089443,-0.239094466,0.0082270056,0.0357939377,0.4321314096,0.0995691568,0.138274163,-0.0019154998,-0.190864414,0.0582643151,-0.0400956571,0.0514679737,0.0135782566,-0.2260534912,0.4534671605,-0.0086119361,-0.0035447541,-0.2412652671,-0.0820137113,-0.0630130991,0.1975128502,-0.1947797537,-0.2986234725,0.2238184363,-0.0007780976,-0.2992435992,0.4079233706,-0.000105777,0.1115242764,0.3177174926,-0.0546918698,-0.2763531208,-0.3126670718,-0.1179979593,-0.1995790899,-0.1801231951,-0.3108087778,-0.224489361,-0.4072842896,-0.0359193645,-0.1976814717,0.4339139163,0.2254712582,0.288928777,0.0939777195,0.3655744493,0.0138632851,0.1234832406,0.3481938839,-0.0990860835,0.1498708278,-0.0261066817,0.2428706437,0.0618628412,0.5340320468,0.0121384664,-0.1359423101,0.3161493242,-0.0881342813,-0.1343109161,-0.2345810682,0.3665965199,0.2039393783,0.5051333308,-0.3382620215,-0.2950498462,-0.3582285941,-0.0540362,0.0441976264,0.3174577951,0.1573789716,-0.0850032195,0.2246190459,-0.1565230787,-0.128318727,-0.0585856065,0.1186569482,0.0848706737,0.1772662103,-0.1670170426,0.116541326,-0.1810568422,-0.2725031674,-0.0729672238,-0.3031312525,-0.2434514761,0.1057290584,-0.1498844028,-0.1740510315,0.1840366125,-0.1226231456,0.3561204076,0.0199349467,-0.0834137425,0.030259829,-0.1282664686,0.0784088969,0.2941545248,0.0664669797,0.2865293026,0.1005908623,0.0432852842,0.3054365218,-0.0957473516,-0.0153024169,0.2361126393,-0.2029713392,-0.0158883203,-0.0455252007,0.3907977045,-0.1099102944,-0.1303402781,0.316121757,-0.3122202456,-0.137791276,-0.0887379572,0.3365528882,-0.1041637957,-0.1737522781,0.0579555221,0.1500877142,-0.0759791508,-0.1040373445,-0.2383497059,-0.0655816421,-0.0106331585,0.0740007833,0.048797749,-0.3156566024,0.5672348738,0.1093741059,-0.0547511838,-0.407150954,0.089046672,0.0703376606,0.0181154348,0.2034135014,-0.1953361928,0.2121665329,0.0029763628,-0.2547075152,-0.1380570531,0.0224881601,-0.3546994925,-0.2375632823,-0.1867226958,0.2677464783,-0.0865126401,0.1017250717,-0.1889074743,0.2865265608,0.2674294412,-0.0266436636,0.0112008266,0.1325137913,0.0306418482,-0.2472921759,0.1174986064,0.3622021377,-0.2075220048,-0.1408250928,0.1055187657,0.1386903077,-0.0218916573,0.0166031532,-0.1378027946,0.2620957494,-0.206310302,0.1565947533,-0.0173679851,-0.350587666,-0.2432554811,0.2392064482,0.0047564222,0.2748712003,0.1912210286,-0.0517991073,-0.0962357223,0.1135012731,0.3098188341,0.0791999847,0.1609462202,0.2221574336,-0.2047823519,-0.4019450247,-0.0886192471,0.0951185599,-0.1413315386,0.0513563305,0.1474854946,-0.0433722213,0.3624107838,-0.0345611423,0.0621011369,0.3707377613,-0.0492520295,-0.089791216,0.2745544314,-0.0244210009,-0.5704759955,0.1188770086,0.0124510657,0.2495318949,-0.2598682046,-0.2070997953,-0.3331198394,-0.1166165248,-0.3170585036,-0.2710929513,0.2438492924,0.1982493848,0.0582692213,-0.0029124236,-0.1119038165,-0.068286553,0.1345844269,0.1485811025,-0.8418689966,0.0498394817,-0.2129893899,0.0804177523,0.2350145429,0.1320937872,0.1785342991,-0.1546062678,-0.0937623978,0.3380415142,0.0430230461,0.2134327143,-0.0276991129,0.2834200263,0.1190686226,-0.1272166818,0.0026943104,0.328243047,0.0061988663,-0.07700409,-0.1966425329,0.5660506487,0.1556259543,0.1487797946,-0.1113553196,-0.0827234909,0.3291027844,0.0196899921,-0.1956842542,-0.3322224915,0.1007923558,-0.0287543554,0.0456186794,0.0467657298,-0.3251135349,0.1408963799,0.6347138882,0.0261397306,0.0073852828,0.0732497573,-0.2073472589,-0.0869325474,0.0815157443,0.2034507245,0.3029956818,0.2047734112,-0.2401095629,0.1650703996,0.1505337656,-0.2791338861,0.2635053992,-0.0394194722,0.1640027463,0.2611382306,0.2849473953,-0.063887924,-0.3041082621,0.1152425036,-0.037274763,0.1826793104,-0.3774652779,0.2238548547,-0.1475861669,-0.1229848489,-0.2147876024,-0.2431918532,-0.3269706964,-0.2343281209,0.1149376258,-0.0334599651,-0.2041691989,0.1390652508,0.2810516655,0.0298622083,-0.0279995073,-0.0193298943,-0.417232126,-0.2008635998,-0.1176668927,0.0509741902,0.1550177485,-0.1096065491,0.2130268812,-0.3792505562,-0.1458819658,-0.3154368401,-0.4999270737,0.252818346,-0.1550461501,0.5203313231,0.2081254721,0.2512639761,-0.0175760482,-0.2799645364,0.3669069409,0.2514350712,-0.3487674892,0.1275802404,-0.1383172423,-0.0640730932,-0.0629763976,-0.4735097885,-0.1007245705,-0.2338860184,0.3656719029,0.1424440444,0.2632218003,0.6181697249,0.0286191851,-0.1326425225,-0.3063797653,0.4378258586,-0.3224540055,-0.6928203702,0.1458532512,-0.2717385292,-0.3099220693,-0.0869625881,0.1160565987,0.2011223435,-0.3413428068,-0.5350948572,-0.2200455219,-0.0060969437,0.1768632382,0.0011520705,0.219134897,0.1107885689,0.1014790908,-0.0868371278,-0.2585103214,-0.3038441241,0.0307266228,-0.046148967,0.4130176008,-0.0031142351,0.2360234261,-0.0675421655,0.2807981074,0.4191533625,-0.1038315371,0.1003282592,-0.1449592561,0.1923125088,-0.2088390738,-0.3092525303,0.1237522066,0.0046162498,0.067847833,0.3873582482,0.1494342089,0.0962555259,-0.1694509834,-0.0993091688,-0.6170864105,-0.3313489854,-0.0105502252,0.0282097794,0.0156983677,0.2296839356,0.1860480905,-0.1981260926,-0.0875743255,0.1677749157,0.357449472,0.0115138832,0.0506597124,-0.3571624756,-0.208537519,-0.3496344686,0.2500419021,-0.1260148734,0.073884055,-0.0058296565,0.1490647495,0.0586058088,-0.1595754474,0.6097591519,-0.1528534293,0.1568210125,0.0101711918,0.2056982666,-0.0279196817,-0.0460854135,0.0388493277,0.4127778411,0.1679823995,0.7363842726,-0.443421483,-0.1499179155,0.1074354276,0.0399351418,-0.2290531546,-0.3228271902,-0.2562301755,-0.4385033548,-0.378490299,-0.2462122589,-0.1504314989,0.3239048123,0.1287073046,-0.1380190551,0.2028118372,0.147905618,-0.0279024597,0.3254078627,0.2495107353,0.1507766545,0.2005638927,0.0883947238,0.3398102224,0.4530973732,0.6804464459,0.1602423191,-0.3058867157,-0.1854248345,-0.1778449416,-0.0007927134,0.1976082176,-0.116481252,0.1874600947,0.2400570661,0.1372124553,-0.1013553664,-0.2747013867,0.3978065848,0.0894651264,-0.3039075434,-0.3379624784,0.4962552786,-0.1628100425,-0.0984534472,-0.0544000715,0.1810122728,-0.3893883228,0.497027725,-0.0196782053,0.9072646499,0.1717189848,-0.0363116376,0.1672073752,-0.2376786619,0.6262361407,-0.2435285598,0.0277066771,-0.435931623,-0.0159377046,0.0963889807,0.0022768257,0.1277958006,0.2750010192,-0.19942303,0.1043997481,-0.1537529081,0.0744260773,-0.2134490907,0.1527742296,-0.1339101046,-0.1001506671,-0.6485547423,0.2162671536,-0.1340774447,0.2843234241,-0.0337457508,-0.1659705639,-0.159703359,-0.3566295803,-0.2164776176,0.0275530275,-0.2082719058,-0.1144604161,0.4093922377,-0.1263098121,0.0895675421,-0.0023254508,0.4025224447,0.2011943609,-0.2376104295,0.1958991736,0.026019033,-0.0614312552,0.1079354808,0.0391659103,0.3096684217,-0.0769102722,-0.1865866333,-0.0448819622,-0.1477648467,-0.1405668557,-0.1647745669,-0.1890875548,-0.3345377445,-0.2339899093,-0.1377707869,-0.0966911986,0.021870451,-0.3250769079,0.2027549446,0.0051447707,-0.2541185617,0.0921353176,-0.0458557904,-0.1987372935,-0.0721427351,0.3532044888,0.2008492649,0.0216386281,0.3913569748,0.3041118383,-0.1193474531,-0.2121891379,0.1043156534,0.3070051968,-0.3281638622,0.0105466489,0.0983090997,0.0214080121,-0.0797631443,0.3965919018,0.228485018,0.0951253995,0.1138253585,-0.5196157098,-0.1872859001,0.2986902893,0.1129778624,-0.0171122085,0.1456822604,0.0790126696,-0.1356936991,0.2035490423,-0.4075133801,-0.0666664019,-0.1816495657,0.1701082885,0.1603744328,0.0799802169,0.1320695877,-0.0601366051,0.1869911253,-0.0830249339,-0.2394561917,-0.2734751105,-0.1484698206,0.1039514467,0.0708254501,-0.0348592401,-0.073783569,-0.2019998878,-0.1209958643,-0.050471127,0.1953824311,0.0575226992,0.1782270372,0.221827209,-0.1442469507,0.0418969356,0.0315932445,-0.1338862777,0.1066339165,0.2410719544,-0.0715849027,0.1054516882,-0.4267304838,0.0500704907,-0.0783985928,0.2025318742,-0.1733056754,0.4281407297,0.3976956904,-0.3567109704,0.0561244041,0.3893373907,0.3376652598,0.3005205393,-0.2515506148,0.1534404308,-0.0235086959,0.2705777884,-0.3498666883,-0.2758151591,0.2064145207,-0.1186000109,0.0424193442,0.023266485,0.0984015986,0.1095156893,-0.2196794897,0.0223608818,0.4810911417,-0.1223142222,0.0450240709,0.3552244306,-0.0936720297,0.0441394038,0.4439209104,0.261541605,0.1219502687,0.3352989554,-0.0303632952,0.2382009625,0.2425151318,0.1473100036,-0.003946777,-0.1024478823,0.4072240293,0.3074493408,0.0694724023,0.0872902125,0.1084565371,0.2912021577,-0.0150106782,-0.0678684711,-0.2059287429,0.1513767391,-0.0357830711,-0.2045345157,-0.45394665,-0.2694568336,-0.0761117935,-0.158104524,-0.3115896583,-0.2052867711,0.085333772,-0.1223781854,-0.2242933512,-0.3759257495,-0.0284937732,-0.0337759592,0.304554522,-0.2974302471,0.4509523511,0.233394891,-0.150089711,0.1736545414,0.4973620176,0.469157964,0.1394221038,0.1619143784,0.0425576866,-0.1611885875,-0.1774737686,-0.0831258669,0.3151035905,0.2066232562,-0.3417832851,0.2389533073,0.1919294298,-0.2252561003,-0.0630818307,0.3651885092,0.2124387771,-0.0102097252,-0.1332978606,-0.5007297993,-0.1973343641,-0.0978454426,-0.1543733627,-0.4194020033,-0.2621977925,0.3703799844,0.1630395204,0.2637903392,-0.1100215688,0.160548687,-0.0822906047,0.4584310353,0.2943876386,-0.131966114,-0.1145555303,-0.1008973941,-0.3746994734,0.1558760256,0.1484064162,-0.0309010874,0.0663360506,0.1782248914,0.1988012046,0.0045799073,-0.0795115456,0.0059818351,0.0064124796,0.273647815,-0.2959456146,-0.1168020219,0.0176428165,0.0351328589,0.0999381989,-0.024298938,0.278688401,-0.0703230202,0.0238399506,-0.2231590599,0.1699506342,0.1513523459,-0.1464562565,0.1923555136,0.098725155,0.1606476456,-0.2993290722,-0.3821372688,-0.1772750914,-0.2946637571,-0.2435605824,0.4900181592,0.012585097,0.5237448812,0.1454437971,-0.0820448473,-0.1965848655,0.0187697597,-0.0781199336,-0.3165510893,-0.4513062537,0.217318058,-0.1105308458,0.0250769481,0.02822081,0.3956522942,-0.0075390348,0.2020113915,-0.2545998991,-0.4512335956,0.5660297275,-0.3079728186,-0.1830760688,0.1965762824,0.1448823512,-0.1755110174,-0.308901608,-0.4869861603,0.2062945366,0.3145409524,-0.1656982899,-0.0797581598,0.0009095652,-0.2106129825,0.0502120294,0.0037505901,0.0310067181,0.2694352269,-0.2553669512,0.0571246743,-0.0974856243]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2743","title":"Dataset JSON is incorrect","comments":"Thanks!","body":"## Describe the bug\r\n\r\nThe JSON file generated for https:\/\/github.com\/huggingface\/datasets\/blob\/573f3d35081cee239d1b962878206e9abe6cde91\/datasets\/journalists_questions\/journalists_questions.py is https:\/\/github.com\/huggingface\/datasets\/blob\/573f3d35081cee239d1b962878206e9abe6cde91\/datasets\/journalists_questions\/dataset_infos.json.\r\n\r\nThe only config should be `plain_text`, but the first key in the JSON is `journalists_questions` (the dataset id) instead.\r\n\r\n```json\r\n{\r\n \"journalists_questions\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nLook at the files.\r\n\r\n## Expected results\r\n\r\nThe first key should be `plain_text`:\r\n\r\n```json\r\n{\r\n \"plain_text\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n## Actual results\r\n\r\n```json\r\n{\r\n \"journalists_questions\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n","comment_length":1,"text":"Dataset JSON is incorrect\n## Describe the bug\r\n\r\nThe JSON file generated for https:\/\/github.com\/huggingface\/datasets\/blob\/573f3d35081cee239d1b962878206e9abe6cde91\/datasets\/journalists_questions\/journalists_questions.py is https:\/\/github.com\/huggingface\/datasets\/blob\/573f3d35081cee239d1b962878206e9abe6cde91\/datasets\/journalists_questions\/dataset_infos.json.\r\n\r\nThe only config should be `plain_text`, but the first key in the JSON is `journalists_questions` (the dataset id) instead.\r\n\r\n```json\r\n{\r\n \"journalists_questions\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nLook at the files.\r\n\r\n## Expected results\r\n\r\nThe first key should be `plain_text`:\r\n\r\n```json\r\n{\r\n \"plain_text\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n## Actual results\r\n\r\n```json\r\n{\r\n \"journalists_questions\": {\r\n \"description\": \"The journalists_questions corpus (version 1.0) is a collection of 10K human-written Arabic\\ntweets manually labeled for question identification over Arabic tweets posted by journalists.\\n\",\r\n ...\r\n```\r\n\r\n\nThanks!","embeddings":[0.2763386071,-0.1870005578,-0.0442819297,0.407626003,-0.0410818495,0.2657962143,0.1620491594,0.3418322802,-0.0370079726,-0.1540858299,0.0795116499,0.4171981215,0.22155568,0.1990771145,0.0392150134,-0.1319586635,0.157664001,0.044743333,0.1465996206,-0.0875437781,-0.215912953,0.4518757164,-0.065444231,-0.0322942175,-0.1898039877,-0.1179384589,-0.1129111573,0.1811296046,-0.2165281922,-0.1850787252,0.4254396558,-0.0219698586,-0.1396995634,0.377404511,-0.0001091793,0.0024298369,0.2350562662,-0.0651909485,-0.1033503041,-0.3208450675,-0.1811095625,-0.2750478983,-0.1061829925,-0.2762231529,-0.1914093196,-0.405890137,-0.0386308171,-0.3047762215,0.5385158658,0.2817671299,0.2475179583,-0.0569340102,0.275228709,0.0009362329,0.0759283528,0.4714987576,-0.031575527,0.1360178739,-0.0353770852,0.2111119628,0.136027962,0.4819524884,0.0713413283,-0.1360028535,0.2385741323,-0.043090418,-0.2185265273,-0.2446662337,0.2821783125,0.1595931798,0.4583485425,-0.3781346679,-0.282292068,-0.2478655875,-0.0625957549,-0.0758792609,0.3254230917,0.2855404317,-0.0088154133,0.3127022982,-0.0967759192,-0.138583377,-0.0211766865,0.1688838154,0.064059414,0.0196432006,-0.1044973433,0.0667614564,-0.2197429389,-0.2365780324,-0.106675297,-0.2075814009,-0.295735985,0.1076041386,-0.2705905437,-0.2180721909,0.1002975628,-0.081337668,0.3877728879,-0.0063155084,-0.0269538611,-0.0749846697,-0.243229866,0.0930873454,0.3682030439,0.1428861171,0.2288916111,0.1613356471,-0.0073908484,0.2473770678,-0.0511233509,-0.0583564192,0.3585169613,-0.1992641389,0.0620280951,-0.075155817,0.3873098195,-0.1298118234,-0.15206559,0.2931955159,-0.3960021734,-0.1225351319,-0.1789177358,0.2649176419,-0.0624859631,-0.0285009556,0.1317119747,0.2349534333,-0.14115493,-0.202816084,-0.178932637,-0.0158058796,0.0630089194,0.0426517501,0.0230174158,-0.2234223485,0.5575652122,0.2651877999,-0.1135465205,-0.484927088,-0.0110932942,0.0140512111,0.0293477867,0.144857496,-0.0987099856,0.2542592585,0.0427970216,-0.2654460669,-0.1135866493,0.0866031349,-0.4762566388,-0.0944344178,-0.2642331719,0.2204362005,-0.1474773884,0.013644998,-0.2686561346,0.3051685691,0.2563047409,-0.0590135567,0.1265516728,0.0053086793,0.0615078174,-0.1813428551,0.1257400364,0.373139441,-0.2541909218,-0.1221368015,0.0002802596,0.1784973741,0.1479703784,0.1499563754,-0.0738711581,0.3318722546,-0.2539100945,0.2768107057,-0.0202343166,-0.4108938575,-0.2943520248,0.2642801702,-0.0241480432,0.269244194,0.1525713205,0.0526410006,0.0056878347,0.1765809357,0.3532582223,0.1082353964,0.1817341,0.2521625161,-0.2884557247,-0.3833418489,0.0240125693,0.0847835168,-0.2431942225,0.1748542041,0.1407098472,-0.1532471031,0.4298832417,0.0247981865,0.0093864193,0.3231854141,0.084752284,-0.1037233248,0.2841641605,0.1082450375,-0.5030856729,-0.0002106278,-0.0451996177,0.1759178787,-0.2606269121,-0.2217311263,-0.372458607,-0.1624333262,-0.342628926,-0.2865441144,0.1988655627,0.1243157983,0.0291213617,0.0460106917,-0.0747565329,0.0529586971,0.1025942564,0.1716515124,-0.8958239555,0.0780738741,-0.1546040475,0.1218426153,0.2447034866,0.0341553502,0.1327334195,-0.1467302144,-0.0797214732,0.3294986784,0.1534478962,0.3290278614,0.0631544888,0.124457173,0.1388436705,-0.2538464963,-0.0954887569,0.4551408887,0.0048416969,-0.0079166489,-0.237774983,0.5702602267,0.148602888,-0.0121198697,-0.2840444148,-0.0176571812,0.3221400976,-0.0045085526,-0.2032716572,-0.2754819095,0.1755943745,-0.0316554271,-0.02648356,-0.0084977513,-0.3502654731,0.1019001603,0.5568363667,0.0190842319,0.0455955826,0.2253407389,-0.1306995153,-0.1752393395,0.0795576647,0.2697910964,0.3627814353,0.1975242049,-0.3445406854,0.1653843224,0.252338618,-0.2665740252,0.2286050767,-0.1621919572,0.1038042009,0.168421492,0.3722675741,-0.0599948093,-0.2368079126,0.2044794708,-0.172859326,0.258341223,-0.4501717091,0.2120753527,-0.1841193885,-0.1165510118,-0.2577286363,-0.1211772636,-0.2169114053,-0.1676380932,0.093074739,-0.193168059,-0.1482730806,0.0781644285,0.1039282382,0.1061245129,-0.0248675551,-0.0984801427,-0.421990782,-0.0465409718,-0.1611760855,0.0142306387,0.0497973002,-0.008851381,0.2103898823,-0.3796075284,-0.1553637087,-0.2037308663,-0.3727175295,0.1749702841,-0.1227600649,0.5141067505,0.1311702281,0.3978467584,-0.2395269275,-0.185652256,0.3221943378,0.2188957483,-0.410176903,0.0905110389,-0.0901261941,0.0724487826,-0.0461093113,-0.5032836795,-0.1231115982,-0.1176912338,0.4109194875,0.1142066643,0.2469187826,0.4129779637,0.0437036231,-0.2019599378,-0.2289550602,0.3185367584,-0.3129992783,-0.6624578238,0.2252908796,-0.2194020748,-0.3357425332,-0.0827638134,0.0969848782,0.0408418067,-0.243085593,-0.6076768637,-0.1474649906,-0.0531961694,0.1341586262,-0.0584693328,0.2005397677,0.1830683947,0.0096657453,-0.0242428519,-0.2593369782,-0.3285947442,0.0356884934,0.0144969458,0.4196034372,0.0675639212,0.2860105932,-0.0432987548,0.2063853741,0.5998988748,-0.0436498858,0.1571963429,-0.2029257417,0.1540182531,-0.21242854,-0.2965936661,0.0566781461,0.0520893671,0.0165321063,0.3794567883,0.1639703363,0.0354894698,-0.1468377411,-0.14512977,-0.8064667583,-0.2377908081,0.0010477355,0.0000278749,0.0320300981,0.126957342,0.1778390855,-0.0911892727,-0.0787011236,0.1206969544,0.3572993875,0.0232731905,0.1412600428,-0.3932928145,-0.3038968146,-0.3171389699,0.2682339847,-0.1168934926,0.2722298801,0.0558237657,0.2296586037,0.0027945603,-0.1452251226,0.6620813608,-0.1058546752,0.2064929008,-0.1560207903,0.2560768127,-0.0147176264,-0.0243399795,0.0411973521,0.4568590224,0.2100914121,0.5894572139,-0.3513887227,-0.0585245192,0.0445800833,0.0542603284,-0.2630439401,-0.3154257536,-0.2048909217,-0.4716860652,-0.1956932694,-0.277294904,-0.0873712599,0.2571139336,0.1058576405,-0.0901162177,0.2556057274,0.09201435,0.0071148314,0.1600325853,0.360414207,0.0278196335,0.1819703579,0.0906604677,0.2404604703,0.3688274324,0.6738234758,0.1725099683,-0.3916538954,-0.3155934513,-0.2048491091,0.0675530061,0.1299189478,-0.0654389858,0.1471871436,0.1937481463,0.0490793139,-0.0540709421,-0.3947614133,0.4488773346,0.0379599221,-0.2776376605,-0.3739134073,0.4854815304,-0.1246425211,-0.0177218672,0.0863299817,0.1848293394,-0.4271223545,0.4670529068,-0.0915385559,0.8921182156,0.2600591481,-0.0855001956,0.1552318186,-0.3350661993,0.6181346178,-0.2015012056,0.0233249739,-0.3984470963,0.0329972841,0.047324881,-0.036693614,0.092873089,0.1118503138,-0.273070842,0.1259077489,-0.2581732571,-0.0174804088,-0.1335442215,0.1170964316,-0.0813531652,-0.0792218521,-0.6786482334,0.1681067497,-0.1945411712,0.2783643901,-0.027306959,-0.2411403656,-0.2786434591,-0.410205096,-0.2810570896,-0.0073112412,-0.1863369495,-0.1564063281,0.4151725471,-0.2232173383,0.0948675051,0.1468624771,0.3337943852,0.1524889618,-0.2664643824,0.14784199,0.0973524824,-0.0149742747,0.1850735098,-0.0030817378,0.2557352483,-0.0779516771,-0.2482703924,-0.1491046101,-0.1119105518,-0.1458811164,-0.1311544031,-0.1311022937,-0.4634431303,-0.2198876739,-0.2276371717,-0.0600866266,0.1080849841,-0.3005936444,0.1669286191,0.1569406539,-0.1629786789,0.0831947699,-0.0253686514,-0.1714862138,-0.1672056615,0.3355053961,0.1518009901,0.0114226015,0.388882935,0.3206914663,-0.0701230764,-0.1887301654,0.2758917212,0.3000027537,-0.3457103372,0.0102873966,-0.097383596,0.1676986814,-0.163330555,0.3757553399,0.1687905192,0.0393196121,0.2159075886,-0.5709873438,-0.1153083742,0.3067803085,0.0989307761,-0.0920952857,0.161207214,0.1025975123,-0.0170323197,0.1434928477,-0.3595798612,-0.0209886823,-0.0936325416,0.0699484646,0.0837639049,0.2347892076,0.1860096008,-0.0874899924,0.1338405162,-0.0500248671,-0.1343640387,-0.2385222465,-0.0727618486,0.1245944798,0.0348651335,-0.1112159863,-0.081876114,-0.2490510046,-0.1366243064,-0.0867222846,-0.0192461722,0.1307604015,0.1658524722,0.2684182525,-0.1278119385,-0.0042246818,-0.0324048325,-0.2189395875,0.1399223655,0.1538141221,-0.1024638489,0.1300072968,-0.2810641229,0.0518864617,-0.1485747099,0.1833717823,-0.0840912461,0.4844384193,0.3655016422,-0.4692210555,0.1554898024,0.3579053879,0.4087373912,0.2827601731,-0.3853673935,0.2795476615,-0.0801372603,0.2370885909,-0.3113476932,-0.1416884363,0.2537001371,0.0156096993,-0.1236363873,0.0775878504,0.0700062364,0.0232877266,-0.2491526455,-0.0002462717,0.5371670127,-0.195385024,0.1441347748,0.3077505231,-0.1093810052,0.0742116794,0.2790153027,0.2399788052,0.0510146692,0.3567482531,0.0432096496,0.1967734993,0.3604121506,0.0988831297,-0.0583216958,-0.1330813915,0.3819454312,0.3297820091,-0.0092918919,0.1011208594,0.2609081268,0.4063321352,-0.0231342297,-0.1279318631,-0.2298607528,0.0614338182,0.022099467,-0.0994303226,-0.3336335719,-0.1872003376,-0.0781503841,-0.2278051972,-0.3177321553,-0.1847807765,0.0895323083,-0.0442164652,-0.3657048047,-0.376178056,0.0167053975,-0.0497651286,0.3209091723,-0.3023212254,0.3749371469,0.2134574503,-0.127351597,0.2228875905,0.5421250463,0.4768697619,0.2172746956,0.2086269259,0.0462340191,-0.1445410401,-0.2356981188,0.0492656454,0.3607959151,0.2301449329,-0.3650345504,0.1672531515,0.1918206066,-0.1808907837,0.1152353659,0.4444406629,0.2615278661,-0.0946971923,-0.0703477561,-0.6244173646,-0.2751363814,-0.1259046048,-0.2132689655,-0.4043039382,-0.1522764564,0.1948531419,0.1811484694,0.184172526,-0.0651250258,0.1448996514,-0.0161793604,0.3213069737,0.3708244264,-0.1143952012,-0.1217101216,-0.1612512767,-0.2370437831,0.1077511162,0.1058258489,-0.0404801741,0.0268688351,0.1987745166,0.1792213917,0.0022575664,-0.0607924163,-0.0148579022,-0.1067958698,0.3512498736,-0.1696942449,-0.1493326724,0.0145510416,0.1148420572,0.0781341419,-0.0076653603,0.3810641468,-0.1060045436,0.0304743312,-0.1823344827,0.1281417012,0.0858380422,-0.2459495813,0.2117184997,0.2673537433,0.1719672382,-0.3450753391,-0.2987164557,-0.1686592549,-0.2622513771,-0.2349951267,0.5468178391,-0.0455766469,0.5609163046,0.1309108734,-0.203952834,-0.1681088656,0.0623982772,-0.0860000402,-0.3723843694,-0.4908148348,0.2063186616,-0.0859101564,-0.1344913393,-0.0102321086,0.4096661508,-0.1158768833,0.22895585,-0.3149545491,-0.4228952825,0.4345666766,-0.4283792973,-0.0903008059,0.1766343266,0.2398056835,-0.0828034878,-0.2642293274,-0.4509426355,0.1807370633,0.3166961074,-0.1220475882,-0.0432305448,0.0711853504,-0.1903482527,0.0520630889,0.0735962316,0.2209400386,0.3254185319,-0.2291904837,0.1845567971,-0.1206768453]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2742","title":"Improve detection of streamable file types","comments":"maybe we should rather attempt to download a `Range` from the server and see if it works?","body":"**Is your feature request related to a problem? Please describe.**\r\n\r\n```python\r\nfrom datasets import load_dataset_builder\r\nfrom datasets.utils.streaming_download_manager import StreamingDownloadManager\r\nbuilder = load_dataset_builder(\"journalists_questions\", name=\"plain_text\")\r\nbuilder._split_generators(StreamingDownloadManager(base_path=builder.base_path))\r\n```\r\n\r\nraises\r\n\r\n```\r\nNotImplementedError: Extraction protocol for file at https:\/\/drive.google.com\/uc?export=download&id=1CBrh-9OrSpKmPQBxTK_ji6mq6WTN_U9U is not implemented yet\r\n```\r\n\r\nBut the file at https:\/\/drive.google.com\/uc?export=download&id=1CBrh-9OrSpKmPQBxTK_ji6mq6WTN_U9U is a text file and it can be streamed:\r\n\r\n```bash\r\ncurl --header \"Range: bytes=0-100\" -L https:\/\/drive.google.com\/uc\\?export\\=download\\&id\\=1CBrh-9OrSpKmPQBxTK_ji6mq6WTN_U9U\r\n506938088174940160 yes 1\r\n302221719412830209 yes 1\r\n289761704907268096 yes 1\r\n513820885032378369 yes %\r\n```\r\n\r\nYet, it's wrongly categorized as a file type that cannot be streamed because the test is currently based on 1. the presence of a file extension at the end of the URL (here: no extension), and 2. the inclusion of this extension in a list of supported formats.\r\n\r\n**Describe the solution you'd like**\r\n\r\nIn the case of an URL (instead of a local path), ask for the MIME type, and decide on that value? Note that it would not work in that case, because the value of `content_type` is `text\/html; charset=UTF-8`.\r\n\r\n**Describe alternatives you've considered**\r\n\r\nAdd a variable in the dataset script to set the data format by hand.\r\n","comment_length":17,"text":"Improve detection of streamable file types\n**Is your feature request related to a problem? Please describe.**\r\n\r\n```python\r\nfrom datasets import load_dataset_builder\r\nfrom datasets.utils.streaming_download_manager import StreamingDownloadManager\r\nbuilder = load_dataset_builder(\"journalists_questions\", name=\"plain_text\")\r\nbuilder._split_generators(StreamingDownloadManager(base_path=builder.base_path))\r\n```\r\n\r\nraises\r\n\r\n```\r\nNotImplementedError: Extraction protocol for file at https:\/\/drive.google.com\/uc?export=download&id=1CBrh-9OrSpKmPQBxTK_ji6mq6WTN_U9U is not implemented yet\r\n```\r\n\r\nBut the file at https:\/\/drive.google.com\/uc?export=download&id=1CBrh-9OrSpKmPQBxTK_ji6mq6WTN_U9U is a text file and it can be streamed:\r\n\r\n```bash\r\ncurl --header \"Range: bytes=0-100\" -L https:\/\/drive.google.com\/uc\\?export\\=download\\&id\\=1CBrh-9OrSpKmPQBxTK_ji6mq6WTN_U9U\r\n506938088174940160 yes 1\r\n302221719412830209 yes 1\r\n289761704907268096 yes 1\r\n513820885032378369 yes %\r\n```\r\n\r\nYet, it's wrongly categorized as a file type that cannot be streamed because the test is currently based on 1. the presence of a file extension at the end of the URL (here: no extension), and 2. the inclusion of this extension in a list of supported formats.\r\n\r\n**Describe the solution you'd like**\r\n\r\nIn the case of an URL (instead of a local path), ask for the MIME type, and decide on that value? Note that it would not work in that case, because the value of `content_type` is `text\/html; charset=UTF-8`.\r\n\r\n**Describe alternatives you've considered**\r\n\r\nAdd a variable in the dataset script to set the data format by hand.\r\n\nmaybe we should rather attempt to download a `Range` from the server and see if it works?","embeddings":[-0.4683198929,-0.0467546023,-0.0979575142,0.1375530511,0.1663962454,-0.2239752114,0.1926611513,0.4674963355,0.0628534257,0.2362468243,-0.0067694117,-0.0470789447,-0.4149926007,0.2554931045,0.0666618496,-0.1503152698,0.0009771778,0.0538053922,0.2009106427,0.0540023446,-0.204148069,-0.060409423,-0.0322798118,-0.3867236376,0.1372658163,0.1212853044,0.1237365007,-0.1005211771,-0.5439122319,-0.474157244,-0.0997944027,0.3655941784,0.3324325681,0.445081681,-0.0001090667,-0.1867188662,0.3261322081,-0.2225363702,-0.3541294932,-0.3582729399,-0.3125940263,-0.0820247084,0.1180793792,-0.0586481243,-0.0832975954,-0.0529165454,0.0051997509,-0.4320715666,0.2591350675,0.4835053086,0.1480431259,0.2258724123,-0.0629004538,0.3467933834,0.3124316335,0.3924538493,-0.1639557183,0.0028929068,0.2594112754,0.1911760718,-0.0581254363,0.1383457035,-0.1698191315,0.103310965,-0.0566330105,0.0279623643,-0.2206067294,-0.5073229074,-0.0161742251,0.5063699484,0.4199713469,0.0651927963,-0.4760579169,-0.2841737866,0.1523337513,0.0003063231,0.2511096299,0.1182074249,-0.3653725684,0.1618278921,-0.3606187999,-0.0820078999,-0.2226942927,0.1248639673,-0.1203310639,0.2459635735,-0.043402236,-0.0680061206,0.0655301437,0.1497651935,0.4056751728,-0.2479868233,0.3087380528,0.0348310992,-0.0393290222,-0.2037237287,-0.1606526524,0.0232138038,0.139927879,0.2566772699,0.3086072505,0.2102730125,-0.1491983533,0.2135013491,0.0305546075,-0.0383743644,-0.0746152326,-0.0697007105,0.2174675316,0.5407654643,-0.0032599913,-0.0687255338,-0.0720643029,-0.0631122068,0.1853893548,0.1223772541,0.4066690803,-0.0915880501,-0.3426751196,-0.150570482,-0.3363800049,-0.253600508,0.1648764163,0.256169498,-0.1719850004,0.426687181,-0.0319470987,0.3137049079,0.0381046534,-0.2230694443,0.1180704385,0.065703012,0.2028967142,-0.0114821773,0.3926180303,-0.2817879915,-0.2386939377,-0.0895161629,-0.0203915052,0.070417285,0.1470766217,-0.0729336962,0.4420581162,-0.0253235754,-0.1359640807,-0.0063474299,0.0284750815,-0.1936660707,-0.263530761,0.0223022848,-0.2017050236,-0.0364400446,0.0798283145,0.2122864425,-0.196270898,-0.1137211472,-0.2630351782,0.336170435,-0.3221484721,0.0124914376,0.033613544,-0.0683287829,-0.041918356,-0.0847013518,0.2602454722,0.5462822914,-0.0980606303,-0.043349836,-0.2405214012,-0.1707349867,0.5798384547,-0.0437973626,-0.0083254362,-0.0659563169,-0.1838844419,0.1395738423,0.4809137881,-0.1480768174,-0.1279195249,0.4295482635,-0.0217103604,0.4528400004,0.608165741,-0.1676059514,0.2023663819,-0.1279568076,-0.0170407351,0.5179781318,-0.161183551,-0.0107255802,-0.3075853288,-0.3225665689,0.2332601845,0.1089288965,0.2063652277,0.1164506003,0.3067207932,-0.1911922246,0.1745196581,-0.180751577,0.094783254,-0.2880555093,0.5135552287,-0.1454490572,-0.1516349018,-0.2969726622,-0.2235253751,-0.0345313996,-0.0053609889,0.1067875773,-0.5588930845,-0.3448258936,-0.1206655875,-0.1046227217,-0.1249413565,0.0593878478,0.194429189,0.2214226276,-0.2366072237,0.0891409293,-0.2099660188,-0.0000200507,-0.1156358197,0.1792646646,0.109158054,0.2531509995,0.158861354,0.0651161149,0.4046080112,-0.0879118964,0.2208455652,-0.0490316004,-0.2288409323,0.1611866504,0.1499025971,0.4502606988,0.1342955381,0.1967618763,0.3344070911,-0.2108052522,-0.1991127133,0.5702888966,0.1838453412,0.1385805905,-0.1338566691,0.2535959482,0.072673738,0.0099196509,0.2294880599,-0.0721119195,0.1688815355,0.0099147763,-0.1288588196,-0.2819772363,0.1058612242,-0.012520642,-0.1538639814,0.0313940495,-0.1393785477,0.0554552376,0.344843179,-0.1956211776,0.0726768151,0.3713411391,-0.0236742068,-0.2375599891,0.0598795488,0.481107235,0.2880631983,0.0595036335,0.3818356693,-0.1275047213,0.1442793757,-0.2373071015,0.2923407555,0.0002398107,-0.2734445333,0.1132262871,0.0885429978,-0.0658793598,-0.3768078983,0.0178051908,-0.003829428,-0.0365524292,-0.4186419845,0.0290679652,-0.5251237154,-0.0360287502,-0.24801974,-0.1389541775,-0.0589400642,-0.1279436797,0.2919638753,0.2840204835,-0.1698154211,-0.0176665634,-0.2772929072,0.4388016462,0.0414904952,-0.206550926,-0.0967263281,0.1288214326,0.0043228273,0.1860977262,0.6305446029,0.1073832139,0.4484096169,-0.3081641793,0.0962823257,-0.624953866,-0.2172114253,0.0335564986,0.0587058291,0.4059967697,0.0810813755,0.2308991998,0.1287959367,-0.3905832469,-0.0668752044,-0.1540997922,-0.000573795,0.2530429363,-0.0182126667,0.0828403682,-0.1011619419,-0.6140940785,-0.1934521198,-0.3065446913,0.1677130163,0.0852085128,0.1408643723,0.0646377355,-0.0652581304,-0.1110421121,0.0727033243,0.2945084274,-0.0126775391,-0.3376013935,0.4186655879,-0.1411740929,-0.2259037048,0.0813955963,-0.0794175118,-0.155160293,0.7107824087,-0.0593888238,-0.1276237816,-0.0179897584,-0.0429530255,0.1628359109,-0.2291655391,0.1578842551,0.0400509462,-0.0961626768,-0.1579543501,0.4025491774,-0.1899679899,0.1545591652,0.3249195218,0.1847119629,0.0663637668,0.1059691384,0.2219640762,-0.0338329673,-0.0687349513,0.6299155354,0.1036639512,-0.0545411929,0.1048766822,0.0151892668,0.3104054332,-0.2058427185,-0.1948582828,0.3442343771,0.1803891361,-0.1539153755,-0.307698667,0.2331072688,-0.4765738845,-0.1373844445,0.3630081415,0.1020590514,0.3724046052,0.0023026918,0.0761036202,0.1107167825,-0.226127103,0.1746571213,0.5635553002,0.4419566989,-0.0829791725,-0.1235202551,0.0987975895,-0.3718417585,0.3634104431,0.3945052028,0.1895539463,-0.1086297184,-0.1734787822,0.1140322536,0.0420814492,-0.0726903751,-0.6443425417,-0.2066244185,-0.1155284718,0.209479481,0.1931023747,-0.0305153187,-0.3284288943,-0.1226562038,-0.0249666739,0.1300157309,-0.2844111621,-0.0322236158,-0.030634366,0.2318109721,0.1423321217,-0.0509388708,-0.117979601,-0.111474596,-0.2214980125,-0.0873977169,0.1017466262,-0.0670050681,-0.5361403227,-0.2599620223,-0.0246616807,0.2275419086,0.4380741119,0.2345821559,0.1742513776,-0.1007441804,0.0362849534,-0.0122565664,0.1975365579,0.4364739358,0.3614235818,0.0574342944,-0.4678523242,0.0148496721,-0.0764242485,-0.0380597748,0.2395375669,-0.206187278,0.3483464718,0.4832243025,0.3337840736,-0.4939823151,0.1397078186,0.4016370475,-0.2434222698,-0.4077292979,-0.7770465016,0.3561581373,-0.0597181432,-0.0468620993,0.2563317418,-0.017340485,-0.2218229026,0.0859864056,0.1644852161,0.9456913471,0.0091849985,-0.2788423598,0.1633626372,-0.4143882394,0.0345199592,-0.3965308964,-0.0209676959,0.0810821876,-0.1224766672,-0.3990935385,-0.0608599186,0.4578854144,0.0664067715,-0.5342295766,0.0528476797,-0.0131357322,0.2774665654,0.1687204987,0.4378339648,-0.6074435711,-0.0628857762,-0.359673202,0.1476643831,-0.1431397945,0.122768335,0.0936427191,-0.1085794792,0.0897913724,-0.0838833153,-0.209005028,-0.1965206712,-0.442707181,-0.0325450152,-0.2860933244,-0.532292068,0.2656708658,0.0808726102,0.0613881871,-0.1452040821,-0.2330865264,0.2294158489,-0.0309782289,0.2384590656,0.0420838855,-0.3452257514,0.4673247337,-0.1069621593,-0.0234375671,0.0204798244,0.0964673981,-0.2299922854,0.2641569078,-0.2156568319,0.2669360638,-0.1431202739,-0.1373481899,-0.4070110619,-0.2420139909,-0.1871687919,0.1247341335,-0.0711027458,-0.0758292452,-0.0959943607,0.2370156348,-0.1377304196,-0.2342478186,0.1525318027,-0.0162387993,-0.0108403238,0.106064029,-0.101411283,-0.3778634965,-0.3085157573,0.0213307086,-0.0295632202,-0.0453520231,-0.0875547603,0.156172365,0.0267271549,-0.1257254332,0.2406300157,-0.1747754812,0.1003924087,-0.2013082951,-0.3780125082,0.0562128536,0.0438468307,-0.3011311889,0.1458930373,-0.1954984516,-0.096965082,0.2290513068,-0.1336568147,-0.3066675663,-0.0459633842,0.0295871515,-0.0294143967,-0.1491422057,0.2581755817,0.2873902619,-0.1312532723,0.1169518605,0.0374178737,-0.1622751653,-0.1400841326,-0.1339197606,0.1999059021,0.1338274777,-0.1215900406,-0.0150664952,0.0449643768,-0.1110612452,-0.1542440206,-0.0489553399,0.0220856592,-0.0585149713,-0.0492316075,-0.0165112596,-0.2595730722,-0.3069768846,-0.0104216933,-0.2838962972,0.4272767305,0.1495513022,0.0115612559,-0.0213054623,0.018067047,-0.0427941829,0.2023470104,0.2898604274,0.2096706927,0.3697500825,0.0967873931,-0.0352621824,0.0552357696,0.1889383495,0.4606357217,-0.0955193713,-0.0238524452,0.2567819953,0.2021662295,-0.0994021595,0.2095018923,-0.1095352545,-0.3000882268,0.1835598797,0.1834964454,0.2072505057,-0.1195419058,-0.0104238559,0.0224421341,0.0373197235,-0.0289172642,0.4075638056,0.3415367603,0.1775428057,0.2778352499,-0.1512511969,-0.3581095934,-0.0371352509,-0.227177605,-0.2344058603,0.4020757079,-0.0602521226,0.2089105994,0.0475345924,-0.1020175368,-0.03085264,0.0145247988,-0.2661497593,0.5484508276,0.4545841217,0.5620856881,-0.2000777423,0.0056185396,-0.2480081618,-0.0244377535,0.137558952,0.0045949467,-0.1206164062,-0.2380751669,-0.4402795136,0.0100140497,0.128476128,0.0774767399,-0.1824942827,0.4470052421,0.1195328534,-0.1271875352,-0.0981833488,-0.215685308,0.5371815562,-0.4945754111,-0.0089918496,0.2984811664,0.1518823802,0.4640013576,0.0631291047,0.2702628076,0.2201014161,-0.1132467762,-0.0647803098,-0.0853086188,-0.1017664,0.0401635282,0.1459961534,0.1537422538,0.1427914947,0.3194053769,0.2111825049,-0.0777760595,0.2214153409,-0.0933594033,-0.203818202,-0.3882759213,0.1006372124,0.1147580966,0.2198564112,-0.3925493062,-0.4492163658,-0.2825176418,0.1275813729,0.0402086005,-0.0920502022,0.1196238995,0.0007096925,0.0661054403,0.2404972315,-0.0128430566,0.0600630641,0.0689689592,-0.0999471694,-0.2770583034,-0.7179276347,-0.2123480141,0.0953014642,-0.0521784425,-0.0723297298,-0.0959768593,-0.0011964095,0.3992811143,-0.0945917815,-0.0438439511,0.2593010366,0.2674329579,-0.2682593465,-0.3017503023,0.2872591615,0.1833754331,-0.4226427972,-0.3086621165,0.2992701232,-0.1995315552,0.0748497844,0.0298304521,0.0061795008,-0.0273449812,-0.0671395883,0.4175612032,0.2814015746,0.5260300636,-0.3578933775,-0.0784129575,-0.0858024731,0.1832464337,-0.0490067564,0.0498794541,-0.0777489766,0.6659365296,0.2426900566,-0.0229002908,-0.2919483483,0.5941691399,-0.0465590879,-0.1776711494,-0.4388510585,-0.0921091065,-0.1107232124,0.0922905207,0.1754117608,-0.013387626,-0.0666915327,-0.2000308931,-0.1337883621,-0.450107187,0.249895975,-0.300894022,-0.2121913284,-0.0282351542,0.1492001563,0.1342182904,0.4660341442,-0.599712193,0.046144411,0.2094112635,-0.2047268152,0.3059102297,0.1697495282,0.2519086599,0.0730286464,-0.1379055679,0.0984833464,0.0926523358,0.0084118396,-0.029790638,-0.2104875296]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2737","title":"SacreBLEU update","comments":"Hi @devrimcavusoglu, \r\nI tried your code with latest version of `datasets`and `sacrebleu==1.5.1` and it's running fine after changing one small thing:\r\n```\r\nsacrebleu = datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [[\"It is a guide to action that ensures that the military will forever heed Party commands\"]] # double brackets here should do the work\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\noutput: {'score': 41.180376356915765, 'counts': [11, 8, 6, 4], 'totals': [18, 17, 16, 15], 'precisions': [61.111111111111114, 47.05882352941177, 37.5, 26.666666666666668], 'bp': 1.0, 'sys_len': 18, 'ref_len': 16}\r\n```","body":"With the latest release of [sacrebleu](https:\/\/github.com\/mjpost\/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error.\r\n\r\n AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER'\r\n\r\nthis happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nsacrebleu= datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: Python 3.8.0\r\n- PyArrow version: 5.0.0\r\n","comment_length":101,"text":"SacreBLEU update\nWith the latest release of [sacrebleu](https:\/\/github.com\/mjpost\/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error.\r\n\r\n AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER'\r\n\r\nthis happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nsacrebleu= datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: Python 3.8.0\r\n- PyArrow version: 5.0.0\r\n\nHi @devrimcavusoglu, \r\nI tried your code with latest version of `datasets`and `sacrebleu==1.5.1` and it's running fine after changing one small thing:\r\n```\r\nsacrebleu = datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [[\"It is a guide to action that ensures that the military will forever heed Party commands\"]] # double brackets here should do the work\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\noutput: {'score': 41.180376356915765, 'counts': [11, 8, 6, 4], 'totals': [18, 17, 16, 15], 'precisions': [61.111111111111114, 47.05882352941177, 37.5, 26.666666666666668], 'bp': 1.0, 'sys_len': 18, 'ref_len': 16}\r\n```","embeddings":[-0.3851117492,0.180214569,-0.0012566224,-0.0679098815,0.5497623682,-0.2189842165,0.0627019629,0.3480107784,-0.2101948857,0.1243230999,0.0086564887,0.3338840604,-0.0440222882,-0.0246399418,-0.1152012944,-0.0089820018,0.1046468839,0.0598538108,0.3436266184,0.0075876298,-0.3184528649,0.1174141839,-0.1806857735,-0.0027261802,-0.2734516859,0.0927983373,-0.0575441308,-0.0664327666,-0.4653164446,-0.7149304748,0.173110649,0.0709628239,-0.0604935363,0.4433312714,-0.000109066,-0.1288360953,0.2636282146,-0.2208125442,-0.4332722723,-0.4529219866,-0.1555378437,-0.2812544405,0.0101911267,-0.1026846096,-0.0918698385,-0.3785425127,-0.0429842286,-0.0021675148,0.2421511263,0.3916675746,0.256303072,0.2788763046,0.0878582969,-0.3278107643,-0.0474800244,0.0717536137,-0.0416628681,0.2062654793,-0.1564477384,0.0919184536,0.069754228,0.1396521628,-0.2196805179,0.1190987453,0.025156932,-0.0775799155,-0.1114928201,-0.0601761714,0.1122960523,-0.047787834,0.3497331738,-0.4587009549,-0.382728368,0.1970104724,0.0173574872,-0.4542078376,-0.0254984181,-0.1006393731,0.0167840701,-0.1670011282,0.0344170108,0.0062938989,-0.0445957854,0.1155721471,-0.159620747,0.3486795425,-0.2516730428,0.0796109736,0.0394845605,-0.0542566106,-0.4969714582,0.107368283,-0.238305971,0.0230437443,-0.3772975802,-0.0650168061,-0.0828539804,-0.3516820669,0.0969957709,0.3739204109,0.0210683458,0.2644359767,0.1377246678,0.1189339012,-0.0622321777,0.5510037541,0.3230520189,0.1967891306,0.0707692876,0.2654051781,-0.0412965603,-0.0453923941,0.0042824158,-0.2698685825,0.2463271171,0.106711179,0.2557301223,-0.1353619397,-0.3971830904,0.1474429816,0.237819761,-0.0447644703,-0.1319333911,0.4307968915,-0.2377204746,0.1725683957,-0.0958559588,0.2664754987,-0.2389557958,-0.3424350023,-0.2415787727,0.0390883572,-0.3046321571,-0.1079160646,0.1816323996,0.2411848903,0.1305253953,-0.0414056592,0.4660554528,-0.2226983905,0.2804109454,-0.230205372,0.0603220426,-0.018826507,-0.119918853,0.0238396004,0.516158402,-0.610388577,0.0151248639,-0.1123187244,-0.2561796904,-0.2588275671,0.343396157,0.2879838645,-0.2528564334,-0.3133336008,-0.4201458991,0.2051164061,0.0603080615,0.0521517247,-0.001207719,-0.0462268293,-0.1970331371,-0.2614487112,0.3741241395,0.320263952,-0.1077721417,-0.1051594615,0.135828808,-0.1713021845,0.0858757123,-0.2513420284,0.0804902092,0.2804731429,0.1020598337,-0.0613257922,0.2927131057,-0.357960999,-0.5469436646,-0.0862499177,0.0579455532,-0.2599137723,0.0062024249,-0.110605292,0.5220208168,0.015823571,-0.0154365925,-0.0312897712,0.1183197945,-0.137460351,-0.2333331704,-0.4542905092,-0.0332773179,0.0685944259,0.5740548372,-0.0187632386,0.2364591807,0.3137117922,0.2896054089,0.0075053233,0.0306334458,0.1310765445,0.595379889,-0.1788109094,0.2888144553,-0.4382771552,0.0940980241,0.1595152766,0.0832317844,0.0736056566,0.0023776656,-0.2065862715,-0.5189365745,0.01281083,-0.2269904166,0.024094034,0.2367413193,0.1432618499,-0.0748593882,0.1702120453,-0.2773281932,-0.0676012337,-0.1973088235,0.2919341028,0.117482841,0.2413860857,-0.0157091059,-0.3451884985,0.2164435387,0.3111023903,0.294957608,-0.0879629478,-0.1517492831,0.4523438215,0.3043161929,0.1105450764,-0.0066715823,0.2907354534,0.0779274553,-0.2848576605,-0.0822741091,0.4827958941,0.0123695042,-0.0732993707,0.0758537129,0.3865855336,0.0324071869,0.1609558612,0.0566230379,0.2090730816,0.0145331286,-0.044533886,-0.1115885377,-0.2812256515,-0.1029724181,-0.2905320227,-0.1413835585,-0.244855538,0.0898525119,0.0704853386,0.2296757549,0.1971512437,0.1057656109,0.0084909536,-0.1468926072,0.0489098392,-0.1874101758,0.2557059824,0.2897592187,0.1384315789,0.0444402024,0.0740775838,0.0620137081,0.0123118209,0.1836156994,0.146964848,-0.3406762183,0.2635623217,0.0756593868,0.0884651467,-0.1997726709,0.0275216233,-0.0363636836,0.3163969815,-0.0755243748,0.1237703562,-0.0936109424,0.194595322,-0.0755663365,-0.2289679497,-0.0397202075,-0.201972723,-0.0094109001,-0.1007340997,0.1501653939,0.4106091559,0.0232313927,0.2186760455,0.0232600998,-0.151840359,0.1562186629,-0.1369847953,0.0295171738,0.0613607392,-0.0998253226,-0.2353059351,0.3304955363,-0.1720774174,-0.0926687866,-0.1907322705,-0.3452540934,0.0945268869,0.0551952235,0.3043855429,0.172596246,-0.1930793077,-0.0501054414,-0.0340395384,0.3287703693,-0.2375615537,0.0633648634,0.3274709284,-0.1111545712,-0.1580805331,-0.0383770168,-0.5256832242,-0.0342421457,-0.3555924296,-0.0401233099,0.3284459412,-0.0220830236,0.1165082529,0.3360468149,0.1921483576,0.1437692046,0.1019056812,-0.215679273,-0.0004950132,0.4178360105,-0.2397291362,-0.407648921,0.2128310055,-0.0815080181,0.2679240704,-0.0091952663,-0.3185970783,-0.4406927526,0.0532261543,-0.0969674885,-0.0705453008,0.1420745254,0.0384497531,0.0709822178,-0.2444822937,-0.1577114761,-0.1952126473,0.0239673052,-0.1525555998,0.288175106,-0.2231129408,-0.0932206884,0.5048215985,0.8064700961,0.3794753551,-0.1979299188,-0.052221451,0.0602564961,0.359092325,0.148836568,-0.3842603564,0.159408778,-0.3968274891,0.1137551516,0.3308050036,-0.040775083,0.0853855535,-0.0770144463,0.0330929756,0.0646861196,-0.2049100697,-0.1893302351,-0.0065218806,0.1907670945,-0.0057825474,0.158508122,-0.2624436021,-0.0677730367,0.2687343657,0.3583220541,-0.3996655047,0.0975513756,-0.3530916274,0.2276314199,-0.2536460161,0.3831908405,-0.3001421988,0.2082966268,0.1580155343,0.0468260497,-0.0848774537,-0.2429382652,0.162378341,-0.0915283784,0.0183954109,0.1837434918,0.0416089818,-0.1694903076,-0.0087943636,-0.191497907,0.1758914739,0.2852979302,0.2649518847,-0.3108689189,0.0797827691,0.1394042373,0.2556314468,-0.2582017779,0.0831907988,-0.4989479184,-0.0726923868,-0.2127495706,0.1450119317,-0.1498376131,0.0575319342,-0.0410669595,0.188061595,-0.449863106,-0.2235033065,0.3454740345,-0.1285301,0.2572482228,-0.1535093486,0.1203255504,-0.008211026,-0.0148126082,0.4275590479,0.281144172,-0.0410138927,-0.4183303118,-0.1800054461,-0.0602440126,0.3975872099,0.2043886483,-0.289737612,0.1413937211,-0.0571306273,0.2844453454,-0.3118420541,0.235816896,0.1697670072,0.115332596,-0.0460016951,-0.2965555787,0.0850745514,-0.1428477317,-0.0698100105,0.1690176427,0.1838866025,-0.0244662799,0.4124061763,0.1249623001,1.0602535009,0.246042192,-0.1515908241,0.3019744456,-0.1289837211,0.1023869887,-0.3847403526,0.1739914864,-0.2857696116,-0.2945435345,0.0644802526,-0.054352358,0.1569506377,0.0145119252,-0.4472410083,0.1225608289,-0.1259525865,0.0808319673,0.1830906123,0.2007603347,0.0973140672,-0.055793602,0.1291324049,0.257162869,-0.132704258,-0.092031464,-0.1239515319,0.1330197603,0.1029102877,-0.1203283668,-0.1604497731,0.0695376396,-0.2480961978,0.1179205179,0.1963865757,-0.1051386371,0.0773166642,0.3369865716,0.6103708744,0.2923319042,-0.1190743446,0.1135116816,-0.0668095797,-0.0486636646,-0.0258172881,-0.081144698,0.3567357063,-0.0724526793,-0.1762507558,0.1584069133,0.1312167346,0.0021974423,0.0420188121,-0.0597675294,0.3281687796,-0.2919075787,0.1386495382,0.0143049089,-0.1136594787,-0.3486387432,0.1937875003,0.0259788614,-0.3033918142,0.3734673262,0.0082068993,-0.2121108174,-0.0440055393,0.5273148417,0.0481280647,-0.3048398495,0.3915407062,-0.2656289339,-0.1693080664,-0.2985632122,-0.1753320694,-0.1514132619,-0.2239667475,0.1547166705,-0.1720940322,0.1406965554,0.2833084464,0.345992595,0.3237282336,-0.0617425516,-0.0836454108,-0.1993489563,-0.319927156,0.0277674925,0.1979090124,0.0105666742,-0.2704561353,0.0705580562,0.0064481446,0.3164363801,-0.3936085403,0.2027153969,-0.2121348977,0.0717591271,0.0528453737,-0.2563829422,0.0203098301,-0.0435715206,0.072755225,0.0188470166,-0.3767015934,-0.1912934035,-0.363417089,0.1428383738,0.1730637252,0.164602831,-0.0813721344,0.1350979805,-0.1678637415,-0.1653146148,0.0640629157,0.3893417418,-0.1728258282,0.2832193673,-0.0623189397,0.3497031927,0.0564676598,-0.1807607859,0.2696948946,0.2693118751,0.1463217288,0.0900451019,-0.129073903,-0.0316120572,-0.061511714,0.1419770867,0.2718562484,0.0525010489,0.3330371082,-0.378654331,-0.021335857,-0.0195415225,0.3227123618,0.2083992511,-0.1975060254,-0.1391373873,0.0594752729,0.2730633616,-0.3722345829,-0.0629851893,-0.1746574491,0.1714313626,0.1159251928,0.2320188284,0.323971808,-0.33073771,0.2126148045,0.1382057816,0.2896261811,-0.2799564004,0.5085770488,-0.3556696177,0.0754972249,0.0331169516,0.2941718996,0.0956103206,0.2405014634,0.2398037314,-0.1697717458,0.639395237,0.1815484315,0.0310278125,-0.1532478034,0.0329866372,0.2180238962,0.0131992027,-0.0650574863,-0.1188617647,-0.0887908414,0.4944628179,-0.170145005,-0.025069857,-0.1897780746,-0.3370122015,0.0346333161,-0.0128553296,0.0006822802,-0.2497465312,-0.2808271646,-0.2134658545,0.2274001986,-0.028344037,0.1051588058,-0.0421129838,-0.0221694801,-0.1650319397,-0.0216455869,0.2968192995,0.2125163823,-0.3541487157,-0.1242636591,0.1371029913,0.1513316929,-0.0172597766,0.3668436408,0.4245293736,0.3550616503,-0.1583877057,0.0007658497,-0.097564213,-0.0061910334,0.0173365157,0.2649992704,0.0491373278,-0.4225720763,0.2388216853,0.2324979007,-0.2302225381,-0.1320457608,0.1067541167,-0.1956955343,-0.3676971793,0.6305989623,0.1942391247,-0.0205059983,-0.16048114,-0.0636261478,-0.5861011744,-0.0131741697,0.3944792747,-0.0346701294,0.1571437865,-0.2207562774,0.1260533184,0.0516832471,0.1006783098,0.1210253164,-0.0353401564,-0.3321741819,-0.0545523465,-0.6144710779,0.2779050171,0.2257882804,0.0675522685,0.0397285223,0.2072090656,-0.1789622158,0.1978862733,0.001036764,0.0218485259,-0.0958745778,-0.2136630863,-0.6087864041,0.08250276,-0.2039693594,0.1171341836,0.0295405965,-0.2681427002,-0.2384239286,-0.3822876811,0.1150824055,0.0105734793,-0.38340801,0.2427310795,0.0775851309,0.4692176878,-0.0435158461,0.1849689633,0.0606192797,-0.1333983392,-0.1014843509,-0.0000468846,-0.2621169984,0.3020838201,-0.0905428752,0.4197928905,-0.2192395478,0.0498056412,-0.2876784801,0.1709616482,0.3208837211,-0.5431158543,-0.2115828842,0.3529495001,-0.0135220317,0.1922452748,0.0165412445,0.3320554197,-0.0361461826,0.5384612083,-0.3094086647,-0.3373981416,0.5012769699,-0.2783643901,-0.3294884264,-0.2559570074,0.6030339599,0.2151721418,-0.0788856149,-0.5270065069,0.0631274655,0.3534007668,0.0524173751,-0.2306404412,0.1097534448,-0.369494319,0.1211507469,-0.0920549706,0.3362956941,-0.0609118119,0.0165720098,0.0852710009,-0.1353364289]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2737","title":"SacreBLEU update","comments":"@bhavitvyamalik hmm. I forgot double brackets, but still didn't work when used it with double brackets. It may be an isseu with platform (using win-10 currently), or versions. What is your platform and your version info for datasets, python, and sacrebleu ?","body":"With the latest release of [sacrebleu](https:\/\/github.com\/mjpost\/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error.\r\n\r\n AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER'\r\n\r\nthis happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nsacrebleu= datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: Python 3.8.0\r\n- PyArrow version: 5.0.0\r\n","comment_length":42,"text":"SacreBLEU update\nWith the latest release of [sacrebleu](https:\/\/github.com\/mjpost\/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error.\r\n\r\n AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER'\r\n\r\nthis happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nsacrebleu= datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: Python 3.8.0\r\n- PyArrow version: 5.0.0\r\n\n@bhavitvyamalik hmm. I forgot double brackets, but still didn't work when used it with double brackets. It may be an isseu with platform (using win-10 currently), or versions. What is your platform and your version info for datasets, python, and sacrebleu ?","embeddings":[-0.3610989153,0.1312095225,-0.0056600953,-0.1026852429,0.4761410058,-0.1557441801,0.1335223466,0.3350856304,-0.0614675321,0.1180700734,0.0635044575,0.414668411,-0.0860366523,-0.1110804603,-0.0810954571,-0.0622073822,0.132537201,0.0672766715,0.3208294213,-0.0251739398,-0.2997123003,0.1096768677,-0.2105325013,0.0695405751,-0.2469487488,0.0275084693,-0.0339476764,-0.0466949157,-0.4527229071,-0.6465780139,0.2405426353,0.1384067237,-0.0376656353,0.4237561822,-0.0001106591,-0.1028935835,0.303573072,-0.1960429102,-0.4353863597,-0.4215492606,-0.151994884,-0.3628089428,0.0174232349,-0.1062814593,-0.0368342102,-0.3589268029,-0.0217846669,-0.0480082631,0.2329918742,0.4084700346,0.2475418895,0.3381185234,0.0511740781,-0.3478782177,-0.1333429813,0.0854820535,-0.045865424,0.20325239,-0.0625104606,0.0833064616,0.0917622522,0.0576002821,-0.2753891647,0.118760936,0.1333260685,-0.1312291473,-0.1327502429,-0.0118353097,0.083412528,0.0151524441,0.3650402725,-0.3823320866,-0.3735580742,0.2127169222,0.0366065875,-0.3824598491,0.009144906,-0.1123226285,-0.0093321521,-0.1700323224,0.0664357245,-0.1201473475,-0.0205986146,0.0703741759,-0.2472744286,0.4081158936,-0.2506107986,0.0469713584,0.0798312947,-0.0848306268,-0.4981856346,0.0868942365,-0.1984250247,0.0335591696,-0.340089798,-0.0250237826,-0.0763024688,-0.3464218378,0.0813369378,0.3945920169,0.048743818,0.2591862977,0.1699689478,0.20894216,-0.1301805824,0.4959830642,0.2420791984,0.220037058,0.1657265276,0.201977089,-0.066154331,-0.0687891021,-0.1287662685,-0.2297536731,0.2602816522,0.0804346278,0.3282777965,-0.0755720735,-0.513369441,0.1752422154,0.2400903702,-0.0607881695,-0.0440927558,0.3927013874,-0.2399066389,0.1500324458,-0.153525725,0.2635989785,-0.2359527349,-0.366746217,-0.2468986809,0.0178700257,-0.2878113389,-0.1278100163,0.1567808092,0.1984203309,0.1318725646,-0.0593957193,0.3452292979,-0.1717815995,0.2690585852,-0.2084930241,-0.0340363085,-0.0009564048,-0.1924007088,0.0749266669,0.4489347041,-0.731613934,-0.1049167067,-0.1289740205,-0.3792109191,-0.2902465165,0.2612980604,0.2740133703,-0.2312810719,-0.3622233272,-0.4179069698,0.2116141766,-0.0123012662,0.0543336086,-0.0246524755,-0.037327271,-0.2178055197,-0.265399307,0.3607482016,0.3120439947,-0.1571212262,-0.1470865756,0.0469976664,-0.1452475488,0.046555765,-0.1418345124,0.0897663683,0.3025019765,0.0039869021,-0.0378518663,0.3772502244,-0.3398386836,-0.4797562063,-0.0732089207,0.0153638031,-0.2394689918,0.0264777839,-0.0947665423,0.5503355265,-0.011440808,0.014831217,0.0521674044,0.118539989,-0.2458545566,-0.1919911206,-0.3879248798,0.000823137,0.0664248466,0.5991947055,0.0690891594,0.2336752415,0.3648743033,0.2155066431,0.0335599668,-0.0352310091,0.2000242025,0.625720799,-0.1667003334,0.2624933422,-0.5427994728,0.0894249454,0.139965117,0.1393514425,0.0019271293,0.0035258583,-0.146954596,-0.5078863502,0.0668669492,-0.1458019167,0.0626409575,0.1951815188,0.1813527048,-0.1051507071,0.2351346016,-0.2970178723,-0.0262000915,-0.113038525,0.2730753422,0.1114348769,0.2101031691,-0.0158116538,-0.2850116491,0.2238170207,0.2779830098,0.2404205501,-0.0966570005,-0.1334513724,0.4079409242,0.2851322889,0.1962303668,0.0060256477,0.3215174675,0.0572762825,-0.3230302334,-0.1335025281,0.4925211966,0.0297270156,-0.0744329318,0.0689793304,0.3435507715,0.1100498289,0.1087410003,0.0434566252,0.2713367939,0.0451424047,-0.0510619655,-0.0742415488,-0.2531876862,-0.0879332051,-0.2256300002,-0.0866144001,-0.1809440106,0.0871166214,0.0022355102,0.3021745682,0.1648402512,0.1735638082,0.0365758501,-0.2439983338,0.0406366624,-0.2193615437,0.2514435053,0.3448076546,0.1203919202,0.0628764853,0.0848325193,-0.0402030684,-0.0243767574,0.2261278778,0.1047915518,-0.4793460071,0.2292708904,0.035115622,0.1252464056,-0.1896268278,0.021619333,0.0023161245,0.3032466173,-0.1119338125,0.1487003863,-0.1841072887,0.2805725038,-0.0352871083,-0.1682987064,-0.0367314555,-0.1768313199,0.0316732451,0.0094892653,0.1756763905,0.4028092325,-0.015124619,0.3465368748,0.0519646257,-0.2131378949,0.1350440234,-0.0875090808,-0.0391141139,0.0214336626,0.0093573881,-0.2057677805,0.2946458757,-0.1813661158,0.0153842447,-0.3245871663,-0.3740167916,0.072844103,0.0797597468,0.4360913038,0.1809325963,-0.1662887633,-0.0249089375,-0.0552834235,0.3301906884,-0.2599101961,0.0595741384,0.3426205814,-0.0848998129,-0.2088809013,0.0111935334,-0.4363763332,-0.0135472305,-0.394340992,-0.0577675626,0.2302500308,-0.1176094115,0.0639883652,0.2232749313,0.1488091648,0.1653535217,0.1498259902,-0.1831990778,-0.1005549282,0.5080993176,-0.192951858,-0.4019785821,0.1647427082,-0.1132767797,0.2334119231,0.0597606115,-0.286100775,-0.3820084035,0.0010044457,-0.005481902,-0.0573163442,0.1456347257,0.075875476,0.0291795786,-0.2026308328,-0.1702332199,-0.2736971974,-0.0046370388,-0.1380315125,0.253814131,-0.1768521816,-0.1441241354,0.4649412036,0.8005051613,0.3342849612,-0.2882514,0.0038186598,0.0407599658,0.3907130659,0.2285870016,-0.4275695682,0.0983100608,-0.3324217796,0.1017847508,0.3215208054,-0.0136995977,0.0427913703,-0.1092609391,-0.0366049334,0.0721485466,-0.1894226819,-0.1876776069,-0.0659533069,0.195995912,0.0204110686,0.1392490268,-0.2542547584,-0.1077836379,0.3225866854,0.4397409856,-0.3692904711,0.1287804097,-0.4256255627,0.2170369029,-0.1380232722,0.3367150426,-0.3182200491,0.2477906048,0.0971426442,0.0606601909,-0.1070528999,-0.2416164577,0.2020468712,-0.1269612759,-0.0091188317,0.1536361724,0.0883920863,-0.1997353584,0.0142824464,-0.1839622557,0.1971755624,0.218321532,0.2564291656,-0.2091741264,0.0493797921,0.0505619943,0.2396542877,-0.2843267024,0.117230922,-0.4796839952,-0.0815596208,-0.0956766531,0.1743693054,-0.2040808499,0.031308271,0.0708277375,0.1279572994,-0.5701898336,-0.2751623094,0.3760758936,-0.1228814945,0.2080383152,-0.2716874778,0.1411860734,0.0255744662,0.0295209773,0.4906158447,0.2761420906,-0.068502903,-0.4534592032,-0.1442547292,-0.0579504259,0.3826785982,0.1869017929,-0.3189704716,0.1494172513,-0.0263446737,0.2732057869,-0.4339578748,0.2760280371,0.2799223065,0.153854847,-0.0894289091,-0.2863087058,0.0404827781,-0.0427840576,-0.1177441403,0.1330591589,0.1954035908,-0.0473233536,0.3297823668,0.1252315044,1.0295549631,0.2497766316,-0.1450731605,0.2048953027,-0.1225552857,0.1770911068,-0.3889919519,0.1491993368,-0.2145756632,-0.3497609496,-0.002797721,-0.0786144733,0.2461260706,-0.0812052488,-0.5005117059,0.1482785344,-0.1056859419,0.1229951829,0.2017677873,0.1952265203,-0.0193815939,-0.0214128178,0.1440591067,0.2238280624,-0.0857426748,-0.0018110197,-0.1456819922,0.134967342,0.1607819051,-0.128990829,-0.2532967329,-0.0202071685,-0.2226053476,0.156037271,0.1556514502,-0.054811012,0.0538812876,0.3182267845,0.6783843637,0.2679536939,-0.1529315561,0.174321726,-0.1608425528,-0.0177353192,-0.0590679385,-0.1327262223,0.4723624289,-0.0432027094,-0.1407504231,0.1197521165,0.1525696963,0.0058705467,-0.0600472353,-0.0463878736,0.223593697,-0.2058603764,0.1211573407,-0.0558937639,-0.04078383,-0.3475443125,0.1786043495,0.0588857643,-0.236081779,0.4113484025,0.0066553582,-0.2602074146,-0.0435961038,0.5117907524,-0.0100947646,-0.3750856519,0.3648047149,-0.3749400079,-0.1787813008,-0.3353444636,-0.1208834797,-0.1823654026,-0.1781058162,0.2070338279,-0.205636397,0.1281830817,0.2462080717,0.3060655594,0.3564524949,-0.1828078181,-0.0865226611,-0.261646837,-0.2805154622,0.0718395486,0.0743194669,-0.0332876109,-0.3151581883,0.0752737001,0.0031480396,0.3176880479,-0.3684154749,0.3009012938,-0.2032568157,-0.0049921661,0.0438934527,-0.1541394293,0.0154811749,-0.0865895972,0.0496667363,-0.0569045544,-0.4074097276,-0.1601004153,-0.377568692,0.1592615247,0.1250626147,0.1932138056,0.026591612,0.0696501434,-0.1082244515,-0.1402436793,0.0531243645,0.4404630363,-0.1807347238,0.2637675107,-0.0732503235,0.3241834939,0.0072246972,-0.160070017,0.275486201,0.24475196,0.1387672871,0.1572490931,-0.1430192143,-0.0202983264,-0.0802921876,0.2010243982,0.3011232316,-0.0488017127,0.3822992146,-0.3700407445,-0.0139493179,-0.0711531937,0.2680604458,0.3289753497,-0.1407838911,-0.0768894479,0.2062951922,0.2511440516,-0.3484357595,-0.0836771429,-0.0977708772,0.1452624649,0.1164647862,0.1948406845,0.3775343895,-0.2665851712,0.1853557825,0.0541433208,0.362493068,-0.2315696031,0.5102029443,-0.2928181291,0.1182635054,0.0621241704,0.2977143526,0.146198526,0.2570770383,0.3735711575,-0.1151320413,0.7012067437,0.2368116826,0.0497611873,-0.0939944535,-0.0247637145,0.3053646088,-0.0334919356,-0.0513393767,-0.1046817377,-0.0802335367,0.4766089618,-0.1945395619,-0.0436305515,-0.1961787939,-0.2771184742,0.0100916978,-0.1029689535,-0.032396242,-0.2655652165,-0.2995892167,-0.2010114938,0.2587386668,-0.0067119733,0.0845404193,-0.0583206117,0.0490168482,-0.1965824366,-0.0169123225,0.2903335392,0.1799899042,-0.3022832572,-0.1207773462,0.0623317696,0.0667870566,0.0497853644,0.4140069485,0.4629575014,0.3816129863,-0.2034565508,-0.0048304414,-0.1843127757,-0.034267135,-0.0191120729,0.3144958317,-0.0050156796,-0.4414651096,0.2036404759,0.2073160559,-0.1920169145,0.0766581297,0.0165652372,-0.2133695185,-0.4664843082,0.6189284325,0.1990565509,-0.0059289359,-0.1693203896,-0.1364178807,-0.5505839586,0.0358369574,0.4961479306,-0.0412426144,0.1906968355,-0.2188341171,0.112923786,0.1664034277,0.073709555,0.2321347892,-0.0000419277,-0.180194661,-0.0235993098,-0.6460703015,0.3164497316,0.1733724326,0.0906195939,0.0722572282,0.1077197269,-0.130611822,0.1696069986,0.0667063817,0.0626215562,-0.0957418606,-0.2245657444,-0.6295440793,0.0424462371,-0.224696219,0.0860632882,-0.0100434395,-0.3156252503,-0.2248573154,-0.4736671746,0.0867957622,0.0383335352,-0.2752350569,0.1972084939,0.0962522328,0.5623537302,-0.0131494077,0.2577796578,0.1092424467,-0.1006504893,-0.126923278,0.008061179,-0.2769252062,0.2525502443,-0.047370851,0.3885968029,-0.3237917423,-0.0190346092,-0.2622708082,0.2190769166,0.3211129904,-0.5104786158,-0.2979589701,0.3339656293,-0.0716052651,0.1641744971,0.0936259255,0.4034532905,-0.0830619633,0.5707306862,-0.3010047376,-0.3583704233,0.4811254144,-0.2347842008,-0.3535979092,-0.2643517852,0.6121694446,0.2160151005,-0.0946953967,-0.5384448171,0.00122475,0.4085337222,-0.0195129141,-0.2371706218,0.1611612886,-0.2878205478,0.1045451984,-0.1034156606,0.3256404102,-0.0482619479,-0.0670940876,0.150183931,-0.1195791662]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2737","title":"SacreBLEU update","comments":"You can check that here, I've reproduced your code in [Google colab](https:\/\/colab.research.google.com\/drive\/1X90fHRgMLKczOVgVk7NDEw_ciZFDjaCM?usp=sharing). Looks like there was some issue in `sacrebleu` which was fixed later from what I've found [here](https:\/\/github.com\/pytorch\/fairseq\/issues\/2049#issuecomment-622367967). Upgrading `sacrebleu` to latest version should work.","body":"With the latest release of [sacrebleu](https:\/\/github.com\/mjpost\/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error.\r\n\r\n AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER'\r\n\r\nthis happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nsacrebleu= datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: Python 3.8.0\r\n- PyArrow version: 5.0.0\r\n","comment_length":36,"text":"SacreBLEU update\nWith the latest release of [sacrebleu](https:\/\/github.com\/mjpost\/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error.\r\n\r\n AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER'\r\n\r\nthis happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nsacrebleu= datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: Python 3.8.0\r\n- PyArrow version: 5.0.0\r\n\nYou can check that here, I've reproduced your code in [Google colab](https:\/\/colab.research.google.com\/drive\/1X90fHRgMLKczOVgVk7NDEw_ciZFDjaCM?usp=sharing). Looks like there was some issue in `sacrebleu` which was fixed later from what I've found [here](https:\/\/github.com\/pytorch\/fairseq\/issues\/2049#issuecomment-622367967). Upgrading `sacrebleu` to latest version should work.","embeddings":[-0.3578206599,0.1911823153,-0.0130563695,-0.1219669282,0.4591220915,-0.2587324679,0.1023559123,0.3462676704,-0.1047268733,0.1702917069,0.0292684827,0.387314558,-0.0286068097,-0.1296031922,-0.0896717906,-0.0818408206,0.0223368928,0.1103716716,0.2108440399,-0.0650736988,-0.2887302041,0.0873918533,-0.1631985605,0.0852666646,-0.2492381483,-0.0023167096,-0.0760261863,-0.0779753551,-0.4405121207,-0.606107235,0.1186018363,0.1210269481,-0.0479873531,0.3743449152,-0.0001058455,-0.1107484773,0.305143863,-0.1988532692,-0.4026468098,-0.3009606898,-0.1276697367,-0.2364968956,0.1063758135,-0.1455808282,-0.1818398535,-0.3046250343,0.0252749342,0.0940319672,0.2512167692,0.4529435635,0.3059908152,0.4062624276,0.0230126735,-0.3524657786,-0.0199980736,-0.0242332816,-0.0131464573,0.2505607605,-0.0646727458,-0.0202181023,-0.0520274863,0.0627775341,-0.1706924438,0.1623967141,-0.0237090085,-0.0873843431,-0.1177439317,-0.0695197657,0.0494956113,-0.0812642649,0.2804450393,-0.38291201,-0.4305168688,0.2086233795,0.0718514994,-0.2917512655,-0.0081197824,-0.118237175,-0.0117069725,-0.1671764553,0.0797321945,-0.0309032053,-0.0469655395,0.0218466222,-0.2776467204,0.3991086185,-0.2167878002,0.0295434184,0.1019448638,-0.0798586756,-0.5395296812,0.1725980639,-0.1630486846,-0.0472776443,-0.3372485936,-0.0520965829,-0.035220474,-0.3039965332,0.079724595,0.2673193514,0.0789336413,0.3234218955,0.172597751,0.2088111937,-0.1074544266,0.5174102187,0.1569871008,0.2109557837,0.1481880993,0.2084748149,-0.1582446992,0.0132106123,-0.0868140534,-0.2773756385,0.1811378449,0.0504109226,0.2989059985,-0.1296224296,-0.5161451101,0.1212263703,0.3249453306,0.0074683409,-0.0784414336,0.460008651,-0.275907129,0.0837625787,-0.1150488257,0.2334918678,-0.3387951851,-0.3230011165,-0.2766523659,0.0408223793,-0.37425071,-0.0881016254,0.2087676078,0.2642614841,0.1801964492,-0.0865474418,0.3493546546,-0.1316841245,0.258661449,-0.2561262548,-0.0517255925,-0.0069767064,-0.1430691183,-0.0315161273,0.4688054621,-0.6108638644,-0.0566984527,-0.1254149824,-0.3261383176,-0.2362999618,0.2884625196,0.3258454204,-0.2371550053,-0.2956566513,-0.3317900896,0.2374335825,0.0726340786,0.027764719,-0.0176040698,-0.122636795,-0.2025411278,-0.2438831478,0.3008957803,0.3525292873,-0.1587591618,-0.1621088684,0.1584931165,-0.1695159525,0.0624839589,-0.0763147175,0.0115527418,0.3146710992,0.0382410213,-0.0237806197,0.2955042422,-0.2857785523,-0.5236398578,-0.0819974169,-0.0154975886,-0.2500258684,-0.0530064143,-0.0627646521,0.6092697382,0.0600496829,-0.0038255965,-0.0115368627,0.1557662636,-0.0427967533,-0.2675361335,-0.3750508428,-0.0079508908,0.0790766925,0.5792466998,0.0346505754,0.2244628966,0.1579549909,0.1856236309,0.0233877096,0.0189662781,0.2308950126,0.6588175297,-0.1663859934,0.247043848,-0.4124954045,0.0987278447,0.1414683759,-0.0113154342,0.0937688276,0.0519420467,-0.1947705895,-0.4691921771,-0.0183418598,-0.2268976122,0.0119521422,0.2715171576,0.1378232688,-0.0797377303,0.2032039464,-0.2275076956,0.0311874021,-0.1075829268,0.2439023256,0.1114137322,0.2018918395,-0.1064896807,-0.3009128273,0.3025524914,0.2251977921,0.2552266419,-0.026972048,-0.1187357083,0.4040307999,0.2718547285,0.1211067364,0.1103964373,0.2345688492,0.0470449403,-0.3110350668,-0.1486893743,0.4316517413,0.0484483428,-0.0514895581,0.0397672914,0.4062992334,0.1410156786,0.0460959598,0.067911312,0.1683560163,0.0360379927,-0.1195729524,-0.0889129713,-0.1963648051,-0.0799371526,-0.1591659784,-0.0623397231,-0.1417454928,0.0780003294,0.0879584253,0.3270541131,0.248662591,0.2066837102,0.076920718,-0.1798181534,-0.0212408919,-0.17702052,0.2161441892,0.3647597432,0.1700563878,0.0438454151,0.0763294026,-0.0463225432,-0.0378088616,0.179175958,0.0653135777,-0.4200987816,0.1537058055,0.1766033918,0.0890044197,-0.1963602304,0.0585504919,0.0622049607,0.2806884348,-0.0912156478,0.1762315631,-0.1289596409,0.3334774971,0.0158991348,-0.2181564569,-0.0460447557,-0.170931235,0.1734342575,0.0025810634,0.0736619085,0.3721660078,0.0109238736,0.3784334064,0.2001804113,-0.2112026513,0.1066684648,-0.2119735032,-0.0180871375,0.0760566518,-0.1197405308,-0.0421262793,0.3133828342,-0.1565406173,-0.0128217386,-0.3236809075,-0.3269429803,0.1113748699,0.0142587032,0.455611378,0.1397294104,-0.1708728969,-0.1437091231,-0.0577829257,0.3358262777,-0.2341038287,-0.0047566285,0.3123632967,-0.1245187744,-0.1732459068,-0.1278391331,-0.4561173618,0.0026664534,-0.3819520175,-0.0850622058,0.1646662205,-0.0369072109,0.0480373465,0.2475905418,0.1148171499,0.0956451893,0.0587986745,-0.2714680135,-0.132200256,0.5068076253,-0.3097522855,-0.4754901528,0.1518469155,-0.1226331145,0.2845422626,0.0372845605,-0.2821929157,-0.5299497247,0.0195122976,0.0280244816,-0.0609569177,0.0801196769,0.1169326678,0.0231861994,-0.2826243043,-0.1794141382,-0.2012349367,0.1383204758,-0.0638355538,0.2962293327,-0.2882066965,-0.1492156088,0.3420609236,0.8126741648,0.2221950144,-0.2842615247,0.028703196,-0.0046370728,0.3649331629,0.1545112431,-0.3445688188,0.1804588586,-0.2892850041,0.1449608058,0.3699704111,-0.0995687917,0.0257931612,-0.0707531422,0.0904710442,0.0109626269,-0.275506556,-0.1829292774,0.0704107806,0.173681587,0.0268890336,0.1381925792,-0.252094537,-0.0163275134,0.2843599319,0.3995743096,-0.278247565,0.0710222051,-0.2709499896,0.1875791699,-0.1672024131,0.2655047178,-0.3068066537,0.2081258446,0.10713882,0.0699549988,-0.064852111,-0.2380272001,0.2308620512,0.0154371578,0.0458044931,0.1292300224,0.0159412883,-0.1366559565,0.0047389762,-0.2175615728,0.1586100608,0.2181323916,0.2016923279,-0.3195238709,0.0964602232,0.0617580079,0.1553493589,-0.342092365,0.0923144966,-0.5545898676,-0.1446536779,-0.1431479454,0.2754791975,-0.2147542238,0.0287505444,-0.0732056499,0.1491055191,-0.4761307836,-0.2124082595,0.2964614928,-0.0050976919,0.2399255782,-0.2561050653,0.1337241381,-0.0347923711,0.0315443948,0.4541696906,0.3160787821,0.0361009277,-0.3791028857,-0.0587961525,-0.0066062883,0.2954760194,0.1132595763,-0.2193434983,0.126214698,-0.097373046,0.3588213623,-0.2885735333,0.3199592829,0.2117467672,0.0973447189,-0.0223521721,-0.1476567239,0.059477631,-0.209419027,-0.0982036516,0.0274583735,0.1805227995,-0.0267300364,0.3437111676,0.1284795851,0.9728658795,0.1545346379,-0.169800505,0.3814636171,-0.0299465097,0.1575747132,-0.3214428127,0.182210952,-0.2842508256,-0.2830302119,0.0590342917,-0.0460111238,0.1815724522,-0.055311434,-0.5791159272,0.1761922687,-0.0546644777,0.0002065218,0.2185235322,0.1534944475,0.0543319359,-0.0398474447,0.067912288,0.2623175085,-0.1888812333,0.0354267918,-0.1071687192,0.0517594479,0.083190091,-0.1294439286,-0.074579142,0.0736999586,-0.296859473,0.1341708452,0.0989097506,-0.0511453487,0.0143608535,0.2612830997,0.549726069,0.3040993512,-0.1275848299,0.2080315948,-0.1019929573,-0.0971850678,-0.0760592744,-0.1031070501,0.4004956782,-0.0556434542,-0.2246282101,0.1341568679,0.1070689857,-0.031657178,-0.0437450595,-0.0109527251,0.2206584662,-0.2373544127,0.0671290755,-0.0911274776,-0.1662378013,-0.325586915,0.2039810419,0.0799049065,-0.2308877558,0.426597774,0.1205413267,-0.282465905,-0.0549682528,0.470836103,-0.022267485,-0.2782503068,0.3541187942,-0.3069104552,-0.1881445348,-0.3704710603,-0.3024503291,-0.1584444344,-0.2403984666,0.1294461489,-0.1099829748,0.0756377727,0.3476689458,0.2972694635,0.2949798405,-0.20069547,-0.0764429793,-0.1964136809,-0.3316259384,0.0594496168,0.1357143819,-0.0173006766,-0.1900379807,0.1527743787,0.0276921261,0.3771696091,-0.431058377,0.2393395603,-0.3082424998,0.1350147575,0.0507780761,-0.2673182786,0.0164036788,0.0019291758,0.0846145749,0.0695557445,-0.3597666919,-0.2404734045,-0.3186885417,0.1150120348,0.1184397712,0.2952625155,-0.0802582726,0.134675473,-0.1008252278,-0.1417819262,0.0732925758,0.3997766376,-0.1848661453,0.260012269,-0.0433524922,0.2943043411,0.1377822459,-0.0841608718,0.2218560129,0.197998181,0.1352068335,0.1358502805,-0.2521337271,-0.0349322855,0.0080188951,0.1639234424,0.2954921126,0.0450421646,0.3262088895,-0.3317493796,-0.0472830907,-0.1157065183,0.2377041578,0.1594416499,-0.1561021656,-0.1854166687,0.1604343504,0.3080506027,-0.3771663308,-0.0732603297,-0.169832468,0.106968984,0.2118207663,0.2212181985,0.2498070747,-0.2954524159,0.2389974147,0.1047464013,0.3551614285,-0.1147335917,0.3196256757,-0.4188706577,0.0481441617,0.0339866988,0.3200391531,0.1382942945,0.2413092256,0.2761481702,-0.1308549047,0.7322460413,0.2012879699,0.0283845905,-0.1049313471,-0.0192275662,0.192839548,0.0340167135,0.0155687463,-0.1041196883,-0.1250298768,0.4388267398,-0.20569399,-0.0482139215,-0.2144743651,-0.2102614939,0.0167076029,-0.0024717108,-0.0432861112,-0.2329721302,-0.201605022,-0.1873738915,0.2655162215,-0.0032351566,0.1305630803,-0.0682545081,0.0290058255,-0.1451794654,-0.0951116085,0.2740392387,0.1035442799,-0.2165935487,-0.0157151893,0.0546055362,0.1020379364,0.0257496741,0.3504705727,0.3399099112,0.4087024927,-0.2074266374,0.0316676907,-0.175041303,0.0038703505,-0.0499920547,0.1995204836,0.0798431486,-0.4812790155,0.2994626462,0.2601359785,-0.2384610176,-0.0848105699,0.0219516903,-0.1837273836,-0.3299479783,0.4343504906,0.2688827813,0.0057522366,-0.1743471324,-0.1197386906,-0.6330940723,0.0110938726,0.3685867786,-0.0753886029,0.1884405762,-0.2897850275,0.1485455334,0.0545619354,0.1447160393,0.1944023073,0.0591149852,-0.2636816204,-0.026903417,-0.7193476558,0.3840250075,0.2125400305,0.087372072,0.0387340486,0.2004411817,-0.1806336492,0.1441377401,0.0589213446,0.0260143913,-0.0570512675,-0.2295215875,-0.6657166481,0.1136710271,-0.3540846705,0.0968133882,0.0587343462,-0.3082034588,-0.1885084063,-0.3494607806,0.1655978411,0.0418650247,-0.2623753548,0.1670268029,0.051426921,0.5181002021,-0.0106914574,0.2291044742,0.0399087891,-0.0865099579,-0.133088842,-0.0327972397,-0.1995065957,0.1676523089,-0.0212184228,0.3710514903,-0.1306427717,0.0637276098,-0.2664960325,0.2270196676,0.3711467087,-0.4613429606,-0.1648534387,0.3245267868,0.0407230407,0.1466799229,0.0596735775,0.3735426068,-0.0053757336,0.5238209367,-0.3329616785,-0.3023942113,0.545835495,-0.3192130923,-0.3209404051,-0.2270155549,0.6169535518,0.2041413486,-0.1760114431,-0.5343707204,0.0416532978,0.3197528422,0.0008442955,-0.16479747,0.1102187634,-0.3402335942,0.0909263566,-0.1789981425,0.406336695,-0.0438836291,0.0316482149,0.0902723894,-0.1580961645]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2737","title":"SacreBLEU update","comments":"It seems that next release of `sacrebleu` (v2.0.0) will break our `datasets` implementation to compute it. See my Google Colab: https:\/\/colab.research.google.com\/drive\/1SKmvvjQi6k_3OHsX5NPkZdiaJIfXyv9X?usp=sharing\r\n\r\nI'm reopening this Issue and making a Pull Request to fix it.","body":"With the latest release of [sacrebleu](https:\/\/github.com\/mjpost\/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error.\r\n\r\n AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER'\r\n\r\nthis happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nsacrebleu= datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: Python 3.8.0\r\n- PyArrow version: 5.0.0\r\n","comment_length":33,"text":"SacreBLEU update\nWith the latest release of [sacrebleu](https:\/\/github.com\/mjpost\/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error.\r\n\r\n AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER'\r\n\r\nthis happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nsacrebleu= datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\r\nreferences = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\r\nresults = sacrebleu.compute(predictions=predictions, references=references)\r\nprint(results)\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.11.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: Python 3.8.0\r\n- PyArrow version: 5.0.0\r\n\nIt seems that next release of `sacrebleu` (v2.0.0) will break our `datasets` implementation to compute it. See my Google Colab: https:\/\/colab.research.google.com\/drive\/1SKmvvjQi6k_3OHsX5NPkZdiaJIfXyv9X?usp=sharing\r\n\r\nI'm reopening this Issue and making a Pull Request to fix it.","embeddings":[-0.3739028871,0.2329158187,0.0166788381,-0.1529369652,0.3844794631,-0.228001982,0.1411173046,0.3971856236,-0.1434217542,0.1172362044,0.0634808242,0.367531687,-0.0747411028,-0.0431298316,-0.0400885046,0.0182079747,0.1286734492,0.0568156801,0.2899592519,0.0479621105,-0.300147146,0.1294311881,-0.1781707853,-0.0137694543,-0.229881078,0.0428284891,-0.0431472957,-0.0255484059,-0.4351977408,-0.6946918368,0.2592553496,0.1504914165,-0.1331426054,0.4043775201,-0.0001106694,-0.0649239719,0.2988054156,-0.1715772152,-0.4254031479,-0.4097581804,-0.1907384396,-0.3609961271,0.0138522377,-0.0621109158,0.0258586202,-0.3896156251,-0.0413792729,0.0150133381,0.2358204424,0.3832374215,0.2394982576,0.430285275,-0.0015964248,-0.3362753689,-0.1122719795,0.0617029145,-0.047362607,0.1549793929,-0.0962243304,0.0600672923,-0.0194021594,0.0392365605,-0.186813578,0.1169505566,0.0850505829,-0.1944031566,-0.2114242166,-0.0239331648,0.0887817144,0.0142879719,0.3667894006,-0.3311085999,-0.4713633358,0.2233430147,0.0124881687,-0.3764422238,0.0764675438,-0.1040389091,-0.0425245799,-0.1320866495,0.0299374238,-0.0979778469,-0.0454691686,0.0143784024,-0.2644269466,0.3045631051,-0.2460582405,0.031894356,0.0465924069,-0.0460272171,-0.4297217131,0.1479371786,-0.2655233145,-0.0081698149,-0.4187926054,-0.0802067667,-0.0276516937,-0.2785298526,0.1515285224,0.4640417397,0.0794282258,0.2926971018,0.1231563091,0.160395056,-0.0903690904,0.5370455384,0.1759420782,0.1939258277,0.206447497,0.2096591294,-0.0102406796,-0.0219156723,0.0056200782,-0.2501822114,0.1966059357,0.0381130204,0.2866578996,-0.102828823,-0.467641592,0.1577839106,0.2295634151,-0.0384413898,-0.0725245103,0.3364729285,-0.2498460859,0.2119585127,-0.19204171,0.2145524919,-0.2049995661,-0.3765915334,-0.2550670505,-0.0666445047,-0.3085423708,-0.0407444835,0.1703774184,0.1910991967,0.172594294,-0.0540474169,0.3929164112,-0.1610344052,0.2505149841,-0.1740705371,0.0372872762,0.0719731227,-0.1483312696,-0.0247008689,0.3736082315,-0.6256074905,-0.0404642522,-0.2014552057,-0.329965055,-0.2798798084,0.2722164094,0.2741959095,-0.2737762332,-0.2794387639,-0.4384195507,0.2355144471,-0.0362764634,0.0633847713,-0.0532350242,-0.0661690459,-0.2227806598,-0.2717621624,0.3008926809,0.3167229295,-0.1638348848,-0.1671287566,0.1063645706,-0.1123757958,0.0829697698,-0.1165450588,0.0046931482,0.2391459644,0.0060909223,-0.020558333,0.2737216949,-0.2885116637,-0.5547575951,-0.0408842228,0.0688699484,-0.2400221527,-0.0268476401,-0.118897602,0.620539546,-0.0705594793,-0.0200987514,0.0195240062,0.06320595,-0.246159181,-0.1765811592,-0.4268577993,0.0149049275,0.0077161361,0.6241717339,0.0516901277,0.2833348811,0.3863051236,0.1588709205,0.0476233736,-0.0055336566,0.2308053672,0.5782594681,-0.1743898541,0.3671568036,-0.5228919983,0.0775714293,0.2116926163,0.122955896,0.005603936,0.031119138,-0.1564725488,-0.4288707972,0.0594722442,-0.1472803056,0.042211622,0.2035235167,0.1826676577,-0.1190581098,0.2278323919,-0.2830649018,-0.0059437174,-0.1857161671,0.2656484246,0.1051391289,0.2024466693,0.0008289005,-0.2361797243,0.221500203,0.2695159912,0.2052172273,-0.1491564661,-0.1476213038,0.428235203,0.3534136117,0.1693412662,0.014657395,0.3167547286,0.0962429345,-0.3463172615,-0.1519589424,0.4838857353,-0.0357929021,-0.0805382878,0.0886524469,0.3073081374,0.1158766598,0.0653569996,0.091283612,0.2688135207,0.0663800836,-0.0455183536,-0.1914458573,-0.2590549886,-0.1344322264,-0.2435743809,-0.0288173221,-0.2215945274,0.0934262052,0.0021051702,0.269428581,0.1448980123,0.1267837137,0.0566761009,-0.177900672,-0.0677237585,-0.2078700215,0.3414859176,0.2781054378,0.1037743017,0.1054759771,0.02401153,-0.0426237918,-0.0134245958,0.1805579215,0.1803596765,-0.3909809887,0.2518912256,0.1336865127,0.1309971809,-0.2239989787,0.0206262972,0.0429931916,0.2813043892,-0.0507084541,0.1824110001,-0.2047246248,0.299683094,-0.0462534949,-0.1584311575,-0.0436794795,-0.2114619762,0.009243397,-0.0766247064,0.1495618373,0.4088260233,-0.0732618421,0.3769261837,0.0553985387,-0.0642862841,0.1847970337,-0.067712456,0.0117821423,0.026519049,-0.0211126916,-0.2400423437,0.2869223654,-0.1938946247,-0.098600179,-0.3416166902,-0.4383988976,0.1606461853,0.0614481717,0.4100453854,0.2410633117,-0.2603682876,-0.0223946311,0.0411864892,0.1884096414,-0.2412289381,-0.0302342027,0.3084082603,-0.1350417882,-0.1841018647,-0.0438066795,-0.454743892,0.0212295149,-0.3150006831,-0.032749787,0.2424473166,-0.0739828199,0.0971793681,0.190607518,0.1345986426,0.1341573,0.0485786386,-0.2081841081,-0.1073848754,0.4198906422,-0.1874568015,-0.414606601,0.1693682373,-0.0941706821,0.3023670614,0.1207817942,-0.2617169619,-0.3486177921,0.0223566964,0.0841832682,-0.1161616594,0.0005431383,0.0513884872,0.0615979768,-0.1983550191,-0.220217824,-0.1936366409,-0.0119821178,-0.121142149,0.1930122972,-0.1696726233,-0.1834976971,0.5552967191,0.8563187718,0.3107652962,-0.2941723168,-0.0198015105,0.0878220648,0.3871375024,0.2270503342,-0.3990766704,0.1633845121,-0.4135036469,0.1365126967,0.3416533172,-0.00661373,-0.0442048684,-0.0444787517,0.0516215861,0.0912019759,-0.1697894633,-0.264424175,-0.0283464883,0.2278472334,0.0533501804,0.2036978304,-0.2415156662,-0.0809582025,0.3237988353,0.3613370359,-0.3245861828,0.119431898,-0.3851375282,0.2987192273,-0.272015065,0.2552255988,-0.3151361644,0.3091472089,0.1574492902,0.0931369737,-0.176647827,-0.2636002302,0.2334956676,-0.0927461162,0.0469977707,0.1676510274,0.0768601075,-0.1789132059,0.0294077732,-0.2039265186,0.1848763973,0.3138547242,0.3291064203,-0.1619696766,0.0284540858,0.0972550213,0.2815677822,-0.2345641553,0.1191424727,-0.4278559983,-0.0747590065,-0.0130678462,0.0715851858,-0.2045690715,-0.0227995533,-0.0277721975,0.1181158498,-0.5347546935,-0.2161837667,0.3146809638,-0.1297747195,0.1752554923,-0.2964365184,0.1668875962,0.0545072481,0.0477736369,0.5340453982,0.2614954114,-0.0147959245,-0.4740161002,-0.0751681626,0.0103574218,0.2893353403,0.1811861992,-0.3556497395,0.1588558257,-0.0256614983,0.1911879033,-0.3821746111,0.3040134013,0.2679178417,0.125944227,-0.0786570534,-0.2530210614,0.0703307614,-0.0784636363,-0.1001707464,0.1268384606,0.1722682714,-0.0530270264,0.3745054901,0.1805081815,0.9338230491,0.3048354685,-0.117640309,0.311278224,-0.0014501766,0.149389416,-0.3385812938,0.168480739,-0.1839235574,-0.3425354958,-0.0159010626,-0.0749692097,0.212080121,-0.1406523734,-0.4320197999,0.1974206865,-0.0337059461,0.146115005,0.2102958858,0.1458112299,-0.019056147,-0.0646474585,0.2706273198,0.2279320061,-0.0937656984,-0.0416848511,-0.1503835768,0.1288317442,0.1888365,-0.0489152372,-0.328735888,0.033273641,-0.2181822956,0.1421864033,0.1775968373,-0.0285909064,0.080340147,0.2912404537,0.6542518735,0.2708209455,-0.2000180036,0.1803056449,-0.1543882191,-0.0423724726,0.0070915837,-0.0903487578,0.4447782636,-0.0496390611,-0.1026115119,0.0386732146,0.1057105884,0.0518939979,-0.0228476766,-0.0338555127,0.2343965322,-0.212132737,0.1545855552,-0.0428687669,-0.0333765224,-0.4309327304,0.1695499867,0.017225787,-0.2008114159,0.3604368567,-0.0283382498,-0.2871698439,-0.0228945874,0.5290211439,0.0331323892,-0.3315867186,0.3670952618,-0.3462993205,-0.1792690158,-0.3073848188,-0.1229000166,-0.1924333721,-0.2215667963,0.1932686716,-0.3249643743,0.1601902395,0.2408547848,0.3278456628,0.3561698496,-0.2040894479,-0.139381513,-0.2308539748,-0.2793035507,0.0768874064,0.0703208968,-0.0398708247,-0.3134394586,0.0283526815,-0.0936166719,0.3216765523,-0.3566919565,0.294423759,-0.2056770921,0.005675849,-0.0150221651,-0.1854879409,0.0732490942,-0.1046479791,0.0243967641,-0.040467903,-0.4105326533,-0.1462352425,-0.3995324969,0.1405790001,0.1378885359,0.1581653953,0.0128124719,0.1462972611,-0.1491273642,-0.1367696077,0.1455403566,0.4514707029,-0.1399774551,0.2465093434,-0.0463888384,0.3047938049,0.1006535217,-0.1171557531,0.3308507502,0.2518533468,0.1103929281,0.2383760959,-0.1055430248,-0.0354177579,-0.1213158444,0.2168011665,0.2368120253,-0.0050427099,0.3941176534,-0.4078185558,-0.016748013,-0.1176989749,0.2764633298,0.2374686301,-0.100190714,-0.1347438842,0.2228135914,0.2545295954,-0.3505980074,-0.1477607191,-0.1271717995,0.2385991514,0.1553158313,0.1561452597,0.4248201549,-0.3391849101,0.1903369129,0.0715982839,0.3079235554,-0.2116219252,0.5018622875,-0.2749926448,0.1414739788,0.0851254314,0.3391653895,0.1029282808,0.3265251219,0.3486456573,-0.0819007382,0.7880043387,0.2494936287,0.0495789908,-0.1053637937,-0.1126099601,0.2925438285,-0.0345586538,-0.0738914758,-0.0221150834,-0.0717446357,0.4817574024,-0.1989732087,-0.0967071652,-0.2307794243,-0.2808462083,-0.0241122134,-0.1283454001,-0.0772293806,-0.2852729559,-0.2119911015,-0.1345700622,0.2132540345,0.0115761813,0.082634218,-0.0819599256,-0.0166168176,-0.136575684,-0.0486811139,0.2407119572,0.125700444,-0.2841622233,-0.1357598752,0.0312858671,0.0344064757,-0.004452006,0.4487253428,0.4019379616,0.377294451,-0.1418691427,-0.0686010048,-0.0889090821,-0.0059410073,-0.0337230153,0.3227184713,-0.0299831238,-0.4964961708,0.2722794414,0.1768933088,-0.1794414073,-0.0078893304,0.0196505133,-0.2082688957,-0.4713931084,0.6122552156,0.1800872982,-0.0107263671,-0.1204446629,-0.0348970592,-0.4972096682,0.0367160589,0.4607993364,0.0038615712,0.1777039617,-0.1913175434,0.1193149611,0.1058959067,0.1353339404,0.2601879239,-0.1021371409,-0.1443094909,-0.0543944538,-0.552370131,0.3112111986,0.1401669383,0.1420584917,0.1100060642,0.2123397142,-0.1531261355,0.1321941018,0.0141488062,0.0232289974,-0.0383470841,-0.3030210435,-0.6422166228,0.1211930737,-0.2429501116,0.0441623442,-0.0508458093,-0.3419215977,-0.2108595818,-0.4180002213,0.0630456284,0.0682836995,-0.2858635783,0.2087311894,-0.0372839011,0.528123498,0.0377626717,0.1317009628,0.0672623292,-0.1974610537,-0.2057643533,-0.0905558392,-0.2562341988,0.1957242042,-0.1716297269,0.3562581241,-0.2618242502,0.0302983951,-0.2557788491,0.2457088977,0.3573638201,-0.4469646513,-0.3357787132,0.3720013499,-0.0847398937,0.1395203918,0.1167863086,0.39231354,-0.123745203,0.5088219643,-0.2650341094,-0.2870718539,0.5064149499,-0.3036544621,-0.3115775287,-0.3253057897,0.6123332381,0.2777469158,-0.136125952,-0.5220625997,-0.0217416491,0.3835979402,-0.0126153482,-0.2013018876,0.1189894006,-0.3484034836,0.120727323,-0.0871778503,0.3735612929,-0.0616131723,-0.1087294221,0.1898055077,-0.0780484676]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2736","title":"Add Microsoft Building Footprints dataset","comments":"Motivation: this can be a useful dataset for researchers working on climate change adaptation, urban studies, geography, etc. I'll see if I can figure out how to add it!","body":"## Adding a Dataset\r\n- **Name:** Microsoft Building Footprints\r\n- **Description:** With the goal to increase the coverage of building footprint data available as open data for OpenStreetMap and humanitarian efforts, we have released millions of building footprints as open data available to download free of charge.\r\n- **Paper:** *link to the dataset paper if available*\r\n- **Data:** https:\/\/www.microsoft.com\/en-us\/maps\/building-footprints\r\n- **Motivation:** this can be a useful dataset for researchers working on climate change adaptation, urban studies, geography, etc.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nReported by: @sashavor","comment_length":29,"text":"Add Microsoft Building Footprints dataset\n## Adding a Dataset\r\n- **Name:** Microsoft Building Footprints\r\n- **Description:** With the goal to increase the coverage of building footprint data available as open data for OpenStreetMap and humanitarian efforts, we have released millions of building footprints as open data available to download free of charge.\r\n- **Paper:** *link to the dataset paper if available*\r\n- **Data:** https:\/\/www.microsoft.com\/en-us\/maps\/building-footprints\r\n- **Motivation:** this can be a useful dataset for researchers working on climate change adaptation, urban studies, geography, etc.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nReported by: @sashavor\nMotivation: this can be a useful dataset for researchers working on climate change adaptation, urban studies, geography, etc. I'll see if I can figure out how to add it!","embeddings":[-0.6230252385,0.0684200227,-0.2241073251,-0.0011123391,-0.0653921142,-0.0625759736,0.1025177017,0.2215958387,0.1189827546,0.3236289024,0.2269601375,0.217902258,-0.2382512838,0.4009601474,0.0867363885,-0.0855379552,-0.0039106542,0.0257602166,-0.3132502437,-0.0593195222,-0.2509870827,-0.0065763914,0.0509937145,-0.1066526622,-0.1733564287,-0.2124451697,-0.2157186568,0.0495205931,-0.1937619001,-0.1133068278,-0.1484749913,0.0985472873,-0.0310337208,0.1762388051,-0.0000950589,0.0303419419,0.0873692408,-0.0173286181,-0.1495585293,0.0322935879,-0.1998689026,-0.2357267588,-0.1913833767,-0.0995227471,-0.2482879758,-0.30550313,0.1753943413,-0.1604298651,0.1601273865,0.1716109365,0.4217909276,0.0971787944,-0.0057386984,-0.0885416195,0.1435337365,0.3612351418,-0.1792609245,-0.1117324308,0.2214247882,0.056765236,-0.1190083548,0.3845185637,-0.1306173801,-0.0669638515,0.1537836045,-0.0036901901,0.4531482756,-0.0692371801,0.0502346717,0.2914082706,0.369674325,-0.0297207776,0.0431030579,-0.2419794798,-0.0569375604,-0.1633583754,0.0782722682,0.3049122095,0.0192148574,0.0670702979,-0.3540785015,-0.2523000538,-0.1428038329,0.2731838524,0.0099574272,0.1090829,-0.2636860907,-0.1367986649,-0.0572412089,-0.0115478802,-0.0016431562,-0.0419681743,-0.0403620191,-0.1071382686,0.158935219,-0.2846086919,0.2675792575,0.3740527034,0.2439793497,0.0790721178,0.217136234,-0.01563178,0.1460665911,0.0936232284,0.0302221999,-0.0052096932,-0.2541327178,-0.0691237301,0.0850651637,0.0107921567,-0.036097236,-0.0156645365,0.0807580575,-0.0613360144,-0.3894498348,0.0395667851,0.1997938752,-0.1449997425,-0.2350879014,-0.1658115387,-0.0954438522,0.0673334375,0.2127809972,0.1905244887,0.1502644271,-0.2285726964,-0.0628765076,0.1328581125,0.087707974,0.0078050257,-0.1772198081,-0.022458028,-0.1922894567,0.2782240808,0.1723856628,0.1065273508,0.3462498486,-0.2274960428,0.122201778,0.1039486155,0.2224634141,0.2691270709,0.3936697245,0.0028162545,-0.1251039803,-0.3085380495,-0.0050263433,0.0725640431,-0.3081750572,-0.1603627801,-0.1034094244,-0.1174105555,-0.6414904594,0.325807035,0.0167046953,-0.1306327581,0.345561415,0.4194295108,-0.0807239786,0.1015363485,0.1099148616,0.5468338132,-0.2805459797,-0.1869441122,0.0710544214,0.3174801469,0.2171293497,-0.1784773916,-0.0948300138,0.0587568693,-0.1399818212,-0.2157729864,-0.0245057698,0.3536578715,-0.214051947,-0.1769830585,0.0251720808,-0.0991853923,-0.1977200061,0.0249370094,-0.4823987782,-0.3185199797,0.1451163441,0.0448162667,0.2196041197,-0.0234549902,-0.1469695866,0.2988389432,-0.1710606962,0.0904276967,-0.2635376453,-0.3852531016,-0.074767828,0.2344444841,-0.0112594897,0.0674141943,0.2207692564,0.1787233651,0.0239800196,-0.1878061444,0.225879699,0.2710601985,0.1962817013,-0.068755284,-0.3588473201,-0.1836929917,-0.2154411227,-0.2696718276,-0.3207305074,-0.1458588094,0.0978271216,-0.1733013391,0.0119499546,-0.1436832547,-0.1488265246,-0.0343428142,0.4483795464,0.1232353747,0.1259393245,0.0121793225,0.1993558109,-0.1470786184,-0.0002672509,-0.0755678192,0.1166488379,0.2808676064,-0.1394074708,-0.0229548384,0.1239664704,0.3052729368,0.0543753803,-0.0755761638,0.2510988712,0.247122407,-0.4497462809,0.1395659,0.2692772746,0.6203433275,0.0687890872,-0.2716166675,0.2816954255,0.0852933526,-0.1750495434,0.0361091383,-0.3376236856,0.1835892648,0.2295089662,0.0209701434,-0.1751335412,0.3101272583,-0.0707540885,-0.2436029762,-0.1267685443,-0.3893863261,0.085204199,0.116021201,0.0240438227,-0.0940998048,0.0616934486,0.2042934746,0.1607656926,0.041861482,0.3469276428,0.0788425729,-0.3081942499,0.0382645391,0.1487524062,-0.0959710702,0.0808029398,0.3439022303,0.227925539,0.009722596,0.0924154297,-0.0774166137,0.1614506692,0.1702503711,-0.0818793252,0.1255161911,0.0986687839,0.0706480071,-0.6637936234,0.0366603695,0.0985401422,-0.0511817485,0.0899757817,-0.1714110225,-0.0347530991,-0.4059104919,-0.0967215896,-0.0804718807,-0.0178093724,-0.2373978645,0.2716466486,0.021604253,-0.2388008833,0.2244889885,0.1554945856,0.3148770034,-0.1643851101,0.0231538918,-0.0639288798,-0.3015567064,0.0390949585,0.3722415566,0.1592794657,0.0796657652,0.5405745506,-0.2245623916,0.5198444128,-0.5102851391,-0.6112650633,0.0957111418,0.1514920592,-0.2901204526,0.2286679149,0.1449795961,0.1110247672,-0.193341732,0.0316030979,-0.0301307458,-0.0487857945,-0.2493214011,-0.2277149707,-0.1805243194,-0.2456382513,0.1113199666,-0.0120784584,-0.5066966414,0.2648170292,0.2403164953,0.0354922302,0.1074682325,0.1158182621,0.2038116604,-0.6048650742,0.2234700918,-0.1103955284,-0.3730153143,0.0482158922,-0.3968025446,-0.0938651934,-0.150873974,-0.1597165614,-0.0215467401,-0.064376466,-0.3661301136,-0.2579316795,-0.1849817038,0.2595286071,0.3191812336,0.0671480596,0.4511183798,0.3124196231,-0.2581288218,-0.0313524641,-0.0152908619,-0.0152023891,0.3127680123,0.3218161762,-0.110319756,-0.1075894684,-0.0312795974,0.2447682768,0.27085343,-0.2615627646,0.3958711624,-0.0930594057,0.5226427317,-0.0089004543,-0.1838506162,0.3940806687,0.0300675631,0.0997569114,0.4693876803,0.2889024019,-0.0597190075,-0.0623610504,0.0250974465,-0.326058805,-0.2032308578,-0.0565259159,0.0013795884,0.1906849295,-0.0734798536,-0.1845312119,-0.3187709451,-0.1213813424,0.4813637137,0.1417661756,0.1107131094,-0.1906781942,0.033068452,0.2859850824,-0.4778053761,0.1688167304,-0.0384078138,-0.6828498244,-0.5322356224,-0.0232678838,-0.1445318907,0.0450762585,0.18625848,-0.1528743058,-0.0885678232,-0.1438151449,0.1635587364,0.0843688473,0.1585358232,-0.0749703199,0.0854639933,0.2112661451,0.4735501111,-0.2738245428,0.0052979048,0.2657738626,0.2148851752,-0.0757661462,0.2871136367,0.1648852527,-0.3182855248,-0.2095270306,0.2221402526,-0.0155639192,0.1223189011,0.4529211819,-0.182093963,0.1176412031,0.0006023564,-0.071159102,-0.0766968429,0.330912292,0.2759955823,0.0682696775,0.140766561,0.3495258689,0.1246495321,0.4100805223,-0.3319398463,0.0867276043,0.1538970321,-0.3069647253,0.2242828459,0.2957182229,0.160738796,0.0075138849,-0.0488596372,0.073808074,-0.5192338824,0.1320587993,0.0601366051,0.2293597162,0.0801776648,0.1553262025,0.3070637286,0.0583862588,-0.2221197784,0.2683374286,-0.027530212,-0.2147562951,0.2057554275,0.2147283405,0.7943812609,-0.0279020965,0.058598768,0.3563005328,-0.1825183183,-0.1278189719,0.0548621714,-0.1430566311,0.0284267478,-0.2834557891,-0.0255118143,0.0109054763,-0.0159111395,0.2491575778,-0.3056859374,-0.1912502348,-0.0228167381,0.1667383015,-0.0822265223,0.1199410781,0.1780029535,-0.258058399,0.0419802703,0.3108378053,0.1387299448,0.268890053,-0.0310552921,-0.1904049367,0.3126851618,-0.1418596357,-0.265437454,-0.082591176,-0.0297539532,-0.2243076116,-0.1122974232,0.0149105927,-0.0759324357,0.0886700302,0.3624705672,-0.08031822,-0.3157357574,0.2020474374,0.1063031554,0.0051082387,-0.1376444548,0.0983411148,-0.1380514503,-0.1262558103,0.163084209,0.0074679549,-0.0462046191,-0.1249515936,-0.4165939987,-0.113306731,0.2288375348,-0.0053538489,0.5253333449,0.0532197356,-0.0698421001,-0.2707270682,0.2843101323,0.1614391208,0.1358452588,0.0491368547,0.283780545,-0.1220448539,-0.1672076881,0.1038564593,0.0259808023,0.1419001222,-0.1899153441,-0.0006591826,-0.213888526,-0.4174348712,0.0332389176,-0.0608518347,-0.0457501113,0.1162158027,0.1164599136,-0.2296576202,0.1569873542,0.3926699162,0.1478805691,-0.0292283967,-0.0863290876,-0.1877082288,-0.3688642979,0.2121437788,-0.1260418892,0.4057597518,0.0060098069,0.018742891,0.1353942007,0.0341270491,-0.5436127186,-0.0257353541,-0.2212553769,0.0951055065,0.1314076632,-0.2149346173,0.2341576368,0.0308509264,0.201445654,-0.2828033864,-0.2702720761,-0.3398702443,-0.1597587168,0.0696216375,-0.1033733934,0.1781076193,0.3261275589,0.0630131587,-0.1480593234,-0.1249708012,0.0129693048,-0.1350703686,0.0083862636,-0.1960857958,0.0350233689,-0.0264251046,-0.0611907616,-0.0173747484,-0.0235324167,0.0424221866,-0.1137643605,0.1766500026,-0.1386779994,-0.12076208,0.1893794537,0.2129257768,0.3505797684,-0.2509464622,0.2759134471,0.2776060104,-0.139607802,0.3650393188,0.0073039914,-0.404553324,-0.0795770735,0.1772398651,0.0323739871,0.4277229011,-0.408665359,0.0117176296,0.1282572597,-0.13191396,0.4178147912,0.1623308361,0.2632777393,-0.1911301166,0.086937584,-0.0108884778,-0.0888862908,0.0302265882,-0.0044235252,0.2714774609,-0.007616207,-0.1423957795,0.3776353002,0.1821473688,0.128664881,0.0805473477,0.0678061247,0.3705387115,-0.1380268782,0.0275955666,0.1527680308,0.0977484882,0.5251341462,0.0518852808,0.2139781117,0.0660587922,0.0854440033,0.208162114,-0.0804193094,-0.0528518595,-0.3378164768,0.4252868295,-0.0080623701,-0.1204302087,-0.1732714474,-0.128067717,-0.0152622089,-0.1200764179,0.2446441352,-0.2435300648,0.1986077875,-0.2377636284,0.2374536991,-0.0236413181,-0.047359135,0.0158296041,0.2397976369,-0.1980521083,0.2210666835,-0.2560222149,-0.1375742853,0.0605704114,0.4935972989,0.1389359385,-0.1153158396,-0.142337516,0.0269992501,-0.1033093557,-0.1255057156,-0.3297213614,0.0240418613,0.0214967262,-0.1325965226,0.4272669256,0.2687676847,-0.3066296279,-0.0320828706,-0.1506853402,-0.0244073123,0.2835618556,-0.2537994385,-0.0056195427,-0.0525398441,-0.2152953446,-0.0921746418,-0.2154505402,-0.2862652838,-0.0455491394,-0.2045318633,0.0657591969,-0.0375814401,0.1787768304,-0.0093694329,0.2979604304,0.2489988506,-0.0259637758,0.0779221654,-0.4312407672,-0.2481151074,0.2071688771,-0.2502091825,0.0365793109,0.050191097,-0.0540026277,0.2050937861,0.1456216574,-0.0717368573,0.0459604822,0.0319520496,-0.1186893135,-0.4050327837,0.0898733139,0.1561270207,0.1325041205,0.1925550699,-0.1578140706,0.0151295373,0.049785547,0.288323611,-0.3631313145,0.0406081192,-0.5141167045,-0.1600818187,0.1661547273,-0.2204020619,-0.0455409139,-0.3230528831,-0.3244899213,-0.0422530919,-0.0764980912,-0.092365399,-0.0277595613,0.2801039219,-0.0329690948,-0.118993476,0.0649994016,0.0712719783,0.1353490949,0.2421407551,0.1140562817,-0.0565737784,0.2857164443,-0.1097848937,0.1879209578,0.0829139724,0.0898013189,-0.0343560167,-0.209067598,0.1867342889,-0.2917206287,0.3974737227,-0.32621786,0.0062650763,0.0756154358,0.0850985125,-0.1385529339,-0.0676285103,-0.1220238283,0.0227510873,0.1825362742,-0.0107599571,0.081734553,0.1383603811,-0.1436312795,-0.1232248545,0.0074330783,0.2546879649,0.0474912189,-0.1564383656,-0.1412763894,-0.300254494]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2730","title":"Update CommonVoice with new release","comments":"cc @patrickvonplaten?","body":"## Adding a Dataset\r\n- **Name:** CommonVoice mid-2021 release\r\n- **Description:** more data in CommonVoice: Languages that have increased the most by percentage are Thai (almost 20x growth, from 12 hours to 250 hours), Luganda (almost 9x growth, from 8 to 80), Esperanto (7x growth, from 100 to 840), and Tamil (almost 8x, from 24 to 220).\r\n- **Paper:** https:\/\/discourse.mozilla.org\/t\/common-voice-2021-mid-year-dataset-release\/83812\r\n- **Data:** https:\/\/commonvoice.mozilla.org\/en\/datasets\r\n- **Motivation:** More data and more varied. I think we just need to add configs in the existing dataset script.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":2,"text":"Update CommonVoice with new release\n## Adding a Dataset\r\n- **Name:** CommonVoice mid-2021 release\r\n- **Description:** more data in CommonVoice: Languages that have increased the most by percentage are Thai (almost 20x growth, from 12 hours to 250 hours), Luganda (almost 9x growth, from 8 to 80), Esperanto (7x growth, from 100 to 840), and Tamil (almost 8x, from 24 to 220).\r\n- **Paper:** https:\/\/discourse.mozilla.org\/t\/common-voice-2021-mid-year-dataset-release\/83812\r\n- **Data:** https:\/\/commonvoice.mozilla.org\/en\/datasets\r\n- **Motivation:** More data and more varied. I think we just need to add configs in the existing dataset script.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\ncc @patrickvonplaten?","embeddings":[-0.3344433606,0.1367919743,-0.1677779853,-0.2020456642,0.0964258611,0.0142566301,0.0312099736,0.4432871044,-0.1097463742,0.0864119902,-0.2006262094,0.0304484237,-0.0000929219,0.4648815095,0.0485813245,-0.2004388571,0.0469747745,-0.0847056583,0.1775542647,-0.2414996326,-0.1956805587,0.159284234,0.1002675742,-0.0610139109,-0.1421611458,-0.0112974877,-0.1180986166,-0.0779026225,-0.1033025458,-0.4212245047,-0.0557945669,0.3998802602,0.1889361739,0.3795855045,-0.0001140729,-0.267732203,0.1291775852,-0.282938689,-0.3162460625,0.0809561089,-0.122398138,-0.2267675996,-0.3007515371,0.1698474437,-0.0335824676,-0.3076707423,-0.0783497766,-0.5394903421,0.1346291155,0.2862784863,0.1324923337,-0.2392212301,-0.0495986417,-0.0847590417,0.1743757278,0.1226275861,-0.2331611216,-0.0224590674,0.7200750709,-0.0471367985,-0.1283523887,0.1861952245,0.2053717077,-0.3689831793,-0.0206042659,0.1148260087,-0.180185765,-0.4507863224,0.2895863354,0.3878348768,0.6003500223,-0.2976606488,-0.332882762,-0.1784917414,-0.0349502824,-0.6233126521,-0.0444373116,-0.0668457747,0.087276794,0.1557532549,0.330816865,-0.1754645854,0.0489841737,0.376817435,0.2000891268,0.2939179242,0.0212639589,-0.0807376802,0.0178868342,-0.2513892055,-0.0957273766,0.0418823287,0.1345462352,0.2704121172,-0.0616524853,-0.2536797523,0.2781873643,-0.4691815674,0.5979049206,-0.1382641345,0.0221846476,0.1636906266,-0.7177422047,-0.0710466057,0.261369735,0.0010581794,0.2224807292,0.0323292911,0.2501889467,0.1384074986,0.5756569505,0.0459384136,0.0056037353,0.0219266713,0.0275007002,0.2551476359,0.2305422723,-0.2179806978,0.0129007827,0.1369679421,0.0933537409,-0.160830453,-0.1919325739,0.1917341948,0.2301498502,0.3446949422,-0.1619316489,0.3014247417,0.0845178515,-0.3756699562,0.0009843463,0.0238367226,-0.162093088,0.2005880326,0.174376443,-0.4192233384,0.0287586655,0.239895165,0.0376986191,-0.0941615999,-0.2154908627,0.1854188293,-0.1295573711,-0.2676977217,-0.1481769979,0.0763317049,-0.1651387513,0.3855668902,-0.4087771475,0.0600954965,0.2268037349,-0.2025876343,-0.3415330052,0.1475694031,0.2609882355,-0.2814308405,0.4178946614,0.4529577494,0.0924673751,-0.2947820723,0.0646596402,0.0851375684,-0.3823174536,0.0566599555,-0.027758643,0.4344403446,-0.4063884318,0.0129419519,-0.2395959198,0.0111388676,-0.1693868786,0.1435539722,0.0543482304,-0.0819815472,-0.1637581885,0.0219902936,0.3144164383,-0.350440681,-0.292234093,0.0804890245,-0.0908804908,-0.2232459188,0.3926413059,-0.0812949017,0.5469364524,0.0346584134,0.278368026,-0.0207253415,-0.3768345714,0.0768553242,-0.2014626712,-0.4410846233,-0.0562710464,0.0877974108,0.2366684228,-0.2129189372,0.0985380411,-0.1833524555,0.5046409369,-0.0508488156,0.3047119081,0.0537162386,0.4292595983,0.0739869773,0.122822158,0.0732175186,-0.3272367716,0.0192208532,0.1604508609,0.1891480088,0.306604594,-0.2408553958,-0.2738653421,-0.2878299057,-0.0695354789,-0.0827527344,0.0998057425,0.4362943769,0.184272781,-0.0548631325,-0.1592609137,0.0031293489,0.1907995194,-0.1224926636,0.045360256,0.1133449078,-0.1127025187,-0.0792486593,0.2075977772,0.1529584527,0.1828159541,-0.1304735988,-0.1940593868,0.1262128949,0.1396103948,0.5812410116,0.3228493035,0.3487066627,0.073867701,0.1276845038,0.5197438002,0.0671397448,-0.091869913,0.03536769,-0.0915923789,0.3888752162,0.4659537673,0.0187825132,-0.052888304,-0.0497652106,0.3332282603,0.074479118,-0.0592496358,-0.2115047127,-0.1237637997,-0.1668100804,0.2334382236,0.0160846673,-0.5571590662,0.3369103372,0.6944016814,-0.1498375833,0.1244730726,0.2302904427,-0.1014405563,-0.1164009646,0.0451240465,-0.0653105974,0.0819583982,0.2474473864,-0.0697759911,0.1164862886,0.288767159,-0.163051188,0.0887035504,0.1475792378,0.0126935039,0.4100949466,0.162103951,-0.0118783256,-0.5632206202,-0.2261361182,-0.09558025,-0.0067465701,0.1422468126,-0.2552599013,0.2077426165,0.1931276768,-0.26526317,0.0269205254,-0.1001066417,-0.2851304412,0.3025777936,0.123479329,-0.1200559884,0.4002256095,-0.211366117,0.1833987385,-0.1288378686,-0.0283519141,-0.2501919568,-0.1376788765,-0.1516355574,0.1108105704,0.2408805639,-0.2609279156,0.4083865583,-0.3563155532,-0.0359086767,-0.3484650552,-0.5124620199,0.1525961012,-0.1924049109,-0.3262976706,0.052773945,-0.3535558581,0.2443952858,0.1312736422,-0.0378578492,-0.2040374875,-0.2033956498,-0.282392174,-0.210174039,0.2242453694,0.1710597724,-0.7559931278,-0.1902291626,-0.5035305023,0.104781054,-0.2272010595,0.0806431696,0.0173982754,-0.1155173257,0.0003744769,-0.4605838954,0.3241290152,-0.3459461331,0.1970504522,0.1474667788,-0.1574266553,-0.2708445191,-0.0511584394,-0.214127183,0.1529091299,-0.0811918303,-0.1359440833,0.115237996,-0.1354601383,0.2170465887,-0.0957715362,0.0103503382,0.3839887679,0.2060741037,-0.2173097134,0.0678657144,-0.2216375172,-0.2839256227,0.2433210611,-0.0259434246,0.0332251787,0.4116035402,-0.071221076,0.812466383,0.0670570284,0.0506800637,0.2948823869,-0.0833784938,0.3705321848,-0.1688691229,-0.2501510382,0.1726172864,0.0763663575,0.1392796934,0.4543424249,-0.0127533358,-0.0318457112,0.0811907798,-0.2857671976,-0.202098906,-0.117915459,0.3965143859,-0.0427138917,0.5259312391,0.1671545208,0.0172552858,-0.2699294388,-0.0045405985,0.0861032084,-0.0197843648,-0.0457516536,-0.1030592173,0.0919369236,0.3861805499,-0.481895715,0.3823921084,0.3076646328,0.1991609335,-0.3786824048,0.0638714209,-0.0802323818,0.1490297168,0.1486767232,-0.1346458644,-0.2995654047,0.0830498785,-0.1100992486,-0.0795523301,0.1137731597,-0.456258446,0.0029105854,0.251947552,0.1557487249,-0.1381316781,-0.1692132205,0.3171076775,0.371419847,-0.0721094757,0.2151260823,0.0709825382,-0.2110930681,-0.5792369843,0.2619172335,-0.041634731,0.0307109468,-0.1926037818,-0.0724933222,0.4968529642,-0.1916741431,0.0275201611,-0.0788383335,0.2381486297,0.3523738384,0.2117387205,-0.1222801432,-0.0091544725,0.5017309785,0.278023392,0.0340096243,-0.1309440434,0.1363482475,-0.4806323946,0.4230264425,0.4341315329,0.2003853768,-0.0370222367,0.0921027511,-0.1105132774,0.0726522356,-0.1926946044,0.4042842984,0.322286129,-0.4913266599,-0.3872855902,0.1299975067,0.0918696299,-0.2914735079,0.0185045656,-0.1125743762,-0.166188702,0.1027377695,-0.2348349392,0.9652076364,0.0743397102,0.1491340399,0.0651347265,-0.1421101987,-0.0370048881,-0.1882981509,-0.151205495,-0.173494041,-0.0238069892,0.01783856,0.150424704,-0.0282417219,-0.121127896,-0.3157284558,0.3495641351,-0.094627142,-0.1736471951,0.0984246507,0.5194648504,-0.4600735307,-0.3413543403,0.0830444768,0.1000958607,-0.0559398942,0.3566455841,0.128812179,-0.1611159444,-0.3174955547,-0.3083710074,-0.4609496295,-0.0037506199,-0.0735678673,-0.1076171771,-0.1092464104,-0.0291870758,0.310397923,-0.2055137157,0.4156777263,-0.0782406479,-0.0743852779,-0.0079337433,0.2793928981,-0.0703237355,0.1390240788,0.0150966654,0.1650548428,-0.2767977715,-0.5002015233,0.168060258,0.0522965156,0.1153576672,-0.5897407532,-0.1492176652,0.2922712266,-0.2540206015,-0.030844273,0.1800870448,-0.2753941119,-0.3876789808,0.1313883364,0.3086997271,-0.1784861982,0.009012213,0.2431482971,-0.0350060463,-0.2756429017,0.237971887,0.3773227036,0.0300548002,0.4704969823,0.2398351282,-0.0137141915,-0.3040874302,0.1382157803,0.3552376628,-0.3208103478,-0.232944563,0.0143037513,-0.2210540324,-0.1700844318,0.4825772047,-0.0368716046,0.2883133292,-0.0561532155,-0.1011507064,-0.2716298103,0.2263635397,0.083847858,0.2843986154,-0.0455170907,0.0157150496,0.2297540009,-0.0846171603,-0.3322087526,-0.0348845199,0.3108028769,0.0702059045,0.6701047421,0.0064835404,0.1426776052,-0.068656072,0.1116203964,-0.0265022833,-0.2357502878,-0.1894302666,-0.169453904,0.1417532414,0.0780743957,0.0820744038,-0.1615343243,0.0513450466,-0.0825088918,-0.070580706,0.0375220366,0.0751338825,0.1411623806,0.2728959024,-0.134295851,-0.1990285665,0.3955514133,-0.0169454236,-0.0018367422,0.1460982412,0.1718902439,0.4233850539,0.0492091328,-0.2094766349,0.0228030365,0.2065178454,0.2238121778,0.076720126,0.0271869954,0.0351200961,-0.0017183846,0.4610902369,0.2864513695,0.2768828273,-0.1181716397,-0.0541617237,0.0058434042,0.2224864215,-0.4577184021,0.0824443027,-0.013849264,-0.2631269693,0.2003635168,0.221513927,0.2845505476,0.0364739038,0.3793738186,-0.0119781466,-0.1015049815,-0.214253366,0.4663233161,0.2083031982,0.0392787084,0.231543228,0.263887912,-0.1337706447,-0.0174935516,0.2087703049,-0.4163952172,0.1392759085,-0.0135165295,0.0751485005,0.3334106207,-0.3391009867,0.4097795486,0.2848491371,-0.1656517237,0.0432805978,0.297216326,0.6120655537,0.1946202964,-0.1469570845,0.0407248735,0.1514866054,-0.0461337529,-0.0106268935,0.0522359088,-0.2172541767,-0.3835811019,-0.024474863,-0.1312777996,0.2107817978,0.3883506954,-0.0108098779,-0.2087029219,-0.3109368682,-0.0716815665,-0.0813184083,0.1359789222,-0.1558258533,0.2835872471,-0.1272656918,-0.259332478,0.0508952364,0.5944473147,0.1336429864,-0.0580289923,0.2644755244,0.2690600157,-0.0195668731,0.1083438993,-0.0111219147,-0.2034238875,0.1833127737,0.3238346577,-0.0131067345,0.1303437203,-0.0799688771,-0.1848501116,0.2316902429,-0.3086585701,-0.0963139087,0.287144959,0.1119072512,-0.0476732664,-0.4469649792,-0.0118696019,-0.4674573243,-0.2269487381,0.25201419,0.1364136785,0.1058524027,-0.3022642136,0.0566924624,0.085597083,-0.1439604908,0.243376255,-0.1076528132,-0.0040238299,-0.0531054735,-0.4185459018,0.1399836391,-0.0516982377,0.0212978907,-0.3105081618,-0.0732347667,0.4565008283,0.2859395146,0.6305113435,0.0251563247,-0.5504930019,-0.2937317491,-0.2035820186,0.0413416326,0.142634362,0.4413681626,0.1180464998,-0.1293048263,0.086598359,-0.2338100672,0.0486372896,-0.3601699471,-0.1475121379,-0.1825370789,-0.1540608853,0.3786067963,-0.0406318866,0.3071018159,-0.3495566249,-0.2846395969,-0.1639335454,-0.1889612079,-0.1816720665,0.3997537792,0.1387470365,0.1688544303,-0.3066754043,-0.5365449786,-0.190332219,0.3195205927,-0.1730363518,-0.1893758476,-0.0064439122,0.2709950209,0.1330003589,-0.2241257429,-0.0161836203,0.0851400867,-0.0203951374,-0.0661897659,-0.3780663908,-0.2393779457,0.0756154135,0.294570595,-0.2629236579,-0.3322130442,0.1792128682,0.3450538218,-0.0582245328,-0.094053477,0.3028229773,0.1311744452,0.2679727674,0.1364177614,-0.0159901232,-0.0095686056,-0.2763176262,-0.1753933579,0.3543057442,0.232315734,-0.2786409259,-0.0044126394,-0.2495384514]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2730","title":"Update CommonVoice with new release","comments":"Does anybody know if there is a bundled link, which would allow direct data download instead of manual? \r\nSomething similar to: `https:\/\/voice-prod-bundler-ee1969a6ce8178826482b88e843c335139bd3fb4.s3.amazonaws.com\/cv-corpus-6.1-2020-12-11\/ab.tar.gz` ? cc @patil-suraj \r\n","body":"## Adding a Dataset\r\n- **Name:** CommonVoice mid-2021 release\r\n- **Description:** more data in CommonVoice: Languages that have increased the most by percentage are Thai (almost 20x growth, from 12 hours to 250 hours), Luganda (almost 9x growth, from 8 to 80), Esperanto (7x growth, from 100 to 840), and Tamil (almost 8x, from 24 to 220).\r\n- **Paper:** https:\/\/discourse.mozilla.org\/t\/common-voice-2021-mid-year-dataset-release\/83812\r\n- **Data:** https:\/\/commonvoice.mozilla.org\/en\/datasets\r\n- **Motivation:** More data and more varied. I think we just need to add configs in the existing dataset script.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":25,"text":"Update CommonVoice with new release\n## Adding a Dataset\r\n- **Name:** CommonVoice mid-2021 release\r\n- **Description:** more data in CommonVoice: Languages that have increased the most by percentage are Thai (almost 20x growth, from 12 hours to 250 hours), Luganda (almost 9x growth, from 8 to 80), Esperanto (7x growth, from 100 to 840), and Tamil (almost 8x, from 24 to 220).\r\n- **Paper:** https:\/\/discourse.mozilla.org\/t\/common-voice-2021-mid-year-dataset-release\/83812\r\n- **Data:** https:\/\/commonvoice.mozilla.org\/en\/datasets\r\n- **Motivation:** More data and more varied. I think we just need to add configs in the existing dataset script.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nDoes anybody know if there is a bundled link, which would allow direct data download instead of manual? \r\nSomething similar to: `https:\/\/voice-prod-bundler-ee1969a6ce8178826482b88e843c335139bd3fb4.s3.amazonaws.com\/cv-corpus-6.1-2020-12-11\/ab.tar.gz` ? cc @patil-suraj \r\n","embeddings":[-0.3897120953,0.0689004436,-0.0436255857,-0.1407739818,-0.0075453636,0.2361044884,-0.0866278261,0.4099915326,-0.0332524851,0.0552654713,-0.2268214375,0.1267442852,-0.1374958158,0.3757566512,0.3153657019,-0.102551654,-0.0221795943,0.0116232457,0.3541306555,-0.2009497434,-0.1525128484,0.15630804,0.0348923542,-0.0708511695,-0.067562893,0.0845847875,-0.1082298458,0.0526001304,-0.1277033687,-0.3679584563,0.1854728758,0.3187081516,0.2517135143,0.2469440848,-0.0001240131,-0.4164184928,0.1370145082,-0.1946643889,-0.5836836696,-0.1107830554,-0.2506595254,0.0372974947,-0.1177459136,0.1902164668,0.1305427402,-0.0199203938,0.0473647751,-0.3805799186,0.0396899171,0.1962828189,0.0604044572,-0.1646220237,0.1000424176,-0.2372972071,0.0352670588,0.1634634286,-0.0934525654,0.3102356791,0.6635541916,0.2257751524,0.2486783713,0.0445476882,0.1510121226,-0.3072907627,-0.0004820801,0.1340319216,-0.2501448095,-0.4530813098,0.1524716914,0.5523585677,0.6914018989,-0.1817145646,-0.3851211667,-0.1164635122,0.1334374994,-0.5129933953,-0.1119492501,0.0647773072,-0.0460774861,0.1696078032,0.2209832817,-0.4564382136,-0.008320705,0.453556031,0.2018192112,0.18715702,0.0545919687,-0.1376045197,0.3618970513,-0.0942300037,-0.0713778585,0.0558883622,0.0256280061,0.434869498,0.0762254372,-0.3975177109,-0.0114526153,-0.5160592198,0.4938572943,-0.2566072047,0.0837338194,0.1518265754,-0.6276643276,-0.1240408123,0.3796098232,-0.260255456,-0.1462113559,0.0701046214,0.3797554076,0.1778353751,0.816459775,0.1496021003,-0.2681249678,0.066015549,-0.2071320713,0.1514522433,-0.0152980229,-0.3966103792,-0.1177079305,-0.0770759806,0.0033965383,0.0795907155,-0.3144282997,0.0444237068,0.0949636102,0.121604152,-0.0240253303,0.1572534144,0.1144951209,-0.2641264498,0.1021916121,0.1461638063,0.0444499068,0.1630934477,0.0774115473,-0.6380671263,0.0543287769,0.0733752102,0.3085193634,-0.1125856936,-0.3097593188,0.2996527851,-0.3038833141,-0.14780128,0.0384144932,0.1520600915,-0.2544485331,0.3241442442,-0.3895079494,0.12453641,-0.0625213608,-0.4871847928,-0.1174947172,0.053168539,0.1520292163,-0.3868251145,0.317874521,0.2663354576,-0.216191873,-0.3090520203,-0.0580613464,0.2228986472,-0.3956089616,0.029081177,0.0688039139,0.5167530179,-0.6113048792,-0.0970230773,-0.3922154605,0.1342055053,-0.2091183215,0.1252496839,-0.028499756,0.0580733046,-0.4365787208,0.0588036999,0.5816566944,-0.517485559,-0.2799951732,0.2649064064,-0.2213199884,-0.2088127732,0.1940362602,0.127349034,0.4670020342,-0.0127242673,0.1766999662,0.1752267182,-0.255661577,0.0175010562,-0.0913153142,-0.4835682213,-0.2657375336,0.16387932,0.1479585022,-0.1275347322,0.1275569052,-0.0209599696,0.4738978148,-0.0539265797,0.2950321734,0.2058830559,0.324365139,0.0320440307,-0.0217146464,0.104077816,-0.3587940633,0.0817736164,0.0326973535,0.1676013321,0.0703762844,-0.3706407249,-0.2648741603,-0.3648478389,-0.0455134474,-0.1106761172,-0.0740583986,0.2317182273,0.2323179394,-0.199136734,-0.2143363655,0.2044940144,0.2810825706,0.1102151275,0.0540390611,0.013871097,-0.0384281725,0.1122697815,0.2321036905,0.1626625806,0.2238751352,-0.2100117803,-0.107539326,0.2778187394,-0.0040991791,0.6636211872,0.2237515748,0.210928455,0.2058418989,0.0763214827,0.2137951404,0.2449189126,-0.1573190689,0.0034839413,-0.2436186671,0.2828460932,0.3007521629,0.0037648624,0.1546221972,0.230799377,0.1304882616,0.0746317357,0.0362140536,-0.1223573312,-0.1598810703,0.0585057773,0.3595196605,-0.1486743391,-0.5801500082,0.3750169873,0.4067924619,-0.1300884634,0.1598211974,0.2438954562,-0.3069536686,-0.1593468338,-0.0127577512,0.1984671205,0.1514736265,0.2299783975,0.0910491422,0.1847023964,0.2917860746,-0.0427496321,0.2333383262,-0.0779767632,-0.0976061672,0.423174262,-0.1869979948,0.1300887764,-0.3433693349,-0.1746968776,-0.0153600946,-0.0549228825,0.0351476632,-0.460986346,0.0200646538,0.1403170824,-0.4364106357,-0.040547926,-0.0631513,-0.3243238032,0.2186037749,0.3028701246,-0.0349642709,0.0587985627,-0.3607234359,0.2980546951,-0.3364134729,-0.4129265845,-0.0940513387,-0.0122429384,-0.248903513,0.0277524069,0.0449794382,-0.2859221399,0.2350471169,-0.2272629589,0.1947665662,-0.4353841245,-0.4220513403,0.1608829796,0.0167447273,-0.4277884662,0.1313138604,-0.2983762026,0.2825034261,0.0157571062,-0.3078204691,-0.1229287907,-0.1884716004,-0.4340620935,-0.137435779,0.2221311927,0.1926539242,-0.7328444123,-0.3953492641,-0.4587977529,0.1181819886,-0.2502526641,0.0705011189,0.0066022901,-0.0758031905,-0.1232876405,-0.2171438932,0.2414301634,-0.2790500224,0.2103625238,0.2944696844,-0.1946478039,-0.3588214517,0.0477154218,-0.2169554383,0.2603701055,0.0520536229,-0.3211286366,0.429033339,0.1537174284,0.231214121,-0.0417889804,-0.0362404361,0.5302618742,0.2607545555,-0.1030115783,0.0379973166,-0.0504230298,-0.2390133739,0.2745098472,0.2551525235,0.046681162,0.2509114444,0.1000865921,0.7929123044,0.0699423254,0.0917326212,0.0113678053,-0.1323479712,0.3699173331,0.0880693719,-0.16677095,-0.0713275969,-0.1005616486,0.1786989868,0.335919112,0.2163038999,-0.0623995773,-0.1944150478,-0.1485823542,-0.1240454093,-0.0508000031,0.0870498195,-0.1279826611,0.3972432911,0.2608755827,-0.0262057222,-0.1185222343,-0.004402441,0.2155898213,0.1549451351,0.0986184627,-0.0102555482,0.3189973831,0.4521585405,-0.346770376,0.1879079491,0.329207629,0.172817111,-0.3932293057,0.0149850631,-0.0070268582,0.1347649246,0.2318745106,-0.2194045037,-0.1079743505,-0.0550622493,-0.0080589904,-0.0717712715,0.1767874211,-0.2748183012,0.0279459227,0.4873165786,-0.1069932505,-0.3011751175,-0.0860540569,0.509532094,0.1847725958,-0.0767919496,0.1322140396,0.2080124319,-0.1358364522,-0.3548860252,0.056572143,-0.159571439,0.0538226962,-0.1640255451,-0.1018120125,0.4033177197,-0.036718037,0.0523289144,-0.3107072115,0.1715804487,0.3582842648,0.255602777,0.1865482032,0.0445892774,0.432236433,0.4311977029,0.0220359974,-0.2229163498,0.0372350924,-0.5398786068,0.391156882,0.4940482974,0.1206308156,-0.1265556961,-0.0328858197,-0.2956703305,0.1037103757,-0.0365500189,0.3544361591,0.2191474438,-0.5651629567,-0.43532148,0.1628646702,0.1665744483,-0.2549287379,0.257956177,-0.1141850427,-0.1582359225,-0.0315858088,-0.280239284,1.1013661623,0.2786725461,0.3372337818,-0.0093836533,-0.3418526053,0.1972944885,-0.2426375449,-0.0923856124,-0.1486104578,0.2155482471,-0.2085614055,0.0637803748,-0.2620514929,0.0314011127,-0.2880208194,0.4694892168,0.0665203854,-0.0216721706,0.2636146247,0.4605933428,-0.2247128636,-0.3435023725,0.2273357213,-0.0084929327,0.1254480034,0.6293458343,0.0409239754,-0.3916577697,-0.2945464253,-0.3658154011,-0.4083197415,0.2242784351,0.0314943902,-0.0602370985,-0.1336973459,-0.0498057008,0.3182790577,-0.1848175824,0.3094366193,-0.0069451174,-0.26862517,0.0355223082,0.1778978407,-0.12503618,0.310616225,0.0363713205,0.0862060189,-0.0954761133,-0.2484400421,0.1662700176,-0.0545728505,-0.2792828381,-0.1199136823,-0.1825055331,0.035206262,-0.2328575552,-0.2651265264,0.3359253109,-0.2981410027,-0.2906297147,0.0317697003,0.1675308943,-0.1748851091,-0.0792971328,0.3043962121,0.0575806312,-0.1693570018,0.6294350624,0.0792630017,-0.2609728873,0.4255687296,0.0130989095,-0.0996785611,-0.165710628,0.0735646039,0.2148167193,-0.5074220896,-0.0870023444,-0.2099504471,-0.0243628155,-0.2353974432,0.3598627746,0.0052564302,0.3169464171,-0.3262867033,-0.2382825166,-0.3997449279,0.137104854,0.0130982017,0.3953421116,-0.0526326932,0.0024275363,0.2174752802,-0.174948737,-0.1994874924,0.1233490035,0.3694260418,-0.127379179,0.6541806459,0.2961562574,0.2610889971,-0.0086470144,-0.0815133303,-0.1322223097,-0.2986318469,0.0087532774,-0.0757567361,0.1869587302,0.0391259864,-0.097215794,-0.0929520279,-0.0846462622,-0.2465285957,0.290964514,0.0212851949,0.0734152496,0.1227353588,0.031334091,-0.0714001656,-0.0764033571,0.0340361185,0.2860245705,0.0586288869,0.4684439898,0.1377413422,0.4391714036,-0.1340589225,-0.2675172687,0.1980863363,0.2103239447,0.3753091097,-0.2128186673,0.306858629,-0.0254090093,-0.0645927265,0.1863846034,0.2407657057,0.3072826862,-0.2076263726,0.0092561627,-0.0353048742,0.0691498294,-0.4054576159,-0.0847057402,-0.2982486784,-0.0039069708,0.0256102756,0.1183420792,0.3522495925,-0.0350687429,0.5823600292,-0.0325137191,0.0720368996,-0.0760496259,0.3950070441,0.4880948663,-0.0044317609,0.2583204508,0.0836246982,0.0684489086,-0.0589560419,-0.0442601703,-0.5513554811,-0.1905621737,0.1594310105,0.0079047466,0.4129476845,-0.3530224562,0.2965228558,0.4625869393,0.1082484871,-0.1131673828,0.3864281476,0.5600408316,0.1485456973,-0.0364757665,0.129424572,0.1728716046,0.1189906374,-0.1233244464,-0.0667541549,-0.2481359392,-0.5411581993,0.2119107991,0.0681435615,-0.126234144,0.1150538027,0.0577244423,0.1066495255,-0.2904767692,-0.2527491748,0.0812533647,-0.0183547642,-0.1742881387,0.0859901235,-0.1135554537,-0.1134597734,0.2942744493,0.5817174911,0.142103374,-0.1389169842,0.2913804948,0.3578938246,-0.0027548566,0.1183114499,0.1787410975,-0.1841954738,0.183446154,0.1833059639,0.1276133358,0.0095172506,0.0336618274,0.0910174772,0.1196576133,-0.2098664939,-0.0627054796,0.4658003151,0.2066520602,-0.0252022371,-0.3366096616,0.0385637321,-0.2646123767,-0.1521443129,0.457765162,-0.0210623052,0.0115582105,-0.1262844056,0.0023984772,0.2506904602,0.0299122892,0.3688190579,-0.0540842935,-0.0706751496,-0.1639261842,-0.3195563853,0.077921316,0.0733777583,-0.0143795358,-0.366042614,-0.0116921971,0.2468598783,0.2482654154,0.4047496021,0.1105215102,-0.276094079,-0.0818060338,0.1156419888,-0.1208294183,0.2238270491,0.475220263,-0.0172497258,-0.1476756781,0.0066567902,-0.1145586073,-0.1547754705,-0.1602435261,-0.1522589624,-0.2332673073,-0.2176443785,0.3434557915,-0.2784252167,0.3796603382,-0.289849788,-0.2737599015,-0.153491199,0.0320383012,-0.1974920481,0.2600558698,0.0700506121,-0.0445153899,-0.2142017633,-0.267490834,-0.2769858539,0.1053996533,-0.0668181553,-0.1746442616,-0.1279671192,0.0851644129,-0.1741345227,-0.0723670796,0.1871174574,0.2277427018,-0.2829019129,-0.1032679901,-0.4452147484,-0.2198555321,0.2667154372,0.1583933681,0.1945142299,-0.2978351414,0.2802817822,0.2864385545,-0.1136660352,-0.1449763477,0.2475972027,0.1035009399,0.1211746037,0.0664869845,0.0072770221,-0.1020929217,0.0491352603,-0.2674430609,0.5598204136,0.3930937648,-0.1803131104,-0.2255277932,-0.4063813984]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2730","title":"Update CommonVoice with new release","comments":"Also see: https:\/\/github.com\/common-voice\/common-voice-bundler\/issues\/15","body":"## Adding a Dataset\r\n- **Name:** CommonVoice mid-2021 release\r\n- **Description:** more data in CommonVoice: Languages that have increased the most by percentage are Thai (almost 20x growth, from 12 hours to 250 hours), Luganda (almost 9x growth, from 8 to 80), Esperanto (7x growth, from 100 to 840), and Tamil (almost 8x, from 24 to 220).\r\n- **Paper:** https:\/\/discourse.mozilla.org\/t\/common-voice-2021-mid-year-dataset-release\/83812\r\n- **Data:** https:\/\/commonvoice.mozilla.org\/en\/datasets\r\n- **Motivation:** More data and more varied. I think we just need to add configs in the existing dataset script.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":3,"text":"Update CommonVoice with new release\n## Adding a Dataset\r\n- **Name:** CommonVoice mid-2021 release\r\n- **Description:** more data in CommonVoice: Languages that have increased the most by percentage are Thai (almost 20x growth, from 12 hours to 250 hours), Luganda (almost 9x growth, from 8 to 80), Esperanto (7x growth, from 100 to 840), and Tamil (almost 8x, from 24 to 220).\r\n- **Paper:** https:\/\/discourse.mozilla.org\/t\/common-voice-2021-mid-year-dataset-release\/83812\r\n- **Data:** https:\/\/commonvoice.mozilla.org\/en\/datasets\r\n- **Motivation:** More data and more varied. I think we just need to add configs in the existing dataset script.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nAlso see: https:\/\/github.com\/common-voice\/common-voice-bundler\/issues\/15","embeddings":[-0.3287990093,0.0056431592,-0.1429676861,-0.136348173,0.0628770068,-0.0556516834,0.0224146005,0.4327798784,-0.1923968494,0.1837512702,-0.193862617,0.0194020867,-0.0095699569,0.2673483193,-0.1464892626,-0.1913931817,0.053161785,0.0027607854,0.1654707938,-0.2329694182,-0.0372840837,0.1840407252,0.057953991,0.0246979259,-0.2935558259,-0.0396808237,-0.0605300292,-0.089106001,-0.068525061,-0.3911475539,-0.0233007185,0.3705922663,0.1399416476,0.4233662486,-0.0001093865,-0.2921771109,0.1471693516,-0.2058456391,-0.3941989243,0.1961703449,-0.0812218934,-0.1016451716,-0.2652498186,0.2674520016,-0.0806330442,-0.124721922,-0.1127621457,-0.441115886,0.2183479369,0.27924335,0.1529974937,-0.1604474932,-0.0829586983,0.0198848341,0.0261242595,0.1122264788,-0.192963466,0.0595030859,0.6804491878,-0.0657623708,-0.2130746543,0.1971845925,0.3200361431,-0.3722434938,-0.1842878312,0.1534989923,-0.1679836065,-0.4465769231,0.256474942,0.4199110866,0.4171394706,-0.2316382825,-0.2451350987,-0.2401668876,-0.0442307293,-0.5980993509,0.0448887013,-0.131061241,0.0454460979,0.1546523124,0.3375899494,-0.1187522039,0.0397706293,0.3219485581,0.1400587112,0.2067350149,-0.0013422724,-0.1634146571,0.1121886,-0.2705281079,-0.202791512,0.2048959732,0.1622344106,0.2746995389,-0.1196292192,-0.233962208,0.1979301423,-0.5092322826,0.4340422153,-0.2187785357,0.1262159348,0.2461792976,-0.6279563308,0.0006841497,0.2528567016,0.0853270739,0.2757959068,0.0058346535,0.3527455032,0.134414658,0.5356166959,0.1203448623,-0.1088360026,0.042500671,-0.0164747052,0.2085831314,0.0779291466,-0.1660064459,-0.0047889967,0.1155418679,0.1316093802,-0.1156344414,-0.1696537733,0.1402976662,0.1531683654,0.4080667496,-0.1286909878,0.1629196256,0.0818507001,-0.331505686,-0.0047363155,-0.0114259701,-0.1262592375,0.1622821391,0.1918677539,-0.5183823705,0.0101346802,0.2009649277,0.2281073481,-0.1432353109,-0.214575693,0.2270039916,-0.2798383236,-0.2979322076,-0.2365250289,0.07733991,-0.1451345831,0.3019645512,-0.3803790212,-0.0956063792,0.1123392805,-0.1741402,-0.2607727349,0.1932162344,0.2993985713,-0.220146805,0.408352375,0.3808480203,0.1286824495,-0.2547202408,0.0745955631,0.011111862,-0.2792163491,0.141075775,-0.0286298022,0.3412098885,-0.2390422076,-0.1498233676,-0.1999104172,-0.0263885465,-0.206585601,0.2132576257,0.0475485176,-0.2493336648,-0.1844452471,0.133523941,0.4131658077,-0.4571498632,-0.2381730825,0.109739162,-0.1660911888,-0.2226671129,0.3503072262,-0.1829272211,0.5053630471,0.0122358156,0.3081395328,0.0919673741,-0.2650010586,0.129248172,-0.2493554354,-0.3696264327,-0.146650508,0.1400948465,0.1465711594,-0.2267912626,0.0486244783,-0.0898158103,0.5055962205,-0.0895162299,0.3797619641,0.0435519442,0.4031324089,-0.0687719956,0.1943478882,0.1243795902,-0.1789497584,0.0022625795,0.0113008115,0.2435719818,0.4047494829,-0.263845861,-0.204265371,-0.306307286,-0.086904265,-0.0110088913,0.1587200314,0.4082472026,0.1919414103,-0.0767723173,-0.1107162088,0.0703116283,0.1728557199,-0.1154650897,0.0578090698,0.0820868239,-0.0727133006,-0.0934978202,0.304913789,0.1380336732,0.1314236224,-0.0730197802,-0.230270043,0.1178233474,0.0340612642,0.6732473373,0.2067365348,0.2734983563,0.07013271,0.1343422085,0.4396686852,0.1539883167,-0.1066209748,0.0405678041,-0.0461065359,0.415803194,0.5396988988,0.0679855123,0.023162052,0.0004799633,0.3002759516,0.0791260004,-0.0789188072,-0.2238884717,-0.0869660899,-0.233853206,0.2313191891,0.1419044137,-0.5391505361,0.3654895425,0.7573713064,-0.1148790047,0.1608107686,0.3280438483,-0.1346707046,-0.1927901953,0.0709903762,-0.1781436652,-0.0301248338,0.181874305,0.0218397118,0.023781199,0.2655636668,-0.1575941741,0.0733333156,0.1193032861,0.055251874,0.4313370585,0.2193501443,-0.1149060577,-0.5601583719,-0.2026019692,-0.1322214901,-0.0244070571,0.1491654515,-0.2236520648,0.0853072032,0.2968565226,-0.2935118377,0.0374068283,-0.1470390707,-0.2467741817,0.3771474659,0.1563028693,-0.1143167615,0.3223186135,-0.0689278468,0.1772903949,-0.0921771675,-0.0243746322,-0.2483328581,-0.0697153881,-0.3159554899,0.1675098389,0.1274507344,-0.3111944199,0.2936020792,-0.2800382376,-0.0318052694,-0.1229453459,-0.4287279248,0.2436238527,-0.1755266786,-0.4185717404,-0.0657474101,-0.4382660687,0.1236356273,0.0648989901,-0.0881204084,-0.1515877247,-0.2005161494,-0.1713787019,-0.2187151164,0.3064626157,0.1296020001,-0.7362055182,-0.2766322196,-0.5589891076,0.1818588078,-0.2512442768,0.039935872,-0.0363409407,-0.123635456,0.0193104539,-0.4042692482,0.4168932736,-0.3513768613,0.2028088421,0.0610142536,-0.168288216,-0.2349072844,-0.1149634495,-0.238270998,0.1909424812,-0.0960951969,-0.1373543739,0.1248797253,-0.0709061325,0.2104311138,-0.0365030169,-0.1126780733,0.3324787021,0.2923232913,-0.2397103757,0.1639590263,-0.097194083,-0.2600665092,0.3527743518,-0.0174686536,-0.0166628603,0.3119572401,-0.0806291625,0.8350934982,0.1460615546,0.0674578175,0.2415838987,-0.0225404613,0.2422726601,-0.1472086012,-0.098453626,0.251768738,0.1373865008,0.1796797812,0.495999068,0.0646237582,-0.0960716382,0.096775949,-0.3139930964,-0.1701499522,-0.1477795094,0.3515248597,-0.0502320975,0.5294163227,0.0715489611,-0.0169599112,-0.1150829345,-0.0272200331,0.1273702979,-0.0302324537,-0.1083531156,-0.1339220554,0.1667931676,0.3548084199,-0.4157821536,0.2468549162,0.2864933014,0.3051741421,-0.4020054936,0.0072022839,-0.0896017328,0.0962998122,0.0240856763,-0.0342513956,-0.2945667207,0.0779169127,-0.0958583578,-0.0295087621,0.1288964301,-0.4385118484,-0.0851936191,0.2515802383,0.0144961271,-0.1470246315,-0.1868223548,0.2074781358,0.407224834,-0.1182112098,0.29727754,0.0548949614,-0.0827758908,-0.6146731973,0.2879380882,-0.0681238919,0.0267771333,-0.2711752057,-0.0888249129,0.5392577648,-0.2731227875,0.1221195832,-0.0107477019,0.2702599168,0.2124397904,0.2449471951,-0.0789385736,0.0047379723,0.3941143453,0.2100632787,0.1242603883,-0.1026957855,0.1740561873,-0.4033103287,0.3960134089,0.4050860107,0.1484668106,0.0831728131,0.0694687292,0.0201945268,0.0628895313,-0.0472131148,0.4113476872,0.3688848019,-0.4769422114,-0.2611274421,0.1361936331,0.1341066509,-0.3131593764,-0.0157942921,-0.2158885449,-0.1703072041,0.100384824,-0.2168703824,0.8932237625,0.19080697,0.1610849351,0.1036306322,-0.0545644239,-0.1326561868,-0.260589987,-0.1413445026,-0.1152132973,0.1293559819,0.0218563415,0.1917188019,-0.0612109452,-0.1193140075,-0.4724832475,0.3740748465,-0.1531548202,-0.1923925877,0.0807363391,0.4558431804,-0.5018402934,-0.4145794809,0.1120151356,0.116994828,-0.0113398703,0.3453491032,0.2392480522,-0.154799521,-0.310315758,-0.3098520637,-0.414847523,0.0612624176,0.0768342689,-0.1139582396,-0.0044787894,-0.0162992775,0.2987708449,-0.2438541651,0.3647650182,0.0415673554,-0.0659847036,0.0414274596,0.2374843806,-0.0266193412,0.0787119716,-0.021862831,0.1971433312,-0.3036467731,-0.4409402907,0.1991460323,0.0441924632,0.1682037115,-0.4956968427,-0.0839928016,0.2335773259,-0.1415428668,0.0871969387,0.1798583567,-0.4254273772,-0.3390347362,0.1779562533,0.3150655925,-0.204813987,-0.0269224308,0.3536845744,0.0494528897,-0.2552163303,0.2695758939,0.4373168051,-0.0931038111,0.4643194973,0.1929383129,-0.038350217,-0.3709061146,-0.0011142773,0.3234229386,-0.3110635579,-0.2168284506,0.0972159654,-0.3706495762,-0.1822660863,0.4841510952,-0.1027794033,0.3393759131,-0.1822178364,-0.0605544187,-0.4040304422,0.1095236838,0.0963064134,0.2334157526,-0.1214052439,-0.0650301427,0.1852952987,-0.0161694922,-0.3843142986,-0.0732570961,0.1566679925,0.0625991151,0.6012382507,-0.0641983449,0.196011126,0.0427998789,0.1609281152,0.1258278489,-0.1799771637,-0.2301522493,-0.1852059364,0.1160403043,0.088419646,0.0634182021,-0.0851716548,0.2325888425,-0.062360473,-0.0845829397,0.0560090542,-0.0386260897,0.1456935853,0.2077568024,-0.1905601174,-0.1926707327,0.5110533834,-0.0827158764,0.0080854557,0.1231429949,0.1880858988,0.395067662,0.0197966881,-0.2576962113,0.1387117952,0.2802461982,0.1808000505,-0.0453108884,0.0062609236,0.0241348241,-0.083990179,0.2567949295,0.274128288,0.2085115314,-0.062448658,-0.0794920251,0.0546278581,0.2632263303,-0.3512751758,-0.0089004552,-0.1561014056,-0.303018719,0.2286928445,0.1973361373,0.2861754,-0.055749286,0.3852461576,-0.0163100753,-0.1486070007,-0.1884242594,0.3659958243,0.1239397004,-0.0280966628,0.2040688097,0.2646134198,-0.1684965789,0.0656731874,0.0052803899,-0.5093812943,0.1318530738,-0.0262814518,0.1865128875,0.3468427658,-0.2553498149,0.3466738164,0.2517774999,-0.0721173584,0.0881051347,0.2651171386,0.7147738338,0.1331702024,-0.177725181,-0.0435140319,0.1901111901,-0.0330490991,-0.0657976419,0.0960084498,-0.1697413921,-0.4783178866,0.0797436684,-0.033891838,0.1399851739,0.2834600508,-0.0349504575,-0.1653536558,-0.2518930435,-0.1317812651,0.0090921745,0.1179957017,-0.0762398317,0.2979882658,-0.1639651656,-0.2183935195,0.0603000633,0.4191316366,0.1132679582,-0.1318572164,0.2308754027,0.3432540298,-0.0474777073,0.1194513813,-0.0038680846,-0.1258392334,0.2070593834,0.2787878215,-0.0435881279,0.1547983289,-0.1914012432,-0.2108052224,0.0758072212,-0.2722718418,-0.0455792136,0.2618606091,0.1050414369,-0.0254453477,-0.5083024502,-0.113676548,-0.5223756433,-0.2293798476,0.3155860901,0.1566532105,0.1086348891,-0.3688565493,0.0733937994,0.1056943983,-0.0807914659,0.1661311835,-0.0437401235,0.0862122029,-0.1031218171,-0.3366262913,0.1758632809,0.0127496878,0.0988838524,-0.3679398596,-0.164902553,0.433216393,0.2149033546,0.5095219016,0.0548311844,-0.3888895512,-0.402870208,-0.1787259579,0.0967973024,0.229982391,0.4075419009,0.1385770589,-0.2034629881,0.1544445306,-0.2500110567,0.0800548792,-0.3775940239,-0.1567913443,-0.2731373012,-0.2130005956,0.3291234374,0.027162889,0.1550955772,-0.3603173196,-0.2617180049,-0.1786272675,-0.2004200518,-0.1505511999,0.1152905077,0.1311967373,0.2428105474,-0.1450206935,-0.4639286697,-0.2497962117,0.193587929,-0.0641472042,-0.3198407292,0.0693906769,0.2402732968,0.0509391166,-0.2182654738,-0.0014241978,0.084653303,0.0598050617,-0.2051076442,-0.3530311286,-0.2645479143,0.1153759584,0.2619373202,-0.1637193412,-0.3841892779,0.2137874067,0.4715111852,-0.086365521,-0.0314058028,0.2566943467,0.0763224885,0.2563596964,0.0487948097,-0.0852594525,0.0293449275,-0.2377815992,-0.2034143955,0.3851200342,0.2725175321,-0.2417205274,0.0568421483,-0.2572908998]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2728","title":"Concurrent use of same dataset (already downloaded)","comments":"Launching simultaneous job relying on the same datasets try some writing issue. I guess it is unexpected since I only need to load some already downloaded file.","body":"## Describe the bug\r\nWhen launching several jobs at the same time loading the same dataset trigger some errors see (last comments).\r\n\r\n## Steps to reproduce the bug\r\nexport HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets\r\nfor MODEL in \"bert-base-uncased\" \"roberta-base\" \"distilbert-base-cased\"; do # \"bert-base-uncased\" \"bert-large-cased\" \"roberta-large\" \"albert-base-v1\" \"albert-large-v1\"; do\r\n for TASK_NAME in \"mrpc\" \"rte\" 'imdb' \"paws\" \"mnli\"; do\r\n export OUTPUT_DIR=${MODEL}_${TASK_NAME}\r\n sbatch --job-name=${OUTPUT_DIR} \\\r\n --gres=gpu:1 \\\r\n --no-requeue \\\r\n --cpus-per-task=10 \\\r\n --hint=nomultithread \\\r\n --time=1:00:00 \\\r\n --output=jobinfo\/${OUTPUT_DIR}_%j.out \\\r\n --error=jobinfo\/${OUTPUT_DIR}_%j.err \\\r\n --qos=qos_gpu-t4 \\\r\n --wrap=\"module purge; module load pytorch-gpu\/py3\/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=\/gpfswork\/rech\/toto\/transformers_models\/$MODEL\"\r\n\r\n done\r\ndone\r\n\r\n\r\n\r\n```python\r\n# Sample code to reproduce the bug\r\n dataset_train = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter)))\r\n\r\n dataset_val = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter, args.filter + 5000)))\r\n\r\n dataset_test = load_dataset('imdb', split='test', download_mode=\"reuse_cache_if_exists\")\r\n dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True)\r\n```\r\n\r\n## Expected results\r\nI believe I am doing something wrong with the objects. \r\n\r\n## Actual results\r\nTraceback (most recent call last):\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 983, in _prepare_split\r\n check_duplicates=True,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/gpfswork\/rech\/tts\/unm25jp\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets==1.8.0\r\n- Platform: linux (jeanzay)\r\n- Python version: pyarrow==2.0.0\r\n- PyArrow version: 3.7.8\r\n","comment_length":27,"text":"Concurrent use of same dataset (already downloaded)\n## Describe the bug\r\nWhen launching several jobs at the same time loading the same dataset trigger some errors see (last comments).\r\n\r\n## Steps to reproduce the bug\r\nexport HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets\r\nfor MODEL in \"bert-base-uncased\" \"roberta-base\" \"distilbert-base-cased\"; do # \"bert-base-uncased\" \"bert-large-cased\" \"roberta-large\" \"albert-base-v1\" \"albert-large-v1\"; do\r\n for TASK_NAME in \"mrpc\" \"rte\" 'imdb' \"paws\" \"mnli\"; do\r\n export OUTPUT_DIR=${MODEL}_${TASK_NAME}\r\n sbatch --job-name=${OUTPUT_DIR} \\\r\n --gres=gpu:1 \\\r\n --no-requeue \\\r\n --cpus-per-task=10 \\\r\n --hint=nomultithread \\\r\n --time=1:00:00 \\\r\n --output=jobinfo\/${OUTPUT_DIR}_%j.out \\\r\n --error=jobinfo\/${OUTPUT_DIR}_%j.err \\\r\n --qos=qos_gpu-t4 \\\r\n --wrap=\"module purge; module load pytorch-gpu\/py3\/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=\/gpfswork\/rech\/toto\/transformers_models\/$MODEL\"\r\n\r\n done\r\ndone\r\n\r\n\r\n\r\n```python\r\n# Sample code to reproduce the bug\r\n dataset_train = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter)))\r\n\r\n dataset_val = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter, args.filter + 5000)))\r\n\r\n dataset_test = load_dataset('imdb', split='test', download_mode=\"reuse_cache_if_exists\")\r\n dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True)\r\n```\r\n\r\n## Expected results\r\nI believe I am doing something wrong with the objects. \r\n\r\n## Actual results\r\nTraceback (most recent call last):\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 983, in _prepare_split\r\n check_duplicates=True,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/gpfswork\/rech\/tts\/unm25jp\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets==1.8.0\r\n- Platform: linux (jeanzay)\r\n- Python version: pyarrow==2.0.0\r\n- PyArrow version: 3.7.8\r\n\nLaunching simultaneous job relying on the same datasets try some writing issue. I guess it is unexpected since I only need to load some already downloaded file.","embeddings":[-0.5779120326,-0.0810000375,-0.0453643724,0.4501565397,0.2459626794,0.0139751369,0.5565331578,0.2723817527,0.1843708605,0.1907434464,-0.0311780032,0.1464349478,0.0212427545,0.1441881359,-0.0885511786,0.0235604327,0.1309166104,-0.2257413864,-0.3968794644,-0.0203631092,-0.2406820655,0.1446244419,0.0855503231,0.1932705194,-0.5113247037,-0.3433853984,-0.0541765466,0.0557928272,0.2590035498,-0.0671327934,0.1664633751,0.2859389484,-0.0116243837,0.8678808808,-0.0001090042,0.1391353309,0.1289675683,0.1109726653,-0.0634465516,-0.2272718996,0.0330815762,-0.107974872,0.1186949238,-0.313777715,0.1189511046,0.0584998429,-0.030055631,-0.2093538344,0.5382676721,0.0736098737,0.2384028584,0.3819690347,-0.3065969348,-0.2297079861,0.1842454374,0.0569794737,0.0860361606,0.0703067929,0.1006150246,-0.160428226,-0.0243540276,0.4298269153,-0.128874436,0.205970034,-0.0231342856,0.0869602188,0.1051392555,-0.5513936877,0.0767241269,0.0688892454,-0.0115327006,-0.3123033047,-0.3542192578,-0.3473508954,-0.0639272109,-0.2551651597,0.0928215683,0.345266372,-0.2033960521,-0.1923074573,-0.0920517072,0.1867493987,-0.0755044073,-0.1349960566,0.2415366471,0.330652684,0.0830517709,0.2387527227,0.0692323297,0.0927333981,0.3280492723,-0.1612951756,-0.1014361978,-0.0775532722,-0.6200729012,0.0457460396,-0.0160569604,-0.465223223,0.1843742728,-0.1858731359,-0.1885086894,0.0499862283,0.1717092693,0.1031510457,0.2989788651,0.0960013494,-0.3177445531,0.2705483735,0.3438619673,0.1331980228,-0.5029964447,0.0652742609,-0.0156139219,-0.2567523718,0.2176098078,0.1873345226,0.2414573133,-0.0829889849,-0.3572193086,-0.0523138382,-0.0511178933,-0.0558524579,-0.1014149934,0.4524842501,0.1777906567,0.1567742676,0.0913131386,0.1969440281,-0.1822924018,0.0624874271,-0.3012304008,-0.0290754717,-0.0146923279,0.2143333256,0.115738757,-0.0643138587,0.2248848528,0.337944448,0.0498892926,-0.4685381353,0.262468487,-0.3657260239,-0.2471660674,0.1961344182,0.1851789206,0.1372065544,-0.1051604077,0.3842988014,0.0134874908,0.2587327957,-0.3288044035,-0.1393946558,0.080968149,0.2727389634,0.0137452884,0.1152774841,-0.3554980457,0.11369133,0.31862849,-0.0416694805,0.1086383611,-0.3720645308,-0.3164130151,-0.1393729448,-0.0205443148,0.5279052258,0.0768126845,0.0268207006,0.1770918816,-0.1090108529,0.0915662721,0.3510977328,-0.3423242867,0.4966166019,-0.3493314087,0.1288179904,0.2440502197,-0.3719643652,-0.1516722888,0.3383845389,-0.1646658182,0.1252206415,0.2112877369,-0.0568555631,0.2122339308,-0.135921821,-0.0084830578,0.1491883695,-0.1079862937,0.4015436471,-0.1918611526,-0.2839153409,0.0489497297,0.1688934714,0.3223809302,-0.093997553,0.0269821547,-0.2874779701,0.3875817657,-0.3304336965,-0.073193185,0.2158527076,-0.0103004053,0.0795356706,0.0810648277,-0.1096467897,-0.5169209242,0.1978815943,-0.3624865711,0.0432497934,0.0342297591,-0.1991609484,0.1067375392,-0.041770719,-0.2488535792,-0.188830331,0.1706272364,0.23171404,-0.1502679437,-0.1181374565,-0.1473599523,0.3599050343,-0.1582805365,0.0234616175,-0.0023934881,0.1705364734,-0.1951894164,-0.2185277343,0.1162100509,-0.0603484437,0.4591343403,-0.1816609353,-0.1646382064,0.234481588,0.1979359686,0.3151443601,-0.060830038,0.0168911535,0.1874209791,0.155690819,-0.0017908275,0.1425172091,0.264064461,-0.1452902257,0.0974556208,0.3178672493,0.0033611003,0.1383701563,0.0590600222,-0.2233062983,0.193423599,-0.1233381256,0.1891794056,0.1224577054,0.2253106385,0.1255556643,0.0167566203,-0.0816582367,-0.5741452575,0.0836443007,0.3775579929,0.1669413447,0.0734937191,0.0733089,-0.0474899113,0.0227888245,-0.1682934314,0.299851954,0.5353221893,0.1349111348,0.0457241945,0.0882089511,-0.1451271623,-0.0892191976,0.0666768849,0.1546055675,0.1948834658,0.5548338294,0.1994925439,-0.0699805468,-0.2803615928,-0.1119266897,0.2671932578,0.0941753089,-0.3748239875,0.1850156933,0.1274146289,-0.1747792363,-0.4213629663,0.0054045664,-0.0335235819,-0.1470950544,-0.0398527905,0.4568405449,-0.1997360289,0.3671449721,-0.0299287681,0.1571482718,-0.1867981851,0.2361011505,-0.1471163779,-0.0711986795,-0.1141485125,-0.0113953063,0.2849563956,0.0669397041,0.3187500536,0.0719447806,-0.3169528842,-0.3599549234,-0.0033945311,0.0783701614,0.0017772753,0.4231902361,-0.0844049677,-0.1965941489,0.0808480829,-0.4469995499,0.2024542987,-0.3591762781,-0.0702030137,-0.1327766627,-0.2264376134,-0.0671751946,-0.1766142398,-0.3789281249,-0.2513652146,-0.2676903903,-0.0906180367,-0.1727518439,0.2116888314,0.0763652846,-0.1411370635,-0.0400552638,0.1537126452,-0.0046703178,-0.3079814911,-0.3230585456,0.069326885,-0.1893555969,-0.3206602931,-0.0872680172,-0.14780882,0.0682690814,0.226744622,-0.4916568995,-0.3557165861,-0.1071279123,0.0577950142,-0.1461234093,-0.0371046998,0.0506975763,0.0610218383,-0.1380059868,-0.1171887517,-0.1728388518,0.2016058713,-0.0312829688,0.0052557569,-0.165980652,0.2671523094,-0.1580260098,0.6628261209,0.2449266016,-0.0611025915,0.3297878206,-0.1419707984,0.0786561295,-0.0182802659,-0.4407642186,-0.0686313659,-0.4300889671,-0.2329535484,0.1423813105,-0.3390009105,-0.1655286103,-0.0883629695,0.0207866021,-0.3483386338,-0.3111569285,0.1051787734,-0.156896323,0.3320013583,0.0473193973,0.1608238518,0.0032935208,0.0731304511,0.2852947414,0.0372188538,0.2085292339,-0.18799676,-0.0320747495,0.0756293982,-0.1916020811,0.3207663596,0.0992457867,0.1881765425,-0.0998482257,0.211159125,0.0173973516,-0.1154941842,0.8625531793,-0.1944365799,0.3259494305,0.1943781078,-0.3524224162,-0.0848695338,-0.1915691793,-0.3549926281,-0.211286366,0.2583167553,0.6643010974,-0.3218054175,-0.1237827912,0.1433633864,-0.0208376907,-0.2824375927,-0.3151057065,-0.1605422795,-0.3001244068,-0.2805768847,0.1758775413,-0.0905000865,0.1202875748,-0.5292029977,-0.0838713869,-0.1814909875,0.2068257481,-0.0105721066,0.058843255,0.2804206014,-0.0604163967,0.4039429128,0.1262611002,-0.0301582571,0.27479738,0.446331203,0.0201088637,-0.2770265043,0.00195421,0.0463416502,0.0897724777,0.4556978643,0.0272614881,0.1352091879,0.067100428,0.2680851817,-0.0672242194,0.1780904233,0.3836325109,0.2200000435,-0.1989115477,-0.3477317393,0.0931109264,0.2356354594,-0.1319556832,0.2889435291,-0.5127384663,-0.2068072408,0.3026010692,-0.0046676905,0.6622441411,0.014262584,0.0911543891,0.3084231317,-0.30280599,0.2865456343,-0.2150480598,0.3221714497,-0.58214885,-0.2683508396,0.1158726588,-0.1270323396,-0.0752488151,0.0526764467,-0.1140703261,0.3120547235,-0.0008438026,0.0324366987,-0.247971788,0.2446663976,0.0065662968,-0.3198530078,0.0531232022,0.1187562793,-0.2992705405,0.2569264472,-0.1400966644,-0.1082457006,0.2545025647,-0.2358598262,-0.2938596904,0.2325545698,-0.3484944701,0.280626297,-0.1433753818,-0.1927305907,0.0463992357,-0.0866825432,0.1560348868,-0.228722617,0.0731344894,0.0957627669,0.0568758026,0.1381038725,-0.2318829894,-0.1537003666,0.4277762473,0.0211631749,-0.3620678484,-0.0820333734,-0.037336532,-0.0148330815,-0.1281271279,0.0481154695,-0.1951638013,-0.1983331442,-0.1131099835,0.2236150652,-0.0742343813,-0.3154717684,0.1518261433,-0.2507011592,-0.2737055421,0.3905716538,-0.1238662377,-0.3739267588,-0.0774779171,0.2204256654,0.1755113006,0.0789951086,0.4106520414,0.3367050588,-0.0640727431,-0.3133040071,-0.238447547,0.1598051339,-0.4164747894,0.1772857159,-0.1435779184,0.1515220255,-0.0273460206,0.3164765239,0.0630009547,0.1205942333,0.0206558891,-0.2940013111,-0.2153663486,-0.0177845843,-0.0637846068,0.2160811573,0.2012470365,0.2618279457,0.0282044765,0.3244502544,-0.3282782733,0.2335392833,-0.2057464421,0.1308392584,0.0513715334,-0.131349355,0.0827395394,0.2397816777,0.0975585505,0.2197265029,-0.1647183448,-0.2612218559,-0.2263840884,0.0987549499,-0.0208422579,0.0202253181,-0.1477030218,-0.1334641278,0.1280261278,-0.0183768459,0.4289841652,0.2284487486,-0.164397791,0.1061610952,0.1367555261,0.1648219824,-0.0818796307,0.1702706814,0.00027305,-0.0996422619,0.1888261735,0.182882145,-0.0026518188,0.0396514907,-0.2979305983,-0.1828772873,0.41345644,0.1560270488,0.0929486826,-0.3851540387,-0.2239552438,0.0314513445,0.4500091374,0.3133749664,-0.3114973307,-0.0574878529,-0.0370501876,0.2802079916,-0.1861449033,-0.0252606235,-0.0004438632,-0.1982536614,0.0170283131,0.3708457351,0.322214067,0.0491976589,-0.0257169344,0.1577520519,0.3708704114,-0.3861029744,0.2404301912,-0.0043916507,-0.1260692775,0.1273742318,0.4078056812,0.2264866531,0.0797451213,0.178853035,0.0911041051,0.1854630262,-0.0301674884,-0.1411308199,0.1116839126,-0.3086596727,-0.0381123275,0.2016683668,-0.2263548374,-0.1244863719,-0.2119970918,-0.0688010901,-0.187140882,-0.2287652493,-0.3681076467,0.2676638961,-0.3974691629,-0.0481058285,0.1096312627,-0.1442805678,0.1278003901,0.2010710239,0.1140565127,0.1347726285,0.4494643509,0.133113876,-0.2817566395,-0.1104476377,-0.3163566291,0.2716574669,0.1550415754,-0.3163852394,0.2832721472,0.3418689966,-0.0504980497,-0.2370741516,0.552675724,0.6522240639,0.3537182212,-0.145219028,0.0072202305,0.0328731798,0.1318574995,-0.0833300203,0.0924024582,0.1320800632,0.2738418281,0.3750738502,0.1627761424,-0.1621029526,0.1509012282,0.2157594413,0.1737514883,-0.2635540366,0.2917876244,-0.0475654304,-0.0709339604,-0.0687781051,0.0109695336,-0.3325090408,0.1167241484,0.4725090563,-0.0993127301,0.2256483436,0.0002578047,0.1222921833,0.0116997818,0.3831605613,0.1213125065,0.100687854,-0.6743170619,0.0617796257,-0.4572269917,0.2957102358,-0.1418879032,0.1963327527,-0.0430551097,0.1047224849,-0.2969474494,-0.0652521476,0.2955662906,-0.0407785922,0.0765402913,0.4635721445,-0.5969395638,0.0495046973,-0.3352076411,-0.1262418479,0.3455519378,-0.4687039554,0.2194332629,-0.1522485614,0.0916097686,0.1163392141,-0.0036385118,0.3093613982,0.1244528666,0.5834572911,0.1750371456,0.6061421037,-0.1260510236,0.1312722862,-0.1131600216,-0.4133682251,-0.3300965428,-0.006951638,0.1493000537,0.1066098064,-0.0756934956,-0.0927477553,-0.3385125399,-0.1523738354,0.0260592829,-0.1396056712,-0.0495962091,0.0400649644,0.0016709572,0.0484994873,0.1314331293,0.2882748544,0.2544392049,0.2327956855,-0.4348084629,-0.3158876002,0.5066917539,-0.2032601684,-0.5028808713,-0.3684080839,0.3902992606,-0.0260889083,0.0840148702,-0.625706315,0.1361856461,0.3196178377,-0.0780654475,-0.1940839589,0.1666283607,0.034007255,-0.1527427286,0.1009978428,-0.0439530723,0.0453941226,-0.3636941314,0.0277081151,-0.0606810227]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2728","title":"Concurrent use of same dataset (already downloaded)","comments":"If i have two jobs that use the same dataset. I got :\r\n\r\n\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 582, in download_and_prepare\r\n self._save_info()\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 690, in _save_info\r\n self.info.write_to_directory(self._cache_dir)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/info.py\", line 195, in write_to_directory\r\n with open(os.path.join(dataset_info_dir, config.LICENSE_FILENAME), \"wb\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/LICENSE'","body":"## Describe the bug\r\nWhen launching several jobs at the same time loading the same dataset trigger some errors see (last comments).\r\n\r\n## Steps to reproduce the bug\r\nexport HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets\r\nfor MODEL in \"bert-base-uncased\" \"roberta-base\" \"distilbert-base-cased\"; do # \"bert-base-uncased\" \"bert-large-cased\" \"roberta-large\" \"albert-base-v1\" \"albert-large-v1\"; do\r\n for TASK_NAME in \"mrpc\" \"rte\" 'imdb' \"paws\" \"mnli\"; do\r\n export OUTPUT_DIR=${MODEL}_${TASK_NAME}\r\n sbatch --job-name=${OUTPUT_DIR} \\\r\n --gres=gpu:1 \\\r\n --no-requeue \\\r\n --cpus-per-task=10 \\\r\n --hint=nomultithread \\\r\n --time=1:00:00 \\\r\n --output=jobinfo\/${OUTPUT_DIR}_%j.out \\\r\n --error=jobinfo\/${OUTPUT_DIR}_%j.err \\\r\n --qos=qos_gpu-t4 \\\r\n --wrap=\"module purge; module load pytorch-gpu\/py3\/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=\/gpfswork\/rech\/toto\/transformers_models\/$MODEL\"\r\n\r\n done\r\ndone\r\n\r\n\r\n\r\n```python\r\n# Sample code to reproduce the bug\r\n dataset_train = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter)))\r\n\r\n dataset_val = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter, args.filter + 5000)))\r\n\r\n dataset_test = load_dataset('imdb', split='test', download_mode=\"reuse_cache_if_exists\")\r\n dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True)\r\n```\r\n\r\n## Expected results\r\nI believe I am doing something wrong with the objects. \r\n\r\n## Actual results\r\nTraceback (most recent call last):\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 983, in _prepare_split\r\n check_duplicates=True,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/gpfswork\/rech\/tts\/unm25jp\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets==1.8.0\r\n- Platform: linux (jeanzay)\r\n- Python version: pyarrow==2.0.0\r\n- PyArrow version: 3.7.8\r\n","comment_length":78,"text":"Concurrent use of same dataset (already downloaded)\n## Describe the bug\r\nWhen launching several jobs at the same time loading the same dataset trigger some errors see (last comments).\r\n\r\n## Steps to reproduce the bug\r\nexport HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets\r\nfor MODEL in \"bert-base-uncased\" \"roberta-base\" \"distilbert-base-cased\"; do # \"bert-base-uncased\" \"bert-large-cased\" \"roberta-large\" \"albert-base-v1\" \"albert-large-v1\"; do\r\n for TASK_NAME in \"mrpc\" \"rte\" 'imdb' \"paws\" \"mnli\"; do\r\n export OUTPUT_DIR=${MODEL}_${TASK_NAME}\r\n sbatch --job-name=${OUTPUT_DIR} \\\r\n --gres=gpu:1 \\\r\n --no-requeue \\\r\n --cpus-per-task=10 \\\r\n --hint=nomultithread \\\r\n --time=1:00:00 \\\r\n --output=jobinfo\/${OUTPUT_DIR}_%j.out \\\r\n --error=jobinfo\/${OUTPUT_DIR}_%j.err \\\r\n --qos=qos_gpu-t4 \\\r\n --wrap=\"module purge; module load pytorch-gpu\/py3\/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=\/gpfswork\/rech\/toto\/transformers_models\/$MODEL\"\r\n\r\n done\r\ndone\r\n\r\n\r\n\r\n```python\r\n# Sample code to reproduce the bug\r\n dataset_train = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter)))\r\n\r\n dataset_val = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter, args.filter + 5000)))\r\n\r\n dataset_test = load_dataset('imdb', split='test', download_mode=\"reuse_cache_if_exists\")\r\n dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True)\r\n```\r\n\r\n## Expected results\r\nI believe I am doing something wrong with the objects. \r\n\r\n## Actual results\r\nTraceback (most recent call last):\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 983, in _prepare_split\r\n check_duplicates=True,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/gpfswork\/rech\/tts\/unm25jp\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets==1.8.0\r\n- Platform: linux (jeanzay)\r\n- Python version: pyarrow==2.0.0\r\n- PyArrow version: 3.7.8\r\n\nIf i have two jobs that use the same dataset. I got :\r\n\r\n\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 582, in download_and_prepare\r\n self._save_info()\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 690, in _save_info\r\n self.info.write_to_directory(self._cache_dir)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/info.py\", line 195, in write_to_directory\r\n with open(os.path.join(dataset_info_dir, config.LICENSE_FILENAME), \"wb\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/LICENSE'","embeddings":[-0.5779120326,-0.0810000375,-0.0453643724,0.4501565397,0.2459626794,0.0139751369,0.5565331578,0.2723817527,0.1843708605,0.1907434464,-0.0311780032,0.1464349478,0.0212427545,0.1441881359,-0.0885511786,0.0235604327,0.1309166104,-0.2257413864,-0.3968794644,-0.0203631092,-0.2406820655,0.1446244419,0.0855503231,0.1932705194,-0.5113247037,-0.3433853984,-0.0541765466,0.0557928272,0.2590035498,-0.0671327934,0.1664633751,0.2859389484,-0.0116243837,0.8678808808,-0.0001090042,0.1391353309,0.1289675683,0.1109726653,-0.0634465516,-0.2272718996,0.0330815762,-0.107974872,0.1186949238,-0.313777715,0.1189511046,0.0584998429,-0.030055631,-0.2093538344,0.5382676721,0.0736098737,0.2384028584,0.3819690347,-0.3065969348,-0.2297079861,0.1842454374,0.0569794737,0.0860361606,0.0703067929,0.1006150246,-0.160428226,-0.0243540276,0.4298269153,-0.128874436,0.205970034,-0.0231342856,0.0869602188,0.1051392555,-0.5513936877,0.0767241269,0.0688892454,-0.0115327006,-0.3123033047,-0.3542192578,-0.3473508954,-0.0639272109,-0.2551651597,0.0928215683,0.345266372,-0.2033960521,-0.1923074573,-0.0920517072,0.1867493987,-0.0755044073,-0.1349960566,0.2415366471,0.330652684,0.0830517709,0.2387527227,0.0692323297,0.0927333981,0.3280492723,-0.1612951756,-0.1014361978,-0.0775532722,-0.6200729012,0.0457460396,-0.0160569604,-0.465223223,0.1843742728,-0.1858731359,-0.1885086894,0.0499862283,0.1717092693,0.1031510457,0.2989788651,0.0960013494,-0.3177445531,0.2705483735,0.3438619673,0.1331980228,-0.5029964447,0.0652742609,-0.0156139219,-0.2567523718,0.2176098078,0.1873345226,0.2414573133,-0.0829889849,-0.3572193086,-0.0523138382,-0.0511178933,-0.0558524579,-0.1014149934,0.4524842501,0.1777906567,0.1567742676,0.0913131386,0.1969440281,-0.1822924018,0.0624874271,-0.3012304008,-0.0290754717,-0.0146923279,0.2143333256,0.115738757,-0.0643138587,0.2248848528,0.337944448,0.0498892926,-0.4685381353,0.262468487,-0.3657260239,-0.2471660674,0.1961344182,0.1851789206,0.1372065544,-0.1051604077,0.3842988014,0.0134874908,0.2587327957,-0.3288044035,-0.1393946558,0.080968149,0.2727389634,0.0137452884,0.1152774841,-0.3554980457,0.11369133,0.31862849,-0.0416694805,0.1086383611,-0.3720645308,-0.3164130151,-0.1393729448,-0.0205443148,0.5279052258,0.0768126845,0.0268207006,0.1770918816,-0.1090108529,0.0915662721,0.3510977328,-0.3423242867,0.4966166019,-0.3493314087,0.1288179904,0.2440502197,-0.3719643652,-0.1516722888,0.3383845389,-0.1646658182,0.1252206415,0.2112877369,-0.0568555631,0.2122339308,-0.135921821,-0.0084830578,0.1491883695,-0.1079862937,0.4015436471,-0.1918611526,-0.2839153409,0.0489497297,0.1688934714,0.3223809302,-0.093997553,0.0269821547,-0.2874779701,0.3875817657,-0.3304336965,-0.073193185,0.2158527076,-0.0103004053,0.0795356706,0.0810648277,-0.1096467897,-0.5169209242,0.1978815943,-0.3624865711,0.0432497934,0.0342297591,-0.1991609484,0.1067375392,-0.041770719,-0.2488535792,-0.188830331,0.1706272364,0.23171404,-0.1502679437,-0.1181374565,-0.1473599523,0.3599050343,-0.1582805365,0.0234616175,-0.0023934881,0.1705364734,-0.1951894164,-0.2185277343,0.1162100509,-0.0603484437,0.4591343403,-0.1816609353,-0.1646382064,0.234481588,0.1979359686,0.3151443601,-0.060830038,0.0168911535,0.1874209791,0.155690819,-0.0017908275,0.1425172091,0.264064461,-0.1452902257,0.0974556208,0.3178672493,0.0033611003,0.1383701563,0.0590600222,-0.2233062983,0.193423599,-0.1233381256,0.1891794056,0.1224577054,0.2253106385,0.1255556643,0.0167566203,-0.0816582367,-0.5741452575,0.0836443007,0.3775579929,0.1669413447,0.0734937191,0.0733089,-0.0474899113,0.0227888245,-0.1682934314,0.299851954,0.5353221893,0.1349111348,0.0457241945,0.0882089511,-0.1451271623,-0.0892191976,0.0666768849,0.1546055675,0.1948834658,0.5548338294,0.1994925439,-0.0699805468,-0.2803615928,-0.1119266897,0.2671932578,0.0941753089,-0.3748239875,0.1850156933,0.1274146289,-0.1747792363,-0.4213629663,0.0054045664,-0.0335235819,-0.1470950544,-0.0398527905,0.4568405449,-0.1997360289,0.3671449721,-0.0299287681,0.1571482718,-0.1867981851,0.2361011505,-0.1471163779,-0.0711986795,-0.1141485125,-0.0113953063,0.2849563956,0.0669397041,0.3187500536,0.0719447806,-0.3169528842,-0.3599549234,-0.0033945311,0.0783701614,0.0017772753,0.4231902361,-0.0844049677,-0.1965941489,0.0808480829,-0.4469995499,0.2024542987,-0.3591762781,-0.0702030137,-0.1327766627,-0.2264376134,-0.0671751946,-0.1766142398,-0.3789281249,-0.2513652146,-0.2676903903,-0.0906180367,-0.1727518439,0.2116888314,0.0763652846,-0.1411370635,-0.0400552638,0.1537126452,-0.0046703178,-0.3079814911,-0.3230585456,0.069326885,-0.1893555969,-0.3206602931,-0.0872680172,-0.14780882,0.0682690814,0.226744622,-0.4916568995,-0.3557165861,-0.1071279123,0.0577950142,-0.1461234093,-0.0371046998,0.0506975763,0.0610218383,-0.1380059868,-0.1171887517,-0.1728388518,0.2016058713,-0.0312829688,0.0052557569,-0.165980652,0.2671523094,-0.1580260098,0.6628261209,0.2449266016,-0.0611025915,0.3297878206,-0.1419707984,0.0786561295,-0.0182802659,-0.4407642186,-0.0686313659,-0.4300889671,-0.2329535484,0.1423813105,-0.3390009105,-0.1655286103,-0.0883629695,0.0207866021,-0.3483386338,-0.3111569285,0.1051787734,-0.156896323,0.3320013583,0.0473193973,0.1608238518,0.0032935208,0.0731304511,0.2852947414,0.0372188538,0.2085292339,-0.18799676,-0.0320747495,0.0756293982,-0.1916020811,0.3207663596,0.0992457867,0.1881765425,-0.0998482257,0.211159125,0.0173973516,-0.1154941842,0.8625531793,-0.1944365799,0.3259494305,0.1943781078,-0.3524224162,-0.0848695338,-0.1915691793,-0.3549926281,-0.211286366,0.2583167553,0.6643010974,-0.3218054175,-0.1237827912,0.1433633864,-0.0208376907,-0.2824375927,-0.3151057065,-0.1605422795,-0.3001244068,-0.2805768847,0.1758775413,-0.0905000865,0.1202875748,-0.5292029977,-0.0838713869,-0.1814909875,0.2068257481,-0.0105721066,0.058843255,0.2804206014,-0.0604163967,0.4039429128,0.1262611002,-0.0301582571,0.27479738,0.446331203,0.0201088637,-0.2770265043,0.00195421,0.0463416502,0.0897724777,0.4556978643,0.0272614881,0.1352091879,0.067100428,0.2680851817,-0.0672242194,0.1780904233,0.3836325109,0.2200000435,-0.1989115477,-0.3477317393,0.0931109264,0.2356354594,-0.1319556832,0.2889435291,-0.5127384663,-0.2068072408,0.3026010692,-0.0046676905,0.6622441411,0.014262584,0.0911543891,0.3084231317,-0.30280599,0.2865456343,-0.2150480598,0.3221714497,-0.58214885,-0.2683508396,0.1158726588,-0.1270323396,-0.0752488151,0.0526764467,-0.1140703261,0.3120547235,-0.0008438026,0.0324366987,-0.247971788,0.2446663976,0.0065662968,-0.3198530078,0.0531232022,0.1187562793,-0.2992705405,0.2569264472,-0.1400966644,-0.1082457006,0.2545025647,-0.2358598262,-0.2938596904,0.2325545698,-0.3484944701,0.280626297,-0.1433753818,-0.1927305907,0.0463992357,-0.0866825432,0.1560348868,-0.228722617,0.0731344894,0.0957627669,0.0568758026,0.1381038725,-0.2318829894,-0.1537003666,0.4277762473,0.0211631749,-0.3620678484,-0.0820333734,-0.037336532,-0.0148330815,-0.1281271279,0.0481154695,-0.1951638013,-0.1983331442,-0.1131099835,0.2236150652,-0.0742343813,-0.3154717684,0.1518261433,-0.2507011592,-0.2737055421,0.3905716538,-0.1238662377,-0.3739267588,-0.0774779171,0.2204256654,0.1755113006,0.0789951086,0.4106520414,0.3367050588,-0.0640727431,-0.3133040071,-0.238447547,0.1598051339,-0.4164747894,0.1772857159,-0.1435779184,0.1515220255,-0.0273460206,0.3164765239,0.0630009547,0.1205942333,0.0206558891,-0.2940013111,-0.2153663486,-0.0177845843,-0.0637846068,0.2160811573,0.2012470365,0.2618279457,0.0282044765,0.3244502544,-0.3282782733,0.2335392833,-0.2057464421,0.1308392584,0.0513715334,-0.131349355,0.0827395394,0.2397816777,0.0975585505,0.2197265029,-0.1647183448,-0.2612218559,-0.2263840884,0.0987549499,-0.0208422579,0.0202253181,-0.1477030218,-0.1334641278,0.1280261278,-0.0183768459,0.4289841652,0.2284487486,-0.164397791,0.1061610952,0.1367555261,0.1648219824,-0.0818796307,0.1702706814,0.00027305,-0.0996422619,0.1888261735,0.182882145,-0.0026518188,0.0396514907,-0.2979305983,-0.1828772873,0.41345644,0.1560270488,0.0929486826,-0.3851540387,-0.2239552438,0.0314513445,0.4500091374,0.3133749664,-0.3114973307,-0.0574878529,-0.0370501876,0.2802079916,-0.1861449033,-0.0252606235,-0.0004438632,-0.1982536614,0.0170283131,0.3708457351,0.322214067,0.0491976589,-0.0257169344,0.1577520519,0.3708704114,-0.3861029744,0.2404301912,-0.0043916507,-0.1260692775,0.1273742318,0.4078056812,0.2264866531,0.0797451213,0.178853035,0.0911041051,0.1854630262,-0.0301674884,-0.1411308199,0.1116839126,-0.3086596727,-0.0381123275,0.2016683668,-0.2263548374,-0.1244863719,-0.2119970918,-0.0688010901,-0.187140882,-0.2287652493,-0.3681076467,0.2676638961,-0.3974691629,-0.0481058285,0.1096312627,-0.1442805678,0.1278003901,0.2010710239,0.1140565127,0.1347726285,0.4494643509,0.133113876,-0.2817566395,-0.1104476377,-0.3163566291,0.2716574669,0.1550415754,-0.3163852394,0.2832721472,0.3418689966,-0.0504980497,-0.2370741516,0.552675724,0.6522240639,0.3537182212,-0.145219028,0.0072202305,0.0328731798,0.1318574995,-0.0833300203,0.0924024582,0.1320800632,0.2738418281,0.3750738502,0.1627761424,-0.1621029526,0.1509012282,0.2157594413,0.1737514883,-0.2635540366,0.2917876244,-0.0475654304,-0.0709339604,-0.0687781051,0.0109695336,-0.3325090408,0.1167241484,0.4725090563,-0.0993127301,0.2256483436,0.0002578047,0.1222921833,0.0116997818,0.3831605613,0.1213125065,0.100687854,-0.6743170619,0.0617796257,-0.4572269917,0.2957102358,-0.1418879032,0.1963327527,-0.0430551097,0.1047224849,-0.2969474494,-0.0652521476,0.2955662906,-0.0407785922,0.0765402913,0.4635721445,-0.5969395638,0.0495046973,-0.3352076411,-0.1262418479,0.3455519378,-0.4687039554,0.2194332629,-0.1522485614,0.0916097686,0.1163392141,-0.0036385118,0.3093613982,0.1244528666,0.5834572911,0.1750371456,0.6061421037,-0.1260510236,0.1312722862,-0.1131600216,-0.4133682251,-0.3300965428,-0.006951638,0.1493000537,0.1066098064,-0.0756934956,-0.0927477553,-0.3385125399,-0.1523738354,0.0260592829,-0.1396056712,-0.0495962091,0.0400649644,0.0016709572,0.0484994873,0.1314331293,0.2882748544,0.2544392049,0.2327956855,-0.4348084629,-0.3158876002,0.5066917539,-0.2032601684,-0.5028808713,-0.3684080839,0.3902992606,-0.0260889083,0.0840148702,-0.625706315,0.1361856461,0.3196178377,-0.0780654475,-0.1940839589,0.1666283607,0.034007255,-0.1527427286,0.1009978428,-0.0439530723,0.0453941226,-0.3636941314,0.0277081151,-0.0606810227]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2728","title":"Concurrent use of same dataset (already downloaded)","comments":"You can probably have a solution much faster than me (first time I use the library). But I suspect some write function are used when loading the dataset from cache.","body":"## Describe the bug\r\nWhen launching several jobs at the same time loading the same dataset trigger some errors see (last comments).\r\n\r\n## Steps to reproduce the bug\r\nexport HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets\r\nfor MODEL in \"bert-base-uncased\" \"roberta-base\" \"distilbert-base-cased\"; do # \"bert-base-uncased\" \"bert-large-cased\" \"roberta-large\" \"albert-base-v1\" \"albert-large-v1\"; do\r\n for TASK_NAME in \"mrpc\" \"rte\" 'imdb' \"paws\" \"mnli\"; do\r\n export OUTPUT_DIR=${MODEL}_${TASK_NAME}\r\n sbatch --job-name=${OUTPUT_DIR} \\\r\n --gres=gpu:1 \\\r\n --no-requeue \\\r\n --cpus-per-task=10 \\\r\n --hint=nomultithread \\\r\n --time=1:00:00 \\\r\n --output=jobinfo\/${OUTPUT_DIR}_%j.out \\\r\n --error=jobinfo\/${OUTPUT_DIR}_%j.err \\\r\n --qos=qos_gpu-t4 \\\r\n --wrap=\"module purge; module load pytorch-gpu\/py3\/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=\/gpfswork\/rech\/toto\/transformers_models\/$MODEL\"\r\n\r\n done\r\ndone\r\n\r\n\r\n\r\n```python\r\n# Sample code to reproduce the bug\r\n dataset_train = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter)))\r\n\r\n dataset_val = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter, args.filter + 5000)))\r\n\r\n dataset_test = load_dataset('imdb', split='test', download_mode=\"reuse_cache_if_exists\")\r\n dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True)\r\n```\r\n\r\n## Expected results\r\nI believe I am doing something wrong with the objects. \r\n\r\n## Actual results\r\nTraceback (most recent call last):\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 983, in _prepare_split\r\n check_duplicates=True,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/gpfswork\/rech\/tts\/unm25jp\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets==1.8.0\r\n- Platform: linux (jeanzay)\r\n- Python version: pyarrow==2.0.0\r\n- PyArrow version: 3.7.8\r\n","comment_length":30,"text":"Concurrent use of same dataset (already downloaded)\n## Describe the bug\r\nWhen launching several jobs at the same time loading the same dataset trigger some errors see (last comments).\r\n\r\n## Steps to reproduce the bug\r\nexport HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets\r\nfor MODEL in \"bert-base-uncased\" \"roberta-base\" \"distilbert-base-cased\"; do # \"bert-base-uncased\" \"bert-large-cased\" \"roberta-large\" \"albert-base-v1\" \"albert-large-v1\"; do\r\n for TASK_NAME in \"mrpc\" \"rte\" 'imdb' \"paws\" \"mnli\"; do\r\n export OUTPUT_DIR=${MODEL}_${TASK_NAME}\r\n sbatch --job-name=${OUTPUT_DIR} \\\r\n --gres=gpu:1 \\\r\n --no-requeue \\\r\n --cpus-per-task=10 \\\r\n --hint=nomultithread \\\r\n --time=1:00:00 \\\r\n --output=jobinfo\/${OUTPUT_DIR}_%j.out \\\r\n --error=jobinfo\/${OUTPUT_DIR}_%j.err \\\r\n --qos=qos_gpu-t4 \\\r\n --wrap=\"module purge; module load pytorch-gpu\/py3\/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=\/gpfswork\/rech\/toto\/transformers_models\/$MODEL\"\r\n\r\n done\r\ndone\r\n\r\n\r\n\r\n```python\r\n# Sample code to reproduce the bug\r\n dataset_train = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter)))\r\n\r\n dataset_val = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter, args.filter + 5000)))\r\n\r\n dataset_test = load_dataset('imdb', split='test', download_mode=\"reuse_cache_if_exists\")\r\n dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True)\r\n```\r\n\r\n## Expected results\r\nI believe I am doing something wrong with the objects. \r\n\r\n## Actual results\r\nTraceback (most recent call last):\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 983, in _prepare_split\r\n check_duplicates=True,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/gpfswork\/rech\/tts\/unm25jp\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets==1.8.0\r\n- Platform: linux (jeanzay)\r\n- Python version: pyarrow==2.0.0\r\n- PyArrow version: 3.7.8\r\n\nYou can probably have a solution much faster than me (first time I use the library). But I suspect some write function are used when loading the dataset from cache.","embeddings":[-0.5779120326,-0.0810000375,-0.0453643724,0.4501565397,0.2459626794,0.0139751369,0.5565331578,0.2723817527,0.1843708605,0.1907434464,-0.0311780032,0.1464349478,0.0212427545,0.1441881359,-0.0885511786,0.0235604327,0.1309166104,-0.2257413864,-0.3968794644,-0.0203631092,-0.2406820655,0.1446244419,0.0855503231,0.1932705194,-0.5113247037,-0.3433853984,-0.0541765466,0.0557928272,0.2590035498,-0.0671327934,0.1664633751,0.2859389484,-0.0116243837,0.8678808808,-0.0001090042,0.1391353309,0.1289675683,0.1109726653,-0.0634465516,-0.2272718996,0.0330815762,-0.107974872,0.1186949238,-0.313777715,0.1189511046,0.0584998429,-0.030055631,-0.2093538344,0.5382676721,0.0736098737,0.2384028584,0.3819690347,-0.3065969348,-0.2297079861,0.1842454374,0.0569794737,0.0860361606,0.0703067929,0.1006150246,-0.160428226,-0.0243540276,0.4298269153,-0.128874436,0.205970034,-0.0231342856,0.0869602188,0.1051392555,-0.5513936877,0.0767241269,0.0688892454,-0.0115327006,-0.3123033047,-0.3542192578,-0.3473508954,-0.0639272109,-0.2551651597,0.0928215683,0.345266372,-0.2033960521,-0.1923074573,-0.0920517072,0.1867493987,-0.0755044073,-0.1349960566,0.2415366471,0.330652684,0.0830517709,0.2387527227,0.0692323297,0.0927333981,0.3280492723,-0.1612951756,-0.1014361978,-0.0775532722,-0.6200729012,0.0457460396,-0.0160569604,-0.465223223,0.1843742728,-0.1858731359,-0.1885086894,0.0499862283,0.1717092693,0.1031510457,0.2989788651,0.0960013494,-0.3177445531,0.2705483735,0.3438619673,0.1331980228,-0.5029964447,0.0652742609,-0.0156139219,-0.2567523718,0.2176098078,0.1873345226,0.2414573133,-0.0829889849,-0.3572193086,-0.0523138382,-0.0511178933,-0.0558524579,-0.1014149934,0.4524842501,0.1777906567,0.1567742676,0.0913131386,0.1969440281,-0.1822924018,0.0624874271,-0.3012304008,-0.0290754717,-0.0146923279,0.2143333256,0.115738757,-0.0643138587,0.2248848528,0.337944448,0.0498892926,-0.4685381353,0.262468487,-0.3657260239,-0.2471660674,0.1961344182,0.1851789206,0.1372065544,-0.1051604077,0.3842988014,0.0134874908,0.2587327957,-0.3288044035,-0.1393946558,0.080968149,0.2727389634,0.0137452884,0.1152774841,-0.3554980457,0.11369133,0.31862849,-0.0416694805,0.1086383611,-0.3720645308,-0.3164130151,-0.1393729448,-0.0205443148,0.5279052258,0.0768126845,0.0268207006,0.1770918816,-0.1090108529,0.0915662721,0.3510977328,-0.3423242867,0.4966166019,-0.3493314087,0.1288179904,0.2440502197,-0.3719643652,-0.1516722888,0.3383845389,-0.1646658182,0.1252206415,0.2112877369,-0.0568555631,0.2122339308,-0.135921821,-0.0084830578,0.1491883695,-0.1079862937,0.4015436471,-0.1918611526,-0.2839153409,0.0489497297,0.1688934714,0.3223809302,-0.093997553,0.0269821547,-0.2874779701,0.3875817657,-0.3304336965,-0.073193185,0.2158527076,-0.0103004053,0.0795356706,0.0810648277,-0.1096467897,-0.5169209242,0.1978815943,-0.3624865711,0.0432497934,0.0342297591,-0.1991609484,0.1067375392,-0.041770719,-0.2488535792,-0.188830331,0.1706272364,0.23171404,-0.1502679437,-0.1181374565,-0.1473599523,0.3599050343,-0.1582805365,0.0234616175,-0.0023934881,0.1705364734,-0.1951894164,-0.2185277343,0.1162100509,-0.0603484437,0.4591343403,-0.1816609353,-0.1646382064,0.234481588,0.1979359686,0.3151443601,-0.060830038,0.0168911535,0.1874209791,0.155690819,-0.0017908275,0.1425172091,0.264064461,-0.1452902257,0.0974556208,0.3178672493,0.0033611003,0.1383701563,0.0590600222,-0.2233062983,0.193423599,-0.1233381256,0.1891794056,0.1224577054,0.2253106385,0.1255556643,0.0167566203,-0.0816582367,-0.5741452575,0.0836443007,0.3775579929,0.1669413447,0.0734937191,0.0733089,-0.0474899113,0.0227888245,-0.1682934314,0.299851954,0.5353221893,0.1349111348,0.0457241945,0.0882089511,-0.1451271623,-0.0892191976,0.0666768849,0.1546055675,0.1948834658,0.5548338294,0.1994925439,-0.0699805468,-0.2803615928,-0.1119266897,0.2671932578,0.0941753089,-0.3748239875,0.1850156933,0.1274146289,-0.1747792363,-0.4213629663,0.0054045664,-0.0335235819,-0.1470950544,-0.0398527905,0.4568405449,-0.1997360289,0.3671449721,-0.0299287681,0.1571482718,-0.1867981851,0.2361011505,-0.1471163779,-0.0711986795,-0.1141485125,-0.0113953063,0.2849563956,0.0669397041,0.3187500536,0.0719447806,-0.3169528842,-0.3599549234,-0.0033945311,0.0783701614,0.0017772753,0.4231902361,-0.0844049677,-0.1965941489,0.0808480829,-0.4469995499,0.2024542987,-0.3591762781,-0.0702030137,-0.1327766627,-0.2264376134,-0.0671751946,-0.1766142398,-0.3789281249,-0.2513652146,-0.2676903903,-0.0906180367,-0.1727518439,0.2116888314,0.0763652846,-0.1411370635,-0.0400552638,0.1537126452,-0.0046703178,-0.3079814911,-0.3230585456,0.069326885,-0.1893555969,-0.3206602931,-0.0872680172,-0.14780882,0.0682690814,0.226744622,-0.4916568995,-0.3557165861,-0.1071279123,0.0577950142,-0.1461234093,-0.0371046998,0.0506975763,0.0610218383,-0.1380059868,-0.1171887517,-0.1728388518,0.2016058713,-0.0312829688,0.0052557569,-0.165980652,0.2671523094,-0.1580260098,0.6628261209,0.2449266016,-0.0611025915,0.3297878206,-0.1419707984,0.0786561295,-0.0182802659,-0.4407642186,-0.0686313659,-0.4300889671,-0.2329535484,0.1423813105,-0.3390009105,-0.1655286103,-0.0883629695,0.0207866021,-0.3483386338,-0.3111569285,0.1051787734,-0.156896323,0.3320013583,0.0473193973,0.1608238518,0.0032935208,0.0731304511,0.2852947414,0.0372188538,0.2085292339,-0.18799676,-0.0320747495,0.0756293982,-0.1916020811,0.3207663596,0.0992457867,0.1881765425,-0.0998482257,0.211159125,0.0173973516,-0.1154941842,0.8625531793,-0.1944365799,0.3259494305,0.1943781078,-0.3524224162,-0.0848695338,-0.1915691793,-0.3549926281,-0.211286366,0.2583167553,0.6643010974,-0.3218054175,-0.1237827912,0.1433633864,-0.0208376907,-0.2824375927,-0.3151057065,-0.1605422795,-0.3001244068,-0.2805768847,0.1758775413,-0.0905000865,0.1202875748,-0.5292029977,-0.0838713869,-0.1814909875,0.2068257481,-0.0105721066,0.058843255,0.2804206014,-0.0604163967,0.4039429128,0.1262611002,-0.0301582571,0.27479738,0.446331203,0.0201088637,-0.2770265043,0.00195421,0.0463416502,0.0897724777,0.4556978643,0.0272614881,0.1352091879,0.067100428,0.2680851817,-0.0672242194,0.1780904233,0.3836325109,0.2200000435,-0.1989115477,-0.3477317393,0.0931109264,0.2356354594,-0.1319556832,0.2889435291,-0.5127384663,-0.2068072408,0.3026010692,-0.0046676905,0.6622441411,0.014262584,0.0911543891,0.3084231317,-0.30280599,0.2865456343,-0.2150480598,0.3221714497,-0.58214885,-0.2683508396,0.1158726588,-0.1270323396,-0.0752488151,0.0526764467,-0.1140703261,0.3120547235,-0.0008438026,0.0324366987,-0.247971788,0.2446663976,0.0065662968,-0.3198530078,0.0531232022,0.1187562793,-0.2992705405,0.2569264472,-0.1400966644,-0.1082457006,0.2545025647,-0.2358598262,-0.2938596904,0.2325545698,-0.3484944701,0.280626297,-0.1433753818,-0.1927305907,0.0463992357,-0.0866825432,0.1560348868,-0.228722617,0.0731344894,0.0957627669,0.0568758026,0.1381038725,-0.2318829894,-0.1537003666,0.4277762473,0.0211631749,-0.3620678484,-0.0820333734,-0.037336532,-0.0148330815,-0.1281271279,0.0481154695,-0.1951638013,-0.1983331442,-0.1131099835,0.2236150652,-0.0742343813,-0.3154717684,0.1518261433,-0.2507011592,-0.2737055421,0.3905716538,-0.1238662377,-0.3739267588,-0.0774779171,0.2204256654,0.1755113006,0.0789951086,0.4106520414,0.3367050588,-0.0640727431,-0.3133040071,-0.238447547,0.1598051339,-0.4164747894,0.1772857159,-0.1435779184,0.1515220255,-0.0273460206,0.3164765239,0.0630009547,0.1205942333,0.0206558891,-0.2940013111,-0.2153663486,-0.0177845843,-0.0637846068,0.2160811573,0.2012470365,0.2618279457,0.0282044765,0.3244502544,-0.3282782733,0.2335392833,-0.2057464421,0.1308392584,0.0513715334,-0.131349355,0.0827395394,0.2397816777,0.0975585505,0.2197265029,-0.1647183448,-0.2612218559,-0.2263840884,0.0987549499,-0.0208422579,0.0202253181,-0.1477030218,-0.1334641278,0.1280261278,-0.0183768459,0.4289841652,0.2284487486,-0.164397791,0.1061610952,0.1367555261,0.1648219824,-0.0818796307,0.1702706814,0.00027305,-0.0996422619,0.1888261735,0.182882145,-0.0026518188,0.0396514907,-0.2979305983,-0.1828772873,0.41345644,0.1560270488,0.0929486826,-0.3851540387,-0.2239552438,0.0314513445,0.4500091374,0.3133749664,-0.3114973307,-0.0574878529,-0.0370501876,0.2802079916,-0.1861449033,-0.0252606235,-0.0004438632,-0.1982536614,0.0170283131,0.3708457351,0.322214067,0.0491976589,-0.0257169344,0.1577520519,0.3708704114,-0.3861029744,0.2404301912,-0.0043916507,-0.1260692775,0.1273742318,0.4078056812,0.2264866531,0.0797451213,0.178853035,0.0911041051,0.1854630262,-0.0301674884,-0.1411308199,0.1116839126,-0.3086596727,-0.0381123275,0.2016683668,-0.2263548374,-0.1244863719,-0.2119970918,-0.0688010901,-0.187140882,-0.2287652493,-0.3681076467,0.2676638961,-0.3974691629,-0.0481058285,0.1096312627,-0.1442805678,0.1278003901,0.2010710239,0.1140565127,0.1347726285,0.4494643509,0.133113876,-0.2817566395,-0.1104476377,-0.3163566291,0.2716574669,0.1550415754,-0.3163852394,0.2832721472,0.3418689966,-0.0504980497,-0.2370741516,0.552675724,0.6522240639,0.3537182212,-0.145219028,0.0072202305,0.0328731798,0.1318574995,-0.0833300203,0.0924024582,0.1320800632,0.2738418281,0.3750738502,0.1627761424,-0.1621029526,0.1509012282,0.2157594413,0.1737514883,-0.2635540366,0.2917876244,-0.0475654304,-0.0709339604,-0.0687781051,0.0109695336,-0.3325090408,0.1167241484,0.4725090563,-0.0993127301,0.2256483436,0.0002578047,0.1222921833,0.0116997818,0.3831605613,0.1213125065,0.100687854,-0.6743170619,0.0617796257,-0.4572269917,0.2957102358,-0.1418879032,0.1963327527,-0.0430551097,0.1047224849,-0.2969474494,-0.0652521476,0.2955662906,-0.0407785922,0.0765402913,0.4635721445,-0.5969395638,0.0495046973,-0.3352076411,-0.1262418479,0.3455519378,-0.4687039554,0.2194332629,-0.1522485614,0.0916097686,0.1163392141,-0.0036385118,0.3093613982,0.1244528666,0.5834572911,0.1750371456,0.6061421037,-0.1260510236,0.1312722862,-0.1131600216,-0.4133682251,-0.3300965428,-0.006951638,0.1493000537,0.1066098064,-0.0756934956,-0.0927477553,-0.3385125399,-0.1523738354,0.0260592829,-0.1396056712,-0.0495962091,0.0400649644,0.0016709572,0.0484994873,0.1314331293,0.2882748544,0.2544392049,0.2327956855,-0.4348084629,-0.3158876002,0.5066917539,-0.2032601684,-0.5028808713,-0.3684080839,0.3902992606,-0.0260889083,0.0840148702,-0.625706315,0.1361856461,0.3196178377,-0.0780654475,-0.1940839589,0.1666283607,0.034007255,-0.1527427286,0.1009978428,-0.0439530723,0.0453941226,-0.3636941314,0.0277081151,-0.0606810227]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2728","title":"Concurrent use of same dataset (already downloaded)","comments":"I have the same issue:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 1040, in _prepare_split\r\n with ArrowWriter(features=self.info.features, path=fpath) as writer:\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/dccstor\/tslm-gen\/.cache\/csv\/default-387f1f95c084d4df\/0.0.0\/2dc6629a9ff6b5697d82c25b73731dd440507a69cbce8b425db50b751e8fcfd0.incomplete\/csv-validation.arrow'. Detail: [errno 2] No such file or directory\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\/dccstor\/tslm\/elron\/tslm-gen\/train.py\", line 510, in \r\n main()\r\n File \"\/dccstor\/tslm\/elron\/tslm-gen\/train.py\", line 246, in main\r\n datasets = prepare_dataset(dataset_args, logger)\r\n File \"\/dccstor\/tslm\/elron\/tslm-gen\/data.py\", line 157, in prepare_dataset\r\n datasets = load_dataset(extension, data_files=data_files, split=dataset_split, cache_dir=dataset_args.dataset_cache_dir, na_filter=False, download_mode=dataset_args.dataset_generate_mode)\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 742, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 654, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file. \r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/dccstor\/tslm-gen\/.cache\/csv\/default-387f1f95c084d4df\/0.0.0\/2dc6629a9ff6b5697d82c25b73731dd440507a69cbce8b425db50b751e8fcfd0.incomplete\/csv-validation.arrow'. Detail: [errno 2] No such file or directory\r\n```","body":"## Describe the bug\r\nWhen launching several jobs at the same time loading the same dataset trigger some errors see (last comments).\r\n\r\n## Steps to reproduce the bug\r\nexport HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets\r\nfor MODEL in \"bert-base-uncased\" \"roberta-base\" \"distilbert-base-cased\"; do # \"bert-base-uncased\" \"bert-large-cased\" \"roberta-large\" \"albert-base-v1\" \"albert-large-v1\"; do\r\n for TASK_NAME in \"mrpc\" \"rte\" 'imdb' \"paws\" \"mnli\"; do\r\n export OUTPUT_DIR=${MODEL}_${TASK_NAME}\r\n sbatch --job-name=${OUTPUT_DIR} \\\r\n --gres=gpu:1 \\\r\n --no-requeue \\\r\n --cpus-per-task=10 \\\r\n --hint=nomultithread \\\r\n --time=1:00:00 \\\r\n --output=jobinfo\/${OUTPUT_DIR}_%j.out \\\r\n --error=jobinfo\/${OUTPUT_DIR}_%j.err \\\r\n --qos=qos_gpu-t4 \\\r\n --wrap=\"module purge; module load pytorch-gpu\/py3\/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=\/gpfswork\/rech\/toto\/transformers_models\/$MODEL\"\r\n\r\n done\r\ndone\r\n\r\n\r\n\r\n```python\r\n# Sample code to reproduce the bug\r\n dataset_train = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter)))\r\n\r\n dataset_val = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter, args.filter + 5000)))\r\n\r\n dataset_test = load_dataset('imdb', split='test', download_mode=\"reuse_cache_if_exists\")\r\n dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True)\r\n```\r\n\r\n## Expected results\r\nI believe I am doing something wrong with the objects. \r\n\r\n## Actual results\r\nTraceback (most recent call last):\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 983, in _prepare_split\r\n check_duplicates=True,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/gpfswork\/rech\/tts\/unm25jp\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets==1.8.0\r\n- Platform: linux (jeanzay)\r\n- Python version: pyarrow==2.0.0\r\n- PyArrow version: 3.7.8\r\n","comment_length":172,"text":"Concurrent use of same dataset (already downloaded)\n## Describe the bug\r\nWhen launching several jobs at the same time loading the same dataset trigger some errors see (last comments).\r\n\r\n## Steps to reproduce the bug\r\nexport HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets\r\nfor MODEL in \"bert-base-uncased\" \"roberta-base\" \"distilbert-base-cased\"; do # \"bert-base-uncased\" \"bert-large-cased\" \"roberta-large\" \"albert-base-v1\" \"albert-large-v1\"; do\r\n for TASK_NAME in \"mrpc\" \"rte\" 'imdb' \"paws\" \"mnli\"; do\r\n export OUTPUT_DIR=${MODEL}_${TASK_NAME}\r\n sbatch --job-name=${OUTPUT_DIR} \\\r\n --gres=gpu:1 \\\r\n --no-requeue \\\r\n --cpus-per-task=10 \\\r\n --hint=nomultithread \\\r\n --time=1:00:00 \\\r\n --output=jobinfo\/${OUTPUT_DIR}_%j.out \\\r\n --error=jobinfo\/${OUTPUT_DIR}_%j.err \\\r\n --qos=qos_gpu-t4 \\\r\n --wrap=\"module purge; module load pytorch-gpu\/py3\/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=\/gpfswork\/rech\/toto\/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=\/gpfswork\/rech\/toto\/transformers_models\/$MODEL\"\r\n\r\n done\r\ndone\r\n\r\n\r\n\r\n```python\r\n# Sample code to reproduce the bug\r\n dataset_train = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter)))\r\n\r\n dataset_val = load_dataset('imdb', split='train', download_mode=\"reuse_cache_if_exists\")\r\n dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True).select(list(range(args.filter, args.filter + 5000)))\r\n\r\n dataset_test = load_dataset('imdb', split='test', download_mode=\"reuse_cache_if_exists\")\r\n dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'),\r\n batched=True)\r\n```\r\n\r\n## Expected results\r\nI believe I am doing something wrong with the objects. \r\n\r\n## Actual results\r\nTraceback (most recent call last):\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 983, in _prepare_split\r\n check_duplicates=True,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/gpfswork\/rech\/tts\/unm25jp\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"compute_measures.py\", line 181, in \r\n train_loader, val_loader, test_loader = get_dataloader(args)\r\n File \"\/gpfsdswork\/projects\/rech\/toto\/intRAOcular\/dataset_utils.py\", line 69, in get_dataloader\r\n dataset_train = load_dataset('paws', \"labeled_final\", split='train', download_mode=\"reuse_cache_if_exists\")\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/gpfslocalsup\/pub\/anaconda-py3\/2020.02\/envs\/pytorch-gpu-1.7.0\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/gpfswork\/rech\/toto\/datasets\/paws\/labeled_final\/1.1.0\/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete\/paws-test.arrow'. Detail: [errno 2] No such file or directory\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets==1.8.0\r\n- Platform: linux (jeanzay)\r\n- Python version: pyarrow==2.0.0\r\n- PyArrow version: 3.7.8\r\n\nI have the same issue:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 1040, in _prepare_split\r\n with ArrowWriter(features=self.info.features, path=fpath) as writer:\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/arrow_writer.py\", line 192, in __init__\r\n self.stream = pa.OSFile(self._path, \"wb\")\r\n File \"pyarrow\/io.pxi\", line 829, in pyarrow.lib.OSFile.__cinit__\r\n File \"pyarrow\/io.pxi\", line 844, in pyarrow.lib.OSFile._open_writable\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/dccstor\/tslm-gen\/.cache\/csv\/default-387f1f95c084d4df\/0.0.0\/2dc6629a9ff6b5697d82c25b73731dd440507a69cbce8b425db50b751e8fcfd0.incomplete\/csv-validation.arrow'. Detail: [errno 2] No such file or directory\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\/dccstor\/tslm\/elron\/tslm-gen\/train.py\", line 510, in \r\n main()\r\n File \"\/dccstor\/tslm\/elron\/tslm-gen\/train.py\", line 246, in main\r\n datasets = prepare_dataset(dataset_args, logger)\r\n File \"\/dccstor\/tslm\/elron\/tslm-gen\/data.py\", line 157, in prepare_dataset\r\n datasets = load_dataset(extension, data_files=data_files, split=dataset_split, cache_dir=dataset_args.dataset_cache_dir, na_filter=False, download_mode=dataset_args.dataset_generate_mode)\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 742, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/tslm\/envs\/anaconda3\/envs\/trf-a100\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 654, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file. \r\nOriginal error:\r\n[Errno 2] Failed to open local file '\/dccstor\/tslm-gen\/.cache\/csv\/default-387f1f95c084d4df\/0.0.0\/2dc6629a9ff6b5697d82c25b73731dd440507a69cbce8b425db50b751e8fcfd0.incomplete\/csv-validation.arrow'. Detail: [errno 2] No such file or directory\r\n```","embeddings":[-0.5779120326,-0.0810000375,-0.0453643724,0.4501565397,0.2459626794,0.0139751369,0.5565331578,0.2723817527,0.1843708605,0.1907434464,-0.0311780032,0.1464349478,0.0212427545,0.1441881359,-0.0885511786,0.0235604327,0.1309166104,-0.2257413864,-0.3968794644,-0.0203631092,-0.2406820655,0.1446244419,0.0855503231,0.1932705194,-0.5113247037,-0.3433853984,-0.0541765466,0.0557928272,0.2590035498,-0.0671327934,0.1664633751,0.2859389484,-0.0116243837,0.8678808808,-0.0001090042,0.1391353309,0.1289675683,0.1109726653,-0.0634465516,-0.2272718996,0.0330815762,-0.107974872,0.1186949238,-0.313777715,0.1189511046,0.0584998429,-0.030055631,-0.2093538344,0.5382676721,0.0736098737,0.2384028584,0.3819690347,-0.3065969348,-0.2297079861,0.1842454374,0.0569794737,0.0860361606,0.0703067929,0.1006150246,-0.160428226,-0.0243540276,0.4298269153,-0.128874436,0.205970034,-0.0231342856,0.0869602188,0.1051392555,-0.5513936877,0.0767241269,0.0688892454,-0.0115327006,-0.3123033047,-0.3542192578,-0.3473508954,-0.0639272109,-0.2551651597,0.0928215683,0.345266372,-0.2033960521,-0.1923074573,-0.0920517072,0.1867493987,-0.0755044073,-0.1349960566,0.2415366471,0.330652684,0.0830517709,0.2387527227,0.0692323297,0.0927333981,0.3280492723,-0.1612951756,-0.1014361978,-0.0775532722,-0.6200729012,0.0457460396,-0.0160569604,-0.465223223,0.1843742728,-0.1858731359,-0.1885086894,0.0499862283,0.1717092693,0.1031510457,0.2989788651,0.0960013494,-0.3177445531,0.2705483735,0.3438619673,0.1331980228,-0.5029964447,0.0652742609,-0.0156139219,-0.2567523718,0.2176098078,0.1873345226,0.2414573133,-0.0829889849,-0.3572193086,-0.0523138382,-0.0511178933,-0.0558524579,-0.1014149934,0.4524842501,0.1777906567,0.1567742676,0.0913131386,0.1969440281,-0.1822924018,0.0624874271,-0.3012304008,-0.0290754717,-0.0146923279,0.2143333256,0.115738757,-0.0643138587,0.2248848528,0.337944448,0.0498892926,-0.4685381353,0.262468487,-0.3657260239,-0.2471660674,0.1961344182,0.1851789206,0.1372065544,-0.1051604077,0.3842988014,0.0134874908,0.2587327957,-0.3288044035,-0.1393946558,0.080968149,0.2727389634,0.0137452884,0.1152774841,-0.3554980457,0.11369133,0.31862849,-0.0416694805,0.1086383611,-0.3720645308,-0.3164130151,-0.1393729448,-0.0205443148,0.5279052258,0.0768126845,0.0268207006,0.1770918816,-0.1090108529,0.0915662721,0.3510977328,-0.3423242867,0.4966166019,-0.3493314087,0.1288179904,0.2440502197,-0.3719643652,-0.1516722888,0.3383845389,-0.1646658182,0.1252206415,0.2112877369,-0.0568555631,0.2122339308,-0.135921821,-0.0084830578,0.1491883695,-0.1079862937,0.4015436471,-0.1918611526,-0.2839153409,0.0489497297,0.1688934714,0.3223809302,-0.093997553,0.0269821547,-0.2874779701,0.3875817657,-0.3304336965,-0.073193185,0.2158527076,-0.0103004053,0.0795356706,0.0810648277,-0.1096467897,-0.5169209242,0.1978815943,-0.3624865711,0.0432497934,0.0342297591,-0.1991609484,0.1067375392,-0.041770719,-0.2488535792,-0.188830331,0.1706272364,0.23171404,-0.1502679437,-0.1181374565,-0.1473599523,0.3599050343,-0.1582805365,0.0234616175,-0.0023934881,0.1705364734,-0.1951894164,-0.2185277343,0.1162100509,-0.0603484437,0.4591343403,-0.1816609353,-0.1646382064,0.234481588,0.1979359686,0.3151443601,-0.060830038,0.0168911535,0.1874209791,0.155690819,-0.0017908275,0.1425172091,0.264064461,-0.1452902257,0.0974556208,0.3178672493,0.0033611003,0.1383701563,0.0590600222,-0.2233062983,0.193423599,-0.1233381256,0.1891794056,0.1224577054,0.2253106385,0.1255556643,0.0167566203,-0.0816582367,-0.5741452575,0.0836443007,0.3775579929,0.1669413447,0.0734937191,0.0733089,-0.0474899113,0.0227888245,-0.1682934314,0.299851954,0.5353221893,0.1349111348,0.0457241945,0.0882089511,-0.1451271623,-0.0892191976,0.0666768849,0.1546055675,0.1948834658,0.5548338294,0.1994925439,-0.0699805468,-0.2803615928,-0.1119266897,0.2671932578,0.0941753089,-0.3748239875,0.1850156933,0.1274146289,-0.1747792363,-0.4213629663,0.0054045664,-0.0335235819,-0.1470950544,-0.0398527905,0.4568405449,-0.1997360289,0.3671449721,-0.0299287681,0.1571482718,-0.1867981851,0.2361011505,-0.1471163779,-0.0711986795,-0.1141485125,-0.0113953063,0.2849563956,0.0669397041,0.3187500536,0.0719447806,-0.3169528842,-0.3599549234,-0.0033945311,0.0783701614,0.0017772753,0.4231902361,-0.0844049677,-0.1965941489,0.0808480829,-0.4469995499,0.2024542987,-0.3591762781,-0.0702030137,-0.1327766627,-0.2264376134,-0.0671751946,-0.1766142398,-0.3789281249,-0.2513652146,-0.2676903903,-0.0906180367,-0.1727518439,0.2116888314,0.0763652846,-0.1411370635,-0.0400552638,0.1537126452,-0.0046703178,-0.3079814911,-0.3230585456,0.069326885,-0.1893555969,-0.3206602931,-0.0872680172,-0.14780882,0.0682690814,0.226744622,-0.4916568995,-0.3557165861,-0.1071279123,0.0577950142,-0.1461234093,-0.0371046998,0.0506975763,0.0610218383,-0.1380059868,-0.1171887517,-0.1728388518,0.2016058713,-0.0312829688,0.0052557569,-0.165980652,0.2671523094,-0.1580260098,0.6628261209,0.2449266016,-0.0611025915,0.3297878206,-0.1419707984,0.0786561295,-0.0182802659,-0.4407642186,-0.0686313659,-0.4300889671,-0.2329535484,0.1423813105,-0.3390009105,-0.1655286103,-0.0883629695,0.0207866021,-0.3483386338,-0.3111569285,0.1051787734,-0.156896323,0.3320013583,0.0473193973,0.1608238518,0.0032935208,0.0731304511,0.2852947414,0.0372188538,0.2085292339,-0.18799676,-0.0320747495,0.0756293982,-0.1916020811,0.3207663596,0.0992457867,0.1881765425,-0.0998482257,0.211159125,0.0173973516,-0.1154941842,0.8625531793,-0.1944365799,0.3259494305,0.1943781078,-0.3524224162,-0.0848695338,-0.1915691793,-0.3549926281,-0.211286366,0.2583167553,0.6643010974,-0.3218054175,-0.1237827912,0.1433633864,-0.0208376907,-0.2824375927,-0.3151057065,-0.1605422795,-0.3001244068,-0.2805768847,0.1758775413,-0.0905000865,0.1202875748,-0.5292029977,-0.0838713869,-0.1814909875,0.2068257481,-0.0105721066,0.058843255,0.2804206014,-0.0604163967,0.4039429128,0.1262611002,-0.0301582571,0.27479738,0.446331203,0.0201088637,-0.2770265043,0.00195421,0.0463416502,0.0897724777,0.4556978643,0.0272614881,0.1352091879,0.067100428,0.2680851817,-0.0672242194,0.1780904233,0.3836325109,0.2200000435,-0.1989115477,-0.3477317393,0.0931109264,0.2356354594,-0.1319556832,0.2889435291,-0.5127384663,-0.2068072408,0.3026010692,-0.0046676905,0.6622441411,0.014262584,0.0911543891,0.3084231317,-0.30280599,0.2865456343,-0.2150480598,0.3221714497,-0.58214885,-0.2683508396,0.1158726588,-0.1270323396,-0.0752488151,0.0526764467,-0.1140703261,0.3120547235,-0.0008438026,0.0324366987,-0.247971788,0.2446663976,0.0065662968,-0.3198530078,0.0531232022,0.1187562793,-0.2992705405,0.2569264472,-0.1400966644,-0.1082457006,0.2545025647,-0.2358598262,-0.2938596904,0.2325545698,-0.3484944701,0.280626297,-0.1433753818,-0.1927305907,0.0463992357,-0.0866825432,0.1560348868,-0.228722617,0.0731344894,0.0957627669,0.0568758026,0.1381038725,-0.2318829894,-0.1537003666,0.4277762473,0.0211631749,-0.3620678484,-0.0820333734,-0.037336532,-0.0148330815,-0.1281271279,0.0481154695,-0.1951638013,-0.1983331442,-0.1131099835,0.2236150652,-0.0742343813,-0.3154717684,0.1518261433,-0.2507011592,-0.2737055421,0.3905716538,-0.1238662377,-0.3739267588,-0.0774779171,0.2204256654,0.1755113006,0.0789951086,0.4106520414,0.3367050588,-0.0640727431,-0.3133040071,-0.238447547,0.1598051339,-0.4164747894,0.1772857159,-0.1435779184,0.1515220255,-0.0273460206,0.3164765239,0.0630009547,0.1205942333,0.0206558891,-0.2940013111,-0.2153663486,-0.0177845843,-0.0637846068,0.2160811573,0.2012470365,0.2618279457,0.0282044765,0.3244502544,-0.3282782733,0.2335392833,-0.2057464421,0.1308392584,0.0513715334,-0.131349355,0.0827395394,0.2397816777,0.0975585505,0.2197265029,-0.1647183448,-0.2612218559,-0.2263840884,0.0987549499,-0.0208422579,0.0202253181,-0.1477030218,-0.1334641278,0.1280261278,-0.0183768459,0.4289841652,0.2284487486,-0.164397791,0.1061610952,0.1367555261,0.1648219824,-0.0818796307,0.1702706814,0.00027305,-0.0996422619,0.1888261735,0.182882145,-0.0026518188,0.0396514907,-0.2979305983,-0.1828772873,0.41345644,0.1560270488,0.0929486826,-0.3851540387,-0.2239552438,0.0314513445,0.4500091374,0.3133749664,-0.3114973307,-0.0574878529,-0.0370501876,0.2802079916,-0.1861449033,-0.0252606235,-0.0004438632,-0.1982536614,0.0170283131,0.3708457351,0.322214067,0.0491976589,-0.0257169344,0.1577520519,0.3708704114,-0.3861029744,0.2404301912,-0.0043916507,-0.1260692775,0.1273742318,0.4078056812,0.2264866531,0.0797451213,0.178853035,0.0911041051,0.1854630262,-0.0301674884,-0.1411308199,0.1116839126,-0.3086596727,-0.0381123275,0.2016683668,-0.2263548374,-0.1244863719,-0.2119970918,-0.0688010901,-0.187140882,-0.2287652493,-0.3681076467,0.2676638961,-0.3974691629,-0.0481058285,0.1096312627,-0.1442805678,0.1278003901,0.2010710239,0.1140565127,0.1347726285,0.4494643509,0.133113876,-0.2817566395,-0.1104476377,-0.3163566291,0.2716574669,0.1550415754,-0.3163852394,0.2832721472,0.3418689966,-0.0504980497,-0.2370741516,0.552675724,0.6522240639,0.3537182212,-0.145219028,0.0072202305,0.0328731798,0.1318574995,-0.0833300203,0.0924024582,0.1320800632,0.2738418281,0.3750738502,0.1627761424,-0.1621029526,0.1509012282,0.2157594413,0.1737514883,-0.2635540366,0.2917876244,-0.0475654304,-0.0709339604,-0.0687781051,0.0109695336,-0.3325090408,0.1167241484,0.4725090563,-0.0993127301,0.2256483436,0.0002578047,0.1222921833,0.0116997818,0.3831605613,0.1213125065,0.100687854,-0.6743170619,0.0617796257,-0.4572269917,0.2957102358,-0.1418879032,0.1963327527,-0.0430551097,0.1047224849,-0.2969474494,-0.0652521476,0.2955662906,-0.0407785922,0.0765402913,0.4635721445,-0.5969395638,0.0495046973,-0.3352076411,-0.1262418479,0.3455519378,-0.4687039554,0.2194332629,-0.1522485614,0.0916097686,0.1163392141,-0.0036385118,0.3093613982,0.1244528666,0.5834572911,0.1750371456,0.6061421037,-0.1260510236,0.1312722862,-0.1131600216,-0.4133682251,-0.3300965428,-0.006951638,0.1493000537,0.1066098064,-0.0756934956,-0.0927477553,-0.3385125399,-0.1523738354,0.0260592829,-0.1396056712,-0.0495962091,0.0400649644,0.0016709572,0.0484994873,0.1314331293,0.2882748544,0.2544392049,0.2327956855,-0.4348084629,-0.3158876002,0.5066917539,-0.2032601684,-0.5028808713,-0.3684080839,0.3902992606,-0.0260889083,0.0840148702,-0.625706315,0.1361856461,0.3196178377,-0.0780654475,-0.1940839589,0.1666283607,0.034007255,-0.1527427286,0.1009978428,-0.0439530723,0.0453941226,-0.3636941314,0.0277081151,-0.0606810227]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2727","title":"Error in loading the Arabic Billion Words Corpus","comments":"I modified the dataset loading script to catch the `IndexError` and inspect the records at which the error is happening, and I found this:\r\nFor the `Techreen` config, the error happens in 36 records when trying to find the `Text` or `Dateline` tags. All these 36 records look something like:\r\n```\r\n\r\n TRN_ARB_0248167<\/ID>\r\n http:\/\/tishreen.news.sy\/tishreen\/public\/read\/248240<\/URL>\r\n Removed, because the original articles was in English<\/Headline>\r\n<\/Techreen>\r\n```\r\n\r\nand all the 288 faulty records in the `Almustaqbal` config look like:\r\n```\r\n\r\n MTL_ARB_0028398<\/ID>\r\n \r\n http:\/\/www.almustaqbal.com\/v4\/article.aspx?type=NP&ArticleID=179015<\/URL>\r\n Removed because it is not available in the original site<\/Headline>\r\n<\/Almustaqbal>\r\n```\r\n\r\nso the error is happening because the articles were removed and so the associated records lack the `Text` tag.\r\n\r\nIn this case, I think we just need to catch the `IndexError` and ignore (pass) it.\r\n","body":"## Describe the bug\r\nI get `IndexError: list index out of range` when trying to load the `Techreen` and `Almustaqbal` configs of the dataset.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nload_dataset(\"arabic_billion_words\", \"Techreen\")\r\nload_dataset(\"arabic_billion_words\", \"Almustaqbal\")\r\n```\r\n\r\n## Expected results\r\nThe datasets load succefully.\r\n\r\n## Actual results\r\n```python\r\n_extract_tags(self, sample, tag)\r\n 139 if len(out) > 0:\r\n 140 break\r\n--> 141 return out[0]\r\n 142 \r\n 143 def _clean_text(self, text):\r\n\r\nIndexError: list index out of range\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.2\r\n- Platform: Ubuntu 18.04.5 LTS\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":128,"text":"Error in loading the Arabic Billion Words Corpus\n## Describe the bug\r\nI get `IndexError: list index out of range` when trying to load the `Techreen` and `Almustaqbal` configs of the dataset.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nload_dataset(\"arabic_billion_words\", \"Techreen\")\r\nload_dataset(\"arabic_billion_words\", \"Almustaqbal\")\r\n```\r\n\r\n## Expected results\r\nThe datasets load succefully.\r\n\r\n## Actual results\r\n```python\r\n_extract_tags(self, sample, tag)\r\n 139 if len(out) > 0:\r\n 140 break\r\n--> 141 return out[0]\r\n 142 \r\n 143 def _clean_text(self, text):\r\n\r\nIndexError: list index out of range\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.2\r\n- Platform: Ubuntu 18.04.5 LTS\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nI modified the dataset loading script to catch the `IndexError` and inspect the records at which the error is happening, and I found this:\r\nFor the `Techreen` config, the error happens in 36 records when trying to find the `Text` or `Dateline` tags. All these 36 records look something like:\r\n```\r\n\r\n TRN_ARB_0248167<\/ID>\r\n http:\/\/tishreen.news.sy\/tishreen\/public\/read\/248240<\/URL>\r\n Removed, because the original articles was in English<\/Headline>\r\n<\/Techreen>\r\n```\r\n\r\nand all the 288 faulty records in the `Almustaqbal` config look like:\r\n```\r\n\r\n MTL_ARB_0028398<\/ID>\r\n \r\n http:\/\/www.almustaqbal.com\/v4\/article.aspx?type=NP&ArticleID=179015<\/URL>\r\n Removed because it is not available in the original site<\/Headline>\r\n<\/Almustaqbal>\r\n```\r\n\r\nso the error is happening because the articles were removed and so the associated records lack the `Text` tag.\r\n\r\nIn this case, I think we just need to catch the `IndexError` and ignore (pass) it.\r\n","embeddings":[-0.1755402386,0.106749475,-0.1183999553,0.4231290519,-0.1488431394,0.2618189752,0.2323353887,0.3904677033,0.3619009554,-0.0010758634,-0.2331460863,-0.0116897542,0.058974117,0.0111060264,-0.0046201525,-0.2025989443,-0.030889973,-0.1021867767,0.2275917828,0.14214845,-0.1725380868,0.088298142,-0.2301309407,-0.0897333324,0.1276371777,-0.0596874207,0.0723104551,0.0443104617,0.2043147385,-0.5446294546,-0.0048863036,-0.4125088155,0.4355430603,0.4883056283,-0.0001138216,0.0546824187,0.3437693715,-0.0272877552,-0.3930965066,-0.4340239763,-0.3352651596,-0.2002716511,0.0597978495,-0.2232355028,-0.0848990083,-0.2553299963,-0.1447215676,-0.3347806036,0.1664268374,0.5260614753,0.1590858251,0.1715174317,0.1333598644,0.106060259,0.3838208318,-0.2292442918,-0.0983324423,0.2020831853,0.5291868448,0.1876037419,-0.0844436958,0.2945673764,-0.1038550511,-0.0169019997,0.1182348877,-0.0495264716,0.1696146429,-0.4821276665,0.0659714937,0.2115795463,0.3448188007,-0.3105709553,-0.4523892105,-0.2377697527,-0.1179611385,-0.3698241711,0.2565723658,0.2197514176,-0.195923686,0.129543364,0.1145651564,-0.0898964331,-0.0348887108,0.3676132858,-0.3026773036,0.3926457763,-0.077249445,-0.0298966225,0.1111189947,-0.1466866583,0.0816025436,0.002760611,0.0597463623,0.2146911174,-0.3444157541,0.2084279954,0.0736802295,-0.1433285624,0.0144621581,0.0827068537,0.2570385337,-0.034107022,-0.0585579835,0.0653540716,0.3505274355,0.1814717501,-0.1095268726,0.0925202146,-0.0106690628,0.303013593,0.0932591036,0.0778528303,0.1683073938,-0.2432312667,0.0530864,-0.0370576158,0.2929569185,-0.3523527086,-0.265804857,0.2261015177,-0.3138167858,-0.1832838506,0.0655586943,0.1817698777,-0.2392186075,-0.1005868986,0.2253467143,0.0679831579,-0.27737239,-0.0633981153,-0.1548104882,0.0935432166,-0.0421134941,0.1769756973,0.4418578446,-0.5744714141,0.1978181005,0.043781396,0.2652725875,-0.0362150185,0.0188351832,-0.4248972237,0.2651003003,0.4377249777,0.0096786022,0.0718528181,0.0458108075,0.0313299224,-0.1299722046,0.309678942,-0.2886221409,-0.1968882084,-0.0432505459,0.12648274,0.0378431939,0.1340029985,-0.1635893285,0.3765878677,0.5087185502,-0.1100702062,0.0502905846,-0.1411720365,-0.2798919678,-0.0035947154,0.1736454964,0.4014997184,-0.3967858851,-0.105560407,0.0184918921,0.565326333,-0.0560430773,0.1985335201,-0.2315893471,0.4627299011,-0.1715318412,0.2583889663,0.2030450553,-0.1243126094,-0.1949735135,0.1152663976,0.2389275134,-0.0143756084,-0.0578818396,-0.1760071516,0.4688828886,-0.1368215531,0.6135954261,0.344940871,-0.0424385816,-0.0966618583,-0.347731024,-0.1141509861,0.1973970234,0.1627766341,0.0313840397,-0.048485674,-0.0097026713,-0.2665247619,0.0788684636,0.1392221302,0.0231262427,0.3674274087,-0.6418525577,0.0916571841,0.2857526541,-0.2323861867,-0.1339323819,0.1271806806,-0.1287386864,0.2774348557,-0.3100445569,0.1869035959,-0.4867818058,0.015705755,-0.1359554082,0.1484376341,0.0637337714,0.0105744414,-0.1396457851,0.2280972898,-0.2554454803,-0.1438363194,0.068384625,-0.0334545262,-0.6968354583,0.2373879552,-0.2281327993,-0.13617751,0.1577715576,0.1020079553,0.140430674,-0.1101043522,-0.1807846725,-0.020392159,-0.1228406578,-0.1708034873,-0.3455619216,0.3240127861,0.3131725788,-0.3127225637,0.2518367469,0.405628264,0.2708936036,-0.0836238265,0.3261787295,0.2131666988,-0.1609645039,0.3445234895,-0.0217991453,-0.2369894981,0.2463639528,0.0210876931,-0.0098291654,-0.2796858251,0.2190893143,-0.1045974866,0.3646546006,0.0616709776,-0.0477095805,-0.058134757,0.1234448552,-0.0161477383,0.1691765636,0.2042669356,-0.0738272592,-0.1680463105,-0.1717485487,-0.3292060196,0.1732795835,0.107717663,-0.0089457426,-0.083401978,0.1157973781,-0.1058290303,0.3423880935,0.255276531,0.2736132741,0.2762809098,0.0783206522,0.1315883994,-0.3082402945,-0.4225620031,0.0142450947,0.2152878642,-0.4495734572,0.4153724015,0.1764909774,-0.3796938658,0.0393821411,-0.6723875403,-0.2069754153,-0.2337341309,0.0206673089,-0.1468473375,-0.2473726422,0.0721461475,-0.1190406457,0.09507218,0.2156872749,0.0018519799,-0.0147829652,0.0973429382,-0.3829988539,-0.0383320674,0.2576623559,-0.0327397026,0.095439449,-0.0960054174,-0.1535268277,-0.1855542362,-0.003608952,0.1915376484,0.0029507354,0.4744865894,0.0214510057,0.3548190296,-0.0678444281,-0.3108948767,0.1428106278,0.284109205,-0.0027109555,0.0835413486,0.0051922076,0.3517223001,0.0820932537,-0.6314830184,-0.2765565813,-0.3663812876,-0.2867331505,-0.0545874983,0.2838801742,0.3187383115,-0.0734038353,-0.1046390012,0.2247227281,0.3155885935,-0.2177784741,-0.0706695691,0.2444880903,-0.0777041316,-0.2715021968,-0.1034858078,-0.0088611031,0.0229867306,0.0225059278,-0.4185804427,0.3741729259,-0.2116332948,-0.1099837944,-0.2799408734,0.1427977234,0.1195821241,-0.1057285815,-0.0864008814,0.2175137401,-0.1457337588,-0.0083484845,-0.0701671541,0.1588692367,0.2455358207,0.3019486666,-0.150668174,0.2919903696,-0.0586923361,-0.040083617,0.6439816952,0.2175884843,0.1514039934,-0.2557828128,-0.3873148263,-0.1864755899,0.0446603149,-0.322828263,0.3679369986,0.0021882905,-0.4897554815,-0.2030827552,-0.3399727345,-0.4267130196,-0.1913820356,0.2030257285,-0.1593590677,0.2108831704,0.0237512253,0.0421076417,-0.0633242056,-0.1329182833,0.1005292386,0.0311089866,-0.0726397038,-0.0021495046,0.2656378448,-0.2039828449,-0.3477562368,0.1620821357,0.2984673977,0.4872823358,-0.055518087,-0.0622755252,0.0048182392,0.1248934716,0.6233374476,0.0114794467,0.2188987285,0.3644364178,0.1261979938,-0.1252692491,-0.2462923229,-0.1677238345,0.2955403924,0.1296832412,0.0469400845,-0.2580867112,0.0334393606,0.0349084251,0.5183107257,-0.1502606273,-0.1503088623,-0.2134931982,-0.2702845633,-0.1576328427,0.0489906408,0.2762345374,0.2790073454,0.1163906455,0.0565570034,0.1449835896,-0.0082677305,0.1182349697,0.0302531123,0.2791365683,-0.0354128219,0.2316610217,-0.0505239516,0.1022323072,0.1699858308,0.4633558691,-0.0541026555,-0.2049700469,0.0626578629,0.0439115539,-0.0007969564,0.2750033438,0.0230817329,-0.0759900287,0.1049807519,0.0703089163,0.0560400039,-0.2724578381,-0.0922067016,-0.0225837156,-0.543132782,-0.4410663545,0.4706727862,-0.0079253893,-0.1015082225,0.3908186257,-0.0670434907,-0.3913850188,0.8489490747,-0.0848357305,0.9151951671,-0.1068367735,0.0912295207,-0.0729310513,0.3846330047,0.2064215839,0.0115282154,0.1456759423,-0.2525363266,-0.298453033,-0.1454464644,-0.012256355,-0.0232700668,0.0328773223,-0.1993303746,0.320872128,-0.2007212043,-0.1426831037,0.0873766169,-0.0067092902,-0.1424009949,-0.3256973624,-0.7548496127,0.0061009666,-0.0925030112,0.0177585036,-0.1283669025,-0.0647783652,-0.0522778556,0.0719934851,-0.3233529925,0.0826198757,-0.2926440835,0.1916070729,-0.0383144841,0.0786379874,0.1929414421,-0.1405192018,0.244963035,0.0390207767,-0.1484766304,-0.1100704521,0.0302471835,-0.1444115639,0.1483397484,-0.1872265637,0.5013310313,-0.0442665927,-0.2843292654,0.1605215818,-0.0077373059,-0.1709920168,-0.1163547412,0.339533031,0.1112183854,-0.3739176691,-0.2430127263,-0.1658253223,0.1200292185,-0.1154507846,0.1192825288,0.1934636831,-0.1843617409,0.112421453,0.1926915348,-0.4189243019,-0.1223241836,0.4717422128,0.2463549078,0.1172112226,0.3903487325,0.6143891215,-0.1637526304,-0.0698994547,-0.1813173294,0.1079720631,-0.4279260933,0.1673279852,-0.0595007278,0.0256694928,0.0052667982,0.5612253547,0.2377902418,-0.0229390804,-0.1789640188,-0.6683347225,0.1777802259,0.3081204891,0.0797463804,-0.0309090503,0.190864265,-0.1966795921,-0.0598718114,-0.1727070808,-0.2727920115,0.1352400631,-0.3521441221,0.380798161,-0.5048600435,-0.0210790336,0.1642596573,-0.0796083882,0.0516792536,-0.0498385914,-0.1113034338,-0.2304140478,0.0514916629,0.1193002015,-0.1328906119,-0.2133136094,-0.3685189188,-0.2479917109,0.0559157878,-0.1663181633,0.096526593,0.253727138,0.0825098678,-0.0666557103,0.1753320545,-0.2093826979,-0.0640961975,0.2392829061,-0.1072063893,0.2693241835,0.1828237176,0.2660240233,-0.1483020633,-0.0021981644,-0.3165096045,0.1672480106,-0.1971392035,0.1155363247,0.2080385536,-0.3303750753,-0.0588245131,0.4746849537,0.2639428377,0.2856270075,-0.3358512521,-0.2472344786,-0.2202569693,0.1564177424,-0.227623418,-0.2707375586,-0.0806937665,-0.2140191644,0.0222404338,0.3439472616,0.3728134632,-0.317720592,-0.2651559711,0.0304209441,0.3469939232,0.0293370541,-0.0581984222,0.3112131059,0.1939305365,0.073518768,0.0554657876,-0.0581345744,0.2517088354,0.35873577,-0.0427223258,0.4240315855,0.5910558701,0.2392537147,-0.2086254805,-0.5564651489,0.4546756446,0.3678334355,0.1437536925,0.0243156049,0.1914880723,-0.3231008649,0.1907578856,-0.0923769102,-0.0231581293,-0.0302627981,-0.0183196161,0.0349920467,0.0932228342,-0.2088588774,0.141479969,-0.2800648212,-0.0208464339,-0.0735580623,0.0253787898,0.1334732771,-0.2514319122,-0.0405881815,0.0336346664,0.0703931525,0.1245529428,-0.159872368,0.314004302,0.2583769858,-0.1721871942,0.0302774105,0.4492284358,0.4798878133,0.4027010202,-0.0890972465,0.1645621508,-0.0922213271,-0.1725895405,-0.1757374704,0.0987889171,0.0299801119,0.1102682501,0.2625525892,0.0869524628,-0.0729662105,0.1525919586,0.3058074415,0.1193872318,0.0408316515,-0.0341429487,-0.1013062522,-0.0667460635,-0.2574473917,0.0409120657,-0.4404973388,-0.0617010742,0.5526071191,-0.4645683765,0.1704440713,0.1823160648,0.0622186549,-0.0459920913,0.470972389,0.596745491,0.1878376901,-0.4081811905,0.1642002463,-0.3384467959,0.1257899702,-0.1452896446,0.1523465365,0.5619260669,0.2891363502,0.0298849139,0.2199967355,0.1524373293,0.3632757962,-0.0070694583,0.0890556946,-0.501711905,-0.3325832784,-0.0525973402,0.3393879235,-0.335118711,-0.1767838746,0.19476448,-0.5119805336,0.0272406824,-0.2343552858,-0.3554559648,0.139610216,0.4828724861,0.2694965005,0.0999229029,-0.0498244241,-0.1282318234,-0.0530842021,-0.1693366617,-0.0467171595,0.0469614938,0.5258587599,0.092763558,0.1517028511,-0.1340388954,-0.1395256072,-0.1057453156,0.3818728328,-0.143522501,-0.0104673337,-0.2730925381,0.0261897985,-0.1786254048,0.1172598824,-0.111997433,0.2092107683,0.126393795,0.3173687458,-0.0004321454,-0.4568518698,0.5289966464,-0.4251689911,-0.230963707,-0.35997504,0.222649619,0.3706854582,-0.0084141027,-0.5656880736,0.0114936586,0.3054683805,0.0736179724,-0.5129502416,0.0012242422,-0.1819702834,0.2164076269,-0.0813525096,0.2238122076,0.1676720679,-0.2962998748,-0.2643440068,-0.1033178791]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2727","title":"Error in loading the Arabic Billion Words Corpus","comments":"Thanks @M-Salti for reporting this issue and for your investigation.\r\n\r\nIndeed, those `IndexError` should be catched and the corresponding record should be ignored.\r\n\r\nI'm opening a Pull Request to fix it.","body":"## Describe the bug\r\nI get `IndexError: list index out of range` when trying to load the `Techreen` and `Almustaqbal` configs of the dataset.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nload_dataset(\"arabic_billion_words\", \"Techreen\")\r\nload_dataset(\"arabic_billion_words\", \"Almustaqbal\")\r\n```\r\n\r\n## Expected results\r\nThe datasets load succefully.\r\n\r\n## Actual results\r\n```python\r\n_extract_tags(self, sample, tag)\r\n 139 if len(out) > 0:\r\n 140 break\r\n--> 141 return out[0]\r\n 142 \r\n 143 def _clean_text(self, text):\r\n\r\nIndexError: list index out of range\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.2\r\n- Platform: Ubuntu 18.04.5 LTS\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":31,"text":"Error in loading the Arabic Billion Words Corpus\n## Describe the bug\r\nI get `IndexError: list index out of range` when trying to load the `Techreen` and `Almustaqbal` configs of the dataset.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nload_dataset(\"arabic_billion_words\", \"Techreen\")\r\nload_dataset(\"arabic_billion_words\", \"Almustaqbal\")\r\n```\r\n\r\n## Expected results\r\nThe datasets load succefully.\r\n\r\n## Actual results\r\n```python\r\n_extract_tags(self, sample, tag)\r\n 139 if len(out) > 0:\r\n 140 break\r\n--> 141 return out[0]\r\n 142 \r\n 143 def _clean_text(self, text):\r\n\r\nIndexError: list index out of range\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.2\r\n- Platform: Ubuntu 18.04.5 LTS\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nThanks @M-Salti for reporting this issue and for your investigation.\r\n\r\nIndeed, those `IndexError` should be catched and the corresponding record should be ignored.\r\n\r\nI'm opening a Pull Request to fix it.","embeddings":[-0.1877755523,-0.046449706,-0.0886767581,0.4570911825,-0.1361012012,0.2721578479,0.2241880596,0.3228879273,0.4278388917,-0.0132559789,-0.2501672208,-0.0138163101,0.1214889213,0.0700821951,-0.0400824957,-0.2980046272,0.0483283624,-0.098388046,0.2569685876,0.0944317132,-0.2235391736,0.0493709408,-0.3123707175,-0.0771447942,0.1526142508,-0.0772051439,0.0157806892,-0.0589512959,0.209041208,-0.4896107316,0.086100854,-0.3483731449,0.5297790766,0.5618437529,-0.0001174112,-0.0130771194,0.4175146818,0.0011641483,-0.2492554486,-0.5517652631,-0.191778481,-0.1388527453,0.1426743865,-0.1907674372,-0.1093662828,-0.268625766,-0.1455368847,-0.2430966496,0.1646284759,0.3619323671,0.1498288214,0.1506914049,0.1792136282,0.0761944503,0.2918259203,-0.2640067935,-0.0309772193,0.1672476977,0.5540938973,0.0845655128,0.0206550229,0.1773422956,-0.1420636475,0.019520076,0.1509270966,-0.0254509784,0.1860082895,-0.6007181406,0.0481732525,0.2375650257,0.4379049242,-0.2943656743,-0.4859935641,-0.203239426,-0.0306405481,-0.3684281707,0.2340958267,0.3169795871,-0.1875098348,0.1218396127,0.1365371197,-0.163762331,-0.1190502718,0.3956000805,-0.2622692585,0.3255661428,-0.0335666724,0.0091807358,0.1443209648,-0.1356599182,0.0591473952,-0.0627769157,0.0089818519,0.2237094939,-0.3714162111,0.24262923,0.0462795645,-0.1512625068,0.0464232042,0.1270489544,0.1780865192,-0.0431698486,-0.0866993442,0.1870376319,0.3773936927,0.1415175349,-0.0474852845,0.1577919126,0.0590441376,0.2584305108,-0.036474321,0.0688213184,0.1227201745,-0.1632344127,0.0382446647,-0.1630247384,0.2272497565,-0.3855598569,-0.2166520804,0.2503850758,-0.400190413,-0.216493547,0.0175691918,0.1911917776,-0.2175972015,-0.1130019203,0.1756854653,0.2376640737,-0.1939764917,-0.0253841691,-0.1380524337,0.0888229981,-0.0909494609,0.0913553238,0.4032443464,-0.60462147,0.2357470393,-0.0026580137,0.3509509861,-0.0137081342,0.114214994,-0.3825574219,0.1943871826,0.4071758091,0.1387536079,0.079913117,0.1162825003,0.0233889893,-0.1573423296,0.4012973011,-0.391155839,-0.1710174978,-0.0674691871,0.102638647,0.000966474,0.0878515691,-0.2276010215,0.3403206468,0.5037471652,-0.0629233569,0.00321158,-0.1481618732,-0.3219294548,-0.0733049363,0.1235521734,0.3431198597,-0.3512627482,-0.0584017523,0.0450982563,0.4893796444,-0.0376650393,0.2192805856,-0.2559595108,0.4304127395,-0.1774687618,0.3240619898,0.2467717826,-0.1277134866,-0.2421393543,0.1979265958,0.2657414377,-0.0758553892,-0.1030729339,-0.2082653493,0.4517375231,-0.0546392873,0.6274052262,0.4939140677,-0.0511970967,-0.0900520161,-0.3175869882,-0.1195221618,0.3496678174,0.125807032,0.1293507218,-0.0166825335,0.0039982693,-0.2633425295,0.1419402808,0.1173850745,-0.0173712447,0.3858934045,-0.4759050906,0.0965217575,0.3008753657,-0.2243798375,0.0428482927,0.1723197103,-0.1413490325,0.2480078042,-0.2273525894,0.2863011956,-0.4168075323,0.0402079336,-0.2054211795,0.0908534005,0.0146368677,-0.0775188431,-0.0872636363,0.2759567499,-0.2865067124,-0.1691031456,0.0641296208,0.0150775602,-0.6049718261,0.3342185915,-0.176308617,-0.2357383072,0.0998833701,-0.0123946629,0.1410332173,-0.1680068374,-0.213278085,0.0060990844,-0.1226605475,-0.2166849971,-0.3014376163,0.2852753103,0.295688808,-0.373118192,0.3329123259,0.4022060037,0.1639844328,-0.0809130594,0.3661567569,0.2040123641,-0.1086356044,0.3784812987,-0.1051313728,-0.2154235393,0.2266115993,0.0307549872,0.0170902889,-0.1032828689,0.2849919796,-0.1050957665,0.3402002156,-0.0231685769,-0.131332323,-0.1196971312,0.0099507049,0.0059377626,0.2142357081,0.2765231133,-0.0754362345,-0.1280435026,-0.1962985843,-0.4573308527,0.2315344214,0.1442405879,-0.0854848549,-0.089523755,0.124490425,-0.0630655214,0.3120190203,0.1077940091,0.294393599,0.2502011359,0.1347865462,0.1014078036,-0.3171717227,-0.3249179423,0.0405472703,0.1729808897,-0.3517558277,0.4300263524,0.2592787147,-0.3555389047,0.0915019512,-0.6388462186,-0.2775249779,-0.1306153536,-0.0360215791,0.0063877609,-0.1150352284,0.1680921614,-0.1682744324,-0.0864626616,0.2360263318,-0.1264009923,0.0223936588,0.0920129195,-0.3701221347,-0.0770025775,0.2550441027,-0.074680008,0.1576037258,0.0727805719,-0.1214017197,-0.1043154299,-0.0292145535,0.1853601933,-0.0183185749,0.4640192986,-0.1149681211,0.3587628901,-0.1403543502,-0.2421628535,0.2271582037,0.18570669,-0.0218745098,0.1637769938,0.0100992899,0.2391341776,0.0944108143,-0.4680998027,-0.314943105,-0.3170129359,-0.2879398763,-0.0266062431,0.1678608358,0.3233623207,-0.0039898078,-0.2117659599,0.2140717953,0.2239971906,-0.174762398,-0.1660606414,0.2782445848,-0.0394990407,-0.3239015937,-0.0820351094,0.0130388215,-0.0526661798,0.0465412848,-0.4170078933,0.3313923776,-0.2046108246,-0.0389936902,-0.2311557084,0.1742732525,0.0914127827,-0.1448886395,-0.0160810445,0.1682032496,-0.0429368019,-0.0291244797,0.0024891554,0.1942388415,0.3245144188,0.2126039714,-0.1889989525,0.2093219459,-0.0382447913,0.0387954004,0.6535492539,0.2737876773,0.07669615,-0.2376984358,-0.4639846981,-0.1501168311,0.1125407815,-0.4123540521,0.3311195672,-0.049803108,-0.4764990509,-0.2501311898,-0.3851373792,-0.3440397084,-0.1309672743,0.1390445977,-0.1468767077,0.2006634176,-0.0047838348,0.0029348929,-0.0379979722,-0.180661276,0.0935202166,-0.0534730256,-0.0655059814,0.026960101,0.3056277037,-0.2745108306,-0.3386605382,0.1035830304,0.2288863957,0.4502712488,-0.0063825543,0.0100238584,0.0519626178,0.1529523134,0.6871073842,0.0248991027,0.207333684,0.3312928379,0.1308223903,-0.1347749084,-0.341904521,-0.1734804511,0.3713223636,0.0345738046,0.1124416292,-0.2644976079,0.0738708824,-0.0232192073,0.508646667,-0.1613170058,-0.2682159543,-0.3335083723,-0.2844643593,-0.2667708695,0.0898740143,0.1821268201,0.3325804472,0.0521926098,0.068613559,0.1627583504,-0.0147584705,0.1783933043,0.044834014,0.2641827762,-0.0901436284,0.2104614377,-0.1074535251,0.0633915141,0.1904766858,0.4232784808,-0.0521410145,-0.2642839551,0.0136034982,-0.0258189868,0.1289604008,0.1327117532,0.0336932838,0.0372574441,-0.0419184528,0.0932733789,0.055029694,-0.3299226463,-0.0015833856,0.0350292698,-0.5425880551,-0.456977874,0.5198664069,-0.0449460112,-0.0834904239,0.4464785755,-0.1348859221,-0.3541570008,0.8186009526,-0.1164209768,0.9402406812,-0.1575017273,0.0902892798,-0.0986625552,0.4318521619,0.2294859439,0.0170629565,0.1314001828,-0.3510845006,-0.3513865471,-0.1283218265,-0.06531059,-0.0176750161,-0.0540253036,-0.1974101663,0.2823766172,-0.2762391269,-0.1637961268,0.1032770574,0.0036852406,-0.0383964069,-0.3015965223,-0.7697243094,-0.0044940934,-0.121173948,0.0640016198,-0.1528094262,-0.0446805507,-0.0569314882,0.0623655617,-0.3020882607,0.1786654294,-0.1697915196,0.2599480748,-0.0245680343,-0.0733057261,0.2296825349,0.0374518894,0.1489694417,0.0733140334,-0.168074429,-0.1542636007,0.0181144364,-0.2912916839,0.0807126909,-0.1277768761,0.4687170982,-0.0247424152,-0.4387729168,0.2530715466,0.040891666,-0.0934379101,-0.1768903285,0.3172102571,0.1629773527,-0.3863539398,-0.3148027956,-0.2251956165,0.1760262847,-0.086022675,0.1034947857,0.2724947929,-0.2727386057,0.1300434172,0.2212480009,-0.3995957673,-0.0710094422,0.4882135391,0.2661291361,0.1050897613,0.4344873726,0.690943718,-0.1168324053,-0.0621257126,-0.0925751701,-0.0260300841,-0.2954339385,0.1991962194,0.0361020714,0.1139544323,-0.016301645,0.5044162869,0.2366807759,-0.0716008767,-0.1014107317,-0.6054025888,0.1177413166,0.2298295796,-0.0089505278,-0.1208771318,0.1899375021,-0.2474105805,-0.0090366649,-0.2172664553,-0.253660202,0.1492655426,-0.3593368232,0.3357177973,-0.4784272015,0.0297559202,0.1237493828,-0.1205692217,0.0153232422,-0.0598338991,-0.0652334616,-0.2206036896,-0.0253338199,0.1380934566,-0.0656346306,-0.1423896849,-0.2752308249,-0.32111606,0.0448403172,-0.1300245076,0.0392535292,0.2385995388,0.0715834647,-0.0087841963,0.2596637607,-0.1693479568,-0.1735746861,0.321518451,-0.1807611138,0.2754142582,0.1087808385,0.1767297983,-0.0140157295,-0.0061274553,-0.41254884,0.0968423486,-0.1366616786,0.095659554,0.2184815705,-0.4270799756,0.0428607091,0.3978970945,0.302223742,0.2615050972,-0.3576152325,-0.1294261217,-0.1006743386,0.1508114636,-0.2962694764,-0.2760988176,0.0981950462,-0.1901113242,-0.100460954,0.3018140793,0.3105677068,-0.2831620574,-0.3916255534,0.0541898012,0.3112674654,0.0897267312,-0.0206472501,0.3724783659,0.2655084431,0.1119907498,0.1267893612,0.0331316032,0.2198856622,0.3703356087,0.0196165089,0.383146137,0.5232915878,0.1656560004,-0.2105608582,-0.5987141132,0.4757802784,0.3198189437,0.2372417003,-0.027561754,0.0966655388,-0.1578449309,0.1674953848,-0.1574896723,-0.0907713994,0.0042773653,-0.0026952235,0.0463196337,0.1055219769,-0.2007157654,0.1826443821,-0.2155261189,-0.0251817163,-0.0862760246,-0.0452749729,0.1951213777,-0.2231222093,-0.1429708451,0.0516928993,0.1707570106,0.1168573201,-0.2242253125,0.2616122365,0.2061665058,-0.2228832394,0.0259971302,0.4006112218,0.5047866702,0.4441466928,-0.1801774204,0.1704861075,-0.0707329512,-0.1570310444,-0.1112940013,0.1080039516,0.0683770254,0.2420630008,0.2476900369,0.0788133442,-0.0633482337,0.2042130381,0.2919194996,0.1018157899,-0.0199516937,-0.0103593832,-0.0866064131,-0.1367335916,-0.3233426213,-0.0731445402,-0.4708897769,0.0395666175,0.6526071429,-0.3489195108,0.0747141466,0.1700440645,0.0357097015,-0.0809754655,0.4416025579,0.699509263,0.2822183073,-0.4421756268,0.200906992,-0.3303514123,0.2510177195,-0.1868999153,0.2030410469,0.5098544955,0.2263411134,0.0117249554,0.2636482418,0.1323353648,0.3015119433,-0.0514408387,0.2220498472,-0.4766448438,-0.3353193402,-0.1140908599,0.3805034161,-0.3412327468,-0.1062499881,0.1164386198,-0.5912806988,0.0090655545,-0.3235999942,-0.2978065908,0.1466219872,0.3390180767,0.3214837909,0.1121080369,0.0966386348,-0.0456193686,0.0239851326,-0.0831463635,-0.1393451542,0.0083726477,0.4147399664,0.0642857105,0.1914910823,-0.1946295947,-0.2010737509,-0.1859124899,0.373742193,-0.1076294482,-0.0778105482,-0.2968203127,-0.0362857953,-0.140078932,0.0614998192,-0.1188210696,0.1798106879,0.1099344268,0.3384969831,0.0274356771,-0.3561402559,0.5066059232,-0.5102148056,-0.337693125,-0.3747169375,0.135424614,0.2817041874,-0.0234075692,-0.5694018006,0.0318839177,0.3264682889,0.113718845,-0.4719185531,0.0395713225,-0.1244498342,0.2546509206,-0.0420337059,0.1964077502,0.0950779691,-0.2324771732,-0.2083833218,-0.0419428796]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2724","title":"404 Error when loading remote data files from private repo","comments":"I guess the issue is when computing the ETags of the remote files. Indeed `use_auth_token` must be passed to `request_etags` here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/35b5e4bc0cb2ed896e40f3eb2a4aa3de1cb1a6c5\/src\/datasets\/builder.py#L160-L160","body":"## Describe the bug\r\nWhen loading remote data files from a private repo, a 404 error is raised.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nurl = hf_hub_url(\"lewtun\/asr-preds-test\", \"preds.jsonl\", repo_type=\"dataset\")\r\ndset = load_dataset(\"json\", data_files=url, use_auth_token=True)\r\n# HTTPError: 404 Client Error: Not Found for url: https:\/\/huggingface.co\/datasets\/lewtun\/asr-preds-test\/resolve\/main\/preds.jsonl\r\n```\r\n\r\n## Expected results\r\nLoad dataset.\r\n\r\n## Actual results\r\n404 Error.\r\n\r\n","comment_length":22,"text":"404 Error when loading remote data files from private repo\n## Describe the bug\r\nWhen loading remote data files from a private repo, a 404 error is raised.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nurl = hf_hub_url(\"lewtun\/asr-preds-test\", \"preds.jsonl\", repo_type=\"dataset\")\r\ndset = load_dataset(\"json\", data_files=url, use_auth_token=True)\r\n# HTTPError: 404 Client Error: Not Found for url: https:\/\/huggingface.co\/datasets\/lewtun\/asr-preds-test\/resolve\/main\/preds.jsonl\r\n```\r\n\r\n## Expected results\r\nLoad dataset.\r\n\r\n## Actual results\r\n404 Error.\r\n\r\n\nI guess the issue is when computing the ETags of the remote files. Indeed `use_auth_token` must be passed to `request_etags` here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/35b5e4bc0cb2ed896e40f3eb2a4aa3de1cb1a6c5\/src\/datasets\/builder.py#L160-L160","embeddings":[0.0871632025,0.025846079,0.0274070129,0.6045913696,-0.1784682125,-0.0898554325,0.2882218957,0.3268503845,0.0184905343,0.1675235331,-0.3693999052,-0.0839862376,0.2241544724,-0.04085657,0.0888220221,-0.0422795936,0.0376422778,0.1200491264,0.3278564811,-0.2406732142,-0.1716177166,0.1605942547,-0.0282529909,0.2701848447,0.069742389,0.1259597391,0.0421590768,0.5915513635,-0.1242948249,-0.1665705591,0.2873155773,-0.0173066445,0.5707443357,0.8942393064,-0.0001184026,0.168162629,0.2595363259,0.010358192,-0.2348511517,-0.1142685115,-0.4038441181,0.1302046031,0.0836360753,-0.1607075483,-0.1807733178,-0.0938127637,-0.3060295582,0.0864418074,0.4408935308,0.3926028609,0.1661800742,0.4484367371,0.0907083973,-0.1656462997,0.4235145748,0.0348576717,0.0053401585,0.2900849581,-0.0708383024,0.0867628902,0.0132969748,-0.0068785753,0.0141346008,0.0842933804,0.2092770934,0.0134577155,-0.0423866138,-0.0992748365,-0.1002172008,0.0658624619,0.4129537046,0.0002570056,-0.2669299543,0.026682891,-0.0367544442,-0.1364852935,0.158232078,0.2569408715,-0.1341071278,0.1208953932,-0.1735553145,-0.3134871423,-0.4048245251,0.3338383734,0.1114476398,-0.0551392734,0.0097003402,0.0252807047,-0.2526357472,-0.2069942504,-0.0964903906,-0.1750052273,0.021321848,0.1697475761,-0.0287924297,-0.0907949507,0.085866265,-0.031663157,0.2942894399,0.3939569592,-0.2702388763,-0.0663872436,-0.4752777517,0.1954914033,-0.017346004,0.3439953625,0.0193969086,0.2792779803,0.4515994191,0.4925678074,0.0249252655,0.0916304067,-0.3378535211,-0.0051901876,-0.1679282337,-0.1927329153,0.4646197557,-0.4979287684,-0.0039766082,0.1847059727,-0.2357386053,0.0277586654,-0.0260837879,0.3406423032,-0.0719407499,-0.1338147968,0.0686857253,0.0375015922,-0.2768373191,-0.0923992842,-0.1704371572,-0.1632038951,0.0213641208,0.259172529,0.1316786557,-0.0267508589,0.2024156749,-0.0369497463,0.3569190204,-0.2349615395,0.0001223852,0.1925103962,-0.1851204485,0.2597736418,0.1659449488,-0.0253740828,0.1190524772,-0.392562747,-0.0547073521,-0.1396570802,-0.6864687204,-0.4307950139,-0.0843315795,0.1072939336,-0.1942071468,0.0400028341,-0.5282006264,-0.3732130527,-0.1112889051,-0.0995284989,0.0376247726,0.0282477625,-0.0483968221,-0.1158322841,0.1585278064,0.585699439,0.3585839868,-0.0071596056,0.3294963539,-0.1886004359,0.2019345164,0.3431803584,-0.1851004809,0.0839203298,-0.2878257632,0.31100142,0.1597580165,-0.4815152586,-0.4112557471,0.645683229,-0.26988554,0.0339968279,0.098229751,-0.1084417254,0.1172511205,-0.2085673362,0.5326524377,0.4657271802,0.2107895911,-0.1300878823,-0.288569659,-0.2136393338,-0.3629421592,0.3998591304,-0.0017962889,0.2109307945,0.3625538647,-0.2569054961,0.2932198346,-0.0458183661,-0.1872207522,0.3072160184,0.3584540188,0.0192747992,0.2242029458,0.0763057843,-0.2873437405,0.1927688122,-0.3585609794,0.1491599977,-0.0419801287,-0.0627901629,-0.3100636303,-0.0910803974,-0.4545488954,0.186371997,0.0425487794,-0.0372703671,-0.1701895148,0.1651738584,0.0201012958,0.2720374167,-0.1222627088,0.2718659937,-0.2868469059,0.6367555857,0.0267659351,0.0750541836,0.1860989779,0.2177869529,0.3806755841,-0.0630792081,-0.1942438781,0.3124832213,0.1950450838,0.1324029118,0.7725487947,-0.0686830506,0.2205391824,-0.0882269591,0.0287845023,0.4030066431,0.0595477112,0.2151002139,0.1824583113,0.1951798499,0.0441593677,0.3703363836,0.0991877913,-0.108142063,0.1792208999,-0.1274946332,-0.1071734875,-0.0383460037,0.3741497993,0.0318130068,-0.2007786632,-0.0603227839,-0.1172189116,0.0278729647,0.2025966495,-0.3027195632,-0.0415041894,0.3303074241,-0.1391055584,-0.0167186912,0.2410385013,0.4205012023,0.2979258597,-0.0170525722,-0.1817577481,0.0656797588,-0.0896785483,-0.0831211284,-0.0638893247,0.0220544692,0.0340768434,0.1687859297,0.1015623882,-0.0376967192,-0.0569456555,-0.4256904721,-0.082486622,0.2719039619,-0.4516474009,-0.0140944477,-0.3728094697,-0.13232328,-0.076727353,-0.2615007162,-0.5702645779,-0.0763109252,0.028902594,-0.0809274614,-0.10595496,-0.1590922326,0.0347101949,0.2810231447,0.1168027893,-0.3547641933,-0.1858559549,-0.0121848453,-0.294015795,-0.0945232734,0.2922323942,0.020486027,0.3175248802,-0.4767365754,-0.0902839676,-0.4070107639,-0.1112848222,0.0356454179,0.1565070599,0.3482768238,-0.0398228914,0.072316125,0.2941636741,0.091648683,0.2570353448,-0.1088714674,-0.2068011612,0.2457992285,-0.1831875741,-0.0909194797,-0.1320876181,-0.0294611044,-0.2550732195,-0.1853720546,0.2371392697,0.3771695495,0.0953637883,-0.2680513561,0.07060536,0.2262733728,-0.1531769484,-0.1468227059,-0.171980992,-0.5023723245,0.2743113935,-0.4032094777,-0.2306600511,0.2652857006,0.2256273478,0.0434053689,-0.1237591431,-0.6602146029,-0.5772987008,-0.0836366713,0.2650004029,-0.2728845179,0.0472681038,0.0438520834,-0.1540024579,0.0778176486,0.04580402,-0.2724538147,0.1395338178,0.047789719,0.1411540806,0.0129360519,0.1984617561,-0.0642819479,0.5992151499,0.1762917489,0.1972340941,0.2701568007,0.0350492485,0.2159227431,-0.0271467417,0.041646257,-0.0525212735,-0.0180699863,-0.16071558,0.2109569311,0.4117851853,-0.0270068999,-0.294361949,-0.7045912743,-0.1905461848,-0.2729957104,-0.4170337021,0.346999824,0.080554761,0.1777725071,-0.0354026482,-0.1152432188,-0.1952486336,0.2496605515,0.4584136307,0.2080112249,0.2246272862,0.0544974878,0.0126659535,-0.4743277729,0.3333809078,0.0314420536,0.2829281688,0.0254774876,0.2000863999,0.1383213103,-0.1713120192,0.0101335393,0.1000095457,0.4979896247,-0.018564282,-0.1639068127,-0.085850805,-0.0535558946,0.0078166742,0.0035866494,-0.1175594479,0.8319283128,-0.2911431789,-0.2107410282,0.002134152,-0.1326838285,-0.1245746836,-0.4110311568,-0.3006387353,-0.3991363943,-0.3001538515,-0.036026001,-0.1755881906,0.1223568469,-0.1613682359,0.1208265871,0.3628306091,0.2604140043,0.0692803264,-0.0567805842,0.3072290123,-0.199885875,-0.0738416538,0.3068895936,0.2507615983,0.4206142128,0.640981555,0.0130596953,-0.4118212461,0.0338695832,-0.1721885353,0.0882983357,0.4270452261,-0.0903204829,0.0549750812,0.2732680738,0.4028321207,-0.107472688,0.1076908931,0.0753507912,-0.2576527297,-0.1100757793,-0.1417010576,0.4426443875,-0.294550091,0.0482027382,0.0306481868,0.5616435409,-0.161317572,0.4553392231,-0.2532244623,0.865635097,0.0318805613,0.3191570938,0.4125316739,-0.2175780386,0.1534341723,-0.6794502139,-0.0327020995,-0.1211867929,0.2353567183,-0.1737674475,-0.0629226342,-0.0326856486,-0.2224305272,0.069711104,0.3440417051,-0.1073444486,0.5143914819,0.2266750187,0.0482462533,0.1429788321,-0.070987463,-0.3608293831,0.0868044719,-0.0884208307,0.0974174961,-0.0813911408,-0.3059489429,-0.3581776619,-0.1596123278,0.0168610848,0.1965004504,0.1090874597,-0.0575448535,0.0566226058,-0.2959418297,0.0211627427,-0.0125352424,0.134713307,0.0030472972,-0.520749867,0.1371695697,-0.1726103127,-0.0293730758,-0.0564425923,-0.0933094174,0.1072477698,-0.2675108016,-0.1880783439,-0.0298242476,-0.1000051796,-0.4531242549,-0.2585156262,0.1860947609,-0.0081695709,-0.1558696777,0.0714388266,-0.0161195863,-0.03413729,-0.1082227975,0.0867442712,0.1800412834,-0.3045523167,-0.3173415661,0.1646490693,-0.2764060795,-0.0275206529,0.4022209048,-0.0957700163,-0.1138910726,0.3496786654,0.1088127717,-0.0804256424,-0.1351315677,-0.0554854199,0.3459487557,-0.7139821649,-0.0095988037,-0.0371179767,0.2363055348,-0.0731430352,-0.0166622754,0.1022861823,-0.4073726833,0.2560967803,-0.5434626341,-0.0538387671,0.0893286914,-0.3175041974,-0.1462234557,0.1050830707,0.1062572896,-0.1318742037,-0.3294658363,-0.242168054,0.3742593527,-0.3990373611,-0.0887278989,-0.0879354253,0.0046875654,0.2556555569,0.1062884182,0.0472144894,0.0730124041,-0.2092871517,-0.0989163294,-0.3098873198,0.2121884376,0.0714150816,-0.1748319715,0.0320985466,-0.1938273758,-0.213830173,0.0898280442,0.0093899993,0.1445708722,0.0404258929,-0.1555693001,0.2398986518,0.032525003,-0.3132787049,0.4965496063,0.1967858672,0.5202296972,-0.1744225472,0.0494262837,-0.6694009304,0.0276605617,0.123119764,0.0687427223,0.2977377772,0.3116061389,0.6110662222,-0.5101627707,0.1276436448,-0.0901973099,0.1447168291,0.1153296307,-0.2795445025,0.0343907997,0.1356980354,0.0908344686,-0.0185945798,-0.2194552422,0.5809531212,-0.0193753131,-0.2160779089,0.1184569672,0.2108077705,0.0717757046,0.0618609786,0.1158037558,0.3747921586,0.1398815066,-0.2101061046,0.0496849753,-0.1048738658,0.2838538885,-0.0839252919,0.0560892858,-0.0317067541,0.3715334535,-0.1107558608,0.1075450107,0.0834409222,0.2526920736,0.0696458668,-0.5230891109,-0.2532204092,-0.0236061476,0.3299840689,0.0113293175,-0.2905017436,0.2662549019,-0.1612525433,0.2738799453,-0.3528705835,-0.043729838,0.0542654246,0.1502536833,-0.0497360863,-0.2233939469,-0.0962868929,-0.0370934978,-0.107738331,-0.1320926547,-0.3277083635,0.1041402668,-0.1485617459,-0.2872103751,-0.2418294698,0.2078895122,0.0006502005,-0.1109865829,0.2159478515,0.1630753428,-0.0378808044,0.1516307145,-0.2827296257,0.3619365692,0.1572063565,-0.0038279509,0.0017600444,0.1595134139,0.0488212854,0.2324170619,0.2722708285,0.1974627674,0.0793102682,0.2943634093,0.0726412535,-0.1164298505,-0.0446594357,-0.2403579205,0.2847384214,-0.1660988629,0.5623672009,-0.5071780682,-0.0439348929,-0.34369874,-0.1898684204,-0.3341602683,0.1306713074,0.2980681956,0.2615304589,-0.0225224998,-0.1373253763,0.0179900397,-0.3996053934,0.4213399291,0.490042001,0.4001396596,-0.34164837,-0.1488673687,-0.4380819798,0.4692400694,-0.0023397366,-0.1144668832,0.1430308223,0.2080901712,-0.1222667322,-0.0149511807,0.2360367328,-0.2913753092,0.2221712619,-0.0908003449,-0.0849624649,-0.0882914886,-0.0792946965,0.0259992294,-0.0253751613,-0.2910265625,0.1560931504,-0.3079158366,-0.021579111,0.1668858081,-0.06080826,0.1342535913,-0.2701957822,0.4539769888,0.1228079498,0.1914724261,0.169109419,-0.337364763,-0.1320326924,-0.1517775357,-0.2058751732,-0.422677815,-0.3909520805,0.1235793829,-0.122985661,0.0895022303,-0.3886905909,0.2831037343,-0.036288593,0.2037679106,-0.1012543961,0.1550435573,-0.2333901972,-0.1216567233,0.1989433318,0.0616157912,0.0581451282,0.0967615098,-0.0836428031,0.0407303758,0.675365448,-0.4463759661,-0.0966645107,0.0911665633,0.3124555349,0.2603935897,-0.3580765128,-0.1640055925,0.1470507085,0.2772576809,0.2774128914,-0.240974918,0.1842027903,-0.266630888,0.1258785427,0.0475342013,0.0952605978,0.0432057567,-0.1891228408,0.3335573077,-0.0526452363]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2724","title":"404 Error when loading remote data files from private repo","comments":"Yes, I remember having properly implemented that: \r\n- https:\/\/github.com\/huggingface\/datasets\/commit\/7a9c62f7cef9ecc293f629f859d4375a6bd26dc8#diff-f933ce41f71c6c0d1ce658e27de62cbe0b45d777e9e68056dd012ac3eb9324f7R160\r\n- https:\/\/github.com\/huggingface\/datasets\/pull\/2628\/commits\/6350a03b4b830339a745f7b1da46ece784ca734c\r\n\r\nBut a subsequent refactoring accidentally removed it...","body":"## Describe the bug\r\nWhen loading remote data files from a private repo, a 404 error is raised.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nurl = hf_hub_url(\"lewtun\/asr-preds-test\", \"preds.jsonl\", repo_type=\"dataset\")\r\ndset = load_dataset(\"json\", data_files=url, use_auth_token=True)\r\n# HTTPError: 404 Client Error: Not Found for url: https:\/\/huggingface.co\/datasets\/lewtun\/asr-preds-test\/resolve\/main\/preds.jsonl\r\n```\r\n\r\n## Expected results\r\nLoad dataset.\r\n\r\n## Actual results\r\n404 Error.\r\n\r\n","comment_length":18,"text":"404 Error when loading remote data files from private repo\n## Describe the bug\r\nWhen loading remote data files from a private repo, a 404 error is raised.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nurl = hf_hub_url(\"lewtun\/asr-preds-test\", \"preds.jsonl\", repo_type=\"dataset\")\r\ndset = load_dataset(\"json\", data_files=url, use_auth_token=True)\r\n# HTTPError: 404 Client Error: Not Found for url: https:\/\/huggingface.co\/datasets\/lewtun\/asr-preds-test\/resolve\/main\/preds.jsonl\r\n```\r\n\r\n## Expected results\r\nLoad dataset.\r\n\r\n## Actual results\r\n404 Error.\r\n\r\n\nYes, I remember having properly implemented that: \r\n- https:\/\/github.com\/huggingface\/datasets\/commit\/7a9c62f7cef9ecc293f629f859d4375a6bd26dc8#diff-f933ce41f71c6c0d1ce658e27de62cbe0b45d777e9e68056dd012ac3eb9324f7R160\r\n- https:\/\/github.com\/huggingface\/datasets\/pull\/2628\/commits\/6350a03b4b830339a745f7b1da46ece784ca734c\r\n\r\nBut a subsequent refactoring accidentally removed it...","embeddings":[0.1519906223,0.0217427462,0.042152416,0.4689883292,-0.1487777978,-0.0837455839,0.1870218068,0.3314785063,0.0251010451,0.1321308613,-0.4147609174,-0.0381756052,0.117717348,-0.084571287,0.2274655551,-0.0939541757,0.0643756837,0.2031136006,0.1788401008,-0.1964158416,-0.1416362226,0.1891839206,-0.0267064814,0.2135807723,0.0884627029,0.1626259089,-0.0169534944,0.5753366947,-0.1334585994,-0.2530795634,0.4264841974,0.0575155728,0.4901399016,0.8685284257,-0.0001199092,0.2189432532,0.2850868106,0.0630401894,-0.2731933594,-0.1770015359,-0.430652827,0.1696650833,0.1902304739,-0.0577684343,-0.1480035186,-0.0856637135,-0.2231205404,0.103069745,0.3863911331,0.2849093974,0.1712985784,0.3798895776,0.0566889867,-0.1431237161,0.2864226103,0.0662521943,-0.0473320037,0.4522379339,-0.0943674222,0.1032377481,-0.064959161,-0.0003474524,-0.0010124088,0.0269993339,0.2268557847,-0.0060468032,-0.1101498604,-0.0566704497,-0.1633121222,0.1307189167,0.3496461809,-0.1119890437,-0.2637358904,-0.0768798217,-0.0070292028,-0.23805134,0.2104153484,0.1895542443,-0.1170151904,0.2497235984,-0.2861321867,-0.418076843,-0.2974921167,0.3176789284,0.064110443,0.0484509431,-0.0318715237,-0.0522889346,-0.3012050688,-0.0576599389,-0.1049260423,-0.125702545,-0.0357950516,0.1374064833,0.0101503776,-0.0649466217,0.262286216,-0.0738000795,0.3891857862,0.2772269845,-0.0017992326,-0.0349763781,-0.4886959195,0.1327638626,0.0280533563,0.3387026489,-0.0982748345,0.222291261,0.5048391819,0.547963798,0.0648633912,0.1237414554,-0.186521098,0.1162764356,-0.198296994,-0.2018738836,0.5696320534,-0.4419042468,-0.0255237557,0.3089796007,-0.3250632882,0.025451865,-0.0543165281,0.2940502763,-0.0640186071,-0.005388211,0.1105384529,0.0006798317,-0.2913879454,0.0378874876,-0.1808711588,-0.126337871,-0.0610318184,0.171999529,0.149998799,-0.0761433095,0.1993522346,-0.0468536206,0.3632746935,-0.2024489939,-0.1399954557,0.2448552549,-0.1660929024,0.2702893317,0.1124734357,-0.0482192822,0.0312759876,-0.3952690661,-0.0608169958,-0.1533695161,-0.6331657767,-0.4467029572,-0.1026829109,0.0789961293,-0.2454281896,0.0484480076,-0.5156443715,-0.2482574135,-0.2269541025,-0.1161455214,0.1025046259,0.0731713697,-0.2155027092,-0.1725380421,0.2357490361,0.6003877521,0.2160528153,-0.0169517063,0.304343909,-0.2247075737,0.1536226869,0.1647028178,-0.1423661709,0.2290190607,-0.3856140971,0.1902227849,0.2130478024,-0.4618348777,-0.4828761518,0.6060372591,-0.3202258945,0.246334523,0.0187543947,-0.2113689333,-0.0374567397,-0.1387769878,0.5398167372,0.4439108372,0.2286476493,-0.1108342558,-0.3019841313,-0.2486560196,-0.4080145359,0.3468935788,0.0490517467,0.2788874209,0.3681977093,-0.1649483293,0.4139636755,-0.0699726567,-0.1263872832,0.3668572903,0.419595927,0.0625421554,0.1435604542,-0.0797368288,-0.3744698763,0.2066181898,-0.3822329044,0.1196452826,0.0238651987,-0.0368936583,-0.3012804687,-0.1524414718,-0.50297153,0.1053224504,0.0548011735,0.0566113442,-0.1741876602,0.1656812727,-0.114625901,0.4241328835,-0.0097633628,0.2932831347,-0.3459373713,0.6859853268,0.0520669818,0.0212307069,0.0752310902,0.3377068937,0.2700014412,-0.1151409075,-0.1538033038,0.4489713907,0.0573419221,0.1254339367,0.6857694387,-0.1921598315,0.3496061265,-0.1284115762,0.0318716057,0.3469433188,-0.0006147947,0.2657709122,0.1878751665,0.1333737522,-0.0380102396,0.2506723106,0.0925961584,-0.1911727041,0.2122319043,-0.1556424797,-0.172091499,-0.1235758364,0.4104898572,-0.0062140287,-0.1908023059,-0.0735201463,-0.1987667382,-0.0376580022,0.1191855446,-0.287922889,0.0572980158,0.3348961473,-0.0859449878,-0.0173967294,0.1962102503,0.463783294,0.3661141396,-0.0279386695,-0.1758410335,0.2077221423,-0.0307257306,-0.0478536636,-0.0640150011,0.0621606186,0.0942825004,0.2740388215,0.1305274516,0.0253582932,-0.0931051672,-0.3885853291,-0.1760837585,0.2046646625,-0.4924301505,-0.0846224949,-0.4646685123,0.0009307872,-0.084841609,-0.2139741778,-0.4650982022,-0.0593363605,-0.0064655407,-0.079875797,-0.1846033484,-0.1233486459,-0.1319295168,0.206031248,-0.0101575637,-0.3751419187,-0.2010852844,0.0916013271,-0.2265715599,-0.1003581136,0.125613153,-0.0187422317,0.3192822933,-0.4008626044,-0.1340949535,-0.3375827372,-0.1568830907,0.1196986064,0.1543842256,0.2986592352,0.1002826989,-0.0388111435,0.2130705118,0.0275541749,0.2490252852,-0.1115017757,-0.2384189367,0.2173651606,-0.1269457787,0.0206025187,-0.1685043424,-0.0235138889,-0.2968161106,-0.3472931981,0.1988022327,0.2847707272,0.0736470222,-0.2436434329,0.0452049673,0.2794009149,-0.2056296766,-0.0306539237,-0.0997859538,-0.4469516873,0.1976716518,-0.3998927176,-0.1846944392,0.2627224028,0.1623168141,0.0497268476,-0.0955392495,-0.7413303256,-0.5427148938,0.0072414684,0.2825992107,-0.2767823637,0.0239467733,0.1121171638,-0.0721712708,0.0861302838,-0.0010128813,-0.2558847666,-0.0224128682,0.0404698029,0.2527865767,0.1034074277,0.2775860727,-0.0809155926,0.5106106997,0.1753750294,0.2900905013,0.3952641487,0.0623344742,0.3486857414,-0.1234561279,0.0279973596,-0.0588812679,-0.1017153412,-0.1672573537,0.1642962247,0.4003346562,0.0011567002,-0.3944430053,-0.702341795,-0.2399015725,-0.2511533201,-0.3958387971,0.2358167022,0.0389198475,0.1524688452,-0.1011965647,0.0648010969,-0.0949854329,0.241015777,0.3989120722,0.2297189236,0.1988522261,0.0865757763,0.0221505743,-0.6100174785,0.1995530576,0.0308054406,0.2061389387,0.0316681415,0.1479301006,0.0330291875,-0.1607430428,0.1166279688,0.1354092062,0.5210430026,-0.0774564072,-0.2037371099,-0.0485219285,-0.0565958023,0.0715441927,0.0715448707,-0.0739260688,0.7497166395,-0.2757319212,-0.1843092889,0.1453226358,-0.2294542193,-0.0722525418,-0.4218307436,-0.3218511939,-0.4020782709,-0.2374112606,-0.0845634863,-0.1632716358,0.1329292506,-0.0823043585,0.0034868375,0.3989427984,0.1448921859,0.0388552435,-0.0921778008,0.3415281475,-0.2010949701,-0.0201292876,0.3490425944,0.3559367359,0.3721437752,0.6303964853,-0.0740082785,-0.2629466355,0.0912154242,-0.0916441828,0.1938480884,0.4522798657,-0.0198248252,0.0325107761,0.2730745077,0.3494525552,-0.2121815979,0.0197356492,0.0629932359,-0.3600474,-0.1444817185,-0.0917157009,0.5425668955,-0.242387712,0.0203025695,0.0324037932,0.6648786068,-0.2799300551,0.3975714743,-0.2922760248,0.8610058427,0.1289741099,0.315233618,0.4537091851,-0.3615630567,0.2927815914,-0.6476732492,0.0116480021,-0.0709133819,0.2144019604,-0.1924668401,-0.0844997689,-0.0397077166,-0.3045714796,0.0499337129,0.3462190032,-0.0580283701,0.4351762235,0.2092389017,-0.0474814288,0.1175565049,-0.0445765071,-0.4515303969,0.0559131056,-0.1300706863,0.1149350554,-0.1449425071,-0.1952134222,-0.3324146569,-0.1463126242,0.0214680936,0.2302918732,0.2007448524,-0.1235348731,0.1145106107,-0.4036968648,0.0324662142,0.0046577733,0.0153040178,0.0796422064,-0.4880464971,0.1559609622,-0.2170826644,0.0957424045,-0.0138128586,-0.0523207113,0.1070038006,-0.1990105808,-0.2001740485,-0.0736115724,-0.176739186,-0.4683381617,-0.2065043598,0.200097248,0.0447459631,-0.1383116394,0.0786036402,0.0633991212,0.1148078293,-0.1882505715,0.0761443898,0.2264969945,-0.3660918772,-0.3395538032,0.1169341654,-0.2562228739,-0.0816713795,0.4282543361,-0.0658488423,-0.1398654431,0.2938952446,0.0974347219,-0.0940848887,-0.102046825,-0.0796987489,0.4889606237,-0.5784721375,0.0619343668,-0.0959531441,0.0981809422,-0.098592326,-0.0332174338,0.1336444318,-0.3805674016,0.12734209,-0.5043212771,-0.1131138057,0.1486774981,-0.3034216166,-0.0561271459,0.1187534854,0.1008640677,-0.0790763646,-0.3572710752,-0.2249250859,0.2928539813,-0.3823474944,-0.1863855869,-0.1235091984,0.0516400784,0.2517188489,0.0905525312,0.044312302,0.1776967496,-0.1890117228,-0.0639829934,-0.3440085649,0.2022429258,0.0772347599,-0.1617627293,0.1021397933,-0.1355783492,-0.1508629769,0.0877779201,0.0712465197,0.154317975,0.094725512,-0.2369141579,0.2930661738,-0.116352208,-0.2011564523,0.4889200628,0.2483322769,0.5059658885,-0.0493224077,0.0667501315,-0.6707577705,0.0313746072,0.1665521562,0.0557229519,0.2788781524,0.3210637867,0.6118314266,-0.5756316781,0.0329907238,0.0060395719,0.3109242022,0.1347319633,-0.2945578694,0.0191926379,0.04291486,0.0752069727,-0.0013580923,-0.2630735934,0.585603416,0.0490428843,-0.1889057457,0.0820923373,0.1229203716,0.050972119,0.0173319336,0.1745790392,0.4702454209,0.1420276165,-0.1721402556,0.1295566708,-0.1266873181,0.282738775,-0.0844974294,0.0468947813,-0.0692189634,0.4968429208,-0.0335238464,0.1390380114,-0.0094808433,0.3376964629,0.1195746064,-0.5196663141,-0.1551904231,0.0363381542,0.1979641914,0.0586456805,-0.2508135736,0.1738142669,-0.1820995957,0.2708267272,-0.3055837452,-0.0988203362,-0.0119951423,0.1081988737,-0.0961680859,-0.2348445058,-0.0236999113,0.0187402572,-0.0866906866,-0.1623461097,-0.2282876223,0.0780910924,-0.1180660725,-0.3563401401,-0.1524612904,0.1279537231,0.0152660618,-0.0656815544,0.2394930869,0.246967569,-0.1023223251,0.0706755519,-0.1640631557,0.2949666083,0.1147912294,0.0527335592,-0.0553489402,0.1584404111,0.0689917803,0.1454928368,0.2391982675,0.2414409667,0.0733418018,0.1912483722,0.085417971,-0.1016418934,0.1510270834,-0.1452060938,0.390814513,-0.240869388,0.5807546973,-0.4651973248,-0.0101849688,-0.2176097184,-0.1564030945,-0.3127634525,0.1948426217,0.3365879953,0.1935874373,-0.0448621511,-0.0810688362,0.0199351404,-0.1868435889,0.3000537455,0.5435923338,0.3429712653,-0.2879975736,-0.1429525018,-0.4444923103,0.523022294,0.02851828,-0.0338279307,0.1645580381,0.2104057372,-0.2261794209,-0.0194597337,0.2304650396,-0.2229701728,0.2771526873,-0.1384963542,-0.043558225,-0.1260959953,-0.0966976434,0.0746505111,-0.0441109166,-0.2333352417,0.1888862401,-0.1680622846,-0.0500559472,0.1312600374,-0.065035291,0.0932962596,-0.2972727418,0.4525432587,0.1580045372,0.0668030009,0.0957956314,-0.4487333894,-0.2434641421,-0.1833722442,-0.1389517337,-0.4815650284,-0.3514602482,0.1729985625,-0.1922954917,0.0325003974,-0.357634306,0.2487705946,-0.0688276291,0.1517855376,-0.1557130516,0.181523785,-0.2259210199,-0.1300972849,0.2394524813,0.1117601991,0.0049152547,0.1079540253,-0.1028830558,-0.0198182855,0.6600618958,-0.4209561944,-0.0405635573,0.0502652861,0.330349803,0.1436003894,-0.4629877806,-0.2183720917,0.0778384432,0.2384808064,0.2662739456,-0.2090523392,0.2086628973,-0.3058154285,0.0186633207,0.0485654883,0.1629762501,0.1390302926,-0.2202229798,0.4223588705,-0.0216658469]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2724","title":"404 Error when loading remote data files from private repo","comments":"I have opened a PR to fix it @lewtun.","body":"## Describe the bug\r\nWhen loading remote data files from a private repo, a 404 error is raised.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nurl = hf_hub_url(\"lewtun\/asr-preds-test\", \"preds.jsonl\", repo_type=\"dataset\")\r\ndset = load_dataset(\"json\", data_files=url, use_auth_token=True)\r\n# HTTPError: 404 Client Error: Not Found for url: https:\/\/huggingface.co\/datasets\/lewtun\/asr-preds-test\/resolve\/main\/preds.jsonl\r\n```\r\n\r\n## Expected results\r\nLoad dataset.\r\n\r\n## Actual results\r\n404 Error.\r\n\r\n","comment_length":9,"text":"404 Error when loading remote data files from private repo\n## Describe the bug\r\nWhen loading remote data files from a private repo, a 404 error is raised.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nurl = hf_hub_url(\"lewtun\/asr-preds-test\", \"preds.jsonl\", repo_type=\"dataset\")\r\ndset = load_dataset(\"json\", data_files=url, use_auth_token=True)\r\n# HTTPError: 404 Client Error: Not Found for url: https:\/\/huggingface.co\/datasets\/lewtun\/asr-preds-test\/resolve\/main\/preds.jsonl\r\n```\r\n\r\n## Expected results\r\nLoad dataset.\r\n\r\n## Actual results\r\n404 Error.\r\n\r\n\nI have opened a PR to fix it @lewtun.","embeddings":[0.0712917,-0.0058958577,0.0056968681,0.5510850549,-0.1494427323,-0.0265011545,0.2285207957,0.4311418831,0.08946006,0.1929305941,-0.4208966196,0.0081157908,0.2076868862,-0.0518906005,0.21647273,-0.1607126296,0.0353167914,0.1799825728,0.2824764252,-0.2595870793,-0.177896589,0.137985602,-0.0538850538,0.3158785999,0.0962068215,0.0927776322,-0.0080373548,0.5909199715,-0.1477114558,-0.2709169984,0.3267617822,-0.0103071127,0.6517696381,0.8622717261,-0.0001145948,0.1480270028,0.2751790583,0.0864261612,-0.2178334147,-0.2196812183,-0.4140685201,0.1352059394,0.1286899745,-0.1258694232,-0.1592705548,-0.1347341686,-0.3020302951,0.2231125534,0.3179853857,0.3022303581,0.2062232047,0.2937770486,0.045304805,-0.2246668935,0.3055076897,-0.0233266652,-0.0246097464,0.4281001091,-0.0949198902,0.0712294653,-0.0279847682,0.0319528133,-0.0515540466,0.1331986189,0.1175860837,0.023218194,0.0301214289,-0.115299724,-0.0912810415,0.1707389951,0.4557885528,-0.0424045771,-0.1236855313,0.1272473782,-0.005594919,-0.2066437304,0.1167937145,0.3084259331,-0.1201704293,0.1909669042,-0.165840432,-0.4075729847,-0.3745573163,0.2887375951,0.1049820781,0.0034932778,0.0110238018,-0.037017528,-0.2777193189,-0.1479655057,0.0257504079,-0.192285791,-0.0764363334,0.1115421802,-0.0324635766,-0.0991886556,0.0933786407,-0.0600141659,0.3666199744,0.2095322609,-0.052750688,-0.0286936592,-0.4376348853,0.1852526218,0.0340921618,0.1648558527,-0.1191619486,0.3575214744,0.4924143553,0.4630121887,0.0169262942,0.1344595104,-0.266084522,0.1149586588,-0.1203830242,-0.2120414525,0.5364689827,-0.5240252614,0.0415623337,0.2673889399,-0.2513974309,0.030911494,-0.1081568077,0.3857171237,-0.105840221,0.002021546,0.1284597665,-0.025521785,-0.2464022785,0.0009908249,-0.1615304649,-0.1227628291,0.0280633252,0.1404120922,0.0828421786,-0.0892398432,0.1351307482,-0.0528042801,0.2862747014,-0.2229463458,-0.031881284,0.1365621835,-0.2249449342,0.2891194522,0.1576158255,-0.0658318773,0.0458842479,-0.2859483659,-0.0285500251,-0.0604846254,-0.7126704454,-0.4432533383,-0.1550359577,0.1526863426,-0.0801785216,-0.0131394276,-0.5630663037,-0.349994123,-0.2485773116,-0.1043686196,0.0599101745,-0.0011675559,-0.1250948012,-0.2015059441,0.165430963,0.5151728392,0.2380943447,0.0732332021,0.2899857163,-0.1883295327,0.2371861786,0.2451305687,-0.1567310095,0.2397658527,-0.3216828108,0.223928228,0.3520272672,-0.4791131616,-0.4761376977,0.6302224398,-0.3005096018,0.0485526323,0.0311094485,-0.1109769791,0.0631384701,-0.1202049777,0.5531357527,0.5033393502,0.1985224038,-0.0758811906,-0.3213317394,-0.2205793709,-0.3712443709,0.3887878358,0.0446488746,0.2431681156,0.32716313,-0.1752168685,0.4163027108,-0.0631747022,-0.2554619312,0.3558324873,0.4630145133,0.0490369983,0.216027528,-0.0766673237,-0.2324349582,0.1595869064,-0.417950511,0.1092520356,0.1248401329,0.0283907354,-0.3430932164,-0.0835298076,-0.4625395834,0.1354681998,0.0935154632,-0.0106914192,-0.2309899479,0.1119606122,-0.1363403946,0.448097527,-0.0777453259,0.2707433105,-0.1899019927,0.7521016002,-0.0140267937,0.0216276161,0.1388342232,0.2381316572,0.3733649254,-0.0674215257,-0.2315668911,0.3162270188,0.0854791999,0.054991398,0.7996613383,-0.2175320536,0.2576076686,-0.1705595255,0.0914442614,0.3963867724,0.1019060761,0.2840932906,0.2933969498,0.0532120131,0.0201182626,0.2820887268,-0.0100076245,-0.1659930795,0.1887242049,-0.1711800545,0.0001009119,0.0173523054,0.4308819473,0.0322616212,-0.2693648338,-0.0807472914,-0.1750596613,-0.043605078,0.189937368,-0.3221983612,0.103941381,0.3912779689,-0.0921079963,0.0149837425,0.1902210712,0.4368773997,0.2917283773,-0.0119424071,-0.2871611416,0.1384393275,-0.0360782966,-0.0098945135,-0.088818185,-0.008166384,0.1515764296,0.23627235,0.1220305264,0.0095654735,-0.0522959642,-0.4114744067,-0.174485907,0.2945167124,-0.3782201111,-0.092864126,-0.3838344514,0.0120319854,-0.0187726803,-0.222516343,-0.4318560064,0.0211320743,0.0102091944,-0.0336456373,-0.0828579664,-0.1378646046,-0.070489265,0.1969862729,0.0842645541,-0.2907545567,-0.1961331517,0.0306427255,-0.2443167418,-0.0588637702,0.1088309661,0.060475599,0.3242200911,-0.3790839911,-0.0734975934,-0.3865015805,-0.0106283473,0.0723735541,0.1662923098,0.2783843577,0.0176141858,0.1253294796,0.2831090093,0.1080457196,0.2510380745,-0.1643634886,-0.2009060532,0.2409310639,-0.1929204613,0.0178474728,-0.0302096307,-0.0647575706,-0.3637214601,-0.3206712604,0.0879066363,0.2696338296,0.1096406654,-0.3252704144,0.0601022765,0.2448760867,-0.1284328848,-0.089616634,-0.1150573939,-0.4633310735,0.2067125142,-0.3612317145,-0.24337475,0.2750439644,0.1787566245,-0.009640744,-0.1968368292,-0.7362080812,-0.546394825,0.0091494052,0.3281366825,-0.2830449641,-0.0138785327,0.0403666608,-0.1134740412,0.0461454839,0.0556835644,-0.2905862927,-0.0588692613,0.1242895946,0.2180033922,0.0700548962,0.2489507198,-0.168210417,0.4712434113,0.1406030506,0.2339785099,0.2989763916,0.0952111781,0.1760406643,-0.0963737518,-0.0063723614,-0.1188534945,-0.0447358452,-0.1703791469,0.1802875549,0.3123578429,-0.0383438393,-0.3510605991,-0.7951014042,-0.1874840111,-0.1440749168,-0.374894768,0.2968627214,0.0527787693,0.085448578,-0.1198735014,-0.0439892672,-0.1528048068,0.231516242,0.4026426971,0.1484474689,0.1749891639,0.1199255139,-0.0332212076,-0.4203334749,0.2535649836,0.0755813345,0.2377640754,-0.0301233921,0.2759259939,0.0985163227,-0.2100084424,0.1058966741,0.0361049697,0.5103294253,-0.016626643,-0.2256902009,0.028643189,-0.0576430224,-0.0102138082,0.0578977764,-0.2032496333,0.6543793082,-0.2079611868,-0.1942207068,0.1763012856,-0.1809899211,-0.1398560107,-0.4539441168,-0.326366812,-0.365655452,-0.269988358,-0.0750061944,-0.2250902504,0.1260140389,-0.1253050119,-0.0075171241,0.4114969969,0.2358535677,-0.0060951593,-0.0226975847,0.3146348,-0.2215196788,-0.0919528827,0.2925110757,0.263174504,0.3979368508,0.5468533039,-0.0179628916,-0.303809911,0.0546366572,-0.085270308,0.1474201679,0.3259739578,-0.0648432076,0.0217198227,0.2393686622,0.3835688233,-0.131944418,-0.0624960512,0.0948365778,-0.2763794661,-0.2218275815,-0.0904504433,0.5891396403,-0.2411394864,0.0365624949,0.0244708583,0.4571083784,-0.1401205957,0.4558717906,-0.2919656336,0.8274775147,0.0049737254,0.2439286709,0.4101966023,-0.3547481,0.1564416438,-0.7397397161,-0.0099833049,-0.125990212,0.1697987318,-0.1776978076,-0.0729514584,-0.0304346904,-0.3032511473,0.0024905889,0.3417636454,-0.1409403682,0.3539195359,0.2244943827,-0.0428399779,0.2108807266,-0.0116830962,-0.4715812504,0.119798027,-0.2007225603,0.109686397,-0.1476556808,-0.2658830285,-0.3184243143,-0.1552838534,0.0056647193,0.2628322542,0.2037681192,-0.1041891873,0.1269384623,-0.3423784673,-0.0262436979,0.0112964734,0.0307835843,0.0261253137,-0.4449074268,0.1467685103,-0.1597562134,0.0080905175,-0.0848922208,0.0291435178,0.0747098401,-0.2843890786,-0.3258242905,0.0543164127,-0.1399104297,-0.4502321184,-0.1715301275,0.2439087927,0.1090548262,-0.13749367,0.0408451781,-0.0244235247,0.0223379079,-0.1011320725,0.137578547,0.2148730308,-0.4253060818,-0.3450867236,0.2460391223,-0.2609076798,-0.0411151499,0.3521239161,0.0256753098,-0.0404810905,0.3096267283,0.1273037791,-0.0778211877,-0.1986033767,-0.0606390201,0.3912292123,-0.5329668522,0.0730199143,-0.0242915489,0.1300145239,-0.0429483876,0.030011395,0.0836008713,-0.4259628356,0.1155352816,-0.5240186453,-0.1020196155,0.0804446936,-0.3375594616,-0.1324581951,0.1408033818,0.1126266569,-0.0799625069,-0.3126099706,-0.2743678391,0.3055468798,-0.4892607033,-0.1393162012,-0.1846093684,-0.0006516779,0.182583198,0.1373849064,0.0930228084,0.1014142185,-0.1770511568,-0.1569723189,-0.2640835345,0.1940206587,0.0689002648,-0.1099165827,0.0154404165,-0.2093139887,-0.1625881046,0.1785517782,0.0698493198,0.0897794142,0.1022437066,-0.1386932731,0.2834094167,-0.1624087095,-0.2582081556,0.5439218879,0.1645564586,0.4544902444,-0.1603737026,-0.0083428025,-0.622023344,-0.0053134914,0.0474296585,-0.0015525603,0.2987638712,0.2713577747,0.6284823418,-0.5930604339,0.1254104674,0.0203015618,0.1417461187,0.0486497581,-0.3199189901,0.0865556151,0.0494384766,0.1315421015,-0.0468404926,-0.2668480873,0.5302373767,-0.0369671285,-0.2116624862,0.0265557282,0.124712199,0.0365080535,-0.0383878723,0.1480189562,0.4463038445,0.0528029986,-0.1928026378,0.1322217435,-0.1050554141,0.2909115255,-0.0960004032,0.1253875941,-0.1180556715,0.4324190617,0.0278544333,0.1020179242,0.1395347416,0.3179032803,0.123146221,-0.5315064192,-0.111193046,0.0352512226,0.2894666493,-0.0429254882,-0.2876443267,0.0758811757,-0.152910322,0.2027826756,-0.3038936257,-0.1328532845,-0.0116186077,0.1952102035,-0.0689476952,-0.1155721918,-0.0092683202,-0.0228974782,-0.0482824743,-0.0826075003,-0.2994631529,0.1374668777,-0.179945901,-0.3190341592,-0.1994658411,0.2556938231,-0.0502429567,-0.0652013198,0.2316329032,0.0840133354,-0.0877146944,0.0594907813,-0.138179183,0.317481786,0.2388736308,-0.0695733428,-0.031616684,0.1169638708,0.0013740589,0.297709316,0.1734185815,0.19632186,0.0938051343,0.1639337987,0.1328978837,-0.1214764863,0.1473803669,-0.1868278235,0.3492898643,-0.1421408057,0.5385420918,-0.3821104169,-0.0176974349,-0.3066322207,-0.2305183113,-0.3028339446,0.201814279,0.4240525961,0.1565279067,-0.0429730788,-0.0839135945,0.0513067171,-0.3016496301,0.3560424447,0.506207943,0.4123966396,-0.3134105504,-0.058559034,-0.3122221828,0.5392264724,-0.0847798958,-0.1129407212,0.2175698131,0.120184131,-0.2038061023,0.0475724265,0.2924028039,-0.1931979358,0.1637337804,-0.0409887023,-0.1292060167,-0.1509091258,-0.0806038752,0.1512299329,-0.0618643612,-0.3303131163,0.2133118212,-0.1874076128,0.0025669774,0.1399852186,-0.1163804606,0.1136191636,-0.203898564,0.4269225597,0.0840165168,0.0981647372,0.1510240734,-0.3223833144,-0.1258986443,-0.2285911292,-0.1086648479,-0.4821895957,-0.2606276274,0.0803881586,-0.1048964113,0.1089777574,-0.3725934923,0.2254531235,-0.1429125667,0.232692495,-0.056006182,0.1004685163,-0.1115435511,-0.2008741349,0.2023501992,0.0506413281,0.0410277694,0.0460924134,-0.1116688251,0.0726584345,0.6951723099,-0.3801924884,-0.134452939,0.0661311001,0.3405743241,0.0694084018,-0.3572324812,-0.1589224488,0.1254814565,0.2642308474,0.2726829052,-0.198208198,0.1934780329,-0.2318903804,0.112258926,0.0731371939,0.0944168121,0.008567296,-0.1341049671,0.3107875884,-0.0236611273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2722","title":"Missing cache file","comments":"This could be solved by going to the glue\/ directory and delete sst2 directory, then load the dataset again will help you redownload the dataset.","body":"Strangely missing cache file after I restart my program again.\r\n\r\n`glue_dataset = datasets.load_dataset('glue', 'sst2')`\r\n\r\n`FileNotFoundError: [Errno 2] No such file or directory: \/Users\/chris\/.cache\/huggingface\/datasets\/glue\/sst2\/1.0.0\/dacbe3125aa31d7f70367a07a8a9e72a5a0bfeb5fc42e75c9db75b96d6053ad\/dataset_info.json'`\r\n","comment_length":25,"text":"Missing cache file\nStrangely missing cache file after I restart my program again.\r\n\r\n`glue_dataset = datasets.load_dataset('glue', 'sst2')`\r\n\r\n`FileNotFoundError: [Errno 2] No such file or directory: \/Users\/chris\/.cache\/huggingface\/datasets\/glue\/sst2\/1.0.0\/dacbe3125aa31d7f70367a07a8a9e72a5a0bfeb5fc42e75c9db75b96d6053ad\/dataset_info.json'`\r\n\nThis could be solved by going to the glue\/ directory and delete sst2 directory, then load the dataset again will help you redownload the dataset.","embeddings":[-0.1489497572,-0.2841119766,-0.090395011,0.146070689,0.2751778662,0.0493270718,0.0770320743,0.2098800093,0.2618972063,0.0594649278,-0.085697338,0.0969992578,0.0986218303,0.2517307997,0.1167899966,-0.14294447,-0.1188406721,0.2136434615,-0.3757638633,0.1028596535,-0.0300949439,0.4992606342,0.0187852196,-0.0301720798,-0.3194895089,0.015822893,-0.0725437552,0.3714369833,-0.0625675693,-0.2505569458,0.2042683363,-0.0081873154,0.1753682047,0.2365752012,-0.0001000679,-0.0085890358,0.1501080245,-0.0443094373,-0.131183669,0.1888656318,-0.123258099,-0.1802763939,-0.0042734607,-0.0790977329,-0.4431518614,0.1293307543,0.149391681,-0.2958288491,0.5744659305,-0.026167728,0.3707997501,0.28149876,0.1990785748,-0.2768567502,0.1207759976,-0.1539129466,-0.0656856894,0.1609669179,0.0918901637,-0.1530089974,0.134279713,0.4699548781,0.0848045349,0.1188219041,0.2851537168,-0.0147726713,-0.2164845616,-0.159825772,0.3803760409,-0.002382593,0.3897214234,-0.2139130086,-0.1837674677,-0.1025153399,-0.0050807288,-0.0097253323,0.1055578887,-0.0117374668,0.1646526754,0.3102853298,0.0797114074,-0.3766746223,0.134093985,-0.1535574496,0.1237724796,0.0248707775,-0.2039152235,0.0251798332,0.2108683288,-0.0817251205,-0.3896350265,0.2301623225,-0.4329809844,0.1080717221,-0.2919595838,0.0540039763,0.2727687061,0.0597484149,-0.0113160992,0.1861447245,0.0120569412,0.0881809369,0.2470487207,0.0937485844,0.1267062575,0.2551039755,0.2777359486,-0.0926357508,0.2941237688,-0.0005995972,-0.1350926906,-0.1442201883,0.0105527854,-0.0579530597,0.1486091614,-0.0236186329,0.0687815174,-0.1440523863,-0.0859181285,-0.0094244853,0.0210778452,0.0438704528,-0.0676199943,0.298885107,-0.1251085848,-0.0440157801,0.1222792491,-0.0032418079,-0.1049566492,-0.1209477261,-0.4230038822,0.0173482299,-0.1510064155,0.1446428597,0.2698577642,-0.2826415896,0.4717916846,-0.0908560678,-0.1431017667,-0.0054614809,0.1562077701,0.1193652377,-0.1149761304,0.3126595616,0.0720781311,0.2043959498,0.1971702725,-0.2255741358,0.0935301557,-0.0622500405,-0.3156168163,-0.3906930387,-0.2478711456,0.2746995389,-0.2598507702,0.0476593114,-0.1760026366,-0.0673485398,0.0220543332,-0.131791383,-0.0509734936,-0.0088433828,-0.105749093,-0.3027727306,0.3504111469,0.4563004076,-0.384116441,-0.3124049604,0.0743108615,-0.3012655675,-0.2867965102,0.2582087517,-0.069087632,0.0903619528,-0.3982503116,0.3396105766,0.086900495,-0.2457402796,-0.441323936,0.1408777982,-0.1561100483,-0.0168532152,0.0004647762,0.0846227854,0.0410560891,0.1609211564,-0.0110930866,0.0727054477,0.129570052,0.1198460981,-0.2933918834,-0.2234240025,-0.1068688482,0.1781915426,-0.0184352919,0.048017703,0.0894301012,-0.053266339,-0.0724112988,-0.0346956961,0.1421084404,0.5428279042,0.3240688145,0.1759873629,0.0710827932,-0.1112197042,-0.4236648977,0.2405983955,0.0110830301,0.19942756,-0.072741203,-0.1040809453,-0.1806593537,-0.1200453416,-0.1043450683,-0.3273666799,0.2714445591,0.2450683862,0.2367115617,0.1192792505,-0.1000405326,0.5312882066,0.2279617637,0.1042723954,-0.5642830729,0.2393198311,-0.1372600794,-0.3061442673,0.0747153759,0.0397150628,0.0365927853,-0.2077722847,-0.0809380114,0.50306952,-0.2509528399,0.1525256038,0.2653539777,0.1040038317,-0.0787997693,-0.2012809813,0.0443733335,-0.1353620589,0.0737980455,0.0207623709,0.1207825392,0.0742132217,0.0406173207,-0.0525603667,0.0079512419,0.0271904189,0.1815289557,-0.1884570271,-0.1001354456,-0.1488170475,0.1305884868,-0.220273152,0.2337646782,-0.0128627382,-0.1599577516,0.4027630091,0.5056480765,0.1514839828,0.1703585982,-0.0077866623,-0.3467351794,-0.0944548994,0.0424062833,0.3587125242,0.5050246119,0.2802165449,-0.0021780601,-0.0187231414,0.0887213126,-0.0561165623,0.1006643325,-0.2698319256,-0.059102416,0.0573711172,0.1500862986,-0.1143287048,-0.3618562222,0.1692366749,-0.1863103956,0.2012964934,-0.1493307799,-0.1143300906,-0.2247170806,-0.2197960317,0.1550284773,0.0793028772,-0.2900716364,-0.1471260786,0.0979439169,0.0780382156,0.0648432449,0.1727213264,0.140134871,0.1194718778,0.0232448783,0.0420770533,-0.2549952567,-0.1949689835,-0.1610857546,0.1396798491,-0.1114572063,0.0598114878,0.3552449942,-0.2490369231,0.2142373472,-0.4008228183,-0.3154816926,0.0011929446,-0.1015660465,0.2772091925,0.3434364498,0.480493933,-0.1362862736,-0.2463300675,0.4448930323,-0.3510623276,-0.1793381274,-0.0874464512,0.0341214389,-0.1851829588,-0.3269742727,-0.1720096916,-0.162594676,-0.3666308224,0.4123755693,-0.0911129639,-0.0702770799,-0.0009885175,-0.0258267056,0.1959026456,-0.2662552595,0.0188400969,-0.4876939952,-0.5863960981,0.2309202701,-0.3134086132,-0.3265423477,0.1614696234,0.1782240272,0.2323613465,-0.3869133294,-0.6166652441,-0.5167448521,-0.261787802,0.2029007673,-0.2004734427,0.1902625561,0.3913955986,-0.0382905975,-0.1795982569,-0.1816891283,-0.4716680944,0.2314304858,0.0114416629,0.2168355286,-0.2339811325,0.3932483494,0.04659044,0.4432985485,0.0131256552,-0.0416375548,0.321998477,-0.082459271,0.456438005,-0.2917788923,-0.3831224144,0.1857813448,-0.0060609407,0.2260977477,0.268130511,0.1443547606,0.1405829042,-0.2879011333,0.1222026944,-0.237661913,-0.1655715555,-0.108788304,0.1454121917,-0.0104357731,0.1099621058,0.0119748609,-0.2576724589,-0.0426460132,0.3268521428,0.2962163091,0.1599027812,0.1903889179,-0.4206802249,-0.1156720594,-0.43706128,0.2657991946,-0.1600761563,-0.0160808023,-0.3290892839,-0.0653525367,0.0163332783,-0.0760554597,0.7218391895,-0.2210390866,-0.035433352,-0.0009087791,0.0285821464,-0.3047817051,-0.1213600263,-0.0297413226,0.0629951954,0.4729382396,0.3769618571,-0.3789212108,-0.1535685509,0.1134478748,-0.052930776,-0.1644910574,-0.1424074024,-0.3487765193,-0.1971343607,-0.2680314779,-0.1303834319,-0.0994106084,0.0301688481,0.2141271234,0.0713193938,-0.0748289004,-0.2438514978,0.0058636931,0.2063797116,0.4361850321,0.0788858533,0.2069233954,0.2102936953,0.3399574161,-0.1238630414,0.449716866,-0.1848240048,-0.0555162914,0.2116197795,-0.0791400522,0.0530190542,0.1041440442,-0.087690942,0.1449717283,-0.030471433,0.4150407314,-0.0847370327,0.1173214391,0.2567462921,0.1363582015,0.0978960767,-0.0575574376,0.1750542223,0.001387923,0.0433513038,0.1693202257,0.2645578384,-0.1397044361,0.0899081379,-0.3566861451,0.7948881984,0.0072723054,-0.0595920086,0.1683618277,-0.266959548,0.4844324887,0.0303155649,0.0382847227,-0.2218044549,-0.3330695927,0.1039323881,-0.0075712409,0.0718081445,-0.0013675725,-0.28924793,0.155691728,0.0943648219,0.1527907699,-0.237653926,0.2912339866,-0.0004337842,-0.0680206269,-0.0517967641,0.3398604393,-0.2662387788,0.3884881139,-0.1627099514,-0.1161806732,0.1328404695,-0.1030165553,-0.2831183672,-0.0023750577,-0.3912834823,0.167473793,0.2133184075,0.0145383403,-0.1710274518,0.0050387294,0.3181121349,0.180662185,-0.357966423,0.1715006381,0.0942139104,-0.1630607992,-0.1737090349,0.1159992367,0.1963389367,-0.111811325,-0.2726132572,0.0378258601,-0.1886433214,-0.2516732216,-0.2021656185,-0.0079317903,-0.1417811513,-0.1820685416,-0.2065447569,0.1918259859,0.0782819092,-0.1552026123,0.2192503661,0.2732061446,-0.0200236198,-0.0007663951,0.0992553607,-0.2341623604,-0.1573440582,0.5235729218,-0.2910884619,-0.0991078764,0.3289558887,0.0550471619,-0.1744873226,-0.2369523942,-0.1459594518,0.0769423172,-0.3101457059,0.0191043243,0.0729071498,-0.212857455,0.2415010035,-0.0643308684,0.2172026038,-0.1466227323,0.0190266706,-0.280313611,-0.2939015925,0.178979516,0.0187826492,0.2110987902,-0.0115306294,0.0773663148,0.0362934172,0.2026423067,-0.4853777289,0.1817747802,-0.1482910067,0.0133206798,0.4220645428,-0.2894528508,0.1827636957,-0.1572646797,0.2635748386,0.100976482,-0.5141131282,-0.3665655553,-0.0135366926,0.0669627264,-0.210591197,0.0188264959,-0.0643627495,-0.0037206442,-0.0521365963,-0.0427093171,0.2446481436,0.0344352759,0.0141831366,0.1515270621,0.0893867686,-0.0311039072,0.2417569607,0.0825499445,0.0548415594,0.0844556764,-0.0832610205,-0.0351253301,-0.2291374952,0.0378641486,0.0491033681,0.1135112047,0.0299337003,0.2438949496,0.0428875685,-0.2724184692,-0.2917987704,-0.0545394607,0.3474995792,0.0014477292,-0.0882969424,0.0474031717,0.0823991746,0.378772974,-0.2625285983,0.0694124252,0.2650741339,0.013172077,0.1829775572,0.1056490317,0.0961884856,0.0241867993,0.153967768,0.2842822075,0.5918827057,-0.2418418676,-0.0685136244,-0.0554467179,-0.1624848098,-0.2928983271,0.2425117046,-0.0054918597,0.0663073659,0.1724504679,-0.0896003321,0.1781722009,0.1148407534,0.2551929653,0.0139945438,-0.2192768306,-0.0379599519,0.1845485717,-0.1068760306,0.0447755158,-0.0614664368,0.6470045447,-0.4179852605,-0.3641883135,-0.2113063782,0.2803063691,-0.2529311776,-0.1189958304,0.1990026087,-0.0641167238,-0.0293624867,0.060433235,0.1364051104,-0.0747184306,0.3099988401,-0.005131613,0.0145470845,-0.1843179911,-0.260073036,0.1281093657,0.095344767,-0.0213207584,0.3233246207,0.0560827106,-0.227941826,-0.0128828762,0.5119288564,0.3439208269,0.1031911746,-0.0047943369,0.2766623497,0.0499539524,-0.135627985,-0.2410331368,0.2170783132,0.0172556397,-0.2678769827,0.1144609302,0.2914259434,-0.270986706,-0.031479422,-0.2785282731,0.2457702011,-0.0369521491,0.105526641,-0.2938429117,-0.1737419516,-0.1535031646,0.0162719432,-0.4971681237,-0.092129223,0.1305255443,-0.0643552989,0.0872596055,-0.3384013474,0.1706446856,-0.124982819,0.5146690011,0.402697444,0.0658305362,-0.0437743068,-0.2727631927,-0.6360180378,0.1922902912,-0.0027035098,-0.0223698877,-0.2565234303,0.2879667282,0.0265969392,-0.0453138389,0.2262425423,0.2340507358,0.05070379,-0.1486345977,-0.4697645009,0.0955657586,0.0967520326,0.0169140957,0.3438380659,-0.2005909234,0.0509592742,0.0119028818,0.237092182,0.0370290093,0.0664552301,-0.1040434912,-0.1123721302,0.508269012,0.1078963131,0.0672490895,-0.0397852324,-0.103502281,-0.1326908916,-0.397380203,-0.0432120003,0.0439188518,0.135822162,0.4293405712,-0.029446343,-0.1377387047,-0.2468767315,0.157139644,0.169945851,0.2434619814,-0.2193081081,0.0672191307,0.0822719857,0.0717636496,0.1030419171,0.2221031934,-0.0175563674,0.1006819084,-0.2025217265,-0.1591154933,0.6115162373,-0.1784405559,-0.0553496033,0.0610221773,0.2773207724,0.2382973731,-0.3452547491,-0.5500255823,0.0908088833,0.1909855455,0.1230885237,-0.1996174455,0.363550663,-0.3064097762,0.1511594653,0.0829545483,0.0998395234,0.2783595622,-0.1304607689,0.1929554343,-0.0530243479]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2722","title":"Missing cache file","comments":"Hi ! Not sure why this file was missing, but yes the way to fix this is to delete the sst2 directory and to reload the dataset","body":"Strangely missing cache file after I restart my program again.\r\n\r\n`glue_dataset = datasets.load_dataset('glue', 'sst2')`\r\n\r\n`FileNotFoundError: [Errno 2] No such file or directory: \/Users\/chris\/.cache\/huggingface\/datasets\/glue\/sst2\/1.0.0\/dacbe3125aa31d7f70367a07a8a9e72a5a0bfeb5fc42e75c9db75b96d6053ad\/dataset_info.json'`\r\n","comment_length":27,"text":"Missing cache file\nStrangely missing cache file after I restart my program again.\r\n\r\n`glue_dataset = datasets.load_dataset('glue', 'sst2')`\r\n\r\n`FileNotFoundError: [Errno 2] No such file or directory: \/Users\/chris\/.cache\/huggingface\/datasets\/glue\/sst2\/1.0.0\/dacbe3125aa31d7f70367a07a8a9e72a5a0bfeb5fc42e75c9db75b96d6053ad\/dataset_info.json'`\r\n\nHi ! Not sure why this file was missing, but yes the way to fix this is to delete the sst2 directory and to reload the dataset","embeddings":[-0.0967947617,-0.2115666568,-0.0807496384,0.1962106824,0.3262526095,0.2369837016,0.0145864468,0.2186023891,0.304212749,0.0247304682,0.0359603092,0.0324181393,0.0695079118,0.0254838355,0.185232088,-0.1839707792,-0.0519994311,0.2447137982,-0.2526428699,0.1134489998,0.0262401756,0.5897667408,-0.016213784,-0.0534840636,-0.4070296288,0.0637794733,-0.1067256927,0.386438787,0.0837384537,-0.3092555702,0.3094595671,-0.1223242581,0.1665251106,0.2378737181,-0.0001069555,-0.0172668993,0.1097328812,-0.0738142133,-0.2003081441,0.193641603,-0.1755623519,-0.2121768892,-0.0431495495,-0.1683145612,-0.3127231598,0.2077138275,0.2080174387,-0.360209614,0.4669828415,-0.0779851526,0.3034725189,0.0734519064,0.1870388538,-0.2253172845,0.0996841341,-0.064716883,0.0180001147,0.1627905965,0.0280185528,-0.0881333426,0.2278116643,0.5137569308,0.0191269275,0.066894874,0.3252755105,0.0058021583,-0.2148669511,-0.1175146177,0.4294656217,0.0252462607,0.5012630224,-0.284591794,-0.2556747496,-0.1477597654,0.0607869476,-0.0376200266,0.1306268275,0.0382047892,0.240204677,0.3680465817,0.1573650092,-0.4963627458,0.1782764643,-0.1671724617,0.2400638312,-0.0252192542,-0.2089725286,0.0257383753,-0.0523489974,0.0602182299,-0.4202076197,0.09361808,-0.4452365041,0.2293759733,-0.3384398818,0.0177128166,0.1564031392,0.0586689785,-0.0672714189,0.1756927818,-0.0096155424,-0.01781529,0.1846621186,0.0050517819,0.1165803894,0.3475584388,0.3002185225,-0.0608562641,0.2772581577,-0.0344318226,-0.112605691,-0.2116935551,-0.0267669167,-0.0445203669,0.3394462168,0.0709249079,0.0471464656,-0.1273613423,-0.0748827234,0.0052467301,0.0094822235,0.0462573543,-0.0173616577,0.2279712409,-0.1619673818,0.0669734329,0.0752746016,-0.1120772958,-0.1022966504,-0.2396534979,-0.4385360777,0.0857379884,-0.1416300684,0.2605865002,0.2251195014,-0.3529346883,0.5194398165,-0.15446347,-0.1339461207,-0.0671917573,0.1733881682,0.0164418183,-0.111294508,0.3600179553,-0.0070324023,0.2911078334,0.2413915545,-0.2093438208,0.0329281725,0.0017530004,-0.4485351145,-0.4871952534,-0.2713923156,0.205846861,-0.3041187823,0.1193074808,-0.2237885147,-0.1324516982,0.1100723818,-0.3623006642,0.0665284321,-0.0316081457,-0.1814624518,-0.3980076015,0.318770647,0.5888073444,-0.4055528939,-0.3779479563,-0.0020709571,-0.2126722187,-0.3125277162,0.2282436341,-0.083651334,0.1937862188,-0.5655229092,0.328269273,0.2421826869,-0.4067055583,-0.5046402216,0.1763454527,-0.0761452243,0.0242318809,0.0388752893,0.1862749904,-0.048539076,0.0908513293,-0.0068641757,-0.0596245863,0.0911441445,0.0738905072,-0.2160988748,-0.185294494,-0.0194679964,0.1719608158,0.1014139429,0.0600856431,0.1387099177,-0.0865883604,-0.1059987321,0.0313375369,0.1604241729,0.6651590466,0.2728141248,0.1942803115,0.1240620986,-0.1508184969,-0.4816050828,0.3217637241,0.0186104774,-0.0582429692,-0.1431977302,-0.1247108653,-0.0210408289,-0.2005503923,-0.1294278651,-0.4022967219,0.1712466925,0.145832777,0.2671712935,0.1689704508,-0.0652441531,0.6422885656,0.1865052134,0.0755148605,-0.5425470471,0.2478586435,-0.0030789503,-0.2368048728,-0.0396937653,0.1197547689,0.091819033,-0.1683971435,-0.1228633597,0.4949476421,-0.2293965816,0.0498527549,0.3663037717,0.1016435996,-0.104260169,-0.137436837,0.036056675,-0.1573877782,0.1217031702,-0.0146611696,0.1188355982,-0.0206106734,0.0581252538,-0.1098696142,-0.0893905312,0.0144160241,0.2158410251,-0.1967654079,-0.1474705786,-0.132848531,0.2166163027,-0.2527221739,0.2928239405,0.0272486284,-0.0646641478,0.3660409749,0.5254284739,0.1012593806,0.1509121954,-0.0414987616,-0.3780709803,-0.0576856919,0.0197057799,0.371759057,0.6975572705,0.2514449656,-0.0356979221,-0.1242030114,0.1776922792,0.0119530903,0.0613102689,-0.2530559301,-0.0369476229,0.1276064068,0.1222911775,0.0318813436,-0.2021453679,0.0858177394,-0.1157987565,0.1562950611,-0.0868853554,-0.0478942692,-0.2492735535,-0.2735444903,0.1308374554,0.1976600587,-0.2253232002,-0.1098191813,0.0344392359,-0.050612133,0.235156551,0.1227781698,0.0337840356,0.1383581311,0.0554196648,-0.0962215737,-0.2523566484,-0.1669972092,-0.1882661432,0.035055209,-0.1320614517,-0.0566251986,0.3350159228,-0.3225826919,0.2025213242,-0.5201035738,-0.2556633353,0.0153616453,-0.0054264311,0.1994890124,0.3850772381,0.3450616002,-0.1408618689,-0.1468842179,0.4020046592,-0.3022862971,-0.1455642134,-0.0385316126,0.0002109645,-0.2106548846,-0.2796366811,-0.1135790125,-0.2354622334,-0.2539646029,0.2946382165,-0.177436769,-0.1345800757,-0.0957031623,-0.1538654715,0.1646320075,-0.2699987888,-0.0264159776,-0.471652329,-0.6318153143,0.2939579785,-0.2295247018,-0.2231517732,0.2445824891,0.0589947253,0.314601928,-0.3103459477,-0.6782829165,-0.4436450303,-0.2228266448,0.285623461,-0.4130557477,0.1635397077,0.4808389246,-0.1488039345,-0.1070157886,-0.1414729506,-0.4869025052,0.0894894451,0.0037858344,0.3139515817,-0.2787528932,0.4786124527,0.0733766481,0.5729888678,0.0530609712,0.0298595708,0.2796929777,0.0171320941,0.5509269834,-0.2649046481,-0.3026599884,0.2523865104,-0.0624427572,0.1633922011,0.2103696167,0.1810739487,0.1441466212,-0.2752344012,0.0400380716,-0.3271701634,-0.0845713094,-0.1091479287,0.0389300138,0.0367598906,0.1565857083,-0.0447064489,-0.3745788038,-0.1370151937,0.3229603469,0.4588408172,0.2946390808,0.2685701251,-0.4339428544,-0.2006481439,-0.3611645401,0.2375063598,-0.0376680791,-0.0321627669,-0.2462547719,0.0461245403,-0.0542898253,0.0707106218,0.8077054024,-0.2146589607,-0.0551431663,0.0144533329,0.0718351752,-0.2178855538,-0.1947081983,0.1727670729,0.292131722,0.460299015,0.3915625811,-0.2678531408,-0.2193643153,0.1635451615,-0.0171571635,-0.2304357439,-0.2725433111,-0.3794245422,-0.039046891,-0.0797814056,-0.2582912743,-0.0535675175,-0.0432490408,0.2502156496,0.0968960077,-0.0923524648,-0.2009296864,-0.1086868644,0.02671206,0.4955202937,0.0252681244,0.1935300827,0.1733480841,0.3421194553,-0.1140537262,0.3577986658,-0.2585392296,-0.1132804006,0.3245096803,-0.159833312,0.0741028786,0.0861833021,-0.1993035227,0.1636456698,0.0218278654,0.3937485516,-0.0986721143,-0.0438540913,0.1909769475,0.1791538298,-0.0297380909,-0.1486721486,0.1355152279,0.11216598,0.0918472186,0.2919159532,0.1278932244,-0.178031981,-0.0020839772,-0.4114621282,0.7303284407,0.06027776,0.0234785657,0.0471579097,-0.2416367382,0.6816242933,-0.1125278175,0.1529457867,-0.1582616866,-0.4088335037,0.0609312914,-0.0639162734,0.1055738851,0.0336703621,-0.2475383282,0.3832395673,0.1345043182,0.1439774334,-0.1651840657,0.2133222669,0.0426835194,-0.0096132234,0.0558951534,0.238257885,-0.3472246826,0.4295883179,-0.2156779766,-0.0095860511,0.1674843282,-0.0825667232,-0.3599324524,-0.0465161279,-0.426512599,0.2485105395,0.2785868645,-0.0668641105,-0.1309327334,0.0868378505,0.3556089103,0.1657104045,-0.3290106952,0.2240129709,0.0849867091,-0.0657548383,-0.2196584642,0.0317459777,0.2520684004,-0.0553100407,-0.157159254,0.0981689692,-0.3675017655,-0.3502759933,-0.1498161256,-0.0038483324,-0.1148574352,-0.164468959,-0.2608122528,0.1685831994,0.2005791217,-0.1329322904,0.1327245533,0.3589597642,-0.0083718849,-0.0744090006,-0.0310706738,-0.2876268625,-0.1363266706,0.5017422438,-0.2428730726,-0.2337963283,0.3761483729,-0.0480007492,-0.2009577155,-0.1655632704,-0.0998668224,0.0081308233,-0.3345195353,-0.0508026369,0.0333758034,-0.0201608893,0.2324299067,-0.328828156,0.2337896526,-0.1422808468,0.0405851416,-0.4315919578,-0.2289014459,0.1564459503,-0.0141130555,0.1237567365,0.0188667029,-0.0220025294,0.0435594767,0.1835611165,-0.3920586407,0.2403355092,-0.1512723416,0.0164126139,0.3780587018,-0.1453934908,0.1978179961,-0.2091195881,0.1878336519,-0.0070134196,-0.4165580571,-0.2812808752,0.0399882123,0.108897455,-0.3066791296,0.1190101579,-0.102370277,-0.0167370737,0.0488022268,-0.0414369553,0.2671299577,0.0853449181,0.0666112751,0.1731099188,0.0559291542,-0.0533682406,0.0554095283,0.1705226451,-0.060293261,0.0776348859,0.0085270023,0.0390934311,-0.1214930192,0.1347723603,-0.0502778739,-0.013681652,-0.1389729977,0.1148260087,0.0675122812,-0.4156566858,-0.2959529161,0.0221160464,0.2257870436,-0.1327262819,-0.1999205202,0.1093705818,0.1244585961,0.2931163907,-0.2306100875,0.0554578118,0.410641104,0.1473627239,0.0764012709,0.0539946109,0.0977285653,-0.0133674527,0.3058975041,0.2899704576,0.6977534294,-0.3106092215,0.0015331106,-0.09254051,0.0339300521,-0.207159102,0.229062885,0.0474291593,-0.0009526642,0.2722170949,0.015616009,0.1789283752,0.2054872811,0.3731603026,-0.110072352,-0.4000127614,-0.1372208446,0.2010285705,-0.2470475733,-0.0090147844,-0.0738268495,0.6431599259,-0.3798422515,-0.2961468995,-0.1248712391,0.1860912442,-0.291266799,-0.0150754116,0.3056092858,0.0000539932,0.1109788343,0.0460697189,0.1823921353,0.0323167779,0.253801465,0.0520848036,-0.0765968859,-0.0945824459,-0.2046376467,0.2194266766,0.1017228216,-0.0381378122,0.2479143143,0.0605676025,-0.3091121316,-0.0519677363,0.5426909328,0.2848141491,0.0730677769,0.0074856137,0.1265549213,0.171031937,-0.1292843372,-0.1871746778,0.2179501355,0.0674464405,-0.3011257052,-0.0585716143,0.2554777563,-0.161511749,0.1505515128,-0.2555468976,0.2517746389,-0.1435961872,0.0600674264,-0.3752918541,-0.1392558813,-0.1722235829,0.2159517705,-0.4729170501,-0.0914306417,0.2356639653,-0.014075961,-0.0139210979,-0.2724198401,0.1286889166,-0.1506770551,0.3211682141,0.5002917647,0.0337833017,-0.1693235934,-0.2195383906,-0.5923278928,0.2428310066,0.0272281598,0.0931195691,-0.1942406744,0.2648688555,-0.0299212337,-0.0864042416,0.2662326992,0.1924630404,0.0803682283,-0.0920396671,-0.47814399,-0.0566295274,0.117373392,0.0293791518,0.2845123708,-0.2394271642,0.0562579408,0.13288863,0.1510089487,0.0085605765,0.0202639252,-0.0801181644,0.0257637296,0.5730304122,0.1187081337,0.0600104667,0.0495250672,-0.1236980408,-0.0332175419,-0.3297727704,0.113004297,0.0761105344,0.0947509184,0.3451061845,-0.244691059,-0.1119472235,-0.2737613916,0.1665126383,0.1354671568,0.2840458453,-0.3247120082,0.0768084079,-0.0207880344,0.1927497238,0.1993477941,0.1823449433,-0.0981678367,0.108079046,-0.1385545135,0.0620515831,0.5860341191,-0.2508491576,0.04632039,-0.0526143163,0.3319084942,0.27831164,-0.2728855014,-0.3691662848,0.0104999803,0.3397522867,0.1677312851,-0.2306310833,0.3144063056,-0.3140178919,0.2389194369,0.141329065,0.0455365442,0.1619603485,-0.1154981479,0.1430302858,-0.1168244705]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2716","title":"Calling shuffle on IterableDataset will disable batching in case any functions were mapped","comments":"Hi :) Good catch ! Feel free to open a PR if you want to contribute, this would be very welcome ;)","body":"When using dataset in streaming mode, if one applies `shuffle` method on the dataset and `map` method for which `batched=True` than the batching operation will not happen, instead `batched` will be set to `False`\r\n\r\nI did RCA on the dataset codebase, the problem is emerging from [this line of code](https:\/\/github.com\/huggingface\/datasets\/blob\/d25a0bf94d9f9a9aa6cabdf5b450b9c327d19729\/src\/datasets\/iterable_dataset.py#L197) here as it is\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batch_size=self.batch_size`, as one can see it is missing batched argument, which means that the iterator fallsback to default constructor value, which in this case is `False`.\r\nTo remedy the problem we can change this line to\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batched=self.batched, batch_size=self.batch_size`\r\n","comment_length":22,"text":"Calling shuffle on IterableDataset will disable batching in case any functions were mapped\nWhen using dataset in streaming mode, if one applies `shuffle` method on the dataset and `map` method for which `batched=True` than the batching operation will not happen, instead `batched` will be set to `False`\r\n\r\nI did RCA on the dataset codebase, the problem is emerging from [this line of code](https:\/\/github.com\/huggingface\/datasets\/blob\/d25a0bf94d9f9a9aa6cabdf5b450b9c327d19729\/src\/datasets\/iterable_dataset.py#L197) here as it is\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batch_size=self.batch_size`, as one can see it is missing batched argument, which means that the iterator fallsback to default constructor value, which in this case is `False`.\r\nTo remedy the problem we can change this line to\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batched=self.batched, batch_size=self.batch_size`\r\n\nHi :) Good catch ! Feel free to open a PR if you want to contribute, this would be very welcome ;)","embeddings":[-0.4098215103,-0.2538074851,0.0164554343,-0.0311614387,0.3398851156,-0.0579713434,0.3023079932,0.0674438328,-0.2175251842,0.3773573041,-0.2310502231,0.2912554741,-0.1922606975,0.2657689154,-0.0942452475,0.1088107154,0.0774022713,0.0824446902,-0.4071667492,-0.0259761699,-0.2257343829,-0.0934465826,-0.5538959503,-0.2494303882,-0.0387422554,-0.1291296035,0.0840008408,0.069073014,-0.2499088496,-0.1048584506,0.1157220602,0.5466902852,-0.0666164756,0.4911211431,-0.0001046605,-0.0466864854,0.220941633,-0.1478953958,-0.1971089542,-0.0162350703,-0.2287014872,0.3616317511,-0.1729200929,-0.0470902249,-0.1105783731,-0.3326943219,0.0179095548,-0.3529171646,0.1090987101,-0.0607379526,0.250481993,0.1520516723,-0.1833851635,-0.0110956132,0.3300379515,0.1602014601,-0.1887402982,0.2795745432,0.3021202087,0.1214002371,-0.0215312038,0.3396961689,-0.0457591675,0.0973074287,0.2876346111,-0.1221546829,-0.0756907761,-0.4055928886,0.1629078388,0.1327944696,0.4041192234,-0.3270388246,-0.2338447571,-0.3728273809,0.0683641359,-0.2494541854,0.1298157573,-0.0826154873,-0.1251394153,-0.1127450392,-0.4295096397,0.1275163591,-0.0135839991,-0.033932738,0.5006251931,0.4225344658,-0.1381253302,0.1263806075,0.2588524818,-0.206630975,0.4659806788,-0.0802471787,-0.2135605961,0.3368564248,-0.4075320661,-0.130137682,0.1735796779,0.2387518883,0.346085012,0.0446178056,0.2171569765,0.2294618189,0.1571498662,0.0546244234,0.4331026077,0.0627880916,-0.0688361675,0.0454046875,0.1574959159,0.3058146536,0.023727756,-0.02492515,0.2888559699,-0.0860852078,0.002355594,-0.0567303747,0.1077041775,-0.1669803709,-0.0952625722,-0.3858787715,-0.1767874956,-0.1537875533,0.2711011767,0.3885038793,0.0594843924,0.0154892504,-0.3158887625,0.2343372405,0.0019726157,0.1345427632,-0.2632302344,-0.2480551451,-0.2327021211,0.0905819163,0.2369067967,-0.4453908801,0.3494920433,0.3258509934,0.0771219358,0.0918583646,0.1089512184,-0.1570501477,0.2872882783,0.0994779319,0.0974388123,0.056982927,0.1213003322,0.1756840199,0.0762383863,0.088006489,-0.0508972406,-0.0789761767,0.4025254846,0.2578459382,-0.1870023608,-0.0908630863,-0.5048840046,0.1181122959,0.0382432751,-0.3109931052,0.1444637179,-0.2712101638,0.0033719011,-0.254860729,0.1507634073,0.1824268699,0.0524112433,-0.0907988623,0.0347349122,-0.2587924302,0.6150118709,0.4008066952,-0.2116879225,0.0417136326,-0.2732911408,0.077940397,0.1313033402,0.1164889857,-0.3975092173,0.1741449237,-0.1221157238,0.4353640974,0.059893094,0.0428745709,0.3480425477,-0.1915022582,0.2754980326,0.2648024857,-0.2541203499,0.2964873016,-0.1950291097,-0.0763328299,-0.0320449248,0.0939136818,-0.2064952999,0.3401971459,0.1230375394,-0.1803273708,0.3168271184,-0.2674479485,0.1347766817,-0.0228714179,0.2825080156,0.0961350203,-0.0171809867,-0.1778329909,-0.243828997,0.0989337862,0.163568303,-0.0619942807,-0.0031122109,-0.3148007393,0.1615255326,-0.1159022003,-0.4890494049,-0.2133393586,0.21111314,0.2268390507,0.0447983257,-0.1561160833,-0.1767099053,0.0718606934,-0.2924746573,-0.1943624467,-0.3770171106,-0.0943284109,0.1136750355,-0.2265723348,0.1598410457,0.1383999735,0.1202635765,-0.1816586554,-0.08545243,0.4552502334,-0.0204102639,0.0224799942,-0.0471611284,0.2453854084,-0.1095877588,0.0012365161,-0.1277897954,-0.0703821704,0.2200995535,-0.1837346107,-0.058323849,0.4448705614,-0.4544275701,0.2487808019,0.2326847017,-0.2751173377,0.4183245897,0.2991345525,-0.2019591331,-0.1013117731,0.0643185824,-0.1856444031,0.1916487515,0.0225364417,-0.3827545345,0.0753574073,0.4850607216,0.1322100163,-0.1092718691,-0.0830761939,-0.1139723659,-0.2782618403,-0.0191041678,0.3351629972,0.406793952,0.238962993,0.2084313184,0.1073293909,0.1976170242,-0.1879748553,0.1229329035,0.0111799175,0.1723272353,0.2312593609,0.1828304231,-0.1218916997,-0.4867269099,-0.3557829261,0.0885691568,0.3144856095,-0.2118907422,-0.2529515922,-0.1668187231,-0.2122499049,-0.1276170015,-0.2103358656,-0.0831772387,-0.220365569,0.13692303,0.7543545365,-0.1674434692,0.2527231276,0.3151095808,0.1411622614,-0.0560657084,-0.09979821,-0.1466866136,-0.0236520562,-0.0085362075,0.1970877498,0.1575423181,-0.1427107006,0.3563771844,-0.0000349335,-0.3080062866,-0.3034623563,0.130791381,0.0878003165,-0.1911070347,0.0158843715,0.0466463268,0.1843953878,-0.0100606075,-0.2569463253,0.2034876943,-0.4171120524,-0.0513277687,0.0685134158,-0.1702582985,-0.0884037316,-0.2802341878,-0.09832526,-0.1416513622,-0.3688466549,0.2363346666,-0.1134562641,0.0388994925,0.2580617368,-0.0042469669,-0.1164416224,-0.1947188824,-0.2532929778,-0.3594523072,-0.6516541839,0.0104260119,-0.2006954849,-0.2527257204,0.0317443088,0.1986943781,0.178017959,0.6395937204,-0.032256078,0.1754576862,-0.1056497917,0.0781790614,-0.0048312638,0.1112322807,0.4022190869,0.0161642283,-0.1008445621,-0.2005631179,-0.0300819241,0.1122007668,0.0311471485,0.2318634093,0.1879943609,0.4561055303,-0.055017069,0.9363186359,0.1427913159,0.1208270714,0.3089776635,-0.0406129807,0.294936806,-0.3398709893,-0.1727134287,0.4189908206,0.1589230299,-0.2530127168,0.2699559629,-0.1785174608,-0.3480798006,-0.142856881,0.1436568797,-0.0000630574,-0.0704254583,0.2039430588,0.1525276303,0.2158192545,0.0193438623,0.1884668171,-0.1212709397,-0.0508087873,-0.2181769013,-0.0350102708,0.3387905955,-0.0903509632,-0.4391949177,-0.38929829,-0.4123508036,0.0667664409,0.0623018816,0.5975311995,-0.1478609294,-0.1179704666,0.2101566046,0.1820258051,0.571148634,-0.4470500052,-0.1075448766,0.1274787188,0.0622721501,-0.0950158983,-0.153840363,-0.273807317,0.2109864652,0.0970169976,0.2901652753,-0.2299704999,0.0009556888,-0.0570670441,-0.0818059444,0.1513829678,-0.0959705934,-0.2688922882,-0.2532071471,-0.2367133796,0.1879086941,0.0725013986,0.0114183333,-0.2773074806,-0.038711708,0.0183910429,0.2690772712,-0.1216251925,0.340726465,0.1726690233,0.2733711004,0.1691559851,0.3852129877,-0.0037226311,0.3598803878,0.1930728704,0.2869406939,-0.1482043117,0.1700377911,-0.0655322075,0.1006533727,0.2456893623,-0.156115517,0.1534910649,0.1478286535,0.2866065204,-0.1028925255,-0.2666449249,0.3975700438,-0.0696273744,-0.4378696382,-0.8425217867,0.1484270096,0.0468965098,-0.138605848,0.5356468558,-0.3040621579,-0.1331494004,0.3750414252,-0.1486594975,0.6218429208,-0.0143620912,-0.1162524223,0.3653719127,0.2237394154,0.1280849874,-0.010589866,0.3386755884,-0.0961945802,-0.5596523285,0.1815623194,-0.105148308,0.0270031542,0.06616503,-0.2295008451,0.2785800695,-0.0691477656,0.16584526,-0.1683233082,-0.1908369809,-0.2544879913,-0.323390305,-0.3023680449,0.2215269655,-0.0260744318,0.0651989356,0.0785662308,-0.0854962915,0.0607560389,0.0981946513,-0.1658928543,0.1918345541,-0.2944265604,-0.1783633083,-0.3193566501,0.1619616449,0.0704469979,0.0123557867,0.0561574474,0.0476905368,-0.09444464,0.1550245285,0.3336307108,0.2662809193,-0.1004927531,0.0957866535,0.1077769548,-0.0701819211,-0.0577214174,0.0543721803,0.1306613088,-0.3757499754,-0.3450508714,-0.1822017431,0.2062250972,-0.142949149,-0.1285928339,0.0668815523,-0.0323937498,-0.2168657184,0.1690602601,-0.0859127939,-0.1933259815,0.4004814327,-0.2756020725,-0.0968619958,-0.2272102982,0.3841655254,0.2000458986,0.180243656,0.3196738362,-0.0627434701,-0.1752788275,-0.2885640562,-0.1654691845,-0.1228657514,-0.1576548517,0.1163402051,0.1648496836,0.03248908,0.0459673963,0.2720926702,-0.2012600601,0.1393444538,-0.3060130179,-0.174759984,-0.3050630987,0.0089874566,0.0825859383,0.1315063983,-0.087179482,0.2274681926,0.0897246525,-0.1020629629,-0.41252321,-0.323641032,-0.0959167853,0.1910414696,0.0058489298,0.0290751401,0.210662514,-0.028109258,0.1836047322,0.1174014956,-0.2297205925,-0.2758109868,0.1535899192,0.0850452036,0.1756462604,-0.0935233906,-0.0710833147,-0.2177160233,-0.5254070759,-0.2647674084,0.4250843525,0.1822078079,0.0167900231,-0.1717594117,0.4655924141,0.0090399589,0.0640558675,0.0469966605,-0.1677904725,-0.4819269776,-0.0941471234,-0.1503874362,-0.188601613,-0.2323779762,-0.1075756922,-0.0179499313,0.2478491962,0.313682884,0.1901548505,-0.2606304884,0.0899412185,0.2874260545,0.4517199993,0.5097648501,0.0900295228,-0.4366600215,0.2320919484,0.2794464231,-0.1522327214,-0.1554152369,-0.0217237715,0.0845896229,0.2006819844,0.0954713598,-0.1406143308,-0.2623805702,0.0018222801,0.0667469949,0.2619022429,-0.4957255125,0.2348495722,0.0708033442,0.0209592301,-0.0810224339,0.2266281992,0.0962515324,-0.0060231835,0.3671919107,-0.0813909248,0.2392350137,0.2457403541,-0.0151901254,-0.0518158153,-0.5209727287,-0.1195381209,0.5239408612,-0.132107228,0.4228522182,0.1253114492,0.0202691108,-0.2050748616,0.0062793498,-0.0397869237,0.0353534967,-0.1845470518,-0.0327133611,-0.0692916587,-0.2562203407,-0.0385797918,0.0194981713,-0.0797346085,0.247082144,0.2956745923,-0.1761798114,-0.0751433223,-0.4886924624,-0.0132662943,0.1551084369,0.4034704864,-0.1090623364,0.0698330998,-0.1169677526,0.0588612296,0.2569010556,0.3191128969,0.1653276682,0.3827476799,0.1881957054,0.1194787249,-0.1357768327,-0.1633858979,-0.1559182405,0.2363310754,0.1890111715,-0.0869611204,0.2447089106,0.2143074721,-0.1717334092,0.041557882,0.0931632295,0.207800895,0.0550413765,-0.0129984682,-0.1015723273,0.0064411722,0.0086368956,-0.0451540053,-0.314696312,-0.0214760583,-0.0340610445,-0.1987436414,0.1235641167,-0.2202576995,0.1245720387,0.0243965071,0.4130278528,0.1606728435,-0.3146054745,-0.1001112089,-0.3307814598,-0.5856945515,-0.0620807372,-0.0364675336,0.1141976416,0.0643222108,0.3521094024,0.076324597,-0.040760871,-0.0060292776,-0.0073859091,-0.1809274852,0.5266046524,-0.2156367451,-0.0458279774,-0.3805638254,0.120419763,0.1375428289,-0.274738282,0.2784750164,-0.0980341509,0.1694284081,0.12343245,0.2686367035,-0.0671863556,-0.3670603931,0.0654961616,0.195461452,0.0584265888,-0.1546788663,0.1712017357,-0.2905956805,-0.2137195319,-0.3174715936,-0.0259923059,0.1941355616,0.4024059772,0.1754822135,-0.085753724,-0.3212851286,0.17494452,0.1112530679,0.1403624713,-0.2917781472,0.375289619,0.0799085274,0.2514567673,-0.1469298154,0.0690483674,-0.0965512991,0.0188273806,-0.361923635,-0.3267616034,0.6507821679,-0.3065678775,-0.3087317646,-0.1929968745,0.3111165762,0.1510482579,-0.1815504581,-0.1616997868,0.3057670295,0.1714146286,-0.077444151,0.0539456494,0.2937753201,-0.2106450945,0.0878423825,-0.1030068174,0.0475121289,0.1259880066,-0.3081816435,0.3053467274,-0.2104217857]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2716","title":"Calling shuffle on IterableDataset will disable batching in case any functions were mapped","comments":"Have raised the PR [here](https:\/\/github.com\/huggingface\/datasets\/pull\/2717)","body":"When using dataset in streaming mode, if one applies `shuffle` method on the dataset and `map` method for which `batched=True` than the batching operation will not happen, instead `batched` will be set to `False`\r\n\r\nI did RCA on the dataset codebase, the problem is emerging from [this line of code](https:\/\/github.com\/huggingface\/datasets\/blob\/d25a0bf94d9f9a9aa6cabdf5b450b9c327d19729\/src\/datasets\/iterable_dataset.py#L197) here as it is\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batch_size=self.batch_size`, as one can see it is missing batched argument, which means that the iterator fallsback to default constructor value, which in this case is `False`.\r\nTo remedy the problem we can change this line to\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batched=self.batched, batch_size=self.batch_size`\r\n","comment_length":5,"text":"Calling shuffle on IterableDataset will disable batching in case any functions were mapped\nWhen using dataset in streaming mode, if one applies `shuffle` method on the dataset and `map` method for which `batched=True` than the batching operation will not happen, instead `batched` will be set to `False`\r\n\r\nI did RCA on the dataset codebase, the problem is emerging from [this line of code](https:\/\/github.com\/huggingface\/datasets\/blob\/d25a0bf94d9f9a9aa6cabdf5b450b9c327d19729\/src\/datasets\/iterable_dataset.py#L197) here as it is\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batch_size=self.batch_size`, as one can see it is missing batched argument, which means that the iterator fallsback to default constructor value, which in this case is `False`.\r\nTo remedy the problem we can change this line to\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batched=self.batched, batch_size=self.batch_size`\r\n\nHave raised the PR [here](https:\/\/github.com\/huggingface\/datasets\/pull\/2717)","embeddings":[-0.4272731245,-0.2991474271,0.0195668675,-0.01638785,0.3683030009,-0.0742568448,0.2522166073,0.0319571532,-0.2550874352,0.3826767206,-0.2251272947,0.3091110289,-0.156638518,0.2613864839,-0.0809769854,0.0484042093,0.0698413104,0.0641888082,-0.3650937676,0.0038317472,-0.2292107493,-0.0802459195,-0.5886576772,-0.2462791651,-0.0541951768,-0.1444052458,0.0832589865,0.0799205154,-0.277500093,-0.1343937367,0.1173299327,0.5228343606,-0.0547915548,0.4918593764,-0.0001062007,-0.0526980497,0.2312622666,-0.1509868503,-0.158042863,-0.0400756672,-0.1848453283,0.361739248,-0.2194947153,-0.0306547303,-0.1389433891,-0.3448046446,0.0368652903,-0.3153303564,0.0967374519,-0.0347801149,0.237539798,0.191700846,-0.1212639734,-0.0280031301,0.2924160361,0.1719074547,-0.1824315041,0.2875457108,0.2676300704,0.1521944851,-0.0238718614,0.3204404116,-0.0071707298,0.1072859243,0.2853108943,-0.1230788156,-0.1001666635,-0.349091202,0.1728558689,0.0927017331,0.3983657658,-0.3476824164,-0.2575695515,-0.3774476051,0.1001721472,-0.2640185654,0.1543089449,-0.0545621626,-0.1439808458,-0.1106420532,-0.465111047,0.1582112908,-0.0158581324,-0.0570740178,0.4968633652,0.4271092415,-0.1395782381,0.1440620571,0.302441746,-0.2181859612,0.4706703126,-0.0868221521,-0.1892041266,0.2885788381,-0.3797154725,-0.0865114108,0.1859757006,0.3104533851,0.3331645429,0.0668107197,0.1924347132,0.1984959245,0.1755506843,0.0483144931,0.4201283157,0.0409259759,-0.0753781348,0.1265605688,0.1535527557,0.3631804585,0.0221799817,-0.0169230737,0.3066366315,-0.1186105758,-0.0198443979,-0.062076252,0.1078238338,-0.1638439745,-0.1234035194,-0.3635532558,-0.1798371375,-0.1576935798,0.2531674802,0.3850029707,0.036850974,0.0222727936,-0.3015725017,0.2164818347,-0.0244715158,0.1341793537,-0.2679838836,-0.2204657346,-0.2469240725,0.0683018193,0.2176553458,-0.4639342129,0.3998603821,0.3153969944,0.0841448307,0.0757383481,0.0820631608,-0.1484253705,0.2845654488,0.1241616756,0.1023881063,0.0033236959,0.1186338738,0.1691438407,0.0723102838,0.0806405395,-0.0357542858,-0.0816931054,0.4457953572,0.2423689365,-0.2178844512,-0.1033536345,-0.505785346,0.1134827137,0.0618128777,-0.2920875251,0.1380295753,-0.2787775993,-0.0286093652,-0.2363350987,0.1620858461,0.1968349814,0.0631091148,-0.1123448685,0.0947522298,-0.2798263729,0.6055294871,0.3906852603,-0.1914422363,0.0441268384,-0.275184691,0.0175656229,0.1056244224,0.1124073714,-0.4302379489,0.140662536,-0.1127388999,0.4054118395,0.1039632633,0.0190833639,0.2940679491,-0.1907298565,0.300590694,0.1974322796,-0.2735237181,0.2857919931,-0.2215472311,-0.0772934705,-0.0801680237,0.0943416581,-0.182755366,0.351094991,0.1178758517,-0.1916355044,0.3264448047,-0.230678618,0.1404335648,-0.0187325962,0.269325465,0.0709471405,-0.0226345528,-0.194473058,-0.1946078539,0.0964828804,0.1274867654,-0.0395692624,0.0174686275,-0.3110537231,0.133405745,-0.0777347609,-0.4675628245,-0.213980481,0.1884703487,0.2690514028,0.1069603711,-0.1713611037,-0.1842455268,0.04976115,-0.3032926619,-0.1984602362,-0.4024328589,-0.0778271779,0.1093848646,-0.232486397,0.168279171,0.152189672,0.1157288104,-0.1488551944,-0.0577872694,0.4318814576,-0.0164120365,0.0250277743,-0.0024147534,0.1934576184,-0.1006301269,-0.0085080396,-0.1508764029,-0.0953718722,0.1943990886,-0.1137892902,-0.1086312532,0.4717007577,-0.4289268553,0.237362951,0.2320123464,-0.2679853737,0.4225678444,0.3004454374,-0.1771286577,-0.0828805491,0.0938339904,-0.1535351872,0.2155321091,0.0340240374,-0.4031252265,0.0773076043,0.4880266488,0.130141139,-0.1670159101,-0.0610636696,-0.0983368531,-0.2736434639,-0.0044806823,0.3481802344,0.4358070493,0.2623057961,0.2008986771,0.1217723489,0.1682746559,-0.196401298,0.1236011758,0.0139807742,0.1297875196,0.2391649634,0.2030436248,-0.1166223213,-0.4954636693,-0.3624055982,0.0904514343,0.3506451845,-0.2163149267,-0.2871666253,-0.1489726454,-0.1953310668,-0.140928328,-0.1437116563,-0.0571769923,-0.2207085639,0.1156101972,0.7453647852,-0.1563911289,0.2647349536,0.3040599823,0.1183139384,-0.0965170562,-0.1045953557,-0.1729465872,0.0155598158,-0.0261045508,0.1858373284,0.1454173476,-0.1560329199,0.3879520297,-0.0330667906,-0.3150498271,-0.3274518251,0.1140258089,0.1068755686,-0.1866953224,0.0165050402,0.0800929666,0.1169020236,-0.064924784,-0.2500439584,0.2245220393,-0.4285452366,-0.049407471,0.0648396984,-0.1821566075,-0.1111546159,-0.2636707723,-0.1237875149,-0.1443320513,-0.3541789353,0.3134951293,-0.0542375594,0.0126569383,0.2356290966,0.0210119821,-0.1136887297,-0.1966167539,-0.2750756741,-0.3700900972,-0.6795936823,-0.0143041918,-0.1872697026,-0.2510687709,-0.0029305809,0.2057149708,0.1311248243,0.6507316232,-0.0472378023,0.1140832976,-0.1225735545,0.1170194969,0.0047674603,0.1625114232,0.4045341909,0.0173028354,-0.1128574163,-0.1967088133,-0.0432602465,0.1491787136,0.0026647958,0.2568701208,0.1885203272,0.4343575835,-0.0448154956,0.9318587184,0.2090218514,0.0778398663,0.3147222996,-0.0572273321,0.3051575124,-0.3302345276,-0.1274739951,0.4436502159,0.1990623623,-0.2327396274,0.2648996413,-0.1420010775,-0.2722092569,-0.1420608908,0.1100354865,0.0199179128,-0.0591521747,0.1886474937,0.149256438,0.2408306897,0.0250429828,0.1920697689,-0.1574570835,-0.0457435064,-0.2449866533,-0.0556851812,0.3707216978,-0.0867130905,-0.4633236825,-0.3608346283,-0.4770020843,0.0765180662,0.084272854,0.5830452442,-0.1535139382,-0.069831714,0.2280002534,0.1688476354,0.5623604059,-0.4600425959,-0.0862117484,0.1134445146,0.0599834956,-0.071795322,-0.1283770651,-0.2897566557,0.2368079573,0.1028440669,0.3223494291,-0.2343145907,-0.0014540664,-0.0137271425,-0.1041437089,0.1541000754,-0.1282398999,-0.2935856283,-0.2976155281,-0.2457202822,0.2349117249,0.0851004273,0.023141237,-0.2734645605,-0.0238589942,0.0119232489,0.2733870149,-0.1169074774,0.3201583922,0.1877637208,0.2807949781,0.1656048,0.3887069523,-0.0139208073,0.4119775891,0.2298536152,0.2563448548,-0.1710545272,0.1790610552,-0.0441637561,0.0750183687,0.2353706658,-0.174800843,0.1367013305,0.1433432698,0.3488438725,-0.1215030625,-0.273653239,0.3809931874,-0.0802723616,-0.4285156727,-0.8337634206,0.1430363953,0.0484443083,-0.128649801,0.4761430621,-0.2529376149,-0.1296059489,0.3569486737,-0.1100163907,0.6182903647,-0.0374592282,-0.1220426187,0.3854730725,0.2120725662,0.1753130406,-0.0112520037,0.3247112036,-0.1200445443,-0.5671272874,0.176795736,-0.1060477048,0.0524615459,0.0513273515,-0.2447343171,0.3069968522,-0.0310998131,0.100846529,-0.132738173,-0.1770223677,-0.287864238,-0.3074367642,-0.3656770587,0.209245339,0.0033235706,0.0543825999,0.0924220011,-0.0801995769,0.0450223461,0.0829115808,-0.2076682001,0.1528564245,-0.2956769168,-0.1994718015,-0.2720076442,0.1640005559,0.0806443468,0.0640510246,0.0886751935,0.047665216,-0.1158013791,0.1994510889,0.3068114817,0.238757208,-0.063759096,0.0946014673,0.0916570351,-0.0648654625,-0.1095779166,0.0293083154,0.1034225821,-0.3887808621,-0.337998718,-0.1809224486,0.2157325596,-0.148026064,-0.1163280159,0.0281401221,-0.0605476499,-0.2440546304,0.1575246602,-0.0771222264,-0.1774614006,0.4190930426,-0.2543261647,-0.1035611406,-0.2397954613,0.4004264772,0.15599069,0.1724576205,0.2922992706,-0.0777603611,-0.1787797362,-0.2645465732,-0.1750203967,-0.1075892746,-0.1727489829,0.1153572053,0.1915488094,0.045164153,0.0578019246,0.2788959742,-0.1802144647,0.1132663116,-0.2797399461,-0.1713662595,-0.3364593387,-0.0181838088,0.0967242569,0.104348965,-0.0338134132,0.2767534554,0.1155722737,-0.1060873345,-0.4011127949,-0.3242248297,-0.1217625514,0.194077909,-0.0009890775,0.0020601489,0.1829856038,-0.0472668856,0.1802376807,0.1495660841,-0.2257641256,-0.2675334811,0.1299355179,0.0907683894,0.1750325412,-0.1236511618,-0.0871486068,-0.1924272478,-0.5336468816,-0.2714109421,0.3835996985,0.2163622081,0.0213812105,-0.1981917918,0.4558108449,0.0383652374,0.1087623015,0.0306230467,-0.1636762023,-0.4742843807,-0.0842974931,-0.1117584333,-0.2124139518,-0.2338621765,-0.112845473,-0.0109552471,0.2816009521,0.3137670457,0.2019557655,-0.2596470714,0.0610503107,0.260420233,0.4974555671,0.5223846436,0.10888464,-0.4005862772,0.2400654107,0.2642013133,-0.1962271333,-0.1504833251,-0.0039151469,0.1017635688,0.2101852298,0.1125913486,-0.1051113904,-0.2743268311,-0.0139406323,0.0641247854,0.300909996,-0.5234748721,0.1909853965,0.0377800427,0.059064094,-0.0551267974,0.2486810684,0.0907467678,0.0175588094,0.3432897627,-0.0917360261,0.2487661541,0.1816679239,0.0091002714,-0.0872705504,-0.563749373,-0.0648865551,0.5042432547,-0.1341985911,0.4263146818,0.1246122569,0.0802719295,-0.2331108153,0.025776593,-0.0595922917,0.0398360714,-0.1654189378,-0.0956293419,-0.0834148601,-0.2713201046,-0.0503459536,0.0783917755,-0.0710191429,0.2662337422,0.3442177176,-0.2110951394,-0.089491941,-0.5448992252,-0.024490878,0.1700504422,0.4162400067,-0.1242294014,0.0985096842,-0.0879343003,0.0305093117,0.254981786,0.3231106699,0.1940374076,0.3578949869,0.2437226027,0.0970356315,-0.1320351064,-0.1293485314,-0.1917389184,0.2409130633,0.1920018047,-0.06116179,0.2474955171,0.2056084424,-0.1704034358,0.0398506261,0.0844797939,0.2230893075,0.0494116023,0.0032020491,-0.1147306859,-0.0056809275,0.0459056534,-0.0318039693,-0.30720523,0.0152608501,-0.0817251131,-0.1991679668,0.1402749419,-0.2269389927,0.1203733906,0.0405086428,0.3888943791,0.1740291864,-0.3196875751,-0.1247259453,-0.3080353737,-0.5856832862,-0.0498037152,0.0003692443,0.0952776968,0.0602064766,0.3335166872,0.0458576605,-0.0793606713,-0.0271414556,0.009333632,-0.138519913,0.5552552938,-0.1867527664,-0.0734026507,-0.3905340135,0.1168008596,0.1728201956,-0.2786750495,0.2651463151,-0.0843021721,0.1487897933,0.0735833123,0.2755590081,-0.0609443784,-0.4022404552,0.0566847287,0.2168806046,0.0868055299,-0.1868018657,0.163871929,-0.3217539489,-0.1934355199,-0.3507672846,-0.0127855279,0.1770691574,0.4295945466,0.1623412222,-0.1155990213,-0.3018893003,0.1865535527,0.1334942579,0.142584011,-0.2665644288,0.3752622902,0.0570392944,0.2557085752,-0.1283230484,0.090284355,-0.0824771374,0.0107689779,-0.3692291677,-0.3205857277,0.6512409449,-0.3304809332,-0.2630523145,-0.1675060242,0.2789848745,0.1872299612,-0.24729985,-0.1981330067,0.3085832894,0.1872859001,-0.0839319453,0.0347228162,0.3118632734,-0.2490521967,0.0828082711,-0.0956064239,0.1276667565,0.1029315516,-0.3414039016,0.3555518985,-0.1924300343]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2716","title":"Calling shuffle on IterableDataset will disable batching in case any functions were mapped","comments":"Fixed by #2717.","body":"When using dataset in streaming mode, if one applies `shuffle` method on the dataset and `map` method for which `batched=True` than the batching operation will not happen, instead `batched` will be set to `False`\r\n\r\nI did RCA on the dataset codebase, the problem is emerging from [this line of code](https:\/\/github.com\/huggingface\/datasets\/blob\/d25a0bf94d9f9a9aa6cabdf5b450b9c327d19729\/src\/datasets\/iterable_dataset.py#L197) here as it is\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batch_size=self.batch_size`, as one can see it is missing batched argument, which means that the iterator fallsback to default constructor value, which in this case is `False`.\r\nTo remedy the problem we can change this line to\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batched=self.batched, batch_size=self.batch_size`\r\n","comment_length":3,"text":"Calling shuffle on IterableDataset will disable batching in case any functions were mapped\nWhen using dataset in streaming mode, if one applies `shuffle` method on the dataset and `map` method for which `batched=True` than the batching operation will not happen, instead `batched` will be set to `False`\r\n\r\nI did RCA on the dataset codebase, the problem is emerging from [this line of code](https:\/\/github.com\/huggingface\/datasets\/blob\/d25a0bf94d9f9a9aa6cabdf5b450b9c327d19729\/src\/datasets\/iterable_dataset.py#L197) here as it is\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batch_size=self.batch_size`, as one can see it is missing batched argument, which means that the iterator fallsback to default constructor value, which in this case is `False`.\r\nTo remedy the problem we can change this line to\r\n`self.ex_iterable.shuffle_data_sources(seed), function=self.function, batched=self.batched, batch_size=self.batch_size`\r\n\nFixed by #2717.","embeddings":[-0.395285517,-0.2580549717,0.0122515624,-0.0444010645,0.3418545127,-0.0624185801,0.3407534957,0.0728289783,-0.2151868045,0.3816896975,-0.2420467436,0.3263010085,-0.1489494294,0.2330623865,-0.105690375,0.0856827125,0.0678082034,0.0656198934,-0.3954455554,-0.0488205403,-0.199670136,-0.094870165,-0.5720129013,-0.2556624115,-0.0677141473,-0.0846328884,0.0704771057,0.1024620309,-0.2154195309,-0.1075428501,0.0926193446,0.4954774976,-0.074690409,0.5054728985,-0.0001026682,-0.0574065819,0.2356432378,-0.1376038194,-0.2195411026,-0.0302493908,-0.1989946365,0.325271219,-0.1885341704,-0.0532282367,-0.1088210046,-0.3041016757,0.0194372442,-0.3227274418,0.1237240434,-0.043613527,0.2722527981,0.184152633,-0.1845273674,-0.0124492077,0.322773844,0.1755822599,-0.1742073745,0.245048061,0.2785119712,0.1300680339,-0.0256195981,0.3138365149,-0.033237081,0.1198655814,0.248374179,-0.1133656129,-0.0957921669,-0.3572365642,0.1694220155,0.120104149,0.41530177,-0.3109550178,-0.2502095699,-0.3557595909,0.0617517345,-0.2333517522,0.1025430486,-0.0588076524,-0.1382247061,-0.137227416,-0.3907602131,0.1516192108,-0.0142845139,-0.0800749809,0.4868149757,0.4131761193,-0.1505860686,0.1337576658,0.2426598817,-0.208231464,0.4612360597,-0.0844351649,-0.1896630079,0.3290823698,-0.4131712019,-0.1198865473,0.189525187,0.2346197069,0.3065702617,0.0274492372,0.2186052054,0.2455582768,0.1560157984,0.0664189309,0.399307102,0.0655703619,-0.073348552,0.0599519946,0.1574373841,0.2660006583,0.0010127411,-0.027641505,0.2829092443,-0.1002530828,0.0191120971,-0.0642546117,0.0646939203,-0.1699994355,-0.1093565002,-0.3937206268,-0.1393920481,-0.157657966,0.2328296155,0.382784903,0.04573806,-0.0061983573,-0.3150723577,0.2183519006,-0.0133610517,0.1407299191,-0.2905318439,-0.2210107297,-0.2341721654,0.0672808439,0.2326530814,-0.4306276739,0.363615483,0.3096279204,0.0580112636,0.0988061726,0.0974460542,-0.131081149,0.2844367623,0.1003097072,0.0907799527,0.0529634356,0.1147043407,0.1794023663,0.0977330729,0.1009168848,-0.0389730819,-0.077579014,0.4097711444,0.2797968388,-0.1621289849,-0.1166302264,-0.4966014028,0.0977714732,0.0264714677,-0.2969416082,0.1220799237,-0.2623833716,-0.0040301029,-0.2655797601,0.1369915307,0.1750949025,0.0480485447,-0.0991999358,0.0420167521,-0.2386150807,0.6055750847,0.3968250453,-0.1979936212,0.025389744,-0.26798594,0.0942711532,0.1387880296,0.1055332348,-0.3805750608,0.1624964029,-0.116435118,0.4029414356,0.0591039583,0.0283916034,0.3533095717,-0.1723694205,0.255012989,0.2619579434,-0.2354220897,0.289771378,-0.2084997594,-0.0370955281,-0.0434016697,0.1031185836,-0.2168055326,0.3207089901,0.1404403001,-0.1597982943,0.3249726295,-0.2428826988,0.1307753026,0.0065736137,0.267475009,0.1109958291,0.0094419746,-0.1757867932,-0.1993084699,0.1043736041,0.1646412611,-0.0274507962,0.0260654893,-0.2968165278,0.1378114969,-0.1123035178,-0.4743850529,-0.215809375,0.2332164794,0.2692168951,0.0653190315,-0.1409410089,-0.1788410693,0.0797793195,-0.2837210596,-0.1925187856,-0.3703306019,-0.0724748373,0.0935656279,-0.2396397591,0.1416019946,0.1322721988,0.1111255065,-0.1848155558,-0.0992142484,0.4452389479,-0.0182956923,0.0225880463,-0.0215189941,0.2109621465,-0.100863345,0.0204873569,-0.1206693798,-0.0516751371,0.2375179529,-0.1775380522,-0.0810049921,0.4481786191,-0.4231111109,0.2334509492,0.2496015728,-0.2849743664,0.4315178096,0.3159383833,-0.1791001558,-0.0859930515,0.0594463311,-0.1643273383,0.1921038032,0.0120577561,-0.3628150225,0.0828747228,0.5004800558,0.124278985,-0.1139405295,-0.0790875331,-0.1050639376,-0.2931163609,-0.040204078,0.3298758268,0.4175722897,0.2348937839,0.2167894691,0.0916210264,0.1825415045,-0.1837881505,0.10601677,-0.0160710886,0.155884549,0.2231678218,0.1933310628,-0.1334583312,-0.5111271739,-0.358748585,0.1189773157,0.3281117678,-0.174154669,-0.2588981986,-0.1502901763,-0.185436666,-0.1441850066,-0.2180285603,-0.0639958456,-0.2242773622,0.1182956249,0.7470705509,-0.1694691181,0.2612163424,0.3097600043,0.122838147,-0.0258654244,-0.0698905736,-0.1623780429,0.0034671903,-0.0115190288,0.2090442479,0.136212796,-0.1282136589,0.327413857,-0.0055159922,-0.2933249176,-0.3096879125,0.1236642823,0.0742519051,-0.1981485188,0.003895611,0.0332587659,0.1694477797,-0.0203617085,-0.2203618288,0.2083944976,-0.4253174663,-0.0580276065,0.0841577202,-0.1362950057,-0.1155143157,-0.2796353102,-0.1106356159,-0.1303305328,-0.3811162412,0.2237794101,-0.119306393,0.0270372741,0.2457327396,-0.0125639709,-0.0998514071,-0.2004968971,-0.2281327695,-0.3536350429,-0.6426688433,0.004472774,-0.2282954156,-0.2500213981,0.0197868329,0.2030038238,0.1708155423,0.5690754652,-0.0140734091,0.1752255112,-0.1251138598,0.081649527,-0.005846221,0.1514060199,0.3951352835,0.009151875,-0.1177734956,-0.2075772285,-0.029656494,0.1155222878,0.0581953526,0.1880676001,0.1668992937,0.4367709756,-0.0494017266,0.8950824738,0.1428708881,0.132152155,0.2985138595,-0.0368691646,0.2923955321,-0.3575330973,-0.1551387161,0.4298119843,0.1756296903,-0.237230584,0.2805807292,-0.1754112542,-0.3373637795,-0.1380339414,0.1053652763,0.0075833616,-0.0673959032,0.1851004064,0.1751855016,0.2139600962,0.039593067,0.2124470621,-0.1274129599,-0.0129575422,-0.2204785794,-0.0379610024,0.3570187688,-0.1004811376,-0.4400269985,-0.3633901179,-0.4150956869,0.0617999323,0.0704475045,0.5814362764,-0.1244256347,-0.1370155364,0.2046886832,0.1702250987,0.5390096307,-0.4211828411,-0.0866010636,0.1404201537,0.0471746437,-0.1131010056,-0.1556282043,-0.2848849297,0.201935336,0.0943595767,0.2835620642,-0.2456188947,-0.0066402899,-0.0669973567,-0.0600547567,0.1638443023,-0.0955201387,-0.2584123611,-0.2532491386,-0.2427530438,0.1917889863,0.0562136993,0.0321590342,-0.3028338552,-0.0258855503,0.0138671258,0.2470598966,-0.1462973058,0.3139876723,0.1693783402,0.2529429793,0.1710644364,0.3494741917,-0.0021146215,0.3569175899,0.200405553,0.2610590756,-0.1407453716,0.1497525871,-0.0839250833,0.0581739359,0.2399302423,-0.1482049674,0.1583794951,0.1450879425,0.2906777859,-0.077172786,-0.2639668882,0.3942924738,-0.0522100031,-0.4087522924,-0.8269668221,0.1584316194,0.0012896541,-0.113835074,0.5107110143,-0.2877533436,-0.1037517786,0.3732539415,-0.1829810739,0.6333636642,0.0208031218,-0.084655568,0.3701159358,0.2117662877,0.1002806053,0.0207864381,0.3368802369,-0.1264138818,-0.5507479906,0.190042004,-0.0971868634,0.0142631754,0.0649558976,-0.2067329139,0.2744077444,-0.0512755886,0.1546070278,-0.1651279777,-0.19133313,-0.2651954591,-0.3068634272,-0.2825239301,0.2399929166,0.0116385883,0.067353107,0.0994033962,-0.0965883285,0.0617716312,0.0980949327,-0.1448837966,0.190747574,-0.2715595067,-0.1623438746,-0.2974942625,0.1488649547,0.0373670235,-0.0102249198,0.0628374591,0.0633740351,-0.0785243362,0.1623419374,0.3306056559,0.2739171684,-0.1024475992,0.0986648127,0.1362906545,-0.0703559816,-0.0849795789,0.0470862053,0.1127080023,-0.3892954886,-0.3255115151,-0.1574763805,0.2021942735,-0.1392416954,-0.1516213268,0.0626052544,-0.0882560909,-0.2405367643,0.1920385063,-0.0699578971,-0.2342935503,0.4180731475,-0.2685998082,-0.0980516747,-0.2461550981,0.3749428689,0.1942438036,0.1805411875,0.329572767,-0.0672459155,-0.1752368361,-0.2816379666,-0.1873762757,-0.1206653491,-0.1400337219,0.1451853365,0.1625614017,0.0707450733,0.0717012137,0.2692719698,-0.2494173646,0.1329619586,-0.3423103392,-0.1837316304,-0.2813356519,-0.0252705421,0.0823801011,0.1219394803,-0.0824993551,0.2444920093,0.0827338398,-0.1115204468,-0.4289260805,-0.2870428264,-0.0734982416,0.2168172896,0.020852169,0.0039178692,0.1856698543,-0.0035479369,0.212941885,0.1412555724,-0.2241715938,-0.3049322069,0.1624162495,0.0691655949,0.1946647614,-0.0795007423,-0.0730432272,-0.2223231494,-0.5550952554,-0.2659615278,0.4183388054,0.1838449389,-0.0151675604,-0.1432150453,0.4488349855,0.0254073683,0.1092382148,0.0484921746,-0.1501758248,-0.4723831713,-0.1114039272,-0.1350842416,-0.1840907931,-0.2128423303,-0.1175190583,-0.0332937315,0.2098447084,0.2955594659,0.2015798092,-0.2866376042,0.0806192234,0.2509213686,0.4870543778,0.511520505,0.1191678271,-0.4385605752,0.2362196296,0.2980528474,-0.1640665382,-0.1484626979,-0.0497935601,0.0783882141,0.238041088,0.091958046,-0.1407213658,-0.2618106604,-0.0074829976,0.0745636001,0.2535369694,-0.4904226959,0.201073736,0.0771134123,0.0083251242,-0.0626438186,0.2326396555,0.0909141749,-0.0016183942,0.4050097167,-0.0754602924,0.2788915038,0.2243276834,-0.0479754135,-0.0657507777,-0.5257908106,-0.0847144276,0.4974823296,-0.1178354993,0.4256037772,0.1208201945,-0.0085019702,-0.2198422104,-0.0026530703,-0.0617936999,0.0304682292,-0.2202529609,-0.0546375997,-0.0514200702,-0.2534153163,-0.0065544234,0.0450404286,-0.0675662234,0.2674799562,0.2610012293,-0.1945774257,-0.0570329875,-0.4699365497,0.0056656012,0.147436142,0.3894799352,-0.0841645002,0.0609901659,-0.0977733433,0.0293720979,0.2505256534,0.290519774,0.1760125458,0.3733549416,0.1806573123,0.1349145174,-0.1499205381,-0.1512183845,-0.1504082382,0.240720883,0.1586939692,-0.1102179438,0.253356576,0.2345554978,-0.1958837509,0.0268632378,0.0911846235,0.1904039383,0.0310521368,-0.0189555697,-0.1127010807,-0.0085153906,-0.0194328725,-0.0459788479,-0.3121986389,-0.0312043801,-0.0265318286,-0.1948318928,0.1080226675,-0.2485329509,0.1310255229,-0.0222970806,0.3993190229,0.1608640403,-0.3187510967,-0.0678831339,-0.3199674785,-0.5710173249,-0.0623152182,-0.0096944142,0.104960613,0.0389434323,0.3603206277,0.0703562647,-0.0854467377,0.007664897,-0.0224442296,-0.1641293913,0.5294302702,-0.2095310837,-0.0493511818,-0.3857730329,0.0950081125,0.1377560049,-0.2934909463,0.2628519535,-0.1146349311,0.1892690063,0.1227997765,0.2842699587,-0.06275668,-0.36529091,0.0593192726,0.1677443087,0.0446486734,-0.1589602977,0.1774675399,-0.2936632037,-0.1825018227,-0.3166127801,-0.0393249393,0.193033725,0.4018985629,0.1966043115,-0.0932723135,-0.2921393812,0.1905105859,0.133834675,0.1184573025,-0.2334136963,0.3482763767,0.0693167895,0.2110519409,-0.1367331743,0.0856811106,-0.0985412896,-0.0003547508,-0.3763158917,-0.3303907514,0.6369723082,-0.30311957,-0.3353652954,-0.2112795115,0.2951059639,0.1413862407,-0.1821120232,-0.1679439992,0.3190461099,0.1603797972,-0.0773841366,0.0378353596,0.2933623493,-0.2244141996,0.0770396516,-0.1009260863,0.0484861955,0.1310122758,-0.2831964493,0.3008265197,-0.1869720966]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2714","title":"add more precise information for size","comments":"We already have this information in the dataset_infos.json files of each dataset.\r\nMaybe we can parse these files in the backend to return their content with the endpoint at huggingface.co\/api\/datasets\r\n\r\nFor now if you want to access this info you have to load the json for each dataset. For example:\r\n- for a dataset on github like `squad` \r\n- https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/squad\/dataset_infos.json\r\n- for a community dataset on the hub like `lhoestq\/squad`:\r\n https:\/\/huggingface.co\/datasets\/lhoestq\/squad\/resolve\/main\/dataset_infos.json","body":"For the import into ELG, we would like a more precise description of the size of the dataset, instead of the current size categories. The size can be expressed in bytes, or any other preferred size unit. As suggested in the slack channel, perhaps this could be computed with a regex for existing datasets.","comment_length":71,"text":"add more precise information for size\nFor the import into ELG, we would like a more precise description of the size of the dataset, instead of the current size categories. The size can be expressed in bytes, or any other preferred size unit. As suggested in the slack channel, perhaps this could be computed with a regex for existing datasets.\nWe already have this information in the dataset_infos.json files of each dataset.\r\nMaybe we can parse these files in the backend to return their content with the endpoint at huggingface.co\/api\/datasets\r\n\r\nFor now if you want to access this info you have to load the json for each dataset. For example:\r\n- for a dataset on github like `squad` \r\n- https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/squad\/dataset_infos.json\r\n- for a community dataset on the hub like `lhoestq\/squad`:\r\n https:\/\/huggingface.co\/datasets\/lhoestq\/squad\/resolve\/main\/dataset_infos.json","embeddings":[-0.041354809,-0.5577720404,-0.1589672863,0.3919742107,0.2053750604,-0.0639249161,-0.0737590641,0.0503646508,0.1064483523,0.0798681453,-0.4394702017,0.0881778374,-0.0656972155,0.4548651874,0.0609381124,0.0388585962,-0.2295204401,-0.112665996,-0.1492429972,-0.2198024839,0.0234504249,0.2798558772,0.2534013987,-0.0295910668,-0.0154504152,0.1214822605,-0.254406631,0.2946007848,-0.5448058248,-0.2168553919,0.162506178,-0.0575589314,0.1592405736,-0.0210057553,-0.0001038512,-0.0647772849,0.1080954596,-0.2295954525,-0.3205727339,-0.0383348875,0.0025480704,-0.3959295452,0.1042662635,-0.0850442499,-0.1412745118,-0.1166362092,0.0335714296,-0.1261952221,0.4541440308,0.0918545425,0.2733739913,0.4500881135,0.1584033072,-0.1389549822,0.0338327177,0.4866320789,0.0143970624,0.2679385543,0.3840728402,0.3003026545,-0.0922307894,0.405079335,0.2523309886,-0.0259290654,0.2601458132,0.0314368606,-0.3214984238,-0.1370878816,0.1025048047,0.4285010099,0.1674049348,-0.3730247617,-0.5177248716,-0.4534083009,0.1588495076,-0.254727155,0.0493003577,0.2319036126,0.030197721,0.1767473519,-0.2547946274,-0.1507339627,-0.2382446975,-0.024006838,-0.1776671559,0.0789698511,-0.2915142477,-0.1185309142,0.0026724185,-0.4049320817,-0.6295150518,0.0317048617,-0.1120769009,0.204180643,-0.1458263695,-0.382292062,0.3146906793,0.2405808419,0.5023838878,0.1578407586,0.0075038429,0.1524950713,0.0011088211,0.1947938204,0.1977786869,0.3022005856,0.3661184311,-0.0779799819,0.1031197831,0.0316854157,0.1499069482,-0.1356841922,-0.213738367,0.0816751868,-0.3636180162,-0.2070543468,0.0226015691,-0.1112502217,-0.0727842227,-0.0099661201,0.1764284372,0.0235274211,0.0794332996,0.439917773,0.0285252072,0.0114363786,-0.094827041,0.2278685868,0.1659278572,-0.2580323517,-0.1681976616,0.0100567983,-0.0577120967,0.1395848393,0.0128062749,-0.3011850715,0.3934868276,-0.1103335544,0.1721936017,-0.0499964878,-0.0783385783,0.0496042967,0.1866596937,0.0405115597,0.0921569839,-0.2169891447,0.0278592855,-0.1454958767,-0.3767353296,-0.2053822875,-0.1997884065,-0.2601394951,-0.4404108524,0.2663024664,-0.234716177,-0.0646486208,-0.4436706305,0.3297384381,-0.1481907368,0.138103798,0.0043476312,0.089858152,-0.0293878783,-0.1064031199,0.0048883492,0.2305946648,-0.0331626125,-0.275741905,-0.0361056812,-0.2553452849,-0.2473154515,0.4085282981,-0.0546536222,-0.278449744,-0.1099671051,0.2847639918,0.3449617624,-0.4075211585,-0.0638781562,0.2323251367,0.0420696475,0.0920103788,0.0441864021,0.0051910719,0.2295906395,0.1645947546,0.0175024141,0.3218150139,-0.1843339354,0.1228394881,-0.1066689193,-0.4086617827,-0.175804019,0.1705343723,0.0425586514,-0.3353559673,0.2083470821,0.0844615996,0.2212419808,-0.1525700092,0.0652610213,0.2249132395,-0.0162420589,-0.0485722311,0.0501228198,-0.1123514995,-0.6196359992,0.1310897619,-0.0196532421,0.237208575,-0.0399618447,-0.4177995026,-0.3142914176,0.1552389115,0.0596921146,-0.0916106254,0.2139571309,-0.0335062295,0.2626999021,-0.0538792685,-0.1323897243,-0.0153961927,-0.1248895526,0.299482137,-0.6840795875,0.0677038357,-0.0187636875,0.0317224115,0.1639617532,0.1434708983,-0.2023173273,-0.0474745519,0.0281766895,0.0756473839,-0.1749101728,0.4565301239,0.260379523,0.4778719544,0.2401754558,-0.0033623995,0.2902507484,-0.1425657272,0.0222379602,0.1272934526,-0.5046000481,0.5015555024,0.3156271279,-0.0389141738,0.0748621151,-0.058989089,0.0190481506,-0.0133134108,-0.2305651307,-0.3546416163,0.0251254924,-0.0404179804,0.2994695902,0.0617636181,-0.3438564837,0.0625477806,0.5101698637,0.0132961441,-0.0046376036,0.345225811,-0.31604442,-0.2252226025,0.2095775306,0.055681482,0.1830606908,0.373601824,0.1089994982,0.1292484403,0.211334765,-0.0286117196,0.1002286598,0.1177811474,-0.0578761175,0.0437692255,0.3493605852,-0.0645666197,-0.4000751078,-0.1172626019,-0.2899731696,0.0852834731,-0.2282990217,-0.0175707508,-0.1347502768,-0.364479512,-0.1586054862,-0.099081628,-0.5645573139,-0.2226260751,0.185556218,-0.247871384,-0.3137249947,-0.2181916833,0.1011981815,0.2453226298,-0.0683198795,0.0089634182,-0.0908100903,-0.085657604,-0.0101825539,0.1782536358,0.1776801795,0.0589755811,0.3929869831,-0.2609719634,0.1044503301,0.006272601,-0.5086253285,0.2717722952,-0.1840400398,0.1842017323,0.0771672577,0.1117586568,0.0895381868,0.0349557251,0.1722739339,-0.0007507968,-0.3836705983,-0.093941763,-0.0080974372,0.1394441426,-0.161365211,-0.1860440522,-0.3122625053,-0.3296838999,0.3015921712,0.0786694139,0.2122530043,0.211279273,-0.0106809111,0.1425525248,-0.1424604058,0.0097210035,-0.3036101758,-0.4350679219,0.2735942304,-0.4120613337,-0.1951003671,-0.057308048,0.0915092304,0.2012043446,-0.2210039496,-0.1597380191,-0.5394004583,-0.122895591,-0.027982872,0.1173581108,0.0357880518,0.255788058,-0.1290501505,-0.1456703097,0.0228043478,-0.2735737562,-0.0556732379,0.2504128814,0.0304258671,-0.2115626633,0.2764234245,0.1665291935,0.2871811688,0.2273597419,-0.2123124897,0.1203336269,-0.1358999312,0.2125939727,0.0071700527,-0.0339481533,0.2497866601,0.119216755,0.0755948722,0.483152926,0.395652622,0.28299734,0.0230872165,-0.0926339403,-0.1653816104,-0.1985621154,0.3444508016,0.2088212073,-0.1278043985,-0.0347601585,-0.0861575007,-0.1628296375,-0.1396651566,-0.1885163188,0.313739419,0.0045145904,0.1377198994,-0.5218554139,-0.1015992314,-0.2448652685,0.5750691891,-0.0382529572,-0.3285715878,-0.0430589579,-0.2700404823,0.1268565655,0.0676165447,0.0229030717,-0.0856686383,-0.3362790048,-0.1821859926,-0.2599523365,-0.3137333393,0.2416992486,0.009534278,-0.1177260429,0.4289170206,0.4547967315,-0.5648524165,-0.2424123883,0.1274438053,0.2162189186,-0.1528227031,-0.0360643193,-0.1462424695,-0.0884627551,-0.492646426,0.0405902229,0.0551720299,0.1614610106,0.0294254068,0.0009253473,0.0585434102,0.0606929846,0.405000031,0.605887711,0.1253755838,0.137028411,-0.1399495304,0.1495877951,0.4613379538,0.27388978,0.6469154358,0.1012863517,-0.1358736157,0.0203396678,0.04022168,0.2384429276,0.3605240285,0.1815389544,-0.0088942628,0.0315243118,0.1655163169,-0.3032897711,0.100359939,-0.048551064,0.0017776511,-0.1444608867,0.0508659258,0.4109845161,-0.0491495095,0.0982712582,0.1295281649,0.3705095351,-0.2069345564,0.214183405,0.0672574565,1.064089179,-0.2176128328,0.1019983664,0.0588854775,-0.19377096,0.4219002724,-0.0598734058,-0.0264882594,-0.2077640146,0.256903708,0.1574484259,0.1491709948,0.145873338,0.0992430896,-0.1668533832,0.0574493036,0.0623663366,0.1175973415,-0.3965123296,0.590929687,-0.4135976434,-0.5295298696,-0.5926879644,0.2633858323,-0.0887999982,0.3599303663,0.0462489128,-0.278126359,-0.1870384663,-0.2317658961,-0.3549026847,-0.1536562741,-0.0324338228,-0.2425073981,0.1593876034,-0.3184924126,0.0122611402,0.328658402,0.0820262209,0.1905644238,-0.2280524373,0.159279868,-0.5026181936,0.1174357235,0.1281570345,-0.0831608251,0.0958847851,-0.178292349,0.0170137156,0.0572190471,0.315084219,0.0120746214,-0.0044423463,-0.0037434944,-0.2274231464,-0.2892987132,-0.0151894847,-0.0078870989,0.0582145117,-0.0911117941,0.1713141799,0.197720632,0.0618448108,-0.0394436046,0.3849599063,-0.0977170914,-0.2396054119,0.2264370173,-0.0940747187,-0.3045296073,0.0304970108,0.1447835416,-0.0074710511,-0.1658855677,0.2489944547,0.3029513657,-0.3718908429,-0.0993817672,0.0348030366,-0.2284687161,-0.098403275,0.1012171358,0.2532755435,0.0787533075,0.2638927102,-0.1840360761,-0.377127111,0.4878149331,-0.0138024008,0.1850367486,-0.2188996226,0.0506912582,0.0482552312,0.4039393663,-0.4329450727,0.092218779,-0.2767193615,-0.0288183857,0.3223300874,0.0294945799,0.3968377113,-0.176538229,0.0903924033,0.015761761,-0.296946317,-0.2389743179,-0.1329906136,0.0942663476,-0.0950658545,0.2489352226,0.0663092807,0.1162624434,-0.0040642619,-0.2708457112,0.3166065812,0.1042487547,0.1591234952,0.0767232552,-0.069843635,0.0723240897,0.1301015466,-0.0857838392,0.345618844,0.3020139933,0.0362830348,0.0966205746,-0.2138929367,-0.1111759394,0.1031236127,0.2083468139,0.2305243462,0.0727463365,0.0921946988,0.2423118949,-0.2250280827,0.2563063204,0.3816102743,0.0709071904,-0.1218112037,0.2718301415,-0.0183460489,0.2761364877,-0.1224985197,0.1581377238,0.183390379,-0.0879193917,0.1031722873,0.3402597606,0.1969461292,0.4780899584,0.255649507,0.0159268212,0.3624537587,0.0430255271,-0.012007718,-0.0018365071,-0.1979355067,-0.0642859265,0.4485360384,0.206879586,0.5565649271,-0.1806066185,-0.322958678,0.2026271671,-0.144753769,0.1746148765,0.3417482078,-0.2067809254,-0.2872588336,0.1873378158,0.3389520347,0.2000156492,0.1048683673,0.4568769634,0.1618061811,-0.3141474724,-0.2933790684,0.4585602283,-0.0922621712,-0.0835779533,0.0667582974,-0.2180444598,-0.3744446337,0.03840537,-0.0629543513,-0.4746851623,0.1443571597,-0.0760225877,0.1167377159,-0.2503052652,-0.0875759274,0.0984767377,0.0628845692,-0.2830202878,0.1778156161,0.2792690992,0.0524337105,0.1525217295,0.2958326638,0.3038707376,-0.1640508324,0.1301806271,0.3555209041,-0.0374775715,-0.0991042778,-0.2428997308,-0.0109764729,0.08004985,-0.0562434271,0.3106294572,0.2558509111,-0.2411569506,-0.1995712519,-0.3234943748,0.0794741586,0.013972654,0.3263380229,-0.0467860848,-0.0423301272,-0.3483895957,-0.2470829338,-0.4944185913,0.0992266983,-0.0236442126,0.1314777583,0.0442094542,-0.2457542121,0.1170220152,0.1964775026,0.2438388318,0.5422875285,0.003423501,-0.0112321917,-0.2232891768,-0.4496414065,0.0686887577,0.0492492616,-0.1442249715,0.0065491698,0.3821818829,0.2754144371,0.1455078721,-0.1040340289,-0.2930573523,-0.0118301054,0.0118323443,-0.3173431456,0.1339316815,-0.0657759532,-0.1426742524,0.2859062254,0.0234383531,0.0873940289,-0.2667748928,0.0612366684,-0.3607884645,0.2922792733,-0.1789945364,-0.1232889369,0.2861504257,0.0752062201,0.1630211622,-0.259696871,-0.1424851567,-0.1944019049,-0.2954752147,-0.2570571005,0.423702836,-0.0181251261,0.3955463469,-0.0450453088,-0.263850987,-0.1138282493,0.0100523978,-0.0870554522,-0.2136553824,0.0140519738,0.1094871983,-0.0256686993,-0.1050348207,-0.0054027052,0.2724955976,0.0359710008,-0.0680934265,-0.0653144643,-0.3072283268,0.5054529905,-0.025777841,-0.0580164529,0.0878132731,0.2606489956,0.4709808826,-0.1615628749,-0.6479039192,0.4406293929,0.2577060163,0.1593430787,0.02677406,0.1952493042,0.0265608374,-0.1362664402,-0.2946957946,-0.1092549935,0.1015441939,-0.2585722506,-0.0192036666,-0.1878204048]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2709","title":"Missing documentation for wnut_17 (ner_tags)","comments":"Hi @maxpel, thanks for reporting this issue.\r\n\r\nIndeed, the documentation in the dataset card is not complete. I\u2019m opening a Pull Request to fix it.\r\n\r\nAs the paper explains, there are 6 entity types and we have ordered them alphabetically: `corporation`, `creative-work`, `group`, `location`, `person` and `product`. \r\n\r\nEach of these entity types has 2 possible IOB2 format tags: \r\n- `B-`: to indicate that the token is the beginning of an entity name, and the \r\n- `I-`: to indicate that the token is inside an entity name. \r\n\r\nAdditionally, there is the standalone IOB2 tag \r\n- `O`: that indicates that the token belongs to no named entity. \r\n\r\nIn total there are 13 possible tags, which correspond to the following integer numbers:\r\n\r\n0. `O`\r\n1. `B-corporation`\r\n2. `I-corporation`\r\n3. `B-creative-work`\r\n4. `I-creative-work`\r\n5. `B-group`\r\n6. `I-group`\r\n7. `B-location`\r\n8. `I-location`\r\n9. `B-person`\r\n10. `I-person`\r\n11. `B-product`\r\n12. `I-product`","body":"On the info page of the wnut_17 data set (https:\/\/huggingface.co\/datasets\/wnut_17), the model output of ner-tags is only documented for these 5 cases:\r\n\r\n`ner_tags: a list of classification labels, with possible values including O (0), B-corporation (1), I-corporation (2), B-creative-work (3), I-creative-work (4).`\r\n\r\nI trained a model with the data and it gives me 13 classes:\r\n\r\n```\r\n\"id2label\": {\r\n \"0\": 0,\r\n \"1\": 1,\r\n \"2\": 2,\r\n \"3\": 3,\r\n \"4\": 4,\r\n \"5\": 5,\r\n \"6\": 6,\r\n \"7\": 7,\r\n \"8\": 8,\r\n \"9\": 9,\r\n \"10\": 10,\r\n \"11\": 11,\r\n \"12\": 12\r\n }\r\n\r\n \"label2id\": {\r\n \"0\": 0,\r\n \"1\": 1,\r\n \"10\": 10,\r\n \"11\": 11,\r\n \"12\": 12,\r\n \"2\": 2,\r\n \"3\": 3,\r\n \"4\": 4,\r\n \"5\": 5,\r\n \"6\": 6,\r\n \"7\": 7,\r\n \"8\": 8,\r\n \"9\": 9\r\n }\r\n```\r\nThe paper (https:\/\/www.aclweb.org\/anthology\/W17-4418.pdf) explains those 6 categories, but the ordering does not match:\r\n\r\n```\r\n1. person\r\n2. location (including GPE, facility)\r\n3. corporation\r\n4. product (tangible goods, or well-defined\r\nservices)\r\n5. creative-work (song, movie, book and\r\nso on)\r\n6. group (subsuming music band, sports team,\r\nand non-corporate organisations)\r\n```\r\nI would be very helpful for me, if somebody could clarify the model ouputs and explain the \"B-\" and \"I-\" prefixes to me.\r\n\r\nReally great work with that and the other packages, I couldn't believe that training the model with that data was basically a one-liner!","comment_length":145,"text":"Missing documentation for wnut_17 (ner_tags)\nOn the info page of the wnut_17 data set (https:\/\/huggingface.co\/datasets\/wnut_17), the model output of ner-tags is only documented for these 5 cases:\r\n\r\n`ner_tags: a list of classification labels, with possible values including O (0), B-corporation (1), I-corporation (2), B-creative-work (3), I-creative-work (4).`\r\n\r\nI trained a model with the data and it gives me 13 classes:\r\n\r\n```\r\n\"id2label\": {\r\n \"0\": 0,\r\n \"1\": 1,\r\n \"2\": 2,\r\n \"3\": 3,\r\n \"4\": 4,\r\n \"5\": 5,\r\n \"6\": 6,\r\n \"7\": 7,\r\n \"8\": 8,\r\n \"9\": 9,\r\n \"10\": 10,\r\n \"11\": 11,\r\n \"12\": 12\r\n }\r\n\r\n \"label2id\": {\r\n \"0\": 0,\r\n \"1\": 1,\r\n \"10\": 10,\r\n \"11\": 11,\r\n \"12\": 12,\r\n \"2\": 2,\r\n \"3\": 3,\r\n \"4\": 4,\r\n \"5\": 5,\r\n \"6\": 6,\r\n \"7\": 7,\r\n \"8\": 8,\r\n \"9\": 9\r\n }\r\n```\r\nThe paper (https:\/\/www.aclweb.org\/anthology\/W17-4418.pdf) explains those 6 categories, but the ordering does not match:\r\n\r\n```\r\n1. person\r\n2. location (including GPE, facility)\r\n3. corporation\r\n4. product (tangible goods, or well-defined\r\nservices)\r\n5. creative-work (song, movie, book and\r\nso on)\r\n6. group (subsuming music band, sports team,\r\nand non-corporate organisations)\r\n```\r\nI would be very helpful for me, if somebody could clarify the model ouputs and explain the \"B-\" and \"I-\" prefixes to me.\r\n\r\nReally great work with that and the other packages, I couldn't believe that training the model with that data was basically a one-liner!\nHi @maxpel, thanks for reporting this issue.\r\n\r\nIndeed, the documentation in the dataset card is not complete. I\u2019m opening a Pull Request to fix it.\r\n\r\nAs the paper explains, there are 6 entity types and we have ordered them alphabetically: `corporation`, `creative-work`, `group`, `location`, `person` and `product`. \r\n\r\nEach of these entity types has 2 possible IOB2 format tags: \r\n- `B-`: to indicate that the token is the beginning of an entity name, and the \r\n- `I-`: to indicate that the token is inside an entity name. \r\n\r\nAdditionally, there is the standalone IOB2 tag \r\n- `O`: that indicates that the token belongs to no named entity. \r\n\r\nIn total there are 13 possible tags, which correspond to the following integer numbers:\r\n\r\n0. `O`\r\n1. `B-corporation`\r\n2. `I-corporation`\r\n3. `B-creative-work`\r\n4. `I-creative-work`\r\n5. `B-group`\r\n6. `I-group`\r\n7. `B-location`\r\n8. `I-location`\r\n9. `B-person`\r\n10. `I-person`\r\n11. `B-product`\r\n12. `I-product`","embeddings":[0.2751238048,-0.2970445752,0.0208108258,0.4673909247,-0.05594486,0.0520152785,-0.0117354188,-0.2991321087,-0.2569106817,-0.218673259,0.0983796492,0.3813357353,-0.1071820334,0.2050385922,0.196168229,-0.0721679255,0.1684128791,-0.1887154281,0.3061253726,-0.1100955382,-0.3069913387,0.3437999487,-0.2482835054,0.4954278171,-0.3762372434,-0.0859183073,-0.1869406849,-0.1539729685,-0.404643625,-0.595091939,0.2896220684,0.1556397229,0.0723897889,-0.0579099618,-0.0001202806,-0.2460776716,0.0285192654,-0.1889371872,-0.1224052832,-0.1234155074,-0.0926301107,-0.2060739398,0.2835662961,-0.0094313715,-0.1429478377,-0.1868063211,0.2751163244,0.5176041126,-0.2508018613,0.1197170019,0.0497632809,-0.1621630192,0.2689089179,0.0288132094,0.3519521058,0.0908359066,-0.1232602298,0.3104263544,0.4276799262,-0.0534058437,-0.2571891546,0.4627604485,0.0963761806,-0.2820251882,0.2589055598,0.2285993099,0.3358659446,-0.4192578197,-0.1301675141,0.3817865849,0.2516928613,0.1147544533,-0.2154103816,-0.0157400407,-0.0526016802,-0.6351476908,0.0023943272,0.4393245876,0.0330131762,0.0414246507,-0.0036089518,0.1708147973,-0.2163026631,0.0360443704,-0.0949905366,0.6974461675,-0.1429927051,-0.1101660803,-0.0203673504,-0.0645245537,-0.3136532903,0.0532633699,0.0026731857,0.0485409833,-0.3527156413,0.0821140707,0.1113006175,0.6753001809,0.1034720987,0.1295490712,-0.1083086431,-0.1619324684,-0.1246846244,0.0191322826,0.0126678748,0.0671091676,0.1944057345,0.2011132985,-0.0446371771,0.3469562232,0.0323721841,-0.074082382,-0.1044222638,0.0455814935,-0.2820935845,0.0927808434,0.06808424,-0.0183761325,-0.296031177,0.2254844457,-0.3230934441,-0.0889051035,0.14402318,-0.0752068758,-0.0731510445,0.0357110389,0.1876580566,-0.0504871607,-0.0379349701,-0.482126683,-0.1482041627,0.0386484899,-0.4639268816,0.2523362935,0.1713847667,0.0126344152,0.2062462568,-0.1111709774,-0.4054633081,0.0354569443,-0.0658331662,-0.0313021578,0.0889694542,0.1795018017,-0.4525698423,0.3838886023,0.0929943621,-0.1642804146,-0.2671718597,0.0119260168,-0.4921444952,-0.3080555201,0.0690308735,0.1287891418,-0.1365886331,0.0609575473,0.1897367388,0.3318854868,-0.1153175086,0.1990909874,-0.1007408574,-0.0983121544,-0.4269200861,-0.2644012868,0.0072038556,0.0409698226,-0.3341094851,-0.1623819619,0.137377739,0.1871725172,0.311260432,-0.022741653,-0.0241282787,0.4247220755,-0.1090937629,0.1467428654,0.2499331087,-0.346585989,-0.0087535521,-0.1740071028,-0.0258988328,0.1329857558,-0.0895642564,0.166426003,-0.1744971722,-0.092697978,0.0160262547,-0.0840970501,-0.0957265422,-0.1990121007,0.0502361655,-0.1406760514,0.6332871318,0.1118355095,-0.1572697759,-0.1110158861,-0.373549819,0.1674022377,0.3539595902,-0.0071341656,0.0449136905,0.3623041809,0.087197952,0.3109552562,0.3268584311,0.05157727,-0.1253266633,-0.0166781731,0.1232129708,0.7216901183,0.1610009074,-0.1609766632,-0.059894897,-0.0297032725,-0.0421017967,-0.5228356719,0.1638392508,-0.0354805775,-0.0927694142,-0.0156906638,-0.2345708907,0.3003310561,-0.7553529739,0.1253665686,-0.5494101048,-0.2719165981,-0.3093493283,0.1723164767,0.2604398429,0.16060175,0.3212370872,0.2714565992,0.0924715698,-0.0461571552,0.1488591433,-0.6642785668,0.0911436304,0.6776841879,0.0883733034,-0.5249889493,-0.0968010277,0.0926435068,-0.1376922876,-0.0823116377,-0.1829960197,0.0447826535,-0.0842518657,0.2933146358,0.1141288579,0.0089593967,0.0623177178,0.0631054342,-0.3090403378,0.0669122711,0.1034506932,-0.1360894889,0.0211302061,0.1199387908,-0.0387462042,0.441496253,0.5638064742,-0.1863643229,-0.0692087114,0.2521673441,-0.1829850376,-0.0577115454,-0.2707814872,0.1821918637,0.2422930598,0.0731470287,0.2016145736,-0.1423315406,-0.2664191127,-0.210128665,0.1479953974,0.0713569671,-0.367538482,0.1322910786,-0.0141255315,0.0283213109,-0.100118272,-0.2369363159,-0.0526684187,0.1579366326,-0.2895200849,-0.0486256331,-0.0847701505,-0.0414972827,0.0493923016,-0.1332643181,-0.3768906295,-0.4618844688,0.0604123473,-0.2065777481,0.104614079,0.0916574374,0.0267902054,0.2577775121,0.0078206351,0.4968671501,-0.0199225992,-0.3135142326,-0.418705374,0.0002778663,-0.1932315826,-0.0689466745,-0.2793949246,-0.4715572894,-0.1962037534,0.1891828179,-0.6522645354,0.1992064714,-0.0307924859,0.0423557907,0.3619500101,0.0384898447,-0.2372522801,0.1165786311,0.012894474,-0.0482755043,-0.2377350032,-0.0630051047,-0.1078974307,0.0475881621,-0.2777667046,-0.3997900188,-0.4137221575,-0.0360726044,0.1471912116,-0.0564060211,0.0793977603,0.5000622272,-0.3187740743,0.0542291403,-0.0402783342,0.4291716218,-0.0579381771,-0.0362953581,0.3741432428,-0.1225995943,-0.2875945568,0.2860401571,-0.2776125073,-0.0642166361,-0.2167772502,-0.1659706831,0.0128629273,0.124749057,-0.2548129261,0.1925065517,0.1471295506,0.1604904383,-0.0510471091,0.1826126873,-0.1810187548,-0.135969758,0.1485753506,-0.1738654226,0.1625648886,-0.1120582893,-0.2352698743,0.1533918083,-0.0981056169,0.0958108157,-0.1561302841,-0.1958439052,0.1817343682,0.1204219311,0.1285723746,0.091467604,-0.1207999215,0.0656292215,0.067137748,0.1672758609,0.1578970402,0.4212943017,0.2485266775,0.4521144927,-0.2304109782,-0.1518731117,0.0616700463,0.5299965143,0.3666865528,-0.1315763146,-0.1318488568,-0.0041670729,-0.2918209434,0.0320729353,0.3986603916,0.2485455126,0.1270177066,-0.1801849753,-0.1087058783,-0.3199878037,-0.3875414133,-0.0180185866,-0.0942424014,-0.103228651,0.0734536126,-0.0681673735,0.0887591168,0.2113142312,-0.4619136155,0.3038245738,0.3779914975,0.1060475186,0.0458209068,-0.2179712206,-0.406691432,-0.3208166957,0.2544103563,0.3007873893,-0.2096461058,-0.3157261014,0.6650323272,-0.1623007357,-0.2146286368,0.0231157877,-0.0798238069,-0.1886558682,0.2049299926,-0.3243944943,0.2717305124,0.1194406673,-0.5196067691,0.3327739537,0.2235525399,0.4568699002,0.2948155701,-0.0084767966,0.1713099629,0.0197428968,-0.1269915104,0.0451138467,0.330783397,-0.0008959325,0.3283192515,0.2983852327,-0.5064970255,0.3612937629,-0.2267688215,0.2376775742,0.1006023958,-0.0175062921,0.3284785151,-0.2298382968,0.0035501444,-0.3467365503,0.14709723,0.4200781882,-0.1543995142,-0.0777294859,-0.1305728704,0.3563929796,0.2603345811,-0.3048217297,0.0040715919,0.257317245,-0.2401522696,0.0644239336,0.4713015854,1.0378038883,0.3031315207,0.1885530353,0.1219178885,-0.3590347469,0.6271583438,-0.441362679,0.2939230204,-0.4917212427,-0.2597663999,-0.0798554495,-0.0532474853,-0.3230319321,0.4851151705,-0.01338151,0.2799533904,0.0795682147,0.1533133537,0.0762822852,0.0381997898,0.4392774105,-0.1355609447,0.1449448168,0.015881516,-0.0943060294,0.0506515801,-0.2155335397,-0.0309181176,-0.058517091,0.1497455835,-0.4512604177,-0.0009578937,0.0739385039,-0.1980791092,0.2527949512,0.256483674,0.2726525664,0.3124995232,0.4582827091,-0.1184878051,-0.1011570469,-0.0288455728,-0.3781509995,-0.0451947786,0.0228756592,0.0239693727,0.1892852783,0.1026283801,-0.0556614883,-0.2188269496,0.1123506129,0.2670868635,-0.3358884752,-0.0304058287,-0.0276488438,-0.1019468382,0.3579661548,0.2216020674,-0.0156743918,-0.0714718401,0.0677612722,0.2044016123,-0.3964378536,-0.0368341692,-0.1398335248,0.0013342832,0.0931503177,-0.053961318,0.2500539422,-0.0573750027,-0.0314918384,0.1672793627,-0.1308667213,-0.1319426447,0.0078255627,0.3666982055,-0.4888992906,0.0116822282,0.1734939665,0.0562343597,0.1463616788,0.5145270228,0.127876699,0.2062714547,0.0274106916,-0.2807463706,-0.1172427908,0.3828388155,-0.1685975939,-0.0338449925,-0.4350360334,-0.4497326314,0.1195739955,0.2847349048,-0.2097821683,-0.0833582506,-0.2161239088,0.2336940467,0.127509132,-0.0361236073,0.0623651892,0.3726071417,-0.088638477,-0.0603287928,0.0074930689,-0.060377501,-0.2313486934,0.1688095629,0.2645080984,0.2135521472,-0.0609406605,0.0440987349,0.0316553712,0.0133301402,-0.0719900131,-0.0996455103,0.2574821115,0.5117637515,-0.1896251738,0.4677595496,-0.051345665,-0.0599408969,0.2764768004,0.5056192279,0.4498587251,-0.2050456405,0.1272488087,-0.0422146954,-0.0180224143,0.3335677087,-0.1050388888,-0.1160612777,0.4081159234,-0.2138755471,0.0616924018,-0.2944965065,0.2523398399,-0.0519592538,0.0903062299,-0.1736741811,0.0929380506,0.1015454009,-0.209033668,-0.1755045205,0.7825032473,0.0604797266,-0.0400784798,0.1015179455,0.2406181246,0.1557171345,0.0507115275,-0.1725818515,0.3301270008,0.137245059,0.1113738716,-0.1051011458,-0.046351362,0.2719040513,0.4097805321,0.0255898926,0.1206087396,0.1058517173,0.5922996402,0.1977982372,0.1648868918,0.0666492656,0.2030246705,-0.0697353482,-0.3384664953,0.2676410079,0.2246956527,-0.045502048,-0.2587059736,-0.5041482449,0.1658724993,-0.3136119843,-0.1184867769,-0.1274348944,-0.1070163026,-0.0739996433,-0.6019303203,-0.1222601309,-0.2403332144,0.0025608169,0.0618584156,-0.3346046507,0.2231557369,-0.0902730674,0.1559700668,-0.1328747869,-0.3667320907,0.3920846581,0.305487752,-0.0074456432,0.0691752732,0.0800209865,0.0280432627,0.0133772409,0.4532531798,0.1409292072,0.2674154341,-0.0482878275,-0.1304044425,0.0924437493,-0.0631448701,0.1392107457,0.1444191784,0.1311929524,-0.2176302373,0.1192922071,-0.047057841,0.0623308048,0.1783405989,0.1782845557,0.1870180815,0.2196922749,-0.1662826985,0.0244083814,-0.2889586389,0.058966659,0.1931678653,-0.1010687724,0.0312129352,0.5167577267,0.5962797403,0.163616091,-0.0893286839,0.0574168041,0.1213992909,0.3225640655,0.5040453076,-0.0957842022,-0.5049678087,-0.4385409653,-0.2283805013,0.1354018748,-0.153917551,-0.154836148,0.0497736484,0.0315377936,0.1469531953,-0.0454765968,-0.2531865835,0.103418678,0.1706099361,0.4527874291,-0.2479971796,-0.2641298175,0.1756777912,0.1481628269,-0.3362433314,0.2616692483,-0.2128396481,-0.3010275066,-0.0757124051,0.1897960454,0.0888680369,-0.0190403257,0.601087451,-0.045583792,0.3202244937,0.2428416014,-0.0212925151,-0.2903785408,-0.2750293612,0.2024097294,-0.0744128898,0.4421254098,-0.3846025765,0.2418860793,0.0347098894,0.0626789629,-0.2775330842,-0.278660804,0.0277651846,-0.21431306,-0.3709549904,0.279589653,-0.0979459956,0.0287372693,0.4505161345,0.1107024476,0.0913536176,-0.1294455081,-0.1693636328,-0.2396493852,0.212036863,-0.3755996823,-0.0540664308,-0.0467674881,0.1419933438,0.535436511,0.4785678089,-0.3146364391,-0.205306977,0.3774160445,-0.1149438918,-0.2709685266,0.1097267568,0.2885722816,-0.1728775948,-0.0676889494,-0.4764373302,0.2064547688,-0.1344137788,-0.238926366,-0.3001998067]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2708","title":"QASC: incomplete training set ","comments":"Hi @danyaljj, thanks for reporting.\r\n\r\nUnfortunately, I have not been able to reproduce your problem. My train split has 8134 examples:\r\n```ipython\r\nIn [10]: ds[\"train\"]\r\nOut[10]:\r\nDataset({\r\n features: ['id', 'question', 'choices', 'answerKey', 'fact1', 'fact2', 'combinedfact', 'formatted_question'],\r\n num_rows: 8134\r\n})\r\n\r\nIn [11]: ds[\"train\"].shape\r\nOut[11]: (8134, 8)\r\n```\r\nand the content of the last 5 examples is:\r\n```ipython\r\nIn [12]: for i in range(8129, 8134):\r\n ...: print(json.dumps(ds[\"train\"][i]))\r\n ...:\r\n{\"id\": \"3KAKFY4PGU1LGXM77JAK2700NGCI3X\", \"question\": \"Chitin can be used for protection by whom?\", \"choices\": {\"text\": [\"Fungi\", \"People\", \"Man\", \"Fish\", \"trees\", \"Dogs\", \"animal\", \"Birds\"], \"label\": [\"A\", \"B\",\r\n \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"D\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Fish scales are also composed of chitin.\", \"combinedfact\": \"Chitin can be used for prote\r\nction by fish.\", \"formatted_question\": \"Chitin can be used for protection by whom? (A) Fungi (B) People (C) Man (D) Fish (E) trees (F) Dogs (G) animal (H) Birds\"}\r\n{\"id\": \"336YQZE83VDAQVZ26HW59X51JZ9M5M\", \"question\": \"Which type of animal uses plates for protection?\", \"choices\": {\"text\": [\"squids\", \"reptiles\", \"sea urchins\", \"fish\", \"amphibians\", \"Frogs\", \"mammals\", \"salm\r\non\"], \"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"B\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Reptiles have scales or plates.\", \"combinedfact\": \"Reptiles use\r\n their plates for protection.\", \"formatted_question\": \"Which type of animal uses plates for protection? (A) squids (B) reptiles (C) sea urchins (D) fish (E) amphibians (F) Frogs (G) mammals (H) salmon\"}\r\n{\"id\": \"3WZ36BJEV3FGS66VGOOUYX0LN8GTBU\", \"question\": \"What are used for protection by fish?\", \"choices\": {\"text\": [\"scales\", \"fins\", \"streams.\", \"coral\", \"gills\", \"Collagen\", \"mussels\", \"whiskers\"], \"label\": [\"\r\nA\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"A\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Fish are backboned aquatic animals.\", \"combinedfact\": \"scales are used for prote\r\nction by fish \", \"formatted_question\": \"What are used for protection by fish? (A) scales (B) fins (C) streams. (D) coral (E) gills (F) Collagen (G) mussels (H) whiskers\"}\r\n{\"id\": \"3Z2R0DQ0JHDKFAO2706OYIXGNA4E28\", \"question\": \"What are pangolins covered in?\", \"choices\": {\"text\": [\"tunicates\", \"Echinoids\", \"shells\", \"exoskeleton\", \"blastoids\", \"barrel-shaped\", \"protection\", \"white\"\r\n], \"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"G\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Pangolins have an elongate and tapering body covered above with ov\r\nerlapping scales.\", \"combinedfact\": \"Pangolins are covered in overlapping protection.\", \"formatted_question\": \"What are pangolins covered in? (A) tunicates (B) Echinoids (C) shells (D) exoskeleton (E) blastoids\r\n (F) barrel-shaped (G) protection (H) white\"}\r\n{\"id\": \"3PMBY0YE272GIWPNWIF8IH5RBHVC9S\", \"question\": \"What are covered with protection?\", \"choices\": {\"text\": [\"apples\", \"trees\", \"coral\", \"clams\", \"roses\", \"wings\", \"hats\", \"fish\"], \"label\": [\"A\", \"B\", \"C\", \"D\r\n\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"H\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Fish are covered with scales.\", \"combinedfact\": \"Fish are covered with protection\", \"formatted_q\r\nuestion\": \"What are covered with protection? (A) apples (B) trees (C) coral (D) clams (E) roses (F) wings (G) hats (H) fish\"}\r\n```\r\n\r\nCould you please load again your dataset and print its shape, like this:\r\n```python\r\nds = load_dataset(\"qasc\", split=\"train)\r\nprint(ds.shape)\r\n```\r\nand confirm which is your output?","body":"## Describe the bug\r\nThe training instances are not loaded properly. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"qasc\", script_version='1.10.2')\r\n \r\ndef load_instances(split): \r\n instances = dataset[split]\r\n print(f\"split: {split} - size: {len(instances)}\")\r\n for x in instances:\r\n print(json.dumps(x))\r\n\r\n\r\nload_instances('test')\r\nload_instances('validation')\r\nload_instances('train')\r\n```\r\n\r\n## results\r\nFor test and validation, we can see the examples in the output (which is good!): \r\n```\r\nsplit: test - size: 920\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Anthax\", \"under water\", \"uterus\", \"wombs\", \"two\", \"moles\", \"live\", \"embryo\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"What type of birth do therian mammals have? (A) Anthax (B) under water (C) uterus (D) wombs (E) two (F) moles (G) live (H) embryo\", \"id\": \"3C44YUNSI1OBFBB8D36GODNOZN9DPA\", \"question\": \"What type of birth do therian mammals have?\"}\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Corvidae\", \"arthropods\", \"birds\", \"backbones\", \"keratin\", \"Jurassic\", \"front paws\", \"Parakeets.\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"By what time had mouse-sized viviparous mammals evolved? (A) Corvidae (B) arthropods (C) birds (D) backbones (E) keratin (F) Jurassic (G) front paws (H) Parakeets.\", \"id\": \"3B1NLC6UGZVERVLZFT7OUYQLD1SGPZ\", \"question\": \"By what time had mouse-sized viviparous mammals evolved?\"}\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Reduced friction\", \"causes infection\", \"vital to a good life\", \"prevents water loss\", \"camouflage from consumers\", \"Protection against predators\", \"spur the growth of the plant\", \"a smooth surface\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"What does a plant's skin do? (A) Reduced friction (B) causes infection (C) vital to a good life (D) prevents water loss (E) camouflage from consumers (F) Protection against predators (G) spur the growth of the plant (H) a smooth surface\", \"id\": \"3QRYMNZ7FYGITFVSJET3PS0F4S0NT9\", \"question\": \"What does a plant's skin do?\"}\r\n...\r\n```\r\nHowever, only a few instances are loaded for the training split, which is not correct. \r\n\r\n## Environment info\r\n- `datasets` version: '1.10.2' \r\n- Platform: MaxOS \r\n- Python version:3.7\r\n- PyArrow version: 3.0.0\r\n","comment_length":496,"text":"QASC: incomplete training set \n## Describe the bug\r\nThe training instances are not loaded properly. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"qasc\", script_version='1.10.2')\r\n \r\ndef load_instances(split): \r\n instances = dataset[split]\r\n print(f\"split: {split} - size: {len(instances)}\")\r\n for x in instances:\r\n print(json.dumps(x))\r\n\r\n\r\nload_instances('test')\r\nload_instances('validation')\r\nload_instances('train')\r\n```\r\n\r\n## results\r\nFor test and validation, we can see the examples in the output (which is good!): \r\n```\r\nsplit: test - size: 920\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Anthax\", \"under water\", \"uterus\", \"wombs\", \"two\", \"moles\", \"live\", \"embryo\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"What type of birth do therian mammals have? (A) Anthax (B) under water (C) uterus (D) wombs (E) two (F) moles (G) live (H) embryo\", \"id\": \"3C44YUNSI1OBFBB8D36GODNOZN9DPA\", \"question\": \"What type of birth do therian mammals have?\"}\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Corvidae\", \"arthropods\", \"birds\", \"backbones\", \"keratin\", \"Jurassic\", \"front paws\", \"Parakeets.\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"By what time had mouse-sized viviparous mammals evolved? (A) Corvidae (B) arthropods (C) birds (D) backbones (E) keratin (F) Jurassic (G) front paws (H) Parakeets.\", \"id\": \"3B1NLC6UGZVERVLZFT7OUYQLD1SGPZ\", \"question\": \"By what time had mouse-sized viviparous mammals evolved?\"}\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Reduced friction\", \"causes infection\", \"vital to a good life\", \"prevents water loss\", \"camouflage from consumers\", \"Protection against predators\", \"spur the growth of the plant\", \"a smooth surface\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"What does a plant's skin do? (A) Reduced friction (B) causes infection (C) vital to a good life (D) prevents water loss (E) camouflage from consumers (F) Protection against predators (G) spur the growth of the plant (H) a smooth surface\", \"id\": \"3QRYMNZ7FYGITFVSJET3PS0F4S0NT9\", \"question\": \"What does a plant's skin do?\"}\r\n...\r\n```\r\nHowever, only a few instances are loaded for the training split, which is not correct. \r\n\r\n## Environment info\r\n- `datasets` version: '1.10.2' \r\n- Platform: MaxOS \r\n- Python version:3.7\r\n- PyArrow version: 3.0.0\r\n\nHi @danyaljj, thanks for reporting.\r\n\r\nUnfortunately, I have not been able to reproduce your problem. My train split has 8134 examples:\r\n```ipython\r\nIn [10]: ds[\"train\"]\r\nOut[10]:\r\nDataset({\r\n features: ['id', 'question', 'choices', 'answerKey', 'fact1', 'fact2', 'combinedfact', 'formatted_question'],\r\n num_rows: 8134\r\n})\r\n\r\nIn [11]: ds[\"train\"].shape\r\nOut[11]: (8134, 8)\r\n```\r\nand the content of the last 5 examples is:\r\n```ipython\r\nIn [12]: for i in range(8129, 8134):\r\n ...: print(json.dumps(ds[\"train\"][i]))\r\n ...:\r\n{\"id\": \"3KAKFY4PGU1LGXM77JAK2700NGCI3X\", \"question\": \"Chitin can be used for protection by whom?\", \"choices\": {\"text\": [\"Fungi\", \"People\", \"Man\", \"Fish\", \"trees\", \"Dogs\", \"animal\", \"Birds\"], \"label\": [\"A\", \"B\",\r\n \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"D\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Fish scales are also composed of chitin.\", \"combinedfact\": \"Chitin can be used for prote\r\nction by fish.\", \"formatted_question\": \"Chitin can be used for protection by whom? (A) Fungi (B) People (C) Man (D) Fish (E) trees (F) Dogs (G) animal (H) Birds\"}\r\n{\"id\": \"336YQZE83VDAQVZ26HW59X51JZ9M5M\", \"question\": \"Which type of animal uses plates for protection?\", \"choices\": {\"text\": [\"squids\", \"reptiles\", \"sea urchins\", \"fish\", \"amphibians\", \"Frogs\", \"mammals\", \"salm\r\non\"], \"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"B\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Reptiles have scales or plates.\", \"combinedfact\": \"Reptiles use\r\n their plates for protection.\", \"formatted_question\": \"Which type of animal uses plates for protection? (A) squids (B) reptiles (C) sea urchins (D) fish (E) amphibians (F) Frogs (G) mammals (H) salmon\"}\r\n{\"id\": \"3WZ36BJEV3FGS66VGOOUYX0LN8GTBU\", \"question\": \"What are used for protection by fish?\", \"choices\": {\"text\": [\"scales\", \"fins\", \"streams.\", \"coral\", \"gills\", \"Collagen\", \"mussels\", \"whiskers\"], \"label\": [\"\r\nA\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"A\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Fish are backboned aquatic animals.\", \"combinedfact\": \"scales are used for prote\r\nction by fish \", \"formatted_question\": \"What are used for protection by fish? (A) scales (B) fins (C) streams. (D) coral (E) gills (F) Collagen (G) mussels (H) whiskers\"}\r\n{\"id\": \"3Z2R0DQ0JHDKFAO2706OYIXGNA4E28\", \"question\": \"What are pangolins covered in?\", \"choices\": {\"text\": [\"tunicates\", \"Echinoids\", \"shells\", \"exoskeleton\", \"blastoids\", \"barrel-shaped\", \"protection\", \"white\"\r\n], \"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"G\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Pangolins have an elongate and tapering body covered above with ov\r\nerlapping scales.\", \"combinedfact\": \"Pangolins are covered in overlapping protection.\", \"formatted_question\": \"What are pangolins covered in? (A) tunicates (B) Echinoids (C) shells (D) exoskeleton (E) blastoids\r\n (F) barrel-shaped (G) protection (H) white\"}\r\n{\"id\": \"3PMBY0YE272GIWPNWIF8IH5RBHVC9S\", \"question\": \"What are covered with protection?\", \"choices\": {\"text\": [\"apples\", \"trees\", \"coral\", \"clams\", \"roses\", \"wings\", \"hats\", \"fish\"], \"label\": [\"A\", \"B\", \"C\", \"D\r\n\", \"E\", \"F\", \"G\", \"H\"]}, \"answerKey\": \"H\", \"fact1\": \"scales are used for protection by scaled animals\", \"fact2\": \"Fish are covered with scales.\", \"combinedfact\": \"Fish are covered with protection\", \"formatted_q\r\nuestion\": \"What are covered with protection? (A) apples (B) trees (C) coral (D) clams (E) roses (F) wings (G) hats (H) fish\"}\r\n```\r\n\r\nCould you please load again your dataset and print its shape, like this:\r\n```python\r\nds = load_dataset(\"qasc\", split=\"train)\r\nprint(ds.shape)\r\n```\r\nand confirm which is your output?","embeddings":[-0.2450181693,-0.3615007997,-0.1671689451,0.3723612428,0.1236955449,0.1451464146,0.0164389201,0.4380207956,-0.1213213205,0.0934934169,0.1632602215,0.124971658,0.0430466495,0.2471365929,-0.0943753049,-0.1647269875,-0.0069864769,0.2966333032,-0.2305759043,-0.0744921416,-0.2418487519,0.3023805916,-0.2500701845,-0.0736244619,-0.1649780124,-0.2872940302,-0.0491756722,0.0133106634,0.0458690599,-0.320064187,-0.0480261333,-0.3247571588,-0.1014455259,0.4697218835,-0.0001062213,0.0782134309,0.1270972788,-0.1408859044,-0.3300843835,0.1074304581,-0.2933759093,-0.1728284657,0.0326805972,-0.0973909572,-0.2557199597,-0.2729076743,-0.0496537611,-0.0854506865,0.5871784687,0.289850086,0.2529067695,0.3633326888,0.033648029,-0.0832497925,-0.221334666,-0.504969418,-0.0405405499,0.0328143686,0.2264511585,0.0457118303,0.2468177974,0.2307891995,0.2760746777,-0.100784339,0.1479250938,-0.2171174735,0.1146207899,-0.1962149888,0.026363261,0.1972755641,0.3719688356,0.1936610043,-0.3621879816,-0.1538570821,-0.0319558084,-0.4144759178,-0.0372161455,0.2690791488,-0.0616779961,0.20732674,-0.1861333698,0.2345629483,-0.1997883022,0.1442641318,-0.1462110281,0.2244074494,-0.0654697195,-0.0954147652,-0.2495789528,0.1812457442,-0.0705581754,0.1260930896,-0.0084403893,0.1862768978,-0.5876232982,-0.1518306881,0.1735303998,-0.2187470049,0.05366778,-0.0276974067,0.4717944562,-0.1412032098,0.0702639893,0.0977402925,0.2745475769,0.510950923,0.201284498,0.1008572876,-0.1483419985,0.2257253379,-0.1809088588,0.1172315031,-0.1104004085,-0.1220034808,0.1084840372,0.2392062545,-0.1832301319,-0.1661884338,-0.4356065989,0.4669902325,-0.4517908394,0.1110789999,-0.2033956498,0.3239640892,-0.2341742665,0.2661927342,0.0968197584,0.2421134561,-0.4497469068,-0.1726887971,-0.3069637716,-0.0053717731,-0.1520239562,-0.1448992342,0.0667963028,-0.0198877677,0.6864049435,-0.1905841082,0.2406973541,-0.1494981796,0.1166925207,-0.1677237153,-0.1573780477,0.2168618292,0.0933346674,-0.1515265107,0.0986844525,0.1273582578,0.073553957,-0.0123291211,-0.1933779418,-0.0778044537,0.0731493235,0.3307126164,0.0673327073,0.1481941789,-0.0800899193,0.3283025026,0.3131912947,0.2075050324,-0.1654668152,-0.1291126758,-0.0672498494,-0.0664217025,0.2284736335,0.1068682075,-0.3879636526,-0.2831277251,0.2111899704,0.1523139328,-0.1277307868,0.1663789004,-0.0637842715,0.4463846385,-0.0574760288,0.1228245422,-0.1606709957,-0.0278364383,-0.0758509561,-0.0808564126,-0.0359562635,0.0637387708,0.0771537423,-0.3781818449,0.3808516264,0.1721238941,0.2533078194,0.2878333032,-0.0454363003,-0.0531011745,-0.3692813516,-0.333049655,-0.1322299391,0.0303953551,0.1376107484,-0.0659362376,-0.0196515061,-0.2123082876,0.4386965334,0.0753439441,-0.1268944889,0.2475073934,0.2016558796,-0.175553605,0.134940505,-0.0882637501,-0.4648680985,0.3856156766,-0.2136990875,0.1805250943,0.1455844641,0.0752157122,-0.5313087106,-0.0354604274,-0.0576039143,-0.3980095685,0.2089185566,-0.04827407,0.1521581262,-0.0416214764,-0.1192983985,0.0717149377,-0.2015973032,-0.0051512355,-0.2455405742,0.2948980629,-0.1022188514,-0.0338154435,0.2529664934,0.1713653058,0.0838688239,-0.2252237201,-0.0535434745,0.3246357739,0.1384629458,-0.0993399248,-0.1715636551,-0.1226139367,0.1180634499,-0.131923154,-0.2020104975,0.2057533711,0.1158741117,-0.0233052354,-0.1940400749,0.3919537663,0.1703977436,0.2448907644,0.2319546491,0.0529230796,0.1448597759,-0.0916807204,0.0975734591,-0.0638853163,0.189599961,-0.0497281142,-0.0844068453,-0.0868861899,-0.2769928873,0.1296674013,0.1336871833,0.0348882228,-0.0792626292,-0.162670821,-0.1537512541,0.1644689143,0.1590701044,0.1174459532,0.1510914415,0.30799824,-0.2536084652,-0.199834615,0.0356864035,0.0004941155,0.1626493484,0.0426238924,0.2464949638,0.2847855985,0.1936950982,-0.0943688676,-0.1065967083,0.1231058016,0.1845837831,0.2153210491,-0.3730052114,0.1432093829,-0.1024540812,0.2950365245,-0.1279529482,0.235516116,0.159334287,-0.2550415099,0.2276655138,0.0367487222,-0.053009022,0.264975369,-0.254773885,0.2745153606,0.1846229136,-0.1531972736,0.1978167146,0.1969082803,-0.3121934831,0.2985317111,-0.0876592323,0.1244520769,0.2559636235,-0.0741104484,-0.298353672,0.1935950369,-0.2393571287,0.1780943573,-0.1836472899,0.483614713,-0.141664952,0.1909188926,-0.3466756642,-0.3810741007,0.1948749274,-0.0581267588,-0.2577676773,0.1504852474,0.0410655998,-0.2013731003,-0.1593966931,-0.7575306296,0.0354381427,-0.1635408401,0.1972049624,0.1563578099,0.2769849002,0.0308583379,0.2668566406,-0.0354165621,-0.1260354668,-0.1651287824,-0.1709091961,-0.194642216,0.3205131292,-0.1530383378,-0.3015633821,-0.1450711936,-0.1393867731,0.5834782124,-0.0346679166,-0.3780317307,-0.2207747847,-0.0099999337,0.1090353206,-0.041204527,-0.1890207827,0.3572488725,-0.0990521386,-0.202939555,-0.0918208286,-0.0775542781,0.308878094,0.1910594404,0.258572638,-0.4090474844,0.2854526043,-0.0080062728,1.0584617853,0.3723424077,0.0747849196,0.0401264355,-0.0044042636,-0.1289398372,-0.1965896934,-0.1487231553,0.1607521623,-0.0593115315,-0.2840839028,0.2656474113,0.1344437301,-0.0197183974,-0.1355739981,0.2519261539,-0.6226943731,-0.1467548907,-0.0441396087,-0.067962192,0.5081368685,0.026144499,0.2575253248,0.1802202463,0.0982106775,0.0189101733,-0.0221583322,0.1799109429,-0.0480811037,-0.384996593,-0.0720152259,-0.06307327,0.1918585002,-0.1368911713,0.2850539684,-0.033641085,-0.0315195583,-0.0155024678,0.019306384,0.491650492,0.1030349955,0.0759638697,0.2137275189,0.0983606428,-0.0879001468,-0.155747667,-0.1249988601,-0.3020196259,-0.1005191281,0.3204067945,-0.3774962127,0.0301839281,-0.0500864834,0.0213078782,-0.3617132008,0.0159993488,-0.3647143245,0.0236450788,-0.1130777672,0.1736128032,0.2534516156,0.239273876,-0.0875733495,-0.0866843015,0.0384343378,-0.3197672069,0.2297399342,0.3085214198,0.3850121796,0.0758537948,0.5038484335,0.1059484705,-0.260270685,0.2349551767,0.5228100419,-0.1499752849,-0.3038350642,0.1117307395,-0.1718555242,0.4033382237,0.2430368215,-0.1971922219,0.0012441196,-0.0755827427,0.2412461638,0.2151294053,0.082520023,0.0790469274,-0.0239105821,0.2027700543,-0.3584736586,0.0560042709,-0.1665772796,-0.0322501585,-0.0816431642,0.1298967302,-0.1747607291,0.333870858,0.1127267703,0.6330314875,0.130593732,-0.0383033939,0.4832280874,0.0490094945,0.122882776,-0.3597963154,0.2030264735,-0.2173835784,-0.1108087525,0.0887389481,-0.0648304,0.1883213073,0.3287271559,-0.1439502239,0.4234223962,-0.448482275,-0.0406923443,0.0738431811,-0.0510886908,0.03146483,-0.0251645409,-0.2900235653,0.2359438986,-0.0015879149,-0.0406085663,-0.1033929139,-0.0731634647,-0.1227568015,-0.1274030656,-0.0092289634,0.2474857122,-0.311883688,-0.1006114334,0.3116766512,-0.1610029638,0.0762248859,0.3196717501,-0.120197773,0.2709676325,-0.0974705294,0.2468037605,-0.046355471,0.057972163,0.0048045497,-0.0912673771,0.4631349444,-0.1395934969,-0.0712312311,0.4149790406,-0.2818175852,0.0176113918,0.0510923229,0.0648634806,0.0080351215,-0.3585578799,-0.1104812324,0.1376360208,0.0029345215,-0.3381873369,0.2300059199,0.0572285652,-0.189274773,0.1670876443,0.0865444615,-0.1463118047,0.0427084751,0.1361495554,0.3570473194,-0.1305964291,0.7882820964,0.1528767347,-0.0398471579,-0.3938729167,-0.0726180077,0.2945176363,-0.435719192,0.3019652069,-0.0136567447,-0.0246525276,-0.0521612875,0.4134021997,0.1887339801,0.0399365388,0.2184929401,-0.4894676507,-0.1931098551,0.2399053276,-0.0145094646,0.0302345157,-0.1322632134,0.290312022,-0.1278672963,0.2097464204,-0.4214288592,-0.1433736831,-0.3985895216,0.1878068149,-0.5181497335,-0.1457366645,0.2272179425,-0.1095385626,0.2473854423,0.0375558138,-0.256757766,-0.3200960159,-0.2861353457,0.0617648736,0.1927329302,0.1127406061,0.288939625,0.041668918,-0.2834427953,0.0178245381,0.1013175398,-0.1948021054,0.0467673764,0.0871966854,0.3864264786,0.0219777226,-0.19198744,-0.1358885318,-0.0882558972,0.0904601142,0.0601679422,-0.2023030221,0.0523505248,-0.012818356,-0.1507111937,0.2215189189,-0.1231176332,0.2737950087,0.2905271351,-0.1926757544,-0.0184266027,-0.1168123558,0.1618770808,0.1685914248,-0.1445003748,0.1315016448,0.0135789206,0.3528312147,-0.3844121397,0.0049668453,-0.3526668847,-0.0184044205,0.0893212184,0.1673433781,0.0437684692,-0.3547252417,-0.0860331729,0.2811055779,-0.2622293532,-0.1081992239,0.2802663743,0.5731699467,0.0275648199,-0.0002921917,0.1819287091,-0.3897887468,-0.0526641682,-0.0991594121,0.07343328,0.1352884918,0.2290858924,0.020084573,-0.0661902726,-0.1742319316,0.011692862,0.0469578169,-0.0897581652,0.3786462843,-0.0138229989,0.2987894714,-0.0638646409,-0.1497026831,-0.3668247461,-0.168053031,-0.0015945042,0.0645888075,0.0085983807,-0.1513916403,-0.1663929671,-0.08183337,0.0792207792,-0.3283424973,0.154255718,-0.1284558624,-0.4417531788,-0.2323850244,-0.3880307078,0.4857454002,0.258284539,-0.2795485556,0.2843214273,0.1506462991,-0.1825749129,-0.1176908165,0.4338412881,0.128068015,0.0845422074,-0.0463073887,-0.0871170908,0.0822568536,-0.1040437296,0.1117975786,0.2822864652,0.1642481089,-0.0767761096,0.2574383914,0.1753002852,-0.2762508988,0.1733468771,0.081053704,0.2326996773,-0.2533421516,0.1304880679,-0.1144494787,-0.0447532646,-0.0553241372,0.1883161962,-0.401135236,0.0053416523,0.1108732074,0.1102420613,0.1241546795,-0.0843065232,0.1275138706,-0.3049993217,0.5124530792,0.0103119044,0.0229149908,-0.4916670918,-0.064716503,-0.2829300165,-0.0501294881,-0.2130349278,0.2104375511,0.0346973427,0.209083125,0.1262042075,0.1017698646,-0.0141174598,0.0250880383,-0.2571706474,0.3189039826,-0.7061988115,-0.135068506,-0.2019830495,0.3497312963,0.1605273336,-0.3106667995,0.0155651346,-0.1239759475,0.1142854393,0.1948511302,-0.1609454155,0.3158736825,-0.0857972428,0.2146507949,0.1274192929,0.1434754282,-0.0719684511,-0.1633684337,-0.1401197612,-0.3524002433,-0.2526800036,0.2253606617,0.2688834667,0.4694870114,0.1029353961,-0.1061380208,0.0337461829,0.0445147716,0.2115573585,-0.4895721972,-0.3758788407,0.2332729399,0.0127202421,0.2232289761,0.1194653288,0.2636312544,-0.1482480466,-0.2476286143,-0.3535003662,-0.4347568154,0.4308348596,-0.07013385,-0.0934192091,-0.1445526779,0.2189143002,0.0158601161,-0.0631370321,-0.6744795442,-0.0659779683,0.409499377,-0.1132725105,-0.2585168481,0.1806693077,0.1244951263,-0.1271043271,-0.0564316101,0.2745738626,0.0278273188,-0.2310827225,0.0818093717,-0.0966653824]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2708","title":"QASC: incomplete training set ","comments":"Hmm .... it must have been a mistake on my side. Sorry for the hassle! ","body":"## Describe the bug\r\nThe training instances are not loaded properly. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"qasc\", script_version='1.10.2')\r\n \r\ndef load_instances(split): \r\n instances = dataset[split]\r\n print(f\"split: {split} - size: {len(instances)}\")\r\n for x in instances:\r\n print(json.dumps(x))\r\n\r\n\r\nload_instances('test')\r\nload_instances('validation')\r\nload_instances('train')\r\n```\r\n\r\n## results\r\nFor test and validation, we can see the examples in the output (which is good!): \r\n```\r\nsplit: test - size: 920\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Anthax\", \"under water\", \"uterus\", \"wombs\", \"two\", \"moles\", \"live\", \"embryo\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"What type of birth do therian mammals have? (A) Anthax (B) under water (C) uterus (D) wombs (E) two (F) moles (G) live (H) embryo\", \"id\": \"3C44YUNSI1OBFBB8D36GODNOZN9DPA\", \"question\": \"What type of birth do therian mammals have?\"}\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Corvidae\", \"arthropods\", \"birds\", \"backbones\", \"keratin\", \"Jurassic\", \"front paws\", \"Parakeets.\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"By what time had mouse-sized viviparous mammals evolved? (A) Corvidae (B) arthropods (C) birds (D) backbones (E) keratin (F) Jurassic (G) front paws (H) Parakeets.\", \"id\": \"3B1NLC6UGZVERVLZFT7OUYQLD1SGPZ\", \"question\": \"By what time had mouse-sized viviparous mammals evolved?\"}\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Reduced friction\", \"causes infection\", \"vital to a good life\", \"prevents water loss\", \"camouflage from consumers\", \"Protection against predators\", \"spur the growth of the plant\", \"a smooth surface\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"What does a plant's skin do? (A) Reduced friction (B) causes infection (C) vital to a good life (D) prevents water loss (E) camouflage from consumers (F) Protection against predators (G) spur the growth of the plant (H) a smooth surface\", \"id\": \"3QRYMNZ7FYGITFVSJET3PS0F4S0NT9\", \"question\": \"What does a plant's skin do?\"}\r\n...\r\n```\r\nHowever, only a few instances are loaded for the training split, which is not correct. \r\n\r\n## Environment info\r\n- `datasets` version: '1.10.2' \r\n- Platform: MaxOS \r\n- Python version:3.7\r\n- PyArrow version: 3.0.0\r\n","comment_length":15,"text":"QASC: incomplete training set \n## Describe the bug\r\nThe training instances are not loaded properly. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"qasc\", script_version='1.10.2')\r\n \r\ndef load_instances(split): \r\n instances = dataset[split]\r\n print(f\"split: {split} - size: {len(instances)}\")\r\n for x in instances:\r\n print(json.dumps(x))\r\n\r\n\r\nload_instances('test')\r\nload_instances('validation')\r\nload_instances('train')\r\n```\r\n\r\n## results\r\nFor test and validation, we can see the examples in the output (which is good!): \r\n```\r\nsplit: test - size: 920\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Anthax\", \"under water\", \"uterus\", \"wombs\", \"two\", \"moles\", \"live\", \"embryo\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"What type of birth do therian mammals have? (A) Anthax (B) under water (C) uterus (D) wombs (E) two (F) moles (G) live (H) embryo\", \"id\": \"3C44YUNSI1OBFBB8D36GODNOZN9DPA\", \"question\": \"What type of birth do therian mammals have?\"}\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Corvidae\", \"arthropods\", \"birds\", \"backbones\", \"keratin\", \"Jurassic\", \"front paws\", \"Parakeets.\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"By what time had mouse-sized viviparous mammals evolved? (A) Corvidae (B) arthropods (C) birds (D) backbones (E) keratin (F) Jurassic (G) front paws (H) Parakeets.\", \"id\": \"3B1NLC6UGZVERVLZFT7OUYQLD1SGPZ\", \"question\": \"By what time had mouse-sized viviparous mammals evolved?\"}\r\n{\"answerKey\": \"\", \"choices\": {\"label\": [\"A\", \"B\", \"C\", \"D\", \"E\", \"F\", \"G\", \"H\"], \"text\": [\"Reduced friction\", \"causes infection\", \"vital to a good life\", \"prevents water loss\", \"camouflage from consumers\", \"Protection against predators\", \"spur the growth of the plant\", \"a smooth surface\"]}, \"combinedfact\": \"\", \"fact1\": \"\", \"fact2\": \"\", \"formatted_question\": \"What does a plant's skin do? (A) Reduced friction (B) causes infection (C) vital to a good life (D) prevents water loss (E) camouflage from consumers (F) Protection against predators (G) spur the growth of the plant (H) a smooth surface\", \"id\": \"3QRYMNZ7FYGITFVSJET3PS0F4S0NT9\", \"question\": \"What does a plant's skin do?\"}\r\n...\r\n```\r\nHowever, only a few instances are loaded for the training split, which is not correct. \r\n\r\n## Environment info\r\n- `datasets` version: '1.10.2' \r\n- Platform: MaxOS \r\n- Python version:3.7\r\n- PyArrow version: 3.0.0\r\n\nHmm .... it must have been a mistake on my side. Sorry for the hassle! ","embeddings":[-0.2450181693,-0.3615007997,-0.1671689451,0.3723612428,0.1236955449,0.1451464146,0.0164389201,0.4380207956,-0.1213213205,0.0934934169,0.1632602215,0.124971658,0.0430466495,0.2471365929,-0.0943753049,-0.1647269875,-0.0069864769,0.2966333032,-0.2305759043,-0.0744921416,-0.2418487519,0.3023805916,-0.2500701845,-0.0736244619,-0.1649780124,-0.2872940302,-0.0491756722,0.0133106634,0.0458690599,-0.320064187,-0.0480261333,-0.3247571588,-0.1014455259,0.4697218835,-0.0001062213,0.0782134309,0.1270972788,-0.1408859044,-0.3300843835,0.1074304581,-0.2933759093,-0.1728284657,0.0326805972,-0.0973909572,-0.2557199597,-0.2729076743,-0.0496537611,-0.0854506865,0.5871784687,0.289850086,0.2529067695,0.3633326888,0.033648029,-0.0832497925,-0.221334666,-0.504969418,-0.0405405499,0.0328143686,0.2264511585,0.0457118303,0.2468177974,0.2307891995,0.2760746777,-0.100784339,0.1479250938,-0.2171174735,0.1146207899,-0.1962149888,0.026363261,0.1972755641,0.3719688356,0.1936610043,-0.3621879816,-0.1538570821,-0.0319558084,-0.4144759178,-0.0372161455,0.2690791488,-0.0616779961,0.20732674,-0.1861333698,0.2345629483,-0.1997883022,0.1442641318,-0.1462110281,0.2244074494,-0.0654697195,-0.0954147652,-0.2495789528,0.1812457442,-0.0705581754,0.1260930896,-0.0084403893,0.1862768978,-0.5876232982,-0.1518306881,0.1735303998,-0.2187470049,0.05366778,-0.0276974067,0.4717944562,-0.1412032098,0.0702639893,0.0977402925,0.2745475769,0.510950923,0.201284498,0.1008572876,-0.1483419985,0.2257253379,-0.1809088588,0.1172315031,-0.1104004085,-0.1220034808,0.1084840372,0.2392062545,-0.1832301319,-0.1661884338,-0.4356065989,0.4669902325,-0.4517908394,0.1110789999,-0.2033956498,0.3239640892,-0.2341742665,0.2661927342,0.0968197584,0.2421134561,-0.4497469068,-0.1726887971,-0.3069637716,-0.0053717731,-0.1520239562,-0.1448992342,0.0667963028,-0.0198877677,0.6864049435,-0.1905841082,0.2406973541,-0.1494981796,0.1166925207,-0.1677237153,-0.1573780477,0.2168618292,0.0933346674,-0.1515265107,0.0986844525,0.1273582578,0.073553957,-0.0123291211,-0.1933779418,-0.0778044537,0.0731493235,0.3307126164,0.0673327073,0.1481941789,-0.0800899193,0.3283025026,0.3131912947,0.2075050324,-0.1654668152,-0.1291126758,-0.0672498494,-0.0664217025,0.2284736335,0.1068682075,-0.3879636526,-0.2831277251,0.2111899704,0.1523139328,-0.1277307868,0.1663789004,-0.0637842715,0.4463846385,-0.0574760288,0.1228245422,-0.1606709957,-0.0278364383,-0.0758509561,-0.0808564126,-0.0359562635,0.0637387708,0.0771537423,-0.3781818449,0.3808516264,0.1721238941,0.2533078194,0.2878333032,-0.0454363003,-0.0531011745,-0.3692813516,-0.333049655,-0.1322299391,0.0303953551,0.1376107484,-0.0659362376,-0.0196515061,-0.2123082876,0.4386965334,0.0753439441,-0.1268944889,0.2475073934,0.2016558796,-0.175553605,0.134940505,-0.0882637501,-0.4648680985,0.3856156766,-0.2136990875,0.1805250943,0.1455844641,0.0752157122,-0.5313087106,-0.0354604274,-0.0576039143,-0.3980095685,0.2089185566,-0.04827407,0.1521581262,-0.0416214764,-0.1192983985,0.0717149377,-0.2015973032,-0.0051512355,-0.2455405742,0.2948980629,-0.1022188514,-0.0338154435,0.2529664934,0.1713653058,0.0838688239,-0.2252237201,-0.0535434745,0.3246357739,0.1384629458,-0.0993399248,-0.1715636551,-0.1226139367,0.1180634499,-0.131923154,-0.2020104975,0.2057533711,0.1158741117,-0.0233052354,-0.1940400749,0.3919537663,0.1703977436,0.2448907644,0.2319546491,0.0529230796,0.1448597759,-0.0916807204,0.0975734591,-0.0638853163,0.189599961,-0.0497281142,-0.0844068453,-0.0868861899,-0.2769928873,0.1296674013,0.1336871833,0.0348882228,-0.0792626292,-0.162670821,-0.1537512541,0.1644689143,0.1590701044,0.1174459532,0.1510914415,0.30799824,-0.2536084652,-0.199834615,0.0356864035,0.0004941155,0.1626493484,0.0426238924,0.2464949638,0.2847855985,0.1936950982,-0.0943688676,-0.1065967083,0.1231058016,0.1845837831,0.2153210491,-0.3730052114,0.1432093829,-0.1024540812,0.2950365245,-0.1279529482,0.235516116,0.159334287,-0.2550415099,0.2276655138,0.0367487222,-0.053009022,0.264975369,-0.254773885,0.2745153606,0.1846229136,-0.1531972736,0.1978167146,0.1969082803,-0.3121934831,0.2985317111,-0.0876592323,0.1244520769,0.2559636235,-0.0741104484,-0.298353672,0.1935950369,-0.2393571287,0.1780943573,-0.1836472899,0.483614713,-0.141664952,0.1909188926,-0.3466756642,-0.3810741007,0.1948749274,-0.0581267588,-0.2577676773,0.1504852474,0.0410655998,-0.2013731003,-0.1593966931,-0.7575306296,0.0354381427,-0.1635408401,0.1972049624,0.1563578099,0.2769849002,0.0308583379,0.2668566406,-0.0354165621,-0.1260354668,-0.1651287824,-0.1709091961,-0.194642216,0.3205131292,-0.1530383378,-0.3015633821,-0.1450711936,-0.1393867731,0.5834782124,-0.0346679166,-0.3780317307,-0.2207747847,-0.0099999337,0.1090353206,-0.041204527,-0.1890207827,0.3572488725,-0.0990521386,-0.202939555,-0.0918208286,-0.0775542781,0.308878094,0.1910594404,0.258572638,-0.4090474844,0.2854526043,-0.0080062728,1.0584617853,0.3723424077,0.0747849196,0.0401264355,-0.0044042636,-0.1289398372,-0.1965896934,-0.1487231553,0.1607521623,-0.0593115315,-0.2840839028,0.2656474113,0.1344437301,-0.0197183974,-0.1355739981,0.2519261539,-0.6226943731,-0.1467548907,-0.0441396087,-0.067962192,0.5081368685,0.026144499,0.2575253248,0.1802202463,0.0982106775,0.0189101733,-0.0221583322,0.1799109429,-0.0480811037,-0.384996593,-0.0720152259,-0.06307327,0.1918585002,-0.1368911713,0.2850539684,-0.033641085,-0.0315195583,-0.0155024678,0.019306384,0.491650492,0.1030349955,0.0759638697,0.2137275189,0.0983606428,-0.0879001468,-0.155747667,-0.1249988601,-0.3020196259,-0.1005191281,0.3204067945,-0.3774962127,0.0301839281,-0.0500864834,0.0213078782,-0.3617132008,0.0159993488,-0.3647143245,0.0236450788,-0.1130777672,0.1736128032,0.2534516156,0.239273876,-0.0875733495,-0.0866843015,0.0384343378,-0.3197672069,0.2297399342,0.3085214198,0.3850121796,0.0758537948,0.5038484335,0.1059484705,-0.260270685,0.2349551767,0.5228100419,-0.1499752849,-0.3038350642,0.1117307395,-0.1718555242,0.4033382237,0.2430368215,-0.1971922219,0.0012441196,-0.0755827427,0.2412461638,0.2151294053,0.082520023,0.0790469274,-0.0239105821,0.2027700543,-0.3584736586,0.0560042709,-0.1665772796,-0.0322501585,-0.0816431642,0.1298967302,-0.1747607291,0.333870858,0.1127267703,0.6330314875,0.130593732,-0.0383033939,0.4832280874,0.0490094945,0.122882776,-0.3597963154,0.2030264735,-0.2173835784,-0.1108087525,0.0887389481,-0.0648304,0.1883213073,0.3287271559,-0.1439502239,0.4234223962,-0.448482275,-0.0406923443,0.0738431811,-0.0510886908,0.03146483,-0.0251645409,-0.2900235653,0.2359438986,-0.0015879149,-0.0406085663,-0.1033929139,-0.0731634647,-0.1227568015,-0.1274030656,-0.0092289634,0.2474857122,-0.311883688,-0.1006114334,0.3116766512,-0.1610029638,0.0762248859,0.3196717501,-0.120197773,0.2709676325,-0.0974705294,0.2468037605,-0.046355471,0.057972163,0.0048045497,-0.0912673771,0.4631349444,-0.1395934969,-0.0712312311,0.4149790406,-0.2818175852,0.0176113918,0.0510923229,0.0648634806,0.0080351215,-0.3585578799,-0.1104812324,0.1376360208,0.0029345215,-0.3381873369,0.2300059199,0.0572285652,-0.189274773,0.1670876443,0.0865444615,-0.1463118047,0.0427084751,0.1361495554,0.3570473194,-0.1305964291,0.7882820964,0.1528767347,-0.0398471579,-0.3938729167,-0.0726180077,0.2945176363,-0.435719192,0.3019652069,-0.0136567447,-0.0246525276,-0.0521612875,0.4134021997,0.1887339801,0.0399365388,0.2184929401,-0.4894676507,-0.1931098551,0.2399053276,-0.0145094646,0.0302345157,-0.1322632134,0.290312022,-0.1278672963,0.2097464204,-0.4214288592,-0.1433736831,-0.3985895216,0.1878068149,-0.5181497335,-0.1457366645,0.2272179425,-0.1095385626,0.2473854423,0.0375558138,-0.256757766,-0.3200960159,-0.2861353457,0.0617648736,0.1927329302,0.1127406061,0.288939625,0.041668918,-0.2834427953,0.0178245381,0.1013175398,-0.1948021054,0.0467673764,0.0871966854,0.3864264786,0.0219777226,-0.19198744,-0.1358885318,-0.0882558972,0.0904601142,0.0601679422,-0.2023030221,0.0523505248,-0.012818356,-0.1507111937,0.2215189189,-0.1231176332,0.2737950087,0.2905271351,-0.1926757544,-0.0184266027,-0.1168123558,0.1618770808,0.1685914248,-0.1445003748,0.1315016448,0.0135789206,0.3528312147,-0.3844121397,0.0049668453,-0.3526668847,-0.0184044205,0.0893212184,0.1673433781,0.0437684692,-0.3547252417,-0.0860331729,0.2811055779,-0.2622293532,-0.1081992239,0.2802663743,0.5731699467,0.0275648199,-0.0002921917,0.1819287091,-0.3897887468,-0.0526641682,-0.0991594121,0.07343328,0.1352884918,0.2290858924,0.020084573,-0.0661902726,-0.1742319316,0.011692862,0.0469578169,-0.0897581652,0.3786462843,-0.0138229989,0.2987894714,-0.0638646409,-0.1497026831,-0.3668247461,-0.168053031,-0.0015945042,0.0645888075,0.0085983807,-0.1513916403,-0.1663929671,-0.08183337,0.0792207792,-0.3283424973,0.154255718,-0.1284558624,-0.4417531788,-0.2323850244,-0.3880307078,0.4857454002,0.258284539,-0.2795485556,0.2843214273,0.1506462991,-0.1825749129,-0.1176908165,0.4338412881,0.128068015,0.0845422074,-0.0463073887,-0.0871170908,0.0822568536,-0.1040437296,0.1117975786,0.2822864652,0.1642481089,-0.0767761096,0.2574383914,0.1753002852,-0.2762508988,0.1733468771,0.081053704,0.2326996773,-0.2533421516,0.1304880679,-0.1144494787,-0.0447532646,-0.0553241372,0.1883161962,-0.401135236,0.0053416523,0.1108732074,0.1102420613,0.1241546795,-0.0843065232,0.1275138706,-0.3049993217,0.5124530792,0.0103119044,0.0229149908,-0.4916670918,-0.064716503,-0.2829300165,-0.0501294881,-0.2130349278,0.2104375511,0.0346973427,0.209083125,0.1262042075,0.1017698646,-0.0141174598,0.0250880383,-0.2571706474,0.3189039826,-0.7061988115,-0.135068506,-0.2019830495,0.3497312963,0.1605273336,-0.3106667995,0.0155651346,-0.1239759475,0.1142854393,0.1948511302,-0.1609454155,0.3158736825,-0.0857972428,0.2146507949,0.1274192929,0.1434754282,-0.0719684511,-0.1633684337,-0.1401197612,-0.3524002433,-0.2526800036,0.2253606617,0.2688834667,0.4694870114,0.1029353961,-0.1061380208,0.0337461829,0.0445147716,0.2115573585,-0.4895721972,-0.3758788407,0.2332729399,0.0127202421,0.2232289761,0.1194653288,0.2636312544,-0.1482480466,-0.2476286143,-0.3535003662,-0.4347568154,0.4308348596,-0.07013385,-0.0934192091,-0.1445526779,0.2189143002,0.0158601161,-0.0631370321,-0.6744795442,-0.0659779683,0.409499377,-0.1132725105,-0.2585168481,0.1806693077,0.1244951263,-0.1271043271,-0.0564316101,0.2745738626,0.0278273188,-0.2310827225,0.0818093717,-0.0966653824]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2707","title":"404 Not Found Error when loading LAMA dataset","comments":"Hi @dwil2444! I was able to reproduce your error when I downgraded to v1.1.2. Updating to the latest version of Datasets fixed the error for me :)","body":"The [LAMA](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=lama) probing dataset is not available for download: \r\n\r\nSteps to Reproduce: \r\n\r\n1. `from datasets import load_dataset`\r\n2. `dataset = load_dataset('lama', 'trex')`. \r\n\r\n\r\nResults: \r\n`FileNotFoundError: Couldn't find file locally at lama\/lama.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/lama\/lama.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/lama\/lama.py`","comment_length":27,"text":"404 Not Found Error when loading LAMA dataset\nThe [LAMA](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=lama) probing dataset is not available for download: \r\n\r\nSteps to Reproduce: \r\n\r\n1. `from datasets import load_dataset`\r\n2. `dataset = load_dataset('lama', 'trex')`. \r\n\r\n\r\nResults: \r\n`FileNotFoundError: Couldn't find file locally at lama\/lama.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/lama\/lama.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/lama\/lama.py`\nHi @dwil2444! I was able to reproduce your error when I downgraded to v1.1.2. Updating to the latest version of Datasets fixed the error for me :)","embeddings":[0.0275762063,-0.1798952222,-0.0416179113,0.5950444341,0.4188714027,0.1276884079,-0.1021334752,0.4045758247,-0.4222145975,0.2916164994,-0.3122213781,0.1641525775,-0.159445703,-0.2467197478,0.0084447721,-0.3867286146,-0.1002725586,-0.0234841295,-0.2141502798,0.1703372449,-0.0552273877,0.3564765453,-0.1458574384,0.1938980818,0.0413480662,0.1218298301,-0.1561390162,0.3231827319,-0.2132577449,-0.3137521744,0.376162082,-0.1966380328,0.2636945844,0.7459003329,-0.000113161,0.170933485,0.2560487688,-0.1524574012,-0.2275671512,-0.3300167024,-0.2031929493,0.020837808,0.0260597821,0.0620778389,-0.327978462,-0.0541198514,0.0296569485,-0.106870763,0.0726983771,0.2427031249,0.2222920507,0.5647499561,0.1862714142,-0.3274774551,0.155467093,-0.1818133444,-0.1400780976,0.6799879074,0.1498614252,0.4447278678,0.1170044169,0.4727893472,0.0081939111,-0.0207349416,0.2193838507,-0.1606039107,0.3265073597,-0.3726662099,0.2364120185,0.3105343878,0.398439914,-0.1125394702,-0.3003064394,-0.188016057,0.0631653368,-0.1206552461,0.2855997384,0.1726784557,-0.2515365183,0.2372489721,0.064684689,-0.3774136603,-0.030924784,0.2397215515,-0.0525427237,-0.3283061683,-0.1941554397,-0.0440329574,-0.0428908393,-0.1177555099,-0.1223740503,-0.1022243947,-0.0604988597,0.4073463678,-0.1801186353,-0.0222832523,0.1310425252,0.1632917374,0.0916107669,0.2455806285,-0.0600872375,0.107350491,-0.4293561876,0.1083620638,0.2306107134,0.0193927865,0.0387212373,0.0064264713,0.1022538543,0.559654057,0.1537634581,-0.0669870302,-0.3494531512,-0.3994399905,-0.4301458895,0.0453638174,0.3619140089,-0.2745935321,-0.1926773489,-0.2255994081,-0.1761281192,-0.1179646626,0.328471154,0.4818301797,-0.0135623077,0.1518219262,-0.2231666446,0.1077072024,-0.0843086615,-0.1456580907,-0.2482452691,0.0794835985,-0.2387344837,0.1790161431,0.1133854911,-0.6890286803,0.3001265824,-0.305708766,-0.1137952954,0.008293299,0.0363545194,-0.0737774596,-0.1109906659,0.2830492258,0.0126989959,-0.0248338543,-0.0502680764,-0.123763524,-0.135119915,0.0123523436,-0.2292298079,-0.2775953412,-0.0139294909,0.1285256743,-0.0168099627,0.203304857,0.1159618348,0.1686997861,-0.2712258697,-0.128826499,-0.0279897451,0.139417395,-0.0879680589,-0.1676285118,0.4258745015,0.5617586374,-0.1878831685,-0.0116636874,-0.0652727038,-0.4448719323,0.021887565,0.2667708099,-0.1007739604,0.2799500227,-0.5007410049,0.0761442259,0.3697856963,-0.3822958767,-0.2424640507,-0.017970413,-0.1458064318,-0.0362399928,0.045597408,-0.1990565956,-0.0711607784,-0.1117392629,0.4304517806,-0.0396544561,0.108241044,-0.2827382386,-0.1561469436,-0.1705923676,-0.0402733684,0.0282070879,0.0869352147,0.0649742782,0.1890324503,-0.1641197652,0.1505366564,0.017725084,0.0044502984,0.33847785,0.333594054,0.2531719208,0.0434678271,-0.1701409817,-0.0906351954,0.1604188234,-0.1763248742,0.2604502439,-0.3216400743,-0.0586831719,-0.3983004391,-0.0621902123,-0.2430595011,-0.1129879728,0.1386483759,0.0277117025,0.1554589719,0.3996631205,-0.1036331877,0.0866570175,0.0413638577,0.3868951499,-0.5217029452,0.2882936299,-0.1621196568,-0.0569980741,0.0447353981,0.1650550961,0.0889760703,-0.2264202088,0.0879154354,0.3215096891,0.1480682343,0.0254212078,0.6113657951,-0.2260311991,0.1211305782,-0.272159636,0.0900689587,-0.0116431955,0.1313446164,0.1288069934,0.063158825,0.1652727276,0.1123038009,0.0762929022,0.1483362764,0.0277672596,0.3416911662,-0.0324320681,-0.072627984,-0.2988233566,0.2953135371,0.1560999155,0.2211470455,-0.0873381495,-0.2797177434,-0.2418055683,-0.2270907015,-0.105482161,-0.0097520975,0.1181006506,-0.3057870865,0.2041134685,0.2184486538,0.3560207188,0.199327752,0.1325501353,0.1636909544,0.1632709503,0.1215503141,-0.1284631938,0.0921372399,0.1699588597,0.0739580318,0.1871473789,-0.1497145146,0.0457578301,-0.3276008964,-0.0989335477,-0.159858793,0.2108633518,-0.1368249655,-0.1308034062,-0.3650540113,-0.5920037031,-0.2478825003,-0.0549499802,-0.4640192091,-0.2258246094,-0.129240036,-0.0146912904,-0.0487080924,0.0384634621,-0.3173106909,0.1435844749,-0.08266332,-0.0794701278,-0.2091592699,0.0474949628,-0.0480669215,0.0619082414,0.3312280476,-0.0067720227,0.3388998806,-0.3461804688,0.110254623,-0.2857080996,-0.3271134496,0.1754175723,0.1256680638,0.2020720541,0.1722787917,0.1957675517,0.0749968067,-0.1231173426,0.3749608397,0.0885668844,-0.1042628959,0.0144845201,-0.022987701,0.2079499364,0.0007505329,-0.2067741305,-0.1117951125,-0.2796300352,-0.0600172356,0.075303629,-0.0136360526,0.0379966013,0.2016094923,0.3852606118,-0.5088940263,-0.0519054495,-0.0848602504,-0.3770296276,0.3114745021,-0.2667360902,-0.2900529802,0.3200337589,-0.2356281877,0.5331386328,-0.1657560617,-0.476556927,0.0859394297,0.0523845255,0.1605126709,0.2272054851,-0.1370075196,0.0498042554,0.0314962864,-0.113392286,0.0020012052,-0.0747741535,-0.3300971389,-0.0530968681,0.4950979054,-0.1138645932,0.0503144599,0.0733376518,0.4199269414,0.3726042509,0.1534758508,0.4652198553,0.1167078093,0.5983056426,-0.0907256156,-0.1926039904,-0.1169227883,0.0117179872,0.127961874,0.0744697303,0.5465224981,0.1982722133,-0.5745587349,-0.3914300501,-0.2849505842,-0.3250796497,-0.3052753508,-0.1911684722,0.1338606328,0.0820843056,0.0130695757,0.083648473,-0.0033110592,0.221497789,0.6827551723,0.0313965753,0.0096505256,0.0218118448,0.4009024501,-0.7738465667,0.2067864388,-0.2102185637,0.0063059945,0.0255133584,0.0963204876,0.2056538463,-0.0184103902,0.6278686523,0.0996665284,0.080792971,-0.0246407073,-0.0494294986,-0.2527091503,0.0870694816,0.0038776551,-0.0025452827,0.1338184774,0.6096508503,-0.2629626393,-0.1945953518,0.1487480104,0.090038985,-0.1467810422,-0.2016202807,-0.2816533744,-0.2833282053,-0.1463649869,-0.3225822747,-0.1948909461,0.3163625002,0.4110971689,-0.1206556633,0.120449841,0.1290352345,0.1325499862,0.0731241554,0.2535375059,0.1423864812,0.1947260946,0.4189747274,0.272995472,0.2054225653,0.6228638887,-0.2285934687,-0.5557920337,-0.1378269196,-0.0102492897,0.0661034435,0.2842209339,-0.1786153913,0.1075807959,0.20942083,0.2447222769,-0.3189245164,0.2377249002,0.4166227579,-0.0420566015,-0.2295139879,-0.2105554491,0.175154686,-0.1090682223,0.0142891174,0.5919674039,0.2655681372,-0.0308937244,-0.0533246882,-0.0762447938,0.8943354487,-0.305203408,-0.1355106533,0.4126397669,-0.1429207474,0.3268454671,0.1835494041,0.1597992033,-0.0332711376,0.1485086977,-0.0339826383,0.0062116575,0.1509281993,-0.0666003227,-0.0723304078,0.370077014,0.1185522005,-0.199647069,-0.1146825701,0.0298474673,-0.0864053518,0.157728821,-0.1449265927,0.1460780352,0.0016581728,0.6845747828,-0.1613355279,-0.1802317351,-0.2283399999,-0.1184443757,-0.0732770562,0.0116945524,-0.1795803607,0.0108829895,0.0359085351,-0.4694708288,0.0718540996,0.1127227694,-0.0739290416,-0.1057820618,-0.4312487543,0.2365910113,-0.2660598755,-0.4037063718,-0.0257567,0.0901981816,0.2451154739,-0.0258612446,-0.1520773023,0.0672041401,0.0271411166,0.0056663677,0.2813135982,-0.0833319053,-0.1544920057,-0.0350880437,-0.1870540679,0.0724634007,0.1514791548,-0.1914689839,0.1532903761,0.263368994,-0.1594762802,-0.4186688066,0.178305909,-0.098926127,0.0503907911,0.392462641,-0.2345751524,-0.3326857686,0.1719307303,0.3241139948,-0.1872755885,-0.1208377108,0.1790044755,0.111054711,-0.5132271051,0.0620101169,0.1603041738,0.4298988581,-0.0017270922,0.0022817804,0.2556880713,-0.6776071787,-0.1010790542,-0.3675164878,-0.2923965454,0.2684902251,-0.1659120023,0.0773575455,-0.1927930862,0.0267066937,0.2402820885,-0.1025599241,-0.2744699121,0.1755054891,-0.2248121053,0.0137472721,0.1955945939,0.1704115868,0.1359426528,0.0623149239,0.1203891635,-0.1133150458,-0.3900868595,-0.1529636234,0.0052958466,0.0870560631,-0.0175164863,-0.1752778292,0.0797705054,-0.1581255943,-0.0152746998,-0.1603564918,-0.1277949512,0.2234126031,0.1035035774,-0.1417796612,0.129402265,-0.0120223276,-0.270036608,0.3153372109,0.1504018903,0.2719039023,0.2604071796,-0.1854821593,-0.2581343055,-0.059067294,-0.0993146524,0.0291406102,-0.0427688956,-0.1533045024,0.2613158524,-0.006748002,0.037508063,-0.0210290477,0.2382086962,0.0559396148,-0.1683042794,0.2652599514,0.4411591291,0.1108987331,-0.3262587488,0.1049161777,0.1623978168,0.0359196961,-0.0440053716,-0.1589201242,-0.2583518028,0.2208855897,-0.0149282375,0.0792653337,0.1950485408,0.1507975459,0.3202217817,0.4101885557,0.2101357728,-0.0290638432,0.3033574522,0.1490911245,0.0973710939,0.4854170084,-0.1422750801,0.3498802185,-0.6094716191,0.299294889,-0.0194636509,-0.447278738,-0.048181206,0.1635994464,-0.0522701181,-0.1631706953,-0.1516117603,0.3918914497,-0.1910349876,0.1293948591,-0.1471154243,0.0728554279,0.0067059388,0.0518127754,0.1234891191,-0.3167695105,0.0602369234,-0.029231349,0.2890924513,-0.3182105422,-0.120258823,0.0752681196,-0.0915647075,-0.1428079605,-0.2174875587,0.0918717757,-0.0202278811,-0.0118623134,0.1545276493,0.2011739165,0.0429387949,0.021825498,0.190354839,0.4008926451,0.043345809,0.0538102984,-0.1292659193,-0.1521480829,-0.0523978546,0.0550789312,0.2060990483,0.039979361,0.1083694622,0.3821549118,0.1420225054,-0.1837977469,-0.2758408487,0.0675113276,0.2499995381,-0.2783149779,0.4672813416,-0.5629258752,0.0096471068,-0.1725465208,-0.0641685128,-0.5021948218,0.5542028546,0.2244344205,-0.3633378446,-0.1388117075,-0.1607698351,0.0650712773,0.0170879606,0.4634781778,0.3883374035,0.4095773101,-0.244513303,-0.5597167611,-0.7324652672,0.2157923281,-0.1681682765,0.0038510081,-0.0165796857,0.0682508945,-0.1144281104,0.3239782453,0.2871296406,0.0853668228,0.3343954086,-0.1083987504,-0.0686181337,-0.1621775329,-0.2101041973,-0.0269089453,-0.1025996804,-0.1848218441,0.0655351505,-0.4303753972,-0.0103819063,0.1170590147,0.135375604,-0.1436257064,-0.0944080874,0.4596574605,-0.0379392765,0.2285944223,0.0479290485,-0.067418322,-0.1437526494,-0.2410252988,0.2154723555,-0.0430866815,-0.0316426903,0.0373084806,-0.2298004478,-0.2477363795,-0.2015143633,0.5169576406,0.1491552889,0.3214690387,-0.0567174889,0.1079277694,-0.0420835353,-0.1029458046,0.2558766007,-0.0612671934,-0.1464368999,-0.0608653091,-0.1390431672,-0.1754335612,0.661213398,-0.3637351394,0.1629911363,-0.1541460901,0.0348055698,0.0658015236,-0.2408943176,-0.7541190982,0.2883235812,0.2903968692,0.3903473914,-0.1727058291,0.1597911716,-0.2616016269,-0.0970124453,0.1424275041,0.0466230363,-0.0516107976,-0.0518991724,0.1443367302,-0.0248512998]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2707","title":"404 Not Found Error when loading LAMA dataset","comments":"Hi @dwil2444, thanks for reporting.\r\n\r\nCould you please confirm which `datasets` version you were using and if the problem persists after you update it to the latest version: `pip install -U datasets`?\r\n\r\nThanks @stevhliu for the hint to fix this! ;)","body":"The [LAMA](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=lama) probing dataset is not available for download: \r\n\r\nSteps to Reproduce: \r\n\r\n1. `from datasets import load_dataset`\r\n2. `dataset = load_dataset('lama', 'trex')`. \r\n\r\n\r\nResults: \r\n`FileNotFoundError: Couldn't find file locally at lama\/lama.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/lama\/lama.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/lama\/lama.py`","comment_length":41,"text":"404 Not Found Error when loading LAMA dataset\nThe [LAMA](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=lama) probing dataset is not available for download: \r\n\r\nSteps to Reproduce: \r\n\r\n1. `from datasets import load_dataset`\r\n2. `dataset = load_dataset('lama', 'trex')`. \r\n\r\n\r\nResults: \r\n`FileNotFoundError: Couldn't find file locally at lama\/lama.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/lama\/lama.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/lama\/lama.py`\nHi @dwil2444, thanks for reporting.\r\n\r\nCould you please confirm which `datasets` version you were using and if the problem persists after you update it to the latest version: `pip install -U datasets`?\r\n\r\nThanks @stevhliu for the hint to fix this! ;)","embeddings":[0.0376231372,-0.2442803234,-0.0386776887,0.6305751204,0.4072245955,0.0883304179,-0.0716991201,0.3489049077,-0.3947159946,0.264603734,-0.2870087922,0.2205041796,-0.1573990136,-0.2962248325,0.0108721973,-0.3568078876,-0.1149339452,-0.0169953573,-0.1815581769,0.1489533335,-0.0625042245,0.3521650434,-0.1265231669,0.1486698538,0.0735035911,0.1503960192,-0.1259550154,0.2814263105,-0.190541774,-0.324713707,0.4116790593,-0.1972879469,0.2471829355,0.7875105143,-0.0001171694,0.1473741084,0.2983234227,-0.1649733931,-0.2821417749,-0.3706027567,-0.2304017842,-0.0749065652,0.0254260506,0.0481946096,-0.3216995895,-0.0121680219,0.0023315025,-0.0769948438,0.1006459892,0.2473402768,0.1817989498,0.5502545834,0.179993704,-0.2549853325,0.1239052936,-0.1564266682,-0.1432594657,0.6937092543,0.212116614,0.4913322031,0.1655281186,0.466221422,-0.0164648257,-0.000398995,0.1752236336,-0.1469997317,0.3066895306,-0.3588722348,0.2023094743,0.3324887753,0.4127957225,-0.1082109511,-0.3176200092,-0.2034640759,0.0543417782,-0.1205691546,0.2801489532,0.2186776847,-0.2847771049,0.2256470174,0.0706216395,-0.4013845623,-0.0462253094,0.2663989961,-0.0475644954,-0.3123933375,-0.2254801095,-0.0457049645,-0.0561008006,-0.122848995,-0.1137527674,-0.1313110739,-0.050231088,0.4134670496,-0.1436476111,-0.0026620096,0.0764176324,0.196240738,0.0534739718,0.2188102007,-0.0718790889,0.1110744551,-0.3836347163,0.0944969058,0.2684501112,-0.0089992173,0.0116549749,0.0559998192,0.0936570838,0.5667535067,0.180558607,-0.0413758755,-0.3334462047,-0.3940544128,-0.4419755638,0.0216966718,0.3429944813,-0.2445634156,-0.181790486,-0.2328260243,-0.1784704328,-0.1435522437,0.331209898,0.4548911154,-0.0192042291,0.1615790725,-0.1978382319,0.0973857343,-0.07085976,-0.1500709653,-0.2035821974,0.0518461131,-0.227414161,0.1836643666,0.151868999,-0.7167775035,0.3160148561,-0.2461057305,-0.142082423,0.01632227,0.0305065606,-0.0926795825,-0.0823924765,0.2839703858,0.0460856035,0.0165917315,-0.0983635634,-0.1527360976,-0.1185643375,0.0476335138,-0.2459807396,-0.2467961907,-0.0715638027,0.0831063986,-0.0353121869,0.2100958228,0.0569599718,0.1596894711,-0.235934034,-0.1610614061,0.0122283325,0.1079684719,-0.0236636344,-0.1690436602,0.4432069957,0.5722745657,-0.1904916465,-0.0203828923,-0.1001596898,-0.4123748839,0.0347054265,0.2353795916,-0.105083622,0.3270788491,-0.5126182437,0.0094884476,0.4096075594,-0.4043959975,-0.2640866339,-0.0146399327,-0.103657946,0.0169052184,0.061881315,-0.2221135199,-0.0735695362,-0.0733288229,0.4005431831,-0.0504778139,0.1311737448,-0.2857051492,-0.180993855,-0.1583044231,-0.0461685397,0.027168408,0.0974410176,0.0271266028,0.1707299203,-0.1774171591,0.1637755483,0.0285246186,0.015392079,0.360457927,0.3656584024,0.3046260178,0.0483875982,-0.1808129847,-0.0931404829,0.1336420774,-0.1816187501,0.2133657932,-0.3849082887,-0.0144105749,-0.3852227628,-0.0800000578,-0.2602271736,-0.1229649261,0.093259193,0.0460892059,0.1697572619,0.4013627768,-0.1205973551,0.1332776248,0.0645349845,0.3940297067,-0.5741817951,0.2775955498,-0.1454002261,-0.0851296261,-0.0152847981,0.1977677494,0.0863129199,-0.232964769,0.116725333,0.2829639316,0.1624978781,-0.0122364433,0.5958023071,-0.2751281261,0.1441280395,-0.2933014035,0.0742353871,-0.0074059241,0.1238600835,0.1390321702,0.0064978856,0.1745623946,0.0916065499,0.1102492586,0.1289982051,0.0054872818,0.3452596664,-0.0348348208,-0.1165444255,-0.2764839828,0.3460576832,0.186855942,0.2750222981,-0.0800100863,-0.233538121,-0.247148782,-0.224796474,-0.0969630182,0.0205947887,0.1212189049,-0.3004150987,0.2324629277,0.154048413,0.3634408712,0.2016908675,0.1185744628,0.1071733236,0.1104388684,0.1724074483,-0.1105893403,0.0749386176,0.1939235926,0.0081138499,0.223238185,-0.1277025193,0.0503921695,-0.2690105438,-0.0030083468,-0.1563002616,0.1812726855,-0.146503523,-0.0971217304,-0.3710281849,-0.5473519564,-0.2929015756,-0.0843420625,-0.4625463486,-0.198014304,-0.1817056239,-0.0270366874,0.0327184871,0.0591555312,-0.2924275696,0.1223865151,-0.1217258722,-0.1199230775,-0.2157153934,0.0569689013,-0.0498639978,0.0278448034,0.3217616677,0.0231651142,0.331033498,-0.3448909223,0.1146355495,-0.2824027538,-0.3251257539,0.1471241266,0.1241642013,0.2616078854,0.1532170922,0.2105141431,0.072250925,-0.1196710095,0.3903245628,0.1341011971,-0.0772342011,0.007932202,-0.0552426577,0.2397370785,0.0300046932,-0.1942540258,-0.1402244717,-0.2553319037,-0.0677700192,0.0564313084,-0.0548759773,-0.0302682444,0.2058975846,0.353776902,-0.5599407554,-0.043356888,-0.0665352792,-0.3719189167,0.3360426724,-0.2776235938,-0.2710165679,0.3012785912,-0.2774116993,0.5949883461,-0.1596422195,-0.4551080465,0.1483637989,0.0645555556,0.1678711921,0.2163109481,-0.1184497252,0.0491827913,0.0213148855,-0.0523296706,0.0135797393,-0.081472598,-0.3569486141,-0.0511713736,0.5589962006,-0.0684969798,0.0786288455,0.0684092417,0.4786718488,0.402877599,0.1334718466,0.4516437948,0.0872956291,0.6159577966,-0.0439010262,-0.2505469918,-0.1680506021,0.0329114161,0.1802301705,0.0468052253,0.5635349751,0.1887270957,-0.5760725141,-0.3702651858,-0.3555663526,-0.3124454021,-0.2666375935,-0.1730482131,0.1711355448,0.1012065709,0.0234094393,0.0597659424,-0.0263879057,0.2043713778,0.6898043752,0.090715766,0.0081735617,0.0336518027,0.3789706528,-0.7080020308,0.1770663857,-0.1999406815,0.0263014212,0.0229964908,0.127669245,0.216702491,-0.0187887661,0.6216830611,0.0953301042,0.0419659726,-0.048609443,-0.0291160662,-0.3157764077,0.09242744,0.0278220084,-0.0208462756,0.1545674205,0.5928747058,-0.3154581189,-0.1971945018,0.1752249151,0.1359913945,-0.2190579623,-0.2317588776,-0.3073806167,-0.2623599768,-0.121384792,-0.3269092143,-0.2211967409,0.3113202751,0.4118064344,-0.1419180781,0.1385832876,0.1523176879,0.0619500615,0.0166473724,0.2721293867,0.1267293841,0.2077225447,0.4207659364,0.2836778462,0.2216935158,0.6485916376,-0.278450191,-0.5561398268,-0.1726404727,-0.0462086722,0.0759332627,0.2549341023,-0.1942779869,0.0927216411,0.2109360546,0.2335279286,-0.3187153637,0.1677999347,0.424769789,-0.0356991589,-0.2502805591,-0.221352458,0.1825153977,-0.0822789893,0.0254876558,0.6119912863,0.2791707814,-0.0188064184,-0.0784475282,-0.1266102493,0.9042650461,-0.2687930763,-0.1294871867,0.4073129594,-0.1276659071,0.2972317338,0.1540271342,0.1268143654,-0.0180852991,0.2190383077,-0.0297417156,-0.0342446752,0.1612790674,-0.0748640001,-0.0664097592,0.3986872137,0.1374391317,-0.1995113045,-0.1338716447,0.0169897862,-0.0966544971,0.1677269489,-0.1297188997,0.106355764,-0.0125417802,0.7009760141,-0.1536023468,-0.1957854778,-0.249174118,-0.1173383296,-0.1056030095,0.0045499336,-0.1956383586,0.0213312376,0.0547736324,-0.518883884,0.0887021795,0.1006741226,-0.0801484361,-0.1125144809,-0.440089047,0.2601293027,-0.2693246007,-0.3994054794,-0.0319329761,0.0987919718,0.2109645903,-0.0006945121,-0.1603876948,0.0480661504,0.0254752208,0.0005276164,0.2996793091,-0.0604753047,-0.1382225454,-0.0428048111,-0.188185066,0.0786639675,0.1960265636,-0.1662229151,0.1293876022,0.2950854003,-0.1340365708,-0.3536822498,0.1474453062,-0.1143755317,0.0568105355,0.4100630283,-0.2306811213,-0.3865353167,0.1940342337,0.3112259805,-0.1955405921,-0.1101502702,0.2158876508,0.0721121877,-0.524944365,0.0313300714,0.1951513439,0.4888619184,-0.004363629,-0.0386479646,0.2547053099,-0.7374974489,-0.1480491608,-0.3607776463,-0.2593629062,0.2665599883,-0.1515924484,0.0607853606,-0.1825426072,-0.038474936,0.282607317,-0.0899165198,-0.235411182,0.2019675374,-0.193925187,0.0402220264,0.2166129053,0.1754212976,0.1277009696,0.0524839424,0.0962223858,-0.1248779893,-0.3624173105,-0.1364046931,0.0337148272,0.1087865531,-0.0152021274,-0.1299057007,0.0349427015,-0.1888576299,0.0385440178,-0.158751905,-0.1408679783,0.2055956274,0.1142647266,-0.1182226613,0.1314355284,-0.0177488942,-0.3116674721,0.3304662704,0.1340043694,0.2897087634,0.2980369031,-0.1655776054,-0.1955421567,-0.030240925,-0.1134232581,-0.009879658,-0.0035832119,-0.1396647245,0.3032436967,-0.0385259204,0.0621065088,-0.0363337584,0.186191231,0.0858511925,-0.1964636296,0.3042485714,0.4258559942,0.0797969401,-0.331640929,0.1336796731,0.1483485252,0.0621007569,-0.0390260927,-0.1588812172,-0.2443127632,0.1652998775,0.0001059369,0.0717246607,0.203134954,0.1051310748,0.3249289095,0.4476822317,0.244368434,-0.06357342,0.2858736813,0.1709221303,0.042959176,0.4747550488,-0.1251416802,0.373691231,-0.6186774969,0.3184231818,-0.0067492626,-0.4154499769,-0.0951471552,0.1269229054,-0.0572546385,-0.1609792113,-0.159477219,0.3915146887,-0.1774332076,0.10141965,-0.1726039499,0.0407884158,0.008249606,0.0605956241,0.134408623,-0.3056889474,0.0596011728,-0.0243143775,0.2929198444,-0.3246059716,-0.1594113111,0.1317434162,-0.1052236557,-0.1515741497,-0.2282690108,0.1309056282,-0.0233432222,0.0086627454,0.1606980562,0.2212691605,0.0177659858,-0.0087296115,0.2035346627,0.3928328753,0.0582292713,0.0500923172,-0.1159055904,-0.1569764763,-0.0361392051,0.0364915803,0.2557414174,0.0382207111,0.0677729473,0.3389037251,0.1214955822,-0.1665449739,-0.2820751071,0.0614657663,0.2272478789,-0.2657392621,0.5067769289,-0.5798210502,-0.019848628,-0.2129505575,-0.0859582648,-0.5317122936,0.5941915512,0.2756078839,-0.3396126032,-0.1432628632,-0.1595932692,0.040614251,0.0290216543,0.4412983656,0.3815442324,0.4354862571,-0.2590344548,-0.5794159174,-0.6965132952,0.1823433936,-0.1787937731,0.0366058499,-0.0486074015,0.0823875815,-0.1335340887,0.3341455758,0.3072037399,0.0891756117,0.3979126215,-0.0584492087,-0.080756709,-0.1679403037,-0.1781058908,-0.0150630586,-0.106800884,-0.1711028963,0.0387680233,-0.4375113249,-0.0356997363,0.1228637695,0.1446999758,-0.1124499217,-0.0667093098,0.4605039656,-0.0549265221,0.2260173708,0.0524843745,-0.0462036207,-0.1337998211,-0.2298377454,0.2762641609,0.0127297873,-0.0151468283,-0.0203662179,-0.2614908516,-0.2483704984,-0.2103631943,0.4305934012,0.1235957593,0.2976341546,0.0021888805,0.0183614418,-0.0310731586,-0.0940474421,0.244273439,-0.0527429469,-0.1662693918,-0.0464552753,-0.1476934999,-0.1338101178,0.6335508227,-0.3730258048,0.1740248352,-0.1673176885,0.0666938052,0.0493967868,-0.208042562,-0.7280305028,0.2806954086,0.3322626948,0.3692113459,-0.1336748898,0.1337774992,-0.2310562432,-0.0583754554,0.1305673569,0.0492009744,-0.0403046235,-0.0162089411,0.1304627955,0.0073383567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2707","title":"404 Not Found Error when loading LAMA dataset","comments":"@stevhliu @albertvillanova updating to the latest version of datasets did in fact fix this issue. Thanks a lot for your help!","body":"The [LAMA](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=lama) probing dataset is not available for download: \r\n\r\nSteps to Reproduce: \r\n\r\n1. `from datasets import load_dataset`\r\n2. `dataset = load_dataset('lama', 'trex')`. \r\n\r\n\r\nResults: \r\n`FileNotFoundError: Couldn't find file locally at lama\/lama.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/lama\/lama.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/lama\/lama.py`","comment_length":21,"text":"404 Not Found Error when loading LAMA dataset\nThe [LAMA](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=lama) probing dataset is not available for download: \r\n\r\nSteps to Reproduce: \r\n\r\n1. `from datasets import load_dataset`\r\n2. `dataset = load_dataset('lama', 'trex')`. \r\n\r\n\r\nResults: \r\n`FileNotFoundError: Couldn't find file locally at lama\/lama.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/lama\/lama.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/lama\/lama.py`\n@stevhliu @albertvillanova updating to the latest version of datasets did in fact fix this issue. Thanks a lot for your help!","embeddings":[0.0579447038,-0.2256722301,-0.0337308869,0.6018184423,0.396944344,0.1228495389,-0.0771401227,0.392708838,-0.4282576144,0.2528084219,-0.3340739012,0.1994093359,-0.1437425762,-0.263453573,0.0329603218,-0.3631291389,-0.1055584773,-0.035641022,-0.1782314926,0.1402931064,-0.0773811564,0.3491840661,-0.1713868082,0.1461031735,0.0305642188,0.164069742,-0.1523557901,0.2818200886,-0.1676369458,-0.3405278027,0.4034930468,-0.1659760773,0.2667887807,0.7545297146,-0.0001155138,0.1582778245,0.2871517241,-0.180117473,-0.2258957177,-0.3208471537,-0.2181340903,-0.0076235915,0.0299344957,0.0575377792,-0.3109140694,-0.0250622947,0.0147859557,-0.1243565828,0.0784541816,0.2548890114,0.1996642798,0.5669783354,0.15794155,-0.2838997245,0.121087864,-0.1600352079,-0.1550655216,0.6935021877,0.1979849786,0.5019051433,0.1076982692,0.4367845953,0.0512147099,-0.0000367471,0.17038849,-0.1627652496,0.3027929962,-0.362082839,0.2260492891,0.3412948549,0.411247015,-0.0895616338,-0.300208658,-0.2041232288,0.0630146936,-0.1047641039,0.2831819952,0.2124739885,-0.2709573507,0.2240071893,0.0596977212,-0.3989564478,-0.0375748686,0.2220338732,-0.0665908009,-0.3245233297,-0.2249275297,-0.0477187447,-0.0698751584,-0.1235073805,-0.1092792824,-0.0962813273,-0.0449433923,0.4012899101,-0.2000897378,0.0081394352,0.0980038121,0.1805590987,0.071898371,0.2427892238,-0.092317313,0.1385717392,-0.4070573747,0.0528271496,0.2572005093,-0.0139705762,0.0244916715,0.0129902465,0.1060971692,0.5729398131,0.1416838169,-0.0277773663,-0.3655084372,-0.4114148319,-0.4233098328,0.0277434904,0.3314841092,-0.2584204674,-0.1657229066,-0.2332686335,-0.1796723604,-0.1320354342,0.3118084073,0.4670644701,0.0070983572,0.1362685859,-0.194996044,0.1143863797,-0.0724026635,-0.1776781678,-0.2421129495,0.0321830474,-0.2401953042,0.1697179228,0.1667345017,-0.7531397343,0.3183695376,-0.2839638591,-0.1498496383,-0.0083345817,0.0144481035,-0.058931306,-0.0803298876,0.2743904293,0.0371800512,0.0159661993,-0.1051173583,-0.1115520671,-0.1097052768,0.0255913399,-0.2360412329,-0.2386386991,-0.04296856,0.0918830559,-0.0386213176,0.2000920475,0.0625077784,0.1967736185,-0.2370226979,-0.1061923727,-0.0301952213,0.1145451367,-0.0816156417,-0.1677071601,0.4616360664,0.5593075156,-0.2058154792,-0.0246673226,-0.0757972151,-0.4332314432,0.0128629459,0.2552057207,-0.0984490141,0.3181884885,-0.5351845026,0.0659497753,0.3240578473,-0.3493839502,-0.2325602919,-0.0395773239,-0.0958142653,0.0048005632,0.0316052288,-0.2354949266,-0.0602965765,-0.0936399773,0.4298363924,-0.0275842343,0.1246361881,-0.2872694731,-0.1498779356,-0.1584853381,-0.0458613411,0.0432468355,0.1069767177,0.0484397411,0.1735407561,-0.1535310298,0.1848183572,0.0487170815,0.0185187478,0.3353233337,0.3199564219,0.268153131,0.0523090921,-0.1446799934,-0.1134740412,0.1642012894,-0.160792008,0.2338319868,-0.3601338267,-0.0121705504,-0.3953308165,-0.0618144572,-0.2387167364,-0.1459473372,0.108727321,0.0498258807,0.1492777914,0.4205012619,-0.1292730868,0.1097558364,0.0478614606,0.3824838102,-0.5560265779,0.2685636878,-0.1329844892,-0.0726381317,0.0431461222,0.1757280827,0.0637694597,-0.2454117388,0.0792406499,0.2824501395,0.1595969051,0.0057037603,0.5873308182,-0.2710196972,0.1232921034,-0.3047973216,0.0646835789,0.0044439742,0.103925623,0.1204294115,0.0022746688,0.217210412,0.117086716,0.0793338716,0.1173363328,0.0420594774,0.3593285978,-0.0393532701,-0.0884055495,-0.2660077214,0.3219717145,0.1995388567,0.2781038284,-0.0838949308,-0.2605073154,-0.2462642938,-0.2269675434,-0.0955885351,0.0221516397,0.1218470037,-0.3248013258,0.2021924555,0.1915891021,0.3560812175,0.1836343557,0.1089359671,0.1458857656,0.1345479041,0.1329140365,-0.1289663017,0.0620400161,0.1720373183,0.0370383263,0.1893455386,-0.140074119,0.040612068,-0.3171598911,-0.0203471258,-0.1183285713,0.1915402561,-0.1601587385,-0.1411346197,-0.3798510134,-0.5577113032,-0.2689440846,-0.0469791926,-0.4519857466,-0.2289761752,-0.1617726833,-0.0296160374,-0.0148885734,0.0779268891,-0.3301406205,0.1626651287,-0.0877002031,-0.0723210499,-0.2258994728,0.0785861835,-0.0488439091,0.0400497913,0.3276121914,-0.0029401446,0.3127982616,-0.3487844169,0.1278117001,-0.2811443806,-0.3291572034,0.1592871994,0.1283323914,0.2206982523,0.1335715801,0.2108578086,0.0832080916,-0.1141858324,0.3580516875,0.0815688744,-0.0978443846,0.0026842158,-0.0238126554,0.2314572483,0.0320728421,-0.2025589645,-0.1074043065,-0.2520040274,-0.0685241893,0.0493981875,-0.0499446094,0.0068831188,0.1808726192,0.3758721948,-0.5315436721,-0.06250678,-0.0959486216,-0.4061956108,0.3328712285,-0.2945596576,-0.3111495674,0.3056975305,-0.2641014159,0.5637584329,-0.1587353647,-0.4655626714,0.199511081,0.0502379946,0.1819996834,0.2339470088,-0.1430178136,0.0354512893,-0.000908708,-0.0746007115,-0.0022464588,-0.0792549998,-0.3318125606,-0.0562983379,0.5509485006,-0.0979870036,0.04112719,0.1134254709,0.403195709,0.3871581554,0.1285032183,0.4658761621,0.1068727374,0.6118932366,-0.0472184755,-0.205602318,-0.1198549196,0.0198165495,0.1720664352,0.0831649452,0.5636385083,0.1724123955,-0.5466659665,-0.3913594484,-0.3036641181,-0.3372727036,-0.2823136747,-0.2080496699,0.1505243629,0.1003943235,0.0266427696,0.070641771,-0.0146528771,0.2296507061,0.6586833,0.0853421837,-0.0068985363,0.0241842847,0.420714438,-0.7590503097,0.1812446862,-0.2369690388,0.0642023757,0.0293870158,0.1057668626,0.2163592577,-0.0039226878,0.650857389,0.1021037251,0.1004896984,-0.0330602415,-0.0383406468,-0.3130971491,0.0892236903,0.0177642591,-0.0205396265,0.1522286087,0.6132919192,-0.2646584213,-0.2038005292,0.1580061316,0.1116837412,-0.1989924014,-0.2295608521,-0.2762853801,-0.2647717893,-0.1234835386,-0.3163943291,-0.2108536661,0.3021590412,0.4022403955,-0.1270831227,0.1152559519,0.1207247227,0.0784332752,0.0421374179,0.2388522476,0.1956626028,0.2288538963,0.4308677316,0.2876803577,0.2100601792,0.6384994984,-0.2398588061,-0.5860441923,-0.1227778271,-0.051128678,0.069818899,0.2586888373,-0.2100685239,0.0858643502,0.1925001293,0.2209434807,-0.3436311781,0.225848645,0.4244188666,-0.0344654359,-0.2482137829,-0.208871305,0.1694015414,-0.1192342713,0.0178049654,0.5995386243,0.2700920403,-0.0264408812,-0.0489983335,-0.1091582328,0.8627343178,-0.3173603415,-0.1430607736,0.3974967897,-0.1369404495,0.2823217213,0.2000727952,0.1165136024,-0.0252035819,0.174273327,-0.0170885399,-0.0182442404,0.144266963,-0.0740384012,-0.0284070037,0.351041019,0.1294260323,-0.2327616662,-0.1022559106,0.0341761447,-0.0630758926,0.145724833,-0.1192374527,0.1204877794,0.0023406313,0.6824351549,-0.1257649511,-0.1849946082,-0.2092803866,-0.0909756571,-0.1011131704,0.0343016833,-0.1515926421,0.0298920907,0.0362040475,-0.4707596004,0.1246271133,0.1019476503,-0.0547827743,-0.1186410487,-0.4549551308,0.2834399045,-0.2564649284,-0.3910729885,-0.0325380415,0.1019546837,0.2347438782,0.0108646406,-0.1453997791,0.065800488,0.0513233878,-0.007699105,0.2702140212,-0.083414197,-0.1498301774,-0.0395016409,-0.1611671597,0.057236433,0.1878028661,-0.1946089119,0.1414126605,0.2921893895,-0.1358785927,-0.3849092722,0.1558649242,-0.1216332763,0.046303656,0.4255721271,-0.2424821705,-0.3623389006,0.1913564205,0.3384474814,-0.166639775,-0.1054475158,0.2395012081,0.106362395,-0.5258744955,0.0544809513,0.1539388746,0.471028477,-0.0193930641,-0.0037683111,0.2319384068,-0.7504907846,-0.1091424227,-0.3552944958,-0.2899658382,0.2494618744,-0.213390708,0.0785017312,-0.1959079057,0.0190292522,0.2596179247,-0.0680845231,-0.2487382144,0.2140538543,-0.204521358,0.0116548212,0.1629922241,0.1818397641,0.1322844923,0.0485198088,0.1022284403,-0.0928021371,-0.3593691587,-0.1395340115,-0.0055814916,0.0957579836,-0.0088076377,-0.1100880131,0.049773939,-0.175045982,-0.0145621346,-0.1571630538,-0.1505521238,0.2373260707,0.1356777549,-0.1125287116,0.1605954468,-0.0496461913,-0.2892830968,0.309197247,0.155746609,0.2655005455,0.2623186409,-0.1723350883,-0.2354275435,-0.0431110524,-0.1349603087,0.0186056327,-0.0405655243,-0.1381248683,0.2875163555,-0.0203582775,0.04281234,-0.0327242725,0.2053019255,0.0751455799,-0.1998059452,0.2819977999,0.4231081903,0.0853480995,-0.341481775,0.1376573592,0.1228301972,0.0577864908,-0.0528531596,-0.1804600358,-0.247385174,0.1940378845,-0.0186299197,0.0420614406,0.2001725882,0.1539037228,0.3345657289,0.4152137637,0.2362950593,-0.0660931543,0.319345355,0.1658211946,0.0652543828,0.4929676652,-0.1017900929,0.354551971,-0.626198113,0.3119459152,0.0161279086,-0.4276145399,-0.0734484121,0.1491994709,-0.0892674327,-0.1475449055,-0.1491669714,0.4075734019,-0.1782061607,0.0909321234,-0.1822098494,0.0398199074,0.0078866798,0.0420845859,0.1249311864,-0.3194990158,0.069099687,-0.0104809459,0.2825878263,-0.330052793,-0.1321210861,0.1035321057,-0.1019439697,-0.1543005407,-0.2264278978,0.1020388901,-0.0165676922,0.0065342872,0.1416335851,0.1948206276,0.0223910399,0.0422548763,0.18140167,0.3701476753,0.02525164,0.0633056611,-0.1117470711,-0.1309592426,-0.0337581933,0.0416672751,0.2496416867,0.0743027925,0.0741263703,0.3713853359,0.1213888079,-0.1566477716,-0.2803471684,0.0927968547,0.2560225725,-0.2445172817,0.4824366868,-0.5761265159,0.0055789584,-0.185095191,-0.0921313912,-0.4762138426,0.5811566114,0.2458808273,-0.384191066,-0.1372481883,-0.160936892,0.0534714833,0.01090327,0.4640223086,0.3801370263,0.4002385139,-0.2473380715,-0.5967437625,-0.7138707638,0.2029968649,-0.1287078857,0.0140654659,-0.0168941598,0.0979119092,-0.1094402596,0.3211934268,0.2787206769,0.0909260437,0.3800651133,-0.0781399608,-0.0895387679,-0.155993849,-0.2007278651,-0.0063408785,-0.1218645647,-0.1536656469,0.0666208565,-0.4474488199,-0.0313929282,0.1314276159,0.1380232126,-0.1137947366,-0.0584910549,0.4580972195,-0.0454882607,0.2134415805,0.0362046845,-0.034439005,-0.1763321459,-0.2321973592,0.2459519655,-0.023341082,-0.0381546244,0.0040175379,-0.2301503718,-0.2383047491,-0.2027824968,0.4307769537,0.1455104947,0.2988996506,-0.0486849807,0.0791117027,-0.0278558638,-0.0889398903,0.2718877494,-0.0235654637,-0.177682966,-0.0572475903,-0.1270515174,-0.1526107937,0.6423467994,-0.3883252442,0.1675477922,-0.1507979482,0.0455477685,0.0417219587,-0.2181846499,-0.7388224602,0.2959373295,0.3184669614,0.3756459951,-0.1593201011,0.144454211,-0.2346397787,-0.069799386,0.1374533921,0.0760496333,-0.0325060748,-0.0383271053,0.124202475,-0.0154372957]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2705","title":"404 not found error on loading WIKIANN dataset","comments":"Hi @ronbutan, thanks for reporting.\r\n\r\nYou are right: we have recently found that the link to the original PAN-X dataset (also called WikiANN), hosted at Dropbox, is no longer working.\r\n\r\nWe have opened an issue in the GitHub repository of the original dataset (afshinrahimi\/mmner#4) and we have also contacted the author by email to ask if they are planning to fix this issue. See the details here: https:\/\/github.com\/huggingface\/datasets\/issues\/2691#issuecomment-885463027\r\n\r\nI close this issue because it is the same as in #2691. Feel free to subscribe to that other issue to be informed about any updates.","body":"## Describe the bug\r\nUnable to retreive wikiann English dataset\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import list_datasets, load_dataset, list_metrics, load_metric\r\nWIKIANN = load_dataset(\"wikiann\",\"en\")\r\n```\r\n\r\n## Expected results\r\nColab notebook should display successful download status\r\n\r\n## Actual results\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.1\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":94,"text":"404 not found error on loading WIKIANN dataset\n## Describe the bug\r\nUnable to retreive wikiann English dataset\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import list_datasets, load_dataset, list_metrics, load_metric\r\nWIKIANN = load_dataset(\"wikiann\",\"en\")\r\n```\r\n\r\n## Expected results\r\nColab notebook should display successful download status\r\n\r\n## Actual results\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.1\r\n- Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nHi @ronbutan, thanks for reporting.\r\n\r\nYou are right: we have recently found that the link to the original PAN-X dataset (also called WikiANN), hosted at Dropbox, is no longer working.\r\n\r\nWe have opened an issue in the GitHub repository of the original dataset (afshinrahimi\/mmner#4) and we have also contacted the author by email to ask if they are planning to fix this issue. See the details here: https:\/\/github.com\/huggingface\/datasets\/issues\/2691#issuecomment-885463027\r\n\r\nI close this issue because it is the same as in #2691. Feel free to subscribe to that other issue to be informed about any updates.","embeddings":[-0.1983268559,0.0143738203,0.0021061967,0.236056,0.0338601805,0.0631663427,0.054253187,0.1971953064,-0.0232596956,0.1844016463,-0.1990168989,0.3482315838,0.1889100224,-0.0352049246,0.1387394071,-0.2517963946,0.029693706,0.263538152,0.0581075698,-0.097589165,-0.2116294056,0.1820748001,-0.4031837285,0.2455023229,-0.143373549,-0.0048098648,0.027901195,0.109960407,-0.3365817666,-0.2690229416,0.3799149394,-0.2225337178,0.3279035985,0.6063837409,-0.0001109609,-0.0759147629,0.4458803535,0.1000097468,-0.391836077,-0.4055833817,-0.4237225652,-0.0974915251,-0.0088995369,-0.2199463993,-0.033632338,-0.3518974781,-0.0263598748,0.0464493744,0.1087762266,0.408013314,0.2161208391,0.3451708853,0.1994796693,-0.0758984461,0.1112233698,-0.2041801065,-0.2938869298,0.2485886812,0.3716220856,0.0193181764,0.2147423625,0.4076726735,-0.012935644,-0.1191930324,0.0002604528,0.0962099433,-0.2352766991,-0.3271325827,0.1940369904,0.4193383157,0.7381356955,-0.0773248747,-0.3165581822,-0.1083422005,0.1998565197,-0.1682266593,0.2259471714,-0.0250789039,-0.1283596456,0.2104655802,0.1944339275,-0.3695300221,-0.1503367275,0.2880272865,-0.0677539706,0.6986863613,-0.0293509737,0.0186183192,-0.10053204,-0.2014934719,0.6673569679,-0.079500705,0.107167244,0.0952379629,-0.1010599658,-0.007474252,0.0782534406,0.270442307,0.0974567384,-0.5342894197,0.0890707225,0.071390301,-0.1726489216,0.1293272525,0.0966725275,0.4672059119,0.113636516,0.059606567,0.3231092989,0.5634970069,-0.0561737493,-0.0862273499,-0.2202486843,0.036675714,0.0765397027,-0.1212231293,0.2347895652,-0.2797554433,-0.1778755784,0.1110151708,0.0104963854,-0.1117839292,0.1340745986,0.3343401253,-0.0520977639,0.4685741365,-0.1090251505,-0.0353921875,-0.2961293459,-0.0141351288,-0.1227805093,0.2299466133,-0.2976323366,-0.0304367337,0.0177038424,-0.0002731988,0.3745611906,-0.1093472764,0.2685654461,-0.0799273401,0.0472409688,-0.0589807145,-0.1680402756,0.2072970122,0.1856461763,0.0938892514,-0.0393015221,-0.5908218622,-0.0037347227,0.2500285208,-0.5024211407,-0.111079976,-0.3087926805,0.207456544,-0.0394152291,-0.0953390598,-0.2219092399,0.0893992707,-0.1252564192,-0.2733161449,-0.0483465157,-0.1564322859,-0.0715266392,-0.0714866295,0.0706887245,0.1540295482,-0.3954985142,-0.0663112924,-0.0150679545,0.0924511626,0.0102316067,0.2509191632,-0.289742589,0.1805291474,-0.4087093472,0.2994852662,0.4748595357,-0.6277826428,-0.4702953398,-0.1551852077,-0.2183941603,0.1764796674,-0.1732209921,-0.1397213638,0.0623195209,-0.0246238504,0.3741478324,0.2529842854,0.181484431,-0.1765975654,-0.3372808397,-0.262729913,-0.1873220056,0.0544974245,0.2701176405,-0.0965783149,0.2220113426,0.09778779,0.2185599208,-0.3049603403,-0.0352812856,0.3708949983,0.2777453661,0.0029751016,0.0427017324,-0.1631835401,-0.2401775122,0.2356497496,-0.0651689544,0.2263505906,-0.1303499937,-0.1732662022,-0.5030014515,0.0282157417,-0.4124868512,0.0118033681,0.202769503,0.0639729351,-0.0951892212,0.3821229041,-0.1254786551,0.0553834289,-0.0648604855,0.0590664372,-0.1425856501,0.3004005551,-0.2772384584,0.0521437638,0.0099659944,0.3502000272,0.3246391714,0.0650507808,-0.058719255,0.141134426,-0.2738318741,-0.0084212022,0.6433116794,-0.0843106881,0.2559211552,-0.4660770297,0.2186956257,0.0654229969,0.184139654,0.0361105651,-0.0412480719,0.2195616364,0.2949728966,0.1033336893,0.164579019,0.0750200525,0.3924272358,-0.2022890002,0.0477028862,-0.3806104958,0.1593785882,-0.2764974833,0.1574205905,-0.1935892254,-0.206379056,0.2418937832,0.4138443172,-0.2443424612,0.2266424745,0.1589666158,-0.2687552273,-0.057376679,0.2165173739,0.3593724072,0.3308875859,0.157457009,0.3751510978,0.0756046474,-0.0619225949,-0.0286794025,0.2185885161,-0.0385723971,0.1342400312,0.2594102025,0.2180501372,0.0059732315,-0.3377846181,-0.2373338938,-0.1496368498,0.2096709609,-0.0474483185,-0.0373308845,-0.2840379775,-0.2394461334,-0.0997332111,-0.1130313873,-0.1936632395,-0.4324688613,0.0376594253,0.0136799458,0.1823806614,0.1621531695,-0.1680104733,-0.0975627005,-0.0643109456,-0.1069126651,-0.3256938457,-0.0213216618,-0.4002248347,0.0550165251,0.1399589479,0.0118893851,0.0437359549,-0.4289552569,0.0444942527,-0.3928014338,-0.4904257655,0.3152220845,-0.006297125,0.3915443718,-0.0743702799,0.2267785817,-0.0817342699,-0.1939721107,0.2416514754,-0.3259669244,0.0569712035,-0.0268688221,-0.081177406,-0.0201227218,0.1438263357,-0.2476803511,-0.3624427021,-0.098898232,-0.3223589957,-0.1130257398,-0.0283177085,0.166384384,0.0870398358,0.2570314407,-0.0655414239,-0.1978173852,-0.2999257743,-0.2391173542,0.4530379772,-0.2117198259,-0.2243974507,0.2539779544,-0.1193230674,0.2886121273,-0.2418434769,-0.5117617249,-0.4658155143,-0.0748426914,0.4123660028,0.0936156511,0.0990675166,0.2016833872,0.0559695065,-0.0554729663,-0.0769423917,-0.1939579695,-0.3744467795,0.0871133506,0.4336403012,-0.0294841379,0.0110902572,0.0580509566,0.4314289093,0.1037614644,0.3736900389,0.4079248607,0.0739920959,0.4846547842,-0.1832888722,-0.1795473844,-0.004560899,-0.1673390418,-0.184282124,0.1388657838,0.3608135879,-0.488489151,-0.3888809979,-0.2070564628,-0.5439021587,-0.2546105981,-0.2533851266,0.3486425877,-0.0121650146,0.2779347897,0.0811950192,-0.1389020681,-0.1706612557,0.0334696844,0.4500019848,-0.1177516282,0.0116634089,-0.3267324865,-0.0038246622,-0.5948861241,0.1245976537,-0.0487169251,0.4634473026,0.0193504766,0.1649514735,0.1243016496,-0.139349848,0.1326625198,-0.5171256661,0.0715036616,0.0174520761,0.2944749892,-0.1008058563,-0.0299444664,-0.1241526529,-0.0218142234,0.0883306786,0.3394628763,-0.196177125,-0.054536961,0.428260535,0.3095884323,-0.0725229532,-0.0278758667,-0.2378898412,-0.6871281266,-0.0488939248,-0.1485823393,-0.0068676337,0.1395158619,0.0660225376,0.0056035332,0.0002006118,0.1595852673,-0.012886853,0.2413861305,0.1181753948,0.1687461734,-0.0836929604,0.0870018676,-0.0382888094,0.2488385439,0.4720655382,-0.1361257136,-0.5144059062,-0.2439080775,-0.0887471884,0.0590641834,0.2464939952,-0.1293631792,-0.232795909,0.0409936868,0.1670073569,0.1961424947,0.2150189281,0.3455315828,0.2300862074,-0.5121203065,-0.0948594883,0.3955667317,0.0403892621,-0.0215681177,0.4098514915,0.1562862545,-0.2189685404,0.1863830537,-0.0732676908,1.0719259977,-0.0911128744,0.2113978863,0.3607043624,-0.1690214276,0.0808518156,0.0697213188,0.0989099741,-0.1443689466,-0.2136272043,-0.1377226561,0.0493165739,0.0133374119,-0.2496158779,-0.1779064685,0.468197912,0.0500933006,0.2999769747,0.3743363917,0.212405771,-0.1332149953,-0.0835406631,-0.1145633459,0.1167285666,-0.1161132529,0.1636296958,-0.2745772898,-0.0361177698,-0.0691803098,-0.127829805,-0.2191946357,0.1846353561,-0.2150389254,0.0012123552,-0.2951925993,-0.6971206665,-0.0881364942,0.4381736219,-0.081420809,0.1437567323,-0.3321286142,0.3885343373,-0.1205431148,0.0194389634,0.043305289,-0.4239307046,0.0485433675,-0.0740098208,-0.4560860693,0.046534434,-0.002385373,0.086826995,-0.1821135581,0.0933426321,0.0025488669,-0.0098118177,-0.1035377309,0.1148661599,0.1216752604,-0.4113838375,0.1354814172,-0.1152538583,-0.0680663884,-0.0492682606,0.2187410891,-0.2328518927,-0.0105873728,0.265933156,0.0526386015,-0.1703009307,0.2356355488,0.0302280057,-0.2531355023,-0.1517297626,0.0054129027,-0.1749428213,-0.3292524219,-0.1728736013,0.049993325,0.0707301348,-0.2453925014,0.2282730341,0.1984298974,-0.247692436,-0.1196608692,-0.4557782114,-0.0702848807,0.2636732161,0.0156579353,0.1252758652,0.0351103134,-0.1545594484,0.0171417948,-0.1899527311,-0.3249895275,0.1302820444,-0.1726169735,-0.100245595,-0.0611154549,-0.0010172212,0.2794394195,-0.0123632634,0.1607062817,-0.068574369,0.050645519,-0.1711769849,-0.0407179967,0.1101030037,0.1396259367,-0.0997912139,-0.0133232577,-0.2990199029,0.1617022157,-0.2171314657,-0.018371379,0.2007293552,0.0666030422,-0.0076038139,0.2076287866,0.1514821351,-0.0018377332,0.3704827726,0.1704530716,0.2995701432,0.0592878014,0.0869315788,0.1470436752,0.1585347503,0.2571714818,-0.1194196716,-0.1919034123,0.4616591036,0.3368369639,-0.3351193368,0.1151582077,0.1822772622,0.1819808483,0.2565379143,-0.235110119,-0.1691006124,0.3897414207,0.2337058485,-0.4088411331,-0.2705039382,0.705108285,0.1384460032,0.0195328798,0.1763795614,0.1471144706,-0.1689489335,0.0990509242,-0.0430023707,0.4494205415,0.1409205794,0.2215243727,0.2571344376,0.0046798685,0.1889474839,0.2908563614,0.0605816841,-0.0078088148,0.5327864885,-0.0477702916,0.3120147884,-0.0900687277,0.5175080895,-0.3650618196,-0.1609176695,0.1385196298,-0.2704819143,0.0875794441,0.0842292905,0.0132957948,-0.1740366966,-0.034923058,0.1546515673,-0.1705435067,0.0971435308,-0.2005712241,-0.0337072648,-0.1742654592,-0.3630827963,0.3250043988,-0.0705805123,0.0345461667,0.0586424731,-0.2898458242,0.142387718,-0.0971452892,-0.0924586058,0.0728584379,0.3994553089,0.1110661775,-0.0131530473,0.3374659121,0.320985347,0.2688960731,-0.3577322066,0.3398103118,0.3719190955,0.1348428428,-0.2798978984,-0.0526648797,0.0249109976,-0.1245267391,-0.0772819147,0.0438013561,0.1628828943,0.0021946826,0.2933055162,0.1943153292,-0.2242334187,0.3932958245,-0.0111729922,-0.033810433,-0.0537447445,-0.4464098513,-0.0781536549,-0.0792448223,-0.3510479927,-0.0094352765,-0.2264390886,0.6084119081,0.4803203344,-0.1324391365,0.0656326339,-0.1473210901,0.0700647235,0.100764513,0.186669603,0.5567760468,0.0488675423,-0.1949884444,-0.2805478275,-0.8394197226,0.2146116644,-0.4184160233,-0.2192976922,0.1283463389,-0.1578165144,-0.1890799403,0.1042639837,0.4804941714,0.4738730192,0.1469684392,0.2053226978,-0.3954737484,-0.3286801577,0.2986510992,0.0961936638,-0.0835272372,-0.2910257578,-0.0635521561,0.0593112297,-0.0091490755,-0.0210535042,0.0002320054,-0.273673445,-0.478951335,0.6185747981,0.2002348751,0.1302377582,-0.2066409886,-0.2219355106,-0.0627419874,-0.231532827,-0.0579645,0.2054094821,0.1275550574,0.1005102322,-0.0311248694,0.006656073,-0.2374490947,0.1402210444,0.1548164785,0.1914416254,-0.2015694976,0.1249348149,0.0730817765,0.0456830226,0.1138535663,0.3327821493,-0.185922578,0.2362201363,-0.0052302298,-0.4675925374,0.4862321317,-0.0227631815,-0.3595177829,-0.0921959281,0.1079111695,-0.0812979043,-0.1136266813,-0.125821203,0.1503780484,0.267208159,0.1806958467,-0.245393008,0.1420566589,-0.0677608326,0.1050696522,0.0141630014,0.4371634424,0.1009191573,-0.2681321502,-0.025753323,-0.0293717962]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2700","title":"from datasets import Dataset is failing ","comments":"Hi @kswamy15, thanks for reporting.\r\n\r\nWe are fixing this critical issue and making an urgent patch release of the `datasets` library today.\r\n\r\nIn the meantime, you can circumvent this issue by updating the `tqdm` library: `!pip install -U tqdm`","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nfrom datasets import Dataset\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n\r\n---------------------------------------------------------------------------\r\nNOTE: If your import is failing due to a missing package, you can\r\nmanually install dependencies using either !pip or !apt.\r\n\r\nTo view examples of installing some common dependencies, click the\r\n\"Open Examples\" button below.\r\n---------------------------------------------------------------------------\r\n\r\n## Environment info\r\n\r\n- `datasets` version: latest version as of 07\/21\/2021\r\n- Platform: Google Colab\r\n- Python version: 3.7\r\n- PyArrow version:\r\n","comment_length":39,"text":"from datasets import Dataset is failing \n## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nfrom datasets import Dataset\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n\r\n---------------------------------------------------------------------------\r\nNOTE: If your import is failing due to a missing package, you can\r\nmanually install dependencies using either !pip or !apt.\r\n\r\nTo view examples of installing some common dependencies, click the\r\n\"Open Examples\" button below.\r\n---------------------------------------------------------------------------\r\n\r\n## Environment info\r\n\r\n- `datasets` version: latest version as of 07\/21\/2021\r\n- Platform: Google Colab\r\n- Python version: 3.7\r\n- PyArrow version:\r\n\nHi @kswamy15, thanks for reporting.\r\n\r\nWe are fixing this critical issue and making an urgent patch release of the `datasets` library today.\r\n\r\nIn the meantime, you can circumvent this issue by updating the `tqdm` library: `!pip install -U tqdm`","embeddings":[-0.4085453153,-0.1528589129,-0.1239814758,0.0641506314,0.1121254042,0.1035053059,0.3430987298,0.223898381,-0.1365936399,0.0548991077,-0.0883905292,0.2876406014,0.0480185561,0.0611997545,-0.1352098435,0.0198291205,-0.0553882942,-0.0109331338,-0.3337739706,0.1266524345,-0.217322588,0.3459280729,-0.3196708858,-0.0665076375,-0.26490587,-0.2171354145,-0.0884475261,0.2700223327,-0.0890713185,-0.3350606263,0.0057384246,0.140682146,-0.0379140452,0.5401774645,-0.0001011049,-0.0078354618,0.3128858507,0.0935639068,-0.084417522,-0.0710472539,-0.3134652674,-0.4694289863,0.0806844607,-0.1916164458,-0.0022029711,0.1091363952,-0.0123654772,-0.3157416284,0.3191045225,0.3372055292,0.3364604712,0.5836625099,0.001960984,-0.306732744,0.0630692542,-0.3369416893,-0.1633457243,0.2865929902,0.4009721577,-0.1956418753,-0.0135743497,0.1969815046,-0.213966459,0.212846294,0.0626571998,0.0328398421,-0.0416467525,-0.4304128885,0.2624842227,0.1870162934,0.5149672627,-0.3271745145,-0.0373378322,-0.0568678826,0.0038167925,-0.2325145602,0.2190859914,-0.036922235,-0.1316181868,0.2236550599,-0.036217466,0.071047999,0.0526873171,0.0384646766,-0.2217678726,-0.0460676551,-0.0829893947,-0.0408100374,-0.0412158743,-0.136370182,0.3713529706,-0.0078106704,0.0284169465,-0.0238212813,-0.5177696943,-0.0575567298,0.3645920157,0.03556858,0.0987452865,0.0120878918,-0.0397548415,0.1094594598,-0.0714503527,0.3100284934,0.1683249623,-0.1073323935,-0.0055003739,0.0794885755,0.282115221,0.188070327,-0.2165435702,0.0819282457,0.0630704537,-0.4723846912,0.2384412438,0.1042847112,0.2806369662,0.1072816551,-0.3729566634,0.045274362,-0.0440311693,0.1376920044,-0.0906591117,0.4109575748,-0.109149456,0.0150475614,-0.0427163467,0.0600475036,-0.2108016759,-0.3220811188,-0.2802446783,0.0243945587,0.0463382415,-0.1378921568,-0.0416072346,-0.1625710577,0.2116598338,0.0482912175,0.157229349,0.0954252183,0.1668868065,-0.27029562,-0.0576179698,0.40771088,0.192671597,0.0409890004,0.2395561785,-0.0955689326,-0.0758202076,0.0929865688,-0.2596942186,-0.1863327324,-0.0069954796,0.3259136379,0.103960976,-0.2556672096,-0.0916619077,0.1133576035,0.0773157701,-0.1863555759,-0.0654879883,-0.2524957061,-0.0674690306,-0.3504024148,0.0776730925,0.5380800366,-0.4327085614,-0.0149561055,-0.159262374,-0.1078128815,0.0124575635,0.2770961225,-0.0810049549,0.1695788801,-0.1159966663,0.0491245463,0.2401681542,-0.319914937,-0.3626366258,0.0908700749,-0.301079005,-0.0427362472,-0.1367135346,0.093721427,0.295752883,-0.0094474945,0.4482402205,0.2462100089,-0.2692749798,0.0903337151,-0.1365799308,-0.2909478545,-0.0536916591,0.0469712801,0.2362893969,-0.0899253488,0.2524038553,-0.2295201272,0.1484709233,0.133948043,-0.0187071133,0.3950461149,0.3054838181,0.0365398787,-0.0323319994,-0.1834006906,-0.5057476163,0.249874711,0.3089755476,0.0818637311,-0.0234779958,-0.0089072566,-0.4486204088,0.2037529796,-0.1624403596,0.0956389755,0.2882280052,0.0894581974,-0.2714303136,0.0509770662,-0.247302711,0.4360179007,0.0321619958,0.08695364,-0.1087867022,0.288949877,-0.2754172385,-0.2581116557,0.2147305906,-0.0580091849,0.247653529,0.0031081075,-0.0581927523,0.329310447,0.0771936849,0.2450991124,0.1307125241,-0.0577195324,0.1969685405,-0.2697572708,-0.095454812,0.0938056782,0.1041573212,-0.0734270588,0.0805765018,0.1706543565,-0.2142150551,0.0104886293,0.0348369218,0.2288723588,0.3592408895,-0.0549508706,-0.0541916639,-0.1366294175,0.2221429497,0.118401818,0.012584161,-0.1474806666,-0.2470891625,0.2493074983,0.490115732,0.1015379876,0.1599902362,-0.1755920053,-0.2158996016,0.0977206901,0.133570075,0.3544579446,0.4767529964,0.2372669578,0.1458981931,0.181716755,-0.0352559723,-0.2257147729,0.0278197397,0.2088332325,0.0691447631,0.1396366805,0.2791813016,0.0575675294,-0.2701753378,-0.3984820545,0.0493996665,0.23472175,-0.3226417601,0.0203030147,-0.3022142053,-0.0669246912,-0.0339748487,-0.1841361821,-0.1792141944,-0.4988186955,0.0177649166,0.1733527333,-0.0362730883,0.2254777551,-0.0886924714,0.0382607728,0.1581800878,-0.234258011,-0.0838016868,-0.2235110551,-0.085879609,0.2016251981,0.1225927621,0.1275233775,0.4814586937,-0.3473612666,-0.0764092579,-0.3007541299,-0.2729535699,0.1080874056,-0.1079408526,0.5393425822,0.1690939814,0.1429461837,-0.189146474,0.0837831348,0.2729692161,-0.2841287255,-0.0361449495,0.1139917225,-0.0069278525,-0.1247009039,-0.0059449803,-0.2854000628,-0.259416014,-0.5138936043,-0.1164918765,0.0366632454,0.0855601579,0.2803019285,0.1171813011,0.25910905,0.0950798318,0.1597633809,-0.157387957,-0.27479285,0.2889837027,-0.2669566274,-0.4839946032,0.2523922622,-0.0436241478,0.248827219,0.0183746628,-0.291695416,-0.347537905,-0.0735730305,0.4598746002,-0.0146951051,0.139476046,0.2242235839,0.1021410152,-0.2486859262,-0.1503612399,-0.193666473,-0.2438552231,0.2690796256,0.0429665484,-0.0695938766,0.4731346071,0.1293032914,0.5410220623,0.2028915733,0.1337266862,0.3551773429,-0.3133036792,0.4701182246,-0.0137464684,-0.4632149339,-0.2093192041,-0.156620115,-0.2196600139,0.1839844286,-0.1219008714,0.00834483,-0.2008323371,0.0313070118,-0.3126885593,-0.2886296213,-0.109926559,-0.0378073715,0.298792392,0.1247180253,-0.0077970163,-0.173063606,0.076629214,0.0802605599,0.351816088,-0.3658187389,-0.2055609971,-0.0920197293,0.0391973183,-0.5612569451,0.3142528236,-0.0376316421,0.3010166883,-0.1695301682,0.0466060266,-0.021070946,-0.136732772,0.1865988523,-0.2707375586,-0.0610268302,0.3763122559,-0.0375204869,-0.2834613323,0.0316957608,-0.303586781,-0.0872135833,0.5882400274,0.1496819705,-0.2319242954,0.0843925253,0.179876864,0.2301519066,-0.0485718325,-0.030697694,-0.3578571379,-0.6081138253,-0.4169081748,-0.0317580178,0.1946903169,0.1478212327,0.0366777033,-0.1218839437,-0.0814928487,-0.2642626762,0.1014252678,0.2987629771,0.0152712651,0.0514461584,0.3791185617,-0.0545986556,-0.308526963,0.4553217292,0.5644432306,0.0935865715,-0.3636811376,-0.0350371301,-0.029774541,0.1475262493,0.1547461301,-0.1215577498,0.0153696928,0.0256638806,0.1582846195,0.3126121163,0.0927452371,0.3412407339,-0.0211885609,-0.104129754,-0.1698946506,0.1131926626,0.1280115098,0.0611299239,0.2848446071,-0.239072606,0.0123010492,0.0189965274,0.0477190316,0.6743601561,-0.005851713,-0.1335733384,0.3497028351,-0.2442308664,0.2312913835,-0.0070458991,0.1458289027,-0.3999635279,-0.418032825,-0.0113902446,-0.140847519,0.3766775131,0.0181208048,-0.2749498785,0.233726114,0.0567082353,0.1694542915,0.004825592,-0.0859140232,-0.1994823515,-0.2102622092,-0.3424766958,0.2780530453,0.0304985475,0.1002307236,-0.2302431166,-0.0756507814,-0.0022717945,-0.1145536751,-0.2208555341,-0.0430641137,-0.3907533288,0.1365590692,0.1541351676,-0.1798406839,0.0128810378,-0.0068833698,0.1035193577,0.0739508048,-0.1130787283,0.1491711438,0.1311754882,0.0251975935,0.1456773132,-0.1355884671,0.3613076806,-0.1701887846,-0.1081273407,0.2956542075,-0.0282272063,-0.1445137113,0.1549284756,0.3218530715,0.0484354459,-0.0987300277,-0.0249714516,0.0726727769,0.2177845687,-0.4324820638,0.2274699062,0.1098215654,0.0280082282,0.4482869506,-0.2745149732,-0.075728409,-0.169284597,0.2780625522,-0.0978034586,-0.0683534667,0.2635719776,0.335036844,-0.2140742093,-0.2988041341,0.3023134172,-0.2714876533,-0.115942575,0.2750825882,-0.1219210848,-0.0068427376,0.1769754142,0.5146286488,0.0672650337,-0.1476502568,-0.1188546121,-0.2926462591,-0.3343289495,0.0866132006,0.0571965165,0.0675611272,0.171148926,0.2797250748,0.0978310928,0.0337822102,-0.4624916911,0.0146880988,-0.4128829241,0.0474854559,0.2357289195,-0.0974387228,0.1013843641,0.0244064629,0.1799715012,0.1052236483,-0.1337603182,-0.3522090912,-0.0336403474,0.0878045037,-0.0039440109,0.0217030104,0.0412187614,-0.196362108,-0.2757995725,-0.1250045598,0.096073173,0.0712824911,-0.1034788564,-0.0618352704,0.1638186276,0.0417050309,-0.1251419783,0.096534349,0.0480588563,0.1241119206,-0.0932389796,0.0561798923,0.0966351628,-0.0530851297,-0.0538892932,0.1123823076,0.0803088844,-0.0846450478,0.4036474526,-0.2934017479,-0.0837887079,0.2388861924,0.4376217127,0.2970944345,-0.1042974964,-0.1326535344,0.131798923,0.2998235226,-0.4269777834,-0.0797768682,0.2276557535,0.031215461,0.2894967794,0.1620601416,0.0091705918,0.0803849995,0.1102855653,0.2927851677,0.3749345243,-0.2685065567,0.0089510148,0.0542334691,0.0661037862,0.162081942,0.3600085378,0.1112756655,0.1788648218,0.4138869047,-0.0186649356,0.1593809128,0.063314043,-0.1102110222,0.0784742981,-0.4868064225,0.376078397,-0.0723925307,0.1060393751,-0.1730608791,0.0226887278,0.0955680162,-0.2848626971,0.1331896782,-0.2175019979,0.3178294599,-0.1695959568,0.2080058753,0.2158291936,0.047075808,-0.004189122,-0.0685386956,0.1173429191,0.0952015519,0.3830111921,0.1151951328,0.0234845281,-0.5718722343,-0.1943042129,0.1402424425,-0.050019484,-0.3423666954,0.0659656227,0.0148209548,-0.0683904216,-0.2042538673,0.4518049955,0.4655418396,0.1875554472,-0.0217461139,-0.0537224747,-0.0694717914,-0.1414836198,-0.1546762139,0.1211891323,-0.0825648829,0.2094569057,0.3394080997,0.251416415,-0.3259422183,-0.1645158678,0.1148011163,-0.163451612,-0.3844025731,0.2955069542,-0.1409961581,-0.1607819945,-0.1595876366,-0.0308328662,-0.3114374578,0.0526956059,0.4397566617,-0.2856874764,0.0353569277,-0.2400782108,0.1659233272,-0.0764081478,0.3434974849,0.3479082286,0.0946846381,-0.3236079514,-0.160908699,-0.8212377429,0.2509789169,-0.1219769567,-0.2337376624,0.0354433432,0.2488986105,0.054354459,0.3188182414,0.3575403094,0.0408062376,-0.0032815549,0.1275628209,-0.3231613934,-0.2139832377,-0.0753674284,-0.2632588744,-0.0036477232,-0.3058440089,0.1710049212,-0.2677228749,0.1970319003,-0.2112950236,-0.4058314562,0.2146857232,-0.0597279854,0.4275778532,0.0655835941,0.3013673127,-0.2161146551,-0.0543659404,-0.1791131496,-0.3456138372,-0.1444771886,0.2212970406,0.1074012965,0.2834338844,-0.0556844175,-0.1634288877,-0.1055841818,0.1816213429,0.1836780608,0.0489662662,-0.0327337161,0.2775433362,-0.0131125534,0.2347469479,-0.0551852286,0.2331456691,0.0064923572,0.3684497774,-0.4001820683,-0.534904182,0.2997890711,-0.327001214,-0.4098891318,-0.3604010046,0.2793195844,0.0132646495,-0.1624010801,-0.1239522994,0.0338407531,0.2871501148,0.0694511458,-0.1881322563,0.4414680004,-0.0199173316,0.0468638577,0.0082707414,0.2054447234,0.0405816101,-0.2055842727,0.1134549603,0.0200569611]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2699","title":"cannot combine splits merging and streaming?","comments":"Hi ! That's missing indeed. We'll try to implement this for the next version :)\r\n\r\nI guess we just need to implement #2564 first, and then we should be able to add support for splits combinations","body":"this does not work:\r\n`dataset = datasets.load_dataset('mc4','iw',split='train+validation',streaming=True)`\r\nwith error:\r\n`ValueError: Bad split: train+validation. Available splits: ['train', 'validation']`\r\n\r\nthese work:\r\n`dataset = datasets.load_dataset('mc4','iw',split='train+validation')`\r\n`dataset = datasets.load_dataset('mc4','iw',split='train',streaming=True)`\r\n`dataset = datasets.load_dataset('mc4','iw',split='validation',streaming=True)`\r\n\r\ni could not find a reference to this in the documentation and the error message is confusing. also would be nice to allow streaming for the merged splits","comment_length":36,"text":"cannot combine splits merging and streaming?\nthis does not work:\r\n`dataset = datasets.load_dataset('mc4','iw',split='train+validation',streaming=True)`\r\nwith error:\r\n`ValueError: Bad split: train+validation. Available splits: ['train', 'validation']`\r\n\r\nthese work:\r\n`dataset = datasets.load_dataset('mc4','iw',split='train+validation')`\r\n`dataset = datasets.load_dataset('mc4','iw',split='train',streaming=True)`\r\n`dataset = datasets.load_dataset('mc4','iw',split='validation',streaming=True)`\r\n\r\ni could not find a reference to this in the documentation and the error message is confusing. also would be nice to allow streaming for the merged splits\nHi ! That's missing indeed. We'll try to implement this for the next version :)\r\n\r\nI guess we just need to implement #2564 first, and then we should be able to add support for splits combinations","embeddings":[-0.508149147,-0.3373530209,-0.1102792025,-0.0140300198,0.0330486335,0.0741378814,0.0743084773,0.4854656756,-0.0663450882,0.1295401752,-0.2676171958,0.231549263,-0.0514129102,0.5110136867,0.030207349,-0.3258370161,0.1286161244,0.0924145356,-0.3972883523,0.2039903551,0.0156943426,0.1319799721,-0.2798555493,-0.3620299101,-0.3513978422,0.3136510849,-0.0559827052,0.0434776768,-0.1932207793,-0.1684154272,-0.0865589157,0.124164544,-0.0005709948,0.5355129838,-0.0001148048,-0.0435678847,0.5377911925,0.055512201,-0.1230915189,0.0520752966,-0.431738019,-0.0420410596,-0.0881977528,0.1587152481,-0.2271650732,0.1198302954,-0.1076327264,-0.1590533257,0.374304235,0.0940160677,0.0899581835,0.1013053209,0.0752255768,-0.2277186215,0.0327020101,0.040793106,-0.075575389,-0.0672890618,0.162594825,0.0533763878,0.1702077836,0.2050912976,0.3278670907,-0.2385341525,0.0529790446,-0.0781996325,-0.2163973451,-0.3731330335,-0.0300137717,0.2754046917,0.417173028,0.1806240976,-0.3277600408,0.0924969837,-0.0132379355,-0.4786670804,0.0948752761,0.0755033493,-0.1146731228,0.131660834,-0.5122500062,-0.2017859966,-0.2065028995,0.2524446547,-0.1153546125,0.6995693445,0.106439501,-0.0259962287,0.5007662177,0.0095833829,0.159890607,-0.0029389681,-0.2509086728,0.0968374833,-0.3748495281,-0.1022785455,-0.4304015934,-0.0613726676,0.0011968015,0.1479566395,-0.0523992777,0.0594878122,0.1386433691,0.100441426,0.4026205838,-0.1157404631,-0.0750735253,0.2515940964,0.2232343853,0.1167275831,0.2711423039,-0.0339947082,0.0628528893,-0.2844311595,-0.4857780039,0.0061910166,-0.0998201147,-0.1338426918,-0.1590296775,-0.3384344578,-0.1455878019,-0.0286702234,0.0967051238,0.0855218098,-0.2533042431,0.1140838787,-0.1045275182,0.2969155908,0.0216697063,-0.34279567,-0.1387792379,-0.0145829609,-0.2554706931,-0.0818123072,0.256593585,-0.2658787966,0.1501655877,0.3329008818,0.2861681879,0.261210978,-0.2194491923,0.0029604794,0.0976145715,0.3461606801,0.2332958281,-0.216462329,-0.1802343577,0.2645345926,0.3113659024,-0.0113179889,0.0223508533,-0.4372555912,-0.0999160856,0.1921676248,0.0101411976,-0.138602972,-0.1920822114,0.5086087584,0.0677137375,-0.1515938789,-0.1903993487,-0.0040798513,0.0967611223,-0.1116227433,0.3906470835,-0.0287303496,-0.487649411,-0.0572691038,-0.1783780605,-0.1254671514,0.2866674066,0.3830592036,-0.268416822,-0.3105136454,0.011701379,0.2296999991,0.2801138461,0.0356148295,-0.1620114148,0.235798195,0.1493208259,-0.0552922003,0.2690360248,-0.1544439793,0.1483695656,-0.0233156532,0.1964316219,0.5279390216,-0.147190541,-0.3320834935,-0.2287954837,-0.0443872958,0.25294137,0.2439569235,0.3917976916,-0.2100874782,0.1873402894,0.1677793115,0.1789098382,-0.1880307198,0.1267259717,-0.0284449384,0.1899458617,-0.0645640194,0.0734615028,-0.2169759721,-0.2115378529,-0.0437554717,-0.0524961725,-0.0564150847,0.5228712559,-0.180092141,0.0138315922,-0.34558478,-0.1949624568,-0.0930152833,0.1002580523,-0.1318469048,0.0416362882,-0.3252920508,-0.3905947208,0.0545199998,-0.063839443,0.0676630884,-0.0878763422,0.4246716499,-0.1991048604,-0.2303444594,-0.0079404404,0.2666900158,0.1580345333,-0.0920138285,-0.044987265,0.6052393913,0.0240022521,0.3310446441,-0.1366328299,0.0064038653,0.4401605427,-0.2197292894,-0.074711211,0.3792750537,-0.0911200345,-0.177829355,-0.3126198351,0.0908949599,-0.3767005205,0.0743770599,0.1943710595,0.0916932002,0.0833508298,0.0190159176,-0.2693000436,-0.1221593693,0.1034233496,-0.5749575496,0.0316056237,-0.1716374457,-0.3143163025,0.1330609024,0.2696282268,0.085392192,-0.1447693259,-0.2021974921,-0.146671176,-0.067652531,0.0505087264,0.2840463519,0.4919460416,0.2036001384,0.2562606633,0.281429857,-0.0278464518,-0.2152473629,0.3708739877,0.0582859963,0.0924872458,0.3329788744,-0.0673544779,-0.1376367062,-0.163920477,0.160938412,0.2921339273,-0.1012122259,-0.1680911481,-0.1960756928,-0.4404210746,-0.3401387334,-0.143660754,-0.2690732777,-0.1394087523,-0.4952287376,0.0970957577,0.2488377541,-0.1684311479,0.1095261723,0.0756291375,0.0112142088,-0.1120052487,0.2869213223,-0.0890295729,0.148209691,-0.0353774391,0.1844819933,0.1746096015,0.0170832444,0.1113235354,0.1222266406,-0.1454949975,-0.2003287077,-0.5325547457,0.1181261688,-0.1742520332,-0.0486428849,0.2068660706,-0.1350508481,0.2030083835,-0.5070753694,-0.0250827447,-0.1179828197,-0.0727683604,-0.0896503925,-0.0503468923,-0.1196975112,-0.4618372023,-0.6362450719,-0.249064371,-0.5404444933,0.2634406388,-0.2554797232,-0.1121743098,0.1065882742,0.0326893181,-0.1179726273,-0.0065717758,-0.429936856,-0.289236784,-0.058936514,0.40229249,-0.4461688697,-0.3850897849,0.0921236649,0.1072704196,0.2493472695,0.2030596435,0.0808837637,0.3455433249,-0.0296856202,0.0870420262,-0.0778548717,-0.0007338915,0.2698391974,-0.1681799293,0.0442819372,-0.1913453788,0.1416907907,0.0298002046,0.3293230236,0.5940420032,-0.0922732279,0.2602944076,0.1720615029,0.8503627181,0.2213719189,0.0338623449,0.0265253801,0.1010485888,0.2220971584,-0.1088563949,-0.2199733704,0.2636600435,-0.1620946378,0.0602046475,0.0874212608,0.0236197356,-0.4355467856,-0.2725833058,0.2202933729,-0.0500405878,0.0048694643,-0.0267521795,-0.2231230289,0.1030633301,0.0437374748,0.0180827715,0.3688739538,-0.0164264999,0.0451072715,0.1667733639,0.2742874026,0.1019531637,-0.3018216491,-0.1953759789,-0.1486048996,0.0285658818,0.0079724863,0.1342537552,-0.2335392833,-0.1523537338,-0.0622507706,-0.0333437547,0.7308999896,-0.4700448215,-0.1862261295,0.0893395096,-0.0316740274,0.0784576833,-0.300265938,-0.1464820951,0.060612414,0.5181647539,0.3313443661,-0.2025873959,-0.0085706217,0.4726244509,-0.1199384704,0.0429338627,-0.2671265602,-0.1636566073,0.222685352,-0.2541056573,-0.1159056872,0.3159099221,0.0911070183,-0.5133219957,-0.1697620004,0.1480538994,-0.3351945877,0.1450450271,0.1714393348,0.3846229017,-0.0806669965,0.1510412991,0.2846970558,-0.0170547776,0.3582703173,0.3160247803,0.4130516648,-0.3127906919,0.2898934186,-0.2952032685,0.6024835706,0.1973261386,0.0268833805,-0.0061180284,0.1774464846,-0.0012975316,0.0833522081,0.2481266111,0.2944794893,-0.0753845125,-0.2934322953,-0.2479617596,0.2437049896,0.0967630595,0.1570513099,0.3665709496,-0.0080958679,-0.2202667445,-0.1802087277,-0.0931445211,0.8311536908,0.3514553905,-0.0511215441,0.0822233632,-0.3420526087,0.1145344228,-0.2410764694,0.349049747,-0.0499596857,-0.1900230348,-0.0996515751,-0.0661770403,0.2365009487,-0.1206925362,-0.5448081493,0.344171375,-0.2958672643,0.4792375863,-0.1160143539,0.182606563,-0.1522690356,-0.1442124546,-0.2476084083,0.1151132733,0.0034280687,-0.0758744925,-0.0651074126,-0.2912391722,0.156796366,0.0651633739,-0.220771268,0.3643645346,0.0778070092,0.1872264147,-0.0820422992,-0.1057078913,0.0941797271,0.1624276787,0.1093289107,0.1021917686,-0.2916803658,0.2685702145,-0.1199024096,-0.0704733282,-0.35296911,-0.0182664189,0.1041846201,-0.3264843822,-0.2080943137,0.210782513,0.1126905307,-0.5092000365,-0.2521855831,-0.0672903955,0.0317982808,-0.0281098001,0.0520917065,0.2679309845,-0.1196016371,-0.3213804364,0.1105262339,-0.1861271709,-0.4356997907,0.1394524872,0.1880930364,0.0779349953,-0.2295378,0.3991298974,0.0167135857,-0.177470088,0.3303194642,0.1787612736,-0.0833467171,-0.1643690318,-0.2420061082,0.2020904422,-0.0776381195,0.1237071827,-0.2825100422,-0.2111569494,0.2983008027,0.1103871316,0.0154013159,0.1366528422,-0.1931951195,-0.1865273863,-0.1653412282,0.260104686,-0.2842327654,0.3302908242,-0.1422861814,0.4343191087,0.0736775324,0.2044568807,-0.2684248984,-0.0025951129,0.1188780069,0.1553227901,-0.2260901779,-0.3581096828,0.0110794995,-0.0739915371,0.042797599,0.2439488173,-0.1716938019,-0.1312972605,-0.0300478414,0.0900147855,0.3175625801,0.031015167,-0.111952126,0.0074076387,-0.4620462358,0.1675175279,0.2413915247,-0.1604738683,0.0731808022,0.046483133,0.7694675326,0.0026706778,0.0763214752,-0.0761987269,0.1711457074,0.1486604512,-0.0044563236,0.0859764963,0.1657081842,-0.1427558661,-0.0114714727,-0.0064311228,0.3535856307,0.0894694626,0.3557210863,-0.3032308817,-0.0499160103,-0.3494810164,0.4437382817,0.4431462586,0.1119622365,0.0514031723,0.0177690294,0.1722629815,-0.1517116874,-0.1266009808,-0.1626868546,0.223469466,-0.0318373032,-0.0154609801,0.3882978857,0.1365392804,-0.0452422649,0.3940407038,-0.1873163879,-0.1082058847,0.6243634224,0.458996743,-0.2730157673,0.079839021,-0.0518252924,-0.1712730229,-0.11285422,0.048309315,-0.1877721697,0.3462684751,0.4000031054,-0.0684837624,0.5021162033,-0.0291217566,-0.4155102372,-0.1350046247,-0.1447894126,0.3166702092,0.2044386417,0.3156459332,-0.2462448329,-0.4899681509,-0.2448236793,-0.0230597313,-0.0245444123,-0.0202680789,-0.1529137939,-0.2530764043,-0.1895494163,0.1706694961,0.1991119832,0.1580578238,0.6029378176,0.0041990434,0.0793794245,-0.216099292,0.2362811118,0.3822194636,0.3558267653,-0.1452438831,0.0076426333,0.3794394732,-0.2650616169,0.3523432314,0.1705774069,0.4073833525,0.2343980372,-0.1795746982,0.0125421388,0.1997924894,-0.2044632286,0.1789812893,0.2393650264,0.5440564156,-0.1517498642,0.1934889853,0.0924246386,-0.0880415216,0.2542188764,-0.282786727,0.3401773274,-0.3977621198,0.3002812564,0.1573112905,0.0152101917,0.1106261015,-0.068754144,-0.4797086716,-0.2704090774,0.3754624426,0.0007999588,-0.0935252458,-0.1560476869,0.074408628,0.2217357755,0.644482553,0.0328457169,-0.1067725644,-0.0711753964,-0.2070062608,-0.3353720307,-0.2207002491,0.0591468513,0.3860920966,-0.1026649252,0.2361465096,0.0716346875,0.2721897662,0.3447860777,-0.1758826524,0.2958407998,-0.0070240041,-0.3321448565,0.036543332,0.1879817396,0.1128573045,0.0031113771,-0.0522038303,0.2949273288,0.0324193202,-0.0187553484,0.1324229687,-0.0701425225,0.2797319591,0.1121880859,-0.2236459851,0.2178762853,0.3656515181,-0.0962721631,-0.0768083781,-0.2699718475,-0.1720238328,-0.4505991936,-0.0722967833,0.0526680201,0.4325370193,-0.0800550804,-0.1802476645,-0.1298143119,0.4174408913,0.013185367,-0.0718945488,-0.1770033836,-0.1322391927,-0.0359180644,0.0519061461,0.3032900989,0.4208347499,-0.2203768045,-0.191725716,-0.0738460869,-0.3431105316,0.467377156,-0.2923311293,-0.2383927256,-0.4524562359,-0.1654773653,0.6991724968,0.2340482324,-0.6319128871,-0.2952767015,-0.008873933,-0.3797499835,0.2310719937,0.5059241652,0.2337808311,-0.1236180663,-0.1809200794,-0.034640789,0.415389806,-0.188378796,0.0353063643,-0.2904878259]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2695","title":"Cannot import load_dataset on Colab","comments":"I'm facing the same issue on Colab today too.\r\n\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n 3 \r\n 4 from ray import tune\r\n----> 5 from datasets import DatasetDict, Dataset\r\n 6 from datasets import load_dataset, load_metric\r\n 7 from dataclasses import dataclass\r\n\r\n7 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n\r\n---------------------------------------------------------------------------\r\nNOTE: If your import is failing due to a missing package, you can\r\nmanually install dependencies using either !pip or !apt.\r\n\r\nTo view examples of installing some common dependencies, click the\r\n\"Open Examples\" button below.\r\n---------------------------------------------------------------------------\r\n```","body":"## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":111,"text":"Cannot import load_dataset on Colab\n## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nI'm facing the same issue on Colab today too.\r\n\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n 3 \r\n 4 from ray import tune\r\n----> 5 from datasets import DatasetDict, Dataset\r\n 6 from datasets import load_dataset, load_metric\r\n 7 from dataclasses import dataclass\r\n\r\n7 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n\r\n---------------------------------------------------------------------------\r\nNOTE: If your import is failing due to a missing package, you can\r\nmanually install dependencies using either !pip or !apt.\r\n\r\nTo view examples of installing some common dependencies, click the\r\n\"Open Examples\" button below.\r\n---------------------------------------------------------------------------\r\n```","embeddings":[-0.4603253901,-0.2148275971,-0.024375163,0.2682137489,0.1354128718,0.051994808,0.4973769188,-0.1074850261,0.1983702779,0.0849533305,-0.3266307712,0.4753232598,-0.1742698699,0.1487485617,-0.2584459782,0.0984701291,-0.0661861375,0.058203578,-0.2907032669,0.0956151709,0.0202920046,0.3308885992,-0.6335685253,-0.3468577266,-0.5786883831,-0.2100416869,-0.0277418792,0.168791458,-0.2702196836,-0.292011261,0.3478223681,0.1770586967,0.0921896547,0.5697174668,-0.0001172403,0.0715135038,0.3933092654,-0.0787535235,-0.2456898838,-0.3918715715,-0.3461531103,-0.3291964531,0.3686217368,-0.1073928103,0.1058271378,0.404165864,-0.1357311457,-0.133275196,0.2626952529,0.1352794766,0.1936342865,0.549197793,0.1630603671,-0.192492187,-0.1125865728,-0.193378076,-0.2020825595,0.7461523414,0.4174177051,0.0168563686,-0.0510248244,0.2604947388,-0.0992677361,0.2890874445,0.1862412095,-0.0334219635,0.1051012799,-0.4805828929,0.1006372571,0.1743771285,0.4806717336,-0.233426556,0.0443901122,-0.1220246106,0.2094738036,-0.2924674749,0.2619004846,-0.0985698551,-0.2118597776,0.0490080118,-0.0226965509,-0.0896168202,-0.0123074474,0.0164376963,-0.0933856592,0.3655072451,-0.1430076212,-0.1118264869,0.0698661357,-0.1096417829,0.5018094778,-0.0161823332,-0.0268674623,0.2063278407,-0.468000561,-0.0093295779,-0.098870948,-0.1408492327,-0.0516990833,-0.1014948189,-0.0361867324,0.0174666177,-0.1110087708,0.303483665,0.1848023385,0.1422498971,0.0347587094,0.0182988103,0.1226537004,0.2864284515,-0.2007629573,0.227394864,-0.259282738,-0.3118798137,0.1845084578,-0.029214235,0.2715788782,-0.0037106252,-0.319062084,-0.1818456352,0.0095681679,0.105042167,0.0446227603,0.3990536928,0.055795569,0.1057089195,-0.0346219428,0.0016938471,-0.3566534519,-0.2394378334,-0.0538450666,0.1936965287,-0.0642801523,-0.1152990013,0.0678034425,-0.1549987495,0.2601871789,-0.0124206394,0.3092051148,-0.0599603131,0.1036741883,-0.3025387824,-0.0606454387,0.331908226,0.1287349612,0.0872022063,-0.0623135045,-0.2055783719,-0.0778879523,0.2355169952,-0.4770620167,-0.1442362964,-0.0243145749,0.1429465264,-0.1339146942,-0.2244311124,-0.5850232244,0.1648343056,0.1447636336,-0.0812816098,-0.0849623531,-0.3062829673,-0.2330714166,-0.1535831094,0.1236880049,0.4482834041,-0.2309288979,-0.142314136,-0.2176906466,0.1205273345,0.1272705644,0.234823361,-0.2021541893,0.1736088246,-0.0962642506,-0.1171419844,0.490647763,-0.4326361418,-0.4671251178,-0.0203478895,-0.2660758793,-0.0052545536,-0.0897063687,0.0485790744,0.3577234745,-0.0669410899,0.5458390117,0.2629367113,-0.1390330195,-0.036604546,-0.1640905291,-0.4099052548,0.1294252127,0.0418195315,0.1834629029,-0.2177709639,0.203032732,-0.0533746518,0.1012382284,0.1859944463,-0.1211299896,0.3843335509,0.2429013252,-0.2223784029,0.0372738875,-0.1859572828,-0.4775209725,0.2240482867,0.409294188,0.0799716562,-0.0674993992,0.05132493,-0.4381961823,0.0791065395,-0.2594439685,0.0676428005,0.0994127542,-0.0666698813,-0.1751303822,0.0889260471,-0.2604830265,0.7403804064,0.0377074406,0.2175934911,-0.3183411956,0.2157882154,-0.1859039217,-0.3081798851,0.1762920618,0.1361715198,0.0753254518,-0.092272602,-0.2389075756,0.1421211809,0.2554138899,0.1713412106,0.061892774,-0.2776550353,0.4123665094,-0.274011761,0.0123248948,-0.1066840887,0.208570227,0.0058228662,0.2551185191,0.1565714628,-0.3707030118,-0.0410736091,-0.0305118244,0.0905382559,0.3726439178,0.1239964738,-0.0050928039,-0.1856487691,0.325183779,0.2222952098,0.192474246,0.0937420502,-0.1297319233,0.2881348133,0.3061757982,0.2452441901,0.038779363,-0.2685610354,-0.2554307878,0.0633648261,0.1922871321,0.2518091798,0.4446454048,0.1803891063,0.1049195379,0.0316309184,0.0318480134,-0.1561172903,0.1151199043,0.2919997871,0.2558412552,0.1311028153,0.4120824635,0.0967440531,-0.078998588,-0.3894120455,0.1753250211,0.1503109038,-0.3317721486,0.0766265243,-0.2146017402,-0.113947168,-0.1419459134,-0.4253391027,-0.1464539021,-0.1978348494,-0.1058005691,0.2810257375,0.2650622427,0.1936620623,-0.0239940546,0.0500169992,0.0625045747,-0.142483592,-0.0558168627,-0.2506769001,-0.0925878361,0.0015159415,0.327739507,0.246811226,0.3685063124,-0.3038308024,-0.0695626289,0.1437006891,-0.332098037,0.1763451695,-0.005682677,0.2919798791,-0.0134428795,-0.0779709816,-0.3206300437,-0.1939175874,0.0215363186,-0.1346231997,-0.1827463955,0.0472795218,-0.1800287366,-0.2297740728,-0.0019908124,-0.2926879525,-0.4344463348,-0.3998571932,-0.3682101369,0.0472466946,-0.0089030508,0.4603004754,-0.0427159443,0.2933925688,0.2250700146,0.0508381985,0.0179356467,-0.3330911994,0.4733818173,-0.2865113616,-0.2936578393,0.2801209092,-0.2680461109,0.3836089373,0.0181830842,-0.1104436517,-0.2135308832,-0.0214798469,0.3837370574,-0.1009398997,0.0416830219,0.0846360251,0.1136619002,0.0782582983,-0.092406489,-0.1815526336,-0.2040507346,0.2196319252,0.1930779815,-0.1657154113,0.2265984565,-0.0363357775,0.8421141505,0.0771993697,-0.0400505289,0.3015678525,-0.061655391,0.3064921498,-0.1064245924,-0.5121492743,-0.0696846843,-0.1073111445,-0.039284341,0.0819721445,-0.1839754134,-0.1112223938,-0.2847859561,-0.2727975249,-0.3160193861,-0.291852355,-0.1147850677,-0.2003910989,0.1308543533,0.1986790299,-0.0612915456,-0.1620080918,-0.1124967635,0.0489822105,0.4355244339,-0.4208028316,-0.0186146274,0.3795593679,-0.0728556365,-0.3464918733,0.1986530572,0.1317849904,0.3303270042,0.0766258389,0.1182002276,0.0601407997,-0.0493247956,0.1122457087,0.05919642,-0.0800003186,0.2366157919,-0.0146941952,-0.4012733102,-0.0577438287,-0.2255629599,0.2240988016,0.4793192744,0.0861525759,-0.2667843401,0.2115175128,0.2363318503,0.1412735432,-0.114222765,-0.1485226005,-0.4387521744,-0.3134378195,-0.378410399,-0.0600447692,0.1850457788,0.2391748279,0.0416690968,-0.0708694831,-0.0640273839,-0.2169730812,0.0610414371,0.3596898913,0.1482839882,0.270448029,0.5801549554,0.1271515787,-0.2473592013,0.4470187426,0.5865612626,0.0028329962,-0.5835172534,-0.0279992204,-0.1797391474,0.2348893136,0.1315195262,-0.0706735849,0.0163302217,0.0595659725,-0.1358518749,0.3848748207,0.3812657893,-0.0635423213,0.1262477934,-0.4026652873,-0.1285912842,0.0958481207,0.0007427831,0.1376991123,0.4601660669,-0.3796385229,-0.0598220304,0.047099065,-0.1843536198,0.686365068,-0.0381890498,-0.2390154898,0.2732535005,-0.1615026444,0.2652589977,0.1131904423,0.1857605428,-0.2762228251,-0.3076212406,0.0923932046,-0.2266215682,0.1900261045,-0.1945811212,-0.3511314392,0.2596141994,0.0564561076,-0.1498299241,-0.0185761657,0.0406053141,0.0024130014,-0.2937503755,-0.3525202274,0.0517692231,-0.1905304193,0.2061717808,-0.1972419322,-0.1833461672,-0.1389480531,-0.1865322441,-0.0331040733,-0.1090358868,-0.6445795894,0.2370983511,0.3314474523,-0.3087444007,0.2005167454,-0.1111385226,0.0950988606,-0.0293304492,-0.1739497632,0.1265123934,0.1993344128,0.0166862272,-0.1108888164,-0.172952801,0.4197721481,-0.0839659944,0.0050038416,0.0576662868,0.0366229266,-0.0225345157,0.4702349007,0.3512617946,0.3681568801,0.150692597,-0.0755758658,0.1810477078,0.26167503,-0.1958780736,0.0780422837,0.1019752249,-0.1530845314,0.2558067143,-0.107513763,-0.1003244594,-0.0019221834,0.3397724032,0.1296715736,-0.0161001608,0.1832213998,0.1551724821,-0.2558426857,-0.0844762251,0.1688349247,-0.3087925613,-0.3004136086,0.2412767708,-0.0199141074,0.0491270311,0.0744944811,0.7340957522,0.2456151992,-0.3685689867,-0.132078737,-0.3772931397,-0.5326016545,0.1417692751,-0.1205238625,-0.0949431732,0.2369502336,0.3836895525,0.120868288,0.1284108609,-0.2208905369,0.1052405685,-0.1584761292,0.1540038735,0.2756511569,-0.0960519463,0.0372927338,0.0859900042,0.0042420314,0.1529782265,0.0671002194,-0.2249542475,-0.1146583706,0.1326806545,-0.1025531739,0.0408711247,-0.1180261746,-0.1512853205,-0.2300352007,-0.1715027094,0.1590514779,0.0351236463,-0.0818419009,-0.1321561933,0.2920054197,-0.0966290534,-0.2174018025,0.4025671184,0.1177951023,0.0810081363,0.0776668489,0.1660300344,0.1665427387,0.1832786202,0.017530106,-0.2253397107,-0.091671519,-0.1765127629,0.0942778215,-0.3320890367,-0.0140226921,0.0699948967,-0.0275069047,0.4203799963,-0.128956452,-0.2099847645,0.0472415946,0.0857807249,-0.3888076842,0.0311133061,0.3042639792,0.154858917,0.175432682,0.2722882032,0.08177194,0.0818850547,0.1544100046,0.3952516615,0.2447241694,-0.0228369981,0.2661114633,-0.2058007568,0.0857820436,0.2306343317,0.2361091226,0.118432261,0.0381054915,0.1526700258,-0.1501827389,0.295978874,0.1446438283,-0.1672739387,0.2365310639,-0.3596028984,0.2424565107,-0.0986769199,0.0569087155,-0.1363285184,0.141786322,0.3276461959,-0.0491868742,-0.0884468332,-0.1608537436,-0.0030408807,-0.300190568,0.2709409893,0.4638240635,-0.0976310596,-0.0931242406,-0.043175891,0.2145775259,0.217528671,0.5374786854,0.3159116209,-0.1668998748,-0.4896293879,-0.1224408448,0.5082057714,-0.1266734451,-0.1844531447,0.4623732269,-0.134871155,0.0100118555,-0.165280953,0.3222151995,0.4884846807,0.3006311953,-0.1383483261,-0.2075260133,0.2566275001,-0.1801053286,-0.1383739263,0.2107214928,0.0596846007,0.2941676378,0.2981710732,0.0842535272,-0.1690877378,-0.0320378505,0.2207306325,-0.3357616365,-0.4409199655,0.206428498,-0.1567060947,-0.0026589187,-0.1897845715,0.0980003253,-0.4563895464,0.2288242429,0.315672189,-0.2224390507,-0.0670025125,-0.1489631385,0.0449592993,-0.0307994634,0.3605282903,0.2467247546,0.3562105,-0.2011301219,-0.1037739813,-0.8396953344,0.3146013319,0.0336657204,-0.0117303068,-0.0385530032,0.1393253952,-0.0011307682,0.3212940991,0.420112431,0.0454290248,0.1901207864,0.0372285321,-0.462015748,-0.2717567384,0.0900407732,-0.445476234,0.0050567961,-0.3917895555,-0.0040985593,-0.2824522853,-0.0414680019,-0.0273984112,-0.4465276599,0.2701758146,0.3022372723,0.2959749699,-0.0437552482,0.6722488999,-0.0536569208,0.099635683,-0.0690652728,-0.3176445663,-0.0269642416,0.1578307003,0.3953226507,0.2114282846,-0.2491105497,0.0748830065,-0.1284852326,0.1337044984,0.1330321282,0.0128994184,-0.2782158554,0.0685418844,-0.1654132903,0.4030524194,-0.1301134676,0.1823724061,0.0637911186,0.2705146372,-0.3036566377,-0.3571579456,0.3911989331,-0.3434489071,-0.4585971534,-0.4276721179,0.2904642522,0.159820199,-0.2049567997,-0.4742090106,0.0093041481,0.2549891472,-0.1533083916,-0.1555007994,0.5398080945,0.0699544102,0.2142618448,-0.1219963953,0.2098635286,0.0046629333,-0.1189015582,-0.1165047884,-0.0643654764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2695","title":"Cannot import load_dataset on Colab","comments":"@phosseini \r\nI think it is related to [1.10.0](https:\/\/github.com\/huggingface\/datasets\/actions\/runs\/1052653701) release done 3 hours ago. (cc: @lhoestq )\r\nFor now I just downgraded to 1.9.0 and it is working fine.","body":"## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":28,"text":"Cannot import load_dataset on Colab\n## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\n@phosseini \r\nI think it is related to [1.10.0](https:\/\/github.com\/huggingface\/datasets\/actions\/runs\/1052653701) release done 3 hours ago. (cc: @lhoestq )\r\nFor now I just downgraded to 1.9.0 and it is working fine.","embeddings":[-0.4603253901,-0.2148275971,-0.024375163,0.2682137489,0.1354128718,0.051994808,0.4973769188,-0.1074850261,0.1983702779,0.0849533305,-0.3266307712,0.4753232598,-0.1742698699,0.1487485617,-0.2584459782,0.0984701291,-0.0661861375,0.058203578,-0.2907032669,0.0956151709,0.0202920046,0.3308885992,-0.6335685253,-0.3468577266,-0.5786883831,-0.2100416869,-0.0277418792,0.168791458,-0.2702196836,-0.292011261,0.3478223681,0.1770586967,0.0921896547,0.5697174668,-0.0001172403,0.0715135038,0.3933092654,-0.0787535235,-0.2456898838,-0.3918715715,-0.3461531103,-0.3291964531,0.3686217368,-0.1073928103,0.1058271378,0.404165864,-0.1357311457,-0.133275196,0.2626952529,0.1352794766,0.1936342865,0.549197793,0.1630603671,-0.192492187,-0.1125865728,-0.193378076,-0.2020825595,0.7461523414,0.4174177051,0.0168563686,-0.0510248244,0.2604947388,-0.0992677361,0.2890874445,0.1862412095,-0.0334219635,0.1051012799,-0.4805828929,0.1006372571,0.1743771285,0.4806717336,-0.233426556,0.0443901122,-0.1220246106,0.2094738036,-0.2924674749,0.2619004846,-0.0985698551,-0.2118597776,0.0490080118,-0.0226965509,-0.0896168202,-0.0123074474,0.0164376963,-0.0933856592,0.3655072451,-0.1430076212,-0.1118264869,0.0698661357,-0.1096417829,0.5018094778,-0.0161823332,-0.0268674623,0.2063278407,-0.468000561,-0.0093295779,-0.098870948,-0.1408492327,-0.0516990833,-0.1014948189,-0.0361867324,0.0174666177,-0.1110087708,0.303483665,0.1848023385,0.1422498971,0.0347587094,0.0182988103,0.1226537004,0.2864284515,-0.2007629573,0.227394864,-0.259282738,-0.3118798137,0.1845084578,-0.029214235,0.2715788782,-0.0037106252,-0.319062084,-0.1818456352,0.0095681679,0.105042167,0.0446227603,0.3990536928,0.055795569,0.1057089195,-0.0346219428,0.0016938471,-0.3566534519,-0.2394378334,-0.0538450666,0.1936965287,-0.0642801523,-0.1152990013,0.0678034425,-0.1549987495,0.2601871789,-0.0124206394,0.3092051148,-0.0599603131,0.1036741883,-0.3025387824,-0.0606454387,0.331908226,0.1287349612,0.0872022063,-0.0623135045,-0.2055783719,-0.0778879523,0.2355169952,-0.4770620167,-0.1442362964,-0.0243145749,0.1429465264,-0.1339146942,-0.2244311124,-0.5850232244,0.1648343056,0.1447636336,-0.0812816098,-0.0849623531,-0.3062829673,-0.2330714166,-0.1535831094,0.1236880049,0.4482834041,-0.2309288979,-0.142314136,-0.2176906466,0.1205273345,0.1272705644,0.234823361,-0.2021541893,0.1736088246,-0.0962642506,-0.1171419844,0.490647763,-0.4326361418,-0.4671251178,-0.0203478895,-0.2660758793,-0.0052545536,-0.0897063687,0.0485790744,0.3577234745,-0.0669410899,0.5458390117,0.2629367113,-0.1390330195,-0.036604546,-0.1640905291,-0.4099052548,0.1294252127,0.0418195315,0.1834629029,-0.2177709639,0.203032732,-0.0533746518,0.1012382284,0.1859944463,-0.1211299896,0.3843335509,0.2429013252,-0.2223784029,0.0372738875,-0.1859572828,-0.4775209725,0.2240482867,0.409294188,0.0799716562,-0.0674993992,0.05132493,-0.4381961823,0.0791065395,-0.2594439685,0.0676428005,0.0994127542,-0.0666698813,-0.1751303822,0.0889260471,-0.2604830265,0.7403804064,0.0377074406,0.2175934911,-0.3183411956,0.2157882154,-0.1859039217,-0.3081798851,0.1762920618,0.1361715198,0.0753254518,-0.092272602,-0.2389075756,0.1421211809,0.2554138899,0.1713412106,0.061892774,-0.2776550353,0.4123665094,-0.274011761,0.0123248948,-0.1066840887,0.208570227,0.0058228662,0.2551185191,0.1565714628,-0.3707030118,-0.0410736091,-0.0305118244,0.0905382559,0.3726439178,0.1239964738,-0.0050928039,-0.1856487691,0.325183779,0.2222952098,0.192474246,0.0937420502,-0.1297319233,0.2881348133,0.3061757982,0.2452441901,0.038779363,-0.2685610354,-0.2554307878,0.0633648261,0.1922871321,0.2518091798,0.4446454048,0.1803891063,0.1049195379,0.0316309184,0.0318480134,-0.1561172903,0.1151199043,0.2919997871,0.2558412552,0.1311028153,0.4120824635,0.0967440531,-0.078998588,-0.3894120455,0.1753250211,0.1503109038,-0.3317721486,0.0766265243,-0.2146017402,-0.113947168,-0.1419459134,-0.4253391027,-0.1464539021,-0.1978348494,-0.1058005691,0.2810257375,0.2650622427,0.1936620623,-0.0239940546,0.0500169992,0.0625045747,-0.142483592,-0.0558168627,-0.2506769001,-0.0925878361,0.0015159415,0.327739507,0.246811226,0.3685063124,-0.3038308024,-0.0695626289,0.1437006891,-0.332098037,0.1763451695,-0.005682677,0.2919798791,-0.0134428795,-0.0779709816,-0.3206300437,-0.1939175874,0.0215363186,-0.1346231997,-0.1827463955,0.0472795218,-0.1800287366,-0.2297740728,-0.0019908124,-0.2926879525,-0.4344463348,-0.3998571932,-0.3682101369,0.0472466946,-0.0089030508,0.4603004754,-0.0427159443,0.2933925688,0.2250700146,0.0508381985,0.0179356467,-0.3330911994,0.4733818173,-0.2865113616,-0.2936578393,0.2801209092,-0.2680461109,0.3836089373,0.0181830842,-0.1104436517,-0.2135308832,-0.0214798469,0.3837370574,-0.1009398997,0.0416830219,0.0846360251,0.1136619002,0.0782582983,-0.092406489,-0.1815526336,-0.2040507346,0.2196319252,0.1930779815,-0.1657154113,0.2265984565,-0.0363357775,0.8421141505,0.0771993697,-0.0400505289,0.3015678525,-0.061655391,0.3064921498,-0.1064245924,-0.5121492743,-0.0696846843,-0.1073111445,-0.039284341,0.0819721445,-0.1839754134,-0.1112223938,-0.2847859561,-0.2727975249,-0.3160193861,-0.291852355,-0.1147850677,-0.2003910989,0.1308543533,0.1986790299,-0.0612915456,-0.1620080918,-0.1124967635,0.0489822105,0.4355244339,-0.4208028316,-0.0186146274,0.3795593679,-0.0728556365,-0.3464918733,0.1986530572,0.1317849904,0.3303270042,0.0766258389,0.1182002276,0.0601407997,-0.0493247956,0.1122457087,0.05919642,-0.0800003186,0.2366157919,-0.0146941952,-0.4012733102,-0.0577438287,-0.2255629599,0.2240988016,0.4793192744,0.0861525759,-0.2667843401,0.2115175128,0.2363318503,0.1412735432,-0.114222765,-0.1485226005,-0.4387521744,-0.3134378195,-0.378410399,-0.0600447692,0.1850457788,0.2391748279,0.0416690968,-0.0708694831,-0.0640273839,-0.2169730812,0.0610414371,0.3596898913,0.1482839882,0.270448029,0.5801549554,0.1271515787,-0.2473592013,0.4470187426,0.5865612626,0.0028329962,-0.5835172534,-0.0279992204,-0.1797391474,0.2348893136,0.1315195262,-0.0706735849,0.0163302217,0.0595659725,-0.1358518749,0.3848748207,0.3812657893,-0.0635423213,0.1262477934,-0.4026652873,-0.1285912842,0.0958481207,0.0007427831,0.1376991123,0.4601660669,-0.3796385229,-0.0598220304,0.047099065,-0.1843536198,0.686365068,-0.0381890498,-0.2390154898,0.2732535005,-0.1615026444,0.2652589977,0.1131904423,0.1857605428,-0.2762228251,-0.3076212406,0.0923932046,-0.2266215682,0.1900261045,-0.1945811212,-0.3511314392,0.2596141994,0.0564561076,-0.1498299241,-0.0185761657,0.0406053141,0.0024130014,-0.2937503755,-0.3525202274,0.0517692231,-0.1905304193,0.2061717808,-0.1972419322,-0.1833461672,-0.1389480531,-0.1865322441,-0.0331040733,-0.1090358868,-0.6445795894,0.2370983511,0.3314474523,-0.3087444007,0.2005167454,-0.1111385226,0.0950988606,-0.0293304492,-0.1739497632,0.1265123934,0.1993344128,0.0166862272,-0.1108888164,-0.172952801,0.4197721481,-0.0839659944,0.0050038416,0.0576662868,0.0366229266,-0.0225345157,0.4702349007,0.3512617946,0.3681568801,0.150692597,-0.0755758658,0.1810477078,0.26167503,-0.1958780736,0.0780422837,0.1019752249,-0.1530845314,0.2558067143,-0.107513763,-0.1003244594,-0.0019221834,0.3397724032,0.1296715736,-0.0161001608,0.1832213998,0.1551724821,-0.2558426857,-0.0844762251,0.1688349247,-0.3087925613,-0.3004136086,0.2412767708,-0.0199141074,0.0491270311,0.0744944811,0.7340957522,0.2456151992,-0.3685689867,-0.132078737,-0.3772931397,-0.5326016545,0.1417692751,-0.1205238625,-0.0949431732,0.2369502336,0.3836895525,0.120868288,0.1284108609,-0.2208905369,0.1052405685,-0.1584761292,0.1540038735,0.2756511569,-0.0960519463,0.0372927338,0.0859900042,0.0042420314,0.1529782265,0.0671002194,-0.2249542475,-0.1146583706,0.1326806545,-0.1025531739,0.0408711247,-0.1180261746,-0.1512853205,-0.2300352007,-0.1715027094,0.1590514779,0.0351236463,-0.0818419009,-0.1321561933,0.2920054197,-0.0966290534,-0.2174018025,0.4025671184,0.1177951023,0.0810081363,0.0776668489,0.1660300344,0.1665427387,0.1832786202,0.017530106,-0.2253397107,-0.091671519,-0.1765127629,0.0942778215,-0.3320890367,-0.0140226921,0.0699948967,-0.0275069047,0.4203799963,-0.128956452,-0.2099847645,0.0472415946,0.0857807249,-0.3888076842,0.0311133061,0.3042639792,0.154858917,0.175432682,0.2722882032,0.08177194,0.0818850547,0.1544100046,0.3952516615,0.2447241694,-0.0228369981,0.2661114633,-0.2058007568,0.0857820436,0.2306343317,0.2361091226,0.118432261,0.0381054915,0.1526700258,-0.1501827389,0.295978874,0.1446438283,-0.1672739387,0.2365310639,-0.3596028984,0.2424565107,-0.0986769199,0.0569087155,-0.1363285184,0.141786322,0.3276461959,-0.0491868742,-0.0884468332,-0.1608537436,-0.0030408807,-0.300190568,0.2709409893,0.4638240635,-0.0976310596,-0.0931242406,-0.043175891,0.2145775259,0.217528671,0.5374786854,0.3159116209,-0.1668998748,-0.4896293879,-0.1224408448,0.5082057714,-0.1266734451,-0.1844531447,0.4623732269,-0.134871155,0.0100118555,-0.165280953,0.3222151995,0.4884846807,0.3006311953,-0.1383483261,-0.2075260133,0.2566275001,-0.1801053286,-0.1383739263,0.2107214928,0.0596846007,0.2941676378,0.2981710732,0.0842535272,-0.1690877378,-0.0320378505,0.2207306325,-0.3357616365,-0.4409199655,0.206428498,-0.1567060947,-0.0026589187,-0.1897845715,0.0980003253,-0.4563895464,0.2288242429,0.315672189,-0.2224390507,-0.0670025125,-0.1489631385,0.0449592993,-0.0307994634,0.3605282903,0.2467247546,0.3562105,-0.2011301219,-0.1037739813,-0.8396953344,0.3146013319,0.0336657204,-0.0117303068,-0.0385530032,0.1393253952,-0.0011307682,0.3212940991,0.420112431,0.0454290248,0.1901207864,0.0372285321,-0.462015748,-0.2717567384,0.0900407732,-0.445476234,0.0050567961,-0.3917895555,-0.0040985593,-0.2824522853,-0.0414680019,-0.0273984112,-0.4465276599,0.2701758146,0.3022372723,0.2959749699,-0.0437552482,0.6722488999,-0.0536569208,0.099635683,-0.0690652728,-0.3176445663,-0.0269642416,0.1578307003,0.3953226507,0.2114282846,-0.2491105497,0.0748830065,-0.1284852326,0.1337044984,0.1330321282,0.0128994184,-0.2782158554,0.0685418844,-0.1654132903,0.4030524194,-0.1301134676,0.1823724061,0.0637911186,0.2705146372,-0.3036566377,-0.3571579456,0.3911989331,-0.3434489071,-0.4585971534,-0.4276721179,0.2904642522,0.159820199,-0.2049567997,-0.4742090106,0.0093041481,0.2549891472,-0.1533083916,-0.1555007994,0.5398080945,0.0699544102,0.2142618448,-0.1219963953,0.2098635286,0.0046629333,-0.1189015582,-0.1165047884,-0.0643654764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2695","title":"Cannot import load_dataset on Colab","comments":"> @phosseini\r\n> I think it is related to [1.10.0](https:\/\/github.com\/huggingface\/datasets\/actions\/runs\/1052653701) release done 3 hours ago. (cc: @lhoestq )\r\n> For now I just downgraded to 1.9.0 and it is working fine.\r\n\r\nSame here, downgraded to 1.9.0 for now and works fine.","body":"## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":41,"text":"Cannot import load_dataset on Colab\n## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\n> @phosseini\r\n> I think it is related to [1.10.0](https:\/\/github.com\/huggingface\/datasets\/actions\/runs\/1052653701) release done 3 hours ago. (cc: @lhoestq )\r\n> For now I just downgraded to 1.9.0 and it is working fine.\r\n\r\nSame here, downgraded to 1.9.0 for now and works fine.","embeddings":[-0.4603253901,-0.2148275971,-0.024375163,0.2682137489,0.1354128718,0.051994808,0.4973769188,-0.1074850261,0.1983702779,0.0849533305,-0.3266307712,0.4753232598,-0.1742698699,0.1487485617,-0.2584459782,0.0984701291,-0.0661861375,0.058203578,-0.2907032669,0.0956151709,0.0202920046,0.3308885992,-0.6335685253,-0.3468577266,-0.5786883831,-0.2100416869,-0.0277418792,0.168791458,-0.2702196836,-0.292011261,0.3478223681,0.1770586967,0.0921896547,0.5697174668,-0.0001172403,0.0715135038,0.3933092654,-0.0787535235,-0.2456898838,-0.3918715715,-0.3461531103,-0.3291964531,0.3686217368,-0.1073928103,0.1058271378,0.404165864,-0.1357311457,-0.133275196,0.2626952529,0.1352794766,0.1936342865,0.549197793,0.1630603671,-0.192492187,-0.1125865728,-0.193378076,-0.2020825595,0.7461523414,0.4174177051,0.0168563686,-0.0510248244,0.2604947388,-0.0992677361,0.2890874445,0.1862412095,-0.0334219635,0.1051012799,-0.4805828929,0.1006372571,0.1743771285,0.4806717336,-0.233426556,0.0443901122,-0.1220246106,0.2094738036,-0.2924674749,0.2619004846,-0.0985698551,-0.2118597776,0.0490080118,-0.0226965509,-0.0896168202,-0.0123074474,0.0164376963,-0.0933856592,0.3655072451,-0.1430076212,-0.1118264869,0.0698661357,-0.1096417829,0.5018094778,-0.0161823332,-0.0268674623,0.2063278407,-0.468000561,-0.0093295779,-0.098870948,-0.1408492327,-0.0516990833,-0.1014948189,-0.0361867324,0.0174666177,-0.1110087708,0.303483665,0.1848023385,0.1422498971,0.0347587094,0.0182988103,0.1226537004,0.2864284515,-0.2007629573,0.227394864,-0.259282738,-0.3118798137,0.1845084578,-0.029214235,0.2715788782,-0.0037106252,-0.319062084,-0.1818456352,0.0095681679,0.105042167,0.0446227603,0.3990536928,0.055795569,0.1057089195,-0.0346219428,0.0016938471,-0.3566534519,-0.2394378334,-0.0538450666,0.1936965287,-0.0642801523,-0.1152990013,0.0678034425,-0.1549987495,0.2601871789,-0.0124206394,0.3092051148,-0.0599603131,0.1036741883,-0.3025387824,-0.0606454387,0.331908226,0.1287349612,0.0872022063,-0.0623135045,-0.2055783719,-0.0778879523,0.2355169952,-0.4770620167,-0.1442362964,-0.0243145749,0.1429465264,-0.1339146942,-0.2244311124,-0.5850232244,0.1648343056,0.1447636336,-0.0812816098,-0.0849623531,-0.3062829673,-0.2330714166,-0.1535831094,0.1236880049,0.4482834041,-0.2309288979,-0.142314136,-0.2176906466,0.1205273345,0.1272705644,0.234823361,-0.2021541893,0.1736088246,-0.0962642506,-0.1171419844,0.490647763,-0.4326361418,-0.4671251178,-0.0203478895,-0.2660758793,-0.0052545536,-0.0897063687,0.0485790744,0.3577234745,-0.0669410899,0.5458390117,0.2629367113,-0.1390330195,-0.036604546,-0.1640905291,-0.4099052548,0.1294252127,0.0418195315,0.1834629029,-0.2177709639,0.203032732,-0.0533746518,0.1012382284,0.1859944463,-0.1211299896,0.3843335509,0.2429013252,-0.2223784029,0.0372738875,-0.1859572828,-0.4775209725,0.2240482867,0.409294188,0.0799716562,-0.0674993992,0.05132493,-0.4381961823,0.0791065395,-0.2594439685,0.0676428005,0.0994127542,-0.0666698813,-0.1751303822,0.0889260471,-0.2604830265,0.7403804064,0.0377074406,0.2175934911,-0.3183411956,0.2157882154,-0.1859039217,-0.3081798851,0.1762920618,0.1361715198,0.0753254518,-0.092272602,-0.2389075756,0.1421211809,0.2554138899,0.1713412106,0.061892774,-0.2776550353,0.4123665094,-0.274011761,0.0123248948,-0.1066840887,0.208570227,0.0058228662,0.2551185191,0.1565714628,-0.3707030118,-0.0410736091,-0.0305118244,0.0905382559,0.3726439178,0.1239964738,-0.0050928039,-0.1856487691,0.325183779,0.2222952098,0.192474246,0.0937420502,-0.1297319233,0.2881348133,0.3061757982,0.2452441901,0.038779363,-0.2685610354,-0.2554307878,0.0633648261,0.1922871321,0.2518091798,0.4446454048,0.1803891063,0.1049195379,0.0316309184,0.0318480134,-0.1561172903,0.1151199043,0.2919997871,0.2558412552,0.1311028153,0.4120824635,0.0967440531,-0.078998588,-0.3894120455,0.1753250211,0.1503109038,-0.3317721486,0.0766265243,-0.2146017402,-0.113947168,-0.1419459134,-0.4253391027,-0.1464539021,-0.1978348494,-0.1058005691,0.2810257375,0.2650622427,0.1936620623,-0.0239940546,0.0500169992,0.0625045747,-0.142483592,-0.0558168627,-0.2506769001,-0.0925878361,0.0015159415,0.327739507,0.246811226,0.3685063124,-0.3038308024,-0.0695626289,0.1437006891,-0.332098037,0.1763451695,-0.005682677,0.2919798791,-0.0134428795,-0.0779709816,-0.3206300437,-0.1939175874,0.0215363186,-0.1346231997,-0.1827463955,0.0472795218,-0.1800287366,-0.2297740728,-0.0019908124,-0.2926879525,-0.4344463348,-0.3998571932,-0.3682101369,0.0472466946,-0.0089030508,0.4603004754,-0.0427159443,0.2933925688,0.2250700146,0.0508381985,0.0179356467,-0.3330911994,0.4733818173,-0.2865113616,-0.2936578393,0.2801209092,-0.2680461109,0.3836089373,0.0181830842,-0.1104436517,-0.2135308832,-0.0214798469,0.3837370574,-0.1009398997,0.0416830219,0.0846360251,0.1136619002,0.0782582983,-0.092406489,-0.1815526336,-0.2040507346,0.2196319252,0.1930779815,-0.1657154113,0.2265984565,-0.0363357775,0.8421141505,0.0771993697,-0.0400505289,0.3015678525,-0.061655391,0.3064921498,-0.1064245924,-0.5121492743,-0.0696846843,-0.1073111445,-0.039284341,0.0819721445,-0.1839754134,-0.1112223938,-0.2847859561,-0.2727975249,-0.3160193861,-0.291852355,-0.1147850677,-0.2003910989,0.1308543533,0.1986790299,-0.0612915456,-0.1620080918,-0.1124967635,0.0489822105,0.4355244339,-0.4208028316,-0.0186146274,0.3795593679,-0.0728556365,-0.3464918733,0.1986530572,0.1317849904,0.3303270042,0.0766258389,0.1182002276,0.0601407997,-0.0493247956,0.1122457087,0.05919642,-0.0800003186,0.2366157919,-0.0146941952,-0.4012733102,-0.0577438287,-0.2255629599,0.2240988016,0.4793192744,0.0861525759,-0.2667843401,0.2115175128,0.2363318503,0.1412735432,-0.114222765,-0.1485226005,-0.4387521744,-0.3134378195,-0.378410399,-0.0600447692,0.1850457788,0.2391748279,0.0416690968,-0.0708694831,-0.0640273839,-0.2169730812,0.0610414371,0.3596898913,0.1482839882,0.270448029,0.5801549554,0.1271515787,-0.2473592013,0.4470187426,0.5865612626,0.0028329962,-0.5835172534,-0.0279992204,-0.1797391474,0.2348893136,0.1315195262,-0.0706735849,0.0163302217,0.0595659725,-0.1358518749,0.3848748207,0.3812657893,-0.0635423213,0.1262477934,-0.4026652873,-0.1285912842,0.0958481207,0.0007427831,0.1376991123,0.4601660669,-0.3796385229,-0.0598220304,0.047099065,-0.1843536198,0.686365068,-0.0381890498,-0.2390154898,0.2732535005,-0.1615026444,0.2652589977,0.1131904423,0.1857605428,-0.2762228251,-0.3076212406,0.0923932046,-0.2266215682,0.1900261045,-0.1945811212,-0.3511314392,0.2596141994,0.0564561076,-0.1498299241,-0.0185761657,0.0406053141,0.0024130014,-0.2937503755,-0.3525202274,0.0517692231,-0.1905304193,0.2061717808,-0.1972419322,-0.1833461672,-0.1389480531,-0.1865322441,-0.0331040733,-0.1090358868,-0.6445795894,0.2370983511,0.3314474523,-0.3087444007,0.2005167454,-0.1111385226,0.0950988606,-0.0293304492,-0.1739497632,0.1265123934,0.1993344128,0.0166862272,-0.1108888164,-0.172952801,0.4197721481,-0.0839659944,0.0050038416,0.0576662868,0.0366229266,-0.0225345157,0.4702349007,0.3512617946,0.3681568801,0.150692597,-0.0755758658,0.1810477078,0.26167503,-0.1958780736,0.0780422837,0.1019752249,-0.1530845314,0.2558067143,-0.107513763,-0.1003244594,-0.0019221834,0.3397724032,0.1296715736,-0.0161001608,0.1832213998,0.1551724821,-0.2558426857,-0.0844762251,0.1688349247,-0.3087925613,-0.3004136086,0.2412767708,-0.0199141074,0.0491270311,0.0744944811,0.7340957522,0.2456151992,-0.3685689867,-0.132078737,-0.3772931397,-0.5326016545,0.1417692751,-0.1205238625,-0.0949431732,0.2369502336,0.3836895525,0.120868288,0.1284108609,-0.2208905369,0.1052405685,-0.1584761292,0.1540038735,0.2756511569,-0.0960519463,0.0372927338,0.0859900042,0.0042420314,0.1529782265,0.0671002194,-0.2249542475,-0.1146583706,0.1326806545,-0.1025531739,0.0408711247,-0.1180261746,-0.1512853205,-0.2300352007,-0.1715027094,0.1590514779,0.0351236463,-0.0818419009,-0.1321561933,0.2920054197,-0.0966290534,-0.2174018025,0.4025671184,0.1177951023,0.0810081363,0.0776668489,0.1660300344,0.1665427387,0.1832786202,0.017530106,-0.2253397107,-0.091671519,-0.1765127629,0.0942778215,-0.3320890367,-0.0140226921,0.0699948967,-0.0275069047,0.4203799963,-0.128956452,-0.2099847645,0.0472415946,0.0857807249,-0.3888076842,0.0311133061,0.3042639792,0.154858917,0.175432682,0.2722882032,0.08177194,0.0818850547,0.1544100046,0.3952516615,0.2447241694,-0.0228369981,0.2661114633,-0.2058007568,0.0857820436,0.2306343317,0.2361091226,0.118432261,0.0381054915,0.1526700258,-0.1501827389,0.295978874,0.1446438283,-0.1672739387,0.2365310639,-0.3596028984,0.2424565107,-0.0986769199,0.0569087155,-0.1363285184,0.141786322,0.3276461959,-0.0491868742,-0.0884468332,-0.1608537436,-0.0030408807,-0.300190568,0.2709409893,0.4638240635,-0.0976310596,-0.0931242406,-0.043175891,0.2145775259,0.217528671,0.5374786854,0.3159116209,-0.1668998748,-0.4896293879,-0.1224408448,0.5082057714,-0.1266734451,-0.1844531447,0.4623732269,-0.134871155,0.0100118555,-0.165280953,0.3222151995,0.4884846807,0.3006311953,-0.1383483261,-0.2075260133,0.2566275001,-0.1801053286,-0.1383739263,0.2107214928,0.0596846007,0.2941676378,0.2981710732,0.0842535272,-0.1690877378,-0.0320378505,0.2207306325,-0.3357616365,-0.4409199655,0.206428498,-0.1567060947,-0.0026589187,-0.1897845715,0.0980003253,-0.4563895464,0.2288242429,0.315672189,-0.2224390507,-0.0670025125,-0.1489631385,0.0449592993,-0.0307994634,0.3605282903,0.2467247546,0.3562105,-0.2011301219,-0.1037739813,-0.8396953344,0.3146013319,0.0336657204,-0.0117303068,-0.0385530032,0.1393253952,-0.0011307682,0.3212940991,0.420112431,0.0454290248,0.1901207864,0.0372285321,-0.462015748,-0.2717567384,0.0900407732,-0.445476234,0.0050567961,-0.3917895555,-0.0040985593,-0.2824522853,-0.0414680019,-0.0273984112,-0.4465276599,0.2701758146,0.3022372723,0.2959749699,-0.0437552482,0.6722488999,-0.0536569208,0.099635683,-0.0690652728,-0.3176445663,-0.0269642416,0.1578307003,0.3953226507,0.2114282846,-0.2491105497,0.0748830065,-0.1284852326,0.1337044984,0.1330321282,0.0128994184,-0.2782158554,0.0685418844,-0.1654132903,0.4030524194,-0.1301134676,0.1823724061,0.0637911186,0.2705146372,-0.3036566377,-0.3571579456,0.3911989331,-0.3434489071,-0.4585971534,-0.4276721179,0.2904642522,0.159820199,-0.2049567997,-0.4742090106,0.0093041481,0.2549891472,-0.1533083916,-0.1555007994,0.5398080945,0.0699544102,0.2142618448,-0.1219963953,0.2098635286,0.0046629333,-0.1189015582,-0.1165047884,-0.0643654764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2695","title":"Cannot import load_dataset on Colab","comments":"Hi, \r\n\r\nupdating tqdm to the newest version resolves the issue for me. You can do this as follows in Colab:\r\n```\r\n!pip install tqdm --upgrade\r\n```","body":"## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":26,"text":"Cannot import load_dataset on Colab\n## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nHi, \r\n\r\nupdating tqdm to the newest version resolves the issue for me. You can do this as follows in Colab:\r\n```\r\n!pip install tqdm --upgrade\r\n```","embeddings":[-0.4603253901,-0.2148275971,-0.024375163,0.2682137489,0.1354128718,0.051994808,0.4973769188,-0.1074850261,0.1983702779,0.0849533305,-0.3266307712,0.4753232598,-0.1742698699,0.1487485617,-0.2584459782,0.0984701291,-0.0661861375,0.058203578,-0.2907032669,0.0956151709,0.0202920046,0.3308885992,-0.6335685253,-0.3468577266,-0.5786883831,-0.2100416869,-0.0277418792,0.168791458,-0.2702196836,-0.292011261,0.3478223681,0.1770586967,0.0921896547,0.5697174668,-0.0001172403,0.0715135038,0.3933092654,-0.0787535235,-0.2456898838,-0.3918715715,-0.3461531103,-0.3291964531,0.3686217368,-0.1073928103,0.1058271378,0.404165864,-0.1357311457,-0.133275196,0.2626952529,0.1352794766,0.1936342865,0.549197793,0.1630603671,-0.192492187,-0.1125865728,-0.193378076,-0.2020825595,0.7461523414,0.4174177051,0.0168563686,-0.0510248244,0.2604947388,-0.0992677361,0.2890874445,0.1862412095,-0.0334219635,0.1051012799,-0.4805828929,0.1006372571,0.1743771285,0.4806717336,-0.233426556,0.0443901122,-0.1220246106,0.2094738036,-0.2924674749,0.2619004846,-0.0985698551,-0.2118597776,0.0490080118,-0.0226965509,-0.0896168202,-0.0123074474,0.0164376963,-0.0933856592,0.3655072451,-0.1430076212,-0.1118264869,0.0698661357,-0.1096417829,0.5018094778,-0.0161823332,-0.0268674623,0.2063278407,-0.468000561,-0.0093295779,-0.098870948,-0.1408492327,-0.0516990833,-0.1014948189,-0.0361867324,0.0174666177,-0.1110087708,0.303483665,0.1848023385,0.1422498971,0.0347587094,0.0182988103,0.1226537004,0.2864284515,-0.2007629573,0.227394864,-0.259282738,-0.3118798137,0.1845084578,-0.029214235,0.2715788782,-0.0037106252,-0.319062084,-0.1818456352,0.0095681679,0.105042167,0.0446227603,0.3990536928,0.055795569,0.1057089195,-0.0346219428,0.0016938471,-0.3566534519,-0.2394378334,-0.0538450666,0.1936965287,-0.0642801523,-0.1152990013,0.0678034425,-0.1549987495,0.2601871789,-0.0124206394,0.3092051148,-0.0599603131,0.1036741883,-0.3025387824,-0.0606454387,0.331908226,0.1287349612,0.0872022063,-0.0623135045,-0.2055783719,-0.0778879523,0.2355169952,-0.4770620167,-0.1442362964,-0.0243145749,0.1429465264,-0.1339146942,-0.2244311124,-0.5850232244,0.1648343056,0.1447636336,-0.0812816098,-0.0849623531,-0.3062829673,-0.2330714166,-0.1535831094,0.1236880049,0.4482834041,-0.2309288979,-0.142314136,-0.2176906466,0.1205273345,0.1272705644,0.234823361,-0.2021541893,0.1736088246,-0.0962642506,-0.1171419844,0.490647763,-0.4326361418,-0.4671251178,-0.0203478895,-0.2660758793,-0.0052545536,-0.0897063687,0.0485790744,0.3577234745,-0.0669410899,0.5458390117,0.2629367113,-0.1390330195,-0.036604546,-0.1640905291,-0.4099052548,0.1294252127,0.0418195315,0.1834629029,-0.2177709639,0.203032732,-0.0533746518,0.1012382284,0.1859944463,-0.1211299896,0.3843335509,0.2429013252,-0.2223784029,0.0372738875,-0.1859572828,-0.4775209725,0.2240482867,0.409294188,0.0799716562,-0.0674993992,0.05132493,-0.4381961823,0.0791065395,-0.2594439685,0.0676428005,0.0994127542,-0.0666698813,-0.1751303822,0.0889260471,-0.2604830265,0.7403804064,0.0377074406,0.2175934911,-0.3183411956,0.2157882154,-0.1859039217,-0.3081798851,0.1762920618,0.1361715198,0.0753254518,-0.092272602,-0.2389075756,0.1421211809,0.2554138899,0.1713412106,0.061892774,-0.2776550353,0.4123665094,-0.274011761,0.0123248948,-0.1066840887,0.208570227,0.0058228662,0.2551185191,0.1565714628,-0.3707030118,-0.0410736091,-0.0305118244,0.0905382559,0.3726439178,0.1239964738,-0.0050928039,-0.1856487691,0.325183779,0.2222952098,0.192474246,0.0937420502,-0.1297319233,0.2881348133,0.3061757982,0.2452441901,0.038779363,-0.2685610354,-0.2554307878,0.0633648261,0.1922871321,0.2518091798,0.4446454048,0.1803891063,0.1049195379,0.0316309184,0.0318480134,-0.1561172903,0.1151199043,0.2919997871,0.2558412552,0.1311028153,0.4120824635,0.0967440531,-0.078998588,-0.3894120455,0.1753250211,0.1503109038,-0.3317721486,0.0766265243,-0.2146017402,-0.113947168,-0.1419459134,-0.4253391027,-0.1464539021,-0.1978348494,-0.1058005691,0.2810257375,0.2650622427,0.1936620623,-0.0239940546,0.0500169992,0.0625045747,-0.142483592,-0.0558168627,-0.2506769001,-0.0925878361,0.0015159415,0.327739507,0.246811226,0.3685063124,-0.3038308024,-0.0695626289,0.1437006891,-0.332098037,0.1763451695,-0.005682677,0.2919798791,-0.0134428795,-0.0779709816,-0.3206300437,-0.1939175874,0.0215363186,-0.1346231997,-0.1827463955,0.0472795218,-0.1800287366,-0.2297740728,-0.0019908124,-0.2926879525,-0.4344463348,-0.3998571932,-0.3682101369,0.0472466946,-0.0089030508,0.4603004754,-0.0427159443,0.2933925688,0.2250700146,0.0508381985,0.0179356467,-0.3330911994,0.4733818173,-0.2865113616,-0.2936578393,0.2801209092,-0.2680461109,0.3836089373,0.0181830842,-0.1104436517,-0.2135308832,-0.0214798469,0.3837370574,-0.1009398997,0.0416830219,0.0846360251,0.1136619002,0.0782582983,-0.092406489,-0.1815526336,-0.2040507346,0.2196319252,0.1930779815,-0.1657154113,0.2265984565,-0.0363357775,0.8421141505,0.0771993697,-0.0400505289,0.3015678525,-0.061655391,0.3064921498,-0.1064245924,-0.5121492743,-0.0696846843,-0.1073111445,-0.039284341,0.0819721445,-0.1839754134,-0.1112223938,-0.2847859561,-0.2727975249,-0.3160193861,-0.291852355,-0.1147850677,-0.2003910989,0.1308543533,0.1986790299,-0.0612915456,-0.1620080918,-0.1124967635,0.0489822105,0.4355244339,-0.4208028316,-0.0186146274,0.3795593679,-0.0728556365,-0.3464918733,0.1986530572,0.1317849904,0.3303270042,0.0766258389,0.1182002276,0.0601407997,-0.0493247956,0.1122457087,0.05919642,-0.0800003186,0.2366157919,-0.0146941952,-0.4012733102,-0.0577438287,-0.2255629599,0.2240988016,0.4793192744,0.0861525759,-0.2667843401,0.2115175128,0.2363318503,0.1412735432,-0.114222765,-0.1485226005,-0.4387521744,-0.3134378195,-0.378410399,-0.0600447692,0.1850457788,0.2391748279,0.0416690968,-0.0708694831,-0.0640273839,-0.2169730812,0.0610414371,0.3596898913,0.1482839882,0.270448029,0.5801549554,0.1271515787,-0.2473592013,0.4470187426,0.5865612626,0.0028329962,-0.5835172534,-0.0279992204,-0.1797391474,0.2348893136,0.1315195262,-0.0706735849,0.0163302217,0.0595659725,-0.1358518749,0.3848748207,0.3812657893,-0.0635423213,0.1262477934,-0.4026652873,-0.1285912842,0.0958481207,0.0007427831,0.1376991123,0.4601660669,-0.3796385229,-0.0598220304,0.047099065,-0.1843536198,0.686365068,-0.0381890498,-0.2390154898,0.2732535005,-0.1615026444,0.2652589977,0.1131904423,0.1857605428,-0.2762228251,-0.3076212406,0.0923932046,-0.2266215682,0.1900261045,-0.1945811212,-0.3511314392,0.2596141994,0.0564561076,-0.1498299241,-0.0185761657,0.0406053141,0.0024130014,-0.2937503755,-0.3525202274,0.0517692231,-0.1905304193,0.2061717808,-0.1972419322,-0.1833461672,-0.1389480531,-0.1865322441,-0.0331040733,-0.1090358868,-0.6445795894,0.2370983511,0.3314474523,-0.3087444007,0.2005167454,-0.1111385226,0.0950988606,-0.0293304492,-0.1739497632,0.1265123934,0.1993344128,0.0166862272,-0.1108888164,-0.172952801,0.4197721481,-0.0839659944,0.0050038416,0.0576662868,0.0366229266,-0.0225345157,0.4702349007,0.3512617946,0.3681568801,0.150692597,-0.0755758658,0.1810477078,0.26167503,-0.1958780736,0.0780422837,0.1019752249,-0.1530845314,0.2558067143,-0.107513763,-0.1003244594,-0.0019221834,0.3397724032,0.1296715736,-0.0161001608,0.1832213998,0.1551724821,-0.2558426857,-0.0844762251,0.1688349247,-0.3087925613,-0.3004136086,0.2412767708,-0.0199141074,0.0491270311,0.0744944811,0.7340957522,0.2456151992,-0.3685689867,-0.132078737,-0.3772931397,-0.5326016545,0.1417692751,-0.1205238625,-0.0949431732,0.2369502336,0.3836895525,0.120868288,0.1284108609,-0.2208905369,0.1052405685,-0.1584761292,0.1540038735,0.2756511569,-0.0960519463,0.0372927338,0.0859900042,0.0042420314,0.1529782265,0.0671002194,-0.2249542475,-0.1146583706,0.1326806545,-0.1025531739,0.0408711247,-0.1180261746,-0.1512853205,-0.2300352007,-0.1715027094,0.1590514779,0.0351236463,-0.0818419009,-0.1321561933,0.2920054197,-0.0966290534,-0.2174018025,0.4025671184,0.1177951023,0.0810081363,0.0776668489,0.1660300344,0.1665427387,0.1832786202,0.017530106,-0.2253397107,-0.091671519,-0.1765127629,0.0942778215,-0.3320890367,-0.0140226921,0.0699948967,-0.0275069047,0.4203799963,-0.128956452,-0.2099847645,0.0472415946,0.0857807249,-0.3888076842,0.0311133061,0.3042639792,0.154858917,0.175432682,0.2722882032,0.08177194,0.0818850547,0.1544100046,0.3952516615,0.2447241694,-0.0228369981,0.2661114633,-0.2058007568,0.0857820436,0.2306343317,0.2361091226,0.118432261,0.0381054915,0.1526700258,-0.1501827389,0.295978874,0.1446438283,-0.1672739387,0.2365310639,-0.3596028984,0.2424565107,-0.0986769199,0.0569087155,-0.1363285184,0.141786322,0.3276461959,-0.0491868742,-0.0884468332,-0.1608537436,-0.0030408807,-0.300190568,0.2709409893,0.4638240635,-0.0976310596,-0.0931242406,-0.043175891,0.2145775259,0.217528671,0.5374786854,0.3159116209,-0.1668998748,-0.4896293879,-0.1224408448,0.5082057714,-0.1266734451,-0.1844531447,0.4623732269,-0.134871155,0.0100118555,-0.165280953,0.3222151995,0.4884846807,0.3006311953,-0.1383483261,-0.2075260133,0.2566275001,-0.1801053286,-0.1383739263,0.2107214928,0.0596846007,0.2941676378,0.2981710732,0.0842535272,-0.1690877378,-0.0320378505,0.2207306325,-0.3357616365,-0.4409199655,0.206428498,-0.1567060947,-0.0026589187,-0.1897845715,0.0980003253,-0.4563895464,0.2288242429,0.315672189,-0.2224390507,-0.0670025125,-0.1489631385,0.0449592993,-0.0307994634,0.3605282903,0.2467247546,0.3562105,-0.2011301219,-0.1037739813,-0.8396953344,0.3146013319,0.0336657204,-0.0117303068,-0.0385530032,0.1393253952,-0.0011307682,0.3212940991,0.420112431,0.0454290248,0.1901207864,0.0372285321,-0.462015748,-0.2717567384,0.0900407732,-0.445476234,0.0050567961,-0.3917895555,-0.0040985593,-0.2824522853,-0.0414680019,-0.0273984112,-0.4465276599,0.2701758146,0.3022372723,0.2959749699,-0.0437552482,0.6722488999,-0.0536569208,0.099635683,-0.0690652728,-0.3176445663,-0.0269642416,0.1578307003,0.3953226507,0.2114282846,-0.2491105497,0.0748830065,-0.1284852326,0.1337044984,0.1330321282,0.0128994184,-0.2782158554,0.0685418844,-0.1654132903,0.4030524194,-0.1301134676,0.1823724061,0.0637911186,0.2705146372,-0.3036566377,-0.3571579456,0.3911989331,-0.3434489071,-0.4585971534,-0.4276721179,0.2904642522,0.159820199,-0.2049567997,-0.4742090106,0.0093041481,0.2549891472,-0.1533083916,-0.1555007994,0.5398080945,0.0699544102,0.2142618448,-0.1219963953,0.2098635286,0.0046629333,-0.1189015582,-0.1165047884,-0.0643654764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2695","title":"Cannot import load_dataset on Colab","comments":"Hi @bayartsogt-ya and @phosseini, thanks for reporting.\r\n\r\nWe are fixing this critical issue and making an urgent patch release of the `datasets` library today.\r\n\r\nIn the meantime, as pointed out by @mariosasko, you can circumvent this issue by updating the `tqdm` library: \r\n```\r\n!pip install -U tqdm\r\n```","body":"## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n","comment_length":48,"text":"Cannot import load_dataset on Colab\n## Describe the bug\r\nGot tqdm concurrent module not found error during importing load_dataset from datasets.\r\n\r\n## Steps to reproduce the bug\r\nHere [colab notebook](https:\/\/colab.research.google.com\/drive\/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error\r\n\r\nOn colab:\r\n```python\r\n!pip install datasets\r\nfrom datasets import load_dataset\r\n```\r\n\r\n## Expected results\r\nWorks without error\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n```\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value\r\n 2 from sklearn.metrics import mean_squared_error\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/__init__.py in ()\r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in ()\r\n 40 from tqdm.auto import tqdm\r\n 41 \r\n---> 42 from datasets.tasks.text_classification import TextClassification\r\n 43 \r\n 44 from . import config, utils\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/tasks\/__init__.py in ()\r\n 1 from typing import Optional\r\n 2 \r\n----> 3 from ..utils.logging import get_logger\r\n 4 from .automatic_speech_recognition import AutomaticSpeechRecognition\r\n 5 from .base import TaskTemplate\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/__init__.py in ()\r\n 19 \r\n 20 from . import logging\r\n---> 21 from .download_manager import DownloadManager, GenerateMode\r\n 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed\r\n 23 from .mock_download_manager import MockDownloadManager\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/download_manager.py in ()\r\n 24 \r\n 25 from .. import config\r\n---> 26 from .file_utils import (\r\n 27 DownloadConfig,\r\n 28 cached_path,\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in ()\r\n 25 import posixpath\r\n 26 import requests\r\n---> 27 from tqdm.contrib.concurrent import thread_map\r\n 28 \r\n 29 from .. import __version__, config, utils\r\n\r\nModuleNotFoundError: No module named 'tqdm.contrib.concurrent'\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.10.0\r\n- Platform: Colab\r\n- Python version: 3.7.11\r\n- PyArrow version: 3.0.0\r\n\nHi @bayartsogt-ya and @phosseini, thanks for reporting.\r\n\r\nWe are fixing this critical issue and making an urgent patch release of the `datasets` library today.\r\n\r\nIn the meantime, as pointed out by @mariosasko, you can circumvent this issue by updating the `tqdm` library: \r\n```\r\n!pip install -U tqdm\r\n```","embeddings":[-0.4603253901,-0.2148275971,-0.024375163,0.2682137489,0.1354128718,0.051994808,0.4973769188,-0.1074850261,0.1983702779,0.0849533305,-0.3266307712,0.4753232598,-0.1742698699,0.1487485617,-0.2584459782,0.0984701291,-0.0661861375,0.058203578,-0.2907032669,0.0956151709,0.0202920046,0.3308885992,-0.6335685253,-0.3468577266,-0.5786883831,-0.2100416869,-0.0277418792,0.168791458,-0.2702196836,-0.292011261,0.3478223681,0.1770586967,0.0921896547,0.5697174668,-0.0001172403,0.0715135038,0.3933092654,-0.0787535235,-0.2456898838,-0.3918715715,-0.3461531103,-0.3291964531,0.3686217368,-0.1073928103,0.1058271378,0.404165864,-0.1357311457,-0.133275196,0.2626952529,0.1352794766,0.1936342865,0.549197793,0.1630603671,-0.192492187,-0.1125865728,-0.193378076,-0.2020825595,0.7461523414,0.4174177051,0.0168563686,-0.0510248244,0.2604947388,-0.0992677361,0.2890874445,0.1862412095,-0.0334219635,0.1051012799,-0.4805828929,0.1006372571,0.1743771285,0.4806717336,-0.233426556,0.0443901122,-0.1220246106,0.2094738036,-0.2924674749,0.2619004846,-0.0985698551,-0.2118597776,0.0490080118,-0.0226965509,-0.0896168202,-0.0123074474,0.0164376963,-0.0933856592,0.3655072451,-0.1430076212,-0.1118264869,0.0698661357,-0.1096417829,0.5018094778,-0.0161823332,-0.0268674623,0.2063278407,-0.468000561,-0.0093295779,-0.098870948,-0.1408492327,-0.0516990833,-0.1014948189,-0.0361867324,0.0174666177,-0.1110087708,0.303483665,0.1848023385,0.1422498971,0.0347587094,0.0182988103,0.1226537004,0.2864284515,-0.2007629573,0.227394864,-0.259282738,-0.3118798137,0.1845084578,-0.029214235,0.2715788782,-0.0037106252,-0.319062084,-0.1818456352,0.0095681679,0.105042167,0.0446227603,0.3990536928,0.055795569,0.1057089195,-0.0346219428,0.0016938471,-0.3566534519,-0.2394378334,-0.0538450666,0.1936965287,-0.0642801523,-0.1152990013,0.0678034425,-0.1549987495,0.2601871789,-0.0124206394,0.3092051148,-0.0599603131,0.1036741883,-0.3025387824,-0.0606454387,0.331908226,0.1287349612,0.0872022063,-0.0623135045,-0.2055783719,-0.0778879523,0.2355169952,-0.4770620167,-0.1442362964,-0.0243145749,0.1429465264,-0.1339146942,-0.2244311124,-0.5850232244,0.1648343056,0.1447636336,-0.0812816098,-0.0849623531,-0.3062829673,-0.2330714166,-0.1535831094,0.1236880049,0.4482834041,-0.2309288979,-0.142314136,-0.2176906466,0.1205273345,0.1272705644,0.234823361,-0.2021541893,0.1736088246,-0.0962642506,-0.1171419844,0.490647763,-0.4326361418,-0.4671251178,-0.0203478895,-0.2660758793,-0.0052545536,-0.0897063687,0.0485790744,0.3577234745,-0.0669410899,0.5458390117,0.2629367113,-0.1390330195,-0.036604546,-0.1640905291,-0.4099052548,0.1294252127,0.0418195315,0.1834629029,-0.2177709639,0.203032732,-0.0533746518,0.1012382284,0.1859944463,-0.1211299896,0.3843335509,0.2429013252,-0.2223784029,0.0372738875,-0.1859572828,-0.4775209725,0.2240482867,0.409294188,0.0799716562,-0.0674993992,0.05132493,-0.4381961823,0.0791065395,-0.2594439685,0.0676428005,0.0994127542,-0.0666698813,-0.1751303822,0.0889260471,-0.2604830265,0.7403804064,0.0377074406,0.2175934911,-0.3183411956,0.2157882154,-0.1859039217,-0.3081798851,0.1762920618,0.1361715198,0.0753254518,-0.092272602,-0.2389075756,0.1421211809,0.2554138899,0.1713412106,0.061892774,-0.2776550353,0.4123665094,-0.274011761,0.0123248948,-0.1066840887,0.208570227,0.0058228662,0.2551185191,0.1565714628,-0.3707030118,-0.0410736091,-0.0305118244,0.0905382559,0.3726439178,0.1239964738,-0.0050928039,-0.1856487691,0.325183779,0.2222952098,0.192474246,0.0937420502,-0.1297319233,0.2881348133,0.3061757982,0.2452441901,0.038779363,-0.2685610354,-0.2554307878,0.0633648261,0.1922871321,0.2518091798,0.4446454048,0.1803891063,0.1049195379,0.0316309184,0.0318480134,-0.1561172903,0.1151199043,0.2919997871,0.2558412552,0.1311028153,0.4120824635,0.0967440531,-0.078998588,-0.3894120455,0.1753250211,0.1503109038,-0.3317721486,0.0766265243,-0.2146017402,-0.113947168,-0.1419459134,-0.4253391027,-0.1464539021,-0.1978348494,-0.1058005691,0.2810257375,0.2650622427,0.1936620623,-0.0239940546,0.0500169992,0.0625045747,-0.142483592,-0.0558168627,-0.2506769001,-0.0925878361,0.0015159415,0.327739507,0.246811226,0.3685063124,-0.3038308024,-0.0695626289,0.1437006891,-0.332098037,0.1763451695,-0.005682677,0.2919798791,-0.0134428795,-0.0779709816,-0.3206300437,-0.1939175874,0.0215363186,-0.1346231997,-0.1827463955,0.0472795218,-0.1800287366,-0.2297740728,-0.0019908124,-0.2926879525,-0.4344463348,-0.3998571932,-0.3682101369,0.0472466946,-0.0089030508,0.4603004754,-0.0427159443,0.2933925688,0.2250700146,0.0508381985,0.0179356467,-0.3330911994,0.4733818173,-0.2865113616,-0.2936578393,0.2801209092,-0.2680461109,0.3836089373,0.0181830842,-0.1104436517,-0.2135308832,-0.0214798469,0.3837370574,-0.1009398997,0.0416830219,0.0846360251,0.1136619002,0.0782582983,-0.092406489,-0.1815526336,-0.2040507346,0.2196319252,0.1930779815,-0.1657154113,0.2265984565,-0.0363357775,0.8421141505,0.0771993697,-0.0400505289,0.3015678525,-0.061655391,0.3064921498,-0.1064245924,-0.5121492743,-0.0696846843,-0.1073111445,-0.039284341,0.0819721445,-0.1839754134,-0.1112223938,-0.2847859561,-0.2727975249,-0.3160193861,-0.291852355,-0.1147850677,-0.2003910989,0.1308543533,0.1986790299,-0.0612915456,-0.1620080918,-0.1124967635,0.0489822105,0.4355244339,-0.4208028316,-0.0186146274,0.3795593679,-0.0728556365,-0.3464918733,0.1986530572,0.1317849904,0.3303270042,0.0766258389,0.1182002276,0.0601407997,-0.0493247956,0.1122457087,0.05919642,-0.0800003186,0.2366157919,-0.0146941952,-0.4012733102,-0.0577438287,-0.2255629599,0.2240988016,0.4793192744,0.0861525759,-0.2667843401,0.2115175128,0.2363318503,0.1412735432,-0.114222765,-0.1485226005,-0.4387521744,-0.3134378195,-0.378410399,-0.0600447692,0.1850457788,0.2391748279,0.0416690968,-0.0708694831,-0.0640273839,-0.2169730812,0.0610414371,0.3596898913,0.1482839882,0.270448029,0.5801549554,0.1271515787,-0.2473592013,0.4470187426,0.5865612626,0.0028329962,-0.5835172534,-0.0279992204,-0.1797391474,0.2348893136,0.1315195262,-0.0706735849,0.0163302217,0.0595659725,-0.1358518749,0.3848748207,0.3812657893,-0.0635423213,0.1262477934,-0.4026652873,-0.1285912842,0.0958481207,0.0007427831,0.1376991123,0.4601660669,-0.3796385229,-0.0598220304,0.047099065,-0.1843536198,0.686365068,-0.0381890498,-0.2390154898,0.2732535005,-0.1615026444,0.2652589977,0.1131904423,0.1857605428,-0.2762228251,-0.3076212406,0.0923932046,-0.2266215682,0.1900261045,-0.1945811212,-0.3511314392,0.2596141994,0.0564561076,-0.1498299241,-0.0185761657,0.0406053141,0.0024130014,-0.2937503755,-0.3525202274,0.0517692231,-0.1905304193,0.2061717808,-0.1972419322,-0.1833461672,-0.1389480531,-0.1865322441,-0.0331040733,-0.1090358868,-0.6445795894,0.2370983511,0.3314474523,-0.3087444007,0.2005167454,-0.1111385226,0.0950988606,-0.0293304492,-0.1739497632,0.1265123934,0.1993344128,0.0166862272,-0.1108888164,-0.172952801,0.4197721481,-0.0839659944,0.0050038416,0.0576662868,0.0366229266,-0.0225345157,0.4702349007,0.3512617946,0.3681568801,0.150692597,-0.0755758658,0.1810477078,0.26167503,-0.1958780736,0.0780422837,0.1019752249,-0.1530845314,0.2558067143,-0.107513763,-0.1003244594,-0.0019221834,0.3397724032,0.1296715736,-0.0161001608,0.1832213998,0.1551724821,-0.2558426857,-0.0844762251,0.1688349247,-0.3087925613,-0.3004136086,0.2412767708,-0.0199141074,0.0491270311,0.0744944811,0.7340957522,0.2456151992,-0.3685689867,-0.132078737,-0.3772931397,-0.5326016545,0.1417692751,-0.1205238625,-0.0949431732,0.2369502336,0.3836895525,0.120868288,0.1284108609,-0.2208905369,0.1052405685,-0.1584761292,0.1540038735,0.2756511569,-0.0960519463,0.0372927338,0.0859900042,0.0042420314,0.1529782265,0.0671002194,-0.2249542475,-0.1146583706,0.1326806545,-0.1025531739,0.0408711247,-0.1180261746,-0.1512853205,-0.2300352007,-0.1715027094,0.1590514779,0.0351236463,-0.0818419009,-0.1321561933,0.2920054197,-0.0966290534,-0.2174018025,0.4025671184,0.1177951023,0.0810081363,0.0776668489,0.1660300344,0.1665427387,0.1832786202,0.017530106,-0.2253397107,-0.091671519,-0.1765127629,0.0942778215,-0.3320890367,-0.0140226921,0.0699948967,-0.0275069047,0.4203799963,-0.128956452,-0.2099847645,0.0472415946,0.0857807249,-0.3888076842,0.0311133061,0.3042639792,0.154858917,0.175432682,0.2722882032,0.08177194,0.0818850547,0.1544100046,0.3952516615,0.2447241694,-0.0228369981,0.2661114633,-0.2058007568,0.0857820436,0.2306343317,0.2361091226,0.118432261,0.0381054915,0.1526700258,-0.1501827389,0.295978874,0.1446438283,-0.1672739387,0.2365310639,-0.3596028984,0.2424565107,-0.0986769199,0.0569087155,-0.1363285184,0.141786322,0.3276461959,-0.0491868742,-0.0884468332,-0.1608537436,-0.0030408807,-0.300190568,0.2709409893,0.4638240635,-0.0976310596,-0.0931242406,-0.043175891,0.2145775259,0.217528671,0.5374786854,0.3159116209,-0.1668998748,-0.4896293879,-0.1224408448,0.5082057714,-0.1266734451,-0.1844531447,0.4623732269,-0.134871155,0.0100118555,-0.165280953,0.3222151995,0.4884846807,0.3006311953,-0.1383483261,-0.2075260133,0.2566275001,-0.1801053286,-0.1383739263,0.2107214928,0.0596846007,0.2941676378,0.2981710732,0.0842535272,-0.1690877378,-0.0320378505,0.2207306325,-0.3357616365,-0.4409199655,0.206428498,-0.1567060947,-0.0026589187,-0.1897845715,0.0980003253,-0.4563895464,0.2288242429,0.315672189,-0.2224390507,-0.0670025125,-0.1489631385,0.0449592993,-0.0307994634,0.3605282903,0.2467247546,0.3562105,-0.2011301219,-0.1037739813,-0.8396953344,0.3146013319,0.0336657204,-0.0117303068,-0.0385530032,0.1393253952,-0.0011307682,0.3212940991,0.420112431,0.0454290248,0.1901207864,0.0372285321,-0.462015748,-0.2717567384,0.0900407732,-0.445476234,0.0050567961,-0.3917895555,-0.0040985593,-0.2824522853,-0.0414680019,-0.0273984112,-0.4465276599,0.2701758146,0.3022372723,0.2959749699,-0.0437552482,0.6722488999,-0.0536569208,0.099635683,-0.0690652728,-0.3176445663,-0.0269642416,0.1578307003,0.3953226507,0.2114282846,-0.2491105497,0.0748830065,-0.1284852326,0.1337044984,0.1330321282,0.0128994184,-0.2782158554,0.0685418844,-0.1654132903,0.4030524194,-0.1301134676,0.1823724061,0.0637911186,0.2705146372,-0.3036566377,-0.3571579456,0.3911989331,-0.3434489071,-0.4585971534,-0.4276721179,0.2904642522,0.159820199,-0.2049567997,-0.4742090106,0.0093041481,0.2549891472,-0.1533083916,-0.1555007994,0.5398080945,0.0699544102,0.2142618448,-0.1219963953,0.2098635286,0.0046629333,-0.1189015582,-0.1165047884,-0.0643654764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2691","title":"xtreme \/ pan-x cannot be downloaded","comments":"Hi @severo, thanks for reporting.\r\n\r\nHowever I have not been able to reproduce this issue. Could you please confirm if the problem persists for you?\r\n\r\nMaybe Dropbox (where the data source is hosted) was temporarily unavailable when you tried.","body":"## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":39,"text":"xtreme \/ pan-x cannot be downloaded\n## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nHi @severo, thanks for reporting.\r\n\r\nHowever I have not been able to reproduce this issue. Could you please confirm if the problem persists for you?\r\n\r\nMaybe Dropbox (where the data source is hosted) was temporarily unavailable when you tried.","embeddings":[-0.352686584,-0.386775732,-0.0510648638,0.2910829186,0.2072106749,0.0486671701,-0.116992861,0.2295708954,0.1450724155,0.0749082938,-0.2071777731,0.2856611907,0.0034074627,0.0860847235,0.196489796,-0.2445528358,0.070219554,0.1196101531,0.0714935958,-0.0395350382,-0.0941303298,0.1315491349,-0.2260093987,0.0777683258,-0.2466472387,0.126108855,0.0916940719,0.2066926062,-0.3236722052,-0.4029152989,0.6594854593,-0.0882715508,0.2548996508,0.6138737202,-0.0001095451,0.0444418713,0.2389015704,-0.0772670582,0.008579297,-0.5894992352,-0.3442205787,-0.4234595001,-0.0681286678,-0.182247594,-0.0317808576,-0.1360592842,-0.1195598543,-0.126975894,0.304120928,0.2758345604,0.2564079463,0.4237165451,0.2048839927,-0.1892297268,0.424253881,0.0090922257,-0.3090731204,0.2538466454,0.4219304025,0.1650082916,0.4848668873,0.2082423866,-0.0695545673,0.0420786776,0.129794836,0.0679554939,-0.1117196828,-0.4720572829,0.0596620217,0.3128014803,0.4577904046,-0.2595132291,-0.3999505937,-0.1368682683,0.1110483035,-0.2919364572,0.3227855265,0.2403816432,-0.0112175364,0.0619956963,0.0464067198,-0.0516867079,-0.0386229977,0.1513980031,-0.042721048,0.2568822503,-0.2622476518,-0.0077699278,0.1031031981,-0.0465255752,0.3317069113,-0.0003910958,-0.0721538886,0.1808130741,-0.2971533835,0.033614397,-0.081234023,-0.1032688469,0.325927943,-0.0659449324,-0.1524967253,-0.0552354939,-0.1848699003,0.0646039173,0.2071894258,0.2051487714,0.0498812981,0.0647757202,0.2026147544,0.1957772225,0.0656497926,-0.0646910891,-0.1810179353,-0.0279590804,0.0963962451,-0.0940713957,0.4540398717,-0.3807743788,-0.3451695442,-0.0102128312,-0.0039379629,0.0608344525,0.1219215095,0.1474575251,-0.0968996659,0.3871399462,-0.104413569,0.3990936875,-0.2333694994,-0.042109821,-0.1832315773,0.353790462,-0.2850381434,-0.201285556,0.2524963915,-0.2233882695,0.3617868721,0.1000485346,0.3133207262,-0.20624955,0.042804569,-0.0248363614,-0.352999717,0.4104738533,0.1801795512,0.2340242118,-0.1464570761,0.0782419667,-0.1296510547,0.1654407978,-0.6251087785,-0.0875181332,-0.098489657,0.1942233145,-0.2081999928,-0.1046866104,-0.6656982303,-0.1676555723,-0.0900125429,0.1431583315,0.0834209844,-0.2053960711,0.1542958617,-0.1790491939,0.2005880773,0.2411441654,-0.2552346289,0.0609929524,-0.1533209234,-0.2647586763,-0.0765785873,0.126275599,-0.2085569054,0.010368458,-0.3422111869,0.323212862,0.4552656114,-0.6140804887,-0.5974395871,-0.1052325368,-0.2622517943,0.0898787528,-0.0617282391,0.1077232808,-0.0357164182,-0.0788005739,0.1014333293,0.2451176792,0.0692812279,-0.0631380528,-0.1651225388,-0.2046420574,-0.3625925779,0.3529902399,0.2083215714,-0.0578963794,0.078610681,0.0761333331,0.0877819508,-0.0318486802,-0.0306234825,0.4155043662,0.3037109375,-0.0001337734,-0.1492662728,-0.36823228,-0.3647809327,0.3155393898,-0.0051267017,-0.2266247869,-0.016196372,-0.0862171724,-0.4500852823,0.1982221305,-0.2209279835,-0.0606530905,0.0769481286,0.0073385714,0.2201862037,0.0836802647,-0.2290256172,0.1458614469,0.0163225271,0.1431554854,-0.4204463959,0.3151150942,-0.048910588,-0.1553777009,0.0702609345,0.0526242554,0.2588835657,-0.2040776908,-0.1541559994,0.2960607111,-0.1171703413,0.0546539836,0.2451883256,0.1121123135,0.1407469213,-0.5683162808,0.2428395599,0.0568337068,0.0048016249,0.1328017712,-0.080193989,0.3285500109,0.0742878169,-0.1284792274,-0.144333154,0.1246474534,0.3701762557,-0.1742715091,0.0623171479,-0.0386976339,0.3798928857,-0.1570974737,0.2001883686,-0.1817019731,-0.3316303194,0.0980749577,0.2502304316,-0.1069736704,0.1860957593,0.1444752514,-0.1026409641,0.1593287289,-0.0966135561,0.1676412821,0.5179623365,0.1285328418,-0.006795112,0.0641476959,0.008289258,-0.1184643656,0.0590544976,0.1046777964,-0.323988378,0.0827619657,0.1612743586,0.1036033481,-0.225023374,-0.1348782033,-0.1345205456,0.0970061123,-0.1478446126,-0.0257360358,-0.1935744584,-0.1062394008,-0.0235615224,-0.062759541,-0.26309219,-0.1395761967,-0.0624651127,0.321655184,0.3296003342,0.1525023878,-0.3904215693,-0.0954002142,0.0352769271,-0.3149357736,-0.2127218843,0.3404073715,-0.0474719182,0.0662847012,0.0962966308,0.0808142498,0.2356974632,-0.370800525,0.2722423375,-0.4419664443,-0.2029943317,0.0966422856,-0.2410521209,0.4696161151,0.1120234504,0.126544103,-0.0322053209,0.2229399383,0.3095308542,-0.4043672681,-0.0039248331,0.0307903998,0.0260589682,-0.1758607477,0.0369341634,-0.1270143539,-0.3618342578,-0.3396022618,0.0283496343,0.080137223,-0.0221556723,0.0101633742,0.0757040083,-0.0299586933,-0.1044041738,-0.1533589661,-0.2170976549,-0.7377814054,0.5196609497,-0.2019710839,-0.4470081627,0.3016469777,0.0556515008,0.2421492785,0.0431420989,-0.5795800686,-0.3328723907,-0.20937334,0.3857634068,0.0262414403,0.172588855,0.179884553,-0.3582758904,0.0249791555,-0.2073284984,-0.240735516,-0.3274493814,0.1376802623,0.3069241941,0.0616082996,0.1685754955,-0.1148670241,0.4480379224,0.1403064281,0.1453653872,0.5500609875,-0.0225298703,0.325284332,-0.1545571536,-0.3528732061,-0.1701274067,0.0941827074,0.119336538,-0.0195563305,0.2615587115,0.0613196641,-0.3054994345,-0.1316119879,-0.0387074724,-0.0832650438,-0.0314533338,0.1191357151,0.1137526259,0.0811114311,0.2296646982,-0.1488065571,-0.1825037003,0.0221286044,0.4156580269,0.1601018459,0.2835165858,-0.5198410749,0.1627908796,-0.346342504,0.2180436999,-0.0051472425,0.3751677871,0.0005981216,0.1822769344,0.2526132166,0.0190845039,0.3365484476,-0.1874768883,0.2626715004,0.1567421108,-0.1884523928,-0.2643436491,-0.1834802032,-0.0880471542,0.1510481238,0.2219396681,0.4319796264,-0.2872444391,-0.2112279087,0.2268488109,0.2067936063,0.0241378546,-0.1701164395,-0.2435171902,-0.5633920431,-0.1476865113,-0.1200377643,0.2404229939,0.055602394,0.1601883769,0.0037757247,-0.2464236468,-0.0884156972,0.1463684291,0.1705907583,0.3395799696,-0.1337962002,0.3536367416,0.1435116082,-0.3297019303,0.2670310438,0.7374875546,0.0058287084,-0.4494514167,0.0718678534,-0.0187181961,-0.0781944618,0.1143369526,-0.2037746608,-0.0973353833,-0.2186990976,0.0812803879,0.0257368647,0.1539977491,0.3771669567,0.0625731349,-0.5213450193,-0.3220100701,0.5426341295,0.2666315138,0.1196527258,0.2263086885,-0.0863892734,0.0470355637,-0.225603655,-0.0838141739,0.9448820353,-0.0112380087,0.0339338742,0.1245707795,0.0145492079,0.1615138054,-0.1924783438,0.0699113756,-0.1522158235,-0.4263075292,-0.1463764012,-0.1881897599,0.1286133081,-0.057763584,0.0621434599,0.2255205661,-0.072600022,0.1192601398,0.0648187473,0.0723348632,-0.1526958644,-0.1707778722,-0.517134726,0.1991241872,-0.0697473362,0.4064856768,-0.1388291866,-0.1040899828,-0.1983234137,-0.0959432647,-0.1218509376,0.2916822731,0.0023223299,0.071559839,0.0579023622,-0.6102405787,0.0385804325,0.052686926,0.0162253845,0.1011554897,-0.265935123,0.3533175886,-0.0021148229,-0.2266723812,0.1472969353,-0.1359361559,-0.153122738,-0.0944726095,-0.1086312085,0.1762408763,-0.0828933343,-0.1928375512,-0.2500218451,0.0991540179,-0.1647954732,-0.0987405181,-0.1774946898,-0.1970985234,0.3366283178,-0.2021453232,0.1516023129,0.2968964875,0.0558715612,0.0631206557,0.0386187993,-0.2568100393,-0.147399798,0.4189608097,-0.0821872801,-0.0837995782,0.113778539,0.1257390678,0.0002201169,-0.1332896799,0.1391722262,0.058611054,-0.6309772134,0.0441202596,-0.1751818359,0.088024877,-0.1130462959,0.3112160563,0.2554554343,-0.3791319132,0.2418451756,-0.6552193761,-0.0697513446,0.094566375,0.02273551,0.2029662281,0.0134273432,0.076626651,0.1292037219,-0.2599191666,-0.3166564405,0.1748611331,-0.0904707015,-0.1386951953,0.2489341348,0.1542686224,0.2799359858,-0.0008410558,0.1445420086,0.0790599883,-0.3133639395,-0.1696363091,-0.0390883312,0.0975604132,0.0073128967,-0.1443008631,0.0250957478,-0.1934474558,-0.2018469274,-0.0162708741,0.0714772269,0.2128660381,0.0548099764,-0.17678985,-0.1202124283,0.0524005145,-0.2712775171,0.3782560825,0.1526099443,0.1362899691,-0.0707215443,0.1698892564,-0.0045894925,0.0377819911,-0.3507540822,-0.1537237018,-0.0099876132,0.323434025,0.3173137009,-0.2338351011,0.3399527669,0.5180972219,0.2589370608,0.3506008685,-0.3326706886,0.2288190275,0.2044676542,0.2669626176,-0.3421081901,-0.197278887,0.3491127789,0.3381572962,-0.0222106166,0.3334654868,0.2860367,0.099935174,0.2936148942,0.1892458647,0.7017909288,0.1483001411,0.3616818786,0.4879673421,0.045340661,0.0547546484,0.4047198594,0.2209099531,0.0589066148,0.5563641191,-0.1501090676,0.3705192208,0.0276990198,0.1410739273,-0.1736342758,-0.2919469178,-0.2292747796,-0.0300019234,0.0328438506,0.0066141635,-0.1024404988,0.0717507079,-0.3785607517,-0.0442164168,-0.372066319,0.2042671293,-0.1073924899,0.1628200859,0.1005004719,-0.2054411918,-0.0019082302,0.0046277586,-0.010632338,0.0236769319,0.2068842798,0.0598007068,-0.2006712854,-0.2601106167,0.014614257,0.3534215987,-0.0258855745,-0.1044208482,0.4160717726,0.0979037359,-0.1558293104,0.0114982836,0.3531677127,0.3471859694,0.1551682949,0.1355300844,-0.0455287211,0.087898083,-0.0654704198,-0.0037283888,0.2300325334,0.0325547531,0.3787957132,0.213646695,0.2474648207,-0.2060374171,0.2166291624,-0.1780153364,0.1044225991,-0.3161133528,0.0276570953,-0.3075526059,-0.1017176583,-0.2598287761,-0.2114053965,-0.1291494519,0.080933392,0.2977313995,0.0193513911,0.0042310371,-0.3231406212,0.0737729371,-0.1102203131,0.1380382031,0.4347139895,-0.0541046672,-0.1887927055,-0.1329211593,-0.8249322176,0.2049012333,-0.3134781718,-0.3593265414,0.2230780721,-0.1967066079,-0.1337318569,0.2950107455,0.3823513389,0.1986643225,0.1417559981,0.2696788907,-0.3786462247,-0.1755008548,0.11647515,0.0805224925,0.1550345868,-0.196610257,0.002044295,-0.0605066828,0.0296265688,-0.1787052453,-0.1014791131,-0.1546016037,-0.2730810344,0.7128177285,-0.0265821442,0.1818777025,0.0089348666,-0.1415954083,-0.3479325175,-0.3359898627,-0.1717506349,0.1654935479,0.0380567461,0.2271931022,-0.1851948351,0.0454304218,-0.4097551405,0.11654035,0.0190046187,-0.0313896202,-0.1484308541,0.1617601812,-0.3718978465,0.1241318285,0.2702310383,0.2067091763,-0.0751156583,0.2049667686,-0.137139976,-0.4007472396,0.3941509724,-0.0630495399,-0.2641801238,-0.0338023342,0.0910274461,-0.0175066758,-0.2228853703,-0.2046363056,0.322298497,0.3259679377,0.0417653024,-0.0742104501,0.1801104099,-0.1593890488,-0.0055480492,0.1667849571,0.56791085,-0.0529464856,-0.166253984,0.4779118598,0.0644990727]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2691","title":"xtreme \/ pan-x cannot be downloaded","comments":"Hmmm, the file (https:\/\/www.dropbox.com\/s\/dl\/12h3qqog6q4bjve\/panx_dataset.tar) really seems to be unavailable... I tried from various connexions and machines and got the same 404 error. Maybe the dataset has been loaded from the cache in your case?","body":"## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":34,"text":"xtreme \/ pan-x cannot be downloaded\n## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nHmmm, the file (https:\/\/www.dropbox.com\/s\/dl\/12h3qqog6q4bjve\/panx_dataset.tar) really seems to be unavailable... I tried from various connexions and machines and got the same 404 error. Maybe the dataset has been loaded from the cache in your case?","embeddings":[-0.2679742277,-0.4418168366,-0.1176263168,0.2636517584,0.2272869796,0.1094410568,-0.1848932803,0.2379712462,0.1063496396,0.1472171843,-0.1563756764,0.2612072229,0.0340300761,0.0054809009,0.1887651682,-0.29292804,-0.0231135618,0.157192722,0.085480608,-0.0127716027,-0.04482596,0.1623114944,-0.1777581871,0.1660579741,-0.2347056568,0.0610586517,0.0814340189,0.1718645096,-0.3454205096,-0.4189862609,0.5661595464,-0.0949465558,0.2182051539,0.7195156813,-0.0001037239,0.0292130355,0.2176596671,-0.0261945613,0.0566013306,-0.4336361289,-0.3120516241,-0.2680827677,-0.0547174066,-0.2174528837,-0.1400291026,-0.1007528901,-0.0880559012,-0.1750144213,0.2505108416,0.3105109632,0.3232139647,0.3803195655,0.1769934148,-0.20664078,0.3996114135,-0.094190076,-0.3172445893,0.0890595615,0.2912508249,0.1761108935,0.5273684859,0.3237523437,-0.0359906144,0.0154759418,0.1438701749,0.0410690084,-0.062304955,-0.4709228575,0.0612652376,0.3000105917,0.4271570444,-0.2391207963,-0.3885632753,-0.1054575145,0.1049027145,-0.4078127444,0.2444296181,0.1799363494,-0.016278049,0.1052497625,0.0096645588,-0.0978146642,-0.0419347472,0.0619605519,-0.0776690245,0.2221194804,-0.2184873968,-0.0683026388,0.0588096492,-0.0759769455,0.2078842819,0.0511512533,-0.0275817458,0.1276533753,-0.2452935725,0.0600650944,0.0632193387,-0.1291561872,0.2096273601,-0.0791392401,-0.0892370194,0.0114943245,-0.1473243833,0.0619761087,0.0659476146,0.2439481467,0.0383303538,0.0237593241,0.2091334462,0.1862054467,0.0814153999,-0.1025757566,-0.1737058163,0.0113051254,0.1954597384,-0.0294615906,0.3959265947,-0.3529497981,-0.2808386385,0.0221817661,0.0397591889,0.0408062376,0.0866130665,0.2145688236,-0.0824295655,0.4011310637,-0.1501498818,0.3440098464,-0.2506442964,0.0770461783,-0.2169536054,0.2924671173,-0.2931419313,-0.1526376605,0.1920557916,-0.0812762156,0.3481796086,-0.0350379422,0.3502663672,-0.1696178615,0.1593687534,0.0612697862,-0.2728770971,0.4204449952,0.1726548076,0.1674812585,-0.0856520236,0.0686986148,-0.1584209353,0.0594495535,-0.5414469242,-0.2124148011,-0.0291898306,0.298630625,-0.1243534461,-0.0292220488,-0.5017438531,-0.2424408942,-0.0926839411,0.2001235187,0.0688052848,-0.2365992665,0.1697010547,-0.2065946758,0.1218228713,0.1404036134,-0.1871487647,0.0682941899,-0.0762738213,-0.2601469457,-0.1281082034,0.0846763924,-0.1892635971,0.0538473502,-0.3212268054,0.2798063755,0.5203265548,-0.6607077718,-0.5881974101,-0.1088319495,-0.3328568339,0.0381410457,-0.032900013,0.1168181747,-0.098201707,0.0026601506,0.2206364572,0.2092215717,0.0729324967,-0.0459184907,-0.1469891369,-0.1996123493,-0.3954456747,0.2889850438,0.2185926884,-0.115557,0.0630726442,0.0086837495,0.1266668141,-0.0867342278,-0.0569839217,0.4788995385,0.3095561564,0.0865387619,-0.0792881995,-0.3216750026,-0.2247870266,0.2472707182,-0.0705728233,-0.0963476226,-0.0327415653,-0.1887047589,-0.4533225,0.143108353,-0.3158530295,-0.1230281666,0.1752057672,0.1260690987,0.2419695854,0.0138474563,-0.2411731631,0.0591528639,0.1169226617,0.1251550317,-0.4320315719,0.2213039547,-0.1491878033,-0.1053038836,0.0922812745,0.1673108488,0.2437001467,-0.2250935286,-0.0908466578,0.324670881,-0.096702151,0.0355161875,0.3533547521,0.0978430361,0.0845166221,-0.5908098221,0.2334867269,0.1035210714,-0.0187885165,0.1300654113,-0.0431514904,0.3024262488,0.1514284164,-0.1150512546,-0.0487053618,0.0742619485,0.4298829436,-0.1736487895,0.0714132115,-0.1306804419,0.2847753465,-0.1419808865,0.1040666774,-0.229980275,-0.2976059914,0.2326201349,0.3081139028,-0.1277648509,0.1826978624,0.0562879071,-0.0640333146,0.1722058952,-0.0551215112,0.1123171449,0.3923814595,0.1955979764,0.0783934593,0.0904865712,0.0251341909,-0.0745071098,0.0514602475,0.0309503525,-0.2735628784,0.007225696,0.1577754319,0.0551329926,-0.2244767994,-0.219876349,-0.1738971472,0.126578778,-0.0564300008,-0.0923083276,-0.2825479507,-0.1587558985,-0.0015430114,-0.0670296699,-0.2526369691,-0.2226977199,0.0491071939,0.2205242068,0.2909301519,0.1391824782,-0.4456433356,-0.1586828381,0.1205253229,-0.3013077974,-0.1710583866,0.2445183694,-0.0701966882,0.1309379637,-0.0495745577,-0.0147706587,0.2298841774,-0.403102845,0.2660707533,-0.4210811555,-0.2356017977,0.1419922858,-0.2103907466,0.2974480689,0.1424486339,0.0910667479,-0.0189679265,0.1650261581,0.3228319585,-0.4411507547,-0.0466851778,0.0402265303,0.0551933236,-0.2243185937,0.0146075087,-0.1893948317,-0.3284139037,-0.3596836627,-0.0406748131,0.1506467164,0.0344251506,-0.018506743,0.0803386122,0.0885249823,-0.0685613677,-0.0511880368,-0.2514695823,-0.7017880082,0.4636758566,-0.2644920349,-0.4040011168,0.403811574,0.0343751051,0.2874763906,0.0281600729,-0.6414683461,-0.426507473,-0.2081430703,0.4057197571,0.0910308436,0.1623046994,0.248888731,-0.1939970553,-0.0804983974,-0.1296487451,-0.2731976211,-0.3393236399,0.1231079474,0.351418674,0.0049005449,0.1693899184,-0.0566528216,0.4797626734,0.1411705464,0.2660491765,0.4828196764,0.0074648089,0.4104550183,-0.1721997857,-0.2594890296,-0.1180761755,0.0553667545,0.0891867504,-0.0057335733,0.3468691707,0.032502301,-0.3200868368,-0.082382232,-0.1269824952,-0.1169473007,-0.13468045,0.1619885862,0.0826363564,0.1031515896,0.2671741843,-0.2168583423,-0.056906905,0.100930728,0.419354856,0.0583784021,0.2410030514,-0.6325910687,0.1531001627,-0.505128324,0.2458711714,-0.0115502262,0.2705217302,0.0215680413,0.0908030123,0.2392057031,-0.029436009,0.2986411452,-0.2792760134,0.2141706347,0.2215055376,-0.1775688082,-0.1594884992,-0.1793937981,-0.1052881777,0.1101509929,0.2103469819,0.357773602,-0.2909290493,-0.2092893422,0.2081021965,0.1121468693,-0.0364820845,-0.1105238721,-0.2614931166,-0.6037278771,-0.1307971627,-0.1397306919,0.1818604767,0.0638291836,0.1908940077,0.009821916,-0.2043603957,-0.0811322331,0.2022408694,0.2132593393,0.3442692459,-0.1325383782,0.1247722432,0.0889199525,-0.2817154527,0.1870243847,0.6367658973,-0.1025065929,-0.3539457917,0.1042637825,-0.0760079324,-0.0201288592,0.1653687656,-0.1158909351,-0.0652608424,-0.2003659308,0.1965849102,0.0341976173,0.2208952904,0.3908571601,0.0154877417,-0.4292639792,-0.2944345474,0.4085500836,0.2411628366,0.0959233195,0.2109410763,-0.0355041549,-0.0108924098,-0.1219793484,-0.0640270039,0.9632140994,-0.0313608646,0.0339886099,0.1607781947,0.0260238368,0.1442558616,-0.1871311665,0.228703469,-0.1123209149,-0.4142073095,-0.1158689633,-0.1293540895,0.0189195257,-0.066495657,0.04764314,0.3310646415,-0.0027626336,0.1432365924,0.0629528239,0.0580046326,-0.1306233853,-0.1149516702,-0.5382687449,0.2539910078,0.0254720133,0.3790527284,-0.1656429023,-0.1075836569,-0.1945652962,-0.1265368611,-0.0394811109,0.2701869309,0.0465234853,0.0725009441,-0.0001520187,-0.5629300475,-0.1786333472,0.0207466464,0.0313843228,0.1240782365,-0.2292952538,0.3383003473,0.0363531224,-0.256144315,0.0899369046,-0.1675926,-0.1227650642,-0.1348508745,-0.1696887612,0.2684601545,-0.0428851992,-0.2032649368,-0.2238265574,0.0783627629,-0.1164328083,-0.0543870814,-0.0939944386,-0.1266309023,0.2196064442,-0.2141880244,0.2009350657,0.1866669208,-0.072964184,0.0575855821,0.0294680838,-0.2769974172,-0.1553137153,0.3060648143,-0.0045216689,-0.1293376237,0.0959784463,0.1144563407,-0.086870715,-0.184484303,-0.047410693,0.1227229014,-0.4821636379,-0.0282458495,-0.1865506321,-0.0069318614,0.0028719665,0.1874542832,0.2820071578,-0.3644230962,0.1617372781,-0.5957939625,-0.1318205148,0.132511571,0.0234612655,0.2270470709,0.0316791236,0.0467218161,0.1196061969,-0.237975657,-0.401240766,0.1236587316,-0.1498193145,-0.098262094,0.1478442848,0.0064231181,0.2935066223,0.0275849495,0.2196459472,0.0600194484,-0.3998734355,-0.2403431684,0.0552065335,0.0627825782,-0.0383056253,-0.119644098,0.0185990967,-0.1285147816,-0.1325446814,-0.0398795232,0.0891960189,0.1925429106,0.0986529738,-0.1837009639,-0.0692941621,0.0167424213,-0.1642524004,0.325841397,0.1398624033,0.231858179,-0.01449777,0.1323750466,-0.0856100172,0.0066349255,-0.1870064437,-0.1457729042,0.0546046719,0.334092319,0.2917370796,-0.2557477653,0.2778270245,0.5029010773,0.2845908701,0.2882720828,-0.2615942359,0.200641349,0.1876648515,0.3432891369,-0.2725106478,-0.2151306123,0.4091151655,0.3377297819,0.0234304555,0.3495349884,0.2278743237,0.0965239331,0.3864113688,0.2378531098,0.700260818,0.0109213432,0.3064975739,0.3967801332,0.0297815446,0.027974084,0.4111215174,0.2453273535,0.057196565,0.5404456258,-0.1116680428,0.3812044859,0.0508449413,0.1516123414,-0.202122882,-0.1842542887,-0.2625824809,-0.0538458042,0.1176701412,0.05690125,-0.020485986,-0.0404535867,-0.3588003516,-0.0177030507,-0.3818139434,0.1952448934,-0.1012210548,0.1443655938,0.1184099689,-0.2016306967,0.039574232,-0.0301492419,0.0656935722,0.0278113112,0.1514737159,-0.0215782505,-0.1175261959,-0.2632604837,-0.0096970303,0.2329890579,0.0319779329,-0.0951244384,0.3985487819,0.1374702454,-0.0519793443,-0.0433443151,0.3460857868,0.3359360397,0.1591624171,0.0225494057,-0.0402083695,0.0929913297,-0.1229498312,-0.0861296952,0.1511770189,-0.0349759422,0.3215088844,0.2348561585,0.3064635992,-0.2580747008,0.1845092922,-0.2394277155,0.1145098582,-0.2463303208,0.0031750936,-0.2701848745,-0.0975320712,-0.2701835334,-0.1282821,-0.2590593994,0.0486110374,0.2965447307,-0.0721521527,0.0045251325,-0.3530624211,0.1089875698,-0.1153899208,0.1304292679,0.4215750992,0.029853899,-0.1419176906,-0.0996891037,-0.8327545524,0.2654565573,-0.3724475205,-0.3370615244,0.1744967699,-0.1068009287,-0.1851877421,0.2556070685,0.4662512243,0.153420493,0.1642299294,0.2508543432,-0.3839328885,-0.0946437493,0.2254661471,0.0352408737,0.1736327857,-0.2499010712,0.0057578972,-0.0723202229,0.1074136198,-0.150979653,-0.0982134789,-0.1745828539,-0.2246095389,0.7090950608,0.0415948518,0.049666401,-0.0144844456,-0.2729943693,-0.3072059453,-0.2880754471,-0.1513740569,0.0844437703,0.0616011992,0.2224786431,-0.1881359816,0.0445704013,-0.4247748256,0.2465502173,0.1028693095,-0.0479445979,-0.0912304297,0.2359804064,-0.3780378699,0.1200850382,0.2901973426,0.1547181606,0.0404204689,0.1829795539,-0.0811453313,-0.4246262908,0.5179020762,-0.0178216472,-0.2735965848,-0.0381783098,0.1045881063,0.0730181932,-0.2387007177,-0.2471564859,0.3185578287,0.258102566,0.1204383299,-0.1316837966,0.1669016331,-0.2302278876,-0.0435867347,0.1229614243,0.5427109003,-0.0923760012,-0.1214353889,0.3414204419,0.0096555706]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2691","title":"xtreme \/ pan-x cannot be downloaded","comments":"Yes @severo, weird... I could access the file when I answered to you, but now I cannot longer access it either... Maybe it was from the cache as you point out.\r\n\r\nAnyway, I have opened an issue in the GitHub repository responsible for the original dataset: https:\/\/github.com\/afshinrahimi\/mmner\/issues\/4\r\nI have also contacted the maintainer by email.\r\n\r\nI'll keep you informed with their answer.","body":"## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":62,"text":"xtreme \/ pan-x cannot be downloaded\n## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nYes @severo, weird... I could access the file when I answered to you, but now I cannot longer access it either... Maybe it was from the cache as you point out.\r\n\r\nAnyway, I have opened an issue in the GitHub repository responsible for the original dataset: https:\/\/github.com\/afshinrahimi\/mmner\/issues\/4\r\nI have also contacted the maintainer by email.\r\n\r\nI'll keep you informed with their answer.","embeddings":[-0.3182061911,-0.3272892833,-0.1103755236,0.2835794687,0.2328954488,0.0658764243,-0.0919090956,0.1860670298,0.0463901907,0.2400702983,-0.2260717303,0.2097327858,-0.0343383439,-0.148910597,0.1411536783,-0.233079657,-0.0239754748,0.0649312139,0.0869294629,-0.0030195133,-0.0435504913,0.0999886692,-0.1346054673,0.1627422869,-0.2816883326,0.1641907394,0.1339256316,0.1528057605,-0.361030668,-0.5697411299,0.5499168038,-0.0718998834,0.2162335813,0.6330673099,-0.0001045263,0.0274301097,0.2513790429,-0.0701240972,0.0219657589,-0.3303210735,-0.1703127176,-0.3326670825,-0.0559186675,-0.2326419502,-0.0492845885,-0.1610853821,-0.0430011265,-0.2736641169,0.1877941489,0.3142366409,0.3028452396,0.3407095969,0.197421059,-0.1840539575,0.4177657962,-0.1173002496,-0.2915065885,0.0986323431,0.2888224125,0.2544907928,0.3358385861,0.3467045724,-0.0624667481,0.136175558,0.1152224466,0.1114344373,0.0359211527,-0.4036871791,0.1034546196,0.3053645194,0.3668734133,-0.2257569581,-0.3703781068,-0.128019318,-0.0011336268,-0.413595587,0.2864317298,0.1885604411,-0.0614449792,0.0487341769,0.0590127409,-0.0640952811,-0.0094630001,0.0822293609,-0.236755088,0.3045526147,-0.2398663014,-0.0810085982,0.0611778311,-0.0976329967,0.4266878366,0.1055221632,-0.049861975,0.1894536316,-0.1910706609,0.0861553177,0.0916260183,-0.031315688,0.187951237,-0.0939677805,-0.0870378762,0.0212328956,-0.1213041991,0.0533422381,-0.0236649718,0.2236055732,0.0136994431,0.0123606063,0.3364660442,0.1675424278,-0.0252039582,-0.1039092392,-0.0610516407,0.0210201573,0.3471226394,0.0151840132,0.2679439485,-0.4040792882,-0.1462789476,0.108835727,0.0811567977,0.0040619867,0.0753683075,0.2696229219,-0.1310441941,0.5004304647,-0.143953681,0.277445972,-0.1786134094,-0.0545943379,-0.2305138558,0.287528187,-0.3185658455,-0.1640536487,0.268139869,-0.0885947198,0.342446655,0.0193034634,0.4485731423,-0.2276663333,0.1455594152,-0.039132528,-0.1567942351,0.3950023055,0.2186512351,0.2417788953,-0.0570860915,0.1070095226,-0.1398312002,0.054744076,-0.5332157612,-0.1445197165,0.0484364741,0.2813412249,-0.1109327674,0.0443463624,-0.4615261555,-0.2585163414,0.0184416547,0.0368481353,0.0654701218,-0.2370756119,0.1036681756,-0.2727521062,0.1194799766,0.1861193627,-0.1496269554,0.0811404139,-0.1508214325,-0.1840720326,-0.1036881953,0.0949762687,-0.2324253768,-0.0280532092,-0.343136102,0.1788948774,0.4966636598,-0.5603690147,-0.6766209602,-0.1945018023,-0.216724813,0.1057187915,0.0522407591,0.1284447461,-0.1195854843,-0.075569354,0.116512984,0.2314845324,0.0689336732,-0.0431960672,-0.2029492408,-0.1580140293,-0.3941799104,0.2986288667,0.3304900229,-0.0866440758,0.1213661656,-0.0463840812,0.0500180684,-0.1777494848,0.0538886599,0.4001027644,0.0876315162,0.0499624014,-0.1079476029,-0.3960916698,-0.26929003,0.3049806654,-0.231490761,-0.0901017115,0.042286627,-0.1657970548,-0.3381282389,0.1252348125,-0.2549578846,-0.1380519867,0.19801943,0.217994526,0.210733667,0.0432911031,-0.1856295615,0.2093699574,0.1347272545,0.0454653166,-0.393473357,0.0988777801,-0.1647421569,-0.0848025233,-0.0186090562,0.0318331234,0.2233188301,-0.1131895483,-0.1313253194,0.2966018319,-0.1410743296,0.1947293133,0.1766807586,0.1290343851,0.0698762462,-0.5461265445,0.3711177111,0.0836980268,0.0137755852,0.1292071939,-0.1238272563,0.3975926638,0.2258637846,-0.1046828404,-0.0073532304,0.0574757345,0.400721401,-0.1839652807,0.137681365,-0.2009602338,0.3223162889,-0.0778860152,0.0890906081,-0.1408875883,-0.3618681431,0.2239440829,0.2993720174,-0.0730524808,0.27456972,0.0513551347,-0.2327725291,0.1006511152,-0.1656292975,0.1345457137,0.4002487957,0.1655014306,0.1585409492,0.1127974987,0.0303282104,-0.1219317317,0.0640179291,0.0798129663,-0.3082915545,0.138644591,0.2200647295,0.0478178822,-0.2584237158,-0.0435232669,-0.1759923846,0.0869173855,-0.1290674359,-0.054927662,-0.2891584337,-0.0187092144,-0.0212734994,-0.1296854913,-0.326795578,-0.2894042134,0.0833147317,0.301923573,0.3182380199,0.0738872439,-0.4791664779,-0.1091631353,-0.0258458965,-0.1513308287,-0.2301011682,0.1849558353,-0.0872135237,0.1294902563,-0.0128820883,-0.0319500677,0.2024096251,-0.3810040355,0.2993235588,-0.3908240497,-0.2395477742,0.1227841303,-0.096419245,0.3663290143,0.1242185086,0.0071552503,0.0495898463,0.2909899652,0.3144429326,-0.5699630976,-0.0599179268,-0.0251489207,-0.0103814062,-0.113144584,-0.002694956,-0.213219583,-0.4026871622,-0.3585482538,-0.1287237555,0.2220827937,0.0845836923,-0.1363319308,-0.0353298225,-0.0234601684,-0.0528173819,0.0132873515,-0.3657125235,-0.7164649367,0.5234928131,-0.3666719198,-0.3437446356,0.3153213263,0.0493955836,0.3250024021,0.0266776681,-0.5940799117,-0.4159550369,-0.192096591,0.2497019023,0.07498952,0.1958060116,0.2518181801,-0.2047294676,-0.1169368848,-0.1770107299,-0.1927513331,-0.3299808204,-0.0104139606,0.274867475,-0.0486648791,0.2104867995,-0.0205360781,0.5043942928,-0.0040310412,0.2921929657,0.5138947964,-0.0157214403,0.387863189,-0.1739789844,-0.2270195484,-0.100549981,0.0359631404,0.1180128828,-0.0814062655,0.3046616912,-0.0297744572,-0.3115225136,0.0018651439,0.0753884241,-0.1000457555,-0.020178698,0.2238378674,0.1390812993,0.1191479415,0.217791304,-0.3337518573,-0.1647444367,0.0646175444,0.2860420942,0.0719969347,0.2217719555,-0.6987003684,0.294508487,-0.35494712,0.2589458823,0.1331325769,0.2972766161,0.0891900063,-0.0356935598,0.2496795207,-0.0314283557,0.2928499579,-0.2753566206,0.1924650073,0.2488081455,-0.1890153438,-0.2577939928,-0.1476651281,-0.0607869364,0.0950525105,0.074111864,0.2206853777,-0.2511303723,-0.2098118812,0.1976277232,0.1668662578,-0.1227129996,-0.0575692691,-0.2543337941,-0.3855942488,-0.2199773788,-0.1009286568,0.1631483436,0.0240852218,0.1836432666,0.1339164972,-0.2818678916,0.0226787385,0.2182982415,0.2294731289,0.4041041136,-0.1779534966,0.0862167254,0.1354383677,-0.2517170012,0.2359964252,0.6561262012,-0.0692821369,-0.197524637,0.2336289436,-0.0230962485,-0.1156215519,0.1739465892,-0.1396587938,-0.0992723107,-0.231638357,0.1636505723,0.0345823802,0.2379450947,0.2754204571,0.0214605536,-0.4427703321,-0.2652609944,0.4133498073,0.2995727062,0.0854316279,0.1106899828,-0.1749097556,-0.0096168919,-0.2360402048,0.0575964451,1.0173091888,0.0572347865,0.2175719887,0.0290108882,0.0634133294,0.0565520823,-0.2485342175,0.2046530694,-0.1223321036,-0.3878816664,-0.1455356777,-0.1915266514,0.0145887081,-0.1603752971,0.071068719,0.3486486971,-0.0586816818,0.219733417,0.1215181798,0.0530670732,-0.2179558426,-0.1773559153,-0.4743056297,0.2352401316,0.0248241406,0.3444925845,-0.1160426512,-0.0381646641,-0.1843739003,-0.0278281551,0.0443977006,0.296023339,-0.0469839014,0.1254895478,0.0148102706,-0.5127177835,-0.1716361046,-0.0126479268,0.085618712,0.0979857296,-0.1677841544,0.3572310209,0.0622941963,-0.1430040896,0.1259832531,-0.1241479516,-0.0142250303,-0.1708570272,-0.0881156251,0.336365819,-0.0510585122,-0.2698979676,-0.337926805,0.108613342,-0.0641647354,-0.0367735177,-0.0996810943,-0.18450284,0.2647095919,-0.1955665052,0.195883289,0.2199278325,-0.0514431819,0.0763026103,-0.0234754607,-0.271589011,-0.1240641698,0.4434834719,0.1363618672,-0.1830368191,0.1523014009,0.1323874295,-0.0747344643,-0.1610767543,-0.07924629,0.1676333249,-0.5585591197,0.0591032393,-0.2399000227,-0.0812933967,0.0603834838,0.1577664763,0.2505898178,-0.2583154142,0.0715049505,-0.5780913234,-0.2697542906,0.0295937639,-0.0586963631,0.1819916815,0.0045941146,0.0138159888,0.0133006377,-0.1490949839,-0.3993140161,0.2320013791,-0.1090359613,-0.1081174612,0.0873617828,-0.0078234142,0.2728002071,0.0636747479,0.1723724008,0.1119483188,-0.4826051593,-0.2258473635,-0.0068131313,0.091032289,-0.0355651565,-0.1436185092,-0.0072475392,-0.0998624414,-0.1102187186,-0.0084548509,0.1170772985,0.2097507864,0.0412967876,-0.2090098411,-0.1159467995,0.0112967454,-0.1504196078,0.3645819724,0.1342124045,0.0499597192,-0.0004250167,0.1925401092,-0.1359707713,-0.0129971942,-0.1121398509,-0.08918152,0.0612369329,0.2666777968,0.3983950615,-0.214110598,0.2144770324,0.5816690922,0.2518544197,0.2296134084,-0.2959100306,0.1607377231,0.1413059086,0.3287953436,-0.3148750067,-0.2781592309,0.419544816,0.2699165046,0.0607740358,0.3457850516,0.3644168973,0.1267200112,0.4909858406,0.2410901934,0.707395196,-0.0904365107,0.3090626001,0.3332034051,-0.031114962,0.088522926,0.5610823035,0.2646366656,0.1068781614,0.5835371614,-0.0724654794,0.4266375899,0.0914976373,0.0557041615,-0.1497879773,-0.1735351682,-0.2629809678,0.0968464464,0.1215830967,0.1593314707,-0.1135704964,-0.102587387,-0.2767374218,-0.074907966,-0.4128287435,0.2028881758,-0.1576404572,0.0098309787,0.0702229962,-0.2928049862,-0.0622260794,0.0822264031,0.0259678699,0.0887751877,0.1094325557,0.0419868678,-0.1684209406,-0.2736288607,-0.1572637856,0.2412972748,0.081709452,-0.1552295536,0.3693655431,0.1513490528,-0.1085919514,-0.0205730014,0.3077450693,0.3604052365,0.115573965,0.0611661784,-0.0524355359,0.0979602933,-0.0790835321,-0.1598639935,0.1252261847,0.0504928343,0.3037779927,0.1888393164,0.2769693434,-0.297316581,0.2524169981,-0.2654285431,0.1609306037,-0.2756276131,-0.1300503165,-0.143550083,-0.083306931,-0.2492133081,-0.1099411845,-0.2449305058,-0.0160888229,0.2885898352,-0.1056004912,0.0203260221,-0.4095277786,0.1032079086,-0.0276090708,0.2106093913,0.4602036178,-0.0942182615,-0.115659155,-0.1780621707,-0.9365273714,0.3066981733,-0.2836107314,-0.2083441764,0.1622201353,-0.1118565127,-0.0682985187,0.2574859858,0.4689860344,0.1569052339,0.2140833288,0.2086200714,-0.3969550431,-0.0938299969,0.3109645844,0.0110946316,0.166164577,-0.280945152,-0.0325659588,-0.1328875721,0.1413195431,-0.1511570811,-0.0938482657,-0.1022334322,-0.0718074962,0.723259151,0.0636356547,0.0494971126,0.006365465,-0.3130992651,-0.3735244572,-0.2301961035,-0.1844259948,-0.0065053417,0.0904605985,0.3327545822,-0.2201779187,-0.028315641,-0.4172604084,0.1894256771,0.1188802049,-0.0296963938,-0.1313203573,0.1838798374,-0.3572983146,0.1247301847,0.3231872916,0.2014229298,0.0441381484,0.2464498132,-0.1630646735,-0.3918913007,0.456047833,0.0383987278,-0.2726228833,-0.1572235227,0.0500914864,0.0204189662,-0.1961495578,-0.2693953514,0.2729896903,0.2459129691,0.1181409806,-0.1239314526,0.1538918018,-0.2061372846,-0.0190597102,0.0374165475,0.4905184507,0.0262595713,-0.1366614848,0.2494590133,0.0027828258]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2691","title":"xtreme \/ pan-x cannot be downloaded","comments":"Reply from the author\/maintainer: \r\n> Will fix the issue and let you know during the weekend.","body":"## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":16,"text":"xtreme \/ pan-x cannot be downloaded\n## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nReply from the author\/maintainer: \r\n> Will fix the issue and let you know during the weekend.","embeddings":[-0.3508202434,-0.4881745875,-0.1012572348,0.2679373324,0.242426917,0.0743881837,-0.1332196444,0.2452753931,0.1442698389,0.1187140942,-0.2436651886,0.2646325231,0.0461617969,0.0465012193,0.1917267889,-0.326759398,0.0447652899,0.1279419363,0.0111566251,-0.0376448408,-0.1420017928,0.1677828431,-0.2066324651,0.1202998906,-0.2243205905,0.1020483896,0.0773360208,0.1569079757,-0.310025692,-0.4581514299,0.6455733776,-0.0633814484,0.2681769133,0.604803741,-0.0001053799,0.0210377499,0.2643092871,-0.0464517288,0.164749071,-0.4361985624,-0.285949707,-0.3566252887,-0.0301674232,-0.2249145359,-0.0407119244,-0.2227444649,-0.1089594513,-0.114901796,0.2539050877,0.224211067,0.294454962,0.3886256516,0.1530675143,-0.2150014043,0.4348991215,-0.0721100867,-0.3141875565,0.1084877253,0.3806249797,0.1838413328,0.464098841,0.2748147547,-0.0413430333,0.0587843582,0.158915177,0.0977327004,-0.0495259948,-0.4637801945,0.038880758,0.3281199038,0.387986362,-0.2812048197,-0.3449712694,-0.0700660646,0.0634025484,-0.3715293407,0.2863736153,0.1802633405,-0.0139233274,0.0406311527,0.0181270111,-0.048152443,-0.0553825833,0.110701099,-0.0341184102,0.2764595449,-0.2252991349,-0.0331717096,0.1091583148,-0.0374769866,0.2385877073,0.086752139,-0.0555851869,0.1689654291,-0.2702018917,0.0684053004,0.007478517,-0.0838781968,0.3481557667,-0.077706039,-0.1982533634,0.0007951256,-0.1292144805,0.0622540116,0.0946067944,0.1408195496,0.0807005838,0.0324028134,0.2074715644,0.1673329324,0.084922336,-0.0790967569,-0.1301611364,-0.0352247357,0.1570245177,-0.041583214,0.3730689585,-0.393217653,-0.2654044032,0.0548107438,0.0834704265,0.0691127703,0.0720816329,0.2562322915,-0.1046686918,0.337223798,-0.1489224434,0.4322533607,-0.220555678,0.0384466723,-0.2084830403,0.3019188046,-0.2957861722,-0.2560020089,0.2096585184,-0.1504617035,0.3913116455,0.0729194656,0.3850796223,-0.199715212,0.0614920445,0.0084678633,-0.309230268,0.3815722466,0.191742301,0.1878200024,-0.1222999021,0.1563587487,-0.1546243429,0.0993809178,-0.5259683728,-0.0668005794,-0.0467678905,0.260764122,-0.1234370545,-0.0680576935,-0.5894210339,-0.1413500905,-0.0553176366,0.1839159578,0.1050052494,-0.2398908734,0.1146361604,-0.2036355883,0.2371749133,0.187357828,-0.2137732953,-0.0268489625,-0.0182714853,-0.3518374264,-0.0915610716,0.1520219445,-0.192415148,-0.019013904,-0.2748942077,0.3359982967,0.475649029,-0.5591967106,-0.5651315451,-0.0963319466,-0.2180439532,0.0007037257,-0.0200067069,0.0962054431,-0.1117716432,-0.0004895465,0.0994872823,0.2488730401,0.0631219074,-0.0226262789,-0.1768683344,-0.2266003489,-0.3862449527,0.3425415158,0.2518791258,-0.0944090635,0.0663152635,-0.0002348693,0.1595528126,-0.083132036,-0.0486579649,0.3730819225,0.2286723554,0.0447254702,-0.1036694273,-0.3814466298,-0.3556773961,0.2895233035,-0.092832014,-0.1557418108,0.0468271151,-0.154229328,-0.4698528647,0.2062272131,-0.2509517968,-0.0803833529,0.132697165,0.0898680985,0.2055639625,0.0326894782,-0.2376073897,0.07682693,0.0526436307,0.1150268391,-0.3940125406,0.2354759574,-0.0872816145,-0.1635206789,0.0684082732,0.0649870113,0.212233156,-0.2378917485,-0.1321241409,0.3139254153,-0.0963537097,0.1405933201,0.241208896,0.0367050841,0.06820032,-0.6196578145,0.2208238542,0.0719355792,-0.0106893228,0.1309360862,-0.0878985301,0.3659490645,0.0959123895,-0.112328954,-0.0786752477,0.0964418203,0.3741547763,-0.1869223863,0.1258788109,-0.0770237446,0.352786392,-0.087723054,0.130599454,-0.2188171744,-0.3352217674,0.0951015875,0.2439258397,-0.0809091628,0.2046398371,0.0871258527,-0.1089674085,0.1696591526,-0.0836644247,0.0586608984,0.4200647771,0.1910061389,0.0222819075,0.109670192,-0.0126072252,-0.1149557158,0.0522856936,0.0696804002,-0.2818588614,0.0545434989,0.110459581,0.0407713465,-0.3076771498,-0.1345253289,-0.1521248221,0.1130894795,-0.1194994673,-0.1115960851,-0.1741676033,-0.0930008516,-0.0067214267,-0.0441843234,-0.2230140716,-0.1807046831,-0.0032339063,0.2508816123,0.3006366789,0.1857052892,-0.4258006215,-0.1767810285,0.0492325723,-0.3311206698,-0.241443634,0.2739618719,-0.0648465529,0.1281394213,0.0012627601,0.1026067287,0.2570173144,-0.3908737004,0.2834922373,-0.3780142367,-0.2129531354,0.0874035731,-0.2608731687,0.3951687813,0.1670734584,0.0618837886,-0.1002265811,0.2385386676,0.3719349504,-0.4895252287,-0.0249968544,0.0311351214,0.0116852922,-0.2028901577,0.0012992581,-0.2291317284,-0.3406861126,-0.3332809806,0.0251754094,0.1601017863,0.0542623773,-0.0320996121,0.0846425891,0.0127804996,-0.1503263116,-0.1215310022,-0.24921,-0.6595410109,0.4708841145,-0.2177618295,-0.4426346123,0.285890162,0.0693380013,0.274273634,0.0797217265,-0.5939115286,-0.3438007236,-0.2070501447,0.4223075509,0.0485908538,0.2106080204,0.1916267872,-0.2807761729,-0.0558228046,-0.1982801408,-0.2254240066,-0.3088796735,0.0396661535,0.2459839135,-0.0177332219,0.2006799877,-0.1411966532,0.4559248388,0.1566533744,0.2250502259,0.5607686043,-0.0784542561,0.2969378531,-0.1781841666,-0.3331464827,-0.0779972598,0.1333859414,0.1393454522,0.004903838,0.2621773183,0.0415848009,-0.2719686031,-0.0684316307,-0.0087972749,-0.0761468783,-0.0551777743,0.1933762431,0.0745694488,0.0178926103,0.2722102404,-0.1637096107,-0.1097764522,0.0779529363,0.3760089576,0.0609090365,0.2542284727,-0.6363356113,0.1994326264,-0.4225559533,0.2633749247,0.0229356289,0.3032845855,0.0271450002,0.1275614649,0.2268926501,0.0173959117,0.3738657236,-0.2115794569,0.2356037199,0.1690299213,-0.1563702524,-0.3037767112,-0.1220082492,-0.1556054652,0.0848981142,0.184973374,0.4261040688,-0.292320013,-0.2029683888,0.2418842614,0.1523060501,0.0301390123,-0.1369530559,-0.3014231026,-0.5566600561,-0.197675854,-0.0641257688,0.2056791782,0.0641507506,0.1885637194,-0.0287858918,-0.2342075109,-0.0634588748,0.1824845225,0.2650254667,0.3496652544,-0.1581734866,0.2069968879,0.088727057,-0.3461327851,0.2118749619,0.67125386,-0.0343951881,-0.4521394372,0.0967701823,-0.0372718796,-0.0049409196,0.0629565641,-0.1753768623,-0.0580417141,-0.2222938389,0.1522696316,0.0563974194,0.2417753637,0.3856702149,0.0768091232,-0.4572806358,-0.2930099964,0.5171843171,0.2759591639,0.1106105596,0.2264081389,-0.0755702481,0.005940577,-0.1686211824,-0.0367956534,0.8605223894,-0.0200942196,-0.0090475492,0.204969883,0.0234196726,0.1926706284,-0.1032149345,0.0858312398,-0.1676981449,-0.4318582118,-0.1242151335,-0.1855196655,0.144000113,-0.0584128164,-0.0112075945,0.2490838319,-0.017395284,0.1041807979,0.0706710443,0.072929129,-0.1568382829,-0.1870139241,-0.592032969,0.2291752994,-0.018918613,0.3568812907,-0.1645397693,-0.0627130717,-0.1835575104,-0.1184018999,-0.0058981748,0.3321668804,-0.0169966053,0.0666977316,0.0330717042,-0.5403254032,-0.0495046414,0.0661531836,0.0649540275,0.1241135895,-0.2367528975,0.3196366429,0.0056379749,-0.2470935732,0.1195449159,-0.1384114623,-0.186273396,-0.1169377342,-0.1802086979,0.1975175589,-0.0473877788,-0.2174139172,-0.2288230062,0.0365486294,-0.1086302623,-0.1126614287,-0.1459478736,-0.2184648961,0.2761242092,-0.2277489156,0.1800664961,0.2461976111,-0.0148837138,0.1121527702,0.0897893757,-0.3175094426,-0.180065766,0.348076582,-0.0394449271,-0.1205917671,0.121854119,0.1929723769,-0.0352570117,-0.1940976083,0.0460470654,0.2011941522,-0.6195076704,0.0208561327,-0.1373987943,0.0142478952,-0.0530654415,0.2701609135,0.2481817901,-0.3544403017,0.2466700971,-0.5972566605,-0.2012000382,0.143312946,-0.0129746636,0.2136299014,0.1011096016,0.0865785256,0.1163206771,-0.2084697187,-0.3718106151,0.1484270394,-0.0851334035,-0.0941162184,0.1408212185,0.0926844701,0.2820590436,0.0364788659,0.1912924945,0.090355143,-0.3732386827,-0.2340145111,-0.0399684124,0.070935227,-0.0267689843,-0.1522597671,0.0009426704,-0.0621790662,-0.2029338777,-0.0487754531,0.1483901441,0.1990872025,0.0471791476,-0.1144540682,0.0083835963,0.03039838,-0.2331443578,0.2854089737,0.1517047137,0.0903529227,-0.1003661975,0.1442606747,-0.0590410568,-0.0432401113,-0.2194743454,-0.102657482,0.1456827372,0.2854143083,0.2501860559,-0.1964508593,0.3128254712,0.513657093,0.2690925598,0.3190071881,-0.2770457864,0.2185478657,0.1570942551,0.3223168552,-0.3402834535,-0.1649016291,0.374347955,0.3258282244,0.0282959323,0.3628846407,0.245329991,0.1605166197,0.3300941586,0.1862071157,0.6517935991,0.0480827279,0.3317585886,0.4907902777,-0.009580316,0.0681622922,0.46288234,0.2584774792,0.0624931715,0.5506413579,-0.1220129505,0.3823799491,0.0109417439,0.122721374,-0.1426023692,-0.2383998781,-0.2924526334,-0.0140767107,0.0763157755,0.046427533,-0.0517435409,0.1019366533,-0.3627051115,-0.0767847449,-0.408557117,0.1955700368,-0.1277136803,0.06112241,0.0815946981,-0.1939544678,-0.024851175,0.000370468,-0.0396142304,-0.0183505099,0.2154421061,0.0037962089,-0.1877613813,-0.2962321043,0.0072554192,0.3406334817,-0.0216184277,-0.110014461,0.3868371546,0.0700776726,-0.090948686,-0.0202033315,0.384165436,0.3209081888,0.1881233454,0.1001429781,-0.0689264089,0.1008488834,-0.0525463484,-0.062636964,0.1833415776,0.0194405541,0.4059861004,0.2199334949,0.2920586169,-0.260191679,0.1627619267,-0.1733614802,0.127404362,-0.2675946057,0.0019139214,-0.3240398169,-0.116640605,-0.2153551728,-0.1564308405,-0.187963292,0.0399812236,0.2971614897,-0.0081563471,0.0440369211,-0.3504026234,0.0983129442,-0.1131945997,0.1918169409,0.3956769407,0.0221832562,-0.2351934016,-0.1102324203,-0.8148165941,0.2829460502,-0.2405501604,-0.3548552692,0.2154842764,-0.1633161306,-0.1039040834,0.3233462274,0.4029321671,0.1540603489,0.2269130051,0.2429272234,-0.36439991,-0.1051161885,0.1346461773,0.0678052977,0.1877437085,-0.2748457193,0.0216611754,-0.0715666786,0.0703181252,-0.217780441,-0.1180777997,-0.1292556822,-0.2397355586,0.7350841761,0.0209610332,0.1643328518,-0.0689478368,-0.1507926881,-0.3329709768,-0.3756676018,-0.1921012253,0.1744665504,0.0068289819,0.2460232377,-0.1461543739,0.0154092675,-0.3872994184,0.1798046678,0.0797630548,-0.0587992668,-0.1046175957,0.1820996851,-0.3372632265,0.0582620353,0.2378584146,0.1692324728,-0.0135631645,0.2319483608,-0.1146038994,-0.4429401159,0.4428634644,0.0188659038,-0.2732521296,-0.0701511577,0.060748376,0.0788189694,-0.2855609357,-0.3377626836,0.339584589,0.281183511,0.1168807819,-0.0960107893,0.1693703681,-0.1990393102,-0.0334395468,0.1144596338,0.5657114387,-0.0420864448,-0.1423905492,0.4052657187,-0.0098423501]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2691","title":"xtreme \/ pan-x cannot be downloaded","comments":"The author told that apparently Dropbox has changed their policy and no longer allow downloading the file without having signed in first. The author asked Hugging Face to host their dataset.","body":"## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n","comment_length":31,"text":"xtreme \/ pan-x cannot be downloaded\n## Describe the bug\r\n\r\nDataset xtreme \/ pan-x cannot be loaded\r\n\r\nSeems related to https:\/\/github.com\/huggingface\/datasets\/pull\/2326\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\")\r\n```\r\n\r\n## Expected results\r\n\r\nLoad the dataset\r\n\r\n## Actual results\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/www.dropbox.com\/s\/12h3qqog6q4bjve\/panx_dataset.tar?dl=1\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: macOS-11.4-x86_64-i386-64bit\r\n- Python version: 3.8.11\r\n- PyArrow version: 4.0.1\r\n\nThe author told that apparently Dropbox has changed their policy and no longer allow downloading the file without having signed in first. The author asked Hugging Face to host their dataset.","embeddings":[-0.2759372294,-0.3480373025,-0.059366364,0.3470798731,0.1086417437,0.0554847009,0.0214382652,0.2132673413,0.3107877672,0.068113707,-0.2372275144,0.2172878534,-0.0742678568,0.1829396188,0.2306470424,-0.1670987904,0.018207306,0.0383961536,0.1040748507,-0.1059666947,-0.0068049063,0.0688059554,-0.2352527827,0.0341514982,-0.2067681998,-0.0287322812,0.0926772133,0.3318829536,-0.3699327707,-0.375195086,0.6139099002,-0.0842426792,0.1851760298,0.5841585398,-0.0001120175,0.0652964935,0.1374395192,-0.1021743938,-0.077500619,-0.5789530277,-0.3472652435,-0.4007537663,-0.0861783698,-0.1636051387,-0.0511669517,-0.1117286012,-0.0150200641,-0.1729518473,0.5082504153,0.3695927262,0.2240834236,0.5469501019,0.1720808595,-0.1378317922,0.3611240685,0.1562046111,-0.2377878875,0.4367509186,0.3216775954,0.1157636195,0.411288619,0.17754592,-0.0996762365,-0.1689160168,0.1308137774,0.0133560635,-0.2266700417,-0.4470198154,0.0784026906,0.2077765316,0.3536137342,-0.1338850856,-0.4127598405,-0.3614983559,0.1108664125,-0.1548257917,0.2727961242,0.1956167519,-0.0359452404,0.1448778957,0.0599417686,-0.1014324799,-0.0129188905,0.0593793765,0.0173093267,0.3373428583,-0.2577695251,0.0144006712,0.1092236415,-0.0751527324,0.3978737295,0.0532085076,-0.0479291938,0.1232226193,-0.1411840767,0.0053743836,-0.0540445745,0.0137527417,0.2482397109,0.0194021352,-0.0785555169,-0.080885604,-0.2006908357,0.0241392329,0.1848617345,0.304869771,0.1491407752,-0.1033317223,0.1745783836,0.3312750459,0.196003139,-0.0434585921,-0.2963788509,0.0445669107,-0.0722321942,-0.1512659788,0.5243734717,-0.2715309858,-0.3970927298,-0.019526897,0.0200423934,-0.0885450915,0.1340081245,0.2182936966,-0.1068647355,0.2549786568,-0.07606861,0.2554853559,-0.163050577,-0.1515774131,-0.1914504617,0.405141592,-0.2343253791,-0.1603657901,0.1848150939,-0.2435047626,0.4186853468,0.1288653165,0.3946613073,-0.150030002,0.1027104482,0.0922761708,-0.1681749523,0.3451004326,0.1561530232,0.1269650757,-0.0141464854,-0.234855175,-0.1157528386,0.0571507141,-0.5933461785,-0.1403056085,-0.1993746758,0.1679313332,-0.2249724567,-0.130701378,-0.4889059663,-0.1503004879,-0.1152251363,0.0413695909,0.1745775342,-0.093983531,0.2278492898,-0.1294245869,0.1880383492,0.4361646771,-0.2327737212,0.0334096067,-0.1792382747,-0.099394165,-0.0730862916,0.0946990103,-0.2904944718,-0.0520451032,-0.4483394921,0.1997848898,0.2893691659,-0.7500990033,-0.5598580837,-0.089861609,-0.4706421494,0.2330068946,0.0047258893,0.0929392725,-0.0257564187,-0.1167187244,0.1843946427,0.0955513045,0.1699332446,-0.1224311441,-0.2045544237,-0.2444781512,-0.2624104321,0.2842634618,0.0622716285,0.0681209564,0.0910937339,0.2827214301,0.2098217756,-0.0228558481,-0.0933938175,0.3326901197,0.2923719287,0.0233746152,-0.28684026,-0.2477786839,-0.423281312,0.3309375346,0.0280762967,-0.105102919,-0.0007369105,-0.2265768051,-0.5678087473,0.1613327563,-0.188889429,-0.1785442084,0.067773439,0.1239690706,0.1513993591,0.0496586636,-0.2474329919,0.2233787626,0.0188436583,0.1543605179,-0.5641939044,0.2311494946,-0.1155117154,-0.0604094341,0.035268791,0.0466389135,0.2501816452,-0.2004779428,-0.0556469597,0.3803123236,-0.1920132637,0.0159702394,0.2529326081,0.0522421375,0.1939553469,-0.4988432229,0.243588537,0.0546597093,0.0235157963,0.1672837436,-0.2205287069,0.4448288381,0.0305797495,-0.2373160869,-0.1401200145,0.1262514889,0.2230827957,-0.1306058466,-0.0272244941,-0.1473646462,0.2565690875,-0.2958555818,0.4428711832,-0.1918583959,-0.3940776289,0.0924434811,0.2551046014,-0.0954708979,0.1634525359,0.1062978208,0.0065807784,0.0754789263,0.0918328539,0.1748152524,0.4484563172,0.1287842691,0.0455228947,0.1220075339,0.1126532406,-0.1875597686,0.1822135746,-0.0599472448,-0.4153219759,0.0525691845,0.2299790233,0.0714469552,-0.3545280993,-0.2406418622,-0.1877214909,0.0300265606,-0.2494356334,-0.0179142952,-0.1286338568,-0.2612890899,-0.0283743665,-0.0056921877,-0.3994728625,-0.141236186,-0.1487310529,0.2708365619,0.249328509,0.0064039682,-0.3584647775,0.1185194328,-0.0075160903,-0.1367571652,-0.3121984899,0.3624312282,-0.0285946559,0.048477158,0.1105538458,-0.0293610375,0.2358980477,-0.415718168,0.1052324548,-0.3573950231,-0.3269350231,0.0451696552,-0.2162965387,0.428937763,0.1453777999,0.2217922211,-0.0008089605,0.0529852659,0.2060077935,-0.21674788,-0.0299977288,-0.210040167,0.0807595924,-0.1460517198,0.0060590436,-0.026528053,-0.4058999717,-0.3775566816,0.2012751102,0.1482295543,-0.0348030217,0.2005346715,0.0097332513,0.1122852936,-0.1734445691,-0.1491319835,-0.1994180381,-0.8692612052,0.4091035426,-0.2042848319,-0.3979110122,0.2263697088,0.1963291615,0.1721195132,-0.1776609868,-0.5201548934,-0.4196139276,-0.2218331993,0.4353312254,-0.0021964402,0.137933746,0.2515825629,-0.2824039757,0.0077370708,-0.1384647042,-0.2788476348,-0.2825559378,0.226845935,0.1722354889,-0.0015180076,0.0224355981,-0.0498864055,0.3794842064,0.1308884919,0.161136061,0.5610560179,-0.0744288862,0.4397812784,-0.1766300052,-0.4628420472,-0.1886907965,0.0963781402,0.0323035717,-0.0544358455,0.2265837789,0.2790605724,-0.356226474,-0.1995026469,-0.0767232552,-0.0645360351,-0.1105739921,0.0093632508,-0.0044459924,0.0796471015,0.2570388317,-0.1833251864,-0.150972724,-0.0236530807,0.4484081268,0.2238715142,0.2405958772,-0.5096655488,0.274063766,-0.3780685663,0.1887898594,-0.0318723284,0.3348709941,-0.1444580108,0.0845750123,0.2295600176,0.0133824516,0.443572998,-0.0775712654,0.3544702232,0.0710437372,-0.1226007938,-0.2085142583,-0.1054210514,-0.0435269736,0.077096723,0.1953823268,0.5760686994,-0.2933995426,-0.1664152443,0.2404119819,0.1298156977,0.0364878997,0.0377409421,-0.2254050821,-0.6075825691,-0.1573698074,-0.0444924496,0.2579651773,-0.0299646053,0.1044436395,0.1127330363,-0.1593716145,-0.0189802218,0.1871154755,0.1539864838,0.2559454441,-0.0247438252,0.2165156454,0.1908061057,-0.1383116543,0.2649733424,0.8343393803,0.0106110238,-0.3429009318,0.0427500196,0.0639129803,0.0057261954,0.3877792656,-0.121197708,-0.0853439718,0.0048180819,0.1961670667,-0.0233457368,0.2301374972,0.3775058389,-0.0149423135,-0.494969219,-0.3380520344,0.5363482237,0.1302823871,0.0742522404,0.1904317141,0.1329300106,0.0057330453,-0.1068958044,-0.0971721634,1.1719795465,0.0562254675,0.2492093742,0.2455012351,-0.0852787271,0.2562092245,-0.088460207,-0.0167998318,-0.1013689563,-0.4164631069,-0.1429820806,-0.1163861454,0.1411392838,-0.0901830569,0.0010765336,0.1374599636,-0.1716613173,0.1739597023,0.1153218895,0.1687582731,-0.2411976457,-0.0612041242,-0.4733584821,0.1426097602,-0.0069748694,0.3821437061,-0.2010969669,-0.2299177796,-0.2083753347,-0.1166216657,-0.0660127476,0.2327334136,0.108354032,-0.012380858,-0.0692909956,-0.5590741038,0.0888753384,0.0713525862,0.0060642329,0.1797234267,-0.3232238293,0.3109991252,-0.0692843497,-0.2035991102,0.129031226,-0.1362445801,-0.1140312999,-0.0352116935,-0.1258568764,0.092984885,-0.1436610669,-0.1394315362,-0.385687232,0.2510842383,-0.2772895098,-0.2268613428,-0.164497152,-0.1835376173,0.3595433235,-0.253041178,0.1040874869,0.1485456377,0.1078658402,-0.2641800344,0.1531633139,-0.2987476289,-0.1521572173,0.5257741213,-0.1988991797,0.0334549844,0.1715101749,0.0130757401,0.0204172749,-0.1379290223,0.1412822753,0.061684221,-0.6248799562,0.1188299209,-0.2430006713,0.0227164477,-0.2295223325,0.3468093574,0.3823336065,-0.2665142119,0.2866683304,-0.5378331542,-0.0770435855,0.1276140809,0.0879106224,0.0976175889,0.0959595963,0.1219419837,0.1061660871,-0.1598082483,-0.3132171631,0.1700588614,-0.121995233,-0.1898820996,0.3454614878,0.0432373248,0.3752042353,-0.016101392,0.1467815042,0.0533298701,-0.2933898568,-0.1397865266,-0.0504710227,0.1410354674,0.0521025062,-0.2471048832,0.0549416952,-0.1684382409,-0.1605031937,0.0132015152,0.0367017649,0.2360640913,0.1174361184,-0.1449898928,-0.1602437347,0.1950679719,-0.245007202,0.3418947458,0.2247636467,0.0340525918,-0.1005309224,0.1787566841,-0.1220848337,0.0183274969,-0.2611814439,-0.085661985,0.0776030645,0.3909317255,0.337634325,-0.134683609,0.2393316478,0.3750910461,0.266994983,0.3125608265,-0.2661042213,0.1791471243,0.1509798318,0.2022509724,-0.3251589835,-0.1843992621,0.2895826697,0.2845374942,0.0079278788,0.2907944024,0.2425782681,0.1455316246,0.0103199352,0.1336497813,0.8089545369,0.2793609798,0.367298305,0.5459699631,-0.1114080101,-0.0239295159,0.4142659307,0.2047233433,0.1880605966,0.4659932554,-0.1904216707,0.3810591996,0.0038549975,0.1714942753,-0.1326284856,-0.3277802765,0.0001665298,0.0214361362,0.1399538368,0.0384782292,-0.0150220022,0.1334608048,-0.4537134767,-0.0577390455,-0.2650279701,0.2728067338,-0.1544536054,0.1055009514,0.0780477971,-0.3082312942,-0.0416714326,0.0296847727,-0.0832872242,-0.1457926184,0.2115954012,0.0386279188,-0.134189114,-0.2017861754,-0.1125044003,0.3462385833,0.0870194584,-0.1962681264,0.4236857891,0.2718048096,-0.1568729728,0.049809549,0.3523477912,0.4650503695,0.0559386909,0.2656495273,0.0647218302,0.0683039874,-0.1031052992,-0.0804571956,0.2312495559,0.0351841599,0.4370720088,0.2358403653,0.238889128,-0.2383612543,0.2402344644,-0.2012109756,0.0160367377,-0.4459690154,-0.0041087847,-0.3592950404,-0.1920768172,-0.2463410348,-0.1867140979,-0.1615373194,0.105026342,0.193741262,0.052289512,0.0789947361,-0.3036747873,0.0630608723,-0.0058307,0.2179626375,0.4620446861,-0.0886168703,-0.0866258517,-0.1759013385,-0.8899638057,0.2022525072,-0.1749405116,-0.3776009977,0.1036339626,-0.2651176155,-0.0521906763,0.1655713767,0.2213157564,0.2603115439,0.147296086,0.1278582811,-0.2324408889,-0.0836085752,0.1863343716,0.1379171759,0.1272398829,-0.1580203921,0.1513994336,0.0785692781,0.019998081,-0.058415886,-0.0605281554,-0.1788770258,-0.3329239488,0.6669124365,0.034896329,0.1883095205,-0.1101698056,-0.2008660436,-0.329852283,-0.1550640464,-0.1784971207,0.2253369689,0.1351694018,0.3386968076,-0.0827751309,-0.0310577396,-0.3715772927,0.3031169176,0.0216147956,-0.0353381447,-0.2809809446,0.0684916452,-0.3736380935,0.0630950779,0.248787418,0.2379322052,-0.0699619502,0.0589816384,-0.1878470629,-0.4683721364,0.5749180913,-0.1280221939,-0.2461773902,0.0778383389,0.1704560667,-0.0042454014,-0.2991142273,-0.0887737572,0.2385407239,0.3178573251,-0.0317883715,-0.0475270487,0.1778371036,-0.2720625103,-0.0316481777,0.1030775681,0.7202537656,0.0673562959,-0.2254704386,0.569835484,0.0235495083]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2689","title":"cannot save the dataset to disk after rename_column","comments":"Hi ! That's because you are trying to overwrite a file that is already open and being used.\r\nIndeed `foo\/dataset.arrow` is open and used by your `dataset` object.\r\n\r\nWhen you do `rename_column`, the resulting dataset reads the data from the same arrow file.\r\nIn other cases like when using `map` on the other hand, the resulting dataset reads the data from another arrow file that is the result of the map transform.\r\n\r\nTherefore overwriting a dataset after `rename_column` is not possible, but it is possible after `map`, since `rename_column` doesn't switch to using another arrow file (the actual data stay the same).","body":"## Describe the bug\r\nIf you use `rename_column` and do no other modification, you will be unable to save the dataset using `save_to_disk`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nIn [1]: from datasets import Dataset, load_from_disk\r\nIn [5]: dataset=Dataset.from_dict({'foo': [0]})\r\nIn [7]: dataset.save_to_disk('foo')\r\nIn [8]: dataset=load_from_disk('foo')\r\nIn [10]: dataset=dataset.rename_column('foo', 'bar')\r\nIn [11]: dataset.save_to_disk('foo')\r\n---------------------------------------------------------------------------\r\nPermissionError Traceback (most recent call last)\r\n in \r\n----> 1 dataset.save_to_disk('foo')\r\n\r\n\/mnt\/beegfs\/projects\/meerqat\/anaconda3\/envs\/meerqat\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in save_to_disk(self, dataset_path\r\n, fs)\r\n 597 if Path(dataset_path, config.DATASET_ARROW_FILENAME) in cache_files_paths:\r\n 598 raise PermissionError(\r\n--> 599 f\"Tried to overwrite {Path(dataset_path, config.DATASET_ARROW_FILENAME)} but a dataset can't overwrite itself.\"\r\n 600 )\r\n 601 if Path(dataset_path, config.DATASET_INDICES_FILENAME) in cache_files_paths:\r\n\r\nPermissionError: Tried to overwrite foo\/dataset.arrow but a dataset can't overwrite itself.\r\n```\r\n\r\nN. B. I created the dataset from dict to enable easy reproduction but the same happens if you load an existing dataset (e.g. starting from `In [8]`)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.10.0-1160.11.1.el7.x86_64-x86_64-with-centos-7.9.2009-Core\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\r\n","comment_length":102,"text":"cannot save the dataset to disk after rename_column\n## Describe the bug\r\nIf you use `rename_column` and do no other modification, you will be unable to save the dataset using `save_to_disk`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nIn [1]: from datasets import Dataset, load_from_disk\r\nIn [5]: dataset=Dataset.from_dict({'foo': [0]})\r\nIn [7]: dataset.save_to_disk('foo')\r\nIn [8]: dataset=load_from_disk('foo')\r\nIn [10]: dataset=dataset.rename_column('foo', 'bar')\r\nIn [11]: dataset.save_to_disk('foo')\r\n---------------------------------------------------------------------------\r\nPermissionError Traceback (most recent call last)\r\n in \r\n----> 1 dataset.save_to_disk('foo')\r\n\r\n\/mnt\/beegfs\/projects\/meerqat\/anaconda3\/envs\/meerqat\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in save_to_disk(self, dataset_path\r\n, fs)\r\n 597 if Path(dataset_path, config.DATASET_ARROW_FILENAME) in cache_files_paths:\r\n 598 raise PermissionError(\r\n--> 599 f\"Tried to overwrite {Path(dataset_path, config.DATASET_ARROW_FILENAME)} but a dataset can't overwrite itself.\"\r\n 600 )\r\n 601 if Path(dataset_path, config.DATASET_INDICES_FILENAME) in cache_files_paths:\r\n\r\nPermissionError: Tried to overwrite foo\/dataset.arrow but a dataset can't overwrite itself.\r\n```\r\n\r\nN. B. I created the dataset from dict to enable easy reproduction but the same happens if you load an existing dataset (e.g. starting from `In [8]`)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.10.0-1160.11.1.el7.x86_64-x86_64-with-centos-7.9.2009-Core\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\r\n\nHi ! That's because you are trying to overwrite a file that is already open and being used.\r\nIndeed `foo\/dataset.arrow` is open and used by your `dataset` object.\r\n\r\nWhen you do `rename_column`, the resulting dataset reads the data from the same arrow file.\r\nIn other cases like when using `map` on the other hand, the resulting dataset reads the data from another arrow file that is the result of the map transform.\r\n\r\nTherefore overwriting a dataset after `rename_column` is not possible, but it is possible after `map`, since `rename_column` doesn't switch to using another arrow file (the actual data stay the same).","embeddings":[-0.0523223989,0.2342098802,-0.0231681652,-0.0133467484,0.3893551528,0.2707957029,0.4929154813,0.246301353,0.0219596829,0.1201097146,-0.0814653933,0.4607105553,-0.1719950885,-0.1512238532,-0.0136408731,-0.0714220777,0.2988217771,-0.1098150685,-0.0071087461,0.1580941379,-0.4297677875,-0.0167975575,-0.0632793829,0.0525315329,-0.1448873281,-0.1060153246,-0.0191482995,0.0699789524,-0.0107616372,-0.2462689877,0.0901805609,-0.0555133633,0.1750023365,0.3935277462,-0.0001104696,-0.1174478754,0.0169522669,-0.0254376344,-0.3840672076,-0.1677256376,-0.2794014215,-0.22320804,-0.1534653902,-0.5357638001,0.2765216529,-0.0022124713,-0.1228611171,-0.3125068247,0.1796944588,0.2632053494,0.294092983,0.1083573475,0.273332864,-0.1635734141,0.0736200586,0.0986325443,-0.3788083792,0.2685394287,-0.2693726122,-0.0911083221,0.1226996332,0.2915157378,-0.1082791165,-0.0690102279,0.1018080711,0.0468242764,0.1152402088,-0.2415997535,0.2724038661,-0.1014997885,0.4159988761,-0.545006156,-0.4493023753,-0.0069589685,0.2386427075,-0.3515321016,0.2436310351,0.1155513525,0.0246032923,0.1335383654,0.0393228084,-0.2681631148,-0.1927657425,0.0060169129,0.0300406311,-0.2012541443,-0.1730899066,0.1861053854,0.0252395608,-0.1277827471,0.3110419214,-0.0618427843,-0.0558179356,0.0676211342,-0.4282257557,-0.0441422723,-0.0765138939,-0.1082729995,-0.1432772726,0.234134987,-0.0629033148,-0.1409233361,0.2153322846,0.2729771733,-0.0386802107,0.2238942534,-0.0026314,0.4684874117,0.3518581092,0.119733721,0.0032541524,-0.1048212945,0.078337796,0.0482479073,0.6009219289,-0.0478974693,0.4625107348,-0.0603512265,-0.3091149926,0.3789649606,0.1044593304,-0.0342989787,0.0012844725,0.2447030544,0.1551560313,0.0356650017,0.0988404751,0.3300382793,0.1512357295,0.167356506,-0.1845288128,-0.0235995818,-0.0770769566,0.1205223575,0.1922730654,0.2134322971,0.0603944696,0.1418900788,-0.1296938509,-0.0818710178,0.063822411,-0.1976366341,0.0316390581,0.3230943084,0.0812872648,0.2417824864,0.1332234889,-0.2870061994,0.0473018102,0.219573319,-0.125742957,-0.1329015791,-0.4467063248,0.2365868688,0.1273117065,0.1414909065,-0.4421401322,-0.2893771231,0.434397608,-0.0966240466,0.2320122719,-0.0182891302,0.0811488405,-0.409035027,0.0521090887,0.221598804,-0.1067450941,0.2233543247,-0.3136855066,0.1113037318,0.1221690103,0.158576414,0.0759357885,-0.0761187598,-0.1343706846,0.0381270684,0.4585378468,-0.3772488832,-0.491268158,-0.1910640746,-0.1587199569,-0.0277118497,0.0377318859,0.0575775579,0.2001122385,-0.1936412901,-0.0350448787,0.1537034363,0.1532455683,0.0386600159,-0.2126651406,-0.0145917758,-0.0151801519,-0.1022223905,-0.0853901356,0.1332870275,0.1747178137,-0.0071969456,0.1526002586,-0.0898960382,0.2490469515,0.2958875,0.4298882782,0.0385020413,-0.0166205857,-0.276522994,-0.6019979119,0.1072554365,0.2519675195,-0.2931177616,-0.3241817653,-0.0773591474,-0.2935066819,0.0149025014,-0.2205447853,0.0091186985,0.1636463404,0.2512086332,0.0141901895,-0.0200967919,-0.245231986,0.2676571906,-0.3195354939,0.1948679537,-0.2083370388,0.201210171,-0.2217801511,-0.2969256938,-0.1804705113,0.1420997977,0.2723119855,0.0925014988,-0.2997994125,0.4652596712,0.0059302798,0.0698818564,-0.129803732,-0.0326074213,-0.1389693618,-0.1473561078,0.0775982514,-0.0047066477,0.2419632226,0.0690333918,-0.362295717,0.0900742859,-0.0637699068,0.2205827683,0.0301657431,0.1150783226,0.1708155274,-0.1170947775,-0.0064452961,-0.2951321304,0.0090424838,0.066455327,0.2198874354,-0.2932216227,-0.1216397136,0.0560927466,0.4650214016,0.064876847,0.1760996878,0.1861424446,-0.0901378095,0.0198986512,-0.0121839065,0.614377141,0.3814362288,0.0090051927,-0.0581259727,-0.0293169972,-0.1447795033,-0.0531236529,0.1978241801,0.010228049,0.1976764053,0.1335366517,0.336348325,0.0209108517,-0.1498912573,-0.1165845022,0.1740596294,0.3704490364,-0.3543183208,0.0263497047,-0.1302931756,0.1459719688,0.0057367785,-0.2700253725,-0.139059633,-0.3707582057,-0.3373607099,0.6752127409,0.0063468511,0.0844874233,-0.1134621948,-0.1232812926,0.0266843792,0.1229012609,-0.3791603148,-0.043895971,0.0552867576,0.0447739065,0.0763220713,-0.2286300957,0.5205156803,-0.0397290662,0.0512861572,-0.5295776129,-0.1521557868,-0.2153201699,-0.127969563,0.3062783182,0.1174204946,0.3138974607,-0.1450230181,-0.2756850421,0.2403118312,0.1121567413,-0.1098957136,0.1784870476,0.1534111351,-0.0302763302,-0.0489839315,-0.1432929337,-0.2808938026,-0.3785196543,0.3248309791,-0.0642383248,0.1248255149,0.0986435562,-0.1798856705,0.1114000976,0.0895826668,0.0878933594,-0.2645496428,-0.1945210695,0.4365966916,-0.238999784,-0.2203456312,0.033036951,0.1541690975,0.1376350522,-0.0165131632,-0.248448804,-0.0653736219,-0.248131305,0.4719092846,-0.3034968078,0.1303994358,0.2934890389,0.3040732443,-0.0327981412,-0.0750491172,-0.2942451537,-0.0698595047,0.2907999754,0.2731865048,-0.1038846597,0.4481271505,-0.109040536,0.2552041411,0.1094086394,0.170207873,0.3902416527,0.0170501973,0.1714869291,-0.2735756636,-0.5204055309,-0.0096631926,-0.2813251019,-0.0576935336,-0.2282736003,-0.1018166021,0.0015257325,0.0120262066,0.0294931214,-0.1431079507,-0.2657017112,-0.0998595804,-0.2131252885,0.2652601302,0.0499818362,0.072299026,-0.2405619472,-0.0992755815,0.002741687,0.3389508724,0.1410218328,-0.0870120227,-0.3747829795,-0.0290641636,-0.4143305421,0.2166876793,-0.0847882852,0.521546483,0.0512607731,-0.0871473774,0.0938091725,-0.0112240361,0.5587125421,-0.1291041672,-0.0079693999,0.2835006714,0.3371244967,-0.4498386979,-0.071507901,0.0595117994,0.2570033371,-0.0782480165,0.5526584983,-0.0892554745,0.0265102275,0.1315530241,0.2733247578,-0.2038503438,-0.0962659344,-0.2484713495,-0.2362007052,-0.4278416932,0.016236702,-0.0969549119,0.0553777479,-0.0205207486,0.1808605343,-0.2027261555,-0.2500330508,-0.1371490657,0.1048819423,0.3459596932,-0.1031088307,0.2991208434,-0.0267977621,0.0492437929,0.1733396947,0.2378225327,-0.3225930333,-0.3099865913,0.1373324841,-0.0973120332,0.2458062768,-0.031277094,-0.1000332832,0.1695282459,-0.1443028152,0.0795336589,-0.1888790876,-0.1126142964,0.1532743126,-0.0480418429,-0.29477337,-0.2696514726,0.5598939657,0.1521376818,-0.0256518982,0.3724631071,0.3968023658,-0.2268082201,0.2994340658,-0.0249312464,0.8441890478,0.0668272302,0.1303777695,0.1475183964,-0.4604527056,0.1810048968,0.149248004,0.1174044684,-0.4521838725,-0.1211725846,0.0608230047,-0.0766050369,0.2374438047,-0.0963647366,-0.2902870774,0.1960967332,-0.4051041901,0.1026286408,0.0222633537,-0.0423743203,-0.4003828466,-0.3322468102,-0.0642171875,0.2072711885,-0.0953589007,-0.0329145417,0.0967205167,0.1744714677,-0.1675388962,-0.0488157161,-0.228915289,0.0024294595,-0.0330784917,0.3684213161,-0.3195157349,-0.536077857,0.0869988725,0.0426133946,0.3537866473,0.17561692,-0.1451901644,0.0426505208,0.1901020557,0.3062268794,0.0830521882,-0.065900676,0.2351706028,0.0219595507,-0.1183580682,-0.0441656597,-0.0409800671,-0.300088048,-0.1520019472,0.1203672513,0.3928636909,-0.3056311905,-0.1839401722,-0.1324622929,-0.131596908,-0.5064845085,0.2050891072,-0.061900802,-0.0568435453,0.4300700128,-0.4037243426,-0.158498317,0.0262498166,0.5301278234,-0.0458667688,0.2593889236,0.4973760545,-0.197917819,-0.1586865932,-0.1909323335,0.247181654,-0.0929589123,-0.3724484444,0.2927720845,-0.1032933146,0.1613491923,-0.1380385906,-0.1858425736,-0.0341811031,-0.0735039189,-0.0738415718,-0.4374042749,-0.503583014,-0.0459229723,0.2080612928,0.1226876676,0.0186896101,-0.0717200637,-0.2417615652,-0.1153246164,-0.3387562335,0.0333128311,-0.2050566077,0.2201225609,0.3785303831,-0.0623284318,0.0490255132,-0.3151646554,0.2734014392,0.2116336226,0.1734603047,-0.2822256088,-0.296164006,0.0619238429,0.0030189378,-0.175701037,-0.1491793841,-0.4168162644,-0.027631592,-0.0899941623,-0.1916018873,0.3427523077,0.0935614407,0.1535224617,-0.1112638041,0.1643322706,0.1769546568,0.0509289429,-0.134540841,-0.0925817192,-0.185049817,0.3565359116,-0.0290322918,0.0291560907,-0.6280362606,-0.1532260627,-0.1658821851,0.0737439767,0.2543137074,-0.196634382,-0.1322879046,0.0563972518,0.2729442418,0.3071060479,-0.230272308,-0.3356589973,0.3239755929,0.2371192873,-0.1902782917,-0.1096052751,0.084216997,-0.2513615489,-0.0419559926,0.0426200517,-0.1256052554,-0.0768942684,-0.4428113401,0.0401112847,0.3885582089,-0.433944732,0.1690969467,0.619751513,0.0043319725,0.1773356795,0.3890752494,0.1132771373,0.0138302501,0.6791512966,0.0265437532,0.3208673,0.3201984167,-0.118963629,0.0929332227,-0.0943544284,0.075803034,-0.2698485553,-0.28568995,0.2321783602,0.0891440809,-0.006661355,-0.3313007355,-0.059994448,0.0065475656,0.3166711628,-0.2491111457,-0.0774016455,-0.0508534275,-0.0563096851,0.1967374235,-0.0553785786,-0.2516256273,-0.1057861596,0.4781981111,0.1616422385,0.0480961688,-0.2193864137,-0.0550822318,0.21410878,0.2809601724,-0.2896917164,0.0601712652,0.2971102297,-0.1570805311,-0.2346585542,0.1563749611,0.3924590945,0.1510291249,0.1292324364,0.1491872966,0.046921052,-0.2155083418,0.1399226636,0.1967817992,-0.0572784767,0.0629953668,0.1710461378,0.1972792596,-0.1851183623,0.0799916163,0.1809648275,0.2702733874,-0.094397366,0.4931677878,-0.1464281529,-0.4655370712,0.0049093794,0.0817357153,-0.0488027968,-0.0114344219,0.4544947743,0.0391237512,0.0721523762,-0.1666642129,0.1194605976,0.108324334,0.3438273668,0.373008877,-0.4471422732,-0.3782073557,-0.08310166,-0.4274823368,-0.051516775,-0.01050252,-0.075749658,-0.0498402081,-0.0158979874,-0.0784551427,-0.0717109069,0.2825300694,-0.0765679777,0.031772837,0.4928271174,-0.1742973477,-0.078168571,0.2210267186,0.0540721938,0.0843354166,-0.4384557307,0.3879526258,0.3464070261,0.124876231,-0.280824244,0.0961899236,0.2024360448,0.1928018481,0.4804947078,0.0663611814,0.0273499526,-0.0669154599,-0.3371925354,-0.0151914377,-0.0473074839,-0.0902100876,0.0753803328,0.0914741009,0.422667861,0.0239975005,-0.1941935718,-0.3309696019,0.2139648944,-0.2441732436,0.1477007419,-0.2238080651,0.1157196313,0.2566614151,0.0730567649,0.0829130784,0.0559126474,-0.0265046787,0.6643064022,-0.4486858547,-0.4417956471,0.564638257,-0.0526746437,-0.5221443176,0.0583654419,0.4157539904,-0.1047889963,-0.0768280402,-0.4792670906,-0.1339039505,0.3335670233,-0.138600722,-0.2472850978,0.3773885667,-0.1568926871,0.2129519731,0.0869420469,0.7443235517,-0.0196809433,-0.2162850499,0.211208716,-0.0952281132]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2689","title":"cannot save the dataset to disk after rename_column","comments":"Ok, thanks for clearing it up :)","body":"## Describe the bug\r\nIf you use `rename_column` and do no other modification, you will be unable to save the dataset using `save_to_disk`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nIn [1]: from datasets import Dataset, load_from_disk\r\nIn [5]: dataset=Dataset.from_dict({'foo': [0]})\r\nIn [7]: dataset.save_to_disk('foo')\r\nIn [8]: dataset=load_from_disk('foo')\r\nIn [10]: dataset=dataset.rename_column('foo', 'bar')\r\nIn [11]: dataset.save_to_disk('foo')\r\n---------------------------------------------------------------------------\r\nPermissionError Traceback (most recent call last)\r\n in \r\n----> 1 dataset.save_to_disk('foo')\r\n\r\n\/mnt\/beegfs\/projects\/meerqat\/anaconda3\/envs\/meerqat\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in save_to_disk(self, dataset_path\r\n, fs)\r\n 597 if Path(dataset_path, config.DATASET_ARROW_FILENAME) in cache_files_paths:\r\n 598 raise PermissionError(\r\n--> 599 f\"Tried to overwrite {Path(dataset_path, config.DATASET_ARROW_FILENAME)} but a dataset can't overwrite itself.\"\r\n 600 )\r\n 601 if Path(dataset_path, config.DATASET_INDICES_FILENAME) in cache_files_paths:\r\n\r\nPermissionError: Tried to overwrite foo\/dataset.arrow but a dataset can't overwrite itself.\r\n```\r\n\r\nN. B. I created the dataset from dict to enable easy reproduction but the same happens if you load an existing dataset (e.g. starting from `In [8]`)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.10.0-1160.11.1.el7.x86_64-x86_64-with-centos-7.9.2009-Core\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\r\n","comment_length":7,"text":"cannot save the dataset to disk after rename_column\n## Describe the bug\r\nIf you use `rename_column` and do no other modification, you will be unable to save the dataset using `save_to_disk`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# Sample code to reproduce the bug\r\nIn [1]: from datasets import Dataset, load_from_disk\r\nIn [5]: dataset=Dataset.from_dict({'foo': [0]})\r\nIn [7]: dataset.save_to_disk('foo')\r\nIn [8]: dataset=load_from_disk('foo')\r\nIn [10]: dataset=dataset.rename_column('foo', 'bar')\r\nIn [11]: dataset.save_to_disk('foo')\r\n---------------------------------------------------------------------------\r\nPermissionError Traceback (most recent call last)\r\n in \r\n----> 1 dataset.save_to_disk('foo')\r\n\r\n\/mnt\/beegfs\/projects\/meerqat\/anaconda3\/envs\/meerqat\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in save_to_disk(self, dataset_path\r\n, fs)\r\n 597 if Path(dataset_path, config.DATASET_ARROW_FILENAME) in cache_files_paths:\r\n 598 raise PermissionError(\r\n--> 599 f\"Tried to overwrite {Path(dataset_path, config.DATASET_ARROW_FILENAME)} but a dataset can't overwrite itself.\"\r\n 600 )\r\n 601 if Path(dataset_path, config.DATASET_INDICES_FILENAME) in cache_files_paths:\r\n\r\nPermissionError: Tried to overwrite foo\/dataset.arrow but a dataset can't overwrite itself.\r\n```\r\n\r\nN. B. I created the dataset from dict to enable easy reproduction but the same happens if you load an existing dataset (e.g. starting from `In [8]`)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.10.0-1160.11.1.el7.x86_64-x86_64-with-centos-7.9.2009-Core\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\r\n\nOk, thanks for clearing it up :)","embeddings":[-0.0523223989,0.2342098802,-0.0231681652,-0.0133467484,0.3893551528,0.2707957029,0.4929154813,0.246301353,0.0219596829,0.1201097146,-0.0814653933,0.4607105553,-0.1719950885,-0.1512238532,-0.0136408731,-0.0714220777,0.2988217771,-0.1098150685,-0.0071087461,0.1580941379,-0.4297677875,-0.0167975575,-0.0632793829,0.0525315329,-0.1448873281,-0.1060153246,-0.0191482995,0.0699789524,-0.0107616372,-0.2462689877,0.0901805609,-0.0555133633,0.1750023365,0.3935277462,-0.0001104696,-0.1174478754,0.0169522669,-0.0254376344,-0.3840672076,-0.1677256376,-0.2794014215,-0.22320804,-0.1534653902,-0.5357638001,0.2765216529,-0.0022124713,-0.1228611171,-0.3125068247,0.1796944588,0.2632053494,0.294092983,0.1083573475,0.273332864,-0.1635734141,0.0736200586,0.0986325443,-0.3788083792,0.2685394287,-0.2693726122,-0.0911083221,0.1226996332,0.2915157378,-0.1082791165,-0.0690102279,0.1018080711,0.0468242764,0.1152402088,-0.2415997535,0.2724038661,-0.1014997885,0.4159988761,-0.545006156,-0.4493023753,-0.0069589685,0.2386427075,-0.3515321016,0.2436310351,0.1155513525,0.0246032923,0.1335383654,0.0393228084,-0.2681631148,-0.1927657425,0.0060169129,0.0300406311,-0.2012541443,-0.1730899066,0.1861053854,0.0252395608,-0.1277827471,0.3110419214,-0.0618427843,-0.0558179356,0.0676211342,-0.4282257557,-0.0441422723,-0.0765138939,-0.1082729995,-0.1432772726,0.234134987,-0.0629033148,-0.1409233361,0.2153322846,0.2729771733,-0.0386802107,0.2238942534,-0.0026314,0.4684874117,0.3518581092,0.119733721,0.0032541524,-0.1048212945,0.078337796,0.0482479073,0.6009219289,-0.0478974693,0.4625107348,-0.0603512265,-0.3091149926,0.3789649606,0.1044593304,-0.0342989787,0.0012844725,0.2447030544,0.1551560313,0.0356650017,0.0988404751,0.3300382793,0.1512357295,0.167356506,-0.1845288128,-0.0235995818,-0.0770769566,0.1205223575,0.1922730654,0.2134322971,0.0603944696,0.1418900788,-0.1296938509,-0.0818710178,0.063822411,-0.1976366341,0.0316390581,0.3230943084,0.0812872648,0.2417824864,0.1332234889,-0.2870061994,0.0473018102,0.219573319,-0.125742957,-0.1329015791,-0.4467063248,0.2365868688,0.1273117065,0.1414909065,-0.4421401322,-0.2893771231,0.434397608,-0.0966240466,0.2320122719,-0.0182891302,0.0811488405,-0.409035027,0.0521090887,0.221598804,-0.1067450941,0.2233543247,-0.3136855066,0.1113037318,0.1221690103,0.158576414,0.0759357885,-0.0761187598,-0.1343706846,0.0381270684,0.4585378468,-0.3772488832,-0.491268158,-0.1910640746,-0.1587199569,-0.0277118497,0.0377318859,0.0575775579,0.2001122385,-0.1936412901,-0.0350448787,0.1537034363,0.1532455683,0.0386600159,-0.2126651406,-0.0145917758,-0.0151801519,-0.1022223905,-0.0853901356,0.1332870275,0.1747178137,-0.0071969456,0.1526002586,-0.0898960382,0.2490469515,0.2958875,0.4298882782,0.0385020413,-0.0166205857,-0.276522994,-0.6019979119,0.1072554365,0.2519675195,-0.2931177616,-0.3241817653,-0.0773591474,-0.2935066819,0.0149025014,-0.2205447853,0.0091186985,0.1636463404,0.2512086332,0.0141901895,-0.0200967919,-0.245231986,0.2676571906,-0.3195354939,0.1948679537,-0.2083370388,0.201210171,-0.2217801511,-0.2969256938,-0.1804705113,0.1420997977,0.2723119855,0.0925014988,-0.2997994125,0.4652596712,0.0059302798,0.0698818564,-0.129803732,-0.0326074213,-0.1389693618,-0.1473561078,0.0775982514,-0.0047066477,0.2419632226,0.0690333918,-0.362295717,0.0900742859,-0.0637699068,0.2205827683,0.0301657431,0.1150783226,0.1708155274,-0.1170947775,-0.0064452961,-0.2951321304,0.0090424838,0.066455327,0.2198874354,-0.2932216227,-0.1216397136,0.0560927466,0.4650214016,0.064876847,0.1760996878,0.1861424446,-0.0901378095,0.0198986512,-0.0121839065,0.614377141,0.3814362288,0.0090051927,-0.0581259727,-0.0293169972,-0.1447795033,-0.0531236529,0.1978241801,0.010228049,0.1976764053,0.1335366517,0.336348325,0.0209108517,-0.1498912573,-0.1165845022,0.1740596294,0.3704490364,-0.3543183208,0.0263497047,-0.1302931756,0.1459719688,0.0057367785,-0.2700253725,-0.139059633,-0.3707582057,-0.3373607099,0.6752127409,0.0063468511,0.0844874233,-0.1134621948,-0.1232812926,0.0266843792,0.1229012609,-0.3791603148,-0.043895971,0.0552867576,0.0447739065,0.0763220713,-0.2286300957,0.5205156803,-0.0397290662,0.0512861572,-0.5295776129,-0.1521557868,-0.2153201699,-0.127969563,0.3062783182,0.1174204946,0.3138974607,-0.1450230181,-0.2756850421,0.2403118312,0.1121567413,-0.1098957136,0.1784870476,0.1534111351,-0.0302763302,-0.0489839315,-0.1432929337,-0.2808938026,-0.3785196543,0.3248309791,-0.0642383248,0.1248255149,0.0986435562,-0.1798856705,0.1114000976,0.0895826668,0.0878933594,-0.2645496428,-0.1945210695,0.4365966916,-0.238999784,-0.2203456312,0.033036951,0.1541690975,0.1376350522,-0.0165131632,-0.248448804,-0.0653736219,-0.248131305,0.4719092846,-0.3034968078,0.1303994358,0.2934890389,0.3040732443,-0.0327981412,-0.0750491172,-0.2942451537,-0.0698595047,0.2907999754,0.2731865048,-0.1038846597,0.4481271505,-0.109040536,0.2552041411,0.1094086394,0.170207873,0.3902416527,0.0170501973,0.1714869291,-0.2735756636,-0.5204055309,-0.0096631926,-0.2813251019,-0.0576935336,-0.2282736003,-0.1018166021,0.0015257325,0.0120262066,0.0294931214,-0.1431079507,-0.2657017112,-0.0998595804,-0.2131252885,0.2652601302,0.0499818362,0.072299026,-0.2405619472,-0.0992755815,0.002741687,0.3389508724,0.1410218328,-0.0870120227,-0.3747829795,-0.0290641636,-0.4143305421,0.2166876793,-0.0847882852,0.521546483,0.0512607731,-0.0871473774,0.0938091725,-0.0112240361,0.5587125421,-0.1291041672,-0.0079693999,0.2835006714,0.3371244967,-0.4498386979,-0.071507901,0.0595117994,0.2570033371,-0.0782480165,0.5526584983,-0.0892554745,0.0265102275,0.1315530241,0.2733247578,-0.2038503438,-0.0962659344,-0.2484713495,-0.2362007052,-0.4278416932,0.016236702,-0.0969549119,0.0553777479,-0.0205207486,0.1808605343,-0.2027261555,-0.2500330508,-0.1371490657,0.1048819423,0.3459596932,-0.1031088307,0.2991208434,-0.0267977621,0.0492437929,0.1733396947,0.2378225327,-0.3225930333,-0.3099865913,0.1373324841,-0.0973120332,0.2458062768,-0.031277094,-0.1000332832,0.1695282459,-0.1443028152,0.0795336589,-0.1888790876,-0.1126142964,0.1532743126,-0.0480418429,-0.29477337,-0.2696514726,0.5598939657,0.1521376818,-0.0256518982,0.3724631071,0.3968023658,-0.2268082201,0.2994340658,-0.0249312464,0.8441890478,0.0668272302,0.1303777695,0.1475183964,-0.4604527056,0.1810048968,0.149248004,0.1174044684,-0.4521838725,-0.1211725846,0.0608230047,-0.0766050369,0.2374438047,-0.0963647366,-0.2902870774,0.1960967332,-0.4051041901,0.1026286408,0.0222633537,-0.0423743203,-0.4003828466,-0.3322468102,-0.0642171875,0.2072711885,-0.0953589007,-0.0329145417,0.0967205167,0.1744714677,-0.1675388962,-0.0488157161,-0.228915289,0.0024294595,-0.0330784917,0.3684213161,-0.3195157349,-0.536077857,0.0869988725,0.0426133946,0.3537866473,0.17561692,-0.1451901644,0.0426505208,0.1901020557,0.3062268794,0.0830521882,-0.065900676,0.2351706028,0.0219595507,-0.1183580682,-0.0441656597,-0.0409800671,-0.300088048,-0.1520019472,0.1203672513,0.3928636909,-0.3056311905,-0.1839401722,-0.1324622929,-0.131596908,-0.5064845085,0.2050891072,-0.061900802,-0.0568435453,0.4300700128,-0.4037243426,-0.158498317,0.0262498166,0.5301278234,-0.0458667688,0.2593889236,0.4973760545,-0.197917819,-0.1586865932,-0.1909323335,0.247181654,-0.0929589123,-0.3724484444,0.2927720845,-0.1032933146,0.1613491923,-0.1380385906,-0.1858425736,-0.0341811031,-0.0735039189,-0.0738415718,-0.4374042749,-0.503583014,-0.0459229723,0.2080612928,0.1226876676,0.0186896101,-0.0717200637,-0.2417615652,-0.1153246164,-0.3387562335,0.0333128311,-0.2050566077,0.2201225609,0.3785303831,-0.0623284318,0.0490255132,-0.3151646554,0.2734014392,0.2116336226,0.1734603047,-0.2822256088,-0.296164006,0.0619238429,0.0030189378,-0.175701037,-0.1491793841,-0.4168162644,-0.027631592,-0.0899941623,-0.1916018873,0.3427523077,0.0935614407,0.1535224617,-0.1112638041,0.1643322706,0.1769546568,0.0509289429,-0.134540841,-0.0925817192,-0.185049817,0.3565359116,-0.0290322918,0.0291560907,-0.6280362606,-0.1532260627,-0.1658821851,0.0737439767,0.2543137074,-0.196634382,-0.1322879046,0.0563972518,0.2729442418,0.3071060479,-0.230272308,-0.3356589973,0.3239755929,0.2371192873,-0.1902782917,-0.1096052751,0.084216997,-0.2513615489,-0.0419559926,0.0426200517,-0.1256052554,-0.0768942684,-0.4428113401,0.0401112847,0.3885582089,-0.433944732,0.1690969467,0.619751513,0.0043319725,0.1773356795,0.3890752494,0.1132771373,0.0138302501,0.6791512966,0.0265437532,0.3208673,0.3201984167,-0.118963629,0.0929332227,-0.0943544284,0.075803034,-0.2698485553,-0.28568995,0.2321783602,0.0891440809,-0.006661355,-0.3313007355,-0.059994448,0.0065475656,0.3166711628,-0.2491111457,-0.0774016455,-0.0508534275,-0.0563096851,0.1967374235,-0.0553785786,-0.2516256273,-0.1057861596,0.4781981111,0.1616422385,0.0480961688,-0.2193864137,-0.0550822318,0.21410878,0.2809601724,-0.2896917164,0.0601712652,0.2971102297,-0.1570805311,-0.2346585542,0.1563749611,0.3924590945,0.1510291249,0.1292324364,0.1491872966,0.046921052,-0.2155083418,0.1399226636,0.1967817992,-0.0572784767,0.0629953668,0.1710461378,0.1972792596,-0.1851183623,0.0799916163,0.1809648275,0.2702733874,-0.094397366,0.4931677878,-0.1464281529,-0.4655370712,0.0049093794,0.0817357153,-0.0488027968,-0.0114344219,0.4544947743,0.0391237512,0.0721523762,-0.1666642129,0.1194605976,0.108324334,0.3438273668,0.373008877,-0.4471422732,-0.3782073557,-0.08310166,-0.4274823368,-0.051516775,-0.01050252,-0.075749658,-0.0498402081,-0.0158979874,-0.0784551427,-0.0717109069,0.2825300694,-0.0765679777,0.031772837,0.4928271174,-0.1742973477,-0.078168571,0.2210267186,0.0540721938,0.0843354166,-0.4384557307,0.3879526258,0.3464070261,0.124876231,-0.280824244,0.0961899236,0.2024360448,0.1928018481,0.4804947078,0.0663611814,0.0273499526,-0.0669154599,-0.3371925354,-0.0151914377,-0.0473074839,-0.0902100876,0.0753803328,0.0914741009,0.422667861,0.0239975005,-0.1941935718,-0.3309696019,0.2139648944,-0.2441732436,0.1477007419,-0.2238080651,0.1157196313,0.2566614151,0.0730567649,0.0829130784,0.0559126474,-0.0265046787,0.6643064022,-0.4486858547,-0.4417956471,0.564638257,-0.0526746437,-0.5221443176,0.0583654419,0.4157539904,-0.1047889963,-0.0768280402,-0.4792670906,-0.1339039505,0.3335670233,-0.138600722,-0.2472850978,0.3773885667,-0.1568926871,0.2129519731,0.0869420469,0.7443235517,-0.0196809433,-0.2162850499,0.211208716,-0.0952281132]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2688","title":"hebrew language codes he and iw should be treated as aliases","comments":"Hi @eyaler, thanks for reporting.\r\n\r\nWhile you are true with respect the Hebrew language tag (\"iw\" is deprecated and \"he\" is the preferred value), in the \"mc4\" dataset (which is a derived dataset) we have kept the language tags present in the original dataset: [Google C4](https:\/\/www.tensorflow.org\/datasets\/catalog\/c4).","body":"https:\/\/huggingface.co\/datasets\/mc4 not listed when searching for hebrew datasets (he) as it uses the older language code iw, preventing discoverability. ","comment_length":46,"text":"hebrew language codes he and iw should be treated as aliases\nhttps:\/\/huggingface.co\/datasets\/mc4 not listed when searching for hebrew datasets (he) as it uses the older language code iw, preventing discoverability. \nHi @eyaler, thanks for reporting.\r\n\r\nWhile you are true with respect the Hebrew language tag (\"iw\" is deprecated and \"he\" is the preferred value), in the \"mc4\" dataset (which is a derived dataset) we have kept the language tags present in the original dataset: [Google C4](https:\/\/www.tensorflow.org\/datasets\/catalog\/c4).","embeddings":[-0.1847832352,-0.0222230945,-0.1543211937,0.0329199657,0.0282858852,0.0720350742,0.5509453416,0.3117233217,0.0522171184,0.1043045521,-0.3407955766,-0.2625412643,0.0178138241,0.0098183285,0.1778062731,0.103592433,0.0763270855,0.0028082454,0.0307024941,-0.259934932,-0.2483024001,0.1667094529,0.0356423743,0.230902493,-0.3309554458,0.07476753,0.235079512,-0.0006581638,0.0945436433,-0.2600386441,-0.050563775,-0.0218149852,-0.1151202247,0.4908214509,-0.0001046046,0.200144574,0.3720237911,-0.1461262107,-0.2577262819,-0.402392596,0.0496585295,-0.0652501434,-0.3410946429,-0.1979911625,-0.3883949816,-0.2093005627,-0.0824905932,-0.425860852,-0.0239149425,0.2144082189,0.2111899704,0.4698423445,-0.1391437799,0.106838204,-0.0199652482,-0.2126427442,-0.0785034969,0.2142321914,0.0535802431,0.5240815878,0.2998705506,0.5605412722,0.2111984044,-0.0797846168,-0.122580193,0.2412062436,-0.1161051169,-0.1545200348,0.2309415787,0.2940369546,0.3640371263,-0.3231701851,-0.1078307852,0.0924717933,-0.0515936539,-0.2096336335,0.4878683686,0.1322401017,-0.0833423212,0.2722433805,-0.0162745751,-0.2900920808,0.02577658,0.2554773986,-0.239437148,0.5161865354,0.100475274,0.0479175523,0.102319859,-0.2073668838,0.2490313947,-0.2282048613,0.1534506381,0.094461903,0.0936851203,-0.2620147467,-0.2162770778,0.0145027619,-0.3406135738,-0.3154471219,0.0759732872,0.2595084012,-0.1825710088,0.0420564152,-0.3105461001,0.0559434555,0.430518508,-0.2673490942,0.452997297,-0.2308996022,0.0026237562,-0.1428270787,-0.0379159562,-0.0601268299,-0.3148117661,-0.1479606628,-0.0055956719,0.298242718,0.022936007,-0.0491462201,-0.4983114302,-0.5706431866,-0.0304256361,0.3789435327,0.1616676897,-0.3083958924,-0.1461149007,-0.0396290906,-0.3200576603,-0.541968286,-0.2043112516,0.1096073166,-0.1466782838,-0.0825254545,0.0807791725,-0.1820848584,0.3079834878,-0.0946761668,0.0314200222,0.1600391567,-0.2420831472,0.0018396663,0.3078255057,0.0659999773,-0.1263970733,0.0821514726,-0.0706280172,-0.2200494111,-0.1961161941,0.1325772256,-0.3130845428,0.1174960509,0.2477213144,0.2029940188,-0.1501318961,-0.004773824,0.062870726,0.2855404615,-0.0142905442,0.0629693493,0.3462482393,-0.0515146144,0.0772701204,-0.1047443822,-0.104922384,0.2785845995,-0.4833551645,-0.4456771016,0.1126218736,-0.0339967497,-0.0810493007,0.5016024113,-0.2918809354,0.172396794,-0.174690783,0.3707315326,0.2921551168,-0.305390507,-0.3107973635,-0.0459155217,0.0442110077,-0.0580154881,-0.1685320586,-0.1529696882,0.0498177521,-0.1401434541,0.0799029768,-0.06562832,0.2023756951,-0.1784927249,-0.0507951081,0.0377850682,0.3984745741,-0.012750457,0.1060536429,-0.0219229851,0.1121249497,-0.0090246676,0.2545722127,-0.1508146524,0.2193585634,0.0614949539,0.0400639139,0.1059669182,0.1841665953,-0.0109765185,-0.1952796727,0.0795870051,-0.1208200753,0.261523813,-0.0632150024,0.0669061467,0.0704118013,-0.1375233829,-0.1428229362,-0.1268767118,0.2377082407,0.035947483,-0.1261019558,0.2591712177,-0.2723131776,0.0940321535,-0.3269435465,0.0285445116,-0.1553587168,0.1628706157,-0.1221369654,0.0393177085,-0.2129364759,0.3802617788,0.1243670732,-0.4671669304,-0.0933214128,0.0844818577,-0.2620695233,0.1656294763,0.1111799106,0.5108319521,0.5489289761,-0.3658974767,0.3155207038,0.1858113557,-0.0510408618,-0.1089917719,0.2349341065,0.2308352292,0.3495796919,0.1166996881,0.1217742115,-0.2267134488,0.4270852506,0.0358429216,-0.1411843598,-0.3088137209,-0.0927452818,0.1146655604,0.5891355872,0.2524322271,-0.1919254959,0.0864684582,1.0686205626,-0.2139691412,-0.0719673559,0.0956165865,-0.2867286205,0.1048127189,0.034709055,-0.1156415418,0.0591344461,0.1322740465,0.2981178164,0.1810235977,0.1942354739,-0.2748283744,0.3244132698,0.2057235241,0.0029113751,-0.003260385,0.0801432282,-0.1853951067,-0.4277556837,0.1569376886,0.0582927838,-0.190404743,-0.227196306,0.2576307952,-0.4687194228,-0.1034525409,-0.363117069,-0.4276822507,-0.5842372179,-0.0740773231,0.0689160675,-0.2383491844,-0.0100949146,0.2999588251,0.3809861243,0.1638946831,-0.3429346383,0.0557869337,-0.2732775211,-0.1871110052,-0.2862946093,0.1406190097,0.116354309,-0.3061376214,0.2784307897,-0.3796431124,-0.2748767734,-0.1865570247,-0.4455599785,0.4134855866,-0.168808952,-0.1180184036,-0.1187332571,-0.0272593647,-0.0475958586,-0.0813059881,-0.0034812698,-0.2797719538,-0.2955234349,-0.2735096216,-0.0544816703,0.0387494825,-0.0833983943,-0.5276796222,-0.6048954129,-0.0471697003,0.0081236148,-0.1959532201,-0.0276988689,0.1241637021,-0.2377651036,-0.1558343023,0.2911198437,-0.2068577558,-0.3472431004,-0.092574589,0.203680858,-0.4344355166,-0.0920788348,0.1560345143,0.0218184832,0.4703041017,0.0739345849,-0.0249648634,0.3892793953,0.0256208591,0.2439542264,0.2953723967,-0.0541684367,0.2268235832,0.005541089,-0.1026123986,-0.3346844316,0.0345104076,0.2723380923,0.0560688637,0.3472976387,0.4592175782,0.1106088683,0.1023152396,0.4332106709,0.1335740983,-0.3373230398,0.317191869,0.194379285,0.3283284009,-0.2244925946,-0.0435064398,0.1524136811,0.0891056359,0.240294978,0.3073195815,-0.0582777932,-0.1126121059,-0.0151067944,-0.0689152628,-0.4128376544,-0.1973827779,0.0830010325,0.168519944,0.3129772544,0.4733056128,0.0657677948,-0.1082371399,-0.1908295751,0.3424507678,0.1806497127,-0.1493627727,0.1426171809,-0.3924442232,-0.2826724648,-0.5756276846,0.1183251888,0.1205575615,0.2156473696,-0.0242032204,0.0632096455,0.1214141995,0.1216598898,0.4390310347,-0.3354904056,-0.0578701235,0.1001624838,-0.0477768369,0.1152476072,-0.0892253444,-0.4080406427,0.107246384,0.5992045999,0.2594999969,-0.3122331798,-0.3205437064,-0.0046130177,-0.1259613782,-0.2476203442,-0.2767196894,-0.0570680052,-0.1351335198,-0.1376323849,-0.3238414526,0.1389372498,-0.1093696654,-0.1388468593,0.1005039737,0.3756710589,-0.1748195589,0.2626543641,0.1723303795,0.1697658896,0.3030052185,-0.2415591031,0.4288002849,0.3547996879,0.4820906222,0.3926686347,-0.1090644225,-0.3789312243,-0.0814877003,-0.2688362896,0.2420612574,0.4999493957,0.1379871666,0.0685851052,0.0340784565,-0.1236998513,-0.1190894246,0.1812736839,0.242228806,0.3162495792,-0.2268504053,-0.414228797,0.2150868326,-0.3416469991,-0.0082802949,0.272593081,0.3026970923,-0.3349891305,0.043440111,0.0965402424,1.0030599833,0.1250052154,-0.0432857536,0.1459399313,-0.1284821182,0.2652156353,-0.4718643427,-0.0268328246,-0.0550214499,0.3103027642,-0.1304385662,-0.0171361025,-0.0788591951,0.4574534595,0.0710953921,0.3584782183,0.0126647446,0.2063193321,0.2466016114,0.1002297029,0.0953696594,-0.1974560916,-0.1075203344,0.1276667416,-0.001697113,0.2657646835,-0.1674702018,-0.2553716898,-0.0706623867,0.0818955526,0.0557383597,-0.1231256798,-0.2867535353,0.3074192405,0.0006250644,-0.2795165777,0.2378388345,-0.041066844,0.8017966151,-0.0641528144,-0.1351787597,0.3463771343,0.1039818227,0.0479529351,-0.1543041468,-0.0186003763,0.1850298941,-0.1475439668,-0.4431726635,0.120592244,0.0532843657,-0.2749581039,-0.2868185043,-0.0032006812,-0.2513613403,0.2532125413,0.2322181463,0.2417040616,-0.0840610787,-0.159155935,0.1464568228,-0.3185131848,0.0585341416,-0.3114402294,0.2933880687,-0.1985556632,-0.3302376568,0.0831745267,0.2927169204,-0.0769592449,0.3099735379,0.0649394318,-0.366409421,-0.1230964363,-0.2681157589,0.3906423748,-0.1110104695,-0.0792763233,0.0877892002,-0.0982678607,-0.4158188701,0.1906227171,0.0250191465,0.1156906411,-0.3648862243,-0.0180266052,-0.0166496411,0.2528652251,-0.2255443931,-0.087367475,0.1048659533,0.5610258579,-0.0658606142,-0.0028842185,-0.3408403099,0.0499163941,0.0808831155,0.0166207477,0.1480215788,-0.1096231341,0.2010758668,0.0578284971,0.0989102721,0.0325098708,-0.3431233168,-0.2032050192,-0.1824482232,0.0659679472,-0.037118271,-0.0201197732,-0.0106992386,0.0225328747,0.2019939721,-0.2463760972,0.1390529126,0.1408370733,0.2217383832,0.422120899,0.3345205486,-0.1149063557,0.1129058078,-0.0554931909,0.2061089724,0.4034079909,0.1521256566,0.0237269271,-0.0052109123,0.0482607894,0.0213153195,0.1940066665,0.0163733847,0.1617128551,0.200746119,-0.308359623,0.197500214,0.0398347937,0.1542032361,0.1245033816,-0.3596285284,-0.3416137397,0.1954770535,0.2210427374,-0.3405461311,-0.2511644065,0.2323724329,0.158882156,0.2442333102,0.2402065545,0.6543265581,0.2860357463,0.1946170777,-0.0912278369,0.4273202717,-0.4784704745,0.0240703914,-0.1936154962,0.0144335814,0.2894088626,0.5641240478,0.0060032844,-0.0756480098,0.5502743721,0.0972605199,0.1881621033,0.1705637723,0.2047504783,0.0370847546,-0.1647953391,-0.1329272687,0.0236482676,-0.1242481768,0.3433175087,0.1148875877,-0.1867021471,0.277374804,-0.3252136707,-0.0014708946,-0.0278655458,-0.2706683576,-0.0555339269,-0.1109921634,-0.1249318197,-0.033250995,0.1942179203,0.1114231572,-0.0464754477,0.0796874985,0.1588647813,-0.1216886714,0.0806339309,-0.1010041013,-0.2985258996,0.0116500678,-0.0998261198,0.2514214516,0.0764719322,-0.1048840284,0.099544622,-0.0504339337,0.08873377,-0.1728899032,0.1524936855,0.2156714499,-0.2303614467,-0.1926901191,-0.1122017801,0.0510225557,0.3161513805,-0.1240313575,0.0932135358,0.1494020075,-0.1309225708,-0.1643325239,0.3775977492,0.171593979,-0.1905042976,-0.0233264882,-0.2705205381,-0.0307782963,-0.3539096117,0.0057449574,-0.4128543139,0.3363720179,0.4712307155,-0.3097417653,-0.0023272068,-0.0018184204,0.1136795059,0.0390474387,0.4614466131,-0.0712390989,-0.0401123799,0.010814542,-0.3932800293,-0.3123115301,-0.0498899929,-0.2666439414,0.7237253785,-0.2092271894,0.1534152478,0.4759406149,-0.1284080446,0.5029777288,0.0292533394,-0.0931169763,-0.323589325,-0.3461982012,-0.0431543626,0.0310200509,0.3263247013,0.0789151564,0.1403746456,0.1000128835,-0.276509583,-0.0000702388,-0.2109870613,0.1974184066,0.4074782133,-0.0448526889,0.2684651315,0.1391966045,0.4069573283,-0.2231526971,-0.1018915176,0.1256193817,-0.0264080632,-0.2576507628,0.473819524,-0.0313562937,0.0670869946,-0.1327971518,-0.3989464939,-0.1315800548,0.4251702428,0.1149931103,0.1058412939,-0.2703827024,-0.0304850787,-0.3385732174,0.015402657,0.0925449654,0.3840616047,0.0084443986,-0.1905254275,0.0091775255,-0.2055165321,0.4161725938,-0.28876701,-0.1860843152,-0.1703763455,0.1327438056,0.1181319654,0.1167874485,-0.3195802271,-0.0479250886,0.3436789215,-0.0913560241,-0.3525882959,0.2031983584,0.0227655936,-0.0406860076,-0.0398657955,0.0140538896,0.2593044043,-0.3928056061,-0.0377736911,-0.1988812238]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2688","title":"hebrew language codes he and iw should be treated as aliases","comments":"For discoverability on the website I updated the YAML tags at the top of the mC4 dataset card https:\/\/github.com\/huggingface\/datasets\/commit\/38288087b1b02f97586e0346e8f28f4960f1fd37\r\n\r\nOnce the website is updated, mC4 will be listed in https:\/\/huggingface.co\/datasets?filter=languages:he\r\n\r\n","body":"https:\/\/huggingface.co\/datasets\/mc4 not listed when searching for hebrew datasets (he) as it uses the older language code iw, preventing discoverability. ","comment_length":30,"text":"hebrew language codes he and iw should be treated as aliases\nhttps:\/\/huggingface.co\/datasets\/mc4 not listed when searching for hebrew datasets (he) as it uses the older language code iw, preventing discoverability. \nFor discoverability on the website I updated the YAML tags at the top of the mC4 dataset card https:\/\/github.com\/huggingface\/datasets\/commit\/38288087b1b02f97586e0346e8f28f4960f1fd37\r\n\r\nOnce the website is updated, mC4 will be listed in https:\/\/huggingface.co\/datasets?filter=languages:he\r\n\r\n","embeddings":[-0.1667793095,-0.0835691169,-0.0892025903,0.0611215979,0.2324795276,-0.0180359893,0.2219808996,0.4122384489,-0.0158847161,-0.0184125621,-0.3140928745,-0.1698120981,0.0033469584,0.2545406222,0.1932530552,0.1975350529,0.0163776241,-0.0476836413,-0.0980868042,-0.1150792018,-0.1693115979,0.2598110735,0.0212985426,0.1030976996,-0.3453690708,0.181439966,0.093154043,0.2252604961,-0.1960123926,-0.3190945983,-0.0182252023,0.0939656943,-0.0006310958,0.4357659221,-0.0001059205,0.0330444202,0.3800244629,-0.1992282271,-0.3762791753,-0.2892414033,0.0209638663,-0.2393657118,-0.2559764087,-0.0811617821,-0.4343551099,-0.1628268808,-0.1122972742,-0.4339210689,0.1044049785,0.1989211887,0.2612394691,0.3563450873,-0.032994628,-0.0351646729,-0.117643632,0.0228875559,-0.1134202108,0.1792144179,0.1580500305,0.3489226699,0.1483013779,0.54327178,0.2903345823,-0.0764120519,-0.1122616827,0.1147642285,0.0032077453,0.0572590679,0.3240350485,0.14874883,0.3166532815,-0.2415895164,-0.1554276496,-0.0568174794,-0.1319238693,0.0243273973,0.5333500504,0.174634397,-0.1649092883,0.2587651312,-0.2961873412,-0.3385000825,0.1553282291,0.2498288751,-0.0771950483,0.4242895544,-0.0888629854,-0.0967942774,0.2540954351,-0.3760229051,0.2349133193,-0.0890634507,0.066388838,0.119261086,0.1198623553,-0.1865350008,-0.0233282428,0.4402767122,-0.0654596612,-0.1952186078,-0.0206408892,0.4489215314,0.0282775033,-0.1221648157,-0.2154051512,0.0195822045,0.406290561,-0.2921652198,0.4357168972,0.1794189215,0.0299781561,-0.1222584695,-0.0275853407,0.0099782944,-0.3433042467,-0.1535710096,0.0033574468,-0.0158009231,0.05647083,-0.2361889184,-0.3366084099,-0.5890074372,0.0629421026,0.5665414333,0.131193161,-0.3434547186,-0.1309618205,0.026705008,-0.3744470179,-0.547901094,-0.1846733838,0.0057189474,-0.279368192,0.0802297145,0.2298758477,-0.0261130817,0.2447016388,-0.0560064688,0.0425881706,0.1433309019,-0.2035028934,0.2507776916,0.19294779,-0.0188651159,-0.0510998629,0.0914350227,0.0767901465,-0.0955582783,-0.1163833886,-0.2118172497,-0.0828713104,-0.0378749818,0.3168015182,0.213849172,-0.0961016864,0.0332690775,0.1287437528,0.2874478996,-0.2442082614,-0.0501865856,0.0877918005,0.1112369597,0.1360358894,0.0165860951,0.1065112203,0.5581029058,-0.1508264244,-0.4534163177,0.0059125475,0.0684777871,-0.0526169203,0.2169591486,-0.265027225,-0.0396188609,-0.2780256569,0.199904263,0.2216721624,-0.2611607313,-0.3147504032,0.104289934,-0.0508712083,0.078357622,-0.0582682565,-0.2508698106,-0.0913714021,-0.0312877409,0.14336285,-0.0250134896,0.2120948881,-0.1466607153,-0.1613957137,0.0091552874,0.1078615263,0.0310231112,0.0563399568,0.0332748666,0.2991711795,0.032914944,0.304586798,-0.1429280937,0.3440973759,0.1078212783,0.1012447178,0.1082189232,0.1702220887,0.0276004244,-0.1284164637,-0.0133595234,-0.1266474724,0.4452771246,-0.3261275291,0.0027909689,-0.0843923315,-0.164378792,-0.2925625145,-0.2476252466,0.2804225385,0.2389830053,0.1320311278,0.2433049232,-0.0412337109,0.2225045115,-0.047925584,0.0427353047,-0.1773307621,0.1300029159,-0.1718190014,0.088953115,-0.050500121,0.3366715908,0.2604981959,-0.4208653569,-0.0333489552,0.1978801042,-0.2265546322,0.0181902926,0.4222949445,0.3469171226,0.4615173936,-0.4829180837,0.2479090393,0.2108457834,-0.0170133114,-0.0548951961,0.1468935758,0.2738035917,0.1913130879,0.1775021404,0.2303002328,0.0313565098,0.5397229195,0.0832959339,-0.11530637,-0.5012766123,0.0700699836,-0.1024854556,0.5470167398,0.049325645,-0.2217351049,0.2288285494,0.5540849566,-0.071282573,-0.0024071324,0.0742132291,-0.3126943409,0.0342741758,0.1063324362,-0.0146783413,-0.0114304386,0.2696539462,0.1796233952,0.1794149131,0.1878886521,-0.2812915444,0.3473748565,0.1071499512,-0.0631066337,-0.0326031186,0.1392660737,-0.2676081359,-0.5354726911,0.0588487796,-0.1164289638,-0.240103811,-0.2407671958,0.1507294774,-0.4924585223,-0.2797491848,-0.326993376,-0.4875526428,-0.5873590112,-0.1275563836,0.1273369193,-0.2287453264,-0.0304019973,0.2408620417,0.1011902913,0.0555539355,-0.3660441637,0.0910643041,-0.252566278,-0.11768803,-0.2850552797,0.1581056714,0.2618581355,-0.1413695961,0.3274270296,-0.5285533667,-0.1004667431,-0.403367579,-0.5248311758,0.4519945383,-0.0507598296,0.0864281729,0.1022053137,0.0659651458,-0.1033915803,-0.0635006353,0.3658282459,-0.2740926445,-0.2576243877,-0.0471568182,-0.0331300646,-0.0117174881,-0.1012047157,-0.4515068829,-0.4519301355,-0.1368679404,-0.0734952018,0.0303943362,-0.0678286999,0.0535799935,-0.1967331767,-0.119337514,-0.0210138187,-0.0974937975,-0.3391346037,-0.0635035932,0.3806593716,-0.4907185137,-0.2501942813,0.093572624,0.084723033,0.3325749636,-0.0734401569,-0.0107869888,0.2763570845,-0.0586664565,0.1256780922,0.2447043657,0.1293507218,0.2293200344,0.196904704,-0.1986111999,-0.2833264768,-0.0769396573,0.1819898039,-0.221714735,0.5080351233,0.2685968578,0.0753014386,0.0591795444,0.2819244266,0.2316615582,-0.0314475037,0.5248018503,0.0978564918,0.3908880949,-0.1025177315,-0.1820404083,0.0640388429,0.0860818252,0.2643449903,0.3846732378,0.1578727514,0.0993579477,-0.29878968,-0.0800873041,-0.548273921,-0.2185353488,-0.0431989357,0.1095790416,0.2598342597,0.4453330338,0.0954796597,-0.0852519646,-0.1659410596,0.3163282275,0.2111107111,-0.2077495158,0.2665352821,-0.4891830981,-0.1579335481,-0.4886635244,0.2576249838,0.0615642332,0.1776461601,-0.0852121264,-0.0991786346,0.230993852,-0.0214702841,0.4347790182,-0.3138712049,-0.0723507851,0.0162231661,-0.1698952168,0.040079698,0.0643714145,-0.2964466214,0.1876497269,0.5032931566,0.2043594867,-0.4638075233,-0.2545905113,-0.0145864151,-0.167941466,-0.1308717579,-0.244960472,-0.0860677585,-0.1080698669,-0.1544160098,-0.2814127207,-0.0102719441,0.1695223004,-0.0051586581,0.0454042852,0.3910742998,-0.1713421941,0.213414222,0.4395768344,0.3315402865,0.1093600839,-0.1435861737,0.5581197143,0.4516920149,0.4931240082,0.5625644326,0.0371542051,-0.2261767983,-0.127462998,-0.2054809332,0.1498253495,0.4963572919,0.1881473362,-0.0026740746,0.2085567862,0.0794465542,-0.116013363,0.0048473612,0.0848879293,0.15997684,-0.3092438877,-0.2858945429,0.1967434287,-0.4800972342,0.0227121618,0.1360349357,0.5112382174,-0.3989129663,-0.0560483895,0.038833607,1.0443203449,0.0834265426,-0.0888741314,-0.0380065031,-0.1019888818,0.1708086133,-0.31624493,-0.0056047803,-0.014518477,0.4962679446,-0.0540185459,0.0265616961,0.0236657783,0.306555599,-0.1433669776,0.3628177941,-0.0266659148,0.205555886,0.1672716886,0.0893017724,-0.1165191829,-0.1310853809,-0.400187552,0.2069390416,0.0027552778,0.3016537428,-0.1420215666,-0.3613623381,-0.2012772113,-0.1989098638,0.0840998217,-0.0700171813,-0.3903966248,0.1370287687,0.0416743271,-0.1735155582,0.1987258941,-0.1213123575,0.612411797,0.0338329151,-0.3374807239,0.281167686,0.0589028969,-0.0266480837,-0.169726029,0.0062772995,0.0187074617,-0.1443843544,-0.44610852,0.0388985537,0.0695435107,-0.2094898522,-0.2025313526,-0.1152647957,-0.2897153795,0.1632699668,0.0955132693,0.0633459687,0.0501856729,-0.2919585705,0.1602457315,-0.3013709188,0.0230592676,-0.3883053064,0.5311200619,-0.0046143192,-0.3006282449,0.1869466454,-0.0246386006,-0.1538189501,0.1107926592,0.1692267209,-0.3829511404,-0.1993738413,-0.2265612632,0.3500517309,-0.136296019,-0.1332392395,0.1816168427,-0.1482315958,-0.2114912122,0.3260599971,0.0872609541,-0.0083559779,-0.1710880548,-0.1118518636,0.0189984627,0.3421672285,-0.1568466723,-0.1571946591,0.0508595221,0.4591531754,0.0487178527,0.0771264806,-0.3621399403,-0.0012986803,-0.0575299412,-0.06817884,0.0778997466,0.0007973795,0.2496756166,0.0448003076,0.1284946799,0.0621148832,-0.4078616798,-0.1852899492,0.0112541234,0.0906419605,0.0179830901,-0.2081391811,-0.0924299583,0.1249430701,0.1920030862,-0.2983664274,0.0413169004,0.0788110122,0.144127354,0.0869644359,0.3329894543,-0.4388476312,0.0128057487,-0.0985919982,0.1681391746,0.1985366493,0.1789587736,-0.0708852261,-0.0885999724,-0.0354589447,0.2975753844,0.29213503,0.0417821109,0.2276308984,0.1888563931,-0.2756195366,0.1884188205,0.2838941813,0.3251848817,0.1674548835,-0.2208600938,-0.1551878005,0.1859316677,0.2292253375,-0.424551934,-0.2475094497,0.2709746361,-0.0124707166,0.2413324118,0.2041707337,0.6218895316,0.1191562563,0.2023993582,0.0492181666,0.3567411602,-0.2847537994,0.0334805064,-0.1281650811,0.0752186105,0.2980564237,0.5230849385,-0.1781964004,-0.1314155608,0.570114553,-0.2448916584,0.1640878171,0.0527411476,0.2045455426,-0.1392392218,-0.1326400489,0.1763770431,-0.0400700942,0.0864072293,0.2247385085,-0.0004767918,0.17774719,0.058806587,-0.2287235707,-0.1411402971,-0.0480113253,-0.2195735276,-0.0457099602,0.0910505801,-0.378423512,-0.1171992719,0.0576266274,0.0785728246,-0.1144059598,0.2686808705,-0.0178963151,0.0079783173,-0.4327700138,-0.0489109643,-0.0954196602,0.0395792499,-0.1379525959,0.2741239369,0.1732984483,-0.0963452607,0.2165274471,-0.018226631,0.2154125571,-0.0560002066,0.1153895333,0.0815469995,-0.2719690502,-0.2206633091,-0.2305669785,0.1775818616,0.2813595533,-0.1658408046,0.0341940261,0.1749116778,-0.1940271556,-0.2047531009,0.2067912966,0.1573317349,0.0771390274,-0.0748556554,-0.3715395033,0.123229064,-0.2839228809,-0.0474473312,-0.4346115887,0.4222894311,0.3308522701,-0.3410861492,-0.0260790158,-0.2093765438,0.1231810078,0.2381916791,0.404083997,-0.1132202372,-0.1260968149,0.0394418724,-0.5264502168,-0.5373830199,-0.1035016105,-0.2164760083,0.4150047004,-0.2493384928,0.1059721783,0.5088518858,-0.1090674549,0.5761896372,0.1663047075,0.0506911613,-0.3068360686,-0.3791311979,0.0902083665,0.1699810475,0.3119498193,0.0322106555,0.0636631325,0.2861620188,-0.2917149663,0.0544035621,-0.0495796204,0.2915843427,0.0694974363,-0.3489289582,0.197667256,0.2845183909,0.45946154,-0.3592260182,-0.1285206825,-0.1030104607,-0.1203168556,-0.0390725136,0.4512982965,0.0279221609,0.2293422669,-0.1679071933,-0.4276465178,-0.2277383357,0.5062935948,0.1606005281,0.2274732739,-0.1285736561,0.0581640564,-0.2775188982,-0.0609477423,-0.0101724807,0.2517804801,-0.0292970221,-0.167950362,-0.025076814,-0.3688813448,0.4275469184,-0.2679290771,-0.2136548758,0.0051192148,0.1310735047,0.1438259035,-0.135485962,-0.4822457433,0.0707099214,0.1772769988,0.0607000329,-0.2188800871,0.3369490504,-0.1613610536,-0.1116699949,-0.0617903732,0.092372328,0.3087485135,-0.2268968523,0.0151002491,-0.0535204709]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2681","title":"5 duplicate datasets","comments":"Yes this was documented in the PR that added this hf->paperswithcode mapping (https:\/\/github.com\/huggingface\/datasets\/pull\/2404) and AFAICT those are slightly distinct datasets so I think it's a wontfix\r\n\r\nFor context on the paperswithcode mapping you can also refer to https:\/\/github.com\/huggingface\/huggingface_hub\/pull\/43 which contains a lot of background discussion ","body":"## Describe the bug\r\n\r\nIn 5 cases, I could find a dataset on Paperswithcode which references two Hugging Face datasets as dataset loaders. They are:\r\n\r\n- https:\/\/paperswithcode.com\/dataset\/multinli -> https:\/\/huggingface.co\/datasets\/multi_nli and https:\/\/huggingface.co\/datasets\/multi_nli_mismatch\r\n \r\n \"Capture\r\n\r\n- https:\/\/paperswithcode.com\/dataset\/squad -> https:\/\/huggingface.co\/datasets\/squad and https:\/\/huggingface.co\/datasets\/squad_v2\r\n- https:\/\/paperswithcode.com\/dataset\/narrativeqa -> https:\/\/huggingface.co\/datasets\/narrativeqa and https:\/\/huggingface.co\/datasets\/narrativeqa_manual\r\n- https:\/\/paperswithcode.com\/dataset\/hate-speech-and-offensive-language -> https:\/\/huggingface.co\/datasets\/hate_offensive and https:\/\/huggingface.co\/datasets\/hate_speech_offensive\r\n- https:\/\/paperswithcode.com\/dataset\/newsph-nli -> https:\/\/huggingface.co\/datasets\/newsph and https:\/\/huggingface.co\/datasets\/newsph_nli\r\n\r\nPossible solutions:\r\n- don't fix (it works)\r\n- for each pair of duplicate datasets, remove one, and create an alias to the other.\r\n\r\n## Steps to reproduce the bug\r\n\r\nVisit the Paperswithcode links, and look at the \"Dataset Loaders\" section\r\n\r\n## Expected results\r\n\r\nThere should only be one reference to a Hugging Face dataset loader\r\n\r\n## Actual results\r\n\r\nTwo Hugging Face dataset loaders\r\n","comment_length":45,"text":"5 duplicate datasets\n## Describe the bug\r\n\r\nIn 5 cases, I could find a dataset on Paperswithcode which references two Hugging Face datasets as dataset loaders. They are:\r\n\r\n- https:\/\/paperswithcode.com\/dataset\/multinli -> https:\/\/huggingface.co\/datasets\/multi_nli and https:\/\/huggingface.co\/datasets\/multi_nli_mismatch\r\n \r\n \"Capture\r\n\r\n- https:\/\/paperswithcode.com\/dataset\/squad -> https:\/\/huggingface.co\/datasets\/squad and https:\/\/huggingface.co\/datasets\/squad_v2\r\n- https:\/\/paperswithcode.com\/dataset\/narrativeqa -> https:\/\/huggingface.co\/datasets\/narrativeqa and https:\/\/huggingface.co\/datasets\/narrativeqa_manual\r\n- https:\/\/paperswithcode.com\/dataset\/hate-speech-and-offensive-language -> https:\/\/huggingface.co\/datasets\/hate_offensive and https:\/\/huggingface.co\/datasets\/hate_speech_offensive\r\n- https:\/\/paperswithcode.com\/dataset\/newsph-nli -> https:\/\/huggingface.co\/datasets\/newsph and https:\/\/huggingface.co\/datasets\/newsph_nli\r\n\r\nPossible solutions:\r\n- don't fix (it works)\r\n- for each pair of duplicate datasets, remove one, and create an alias to the other.\r\n\r\n## Steps to reproduce the bug\r\n\r\nVisit the Paperswithcode links, and look at the \"Dataset Loaders\" section\r\n\r\n## Expected results\r\n\r\nThere should only be one reference to a Hugging Face dataset loader\r\n\r\n## Actual results\r\n\r\nTwo Hugging Face dataset loaders\r\n\nYes this was documented in the PR that added this hf->paperswithcode mapping (https:\/\/github.com\/huggingface\/datasets\/pull\/2404) and AFAICT those are slightly distinct datasets so I think it's a wontfix\r\n\r\nFor context on the paperswithcode mapping you can also refer to https:\/\/github.com\/huggingface\/huggingface_hub\/pull\/43 which contains a lot of background discussion ","embeddings":[0.0702893212,-0.0303926785,0.0002651882,0.3026076257,0.1615348011,0.0224872138,0.3812558949,0.1487317532,0.0632141978,0.2060803473,-0.1454353333,0.0585726313,0.1328154206,-0.0630716756,0.0719406009,-0.0062969122,0.247424975,-0.1418853402,-0.2275305986,-0.0822188854,-0.1461635381,0.2783472538,-0.0094332611,-0.1467349678,-0.3533268273,0.0687635094,-0.0858611837,0.4032169282,-0.0831234679,-0.2152673304,0.3845347464,0.2588376701,-0.183336556,0.3635749519,-0.0001085839,0.1327501237,0.1191545427,0.1520772427,-0.3095708191,0.0615057647,0.1082122922,-0.0125753414,0.0173136573,-0.0110088121,0.082987383,-0.3159566522,0.0917756036,-0.243176043,0.4125827551,0.1412670761,0.2476853877,0.230679065,0.0286568329,-0.1598045081,0.3622328341,0.0127873514,-0.0677059665,0.1247701049,0.1705979258,0.0258715153,-0.0382995047,0.5596495271,0.0059267292,-0.1049735472,0.0351694301,0.181407243,-0.0339607224,-0.0845704526,0.0711252242,0.0955600739,-0.0416178964,-0.2665373385,-0.2961679101,-0.3810706437,-0.0798291191,0.1512447596,0.3346849978,-0.1314772815,0.0732252896,-0.1004147381,-0.1617500186,0.1153694317,0.1003635302,-0.0374749899,0.0539106987,-0.0181834996,0.0245162081,0.1457730532,0.2791580558,-0.0956728309,0.0412629656,-0.096397616,-0.2437047511,0.1251586974,-0.3186259866,0.1847620159,0.0741514117,-0.014166831,0.4505223632,0.1745672524,0.1664842963,-0.1277234256,-0.1661153883,0.0062896046,0.2641851604,0.1139489114,-0.203925997,0.1781058908,0.1740802824,0.2517185807,-0.0281441826,0.0542662665,0.1593552232,-0.1900847256,0.0065158191,-0.2685298026,0.2132135481,-0.2687272429,-0.3383703232,0.151846081,-0.0451440439,0.0742845982,-0.2210177332,0.2351892143,0.229051441,0.13244991,0.1639441997,-0.0091079175,-0.2775206268,-0.2676825225,-0.3610759676,0.0377252363,-0.16983895,-0.0133068375,0.0501572415,-0.3885450363,0.2597261071,0.3444271684,0.2077880204,-0.204994604,0.0852164328,0.0182680655,-0.046519652,0.1520562619,0.0271938052,0.1776815504,0.1024062037,-0.2849942744,0.078703627,-0.0889813155,-0.0825323537,-0.0154928183,-0.2907681167,0.2522127628,0.007712075,0.0648043007,-0.0191126317,0.1038396135,0.0420782417,-0.1259654462,0.0561734289,0.0037836186,-0.0485528484,-0.0468566678,0.0524536744,0.5406343341,0.1214762032,-0.3457823694,0.2633025348,-0.1628337502,0.1812824905,0.2803883255,-0.2148378938,-0.0444383584,-0.4685078263,-0.2270385772,-0.0370501317,-0.3887757361,-0.4229961932,0.1572689563,-0.2734949291,0.3648278713,0.2113612592,0.0895469412,-0.0669702142,-0.0258573629,-0.1110830083,0.0277507845,0.0342650451,0.1007514894,-0.4402085543,-0.3214183152,0.0769081712,0.1578864604,-0.0070328852,-0.0400566123,0.2076918036,0.0139018679,0.3016479909,-0.0779571384,-0.0256003998,0.1917275488,0.1942148805,-0.0533731952,0.316943258,0.1118596792,-0.5789576173,0.1740713716,-0.2582848072,0.0822262838,0.1265852898,-0.3745349348,-0.126064539,-0.0862090215,-0.4764727652,-0.5406158566,0.1560219526,0.2849034071,-0.0491379164,-0.1052725315,-0.2403944135,0.5512455702,-0.2626474798,0.2151258886,-0.4765309989,0.1979452521,-0.0749793202,0.1527921259,0.117269285,0.1010274887,0.0043886839,0.1111822426,0.0473269969,0.2589428127,0.0552445762,0.3331555128,0.1939175278,-0.0155926934,0.3004575074,-0.1741740555,-0.1487469673,0.1129791513,0.0885687545,0.0601312518,-0.3865839541,0.2396647036,-0.2031966746,0.0032817652,-0.0548340864,-0.1308239549,0.2923994362,-0.1416111887,-0.0034342739,-0.5109378695,0.2535313368,0.0886523053,0.4695344865,0.1972254366,-0.2442840189,0.1927701533,0.0407992937,0.1407828033,-0.1942288876,-0.0162592698,-0.1412751377,-0.1349201798,-0.0798633024,0.1790764034,0.3696672916,0.2851987779,-0.1035690382,0.1510406286,-0.1323779225,-0.1944527477,0.2456227243,-0.2350581884,-0.0799733102,0.203520596,0.321978718,0.1378724575,-0.6552867889,-0.0686195642,0.0124594215,-0.0030521797,-0.486017704,-0.163531512,-0.1157722473,-0.3308982551,-0.450673908,0.1937761307,-0.3726775646,-0.2968254089,0.2177338302,0.1844272166,-0.2909290493,0.055937849,-0.1773156375,0.509455502,-0.1905319691,0.1508851796,-0.1049923375,0.1799394786,-0.3306490183,0.1656581759,0.0899941474,-0.1211987138,0.3523873687,-0.0389506519,-0.2115743905,-0.5472989678,-0.5236175656,0.2084300816,0.1391555518,0.3696638644,0.3066101372,-0.12800093,-0.0571224093,-0.2924894094,0.0273073744,0.1066731438,-0.3492100537,-0.0958679169,-0.3238349855,0.1498928964,-0.1476000696,-0.1449313313,0.0635176674,-0.2053784728,0.0606595092,-0.1372239143,0.0845991597,0.3050029576,-0.1538970321,0.0602918193,-0.3173991144,-0.0388752185,-0.4645876884,-0.4354595244,-0.0108941216,-0.1829256266,-0.2920600772,-0.1587105244,0.0971509442,-0.041164957,0.142351985,-0.687128067,-0.4630126655,-0.1058348864,0.0027167872,0.0282152183,0.1861343682,0.2162626684,0.142097801,-0.2022675276,-0.1953265369,-0.0114189545,0.3829001784,-0.1636298895,0.3035388291,-0.1152472422,0.1026080549,0.0764358491,0.3137539029,0.8434222937,-0.0014699423,0.4338806868,-0.0673912466,0.3124116957,-0.0208490472,-0.1881705821,0.1442436576,-0.0755560026,0.1585820615,0.4018268883,-0.046321027,-0.0364354663,0.0287632924,-0.1463023871,-0.2654136717,-0.2283642739,-0.0789313242,0.0548282191,0.0716530904,0.0920652226,-0.2884059846,-0.0331864096,-0.1515438557,0.157295078,0.0370928049,0.1450941265,0.1348901242,-0.2468572706,0.2602545619,-0.2389468849,0.1679644585,0.1347388178,0.1514575481,-0.0158140231,-0.0577350818,0.0136937648,-0.0228131507,0.7903946638,-0.0340327322,0.2042674422,0.0278102607,-0.1460229605,-0.1756715775,-0.1385099441,0.0538256504,-0.1323562413,0.0421092734,0.4128994942,-0.3416718245,0.0711300671,0.4328715503,-0.0443177298,-0.1570342332,-0.1497314274,-0.1578308791,-0.3200147748,-0.1570153683,0.0271465722,-0.0239898842,0.1375602186,-0.0673896074,-0.1892472953,-0.1364288628,-0.2950667143,0.085898824,0.0667081624,0.3032994568,0.2619000375,0.466904372,0.4708637595,0.3350967169,0.1465926468,1.0519518852,-0.1441850513,-0.51976192,0.0216763727,-0.0926425681,-0.0051405807,0.4899062514,0.0896271989,0.1215441823,0.064738363,0.0467976965,0.0508454591,-0.0193282533,0.2704710662,-0.0013513727,-0.1833829582,-0.1507109404,0.427587986,0.1461550593,-0.1242991462,0.4206101298,0.2107418031,-0.3146691024,0.2819368243,0.0404135548,0.9385852218,0.1558808386,0.2149765491,0.0375557542,-0.269210428,0.5875102878,-0.0289388988,-0.1418969929,-0.2176423371,-0.4918928146,0.1432878822,-0.0101483054,-0.0341528952,-0.0936076716,0.0485005043,0.0905280113,-0.0571779944,0.2710876763,0.0720700845,0.0879282355,-0.2031649798,-0.1366329491,-0.1881481111,0.1276730597,-0.0253962558,0.0104331253,0.0216567889,0.0600931048,-0.0322426781,-0.3847247064,-0.1123057827,0.1674442589,-0.5221056342,0.141577825,0.0915613919,0.0177282896,-0.0306471027,0.0966902301,0.4413726628,-0.1222486496,-0.0585908629,0.226809293,0.249611333,0.2676342428,-0.1077380776,-0.1987310797,0.3762801588,-0.1605619341,-0.0839054734,-0.1093350351,0.1593042165,-0.2135662138,-0.4714793265,0.0761865303,-0.4596892595,-0.045473367,-0.0102323536,0.0326745473,0.2039303482,-0.2811032236,0.1403630823,-0.0378480852,-0.0130375596,-0.0909868926,0.0814655423,-0.3286118209,-0.1354033053,0.5333234072,0.2865935862,-0.2705764472,0.1788006723,0.1192112118,-0.0377222747,-0.0770085603,-0.2253712863,0.5069448352,-0.6518533826,0.1981790513,-0.3616855145,-0.0053390381,-0.0027533618,0.4223929942,0.1842178553,0.0392529927,0.1321129799,-0.3254236877,-0.0498325974,0.2862562835,-0.130844295,0.3340655267,0.1672133505,0.2803391218,-0.1923380941,0.1796861589,-0.3889954984,0.0801327229,-0.0645375401,0.010260528,0.1012523323,-0.015119723,0.1557547897,-0.0029579836,0.1402813345,0.2721829116,-0.1252323836,-0.2485208213,-0.3904340863,0.1317447126,-0.1210007668,-0.0705044717,0.0038770731,-0.0281607471,0.1299735755,-0.042563919,0.5148792863,0.0577580668,0.0089614578,-0.209215045,-0.005176194,-0.0291192103,0.0734870359,0.0426977426,-0.038006749,0.0841839686,0.3089031577,0.0767982453,-0.3439321518,-0.1398139447,0.2705498338,0.1343269497,0.1485643834,0.1832740158,-0.016715141,-0.2120540291,-0.2311356217,0.2464067638,0.3307932913,0.2452835441,-0.4638030827,-0.1886005253,-0.1990245581,0.2618715763,-0.4365540445,-0.4500624239,-0.0400506966,0.0169516969,-0.0279507078,0.1592962593,0.4390650988,-0.2071944624,0.0170791,0.0962801799,0.5881931782,-0.2251541317,0.1943998486,-0.1835407913,-0.102593936,-0.0928413942,-0.1002477258,0.0791403651,0.203596279,0.2097256929,-0.1540225297,0.3255163133,0.0098198075,0.3314209878,0.2330665141,-0.221032843,0.1576393694,0.4859611392,-0.2427341938,0.3098114133,0.0793527439,0.1540991515,0.110753186,-0.110852696,-0.2532792091,0.3032995164,-0.0106300972,-0.0971802473,-0.1020239815,-0.3589640558,-0.5747909546,0.2360834479,-0.2080262005,-0.1438653469,0.3955699801,-0.048891332,-0.3522392213,-0.5194411278,-0.1329230517,0.2853826284,0.3495978117,-0.4062205851,0.3400549591,0.6960601807,-0.2867567241,0.2309230268,0.5722973347,0.4350290895,0.100193508,0.3267898262,0.0076676221,-0.0682522207,0.1063891202,-0.2850441039,0.1740033627,-0.1171878055,-0.2777267098,0.3249938488,0.153609708,-0.2502075136,-0.3245690465,0.0223234501,0.3017281294,-0.4690503478,0.128637135,-0.3570268154,0.2763645947,-0.0031686632,0.0753341615,-0.4512562454,-0.2382504195,0.3177053928,-0.0202254187,0.1521885544,-0.0654857308,0.1148128957,0.1379262656,0.4535553157,0.2049616128,-0.0245894641,-0.0404889509,-0.4924447536,-0.5506696701,0.3423610628,-0.1147316024,0.2573940158,0.1016030461,0.1014395207,-0.0690052956,-0.0847876966,0.2085533738,0.21211496,-0.0459356979,0.1773765236,-0.3282060027,-0.0099743707,-0.1070321873,-0.1118798405,0.0895476416,-0.2577895522,0.2668514252,-0.0040977462,0.0399764031,0.0481396839,0.0566333309,0.1933891922,-0.2851020098,0.2023127824,0.3519156873,0.2769759595,-0.1271612644,-0.3484401703,-0.296931088,-0.1960515231,-0.3339642584,0.2065487355,0.1690260917,0.3120585382,0.006532107,-0.4518444538,-0.0230731089,0.1562701762,0.2045668066,0.1513562799,-0.5145587325,0.1581469923,-0.0708075091,0.0137268286,0.141933009,0.4631643295,0.0257923119,0.0870908573,-0.0072634048,-0.5874010921,0.2039549351,-0.5411157608,0.0610732473,-0.0182316415,0.2844233513,0.0451311953,-0.0510971397,-0.5387969017,0.0153928548,0.3055503964,0.0063499254,-0.0691114217,0.3009802401,0.1168983132,0.0154018849,-0.0079800822,0.0861914158,0.3158290088,-0.2034040689,0.3573434949,-0.2383292764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2681","title":"5 duplicate datasets","comments":"Thanks for the antecedents. I close.","body":"## Describe the bug\r\n\r\nIn 5 cases, I could find a dataset on Paperswithcode which references two Hugging Face datasets as dataset loaders. They are:\r\n\r\n- https:\/\/paperswithcode.com\/dataset\/multinli -> https:\/\/huggingface.co\/datasets\/multi_nli and https:\/\/huggingface.co\/datasets\/multi_nli_mismatch\r\n \r\n \"Capture\r\n\r\n- https:\/\/paperswithcode.com\/dataset\/squad -> https:\/\/huggingface.co\/datasets\/squad and https:\/\/huggingface.co\/datasets\/squad_v2\r\n- https:\/\/paperswithcode.com\/dataset\/narrativeqa -> https:\/\/huggingface.co\/datasets\/narrativeqa and https:\/\/huggingface.co\/datasets\/narrativeqa_manual\r\n- https:\/\/paperswithcode.com\/dataset\/hate-speech-and-offensive-language -> https:\/\/huggingface.co\/datasets\/hate_offensive and https:\/\/huggingface.co\/datasets\/hate_speech_offensive\r\n- https:\/\/paperswithcode.com\/dataset\/newsph-nli -> https:\/\/huggingface.co\/datasets\/newsph and https:\/\/huggingface.co\/datasets\/newsph_nli\r\n\r\nPossible solutions:\r\n- don't fix (it works)\r\n- for each pair of duplicate datasets, remove one, and create an alias to the other.\r\n\r\n## Steps to reproduce the bug\r\n\r\nVisit the Paperswithcode links, and look at the \"Dataset Loaders\" section\r\n\r\n## Expected results\r\n\r\nThere should only be one reference to a Hugging Face dataset loader\r\n\r\n## Actual results\r\n\r\nTwo Hugging Face dataset loaders\r\n","comment_length":6,"text":"5 duplicate datasets\n## Describe the bug\r\n\r\nIn 5 cases, I could find a dataset on Paperswithcode which references two Hugging Face datasets as dataset loaders. They are:\r\n\r\n- https:\/\/paperswithcode.com\/dataset\/multinli -> https:\/\/huggingface.co\/datasets\/multi_nli and https:\/\/huggingface.co\/datasets\/multi_nli_mismatch\r\n \r\n \"Capture\r\n\r\n- https:\/\/paperswithcode.com\/dataset\/squad -> https:\/\/huggingface.co\/datasets\/squad and https:\/\/huggingface.co\/datasets\/squad_v2\r\n- https:\/\/paperswithcode.com\/dataset\/narrativeqa -> https:\/\/huggingface.co\/datasets\/narrativeqa and https:\/\/huggingface.co\/datasets\/narrativeqa_manual\r\n- https:\/\/paperswithcode.com\/dataset\/hate-speech-and-offensive-language -> https:\/\/huggingface.co\/datasets\/hate_offensive and https:\/\/huggingface.co\/datasets\/hate_speech_offensive\r\n- https:\/\/paperswithcode.com\/dataset\/newsph-nli -> https:\/\/huggingface.co\/datasets\/newsph and https:\/\/huggingface.co\/datasets\/newsph_nli\r\n\r\nPossible solutions:\r\n- don't fix (it works)\r\n- for each pair of duplicate datasets, remove one, and create an alias to the other.\r\n\r\n## Steps to reproduce the bug\r\n\r\nVisit the Paperswithcode links, and look at the \"Dataset Loaders\" section\r\n\r\n## Expected results\r\n\r\nThere should only be one reference to a Hugging Face dataset loader\r\n\r\n## Actual results\r\n\r\nTwo Hugging Face dataset loaders\r\n\nThanks for the antecedents. I close.","embeddings":[0.0829851255,-0.1076155156,-0.0309701115,0.3780257106,0.2171273977,0.0425042287,0.4238376021,0.1185694709,0.0298578348,0.2367315441,-0.11034251,0.1014472544,0.1050269678,0.0182530712,0.0259088296,-0.0115713412,0.2270835787,-0.1464574784,-0.1989684105,-0.0486159697,-0.144487977,0.2756378353,-0.0563893616,-0.1630808115,-0.3958544135,0.0806531608,-0.0764266327,0.3414957523,-0.1392159015,-0.1970451325,0.3659976125,0.1712640524,-0.210951075,0.3358832002,-0.0001074394,0.0986455679,0.1678007841,0.1287909597,-0.2695356905,0.0627846867,0.066136159,0.0022383647,0.0897111595,-0.0550277345,0.078158319,-0.2666882873,0.1042815447,-0.2564461231,0.5225718021,0.1103311181,0.2633324265,0.2091964781,-0.007276007,-0.1367125362,0.2754723132,0.008755005,-0.0649417341,0.1575710773,0.2393498272,0.0993351713,0.0113247642,0.5593844056,-0.0711542293,-0.1209552363,0.0270656236,0.157293871,-0.0015271136,-0.1208903939,0.0702085644,0.0576524064,0.0335056558,-0.2908940613,-0.3126444221,-0.3753782213,-0.0313444473,0.1214718893,0.3329713047,-0.1377215087,0.0991565585,-0.0397212394,-0.2359951288,0.1243308485,0.0755200237,-0.0437185988,0.0310093369,-0.0641650483,0.0020592841,0.1569307297,0.2417702079,-0.1621809602,0.0332256146,-0.0525192097,-0.1972812861,0.0840685889,-0.3355811834,0.1515242755,0.0302576181,-0.0039726575,0.4404887557,0.2591186464,0.2485753596,-0.0953096524,-0.1221359074,0.019798398,0.2555045784,0.1678067744,-0.1460874975,0.1148674637,0.1674222797,0.188510105,-0.0935595483,0.0986997634,0.1398757249,-0.1925842017,0.0076341736,-0.1964120567,0.270465374,-0.2263042331,-0.3380233645,0.1265388131,-0.036577329,0.0837661251,-0.1689340025,0.2163385153,0.2037184536,0.0958432108,0.1220968664,0.03388175,-0.2446605116,-0.3587274849,-0.3652580678,0.013679062,-0.204408288,-0.065370366,0.0862493888,-0.4045987725,0.306340605,0.3399194479,0.147747457,-0.2962387204,0.0988697633,0.0602443144,-0.0664924458,0.1010792777,0.0005405656,0.2019430846,0.0827555433,-0.2393064052,0.1017551199,-0.0614273064,-0.0856052041,-0.0618612655,-0.2929818034,0.2632038295,-0.0637470409,0.0582980476,-0.0269044917,0.0810854957,0.0615122877,-0.1324466616,-0.0043954942,0.0397010818,-0.0372084044,-0.1054700613,0.0709785372,0.537042141,0.164591074,-0.3510141671,0.1961857677,-0.169881478,0.18551898,0.2889931202,-0.1668428034,-0.0619312413,-0.4176809192,-0.2187329531,-0.0421584807,-0.3700031936,-0.4381489754,0.0808020905,-0.3149140179,0.3125063777,0.2186381221,0.1261587292,-0.0540282205,0.0072578737,-0.0864982009,0.0556704216,0.0341292843,0.1164614633,-0.437171191,-0.3547561467,0.1262613982,0.1406144947,-0.0576362982,-0.1241005957,0.1643970907,-0.0060342369,0.2664408088,-0.0682631284,-0.0285939574,0.1626154035,0.1637615263,-0.1121459901,0.2983109355,0.0745667294,-0.656317234,0.2078751177,-0.2037440538,0.0626871288,0.0852371603,-0.3362874985,-0.1612803936,-0.0481437519,-0.4903667867,-0.5183190703,0.1625499576,0.2933423519,-0.0219975226,-0.0869539157,-0.2803246677,0.4818663895,-0.2157179415,0.2085496932,-0.5426595211,0.1397224367,-0.0835138634,0.1304191053,0.1452495605,0.1113530397,0.0354260318,0.0879909769,0.0920188874,0.2392628044,0.0914939418,0.3947122097,0.1573657542,-0.027884189,0.302526176,-0.2083860636,-0.1395204961,0.1744082868,0.0927907601,0.0399306491,-0.3922187686,0.2563982904,-0.2268476188,-0.0164900515,-0.117196694,-0.0480373353,0.3050113916,-0.1751668155,-0.0538716726,-0.5137615204,0.2468091249,0.0875725523,0.4690731466,0.2368940264,-0.2649301887,0.1746334136,-0.0093830498,0.1493581086,-0.1572528929,-0.0486216173,-0.1384628713,-0.1321720928,-0.1212716475,0.2458830476,0.3264389634,0.2664453685,-0.0685997009,0.1398857981,-0.0467287078,-0.1981027871,0.2712133825,-0.195199132,-0.1060180292,0.1941611022,0.3234312236,0.1233382747,-0.5904504061,-0.043816518,0.0574592054,0.0070381109,-0.4460994303,-0.2005286813,-0.0941656679,-0.3660759628,-0.4709175825,0.0921204463,-0.3795728385,-0.242471844,0.2095194161,0.1726233512,-0.3526107967,0.0885820389,-0.1255492866,0.5027791858,-0.1776020825,0.1763579249,-0.1012801826,0.1807447821,-0.2632961869,0.1758145541,0.1499112844,-0.092287831,0.3440003395,-0.1263424307,-0.1485261321,-0.5246498585,-0.4854647219,0.1493713856,0.1202674061,0.3866564035,0.2907900512,-0.1087879241,-0.061661914,-0.3028994203,0.0689580515,0.0716643035,-0.3553073704,-0.1015756056,-0.228708759,0.1531338245,-0.1588887274,-0.2012438774,0.0386685804,-0.2064411342,0.113047488,-0.1352057308,0.1026085988,0.3094643056,-0.1484943032,0.0797082186,-0.2895186841,0.0304114278,-0.4099815786,-0.4385839999,-0.0289916005,-0.1782868207,-0.2783553898,-0.1396600604,0.110012345,0.0118729994,0.1340688765,-0.6314336061,-0.4356019497,-0.1265996248,-0.0445878617,0.0589725226,0.1609381139,0.1964104474,0.1636225879,-0.1963183284,-0.199839294,-0.055738356,0.3655629456,-0.1697330773,0.2897583246,-0.114694111,0.0636219978,0.0752329901,0.3535115123,0.8597479463,-0.0238043349,0.3567320406,-0.0554988869,0.2448342592,-0.0060026958,-0.2253536284,0.0965585187,-0.0815113187,0.1061928496,0.4190345407,-0.0184432976,-0.0130523164,-0.0507767759,-0.1050020829,-0.342487067,-0.222238332,-0.1108296812,0.0317750871,0.0745243952,0.0963902399,-0.3184947073,-0.0163788591,-0.1220563799,0.0801777393,-0.0216300618,0.1204551458,0.0990570113,-0.280415833,0.2203644067,-0.2516908348,0.1832295358,0.1314854622,0.1651589721,0.0276942197,-0.0745188072,0.0375763886,0.0263455622,0.8054099083,-0.0349620804,0.2442902178,0.0114073679,-0.1099433452,-0.2418548763,-0.0681631267,0.0553720631,-0.0597113594,0.0881068558,0.3763348162,-0.3201550841,0.1157260686,0.3902792037,0.0590792559,-0.1594683826,-0.1230613962,-0.1550804526,-0.3093456626,-0.1673439592,0.0109198447,0.0353038758,0.1560970545,-0.0369092114,-0.1915078312,-0.1590287089,-0.3005969524,0.1147505566,0.0877751857,0.3695265949,0.230223909,0.541762054,0.4906654358,0.3256964982,0.0998103544,1.0418337584,-0.1086399928,-0.5451667309,-0.06186001,-0.1065241173,0.0041573169,0.4634348452,0.094315134,0.0806133375,0.0097474353,-0.0370810181,0.0961971134,-0.0256482568,0.2518001497,0.053298939,-0.1808661819,-0.1472725123,0.4449433386,0.1768585145,-0.1554090977,0.4044350982,0.1348468363,-0.2862853706,0.2656615376,0.0396112278,0.9905830622,0.0916805416,0.1698267311,0.0088694552,-0.2083519846,0.5911485553,0.1167627275,-0.0857205242,-0.2621173561,-0.482452035,0.1615466923,-0.019635411,-0.0665022507,-0.0926391631,0.0039426326,0.052929122,-0.0621556528,0.2204668671,0.0837866589,0.0733760297,-0.1834898293,-0.0887566358,-0.1639351845,0.1173632592,-0.052461084,0.0307641011,0.0248716343,0.0429947488,-0.0182259269,-0.409275353,-0.1192297265,0.1280943304,-0.4984422326,0.1759338379,0.0896781459,-0.0342097394,-0.0800834,0.2357635647,0.3982121944,-0.1249332279,-0.1150127128,0.2071108222,0.2123713344,0.2407184243,-0.0935419425,-0.1723194569,0.3417997658,-0.1665623635,-0.0669873729,-0.1470486075,0.2141506374,-0.175229013,-0.4635776281,0.1466592699,-0.4125468731,-0.0665328801,-0.0727275833,-0.0009667877,0.163607344,-0.3034588397,0.145427838,-0.0003480094,-0.0309954789,-0.0239560101,0.0339858457,-0.3281144202,-0.1008741707,0.5480856299,0.304615438,-0.3072265983,0.1934396178,0.1354522854,0.0053077391,-0.0870820582,-0.1723820567,0.3778327107,-0.6282155514,0.2383081615,-0.4064598978,0.0480334163,0.0183214936,0.4203711152,0.1938085258,0.0926516205,0.1568671763,-0.3044437766,-0.0804131702,0.3070616424,-0.0940938815,0.3407396078,0.1383229792,0.2470348626,-0.1189785674,0.1373418421,-0.3924759328,0.1460850537,-0.0897486508,0.0512100607,0.1523006111,-0.0391586684,0.0996528193,0.0141981626,0.1301029921,0.2421108335,-0.1120210588,-0.261257112,-0.3284693062,0.1276257187,-0.1364535391,-0.0574812442,-0.018917283,-0.0272912811,0.1620443165,-0.0555376597,0.419893533,0.0971842259,0.0383872464,-0.2042646557,-0.0399198718,0.0299124364,0.0775327012,0.0296080671,-0.0577272661,0.044288449,0.2892661989,0.0800348669,-0.2819232345,-0.1542639285,0.2743318081,0.0483308733,0.1162263379,0.1745982021,-0.0435123704,-0.1903005391,-0.2428722084,0.3194539845,0.3492427766,0.2895379663,-0.4959171116,-0.1170180067,-0.187320888,0.2846493721,-0.433695972,-0.4097294211,0.0019007432,0.0529005863,-0.0669689029,0.1633708477,0.3810757995,-0.2394639552,0.0155893937,0.0939534232,0.6153839827,-0.2089773417,0.2802320123,-0.1751033515,-0.1318558455,-0.1785283387,-0.0562993996,0.1276495755,0.2354510427,0.2324856967,-0.2124322355,0.335160166,-0.041716069,0.2993440628,0.2216992229,-0.1610769331,0.151435554,0.4801160693,-0.2486007661,0.3099756241,0.0677457824,0.2323823422,0.0335334204,-0.0816499963,-0.2736996412,0.2762947977,-0.0021485547,-0.1356745064,-0.0896537974,-0.3763157725,-0.5219728947,0.2295209914,-0.2288047969,-0.1153228953,0.4203801155,-0.0197891966,-0.3682139814,-0.5339972377,-0.1812485605,0.365495801,0.3424223661,-0.4848727584,0.3353043497,0.7156754136,-0.2647434473,0.2283589989,0.5817952752,0.4783472121,0.0858240873,0.2846254408,0.0636030734,-0.1314548701,0.0923703387,-0.3130145371,0.2244254649,-0.1200171262,-0.2438609302,0.3321384788,0.1652647555,-0.2445125729,-0.2826802135,0.0731149837,0.3491365612,-0.4898867309,0.0480762273,-0.399865061,0.2101245075,-0.0723326206,0.0704743639,-0.5086477399,-0.2518310845,0.2917710245,0.0008953735,0.1208568811,-0.0715812519,0.1156040281,0.0985105336,0.4622027576,0.222587347,0.0184331182,-0.059199173,-0.5027386546,-0.5635710359,0.3672677279,-0.0595680065,0.2695916593,0.0632983446,0.0966918319,-0.1360212713,-0.0175968353,0.1647987962,0.2913304567,-0.1034862697,0.172163561,-0.3884851336,-0.0391867049,-0.0216259249,-0.1081611812,0.1590933651,-0.239686355,0.2902083695,-0.0771125481,0.0461586267,0.0264537018,0.0443836413,0.1431119144,-0.1878481209,0.2072128057,0.3852783144,0.2910854816,-0.0911953896,-0.3596329391,-0.3491081893,-0.2067830861,-0.2884871066,0.2503961921,0.2750522196,0.2973109484,-0.0545589142,-0.3415000737,-0.0031443418,0.1271832436,0.1904280633,0.127393052,-0.4955118895,0.0618522801,-0.0486388691,0.0122815166,0.0864730775,0.4840242863,-0.0431540497,0.0674273968,-0.0387280472,-0.597808063,0.2478353083,-0.5086542964,0.0293231327,-0.0347541869,0.2833380997,0.0308676846,-0.001782098,-0.5506199002,0.0432116501,0.2933330238,0.0014058085,-0.0846149623,0.2946508229,0.1380811334,-0.0134432651,-0.0041934787,0.0862083882,0.387219578,-0.1752901077,0.3821578324,-0.2379760146]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2679","title":"Cannot load the blog_authorship_corpus due to codec errors","comments":"Hi @izaskr, thanks for reporting.\r\n\r\nHowever the traceback you joined does not correspond to the codec error message: it is about other error `NonMatchingSplitsSizesError`. Maybe you missed some important part of your traceback...\r\n\r\nI'm going to have a look at the dataset anyway...","body":"## Describe the bug\r\nA codec error is raised while loading the blog_authorship_corpus. \r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nraw_datasets = load_dataset(\"blog_authorship_corpus\")\r\n```\r\n\r\n\r\n## Expected results\r\nLoading the dataset without errors.\r\n\r\n## Actual results\r\nAn error similar to the one below was raised for (what seems like) every XML file.\r\n\/home\/izaskr\/.cache\/huggingface\/datasets\/downloads\/extracted\/7cf52524f6517e168604b41c6719292e8f97abbe8f731e638b13423f4212359a\/blogs\/788358.male.24.Arts.Libra.xml cannot be loaded. Error message: 'utf-8' codec can't decode byte 0xe7 in position 7551: invalid continuation byte\r\n\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 856, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 583, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 671, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.15.0-132-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyArrow version: 4.0.1\r\n\r\n","comment_length":43,"text":"Cannot load the blog_authorship_corpus due to codec errors\n## Describe the bug\r\nA codec error is raised while loading the blog_authorship_corpus. \r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nraw_datasets = load_dataset(\"blog_authorship_corpus\")\r\n```\r\n\r\n\r\n## Expected results\r\nLoading the dataset without errors.\r\n\r\n## Actual results\r\nAn error similar to the one below was raised for (what seems like) every XML file.\r\n\/home\/izaskr\/.cache\/huggingface\/datasets\/downloads\/extracted\/7cf52524f6517e168604b41c6719292e8f97abbe8f731e638b13423f4212359a\/blogs\/788358.male.24.Arts.Libra.xml cannot be loaded. Error message: 'utf-8' codec can't decode byte 0xe7 in position 7551: invalid continuation byte\r\n\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 856, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 583, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 671, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.15.0-132-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyArrow version: 4.0.1\r\n\r\n\nHi @izaskr, thanks for reporting.\r\n\r\nHowever the traceback you joined does not correspond to the codec error message: it is about other error `NonMatchingSplitsSizesError`. Maybe you missed some important part of your traceback...\r\n\r\nI'm going to have a look at the dataset anyway...","embeddings":[-0.2084893584,0.4780819416,-0.0540075116,0.3998489976,0.3617135882,0.2268402874,0.0785168484,0.4288438559,-0.1586735994,0.2435750514,0.0468282923,0.2679276764,-0.0666647777,-0.210872665,0.0490831248,0.0175640285,-0.0882997364,0.1426831633,0.2282305956,-0.2342811376,-0.2054748684,0.1818809807,-0.1930260658,0.0730281472,-0.2108267397,-0.0789327174,0.1461101174,0.6233423352,-0.1766088754,-0.6049428582,0.1305889636,0.0211400893,0.1271802336,0.3438818455,-0.0001129081,-0.0840206072,0.5609151125,-0.0908816382,-0.2077678293,-0.2428600639,-0.2933349609,0.0893766582,-0.1876998544,-0.2272577435,0.0519394279,0.1707786471,0.043988727,-0.5675269365,0.3152764142,0.5750642419,0.2217924744,-0.0919267312,0.0331094116,-0.0800493881,0.3778606355,0.045461297,0.1114239097,0.4461227059,0.4049824476,-0.106069237,-0.0645619035,0.3080867231,-0.3914480805,-0.0931516141,0.0810792297,-0.1104288027,0.2468281239,-0.3147915006,0.6316772699,0.1682745665,0.7650758624,-0.2163730562,-0.081857644,0.0109544322,0.1951058507,-0.273940444,0.1889482737,0.3846224546,-0.3076028824,0.0981161445,-0.0437706299,-0.0178626496,-0.2440257967,-0.0302386694,-0.1594134271,0.0686632022,-0.0978547037,0.0309896469,-0.2485807091,0.0831889808,0.3431307375,-0.087953411,-0.2779461741,0.3806451559,0.0356344506,0.1754511893,0.1777861267,0.0299991332,-0.0623696409,0.1734724194,-0.0929429084,0.1898393929,0.1581863463,0.2450632006,0.0953593701,-0.052051533,0.1189913675,0.3499027193,0.2957804501,-0.0092535857,-0.0625836328,-0.0884831399,-0.3842275739,-0.385307163,-0.113360554,0.2477751374,0.2131842673,-0.3699801266,-0.4927609563,0.3212743998,-0.0740120709,0.0194718782,-0.0607549623,0.485186547,0.1276381463,0.0470056459,0.2408136576,0.2430116534,-0.1895996183,-0.0786771104,-0.1897735894,0.1530003697,-0.0485872775,-0.0438676737,0.1317065358,-0.5761977434,0.1176543161,-0.0184416268,0.3553464711,-0.320558995,-0.3360753059,-0.0648012385,-0.1094540656,0.0805431306,-0.0559896976,0.0071818228,0.1801805049,-0.3210016191,-0.0618890561,-0.0260087419,-0.1589974463,-0.3241128623,-0.0479485355,0.2177026421,0.0325050503,0.1240164414,-0.2615481317,0.098749727,0.4569669068,-0.1838753968,0.0121197803,-0.0366272479,-0.1992596835,-0.1116067916,0.0685284734,0.7190669775,0.0927176028,0.0617560372,0.4470205903,0.220451206,0.2396672964,-0.0346576385,-0.0303007979,0.194758296,-0.2161441147,0.1940477937,0.173850432,-0.6035790443,-0.0017290334,0.1910260767,0.0441783108,0.1295305341,0.0547363125,-0.3995582759,0.4223296642,-0.3531521559,0.1809348166,0.4003559351,-0.0078087784,0.0422336236,-0.2096998394,-0.0094913309,-0.0592940226,0.1407908052,0.0956548601,0.0371472165,0.4447281957,0.0195302796,0.5081382394,-0.2271026224,-0.0435088165,0.2044648826,-0.0637107193,0.0057082479,-0.0157999657,-0.2292633206,-0.1008971632,0.2268027067,-0.0385448001,-0.0186795779,-0.3047531843,-0.1203672215,-0.3363896608,0.0941970125,-0.1724899709,-0.0096021341,0.1468279064,0.0825056061,0.1227488443,0.211104691,-0.2485478967,0.2506260276,-0.3510062695,0.1294386089,-0.5143997073,0.2901625633,0.0366383158,-0.1802812517,0.3910512626,0.5273911357,0.2063739449,-0.1315859556,-0.2450412661,0.3295736909,-0.2906770408,0.6357571483,-0.3684304059,-0.2102894187,0.1556368917,-0.3921348453,0.1912713796,0.1315571219,0.2919704914,-0.1065121442,0.2268299609,0.1296588629,0.0019443497,0.0616183653,0.1570881456,-0.0124361692,0.3217045665,-0.0014806424,-0.1980579346,-0.306335181,0.6450024843,0.2185741812,0.1565416306,-0.105205372,-0.2416180521,0.0593610816,0.4993925989,-0.0996020436,-0.0267609991,0.035702005,-0.0107226865,0.0234232377,0.0127148628,-0.1461623758,0.1471957862,0.0217034426,-0.1039544493,0.0323545709,0.0716076419,-0.3352051079,0.3919107318,-0.1687837392,0.0142022185,0.4510241747,0.0282918811,0.0123838438,-0.3308029473,-0.1881506145,0.0471752398,0.0889911354,-0.5186877847,0.0840452239,-0.2582147419,-0.2728693783,-0.2343662679,-0.3481079042,-0.0588170327,-0.1013338491,-0.061281193,0.1418228,-0.0765725076,0.0614699721,-0.3226076365,0.1384917498,-0.1622715741,0.1579311192,-0.1613741368,-0.1625829339,-0.1952576339,-0.0374632478,0.4106931388,0.2090358436,0.145612061,-0.3906173706,0.0756685361,-0.1550291479,-0.1967244893,0.1922615021,-0.0072839498,0.0445013195,-0.096276179,-0.1030256748,0.6314195991,-0.1873525977,0.0774794072,0.0403758325,-0.00959244,0.4120475352,0.1020493954,-0.4308469594,-0.0713257641,-0.239278838,-0.3840438426,-0.5113678575,-0.2183594406,0.0891492143,0.2631490529,0.4002134502,-0.2125230432,-0.0471867919,-0.0296742711,0.3316637278,-0.1299691945,-0.2295024246,0.203759864,-0.2909738719,-0.3221518993,0.1063784137,0.0682759359,0.3978834748,-0.0266790073,-0.2990857959,0.200989604,0.0736920312,0.0807962045,-0.1829186678,-0.0326646566,0.1681961864,0.0659452304,0.0074228281,-0.076543875,0.0337187611,-0.3264494538,-0.0253692903,-0.1287384629,-0.0559628531,0.2026890516,-0.1725133657,0.061919529,0.4852911234,0.0982005894,0.2867949903,0.0515121557,0.21362333,0.0147207957,-0.2460832298,0.4707962275,-0.1930654645,-0.224330157,0.3553866446,0.0529203303,-0.1348503232,-0.2152939588,0.0723881498,-0.1209119037,-0.2952137887,0.0524808206,-0.3898859024,0.2485878766,0.0668382794,0.3902979493,0.0765489116,-0.2069254071,-0.0084624765,0.3520234525,0.1568837911,-0.125199005,0.014238039,0.1312190145,-0.4801469743,0.344866395,0.2102483958,-0.0640166476,-0.0432956107,0.2608812153,0.0587284379,-0.1514606029,0.0913045034,-0.2261163145,0.267290622,0.2079252452,0.3854580224,-0.2534746528,-0.0441583097,-0.0180206317,0.4818536937,-0.0801013783,-0.0459683985,-0.0982749909,0.1211590767,0.0731569082,0.1224754825,-0.0888519958,-0.1877179295,-0.2884795368,-0.3591709733,-0.3103108108,-0.2335228026,0.0687817186,0.3102971613,-0.1312476397,-0.1198605895,-0.0953507051,-0.0344561301,0.1942464113,0.0822998956,0.0921882316,-0.075224705,0.1156871542,-0.0131787956,-0.1346069872,0.3768363297,0.7526932359,-0.1334387064,-0.5529475808,-0.1159119308,-0.3700479567,0.0784900784,0.4450315237,-0.2124755681,-0.1339324266,0.0593456626,-0.0505468845,-0.3437484801,-0.2992713153,0.0504755303,0.0993438736,-0.2784160078,-0.5206426978,0.4623544812,-0.1271226257,0.0311399177,0.1468494684,0.0418167189,-0.3473612964,0.1372857243,0.0064516137,0.9900810122,0.214081645,-0.1357927918,0.075341925,-0.2174910456,0.0528732426,-0.2248955965,0.1980188191,-0.2843383253,0.0222397074,-0.151835233,0.2045907527,0.2688502371,0.062127769,-0.2158529311,0.1995509863,-0.0714000538,-0.2375721633,-0.0485025756,0.0031614071,-0.4199114144,-0.2580136359,-0.5787045956,0.1172822863,-0.1179324016,0.1192467287,-0.118909888,0.0359294116,0.0455081798,-0.3266947865,-0.4608321786,0.3758393228,-0.2024473846,-0.1504061669,-0.0577517897,-0.3814416826,0.3199534714,0.0561904982,0.2504869401,0.0391977765,-0.0429419242,0.1452626884,0.198122099,-0.0339066684,-0.0243588984,0.0854877979,0.2385652214,-0.0663192645,-0.1132121161,0.2658363581,-0.0436588787,-0.1837543845,0.2556534111,0.066129759,-0.0826183781,-0.0395169929,-0.043281734,0.0955182016,0.1809807569,-0.0629983768,0.1403061599,0.1276887804,-0.2898816168,-0.054411184,-0.110742934,-0.2697612345,0.0812905729,0.2204324752,-0.0361849926,-0.0838951319,0.4614356756,0.4005638361,-0.2099898458,-0.1176223457,-0.3479477763,0.1872604489,-0.5638506413,0.0368575789,-0.0007268398,0.0062479069,-0.1586897224,0.4170023203,-0.1392202228,-0.0937972292,0.0625543371,-0.5232264996,-0.0387310758,0.1154574826,-0.010349893,0.032931786,-0.0487522334,0.000144313,-0.2369749844,0.1144792289,-0.3007834852,0.1038622558,-0.2077763975,0.1372835636,0.4296366274,-0.0340136513,-0.1030200869,-0.2039479315,0.1335907727,0.3360125124,-0.1127493456,-0.150005728,-0.0085570309,0.1377923191,0.0324160047,-0.0805118382,0.0895498246,-0.0501768626,-0.1057639048,-0.1934644133,0.2179796398,0.1474811435,0.0261413902,-0.0277224965,-0.4482190609,0.2940060794,-0.0006671938,0.2398097217,-0.1485024691,0.4302434027,-0.1185299233,0.1239034459,-0.2088452727,-0.0353734829,-0.1150486767,0.2561821043,-0.096136108,0.0655865893,0.436314702,-0.2451550215,0.3256570995,0.4871932566,-0.0835472047,0.3445193172,-0.409155339,0.0281194597,0.0262620784,0.2102447897,-0.1836768985,-0.0181143712,-0.4960899055,-0.0479394235,-0.0900042951,0.2297486365,0.1105230153,-0.3707810938,-0.2305627912,0.0906378776,0.4767543972,-0.0063164178,0.2191631645,0.0231047086,0.0493144915,-0.0192854162,-0.0702998638,0.2188107371,0.0461868681,0.2720300555,-0.4646321833,0.3093931973,0.2494731247,0.1801015735,-0.0085307807,-0.4411627054,0.048236642,0.2103961706,0.0852203667,0.4002787471,0.1609909683,0.1586003602,-0.1475254446,-0.1259341389,-0.1613467187,0.0410661325,-0.0583181717,0.0472529568,-0.3145262897,-0.2136671394,-0.1724198312,-0.1043776944,-0.1697824746,0.1193011403,0.0753186196,0.1556896269,-0.3476271927,-0.3463916481,0.137742281,-0.2096372694,0.244703114,-0.1253280193,0.5787795782,0.4382135272,0.0846361071,0.4258896708,0.3017472923,0.7615882754,0.305842042,0.0535050742,0.1927803457,-0.1810554266,-0.1250636578,0.1044058949,0.3984232247,0.0448813178,0.3953417838,0.0439625829,0.1863657087,-0.138005659,0.1443944722,0.2655777335,0.4306209385,-0.191459626,0.2662214637,-0.1739098132,-0.0572185442,-0.419878304,-0.0946548581,-0.3771162927,0.1363275051,0.2699846923,-0.0920646265,0.095140107,-0.0083605088,0.1025270373,0.0740036219,0.3087226152,0.4034377933,0.1357361525,0.0255365111,-0.2081612647,-0.4297647178,0.3247804344,-0.0908724591,-0.3741289973,-0.06660413,0.0824070871,0.1919348389,0.1155832782,0.1238217652,-0.0432212763,0.383332938,0.0597238839,-0.1724222451,-0.0146099431,0.2414140701,-0.0620328858,-0.3025852442,-0.3642935455,0.0325732566,-0.2273191065,0.0701559708,0.0482755043,-0.0399629399,-0.0922599137,-0.2646709979,0.3722178042,0.0610680655,0.2166769505,-0.3107425272,-0.2267269194,-0.366376698,-0.1446740031,-0.15449965,0.4856240153,0.1616640836,0.2741329968,-0.0787693486,0.0253111199,-0.2682861984,0.1879413426,-0.1179688945,0.2693620622,-0.3983275592,-0.0619250089,-0.130013749,0.063481167,-0.0002447377,0.2416365594,0.013725874,0.0728774592,-0.1267946959,-0.6861604452,0.4283364117,-0.5565597415,-0.4156657457,0.0878274739,0.6013883948,0.0051166438,0.0725319907,-0.642991662,0.2427362651,0.0948262587,-0.2015708536,-0.1887482554,-0.0072774659,-0.1239597574,-0.1315533072,-0.0403240211,0.0875855461,0.2611950338,-0.1649687439,0.1641228795,-0.408061862]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2679","title":"Cannot load the blog_authorship_corpus due to codec errors","comments":"Hi @izaskr, thanks again for having reported this issue.\r\n\r\nAfter investigation, I have created a Pull Request (#2685) to fix several issues with this dataset:\r\n- the `NonMatchingSplitsSizesError`\r\n- the `UnicodeDecodeError`\r\n\r\nOnce the Pull Request merged into master, you will be able to load this dataset if you install `datasets` from our GitHub repository master branch. Otherwise, you will be able to use it after our next release, by updating `datasets`: `pip install -U datasets`.","body":"## Describe the bug\r\nA codec error is raised while loading the blog_authorship_corpus. \r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nraw_datasets = load_dataset(\"blog_authorship_corpus\")\r\n```\r\n\r\n\r\n## Expected results\r\nLoading the dataset without errors.\r\n\r\n## Actual results\r\nAn error similar to the one below was raised for (what seems like) every XML file.\r\n\/home\/izaskr\/.cache\/huggingface\/datasets\/downloads\/extracted\/7cf52524f6517e168604b41c6719292e8f97abbe8f731e638b13423f4212359a\/blogs\/788358.male.24.Arts.Libra.xml cannot be loaded. Error message: 'utf-8' codec can't decode byte 0xe7 in position 7551: invalid continuation byte\r\n\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 856, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 583, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 671, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.15.0-132-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyArrow version: 4.0.1\r\n\r\n","comment_length":75,"text":"Cannot load the blog_authorship_corpus due to codec errors\n## Describe the bug\r\nA codec error is raised while loading the blog_authorship_corpus. \r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nraw_datasets = load_dataset(\"blog_authorship_corpus\")\r\n```\r\n\r\n\r\n## Expected results\r\nLoading the dataset without errors.\r\n\r\n## Actual results\r\nAn error similar to the one below was raised for (what seems like) every XML file.\r\n\/home\/izaskr\/.cache\/huggingface\/datasets\/downloads\/extracted\/7cf52524f6517e168604b41c6719292e8f97abbe8f731e638b13423f4212359a\/blogs\/788358.male.24.Arts.Libra.xml cannot be loaded. Error message: 'utf-8' codec can't decode byte 0xe7 in position 7551: invalid continuation byte\r\n\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 856, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 583, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 671, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.15.0-132-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyArrow version: 4.0.1\r\n\r\n\nHi @izaskr, thanks again for having reported this issue.\r\n\r\nAfter investigation, I have created a Pull Request (#2685) to fix several issues with this dataset:\r\n- the `NonMatchingSplitsSizesError`\r\n- the `UnicodeDecodeError`\r\n\r\nOnce the Pull Request merged into master, you will be able to load this dataset if you install `datasets` from our GitHub repository master branch. Otherwise, you will be able to use it after our next release, by updating `datasets`: `pip install -U datasets`.","embeddings":[-0.2084893584,0.4780819416,-0.0540075116,0.3998489976,0.3617135882,0.2268402874,0.0785168484,0.4288438559,-0.1586735994,0.2435750514,0.0468282923,0.2679276764,-0.0666647777,-0.210872665,0.0490831248,0.0175640285,-0.0882997364,0.1426831633,0.2282305956,-0.2342811376,-0.2054748684,0.1818809807,-0.1930260658,0.0730281472,-0.2108267397,-0.0789327174,0.1461101174,0.6233423352,-0.1766088754,-0.6049428582,0.1305889636,0.0211400893,0.1271802336,0.3438818455,-0.0001129081,-0.0840206072,0.5609151125,-0.0908816382,-0.2077678293,-0.2428600639,-0.2933349609,0.0893766582,-0.1876998544,-0.2272577435,0.0519394279,0.1707786471,0.043988727,-0.5675269365,0.3152764142,0.5750642419,0.2217924744,-0.0919267312,0.0331094116,-0.0800493881,0.3778606355,0.045461297,0.1114239097,0.4461227059,0.4049824476,-0.106069237,-0.0645619035,0.3080867231,-0.3914480805,-0.0931516141,0.0810792297,-0.1104288027,0.2468281239,-0.3147915006,0.6316772699,0.1682745665,0.7650758624,-0.2163730562,-0.081857644,0.0109544322,0.1951058507,-0.273940444,0.1889482737,0.3846224546,-0.3076028824,0.0981161445,-0.0437706299,-0.0178626496,-0.2440257967,-0.0302386694,-0.1594134271,0.0686632022,-0.0978547037,0.0309896469,-0.2485807091,0.0831889808,0.3431307375,-0.087953411,-0.2779461741,0.3806451559,0.0356344506,0.1754511893,0.1777861267,0.0299991332,-0.0623696409,0.1734724194,-0.0929429084,0.1898393929,0.1581863463,0.2450632006,0.0953593701,-0.052051533,0.1189913675,0.3499027193,0.2957804501,-0.0092535857,-0.0625836328,-0.0884831399,-0.3842275739,-0.385307163,-0.113360554,0.2477751374,0.2131842673,-0.3699801266,-0.4927609563,0.3212743998,-0.0740120709,0.0194718782,-0.0607549623,0.485186547,0.1276381463,0.0470056459,0.2408136576,0.2430116534,-0.1895996183,-0.0786771104,-0.1897735894,0.1530003697,-0.0485872775,-0.0438676737,0.1317065358,-0.5761977434,0.1176543161,-0.0184416268,0.3553464711,-0.320558995,-0.3360753059,-0.0648012385,-0.1094540656,0.0805431306,-0.0559896976,0.0071818228,0.1801805049,-0.3210016191,-0.0618890561,-0.0260087419,-0.1589974463,-0.3241128623,-0.0479485355,0.2177026421,0.0325050503,0.1240164414,-0.2615481317,0.098749727,0.4569669068,-0.1838753968,0.0121197803,-0.0366272479,-0.1992596835,-0.1116067916,0.0685284734,0.7190669775,0.0927176028,0.0617560372,0.4470205903,0.220451206,0.2396672964,-0.0346576385,-0.0303007979,0.194758296,-0.2161441147,0.1940477937,0.173850432,-0.6035790443,-0.0017290334,0.1910260767,0.0441783108,0.1295305341,0.0547363125,-0.3995582759,0.4223296642,-0.3531521559,0.1809348166,0.4003559351,-0.0078087784,0.0422336236,-0.2096998394,-0.0094913309,-0.0592940226,0.1407908052,0.0956548601,0.0371472165,0.4447281957,0.0195302796,0.5081382394,-0.2271026224,-0.0435088165,0.2044648826,-0.0637107193,0.0057082479,-0.0157999657,-0.2292633206,-0.1008971632,0.2268027067,-0.0385448001,-0.0186795779,-0.3047531843,-0.1203672215,-0.3363896608,0.0941970125,-0.1724899709,-0.0096021341,0.1468279064,0.0825056061,0.1227488443,0.211104691,-0.2485478967,0.2506260276,-0.3510062695,0.1294386089,-0.5143997073,0.2901625633,0.0366383158,-0.1802812517,0.3910512626,0.5273911357,0.2063739449,-0.1315859556,-0.2450412661,0.3295736909,-0.2906770408,0.6357571483,-0.3684304059,-0.2102894187,0.1556368917,-0.3921348453,0.1912713796,0.1315571219,0.2919704914,-0.1065121442,0.2268299609,0.1296588629,0.0019443497,0.0616183653,0.1570881456,-0.0124361692,0.3217045665,-0.0014806424,-0.1980579346,-0.306335181,0.6450024843,0.2185741812,0.1565416306,-0.105205372,-0.2416180521,0.0593610816,0.4993925989,-0.0996020436,-0.0267609991,0.035702005,-0.0107226865,0.0234232377,0.0127148628,-0.1461623758,0.1471957862,0.0217034426,-0.1039544493,0.0323545709,0.0716076419,-0.3352051079,0.3919107318,-0.1687837392,0.0142022185,0.4510241747,0.0282918811,0.0123838438,-0.3308029473,-0.1881506145,0.0471752398,0.0889911354,-0.5186877847,0.0840452239,-0.2582147419,-0.2728693783,-0.2343662679,-0.3481079042,-0.0588170327,-0.1013338491,-0.061281193,0.1418228,-0.0765725076,0.0614699721,-0.3226076365,0.1384917498,-0.1622715741,0.1579311192,-0.1613741368,-0.1625829339,-0.1952576339,-0.0374632478,0.4106931388,0.2090358436,0.145612061,-0.3906173706,0.0756685361,-0.1550291479,-0.1967244893,0.1922615021,-0.0072839498,0.0445013195,-0.096276179,-0.1030256748,0.6314195991,-0.1873525977,0.0774794072,0.0403758325,-0.00959244,0.4120475352,0.1020493954,-0.4308469594,-0.0713257641,-0.239278838,-0.3840438426,-0.5113678575,-0.2183594406,0.0891492143,0.2631490529,0.4002134502,-0.2125230432,-0.0471867919,-0.0296742711,0.3316637278,-0.1299691945,-0.2295024246,0.203759864,-0.2909738719,-0.3221518993,0.1063784137,0.0682759359,0.3978834748,-0.0266790073,-0.2990857959,0.200989604,0.0736920312,0.0807962045,-0.1829186678,-0.0326646566,0.1681961864,0.0659452304,0.0074228281,-0.076543875,0.0337187611,-0.3264494538,-0.0253692903,-0.1287384629,-0.0559628531,0.2026890516,-0.1725133657,0.061919529,0.4852911234,0.0982005894,0.2867949903,0.0515121557,0.21362333,0.0147207957,-0.2460832298,0.4707962275,-0.1930654645,-0.224330157,0.3553866446,0.0529203303,-0.1348503232,-0.2152939588,0.0723881498,-0.1209119037,-0.2952137887,0.0524808206,-0.3898859024,0.2485878766,0.0668382794,0.3902979493,0.0765489116,-0.2069254071,-0.0084624765,0.3520234525,0.1568837911,-0.125199005,0.014238039,0.1312190145,-0.4801469743,0.344866395,0.2102483958,-0.0640166476,-0.0432956107,0.2608812153,0.0587284379,-0.1514606029,0.0913045034,-0.2261163145,0.267290622,0.2079252452,0.3854580224,-0.2534746528,-0.0441583097,-0.0180206317,0.4818536937,-0.0801013783,-0.0459683985,-0.0982749909,0.1211590767,0.0731569082,0.1224754825,-0.0888519958,-0.1877179295,-0.2884795368,-0.3591709733,-0.3103108108,-0.2335228026,0.0687817186,0.3102971613,-0.1312476397,-0.1198605895,-0.0953507051,-0.0344561301,0.1942464113,0.0822998956,0.0921882316,-0.075224705,0.1156871542,-0.0131787956,-0.1346069872,0.3768363297,0.7526932359,-0.1334387064,-0.5529475808,-0.1159119308,-0.3700479567,0.0784900784,0.4450315237,-0.2124755681,-0.1339324266,0.0593456626,-0.0505468845,-0.3437484801,-0.2992713153,0.0504755303,0.0993438736,-0.2784160078,-0.5206426978,0.4623544812,-0.1271226257,0.0311399177,0.1468494684,0.0418167189,-0.3473612964,0.1372857243,0.0064516137,0.9900810122,0.214081645,-0.1357927918,0.075341925,-0.2174910456,0.0528732426,-0.2248955965,0.1980188191,-0.2843383253,0.0222397074,-0.151835233,0.2045907527,0.2688502371,0.062127769,-0.2158529311,0.1995509863,-0.0714000538,-0.2375721633,-0.0485025756,0.0031614071,-0.4199114144,-0.2580136359,-0.5787045956,0.1172822863,-0.1179324016,0.1192467287,-0.118909888,0.0359294116,0.0455081798,-0.3266947865,-0.4608321786,0.3758393228,-0.2024473846,-0.1504061669,-0.0577517897,-0.3814416826,0.3199534714,0.0561904982,0.2504869401,0.0391977765,-0.0429419242,0.1452626884,0.198122099,-0.0339066684,-0.0243588984,0.0854877979,0.2385652214,-0.0663192645,-0.1132121161,0.2658363581,-0.0436588787,-0.1837543845,0.2556534111,0.066129759,-0.0826183781,-0.0395169929,-0.043281734,0.0955182016,0.1809807569,-0.0629983768,0.1403061599,0.1276887804,-0.2898816168,-0.054411184,-0.110742934,-0.2697612345,0.0812905729,0.2204324752,-0.0361849926,-0.0838951319,0.4614356756,0.4005638361,-0.2099898458,-0.1176223457,-0.3479477763,0.1872604489,-0.5638506413,0.0368575789,-0.0007268398,0.0062479069,-0.1586897224,0.4170023203,-0.1392202228,-0.0937972292,0.0625543371,-0.5232264996,-0.0387310758,0.1154574826,-0.010349893,0.032931786,-0.0487522334,0.000144313,-0.2369749844,0.1144792289,-0.3007834852,0.1038622558,-0.2077763975,0.1372835636,0.4296366274,-0.0340136513,-0.1030200869,-0.2039479315,0.1335907727,0.3360125124,-0.1127493456,-0.150005728,-0.0085570309,0.1377923191,0.0324160047,-0.0805118382,0.0895498246,-0.0501768626,-0.1057639048,-0.1934644133,0.2179796398,0.1474811435,0.0261413902,-0.0277224965,-0.4482190609,0.2940060794,-0.0006671938,0.2398097217,-0.1485024691,0.4302434027,-0.1185299233,0.1239034459,-0.2088452727,-0.0353734829,-0.1150486767,0.2561821043,-0.096136108,0.0655865893,0.436314702,-0.2451550215,0.3256570995,0.4871932566,-0.0835472047,0.3445193172,-0.409155339,0.0281194597,0.0262620784,0.2102447897,-0.1836768985,-0.0181143712,-0.4960899055,-0.0479394235,-0.0900042951,0.2297486365,0.1105230153,-0.3707810938,-0.2305627912,0.0906378776,0.4767543972,-0.0063164178,0.2191631645,0.0231047086,0.0493144915,-0.0192854162,-0.0702998638,0.2188107371,0.0461868681,0.2720300555,-0.4646321833,0.3093931973,0.2494731247,0.1801015735,-0.0085307807,-0.4411627054,0.048236642,0.2103961706,0.0852203667,0.4002787471,0.1609909683,0.1586003602,-0.1475254446,-0.1259341389,-0.1613467187,0.0410661325,-0.0583181717,0.0472529568,-0.3145262897,-0.2136671394,-0.1724198312,-0.1043776944,-0.1697824746,0.1193011403,0.0753186196,0.1556896269,-0.3476271927,-0.3463916481,0.137742281,-0.2096372694,0.244703114,-0.1253280193,0.5787795782,0.4382135272,0.0846361071,0.4258896708,0.3017472923,0.7615882754,0.305842042,0.0535050742,0.1927803457,-0.1810554266,-0.1250636578,0.1044058949,0.3984232247,0.0448813178,0.3953417838,0.0439625829,0.1863657087,-0.138005659,0.1443944722,0.2655777335,0.4306209385,-0.191459626,0.2662214637,-0.1739098132,-0.0572185442,-0.419878304,-0.0946548581,-0.3771162927,0.1363275051,0.2699846923,-0.0920646265,0.095140107,-0.0083605088,0.1025270373,0.0740036219,0.3087226152,0.4034377933,0.1357361525,0.0255365111,-0.2081612647,-0.4297647178,0.3247804344,-0.0908724591,-0.3741289973,-0.06660413,0.0824070871,0.1919348389,0.1155832782,0.1238217652,-0.0432212763,0.383332938,0.0597238839,-0.1724222451,-0.0146099431,0.2414140701,-0.0620328858,-0.3025852442,-0.3642935455,0.0325732566,-0.2273191065,0.0701559708,0.0482755043,-0.0399629399,-0.0922599137,-0.2646709979,0.3722178042,0.0610680655,0.2166769505,-0.3107425272,-0.2267269194,-0.366376698,-0.1446740031,-0.15449965,0.4856240153,0.1616640836,0.2741329968,-0.0787693486,0.0253111199,-0.2682861984,0.1879413426,-0.1179688945,0.2693620622,-0.3983275592,-0.0619250089,-0.130013749,0.063481167,-0.0002447377,0.2416365594,0.013725874,0.0728774592,-0.1267946959,-0.6861604452,0.4283364117,-0.5565597415,-0.4156657457,0.0878274739,0.6013883948,0.0051166438,0.0725319907,-0.642991662,0.2427362651,0.0948262587,-0.2015708536,-0.1887482554,-0.0072774659,-0.1239597574,-0.1315533072,-0.0403240211,0.0875855461,0.2611950338,-0.1649687439,0.1641228795,-0.408061862]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2679","title":"Cannot load the blog_authorship_corpus due to codec errors","comments":"@albertvillanova \r\nCan you shed light on how this fix works?\r\n\r\nWe're experiencing a similar issue. \r\n\r\nIf we run several runs (eg in a Wandb sweep) the first run \"works\" but then we get `NonMatchingSplitsSizesError`\r\n\r\n| run num | actual train examples # | expected example # | recorded example # |\r\n| ------- | -------------- | ----------------- | -------- |\r\n| 1 | 100 | 100 | 100 |\r\n| 2 | 102 | 100 | 102 |\r\n| 3 | 100 | 100 | 202 | \r\n| 4 | 40 | 100 | 40 |\r\n| 5 | 40 | 100 | 40 |\r\n| 6 | 40 | 100 | 40 | \r\n\r\n\r\nThe second through the nth all crash with \r\n\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=19980970, num_examples=100, dataset_name='cies'), 'recorded': SplitInfo(name='train', num_bytes=40163811, num_examples=202, dataset_name='cies')}]\r\n\r\n```","body":"## Describe the bug\r\nA codec error is raised while loading the blog_authorship_corpus. \r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nraw_datasets = load_dataset(\"blog_authorship_corpus\")\r\n```\r\n\r\n\r\n## Expected results\r\nLoading the dataset without errors.\r\n\r\n## Actual results\r\nAn error similar to the one below was raised for (what seems like) every XML file.\r\n\/home\/izaskr\/.cache\/huggingface\/datasets\/downloads\/extracted\/7cf52524f6517e168604b41c6719292e8f97abbe8f731e638b13423f4212359a\/blogs\/788358.male.24.Arts.Libra.xml cannot be loaded. Error message: 'utf-8' codec can't decode byte 0xe7 in position 7551: invalid continuation byte\r\n\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 856, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 583, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 671, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.15.0-132-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyArrow version: 4.0.1\r\n\r\n","comment_length":135,"text":"Cannot load the blog_authorship_corpus due to codec errors\n## Describe the bug\r\nA codec error is raised while loading the blog_authorship_corpus. \r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nraw_datasets = load_dataset(\"blog_authorship_corpus\")\r\n```\r\n\r\n\r\n## Expected results\r\nLoading the dataset without errors.\r\n\r\n## Actual results\r\nAn error similar to the one below was raised for (what seems like) every XML file.\r\n\/home\/izaskr\/.cache\/huggingface\/datasets\/downloads\/extracted\/7cf52524f6517e168604b41c6719292e8f97abbe8f731e638b13423f4212359a\/blogs\/788358.male.24.Arts.Libra.xml cannot be loaded. Error message: 'utf-8' codec can't decode byte 0xe7 in position 7551: invalid continuation byte\r\n\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 856, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 583, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 671, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/home\/izaskr\/anaconda3\/envs\/local_vae_older\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 74, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.15.0-132-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyArrow version: 4.0.1\r\n\r\n\n@albertvillanova \r\nCan you shed light on how this fix works?\r\n\r\nWe're experiencing a similar issue. \r\n\r\nIf we run several runs (eg in a Wandb sweep) the first run \"works\" but then we get `NonMatchingSplitsSizesError`\r\n\r\n| run num | actual train examples # | expected example # | recorded example # |\r\n| ------- | -------------- | ----------------- | -------- |\r\n| 1 | 100 | 100 | 100 |\r\n| 2 | 102 | 100 | 102 |\r\n| 3 | 100 | 100 | 202 | \r\n| 4 | 40 | 100 | 40 |\r\n| 5 | 40 | 100 | 40 |\r\n| 6 | 40 | 100 | 40 | \r\n\r\n\r\nThe second through the nth all crash with \r\n\r\n```\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=19980970, num_examples=100, dataset_name='cies'), 'recorded': SplitInfo(name='train', num_bytes=40163811, num_examples=202, dataset_name='cies')}]\r\n\r\n```","embeddings":[-0.2084893584,0.4780819416,-0.0540075116,0.3998489976,0.3617135882,0.2268402874,0.0785168484,0.4288438559,-0.1586735994,0.2435750514,0.0468282923,0.2679276764,-0.0666647777,-0.210872665,0.0490831248,0.0175640285,-0.0882997364,0.1426831633,0.2282305956,-0.2342811376,-0.2054748684,0.1818809807,-0.1930260658,0.0730281472,-0.2108267397,-0.0789327174,0.1461101174,0.6233423352,-0.1766088754,-0.6049428582,0.1305889636,0.0211400893,0.1271802336,0.3438818455,-0.0001129081,-0.0840206072,0.5609151125,-0.0908816382,-0.2077678293,-0.2428600639,-0.2933349609,0.0893766582,-0.1876998544,-0.2272577435,0.0519394279,0.1707786471,0.043988727,-0.5675269365,0.3152764142,0.5750642419,0.2217924744,-0.0919267312,0.0331094116,-0.0800493881,0.3778606355,0.045461297,0.1114239097,0.4461227059,0.4049824476,-0.106069237,-0.0645619035,0.3080867231,-0.3914480805,-0.0931516141,0.0810792297,-0.1104288027,0.2468281239,-0.3147915006,0.6316772699,0.1682745665,0.7650758624,-0.2163730562,-0.081857644,0.0109544322,0.1951058507,-0.273940444,0.1889482737,0.3846224546,-0.3076028824,0.0981161445,-0.0437706299,-0.0178626496,-0.2440257967,-0.0302386694,-0.1594134271,0.0686632022,-0.0978547037,0.0309896469,-0.2485807091,0.0831889808,0.3431307375,-0.087953411,-0.2779461741,0.3806451559,0.0356344506,0.1754511893,0.1777861267,0.0299991332,-0.0623696409,0.1734724194,-0.0929429084,0.1898393929,0.1581863463,0.2450632006,0.0953593701,-0.052051533,0.1189913675,0.3499027193,0.2957804501,-0.0092535857,-0.0625836328,-0.0884831399,-0.3842275739,-0.385307163,-0.113360554,0.2477751374,0.2131842673,-0.3699801266,-0.4927609563,0.3212743998,-0.0740120709,0.0194718782,-0.0607549623,0.485186547,0.1276381463,0.0470056459,0.2408136576,0.2430116534,-0.1895996183,-0.0786771104,-0.1897735894,0.1530003697,-0.0485872775,-0.0438676737,0.1317065358,-0.5761977434,0.1176543161,-0.0184416268,0.3553464711,-0.320558995,-0.3360753059,-0.0648012385,-0.1094540656,0.0805431306,-0.0559896976,0.0071818228,0.1801805049,-0.3210016191,-0.0618890561,-0.0260087419,-0.1589974463,-0.3241128623,-0.0479485355,0.2177026421,0.0325050503,0.1240164414,-0.2615481317,0.098749727,0.4569669068,-0.1838753968,0.0121197803,-0.0366272479,-0.1992596835,-0.1116067916,0.0685284734,0.7190669775,0.0927176028,0.0617560372,0.4470205903,0.220451206,0.2396672964,-0.0346576385,-0.0303007979,0.194758296,-0.2161441147,0.1940477937,0.173850432,-0.6035790443,-0.0017290334,0.1910260767,0.0441783108,0.1295305341,0.0547363125,-0.3995582759,0.4223296642,-0.3531521559,0.1809348166,0.4003559351,-0.0078087784,0.0422336236,-0.2096998394,-0.0094913309,-0.0592940226,0.1407908052,0.0956548601,0.0371472165,0.4447281957,0.0195302796,0.5081382394,-0.2271026224,-0.0435088165,0.2044648826,-0.0637107193,0.0057082479,-0.0157999657,-0.2292633206,-0.1008971632,0.2268027067,-0.0385448001,-0.0186795779,-0.3047531843,-0.1203672215,-0.3363896608,0.0941970125,-0.1724899709,-0.0096021341,0.1468279064,0.0825056061,0.1227488443,0.211104691,-0.2485478967,0.2506260276,-0.3510062695,0.1294386089,-0.5143997073,0.2901625633,0.0366383158,-0.1802812517,0.3910512626,0.5273911357,0.2063739449,-0.1315859556,-0.2450412661,0.3295736909,-0.2906770408,0.6357571483,-0.3684304059,-0.2102894187,0.1556368917,-0.3921348453,0.1912713796,0.1315571219,0.2919704914,-0.1065121442,0.2268299609,0.1296588629,0.0019443497,0.0616183653,0.1570881456,-0.0124361692,0.3217045665,-0.0014806424,-0.1980579346,-0.306335181,0.6450024843,0.2185741812,0.1565416306,-0.105205372,-0.2416180521,0.0593610816,0.4993925989,-0.0996020436,-0.0267609991,0.035702005,-0.0107226865,0.0234232377,0.0127148628,-0.1461623758,0.1471957862,0.0217034426,-0.1039544493,0.0323545709,0.0716076419,-0.3352051079,0.3919107318,-0.1687837392,0.0142022185,0.4510241747,0.0282918811,0.0123838438,-0.3308029473,-0.1881506145,0.0471752398,0.0889911354,-0.5186877847,0.0840452239,-0.2582147419,-0.2728693783,-0.2343662679,-0.3481079042,-0.0588170327,-0.1013338491,-0.061281193,0.1418228,-0.0765725076,0.0614699721,-0.3226076365,0.1384917498,-0.1622715741,0.1579311192,-0.1613741368,-0.1625829339,-0.1952576339,-0.0374632478,0.4106931388,0.2090358436,0.145612061,-0.3906173706,0.0756685361,-0.1550291479,-0.1967244893,0.1922615021,-0.0072839498,0.0445013195,-0.096276179,-0.1030256748,0.6314195991,-0.1873525977,0.0774794072,0.0403758325,-0.00959244,0.4120475352,0.1020493954,-0.4308469594,-0.0713257641,-0.239278838,-0.3840438426,-0.5113678575,-0.2183594406,0.0891492143,0.2631490529,0.4002134502,-0.2125230432,-0.0471867919,-0.0296742711,0.3316637278,-0.1299691945,-0.2295024246,0.203759864,-0.2909738719,-0.3221518993,0.1063784137,0.0682759359,0.3978834748,-0.0266790073,-0.2990857959,0.200989604,0.0736920312,0.0807962045,-0.1829186678,-0.0326646566,0.1681961864,0.0659452304,0.0074228281,-0.076543875,0.0337187611,-0.3264494538,-0.0253692903,-0.1287384629,-0.0559628531,0.2026890516,-0.1725133657,0.061919529,0.4852911234,0.0982005894,0.2867949903,0.0515121557,0.21362333,0.0147207957,-0.2460832298,0.4707962275,-0.1930654645,-0.224330157,0.3553866446,0.0529203303,-0.1348503232,-0.2152939588,0.0723881498,-0.1209119037,-0.2952137887,0.0524808206,-0.3898859024,0.2485878766,0.0668382794,0.3902979493,0.0765489116,-0.2069254071,-0.0084624765,0.3520234525,0.1568837911,-0.125199005,0.014238039,0.1312190145,-0.4801469743,0.344866395,0.2102483958,-0.0640166476,-0.0432956107,0.2608812153,0.0587284379,-0.1514606029,0.0913045034,-0.2261163145,0.267290622,0.2079252452,0.3854580224,-0.2534746528,-0.0441583097,-0.0180206317,0.4818536937,-0.0801013783,-0.0459683985,-0.0982749909,0.1211590767,0.0731569082,0.1224754825,-0.0888519958,-0.1877179295,-0.2884795368,-0.3591709733,-0.3103108108,-0.2335228026,0.0687817186,0.3102971613,-0.1312476397,-0.1198605895,-0.0953507051,-0.0344561301,0.1942464113,0.0822998956,0.0921882316,-0.075224705,0.1156871542,-0.0131787956,-0.1346069872,0.3768363297,0.7526932359,-0.1334387064,-0.5529475808,-0.1159119308,-0.3700479567,0.0784900784,0.4450315237,-0.2124755681,-0.1339324266,0.0593456626,-0.0505468845,-0.3437484801,-0.2992713153,0.0504755303,0.0993438736,-0.2784160078,-0.5206426978,0.4623544812,-0.1271226257,0.0311399177,0.1468494684,0.0418167189,-0.3473612964,0.1372857243,0.0064516137,0.9900810122,0.214081645,-0.1357927918,0.075341925,-0.2174910456,0.0528732426,-0.2248955965,0.1980188191,-0.2843383253,0.0222397074,-0.151835233,0.2045907527,0.2688502371,0.062127769,-0.2158529311,0.1995509863,-0.0714000538,-0.2375721633,-0.0485025756,0.0031614071,-0.4199114144,-0.2580136359,-0.5787045956,0.1172822863,-0.1179324016,0.1192467287,-0.118909888,0.0359294116,0.0455081798,-0.3266947865,-0.4608321786,0.3758393228,-0.2024473846,-0.1504061669,-0.0577517897,-0.3814416826,0.3199534714,0.0561904982,0.2504869401,0.0391977765,-0.0429419242,0.1452626884,0.198122099,-0.0339066684,-0.0243588984,0.0854877979,0.2385652214,-0.0663192645,-0.1132121161,0.2658363581,-0.0436588787,-0.1837543845,0.2556534111,0.066129759,-0.0826183781,-0.0395169929,-0.043281734,0.0955182016,0.1809807569,-0.0629983768,0.1403061599,0.1276887804,-0.2898816168,-0.054411184,-0.110742934,-0.2697612345,0.0812905729,0.2204324752,-0.0361849926,-0.0838951319,0.4614356756,0.4005638361,-0.2099898458,-0.1176223457,-0.3479477763,0.1872604489,-0.5638506413,0.0368575789,-0.0007268398,0.0062479069,-0.1586897224,0.4170023203,-0.1392202228,-0.0937972292,0.0625543371,-0.5232264996,-0.0387310758,0.1154574826,-0.010349893,0.032931786,-0.0487522334,0.000144313,-0.2369749844,0.1144792289,-0.3007834852,0.1038622558,-0.2077763975,0.1372835636,0.4296366274,-0.0340136513,-0.1030200869,-0.2039479315,0.1335907727,0.3360125124,-0.1127493456,-0.150005728,-0.0085570309,0.1377923191,0.0324160047,-0.0805118382,0.0895498246,-0.0501768626,-0.1057639048,-0.1934644133,0.2179796398,0.1474811435,0.0261413902,-0.0277224965,-0.4482190609,0.2940060794,-0.0006671938,0.2398097217,-0.1485024691,0.4302434027,-0.1185299233,0.1239034459,-0.2088452727,-0.0353734829,-0.1150486767,0.2561821043,-0.096136108,0.0655865893,0.436314702,-0.2451550215,0.3256570995,0.4871932566,-0.0835472047,0.3445193172,-0.409155339,0.0281194597,0.0262620784,0.2102447897,-0.1836768985,-0.0181143712,-0.4960899055,-0.0479394235,-0.0900042951,0.2297486365,0.1105230153,-0.3707810938,-0.2305627912,0.0906378776,0.4767543972,-0.0063164178,0.2191631645,0.0231047086,0.0493144915,-0.0192854162,-0.0702998638,0.2188107371,0.0461868681,0.2720300555,-0.4646321833,0.3093931973,0.2494731247,0.1801015735,-0.0085307807,-0.4411627054,0.048236642,0.2103961706,0.0852203667,0.4002787471,0.1609909683,0.1586003602,-0.1475254446,-0.1259341389,-0.1613467187,0.0410661325,-0.0583181717,0.0472529568,-0.3145262897,-0.2136671394,-0.1724198312,-0.1043776944,-0.1697824746,0.1193011403,0.0753186196,0.1556896269,-0.3476271927,-0.3463916481,0.137742281,-0.2096372694,0.244703114,-0.1253280193,0.5787795782,0.4382135272,0.0846361071,0.4258896708,0.3017472923,0.7615882754,0.305842042,0.0535050742,0.1927803457,-0.1810554266,-0.1250636578,0.1044058949,0.3984232247,0.0448813178,0.3953417838,0.0439625829,0.1863657087,-0.138005659,0.1443944722,0.2655777335,0.4306209385,-0.191459626,0.2662214637,-0.1739098132,-0.0572185442,-0.419878304,-0.0946548581,-0.3771162927,0.1363275051,0.2699846923,-0.0920646265,0.095140107,-0.0083605088,0.1025270373,0.0740036219,0.3087226152,0.4034377933,0.1357361525,0.0255365111,-0.2081612647,-0.4297647178,0.3247804344,-0.0908724591,-0.3741289973,-0.06660413,0.0824070871,0.1919348389,0.1155832782,0.1238217652,-0.0432212763,0.383332938,0.0597238839,-0.1724222451,-0.0146099431,0.2414140701,-0.0620328858,-0.3025852442,-0.3642935455,0.0325732566,-0.2273191065,0.0701559708,0.0482755043,-0.0399629399,-0.0922599137,-0.2646709979,0.3722178042,0.0610680655,0.2166769505,-0.3107425272,-0.2267269194,-0.366376698,-0.1446740031,-0.15449965,0.4856240153,0.1616640836,0.2741329968,-0.0787693486,0.0253111199,-0.2682861984,0.1879413426,-0.1179688945,0.2693620622,-0.3983275592,-0.0619250089,-0.130013749,0.063481167,-0.0002447377,0.2416365594,0.013725874,0.0728774592,-0.1267946959,-0.6861604452,0.4283364117,-0.5565597415,-0.4156657457,0.0878274739,0.6013883948,0.0051166438,0.0725319907,-0.642991662,0.2427362651,0.0948262587,-0.2015708536,-0.1887482554,-0.0072774659,-0.1239597574,-0.1315533072,-0.0403240211,0.0875855461,0.2611950338,-0.1649687439,0.1641228795,-0.408061862]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2678","title":"Import Error in Kaggle notebook","comments":"This looks like an issue with PyArrow. Did you try reinstalling it ?","body":"## Describe the bug\r\nNot able to import datasets library in kaggle notebooks\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n!pip install datasets\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nNo such error\r\n\r\n## Actual results\r\n```\r\nImportError Traceback (most recent call last)\r\n in \r\n----> 1 import datasets\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/__init__.py in \r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in \r\n 36 import pandas as pd\r\n 37 import pyarrow as pa\r\n---> 38 import pyarrow.compute as pc\r\n 39 from multiprocess import Pool, RLock\r\n 40 from tqdm.auto import tqdm\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in \r\n 16 # under the License.\r\n 17 \r\n---> 18 from pyarrow._compute import ( # noqa\r\n 19 Function,\r\n 20 FunctionOptions,\r\n\r\nImportError: \/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Kaggle\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":13,"text":"Import Error in Kaggle notebook\n## Describe the bug\r\nNot able to import datasets library in kaggle notebooks\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n!pip install datasets\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nNo such error\r\n\r\n## Actual results\r\n```\r\nImportError Traceback (most recent call last)\r\n in \r\n----> 1 import datasets\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/__init__.py in \r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in \r\n 36 import pandas as pd\r\n 37 import pyarrow as pa\r\n---> 38 import pyarrow.compute as pc\r\n 39 from multiprocess import Pool, RLock\r\n 40 from tqdm.auto import tqdm\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in \r\n 16 # under the License.\r\n 17 \r\n---> 18 from pyarrow._compute import ( # noqa\r\n 19 Function,\r\n 20 FunctionOptions,\r\n\r\nImportError: \/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Kaggle\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n\nThis looks like an issue with PyArrow. Did you try reinstalling it ?","embeddings":[-0.3311520219,-0.0346449129,-0.1113157496,0.1136168614,0.2160874158,-0.0077066058,0.3467618227,0.3172732592,-0.0728922784,-0.0299229212,-0.1647050381,0.8303610682,0.0918386281,0.2034669518,0.0633266047,-0.0786100924,0.1253454238,0.2315590531,-0.0554434955,0.0687854514,-0.0982853994,0.041001223,-0.2998902202,0.2606538236,-0.0589990541,-0.3248463869,0.1387410313,-0.0481898934,-0.2148354352,-0.3897067904,0.3666331768,-0.2054637969,0.1483573169,0.3310925663,-0.0001053943,0.0461394116,0.3383732736,0.0213731006,-0.3294006884,-0.1116867289,-0.1150480658,-0.4699164331,0.1610479802,-0.2939140201,0.439299047,-0.3512535393,-0.2151754647,-0.0582361817,0.2059800625,0.1730158776,0.2564456761,0.4804010987,0.4850061834,0.0228538681,0.4336473942,-0.1355711818,-0.1660907418,0.4058332145,0.2513929605,-0.2714586854,0.0112680187,0.0794472769,-0.081440039,0.0862456486,0.1276094019,0.0741997808,-0.1207379997,-0.349201411,-0.0577981696,0.0286440179,0.4813690484,-0.5208933949,-0.1776010543,-0.0990070999,0.1325976253,-0.0519051552,0.1250441074,0.2178072929,-0.0520337485,-0.1047107205,0.2848064303,-0.1860972345,-0.2418589443,0.2095439732,-0.1712005734,0.4520232379,-0.0350294188,0.1072933972,0.2992790639,-0.131762892,0.5775826573,-0.0420558378,-0.1079123765,0.1141147986,-0.3564469814,0.1612535715,0.0754735246,-0.0356080681,0.1269459575,0.0571997985,0.0884838104,0.1717464626,0.1675661802,0.3098486066,0.1394140124,0.3081650734,-0.026280392,0.240899846,0.2992525399,0.2408293933,0.1383658797,-0.0559013151,-0.0657399073,-0.3916476965,0.1576513499,-0.105054833,0.3078275919,0.1841371208,-0.3587899208,-0.0575626455,-0.3739744425,-0.106143035,-0.0526914038,0.1579804718,-0.2192387581,0.0793302953,0.1513368338,0.1610507667,-0.1558080018,-0.062730439,-0.156455785,0.262708962,-0.0840767696,-0.2171460241,0.0356447808,-0.3135081828,0.1715532094,0.0884471536,0.1957621425,0.1463638991,0.193526119,-0.114275828,-0.2211187184,0.3127185702,0.0758652762,0.0817181021,0.0919212103,-0.3852344751,-0.1201685518,0.1241520941,-0.2185050398,-0.1260708869,-0.2826134562,0.2253304571,-0.2195148319,0.0472952239,-0.5390341878,-0.2033290267,-0.0355586819,-0.1069118828,-0.0238412004,-0.2577725351,0.251953423,-0.3059752285,0.2445448041,0.0206875745,-0.3586955965,0.0174879096,-0.2140699625,-0.216658622,-0.0499044843,0.1814651042,-0.317650944,0.2158074975,-0.2199999839,0.2172692567,0.6034314036,-0.4509723783,-0.501324296,-0.0146605149,-0.111095421,-0.0385056585,-0.0815515444,0.0978181437,-0.1336588711,0.2454488277,0.0614740141,0.1481057405,-0.0120802261,-0.1089448184,-0.2383634597,-0.2573354244,0.154750064,0.1797798574,0.0888211876,-0.3220164776,0.1798798442,-0.1596307904,0.0999965593,0.0135519747,0.0888313577,0.0106155565,0.3654177189,0.0573742576,0.0431770422,-0.1443686783,-0.2541452348,0.2076940984,-0.2042830884,0.0554628521,-0.3427084684,-0.0785161778,-0.332668066,0.2844836712,-0.2056761682,0.0662623793,0.1995526105,-0.1054867133,-0.006850048,0.3752055466,-0.0380990282,0.223977983,0.1703836918,0.1347755641,-0.2951853871,0.370777756,-0.3771543503,-0.2649771273,0.1128828153,0.0975651145,-0.0902376026,0.039896559,-0.0091691371,0.1615402848,0.0327153839,0.062183097,-0.0280498546,-0.1954877228,0.1557335109,-0.4090655148,0.1016246527,0.1074901372,0.2792437673,0.084936671,0.1338961124,0.1193652675,-0.0569445305,-0.1091520488,-0.0003105028,0.1789021045,0.119720839,0.1447720826,0.1295149624,-0.0427674502,0.1811077148,-0.0599599704,0.1227110699,-0.0826661289,-0.1746010184,-0.0156738535,0.4049962759,0.0170012973,0.2242792994,0.0410494022,-0.4309890866,0.1457851529,0.2109175473,-0.0895459726,0.3606088161,0.2290539294,-0.2024454176,-0.0511369817,-0.0836352333,-0.1683250368,0.1141008213,0.2657561302,0.3559385836,0.1905687004,0.3204768896,-0.0573273264,-0.117493175,-0.1583665311,-0.0629875436,0.2429005355,-0.1523339599,0.2514917552,-0.2236230224,-0.2371555567,-0.1255450398,-0.2121432424,-0.0406207517,-0.4249206483,-0.1771907061,0.2521310449,0.1457414478,0.2258049548,0.2357546687,-0.1407993436,0.0676147342,-0.2078075558,-0.2566402853,-0.42281124,-0.3970933259,0.1325372458,0.3020269573,0.1665483415,0.2187388092,-0.1542665213,0.2036151737,-0.4213411212,-0.4639923871,-0.1860967427,-0.0976047739,0.3936481476,0.2713234425,0.1565301418,-0.2673632503,-0.3296768963,0.4015653729,0.0559328049,-0.2371349484,0.1609768569,-0.1273807734,-0.1603078842,0.0265629608,-0.3340500593,-0.3674561977,-0.38889575,0.3022526205,0.1425952166,-0.1501105428,-0.1753434539,0.307924062,0.2518005073,0.0867730752,-0.0189682655,0.0362010188,0.2769465148,0.5777769685,-0.0785522982,-0.4402299523,0.1778320074,-0.2454897612,0.2321709991,-0.0807304904,-0.4525515735,0.0137244817,-0.1428705007,0.6574349999,-0.1949453801,0.2248612195,0.1461232901,0.19440718,0.0050085303,-0.0748384297,-0.2804314494,-0.1694361418,-0.0543850437,0.0174455885,0.0070671826,0.083417654,-0.0012343155,0.6059474349,-0.1022099927,-0.0920615494,0.3021646142,-0.1603866071,0.192168057,0.0674042478,-0.3959595561,-0.0129337115,-0.1948434263,0.2733744979,-0.1103689373,-0.2912972271,-0.3038033247,-0.2947912812,-0.2213208675,-0.1946087778,-0.2214640975,0.0284014568,-0.0217632521,0.0578353703,0.0610600822,-0.0809577256,-0.1692267805,-0.1915280968,0.2128890306,-0.0491749682,-0.2417472303,0.0273728706,-0.3530057967,-0.4973076284,-0.1512071788,0.2985376716,0.1216965616,0.1690391898,0.129118517,-0.2048423141,0.1314182729,-0.2109707147,0.4666034877,0.0475346521,-0.1730989665,0.2648140192,0.2524162233,-0.7619399428,-0.2930575907,-0.247046262,0.0344015695,0.1267065108,0.0354574472,-0.2741528153,0.2073599547,0.0339644849,0.4462114871,-0.0417002253,0.0460819118,-0.4358468354,-0.3943654597,-0.3359904885,-0.1116414741,-0.0931311697,0.3062421978,-0.0162840877,0.0206967313,-0.1978460401,-0.0775626674,-0.182386294,0.2842848897,0.2198854238,-0.1273007244,0.2286867648,-0.1416841,-0.0619712621,0.5114405751,0.3834291697,0.1248319522,-0.6873646379,-0.1233930886,-0.0456460305,0.1124688983,-0.1834222972,-0.1461042017,0.2096929699,0.0150978705,0.065639317,0.0241867676,0.0618025661,0.2626633346,0.0865573883,-0.0501556545,-0.1200991049,0.489721477,0.2139827311,0.031509243,0.3634704649,0.0289658587,-0.1240202114,0.1110695824,0.1447690129,0.6335715652,-0.0350809693,0.0789721757,0.6067389846,-0.3013091683,0.5186696649,-0.1109880358,0.0873192623,-0.480286181,-0.2186362147,0.0214732997,-0.3365723491,0.2853909135,-0.321480155,-0.2422941774,0.0244872998,-0.2415529341,0.2196568996,0.0812886283,0.1670666933,-0.2723799348,0.0136714354,-0.1436060071,0.1400235593,-0.1867625117,0.1916611642,-0.0730477199,-0.0536134653,-0.4096933007,-0.1535685956,-0.1789149195,0.0516518168,-0.286683768,0.4776212573,0.2005763948,-0.352234453,0.276370883,0.381475687,0.131839186,-0.0443640985,-0.0398015417,0.1345399171,0.1783033162,-0.2488502115,0.2056542188,0.0901162326,0.1777213067,-0.1697764844,-0.3576977551,0.5502688289,-0.0464630276,0.0084914621,-0.0911787972,0.5004709959,0.0921704248,-0.1599373072,-0.0630561933,-0.160787791,0.1178363338,-0.1444625854,0.1528141946,0.0997176543,0.0919491425,0.3907577395,-0.133942157,0.0157478042,-0.0570079535,0.2607944608,-0.0886433572,0.1141196936,0.2365716249,0.1417631209,-0.308662802,-0.184094578,0.1364180446,-0.0827045292,-0.3662291467,0.0282002166,0.1428482085,0.1320424378,-0.2003016919,0.3363578916,0.1547127366,-0.1210276335,0.0899886712,-0.3515769243,-0.2342706323,0.3379463255,0.0699090734,-0.1213811189,0.3330337405,0.0952664316,0.0492517278,-0.0013660407,-0.3063868582,0.2259493321,-0.2876387239,0.0289584547,0.0536580458,0.0481787212,-0.0149588622,-0.0650198832,0.1885053068,0.0682474449,-0.0734772086,-0.296530515,-0.0607020147,0.044362478,-0.0705435351,-0.1305528283,0.05146816,-0.2485475242,0.1101218313,-0.0127899433,-0.0247866735,0.0288005657,-0.150307253,0.0698307455,-0.0029570677,0.0247968547,-0.2099232227,0.2090702802,-0.2179846019,-0.0287424885,0.0190586876,0.1830332428,0.2522783875,0.1403273493,-0.0176488999,-0.1722995788,-0.2750736177,-0.0435595065,0.2539570928,-0.3827710152,-0.0337501615,-0.1890946776,0.0967165679,0.5022469163,-0.2665284574,0.0024989152,0.3398680389,0.233772397,-0.4564214349,-0.358672291,0.1599986404,-0.0536384694,-0.0029916666,0.0004135891,0.1760163307,-0.0249031559,-0.0723086521,0.2437831759,0.1339891702,-0.1173607185,0.221595034,0.4241959155,-0.0721441582,-0.0086833881,0.2807653248,0.2971855402,0.1682372391,0.3244005442,-0.3154556751,0.0702838972,-0.077317439,0.0394295342,-0.0290027484,-0.2539077401,-0.1211656854,0.0325723067,-0.0957030654,-0.2007051408,0.1203617379,0.5001189709,0.1547101438,0.0383893363,0.0909321308,0.0775668472,-0.1244110093,0.2770017385,-0.0674759895,-0.12542364,0.0961906388,-0.0053253025,0.0825082511,-0.2226887345,0.445645988,0.2423892468,-0.2123492509,-0.260347873,0.1832074523,0.3811062574,0.0669498146,-0.1098572165,-0.1104389578,0.2248701304,-0.21776402,-0.3534426391,0.2441451699,0.5869365335,0.2603342831,-0.3102401197,0.0944029316,0.0669887513,-0.1572199911,-0.0314721949,0.0495271794,0.1830658466,0.1584377885,0.2474005818,0.2550219297,-0.210365966,-0.0762049407,0.1017942876,0.1866766512,-0.1842792183,0.0747471228,-0.2892583311,-0.2940638661,-0.0496010259,0.0403096192,-0.2266136706,0.4456878006,0.2843503952,0.2797303498,-0.0189623833,-0.3126815557,0.1155215055,0.0669810772,0.5089535713,0.2625835538,0.4421732128,-0.1770849973,-0.1440308839,-0.505704999,0.1786835045,-0.0854798779,-0.26680246,-0.0114587909,0.0403195098,-0.0444344357,0.2729148567,0.2355618626,0.0073267017,0.282795459,0.1371320784,-0.3872559667,-0.2417166084,0.2488875538,-0.1888399124,0.0287327934,-0.3225711286,0.1298228651,-0.4148222208,0.1735837758,-0.3441610336,-0.0993942544,0.2711322308,0.1419683397,0.6289787889,-0.018064443,0.5183414221,0.1692893058,0.1518971324,0.0126480022,-0.2340146154,-0.1018498614,0.2416532487,0.1397093683,0.1854530573,-0.2032283396,-0.1038618982,-0.2030742317,0.4040062129,0.2426775992,0.1769190133,0.0183933768,0.0665852726,0.07225658,-0.0664942637,0.0682139993,0.0553854853,-0.141695708,0.1172400564,-0.3867329955,-0.4122762382,0.2657864392,-0.2571204901,-0.3173795938,-0.0369095057,0.0078407209,0.0998368785,-0.0957258344,-0.3347949088,0.0125862304,0.4346929491,-0.0102515481,-0.2399895787,0.2506608069,0.1179791167,0.1057687327,-0.072497189,0.366053164,-0.0585739277,-0.1416140497,0.0302562416,-0.1465342194]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2678","title":"Import Error in Kaggle notebook","comments":"@lhoestq I did, and then let pip handle the installation in `pip import datasets`. I also tried using conda but it gives the same error.\r\n\r\nEdit: pyarrow version on kaggle is 4.0.0, it gets replaced with 4.0.1. So, I don't think uninstalling will change anything.\r\n```\r\nInstall Trace of datasets:\r\n\r\nCollecting datasets\r\n Downloading datasets-1.9.0-py3-none-any.whl (262 kB)\r\n |\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 262 kB 834 kB\/s eta 0:00:01\r\nRequirement already satisfied: dill in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (0.3.4)\r\nCollecting pyarrow!=4.0.0,>=1.0.0\r\n Downloading pyarrow-4.0.1-cp37-cp37m-manylinux2014_x86_64.whl (21.8 MB)\r\n |\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 21.8 MB 6.2 MB\/s eta 0:00:01\r\nRequirement already satisfied: importlib-metadata in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (3.4.0)\r\nRequirement already satisfied: huggingface-hub<0.1.0 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (0.0.8)\r\nRequirement already satisfied: pandas in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (1.2.4)\r\nRequirement already satisfied: requests>=2.19.0 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (2.25.1)\r\nRequirement already satisfied: fsspec>=2021.05.0 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (2021.6.1)\r\nRequirement already satisfied: multiprocess in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (0.70.12.2)\r\nRequirement already satisfied: packaging in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (20.9)\r\nCollecting xxhash\r\n Downloading xxhash-2.0.2-cp37-cp37m-manylinux2010_x86_64.whl (243 kB)\r\n |\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 243 kB 23.7 MB\/s eta 0:00:01\r\nRequirement already satisfied: numpy>=1.17 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (1.19.5)\r\nRequirement already satisfied: tqdm>=4.27 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (4.61.1)\r\nRequirement already satisfied: filelock in \/opt\/conda\/lib\/python3.7\/site-packages (from huggingface-hub<0.1.0->datasets) (3.0.12)\r\nRequirement already satisfied: urllib3<1.27,>=1.21.1 in \/opt\/conda\/lib\/python3.7\/site-packages (from requests>=2.19.0->datasets) (1.26.5)\r\nRequirement already satisfied: idna<3,>=2.5 in \/opt\/conda\/lib\/python3.7\/site-packages (from requests>=2.19.0->datasets) (2.10)\r\nRequirement already satisfied: certifi>=2017.4.17 in \/opt\/conda\/lib\/python3.7\/site-packages (from requests>=2.19.0->datasets) (2021.5.30)\r\nRequirement already satisfied: chardet<5,>=3.0.2 in \/opt\/conda\/lib\/python3.7\/site-packages (from requests>=2.19.0->datasets) (4.0.0)\r\nRequirement already satisfied: typing-extensions>=3.6.4 in \/opt\/conda\/lib\/python3.7\/site-packages (from importlib-metadata->datasets) (3.7.4.3)\r\nRequirement already satisfied: zipp>=0.5 in \/opt\/conda\/lib\/python3.7\/site-packages (from importlib-metadata->datasets) (3.4.1)\r\nRequirement already satisfied: pyparsing>=2.0.2 in \/opt\/conda\/lib\/python3.7\/site-packages (from packaging->datasets) (2.4.7)\r\nRequirement already satisfied: python-dateutil>=2.7.3 in \/opt\/conda\/lib\/python3.7\/site-packages (from pandas->datasets) (2.8.1)\r\nRequirement already satisfied: pytz>=2017.3 in \/opt\/conda\/lib\/python3.7\/site-packages (from pandas->datasets) (2021.1)\r\nRequirement already satisfied: six>=1.5 in \/opt\/conda\/lib\/python3.7\/site-packages (from python-dateutil>=2.7.3->pandas->datasets) (1.15.0)\r\nInstalling collected packages: xxhash, pyarrow, datasets\r\n Attempting uninstall: pyarrow\r\n Found existing installation: pyarrow 4.0.0\r\n Uninstalling pyarrow-4.0.0:\r\n Successfully uninstalled pyarrow-4.0.0\r\nSuccessfully installed datasets-1.9.0 pyarrow-4.0.1 xxhash-2.0.2\r\nWARNING: Running pip as root will break packages and permissions. You should install packages reliably by using venv: https:\/\/pip.pypa.io\/warnings\/venv\r\n```","body":"## Describe the bug\r\nNot able to import datasets library in kaggle notebooks\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n!pip install datasets\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nNo such error\r\n\r\n## Actual results\r\n```\r\nImportError Traceback (most recent call last)\r\n in \r\n----> 1 import datasets\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/__init__.py in \r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in \r\n 36 import pandas as pd\r\n 37 import pyarrow as pa\r\n---> 38 import pyarrow.compute as pc\r\n 39 from multiprocess import Pool, RLock\r\n 40 from tqdm.auto import tqdm\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in \r\n 16 # under the License.\r\n 17 \r\n---> 18 from pyarrow._compute import ( # noqa\r\n 19 Function,\r\n 20 FunctionOptions,\r\n\r\nImportError: \/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Kaggle\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":322,"text":"Import Error in Kaggle notebook\n## Describe the bug\r\nNot able to import datasets library in kaggle notebooks\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n!pip install datasets\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nNo such error\r\n\r\n## Actual results\r\n```\r\nImportError Traceback (most recent call last)\r\n in \r\n----> 1 import datasets\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/__init__.py in \r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in \r\n 36 import pandas as pd\r\n 37 import pyarrow as pa\r\n---> 38 import pyarrow.compute as pc\r\n 39 from multiprocess import Pool, RLock\r\n 40 from tqdm.auto import tqdm\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in \r\n 16 # under the License.\r\n 17 \r\n---> 18 from pyarrow._compute import ( # noqa\r\n 19 Function,\r\n 20 FunctionOptions,\r\n\r\nImportError: \/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Kaggle\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n\n@lhoestq I did, and then let pip handle the installation in `pip import datasets`. I also tried using conda but it gives the same error.\r\n\r\nEdit: pyarrow version on kaggle is 4.0.0, it gets replaced with 4.0.1. So, I don't think uninstalling will change anything.\r\n```\r\nInstall Trace of datasets:\r\n\r\nCollecting datasets\r\n Downloading datasets-1.9.0-py3-none-any.whl (262 kB)\r\n |\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 262 kB 834 kB\/s eta 0:00:01\r\nRequirement already satisfied: dill in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (0.3.4)\r\nCollecting pyarrow!=4.0.0,>=1.0.0\r\n Downloading pyarrow-4.0.1-cp37-cp37m-manylinux2014_x86_64.whl (21.8 MB)\r\n |\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 21.8 MB 6.2 MB\/s eta 0:00:01\r\nRequirement already satisfied: importlib-metadata in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (3.4.0)\r\nRequirement already satisfied: huggingface-hub<0.1.0 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (0.0.8)\r\nRequirement already satisfied: pandas in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (1.2.4)\r\nRequirement already satisfied: requests>=2.19.0 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (2.25.1)\r\nRequirement already satisfied: fsspec>=2021.05.0 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (2021.6.1)\r\nRequirement already satisfied: multiprocess in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (0.70.12.2)\r\nRequirement already satisfied: packaging in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (20.9)\r\nCollecting xxhash\r\n Downloading xxhash-2.0.2-cp37-cp37m-manylinux2010_x86_64.whl (243 kB)\r\n |\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 243 kB 23.7 MB\/s eta 0:00:01\r\nRequirement already satisfied: numpy>=1.17 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (1.19.5)\r\nRequirement already satisfied: tqdm>=4.27 in \/opt\/conda\/lib\/python3.7\/site-packages (from datasets) (4.61.1)\r\nRequirement already satisfied: filelock in \/opt\/conda\/lib\/python3.7\/site-packages (from huggingface-hub<0.1.0->datasets) (3.0.12)\r\nRequirement already satisfied: urllib3<1.27,>=1.21.1 in \/opt\/conda\/lib\/python3.7\/site-packages (from requests>=2.19.0->datasets) (1.26.5)\r\nRequirement already satisfied: idna<3,>=2.5 in \/opt\/conda\/lib\/python3.7\/site-packages (from requests>=2.19.0->datasets) (2.10)\r\nRequirement already satisfied: certifi>=2017.4.17 in \/opt\/conda\/lib\/python3.7\/site-packages (from requests>=2.19.0->datasets) (2021.5.30)\r\nRequirement already satisfied: chardet<5,>=3.0.2 in \/opt\/conda\/lib\/python3.7\/site-packages (from requests>=2.19.0->datasets) (4.0.0)\r\nRequirement already satisfied: typing-extensions>=3.6.4 in \/opt\/conda\/lib\/python3.7\/site-packages (from importlib-metadata->datasets) (3.7.4.3)\r\nRequirement already satisfied: zipp>=0.5 in \/opt\/conda\/lib\/python3.7\/site-packages (from importlib-metadata->datasets) (3.4.1)\r\nRequirement already satisfied: pyparsing>=2.0.2 in \/opt\/conda\/lib\/python3.7\/site-packages (from packaging->datasets) (2.4.7)\r\nRequirement already satisfied: python-dateutil>=2.7.3 in \/opt\/conda\/lib\/python3.7\/site-packages (from pandas->datasets) (2.8.1)\r\nRequirement already satisfied: pytz>=2017.3 in \/opt\/conda\/lib\/python3.7\/site-packages (from pandas->datasets) (2021.1)\r\nRequirement already satisfied: six>=1.5 in \/opt\/conda\/lib\/python3.7\/site-packages (from python-dateutil>=2.7.3->pandas->datasets) (1.15.0)\r\nInstalling collected packages: xxhash, pyarrow, datasets\r\n Attempting uninstall: pyarrow\r\n Found existing installation: pyarrow 4.0.0\r\n Uninstalling pyarrow-4.0.0:\r\n Successfully uninstalled pyarrow-4.0.0\r\nSuccessfully installed datasets-1.9.0 pyarrow-4.0.1 xxhash-2.0.2\r\nWARNING: Running pip as root will break packages and permissions. You should install packages reliably by using venv: https:\/\/pip.pypa.io\/warnings\/venv\r\n```","embeddings":[-0.3157348335,-0.0831313878,-0.1109722927,0.1658830196,0.1794455051,-0.0220792536,0.3208149672,0.2708318233,-0.0284393821,-0.0525305681,-0.1436260194,0.85928303,0.0922810957,0.2378659844,0.0209103543,-0.0479097851,0.161305666,0.2256399542,-0.0434857272,0.0514687337,-0.0643853098,0.0771296248,-0.2960688174,0.2006766051,-0.059248656,-0.3657685816,0.1798947603,-0.014297124,-0.2524811327,-0.3473741114,0.4036728442,-0.1720401645,0.1640073508,0.3787457347,-0.0001087982,0.0669417232,0.3618468642,-0.0291197337,-0.3702855408,-0.1077230796,-0.2176077515,-0.6008007526,0.0832865685,-0.2847102284,0.4702086747,-0.2857757211,-0.2630595267,-0.1921920329,0.2056123465,0.1816900522,0.220649913,0.4774709642,0.4807039201,0.0092315031,0.2772545218,-0.1289513111,-0.2072615474,0.302177608,0.2401313484,-0.2112718225,0.1166253239,0.1265563369,-0.0676916912,0.0300699398,0.1025223359,0.0168172158,-0.1558161676,-0.3838818073,-0.0678934753,0.0581074692,0.4930505753,-0.5021786094,-0.2771857083,-0.1238675043,0.1140542701,0.0022397963,0.1400085837,0.1730056405,-0.0677028969,-0.0962271243,0.2432511002,-0.2378110439,-0.2651634812,0.1864016056,-0.1860895604,0.4596527219,-0.0485247113,0.0963278115,0.2849809527,-0.1344873905,0.6502164602,-0.0968685225,-0.1329707056,0.1146485433,-0.4117366672,0.0971322283,0.010409127,-0.0425877497,0.1165604293,0.1059470251,0.1066146642,0.1188913435,0.0963285193,0.3061642647,0.0781596452,0.3834520876,0.0288374443,0.3633151054,0.2639465928,0.2905315459,0.245966047,-0.0625741929,-0.0748769566,-0.3948988616,0.0809105411,-0.0593818463,0.2644909322,0.1566989869,-0.4245138466,-0.0065868539,-0.2908203006,-0.0526703186,0.0005867296,0.1260784417,-0.2189185619,0.1570490748,0.153926149,0.1813415289,-0.1178863421,-0.0593994446,-0.1201268286,0.3064025939,-0.0218912102,-0.2258309722,0.0449589416,-0.2936345041,0.2442829013,0.0245810207,0.2295258194,0.1750361025,0.1705325842,-0.1257321984,-0.1919405162,0.3506008089,0.0321080126,0.115392603,0.1277721226,-0.3931867778,-0.1246997267,0.0788087621,-0.1662564874,-0.0804766044,-0.2768648565,0.1809964627,-0.290689826,0.051424928,-0.5761383176,-0.2516323924,0.0038182673,-0.0501182638,-0.0043983608,-0.1965858638,0.234431088,-0.2544774711,0.2506346703,0.0534614958,-0.3594920933,0.0474446714,-0.2210427821,-0.2070468217,-0.0644442812,0.1289324909,-0.2450414598,0.1776161641,-0.2288636714,0.134849295,0.6730281711,-0.6061068773,-0.5256843567,0.0257641803,-0.0934768692,0.0148919905,-0.0130191706,0.0944576636,-0.1029906943,0.2157653421,0.0677377284,0.1622854322,-0.0314902738,-0.1910270452,-0.2592289746,-0.2871662676,0.0578858815,0.1806648374,0.0925267786,-0.2924076617,0.1847915649,-0.1539341211,0.0381951183,0.0600379854,0.1202748492,0.0565657131,0.3603731394,0.0527839586,-0.0124018872,-0.1248254329,-0.2917067111,0.268514514,-0.1398256272,0.0348941907,-0.406406641,-0.0832441077,-0.3611127138,0.3012945354,-0.217399776,0.0525478832,0.1697994173,-0.123736605,0.0105001181,0.3521030843,-0.0077052573,0.3821054101,0.1918429434,0.1550069898,-0.3192882836,0.3061139882,-0.3814661801,-0.1457504779,0.0615457892,0.1301296651,-0.1030459404,0.0251288638,-0.0443161391,0.1139586046,0.077213712,0.0239458177,0.010299528,-0.2309055328,0.1762903035,-0.3153473437,0.1356556416,0.1335422546,0.3049322069,0.0928746164,0.0961892903,0.095304653,-0.0592509471,-0.095798105,0.014867017,0.257581383,0.2067648321,0.2018836439,0.061224442,-0.1222234517,0.1843215227,-0.0238401052,0.1030386165,-0.0872515887,-0.1367312521,-0.023938071,0.360041827,0.0096609853,0.1436843425,0.0097316708,-0.3445076942,0.2186189443,0.2709667087,0.0152541101,0.3870508969,0.2118288875,-0.2134576887,-0.0735057443,-0.0510983989,-0.1095043346,0.1225212216,0.3149383664,0.391687423,0.1771174818,0.245856002,-0.0654154643,0.0114169847,-0.2596939802,-0.04270285,0.2147675902,-0.1877250969,0.2559342682,-0.2284218222,-0.2371326089,-0.162731111,-0.1992757618,0.0117198545,-0.4043591022,-0.2368007302,0.2449380606,0.1685368866,0.1849227846,0.1470604688,-0.1181396544,0.0749962181,-0.2854743898,-0.2432315946,-0.3484172225,-0.3759360015,0.1030787006,0.3510778248,0.1604800373,0.2290232033,-0.1745557785,0.2487835139,-0.4051178396,-0.4474343657,-0.210863471,-0.0932771042,0.4236423969,0.266956836,0.1613865495,-0.2377539724,-0.2812199891,0.3301022947,0.1585963517,-0.2695578933,0.1333747208,-0.1104923189,-0.2200868726,0.0303354114,-0.3240668774,-0.3991329074,-0.4153462648,0.3101412356,0.1853808314,-0.1316485703,-0.1519116759,0.3288983107,0.2101772875,0.1746779233,0.0226537231,0.0737200826,0.2688193619,0.5995613337,-0.1175564229,-0.4101842046,0.2229814231,-0.2326618135,0.3032893836,-0.0754950792,-0.4256069362,0.0147363963,-0.173968941,0.650064826,-0.2176861614,0.2231082469,0.1810691208,0.2036795765,0.0705053508,-0.1243862733,-0.2514392436,-0.1216403469,-0.1191993132,0.0573351607,0.0553836375,0.1176579595,0.0425810367,0.5509263277,0.0537305214,-0.0652405843,0.2583059371,-0.0818953067,0.2138422579,0.0764959231,-0.4019142687,-0.0495567694,-0.1408323646,0.2763573229,-0.1539872587,-0.2508073449,-0.2186053246,-0.3311209679,-0.1756257564,-0.1506525129,-0.1933946162,0.0008879229,-0.1569964439,0.0058023906,0.0806929842,-0.0736643448,-0.1882080883,-0.1347945929,0.1647774279,0.030728396,-0.3172053695,0.1159444898,-0.346139133,-0.4914352596,-0.1553007364,0.302551955,0.1290160716,0.2815119624,0.1602945328,-0.1667509675,0.0705202147,-0.193551898,0.4416457415,-0.0341807157,-0.2406097054,0.2495842278,0.2823726237,-0.792556107,-0.3175819218,-0.2055811733,0.0780888572,0.1872676164,0.1108763516,-0.2535757422,0.1550148875,-0.0309476759,0.4460534751,-0.0348791964,0.0585274696,-0.4713844359,-0.4379486144,-0.3892823756,-0.1156110764,-0.0409338325,0.3213148117,-0.0254295487,-0.0410624593,-0.2078595906,-0.0732220039,-0.1497963369,0.2925761342,0.225196749,-0.1615441144,0.2594017386,-0.1084294468,-0.0424697399,0.4509440362,0.4091886878,0.0962642357,-0.7135218382,-0.1644943357,-0.0819562748,0.105110094,-0.1693468094,-0.1602805704,0.1434170008,-0.0038493138,0.0208178703,0.0615006723,0.0626365319,0.1762912422,0.147881642,-0.1285622716,-0.1909573823,0.478181988,0.3002482653,0.0286416765,0.4108391702,0.0258860588,-0.147750482,0.0995122567,0.1331348717,0.6035118103,0.0099610621,0.0515038557,0.5848056674,-0.3137616515,0.5632751584,-0.1316949576,0.039374955,-0.4316044152,-0.2410881966,-0.0023158144,-0.3130731285,0.3180978894,-0.3315046728,-0.2583664656,0.0785926729,-0.2612192035,0.1756118834,0.1027451605,0.1611453593,-0.3093280494,-0.0600255355,-0.1455919892,0.1129977778,-0.130908668,0.2148073018,-0.0661608651,-0.0331405438,-0.5395659208,-0.1773793995,-0.1691561043,0.0144573394,-0.2888350189,0.5264725685,0.1666902006,-0.32030496,0.3427615166,0.4244214296,0.2471299469,0.0057721124,-0.0455859154,0.088987954,0.2085776478,-0.2371683121,0.2529144287,0.0462157205,0.2042732239,-0.1563991904,-0.3190074861,0.4772299528,-0.0846555009,-0.0295606349,0.0000968873,0.4971589744,0.1131919548,-0.1217275634,-0.1202549413,-0.1595301032,0.1249535009,-0.101598233,0.1322087646,0.1119273975,0.0466097221,0.4556714594,-0.1267765015,0.0519725196,-0.0274901073,0.2234774679,-0.1149354428,0.062030647,0.212306127,0.0696586296,-0.3292407095,-0.1683603227,0.1508702934,-0.1530785412,-0.3438253999,0.0077211903,0.1318709254,-0.0082277087,-0.2520041168,0.3728192449,0.2031218261,-0.1084895879,0.1279940009,-0.4349561632,-0.2474517673,0.348454088,-0.0220291503,-0.1178769693,0.2887565792,0.0877575129,-0.0058842935,-0.0320914164,-0.2641662359,0.1931818873,-0.3227640688,-0.0316981152,0.1425732076,0.1127643809,0.0105734523,-0.0856628641,0.1398329586,0.0461558588,-0.0437212959,-0.256709367,-0.0659087598,0.0574810281,-0.0343529843,-0.1474078149,0.074134551,-0.264393568,0.0659722835,-0.0042302916,-0.0947163105,0.0316515788,-0.1309354901,0.0467784517,-0.0807721391,0.119444944,-0.2579266429,0.2794899344,-0.1730121225,-0.0186980069,-0.0069067883,0.225232631,0.2478505671,0.1844674945,-0.0095565571,-0.1614102423,-0.2980037332,-0.0180742815,0.2375543714,-0.3787600696,0.1076794118,-0.0900494531,0.0517655872,0.5189228654,-0.2699650228,0.0082829054,0.3809455931,0.1989725828,-0.4422296584,-0.3729238212,0.178609699,-0.070999518,-0.0171532836,0.0885272771,0.202518791,-0.0804912969,-0.0201552715,0.2654154301,0.125896126,-0.1370826662,0.3140000403,0.4234543443,-0.0378350914,0.023216635,0.2812746167,0.2963754237,0.1855885386,0.404745698,-0.3500570655,0.0282888897,-0.0369498804,0.1045031399,-0.0547996163,-0.2855793238,-0.0532268323,-0.0571664646,-0.1070988923,-0.1845387071,0.1097596288,0.4301123321,0.0574706718,0.030785488,0.0520781502,0.0177019034,-0.102380015,0.2919911742,-0.0336424932,-0.1272791028,0.1321565062,-0.0617207624,0.1519883871,-0.2190744728,0.3816211522,0.2350091636,-0.2098590583,-0.2127113938,0.2184395194,0.4200309813,0.0760444477,-0.1324736625,-0.1067286879,0.2407669127,-0.1935664713,-0.3823975027,0.2315094173,0.6090102792,0.3060406446,-0.2581614852,0.1225211918,0.087199688,-0.1709100753,-0.0398911312,0.0745106414,0.1315118074,0.0693416893,0.2778826356,0.2114032507,-0.1768911034,-0.0950515121,0.1358908266,0.1936742812,-0.1657042056,0.161897108,-0.3206715584,-0.2677470744,-0.0531426743,0.048151046,-0.2507852912,0.4375481606,0.2409529388,0.2532978952,0.0314250328,-0.3065511286,0.0889737979,0.057438612,0.452807039,0.2309438735,0.3824341595,-0.1360666156,-0.1408254206,-0.531080544,0.0665639415,-0.0521799922,-0.30100438,-0.0128997872,0.007157627,-0.0645844266,0.2349253893,0.1780211031,0.0269659236,0.3565102518,0.1520022601,-0.3744759262,-0.3461716771,0.3202928603,-0.1987849623,0.0454253517,-0.3909824789,0.053732723,-0.3766963482,0.1434802711,-0.349260658,-0.0599466637,0.2764985561,0.0935333446,0.566752851,-0.0021582509,0.5926168561,0.1934985816,0.0967657343,-0.0043814457,-0.1707550883,-0.1537618786,0.2509750724,0.1735661924,0.1745191216,-0.2626297772,-0.0890750661,-0.1791506708,0.404606849,0.2431994528,0.1485372782,-0.0523443706,0.0902898908,-0.0300599188,-0.0034671728,0.0776304081,0.0713858753,-0.173975274,0.139273718,-0.3855036497,-0.4081261754,0.2425971925,-0.2250495553,-0.2957756817,-0.0146500422,-0.0010494281,0.0575169623,-0.0983297899,-0.3163691163,0.0118404776,0.3954785168,-0.0599404685,-0.2244613916,0.3057887554,0.1292082965,0.0564274788,-0.0965944007,0.4190945029,-0.1479633451,-0.1125377864,0.0788376033,-0.1065108627]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2678","title":"Import Error in Kaggle notebook","comments":"You may need to restart your kaggle notebook after installing a newer version of `pyarrow`.\r\n\r\nIf it doesn't work we'll probably have to create an issue on [arrow's JIRA](https:\/\/issues.apache.org\/jira\/projects\/ARROW\/issues\/), and maybe ask kaggle why it could fail","body":"## Describe the bug\r\nNot able to import datasets library in kaggle notebooks\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n!pip install datasets\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nNo such error\r\n\r\n## Actual results\r\n```\r\nImportError Traceback (most recent call last)\r\n in \r\n----> 1 import datasets\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/__init__.py in \r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in \r\n 36 import pandas as pd\r\n 37 import pyarrow as pa\r\n---> 38 import pyarrow.compute as pc\r\n 39 from multiprocess import Pool, RLock\r\n 40 from tqdm.auto import tqdm\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in \r\n 16 # under the License.\r\n 17 \r\n---> 18 from pyarrow._compute import ( # noqa\r\n 19 Function,\r\n 20 FunctionOptions,\r\n\r\nImportError: \/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Kaggle\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":37,"text":"Import Error in Kaggle notebook\n## Describe the bug\r\nNot able to import datasets library in kaggle notebooks\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n!pip install datasets\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nNo such error\r\n\r\n## Actual results\r\n```\r\nImportError Traceback (most recent call last)\r\n in \r\n----> 1 import datasets\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/__init__.py in \r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in \r\n 36 import pandas as pd\r\n 37 import pyarrow as pa\r\n---> 38 import pyarrow.compute as pc\r\n 39 from multiprocess import Pool, RLock\r\n 40 from tqdm.auto import tqdm\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in \r\n 16 # under the License.\r\n 17 \r\n---> 18 from pyarrow._compute import ( # noqa\r\n 19 Function,\r\n 20 FunctionOptions,\r\n\r\nImportError: \/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Kaggle\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n\nYou may need to restart your kaggle notebook after installing a newer version of `pyarrow`.\r\n\r\nIf it doesn't work we'll probably have to create an issue on [arrow's JIRA](https:\/\/issues.apache.org\/jira\/projects\/ARROW\/issues\/), and maybe ask kaggle why it could fail","embeddings":[-0.3392023742,-0.0755517036,-0.1140920669,0.1258439571,0.200350374,0.0130886594,0.3512435555,0.3286038935,-0.0306172576,-0.0346199907,-0.1804621667,0.8273153305,0.0816761181,0.2334335148,0.0459386706,-0.0699091032,0.1140711233,0.2013818324,-0.0490400568,0.0703774542,-0.1178435981,0.0834914967,-0.2884147167,0.2321556211,-0.0584075451,-0.3381468058,0.1671685874,-0.0141577022,-0.243008405,-0.3790449798,0.3784846365,-0.1966231167,0.1817374527,0.3294718564,-0.0001053689,0.0604676642,0.3373236358,-0.0124659725,-0.3102439344,-0.1045579687,-0.144251138,-0.5257630944,0.102859512,-0.3128523231,0.4279817939,-0.3194990158,-0.2402214706,-0.1276630908,0.2308556437,0.1961776018,0.2656954825,0.4633553624,0.4862199128,0.0011487751,0.3741651773,-0.114361994,-0.1743800789,0.3628072143,0.240282014,-0.2248286754,0.0500849783,0.1181876883,-0.0509729721,0.0626084879,0.1312723756,0.0576717444,-0.105301097,-0.3694544435,-0.0583921708,0.0525120683,0.5115424991,-0.5075426698,-0.217743516,-0.0814533979,0.1224679798,-0.0286665689,0.1378105432,0.205057472,-0.0501832552,-0.121019423,0.2834636867,-0.2121216506,-0.2584421635,0.2005665153,-0.1483705193,0.4371291399,-0.0406336375,0.0960493162,0.2935906649,-0.1535814255,0.607909739,-0.0881311968,-0.1116075665,0.1210281327,-0.3806024194,0.1239959821,0.0679238811,-0.0673968717,0.0992070884,0.0918953493,0.0935985371,0.1550542563,0.1212946326,0.3166662157,0.1174943596,0.3261029124,0.0165996272,0.2997159362,0.2864753306,0.301731199,0.1729280204,-0.080028072,-0.0747301728,-0.396931529,0.1270083636,-0.0698432997,0.3046585619,0.1747567803,-0.3466577232,-0.0529749878,-0.3350789249,-0.0637516901,-0.0116905728,0.1123064607,-0.2094092816,0.1152171567,0.1874995381,0.1460001171,-0.1666171998,-0.0682867989,-0.152149722,0.3062852621,-0.0251900125,-0.2267778218,0.0288292114,-0.3074620068,0.1900967658,0.0459270477,0.2293920666,0.1597653776,0.1890352219,-0.1123431772,-0.2045965642,0.3003136516,0.0601037517,0.0712902844,0.1299431026,-0.4086226523,-0.1351532787,0.1063650399,-0.2069546431,-0.111725606,-0.2594716251,0.2262059152,-0.2221556306,0.0504627675,-0.5596429706,-0.2054488957,-0.0173047706,-0.0858525783,0.0026646075,-0.2396202087,0.2015488595,-0.2670944035,0.2504836023,0.0302170031,-0.3585840762,0.0127391433,-0.2459341139,-0.2216100693,-0.0705787316,0.1604956985,-0.2848051488,0.1845775843,-0.2264408767,0.1887181103,0.6346998215,-0.5204713941,-0.4979526699,-0.018640466,-0.1067583635,-0.0413649306,-0.0616752915,0.1173457056,-0.1233747005,0.2213110924,0.0595052391,0.1761829704,-0.0356127433,-0.1629237086,-0.2430763394,-0.2509732842,0.1073993966,0.1639567465,0.0860650688,-0.3046607077,0.1918822378,-0.1370863914,0.078192696,0.0483362637,0.0952313989,0.0352379233,0.3661607206,0.0145554803,0.0464886092,-0.1311745197,-0.2734666765,0.2382600307,-0.1955827624,0.0514372028,-0.3236394525,-0.0954131186,-0.3622351587,0.2734937668,-0.2208878994,0.0895869732,0.2002182305,-0.1121829674,-0.0092320796,0.3617269993,-0.0242020935,0.2681074142,0.1552404314,0.1296823621,-0.3341867626,0.3652006984,-0.3834288716,-0.1805645972,0.1077799574,0.1040864587,-0.1019412577,0.038060829,-0.0430635437,0.1366348267,0.0826368034,0.0447767116,-0.0120110819,-0.1960675567,0.156398043,-0.3880776465,0.0727496743,0.1435055882,0.3162689805,0.0837149397,0.0832008496,0.1296728551,-0.0675327405,-0.0983568504,0.0049291076,0.2275876254,0.1567096561,0.1868839711,0.1146123335,-0.0986970663,0.1540551186,-0.0639605597,0.0871819779,-0.0876762643,-0.1409797519,-0.0280204862,0.4053031504,0.0233542472,0.180837363,0.0304343924,-0.3962771297,0.1644174457,0.2427372932,-0.0300172195,0.3535702527,0.2256231159,-0.2056582123,-0.0412726216,-0.028223414,-0.1469460726,0.133950457,0.2999617755,0.3659511805,0.1645154357,0.2697115541,-0.0871521533,-0.0774467736,-0.2046924382,-0.065367952,0.2356103808,-0.1783069372,0.2326025069,-0.2168598175,-0.2743118703,-0.1611234993,-0.1632225811,-0.0215983372,-0.4483352304,-0.2108484209,0.2484633029,0.1435390264,0.1773296148,0.1788655668,-0.1151916981,0.0936763436,-0.2396900952,-0.2538070381,-0.4000017345,-0.3790816665,0.1432716846,0.3257195055,0.1581729054,0.226628229,-0.1519053876,0.253231436,-0.4079263508,-0.470117867,-0.2024706751,-0.0813918486,0.3739663064,0.2682346404,0.1427751929,-0.2532339096,-0.3090758622,0.383012563,0.0678421855,-0.2730149031,0.1693322957,-0.1098784879,-0.1784485281,0.0253319629,-0.3571756482,-0.3664132655,-0.4171212912,0.3046458364,0.1833467036,-0.1222848222,-0.1501722783,0.2870305479,0.2386408597,0.1402593553,-0.0331910923,0.043846257,0.2569077611,0.5885706544,-0.1185474172,-0.4215679169,0.2081314772,-0.2530041635,0.2879258394,-0.121299237,-0.4119750261,0.0059753773,-0.1581647396,0.6464039087,-0.193762362,0.2262534499,0.1384783089,0.188379541,0.0149807138,-0.0934884846,-0.2597441971,-0.1256247759,-0.0381724127,0.0162672307,0.0328296721,0.1183337048,0.0268893428,0.5717334747,-0.0252747666,-0.0560583211,0.2782263756,-0.1158827692,0.2086124122,0.0487818234,-0.3988294899,-0.0288758185,-0.1440940499,0.264315784,-0.1259354353,-0.2704821229,-0.247497052,-0.2997135222,-0.2148075551,-0.1263627261,-0.2221943885,-0.0054586139,-0.0757259354,0.0108818356,0.0595684722,-0.1000430137,-0.2002983689,-0.1745503247,0.2006658465,-0.0082941316,-0.280613333,0.1012868434,-0.3764595091,-0.4813119173,-0.1698222607,0.3195928335,0.1419187039,0.2067774236,0.1361561865,-0.1811806858,0.0918796211,-0.1955461651,0.3912837803,-0.0000506613,-0.1831399351,0.2893381715,0.2120577395,-0.7940057516,-0.298402369,-0.2168442458,0.0245926883,0.1518408954,0.0556017384,-0.2358745784,0.1825493127,-0.0067499988,0.425568521,-0.05207517,0.0466572605,-0.4544198513,-0.4137452841,-0.3685896099,-0.1270058751,-0.0514204204,0.3349300623,0.0022670631,0.0053239567,-0.1855451316,-0.1119637936,-0.1576914042,0.267035991,0.2146258652,-0.1115878597,0.2346787304,-0.1138721406,-0.0742620006,0.455401212,0.4173569381,0.1206770092,-0.7068604231,-0.1579635441,-0.1026354134,0.0967423692,-0.2202046365,-0.1588289738,0.1977932155,0.0221880116,0.0437975638,0.0417434424,0.0806979835,0.1985219568,0.0862323418,-0.0655687675,-0.1619855314,0.4999957085,0.2260810286,0.0486237258,0.4143022001,0.014600399,-0.1390304714,0.092745848,0.1526046097,0.5871587396,-0.0301652849,0.0560948513,0.590744555,-0.3014115691,0.5513144732,-0.123046577,0.0741301998,-0.4628839195,-0.2247702777,0.00550899,-0.3155349195,0.2969496846,-0.3124720156,-0.2542790473,0.032956142,-0.2659881711,0.2080370039,0.0553779416,0.1781307608,-0.2830354869,-0.0118617183,-0.1498598456,0.1494500041,-0.1639214307,0.1959712207,-0.0625473335,-0.05027413,-0.4486037493,-0.1684483588,-0.1643163115,0.0441637635,-0.269161284,0.5237214565,0.1890530884,-0.3010551929,0.3148763776,0.3913817704,0.1751127392,-0.0387776308,-0.0569794588,0.1074403822,0.1616203338,-0.2480660528,0.2128370702,0.0806438103,0.2055585235,-0.1866948158,-0.3298456669,0.5272290707,-0.04522283,-0.0355114639,-0.0341983847,0.4944119155,0.1025414616,-0.1520354003,-0.0937952921,-0.1552090943,0.0723394677,-0.1172850952,0.1601324379,0.1095342264,0.0587411448,0.4139998555,-0.1057189405,0.0269063506,-0.0296010375,0.2540785372,-0.1155892238,0.1067097336,0.2500082254,0.1303191483,-0.3078186512,-0.1945816576,0.180653587,-0.0622145012,-0.3541774154,0.0479429215,0.1496587098,0.1058175713,-0.2028281391,0.3725810647,0.1533964425,-0.1213793233,0.1287823915,-0.3833631873,-0.2656586766,0.3385193944,-0.009002313,-0.1103962287,0.2926871479,0.0781982765,0.0313309953,-0.0062161991,-0.3084301651,0.2052085251,-0.3302122056,0.0083814608,0.1011927873,0.0794145018,-0.0308862329,-0.0641281456,0.1722777933,0.0438548177,-0.0863537937,-0.2966591716,-0.073249042,0.0432248786,-0.0479828864,-0.1568133235,0.0791949928,-0.2420348227,0.0501923636,-0.0129745714,-0.0702720359,0.0301929265,-0.1627746224,0.0894445181,-0.0856829137,0.0408185832,-0.2453520149,0.2226581126,-0.1917177886,-0.0080792122,0.0023848678,0.18243514,0.2653460205,0.145166263,-0.0172718894,-0.1375804096,-0.3005637527,-0.0740327239,0.2530136406,-0.3708564043,0.0302670598,-0.1211541966,0.0681599006,0.4832296968,-0.2564840317,0.0227945708,0.3177834451,0.2227733135,-0.4727803469,-0.3402053416,0.1669091135,-0.0276112594,-0.0059124553,0.0402585045,0.1804118901,-0.0214571785,-0.0464722104,0.2500554025,0.12498229,-0.1655674279,0.2797202766,0.4209264219,-0.0494952276,0.0264377594,0.2675212324,0.2963806987,0.2045336366,0.3570425212,-0.3276477158,0.0601757355,-0.0562297702,0.0378401317,-0.0056671384,-0.313868314,-0.108388707,-0.0105359964,-0.1038451344,-0.1855692714,0.1291917413,0.4432267845,0.1279405653,0.0593539029,0.0844720975,0.0607650094,-0.1108182892,0.2886046469,-0.0729117841,-0.1020413414,0.0874914303,-0.0401765443,0.121472314,-0.205082342,0.408272475,0.242234692,-0.196959421,-0.2069229633,0.20369187,0.4047403634,0.0780595914,-0.1416164339,-0.1425926238,0.2495490015,-0.1884701848,-0.3304874003,0.2276614308,0.5777513981,0.2577286363,-0.2740432322,0.1154701635,0.0717173368,-0.1743362695,-0.0180987623,0.0540499166,0.1704169214,0.1248132586,0.2793814838,0.2489602119,-0.1899071783,-0.1014503315,0.1192901731,0.1714862436,-0.1778189093,0.1440205574,-0.3029229045,-0.2679336071,-0.0933262184,0.0538843498,-0.2071690708,0.401553005,0.2685337067,0.2801018357,0.0139480932,-0.312048763,0.1121473685,0.0517153479,0.4743934572,0.243885681,0.3920080364,-0.1502237171,-0.122665979,-0.5042753816,0.1037847102,-0.0440829583,-0.3021405041,0.0058270576,0.0311485231,-0.038629964,0.2410347611,0.2164308131,-0.0111482218,0.3305140734,0.1124736518,-0.3594214916,-0.2880426049,0.3069213927,-0.2013653964,0.0282324553,-0.3520439565,0.0942914411,-0.4053052366,0.1781670898,-0.3622369468,-0.1047142148,0.2711576819,0.1081959903,0.5927200317,-0.0197991524,0.5346366167,0.1709815115,0.1650644988,0.0199482013,-0.2127310336,-0.1448093504,0.2392459214,0.1270064414,0.2343883216,-0.2212746739,-0.0960163474,-0.2016210109,0.4118447006,0.2190938294,0.1606915444,-0.0283426959,0.0924783126,-0.0056446786,-0.0250828415,0.0812172219,0.0568823479,-0.1496459246,0.1283640414,-0.3866488934,-0.4248799384,0.2752146125,-0.217607379,-0.3154255152,-0.0374940857,0.0134215457,0.1010604128,-0.1023396179,-0.3162206709,0.0120670311,0.4264295101,-0.0359578021,-0.2100612968,0.2819170952,0.1119932607,0.1220196411,-0.0898990557,0.3843321204,-0.0953404531,-0.1146392003,0.0220286213,-0.1456835568]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2678","title":"Import Error in Kaggle notebook","comments":"> You may need to restart your kaggle notebook before after installing a newer version of `pyarrow`.\r\n> \r\n> If it doesn't work we'll probably have to create an issue on [arrow's JIRA](https:\/\/issues.apache.org\/jira\/projects\/ARROW\/issues\/), and maybe ask kaggle why it could fail\r\n\r\nIt works after restarting.\r\nMy bad, I forgot to restart the notebook. Sorry for the trouble!","body":"## Describe the bug\r\nNot able to import datasets library in kaggle notebooks\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n!pip install datasets\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nNo such error\r\n\r\n## Actual results\r\n```\r\nImportError Traceback (most recent call last)\r\n in \r\n----> 1 import datasets\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/__init__.py in \r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in \r\n 36 import pandas as pd\r\n 37 import pyarrow as pa\r\n---> 38 import pyarrow.compute as pc\r\n 39 from multiprocess import Pool, RLock\r\n 40 from tqdm.auto import tqdm\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in \r\n 16 # under the License.\r\n 17 \r\n---> 18 from pyarrow._compute import ( # noqa\r\n 19 Function,\r\n 20 FunctionOptions,\r\n\r\nImportError: \/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Kaggle\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":57,"text":"Import Error in Kaggle notebook\n## Describe the bug\r\nNot able to import datasets library in kaggle notebooks\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n!pip install datasets\r\nimport datasets\r\n```\r\n\r\n## Expected results\r\nNo such error\r\n\r\n## Actual results\r\n```\r\nImportError Traceback (most recent call last)\r\n in \r\n----> 1 import datasets\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/__init__.py in \r\n 31 )\r\n 32 \r\n---> 33 from .arrow_dataset import Dataset, concatenate_datasets\r\n 34 from .arrow_reader import ArrowReader, ReadInstruction\r\n 35 from .arrow_writer import ArrowWriter\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in \r\n 36 import pandas as pd\r\n 37 import pyarrow as pa\r\n---> 38 import pyarrow.compute as pc\r\n 39 from multiprocess import Pool, RLock\r\n 40 from tqdm.auto import tqdm\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in \r\n 16 # under the License.\r\n 17 \r\n---> 18 from pyarrow._compute import ( # noqa\r\n 19 Function,\r\n 20 FunctionOptions,\r\n\r\nImportError: \/opt\/conda\/lib\/python3.7\/site-packages\/pyarrow\/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Kaggle\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n\n> You may need to restart your kaggle notebook before after installing a newer version of `pyarrow`.\r\n> \r\n> If it doesn't work we'll probably have to create an issue on [arrow's JIRA](https:\/\/issues.apache.org\/jira\/projects\/ARROW\/issues\/), and maybe ask kaggle why it could fail\r\n\r\nIt works after restarting.\r\nMy bad, I forgot to restart the notebook. Sorry for the trouble!","embeddings":[-0.3438675702,-0.0770802647,-0.1142193824,0.1320918351,0.1916217655,0.0109465495,0.340642184,0.3310463727,-0.0405511446,-0.0315704569,-0.1770029217,0.8259581327,0.0784066692,0.2167555988,0.042296242,-0.0751289576,0.1151352748,0.2059997469,-0.0540337525,0.0684145167,-0.1204729155,0.0821902528,-0.2884452641,0.2308666408,-0.0536760129,-0.335398227,0.1634015739,-0.0156411771,-0.2512480915,-0.389315486,0.3791489899,-0.2030484974,0.1786520332,0.3298630118,-0.0001054691,0.0557772815,0.3331196308,-0.001695561,-0.3189184666,-0.0971903503,-0.1449750513,-0.5241268873,0.1139851436,-0.3169448674,0.4252889752,-0.3302152753,-0.2376512587,-0.1207024008,0.2349157035,0.2009016424,0.264329344,0.4617816806,0.4861875176,-0.004376133,0.3844998479,-0.1263221502,-0.167839393,0.3752985597,0.2364216447,-0.2263902128,0.0413371772,0.1175254211,-0.0614103116,0.0632822588,0.1276656091,0.0492599458,-0.1038790867,-0.3674464226,-0.0683321431,0.0503972881,0.5075068474,-0.5111188889,-0.2169642001,-0.0854614452,0.1243208125,-0.0331041217,0.1380392164,0.2016976625,-0.053589806,-0.1101543531,0.2808272839,-0.2064090818,-0.2585773468,0.1937158555,-0.1518905312,0.4403881431,-0.0387481116,0.0953781009,0.2886867523,-0.1460981816,0.6073461771,-0.0855476558,-0.1203497052,0.1216696128,-0.3787596524,0.1202813387,0.068331331,-0.0643597394,0.0970348492,0.0882156864,0.0919665843,0.1456046999,0.1305030137,0.3127405643,0.1170410216,0.3272093236,0.0092065716,0.291288048,0.2798186541,0.2912294865,0.1691599041,-0.0766246244,-0.0699370801,-0.3918393254,0.1344002932,-0.0694812238,0.3126778305,0.1718910336,-0.3556835651,-0.0494576171,-0.3413330615,-0.0610720553,-0.018745685,0.1171982065,-0.2134622037,0.1104203761,0.1844510287,0.1440460533,-0.1670670509,-0.0719212443,-0.1517760754,0.3045976162,-0.0294002872,-0.2297071368,0.0208934434,-0.3086881638,0.1870203167,0.0515802316,0.2191489041,0.1520774961,0.190795511,-0.1156799346,-0.1989659369,0.3034534454,0.0597290881,0.075162217,0.1255255342,-0.4011129141,-0.129481256,0.1090210378,-0.2078445405,-0.1130267158,-0.2632022798,0.2298652679,-0.2152012885,0.0539015308,-0.5561850667,-0.204222247,-0.0161544625,-0.0957035124,0.0088539263,-0.2448432595,0.2065392137,-0.2754352987,0.2436387539,0.0379972048,-0.3599003255,0.019681843,-0.2543708384,-0.2262233198,-0.0632984191,0.159200877,-0.2863209844,0.19538486,-0.2295816392,0.1917762458,0.638063252,-0.5143753886,-0.4943232834,-0.0183402691,-0.1120874807,-0.0296253059,-0.056907054,0.1142123267,-0.1233816445,0.2207448334,0.0638165101,0.1699175239,-0.03144373,-0.1549661905,-0.2469737381,-0.2476423383,0.1074169576,0.1619285345,0.0861109868,-0.3101319671,0.1939121932,-0.1340430975,0.0820696577,0.0488968976,0.0861349106,0.0410352498,0.3639855683,0.0073456755,0.0412799157,-0.1453770697,-0.2789386213,0.2334987968,-0.2037810087,0.043385569,-0.3181882203,-0.0976251066,-0.3575016558,0.2770489454,-0.2133723199,0.0811754838,0.2015091926,-0.1052717045,-0.0017692286,0.3523833752,-0.0231055394,0.2652153671,0.1516367644,0.1324870586,-0.33897686,0.3570480943,-0.3791424334,-0.1911290884,0.1065430865,0.1010324731,-0.0990490168,0.0375528336,-0.0400208198,0.1391092241,0.0869648904,0.0505159572,-0.0299680922,-0.197322458,0.1561044604,-0.3874383867,0.0762953386,0.1411544085,0.3138978481,0.0901013389,0.0919348896,0.1251188368,-0.0571356192,-0.1056846157,0.0100074066,0.2146949768,0.1489283592,0.1847998351,0.1219904274,-0.0952658504,0.1637495458,-0.0575594567,0.0876649097,-0.0831241906,-0.1441524774,-0.0253729504,0.4067189991,0.0275045056,0.1840658188,0.0293449238,-0.3974842131,0.1613838822,0.2377908081,-0.0388252065,0.3521713614,0.2281310111,-0.2038653791,-0.0367030092,-0.0257220864,-0.1459667534,0.1384558678,0.2991797626,0.3653391302,0.1721170843,0.2751637101,-0.0793751255,-0.0749894753,-0.2025559545,-0.0662610829,0.2388169616,-0.1793177426,0.2328462303,-0.2155467868,-0.2813491225,-0.1677555144,-0.168367967,-0.0178475194,-0.4374024868,-0.2076959312,0.2531118691,0.1370605528,0.1795032322,0.1891433746,-0.1191048995,0.0914462358,-0.2401306182,-0.2524135709,-0.3897508681,-0.3771707416,0.1407493949,0.3231416643,0.1610304266,0.2322367877,-0.1523366272,0.2447451055,-0.4053628743,-0.4641484022,-0.1996776909,-0.0798342302,0.3753569424,0.2712040544,0.1453299969,-0.2507660091,-0.3088523746,0.3701548576,0.0678835213,-0.2780373991,0.1633874774,-0.110238649,-0.1721238196,0.0301491357,-0.3643354177,-0.3662023544,-0.4178311229,0.3031972051,0.1823398471,-0.1202225611,-0.1583278924,0.2869707346,0.2404126525,0.1297005862,-0.0267205127,0.0517497584,0.2594495714,0.5824719071,-0.1134833768,-0.4176762998,0.1986724287,-0.2575908601,0.2834931612,-0.120212473,-0.4212935269,0.0007623154,-0.1604905128,0.6609592438,-0.1991951168,0.21468997,0.1496120393,0.1904706359,0.0129404534,-0.0890318006,-0.2557511628,-0.1402997077,-0.0363350734,0.0108378865,0.0229870286,0.1207668483,0.015704697,0.5826402903,-0.0225849897,-0.0533317551,0.27514112,-0.1212432683,0.2058857828,0.0454710908,-0.4002604187,-0.0308185071,-0.1482352465,0.2534759641,-0.1238636374,-0.2759883702,-0.2487933934,-0.3007619083,-0.2217008024,-0.1351772994,-0.2148837894,-0.000789777,-0.0668296516,0.0191461127,0.0608749948,-0.0958221704,-0.1899784952,-0.1774168164,0.1992028058,-0.0126465997,-0.2787075043,0.0954815596,-0.364117533,-0.485080421,-0.1638602018,0.3125023544,0.1482526064,0.1970622092,0.1409003139,-0.1766305268,0.0947585627,-0.1950093657,0.3924500346,0.0057315957,-0.1782677025,0.2852152586,0.2199833095,-0.7880539298,-0.2938235998,-0.2227154523,0.0238452721,0.1574361622,0.0574895889,-0.2375729084,0.187251389,0.0052277949,0.4398751259,-0.0548214205,0.0437475853,-0.453920722,-0.4103513658,-0.3626693785,-0.1204480529,-0.0482040159,0.3372283876,0.003894761,0.0014654777,-0.184218362,-0.1064535156,-0.1629242599,0.2704035938,0.2156515419,-0.1212681383,0.236635074,-0.1229669303,-0.0786359161,0.4644381404,0.4202221632,0.1237812787,-0.6989777088,-0.1592781395,-0.0817470998,0.1029516906,-0.2168052644,-0.1642383039,0.2025846392,0.0215025079,0.0443370938,0.0403110906,0.0755984336,0.2012656033,0.0874472037,-0.0733746216,-0.1510349065,0.5053038001,0.2279075384,0.0496598817,0.4068602622,0.0133208232,-0.1364239603,0.0958854482,0.1526998729,0.6006861925,-0.027830353,0.0508123748,0.6026001573,-0.3067754507,0.553940773,-0.1300061196,0.0840124339,-0.4649183154,-0.2181677967,0.0070012673,-0.3166860342,0.295065105,-0.3199923337,-0.2508445084,0.035638988,-0.2634040415,0.2071664184,0.0573786758,0.1801076531,-0.2765654922,-0.0084586227,-0.152601108,0.1460164189,-0.1678945273,0.1933187246,-0.06967628,-0.051618699,-0.4364555478,-0.1682930142,-0.1661471575,0.0437589549,-0.2752037942,0.5126556754,0.1887407154,-0.3083454072,0.3038027287,0.3915411532,0.1575747579,-0.0325702615,-0.0551358685,0.1062993333,0.1592918038,-0.2499812245,0.2163820267,0.0822423846,0.2063043416,-0.1870326996,-0.3244163692,0.530528605,-0.0557478294,-0.0302156117,-0.0245098546,0.4902910888,0.1098913029,-0.1539049298,-0.0912336186,-0.1535329968,0.0709805191,-0.1260724217,0.1597077399,0.1015518159,0.0588010624,0.4141299427,-0.1115617231,0.0191840883,-0.0419341363,0.2632978857,-0.1064587086,0.1108459607,0.2457139492,0.1302165687,-0.3101195395,-0.1941179484,0.1747112572,-0.0625777021,-0.3609369695,0.0510273762,0.1442755461,0.1181773469,-0.2049164772,0.3655370176,0.1593819559,-0.116038695,0.1217262,-0.3830699921,-0.2565536201,0.3387972414,0.0037231192,-0.1163126826,0.3004075587,0.0813373923,0.0303024705,-0.0058226306,-0.307453692,0.2028250992,-0.3254200518,0.0115600508,0.0950656086,0.0824567452,-0.0308389813,-0.0642866194,0.1781428456,0.0478643961,-0.0828657672,-0.2979176939,-0.0692147166,0.0436687842,-0.0529871471,-0.1559217721,0.0789398998,-0.2427777052,0.0613806844,-0.0118203266,-0.0639528409,0.0244867448,-0.1600545198,0.0814574957,-0.0773382336,0.0409088954,-0.2512698174,0.2282156944,-0.1928745359,-0.0163538773,-0.0043092957,0.1892685443,0.2621607184,0.1422040164,-0.0126650352,-0.1455067247,-0.2946572602,-0.0714451596,0.2462890893,-0.3729647994,0.0193480905,-0.1245205551,0.0701197386,0.4878968,-0.2644610107,0.0218162015,0.3213043511,0.2248581499,-0.4720913768,-0.3357239366,0.1608940512,-0.028791545,-0.0033933909,0.0460184254,0.1743441224,-0.0250529163,-0.0399036258,0.2498086989,0.1167117506,-0.1551566422,0.2698944211,0.4228686988,-0.0452196859,0.0207455065,0.2719663084,0.3008873761,0.2052327245,0.3551146686,-0.3248211443,0.064556472,-0.0516876876,0.0377670564,-0.011970263,-0.3060587943,-0.1069976464,-0.0035317303,-0.1044231281,-0.1902880967,0.1279561073,0.4319310784,0.131954357,0.0557391271,0.0828470662,0.0617178231,-0.1190722138,0.296446383,-0.0605186261,-0.1037477329,0.0912166759,-0.0383051038,0.1162862778,-0.2087968141,0.4120825231,0.2387765646,-0.2003590018,-0.208963111,0.2051567435,0.4049689174,0.0770315528,-0.144757241,-0.1308675408,0.253271848,-0.1904788762,-0.3400830328,0.2362891287,0.5830227137,0.2580747604,-0.2751707137,0.1093604192,0.0731223375,-0.1765502393,-0.0219403151,0.0576350801,0.160172835,0.1368487179,0.2793982923,0.2509010136,-0.19371894,-0.0918068737,0.1245842353,0.162954554,-0.1905058771,0.1447398067,-0.2970665991,-0.2748733461,-0.0879247189,0.0597498417,-0.2057939917,0.4084696174,0.274512887,0.2821597159,0.0049594394,-0.3078812957,0.1118856892,0.0549869463,0.4759660363,0.2466983199,0.3976466358,-0.1549991816,-0.1284190118,-0.501683712,0.1108886227,-0.0524985567,-0.3014678061,0.0038353808,0.0309390258,-0.0469313785,0.2556980252,0.2194003165,-0.0072288988,0.3248770833,0.1229475215,-0.356780827,-0.2917700708,0.3050172031,-0.2067124248,0.0295966174,-0.3504011035,0.1054066941,-0.4134699702,0.1766074598,-0.359098196,-0.1072512642,0.2813653648,0.114610374,0.5974835753,-0.0193357673,0.5290721059,0.1665176451,0.1585889608,0.0253884699,-0.2136002183,-0.1434469223,0.2446595281,0.1301643997,0.2302502096,-0.2169490755,-0.0975293517,-0.1996528804,0.4094184935,0.2201499343,0.1637845486,-0.0218733884,0.0871207118,-0.0011448949,-0.0236402415,0.0805426985,0.0555492826,-0.1418817937,0.1273859143,-0.3887636662,-0.4281505048,0.2744404972,-0.2218314707,-0.3107546568,-0.0388694443,0.024493156,0.1010354012,-0.1053964272,-0.3233850896,0.0116408402,0.4379050434,-0.0378320962,-0.2044843733,0.2812786698,0.1053229347,0.1228214279,-0.0860601142,0.3797081709,-0.0952712223,-0.1166139916,0.0315258242,-0.1509275883]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2677","title":"Error when downloading C4","comments":"Hi Thanks for reporting !\r\nIt looks like these files are not correctly reported in the list of expected files to download, let me fix that ;)","body":"Hi,\r\nI am trying to download `en` corpus from C4 dataset. However, I get an error caused by validation files download (see image). My code is very primitive:\r\n`datasets.load_dataset('c4', 'en')`\r\n\r\nIs this a bug or do I have some configurations missing on my server? \r\nThanks!\r\n\r\n\r\n\"\u0421\u043d\u0438\u043c\u043e\u043a","comment_length":27,"text":"Error when downloading C4\nHi,\r\nI am trying to download `en` corpus from C4 dataset. However, I get an error caused by validation files download (see image). My code is very primitive:\r\n`datasets.load_dataset('c4', 'en')`\r\n\r\nIs this a bug or do I have some configurations missing on my server? \r\nThanks!\r\n\r\n\r\n\"\u0421\u043d\u0438\u043c\u043e\u043a\nHi Thanks for reporting !\r\nIt looks like these files are not correctly reported in the list of expected files to download, let me fix that ;)","embeddings":[0.096845977,-0.0686112046,-0.0539066978,0.4750131071,0.2311235666,0.2395801693,0.0058158985,0.1676318944,-0.0170363579,-0.1058175042,-0.0259849485,-0.2992729247,0.1618457735,-0.0318008177,0.0029821228,-0.1193322763,0.135855034,0.084860906,-0.1391199231,-0.3514162302,-0.1153822467,0.0911915377,-0.1484110653,0.1158548519,-0.026407348,-0.0274939761,0.0056613502,0.0816003084,-0.3496825397,-0.3620357513,0.555793941,0.1042395383,0.1785736382,0.0744807795,-0.0001167735,-0.1865388751,0.3024600446,-0.3341219127,-0.1340886652,-0.121632129,-0.3307465017,-0.2062899321,-0.2419678271,-0.3494567275,-0.0442584194,-0.140995115,0.1553079635,-0.1196931154,0.0351108015,0.4630728066,0.1809932441,-0.1043740585,0.0083733583,-0.1809168756,-0.0925607607,0.0102551347,0.0669598803,0.1523742378,0.1125920042,0.295470953,0.3457050622,0.1438502967,0.0234140679,-0.0461742245,0.2648062408,-0.1932270974,0.1687993854,-0.3840813935,0.3017909229,0.2715108693,0.9384877682,-0.1433071792,-0.1691133976,0.2467643768,0.1892973334,-0.1855994314,0.294626236,0.6772632599,-0.0679940879,0.1251312643,-0.3260817826,-0.1865656674,-0.1197463498,0.2777039707,-0.1292812526,-0.2179443389,0.0771927834,-0.051841978,0.1002988219,0.2046857923,0.0032096968,-0.4079837203,-0.177982986,0.2119580954,-0.1153169423,0.0007458262,-0.4127675593,0.3904469013,-0.0880833864,-0.0488232821,-0.2443313748,-0.0720183104,-0.1606342196,0.0810397491,0.1806429774,-0.0447047949,0.1368902624,0.1567291468,0.0802775994,-0.1490299851,0.1545657068,-0.0944330469,-0.1808453649,-0.3186020255,-0.0616494119,0.2068343312,-0.1288014203,-0.4318065941,-0.2298416942,0.2312493473,-0.3119583428,0.0769578367,-0.1125138998,0.4242171347,-0.4325264394,-0.0528689064,-0.0815890431,0.1484345198,-0.0591337755,-0.5108196139,-0.1177364141,0.0567053556,-0.2883011997,-0.2188446075,0.2805007994,-0.5110835433,0.2950258553,0.0521375276,0.1721790731,-0.1180123091,-0.0835570171,-0.3251758814,0.0874162763,0.2726626396,-0.1362612098,0.1245287657,0.2622080147,-0.2980207503,0.0288882442,-0.0410171784,-0.3503580689,-0.120409824,-0.1634520143,0.1190706417,-0.2125154436,-0.0093940236,-0.349647373,-0.2559763491,0.2443347275,-0.0921485797,0.1594803333,-0.3166207373,0.031773597,-0.1681563109,-0.0448281765,0.3502218127,-0.3620434105,0.0180544611,-0.0329119675,-0.1109365821,0.361064136,0.2010711879,-0.1729782671,0.3088462949,-0.2813074589,0.2549065053,0.0488998368,-0.5429531932,-0.4550856054,0.2392617166,0.1014861241,-0.0787607357,-0.226033479,-0.0706493407,0.388184011,-0.1845718324,-0.0514269136,0.0685296431,0.0910026878,0.0454453379,-0.1369255334,0.0582367703,0.1933749467,-0.0361915529,0.2007416189,-0.0915700868,0.2835775614,-0.1294594705,0.2464183122,-0.190301314,0.1727260649,0.2742322385,0.1823205501,-0.1228114814,0.1321302801,0.3525640666,-0.4031892121,0.1628526002,-0.043868646,-0.000149133,-0.0202082377,0.0030592661,-0.2633752823,-0.1960603148,-0.2828921676,-0.0816282704,0.029565528,-0.1728123724,0.0364455581,0.1513742357,0.1686381549,0.5813813806,-0.2374359816,0.0732231513,0.1647721529,0.3052451909,-0.1146178693,0.3243879974,0.0605081208,-0.197623983,0.301115036,-0.1183037311,-0.1741787195,0.1802297086,0.2748375535,0.2119431049,-0.3257678449,0.0051281089,0.4052165747,-0.3786321282,0.1828547269,0.4098116457,0.1813715845,-0.0424861833,0.0105122207,-0.0025696948,-0.0898008645,-0.2186424285,0.0273648687,0.1081600562,0.3547260463,-0.0161224678,-0.2691572905,-0.1640057266,0.3425188959,-0.2733933926,0.0401486084,-0.0079950253,-0.1774633229,-0.3209866583,0.4895766377,0.0607112646,-0.0399639159,0.1609406471,-0.0280320551,0.0574853681,0.3541265428,0.2005570829,0.5117332935,-0.0272196513,-0.2033050656,0.2647412419,0.1374890059,-0.1859330535,0.3109610081,-0.2197519541,0.1099894643,0.1862859279,-0.0879818574,-0.1205739602,-0.3998787403,0.150488466,0.3180182278,0.0884106085,-0.3093161583,0.1285469532,-0.4081644118,-0.2306700349,-0.2915096581,-0.1382839978,-0.0635014698,0.0557817519,0.0039454689,-0.1081699133,-0.1405950338,-0.0048358324,-0.6207512021,-0.0432001874,-0.0407774523,-0.1709545106,0.0394609161,0.2573703825,-0.235153839,0.0094771134,0.1098938584,0.0883889273,0.3553434014,-0.3984425068,0.262776792,-0.3834376931,-0.2011980861,0.2526563704,0.0835286155,-0.1793435067,-0.0235761423,0.2764116228,-0.3303965926,-0.1471963376,-0.0817749277,0.05934266,-0.159355104,0.198516652,0.0402036086,-0.1698155254,-0.1646369845,-0.5502979755,-0.4189038277,-0.0854859203,-0.1607704014,-0.1752345115,0.0396212041,0.4361684918,0.0182257835,-0.0780427381,0.1090334654,-0.1596165448,-0.2542777956,-0.1209918112,0.4298145175,-0.2539363503,-0.5836586952,0.1029922739,0.0107242055,0.1744459718,0.3262527287,-0.3350965381,0.3499177694,0.1039251313,0.0213261619,0.0253066868,-0.2157391608,0.3457696736,-0.2229926735,0.1244463027,-0.1363958716,0.2712985277,0.1059970111,0.2097367644,0.5329785943,-0.4421275258,-0.0302732196,0.1538648307,0.2319278717,0.1277960241,-0.2935450375,0.5387865901,0.2865954041,0.3581418395,0.2793970704,-0.3494909108,0.3341584504,-0.2343039513,-0.0606031269,0.3306901753,-0.1923257411,-0.3205670416,-0.154357031,0.1352211833,-0.3458194137,-0.125665307,-0.0995995998,-0.0232364833,0.36725384,0.3243452609,0.3868780136,0.2190692425,-0.2529470026,0.0879832804,0.3665077686,0.192277953,0.1218931153,-0.1594341844,-0.2385463119,-0.0137447994,0.2690027654,0.3512971997,0.1429186463,0.0086921025,0.3366684616,-0.0169546269,0.0273881294,0.5130409598,-0.3365210891,0.1041382402,-0.1199322045,0.3165088892,-0.1912990957,-0.2906217873,-0.1121694893,0.0052343556,0.2563720345,0.2647580206,-0.3294284642,0.1596473753,-0.1577940583,0.3297686875,-0.071708329,-0.3489955366,-0.1800928265,-0.3787818551,-0.2861402333,0.2097589672,0.0344523117,0.109716095,-0.252162993,0.0800885037,0.1153602451,0.4076044559,0.1858725697,0.0648535267,0.2890301943,0.0751469508,-0.0058169202,-0.0153915156,0.0391469225,0.7092399001,0.3661020398,-0.1685134619,-0.280257076,-0.1138140485,-0.2125694603,0.1873963773,0.1063665897,0.0132571263,-0.1825789362,0.238349691,-0.2654716372,0.0752489492,0.3841482997,0.1735219657,-0.0579014793,-0.6417623758,-0.4279414117,0.0271733105,-0.165762946,0.0767464563,0.1531163156,0.218525663,-0.1480732262,-0.2544135749,0.2818487287,0.8829857111,0.1032863334,-0.1745401472,-0.0468322299,-0.1570571959,0.1641242802,-0.6999188066,0.0634924173,-0.2291508615,-0.4758976996,-0.2166135162,-0.1454949975,0.3087323308,0.480378896,-0.2660745978,0.3267466128,0.1208245978,0.2696650326,0.0956342369,0.1172953472,0.0867618397,-0.2291977406,-0.3270386755,0.1372856796,-0.2278520167,0.3816149533,-0.3304398656,-0.0874117911,-0.035494063,-0.343280077,-0.2652930617,0.3194360137,0.0433429964,0.0522038452,0.0054626083,-0.0803291872,0.3363468349,0.3796256781,0.3059648275,0.5810824037,-0.1137592793,0.0904678851,0.0829676017,-0.214010939,-0.0017038669,-0.1863299459,0.0727234706,-0.2209589928,-0.163743481,0.1252422333,0.025680488,-0.1554332376,0.3542510867,-0.0052304799,-0.3836031556,-0.048526939,-0.318536669,0.0018296841,0.0806474313,0.0741632953,0.0542365238,0.0092410017,-0.0484004617,0.2165422142,0.2888498306,-0.2738632262,-0.1093748659,0.4820823669,0.0069062877,-0.422577858,0.3957419693,0.0692836121,-0.2859167755,-0.2346082777,-0.2766028047,-0.3226013184,-0.437826246,-0.2122822702,-0.1211898923,0.2519947886,-0.350102365,0.2291532308,0.1092435792,-0.0886007473,0.3342339993,-0.4625097215,-0.1260686964,0.0029643304,-0.1780199111,-0.0232610498,0.1158497855,-0.0608183555,0.1773510128,0.0480661169,-0.2034521848,0.0295510907,-0.2206774652,0.0934600681,0.1844990849,0.0291805509,-0.0149676986,-0.2728755772,0.0106825884,0.2006671876,-0.1280092895,-0.1032630652,-0.043616347,0.1919803172,0.1563544422,0.0032915976,0.3131914735,-0.2257286906,-0.1605040431,-0.0429567434,-0.0485746227,0.2598894536,-0.0353706256,0.2518923581,0.2516219914,0.1158096716,-0.2370556444,0.0919012427,0.1132134423,0.4768443108,-0.0105123362,0.0880640224,0.1420779675,0.0484816581,-0.233406201,0.0994983613,0.2594985068,-0.0611053407,0.3047938347,-0.3842670321,0.4487786591,0.1527723372,0.0259962138,0.2158565521,-0.2095505148,-0.0445937626,0.4983430207,0.1671554446,-0.2295721471,-0.0541868992,0.1648519486,0.2072173357,-0.2720639706,-0.0493467934,0.1995109916,-0.3289795816,0.2600112259,0.0632012114,0.3941882551,0.006607607,0.265260309,0.4170495868,-0.0522826277,0.1995477676,0.0526961833,-0.136483416,-0.1017281711,0.131113112,0.1044999361,-0.2310553044,0.3705916703,0.2657863498,0.0234879274,-0.4831438959,-0.0467823148,0.3182042837,-0.0425548479,0.1746276915,0.424246192,-0.1332391798,0.3142977655,-0.197224021,-0.8147056103,0.2134571671,-0.113661088,0.2533870041,-0.0570821501,-0.2354959846,0.1878345162,0.3116334379,-0.1242853403,-0.0830523968,0.1101040095,0.308849901,-0.2186730057,-0.2400930375,0.2889841199,0.0278465282,0.0292526707,-0.2123586684,0.231757611,0.2056273967,0.0963573456,0.1508677006,0.0168426093,0.2187575698,0.3880805671,0.2287818491,0.0126237841,0.0145515623,-0.0118375542,0.1728646308,0.5091121793,0.4928583205,0.1775871664,0.363219589,0.0782714784,-0.0588837564,0.1305732578,0.0597452037,0.1148815975,-0.1486169547,0.2104657739,-0.0780468881,0.0869636312,-0.3805343211,-0.0649956986,-0.2805466056,0.3111959696,0.5664162636,-0.0162147805,0.1151912138,-0.1281662583,0.0291566476,-0.1101016328,0.3644055724,0.2182445377,0.498585254,-0.3034207821,-0.2396450937,-0.617010057,0.3663160503,-0.3759625852,0.2365921438,-0.1307976842,-0.0122162793,0.3323756456,0.370773524,-0.0665928349,0.3085778356,0.1430294067,0.4748997986,-0.0532182977,-0.0703132525,0.2046019733,0.0699485466,-0.1227095053,-0.4148941338,0.0504748672,-0.1069609672,0.0798351467,-0.180724889,-0.0365894921,-0.0613417439,0.2318529636,0.3578616083,0.1050796956,0.883009553,-0.0567258224,-0.0103408061,-0.0094702449,0.0541442782,-0.142879039,0.2769037783,0.020986326,0.1057195812,0.0254502129,0.0194994099,-0.1232581213,0.4462434649,0.1924265176,0.1385368109,-0.3047604859,-0.2555625439,-0.366092205,-0.2397816926,-0.0890367329,-0.1024262607,-0.012979541,-0.0846836045,0.0473409668,-0.404579252,0.5043489933,-0.5797134638,-0.1596442908,-0.0519169085,0.1902922243,-0.0386947244,-0.1701991707,-0.1726155728,0.2493821681,0.3882625401,-0.2096813768,-0.2787168026,-0.0153385336,0.0080338167,-0.0208508857,-0.051716812,0.7078829408,0.0515536666,-0.4494343102,-0.1627371758,-0.4089475572]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2677","title":"Error when downloading C4","comments":"Alright this is fixed now. We'll do a new release soon to make the fix available.\r\n\r\nIn the meantime feel free to simply pass `ignore_verifications=True` to `load_dataset` to skip this error","body":"Hi,\r\nI am trying to download `en` corpus from C4 dataset. However, I get an error caused by validation files download (see image). My code is very primitive:\r\n`datasets.load_dataset('c4', 'en')`\r\n\r\nIs this a bug or do I have some configurations missing on my server? \r\nThanks!\r\n\r\n\r\n\"\u0421\u043d\u0438\u043c\u043e\u043a","comment_length":31,"text":"Error when downloading C4\nHi,\r\nI am trying to download `en` corpus from C4 dataset. However, I get an error caused by validation files download (see image). My code is very primitive:\r\n`datasets.load_dataset('c4', 'en')`\r\n\r\nIs this a bug or do I have some configurations missing on my server? \r\nThanks!\r\n\r\n\r\n\"\u0421\u043d\u0438\u043c\u043e\u043a\nAlright this is fixed now. We'll do a new release soon to make the fix available.\r\n\r\nIn the meantime feel free to simply pass `ignore_verifications=True` to `load_dataset` to skip this error","embeddings":[0.0360019878,-0.0390101783,-0.0446487889,0.4204434156,0.2243454158,0.2426010668,-0.0268920865,0.1530843377,-0.0672710538,-0.0291624796,0.0164733157,-0.2508128285,0.1566776633,0.0499308072,0.0004604018,-0.0572996549,0.1233928427,0.0768573433,-0.1300052702,-0.2965222001,-0.1279532611,0.0637142062,-0.1454876363,0.0978287309,-0.0157246217,0.0041988972,-0.0361221768,0.112853542,-0.3306386173,-0.4125275612,0.5867489576,0.0864503309,0.2514703274,0.0607930385,-0.0001131244,-0.1866180599,0.3207838237,-0.2820208967,-0.1749012768,-0.1334163994,-0.325238049,-0.1429717839,-0.2270385921,-0.3486924767,-0.0586523786,-0.0870727822,0.1525995135,-0.1404471695,0.0293257777,0.3692456782,0.2102141678,-0.1186661646,0.0342105515,-0.2088637203,-0.146671623,-0.0970341563,0.0677079633,0.1624719352,0.136817649,0.2822684348,0.3550719321,0.2087816447,-0.041719079,-0.0173644219,0.3091590405,-0.2235092521,0.2251718789,-0.3702123761,0.3360903561,0.279658556,0.965049088,-0.1046360657,-0.1588131785,0.2854292095,0.2251126915,-0.1731430739,0.3463103771,0.6254925132,-0.0934754759,0.1446568221,-0.3239200711,-0.2792814374,-0.1134175807,0.2988192141,-0.1196650937,-0.1817182899,0.1240483895,-0.0344568565,0.1894722432,0.2115551978,0.051272653,-0.3572794497,-0.229422465,0.2440073341,-0.1396593153,-0.0109471083,-0.3934450746,0.4090083539,-0.1072911099,-0.0308373738,-0.3079855144,-0.0492989607,-0.1120638251,0.0944594964,0.2030055076,-0.0697778836,0.1198696271,0.1149196029,0.1016233638,-0.1529776305,0.2040787488,-0.0545782261,-0.1837911159,-0.3530216813,-0.0942231119,0.155848965,-0.1233038604,-0.3783267438,-0.1654168069,0.1764023751,-0.3575060666,0.1171477735,-0.137021333,0.4213334322,-0.3903787732,-0.050491225,-0.0359500907,0.1107082516,-0.1120523885,-0.5132819414,-0.1210590228,0.0789467245,-0.2682026923,-0.2268249393,0.2764569819,-0.51893121,0.2074765712,0.0441430844,0.1534954309,-0.0671436042,-0.0873310044,-0.282404989,0.0264247432,0.2698502243,-0.120303981,0.0706778541,0.2335549742,-0.3244762123,0.0129573429,0.0130953621,-0.3358463943,-0.1569307148,-0.2096270174,0.1417196691,-0.1472776085,-0.0375684462,-0.3525378704,-0.2765506208,0.2389992476,-0.0917323455,0.1340071261,-0.2895190418,-0.012052021,-0.1646085829,-0.0726770982,0.3515331745,-0.3475580812,0.0224019084,-0.063313365,-0.0984154344,0.3028386533,0.1756015271,-0.1984318197,0.2672305703,-0.2671558261,0.2501603365,-0.0137765417,-0.5401049852,-0.509370625,0.2695917487,0.055113297,-0.0686934739,-0.2188578397,-0.035527721,0.3348928094,-0.192320019,0.0134166107,0.1395029426,0.0648256168,0.0564677566,-0.1089969426,-0.0176786371,0.2234360576,-0.0455377512,0.2158854306,-0.0558993369,0.2599472702,-0.0973496363,0.2286307067,-0.2337479144,0.1718613952,0.3047979772,0.2269141823,-0.1765137315,0.0588294864,0.3398589194,-0.4050041735,0.2001463026,0.0163853783,0.0992064029,-0.0325783491,-0.0069131684,-0.194969207,-0.1770458966,-0.2826213539,-0.099384889,0.0356473997,-0.2061429918,0.0358737074,0.1314911544,0.190925315,0.5959507227,-0.2169481367,0.0794818178,0.136102736,0.2638708055,-0.1812299639,0.3510243893,0.0228076521,-0.1756054759,0.2562204301,-0.1429301947,-0.1528047174,0.2349878699,0.2168685496,0.2181942463,-0.3224677145,-0.0151658105,0.3912174702,-0.3939552009,0.142312929,0.3844948113,0.2187794298,-0.016003333,0.0914730579,-0.046353627,-0.1119231656,-0.2208696008,0.0427727215,0.0983948335,0.3558832705,0.002010664,-0.2500748932,-0.1520540118,0.3440111876,-0.3176958561,-0.0068033431,-0.0388007686,-0.1095327437,-0.3359957337,0.4060756862,0.0946510211,-0.0373951457,0.1653100997,-0.0237267278,0.0274253991,0.3157754838,0.248559311,0.447175622,-0.0136000365,-0.2143898904,0.2998844683,0.1573820561,-0.1602239013,0.3488217294,-0.2297730446,0.1348233819,0.0508116186,-0.1089659184,-0.0918366611,-0.3125115931,0.116694212,0.3133603334,0.1552191228,-0.2792727649,0.1318763644,-0.4109159708,-0.2489753813,-0.1848032624,-0.1850156784,-0.0852527842,0.0458248407,-0.0152919302,-0.0633327588,-0.1338995546,-0.0200535469,-0.5972525477,-0.0605826043,-0.0365189649,-0.2365607917,0.0650054887,0.1695266664,-0.2182215899,0.0348358937,0.1568024307,0.0094862776,0.3665184677,-0.3867044449,0.2242770791,-0.3541082144,-0.234388411,0.2437724024,0.0588889793,-0.1631189287,-0.0350466035,0.2716336548,-0.3771987855,-0.1262652874,-0.1053924859,0.0165898465,-0.164206937,0.2119745761,0.1110279858,-0.1994582415,-0.1752018034,-0.5475433469,-0.4019851089,-0.0843536481,-0.1555342674,-0.1761154383,0.0200413894,0.4436435997,0.0465703383,-0.0550639629,0.1353532821,-0.2508653402,-0.3042049706,-0.106329754,0.319399029,-0.2624073327,-0.5575564504,0.1595307887,0.0274067931,0.1751662493,0.3157703578,-0.3637105525,0.3887701333,0.0978396684,0.1054476723,-0.0301313512,-0.2219379991,0.4351092875,-0.2020962983,0.0946493298,-0.0956711918,0.276397109,0.0846992806,0.1521805078,0.5275927186,-0.428185761,-0.03821465,0.1636305153,0.176832363,0.0551307313,-0.2453256696,0.5363041162,0.2724509239,0.396664083,0.2414665669,-0.3665806055,0.337654233,-0.2462864965,-0.0973270014,0.305460155,-0.2202027589,-0.3122652769,-0.2184076905,0.0671336204,-0.2943171859,-0.1775604039,-0.1228683367,-0.0551597141,0.3067913949,0.3584987819,0.3822774887,0.2345034778,-0.2700302005,0.0280062724,0.328011781,0.1204863042,0.1090977937,-0.1994676888,-0.2411132306,-0.0692342371,0.3282172084,0.3817484975,0.1986070126,0.0327287801,0.3442434669,-0.0016995905,-0.0023001926,0.509965241,-0.2901915014,0.1360222697,-0.1347065568,0.295514822,-0.2025738209,-0.2343929559,-0.0494216345,0.0251400024,0.2978613973,0.2174991816,-0.279897809,0.1700639874,-0.1417282522,0.3046318889,-0.0283361599,-0.3298509419,-0.184249416,-0.371111095,-0.3130017817,0.2186618149,0.0056444854,0.1686593443,-0.2048041224,0.1331519485,0.069401443,0.4012479484,0.1360420734,0.0466868281,0.2689102888,0.0711444095,-0.0031647035,-0.0124653997,0.0292657018,0.7025738358,0.3302221894,-0.1346909702,-0.2716224194,-0.1439027488,-0.124485977,0.1391993612,0.1108241901,0.0168137811,-0.1897696257,0.2859244049,-0.2699915767,0.0322893411,0.2813590169,0.1306811571,-0.0307326261,-0.6320250034,-0.4210796356,0.0609079227,-0.1813705415,0.0958567038,0.1489226222,0.2680442035,-0.1285711378,-0.2174446583,0.2188282311,0.8750144243,0.129087314,-0.1301256269,-0.0264667366,-0.088605687,0.216375649,-0.6644521952,0.1195862144,-0.225665465,-0.3912491202,-0.2204607427,-0.128436327,0.3262503743,0.4141201079,-0.2569352388,0.3033985198,0.0509971492,0.2380173206,0.0988602787,0.088677384,0.0070952317,-0.2492571622,-0.2533776462,0.1659843773,-0.2577988207,0.3493568003,-0.3405812979,-0.0519937202,-0.0479628369,-0.3674954176,-0.3082837462,0.369809866,0.0263160206,0.1083767712,0.002481838,-0.0061472631,0.3616018593,0.3097478449,0.3440255225,0.5899108648,-0.1518326551,0.1053577438,0.0975242406,-0.2661140263,0.0407299511,-0.2164760232,0.1155538335,-0.1921957731,-0.2545380592,0.1187404692,0.0392115042,-0.2071434706,0.4058914483,0.0351617336,-0.300288111,0.0039731245,-0.3729739785,0.009060137,0.1176952124,0.031693168,0.088812083,-0.0114984075,-0.0897939205,0.1707094759,0.3015867472,-0.2777350247,-0.0811748877,0.5380644202,-0.0165195428,-0.4363062978,0.406716764,0.0910997465,-0.265063405,-0.2602672875,-0.2303221077,-0.3599171042,-0.3782087266,-0.231896773,-0.0780696198,0.2442931384,-0.3475922346,0.212142989,0.0534669124,-0.032672137,0.3668103814,-0.4717726409,-0.1038790047,-0.0051591131,-0.2411713898,-0.0263197701,0.1795349717,0.0173060559,0.1448735297,0.0666748285,-0.2324116677,0.0188774355,-0.2115794867,0.0607994869,0.2579632103,0.0157741904,0.0331682377,-0.3173055649,0.0562631302,0.2354151905,-0.1383284181,-0.1442530304,-0.04354297,0.1712250113,0.1415520608,-0.0547130741,0.3798898458,-0.2556415498,-0.182508558,-0.0047949753,-0.028568672,0.3092058003,-0.0480012484,0.1978426129,0.2315911204,0.1624959707,-0.2306028605,0.1384638548,0.0822305158,0.4809704423,-0.0767595395,0.0964095816,0.0872657672,0.0559793003,-0.2521456778,0.0656088069,0.2322040498,-0.0088385735,0.2820625603,-0.4166627228,0.4334471226,0.1091429368,-0.0127714677,0.2232515961,-0.2425817698,-0.10637144,0.5650658607,0.1957122684,-0.2459315211,-0.0448807441,0.10908328,0.1616179049,-0.2950553894,-0.0200713277,0.1439014375,-0.3322841525,0.1799618304,0.0627885684,0.3500887156,-0.0705460608,0.2851703763,0.4285671711,-0.1124065295,0.158018291,0.0175332222,-0.0901254937,-0.0449869111,0.1519815028,0.0207853839,-0.2855881155,0.3917182088,0.3113312721,-0.0392871425,-0.5306161046,-0.0169720296,0.3337498903,-0.057357043,0.1362052113,0.4013494551,-0.1384161115,0.2527980506,-0.1455336511,-0.780780375,0.2933789492,-0.1241624728,0.199379921,0.044795461,-0.1824665815,0.1992432177,0.2737719417,-0.1125888899,-0.1241302788,0.0894679129,0.3351041675,-0.2027253509,-0.2588581741,0.2589485645,0.0043815561,-0.0228770245,-0.2204963863,0.1967337877,0.1613188982,0.1165984794,0.1213727444,0.0278691314,0.281105727,0.380027622,0.1551166177,0.0051991115,-0.0512977429,0.0308091771,0.2106678337,0.4682571888,0.5239808559,0.1957229376,0.3474487364,0.0808404014,-0.061709445,0.1058755741,0.0885473713,0.1906840354,-0.1326034814,0.2188703269,-0.0792564005,0.1500311196,-0.4193201363,0.0314085893,-0.2548205256,0.3473964334,0.5314039588,-0.0558549166,0.1173988879,-0.1215582266,0.0454474092,-0.0932245925,0.3865903914,0.240059495,0.4453108013,-0.2861812115,-0.1872860342,-0.6304110289,0.318595171,-0.3895478547,0.1897372156,-0.1382484883,0.020808626,0.3115709722,0.3615578413,-0.130706802,0.3410048187,0.1538964957,0.444005698,-0.0241616517,-0.0479540974,0.2248769999,0.0106653953,-0.1028805003,-0.4751072228,0.0300566461,-0.1046135724,0.0870180354,-0.1508562416,-0.0130428681,-0.0212868005,0.1397132725,0.4028874338,0.082702443,0.8158760667,-0.0312414486,-0.0269636363,-0.06477236,0.0436931103,-0.1682123691,0.2752511203,0.0953673571,0.0657407418,-0.0277018417,0.0265627801,-0.1445897371,0.4608074129,0.1300292462,0.164918229,-0.3377990127,-0.2344678342,-0.3548519015,-0.1299698353,-0.1278645396,-0.0809602588,-0.0189350042,-0.0265213288,0.0469351523,-0.402772516,0.5305594802,-0.5888613462,-0.1571630538,-0.0297350697,0.1984483749,-0.0414125323,-0.2330597937,-0.1098873019,0.2802364826,0.3407003284,-0.1172643974,-0.2922654748,0.0082283206,-0.0437997766,0.0204540771,-0.0709812641,0.7276521921,0.0326760039,-0.4151473343,-0.1617050767,-0.3914832771]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2677","title":"Error when downloading C4","comments":"@lhoestq thank you for such a quick feedback!","body":"Hi,\r\nI am trying to download `en` corpus from C4 dataset. However, I get an error caused by validation files download (see image). My code is very primitive:\r\n`datasets.load_dataset('c4', 'en')`\r\n\r\nIs this a bug or do I have some configurations missing on my server? \r\nThanks!\r\n\r\n\r\n\"\u0421\u043d\u0438\u043c\u043e\u043a","comment_length":8,"text":"Error when downloading C4\nHi,\r\nI am trying to download `en` corpus from C4 dataset. However, I get an error caused by validation files download (see image). My code is very primitive:\r\n`datasets.load_dataset('c4', 'en')`\r\n\r\nIs this a bug or do I have some configurations missing on my server? \r\nThanks!\r\n\r\n\r\n\"\u0421\u043d\u0438\u043c\u043e\u043a\n@lhoestq thank you for such a quick feedback!","embeddings":[0.061793033,-0.0860687941,-0.0479444414,0.448982805,0.2215530276,0.23021245,-0.0330859609,0.157870844,-0.0889108628,-0.0389548168,-0.0043446613,-0.2891196311,0.1426427066,0.0200259537,-0.0306665823,-0.1112219468,0.1224222854,0.1023219675,-0.1053356305,-0.323130399,-0.1176407859,0.0106312279,-0.1648942828,0.110300079,0.026287904,-0.0097798649,0.0003426506,0.1158432811,-0.3585418165,-0.3621957898,0.5530734062,0.0685187653,0.201746136,0.0887458697,-0.0001158856,-0.1960374564,0.297921598,-0.291267395,-0.152903825,-0.1510142684,-0.3564656079,-0.1293659508,-0.2348695695,-0.3450273871,-0.1182055771,-0.0631506965,0.1539894044,-0.1570960134,0.048681099,0.4164724648,0.1830768734,-0.1071507186,0.0137168737,-0.231496498,-0.201387763,-0.0305223744,0.0659266263,0.1705687046,0.1287416369,0.2572378516,0.3474600911,0.1605255306,0.0132957073,-0.0228848197,0.2810993791,-0.2411866188,0.1984723657,-0.3189607263,0.3368746042,0.2791511416,0.9787775874,-0.1083573401,-0.1444306374,0.3548652828,0.2134022862,-0.2010215074,0.3309968114,0.6487460136,-0.0826108456,0.160210669,-0.3288379908,-0.2522948682,-0.1675927639,0.2944012582,-0.1549027711,-0.1467129141,0.1106017604,-0.0849674568,0.1559390724,0.1889202744,0.0034310948,-0.3357762098,-0.2184631675,0.2485887259,-0.1503378153,-0.0072991811,-0.4373141229,0.4134985209,-0.0975915566,-0.035989441,-0.2706240118,-0.0362013839,-0.1111999974,0.0831497982,0.2124078125,-0.0668588057,0.147208944,0.1679933071,0.0820573419,-0.1831337363,0.162457943,-0.0558546558,-0.1577476114,-0.3487668633,-0.0345256217,0.1559273899,-0.1589246392,-0.4140318632,-0.1817277819,0.250492692,-0.33932212,0.0917578936,-0.0993224382,0.4581967592,-0.3990583122,-0.0729397014,-0.0744403377,0.1351937652,-0.0927351937,-0.5417319536,-0.11846672,0.0874820426,-0.3005711436,-0.2423428893,0.2378648221,-0.5039455891,0.2161420286,0.043105863,0.161673367,-0.0964144021,-0.1213472635,-0.3071010411,0.0634326413,0.2239746302,-0.1413739324,0.0777177513,0.252430886,-0.318120271,0.046370659,-0.0308066811,-0.3372936845,-0.1450243741,-0.2296413481,0.1240482852,-0.1476361006,-0.0292540174,-0.3287064135,-0.2533218861,0.2399029434,-0.1038007215,0.179541111,-0.2657217383,-0.018968232,-0.1591020077,-0.0753210783,0.3343384862,-0.3423299491,0.0376772508,-0.0311223995,-0.0832238421,0.2741538286,0.2242904007,-0.1951658428,0.2515410483,-0.2415113002,0.2791239917,0.0786271021,-0.5521119237,-0.51532197,0.2430380285,0.0717346966,-0.1472506672,-0.2335861325,-0.089446485,0.3533758521,-0.1477275938,-0.0257030353,0.102032043,0.0941109434,0.0379433744,-0.1204294264,0.0604737215,0.2159897536,0.002369106,0.2322340459,-0.098856017,0.2667962015,-0.1138730496,0.2359087467,-0.2393009514,0.176080361,0.2788185179,0.2383302897,-0.1242683902,0.077226229,0.2876140475,-0.3810850382,0.142351687,0.0107331965,0.0762486979,-0.0011442197,0.0394244641,-0.2413472682,-0.1907977313,-0.3075550497,-0.0608377643,0.028234452,-0.2354625762,0.0402419716,0.1156416982,0.1702338159,0.5983549356,-0.1753506511,0.0735518336,0.1153532639,0.2784141898,-0.1839047223,0.3289487362,-0.0176968351,-0.2116993666,0.2687041461,-0.1300704777,-0.1990704387,0.2145781219,0.2418289632,0.1999931186,-0.3405604959,-0.0006395292,0.3757499754,-0.4037655592,0.1474549919,0.4057971835,0.2090400457,-0.0111804679,0.0618812181,-0.0319071896,-0.0714937299,-0.2536828518,0.0315615125,0.0890021324,0.3211630285,-0.0234179869,-0.2340928316,-0.1785120219,0.3596225381,-0.313668102,0.0462786742,-0.0079218261,-0.1496249437,-0.3768742979,0.3946968317,0.0406965986,-0.0096956585,0.1591360867,-0.0168432202,0.0523296036,0.306360513,0.2030296028,0.4934886098,-0.0403367244,-0.257648766,0.3089373708,0.1298175603,-0.1598781943,0.3139909208,-0.2656512856,0.1835121661,0.0851710737,-0.0951695144,-0.1107219979,-0.3297852576,0.1406659484,0.3285274506,0.154752627,-0.2718478739,0.1393106133,-0.4375753403,-0.2050128579,-0.2197473049,-0.160150215,-0.077958636,0.0839445665,-0.0291766245,-0.0515235998,-0.1238136962,-0.0453554913,-0.6379559636,-0.063850902,0.0033960871,-0.1954630762,0.060765259,0.18819125,-0.1948238462,0.0129040759,0.1117704511,0.0700679794,0.3571193516,-0.3355754912,0.2499948442,-0.3134472668,-0.2200897336,0.225077495,0.1241410524,-0.1784747541,-0.0144987265,0.2775062621,-0.4003030658,-0.1126248762,-0.0480323657,0.0182880256,-0.1763126105,0.1902129799,0.1043609753,-0.1778399497,-0.1307142079,-0.5133389235,-0.420014739,-0.0903046131,-0.1963710189,-0.2389082611,0.0292250384,0.4292242229,0.0397744775,-0.0888050869,0.1379049867,-0.1719043702,-0.3215815723,-0.0751597136,0.3428555429,-0.2622688115,-0.5823302269,0.0917848423,0.0202938188,0.2089447528,0.3315109909,-0.3047239184,0.377053082,0.1170597151,0.0817025974,-0.0305141564,-0.236500293,0.4115386307,-0.1906617135,0.1037766039,-0.0897646323,0.2673054338,0.1189260185,0.1689331383,0.5265540481,-0.4418097734,-0.0138422819,0.171447143,0.1718893945,0.0195273794,-0.302093327,0.5654307008,0.2541282475,0.3535559773,0.2548460066,-0.3889175057,0.3400193751,-0.1631122977,-0.0758549348,0.2882586718,-0.2203367949,-0.3463083804,-0.2030228078,0.0251469072,-0.30413872,-0.1341252923,-0.1127750203,-0.0540085845,0.3440546691,0.337962687,0.3910868168,0.2379888892,-0.2618125975,0.0389099345,0.3513302505,0.1129975021,0.0674337745,-0.1756090224,-0.1957666427,-0.0044675088,0.3184635341,0.3720580041,0.2269849926,-0.0122782988,0.3328892291,-0.0246064551,-0.0068683648,0.5469513535,-0.3316878378,0.1516066194,-0.1678111553,0.2929637134,-0.2166442573,-0.238187328,-0.0667239055,0.0023411547,0.2599190474,0.2279821783,-0.2823282182,0.1462491006,-0.1359122545,0.2812462747,-0.0674062446,-0.3850528896,-0.1707847267,-0.3663870692,-0.2679974139,0.2509695292,-0.0149657503,0.1729479879,-0.1361280829,0.1859411895,0.056435816,0.3840178251,0.1817738712,0.0641247109,0.2636218369,0.0188541654,-0.0355414264,-0.0235112384,0.0389170758,0.6832143068,0.2889851928,-0.1421404332,-0.3107333779,-0.1467391104,-0.0891540274,0.206872955,0.0847630799,0.0420360379,-0.1791973263,0.2700235844,-0.2854351699,0.0344001055,0.3133345246,0.1299700588,0.0094489083,-0.6830766797,-0.408131063,0.041494675,-0.1695488691,0.0794283822,0.1271471083,0.2720929086,-0.1582925767,-0.214627713,0.2578505874,0.8839765191,0.1201632023,-0.1377265155,-0.0055365167,-0.1022127867,0.1310428381,-0.6483941078,0.0752729475,-0.2543719709,-0.4459970593,-0.2387423217,-0.1273019761,0.3382916451,0.4829192162,-0.2340260297,0.334046334,0.0745308921,0.3073465824,0.1232575104,0.0926441029,0.0427180119,-0.2260013968,-0.36770913,0.1503975838,-0.250657469,0.348028928,-0.3459722102,-0.051753331,-0.0529068746,-0.3507084846,-0.2708226144,0.3497807086,0.0319420248,0.1055724174,0.0227170624,-0.0716295913,0.4009978175,0.3297452033,0.3512077332,0.6259921193,-0.1110822111,0.1285103858,0.1160710305,-0.2537970543,0.0313491076,-0.1805157363,0.1290529221,-0.2139514685,-0.2606276274,0.1519829631,0.0679557025,-0.1866689622,0.3972842097,0.0349716023,-0.3882435262,-0.0579943806,-0.3930203617,-0.0369865745,0.1220059693,0.0647633895,0.0654463246,-0.0072424076,-0.1034830064,0.1688363254,0.3306130469,-0.2705644071,-0.0784927607,0.5468775034,-0.0162950568,-0.4279412031,0.4092590511,0.0641053021,-0.2503783703,-0.2655463219,-0.2549596727,-0.2916479707,-0.3708844781,-0.21769768,-0.068881318,0.2159407437,-0.3092800677,0.2005383521,0.0609992743,-0.0503341183,0.3529225588,-0.4477879107,-0.1201916635,0.0010904372,-0.1915726811,-0.0175480973,0.1768988818,-0.0191283394,0.1160636917,0.0776636079,-0.2125132829,0.0362027064,-0.1920768321,0.09290196,0.2502324283,0.0139188375,-0.0190984737,-0.3128298223,0.0507648997,0.2164537907,-0.121300213,-0.124783203,-0.0560708828,0.1787258536,0.1355506182,-0.0302759409,0.3259413838,-0.2540028989,-0.1473353654,0.0072546648,-0.0330965333,0.2946283519,-0.0339699909,0.2394437343,0.2651593387,0.1014432013,-0.2220606804,0.1322997659,0.0974686593,0.4753218591,-0.0568394102,0.0945643857,0.113960743,0.0484808423,-0.2590867877,0.0870016813,0.2856717706,-0.0678805187,0.2868306935,-0.4372898638,0.414825201,0.0895021856,0.0327927098,0.2146599889,-0.2378216833,-0.0892557427,0.5437760353,0.1862935275,-0.2474372685,-0.040280249,0.1428750902,0.1573577821,-0.2544677258,-0.0222209468,0.1325211078,-0.2841832042,0.1643624008,0.065027453,0.3870403767,-0.0519740917,0.2850809693,0.4290046394,-0.1155670285,0.1540479064,0.06790144,-0.1084373593,-0.0721918643,0.1190348268,0.0827298537,-0.2359780073,0.3868058324,0.3226931691,0.0043415632,-0.4696398675,-0.0599855483,0.3564190865,-0.0757508129,0.1628065258,0.400480628,-0.131282106,0.280667156,-0.2020101994,-0.7752689719,0.2807654142,-0.1118871719,0.1898418665,0.00367175,-0.2083560079,0.2204443812,0.3259847462,-0.13484402,-0.0885859132,0.1156248823,0.3094391525,-0.1769857854,-0.2432174981,0.3013751507,0.0469480306,-0.0082634222,-0.2193351835,0.2264680117,0.1919235885,0.0994670242,0.127913937,0.0134939477,0.2410575897,0.3891017437,0.1815999746,-0.0087515507,-0.0516346097,-0.0122014359,0.2067151219,0.4715600312,0.5478778481,0.1572024375,0.3482261598,0.0772710443,-0.0450771786,0.1518105865,0.0780804008,0.1895386279,-0.1443981826,0.1798229367,-0.0509747192,0.0969415903,-0.4217994809,-0.0005541774,-0.2637136579,0.3475390673,0.5328541398,-0.0414373912,0.1269885898,-0.1195239723,0.0287733078,-0.0405018516,0.4333681464,0.2203720063,0.4670423567,-0.3202743232,-0.2354932725,-0.6198289394,0.3291143477,-0.3924138844,0.2170561701,-0.1537842304,-0.0217916984,0.3249185681,0.3477237523,-0.1196467578,0.3593724668,0.1598127484,0.5093191266,-0.0366099849,-0.0136881899,0.2184744775,0.0494544841,-0.0992553532,-0.4432501793,0.0374353342,-0.0969335884,0.0854400918,-0.1965255439,0.0297669042,-0.0916981846,0.1536136419,0.3702183068,0.0529956743,0.8496043682,-0.0349790379,-0.0088776294,-0.0347436927,0.0949283689,-0.1796005815,0.3021279573,0.1125744209,0.0731272474,0.016194215,0.026113024,-0.1052408144,0.4595661759,0.1444176733,0.1321722418,-0.3057355285,-0.2582826316,-0.3193020821,-0.2188503295,-0.1401188523,-0.0832340419,-0.0117529863,-0.0306287464,0.024981007,-0.3807771802,0.5199109912,-0.5765818954,-0.1807070673,-0.0550878271,0.1669806093,-0.0493698791,-0.22153157,-0.181013003,0.2935829163,0.3462524116,-0.1344404072,-0.312897563,0.0457644537,-0.0014580524,-0.0322759002,-0.0641572922,0.7480258346,0.0406709462,-0.456993103,-0.1798424125,-0.3427414894]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2669","title":"Metric kwargs are not passed to underlying external metric f1_score","comments":"Hi @BramVanroy, thanks for reporting.\r\n\r\nFirst, note that `\"min\"` is not an allowed value for `average`. According to scikit-learn [documentation](https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html), `average` can only take the values: `{\"micro\", \"macro\", \"samples\", \"weighted\", \"binary\"} or None, default=\"binary\"`.\r\n\r\nSecond, you should take into account that all additional metric-specific argument should be passed in the method `compute` (and not in the method `load_metric`). You can find more information in our documentation: https:\/\/huggingface.co\/docs\/datasets\/using_metrics.html#computing-the-metric-scores\r\n\r\nSo for example, if you would like to calculate the macro-averaged F1 score, you should use:\r\n```python\r\nimport datasets\r\n\r\nf1 = datasets.load_metric(\"f1\", keep_in_memory=True)\r\nf1.add_batch(predictions=[0,2,3], references=[1, 2, 3])\r\nf1.compute(average=\"macro\")\r\n```","body":"## Describe the bug\r\nWhen I want to use F1 score with average=\"min\", this keyword argument does not seem to be passed through to the underlying sklearn metric. This is evident because [sklearn](https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html) throws an error telling me so.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\nf1 = datasets.load_metric(\"f1\", keep_in_memory=True, average=\"min\")\r\nf1.add_batch(predictions=[0,2,3], references=[1, 2, 3])\r\nf1.compute()\r\n```\r\n\r\n## Expected results\r\nNo error, because `average=\"min\"` should be passed correctly to f1_score in sklearn.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\datasets\\metric.py\", line 402, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\metrics\\f1\\82177930a325d4c28342bba0f116d73f6d92fb0c44cd67be32a07c1262b61cfe\\f1.py\", line 97, in _compute\r\n \"f1\": f1_score(\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1071, in f1_score\r\n return fbeta_score(y_true, y_pred, beta=1, labels=labels,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1195, in fbeta_score\r\n _, _, f, _ = precision_recall_fscore_support(y_true, y_pred,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1464, in precision_recall_fscore_support\r\n labels = _check_set_wise_labels(y_true, y_pred, average, labels,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1294, in _check_set_wise_labels\r\n raise ValueError(\"Target is %s but average='binary'. Please \"\r\nValueError: Target is multiclass but average='binary'. Please choose another average setting, one of [None, 'micro', 'macro', 'weighted'].\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.9.2\r\n- PyArrow version: 4.0.1","comment_length":96,"text":"Metric kwargs are not passed to underlying external metric f1_score\n## Describe the bug\r\nWhen I want to use F1 score with average=\"min\", this keyword argument does not seem to be passed through to the underlying sklearn metric. This is evident because [sklearn](https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html) throws an error telling me so.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\nf1 = datasets.load_metric(\"f1\", keep_in_memory=True, average=\"min\")\r\nf1.add_batch(predictions=[0,2,3], references=[1, 2, 3])\r\nf1.compute()\r\n```\r\n\r\n## Expected results\r\nNo error, because `average=\"min\"` should be passed correctly to f1_score in sklearn.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\datasets\\metric.py\", line 402, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\metrics\\f1\\82177930a325d4c28342bba0f116d73f6d92fb0c44cd67be32a07c1262b61cfe\\f1.py\", line 97, in _compute\r\n \"f1\": f1_score(\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1071, in f1_score\r\n return fbeta_score(y_true, y_pred, beta=1, labels=labels,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1195, in fbeta_score\r\n _, _, f, _ = precision_recall_fscore_support(y_true, y_pred,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1464, in precision_recall_fscore_support\r\n labels = _check_set_wise_labels(y_true, y_pred, average, labels,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1294, in _check_set_wise_labels\r\n raise ValueError(\"Target is %s but average='binary'. Please \"\r\nValueError: Target is multiclass but average='binary'. Please choose another average setting, one of [None, 'micro', 'macro', 'weighted'].\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.9.2\r\n- PyArrow version: 4.0.1\nHi @BramVanroy, thanks for reporting.\r\n\r\nFirst, note that `\"min\"` is not an allowed value for `average`. According to scikit-learn [documentation](https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html), `average` can only take the values: `{\"micro\", \"macro\", \"samples\", \"weighted\", \"binary\"} or None, default=\"binary\"`.\r\n\r\nSecond, you should take into account that all additional metric-specific argument should be passed in the method `compute` (and not in the method `load_metric`). You can find more information in our documentation: https:\/\/huggingface.co\/docs\/datasets\/using_metrics.html#computing-the-metric-scores\r\n\r\nSo for example, if you would like to calculate the macro-averaged F1 score, you should use:\r\n```python\r\nimport datasets\r\n\r\nf1 = datasets.load_metric(\"f1\", keep_in_memory=True)\r\nf1.add_batch(predictions=[0,2,3], references=[1, 2, 3])\r\nf1.compute(average=\"macro\")\r\n```","embeddings":[-0.0516288541,-0.5501691103,0.0873140618,0.2038566321,0.4681911767,-0.0643797889,0.1748601645,-0.1953885257,0.3451755941,0.2828977704,0.0679134801,0.4462169111,0.1783719808,0.0242547859,-0.0182039514,0.0616982467,0.1058040708,-0.3587132096,-0.0846228451,-0.2987218797,-0.4068837166,0.0102044912,-0.4048623741,0.0202550888,-0.2006222904,0.0613628998,-0.0347853079,-0.1700376719,0.0016798849,-0.003900961,0.3077568412,0.1265656352,-0.0970672965,0.3663867116,-0.0001235781,-0.0986129567,0.4526257515,-0.1368988007,-0.0371827036,-0.0111093875,-0.1764513552,0.014776093,-0.1244774237,-0.3509167433,0.1918561906,0.1058521941,-0.1739191711,-0.3824925721,0.2338166088,0.3685019314,0.0636828318,0.1573851407,-0.2092087269,-0.1720214784,-0.2959969938,-0.1044454128,-0.1850389093,0.4786297977,-0.6841754317,-0.0733038336,-0.1676173359,0.0917701647,0.1501622796,-0.0126522053,0.7468219995,0.0399315543,0.3019666672,0.1136633381,0.1792099476,-0.1155982167,0.0679363832,-0.3869571686,-0.403575331,-0.1462424695,0.0940086693,-0.7693849206,0.1077826694,0.0144160725,-0.2267361879,-0.092972286,-0.1046825647,-0.1168835163,-0.3561751246,0.0764986202,-0.050303366,0.6123213172,0.0015167097,0.0829183236,0.192719996,0.1838299781,0.0406339429,-0.2930784225,0.0284065641,0.071691215,-0.2631740272,0.2108209878,-0.0395305119,-0.0636173189,0.27936849,0.0515006147,-0.0760850012,0.0437829681,0.430503428,0.0532589592,0.197408691,0.5313013792,0.4781782627,0.3300953805,-0.0164008401,0.0775587261,-0.1973581165,-0.2480203807,0.271512419,-0.0058645201,0.2495119125,0.1664949208,0.0198457837,-0.07760185,-0.2983479202,-0.0815996081,0.0488367118,-0.013127977,0.0346073136,0.159644261,-0.0479189232,0.3419955671,0.0582272336,0.3695012033,-0.112433292,-0.0151641183,-0.2224202901,0.1949929297,-0.4361763597,0.1441707015,-0.2022401094,-0.1239357069,0.5087715983,-0.0023832025,0.606444478,-0.4638635218,-0.0930645168,0.0385823548,-0.0192406494,-0.012212811,-0.4288246334,0.2412556112,0.033799082,0.0614981018,0.0438620262,-0.0164111089,-0.3245539367,-0.0843018368,0.4818834066,0.041315563,-0.2956653833,-0.132889986,-0.1859076768,0.2178625315,0.0208467692,0.114158757,0.0367673673,-0.1453685611,-0.5320333838,-0.2027044743,0.7653902769,0.0672897398,0.1244359612,-0.2514540553,-0.1332684457,0.0805685669,-0.0005603431,0.0552252233,0.0327283815,0.2792529762,-0.0457489416,0.0623041987,0.2406337559,-0.6767405272,-0.4408147037,0.0251210853,0.046009697,-0.0332321338,0.2000237256,-0.1751340479,0.3521326184,-0.0047277925,0.1999831498,0.430185914,-0.1823700368,-0.0675364807,-0.421084106,-0.0479820445,0.048952315,-0.1104324088,0.4257763326,0.0662507042,0.0844967216,0.2618041635,0.0360832065,0.1915479749,-0.1473145038,0.0495125614,0.0701558739,-0.3530410528,0.0856516659,0.055507049,-0.1457139701,0.1035256162,-0.1434553862,-0.168143332,0.1927152425,-0.3029265404,-0.7353894711,0.2379848659,-0.0126784286,-0.1675952822,0.027924424,-0.0007104453,-0.1402493715,0.2118350416,0.1246928424,-0.2877166867,-0.2155011594,0.0689764768,0.2514443398,0.0879586414,-0.0275266338,-0.1178222597,-0.0540140606,0.26175946,0.4194566607,0.0281710755,-0.0429728068,0.5366603136,0.2259910554,0.1409419775,0.1608157605,0.0340641066,0.24411273,-0.0328771435,0.0478120819,0.3331083655,0.1005686298,-0.1979617476,-0.0977448374,0.4207776785,0.1251930445,0.3327291906,-0.0869361833,0.2726321518,-0.2567584515,0.0971228853,-0.1023771539,-0.2068930119,-0.0779832676,-0.4906929433,-0.0203643702,-0.2348927855,0.0249764565,0.1215319857,0.4332318902,0.1995496154,0.1523185372,-0.076057896,-0.2277942002,-0.0624574721,0.0835695267,-0.1111955717,0.4364539385,0.1562025398,0.0343498029,0.2055181116,-0.1461104602,-0.095109567,0.1814320683,0.4817385674,-0.0805645362,0.2435981035,-0.0937529802,-0.1131483465,-0.0635224879,0.2218562812,0.0521538183,-0.0792746693,-0.5283737779,0.2364412844,-0.0392129347,0.2011196017,-0.2159098983,0.0986009464,-0.1711424738,-0.7540907264,-0.1033217013,0.0821495727,-0.1489276886,0.3635009825,-0.2049851418,0.2673458159,-0.0496508144,0.0223412737,0.036626853,-0.3193083107,-0.1708577275,-0.1480315328,0.2124335468,-0.0287868083,0.3443077803,0.3873269558,0.2164435089,0.0731538311,-0.4658792913,-0.1431510746,0.081477575,0.918420434,-0.0383187644,-0.1673643887,-0.3328225613,-0.0775397047,0.354608506,-0.2792412341,0.0753313825,0.1667623371,0.0544317216,0.1211809143,-0.3318283856,-0.1408891678,0.0599886477,-0.1981328428,-0.1365239322,-0.1284700185,-0.2889370024,0.3429214656,0.1867969185,0.0009099875,-0.0350905024,-0.0836620182,-0.097604759,0.0677944645,0.3448297381,-0.1971108466,-0.2251143754,0.0398046635,-0.0092014624,0.2385275364,-0.0336470567,-0.3195250928,-0.4439285696,0.2071544677,0.1057022214,0.0439924188,0.1653718948,-0.0892879516,-0.4593803585,0.1171314344,-0.2870640457,-0.219100967,0.3560729325,-0.2641616166,0.1047707945,-0.1369574219,0.3790746927,0.2572436631,0.6966174245,0.0503653102,0.1582950801,-0.0299590658,0.0022024186,0.2247810364,0.0723269656,-0.3129124343,0.1803696454,0.0302197263,-0.1768461466,0.2460018247,0.1162326187,0.2928936779,-0.0578404367,-0.3579824567,-0.2600935698,-0.3655110598,-0.1559731811,-0.0219270606,0.2030694187,0.1455339044,0.5627012253,-0.4348830283,-0.0822068229,0.3034934998,-0.1214675456,-0.026632091,0.0871555507,-0.2401306182,-0.3764371574,-0.1923998743,0.169432506,-0.1273618788,0.4058916271,0.0894154757,0.2031165957,0.1276355237,-0.1006715149,0.2976213396,-0.2545888424,0.1782376915,0.188301146,0.1651884317,-0.3848178983,-0.1694618165,-0.1441166401,-0.1351506412,0.2666935921,0.0691986829,-0.5115234256,-0.1317024082,0.1223711744,-0.0508822687,-0.2466030419,-0.1446665525,-0.2461933047,-0.2586558759,0.1532775164,0.0380989574,0.1500681639,0.3929439187,-0.0927803889,0.5734653473,-0.1120118648,-0.2354331166,0.2397040278,-0.0630112514,0.55297333,-0.019410111,0.0340885743,0.122265406,-0.1265092641,-0.2053155899,0.1226585954,0.1356473714,-0.3416363299,-0.1437759697,-0.3832884729,0.5200098157,0.2293066233,-0.2753112316,0.1476996839,-0.3316442668,-0.1977547258,-0.01063303,-0.0570235774,0.2516385615,-0.0470467024,0.117916353,-0.2057297826,0.53506428,0.2149329782,-0.1624639779,0.4008515775,0.1537148505,-0.4334540963,0.1491287351,0.3389035463,0.7175228,0.4781146049,-0.0385890417,0.4263069034,-0.2586626112,0.1159346029,0.0563809834,-0.004118057,-0.2077731639,-0.3220458925,0.1919017732,0.0116220526,0.105819352,0.0006326091,-0.0864815414,0.1789881289,-0.2033778429,0.386985451,0.1023421884,0.2675025761,0.0608242601,0.0337739214,-0.1049404293,0.0172547027,0.0907756463,-0.1177081317,0.0535145216,-0.3840714991,0.0313193277,-0.1755132377,-0.0016678156,0.052239202,-0.2561365664,0.0642953962,0.1166628972,-0.000884136,0.1194904894,0.0016806067,0.5374199748,-0.035936296,-0.1339445859,0.2569860518,-0.4386573136,0.0324886031,0.1740621477,0.1364796907,0.0406217873,-0.0596820042,-0.2285199016,0.3244857788,0.0586300716,-0.1731661111,-0.2100576758,-0.2583572268,-0.2615670562,-0.1521653831,-0.2196034491,-0.0531676374,-0.0192706417,-0.1628288925,0.0028900355,-0.1240886897,0.0244238637,0.4010928869,-0.1081084684,-0.1133654341,0.0903801695,0.3428358436,-0.2323462665,-0.2853524685,0.4727140963,-0.3125245571,-0.0826675519,-0.0505311377,0.174883455,-0.1998937875,-0.2814690769,0.2686016858,-0.0161281656,0.1039671749,0.0345945507,0.404282093,0.1158282235,-0.1011151522,-0.0572219491,-0.4350415468,0.1116835922,0.408542186,-0.2350719571,-0.0952422991,-0.4258363247,0.1973529905,0.077968359,0.1975916922,-0.2119119465,-0.2755135596,0.1248587742,-0.1465166807,0.267093569,0.0946127996,0.045810055,0.0242839176,-0.0631357431,0.0884652212,-0.0965996236,-0.0297743138,-0.2549293935,0.168671295,0.031317763,0.3304584324,0.087002039,-0.0451474674,-0.1597486585,-0.2132950574,0.0232998449,0.1865111142,-0.1165359169,0.5392251015,0.1041021645,0.8976370692,0.2134967893,-0.053431619,0.2053504586,0.2913578451,-0.0549782775,0.1057695076,0.313290447,0.136546433,-0.2882294655,-0.0891074315,-0.3128887713,0.0653656051,0.2927876711,-0.1010156274,0.0120883835,0.4501214027,0.0216898844,0.2443935424,-0.1739862859,0.0332906954,-0.3316778243,0.108799845,-0.1615259051,0.1517754495,0.1381608844,0.1890217066,0.0372141413,0.2193840146,0.3896560371,-0.1826103628,0.1538563818,0.0071755741,0.2480259389,-0.2777958214,0.3476074338,0.1518511772,0.2368544191,0.2364320755,0.278834939,0.0636553392,0.0487269908,0.4418359697,0.1506568491,0.2316164076,0.1368170083,-0.0283415467,-0.1424889117,-0.0037023097,-0.057564225,0.0473345183,-0.0968374386,0.1636189669,-0.3007003665,0.2142596692,-0.0032683706,-0.3228293955,-0.2563185096,-0.2440361679,-0.0155125642,-0.2392959148,-0.5574415922,-0.2417226434,-0.0506323054,-0.1181863248,0.1758486778,-0.0051929546,0.2497843951,-0.2485955507,-0.1441931427,0.2668414414,-0.26188007,0.3737464547,0.5232933164,-0.3483285308,0.0587759279,-0.1005220339,0.2709252238,0.21099931,0.3054180145,0.2390975207,0.3020229638,-0.3668431342,-0.2365858406,0.1456976682,-0.2033136785,0.114478454,0.3290289938,0.2814850509,-0.2071492225,0.202794373,0.0467125736,-0.1556740701,0.1896248013,0.2759167552,0.2000056803,-0.4089815021,0.3424044251,-0.0396012291,-0.0322977342,0.1221435815,-0.1428463608,-0.042736847,-0.2502233982,0.0933625996,0.2045282722,-0.0232299678,-0.5147492886,0.0396317132,0.0295872055,0.3810938299,0.0585155897,0.0123011433,-0.4914895594,-0.0101353619,-0.6506745219,-0.0930294916,0.2496894598,-0.1495498121,0.1456627101,0.1661141962,-0.0759351775,0.0700000003,0.0718906522,0.5094356537,0.1752088219,-0.0033009178,-0.4087581933,-0.1453213394,-0.2016129643,-0.2168287337,0.2321971655,-0.1759064496,-0.0097852089,-0.3974433839,-0.0150046283,0.0600623563,0.0033534041,0.1645422727,0.6511238813,0.1873810589,0.0206307843,0.7741081119,0.0136421323,0.0299884081,-0.4132049978,-0.1902291775,-0.2227948457,-0.0953332484,-0.1362854838,0.3383519053,-0.0151670501,-0.4483858347,-0.0393702649,0.211392343,0.3170768023,-0.4165564477,-0.0193999615,0.3332815766,-0.1906142682,-0.0921211764,0.0234425347,0.3206248879,-0.0995698422,0.2311321348,0.0546547771,-0.2851710618,0.2310818583,-0.3407650292,-0.3436341882,-0.2862753868,0.3049652874,-0.0992926285,-0.1729005277,-0.6666846275,-0.1416708827,0.2823508978,0.1122600064,-0.0339417011,0.2595966458,-0.0015445829,0.0506242812,-0.0288966801,0.0143096223,-0.1536698639,-0.1183005571,0.080745019,-0.1149214879]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2669","title":"Metric kwargs are not passed to underlying external metric f1_score","comments":"Thanks, that was it. A bit strange though, since `load_metric` had an argument `metric_init_kwargs`. I assume that that's for specific initialisation arguments whereas `average` is for the function itself.","body":"## Describe the bug\r\nWhen I want to use F1 score with average=\"min\", this keyword argument does not seem to be passed through to the underlying sklearn metric. This is evident because [sklearn](https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html) throws an error telling me so.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\nf1 = datasets.load_metric(\"f1\", keep_in_memory=True, average=\"min\")\r\nf1.add_batch(predictions=[0,2,3], references=[1, 2, 3])\r\nf1.compute()\r\n```\r\n\r\n## Expected results\r\nNo error, because `average=\"min\"` should be passed correctly to f1_score in sklearn.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\datasets\\metric.py\", line 402, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\metrics\\f1\\82177930a325d4c28342bba0f116d73f6d92fb0c44cd67be32a07c1262b61cfe\\f1.py\", line 97, in _compute\r\n \"f1\": f1_score(\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1071, in f1_score\r\n return fbeta_score(y_true, y_pred, beta=1, labels=labels,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1195, in fbeta_score\r\n _, _, f, _ = precision_recall_fscore_support(y_true, y_pred,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1464, in precision_recall_fscore_support\r\n labels = _check_set_wise_labels(y_true, y_pred, average, labels,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1294, in _check_set_wise_labels\r\n raise ValueError(\"Target is %s but average='binary'. Please \"\r\nValueError: Target is multiclass but average='binary'. Please choose another average setting, one of [None, 'micro', 'macro', 'weighted'].\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.9.2\r\n- PyArrow version: 4.0.1","comment_length":29,"text":"Metric kwargs are not passed to underlying external metric f1_score\n## Describe the bug\r\nWhen I want to use F1 score with average=\"min\", this keyword argument does not seem to be passed through to the underlying sklearn metric. This is evident because [sklearn](https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html) throws an error telling me so.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\nf1 = datasets.load_metric(\"f1\", keep_in_memory=True, average=\"min\")\r\nf1.add_batch(predictions=[0,2,3], references=[1, 2, 3])\r\nf1.compute()\r\n```\r\n\r\n## Expected results\r\nNo error, because `average=\"min\"` should be passed correctly to f1_score in sklearn.\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\datasets\\metric.py\", line 402, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\metrics\\f1\\82177930a325d4c28342bba0f116d73f6d92fb0c44cd67be32a07c1262b61cfe\\f1.py\", line 97, in _compute\r\n \"f1\": f1_score(\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1071, in f1_score\r\n return fbeta_score(y_true, y_pred, beta=1, labels=labels,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1195, in fbeta_score\r\n _, _, f, _ = precision_recall_fscore_support(y_true, y_pred,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\utils\\validation.py\", line 63, in inner_f\r\n return f(*args, **kwargs)\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1464, in precision_recall_fscore_support\r\n labels = _check_set_wise_labels(y_true, y_pred, average, labels,\r\n File \"C:\\Users\\bramv\\.virtualenvs\\pipeline-TpEsXVex\\lib\\site-packages\\sklearn\\metrics\\_classification.py\", line 1294, in _check_set_wise_labels\r\n raise ValueError(\"Target is %s but average='binary'. Please \"\r\nValueError: Target is multiclass but average='binary'. Please choose another average setting, one of [None, 'micro', 'macro', 'weighted'].\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.9.2\r\n- PyArrow version: 4.0.1\nThanks, that was it. A bit strange though, since `load_metric` had an argument `metric_init_kwargs`. I assume that that's for specific initialisation arguments whereas `average` is for the function itself.","embeddings":[-0.0516288541,-0.5501691103,0.0873140618,0.2038566321,0.4681911767,-0.0643797889,0.1748601645,-0.1953885257,0.3451755941,0.2828977704,0.0679134801,0.4462169111,0.1783719808,0.0242547859,-0.0182039514,0.0616982467,0.1058040708,-0.3587132096,-0.0846228451,-0.2987218797,-0.4068837166,0.0102044912,-0.4048623741,0.0202550888,-0.2006222904,0.0613628998,-0.0347853079,-0.1700376719,0.0016798849,-0.003900961,0.3077568412,0.1265656352,-0.0970672965,0.3663867116,-0.0001235781,-0.0986129567,0.4526257515,-0.1368988007,-0.0371827036,-0.0111093875,-0.1764513552,0.014776093,-0.1244774237,-0.3509167433,0.1918561906,0.1058521941,-0.1739191711,-0.3824925721,0.2338166088,0.3685019314,0.0636828318,0.1573851407,-0.2092087269,-0.1720214784,-0.2959969938,-0.1044454128,-0.1850389093,0.4786297977,-0.6841754317,-0.0733038336,-0.1676173359,0.0917701647,0.1501622796,-0.0126522053,0.7468219995,0.0399315543,0.3019666672,0.1136633381,0.1792099476,-0.1155982167,0.0679363832,-0.3869571686,-0.403575331,-0.1462424695,0.0940086693,-0.7693849206,0.1077826694,0.0144160725,-0.2267361879,-0.092972286,-0.1046825647,-0.1168835163,-0.3561751246,0.0764986202,-0.050303366,0.6123213172,0.0015167097,0.0829183236,0.192719996,0.1838299781,0.0406339429,-0.2930784225,0.0284065641,0.071691215,-0.2631740272,0.2108209878,-0.0395305119,-0.0636173189,0.27936849,0.0515006147,-0.0760850012,0.0437829681,0.430503428,0.0532589592,0.197408691,0.5313013792,0.4781782627,0.3300953805,-0.0164008401,0.0775587261,-0.1973581165,-0.2480203807,0.271512419,-0.0058645201,0.2495119125,0.1664949208,0.0198457837,-0.07760185,-0.2983479202,-0.0815996081,0.0488367118,-0.013127977,0.0346073136,0.159644261,-0.0479189232,0.3419955671,0.0582272336,0.3695012033,-0.112433292,-0.0151641183,-0.2224202901,0.1949929297,-0.4361763597,0.1441707015,-0.2022401094,-0.1239357069,0.5087715983,-0.0023832025,0.606444478,-0.4638635218,-0.0930645168,0.0385823548,-0.0192406494,-0.012212811,-0.4288246334,0.2412556112,0.033799082,0.0614981018,0.0438620262,-0.0164111089,-0.3245539367,-0.0843018368,0.4818834066,0.041315563,-0.2956653833,-0.132889986,-0.1859076768,0.2178625315,0.0208467692,0.114158757,0.0367673673,-0.1453685611,-0.5320333838,-0.2027044743,0.7653902769,0.0672897398,0.1244359612,-0.2514540553,-0.1332684457,0.0805685669,-0.0005603431,0.0552252233,0.0327283815,0.2792529762,-0.0457489416,0.0623041987,0.2406337559,-0.6767405272,-0.4408147037,0.0251210853,0.046009697,-0.0332321338,0.2000237256,-0.1751340479,0.3521326184,-0.0047277925,0.1999831498,0.430185914,-0.1823700368,-0.0675364807,-0.421084106,-0.0479820445,0.048952315,-0.1104324088,0.4257763326,0.0662507042,0.0844967216,0.2618041635,0.0360832065,0.1915479749,-0.1473145038,0.0495125614,0.0701558739,-0.3530410528,0.0856516659,0.055507049,-0.1457139701,0.1035256162,-0.1434553862,-0.168143332,0.1927152425,-0.3029265404,-0.7353894711,0.2379848659,-0.0126784286,-0.1675952822,0.027924424,-0.0007104453,-0.1402493715,0.2118350416,0.1246928424,-0.2877166867,-0.2155011594,0.0689764768,0.2514443398,0.0879586414,-0.0275266338,-0.1178222597,-0.0540140606,0.26175946,0.4194566607,0.0281710755,-0.0429728068,0.5366603136,0.2259910554,0.1409419775,0.1608157605,0.0340641066,0.24411273,-0.0328771435,0.0478120819,0.3331083655,0.1005686298,-0.1979617476,-0.0977448374,0.4207776785,0.1251930445,0.3327291906,-0.0869361833,0.2726321518,-0.2567584515,0.0971228853,-0.1023771539,-0.2068930119,-0.0779832676,-0.4906929433,-0.0203643702,-0.2348927855,0.0249764565,0.1215319857,0.4332318902,0.1995496154,0.1523185372,-0.076057896,-0.2277942002,-0.0624574721,0.0835695267,-0.1111955717,0.4364539385,0.1562025398,0.0343498029,0.2055181116,-0.1461104602,-0.095109567,0.1814320683,0.4817385674,-0.0805645362,0.2435981035,-0.0937529802,-0.1131483465,-0.0635224879,0.2218562812,0.0521538183,-0.0792746693,-0.5283737779,0.2364412844,-0.0392129347,0.2011196017,-0.2159098983,0.0986009464,-0.1711424738,-0.7540907264,-0.1033217013,0.0821495727,-0.1489276886,0.3635009825,-0.2049851418,0.2673458159,-0.0496508144,0.0223412737,0.036626853,-0.3193083107,-0.1708577275,-0.1480315328,0.2124335468,-0.0287868083,0.3443077803,0.3873269558,0.2164435089,0.0731538311,-0.4658792913,-0.1431510746,0.081477575,0.918420434,-0.0383187644,-0.1673643887,-0.3328225613,-0.0775397047,0.354608506,-0.2792412341,0.0753313825,0.1667623371,0.0544317216,0.1211809143,-0.3318283856,-0.1408891678,0.0599886477,-0.1981328428,-0.1365239322,-0.1284700185,-0.2889370024,0.3429214656,0.1867969185,0.0009099875,-0.0350905024,-0.0836620182,-0.097604759,0.0677944645,0.3448297381,-0.1971108466,-0.2251143754,0.0398046635,-0.0092014624,0.2385275364,-0.0336470567,-0.3195250928,-0.4439285696,0.2071544677,0.1057022214,0.0439924188,0.1653718948,-0.0892879516,-0.4593803585,0.1171314344,-0.2870640457,-0.219100967,0.3560729325,-0.2641616166,0.1047707945,-0.1369574219,0.3790746927,0.2572436631,0.6966174245,0.0503653102,0.1582950801,-0.0299590658,0.0022024186,0.2247810364,0.0723269656,-0.3129124343,0.1803696454,0.0302197263,-0.1768461466,0.2460018247,0.1162326187,0.2928936779,-0.0578404367,-0.3579824567,-0.2600935698,-0.3655110598,-0.1559731811,-0.0219270606,0.2030694187,0.1455339044,0.5627012253,-0.4348830283,-0.0822068229,0.3034934998,-0.1214675456,-0.026632091,0.0871555507,-0.2401306182,-0.3764371574,-0.1923998743,0.169432506,-0.1273618788,0.4058916271,0.0894154757,0.2031165957,0.1276355237,-0.1006715149,0.2976213396,-0.2545888424,0.1782376915,0.188301146,0.1651884317,-0.3848178983,-0.1694618165,-0.1441166401,-0.1351506412,0.2666935921,0.0691986829,-0.5115234256,-0.1317024082,0.1223711744,-0.0508822687,-0.2466030419,-0.1446665525,-0.2461933047,-0.2586558759,0.1532775164,0.0380989574,0.1500681639,0.3929439187,-0.0927803889,0.5734653473,-0.1120118648,-0.2354331166,0.2397040278,-0.0630112514,0.55297333,-0.019410111,0.0340885743,0.122265406,-0.1265092641,-0.2053155899,0.1226585954,0.1356473714,-0.3416363299,-0.1437759697,-0.3832884729,0.5200098157,0.2293066233,-0.2753112316,0.1476996839,-0.3316442668,-0.1977547258,-0.01063303,-0.0570235774,0.2516385615,-0.0470467024,0.117916353,-0.2057297826,0.53506428,0.2149329782,-0.1624639779,0.4008515775,0.1537148505,-0.4334540963,0.1491287351,0.3389035463,0.7175228,0.4781146049,-0.0385890417,0.4263069034,-0.2586626112,0.1159346029,0.0563809834,-0.004118057,-0.2077731639,-0.3220458925,0.1919017732,0.0116220526,0.105819352,0.0006326091,-0.0864815414,0.1789881289,-0.2033778429,0.386985451,0.1023421884,0.2675025761,0.0608242601,0.0337739214,-0.1049404293,0.0172547027,0.0907756463,-0.1177081317,0.0535145216,-0.3840714991,0.0313193277,-0.1755132377,-0.0016678156,0.052239202,-0.2561365664,0.0642953962,0.1166628972,-0.000884136,0.1194904894,0.0016806067,0.5374199748,-0.035936296,-0.1339445859,0.2569860518,-0.4386573136,0.0324886031,0.1740621477,0.1364796907,0.0406217873,-0.0596820042,-0.2285199016,0.3244857788,0.0586300716,-0.1731661111,-0.2100576758,-0.2583572268,-0.2615670562,-0.1521653831,-0.2196034491,-0.0531676374,-0.0192706417,-0.1628288925,0.0028900355,-0.1240886897,0.0244238637,0.4010928869,-0.1081084684,-0.1133654341,0.0903801695,0.3428358436,-0.2323462665,-0.2853524685,0.4727140963,-0.3125245571,-0.0826675519,-0.0505311377,0.174883455,-0.1998937875,-0.2814690769,0.2686016858,-0.0161281656,0.1039671749,0.0345945507,0.404282093,0.1158282235,-0.1011151522,-0.0572219491,-0.4350415468,0.1116835922,0.408542186,-0.2350719571,-0.0952422991,-0.4258363247,0.1973529905,0.077968359,0.1975916922,-0.2119119465,-0.2755135596,0.1248587742,-0.1465166807,0.267093569,0.0946127996,0.045810055,0.0242839176,-0.0631357431,0.0884652212,-0.0965996236,-0.0297743138,-0.2549293935,0.168671295,0.031317763,0.3304584324,0.087002039,-0.0451474674,-0.1597486585,-0.2132950574,0.0232998449,0.1865111142,-0.1165359169,0.5392251015,0.1041021645,0.8976370692,0.2134967893,-0.053431619,0.2053504586,0.2913578451,-0.0549782775,0.1057695076,0.313290447,0.136546433,-0.2882294655,-0.0891074315,-0.3128887713,0.0653656051,0.2927876711,-0.1010156274,0.0120883835,0.4501214027,0.0216898844,0.2443935424,-0.1739862859,0.0332906954,-0.3316778243,0.108799845,-0.1615259051,0.1517754495,0.1381608844,0.1890217066,0.0372141413,0.2193840146,0.3896560371,-0.1826103628,0.1538563818,0.0071755741,0.2480259389,-0.2777958214,0.3476074338,0.1518511772,0.2368544191,0.2364320755,0.278834939,0.0636553392,0.0487269908,0.4418359697,0.1506568491,0.2316164076,0.1368170083,-0.0283415467,-0.1424889117,-0.0037023097,-0.057564225,0.0473345183,-0.0968374386,0.1636189669,-0.3007003665,0.2142596692,-0.0032683706,-0.3228293955,-0.2563185096,-0.2440361679,-0.0155125642,-0.2392959148,-0.5574415922,-0.2417226434,-0.0506323054,-0.1181863248,0.1758486778,-0.0051929546,0.2497843951,-0.2485955507,-0.1441931427,0.2668414414,-0.26188007,0.3737464547,0.5232933164,-0.3483285308,0.0587759279,-0.1005220339,0.2709252238,0.21099931,0.3054180145,0.2390975207,0.3020229638,-0.3668431342,-0.2365858406,0.1456976682,-0.2033136785,0.114478454,0.3290289938,0.2814850509,-0.2071492225,0.202794373,0.0467125736,-0.1556740701,0.1896248013,0.2759167552,0.2000056803,-0.4089815021,0.3424044251,-0.0396012291,-0.0322977342,0.1221435815,-0.1428463608,-0.042736847,-0.2502233982,0.0933625996,0.2045282722,-0.0232299678,-0.5147492886,0.0396317132,0.0295872055,0.3810938299,0.0585155897,0.0123011433,-0.4914895594,-0.0101353619,-0.6506745219,-0.0930294916,0.2496894598,-0.1495498121,0.1456627101,0.1661141962,-0.0759351775,0.0700000003,0.0718906522,0.5094356537,0.1752088219,-0.0033009178,-0.4087581933,-0.1453213394,-0.2016129643,-0.2168287337,0.2321971655,-0.1759064496,-0.0097852089,-0.3974433839,-0.0150046283,0.0600623563,0.0033534041,0.1645422727,0.6511238813,0.1873810589,0.0206307843,0.7741081119,0.0136421323,0.0299884081,-0.4132049978,-0.1902291775,-0.2227948457,-0.0953332484,-0.1362854838,0.3383519053,-0.0151670501,-0.4483858347,-0.0393702649,0.211392343,0.3170768023,-0.4165564477,-0.0193999615,0.3332815766,-0.1906142682,-0.0921211764,0.0234425347,0.3206248879,-0.0995698422,0.2311321348,0.0546547771,-0.2851710618,0.2310818583,-0.3407650292,-0.3436341882,-0.2862753868,0.3049652874,-0.0992926285,-0.1729005277,-0.6666846275,-0.1416708827,0.2823508978,0.1122600064,-0.0339417011,0.2595966458,-0.0015445829,0.0506242812,-0.0288966801,0.0143096223,-0.1536698639,-0.1183005571,0.080745019,-0.1149214879]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2663","title":"[`to_json`] add multi-proc sharding support","comments":"Hi @stas00, \r\nI want to work on this issue and I was thinking why don't we use `imap` [in this loop](https:\/\/github.com\/huggingface\/datasets\/blob\/440b14d0dd428ae1b25881aa72ba7bbb8ad9ff84\/src\/datasets\/io\/json.py#L99)? This way, using offset (which is being used to slice the pyarrow table) we can convert pyarrow table to `json` using multiprocessing. I've a small code snippet for some clarity:\r\n```\r\nresult = list(\r\n pool.imap(self._apply_df, [(offset, batch_size) for offset in range(0, len(self.dataset), batch_size)])\r\n )\r\n```\r\n`_apply_df` is a function which will return `batch.to_pandas().to_json(path_or_buf=None, orient=\"records\", lines=True)` which is basically json version of the batched pyarrow table. Later on we can concatenate it to form json file? \r\n\r\nI think the only downside here is to write file from `imap` output (output would be a list and we'll need to iterate over it and write in a file) which might add a little overhead cost. What do you think about this?","body":"As discussed on slack it appears that `to_json` is quite slow on huge datasets like OSCAR.\r\n\r\nI implemented sharded saving, which is much much faster - but the tqdm bars all overwrite each other, so it's hard to make sense of the progress, so if possible ideally this multi-proc support could be implemented internally in `to_json` via `num_proc` argument. I guess `num_proc` will be the number of shards?\r\n\r\nI think the user will need to use this feature wisely, since too many processes writing to say normal style HD is likely to be slower than one process.\r\n\r\nI'm not sure whether the user should be responsible to concatenate the shards at the end or `datasets`, either way works for my needs.\r\n\r\nThe code I was using:\r\n\r\n```\r\nfrom multiprocessing import cpu_count, Process, Queue\r\n\r\n[...]\r\n\r\nfiltered_dataset = concat_dataset.map(filter_short_documents, batched=True, batch_size=256, num_proc=cpu_count())\r\n\r\nDATASET_NAME = \"oscar\"\r\nSHARDS = 10\r\ndef process_shard(idx):\r\n print(f\"Sharding {idx}\")\r\n ds_shard = filtered_dataset.shard(SHARDS, idx, contiguous=True)\r\n # ds_shard = ds_shard.shuffle() # remove contiguous=True above if shuffling\r\n print(f\"Saving {DATASET_NAME}-{idx}.jsonl\")\r\n ds_shard.to_json(f\"{DATASET_NAME}-{idx}.jsonl\", orient=\"records\", lines=True, force_ascii=False)\r\n\r\nqueue = Queue()\r\nprocesses = [Process(target=process_shard, args=(idx,)) for idx in range(SHARDS)]\r\nfor p in processes:\r\n p.start()\r\n\r\nfor p in processes:\r\n p.join()\r\n```\r\n\r\nThank you!\r\n\r\n@lhoestq ","comment_length":139,"text":"[`to_json`] add multi-proc sharding support\nAs discussed on slack it appears that `to_json` is quite slow on huge datasets like OSCAR.\r\n\r\nI implemented sharded saving, which is much much faster - but the tqdm bars all overwrite each other, so it's hard to make sense of the progress, so if possible ideally this multi-proc support could be implemented internally in `to_json` via `num_proc` argument. I guess `num_proc` will be the number of shards?\r\n\r\nI think the user will need to use this feature wisely, since too many processes writing to say normal style HD is likely to be slower than one process.\r\n\r\nI'm not sure whether the user should be responsible to concatenate the shards at the end or `datasets`, either way works for my needs.\r\n\r\nThe code I was using:\r\n\r\n```\r\nfrom multiprocessing import cpu_count, Process, Queue\r\n\r\n[...]\r\n\r\nfiltered_dataset = concat_dataset.map(filter_short_documents, batched=True, batch_size=256, num_proc=cpu_count())\r\n\r\nDATASET_NAME = \"oscar\"\r\nSHARDS = 10\r\ndef process_shard(idx):\r\n print(f\"Sharding {idx}\")\r\n ds_shard = filtered_dataset.shard(SHARDS, idx, contiguous=True)\r\n # ds_shard = ds_shard.shuffle() # remove contiguous=True above if shuffling\r\n print(f\"Saving {DATASET_NAME}-{idx}.jsonl\")\r\n ds_shard.to_json(f\"{DATASET_NAME}-{idx}.jsonl\", orient=\"records\", lines=True, force_ascii=False)\r\n\r\nqueue = Queue()\r\nprocesses = [Process(target=process_shard, args=(idx,)) for idx in range(SHARDS)]\r\nfor p in processes:\r\n p.start()\r\n\r\nfor p in processes:\r\n p.join()\r\n```\r\n\r\nThank you!\r\n\r\n@lhoestq \nHi @stas00, \r\nI want to work on this issue and I was thinking why don't we use `imap` [in this loop](https:\/\/github.com\/huggingface\/datasets\/blob\/440b14d0dd428ae1b25881aa72ba7bbb8ad9ff84\/src\/datasets\/io\/json.py#L99)? This way, using offset (which is being used to slice the pyarrow table) we can convert pyarrow table to `json` using multiprocessing. I've a small code snippet for some clarity:\r\n```\r\nresult = list(\r\n pool.imap(self._apply_df, [(offset, batch_size) for offset in range(0, len(self.dataset), batch_size)])\r\n )\r\n```\r\n`_apply_df` is a function which will return `batch.to_pandas().to_json(path_or_buf=None, orient=\"records\", lines=True)` which is basically json version of the batched pyarrow table. Later on we can concatenate it to form json file? \r\n\r\nI think the only downside here is to write file from `imap` output (output would be a list and we'll need to iterate over it and write in a file) which might add a little overhead cost. What do you think about this?","embeddings":[-0.242046684,-0.3245055377,-0.0334172063,-0.0440206081,-0.0977220312,-0.0860110968,0.4022955596,0.110401459,-0.0120916022,0.3323682249,-0.0399441943,0.2764456272,-0.1313705295,0.220484823,-0.2396433651,-0.0810559019,0.1407508403,-0.1243918613,0.4335382879,0.1816553771,-0.2526787221,0.3090566695,-0.1334240586,-0.2920532227,-0.1443659961,-0.1116204932,0.1702491939,0.1693739146,0.1813992709,-0.2050732076,-0.2966971099,0.1466355026,-0.1574244648,-0.0663380995,-0.0001212066,-0.0165756941,0.3360563517,-0.0251887646,0.1674968004,0.0298161469,-0.2610717416,-0.2657977343,-0.1118150875,-0.3679598272,0.1589178443,0.2573316395,0.079610385,-0.1090120748,0.2735993862,0.1824122667,0.0266379192,-0.0381312259,0.2731877565,0.1871727109,-0.1231689528,-0.0584274754,-0.134199053,-0.1373107284,0.1717839539,0.2883924544,0.1722884029,0.2656449974,-0.071409367,-0.0292127803,-0.0493074544,0.0790251568,0.0676737875,-0.2999537587,-0.0728117675,0.1627262384,-0.1009485945,-0.0941404179,-0.3561589122,-0.2951528728,-0.1113355607,-0.5862678289,-0.0910202265,-0.0569902956,-0.0046954835,0.0016200872,-0.2583362758,-0.1352896541,-0.1115969121,-0.0975015834,0.055972673,-0.0492389314,0.1559401751,0.2313601971,0.4239072502,-0.116016157,-0.2813234031,-0.036755722,0.1901650429,-0.0995850414,-0.318210125,-0.2331991494,0.067072466,-0.2741035819,-0.1916322261,0.2405323833,-0.0652907044,0.3501672149,0.1451126337,0.2800191641,-0.0082594501,-0.3676018119,0.1055693775,-0.2028030753,0.0866219476,0.0854692459,-0.0438373163,0.0142234955,0.1041972712,-0.1864198446,0.2379478663,0.0610435084,-0.224745959,0.0134621486,-0.1326915771,0.1245782599,0.0409150533,-0.1848429441,-0.1038549617,0.1181193963,0.1211151928,0.45546031,-0.2897269428,0.1267385185,-0.1991168708,0.0845798627,-0.0060030092,0.1752078086,-0.1212147102,0.3038285971,-0.0945648476,0.3363530636,-0.1660330445,-0.1813056916,0.4267489612,0.0143280402,0.0071663768,-0.16150783,-0.0277643669,0.3133506477,-0.052582223,-0.0263768546,-0.0027026199,-0.6090400219,-0.5215034485,0.0872401595,-0.424731642,-0.5617137551,0.0674493909,-0.0105560878,0.0413319431,0.053050071,-0.7100686431,0.1912888587,0.1468509734,0.0172505639,-0.2996538877,-0.0313656293,-0.3974209726,-0.3379457891,0.0040908754,0.1846761703,-0.0345499441,0.1666767001,-0.1057315916,0.1573319584,0.4911988974,0.3862863481,0.0417262204,0.1512308419,0.1384150386,0.4020533562,0.1216073856,-0.2359776348,-0.3532421887,0.343623966,-0.1950227618,0.1283876598,0.58977139,0.1619928926,0.1941811889,-0.1963891238,0.5543405414,0.084137179,-0.0459819511,0.2191111892,-0.288382709,-0.0190754812,-0.6522377133,0.2896907926,0.0680862591,0.0845365077,0.1911191493,-0.2741482556,0.2790185213,-0.1253771037,0.0269500688,0.1180229038,0.2163124979,-0.1152399108,-0.263433367,0.1062837616,-0.0953335464,0.3388625681,0.2482239306,0.1104214638,-0.0917066708,-0.1606258154,0.2462689579,0.1796581149,0.0966826454,0.1453395188,-0.0540784188,-0.2201896161,-0.148558557,-0.0962636843,-0.0684250891,0.5936162472,0.1674351841,-0.0775221288,-0.4122172594,0.1772023886,0.0901107192,-0.1776477545,0.0167182907,0.1800415516,0.0830760226,-0.1811238974,0.2029942572,0.0858528763,0.2311604619,0.1759352535,-0.020178318,0.3995901942,0.058495447,0.1984225512,0.0361033008,-0.2241056561,0.0027489795,-0.2517606318,-0.1687451601,0.7329337001,0.2339997739,0.5969922543,0.1784469783,-0.1098150834,0.2974082232,0.0058557852,0.0032608882,-0.0948940367,0.1027392,0.433375895,-0.4786273241,-0.2609365582,-0.3565048575,0.4948657155,0.325558573,0.161405161,-0.2580355406,0.29643327,0.3624889851,0.0195348449,0.1627064049,0.3779317737,0.6566362381,0.2248315811,-0.1344006956,-0.0622735098,0.0452874824,-0.3232685626,0.050376039,0.09435381,0.369332552,0.0394429304,0.207812205,0.015269028,0.0442837626,-0.5829306841,-0.0670747459,0.0321427658,-0.147529617,0.1656359881,-0.1907572448,0.5931070447,0.0694528222,0.0002796644,0.1994153708,-0.4266071022,-0.018969696,0.5152558684,-0.15105775,-0.1607409716,0.1470748484,0.0544022396,-0.3848678172,-0.3304622769,-0.2615265548,-0.0200198349,0.1954409033,-0.0881892592,0.3558399677,-0.0667843595,0.2389365733,-0.0046784831,0.0458213799,0.0259417854,-0.189389497,-0.0488842279,-0.1514607072,0.1716158539,-0.0206062961,-0.1644494385,-0.0581386164,-0.2877838314,0.2799780369,0.1137813032,-0.0842128918,0.2283849418,-0.2359640598,-0.1618583351,0.1569346488,0.0827050433,-0.2531999052,-0.2911528945,0.5311349034,0.1381748021,0.3013705313,-0.4351008534,0.1738480628,-0.1655333787,0.0203211103,0.0540141091,0.0509408452,-0.4110952616,0.3229976892,-0.0502166562,-0.0940238982,-0.1275964826,-0.0198118761,0.1817742884,0.1251268536,-0.2338487804,-0.386282891,-0.2070031315,0.1083227545,-0.0194043163,-0.2067757547,0.4428941905,0.1868667454,-0.1519838125,0.1370428652,-0.1170813367,0.1773805022,-0.1496586055,-0.1970183253,-0.0474350415,0.5124171376,0.1980293244,0.6017048359,0.256328851,-0.0557649992,0.2804168165,0.1276686937,-0.1844818592,-0.1950068325,-0.1390612125,-0.0172410626,-0.1265537441,0.0335102081,-0.0515082888,0.027430065,-0.4436281025,-0.0378677733,0.1597196609,-0.3875334859,-0.3500620425,0.2630936503,0.203405112,-0.1624292284,-0.1210547611,-0.2143114805,-0.2355105728,0.0491949208,-0.2517802119,0.4220284522,0.1440107226,-0.3891451657,-0.3714787662,-0.4610387683,-0.6336141825,0.4533956945,0.1017774194,0.1259099692,0.0799116194,-0.1299972683,0.1769379228,-0.4740127623,0.4230646491,-0.2788916826,-0.3419285715,0.441165328,-0.2057162076,-0.1236956567,-0.0165876187,0.0198714342,0.2513070107,0.4258913696,0.7600981593,-0.3608616292,-0.1523099542,-0.0209073517,-0.2820302546,-0.1159221008,-0.3202486336,-0.1753597409,-0.3228887916,-0.4666146338,0.1101015136,0.1017075926,0.4080095291,-0.0057399729,-0.0836375579,-0.0733939856,-0.2006114125,0.3292187154,-0.1313379556,-0.1254904121,0.0221712757,0.2492970973,-0.4836689234,0.0671683922,0.2951049805,0.2029377222,-0.0138543118,-0.0998578072,-0.1386744529,0.1534721404,0.3938668668,0.2586364448,0.0421480052,0.149000138,-0.2551950812,0.6260348558,-0.0486512743,0.0141149834,-0.0253271721,0.152886048,-0.218819499,0.0956200436,0.3878549635,0.1129258573,-0.0957068428,0.0178133938,0.1687095016,-0.0675470755,0.1491490453,0.225141257,0.9898107052,-0.3000160456,0.1346515566,-0.1427120268,0.0296357404,-0.0609623007,-0.5919108391,0.1342042983,-0.0659704059,-0.1331831217,0.0868166536,-0.0425486378,0.2160753161,0.3780532181,-0.0626917332,0.3177190125,0.4126956761,-0.4591330886,-0.0893737823,0.3414219618,-0.2101225555,-0.5266346931,-0.06003014,-0.0211201925,0.2636204362,-0.4082714021,0.0089062965,-0.1239412799,-0.1102142185,0.0826623812,-0.303912878,-0.0807647482,0.143129006,0.1746755242,0.117700927,-0.1286086291,0.00480593,0.1438336223,0.3051338494,-0.0194032863,-0.0641005039,-0.150350973,-0.0903271362,-0.0649671406,0.1151676029,-0.4983928204,0.3140555918,-0.2099169493,-0.1760046929,-0.1197826266,0.1505512595,-0.5108479261,-0.1043047383,0.2440752238,0.2966704369,0.2887217402,-0.2108632326,0.3275487721,-0.4023462236,-0.2975877821,-0.0075662476,0.27863428,-0.1586078852,0.5502755046,-0.5499783158,-0.1529865712,0.0382565781,0.1312533766,0.1465375125,-0.0624503717,0.0038701266,-0.09766303,-0.2169619948,-0.0971847922,0.1312489361,-0.0407382213,0.055336915,0.1016310304,-0.1436048597,-0.181569472,0.0639711991,-0.1180490032,-0.0877136886,-0.0725387856,-0.0936049595,-0.3099119365,-0.2076792121,0.1351599693,-0.1702670306,-0.0460894294,-0.1275122613,0.174594149,-0.2650547624,0.2505453229,-0.1761873811,0.0518844426,-0.2666443586,0.0531452671,0.2940184474,-0.1134411842,0.1016727313,0.1896566302,0.0879807845,0.2687368095,-0.2080687881,-0.1312021166,0.1138236895,0.1918836236,-0.0635316074,0.0977950171,0.1916431934,0.2487067878,-0.3594392538,-0.1701643765,0.0609475262,0.0068934439,-0.2717464268,-0.142234832,0.7559414506,0.067043893,0.0869190469,0.1275442243,0.0370017588,0.6623572111,-0.2667343915,0.2846893966,-0.001667654,-0.171342954,0.0210855026,0.1079128832,0.2110730261,0.1510643512,0.4081349373,-0.0054548155,-0.2449522465,-0.187713176,0.2510827482,0.0112663759,-0.0196200125,0.0264090914,0.1363794804,0.0955168605,0.0654900372,-0.3602882624,0.2202573568,-0.0914757401,0.1921834648,0.3553713858,0.0795507357,0.1364493072,-0.1006603464,0.0064262394,0.1384610832,-0.2223738283,-0.4793302715,-0.443996191,0.162861824,0.1356537342,0.3500679433,-0.0445967093,0.2138268203,0.139121145,-0.0987010673,0.9213707447,0.2334339172,0.2134949714,-0.0153667312,-0.1005678326,0.0820200145,0.1242319047,0.0911824629,0.3882047236,0.1295160055,0.2498442233,0.0221476667,0.0599887408,-0.2447606474,0.4115552902,0.1417625993,-0.482548207,0.1349291354,0.0269402228,-0.1517360806,0.0484983549,-0.043924503,0.1780866683,0.4796350598,0.1804549992,0.4221502841,0.060802523,-0.1545421481,-0.1378033906,0.4936316311,-0.2833922505,0.0602761395,0.2513494194,-0.2618722022,-0.0639844313,-0.0025814138,0.3616571724,0.0012852149,0.0663288683,0.2202922255,-0.1101968437,0.0451882184,-0.0529714338,0.3104578555,0.0357825942,0.2876131535,0.1312081367,0.0291609187,-0.1718657017,0.0844169483,-0.048168093,0.0933858156,-0.04881192,0.4057820141,0.0755932853,-0.5193164945,0.1798733622,0.2910909057,-0.4833246768,-0.2664075494,0.1483105719,-0.642766118,0.0150821013,0.0711624697,0.0286361035,-0.071116589,0.3917168379,-0.4299818873,0.4232682586,-0.3951330185,-0.0731770694,-0.1655282676,-0.0722793713,-0.1821724325,0.1409935802,0.0246726573,0.4999508858,0.2592582703,0.4662968814,-0.0587026477,0.0591153763,-0.3290489316,0.5914949179,-0.3510633409,0.4751754999,0.3905496597,-0.2426531017,0.0615285821,-0.2855625451,0.4005372822,0.093343094,0.0339241959,-0.1799997985,0.2867391109,0.0044836863,0.210711956,0.1944090426,0.1933003664,0.0686054826,-0.3118375242,-0.1851777434,-0.0143256374,0.3048584163,-0.1415357441,-0.0948515907,-0.1159189194,0.513730824,-0.2455666661,0.3382499814,0.059224125,-0.1040126607,-0.1777090132,-0.4909538925,-0.1758880615,0.0819056407,0.0400156081,0.3124935329,-0.1231914088,0.4903734028,0.2959265709,0.0327536799,-0.1817988157,-0.132129997,0.2517085373,-0.1657280177,-0.1034460962,-0.0664739907,0.2598737478,-0.4155711234,0.5195026994,-0.249162823,-0.1954654306,0.2228465825,0.0593374856,-0.3831810653,0.5472032428,0.0982533172,-0.6427063942,-0.0774774477,-0.1920465678,-0.0915123001,-0.2088404745,-0.2902126014,-0.2356254309]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2663","title":"[`to_json`] add multi-proc sharding support","comments":"Followed up in https:\/\/github.com\/huggingface\/datasets\/pull\/2747","body":"As discussed on slack it appears that `to_json` is quite slow on huge datasets like OSCAR.\r\n\r\nI implemented sharded saving, which is much much faster - but the tqdm bars all overwrite each other, so it's hard to make sense of the progress, so if possible ideally this multi-proc support could be implemented internally in `to_json` via `num_proc` argument. I guess `num_proc` will be the number of shards?\r\n\r\nI think the user will need to use this feature wisely, since too many processes writing to say normal style HD is likely to be slower than one process.\r\n\r\nI'm not sure whether the user should be responsible to concatenate the shards at the end or `datasets`, either way works for my needs.\r\n\r\nThe code I was using:\r\n\r\n```\r\nfrom multiprocessing import cpu_count, Process, Queue\r\n\r\n[...]\r\n\r\nfiltered_dataset = concat_dataset.map(filter_short_documents, batched=True, batch_size=256, num_proc=cpu_count())\r\n\r\nDATASET_NAME = \"oscar\"\r\nSHARDS = 10\r\ndef process_shard(idx):\r\n print(f\"Sharding {idx}\")\r\n ds_shard = filtered_dataset.shard(SHARDS, idx, contiguous=True)\r\n # ds_shard = ds_shard.shuffle() # remove contiguous=True above if shuffling\r\n print(f\"Saving {DATASET_NAME}-{idx}.jsonl\")\r\n ds_shard.to_json(f\"{DATASET_NAME}-{idx}.jsonl\", orient=\"records\", lines=True, force_ascii=False)\r\n\r\nqueue = Queue()\r\nprocesses = [Process(target=process_shard, args=(idx,)) for idx in range(SHARDS)]\r\nfor p in processes:\r\n p.start()\r\n\r\nfor p in processes:\r\n p.join()\r\n```\r\n\r\nThank you!\r\n\r\n@lhoestq ","comment_length":4,"text":"[`to_json`] add multi-proc sharding support\nAs discussed on slack it appears that `to_json` is quite slow on huge datasets like OSCAR.\r\n\r\nI implemented sharded saving, which is much much faster - but the tqdm bars all overwrite each other, so it's hard to make sense of the progress, so if possible ideally this multi-proc support could be implemented internally in `to_json` via `num_proc` argument. I guess `num_proc` will be the number of shards?\r\n\r\nI think the user will need to use this feature wisely, since too many processes writing to say normal style HD is likely to be slower than one process.\r\n\r\nI'm not sure whether the user should be responsible to concatenate the shards at the end or `datasets`, either way works for my needs.\r\n\r\nThe code I was using:\r\n\r\n```\r\nfrom multiprocessing import cpu_count, Process, Queue\r\n\r\n[...]\r\n\r\nfiltered_dataset = concat_dataset.map(filter_short_documents, batched=True, batch_size=256, num_proc=cpu_count())\r\n\r\nDATASET_NAME = \"oscar\"\r\nSHARDS = 10\r\ndef process_shard(idx):\r\n print(f\"Sharding {idx}\")\r\n ds_shard = filtered_dataset.shard(SHARDS, idx, contiguous=True)\r\n # ds_shard = ds_shard.shuffle() # remove contiguous=True above if shuffling\r\n print(f\"Saving {DATASET_NAME}-{idx}.jsonl\")\r\n ds_shard.to_json(f\"{DATASET_NAME}-{idx}.jsonl\", orient=\"records\", lines=True, force_ascii=False)\r\n\r\nqueue = Queue()\r\nprocesses = [Process(target=process_shard, args=(idx,)) for idx in range(SHARDS)]\r\nfor p in processes:\r\n p.start()\r\n\r\nfor p in processes:\r\n p.join()\r\n```\r\n\r\nThank you!\r\n\r\n@lhoestq \nFollowed up in https:\/\/github.com\/huggingface\/datasets\/pull\/2747","embeddings":[-0.2589680254,-0.0918775871,-0.0576498061,-0.0704856366,-0.0612732619,-0.1120798886,0.1844341904,0.2372895777,-0.1721450388,0.2867182195,-0.0419638082,0.2308939695,-0.1627990752,0.3421465456,-0.1984911114,-0.1427096128,0.0648157671,-0.0671752021,0.3733898401,0.1362454444,-0.1863879263,0.3017354012,-0.0441765226,-0.3480331898,-0.2585121989,-0.1721113324,0.160213232,0.2564293146,0.0634392798,-0.2405580729,-0.2433749735,0.3193820715,-0.1165914014,0.0491569079,-0.0001134464,-0.1463132799,0.1883964241,-0.0643665791,0.1147759184,0.0870793536,-0.2839747071,-0.3492469788,-0.0697029606,-0.3090184927,0.036645934,0.2163705677,-0.0231226757,0.0443896428,0.2931517363,0.121189028,0.1135433465,0.0337900929,0.2684995234,0.1284158677,0.0301001705,0.047791563,-0.2120207399,-0.2040484697,0.1391547024,0.1760188937,0.0613458753,0.3060928285,0.0417647995,0.0619020127,-0.0620846637,0.0019832454,-0.0662469044,-0.2511831522,0.0135586923,0.2696016133,-0.1173676401,-0.1742272079,-0.3483434319,-0.2858772874,-0.1442148238,-0.6216781139,-0.0130536826,-0.0098345829,-0.0494724847,-0.1249613985,-0.2632095218,-0.1180953607,-0.0632162541,-0.1591652483,0.0824892521,-0.1103856117,0.0552253835,0.0782165676,0.4915221035,-0.2110628784,-0.3899771571,-0.0411269218,0.0796566308,-0.0946010351,-0.4593486488,-0.304228723,0.0852270052,-0.1985126436,-0.0818688497,0.2214163691,-0.0954067856,0.3724784851,0.1753745079,0.2817041874,-0.0435911529,-0.335526675,0.0629678443,-0.2974533141,0.1588410437,0.0986794606,-0.1107533723,0.0916972682,0.2281201482,-0.3668832481,0.1427673846,-0.0608793646,0.0057003847,-0.0442992263,-0.1416345388,0.1909582168,0.1729401201,-0.1656423658,-0.0985864922,0.2449750006,0.1509555876,0.3667424619,-0.2883856595,0.0022954706,-0.2577789426,0.0565467477,-0.073653996,0.1234186068,-0.0789595768,0.3007636666,0.0117143262,0.2879104614,-0.1176278889,-0.0087823775,0.410782218,0.0427646562,0.0713542849,-0.1149643958,-0.0356356874,0.2949905097,0.0197471343,0.0048119319,-0.0060105543,-0.5990056396,-0.4445469081,-0.0252399836,-0.2807484269,-0.4899550974,-0.096292831,0.1003282964,-0.0485349223,0.0810248479,-0.6613173485,0.1621778607,0.2364887893,0.0902986005,-0.2587816417,-0.1060409993,-0.3017772138,-0.3300236762,0.14380081,0.2550338805,-0.0137519194,0.2111866176,-0.0422421731,0.0739897192,0.5260491371,0.3500675857,0.0388346016,-0.0060482156,0.1456352025,0.265100956,0.2407912612,-0.0609561652,-0.3280103207,0.3941291869,-0.2001901418,0.176992774,0.435189575,0.0772903115,0.3361689746,-0.1255822629,0.3731564581,0.114913933,-0.0762932226,0.2278157026,-0.3913858831,0.0391861908,-0.5394315124,0.3452333808,0.0435075164,0.0058671711,0.1637889594,-0.1936094314,0.386806488,-0.1124520749,0.1601219773,0.1893366575,0.3028890491,-0.0440622903,-0.2237871736,-0.0190573242,-0.0899518505,0.3437098265,0.1706651002,-0.0078226123,-0.0392822661,-0.2486129403,0.1950482428,0.1261149794,-0.0771168545,0.0366546772,0.0522872843,-0.0196360108,-0.0568970107,-0.0804030746,-0.0480247922,0.3600597978,0.0624312796,-0.1266635805,-0.3918425143,0.1445315033,0.0346137621,-0.2377437055,0.0963305905,0.2672162652,0.1199900955,-0.1684671044,0.0765360147,0.1502855271,0.2660882771,0.2157917619,0.12018805,0.3985857666,0.0166479237,0.2176698446,0.011158417,-0.0748690069,0.0437266827,-0.1662618518,-0.2861901224,0.7992159128,0.0389492214,0.4255234897,0.1560093164,-0.0184106939,0.4775709212,0.0324083418,0.0810338333,-0.0587894805,0.2202866822,0.257635355,-0.4507189095,-0.115712367,-0.457080245,0.5517137647,0.3584005237,0.1330431551,-0.2613343596,0.2976892292,0.3658581078,-0.0608048066,0.1840056628,0.5441958308,0.5139400959,0.2723759711,-0.1242992282,-0.1757538915,-0.0000516868,-0.3410226107,0.0358858742,-0.0825001374,0.3095395863,0.1206847727,0.25117594,-0.0716604888,-0.0488715097,-0.4492748976,0.0135223549,0.0970034674,-0.1357730478,0.067643635,-0.2015535533,0.5673757195,-0.0400692038,0.0453298241,0.1000094861,-0.4185921848,0.0607100539,0.5298936963,-0.2244472802,-0.0805164948,0.1709019542,0.0041539585,-0.2994757593,-0.352668196,-0.1902913749,-0.0507390536,0.1903806627,-0.0146614825,0.2862617075,0.0409984998,0.3056821525,-0.0013201875,0.041760955,-0.2032198906,-0.2093817592,-0.0910733864,-0.2177374959,0.1406840533,0.0251620114,-0.0886580646,-0.0177857019,-0.3122074902,0.348960489,0.1336126626,-0.0720665529,0.2245164216,-0.1749510169,-0.2223935276,0.1208523512,0.0281568896,-0.2387424707,-0.3585073948,0.5198842287,0.0947335884,0.3278203607,-0.504627347,0.185286656,-0.1710387021,0.0062035676,0.0299506355,-0.0074030384,-0.439620018,0.3440851271,-0.2162673473,-0.2472946793,0.0050764387,0.0275440626,0.0426644906,0.0977484882,-0.1366821826,-0.3434458077,-0.3470418453,0.2054468542,0.0024989238,-0.2999646366,0.4730596542,0.1646277457,-0.1680541188,0.0729414448,-0.2257523388,0.1702606827,-0.1517709345,-0.0062310561,-0.1488639116,0.5266501904,0.1976227015,0.8174567223,0.2063963562,-0.0895163193,0.2693181932,0.0992315933,-0.1557885408,-0.0001181565,-0.1991666704,0.1828428954,-0.1033343226,0.0511966497,0.0288370159,0.0008598319,-0.307137996,0.0801747292,0.1692079455,-0.3501419723,-0.3318415582,0.240520224,0.099088192,-0.0563399866,-0.0980466679,-0.1484862864,-0.1104917675,0.0366098098,-0.2518438399,0.4723474681,0.1173130199,-0.3247686327,-0.3189881742,-0.2893016338,-0.6385824084,0.4125442505,0.0620582923,0.0785127878,0.0543797053,-0.1721871793,0.1240257621,-0.487157166,0.5785815716,-0.2515177727,-0.129605785,0.3922459483,-0.3490431309,-0.1943314523,-0.1097179875,0.0015705039,0.2381868213,0.3365282416,0.5641977787,-0.3093546033,-0.2194643617,0.0180108752,-0.2969848514,-0.1985451132,-0.1894515306,-0.1720685363,-0.4607029259,-0.4661668837,0.1895992905,-0.0083334781,0.4546500444,-0.1556004733,-0.1550741494,-0.1105300635,-0.0970676467,0.3385181725,0.0278036147,0.0083560273,0.0546867959,0.3027765751,-0.5568296909,0.1546294093,0.3656754792,0.0650907159,-0.0925740898,-0.0643983483,-0.0771190301,0.11476437,0.3390805125,0.2223137766,0.0042771841,0.2374492288,-0.3448750079,0.6518741846,-0.1514120251,0.1159031838,0.0627114549,0.2201201171,-0.2186619192,0.0784870088,0.4713288546,0.038898278,-0.0745324865,-0.0823944137,0.2758824527,-0.0258192066,0.2147291452,0.2543294728,0.90400213,-0.3341965675,0.0966537297,-0.1367000639,-0.1883847266,-0.021849867,-0.5390428305,0.147691533,-0.0866924897,-0.1312883943,0.103168346,0.0363894887,0.1505188644,0.2571533918,-0.1399183869,0.2715266049,0.2507302463,-0.3560775518,-0.1192840263,0.3339233696,-0.1422005296,-0.5442208052,-0.1964393556,0.0923828855,0.146931082,-0.3542996049,-0.0082650846,-0.2251112908,0.018463945,0.1209683493,-0.3432137072,-0.0250184927,0.1527830362,0.0871875137,0.0166246593,-0.2674011588,0.0305303708,0.2301599681,0.2422534078,0.0989266485,-0.1039030924,-0.0019325573,-0.1371763051,-0.019805517,0.0088288337,-0.4079633951,0.2255508602,-0.1861819029,-0.1813757271,-0.1066384092,0.1574758738,-0.472532481,-0.0860512108,0.2205504328,0.0314872488,0.2007463723,-0.2378676087,0.1620062292,-0.4257148206,-0.3968725204,0.0981983542,0.2481881976,-0.2990430295,0.5140537024,-0.4327147901,-0.1405060291,0.0630534589,0.1855185032,0.0745097697,0.1287233084,0.0382935554,0.0663165972,-0.1808655411,-0.2193683684,0.0674226657,-0.0805138722,0.0529032648,0.2355923653,-0.1621084809,-0.1756132841,0.1157682389,0.0039531016,-0.0621184222,-0.0920534357,-0.0410789102,-0.3032677472,-0.1788046211,0.0014725219,-0.1898106784,0.0482676327,-0.1112307608,0.2778709829,-0.285746038,0.4806691408,-0.3021123707,0.0671323538,-0.3638429046,0.0995892435,0.3331357837,-0.0893965289,0.088524431,0.1734143943,0.1287761182,0.2030022442,-0.2548954487,-0.2278966755,0.0959419757,0.133914873,-0.0759692267,0.0851098895,0.1287074983,0.3423366547,-0.4360066056,-0.2556177378,0.1508336514,-0.0717233866,-0.2614281476,-0.0541964918,0.5440743566,-0.0900605544,0.1332993507,-0.0115436809,0.0604784228,0.5467866063,-0.2553377151,0.2194972038,0.0156388544,-0.2189732641,0.0611132793,0.0772077218,0.193208158,0.0229556933,0.3467494547,-0.0593452603,-0.3773351312,-0.1065883785,0.4373129606,0.0407470576,0.0138168726,0.064128764,0.0743581951,0.1769029498,-0.0028305359,-0.2559142709,0.2584605515,-0.0707444251,0.2269387245,0.2541235089,0.0339746922,0.0889403895,-0.1701530814,0.1415061057,0.0453248098,-0.3516990542,-0.3000396788,-0.323405385,0.1455805898,0.0664978623,0.3822084367,-0.0270418748,0.2925457656,0.207997337,-0.0498952456,0.9926770926,0.2040609717,0.2599684894,0.0121290935,-0.1198292375,0.0648699924,0.1807128787,0.0718857497,0.3505195379,0.0495096967,0.3022670448,0.0501169227,-0.1044208407,-0.3412741125,0.4734872878,0.0961911678,-0.4990324676,0.1238430589,-0.142181918,-0.1253979951,0.090423651,0.0067218943,0.1236199215,0.5827206373,0.0807865784,0.3513689041,-0.1734034568,0.0706836432,-0.1013105065,0.3962741196,-0.228464365,0.1180418357,0.312005192,-0.3629868329,-0.0978060737,-0.0150598744,0.3627075851,0.0798354074,-0.0547518693,0.270739764,-0.17515634,0.0706452802,0.0104782386,0.3343847692,-0.0500765443,0.1918580383,0.2520774603,0.11894425,-0.2285446823,-0.0361148976,-0.0317626409,0.0966361687,-0.0385205969,0.4121084809,0.1226985455,-0.345122695,0.085933052,0.2492906302,-0.474761039,-0.4300922155,0.2122519463,-0.4477131367,0.0929128975,0.0080875224,0.0892013758,-0.1303339601,0.4167878032,-0.389747858,0.3080649376,-0.3478427529,-0.1400349289,-0.3211285174,-0.0594255775,-0.1840226501,0.0174644962,0.008437695,0.6154229045,0.1432597786,0.2925439477,-0.0710944533,0.0558931567,-0.0896000341,0.5094795227,-0.3882248998,0.4949989021,0.1684991717,-0.2269578129,0.0897995681,-0.3326002359,0.4390733242,0.096319519,0.0859970972,-0.1896385103,0.3107280731,0.0721153617,-0.0355978943,0.1458602697,0.2387641668,0.0365365334,-0.365978539,-0.1196603626,-0.0404097214,0.3039698303,-0.2112697959,-0.217676878,-0.066395469,0.5828909278,-0.1233704761,0.4060660005,-0.0406735614,0.114686057,-0.1667162031,-0.4330833852,-0.0492257774,0.0910193846,0.161679104,0.3147015572,-0.0366392247,0.438061595,0.1792207807,0.1771707833,-0.2501685917,-0.20222269,0.347094208,-0.1991768926,-0.1193009913,-0.0769897327,0.3022851646,-0.3155647814,0.4542097151,-0.2544998825,-0.1181887835,0.1468921751,0.1384284198,-0.3123249412,0.5528783202,0.0068337419,-0.5275618434,-0.089308098,-0.1818077117,-0.0406935774,-0.1866644323,-0.2474841774,-0.2761361599]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2655","title":"Allow the selection of multiple columns at once","comments":"Hi! I was looking into this and hope you can clarify a point. Your my_dataset variable would be of type DatasetDict which means the alternative you've described (dict comprehension) is what makes sense. \r\nIs there a reason why you wouldn't want to convert my_dataset to a pandas df if you'd like to use it like one? Please let me know if I'm missing something.","body":"**Is your feature request related to a problem? Please describe.**\r\n\r\nSimilar to pandas, it would be great if we could select multiple columns at once.\r\n\r\n\r\n**Describe the solution you'd like**\r\n```python\r\nmy_dataset = ... # Has columns ['idx', 'sentence', 'label']\r\nidx, label = my_dataset[['idx', 'label']]\r\n```\r\n\r\n**Describe alternatives you've considered**\r\nwe can do `[dataset[col] for col in ('idx', 'label')]`\r\n\r\n**Additional context**\r\nThis is of course very minor.\r\n","comment_length":64,"text":"Allow the selection of multiple columns at once\n**Is your feature request related to a problem? Please describe.**\r\n\r\nSimilar to pandas, it would be great if we could select multiple columns at once.\r\n\r\n\r\n**Describe the solution you'd like**\r\n```python\r\nmy_dataset = ... # Has columns ['idx', 'sentence', 'label']\r\nidx, label = my_dataset[['idx', 'label']]\r\n```\r\n\r\n**Describe alternatives you've considered**\r\nwe can do `[dataset[col] for col in ('idx', 'label')]`\r\n\r\n**Additional context**\r\nThis is of course very minor.\r\n\nHi! I was looking into this and hope you can clarify a point. Your my_dataset variable would be of type DatasetDict which means the alternative you've described (dict comprehension) is what makes sense. \r\nIs there a reason why you wouldn't want to convert my_dataset to a pandas df if you'd like to use it like one? Please let me know if I'm missing something.","embeddings":[-0.0573705472,-0.2296809554,-0.196985513,0.0512581915,0.2014224976,0.2186503857,0.5181803107,0.1173711866,0.3343811333,0.4082842171,-0.1473215073,0.3830779195,0.0007235444,0.2195994854,-0.2610401511,-0.3490301371,-0.1463272572,0.1411355287,0.0857272297,0.0135873593,-0.1727700979,-0.3008389771,-0.34792009,0.0155922901,-0.0639111847,-0.2046385854,-0.2487484068,-0.2731154859,-0.374528259,-0.2284196466,-0.0509770252,0.4818705618,-0.0998930559,0.0420988575,-0.0001022023,-0.2425405979,0.0494161844,0.0450042561,-0.372803688,-0.3197800517,-0.1985395104,-0.2248339653,0.3259703517,-0.3805671632,-0.1459850371,-0.176906988,-0.2034198344,-0.2526134551,-0.0332642533,0.087867476,0.3098186553,-0.014179701,-0.212574631,0.0007960759,-0.0395887345,0.1632219702,-0.1948770136,-0.0862278566,0.4251423776,-0.022268353,0.2404662669,0.1023098603,-0.2389610261,0.0799369439,0.0789678991,0.1655120254,0.2781155109,-0.2613496184,0.0809774324,0.2927647829,0.7717722058,-0.1894667894,-0.4718903303,-0.2102407068,0.4684536457,-0.1977852434,-0.3191841245,0.2664964199,-0.2287949473,0.2275188863,0.1129403636,-0.2688922286,-0.1090144664,0.1486848146,-0.155721277,0.3326571584,-0.0430136062,0.0551568158,0.131431073,-0.0222349931,0.1233634427,-0.1049084663,0.1292737126,0.1556303203,-0.6503581405,-0.2172845006,-0.2239073962,-0.2029287517,0.1681587398,0.2159994692,-0.0236092377,-0.0231380388,-0.1188300475,0.316463232,0.4838551879,-0.0458058529,-0.0439748168,0.389067024,-0.0343437567,-0.1517443806,0.0094906045,-0.0180744138,-0.091057241,0.0409744494,0.0095805982,0.2227684557,0.0332288779,0.0254826564,-0.2143536806,-0.2053003162,-0.1048714966,-0.2566322088,-0.06128674,0.4625502229,0.1120596603,0.3907713294,-0.1936305612,0.2314957231,0.2148041874,-0.2290373892,-0.2068743706,0.2864536047,0.1582296789,-0.1184982359,0.1220543459,0.0670838952,-0.1788309067,0.3173830509,0.0300921556,-0.3956840336,-0.1217478663,-0.3682628572,0.0412878841,0.1645923257,-0.0171067212,0.1042750329,0.2690171897,-0.2168727666,-0.3145562708,0.3942932487,0.0467404574,-0.1573522538,-0.2171412557,0.1881261617,0.1842950583,-0.3066887558,-0.0916226506,0.5061647296,-0.0624875426,0.3623635471,0.1700692475,0.0290999245,0.0495873243,-0.3428075016,-0.0607744679,0.3431543112,-0.5089313388,0.0842315033,0.1649499536,-0.0967291221,-0.0080227749,0.1633627415,-0.2780387402,0.086305052,0.2251906097,0.0358405933,0.3047424257,-0.3884338439,-0.1144216284,0.1912813038,0.0785425976,0.1773522496,0.458563298,0.1077365652,0.1881625503,0.2333642244,0.057728298,0.1933408678,0.0248733461,0.0773601532,0.2218871117,-0.4066192508,0.2764487863,0.1477501839,-0.1539059281,0.1683893204,-0.1599247307,-0.1816745698,0.3445219994,-0.2589389384,-0.1385620385,0.1810883731,0.3131214976,0.1123533547,-0.0972969234,-0.4043030441,-0.2775419056,0.1191289872,0.4883978069,0.1380021274,0.0122433724,-0.4381889105,-0.1534580588,0.0582767762,-0.1595164239,0.1654387265,0.2506510615,-0.2094160467,-0.1075731367,-0.1019803286,-0.5070766211,0.3666983247,-0.1956880093,0.1371779293,0.1988463253,0.1711434573,0.0042811586,0.1743839979,0.0299108475,0.0195631087,0.200883016,0.1793606728,0.1491049379,0.0841175914,0.1976446509,-0.0701240003,-0.3016873598,-0.0303100739,0.074177511,0.1923077404,-0.0160453785,-0.2838518322,0.1254215091,0.0264123604,-0.1852659434,0.0936241671,-0.1243551448,0.231679529,0.0130066602,0.0152296619,0.1907401383,-0.1493622959,-0.1304418594,-0.3635964096,0.0054596518,-0.0161964651,-0.0854954123,-0.0368625186,-0.4798964262,-0.023961775,0.0446463041,0.1415292472,0.1112430543,0.0083408942,0.206122309,0.1885337234,-0.0144463023,0.2635808587,0.3555055261,0.2335148752,-0.2185611576,-0.1486164629,-0.0873393267,0.0437040031,0.177231133,0.1707769781,0.0495433249,0.198735714,0.1211190522,0.0136511913,-0.156597212,-0.1618948281,0.2901345193,0.1482460499,-0.0390661284,-0.0574769676,-0.4436930418,-0.3970943987,0.2507375777,-0.1709859669,0.2013518363,-0.2210760862,-0.0407893099,-0.0284860525,0.0827126801,-0.1632941365,-0.1330616027,0.3956558108,-0.0260782279,-0.0435184315,0.1253520697,-0.0867817253,0.2436129451,0.1832947284,0.0363313928,0.0684396252,0.1590244025,0.509021461,-0.1408690959,-0.282323271,0.0954570919,0.0301749706,-0.1105845273,0.157165274,0.2084631026,0.2270528078,-0.0286244433,-0.0005949585,0.1542893797,0.5449559689,-0.0339258127,-0.1696295887,0.293104887,-0.0085181464,-0.0150808459,-0.1630134434,-0.4817564189,-0.3427665234,0.2621435821,-0.0571562424,0.1788183749,-0.0686620474,-0.0465505086,-0.3099000156,0.1274475306,0.3333838582,-0.3092621863,0.1138573214,0.1015176699,-0.1127338856,-0.1350675225,-0.2750650048,0.0670812726,0.0355206728,-0.1064128503,-0.0677800998,-0.2876259089,-0.2491898388,-0.1500188261,-0.223416701,0.145767495,0.0861834884,0.1993306428,-0.0843114704,-0.0341170803,-0.1865898669,0.0430478305,0.3065355122,-0.1867840141,0.1513662636,0.2028809041,0.0744919926,0.0181314535,0.4157295823,0.2260370255,0.2796542943,-0.0739429668,-0.037148688,0.0357310697,-0.3922090232,-0.3193703592,-0.2047726512,0.0584675558,0.0416636989,-0.0320725292,-0.0897012353,-0.2026488632,0.2067473084,-0.1931208074,-0.1450480074,0.0521599725,-0.4818813503,0.0891080871,-0.1158978939,-0.1679169685,-0.1721760035,0.219962135,-0.1573328823,0.0282431245,0.1448344439,-0.3076137006,-0.1853602976,-0.0703724846,-0.2066439092,0.0887705237,0.1710277051,0.2410681546,0.1262309551,0.1908347756,0.2949064076,0.191714108,0.6330702901,-0.3205320835,-0.4039381444,0.0916703045,-0.2199680656,-0.2654644549,0.0371432118,-0.214447692,-0.1425646842,-0.272359699,0.2915302217,-0.4198493659,-0.1505189985,0.3161565661,0.3153994679,-0.2118120044,-0.453820765,0.0145483138,0.1172990128,-0.221543327,-0.0664238781,0.2788385153,0.1231231242,0.1009590477,0.0123831471,-0.1080928519,-0.1009779349,0.2842496336,0.0370031781,-0.0776291117,-0.0487520881,0.1209403574,0.0725595206,-0.0783501565,-0.1066666991,0.5034688115,-0.2633043826,-0.3774503171,-0.1949518472,-0.0676825792,0.6232670546,0.3686952591,0.1476385593,0.7594145536,-0.3097069263,-0.0316558443,-0.0905638039,-0.4885081947,0.3002042472,-0.0555301085,-0.3722087741,-0.5473954082,0.4530599117,-0.0088565135,-0.0950873196,0.3091556132,0.3756289184,-0.065113768,0.322409451,0.1247900948,0.8138467669,-0.2744164169,-0.1229395196,-0.3575757444,-0.1176925004,0.4589078724,-0.0823843032,0.1852676421,-0.2941487134,-0.3110412061,-0.0497848392,-0.0187119506,0.2967815995,0.2600573003,-0.3409377933,0.0782409981,-0.4393923283,0.0698702037,0.1553504169,0.2486589253,-0.1398492157,-0.2777968049,-0.1104580313,0.1425046623,-0.11223685,-0.146151185,0.1312573701,0.1054700986,-0.0545460805,0.0209372453,-0.4702295363,-0.2768166959,-0.0765133724,0.1107345745,0.0855398625,-0.1964536905,0.332731694,0.2712892294,0.1146991998,0.2238538414,0.1099856868,0.0301439688,0.1507939696,0.0676314831,-0.2251382917,0.0081456276,0.2617019713,0.1127973422,-0.3346254528,0.3105539083,0.1175569221,-0.5033831,-0.3143616021,0.2931494713,0.3977465332,-0.2647148073,-0.4058215022,-0.0048082331,0.1831255108,-0.3095092177,0.1849841177,0.0178682189,-0.019305598,0.3045940697,0.0085479701,0.1418853402,-0.0627700463,0.076244317,0.1791882366,-0.0277929716,0.2033109963,0.15646559,-0.2664813995,-0.2013559043,-0.0680582151,-0.1449985355,0.0511512049,0.1259911805,-0.0413151272,0.0199130606,0.0658484623,0.3281392157,0.0102786254,0.2176457644,-0.3694007397,-0.3393454254,0.0748286024,0.1342157125,-0.065999113,0.2429776192,-0.1170239225,0.0776227117,-0.3198787272,0.2393063903,-0.3837425411,-0.1677518636,-0.3384577036,0.0895199627,0.4663958549,0.0885883197,-0.3127020597,-0.0063825767,0.3497034907,0.1977303475,0.364748925,-0.3827804923,-0.2339254916,0.0406805165,0.1689045131,0.329472214,-0.0043327007,-0.2240867764,-0.2607189417,-0.1267700642,0.1428365707,0.2729929388,0.0092670992,0.1664817035,0.0210776255,-0.2067384571,-0.2106310874,0.1579297483,-0.1462679803,-0.1051932797,0.2947708666,0.1650696695,0.0194176808,-0.011125898,-0.4132359922,0.1817750633,-0.0364187844,-0.0044777188,-0.0662928745,0.0791309252,0.1916416585,-0.0268952362,0.3772665858,0.1328653693,-0.1443950981,0.2026526779,0.2215274274,0.2833153009,-0.3196436763,-0.3168649971,-0.1006961167,0.0153604066,0.3064985573,0.6884785295,0.0813970938,0.2372498363,0.1005137116,-0.3109853566,0.2229533494,-0.0204653107,0.0466402732,0.5463964939,0.1399906129,-0.3324565887,0.2372800559,0.251973629,0.0924372971,0.2494698763,-0.0068376563,-0.0004596483,0.1208663285,0.0719119832,0.3791204393,-0.0603928491,0.5356019735,-0.0790693909,-0.1834257394,-0.1936292648,-0.1061276421,0.0947581455,-0.0275884084,0.0609730035,-0.2364515066,0.3580083847,-0.0336460695,-0.2596780956,0.0461702719,-0.1539089382,0.1392641366,-0.1982196569,-0.0651953742,-0.3060320318,0.0937927514,-0.1541672945,0.404058665,-0.3358256221,0.2959905565,-0.0621732511,0.4309237003,-0.3050548434,-0.0093300175,0.0279920604,0.0790503025,0.1278961897,0.1090109125,0.0680753961,0.2636183202,0.0589405932,-0.1109256074,-0.4688360095,-0.1591546535,0.2030151933,-0.0085284207,0.0494161397,-0.0306812748,0.4116610885,0.2149247229,-0.1834035367,-0.0107745184,0.2067624927,0.3893941939,-0.6877523661,0.1380573809,0.3034943938,-0.2025026083,0.1468922794,-0.0361430272,-0.417655915,-0.2483056039,0.1312694997,-0.0821692646,0.070200935,0.1576975733,0.1691781878,-0.032006301,-0.1209987104,-0.0300240237,0.1371437013,-0.1160302982,0.0243169796,-0.2183522284,-0.2465981692,-0.0963772014,-0.5737490058,0.319041729,0.0891867727,0.1561410874,0.2560800612,0.111494951,-0.0466076434,-0.1248687133,0.5571189523,-0.1855019331,-0.3769250214,0.1967999637,0.0130888568,-0.0289813485,-0.1083410233,0.1454726011,0.1599441022,0.1609099954,-0.3709370196,0.2120975852,0.1374091804,0.0904128477,0.0504023544,-0.0397355705,0.4085710347,0.0949267596,0.144052878,0.0353954621,0.0942068771,-0.2229650468,-0.0063293646,0.1614854336,0.237297833,-0.1276751906,-0.0223604012,0.1613538116,-0.0775214732,0.0235079601,-0.0818515122,-0.1837727278,0.1447621882,0.332929492,-0.13485533,-0.0858381391,0.0106967473,-0.0539036021,0.1986066699,0.1370076835,-0.3265919387,0.2328226268,0.0593990237,-0.338838011,0.0896390453,0.0826264545,-0.0613287501,0.1992231607,-0.2359579057,-0.1262844652,0.3931283057,-0.0934765413,-0.1359141618,0.4957374632,0.1258692592,0.0235412996,-0.1318910122,-0.2299885899,-0.0278316401,0.1875175238,-0.4701178968,-0.2491233945]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2655","title":"Allow the selection of multiple columns at once","comments":"Hi! Sorry for the delay.\r\n\r\nIn this case, the dataset would be a `datasets.Dataset` and we want to select multiple columns, the `idx` and `label` columns for example.\r\n\r\nMy issue is that my dataset is too big for memory if I load everything into pandas.","body":"**Is your feature request related to a problem? Please describe.**\r\n\r\nSimilar to pandas, it would be great if we could select multiple columns at once.\r\n\r\n\r\n**Describe the solution you'd like**\r\n```python\r\nmy_dataset = ... # Has columns ['idx', 'sentence', 'label']\r\nidx, label = my_dataset[['idx', 'label']]\r\n```\r\n\r\n**Describe alternatives you've considered**\r\nwe can do `[dataset[col] for col in ('idx', 'label')]`\r\n\r\n**Additional context**\r\nThis is of course very minor.\r\n","comment_length":45,"text":"Allow the selection of multiple columns at once\n**Is your feature request related to a problem? Please describe.**\r\n\r\nSimilar to pandas, it would be great if we could select multiple columns at once.\r\n\r\n\r\n**Describe the solution you'd like**\r\n```python\r\nmy_dataset = ... # Has columns ['idx', 'sentence', 'label']\r\nidx, label = my_dataset[['idx', 'label']]\r\n```\r\n\r\n**Describe alternatives you've considered**\r\nwe can do `[dataset[col] for col in ('idx', 'label')]`\r\n\r\n**Additional context**\r\nThis is of course very minor.\r\n\nHi! Sorry for the delay.\r\n\r\nIn this case, the dataset would be a `datasets.Dataset` and we want to select multiple columns, the `idx` and `label` columns for example.\r\n\r\nMy issue is that my dataset is too big for memory if I load everything into pandas.","embeddings":[-0.112305522,-0.3079831898,-0.1617852896,0.17450881,0.276807785,0.2954530716,0.4830079079,0.1078617051,0.2709566057,0.4354278445,-0.1071548834,0.1811603904,-0.0057086572,0.1110229865,-0.1205765307,-0.4149233699,-0.1460112482,0.1239385754,0.1093645617,0.1609404683,-0.1801461279,-0.4086984694,-0.2840020657,-0.0011803384,-0.092296347,-0.1219144464,-0.3127509058,-0.3032180369,-0.3770621419,-0.2049789727,-0.0196613744,0.3044013083,-0.0134685794,0.1382125765,-0.0001026555,-0.2590750456,0.1286849827,0.1175028831,-0.244921267,-0.1337271929,-0.2202558815,-0.339571625,0.2827405035,-0.3361616433,-0.0875567272,-0.1732636839,-0.2556638122,-0.1358377486,-0.260029763,0.0658948272,0.3234616816,0.152897656,-0.1889439076,0.0023041901,0.0063465508,0.1277681589,-0.2579077184,-0.1608469784,0.4758292139,-0.1474863589,0.3507170975,-0.0039973366,-0.1867642105,0.0326898396,0.0836206153,0.1034264043,0.2566024959,-0.2503734529,0.1066819206,0.2868083417,0.7255767584,-0.1725481153,-0.4039957821,-0.1537713856,0.5212610364,-0.3376422524,-0.2131011933,0.5081314445,-0.1903934181,0.1000797153,-0.0466958694,-0.3265556097,-0.0729803517,0.2228528261,-0.1081898361,0.2732259333,0.0282639675,0.03926358,0.3872614205,-0.0016128597,0.2152474821,-0.1840659678,0.1768385917,0.2321210653,-0.7773068547,-0.2501906157,-0.2356759906,-0.1116905734,0.2165343165,0.1640052348,0.027720917,-0.1002437547,-0.182840392,0.2871480584,0.4455347359,-0.0151288416,-0.3066451252,0.3155571222,0.1022447273,-0.0562237911,0.1859585643,-0.0062980005,0.0271456502,0.0207334477,0.2195993811,0.1581209749,-0.1780441105,0.0976189449,-0.0328994766,-0.26516518,-0.0894478783,-0.2567326725,-0.0282300059,0.367639184,0.0666408986,0.3588038385,-0.1916099191,0.1649855226,0.1718434393,-0.2895754874,-0.2468194515,0.3928829134,0.0549767315,-0.0291879643,0.2498265207,-0.0250980668,-0.2088542879,0.373406291,-0.0406710505,-0.3881934583,-0.049061574,-0.413400948,-0.014645732,0.2703150511,0.0037309716,0.0997229442,0.1696856171,-0.0508779138,-0.3953902125,0.4790834486,-0.002467064,-0.3585241437,-0.0582644343,0.1234737337,0.2299954742,-0.125632748,-0.2516845167,0.3921030462,0.0448530354,0.4098561406,0.2616648972,0.0018020573,0.0157517269,-0.2986845672,-0.0468573235,0.3674305379,-0.3242014647,0.1272618324,0.0768462941,-0.1038113534,-0.0715092868,0.2260803282,-0.2931676209,-0.0978576913,0.1805267334,-0.013063659,0.2161767185,-0.2442466021,-0.090066433,0.0857360959,0.0422934406,0.0927949771,0.4033911824,0.2204860449,-0.0306513738,0.2627088726,0.057953164,0.3283009529,-0.052146107,0.2525752187,0.1994085312,-0.4831075668,0.2263852805,0.1378232837,0.0445587449,0.0151862036,-0.1832596809,-0.0865488648,0.2737626135,-0.1824587137,-0.1385209858,0.2540957034,0.1792686433,0.0199871175,-0.1123359725,-0.394777298,-0.301106602,0.1568817049,0.4884350598,0.0642902479,0.1696430743,-0.2737737894,0.039078299,0.0045270924,-0.0981860906,0.2652836144,0.1543953717,-0.2062114626,-0.0029782839,-0.1011045724,-0.3993161023,0.4738367796,-0.229688406,0.1321178377,0.0275053978,0.2511871159,0.0649993196,0.124628894,0.164604485,-0.1319043934,0.2843926549,0.1517538428,0.0908201858,0.1125609502,0.1458586454,0.0646826997,-0.3552847505,0.1177156195,0.1671918184,0.1794776767,0.0260349084,-0.3332150877,0.0846312642,0.0246968698,-0.123212263,0.0588422939,-0.3414193094,0.2632939517,0.1129575223,-0.0123408604,0.1553319991,-0.1083925366,-0.1047378629,-0.1918958426,0.1666090488,0.0842899159,-0.2391910702,-0.0248147231,-0.6293264627,-0.1653970778,-0.0721647069,0.1629372388,0.0634677485,-0.0024512354,0.1556061655,0.20158723,0.0425239652,0.1142363697,0.47955966,0.1736518294,-0.2230082452,-0.1962168813,-0.0580710843,0.0025859727,0.183200106,0.2066932321,0.0457415655,0.1509783417,0.2637327015,-0.0220886581,-0.1352917701,-0.0007264406,0.4631795585,0.1383932978,0.0121772531,-0.0780441687,-0.4353890717,-0.2944473624,0.1450490654,-0.2059286833,0.1311485469,-0.0954132006,-0.0634456277,0.1592476219,0.1958408058,-0.2035662979,-0.292973578,0.5270133018,-0.0790494829,0.0136274798,0.1328639686,0.1082940251,0.2760705352,0.1199759021,0.1261203885,0.0356749892,0.1787355095,0.6488932371,-0.1277385652,-0.3254690766,0.1147138327,-0.0649893209,-0.0020178144,0.0803012699,0.0581910424,0.1841234267,0.0059469636,-0.0061487369,0.1868944615,0.4603095055,0.0214347299,-0.1087111086,0.2561409175,0.0313772298,-0.003277506,-0.0456619225,-0.4483949244,-0.3211632371,0.232732892,-0.1309588999,0.1882290244,-0.1114174053,-0.0094253719,-0.4182296693,0.0695729703,0.100777097,-0.2108718604,0.0373317115,0.1368307322,-0.1218452677,-0.1473861635,-0.2751776278,0.1980687529,0.0036962135,-0.0153313,-0.1016477272,-0.18618159,-0.365926832,-0.2651929557,-0.2874382138,0.1633550972,0.1649812758,0.1100185066,-0.0244297124,0.0806228667,-0.2668939233,-0.0728356987,0.345305264,-0.0653501973,0.203998521,0.2225221545,0.0862639323,0.0524850711,0.364397645,0.2674494982,0.3109446466,-0.0122172162,-0.1367586702,0.0528562814,-0.5359568596,-0.3218106329,-0.203645587,0.1261132509,0.0635130405,-0.0032934542,-0.1805378646,-0.2337546349,0.2082329392,-0.0170642193,-0.0829480588,0.0652038231,-0.3010801077,0.2161557972,-0.207605511,-0.349235028,-0.3170901239,0.0577970296,-0.1490228176,0.0522461832,0.0873313174,-0.2652096152,-0.1717193425,-0.0842929333,-0.2234997004,0.0026798104,0.1123490557,0.2241793126,0.0920044035,0.2355044931,0.3232548833,0.2759321928,0.6166089177,-0.1745446026,-0.3837954402,0.0508940071,-0.3853891492,-0.3669198155,-0.0413617902,-0.1233104989,-0.156922549,-0.3198365271,0.4189449847,-0.2965467572,-0.1368431449,0.3415648639,0.3486963212,-0.2732387781,-0.5648306012,-0.0781126842,0.1573353857,-0.1444006711,0.0061028148,0.2095869929,0.1210958734,0.0263989251,-0.0221582688,-0.0957031474,-0.0920027569,0.2835134566,-0.16392681,-0.147869885,0.0068457094,0.2150204778,0.0852211565,-0.2006318271,-0.1354100853,0.4599723816,-0.2418032289,-0.3409645855,-0.1602625996,-0.058792267,0.5828820467,0.3401700854,0.2183081955,0.7634493113,-0.3323463202,0.0773626044,-0.1714491844,-0.4547928274,0.2537869513,-0.1395791471,-0.3544489145,-0.4438308179,0.5387898684,0.0798114985,-0.049132131,0.3948583305,0.3442941904,-0.0674315095,0.0000406218,0.1498573273,0.8275064826,-0.431555301,-0.2111641467,-0.4902248085,-0.0841286555,0.3179245293,-0.1431909353,0.338186264,-0.2494236976,-0.5191368461,0.0597460084,-0.0786459669,0.2748971581,0.1233871654,-0.2035630345,0.0106944358,-0.4162721038,0.0267461576,0.0781420991,0.2290372998,-0.1385079324,-0.3053470254,-0.07368204,0.1564331055,-0.1514736563,-0.1041837409,0.1494265199,0.2186588347,-0.0521988347,0.0623728819,-0.4842542112,-0.1886460632,-0.0390997417,0.2016901523,-0.0052032634,-0.1411771774,0.2614848614,0.2431992292,0.0940678492,0.1636529714,0.1816764325,-0.0763617009,0.0898234993,-0.0293336045,-0.3864740729,0.0532244183,0.1954735667,0.0523706935,-0.3014811277,0.4472315907,0.1216009185,-0.5736987591,-0.4057835042,0.3636288345,0.356204778,-0.1317463368,-0.4770765305,0.0079827765,0.2676302791,-0.2343889475,0.1771584004,0.0703547224,0.0031156442,0.3291621804,-0.0134824011,0.1683643907,-0.1278030276,0.2369877398,0.1456730068,0.0540696345,0.1863732189,0.2619101703,-0.2044371367,-0.1409169883,-0.0470538363,-0.0204169489,0.0557820164,0.0369983278,-0.2269259691,0.0865379274,0.1522174925,0.2939559817,-0.0061811805,0.176974684,-0.2811827958,-0.2921189666,0.0165202785,0.0989243835,-0.0559586883,0.2511584461,-0.1503310055,-0.0370121673,-0.3057663143,0.3603931963,-0.3364816308,-0.090523012,-0.2678503692,0.1122492626,0.4826426804,0.1033248156,-0.4487980902,-0.0426026918,0.2752297521,0.1906019002,0.3579068482,-0.3008052409,-0.2160656154,0.0401901156,0.1674356163,0.282328248,0.0328253806,-0.3111070991,-0.2416770905,-0.0723744109,0.1359636039,0.2352171987,-0.0181063544,0.0346982479,0.1205165908,-0.4068869054,-0.2512097061,0.3073613644,-0.1420863569,-0.1013443321,0.3795372546,0.162858218,0.1169068739,-0.0098360674,-0.5943379402,0.1291301101,-0.0624103248,-0.0212279726,-0.0049673454,0.1171779484,0.22630997,0.0395613648,0.2011488676,0.0559255295,-0.1185511872,0.3107331693,0.1497648507,0.2267699093,-0.238706097,-0.4814903736,-0.2971482873,0.1094649434,0.1556464136,0.6094633937,0.1171147823,0.2621331513,0.041783534,-0.2218045741,0.2776869237,-0.0027745524,0.0446519665,0.5172438622,0.1084972918,-0.3845739961,0.3676688075,0.2722573876,0.032754723,0.0297846757,-0.0118317343,0.0362432823,0.0675960407,0.0136895012,0.368784517,-0.1322773397,0.2324125618,-0.0362096205,-0.1461420208,-0.1416944712,-0.1570498645,0.1373042911,-0.0495222434,0.0644794777,-0.3368619978,0.3564508259,-0.014976386,-0.1321482807,0.1006447747,-0.1782672256,0.0650985017,0.0190557502,0.0225722529,-0.2802747786,0.3539918959,-0.0896716267,0.3283382356,-0.2615713179,0.3097865283,0.2304595709,0.3455485404,-0.3301073909,-0.1253711581,-0.0218322538,-0.0188373867,-0.0075642946,0.0874795765,-0.1043459028,0.2899871767,0.2223846912,-0.0856919065,-0.4326574504,-0.1236099005,0.245962888,0.0005929975,0.1164963394,0.010341634,0.3054492474,0.1678983569,-0.1868195385,0.0468549728,0.1130276173,0.4121386409,-0.8193488121,0.099372685,0.3142160475,-0.2676081061,0.0531292967,0.0157383345,-0.3119893968,-0.1369479448,0.1825878769,-0.0499920174,-0.0048553268,0.1363015324,0.1282060891,-0.1264782399,-0.035324771,-0.04395172,-0.0469574071,-0.1979511231,-0.1068289578,-0.2249566615,-0.2449551076,-0.2073522061,-0.4267042875,0.4060287774,0.0711869374,0.0132246427,0.2281348109,0.1053836942,-0.1622509658,0.0006872828,0.5898199677,-0.2509422898,-0.3970303237,0.2623259723,0.0200808346,-0.0306449514,-0.089848727,0.2543352544,0.1397684515,0.1256917417,-0.4077950418,0.2397902608,0.0784811303,0.2195717543,0.1887874156,-0.1125884056,0.2785730958,0.1620334834,0.1520570666,-0.048220966,0.1639133096,-0.1264466941,-0.0584520623,0.1564704031,0.2043753415,-0.2094189823,-0.0072863111,0.1009927243,0.0564740859,-0.1597691923,-0.1402171403,-0.0518520549,-0.0705242157,0.3543420732,-0.0740791261,0.0884364173,-0.0341247879,-0.070260033,0.1476794481,0.1522465944,-0.2481177151,0.0753122419,0.0907553807,-0.1930681467,-0.0810833052,0.1452884674,-0.1062946767,0.2746168375,-0.2596270442,-0.1857055873,0.3158004284,-0.1611783803,-0.0726443827,0.5669779181,0.1610433459,0.0384031869,-0.1069220752,-0.1774678081,0.0077993493,0.2639756203,-0.4143879414,-0.1318200827]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2654","title":"Give a user feedback if the dataset he loads is streamable or not","comments":"#self-assign","body":"**Is your feature request related to a problem? Please describe.**\r\nI would love to know if a `dataset` is with the current implementation streamable or not. \r\n\r\n**Describe the solution you'd like**\r\nWe could show a warning when a dataset is loaded with `load_dataset('...',streaming=True)` when its lot streamable, e.g. if it is an archive. \r\n\r\n**Describe alternatives you've considered**\r\nAdd a new metadata tag for \"streaming\"\r\n","comment_length":1,"text":"Give a user feedback if the dataset he loads is streamable or not\n**Is your feature request related to a problem? Please describe.**\r\nI would love to know if a `dataset` is with the current implementation streamable or not. \r\n\r\n**Describe the solution you'd like**\r\nWe could show a warning when a dataset is loaded with `load_dataset('...',streaming=True)` when its lot streamable, e.g. if it is an archive. \r\n\r\n**Describe alternatives you've considered**\r\nAdd a new metadata tag for \"streaming\"\r\n\n#self-assign","embeddings":[-0.386490494,-0.0753532574,-0.1559201628,-0.0280467253,0.0786294937,-0.0864201337,0.0634772778,0.2331114858,0.0479878187,0.1733263731,0.242780298,0.2553816438,-0.3610651791,0.2538989484,-0.2390858084,-0.0682933405,-0.1925243139,0.2463821322,0.0271420646,0.0251488984,-0.0089142742,-0.3277525604,0.1594410092,-0.2067571729,0.1384985596,0.0589002632,0.2777647078,0.0063075176,-0.414856106,-0.5290967226,-0.1042208523,0.6001194715,-0.0037718576,0.153514564,-0.0001013907,-0.1611902416,0.4878281057,-0.0567288809,-0.4370283484,-0.0489433035,-0.324377954,-0.0845014602,0.4111237824,-0.1137456596,-0.1491480172,-0.0755111799,0.1785855442,-0.3399660289,0.3254101276,0.1559751481,0.2582333684,0.1479986459,-0.1959843785,0.0402494185,0.181825459,0.3420345783,-0.28776595,0.1714212149,0.1897300333,0.3205068409,-0.289975673,0.4419681728,-0.1764093637,0.0894802585,0.4129264057,-0.2009213418,-0.1425331235,-0.1245672479,-0.1643987,0.358284384,0.8423913121,-0.1444597393,-0.1336076856,-0.2743810415,0.1760063171,-0.1738714427,0.0280286893,-0.0029780415,-0.2524289489,0.1886110008,-0.2293714285,-0.2240005732,-0.3321895301,0.0384190269,0.0779398605,0.2990438938,0.0845322981,-0.0571222641,0.1952447146,0.1180684939,0.3494934738,0.1036706492,-0.0743715018,0.0327841416,-0.3043943048,-0.269036442,0.1546379477,-0.1140342429,0.2391618043,0.2554944158,0.4104701877,0.1414100528,-0.0745477602,0.1982324868,0.23473607,-0.0366856977,0.2107940614,-0.3249052465,0.2578808963,0.0731756762,0.0688936338,-0.1524994075,0.2214730233,0.2359937131,0.2969708145,0.1303084046,0.369269073,-0.2136805803,-0.1043363065,0.0472263992,-0.0618871637,-0.3613417447,0.0947753713,0.2525915802,-0.1362878978,0.2417262048,-0.1015288085,0.0339214318,0.2017615438,-0.4130122662,-0.0291694608,-0.1372924745,-0.308766067,0.1712158322,0.1471936703,-0.4777767956,0.0035696139,0.0281289648,-0.0438351892,0.0312227197,0.3571619391,-0.0611580946,0.0261421297,0.1617662013,-0.2072983235,-0.0787619948,0.1240055114,0.2488883883,-0.2877151966,0.2110588998,0.129071936,-0.2976518869,0.04873088,0.3097337484,-0.1082197875,-0.1297412217,-0.1469227672,0.5340294242,-0.4985406995,-0.2949351668,0.1016642377,0.3449594975,-0.0501868166,-0.2100450546,-0.06628225,0.4946549833,-0.2345071286,-0.1876344979,-0.4066006243,-0.453215301,0.2455570698,0.0174971297,-0.2248364836,0.0151387788,0.1544627845,-0.0119283339,0.6330034733,-0.0929674655,-0.0220277309,0.5101202726,-0.2103195041,0.2227145731,0.3376653492,0.3194203973,0.1987773478,0.0053143068,-0.042483896,0.2768621445,-0.0648584291,0.0540961362,-0.287160933,-0.3000769913,0.2274936438,0.1707745194,0.0707281381,0.1629011035,0.1399580389,-0.066685155,0.164915204,0.1333294362,-0.0949560255,-0.2224206179,0.418296814,0.0876821429,-0.3319534361,-0.0345390402,-0.2437539399,0.0044144453,0.1239750087,-0.043611493,0.2876324654,-0.2957673371,-0.3009924591,0.0830493569,-0.010914471,0.0295917671,0.286051482,0.3132918477,-0.3593302369,0.0095459688,-0.3415893912,0.2999947965,-0.2890333235,0.0902469605,-0.0713141039,-0.0104485955,0.1581583917,-0.0419881828,0.1381871998,-0.2274726182,0.2333584428,0.143493861,-0.1355366409,0.3913865983,0.0471992567,0.3627124727,-0.0107481014,0.2678534985,0.131773144,-0.1706988364,0.2855980992,0.174322933,0.0315296166,-0.0013318878,-0.2587227821,0.0573218539,-0.0597457774,-0.0432546102,-0.0068890909,-0.2404403687,0.0607716367,-0.0257887393,-0.3275288641,-0.0466867089,0.1257131249,-0.0016350617,-0.1315074861,0.1458145827,-0.3788232803,-0.0767758191,0.2132112235,0.0797405988,0.0439144187,0.4495173097,0.015262072,-0.3189000785,0.3321025074,0.1454902589,0.4745833278,0.336741209,0.1322947294,0.0025153621,0.184167549,-0.1093984097,0.0638225526,-0.1445942223,-0.0559787117,0.1008213162,-0.0075798319,0.0207561683,-0.3501297832,-0.2287765741,0.0013678552,0.1049418971,-0.0414789617,-0.2181842476,-0.0875054896,-0.3902789354,0.056261126,-0.2072051018,-0.0453517362,-0.1313884556,0.3408313394,0.5653298497,-0.2346772552,0.0528414175,-0.2678106129,0.4714451134,-0.1547592729,0.0246415846,-0.3929772377,-0.0915554836,0.1295676082,0.2055664361,-0.0265444405,-0.016119618,0.750626862,0.2778375149,0.366250217,-0.3650811315,-0.2284796685,-0.0036857333,0.1561871916,0.0390927307,-0.0295950733,0.4245865047,0.3441903889,-0.4065552354,0.1335194856,-0.0926809385,0.027522102,0.0707594678,0.0866938755,0.1709905565,-0.0403745472,-0.2682164013,-0.3047307134,-0.6813281178,0.1347262263,-0.0960229114,0.0999438092,0.1018745974,0.0407117829,-0.1468031555,0.0629811361,0.2010446638,-0.1741165817,-0.3040235639,-0.109929353,-0.0905360729,-0.0227535795,0.0280689429,0.1720814407,-0.0355349928,0.421189487,-0.3504660726,0.0179470126,-0.0364635587,-0.1187149361,-0.2084804326,-0.5121216774,0.2066882849,0.3022755384,-0.2924049199,0.0195523743,-0.0547157228,-0.3162848651,-0.0148101486,0.0730249509,0.0388066806,0.4326710105,-0.0085124718,0.4920962453,-0.0124553787,-0.14186728,0.4557244778,0.1649619788,0.3440112472,-0.040091604,-0.1373035163,0.0910381898,0.1746022105,-0.1724781096,0.2750851214,-0.0359121189,-0.3822969496,-0.3068858683,-0.0848566964,-0.2187092006,-0.0986485183,0.3085207343,-0.1086291224,0.1687934548,-0.1506562531,0.016404856,0.1223025396,-0.0766107962,-0.0681617558,0.3103946745,0.3248276711,-0.1768766493,-0.1401697993,0.1725437045,-0.1663061827,0.1186603308,-0.1609476954,0.0915234983,-0.3280895054,-0.1512720585,0.0324046053,0.1340192407,0.2315284014,-0.3784803152,-0.2304242551,-0.078838326,0.1506970078,-0.0414942577,-0.0093249343,0.0967480242,-0.0032674123,-0.1742426306,0.0415550284,-0.3080045581,0.0384602025,0.071390301,-0.0734200776,-0.0917156562,-0.3896146715,-0.0744042024,0.245029971,-0.1937759817,0.2198937982,-0.1852602214,-0.3441869318,-0.1751682609,-0.3560366631,0.1141331047,0.1625955105,0.054021351,0.3290721774,0.0317696072,0.096235849,0.0008069733,0.0986881852,0.0979772657,0.1683328152,0.2112763524,0.2084836066,-0.1788799614,-0.1571086645,0.0959596634,0.1112223491,0.1934203804,-0.0119561926,0.3724042773,0.32226336,0.1456931978,-0.2843781114,0.1671568304,0.2163541466,-0.3190145493,-0.3587409854,-0.6744565368,0.6354106665,-0.1098569483,-0.254442066,0.4636869133,-0.2186401337,-0.207144171,0.1145215034,0.0928117186,1.0428253412,-0.095379144,-0.2611708939,0.0537584983,-0.2616550922,0.2489877641,-0.5175178051,-0.0788507387,-0.1053170338,-0.3536138237,-0.1829101592,-0.0399040729,0.1961517334,0.0577482805,-0.6106126308,-0.1492632031,-0.0620210655,0.1171147674,-0.0216192994,0.5512166023,-0.3129017949,-0.2414316684,-0.5524566174,0.2269751281,-0.3169329762,0.1102489978,-0.1375515312,-0.1036001071,0.5193989277,-0.1348026544,-0.3839791119,-0.1183072254,-0.2465753555,-0.2308535874,-0.5191705823,-0.1036086753,-0.0573422946,-0.069024235,-0.3273700476,0.0388438627,-0.1493523717,0.1899003088,-0.0478644222,0.2089725584,-0.167898342,0.0163508132,0.2850864828,-0.2037527412,-0.1827045977,0.4371204376,-0.0526351593,-0.1962835044,0.0552268401,-0.1813160777,0.4153044224,-0.3604307473,0.010061468,0.0240501761,-0.2315644473,-0.0141730355,0.2120159715,0.1421414465,0.0324524306,-0.0152652217,0.2309588343,-0.0912485644,-0.2930075526,0.1555221677,-0.0019937891,-0.1709388047,-0.0944703817,-0.16888991,-0.2549053133,-0.3477312624,0.0589243919,-0.0221512299,-0.0375199355,-0.0022592482,0.1511002332,-0.0723831877,0.1386438906,-0.0364303328,0.185151577,0.0989611745,-0.2004152983,-0.1056446061,-0.2652685344,0.0690708011,-0.0429967791,0.2820933461,-0.3990847766,-0.354585439,0.0943420008,0.0893232077,-0.4334770143,-0.5305477977,-0.1142167896,0.1001931801,0.1174969673,0.2416098565,0.0852831304,-0.1517283618,0.1409555674,-0.0370960683,-0.1289097965,-0.2452014834,-0.1464660615,0.1514444053,0.1450344026,0.2007514983,-0.2104429305,-0.2172721028,-0.0814302713,-0.088899374,0.2240416408,0.0159167666,-0.1856235564,0.1236376986,0.361046046,0.1959607005,-0.3630832434,0.0647071823,-0.0123617975,0.0503731668,0.0618870743,-0.2005374432,-0.0600579046,-0.1068119258,-0.0835602731,0.156441614,0.4484623671,0.1612438709,0.0590756424,0.2860468924,0.096855253,0.0025860725,0.352507472,0.1642947942,-0.0115747638,-0.0670436546,0.1668884754,0.3464700282,-0.0685666427,0.072416462,-0.1588343084,-0.1151739135,0.2487585098,-0.128529653,-0.007571104,0.2453185767,0.3719063103,0.0056796167,0.0348037705,-0.1722894609,0.4197755754,0.2638269961,-0.0195282344,-0.0444250964,0.1603088528,0.1229333133,0.1593213677,-0.0160727408,0.0469859913,0.4493066072,0.2122904509,0.0261509065,-0.1320697367,-0.1583087146,-0.0228260346,0.3569809198,-0.1052950025,0.403321296,0.3790698349,0.0896330625,0.1688013673,-0.0396662802,-0.2373590767,-0.1408924311,0.0932177603,0.2884298563,-0.1764384508,-0.2027959526,-0.2742438316,-0.0695189834,0.1057040393,0.1660540849,0.0842161328,0.1364714652,0.1353806853,-0.2125350088,0.0534853004,-0.0281233937,0.3186872602,-0.3287459612,-0.108437553,0.1644595712,0.1076296121,0.4938857555,0.2233679891,-0.0701683834,0.317428261,0.1703831404,0.2301032841,-0.0594775937,-0.0538664013,-0.1752513796,0.1997669935,0.3602409065,0.0855414718,0.0493798479,0.2883541286,-0.280139178,0.4506226778,-0.1870567203,-0.1155654937,-0.3012575209,0.0429574214,0.5293895602,-0.2633572519,-0.1550098509,-0.115014188,-0.3124526143,-0.0592349768,0.2556221783,0.0575619042,-0.0949539617,-0.2491535842,0.1370812058,0.255174011,0.3539218307,-0.0366621204,-0.0418637916,-0.1545535773,-0.1143778861,-0.654438436,-0.0306754299,0.0356395952,-0.0060570445,0.2496453971,-0.1033457592,0.1173445359,0.5393981934,-0.1397619545,-0.2003513575,-0.1015393808,0.0693836585,-0.4219610095,-0.2018654794,0.2489633113,0.0128476238,-0.1240536794,-0.0354168713,0.2024315894,-0.0706161708,0.1403754652,-0.1926916391,-0.2105874568,0.084573321,0.2914016545,0.4421469569,-0.0333758891,0.4754070938,-0.2047967762,-0.0165754035,0.0023529767,-0.1319050342,-0.1245098189,-0.0406457782,0.2259453982,0.3850391209,0.1946251988,0.0559416488,-0.4384351075,0.4731435776,-0.1362621486,-0.126369223,-0.2230929881,-0.1108299717,0.1509546489,0.0216308087,-0.1064456627,0.0754647925,0.0327274092,-0.0633347332,-0.302195549,-0.0948974565,0.5520768166,-0.1305140108,-0.0535188206,0.0398500152,0.0884915516,0.1312378645,0.1925837994,-0.5250589252,0.2096817791,0.2981042564,-0.2159511149,0.308264643,0.330135107,0.1423274875,0.0471516587,0.0012471846,0.2194073945,-0.009425737,-0.0672876537,-0.0250651445,-0.2113328427]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2654","title":"Give a user feedback if the dataset he loads is streamable or not","comments":"I understand it already raises a `NotImplementedError` exception, eg:\r\n\r\n```\r\n>>> dataset = load_dataset(\"journalists_questions\", name=\"plain_text\", split=\"train\", streaming=True)\r\n\r\n[...]\r\nNotImplementedError: Extraction protocol for file at https:\/\/drive.google.com\/uc?export=download&id=1CBrh-9OrSpKmPQBxTK_ji6mq6WTN_U9U is not implemented yet\r\n```\r\n","body":"**Is your feature request related to a problem? Please describe.**\r\nI would love to know if a `dataset` is with the current implementation streamable or not. \r\n\r\n**Describe the solution you'd like**\r\nWe could show a warning when a dataset is loaded with `load_dataset('...',streaming=True)` when its lot streamable, e.g. if it is an archive. \r\n\r\n**Describe alternatives you've considered**\r\nAdd a new metadata tag for \"streaming\"\r\n","comment_length":30,"text":"Give a user feedback if the dataset he loads is streamable or not\n**Is your feature request related to a problem? Please describe.**\r\nI would love to know if a `dataset` is with the current implementation streamable or not. \r\n\r\n**Describe the solution you'd like**\r\nWe could show a warning when a dataset is loaded with `load_dataset('...',streaming=True)` when its lot streamable, e.g. if it is an archive. \r\n\r\n**Describe alternatives you've considered**\r\nAdd a new metadata tag for \"streaming\"\r\n\nI understand it already raises a `NotImplementedError` exception, eg:\r\n\r\n```\r\n>>> dataset = load_dataset(\"journalists_questions\", name=\"plain_text\", split=\"train\", streaming=True)\r\n\r\n[...]\r\nNotImplementedError: Extraction protocol for file at https:\/\/drive.google.com\/uc?export=download&id=1CBrh-9OrSpKmPQBxTK_ji6mq6WTN_U9U is not implemented yet\r\n```\r\n","embeddings":[-0.3121467233,0.1084250659,-0.0998005643,0.0617054515,0.1377473921,-0.1033977941,0.1825834215,0.2757679522,-0.0348217525,0.2482772321,0.23722516,0.2212012112,-0.41875422,0.2228206396,-0.2498128265,-0.1057739332,-0.1940762699,0.2266800851,0.1973604113,0.0241289698,0.050080765,-0.1404721588,0.125907585,-0.2275450975,0.0955903977,0.040822193,0.2770479918,0.0032736778,-0.4309591651,-0.4094036222,-0.1264721453,0.5943691134,0.1119763926,0.3047804236,-0.0001064388,-0.1363653243,0.4294652045,-0.0646332726,-0.4166698158,-0.2811459601,-0.2709993124,-0.04569529,0.3828910887,-0.0648992881,-0.2366087437,0.0000304834,0.1893444657,-0.2507543266,0.398091197,0.3040390313,0.1841346622,0.1431601793,-0.1689657271,0.1103191748,0.2620275021,0.2505020797,-0.2353683561,0.1992798001,0.134420082,0.3304212689,-0.3621218204,0.3760129213,-0.1080571488,0.0278845839,0.2396207899,-0.1162741035,-0.2675046921,-0.3587150276,-0.1097213551,0.3916170299,0.8121817112,-0.0677935779,-0.266923964,-0.2799865305,0.2192740738,-0.1652314067,0.2707595229,0.0779115707,-0.3087148368,0.2289714813,-0.3739491403,-0.3058995605,-0.4223290384,0.0608525462,0.0020896047,0.2876546085,0.0454271957,-0.0564265884,0.2633794248,0.184820801,0.22921893,0.0431781076,0.023470154,-0.0227358416,-0.3078721762,-0.3380286992,0.0473389663,-0.1607312709,0.1749161631,0.3495371044,0.3600506186,0.1747979075,-0.0686454102,0.2610440254,0.3030857444,-0.0043789367,0.1163038835,-0.2984133065,0.3219976127,0.1630396098,0.139344126,-0.1453951448,0.1518913358,0.1023415998,0.1967503428,0.1389230788,0.4190607369,-0.17866081,-0.1202374622,0.044085402,-0.2055466771,-0.3528243303,0.0303816646,0.2117040008,-0.2265862525,0.1878919005,-0.1177252755,0.1621883065,0.1279754937,-0.3675708175,0.0093998937,0.0382103808,-0.2182808965,0.203861773,0.1490737647,-0.5337638259,-0.0831713602,0.0130760521,-0.1549495608,-0.0313117877,0.3512848914,-0.1534520835,-0.0033806057,0.1138476729,-0.016180858,-0.0984127149,-0.0082277646,0.1018308923,-0.260963589,0.2310131639,-0.0887562186,-0.2245736569,0.0018799404,0.2684470117,-0.234155044,-0.153254807,-0.1981976628,0.5610991716,-0.4761672318,-0.2225428671,0.1492543668,0.3053433895,-0.0972417444,-0.1726650596,-0.0284204539,0.4456055164,-0.2409359366,-0.1790953726,-0.3171626031,-0.4264673293,0.3974467516,0.0774606764,-0.1762910336,0.0792955533,0.0524514392,0.0932530761,0.5550786257,-0.0697788671,-0.1110180095,0.5365390182,-0.3236947656,0.3182698786,0.3217315972,0.13688308,0.3014834225,0.0313737877,0.0506621562,0.3943997324,-0.097603865,-0.0137693277,-0.313816458,-0.2959581614,0.1970930994,0.2616508305,0.1503327042,0.1946424991,0.1730065495,-0.1789409816,0.1596208215,0.0251363069,-0.0353978425,-0.0746823326,0.4696786702,-0.0059243557,-0.3908084631,0.0882130414,-0.3534206748,0.0092368284,0.1217710376,0.0998350084,0.0756967887,-0.3154946864,-0.1760160923,-0.1409995705,-0.1186433136,0.0458630547,0.1992529184,0.2442771494,-0.2380592078,0.0990362465,-0.2280892432,0.2491263598,-0.2165423781,0.2083421946,-0.2282067388,0.1588588208,0.1062748432,-0.0724145398,0.1579423845,-0.302803129,0.1809504479,0.0156517364,-0.0901722163,0.3711771369,0.0464416705,0.5940019488,0.071555391,0.2712281346,0.2272628695,-0.2836053371,0.1909567863,0.3033688664,0.0110063823,0.0073280735,-0.3751382232,0.2143037617,0.1132769063,-0.1076527461,0.1807173342,-0.1787669212,0.1305251271,-0.0777626261,-0.3311727643,-0.0216072369,0.1609404981,0.0429362245,-0.1634137034,0.0293836836,-0.3633941114,-0.1784788668,0.2748422027,0.0663698167,-0.0496440642,0.5686532855,-0.0119932629,-0.3284745216,0.2952456772,0.215292111,0.3742587268,0.2192719579,0.2608889043,0.1018619016,0.3289047182,-0.1284110546,0.1090840995,0.0433938652,-0.1185555309,0.1851857901,0.0774018094,0.0425153822,-0.4282302558,-0.1988510191,0.0296159647,0.1814826429,-0.0836001039,-0.0599258654,-0.0986200348,-0.1597470939,0.0016256224,-0.1829215139,-0.1229931936,-0.122131899,0.2737377882,0.6344076395,-0.2617421448,-0.0066787251,-0.2472235411,0.4284985065,-0.1542272568,-0.0365397893,-0.3052411079,0.0085203778,0.1265051067,0.1804899424,0.1358112246,-0.0603530928,0.7215875387,0.2235303223,0.2180972099,-0.4833516181,-0.2957260311,0.1400237232,0.0619995482,0.0927764699,-0.0645019412,0.2419243008,0.1780843586,-0.4123814404,0.0639034808,-0.0645541772,-0.014477253,0.2115996182,0.0463888086,0.2912009358,-0.0600490235,-0.4394127429,-0.378875792,-0.528772831,0.2629510462,-0.0132583296,0.0757973269,0.2690798938,0.0719243214,-0.1398300081,0.1498427093,0.0919446871,-0.1639776975,-0.3578228951,0.0425293148,-0.1043366268,-0.0163401142,0.0637978762,0.1183894575,-0.1711734384,0.5343636274,-0.385127157,-0.0583212338,-0.1136077791,-0.2023053318,-0.1155278906,-0.5180271864,0.2743529081,0.157575056,-0.2281766981,0.1127378047,-0.0284653194,-0.2289383858,0.1795026213,0.1157948822,0.0684842169,0.418427527,0.0242061075,0.3933297098,0.0401668958,-0.0478451028,0.4635876119,0.2107439041,0.2624152601,0.0386320874,-0.0146240201,0.1145379841,0.0727314949,-0.3440086544,0.3816343844,-0.0103282277,-0.3619115353,-0.3208196461,-0.0507948287,-0.2861067951,-0.1568214893,0.318870455,-0.0335056856,0.1813454926,-0.1649651378,0.0061604753,0.2549753785,-0.2481225133,0.0024467711,0.3957547843,0.4271703959,-0.2007686645,-0.2212179899,0.166821003,-0.3207583725,0.1439557374,-0.0952143371,0.1345741153,-0.2814964354,-0.1487432122,0.1307980269,0.0471016988,0.1721300036,-0.503760159,-0.3077434599,-0.1230988875,0.1804492921,0.1246237755,-0.1156438366,-0.1346152723,-0.0081135454,-0.0841058418,0.2051527947,-0.329757601,0.052274622,0.1061700508,-0.1360592991,-0.1200311929,-0.3209198415,-0.083951436,0.1141536236,-0.2305552661,0.1754379719,-0.1583168656,-0.2460886091,-0.3209343851,-0.32352826,0.2442778945,0.1414483786,0.2715743184,0.2621294856,0.0992308557,0.0347510315,0.0128474589,0.0679872856,0.0814794824,0.3601877689,0.4002394974,0.1736349016,-0.2835341096,0.0120990025,-0.1100340933,0.1290238202,0.2547206879,-0.0725501478,0.3577904403,0.3873589635,0.2700962722,-0.3172830939,0.1379453093,0.2631407082,-0.3558495939,-0.3819205761,-0.5331365466,0.5077525973,-0.0340490378,-0.2249023318,0.3006294966,-0.1397238672,-0.2113225162,0.1403768212,0.0077697863,1.1393032074,-0.0813356936,-0.2863714993,0.028147636,-0.2087207586,0.3228910565,-0.6179022789,0.0171213895,-0.1327885091,-0.383449167,-0.2388115525,-0.0798731744,0.1985814422,0.1235038638,-0.6832168102,-0.0016386037,-0.0893844813,0.1861293763,0.0043755574,0.637283206,-0.3760340214,-0.1759621948,-0.4483305514,0.139435485,-0.244715035,0.2459938228,-0.1275124401,-0.1912702173,0.4177613854,-0.1814320236,-0.3381588161,-0.1181489453,-0.3152578771,-0.2668800354,-0.461450994,-0.3543007374,0.0680491254,0.0507476665,-0.1740121543,-0.0937679261,-0.244910717,0.1707958579,-0.17932643,0.1476754248,-0.1403568834,-0.0248742569,0.4186308682,-0.2008976936,-0.0720935985,0.4372836351,0.0124198608,-0.3833431602,0.1031074822,-0.2847840488,0.2615044415,-0.287712127,-0.0195643529,-0.0967934355,-0.3447515368,-0.1151564121,0.1612796485,0.0826521292,-0.009265095,-0.0857049078,0.2834558487,-0.0651963577,-0.2721075416,0.3171851933,0.0930131599,-0.2397344559,0.0075635994,-0.025570102,-0.3257597685,-0.3389294446,-0.140324086,0.1201830134,-0.0249485355,-0.0174279865,0.04686784,0.0907500833,0.0668364614,-0.046154391,0.1389410645,0.1220056862,-0.0399709232,-0.1642753035,-0.2067984939,0.0837500393,-0.0763845742,0.2557650805,-0.3581160903,-0.2978525162,0.1445733905,0.0498078093,-0.3824309111,-0.334664613,-0.1456539631,0.0220011361,0.2006407678,0.1348557323,0.2724148035,-0.1450314224,0.0851256028,-0.0014294925,-0.1600330025,-0.1832202524,-0.0749987587,0.1762990803,0.2494759709,0.0802872181,-0.1696700305,-0.154597193,-0.1127115414,0.031631235,0.2693580687,0.0863947645,-0.107670404,0.1121908724,0.2154077142,0.2012629658,-0.4698934257,-0.1330149919,-0.1361059397,0.2858886719,0.1375322789,-0.1758603305,-0.0233396329,-0.0657681525,-0.2573368549,0.1921193302,0.5488527417,0.2320527434,0.1128172055,0.1030887589,0.0854143798,-0.1116761491,0.3619665802,0.3429034352,-0.2478907853,-0.0252952464,0.1975699067,0.3165720999,0.0372548401,-0.0117167709,-0.1420812458,-0.2307102233,0.1755717248,-0.0492806882,0.12661542,0.2202084064,0.3079924285,0.0576988235,0.0436656475,-0.1060422137,0.3654940724,0.2247043252,-0.0537398756,0.0165367164,0.1116277054,0.0952803269,0.0640545711,-0.144019857,-0.0732365698,0.4655786455,0.1421111077,0.0219196193,-0.1203843132,-0.3483928144,-0.0647101998,0.4036747515,-0.1332679391,0.3305848837,0.3621432483,0.2371151447,0.1317166239,0.0474815592,-0.2780638933,-0.1552162319,0.0312146787,0.3018753827,-0.1466694176,-0.3033553064,-0.3367460668,0.089594841,0.053884048,0.1774701625,0.0236671306,0.2555823028,0.0480021723,-0.0502861254,-0.0424842946,-0.1595510989,0.3600115478,-0.40731287,-0.0228468068,0.0944221541,0.1864829808,0.4190039933,0.1505408883,0.0278033912,0.2221666425,0.0606835559,0.1219135448,-0.0456663482,-0.0402871482,0.0180277824,0.2732661664,0.3632723987,-0.0060758744,-0.0276233144,0.2412563711,-0.2499462664,0.4658037424,-0.1158836037,-0.1067913994,-0.3139313757,-0.0614184216,0.5476936698,-0.2058524042,-0.201374948,-0.3312001228,-0.2827060521,-0.056299679,0.2169641405,0.0852731764,-0.0955080241,-0.1160291955,0.0990270749,0.2164182812,0.1388693899,0.1311117709,-0.0507129952,-0.2327347398,-0.1479548067,-0.6127151847,0.1011702716,-0.0141725615,0.0776274875,0.1402823627,0.0924198329,0.0372856408,0.5932241082,-0.2085782588,-0.2035126984,-0.1500034034,-0.0502964184,-0.4235230088,-0.1181755066,0.3140201569,0.035892535,-0.2146659791,-0.1417931318,0.1993782669,-0.1184097752,0.0649343133,-0.2269933969,-0.1133242622,-0.044030305,0.2698156834,0.5836316943,-0.0327829011,0.600011766,-0.2425941378,-0.0778988004,0.0491771996,-0.0037097405,-0.1775400341,-0.0804919973,0.1503085196,0.511254251,0.3086289763,0.0521610454,-0.4756872356,0.5256850719,-0.1372179985,-0.1262235343,-0.3493207395,-0.0594529137,0.2060448825,0.1305942386,-0.0129764853,0.0898264125,0.1220912039,-0.1428586096,-0.4350024164,-0.2368573248,0.4311558306,-0.1696451455,-0.0917342007,0.0775055215,0.1349052042,0.1489223093,0.291703701,-0.6755431294,0.2758554518,0.2459852397,-0.2296231985,0.284226954,0.3128637373,0.2513948083,0.0803821012,0.0008700258,0.2112134099,0.0321697704,-0.1610609442,-0.0838166252,-0.246906206]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2653","title":"Add SD task for SUPERB","comments":"Note that this subset requires us to:\r\n\r\n* generate the LibriMix corpus from LibriSpeech\r\n* prepare the corpus for diarization\r\n\r\nAs suggested by @lhoestq we should perform these steps locally and add the prepared data to this public repo on the Hub: https:\/\/huggingface.co\/datasets\/superb\/superb-data\r\n\r\nThen we can use the URLs for the files to load the data in `superb`'s dataset loading script.\r\n\r\nFor consistency, I suggest we name the folders in `superb-data` in the same way as the configs in the dataset loading script - e.g. use `sd` for speech diarization in both places :)","body":"Include the SD (Speaker Diarization) task as described in the [SUPERB paper](https:\/\/arxiv.org\/abs\/2105.01051) and `s3prl` [instructions](https:\/\/github.com\/s3prl\/s3prl\/tree\/master\/s3prl\/downstream#sd-speaker-diarization).\r\n\r\nSteps:\r\n- [x] Generate the LibriMix corpus\r\n- [x] Prepare the corpus for diarization\r\n- [x] Upload these files to the superb-data repo\r\n- [x] Transcribe the corresponding s3prl processing of these files into our superb loading script\r\n- [ ] README: tags + description sections\r\n\r\nRelated to #2619.\r\n\r\ncc: @lewtun \r\n","comment_length":94,"text":"Add SD task for SUPERB\nInclude the SD (Speaker Diarization) task as described in the [SUPERB paper](https:\/\/arxiv.org\/abs\/2105.01051) and `s3prl` [instructions](https:\/\/github.com\/s3prl\/s3prl\/tree\/master\/s3prl\/downstream#sd-speaker-diarization).\r\n\r\nSteps:\r\n- [x] Generate the LibriMix corpus\r\n- [x] Prepare the corpus for diarization\r\n- [x] Upload these files to the superb-data repo\r\n- [x] Transcribe the corresponding s3prl processing of these files into our superb loading script\r\n- [ ] README: tags + description sections\r\n\r\nRelated to #2619.\r\n\r\ncc: @lewtun \r\n\nNote that this subset requires us to:\r\n\r\n* generate the LibriMix corpus from LibriSpeech\r\n* prepare the corpus for diarization\r\n\r\nAs suggested by @lhoestq we should perform these steps locally and add the prepared data to this public repo on the Hub: https:\/\/huggingface.co\/datasets\/superb\/superb-data\r\n\r\nThen we can use the URLs for the files to load the data in `superb`'s dataset loading script.\r\n\r\nFor consistency, I suggest we name the folders in `superb-data` in the same way as the configs in the dataset loading script - e.g. use `sd` for speech diarization in both places :)","embeddings":[-0.2387332767,-0.1557696015,0.0066516008,0.1112360433,0.3777109385,-0.1991261393,0.008201831,-0.1370199472,0.1117523089,0.3464581966,-0.3380066156,0.4965123236,-0.1501338184,0.4443328083,0.2138432264,0.2012203336,0.0950579941,0.2073054612,-0.2991645336,0.0513462573,-0.1497841179,-0.11977157,-0.1633818448,-0.1436896175,-0.2040832639,0.2193612754,0.0727682114,0.0811632127,-0.2905584872,-0.2938543558,0.162700817,0.5291999578,0.4015758932,0.3271007538,-0.0001083251,-0.1900360137,0.3640153706,-0.4076463878,-0.5083063245,-0.170351848,-0.2225591242,0.2536762953,-0.1591915041,-0.0934045389,-0.6251901388,-0.1363317817,0.0133751472,-0.1100857332,0.2596839964,0.1732192189,0.2259884179,-0.0555906668,-0.0955966935,-0.1510584205,0.1919825822,0.2502139807,-0.1117708832,-0.0226814114,0.1821225435,0.0146341296,0.1411370486,0.2883913517,-0.1353493929,-0.1251489073,0.6657478213,-0.1592933089,-0.0340276808,-0.0476164259,-0.0653878674,0.2897382677,0.6062102914,-0.2307612896,-0.3986759782,0.1317663938,0.0489644669,-0.1367883086,-0.0511328392,-0.1670542508,-0.1804630458,-0.0097852154,-0.159321174,-0.2900246978,0.0168993883,-0.0991930068,0.1257368028,0.4029803574,-0.2149695754,-0.2486239821,0.1087753102,-0.1918498725,-0.2503000498,0.2007246464,0.2491452545,0.1213822961,-0.2323536426,-0.1240352243,0.2190023512,-0.2282164246,0.0134818926,0.0134721603,0.2555201948,0.0663685948,0.0736561865,0.0467751846,0.2122075111,0.3263137639,0.3418327272,-0.0758142918,0.1266981065,0.3235274851,0.1213152185,-0.0469559021,-0.4792706966,0.0506935194,-0.3832933605,0.12917687,0.0326539949,-0.2910527587,-0.175021708,-0.1362396181,0.2221731246,0.0894525126,0.1904240847,0.6082089543,-0.1698205918,0.01628289,-0.1385700107,0.1991364509,-0.2245851606,0.3818306923,-0.1250522435,0.1218158081,-0.2075890452,0.2060779929,0.155845657,-0.3326058686,0.135431245,-0.036553368,0.0925743058,-0.0350978263,-0.3106779158,0.2560380697,0.1394774467,0.011242196,0.2154384404,-0.024268534,-0.4070394933,-0.0185336247,-0.2668024898,0.0970803052,0.2298041582,-0.3090216815,-0.2402658314,0.2687863111,-0.2112889886,-0.2399874479,-0.2093415409,0.2070339173,-0.025995804,-0.3565532267,-0.0472566225,-0.1550932378,-0.1873577982,0.0806875154,0.3618192077,0.0493871719,0.1410747916,-0.2385227829,-0.0613222308,-0.0213319585,0.1526767313,-0.0223494153,0.1086406261,0.1549884528,0.0454300493,0.3059645891,0.0213385373,-0.7309918404,-0.295655489,-0.0253892522,-0.0704014972,0.0473956242,0.0758811012,0.0266980901,0.5626787543,-0.3666591644,0.0267688185,0.4169299006,0.1045972407,0.0271830522,-0.1137812883,-0.1186165512,0.1551050991,0.333193481,-0.1412375122,-0.0132466843,-0.0201382544,0.5371373892,0.1115940213,-0.2153895199,0.1351972073,-0.0304381102,0.1972648948,0.2505992651,0.0189217851,-0.1343054771,-0.1566025317,-0.0714810267,0.3544766307,0.364461422,-0.2444358766,-0.4812817872,-0.0186536182,-0.2085857242,-0.2348953933,-0.0008711324,0.1966120899,0.1794876307,0.4501173198,-0.2948080301,-0.103123419,-0.4538341463,-0.1539292485,-0.0776344165,-0.0457611158,0.0276464708,-0.0710129365,-0.0754553452,0.2740697563,0.4282013476,-0.0134114316,0.0006371762,0.0198208001,0.360601902,-0.1204632446,0.2716153562,-0.0330983698,0.4144019186,0.0200404413,-0.5199398994,0.1516585201,0.060767103,0.0611061342,-0.0336799286,-0.085207656,0.0801730677,-0.215658471,0.1141562238,0.2457428873,-0.2012441456,0.2822172642,0.0160779636,-0.1429870427,-0.1003037095,-0.0837649554,-0.0753232315,0.1822261214,0.0374899656,-0.295640856,0.0134174684,0.1434490085,-0.1508319676,0.2064888775,-0.00133868,-0.0963153392,-0.2068926096,0.1192235276,-0.1725337356,0.0588459037,0.2269429713,0.2886661291,-0.0056762486,-0.08356978,-0.1300175935,0.1359495372,-0.0962685645,-0.0186919682,0.3149695694,0.0194331501,-0.0111387698,-0.4310698211,0.0500695966,-0.0565520786,-0.0238469187,-0.0844350681,-0.4412799776,-0.0452111848,-0.6270555258,0.0019319929,-0.4387516379,0.1230187714,0.1873015165,0.254145354,0.0001312008,-0.3270376027,0.2696534395,0.3459562659,0.2631192207,-0.2504715025,-0.0477190204,-0.3070183098,-0.2446109354,-0.0061026989,0.1591077894,0.1529820859,0.0119618867,0.0585236326,0.0105655249,0.2202136815,-0.2467608601,-0.1310711056,0.2598434389,-0.0594292767,0.1896293908,0.0486859381,-0.0084419986,0.2503096461,-0.0831067935,0.0541906469,-0.3319964409,-0.1546895653,-0.0413330644,0.0950936154,-0.005971218,-0.3511022329,-0.3864653111,-0.1713501513,-0.4853638709,-0.2212005705,-0.1197456941,0.097088635,-0.0555566549,-0.3157558143,0.0252460707,0.1532766223,0.4391430616,-0.2017246932,-0.0928222016,0.2107275128,-0.4319680035,-0.1795963496,-0.162950471,-0.0613188371,0.2938212454,0.0343351439,-0.1930086911,-0.0772728845,0.1486816108,-0.1429511905,0.0608095899,-0.0430621617,0.3532962203,0.1241763309,-0.0949400961,-0.1835152358,-0.2840723395,0.0802196786,0.3533506095,0.1515348107,-0.0702311844,0.024417419,-0.0395405479,0.3105427623,0.0657296479,-0.030512413,0.2941798568,0.0216222163,0.1571723819,-0.0036712699,0.1609631628,0.411649555,0.0916723311,-0.2100939006,0.5984882712,0.2012268752,0.072218962,-0.1221123412,-0.0430777632,-0.496040225,-0.1229056716,-0.1275876015,-0.1947238594,0.0066919518,-0.0712367743,0.1688202173,0.1662697494,-0.0502689146,0.3053994477,0.1846429408,0.3476548195,0.0191542972,-0.1670770049,-0.2236253023,-0.2916164994,0.1689397246,0.2247604579,0.1301315278,-0.1703353524,-0.3254180551,-0.0393572859,-0.1953029782,0.0927153304,-0.3638089597,0.1425368935,0.0832579881,0.0921076983,-0.1773830056,0.0022863583,-0.1171138212,-0.0452600792,-0.0477973968,0.036805883,-0.4489265978,-0.081379734,0.2349191457,-0.0324411467,-0.1318274289,-0.1953958869,-0.0375424065,-0.6822195649,-0.0587642044,0.1167947054,0.0817241371,0.2942719758,-0.0501620136,-0.1117715314,0.2129606903,0.1041549146,0.1387806982,0.111742653,0.2444797307,0.2800198197,0.1165295243,0.2486374825,0.0691628158,0.4142351449,0.5681797862,-0.1052540019,-0.0624382831,0.002276131,-0.1300584376,0.3896211982,0.1206476688,0.20801346,0.1284282655,-0.010366112,-0.1830569953,-0.7336474061,0.3549474478,0.1393647194,0.0852793306,-0.1830883026,-0.6046622992,0.3944536746,0.1575481445,-0.1495505422,0.0865622014,-0.1425365955,-0.3577366173,0.2706686258,0.1563920677,1.0518200397,-0.2577362955,-0.0205295756,-0.0682568774,-0.110253267,0.2670953572,-0.2811877728,0.1263157278,-0.0363979377,-0.0499609932,0.0489073545,0.2216233313,-0.0825650766,0.1110292897,-0.3840634227,0.2460258901,0.1215831265,0.0690218657,0.0188957117,0.2989875674,0.0200606082,-0.3731887639,-0.2414264381,0.1576666534,-0.1448145658,0.2374411672,-0.1495661139,0.0253533982,0.1036651656,-0.1133126542,-0.0152338324,0.2427361459,0.0636692196,-0.315256089,-0.2060832232,-0.1548039913,0.0955947861,0.3037030697,-0.1515712291,0.2297077924,-0.1060803235,0.1211513802,0.0687353164,0.3163684309,0.0635415465,0.0734943226,0.3956197202,-0.1521241218,0.2492260188,0.0123690628,0.2981196344,-0.1289328635,-0.2205707431,-0.3212549388,0.0293264594,-0.0470050983,0.0864681676,0.2525437474,-0.0306523852,-0.1810584068,0.2006276846,-0.0628491789,-0.4676531255,0.2213156521,-0.0301249847,-0.0668001994,-0.0389805771,0.219836548,0.1138350219,-0.3127253056,-0.0027811402,0.1530316919,-0.0303973183,-0.3113151193,-0.4853436351,-0.2333031297,0.0084999902,0.0897154883,0.1098415032,-0.3046223223,-0.0016213813,0.6208372712,0.0262152795,-0.2126692683,-0.1955835819,0.2060410082,-0.2056098431,0.2131326795,0.1754969209,0.4463225901,-0.0294495672,0.232604593,-0.0460700616,0.1175038815,-0.4152622223,-0.1724226922,-0.4506853819,0.368912518,0.0289844964,0.0565157533,0.1596822739,-0.1116546765,0.2143555582,-0.0604889765,0.0054145162,-0.1985801607,-0.0093509657,0.0974455029,0.2321246862,0.0716449171,-0.2855545282,0.0776179433,-0.2219574749,-0.2022411525,0.3412209451,0.2695657313,-0.0611862279,0.3556402922,0.0330335535,-0.0432591885,-0.0384666547,0.1293978095,0.1208643913,0.3543022275,0.0388860516,0.1090509444,0.1320512891,-0.188611865,0.0620809607,0.1777036637,0.4266299605,-0.0461703129,0.1700850874,-0.001513562,-0.0158827733,0.2375985086,0.6916753054,0.0541597381,-0.0590881407,-0.2120251507,0.1609108001,0.3708564341,-0.1485460401,-0.0129187768,-0.5179472566,-0.2849593759,0.1003659144,0.2510245442,0.0115845101,0.0169342812,-0.2469286621,-0.0236691106,-0.0586136468,0.0771436617,0.0785495937,0.1035350859,-0.0080735246,-0.0205560513,0.1679334342,-0.1570155621,-0.0534772575,0.432836026,0.1236586571,0.5165599585,0.2849929929,0.4423142076,0.0261055026,-0.0538114607,0.0328545906,0.1794987619,-0.1046281978,0.2635351419,-0.1521062106,0.5993239284,-0.3265370727,-0.0833450407,-0.1794980913,0.1200798675,-0.2615887225,-0.2134371698,-0.0151997516,-0.255702287,0.1766146868,0.0908825845,-0.0540149249,0.1169527322,-0.0005367667,-0.0073793009,0.0711869821,-0.0142641487,0.0348352939,0.0919852331,0.1931048334,-0.2312734127,0.4226826727,0.2813998461,0.0357513018,0.3122564554,0.4274140894,0.2777602077,0.2222013474,-0.3424446285,-0.0684827045,0.0338017344,0.0183839221,-0.0938161761,0.4202004969,0.2514146864,0.3408825696,0.2904654145,0.2242545187,-0.244018361,0.1784814298,-0.0694838986,-0.0060644317,0.0527358539,0.4793894291,0.1439281702,-0.0471906401,-0.0995904282,0.1491744965,-0.790872395,-0.2610087693,0.11874488,-0.3048148155,0.1629015058,-0.3607211709,0.1167136729,0.0910066664,0.5737715364,0.200151965,0.2887257934,0.065725185,-0.2745932341,-0.3580307364,-0.0606538653,0.2796656489,0.03790842,-0.2496096939,0.1469863355,0.1392485946,0.0493827797,-0.4266400337,0.0186102763,-0.1071025953,0.2197745591,-0.1197674572,0.3733874261,-0.1624426842,0.1714730412,-0.0584266447,-0.1896253824,-0.1771630496,-0.225442946,0.0215702392,0.0829843581,0.352686584,0.14508982,-0.2686474919,0.3546158373,-0.2303533554,0.5316294432,-0.3417062759,0.0261851456,0.1350892931,0.0329936892,-0.3262867928,-0.3018455207,-0.2989817262,0.1844819337,0.1585558951,0.4981390834,-0.255497396,-0.2075786442,-0.050176546,0.3235316575,-0.1857936233,-0.0170856975,-0.4260116816,-0.2550640702,-0.0187295284,0.214742437,0.0881758705,-0.1216906682,-0.4048175514,-0.5538058877,0.324054569,0.2493555397,-0.2571322918,-0.4015716016,0.2706722021,0.1335133314,0.0151636209,-0.6682818532,0.2751118541,0.0016302431,0.1941598654,-0.0660212561,0.4154755175,-0.0731355324,-0.1756197661,-0.1833537817,-0.1522699147,0.1328650564,0.0113620646,-0.1448207051,-0.3962369561]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2653","title":"Add SD task for SUPERB","comments":"@lewtun @lhoestq: \r\n\r\nI have already generated the LibriMix corpus and prepared the corpus for diarization. The output is 3 dirs (train, dev, test), each one containing 6 files: reco2dur rttm segments spk2utt utt2spk wav.scp\r\n\r\nNext steps:\r\n- Upload these files to the superb-data repo\r\n- Transcribe the corresponding s3prl processing of these files into our superb loading script\r\n\r\nNote that processing of these files is a bit more intricate than usual datasets: https:\/\/github.com\/s3prl\/s3prl\/blob\/master\/s3prl\/downstream\/diarization\/dataset.py#L233\r\n\r\n","body":"Include the SD (Speaker Diarization) task as described in the [SUPERB paper](https:\/\/arxiv.org\/abs\/2105.01051) and `s3prl` [instructions](https:\/\/github.com\/s3prl\/s3prl\/tree\/master\/s3prl\/downstream#sd-speaker-diarization).\r\n\r\nSteps:\r\n- [x] Generate the LibriMix corpus\r\n- [x] Prepare the corpus for diarization\r\n- [x] Upload these files to the superb-data repo\r\n- [x] Transcribe the corresponding s3prl processing of these files into our superb loading script\r\n- [ ] README: tags + description sections\r\n\r\nRelated to #2619.\r\n\r\ncc: @lewtun \r\n","comment_length":73,"text":"Add SD task for SUPERB\nInclude the SD (Speaker Diarization) task as described in the [SUPERB paper](https:\/\/arxiv.org\/abs\/2105.01051) and `s3prl` [instructions](https:\/\/github.com\/s3prl\/s3prl\/tree\/master\/s3prl\/downstream#sd-speaker-diarization).\r\n\r\nSteps:\r\n- [x] Generate the LibriMix corpus\r\n- [x] Prepare the corpus for diarization\r\n- [x] Upload these files to the superb-data repo\r\n- [x] Transcribe the corresponding s3prl processing of these files into our superb loading script\r\n- [ ] README: tags + description sections\r\n\r\nRelated to #2619.\r\n\r\ncc: @lewtun \r\n\n@lewtun @lhoestq: \r\n\r\nI have already generated the LibriMix corpus and prepared the corpus for diarization. The output is 3 dirs (train, dev, test), each one containing 6 files: reco2dur rttm segments spk2utt utt2spk wav.scp\r\n\r\nNext steps:\r\n- Upload these files to the superb-data repo\r\n- Transcribe the corresponding s3prl processing of these files into our superb loading script\r\n\r\nNote that processing of these files is a bit more intricate than usual datasets: https:\/\/github.com\/s3prl\/s3prl\/blob\/master\/s3prl\/downstream\/diarization\/dataset.py#L233\r\n\r\n","embeddings":[-0.2306326032,-0.3324863613,-0.0547317527,0.0661656931,0.3352736831,-0.2812387645,0.0837412626,-0.061148867,-0.11445418,0.4088612795,-0.4140782356,0.4271566272,-0.1571262032,0.3467774987,0.157153517,0.0596581884,0.0408407263,0.2455734313,-0.3144540787,-0.0464579388,-0.0659857914,0.0155268218,-0.1118666157,-0.287522912,-0.1676342189,0.3125640154,0.0159357525,-0.0790378004,-0.3433384299,-0.2468428016,0.1256893575,0.4726737738,0.281218946,0.3263559341,-0.0001044987,-0.2434346527,0.1248142496,-0.4071696699,-0.3041563332,0.0205105282,-0.1400114894,0.2529032528,-0.2431382239,-0.1004232243,-0.621483326,-0.1066604331,-0.0715739205,0.0587374866,0.3838650882,0.2615119219,0.252238214,-0.1253829151,0.0187662579,-0.0416322239,0.3106273413,0.237228334,-0.1541283429,-0.0043616295,0.2208904028,0.0400725566,0.0306639131,0.2973972261,-0.2016076446,-0.0407217927,0.4242485166,-0.3073334396,0.0961292982,-0.1444908977,0.0631378889,0.4065577984,0.4522716999,-0.2049413025,-0.2887031436,0.2152059525,0.0077375132,-0.250554055,-0.2491929084,-0.0404339358,-0.2725820839,-0.0228523165,-0.2060943842,-0.1376397312,0.0324130356,-0.1914403737,0.019091947,0.3704536855,-0.2671468258,-0.3009182811,0.2260408401,-0.1345985085,-0.3198322654,0.3320156038,0.2018844038,0.1263996065,-0.2140351385,-0.184205547,0.2491409332,-0.3830596209,-0.0471485704,-0.0462135226,0.2658673823,0.1086573899,0.2545675039,0.0696091652,0.0248165876,0.1501286775,0.2364243269,-0.035419587,-0.0448732153,0.1341962665,-0.0506398194,-0.0012747318,-0.4389600158,0.0330511592,-0.3648111522,0.1746805459,-0.0583863072,-0.310341537,-0.2033723295,-0.1287636161,-0.0196185149,0.2213390023,0.0794401541,0.4776504636,-0.0948933885,0.0649780482,-0.0509123318,0.2270583659,-0.2490281165,0.3620907068,-0.1059040204,0.1659913957,-0.1457807273,0.0199768879,0.2315411419,-0.3207785785,0.1398188919,-0.049939312,0.1733028889,-0.1151324287,-0.2575062513,0.1903565079,0.2721084058,-0.0316013694,0.1445753127,-0.1708843857,-0.3705866635,0.0417678021,-0.2342026979,-0.0167721901,0.2933934927,-0.228061229,-0.2369891256,0.3036904633,-0.1506552696,-0.1359421462,0.0305428952,0.2015447617,0.0977666527,-0.3038559556,-0.0548577122,-0.2512845695,-0.2239756733,0.1178684458,0.3892634213,-0.070179753,0.1160391346,-0.2245630324,-0.0051192115,-0.0084878551,0.2964225709,-0.0331700109,0.2258887142,0.0678634122,0.032679975,0.1900208592,-0.0479926355,-0.5481700301,-0.2245559543,-0.0074401433,0.0226993319,0.0313451886,0.1355641186,0.0045564678,0.6181932688,-0.3744726479,-0.0381599963,0.3380509019,0.0725140944,0.0997640416,-0.1310929358,-0.0191139672,0.0118534891,0.2393464893,-0.0813841149,-0.0808948353,-0.0676738396,0.497025758,0.2334402651,-0.1725580692,0.1217373312,-0.0119452635,0.2894847691,0.0910597369,0.0217040274,-0.1855188459,0.0678440779,-0.1747249216,0.2588680089,0.383089751,-0.0389183201,-0.4150964022,-0.0010084159,-0.1662146896,-0.2090087384,-0.0032301026,0.2927136123,0.2428493798,0.4071397781,-0.3336694539,-0.0167746805,-0.6764424443,-0.0711311549,-0.0923162699,-0.0045005032,-0.0199995451,-0.0798712224,-0.1189093962,0.3851715624,0.5336657166,-0.0507785231,0.0314621627,0.0887996107,0.2988008857,-0.072608754,0.2658647895,-0.2084178776,0.4722294211,-0.0895455033,-0.5653336644,0.0470015146,0.1999008805,0.0958383605,-0.0654768944,0.0055325581,0.1372514814,-0.0953737646,0.0778418556,0.2655323148,-0.2062670887,0.2208839804,-0.0152462181,-0.1619107425,0.0088333944,0.0560737848,-0.2033192366,0.0120154358,0.0018841662,-0.2424480915,0.041762691,0.1776093543,-0.1071049571,0.1440399289,-0.0182286855,-0.0981846228,-0.2613983452,-0.0173643734,-0.0852323174,-0.102264747,0.2748094499,0.3855220675,-0.1190942526,-0.1484159082,-0.0794157386,0.193910867,-0.0809500739,-0.1137179211,0.4792473912,0.0618225522,-0.0789260715,-0.4903302193,0.0071991622,-0.0477589369,-0.0150727332,-0.0146992607,-0.3657634556,0.0245484039,-0.5591490269,-0.0659198239,-0.5107609034,0.1796441078,0.1567976028,0.3538485467,0.1014042571,-0.2838631868,0.3210274577,0.3669550717,0.2201403975,-0.0898769721,0.0792877004,-0.1918426752,-0.194029212,0.0848434642,0.2379512787,0.1225406379,0.0649443045,0.0667414218,-0.0530556776,0.0843598172,-0.1486164927,-0.1039231494,0.2008406073,-0.0775581524,0.087992534,0.0492757745,0.0500362925,0.1740358621,-0.1463154107,-0.0014404202,-0.2926126719,-0.1680888534,0.1636350304,0.0077726361,-0.0098427515,-0.3968880773,-0.3802548349,-0.0190264583,-0.4988487661,-0.1282486618,-0.0362591445,0.2246869355,0.0443339944,-0.1743770689,0.0240723472,0.1243400872,0.4566859305,-0.1572700292,-0.1098150015,0.1976679713,-0.3751398027,-0.2254721224,-0.2104405314,-0.0910398439,0.3619688451,0.0524688028,-0.0941483155,-0.0015293293,0.1633959115,-0.2450573891,-0.0099363634,0.0543501079,0.265822053,0.114933081,-0.116809085,-0.2195414603,0.0235201474,0.0505326279,0.2868507504,0.230368495,-0.1580718756,-0.0439049937,0.0079755094,0.3405783176,-0.0335665382,-0.139427945,0.191802904,-0.0780912489,0.0999262482,0.0838366523,0.1801462322,0.4542690217,0.0080785472,-0.1352082938,0.6013185978,0.1653676033,0.0473377816,-0.0611419864,0.2469249815,-0.3664108217,-0.0859006047,-0.0229577441,-0.1226144955,0.0805792212,-0.0697810277,0.1989886165,0.3178986609,-0.0515469722,0.2662784755,0.0324857384,0.3358448744,0.0018380027,-0.0721064433,-0.1589868218,-0.3156154156,0.0279980544,0.3300453424,0.0331320576,-0.098175317,-0.3514757454,-0.0617528111,-0.2059909254,0.0241250023,-0.2224295586,0.2713911235,0.2328320444,0.0883112475,-0.0381249785,-0.0083952555,-0.3488220572,-0.1965794265,0.0651907399,-0.1453956813,-0.4111145735,-0.0551003218,0.2846323252,0.0360940322,0.0026893185,-0.0581408925,-0.1157532111,-0.5529296994,-0.0287264045,0.1698338538,0.1482663751,0.2186100632,-0.1743785739,-0.127317965,0.1781241596,0.0050887237,0.1613935977,0.1702770591,0.2613531351,0.1952197999,0.1688318104,0.0751852915,-0.0710592493,0.4589055777,0.3865032494,-0.0444380157,0.0943175405,0.0667774007,-0.1656968147,0.3028517067,0.0504183024,0.3310027421,0.1102581322,-0.1322111487,-0.055098366,-0.6019122005,0.4275262654,0.1789315492,0.0574011318,-0.0100485552,-0.4498083889,0.36644274,-0.0553249605,-0.0761681646,0.0277816989,-0.1592036188,-0.2261774689,0.3278327286,0.22649014,1.0557332039,-0.1432823539,-0.0407687724,0.0989282429,-0.0821180791,0.0767414644,-0.3198874891,0.2381500602,-0.0709342062,-0.0617594235,-0.0001868081,0.2412300557,-0.1544336677,0.0569787696,-0.4224568009,0.0385999754,0.2103670239,-0.0116842128,-0.0343216658,-0.0049935468,0.0523136333,-0.2225150764,-0.1930847615,0.2036749721,-0.1351488829,0.0884169489,-0.0963512734,-0.0180269964,0.1538256407,-0.140937537,-0.0319010802,0.2230220586,-0.0113277193,-0.2855211794,-0.0643817559,-0.1961159855,-0.0370364338,0.2270756662,-0.3433634639,0.2473928183,-0.0330451801,0.0383865461,0.1383408159,0.2762387693,0.0389023982,0.1228723079,0.2784736156,-0.1656940132,0.2597303987,-0.0495828874,0.2597331703,0.0222773887,-0.0376325063,-0.2906152606,0.0964339375,0.0514145121,0.1492400467,0.2572363317,-0.1945392191,-0.168217808,0.2234095782,-0.0601494536,-0.6823365092,0.2686046064,0.0336509198,-0.0525873229,-0.0146580236,0.1720029116,0.1937297881,-0.2056248039,-0.03150497,0.1955662966,-0.0608237982,-0.3801860809,-0.4288159907,-0.234147802,-0.0988483354,0.029470535,0.097923167,-0.3611582816,0.1473832428,0.5644745827,-0.0358488075,-0.1570383757,-0.2788638175,0.2000600547,-0.1701855212,0.0973264128,0.0860838518,0.532431066,-0.1332002431,0.2280185223,-0.0996868014,0.2533307076,-0.4481813014,-0.1555628181,-0.4357386231,0.4334950745,-0.0423927046,-0.1146885604,0.0695859343,-0.0641074926,0.275678426,0.0756593645,-0.0946283787,-0.2677778602,-0.0259269103,0.0798231065,0.14190346,0.0841582343,-0.1523618698,0.3329120874,-0.3364926577,-0.2200277597,0.3265501261,0.2005949318,-0.0533113107,0.323287189,-0.0727536008,-0.0206120107,0.0121215256,-0.083904542,0.1334696114,0.1765251756,0.0147238979,0.109528169,0.1358218044,-0.208904475,0.1125435978,0.2479685098,0.3644374907,-0.2571533918,0.2628558874,0.0926766917,0.0390591286,0.2321558148,0.6654078364,0.0981388539,-0.0007727146,-0.2543886304,0.1583813131,0.4445456266,-0.2427861392,0.0042162864,-0.5556681156,-0.26296857,0.1984889805,0.2155687809,-0.063848041,-0.2041838169,-0.1155447513,0.0124320155,-0.1126069129,-0.0285255853,0.0075037112,0.2312056869,0.0835143179,-0.0903502032,-0.0491828211,-0.2032928616,0.0378117263,0.4039697647,0.0077075846,0.566360414,0.2026461661,0.2588069439,0.0613277331,0.0391851589,-0.0602189116,0.2987784743,-0.1448058188,0.3799729943,-0.123351045,0.4566288888,-0.2239579409,-0.1622548103,-0.2535138726,0.1247854158,-0.2302413136,-0.1228770018,0.0045669181,-0.2692739964,-0.0339429826,0.0132917585,0.0711433068,0.2759763002,-0.0217501801,-0.0007499849,0.1304038465,-0.0083941827,-0.1292393059,0.0283902418,0.3064989746,-0.2977680862,0.3046827018,0.3741045594,-0.0208565798,0.2079509795,0.3667666614,0.2912707329,0.1162853464,-0.2981622815,0.0292447153,-0.0771596208,-0.0221265648,-0.1009698883,0.3379071057,0.2625353038,0.3141921759,0.2241152376,0.2191426754,-0.2802014053,0.1896580905,-0.0158354528,-0.14960666,0.0556346104,0.304038018,0.0036888919,0.0052481079,-0.1746345609,0.0441916622,-0.674569726,-0.219366461,0.0320769548,-0.2078758776,0.1861616224,-0.1932148337,0.1298995167,0.0137166996,0.5280652046,0.1954897344,0.184864819,0.093573533,-0.2953617275,-0.2751648128,0.0423691086,0.3151341677,0.0227445774,-0.2715735137,0.1354252547,0.3042470813,0.1386160851,-0.5134697556,-0.0223846231,0.0471751951,0.3129724562,-0.2073919624,0.36067909,-0.148221001,0.132529676,-0.0403504409,-0.2700538039,-0.2355543971,-0.3604171276,0.1735894829,0.1764336377,0.2119522244,0.0678332895,-0.0882794186,0.0157065373,-0.1559094787,0.4464556575,-0.3855407536,-0.0427055955,0.1393243074,0.0659097359,-0.2002022117,-0.3122700155,-0.1709507257,0.3210387826,0.2248520106,0.3830829859,-0.1534941792,-0.1580290198,0.0711329207,0.3729661703,-0.2364521325,0.0506322719,-0.3769299388,-0.2457831502,0.0239396058,0.1878746152,0.1292333007,-0.1024035662,-0.4255013764,-0.5281770825,0.304894805,0.1131717041,-0.2440490723,-0.4501844049,0.322385639,0.1904883683,0.0026973912,-0.6447272301,0.2225141078,-0.0392062925,0.1735617071,-0.0894950628,0.3659565449,-0.0188618265,-0.2235411853,-0.1541429907,-0.16731821,0.2073861808,0.0264924932,-0.1550417691,-0.4224494994]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2651","title":"Setting log level higher than warning does not suppress progress bar","comments":"Hi,\r\n\r\nyou can suppress progress bars by patching logging as follows:\r\n```python\r\nimport datasets\r\nimport logging\r\ndatasets.logging.get_verbosity = lambda: logging.NOTSET\r\n# map call ...\r\n```","body":"## Describe the bug\r\nI would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well).\r\nAccording to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0.\r\n\r\nI also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\nfrom datasets.utils.logging import set_verbosity_error\r\n\r\nset_verbosity_error()\r\n\r\ndef dummy_map(batch):\r\n return batch\r\n\r\ncommon_voice_train = datasets.load_dataset(\"common_voice\", \"de\", split=\"train\")\r\ncommon_voice_test = datasets.load_dataset(\"common_voice\", \"de\", split=\"test\")\r\n\r\ncommon_voice_train.map(dummy_map)\r\n```\r\n\r\n## Expected results\r\n- The progress bar for `.map` call won't be shown\r\n\r\n## Actual results\r\n- The progress bar for `.map` is still shown \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.5\r\n- PyArrow version: 4.0.1\r\n","comment_length":25,"text":"Setting log level higher than warning does not suppress progress bar\n## Describe the bug\r\nI would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well).\r\nAccording to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0.\r\n\r\nI also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\nfrom datasets.utils.logging import set_verbosity_error\r\n\r\nset_verbosity_error()\r\n\r\ndef dummy_map(batch):\r\n return batch\r\n\r\ncommon_voice_train = datasets.load_dataset(\"common_voice\", \"de\", split=\"train\")\r\ncommon_voice_test = datasets.load_dataset(\"common_voice\", \"de\", split=\"test\")\r\n\r\ncommon_voice_train.map(dummy_map)\r\n```\r\n\r\n## Expected results\r\n- The progress bar for `.map` call won't be shown\r\n\r\n## Actual results\r\n- The progress bar for `.map` is still shown \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.5\r\n- PyArrow version: 4.0.1\r\n\nHi,\r\n\r\nyou can suppress progress bars by patching logging as follows:\r\n```python\r\nimport datasets\r\nimport logging\r\ndatasets.logging.get_verbosity = lambda: logging.NOTSET\r\n# map call ...\r\n```","embeddings":[-0.4419386983,-0.1637800336,0.0876798704,-0.1625066102,0.1394929886,-0.0227062833,0.4293670058,0.2288739532,0.0989456922,0.1445457786,0.1754810065,0.6159744263,-0.1407675892,0.1008445546,-0.1938808262,0.1940040886,0.0229355711,0.0618275292,0.097369045,-0.0051870239,-0.2738218009,-0.2124589086,-0.4537506104,0.1177136377,-0.1908455044,-0.0994138941,0.2819594145,-0.0676383972,-0.1347370744,-0.6845767498,0.3213669062,0.065580003,0.1590007544,0.6035616398,-0.0001319013,-0.062754631,0.5607278943,0.1790860295,-0.2247193009,-0.1718731225,-0.5930761695,-0.3090308607,0.3925269246,-0.0706966892,0.029787939,-0.3474459648,-0.1272029579,-0.2612386048,0.2345679253,0.1812126786,-0.0071892273,0.2482845783,-0.0372400172,0.5450522304,0.2635576129,0.1993847489,-0.2768193781,0.1551691741,0.4623755217,-0.2347632349,-0.3073786795,0.3043422997,-0.3391894996,0.0891906247,0.364815563,-0.1111718714,0.6513959169,-0.5470018983,0.3078687191,0.1294251531,0.3183060586,-0.2856887281,-0.388346523,-0.1922209859,-0.0462128073,-0.0714263916,0.1643112451,-0.0450842045,-0.3889573812,0.1488934159,-0.7792407274,-0.1206325367,-0.2446362525,0.2550941408,0.1909737736,0.2744412422,-0.0276818555,0.1187562421,0.1178881973,-0.1365887076,0.0607780181,-0.2545288503,0.0116920443,0.2483185679,-0.4572415054,-0.0615720265,0.3918232322,0.1325479448,0.1498936415,-0.0930807739,-0.1562084556,0.1936067343,0.1616427749,0.1653095484,0.2875440717,0.149799183,0.5941331983,0.2832121849,0.2572874129,0.1503210217,0.1399834305,0.2325322032,0.3986858726,-0.4250461757,0.4426813722,0.3641909361,0.3208539784,-0.2956034541,0.0545899868,0.0405408517,-0.4391988218,-0.0771040246,0.2471935302,-0.1380392611,0.4477908313,-0.3311774433,0.1545620263,0.2359691113,-0.1745847017,0.0984209254,0.0096889799,-0.2765092254,-0.2854017317,0.184328407,0.2677907348,-0.2136614025,0.2089000493,0.1451055557,-0.2515017092,0.0489353426,0.305675149,0.0159485061,0.5277969837,0.4136826694,-0.2733989954,0.2552737594,-0.011760084,0.0148656834,-0.1286605746,0.4761300683,-0.0766501874,-0.4540790319,-0.181950748,-0.0578345284,-0.4170138538,0.1642577648,-0.5429635048,0.3450361192,0.2866149247,-0.0346266739,0.3771919906,-0.1488109231,-0.5182569027,-0.098286815,0.1885202378,0.4915161431,-0.6797257662,-0.1832246631,-0.3486949801,-0.255328685,0.2047462612,0.0311415233,-0.235932216,0.0198393911,-0.0590228215,0.3680252433,-0.0158654414,-0.1317313015,-0.3091971874,0.3006400168,-0.051639691,0.1174315512,0.0648248047,-0.2404747754,-0.1172710061,-0.2352151722,0.0669461638,-0.0542757884,0.2040422261,-0.193476662,-0.2301561236,-0.2625632882,0.0543837547,0.0057217423,-0.1277088225,0.0963604599,0.1668698341,-0.4069348872,0.4588558376,0.0186784789,0.175115183,0.2314312011,0.3669462502,-0.4304539859,-0.0262334719,-0.0806005299,-0.1245752275,0.1955326349,-0.1073027998,-0.0875074416,-0.3066435456,-0.3172891438,0.0935739428,0.0031500207,-0.2190496922,-0.2050161064,-0.1374476701,-0.2400769889,-0.1346881092,-0.1813044697,-0.2845744193,-0.2089660466,-0.3723792732,0.0575660132,-0.1472546011,0.0575117543,0.2907574475,-0.3923129439,0.0755687431,0.1688372046,0.2368117571,0.1207592487,-0.125924021,0.3243703246,-0.1417277902,0.5567009449,-0.100467436,0.2094581425,0.121137239,-0.1209335476,0.026592046,0.1609217376,-0.2660254836,0.2535660267,0.1031406373,0.0171375033,0.4108292758,0.2966626287,-0.0813289285,-0.0227405485,0.0175928492,-0.1252802014,-0.2696110904,-0.407341212,0.2404932231,0.274861604,0.4156718552,-0.0455322601,0.1888778955,0.0303593129,0.5161691904,0.0442289338,0.1040256396,-0.0162763689,-0.206837967,-0.1898768693,0.3138594031,-0.0990591645,0.3751361072,0.1118451878,0.2199739665,0.149659425,-0.0212915689,-0.0491001941,0.1307388395,0.2823371887,0.0909593552,0.2894970477,0.0859388933,-0.1233443096,-0.1833081543,-0.1638492942,-0.1781530082,0.0441813581,-0.500053525,0.0947351381,-0.2914811969,-0.1945669055,-0.0300083589,0.0856101364,-0.2602653801,0.0320401378,0.1747450978,0.0906224549,-0.2227553278,0.4843930006,0.2578085363,-0.1074472219,0.1933289915,-0.1457282752,-0.4724210799,-0.0866331384,-0.1253925562,-0.1823430657,-0.0189420171,-0.2148829699,-0.1284613907,0.2228270024,-0.1062044948,-0.1972326785,-0.0630908608,-0.0496863946,-0.1086649671,0.2292987555,0.0186894909,0.068712309,-0.0976074189,0.1998907328,-0.0956420973,-0.3264266253,-0.0166796353,-0.1359514594,-0.2407793403,0.2194422483,-0.1885677576,0.066867575,-0.2478818446,-0.1415907145,-0.0791148022,-0.358297646,0.1285268664,0.2675815225,0.1696363389,0.3616370559,-0.1621623784,-0.2807433903,0.0995376557,-0.5332383513,0.2486508191,-0.0036411504,0.0805034637,-0.0706206337,0.0875727832,-0.0307491459,0.3047397435,-0.3325500786,-0.0666837692,-0.2412404567,0.0422770791,0.0657984614,0.0382698439,0.2339995056,0.0500805117,0.0235204082,-0.0735215843,-0.3944574893,0.0005847008,0.2307642847,0.3519811332,0.1808226705,0.2470562011,0.0638627782,0.4577203989,0.5331856012,-0.0457769856,0.0876507014,0.2052827477,0.2106728107,0.0915259421,-0.1375364959,-0.2229065597,0.0465073735,-0.1945648342,0.1077928916,0.0719749331,-0.2971440256,0.15068385,-0.509393394,0.0052218912,-0.3185647726,-0.227126196,-0.1176449582,0.360894829,-0.1305042058,0.2869966924,-0.1469755769,-0.2282096446,-0.1503717005,0.0929701254,0.3258207142,-0.0075670262,-0.0222309269,-0.1427607238,-0.7272332907,0.353939563,0.0946092084,0.1905517727,-0.232779935,0.0496623144,0.1037776098,0.362875402,0.2293982655,-0.5157282948,0.5419458747,-0.1000210866,0.068157047,-0.4487811625,-0.128435269,-0.0385266319,0.4695380032,0.2855923474,0.3334144652,0.030608749,-0.2551819086,-0.0336538032,0.1708623022,0.102869451,-0.2805161774,-0.477993995,0.112444669,-0.1199508756,0.1532053351,0.1270662397,-0.1490374655,-0.2905454636,-0.4047369659,0.4099102318,-0.2108185589,-0.0392247066,0.0343888663,0.0404037163,0.0293979459,-0.1926476061,0.3561835587,-0.1313863546,0.1562406123,0.3615219593,0.1200857684,-0.3876297176,-0.0415249541,-0.2409797758,0.2633204162,0.3591601253,-0.3182227314,0.2258417904,0.2511283755,0.2178796679,-0.0568475015,-0.3926560581,0.3767252564,0.0825198442,-0.2441381365,0.0366806574,0.3615198135,0.2846278846,0.0048829592,0.3582075834,-0.1439699978,-0.3557717502,0.1363275945,0.192844972,0.670355022,0.0464235656,-0.0555054024,-0.1147095412,-0.0216103829,0.2048896849,0.2880737185,0.3241170645,-0.0797359645,-0.0270820521,-0.0731248111,-0.2170137912,0.3660193682,0.4260587394,-0.0611474216,0.1053868383,-0.3908004761,0.1868422478,0.021676803,0.0957271159,0.0689627305,-0.1061701402,-0.5644488335,-0.103889145,0.138389498,0.1261152774,0.0942866504,0.0276049618,0.2706151307,-0.1854799837,-0.2928012908,0.1106323674,-0.0034728136,-0.0112966876,-0.0601467602,-0.4210722148,0.3573764861,0.1403993666,-0.004003,0.0417735651,-0.3574367166,-0.1172961444,0.1370169371,0.3045192659,-0.0825027972,-0.0577627905,0.3106999099,-0.0192489922,-0.3764193952,0.3355257809,-0.0487450026,-0.3803371489,-0.1026892588,0.0091499798,0.1832506359,-0.0248408169,0.3616975248,-0.0915992633,0.2164245993,-0.075452134,-0.0184809212,0.2174575776,-0.0591079593,0.4294327796,-0.4074733853,-0.5229222775,-0.092700839,0.439002037,-0.0634731352,0.1430631727,0.420136869,0.1718601882,-0.0951449648,-0.1441233903,-0.0777489543,0.5916374922,-0.2072825879,0.0725369453,0.2695091069,0.2778219879,0.1658399403,0.0501280464,-0.2087292522,0.2175135761,-0.4212330878,-0.2543781996,-0.3175611794,-0.0088604754,-0.0737412646,0.2421280444,-0.1627411991,0.2145867944,0.3262541294,-0.1238767281,-0.0687061325,-0.150385648,-0.0415361337,-0.0758055598,0.0196595155,0.4972249568,0.2590437233,-0.058503639,-0.2144831568,-0.0271567944,-0.245652318,0.0011471537,-0.0487663448,0.2012062818,0.0231991336,-0.1960074008,0.1144611686,-0.1625995785,-0.0466735177,-0.061079856,0.3420162201,0.4181254506,0.2039483935,0.4495308399,-0.1516053975,0.0352698304,-0.1840051264,-0.1711871177,-0.1577925682,0.0731631517,-0.0218010042,0.2425388396,0.2466847003,0.0664247051,0.1573636383,-0.1460386515,-0.0087375045,0.5891343355,-0.1070055142,0.0546476953,-0.0847922638,-0.1332147866,0.2813197374,0.1972290277,-0.2099963576,-0.3687160611,-0.1818748713,-0.0137198996,-0.1013230458,-0.2978684604,0.2876681983,0.1761937588,0.0931803882,0.0762057826,-0.3027901649,0.0076991282,0.2502682209,0.1959593594,-0.1139157936,-0.2009650171,0.2130936384,0.387919277,-0.1446015239,-0.0477547161,0.4281346798,0.2170432508,0.2008768618,0.2702795267,0.2117697746,0.4954335392,0.1691818088,0.1041399017,-0.211442098,-0.2212606072,-0.310947597,0.2509047389,-0.334071368,-0.1139445528,0.1798522323,0.2877365351,-0.1051025242,-0.2036352903,0.0761896297,-0.1378473192,-0.0804357454,0.0155550204,-0.1261262,0.0717876479,0.1098936126,-0.072645776,-0.0561332703,0.3540784419,0.3864502609,-0.0885678381,-0.0087136272,-0.0395477228,0.001240251,0.1779047847,0.3660145998,0.0556320176,-0.2327338159,-0.203435868,0.0603443161,0.1389932036,0.2818627656,0.3087726533,0.2892176211,0.1288416237,-0.0372468196,0.1457116455,-0.0828652531,-0.0870758221,0.4028379023,0.0156336427,-0.1129826605,0.1628928632,-0.1151732728,0.1128830016,0.2430272102,0.0932037905,0.3082069457,-0.3213830888,0.4500484169,0.1936810315,-0.2729335129,0.2564978302,-0.1527054906,-0.1216280013,-0.0642864332,0.6507233381,0.0050389403,-0.0175394379,-0.370709002,-0.0279638916,-0.2767977417,0.1902830601,0.4077703059,0.1590858549,-0.2476168871,0.1171312779,-0.4242975712,0.0680205375,0.1552351862,-0.0041450923,0.066938825,-0.0209292974,-0.3779889643,0.3652943075,0.459926486,-0.6130456328,-0.1873304397,-0.2289879322,-0.0798624456,-0.0266409703,0.0073652142,0.1018709019,-0.1415021569,-0.2839627862,0.2648752928,-0.2523229718,-0.1485255808,-0.1704113334,-0.3014567792,0.3983106017,0.1424654573,0.504865706,0.2264357507,0.0543133914,-0.0375089049,-0.0292079784,-0.1785160154,0.1687036008,-0.2268446386,0.2816552222,-0.0607071295,0.4257074893,-0.506567955,0.1821231395,-0.1935373545,0.0765630826,-0.257101804,-0.0011248698,0.0492265299,0.1285170913,0.1317146271,-0.1435742676,-0.2119310498,0.0161989722,0.0666766539,0.3037759066,-0.2902284861,-0.1640961617,0.394685626,-0.5808702111,-0.2545497119,-0.1875020415,0.4973037541,0.0882049277,-0.2843635976,-0.5184586644,-0.1538226604,0.2201806754,0.0405524597,-0.2124414891,0.3295530975,-0.2428317964,0.3919376731,-0.0787818283,0.1403305978,0.2935989797,0.1744740307,0.0523517579,-0.346852839]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2651","title":"Setting log level higher than warning does not suppress progress bar","comments":"Thank you, it worked :)","body":"## Describe the bug\r\nI would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well).\r\nAccording to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0.\r\n\r\nI also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\nfrom datasets.utils.logging import set_verbosity_error\r\n\r\nset_verbosity_error()\r\n\r\ndef dummy_map(batch):\r\n return batch\r\n\r\ncommon_voice_train = datasets.load_dataset(\"common_voice\", \"de\", split=\"train\")\r\ncommon_voice_test = datasets.load_dataset(\"common_voice\", \"de\", split=\"test\")\r\n\r\ncommon_voice_train.map(dummy_map)\r\n```\r\n\r\n## Expected results\r\n- The progress bar for `.map` call won't be shown\r\n\r\n## Actual results\r\n- The progress bar for `.map` is still shown \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.5\r\n- PyArrow version: 4.0.1\r\n","comment_length":5,"text":"Setting log level higher than warning does not suppress progress bar\n## Describe the bug\r\nI would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well).\r\nAccording to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0.\r\n\r\nI also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\nfrom datasets.utils.logging import set_verbosity_error\r\n\r\nset_verbosity_error()\r\n\r\ndef dummy_map(batch):\r\n return batch\r\n\r\ncommon_voice_train = datasets.load_dataset(\"common_voice\", \"de\", split=\"train\")\r\ncommon_voice_test = datasets.load_dataset(\"common_voice\", \"de\", split=\"test\")\r\n\r\ncommon_voice_train.map(dummy_map)\r\n```\r\n\r\n## Expected results\r\n- The progress bar for `.map` call won't be shown\r\n\r\n## Actual results\r\n- The progress bar for `.map` is still shown \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.5\r\n- PyArrow version: 4.0.1\r\n\nThank you, it worked :)","embeddings":[-0.4366222322,-0.1895044595,0.0743782967,-0.1453161091,0.1211536825,-0.016663976,0.4509515166,0.1940404773,0.0887316912,0.1508802325,0.1999914795,0.561342299,-0.1248402894,0.1077278703,-0.1963586807,0.1819085777,0.0539611056,0.0602353439,0.1040947512,-0.0412896089,-0.283100903,-0.1897912174,-0.4784622192,0.1258878112,-0.2004112601,-0.0713027194,0.2745776176,-0.0329547897,-0.1289592087,-0.6764261127,0.3057643473,0.0324908718,0.1470482349,0.6435681581,-0.0001345348,-0.091379419,0.5871583223,0.1834015548,-0.2203285247,-0.1844286323,-0.5890929699,-0.3078994751,0.4234398901,-0.0559209548,0.032989651,-0.3635881841,-0.1030680537,-0.2145153582,0.2370788455,0.1808652282,-0.0269955229,0.2440971136,-0.0404821709,0.5187543035,0.2733104229,0.2085676789,-0.2833001614,0.1979553849,0.4370087385,-0.2615445852,-0.2856270075,0.2866729796,-0.3134352267,0.0799416006,0.3231032193,-0.1139236987,0.6235026121,-0.5465793014,0.3444277346,0.115679808,0.3415282369,-0.2682231367,-0.3771683574,-0.1847592145,-0.0501209721,-0.0739107653,0.1773614883,-0.0129236635,-0.3812271059,0.1733194441,-0.7525091767,-0.114568904,-0.2885511518,0.2656427324,0.1739243418,0.284184128,-0.0193002746,0.1282871813,0.101485759,-0.0989665687,0.0671404749,-0.2542367578,0.025395073,0.2150278389,-0.456751436,-0.0580200665,0.3556349576,0.1373711228,0.1577067375,-0.0898790285,-0.1611046791,0.1899184883,0.1817793548,0.1458337903,0.3057151437,0.1561451703,0.5800405741,0.2934983075,0.2502754629,0.1172977164,0.1357527375,0.2152757049,0.3767789006,-0.4010928571,0.4244675934,0.3494880795,0.3014306128,-0.3070563972,0.0448281765,0.0363124609,-0.4337946177,-0.0822166726,0.2316437215,-0.0993990526,0.457093358,-0.3252272904,0.1537739336,0.2615848482,-0.1866567731,0.1096090749,0.0359857753,-0.2741207182,-0.3176110983,0.1769816428,0.263289094,-0.1888583302,0.2181212008,0.130034104,-0.2113429457,0.0460226499,0.3035015166,0.0041000759,0.535372436,0.4379402697,-0.2564878464,0.2543399632,-0.003224425,-0.0019865099,-0.1187907234,0.4955053329,-0.0472404882,-0.4734765291,-0.1706902683,-0.0787167624,-0.4231211543,0.1825542748,-0.5122413039,0.3531803489,0.2957515717,-0.0337536223,0.3916368484,-0.1298916638,-0.5204783082,-0.1054702923,0.160868749,0.4821623862,-0.7146718502,-0.1819467098,-0.3525137901,-0.2679431736,0.1923442185,0.0453168079,-0.2470272779,0.0181159936,-0.0643010885,0.3563184142,0.0087004201,-0.1241793111,-0.315777719,0.2828499079,-0.0569979213,0.1253511906,0.0801649392,-0.2366905063,-0.1426499933,-0.2320195585,0.0428843871,-0.0828139633,0.2093628645,-0.227692917,-0.2336182147,-0.2623175979,0.0865076482,0.0143217584,-0.0898736343,0.1308040172,0.1914979219,-0.4275005162,0.4622023702,0.0153928781,0.1525639594,0.2305032015,0.3655529916,-0.4179873466,-0.0053507634,-0.0861879662,-0.1011683121,0.183853507,-0.1167273819,-0.0882472545,-0.2727741003,-0.3062525094,0.0884497315,0.0263484791,-0.2280812562,-0.2118432075,-0.1654342413,-0.2109457254,-0.1372637749,-0.1789101213,-0.3033931255,-0.203352809,-0.3971097767,0.0537261367,-0.1485006362,0.0728967711,0.2654079199,-0.4141547978,0.0374399982,0.1416274607,0.2437844127,0.1319215149,-0.1192876324,0.3179824054,-0.1445930004,0.5306198597,-0.0908385366,0.2002786696,0.0963924155,-0.1593496799,0.0415081307,0.1784121394,-0.2656928897,0.2534093559,0.1261481196,0.0174092166,0.3971624672,0.3040692806,-0.0512763336,-0.0347845331,-0.0087360749,-0.1126908958,-0.2915476263,-0.3830394149,0.286636591,0.2308032066,0.3962568343,-0.0740749165,0.1192931533,0.0277722627,0.5042585135,0.0256615393,0.1083011776,-0.0035510941,-0.1909963638,-0.1720311642,0.3263608515,-0.1283919066,0.4036282599,0.1070919484,0.2295198441,0.1590770632,-0.0231768321,-0.0513837449,0.1276387572,0.2856346369,0.076255843,0.3203601241,0.1051395088,-0.1456760913,-0.1609080434,-0.1191010773,-0.1783548445,0.0295974314,-0.5095726252,0.1181132495,-0.2663790584,-0.1827956736,-0.0412399285,0.0598365702,-0.2675733268,0.0438957512,0.1903299093,0.1057748944,-0.2147472203,0.4609238505,0.2445789427,-0.0959711075,0.1808788627,-0.121426262,-0.4633196592,-0.044868622,-0.1306193024,-0.1998141259,-0.0267205965,-0.1969577074,-0.1358225793,0.2057523429,-0.1196157262,-0.1972937882,-0.0667202473,-0.02934983,-0.0904197097,0.2442874908,0.0217878744,0.0617012754,-0.1198640689,0.1975722313,-0.0573624186,-0.3322888017,-0.0023738358,-0.1274659187,-0.2256227583,0.2091623396,-0.16722399,0.0933271348,-0.2575122416,-0.1156263426,-0.074849464,-0.3515520394,0.1344306618,0.2727233768,0.1704082191,0.3361809552,-0.1727546453,-0.2849339247,0.1088538319,-0.5221973062,0.2548729181,-0.0329003632,0.0567956567,-0.0533744432,0.1012202948,-0.0807541385,0.3114565015,-0.3402623534,-0.0751883537,-0.2442617714,0.0135638295,0.0533219017,0.0363557935,0.247253418,0.0145042324,0.0490928888,-0.0857396722,-0.392418772,0.0097812088,0.2281423807,0.3786133528,0.1612499207,0.2444765419,0.0446085595,0.4915258884,0.5163927674,-0.0024622383,0.1002281904,0.2015589029,0.2330330312,0.0797830001,-0.1610596627,-0.2486424148,0.0500095487,-0.2048521191,0.1257278323,0.0694046095,-0.2729186416,0.1270224601,-0.5159067512,-0.0320396945,-0.2876974344,-0.2163576633,-0.0940592587,0.3546862006,-0.126742065,0.2936576009,-0.1175334156,-0.2145199925,-0.171711117,0.1462939531,0.3044540882,-0.020254679,-0.0263067577,-0.2009218186,-0.725900948,0.3370814919,0.0761053711,0.2099784911,-0.2474346459,0.0292637795,0.1185441241,0.3390406668,0.2738924623,-0.4972936213,0.5712028146,-0.1241491213,0.0329964794,-0.4254927337,-0.1400945038,-0.0470325649,0.4806449711,0.2772702277,0.3089494407,0.0401758552,-0.2596902251,-0.0157185495,0.1729172766,0.0971239582,-0.3175732791,-0.5171657205,0.1440287679,-0.0981356651,0.1447695792,0.1204323098,-0.1497597247,-0.3155136704,-0.3936460316,0.4108857214,-0.2200438678,-0.0275793839,0.0164919812,0.0314443596,-0.0021013541,-0.2182169855,0.3401775658,-0.0833299384,0.1575746089,0.3678264022,0.1298787445,-0.3685784638,-0.0039333678,-0.2523728907,0.26440835,0.4159534574,-0.3378778994,0.2192041278,0.2431717217,0.2255040109,-0.0358333364,-0.3956553638,0.3884732723,0.0800477862,-0.2540038824,0.0725561753,0.384901017,0.2306375206,0.0165652093,0.3709304631,-0.1172259226,-0.349442333,0.1379201412,0.1485601515,0.739689827,0.0700373501,-0.0203488581,-0.1114272028,-0.0323832408,0.1619278193,0.2905729413,0.3374275267,-0.0904777274,-0.0497136004,-0.0842127576,-0.2353036851,0.3700191081,0.436014533,-0.0459268615,0.1339512318,-0.4334485829,0.2030454576,0.0289395172,0.1065952033,0.0690995827,-0.0396557562,-0.5679023266,-0.1259214729,0.1884072721,0.137459442,0.0665205568,0.0161108468,0.2304401547,-0.1749318093,-0.2970417738,0.1078064069,0.0061557782,-0.0587684512,-0.0622148663,-0.4939040244,0.3695282936,0.1299766302,-0.0068509509,0.0269065797,-0.3696957231,-0.1247505024,0.104847908,0.2793019414,-0.093085818,-0.0561046377,0.2871667147,-0.0319346935,-0.3887970448,0.327473551,-0.053651467,-0.3810686767,-0.1186726689,0.0216105282,0.2097165287,-0.0428618006,0.3449860513,-0.0885884687,0.2159010321,-0.0754101723,-0.0332036503,0.2230555117,-0.0527375042,0.4099672735,-0.3975879252,-0.5194176435,-0.1001872197,0.426129669,-0.0681529343,0.1508343518,0.4230159223,0.184945181,-0.092615217,-0.1267095655,-0.0956690535,0.6199712157,-0.2234287262,0.0520470217,0.2567071021,0.2721745372,0.1533064395,0.0670570135,-0.1969036758,0.2230751216,-0.4461251497,-0.2744704187,-0.2919118404,0.0049771606,-0.0395389833,0.2355494946,-0.2009961307,0.2263100296,0.3504720628,-0.1619743556,-0.0465516932,-0.1710611135,-0.0167153683,-0.0795968771,0.00774054,0.5189781189,0.2658918202,-0.0415309928,-0.2269735783,-0.0420698859,-0.2279737741,0.0225804579,-0.0071748053,0.2283177525,0.0378097259,-0.1916879714,0.1173202768,-0.171503976,-0.0544584543,-0.040199969,0.3170288205,0.4192363918,0.2088205963,0.4509541094,-0.1462937295,0.0399448574,-0.1798142046,-0.1421520859,-0.1549686491,0.1066516638,-0.0259164385,0.2351192236,0.2743012905,0.0986631662,0.1587776095,-0.1500926614,-0.018521836,0.5697210431,-0.1254421473,0.0389681123,-0.0631025955,-0.1649565399,0.3101995289,0.2201516926,-0.225717932,-0.3638964593,-0.1483884305,-0.0364217795,-0.1135988906,-0.2758148313,0.2651437223,0.1681870371,0.0704304576,0.0956298262,-0.3088129461,-0.009747657,0.2722441852,0.2013941258,-0.1357320696,-0.1686853021,0.2026935816,0.383618623,-0.1374166012,-0.0205119886,0.4279457331,0.2303544283,0.1847298294,0.276330024,0.2287867814,0.4875734746,0.1613244712,0.0903142095,-0.2241203636,-0.1987064481,-0.302898109,0.2729904652,-0.3013044,-0.1466984153,0.1599042416,0.2988327146,-0.1062285304,-0.1810869575,0.0839490965,-0.1546312571,-0.0483906157,0.0266083684,-0.1023123637,0.0513085425,0.1330785006,-0.0770653859,-0.0599723868,0.355501771,0.3546187878,-0.0932412818,-0.0098798536,-0.0368990898,0.026849376,0.2082889527,0.3374266922,0.0508362055,-0.2282774448,-0.1850134134,0.0792355165,0.1585012078,0.287923187,0.3144430518,0.3152282536,0.1287280172,-0.0597806945,0.1287047565,-0.0977703407,-0.0650639459,0.4125493765,0.0639128238,-0.115642257,0.1631851494,-0.1343089491,0.1177459881,0.2320352644,0.1356163472,0.2673477232,-0.3518297672,0.4342441261,0.1628826559,-0.3047359884,0.2614014149,-0.1643307805,-0.1494656801,-0.0396311805,0.63878721,0.0062992363,-0.0247333013,-0.3449418545,-0.0422095545,-0.2876565754,0.1764049083,0.4158902168,0.2149531394,-0.2339255363,0.0803675726,-0.407050252,0.0977935344,0.1503385007,0.0114642559,0.0591475889,-0.0098431678,-0.3878220022,0.3545519412,0.5087055564,-0.5996223092,-0.1852422059,-0.1791339964,-0.065855436,-0.0448382013,0.0230244137,0.1399784386,-0.1736660749,-0.2680561543,0.2474185228,-0.2441124767,-0.1802080423,-0.1695580035,-0.2874850631,0.3598207235,0.1391485333,0.5063350201,0.2199589014,0.0921766311,-0.018493209,-0.0397837237,-0.1561372578,0.1596335322,-0.2182407677,0.2920566797,-0.0388845801,0.4051301479,-0.5195279717,0.1790847927,-0.175137803,0.0605585724,-0.2603450418,-0.0069156941,0.0544952638,0.1411406994,0.1439352781,-0.1631395072,-0.2179023474,0.0221333224,0.0596022271,0.2734459043,-0.3245607018,-0.1242209449,0.3923085332,-0.6037874818,-0.2861418426,-0.1776348948,0.5171576142,0.0850820392,-0.2996498942,-0.5478132367,-0.1730964333,0.2138219923,0.0722021982,-0.1775416732,0.3376224637,-0.269828856,0.386503309,-0.0750030354,0.1426694393,0.2976945341,0.1622108519,0.0419383906,-0.36179474]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2651","title":"Setting log level higher than warning does not suppress progress bar","comments":"See https:\/\/github.com\/huggingface\/datasets\/issues\/2528 for reference","body":"## Describe the bug\r\nI would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well).\r\nAccording to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0.\r\n\r\nI also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\nfrom datasets.utils.logging import set_verbosity_error\r\n\r\nset_verbosity_error()\r\n\r\ndef dummy_map(batch):\r\n return batch\r\n\r\ncommon_voice_train = datasets.load_dataset(\"common_voice\", \"de\", split=\"train\")\r\ncommon_voice_test = datasets.load_dataset(\"common_voice\", \"de\", split=\"test\")\r\n\r\ncommon_voice_train.map(dummy_map)\r\n```\r\n\r\n## Expected results\r\n- The progress bar for `.map` call won't be shown\r\n\r\n## Actual results\r\n- The progress bar for `.map` is still shown \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.5\r\n- PyArrow version: 4.0.1\r\n","comment_length":4,"text":"Setting log level higher than warning does not suppress progress bar\n## Describe the bug\r\nI would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well).\r\nAccording to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0.\r\n\r\nI also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\nfrom datasets.utils.logging import set_verbosity_error\r\n\r\nset_verbosity_error()\r\n\r\ndef dummy_map(batch):\r\n return batch\r\n\r\ncommon_voice_train = datasets.load_dataset(\"common_voice\", \"de\", split=\"train\")\r\ncommon_voice_test = datasets.load_dataset(\"common_voice\", \"de\", split=\"test\")\r\n\r\ncommon_voice_train.map(dummy_map)\r\n```\r\n\r\n## Expected results\r\n- The progress bar for `.map` call won't be shown\r\n\r\n## Actual results\r\n- The progress bar for `.map` is still shown \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.5\r\n- PyArrow version: 4.0.1\r\n\nSee https:\/\/github.com\/huggingface\/datasets\/issues\/2528 for reference","embeddings":[-0.4194480479,-0.2166283131,0.0877702013,-0.1140532419,0.1437677443,-0.0061434344,0.4659480453,0.1874903589,0.1150936633,0.1609269083,0.1710264683,0.5213401318,-0.1191446036,0.1564169377,-0.179561913,0.163658455,0.0363907404,0.0399048701,0.0673741773,-0.0241544601,-0.2623018622,-0.1510341316,-0.4679849148,0.1196368635,-0.2278078645,-0.08566726,0.2721702754,0.006182753,-0.1426071525,-0.6765163541,0.3204423189,0.06615071,0.127334848,0.6453863382,-0.0001329726,-0.0949901268,0.5629515052,0.1842794865,-0.2239541262,-0.1839147955,-0.5513772964,-0.3232943118,0.4023686349,-0.0338120013,0.0208013412,-0.3501756191,-0.0920740962,-0.23931624,0.2670522034,0.1653190702,-0.0192766171,0.2831249535,-0.0586092174,0.4998004138,0.2234582007,0.2457610816,-0.2917155623,0.2161928713,0.3965168595,-0.2511083782,-0.2980992198,0.333540082,-0.3044558465,0.0793330744,0.3228489161,-0.0942102596,0.5482657552,-0.5429745317,0.3560597301,0.1638410538,0.313989222,-0.275775224,-0.4008262455,-0.2374729216,-0.0689214841,-0.0841397867,0.2074192017,-0.0569322035,-0.3642187417,0.1841091514,-0.7559099197,-0.1116252095,-0.2625908554,0.2257198244,0.158071667,0.197280243,-0.0735766441,0.115197368,0.1341672689,-0.1101859957,-0.0222957116,-0.2534662485,0.025161013,0.2113279402,-0.4722692966,-0.0633324459,0.3794118464,0.1291335225,0.2034267783,-0.0867183432,-0.1818345338,0.1976144165,0.1256779283,0.1203658655,0.3252741396,0.181173414,0.5801015496,0.2417636067,0.2746009231,0.1342278272,0.1692243367,0.1933048666,0.3915659785,-0.4095592797,0.3914437592,0.3446343243,0.3438452482,-0.3137471378,0.0155794183,0.0509666614,-0.4092164338,-0.0681180805,0.2430140078,-0.1016196534,0.435893327,-0.3065005839,0.1776935905,0.2534344196,-0.1883926541,0.1004866213,0.0123936953,-0.2854112387,-0.301368475,0.1843718737,0.2685053349,-0.2216465324,0.2476360947,0.1354619861,-0.181137532,0.0428304076,0.266035378,-0.0139230303,0.5266014934,0.4214462638,-0.2664098442,0.2670305073,0.0083891284,0.0135655217,-0.1076314971,0.4208973348,-0.0334603302,-0.4324358404,-0.1698894948,-0.0774066001,-0.4194755852,0.1852019131,-0.5703686476,0.3945088685,0.2625325024,-0.0061020432,0.3849343657,-0.1200000644,-0.5595036745,-0.0686792657,0.1874358505,0.5323165059,-0.6689030528,-0.1983337104,-0.302010417,-0.2982584536,0.1592373848,0.0569975749,-0.2219586968,0.0172005482,-0.0877886638,0.3599467278,-0.0123618767,-0.1341806203,-0.3060700595,0.2788533866,-0.0574235804,0.1598239988,0.0783697739,-0.2394248098,-0.1313073039,-0.2272685021,0.0499346927,-0.0520346835,0.2066144645,-0.2099166065,-0.2178607285,-0.2718025744,0.0620777458,0.0047339932,-0.1030277312,0.1302037537,0.1705186516,-0.4230135083,0.4602574706,0.0050961762,0.1662645489,0.2522232234,0.3600953221,-0.3725225627,-0.0190367922,-0.0989761353,-0.1513772905,0.1950443089,-0.0961543098,-0.0752197653,-0.258448869,-0.335549742,0.0888508782,0.0037014983,-0.2363209128,-0.2194863856,-0.1652313918,-0.2052398026,-0.1206945032,-0.1470932961,-0.3188953698,-0.1347645521,-0.3923554122,0.0856432915,-0.1959695369,0.0706903189,0.2787785828,-0.3889972866,0.0497435816,0.1792114973,0.2545569241,0.0958552361,-0.1060143039,0.3385999203,-0.1613691449,0.5349557996,-0.076564312,0.199458614,0.1310655773,-0.1394268274,0.0066059418,0.1657056361,-0.2604729533,0.2606535256,0.1015329361,0.0371911973,0.3780524433,0.3061613739,-0.0579855554,-0.0211421959,-0.0115475999,-0.1071384326,-0.3096522093,-0.3777331412,0.2929332256,0.2169059664,0.3872652054,-0.0725138262,0.0711777136,0.0179527607,0.5371723771,0.0275604054,0.0961741656,0.0039968612,-0.1861436069,-0.1759094447,0.309553802,-0.1377416104,0.3949810565,0.1037656367,0.234185636,0.1701298058,-0.0120425513,-0.0249662343,0.1237814575,0.2725049257,0.0394235104,0.308265835,0.0963008031,-0.1269568354,-0.1867012531,-0.113029778,-0.1921275407,0.0026033306,-0.5119529366,0.1249052361,-0.2727039158,-0.2093120366,-0.0580459423,0.0679945722,-0.3036392629,0.0245304555,0.1856606603,0.1214190796,-0.2246751338,0.4604810774,0.2647572756,-0.0691636577,0.1800733209,-0.0982146785,-0.4667694569,-0.0165449735,-0.1305678189,-0.2023025304,-0.0306870975,-0.1962864101,-0.1247068867,0.1705867499,-0.1275937408,-0.2164132595,-0.091236636,0.0031766563,-0.1149679199,0.2591285408,0.0458533019,0.0846919045,-0.1029759198,0.2210692167,-0.0422953144,-0.339094013,-0.0236617029,-0.1512133628,-0.2036243379,0.2260996252,-0.1768358648,0.0907532871,-0.2339164615,-0.1362813115,-0.0044371118,-0.3908513188,0.1155016944,0.3416927755,0.1535814255,0.3538263142,-0.1998694539,-0.2381340861,0.0656993911,-0.5592921376,0.2181504667,-0.0225913767,0.0331592597,-0.0761664212,0.1021731645,-0.0918191448,0.2675778568,-0.357031852,-0.1622637659,-0.2736046612,0.0350182131,0.0648664013,0.0707481205,0.2284070104,-0.0030902766,0.0444658771,-0.0733879581,-0.4413124025,0.0344592817,0.2094848752,0.3510352075,0.1781585068,0.2788648009,0.0500062034,0.4938481152,0.5491995811,-0.0072918548,0.1220052093,0.1980454624,0.3001471758,0.046387285,-0.1811504811,-0.2339049727,0.0332602412,-0.1602295488,0.1277446747,0.0749926418,-0.1945340782,0.130738616,-0.5043548346,-0.012166149,-0.3147657514,-0.2307464778,-0.0638118908,0.3222165406,-0.110972479,0.3051475585,-0.1192508936,-0.1826013923,-0.1384406984,0.1604781747,0.324493736,-0.0177375302,-0.0665804893,-0.1565974355,-0.7497754097,0.3291145563,0.0784392431,0.2200621217,-0.281272918,0.030507531,0.1147278175,0.3535027802,0.3254816234,-0.4611082375,0.5472042561,-0.156603992,-0.007216536,-0.4187338054,-0.1097273454,-0.0313821509,0.4772336483,0.3027766049,0.3317657411,0.0044216407,-0.2987476885,0.0217907485,0.1919641495,0.1006112546,-0.2874781489,-0.5288088322,0.1305342466,-0.1117770746,0.1583827734,0.1379870772,-0.1541638523,-0.2864658535,-0.3668367565,0.4029609561,-0.2446129769,-0.0241480451,0.0089512356,0.0326549448,0.029357912,-0.1925097406,0.396454066,-0.1021244824,0.1322388649,0.4131764472,0.0995199606,-0.3907552361,-0.0037936848,-0.2269957364,0.263566047,0.4420370162,-0.3409429193,0.2153915167,0.2474151403,0.2321072519,-0.1077117324,-0.3518369496,0.4158522189,0.1311853528,-0.2641296089,0.0554407351,0.3793281913,0.2485082448,0.0219595209,0.3741231561,-0.0422271714,-0.3474364281,0.1061656103,0.1026434749,0.7153917551,0.0595244281,-0.0043461332,-0.103656128,-0.0146664577,0.180541724,0.2902141213,0.3213528693,-0.0837808922,-0.0792642608,-0.0755913705,-0.2176297158,0.3704991043,0.3794518411,-0.0272791013,0.1546705365,-0.3713064492,0.2048707306,-0.0110497558,0.1191859916,0.0478529148,-0.0680252165,-0.5839021206,-0.1201954409,0.1824703217,0.1717062443,0.0825518668,0.0070218286,0.2505727708,-0.1865665019,-0.3046936393,0.0991005525,-0.014890356,-0.074360311,-0.0033571341,-0.4704345465,0.3738690019,0.1327674389,0.0532013066,0.0329340398,-0.3798650205,-0.1416080892,0.0609789118,0.3000295758,-0.1093980744,-0.0563324988,0.2883348167,-0.0080739455,-0.3495657444,0.3067298234,-0.0411277749,-0.3610078096,-0.0967257395,0.0049131839,0.1584924906,-0.0767803788,0.3591016233,-0.0942308009,0.2631840408,-0.0867772624,-0.0315935016,0.2539364696,-0.0566937923,0.3660244942,-0.3707057536,-0.5183649659,-0.107042104,0.4345510304,-0.1286250204,0.1439403743,0.4312184453,0.2185925245,-0.1005426794,-0.1184666902,-0.0901513696,0.6416493654,-0.2571544647,0.0688189268,0.2737260461,0.2366262227,0.1115942746,0.110775426,-0.2227440029,0.1716037542,-0.4190158844,-0.2762865126,-0.3318865597,-0.0077505009,-0.0748814866,0.2499898374,-0.1560772061,0.2026515007,0.3454889059,-0.1640282273,-0.0586745813,-0.1611113101,-0.035816852,-0.1080293506,0.0661339834,0.4837933183,0.3274909258,-0.0393911861,-0.2212652862,-0.0230292305,-0.263265878,0.0215726402,-0.0174621698,0.2178884149,0.0027966788,-0.2193585783,0.1431203634,-0.1452874541,-0.0603825524,-0.0859676003,0.3619781137,0.4405718148,0.2226716578,0.4477665126,-0.2011737972,0.0478123873,-0.1557096094,-0.1546194553,-0.1523964703,0.0863650963,-0.0099747181,0.2225942016,0.2048904598,0.0967516229,0.1289904863,-0.1224995926,0.0318545178,0.5156710744,-0.1409992576,0.0301239584,-0.0870249048,-0.1059626713,0.3522708416,0.1842837036,-0.2437331229,-0.3428603709,-0.1587504894,-0.0292355679,-0.0993852392,-0.2497040629,0.2789404988,0.1552611589,0.0966156945,0.1073512137,-0.2759996355,0.0214605164,0.291249305,0.1997137368,-0.1269546598,-0.144477576,0.2368112803,0.3344915807,-0.1610995382,-0.0098808939,0.4814530015,0.2492060214,0.1974928826,0.2881422341,0.216498524,0.4845138788,0.1097924784,0.1067037806,-0.2308211774,-0.2364524305,-0.2510317862,0.2623923421,-0.322637707,-0.1677157283,0.1167309955,0.3301038742,-0.166814357,-0.2044312358,0.066721037,-0.1337678134,-0.068150878,-0.0081172306,-0.0525425635,0.0584037043,0.1287471354,-0.0413550027,-0.0567552894,0.3195120692,0.3737288713,-0.1247482747,-0.0308314171,-0.0825385824,-0.0177627765,0.2014260143,0.3514150083,0.0771127865,-0.2032427937,-0.1703777611,0.0721133351,0.1613259017,0.3464232683,0.3160504401,0.2732931077,0.1529667675,-0.0767317563,0.1059675515,-0.0837308019,-0.0649701059,0.4270730019,-0.0024564299,-0.1129548922,0.1872640252,-0.1234095916,0.1056629643,0.1956489235,0.1061671302,0.2706530988,-0.394570142,0.4324447215,0.1137485728,-0.2513903081,0.2198579907,-0.1841418445,-0.1515815109,-0.0484459884,0.631777823,0.008543577,-0.013667169,-0.3724759817,-0.0293097422,-0.2737779021,0.1569490135,0.4406691492,0.1815847009,-0.2336033434,0.1212950349,-0.4157107472,0.0784173161,0.1878456473,0.0082267113,0.054933995,-0.013743802,-0.3882050514,0.3504621685,0.4890019298,-0.6110494137,-0.1095660254,-0.2364737839,-0.0730164871,-0.0528270155,0.013632115,0.1175284088,-0.1437648237,-0.2704066336,0.2419891655,-0.246841222,-0.1820782274,-0.1693627238,-0.2399415225,0.2725305259,0.0915756375,0.4885508716,0.2464660704,0.0629322007,-0.0193991344,-0.0721851736,-0.1774038374,0.1586482674,-0.2272848487,0.2812425196,-0.0522916391,0.4309611619,-0.5238478184,0.1463978738,-0.1905011088,0.0815974325,-0.2122621387,-0.0034652699,0.0155705903,0.1659046859,0.1260396093,-0.1721071899,-0.1848791838,0.0498261414,0.0452929251,0.3013288081,-0.3117759526,-0.1462138593,0.414262712,-0.6244522333,-0.2566308379,-0.158139959,0.5108258128,0.1120488048,-0.3501449823,-0.5960427523,-0.1468214244,0.1894258857,0.0729436502,-0.2026336044,0.3665900528,-0.3010776937,0.3585597873,-0.0689898282,0.1680401266,0.3208993077,0.1663098782,0.1175304204,-0.339356631]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2651","title":"Setting log level higher than warning does not suppress progress bar","comments":"Note also that you can disable the progress bar with\r\n\r\n```python\r\nfrom datasets.utils import disable_progress_bar\r\ndisable_progress_bar()\r\n```\r\n\r\nSee https:\/\/github.com\/huggingface\/datasets\/blob\/8814b393984c1c2e1800ba370de2a9f7c8644908\/src\/datasets\/utils\/tqdm_utils.py#L84","body":"## Describe the bug\r\nI would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well).\r\nAccording to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0.\r\n\r\nI also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\nfrom datasets.utils.logging import set_verbosity_error\r\n\r\nset_verbosity_error()\r\n\r\ndef dummy_map(batch):\r\n return batch\r\n\r\ncommon_voice_train = datasets.load_dataset(\"common_voice\", \"de\", split=\"train\")\r\ncommon_voice_test = datasets.load_dataset(\"common_voice\", \"de\", split=\"test\")\r\n\r\ncommon_voice_train.map(dummy_map)\r\n```\r\n\r\n## Expected results\r\n- The progress bar for `.map` call won't be shown\r\n\r\n## Actual results\r\n- The progress bar for `.map` is still shown \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.5\r\n- PyArrow version: 4.0.1\r\n","comment_length":19,"text":"Setting log level higher than warning does not suppress progress bar\n## Describe the bug\r\nI would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well).\r\nAccording to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0.\r\n\r\nI also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\nfrom datasets.utils.logging import set_verbosity_error\r\n\r\nset_verbosity_error()\r\n\r\ndef dummy_map(batch):\r\n return batch\r\n\r\ncommon_voice_train = datasets.load_dataset(\"common_voice\", \"de\", split=\"train\")\r\ncommon_voice_test = datasets.load_dataset(\"common_voice\", \"de\", split=\"test\")\r\n\r\ncommon_voice_train.map(dummy_map)\r\n```\r\n\r\n## Expected results\r\n- The progress bar for `.map` call won't be shown\r\n\r\n## Actual results\r\n- The progress bar for `.map` is still shown \r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.5\r\n- PyArrow version: 4.0.1\r\n\nNote also that you can disable the progress bar with\r\n\r\n```python\r\nfrom datasets.utils import disable_progress_bar\r\ndisable_progress_bar()\r\n```\r\n\r\nSee https:\/\/github.com\/huggingface\/datasets\/blob\/8814b393984c1c2e1800ba370de2a9f7c8644908\/src\/datasets\/utils\/tqdm_utils.py#L84","embeddings":[-0.4388272762,-0.2037361413,0.0796426833,-0.152261883,0.1692188829,-0.0159658007,0.4777553976,0.2021972537,0.0318350866,0.1475353092,0.1682003886,0.5794948339,-0.152897656,0.10003113,-0.1835140139,0.1944968998,0.0263119973,0.0424377173,0.0457620323,0.0004414686,-0.2606731355,-0.1466900259,-0.4761810601,0.0941639617,-0.2263159901,-0.0694701001,0.2926193774,0.0035402456,-0.1176254302,-0.6455771327,0.3015622199,0.0900377035,0.1431134343,0.6660018563,-0.000133249,-0.0878026485,0.5361055136,0.1657677442,-0.1852545738,-0.1453343928,-0.5602222681,-0.3629795909,0.390527755,-0.0617991835,0.0420391709,-0.3005341291,-0.1574251056,-0.2664179504,0.2540456951,0.1718673557,-0.0169129483,0.3063340485,-0.1073400155,0.5482039452,0.2376393527,0.1451119483,-0.3086778224,0.1676180363,0.4750162661,-0.2120764256,-0.2969590425,0.3610367775,-0.3199104667,0.1028575674,0.3066386878,-0.0769179389,0.6109765172,-0.562685132,0.3343413472,0.1769725382,0.3472414911,-0.3141160011,-0.3868072629,-0.1731614619,-0.0694993213,-0.092866905,0.205878377,-0.0921727419,-0.3756278753,0.1914187521,-0.798966229,-0.107239224,-0.2327204645,0.2482954711,0.1075972766,0.2289103717,-0.0718901008,0.0963486731,0.0785074681,-0.1218960881,0.1031514108,-0.2171088457,0.004403878,0.2491951883,-0.5141398311,-0.1063526124,0.3615424633,0.0719984099,0.1568298489,-0.0595684089,-0.1863827854,0.2191510051,0.1158890873,0.1340003759,0.2554060221,0.1776839197,0.5470439196,0.2377239168,0.2636432648,0.1237090528,0.1506489515,0.2026610225,0.3975674808,-0.435788244,0.4497306943,0.3733786643,0.3258314729,-0.278327316,0.007087599,0.0931878909,-0.3374129832,-0.0380666889,0.2372532934,-0.0866661444,0.4455689192,-0.2540481091,0.1459576041,0.267873168,-0.1896029413,0.0222190097,0.0186259449,-0.2887332141,-0.2559722364,0.1881934255,0.2404662967,-0.1552522779,0.2149009109,0.1360751688,-0.1472955197,0.0388564803,0.3051345646,-0.0191650707,0.5638685226,0.4335379303,-0.2885813117,0.2676416039,0.0248658936,0.0252257604,-0.0963666365,0.4703077972,-0.0585900992,-0.4574696422,-0.1870668828,-0.0671591312,-0.4046756625,0.1836805195,-0.5151902437,0.3845124543,0.2827964723,-0.0090500852,0.402459681,-0.1470163465,-0.5378506184,-0.088026017,0.2301702052,0.5276051164,-0.6987097859,-0.1768608838,-0.3421043754,-0.2491247952,0.1895651221,0.0326977149,-0.209911719,0.0540279038,-0.0955106393,0.3148594201,0.072308585,-0.1670954973,-0.3650942743,0.3568647206,-0.0652225092,0.1153442338,0.0775962546,-0.2185321301,-0.1156786755,-0.2229297906,0.0675712153,-0.0615469627,0.1715471298,-0.2080003768,-0.2356216311,-0.2607070506,0.0544957444,-0.0067107948,-0.0851498842,0.1419218928,0.1761008352,-0.435642153,0.4553629458,0.073969692,0.1693066806,0.2582277656,0.333901763,-0.4463591576,-0.0426594056,-0.1058060229,-0.1639454812,0.2082848698,-0.1032622904,-0.0945881158,-0.2895023525,-0.3256093562,0.070658274,0.0064348616,-0.2393781543,-0.2069596052,-0.1521600932,-0.2338041216,-0.1510626227,-0.2036368847,-0.3345766664,-0.1761711389,-0.3320506811,0.0417802036,-0.1015787125,0.0562134869,0.3076057732,-0.3617397547,0.0521020964,0.1781619638,0.2282764465,0.1035378501,-0.1218878329,0.3652700186,-0.1437883079,0.5218519568,-0.0476370454,0.1986086071,0.1101089269,-0.1610442549,0.0058754878,0.1750309169,-0.2611900866,0.2790229619,0.1503447443,0.0661970675,0.3930872083,0.282735765,-0.0947121605,-0.0167267304,0.0013401011,-0.1263046265,-0.2820746303,-0.4369442165,0.2476718724,0.2622611523,0.3680646718,-0.1068669483,0.1074983478,0.0735359117,0.4914260507,0.0519005768,0.1097518578,-0.0605801828,-0.2102842778,-0.1543627232,0.2606515288,-0.0956025198,0.4311898649,0.093353264,0.2250624746,0.1879181713,0.0072549391,-0.0469730012,0.1144525483,0.284550786,0.0379744954,0.3296860456,0.0763445422,-0.1105656847,-0.156406194,-0.1601752043,-0.1460017413,0.0240580291,-0.5067870021,0.1301323175,-0.3039036393,-0.1741355062,-0.029542502,0.0431750864,-0.2786378264,-0.0110479686,0.1811807007,0.0700712129,-0.2074576765,0.4660838246,0.2094022036,-0.087380223,0.1957890391,-0.1263174862,-0.4703644216,-0.0578536466,-0.1195598319,-0.191275835,-0.0444651283,-0.1912553757,-0.1227570251,0.1609010994,-0.1007062122,-0.2174821645,-0.0513029061,-0.042095691,-0.0791766122,0.3041088283,0.0238878708,0.065711543,-0.1246171296,0.2760334909,-0.0570036396,-0.3113236725,-0.0033401474,-0.1241060272,-0.2416000217,0.1921311766,-0.1650097072,0.0801849514,-0.2275784761,-0.1445658803,-0.1452714205,-0.4155721962,0.1291881055,0.27291888,0.1285600066,0.3616063893,-0.17458345,-0.2520793378,0.0768212676,-0.5869483948,0.2583813071,-0.0286887027,0.0308609623,-0.0457013771,0.0982180908,-0.0320495032,0.2712278366,-0.3657181561,-0.1137825698,-0.2194886208,0.0420212746,0.0528603457,0.0495054238,0.2127379775,0.017098818,0.0331126489,-0.0813796222,-0.3952858746,-0.054773204,0.2042000294,0.3584367335,0.1497317106,0.2788911164,0.0666010082,0.5233482122,0.5273795724,0.0182266161,0.0799957961,0.2011780143,0.2599047422,0.0884442925,-0.1507943869,-0.254309833,0.0642150715,-0.1609962881,0.1307298541,0.0551926494,-0.154472962,0.1447830349,-0.4346088767,-0.0049787955,-0.3219949901,-0.2329060137,-0.1715308279,0.3573040366,-0.1110612825,0.2786035538,-0.1246926859,-0.1973200291,-0.1406462789,0.19760786,0.2490927577,-0.0153043447,-0.0503521934,-0.1122866869,-0.7124314308,0.3519894779,0.0758896545,0.2859663367,-0.2444900721,0.0119809797,0.1227614135,0.3503416181,0.2800548673,-0.4966639578,0.5247847438,-0.0929265693,0.0194527004,-0.4260508418,-0.1460878402,-0.0423696265,0.4549838603,0.3069401681,0.3457627296,0.0729792416,-0.2809364796,-0.0387250856,0.1553349793,0.0693569928,-0.3266054988,-0.5476042628,0.1364569664,-0.098186098,0.1574928612,0.1247021034,-0.1416090131,-0.2728050351,-0.376873076,0.3726088107,-0.2567854822,-0.0151186306,0.0430185646,0.033469405,0.0162063874,-0.1553938687,0.327055335,-0.1485353112,0.0930717066,0.3858939707,0.1286588162,-0.3814192414,0.009561196,-0.2251478583,0.2644339204,0.4158936143,-0.3462156653,0.2357854545,0.275482446,0.2510716617,-0.019245984,-0.3622413576,0.3614005446,0.0845031589,-0.2436749488,0.0730552375,0.3501393795,0.2230266184,0.0315576792,0.3476161659,-0.1459113657,-0.298895061,0.0685529932,0.140094012,0.7327116132,0.0617937297,-0.0144041358,-0.0860870704,0.0180072356,0.1429748982,0.2686703205,0.3113036752,-0.0786021128,-0.0717868209,-0.1081068888,-0.2314878702,0.3762955964,0.4070205688,-0.0584800802,0.1555220038,-0.3735044599,0.1973440945,0.0259927381,0.0688594803,0.0570675656,-0.0909051672,-0.5662988424,-0.1167453304,0.2005320638,0.1018138379,0.0523024388,0.032077428,0.2647800446,-0.1404699832,-0.2725673318,0.0908145458,-0.0659794882,-0.0348456092,0.0290495958,-0.478143841,0.3578082621,0.1115507558,0.0508481003,0.0253510792,-0.354246676,-0.1177039593,0.166096434,0.3358995616,-0.0619470254,-0.0807975307,0.320784986,-0.0067279204,-0.3636156023,0.3760668933,-0.0415591411,-0.3307175338,-0.0405644886,0.0347797275,0.205036357,-0.0230027102,0.3666189313,-0.0951414481,0.267087549,-0.0923197791,-0.0294135138,0.2634598017,-0.0588514842,0.4001792371,-0.4039235413,-0.5432715416,-0.088204734,0.3998919725,-0.0773521215,0.1587757766,0.398955524,0.1926053315,-0.090589039,-0.1487978101,-0.0242899936,0.572453618,-0.2445891798,0.0855640396,0.2558706403,0.2815245092,0.1719108075,0.0982372016,-0.2300800234,0.1694042683,-0.4567411244,-0.2837529778,-0.3306879401,-0.0320520177,-0.1125072837,0.2195500731,-0.1810237914,0.2398946136,0.3140030503,-0.1277663708,-0.0633230954,-0.1704508066,-0.0580626428,-0.0997020453,0.0153058609,0.4879112542,0.3068168163,-0.0222678464,-0.2422080487,-0.0411101989,-0.2885691822,0.0003656921,-0.0472630858,0.2228946835,0.0553663038,-0.2178845257,0.1097008958,-0.126944229,-0.0509038158,-0.0607409552,0.3688483834,0.4017429352,0.2120190412,0.4640012681,-0.1659378707,0.0108392769,-0.2196334451,-0.1615648568,-0.1266148686,0.1038639843,-0.0171624403,0.2366272956,0.2375610471,0.0850827843,0.1205140129,-0.1252826899,0.0252885129,0.5037805438,-0.1073520854,0.0316263214,-0.0593668483,-0.137345612,0.2653338015,0.2075216174,-0.2695911825,-0.4000188112,-0.1547537446,-0.0291503426,-0.1095416322,-0.2723490298,0.2048151046,0.1551049799,0.0983933806,0.0957409739,-0.2827530205,0.0102355918,0.279899925,0.2201735079,-0.1466089785,-0.2405674905,0.1962162852,0.329564929,-0.1332354993,0.031626232,0.4858592153,0.2517426312,0.1496288925,0.2734567821,0.2190313935,0.4766215086,0.1542240381,0.0910504162,-0.2495795339,-0.2548619807,-0.2787213027,0.2103088349,-0.3467278183,-0.155412674,0.1416518241,0.2543242574,-0.1171524525,-0.177302748,0.083325319,-0.1662633717,-0.0445052832,-0.0322798863,-0.0387063064,0.0699214861,0.1147415042,-0.0731822625,-0.0436077677,0.3693082035,0.3610862195,-0.079122819,-0.0125409262,-0.0775739029,-0.0477960557,0.2393307388,0.292709887,0.0355951674,-0.2087331712,-0.1845273674,0.0805376321,0.1834364831,0.3069001734,0.3027544618,0.2926495969,0.1202861592,-0.0902814195,0.1253693998,-0.0883065388,-0.0626058877,0.41165483,-0.0082992176,-0.1507256329,0.1903624684,-0.1247329861,0.1203603148,0.1995243728,0.1233226508,0.3067184985,-0.387367487,0.4715417027,0.1450249255,-0.2299478501,0.224338606,-0.1097464934,-0.1656803191,-0.061749842,0.6827004552,-0.0302160792,0.0076491456,-0.3466838598,-0.034598846,-0.2770442665,0.1914263815,0.4415136576,0.147093147,-0.26286906,0.1024152339,-0.4719314873,0.0841681585,0.2158723325,-0.0150023978,0.0823911279,0.0288239159,-0.3979296684,0.3991671503,0.5172958374,-0.6161106825,-0.1561103612,-0.2329222411,-0.0829843208,-0.0420483537,-0.0173367113,0.1415510178,-0.1652452946,-0.3095965683,0.2503563166,-0.2092673779,-0.1735659093,-0.135693565,-0.2915956974,0.3595241308,0.1325262785,0.4981567264,0.2361743301,0.0376023352,-0.0176861584,-0.0324681029,-0.1756130755,0.1570251137,-0.2105830908,0.2739689946,-0.0459046625,0.370100677,-0.5149692893,0.1676447093,-0.155926764,0.065442197,-0.228965193,0.0069134124,0.0399497561,0.1737319678,0.1414284557,-0.0992364213,-0.1870295405,0.0374041423,0.0606700927,0.3610664606,-0.311211586,-0.1599254012,0.3900365233,-0.5684536695,-0.2624579668,-0.2146308273,0.5193513632,0.064094089,-0.2825242579,-0.5841557384,-0.1547001451,0.2183987349,0.0663641244,-0.2001501322,0.3623512387,-0.2834407985,0.394987911,-0.0630562082,0.153617084,0.2650113404,0.1830284297,0.021702107,-0.3092701137]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2648","title":"Add web_split dataset for Paraphase and Rephrase benchmark","comments":"#take","body":"## Describe:\r\nFor getting simple sentences from complex sentence there are dataset and task like wiki_split that is available in hugging face datasets. This web_split is a very similar dataset. There some research paper which states that by combining these two datasets we if we train the model it will yield better results on both tests data.\r\n\r\nThis dataset is made from web NLG data.\r\n\r\nAll the dataset related details are provided in the below repository\r\n\r\nGithub link: https:\/\/github.com\/shashiongithub\/Split-and-Rephrase\r\n\r\n","comment_length":1,"text":"Add web_split dataset for Paraphase and Rephrase benchmark\n## Describe:\r\nFor getting simple sentences from complex sentence there are dataset and task like wiki_split that is available in hugging face datasets. This web_split is a very similar dataset. There some research paper which states that by combining these two datasets we if we train the model it will yield better results on both tests data.\r\n\r\nThis dataset is made from web NLG data.\r\n\r\nAll the dataset related details are provided in the below repository\r\n\r\nGithub link: https:\/\/github.com\/shashiongithub\/Split-and-Rephrase\r\n\r\n\n#take","embeddings":[-0.0288645159,-0.1778773814,-0.1873570085,-0.0683549643,-0.1398036629,-0.1547354758,0.3074898422,0.4715948105,-0.1738998145,-0.0117895892,0.0453259833,0.1378726661,-0.0555660985,0.1940241456,0.1701168716,-0.6055010557,-0.0123983268,0.074451454,0.2643257082,0.1444887072,0.1512328833,0.0593265444,-0.0403773151,-0.0020358979,0.157785356,-0.2311220765,-0.0768359378,0.1669075191,0.0414684266,-0.1626915187,-0.1479950398,0.0482937805,0.0278629288,0.332097739,-0.0001016112,-0.0322811715,-0.0505995676,-0.2499746084,-0.1838691533,0.0242072549,0.0995488092,0.0070744916,0.0803128183,-0.2609416842,-0.4809730947,0.0576129742,-0.1581419557,-0.1086849645,0.7326014638,0.3654152155,0.2578417659,0.3325785398,-0.2626753747,0.0687723532,0.0968215093,-0.1221051589,-0.2051911354,-0.2347332537,0.2340823859,0.0629764497,-0.3238230348,0.3008442521,-0.0755989403,0.0241029691,-0.3062632978,-0.0810972452,0.2578151524,-0.1594085097,0.0124270581,0.4539518654,0.0650076792,-0.0049667917,-0.2543866932,-0.2598276138,-0.0935717672,0.0525875241,-0.3415593803,0.0962758958,-0.132599771,0.2999708056,-0.1140079126,-0.0786191523,-0.1031865478,-0.2102257311,0.0743656158,0.2541102469,0.0881659761,-0.1330318153,0.2695815563,-0.0582153238,-0.0935087278,-0.0538209304,-0.1066537872,-0.0329912938,-0.2511689365,-0.2395282835,0.0351136439,-0.2438716143,0.1491824985,0.0371046588,0.3019566238,0.181977734,0.3964828551,-0.025000887,-0.0477328263,0.376462847,0.2004141212,0.1573153287,0.0453593992,0.0831126869,-0.0662739873,0.1469591707,0.0972364247,0.0546463244,0.0605073236,-0.0782834738,-0.4010339081,-0.0759979561,-0.3309721351,0.077511169,-0.1026607007,-0.0758104026,-0.1371463686,0.317525208,0.0926714689,0.1095987409,-0.0086928839,0.2597160339,-0.0580199435,-0.2749015093,-0.1794258058,-0.0557687804,0.0618840232,0.1956197768,0.1241040975,0.1833944917,0.1779671907,-0.0789642185,-0.0430732332,-0.2905367911,0.0696678981,-0.1243372932,0.1168813482,-0.2031008303,0.1822605878,-0.22881262,-0.1099798232,0.0091430331,-0.1907832623,0.043426197,-0.0514515638,-0.1618723422,0.0996923149,0.2647004724,-0.286067456,-0.2004657388,-0.0349415578,0.6340510249,0.1152209789,0.0862558186,0.034409605,0.2213574499,-0.2979502082,0.0278708525,0.1528502703,-0.0018394464,-0.1418185383,-0.1924065202,0.2306130379,0.1393484771,0.0075402856,0.4039290547,-0.1078338325,0.3824608326,0.2307473719,0.268301487,0.2636847198,-0.4192498326,-0.2203278244,0.3674188554,-0.0908930004,0.2000876367,0.1353859603,-0.2716650665,0.259272337,0.0762775242,-0.2066711187,0.3074487746,0.167615965,0.1279236972,-0.2651004195,-0.2885113358,0.2652745843,0.2443956286,0.0962305591,-0.3033280671,-0.2748201787,0.0347330049,0.3110316992,-0.3344899118,0.0566331595,0.0805408806,-0.1223563924,-0.3176546097,0.1923390925,-0.1843535155,0.0477645621,0.0213413555,0.0610602908,0.3190344274,-0.0704525858,-0.1634536088,-0.1535840929,-0.2407300919,0.0354612768,-0.294726193,0.289334774,-0.1496075541,0.1090677306,-0.0121676028,0.1619061381,0.0069165071,-0.1217783242,-0.1882754862,0.1360917389,0.2316889018,-0.1504922658,-0.1146272197,0.135905683,0.0571518131,-0.1382952482,0.0359300524,-0.06093302,0.1827898175,0.0002817136,0.4939135313,0.2210204154,-0.1895731539,-0.0013128637,-0.5283855796,0.0383531079,0.3542421162,-0.0071478216,-0.0471391864,-0.2811876833,0.2770487964,0.2664055228,0.13990964,0.3506095409,-0.3057734072,0.1371359527,-0.3607771993,-0.101715818,0.0006264256,0.3120720983,-0.1505287588,0.0750688538,0.0346065052,-0.317244947,0.0821414068,0.3548220098,-0.1794295013,0.2598607242,0.1075433046,-0.0233319066,-0.3381915092,-0.1486577392,-0.1655456126,-0.0093622645,0.3759807646,0.076067768,-0.0149339717,-0.0383888558,-0.0058002593,0.2012723982,0.1213991493,0.4316959083,0.1705682576,0.25036183,-0.1054296792,-0.2583144903,-0.0815766379,-0.0435520709,-0.0130828423,-0.18481341,0.0343692005,0.0855949447,-0.0866212696,-0.0450315028,0.0447679833,-0.1337621957,0.0039187772,0.324846983,-0.2313073874,-0.1931544095,-0.0453141443,-0.1339714676,0.5653936267,0.0915300846,-0.1782948971,-0.0446423516,-0.460169822,-0.1720284522,0.2701002359,-0.118408002,0.1765006781,-0.0300717019,-0.3637126088,-0.0787173584,0.209602043,-0.2971865535,0.2753501236,-0.1758042425,0.4350529909,0.0052438919,-0.0065441034,-0.1625712812,-0.1203380674,0.0676960647,-0.4252096117,-0.2726360559,-0.1748469025,0.1529573649,0.0559234582,-0.0879024714,-0.5407210588,-0.1277888715,-0.2784256935,0.2440552115,0.1151243821,0.0416758917,-0.13632375,-0.0630194023,0.0436252542,-0.04847892,0.259552002,-0.222953245,-0.3371585011,-0.1142389402,-0.1009931564,-0.384496659,-0.5984056592,-0.0945867524,0.1917591542,-0.0960440785,-0.2096124589,0.0904046968,0.0953291431,0.2000770569,-0.1439581364,-0.396882236,0.212722078,0.0476872399,-0.1810710728,-0.2729285061,-0.0221275929,0.4839444458,0.0773049891,0.2002920359,-0.2692375481,0.1418515593,0.3634506762,0.7001590133,-0.0068807765,-0.1521413028,0.0765575022,-0.2236717194,-0.2702566385,-0.0054342854,-0.0646548122,0.0898555517,-0.1115946323,0.0488973856,0.5577968955,-0.0455697849,-0.3598865271,-0.1997414231,-0.4661801755,-0.2200716138,-0.2364166528,0.2185426354,0.1340625435,0.2547062933,-0.0431562588,0.1331527531,-0.1190163791,-0.0731991082,0.050722003,-0.142326057,0.1256814748,-0.0965836123,-0.0992510393,0.3893972039,-0.4315908253,0.2374466509,0.4231916666,0.1090402082,-0.0082071926,-0.3092043102,-0.0747319385,-0.1300977021,0.247283265,-0.2077325583,0.081879057,-0.2114648223,-0.1098071188,-0.6485144496,0.1982103139,-0.2844995558,-0.1247802675,0.4377308786,0.3685232997,-0.3356840014,0.0158970468,0.1908910871,0.0203118566,-0.135594219,-0.327394098,-0.1462171525,-0.103047505,-0.1509972513,0.3998709619,0.0547210164,0.0646210089,-0.0989707634,-0.2360146195,0.1333415061,-0.173157528,0.0774802417,0.3515531421,0.1677058339,-0.4024573565,0.1037460491,-0.0052378271,-0.0192078166,-0.0664623082,0.1068519577,-0.1637498885,0.2946846485,-0.1094563305,-0.1812939197,0.3922466934,0.2378596812,0.0673123524,0.2950683534,0.1392222345,0.0682529137,-0.2878331244,-0.0494544394,0.2481822819,0.2025791109,-0.1293972731,-0.0167936608,0.2789422572,0.185532257,-0.1612981856,-0.3533717096,-0.2528837621,-0.3826754093,0.5741517544,0.1725623608,0.9510231614,0.040142417,0.1192300245,-0.0685308203,-0.0095045008,0.1740821451,-0.2570624053,0.1134187281,0.0312912017,-0.2134619802,-0.0028258869,0.1385058165,0.0647217184,0.0127429552,-0.3681533933,0.1987541616,0.0916063339,-0.015467451,0.0585930422,0.2449419349,0.0279828012,-0.1511097848,-0.424300164,0.1656162441,0.1358497888,-0.272859782,0.0953239277,-0.2028434873,-0.2224569619,-0.0600441061,-0.101081796,0.207677424,0.0054452321,-0.0056309593,-0.207915619,-0.3145811856,-0.0385752432,-0.121409066,-0.1192197353,0.0748970956,-0.173940137,0.306799382,-0.1545059532,0.1264387071,0.1419201195,-0.1011576504,0.3497188389,-0.2723554373,-0.588506043,-0.1119110808,0.0552846007,-0.1897664517,-0.1574540883,-0.3328258395,-0.0600225553,-0.1678334028,0.02717047,0.1936513484,-0.2732265294,-0.3572577536,0.2620496452,-0.0516159646,-0.4133223891,0.2437613159,0.2495730221,-0.3428574204,-0.2143488675,0.2475894094,0.3020260334,0.6073715091,0.3428602219,0.1523388475,-0.0105126323,-0.4468329847,-0.2356868684,0.2599701285,-0.2576028407,0.1929806918,-0.267313838,-0.2022397667,0.1040418074,0.2625377178,0.1717937142,0.1426449418,-0.1449778974,-0.035528779,-0.1088090464,0.0937881544,-0.1835062802,0.0528060086,0.1925470382,0.1307764351,0.0495259352,0.4649589956,-0.4707789421,-0.2025227547,-0.2250685096,0.2701049447,0.0133786276,-0.2236696631,0.209790349,0.1228627935,0.2570864558,0.39416942,-0.0175655261,-0.3999777734,-0.1368736178,0.0712369233,-0.0494793653,0.1875357926,0.0325002186,0.2323516458,0.0314443968,-0.1378849447,0.0275773033,-0.2846364677,0.2346347719,0.2725094855,-0.0108135594,-0.0919205099,-0.0462215692,-0.1067359671,-0.2390510738,0.0807376802,-0.0844341666,0.3460864127,0.2110967487,-0.0543783903,0.2725384533,-0.0298739281,0.4509206116,0.1376706511,0.0862305462,0.106643796,-0.2038817853,-0.2540319264,0.4175493121,0.2055110782,-0.0973763689,-0.1110120714,-0.1177562773,0.3727229536,-0.3402031362,0.0056756753,0.2115814537,-0.304625988,0.2724184096,0.3161087036,0.3333856463,0.1434262991,0.0988074616,-0.0227966476,0.2932492197,-0.0683169886,-0.0540904254,0.1922499686,0.066083774,-0.0783101544,0.2287473232,-0.1971527189,0.4330301583,-0.0871824026,-0.0015565783,0.4997726977,0.1681853384,0.6007428765,0.3325848579,0.4729866683,0.233653903,0.1346346289,-0.0257221218,0.247320354,0.1585436165,0.5062537193,0.2912385762,-0.0927703157,-0.2964932621,0.3814239204,-0.0867841691,-0.4107515514,0.0540961772,-0.2003592849,-0.1865158379,-0.0157227311,0.1464958191,-0.0033718215,0.132966429,0.0752055421,0.1946931928,-0.2048309743,-0.1958883107,0.0493690148,0.3956137598,-0.3078658879,0.2450905442,0.2459358275,-0.1614159048,0.0299472641,0.4961176217,0.2635521591,-0.2251489609,-0.2059305608,0.015206622,0.0356724039,-0.0286545064,-0.1295115799,0.2021036744,0.1344641745,-0.0404758751,0.4872911274,0.281457305,-0.2656034231,0.0839111134,0.1244581565,-0.1256856769,-0.3106482029,0.0880922824,0.2281372398,0.003419243,-0.2377467453,-0.054123003,-0.2702907324,0.2418325841,-0.0411070958,-0.42054075,0.0995845795,-0.1404414475,0.1517601311,-0.08107844,0.6691503525,0.0923091546,0.3051502109,-0.0049434132,-0.0363647006,-0.1301757395,0.0823583528,-0.0756552592,-0.0053885779,0.0429079868,0.2436386943,0.0063887052,0.2874322832,0.2141180187,0.1240224689,-0.2050896734,-0.0458542742,-0.4530299604,0.3129820526,0.0288706925,0.0462282933,0.0829254985,-0.1205852628,-0.0059739747,0.0792059302,0.1457206309,0.0022077421,-0.2352814078,0.3078123927,-0.1351796091,0.2736637592,0.1232652962,0.3387795687,-0.316750735,0.0777238235,0.1224116012,0.0003998667,-0.2735252976,0.0903666168,0.3385608196,0.0584261343,-0.0271695182,-0.1829447001,0.1358652562,0.040142972,0.2021953911,-0.063946642,-0.1833394915,-0.2195876092,0.0539337359,-0.2347049862,-0.0475318581,0.289600879,-0.0653934255,-0.1519202143,0.0057238545,-0.3085969388,0.1166794151,-0.1338546723,-0.0929705575,-0.0879138857,0.0845828354,0.3508962691,-0.0446807556,-0.7940334678,-0.1043574363,0.2865952253,0.1914792806,0.1763786972,0.3546121418,0.2080398947,0.0607460067,-0.0587291457,-0.2226141989,0.1205729619,-0.0617412291,-0.170849219,-0.4341712594]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2646","title":"downloading of yahoo_answers_topics dataset failed","comments":"Hi ! I just tested and it worked fine today for me.\r\n\r\nI think this is because the dataset is stored on Google Drive which has a quota limit for the number of downloads per day, see this similar issue https:\/\/github.com\/huggingface\/datasets\/issues\/996 \r\n\r\nFeel free to try again today, now that the quota was reset","body":"## Describe the bug\r\nI get an error datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files when I try to download the yahoo_answers_topics dataset\r\n\r\n## Steps to reproduce the bug\r\n self.dataset = load_dataset(\r\n 'yahoo_answers_topics', cache_dir=self.config['yahoo_cache_dir'], split='train[:90%]')\r\n# Sample code to reproduce the bug\r\n self.dataset = load_dataset(\r\n 'yahoo_answers_topics', cache_dir=self.config['yahoo_cache_dir'], split='train[:90%]')\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files\r\n\r\n","comment_length":53,"text":"downloading of yahoo_answers_topics dataset failed\n## Describe the bug\r\nI get an error datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files when I try to download the yahoo_answers_topics dataset\r\n\r\n## Steps to reproduce the bug\r\n self.dataset = load_dataset(\r\n 'yahoo_answers_topics', cache_dir=self.config['yahoo_cache_dir'], split='train[:90%]')\r\n# Sample code to reproduce the bug\r\n self.dataset = load_dataset(\r\n 'yahoo_answers_topics', cache_dir=self.config['yahoo_cache_dir'], split='train[:90%]')\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files\r\n\r\n\nHi ! I just tested and it worked fine today for me.\r\n\r\nI think this is because the dataset is stored on Google Drive which has a quota limit for the number of downloads per day, see this similar issue https:\/\/github.com\/huggingface\/datasets\/issues\/996 \r\n\r\nFeel free to try again today, now that the quota was reset","embeddings":[-0.4173349738,0.143848151,-0.0522662066,0.2103923559,0.2381456643,-0.0566483326,0.2377109826,0.2905799747,0.1442447603,0.0573149025,-0.0823222697,-0.0347232893,0.0095932391,0.2967808545,-0.1141854674,0.1532201618,0.1207118854,-0.2423342615,-0.3290193081,0.187282294,0.0082971463,0.3475176692,-0.1365118027,-0.186028257,-0.2225944102,0.0802132934,0.0324274227,0.1877681762,-0.3060182929,-0.1939620823,0.499863416,0.1681091785,0.048498448,0.2420327514,-0.0001159498,-0.0456452072,0.3564919531,-0.023051843,-0.2754577696,-0.171936661,-0.6136645675,-0.3641073406,-0.162503168,-0.0510163009,0.2578029037,0.4485144913,-0.0789878741,-0.0727403313,0.2911607027,0.3613387048,0.2246932238,0.5078260899,0.5233133435,0.0212832503,0.6007751226,0.0284277834,-0.1523155123,0.3020832241,0.0485347733,0.0955684856,0.0357895978,0.2261752188,-0.0547714457,0.2072292268,0.0509779118,-0.131506905,0.0101163024,-0.2643214762,0.18099913,0.3761467338,0.7534170747,-0.0676243082,-0.3866149187,-0.2335674614,-0.1377534717,-0.0953435823,0.3624245822,0.1967544854,-0.1790942848,0.1182354912,-0.3431764841,-0.0559018403,-0.1109952107,0.229899779,0.1896436065,-0.0589184761,-0.2049210817,-0.0243713334,-0.0747148991,-0.1185131967,0.3572820723,-0.277400285,-0.0273528006,0.1389751434,-0.5924146175,-0.1630174518,-0.147222802,0.3785592318,0.5116468072,0.5190519094,0.0123613179,0.1149433553,-0.27763924,-0.1024332941,0.3908933997,0.4025636017,-0.0215438753,0.1509063542,0.4378218949,0.3599639237,0.0442593172,0.0989443734,0.0606571287,-0.1273132861,0.3956009448,-0.0489524379,0.2884015441,-0.4871147275,-0.3083412647,0.2666872144,-0.2881489694,-0.2556790113,-0.034938097,0.1287734658,-0.2728326321,0.3193659782,-0.1581632942,0.231635198,0.0761210695,-0.1068308949,-0.2661254108,-0.0457298346,-0.1917149127,0.1251797229,0.409027487,-0.3089469969,0.3847400844,0.0583134443,-0.0366222151,-0.1010764167,0.0702504516,-0.0097570438,-0.3626892865,0.4796497226,-0.0283761043,0.3199997842,0.0633891225,0.205698505,-0.0027630047,0.188662529,-0.3618530929,-0.3521518409,0.0451591946,0.1900095344,-0.3650646806,-0.1490500867,-0.2103958875,-0.0985899344,0.0886896625,-0.0989632681,-0.0011395429,-0.0808662251,-0.1798315793,-0.1877604127,0.1709344387,0.2947294712,-0.1214032397,0.0492805317,-0.2126547396,-0.2295565754,0.0075604594,0.0988980085,-0.102120772,0.0516035929,-0.3121572137,0.1910166144,-0.0304746944,-0.044129435,-0.8213795424,0.112007983,0.2127331048,0.1995853484,0.0226960126,0.1158608794,-0.0115395924,-0.0558772534,0.0683128536,0.2879424691,-0.111356616,0.0531982332,-0.2135489434,-0.2146818191,-0.1412281245,0.1924072504,0.2178845257,0.179527387,0.1498346031,-0.0014782615,0.2000293434,0.0601047017,0.02646528,0.4225367606,0.2325090021,0.1264427751,-0.179184556,-0.4266201854,-0.6495069861,0.2739134729,-0.1510443538,-0.1101723239,-0.0169516113,-0.1082540005,-0.4812152982,-0.1396791339,-0.0671041086,-0.120860599,0.0543915816,0.1184010804,0.1921287328,0.0473960415,-0.0808084756,0.6759991646,-0.2277866602,0.1251005828,-0.2809175849,0.3650693893,-0.1465261579,-0.1149823815,0.1866227388,-0.0570824929,0.1539196968,0.0262796674,-0.1638719887,0.3405667543,0.0817586109,-0.0481769145,0.1589739025,0.3192985952,0.1357056648,-0.2003445774,-0.1071018875,0.2686326802,0.0888122767,-0.0193650126,-0.3108480275,0.1076376289,-0.3952946365,0.0862755626,-0.0847249702,0.2748892605,0.2487687469,-0.1529650986,-0.2354714423,0.0756012946,0.4902819693,0.1661131233,-0.015917832,0.0276128631,-0.2370831817,0.0720305815,0.3816649318,-0.1854065806,-0.0236775093,0.231529519,-0.0424432643,-0.084859401,0.0524934307,0.3492453992,0.4121186733,0.0774499029,0.1270791888,0.0192581601,-0.0698235929,-0.2119279206,-0.0097266156,0.1050680801,0.0516628176,0.3873591721,0.0891756043,0.0201651081,-0.3156518042,0.017886471,0.0662738904,0.2733680606,-0.4959610701,-0.0993697047,-0.2257489115,-0.0083330069,-0.1796372533,-0.2272870094,-0.4338906407,-0.4144547582,-0.0414733104,0.4321182966,-0.0073697041,0.1584973484,-0.1760055721,0.0766011328,0.1298984587,-0.1298920363,-0.0110590551,0.1239375621,-0.030557489,0.0674769878,0.4870757163,-0.1200205684,0.2683666348,-0.3797082305,0.0057567521,-0.5876286626,-0.0252418835,0.031945128,-0.0429755375,0.329652667,0.0681997612,0.3912466168,-0.3351227045,0.089258641,0.0412826017,0.0300798323,-0.2807072699,0.172067076,0.0588248335,0.0429034531,0.0793969035,-0.1639138311,-0.1592887789,-0.2847833335,0.0246164221,-0.1046027541,0.0429372713,0.3333446681,-0.0596335828,-0.0055995085,0.0306974165,0.0233642235,-0.3865577877,-0.6483180523,0.5215483904,0.0017794424,-0.4173825085,0.0635498688,0.1082302183,-0.1324106157,0.3234136701,-0.5212422609,-0.0796859786,-0.3460487425,0.2091495544,0.1091664284,0.1893132478,0.4114702046,-0.2102251202,-0.0587379932,-0.0052901902,-0.1677402854,0.0951371267,-0.0565814525,0.2765411735,0.0760917589,0.0933395401,-0.0337052345,0.6843774319,0.2569456398,0.3306367099,0.4992987216,0.1553970277,0.3406203985,0.0140705667,-0.1394452304,0.0300782286,-0.2988120019,0.1110088229,0.0800207779,0.1796587855,0.210477069,-0.2474083453,-0.1085623056,-0.2964667678,-0.1631582081,0.0617194623,-0.1888989806,0.3915340304,0.22359474,0.0087210583,0.1550948769,-0.3679841459,0.1491588205,0.1619505584,-0.0454495959,-0.0742422566,-0.5089077353,-0.0537511483,-0.3977300525,0.2291778177,0.0653380007,0.4077839553,-0.080859229,0.0933067128,-0.0127246631,-0.0308022872,0.4612418413,-0.1670759618,0.2261899412,-0.05589284,-0.0066081695,-0.2543902397,-0.2947322428,-0.0410116352,0.0976036489,0.0445083529,0.2930547893,-0.5040320754,0.0814427435,0.2037864923,0.3086685538,0.0089985374,-0.2055960745,-0.2282193899,-0.2156246006,-0.4626051486,0.0130481543,0.1228732616,0.191258207,-0.0961637199,-0.0554634035,0.0683327392,-0.0295074414,0.0314823836,-0.3027850986,0.2211989164,0.1362633556,0.2531533539,0.0497703888,-0.0033726569,0.0210347697,0.8041908741,-0.0797758326,-0.2488984317,0.1816279739,-0.238103956,-0.2277027369,0.2497159839,-0.1022694483,-0.1096461639,0.2862842977,0.0119943703,-0.004020642,-0.08686883,0.356418699,0.0856344029,-0.2989564836,-0.3246103823,0.3209982216,0.0831265002,-0.0425068401,0.3991122544,0.2725443542,0.0491184928,-0.1712504476,-0.0399865173,0.8578272462,0.0498328321,0.085219115,0.0202612877,-0.0567888767,0.4342235327,-0.2464381903,-0.0729638934,-0.2406994998,-0.3326307535,-0.0846639425,-0.209235087,0.082740441,-0.0030932317,-0.0626889318,0.3957354724,-0.2062758654,0.3918378949,0.1450444907,0.1694121063,-0.3521957994,0.1240327954,-0.1580896676,0.1550447196,-0.2077642232,0.4187972546,-0.1372315288,-0.0071842307,-0.3030921221,-0.1701449156,-0.7933596373,0.1383823007,0.0477104522,0.0148500502,0.3436414301,-0.1029467061,0.0108632185,0.2192789465,0.5752388835,0.0204204991,-0.3639788926,0.198046267,-0.0757807121,-0.0254166815,-0.0132437302,-0.0763356686,0.3167336285,-0.333546102,-0.2046466023,0.1010549143,-0.1129523888,-0.404710412,-0.165486455,-0.1741244197,0.1597436517,-0.4395293295,-0.11076051,-0.1657620519,0.114732787,-0.2908715308,0.1486520618,0.2463239878,0.0042935251,0.1918426007,0.1145737842,-0.0875285342,-0.2222450972,0.5563127398,-0.1080517918,0.0043464587,0.3435761631,0.2170304954,-0.2311031222,-0.1621931791,0.1701336056,-0.0165814329,-0.5097993016,0.0385338403,-0.3021158874,0.2016075701,-0.2905807495,0.4739763737,0.3647392094,0.2416931242,0.3363366723,-0.7355579734,0.038638968,-0.0377286896,0.0533793829,0.2291826755,-0.2330833375,-0.3901047111,0.2119872123,-0.2208040357,-0.2780774832,-0.0557719097,-0.3257577717,-0.0768239647,0.0049985326,0.0173162166,0.3589075208,-0.2168374807,0.1106583178,0.0186862499,-0.1566238105,-0.2169034332,-0.1430363655,0.1212925464,0.0578184351,-0.2803612649,0.1490190774,-0.2365238816,-0.129247576,0.1539893895,0.0131596504,0.2773077786,0.0367566645,0.0113030197,-0.0722257048,-0.060302902,-0.2676008046,0.1169219017,-0.1861576289,0.2702559233,0.0370691121,0.2360322773,0.3051555157,-0.0430770554,-0.4035196006,0.0967597365,0.0711774528,-0.086765714,0.2332345396,-0.4593537748,-0.0846240371,0.2411811799,0.4229108095,0.3079669774,-0.2983403206,0.0382731929,0.2060984224,0.2426954806,-0.2796031237,-0.1308420002,0.2574138641,-0.0331753492,-0.0699062645,-0.0807468742,0.3395253718,-0.2981320918,0.2762935162,0.1378074735,0.1826014668,0.2348324358,0.1614445746,0.3945924342,0.0780492201,0.1083945408,0.2343331724,-0.1447662264,-0.0748584047,0.0626418516,0.0293496121,0.0213523097,-0.2543692589,0.2525362372,-0.1574708968,-0.4455317259,0.2843333781,0.1948120892,-0.1493375301,-0.0774936229,-0.0952652693,0.4345990419,-0.196172297,0.0277624372,-0.449728936,0.0599914677,-0.028885711,0.1380897015,-0.0134115499,-0.3608105183,-0.1873149276,0.3162416518,-0.1222870201,-0.088707298,0.4075271785,0.0635419115,-0.3009321988,-0.5092539787,0.2208693922,0.0110245962,0.0231642462,-0.1912496686,0.1664220095,-0.0332765654,-0.1903323382,0.0944038108,0.4231260717,0.2981813252,0.1958506554,0.0610541254,-0.0307475589,0.078722775,-0.0598154664,0.0852559581,0.3819755316,0.1442199498,-0.1878294796,0.3243119121,0.1266505271,-0.2077419013,0.1612347066,-0.164061442,-0.0760766864,-0.1665035039,0.1068913415,-0.2139305323,0.2273306102,-0.0371264592,-0.0536025614,-0.247753486,-0.2532756329,0.1240269914,-0.0356305018,0.0532443263,-0.2184621841,0.0732231885,-0.2117638141,0.4391584396,0.501201272,0.2203856558,-0.1989354044,-0.3539739549,-0.7645021677,0.0755132437,-0.1055578962,0.179570362,0.1648111492,0.0590529218,-0.2350423485,0.3077589869,0.2918091416,-0.1663728803,-0.2167537808,0.1551816463,-0.2801144123,0.027818542,-0.0017781905,0.0392916091,-0.1063555405,-0.3270182312,0.4496087432,-0.0174866188,-0.0518388972,0.057967104,-0.157375291,-0.0224293619,-0.0660107285,0.1293937415,0.1314204186,0.4132753313,0.1115584448,-0.2058252543,-0.1611676961,-0.1233502701,-0.088525407,0.0245666113,-0.1909865737,0.4489951134,-0.0743558854,0.01403658,-0.2581718862,0.4458042383,0.0753708035,-0.1675117612,-0.2399927378,-0.0571002364,-0.0309006628,0.0237363838,0.1177297384,-0.0577359535,-0.166467458,0.207793355,-0.2280246168,-0.1409021616,0.2995199263,-0.0260362905,-0.1205229536,-0.2113777101,0.0801455677,0.1370503753,-0.2118641734,-0.4640899599,0.1963191777,0.3802842498,0.1471931189,-0.2640032768,0.2581161261,-0.1581685692,0.1585735381,0.1285286248,0.3842264116,0.0633042827,-0.2160658389,0.319385916,-0.1347644627]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2645","title":"load_dataset processing failed with OS error after downloading a dataset","comments":"Hi ! It looks like an issue with pytorch.\r\n\r\nCould you try to run `import torch` and see if it raises an error ?","body":"## Describe the bug\r\nAfter downloading a dataset like opus100, there is a bug that \r\nOSError: Cannot find data file.\r\nOriginal error:\r\ndlopen: cannot load any more object with static TLS\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nthis_dataset = load_dataset('opus100', 'af-en')\r\n```\r\n\r\n## Expected results\r\nthere is no error when running load_dataset.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prep\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 989, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 952, in encode_example\r\n example = cast_to_python_objects(example)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 219, in cast_to_python_ob\r\n return _cast_to_python_objects(obj)[0]\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 165, in _cast_to_python_o\r\n import torch\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/torch\/__init__.py\", line 188, in \r\n _load_global_deps()\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/torch\/__init__.py\", line 141, in _load_global_deps\r\n ctypes.CDLL(lib_path, mode=ctypes.RTLD_GLOBAL)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/ctypes\/__init__.py\", line 348, in __init__\r\n self._handle = _dlopen(self._name, mode)\r\nOSError: dlopen: cannot load any more object with static TLS\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"download_hub_opus100.py\", line 9, in \r\n this_dataset = load_dataset('opus100', language_pair)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepa\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prep\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\ndlopen: cannot load any more object with static TLS\r\n\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.13.0-32-generic-x86_64-with-debian-jessie-sid\r\n- Python version: 3.6.6\r\n- PyArrow version: 3.0.0\r\n\r\n\r\n","comment_length":24,"text":"load_dataset processing failed with OS error after downloading a dataset\n## Describe the bug\r\nAfter downloading a dataset like opus100, there is a bug that \r\nOSError: Cannot find data file.\r\nOriginal error:\r\ndlopen: cannot load any more object with static TLS\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nthis_dataset = load_dataset('opus100', 'af-en')\r\n```\r\n\r\n## Expected results\r\nthere is no error when running load_dataset.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prep\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 989, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 952, in encode_example\r\n example = cast_to_python_objects(example)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 219, in cast_to_python_ob\r\n return _cast_to_python_objects(obj)[0]\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 165, in _cast_to_python_o\r\n import torch\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/torch\/__init__.py\", line 188, in \r\n _load_global_deps()\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/torch\/__init__.py\", line 141, in _load_global_deps\r\n ctypes.CDLL(lib_path, mode=ctypes.RTLD_GLOBAL)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/ctypes\/__init__.py\", line 348, in __init__\r\n self._handle = _dlopen(self._name, mode)\r\nOSError: dlopen: cannot load any more object with static TLS\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"download_hub_opus100.py\", line 9, in \r\n this_dataset = load_dataset('opus100', language_pair)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepa\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prep\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\ndlopen: cannot load any more object with static TLS\r\n\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.13.0-32-generic-x86_64-with-debian-jessie-sid\r\n- Python version: 3.6.6\r\n- PyArrow version: 3.0.0\r\n\r\n\r\n\nHi ! It looks like an issue with pytorch.\r\n\r\nCould you try to run `import torch` and see if it raises an error ?","embeddings":[-0.4430502951,0.2031198442,-0.0454366133,0.484775424,0.30398193,-0.0233756881,0.2349691838,0.3660457134,-0.1082289293,0.0927524492,-0.1625688374,0.4678144753,-0.0074352133,-0.0836081207,-0.0174168609,-0.1323251277,0.0387222506,0.2778642178,-0.6045764089,-0.0482679792,-0.4312471151,0.3115380704,-0.1030293554,-0.0990164578,0.0378610119,0.0749130622,-0.1906295121,0.3197152019,-0.0947884694,-0.260250777,-0.020933656,-0.1710368693,0.2840125263,0.5108674169,-0.0001183986,0.2902742326,0.0925462469,-0.0907713547,-0.402387321,-0.1111452058,-0.0832587332,-0.1031597853,0.148237586,-0.3790362477,-0.0424588993,-0.2175002843,-0.0282115731,-0.5162989497,0.2104209661,0.3379337788,0.1746215373,0.374222666,0.2902139723,-0.148411572,0.254062593,-0.0283484124,-0.0414075442,0.048455026,0.1556954682,-0.0565809533,0.0651961565,0.292029947,-0.1614065915,0.1469640583,0.2089904547,0.0581361279,0.2701870203,-0.33701846,0.3568376005,0.1577442288,0.5403915048,-0.0280469693,-0.3331862986,-0.187848255,0.0130488845,-0.4298108518,0.2435779721,0.2214604914,-0.0285913814,0.0868347734,0.1412747353,-0.1238947809,-0.0940341353,0.3054019213,-0.042060364,0.2001214623,-0.0576187335,-0.0591642372,0.0631149262,0.2239584625,0.2793515921,0.0497795753,-0.0100777987,0.3172575235,-0.4745251834,0.1773097366,0.1719004661,0.0040303003,0.0933709666,-0.0650276318,0.1036134586,-0.2318625599,-0.0230106711,-0.0736536384,0.2052522302,0.0471473411,-0.2928019464,-0.1984692216,0.1404688209,0.0625927523,-0.1818164736,-0.0956845358,-0.1932927221,-0.3410633504,0.0759849101,0.1347163916,0.5756460428,-0.1788666844,-0.3567049801,-0.0537942536,-0.1881234646,-0.0532148778,0.106923528,0.2891930044,-0.0607561283,0.4205152392,0.2440235764,0.1303139478,-0.3833307922,0.0157029722,-0.1193970069,-0.2396061271,-0.1751903147,-0.018225871,0.064439863,-0.4865529239,0.3154323995,0.1732182652,-0.0735439509,-0.3623705506,0.3777589798,-0.079421401,-0.0954614356,0.2311086208,-0.0021412387,-0.1261712462,0.1479644179,-0.0137931965,0.0977248177,0.4053345621,-0.4959002435,-0.5249584317,-0.4787271023,0.171328932,-0.3053381145,0.1398524195,0.3070881665,-0.1751901656,0.1507404149,-0.1544900537,-0.098591283,-0.2091853321,-0.3328261971,-0.1935880631,0.4081570506,0.4952768385,-0.3194479048,0.0553343706,-0.252432704,-0.1881270856,-0.1042026207,-0.0230471604,-0.0677800998,0.3877436221,-0.5029060841,-0.2613579333,0.2853173614,-0.442951858,-0.6753488779,0.0294269733,-0.2476943135,0.0424231142,0.3182734847,0.0589966103,0.3740208149,-0.0094537772,-0.1303220242,0.2131943554,0.1354910731,0.0185981672,-0.1482967734,-0.1480104178,0.0453190804,0.2436876893,0.0738346502,0.0752403513,0.1276132911,0.1137061119,0.2993844151,-0.1438609511,-0.1832547486,0.1096934155,0.2658593357,0.1396879703,0.2437562943,-0.0872695222,-0.4674577117,0.0827427655,-0.0533501357,0.0326676778,-0.2187159657,-0.0448126234,-0.1293059886,0.0956431851,-0.2936310172,-0.3219188452,0.0730179325,0.1981949508,0.0066885557,0.0158333015,-0.086826086,0.3457705677,0.1164511517,0.1674299091,-0.2508371174,0.330353111,0.1207666919,-0.0379181989,-0.0812918246,0.3204132915,0.0453062765,-0.2102501094,-0.2788046896,0.4593667984,0.1484796256,-0.0225258488,-0.0986613557,0.1753183156,0.2358965874,-0.2645653784,-0.098218061,0.0409373902,0.2143257558,0.0708438754,0.1004788578,0.0042706537,-0.1008393019,0.1319384128,0.1286811531,-0.004212575,0.0599547587,-0.2399789393,0.035048604,-0.0927790403,0.364774555,-0.0100674117,0.3398686349,0.1148416102,-0.2486354411,-0.1168410107,0.4487637579,-0.0409491025,0.2245076895,0.1049687639,-0.2105497122,0.2330356687,-0.0582606643,0.1338374168,0.763418138,0.1158927903,-0.3087420464,0.0303173065,0.1568048,-0.0560591407,0.3047151864,0.2135780603,0.5176039934,0.4227158427,0.0548551455,0.2111869901,-0.1041079387,-0.4372707903,0.1301827282,0.2429524511,-0.3268970549,-0.0282367282,-0.2275322825,-0.0035175544,-0.17453444,-0.0063033402,-0.1977397352,-0.1105283424,-0.003894232,0.11990612,-0.0541347824,0.0521627367,-0.2768771946,0.0833762586,0.2396606207,-0.2275040299,0.1765099317,-0.0787875429,-0.1228342876,-0.0798948407,0.3425710499,-0.1298445314,0.2565789223,-0.3259460628,0.1347655356,-0.1751976609,-0.212033093,0.1560975164,0.2870919406,0.1585514694,0.1268410534,0.212009266,0.1711961478,-0.2368269414,0.2334577441,-0.1223337799,-0.0551896133,0.111345306,-0.0120252324,0.2442098558,-0.047289893,-0.4643704593,-0.3488773406,-0.4888262451,-0.0874160677,-0.0671280771,0.2688062191,0.1994024813,0.1261017323,0.2629390359,0.2930875719,0.0473046899,-0.0196140371,-0.1887223721,0.2164980024,-0.303324461,-0.2178200781,-0.2564584911,0.0712681711,0.2940013111,0.0316479616,-0.6996409893,-0.3688041568,-0.0958594903,0.1920544207,-0.4736225605,-0.0218408015,0.2243337035,0.0982552692,0.023132002,-0.3120465875,0.0890534148,0.0811186284,-0.1509288847,0.2234877795,-0.1041746438,0.4505559206,-0.2182901651,0.5835216045,-0.0896609947,-0.2326279879,0.494356215,-0.1522163749,0.3561162949,-0.1552245617,-0.2526461184,-0.204453364,-0.0917278007,0.0118409358,0.2949996889,0.0903794318,0.009650996,-0.2692483664,0.0204118025,-0.447409302,-0.2161026299,0.1328551918,-0.1035332307,0.2543566525,-0.0747548789,0.2682151794,-0.1886687428,-0.0850550905,0.1154112816,0.2735349834,0.1169504449,-0.0427167229,0.0068599954,-0.0428304709,-0.1197035462,0.2423281372,-0.1672645509,0.3011179268,-0.0396490991,-0.103893064,0.0168126728,0.0017572303,0.8671812415,0.1616336852,0.0611756556,0.0744498223,0.0017681328,-0.2096125484,-0.1531897336,-0.0486999862,0.559866786,0.4752474129,0.610046804,-0.1361040026,-0.2062624395,0.1989943981,0.2431159616,-0.0953741521,-0.4367493689,-0.32781896,-0.3320827186,-0.2650568485,-0.0560366884,-0.0521913506,0.4572674334,-0.1983002126,0.0208346862,-0.1782423705,-0.4005160332,-0.1174488962,0.0723976493,0.4080714583,-0.0451458469,0.3436599374,0.1963219792,0.3403641284,0.3059301972,0.630396843,-0.1995369792,-0.5900688171,0.3027574718,-0.1965070069,0.0657598078,0.1468281299,-0.0596202798,-0.1145514175,0.0031646176,0.1844054163,-0.0919595808,0.0092866486,0.2338160425,0.0345937759,-0.3580452204,-0.3023239374,0.3559458554,-0.0250899177,0.0034085766,0.3342488408,-0.2428393662,-0.1591667384,0.5583699942,-0.0221915655,0.8753305674,-0.1954441816,-0.0461548939,0.0339419879,-0.2444019169,0.7470536828,-0.4442422986,0.1304815114,-0.3095660508,-0.3840742409,0.041841343,-0.2273294628,0.191700533,0.0126245022,-0.1538767964,0.1041535586,-0.1464486718,0.1213339791,0.1688227654,0.3585916162,-0.192875877,-0.154781878,-0.2013199478,-0.0067660511,0.0725001022,0.4714586437,-0.1576274037,-0.2567614615,-0.1816907227,-0.0251296349,-0.0580865145,0.1820368767,-0.5054666996,0.2635750771,-0.0394345596,-0.5898033381,0.0116450442,0.7242991924,0.0000022443,0.1009300277,-0.0902633443,0.2722335458,0.1893350184,0.0081513533,0.2704310715,-0.0907730907,0.248292774,-0.07753966,-0.1894851923,0.0635934547,-0.2281297147,-0.240160495,0.4469697773,-0.0698916465,0.3370924592,-0.2419395894,-0.1958490312,-0.1846798956,-0.0030680869,-0.106199488,0.0852934197,0.0861071572,-0.2696993053,0.0032323056,-0.1789136976,-0.4170422256,-0.030387206,0.3435989022,-0.1027674228,0.1958585978,0.5353897214,0.2496230006,0.0700473338,-0.1286619008,0.0876137018,-0.0788822025,-0.8638370633,0.0030893139,0.1329429448,0.1910300851,-0.1324961036,-0.3880008459,-0.0416091643,0.0246454943,-0.2592012584,-0.8063874841,-0.1279871166,0.0003277925,-0.0555134118,0.1209058613,0.0125717521,0.1814002395,-0.1055719927,-0.1266915947,-0.2139290869,0.113836728,-0.0415839553,-0.0350903012,0.208737269,-0.0261919517,0.0522150025,-0.1148363352,0.1224132404,0.132281661,-0.0545900911,-0.1592982262,-0.1986370534,0.1914595813,-0.1697370559,-0.1901634187,-0.069498904,-0.2603869736,-0.082346417,-0.0308069028,0.1341706663,-0.1280405223,-0.134207353,-0.1807747185,0.0529537685,0.1350380629,-0.2587348819,0.3014797568,-0.1523777694,0.1533981562,-0.0465169363,0.3444420993,-0.1278180778,0.0601228029,-0.2831707895,-0.2150336951,-0.0194683336,0.0541305691,0.2921988666,-0.2612036169,0.0437421277,0.2345066369,0.0612110458,0.2389376014,-0.2238367051,0.0480348058,0.1548053175,0.2138174176,-0.2007173747,0.0465119332,0.2298102826,0.2254763693,-0.0167983789,0.1226183176,0.1225849539,-0.0604850501,-0.0298845451,0.0791068822,0.262686789,0.4071723223,0.3707687259,0.8125104308,0.031991221,0.2267281264,-0.1681687981,-0.2003173828,-0.013751138,0.5678642392,-0.1570239961,0.148841843,0.1619775742,0.14955917,0.073627606,-0.4403559268,-0.1011140421,-0.0547245592,-0.0171095449,0.1752587408,-0.2550035715,0.5106303692,-0.2262840271,0.0769669712,-0.2533049881,0.3183328509,-0.1050111279,0.0823977515,-0.0645420402,-0.1711417139,0.1584534347,0.0610035509,-0.009204614,0.0408185162,-0.2774723768,0.2281303555,-0.2549317181,-0.1301789433,-0.3134045005,0.3595224619,-0.0539512411,-0.1707004458,0.359572202,0.3049595058,0.1747180372,0.1891994178,0.3637140691,0.4831361175,0.4503183961,-0.3225102127,0.0358875319,0.1269645691,0.0891217515,-0.0236321278,-0.006765008,0.0139688142,0.2668419778,0.3173691332,0.0636152476,-0.1635799557,0.050705675,0.0274479333,0.1804313809,-0.2532272935,-0.0266466737,-0.0367832556,-0.1780875623,-0.1143435463,0.0704004765,-0.2277838141,-0.1253740937,0.3511192501,-0.0495121218,-0.0020054688,0.0596436821,0.0218053944,-0.2592401505,0.3652291894,0.2182929814,0.5810967684,-0.2674711049,-0.2173576206,-0.6972265244,0.3747780025,-0.0192843415,-0.2902251184,0.2166699022,-0.0611446574,-0.0355450101,0.155581966,0.3550474346,-0.0920129195,0.359595418,0.2411309183,-0.4858283699,-0.3335185349,-0.4250916541,0.2082862556,-0.0554010309,-0.4326459765,0.0748032629,0.0352406353,0.0519154668,-0.190416038,-0.1346190423,-0.0248721167,-0.0795990899,0.4432128966,-0.0042051258,0.4430550337,-0.0852830783,-0.0573904589,0.0039829533,-0.1729486436,-0.1389403939,0.1609298438,0.3259431422,0.0645698756,-0.1252213717,-0.3071177006,-0.2925894558,0.414791584,-0.0671640635,0.3546156585,-0.3536361754,0.0940052494,-0.1247276142,0.0684199706,0.2927663028,0.1802985221,-0.1506899148,0.1551077664,-0.2595286369,-0.1846695691,0.4124612212,-0.6139751673,0.065929316,-0.0619227327,0.3500474393,0.0316230431,0.1475655586,-0.3610766828,0.0741277635,0.4486697316,0.0358206481,-0.3080541193,-0.0032146326,0.1185030639,0.1700793505,0.0924642012,0.310846746,-0.0426308848,-0.3124408424,0.0873371586,-0.1259962916]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2645","title":"load_dataset processing failed with OS error after downloading a dataset","comments":"> Hi ! It looks like an issue with pytorch.\r\n> \r\n> Could you try to run `import torch` and see if it raises an error ?\r\n\r\nIt works. Thank you!","body":"## Describe the bug\r\nAfter downloading a dataset like opus100, there is a bug that \r\nOSError: Cannot find data file.\r\nOriginal error:\r\ndlopen: cannot load any more object with static TLS\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nthis_dataset = load_dataset('opus100', 'af-en')\r\n```\r\n\r\n## Expected results\r\nthere is no error when running load_dataset.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prep\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 989, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 952, in encode_example\r\n example = cast_to_python_objects(example)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 219, in cast_to_python_ob\r\n return _cast_to_python_objects(obj)[0]\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 165, in _cast_to_python_o\r\n import torch\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/torch\/__init__.py\", line 188, in \r\n _load_global_deps()\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/torch\/__init__.py\", line 141, in _load_global_deps\r\n ctypes.CDLL(lib_path, mode=ctypes.RTLD_GLOBAL)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/ctypes\/__init__.py\", line 348, in __init__\r\n self._handle = _dlopen(self._name, mode)\r\nOSError: dlopen: cannot load any more object with static TLS\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"download_hub_opus100.py\", line 9, in \r\n this_dataset = load_dataset('opus100', language_pair)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepa\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prep\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\ndlopen: cannot load any more object with static TLS\r\n\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.13.0-32-generic-x86_64-with-debian-jessie-sid\r\n- Python version: 3.6.6\r\n- PyArrow version: 3.0.0\r\n\r\n\r\n","comment_length":31,"text":"load_dataset processing failed with OS error after downloading a dataset\n## Describe the bug\r\nAfter downloading a dataset like opus100, there is a bug that \r\nOSError: Cannot find data file.\r\nOriginal error:\r\ndlopen: cannot load any more object with static TLS\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nthis_dataset = load_dataset('opus100', 'af-en')\r\n```\r\n\r\n## Expected results\r\nthere is no error when running load_dataset.\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prep\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 989, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 952, in encode_example\r\n example = cast_to_python_objects(example)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 219, in cast_to_python_ob\r\n return _cast_to_python_objects(obj)[0]\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/features.py\", line 165, in _cast_to_python_o\r\n import torch\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/torch\/__init__.py\", line 188, in \r\n _load_global_deps()\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/torch\/__init__.py\", line 141, in _load_global_deps\r\n ctypes.CDLL(lib_path, mode=ctypes.RTLD_GLOBAL)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/ctypes\/__init__.py\", line 348, in __init__\r\n self._handle = _dlopen(self._name, mode)\r\nOSError: dlopen: cannot load any more object with static TLS\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"download_hub_opus100.py\", line 9, in \r\n this_dataset = load_dataset('opus100', language_pair)\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepa\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/anaconda3\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 658, in _download_and_prep\r\n + str(e)\r\nOSError: Cannot find data file.\r\nOriginal error:\r\ndlopen: cannot load any more object with static TLS\r\n\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.13.0-32-generic-x86_64-with-debian-jessie-sid\r\n- Python version: 3.6.6\r\n- PyArrow version: 3.0.0\r\n\r\n\r\n\n> Hi ! It looks like an issue with pytorch.\r\n> \r\n> Could you try to run `import torch` and see if it raises an error ?\r\n\r\nIt works. Thank you!","embeddings":[-0.4430502951,0.2031198442,-0.0454366133,0.484775424,0.30398193,-0.0233756881,0.2349691838,0.3660457134,-0.1082289293,0.0927524492,-0.1625688374,0.4678144753,-0.0074352133,-0.0836081207,-0.0174168609,-0.1323251277,0.0387222506,0.2778642178,-0.6045764089,-0.0482679792,-0.4312471151,0.3115380704,-0.1030293554,-0.0990164578,0.0378610119,0.0749130622,-0.1906295121,0.3197152019,-0.0947884694,-0.260250777,-0.020933656,-0.1710368693,0.2840125263,0.5108674169,-0.0001183986,0.2902742326,0.0925462469,-0.0907713547,-0.402387321,-0.1111452058,-0.0832587332,-0.1031597853,0.148237586,-0.3790362477,-0.0424588993,-0.2175002843,-0.0282115731,-0.5162989497,0.2104209661,0.3379337788,0.1746215373,0.374222666,0.2902139723,-0.148411572,0.254062593,-0.0283484124,-0.0414075442,0.048455026,0.1556954682,-0.0565809533,0.0651961565,0.292029947,-0.1614065915,0.1469640583,0.2089904547,0.0581361279,0.2701870203,-0.33701846,0.3568376005,0.1577442288,0.5403915048,-0.0280469693,-0.3331862986,-0.187848255,0.0130488845,-0.4298108518,0.2435779721,0.2214604914,-0.0285913814,0.0868347734,0.1412747353,-0.1238947809,-0.0940341353,0.3054019213,-0.042060364,0.2001214623,-0.0576187335,-0.0591642372,0.0631149262,0.2239584625,0.2793515921,0.0497795753,-0.0100777987,0.3172575235,-0.4745251834,0.1773097366,0.1719004661,0.0040303003,0.0933709666,-0.0650276318,0.1036134586,-0.2318625599,-0.0230106711,-0.0736536384,0.2052522302,0.0471473411,-0.2928019464,-0.1984692216,0.1404688209,0.0625927523,-0.1818164736,-0.0956845358,-0.1932927221,-0.3410633504,0.0759849101,0.1347163916,0.5756460428,-0.1788666844,-0.3567049801,-0.0537942536,-0.1881234646,-0.0532148778,0.106923528,0.2891930044,-0.0607561283,0.4205152392,0.2440235764,0.1303139478,-0.3833307922,0.0157029722,-0.1193970069,-0.2396061271,-0.1751903147,-0.018225871,0.064439863,-0.4865529239,0.3154323995,0.1732182652,-0.0735439509,-0.3623705506,0.3777589798,-0.079421401,-0.0954614356,0.2311086208,-0.0021412387,-0.1261712462,0.1479644179,-0.0137931965,0.0977248177,0.4053345621,-0.4959002435,-0.5249584317,-0.4787271023,0.171328932,-0.3053381145,0.1398524195,0.3070881665,-0.1751901656,0.1507404149,-0.1544900537,-0.098591283,-0.2091853321,-0.3328261971,-0.1935880631,0.4081570506,0.4952768385,-0.3194479048,0.0553343706,-0.252432704,-0.1881270856,-0.1042026207,-0.0230471604,-0.0677800998,0.3877436221,-0.5029060841,-0.2613579333,0.2853173614,-0.442951858,-0.6753488779,0.0294269733,-0.2476943135,0.0424231142,0.3182734847,0.0589966103,0.3740208149,-0.0094537772,-0.1303220242,0.2131943554,0.1354910731,0.0185981672,-0.1482967734,-0.1480104178,0.0453190804,0.2436876893,0.0738346502,0.0752403513,0.1276132911,0.1137061119,0.2993844151,-0.1438609511,-0.1832547486,0.1096934155,0.2658593357,0.1396879703,0.2437562943,-0.0872695222,-0.4674577117,0.0827427655,-0.0533501357,0.0326676778,-0.2187159657,-0.0448126234,-0.1293059886,0.0956431851,-0.2936310172,-0.3219188452,0.0730179325,0.1981949508,0.0066885557,0.0158333015,-0.086826086,0.3457705677,0.1164511517,0.1674299091,-0.2508371174,0.330353111,0.1207666919,-0.0379181989,-0.0812918246,0.3204132915,0.0453062765,-0.2102501094,-0.2788046896,0.4593667984,0.1484796256,-0.0225258488,-0.0986613557,0.1753183156,0.2358965874,-0.2645653784,-0.098218061,0.0409373902,0.2143257558,0.0708438754,0.1004788578,0.0042706537,-0.1008393019,0.1319384128,0.1286811531,-0.004212575,0.0599547587,-0.2399789393,0.035048604,-0.0927790403,0.364774555,-0.0100674117,0.3398686349,0.1148416102,-0.2486354411,-0.1168410107,0.4487637579,-0.0409491025,0.2245076895,0.1049687639,-0.2105497122,0.2330356687,-0.0582606643,0.1338374168,0.763418138,0.1158927903,-0.3087420464,0.0303173065,0.1568048,-0.0560591407,0.3047151864,0.2135780603,0.5176039934,0.4227158427,0.0548551455,0.2111869901,-0.1041079387,-0.4372707903,0.1301827282,0.2429524511,-0.3268970549,-0.0282367282,-0.2275322825,-0.0035175544,-0.17453444,-0.0063033402,-0.1977397352,-0.1105283424,-0.003894232,0.11990612,-0.0541347824,0.0521627367,-0.2768771946,0.0833762586,0.2396606207,-0.2275040299,0.1765099317,-0.0787875429,-0.1228342876,-0.0798948407,0.3425710499,-0.1298445314,0.2565789223,-0.3259460628,0.1347655356,-0.1751976609,-0.212033093,0.1560975164,0.2870919406,0.1585514694,0.1268410534,0.212009266,0.1711961478,-0.2368269414,0.2334577441,-0.1223337799,-0.0551896133,0.111345306,-0.0120252324,0.2442098558,-0.047289893,-0.4643704593,-0.3488773406,-0.4888262451,-0.0874160677,-0.0671280771,0.2688062191,0.1994024813,0.1261017323,0.2629390359,0.2930875719,0.0473046899,-0.0196140371,-0.1887223721,0.2164980024,-0.303324461,-0.2178200781,-0.2564584911,0.0712681711,0.2940013111,0.0316479616,-0.6996409893,-0.3688041568,-0.0958594903,0.1920544207,-0.4736225605,-0.0218408015,0.2243337035,0.0982552692,0.023132002,-0.3120465875,0.0890534148,0.0811186284,-0.1509288847,0.2234877795,-0.1041746438,0.4505559206,-0.2182901651,0.5835216045,-0.0896609947,-0.2326279879,0.494356215,-0.1522163749,0.3561162949,-0.1552245617,-0.2526461184,-0.204453364,-0.0917278007,0.0118409358,0.2949996889,0.0903794318,0.009650996,-0.2692483664,0.0204118025,-0.447409302,-0.2161026299,0.1328551918,-0.1035332307,0.2543566525,-0.0747548789,0.2682151794,-0.1886687428,-0.0850550905,0.1154112816,0.2735349834,0.1169504449,-0.0427167229,0.0068599954,-0.0428304709,-0.1197035462,0.2423281372,-0.1672645509,0.3011179268,-0.0396490991,-0.103893064,0.0168126728,0.0017572303,0.8671812415,0.1616336852,0.0611756556,0.0744498223,0.0017681328,-0.2096125484,-0.1531897336,-0.0486999862,0.559866786,0.4752474129,0.610046804,-0.1361040026,-0.2062624395,0.1989943981,0.2431159616,-0.0953741521,-0.4367493689,-0.32781896,-0.3320827186,-0.2650568485,-0.0560366884,-0.0521913506,0.4572674334,-0.1983002126,0.0208346862,-0.1782423705,-0.4005160332,-0.1174488962,0.0723976493,0.4080714583,-0.0451458469,0.3436599374,0.1963219792,0.3403641284,0.3059301972,0.630396843,-0.1995369792,-0.5900688171,0.3027574718,-0.1965070069,0.0657598078,0.1468281299,-0.0596202798,-0.1145514175,0.0031646176,0.1844054163,-0.0919595808,0.0092866486,0.2338160425,0.0345937759,-0.3580452204,-0.3023239374,0.3559458554,-0.0250899177,0.0034085766,0.3342488408,-0.2428393662,-0.1591667384,0.5583699942,-0.0221915655,0.8753305674,-0.1954441816,-0.0461548939,0.0339419879,-0.2444019169,0.7470536828,-0.4442422986,0.1304815114,-0.3095660508,-0.3840742409,0.041841343,-0.2273294628,0.191700533,0.0126245022,-0.1538767964,0.1041535586,-0.1464486718,0.1213339791,0.1688227654,0.3585916162,-0.192875877,-0.154781878,-0.2013199478,-0.0067660511,0.0725001022,0.4714586437,-0.1576274037,-0.2567614615,-0.1816907227,-0.0251296349,-0.0580865145,0.1820368767,-0.5054666996,0.2635750771,-0.0394345596,-0.5898033381,0.0116450442,0.7242991924,0.0000022443,0.1009300277,-0.0902633443,0.2722335458,0.1893350184,0.0081513533,0.2704310715,-0.0907730907,0.248292774,-0.07753966,-0.1894851923,0.0635934547,-0.2281297147,-0.240160495,0.4469697773,-0.0698916465,0.3370924592,-0.2419395894,-0.1958490312,-0.1846798956,-0.0030680869,-0.106199488,0.0852934197,0.0861071572,-0.2696993053,0.0032323056,-0.1789136976,-0.4170422256,-0.030387206,0.3435989022,-0.1027674228,0.1958585978,0.5353897214,0.2496230006,0.0700473338,-0.1286619008,0.0876137018,-0.0788822025,-0.8638370633,0.0030893139,0.1329429448,0.1910300851,-0.1324961036,-0.3880008459,-0.0416091643,0.0246454943,-0.2592012584,-0.8063874841,-0.1279871166,0.0003277925,-0.0555134118,0.1209058613,0.0125717521,0.1814002395,-0.1055719927,-0.1266915947,-0.2139290869,0.113836728,-0.0415839553,-0.0350903012,0.208737269,-0.0261919517,0.0522150025,-0.1148363352,0.1224132404,0.132281661,-0.0545900911,-0.1592982262,-0.1986370534,0.1914595813,-0.1697370559,-0.1901634187,-0.069498904,-0.2603869736,-0.082346417,-0.0308069028,0.1341706663,-0.1280405223,-0.134207353,-0.1807747185,0.0529537685,0.1350380629,-0.2587348819,0.3014797568,-0.1523777694,0.1533981562,-0.0465169363,0.3444420993,-0.1278180778,0.0601228029,-0.2831707895,-0.2150336951,-0.0194683336,0.0541305691,0.2921988666,-0.2612036169,0.0437421277,0.2345066369,0.0612110458,0.2389376014,-0.2238367051,0.0480348058,0.1548053175,0.2138174176,-0.2007173747,0.0465119332,0.2298102826,0.2254763693,-0.0167983789,0.1226183176,0.1225849539,-0.0604850501,-0.0298845451,0.0791068822,0.262686789,0.4071723223,0.3707687259,0.8125104308,0.031991221,0.2267281264,-0.1681687981,-0.2003173828,-0.013751138,0.5678642392,-0.1570239961,0.148841843,0.1619775742,0.14955917,0.073627606,-0.4403559268,-0.1011140421,-0.0547245592,-0.0171095449,0.1752587408,-0.2550035715,0.5106303692,-0.2262840271,0.0769669712,-0.2533049881,0.3183328509,-0.1050111279,0.0823977515,-0.0645420402,-0.1711417139,0.1584534347,0.0610035509,-0.009204614,0.0408185162,-0.2774723768,0.2281303555,-0.2549317181,-0.1301789433,-0.3134045005,0.3595224619,-0.0539512411,-0.1707004458,0.359572202,0.3049595058,0.1747180372,0.1891994178,0.3637140691,0.4831361175,0.4503183961,-0.3225102127,0.0358875319,0.1269645691,0.0891217515,-0.0236321278,-0.006765008,0.0139688142,0.2668419778,0.3173691332,0.0636152476,-0.1635799557,0.050705675,0.0274479333,0.1804313809,-0.2532272935,-0.0266466737,-0.0367832556,-0.1780875623,-0.1143435463,0.0704004765,-0.2277838141,-0.1253740937,0.3511192501,-0.0495121218,-0.0020054688,0.0596436821,0.0218053944,-0.2592401505,0.3652291894,0.2182929814,0.5810967684,-0.2674711049,-0.2173576206,-0.6972265244,0.3747780025,-0.0192843415,-0.2902251184,0.2166699022,-0.0611446574,-0.0355450101,0.155581966,0.3550474346,-0.0920129195,0.359595418,0.2411309183,-0.4858283699,-0.3335185349,-0.4250916541,0.2082862556,-0.0554010309,-0.4326459765,0.0748032629,0.0352406353,0.0519154668,-0.190416038,-0.1346190423,-0.0248721167,-0.0795990899,0.4432128966,-0.0042051258,0.4430550337,-0.0852830783,-0.0573904589,0.0039829533,-0.1729486436,-0.1389403939,0.1609298438,0.3259431422,0.0645698756,-0.1252213717,-0.3071177006,-0.2925894558,0.414791584,-0.0671640635,0.3546156585,-0.3536361754,0.0940052494,-0.1247276142,0.0684199706,0.2927663028,0.1802985221,-0.1506899148,0.1551077664,-0.2595286369,-0.1846695691,0.4124612212,-0.6139751673,0.065929316,-0.0619227327,0.3500474393,0.0316230431,0.1475655586,-0.3610766828,0.0741277635,0.4486697316,0.0358206481,-0.3080541193,-0.0032146326,0.1185030639,0.1700793505,0.0924642012,0.310846746,-0.0426308848,-0.3124408424,0.0873371586,-0.1259962916]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2644","title":"Batched `map` not allowed to return 0 items","comments":"Hi ! Thanks for reporting. Indeed it looks like type inference makes it fail. We should probably just ignore this step until a non-empty batch is passed.","body":"## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":27,"text":"Batched `map` not allowed to return 0 items\n## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n\nHi ! Thanks for reporting. Indeed it looks like type inference makes it fail. We should probably just ignore this step until a non-empty batch is passed.","embeddings":[-0.2328359336,-0.3903888166,-0.0444566682,0.1644899547,-0.1085444242,-0.0687690005,0.1163394973,0.4056320786,0.6400417089,0.1178812236,-0.0565718673,0.2460853755,-0.3954412043,0.1043543741,-0.1629210711,0.1608026177,-0.0449331626,0.1405789405,-0.1691175699,-0.1914925724,-0.3362929821,0.163199544,-0.3979255855,-0.0501046143,-0.1487712115,-0.2919759154,0.286480099,-0.0362105146,-0.162814334,-0.2454052418,0.206091553,-0.2328298539,-0.2091356516,0.5783210993,-0.0001299276,-0.0231365338,0.1505937427,-0.1851885766,-0.0897614583,-0.3029606342,-0.1908181012,-0.1715300977,-0.1157183424,-0.2824648917,0.1819460839,-0.2716506422,-0.1956346184,-0.4482398331,0.1071964726,0.2587007582,0.043321602,0.4188180268,0.003066103,0.1764480174,0.3312488496,0.2664009035,-0.0753516629,0.1288110614,0.7205767035,-0.2881007195,0.0226270352,0.3385902047,-0.3921834528,0.0709065422,0.1367558688,-0.0309035499,0.230168879,-0.5515586734,0.2697506845,0.3675674498,-0.0899998844,0.0607856587,-0.433337301,-0.6687820554,-0.0403036699,-0.0954985172,-0.0388978831,0.2810007632,-0.4210395217,-0.0564594232,-0.4846186638,0.1217950135,0.0078949472,0.1618547142,0.0306755304,0.4589543343,-0.0930409953,0.2003241777,0.3451099396,-0.1618324965,-0.0994021893,-0.1672711968,-0.2236276567,0.5847230554,-0.0366987288,-0.1459876001,0.3893203735,0.2197681516,0.2230358422,-0.3132426441,-0.050306201,-0.0742755681,0.2646799386,-0.0234315637,0.4159452617,0.0793761313,0.2686936855,0.3493200243,-0.005052017,-0.0590938851,0.0676903278,0.1131693944,0.113442108,-0.0890844166,-0.0096516423,-0.0764087662,0.2567855418,-0.0787523314,-0.1218986362,0.2341007143,-0.4100169539,0.1369067729,0.0620478503,0.1250830591,0.1700579822,-0.1776706576,0.1202128828,0.1126287505,-0.1570285559,0.1897115409,-0.0772467405,-0.1316168755,-0.192502588,-0.3208249807,0.1964786202,-0.3342862427,0.2418166101,-0.1585832834,0.0198275335,-0.0615327619,0.0957399309,-0.2645617723,0.7512272596,0.4937881231,-0.119511351,0.2913580537,0.3966238201,-0.2185574174,-0.1752834171,0.4062903821,-0.3826135099,-0.2777348757,0.2100417018,-0.0401864871,-0.2380758375,0.346570462,-0.5073054433,0.2185845971,0.1782458276,0.0822512209,0.2914729118,-0.2697472274,-0.105394572,-0.2360114455,0.0425621271,0.6938397884,-0.63877213,-0.049924884,0.1101892963,0.1013790518,-0.0613662116,0.3478800058,-0.0454604253,0.3481105566,-0.4107360542,0.3059869707,0.0873124897,-0.1264710426,-0.3275630772,0.3134204149,-0.0358151942,0.2356395572,-0.0210545417,-0.0642482191,0.5519376397,-0.1873890162,0.3676523268,-0.1700588763,-0.279956907,0.1041627601,-0.0503817089,-0.039843861,0.0603797175,-0.2420822233,0.19492428,0.0165877249,0.0390782356,-0.5488622189,0.1286391169,-0.1508790404,0.3383628428,0.2401701212,0.262108326,0.1233314499,0.0182514824,-0.2563515604,-0.4713618159,0.3001765013,-0.4880333245,0.021248579,-0.4338338673,-0.1033039168,0.0309646595,0.1834742576,-0.0350200944,0.0330956019,-0.0749208331,-0.4975802898,0.1765710711,-0.0213433541,-0.2262366265,-0.2370698303,-0.1358649582,-0.0084387707,-0.3567225337,0.071359843,0.2501084507,-0.2921575904,-0.2542045712,0.2466300279,0.1149316132,-0.1176109165,0.117354244,0.1758526713,-0.0348161124,-0.1673572958,-0.1400127411,0.131253615,0.0117278975,-0.0458009467,0.101837188,0.1450134069,0.3444767892,-0.2378433347,-0.0694068223,0.5054207444,-0.1876149476,0.5359300375,-0.1796862483,0.0512384214,-0.1996322572,0.042606812,-0.2059215009,-0.3850508034,0.0762836039,-0.1577371508,0.0513374284,0.0319146588,0.0012563785,-0.1989199072,0.0856170282,-0.046962332,0.16477938,0.0144881429,-0.0630519465,0.2266895324,0.0446924791,-0.0142778121,0.3432699442,0.2191425711,-0.0887260064,-0.1521824449,0.250440836,0.0001223587,0.4139182568,0.18166776,0.0260531791,0.1550095528,-0.003640976,-0.0657203272,-0.2528341115,-0.1814302802,0.005715027,0.149533838,-0.3952980638,-0.0354201086,0.1101150513,-0.158504501,0.0689762086,-0.0765032619,-0.290856123,-0.3032285273,-0.0453336127,0.2601582408,-0.2722713947,-0.02400795,-0.4013136625,0.0116532287,0.1975094527,-0.2747197151,-0.2812999189,-0.1685465872,-0.0756110176,-0.0741591007,0.0731139854,-0.1339878589,0.3615631759,0.2521368861,-0.2612816691,-0.243771404,-0.0951137543,0.0065136496,-0.5370389223,0.2054035813,0.4100922644,0.3981708586,-0.2923846245,0.1470615864,0.2793858051,-0.1071430445,-0.2480682433,-0.0079056751,0.0223337524,0.153262645,-0.179215163,0.1474315226,0.0644616261,-0.3209299445,0.1781849116,-0.1891590655,0.2615512013,0.2194533795,0.2780077457,0.1538574994,-0.3693777621,-0.4285494387,0.1223512962,-0.1416769028,0.4460211396,0.0390229337,-0.1469308287,0.1844055504,0.0262182392,-0.1853193641,0.2405568659,-0.0966859683,0.0939638689,-0.0562738925,0.1234829947,0.190006569,0.3668004274,0.4985276759,0.0864884332,0.0460447706,-0.2846722603,-0.1089555398,0.0812640563,0.0222764555,-0.064514026,0.422529459,0.3217896223,0.0287001207,0.7504062057,0.3654765189,0.1272135675,0.0946453437,0.0725369975,0.3822363317,-0.1140064001,-0.379692018,-0.2082244754,0.1538844705,-0.0840781853,0.0325130187,0.0550553165,-0.0620342046,-0.0012852303,0.0619004257,-0.4879447818,-0.1890367717,0.4240106642,-0.3093628883,0.0761560276,-0.0749146789,0.095616892,-0.2788098752,-0.1820272505,-0.1637149751,-0.3036242127,0.642676115,-0.0823247358,-0.0753789768,-0.1154965758,-0.6797434092,0.3325339854,0.2211797684,-0.0848115087,0.1067164391,-0.1852946132,0.0687569454,0.3100571632,0.9494889379,-0.1932723969,0.1566554606,0.0814261511,-0.0248434823,-0.4435347021,0.099806577,-0.3279415369,0.2550869286,0.249542281,0.5071648955,-0.3575258851,-0.072344102,-0.0420534089,-0.0277328677,0.0347220786,-0.0264866166,-0.1733805835,-0.1495417058,-0.138031885,0.3645800352,0.4469464123,0.072685346,0.1356703341,-0.2094711065,-0.0988867208,0.0481703952,0.017599402,-0.1340299547,0.4707481861,0.0158333834,0.0505986102,0.2701371014,0.111734964,0.2206298858,0.490062803,-0.153430447,0.2074077874,-0.1815864742,0.0305913817,0.3123363554,0.4398168921,0.0099648293,0.0579147972,0.381275177,0.3522175848,-0.0515045896,-0.2942019403,0.1262360066,0.0464040264,-0.5417702198,0.0036563263,0.115878664,-0.0995823666,0.0849705786,0.6082314849,0.5790846348,-0.3638046384,0.3609662056,0.0432918593,0.8923068643,-0.0148932636,0.1076276153,0.1486807168,0.0709492192,0.2498285174,0.5552585125,0.0943638608,0.0084218662,0.0974387005,-0.0407663956,-0.008215704,0.1668746769,0.3400380611,0.1027865037,0.2708595991,0.0819577202,0.1208938435,-0.150346145,-0.2895895541,0.2877454758,-0.33018139,-0.344931066,-0.0755386129,0.0785713568,0.2657673061,-0.0579297617,-0.1058492661,-0.0309242699,-0.4037671387,-0.1928453594,-0.1852022856,-0.4449900985,0.0211743154,0.2638934255,-0.2862427831,-0.0328586698,0.05158443,-0.4806175828,0.1447317004,-0.1111120284,-0.1330832839,0.3892788291,0.1719237566,0.0218227748,-0.152192831,0.1726180464,0.1998168081,-0.1803179681,0.0575537086,-0.4518596828,-0.172994867,0.0644659325,0.4083662331,-0.2023411393,-0.222900331,-0.3968209326,-0.1921593547,0.1015419289,0.0394879319,-0.0883383155,0.2670643032,-0.0757657439,0.3907501996,-0.192009002,-0.4534187615,-0.0437421575,0.5255671144,0.0346293263,-0.1923187524,0.412570715,0.1591852754,-0.1644421071,-0.0695428327,0.1383889616,0.3426440954,-0.6278496981,0.5433270335,0.2498693764,0.0369901843,0.137832284,0.4185103774,0.1368647516,0.2798647583,-0.3155887127,-0.1867186725,-0.2580824792,0.4086378217,0.1035424992,0.0359807424,-0.142046541,-0.2367399484,0.328837961,-0.1956318468,-0.1520757377,0.0660050362,-0.0865769833,0.3288319707,0.2184335589,0.2640419304,0.3569841981,-0.081811808,-0.1028169468,0.131531179,-0.279350847,-0.0736056119,-0.0467084982,0.176355049,0.1042599976,-0.220398739,0.1627230346,-0.4074191153,-0.0460065827,-0.4062663615,0.261949718,0.490960151,-0.0040139835,-0.2378341705,0.3719460666,0.0600784533,-0.0297455657,0.2064462304,-0.1776874214,-0.0705968589,-0.067778334,0.3043535054,0.1634591073,0.0337509252,0.0829738677,0.0720483661,0.014262286,0.0030940275,0.4227865934,-0.20461151,-0.2124462128,0.3322144449,0.2543900013,0.0184191465,-0.0208611134,-0.1518290639,0.2098122537,0.0060319579,-0.1523621529,0.0288437009,0.1550513953,0.3418107033,0.2605259717,0.3646138608,-0.1685419381,-0.2591929734,0.0174677297,0.1886160076,0.0211036801,0.0309402049,0.1880581975,0.3906849027,0.333427757,-0.0989431292,0.3222150207,0.0788699687,0.2825941443,0.2901285887,0.2377108186,0.2660039365,0.0139707457,0.0864454508,-0.2048538476,-0.5083088875,0.0377854444,0.0890346915,-0.0027337347,0.3360840082,0.1234634519,0.1538118124,-0.1924221218,-0.2750574648,0.0418601818,0.0021718226,-0.3454618752,-0.0618339814,-0.3136749864,-0.2561423182,0.4150924981,-0.1609165072,0.0627365261,-0.3454859853,0.3650220633,-0.077607587,-0.0953704789,-0.1441647857,-0.3677257597,0.0240027625,0.2292592824,-0.1103788093,0.1378632039,-0.0913239047,-0.0463289171,0.0504921414,0.5798316002,0.5111631751,0.3985682726,-0.0246103872,0.4008245468,0.130157128,-0.2436257899,-0.2861928642,0.3631888032,-0.0595287718,-0.0250939578,0.2001091689,-0.0104520181,-0.0469259843,-0.0630542859,-0.0026363027,0.1076258346,0.1635151505,-0.1908553392,-0.197421819,-0.2362531722,-0.2687476575,-0.0491944067,-0.0817951337,0.0160535164,0.29386428,-0.0996417478,0.0463557616,0.0778610259,-0.0013327353,-0.0415194184,0.2583118975,0.4536593854,0.0035883174,-0.4841513932,0.1772225052,-0.3718905747,-0.0301732477,-0.4928691983,0.2783645988,0.0389187187,0.197258383,0.1198171675,-0.1226420105,0.3183076978,-0.050173346,-0.0438348092,0.3545496762,0.0648137182,-0.2790630162,-0.1928181499,-0.0135354837,0.0101545881,-0.3842709064,0.1062459424,-0.11351078,-0.0753601417,-0.105609633,0.1699269265,-0.1576716304,0.0275927205,0.1774147898,0.3360162675,-0.0418955907,0.0294316802,-0.0177058894,-0.3504039645,0.1464190781,-0.126489982,0.1394371688,0.359600842,0.4612443745,-0.3284623921,-0.2027059644,-0.0240753405,0.0423464887,0.0046442109,0.0361618474,-0.1818929613,0.1435846388,-0.3938452899,-0.0358402953,-0.1220736802,0.0222712085,0.0860232413,0.0270857383,-0.252587229,-0.3549436331,0.3827648461,-0.6531078815,-0.2620070577,-0.0693184882,0.3420590162,0.2115111947,-0.2256407142,-0.3928543925,-0.1719201207,0.2626016438,-0.0693581328,-0.0616543554,-0.1847913414,-0.5037385225,-0.0344289057,-0.1576237828,0.1152830869,0.0357203335,0.0066587133,0.231196329,-0.3098658323]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2644","title":"Batched `map` not allowed to return 0 items","comments":"Sounds good! Do you want me to propose a PR? I'm quite busy right now, but if it's not too urgent I could take a look next week.","body":"## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":28,"text":"Batched `map` not allowed to return 0 items\n## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n\nSounds good! Do you want me to propose a PR? I'm quite busy right now, but if it's not too urgent I could take a look next week.","embeddings":[-0.2328359336,-0.3903888166,-0.0444566682,0.1644899547,-0.1085444242,-0.0687690005,0.1163394973,0.4056320786,0.6400417089,0.1178812236,-0.0565718673,0.2460853755,-0.3954412043,0.1043543741,-0.1629210711,0.1608026177,-0.0449331626,0.1405789405,-0.1691175699,-0.1914925724,-0.3362929821,0.163199544,-0.3979255855,-0.0501046143,-0.1487712115,-0.2919759154,0.286480099,-0.0362105146,-0.162814334,-0.2454052418,0.206091553,-0.2328298539,-0.2091356516,0.5783210993,-0.0001299276,-0.0231365338,0.1505937427,-0.1851885766,-0.0897614583,-0.3029606342,-0.1908181012,-0.1715300977,-0.1157183424,-0.2824648917,0.1819460839,-0.2716506422,-0.1956346184,-0.4482398331,0.1071964726,0.2587007582,0.043321602,0.4188180268,0.003066103,0.1764480174,0.3312488496,0.2664009035,-0.0753516629,0.1288110614,0.7205767035,-0.2881007195,0.0226270352,0.3385902047,-0.3921834528,0.0709065422,0.1367558688,-0.0309035499,0.230168879,-0.5515586734,0.2697506845,0.3675674498,-0.0899998844,0.0607856587,-0.433337301,-0.6687820554,-0.0403036699,-0.0954985172,-0.0388978831,0.2810007632,-0.4210395217,-0.0564594232,-0.4846186638,0.1217950135,0.0078949472,0.1618547142,0.0306755304,0.4589543343,-0.0930409953,0.2003241777,0.3451099396,-0.1618324965,-0.0994021893,-0.1672711968,-0.2236276567,0.5847230554,-0.0366987288,-0.1459876001,0.3893203735,0.2197681516,0.2230358422,-0.3132426441,-0.050306201,-0.0742755681,0.2646799386,-0.0234315637,0.4159452617,0.0793761313,0.2686936855,0.3493200243,-0.005052017,-0.0590938851,0.0676903278,0.1131693944,0.113442108,-0.0890844166,-0.0096516423,-0.0764087662,0.2567855418,-0.0787523314,-0.1218986362,0.2341007143,-0.4100169539,0.1369067729,0.0620478503,0.1250830591,0.1700579822,-0.1776706576,0.1202128828,0.1126287505,-0.1570285559,0.1897115409,-0.0772467405,-0.1316168755,-0.192502588,-0.3208249807,0.1964786202,-0.3342862427,0.2418166101,-0.1585832834,0.0198275335,-0.0615327619,0.0957399309,-0.2645617723,0.7512272596,0.4937881231,-0.119511351,0.2913580537,0.3966238201,-0.2185574174,-0.1752834171,0.4062903821,-0.3826135099,-0.2777348757,0.2100417018,-0.0401864871,-0.2380758375,0.346570462,-0.5073054433,0.2185845971,0.1782458276,0.0822512209,0.2914729118,-0.2697472274,-0.105394572,-0.2360114455,0.0425621271,0.6938397884,-0.63877213,-0.049924884,0.1101892963,0.1013790518,-0.0613662116,0.3478800058,-0.0454604253,0.3481105566,-0.4107360542,0.3059869707,0.0873124897,-0.1264710426,-0.3275630772,0.3134204149,-0.0358151942,0.2356395572,-0.0210545417,-0.0642482191,0.5519376397,-0.1873890162,0.3676523268,-0.1700588763,-0.279956907,0.1041627601,-0.0503817089,-0.039843861,0.0603797175,-0.2420822233,0.19492428,0.0165877249,0.0390782356,-0.5488622189,0.1286391169,-0.1508790404,0.3383628428,0.2401701212,0.262108326,0.1233314499,0.0182514824,-0.2563515604,-0.4713618159,0.3001765013,-0.4880333245,0.021248579,-0.4338338673,-0.1033039168,0.0309646595,0.1834742576,-0.0350200944,0.0330956019,-0.0749208331,-0.4975802898,0.1765710711,-0.0213433541,-0.2262366265,-0.2370698303,-0.1358649582,-0.0084387707,-0.3567225337,0.071359843,0.2501084507,-0.2921575904,-0.2542045712,0.2466300279,0.1149316132,-0.1176109165,0.117354244,0.1758526713,-0.0348161124,-0.1673572958,-0.1400127411,0.131253615,0.0117278975,-0.0458009467,0.101837188,0.1450134069,0.3444767892,-0.2378433347,-0.0694068223,0.5054207444,-0.1876149476,0.5359300375,-0.1796862483,0.0512384214,-0.1996322572,0.042606812,-0.2059215009,-0.3850508034,0.0762836039,-0.1577371508,0.0513374284,0.0319146588,0.0012563785,-0.1989199072,0.0856170282,-0.046962332,0.16477938,0.0144881429,-0.0630519465,0.2266895324,0.0446924791,-0.0142778121,0.3432699442,0.2191425711,-0.0887260064,-0.1521824449,0.250440836,0.0001223587,0.4139182568,0.18166776,0.0260531791,0.1550095528,-0.003640976,-0.0657203272,-0.2528341115,-0.1814302802,0.005715027,0.149533838,-0.3952980638,-0.0354201086,0.1101150513,-0.158504501,0.0689762086,-0.0765032619,-0.290856123,-0.3032285273,-0.0453336127,0.2601582408,-0.2722713947,-0.02400795,-0.4013136625,0.0116532287,0.1975094527,-0.2747197151,-0.2812999189,-0.1685465872,-0.0756110176,-0.0741591007,0.0731139854,-0.1339878589,0.3615631759,0.2521368861,-0.2612816691,-0.243771404,-0.0951137543,0.0065136496,-0.5370389223,0.2054035813,0.4100922644,0.3981708586,-0.2923846245,0.1470615864,0.2793858051,-0.1071430445,-0.2480682433,-0.0079056751,0.0223337524,0.153262645,-0.179215163,0.1474315226,0.0644616261,-0.3209299445,0.1781849116,-0.1891590655,0.2615512013,0.2194533795,0.2780077457,0.1538574994,-0.3693777621,-0.4285494387,0.1223512962,-0.1416769028,0.4460211396,0.0390229337,-0.1469308287,0.1844055504,0.0262182392,-0.1853193641,0.2405568659,-0.0966859683,0.0939638689,-0.0562738925,0.1234829947,0.190006569,0.3668004274,0.4985276759,0.0864884332,0.0460447706,-0.2846722603,-0.1089555398,0.0812640563,0.0222764555,-0.064514026,0.422529459,0.3217896223,0.0287001207,0.7504062057,0.3654765189,0.1272135675,0.0946453437,0.0725369975,0.3822363317,-0.1140064001,-0.379692018,-0.2082244754,0.1538844705,-0.0840781853,0.0325130187,0.0550553165,-0.0620342046,-0.0012852303,0.0619004257,-0.4879447818,-0.1890367717,0.4240106642,-0.3093628883,0.0761560276,-0.0749146789,0.095616892,-0.2788098752,-0.1820272505,-0.1637149751,-0.3036242127,0.642676115,-0.0823247358,-0.0753789768,-0.1154965758,-0.6797434092,0.3325339854,0.2211797684,-0.0848115087,0.1067164391,-0.1852946132,0.0687569454,0.3100571632,0.9494889379,-0.1932723969,0.1566554606,0.0814261511,-0.0248434823,-0.4435347021,0.099806577,-0.3279415369,0.2550869286,0.249542281,0.5071648955,-0.3575258851,-0.072344102,-0.0420534089,-0.0277328677,0.0347220786,-0.0264866166,-0.1733805835,-0.1495417058,-0.138031885,0.3645800352,0.4469464123,0.072685346,0.1356703341,-0.2094711065,-0.0988867208,0.0481703952,0.017599402,-0.1340299547,0.4707481861,0.0158333834,0.0505986102,0.2701371014,0.111734964,0.2206298858,0.490062803,-0.153430447,0.2074077874,-0.1815864742,0.0305913817,0.3123363554,0.4398168921,0.0099648293,0.0579147972,0.381275177,0.3522175848,-0.0515045896,-0.2942019403,0.1262360066,0.0464040264,-0.5417702198,0.0036563263,0.115878664,-0.0995823666,0.0849705786,0.6082314849,0.5790846348,-0.3638046384,0.3609662056,0.0432918593,0.8923068643,-0.0148932636,0.1076276153,0.1486807168,0.0709492192,0.2498285174,0.5552585125,0.0943638608,0.0084218662,0.0974387005,-0.0407663956,-0.008215704,0.1668746769,0.3400380611,0.1027865037,0.2708595991,0.0819577202,0.1208938435,-0.150346145,-0.2895895541,0.2877454758,-0.33018139,-0.344931066,-0.0755386129,0.0785713568,0.2657673061,-0.0579297617,-0.1058492661,-0.0309242699,-0.4037671387,-0.1928453594,-0.1852022856,-0.4449900985,0.0211743154,0.2638934255,-0.2862427831,-0.0328586698,0.05158443,-0.4806175828,0.1447317004,-0.1111120284,-0.1330832839,0.3892788291,0.1719237566,0.0218227748,-0.152192831,0.1726180464,0.1998168081,-0.1803179681,0.0575537086,-0.4518596828,-0.172994867,0.0644659325,0.4083662331,-0.2023411393,-0.222900331,-0.3968209326,-0.1921593547,0.1015419289,0.0394879319,-0.0883383155,0.2670643032,-0.0757657439,0.3907501996,-0.192009002,-0.4534187615,-0.0437421575,0.5255671144,0.0346293263,-0.1923187524,0.412570715,0.1591852754,-0.1644421071,-0.0695428327,0.1383889616,0.3426440954,-0.6278496981,0.5433270335,0.2498693764,0.0369901843,0.137832284,0.4185103774,0.1368647516,0.2798647583,-0.3155887127,-0.1867186725,-0.2580824792,0.4086378217,0.1035424992,0.0359807424,-0.142046541,-0.2367399484,0.328837961,-0.1956318468,-0.1520757377,0.0660050362,-0.0865769833,0.3288319707,0.2184335589,0.2640419304,0.3569841981,-0.081811808,-0.1028169468,0.131531179,-0.279350847,-0.0736056119,-0.0467084982,0.176355049,0.1042599976,-0.220398739,0.1627230346,-0.4074191153,-0.0460065827,-0.4062663615,0.261949718,0.490960151,-0.0040139835,-0.2378341705,0.3719460666,0.0600784533,-0.0297455657,0.2064462304,-0.1776874214,-0.0705968589,-0.067778334,0.3043535054,0.1634591073,0.0337509252,0.0829738677,0.0720483661,0.014262286,0.0030940275,0.4227865934,-0.20461151,-0.2124462128,0.3322144449,0.2543900013,0.0184191465,-0.0208611134,-0.1518290639,0.2098122537,0.0060319579,-0.1523621529,0.0288437009,0.1550513953,0.3418107033,0.2605259717,0.3646138608,-0.1685419381,-0.2591929734,0.0174677297,0.1886160076,0.0211036801,0.0309402049,0.1880581975,0.3906849027,0.333427757,-0.0989431292,0.3222150207,0.0788699687,0.2825941443,0.2901285887,0.2377108186,0.2660039365,0.0139707457,0.0864454508,-0.2048538476,-0.5083088875,0.0377854444,0.0890346915,-0.0027337347,0.3360840082,0.1234634519,0.1538118124,-0.1924221218,-0.2750574648,0.0418601818,0.0021718226,-0.3454618752,-0.0618339814,-0.3136749864,-0.2561423182,0.4150924981,-0.1609165072,0.0627365261,-0.3454859853,0.3650220633,-0.077607587,-0.0953704789,-0.1441647857,-0.3677257597,0.0240027625,0.2292592824,-0.1103788093,0.1378632039,-0.0913239047,-0.0463289171,0.0504921414,0.5798316002,0.5111631751,0.3985682726,-0.0246103872,0.4008245468,0.130157128,-0.2436257899,-0.2861928642,0.3631888032,-0.0595287718,-0.0250939578,0.2001091689,-0.0104520181,-0.0469259843,-0.0630542859,-0.0026363027,0.1076258346,0.1635151505,-0.1908553392,-0.197421819,-0.2362531722,-0.2687476575,-0.0491944067,-0.0817951337,0.0160535164,0.29386428,-0.0996417478,0.0463557616,0.0778610259,-0.0013327353,-0.0415194184,0.2583118975,0.4536593854,0.0035883174,-0.4841513932,0.1772225052,-0.3718905747,-0.0301732477,-0.4928691983,0.2783645988,0.0389187187,0.197258383,0.1198171675,-0.1226420105,0.3183076978,-0.050173346,-0.0438348092,0.3545496762,0.0648137182,-0.2790630162,-0.1928181499,-0.0135354837,0.0101545881,-0.3842709064,0.1062459424,-0.11351078,-0.0753601417,-0.105609633,0.1699269265,-0.1576716304,0.0275927205,0.1774147898,0.3360162675,-0.0418955907,0.0294316802,-0.0177058894,-0.3504039645,0.1464190781,-0.126489982,0.1394371688,0.359600842,0.4612443745,-0.3284623921,-0.2027059644,-0.0240753405,0.0423464887,0.0046442109,0.0361618474,-0.1818929613,0.1435846388,-0.3938452899,-0.0358402953,-0.1220736802,0.0222712085,0.0860232413,0.0270857383,-0.252587229,-0.3549436331,0.3827648461,-0.6531078815,-0.2620070577,-0.0693184882,0.3420590162,0.2115111947,-0.2256407142,-0.3928543925,-0.1719201207,0.2626016438,-0.0693581328,-0.0616543554,-0.1847913414,-0.5037385225,-0.0344289057,-0.1576237828,0.1152830869,0.0357203335,0.0066587133,0.231196329,-0.3098658323]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2644","title":"Batched `map` not allowed to return 0 items","comments":"Sure if you're interested feel free to open a PR :)\r\n\r\nYou can also ping me anytime if you have questions or if I can help !","body":"## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":27,"text":"Batched `map` not allowed to return 0 items\n## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n\nSure if you're interested feel free to open a PR :)\r\n\r\nYou can also ping me anytime if you have questions or if I can help !","embeddings":[-0.2328359336,-0.3903888166,-0.0444566682,0.1644899547,-0.1085444242,-0.0687690005,0.1163394973,0.4056320786,0.6400417089,0.1178812236,-0.0565718673,0.2460853755,-0.3954412043,0.1043543741,-0.1629210711,0.1608026177,-0.0449331626,0.1405789405,-0.1691175699,-0.1914925724,-0.3362929821,0.163199544,-0.3979255855,-0.0501046143,-0.1487712115,-0.2919759154,0.286480099,-0.0362105146,-0.162814334,-0.2454052418,0.206091553,-0.2328298539,-0.2091356516,0.5783210993,-0.0001299276,-0.0231365338,0.1505937427,-0.1851885766,-0.0897614583,-0.3029606342,-0.1908181012,-0.1715300977,-0.1157183424,-0.2824648917,0.1819460839,-0.2716506422,-0.1956346184,-0.4482398331,0.1071964726,0.2587007582,0.043321602,0.4188180268,0.003066103,0.1764480174,0.3312488496,0.2664009035,-0.0753516629,0.1288110614,0.7205767035,-0.2881007195,0.0226270352,0.3385902047,-0.3921834528,0.0709065422,0.1367558688,-0.0309035499,0.230168879,-0.5515586734,0.2697506845,0.3675674498,-0.0899998844,0.0607856587,-0.433337301,-0.6687820554,-0.0403036699,-0.0954985172,-0.0388978831,0.2810007632,-0.4210395217,-0.0564594232,-0.4846186638,0.1217950135,0.0078949472,0.1618547142,0.0306755304,0.4589543343,-0.0930409953,0.2003241777,0.3451099396,-0.1618324965,-0.0994021893,-0.1672711968,-0.2236276567,0.5847230554,-0.0366987288,-0.1459876001,0.3893203735,0.2197681516,0.2230358422,-0.3132426441,-0.050306201,-0.0742755681,0.2646799386,-0.0234315637,0.4159452617,0.0793761313,0.2686936855,0.3493200243,-0.005052017,-0.0590938851,0.0676903278,0.1131693944,0.113442108,-0.0890844166,-0.0096516423,-0.0764087662,0.2567855418,-0.0787523314,-0.1218986362,0.2341007143,-0.4100169539,0.1369067729,0.0620478503,0.1250830591,0.1700579822,-0.1776706576,0.1202128828,0.1126287505,-0.1570285559,0.1897115409,-0.0772467405,-0.1316168755,-0.192502588,-0.3208249807,0.1964786202,-0.3342862427,0.2418166101,-0.1585832834,0.0198275335,-0.0615327619,0.0957399309,-0.2645617723,0.7512272596,0.4937881231,-0.119511351,0.2913580537,0.3966238201,-0.2185574174,-0.1752834171,0.4062903821,-0.3826135099,-0.2777348757,0.2100417018,-0.0401864871,-0.2380758375,0.346570462,-0.5073054433,0.2185845971,0.1782458276,0.0822512209,0.2914729118,-0.2697472274,-0.105394572,-0.2360114455,0.0425621271,0.6938397884,-0.63877213,-0.049924884,0.1101892963,0.1013790518,-0.0613662116,0.3478800058,-0.0454604253,0.3481105566,-0.4107360542,0.3059869707,0.0873124897,-0.1264710426,-0.3275630772,0.3134204149,-0.0358151942,0.2356395572,-0.0210545417,-0.0642482191,0.5519376397,-0.1873890162,0.3676523268,-0.1700588763,-0.279956907,0.1041627601,-0.0503817089,-0.039843861,0.0603797175,-0.2420822233,0.19492428,0.0165877249,0.0390782356,-0.5488622189,0.1286391169,-0.1508790404,0.3383628428,0.2401701212,0.262108326,0.1233314499,0.0182514824,-0.2563515604,-0.4713618159,0.3001765013,-0.4880333245,0.021248579,-0.4338338673,-0.1033039168,0.0309646595,0.1834742576,-0.0350200944,0.0330956019,-0.0749208331,-0.4975802898,0.1765710711,-0.0213433541,-0.2262366265,-0.2370698303,-0.1358649582,-0.0084387707,-0.3567225337,0.071359843,0.2501084507,-0.2921575904,-0.2542045712,0.2466300279,0.1149316132,-0.1176109165,0.117354244,0.1758526713,-0.0348161124,-0.1673572958,-0.1400127411,0.131253615,0.0117278975,-0.0458009467,0.101837188,0.1450134069,0.3444767892,-0.2378433347,-0.0694068223,0.5054207444,-0.1876149476,0.5359300375,-0.1796862483,0.0512384214,-0.1996322572,0.042606812,-0.2059215009,-0.3850508034,0.0762836039,-0.1577371508,0.0513374284,0.0319146588,0.0012563785,-0.1989199072,0.0856170282,-0.046962332,0.16477938,0.0144881429,-0.0630519465,0.2266895324,0.0446924791,-0.0142778121,0.3432699442,0.2191425711,-0.0887260064,-0.1521824449,0.250440836,0.0001223587,0.4139182568,0.18166776,0.0260531791,0.1550095528,-0.003640976,-0.0657203272,-0.2528341115,-0.1814302802,0.005715027,0.149533838,-0.3952980638,-0.0354201086,0.1101150513,-0.158504501,0.0689762086,-0.0765032619,-0.290856123,-0.3032285273,-0.0453336127,0.2601582408,-0.2722713947,-0.02400795,-0.4013136625,0.0116532287,0.1975094527,-0.2747197151,-0.2812999189,-0.1685465872,-0.0756110176,-0.0741591007,0.0731139854,-0.1339878589,0.3615631759,0.2521368861,-0.2612816691,-0.243771404,-0.0951137543,0.0065136496,-0.5370389223,0.2054035813,0.4100922644,0.3981708586,-0.2923846245,0.1470615864,0.2793858051,-0.1071430445,-0.2480682433,-0.0079056751,0.0223337524,0.153262645,-0.179215163,0.1474315226,0.0644616261,-0.3209299445,0.1781849116,-0.1891590655,0.2615512013,0.2194533795,0.2780077457,0.1538574994,-0.3693777621,-0.4285494387,0.1223512962,-0.1416769028,0.4460211396,0.0390229337,-0.1469308287,0.1844055504,0.0262182392,-0.1853193641,0.2405568659,-0.0966859683,0.0939638689,-0.0562738925,0.1234829947,0.190006569,0.3668004274,0.4985276759,0.0864884332,0.0460447706,-0.2846722603,-0.1089555398,0.0812640563,0.0222764555,-0.064514026,0.422529459,0.3217896223,0.0287001207,0.7504062057,0.3654765189,0.1272135675,0.0946453437,0.0725369975,0.3822363317,-0.1140064001,-0.379692018,-0.2082244754,0.1538844705,-0.0840781853,0.0325130187,0.0550553165,-0.0620342046,-0.0012852303,0.0619004257,-0.4879447818,-0.1890367717,0.4240106642,-0.3093628883,0.0761560276,-0.0749146789,0.095616892,-0.2788098752,-0.1820272505,-0.1637149751,-0.3036242127,0.642676115,-0.0823247358,-0.0753789768,-0.1154965758,-0.6797434092,0.3325339854,0.2211797684,-0.0848115087,0.1067164391,-0.1852946132,0.0687569454,0.3100571632,0.9494889379,-0.1932723969,0.1566554606,0.0814261511,-0.0248434823,-0.4435347021,0.099806577,-0.3279415369,0.2550869286,0.249542281,0.5071648955,-0.3575258851,-0.072344102,-0.0420534089,-0.0277328677,0.0347220786,-0.0264866166,-0.1733805835,-0.1495417058,-0.138031885,0.3645800352,0.4469464123,0.072685346,0.1356703341,-0.2094711065,-0.0988867208,0.0481703952,0.017599402,-0.1340299547,0.4707481861,0.0158333834,0.0505986102,0.2701371014,0.111734964,0.2206298858,0.490062803,-0.153430447,0.2074077874,-0.1815864742,0.0305913817,0.3123363554,0.4398168921,0.0099648293,0.0579147972,0.381275177,0.3522175848,-0.0515045896,-0.2942019403,0.1262360066,0.0464040264,-0.5417702198,0.0036563263,0.115878664,-0.0995823666,0.0849705786,0.6082314849,0.5790846348,-0.3638046384,0.3609662056,0.0432918593,0.8923068643,-0.0148932636,0.1076276153,0.1486807168,0.0709492192,0.2498285174,0.5552585125,0.0943638608,0.0084218662,0.0974387005,-0.0407663956,-0.008215704,0.1668746769,0.3400380611,0.1027865037,0.2708595991,0.0819577202,0.1208938435,-0.150346145,-0.2895895541,0.2877454758,-0.33018139,-0.344931066,-0.0755386129,0.0785713568,0.2657673061,-0.0579297617,-0.1058492661,-0.0309242699,-0.4037671387,-0.1928453594,-0.1852022856,-0.4449900985,0.0211743154,0.2638934255,-0.2862427831,-0.0328586698,0.05158443,-0.4806175828,0.1447317004,-0.1111120284,-0.1330832839,0.3892788291,0.1719237566,0.0218227748,-0.152192831,0.1726180464,0.1998168081,-0.1803179681,0.0575537086,-0.4518596828,-0.172994867,0.0644659325,0.4083662331,-0.2023411393,-0.222900331,-0.3968209326,-0.1921593547,0.1015419289,0.0394879319,-0.0883383155,0.2670643032,-0.0757657439,0.3907501996,-0.192009002,-0.4534187615,-0.0437421575,0.5255671144,0.0346293263,-0.1923187524,0.412570715,0.1591852754,-0.1644421071,-0.0695428327,0.1383889616,0.3426440954,-0.6278496981,0.5433270335,0.2498693764,0.0369901843,0.137832284,0.4185103774,0.1368647516,0.2798647583,-0.3155887127,-0.1867186725,-0.2580824792,0.4086378217,0.1035424992,0.0359807424,-0.142046541,-0.2367399484,0.328837961,-0.1956318468,-0.1520757377,0.0660050362,-0.0865769833,0.3288319707,0.2184335589,0.2640419304,0.3569841981,-0.081811808,-0.1028169468,0.131531179,-0.279350847,-0.0736056119,-0.0467084982,0.176355049,0.1042599976,-0.220398739,0.1627230346,-0.4074191153,-0.0460065827,-0.4062663615,0.261949718,0.490960151,-0.0040139835,-0.2378341705,0.3719460666,0.0600784533,-0.0297455657,0.2064462304,-0.1776874214,-0.0705968589,-0.067778334,0.3043535054,0.1634591073,0.0337509252,0.0829738677,0.0720483661,0.014262286,0.0030940275,0.4227865934,-0.20461151,-0.2124462128,0.3322144449,0.2543900013,0.0184191465,-0.0208611134,-0.1518290639,0.2098122537,0.0060319579,-0.1523621529,0.0288437009,0.1550513953,0.3418107033,0.2605259717,0.3646138608,-0.1685419381,-0.2591929734,0.0174677297,0.1886160076,0.0211036801,0.0309402049,0.1880581975,0.3906849027,0.333427757,-0.0989431292,0.3222150207,0.0788699687,0.2825941443,0.2901285887,0.2377108186,0.2660039365,0.0139707457,0.0864454508,-0.2048538476,-0.5083088875,0.0377854444,0.0890346915,-0.0027337347,0.3360840082,0.1234634519,0.1538118124,-0.1924221218,-0.2750574648,0.0418601818,0.0021718226,-0.3454618752,-0.0618339814,-0.3136749864,-0.2561423182,0.4150924981,-0.1609165072,0.0627365261,-0.3454859853,0.3650220633,-0.077607587,-0.0953704789,-0.1441647857,-0.3677257597,0.0240027625,0.2292592824,-0.1103788093,0.1378632039,-0.0913239047,-0.0463289171,0.0504921414,0.5798316002,0.5111631751,0.3985682726,-0.0246103872,0.4008245468,0.130157128,-0.2436257899,-0.2861928642,0.3631888032,-0.0595287718,-0.0250939578,0.2001091689,-0.0104520181,-0.0469259843,-0.0630542859,-0.0026363027,0.1076258346,0.1635151505,-0.1908553392,-0.197421819,-0.2362531722,-0.2687476575,-0.0491944067,-0.0817951337,0.0160535164,0.29386428,-0.0996417478,0.0463557616,0.0778610259,-0.0013327353,-0.0415194184,0.2583118975,0.4536593854,0.0035883174,-0.4841513932,0.1772225052,-0.3718905747,-0.0301732477,-0.4928691983,0.2783645988,0.0389187187,0.197258383,0.1198171675,-0.1226420105,0.3183076978,-0.050173346,-0.0438348092,0.3545496762,0.0648137182,-0.2790630162,-0.1928181499,-0.0135354837,0.0101545881,-0.3842709064,0.1062459424,-0.11351078,-0.0753601417,-0.105609633,0.1699269265,-0.1576716304,0.0275927205,0.1774147898,0.3360162675,-0.0418955907,0.0294316802,-0.0177058894,-0.3504039645,0.1464190781,-0.126489982,0.1394371688,0.359600842,0.4612443745,-0.3284623921,-0.2027059644,-0.0240753405,0.0423464887,0.0046442109,0.0361618474,-0.1818929613,0.1435846388,-0.3938452899,-0.0358402953,-0.1220736802,0.0222712085,0.0860232413,0.0270857383,-0.252587229,-0.3549436331,0.3827648461,-0.6531078815,-0.2620070577,-0.0693184882,0.3420590162,0.2115111947,-0.2256407142,-0.3928543925,-0.1719201207,0.2626016438,-0.0693581328,-0.0616543554,-0.1847913414,-0.5037385225,-0.0344289057,-0.1576237828,0.1152830869,0.0357203335,0.0066587133,0.231196329,-0.3098658323]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2644","title":"Batched `map` not allowed to return 0 items","comments":"Sorry to ping you, @lhoestq, did you have a chance to take a look at the proposed PR? Thank you!","body":"## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":20,"text":"Batched `map` not allowed to return 0 items\n## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n\nSorry to ping you, @lhoestq, did you have a chance to take a look at the proposed PR? Thank you!","embeddings":[-0.2328359336,-0.3903888166,-0.0444566682,0.1644899547,-0.1085444242,-0.0687690005,0.1163394973,0.4056320786,0.6400417089,0.1178812236,-0.0565718673,0.2460853755,-0.3954412043,0.1043543741,-0.1629210711,0.1608026177,-0.0449331626,0.1405789405,-0.1691175699,-0.1914925724,-0.3362929821,0.163199544,-0.3979255855,-0.0501046143,-0.1487712115,-0.2919759154,0.286480099,-0.0362105146,-0.162814334,-0.2454052418,0.206091553,-0.2328298539,-0.2091356516,0.5783210993,-0.0001299276,-0.0231365338,0.1505937427,-0.1851885766,-0.0897614583,-0.3029606342,-0.1908181012,-0.1715300977,-0.1157183424,-0.2824648917,0.1819460839,-0.2716506422,-0.1956346184,-0.4482398331,0.1071964726,0.2587007582,0.043321602,0.4188180268,0.003066103,0.1764480174,0.3312488496,0.2664009035,-0.0753516629,0.1288110614,0.7205767035,-0.2881007195,0.0226270352,0.3385902047,-0.3921834528,0.0709065422,0.1367558688,-0.0309035499,0.230168879,-0.5515586734,0.2697506845,0.3675674498,-0.0899998844,0.0607856587,-0.433337301,-0.6687820554,-0.0403036699,-0.0954985172,-0.0388978831,0.2810007632,-0.4210395217,-0.0564594232,-0.4846186638,0.1217950135,0.0078949472,0.1618547142,0.0306755304,0.4589543343,-0.0930409953,0.2003241777,0.3451099396,-0.1618324965,-0.0994021893,-0.1672711968,-0.2236276567,0.5847230554,-0.0366987288,-0.1459876001,0.3893203735,0.2197681516,0.2230358422,-0.3132426441,-0.050306201,-0.0742755681,0.2646799386,-0.0234315637,0.4159452617,0.0793761313,0.2686936855,0.3493200243,-0.005052017,-0.0590938851,0.0676903278,0.1131693944,0.113442108,-0.0890844166,-0.0096516423,-0.0764087662,0.2567855418,-0.0787523314,-0.1218986362,0.2341007143,-0.4100169539,0.1369067729,0.0620478503,0.1250830591,0.1700579822,-0.1776706576,0.1202128828,0.1126287505,-0.1570285559,0.1897115409,-0.0772467405,-0.1316168755,-0.192502588,-0.3208249807,0.1964786202,-0.3342862427,0.2418166101,-0.1585832834,0.0198275335,-0.0615327619,0.0957399309,-0.2645617723,0.7512272596,0.4937881231,-0.119511351,0.2913580537,0.3966238201,-0.2185574174,-0.1752834171,0.4062903821,-0.3826135099,-0.2777348757,0.2100417018,-0.0401864871,-0.2380758375,0.346570462,-0.5073054433,0.2185845971,0.1782458276,0.0822512209,0.2914729118,-0.2697472274,-0.105394572,-0.2360114455,0.0425621271,0.6938397884,-0.63877213,-0.049924884,0.1101892963,0.1013790518,-0.0613662116,0.3478800058,-0.0454604253,0.3481105566,-0.4107360542,0.3059869707,0.0873124897,-0.1264710426,-0.3275630772,0.3134204149,-0.0358151942,0.2356395572,-0.0210545417,-0.0642482191,0.5519376397,-0.1873890162,0.3676523268,-0.1700588763,-0.279956907,0.1041627601,-0.0503817089,-0.039843861,0.0603797175,-0.2420822233,0.19492428,0.0165877249,0.0390782356,-0.5488622189,0.1286391169,-0.1508790404,0.3383628428,0.2401701212,0.262108326,0.1233314499,0.0182514824,-0.2563515604,-0.4713618159,0.3001765013,-0.4880333245,0.021248579,-0.4338338673,-0.1033039168,0.0309646595,0.1834742576,-0.0350200944,0.0330956019,-0.0749208331,-0.4975802898,0.1765710711,-0.0213433541,-0.2262366265,-0.2370698303,-0.1358649582,-0.0084387707,-0.3567225337,0.071359843,0.2501084507,-0.2921575904,-0.2542045712,0.2466300279,0.1149316132,-0.1176109165,0.117354244,0.1758526713,-0.0348161124,-0.1673572958,-0.1400127411,0.131253615,0.0117278975,-0.0458009467,0.101837188,0.1450134069,0.3444767892,-0.2378433347,-0.0694068223,0.5054207444,-0.1876149476,0.5359300375,-0.1796862483,0.0512384214,-0.1996322572,0.042606812,-0.2059215009,-0.3850508034,0.0762836039,-0.1577371508,0.0513374284,0.0319146588,0.0012563785,-0.1989199072,0.0856170282,-0.046962332,0.16477938,0.0144881429,-0.0630519465,0.2266895324,0.0446924791,-0.0142778121,0.3432699442,0.2191425711,-0.0887260064,-0.1521824449,0.250440836,0.0001223587,0.4139182568,0.18166776,0.0260531791,0.1550095528,-0.003640976,-0.0657203272,-0.2528341115,-0.1814302802,0.005715027,0.149533838,-0.3952980638,-0.0354201086,0.1101150513,-0.158504501,0.0689762086,-0.0765032619,-0.290856123,-0.3032285273,-0.0453336127,0.2601582408,-0.2722713947,-0.02400795,-0.4013136625,0.0116532287,0.1975094527,-0.2747197151,-0.2812999189,-0.1685465872,-0.0756110176,-0.0741591007,0.0731139854,-0.1339878589,0.3615631759,0.2521368861,-0.2612816691,-0.243771404,-0.0951137543,0.0065136496,-0.5370389223,0.2054035813,0.4100922644,0.3981708586,-0.2923846245,0.1470615864,0.2793858051,-0.1071430445,-0.2480682433,-0.0079056751,0.0223337524,0.153262645,-0.179215163,0.1474315226,0.0644616261,-0.3209299445,0.1781849116,-0.1891590655,0.2615512013,0.2194533795,0.2780077457,0.1538574994,-0.3693777621,-0.4285494387,0.1223512962,-0.1416769028,0.4460211396,0.0390229337,-0.1469308287,0.1844055504,0.0262182392,-0.1853193641,0.2405568659,-0.0966859683,0.0939638689,-0.0562738925,0.1234829947,0.190006569,0.3668004274,0.4985276759,0.0864884332,0.0460447706,-0.2846722603,-0.1089555398,0.0812640563,0.0222764555,-0.064514026,0.422529459,0.3217896223,0.0287001207,0.7504062057,0.3654765189,0.1272135675,0.0946453437,0.0725369975,0.3822363317,-0.1140064001,-0.379692018,-0.2082244754,0.1538844705,-0.0840781853,0.0325130187,0.0550553165,-0.0620342046,-0.0012852303,0.0619004257,-0.4879447818,-0.1890367717,0.4240106642,-0.3093628883,0.0761560276,-0.0749146789,0.095616892,-0.2788098752,-0.1820272505,-0.1637149751,-0.3036242127,0.642676115,-0.0823247358,-0.0753789768,-0.1154965758,-0.6797434092,0.3325339854,0.2211797684,-0.0848115087,0.1067164391,-0.1852946132,0.0687569454,0.3100571632,0.9494889379,-0.1932723969,0.1566554606,0.0814261511,-0.0248434823,-0.4435347021,0.099806577,-0.3279415369,0.2550869286,0.249542281,0.5071648955,-0.3575258851,-0.072344102,-0.0420534089,-0.0277328677,0.0347220786,-0.0264866166,-0.1733805835,-0.1495417058,-0.138031885,0.3645800352,0.4469464123,0.072685346,0.1356703341,-0.2094711065,-0.0988867208,0.0481703952,0.017599402,-0.1340299547,0.4707481861,0.0158333834,0.0505986102,0.2701371014,0.111734964,0.2206298858,0.490062803,-0.153430447,0.2074077874,-0.1815864742,0.0305913817,0.3123363554,0.4398168921,0.0099648293,0.0579147972,0.381275177,0.3522175848,-0.0515045896,-0.2942019403,0.1262360066,0.0464040264,-0.5417702198,0.0036563263,0.115878664,-0.0995823666,0.0849705786,0.6082314849,0.5790846348,-0.3638046384,0.3609662056,0.0432918593,0.8923068643,-0.0148932636,0.1076276153,0.1486807168,0.0709492192,0.2498285174,0.5552585125,0.0943638608,0.0084218662,0.0974387005,-0.0407663956,-0.008215704,0.1668746769,0.3400380611,0.1027865037,0.2708595991,0.0819577202,0.1208938435,-0.150346145,-0.2895895541,0.2877454758,-0.33018139,-0.344931066,-0.0755386129,0.0785713568,0.2657673061,-0.0579297617,-0.1058492661,-0.0309242699,-0.4037671387,-0.1928453594,-0.1852022856,-0.4449900985,0.0211743154,0.2638934255,-0.2862427831,-0.0328586698,0.05158443,-0.4806175828,0.1447317004,-0.1111120284,-0.1330832839,0.3892788291,0.1719237566,0.0218227748,-0.152192831,0.1726180464,0.1998168081,-0.1803179681,0.0575537086,-0.4518596828,-0.172994867,0.0644659325,0.4083662331,-0.2023411393,-0.222900331,-0.3968209326,-0.1921593547,0.1015419289,0.0394879319,-0.0883383155,0.2670643032,-0.0757657439,0.3907501996,-0.192009002,-0.4534187615,-0.0437421575,0.5255671144,0.0346293263,-0.1923187524,0.412570715,0.1591852754,-0.1644421071,-0.0695428327,0.1383889616,0.3426440954,-0.6278496981,0.5433270335,0.2498693764,0.0369901843,0.137832284,0.4185103774,0.1368647516,0.2798647583,-0.3155887127,-0.1867186725,-0.2580824792,0.4086378217,0.1035424992,0.0359807424,-0.142046541,-0.2367399484,0.328837961,-0.1956318468,-0.1520757377,0.0660050362,-0.0865769833,0.3288319707,0.2184335589,0.2640419304,0.3569841981,-0.081811808,-0.1028169468,0.131531179,-0.279350847,-0.0736056119,-0.0467084982,0.176355049,0.1042599976,-0.220398739,0.1627230346,-0.4074191153,-0.0460065827,-0.4062663615,0.261949718,0.490960151,-0.0040139835,-0.2378341705,0.3719460666,0.0600784533,-0.0297455657,0.2064462304,-0.1776874214,-0.0705968589,-0.067778334,0.3043535054,0.1634591073,0.0337509252,0.0829738677,0.0720483661,0.014262286,0.0030940275,0.4227865934,-0.20461151,-0.2124462128,0.3322144449,0.2543900013,0.0184191465,-0.0208611134,-0.1518290639,0.2098122537,0.0060319579,-0.1523621529,0.0288437009,0.1550513953,0.3418107033,0.2605259717,0.3646138608,-0.1685419381,-0.2591929734,0.0174677297,0.1886160076,0.0211036801,0.0309402049,0.1880581975,0.3906849027,0.333427757,-0.0989431292,0.3222150207,0.0788699687,0.2825941443,0.2901285887,0.2377108186,0.2660039365,0.0139707457,0.0864454508,-0.2048538476,-0.5083088875,0.0377854444,0.0890346915,-0.0027337347,0.3360840082,0.1234634519,0.1538118124,-0.1924221218,-0.2750574648,0.0418601818,0.0021718226,-0.3454618752,-0.0618339814,-0.3136749864,-0.2561423182,0.4150924981,-0.1609165072,0.0627365261,-0.3454859853,0.3650220633,-0.077607587,-0.0953704789,-0.1441647857,-0.3677257597,0.0240027625,0.2292592824,-0.1103788093,0.1378632039,-0.0913239047,-0.0463289171,0.0504921414,0.5798316002,0.5111631751,0.3985682726,-0.0246103872,0.4008245468,0.130157128,-0.2436257899,-0.2861928642,0.3631888032,-0.0595287718,-0.0250939578,0.2001091689,-0.0104520181,-0.0469259843,-0.0630542859,-0.0026363027,0.1076258346,0.1635151505,-0.1908553392,-0.197421819,-0.2362531722,-0.2687476575,-0.0491944067,-0.0817951337,0.0160535164,0.29386428,-0.0996417478,0.0463557616,0.0778610259,-0.0013327353,-0.0415194184,0.2583118975,0.4536593854,0.0035883174,-0.4841513932,0.1772225052,-0.3718905747,-0.0301732477,-0.4928691983,0.2783645988,0.0389187187,0.197258383,0.1198171675,-0.1226420105,0.3183076978,-0.050173346,-0.0438348092,0.3545496762,0.0648137182,-0.2790630162,-0.1928181499,-0.0135354837,0.0101545881,-0.3842709064,0.1062459424,-0.11351078,-0.0753601417,-0.105609633,0.1699269265,-0.1576716304,0.0275927205,0.1774147898,0.3360162675,-0.0418955907,0.0294316802,-0.0177058894,-0.3504039645,0.1464190781,-0.126489982,0.1394371688,0.359600842,0.4612443745,-0.3284623921,-0.2027059644,-0.0240753405,0.0423464887,0.0046442109,0.0361618474,-0.1818929613,0.1435846388,-0.3938452899,-0.0358402953,-0.1220736802,0.0222712085,0.0860232413,0.0270857383,-0.252587229,-0.3549436331,0.3827648461,-0.6531078815,-0.2620070577,-0.0693184882,0.3420590162,0.2115111947,-0.2256407142,-0.3928543925,-0.1719201207,0.2626016438,-0.0693581328,-0.0616543554,-0.1847913414,-0.5037385225,-0.0344289057,-0.1576237828,0.1152830869,0.0357203335,0.0066587133,0.231196329,-0.3098658323]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2644","title":"Batched `map` not allowed to return 0 items","comments":"Yes and it's all good, thank you :)\r\n\r\nFeel free to close this issue if it's good for you","body":"## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":19,"text":"Batched `map` not allowed to return 0 items\n## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n\nYes and it's all good, thank you :)\r\n\r\nFeel free to close this issue if it's good for you","embeddings":[-0.2328359336,-0.3903888166,-0.0444566682,0.1644899547,-0.1085444242,-0.0687690005,0.1163394973,0.4056320786,0.6400417089,0.1178812236,-0.0565718673,0.2460853755,-0.3954412043,0.1043543741,-0.1629210711,0.1608026177,-0.0449331626,0.1405789405,-0.1691175699,-0.1914925724,-0.3362929821,0.163199544,-0.3979255855,-0.0501046143,-0.1487712115,-0.2919759154,0.286480099,-0.0362105146,-0.162814334,-0.2454052418,0.206091553,-0.2328298539,-0.2091356516,0.5783210993,-0.0001299276,-0.0231365338,0.1505937427,-0.1851885766,-0.0897614583,-0.3029606342,-0.1908181012,-0.1715300977,-0.1157183424,-0.2824648917,0.1819460839,-0.2716506422,-0.1956346184,-0.4482398331,0.1071964726,0.2587007582,0.043321602,0.4188180268,0.003066103,0.1764480174,0.3312488496,0.2664009035,-0.0753516629,0.1288110614,0.7205767035,-0.2881007195,0.0226270352,0.3385902047,-0.3921834528,0.0709065422,0.1367558688,-0.0309035499,0.230168879,-0.5515586734,0.2697506845,0.3675674498,-0.0899998844,0.0607856587,-0.433337301,-0.6687820554,-0.0403036699,-0.0954985172,-0.0388978831,0.2810007632,-0.4210395217,-0.0564594232,-0.4846186638,0.1217950135,0.0078949472,0.1618547142,0.0306755304,0.4589543343,-0.0930409953,0.2003241777,0.3451099396,-0.1618324965,-0.0994021893,-0.1672711968,-0.2236276567,0.5847230554,-0.0366987288,-0.1459876001,0.3893203735,0.2197681516,0.2230358422,-0.3132426441,-0.050306201,-0.0742755681,0.2646799386,-0.0234315637,0.4159452617,0.0793761313,0.2686936855,0.3493200243,-0.005052017,-0.0590938851,0.0676903278,0.1131693944,0.113442108,-0.0890844166,-0.0096516423,-0.0764087662,0.2567855418,-0.0787523314,-0.1218986362,0.2341007143,-0.4100169539,0.1369067729,0.0620478503,0.1250830591,0.1700579822,-0.1776706576,0.1202128828,0.1126287505,-0.1570285559,0.1897115409,-0.0772467405,-0.1316168755,-0.192502588,-0.3208249807,0.1964786202,-0.3342862427,0.2418166101,-0.1585832834,0.0198275335,-0.0615327619,0.0957399309,-0.2645617723,0.7512272596,0.4937881231,-0.119511351,0.2913580537,0.3966238201,-0.2185574174,-0.1752834171,0.4062903821,-0.3826135099,-0.2777348757,0.2100417018,-0.0401864871,-0.2380758375,0.346570462,-0.5073054433,0.2185845971,0.1782458276,0.0822512209,0.2914729118,-0.2697472274,-0.105394572,-0.2360114455,0.0425621271,0.6938397884,-0.63877213,-0.049924884,0.1101892963,0.1013790518,-0.0613662116,0.3478800058,-0.0454604253,0.3481105566,-0.4107360542,0.3059869707,0.0873124897,-0.1264710426,-0.3275630772,0.3134204149,-0.0358151942,0.2356395572,-0.0210545417,-0.0642482191,0.5519376397,-0.1873890162,0.3676523268,-0.1700588763,-0.279956907,0.1041627601,-0.0503817089,-0.039843861,0.0603797175,-0.2420822233,0.19492428,0.0165877249,0.0390782356,-0.5488622189,0.1286391169,-0.1508790404,0.3383628428,0.2401701212,0.262108326,0.1233314499,0.0182514824,-0.2563515604,-0.4713618159,0.3001765013,-0.4880333245,0.021248579,-0.4338338673,-0.1033039168,0.0309646595,0.1834742576,-0.0350200944,0.0330956019,-0.0749208331,-0.4975802898,0.1765710711,-0.0213433541,-0.2262366265,-0.2370698303,-0.1358649582,-0.0084387707,-0.3567225337,0.071359843,0.2501084507,-0.2921575904,-0.2542045712,0.2466300279,0.1149316132,-0.1176109165,0.117354244,0.1758526713,-0.0348161124,-0.1673572958,-0.1400127411,0.131253615,0.0117278975,-0.0458009467,0.101837188,0.1450134069,0.3444767892,-0.2378433347,-0.0694068223,0.5054207444,-0.1876149476,0.5359300375,-0.1796862483,0.0512384214,-0.1996322572,0.042606812,-0.2059215009,-0.3850508034,0.0762836039,-0.1577371508,0.0513374284,0.0319146588,0.0012563785,-0.1989199072,0.0856170282,-0.046962332,0.16477938,0.0144881429,-0.0630519465,0.2266895324,0.0446924791,-0.0142778121,0.3432699442,0.2191425711,-0.0887260064,-0.1521824449,0.250440836,0.0001223587,0.4139182568,0.18166776,0.0260531791,0.1550095528,-0.003640976,-0.0657203272,-0.2528341115,-0.1814302802,0.005715027,0.149533838,-0.3952980638,-0.0354201086,0.1101150513,-0.158504501,0.0689762086,-0.0765032619,-0.290856123,-0.3032285273,-0.0453336127,0.2601582408,-0.2722713947,-0.02400795,-0.4013136625,0.0116532287,0.1975094527,-0.2747197151,-0.2812999189,-0.1685465872,-0.0756110176,-0.0741591007,0.0731139854,-0.1339878589,0.3615631759,0.2521368861,-0.2612816691,-0.243771404,-0.0951137543,0.0065136496,-0.5370389223,0.2054035813,0.4100922644,0.3981708586,-0.2923846245,0.1470615864,0.2793858051,-0.1071430445,-0.2480682433,-0.0079056751,0.0223337524,0.153262645,-0.179215163,0.1474315226,0.0644616261,-0.3209299445,0.1781849116,-0.1891590655,0.2615512013,0.2194533795,0.2780077457,0.1538574994,-0.3693777621,-0.4285494387,0.1223512962,-0.1416769028,0.4460211396,0.0390229337,-0.1469308287,0.1844055504,0.0262182392,-0.1853193641,0.2405568659,-0.0966859683,0.0939638689,-0.0562738925,0.1234829947,0.190006569,0.3668004274,0.4985276759,0.0864884332,0.0460447706,-0.2846722603,-0.1089555398,0.0812640563,0.0222764555,-0.064514026,0.422529459,0.3217896223,0.0287001207,0.7504062057,0.3654765189,0.1272135675,0.0946453437,0.0725369975,0.3822363317,-0.1140064001,-0.379692018,-0.2082244754,0.1538844705,-0.0840781853,0.0325130187,0.0550553165,-0.0620342046,-0.0012852303,0.0619004257,-0.4879447818,-0.1890367717,0.4240106642,-0.3093628883,0.0761560276,-0.0749146789,0.095616892,-0.2788098752,-0.1820272505,-0.1637149751,-0.3036242127,0.642676115,-0.0823247358,-0.0753789768,-0.1154965758,-0.6797434092,0.3325339854,0.2211797684,-0.0848115087,0.1067164391,-0.1852946132,0.0687569454,0.3100571632,0.9494889379,-0.1932723969,0.1566554606,0.0814261511,-0.0248434823,-0.4435347021,0.099806577,-0.3279415369,0.2550869286,0.249542281,0.5071648955,-0.3575258851,-0.072344102,-0.0420534089,-0.0277328677,0.0347220786,-0.0264866166,-0.1733805835,-0.1495417058,-0.138031885,0.3645800352,0.4469464123,0.072685346,0.1356703341,-0.2094711065,-0.0988867208,0.0481703952,0.017599402,-0.1340299547,0.4707481861,0.0158333834,0.0505986102,0.2701371014,0.111734964,0.2206298858,0.490062803,-0.153430447,0.2074077874,-0.1815864742,0.0305913817,0.3123363554,0.4398168921,0.0099648293,0.0579147972,0.381275177,0.3522175848,-0.0515045896,-0.2942019403,0.1262360066,0.0464040264,-0.5417702198,0.0036563263,0.115878664,-0.0995823666,0.0849705786,0.6082314849,0.5790846348,-0.3638046384,0.3609662056,0.0432918593,0.8923068643,-0.0148932636,0.1076276153,0.1486807168,0.0709492192,0.2498285174,0.5552585125,0.0943638608,0.0084218662,0.0974387005,-0.0407663956,-0.008215704,0.1668746769,0.3400380611,0.1027865037,0.2708595991,0.0819577202,0.1208938435,-0.150346145,-0.2895895541,0.2877454758,-0.33018139,-0.344931066,-0.0755386129,0.0785713568,0.2657673061,-0.0579297617,-0.1058492661,-0.0309242699,-0.4037671387,-0.1928453594,-0.1852022856,-0.4449900985,0.0211743154,0.2638934255,-0.2862427831,-0.0328586698,0.05158443,-0.4806175828,0.1447317004,-0.1111120284,-0.1330832839,0.3892788291,0.1719237566,0.0218227748,-0.152192831,0.1726180464,0.1998168081,-0.1803179681,0.0575537086,-0.4518596828,-0.172994867,0.0644659325,0.4083662331,-0.2023411393,-0.222900331,-0.3968209326,-0.1921593547,0.1015419289,0.0394879319,-0.0883383155,0.2670643032,-0.0757657439,0.3907501996,-0.192009002,-0.4534187615,-0.0437421575,0.5255671144,0.0346293263,-0.1923187524,0.412570715,0.1591852754,-0.1644421071,-0.0695428327,0.1383889616,0.3426440954,-0.6278496981,0.5433270335,0.2498693764,0.0369901843,0.137832284,0.4185103774,0.1368647516,0.2798647583,-0.3155887127,-0.1867186725,-0.2580824792,0.4086378217,0.1035424992,0.0359807424,-0.142046541,-0.2367399484,0.328837961,-0.1956318468,-0.1520757377,0.0660050362,-0.0865769833,0.3288319707,0.2184335589,0.2640419304,0.3569841981,-0.081811808,-0.1028169468,0.131531179,-0.279350847,-0.0736056119,-0.0467084982,0.176355049,0.1042599976,-0.220398739,0.1627230346,-0.4074191153,-0.0460065827,-0.4062663615,0.261949718,0.490960151,-0.0040139835,-0.2378341705,0.3719460666,0.0600784533,-0.0297455657,0.2064462304,-0.1776874214,-0.0705968589,-0.067778334,0.3043535054,0.1634591073,0.0337509252,0.0829738677,0.0720483661,0.014262286,0.0030940275,0.4227865934,-0.20461151,-0.2124462128,0.3322144449,0.2543900013,0.0184191465,-0.0208611134,-0.1518290639,0.2098122537,0.0060319579,-0.1523621529,0.0288437009,0.1550513953,0.3418107033,0.2605259717,0.3646138608,-0.1685419381,-0.2591929734,0.0174677297,0.1886160076,0.0211036801,0.0309402049,0.1880581975,0.3906849027,0.333427757,-0.0989431292,0.3222150207,0.0788699687,0.2825941443,0.2901285887,0.2377108186,0.2660039365,0.0139707457,0.0864454508,-0.2048538476,-0.5083088875,0.0377854444,0.0890346915,-0.0027337347,0.3360840082,0.1234634519,0.1538118124,-0.1924221218,-0.2750574648,0.0418601818,0.0021718226,-0.3454618752,-0.0618339814,-0.3136749864,-0.2561423182,0.4150924981,-0.1609165072,0.0627365261,-0.3454859853,0.3650220633,-0.077607587,-0.0953704789,-0.1441647857,-0.3677257597,0.0240027625,0.2292592824,-0.1103788093,0.1378632039,-0.0913239047,-0.0463289171,0.0504921414,0.5798316002,0.5111631751,0.3985682726,-0.0246103872,0.4008245468,0.130157128,-0.2436257899,-0.2861928642,0.3631888032,-0.0595287718,-0.0250939578,0.2001091689,-0.0104520181,-0.0469259843,-0.0630542859,-0.0026363027,0.1076258346,0.1635151505,-0.1908553392,-0.197421819,-0.2362531722,-0.2687476575,-0.0491944067,-0.0817951337,0.0160535164,0.29386428,-0.0996417478,0.0463557616,0.0778610259,-0.0013327353,-0.0415194184,0.2583118975,0.4536593854,0.0035883174,-0.4841513932,0.1772225052,-0.3718905747,-0.0301732477,-0.4928691983,0.2783645988,0.0389187187,0.197258383,0.1198171675,-0.1226420105,0.3183076978,-0.050173346,-0.0438348092,0.3545496762,0.0648137182,-0.2790630162,-0.1928181499,-0.0135354837,0.0101545881,-0.3842709064,0.1062459424,-0.11351078,-0.0753601417,-0.105609633,0.1699269265,-0.1576716304,0.0275927205,0.1774147898,0.3360162675,-0.0418955907,0.0294316802,-0.0177058894,-0.3504039645,0.1464190781,-0.126489982,0.1394371688,0.359600842,0.4612443745,-0.3284623921,-0.2027059644,-0.0240753405,0.0423464887,0.0046442109,0.0361618474,-0.1818929613,0.1435846388,-0.3938452899,-0.0358402953,-0.1220736802,0.0222712085,0.0860232413,0.0270857383,-0.252587229,-0.3549436331,0.3827648461,-0.6531078815,-0.2620070577,-0.0693184882,0.3420590162,0.2115111947,-0.2256407142,-0.3928543925,-0.1719201207,0.2626016438,-0.0693581328,-0.0616543554,-0.1847913414,-0.5037385225,-0.0344289057,-0.1576237828,0.1152830869,0.0357203335,0.0066587133,0.231196329,-0.3098658323]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2644","title":"Batched `map` not allowed to return 0 items","comments":"Everything's good, thanks!","body":"## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":3,"text":"Batched `map` not allowed to return 0 items\n## Describe the bug\r\nI'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`.\r\n\r\nHowever, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L100), but no elements were returned in this case.\r\n\r\nFor this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n # The following line simulates no matches found, so we return an empty batch\r\n result = {'key': []}\r\n return result\r\n\r\nfiltered_dataset = dataset.map(\r\n select_rows,\r\n remove_columns = dataset.column_names,\r\n batched = True,\r\n num_proc = 1,\r\n desc = \"Selecting rows with images that exist\"\r\n)\r\n```\r\n\r\nThe code above immediately triggers the exception. If we use the following instead:\r\n\r\n```python\r\ndef select_rows(examples):\r\n # `key` is a column name that exists in the original dataset\r\n result = {'key': []} # or defaultdict or whatever\r\n \r\n # code to check for condition and append elements to result\r\n # some_items_found will be set to True if there were any matching elements in the batch\r\n \r\n return result if some_items_found else {}\r\n```\r\n\r\nThen it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it.\r\n\r\nIn my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items.\r\n\r\n## Expected results\r\nThe dataset would be filtered and only the matching fields would be returned.\r\n\r\n## Actual results\r\nAn exception is encountered, as described. Using a workaround makes it fail further along the line.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- PyArrow version: 4.0.1\r\n\nEverything's good, thanks!","embeddings":[-0.2328359336,-0.3903888166,-0.0444566682,0.1644899547,-0.1085444242,-0.0687690005,0.1163394973,0.4056320786,0.6400417089,0.1178812236,-0.0565718673,0.2460853755,-0.3954412043,0.1043543741,-0.1629210711,0.1608026177,-0.0449331626,0.1405789405,-0.1691175699,-0.1914925724,-0.3362929821,0.163199544,-0.3979255855,-0.0501046143,-0.1487712115,-0.2919759154,0.286480099,-0.0362105146,-0.162814334,-0.2454052418,0.206091553,-0.2328298539,-0.2091356516,0.5783210993,-0.0001299276,-0.0231365338,0.1505937427,-0.1851885766,-0.0897614583,-0.3029606342,-0.1908181012,-0.1715300977,-0.1157183424,-0.2824648917,0.1819460839,-0.2716506422,-0.1956346184,-0.4482398331,0.1071964726,0.2587007582,0.043321602,0.4188180268,0.003066103,0.1764480174,0.3312488496,0.2664009035,-0.0753516629,0.1288110614,0.7205767035,-0.2881007195,0.0226270352,0.3385902047,-0.3921834528,0.0709065422,0.1367558688,-0.0309035499,0.230168879,-0.5515586734,0.2697506845,0.3675674498,-0.0899998844,0.0607856587,-0.433337301,-0.6687820554,-0.0403036699,-0.0954985172,-0.0388978831,0.2810007632,-0.4210395217,-0.0564594232,-0.4846186638,0.1217950135,0.0078949472,0.1618547142,0.0306755304,0.4589543343,-0.0930409953,0.2003241777,0.3451099396,-0.1618324965,-0.0994021893,-0.1672711968,-0.2236276567,0.5847230554,-0.0366987288,-0.1459876001,0.3893203735,0.2197681516,0.2230358422,-0.3132426441,-0.050306201,-0.0742755681,0.2646799386,-0.0234315637,0.4159452617,0.0793761313,0.2686936855,0.3493200243,-0.005052017,-0.0590938851,0.0676903278,0.1131693944,0.113442108,-0.0890844166,-0.0096516423,-0.0764087662,0.2567855418,-0.0787523314,-0.1218986362,0.2341007143,-0.4100169539,0.1369067729,0.0620478503,0.1250830591,0.1700579822,-0.1776706576,0.1202128828,0.1126287505,-0.1570285559,0.1897115409,-0.0772467405,-0.1316168755,-0.192502588,-0.3208249807,0.1964786202,-0.3342862427,0.2418166101,-0.1585832834,0.0198275335,-0.0615327619,0.0957399309,-0.2645617723,0.7512272596,0.4937881231,-0.119511351,0.2913580537,0.3966238201,-0.2185574174,-0.1752834171,0.4062903821,-0.3826135099,-0.2777348757,0.2100417018,-0.0401864871,-0.2380758375,0.346570462,-0.5073054433,0.2185845971,0.1782458276,0.0822512209,0.2914729118,-0.2697472274,-0.105394572,-0.2360114455,0.0425621271,0.6938397884,-0.63877213,-0.049924884,0.1101892963,0.1013790518,-0.0613662116,0.3478800058,-0.0454604253,0.3481105566,-0.4107360542,0.3059869707,0.0873124897,-0.1264710426,-0.3275630772,0.3134204149,-0.0358151942,0.2356395572,-0.0210545417,-0.0642482191,0.5519376397,-0.1873890162,0.3676523268,-0.1700588763,-0.279956907,0.1041627601,-0.0503817089,-0.039843861,0.0603797175,-0.2420822233,0.19492428,0.0165877249,0.0390782356,-0.5488622189,0.1286391169,-0.1508790404,0.3383628428,0.2401701212,0.262108326,0.1233314499,0.0182514824,-0.2563515604,-0.4713618159,0.3001765013,-0.4880333245,0.021248579,-0.4338338673,-0.1033039168,0.0309646595,0.1834742576,-0.0350200944,0.0330956019,-0.0749208331,-0.4975802898,0.1765710711,-0.0213433541,-0.2262366265,-0.2370698303,-0.1358649582,-0.0084387707,-0.3567225337,0.071359843,0.2501084507,-0.2921575904,-0.2542045712,0.2466300279,0.1149316132,-0.1176109165,0.117354244,0.1758526713,-0.0348161124,-0.1673572958,-0.1400127411,0.131253615,0.0117278975,-0.0458009467,0.101837188,0.1450134069,0.3444767892,-0.2378433347,-0.0694068223,0.5054207444,-0.1876149476,0.5359300375,-0.1796862483,0.0512384214,-0.1996322572,0.042606812,-0.2059215009,-0.3850508034,0.0762836039,-0.1577371508,0.0513374284,0.0319146588,0.0012563785,-0.1989199072,0.0856170282,-0.046962332,0.16477938,0.0144881429,-0.0630519465,0.2266895324,0.0446924791,-0.0142778121,0.3432699442,0.2191425711,-0.0887260064,-0.1521824449,0.250440836,0.0001223587,0.4139182568,0.18166776,0.0260531791,0.1550095528,-0.003640976,-0.0657203272,-0.2528341115,-0.1814302802,0.005715027,0.149533838,-0.3952980638,-0.0354201086,0.1101150513,-0.158504501,0.0689762086,-0.0765032619,-0.290856123,-0.3032285273,-0.0453336127,0.2601582408,-0.2722713947,-0.02400795,-0.4013136625,0.0116532287,0.1975094527,-0.2747197151,-0.2812999189,-0.1685465872,-0.0756110176,-0.0741591007,0.0731139854,-0.1339878589,0.3615631759,0.2521368861,-0.2612816691,-0.243771404,-0.0951137543,0.0065136496,-0.5370389223,0.2054035813,0.4100922644,0.3981708586,-0.2923846245,0.1470615864,0.2793858051,-0.1071430445,-0.2480682433,-0.0079056751,0.0223337524,0.153262645,-0.179215163,0.1474315226,0.0644616261,-0.3209299445,0.1781849116,-0.1891590655,0.2615512013,0.2194533795,0.2780077457,0.1538574994,-0.3693777621,-0.4285494387,0.1223512962,-0.1416769028,0.4460211396,0.0390229337,-0.1469308287,0.1844055504,0.0262182392,-0.1853193641,0.2405568659,-0.0966859683,0.0939638689,-0.0562738925,0.1234829947,0.190006569,0.3668004274,0.4985276759,0.0864884332,0.0460447706,-0.2846722603,-0.1089555398,0.0812640563,0.0222764555,-0.064514026,0.422529459,0.3217896223,0.0287001207,0.7504062057,0.3654765189,0.1272135675,0.0946453437,0.0725369975,0.3822363317,-0.1140064001,-0.379692018,-0.2082244754,0.1538844705,-0.0840781853,0.0325130187,0.0550553165,-0.0620342046,-0.0012852303,0.0619004257,-0.4879447818,-0.1890367717,0.4240106642,-0.3093628883,0.0761560276,-0.0749146789,0.095616892,-0.2788098752,-0.1820272505,-0.1637149751,-0.3036242127,0.642676115,-0.0823247358,-0.0753789768,-0.1154965758,-0.6797434092,0.3325339854,0.2211797684,-0.0848115087,0.1067164391,-0.1852946132,0.0687569454,0.3100571632,0.9494889379,-0.1932723969,0.1566554606,0.0814261511,-0.0248434823,-0.4435347021,0.099806577,-0.3279415369,0.2550869286,0.249542281,0.5071648955,-0.3575258851,-0.072344102,-0.0420534089,-0.0277328677,0.0347220786,-0.0264866166,-0.1733805835,-0.1495417058,-0.138031885,0.3645800352,0.4469464123,0.072685346,0.1356703341,-0.2094711065,-0.0988867208,0.0481703952,0.017599402,-0.1340299547,0.4707481861,0.0158333834,0.0505986102,0.2701371014,0.111734964,0.2206298858,0.490062803,-0.153430447,0.2074077874,-0.1815864742,0.0305913817,0.3123363554,0.4398168921,0.0099648293,0.0579147972,0.381275177,0.3522175848,-0.0515045896,-0.2942019403,0.1262360066,0.0464040264,-0.5417702198,0.0036563263,0.115878664,-0.0995823666,0.0849705786,0.6082314849,0.5790846348,-0.3638046384,0.3609662056,0.0432918593,0.8923068643,-0.0148932636,0.1076276153,0.1486807168,0.0709492192,0.2498285174,0.5552585125,0.0943638608,0.0084218662,0.0974387005,-0.0407663956,-0.008215704,0.1668746769,0.3400380611,0.1027865037,0.2708595991,0.0819577202,0.1208938435,-0.150346145,-0.2895895541,0.2877454758,-0.33018139,-0.344931066,-0.0755386129,0.0785713568,0.2657673061,-0.0579297617,-0.1058492661,-0.0309242699,-0.4037671387,-0.1928453594,-0.1852022856,-0.4449900985,0.0211743154,0.2638934255,-0.2862427831,-0.0328586698,0.05158443,-0.4806175828,0.1447317004,-0.1111120284,-0.1330832839,0.3892788291,0.1719237566,0.0218227748,-0.152192831,0.1726180464,0.1998168081,-0.1803179681,0.0575537086,-0.4518596828,-0.172994867,0.0644659325,0.4083662331,-0.2023411393,-0.222900331,-0.3968209326,-0.1921593547,0.1015419289,0.0394879319,-0.0883383155,0.2670643032,-0.0757657439,0.3907501996,-0.192009002,-0.4534187615,-0.0437421575,0.5255671144,0.0346293263,-0.1923187524,0.412570715,0.1591852754,-0.1644421071,-0.0695428327,0.1383889616,0.3426440954,-0.6278496981,0.5433270335,0.2498693764,0.0369901843,0.137832284,0.4185103774,0.1368647516,0.2798647583,-0.3155887127,-0.1867186725,-0.2580824792,0.4086378217,0.1035424992,0.0359807424,-0.142046541,-0.2367399484,0.328837961,-0.1956318468,-0.1520757377,0.0660050362,-0.0865769833,0.3288319707,0.2184335589,0.2640419304,0.3569841981,-0.081811808,-0.1028169468,0.131531179,-0.279350847,-0.0736056119,-0.0467084982,0.176355049,0.1042599976,-0.220398739,0.1627230346,-0.4074191153,-0.0460065827,-0.4062663615,0.261949718,0.490960151,-0.0040139835,-0.2378341705,0.3719460666,0.0600784533,-0.0297455657,0.2064462304,-0.1776874214,-0.0705968589,-0.067778334,0.3043535054,0.1634591073,0.0337509252,0.0829738677,0.0720483661,0.014262286,0.0030940275,0.4227865934,-0.20461151,-0.2124462128,0.3322144449,0.2543900013,0.0184191465,-0.0208611134,-0.1518290639,0.2098122537,0.0060319579,-0.1523621529,0.0288437009,0.1550513953,0.3418107033,0.2605259717,0.3646138608,-0.1685419381,-0.2591929734,0.0174677297,0.1886160076,0.0211036801,0.0309402049,0.1880581975,0.3906849027,0.333427757,-0.0989431292,0.3222150207,0.0788699687,0.2825941443,0.2901285887,0.2377108186,0.2660039365,0.0139707457,0.0864454508,-0.2048538476,-0.5083088875,0.0377854444,0.0890346915,-0.0027337347,0.3360840082,0.1234634519,0.1538118124,-0.1924221218,-0.2750574648,0.0418601818,0.0021718226,-0.3454618752,-0.0618339814,-0.3136749864,-0.2561423182,0.4150924981,-0.1609165072,0.0627365261,-0.3454859853,0.3650220633,-0.077607587,-0.0953704789,-0.1441647857,-0.3677257597,0.0240027625,0.2292592824,-0.1103788093,0.1378632039,-0.0913239047,-0.0463289171,0.0504921414,0.5798316002,0.5111631751,0.3985682726,-0.0246103872,0.4008245468,0.130157128,-0.2436257899,-0.2861928642,0.3631888032,-0.0595287718,-0.0250939578,0.2001091689,-0.0104520181,-0.0469259843,-0.0630542859,-0.0026363027,0.1076258346,0.1635151505,-0.1908553392,-0.197421819,-0.2362531722,-0.2687476575,-0.0491944067,-0.0817951337,0.0160535164,0.29386428,-0.0996417478,0.0463557616,0.0778610259,-0.0013327353,-0.0415194184,0.2583118975,0.4536593854,0.0035883174,-0.4841513932,0.1772225052,-0.3718905747,-0.0301732477,-0.4928691983,0.2783645988,0.0389187187,0.197258383,0.1198171675,-0.1226420105,0.3183076978,-0.050173346,-0.0438348092,0.3545496762,0.0648137182,-0.2790630162,-0.1928181499,-0.0135354837,0.0101545881,-0.3842709064,0.1062459424,-0.11351078,-0.0753601417,-0.105609633,0.1699269265,-0.1576716304,0.0275927205,0.1774147898,0.3360162675,-0.0418955907,0.0294316802,-0.0177058894,-0.3504039645,0.1464190781,-0.126489982,0.1394371688,0.359600842,0.4612443745,-0.3284623921,-0.2027059644,-0.0240753405,0.0423464887,0.0046442109,0.0361618474,-0.1818929613,0.1435846388,-0.3938452899,-0.0358402953,-0.1220736802,0.0222712085,0.0860232413,0.0270857383,-0.252587229,-0.3549436331,0.3827648461,-0.6531078815,-0.2620070577,-0.0693184882,0.3420590162,0.2115111947,-0.2256407142,-0.3928543925,-0.1719201207,0.2626016438,-0.0693581328,-0.0616543554,-0.1847913414,-0.5037385225,-0.0344289057,-0.1576237828,0.1152830869,0.0357203335,0.0066587133,0.231196329,-0.3098658323]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2643","title":"Enum used in map functions will raise a RecursionError with dill.","comments":"I'm running into this as well. (Thank you so much for reporting @jorgeecardona \u2014 was staring at this massive stack trace and unsure what exactly was wrong!)","body":"## Describe the bug\r\n\r\nEnums used in functions pass to `map` will fail at pickling with a maximum recursion exception as described here: https:\/\/github.com\/uqfoundation\/dill\/issues\/250#issuecomment-852566284\r\n\r\nIn my particular case, I use an enum to define an argument with fixed options using the `TraininigArguments` dataclass as base class and the `HfArgumentParser`. In the same file I use a `ds.map` that tries to pickle the content of the module including the definition of the enum that runs into the dill bug described above.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom enum import Enum\r\n\r\nclass A(Enum):\r\n a = 'a'\r\n\r\ndef main():\r\n a = A.a\r\n \r\n def f(x):\r\n return {} if a == a.a else x\r\n \r\n ds = load_dataset('cnn_dailymail', '3.0.0')['test']\r\n ds = ds.map(f, num_proc=15)\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\n## Expected results\r\nThe known problem with dill could be prevented as explained in the link above (workaround.) Since `HFArgumentParser` nicely uses the enum class for choices it makes sense to also deal with this bug under the hood.\r\n\r\n## Actual results\r\n\r\n```python\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1373, in save_type\r\n pickler.save_reduce(_create_type, (type(obj), obj.__name__,\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 534, in save\r\n self.framer.commit_frame()\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 220, in commit_frame\r\n if f.tell() >= self._FRAME_SIZE_TARGET or force:\r\nRecursionError: maximum recursion depth exceeded while calling a Python object\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.9.0-4-amd64-x86_64-with-glibc2.10\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":27,"text":"Enum used in map functions will raise a RecursionError with dill.\n## Describe the bug\r\n\r\nEnums used in functions pass to `map` will fail at pickling with a maximum recursion exception as described here: https:\/\/github.com\/uqfoundation\/dill\/issues\/250#issuecomment-852566284\r\n\r\nIn my particular case, I use an enum to define an argument with fixed options using the `TraininigArguments` dataclass as base class and the `HfArgumentParser`. In the same file I use a `ds.map` that tries to pickle the content of the module including the definition of the enum that runs into the dill bug described above.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom enum import Enum\r\n\r\nclass A(Enum):\r\n a = 'a'\r\n\r\ndef main():\r\n a = A.a\r\n \r\n def f(x):\r\n return {} if a == a.a else x\r\n \r\n ds = load_dataset('cnn_dailymail', '3.0.0')['test']\r\n ds = ds.map(f, num_proc=15)\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\n## Expected results\r\nThe known problem with dill could be prevented as explained in the link above (workaround.) Since `HFArgumentParser` nicely uses the enum class for choices it makes sense to also deal with this bug under the hood.\r\n\r\n## Actual results\r\n\r\n```python\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1373, in save_type\r\n pickler.save_reduce(_create_type, (type(obj), obj.__name__,\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 534, in save\r\n self.framer.commit_frame()\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 220, in commit_frame\r\n if f.tell() >= self._FRAME_SIZE_TARGET or force:\r\nRecursionError: maximum recursion depth exceeded while calling a Python object\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.9.0-4-amd64-x86_64-with-glibc2.10\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n\nI'm running into this as well. (Thank you so much for reporting @jorgeecardona \u2014 was staring at this massive stack trace and unsure what exactly was wrong!)","embeddings":[0.1053604856,0.1992541552,0.0405323058,0.1406992823,-0.0743715987,-0.0783433765,0.1477604657,0.2337792367,0.1263172179,0.1690644473,0.1453331262,0.9332155585,-0.2908696234,-0.3072986007,-0.0783376396,0.0934459493,0.0223675463,0.0455224887,-0.6181129813,-0.245678708,-0.3375143707,-0.0052422048,-0.1024470925,-0.0100890975,-0.2116018087,0.2239997536,0.0477665961,0.4010353982,0.0841139033,-0.2200670987,0.3333719671,-0.0734497905,0.1974620223,0.3142178059,-0.0001113917,0.0850248039,0.118756108,-0.2274410874,-0.365465194,-0.2395340651,0.1474908143,0.1226186752,0.0726437643,-0.4154757559,0.2226357162,0.3753210306,-0.0025735646,-0.2636497021,-0.0621624291,0.0543684065,0.1718367338,-0.1396808624,0.2170403898,0.0761598274,0.4435100555,-0.0182866361,0.0448160022,0.5206541419,0.1653602272,-0.3055960536,-0.1643484682,0.1722366214,-0.0976958573,-0.0254113507,0.3883353472,-0.1136089861,0.6294152737,-0.0288456883,-0.1640148461,0.2222030163,-0.1290313452,-0.2840961218,-0.2295470983,0.040171016,-0.0794383436,-0.1017234176,0.3087863922,-0.0180283096,-0.0133565376,-0.0775611177,0.124069795,-0.2159891874,0.0056776186,0.4278802872,-0.1641043723,0.5943023562,0.1378127486,0.1661039293,0.2780688107,-0.1684818417,-0.1965375096,0.1178516373,0.1020064875,0.0950327516,-0.2208731472,-0.0939793363,0.2846585512,-0.3601528108,0.3353612125,-0.1542606801,0.0143164145,0.1036642119,-0.0685879216,0.3336909115,0.3899123371,-0.141833812,-0.0516568385,0.5599200726,0.2155424654,-0.112243861,0.0307026245,0.3607044518,0.0600120537,0.091906555,-0.0083617987,-0.1110875979,0.0112875355,-0.0954111367,-0.1062042788,0.4172789752,-0.474712491,-0.1136129498,0.1349558085,0.1240020096,-0.1058503166,0.0211870689,0.1341360658,0.0897011906,-0.0525646433,0.4216713011,-0.2110072672,-0.0060294033,0.0234192684,-0.2131149173,-0.1768059134,0.0548265763,0.2632889152,0.0713272691,0.1323001087,-0.0223733783,0.0825912654,-0.2246848345,0.1783074588,0.102531299,-0.5080521703,-0.083212778,0.2530028224,-0.2604123056,-0.2579993904,0.2515308559,-0.453178972,-0.0923266262,-0.3719526231,0.2141515613,0.1512587517,0.142075941,0.1465840489,-0.1181596518,0.3029899895,-0.1372455209,0.1552291662,-0.4185498357,-0.3111720085,-0.2451112121,0.0752648413,-0.0309842099,-0.072397247,-0.2012919784,-0.2666209936,0.1593035311,0.175672859,-0.049040623,-0.1848199815,-0.1414388418,-0.2404543161,0.2867933214,-0.1029907167,-0.1263521612,-0.3646345437,-0.0630571842,-0.0521467067,0.1691658795,-0.254185915,-0.1637441516,0.2021380663,-0.2736370265,0.5204458833,0.175168708,0.1138640717,-0.0142076705,-0.3202032149,-0.113395296,0.3278877735,0.1209530756,-0.1715014875,-0.0926166549,-0.2653790116,-0.1963817775,0.3977228403,-0.2698430121,0.0405843817,0.131158188,0.108336933,0.0882655382,0.0628296435,-0.1535954922,-0.3434669971,0.1982899308,-0.2558756173,-0.0215071645,0.0164881572,-0.1004741266,0.0963971615,0.2580535114,-0.0351012833,-0.0973073915,0.1891377568,0.0994406566,-0.1080731302,-0.021638101,-0.0003865003,-0.329185307,-0.4009248614,-0.0294388644,0.0524632335,0.302303493,-0.2305079699,-0.2322060466,-0.2058906853,0.1653586477,-0.0919003114,0.0035678879,-0.2516871393,0.0290701613,0.387491405,-0.5933543444,-0.2805407345,-0.0545177348,-0.1167532355,-0.274884969,0.0368423834,0.4684727192,0.1454917938,-0.2771622539,0.2459003329,0.1754849255,0.1758498102,0.1228594929,0.0697168633,0.2272963524,-0.199826479,-0.1135167107,-0.0290244762,-0.3736829162,-0.1019667089,0.0934689865,0.0345105156,0.1148613542,-0.0388505384,-0.2529444695,0.1679152846,0.2472826391,0.2872360945,0.2680383921,-0.1414113045,0.0800695792,0.1683732569,0.056109257,0.1142272428,-0.0879762098,0.097669974,0.1423885375,-0.1171692982,0.1136519611,-0.0736678466,0.0754612312,0.3981694579,0.1457406729,-0.0600115098,0.0496477075,-0.0412557833,-0.2883348167,-0.0536869168,0.102130346,-0.0684398487,0.0745906532,-0.4567013681,0.4747098982,0.1333308071,-0.5235562921,-0.2458880842,-0.2118292898,-0.170154497,0.1294812113,-0.523945272,0.2111630291,-0.1791205406,-0.4466718137,0.1996660978,-0.344062984,0.2858987749,-0.3210588396,-0.2967031598,-0.0235466622,0.1581546515,-0.1790600866,0.0374892913,0.262714833,-0.3319718242,0.0355784371,-0.2012701929,0.146975711,0.0324610099,0.1711266041,0.2742405534,0.0931569561,-0.4386810958,-0.4228983819,0.1325159371,0.1339356452,-0.0586651936,0.060103815,0.2963472307,0.2722679377,-0.2222012132,-0.3065881729,-0.1973068565,-0.3845987916,-0.2084994912,0.0703798085,0.4546933472,0.2742165625,-0.0160644073,0.4758411646,0.3324288726,0.3127490878,-0.3234798014,0.3757663071,0.2017489374,-0.1626095921,-0.2093153149,-0.1637184769,-0.5420069695,0.0844266042,0.1182149425,-0.5454444885,-0.2052110136,0.1891743392,0.0723087415,-0.2491007149,-0.1957806796,0.1190048307,0.2233190536,0.0628091916,-0.0219544377,-0.2726022601,0.2237222791,0.2733663917,0.151839152,0.178386867,-0.0109509388,-0.0432436094,0.3684876859,0.3489730954,-0.0253312588,0.4116002917,0.3529876173,-0.0865499526,0.0425222665,0.0363262221,0.1121196747,0.0395058356,-0.0896948352,0.1329944879,-0.1452875882,-0.6623025537,0.0320489667,0.1480097175,0.3037308156,-0.4923022091,0.2677827179,-0.3939769864,0.0531443954,-0.3042103946,0.2357545048,0.1242341846,0.2263368964,0.1090450808,-0.0914253443,0.1903573573,-0.119236961,-0.0924873874,-0.1650369912,-0.0067229732,0.1185412407,0.3353762925,0.3717659414,-0.0453587957,-0.3760144413,-0.1243962571,-0.1063320711,0.4010097384,-0.348198235,0.0769576058,0.4122265279,-0.2628343105,-0.2899609208,-0.1358057708,0.0748557746,0.04181467,0.3262100518,-0.1238102093,-0.4923030436,0.1185786054,0.4219985008,0.0081256106,0.1181252748,-0.0038845951,-0.0588985309,-0.1878030002,-0.2951493263,-0.0115583511,-0.1458280981,0.1225045398,0.2150284946,-0.0180001073,-0.1149309427,0.1004170701,-0.083802782,-0.1668771803,0.286261797,-0.1239508986,0.1584712863,0.1450294107,0.3025132716,0.2738066912,0.3276348114,-0.3374811709,0.1992717832,0.2932472825,0.225600034,0.1761763394,0.055248104,0.2008027434,0.153792724,0.0105553176,-0.190515995,-0.1788253039,-0.3223810494,0.0872777924,0.0500287153,-0.1219283715,-0.3815679848,0.0662227944,0.1693701446,-0.0110464906,0.0616265237,-0.3657358885,-0.0102443323,0.4981328547,0.3520674407,0.8710961938,0.2196191847,-0.0461065657,0.0972311571,-0.0263179112,-0.2297058851,0.3024886549,-0.1804317832,-0.298661232,0.0040674191,-0.0744967759,-0.1246090904,-0.0215666257,-0.0152552426,-0.3948971331,0.023443792,-0.1997488886,0.0419672355,0.0282367524,-0.052602414,-0.0607006252,-0.1094831228,0.2939590514,0.1608690619,0.092987366,0.0798245817,-0.0656135753,-0.0919522792,-0.2405122072,-0.3805098832,-0.1540364474,0.1441755742,-0.113765955,0.4462915659,0.3543665707,-0.2859048545,0.132515952,0.3625552058,-0.2675669491,0.1217103004,-0.0649961829,-0.0234083273,0.0419259332,0.1990351826,0.256991595,-0.0653651804,0.5327146053,-0.0084034232,-0.232122466,0.1457434893,-0.0940770656,-0.2741780281,-0.0868037418,0.3060580492,0.3108747303,-0.0182180591,0.0556958988,-0.0570490658,-0.1751941741,-0.1722622961,0.1487811357,0.1978515238,-0.1734806895,0.2513152361,0.0907398239,-0.2189705074,-0.0801058486,0.401026696,0.4048638046,0.5142771602,0.3133557141,-0.0660538897,-0.1195442602,-0.152550444,-0.3521785736,-0.353916347,0.2396395355,0.2017319798,0.0184830464,-0.1471089572,-0.3326056302,-0.0950678438,0.0819411352,0.2781379223,-0.4085233808,-0.2797383368,-0.3357233703,0.0974265784,-0.1204471439,0.214709416,-0.2588467002,-0.0375364274,-0.315636605,0.2042546868,-0.3219061494,0.0381848291,-0.3997866213,0.172280848,-0.1483059227,-0.2481526434,0.0182314105,0.030765146,0.16738756,0.2304180712,-0.2873899937,-0.1121997014,-0.2405648828,0.0865457207,0.1410399079,-0.136915043,0.1307747364,-0.1663818508,0.0774900615,-0.1897876561,0.0325981155,-0.1548467577,-0.1185716167,0.3641763628,0.1978075355,0.1557996869,0.1883445233,0.2656750381,-0.1431949884,0.2839567363,-0.1587206721,0.0132913869,-0.2259700447,-0.0838630572,-0.0842180699,0.0726768449,-0.2182113379,-0.2391781211,0.2639332712,-0.0150741013,-0.1912009418,-0.0865056589,0.1725257486,0.2049276382,-0.3796469569,-0.1530585438,0.5407616496,0.2050435245,-0.0046196752,-0.034104757,0.5727972984,-0.0924606994,0.2316608131,0.3061138988,-0.1537572145,-0.1879758388,-0.1503074467,-0.0117991408,-0.347412318,0.0528133772,0.2156009674,0.4500202239,-0.1907598674,0.1537032723,0.1534421295,-0.1342354417,0.543471396,0.014411536,0.13494578,0.1574079692,-0.0015679502,0.0444371216,0.3795166612,-0.4899705648,0.4982151687,0.6172639132,0.2385596186,0.3534963131,0.113644518,0.1797678173,-0.1803993583,0.031092912,-0.1614094526,0.2059415877,-0.347053647,-0.0182736274,-0.0175698288,-0.1287972331,0.0025709434,0.0317995027,-0.0331078768,-0.1463284642,0.2225922793,0.1030733287,-0.2522372901,-0.3613153398,-0.0293793548,-0.1957933158,0.0197423287,-0.2202988565,-0.0594290979,0.1275481731,0.1394574195,-0.3628342152,0.1233991608,0.2021780014,0.5606964827,-0.4515576959,-0.2535941005,0.295152545,0.0871662647,-0.298833847,0.1697481275,-0.2322255969,-0.1714939475,0.442286998,0.0752162784,-0.2449633479,-0.1833675355,0.1342613399,0.4924425185,0.2643300891,0.0972132012,0.2130136937,0.2047800869,0.1633075774,-0.2080276459,-0.2605635226,-0.1555678248,0.1988855749,-0.0063203233,0.319568187,0.072877422,0.0928778946,-0.0425355993,0.2382954061,0.2152564079,0.1745687723,-0.0464246981,-0.0878164098,-0.2692610621,0.3507028818,-0.1388013363,-0.2081338614,0.3059596717,-0.0111727286,0.0657363832,0.1503104717,-0.1154858842,-0.2089880705,0.1607624143,0.4486097991,-0.2906093001,-0.1849911362,-0.1059620678,-0.1510676295,0.2687521875,-0.5409830809,0.1342630833,0.1473646313,0.070543848,-0.2318166494,0.0140693821,-0.011606033,0.3986549675,0.4254052043,0.1147441939,0.4639270306,-0.1939551681,-0.1057455763,0.2056579888,0.0157291461,-0.02135076,-0.169184491,-0.1176791862,0.3652634323,-0.0849698782,-0.2337134629,-0.1284033954,0.5321088433,-0.1069320366,-0.1859683394,-0.1602980644,0.3123152256,-0.0132144419,0.0239587072,-0.4056390226,0.205797568,0.1661835015,0.2969307899,-0.0653434992,-0.2356067747,0.7419953942,0.0621224977,-0.0540980138,-0.1270354241,0.0202787891,-0.5549423099,-0.1697122902,-0.2463076562,-0.1268685609,0.2110933661,0.194737643,-0.5221104622,-0.1502920389,0.1450449228,0.0715027526,-0.2004755437,0.6373544335,-0.2366663516,-0.3341544271,0.2838115394,-0.097884126]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2643","title":"Enum used in map functions will raise a RecursionError with dill.","comments":"Hi ! Thanks for reporting :)\r\n\r\nUntil this is fixed on `dill`'s side, we could implement a custom saving in our Pickler indefined in utils.py_utils.py\r\nThere is already a suggestion in this message about how to do it:\r\nhttps:\/\/github.com\/uqfoundation\/dill\/issues\/250#issuecomment-852566284\r\n\r\nLet me know if such a workaround could help, and feel free to open a PR if you want to contribute !","body":"## Describe the bug\r\n\r\nEnums used in functions pass to `map` will fail at pickling with a maximum recursion exception as described here: https:\/\/github.com\/uqfoundation\/dill\/issues\/250#issuecomment-852566284\r\n\r\nIn my particular case, I use an enum to define an argument with fixed options using the `TraininigArguments` dataclass as base class and the `HfArgumentParser`. In the same file I use a `ds.map` that tries to pickle the content of the module including the definition of the enum that runs into the dill bug described above.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom enum import Enum\r\n\r\nclass A(Enum):\r\n a = 'a'\r\n\r\ndef main():\r\n a = A.a\r\n \r\n def f(x):\r\n return {} if a == a.a else x\r\n \r\n ds = load_dataset('cnn_dailymail', '3.0.0')['test']\r\n ds = ds.map(f, num_proc=15)\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\n## Expected results\r\nThe known problem with dill could be prevented as explained in the link above (workaround.) Since `HFArgumentParser` nicely uses the enum class for choices it makes sense to also deal with this bug under the hood.\r\n\r\n## Actual results\r\n\r\n```python\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1373, in save_type\r\n pickler.save_reduce(_create_type, (type(obj), obj.__name__,\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 534, in save\r\n self.framer.commit_frame()\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 220, in commit_frame\r\n if f.tell() >= self._FRAME_SIZE_TARGET or force:\r\nRecursionError: maximum recursion depth exceeded while calling a Python object\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.9.0-4-amd64-x86_64-with-glibc2.10\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":61,"text":"Enum used in map functions will raise a RecursionError with dill.\n## Describe the bug\r\n\r\nEnums used in functions pass to `map` will fail at pickling with a maximum recursion exception as described here: https:\/\/github.com\/uqfoundation\/dill\/issues\/250#issuecomment-852566284\r\n\r\nIn my particular case, I use an enum to define an argument with fixed options using the `TraininigArguments` dataclass as base class and the `HfArgumentParser`. In the same file I use a `ds.map` that tries to pickle the content of the module including the definition of the enum that runs into the dill bug described above.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom enum import Enum\r\n\r\nclass A(Enum):\r\n a = 'a'\r\n\r\ndef main():\r\n a = A.a\r\n \r\n def f(x):\r\n return {} if a == a.a else x\r\n \r\n ds = load_dataset('cnn_dailymail', '3.0.0')['test']\r\n ds = ds.map(f, num_proc=15)\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\n## Expected results\r\nThe known problem with dill could be prevented as explained in the link above (workaround.) Since `HFArgumentParser` nicely uses the enum class for choices it makes sense to also deal with this bug under the hood.\r\n\r\n## Actual results\r\n\r\n```python\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1373, in save_type\r\n pickler.save_reduce(_create_type, (type(obj), obj.__name__,\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 534, in save\r\n self.framer.commit_frame()\r\n File \"\/home\/xxxx\/miniconda3\/lib\/python3.8\/pickle.py\", line 220, in commit_frame\r\n if f.tell() >= self._FRAME_SIZE_TARGET or force:\r\nRecursionError: maximum recursion depth exceeded while calling a Python object\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.9.0-4-amd64-x86_64-with-glibc2.10\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n\nHi ! Thanks for reporting :)\r\n\r\nUntil this is fixed on `dill`'s side, we could implement a custom saving in our Pickler indefined in utils.py_utils.py\r\nThere is already a suggestion in this message about how to do it:\r\nhttps:\/\/github.com\/uqfoundation\/dill\/issues\/250#issuecomment-852566284\r\n\r\nLet me know if such a workaround could help, and feel free to open a PR if you want to contribute !","embeddings":[0.1053604856,0.1992541552,0.0405323058,0.1406992823,-0.0743715987,-0.0783433765,0.1477604657,0.2337792367,0.1263172179,0.1690644473,0.1453331262,0.9332155585,-0.2908696234,-0.3072986007,-0.0783376396,0.0934459493,0.0223675463,0.0455224887,-0.6181129813,-0.245678708,-0.3375143707,-0.0052422048,-0.1024470925,-0.0100890975,-0.2116018087,0.2239997536,0.0477665961,0.4010353982,0.0841139033,-0.2200670987,0.3333719671,-0.0734497905,0.1974620223,0.3142178059,-0.0001113917,0.0850248039,0.118756108,-0.2274410874,-0.365465194,-0.2395340651,0.1474908143,0.1226186752,0.0726437643,-0.4154757559,0.2226357162,0.3753210306,-0.0025735646,-0.2636497021,-0.0621624291,0.0543684065,0.1718367338,-0.1396808624,0.2170403898,0.0761598274,0.4435100555,-0.0182866361,0.0448160022,0.5206541419,0.1653602272,-0.3055960536,-0.1643484682,0.1722366214,-0.0976958573,-0.0254113507,0.3883353472,-0.1136089861,0.6294152737,-0.0288456883,-0.1640148461,0.2222030163,-0.1290313452,-0.2840961218,-0.2295470983,0.040171016,-0.0794383436,-0.1017234176,0.3087863922,-0.0180283096,-0.0133565376,-0.0775611177,0.124069795,-0.2159891874,0.0056776186,0.4278802872,-0.1641043723,0.5943023562,0.1378127486,0.1661039293,0.2780688107,-0.1684818417,-0.1965375096,0.1178516373,0.1020064875,0.0950327516,-0.2208731472,-0.0939793363,0.2846585512,-0.3601528108,0.3353612125,-0.1542606801,0.0143164145,0.1036642119,-0.0685879216,0.3336909115,0.3899123371,-0.141833812,-0.0516568385,0.5599200726,0.2155424654,-0.112243861,0.0307026245,0.3607044518,0.0600120537,0.091906555,-0.0083617987,-0.1110875979,0.0112875355,-0.0954111367,-0.1062042788,0.4172789752,-0.474712491,-0.1136129498,0.1349558085,0.1240020096,-0.1058503166,0.0211870689,0.1341360658,0.0897011906,-0.0525646433,0.4216713011,-0.2110072672,-0.0060294033,0.0234192684,-0.2131149173,-0.1768059134,0.0548265763,0.2632889152,0.0713272691,0.1323001087,-0.0223733783,0.0825912654,-0.2246848345,0.1783074588,0.102531299,-0.5080521703,-0.083212778,0.2530028224,-0.2604123056,-0.2579993904,0.2515308559,-0.453178972,-0.0923266262,-0.3719526231,0.2141515613,0.1512587517,0.142075941,0.1465840489,-0.1181596518,0.3029899895,-0.1372455209,0.1552291662,-0.4185498357,-0.3111720085,-0.2451112121,0.0752648413,-0.0309842099,-0.072397247,-0.2012919784,-0.2666209936,0.1593035311,0.175672859,-0.049040623,-0.1848199815,-0.1414388418,-0.2404543161,0.2867933214,-0.1029907167,-0.1263521612,-0.3646345437,-0.0630571842,-0.0521467067,0.1691658795,-0.254185915,-0.1637441516,0.2021380663,-0.2736370265,0.5204458833,0.175168708,0.1138640717,-0.0142076705,-0.3202032149,-0.113395296,0.3278877735,0.1209530756,-0.1715014875,-0.0926166549,-0.2653790116,-0.1963817775,0.3977228403,-0.2698430121,0.0405843817,0.131158188,0.108336933,0.0882655382,0.0628296435,-0.1535954922,-0.3434669971,0.1982899308,-0.2558756173,-0.0215071645,0.0164881572,-0.1004741266,0.0963971615,0.2580535114,-0.0351012833,-0.0973073915,0.1891377568,0.0994406566,-0.1080731302,-0.021638101,-0.0003865003,-0.329185307,-0.4009248614,-0.0294388644,0.0524632335,0.302303493,-0.2305079699,-0.2322060466,-0.2058906853,0.1653586477,-0.0919003114,0.0035678879,-0.2516871393,0.0290701613,0.387491405,-0.5933543444,-0.2805407345,-0.0545177348,-0.1167532355,-0.274884969,0.0368423834,0.4684727192,0.1454917938,-0.2771622539,0.2459003329,0.1754849255,0.1758498102,0.1228594929,0.0697168633,0.2272963524,-0.199826479,-0.1135167107,-0.0290244762,-0.3736829162,-0.1019667089,0.0934689865,0.0345105156,0.1148613542,-0.0388505384,-0.2529444695,0.1679152846,0.2472826391,0.2872360945,0.2680383921,-0.1414113045,0.0800695792,0.1683732569,0.056109257,0.1142272428,-0.0879762098,0.097669974,0.1423885375,-0.1171692982,0.1136519611,-0.0736678466,0.0754612312,0.3981694579,0.1457406729,-0.0600115098,0.0496477075,-0.0412557833,-0.2883348167,-0.0536869168,0.102130346,-0.0684398487,0.0745906532,-0.4567013681,0.4747098982,0.1333308071,-0.5235562921,-0.2458880842,-0.2118292898,-0.170154497,0.1294812113,-0.523945272,0.2111630291,-0.1791205406,-0.4466718137,0.1996660978,-0.344062984,0.2858987749,-0.3210588396,-0.2967031598,-0.0235466622,0.1581546515,-0.1790600866,0.0374892913,0.262714833,-0.3319718242,0.0355784371,-0.2012701929,0.146975711,0.0324610099,0.1711266041,0.2742405534,0.0931569561,-0.4386810958,-0.4228983819,0.1325159371,0.1339356452,-0.0586651936,0.060103815,0.2963472307,0.2722679377,-0.2222012132,-0.3065881729,-0.1973068565,-0.3845987916,-0.2084994912,0.0703798085,0.4546933472,0.2742165625,-0.0160644073,0.4758411646,0.3324288726,0.3127490878,-0.3234798014,0.3757663071,0.2017489374,-0.1626095921,-0.2093153149,-0.1637184769,-0.5420069695,0.0844266042,0.1182149425,-0.5454444885,-0.2052110136,0.1891743392,0.0723087415,-0.2491007149,-0.1957806796,0.1190048307,0.2233190536,0.0628091916,-0.0219544377,-0.2726022601,0.2237222791,0.2733663917,0.151839152,0.178386867,-0.0109509388,-0.0432436094,0.3684876859,0.3489730954,-0.0253312588,0.4116002917,0.3529876173,-0.0865499526,0.0425222665,0.0363262221,0.1121196747,0.0395058356,-0.0896948352,0.1329944879,-0.1452875882,-0.6623025537,0.0320489667,0.1480097175,0.3037308156,-0.4923022091,0.2677827179,-0.3939769864,0.0531443954,-0.3042103946,0.2357545048,0.1242341846,0.2263368964,0.1090450808,-0.0914253443,0.1903573573,-0.119236961,-0.0924873874,-0.1650369912,-0.0067229732,0.1185412407,0.3353762925,0.3717659414,-0.0453587957,-0.3760144413,-0.1243962571,-0.1063320711,0.4010097384,-0.348198235,0.0769576058,0.4122265279,-0.2628343105,-0.2899609208,-0.1358057708,0.0748557746,0.04181467,0.3262100518,-0.1238102093,-0.4923030436,0.1185786054,0.4219985008,0.0081256106,0.1181252748,-0.0038845951,-0.0588985309,-0.1878030002,-0.2951493263,-0.0115583511,-0.1458280981,0.1225045398,0.2150284946,-0.0180001073,-0.1149309427,0.1004170701,-0.083802782,-0.1668771803,0.286261797,-0.1239508986,0.1584712863,0.1450294107,0.3025132716,0.2738066912,0.3276348114,-0.3374811709,0.1992717832,0.2932472825,0.225600034,0.1761763394,0.055248104,0.2008027434,0.153792724,0.0105553176,-0.190515995,-0.1788253039,-0.3223810494,0.0872777924,0.0500287153,-0.1219283715,-0.3815679848,0.0662227944,0.1693701446,-0.0110464906,0.0616265237,-0.3657358885,-0.0102443323,0.4981328547,0.3520674407,0.8710961938,0.2196191847,-0.0461065657,0.0972311571,-0.0263179112,-0.2297058851,0.3024886549,-0.1804317832,-0.298661232,0.0040674191,-0.0744967759,-0.1246090904,-0.0215666257,-0.0152552426,-0.3948971331,0.023443792,-0.1997488886,0.0419672355,0.0282367524,-0.052602414,-0.0607006252,-0.1094831228,0.2939590514,0.1608690619,0.092987366,0.0798245817,-0.0656135753,-0.0919522792,-0.2405122072,-0.3805098832,-0.1540364474,0.1441755742,-0.113765955,0.4462915659,0.3543665707,-0.2859048545,0.132515952,0.3625552058,-0.2675669491,0.1217103004,-0.0649961829,-0.0234083273,0.0419259332,0.1990351826,0.256991595,-0.0653651804,0.5327146053,-0.0084034232,-0.232122466,0.1457434893,-0.0940770656,-0.2741780281,-0.0868037418,0.3060580492,0.3108747303,-0.0182180591,0.0556958988,-0.0570490658,-0.1751941741,-0.1722622961,0.1487811357,0.1978515238,-0.1734806895,0.2513152361,0.0907398239,-0.2189705074,-0.0801058486,0.401026696,0.4048638046,0.5142771602,0.3133557141,-0.0660538897,-0.1195442602,-0.152550444,-0.3521785736,-0.353916347,0.2396395355,0.2017319798,0.0184830464,-0.1471089572,-0.3326056302,-0.0950678438,0.0819411352,0.2781379223,-0.4085233808,-0.2797383368,-0.3357233703,0.0974265784,-0.1204471439,0.214709416,-0.2588467002,-0.0375364274,-0.315636605,0.2042546868,-0.3219061494,0.0381848291,-0.3997866213,0.172280848,-0.1483059227,-0.2481526434,0.0182314105,0.030765146,0.16738756,0.2304180712,-0.2873899937,-0.1121997014,-0.2405648828,0.0865457207,0.1410399079,-0.136915043,0.1307747364,-0.1663818508,0.0774900615,-0.1897876561,0.0325981155,-0.1548467577,-0.1185716167,0.3641763628,0.1978075355,0.1557996869,0.1883445233,0.2656750381,-0.1431949884,0.2839567363,-0.1587206721,0.0132913869,-0.2259700447,-0.0838630572,-0.0842180699,0.0726768449,-0.2182113379,-0.2391781211,0.2639332712,-0.0150741013,-0.1912009418,-0.0865056589,0.1725257486,0.2049276382,-0.3796469569,-0.1530585438,0.5407616496,0.2050435245,-0.0046196752,-0.034104757,0.5727972984,-0.0924606994,0.2316608131,0.3061138988,-0.1537572145,-0.1879758388,-0.1503074467,-0.0117991408,-0.347412318,0.0528133772,0.2156009674,0.4500202239,-0.1907598674,0.1537032723,0.1534421295,-0.1342354417,0.543471396,0.014411536,0.13494578,0.1574079692,-0.0015679502,0.0444371216,0.3795166612,-0.4899705648,0.4982151687,0.6172639132,0.2385596186,0.3534963131,0.113644518,0.1797678173,-0.1803993583,0.031092912,-0.1614094526,0.2059415877,-0.347053647,-0.0182736274,-0.0175698288,-0.1287972331,0.0025709434,0.0317995027,-0.0331078768,-0.1463284642,0.2225922793,0.1030733287,-0.2522372901,-0.3613153398,-0.0293793548,-0.1957933158,0.0197423287,-0.2202988565,-0.0594290979,0.1275481731,0.1394574195,-0.3628342152,0.1233991608,0.2021780014,0.5606964827,-0.4515576959,-0.2535941005,0.295152545,0.0871662647,-0.298833847,0.1697481275,-0.2322255969,-0.1714939475,0.442286998,0.0752162784,-0.2449633479,-0.1833675355,0.1342613399,0.4924425185,0.2643300891,0.0972132012,0.2130136937,0.2047800869,0.1633075774,-0.2080276459,-0.2605635226,-0.1555678248,0.1988855749,-0.0063203233,0.319568187,0.072877422,0.0928778946,-0.0425355993,0.2382954061,0.2152564079,0.1745687723,-0.0464246981,-0.0878164098,-0.2692610621,0.3507028818,-0.1388013363,-0.2081338614,0.3059596717,-0.0111727286,0.0657363832,0.1503104717,-0.1154858842,-0.2089880705,0.1607624143,0.4486097991,-0.2906093001,-0.1849911362,-0.1059620678,-0.1510676295,0.2687521875,-0.5409830809,0.1342630833,0.1473646313,0.070543848,-0.2318166494,0.0140693821,-0.011606033,0.3986549675,0.4254052043,0.1147441939,0.4639270306,-0.1939551681,-0.1057455763,0.2056579888,0.0157291461,-0.02135076,-0.169184491,-0.1176791862,0.3652634323,-0.0849698782,-0.2337134629,-0.1284033954,0.5321088433,-0.1069320366,-0.1859683394,-0.1602980644,0.3123152256,-0.0132144419,0.0239587072,-0.4056390226,0.205797568,0.1661835015,0.2969307899,-0.0653434992,-0.2356067747,0.7419953942,0.0621224977,-0.0540980138,-0.1270354241,0.0202787891,-0.5549423099,-0.1697122902,-0.2463076562,-0.1268685609,0.2110933661,0.194737643,-0.5221104622,-0.1502920389,0.1450449228,0.0715027526,-0.2004755437,0.6373544335,-0.2366663516,-0.3341544271,0.2838115394,-0.097884126]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2642","title":"Support multi-worker with streaming dataset (IterableDataset).","comments":"Hi ! This is a great idea :)\r\nI think we could have something similar to what we have in `datasets.Dataset.map`, i.e. a `num_proc` parameter that tells how many processes to spawn to parallelize the data processing. \r\n\r\nRegarding AUTOTUNE, this could be a nice feature as well, we could see how to add it in a second step","body":"**Is your feature request related to a problem? Please describe.**\r\nThe current `.map` does not support multi-process, CPU can become bottleneck if the pre-processing is complex (e.g. t5 span masking).\r\n\r\n**Describe the solution you'd like**\r\nIdeally `.map` should support multi-worker like tfds, with `AUTOTUNE`.\r\n\r\n**Describe alternatives you've considered**\r\nA simpler solution is to shard the dataset and process it in parallel with pytorch dataloader. The shard does not need to be of equal size.\r\n* https:\/\/pytorch.org\/docs\/stable\/data.html#torch.utils.data.IterableDataset\r\n\r\n**Additional context**\r\n","comment_length":58,"text":"Support multi-worker with streaming dataset (IterableDataset).\n**Is your feature request related to a problem? Please describe.**\r\nThe current `.map` does not support multi-process, CPU can become bottleneck if the pre-processing is complex (e.g. t5 span masking).\r\n\r\n**Describe the solution you'd like**\r\nIdeally `.map` should support multi-worker like tfds, with `AUTOTUNE`.\r\n\r\n**Describe alternatives you've considered**\r\nA simpler solution is to shard the dataset and process it in parallel with pytorch dataloader. The shard does not need to be of equal size.\r\n* https:\/\/pytorch.org\/docs\/stable\/data.html#torch.utils.data.IterableDataset\r\n\r\n**Additional context**\r\n\nHi ! This is a great idea :)\r\nI think we could have something similar to what we have in `datasets.Dataset.map`, i.e. a `num_proc` parameter that tells how many processes to spawn to parallelize the data processing. \r\n\r\nRegarding AUTOTUNE, this could be a nice feature as well, we could see how to add it in a second step","embeddings":[-0.6293717623,-0.5262192488,-0.1424407065,-0.0420489386,-0.1927177757,-0.0104304384,0.5165578127,0.1782796085,0.0827181712,0.1480467021,-0.0847553164,0.3020122349,-0.2687706351,0.2669042647,-0.0826474428,-0.2372147441,-0.1250282079,0.064319551,-0.0082344348,0.2308466136,-0.1249568984,0.0046403394,-0.1339984387,-0.2044828087,-0.3420331776,-0.0932869539,-0.0188155528,0.0379000381,0.317248106,-0.2443901002,0.0603744462,0.4343952835,0.1305341572,0.5852185488,-0.0001076541,-0.0930030048,0.1798295975,-0.000388135,-0.0898807794,0.0590006188,-0.0778129026,-0.1035521254,0.0630580485,-0.3442632556,-0.0596426465,-0.2667041421,0.175497666,-0.4042609334,0.2037626058,-0.1677056253,0.1624763161,0.3757341802,-0.1845998466,0.1089050248,-0.3134669662,-0.0016090807,-0.1212616712,0.069438614,0.6291935444,0.0375892594,-0.3766747415,0.3427779973,-0.2514382899,0.3703422844,0.2419779748,-0.1848531663,-0.3477918208,-0.4835861623,0.0880194157,0.3975020647,0.264756918,-0.2432525605,-0.1754490584,-0.4112393558,-0.1313700378,-0.0672614872,-0.0569014587,0.159991309,-0.2204431891,-0.0761081874,-0.3439476192,-0.0879495889,-0.2108642757,0.1095366105,0.1564038545,0.3157585263,0.0834367275,0.2758638561,0.260576278,0.1159337834,0.3285752237,-0.1067808419,0.332206279,0.1372993886,-0.4830509722,-0.2322729528,-0.0932076126,-0.4594946504,0.1724085361,-0.0372719429,0.2354434133,0.3031925261,-0.1467373371,0.3102487326,0.313496381,-0.273981899,-0.2789573073,-0.1171555594,0.4016439021,-0.1800670773,-0.1007044688,0.0791770443,0.1089924052,-0.3574824333,0.3322720528,0.211125344,0.1256034225,0.3532764018,0.1471811533,-0.3941155076,-0.1462440491,-0.2977169752,-0.063694261,0.1053121909,-0.064752005,0.3614375591,-0.1102483645,-0.1895647496,-0.0609045178,-0.26942119,-0.0108173992,-0.2652403414,-0.0306396093,0.1368186623,0.207841903,-0.5171914101,0.2112829685,-0.0188694466,0.3191460967,0.230833739,0.3875932395,-0.2626329362,0.2241920084,0.0336896703,0.026763346,-0.0016433458,-0.184890911,0.51435256,-0.1868384182,0.3826094866,-0.2418973148,-0.53414464,-0.0030775217,0.1260261983,0.0426396318,-0.0261148065,-0.1706091017,0.3209350407,0.2911277115,-0.2322874814,-0.0450492837,-0.0829442665,-0.3704201281,-0.2616883218,0.1099369377,0.1438874751,0.0900046006,-0.0141940154,-0.0176147521,0.052084934,0.0604343526,0.2784424126,-0.2516744733,-0.1148802638,-0.1209258139,0.2565873563,0.2521813214,-0.06233247,-0.1281469017,0.5759640932,-0.101285927,0.0670711622,0.2291383892,0.4602633417,0.4873190224,-0.1004529223,0.0499885306,0.4320580065,-0.3681764603,0.3279699683,-0.1328706741,-0.2349713296,0.2222042829,0.4411790669,0.2141229957,-0.1205196679,-0.0339910425,-0.3242702186,0.3923345506,-0.1728551537,0.2822696865,0.0077500418,-0.018366443,0.0428027883,0.103909038,-0.4952968061,-0.4326003492,0.1770922691,0.1475961655,-0.131574735,0.3082438409,-0.1840863228,0.2749662995,-0.1251863539,-0.1211157218,-0.1684861779,0.0596544668,0.1639610231,-0.2323682159,-0.4080905914,-0.5029201508,0.2063696533,-0.0860888213,-0.2651174664,-0.1824333817,0.1871777475,0.1442110389,-0.0866211578,-0.1398963183,-0.0140008954,-0.0666918233,-0.161142081,-0.0031100495,0.2786311805,-0.1334877312,0.2993046343,0.0858057737,0.5260654688,0.4518821836,-0.1562732011,0.0458431244,0.2426924109,-0.1348728091,-0.2062017322,-0.0090116095,0.1845544726,-0.33687976,0.2194125354,0.041712638,-0.0113582388,0.2030809373,0.0669188723,-0.1730810255,0.1992408782,-0.0525555797,-0.3027154803,0.0736511126,-0.0028769041,-0.5791730285,0.2516581714,0.3946070373,0.1065459698,-0.0127725545,0.0887364596,-0.1118107513,0.0106015094,0.1834913939,0.1043288931,0.3538615108,0.2624371052,0.1418777704,-0.1068482324,0.1929641515,-0.0949223638,-0.0051331879,0.0241868906,0.3479547799,0.2728956044,0.0987079814,-0.0123774242,-0.0414282717,-0.4256393611,0.2487670779,-0.0164841898,0.0392941087,-0.015385841,0.086906217,-0.0173120126,-0.0803795904,-0.1648901552,-0.0341393463,-0.0209993199,0.1612665355,0.7649838328,-0.1086928695,0.336827904,0.1869968772,0.1899265647,0.0012889812,-0.2053337544,-0.1098533422,-0.2062001526,0.2713899314,0.1082150042,0.1666261554,-0.2358656675,0.4359210432,0.3054795861,-0.0445546247,-0.4260461032,-0.0482704118,-0.0027315952,0.0770408586,-0.2029365301,0.0277969167,0.1210204139,0.2024903148,-0.2030792087,0.0591698736,-0.3240244985,0.0467326082,-0.1095950007,0.0144648943,0.010242519,-0.0533558987,-0.2154246122,-0.2746195495,-0.4614111483,0.3068508208,-0.19972682,0.3495417237,-0.1097627804,0.0805519074,-0.1313741058,0.3815685511,-0.1566507071,-0.0187258665,-0.4405144155,0.2183481753,-0.3194648027,-0.1749503016,-0.0687201545,0.0693106279,0.1051425934,0.4950210154,-0.1572958827,-0.1464728862,-0.1428736746,0.2285068035,-0.1550750434,-0.0148039218,0.4700108469,0.2209294885,-0.0559558645,0.0330369845,0.0362328701,-0.0336562544,0.2212250829,-0.1767287254,0.3686834872,0.1956779212,0.2241793126,0.9834297299,0.2730429769,-0.1933690012,0.1726339757,-0.097055681,-0.2342126518,-0.0213304907,-0.4307989478,0.1177888811,-0.3169783056,0.1275519133,0.2472836971,-0.168302387,-0.4030023217,-0.0669212043,-0.150838986,-0.0975206941,-0.0979943275,0.4519287944,-0.2520998418,0.5229313374,-0.2508051097,-0.2353953123,-0.2905541956,-0.0854285732,-0.0017171795,-0.0591426492,0.53251791,-0.216979757,0.1875937432,0.2896521986,-0.3857785761,0.2180109024,0.2700888216,0.1115728766,0.021134736,-0.0308000892,0.1801567823,-0.0222091693,0.5902085304,-0.1809595376,-0.0085114744,-0.1172944978,-0.4298487306,-0.1785355657,0.1208770126,-0.0219998807,0.5310918689,0.1881127656,0.3519924283,-0.3006885052,-0.3387384415,0.1681139171,-0.0491723381,-0.2068436295,-0.5010204315,-0.1466505378,0.0997285321,-0.3303322196,0.0575715154,0.2879675031,-0.0101082092,-0.0430895276,-0.3219867051,0.0266618822,-0.0725414976,-0.0972080082,0.0893383399,-0.1573425978,0.0391403176,0.2516453266,0.2952630222,0.099836953,-0.0102315145,0.1687058508,-0.0394127667,-0.299524039,0.1923433393,0.0215376392,0.3395831585,0.1412065476,0.073301591,0.1217143685,0.3094638288,0.3474085331,-0.3580639064,-0.0697830096,0.1916181594,0.1663687378,-0.2413716167,-0.5865659714,0.1394339353,-0.0614258535,0.0212546717,0.7402999997,-0.6871585846,-0.0097415,0.0306589324,0.2126387507,0.7318966985,-0.1772031188,-0.0073235896,-0.243995145,-0.1093548685,-0.028422799,-0.6782634854,0.2952972651,-0.3901472986,-0.3920981884,0.0123796547,-0.011121735,-0.0493862033,0.3122452199,-0.1722318679,0.0835743025,0.4423432052,0.1314243972,-0.0283049587,0.5326045156,-0.1827601343,-0.4784121215,0.1902934015,0.100715071,-0.1842092425,-0.0690598935,-0.0500683375,-0.2347414196,0.4787383676,0.1893509924,-0.31637761,-0.1800214499,-0.3334517479,0.2092456818,-0.1788118482,-0.1591260433,0.1755997539,0.3089125454,-0.0353319645,-0.2407430112,-0.1726011038,-0.0363512076,0.1060290486,0.1997060627,-0.2126330435,-0.2004948556,0.2052834332,-0.1876517236,-0.3720924556,0.184706375,0.0795229599,-0.7108382583,-0.1137463227,0.1862836778,0.2419992238,-0.053062696,0.0024892446,0.1466775537,-0.3283701241,-0.0868974105,0.0864286199,0.1477681845,-0.2835683227,0.3946521282,-0.1098798439,-0.0146181285,0.0194647107,0.2222323567,0.2573185861,0.1654961705,0.1047625244,0.1918556094,-0.1148463786,-0.1579134762,-0.079526335,-0.0692368969,-0.2391890734,0.1350431591,0.0412491858,-0.0697527602,-0.0137545923,0.0260217246,-0.1451541185,0.3203093708,-0.4574456513,-0.1812987626,-0.3838823736,-0.0492511615,-0.1282189041,0.2407492548,0.1523815691,0.1835390925,0.2417335957,0.277410984,-0.3676072657,0.0033951742,0.0303822048,0.1178159639,0.3157605231,0.0077411798,0.0328884535,0.1178876385,-0.0120737925,0.140061453,-0.2202544361,-0.2053061277,0.0054280791,0.1310355067,-0.002832399,-0.1303898543,-0.1397244036,-0.3036535978,-0.4215177596,-0.2941893339,0.387407124,0.00734577,-0.1853871047,0.1504860818,0.3489686847,-0.1714664698,-0.3820738494,0.1257567257,0.0127149727,-0.0110706585,-0.1302940845,0.1807024628,0.1936528981,-0.0785646215,0.1138292402,-0.0060946005,0.4877407551,0.0790147632,0.2957890928,-0.0931199118,-0.0083848489,-0.0264299829,0.5431421995,0.0852585509,0.0114001101,0.0572850928,0.1059923917,0.2380095273,-0.2243831009,-0.0963295922,0.2588998675,-0.0133982236,0.3597096503,0.0477934182,0.1111506447,0.2732601166,0.2716708481,0.0893561617,0.0652725548,-0.6158900261,0.1082556024,0.0534311794,0.0312711224,-0.007179006,0.1999311298,-0.0107484721,0.1078582257,0.1145010516,-0.0024117895,0.4664043784,0.344442904,-0.0291094705,0.4277687371,-0.1750665903,-0.1497745067,-0.2447574884,0.054422237,0.2061463445,0.1318284422,0.3818688393,-0.1328710467,-0.3123094738,-0.1352181584,0.2957091928,-0.0832959488,0.080159843,0.1402188092,-0.0916441083,0.0491214059,0.1167821139,0.0031895956,0.1527353078,0.5788418651,0.1417350918,0.0299209543,-0.1154498011,-0.1152566299,0.1618232578,0.4018129706,-0.1014020368,0.0352480747,0.260515362,-0.0876454711,-0.1865589768,0.3738208413,-0.0114057465,0.2288345248,-0.2890346646,0.3083408475,-0.1942512691,0.0493062548,-0.0616876483,0.1305252165,0.0310061555,-0.1848922074,0.271843791,0.0862404704,-0.1359890997,0.1400529444,0.0778839439,-0.0533225238,-0.1775750816,-0.0522574633,0.1619805396,-0.3829640746,0.0948392153,-0.1358783692,-0.3094893694,-0.1989818662,0.3970517516,-0.5590798259,-0.0172540061,-0.0677935854,0.0971185192,-0.0362348855,0.5721184611,-0.0464657024,0.1212590188,-0.1624721289,-0.0892015621,-0.4362131059,0.1404664069,0.0510181114,0.1698209345,0.048459582,0.163326323,0.0967913046,0.2467531562,0.177627638,-0.2321387529,0.2296186835,0.5826244354,-0.26575014,0.0082400525,-0.1866099983,0.1839834452,0.2505390644,-0.2351410985,0.1752399206,-0.0001161284,0.0779676288,-0.139490962,-0.1977130473,0.2392979711,-0.1181762889,0.3051758409,-0.0323589481,0.1504871547,-0.1956106126,0.1657877117,0.01889362,0.1386981457,-0.2586944401,0.1583970189,0.4099116325,0.2171963453,-0.1410984844,-0.2294483036,-0.4657680988,0.1876066476,-0.2457248271,-0.0698331818,-0.0427025333,-0.0891591236,-0.1189175546,0.1519983709,-0.0827867985,0.4181878567,-0.0050670542,-0.1724432558,-0.3620939255,-0.1471317261,0.2909373343,-0.3971106708,-0.3432902396,-0.2164865732,0.1205651164,0.2122367769,0.3054523766,-0.3281841874,-0.0952289179,0.227841258,0.0935083553,-0.0681600124,0.2993309796,-0.0034042825,-0.1540495753,-0.0574610196,0.020473443,0.1408082396,-0.183933571,-0.1861020625,-0.3761651814]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2641","title":"load_dataset(\"financial_phrasebank\") NonMatchingChecksumError","comments":"Hi! It's probably because this dataset is stored on google drive and it has a per day quota limit. It should work if you retry, I was able to initiate the download.\r\n\r\nSimilar issue [here](https:\/\/github.com\/huggingface\/datasets\/issues\/2646)","body":"## Describe the bug\r\nAttempting to download the financial_phrasebank dataset results in a NonMatchingChecksumError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"financial_phrasebank\", 'sentences_allagree')\r\n```\r\n\r\n## Expected results\r\nI expect to see the financial_phrasebank dataset downloaded successfully\r\n\r\n## Actual results\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.researchgate.net\/profile\/Pekka_Malo\/publication\/251231364_FinancialPhraseBank-v10\/data\/0c96051eee4fb1d56e000000\/FinancialPhraseBank-v10.zip']\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.14.232-177.418.amzn2.x86_64-x86_64-with-debian-10.6\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":35,"text":"load_dataset(\"financial_phrasebank\") NonMatchingChecksumError\n## Describe the bug\r\nAttempting to download the financial_phrasebank dataset results in a NonMatchingChecksumError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"financial_phrasebank\", 'sentences_allagree')\r\n```\r\n\r\n## Expected results\r\nI expect to see the financial_phrasebank dataset downloaded successfully\r\n\r\n## Actual results\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.researchgate.net\/profile\/Pekka_Malo\/publication\/251231364_FinancialPhraseBank-v10\/data\/0c96051eee4fb1d56e000000\/FinancialPhraseBank-v10.zip']\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.14.232-177.418.amzn2.x86_64-x86_64-with-debian-10.6\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n\nHi! It's probably because this dataset is stored on google drive and it has a per day quota limit. It should work if you retry, I was able to initiate the download.\r\n\r\nSimilar issue [here](https:\/\/github.com\/huggingface\/datasets\/issues\/2646)","embeddings":[-0.1221614629,0.2191510051,-0.0784073547,0.2884528637,0.2203612775,0.1347171366,0.0658889487,0.3310612142,0.2095745206,0.1821722388,-0.1414383203,0.0533875376,0.1674158573,0.0750155449,-0.141532883,0.0439132415,0.0427031741,-0.0287592523,0.0304532889,-0.1089905724,-0.1035575122,0.3996220529,-0.2143702209,-0.2048737407,-0.0571743771,0.0271739345,0.1256504357,0.1493478715,-0.355509609,-0.1062427387,0.3173011243,0.2356399447,-0.0364060998,0.4088571072,-0.0001165979,-0.1570466757,0.2954361141,-0.00232066,-0.3087729216,-0.4117237628,-0.3101617694,-0.5327080488,-0.0006328061,-0.1562370658,0.1472093612,0.2293883711,0.1187414825,-0.0404394418,0.1839256287,0.2563109994,0.1976445317,0.1766314059,0.1544101089,-0.0672572702,0.2356061935,-0.3061351478,-0.0442656465,0.2061338276,0.1339274347,0.1396629959,-0.1116091758,0.0986639559,-0.2409786433,0.3398787975,-0.0248231217,-0.0569328777,-0.1054870635,-0.4350318611,0.3071873486,0.4293161333,0.5019190311,-0.138317883,-0.333804518,-0.2137775719,-0.1314466596,-0.0050602425,0.4133678973,0.2254237831,-0.2224520743,-0.0487492941,-0.4026898146,0.2127029002,0.0522077084,0.1742328107,0.0008497239,-0.0786489025,-0.1465951353,-0.0958845466,0.2525239885,-0.0932485312,0.0778142214,-0.3102552295,-0.0711328387,0.2796263099,-0.3990389109,-0.0756752491,0.0446167625,0.5124953985,0.5575388074,0.372407496,0.1555419564,0.2275237739,-0.2916593254,0.0197752379,0.1210928038,0.2486981153,0.0146071771,-0.1616899967,0.2106146663,0.4556140304,-0.0590558872,0.1310696006,-0.0883604214,-0.074275732,0.5709369779,-0.0558999628,0.0653932542,-0.5341384411,-0.3482496142,0.444601208,-0.1251795739,-0.1410444379,-0.043768432,0.2691244781,-0.251355052,0.1109621525,-0.0669108108,0.1472851187,-0.0376438834,-0.1506907195,-0.2671844661,-0.0550954677,-0.1249775961,-0.0389060453,0.2755370736,-0.4455491304,0.3306753039,-0.0811800137,0.2223319262,-0.2022905946,0.0439216383,-0.1592709571,-0.2756517529,0.2825688422,0.136262551,0.2232762128,0.0382252634,0.0837551206,0.0077711293,0.01986387,-0.2361139506,-0.1198631674,0.0956027061,0.2120611072,-0.5457581282,-0.426435709,-0.0918331668,-0.2004864216,0.1569826752,-0.0646281019,-0.0429721363,-0.0988637283,-0.2170904577,-0.2330788523,-0.0128975147,0.2208419889,-0.1360917836,-0.0196837783,-0.1142555252,-0.0846794471,0.4203273058,0.3317056596,-0.0746124163,0.0467755422,-0.199337706,0.3415529132,0.2876548469,-0.0402506292,-0.8489012718,0.2736243308,-0.0461516269,0.3179981411,0.1135814711,0.0936846882,-0.0120978067,0.0986145288,0.0596022792,0.2110910118,0.0724116638,0.1698853076,-0.1996620446,-0.3329528868,-0.0283887461,0.1590505093,0.0541048683,0.1359466016,0.1838505566,-0.0445343107,0.289162606,-0.0262767635,0.0180040952,0.2336857468,0.2727536857,0.20902583,0.1021457314,-0.2314194143,-0.3615790606,0.3223058879,-0.2195711583,0.1002212018,-0.0158476625,-0.1486970633,-0.3613629639,-0.2824739516,-0.149021998,0.1300820112,0.0898795724,0.3411881328,0.3992163241,0.0405050106,-0.0459838361,0.3040063679,-0.1290018857,0.0751188993,-0.1407593936,0.2749923766,-0.0342067815,-0.0159849785,0.0301264729,-0.0177684445,0.1630057842,0.0362789519,-0.1874522567,0.2895474732,0.2095121741,0.1428240538,-0.0875202417,0.2737728953,-0.159500286,-0.0725642443,0.0289917383,0.4911872149,0.1428546757,-0.0163486991,-0.1632606685,0.3131428361,0.0755077377,0.0060959174,0.0377293304,0.159274742,0.3733704686,-0.1345328391,-0.0825447515,-0.1518207937,0.5095816851,-0.3409509659,0.2042862773,0.0117272995,-0.2768357694,-0.0628604516,0.2434820831,-0.1203336194,0.1307428926,0.185330078,0.0470325053,-0.075967975,-0.1524516791,0.5569287539,0.2706282437,0.2248675376,0.1861108243,0.0303609613,-0.0696507618,-0.2190412581,-0.0496182106,0.2269037813,0.1424941719,0.502320528,0.2711748779,0.0449598208,-0.4356445968,-0.0609106049,0.0420997292,0.327498734,-0.3210887015,-0.1184834614,-0.0740199834,-0.3158183694,-0.2219233215,-0.1268128157,-0.1441177279,-0.4463294744,-0.1008804739,0.3465779722,0.0165852327,0.0658375844,-0.2468461096,0.056430541,-0.0425222591,-0.2394363582,-0.1359230429,0.2502993047,-0.1967727393,0.0108748442,0.2604633868,-0.0329329446,0.3196413219,-0.391744405,-0.0326173566,-0.7137257457,-0.2451633215,0.0635802969,-0.2230637223,0.2376654297,0.2862507403,0.2511952221,-0.2726148367,0.0127000222,0.1277004629,-0.0658945367,-0.3835560977,0.2892594039,-0.1202679649,0.0041038278,0.0760868415,0.0091314269,-0.2212635875,-0.3392722309,0.105578728,0.1535753757,0.174302265,0.3875114918,-0.1369953603,0.0982941538,-0.0144528039,0.4250248969,-0.289943248,-0.5708986521,0.5505203605,0.1047234163,-0.4363671839,0.065043658,-0.0068894234,0.0403068103,0.2376913726,-0.5485623479,-0.1757599711,-0.4750021994,-0.0197928585,0.2069163471,0.0863772184,0.2451558262,-0.1192784235,-0.0731347576,-0.1427673697,-0.2040537,0.1644188315,0.045720946,0.2432695478,-0.2186840475,0.3314571381,0.0300983507,0.2455593199,0.4993488789,0.2258814722,0.3620388806,-0.0157043617,0.1015442535,-0.1246872172,-0.2068131119,-0.2209392935,-0.1449076235,-0.1907991916,0.2884584665,0.0923221707,-0.0258492325,-0.1714098603,-0.2088149041,-0.3271819353,-0.0996694118,-0.0835728869,0.0374356695,0.2761641741,0.214187026,0.1735148281,0.0233187899,-0.4117975533,-0.0359502696,0.1915614456,-0.1432238966,-0.0861984044,-0.1249313354,0.0669990927,-0.1941172332,0.3062373996,0.2048941404,0.3132772446,0.0918785185,-0.056760706,0.0808625817,-0.1370031834,0.2350171506,-0.3195798695,0.241784662,-0.1173866615,-0.0975533724,-0.040049348,-0.1346797645,-0.1839417368,-0.0066519091,0.1781337708,0.2111274898,-0.3331180513,0.1802726686,0.1622035205,0.3555176854,-0.0734445229,-0.2499942482,-0.360252291,-0.3973374367,-0.4730682075,-0.0465526879,-0.0780729055,0.2964992225,0.0057762382,0.0083070863,0.1236909851,-0.0609051883,0.3348438144,-0.117835246,0.2964917421,0.1444339007,0.431088686,0.0208035167,-0.2529490888,0.242824465,0.8506208658,-0.1115588322,-0.3002641499,-0.0208238456,-0.2596761584,-0.1827894449,0.2945639193,-0.0698252693,-0.1580360085,0.5602295399,-0.0459534824,0.0295528043,-0.184143737,0.2151614577,0.1545519233,-0.3614543378,-0.3829713464,0.2304234952,0.1975788772,-0.0046774303,-0.0037391877,0.0719225854,0.0169111509,0.272488445,-0.0937047601,0.8974538445,-0.0552459881,0.1175897419,0.1804947108,-0.0326114744,0.2929914892,-0.1215180606,0.0069582439,-0.3133254051,-0.3222326636,-0.1347278059,-0.1603716612,-0.0290785711,0.0732391998,-0.0866975263,0.4799547791,-0.079834573,0.1357246637,0.1839293391,0.2583578527,-0.3799636662,0.014992808,-0.0880148485,0.1443735808,-0.1540364176,0.483638674,-0.0140656531,-0.0352432057,-0.1993362159,-0.069743827,-0.4931161702,0.1923331171,-0.100921534,-0.0309277624,0.1258205324,-0.0468102619,-0.0820196122,0.141715914,0.3609253764,0.2172038108,-0.3999876678,0.0911652893,0.4621311426,-0.1197995245,0.161677897,-0.1688745022,0.3982613087,-0.0886032805,-0.0883511454,0.0203617103,0.0301736277,-0.2782293558,-0.1040683612,-0.1673217714,-0.220861122,-0.4374098778,-0.3119948804,-0.2206147462,-0.0220591165,-0.3384540677,0.1484306902,0.3153917193,-0.1170987263,0.0452208221,0.1356401742,-0.3526615798,-0.2894956768,0.5082674623,0.0665679798,-0.2172290832,0.4001287818,0.2137259394,-0.3720187545,-0.2060296685,-0.2480863333,0.01548569,-0.5568864942,0.102847442,-0.1460930407,0.1439404935,-0.1928805262,0.6873797774,0.3661993146,0.0697987527,0.0478507765,-0.5624497533,-0.0862551332,0.0078372406,0.1924950182,0.2684269547,-0.1106655896,-0.1940059066,0.118228808,-0.3049824238,-0.3084527254,0.0951333269,-0.3677180409,-0.2362301201,0.0257713441,0.0456887297,0.2157558352,-0.0252677556,0.1524679512,0.1220989674,-0.1302349418,-0.229039222,-0.0952444226,0.1270601153,0.1071467698,-0.1345247179,0.1926581562,-0.2410964072,-0.2747995555,-0.0860038251,-0.107716687,0.2322985679,-0.0064306245,-0.0020665366,0.4033741653,-0.0852802694,-0.0753919482,-0.0103094289,-0.3013207316,0.1477064639,0.0498765074,0.0464972928,0.1096673831,0.0157171991,0.0190072209,0.1012862325,0.2976625562,-0.1261164844,0.2856270671,-0.4933497906,0.0349993668,0.3791495562,0.4438384771,0.3468146026,-0.3567600846,-0.1222538874,0.2193131149,0.2458251268,-0.3080707192,-0.0401919074,0.1209118292,-0.1554643065,0.0858024061,0.1831367165,0.280410558,-0.2571416497,0.3184908926,0.0923778191,0.3009266257,-0.1381949633,-0.0575688109,0.2769338787,0.0487014428,-0.0406655297,0.1509645134,-0.0880107805,-0.2364577204,-0.0612557009,-0.001863122,0.1562165767,-0.1395245343,0.2553448081,-0.3627968431,-0.4343455434,0.5834093094,0.2807840407,-0.1072905064,0.0108279726,0.0934102908,0.4049898088,0.0636731982,0.4803867638,-0.4838216007,0.124295339,-0.1419125795,-0.0481925234,0.0856263563,-0.3457255065,-0.0336673483,0.2599742115,-0.1245321482,-0.0538146533,0.1761938483,0.0260399356,-0.0662005171,-0.5886382461,-0.2144330293,0.0556662157,0.0368917659,-0.368244797,0.3280424476,0.0068312255,-0.1071481705,0.1543097198,0.5266041756,0.4051484168,0.2689424157,0.0497499481,-0.0460616015,-0.1030209064,-0.0639668107,-0.1054773852,0.7308725715,0.0666670501,0.1839330643,0.4525199234,0.1251128167,-0.2385045737,-0.1184504256,-0.1909351498,-0.165915072,-0.2035482377,0.5051550269,-0.054607939,0.3409983218,-0.232700929,0.0303854961,-0.5357776284,-0.2007468343,0.1476441175,0.0456339568,0.1432324499,0.0100595895,0.0948283374,-0.1558019519,0.2976531684,0.4738085866,0.3946791589,-0.2382072955,-0.1704333276,-0.7519157529,0.2489974797,-0.2298056632,0.0174109284,0.0112719359,0.2548753917,0.0063320687,0.1431942433,0.2265385836,0.1472438425,-0.2896079123,0.0963929743,-0.0963002667,-0.1539226919,0.1487318426,-0.0657908842,-0.1417070925,-0.2585194707,0.2292403877,0.0112274736,0.0666423813,-0.099796176,-0.0953951254,-0.1319230795,-0.0698076338,0.1715169549,0.0703996345,0.7023627758,-0.1001646668,-0.1950004548,-0.1363225579,-0.0615544692,-0.196420297,0.2539495528,-0.0353173018,0.3724659383,0.0276157577,-0.1068934202,-0.2079974711,0.1193821654,0.2141568214,-0.3031130433,-0.2412962466,0.0820641145,-0.2160593718,0.1823354661,-0.1776262522,0.0750409514,0.1989728659,0.2210656404,-0.1798107028,-0.4148392081,0.5957764983,-0.1084750518,-0.19664675,-0.0364140421,0.1348127127,0.066078499,-0.3253687918,-0.7558819056,0.1723068953,0.2922219634,0.0433010533,-0.3046835363,0.2104223073,-0.292498827,0.1448926777,0.0048218095,0.3195835948,0.1815310866,-0.2171662301,0.0646396726,-0.1397776157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2641","title":"load_dataset(\"financial_phrasebank\") NonMatchingChecksumError","comments":"Hi ! Loading the dataset works on my side as well.\r\nFeel free to try again and let us know if it works for you know","body":"## Describe the bug\r\nAttempting to download the financial_phrasebank dataset results in a NonMatchingChecksumError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"financial_phrasebank\", 'sentences_allagree')\r\n```\r\n\r\n## Expected results\r\nI expect to see the financial_phrasebank dataset downloaded successfully\r\n\r\n## Actual results\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.researchgate.net\/profile\/Pekka_Malo\/publication\/251231364_FinancialPhraseBank-v10\/data\/0c96051eee4fb1d56e000000\/FinancialPhraseBank-v10.zip']\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.14.232-177.418.amzn2.x86_64-x86_64-with-debian-10.6\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":26,"text":"load_dataset(\"financial_phrasebank\") NonMatchingChecksumError\n## Describe the bug\r\nAttempting to download the financial_phrasebank dataset results in a NonMatchingChecksumError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"financial_phrasebank\", 'sentences_allagree')\r\n```\r\n\r\n## Expected results\r\nI expect to see the financial_phrasebank dataset downloaded successfully\r\n\r\n## Actual results\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.researchgate.net\/profile\/Pekka_Malo\/publication\/251231364_FinancialPhraseBank-v10\/data\/0c96051eee4fb1d56e000000\/FinancialPhraseBank-v10.zip']\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.14.232-177.418.amzn2.x86_64-x86_64-with-debian-10.6\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n\nHi ! Loading the dataset works on my side as well.\r\nFeel free to try again and let us know if it works for you know","embeddings":[-0.1198334545,0.179409802,-0.0883815736,0.3396996856,0.187968567,0.1173867732,0.0060668611,0.3996661901,0.1419468075,0.1976888925,-0.161383763,0.2437268794,0.230403617,-0.1089785248,-0.092774637,0.140614748,0.1338221133,0.029889619,0.0345540233,-0.1668419242,-0.1839316636,0.3492645025,-0.1968892962,-0.1531044543,-0.0255643241,0.1277683973,0.1105171144,0.1647164822,-0.2496967465,-0.1128350124,0.3732021451,0.1607759744,-0.0779852793,0.4326273203,-0.0001198013,-0.1041963696,0.234977439,-0.0170198269,-0.2702702284,-0.3771175444,-0.4048727453,-0.5063537359,0.0147821624,-0.1779724061,0.1007493883,0.267157495,0.1191743761,-0.0723108277,0.0938146859,0.1856914759,0.1936353743,0.1554289162,0.0486355983,-0.0412597843,0.3124481142,-0.2527000308,-0.0651909336,0.1622196436,0.1873314977,0.1267400682,-0.2227674723,0.1186722293,-0.3079533577,0.3004409671,-0.0344962887,0.0108133554,0.0467392728,-0.3264696896,0.2752406299,0.4079024792,0.4325535297,-0.1871513426,-0.294485718,-0.1738485247,-0.1533012241,-0.0023674676,0.4001485109,0.1629086882,-0.2278741598,-0.0463881977,-0.3922587633,0.2648798525,0.0528610013,0.2153360695,0.0590297543,0.0483919531,-0.0609724782,-0.1191069931,0.1370472461,-0.0754750073,0.2253561467,-0.3740883768,-0.1097705364,0.2610442936,-0.4206713438,-0.0614455864,-0.0303055104,0.5376824737,0.5291839838,0.3535023034,0.1522078067,0.2514333427,-0.2445680499,0.0461458862,0.020478759,0.2093227208,0.0655857772,-0.0943782404,0.1245097518,0.3872912824,-0.1391385645,0.1805371195,-0.0368315205,-0.0456456318,0.6784492731,0.0181602072,0.1439567208,-0.5884661674,-0.3368144333,0.4549659789,-0.151741609,-0.1199077368,-0.0062457686,0.3078668714,-0.2831119597,0.1685542166,-0.1050597578,0.1378824711,-0.0804604143,-0.1539838761,-0.2957545519,-0.1397241205,-0.1504713148,-0.0632841587,0.2246820927,-0.3565534651,0.3352398276,-0.0515730828,0.2992823422,-0.1967782378,0.0159074217,-0.1158531681,-0.1921083927,0.2023259848,0.0298571456,0.2005928308,0.1549967825,0.0597271658,0.0133519163,0.0879230276,-0.2705900073,-0.1324612498,0.1108844131,0.2053999752,-0.5836594105,-0.393779397,0.0451772697,-0.2672811747,0.2497830391,-0.1411281079,0.0544548631,-0.109312281,-0.2154649645,-0.2687777579,-0.0199841242,0.1369310617,-0.1675184518,-0.0077661779,-0.0942448378,-0.1026165262,0.5374264717,0.2014993429,-0.0416983888,0.0285321809,-0.2054149061,0.2833956778,0.3906702995,-0.1673758924,-0.8798965812,0.259763062,-0.0273200367,0.4781271517,0.1399355829,0.1600446552,-0.0521174632,0.0658210292,0.1036422625,0.1195324734,0.0777687803,0.1902897507,-0.2474446148,-0.3442918062,0.0394470915,0.1539608985,0.0443249755,0.1407001317,0.1851332635,-0.0210841764,0.3947173357,-0.0458302759,-0.0327074528,0.0945947394,0.3341778517,0.2300528884,0.1512163281,-0.3137423694,-0.3739702404,0.3189919591,-0.2645703256,0.0681858212,0.0149742896,-0.1703307927,-0.3815482259,-0.2504169047,-0.2797917128,0.1239095032,0.0768753961,0.3769244254,0.360422194,-0.0117445569,-0.0419714786,0.3240392804,-0.0933088884,0.1179432124,-0.1803281754,0.2804037631,0.0320097543,-0.0506400988,-0.1284962445,0.0905664787,0.1922347993,0.110051766,-0.247399196,0.3598941863,0.2414757758,0.0673888624,-0.1680870503,0.2145634741,-0.1735956222,-0.1560672373,-0.030370187,0.523876071,0.1082741842,-0.002380654,-0.0148321139,0.2883145213,0.0723006129,0.0380463526,-0.0439890847,0.0857009962,0.3262713253,-0.1612974405,-0.0934867039,-0.2017478943,0.4962559342,-0.4073311985,0.1981087327,0.0379774496,-0.2952640951,-0.0991547853,0.1828975976,-0.1132807732,0.1603105068,0.1191875264,0.0249365289,-0.0544468425,-0.1000170708,0.5423449874,0.3044378459,0.2119982094,0.1004110053,0.1059395745,-0.1225009412,-0.1714084148,-0.0072758668,0.2283169925,0.2661945522,0.54904145,0.2414600253,0.0022512665,-0.409511894,-0.0301399939,-0.0552837364,0.2440394461,-0.4301237464,-0.0997081324,-0.1695824116,-0.3729839027,-0.2735750973,-0.2497572154,-0.1284107119,-0.4585259855,-0.0836521238,0.2773180604,0.0546669215,0.1180557311,-0.377423197,-0.0120092863,-0.0948292315,-0.2251132578,-0.1511166394,0.210032016,-0.1957802922,-0.0154231535,0.1645876467,-0.0427266024,0.2526283562,-0.4117342532,0.0165241696,-0.6264835596,-0.2493328154,0.0706534237,-0.1580813527,0.207870841,0.3840282857,0.1392316073,-0.2632396519,-0.0744720697,0.2209399939,-0.039336957,-0.3720573783,0.3436579108,-0.0934946314,0.014834187,0.1287033558,-0.0040662088,-0.2322083265,-0.2908137143,-0.0185987856,0.1010554507,0.2270297706,0.3653247654,-0.1820604354,0.0977607444,-0.116912134,0.4977903664,-0.2711399198,-0.5762326717,0.5287802219,0.1319577843,-0.3189797997,0.0123399375,-0.0178966019,0.0606252067,0.2794049978,-0.5194585323,-0.1023500934,-0.3720165491,-0.0292396992,0.1933085024,0.0693060234,0.1819202751,0.0034621039,-0.0545215607,-0.2112660259,-0.1410179138,0.1383913159,-0.0830824897,0.3895276487,-0.2822101712,0.295948118,0.0198221058,0.2290988415,0.5105573535,0.2102246732,0.2790658772,-0.022137912,0.0786330849,-0.1357040107,-0.2530153692,-0.3591159284,-0.0960387141,-0.2027603835,0.2613113821,0.0672068223,-0.0426020846,-0.1346848756,-0.158892855,-0.3620484471,-0.1007311344,-0.1116911918,-0.1078892425,0.2217810005,0.2286429554,0.1920685619,-0.007441333,-0.4660685658,-0.0560882688,0.1957435608,-0.2219424844,-0.1190175265,-0.1878309548,0.0527467765,-0.0519200526,0.3516172469,0.2311101854,0.3435094059,0.1457471102,-0.1631283909,0.0708224028,-0.124952592,0.2933343351,-0.3399789631,0.2645270824,-0.091392912,-0.1360713094,-0.0874281526,-0.1689927429,-0.1418350488,0.0066415421,0.197995007,0.2709618509,-0.2072632611,0.1520739198,0.1595966369,0.3525725305,-0.1252574772,-0.3678177297,-0.4135631621,-0.2623330951,-0.3545674384,-0.1133567393,-0.240335837,0.3150263727,0.0603454709,-0.0413846262,0.1093557477,-0.0521601662,0.4151341319,-0.0601407476,0.3245254159,0.0494062006,0.4183500409,-0.0234103259,-0.1006493494,0.2492771745,0.7642992735,-0.1491091102,-0.4221660793,-0.1268511713,-0.2571389377,-0.036469955,0.2482252568,-0.115008153,-0.1268873215,0.53416574,-0.1608298123,0.0693541095,-0.2341209352,0.1269024611,0.1964583099,-0.366114974,-0.3686056137,0.1963641196,0.1817590147,-0.0542320162,-0.0333044827,-0.017497424,-0.0665110201,0.340520829,-0.0597456321,0.9199489951,0.04241696,0.1259318441,0.1951045692,-0.0516662151,0.3047737181,-0.1051316857,0.0203014109,-0.3845819831,-0.3679980636,-0.1337034851,-0.2327865064,0.0154210357,0.0759782344,-0.0633779466,0.491633147,-0.1382278949,0.0897126794,0.1643395275,0.0708203018,-0.2747765481,0.0489215255,-0.0806936547,0.118889764,-0.2044604123,0.4700170457,-0.0599913634,0.0378527939,-0.252348274,-0.0325147994,-0.388263166,0.3136288226,-0.090041317,0.0068345587,0.1810777336,-0.0599018894,-0.1171378195,0.2041873634,0.3489326835,0.2161980271,-0.338876754,0.0361063816,0.6333322525,0.0288186092,0.2552764416,-0.1395580322,0.3817412555,-0.0319739878,-0.1542731225,0.0370163806,0.0370306149,-0.3310627341,-0.0266514141,-0.2256548405,-0.0555973016,-0.4416376054,-0.3367227912,-0.2361729741,0.0836619958,-0.3581354618,0.1454236209,0.2794807553,-0.1910571903,0.1048034579,0.0902285278,-0.4249240458,-0.1986413598,0.5109456182,0.1301188618,-0.2581571639,0.3898133934,0.1299090534,-0.3723615408,-0.1800105721,-0.1882393509,-0.0007799265,-0.5864365697,0.0053129415,-0.1609655619,0.1322835535,-0.1549681425,0.6723003983,0.4406674802,0.0752778798,0.032201644,-0.512088716,-0.1179987341,0.0460898951,0.285842061,0.2809039652,-0.0923958793,-0.2369232178,0.0420583226,-0.3204138875,-0.2894462943,0.027952604,-0.3091941774,-0.1901140064,-0.1275557578,0.1389643699,0.1591062695,-0.0235692039,0.1628994197,0.1089341044,-0.1183577999,-0.2295141667,-0.0552686676,0.1434463263,0.1118168309,-0.1579221934,0.1640967578,-0.1574898809,-0.1348264813,-0.1007241532,-0.1353987157,0.2522682548,0.0209836122,0.1265066713,0.4397986233,-0.0466284715,-0.1041769385,0.0049416046,-0.334520489,0.1264840215,0.0096745696,-0.0096277455,-0.0194081087,-0.0256755576,0.0981447399,0.1202702671,0.2992767096,-0.0745696798,0.3204854727,-0.5192990899,0.0408260599,0.4413575232,0.3220761716,0.3290515542,-0.3597911596,-0.1625899076,0.1890519559,0.2422227412,-0.3330812752,-0.0266632475,0.184420079,-0.1138401181,0.0457813032,0.2178548574,0.2019940168,-0.3238893449,0.3665143549,0.0521347858,0.3090809584,-0.1789855063,0.0376739502,0.4241368771,0.026364442,0.0111826751,0.0806491822,-0.1289931834,-0.2466052622,-0.0109650129,0.0656310767,0.2367641479,-0.0883421898,0.2106245011,-0.3246671855,-0.3685192168,0.5605862141,0.2961850166,-0.1180598363,0.0086457981,0.1643833965,0.411629647,0.122403793,0.4931636453,-0.4444054961,0.0553323254,-0.1762746871,-0.062582165,-0.0243954714,-0.3048769832,0.0143936044,0.1418551207,-0.2141316831,0.0696996674,0.1698767096,0.0180700086,-0.0291241333,-0.5382275581,-0.1806331724,0.0753896236,0.001889037,-0.3647792935,0.3437754214,0.0476115905,-0.074075751,0.1884774268,0.5166020989,0.3630855083,0.3304821849,0.0572223291,-0.1468572021,-0.1063451469,-0.0453084186,-0.106715247,0.7710260153,0.0892901346,0.1964202225,0.4270026386,0.12177255,-0.2061087936,-0.0785388127,-0.0745926425,-0.1373129636,-0.0980082974,0.625608027,-0.0283871889,0.3117575645,-0.1254246384,0.087138705,-0.5139369369,-0.1703628749,0.2215815634,0.1166302115,0.1453156322,0.0713299215,0.0824283957,-0.1198962331,0.2320910096,0.4119341671,0.3059858978,-0.2828147411,-0.1840687096,-0.7421238422,0.381867379,-0.1460011452,0.0460472479,0.0320160128,0.2192141265,0.0551559776,0.1389514953,0.2569927573,0.0552323535,-0.3129651248,0.1226778179,-0.0873489901,-0.174002856,0.1411593556,-0.0759258345,-0.171354115,-0.2909752131,0.200751856,0.1011782512,0.0787931159,-0.0870427489,-0.1461329609,-0.0750289261,-0.0425874889,0.1597317457,0.0204741675,0.6913182139,-0.0852561519,-0.1843104511,-0.0807620287,-0.1366753131,-0.1403537691,0.3539611995,-0.0689117759,0.337916851,0.0062353481,-0.0303952098,-0.1603726,0.0069345133,0.2029080093,-0.4047210813,-0.2947221696,0.1081530899,-0.1863506287,0.1158688217,-0.19770208,0.069220759,0.1868559122,0.3272367716,-0.1260377616,-0.3980073631,0.6327150464,-0.10716407,-0.2257052809,-0.0875340477,0.1851278245,-0.0624419339,-0.2896820307,-0.7371842861,0.052709613,0.3887307644,0.0558813438,-0.2227690071,0.1716944724,-0.2935409546,0.1815584898,-0.0190171599,0.288515985,0.1767713577,-0.1869165003,0.0473232381,-0.2415455878]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2641","title":"load_dataset(\"financial_phrasebank\") NonMatchingChecksumError","comments":"Thank you! I've been trying periodically for the past month, and no luck yet with this particular dataset. Just tried again and still hitting the checksum error.\r\n\r\nCode:\r\n\r\n`dataset = load_dataset(\"financial_phrasebank\", \"sentences_allagree\") `\r\n\r\nTraceback:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset(\"financial_phrasebank\", \"sentences_allagree\")\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 859 ignore_verifications=ignore_verifications,\r\n 860 try_from_hf_gcs=try_from_hf_gcs,\r\n--> 861 use_auth_token=use_auth_token,\r\n 862 )\r\n 863 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 582 if not downloaded_from_gcs:\r\n 583 self._download_and_prepare(\r\n--> 584 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 585 )\r\n 586 # Sync info\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 642 if verify_infos:\r\n 643 verify_checksums(\r\n--> 644 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 645 )\r\n 646 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 38 if len(bad_urls) > 0:\r\n 39 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 40 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 41 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 42 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.researchgate.net\/profile\/Pekka_Malo\/publication\/251231364_FinancialPhraseBank-v10\/data\/0c96051eee4fb1d56e000000\/FinancialPhraseBank-v10.zip']\r\n```","body":"## Describe the bug\r\nAttempting to download the financial_phrasebank dataset results in a NonMatchingChecksumError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"financial_phrasebank\", 'sentences_allagree')\r\n```\r\n\r\n## Expected results\r\nI expect to see the financial_phrasebank dataset downloaded successfully\r\n\r\n## Actual results\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.researchgate.net\/profile\/Pekka_Malo\/publication\/251231364_FinancialPhraseBank-v10\/data\/0c96051eee4fb1d56e000000\/FinancialPhraseBank-v10.zip']\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.14.232-177.418.amzn2.x86_64-x86_64-with-debian-10.6\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n","comment_length":174,"text":"load_dataset(\"financial_phrasebank\") NonMatchingChecksumError\n## Describe the bug\r\nAttempting to download the financial_phrasebank dataset results in a NonMatchingChecksumError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"financial_phrasebank\", 'sentences_allagree')\r\n```\r\n\r\n## Expected results\r\nI expect to see the financial_phrasebank dataset downloaded successfully\r\n\r\n## Actual results\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.researchgate.net\/profile\/Pekka_Malo\/publication\/251231364_FinancialPhraseBank-v10\/data\/0c96051eee4fb1d56e000000\/FinancialPhraseBank-v10.zip']\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.14.232-177.418.amzn2.x86_64-x86_64-with-debian-10.6\r\n- Python version: 3.7.10\r\n- PyArrow version: 4.0.1\r\n\nThank you! I've been trying periodically for the past month, and no luck yet with this particular dataset. Just tried again and still hitting the checksum error.\r\n\r\nCode:\r\n\r\n`dataset = load_dataset(\"financial_phrasebank\", \"sentences_allagree\") `\r\n\r\nTraceback:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset(\"financial_phrasebank\", \"sentences_allagree\")\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs)\r\n 859 ignore_verifications=ignore_verifications,\r\n 860 try_from_hf_gcs=try_from_hf_gcs,\r\n--> 861 use_auth_token=use_auth_token,\r\n 862 )\r\n 863 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 582 if not downloaded_from_gcs:\r\n 583 self._download_and_prepare(\r\n--> 584 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 585 )\r\n 586 # Sync info\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 642 if verify_infos:\r\n 643 verify_checksums(\r\n--> 644 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 645 )\r\n 646 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 38 if len(bad_urls) > 0:\r\n 39 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 40 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 41 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 42 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.researchgate.net\/profile\/Pekka_Malo\/publication\/251231364_FinancialPhraseBank-v10\/data\/0c96051eee4fb1d56e000000\/FinancialPhraseBank-v10.zip']\r\n```","embeddings":[-0.1219035983,0.2504358292,-0.0632824302,0.3291988373,0.1633206755,0.1306015104,0.0294304509,0.3981843889,0.1042740047,0.1001081914,-0.2370298058,0.251907289,0.2408214658,-0.1808868349,-0.1394717395,0.1404775232,0.1074734256,0.043175105,-0.0138186608,-0.0908906534,-0.1074014753,0.2812159956,-0.2145471871,-0.2022111565,-0.0256867241,0.0863929465,0.1104901284,0.0720111281,-0.2658267319,-0.1348887682,0.4920259714,0.1678609848,-0.0726743191,0.4778312445,-0.0001244328,-0.0519944988,0.271574378,-0.0486339889,-0.3567807078,-0.3707872033,-0.4534249604,-0.4662193358,-0.043372035,-0.2014847696,0.1378681362,0.3529549837,0.144514963,-0.1261881292,0.0545095019,0.1745295227,0.1397491992,0.2478656322,0.1467762291,0.0349865034,0.2998352647,-0.190959245,-0.1134894267,0.270352155,0.1460585147,0.0511562563,-0.2077224404,0.108807154,-0.3613056242,0.3707385063,-0.0008351873,0.0689157397,0.0588344298,-0.2889934778,0.2137651891,0.3574343026,0.4351553023,-0.218733266,-0.3650907278,-0.1239956617,-0.2084687203,-0.046321772,0.4173539877,0.1647838354,-0.283115685,-0.0629981011,-0.468390882,0.277602464,-0.0092773754,0.2345029265,0.0326497257,0.0620341748,-0.0185132474,-0.078727223,0.1915030032,-0.0618736297,0.265865773,-0.3590570688,-0.0391858891,0.2633775771,-0.4160850644,-0.1000158042,-0.0744050965,0.4421432912,0.4887327552,0.4190541208,0.1808799803,0.1608913392,-0.1823398769,0.0954611823,0.0509010777,0.1860753894,0.0430604406,0.0192637518,0.1798640043,0.4134500027,-0.1069433987,0.1626017839,-0.0692646354,-0.0316085517,0.7316473722,0.0094839688,0.2204508185,-0.6202270985,-0.3789457381,0.4553828239,-0.1631374806,-0.1740154773,0.0485575385,0.2002506554,-0.3129514456,0.1997944713,-0.0601880364,0.0835547298,-0.0794082284,-0.1734060645,-0.2265704125,-0.0589110851,-0.1227228791,0.0320249833,0.3139738142,-0.3250138164,0.3017095327,-0.0585967302,0.3189172447,-0.1349220276,0.1294149905,-0.14722161,-0.177792415,0.3243793249,-0.0608408451,0.1900531948,0.1779836416,0.0699722171,-0.1106022149,0.1508008689,-0.2157249302,-0.1449101269,0.0084858332,0.1288296133,-0.6136748791,-0.3836396337,0.074727647,-0.417157799,0.3301468492,-0.2364791781,0.0597680807,-0.1355568022,-0.2099694163,-0.2326627523,0.0224018674,0.1867602468,-0.1150826812,0.0642107129,-0.1237894073,-0.0815362334,0.537407577,0.2021664828,-0.0115692494,0.0198319256,-0.1981375515,0.1501951367,0.2272737771,-0.1579652578,-0.8371682167,0.2042014748,0.0438029431,0.5060791373,0.1131729558,0.1906835437,-0.0519395955,0.0921697095,0.1658838689,0.0820414498,-0.0265884344,0.084858343,-0.2368453592,-0.3326882124,0.1446264833,0.2096131593,0.0511388145,0.1455373615,0.1411495954,0.0321554951,0.2999107838,-0.0340311565,-0.0117764939,0.0532305762,0.4530851841,0.172077924,0.0931847841,-0.3631533682,-0.4774758518,0.2887387872,-0.1649101824,0.0705624968,-0.0916769207,-0.0877335221,-0.3856807351,-0.2757385969,-0.2165607363,0.1742729545,0.0195611566,0.3824872673,0.3753988445,-0.0675044581,-0.0457649902,0.422455281,-0.1418426484,0.1107591093,-0.2753636837,0.3411239386,-0.0082150716,-0.0534482971,-0.0830925554,0.0514111891,0.1445264965,0.0604137108,-0.2113299519,0.3504087925,0.2670076489,-0.0531184264,-0.1756032705,0.2846336067,-0.1455773115,-0.16696769,-0.0429787487,0.526586473,0.091803357,-0.0179369021,-0.0586445779,0.3515950143,-0.0240479726,0.0481224805,-0.0688220188,0.1417023689,0.3558284342,-0.1717816144,-0.1814339757,-0.1569634527,0.5086337924,-0.2532770038,0.1340863109,0.0424634553,-0.2152335197,-0.1337844431,0.1058997959,-0.1258063614,0.1081808358,0.1274549067,0.1138325185,-0.0263647847,-0.0446382761,0.497566402,0.3507782221,0.1526300609,0.1489660293,0.1117477864,-0.2153270841,-0.0640217364,-0.0453858338,0.1830278337,0.2811230719,0.5960560441,0.228765592,0.0227123145,-0.389703989,-0.163664937,-0.0485494509,0.3402167559,-0.4080109298,-0.0583371408,-0.2477993071,-0.207867682,-0.297673434,-0.3445234001,-0.2360521555,-0.4660623074,-0.1171728522,0.4518059492,0.0896143392,0.1947255731,-0.5383248329,0.01948178,-0.0496499166,-0.3143502176,-0.1638917327,0.1301636994,-0.1365332901,-0.0599620976,0.328393966,-0.0334941633,0.2559363544,-0.43783167,0.001131805,-0.5693760514,-0.3058415949,-0.0112471888,-0.095018439,0.2326525897,0.3062504232,0.1937238127,-0.2337453365,-0.2111416906,0.1895799041,-0.0036033427,-0.3972535431,0.3469931185,-0.1643625349,0.0363665074,0.1348143369,-0.0390872881,-0.1930673271,-0.2375830412,-0.020973986,0.1128812134,0.2617201805,0.3831792176,-0.1477780193,0.1182787195,-0.1250498742,0.5495764613,-0.222515583,-0.6153051853,0.6148754954,0.1510251164,-0.2769511044,0.060051091,-0.0439410582,0.1141605675,0.3257282972,-0.5195677876,-0.1178133339,-0.3152972758,-0.0033047076,0.1880886704,0.0336292945,0.1843613833,-0.0221023895,0.0029330023,-0.163847506,-0.2238503993,0.1285649091,-0.0007640535,0.4240906835,-0.2705948353,0.2751899958,0.0128684258,0.2359661609,0.5632451177,0.1702878028,0.1899769306,0.0112493979,0.128674373,-0.1215762943,-0.2859448791,-0.3453660905,-0.134869352,-0.2140858024,0.1119889021,0.0240026806,-0.0919812396,-0.1786998808,-0.1164298281,-0.2810487449,-0.1257558912,-0.0735258833,-0.1298160404,0.2713858187,0.173288092,0.1204109192,-0.0861360952,-0.4792509973,-0.0838758722,0.2706709206,-0.2183157802,-0.0733839273,-0.1870017946,0.0389393382,-0.0228251126,0.3018490672,0.2446889132,0.5021722317,0.1392087638,-0.1495168954,0.0045428798,-0.1419413388,0.2533041835,-0.2825844586,0.274772495,0.0071261171,-0.1487609595,-0.029389251,-0.1878370047,-0.1596638709,0.0067626568,0.2366042584,0.3087529242,-0.2637107372,0.1630425453,0.2867021263,0.3816690147,-0.0785213187,-0.350381732,-0.4151213765,-0.2587952614,-0.3248911202,-0.043484848,-0.1886050999,0.3431871235,0.1090604886,0.006323541,0.0852997079,-0.0185558815,0.4009565115,-0.0053486396,0.2346235514,0.0873074159,0.4733595848,0.0183095019,-0.1044885814,0.2621967196,0.7463160753,-0.1263427436,-0.3905002773,-0.0601693913,-0.2182451189,-0.0893031508,0.1572825462,-0.0247590467,-0.1962777376,0.5755331516,-0.2040039301,0.0969354212,-0.2935264707,0.060686063,0.1835066527,-0.4045857489,-0.3151387274,0.174652338,0.1504544318,-0.0043041045,0.1125732213,-0.0374620147,0.0185136218,0.1966507286,0.0452961437,0.9319778681,0.0543331727,0.0852786675,0.1511657089,-0.1225954294,0.1917185932,-0.0383278206,0.0385138839,-0.3248577416,-0.3994269669,-0.1422884911,-0.286881119,0.1556828916,0.0419328287,-0.0673466325,0.42791453,-0.0480772816,0.1722876132,0.1531153321,-0.000400114,-0.248152256,0.0684647635,-0.0544429012,0.0903516933,-0.2329608351,0.4803930223,-0.034605179,0.0329111442,-0.2976822853,-0.0550479665,-0.4033449292,0.2658832073,-0.0128097609,0.0464143604,0.3049309552,-0.0883250609,-0.098607108,0.0667986348,0.3115851581,0.0923191532,-0.337828815,-0.0497367196,0.5200307369,0.0160232689,0.2511173487,-0.1807620525,0.3796581924,-0.0339454412,-0.1229576021,0.0533333905,0.054515142,-0.415699482,0.1536203772,-0.2063109726,-0.0717368498,-0.3756573796,-0.2514221668,-0.1947635114,0.1044772118,-0.3385736048,0.1092134416,0.2938653827,-0.1672352999,0.1314901412,0.0317732655,-0.3552619815,-0.169675976,0.5675472021,0.1759825796,-0.3029550016,0.3590067327,0.0606706627,-0.4059326947,-0.1437843591,-0.1519134045,-0.1260746121,-0.5568310618,0.0036303152,-0.2260167748,0.2084529847,-0.1619078517,0.5562608242,0.4231033325,0.0605460592,0.096371308,-0.5787784457,-0.1846327484,0.0521542877,0.2902796268,0.328240335,-0.2014436871,-0.1680798382,0.0361440815,-0.2816245258,-0.2290261388,0.0165986102,-0.3228541315,-0.1437461227,-0.1001370996,0.0894568786,0.1147092432,0.0107994247,0.1213112026,0.0876619816,-0.1428938359,-0.1518722475,-0.1098080948,0.1690674424,0.1186791807,-0.1569387317,0.2470406592,-0.1336455345,-0.1350231171,-0.0956211016,-0.2308693081,0.2656267583,0.0039489083,0.1658152789,0.4067266583,-0.0667613,-0.1009629443,-0.0175443906,-0.3136405349,0.0998436958,0.085150376,0.0601071529,-0.0241164435,0.0018699007,0.1025781035,0.1002928615,0.1374376118,-0.1723908633,0.3762360215,-0.4806801975,0.0395823196,0.4147235751,0.3139627576,0.4134084284,-0.2964343131,-0.1443814486,0.2171409428,0.1816915125,-0.3661000431,-0.0208087619,0.3057405949,-0.0886870176,0.0459727384,0.2180385739,0.0981330052,-0.4027279317,0.5062954426,0.0732105374,0.2739263475,-0.2543126047,0.0885758027,0.4516227245,0.0508755483,-0.0288515426,0.1816992462,-0.1374536008,-0.1385573596,-0.053874515,0.0023309088,0.1581179351,-0.1749131382,0.2445430607,-0.3738254607,-0.331335485,0.4353528619,0.3104014695,-0.0227358956,0.0182375256,0.1479924619,0.3204092681,0.0382739007,0.4804665446,-0.4871772826,0.0208967663,-0.1371549517,-0.0031390232,-0.126435101,-0.3061621487,-0.0127937216,0.1861976981,-0.1569748372,0.0500199944,0.3224602342,0.0631819665,-0.0963125527,-0.5983762145,-0.1321250796,0.0831595361,-0.0296675507,-0.3643828332,0.348751843,0.0935672075,-0.1344633847,0.1059623808,0.4878091514,0.3832361996,0.355528295,0.1230700389,-0.0582430325,-0.1259444058,-0.0571523272,-0.0442971028,0.8251379728,0.0631786883,0.1144646555,0.3569466472,0.0808132812,-0.1470834464,-0.0094993766,-0.1333684921,-0.2202531397,-0.0671081021,0.6605429053,0.0178663731,0.2788397968,-0.0525649711,0.0635827035,-0.3988302648,-0.2211646736,0.1704478413,0.1211576164,0.0941234678,0.1547725648,0.0418642797,-0.1778076589,0.2235853374,0.3849759996,0.2284827083,-0.3141388595,-0.2495043874,-0.8176966906,0.3130852878,-0.2140867263,-0.0132976277,0.1057875678,0.2192129344,0.0326825678,0.2004255503,0.2210351229,-0.0831796899,-0.2814387083,0.1178940609,-0.1038771421,-0.1848383546,0.1755459905,-0.0894720927,-0.1763823628,-0.3068356514,0.1846025139,0.1732802093,0.0328215137,-0.0366832912,-0.1215738133,-0.0129544185,0.1075286567,0.0982147828,0.0058040638,0.7441561818,0.0030064916,-0.2014527321,-0.1133371592,-0.1136915013,-0.1094554886,0.362321049,-0.0614569932,0.37280792,-0.000126167,-0.0332113355,-0.19792445,0.1834510565,0.1207965538,-0.4166775048,-0.2454323918,0.1411975175,-0.2058520466,0.1800930351,-0.2665816247,0.0322953276,0.1959614903,0.4007237256,-0.2157139331,-0.3244048953,0.6163963079,-0.1873249412,-0.2334503829,-0.0866211727,0.1230190694,-0.22332789,-0.264515996,-0.7100596428,0.0879511014,0.4004306495,0.0140681574,-0.2292504162,0.2437381893,-0.28314659,0.1963928193,-0.0241503585,0.4023310542,0.1048548594,-0.2167249769,0.0047478341,-0.209256351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2630","title":"Progress bars are not properly rendered in Jupyter notebook","comments":"To add my experience when trying to debug this issue:\r\n\r\nSeems like previously the workaround given [here](https:\/\/github.com\/tqdm\/tqdm\/issues\/485#issuecomment-473338308) worked around this issue. But with the latest version of jupyter\/tqdm I still get terminal warnings that IPython tried to send a message from a forked process.","body":"## Describe the bug\r\nThe progress bars are not Jupyter widgets; regular progress bars appear (like in a terminal).\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nds.map(tokenize, num_proc=10)\r\n```\r\n\r\n## Expected results\r\nJupyter widgets displaying the progress bars.\r\n\r\n## Actual results\r\nSimple plane progress bars.\r\n\r\ncc: Reported by @thomwolf ","comment_length":44,"text":"Progress bars are not properly rendered in Jupyter notebook\n## Describe the bug\r\nThe progress bars are not Jupyter widgets; regular progress bars appear (like in a terminal).\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nds.map(tokenize, num_proc=10)\r\n```\r\n\r\n## Expected results\r\nJupyter widgets displaying the progress bars.\r\n\r\n## Actual results\r\nSimple plane progress bars.\r\n\r\ncc: Reported by @thomwolf \nTo add my experience when trying to debug this issue:\r\n\r\nSeems like previously the workaround given [here](https:\/\/github.com\/tqdm\/tqdm\/issues\/485#issuecomment-473338308) worked around this issue. But with the latest version of jupyter\/tqdm I still get terminal warnings that IPython tried to send a message from a forked process.","embeddings":[0.1817371696,0.0266521983,-0.0048842984,0.0565507188,0.1184258536,-0.3483625948,0.556789875,0.4173842669,-0.2593963444,0.0024573819,-0.2125666291,0.6025976539,0.2658938468,0.1672159582,0.0038400963,-0.2379226387,-0.1584529132,0.1035003811,-0.3469713032,0.0024837342,-0.2896267474,0.1918030232,-0.2716622353,0.3635055125,-0.4041291773,0.0030341123,0.0743729547,0.0767616257,0.1247892603,-0.4877848625,0.1010513604,-0.0711862147,0.0366622657,0.3372041583,-0.0001156703,-0.0285918657,0.5814056396,0.0483196415,-0.0534029566,0.3121868372,0.4811497331,-0.4962196648,0.2921688855,-0.2768397629,0.0297931377,-0.1636991352,-0.3128537536,0.1014387682,0.5365033746,-0.1144312248,0.1887732148,0.5778613091,0.0724003091,0.201487571,0.0623273849,-0.0967912152,-0.232330367,0.0826817229,0.4364378154,-0.1705191731,-0.4072221816,0.086250186,-0.0827880129,0.3567173481,-0.0489209443,0.1682757884,0.8467930555,-0.095043242,-0.3010845482,0.1756129712,0.1570870727,-0.2217915058,-0.2138753235,-0.1908809394,-0.2720872462,-0.1241789237,-0.0247892942,-0.2117275,-0.1801390946,0.0630745143,-0.3082980812,0.1876827031,-0.0752101466,0.0611595735,-0.2718208432,0.0522353537,0.0162742119,0.0222461689,-0.4251683652,0.044707261,0.2143156826,-0.0899430662,0.0952503532,0.0319860168,0.0103438683,-0.1096736342,0.1929796934,-0.2212168574,-0.1708805859,0.062981531,-0.1323730499,0.0801317021,-0.1984624714,0.1380109638,0.227278918,-0.0242731273,0.2722266018,0.031876523,0.4079745412,0.1121730804,0.0771005675,0.020398302,0.1639780104,-0.197235167,0.1264158487,0.2029946595,-0.0121485041,-0.0115512963,-0.164659366,0.3339937925,0.3537817895,0.0827571824,0.1300661415,-0.0598980151,-0.112852715,-0.0985690653,0.1420909762,0.2114596516,-0.1650141031,-0.3398205638,-0.0455710255,-0.3156957328,-0.1131617427,-0.0326119512,-0.0819021687,0.5226358771,0.2015721053,0.1167481616,-0.013367529,-0.0608451813,-0.0039571417,-0.0454569608,0.1604146063,0.2736178041,-0.1801854223,0.0645642579,0.3331306279,-0.2147260755,0.1004189998,0.4017115831,0.0019665295,0.014730881,-0.3248888552,0.0271350294,-0.3449071944,0.2925201058,-0.3482396007,-0.2347100675,0.3120609522,0.0191090684,0.0911080837,-0.4993657768,0.0704650953,-0.0417196676,0.2428245693,0.2687725723,-0.5468888879,-0.3661729395,-0.2729978859,-0.1132058352,0.5822642446,0.4652715027,0.0913209617,0.3527902663,-0.0516454615,0.0799935684,-0.0715077668,0.0040799421,-0.3240094185,-0.0506807715,0.0342350155,0.3598086834,-0.2217023969,-0.1391577423,0.5451210141,0.0268424228,0.2225217074,0.1077181548,0.2201520503,0.0050674751,-0.5182207823,0.1198147163,0.1424679011,0.053952083,0.2264510393,-0.0413967334,-0.0760639831,-0.3514195979,0.3787595332,0.4695475399,0.0613986328,-0.1032748073,-0.1235893145,0.0379323103,0.2085112035,0.2291628867,0.1162103117,0.2249444723,-0.2089492977,-0.0343402848,0.0795870051,0.1317459792,-0.4423643649,0.0988320336,-0.3059394658,-0.141505003,0.1749251485,-0.1771274954,-0.0869025737,0.2503200471,0.1665762812,-0.1593328267,-0.0878702104,0.0841876566,0.2446449697,0.6247319579,0.2298951894,-0.6264795661,0.0758993849,0.1546844691,-0.0475362353,0.1062089875,-0.0996648744,0.0255801398,0.4109326303,-0.0742413998,-0.2628404498,-0.2066325843,0.24791646,0.0921126977,-0.0527587496,0.61665833,-0.1669974476,-0.0735857636,0.2651449144,0.108077243,-0.1507932097,0.216676712,-0.1668829918,0.3541098833,-0.0477515906,0.004200194,-0.1646630615,-0.0965929255,0.0865282565,0.051981438,0.1038385108,-0.2703277767,0.2899268568,0.2455769479,0.2856424451,0.0597281158,0.1261171252,-0.108426854,-0.442605257,0.39192909,0.35090819,-0.1244852766,-0.0641749203,-0.005992013,-0.0925087631,0.1071569398,-0.1056128591,-0.1911780536,-0.2419401109,0.2931978405,0.0717580467,0.2813779414,-0.0015513499,-0.0976843238,-0.1878172606,0.1744771004,-0.2419845313,0.0416071415,-0.2596999705,-0.1773075908,-0.2635091245,-0.0708165392,0.0574664697,-0.3950704634,-0.1241569594,-0.2460366786,0.0640075579,0.2776843011,-0.0457182713,0.4009768069,0.4703310728,-0.1269136071,0.2192731798,0.3190117776,-0.2036365867,-0.2075867504,-0.1263897866,0.0597334765,-0.0034549914,-0.0828465149,-0.0726938695,-0.4586654902,0.2137558609,-0.2816906273,-0.0525495075,0.0743593872,-0.0355221555,0.5063041449,-0.2617628574,0.2529340386,-0.1107394397,0.4106066525,0.4045543969,-0.2005270422,-0.1289593428,0.0199467838,0.0778892413,-0.2607973516,-0.0393591039,0.3344100118,-0.2608054578,-0.1424450278,0.2783357501,0.0502967797,0.169492811,-0.5509198308,-0.3159068227,0.0516400151,0.2530952692,-0.1564558446,-0.0282126497,-0.222311452,0.2857842743,-0.2854257226,-0.2039786428,-0.1333153844,-0.0329735503,0.021691991,-0.6995676756,-0.3913334608,-0.016667543,-0.151509434,0.5323093534,0.0880490392,0.0253739655,0.027831845,0.0030348997,-0.0044203573,-0.0663749129,-0.2988654971,0.0168136694,0.0403070152,0.3351403773,-0.0734657943,0.0413795263,-0.1678373963,0.2677576542,0.1905387193,0.0943192318,0.0324877612,-0.0937042907,0.0733191222,0.0126712257,-0.0974273905,-0.1152827963,-0.1858903766,-0.0731040388,0.063105531,0.1063179895,0.1593655497,0.1933678389,0.0347987041,-0.1947152913,-0.2861011624,-0.1182760298,-0.2299228162,-0.1167777628,0.0335355252,0.331215471,-0.0719773769,-0.2232434899,-0.1226547956,0.1832716316,0.4686769247,0.0486777388,0.3824283481,-0.1621936858,-0.2085372508,-0.0009658141,0.1433832943,0.0303803757,-0.0637195334,0.0102410028,0.2103758007,0.1903972179,0.1864861846,0.0376273729,-0.1315955818,0.4868620038,-0.1329531968,-0.5907799602,-0.4485073686,-0.083837226,-0.0461817309,0.4464483261,0.2025932819,0.1214461103,0.1832035184,-0.1629596055,0.1613380462,-0.440679431,-0.4085015953,-0.5296405554,-0.0772080421,-0.1208097488,0.101293005,0.0220485777,-0.2823099494,-0.3443133235,-0.4439450502,0.151032269,-0.2343680859,-0.203864187,0.4826022983,0.1618641466,-0.3691056371,0.2843438089,0.1946534812,0.0703483447,0.2092734128,-0.1213418022,0.1955966204,0.1174235046,0.2129205465,-0.453571409,-0.0542346835,-0.0327297859,-0.0540474355,0.269675374,-0.3494567871,0.1118169352,-0.0713969916,-0.2171407789,0.329379797,-0.1024744138,0.1040697843,0.251262784,0.0664817095,-0.1387920827,-0.0392667502,0.064806141,-0.2958228588,0.2502632141,-0.392426759,0.2538971901,0.4127412438,-0.3073870838,-0.1985529512,0.1135265231,-0.1008401066,0.0674055517,0.3842746317,0.0769118741,-0.4579051733,-0.1355749667,0.0896987543,-0.1970648468,-0.0743465647,0.0143761365,-0.2643181086,0.0920367166,-0.3580286503,-0.0003541387,0.142257452,-0.1866117418,0.4732814729,-0.0945192277,-0.2393090725,0.1698595881,-0.1792891175,0.3100373149,-0.0778423548,0.0323348418,-0.2217236161,0.2593061328,-0.324244976,0.1942670196,0.0868859887,0.3065232933,0.6779507995,-0.2582505047,-0.478056252,-0.1305696815,0.0957541838,-0.0298391953,0.0410921536,0.1004782617,0.2653992772,0.1224285662,0.0419644639,0.2340957224,0.1169517115,-0.2659300864,-0.3217813969,0.230324626,0.0165103897,-0.4008854926,-0.1908876747,0.2124530375,0.5923561454,0.0772768185,0.6071106791,-0.170869112,0.0938695967,-0.2045491636,0.1392642409,0.2038924992,-0.1019264236,0.6167327166,-0.2241156399,-0.0105039477,-0.0011596089,0.0155340517,-0.1059429869,0.2463877797,0.13463898,0.2571750581,-0.2707896829,-0.0400446802,0.390444994,0.0253423881,-0.0369004197,0.4749502242,-0.2261115909,0.2843965292,-0.3062983751,0.204947114,0.3042432964,-0.1515857726,-0.4816409349,-0.1707336605,-0.5891593695,0.0861815661,-0.0511260107,-0.1754157096,-0.0631226078,0.5393032432,-0.0343938582,-0.1166118979,-0.2260076255,-0.0027782065,-0.1965398341,0.1588025987,0.3093067408,-0.1977652907,-0.3245773315,-0.1785841286,-0.050726153,0.0764013901,0.0561276637,-0.1820083559,-0.2545235157,0.0841501728,-0.0078679845,-0.2569127381,-0.0423270091,0.1870150864,-0.0193927884,-0.0605703592,0.1704659015,0.0101391189,0.0925533399,0.2859082222,0.0786387697,0.1287005246,-0.1340811253,-0.0895946473,0.1099940091,0.5431039333,-0.0363829434,0.1304314286,0.2100278288,0.2070265859,-0.0093542328,-0.084357895,-0.3582532108,0.257576257,0.369329989,-0.2755759358,0.1990129352,-0.3448761404,0.2791692317,0.2967947423,-0.0382647552,-0.4296239614,0.1247139648,0.160326317,-0.0035055142,-0.1456474811,0.1596736312,0.3591039777,-0.0659685656,0.23620902,-0.1499637067,0.0523925126,0.1445040703,0.0649675205,0.0041035349,-0.1538078934,-0.2074184418,0.3424953222,-0.0387569666,0.2916079462,0.2955338657,-0.1797022969,-0.2489405572,-0.2290108055,0.0882855356,-0.1418279707,0.1646235287,-0.0915948451,0.0886861533,0.2083358765,-0.3614554703,-0.2503944039,-0.0059378115,0.0110781705,0.0686561987,0.1550808847,-0.1648520827,-0.2216762155,-0.3788727224,0.0777991638,-0.0932442918,0.0874356478,-0.1769072264,-0.0204309896,-0.0972802937,0.0688078031,-0.1281937212,0.3972241879,0.2002463192,-0.0158652551,0.0069478848,-0.3464375138,-0.2009720355,0.1060384363,0.2996792793,-0.0015275384,-0.108637549,0.3010883629,0.2425166368,-0.269839555,0.1807521135,0.2070456296,0.330827713,-0.1841491163,-0.5122314095,-0.0871947706,0.0147968708,0.07656046,0.3983009458,0.0104474034,-0.2393776923,0.037204463,0.0716233253,-0.1399573982,0.1981387287,0.0323799178,0.2398914546,0.1445452273,0.5007633567,-0.2323464155,-0.2111147344,0.1001094356,-0.3531180322,-0.1419487447,0.0075195991,0.2385790199,-0.0802276582,-0.0238477588,-0.14158687,0.0790749788,-0.4065202177,0.3623471558,0.3410251439,0.3175618947,-0.0254955441,-0.0492038354,-0.3845283091,0.3794111609,-0.1781532466,-0.3013525009,0.2541190386,0.0433156677,-0.2651920021,0.3701040745,0.7036457658,-0.2908414304,-0.5082321763,-0.3333394527,-0.2769772708,0.0086110411,0.3501788676,-0.2469127178,-0.0890448913,-0.0982787088,0.2765984535,-0.0749764442,0.1320375055,0.1540165693,-0.1242890954,0.0688353553,0.5911547542,0.0691591874,0.2064617425,0.1213232279,0.0442968942,-0.131938085,0.1508338898,-0.345410347,-0.0218686219,0.3271324039,-0.3355371356,0.4607938826,0.2721303999,-0.0749800056,-0.3710611165,0.4132907093,0.2208662033,-0.004890745,-0.0903680399,-0.093483232,0.0511622019,-0.0792129189,-0.0652787164,0.1165084094,-0.0506937616,0.4800716639,-0.0764005035,-0.1664082855,0.2783628106,-0.183829084,-0.3798915148,-0.1245586649,-0.0217788797,0.0725362897,-0.009623616,-0.0441760309,-0.3208202422,0.2267317921,-0.1420073658,-0.28459391,0.1216770783,-0.1535632312,0.0717051253,0.0470278598,0.1328727603,0.1103113666,0.0145060429,-0.1096885055,-0.0973096192]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2630","title":"Progress bars are not properly rendered in Jupyter notebook","comments":"Hi @mludv, thanks for the hint!!! :) \r\n\r\nWe will definitely take it into account to try to fix this issue... It seems somehow related to `multiprocessing` and `tqdm`...","body":"## Describe the bug\r\nThe progress bars are not Jupyter widgets; regular progress bars appear (like in a terminal).\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nds.map(tokenize, num_proc=10)\r\n```\r\n\r\n## Expected results\r\nJupyter widgets displaying the progress bars.\r\n\r\n## Actual results\r\nSimple plane progress bars.\r\n\r\ncc: Reported by @thomwolf ","comment_length":28,"text":"Progress bars are not properly rendered in Jupyter notebook\n## Describe the bug\r\nThe progress bars are not Jupyter widgets; regular progress bars appear (like in a terminal).\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nds.map(tokenize, num_proc=10)\r\n```\r\n\r\n## Expected results\r\nJupyter widgets displaying the progress bars.\r\n\r\n## Actual results\r\nSimple plane progress bars.\r\n\r\ncc: Reported by @thomwolf \nHi @mludv, thanks for the hint!!! :) \r\n\r\nWe will definitely take it into account to try to fix this issue... It seems somehow related to `multiprocessing` and `tqdm`...","embeddings":[0.0840747207,-0.0174471959,-0.0553433672,0.146748662,0.1418051273,-0.2826747596,0.4365513623,0.3081710339,-0.2813870907,0.1335172951,-0.1734683812,0.4958986938,0.302113533,0.360643357,-0.0934584588,-0.2819553018,-0.0838802904,0.1640948504,-0.3097079098,-0.0013430968,-0.3553539813,0.2289115191,-0.3600171804,0.274898231,-0.3260476887,-0.0157691464,0.0067369333,-0.0339760669,0.2025540769,-0.3707706332,0.0603209287,-0.0869974941,0.1291181743,0.4368245304,-0.0001134229,-0.1087477282,0.4875665009,0.0413400717,0.0329009667,0.2878982425,0.4559450746,-0.4877927303,0.2829349339,-0.2850020528,0.0710600391,-0.1794690639,-0.3954918087,-0.0420616306,0.406507194,-0.1567408293,0.2473642379,0.4405182004,-0.0182326809,0.2076785713,0.0391099975,-0.1461219341,-0.2083894163,-0.0282222517,0.4653917551,-0.3063619435,-0.3879884779,0.1063967794,-0.0486431532,0.4187018573,-0.0393807665,0.2133056968,0.9913772941,-0.0878651738,-0.2944223583,0.2115101814,0.2394176275,-0.105977878,-0.1512372196,-0.1202547997,-0.2735105455,-0.1160391644,-0.1966740787,-0.0861075521,-0.2104892582,0.0214731023,-0.3254862428,0.1474572867,0.0230799317,0.0039320379,-0.2227223516,0.0638749674,-0.0164266899,-0.0254160725,-0.4152092636,0.1107601821,0.1686975807,-0.1647108644,0.1662720144,-0.017423898,-0.0668758824,0.0514467619,0.1705864519,-0.0394774675,-0.1951253414,-0.0893868878,-0.0934441686,0.1065666303,-0.1109584719,0.1176110208,0.1380227953,0.0363824181,0.2670516074,-0.0737986863,0.4157478809,0.0265655257,0.0262287073,0.0510558039,0.2824546099,-0.2843369544,0.1190832481,0.2696622014,-0.1699028164,-0.0240198709,-0.1325654685,0.2839515805,0.3845815063,-0.1172237247,0.050101649,-0.0212007668,-0.1604377031,0.0458948351,0.0982232392,0.2016148567,-0.093886435,-0.2694262862,-0.1279253215,-0.2795524895,-0.2595213652,0.0063250475,-0.1258066744,0.3609222472,0.2234245986,0.1400429904,-0.0556914099,-0.1958279908,-0.0410629772,-0.0504104123,0.1032537669,0.2466241866,-0.0829610452,0.1367767453,0.2960921824,-0.2910871506,0.1265167892,0.425904572,-0.0393438265,0.0393931307,-0.3534792364,0.0832789913,-0.3191151619,0.3744330108,-0.4555279315,-0.2263981551,0.3937739134,0.0776729807,0.0967461243,-0.4163134694,-0.0178411622,-0.1785806417,0.17323713,0.1455981731,-0.5406771898,-0.3143543601,-0.0999097899,-0.1438417584,0.6745646596,0.3311755657,0.0616390258,0.3111010194,-0.0355523303,0.1803805232,-0.0299432483,-0.0998076946,-0.2703015506,-0.0446333215,0.0686812028,0.3434061408,-0.1168109104,-0.1047256365,0.387932837,0.0571679473,0.1605418772,0.1995611638,0.2249500453,-0.0040507358,-0.5000802875,0.1472550035,0.2613554001,0.1368755549,0.2515562475,-0.10519059,-0.1100691408,-0.4471721053,0.4548561275,0.3896095753,0.0735659003,-0.1086159647,-0.2065193206,0.0486687124,0.20959723,0.2003743201,0.0501533598,0.2725409865,-0.2228266299,-0.1505789608,0.1081631556,0.0841673166,-0.2802225947,0.0574537516,-0.3637309074,-0.2123386711,0.2303520441,-0.0214926694,-0.0324373543,0.1609050781,0.1928894222,-0.0381628126,0.0449481718,-0.0241240729,0.2177133262,0.6586594582,0.220101133,-0.5620321035,0.1005694792,0.1098099351,-0.0100126024,0.0599663928,-0.1167561784,0.105545871,0.3706769645,-0.0126156434,-0.2862404287,-0.2531236112,0.1941245794,0.1476222426,-0.1142541096,0.5261477828,-0.058354076,-0.1218652278,0.3449297249,0.087261498,-0.1932433844,0.3505972326,-0.133608982,0.2573168278,-0.0148074897,0.016546404,-0.0542389229,-0.1458518058,0.137533918,0.0343477614,0.0556813329,-0.2651257813,0.3374623358,0.2101247609,0.2718619108,0.0949251875,0.1694935709,-0.1708599627,-0.4523812532,0.3870063424,0.3288426995,-0.1260956228,0.1076468676,0.0887194499,-0.0556902587,0.0388398618,0.0346243121,-0.2729830742,-0.2650905252,0.2690735161,0.2852129042,0.3415223658,-0.0553954542,-0.1524236351,-0.0843226239,0.1791528463,-0.2676634192,0.0621223412,-0.1503647715,-0.216901809,-0.2802102268,-0.0814332962,0.1066746041,-0.2798359394,-0.0973197892,-0.2014974654,-0.0294565782,0.1356530339,0.0548096485,0.3602169752,0.5633695126,-0.1225382686,0.140152514,0.3674173951,-0.184812814,-0.2312560081,-0.2094030827,0.0796075463,0.0218693707,-0.0270249415,-0.1452461034,-0.5066241622,0.2695992291,-0.2906699181,0.0424460471,0.0552971922,0.0010296454,0.4239055812,-0.311807543,0.1914335191,-0.0486043878,0.4684601724,0.3779620826,-0.1273926646,-0.1681383252,0.0822190493,0.0849710479,-0.2132692188,-0.0504145734,0.3898370266,-0.331717968,-0.1092651263,0.2391228974,-0.0295659266,0.1968768239,-0.7200556397,-0.2799727619,-0.0238183253,0.2932691574,-0.0693960413,-0.1043622568,-0.1313590556,0.2333750129,-0.2860134244,-0.2543158233,-0.2362781763,-0.0590562932,0.0311986003,-0.6471823454,-0.4120875895,-0.0859325826,-0.2004361153,0.3781549633,0.0382053107,0.1749142706,0.1584655195,0.0033087095,-0.0538710952,-0.0849443451,-0.4036501348,0.0524333604,-0.016965311,0.3014837205,-0.0230090097,-0.0270725153,-0.1710710675,0.1382593215,0.2992339134,0.0753851905,0.1574761719,-0.0522209629,-0.0141519243,0.1012698188,-0.0767275617,-0.169376418,-0.1942314506,0.0478077605,0.1227429286,0.0846474916,0.0652763098,0.1389370114,-0.0133855026,-0.3803442121,-0.1787613779,-0.1553856134,-0.1738344431,-0.0848157555,0.0064560538,0.2819902003,-0.0114201736,-0.2094163597,-0.1050023735,0.1878012866,0.3450337648,-0.021855576,0.4383287728,-0.2292405516,-0.1616394967,0.1673971266,0.1712918878,-0.0219538417,0.0534426868,0.0723043382,0.1686126739,0.1126676127,0.345429033,-0.1313160807,-0.1081416011,0.4399821758,-0.1932770163,-0.559992969,-0.5287843943,-0.1388813853,-0.0271135122,0.3023209572,0.3238762021,0.1704627275,0.129438296,-0.305863291,0.1342191547,-0.4507177472,-0.5005689859,-0.5428311229,-0.0320379026,-0.0099370442,-0.0275907852,-0.0984961912,-0.1000111476,-0.3013594449,-0.4653050303,0.1186718568,-0.2689271569,-0.1494286358,0.433070302,0.159185186,-0.2284727842,0.1799621135,0.1587318182,0.2233385891,0.1399243772,-0.1881383508,0.001787998,0.2173331827,0.258693397,-0.4733978808,0.0423986576,-0.0806371123,-0.1161955222,0.2704633772,-0.4603219926,0.2149654776,-0.1570174992,-0.2689239681,0.3414726555,-0.04832194,0.013210671,0.2080332041,0.1232946739,0.0050048572,-0.1137180626,-0.0689207762,-0.285761714,0.213022843,-0.3701871037,0.2292711586,0.3312902451,-0.3805872798,-0.1750418693,-0.0680916235,-0.2268899381,0.1177392006,0.3030438721,0.0624668971,-0.3827464283,-0.2603964508,0.0825034454,-0.2630774379,-0.0258884877,-0.0040114694,-0.2764795423,0.2061020434,-0.4202793241,-0.041172266,0.1308260709,-0.1384242475,0.6062145233,-0.1045560092,-0.1399495602,0.1635837555,-0.2762543261,0.3311765492,-0.129890725,0.0665112957,-0.154028371,0.2529998422,-0.334065944,0.1621988714,0.0469716527,0.3277939856,0.5213476419,-0.2467439175,-0.4642066956,0.127959922,0.1828425527,-0.0753348246,0.080231525,0.0282323379,0.1075994521,0.1536436528,-0.0152189806,0.2035623491,0.065937534,-0.3426293731,-0.3608886003,0.2768605351,0.0544736423,-0.3398239017,-0.158008039,0.2512802184,0.515322268,0.1659759432,0.5921123624,-0.0823672414,0.2200642526,-0.231007129,0.1685484052,0.2138753235,-0.2118781805,0.7508046031,-0.206768319,-0.0217721723,0.0426451899,0.0263492931,0.0704070702,0.12680237,0.067162469,0.3412512541,-0.2547578514,-0.0963543057,0.2661041021,0.1276560277,0.0116728861,0.4833355248,-0.2220047116,0.2646160424,-0.2258884311,0.3405117691,0.2906493247,-0.0099985721,-0.4293839931,-0.2461832464,-0.5078002214,0.0547656603,0.0729146078,-0.1993194818,-0.1020386145,0.4341852665,0.0094320737,0.1235913113,-0.2600487173,0.04824806,-0.2103758901,0.1416706592,0.2156020254,-0.1273956597,-0.3874331117,-0.1020736098,0.0111418162,0.0294885319,-0.024992587,-0.2505482137,-0.2373582721,0.0590547882,-0.0618388206,-0.1189650893,0.0829755515,0.1766029298,-0.0137001313,-0.0503276139,0.1899188757,0.0070657958,0.0427719057,0.3488994539,0.1346112639,0.116541557,-0.1515337825,-0.060568992,-0.0059830341,0.5434562564,-0.0660191178,0.1043648198,0.3075919747,0.1612798125,0.0489343926,-0.2327803075,-0.3188105524,0.2774234414,0.2623284459,-0.305025965,0.2351172417,-0.3422241211,0.3326387703,0.1437485814,-0.0338615924,-0.2724226415,0.1375377774,0.2036044896,-0.0683354586,-0.2091984451,0.2082445174,0.3461470902,-0.0081881676,0.3194356263,-0.1421828121,0.0462906882,0.1024083421,0.0240945835,-0.0675438568,0.0091027301,-0.2059791535,0.3505121768,0.065663591,0.2744111121,0.2748164535,-0.0593207106,-0.3289284408,-0.1768021286,0.1106421202,-0.0434987471,0.0747662932,-0.0942509398,0.0331626534,0.1554329842,-0.1741340458,-0.1245347559,-0.0719827563,-0.0183950998,-0.0628900602,0.2263251543,-0.1679671705,-0.2615422308,-0.3279536664,0.0831791386,-0.0947097242,-0.0216262024,-0.0767756104,-0.0971215069,-0.0849745274,0.0793215781,-0.1062789783,0.3211348951,0.3122420013,-0.0279705245,-0.0099640507,-0.1597609371,-0.2634783387,0.1258375645,0.3141278327,0.001135607,-0.176520586,0.2330850214,0.2390505821,-0.3890814781,0.210318774,0.1597009599,0.3547718525,-0.2991580963,-0.5115677118,0.0465826616,0.0431716144,0.074561581,0.3577716947,0.1426044554,-0.2597827613,0.0522157699,0.0783246085,-0.1751221865,0.2820823193,0.0768240988,0.3440591395,0.1252899021,0.449524641,-0.2704695761,-0.1520071775,0.1030547619,-0.2561652064,-0.2923443913,0.0472157486,0.202203393,-0.14733769,-0.0348979607,-0.1349245161,0.1029419079,-0.396171242,0.3771845102,0.2636162639,0.4115664959,-0.1070389375,-0.1958614439,-0.3908257186,0.3388753831,-0.1304858476,-0.2475671917,0.1263137907,0.1302093714,-0.2292502075,0.3551634252,0.770339787,-0.1849017739,-0.4258206189,-0.1495705396,-0.4036408365,0.0335611403,0.317378372,-0.2314184904,-0.0433621518,-0.0700697005,0.2427544594,-0.0539348163,0.2154805958,0.053141728,-0.0059425174,0.0582594648,0.3602853119,0.1891822964,0.1174982041,0.0260107126,0.0147798359,-0.0830277577,0.3153704107,-0.1140944511,0.1464644074,0.2588916719,-0.2921168208,0.3584014773,0.1357862502,0.1304063201,-0.3719023764,0.3412769139,0.273438096,0.1059357226,-0.1206761375,-0.0670828596,0.2311580479,-0.1974951476,0.0351051949,0.102952525,-0.1211241111,0.3408853412,-0.1112642363,-0.0985409915,0.2666783929,-0.1712308377,-0.4160819948,-0.1571352333,-0.0429105386,0.0663789213,0.047920689,-0.0647944808,-0.3053124547,0.2395626456,-0.0582254305,-0.3515852392,0.14495942,-0.325643301,0.0265676007,0.0682527199,-0.0356904119,0.1399087757,-0.0035144549,-0.1854761243,-0.1822355837]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2629","title":"Load datasets from the Hub without requiring a dataset script","comments":"This is so cool, let us know if we can help with anything on the hub side (@Pierrci @elishowk) \ud83c\udf89 ","body":"As a user I would like to be able to upload my csv\/json\/text\/parquet\/etc. files in a dataset repository on the Hugging Face Hub and be able to load this dataset with `load_dataset` without having to implement a dataset script.\r\n\r\nMoreover I would like to be able to specify which file goes into which split using the `data_files` argument.\r\n\r\nThis feature should be compatible with private repositories and dataset streaming.\r\n\r\nThis can be implemented by checking the extension of the files in the dataset repository and then by using the right dataset builder that is already packaged in the library (csv\/json\/text\/parquet\/etc.)","comment_length":20,"text":"Load datasets from the Hub without requiring a dataset script\nAs a user I would like to be able to upload my csv\/json\/text\/parquet\/etc. files in a dataset repository on the Hugging Face Hub and be able to load this dataset with `load_dataset` without having to implement a dataset script.\r\n\r\nMoreover I would like to be able to specify which file goes into which split using the `data_files` argument.\r\n\r\nThis feature should be compatible with private repositories and dataset streaming.\r\n\r\nThis can be implemented by checking the extension of the files in the dataset repository and then by using the right dataset builder that is already packaged in the library (csv\/json\/text\/parquet\/etc.)\nThis is so cool, let us know if we can help with anything on the hub side (@Pierrci @elishowk) \ud83c\udf89 ","embeddings":[-0.4575762749,-0.0447009094,-0.0421369784,0.1991639733,-0.1262193918,0.1635166407,0.3772648275,0.1787040085,0.413285464,0.1388915181,-0.2391807884,0.3387723565,-0.0637952238,0.5494150519,0.2678030729,0.173187837,0.1355401725,0.3373374045,0.0156680774,0.1087146997,-0.147266522,-0.0281416457,0.0509655774,-0.3055997789,-0.4166117907,0.1337530464,-0.0574752875,0.5216122866,-0.2011022568,-0.2119361311,0.4005629122,0.5012663603,0.2502647936,0.1841825992,-0.0001212054,0.043409273,0.1928070486,-0.1468495429,-0.3016461432,-0.1539461315,-0.2993468642,0.0231876429,0.2855739892,-0.1010636985,-0.0823806301,0.2093599439,0.1571625918,-0.1291951388,0.3677399457,0.0976226702,0.0166193526,0.2999813855,-0.259234041,-0.338388443,0.1060967371,0.5227719545,-0.0501799621,0.0753658265,0.446670562,0.1501270384,-0.2684879899,-0.0616010539,-0.0511127971,-0.1306405514,0.6978545785,0.0141173322,-0.4139381647,-0.3369881511,-0.0558858179,0.2696109116,0.4849348664,-0.2226595432,-0.1884816736,-0.5040717721,-0.1010899246,-0.023111308,0.1161088943,0.3829921782,-0.0745234787,0.3300855458,-0.1943240762,-0.4995049238,-0.2894499898,-0.0394161791,0.087032862,0.0510337315,0.0833544731,-0.1139840186,0.2059298754,0.0716915876,0.3125401437,-0.1525991857,-0.1997705102,0.184158653,-0.2591173649,-0.1511709541,-0.1909290105,0.3544423282,0.2726554573,0.4776968658,0.0313507691,0.2450513095,-0.4523226619,0.0596118532,0.3471122384,-0.0049101491,0.1210729033,0.0548969321,0.3135947287,0.2362887412,0.3534661233,-0.0196903087,-0.1949664801,0.0806540176,-0.4910582602,-0.1728901565,0.1783725023,-0.085036315,-0.0836664587,-0.34614259,0.5396619439,0.0528207645,0.2299038172,0.2829740047,0.136417225,0.1290995926,-0.2617965043,0.1894248277,-0.1174897179,-0.2959381342,-0.1057050675,0.1344973892,0.1033403873,0.2694162726,0.3688505292,-0.3593107462,0.1102583632,0.0069277943,0.4422927499,0.0437481254,-0.0864408463,-0.0130618494,0.0163177587,0.1432943791,0.3748804033,-0.0535977893,-0.0228161514,-0.2377948761,-0.3192485273,0.0625885203,-0.4418234229,-0.5183890462,-0.0569747314,0.0121524492,-0.4268384278,0.0280668084,-0.6266530752,0.0605005175,-0.399880439,-0.0416647047,-0.0795237869,0.2116308361,-0.0123844482,-0.0399878994,0.1320928931,0.5962692499,-0.2625654936,-0.1146187037,-0.0567253903,-0.2146859616,-0.3286296725,-0.0449763387,-0.1742637306,0.1860437542,-0.2278230935,0.3033220172,0.4498622715,-0.5796934962,-0.2853997648,0.0064900233,-0.1179377511,0.0732642263,0.1743253171,0.3771463633,0.3957131207,-0.1296545267,-0.1318271756,0.52608639,0.0008256844,0.0909401029,0.1459450424,-0.5183837414,-0.1579161435,0.4319950044,0.1243649349,0.0304883625,0.3115407825,0.2013223916,0.1387485564,-0.2980290055,0.1430205852,-0.101116322,0.0254287664,0.0401319154,-0.068433471,-0.0548522025,-0.6891098022,0.3163145781,0.0964968652,-0.0020367806,-0.0244451053,-0.3710264266,-0.1171226948,-0.1898974627,-0.2818386555,0.0634649992,-0.0909614712,-0.0830148682,0.0350424983,-0.135661006,-0.5921099782,0.6018566489,-0.0149240941,0.1960612983,-0.653110981,0.4063933194,0.3386701345,-0.0153757492,0.2168535143,-0.0285904426,0.1245829239,-0.1221515909,0.1126259491,0.2599679232,0.125773564,0.6147466898,0.2605298162,0.1702958643,0.1483879089,-0.0583523586,0.2140933722,-0.348982662,0.012365656,-0.1390500367,-0.3719674349,0.3414103091,-0.4983025491,0.1190770417,0.0701576546,-0.0204376485,0.0052312729,-0.0468474254,-0.1861120313,-0.1877366155,-0.0455209725,0.104294844,0.1593873799,0.0439118966,0.0219548531,0.02188785,0.4020877182,-0.0820777938,0.0510959625,0.1745688617,-0.1587085277,-0.0367911123,0.1882548481,0.0600359589,0.2340117991,0.1528205723,-0.066654548,0.2553109825,0.3917658329,-0.2559643984,-0.0063171205,-0.0388427414,0.0046937261,0.1148881465,-0.1143906042,-0.1490908116,-0.3427891135,-0.0038192514,-0.0566469505,-0.1197237745,-0.202475369,-0.1974553764,-0.1885620356,0.0441672839,0.1086140275,-0.1390506774,-0.2280121893,-0.0474897027,0.1132057905,0.292614758,0.1356831193,-0.1808515936,0.1021874547,0.8752858639,-0.2077578604,-0.4260379076,-0.2744226754,-0.0052179634,0.2112582475,-0.0357649997,0.2536873519,0.3942889273,0.2186065614,-0.2756740153,0.2006195486,-0.3470382094,0.1797520369,0.1969391853,0.0370648652,0.0286822394,0.1138853133,0.078886807,0.2208523601,-0.0224462468,-0.1317495555,-0.0986444578,-0.0600061901,-0.1949043125,0.0839634165,-0.0586361997,-0.1458729357,-0.1207059771,-0.367988646,-0.4563333094,0.3621260822,0.3808919191,0.027583871,-0.2261142433,-0.0948191062,-0.0128289461,-0.1683928072,-0.2992093563,0.0453890376,-0.5231456161,0.4055296183,-0.3415705264,-0.2106558979,0.2707829773,0.0609666407,0.1544732302,0.0539418235,-0.3392732143,-0.2430040538,0.0532311946,0.08949285,0.0398268215,0.065090932,0.2832045555,-0.1102572456,0.2209951878,-0.0620181859,-0.0730815977,0.088278912,0.2928557396,-0.3506483436,0.1858042628,-0.120524548,-0.1198292375,0.5136722922,0.2791596651,0.2512265146,0.1940324903,-0.0187380556,0.3876357377,-0.3456962407,0.0017486442,-0.1830449402,0.038366437,-0.0451302975,0.4778506756,0.2628038526,0.1216737702,-0.3591516316,-0.4306385517,0.0288198814,-0.1977297217,0.0783412829,-0.0111111626,-0.0104182027,-0.1670587659,0.02052336,-0.1251323223,-0.2814126015,0.0326022245,0.3820087612,0.3374103606,0.1487733573,-0.1922379434,0.0919733122,-0.291688174,0.2229800969,-0.0234763976,-0.3273980916,0.0144211436,0.0260607358,0.043443948,-0.0784057751,0.5525973439,-0.076823622,-0.0256261975,-0.2390150428,-0.2209435701,-0.1558260322,0.099499844,0.13804923,0.0132212332,-0.4880362153,0.7941618562,-0.0995156989,-0.3218313456,0.052417025,-0.0858286992,-0.186013341,-0.1060685366,0.1103658453,-0.0785629973,-0.4641357362,-0.2623743713,0.0945599303,-0.1030065119,-0.2374695539,-0.1350483149,-0.0629906133,0.131077826,0.1900197864,0.1737861335,0.2542228997,0.0578282587,-0.1312521994,0.4289519787,0.0591998696,0.2980826795,0.6782125235,0.1396740824,-0.4402035177,0.0134959929,-0.20106408,0.292177707,0.5223480463,0.0938640907,0.0564075522,0.0585680827,0.0362378173,-0.5819572806,0.1477123201,-0.0381950401,-0.2752726972,-0.4304094613,-0.4510777891,0.6226552725,-0.0689187199,0.0640258566,-0.1069794819,0.4030180275,-0.377048552,0.1231429726,0.0702426657,0.8982323408,-0.0711545944,0.2976281643,0.0360319428,-0.0640621483,0.3835209608,-0.6910941601,-0.0053561637,-0.3164266348,-0.1875433028,-0.3519823253,-0.0240890458,0.1682009697,0.2223049551,-0.2135065496,0.3743098378,0.0577208921,0.4703815281,-0.044575654,0.4272412062,-0.3970298469,-0.5239443779,-0.1813053787,0.0216596145,0.0262456592,-0.0213817097,-0.0455973074,-0.1437030882,0.0020615486,-0.1098015904,-0.0077887373,0.14023754,0.2143561393,-0.2197984159,0.0066357478,-0.0144395251,0.3113361597,0.3582701087,0.2158787698,0.0300085992,-0.3226819336,0.2304143161,-0.2172023356,-0.0157416053,-0.1394653767,-0.0272873063,0.2691299915,0.0546619445,-0.1156827435,-0.0498544984,-0.0464590304,-0.3997415006,-0.1941818297,0.1144541129,0.1813288778,-0.2519482076,-0.1360103488,0.012785499,0.0661640167,0.0536811762,-0.0303368792,0.0336856358,0.0313881598,-0.0099908598,0.1384607404,-0.1035657674,-0.0863178372,0.1303190738,-0.285409987,-0.325083226,0.0819170773,-0.063093476,-0.0948143974,-0.1671719253,0.196656689,0.4302743077,-0.3918291628,0.1148462296,0.0370115265,-0.0390540361,0.0038853923,0.0323107056,-0.039097771,-0.1722213775,-0.1630108207,-0.2594662607,-0.0582631007,0.240571022,0.0249611661,0.3125114143,0.2753748,-0.347641468,-0.2138452083,-0.0816120505,-0.1637906134,0.1681686491,-0.0679172575,0.0338150002,-0.026006056,0.1627786756,0.0351047181,-0.2060480565,-0.108718209,-0.0047195763,-0.3626045883,-0.0274240635,-0.0853212774,0.2391911298,0.0578095429,0.1277653277,-0.1629247069,-0.1107515469,-0.214534089,-0.3260585368,0.5787270069,0.0464808196,-0.0090240892,-0.1731228232,0.5798003078,0.0645233244,-0.3917691708,0.3677161038,0.2882386148,0.493608743,-0.1893949509,0.1349033415,-0.305721581,0.0603804514,0.3331647813,0.6284927726,0.9139981866,0.1212794781,0.2586690485,-0.1242341846,0.1203274727,-0.0284565277,0.2634209991,0.117924206,-0.1473696828,0.1469343305,0.1326860487,-0.0013373005,0.0035384309,-0.1091660485,0.4651365876,-0.2163637578,0.0897275582,0.4169329703,0.3571682572,0.1019718423,-0.2288594842,0.0479043797,0.6168879867,0.1762443781,0.3369440436,0.4259212911,-0.1810424328,0.0998253375,-0.0017925523,0.2226395011,0.0534642749,0.4730301201,-0.3988971114,0.3629610538,0.2224749178,0.1349291652,0.3365784585,-0.3683717847,-0.2057874352,-0.0170305502,0.2325688303,0.0959953591,-0.0759056881,0.3478367031,-0.1701821238,-0.0474816747,-0.4018290937,0.0004564622,0.0625689253,0.1317359507,0.0139117725,-0.2197066545,-0.283659488,0.2622634768,-0.0878894851,-0.3314149678,-0.1918561608,0.1685844362,0.2644779086,0.2347223163,0.3726985753,-0.0108067077,0.0364033729,-0.0733317584,0.0056839227,-0.0789170489,0.1261888593,0.473839879,0.06578058,-0.095952183,0.1417224854,0.0203568954,0.2079632878,0.1000199467,-0.0975631773,0.1016673967,-0.0791088715,-0.1497850567,0.3356243968,0.0141824037,0.0848566294,-0.0308180954,-0.0254233349,-0.169125855,0.0687802136,-0.3932406306,0.3734216392,0.1371650398,-0.0253048688,-0.0013593927,-0.0318555422,-0.1749774069,-0.281285435,0.4291774035,-0.0715984255,0.1156730726,0.1271270365,-0.0008259757,-0.0232295673,0.4580632448,0.1040098593,-0.0665539429,0.0290403794,-0.1584466994,-0.3442613482,-0.0480582491,0.1553260088,-0.0620339513,0.0165732633,-0.0875465721,0.1618356258,0.2478548735,0.0179807097,-0.2271553874,0.2294033021,0.0271107052,-0.0681097284,-0.04274543,0.1231129542,0.1112904549,-0.1210737824,-0.277012229,0.0364801064,0.1902371943,-0.2698022127,0.2569993436,0.2185710669,0.2078666389,-0.4765966535,0.2123611867,-0.2095575035,0.2198643684,-0.0315437168,0.1896931082,-0.2996183038,-0.0461363271,-0.3355704546,-0.0755938366,-0.1748864204,0.2164681852,-0.2423885465,-0.013369035,-0.3065297604,-0.0302063487,-0.0357387178,0.1394870132,-0.2000430077,-0.0881002918,-0.3969622254,0.1654981971,0.2673887014,0.1068121791,-0.1092738509,-0.1776873171,-0.0837421045,0.0489660054,0.4212643206,-0.2280207574,-0.2450335175,0.0018892679,-0.0527397022,0.1859927773,-0.1028055623,-0.1450473517,0.112739794,0.0636365488,-0.1726148725,0.1622059196,0.2499663383,0.1666971445,-0.1502140313,-0.2082291096,-0.2506308854,0.3272816837,-0.1672694385,-0.1096107364,-0.4348281026]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2624","title":"can't set verbosity for `metric.py`","comments":"Thanks @thomas-happify for reporting and thanks @mariosasko for the fix.","body":"## Describe the bug\r\n```\r\n[2021-07-10 20:13:11,528][datasets.utils.filelock][INFO] - Lock 139705371374976 acquired on \/root\/.cache\/huggingface\/metrics\/seqeval\/default\/default_experiment-1-0.arrow.lock\r\n[2021-07-10 20:13:11,529][datasets.arrow_writer][INFO] - Done writing 32 examples in 6100 bytes \/root\/.cache\/huggingface\/metrics\/seqeval\/default\/default_experiment-1-0.arrow.\r\n[2021-07-10 20:13:11,531][datasets.arrow_dataset][INFO] - Set __getitem__(key) output type to python objects for no columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n[2021-07-10 20:13:11,543][\/conda\/envs\/myenv\/lib\/python3.8\/site-packages\/datasets\/metric.py][INFO] - Removing \/root\/.cache\/huggingface\/metrics\/seqeval\/default\/default_experiment-1-0.arrow\r\n```\r\nAs you can see, `datasets` logging come from different places. \r\n`filelock`, `arrow_writer` & `arrow_dataset` comes from `datasets.*` which are expected \r\nHowever, `metric.py` logging comes from `\/conda\/envs\/myenv\/lib\/python3.8\/site-packages\/datasets\/`\r\n\r\nSo when setting `datasets.utils.logging.set_verbosity_error()`, it still logs the last message which is annoying during evaluation. \r\n\r\nI had to do \r\n```\r\nlogging.getLogger(\"\/conda\/envs\/myenv\/lib\/python3.8\/site-packages\/datasets\/metric\").setLevel(logging.ERROR)\r\n``` \r\nto fully mute these messages\r\n\r\n## Expected results\r\nit shouldn't log these messages when setting `datasets.utils.logging.set_verbosity_error()`\r\n\r\n## Environment info\r\n\r\n- `datasets` version: tried both 1.8.0 & 1.9.0\r\n- Platform: Ubuntu 18.04.5 LTS \r\n- Python version: 3.8.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":10,"text":"can't set verbosity for `metric.py`\n## Describe the bug\r\n```\r\n[2021-07-10 20:13:11,528][datasets.utils.filelock][INFO] - Lock 139705371374976 acquired on \/root\/.cache\/huggingface\/metrics\/seqeval\/default\/default_experiment-1-0.arrow.lock\r\n[2021-07-10 20:13:11,529][datasets.arrow_writer][INFO] - Done writing 32 examples in 6100 bytes \/root\/.cache\/huggingface\/metrics\/seqeval\/default\/default_experiment-1-0.arrow.\r\n[2021-07-10 20:13:11,531][datasets.arrow_dataset][INFO] - Set __getitem__(key) output type to python objects for no columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n[2021-07-10 20:13:11,543][\/conda\/envs\/myenv\/lib\/python3.8\/site-packages\/datasets\/metric.py][INFO] - Removing \/root\/.cache\/huggingface\/metrics\/seqeval\/default\/default_experiment-1-0.arrow\r\n```\r\nAs you can see, `datasets` logging come from different places. \r\n`filelock`, `arrow_writer` & `arrow_dataset` comes from `datasets.*` which are expected \r\nHowever, `metric.py` logging comes from `\/conda\/envs\/myenv\/lib\/python3.8\/site-packages\/datasets\/`\r\n\r\nSo when setting `datasets.utils.logging.set_verbosity_error()`, it still logs the last message which is annoying during evaluation. \r\n\r\nI had to do \r\n```\r\nlogging.getLogger(\"\/conda\/envs\/myenv\/lib\/python3.8\/site-packages\/datasets\/metric\").setLevel(logging.ERROR)\r\n``` \r\nto fully mute these messages\r\n\r\n## Expected results\r\nit shouldn't log these messages when setting `datasets.utils.logging.set_verbosity_error()`\r\n\r\n## Environment info\r\n\r\n- `datasets` version: tried both 1.8.0 & 1.9.0\r\n- Platform: Ubuntu 18.04.5 LTS \r\n- Python version: 3.8.10\r\n- PyArrow version: 3.0.0\r\n\nThanks @thomas-happify for reporting and thanks @mariosasko for the fix.","embeddings":[-0.0164450556,-0.1983051002,0.0624923632,0.3260705769,0.3604136407,0.0033402808,0.2497837096,0.2126341164,0.1022727489,0.0168889575,-0.107101582,0.2478218377,-0.4029469788,-0.2203720957,-0.084005557,0.0563643686,-0.096119076,-0.2050801963,-0.2870321572,-0.0052471557,-0.1695397496,0.018277077,-0.2321506739,-0.1455849558,-0.2151203454,-0.3049913645,0.0790183395,0.0643912032,-0.2318696529,-0.6620803475,0.0565500893,0.2524934411,-0.0679500625,0.3298941851,-0.0001217199,-0.1692154557,0.2803701162,0.061549224,-0.2758444846,-0.0645168498,-0.2787055373,-0.3280662298,0.3504136801,-0.1066366136,0.0699354038,-0.3792483807,0.0067071002,-0.8163947463,0.1250315309,0.556129992,0.0772203878,0.482878536,-0.0428112708,-0.1152455583,-0.1727674603,0.029450167,-0.1470813304,0.3210988045,0.0574021526,0.3898583353,-0.1668436676,0.2930928171,-0.0125737488,0.0350346863,0.4341198206,0.0100479536,0.3920857012,-0.046760425,0.2035232037,0.0716290995,0.459723264,-0.308673799,-0.275817275,-0.1650470495,-0.0142459674,-0.6232368946,0.3095535338,0.0530422889,-0.0149922874,0.1309096515,-0.2309750021,-0.1991046071,-0.3325772285,-0.129840076,0.2360067368,-0.1349157989,-0.2652155757,0.034873791,-0.0028513425,-0.0733632296,-0.0032915939,-0.0925337598,-0.2141291797,0.1441205144,-0.4087710083,0.0915540457,0.1826591343,0.1220228672,0.1890175641,0.212641567,-0.1053712219,-0.0897499695,0.2264687419,0.0374114178,0.2668993175,0.5491036177,0.6421540976,0.5446236134,0.1643903702,0.0981422514,0.1792181283,0.0147715556,0.2008109987,-0.6014507413,0.6218041182,0.0650980249,0.2479801327,-0.1337943673,-0.1556895077,0.0989957601,-0.2872766256,-0.0103048868,0.1356708854,0.2623936832,0.2576891482,0.0596928298,0.0203463603,0.6495419145,-0.0984395519,0.148003161,-0.1879936755,0.0255448166,-0.0077158981,0.2401415557,0.1507176161,-0.0763579756,0.3559601307,0.1839520186,-0.0639201105,-0.2492405623,0.1221696883,-0.0861924887,-0.0288186781,0.1430593431,-0.0778643116,0.4392383397,0.1295867562,-0.2383223921,-0.2636347711,-0.049704805,0.0569836684,-0.2842857242,-0.0074634901,0.040415369,-0.4934317768,0.0442594476,0.0842013732,0.2576876879,0.1914454997,0.1904374808,0.2325722873,0.0565516986,-0.0435526483,-0.2317049354,0.1007082984,0.4656771421,-0.3440419436,-0.1920721382,-0.2648237646,0.0152872046,0.253970474,0.215752691,0.0927517638,0.1676219851,-0.156467557,0.1167004853,0.3492186964,-0.7669632435,-0.1473072916,0.1591783911,-0.2447284758,0.17201294,0.6470825672,-0.1939228028,0.157317102,0.0116600711,0.3219506741,-0.060640689,0.3228513896,-0.3583301306,-0.2799353898,-0.0809380934,-0.0578205474,-0.2309534252,-0.029324498,-0.0553553291,0.0651343912,-0.0240133405,0.2654452622,-0.1016368642,0.0123178149,0.3998618424,0.3589005172,-0.3530058861,0.0436522998,-0.2529400885,-0.0618834086,0.0189472735,-0.1010708511,0.0278828926,-0.3602592051,-0.1150591895,-0.346229583,-0.0885238498,-0.2014961541,-0.5932021141,0.0058553587,0.0172886271,-0.0488378629,0.0196845476,-0.2839436531,0.179469943,0.0425922982,0.2003278136,-0.1184783354,-0.3247057199,0.0648921281,-0.2354654074,0.1548301876,0.1198306829,0.1747334599,0.1077245027,-0.1513949633,0.4884937704,0.116710037,0.4465206265,0.1400009394,0.2830547392,0.1160362214,0.0318511724,0.1670981199,-0.0961703286,-0.211820662,0.1421679705,-0.1261021346,0.3970070481,0.1163775399,0.122106567,-0.1616731435,0.1263966262,0.1336526871,0.1298662871,-0.2831715643,-0.6291419268,0.6243052483,0.1267271936,0.2755241394,0.0409941748,0.011163054,-0.0724290311,0.38078776,-0.011126806,0.3513855636,-0.128338173,-0.226719752,-0.1150107309,0.0205685254,-0.1554236859,0.4445778131,0.0867657289,0.1374484003,0.0298621003,0.0261621159,-0.0853510574,0.3984567225,-0.1252631098,0.069249548,-0.0132521568,0.0828045681,-0.2102893293,-0.1034414992,0.2290515155,-0.118354328,0.0429563262,-0.5191911459,-0.0565301403,-0.3778877556,0.2134447545,-0.2643279731,0.0933636129,-0.1556318104,-0.2250060439,0.0252066068,-0.2027360946,0.0254288036,0.3733584881,0.168591857,0.3254945874,0.3283054233,0.1588714719,0.0401892401,-0.1040837541,0.0071758009,-0.0895527899,-0.1507396549,-0.1209394038,0.0437253118,-0.0361163467,-0.0091241198,-0.1600427926,-0.247460112,0.192608878,0.1127900481,0.6947782636,0.2625145614,0.0454062186,0.3110182285,0.0924242884,0.0273315795,0.1597931385,0.0091093723,0.2158938795,-0.221982792,0.294264853,-0.1231253967,-0.1892974675,-0.0673691705,-0.3932366967,0.3603760004,-0.132033065,0.0609931201,0.3443830013,0.0774757266,0.2090590298,-0.2359493226,0.5093499422,0.0043735499,-0.1389270574,0.3678599,-0.2394568175,-0.2037159652,-0.3087920845,0.1644431204,0.3166615665,-0.0275044963,-0.2505586445,-0.1664769053,-0.340723604,0.1677137315,0.0434190631,0.0853528753,0.1394096166,0.1067832708,0.0044752853,-0.2147698849,-0.2878776789,0.2752936184,-0.086995393,0.1809836328,-0.0594176762,0.3070940375,0.1807605028,0.2505643666,0.5291647315,0.2686712146,0.0501484945,0.1129132956,0.5512906313,0.0138282087,-0.3059231341,0.2039941847,0.190837279,-0.1642496884,0.0083469069,0.0239218194,0.1569709778,-0.0434987508,-0.2635162771,-0.0525062643,-0.2673736811,-0.0766462013,-0.3548699319,0.1559651792,-0.1433241516,0.423941046,-0.3618697822,0.0603283867,0.0353110284,0.1243573949,0.1034142599,-0.0129614975,-0.2013385892,0.1229084805,-0.375857234,0.5071482658,-0.2869330049,0.3585317731,-0.1421070695,-0.0593562052,0.0821808428,0.2602353096,0.4357093871,-0.0935138389,0.0460582562,-0.1688995361,0.1624404192,-0.8294293284,0.1357749254,-0.1944236308,0.3039085567,0.2164303362,0.4919110537,-0.4865396321,-0.1049480215,-0.1431339979,0.1641668677,-0.0596312582,-0.2888057232,-0.147220403,0.1133880615,-0.1393871158,-0.1559313387,-0.0204090886,0.3750354052,0.0780899003,0.0639546141,0.1456229538,-0.328140974,0.1105480045,-0.2014809251,0.3712606132,-0.3418799937,0.0552370884,0.3000589311,-0.051558122,0.0202817693,0.2457080036,0.0307868067,-0.4742592871,0.1085763276,-0.0963651389,0.2828398049,0.4608395398,-0.3322256505,0.3634103835,0.0905560777,0.1703660041,-0.1654758453,-0.0881478861,0.6081295609,0.2906722724,-0.0829603598,-0.0700436085,0.2537160516,0.5579833388,0.0253055673,0.5752627254,0.1135116294,-0.4807421863,0.1757535487,0.1608614326,0.8202492595,0.1741283983,-0.0458666347,0.0105153136,-0.5095279813,0.5416184068,0.0345193855,0.1839217693,-0.044546023,-0.0187242217,0.1963469088,-0.2686910331,0.3337648511,-0.2523767352,-0.1560450196,0.1603685617,-0.6277882457,-0.0868003815,-0.2025874257,0.2064725161,-0.6442013383,-0.1125139669,-0.3217775226,-0.0912993997,-0.174662143,0.0087322835,0.0400173888,-0.1610732675,0.0290423594,-0.2320365906,-0.4034992158,-0.2457226068,-0.1206716076,-0.0131660039,0.005651928,-0.361759603,0.0189213175,0.249561131,0.3716038764,0.2450838238,-0.1733157933,0.0846352503,0.0747837648,0.1610556841,0.0978616998,0.0574021973,0.3293653429,-0.0686309934,-0.0357178524,0.0930647776,-0.1056243181,-0.1139383689,0.0893620998,-0.3299735188,-0.237920925,-0.0459102839,0.086157307,-0.0119245257,0.0705703795,-0.1414358616,0.0438312516,0.2544493973,-0.1361564398,0.153183639,-0.252274543,-0.3171855807,-0.1907018423,0.4217742383,0.0212673917,0.0480265878,0.4595507681,0.1551587135,-0.285503149,-0.1368323267,0.0971589983,0.4120389521,-0.3883369863,0.374007076,0.0595178232,-0.0645230636,-0.1687813699,0.0701079667,0.1518827826,0.1414371282,-0.2698029578,-0.3390607834,-0.3413351476,0.2970564067,-0.3959448934,-0.0175308883,-0.3414161503,0.1011019722,-0.216849044,0.2438206673,-0.214143604,-0.3096663058,-0.0889823139,-0.1279851794,0.2567438185,0.002567281,0.3332538307,0.0313410535,0.0169229936,0.0322058275,-0.2285452783,-0.0160173494,-0.2416029721,0.1547051817,-0.2681459486,-0.0506504253,0.1964904666,-0.3203561902,-0.001411253,-0.2386238575,0.1126094311,0.3639178872,0.0614244901,0.1267693341,-0.3446662724,0.2509265244,0.0568228364,-0.2720153928,-0.0771230459,-0.1439490914,-0.0541860946,0.0916921273,-0.1823784262,-0.0074243923,-0.2403686196,-0.0453206748,-0.1089047864,0.1985713094,-0.1327511668,-0.0209933445,-0.2069012225,0.1218763068,0.4348282814,0.2941372693,-0.1354902089,-0.1288892925,-0.2663221061,0.0251052901,0.0696430579,0.0146890339,-0.5752022266,0.033746738,-0.0395466462,0.2828603387,-0.1070483848,-0.0344460793,0.303244859,0.1492841542,0.0232725497,-0.1760827303,0.2588167191,0.0607743561,-0.1147778481,-0.1143609732,0.6400613785,0.2229339629,0.3544741273,0.1318780929,-0.2174153775,0.2804185748,0.1710978001,-0.0069963695,0.2477241009,0.1154584065,-0.0184469279,0.0017336718,-0.2039725184,-0.1320480704,-0.1374344975,0.4389801025,-0.3817619681,0.0001754238,-0.0529538244,0.0204616711,-0.0898575485,0.0627794117,-0.0890953541,-0.2729817629,-0.0195932332,-0.2200428396,-0.0642063916,-0.0565999039,-0.0395338088,0.0206888095,-0.0932507068,-0.5210008025,0.2200365216,-0.0288567394,0.385497123,-0.1773010194,0.1805490106,0.1089988351,0.0472670719,0.231135577,0.5787084699,0.3976887167,0.4378178716,0.0614022166,-0.0246233828,0.127121225,-0.1487811655,0.1082215682,0.3670316637,0.2607315481,-0.2848185301,0.1577142477,-0.0111270258,-0.1224841028,0.3849051595,0.1343702823,0.2021715641,-0.294616729,0.3844841123,-0.1523032933,-0.0137252184,0.0207273364,-0.1844816655,-0.3430219889,-0.0715130344,0.3318836987,0.119039163,-0.0909185931,-0.4288811386,0.0544002317,-0.0594451912,0.1131676063,0.4539165497,-0.108069934,-0.2731710076,0.2395078987,-0.3752695322,0.2776774168,0.2422740459,-0.0534590371,-0.0728696063,0.1608915925,0.0549881049,0.1378500164,0.3992603421,0.0837507397,-0.052310925,-0.1917655319,-0.379558444,0.0569669865,0.1458509862,-0.0470301434,-0.0259787105,-0.0895335898,0.2408657372,-0.4458166063,0.0152827175,-0.064199917,-0.0547806583,0.2946983576,0.646540463,0.2526344657,0.2410595119,0.0822716206,0.1001594365,-0.2657421827,-0.501617074,-0.1268895417,-0.24351044,-0.0370829739,-0.043037463,0.2053964734,-0.3191708624,-0.4217471182,-0.3497889638,0.4676087201,-0.2629002631,-0.4972365499,0.2667987645,0.1246410608,0.1365693212,-0.2957902253,0.2285525352,0.1610539854,-0.0034055787,0.3188881576,-0.3713078499,-0.423151195,0.4177734852,-0.4886701703,-0.1957473159,0.332755059,0.5447626114,0.4989789724,-0.4108757079,-0.4951102138,-0.0041189152,0.3204493225,-0.0719457343,-0.1525457948,0.1386837959,-0.20588094,0.128389582,0.12037763,0.2910915017,0.2711361349,-0.0829176381,-0.1535827816,-0.1976826936]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2622","title":"Integration with AugLy","comments":"Hi,\r\n\r\nyou can define your own custom formatting with `Dataset.set_transform()` and then run the tokenizer with the batches of augmented data as follows:\r\n```python\r\ndset = load_dataset(\"imdb\", split=\"train\") # Let's say we are working with the IMDB dataset\r\ndset.set_transform(lambda ex: {\"text\": augly_text_augmentation(ex[\"text\"])}, columns=\"text\", output_all_columns=True)\r\ndataloader = torch.utils.data.DataLoader(dset, batch_size=32)\r\nfor epoch in range(5):\r\n for batch in dataloader:\r\n tokenizer_output = tokenizer(batch.pop(\"text\"), padding=True, truncation=True, return_tensors=\"pt\")\r\n batch.update(tokenizer_output)\r\n output = model(**batch)\r\n ...\r\n```","body":"**Is your feature request related to a problem? Please describe.**\r\nFacebook recently launched a library, [AugLy](https:\/\/github.com\/facebookresearch\/AugLy) , that has a unified API for augmentations for image, video and text.\r\n\r\nIt would be pretty exciting to have it hooked up to HF libraries so that we can make NLP models robust to misspellings or to punctuation, or emojis etc. Plus, with Transformers supporting more CV use cases, having augmentations support becomes crucial.\r\n\r\n**Describe the solution you'd like**\r\nThe biggest difference between augmentations and preprocessing is that preprocessing happens only once, but you are running augmentations once per epoch. AugLy operates on text directly, so this breaks the typical workflow where we would run the tokenizer once, set format to pt tensors and be ready for the Dataloader.\r\n\r\n**Describe alternatives you've considered**\r\n\r\nOne possible way of implementing these is to make a custom Dataset class where getitem(i) runs the augmentation and the tokenizer every time, though this would slow training down considerably given we wouldn't even run the tokenizer in batches.\r\n","comment_length":68,"text":"Integration with AugLy\n**Is your feature request related to a problem? Please describe.**\r\nFacebook recently launched a library, [AugLy](https:\/\/github.com\/facebookresearch\/AugLy) , that has a unified API for augmentations for image, video and text.\r\n\r\nIt would be pretty exciting to have it hooked up to HF libraries so that we can make NLP models robust to misspellings or to punctuation, or emojis etc. Plus, with Transformers supporting more CV use cases, having augmentations support becomes crucial.\r\n\r\n**Describe the solution you'd like**\r\nThe biggest difference between augmentations and preprocessing is that preprocessing happens only once, but you are running augmentations once per epoch. AugLy operates on text directly, so this breaks the typical workflow where we would run the tokenizer once, set format to pt tensors and be ready for the Dataloader.\r\n\r\n**Describe alternatives you've considered**\r\n\r\nOne possible way of implementing these is to make a custom Dataset class where getitem(i) runs the augmentation and the tokenizer every time, though this would slow training down considerably given we wouldn't even run the tokenizer in batches.\r\n\nHi,\r\n\r\nyou can define your own custom formatting with `Dataset.set_transform()` and then run the tokenizer with the batches of augmented data as follows:\r\n```python\r\ndset = load_dataset(\"imdb\", split=\"train\") # Let's say we are working with the IMDB dataset\r\ndset.set_transform(lambda ex: {\"text\": augly_text_augmentation(ex[\"text\"])}, columns=\"text\", output_all_columns=True)\r\ndataloader = torch.utils.data.DataLoader(dset, batch_size=32)\r\nfor epoch in range(5):\r\n for batch in dataloader:\r\n tokenizer_output = tokenizer(batch.pop(\"text\"), padding=True, truncation=True, return_tensors=\"pt\")\r\n batch.update(tokenizer_output)\r\n output = model(**batch)\r\n ...\r\n```","embeddings":[-0.1749232709,-0.1920810193,-0.1434521526,-0.239678964,0.1912634373,0.0263557434,0.167391479,0.3273121417,-0.3234899342,0.0160070173,0.0447254553,-0.0170348417,-0.2608955204,0.1105004624,-0.0723288879,-0.1290824711,0.1623859406,0.1287831068,0.0835960805,-0.0614518151,0.2458241582,-0.0801938996,-0.1216678768,0.0804531202,-0.4331211746,-0.2005291581,-0.0184633899,-0.1191418543,0.0323870108,-0.0549192354,-0.2768345773,0.2928821146,0.139453426,0.2861319482,-0.0001094861,-0.2278579026,-0.0457839854,-0.1373027265,-0.3240187466,-0.0370470993,0.8906273842,-0.4260549843,0.018437136,-0.1956669092,-0.2086928785,-0.2443619668,0.225955084,-0.1025847942,0.7065521479,-0.1400630176,0.1746794581,0.1642660201,-0.1695449203,0.1748896688,0.0812375322,0.6041795611,-0.0998975858,-0.1248127446,0.3372099102,-0.287638396,-0.1602464765,0.5248727798,0.0838095024,-0.0850363374,0.3582636416,0.0786501989,0.0502489135,-0.5240317583,-0.2224519402,0.1617945582,0.064426139,-0.1844712794,-0.5932481885,-0.3726418614,0.2279145569,-0.0164295323,-0.2347869426,0.1063614562,-0.1818405986,0.0403820574,-0.1904088408,-0.1769422144,-0.0590320975,0.2114746869,0.2574062645,0.2118017375,0.2327120304,-0.1030833274,-0.3152122796,-0.1901496947,-0.017550623,0.1092635542,0.216767475,0.3279650807,-0.0164926499,-0.1686119884,-0.2443469465,-0.3988581598,-0.0008842092,0.1280761212,0.0307479911,0.420787394,-0.4762347043,0.0391779132,0.0057944143,0.0282802861,0.0076813176,0.1638359725,0.3196819425,0.054072842,0.0039827083,-0.0232678968,0.0108465804,0.1630419195,0.0301643293,-0.1091242656,-0.3127089739,0.2283012867,-0.2628072798,0.0098471437,-0.1675759256,0.1402470917,0.0705191717,0.2345561385,-0.2274330705,0.2814390957,0.2564775944,0.1967922896,0.0014192906,0.035521768,-0.0433994904,0.1743155271,-0.3539296091,0.1625114679,0.280423969,0.5648206472,0.1444967091,0.0708900243,0.3205162585,0.0095686428,0.2328449339,-0.1425879747,0.4217463732,-0.0478327163,-0.1146397516,0.2201912552,-0.0657536536,0.0721538365,-0.119618915,-0.1751389802,-0.1528177708,0.1419904232,-0.2186926752,0.2464309633,0.3551001549,-0.0151164867,-0.2094570398,0.3241281509,0.0210421551,-0.3204990327,0.1727177501,-0.1375955194,0.0115348604,-0.0599739328,0.2275060564,0.0374388285,-0.0412992723,-0.2540357709,0.4374252558,0.2148566544,0.1135617793,0.3173480928,-0.2680597603,0.0712589696,-0.0477385707,0.0781117305,0.5492462516,-0.4070941806,0.051066298,0.2031701058,0.0156891905,0.063910991,-0.1526074409,0.3913735449,0.2438090593,0.0073967106,0.1471612751,0.5788085461,-0.0462913401,0.072911121,-0.0735741481,-0.3104627728,0.2083489448,0.0925204605,0.0948226675,-0.3573521674,-0.2048754394,0.0999733284,0.3083713055,-0.2700299323,0.3583818972,-0.0095584178,-0.2125834227,0.3153033555,-0.0342440456,-0.1558937132,0.0081812898,0.0151067451,-0.0870277882,0.3045280874,-0.1963891983,-0.3819638193,0.229483664,-0.0637034997,0.0682513863,-0.3417001963,0.1884030551,0.0611839779,-0.0851469785,-0.0551100858,-0.3363082409,0.1103338376,-0.1135299876,0.1098784581,0.0517388619,0.040688049,-0.03455396,-0.1430622339,-0.3016890287,0.3711787164,-0.0835312009,-0.0199187733,-0.0105550857,0.2912918925,-0.0540971868,0.1201765835,-0.0536797419,0.2401742935,0.3565207422,-0.0962436572,0.1682845652,0.1372682005,-0.0107516479,-0.289251864,-0.0039154035,0.5061603785,0.2224728614,0.2047206461,0.3028242588,-0.0844083056,0.0872405171,0.0437296331,-0.2721039951,-0.3616888821,-0.1609094441,-0.0939855874,0.1342760772,-0.3366361856,-0.5899156928,-0.0114842718,0.2806825638,0.2557839751,-0.0145958262,0.1399237812,-0.183690384,-0.1397485733,0.0756567791,-0.5855828524,-0.0750117525,0.2024073005,-0.0168656539,0.1435841024,-0.0854747966,-0.0052283532,0.0352681763,0.2353792638,-0.0255144238,0.0951068327,0.4083151817,0.3604509532,-0.1396538168,-0.3078117371,-0.1590868384,-0.4041635394,-0.0580309778,0.2476757914,-0.1718835831,-0.4956324697,-0.1368497014,-0.4858056307,-0.2315850407,0.0946224555,0.3938374221,-0.0339780822,-0.1213260442,0.3089965284,-0.0552519336,0.3120442033,-0.0148540866,-0.1032003462,-0.029511502,0.195880264,-0.2042996138,0.0272398032,0.003990253,0.047898218,0.4143897593,0.2367702872,-0.0613217652,-0.368853271,-0.5738261342,0.1989150047,-0.0435740761,-0.0754411444,0.1535942703,-0.2465477586,0.0378198773,-0.5249401331,0.0829260051,-0.2365261763,-0.0877230018,-0.3229259849,0.0898313299,-0.1227382049,-0.2532222569,0.0823828354,-0.3209186792,-0.350895375,-0.0673328415,0.213936165,0.0358337909,-0.1540177017,-0.260230571,-0.1037170589,0.1582040936,0.2489973009,0.0276557952,0.0360089839,0.2866538167,-0.3043049574,-0.1436741948,-0.1987398416,-0.3153315187,0.1066342443,0.2237939388,-0.1450603306,-0.2972123623,0.0744433552,0.0842172951,-0.248432532,-0.1239834875,0.0842127874,-0.0569638982,-0.1526043266,0.0490060262,-0.0183922965,0.1789525002,-0.1769568324,0.1623331755,-0.3505836427,0.2998358309,0.2065783292,0.4156169891,-0.1009587571,-0.1860577464,0.4628739059,-0.1015675366,-0.0547133237,-0.0337048657,-0.069356434,-0.0145598808,-0.2122713029,-0.0271373782,0.2304039747,-0.0089175217,-0.2364802808,-0.1578734964,0.2057914436,-0.3513627648,-0.1594031155,0.3336032927,-0.5718927979,0.2396318167,-0.213025108,0.1591394693,-0.1597550958,-0.2681823969,0.321059674,-0.1664436907,0.267380625,-0.0826042816,-0.0513877571,0.0470828228,-0.1404735148,0.2797584832,0.2064885646,0.1528767794,0.1210501865,-0.17610313,0.1129442081,-0.1383547634,0.4136569798,0.0103406664,-0.3415987492,-0.0356955975,-0.1137947068,-0.1227869838,-0.1780914515,-0.1877986044,-0.0266245157,0.5311674476,0.5997403264,-0.2997632325,-0.2721070051,0.3077866137,-0.1064927801,-0.1589561701,0.1346833259,0.1879262775,0.2151338607,-0.4126152098,0.1211763918,0.0550892204,0.0057670176,0.1323633194,-0.1052786633,-0.0704211071,0.1855515689,0.267701745,0.1676829606,-0.039902959,-0.0350229032,-0.0575038269,-0.0223080646,0.0960756019,0.3651774824,-0.0264440011,0.0704547614,-0.2445635796,-0.0447181351,0.1562063545,0.4355128706,0.4480147362,0.2516478002,0.3036258221,0.0256296732,0.1053525582,-0.3496856093,0.3098253906,0.4465028048,-0.0153545085,-0.497428894,-0.4103819132,0.1762539297,-0.0963573232,-0.0155202346,-0.2599515319,0.0811266229,-0.3926424384,0.6640623808,0.4433370233,1.1297671795,0.3573449254,0.4586040676,-0.042353455,0.1152503565,0.3830606341,-0.673458755,0.0660290271,-0.3117773235,0.0425253846,-0.0314418972,0.0484329611,0.0105464216,0.2368612587,-0.0657507405,0.2240267694,0.0588077642,0.0110488478,-0.108508639,0.4447461963,0.005106749,-0.6270374656,-0.1310868859,0.1267711818,0.0423105732,-0.0223922729,-0.088453114,-0.1645103842,0.1444617063,0.0065420261,-0.135134235,0.0482105166,-0.2696858048,0.1632843465,0.0733539984,-0.0779519007,0.2392180264,-0.2708636224,0.3563738763,0.1940164268,0.1004142389,0.1763637066,0.1171764955,0.2489150316,-0.0542249531,-0.316765368,0.8901433945,-0.0371697061,0.2000305057,0.0912044793,0.2408939451,-0.1644595414,-0.0225650202,-0.1324387491,0.3902791142,-0.1736596972,-0.0781239271,0.1845993549,-0.2183440328,-0.1247097626,0.1335342973,-0.0311878733,-0.1671603471,0.2195684612,0.3298569024,-0.0261120386,-0.1715523601,0.1039448529,-0.2159543335,-0.4586632252,0.560089469,0.1068932042,-0.222066462,-0.1168676466,-0.2498060614,0.2313771546,-0.0732233375,-0.0247418173,0.1908100843,-0.1832662523,0.0654036626,0.2495058626,0.132174179,0.0188958049,-0.3815691471,-0.0789289474,-0.2347981781,0.2478352934,0.200075835,0.2317266464,0.0152797624,0.3859042525,-0.0833211541,0.4019308686,-0.3887147605,-0.0736246258,0.0603842549,0.3245539963,-0.4759229422,-0.1236636117,-0.063530542,0.2932154536,0.095219478,0.2746931612,-0.0704276487,-0.1763802916,-0.3155696988,0.1093037277,0.0088389916,-0.030079402,-0.0055691213,0.0915765464,-0.1552722305,-0.3959919512,0.046936743,-0.057598684,-0.0003967332,0.049351383,-0.0077325245,0.0337227993,0.0584315285,-0.1813467145,0.1635546982,0.2311700881,0.2100100219,0.1916544288,-0.0230313372,-0.1339606345,-0.1600054801,0.2594057322,0.3972215354,-0.1446175724,0.0327139087,0.064067401,-0.2973059416,0.0151833585,0.3159179986,0.1479309201,-0.0713802949,-0.2626295984,0.3240513802,0.2630267739,-0.2260718197,0.138336435,-0.0347756147,-0.5336185694,0.0946138799,0.2589986622,0.343636781,0.4469383955,-0.1533930898,-0.003376601,-0.0396505706,-0.4922482967,-0.0260316059,-0.1051439121,-0.1139655858,0.0825660601,0.2647008598,0.1236176714,0.2827006876,-0.0334637351,0.0462083109,0.2988395095,0.3937669098,0.0245549418,0.1559988707,-0.0895115137,-0.1781278104,-0.1676718742,0.3514990807,0.3922958672,-0.0154848332,0.01065455,0.1252191961,-0.3595495522,-0.0910577253,0.4749238193,-0.1580017358,-0.0369389318,0.0938891023,0.01664396,-0.532045722,-0.0606442355,-0.0281220321,-0.0847434178,-0.3039164245,0.2317366153,-0.0121706817,-0.0917977542,0.1023019478,-0.2134156823,0.5871031284,-0.2450865358,0.0072283843,0.1629156321,0.1580485702,-0.0144391414,0.1343713999,0.3595371246,0.3483406901,-0.2469238192,0.0921181142,-0.3106658459,-0.0230593532,-0.1457124352,0.2029488385,0.318666935,0.1778246909,0.0684296042,0.1633449495,-0.2882708311,-0.0467927568,-0.1142951921,0.3790620267,-0.2499668151,0.1722797602,0.2577950358,-0.1916575432,-0.4491395652,0.0122586591,-0.2413178682,-0.02361935,0.3017308116,-0.2390322983,-0.0330839418,-0.2029784769,0.1085819006,0.0240442604,0.460279882,0.5174255371,0.3198362589,-0.0012373828,-0.0460660011,-0.1639815569,-0.0320615545,0.306517601,-0.0632732362,-0.0743617043,0.3361044526,0.2154158205,0.0053997468,-0.0040307585,-0.237017259,-0.2534395158,0.2219527811,-0.1255366951,0.4652813971,-0.2609788179,0.2266547233,0.1869705021,-0.1629934311,-0.0399313942,-0.2518545985,0.0541983731,0.0949323997,-0.0672334582,0.1191674396,0.000279998,0.5229854584,0.0870423615,0.122985594,-0.3334083855,-0.1373925656,-0.1719631404,-0.1722701341,-0.55221349,0.0878214389,0.0875230506,0.6268986464,0.0971802399,-0.1202661693,-0.124128215,0.0881530792,0.0131391045,-0.5722593665,-0.1499316245,0.0011818724,-0.0878508091,-0.0149726849,-0.0661503524,0.472006917,0.0517066121,-0.0070290994,-0.3324206471,-0.4548497498,0.3370195925,-0.0535528734,-0.2709163725,-0.399135679,0.1516965628,0.3290503025,0.1189180985,-0.6594867706,-0.1559226513,0.2090781033,0.0794972926,-0.0144651067,0.3205159605,0.0903447866,-0.1893980354,-0.2628094554,0.0255032629,-0.198124215,0.0315613933,-0.317602098,-0.2944549024]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2618","title":"`filelock.py` Error","comments":"Hi @liyucheng09, thanks for reporting.\r\n\r\nApparently this issue has to do with your environment setup. One question: is your data in an NFS share? Some people have reported this error when using `fcntl` to write to an NFS share... If this is the case, then it might be that your NFS just may not be set up to provide file locks. You should ask your system administrator, or try these commands in the terminal:\r\n```shell\r\nsudo systemctl enable rpc-statd\r\nsudo systemctl start rpc-statd\r\n```","body":"## Describe the bug\r\n\r\nIt seems that the `filelock.py` went error. \r\n\r\n```\r\n>>> ds=load_dataset('xsum')\r\n\r\n^CTraceback (most recent call last):\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 402, in _acquire\r\n fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\nOSError: [Errno 37] No locks available\r\n```\r\n\r\nAccording to error log, it is OSError, but there is an `except` in the `_acquire` function.\r\n\r\n```\r\n def _acquire(self):\r\n open_mode = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC\r\n try:\r\n fd = os.open(self._lock_file, open_mode)\r\n except (IOError, OSError):\r\n pass\r\n else:\r\n self._lock_file_fd = fd\r\n return None\r\n```\r\n\r\nI don't know why it stucked rather than `pass` directly.\r\n\r\nI am not quite familiar with filelock operation, so any help is highly appriciated.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\nds = load_dataset('xsum')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\n```\r\n>>> ds=load_dataset('xsum')\r\n\r\n^CTraceback (most recent call last):\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 402, in _acquire\r\n fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\nOSError: [Errno 37] No locks available\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 818, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 470, in prepare_module\r\n with FileLock(lock_path):\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 402, in _acquire\r\n fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\nKeyboardInterrupt\r\n```\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.15.0-135-generic-x86_64-with-debian-buster-sid\r\n- Python version: 3.6.13\r\n- PyArrow version: 4.0.1\r\n","comment_length":84,"text":"`filelock.py` Error\n## Describe the bug\r\n\r\nIt seems that the `filelock.py` went error. \r\n\r\n```\r\n>>> ds=load_dataset('xsum')\r\n\r\n^CTraceback (most recent call last):\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 402, in _acquire\r\n fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\nOSError: [Errno 37] No locks available\r\n```\r\n\r\nAccording to error log, it is OSError, but there is an `except` in the `_acquire` function.\r\n\r\n```\r\n def _acquire(self):\r\n open_mode = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC\r\n try:\r\n fd = os.open(self._lock_file, open_mode)\r\n except (IOError, OSError):\r\n pass\r\n else:\r\n self._lock_file_fd = fd\r\n return None\r\n```\r\n\r\nI don't know why it stucked rather than `pass` directly.\r\n\r\nI am not quite familiar with filelock operation, so any help is highly appriciated.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\nds = load_dataset('xsum')\r\n```\r\n\r\n## Expected results\r\nA clear and concise description of the expected results.\r\n\r\n## Actual results\r\n```\r\n>>> ds=load_dataset('xsum')\r\n\r\n^CTraceback (most recent call last):\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 402, in _acquire\r\n fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\nOSError: [Errno 37] No locks available\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 818, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 470, in prepare_module\r\n with FileLock(lock_path):\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 323, in __enter__\r\n self.acquire()\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 272, in acquire\r\n self._acquire()\r\n File \"\/user\/HS502\/yl02706\/.conda\/envs\/lyc\/lib\/python3.6\/site-packages\/datasets\/utils\/filelock.py\", line 402, in _acquire\r\n fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\nKeyboardInterrupt\r\n```\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.0\r\n- Platform: Linux-4.15.0-135-generic-x86_64-with-debian-buster-sid\r\n- Python version: 3.6.13\r\n- PyArrow version: 4.0.1\r\n\nHi @liyucheng09, thanks for reporting.\r\n\r\nApparently this issue has to do with your environment setup. One question: is your data in an NFS share? Some people have reported this error when using `fcntl` to write to an NFS share... If this is the case, then it might be that your NFS just may not be set up to provide file locks. You should ask your system administrator, or try these commands in the terminal:\r\n```shell\r\nsudo systemctl enable rpc-statd\r\nsudo systemctl start rpc-statd\r\n```","embeddings":[0.1390438825,-0.3473012149,0.0151467938,0.0418614037,0.0302764755,0.0907045901,0.1623594612,0.1356668025,0.0592375062,0.0936220959,-0.1010852456,0.340319097,-0.0329612717,-0.4240655601,-0.459251672,0.1141825914,0.0069274814,-0.0698580444,-0.1149571389,-0.0138159692,-0.3595577478,0.3968342245,-0.4249511659,0.0698371232,-0.2793181241,-0.0480348431,-0.1366254836,0.3886517584,-0.1688994318,-0.1395633221,-0.0517245941,0.0867065936,0.0480411835,0.5414582491,-0.0001142323,-0.1104709134,0.0751497373,-0.0566150211,-0.3720932305,-0.0422811881,0.1407606304,-0.0910087451,-0.0822487995,-0.3874730766,0.2373984158,-0.3464207649,0.1434787065,-0.2871147692,0.4022390246,0.2639315426,0.1441263407,0.1482024789,0.4175293744,-0.178009063,0.1095130891,-0.4201950729,-0.0597184002,0.6798739433,0.7193904519,-0.1613604128,0.0171960033,0.0855058357,-0.148249507,-0.2049404085,0.2320636958,0.0193500929,0.3815196753,-0.4401879013,-0.2167775631,-0.0676430166,0.0476562157,-0.0331165791,-0.4759555459,-0.1618117094,0.4275055826,-0.07761994,0.0674172193,0.1474146098,-0.3159746826,0.3756872118,0.3556930125,-0.1153605878,-0.0145976534,0.0721763074,-0.0632616431,0.3086351454,0.0281341132,0.1370660216,0.262724787,-0.030876169,-0.0805696622,-0.0226215124,-0.0786239654,0.1791381687,-0.375010848,0.0532035343,0.1346542835,0.0576871149,0.0491891205,-0.2361772358,-0.3443281651,-0.246875599,-0.2653763294,0.2343066335,-0.1037879363,0.0391341224,0.0564300455,0.1757361293,0.3515813649,0.1015717313,-0.1955176741,0.013869076,-0.0682477951,-0.2627129257,0.1323340237,0.2916414738,0.407789737,-0.1522002071,-0.0398686491,0.233780697,-0.3236961961,0.018107092,0.1310848594,0.2916454077,0.1106775701,-0.0809489489,0.4470187724,0.1683355421,-0.1455666572,0.0815058947,-0.089699015,0.2096744478,-0.0342741646,-0.0005644005,-0.170863539,-0.3630220592,0.0755960718,0.3683515191,0.238421917,-0.3314391673,0.2022261024,-0.1128131896,-0.1492998302,0.1193967164,-0.1394654959,-0.0028866963,0.3368239105,-0.2456145138,0.128629908,0.171438545,-0.3326620162,-0.1702174097,-0.2444072664,0.144668743,-0.0669414401,0.0573142767,0.1982481927,-0.3010265827,0.0458981581,-0.315223217,0.1590230465,-0.3397700787,0.1084697098,-0.1530378461,0.0482599549,0.4075002074,-0.12052425,0.2917505801,0.1815027595,-0.1920057386,0.1625549197,0.0901744589,-0.02133926,0.1109134704,-0.4444103837,0.019757757,0.1067390218,-0.5011029243,-0.5333332419,0.0561696328,-0.4790895581,0.0537881702,0.1695948094,0.4211776853,-0.0173681006,-0.1641702652,0.2718978524,0.0371419527,-0.0135011952,-0.1450970024,-0.1528099775,0.0375398956,-0.0440814942,0.1406921446,-0.0029505112,0.1586375833,0.1399629563,0.0071283313,0.5296826959,-0.146213159,-0.2674141228,0.1628330052,0.4794795811,0.1399137527,0.0850243941,-0.1591180116,-0.189272061,0.2522282302,-0.1989827305,-0.0743363202,-0.0725066662,-0.0764314309,-0.0437692963,-0.0435342342,-0.0420152061,-0.0813650563,0.1175978482,-0.0148879644,-0.1799628735,0.0497146435,-0.1294880211,0.4790835083,-0.4433471859,0.2013792098,-0.1435981095,0.0725171342,-0.1064023823,-0.2321328968,-0.2575782537,-0.0071405102,0.2843230367,-0.1900358796,-0.2792275548,0.4126048088,0.2959444821,-0.1124275327,-0.2172019482,-0.2112116665,0.1744597554,0.3583770394,-0.2304764688,0.0148896417,0.0837068111,-0.0612622015,0.1202413812,0.1285025179,0.0746100321,-0.0397937819,-0.0832265541,0.2109800279,-0.0251819398,0.0652944073,0.0204290822,0.0194804948,0.4611050189,0.0080103828,0.0624385551,-0.1079219654,-0.0518783815,-0.0763699785,0.1907402128,0.0570197999,0.2793467343,0.1884344518,0.3128790557,0.3672698438,-0.193111077,0.5885143876,0.7208219767,0.101684466,-0.2875608802,0.0199206434,0.3027022481,-0.2104957104,-0.0469008498,-0.0330835022,-0.0418750085,0.2120727748,0.0586945824,0.2311490327,0.2032687366,-0.8740941286,-0.0125969276,0.2340481877,-0.004902795,-0.0489193536,-0.1694609076,0.1353217512,0.0601872131,0.3312356472,-0.0317488611,-0.3138143122,0.1344853938,0.4649529457,-0.074478209,0.0039711501,-0.3631457984,0.0643610656,0.1980468035,0.1316606551,0.1112059504,0.0367514491,-0.0958562642,-0.0664244443,0.1315078735,-0.0980742946,0.2080141157,-0.0003201759,-0.081226252,-0.4246044159,-0.1829681247,-0.0314671658,-0.1735013127,0.4597562253,0.056715481,0.3035485148,-0.0233455431,0.0249478146,0.1383039355,-0.0979297757,-0.0782772005,0.0798414126,0.0149578704,-0.0715152249,-0.3844693005,-0.034707766,-0.3898947835,-0.5175123811,0.1433671862,-0.1855088621,0.0664420128,0.3325220942,-0.1523124278,0.4005431831,0.1016704738,0.1809343398,0.1439602524,-0.292670399,0.2969040275,-0.0858423859,-0.4228855968,-0.3153539896,0.0855032727,-0.473782897,-0.1134320945,-0.2983644307,-0.5243878961,-0.3494296968,0.0939467922,-0.5529792309,0.0196183491,0.161808297,-0.2451497763,-0.0653789341,-0.2333385199,0.1794316769,0.1334120482,0.0864112228,0.1263547987,0.0655043945,0.0586766675,-0.0067032562,0.3639445603,0.108551845,0.2269399017,0.2979877591,-0.0920615122,0.268489629,0.0239143018,-0.2958976328,-0.203977406,0.2649236321,-0.2320870757,-0.1843943298,-0.0148587301,0.2902778387,-0.1478251815,-0.3533303142,-0.2893275321,-0.350485146,-0.0850317106,0.4227711856,-0.0989018828,-0.1246487796,0.120150581,0.1084767058,0.1586569399,0.0896879584,0.5906792283,0.412155956,-0.105821602,-0.0802925453,-0.196801424,0.0326542407,0.3254837692,-0.0353327133,0.2591308057,-0.0896734744,0.0270839781,0.10452234,-0.1417988241,0.4386395216,0.3848526776,0.1384619772,0.0464407578,0.1160206944,-0.0257852338,-0.1277820319,-0.2922452986,0.2406879365,0.4588389397,0.3851797283,-0.034746781,-0.0558451898,-0.2660072744,0.1702415347,-0.2192809582,-0.6137136221,-0.0278826393,-0.1019435525,-0.4393648207,0.1196118146,0.0954027474,0.1485133618,-0.017836567,-0.0758678317,0.2558481395,-0.2038924992,0.0230451655,-0.0956969112,0.0668246746,-0.1188011989,0.1291119605,-0.0056663421,-0.3534395993,0.3384052217,0.228788659,0.2727631927,0.0599868633,0.2580750287,-0.2394795865,-0.0094577195,0.2725385427,-0.2674665451,0.1491403431,-0.0385685377,0.2362756282,0.1778629571,-0.2500508726,0.2482469678,0.0410698019,0.038538143,0.0458142795,0.0593945831,0.1177939475,-0.0741813555,0.7165994048,0.4143915474,0.0822594389,0.3143230677,0.0100753345,0.7889759541,0.0115474584,-0.0575823821,0.22419402,-0.0027726183,0.3785954416,-0.1774417013,0.1131703481,-0.3271002173,-0.5434795618,0.1170889512,-0.2865743041,0.4315250814,-0.4660895765,-0.0726576373,-0.0296697058,-0.3799648881,-0.4223923087,0.0227125119,0.0103898197,-0.8001109958,-0.0394977219,-0.1688697189,0.1111011803,0.1646426618,0.3454254866,-0.108151257,-0.1112472415,-0.3464567661,-0.1693930179,-0.3626006246,0.1279752254,0.0296590533,0.2635756731,-0.089251399,0.0973387957,-0.0544373803,0.4668441415,-0.0660160705,0.2516547441,0.0980864763,-0.0213680305,0.032290604,0.0524955355,0.2574519217,-0.0581384152,0.5052600503,-0.0395311825,-0.1861050129,0.2242104858,-0.118209511,-0.2262900174,0.0197783336,0.1649948061,0.0535584986,0.0207640566,-0.3287976682,-0.3234477341,-0.2497397214,-0.1183941811,0.0701757967,0.2645713985,-0.2806156576,0.0412965044,-0.2572526634,-0.2472554594,0.0778117776,0.236946553,-0.1354971379,-0.1562503129,0.7352300286,0.3345285356,0.0765157938,-0.0932231396,0.168274045,-0.1857693046,-0.3107033074,0.121874541,-0.1181877479,0.2336872518,0.1547018737,0.135411486,0.1855074614,-0.1656293422,-0.0478424728,-0.8504985571,-0.3750932217,-0.1020794734,-0.2187472731,0.0418167375,-0.1938312501,-0.0538672507,-0.1249258742,-0.187947765,-0.2676637471,0.0953266546,-0.3377147913,-0.1140251234,0.278742671,0.1326110214,-0.0353812091,0.0669252947,0.1483341604,0.4477180243,0.0377090164,-0.1106635556,-0.238695994,0.1802296937,-0.0075018634,0.0598253459,0.1673072278,-0.4267778397,-0.0159850325,0.1985914558,-0.1804462522,0.169961974,0.0499133095,0.131634891,0.4090231955,0.3865374625,-0.031267006,0.2474783957,-0.1597062945,-0.0531041622,-0.0117535824,-0.0375697128,-0.325842768,0.0814677998,-0.6099715829,-0.2135707289,-0.035285946,0.2961655259,0.228955254,-0.0526553504,0.2311433256,0.3749023974,0.3837078512,0.1908661425,-0.146335721,0.1729480624,0.1554069221,0.1978396177,-0.0282338094,0.0028656295,0.2605311573,0.1168461367,-0.242686227,0.1361647397,-0.5636625886,-0.0344675593,0.0085772937,0.0550476499,-0.0534494482,0.1351453066,-0.0505981408,0.0278801285,-0.0216870457,0.0026913621,0.2283748835,-0.1127719209,0.1167695895,0.2027147859,-0.3984925151,-0.0443702787,0.0387875065,0.0246722512,0.2728337049,-0.3969203234,0.3170889914,0.0065246797,0.1835428327,-0.1422093809,-0.1942432076,0.1280908883,-0.2384287566,0.1914542913,0.2233331501,0.3583884835,-0.1798888892,0.3329331875,0.3796788454,-0.0558151975,0.076756157,-0.2198263556,0.0660909861,0.0864089876,-0.2148165107,0.342777878,0.0418097116,-0.3922552764,-0.4485302866,0.2099495679,0.3679635227,-0.3328234553,-0.2037753612,0.5172691345,0.1183421761,0.3015903533,-0.0462849289,0.3640754521,0.5558555722,0.0813660473,0.3179981709,-0.1306941211,-0.0043299631,0.0670639873,0.3741815388,0.1950092465,0.0765072331,0.2795681953,0.1557217985,-0.2470973432,0.1012527123,0.099015452,0.2535735071,-0.2710577548,0.2363589257,0.1587246358,-0.1288722306,-0.1093393117,0.0350657515,-0.2995659113,-0.0277637448,-0.0843694657,0.2166175842,0.0319533311,-0.013926561,0.0988720953,-0.0031072681,-0.318824172,0.5872225165,0.2256663442,-0.3283103406,0.1795732081,-0.5215361714,0.3325365484,0.1587109715,-0.4344836473,-0.1075680256,-0.0355027504,-0.0812782422,-0.0342814066,0.5500479937,-0.1895498335,0.1838137358,0.4769846201,-0.2052764148,-0.2011806518,0.1557516754,-0.1988181472,-0.0595369451,-0.1038194448,0.0095436731,-0.4522993863,0.1238261312,-0.0814455077,-0.5590344071,-0.0229633916,-0.0602851398,0.2571274936,0.1017161906,0.3643156886,-0.1040783897,-0.3773122728,-0.1476343423,-0.1132521853,0.0484950729,-0.3111388683,0.0286667272,0.3408141434,-0.3522243202,-0.4135697782,-0.4027699828,0.4307859242,-0.4707098603,-0.4059118629,-0.5092315674,0.3551386297,-0.0586614124,0.0347452834,0.2844442427,0.3230705559,0.2075720727,0.3168869615,-0.406381309,-0.1496329308,0.6324794888,-0.5143606067,-0.0433195867,0.1410471201,0.4048153162,-0.0384845026,-0.1438569129,-0.0502866656,0.0080289235,0.3262680471,0.0352990031,-0.11909578,0.0173804164,0.0460166,-0.0275519267,0.0891542584,0.1260186583,0.0651954263,0.1504261047,0.3225906789,0.0388346836]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"Thanks for reporting @TevenLeScao! I'm having a look...","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":8,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\nThanks for reporting @TevenLeScao! I'm having a look...","embeddings":[-0.3835540116,0.1337701827,-0.0253264122,0.2731699347,0.0476374775,0.0902696028,0.1814594418,0.3819395006,-0.0007448968,-0.0454650521,0.3611760437,0.0126897208,0.0861551389,0.1522721052,-0.1370753795,-0.1786783487,0.2108934969,-0.0611494221,0.0313629135,0.2152216285,-0.233312279,0.2721946537,-0.1955180466,-0.1978287399,-0.4112713933,-0.2598757446,0.1422315091,-0.0248247869,0.0180431418,-0.472040385,-0.0665634125,-0.2579340935,0.4506657124,0.5027949214,-0.0001196416,-0.1828569621,0.3196112514,-0.0112326443,-0.2220180929,0.0707937181,-0.4665146768,-0.3899549544,-0.0209787227,-0.3519911468,0.4165434539,-0.1129655987,-0.3517024219,-0.4365029335,0.6440546513,0.4032406509,0.1420957893,-0.0286113713,0.1127021834,0.0050554201,-0.091488339,0.1034186706,0.0035507113,0.1003285497,-0.1626612246,0.2847627997,0.2777496874,-0.0773247108,-0.1468372941,-0.3109425008,-0.1341963112,-0.0922511294,-0.0620680563,-0.2490646988,0.2073388398,0.3660363257,0.1947740763,-0.2055933028,-0.1659695506,-0.3406728506,0.0919058919,-0.6232780814,0.0507699922,0.5413522124,-0.0108529832,0.1067691743,-0.2495455891,-0.104976207,-0.2370153666,-0.0483339801,0.0022924934,-0.2150741518,-0.107801497,0.0177805834,-0.1525956839,-0.1030657142,-0.0639625862,-0.2661671638,-0.5572221279,-0.0044215643,-0.1861709356,0.0565857217,0.1499928236,-0.4187387228,0.1171992421,-0.0632935688,0.3266349733,-0.3773003817,0.2348451614,0.0594772473,0.5592290163,0.1646710634,0.2710618079,-0.0871628746,0.2600767612,-0.0085000023,0.0787558556,-0.080585584,0.0144567313,-0.2083365619,0.1001313254,0.0720483214,-0.0198398773,-0.1044587344,-0.0545963943,0.2214030325,-0.3935464919,-0.1775218695,-0.2575071454,-0.1779677123,-0.0823953152,0.2208130956,0.1698823124,0.2616044283,-0.1637306511,-0.055641707,-0.0403708592,-0.105635047,-0.0870952979,-0.0314172357,-0.0310927555,-0.1502984315,-0.0088989502,0.2479841709,-0.07146056,-0.2792142928,-0.1221548915,-0.0518075824,-0.1592278332,0.1870426089,0.2745894492,0.1416389495,-0.1817220151,-0.1721425205,-0.0368696339,0.1909364015,-0.4652948678,-0.0441438369,-0.3462507129,0.1497993022,-0.174948737,0.137263,-0.5231258273,-0.0600629412,0.091476731,0.1430438757,0.3687433898,-0.281902492,0.5918431878,-0.2670210898,0.0309650507,0.2506024241,-0.7489979267,0.0754328445,0.1577064991,-0.014049314,0.4526615441,0.3861754835,-0.0494974963,0.3784714043,-0.1160002425,0.2891608477,0.0731833652,0.1971013099,0.0908507481,0.5022193789,-0.1005683616,-0.0170064662,0.0110649858,-0.3200435638,0.6402289271,0.0964221358,0.5203152299,0.4036623538,0.0454398729,0.2351219058,-0.1946289837,-0.1849508435,-0.0352446176,-0.2054119706,-0.3632991016,-0.0579631962,0.1466737837,-0.1935555488,0.3089207113,-0.2521445751,0.1233985275,0.4414643645,0.1443438381,0.1289285421,0.0102021992,0.2916257083,-0.4374501705,-0.1749869883,-0.1071721017,-0.1615067124,-0.1290941387,-0.0509144217,0.2104451507,0.2295134962,-0.0363817848,0.1813212186,0.052343525,-0.1555444002,-0.127720505,0.2019660324,-0.0155298254,0.2325048,0.1545464993,0.0372680686,-0.2765450478,0.2934662998,0.2316015363,-0.1960530877,0.2151839733,0.1256998777,0.1772942543,-0.3326705694,-0.1339832544,0.2177257985,0.0099375555,-0.0327435583,-0.1471460015,0.1095461324,0.1222943515,0.0831131116,-0.1773028076,0.2860653102,-0.044825647,0.0646856949,-0.1169398651,0.5757758617,-0.2219541669,0.3352481425,-0.0439807288,-0.0257019214,0.2293725312,-0.115685977,-0.0543876477,-0.1375717372,0.2854988277,0.1654393226,-0.1292898655,-0.0068243244,-0.3151094019,0.0202219207,0.5460851192,-0.0093529318,0.0469403714,0.1808184236,-0.0217730775,-0.0086931242,0.0634510666,0.0896894187,0.2979865074,0.1401878744,-0.0603684112,0.0903683901,0.2223309577,-0.0778003931,0.2852793038,0.1289278418,0.2591320574,0.2978545725,0.2030513138,0.1441600174,-0.1340728253,-0.3822749257,-0.2039360404,0.0924512744,-0.3180256784,0.3180815279,-0.1785783917,-0.0144326668,-0.4827946424,-0.3543928564,-0.2468315214,-0.3017994165,-0.1766790301,-0.120608367,-0.4763222933,-0.1508393288,-0.1570847183,0.1003781781,0.0947614163,-0.0955487862,0.0521110073,-0.0133182742,-0.1134338975,-0.0862653404,0.0303208008,-0.1828777492,0.0727148056,0.1205032095,0.1148072258,-0.1089555994,0.138138175,0.1159391552,0.011363769,-0.0723430514,0.1234691814,0.24436903,-0.0052500223,-0.5180720091,0.0132895932,0.1782093942,0.0012242054,0.1616311073,-0.0921311304,-0.1736256629,-0.0973561108,-0.0708637834,-0.0394526832,-0.3384982646,0.4255334139,-0.1087423339,0.053168159,-0.0728065148,0.1306462288,-0.0991212502,0.0408131406,-0.305167526,0.20312123,-0.3803762496,-0.0099354219,-0.29374066,-0.2352826148,0.1314555556,0.1669387221,-0.0329655521,-0.0055567622,-0.391361475,0.6218963265,-0.0942029431,0.3956906199,-0.1493091881,0.1626509726,0.116283372,-0.0766213313,-0.0299288798,-0.034230981,0.1094638929,0.1634141356,0.0434522554,0.4693659544,-0.0881285369,0.6779676676,-0.2237532884,-0.1832995564,0.7793295383,0.1484701186,0.2064243555,0.0921519697,0.0852276087,-0.14523682,-0.0044066422,0.1173456982,-0.1005705521,-0.1760220528,0.1545746475,0.0283843968,-0.1508064419,0.0348946676,-0.0198912155,-0.4069148898,-0.3647173345,0.2348921746,0.004411554,0.3050079048,-0.3134261966,-0.0526755564,-0.0389939696,0.0619586632,-0.2127388418,-0.0408007987,0.2097734362,0.1192491576,-0.559535861,-0.4525100887,-0.5020563602,0.2077083439,0.0014540483,0.603766799,0.131311655,0.0917432606,-0.0753988251,-0.1140662283,0.6436045766,0.0058109644,-0.1951618195,0.2375844121,0.386336565,-0.1607325822,-0.1332825124,0.1474809647,0.2159446031,0.8325591087,0.0099383388,-0.3132323027,-0.0252103545,0.2512908876,0.3659793735,-0.1320299953,-0.2465665042,-0.1250897348,-0.187350288,-0.0598954447,-0.24716793,0.0945841148,0.4462836683,-0.1834443063,-0.1511525214,0.0803039446,-0.3116677105,-0.0053784442,0.0105255041,0.2616732419,-0.1018575132,0.1477785259,-0.0853325501,0.3078677058,0.4899508655,0.2527694106,0.0647739545,-0.3482472897,-0.0574569814,-0.2685311139,0.403082937,-0.1060034186,-0.2391371727,0.0764265656,-0.1383740306,0.3870101571,0.0959706828,-0.312458545,0.3040596545,0.1025290415,-0.8577453494,0.0645336658,0.3564459383,0.0776784942,0.268807143,0.1547028422,-0.286196053,-0.4084353149,0.2111502141,-0.2155685872,0.5354143381,0.4569797814,0.1446368843,0.0646170899,-0.3966632187,0.5984712243,-0.050855767,0.0498608612,-0.2009818405,0.4494354129,0.1409459114,-0.1563797146,0.1118669957,0.0670143738,-0.1035585776,0.0396783054,-0.5364384055,-0.2564867437,-0.2484100163,0.0004094007,0.227780968,-0.1971032321,-0.4928751588,0.0050188228,-0.0587676689,-0.230747804,-0.0739439428,-0.2108835578,0.0558624938,-0.1373522282,-0.3768112659,0.1327773184,0.056229651,0.1270544529,0.3438416719,-0.1822942048,0.1789266765,-0.2006658018,0.2654878795,0.1668645293,0.0244745221,-0.0340435542,-0.0723750219,-0.0621858686,0.0925074816,-0.0050350293,0.1176006943,-0.0515903942,-0.1284512877,0.1734403819,0.0692504123,-0.2061031759,-0.0875839517,0.2422014177,-0.2733277977,-0.3255246282,-0.148767665,0.1421758533,0.2509216368,-0.1236157045,0.0389809683,0.18133834,-0.0155195361,0.3283270299,-0.3563518226,-0.1639139205,-0.2968806624,0.0772173479,0.0310481451,-0.3043638766,0.2398792058,0.5180940032,-0.1771533042,-0.0685465857,0.4630976319,0.3336755931,-0.4935486615,0.5193856359,-0.0718338788,0.2705619931,-0.0986742526,0.4814068079,0.0181858055,0.0008819696,0.1572981775,-0.3943596184,-0.1781403273,0.0208502337,0.1891297847,-0.0116522452,0.0509905107,0.5198563337,0.0438094698,-0.2241548598,-0.1480784118,0.0029525445,-0.175141573,0.0353267342,-0.0961995721,-0.0064828023,0.0357723348,-0.0661632195,0.0527311638,0.1101113632,-0.0307108182,-0.1241597459,-0.356960237,0.1330173314,0.1364973933,-0.213485539,0.1574960202,-0.0284412932,-0.2115841657,0.0064735119,0.2217430025,0.0088125607,-0.0912610367,-0.0742899328,0.1118884683,-0.132258594,-0.0885815024,0.1118080691,0.2963478863,0.6302074194,0.0133966263,0.434440136,-0.0908531845,-0.0931121185,-0.4442386329,0.053028442,-0.1838452369,0.379684329,0.1136303768,-0.7344948053,-0.0569514558,0.3135048449,0.1440463811,0.3099168837,-0.2705783546,0.0939457268,0.0643387213,0.1858001649,0.0311663691,-0.1638082415,0.0677069053,0.1059025228,-0.1852017343,0.0332047679,-0.0280226544,-0.3989001811,-0.1950727403,-0.0799626336,0.3378843367,0.1998663396,0.2759836912,0.3507215977,0.1426646858,0.1734568626,-0.1532565504,0.2373753935,0.2045940161,0.07292445,0.1856802106,-0.1088075638,0.6076552272,-0.1315846294,-0.1782245338,-0.529419601,-0.0858905241,0.0434539132,-0.1308633238,0.3036104739,0.1525634676,0.1237440184,-0.1487126946,-0.0867675394,-0.1235088259,0.1483534276,-0.1321201771,0.0491536967,-0.2314487696,-0.0158879142,0.1822529137,0.1275390834,0.0311901998,0.2131751776,-0.2341464758,0.2386569977,-0.0199750867,-0.4306848049,0.1308334321,-0.0980624855,0.1991312206,-0.2744833529,0.0886746943,0.1754304767,-0.0585307926,-0.1104223579,0.129961744,0.5970198512,0.297948271,0.2625238299,-0.1205175295,-0.183571592,0.1180144474,0.097236976,0.1499893814,0.2311720401,-0.0357310176,0.1872949153,0.0442537479,-0.0637615472,0.3494564891,0.1730463356,0.0851096511,0.0473708361,0.0976001918,-0.1405543685,-0.199941799,-0.4745958745,-0.2003811598,-0.2328689098,-0.0266454965,0.395747155,-0.1033239439,-0.1024453938,-0.0751891583,0.0330789909,-0.2828265131,0.4124233723,0.2536088526,0.3815958202,-0.1085981354,-0.2012482733,0.1328729689,-0.0740684941,-0.3343696296,0.3943842053,0.3165627718,0.290238589,-0.2557206452,0.0196303912,0.2551514506,0.609011054,-0.4524421394,0.0580099858,-0.4037280977,0.0743918568,0.4380075037,0.0958239734,-0.1570504308,-0.1600841582,0.4561393559,-0.0802299827,-0.0740605593,-0.1446939558,0.1340744644,-0.0343763605,0.0443297401,0.059809085,0.4057221711,0.0951117724,-0.0312908664,-0.2445305288,-0.3060822785,-0.1973097622,0.05498299,0.2535021901,0.2860682011,0.5162214637,-0.1113298088,-0.0169925131,-0.2252942175,0.171865344,-0.0074566673,-0.4126228988,-0.1405547857,0.0241691228,-0.1212096885,0.0533396639,0.2411208749,0.2424975634,-0.090981178,0.2419409603,0.0147854043,-0.2373495996,0.0959351584,-0.4324303865,-0.3282380104,-0.0428509712,0.3079414666,0.2080894113,-0.0619675033,-0.0406340808,0.1646510214,0.3692736626,-0.1553759724,-0.3649173677,0.2631973326,-0.1553139538,-0.0984587297,0.2267612368,0.3285340369,0.2966492176,-0.4224770963,0.2380135059,-0.1086261123]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"(not sure what just happened on the assignations sorry)","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":9,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\n(not sure what just happened on the assignations sorry)","embeddings":[-0.3850312829,0.1336508542,-0.0222392716,0.2809231579,0.0494918488,0.0855972543,0.1737643629,0.3757632971,-0.0042911093,-0.0327521637,0.3524527848,0.0065437555,0.0792659447,0.1389908493,-0.1420239955,-0.176225394,0.2000755072,-0.0662752241,0.0317878909,0.2177524269,-0.2450545579,0.2635695636,-0.1906572133,-0.1829949021,-0.4161287844,-0.2705082595,0.1510763913,-0.0386185274,0.0153434724,-0.4797801077,-0.0791374296,-0.260230273,0.4474053681,0.4944441915,-0.0001198757,-0.1843488514,0.3177504838,-0.0145362904,-0.2182610631,0.0858289972,-0.4662579894,-0.385001123,-0.0246721655,-0.364887923,0.4111401439,-0.1046407297,-0.3370763063,-0.4342477918,0.6482345462,0.4108015895,0.1415202469,-0.0388237871,0.1346869171,-0.012388682,-0.0648260117,0.1095590293,0.0188400242,0.1173433214,-0.1616414934,0.2906214893,0.2640044689,-0.0776768476,-0.1521448493,-0.3226140141,-0.1184219494,-0.1008546352,-0.0571243018,-0.2483423054,0.2025732398,0.353903234,0.1894200891,-0.2152506113,-0.1574115008,-0.3328281343,0.0940524414,-0.6301178932,0.0508323349,0.5465570092,0.0018558241,0.1108500957,-0.2197962403,-0.1112104654,-0.2393992841,-0.0380409174,0.0098487213,-0.2139936686,-0.1000095159,0.0226934087,-0.1494220644,-0.0973783284,-0.0766340122,-0.2599663436,-0.5644696355,-0.0024706377,-0.1934744269,0.0684020743,0.1358557791,-0.4293648899,0.1279627383,-0.0603997931,0.3198116124,-0.3849320114,0.2455670536,0.0673920289,0.5500015616,0.1777159423,0.2722126842,-0.0708324239,0.2516005933,-0.013420919,0.0817655548,-0.083527185,0.0218745153,-0.2030589283,0.0950110704,0.0772202611,-0.015972577,-0.0861388668,-0.06112279,0.2182668,-0.3901385367,-0.1800228059,-0.2523075342,-0.1608887166,-0.0897493958,0.2074467093,0.1661572903,0.2688969672,-0.1522171497,-0.0544557087,-0.0465131924,-0.1007597372,-0.0936505869,-0.0327996463,-0.0457952432,-0.1460981518,-0.0094888369,0.2383956462,-0.0817741677,-0.2882793844,-0.1192425936,-0.064165093,-0.163821891,0.1707389951,0.2650295496,0.1394260973,-0.1795890927,-0.1722781807,-0.0379654467,0.1900897622,-0.4805969298,-0.0492987707,-0.3573263884,0.1473133564,-0.163469851,0.1444828808,-0.5233575106,-0.0663693249,0.0956658274,0.146043539,0.3714286089,-0.2875481546,0.620562911,-0.2777485549,0.0342246071,0.2577074766,-0.7198041081,0.0688427836,0.1442873478,-0.0324333943,0.4723386765,0.3893199563,-0.047522217,0.3747797012,-0.1058199927,0.2845960557,0.0934914649,0.2011472285,0.081085518,0.4924468696,-0.1044526398,-0.0271705166,0.0236796066,-0.3020902574,0.6519228816,0.1099848747,0.521494329,0.3980256617,0.0566930547,0.2394725829,-0.1905330569,-0.1896248609,-0.0330159143,-0.2138754874,-0.354141295,-0.0530352555,0.1533979326,-0.2068889737,0.3096469641,-0.2489516437,0.1344087571,0.4555870891,0.1456087232,0.1409719139,0.0103083998,0.2925568521,-0.4388349056,-0.1872715652,-0.1338905394,-0.160936445,-0.120594129,-0.0512346551,0.2045447677,0.2352252305,-0.0294671282,0.1842304766,0.0504001491,-0.1431951821,-0.1242702305,0.2058728635,-0.0166799203,0.2279495746,0.1462930441,0.045178175,-0.270825386,0.2935502529,0.2251351774,-0.2041480094,0.2169742286,0.1226722822,0.1799077541,-0.3267632425,-0.1335704774,0.2130627036,0.0142314378,-0.0523956232,-0.1565363407,0.106087476,0.1229233146,0.0667832121,-0.1815522313,0.2818332911,-0.0375705212,0.0740894303,-0.1102848649,0.5657496452,-0.2413817793,0.3265841901,-0.0648201779,-0.032751862,0.2320103943,-0.1115841046,-0.0479868315,-0.1392869204,0.2896444201,0.1730006188,-0.1323317736,-0.0003595479,-0.3256942332,-0.0055628177,0.5513115525,-0.0029183482,0.054534018,0.1938970983,-0.023308862,0.0039018749,0.0536086708,0.0748823658,0.2807109952,0.1383168846,-0.0688075051,0.0898337662,0.230294928,-0.0713350475,0.2849713266,0.1261349916,0.256624341,0.3118025064,0.2000617236,0.1549509615,-0.1341328621,-0.3820091188,-0.2045339197,0.1088437736,-0.3189269006,0.3047693968,-0.173079893,-0.02318017,-0.4876317084,-0.3532818556,-0.2595121562,-0.2921046615,-0.1759457439,-0.1288535148,-0.4672134519,-0.1410895437,-0.130323112,0.0979244336,0.0952453539,-0.0954186022,0.0429055616,-0.0209062304,-0.1098111123,-0.0916538015,0.0324942917,-0.1860236377,0.0783969834,0.1259966642,0.1092937663,-0.1107341722,0.1385951489,0.1032965481,0.0127786277,-0.063948907,0.1075756475,0.2519796193,-0.0012037269,-0.5285543799,0.0061785094,0.1885720789,0.0025107549,0.163725391,-0.1106504947,-0.1794643253,-0.0912223905,-0.0787258968,-0.0451821275,-0.3405896723,0.4451104105,-0.1060839444,0.0639376789,-0.0866711885,0.1318916678,-0.1076742038,0.0211135484,-0.2907147408,0.2117835283,-0.3773761094,-0.0206548367,-0.2790070176,-0.2276432663,0.1204475239,0.1580452621,-0.0395865701,-0.0041998946,-0.3872665465,0.6198440194,-0.0979223251,0.3970405161,-0.1527565271,0.1691839695,0.1109840572,-0.0595563874,-0.0250740647,-0.0203645546,0.0961131155,0.1669696271,0.0574165732,0.4714373946,-0.0832632631,0.6724117994,-0.2386009842,-0.184216693,0.7880532742,0.1492850929,0.219355464,0.0877366588,0.0763441622,-0.1440263838,-0.0085032061,0.1176889539,-0.0922783539,-0.1655730605,0.1498000771,0.0018457406,-0.1500566155,0.0389741026,-0.0255914759,-0.4380201101,-0.3653962016,0.2375136167,0.0146769593,0.3148706257,-0.3269868791,-0.0636758357,-0.0385371819,0.0624481067,-0.2145822048,-0.0334292315,0.2076183856,0.1283064783,-0.5585277677,-0.469091922,-0.4884331822,0.2143175155,-0.0053679403,0.5926195979,0.1305274367,0.106354855,-0.0672883019,-0.1183417141,0.6427695155,0.0116234366,-0.1974219531,0.2286448032,0.3985086381,-0.1601184011,-0.124441281,0.1419327259,0.2035093904,0.8284231424,0.0239006225,-0.3348000646,-0.0210007597,0.2567728758,0.3658872247,-0.1377760023,-0.2441083938,-0.1360807866,-0.1880560815,-0.0832224563,-0.2317319661,0.0901478454,0.4567770064,-0.1638862193,-0.1482391506,0.0728263408,-0.3071516752,-0.0058071939,-0.0014285228,0.2642574608,-0.1020807773,0.1350469887,-0.0933150128,0.3134117126,0.5068423748,0.252395153,0.0692903772,-0.365388006,-0.0598377436,-0.2572088242,0.4201459289,-0.1122015715,-0.2310388535,0.0890044272,-0.1339729428,0.3940905631,0.1061567888,-0.2943665683,0.306410104,0.0962400362,-0.8758054376,0.0695648417,0.3481321633,0.0772551894,0.2612735629,0.1553052664,-0.2824742794,-0.4007288218,0.202693224,-0.1960925907,0.5578150749,0.4399380088,0.1302335113,0.0661972314,-0.4116248488,0.6040495038,-0.0571329631,0.0589463674,-0.2209350765,0.4581719041,0.1536133736,-0.1691449881,0.1134070978,0.0547165237,-0.1088650003,0.034288507,-0.5406997204,-0.2531602681,-0.2511471212,0.0046398691,0.2317053378,-0.1918942928,-0.4825690687,0.0034326299,-0.0776367486,-0.2312682718,-0.0759823769,-0.2130475938,0.0559753217,-0.1397358626,-0.3676080406,0.1299230307,0.0314646587,0.118100889,0.3472825885,-0.1753893048,0.1705048084,-0.1958725601,0.2539284527,0.1711360216,0.0439112671,-0.0382990688,-0.0937756374,-0.0769142732,0.0808967501,0.0019744171,0.1068714038,-0.0555433817,-0.1470002383,0.1872856021,0.0597732589,-0.206777826,-0.089513883,0.2416596413,-0.2600375116,-0.3182446361,-0.1365777999,0.1381595284,0.2456513792,-0.1214976907,0.0367315933,0.1739423722,-0.0177370366,0.334043026,-0.3549747169,-0.1737840921,-0.2984444797,0.0784650296,0.0435447432,-0.2981089056,0.2398856729,0.5044230819,-0.17605187,-0.0707599521,0.45676139,0.3281340301,-0.5052844286,0.5219472051,-0.0627808869,0.2634769678,-0.1009146273,0.483628422,0.0315853208,0.004446432,0.1660699993,-0.3849200904,-0.1894566566,0.0319586843,0.1947980374,-0.0084007178,0.065283224,0.5416654944,0.0318242796,-0.2130215317,-0.1481459141,-0.0046157157,-0.1840322018,0.0457329601,-0.0899663419,-0.0163104124,0.0306232907,-0.0538033284,0.0509572066,0.105406113,-0.0251506418,-0.1230851933,-0.3603518605,0.1360474974,0.1325486153,-0.1987965554,0.1411616802,-0.0424824655,-0.2109416276,0.0025160476,0.2281491309,-0.001296035,-0.0976916179,-0.0616121553,0.1148051023,-0.1342394501,-0.0909421295,0.1143978089,0.2908406556,0.610571146,0.0017733116,0.4406329691,-0.0928368047,-0.103048943,-0.4589345753,0.0610997193,-0.1747564226,0.3796329796,0.1077839434,-0.7346025705,-0.0652302951,0.3187016845,0.1218255088,0.291770488,-0.2489936054,0.0967937335,0.0636612996,0.1852182746,0.0268570017,-0.1593598723,0.0649536699,0.0920213461,-0.1827933639,0.0239231195,-0.0371742547,-0.3983514607,-0.2042734027,-0.0860919058,0.3431583643,0.2068064809,0.2700003386,0.346044302,0.1599700004,0.1550939232,-0.1450005174,0.2510818839,0.2142534256,0.0525508001,0.1905857325,-0.1113353968,0.6105297208,-0.1368077397,-0.1807514131,-0.5275430679,-0.0896375328,0.0444049314,-0.1312000453,0.3005880117,0.1448160559,0.132022649,-0.1285357773,-0.0884119496,-0.120670855,0.1622865051,-0.1364792287,0.0542699434,-0.2498229444,-0.0138613284,0.1867912114,0.1191922873,0.0348286889,0.2186621279,-0.2226133943,0.2442909032,-0.0275606867,-0.4321265817,0.1330662668,-0.0776671395,0.206802845,-0.2787876129,0.0989161655,0.1792585403,-0.0493160672,-0.1084615216,0.1296239942,0.5975341797,0.2879384458,0.2630448341,-0.116804339,-0.1898532063,0.1139365435,0.0915242657,0.1449324191,0.2361572981,-0.0262196604,0.1862749308,0.0458872989,-0.0669771433,0.3344222605,0.1679669619,0.0776584372,0.071095407,0.0824128836,-0.1429105699,-0.2161936015,-0.468067795,-0.1997461021,-0.242094487,-0.0212399215,0.3961208463,-0.1061236858,-0.0989791453,-0.0720583051,0.0314672254,-0.2802953124,0.4021660686,0.2535834312,0.3819744289,-0.1173895448,-0.2043876648,0.1432851106,-0.0628123134,-0.3248874843,0.3996943533,0.3189509213,0.2806887627,-0.2690669894,0.0102982344,0.2537681758,0.6098229885,-0.4445378184,0.0664372146,-0.4033975899,0.0906638876,0.4471966326,0.0960636213,-0.1446422189,-0.1575000584,0.4551168382,-0.0880476385,-0.0738983825,-0.1437771469,0.1328723133,-0.0407196656,0.0470245294,0.0651584342,0.4005176723,0.1017626002,-0.0379070826,-0.2413177639,-0.296497196,-0.207551077,0.0579809844,0.2509723604,0.287204653,0.5164981484,-0.0998048037,-0.0131835481,-0.2307984829,0.1811380833,-0.001199634,-0.3998931646,-0.1429832131,0.0243962128,-0.1198367849,0.0517549068,0.2366981655,0.2413293272,-0.0691984817,0.2536450922,0.0123597803,-0.2277589738,0.0981341302,-0.4194343686,-0.3223852217,-0.0399646498,0.3124932349,0.2089057863,-0.0502399579,-0.0374595933,0.1735438704,0.3814381361,-0.1568620354,-0.3749901652,0.2588804662,-0.1553317904,-0.0980313793,0.2286963016,0.3107569814,0.2917194366,-0.4121640027,0.23805435,-0.1093736514]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"For some reason this happens (both `datasets` version are on master) only on Python 3.6 and not Python 3.8.","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":19,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\nFor some reason this happens (both `datasets` version are on master) only on Python 3.6 and not Python 3.8.","embeddings":[-0.3504133224,0.1333585531,-0.0069016316,0.3097095788,0.0389037989,0.0805483982,0.2188645303,0.3883448839,0.0006037177,-0.0402306095,0.3231003284,0.0345943347,0.0673164427,0.0446980782,-0.168024078,-0.1674068421,0.2194828242,-0.0308502372,0.0336684771,0.2225417495,-0.2770715058,0.3025195897,-0.1921600103,-0.2260363847,-0.3988591135,-0.2168714553,0.1474183947,0.0140270805,-0.0263657849,-0.5282495618,-0.0391146727,-0.2388433814,0.4440392554,0.5343334675,-0.0001208482,-0.1544707865,0.333599031,0.0061301352,-0.2074822038,0.0264416467,-0.4390989244,-0.32505247,0.0626767799,-0.3321589828,0.4254887998,-0.043325413,-0.356161654,-0.4652174115,0.6484056711,0.4455591738,0.1272598058,0.0488448367,0.0954690576,-0.0134071819,-0.0303338505,0.1398149431,-0.0270308219,0.1463216394,-0.1673423201,0.2741565406,0.2642328143,-0.0571525879,-0.1586708724,-0.2541062534,-0.119923979,-0.085410364,-0.0593052246,-0.2858139575,0.1833256185,0.4000854194,0.1188576818,-0.1941767186,-0.2122763246,-0.3544609249,0.1109925061,-0.5715143085,0.0593010001,0.5090208054,-0.0011916923,0.1080368087,-0.2637678385,-0.114410229,-0.2154253274,-0.0444664508,0.0033301797,-0.1853815615,-0.1245927736,0.0236596912,-0.1236833036,-0.0916753635,-0.0017678378,-0.2792039514,-0.5480955839,-0.0483755842,-0.1531420797,0.0719255358,0.1683174223,-0.41027686,0.0765405893,-0.1706936359,0.2756137252,-0.3768747449,0.2522144914,0.0399434194,0.5629569292,0.1241348237,0.3155882359,-0.0116557395,0.2348292768,0.0134433247,0.0630376115,-0.0512274355,0.0182550773,-0.1685528904,0.1308631897,0.0527191199,0.0156177739,-0.0881923288,-0.0524228923,0.1799780875,-0.3575305343,-0.1791661978,-0.2505482733,-0.1906233877,-0.0992806479,0.2618553042,0.1635038704,0.2956346869,-0.1717634797,0.0063081384,-0.0204488989,-0.1102865338,-0.0897183716,-0.0236233976,-0.073195003,-0.2337141484,-0.0126970643,0.2607291639,-0.0713082403,-0.2439553589,-0.0935669243,-0.0429422632,-0.1192921028,0.2015205622,0.3051171303,0.1432213038,-0.1321808249,-0.1420604289,-0.0551187806,0.1836929768,-0.44878003,-0.0509726219,-0.3994896412,0.1303583235,-0.1831211299,0.1322419196,-0.4531036913,-0.0757280588,0.1063631997,0.1056371853,0.3775993288,-0.3547270596,0.5717495084,-0.2804208994,0.0224844217,0.2825922072,-0.7132418156,0.0905744135,0.0787857249,-0.018038623,0.4428821206,0.3599768877,-0.0533611178,0.3268298507,-0.1179183424,0.2507768273,0.1492944807,0.1822726578,0.0733188763,0.4902765751,-0.075096488,0.0082652271,-0.0178810563,-0.3451776803,0.6365368962,0.1081678718,0.5238189101,0.378954649,0.0567399897,0.247143209,-0.1749355942,-0.1934174299,-0.058229696,-0.1998623759,-0.3382233083,-0.0477836467,0.1436721087,-0.2043278515,0.3136581779,-0.2468699366,0.1187639832,0.4237191677,0.1386943907,0.1213777065,0.0007520858,0.2297507823,-0.4470793903,-0.1484595537,-0.1031326875,-0.1582561135,-0.1408949494,-0.0262280516,0.2096289694,0.2228936702,-0.0618911646,0.1483442932,0.0206855722,-0.1179929227,-0.0828323588,0.2088543773,-0.0439206995,0.1072661281,0.1892146617,0.0468167737,-0.2518898249,0.3014494181,0.2420794815,-0.2325034887,0.2049761862,0.1209193617,0.1598200649,-0.3047866523,-0.1392458826,0.2068885416,0.0553343631,-0.0437620915,-0.2117630988,0.0500013866,0.1228865013,0.1427299976,-0.1562665254,0.2958237231,-0.0600599162,0.0940000713,-0.0663738996,0.6036195755,-0.1680673957,0.3515254557,-0.0444824845,-0.0093329754,0.204870522,-0.1004328355,-0.0489793755,-0.2392224967,0.2890396416,0.2199492753,-0.1189383641,0.035022106,-0.2944984734,0.0433705114,0.6203617454,0.0286877751,0.0749300122,0.2449778318,-0.0279280301,-0.0106857792,0.0029470548,0.1193481237,0.34180516,0.1002836004,0.0062348549,0.1227353662,0.2248274833,-0.0821276978,0.2882922888,0.1530649364,0.3440402746,0.297332257,0.1898810863,0.1800428331,-0.1405833066,-0.3945007026,-0.2151729912,0.0726259276,-0.3205601275,0.3570815623,-0.181052506,-0.0075610415,-0.4912694693,-0.4257619977,-0.2852296531,-0.3221094012,-0.1869159192,-0.117719993,-0.4756629765,-0.1666974425,-0.1625766903,0.0679481179,0.1090966687,-0.062014658,0.0807449296,0.0071076346,-0.1285309643,-0.1226083934,-0.0314302854,-0.2198490202,0.0281818788,0.0962631404,0.09539105,-0.0799174607,0.1263902634,0.1485719532,0.0225120932,-0.0986567363,0.1772813201,0.1919338256,0.0379259177,-0.5070132017,0.044114504,0.1346117854,0.0048047537,0.1715757996,-0.0695156157,-0.1623263955,-0.0968115181,-0.0530371778,-0.0232678726,-0.3681439459,0.4280723035,-0.105002597,0.0783650875,-0.0927589759,0.1251449585,-0.0520538427,-0.0536761992,-0.2879790962,0.2134601921,-0.440947175,0.0076127052,-0.2837551236,-0.2592487037,0.1172973439,0.1513832062,-0.0292772949,0.0501040556,-0.3940513432,0.507517457,-0.1001565978,0.3795506954,-0.123145014,0.1970722824,0.1210306212,-0.0695032999,-0.0169989299,-0.0407808423,0.110844411,0.1125232428,0.0556513295,0.4821352065,-0.0872536749,0.7164244056,-0.2381183803,-0.1243974194,0.8051139712,0.162354365,0.2087133527,0.0996946692,0.1702729613,-0.1429881901,-0.0274126045,0.0307171643,-0.123238042,-0.1478364617,0.1717360318,0.034560442,-0.1530414373,0.0378615372,-0.0108109387,-0.3851906359,-0.376285851,0.2383833677,-0.1005229503,0.3272312582,-0.3106667995,0.0180654135,-0.0260734912,0.0866981,-0.2485881448,-0.0391858071,0.2262448221,0.1218070462,-0.5017971992,-0.3746839166,-0.4901303947,0.2143228352,0.0604575202,0.6483105421,0.1418084055,0.0654700324,-0.0685712993,-0.0753924996,0.6379687786,-0.0187666211,-0.2247714698,0.245076552,0.3232691586,-0.2799645066,-0.1315702051,0.1515457779,0.2155806869,0.8975403905,0.0597874522,-0.2848918438,-0.0003903492,0.2221137285,0.310757488,-0.1043004692,-0.2652477026,-0.1747551262,-0.1859595776,-0.0746014267,-0.2814921439,0.1014089733,0.4137970805,-0.1576348096,-0.1487754732,0.0687109903,-0.3437816203,-0.0412046425,0.0084780995,0.2232956737,-0.0842653513,0.1385912299,-0.0246471073,0.2563930452,0.4521253705,0.3058808744,0.0356718525,-0.446742177,-0.0822937861,-0.2442297339,0.377129972,-0.0825352669,-0.215109244,0.1014142409,-0.0907824039,0.3834989965,0.1219995022,-0.3035013378,0.3149876297,0.0513521843,-0.8291917443,0.0223533548,0.3281481266,0.0547031276,0.2725344598,0.1803308278,-0.2888335288,-0.3916205764,0.175371021,-0.1807685196,0.5057501197,0.4422088563,0.1994864196,0.0818068609,-0.4618529975,0.6346319318,0.1170891374,0.0573921762,-0.2433636039,0.4591765404,0.1332220584,-0.1751398146,0.1018057093,0.0069232634,-0.101901181,0.0277276337,-0.5292726159,-0.290435791,-0.2342902422,-0.0197921805,0.1661034077,-0.2177004963,-0.519983232,-0.0114516206,-0.0364235081,-0.2229024023,-0.0615968704,-0.2148885578,0.08854267,-0.1690396965,-0.4068056047,0.1753613502,0.0427624658,0.1383465081,0.3813124299,-0.2322726399,0.1784357131,-0.1881958693,0.2385398895,0.1239833757,0.0410763361,-0.0559253395,-0.0737752691,-0.0862133205,0.103248544,-0.0203391947,0.1611801535,-0.0562096834,-0.1563476175,0.162219435,0.0425417423,-0.2514265776,-0.0624584556,0.2822570503,-0.2384875268,-0.3111319244,-0.1231714487,0.0842910483,0.2433489412,-0.1255088598,0.0237615909,0.2197463363,0.0179516934,0.3015409112,-0.3990695775,-0.202407673,-0.2789299488,0.1237437204,0.0286224093,-0.2644253373,0.2547756135,0.5522825122,-0.1724933535,-0.054852806,0.4613583982,0.2994168997,-0.4373067021,0.5468577147,-0.1161938235,0.2572834492,-0.0647379532,0.4393739998,0.0246309582,-0.003930035,0.1056320146,-0.4127843678,-0.1596419662,-0.0561250523,0.1559865028,-0.0151458001,0.0502725765,0.4513688982,0.0733486712,-0.2226619273,-0.1254234463,-0.000381679,-0.1558341235,0.0274691097,-0.0910618454,-0.0516920127,0.0702706277,-0.0788919553,0.0237131529,0.0943355635,-0.042736318,-0.0913175642,-0.3859690726,0.1511209458,0.1576991975,-0.2384630442,0.1840201914,-0.0280541852,-0.2168807536,-0.0194976311,0.2399761826,0.0084568616,-0.1093645468,-0.0348289348,0.041755043,-0.1395688504,-0.0450161919,0.0915858969,0.2478062958,0.6310987473,-0.0039616507,0.4617322385,-0.0837521851,-0.0817984641,-0.4447195232,0.0365296341,-0.1844370812,0.3684317768,0.1258168817,-0.7131573558,-0.0333928876,0.3153312206,0.1713236272,0.3518637121,-0.2561754286,0.0928815231,0.0609294251,0.149395436,0.0489133969,-0.1838422567,-0.0148720909,0.1542601734,-0.2073802799,0.0270512737,-0.105536595,-0.3753692806,-0.2184968442,-0.0562068708,0.3148269355,0.196940586,0.30100739,0.3379613459,0.1170072183,0.1715703309,-0.0953869522,0.2353029698,0.1874186099,0.0852255747,0.2191600353,-0.0530835353,0.5097365379,-0.1636622548,-0.1327753067,-0.5077123642,-0.0629225001,-0.0243233033,-0.1108370796,0.2969662249,0.1731534004,0.1295631975,-0.2328277975,-0.0546791628,-0.1317437738,0.1354822516,-0.114602156,0.0271173827,-0.1979801357,0.0300390851,0.2139832973,0.1403457373,0.0250710025,0.2218130827,-0.1614846736,0.2566998899,-0.014826268,-0.4643138647,0.1446292847,-0.0962818637,0.2030678391,-0.3048176467,0.0698810071,0.1747236103,-0.0594311766,-0.1174420789,0.12700966,0.6228163242,0.3317955434,0.2254386544,-0.1523679197,-0.1855991632,0.1193912774,0.0882173777,0.1380416751,0.2320008874,0.0738767311,0.1868025064,0.0277626645,-0.0597374476,0.3779555261,0.125699237,0.1086775213,0.0675298125,0.2136912495,-0.1539224982,-0.1720199734,-0.4478203654,-0.1431650221,-0.2603188157,-0.0390319787,0.415571332,-0.0776903704,-0.1089188755,-0.111005567,0.0211324766,-0.2312806994,0.4254668355,0.3163938522,0.3585678637,-0.0846127495,-0.1597926468,0.1236490309,-0.0332793184,-0.3323061466,0.3897259235,0.3716337979,0.2889091372,-0.2752666473,0.0416816808,0.2435319126,0.6358326077,-0.5123924613,-0.011497872,-0.4014294147,0.0414297543,0.4235149324,0.0521733575,-0.1359284967,-0.1739708483,0.4652428627,-0.1783279777,-0.0755835921,-0.1921002269,0.1029091924,-0.0304345749,0.0116674704,0.0697250292,0.4269153476,0.0840982422,-0.0229895804,-0.2913000584,-0.346473366,-0.1918723583,0.0432452634,0.2732582092,0.2471267879,0.5369346142,-0.1444123983,-0.0654276311,-0.2570613623,0.2039139122,-0.0345538892,-0.3771499693,-0.0903435275,0.0467704237,-0.0981632322,0.0540509559,0.2709196508,0.2368739098,-0.0699155256,0.2132071704,-0.0019537383,-0.2278083265,0.0939574242,-0.4700144529,-0.3610374331,-0.047168985,0.2895595729,0.1970520169,-0.0877747685,-0.1467585564,0.163264811,0.3798914552,-0.1767860353,-0.3856873214,0.2137864083,-0.1509815902,-0.0710370913,0.2052475214,0.3572631776,0.265460521,-0.4532501698,0.257243365,-0.0953833237]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"@TevenLeScao we are using `pandas` to serialize the dataset to JSON Lines. So it must be due to pandas. Could you please check the pandas version causing the issue?","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":29,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\n@TevenLeScao we are using `pandas` to serialize the dataset to JSON Lines. So it must be due to pandas. Could you please check the pandas version causing the issue?","embeddings":[-0.3529362082,0.1344455481,-0.013971135,0.2699111104,0.0913660303,0.0805774182,0.2758990824,0.4209209979,-0.0012249912,-0.0546875075,0.3439740539,0.0168313999,0.1050095931,0.1663783789,-0.212326929,-0.2198456228,0.2366723865,0.0008698059,-0.0046996996,0.2111837864,-0.2990753651,0.2603419125,-0.1727563441,-0.1527426541,-0.4317580462,-0.3298732638,0.0938417837,-0.0441338681,0.0058596348,-0.472507447,-0.0423331596,-0.2630417645,0.4676561952,0.5261772275,-0.0001195658,-0.1944762319,0.3220655024,0.0135893067,-0.1867634207,0.0544615015,-0.4758793116,-0.3755782247,0.0264103934,-0.3169336021,0.4677455425,-0.1366833597,-0.3192306161,-0.3943061829,0.660025537,0.3750503063,0.1458732486,-0.0116166044,0.1228370667,0.0027563083,-0.1855171621,0.1420014501,0.0120759271,0.1403747201,-0.1949197501,0.2560220659,0.2439377159,-0.0405929051,-0.1299279332,-0.2738689184,-0.1651043743,-0.1033715084,0.0106685227,-0.2650201917,0.222212106,0.3889644742,0.2334923893,-0.1778819114,-0.2301101238,-0.3413921297,0.0876553357,-0.5814914703,0.0522426181,0.4974866509,0.034824919,0.1073150113,-0.1599738449,-0.09862037,-0.2680465281,-0.0571152084,-0.0128136175,-0.2351782918,-0.1353159398,0.0022890316,-0.2164380252,-0.1658328176,-0.0493835434,-0.2810891271,-0.480743289,-0.0021582243,-0.2009210736,0.0675941333,0.1232376322,-0.5326524973,0.0609980859,-0.0570523851,0.3082892299,-0.3318081796,0.184278518,0.0480579287,0.5355818868,0.2027215362,0.2703160346,-0.0536257997,0.2857007086,0.0064849304,0.0928449035,-0.074902907,-0.0085742287,-0.2188671976,0.0825398639,0.0149517981,0.0333818607,-0.0884099975,-0.0457932614,0.2381830364,-0.3975786269,-0.1502700895,-0.2566150725,-0.1565059572,-0.103684634,0.3045494854,0.1440952569,0.2482615411,-0.0999802724,0.0432455242,-0.0405064486,-0.0961541906,-0.0510437451,-0.0417152904,-0.0598883927,-0.1935311109,-0.0223588943,0.2295279652,-0.1399097741,-0.2545557618,-0.1193480939,-0.0958460048,-0.1304164231,0.1919517219,0.2618643045,0.1400866061,-0.1878334135,-0.163624391,-0.0112469327,0.1853886992,-0.4196414351,-0.0066012368,-0.4253169894,0.1383522749,-0.1044838354,0.136747092,-0.4679190814,-0.0312343128,0.1029534191,0.1681160033,0.4023908079,-0.2843471467,0.5852193832,-0.273724854,-0.0134121003,0.2795955837,-0.7273977399,0.0780293494,0.1663562506,-0.0066934167,0.4735563993,0.4228544533,-0.0087973019,0.3183481097,-0.083881624,0.2450965196,0.0956274271,0.1964691579,0.0790034309,0.584322989,-0.0501275435,0.0050995774,0.0808039308,-0.3214190304,0.598856926,0.1140356958,0.4622335732,0.3923937976,0.061467912,0.2183372825,-0.1816774309,-0.2373981625,0.0264629759,-0.2356802225,-0.2884986699,0.0058600386,0.1468087435,-0.1964570284,0.3013794124,-0.2460617423,0.1070937291,0.421831876,0.1388188452,0.1713828593,0.0404157192,0.2613672912,-0.3996379673,-0.2026008815,-0.1755854934,-0.1705565453,-0.1746134162,-0.06112957,0.1994912922,0.2102854401,0.03375021,0.1846895665,0.0434003659,-0.1528932601,-0.1080069318,0.2392733991,-0.0202543754,0.2720686495,0.1215363964,0.0703212246,-0.287720114,0.289414674,0.1951189339,-0.2047723234,0.2890503705,0.1089512631,0.1802468896,-0.3298682272,-0.1243562996,0.2506777346,-0.0237743761,-0.0298403632,-0.1105632186,0.0919209495,0.1511948705,0.1865115464,-0.185196504,0.2800279856,-0.0211608093,0.0737264603,-0.1132216007,0.564217627,-0.2457125485,0.2829084694,-0.0625573918,-0.0128166247,0.3005697131,-0.0625316426,-0.0879970044,-0.1714465022,0.2660799325,0.1631940305,-0.1138843596,0.0562230274,-0.4401580095,-0.0516484156,0.5020251274,-0.0390517041,0.1236082837,0.2182702273,0.023058584,-0.0328239761,0.0820230469,0.0920968652,0.2714562118,0.1238182634,-0.0892974138,0.0432210565,0.1545457244,-0.0774390697,0.2806868553,0.1741724312,0.2688159943,0.2540608048,0.1505658925,0.1775213182,-0.1411738545,-0.3314296007,-0.1864497811,0.111568436,-0.3104502559,0.3321208954,-0.2076403797,-0.0999990553,-0.4491662681,-0.3982090652,-0.2386234999,-0.2300269008,-0.2367467433,-0.1190832257,-0.4333932102,-0.0752878189,-0.1449396312,0.0991034955,0.1697894931,-0.0045734351,0.0295294896,0.0061937319,-0.0869575962,-0.0757521391,0.0253164489,-0.1559317708,0.0153033612,0.1156513169,0.170398429,-0.0613914318,0.1679659337,0.0688275769,-0.0236092228,-0.1093089134,0.1402525008,0.2671134174,-0.05519595,-0.5096024275,-0.0115265194,0.1602485031,0.0281435661,0.1368238926,-0.1092099324,-0.2188175023,-0.0702953041,-0.0341573693,-0.0018332986,-0.3069933653,0.4799240828,-0.0855262056,0.0724559873,-0.0527044535,0.1263266355,-0.0759281069,0.0402933769,-0.2723293304,0.1749473661,-0.4066644013,0.0378724411,-0.2577102184,-0.2774894536,0.1150122434,0.144803077,-0.0473119766,-0.0669016168,-0.3617930412,0.5396341085,-0.1307225674,0.3153801262,-0.1509942561,0.1408164501,0.1063914597,-0.0515443981,-0.0481923334,-0.0095522534,0.1201779321,0.1875929832,0.0448188484,0.4433203042,-0.041380953,0.7773233652,-0.1926977783,-0.2444990724,0.789844811,0.1316350251,0.2115896344,0.1151454747,0.1390173584,-0.1353597343,-0.0279302131,0.0455650203,-0.1468821168,-0.0766203925,0.1834488213,-0.0149411904,-0.118060872,0.0647344217,-0.0425785705,-0.3984566629,-0.3288905025,0.1766891778,-0.0940921307,0.3057628274,-0.2974820733,-0.021245772,-0.0664557368,0.0871310458,-0.1968187988,0.0041389382,0.227421537,0.1077059284,-0.4807265699,-0.4307637513,-0.5369927287,0.2102299929,0.0410189629,0.7110475898,0.1348099858,0.1258124411,-0.0493269786,-0.0679451004,0.7193492651,0.0377152525,-0.2552365363,0.1822211295,0.3052651882,-0.220570311,-0.0752837509,0.1674725115,0.2276556343,0.7511644363,0.0254301522,-0.277977556,-0.017720161,0.2570811212,0.2692537904,-0.1374401152,-0.264275521,-0.1753081977,-0.1455651522,-0.0927020758,-0.2050192654,0.0776114538,0.4493768513,-0.1585988998,-0.1447980553,0.0166420862,-0.4110098779,0.0234444048,0.0332664698,0.2265168428,-0.0339936018,0.1226979122,-0.0917227939,0.3180520833,0.5115271211,0.1914930195,0.0366322175,-0.3783223629,-0.0597475953,-0.3364688158,0.3730210662,-0.0945392475,-0.2697618604,0.0890244991,-0.1225717515,0.3454491198,0.0978458896,-0.3196589649,0.3448731899,0.1271499097,-0.8693082333,0.040328864,0.3497123718,0.0754916295,0.2559961081,0.1632469893,-0.291364342,-0.4001287222,0.1954010725,-0.2215898931,0.5349753499,0.400005579,0.1256385744,0.078060776,-0.4380124211,0.626794517,0.1097606868,0.0620516241,-0.2052246481,0.3821625113,0.0868347362,-0.145893693,0.1996168643,0.0108880233,-0.0975732654,0.0620532669,-0.5207388997,-0.2721308172,-0.2103891224,0.0633196533,0.2039902955,-0.2107553631,-0.570271492,-0.0021740061,-0.060787946,-0.2533459663,-0.0670553073,-0.2045769244,0.0590337999,-0.1251281351,-0.3649452925,0.0907058269,0.1318021119,0.2061830014,0.3368248045,-0.1952323616,0.1447541416,-0.1788713485,0.295581311,0.1077705696,0.0045018042,-0.0674382746,-0.1130898818,-0.1078403145,0.0571676232,0.0388853215,0.1876392365,-0.0155582475,-0.1259191185,0.1938151568,0.0704370812,-0.1935295165,-0.096293956,0.2029899508,-0.2946879864,-0.329508394,-0.1623876095,0.032916002,0.2324554473,-0.1372097433,0.0390672572,0.1924266517,0.0280409027,0.3296630085,-0.3757055402,-0.161416769,-0.3390892148,0.1104764715,0.0222626142,-0.1946795732,0.2751014829,0.5374060273,-0.2416733652,-0.0429454409,0.4625181556,0.3201998472,-0.4603796899,0.5341430306,-0.0770808235,0.2449527532,-0.120336391,0.4825563431,0.077265121,0.0012956185,0.121118404,-0.3984121978,-0.2224571258,0.0035766128,0.1803524792,0.0048175994,0.0767067,0.5469903946,0.0969509184,-0.2241243273,-0.1405323893,-0.0130537627,-0.2337539792,0.0333456285,-0.0965530947,-0.0322135612,0.0356991403,-0.0731009543,0.0366536193,0.1145421937,0.0164362732,-0.1236490011,-0.3727941215,0.1271935105,0.1557142138,-0.2110016197,0.1434066296,-0.0033734473,-0.2501889169,0.0102502154,0.1792968065,0.0619124062,-0.0944190323,-0.0458827205,0.1153456196,-0.161063686,-0.0681879148,0.1086579189,0.2025467455,0.5488921404,0.0119073242,0.4053371251,-0.0716577917,-0.0732147992,-0.4471322,0.0277351066,-0.203941375,0.3001711369,0.1436178088,-0.7271795273,-0.0998679847,0.3671343029,0.1616151333,0.2915070653,-0.2427118868,0.077221334,0.0469421595,0.1624555588,-0.0157322418,-0.1802331805,-0.0087852543,0.0827633068,-0.1923856288,0.0270687398,-0.0308604501,-0.4143474698,-0.1752532125,-0.0408901758,0.3402948976,0.1533829421,0.2932591736,0.3838325143,0.1860468686,0.161279127,-0.083013162,0.3084394634,0.1719073951,0.0413558856,0.2358126938,-0.1518711895,0.5499083996,-0.0797493383,-0.2109907866,-0.4842012823,-0.1013555154,0.0385552123,-0.2127789706,0.2878963351,0.178676337,0.141226247,-0.2053218931,-0.0585112013,-0.1316242665,0.1469825208,-0.1549752504,-0.0191578809,-0.2165710479,0.0453941673,0.2118901163,0.0991670266,0.0438985489,0.2256003767,-0.2876901925,0.2112767845,-0.021983251,-0.4377116561,0.1144837737,-0.1112745553,0.2158975899,-0.2742503583,0.0428081751,0.1431733072,-0.0642303899,-0.0831959993,0.1545310467,0.5885646343,0.269770205,0.2648819685,-0.1208004579,-0.1706359684,0.107335262,0.174276486,0.1686101556,0.2314853668,-0.0136458408,0.1752024144,0.039228268,-0.0672091469,0.3838609159,0.1897573918,0.1100059003,0.0028359594,0.1769027412,-0.1984802634,-0.1982295811,-0.4200226963,-0.1710662693,-0.2558765709,-0.0049822605,0.3394448161,-0.1537775993,-0.1319400668,-0.0998607501,0.0291871186,-0.2728424668,0.410010457,0.3090417087,0.3422877789,-0.0362546705,-0.1375628561,0.1079030931,-0.1108530462,-0.2709168494,0.384044379,0.3714982271,0.2755203247,-0.2814259827,-0.0565951206,0.2465707064,0.6157265306,-0.5056514144,0.0120019978,-0.3906681538,0.1191625446,0.3668136895,0.0803157017,-0.1651151329,-0.122963056,0.4900046587,-0.1206821129,-0.0789343938,-0.1198878661,0.0980636403,-0.109252736,0.0767080784,0.0152672334,0.407881707,0.0250436682,-0.0295509025,-0.208316654,-0.2941704988,-0.192832768,0.0614787675,0.3428522348,0.2553367913,0.5357603431,-0.1673200876,-0.0385081209,-0.1830641925,0.2234833539,-0.03288991,-0.3865860701,-0.1538767815,0.0641760454,-0.1441627741,0.0128575647,0.2859579623,0.2245039791,-0.0872955099,0.2617087364,0.0529523976,-0.2685489357,0.0616977736,-0.4994950891,-0.3638346195,-0.0477742851,0.2398652881,0.2502560019,-0.0480287299,-0.046989128,0.1621945053,0.3588542938,-0.1503745764,-0.3527107537,0.1869351864,-0.1928920895,-0.0597137846,0.178605333,0.2957515419,0.2869983315,-0.4042109251,0.2542159259,-0.124879986]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"@TevenLeScao I have just checked it: this was a bug in `pandas` and it was fixed in version 1.2: https:\/\/github.com\/pandas-dev\/pandas\/pull\/36898","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":20,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\n@TevenLeScao I have just checked it: this was a bug in `pandas` and it was fixed in version 1.2: https:\/\/github.com\/pandas-dev\/pandas\/pull\/36898","embeddings":[-0.3412717879,0.124030225,-0.0055019557,0.2580174208,0.1083403826,0.0757042468,0.3035877347,0.4069502652,0.0264133904,-0.0411831364,0.2745392621,0.0067486395,0.1387162507,0.1741790026,-0.2055495232,-0.2439861596,0.2017897964,-0.0212467574,0.0084578265,0.2275885195,-0.310215354,0.2221112996,-0.1696818024,-0.1478930116,-0.4680121839,-0.3230388761,0.0654656067,-0.0415014438,0.0460146032,-0.4286079705,-0.0355177671,-0.2563436925,0.4413874149,0.5364952683,-0.000119037,-0.2155592889,0.3319274187,0.0182693955,-0.2052394748,0.0686349273,-0.472040087,-0.3954090178,0.0219647828,-0.343072027,0.4541318715,-0.1743062586,-0.3328178525,-0.4177092016,0.6475176811,0.3707948625,0.1486110091,-0.032841567,0.1063491181,-0.0371334702,-0.2036338001,0.0837967247,0.0200115144,0.1159936935,-0.2379019111,0.2098362297,0.2300752401,-0.0716975778,-0.1255659312,-0.2341950536,-0.1767509431,-0.065696083,-0.0222574696,-0.2992710471,0.1961596906,0.4449341595,0.2029707432,-0.1828810722,-0.2681335509,-0.4125076234,0.1113054082,-0.5898334384,0.0529373623,0.4741686285,0.0347095132,0.0870471448,-0.112177439,-0.0808055401,-0.2130061239,-0.0569745228,-0.0503388122,-0.237240538,-0.1518814117,-0.0181154292,-0.203660354,-0.1465211213,0.0027886073,-0.2815826535,-0.4799128473,0.0245663878,-0.19687967,0.0740092546,0.1518745571,-0.5287777185,0.0522209406,-0.0778356344,0.3272016048,-0.3820248842,0.1291124523,0.0760853887,0.5744822621,0.2120935023,0.1829941869,-0.0716198087,0.300283879,0.0313625075,0.0964887291,-0.0893944502,0.0688516125,-0.2353248,0.0683994666,0.0023821266,0.0328889601,-0.0524887182,-0.0941911712,0.2288012058,-0.3604729474,-0.1577097327,-0.2764321566,-0.1858027726,-0.1272218376,0.3287419975,0.1308708787,0.2427649945,-0.0845068991,0.0472011901,-0.0698759109,-0.0375033394,-0.0055307657,-0.0535055138,-0.0542489439,-0.1753414273,0.0107756685,0.2415701598,-0.0861282945,-0.2899804115,-0.1535128206,-0.1127546132,-0.1566496044,0.2403205782,0.3214479983,0.1444984972,-0.1472991407,-0.1794864982,-0.0155875254,0.1483362466,-0.4130090773,-0.0150122447,-0.3744748235,0.1352471709,-0.1423554868,0.148070246,-0.5019357204,-0.031469591,0.077753365,0.1964427382,0.4208015501,-0.2530838549,0.6042643785,-0.2848632336,0.0148763675,0.281612277,-0.7312355638,0.1030590534,0.147030279,0.0361658372,0.4897561371,0.417034179,-0.024581166,0.3368774056,-0.1128997579,0.2621537149,0.0711583421,0.2094028443,0.0982469022,0.5672950745,-0.0708639771,0.0182830803,0.0297987387,-0.3001289666,0.5715810657,0.1137629151,0.4756506979,0.4330525994,0.0384260304,0.2293846905,-0.1646941304,-0.1979120076,0.0348151326,-0.2483368069,-0.2546766698,0.0026689982,0.1543461829,-0.192816779,0.2644361854,-0.2410074472,0.1251212358,0.4268120825,0.132885322,0.141797781,0.0218667574,0.2489672601,-0.3805499077,-0.1901163459,-0.1566674411,-0.1665305197,-0.1608521789,-0.0739144832,0.1734650731,0.1861049533,0.0388094559,0.1554861814,0.0502698794,-0.1581521928,-0.1364510655,0.2717756927,-0.0530617386,0.3020300865,0.136513263,0.05702338,-0.2673314512,0.2672393322,0.2391949296,-0.1706701517,0.3222697973,0.1106031612,0.1830595434,-0.2960664928,-0.1514948606,0.2638213038,-0.0377146713,0.0216184016,-0.1175799146,0.0669658855,0.1620994657,0.194870621,-0.227302596,0.2584483027,-0.0121855205,0.0613592342,-0.1402769387,0.6026515961,-0.2641135752,0.2733408213,-0.0368501619,0.0274138823,0.3373424709,-0.1023813337,-0.1084210873,-0.1580342501,0.2359513342,0.1434629112,-0.1225715801,0.054006055,-0.4003921151,-0.030364627,0.5197226405,0.0007626491,0.117055811,0.1844302863,0.0339210033,-0.0154382503,0.0759516507,0.1254651845,0.293073982,0.1264980435,-0.039446827,0.0525660887,0.115116857,-0.0862425044,0.2848719954,0.1711804271,0.2693832517,0.248569265,0.1466228515,0.197782129,-0.1394346803,-0.3373427093,-0.1755522192,0.1275863945,-0.3030416071,0.3055096567,-0.1936163455,-0.1043043733,-0.3914400339,-0.4014569223,-0.2060244679,-0.2955180705,-0.2415339798,-0.0955116227,-0.4658668339,-0.0853125155,-0.1617810428,0.0963584259,0.1571011841,-0.0004351769,0.0192784946,0.0009038327,-0.0431024805,-0.0549469851,-0.0049517052,-0.1238726228,0.0270979665,0.1601220816,0.1456878185,-0.0970128104,0.1815309674,0.0411112234,-0.0188201144,-0.0832133666,0.1475452334,0.2630810738,-0.0727505088,-0.5481500626,-0.0182919092,0.1664054245,0.0384018235,0.1412490606,-0.1206467077,-0.1862722933,-0.0932749659,-0.0603482835,0.0194407683,-0.3349897861,0.4810393155,-0.105528228,0.02690145,-0.086422205,0.1476101577,-0.0642029271,0.0781240016,-0.2673318684,0.1539460123,-0.4169496894,0.0305447057,-0.2566615641,-0.2721676528,0.1277552396,0.0963209942,0.0096773701,-0.078258507,-0.3716353476,0.5211656094,-0.1465875953,0.3120235503,-0.1196961477,0.1910612881,0.0726432726,-0.0763165951,-0.0702414066,0.0268293731,0.1516522765,0.1500844061,0.0634180829,0.4252824485,-0.0681784153,0.7963733673,-0.2070921212,-0.2428016812,0.7851786017,0.1798446327,0.2147368342,0.1114979833,0.1771980226,-0.1324888319,-0.0516305789,0.0562985651,-0.1535121948,-0.0847821683,0.138390258,-0.0334206969,-0.0991474763,0.0346302204,-0.0383115001,-0.3957054317,-0.337207824,0.1653221548,-0.1299589425,0.2785852551,-0.279636234,-0.0569239035,-0.0546337701,0.1119031981,-0.2268248349,-0.0253560226,0.2500804961,0.0908861309,-0.4192575216,-0.44160074,-0.5331693888,0.1681439131,0.0267068036,0.694568634,0.1523023546,0.1240498871,-0.0636752397,-0.0554272681,0.6983473301,0.0264313705,-0.2924920619,0.1699612588,0.2758511007,-0.2267137915,-0.0662788972,0.1623209417,0.187394172,0.7814658284,-0.0040756771,-0.2857832909,-0.0204492901,0.2629454434,0.2712848783,-0.1538704932,-0.2529331744,-0.1603029668,-0.1997375488,-0.0814147219,-0.1784598976,0.1562794894,0.4531793892,-0.2090243697,-0.159223631,-0.0071684835,-0.4004323184,0.0399247706,0.0399925038,0.1950909346,-0.0108049065,0.150407806,-0.1228339225,0.2694620192,0.4711761177,0.2290257663,0.0092923949,-0.3669433296,-0.0709777474,-0.3196597695,0.380223453,-0.0930656269,-0.2809783816,0.0854495689,-0.1541886479,0.3633084893,0.1215594113,-0.2874757051,0.3155450821,0.1552750468,-0.832490921,0.0266837832,0.3657891154,0.1117782444,0.2571146488,0.1980552077,-0.2689127028,-0.3762990236,0.1702100039,-0.2446850985,0.4795739055,0.3824549317,0.1580195576,0.1130345091,-0.4080251753,0.676199317,0.0891284198,0.0456143841,-0.1887376755,0.3455237448,0.0942925066,-0.1524701267,0.171870932,0.005232058,-0.0884762779,0.0735983029,-0.5113877058,-0.2449371219,-0.193139568,0.0475888774,0.2021462321,-0.2321405113,-0.6011215448,-0.0003101752,-0.0318100415,-0.2616799772,-0.0451884866,-0.1800916791,0.053606119,-0.0957735032,-0.361068666,0.0835338309,0.147834301,0.1861364692,0.3207148015,-0.1669660807,0.1520547569,-0.1513990909,0.3327769637,0.1178430915,0.0078483447,-0.0754910931,-0.0993498117,-0.0884820223,0.0514725745,0.0506722704,0.2107183784,0.0061887163,-0.0954473093,0.253842324,0.0702206865,-0.1800710559,-0.0787266493,0.2651982605,-0.3444180191,-0.3410686553,-0.1577455699,0.0235497728,0.2334502339,-0.1699759066,0.0516469106,0.2095892131,0.014996592,0.3483524024,-0.3776893914,-0.1721515357,-0.3480566442,0.1002505645,0.0455047749,-0.1881159395,0.244630456,0.4835463166,-0.2497799248,-0.0274277516,0.4238178134,0.3446509838,-0.4265529811,0.5984869599,-0.1005206779,0.2218951583,-0.1415479034,0.4189109802,0.0658269227,0.0264702234,0.1126519144,-0.3906410038,-0.2438426614,-0.0110298712,0.198543936,0.0316888951,0.0800386369,0.5649981499,0.0825437829,-0.1961648315,-0.1431964487,-0.0040077032,-0.2632652521,0.0200008973,-0.0976494178,0.0054125255,0.0304383151,-0.0590460375,0.0467738919,0.1368298978,0.0306494776,-0.1308159232,-0.4030769467,0.1121966243,0.1812619269,-0.2131501138,0.1576090604,-0.0253619235,-0.2624441385,-0.0120829586,0.1790465415,0.0902934819,-0.1016900316,-0.0464464873,0.1531558782,-0.1598239541,-0.0562741682,0.1164110899,0.2109647393,0.5365225077,0.0306342151,0.3913279474,-0.0880648866,-0.0555301793,-0.4043053389,-0.0179543439,-0.2343648076,0.2840470374,0.1519258022,-0.6966723204,-0.1060282588,0.3511353433,0.1717192233,0.2847226858,-0.2599787712,0.1179752797,0.0656609684,0.158409521,-0.0572258979,-0.1518693119,-0.0016454891,0.1128946021,-0.1379239857,0.0401751176,-0.0058040484,-0.4350820184,-0.1529610902,-0.0542751811,0.3313514292,0.1398850083,0.2754512429,0.3992125392,0.1721582562,0.176974535,-0.0564028434,0.2889548242,0.1773953289,0.0684148595,0.2014059722,-0.1621271521,0.4717501402,-0.0953748301,-0.2096824497,-0.4819872379,-0.0746587366,0.0079442738,-0.2684753835,0.2657918632,0.1635710448,0.1167763993,-0.2271022797,-0.0101147518,-0.0770572796,0.1764437854,-0.1731632054,0.0039130962,-0.1582723558,0.0181202181,0.2310760766,0.1033621132,0.0914603248,0.1970944107,-0.2558514178,0.1926349252,-0.015990736,-0.3948794901,0.1356339157,-0.1107458025,0.2268871218,-0.272575736,-0.0000765397,0.1675511152,-0.0984948874,-0.1251696199,0.1620348096,0.5886413455,0.2581673861,0.2816857696,-0.1151436418,-0.1393072307,0.0920643359,0.1447825581,0.1526227444,0.230088532,-0.0126196053,0.1981311291,0.0573749691,-0.0856673345,0.3561891615,0.1813800037,0.1266648918,-0.0304752905,0.2105373144,-0.1760418266,-0.1969998479,-0.4345936477,-0.1564471126,-0.2098419815,-0.0391119011,0.3237751126,-0.1878923476,-0.1111786515,-0.1099402383,0.0362065919,-0.2735576928,0.3935525119,0.3105427027,0.3471532762,-0.0577560924,-0.1337611526,0.0504735559,-0.0989685804,-0.3138723671,0.3750606477,0.3629861176,0.2740958333,-0.3101061285,-0.0636712611,0.2408324629,0.6288326979,-0.4764030576,-0.0152283432,-0.3945148289,0.0170960817,0.2993396521,0.0733305365,-0.1631336808,-0.1150184646,0.4784143865,-0.091317609,-0.0644214675,-0.10116072,0.1213784665,-0.086934112,0.117780894,-0.0237041768,0.42740044,0.0045204889,-0.0416510664,-0.1996573061,-0.3047764301,-0.1673973054,0.0448243432,0.2976605296,0.2640576363,0.5109522939,-0.146737352,-0.1196804792,-0.203610152,0.2333160043,0.000966472,-0.3993254006,-0.1593145728,0.0673374906,-0.125060454,0.0530025028,0.2979461253,0.2603336275,-0.1092884243,0.2785044909,0.081472449,-0.2846652865,0.0990646034,-0.4734436572,-0.3631598055,-0.0700144023,0.230744943,0.2449279279,-0.0357334092,-0.0005882207,0.1598483622,0.339705348,-0.1527702659,-0.3735357225,0.1693944037,-0.2099458873,-0.037030559,0.1827775687,0.2476982772,0.3111520112,-0.3687021732,0.2597649992,-0.1153498739]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"Thanks ! I'm creating a PR","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":6,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\nThanks ! I'm creating a PR","embeddings":[-0.427283287,0.0908893943,-0.0069078896,0.30450055,0.1101281121,0.0595720895,0.1178438216,0.3416466713,-0.049976889,-0.0005295728,0.4177300632,-0.0051334971,0.1104985625,0.1496742368,-0.1306799501,-0.2671569884,0.2184585333,-0.0443501845,0.1490712613,0.2499383241,-0.2511847317,0.2489295453,-0.1962835491,-0.144418478,-0.3765289187,-0.3065877855,0.144577831,-0.0244777109,0.0046045659,-0.4617555141,-0.1050651893,-0.2520403862,0.5239130855,0.439138025,-0.0001215967,-0.1864488125,0.3046823442,-0.0167128034,-0.162470594,0.0709542558,-0.4333614409,-0.3496669829,-0.0711162314,-0.3032565713,0.365745455,-0.1953448951,-0.3685683012,-0.3670468032,0.6142721176,0.4207582772,0.124270089,-0.0004392271,0.1086045727,-0.0232647285,-0.0917617157,0.0890177786,-0.0030691079,0.0519080758,-0.1633091122,0.2870160043,0.2951007485,-0.1193699241,-0.1262741983,-0.275118649,-0.1268461347,-0.0821361989,-0.0387657173,-0.2255685329,0.2136830837,0.3227701485,0.1980535239,-0.2041624188,-0.1467292458,-0.3374611437,0.1020307541,-0.6630982757,0.0770436972,0.5751552582,-0.0269491207,0.1183803752,-0.2591252923,-0.0987291858,-0.2543337643,-0.0483053289,0.0238760039,-0.2509928942,-0.0715064108,0.0457668267,-0.1135317832,-0.1182018816,-0.0280984752,-0.2536236048,-0.5682675242,-0.0637095124,-0.125085026,0.0653623492,0.1226912513,-0.3651655018,0.1310784519,0.0152792037,0.3004535139,-0.3975632489,0.2619207799,0.0524008721,0.6028311253,0.1507598758,0.2432179302,0.0218110923,0.2518854141,-0.0080594495,0.0678328201,-0.0875031203,0.0362690315,-0.1566821635,0.1038336903,0.021319801,-0.0513790958,-0.0989597291,-0.0682226494,0.2551515996,-0.4084287584,-0.1725775003,-0.264518559,-0.1474957466,-0.094859533,0.2974821925,0.176979214,0.2306609154,-0.1891366392,-0.0595323667,-0.0104561727,-0.0940298066,-0.1112570092,-0.0344765112,-0.0854777992,-0.0776571557,-0.0170747675,0.2232716531,-0.1031814665,-0.2822414935,-0.091796279,-0.0763899535,-0.1437824667,0.2213103324,0.3065117598,0.0524686873,-0.2240124643,-0.2179869562,-0.042719014,0.2367165238,-0.3958630264,-0.0472608805,-0.3306819201,0.1375654191,-0.1679535508,0.1552869529,-0.4921531081,-0.0402255692,0.0926095322,0.1175833717,0.3760268092,-0.3457896113,0.6102827787,-0.2446608245,0.0306654405,0.2735808492,-0.725980103,0.0610055178,0.2309175879,-0.0679816529,0.5091286302,0.4141031504,-0.0351111703,0.3821699619,-0.0948366523,0.2147140205,0.0573061742,0.2075272501,0.1105481237,0.4847134054,-0.1065026373,-0.0571683832,0.0756175146,-0.3117738068,0.6279677749,0.10456983,0.5055903196,0.400937289,-0.0256258603,0.2636191547,-0.1839706153,-0.1364052445,-0.0567149669,-0.1880374551,-0.3812234998,0.0126026673,0.1430567801,-0.1360194832,0.2870816886,-0.2629028857,0.1157043353,0.4732118845,0.1283840835,0.1588178575,-0.0091828723,0.2756644487,-0.3671340942,-0.2344531715,-0.1369848847,-0.1803825498,-0.0922297016,-0.0359347649,0.198836714,0.2628721893,-0.0306091867,0.1835809499,0.0102931997,-0.1046282127,-0.1097865775,0.1482830197,-0.0142393271,0.228450045,0.1334215701,0.0408790819,-0.2737740576,0.2976744771,0.2044093311,-0.2058689594,0.247402668,0.1623055935,0.1684133708,-0.3126221597,-0.1032860428,0.2229085118,0.0407146439,-0.0734823272,-0.0602834411,0.1231004223,0.1650483161,0.0742956251,-0.1882502437,0.2701932788,-0.052758202,0.0822494999,-0.189459011,0.5972123146,-0.2065780014,0.3039011359,-0.0139302928,-0.0242781639,0.2527696788,-0.1129166707,-0.0237048753,-0.121815227,0.3349547088,0.2374741733,-0.1413132995,0.0228012986,-0.4062594175,0.0101278014,0.5158886313,-0.0330354907,0.0099596549,0.2006329149,0.0035069301,-0.0192346182,0.0773344189,0.0940996483,0.3149075806,0.1565819532,-0.0849291384,0.113056235,0.2031299323,-0.075126186,0.2770595849,0.1314623058,0.2067047358,0.2823491096,0.2304030806,0.1412636489,-0.1400038153,-0.373051703,-0.1978419721,0.1226895973,-0.2970970869,0.3091567755,-0.1641323864,-0.0292561669,-0.4944787025,-0.2956286669,-0.1970687509,-0.2395397425,-0.1868915111,-0.141299665,-0.4715084434,-0.1317484826,-0.0918573067,0.1092467308,0.0924809799,-0.0815825015,0.069021605,-0.0241323933,-0.132557705,-0.1031318679,-0.0300067291,-0.1781229973,0.12546359,0.1220029294,0.103094548,-0.1517464668,0.1403141022,0.1539807469,0.0214326065,-0.086996302,0.131068781,0.2358669788,0.0361981541,-0.5285502672,0.0219422448,0.1750990152,-0.0030436853,0.1526522636,-0.1193801835,-0.2397427857,-0.0919130445,-0.1226349026,0.0074633863,-0.308401823,0.4813194573,-0.040332444,0.0198099893,-0.1608387232,0.1676129699,-0.1592339575,0.0008740981,-0.3384691775,0.2137446404,-0.3577175736,-0.0818505809,-0.2669862509,-0.2489576787,0.1325875372,0.1879355162,-0.1356853843,0.0241384618,-0.4038697183,0.5267026424,-0.0868500769,0.4759992361,-0.1299003661,0.1668464243,0.126900807,-0.0425463095,-0.0277582891,-0.0083943978,0.0103995912,0.1796653867,0.0504578836,0.5497773886,-0.0832432136,0.6674972773,-0.2948662639,-0.1783452779,0.7951195836,0.104775548,0.1767286509,0.050688304,0.0906746164,-0.1347900629,0.0557525977,0.157542035,-0.1172773838,-0.1603123844,0.1570060104,0.0610486604,-0.1472172737,0.0336134657,0.038928844,-0.3868366778,-0.3201098144,0.2016365677,0.0263137016,0.3630300164,-0.3274406195,-0.0321887657,-0.0571922809,0.0790940896,-0.2537908852,-0.0623212345,0.2343730927,0.111774385,-0.5828482509,-0.4280754328,-0.4881945252,0.213320002,0.0272165295,0.5919046998,0.0863692462,0.1157797799,-0.067902334,-0.1450097859,0.6606110334,-0.0358412974,-0.2167564631,0.1980374455,0.3949572742,-0.0860708356,-0.1050386429,0.1352813095,0.223456338,0.8426058888,0.0086417645,-0.3551011682,-0.0079893693,0.2566036284,0.3563587666,-0.1371339411,-0.2523956597,-0.1420277059,-0.2354355156,-0.0811997652,-0.1665876359,0.0558767281,0.4358101785,-0.1945470721,-0.1270201355,0.1366364509,-0.2879390419,-0.0216039121,0.0249188188,0.296669215,-0.1586866677,0.1415983886,-0.0902375653,0.3110305071,0.558138907,0.2475836873,0.0563273393,-0.4013028145,-0.0363358222,-0.2941616774,0.4015272856,-0.1261665225,-0.2133723199,0.0886985213,-0.1410927922,0.4426511824,0.1177091375,-0.3402509689,0.2948023081,0.0843968019,-0.8257698417,0.037120942,0.2916265428,0.1173250377,0.267154634,0.050575085,-0.3116075099,-0.3423392773,0.2135063112,-0.2104618102,0.5550667644,0.4145555198,0.1030885577,0.0470154174,-0.4581383467,0.6308457255,-0.1110874563,0.0349841379,-0.2238029391,0.4713591337,0.1351163834,-0.1899787784,0.1191589385,0.0549397469,-0.1464990824,0.0521863066,-0.525963068,-0.3165816665,-0.2048987746,0.0225507189,0.2469569594,-0.1694769114,-0.5396955013,-0.0229151435,-0.0787071437,-0.273352772,-0.09533941,-0.2132762074,0.0909573361,-0.1448233277,-0.4521335661,0.1374640912,0.0375296995,0.0676475316,0.4206795394,-0.175599739,0.143498674,-0.1026506275,0.2849647403,0.1069429368,0.011192889,0.0182871427,-0.0701853856,-0.0975667164,0.0645761341,0.0151115628,0.0905995741,-0.062747255,-0.1389848739,0.2037156224,0.0266653001,-0.2290484309,-0.1625358611,0.2488311976,-0.2518967688,-0.2954190373,-0.0988538712,0.1285401434,0.2023301125,-0.1375951171,0.017950872,0.1637119055,0.0261464119,0.2948429883,-0.2954943776,-0.1545986086,-0.3183688819,0.036459513,0.0204192065,-0.287871182,0.1689448655,0.537106514,-0.1923359036,-0.0832068399,0.4404427707,0.2990499139,-0.4871707261,0.5459553003,-0.0343533903,0.2318884432,-0.0571891665,0.4806595445,0.0345646665,-0.0434277579,0.1830724776,-0.3553012013,-0.1835713089,-0.0507189855,0.1727003008,-0.0363927595,0.0844593272,0.5871415734,0.0436046645,-0.2138862312,-0.1269806772,-0.0417361893,-0.2694642246,0.0534997396,-0.105471231,-0.0149582857,0.0030812446,-0.0793847963,0.0259404983,0.1127712652,-0.0498055369,-0.0886077881,-0.33116889,0.1628359258,0.1390307397,-0.2030043602,0.1056328267,0.0117361145,-0.1952256411,-0.0111195864,0.2391998172,0.0046146559,-0.1073835343,-0.1318762004,0.1107558906,-0.1051993817,-0.0743600726,0.068584688,0.3006308675,0.6597557068,0.0459951237,0.4207652509,-0.1151629537,-0.1089591607,-0.4420247972,0.0442201719,-0.084157072,0.3956668675,0.1178523451,-0.7017648816,-0.0849473625,0.2941749692,0.1294915527,0.3398170471,-0.2682321966,0.1507746279,0.0203293674,0.1610184461,0.0231103469,-0.1701154113,0.0309860762,0.1126019657,-0.212599203,0.0292618386,0.0093278009,-0.4023272097,-0.2213643789,-0.0760213956,0.3698728383,0.1324811578,0.1944633871,0.3320173323,0.1502299458,0.1931794435,-0.1298972517,0.2663260996,0.1707035005,0.0373050645,0.2301277667,-0.1039807871,0.6213839054,-0.0847157463,-0.198186487,-0.5445185304,-0.0584734045,0.0217999332,-0.0810086504,0.2741642594,0.1828726232,0.1106930003,-0.1215432733,-0.0454161949,-0.1370797604,0.140175879,-0.1145510525,0.0165146533,-0.2394677401,0.0034986644,0.1691620797,0.1881863326,0.0532917827,0.2128890455,-0.206826821,0.2204610109,-0.0251530614,-0.4943318069,0.1854570061,-0.1371632963,0.1844324619,-0.2892076075,0.1356842965,0.1428992301,-0.0849340856,-0.0726473555,0.0813533813,0.6168562174,0.2800317705,0.2366171777,-0.1472115368,-0.2172355652,0.1094167009,0.0892832428,0.143808946,0.2417254895,0.0054477155,0.1803780645,0.0178237129,-0.0427641869,0.3899846375,0.1793584228,0.0640681162,0.0701846257,0.139308244,-0.1160067245,-0.2258557826,-0.4446611106,-0.1999170035,-0.1997150928,0.0382817835,0.3431219757,-0.1018750891,-0.1019686311,-0.0628872961,0.0185563304,-0.2601873577,0.4027629793,0.283429563,0.3490065634,-0.1201572865,-0.2020555586,0.1534915417,-0.0669405684,-0.3531534672,0.3860530257,0.3397310674,0.2968025208,-0.2514864802,0.0368218645,0.1771852225,0.6627504826,-0.4531202912,0.0812006965,-0.4297362566,0.0301161427,0.4862041175,0.1221491173,-0.1450305134,-0.1627401859,0.4524957538,-0.0345538631,-0.0918857753,-0.1662692279,0.1410098672,-0.0515530184,0.0872864276,0.0129746664,0.4337684214,0.1424631029,-0.0556715429,-0.2459709942,-0.3043395579,-0.1997292936,0.0139552895,0.2371140271,0.2515494227,0.4874588549,-0.11129684,-0.0255576428,-0.2164155096,0.1304179728,0.0198761858,-0.359719336,-0.111334689,-0.0011297093,-0.1010388136,0.0537154935,0.2118573338,0.2015973777,-0.0782555789,0.1916049868,0.0285159927,-0.1946606189,0.1129841357,-0.3943724632,-0.3121351898,0.0147479614,0.2617555559,0.2039490193,-0.0784603357,-0.0825137943,0.1651633978,0.3715538979,-0.1971422285,-0.4485737383,0.214173913,-0.1954564303,-0.1085790247,0.2272661626,0.3812497258,0.280511409,-0.4441871047,0.2663842142,-0.0806656256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"Well I though it was me who has taken on this issue... \ud83d\ude05 ","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":13,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\nWell I though it was me who has taken on this issue... \ud83d\ude05 ","embeddings":[-0.381496042,0.1416245848,-0.0314113833,0.2693937421,0.0503840558,0.0776250809,0.1727975458,0.3613765836,-0.0021893056,-0.0321630426,0.3624667823,0.0022020794,0.0744440034,0.1662325263,-0.1462906897,-0.1717242002,0.1870137751,-0.0604586788,0.0400327705,0.2251366824,-0.2347595841,0.2736390233,-0.1921285242,-0.198765561,-0.4124678075,-0.2566157877,0.1424660534,-0.0318162926,0.0023674974,-0.4859756529,-0.0713481829,-0.2452652752,0.4516730309,0.492726922,-0.0001187635,-0.1919668913,0.3233660758,-0.0044267601,-0.2326923907,0.0693785772,-0.4573564231,-0.374694407,-0.0160190295,-0.3471026719,0.4194579124,-0.107924521,-0.3447861075,-0.4215415716,0.6427291632,0.3935449719,0.1514766961,-0.0364488438,0.1382516325,-0.0078407601,-0.08686129,0.1159153357,0.0030568542,0.1058249325,-0.1556789726,0.2756854296,0.265769273,-0.0741082877,-0.1509165913,-0.3164863288,-0.1351798326,-0.1053164452,-0.046082437,-0.2601768672,0.201141566,0.3636045158,0.1958273798,-0.2042473704,-0.1671886742,-0.3338914812,0.1088537499,-0.6169026494,0.0381614119,0.550086081,-0.0224332549,0.1182596609,-0.2470849305,-0.1029288322,-0.2466813028,-0.0554091372,0.0143728433,-0.2161848247,-0.1115105748,0.0113251405,-0.1478096992,-0.1085581779,-0.0698315278,-0.2741244435,-0.5732527971,-0.0036516641,-0.1840486825,0.0570930131,0.1485115737,-0.420024544,0.1287482083,-0.0586175844,0.3299888968,-0.3596474826,0.243800953,0.0650487617,0.5489268303,0.1661251187,0.2953171134,-0.0814832151,0.2668429911,-0.0045106737,0.085246399,-0.0830559209,0.0259351712,-0.1936476231,0.0822770298,0.0818149298,-0.0358221158,-0.0989273638,-0.0552788898,0.2111806273,-0.3862944841,-0.1723247617,-0.2443553358,-0.1923894882,-0.0681470707,0.2308228761,0.1729385555,0.2587677836,-0.1568240225,-0.0641464889,-0.0402877741,-0.1067453995,-0.0943291038,-0.0177833959,-0.0232520681,-0.1438556015,-0.02167424,0.2503770292,-0.0716631487,-0.2844986022,-0.1191941872,-0.0463824309,-0.167800054,0.1828901619,0.2796661258,0.125776276,-0.19006522,-0.1780466735,-0.0314285271,0.19382146,-0.4698239267,-0.0470830947,-0.3575198054,0.1623171866,-0.1738317162,0.1464631259,-0.5239098668,-0.0653836727,0.0814745724,0.1515371799,0.3685851097,-0.2787260115,0.6063600183,-0.2634146512,0.0284443181,0.2474117875,-0.7251703143,0.0762580186,0.1469188035,-0.0281563103,0.4560164809,0.3809763491,-0.0638424307,0.3769870996,-0.1198917702,0.2891185284,0.0731380805,0.2030196935,0.1071213186,0.5030428767,-0.0896408334,-0.0122007933,0.0196162891,-0.3181828856,0.6450618505,0.0994917229,0.5115265846,0.4066801965,0.0402603596,0.2400664538,-0.2018621415,-0.1862701178,-0.0397888012,-0.2100708932,-0.3642632961,-0.0687655136,0.1424001306,-0.2139342427,0.3179022372,-0.271410197,0.1137208492,0.4455432594,0.1467990875,0.1307491213,0.0128579224,0.2901844382,-0.4459474981,-0.1719225049,-0.1293558031,-0.1530877501,-0.1343258619,-0.0626034737,0.212406382,0.2259224504,-0.0493843816,0.1751536727,0.0632646158,-0.1617314667,-0.1421660036,0.2012657523,-0.0174510293,0.2371828854,0.1544125974,0.0398194753,-0.2718606591,0.2936110795,0.2278179228,-0.1895488799,0.2145272046,0.1234789491,0.1750037521,-0.3221664727,-0.1381591409,0.2212661654,0.0109057594,-0.031526044,-0.1282755435,0.1032779515,0.1303922385,0.0728683099,-0.1804379672,0.2812793851,-0.059811417,0.0603696182,-0.1155148596,0.5728577971,-0.2222756296,0.3352801204,-0.0508016534,-0.0303374305,0.2239435166,-0.1197344959,-0.0614091232,-0.1443758905,0.2826527059,0.1780817956,-0.1415268183,0.0040465742,-0.3117997646,0.016899921,0.5378286839,0.0039917328,0.0451456085,0.1923908144,-0.0154893063,-0.0029106075,0.0577978604,0.0889981985,0.2775100172,0.1422639042,-0.0565457381,0.0807956755,0.2270396948,-0.0858456865,0.2903964818,0.1424583197,0.2634243369,0.2909872234,0.2022095621,0.1500596702,-0.1419088989,-0.3790869713,-0.2034108192,0.0844350606,-0.3118015528,0.3062453866,-0.1873486936,-0.0310295168,-0.4690611959,-0.363414079,-0.2565049827,-0.3029571474,-0.1617710441,-0.126706779,-0.4747098386,-0.1561254561,-0.1374550313,0.1017379388,0.0911955759,-0.07906802,0.0467300676,-0.0261167102,-0.1052925959,-0.0826745331,0.0405164696,-0.1907892376,0.0815287009,0.1278831214,0.1246872172,-0.1129030585,0.1253710836,0.1077399179,0.0141091598,-0.0771577582,0.1244019717,0.2509830594,-0.0020655508,-0.5306687951,0.0121474853,0.1947695762,-0.0031994511,0.1690352857,-0.0980581492,-0.1771190315,-0.109101586,-0.0645484403,-0.0356175415,-0.3384291828,0.4346758127,-0.1208576486,0.057850685,-0.0520751439,0.1307850033,-0.1098903865,0.0245693531,-0.2970378399,0.2168162465,-0.3849625885,-0.0130544389,-0.2947540879,-0.2277774364,0.1377084553,0.1697107106,-0.0390621088,-0.0054156883,-0.3901124597,0.6086903214,-0.0899050981,0.3947812915,-0.1476069838,0.1565794647,0.100885354,-0.0641215965,-0.0395076871,-0.0243023131,0.1223501265,0.1637266427,0.0432380438,0.4790418744,-0.0844956711,0.668894887,-0.2137571573,-0.1913267076,0.7753190994,0.1544511914,0.1962610036,0.1045944393,0.0738165826,-0.1455271095,-0.0097335521,0.1107204556,-0.105469875,-0.1601418555,0.1542344838,0.0282873251,-0.1413059384,0.0293497834,-0.0208030287,-0.4031073153,-0.3671022654,0.2359723151,-0.0004721522,0.3159998953,-0.3290997148,-0.0641200915,-0.0348501913,0.0639600754,-0.2113574892,-0.0377853923,0.2221650779,0.1228924692,-0.5501188636,-0.4389676452,-0.4991746545,0.21505633,-0.0008071308,0.5915634632,0.131031841,0.1030094624,-0.0794986337,-0.1053973734,0.6302264929,0.0239202231,-0.2121130526,0.2287736386,0.3834627271,-0.1452225596,-0.1276203245,0.1549945325,0.2213310748,0.8257168531,0.0110287992,-0.3159392774,-0.0360068679,0.2465339452,0.3664432168,-0.1330951154,-0.2601721287,-0.1241481751,-0.1863616258,-0.0653586015,-0.2616892755,0.1070187688,0.4543697238,-0.1919416785,-0.1497575194,0.0925254524,-0.3061181307,-0.0014475702,-0.010422023,0.2638573349,-0.0990473703,0.1435551494,-0.0885409266,0.3110064566,0.4951170385,0.235554561,0.060125798,-0.3539728522,-0.060949821,-0.2848030627,0.3988520205,-0.1066705436,-0.2313575894,0.0943740085,-0.1369714588,0.3956717253,0.0957504436,-0.318449676,0.289847821,0.0829744861,-0.8660159707,0.0810194612,0.363620162,0.0653581023,0.2681982219,0.15500319,-0.2802218199,-0.4108090699,0.2055836916,-0.1993333697,0.538156867,0.4572116733,0.1454359293,0.0509019494,-0.4000753462,0.596706152,-0.0424737111,0.0509336255,-0.2073393464,0.4543745816,0.1409928799,-0.1519444436,0.1217508167,0.0671546683,-0.101749137,0.047395248,-0.5350471139,-0.2699283957,-0.2410467565,0.0119162658,0.2258010805,-0.2014576793,-0.4745540321,0.0109648053,-0.0751565099,-0.2356260866,-0.0779712126,-0.2328521758,0.080690667,-0.133133024,-0.3702541292,0.1250583827,0.0365435928,0.1280607581,0.3556050062,-0.187907055,0.177666977,-0.208127901,0.2655193806,0.1594788581,0.0271721575,-0.033692304,-0.0699795783,-0.0717179105,0.0707579777,0.0039085522,0.1120459735,-0.0564014465,-0.1310946196,0.1632482409,0.0637742653,-0.2078828812,-0.1117601246,0.2319571376,-0.2584581673,-0.3109644949,-0.1370985806,0.1437681466,0.2460392416,-0.1134817451,0.0439893678,0.1835634112,-0.011963414,0.341171205,-0.3455042243,-0.1718343049,-0.2913289368,0.0686649531,0.0165724754,-0.3111600876,0.2369669527,0.5205271244,-0.1870297194,-0.0723367184,0.4607591927,0.3218353689,-0.4903994203,0.5122543573,-0.0538670607,0.2809639275,-0.0790368393,0.4820776582,0.0249269307,-0.002214117,0.1711445898,-0.386749953,-0.1743482649,0.0250051264,0.1852960885,-0.0023715242,0.0493485183,0.5355398059,0.0500433184,-0.2267159075,-0.1599458605,-0.0037271052,-0.1724423468,0.0303908065,-0.0877424181,-0.0226497911,0.0177405216,-0.0613109544,0.0605694428,0.1073039323,-0.0304935127,-0.1304772198,-0.3482083082,0.1285364181,0.1247810721,-0.2067327946,0.1486331373,-0.0257723164,-0.210595876,0.0125834011,0.2161144167,0.0123369079,-0.0984192193,-0.0703566968,0.0929609686,-0.1255117655,-0.0771685466,0.1159131825,0.3050191998,0.6307603121,0.0071912287,0.4278383255,-0.1079897285,-0.0974152535,-0.4439594746,0.0532793254,-0.1969467849,0.3724419177,0.1121944115,-0.7253928781,-0.0635103136,0.3224490285,0.1361628026,0.2982845008,-0.2639113665,0.0958463252,0.0580087714,0.1943989247,0.0376316346,-0.1580631137,0.0484823249,0.1013017073,-0.1880428344,0.0234057344,-0.0285302065,-0.4015385211,-0.1751411706,-0.0678827688,0.3376443982,0.1941541582,0.2729086876,0.3144955933,0.1436522007,0.1535956413,-0.1568152159,0.2404370606,0.1968843937,0.0478410907,0.1732318699,-0.1201198846,0.614027679,-0.132921353,-0.1693854779,-0.5334641337,-0.0926979855,0.054622665,-0.1340294778,0.2981455922,0.1477418244,0.1353397071,-0.1468390524,-0.0931917205,-0.1098138839,0.1542013735,-0.1216845736,0.0659909472,-0.2501848936,-0.0133989463,0.1680980027,0.1161622331,0.0418280475,0.2207736075,-0.236791119,0.2423516512,-0.0307029225,-0.4368209541,0.1188586056,-0.0872430354,0.21204409,-0.2846425772,0.0995701924,0.1902224422,-0.0575734451,-0.1081222445,0.1299331039,0.5976379514,0.3028028011,0.2508352697,-0.1207494661,-0.1952595264,0.1208019108,0.1041657552,0.1275493503,0.2294578999,-0.0426506549,0.187068224,0.0508228391,-0.063208431,0.3476133049,0.1688031256,0.0790319145,0.0559393875,0.0949998498,-0.1233065054,-0.1945202202,-0.475609988,-0.1981648207,-0.2342899144,-0.0374282338,0.3994510472,-0.1010120437,-0.1033794284,-0.0761078298,0.037747737,-0.2820747793,0.40622738,0.268860817,0.3722329736,-0.111432448,-0.2045547068,0.1339937449,-0.0873148814,-0.3267216682,0.4088355601,0.3220164776,0.2847061157,-0.2605342567,0.0110296225,0.2588570714,0.5978187323,-0.4492655993,0.0529889278,-0.4136816263,0.0936464071,0.4415519238,0.0858240947,-0.153981477,-0.1477378309,0.4645514786,-0.0769971535,-0.0638341978,-0.1418452561,0.149162218,-0.0171313416,0.0574150123,0.0441469252,0.4013847709,0.0907098129,-0.0258448478,-0.2388593704,-0.2930375934,-0.205279097,0.0681558773,0.2511330545,0.2939870954,0.515745163,-0.1066200286,-0.0260221828,-0.234261781,0.1801712513,-0.0148445992,-0.4119394124,-0.1430851966,0.0148807699,-0.1037403196,0.0439707264,0.2355529964,0.242524609,-0.0888504907,0.2390815318,0.0053054411,-0.2527035773,0.0994704366,-0.4123465419,-0.3238618076,-0.0328744762,0.3118266463,0.2092553824,-0.0464802608,-0.0502065271,0.1645802855,0.3762979805,-0.1588479131,-0.3540022075,0.2611423731,-0.1475060731,-0.1050851122,0.2282784134,0.3149432242,0.3020013273,-0.410178721,0.2333961725,-0.0974126905]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"Sorry, I was also talking to teven offline so I already had the PR ready before noticing x)","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":18,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\nSorry, I was also talking to teven offline so I already had the PR ready before noticing x)","embeddings":[-0.3764406741,0.0621025078,-0.0249254145,0.286175251,0.0446193144,0.0809495226,0.155846566,0.3971746266,0.0510593839,-0.0476375185,0.3336996734,0.004115683,0.0834176391,0.1582755297,-0.1321929693,-0.1995778978,0.1929605454,-0.0722633898,0.0871973783,0.2302482277,-0.2315653414,0.2663528323,-0.1839067787,-0.2031627148,-0.4015369117,-0.2735066712,0.1577025503,-0.0050987559,0.0436923541,-0.4615628123,-0.0991505682,-0.2882117629,0.4996251166,0.4922945499,-0.0001196299,-0.1714506298,0.2876414359,-0.0389574282,-0.2002032399,0.0875424296,-0.4354243875,-0.3970061839,-0.0468420871,-0.3596850038,0.3823304474,-0.1017141715,-0.348195076,-0.4584693611,0.6558184028,0.4028030336,0.1424899995,0.0162800271,0.1383096427,0.0170791205,-0.0757894963,0.0695119649,-0.0042966921,0.1023576409,-0.1442514062,0.3273944259,0.2477681339,-0.0633077919,-0.1642436981,-0.2750755847,-0.1225755438,-0.0865195766,-0.0925146863,-0.251667887,0.198066026,0.363743782,0.1841873974,-0.2028266639,-0.1457754076,-0.3202906549,0.0713843182,-0.6242572665,0.0502856746,0.5177451968,-0.0037660636,0.1118399054,-0.2235129178,-0.0908083618,-0.2312039435,-0.0297229458,0.0023631344,-0.2277643681,-0.1178186238,0.0040284269,-0.1674533933,-0.0984246954,-0.1111671031,-0.2227237374,-0.5720722079,-0.0140362605,-0.178409636,0.0443736389,0.136865288,-0.4408860207,0.1079220921,-0.0460403375,0.3337441683,-0.3853139877,0.2083592564,0.0500750467,0.5481048822,0.1906710416,0.2801466584,-0.1100368425,0.2598357201,-0.0213279296,0.0885011181,-0.0792431608,0.0471384935,-0.2001504153,0.0954323933,0.0385467969,-0.0271395184,-0.0987015888,-0.059927199,0.2017267793,-0.4078757763,-0.1518074423,-0.2807910442,-0.1851532608,-0.098366864,0.2293490767,0.1623693109,0.2720747292,-0.1685682833,-0.1085861698,-0.0424001999,-0.0828755498,-0.0747891665,-0.0433050208,-0.0337312743,-0.1575301588,-0.0203637742,0.2245582789,-0.0742422193,-0.2805363834,-0.1095870435,-0.0356901884,-0.1673441231,0.1796679348,0.285264045,0.1275248826,-0.1639618129,-0.2139405161,-0.0154010765,0.1968718767,-0.4836391211,-0.0749545172,-0.3604170382,0.1418749243,-0.1785311848,0.1125149131,-0.4827480912,-0.0880973339,0.1005527154,0.1740028113,0.3533980846,-0.2788318396,0.6303645968,-0.2761406004,0.0279032234,0.2305346727,-0.7488188744,0.0958139673,0.1825981885,0.0159449726,0.4496082664,0.3953610361,-0.0237344708,0.3770441413,-0.0881658271,0.2835766077,0.1006708369,0.1974977553,0.1171777025,0.4938797653,-0.072217837,-0.0063227187,0.018416658,-0.2992613018,0.6212943792,0.1084748283,0.5063618422,0.4043264091,0.0377526022,0.2353627086,-0.2143671811,-0.168838948,-0.0559027568,-0.1844295561,-0.3852595389,-0.055643566,0.1588097513,-0.174625054,0.296225518,-0.2451870888,0.118582204,0.4660310149,0.082875967,0.1488805562,0.0076211854,0.3006697297,-0.4398680031,-0.1985074878,-0.0882368535,-0.1634428203,-0.126803726,-0.0337225273,0.19733347,0.2187891901,-0.0381889343,0.1392454058,0.0512021445,-0.1272217929,-0.1436673552,0.17980434,-0.0061849961,0.2388722897,0.1822498143,0.0251253378,-0.3140966594,0.2853173912,0.2368263304,-0.2068870068,0.1841392964,0.1220637634,0.1675075591,-0.3401522636,-0.1555438042,0.2237614989,0.0257933401,-0.0869451091,-0.1305888891,0.124615401,0.128108263,0.0885499716,-0.1277747601,0.2919149697,-0.0615549199,0.0573416911,-0.1319109052,0.5513093472,-0.2694433033,0.3201516867,-0.0676286891,-0.0254177395,0.2279528677,-0.1107776985,-0.0442409292,-0.1289412677,0.2877562642,0.1512959301,-0.1880396307,-0.0283117518,-0.3569781184,0.0031246247,0.5258997679,0.0223967172,0.0300821047,0.1846691966,-0.0212222394,0.009899091,0.0738691986,0.0619092546,0.2687598765,0.1387556195,-0.0613799691,0.0710852742,0.2284822911,-0.067713663,0.2718981206,0.1209923401,0.2285152525,0.3146282434,0.1939519346,0.1624784768,-0.118125841,-0.4355782866,-0.2385646552,0.111990042,-0.3184938133,0.2901818156,-0.1427608579,-0.0482058153,-0.4705456197,-0.3678222001,-0.2575357556,-0.3024018109,-0.1673855931,-0.0986443087,-0.4747144878,-0.1406026036,-0.113705337,0.0823613703,0.0937031657,-0.1105913371,0.0388361849,-0.0006975631,-0.1376157254,-0.0897183269,0.0183231886,-0.1932256073,0.1084228531,0.0978774279,0.1040239632,-0.115459986,0.1239950135,0.1251721829,-0.0150250634,-0.0435006693,0.120246008,0.2838810086,-0.0367315859,-0.5128861666,0.033721026,0.1928918213,0.0333153158,0.1826905161,-0.0985661671,-0.174150452,-0.0930802599,-0.1191139147,-0.0344737433,-0.3550938666,0.4420220852,-0.0710536093,0.0302069243,-0.0339094698,0.1259305328,-0.1103534698,0.067188397,-0.284722507,0.2316463888,-0.4047389627,-0.0109364316,-0.2776227295,-0.2221847177,0.1345082223,0.1779789925,-0.0056694248,0.0136070233,-0.3946296573,0.6360439062,-0.0991959944,0.442253232,-0.1970859468,0.1547565311,0.1013635471,-0.0505038314,-0.0257120151,-0.0238358397,0.1536919326,0.1530253142,0.026020091,0.4894853532,-0.071928218,0.6651268601,-0.2164898962,-0.1908443719,0.7653546929,0.1234655678,0.2171909362,0.1095760241,0.0839432627,-0.1581361145,0.0156226475,0.0867911875,-0.1268138289,-0.1643209904,0.1484076977,0.0400952213,-0.1123043895,0.0120476028,0.0154588083,-0.4064573646,-0.3510844707,0.2497032434,-0.0109538483,0.293851912,-0.3571616411,-0.0692212656,-0.0584152266,0.0109491404,-0.2322793752,-0.0172876529,0.2271164656,0.1395562291,-0.5072889924,-0.4268551767,-0.4834094048,0.2316377163,0.0101298345,0.5988478065,0.1351298094,0.0753418654,-0.0829707608,-0.0961536616,0.6169036031,-0.0279185213,-0.1888392866,0.2580966651,0.382047385,-0.1696608514,-0.1103737652,0.1042626351,0.1699741781,0.8767220378,0.0393479131,-0.3058124483,-0.0314476714,0.2559388876,0.3879103065,-0.1291906834,-0.2496612072,-0.1338583827,-0.1862454563,-0.0875428468,-0.2613996267,0.1268766522,0.444992274,-0.1607404649,-0.152597487,0.0704758391,-0.3098804653,-0.010151566,0.0039964574,0.2558422983,-0.1063281596,0.1659551859,-0.1203412637,0.3574884832,0.4377621114,0.1959051937,0.0653356835,-0.3315350115,-0.082811892,-0.2703194022,0.4001966417,-0.1436281651,-0.213992998,0.0637364537,-0.1926668733,0.4227811694,0.1368255168,-0.2731592059,0.2954876125,0.0849181339,-0.8648973107,0.0916431248,0.3287992179,0.0798564851,0.283190906,0.1733735502,-0.365685761,-0.3700639009,0.2008540332,-0.1924816072,0.5724699497,0.4445646107,0.1451869458,0.0728926361,-0.3968038559,0.6061348319,-0.1380781084,0.0499902181,-0.194458738,0.4044934213,0.148077026,-0.1363782287,0.1013556048,0.0813191012,-0.1146971211,0.0374636315,-0.5104404688,-0.235532403,-0.2661845386,0.0068015018,0.254833132,-0.204574272,-0.4993151426,0.0002572995,-0.0960661247,-0.2361402661,-0.0767511725,-0.2121946961,0.0923323482,-0.1266333014,-0.3447324634,0.1222204119,0.067382127,0.1341661364,0.3516376317,-0.1614585072,0.1418845952,-0.1786920875,0.2616021335,0.1652996242,0.030198751,-0.0494894497,-0.0438409932,-0.0911645442,0.101738207,0.0304714236,0.0966755375,-0.0668495893,-0.1283182651,0.1746311039,0.0405150466,-0.2131315619,-0.1321575046,0.212876454,-0.2299484909,-0.3254306614,-0.0812895745,0.1551062316,0.2293550521,-0.1209973469,0.0432926938,0.1576850116,0.001223683,0.3172678947,-0.3695150614,-0.1802066863,-0.3000696599,0.0828467607,0.0605169125,-0.3104256988,0.2204139084,0.5383705497,-0.1888041049,-0.0755431056,0.4676379859,0.2899444699,-0.5295379758,0.5389208794,-0.0684270412,0.2785640657,-0.101506874,0.5041590333,0.0523462035,0.0035968937,0.195527792,-0.3853607178,-0.1970251054,0.0192924999,0.2155109644,0.0335064493,0.0403320193,0.5567016602,0.0400168262,-0.1997793019,-0.148713097,-0.0358410515,-0.1530419439,0.0441131629,-0.0679204836,0.0058577601,0.044663094,-0.0326013677,0.0643292442,0.115463905,-0.0202104561,-0.1217160001,-0.3338364661,0.1359889507,0.1383677274,-0.2122684568,0.1569171846,-0.0265429001,-0.2173281163,-0.0107315425,0.2423766702,0.0046660895,-0.0858207494,-0.0795536041,0.1157719567,-0.0999290422,-0.0826604664,0.1023366749,0.3121549189,0.6382740736,-0.0223691575,0.4231959283,-0.0875720456,-0.1236261427,-0.4861341417,0.0350304618,-0.168668136,0.3962354362,0.0872164741,-0.7172757387,-0.0945845544,0.2883245349,0.1118499637,0.3069839776,-0.2442878783,0.0710045621,0.0554245152,0.1902160048,0.0319104455,-0.1302335262,0.0334512666,0.106476903,-0.2001512796,0.0308408607,-0.0713485405,-0.3822000027,-0.1464611441,-0.1012207046,0.3103377819,0.1554685831,0.2649972141,0.3487348258,0.1282581091,0.1409234405,-0.1922421455,0.1946353018,0.1932111233,0.0450021587,0.1877515614,-0.1048113257,0.609526813,-0.1501395404,-0.1892778426,-0.5214866996,-0.1123236716,0.0434060916,-0.1085383296,0.2948245704,0.1533850133,0.1485003084,-0.109645389,-0.1047788262,-0.1312397569,0.1712991297,-0.1323463172,0.0783267245,-0.237548098,-0.0311540198,0.1510163248,0.1255936325,0.0470656008,0.2665865123,-0.2235806286,0.2642555535,-0.0198155344,-0.4577933848,0.1466539353,-0.1241188422,0.1841650307,-0.2721953392,0.1180655658,0.2160521001,-0.0673265904,-0.1253814846,0.0946724415,0.5830350518,0.2718704641,0.2764338851,-0.1151664555,-0.2004850209,0.11040923,0.0924615189,0.1398358494,0.2113536149,-0.0534771495,0.2168446332,0.0385827273,-0.0623799376,0.3692184389,0.1508293748,0.0476589873,0.0692039579,0.125108555,-0.0931200236,-0.2209441066,-0.4400243163,-0.2172728926,-0.2373041362,-0.0207324568,0.370136112,-0.0824688599,-0.1008038297,-0.071829766,0.0317697413,-0.2750126421,0.4035339355,0.233179301,0.3617777228,-0.1201839,-0.1971982569,0.149626568,-0.0485246442,-0.3061576188,0.3782445788,0.3052503467,0.2545855045,-0.2659119666,0.0281374548,0.2452299297,0.6335807443,-0.4491251111,0.0829716921,-0.4112226069,0.0556150004,0.4788252115,0.0878123939,-0.1390585154,-0.1592511237,0.4242450595,-0.0451971292,-0.0659674779,-0.1633172929,0.1127593517,-0.0138572147,0.072988674,0.0500523075,0.3976199031,0.0954840258,-0.0486511849,-0.2607992887,-0.2927933335,-0.214659065,0.056714911,0.273206085,0.3186675012,0.5101498961,-0.105161339,0.0070793643,-0.2610447705,0.1790618747,-0.0048993127,-0.3528938293,-0.1002612561,-0.0099243503,-0.1000591069,0.0584652051,0.2610979378,0.2403874248,-0.088178955,0.2841227949,-0.00165466,-0.245039016,0.1022859588,-0.4079575241,-0.3293607831,-0.0242526904,0.310549885,0.1907433271,-0.0387677811,-0.040448349,0.1713301688,0.39404127,-0.1353384405,-0.3708798587,0.2906871736,-0.115710482,-0.1221930832,0.236146301,0.3582156897,0.3031458259,-0.4336032569,0.2544978857,-0.0895151421]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"I was also already working in my PR... Nevermind. Next time we should pay attention if there is somebody (self-)assigned to an issue and if he\/she is still working on it before overtaking it... \ud83d\ude04 ","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":35,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\nI was also already working in my PR... Nevermind. Next time we should pay attention if there is somebody (self-)assigned to an issue and if he\/she is still working on it before overtaking it... \ud83d\ude04 ","embeddings":[-0.3961871862,0.1412499547,-0.0447770953,0.2652050853,0.0552897528,0.0680769011,0.155892238,0.3812042475,-0.0110884532,-0.0350051261,0.3722907901,0.0118637159,0.0816399232,0.1777890325,-0.1461665928,-0.1732498854,0.1780270934,-0.0427158028,0.0344856158,0.2413803339,-0.236317426,0.2680275142,-0.1862008572,-0.1968695521,-0.4044490755,-0.2788290977,0.1452196985,-0.0217896532,-0.0256087296,-0.48308447,-0.0936043188,-0.2329350412,0.4571807683,0.4846563935,-0.0001165353,-0.1735849828,0.3210092187,0.0024078621,-0.2353123873,0.0530448928,-0.4531081319,-0.3837760687,-0.0385644138,-0.3636563718,0.4063388705,-0.1223696694,-0.3413085043,-0.4172587991,0.6630897522,0.4112569392,0.1693600565,-0.0337702073,0.1536696553,-0.0172527619,-0.0696142092,0.1026432589,0.0065902681,0.0938838273,-0.1790848225,0.2880727053,0.2553639412,-0.058424864,-0.1424843371,-0.3091882467,-0.1423947513,-0.1177988574,-0.0604584888,-0.255299449,0.1998627335,0.3594552577,0.1978211254,-0.225909397,-0.1474906355,-0.311699748,0.113338083,-0.6214421391,0.0487781763,0.5457020998,-0.019328516,0.1305297017,-0.2266016304,-0.1034050137,-0.2398211062,-0.0559336543,0.0324959941,-0.2110430002,-0.1272573769,-0.0004993345,-0.1528132111,-0.1180546284,-0.0579757839,-0.2560561001,-0.5472071171,-0.0116200168,-0.1763406694,0.0649497658,0.1567471772,-0.424926132,0.1423417181,-0.0389583297,0.3401736617,-0.35565871,0.2603927255,0.0585185327,0.522449255,0.1673792899,0.2801506519,-0.0627358481,0.2633457482,0.0293107517,0.0720383078,-0.0799785852,0.0121569717,-0.2061021626,0.0648714453,0.0841450468,-0.0219788738,-0.0744552091,-0.0336105488,0.2073490024,-0.3438684642,-0.1791134328,-0.2502126992,-0.1604059935,-0.0843880698,0.2370717525,0.1838051826,0.2299661785,-0.1399749815,-0.0561746173,-0.0527573451,-0.0764699876,-0.0975608602,-0.0274426453,-0.0434101,-0.1516244411,-0.0221036784,0.2452819049,-0.0967631489,-0.2866630852,-0.1144346818,-0.055788435,-0.1505908966,0.1774312109,0.2737610936,0.0941286683,-0.1824358106,-0.1792254746,-0.029098222,0.1685352921,-0.4769117832,-0.0423701033,-0.3549825847,0.1847230196,-0.1505167633,0.1518642008,-0.4985494614,-0.0430408195,0.0819737092,0.1303446889,0.361082375,-0.27511397,0.6181912422,-0.2749441862,0.054061193,0.239700824,-0.7175135016,0.0810841471,0.13168706,-0.0445571654,0.4499322176,0.359482497,-0.0470605791,0.3819954693,-0.1250736266,0.286224395,0.1025873944,0.2136678249,0.0858998448,0.4878996015,-0.0910149962,-0.023390118,0.0445551425,-0.347625494,0.6521735191,0.1151155531,0.5111984015,0.3936977684,0.043070022,0.2332152575,-0.2062816769,-0.1988590509,-0.0399719886,-0.1981981099,-0.3763442338,-0.0651772842,0.1363246739,-0.2017886043,0.3150215745,-0.2656435072,0.1112892628,0.460814476,0.1455947757,0.1242725328,0.0168557186,0.2683566511,-0.4331436157,-0.193739742,-0.1475527585,-0.1563939154,-0.1399597824,-0.060787309,0.1878938526,0.217656523,-0.0309704784,0.1624682844,0.0908312201,-0.1438449323,-0.1194213405,0.2200558186,-0.030387003,0.2314020246,0.1400971413,0.0296045989,-0.2782298326,0.279717654,0.1940128952,-0.2066187263,0.2257094681,0.1197512522,0.1610760987,-0.339623332,-0.1450092643,0.207998842,-0.0224906057,-0.0433610789,-0.131381169,0.1153152585,0.1300602704,0.0743060783,-0.1831325591,0.2659928203,-0.0412702672,0.0790857375,-0.1302531064,0.5811470747,-0.2238218933,0.3129170835,-0.0337442122,-0.0359509028,0.2417221814,-0.132675305,-0.0523958802,-0.1432109028,0.2920071483,0.150668785,-0.1095961332,-0.0038493476,-0.2997734547,0.0130576845,0.5700365305,-0.0009287314,0.0543813966,0.2003084868,-0.010424342,-0.0128898723,0.0481963418,0.0733976588,0.2701425552,0.143492505,-0.0610888116,0.0817700177,0.2391866148,-0.0793039054,0.2866185009,0.1085496023,0.261750102,0.3045495749,0.1869598925,0.1485618353,-0.1427892298,-0.3971889019,-0.1929118484,0.0980407372,-0.2990089357,0.2812851071,-0.1797007471,-0.0317723639,-0.4606834948,-0.3426710665,-0.2431136966,-0.2658654153,-0.1617096663,-0.1251090318,-0.4919681251,-0.1103691682,-0.1233099625,0.1165727898,0.0983998626,-0.0651695579,0.048016537,-0.0258516893,-0.0790171176,-0.0626124367,0.0483229011,-0.1822835505,0.0869537145,0.1118431091,0.1157572567,-0.111856766,0.1259676367,0.1281925738,0.0135444012,-0.07983955,0.1168403402,0.2671394944,0.0046183686,-0.5441419482,0.0093050757,0.1585275382,-0.00065394,0.1666564196,-0.1094910428,-0.1939055622,-0.1040722504,-0.1158947349,-0.0497131273,-0.3589127362,0.4297665656,-0.1107399017,0.0624386147,-0.0566826947,0.14181979,-0.0834655389,0.0264389012,-0.2701607049,0.2086093426,-0.394708842,-0.0383481346,-0.2872618139,-0.2344918549,0.1153705046,0.135106191,-0.0412317701,-0.0148711391,-0.3945074975,0.59854877,-0.0911805034,0.4072619677,-0.1543261409,0.1598561704,0.1082134768,-0.0609392188,-0.0605885424,-0.0171037912,0.1243246719,0.1429907382,0.0405626632,0.4537538886,-0.0769098774,0.6510972381,-0.2171832919,-0.1933772713,0.7931034565,0.1460744441,0.2074933201,0.1000397354,0.089025721,-0.1275724024,-0.0149202254,0.1259258389,-0.1118104607,-0.1663991958,0.1738021076,0.0149771534,-0.1547592282,0.0323641524,-0.0189309437,-0.4336746335,-0.3735411465,0.2495989352,0.017278906,0.3149969578,-0.3155673146,-0.0837371424,-0.0497695245,0.08106938,-0.2002610117,-0.0312317349,0.2059081197,0.1080906913,-0.5688766241,-0.4343657196,-0.4829702079,0.2001951486,0.0003758477,0.5802284479,0.1253155023,0.1026378423,-0.0884906873,-0.1208553389,0.6239224076,0.0179618075,-0.2077336162,0.2325397581,0.3804987073,-0.1194761619,-0.1136853099,0.1311785579,0.204076305,0.8154348135,0.0144137042,-0.3362888396,-0.0315197371,0.2510600984,0.3690201938,-0.1367792934,-0.2473214716,-0.1221588477,-0.2105505466,-0.1063999832,-0.2394168824,0.1065255553,0.4578876793,-0.1845734119,-0.1501585692,0.0830235407,-0.309368819,-0.023828648,-0.0061984942,0.2689370215,-0.1025290489,0.1319540441,-0.1067499369,0.3046076298,0.4929367602,0.255545646,0.0591752715,-0.3818759322,-0.0648600236,-0.2778077424,0.3906009495,-0.0863911882,-0.2303906828,0.0898377821,-0.1147762537,0.4310974777,0.105524525,-0.3060207069,0.2839092016,0.0938793123,-0.868029952,0.0672663301,0.3517992198,0.0851373076,0.266337961,0.1425240189,-0.2993607819,-0.3890745938,0.2268268466,-0.2212848812,0.5617583394,0.4622302651,0.1309824586,0.0501690954,-0.3823641539,0.6059626341,-0.0751360655,0.0496064313,-0.1981856674,0.4180105329,0.159910664,-0.1468543857,0.1330642253,0.0571737476,-0.1249788627,0.0460069776,-0.5140420794,-0.2688521147,-0.2236864865,0.0458565131,0.2291859537,-0.195632562,-0.4780570865,0.0274467021,-0.0778480917,-0.2507957816,-0.0765789524,-0.2212525308,0.0822526217,-0.1366333663,-0.3644062281,0.1354053468,0.0133268964,0.1475530416,0.3296108246,-0.1759182215,0.1654904485,-0.2167600542,0.2632957995,0.1756990105,0.0403590836,-0.035515029,-0.0498479754,-0.0910549834,0.0518333837,0.0180130992,0.1397246122,-0.0720630214,-0.147631079,0.1676947027,0.0531062931,-0.1870756745,-0.0945010632,0.21206671,-0.2272666246,-0.3123402297,-0.136952132,0.1384249032,0.2350001037,-0.1234996766,0.065235965,0.154571712,-0.0345686674,0.3239793479,-0.326908052,-0.1750406474,-0.3134274483,0.0673527569,0.0197510105,-0.2836399078,0.2368383408,0.5338144302,-0.1851442605,-0.0918221548,0.454033941,0.3182340264,-0.5063896775,0.5193547606,-0.0397055373,0.255313307,-0.0921567976,0.4893008769,0.0531962626,0.0110742962,0.1609015912,-0.3807061613,-0.1938465238,0.0168898646,0.1840261668,-0.0162059031,0.0661963075,0.5399193168,0.040498618,-0.1953213811,-0.1874677092,-0.0092314305,-0.1997160316,0.0325760096,-0.0907296687,-0.0542655885,0.020311648,-0.0453841165,0.0894769356,0.1245509535,-0.0444692373,-0.1580920964,-0.3493080735,0.1171221659,0.126257211,-0.2073093504,0.1414673477,-0.0291936789,-0.2103943378,0.0005524177,0.2124820352,-0.011709311,-0.0763837099,-0.0766826347,0.0884387195,-0.1290406138,-0.083490476,0.1204849035,0.2823749185,0.6134369373,-0.0171169471,0.4254063964,-0.1109530032,-0.0925211608,-0.4530909359,0.0277668368,-0.1683810055,0.395414561,0.1107917279,-0.7080252171,-0.0672919378,0.3276768923,0.1321103722,0.2972359955,-0.2673638761,0.1047082171,0.0698422566,0.2139061689,0.0089479564,-0.1369238645,0.0516106673,0.0806510746,-0.1645413935,0.0097802486,-0.0316827111,-0.3678219616,-0.1688465923,-0.0747514367,0.3498595059,0.1809715629,0.2724546194,0.2772687078,0.1594463885,0.1348486245,-0.1623574495,0.2629074454,0.2012910098,0.0330722854,0.1631927937,-0.1277637184,0.5978223085,-0.121980615,-0.1643756181,-0.5197516084,-0.0607057177,0.0547770299,-0.1363772601,0.2861795723,0.1311553568,0.1412822604,-0.1452395022,-0.0822572559,-0.1150156036,0.154047817,-0.116171442,0.0658309758,-0.2405940294,-0.0113795288,0.1586648375,0.1075955853,0.0564217828,0.236765936,-0.2280269414,0.2521001697,-0.029671656,-0.4436670244,0.1266710311,-0.0706359744,0.2083536834,-0.2805272937,0.107360594,0.2040575445,-0.0592349879,-0.1331193298,0.1519403905,0.6114037037,0.2904869616,0.243024841,-0.1104627475,-0.1894607246,0.1082884073,0.0853863806,0.1249112487,0.244223848,-0.0256536007,0.1872606426,0.0772907734,-0.0963657796,0.3418993354,0.1763305068,0.0398245826,0.0606257953,0.0977976024,-0.1357241124,-0.1935822368,-0.4640133381,-0.192735225,-0.2539294362,-0.0277088173,0.4014793932,-0.1180350706,-0.0988111272,-0.0877253935,0.0542042963,-0.2858713269,0.4002655745,0.2522842288,0.36384359,-0.1063379645,-0.1944318861,0.135093227,-0.0689453855,-0.3103598058,0.3789413571,0.3164231479,0.2905765474,-0.2551521063,0.0086134942,0.2256198674,0.5962058306,-0.450405091,0.0481308438,-0.4447442293,0.1119725108,0.4354956746,0.0993830711,-0.1263372153,-0.1538099051,0.4535107613,-0.0902017131,-0.0474685989,-0.151319012,0.1403334141,-0.0010731497,0.0311159007,0.0529586524,0.3920421898,0.0907526687,-0.0732082725,-0.2500061393,-0.2859654129,-0.2042217851,0.0479783118,0.2739373446,0.335398972,0.5070761442,-0.116431959,-0.0498187691,-0.2381437272,0.1675784886,-0.0087683192,-0.39362517,-0.1375326514,0.0164312404,-0.0868346319,0.0525359437,0.2316702008,0.2337376326,-0.0883542448,0.2496708184,0.0207404476,-0.2763338983,0.1088415235,-0.3965429068,-0.3167789876,-0.0161589608,0.3135694563,0.2053137124,-0.0361023135,-0.0450268611,0.1803872138,0.3660028577,-0.1656155586,-0.3628181815,0.2734828889,-0.161482662,-0.0895518363,0.2288848162,0.3258042336,0.2899284959,-0.4333473444,0.2282893509,-0.08570683]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2615","title":"Jsonlines export error","comments":"The fix is available on `master` @TevenLeScao , thanks for reporting","body":"## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1","comment_length":11,"text":"Jsonlines export error\n## Describe the bug\r\nWhen exporting large datasets in jsonlines (c4 in my case) the created file has an error every 9999 lines: the 9999th and 10000th are concatenated, thus breaking the jsonlines format. This sounds like it is related to batching, which is by 10000 by default\r\n\r\n## Steps to reproduce the bug\r\nThis what I'm running:\r\n\r\nin python:\r\n\r\n```\r\nfrom datasets import load_dataset\r\nptb = load_dataset(\"ptb_text_only\")\r\nptb[\"train\"].to_json(\"ptb.jsonl\")\r\n```\r\n\r\nthen out of python:\r\n\r\n```\r\nhead -10000 ptb.jsonl\r\n```\r\n\r\n## Expected results\r\nProperly separated lines\r\n\r\n## Actual results\r\nThe last line is a concatenation of two lines\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Linux-5.4.0-1046-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyArrow version: 4.0.1\nThe fix is available on `master` @TevenLeScao , thanks for reporting","embeddings":[-0.405008316,0.1235859469,-0.0298681334,0.2725079954,0.0405567698,0.0832087025,0.1691810042,0.3880774379,-0.0028881533,-0.0449166968,0.3526557386,0.0118465247,0.0880170986,0.1497633606,-0.1447156519,-0.1722562164,0.1955797523,-0.0450855196,0.0241986774,0.2256097645,-0.2370414734,0.2739497721,-0.1929706782,-0.1880559474,-0.3998797536,-0.265417695,0.1470815837,-0.0317343585,0.0040826979,-0.4822725356,-0.0709139928,-0.2552482188,0.4480537772,0.4993006885,-0.0001194796,-0.1862578243,0.3148252964,-0.0045909937,-0.2177766711,0.0601914935,-0.4617322087,-0.4048300982,-0.0284081269,-0.3505156934,0.4142583311,-0.1208981946,-0.3492536247,-0.4158183932,0.657796979,0.4023522735,0.1435166746,-0.0400328785,0.1272985488,0.0049353405,-0.0746143684,0.1216396987,0.0098734433,0.099699147,-0.1540870667,0.2917856872,0.2802531421,-0.073027432,-0.1479884982,-0.3237630129,-0.1420194507,-0.1080806106,-0.0581665523,-0.2461665273,0.2157996148,0.3662033081,0.1880958229,-0.2147416174,-0.1680997908,-0.3279832006,0.0922151282,-0.6317983866,0.0422940813,0.5419908166,0.0049082595,0.105496183,-0.2468212992,-0.1166446507,-0.2461753935,-0.0548586883,0.0170327481,-0.2143408954,-0.1030853391,0.0154725816,-0.1451327801,-0.1205315739,-0.058422111,-0.2674475908,-0.5615639091,0.0003576541,-0.1826496571,0.0555582009,0.1309987158,-0.4299924672,0.1301409006,-0.0606913082,0.324067682,-0.3654105365,0.2496797293,0.0651632547,0.540812552,0.1556678116,0.2923827469,-0.0993172377,0.2539035082,-0.0060758987,0.0822605267,-0.0790791288,0.0162127595,-0.2027202696,0.0998308137,0.0818818286,-0.033647798,-0.0960862637,-0.0371447653,0.2070786804,-0.3853095472,-0.1794672161,-0.2511393726,-0.1714400351,-0.0819479302,0.2164943814,0.1527491361,0.258400172,-0.1432008147,-0.0608004145,-0.0422608666,-0.1053021029,-0.0851033852,-0.037050005,-0.0292017646,-0.1636219323,-0.007392507,0.2365478277,-0.0824336559,-0.2791728079,-0.101925388,-0.0656405538,-0.1604738832,0.1852317452,0.2863330543,0.1264008135,-0.1883973479,-0.1539654136,-0.0332119018,0.194177866,-0.486875087,-0.0549434647,-0.3442725837,0.1513715386,-0.1686931849,0.1538721025,-0.5203329921,-0.0621901862,0.0909727365,0.1379349083,0.3638911843,-0.2885657847,0.5915679336,-0.2633570135,0.0328185931,0.2447248846,-0.7416411042,0.0727430582,0.1433322579,-0.0289831795,0.4514209032,0.3837051988,-0.0611067079,0.3714471459,-0.1115526035,0.2810883224,0.0980394632,0.2072175592,0.0877893269,0.5092478395,-0.0944817588,-0.0340407602,0.0189527087,-0.3228698075,0.6582375765,0.1010507643,0.5133690834,0.4022120237,0.0486898385,0.2361323684,-0.1882143319,-0.1926927865,-0.0239244793,-0.2018441111,-0.3565918803,-0.0601259582,0.1517152935,-0.1956804097,0.3046914041,-0.2512962818,0.1205796376,0.4577215314,0.1424019635,0.1354188621,0.0149161173,0.2923644781,-0.4403756559,-0.1856592298,-0.1384824216,-0.1751016229,-0.1388922483,-0.0479185097,0.2108189762,0.2307403982,-0.0279270615,0.1985749304,0.055883836,-0.1572674066,-0.1199251264,0.2022698373,-0.0067227962,0.2208934128,0.1492545903,0.0473087728,-0.264901787,0.2993874848,0.2316002101,-0.197079882,0.2262210399,0.1153742969,0.182331115,-0.346088469,-0.1354358792,0.2160686255,0.0085572368,-0.0384906456,-0.1419818848,0.0999190733,0.1285383105,0.0714925602,-0.1756116748,0.273702383,-0.0449263938,0.0702755004,-0.1108802557,0.5756208897,-0.2315481752,0.3267292976,-0.054457508,-0.0390514955,0.2353304625,-0.109222278,-0.0498516969,-0.1329450607,0.2927764654,0.1700849384,-0.1377172172,0.0078718923,-0.3079791069,0.0008201345,0.5528987646,-0.0061938073,0.0509872586,0.1897959411,-0.0245077759,-0.0118631879,0.0489390232,0.0731477588,0.2852787375,0.1429065168,-0.070137687,0.0829501003,0.230766654,-0.071425125,0.2879239321,0.1323756725,0.278067261,0.3093650639,0.2001063377,0.1442139298,-0.1392228752,-0.3962332904,-0.1813122779,0.0995324105,-0.2937791348,0.3056457937,-0.1745990664,-0.0234723892,-0.4829624891,-0.3484551907,-0.2511421442,-0.2831969857,-0.1795460582,-0.1242892444,-0.4628697634,-0.1420505494,-0.1334463805,0.0935841352,0.0989627913,-0.0877343938,0.036773406,-0.0189095996,-0.0949810594,-0.0846271291,0.0215479322,-0.1893379837,0.0760381967,0.1250029057,0.1208595186,-0.1129184291,0.1357406527,0.1078993082,0.0172779802,-0.0921800062,0.1095613316,0.2539359033,-0.0022323544,-0.511523366,0.0121291569,0.1757571697,0.0129172886,0.167890206,-0.1153367609,-0.1633795798,-0.0993216187,-0.0726104379,-0.0444507673,-0.3282046616,0.4324097335,-0.1114400402,0.0662648976,-0.0897882208,0.1328062713,-0.1148564592,0.0188318379,-0.3029354215,0.2101678699,-0.3862332106,-0.0119210547,-0.2858659327,-0.2235918045,0.1261739731,0.1688201129,-0.0404321402,-0.0074139005,-0.3865295947,0.6128473282,-0.0879221857,0.3944288194,-0.1581394225,0.1515704095,0.1097089201,-0.0550442748,-0.0376509912,-0.0200130548,0.1056890488,0.1585512012,0.043169681,0.4706217051,-0.0769215599,0.697093606,-0.220825389,-0.1902220249,0.7875600457,0.1402644813,0.2048187256,0.0987920538,0.0756180435,-0.1386789232,0.0004486932,0.1139988601,-0.0954322517,-0.1477368474,0.1499375552,0.0136428382,-0.1374288797,0.0434505902,-0.0267591067,-0.4090518057,-0.3595278561,0.2341947258,0.0138217043,0.322183758,-0.310695529,-0.0660416782,-0.0452389307,0.0640351772,-0.2116262168,-0.0413259827,0.2116163969,0.1316150725,-0.5474601388,-0.4660968482,-0.4979642034,0.21285972,0.0042523551,0.5995929241,0.1321567446,0.1254973561,-0.0822201893,-0.1085729301,0.6382525563,0.0052620536,-0.1892095804,0.2247469276,0.3766596913,-0.1490101814,-0.1213854253,0.1479465365,0.2141778022,0.8320970535,0.0211837906,-0.3263210654,-0.0284923147,0.2498813272,0.3635897636,-0.134477973,-0.2507669032,-0.1250063181,-0.1887818277,-0.0906164646,-0.2398902327,0.0880281925,0.4583913088,-0.1895028949,-0.1577547044,0.0812819377,-0.3138997555,-0.0098850336,-0.0010502191,0.2586965859,-0.0975641981,0.1388719827,-0.0903831497,0.3072870374,0.5018109679,0.2491776943,0.0612492412,-0.3671596944,-0.0612184405,-0.269784987,0.4074902833,-0.1193003431,-0.2467735857,0.0892221853,-0.1366838366,0.4067941904,0.108351849,-0.3002580702,0.311976701,0.0954423398,-0.8631698489,0.0747869983,0.3389799297,0.0810857192,0.2705108821,0.1383829266,-0.2971597612,-0.4020295441,0.1909779012,-0.2115793377,0.5402980447,0.4408109188,0.1340799928,0.0566079766,-0.3967542052,0.606908977,-0.054130327,0.0473060124,-0.2041593045,0.4511749148,0.1480282247,-0.1569631994,0.1203833818,0.0616617948,-0.1042971462,0.0445175804,-0.5250357389,-0.250277102,-0.2495995164,0.0209385008,0.2398371994,-0.1943864077,-0.486472398,0.0058816983,-0.0766102523,-0.2416293025,-0.0848643929,-0.2178914249,0.086648494,-0.1396303773,-0.3671004474,0.1276575923,0.0373417363,0.1346300393,0.3383458853,-0.1834291369,0.1848900318,-0.2057003975,0.2556527853,0.1756949276,0.0331248417,-0.0398071036,-0.0869682282,-0.0792134404,0.0773596317,0.0081077246,0.0922336727,-0.0523940697,-0.1346260458,0.1799896359,0.0624195375,-0.1951021999,-0.0762695819,0.2278955579,-0.2693978548,-0.3249509037,-0.1447608322,0.1316221505,0.2370776832,-0.1230130866,0.0402577557,0.1823360026,-0.0118777957,0.3279649913,-0.3542585373,-0.1826790124,-0.2977995276,0.0726243854,0.0354694985,-0.3020884395,0.2518725991,0.5221651196,-0.1780214459,-0.0758403167,0.4633221924,0.3301824331,-0.5119114518,0.5208441019,-0.0547270514,0.2744197845,-0.0896200016,0.4796876311,0.0294849761,0.0042985352,0.1708033234,-0.3854738772,-0.1746601015,0.0240966622,0.1879249513,-0.0112349326,0.0715379715,0.5349473357,0.031402804,-0.2109509408,-0.151205346,0.0130630676,-0.1701929718,0.0375986919,-0.1098505706,-0.016135186,0.0309744496,-0.0472330898,0.0527117364,0.1176920384,-0.0368794911,-0.1284762025,-0.345143348,0.1326240599,0.13119331,-0.2080461681,0.1377689093,-0.0293362625,-0.2110309303,0.0145669878,0.215242222,-0.0121053951,-0.0872124285,-0.0669713169,0.1054671928,-0.1365270615,-0.0941986218,0.1122084334,0.2895731628,0.6160725951,-0.0060389657,0.4469953775,-0.0944463685,-0.1031380296,-0.4617595375,0.0454332121,-0.155441761,0.3760178983,0.1151607856,-0.7323026657,-0.0590881184,0.3152715862,0.1347816139,0.2978375852,-0.2660380304,0.0920103118,0.0535863601,0.1844918728,0.0201788675,-0.1720852107,0.0569990426,0.1025939807,-0.1787149906,0.0191148408,-0.0309568867,-0.3914044499,-0.1784975529,-0.0821969733,0.3440998793,0.1997227818,0.2555519938,0.3409826458,0.1627792269,0.1615519226,-0.149455443,0.2628943026,0.2081009895,0.0623160452,0.1892818213,-0.1043664068,0.601788938,-0.1368480921,-0.1821700931,-0.5330044031,-0.0942562446,0.0453863032,-0.1365870386,0.2952599525,0.1411737204,0.1332967877,-0.1303298771,-0.0896984786,-0.1326977015,0.1580142081,-0.1283101439,0.0470174588,-0.2339497656,-0.0030484884,0.1792140603,0.1273174882,0.03237563,0.2298112661,-0.243622303,0.2448901832,-0.0126785496,-0.4355149865,0.13095662,-0.0723523721,0.1937611401,-0.2801180184,0.0993718654,0.1804019213,-0.0686322004,-0.1141299456,0.1239871979,0.5954982042,0.2786192,0.2679579258,-0.1112586558,-0.1955702007,0.1202879027,0.1089356542,0.1356731951,0.2388647199,-0.0357823819,0.1886294335,0.048322048,-0.0646266937,0.3366886675,0.1740183383,0.0634851009,0.0751289651,0.0871341601,-0.1306285858,-0.2107579708,-0.4721159935,-0.1948927492,-0.2264468968,-0.0259539895,0.4003715813,-0.0991885737,-0.1061838046,-0.0792044625,0.0349020995,-0.2885478139,0.4099138081,0.2506617606,0.3784333169,-0.1104174033,-0.2068300098,0.1516787261,-0.0750792846,-0.3335448503,0.3990055919,0.3340448141,0.285682112,-0.2654836476,0.0133764055,0.2545263469,0.6111516356,-0.4400881529,0.0471823402,-0.4072888792,0.0969798863,0.4501033127,0.0994786397,-0.1506197453,-0.1467422843,0.4659294784,-0.0992753357,-0.0714195073,-0.1525705904,0.1305383146,-0.0389366448,0.0370242074,0.0455461964,0.3978365362,0.0849453211,-0.0407514572,-0.2249807864,-0.27852422,-0.1984531581,0.063682206,0.2560340762,0.293315351,0.5096836686,-0.1099316254,-0.0177363101,-0.2223049253,0.1686456352,-0.0195286255,-0.4072673023,-0.1441032141,0.0168474,-0.1150060445,0.0529543124,0.2415007949,0.2285186648,-0.0704910979,0.2415397763,0.0161693394,-0.2320394963,0.0925425962,-0.4091297388,-0.3230266869,-0.0519254953,0.3152573109,0.2143428773,-0.0504888743,-0.0328557305,0.1748898476,0.3723247349,-0.1480707228,-0.3696187139,0.2554557323,-0.1531239748,-0.0919132978,0.2206361443,0.3146368265,0.2839918435,-0.4200247228,0.2236915827,-0.1069282964]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2607","title":"Streaming local gzip compressed JSON line files is not working","comments":"Updating to pyarrow-4.0.1 didn't fix the issue","body":"## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0","comment_length":7,"text":"Streaming local gzip compressed JSON line files is not working\n## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0\nUpdating to pyarrow-4.0.1 didn't fix the issue","embeddings":[-0.1992494613,-0.1339781284,-0.0091445101,0.2391350567,0.0777640566,0.0168800913,0.3976738453,0.5178918839,0.1970213056,0.0789984018,0.1030013189,0.3306827545,-0.0883667096,0.0208737329,0.2288863659,-0.1809338927,-0.0436617956,0.3560613692,-0.0251667686,0.0015848613,-0.1305664033,0.2937984169,-0.1404591352,-0.0185168069,-0.0187981017,-0.0395553522,-0.0471503064,0.1500941366,-0.0939593539,-0.3326378763,0.0627696067,-0.1684151441,0.188022688,0.4454552829,-0.000114536,0.288641125,0.2961105406,-0.1629139781,-0.3886071146,-0.4712482095,-0.229901135,-0.1492993236,-0.1307716817,-0.2858453691,-0.1378189325,-0.3474362791,-0.0337565579,-0.5071999431,0.5416871309,0.2922603488,0.1556637138,0.0419651642,-0.061772272,0.0905790925,0.0390155055,0.1573239863,-0.0886082724,0.3150972426,0.0327884369,0.1149678901,0.0552849993,0.4286454916,-0.2254960537,-0.1747770905,0.1473699361,-0.0008157574,-0.0843830332,-0.1257867962,-0.1683152765,0.085735105,0.1845451146,-0.4122557342,-0.4398344159,-0.3761076331,0.0466741547,-0.44880265,0.2118544728,0.4092858434,-0.250308305,0.1969692856,-0.100665614,-0.130531922,-0.3545617759,0.0854526162,0.36184901,0.3480824232,-0.0946776345,0.0091285398,-0.0919867903,0.1193795577,-0.0839845538,-0.3859618604,-0.1857065707,0.227756381,-0.2421413958,-0.1349226832,0.1443624049,-0.6062148213,0.1511498541,0.1012284085,0.4615779221,-0.1257445961,0.3588074744,0.0546084084,0.2728678286,0.1870484799,0.1769655049,0.2638421357,0.0566591471,0.3898805082,-0.0865667164,-0.1980740726,-0.2363374233,-0.0394049771,-0.0915300399,0.1368107945,0.4833262861,-0.1576653421,-0.089093484,0.081910789,-0.5941808224,-0.1327005029,0.1576721966,0.0986096784,0.0661023408,0.2201749086,0.026484726,0.1610841751,-0.002104495,-0.0618949644,-0.0907956734,0.0235721208,0.0740586743,0.1347989291,0.251897037,-0.5213885903,0.2793646455,0.0087546054,-0.046029456,-0.1562338322,0.0058455556,-0.12555632,0.1504514217,0.146838814,-0.0259293783,0.0193639155,0.2058843374,-0.1102267727,-0.1738193929,0.1183539331,-0.2751955688,-0.2672133446,0.0372359268,0.2010167539,0.0381590351,0.009389149,-0.0081491908,0.0059310379,0.0556991696,-0.3565125763,0.023636166,-0.0094647072,-0.0558016375,-0.0841292068,0.2759922445,0.4360964596,-0.3608719409,-0.1035921797,0.0715648532,0.2175855786,0.7216550708,0.3425383568,-0.2379719913,0.1273078024,-0.3681203723,0.3619052172,0.6931351423,0.0160383359,-0.3624646366,0.8522378206,-0.1086349562,0.3817502558,0.4800842702,-0.20442909,0.1941484958,-0.0105180815,0.3352073431,0.2812868953,-0.0445720889,0.0385273471,-0.464271307,0.1312143207,0.0419497304,0.0378056802,-0.2513179779,0.2954408824,-0.0433923602,0.015032988,0.3520895541,-0.1800565869,0.1081850231,0.2904555202,0.2378064245,0.0199463479,0.2021734118,0.0092641879,-0.2178035229,0.0324927419,-0.3366108537,-0.2629368603,-0.2826167047,-0.0624102913,-0.1401127279,0.1487002075,-0.1843477637,-0.3533777595,0.1464820653,0.4111509025,0.3808980882,0.0224370789,-0.4623301327,0.3006553948,-0.2987254858,-0.1048559621,-0.2611705363,0.429363668,0.0721354857,-0.2631469369,0.1758847386,-0.0589892119,0.1717893928,-0.2416345477,-0.2221167684,0.2450815886,-0.1208557263,0.1709071249,0.1169753596,-0.2226256579,0.2957748175,-0.0062111109,-0.0720564499,0.3445874751,0.2193278074,0.2041208595,-0.2239909172,0.4739874303,-0.0305651519,0.3029977679,0.0209108982,-0.205836162,0.1924333572,-0.0084315231,-0.3421524465,0.0653704852,0.3345847428,-0.1258885711,0.0946051851,-0.0612957813,-0.3791344762,-0.0635569915,0.5271067023,-0.0830585435,-0.1615021825,0.1926877648,-0.1803122163,-0.0555438772,0.2965638041,-0.0067166113,0.3335767984,0.1575452387,0.104602471,0.0166130755,-0.1111166552,-0.2565548122,0.2091605067,-0.0900344998,0.1146262661,0.5449168086,0.0356366746,0.0449167937,-0.3995397389,-0.2428149432,-0.128882125,0.1455808133,-0.0990883484,0.384134531,-0.3911352754,-0.3713185489,-0.1201700941,-0.1128693894,-0.0630642995,-0.0857734159,-0.0269654337,0.2985586822,-0.2004606724,0.2329237461,0.0005526908,0.0394016802,0.251948297,-0.4227379262,-0.4818938971,-0.1630378664,-0.191156432,-0.0104031721,0.2656034231,-0.2878884673,0.3017160296,-0.0459037386,-0.3699751496,-0.2707997561,0.1194068417,0.1435605735,0.0721819699,-0.0227004942,-0.0912801325,0.2280841321,-0.0796872899,-0.3324358761,0.287009567,-0.2174522281,-0.1486556232,0.194851473,-0.0311844051,-0.0018445911,-0.2352986038,-0.2655209303,-0.1472442448,-0.4852220118,0.289681673,-0.0403695032,0.3086782992,0.2429099977,0.1440986991,0.1779596359,0.1556582451,-0.0546927191,-0.1509623826,-0.6362751126,0.3917452991,-0.1101357043,-0.2303782552,-0.0295004416,-0.1555970311,-0.0550973527,0.3612907529,-0.4562184811,0.237912789,0.0702732503,0.0834183842,-0.0950674713,-0.134057045,0.1914462298,-0.1075467318,-0.0603607446,-0.2590092719,0.0759437755,-0.0122350994,0.2527312934,0.1833108217,0.4013436437,0.7469967604,-0.039161507,0.2865437567,0.353391856,0.1458686143,0.5503771305,-0.1300928146,0.16025047,-0.2425827533,-0.2188410312,0.0596462861,0.1050836518,-0.3196414113,0.0274330787,0.0012185025,-0.3406588137,-0.3240846992,0.1244037524,-0.3062015474,-0.1882945746,0.2014841437,-0.0963935554,0.1964119375,-0.0394231714,-0.0296395868,-0.0032317135,0.1153508127,-0.048328694,0.2654097974,0.4285131991,0.1289328188,-0.0543236472,-0.1830666661,-0.3484928906,0.078764379,0.2743041515,0.6020610929,-0.1669999659,-0.1687550992,0.0798476189,-0.1684129536,0.4636302292,-0.1258804947,-0.0341541246,0.3472768664,-0.1242791489,-0.0266533308,-0.1190542877,-0.0431695469,0.3530095518,0.195949018,0.6723818183,-0.1772067547,-0.220200792,0.1132560149,0.0215652157,0.0907032862,-0.156267941,-0.2928073406,-0.1358119994,-0.201933071,-0.0655045658,-0.0717766806,0.2410783023,0.0199561808,-0.131281808,0.0536011457,-0.0730547979,-0.1126663312,0.0078033102,0.294038862,-0.0325312726,0.1593426615,0.057022687,0.5039331317,0.232784465,0.5136659145,0.5314756632,-0.4147776663,0.1325844377,0.0702452734,0.1329945028,0.1649885625,-0.1350931078,-0.1513888985,0.138104111,0.2574620247,-0.0355816707,-0.0832144171,0.2338197678,-0.1555819809,-0.3567377925,-0.4606517553,0.2432612181,-0.0392031707,0.1920537502,0.7787454724,-0.6046575308,-0.1680700779,0.1287034303,-0.327439338,0.6506022215,0.3056919277,-0.01261207,0.4230307937,-0.1530070156,0.4291078448,-0.4506242871,0.3880233169,-0.107545048,-0.1895088851,0.0136250248,-0.1521783918,0.3272462487,0.1736504287,-0.3301243484,0.1560810655,0.0949875191,-0.0261116754,0.056402456,0.0844660252,-0.3410910368,-0.1478554606,-0.6056392193,0.0822727308,-0.1643425077,0.3591601551,0.119864881,-0.3871607482,0.0100027071,-0.2189189792,-0.1640434861,0.4873065948,-0.2205232978,-0.0437984429,-0.1201645508,-0.6457443237,-0.3231777251,-0.1665352583,-0.3695810437,-0.1014870033,-0.0968803316,-0.0140162325,-0.1560026407,0.0815750286,-0.4244860113,0.2442319244,0.1879109144,-0.1447571814,-0.0854346454,0.1730190217,-0.2793938816,-0.55167979,0.2067956775,-0.0945816264,0.2443367243,0.1387685984,-0.014279468,0.0188988578,-0.1952127963,-0.2021060437,0.1315957457,-0.1408141851,-0.3405287564,-0.2667013407,-0.151902318,-0.200861007,-0.1044818312,0.2798337042,0.0701378211,0.3024792969,0.3966619372,-0.0176074822,-0.0633016229,-0.2849562764,0.0822029561,-0.1994450986,-0.0627179369,0.1458605379,-0.0238583386,0.1735637039,-0.2250163257,-0.0105745681,0.1574327648,0.0527799204,0.0311236531,-0.4647310376,-0.1359294653,-0.1071501747,-0.0055007907,0.1117188111,0.1044034883,0.2052558213,0.1506942809,-0.1848126501,-0.287404716,-0.1490447223,-0.1798323691,0.0388781689,0.0197014529,0.1445133686,0.3058988154,0.0417761318,0.1726244986,0.1845808923,0.1145528182,-0.2760638893,-0.1294748783,0.1017443016,0.1475505531,-0.1066455618,-0.2568067312,-0.0784032717,-0.239744544,-0.0903129727,0.1030201986,-0.0890858918,-0.0714164227,0.2204833478,0.3504038453,-0.0414797217,-0.1651158482,-0.0696929991,-0.2029707879,0.3246166408,-0.149192661,0.0957456976,0.0479332507,0.0756747648,-0.5871013403,-0.0813039467,-0.0964607447,0.1543788612,0.59524405,-0.4551479816,0.03112543,0.0455692224,0.1699540019,0.4789012969,-0.1304557621,0.0390133224,0.1729512513,0.1876669079,-0.0484990031,-0.0702733546,0.1377959102,0.1060291827,-0.0637325421,-0.0711321831,-0.1525841057,-0.1502206475,0.0998598635,0.0143192159,0.3330241144,0.0458889343,0.211354658,0.4080850184,0.2017606944,0.0718020722,-0.3809928596,0.0200543348,0.0718998685,-0.1887304932,-0.1524383426,-0.1257040054,0.0813989639,0.1215052903,-0.3927783072,-0.3188354373,-0.3502144217,0.0856014788,-0.1179064736,0.4086128771,-0.0059121046,0.2030270994,-0.2063986659,0.0338467211,0.0438848063,0.16065979,0.0185740534,0.115962036,-0.5277093649,-0.031080272,-0.0583055541,0.0389653742,0.3544371128,0.5155139565,-0.0361992531,0.2214543521,-0.1061519906,-0.1359117925,0.0883783028,0.1771474481,0.3425745368,-0.3739395738,0.2876466513,0.2135393769,0.0820103437,-0.0754379928,0.0099544805,0.3729601204,0.3608309031,0.0665017441,0.0278094839,0.0363483615,-0.0933402553,0.0266610924,0.335478574,0.1904851943,-0.1879787147,0.2221895158,0.1101663113,-0.088712424,0.3944174051,0.0983025655,0.1196607053,-0.1152643189,0.1943055987,-0.247738108,-0.1086110324,0.0640453398,-0.354144752,-0.358769089,-0.0634774119,0.2332483977,0.1124899834,0.0440791547,-0.0250769686,0.0849539489,-0.1289328188,0.279007405,0.2456029654,-0.0799998939,-0.1668445021,-0.1844663471,-0.2015265524,0.1580558419,-0.040410243,-0.1079189107,0.2371376604,0.3279328048,-0.0682448745,0.2060449421,-0.0900718644,-0.1822002977,-0.2946017087,0.2807135284,-0.3546518683,-0.2369975001,0.1787969619,0.2335970402,0.0910652205,-0.4543965757,0.2050732076,-0.3312031925,0.1178356186,0.0553922281,-0.0305205919,0.0156790111,0.0668511316,0.2356533855,0.1621074677,0.1155462116,-0.1349952668,-0.192071259,-0.072820738,-0.176847145,-0.0183110926,0.1458830535,-0.0582501702,0.502874732,-0.0484356917,-0.4014241993,-0.4371389151,0.4285812378,-0.0439626239,-0.4439990819,0.0103937835,0.0140104787,-0.1760369986,0.1818829775,0.0124762263,0.1777729988,-0.0506044216,0.0616825111,-0.3085522056,-0.1021585092,0.5942243338,-0.4224966168,-0.3192535639,0.3241405487,0.1266875267,0.1190700009,0.1298958212,-0.1965908855,0.1324760169,0.437204659,-0.0283873919,0.1215384007,0.1597634852,0.1152027473,0.1851295084,0.1178187653,0.0776488185,0.1725968719,-0.259760052,0.1329277754,-0.214050144]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2607","title":"Streaming local gzip compressed JSON line files is not working","comments":"Here is an exemple dataset with 2 of these compressed JSON files: https:\/\/huggingface.co\/datasets\/thomwolf\/github-python","body":"## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0","comment_length":13,"text":"Streaming local gzip compressed JSON line files is not working\n## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0\nHere is an exemple dataset with 2 of these compressed JSON files: https:\/\/huggingface.co\/datasets\/thomwolf\/github-python","embeddings":[-0.1992494613,-0.1339781284,-0.0091445101,0.2391350567,0.0777640566,0.0168800913,0.3976738453,0.5178918839,0.1970213056,0.0789984018,0.1030013189,0.3306827545,-0.0883667096,0.0208737329,0.2288863659,-0.1809338927,-0.0436617956,0.3560613692,-0.0251667686,0.0015848613,-0.1305664033,0.2937984169,-0.1404591352,-0.0185168069,-0.0187981017,-0.0395553522,-0.0471503064,0.1500941366,-0.0939593539,-0.3326378763,0.0627696067,-0.1684151441,0.188022688,0.4454552829,-0.000114536,0.288641125,0.2961105406,-0.1629139781,-0.3886071146,-0.4712482095,-0.229901135,-0.1492993236,-0.1307716817,-0.2858453691,-0.1378189325,-0.3474362791,-0.0337565579,-0.5071999431,0.5416871309,0.2922603488,0.1556637138,0.0419651642,-0.061772272,0.0905790925,0.0390155055,0.1573239863,-0.0886082724,0.3150972426,0.0327884369,0.1149678901,0.0552849993,0.4286454916,-0.2254960537,-0.1747770905,0.1473699361,-0.0008157574,-0.0843830332,-0.1257867962,-0.1683152765,0.085735105,0.1845451146,-0.4122557342,-0.4398344159,-0.3761076331,0.0466741547,-0.44880265,0.2118544728,0.4092858434,-0.250308305,0.1969692856,-0.100665614,-0.130531922,-0.3545617759,0.0854526162,0.36184901,0.3480824232,-0.0946776345,0.0091285398,-0.0919867903,0.1193795577,-0.0839845538,-0.3859618604,-0.1857065707,0.227756381,-0.2421413958,-0.1349226832,0.1443624049,-0.6062148213,0.1511498541,0.1012284085,0.4615779221,-0.1257445961,0.3588074744,0.0546084084,0.2728678286,0.1870484799,0.1769655049,0.2638421357,0.0566591471,0.3898805082,-0.0865667164,-0.1980740726,-0.2363374233,-0.0394049771,-0.0915300399,0.1368107945,0.4833262861,-0.1576653421,-0.089093484,0.081910789,-0.5941808224,-0.1327005029,0.1576721966,0.0986096784,0.0661023408,0.2201749086,0.026484726,0.1610841751,-0.002104495,-0.0618949644,-0.0907956734,0.0235721208,0.0740586743,0.1347989291,0.251897037,-0.5213885903,0.2793646455,0.0087546054,-0.046029456,-0.1562338322,0.0058455556,-0.12555632,0.1504514217,0.146838814,-0.0259293783,0.0193639155,0.2058843374,-0.1102267727,-0.1738193929,0.1183539331,-0.2751955688,-0.2672133446,0.0372359268,0.2010167539,0.0381590351,0.009389149,-0.0081491908,0.0059310379,0.0556991696,-0.3565125763,0.023636166,-0.0094647072,-0.0558016375,-0.0841292068,0.2759922445,0.4360964596,-0.3608719409,-0.1035921797,0.0715648532,0.2175855786,0.7216550708,0.3425383568,-0.2379719913,0.1273078024,-0.3681203723,0.3619052172,0.6931351423,0.0160383359,-0.3624646366,0.8522378206,-0.1086349562,0.3817502558,0.4800842702,-0.20442909,0.1941484958,-0.0105180815,0.3352073431,0.2812868953,-0.0445720889,0.0385273471,-0.464271307,0.1312143207,0.0419497304,0.0378056802,-0.2513179779,0.2954408824,-0.0433923602,0.015032988,0.3520895541,-0.1800565869,0.1081850231,0.2904555202,0.2378064245,0.0199463479,0.2021734118,0.0092641879,-0.2178035229,0.0324927419,-0.3366108537,-0.2629368603,-0.2826167047,-0.0624102913,-0.1401127279,0.1487002075,-0.1843477637,-0.3533777595,0.1464820653,0.4111509025,0.3808980882,0.0224370789,-0.4623301327,0.3006553948,-0.2987254858,-0.1048559621,-0.2611705363,0.429363668,0.0721354857,-0.2631469369,0.1758847386,-0.0589892119,0.1717893928,-0.2416345477,-0.2221167684,0.2450815886,-0.1208557263,0.1709071249,0.1169753596,-0.2226256579,0.2957748175,-0.0062111109,-0.0720564499,0.3445874751,0.2193278074,0.2041208595,-0.2239909172,0.4739874303,-0.0305651519,0.3029977679,0.0209108982,-0.205836162,0.1924333572,-0.0084315231,-0.3421524465,0.0653704852,0.3345847428,-0.1258885711,0.0946051851,-0.0612957813,-0.3791344762,-0.0635569915,0.5271067023,-0.0830585435,-0.1615021825,0.1926877648,-0.1803122163,-0.0555438772,0.2965638041,-0.0067166113,0.3335767984,0.1575452387,0.104602471,0.0166130755,-0.1111166552,-0.2565548122,0.2091605067,-0.0900344998,0.1146262661,0.5449168086,0.0356366746,0.0449167937,-0.3995397389,-0.2428149432,-0.128882125,0.1455808133,-0.0990883484,0.384134531,-0.3911352754,-0.3713185489,-0.1201700941,-0.1128693894,-0.0630642995,-0.0857734159,-0.0269654337,0.2985586822,-0.2004606724,0.2329237461,0.0005526908,0.0394016802,0.251948297,-0.4227379262,-0.4818938971,-0.1630378664,-0.191156432,-0.0104031721,0.2656034231,-0.2878884673,0.3017160296,-0.0459037386,-0.3699751496,-0.2707997561,0.1194068417,0.1435605735,0.0721819699,-0.0227004942,-0.0912801325,0.2280841321,-0.0796872899,-0.3324358761,0.287009567,-0.2174522281,-0.1486556232,0.194851473,-0.0311844051,-0.0018445911,-0.2352986038,-0.2655209303,-0.1472442448,-0.4852220118,0.289681673,-0.0403695032,0.3086782992,0.2429099977,0.1440986991,0.1779596359,0.1556582451,-0.0546927191,-0.1509623826,-0.6362751126,0.3917452991,-0.1101357043,-0.2303782552,-0.0295004416,-0.1555970311,-0.0550973527,0.3612907529,-0.4562184811,0.237912789,0.0702732503,0.0834183842,-0.0950674713,-0.134057045,0.1914462298,-0.1075467318,-0.0603607446,-0.2590092719,0.0759437755,-0.0122350994,0.2527312934,0.1833108217,0.4013436437,0.7469967604,-0.039161507,0.2865437567,0.353391856,0.1458686143,0.5503771305,-0.1300928146,0.16025047,-0.2425827533,-0.2188410312,0.0596462861,0.1050836518,-0.3196414113,0.0274330787,0.0012185025,-0.3406588137,-0.3240846992,0.1244037524,-0.3062015474,-0.1882945746,0.2014841437,-0.0963935554,0.1964119375,-0.0394231714,-0.0296395868,-0.0032317135,0.1153508127,-0.048328694,0.2654097974,0.4285131991,0.1289328188,-0.0543236472,-0.1830666661,-0.3484928906,0.078764379,0.2743041515,0.6020610929,-0.1669999659,-0.1687550992,0.0798476189,-0.1684129536,0.4636302292,-0.1258804947,-0.0341541246,0.3472768664,-0.1242791489,-0.0266533308,-0.1190542877,-0.0431695469,0.3530095518,0.195949018,0.6723818183,-0.1772067547,-0.220200792,0.1132560149,0.0215652157,0.0907032862,-0.156267941,-0.2928073406,-0.1358119994,-0.201933071,-0.0655045658,-0.0717766806,0.2410783023,0.0199561808,-0.131281808,0.0536011457,-0.0730547979,-0.1126663312,0.0078033102,0.294038862,-0.0325312726,0.1593426615,0.057022687,0.5039331317,0.232784465,0.5136659145,0.5314756632,-0.4147776663,0.1325844377,0.0702452734,0.1329945028,0.1649885625,-0.1350931078,-0.1513888985,0.138104111,0.2574620247,-0.0355816707,-0.0832144171,0.2338197678,-0.1555819809,-0.3567377925,-0.4606517553,0.2432612181,-0.0392031707,0.1920537502,0.7787454724,-0.6046575308,-0.1680700779,0.1287034303,-0.327439338,0.6506022215,0.3056919277,-0.01261207,0.4230307937,-0.1530070156,0.4291078448,-0.4506242871,0.3880233169,-0.107545048,-0.1895088851,0.0136250248,-0.1521783918,0.3272462487,0.1736504287,-0.3301243484,0.1560810655,0.0949875191,-0.0261116754,0.056402456,0.0844660252,-0.3410910368,-0.1478554606,-0.6056392193,0.0822727308,-0.1643425077,0.3591601551,0.119864881,-0.3871607482,0.0100027071,-0.2189189792,-0.1640434861,0.4873065948,-0.2205232978,-0.0437984429,-0.1201645508,-0.6457443237,-0.3231777251,-0.1665352583,-0.3695810437,-0.1014870033,-0.0968803316,-0.0140162325,-0.1560026407,0.0815750286,-0.4244860113,0.2442319244,0.1879109144,-0.1447571814,-0.0854346454,0.1730190217,-0.2793938816,-0.55167979,0.2067956775,-0.0945816264,0.2443367243,0.1387685984,-0.014279468,0.0188988578,-0.1952127963,-0.2021060437,0.1315957457,-0.1408141851,-0.3405287564,-0.2667013407,-0.151902318,-0.200861007,-0.1044818312,0.2798337042,0.0701378211,0.3024792969,0.3966619372,-0.0176074822,-0.0633016229,-0.2849562764,0.0822029561,-0.1994450986,-0.0627179369,0.1458605379,-0.0238583386,0.1735637039,-0.2250163257,-0.0105745681,0.1574327648,0.0527799204,0.0311236531,-0.4647310376,-0.1359294653,-0.1071501747,-0.0055007907,0.1117188111,0.1044034883,0.2052558213,0.1506942809,-0.1848126501,-0.287404716,-0.1490447223,-0.1798323691,0.0388781689,0.0197014529,0.1445133686,0.3058988154,0.0417761318,0.1726244986,0.1845808923,0.1145528182,-0.2760638893,-0.1294748783,0.1017443016,0.1475505531,-0.1066455618,-0.2568067312,-0.0784032717,-0.239744544,-0.0903129727,0.1030201986,-0.0890858918,-0.0714164227,0.2204833478,0.3504038453,-0.0414797217,-0.1651158482,-0.0696929991,-0.2029707879,0.3246166408,-0.149192661,0.0957456976,0.0479332507,0.0756747648,-0.5871013403,-0.0813039467,-0.0964607447,0.1543788612,0.59524405,-0.4551479816,0.03112543,0.0455692224,0.1699540019,0.4789012969,-0.1304557621,0.0390133224,0.1729512513,0.1876669079,-0.0484990031,-0.0702733546,0.1377959102,0.1060291827,-0.0637325421,-0.0711321831,-0.1525841057,-0.1502206475,0.0998598635,0.0143192159,0.3330241144,0.0458889343,0.211354658,0.4080850184,0.2017606944,0.0718020722,-0.3809928596,0.0200543348,0.0718998685,-0.1887304932,-0.1524383426,-0.1257040054,0.0813989639,0.1215052903,-0.3927783072,-0.3188354373,-0.3502144217,0.0856014788,-0.1179064736,0.4086128771,-0.0059121046,0.2030270994,-0.2063986659,0.0338467211,0.0438848063,0.16065979,0.0185740534,0.115962036,-0.5277093649,-0.031080272,-0.0583055541,0.0389653742,0.3544371128,0.5155139565,-0.0361992531,0.2214543521,-0.1061519906,-0.1359117925,0.0883783028,0.1771474481,0.3425745368,-0.3739395738,0.2876466513,0.2135393769,0.0820103437,-0.0754379928,0.0099544805,0.3729601204,0.3608309031,0.0665017441,0.0278094839,0.0363483615,-0.0933402553,0.0266610924,0.335478574,0.1904851943,-0.1879787147,0.2221895158,0.1101663113,-0.088712424,0.3944174051,0.0983025655,0.1196607053,-0.1152643189,0.1943055987,-0.247738108,-0.1086110324,0.0640453398,-0.354144752,-0.358769089,-0.0634774119,0.2332483977,0.1124899834,0.0440791547,-0.0250769686,0.0849539489,-0.1289328188,0.279007405,0.2456029654,-0.0799998939,-0.1668445021,-0.1844663471,-0.2015265524,0.1580558419,-0.040410243,-0.1079189107,0.2371376604,0.3279328048,-0.0682448745,0.2060449421,-0.0900718644,-0.1822002977,-0.2946017087,0.2807135284,-0.3546518683,-0.2369975001,0.1787969619,0.2335970402,0.0910652205,-0.4543965757,0.2050732076,-0.3312031925,0.1178356186,0.0553922281,-0.0305205919,0.0156790111,0.0668511316,0.2356533855,0.1621074677,0.1155462116,-0.1349952668,-0.192071259,-0.072820738,-0.176847145,-0.0183110926,0.1458830535,-0.0582501702,0.502874732,-0.0484356917,-0.4014241993,-0.4371389151,0.4285812378,-0.0439626239,-0.4439990819,0.0103937835,0.0140104787,-0.1760369986,0.1818829775,0.0124762263,0.1777729988,-0.0506044216,0.0616825111,-0.3085522056,-0.1021585092,0.5942243338,-0.4224966168,-0.3192535639,0.3241405487,0.1266875267,0.1190700009,0.1298958212,-0.1965908855,0.1324760169,0.437204659,-0.0283873919,0.1215384007,0.1597634852,0.1152027473,0.1851295084,0.1178187653,0.0776488185,0.1725968719,-0.259760052,0.1329277754,-0.214050144]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2607","title":"Streaming local gzip compressed JSON line files is not working","comments":"Hi @thomwolf, thanks for reporting.\r\n\r\nIt seems this might be due to the fact that the JSON Dataset builder uses `pyarrow.json` (`paj.read_json`) to read the data without using the Python standard `open(file,...` (which is the one patched with `xopen` to work in streaming mode).\r\n\r\nThis has to be fixed.","body":"## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0","comment_length":49,"text":"Streaming local gzip compressed JSON line files is not working\n## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0\nHi @thomwolf, thanks for reporting.\r\n\r\nIt seems this might be due to the fact that the JSON Dataset builder uses `pyarrow.json` (`paj.read_json`) to read the data without using the Python standard `open(file,...` (which is the one patched with `xopen` to work in streaming mode).\r\n\r\nThis has to be fixed.","embeddings":[-0.1992494613,-0.1339781284,-0.0091445101,0.2391350567,0.0777640566,0.0168800913,0.3976738453,0.5178918839,0.1970213056,0.0789984018,0.1030013189,0.3306827545,-0.0883667096,0.0208737329,0.2288863659,-0.1809338927,-0.0436617956,0.3560613692,-0.0251667686,0.0015848613,-0.1305664033,0.2937984169,-0.1404591352,-0.0185168069,-0.0187981017,-0.0395553522,-0.0471503064,0.1500941366,-0.0939593539,-0.3326378763,0.0627696067,-0.1684151441,0.188022688,0.4454552829,-0.000114536,0.288641125,0.2961105406,-0.1629139781,-0.3886071146,-0.4712482095,-0.229901135,-0.1492993236,-0.1307716817,-0.2858453691,-0.1378189325,-0.3474362791,-0.0337565579,-0.5071999431,0.5416871309,0.2922603488,0.1556637138,0.0419651642,-0.061772272,0.0905790925,0.0390155055,0.1573239863,-0.0886082724,0.3150972426,0.0327884369,0.1149678901,0.0552849993,0.4286454916,-0.2254960537,-0.1747770905,0.1473699361,-0.0008157574,-0.0843830332,-0.1257867962,-0.1683152765,0.085735105,0.1845451146,-0.4122557342,-0.4398344159,-0.3761076331,0.0466741547,-0.44880265,0.2118544728,0.4092858434,-0.250308305,0.1969692856,-0.100665614,-0.130531922,-0.3545617759,0.0854526162,0.36184901,0.3480824232,-0.0946776345,0.0091285398,-0.0919867903,0.1193795577,-0.0839845538,-0.3859618604,-0.1857065707,0.227756381,-0.2421413958,-0.1349226832,0.1443624049,-0.6062148213,0.1511498541,0.1012284085,0.4615779221,-0.1257445961,0.3588074744,0.0546084084,0.2728678286,0.1870484799,0.1769655049,0.2638421357,0.0566591471,0.3898805082,-0.0865667164,-0.1980740726,-0.2363374233,-0.0394049771,-0.0915300399,0.1368107945,0.4833262861,-0.1576653421,-0.089093484,0.081910789,-0.5941808224,-0.1327005029,0.1576721966,0.0986096784,0.0661023408,0.2201749086,0.026484726,0.1610841751,-0.002104495,-0.0618949644,-0.0907956734,0.0235721208,0.0740586743,0.1347989291,0.251897037,-0.5213885903,0.2793646455,0.0087546054,-0.046029456,-0.1562338322,0.0058455556,-0.12555632,0.1504514217,0.146838814,-0.0259293783,0.0193639155,0.2058843374,-0.1102267727,-0.1738193929,0.1183539331,-0.2751955688,-0.2672133446,0.0372359268,0.2010167539,0.0381590351,0.009389149,-0.0081491908,0.0059310379,0.0556991696,-0.3565125763,0.023636166,-0.0094647072,-0.0558016375,-0.0841292068,0.2759922445,0.4360964596,-0.3608719409,-0.1035921797,0.0715648532,0.2175855786,0.7216550708,0.3425383568,-0.2379719913,0.1273078024,-0.3681203723,0.3619052172,0.6931351423,0.0160383359,-0.3624646366,0.8522378206,-0.1086349562,0.3817502558,0.4800842702,-0.20442909,0.1941484958,-0.0105180815,0.3352073431,0.2812868953,-0.0445720889,0.0385273471,-0.464271307,0.1312143207,0.0419497304,0.0378056802,-0.2513179779,0.2954408824,-0.0433923602,0.015032988,0.3520895541,-0.1800565869,0.1081850231,0.2904555202,0.2378064245,0.0199463479,0.2021734118,0.0092641879,-0.2178035229,0.0324927419,-0.3366108537,-0.2629368603,-0.2826167047,-0.0624102913,-0.1401127279,0.1487002075,-0.1843477637,-0.3533777595,0.1464820653,0.4111509025,0.3808980882,0.0224370789,-0.4623301327,0.3006553948,-0.2987254858,-0.1048559621,-0.2611705363,0.429363668,0.0721354857,-0.2631469369,0.1758847386,-0.0589892119,0.1717893928,-0.2416345477,-0.2221167684,0.2450815886,-0.1208557263,0.1709071249,0.1169753596,-0.2226256579,0.2957748175,-0.0062111109,-0.0720564499,0.3445874751,0.2193278074,0.2041208595,-0.2239909172,0.4739874303,-0.0305651519,0.3029977679,0.0209108982,-0.205836162,0.1924333572,-0.0084315231,-0.3421524465,0.0653704852,0.3345847428,-0.1258885711,0.0946051851,-0.0612957813,-0.3791344762,-0.0635569915,0.5271067023,-0.0830585435,-0.1615021825,0.1926877648,-0.1803122163,-0.0555438772,0.2965638041,-0.0067166113,0.3335767984,0.1575452387,0.104602471,0.0166130755,-0.1111166552,-0.2565548122,0.2091605067,-0.0900344998,0.1146262661,0.5449168086,0.0356366746,0.0449167937,-0.3995397389,-0.2428149432,-0.128882125,0.1455808133,-0.0990883484,0.384134531,-0.3911352754,-0.3713185489,-0.1201700941,-0.1128693894,-0.0630642995,-0.0857734159,-0.0269654337,0.2985586822,-0.2004606724,0.2329237461,0.0005526908,0.0394016802,0.251948297,-0.4227379262,-0.4818938971,-0.1630378664,-0.191156432,-0.0104031721,0.2656034231,-0.2878884673,0.3017160296,-0.0459037386,-0.3699751496,-0.2707997561,0.1194068417,0.1435605735,0.0721819699,-0.0227004942,-0.0912801325,0.2280841321,-0.0796872899,-0.3324358761,0.287009567,-0.2174522281,-0.1486556232,0.194851473,-0.0311844051,-0.0018445911,-0.2352986038,-0.2655209303,-0.1472442448,-0.4852220118,0.289681673,-0.0403695032,0.3086782992,0.2429099977,0.1440986991,0.1779596359,0.1556582451,-0.0546927191,-0.1509623826,-0.6362751126,0.3917452991,-0.1101357043,-0.2303782552,-0.0295004416,-0.1555970311,-0.0550973527,0.3612907529,-0.4562184811,0.237912789,0.0702732503,0.0834183842,-0.0950674713,-0.134057045,0.1914462298,-0.1075467318,-0.0603607446,-0.2590092719,0.0759437755,-0.0122350994,0.2527312934,0.1833108217,0.4013436437,0.7469967604,-0.039161507,0.2865437567,0.353391856,0.1458686143,0.5503771305,-0.1300928146,0.16025047,-0.2425827533,-0.2188410312,0.0596462861,0.1050836518,-0.3196414113,0.0274330787,0.0012185025,-0.3406588137,-0.3240846992,0.1244037524,-0.3062015474,-0.1882945746,0.2014841437,-0.0963935554,0.1964119375,-0.0394231714,-0.0296395868,-0.0032317135,0.1153508127,-0.048328694,0.2654097974,0.4285131991,0.1289328188,-0.0543236472,-0.1830666661,-0.3484928906,0.078764379,0.2743041515,0.6020610929,-0.1669999659,-0.1687550992,0.0798476189,-0.1684129536,0.4636302292,-0.1258804947,-0.0341541246,0.3472768664,-0.1242791489,-0.0266533308,-0.1190542877,-0.0431695469,0.3530095518,0.195949018,0.6723818183,-0.1772067547,-0.220200792,0.1132560149,0.0215652157,0.0907032862,-0.156267941,-0.2928073406,-0.1358119994,-0.201933071,-0.0655045658,-0.0717766806,0.2410783023,0.0199561808,-0.131281808,0.0536011457,-0.0730547979,-0.1126663312,0.0078033102,0.294038862,-0.0325312726,0.1593426615,0.057022687,0.5039331317,0.232784465,0.5136659145,0.5314756632,-0.4147776663,0.1325844377,0.0702452734,0.1329945028,0.1649885625,-0.1350931078,-0.1513888985,0.138104111,0.2574620247,-0.0355816707,-0.0832144171,0.2338197678,-0.1555819809,-0.3567377925,-0.4606517553,0.2432612181,-0.0392031707,0.1920537502,0.7787454724,-0.6046575308,-0.1680700779,0.1287034303,-0.327439338,0.6506022215,0.3056919277,-0.01261207,0.4230307937,-0.1530070156,0.4291078448,-0.4506242871,0.3880233169,-0.107545048,-0.1895088851,0.0136250248,-0.1521783918,0.3272462487,0.1736504287,-0.3301243484,0.1560810655,0.0949875191,-0.0261116754,0.056402456,0.0844660252,-0.3410910368,-0.1478554606,-0.6056392193,0.0822727308,-0.1643425077,0.3591601551,0.119864881,-0.3871607482,0.0100027071,-0.2189189792,-0.1640434861,0.4873065948,-0.2205232978,-0.0437984429,-0.1201645508,-0.6457443237,-0.3231777251,-0.1665352583,-0.3695810437,-0.1014870033,-0.0968803316,-0.0140162325,-0.1560026407,0.0815750286,-0.4244860113,0.2442319244,0.1879109144,-0.1447571814,-0.0854346454,0.1730190217,-0.2793938816,-0.55167979,0.2067956775,-0.0945816264,0.2443367243,0.1387685984,-0.014279468,0.0188988578,-0.1952127963,-0.2021060437,0.1315957457,-0.1408141851,-0.3405287564,-0.2667013407,-0.151902318,-0.200861007,-0.1044818312,0.2798337042,0.0701378211,0.3024792969,0.3966619372,-0.0176074822,-0.0633016229,-0.2849562764,0.0822029561,-0.1994450986,-0.0627179369,0.1458605379,-0.0238583386,0.1735637039,-0.2250163257,-0.0105745681,0.1574327648,0.0527799204,0.0311236531,-0.4647310376,-0.1359294653,-0.1071501747,-0.0055007907,0.1117188111,0.1044034883,0.2052558213,0.1506942809,-0.1848126501,-0.287404716,-0.1490447223,-0.1798323691,0.0388781689,0.0197014529,0.1445133686,0.3058988154,0.0417761318,0.1726244986,0.1845808923,0.1145528182,-0.2760638893,-0.1294748783,0.1017443016,0.1475505531,-0.1066455618,-0.2568067312,-0.0784032717,-0.239744544,-0.0903129727,0.1030201986,-0.0890858918,-0.0714164227,0.2204833478,0.3504038453,-0.0414797217,-0.1651158482,-0.0696929991,-0.2029707879,0.3246166408,-0.149192661,0.0957456976,0.0479332507,0.0756747648,-0.5871013403,-0.0813039467,-0.0964607447,0.1543788612,0.59524405,-0.4551479816,0.03112543,0.0455692224,0.1699540019,0.4789012969,-0.1304557621,0.0390133224,0.1729512513,0.1876669079,-0.0484990031,-0.0702733546,0.1377959102,0.1060291827,-0.0637325421,-0.0711321831,-0.1525841057,-0.1502206475,0.0998598635,0.0143192159,0.3330241144,0.0458889343,0.211354658,0.4080850184,0.2017606944,0.0718020722,-0.3809928596,0.0200543348,0.0718998685,-0.1887304932,-0.1524383426,-0.1257040054,0.0813989639,0.1215052903,-0.3927783072,-0.3188354373,-0.3502144217,0.0856014788,-0.1179064736,0.4086128771,-0.0059121046,0.2030270994,-0.2063986659,0.0338467211,0.0438848063,0.16065979,0.0185740534,0.115962036,-0.5277093649,-0.031080272,-0.0583055541,0.0389653742,0.3544371128,0.5155139565,-0.0361992531,0.2214543521,-0.1061519906,-0.1359117925,0.0883783028,0.1771474481,0.3425745368,-0.3739395738,0.2876466513,0.2135393769,0.0820103437,-0.0754379928,0.0099544805,0.3729601204,0.3608309031,0.0665017441,0.0278094839,0.0363483615,-0.0933402553,0.0266610924,0.335478574,0.1904851943,-0.1879787147,0.2221895158,0.1101663113,-0.088712424,0.3944174051,0.0983025655,0.1196607053,-0.1152643189,0.1943055987,-0.247738108,-0.1086110324,0.0640453398,-0.354144752,-0.358769089,-0.0634774119,0.2332483977,0.1124899834,0.0440791547,-0.0250769686,0.0849539489,-0.1289328188,0.279007405,0.2456029654,-0.0799998939,-0.1668445021,-0.1844663471,-0.2015265524,0.1580558419,-0.040410243,-0.1079189107,0.2371376604,0.3279328048,-0.0682448745,0.2060449421,-0.0900718644,-0.1822002977,-0.2946017087,0.2807135284,-0.3546518683,-0.2369975001,0.1787969619,0.2335970402,0.0910652205,-0.4543965757,0.2050732076,-0.3312031925,0.1178356186,0.0553922281,-0.0305205919,0.0156790111,0.0668511316,0.2356533855,0.1621074677,0.1155462116,-0.1349952668,-0.192071259,-0.072820738,-0.176847145,-0.0183110926,0.1458830535,-0.0582501702,0.502874732,-0.0484356917,-0.4014241993,-0.4371389151,0.4285812378,-0.0439626239,-0.4439990819,0.0103937835,0.0140104787,-0.1760369986,0.1818829775,0.0124762263,0.1777729988,-0.0506044216,0.0616825111,-0.3085522056,-0.1021585092,0.5942243338,-0.4224966168,-0.3192535639,0.3241405487,0.1266875267,0.1190700009,0.1298958212,-0.1965908855,0.1324760169,0.437204659,-0.0283873919,0.1215384007,0.1597634852,0.1152027473,0.1851295084,0.1178187653,0.0776488185,0.1725968719,-0.259760052,0.1329277754,-0.214050144]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2607","title":"Streaming local gzip compressed JSON line files is not working","comments":"Sorry for reopening this, but I'm having the same issue as @thomwolf when streaming a gzipped JSON Lines file from the hub. Or is that just not possible by definition?\r\nI installed `datasets`in editable mode from source (so probably includes the fix from #2608 ?): \r\n```\r\n>>> datasets.__version__\r\n'1.9.1.dev0'\r\n```\r\n\r\n```\r\n>>> msmarco = datasets.load_dataset(\"webis\/msmarco\", \"corpus\", streaming=True)\r\nUsing custom data configuration corpus-174d3b7155eb68db\r\n>>> msmarco_iter = iter(msmarco['train'])\r\n>>> print(next(msmarco_iter))\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/media\/ssd\/TREC\/msmarco\/datasets\/src\/datasets\/iterable_dataset.py\", line 338, in __iter__\r\n for key, example in self._iter():\r\n File \"\/media\/ssd\/TREC\/msmarco\/datasets\/src\/datasets\/iterable_dataset.py\", line 335, in _iter\r\n yield from ex_iterable\r\n File \"\/media\/ssd\/TREC\/msmarco\/datasets\/src\/datasets\/iterable_dataset.py\", line 78, in __iter__\r\n for key, example in self.generate_examples_fn(**self.kwargs):\r\n File \"\/home\/christopher\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/msmarco\/eb63dff8d83107168e973c7a655a6082d37e08d71b4ac39a0afada479c138745\/msmarco.py\", line 96, in _generate_examples\r\n with gzip.open(file, \"rt\", encoding=\"utf-8\") as f:\r\n File \"\/usr\/lib\/python3.6\/gzip.py\", line 53, in open\r\n binary_file = GzipFile(filename, gz_mode, compresslevel)\r\n File \"\/usr\/lib\/python3.6\/gzip.py\", line 163, in __init__\r\n fileobj = self.myfileobj = builtins.open(filename, mode or 'rb')\r\nFileNotFoundError: [Errno 2] No such file or directory: 'https:\/\/huggingface.co\/datasets\/webis\/msmarco\/resolve\/main\/msmarco_doc_00.gz'\r\n```\r\n\r\nLoading the dataset without streaming set to True, works fine.","body":"## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0","comment_length":167,"text":"Streaming local gzip compressed JSON line files is not working\n## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0\nSorry for reopening this, but I'm having the same issue as @thomwolf when streaming a gzipped JSON Lines file from the hub. Or is that just not possible by definition?\r\nI installed `datasets`in editable mode from source (so probably includes the fix from #2608 ?): \r\n```\r\n>>> datasets.__version__\r\n'1.9.1.dev0'\r\n```\r\n\r\n```\r\n>>> msmarco = datasets.load_dataset(\"webis\/msmarco\", \"corpus\", streaming=True)\r\nUsing custom data configuration corpus-174d3b7155eb68db\r\n>>> msmarco_iter = iter(msmarco['train'])\r\n>>> print(next(msmarco_iter))\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/media\/ssd\/TREC\/msmarco\/datasets\/src\/datasets\/iterable_dataset.py\", line 338, in __iter__\r\n for key, example in self._iter():\r\n File \"\/media\/ssd\/TREC\/msmarco\/datasets\/src\/datasets\/iterable_dataset.py\", line 335, in _iter\r\n yield from ex_iterable\r\n File \"\/media\/ssd\/TREC\/msmarco\/datasets\/src\/datasets\/iterable_dataset.py\", line 78, in __iter__\r\n for key, example in self.generate_examples_fn(**self.kwargs):\r\n File \"\/home\/christopher\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/msmarco\/eb63dff8d83107168e973c7a655a6082d37e08d71b4ac39a0afada479c138745\/msmarco.py\", line 96, in _generate_examples\r\n with gzip.open(file, \"rt\", encoding=\"utf-8\") as f:\r\n File \"\/usr\/lib\/python3.6\/gzip.py\", line 53, in open\r\n binary_file = GzipFile(filename, gz_mode, compresslevel)\r\n File \"\/usr\/lib\/python3.6\/gzip.py\", line 163, in __init__\r\n fileobj = self.myfileobj = builtins.open(filename, mode or 'rb')\r\nFileNotFoundError: [Errno 2] No such file or directory: 'https:\/\/huggingface.co\/datasets\/webis\/msmarco\/resolve\/main\/msmarco_doc_00.gz'\r\n```\r\n\r\nLoading the dataset without streaming set to True, works fine.","embeddings":[-0.1992494613,-0.1339781284,-0.0091445101,0.2391350567,0.0777640566,0.0168800913,0.3976738453,0.5178918839,0.1970213056,0.0789984018,0.1030013189,0.3306827545,-0.0883667096,0.0208737329,0.2288863659,-0.1809338927,-0.0436617956,0.3560613692,-0.0251667686,0.0015848613,-0.1305664033,0.2937984169,-0.1404591352,-0.0185168069,-0.0187981017,-0.0395553522,-0.0471503064,0.1500941366,-0.0939593539,-0.3326378763,0.0627696067,-0.1684151441,0.188022688,0.4454552829,-0.000114536,0.288641125,0.2961105406,-0.1629139781,-0.3886071146,-0.4712482095,-0.229901135,-0.1492993236,-0.1307716817,-0.2858453691,-0.1378189325,-0.3474362791,-0.0337565579,-0.5071999431,0.5416871309,0.2922603488,0.1556637138,0.0419651642,-0.061772272,0.0905790925,0.0390155055,0.1573239863,-0.0886082724,0.3150972426,0.0327884369,0.1149678901,0.0552849993,0.4286454916,-0.2254960537,-0.1747770905,0.1473699361,-0.0008157574,-0.0843830332,-0.1257867962,-0.1683152765,0.085735105,0.1845451146,-0.4122557342,-0.4398344159,-0.3761076331,0.0466741547,-0.44880265,0.2118544728,0.4092858434,-0.250308305,0.1969692856,-0.100665614,-0.130531922,-0.3545617759,0.0854526162,0.36184901,0.3480824232,-0.0946776345,0.0091285398,-0.0919867903,0.1193795577,-0.0839845538,-0.3859618604,-0.1857065707,0.227756381,-0.2421413958,-0.1349226832,0.1443624049,-0.6062148213,0.1511498541,0.1012284085,0.4615779221,-0.1257445961,0.3588074744,0.0546084084,0.2728678286,0.1870484799,0.1769655049,0.2638421357,0.0566591471,0.3898805082,-0.0865667164,-0.1980740726,-0.2363374233,-0.0394049771,-0.0915300399,0.1368107945,0.4833262861,-0.1576653421,-0.089093484,0.081910789,-0.5941808224,-0.1327005029,0.1576721966,0.0986096784,0.0661023408,0.2201749086,0.026484726,0.1610841751,-0.002104495,-0.0618949644,-0.0907956734,0.0235721208,0.0740586743,0.1347989291,0.251897037,-0.5213885903,0.2793646455,0.0087546054,-0.046029456,-0.1562338322,0.0058455556,-0.12555632,0.1504514217,0.146838814,-0.0259293783,0.0193639155,0.2058843374,-0.1102267727,-0.1738193929,0.1183539331,-0.2751955688,-0.2672133446,0.0372359268,0.2010167539,0.0381590351,0.009389149,-0.0081491908,0.0059310379,0.0556991696,-0.3565125763,0.023636166,-0.0094647072,-0.0558016375,-0.0841292068,0.2759922445,0.4360964596,-0.3608719409,-0.1035921797,0.0715648532,0.2175855786,0.7216550708,0.3425383568,-0.2379719913,0.1273078024,-0.3681203723,0.3619052172,0.6931351423,0.0160383359,-0.3624646366,0.8522378206,-0.1086349562,0.3817502558,0.4800842702,-0.20442909,0.1941484958,-0.0105180815,0.3352073431,0.2812868953,-0.0445720889,0.0385273471,-0.464271307,0.1312143207,0.0419497304,0.0378056802,-0.2513179779,0.2954408824,-0.0433923602,0.015032988,0.3520895541,-0.1800565869,0.1081850231,0.2904555202,0.2378064245,0.0199463479,0.2021734118,0.0092641879,-0.2178035229,0.0324927419,-0.3366108537,-0.2629368603,-0.2826167047,-0.0624102913,-0.1401127279,0.1487002075,-0.1843477637,-0.3533777595,0.1464820653,0.4111509025,0.3808980882,0.0224370789,-0.4623301327,0.3006553948,-0.2987254858,-0.1048559621,-0.2611705363,0.429363668,0.0721354857,-0.2631469369,0.1758847386,-0.0589892119,0.1717893928,-0.2416345477,-0.2221167684,0.2450815886,-0.1208557263,0.1709071249,0.1169753596,-0.2226256579,0.2957748175,-0.0062111109,-0.0720564499,0.3445874751,0.2193278074,0.2041208595,-0.2239909172,0.4739874303,-0.0305651519,0.3029977679,0.0209108982,-0.205836162,0.1924333572,-0.0084315231,-0.3421524465,0.0653704852,0.3345847428,-0.1258885711,0.0946051851,-0.0612957813,-0.3791344762,-0.0635569915,0.5271067023,-0.0830585435,-0.1615021825,0.1926877648,-0.1803122163,-0.0555438772,0.2965638041,-0.0067166113,0.3335767984,0.1575452387,0.104602471,0.0166130755,-0.1111166552,-0.2565548122,0.2091605067,-0.0900344998,0.1146262661,0.5449168086,0.0356366746,0.0449167937,-0.3995397389,-0.2428149432,-0.128882125,0.1455808133,-0.0990883484,0.384134531,-0.3911352754,-0.3713185489,-0.1201700941,-0.1128693894,-0.0630642995,-0.0857734159,-0.0269654337,0.2985586822,-0.2004606724,0.2329237461,0.0005526908,0.0394016802,0.251948297,-0.4227379262,-0.4818938971,-0.1630378664,-0.191156432,-0.0104031721,0.2656034231,-0.2878884673,0.3017160296,-0.0459037386,-0.3699751496,-0.2707997561,0.1194068417,0.1435605735,0.0721819699,-0.0227004942,-0.0912801325,0.2280841321,-0.0796872899,-0.3324358761,0.287009567,-0.2174522281,-0.1486556232,0.194851473,-0.0311844051,-0.0018445911,-0.2352986038,-0.2655209303,-0.1472442448,-0.4852220118,0.289681673,-0.0403695032,0.3086782992,0.2429099977,0.1440986991,0.1779596359,0.1556582451,-0.0546927191,-0.1509623826,-0.6362751126,0.3917452991,-0.1101357043,-0.2303782552,-0.0295004416,-0.1555970311,-0.0550973527,0.3612907529,-0.4562184811,0.237912789,0.0702732503,0.0834183842,-0.0950674713,-0.134057045,0.1914462298,-0.1075467318,-0.0603607446,-0.2590092719,0.0759437755,-0.0122350994,0.2527312934,0.1833108217,0.4013436437,0.7469967604,-0.039161507,0.2865437567,0.353391856,0.1458686143,0.5503771305,-0.1300928146,0.16025047,-0.2425827533,-0.2188410312,0.0596462861,0.1050836518,-0.3196414113,0.0274330787,0.0012185025,-0.3406588137,-0.3240846992,0.1244037524,-0.3062015474,-0.1882945746,0.2014841437,-0.0963935554,0.1964119375,-0.0394231714,-0.0296395868,-0.0032317135,0.1153508127,-0.048328694,0.2654097974,0.4285131991,0.1289328188,-0.0543236472,-0.1830666661,-0.3484928906,0.078764379,0.2743041515,0.6020610929,-0.1669999659,-0.1687550992,0.0798476189,-0.1684129536,0.4636302292,-0.1258804947,-0.0341541246,0.3472768664,-0.1242791489,-0.0266533308,-0.1190542877,-0.0431695469,0.3530095518,0.195949018,0.6723818183,-0.1772067547,-0.220200792,0.1132560149,0.0215652157,0.0907032862,-0.156267941,-0.2928073406,-0.1358119994,-0.201933071,-0.0655045658,-0.0717766806,0.2410783023,0.0199561808,-0.131281808,0.0536011457,-0.0730547979,-0.1126663312,0.0078033102,0.294038862,-0.0325312726,0.1593426615,0.057022687,0.5039331317,0.232784465,0.5136659145,0.5314756632,-0.4147776663,0.1325844377,0.0702452734,0.1329945028,0.1649885625,-0.1350931078,-0.1513888985,0.138104111,0.2574620247,-0.0355816707,-0.0832144171,0.2338197678,-0.1555819809,-0.3567377925,-0.4606517553,0.2432612181,-0.0392031707,0.1920537502,0.7787454724,-0.6046575308,-0.1680700779,0.1287034303,-0.327439338,0.6506022215,0.3056919277,-0.01261207,0.4230307937,-0.1530070156,0.4291078448,-0.4506242871,0.3880233169,-0.107545048,-0.1895088851,0.0136250248,-0.1521783918,0.3272462487,0.1736504287,-0.3301243484,0.1560810655,0.0949875191,-0.0261116754,0.056402456,0.0844660252,-0.3410910368,-0.1478554606,-0.6056392193,0.0822727308,-0.1643425077,0.3591601551,0.119864881,-0.3871607482,0.0100027071,-0.2189189792,-0.1640434861,0.4873065948,-0.2205232978,-0.0437984429,-0.1201645508,-0.6457443237,-0.3231777251,-0.1665352583,-0.3695810437,-0.1014870033,-0.0968803316,-0.0140162325,-0.1560026407,0.0815750286,-0.4244860113,0.2442319244,0.1879109144,-0.1447571814,-0.0854346454,0.1730190217,-0.2793938816,-0.55167979,0.2067956775,-0.0945816264,0.2443367243,0.1387685984,-0.014279468,0.0188988578,-0.1952127963,-0.2021060437,0.1315957457,-0.1408141851,-0.3405287564,-0.2667013407,-0.151902318,-0.200861007,-0.1044818312,0.2798337042,0.0701378211,0.3024792969,0.3966619372,-0.0176074822,-0.0633016229,-0.2849562764,0.0822029561,-0.1994450986,-0.0627179369,0.1458605379,-0.0238583386,0.1735637039,-0.2250163257,-0.0105745681,0.1574327648,0.0527799204,0.0311236531,-0.4647310376,-0.1359294653,-0.1071501747,-0.0055007907,0.1117188111,0.1044034883,0.2052558213,0.1506942809,-0.1848126501,-0.287404716,-0.1490447223,-0.1798323691,0.0388781689,0.0197014529,0.1445133686,0.3058988154,0.0417761318,0.1726244986,0.1845808923,0.1145528182,-0.2760638893,-0.1294748783,0.1017443016,0.1475505531,-0.1066455618,-0.2568067312,-0.0784032717,-0.239744544,-0.0903129727,0.1030201986,-0.0890858918,-0.0714164227,0.2204833478,0.3504038453,-0.0414797217,-0.1651158482,-0.0696929991,-0.2029707879,0.3246166408,-0.149192661,0.0957456976,0.0479332507,0.0756747648,-0.5871013403,-0.0813039467,-0.0964607447,0.1543788612,0.59524405,-0.4551479816,0.03112543,0.0455692224,0.1699540019,0.4789012969,-0.1304557621,0.0390133224,0.1729512513,0.1876669079,-0.0484990031,-0.0702733546,0.1377959102,0.1060291827,-0.0637325421,-0.0711321831,-0.1525841057,-0.1502206475,0.0998598635,0.0143192159,0.3330241144,0.0458889343,0.211354658,0.4080850184,0.2017606944,0.0718020722,-0.3809928596,0.0200543348,0.0718998685,-0.1887304932,-0.1524383426,-0.1257040054,0.0813989639,0.1215052903,-0.3927783072,-0.3188354373,-0.3502144217,0.0856014788,-0.1179064736,0.4086128771,-0.0059121046,0.2030270994,-0.2063986659,0.0338467211,0.0438848063,0.16065979,0.0185740534,0.115962036,-0.5277093649,-0.031080272,-0.0583055541,0.0389653742,0.3544371128,0.5155139565,-0.0361992531,0.2214543521,-0.1061519906,-0.1359117925,0.0883783028,0.1771474481,0.3425745368,-0.3739395738,0.2876466513,0.2135393769,0.0820103437,-0.0754379928,0.0099544805,0.3729601204,0.3608309031,0.0665017441,0.0278094839,0.0363483615,-0.0933402553,0.0266610924,0.335478574,0.1904851943,-0.1879787147,0.2221895158,0.1101663113,-0.088712424,0.3944174051,0.0983025655,0.1196607053,-0.1152643189,0.1943055987,-0.247738108,-0.1086110324,0.0640453398,-0.354144752,-0.358769089,-0.0634774119,0.2332483977,0.1124899834,0.0440791547,-0.0250769686,0.0849539489,-0.1289328188,0.279007405,0.2456029654,-0.0799998939,-0.1668445021,-0.1844663471,-0.2015265524,0.1580558419,-0.040410243,-0.1079189107,0.2371376604,0.3279328048,-0.0682448745,0.2060449421,-0.0900718644,-0.1822002977,-0.2946017087,0.2807135284,-0.3546518683,-0.2369975001,0.1787969619,0.2335970402,0.0910652205,-0.4543965757,0.2050732076,-0.3312031925,0.1178356186,0.0553922281,-0.0305205919,0.0156790111,0.0668511316,0.2356533855,0.1621074677,0.1155462116,-0.1349952668,-0.192071259,-0.072820738,-0.176847145,-0.0183110926,0.1458830535,-0.0582501702,0.502874732,-0.0484356917,-0.4014241993,-0.4371389151,0.4285812378,-0.0439626239,-0.4439990819,0.0103937835,0.0140104787,-0.1760369986,0.1818829775,0.0124762263,0.1777729988,-0.0506044216,0.0616825111,-0.3085522056,-0.1021585092,0.5942243338,-0.4224966168,-0.3192535639,0.3241405487,0.1266875267,0.1190700009,0.1298958212,-0.1965908855,0.1324760169,0.437204659,-0.0283873919,0.1215384007,0.1597634852,0.1152027473,0.1851295084,0.1178187653,0.0776488185,0.1725968719,-0.259760052,0.1329277754,-0.214050144]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2607","title":"Streaming local gzip compressed JSON line files is not working","comments":"Hi ! To make the streaming work, we extend `open` in the dataset builder to work with urls.\r\n\r\nTherefore you just need to use `open` before using `gzip.open`:\r\n```diff\r\n- with gzip.open(file, \"rt\", encoding=\"utf-8\") as f:\r\n+ with gzip.open(open(file, \"rb\"), \"rt\", encoding=\"utf-8\") as f:\r\n```\r\n\r\nYou can see that it is the case for oscar.py and c4.py for example:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8814b393984c1c2e1800ba370de2a9f7c8644908\/datasets\/oscar\/oscar.py#L358-L358\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8814b393984c1c2e1800ba370de2a9f7c8644908\/datasets\/c4\/c4.py#L88-L88\r\n\r\n","body":"## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0","comment_length":61,"text":"Streaming local gzip compressed JSON line files is not working\n## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0\nHi ! To make the streaming work, we extend `open` in the dataset builder to work with urls.\r\n\r\nTherefore you just need to use `open` before using `gzip.open`:\r\n```diff\r\n- with gzip.open(file, \"rt\", encoding=\"utf-8\") as f:\r\n+ with gzip.open(open(file, \"rb\"), \"rt\", encoding=\"utf-8\") as f:\r\n```\r\n\r\nYou can see that it is the case for oscar.py and c4.py for example:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8814b393984c1c2e1800ba370de2a9f7c8644908\/datasets\/oscar\/oscar.py#L358-L358\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8814b393984c1c2e1800ba370de2a9f7c8644908\/datasets\/c4\/c4.py#L88-L88\r\n\r\n","embeddings":[-0.1992494613,-0.1339781284,-0.0091445101,0.2391350567,0.0777640566,0.0168800913,0.3976738453,0.5178918839,0.1970213056,0.0789984018,0.1030013189,0.3306827545,-0.0883667096,0.0208737329,0.2288863659,-0.1809338927,-0.0436617956,0.3560613692,-0.0251667686,0.0015848613,-0.1305664033,0.2937984169,-0.1404591352,-0.0185168069,-0.0187981017,-0.0395553522,-0.0471503064,0.1500941366,-0.0939593539,-0.3326378763,0.0627696067,-0.1684151441,0.188022688,0.4454552829,-0.000114536,0.288641125,0.2961105406,-0.1629139781,-0.3886071146,-0.4712482095,-0.229901135,-0.1492993236,-0.1307716817,-0.2858453691,-0.1378189325,-0.3474362791,-0.0337565579,-0.5071999431,0.5416871309,0.2922603488,0.1556637138,0.0419651642,-0.061772272,0.0905790925,0.0390155055,0.1573239863,-0.0886082724,0.3150972426,0.0327884369,0.1149678901,0.0552849993,0.4286454916,-0.2254960537,-0.1747770905,0.1473699361,-0.0008157574,-0.0843830332,-0.1257867962,-0.1683152765,0.085735105,0.1845451146,-0.4122557342,-0.4398344159,-0.3761076331,0.0466741547,-0.44880265,0.2118544728,0.4092858434,-0.250308305,0.1969692856,-0.100665614,-0.130531922,-0.3545617759,0.0854526162,0.36184901,0.3480824232,-0.0946776345,0.0091285398,-0.0919867903,0.1193795577,-0.0839845538,-0.3859618604,-0.1857065707,0.227756381,-0.2421413958,-0.1349226832,0.1443624049,-0.6062148213,0.1511498541,0.1012284085,0.4615779221,-0.1257445961,0.3588074744,0.0546084084,0.2728678286,0.1870484799,0.1769655049,0.2638421357,0.0566591471,0.3898805082,-0.0865667164,-0.1980740726,-0.2363374233,-0.0394049771,-0.0915300399,0.1368107945,0.4833262861,-0.1576653421,-0.089093484,0.081910789,-0.5941808224,-0.1327005029,0.1576721966,0.0986096784,0.0661023408,0.2201749086,0.026484726,0.1610841751,-0.002104495,-0.0618949644,-0.0907956734,0.0235721208,0.0740586743,0.1347989291,0.251897037,-0.5213885903,0.2793646455,0.0087546054,-0.046029456,-0.1562338322,0.0058455556,-0.12555632,0.1504514217,0.146838814,-0.0259293783,0.0193639155,0.2058843374,-0.1102267727,-0.1738193929,0.1183539331,-0.2751955688,-0.2672133446,0.0372359268,0.2010167539,0.0381590351,0.009389149,-0.0081491908,0.0059310379,0.0556991696,-0.3565125763,0.023636166,-0.0094647072,-0.0558016375,-0.0841292068,0.2759922445,0.4360964596,-0.3608719409,-0.1035921797,0.0715648532,0.2175855786,0.7216550708,0.3425383568,-0.2379719913,0.1273078024,-0.3681203723,0.3619052172,0.6931351423,0.0160383359,-0.3624646366,0.8522378206,-0.1086349562,0.3817502558,0.4800842702,-0.20442909,0.1941484958,-0.0105180815,0.3352073431,0.2812868953,-0.0445720889,0.0385273471,-0.464271307,0.1312143207,0.0419497304,0.0378056802,-0.2513179779,0.2954408824,-0.0433923602,0.015032988,0.3520895541,-0.1800565869,0.1081850231,0.2904555202,0.2378064245,0.0199463479,0.2021734118,0.0092641879,-0.2178035229,0.0324927419,-0.3366108537,-0.2629368603,-0.2826167047,-0.0624102913,-0.1401127279,0.1487002075,-0.1843477637,-0.3533777595,0.1464820653,0.4111509025,0.3808980882,0.0224370789,-0.4623301327,0.3006553948,-0.2987254858,-0.1048559621,-0.2611705363,0.429363668,0.0721354857,-0.2631469369,0.1758847386,-0.0589892119,0.1717893928,-0.2416345477,-0.2221167684,0.2450815886,-0.1208557263,0.1709071249,0.1169753596,-0.2226256579,0.2957748175,-0.0062111109,-0.0720564499,0.3445874751,0.2193278074,0.2041208595,-0.2239909172,0.4739874303,-0.0305651519,0.3029977679,0.0209108982,-0.205836162,0.1924333572,-0.0084315231,-0.3421524465,0.0653704852,0.3345847428,-0.1258885711,0.0946051851,-0.0612957813,-0.3791344762,-0.0635569915,0.5271067023,-0.0830585435,-0.1615021825,0.1926877648,-0.1803122163,-0.0555438772,0.2965638041,-0.0067166113,0.3335767984,0.1575452387,0.104602471,0.0166130755,-0.1111166552,-0.2565548122,0.2091605067,-0.0900344998,0.1146262661,0.5449168086,0.0356366746,0.0449167937,-0.3995397389,-0.2428149432,-0.128882125,0.1455808133,-0.0990883484,0.384134531,-0.3911352754,-0.3713185489,-0.1201700941,-0.1128693894,-0.0630642995,-0.0857734159,-0.0269654337,0.2985586822,-0.2004606724,0.2329237461,0.0005526908,0.0394016802,0.251948297,-0.4227379262,-0.4818938971,-0.1630378664,-0.191156432,-0.0104031721,0.2656034231,-0.2878884673,0.3017160296,-0.0459037386,-0.3699751496,-0.2707997561,0.1194068417,0.1435605735,0.0721819699,-0.0227004942,-0.0912801325,0.2280841321,-0.0796872899,-0.3324358761,0.287009567,-0.2174522281,-0.1486556232,0.194851473,-0.0311844051,-0.0018445911,-0.2352986038,-0.2655209303,-0.1472442448,-0.4852220118,0.289681673,-0.0403695032,0.3086782992,0.2429099977,0.1440986991,0.1779596359,0.1556582451,-0.0546927191,-0.1509623826,-0.6362751126,0.3917452991,-0.1101357043,-0.2303782552,-0.0295004416,-0.1555970311,-0.0550973527,0.3612907529,-0.4562184811,0.237912789,0.0702732503,0.0834183842,-0.0950674713,-0.134057045,0.1914462298,-0.1075467318,-0.0603607446,-0.2590092719,0.0759437755,-0.0122350994,0.2527312934,0.1833108217,0.4013436437,0.7469967604,-0.039161507,0.2865437567,0.353391856,0.1458686143,0.5503771305,-0.1300928146,0.16025047,-0.2425827533,-0.2188410312,0.0596462861,0.1050836518,-0.3196414113,0.0274330787,0.0012185025,-0.3406588137,-0.3240846992,0.1244037524,-0.3062015474,-0.1882945746,0.2014841437,-0.0963935554,0.1964119375,-0.0394231714,-0.0296395868,-0.0032317135,0.1153508127,-0.048328694,0.2654097974,0.4285131991,0.1289328188,-0.0543236472,-0.1830666661,-0.3484928906,0.078764379,0.2743041515,0.6020610929,-0.1669999659,-0.1687550992,0.0798476189,-0.1684129536,0.4636302292,-0.1258804947,-0.0341541246,0.3472768664,-0.1242791489,-0.0266533308,-0.1190542877,-0.0431695469,0.3530095518,0.195949018,0.6723818183,-0.1772067547,-0.220200792,0.1132560149,0.0215652157,0.0907032862,-0.156267941,-0.2928073406,-0.1358119994,-0.201933071,-0.0655045658,-0.0717766806,0.2410783023,0.0199561808,-0.131281808,0.0536011457,-0.0730547979,-0.1126663312,0.0078033102,0.294038862,-0.0325312726,0.1593426615,0.057022687,0.5039331317,0.232784465,0.5136659145,0.5314756632,-0.4147776663,0.1325844377,0.0702452734,0.1329945028,0.1649885625,-0.1350931078,-0.1513888985,0.138104111,0.2574620247,-0.0355816707,-0.0832144171,0.2338197678,-0.1555819809,-0.3567377925,-0.4606517553,0.2432612181,-0.0392031707,0.1920537502,0.7787454724,-0.6046575308,-0.1680700779,0.1287034303,-0.327439338,0.6506022215,0.3056919277,-0.01261207,0.4230307937,-0.1530070156,0.4291078448,-0.4506242871,0.3880233169,-0.107545048,-0.1895088851,0.0136250248,-0.1521783918,0.3272462487,0.1736504287,-0.3301243484,0.1560810655,0.0949875191,-0.0261116754,0.056402456,0.0844660252,-0.3410910368,-0.1478554606,-0.6056392193,0.0822727308,-0.1643425077,0.3591601551,0.119864881,-0.3871607482,0.0100027071,-0.2189189792,-0.1640434861,0.4873065948,-0.2205232978,-0.0437984429,-0.1201645508,-0.6457443237,-0.3231777251,-0.1665352583,-0.3695810437,-0.1014870033,-0.0968803316,-0.0140162325,-0.1560026407,0.0815750286,-0.4244860113,0.2442319244,0.1879109144,-0.1447571814,-0.0854346454,0.1730190217,-0.2793938816,-0.55167979,0.2067956775,-0.0945816264,0.2443367243,0.1387685984,-0.014279468,0.0188988578,-0.1952127963,-0.2021060437,0.1315957457,-0.1408141851,-0.3405287564,-0.2667013407,-0.151902318,-0.200861007,-0.1044818312,0.2798337042,0.0701378211,0.3024792969,0.3966619372,-0.0176074822,-0.0633016229,-0.2849562764,0.0822029561,-0.1994450986,-0.0627179369,0.1458605379,-0.0238583386,0.1735637039,-0.2250163257,-0.0105745681,0.1574327648,0.0527799204,0.0311236531,-0.4647310376,-0.1359294653,-0.1071501747,-0.0055007907,0.1117188111,0.1044034883,0.2052558213,0.1506942809,-0.1848126501,-0.287404716,-0.1490447223,-0.1798323691,0.0388781689,0.0197014529,0.1445133686,0.3058988154,0.0417761318,0.1726244986,0.1845808923,0.1145528182,-0.2760638893,-0.1294748783,0.1017443016,0.1475505531,-0.1066455618,-0.2568067312,-0.0784032717,-0.239744544,-0.0903129727,0.1030201986,-0.0890858918,-0.0714164227,0.2204833478,0.3504038453,-0.0414797217,-0.1651158482,-0.0696929991,-0.2029707879,0.3246166408,-0.149192661,0.0957456976,0.0479332507,0.0756747648,-0.5871013403,-0.0813039467,-0.0964607447,0.1543788612,0.59524405,-0.4551479816,0.03112543,0.0455692224,0.1699540019,0.4789012969,-0.1304557621,0.0390133224,0.1729512513,0.1876669079,-0.0484990031,-0.0702733546,0.1377959102,0.1060291827,-0.0637325421,-0.0711321831,-0.1525841057,-0.1502206475,0.0998598635,0.0143192159,0.3330241144,0.0458889343,0.211354658,0.4080850184,0.2017606944,0.0718020722,-0.3809928596,0.0200543348,0.0718998685,-0.1887304932,-0.1524383426,-0.1257040054,0.0813989639,0.1215052903,-0.3927783072,-0.3188354373,-0.3502144217,0.0856014788,-0.1179064736,0.4086128771,-0.0059121046,0.2030270994,-0.2063986659,0.0338467211,0.0438848063,0.16065979,0.0185740534,0.115962036,-0.5277093649,-0.031080272,-0.0583055541,0.0389653742,0.3544371128,0.5155139565,-0.0361992531,0.2214543521,-0.1061519906,-0.1359117925,0.0883783028,0.1771474481,0.3425745368,-0.3739395738,0.2876466513,0.2135393769,0.0820103437,-0.0754379928,0.0099544805,0.3729601204,0.3608309031,0.0665017441,0.0278094839,0.0363483615,-0.0933402553,0.0266610924,0.335478574,0.1904851943,-0.1879787147,0.2221895158,0.1101663113,-0.088712424,0.3944174051,0.0983025655,0.1196607053,-0.1152643189,0.1943055987,-0.247738108,-0.1086110324,0.0640453398,-0.354144752,-0.358769089,-0.0634774119,0.2332483977,0.1124899834,0.0440791547,-0.0250769686,0.0849539489,-0.1289328188,0.279007405,0.2456029654,-0.0799998939,-0.1668445021,-0.1844663471,-0.2015265524,0.1580558419,-0.040410243,-0.1079189107,0.2371376604,0.3279328048,-0.0682448745,0.2060449421,-0.0900718644,-0.1822002977,-0.2946017087,0.2807135284,-0.3546518683,-0.2369975001,0.1787969619,0.2335970402,0.0910652205,-0.4543965757,0.2050732076,-0.3312031925,0.1178356186,0.0553922281,-0.0305205919,0.0156790111,0.0668511316,0.2356533855,0.1621074677,0.1155462116,-0.1349952668,-0.192071259,-0.072820738,-0.176847145,-0.0183110926,0.1458830535,-0.0582501702,0.502874732,-0.0484356917,-0.4014241993,-0.4371389151,0.4285812378,-0.0439626239,-0.4439990819,0.0103937835,0.0140104787,-0.1760369986,0.1818829775,0.0124762263,0.1777729988,-0.0506044216,0.0616825111,-0.3085522056,-0.1021585092,0.5942243338,-0.4224966168,-0.3192535639,0.3241405487,0.1266875267,0.1190700009,0.1298958212,-0.1965908855,0.1324760169,0.437204659,-0.0283873919,0.1215384007,0.1597634852,0.1152027473,0.1851295084,0.1178187653,0.0776488185,0.1725968719,-0.259760052,0.1329277754,-0.214050144]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2607","title":"Streaming local gzip compressed JSON line files is not working","comments":"@lhoestq Sorry I missed that. Thank you Quentin!","body":"## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0","comment_length":8,"text":"Streaming local gzip compressed JSON line files is not working\n## Describe the bug\r\nUsing streaming to iterate on local gzip compressed JSON files raise a file not exist error\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nstreamed_dataset = load_dataset('json', split='train', data_files=data_files, streaming=True)\r\n\r\nnext(iter(streamed_dataset))\r\n```\r\n\r\n## Actual results\r\n```\r\nFileNotFoundError Traceback (most recent call last)\r\n in \r\n----> 1 next(iter(streamed_dataset))\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 336 \r\n 337 def __iter__(self):\r\n--> 338 for key, example in self._iter():\r\n 339 if self.features:\r\n 340 # we encode the example for ClassLabel feature types for example\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in _iter(self)\r\n 333 else:\r\n 334 ex_iterable = self._ex_iterable\r\n--> 335 yield from ex_iterable\r\n 336 \r\n 337 def __iter__(self):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in __iter__(self)\r\n 76 \r\n 77 def __iter__(self):\r\n---> 78 for key, example in self.generate_examples_fn(**self.kwargs):\r\n 79 yield key, example\r\n 80 \r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/iterable_dataset.py in wrapper(**kwargs)\r\n 282 def wrapper(**kwargs):\r\n 283 python_formatter = PythonFormatter()\r\n--> 284 for key, table in generate_tables_fn(**kwargs):\r\n 285 batch = python_formatter.format_batch(table)\r\n 286 for i, example in enumerate(_batch_to_examples(batch)):\r\n\r\n~\/Documents\/GitHub\/datasets\/src\/datasets\/packaged_modules\/json\/json.py in _generate_tables(self, files, original_files)\r\n 85 file,\r\n 86 read_options=self.config.pa_read_options,\r\n---> 87 parse_options=self.config.pa_parse_options,\r\n 88 )\r\n 89 except pa.ArrowInvalid as err:\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json.read_json()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/_json.pyx in pyarrow._json._get_reader()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_input_stream()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.get_native_file()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile.__cinit__()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/io.pxi in pyarrow.lib.OSFile._open_readable()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nFileNotFoundError: [Errno 2] Failed to open local file 'gzip:\/\/file-000000000000.json::\/Users\/thomwolf\/github-dataset\/file-000000000000.json.gz'. Detail: [errno 2] No such file or directory\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.9.1.dev0\r\n- Platform: Darwin-19.6.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyArrow version: 1.0.0\n@lhoestq Sorry I missed that. Thank you Quentin!","embeddings":[-0.1992494613,-0.1339781284,-0.0091445101,0.2391350567,0.0777640566,0.0168800913,0.3976738453,0.5178918839,0.1970213056,0.0789984018,0.1030013189,0.3306827545,-0.0883667096,0.0208737329,0.2288863659,-0.1809338927,-0.0436617956,0.3560613692,-0.0251667686,0.0015848613,-0.1305664033,0.2937984169,-0.1404591352,-0.0185168069,-0.0187981017,-0.0395553522,-0.0471503064,0.1500941366,-0.0939593539,-0.3326378763,0.0627696067,-0.1684151441,0.188022688,0.4454552829,-0.000114536,0.288641125,0.2961105406,-0.1629139781,-0.3886071146,-0.4712482095,-0.229901135,-0.1492993236,-0.1307716817,-0.2858453691,-0.1378189325,-0.3474362791,-0.0337565579,-0.5071999431,0.5416871309,0.2922603488,0.1556637138,0.0419651642,-0.061772272,0.0905790925,0.0390155055,0.1573239863,-0.0886082724,0.3150972426,0.0327884369,0.1149678901,0.0552849993,0.4286454916,-0.2254960537,-0.1747770905,0.1473699361,-0.0008157574,-0.0843830332,-0.1257867962,-0.1683152765,0.085735105,0.1845451146,-0.4122557342,-0.4398344159,-0.3761076331,0.0466741547,-0.44880265,0.2118544728,0.4092858434,-0.250308305,0.1969692856,-0.100665614,-0.130531922,-0.3545617759,0.0854526162,0.36184901,0.3480824232,-0.0946776345,0.0091285398,-0.0919867903,0.1193795577,-0.0839845538,-0.3859618604,-0.1857065707,0.227756381,-0.2421413958,-0.1349226832,0.1443624049,-0.6062148213,0.1511498541,0.1012284085,0.4615779221,-0.1257445961,0.3588074744,0.0546084084,0.2728678286,0.1870484799,0.1769655049,0.2638421357,0.0566591471,0.3898805082,-0.0865667164,-0.1980740726,-0.2363374233,-0.0394049771,-0.0915300399,0.1368107945,0.4833262861,-0.1576653421,-0.089093484,0.081910789,-0.5941808224,-0.1327005029,0.1576721966,0.0986096784,0.0661023408,0.2201749086,0.026484726,0.1610841751,-0.002104495,-0.0618949644,-0.0907956734,0.0235721208,0.0740586743,0.1347989291,0.251897037,-0.5213885903,0.2793646455,0.0087546054,-0.046029456,-0.1562338322,0.0058455556,-0.12555632,0.1504514217,0.146838814,-0.0259293783,0.0193639155,0.2058843374,-0.1102267727,-0.1738193929,0.1183539331,-0.2751955688,-0.2672133446,0.0372359268,0.2010167539,0.0381590351,0.009389149,-0.0081491908,0.0059310379,0.0556991696,-0.3565125763,0.023636166,-0.0094647072,-0.0558016375,-0.0841292068,0.2759922445,0.4360964596,-0.3608719409,-0.1035921797,0.0715648532,0.2175855786,0.7216550708,0.3425383568,-0.2379719913,0.1273078024,-0.3681203723,0.3619052172,0.6931351423,0.0160383359,-0.3624646366,0.8522378206,-0.1086349562,0.3817502558,0.4800842702,-0.20442909,0.1941484958,-0.0105180815,0.3352073431,0.2812868953,-0.0445720889,0.0385273471,-0.464271307,0.1312143207,0.0419497304,0.0378056802,-0.2513179779,0.2954408824,-0.0433923602,0.015032988,0.3520895541,-0.1800565869,0.1081850231,0.2904555202,0.2378064245,0.0199463479,0.2021734118,0.0092641879,-0.2178035229,0.0324927419,-0.3366108537,-0.2629368603,-0.2826167047,-0.0624102913,-0.1401127279,0.1487002075,-0.1843477637,-0.3533777595,0.1464820653,0.4111509025,0.3808980882,0.0224370789,-0.4623301327,0.3006553948,-0.2987254858,-0.1048559621,-0.2611705363,0.429363668,0.0721354857,-0.2631469369,0.1758847386,-0.0589892119,0.1717893928,-0.2416345477,-0.2221167684,0.2450815886,-0.1208557263,0.1709071249,0.1169753596,-0.2226256579,0.2957748175,-0.0062111109,-0.0720564499,0.3445874751,0.2193278074,0.2041208595,-0.2239909172,0.4739874303,-0.0305651519,0.3029977679,0.0209108982,-0.205836162,0.1924333572,-0.0084315231,-0.3421524465,0.0653704852,0.3345847428,-0.1258885711,0.0946051851,-0.0612957813,-0.3791344762,-0.0635569915,0.5271067023,-0.0830585435,-0.1615021825,0.1926877648,-0.1803122163,-0.0555438772,0.2965638041,-0.0067166113,0.3335767984,0.1575452387,0.104602471,0.0166130755,-0.1111166552,-0.2565548122,0.2091605067,-0.0900344998,0.1146262661,0.5449168086,0.0356366746,0.0449167937,-0.3995397389,-0.2428149432,-0.128882125,0.1455808133,-0.0990883484,0.384134531,-0.3911352754,-0.3713185489,-0.1201700941,-0.1128693894,-0.0630642995,-0.0857734159,-0.0269654337,0.2985586822,-0.2004606724,0.2329237461,0.0005526908,0.0394016802,0.251948297,-0.4227379262,-0.4818938971,-0.1630378664,-0.191156432,-0.0104031721,0.2656034231,-0.2878884673,0.3017160296,-0.0459037386,-0.3699751496,-0.2707997561,0.1194068417,0.1435605735,0.0721819699,-0.0227004942,-0.0912801325,0.2280841321,-0.0796872899,-0.3324358761,0.287009567,-0.2174522281,-0.1486556232,0.194851473,-0.0311844051,-0.0018445911,-0.2352986038,-0.2655209303,-0.1472442448,-0.4852220118,0.289681673,-0.0403695032,0.3086782992,0.2429099977,0.1440986991,0.1779596359,0.1556582451,-0.0546927191,-0.1509623826,-0.6362751126,0.3917452991,-0.1101357043,-0.2303782552,-0.0295004416,-0.1555970311,-0.0550973527,0.3612907529,-0.4562184811,0.237912789,0.0702732503,0.0834183842,-0.0950674713,-0.134057045,0.1914462298,-0.1075467318,-0.0603607446,-0.2590092719,0.0759437755,-0.0122350994,0.2527312934,0.1833108217,0.4013436437,0.7469967604,-0.039161507,0.2865437567,0.353391856,0.1458686143,0.5503771305,-0.1300928146,0.16025047,-0.2425827533,-0.2188410312,0.0596462861,0.1050836518,-0.3196414113,0.0274330787,0.0012185025,-0.3406588137,-0.3240846992,0.1244037524,-0.3062015474,-0.1882945746,0.2014841437,-0.0963935554,0.1964119375,-0.0394231714,-0.0296395868,-0.0032317135,0.1153508127,-0.048328694,0.2654097974,0.4285131991,0.1289328188,-0.0543236472,-0.1830666661,-0.3484928906,0.078764379,0.2743041515,0.6020610929,-0.1669999659,-0.1687550992,0.0798476189,-0.1684129536,0.4636302292,-0.1258804947,-0.0341541246,0.3472768664,-0.1242791489,-0.0266533308,-0.1190542877,-0.0431695469,0.3530095518,0.195949018,0.6723818183,-0.1772067547,-0.220200792,0.1132560149,0.0215652157,0.0907032862,-0.156267941,-0.2928073406,-0.1358119994,-0.201933071,-0.0655045658,-0.0717766806,0.2410783023,0.0199561808,-0.131281808,0.0536011457,-0.0730547979,-0.1126663312,0.0078033102,0.294038862,-0.0325312726,0.1593426615,0.057022687,0.5039331317,0.232784465,0.5136659145,0.5314756632,-0.4147776663,0.1325844377,0.0702452734,0.1329945028,0.1649885625,-0.1350931078,-0.1513888985,0.138104111,0.2574620247,-0.0355816707,-0.0832144171,0.2338197678,-0.1555819809,-0.3567377925,-0.4606517553,0.2432612181,-0.0392031707,0.1920537502,0.7787454724,-0.6046575308,-0.1680700779,0.1287034303,-0.327439338,0.6506022215,0.3056919277,-0.01261207,0.4230307937,-0.1530070156,0.4291078448,-0.4506242871,0.3880233169,-0.107545048,-0.1895088851,0.0136250248,-0.1521783918,0.3272462487,0.1736504287,-0.3301243484,0.1560810655,0.0949875191,-0.0261116754,0.056402456,0.0844660252,-0.3410910368,-0.1478554606,-0.6056392193,0.0822727308,-0.1643425077,0.3591601551,0.119864881,-0.3871607482,0.0100027071,-0.2189189792,-0.1640434861,0.4873065948,-0.2205232978,-0.0437984429,-0.1201645508,-0.6457443237,-0.3231777251,-0.1665352583,-0.3695810437,-0.1014870033,-0.0968803316,-0.0140162325,-0.1560026407,0.0815750286,-0.4244860113,0.2442319244,0.1879109144,-0.1447571814,-0.0854346454,0.1730190217,-0.2793938816,-0.55167979,0.2067956775,-0.0945816264,0.2443367243,0.1387685984,-0.014279468,0.0188988578,-0.1952127963,-0.2021060437,0.1315957457,-0.1408141851,-0.3405287564,-0.2667013407,-0.151902318,-0.200861007,-0.1044818312,0.2798337042,0.0701378211,0.3024792969,0.3966619372,-0.0176074822,-0.0633016229,-0.2849562764,0.0822029561,-0.1994450986,-0.0627179369,0.1458605379,-0.0238583386,0.1735637039,-0.2250163257,-0.0105745681,0.1574327648,0.0527799204,0.0311236531,-0.4647310376,-0.1359294653,-0.1071501747,-0.0055007907,0.1117188111,0.1044034883,0.2052558213,0.1506942809,-0.1848126501,-0.287404716,-0.1490447223,-0.1798323691,0.0388781689,0.0197014529,0.1445133686,0.3058988154,0.0417761318,0.1726244986,0.1845808923,0.1145528182,-0.2760638893,-0.1294748783,0.1017443016,0.1475505531,-0.1066455618,-0.2568067312,-0.0784032717,-0.239744544,-0.0903129727,0.1030201986,-0.0890858918,-0.0714164227,0.2204833478,0.3504038453,-0.0414797217,-0.1651158482,-0.0696929991,-0.2029707879,0.3246166408,-0.149192661,0.0957456976,0.0479332507,0.0756747648,-0.5871013403,-0.0813039467,-0.0964607447,0.1543788612,0.59524405,-0.4551479816,0.03112543,0.0455692224,0.1699540019,0.4789012969,-0.1304557621,0.0390133224,0.1729512513,0.1876669079,-0.0484990031,-0.0702733546,0.1377959102,0.1060291827,-0.0637325421,-0.0711321831,-0.1525841057,-0.1502206475,0.0998598635,0.0143192159,0.3330241144,0.0458889343,0.211354658,0.4080850184,0.2017606944,0.0718020722,-0.3809928596,0.0200543348,0.0718998685,-0.1887304932,-0.1524383426,-0.1257040054,0.0813989639,0.1215052903,-0.3927783072,-0.3188354373,-0.3502144217,0.0856014788,-0.1179064736,0.4086128771,-0.0059121046,0.2030270994,-0.2063986659,0.0338467211,0.0438848063,0.16065979,0.0185740534,0.115962036,-0.5277093649,-0.031080272,-0.0583055541,0.0389653742,0.3544371128,0.5155139565,-0.0361992531,0.2214543521,-0.1061519906,-0.1359117925,0.0883783028,0.1771474481,0.3425745368,-0.3739395738,0.2876466513,0.2135393769,0.0820103437,-0.0754379928,0.0099544805,0.3729601204,0.3608309031,0.0665017441,0.0278094839,0.0363483615,-0.0933402553,0.0266610924,0.335478574,0.1904851943,-0.1879787147,0.2221895158,0.1101663113,-0.088712424,0.3944174051,0.0983025655,0.1196607053,-0.1152643189,0.1943055987,-0.247738108,-0.1086110324,0.0640453398,-0.354144752,-0.358769089,-0.0634774119,0.2332483977,0.1124899834,0.0440791547,-0.0250769686,0.0849539489,-0.1289328188,0.279007405,0.2456029654,-0.0799998939,-0.1668445021,-0.1844663471,-0.2015265524,0.1580558419,-0.040410243,-0.1079189107,0.2371376604,0.3279328048,-0.0682448745,0.2060449421,-0.0900718644,-0.1822002977,-0.2946017087,0.2807135284,-0.3546518683,-0.2369975001,0.1787969619,0.2335970402,0.0910652205,-0.4543965757,0.2050732076,-0.3312031925,0.1178356186,0.0553922281,-0.0305205919,0.0156790111,0.0668511316,0.2356533855,0.1621074677,0.1155462116,-0.1349952668,-0.192071259,-0.072820738,-0.176847145,-0.0183110926,0.1458830535,-0.0582501702,0.502874732,-0.0484356917,-0.4014241993,-0.4371389151,0.4285812378,-0.0439626239,-0.4439990819,0.0103937835,0.0140104787,-0.1760369986,0.1818829775,0.0124762263,0.1777729988,-0.0506044216,0.0616825111,-0.3085522056,-0.1021585092,0.5942243338,-0.4224966168,-0.3192535639,0.3241405487,0.1266875267,0.1190700009,0.1298958212,-0.1965908855,0.1324760169,0.437204659,-0.0283873919,0.1215384007,0.1597634852,0.1152027473,0.1851295084,0.1178187653,0.0776488185,0.1725968719,-0.259760052,0.1329277754,-0.214050144]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2606","title":"[Metrics] addition of wiki_split metrics","comments":"#take","body":"**Is your feature request related to a problem? Please describe.**\r\nWhile training the model on sentence split the task in English we require to evaluate the trained model on `Exact Match`, `SARI` and `BLEU` score\r\nlike this \r\n![image](https:\/\/user-images.githubusercontent.com\/26653468\/124746876-ff5a3380-df3e-11eb-9a01-4b48db7a6694.png)\r\nWhile training we require metrics which can give all the output\r\n\r\nCurrently, we don't have an exact match for text normalized data\r\n\r\n**Describe the solution you'd like**\r\nA custom metrics for wiki_split that can calculate these three values and provide it in the form of a single dictionary\r\nFor exact match, we can refer to [this](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/data\/metrics\/squad_metrics.py) \r\n\r\n**Describe alternatives you've considered**\r\nTwo metrics are already present one more can be added for an exact match then we can run all three metrics in training script\r\n\r\n#self-assign","comment_length":1,"text":"[Metrics] addition of wiki_split metrics\n**Is your feature request related to a problem? Please describe.**\r\nWhile training the model on sentence split the task in English we require to evaluate the trained model on `Exact Match`, `SARI` and `BLEU` score\r\nlike this \r\n![image](https:\/\/user-images.githubusercontent.com\/26653468\/124746876-ff5a3380-df3e-11eb-9a01-4b48db7a6694.png)\r\nWhile training we require metrics which can give all the output\r\n\r\nCurrently, we don't have an exact match for text normalized data\r\n\r\n**Describe the solution you'd like**\r\nA custom metrics for wiki_split that can calculate these three values and provide it in the form of a single dictionary\r\nFor exact match, we can refer to [this](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/data\/metrics\/squad_metrics.py) \r\n\r\n**Describe alternatives you've considered**\r\nTwo metrics are already present one more can be added for an exact match then we can run all three metrics in training script\r\n\r\n#self-assign\n#take","embeddings":[-0.0108734518,-0.4354909658,-0.0810058638,-0.0047024167,0.1584804803,0.0482319556,0.2976144552,0.2412920594,0.0471994504,0.0514928512,-0.3600813448,0.2465806752,-0.0850384608,0.3682302535,-0.0116167311,-0.4090156853,-0.0145660974,0.0672861338,0.2245051414,-0.0883490965,-0.0782746151,0.2452747673,-0.1668302715,-0.0830089748,-0.4655505717,0.1755770892,-0.2348630279,-0.3415921926,-0.1821687669,-0.2304375619,0.1858531535,0.2258653641,-0.0856704637,0.0899452567,-0.0001124652,-0.1922752261,0.0371840484,-0.3581243753,-0.0468795449,-0.2123134434,-0.0086784782,-0.2676844299,0.0231268108,-0.3474774957,-0.264344126,0.4214245677,-0.0139227184,0.253911674,0.4268681705,0.2260359675,0.0992535502,-0.1226979718,-0.1266997755,0.0713395998,0.0240931418,0.0118396617,-0.2083725035,0.1993444711,0.0340768285,-0.0793713331,-0.095305346,0.4785133898,0.2631920576,-0.1101806015,0.1705851555,0.1973936558,0.2021975219,-0.1380425692,-0.1045731381,0.2661604583,0.1761663258,0.0021506252,-0.3227240443,-0.1788071096,0.073977448,-0.346756041,-0.0043581128,-0.1597502083,0.0439025238,0.0966747627,-0.1678279787,-0.1711270511,-0.3129262626,0.2143776119,-0.2289673239,0.8462690115,0.0223514568,-0.0876246914,0.1312348247,0.104237102,-0.1838095486,-0.1042055488,-0.0161817111,0.3422540128,-0.3128513098,-0.5118837953,-0.3126981258,-0.3399874568,0.2238509953,-0.3415017724,-0.0783388689,0.0586883016,0.3344822526,-0.1088119,0.0781163946,0.2806843817,0.2978432775,0.4095400274,0.0631273016,0.3250563741,-0.0119261108,-0.1812836081,0.1960469186,0.1150623411,-0.1236703992,0.1590280086,-0.231965363,-0.0287504084,-0.0954474062,0.0075334087,0.1866023093,-0.3153975606,0.0367763638,0.2459415793,0.0382750779,0.1716632247,-0.1766652465,0.3307554126,-0.1169535965,0.0243742093,-0.0514460057,0.1045718715,0.0075544799,0.3440403938,0.1096572652,0.2466457039,0.3307707012,-0.0840446353,0.2290849835,-0.4110360444,-0.1909361035,-0.0869821906,-0.1034576893,-0.0609439649,-0.1829702854,-0.1230828986,-0.0050824257,-0.5211009979,-0.1641472876,-0.149210915,-0.2664138973,-0.2620646954,0.0708805993,0.1893960387,-0.1814319938,-0.1542225629,0.1760816574,0.7476133704,0.2323122174,0.1461353153,-0.0147564439,0.1642362177,-0.5481278896,0.0008413988,0.0926151723,0.087339066,-0.2480805367,-0.4355640411,0.1604940146,-0.0196156595,-0.0493639894,0.0920919031,0.0014264345,0.3924592733,0.2942033112,0.3645392954,0.5195230842,-0.6817261577,-0.0896248519,0.0400216766,0.0123543544,0.0719486773,0.1726577282,-0.0676327422,0.5020038486,0.1433514655,0.2470118701,0.4273153543,0.0223217048,-0.2457630187,-0.1816884428,-0.3276519477,0.0831920281,0.258133769,0.5685046315,-0.4379381239,0.0435401648,0.4242618084,0.2084986567,-0.2558453083,0.0875275731,-0.1042796001,0.091351077,-0.1662482917,0.2745337784,-0.5754967332,-0.0806728527,0.2845152318,-0.4863102436,0.2022362947,0.1472886056,-0.4478150904,-0.3981456757,-0.101802133,-0.2832986414,-0.3267942369,0.1423493177,-0.4428462088,0.1128974929,0.1591225863,-0.2000311911,-0.2113046497,-0.3816462755,0.1403284967,0.3021244705,0.1607596725,-0.1948791146,0.0476152562,-0.0322220735,0.5221585035,0.1589255631,0.093800135,-0.0178318489,0.0916847289,0.089845553,0.0783327594,0.1746248305,-0.0245656446,0.385011673,-0.1361356974,0.0253865868,-0.046365656,0.0587711185,-0.357516855,-0.4234451652,0.4680012763,0.1500843614,0.1935080141,-0.2365969419,0.004178924,0.0282172337,-0.2439463884,0.0162867047,-0.4874413013,-0.0041830982,-0.3844891489,-0.4361751974,-0.3380913436,0.340883702,-0.08529111,0.6905116439,-0.0917510837,0.17493999,0.1817024499,0.1581846327,-0.1576342285,-0.3220058382,0.1084281728,0.3177078366,0.201956585,0.1001745313,0.0949339718,0.1600643843,-0.0286048763,0.264057368,0.119466275,-0.012531531,0.4884136915,0.0895214155,-0.1239243671,0.0296052881,0.116112262,0.0741665363,-0.1066145077,-0.1420397609,-0.1895856708,0.1387465298,0.2250508219,0.0576538853,0.0570085794,-0.3171152771,-0.1993054897,0.4867733419,-0.4071998894,0.0718539953,0.0471871682,0.3667993844,0.2676619291,-0.1902951449,-0.0155196758,-0.016442826,-0.5182403922,-0.0150733674,0.1086686403,-0.0132299364,0.0805531293,0.0289110523,-0.1798404306,-0.063358292,0.0581646152,-0.6479976773,-0.045455873,-0.1209051907,0.4809908867,0.182935372,-0.1582979858,-0.1113382354,-0.1929662824,0.1243660152,0.145489037,-0.0105079198,-0.1031789407,-0.1503875852,-0.0694659725,-0.0842636377,-0.5363788605,-0.3308976293,-0.1420385242,0.1345841438,0.1214548647,0.134393841,-0.0182168055,-0.1027972028,0.2257796824,-0.099556312,-0.0044090957,-0.2622331083,0.2412851602,0.2950268686,-0.2837686837,-0.1603805274,-0.002141247,-0.2870586216,0.237929225,-0.3253953457,-0.1236787587,-0.3656806946,0.0497286953,0.1419606507,-0.067446284,0.0481553115,0.2488096505,-0.0275256243,0.0342696719,-0.1894263923,0.0197124053,0.2812471092,0.0765397996,0.144605279,-0.2293727994,0.291782707,0.3041765392,1.002266407,0.2499503493,0.0869392231,-0.0209035929,0.1040930152,-0.1312124282,0.1490108371,-0.173412174,0.1045545265,-0.1417907625,0.0687133446,0.6525756121,0.1677458286,-0.2328488082,-0.2220606059,-0.4113157988,-0.0526912585,-0.2905477285,0.1367675811,0.3410485685,0.1170229018,-0.0887090042,0.1445301324,-0.1450496763,-0.1359406114,0.122483857,0.3009503186,-0.2792764604,-0.1201574206,-0.1096251756,-0.1464102566,-0.3137571216,0.050702434,0.4211500883,-0.1813813001,-0.1266328692,0.256413728,0.0369607657,-0.0244236123,0.4259935915,-0.4380820394,-0.1031240895,-0.054791294,0.0734477267,0.1480816901,-0.0036034647,-0.2957559526,-0.2477080226,0.3483385444,0.7993957996,-0.3927590549,-0.1666149795,0.3952989876,0.0417071953,-0.3035710752,-0.2420864105,-0.1031031385,0.0213519689,-0.2344149202,0.2348171771,0.239741236,0.1890729666,0.1302186549,-0.1185843796,0.124610886,0.0308943689,0.1568292826,0.1971368194,-0.2235005349,0.1165731251,0.0424999446,0.0134556498,-0.5588419437,-0.4197962284,0.0540450364,-0.2478689253,-0.3487437963,-0.2998293936,-0.3658888936,0.5713354945,0.3803950548,0.0722264946,0.21302028,0.0009595511,0.1132120937,0.0511264689,-0.0553890243,0.1472159922,0.114591293,-0.2027529329,-0.0111352466,0.4667524695,-0.0338670239,-0.0131159863,0.0226206444,0.1809388697,-0.47612077,0.4870862365,0.4023005962,1.009608984,-0.0311623272,-0.0306951292,0.0138004236,-0.2348680198,0.3145507574,-0.5224723816,0.0921685249,-0.2304604799,-0.0754976943,-0.0717494786,0.1820046157,0.1061332449,0.0738433301,-0.4667941332,0.2523726523,0.0158751886,0.1071078107,0.0293071549,0.3842848837,-0.0693093613,-0.2510159314,0.0538304225,0.1073154435,-0.0299580693,0.0555760302,-0.154536128,-0.2059567124,-0.0579645596,-0.0608299561,-0.1603045464,-0.1590043753,-0.0655693039,0.0261786189,-0.1500149816,-0.1758954972,0.2638113499,0.4314410985,0.3160432577,0.1020517647,-0.3059585392,0.2816359401,-0.4747048914,0.1190100461,-0.0653978735,-0.0762350783,0.1911738217,-0.2721585631,-0.3162779212,-0.1791117787,0.1331796497,0.0547424667,0.0480273552,-0.1227551997,0.2591999769,-0.3321759701,0.0491220839,0.0050236499,-0.16660887,-0.2805168331,0.1333911866,-0.0973056778,-0.424882412,0.3512923121,0.0596512295,-0.1164836138,0.0404191427,0.1460100859,0.1099529564,-0.151640594,0.2069425434,-0.0469800718,-0.272033006,-0.2378074974,-0.0318068266,-0.2581899166,-0.3723013401,0.0974514037,0.2059372216,-0.2405399978,-0.0304036178,0.4231791794,0.3888774216,0.0797244236,-0.040164493,-0.1680778265,-0.0582553893,-0.0249085221,-0.1157816425,0.1656404287,-0.1150864586,0.2409557849,0.0271313097,0.6229279637,-0.3118356168,0.0469223224,-0.2905961573,0.1967642605,0.0607736744,-0.3178661764,0.088963002,0.2049255371,0.0500422977,0.0329432935,-0.0947168097,-0.2348494679,-0.3631631732,0.1372924149,0.0730686188,0.4026608467,-0.0084872153,0.3488598168,-0.0928246081,-0.0968764648,0.0894628316,0.082411252,0.1265242696,0.6913731694,0.1860764325,0.3231155872,-0.1931949258,-0.1400911063,-0.069677256,0.2251370996,-0.1034554094,0.2135898322,0.0048790746,0.008710824,-0.1151182652,0.0113798389,0.2831558585,0.065459013,0.2416498363,0.3101626933,-0.0341921635,0.1043158621,-0.0391538776,-0.060704533,-0.106434077,0.1226768419,-0.0397182554,0.2175123841,-0.2944982052,0.0427647978,0.2501755953,-0.2450509518,0.1678249687,0.3118280172,0.3508434296,0.07067734,0.1669103205,0.0311469361,0.0749448612,-0.2445232123,0.3675307631,0.3725689948,0.1773701608,-0.0355921462,0.3742093742,-0.318523705,0.0726473927,0.1118429601,-0.3560016155,0.2426685244,-0.1969230771,0.3990127444,0.2942784131,0.3562266827,0.1945735067,-0.0824012086,0.0547910072,-0.0172640998,-0.053867586,0.2539360523,0.1379588395,-0.0854049996,-0.2546373904,0.2750627398,0.0090842713,-0.0912612081,-0.0732017085,-0.2155235112,-0.2308391929,-0.0596610792,0.1457537413,-0.1032838449,-0.0167070311,0.2443398684,0.2111495882,0.295861274,0.0128967278,0.2044301927,0.6765684485,-0.4111682475,-0.1308387369,0.3252367079,0.2052395046,-0.0149190361,0.1801188439,-0.0178181157,-0.0409122035,-0.2533358037,0.1585061997,0.0297928415,-0.1888443828,0.3041006029,0.2794083953,0.0711143613,-0.0544580184,0.2535883188,0.1221513078,-0.2410747111,0.019018136,0.1118210927,-0.309651792,-0.4022974074,0.2114533633,0.3813875616,-0.0587709099,-0.1250756383,-0.0260023791,-0.334421128,0.1126562953,0.0206005108,0.0206803996,0.2187530845,-0.0964986831,0.0764336139,0.1586755514,0.0332267582,-0.1186452359,-0.0471526645,-0.4126438498,0.2628372908,-0.0615074337,-0.2529619038,0.3107227683,-0.0410642996,0.0616172291,0.1714739799,-0.2872017026,0.2845579386,0.0956055745,-0.1126533747,-0.0388256274,0.0199135318,-0.3316313624,0.0455927327,-0.0690705329,-0.0565653108,0.1821770519,-0.0045478549,-0.2506435215,-0.0098154983,-0.1290107369,-0.0673115104,-0.0528664477,0.3382844627,0.1761284471,-0.1106468886,0.1834387481,0.5330587626,-0.0932064503,0.2823740244,0.1793414205,0.0369024538,-0.4033731818,0.0745892599,-0.0085023008,0.2588573396,-0.1020166278,-0.1850942671,0.0777364224,0.3389540911,0.2206950188,-0.0230118819,-0.1241777688,0.027831357,0.0182170067,-0.1421001405,0.3622922599,0.5205355883,-0.1265090704,0.1521385163,-0.0578896776,-0.2907643616,0.1811594069,-0.17970936,-0.2190361619,0.0565602966,0.2494017631,0.2573389709,0.136787653,-0.8087661862,-0.2768653929,0.3903912008,0.0381560214,0.0500318408,0.5434799194,0.40083009,-0.1310540289,-0.0431869552,-0.3491101563,0.0107251946,-0.0706225485,-0.2817269564,-0.4649561644]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Hi !\r\nIf we want something more general, we could either\r\n1. delete the extracted files after the arrow data generation automatically, or \r\n2. delete each extracted file during the arrow generation right after it has been closed.\r\n\r\nSolution 2 is better to save disk space during the arrow generation. Is it what you had in mind ?\r\n\r\nThe API could look like\r\n```python\r\nload_dataset(..., delete_extracted_files_after_usage=True)\r\n```\r\n\r\nIn terms of implementation, here are some directions we could take for each solution:\r\n1. get the list of the extracted files from the DownloadManager and then delete them after the dataset is processed. This can be implemented in `download_and_prepare` I guess\r\n2. maybe wrap and mock `open` in the builder to make it delete the file when the file is closed.","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":129,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nHi !\r\nIf we want something more general, we could either\r\n1. delete the extracted files after the arrow data generation automatically, or \r\n2. delete each extracted file during the arrow generation right after it has been closed.\r\n\r\nSolution 2 is better to save disk space during the arrow generation. Is it what you had in mind ?\r\n\r\nThe API could look like\r\n```python\r\nload_dataset(..., delete_extracted_files_after_usage=True)\r\n```\r\n\r\nIn terms of implementation, here are some directions we could take for each solution:\r\n1. get the list of the extracted files from the DownloadManager and then delete them after the dataset is processed. This can be implemented in `download_and_prepare` I guess\r\n2. maybe wrap and mock `open` in the builder to make it delete the file when the file is closed.","embeddings":[-0.0422243811,-0.052077923,-0.1375647485,0.249136284,-0.0890204087,0.1040985212,-0.1198280752,0.1847193092,0.234230876,0.2407588214,0.1567286253,0.5178807378,-0.3404889405,-0.0508526862,-0.1349290013,-0.020200653,-0.1992256045,0.3400244117,-0.0047850101,0.2303301692,-0.2209787965,0.0262257326,0.0232040863,-0.1026177779,0.0811791494,-0.3859203756,0.1359873116,-0.1867462993,-0.1622753888,-0.5164051652,0.05417528,0.5212088823,0.2322490513,0.0971321985,-0.0001223348,0.0322482027,0.2591515183,0.079954356,-0.4075611234,0.1244008169,-0.1461262256,-0.2453016639,0.2597362101,-0.1671171933,0.2101428211,-0.564896524,-0.2238956094,-0.6139017344,0.5910028219,-0.0977437869,0.0498433188,0.1384608746,-0.3260321617,0.2452622503,0.2004276514,0.2770322561,-0.0424004868,0.0223673433,0.4151899219,0.0967546105,0.1619669199,0.3807038665,-0.0550241023,-0.1603195667,0.4445157647,0.0631318614,-0.0319245309,-0.4852730334,0.1398253888,-0.0066928174,0.7061523795,-0.4330633283,-0.2672884464,-0.457439065,0.0280466024,-0.592244029,0.0385922231,0.2951566577,-0.1211811528,0.2089288384,-0.110395126,-0.4544366598,-0.3733914196,-0.0871406049,0.6077163815,-0.3243749738,-0.0568619892,-0.0656688809,0.3697440028,0.0736787319,0.2949277461,-0.2698952258,-0.1399102509,0.5059065223,-0.0271433536,-0.1845743209,-0.2607621849,-0.2897108495,0.11943537,0.3029824495,0.3516383767,-0.007010607,0.193062216,0.1110153273,0.1876697838,0.3751322925,0.0078877974,-0.0530541278,0.1721748412,0.2076802552,0.0642983392,-0.0915418714,0.0383336209,-0.1278788298,0.3636780679,0.0154576702,0.0716330782,0.0261886418,-0.1650816947,0.07824976,0.1291304678,-0.135167703,0.1513490379,0.2670413256,0.0587308556,0.1620476097,0.215889141,0.2350441068,-0.0299415961,-0.1465885341,0.1376228184,0.0731804147,-0.2401808649,-0.1884089112,0.3641346991,-0.0083342325,0.0416415893,0.0005885031,-0.3138786256,-0.0641817972,0.6676418781,-0.0993862525,0.2807014883,0.3421969414,-0.1093466878,-0.2436191589,-0.3248208165,-0.1198692247,-0.19655855,0.5875695944,-0.1568873972,-0.3170170784,-0.1368437409,0.1290899366,-0.0270507522,0.2437296063,-0.5556287766,0.1172113717,0.16315265,0.0627298281,-0.0224351045,0.2961559594,-0.0586108044,-0.1787772775,-0.189347744,0.3691442907,-0.5007237196,-0.1223104745,-0.3205631673,-0.1797679216,0.1424594522,-0.0010882767,-0.1239746287,0.37910676,-0.2620406151,0.091597259,0.5124638677,-0.181572035,-0.4231873453,0.1482755989,0.1100984067,0.2711139619,0.3149678707,0.2507965267,0.2787887752,0.0021172836,-0.038488593,0.4016497135,0.0736518428,0.0040990273,-0.2239464074,-0.4116459191,0.0513558574,0.2471985817,-0.0207932256,0.039883431,0.0623965673,-0.1024292186,0.3069228828,0.097825557,0.151983127,0.1327217519,0.1448920667,-0.2485721409,-0.2080590725,0.0697800443,-0.5534030199,0.173768878,-0.5192010403,-0.291768074,-0.4091202915,-0.3444458842,0.0820593238,0.1959255189,-0.2989016473,0.1966816336,0.0327640213,-0.1894039512,0.2831685543,-0.0064171986,-0.4284304976,0.0199055001,-0.1409000307,0.1741943359,-0.4501312375,0.1040370539,0.2124263197,-0.1986983567,-0.1110477,-0.0094022434,-0.2910374701,-0.3422919512,-0.093639046,0.2829712629,0.3651207089,0.268342793,0.1073635519,0.2281978279,0.2203880847,0.2218503952,0.0824293271,0.2590484917,0.0871407762,0.1824860573,-0.2291260958,0.5258798003,-0.0765097216,0.0241309945,-0.1283982992,-0.3022350371,0.2240570784,-0.0017422049,0.0128592802,-0.4730243087,-0.0542870536,0.3683068752,0.1288885623,0.0917666256,0.1476072818,-0.0150869377,0.0925110132,-0.103151001,-0.1026698127,0.3242192566,-0.0696177632,-0.1613177806,0.0573300086,0.4151658118,0.4481205642,0.0921721533,-0.1154637635,-0.078765668,0.4702810943,0.0514211953,0.2684764862,0.1953540295,0.0813863352,0.3618116379,0.1682407856,-0.062261533,-0.2748675942,-0.0914838165,0.0715449303,0.2726363242,0.2208339423,-0.0611376241,-0.1839365214,-0.2929829359,0.2759950757,0.1252204776,0.2172905952,-0.0161953643,0.4093545973,0.0779331475,-0.3893392384,0.0748196319,0.0082118474,0.5284100771,-0.0873366371,-0.2672452927,-0.3404185772,-0.1312678456,0.2366932034,-0.0275300201,0.3818286657,-0.198455438,0.5107133985,0.0785908923,0.0959751084,-0.0302718319,-0.0918449312,0.0615337417,0.4084796011,0.0522785895,0.1331733316,0.1684497297,0.1248740777,0.0609198548,-0.1308309436,-0.2205235362,-0.0963044167,0.1561925262,0.0524337031,0.0963251069,-0.2165250629,-0.0163153131,-0.3071950078,-0.3777163327,0.2557627857,0.1072915643,0.1372369677,0.0845113918,0.2943944335,-0.0587567501,0.1412729025,0.0703832358,0.0453881212,-0.5879281759,0.1189820543,-0.2073395699,0.0725701302,0.3074827492,-0.122104302,0.060423281,0.3388400972,-0.3763493299,-0.0693922415,-0.041573368,0.2056569159,-0.2972852886,-0.0365348272,0.1967595965,0.332067728,-0.0217487123,0.0054507819,0.2608991861,0.1648945659,0.2400366217,0.0658634081,0.2286016941,0.2869085073,0.0574271269,0.4774214923,-0.1240692511,0.0212154537,0.3330622911,0.3224788606,0.3370009363,-0.2226606905,0.2584810257,-0.5167121887,-0.0704164281,-0.1229764894,0.0886602029,0.1452048123,-0.1283564121,0.0905667841,-0.1557757109,-0.2041723281,-0.1858889759,0.3037092984,-0.3123892844,0.0866007134,-0.0249372311,-0.2253086418,-0.1689941138,-0.0579383373,0.0685019717,0.1171312407,0.5481666327,-0.1229562834,-0.0619335175,-0.0753128976,-0.371712625,0.2466869503,-0.1443476826,-0.2700845897,0.144191429,0.0228029434,0.1901639551,0.0920120403,0.2852837443,-0.0214631185,-0.0921425,-0.193748951,-0.0241023228,-0.3323916495,0.0090945959,0.0599585138,0.0449143201,-0.1582286507,0.3985447586,-0.3091912866,-0.1924604326,-0.3509970605,0.6400802732,-0.2645498216,-0.0694291219,-0.1440260857,-0.30620116,-0.6328192353,0.0667906776,0.0798805356,-0.0590765662,0.1539488882,-0.3315378129,-0.1019188687,-0.1806176156,0.018713044,0.014303932,0.6302096844,-0.3630812764,0.0931552202,-0.1307257861,0.023438951,0.4154005051,0.3095870316,-0.0904389024,-0.1662482768,-0.087940149,-0.2519861162,0.0639646649,0.1684660614,-0.2705995739,0.4433526695,-0.1818029433,0.5496017933,-0.1045359522,-0.0213431846,-0.0765382797,-0.2536759377,0.0044941828,-0.4548161626,0.2276911139,0.0446160436,-0.17590639,0.5514417887,-0.0850599185,-0.269974649,0.5203489661,0.0690080449,1.0979140997,-0.0484184437,0.0587538034,-0.0167360753,-0.277587086,0.3807532489,-0.4976661801,0.2090118378,-0.0645463169,-0.0310499091,0.0491861254,-0.0622942112,0.0427290425,0.2365058959,-0.1930373013,0.1230343133,-0.3904757202,-0.1313737035,-0.2855242491,0.3503832519,-0.2633569837,-0.2937426567,-0.242905587,-0.0423143953,0.0437845178,0.0046018329,-0.3125770092,-0.0823141411,0.3401916027,-0.062388435,0.2848976552,-0.1401859671,-0.3396236897,0.3695364892,-0.1837324947,-0.1763905287,-0.2778175771,0.2090361416,-0.2893456519,-0.0132783838,0.1523069143,0.0673097,0.0064265812,0.0233527906,-0.3320411742,-0.0007070292,0.0622992404,-0.1600415558,0.0114372652,0.2037501633,-0.3674548566,-0.3369385004,0.2151462287,0.1448117495,-0.0885298625,0.0572303161,-0.3552044928,-0.2007966489,-0.4958284199,0.0858129933,0.0628234819,0.3129725754,-0.2482839674,0.2814537287,-0.3506737351,-0.1324832439,-0.1104247123,-0.2255333811,0.2954696715,-0.0500770621,0.3738799989,-0.1685556769,-0.2155819088,-0.3375291526,0.0205049217,-0.0178961772,-0.4794527292,0.4637895226,0.1569728553,-0.0721625611,0.03540162,0.0089947144,-0.0942448378,0.0493260697,-0.1572988033,-0.1785902381,-0.5771638751,0.2368875742,0.2657475471,0.4896398187,0.0345769264,-0.1116786599,-0.1058107018,0.0329581499,-0.2004258335,0.2264089286,0.2059653699,0.1356118172,-0.0961439684,0.2747266889,0.0138176642,-0.2292102128,-0.1126270145,0.0334489308,-0.1656562984,-0.1269175857,-0.1983715743,0.1402419209,-0.0914144665,-0.1350338906,-0.0911215916,-0.2899683416,0.0810251012,-0.1910169572,0.2622929215,-0.2783589065,-0.1321274191,-0.1273304522,0.3750869334,0.0331165269,-0.2510362864,0.2264987677,-0.1331769675,0.410457164,0.0275567211,0.0273100492,-0.0992766991,-0.056713298,-0.2499890774,0.0564191639,0.2500882447,0.0263423696,0.1521479487,-0.071245797,-0.0994432867,-0.0005779213,0.1916674972,0.2741944492,-0.0699497312,-0.0008056021,-0.0776972249,0.175611496,-0.2982020378,-0.1586703211,-0.0591925047,-0.0076438049,-0.1619354635,0.2961508334,0.2256487459,0.0899920613,-0.1198943779,0.1513798237,0.3126499951,-0.2347842455,0.2393759489,0.3698675931,0.0780754164,-0.1059271842,-0.0545853153,0.3103908002,0.2109915316,-0.0467230082,-0.2592592537,0.4465030432,0.507286787,0.2389114946,-0.5002034903,-0.3075029552,-0.2232683599,0.519541204,-0.08148399,0.3681805432,0.4082181752,0.2382922769,0.2595690191,-0.2480434626,-0.1358735859,0.5102303028,0.0621812977,-0.0676644742,0.0678967983,-0.1146043912,0.1157870293,0.10465689,0.17962192,0.1921066791,-0.1854077429,0.3415929675,0.073263891,0.2138117999,0.4533448219,0.2281685472,0.4140392542,-0.2492756695,-0.2296032608,0.2814235687,0.2110300958,-0.1215110496,-0.1747248918,0.0849729702,0.1095462963,-0.4609772265,0.0860381052,0.0707164854,0.0760113746,-0.2645121217,0.3912464082,0.2391772568,-0.0945055261,-0.0702006221,0.0254679732,-0.1013296321,0.0878035203,0.1649087965,0.234488979,-0.1868000031,-0.0388952531,0.2961952388,-0.471298635,-0.020624157,-0.1165809706,-0.1584409326,-0.0656100363,0.2926974297,-0.1955597699,-0.0210692696,-0.1047487259,0.0503780842,-0.2902586162,0.2498258352,0.1887843162,-0.1815607697,-0.3060776591,-0.2786382437,-0.3781633973,-0.2545593679,-0.1980644614,-0.0352448411,0.259896487,0.1381271034,-0.035641931,0.1969049722,0.0324109942,0.0291325934,0.0739200711,0.1020731851,-0.4914737642,0.0756241903,0.1766169369,0.1397940218,0.0936074927,-0.5097525716,-0.3088516891,-0.1815595478,-0.0467219353,-0.0911011845,0.2209638953,0.2755781412,0.6003807783,0.7884604931,-0.1086812764,-0.2611299753,-0.0246224497,-0.2064298391,0.062554054,-0.0551166795,-0.0832696483,-0.2104490995,0.3017953336,0.3111482859,-0.2844796777,0.3605870306,-0.3408155739,-0.2436244637,-0.1466899663,0.0409392416,-0.2627132535,-0.0334496722,-0.0489040539,0.400144726,0.0136861205,0.2550777495,0.0295581743,0.0588340871,-0.3597678244,-0.0254772417,0.3839750588,0.1979362965,-0.2956695557,0.4079867899,0.1053552628,0.0487198308,0.0305683669,-0.3500054777,0.2871949971,0.1400469393,-0.3088756502,-0.0610060729,0.2714013457,0.0102574881,0.0061118952,0.1072329655,0.4342740178,0.2631801069,0.0932724327,-0.4791612923,-0.2532896101]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Also, if I delete the extracted files they need to be re-extracted again instead of loading from the Arrow cache files","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":21,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nAlso, if I delete the extracted files they need to be re-extracted again instead of loading from the Arrow cache files","embeddings":[-0.0280625913,-0.1268454641,-0.1462638825,0.1517361701,-0.1755876094,0.1943831593,-0.2197522521,0.2515498996,0.1799505502,0.1638493091,0.1056418791,0.4813863039,-0.2407226562,-0.1040748432,-0.1887400746,-0.0030702341,-0.2355995178,0.2889693975,-0.007609963,0.1621083319,-0.1351312846,0.0150677748,0.0465388186,-0.1541214883,0.0537764616,-0.4090359211,0.2061123103,-0.2602271438,-0.1102860421,-0.3523688614,-0.0030772036,0.5342466831,0.2766074836,0.141957894,-0.0001208869,-0.0142732738,0.2875986993,0.0723435506,-0.4089965522,0.2255584151,-0.243676886,-0.1031287462,0.2598846257,-0.1265426427,0.2946948111,-0.4084398448,-0.2041454315,-0.6497346759,0.5037788749,-0.2386162877,0.0973825306,0.009610286,-0.5454075933,0.1060440689,0.3338082433,0.3475065231,-0.1310423166,-0.0081458325,0.4774487019,0.100996539,0.2767952979,0.3542622626,-0.0738362148,-0.2929219007,0.338668555,-0.0257929079,-0.0745974183,-0.2595772743,0.2915354371,0.0246846136,0.6543719769,-0.3459971547,-0.0916249156,-0.3327183425,0.1247224882,-0.586995244,0.0834856406,0.1749043018,0.0470301099,0.2470870614,-0.1042246521,-0.4791462421,-0.3593364954,-0.1478547901,0.6762899756,-0.4306482971,-0.0751291513,-0.1268769652,0.5420335531,0.1520411521,0.2174185514,-0.1760035306,-0.3039318025,0.4096335173,-0.0656304583,-0.1792774647,-0.4102030396,-0.2998050451,0.1566981822,0.1752879173,0.4254452884,-0.0052143037,0.1747297347,0.0212203227,0.2458343506,0.3867214024,0.0701313987,-0.0618336499,0.1884942502,0.1172566861,0.1380652636,-0.0492788553,0.0289419107,-0.0381849222,0.2580288351,0.0199359134,-0.038412381,-0.1320509911,-0.0718108863,0.0647786483,0.2061200589,-0.1974595934,0.086588949,0.1973827928,0.2360136807,0.0340708494,0.0615368113,0.2982646227,0.0124634709,-0.3605464697,0.1215328351,0.1221913174,-0.3169591129,-0.2377638072,0.4114531279,-0.0058699609,-0.0418000557,-0.1277221888,-0.3367883563,0.0653983206,0.5076660514,-0.1196554154,0.2788880467,0.277174741,-0.1964025646,-0.2270881534,-0.2803883255,-0.0114324261,-0.1321524382,0.5104100704,-0.1253011823,-0.2173408419,-0.066548638,0.1125036553,0.1528291404,0.2723378241,-0.5884199142,0.1173697188,0.1026504189,0.0425826795,0.0408399254,0.3486083448,0.0186311677,-0.0803235322,-0.3321526349,0.2598174214,-0.6571017504,-0.0888553485,-0.3545111716,-0.2567665279,0.2115150094,0.112507388,-0.1175073907,0.2376223356,-0.1577705294,0.1594096273,0.3421454728,-0.0935675725,-0.3924858272,0.1546457261,0.0958269089,0.1922517717,0.3176071346,0.3045813739,0.384672761,0.0386508517,-0.0686706528,0.2615580857,0.0880157351,-0.10837017,-0.3247458935,-0.4044742584,0.0343365036,0.2238271981,-0.0033786683,0.0553450696,0.2231910229,-0.0305488575,0.1710044295,0.149852097,0.1663506776,0.1606840342,0.1390224993,-0.2714569867,-0.2364161164,0.1407706439,-0.6055968404,-0.0088958433,-0.5798181891,-0.3777634799,-0.1828524619,-0.3519939184,0.1315580755,0.0131068761,-0.1980336308,0.1989232302,0.0346161276,-0.0616230592,0.2489040345,-0.0277927574,-0.4232954681,-0.0171468463,-0.0940821245,-0.0000205979,-0.4419443011,0.1254571229,0.1512196362,-0.259088397,-0.019973442,0.0140808979,-0.2875122428,-0.4913887978,-0.1318501681,0.2647311985,0.3056980669,0.2912358046,0.2880557775,0.2514718771,0.1814325601,0.1652215123,0.1045260057,0.1599608064,-0.0176257435,0.0884636194,-0.3933174014,0.3887653053,-0.2242788523,-0.0352116078,-0.1823385656,-0.3922522664,0.2207522392,-0.1020581275,0.1129981726,-0.5385864377,0.0472975224,0.3436099291,-0.1163012832,0.0287124459,0.1037619263,-0.1153912768,-0.0062079015,-0.044994086,-0.1471545547,0.3006251156,-0.0148150921,-0.1625528634,0.0789365247,0.3862743676,0.3511201739,0.1766192466,-0.1846850961,-0.1630650908,0.4579220116,-0.009584832,0.1368839145,0.0368465967,0.018498214,0.3677564859,0.0602166951,-0.0062987711,-0.2424093634,-0.0591289215,0.1128261387,0.2101317495,0.2819328308,-0.2630928755,-0.276941061,-0.4187943935,0.4745363891,0.3225945234,0.258238852,-0.0183715001,0.5235859752,0.0613533333,-0.4129450619,0.06954474,-0.0198271107,0.668682754,-0.1250792444,-0.2156380862,-0.4043356776,-0.1886476129,0.1578585804,0.0167335793,0.3731075823,-0.101245068,0.5151019096,0.1303833127,0.1142640114,-0.0901175216,-0.0323509462,0.0783112571,0.3688560426,-0.0753833354,0.0812374055,0.1757234782,-0.0526804589,0.1216848269,-0.1933170557,-0.2980662882,-0.124831304,0.1577711999,0.0885916725,-0.0213484634,-0.2257019877,-0.0321052261,-0.1681333482,-0.3950457871,0.2272086293,-0.0255394801,0.0000476393,0.1141305417,0.2313457876,0.0122267203,0.0164264757,0.0707674026,-0.0095825437,-0.588660121,0.132973209,-0.208014816,0.0741187483,0.2229170948,0.0186059121,-0.0014545916,0.3487145007,-0.3909091055,0.0475146547,0.0179734323,0.1181501076,-0.3372391164,-0.0764741078,0.214870289,0.2313204259,-0.0852172971,0.0311033241,0.1433147341,0.2431013286,0.3958516717,0.1196727753,0.0834100172,0.2145751119,0.1187930927,0.3881933689,-0.2566867471,0.0381416716,0.3195043206,0.3901723027,0.4156238735,-0.1908151805,0.2777892649,-0.4188162088,0.0556625202,-0.0858858898,0.1190642938,0.0261614453,-0.1459260583,0.1614796817,-0.2239902467,-0.1283926666,-0.0923646688,0.3123985529,-0.2337031066,0.0603834316,-0.005724933,-0.2973092496,-0.0169660896,-0.0230034702,0.1660930812,0.1780434698,0.4248476624,-0.1065800041,-0.2460834384,-0.0765033066,-0.1745066643,0.3237021565,-0.2353146374,-0.2074514031,-0.0801080391,0.0300733391,0.1610285938,0.2512604594,0.3416738808,-0.1118747219,-0.0361465141,-0.1675199866,0.0489435717,-0.2440116554,-0.0252702888,-0.0103160394,0.0381948017,-0.0038875863,0.4042293131,-0.2591198683,-0.1607069522,-0.2055034339,0.5759979486,-0.2591153085,-0.091646485,-0.0329354405,-0.1711935252,-0.5231106281,0.0666255504,0.018275341,-0.1166655868,0.1697268337,-0.3050328791,-0.0972613767,-0.3248197436,0.0477094352,0.0438211299,0.5563672781,-0.2645012438,0.0622149333,-0.1286518872,0.1483029723,0.5444576144,0.2043477446,-0.0919583291,0.0192476548,0.0162746143,-0.3969781697,0.1736557037,0.1599513888,-0.2013671398,0.3397914469,-0.2212539017,0.5212981105,0.0699868947,-0.0226430334,-0.2058652192,-0.303817451,-0.0185768213,-0.1665754616,0.1150055453,0.0280127861,-0.1756879985,0.4709266722,-0.105457522,-0.2438369393,0.4904084802,0.0997257084,1.1869287491,-0.0977287665,-0.037799228,-0.2128304392,-0.2820965052,0.2717681229,-0.5684899688,0.1390948892,0.0268349424,-0.0716800019,0.1070432886,-0.0506122969,0.0369741991,0.2111321837,-0.248174578,0.0501421019,-0.3473038077,-0.1581049562,-0.2884771228,0.3806913793,-0.0826263726,-0.2182884365,-0.048971504,-0.023917431,0.0337180682,-0.0894671381,-0.2412536293,-0.0738157183,0.3701569736,-0.0102715101,0.3230470121,-0.1315903664,-0.2649998069,0.3628649116,-0.1891300231,-0.1291918904,-0.2284797579,0.044004377,-0.2324737608,0.0319456719,0.1362996101,0.0124497525,0.0906266347,-0.0633031279,-0.4203027785,0.0675697774,-0.1030760333,-0.1273811311,0.0338451266,0.1527559757,-0.2764265835,-0.3605756164,0.151114136,0.1476814747,-0.1120914966,0.0539426543,-0.2646898329,-0.1267174035,-0.5518767238,0.0999613777,0.0886899605,0.2987122238,-0.1719677895,0.2865864038,-0.1654573232,-0.1470646709,-0.2052546442,-0.1596971303,0.2845774889,-0.1546224058,0.2459614426,-0.2449815869,-0.2290409207,-0.4594869316,-0.0160061289,-0.0629504994,-0.3975654244,0.4622119069,0.0260998234,-0.1692353934,0.0772049204,-0.0973463431,0.0437741168,0.0307617281,-0.1405878812,-0.1074351296,-0.4702298045,0.2746435404,0.2288825363,0.5749309659,0.0355819128,-0.0146400621,-0.123010546,0.1055092365,-0.2284840941,0.2524847388,0.2734741867,0.1581341773,-0.0451621786,0.2865183055,0.0547466539,-0.2374959439,-0.0978449136,0.0568663515,-0.1206638068,-0.1586706489,-0.107133694,0.112603128,-0.1398201585,-0.1054832786,-0.1901089996,-0.2603553534,0.0911482126,-0.1766229868,0.2107422501,-0.2486180514,-0.140925765,-0.1472358108,0.4561414421,-0.0292984024,-0.0877420828,0.173032999,-0.0714690462,0.350318104,-0.0195524283,-0.0122248968,-0.0435412228,-0.0748710707,-0.3222399652,0.0677052066,0.3231784105,0.0681726262,-0.0155770341,-0.0005591909,-0.0903892964,0.0692130923,0.0420602784,0.2573955953,-0.0696018264,-0.1179596335,-0.0512681417,0.1871215105,-0.2150454819,-0.1542569697,-0.0635144934,0.0174148325,-0.2230870873,0.2765961587,0.3653716445,0.0691098645,0.0669683591,0.0881156102,0.2617764473,-0.2271976918,0.3391412199,0.320699662,0.0553323664,-0.2411572039,-0.1346071064,0.2819794416,0.1234779358,-0.162291646,-0.1784497797,0.3938885033,0.5958599448,0.3432156444,-0.5122553706,-0.2275402397,-0.294673115,0.6062335968,-0.0741070509,0.4444570541,0.4342083633,0.1758130044,0.318977654,-0.2883318961,-0.2041713893,0.4129080474,0.0876334831,0.0666524097,0.028796643,-0.1246611401,0.1117320135,0.1013873965,0.1449334323,0.4016484916,-0.1080211625,0.3295115829,0.0834280401,0.2578529716,0.5938124657,0.1851592511,0.4273990989,-0.1433627605,-0.2068376839,0.2007970363,0.2718924284,0.049302958,-0.1012151912,-0.0775129348,0.1623560339,-0.364290446,-0.0038273663,0.0908987224,0.0632570982,-0.2624429166,0.4751577377,0.2595592439,-0.2415838391,-0.0355233885,0.0558042042,-0.082285203,0.1704405844,0.0924598351,0.203028053,-0.0839178562,-0.2313925177,0.3846432567,-0.4732673168,0.0324267,-0.1434434503,-0.0942875072,0.0122139221,0.2725045383,-0.2859966159,0.0040995516,-0.0526313446,0.0566739403,-0.2232700735,0.2803287208,0.2043563128,-0.3420989811,-0.2589809597,-0.3161487579,-0.2175919563,-0.2054595202,-0.2143557966,0.0308525115,0.2443353385,0.2561444342,-0.0242517516,0.2262441367,0.2455408275,-0.1379126012,-0.0417927653,0.0555960834,-0.4732873142,0.1433893591,0.3838844001,0.2196660787,0.1469879597,-0.4231848121,-0.2671890557,0.0411410742,-0.0139269261,-0.1904661059,0.2165529877,0.1297936738,0.7412775159,0.6710228324,-0.1407821774,-0.3098009229,-0.086834237,-0.1885192245,-0.0123669924,-0.125013262,0.0300796647,-0.1888549179,0.47497648,0.2099930793,-0.2864185274,0.2727656066,-0.2861827016,-0.2152543515,-0.1712091863,0.2367955297,-0.229683876,-0.0501833372,0.0234882534,0.4224823117,-0.0384043828,0.1940822601,0.0690301806,0.0839979053,-0.3601064086,0.0329036787,0.3062989712,0.3926432133,-0.3070211709,0.5346334577,0.1490335166,0.085221298,-0.0019340154,-0.2796180248,0.2397037446,0.1016343683,-0.2820937335,0.043675106,0.2897916138,-0.0310820732,-0.0459957346,0.1021824256,0.2989760041,0.3729035854,0.040529348,-0.4589451253,-0.264326483]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"I think we already opened an issue about this topic (suggested by @stas00): duplicated of #2481?\r\n\r\nThis is in our TODO list... \ud83d\ude05 ","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":23,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nI think we already opened an issue about this topic (suggested by @stas00): duplicated of #2481?\r\n\r\nThis is in our TODO list... \ud83d\ude05 ","embeddings":[-0.0192103572,-0.1269763708,-0.1481829584,0.0798467398,-0.1321689636,0.1765157878,-0.1987018585,0.2431635112,0.2124012709,0.2022783458,0.161542356,0.4751735032,-0.22158885,-0.0703336075,-0.2046113312,0.0402210467,-0.1827221364,0.2646833658,0.002185296,0.182455495,-0.1406535357,0.0362731703,0.0394842103,-0.1689869314,0.0975508094,-0.376784265,0.2147901207,-0.2744055688,-0.120177038,-0.3621692955,0.042904567,0.5484212637,0.2381388396,0.1525383592,-0.0001205524,-0.0380499102,0.2979128957,0.1006247252,-0.4042674899,0.1757793725,-0.2399449795,-0.1098071039,0.2304793,-0.1402706802,0.2754366398,-0.3210353255,-0.1625757962,-0.6223289967,0.4326825738,-0.2468685657,0.0960136876,0.0167482756,-0.5319135785,0.0954349115,0.3050426543,0.4185585976,-0.1824688315,-0.0573912151,0.5388312936,0.0341586135,0.2726577818,0.3712629378,-0.0588131174,-0.2781784236,0.3521537483,-0.0666618049,-0.0870359689,-0.2991166711,0.2780787945,-0.0211910494,0.6806836128,-0.3543694913,-0.1248807833,-0.3069045842,0.0766370445,-0.5461723804,0.1264318228,0.1799128205,-0.0215025805,0.2723221481,-0.1678588539,-0.4257777035,-0.3638321161,-0.1120299175,0.6866944432,-0.3886205256,-0.0511001199,-0.144225657,0.5321463346,0.1710869372,0.2038838565,-0.0841110423,-0.2778005898,0.416785866,-0.0302322358,-0.1976915747,-0.3469560444,-0.3571130335,0.248250246,0.1723577678,0.4253973067,-0.0317957699,0.1785061657,0.0627935454,0.2811262012,0.3864467442,0.0619176887,-0.1031351537,0.2124250531,0.1001006886,0.1506522894,-0.0452674963,0.1101079658,-0.0106062712,0.2505978346,0.0307717994,-0.0374104492,-0.1456220299,-0.0515751578,0.0740639418,0.202386871,-0.1960340738,0.0666680411,0.1708460301,0.2334365845,0.0534014665,0.0713056028,0.3021495342,-0.0025234201,-0.3367939889,0.125625506,0.1341754198,-0.3137151003,-0.2757600248,0.4017677307,-0.0329739787,-0.0375755578,-0.0856447443,-0.3078609407,0.0509177782,0.5031169057,-0.0553310178,0.2366694808,0.3152055144,-0.1929736733,-0.2496155202,-0.2744360566,-0.0515336432,-0.1477893591,0.4894921482,-0.1400498599,-0.2936847806,-0.1132905185,0.1049487591,0.1574423015,0.2329837084,-0.5811308622,0.1551566273,0.0752395317,0.0513779856,-0.0294898245,0.4142984748,-0.0133437542,-0.0749279633,-0.3167106509,0.2698732018,-0.6963501573,-0.0778401792,-0.3436824083,-0.2515718341,0.2094901055,0.1533285379,-0.1195058078,0.182037279,-0.1769447327,0.1597190499,0.2982476056,-0.103884764,-0.3767288625,0.1588909626,0.0618305504,0.1861554384,0.3381636441,0.3027474582,0.3510040343,0.002831195,-0.0717588663,0.3363138437,0.0428750925,-0.0459031463,-0.3771969378,-0.42586115,0.1138878167,0.216146037,0.018699687,0.0835732073,0.2645459175,0.0252894498,0.1387610734,0.1308916509,0.1084435284,0.1359111518,0.1836431921,-0.2509692907,-0.235962227,0.1677518189,-0.5864220262,-0.0330729447,-0.5517913103,-0.2943542004,-0.143564716,-0.3622748852,0.139668107,0.0530983694,-0.147879675,0.1732763499,0.0387975089,-0.088631615,0.2117042392,-0.0261978768,-0.4154306948,-0.0330969319,-0.1128378361,0.002000845,-0.464940995,0.1039662734,0.0835514069,-0.2256426215,-0.0497684032,0.0020140291,-0.25286901,-0.5000703931,-0.1524206996,0.2905792296,0.2850807011,0.2995112836,0.2822110653,0.2653039992,0.2180715054,0.1882403195,0.047434736,0.1773793548,-0.045733124,0.0735766143,-0.4050086439,0.34395051,-0.2409024686,-0.0849303231,-0.2035918385,-0.3892908096,0.251600951,-0.1002887413,0.0859725028,-0.4930270314,0.0859950036,0.3708960414,-0.1917723566,0.0199163519,0.1214128882,-0.1024174243,0.0037558398,-0.0782458112,-0.1640661061,0.3610450029,0.0768096149,-0.2198453695,0.070446752,0.3702918589,0.3388828337,0.1903007776,-0.2203627527,-0.1675284952,0.4414020777,0.0176829249,0.1687619388,0.0462639183,-0.0339810736,0.3442345858,0.1006004661,-0.0012359292,-0.2634166479,-0.0832917765,0.1024767011,0.2208114564,0.261554122,-0.2776215672,-0.243185997,-0.4085784554,0.5601942539,0.3427139223,0.2305352837,-0.0628957376,0.5425235033,0.0828968138,-0.4694440961,0.0687304214,0.0326804258,0.7095785737,-0.1634512544,-0.1467918754,-0.4266526103,-0.1652532071,0.1574263722,-0.0002175611,0.387247473,-0.1314132363,0.5233255625,0.1060379297,0.0948024616,-0.0910519063,-0.1202437282,0.0795937777,0.3784058094,0.0271689575,0.1042734832,0.1606732756,-0.0195349399,0.058591187,-0.2174522281,-0.2745068669,-0.1152909324,0.1672182977,0.1256192774,0.0023228175,-0.2697550356,-0.0511664897,-0.1811908185,-0.4072161615,0.2652530372,-0.0908301175,0.0454159677,0.1008219495,0.2078517377,-0.0895771459,0.0180449113,0.0879817232,-0.070533596,-0.706107676,0.1135700494,-0.1956966817,0.0273814555,0.2443943769,0.0343266912,-0.0326903574,0.324528873,-0.4147774279,-0.0504501574,0.0886664689,0.1621686369,-0.3129668236,-0.1039424092,0.1685178578,0.2275385857,-0.0814688727,0.0571502298,0.0484278426,0.1911776662,0.3824151456,0.1352135092,0.1176510155,0.2851807773,0.2092827857,0.3362184167,-0.2803320885,0.0880405679,0.3748086393,0.3983893096,0.418654114,-0.221598804,0.1879457384,-0.3908182681,0.0196481366,-0.1622592062,0.1251033247,0.0533163436,-0.2120589912,0.1556338817,-0.2383015603,-0.1149868891,-0.0837213993,0.2974512577,-0.2194238901,0.0257989522,0.0152262766,-0.2889419198,-0.0783254206,-0.044106774,0.1652411968,0.2261772305,0.4313418567,-0.075990431,-0.2336569577,-0.0816135705,-0.1685784906,0.3073792756,-0.2065442652,-0.1619284302,-0.1456630826,0.0682426095,0.1803955436,0.2353226691,0.4696421027,-0.0961222947,-0.0197155066,-0.2219821662,0.0241034739,-0.1631164551,0.0119683146,0.0207884293,0.012108787,0.1047578827,0.4172722399,-0.2507812083,-0.1821791679,-0.2067555338,0.5407698154,-0.2291523367,-0.1452026516,-0.0057223034,-0.2123621404,-0.5202825665,0.0625925139,0.0947567001,-0.1236630157,0.2344865501,-0.3365588188,-0.0600591451,-0.3407364786,0.0559318773,0.0333825201,0.5054509044,-0.2519854903,0.0512894765,-0.1779838651,0.1141240224,0.4623526931,0.175274983,-0.1009979621,0.0274560712,0.0044709137,-0.393412292,0.1438115537,0.150895685,-0.1485908628,0.2990285754,-0.2558405697,0.5546305776,0.0296257231,-0.0601641089,-0.1810203344,-0.3058012128,-0.0189943239,-0.1649739146,0.1217677519,0.026150655,-0.1800437123,0.4581396282,-0.1303936541,-0.2181837261,0.5202693343,0.0637353361,1.1716758013,-0.0742964447,-0.0621443726,-0.2309649885,-0.2570972443,0.3165613413,-0.6766657829,0.1244023144,0.073421061,-0.1243324354,0.1023870409,-0.0162254497,0.0199672543,0.2219273001,-0.2478419095,0.0639300197,-0.3001569808,-0.1765232831,-0.2858092487,0.3896434307,-0.1006885096,-0.2490497679,-0.0716716945,-0.0173443053,0.015346989,-0.0213925876,-0.221571207,-0.059590701,0.4131851196,-0.0172294304,0.246873945,-0.1121237874,-0.2360199839,0.3648224175,-0.1647970676,-0.0689614937,-0.2324288189,0.0299131796,-0.17332156,0.0942813754,0.1340536624,-0.0163538698,0.0748664513,-0.0172675382,-0.3920020461,0.0818214938,-0.054944057,-0.1253946871,0.0464399606,0.1012335718,-0.2979835868,-0.3712752461,0.0831151456,0.0987869352,-0.1517133117,0.0230122637,-0.2844040692,-0.1307133585,-0.5083082914,0.0835138857,0.091868259,0.3376482427,-0.1956746429,0.3320388198,-0.1939457506,-0.1492915004,-0.217947796,-0.1215124205,0.2511497438,-0.1914851815,0.1788420528,-0.2094887495,-0.24187693,-0.431669414,-0.0149513539,-0.0827951878,-0.3865464628,0.4302395284,0.049406372,-0.2330640405,0.0752275661,-0.1367808878,0.0579775162,0.0033783542,-0.1083917469,-0.1550700068,-0.4857673943,0.2241017222,0.217526406,0.568844974,0.0788131356,0.0453644954,-0.1192526072,0.0687630624,-0.2306442112,0.2507713437,0.2330899984,0.1651658267,-0.0121291894,0.2815650403,0.0334079526,-0.2585384846,-0.0945775807,0.0784750953,-0.1148144528,-0.1441342831,-0.0973203927,0.1215382144,-0.0991292596,-0.0808286294,-0.1344829202,-0.2419474274,0.0778194219,-0.1644069105,0.2182532251,-0.1838774681,-0.1200954467,-0.1128276065,0.409923017,-0.0394897647,-0.0061417767,0.1524103582,-0.0428900309,0.3734461367,-0.0446032882,0.0133972345,-0.0645923913,-0.0591993965,-0.3752082884,0.0733758733,0.2588769794,0.1090116128,-0.0003994321,0.0225511659,-0.0875493586,0.0966071188,0.039910201,0.1827138066,-0.116992034,-0.0676395446,-0.0235159285,0.1917645335,-0.1976939887,-0.1988266557,-0.0608728044,0.0752246603,-0.2139880508,0.247860074,0.4000194073,0.0741869584,0.1304767281,0.1099632904,0.3531478941,-0.2766364813,0.319308579,0.310467124,-0.0166351292,-0.2100169361,-0.1769705117,0.2833026946,0.1194732562,-0.1427136511,-0.1698103845,0.3863299191,0.600713253,0.3873370588,-0.487527281,-0.1816841662,-0.2639196515,0.6182791591,-0.0514408872,0.4884519279,0.4662978649,0.18845433,0.2922814488,-0.2606734931,-0.2065418363,0.4309628606,0.0768073872,0.0376479402,-0.0007762157,-0.1363193989,0.1158710569,0.0870209783,0.1426931769,0.3840877116,-0.1165056899,0.3271659911,0.065534927,0.288872987,0.5928450823,0.1578070968,0.4371984005,-0.1594477892,-0.1655697674,0.1945153326,0.2201560289,0.0693242475,-0.0579906255,-0.0156213678,0.1615114212,-0.3188797235,-0.0212751571,0.0590967909,0.053390082,-0.2769364417,0.4711276591,0.2056374252,-0.2247938216,-0.0590540357,0.0458208285,-0.0872469619,0.1757305712,0.1273068339,0.2065495253,-0.0734714046,-0.2261738926,0.375502497,-0.434176445,0.0296866558,-0.1775697321,-0.1082248911,-0.0197179746,0.2628769875,-0.2581289113,0.0527019501,-0.0655056387,0.0659321621,-0.1818637848,0.2217471302,0.1854013652,-0.4435361922,-0.2482797503,-0.2721306086,-0.2667448819,-0.1807051748,-0.1906944066,0.0515160151,0.3041940331,0.2105135769,-0.0289457012,0.2490502894,0.1911314577,-0.1328953356,-0.0568950139,0.1076925546,-0.4754080176,0.1072263196,0.3888148963,0.2782554328,0.1642229557,-0.3787922561,-0.2844346762,0.0037730832,-0.0145655777,-0.251518935,0.276555419,0.1154202446,0.7933052778,0.5852683783,-0.1099351123,-0.3783088923,-0.0988554358,-0.2023891956,0.0157392081,-0.129327938,0.0116771394,-0.1971288323,0.450674355,0.1910800487,-0.2832457721,0.2106015831,-0.2632028162,-0.2245255262,-0.2430842221,0.2408836484,-0.2843170166,0.0277252458,0.0219361279,0.4092614651,-0.0663308576,0.1995097399,0.0634391233,0.1137813553,-0.3624653518,-0.017779734,0.2839457691,0.3558633327,-0.3132498264,0.5239496827,0.0752069801,0.0631889403,-0.0131416433,-0.2903980613,0.2310442924,0.0778578669,-0.2676573098,0.0424916372,0.3427077234,-0.0503719598,-0.0736865699,0.1060918048,0.2718821466,0.4296522737,0.020260755,-0.4044409096,-0.2552907765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"I think the deletion of each extracted file could be implemented in our CacheManager and ExtractManager (once merged to master: #2295, #2277). \ud83d\ude09 ","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":23,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nI think the deletion of each extracted file could be implemented in our CacheManager and ExtractManager (once merged to master: #2295, #2277). \ud83d\ude09 ","embeddings":[-0.0226012003,-0.1121031269,-0.1641411185,0.1639864743,-0.140830189,0.156821683,-0.1947341263,0.2873069644,0.2321140617,0.1775878519,0.0999825671,0.4532558918,-0.2333150804,-0.1064827368,-0.2285246402,0.0652216226,-0.218204692,0.2731807232,-0.0273976568,0.2025793642,-0.0816403404,0.0862377882,0.0794725493,-0.1437506974,0.0515122674,-0.4186532199,0.2156175673,-0.2527047694,-0.0921536312,-0.3959024847,0.0014364745,0.5230810046,0.2234824449,0.1859763563,-0.0001158199,-0.0553465709,0.305426836,0.0845195577,-0.3473008871,0.1930393875,-0.1883360595,-0.1002468243,0.2181223482,-0.1417431533,0.2566103041,-0.304568857,-0.1959095746,-0.6068360806,0.4527910054,-0.2366441041,0.1494680345,-0.0148590161,-0.5173233747,0.1346339583,0.3028507829,0.3194510937,-0.110709466,-0.047523845,0.424187094,0.0751272961,0.2008986175,0.3739815652,-0.1134387627,-0.2500633895,0.3243033588,-0.0099846674,-0.1423219591,-0.2449120879,0.3142209947,0.0150832366,0.5898211598,-0.3973073959,-0.0776183307,-0.2895053327,0.0565595329,-0.5846502185,0.1107224002,0.1677326113,0.0058494452,0.2414092422,-0.1400053054,-0.4298626781,-0.3438484669,-0.190056175,0.6732532382,-0.3742928207,-0.0949344262,-0.1413282603,0.453468442,0.1464481205,0.1862225384,-0.12596111,-0.2888209522,0.3707067966,-0.0712986365,-0.2018734813,-0.3689159453,-0.3332546949,0.2370148301,0.1708972007,0.4238263369,0.051885318,0.2351701111,0.0633939952,0.263776511,0.3954226673,0.0883608982,-0.1276811212,0.2007574886,0.111225605,0.0768835247,-0.0542461537,0.092893675,0.0000739668,0.3071662784,0.0206521172,-0.0722566023,-0.155696705,-0.0429361723,0.0741358697,0.2563768029,-0.1725345105,0.0869790241,0.2019751072,0.153233707,0.0464817397,0.0632906482,0.2291132659,0.0240844488,-0.2961390913,0.0856631249,0.1718940437,-0.3100761473,-0.2661429644,0.3942476213,-0.0600514002,-0.0353220105,-0.1283133924,-0.3043464422,0.0390801802,0.5187656283,-0.1317350119,0.2679392695,0.2626909912,-0.1665892303,-0.2599322498,-0.2856931984,-0.0274822693,-0.1568219364,0.4221943617,-0.1159831136,-0.2432418466,-0.0390984714,0.1673652232,0.1655025631,0.3048390448,-0.4949506819,0.1325614452,0.155432269,0.0527000874,-0.0254455041,0.3915011585,-0.0121706454,-0.1084474176,-0.3085303009,0.2602828741,-0.6160305142,-0.0765488148,-0.2943627536,-0.2651220262,0.193464458,0.137913838,-0.0625761971,0.1802577525,-0.1758537292,0.1568830758,0.3047702014,-0.104730159,-0.3668222725,0.1630241573,0.0668906868,0.1986334622,0.3475049436,0.2767180204,0.3986903429,-0.0060511143,-0.0315092094,0.2765131295,0.0772935599,-0.0692836791,-0.3465442955,-0.3954345286,0.0166981407,0.209360674,0.0411174819,0.0486390293,0.2571693957,0.0145428358,0.161527589,0.1562286019,0.1540129185,0.1770214885,0.1292431802,-0.2629849017,-0.217876941,0.1299288571,-0.5425730348,0.0111168921,-0.5635592341,-0.3578417897,-0.1681194603,-0.3451923728,0.1446575075,0.0259726755,-0.1506370753,0.1525574774,0.1201664135,-0.030617103,0.2321802378,-0.0576329865,-0.414154768,0.0821246356,-0.0787246674,0.0214121081,-0.4709240794,0.0398496687,0.1304672509,-0.2229179591,-0.0247545522,0.0109062092,-0.27362746,-0.5044941306,-0.1394358277,0.2698444426,0.2659398913,0.2694551647,0.2403144985,0.2347733825,0.1686521471,0.2529337108,0.078090407,0.1655546427,-0.0219974983,0.0573894754,-0.3692435324,0.3941814303,-0.1629272848,-0.0680424571,-0.2091500908,-0.3963089585,0.2487940341,-0.1182916015,0.1494150758,-0.5282139182,0.0754013509,0.3249076605,-0.0978747457,0.012904698,0.1258081645,-0.0843765736,-0.0029913154,0.0030726942,-0.0814075097,0.3343492746,-0.0005701191,-0.2064735442,0.0179730132,0.3947648704,0.3731953502,0.1863262057,-0.1443179399,-0.1430867463,0.4674766958,-0.0077398438,0.1551846713,0.0356968753,0.0012495855,0.3792584538,0.0827416554,0.009578635,-0.2653331161,-0.090862073,0.1035449356,0.1854015291,0.3281438947,-0.2157000303,-0.2645330131,-0.3503082991,0.4957950711,0.3474987447,0.2267124355,0.0012920999,0.529710412,0.1080581471,-0.4376138747,0.0794259384,0.0360067375,0.634299159,-0.1338887513,-0.1276343018,-0.4051322043,-0.1930312961,0.1563112885,0.0413373262,0.3700557947,-0.0946758166,0.4919627607,0.1280309707,0.1074846536,-0.1013563871,-0.0828962848,0.0904887915,0.3652791381,-0.0422652364,0.1250080317,0.1520941705,-0.0196174011,0.1258698851,-0.216358915,-0.3381376863,-0.1388873011,0.2075248063,0.085205473,-0.0198435262,-0.2818321288,-0.0219430272,-0.1901815236,-0.4262129366,0.2239803672,-0.060232833,0.0722541735,0.0535561889,0.1596736014,-0.0304912999,0.0300519224,0.0889111683,-0.0778837427,-0.6628459096,0.0983512327,-0.1868419647,0.037983451,0.2425360233,0.0444731936,0.0063195541,0.2807715535,-0.387660861,-0.0573442318,0.0818394944,0.104945749,-0.3130922616,-0.118442066,0.2026218474,0.2476184517,-0.1233276799,-0.0007635195,0.1013074517,0.2300355136,0.3424203098,0.1033401042,0.1240427196,0.3077454567,0.2200436592,0.3042977154,-0.2809365988,0.0443847291,0.333258599,0.4032059312,0.3957530856,-0.199958846,0.2320665717,-0.3902228773,-0.0107682878,-0.1272917986,0.1531163752,0.0192082021,-0.237198174,0.1621901393,-0.2351834029,-0.1076337025,-0.0924572572,0.3103538156,-0.1896032393,0.0137225073,0.0337256268,-0.2891180217,-0.0444940105,-0.028092796,0.1488057226,0.166950956,0.3563663363,-0.0836421326,-0.2635137439,-0.0878591239,-0.2644632459,0.3332786262,-0.2223545164,-0.1899109334,-0.1091748253,0.0586028583,0.1371446848,0.1942606419,0.3985287845,-0.0987224802,-0.0259500612,-0.1894541234,-0.0066748327,-0.157893464,0.034721151,0.0147041455,0.0553894117,0.0625592545,0.3513251841,-0.3022957444,-0.2003955394,-0.2519724667,0.5415742397,-0.2682743669,-0.1403572708,-0.0532955416,-0.2300033569,-0.515253067,0.0718514845,0.0248224735,-0.1205939054,0.1934349388,-0.317803055,-0.0970483869,-0.3147116899,0.0356624983,0.0134953726,0.5749182105,-0.3080909252,0.0529579893,-0.1670598984,0.1460404992,0.4614255428,0.1831458062,-0.0740575194,0.0567422435,0.0455558524,-0.360561192,0.0747279301,0.1328524351,-0.1521331519,0.3234331012,-0.237854436,0.5641589761,0.0926813483,-0.0550357513,-0.2353071868,-0.2607458234,-0.019143194,-0.1411494911,0.0838327333,0.0453281738,-0.1757596582,0.4483439922,-0.2364695221,-0.2719984651,0.4784607589,0.0938816443,1.15234375,-0.0630939156,-0.0443608947,-0.2481875569,-0.2877420783,0.2968457341,-0.6555526257,0.1433559507,-0.0354745463,-0.1166598722,0.1045608446,0.0114563173,0.0144009078,0.1993405074,-0.252481848,0.0956661925,-0.287977457,-0.1406229287,-0.2822276056,0.3869046867,-0.1198891476,-0.2729753554,-0.0524286292,0.0311328713,0.0025455994,-0.0838767961,-0.2410697341,-0.1141219512,0.4444569945,-0.0500526838,0.3152216971,-0.0998766944,-0.2177527696,0.4091772735,-0.1741435081,-0.1087461337,-0.322709918,-0.04315136,-0.1769373715,0.0509370416,0.1713897139,-0.0063803578,0.1149518415,-0.0327295102,-0.4030062258,0.0416042879,-0.1148713902,-0.1488921195,0.0588002466,0.1078942269,-0.2950463295,-0.3856203556,0.1859516054,0.141057834,-0.1611456871,0.081183672,-0.2483866066,-0.106498614,-0.525357604,0.0941527486,0.1336240619,0.3176446557,-0.2119374126,0.3105860054,-0.1847433895,-0.1575226337,-0.2423824519,-0.1405697465,0.3364911973,-0.1629328728,0.2128500491,-0.2056051344,-0.250236094,-0.4690576494,-0.0277926885,-0.047667481,-0.3634729683,0.4130420983,0.0049337633,-0.2247474194,0.1111173183,-0.1620084345,0.0459868461,0.0497669615,-0.1696457416,-0.1416980475,-0.4695400596,0.2392649502,0.2084290981,0.5362950563,0.0531935468,-0.0222181194,-0.1506141573,0.1603457183,-0.2843222022,0.2370120734,0.2197365314,0.1688712388,-0.050639458,0.2401752323,0.0455840528,-0.2088134438,-0.0520529635,0.1431405991,-0.1415210664,-0.2087105364,-0.0709005222,0.0874165893,-0.1571533382,-0.0231144223,-0.2286979109,-0.1295991987,0.0958835632,-0.1730569452,0.2635291517,-0.1905890405,-0.1276008636,-0.1356274784,0.4436528385,-0.0629725382,-0.0862333998,0.1529286653,-0.0350121558,0.3017438352,-0.0438307151,-0.0223690514,-0.0828737319,-0.0681064129,-0.302716434,0.0466786101,0.2749306858,0.0693440139,-0.0142248739,-0.0127626257,-0.0826871395,0.1370902658,0.0449917652,0.2047968805,-0.0748962462,-0.122948654,-0.0379080921,0.2374708354,-0.2128136456,-0.2070528418,-0.0718625784,0.0127483848,-0.1917217225,0.2789100409,0.362636894,0.0886571631,0.1699345112,0.1253297925,0.3471669853,-0.2870437205,0.2739583254,0.2154373825,0.0546468869,-0.2092100382,-0.0646427646,0.3144549429,0.1434210688,-0.1145211756,-0.1292232871,0.390596956,0.5757665038,0.3090756238,-0.5367895365,-0.2223836333,-0.2938580215,0.6234041452,-0.1351083517,0.4427971542,0.4038825333,0.1510081887,0.2659566402,-0.3229813874,-0.2195305377,0.4456720948,0.045727782,0.0698690489,0.0785575435,-0.077475749,0.1426015794,0.091416873,0.1999518871,0.4314355552,-0.0809788406,0.3132270873,0.0318873338,0.2796559036,0.5481156707,0.1850630045,0.4165057242,-0.1622707397,-0.1929339767,0.202333957,0.2120334655,0.0040217219,-0.0710147023,-0.065638572,0.1874190569,-0.3410095274,-0.0158141963,0.067799814,0.0633217841,-0.3073609471,0.4482382834,0.2307689339,-0.2887846231,-0.0415346101,0.0883270577,-0.1180122048,0.2189888358,0.1187861115,0.1506620497,-0.0560573526,-0.2042896152,0.4145422876,-0.425697118,0.0216816012,-0.1777397543,-0.2082077414,-0.0141025661,0.2535845041,-0.2977363169,0.026771741,-0.0983163863,0.0933161974,-0.1633647382,0.2756984532,0.1564216018,-0.4175422788,-0.2338576615,-0.2869608998,-0.2420893461,-0.1407167464,-0.2342229187,0.0474550948,0.2318329513,0.2380513102,-0.0537076108,0.2327623814,0.1777539551,-0.1275856644,-0.1051266044,0.0347123034,-0.5523297191,0.1517515779,0.4012066126,0.1970658153,0.1963340193,-0.399938345,-0.2417613268,-0.0255168397,0.0489216633,-0.1619679928,0.2541067004,0.1139992028,0.7210950255,0.630227387,-0.0951025188,-0.3539769948,-0.1152171716,-0.1746417135,-0.0046844357,-0.0954035074,0.0094209975,-0.2118441761,0.4912092388,0.2396741211,-0.2630660832,0.2058998942,-0.2440057546,-0.1443867087,-0.2059300095,0.1979186237,-0.2080575228,0.0010099437,0.0479298048,0.4529615343,-0.0281877629,0.2303513438,0.0885680094,0.0875783786,-0.400724411,-0.0172654465,0.2634523511,0.3387112319,-0.3430575728,0.4724929333,0.120660685,0.1163157523,-0.0524527729,-0.2271259427,0.2592238486,0.0508884378,-0.226113081,-0.0018725937,0.2962721586,-0.0765097216,-0.0646928549,0.1092458665,0.2242138833,0.3844121397,0.013439077,-0.4318150282,-0.2402965724]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Oh yes sorry, I didn't check if this was a duplicate","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":11,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nOh yes sorry, I didn't check if this was a duplicate","embeddings":[-0.0366033129,-0.108371377,-0.1726197451,0.0925714895,-0.1399614066,0.1900848746,-0.237350598,0.2739976645,0.1837249249,0.1817451119,0.1359114647,0.4805269837,-0.2320878953,-0.0363494307,-0.2029769272,0.0387382358,-0.2083782107,0.2729924917,-0.0157038309,0.1570293307,-0.1625794023,0.0492690355,0.0893748924,-0.1754247695,0.0940419883,-0.3830752969,0.2062424719,-0.2454313189,-0.1378386021,-0.340425998,-0.0115231294,0.5402710438,0.2562739253,0.1469141692,-0.000118729,-0.0678280145,0.2854865193,0.0646062791,-0.3990343809,0.1823615283,-0.2473600954,-0.1513553709,0.2410388887,-0.1561985463,0.2664539516,-0.336309284,-0.191858843,-0.6333572268,0.4550628364,-0.2540899217,0.130632937,-0.0069808848,-0.5370433331,0.1025287285,0.3090062737,0.3852315247,-0.147271201,-0.0232016742,0.4680567086,0.0541128255,0.2940069139,0.3427099288,-0.0879760161,-0.2660434544,0.3735465407,-0.0129398163,-0.0674647689,-0.2672745287,0.3080031872,-0.0329011157,0.6756606102,-0.3774119616,-0.0876026228,-0.3117339313,0.0755361915,-0.5802270174,0.1118340343,0.1735110432,0.0203509722,0.2382105738,-0.1597699672,-0.4506089091,-0.3615843058,-0.1576098502,0.6576147676,-0.3961524963,-0.075766854,-0.1675001383,0.5345982313,0.160898596,0.1689013988,-0.1312007308,-0.2892184258,0.3916605115,-0.0372729525,-0.2062174678,-0.3568744957,-0.3141362667,0.2141107768,0.1672184169,0.3993676603,-0.0145107554,0.1932789087,0.0235235933,0.2545385659,0.3701909482,0.0641597882,-0.1073692366,0.1913612187,0.0871494785,0.1282299012,-0.0592924394,0.0726686195,-0.0114675341,0.2735614777,0.00906845,-0.0172698647,-0.1670493484,-0.0533333719,0.0605440512,0.2333821207,-0.2068631053,0.0521364063,0.2194326222,0.2255603224,0.0073946472,0.0834517926,0.2904047966,-0.0015170277,-0.3551248014,0.1008089334,0.1562025249,-0.2973764241,-0.2329398543,0.382396102,0.0029081097,-0.0565900244,-0.1127844602,-0.3232616484,0.0703208447,0.5088139176,-0.0700774193,0.2539819181,0.2741203904,-0.2019420117,-0.2093896419,-0.2533438802,-0.0186691936,-0.1518227756,0.4698552191,-0.1208731532,-0.249109,-0.0934977457,0.1456493139,0.1439117193,0.2464458346,-0.5645948052,0.1426448375,0.0977867395,0.0292856079,-0.0047346102,0.4131239951,-0.0131983077,-0.0929034576,-0.2991570532,0.26981318,-0.6811035872,-0.0782427341,-0.3273492754,-0.2331824154,0.2083088607,0.1325968653,-0.0779609382,0.2182586342,-0.1462598145,0.1569957286,0.3353454471,-0.0705056712,-0.3803434074,0.1660287231,0.0469946563,0.1794262379,0.3371204734,0.2802246511,0.3821362853,0.0036964563,-0.0643220767,0.2823109925,0.0809055865,-0.0796709806,-0.3723615408,-0.4111176729,0.0716264173,0.2158055156,-0.0009899011,0.0935366377,0.2123034,-0.0162333939,0.1563631743,0.1369459927,0.1338671893,0.1538633853,0.1613379568,-0.273861438,-0.2404101342,0.127034843,-0.5532772541,-0.0333113335,-0.5310477018,-0.3229122162,-0.14284949,-0.3509566784,0.1190963462,0.0247076582,-0.1623096466,0.1834697127,0.0734581426,-0.0518079996,0.2375502735,-0.0494480878,-0.3926911354,-0.0270851012,-0.056765981,-0.012296061,-0.4700322449,0.0999993682,0.0983990729,-0.2659701109,-0.0158300493,-0.0018842886,-0.2515043616,-0.4967831671,-0.1496641189,0.2637413442,0.2650123835,0.2750349343,0.2835144699,0.2992465198,0.1750531942,0.2245407254,0.0958364829,0.1464095712,-0.0100443792,0.0917588845,-0.4031938612,0.3789787889,-0.2213358581,-0.0602152273,-0.1975868195,-0.3953844607,0.2228783816,-0.1026969403,0.1207740456,-0.5215615034,0.0830339342,0.3393648565,-0.1848513633,-0.0060059889,0.1291021407,-0.0720065609,-0.0013783743,-0.062707983,-0.1805749983,0.3522246182,0.0345799774,-0.2068740427,0.0665417686,0.3632323146,0.3779363036,0.2090756893,-0.2099640965,-0.1597910076,0.4422496259,-0.0052011241,0.1599645317,0.0270432271,-0.04630569,0.3176550567,0.0630844235,0.0316606611,-0.215373382,-0.08439181,0.0795580596,0.2316459119,0.2774061263,-0.2586554289,-0.2563149035,-0.4206625521,0.543202877,0.3581478298,0.2439655662,-0.0285066012,0.5355093479,0.0912705064,-0.4527863562,0.0882182419,0.0371657535,0.6941345334,-0.1463066936,-0.1761846691,-0.4102514386,-0.1941715032,0.1641775072,0.0362143815,0.3594681025,-0.0730502531,0.5412905812,0.0996026248,0.0950003415,-0.0794828981,-0.0406510942,0.053812854,0.3584539294,-0.0280807409,0.1032210514,0.1765483916,-0.0233778581,0.1088951081,-0.1929932684,-0.2972245514,-0.1074461564,0.1503559351,0.1215612516,-0.0353896022,-0.2441469729,-0.0179016367,-0.1660543084,-0.4211525619,0.2708718479,-0.01993821,0.0266862698,0.1058591977,0.2165023685,-0.0331505425,0.0501429662,0.0986336246,-0.0864810199,-0.6310973763,0.1154087633,-0.1973150223,0.0255981497,0.230828926,0.0389048979,0.0025380892,0.3072097301,-0.3970270455,-0.0016880409,0.0433328934,0.1448012739,-0.3366544247,-0.0752177984,0.2304437906,0.2447564453,-0.1160057858,0.0311427675,0.0587359294,0.215231061,0.3872125149,0.1058475897,0.0741851106,0.2719494402,0.1309217662,0.3443796039,-0.2981349528,0.0541751534,0.3607215285,0.3938388228,0.4097033441,-0.1910880357,0.2054018974,-0.4037862718,0.0391907394,-0.1088281348,0.1233941689,0.0180391185,-0.1765770465,0.1617567837,-0.2774718106,-0.1280798912,-0.0802805722,0.3409782946,-0.2277864814,0.0114694051,0.0162397455,-0.2949706912,-0.0956885815,-0.0183474161,0.1572837085,0.1921064556,0.3895140886,-0.1272989661,-0.2529223859,-0.0964457318,-0.1733344048,0.3196084797,-0.2460700423,-0.2221226394,-0.1338500828,0.0342008509,0.1646733582,0.2422480732,0.4378425181,-0.1536037177,-0.0670116246,-0.1898044646,0.0282038655,-0.2070685029,0.0047856984,-0.0009562664,0.0465451404,0.0790324509,0.4111305773,-0.2496115118,-0.1631666571,-0.1934019029,0.5542837977,-0.2414652556,-0.1423303932,-0.040325582,-0.1871939749,-0.5418896079,0.0779767707,0.0752728209,-0.1279450059,0.1951439232,-0.277097553,-0.0609137043,-0.327201277,0.0613522455,0.0596629195,0.5372853279,-0.25898543,0.0630745515,-0.1553764045,0.1424224973,0.4812816381,0.1921659261,-0.1053431407,0.0466666855,0.0131133171,-0.353772521,0.1718088835,0.1683823317,-0.1227810234,0.3079359531,-0.2480763048,0.5612413883,0.0208325442,-0.0538344942,-0.1981996745,-0.3082002997,-0.0199638885,-0.1728536338,0.1233097464,0.0297602229,-0.1686380059,0.4738834202,-0.1482361853,-0.2449397445,0.5139602423,0.0808211863,1.20223701,-0.062676549,-0.0292581245,-0.2138353735,-0.2736672163,0.3012009561,-0.6480371356,0.1267990172,0.0348622426,-0.1223937124,0.1144246981,-0.0245185029,0.010590801,0.1992421746,-0.2962263823,0.0446326248,-0.3182480931,-0.1504679769,-0.2817161083,0.3974272013,-0.0930453539,-0.2443688363,-0.0617881492,0.0083910488,0.040078897,-0.0777519718,-0.2361030132,-0.078844741,0.369715929,-0.0308240075,0.2628191411,-0.130300954,-0.2262835652,0.3534944057,-0.1771006435,-0.1011271104,-0.2281838208,-0.0052153086,-0.1961507499,0.0974376425,0.1281856447,-0.0378666669,0.065939568,-0.0243063103,-0.401094079,0.0533346497,-0.1071375906,-0.1524758041,0.0201633461,0.1266534477,-0.2878597379,-0.3150972128,0.1274878234,0.1516833603,-0.1376063973,-0.002138607,-0.2780788541,-0.1372189075,-0.5119459033,0.0705494359,0.1146756485,0.2921431363,-0.1903174371,0.3096855581,-0.1575526893,-0.1587186754,-0.2414000332,-0.1363027692,0.2840113044,-0.1326251179,0.2125296742,-0.2551085055,-0.2296179086,-0.4739128947,-0.023028275,-0.0402275622,-0.3628153801,0.4374430776,0.0518407859,-0.2191651464,0.0642379746,-0.1196025386,0.0856859758,0.0455489196,-0.1266820729,-0.122181952,-0.4348492324,0.2440638393,0.2208270729,0.5665878057,0.0369738787,0.054934822,-0.1347083896,0.1407600939,-0.2639674842,0.2275250703,0.2521639764,0.1385228634,-0.0113352612,0.2767440975,0.0374166332,-0.2362938076,-0.0667074099,0.07752949,-0.1398570836,-0.1875730902,-0.0873895735,0.1043969393,-0.1464862823,-0.1022448316,-0.1439996213,-0.2106708586,0.067405723,-0.1571490616,0.1972636729,-0.2463503927,-0.1256453991,-0.1308629662,0.4668634832,-0.0443376675,-0.0373267271,0.1524697393,-0.0812270641,0.3616794646,-0.0460871086,0.0132750263,-0.0732625946,-0.0760644525,-0.3422836065,0.0810257569,0.2861869931,0.1075462103,-0.0271409806,0.0387975574,-0.1140372828,0.1164635122,0.0297708139,0.2615965307,-0.0864254683,-0.1010217443,-0.0347939767,0.2231036127,-0.1939083934,-0.1627264768,-0.0565035827,0.0017468473,-0.19946751,0.2903398275,0.3683210015,0.077700913,0.1262744218,0.120930925,0.3442877531,-0.2505447865,0.3180768788,0.2547177374,0.0236294419,-0.2278176099,-0.1372066587,0.2851800025,0.117855452,-0.169683978,-0.1532771289,0.3532061875,0.5774874687,0.3543054163,-0.5351125598,-0.1937206239,-0.2480241209,0.5989273787,-0.0926124305,0.4647454023,0.4544906914,0.1440783143,0.2688583136,-0.2787222564,-0.2222387791,0.4319068491,0.0799592733,0.0860915333,0.0292090327,-0.1300448477,0.1288218349,0.0683501288,0.1789857596,0.3900782466,-0.0809497014,0.2968951762,0.0866479203,0.2687604725,0.5958080292,0.1438880712,0.4303729236,-0.167288512,-0.1848739386,0.1905025989,0.2523037195,0.0803313479,-0.0874929875,0.0041762884,0.1942662299,-0.3599362671,-0.0176781863,0.0604426116,0.0385148376,-0.2829529941,0.4679288268,0.2617961764,-0.2367435843,-0.0590377375,0.0841996521,-0.1142919138,0.1914262325,0.1306995451,0.1644876599,-0.0929177254,-0.2482554913,0.415923506,-0.4679166973,0.0069217146,-0.1816089153,-0.1371795088,0.0096942578,0.2251676321,-0.2824765444,0.0533664338,-0.0263313036,0.0805213824,-0.1857808083,0.2564826012,0.1769459397,-0.3787438869,-0.2738177478,-0.2875075936,-0.2128015757,-0.1725556999,-0.2185553759,-0.0114332503,0.2425569296,0.2486149073,-0.0393899046,0.2290782481,0.2297435254,-0.1294936985,-0.0754836276,0.0790291801,-0.5022533536,0.1299479008,0.3763021529,0.2481629848,0.1688287556,-0.3566430509,-0.2574618161,0.0144397859,0.0173615981,-0.2374392152,0.2342004627,0.0861952603,0.7500835657,0.6114999056,-0.1273616701,-0.3216583729,-0.1342662275,-0.1880308241,0.0140716061,-0.1020938233,0.0027467571,-0.1804737151,0.5183695555,0.1945342869,-0.2854451537,0.2153972536,-0.259203285,-0.1807630807,-0.213474229,0.2537243664,-0.2264254838,0.0112141669,0.0488784835,0.4291118085,-0.0677484274,0.1981167197,0.0523027442,0.1278932542,-0.3842124045,0.0195424929,0.2855980098,0.3838469684,-0.3417404592,0.5095618367,0.1076676548,0.0323039852,-0.0023807238,-0.2725888789,0.2368979305,0.0908153057,-0.253516227,0.0350076295,0.3668664992,-0.0403196253,-0.0636773333,0.1027953476,0.2654809058,0.3798527718,0.0224836767,-0.4177931249,-0.2699744105]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Nevermind @thomwolf, I just mentioned the other issue so that both appear linked in GitHub and we do not forget to close both once we make the corresponding Pull Request... That was the main reason! \ud83d\ude04 ","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":36,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nNevermind @thomwolf, I just mentioned the other issue so that both appear linked in GitHub and we do not forget to close both once we make the corresponding Pull Request... That was the main reason! \ud83d\ude04 ","embeddings":[-0.0855267942,-0.0671884939,-0.1844741851,0.1047870219,-0.068177104,0.0940928012,-0.2069166601,0.3579802513,0.1929914653,0.1803302169,0.1256364733,0.5178521276,-0.1990656555,-0.0532914735,-0.2303050011,0.1003466919,-0.1710781455,0.2507988214,-0.0966360569,0.1721239835,-0.1182256564,0.1586297452,0.0584242344,-0.2147803605,0.0472866036,-0.3116940856,0.2470157593,-0.1306723356,-0.200181976,-0.3335465491,0.0496934317,0.4665739238,0.1547939777,0.1924553365,-0.0001120515,-0.0588400029,0.3263459802,0.0419635475,-0.2812105119,0.1578843147,-0.1546334028,-0.1296814978,0.2462517768,-0.1716015488,0.1941534132,-0.2860300839,-0.2619872987,-0.5600835681,0.4046288133,-0.2074566036,0.218479991,0.0659196898,-0.477872014,0.0386695005,0.3469431698,0.3130682409,-0.220800519,-0.0241451934,0.4340992272,0.0210293364,0.1328230202,0.4203089178,-0.0427928716,-0.16822806,0.3660376072,-0.0318532847,-0.1320763677,-0.2659948468,0.365187943,-0.041888658,0.6081599593,-0.4039493203,-0.0704969019,-0.2332202345,-0.0010145602,-0.5103725791,0.1964832693,0.1308744848,-0.0294248573,0.2057877779,-0.2156448513,-0.3320873678,-0.2986850441,-0.2275019437,0.5624457002,-0.3171729743,-0.0931841806,-0.1442539543,0.4627858996,0.1177915186,0.1654220372,-0.043840453,-0.309324652,0.2831886113,-0.0899897292,-0.2204525322,-0.2162852585,-0.3502406776,0.2751026154,0.2159093171,0.3916691244,0.0350795686,0.195248127,-0.0696302727,0.2703437805,0.4323017597,0.0422205478,-0.1306513101,0.2671504617,0.0566859804,0.0712897331,0.01436364,0.1246161163,-0.0277856607,0.2098570913,-0.0353090949,-0.022360431,-0.273599267,-0.0027620066,0.0769031569,0.299153477,-0.111663267,0.0472754613,0.2507170141,0.2106908709,0.1088606045,0.0572780743,0.2573492825,-0.0073827258,-0.2515538931,0.0386273153,0.1590304375,-0.3453048468,-0.2071355581,0.3635476232,0.0165214315,0.044596605,-0.087183252,-0.2391292304,0.0957716629,0.4768103957,-0.0525910407,0.1936014742,0.3282550871,-0.1214080602,-0.1968099475,-0.2486093491,-0.0390990563,-0.1242086813,0.3764862716,-0.1488395035,-0.2738034129,-0.0759822577,0.2167507559,0.1099694967,0.2422160208,-0.5487085581,0.1165578514,0.1360514164,0.0423572063,0.0025537878,0.3891156316,0.0077833822,-0.1569076777,-0.2708091736,0.1842839122,-0.627738595,0.0181340445,-0.2419081479,-0.184918493,0.2255009711,0.2400516868,-0.1519252807,0.1250222623,-0.137711376,0.1953272074,0.290351212,-0.1376105547,-0.4272357821,0.1646077335,0.0116296783,0.126774922,0.3082715869,0.1919442564,0.3640693426,0.0199810527,-0.0059916275,0.2740213275,0.0454165302,-0.0088724196,-0.447219044,-0.4030652344,0.0181710143,0.206569761,0.0435483195,-0.0042692809,0.2280374467,-0.0454459749,0.1447422057,0.0252100378,0.1102002263,0.1990105212,0.1636434495,-0.3638190031,-0.1670939326,0.1051172838,-0.5550240278,0.0267475788,-0.5375996828,-0.2535658181,-0.1553022563,-0.3377280235,0.1138793975,-0.0178060979,-0.1368129849,0.0994877517,0.1636171341,-0.0505281277,0.2101665437,-0.1134429798,-0.3678587675,-0.0459869616,-0.0555342175,-0.0171632301,-0.4335534871,0.0670195073,0.0713249147,-0.2523716688,0.0725354925,-0.0116517227,-0.185488537,-0.4729955196,-0.1538630873,0.2603837252,0.2450781614,0.2921971083,0.2877057195,0.262139976,0.212813735,0.2098298073,0.1372448355,0.1424447298,-0.0159271136,0.059455283,-0.4149567485,0.3945383132,-0.1529703885,-0.0591715798,-0.1509149373,-0.3506002426,0.2903297842,-0.1401845664,0.0475266464,-0.4543602467,0.101994954,0.3108375371,-0.11021474,0.0151068009,0.0789266378,0.0402069725,0.1006071568,-0.0401637033,-0.139970392,0.3669389188,0.0554620773,-0.2806324363,0.0664960742,0.3567688167,0.3791134059,0.2107024789,-0.194929108,-0.1907741725,0.3597587049,-0.0184016526,0.2091444284,0.022183774,0.0291000344,0.3412797451,0.1237492785,-0.0266182702,-0.2955551744,-0.1001159623,0.0776936337,0.2316409498,0.2472865582,-0.281629622,-0.2429945767,-0.3020982444,0.5641671419,0.3144172728,0.2222724408,-0.0718772188,0.5412746072,0.0934951305,-0.5482689142,0.1295159161,0.0619637035,0.6242787838,-0.1175669357,-0.0005498112,-0.3989090025,-0.2547470033,0.136145547,0.0873429701,0.3191331327,-0.0864869282,0.4921017289,0.03619048,0.0833235234,-0.148256734,-0.1364146471,0.095189482,0.2864724398,-0.0357893258,0.1129134893,0.1283292025,0.0019894571,0.0689897165,-0.1443587989,-0.3590827882,-0.1567138731,0.1778903157,0.0516764335,-0.1459529996,-0.3483550549,-0.137578398,-0.1167290583,-0.4708235562,0.2861891687,-0.0042631645,0.0580689348,0.1066811085,0.1871888191,-0.010237027,0.101838626,0.0515305176,-0.1312522739,-0.645632267,0.1629998386,-0.2419216633,-0.1073780879,0.2100315839,0.0845394582,0.0674860552,0.2412569374,-0.4375688136,-0.047203064,-0.0146714579,0.1606194526,-0.2070327401,-0.118526876,0.1915573925,0.2598814368,-0.2031121552,-0.0333354995,0.0307549555,0.2139242291,0.3151896894,0.1125565097,0.0599735193,0.289211452,0.1651336402,0.3626134098,-0.3334113955,0.0752129778,0.375272274,0.3635934293,0.4455353618,-0.2741265595,0.2036553025,-0.3434279859,-0.0178960413,-0.0921317115,0.1906603128,0.0659162477,-0.2766295373,0.1496109664,-0.2267407924,-0.0468463264,-0.0985539481,0.3262548149,-0.2088486403,-0.0914773643,0.0621343665,-0.2048475891,-0.0688840151,0.0063673197,0.1837731004,0.1083875969,0.2688840628,-0.1304294318,-0.2644743621,-0.0394006856,-0.2572125494,0.2851050794,-0.2718003094,-0.2192859948,-0.1319360286,-0.0172103178,0.1451883614,0.131049931,0.4179213047,-0.140944317,-0.1006120071,-0.1882832944,0.0125888558,-0.2201875597,0.0189312864,-0.0363561064,0.107901223,0.1314092278,0.3927888572,-0.2881758213,-0.2088010311,-0.2204589248,0.4583599269,-0.203073293,-0.0995665565,-0.0356749073,-0.2675747573,-0.5211288333,0.0827575102,0.1385283917,-0.1120305359,0.1495106816,-0.2988811731,-0.0196059756,-0.3307202756,0.0492195189,0.1010380089,0.587572515,-0.288762033,0.0951905027,-0.201562196,0.1822337657,0.447460115,0.2902311385,-0.0935392678,0.1095428765,0.0888662785,-0.3233518302,0.0258917809,0.1684431434,-0.1167424992,0.2301909029,-0.2475335151,0.6316025257,0.0619255453,-0.1011545882,-0.1581904143,-0.2953957617,0.0368270837,-0.1168675125,0.1441997141,0.0038704437,-0.1517389715,0.4974929094,-0.174226433,-0.2369040698,0.5581269264,0.111887604,1.1223025322,0.066245608,-0.0082102856,-0.1159408912,-0.3521792293,0.2567613721,-0.652798593,0.0919904262,0.0124151642,-0.1366784275,0.1495248973,-0.0076310323,-0.0505187176,0.1721395254,-0.321641475,0.0434333868,-0.2436955869,-0.1264190376,-0.3005656004,0.3972910941,-0.088121146,-0.3036735058,-0.1368499398,0.0695902333,0.0961647108,-0.1027526334,-0.2141225636,-0.1129526868,0.3157936037,-0.0842183679,0.2002479732,-0.0531649329,-0.1764604449,0.3988555074,-0.1700195074,-0.077173382,-0.2630424798,-0.0888826102,-0.1211327091,0.0884734988,0.0152103426,0.0539484024,0.1129603684,0.0923838764,-0.3502333164,-0.0262473058,-0.0631624311,-0.1901107132,0.0450095348,0.0384848081,-0.3371963799,-0.2730949819,0.1275775284,0.1341819763,-0.1687294841,-0.0039641424,-0.2065664679,-0.0774889812,-0.5094299912,0.013762611,0.1816055179,0.2855283916,-0.1449526697,0.3495632112,-0.0768412799,-0.2227308005,-0.2739017308,-0.0561630167,0.3547869921,-0.091887787,0.1497976184,-0.1659592986,-0.2815509439,-0.4964163005,-0.0726443455,0.0225013308,-0.4313990176,0.4540007114,-0.0022811722,-0.3116541505,0.1154088303,-0.0660914704,0.1380863488,0.0372062437,-0.1806195378,-0.0911826566,-0.4733947515,0.2555477321,0.150656566,0.5617165565,0.0009446395,-0.0016481731,-0.0733282864,0.185130626,-0.343716234,0.1866710931,0.1418766081,0.1569520235,-0.0738390237,0.1522551328,0.0887035429,-0.1735325307,0.0005617291,0.118409574,-0.1825791001,-0.2424518019,-0.0360613316,0.0890141428,-0.1598332375,-0.0823260844,-0.1170865074,-0.1031417474,-0.000993015,-0.1549751163,0.3016618788,-0.1995360255,-0.0685699284,-0.2173978239,0.4563748538,-0.0954960808,0.0174306892,0.1671779603,-0.0397904813,0.3143796325,-0.051456105,0.0146957245,-0.0966604799,-0.1061326265,-0.213660121,0.0702321976,0.2113621235,0.1138135269,-0.00930354,0.0091657769,-0.1153390184,0.1331099272,0.1107856855,0.2708093524,-0.0825885758,-0.1567667723,-0.032939326,0.3032121956,-0.2050681412,-0.1960132122,-0.1091993749,0.0081968037,-0.164889425,0.1821452677,0.3783392608,0.0112582073,0.2039268166,0.1890286207,0.3733148873,-0.292996943,0.22335504,0.0937391371,-0.0571074747,-0.2370075881,-0.0850778595,0.2410746664,0.1441717893,-0.0554114357,-0.1522160023,0.3187150657,0.5666558743,0.3326039612,-0.4247448742,-0.0982944816,-0.1980435103,0.5276719928,-0.1208389401,0.4313080907,0.4194306731,0.1199663803,0.2624455988,-0.3267873228,-0.2990862727,0.4584262967,0.0017612294,0.0138817709,0.0278939884,-0.1851598173,0.1503153741,0.0822533518,0.1463255733,0.4236711562,0.005089514,0.2698338628,0.0395110697,0.2251470834,0.5589661002,0.1303586215,0.4065194428,-0.1425887048,-0.0419807062,0.2612911761,0.2220591307,0.0050223274,0.0026297749,0.1577077359,0.2627748549,-0.3247102797,-0.0300761051,0.1915375292,0.0047532809,-0.3440208733,0.4657173455,0.2651111782,-0.2276815325,0.0245036595,0.1511320025,-0.2007097453,0.1950538158,0.0990488529,0.1387065798,-0.0145252338,-0.2503454685,0.3574079573,-0.4714492261,0.0433498137,-0.2280961722,-0.2472657561,-0.0754457563,0.1768155098,-0.2844343483,0.072503604,-0.1067322493,0.1258524954,-0.1753312796,0.3727457225,0.1064315587,-0.4275980294,-0.2481838763,-0.3017343581,-0.2549421489,-0.0955486894,-0.2005655617,0.045793049,0.2084631175,0.3400158882,-0.0459286422,0.155306682,0.2822338343,-0.1257595122,-0.1181766614,0.045445893,-0.6027138829,0.264029175,0.3624567688,0.2131042033,0.2241931707,-0.3873642385,-0.2172870636,0.0356928334,0.0904211923,-0.2374916673,0.2351773828,0.1361281425,0.6803244948,0.5626975298,-0.0797042176,-0.3214339614,-0.1804889292,-0.250641197,-0.1027373224,-0.1903569698,0.0146540832,-0.1661782116,0.5049532056,0.232637167,-0.1965624243,0.0962210968,-0.2331662625,-0.1289421022,-0.1913463473,0.2648508549,-0.1744265854,0.0859469473,0.0570479184,0.3812099993,-0.117163755,0.1815215647,0.0742729455,0.100742355,-0.3683781326,-0.1265716106,0.3467052281,0.2702930868,-0.3461805582,0.5040483475,0.1022237539,0.0673033968,-0.0944453254,-0.3364172876,0.213380456,0.0235660188,-0.1767968535,-0.0001871988,0.3356523216,-0.0955229178,-0.0587279871,0.067130886,0.2493175864,0.3315478265,-0.0811263695,-0.3822200596,-0.2865698934]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Ok yes. I think this is an important feature to be able to use large datasets which are pretty much always compressed files.\r\n\r\nIn particular now this requires to keep the extracted file on the drive if you want to avoid reprocessing the dataset so in my case, this require using always ~400GB of drive instead of just 200GB (which is already significant). \r\n\r\nTwo nice features would be to:\r\n- allow to delete the extracted files without loosing the ability to load the dataset from the cached arrow-file\r\n- streamlined decompression when only the currently read file is extracted - this might require to read the list of files from the extracted archives before processing them?","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":116,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nOk yes. I think this is an important feature to be able to use large datasets which are pretty much always compressed files.\r\n\r\nIn particular now this requires to keep the extracted file on the drive if you want to avoid reprocessing the dataset so in my case, this require using always ~400GB of drive instead of just 200GB (which is already significant). \r\n\r\nTwo nice features would be to:\r\n- allow to delete the extracted files without loosing the ability to load the dataset from the cached arrow-file\r\n- streamlined decompression when only the currently read file is extracted - this might require to read the list of files from the extracted archives before processing them?","embeddings":[-0.0659782663,-0.0675047114,-0.1214437932,0.1361941546,-0.1881691217,0.2371559441,-0.1601299345,0.2473284453,0.1624708176,0.1584105045,0.1043199822,0.4483769238,-0.3028531969,-0.1058051959,-0.1939978004,-0.0391121805,-0.1808364391,0.3018629253,0.060030818,0.1913317889,-0.1481152475,-0.0545217507,0.0214629024,-0.1286898702,0.1161556914,-0.4201310873,0.2055554688,-0.2362951189,-0.0982239917,-0.3271234035,-0.007774367,0.5414354801,0.3300149441,0.1635190398,-0.0001216864,0.052123867,0.2817978859,0.0643243417,-0.4494723678,0.1446421295,-0.319024384,-0.1872631907,0.1448402554,-0.1605795622,0.2758429945,-0.5139070153,-0.1515802592,-0.7546899915,0.4762697518,-0.2039914727,0.0753073767,-0.0049457639,-0.5496518612,0.1562354863,0.3124717474,0.3418077826,-0.1361890435,-0.0434954464,0.4967588782,0.1190313622,0.2489924133,0.3314892054,-0.0464917868,-0.3507281542,0.3594222963,-0.0695533082,-0.0735744238,-0.2807359695,0.2501985431,0.062279094,0.6185549498,-0.29497841,-0.1788841784,-0.3965529203,0.0939029753,-0.6207100153,-0.0016169348,0.2456875294,0.0135154417,0.3100341558,-0.0990560204,-0.4396600723,-0.3894589543,-0.1128680632,0.6326623559,-0.4742105007,-0.0741034672,-0.1111170501,0.5121361017,0.1867254823,0.2198597342,-0.1967426687,-0.216770947,0.3942317665,-0.0830868557,-0.1818786412,-0.3705095351,-0.2833347619,0.1888691038,0.1414779872,0.4201419055,-0.043493446,0.1795663387,0.0853910446,0.2840005159,0.3924605846,-0.0170195773,-0.1081940308,0.164158985,0.1096849069,0.1837197095,-0.0634693876,0.0451794453,-0.0622212812,0.1686196029,0.1019991189,-0.0905887187,-0.1100680009,-0.0399750024,0.0446358025,0.0713075846,-0.2095420957,0.0990588367,0.1798976511,0.2338270098,0.0712885484,0.0895205736,0.2898355424,0.0463679843,-0.3151602149,0.1110785678,0.0767037272,-0.3082774282,-0.2443213016,0.4048590362,-0.1130664423,-0.0529358163,-0.1779991984,-0.2501621842,0.0887633339,0.5159014463,-0.15522708,0.32464239,0.2360241115,-0.2288743705,-0.2514579892,-0.2872161567,-0.0144998934,-0.1715121269,0.5403386354,-0.1308786273,-0.2736297846,-0.0119039081,0.0827526972,0.0778672025,0.2827013135,-0.5317140222,0.2350833267,0.0539898686,-0.0468185395,-0.0410921536,0.3351682425,-0.0156068821,-0.0135693792,-0.3476057351,0.2468978018,-0.6518745422,-0.0602268279,-0.3570014834,-0.2370943874,0.1824218333,0.1142961234,-0.1536521316,0.1675506979,-0.1818363518,0.1057171226,0.382294178,-0.0641675666,-0.3867464364,0.1953918934,0.1077707261,0.2615795434,0.3133714795,0.2150320411,0.4147535861,-0.0729536191,-0.0957668126,0.2799834609,0.0771862268,-0.0969428793,-0.2989012301,-0.4622783959,0.0986724347,0.2479757816,0.0211347342,0.1290628165,0.2614646852,-0.0357314311,0.2037031651,0.1651514918,0.2209406495,0.1400309354,0.0642625839,-0.2695454657,-0.1503100544,0.1737479866,-0.4975177646,0.01263077,-0.5214065313,-0.3172445595,-0.2045380324,-0.2885702252,0.1372274905,-0.0335044265,-0.1988251954,0.2015222907,0.0237849467,-0.1412809044,0.2724556625,0.0653194413,-0.410171032,-0.065781571,-0.0675783232,-0.0595216975,-0.4646087587,0.1204921156,0.15674676,-0.1407268792,-0.1265902072,0.0524367802,-0.3225057423,-0.4410732687,-0.1505691856,0.1877081245,0.2010252923,0.2613529265,0.1614187062,0.3275243342,0.2161498666,0.142862767,0.1739604175,0.1386932433,0.0297906715,0.0958648995,-0.3868570924,0.4077583849,-0.3279556036,-0.02537922,-0.1928204447,-0.4176359773,0.2501657605,-0.1081964374,0.0301855393,-0.5059494376,0.0732085854,0.3495838344,-0.0904177725,-0.0261918902,0.1483157873,-0.2089301199,-0.0180657115,-0.076565899,-0.257807076,0.3139603734,-0.0984877869,-0.1204739064,0.1099255383,0.456631273,0.3945102394,0.213454783,-0.1338796914,-0.1825801432,0.5015577674,0.016985897,0.1319739074,0.0567714237,0.0611689985,0.4169522226,0.1301807612,0.0714413598,-0.2148226947,-0.1260572374,0.176108703,0.1470306665,0.2887785733,-0.1400389522,-0.2087569684,-0.4624943733,0.4350057542,0.384583503,0.2519319952,-0.0065840073,0.521450758,0.007744438,-0.3969198465,0.0599964745,-0.0019939914,0.7599349618,-0.0931037068,-0.3055965304,-0.4187726378,-0.1977101117,0.1883245409,0.0038354045,0.3493969142,-0.1790193319,0.4796026051,0.2143743187,0.0479682013,-0.0809134394,-0.0552876033,0.0849809125,0.3521818817,-0.0478455313,0.0350214019,0.1456884444,0.0863028392,0.0937902629,-0.1733703613,-0.2101908177,-0.1256005913,0.1074935123,0.120885931,0.0952274501,-0.26980111,0.0426243693,-0.2468212098,-0.4235396981,0.309130609,-0.0513284877,-0.0689929351,0.1583266854,0.2341705114,-0.0445756726,0.0575840995,0.0680007637,-0.0336196497,-0.6436601877,0.126938045,-0.194460839,0.0436173677,0.0808591545,-0.0631687641,-0.0740565509,0.4198182523,-0.3769463003,0.0351677239,0.0643253475,0.1337387413,-0.2988145053,-0.0806291699,0.203765735,0.2121611387,-0.0176124293,-0.0105862413,0.1454561204,0.2049527317,0.3793349862,0.1457602978,0.1917964369,0.2467255294,0.2169598043,0.4157980978,-0.2578362525,-0.0076828725,0.3150094151,0.3919858336,0.4485295713,-0.1432017386,0.1977029443,-0.3346395791,-0.0101613402,-0.1489811838,0.0669744015,0.026199406,-0.2776841223,0.1452684999,-0.2031218857,-0.0504853539,-0.1147197783,0.2853794098,-0.2961053252,0.1113294363,0.0297484435,-0.2856206596,-0.0609254465,-0.0604318641,0.1740287393,0.1901502013,0.527521193,-0.0549111515,-0.1452022642,-0.0328800902,-0.167621851,0.2732896805,-0.1726186872,-0.1998912692,0.0356642716,-0.0120085748,0.0921494812,0.2388247699,0.351552397,-0.1252101511,-0.0307965763,-0.1520735621,0.0430063009,-0.2111055553,0.0137661034,0.0080303699,0.0207165964,0.0089398762,0.4495282471,-0.2623148561,-0.1419359148,-0.2325685471,0.5902493,-0.2365605831,-0.1827652156,-0.0008258743,-0.1295437366,-0.558511138,0.0057399222,0.0432744026,-0.1035288572,0.0979587138,-0.2668532729,-0.1903073937,-0.2498408854,0.0896735191,0.0202734731,0.4588965476,-0.2742724121,0.0562027916,-0.035032738,0.2389666736,0.5481042862,0.2142433524,-0.1473025531,-0.0585219599,0.0262672845,-0.3692209423,0.1674478799,0.1663966328,-0.2425571233,0.3254412115,-0.2277772725,0.4670628309,-0.0689684525,0.0652040616,-0.2590865195,-0.2505386174,-0.057197772,-0.2039663345,0.1911697686,0.0541405343,-0.2098721415,0.4351094663,-0.1205914766,-0.2572704256,0.5326521397,0.0590778664,1.1551059484,-0.1128400043,0.0149776274,-0.2680406272,-0.1797770113,0.3728262782,-0.6633321047,0.1816278845,0.0149576878,-0.1384883225,0.0593060404,-0.0376286991,0.0376681797,0.2846598029,-0.2282191813,0.0590172149,-0.2025690228,-0.2804599106,-0.2769897878,0.3981357813,-0.0736177936,-0.2576678097,-0.0592688732,-0.0360440314,0.0596208535,-0.0574105307,-0.2111502886,-0.1065161675,0.3768342733,0.0289752763,0.2776345313,-0.1710879058,-0.3207930028,0.37308532,-0.2118289769,-0.2551985979,-0.1024199054,0.0718890801,-0.2685132921,0.0503335856,0.1471635401,-0.0070664492,0.0903498009,-0.0627790689,-0.3701249063,0.0344958156,-0.0024105327,-0.0782943442,0.0977918729,0.133659333,-0.3397369981,-0.3465418518,0.1758268774,0.1366618127,-0.1244315505,0.0564565919,-0.2980706394,-0.0924467146,-0.4841829538,0.1409901083,0.0586788356,0.3002807796,-0.1433354914,0.2562084496,-0.1935006529,-0.1190173849,-0.149434045,-0.1528203189,0.1490313411,-0.201412037,0.2311071754,-0.2553104162,-0.2430458069,-0.3802861869,-0.0017059805,-0.1441027075,-0.3890040219,0.4247324467,0.0393172652,-0.164671272,0.0566639565,-0.1659507453,0.0021754019,0.0056372131,-0.0699911565,-0.167266041,-0.4592873156,0.2587143183,0.1362250149,0.5790670514,0.0889938995,0.0018238538,-0.1906795204,0.0303710699,-0.2141229808,0.312122643,0.3379603624,0.1776246428,0.0269685257,0.2774659097,0.0403180234,-0.2374209911,-0.1122769862,0.108269237,-0.0424622186,-0.1265537739,-0.1338105947,0.1278323233,-0.1039082706,-0.117332086,-0.1563272923,-0.264405936,0.0292451568,-0.2014670223,0.252779156,-0.252995342,-0.1273670793,-0.1315404326,0.3878419995,-0.0559778735,-0.1523716897,0.1996946633,-0.0371241346,0.4207261503,0.0125948014,0.1074035168,-0.079014048,-0.0557299778,-0.3938583434,0.0815995038,0.3363562822,0.0910331011,0.1453466266,0.0098264096,-0.0346613526,0.111732468,0.0570550896,0.2079506516,-0.1051376984,-0.0590624847,0.0234868582,0.1583219022,-0.2124461383,-0.1063571349,-0.0475777201,-0.0073401737,-0.1604300737,0.2504925728,0.3605211675,0.0801652148,-0.0175501183,0.041823972,0.3067479432,-0.1281606257,0.3053994775,0.3543405533,0.1122724339,-0.1918138415,-0.1634384841,0.3312183619,0.1881328821,-0.1521891952,-0.1744483262,0.4083150625,0.6499543786,0.3493705988,-0.5603497028,-0.2789341509,-0.2383551896,0.6232469678,-0.0590776578,0.4289219975,0.3845160007,0.2122107446,0.3161064684,-0.1366515905,-0.1427980363,0.3982588649,0.0905594006,0.0355214886,0.0893562138,-0.1809120625,0.1088452637,0.069777064,0.1355870515,0.3712620735,-0.1599113196,0.3642855883,0.1624091864,0.2602543235,0.5718260407,0.2083934993,0.4663642049,-0.1589477956,-0.2222063094,0.203652367,0.2428520918,0.1175448969,-0.148644492,-0.057663817,0.1116549447,-0.3451363146,-0.0397500731,0.0243021622,0.0840382203,-0.2572030425,0.4885839522,0.2748909593,-0.1898321211,0.0664420649,0.0044170911,-0.0531427227,0.1279197633,0.106685631,0.2125914693,-0.1347337663,-0.2357351929,0.3795200288,-0.4040885568,0.0092477389,-0.156363517,-0.016412247,0.0293930583,0.3202225566,-0.2720809281,0.0790598243,-0.0039838958,0.0465872064,-0.2235050946,0.2113467604,0.250146836,-0.3231532872,-0.2714592218,-0.2596997023,-0.1967801303,-0.2127230167,-0.2271255255,0.0763290524,0.1694624871,0.255387336,0.0660128817,0.2566123903,0.0694929585,-0.1009575278,0.0615431219,0.0652921945,-0.4469427466,0.0430698469,0.3881988823,0.2435121536,0.1337826848,-0.4015809,-0.3996726573,-0.0001541694,-0.0219380166,-0.1888750941,0.2017796189,0.0450216085,0.6890196204,0.6287796497,-0.1140046865,-0.3001966476,-0.0539143607,-0.2284241915,-0.01189934,-0.0830095932,0.0136514502,-0.1952994615,0.4423668981,0.1934665591,-0.3521393239,0.2612171471,-0.2566983104,-0.2904362977,-0.1876447648,0.1607292444,-0.3335031271,-0.0664425939,0.0190968812,0.4339583516,-0.0229122136,0.2096648663,0.0817559138,0.0818541124,-0.3381944597,0.0530681089,0.3016922474,0.3127169013,-0.2746132314,0.5506646633,0.0493562184,0.0816834494,0.0799990445,-0.2568173409,0.2503429651,0.1334230453,-0.3104121089,0.0269021839,0.22445786,0.0296762194,-0.0604568943,0.0501811132,0.2297381461,0.4168799818,0.0695711449,-0.4529075325,-0.2204630375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Here is a sample dataset with 2 such large compressed JSON files for debugging: https:\/\/huggingface.co\/datasets\/thomwolf\/github-python","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":15,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nHere is a sample dataset with 2 such large compressed JSON files for debugging: https:\/\/huggingface.co\/datasets\/thomwolf\/github-python","embeddings":[-0.0169936661,-0.0682562366,-0.1684834659,0.1376501173,-0.114887163,0.2113289535,-0.2281462401,0.3236738741,0.1736370474,0.1715837121,0.0459124111,0.4963761866,-0.2788704336,-0.0538061522,-0.1982839555,0.0174690802,-0.2327052504,0.2629131079,-0.000834777,0.1391376108,-0.2056843489,0.0763502717,0.095775269,-0.1601253599,0.0850936845,-0.399422884,0.1995611489,-0.1731853336,-0.1389152557,-0.3850656748,0.0341053493,0.4772646129,0.2020599991,0.2446924001,-0.0001157716,-0.0168931186,0.2781267464,0.0621967502,-0.3196890354,0.1111031994,-0.1010666266,-0.2671664953,0.2862975895,-0.2343785167,0.2200739831,-0.4366144836,-0.2372525632,-0.5964100957,0.5392258167,-0.1587032676,0.1493151784,0.0765969381,-0.4646046162,0.2176100016,0.254529953,0.2713605762,-0.0684066489,-0.0163567811,0.4038017988,0.0394244529,0.2270101011,0.3743507862,-0.1088718921,-0.2328777909,0.2765701115,-0.0244357251,-0.1708483547,-0.30673334,0.2547684014,0.0177752133,0.5055035949,-0.4267705381,-0.206065014,-0.3952306509,0.0189415682,-0.6138208508,0.072251156,0.3135051429,0.0087184934,0.2182660401,-0.1577137262,-0.4097868502,-0.3154049516,-0.1699337363,0.5475116372,-0.3331407905,-0.098488085,-0.1441705674,0.4340670407,0.1463348866,0.084492445,-0.1203620434,-0.2248427868,0.3644792736,-0.1286361665,-0.2182502151,-0.2475602925,-0.3634177148,0.2481589615,0.1696466953,0.3965750337,-0.0312122609,0.1674796194,0.0042392625,0.3777911067,0.3392499089,-0.0816996843,-0.0245182812,0.24783355,0.1200791821,0.064010419,-0.0303204954,0.1094782799,-0.0536826216,0.3268526793,-0.0790069774,0.0345332548,-0.1398760974,-0.0631909966,0.1534960866,0.1859401613,-0.1214794442,0.056186527,0.2608822882,0.0880092904,0.0602556765,0.0881378725,0.3040839434,-0.0305862837,-0.1868526638,0.0686677396,0.1571844071,-0.2111539394,-0.318841368,0.4060235322,-0.0792245045,0.0122818993,-0.0984817892,-0.2402355224,0.0544403717,0.5071704388,-0.1121432856,0.2900906503,0.3412407637,-0.149225682,-0.1910929382,-0.1826563478,-0.0261287652,-0.2034787536,0.4042988122,-0.1277468204,-0.2351548225,-0.0458506197,0.1783536375,0.0477979444,0.2497895211,-0.5278394818,0.0684289187,0.1264518499,0.1158713326,0.0041911909,0.3143624067,0.0215372331,-0.1205291674,-0.2801882923,0.2492801547,-0.5500209332,-0.0365078561,-0.1783906966,-0.1092542335,0.1778396368,0.2550433278,-0.0808312967,0.1595889032,-0.1068121046,0.0955895633,0.3396663666,-0.1749308258,-0.456825465,0.1986899227,0.105441764,0.2928672135,0.3374663591,0.2192000747,0.3792721629,-0.0339504071,0.0907950699,0.3430294693,0.0721237138,0.0276341811,-0.363984257,-0.3740795255,0.0892929584,0.267718941,-0.0272140168,0.0129804919,0.147609666,-0.0313696526,0.2347529531,0.1169256717,0.0864295438,0.2102002203,0.1295076013,-0.2718859911,-0.1732898802,0.071410872,-0.4378314912,0.0112419818,-0.538595736,-0.3038817942,-0.3040435314,-0.3855687976,0.0789063275,0.0620402172,-0.2327082455,0.1554632336,0.1304564327,-0.1921911538,0.3403347433,-0.0035167316,-0.3402796686,0.0292939339,-0.0590032525,0.0511920974,-0.4845068753,0.0923854634,0.1839355528,-0.2581029832,0.0771689937,0.034825433,-0.2512815893,-0.4606161714,-0.1254389733,0.2028525621,0.2771537602,0.2229471654,0.1549841911,0.2596005499,0.1791579574,0.2709450722,0.0931848064,0.250045985,0.0113254711,0.1312356144,-0.3801043332,0.5091734529,-0.2080241144,-0.0311560705,-0.1015357524,-0.376611352,0.3098761141,-0.0779668763,0.020364821,-0.4797513187,0.0970776603,0.3106902838,-0.0803746581,-0.0017288779,0.0275561437,-0.1156489477,0.0523684099,-0.0504994541,-0.1193759292,0.3685443699,-0.022044586,-0.1889853776,0.0388121009,0.3843270242,0.3549191058,0.2104460299,-0.1722673476,-0.1049537733,0.4155032039,0.0325230546,0.184768945,0.050443925,0.015313264,0.3043811619,0.2006035298,-0.0044324794,-0.2743135989,-0.0368107036,0.0138317998,0.2311517894,0.2526070476,-0.1381140202,-0.2445079386,-0.3480923474,0.4648372531,0.2443551123,0.2184765935,-0.1068165153,0.4667146802,0.0895445719,-0.421718508,0.0682236925,0.0289665833,0.5414905548,-0.0196305104,-0.2216268927,-0.3330029845,-0.2109463513,0.2685104907,0.0519700423,0.2835614383,-0.1001383886,0.5142947435,0.0631648228,0.0424041934,-0.06173715,-0.061844565,0.0530674979,0.2984448969,0.0031475276,0.16386877,0.1881419271,0.0660543889,0.137511611,-0.1802319884,-0.3209214807,-0.1144834459,0.2043666542,0.0844947845,0.0213908479,-0.2193122506,-0.0722922087,-0.293177247,-0.4365380704,0.3728617132,0.0793152079,0.0671292171,0.1434578151,0.2093574852,-0.0487483777,0.1384535283,0.1523579508,-0.0831187293,-0.6068615317,0.1853674799,-0.2284948379,-0.0573727228,0.2039583176,0.0092333173,-0.0113012167,0.250402838,-0.4052413106,-0.0921040401,-0.1114999652,0.1287299246,-0.2769276798,-0.0043816427,0.1526069939,0.2187811583,-0.1331076175,0.0322530046,0.1409778446,0.1769621521,0.3837744296,0.0174865257,0.0662105158,0.3317128122,0.1933521181,0.3376311362,-0.1656651199,0.0075330455,0.3359537721,0.3517350256,0.4334448576,-0.2395508885,0.179918319,-0.4714204967,-0.0671335086,-0.0689718947,0.1028317064,0.1017300412,-0.1578926444,0.1740158796,-0.2745215297,-0.1388277411,-0.1511255205,0.3018121421,-0.2162437737,0.0104613556,0.0686368421,-0.2729349136,-0.1300522089,-0.032227993,0.1041338593,0.1449568868,0.3493366539,-0.1110005379,-0.1534313858,-0.1122224033,-0.3458511531,0.2769809663,-0.227492407,-0.1854131073,0.0434966609,-0.0086046271,0.1140223071,0.1415327787,0.4429002404,-0.0973188579,-0.0674819648,-0.151565671,-0.0078164292,-0.2977987528,0.062273141,-0.0488107689,0.1298964769,0.0856688917,0.3897455335,-0.2689493001,-0.1939980388,-0.1906466335,0.6191979051,-0.2709014416,-0.1249744892,-0.1746865064,-0.3342669606,-0.5528398752,0.0548357777,0.1354525834,-0.1232485026,0.0804169178,-0.2512666285,-0.1445064247,-0.3039771914,0.122902602,0.0752861425,0.5894860029,-0.3132990301,0.1106707752,-0.2050095499,0.1528285444,0.4619110227,0.3640015423,-0.0571881346,-0.0211222935,-0.0330611803,-0.2187896222,0.1290117651,0.1350948215,-0.1631465256,0.2773213685,-0.3452711403,0.5883857012,-0.0162549,-0.0405322053,-0.1741306633,-0.2406437397,-0.0820339322,-0.2605144083,0.1656965762,0.0793748274,-0.1063420475,0.4098128676,-0.1779149771,-0.2386145741,0.5379610062,0.002636421,1.0563164949,-0.0883757398,0.066152513,-0.1089380756,-0.2795662284,0.3381921351,-0.6869991422,0.0967085287,0.001751809,-0.0950515494,0.1262803972,-0.001609232,0.0169822834,0.1814923584,-0.2697455287,0.1540760845,-0.3199807405,-0.2937957048,-0.3428912461,0.439859122,-0.1278840154,-0.279148072,-0.3186212182,0.0356456712,0.0867928639,-0.1311885417,-0.1559235901,-0.1102386862,0.3341566622,-0.0814117044,0.2514284849,-0.1553727537,-0.1693622172,0.4123905599,-0.1458032578,-0.1786203235,-0.2333274335,-0.017269196,-0.2029901296,0.1408917904,0.1989801377,-0.0188093744,0.0143102081,-0.0316998772,-0.3350197077,0.0901224166,-0.0723351464,-0.1213047132,0.0447090641,0.1458229572,-0.3071435988,-0.3199861944,0.2664019465,0.2024478316,-0.2397292107,0.0016068713,-0.2665775716,-0.1490221322,-0.4852522016,0.0462726578,0.1335788667,0.3378057778,-0.2357253283,0.3232236505,-0.1937749982,-0.221286431,-0.2584616244,-0.0837082639,0.3287581801,-0.052411858,0.2663638592,-0.159684211,-0.2327763587,-0.4453850091,0.0238822605,-0.0758127868,-0.3097715378,0.4812886417,0.0662761554,-0.1360445619,0.0332223698,-0.1099282578,0.0139062759,-0.014641637,-0.1281180978,-0.1300586015,-0.4838713706,0.2004666477,0.2768577039,0.5199813247,0.1086602435,0.0335475244,-0.1974541992,0.2192512453,-0.2890631258,0.2319226861,0.172998026,0.1176061034,-0.0655213669,0.2084127516,0.100701727,-0.2065102756,-0.0194545612,0.1000729203,-0.1160536781,-0.2427807599,-0.108425647,0.091121152,-0.1501511335,-0.0781620815,-0.1085741147,-0.1531800926,0.0593728684,-0.1755370498,0.2849014699,-0.1985510737,-0.0807711184,-0.1160584986,0.4480643868,-0.1509602219,-0.0319529362,0.1118379608,-0.1484582126,0.363899678,-0.0110233156,-0.0164212938,-0.0709534064,-0.0450945534,-0.3389012516,0.0378273837,0.2419276088,0.0599709041,0.0576584116,-0.0539072454,-0.1338953525,0.1282693595,0.1445530355,0.3312636316,-0.0961762518,-0.0910813287,-0.0232603401,0.2364314646,-0.2761449218,-0.2036462277,-0.0390889794,-0.0260876622,-0.1250431985,0.2997507453,0.3590848148,0.1174056903,0.0677081048,0.1324165612,0.3830536306,-0.227650106,0.3243924975,0.2142491043,0.1060079709,-0.2711303532,-0.1087198183,0.3324488997,0.1665827781,-0.1057286784,-0.0883628577,0.3563954234,0.5358122587,0.2586383224,-0.5757492185,-0.2219984829,-0.2463385016,0.5269256234,-0.1570944637,0.500741899,0.4327366054,0.1987486035,0.2180612087,-0.2279835939,-0.2971060574,0.5140675306,0.0754104108,-0.0314709432,0.1933294684,-0.0960355327,0.1717407107,0.1403826475,0.1796314269,0.335164547,-0.1277837753,0.2823382616,0.1393050849,0.1572492123,0.5491549969,0.0761973858,0.4848117828,-0.197527349,-0.204747051,0.2450865656,0.2090682983,-0.0273484513,-0.0971390381,0.1207562387,0.1856268942,-0.4163447917,0.0421207473,0.0412998348,0.0243751183,-0.2815448344,0.436105758,0.201238364,-0.1999259889,0.0006576125,0.1032505557,-0.1542203128,0.1716529131,0.1217003837,0.1460222453,-0.1209237501,-0.1648214161,0.3124657869,-0.4224329889,-0.0443561487,-0.1806552261,-0.2683995664,-0.0632991195,0.2645927072,-0.259144932,0.0685495734,-0.0849260315,0.1093234792,-0.2058179379,0.3559647202,0.1919505447,-0.305415839,-0.2546583116,-0.2305723578,-0.2477012575,-0.1783735603,-0.3294903934,-0.0898144096,0.2122757137,0.2586374581,-0.0514870808,0.2432061285,0.0774542242,-0.0654528812,-0.1187923625,0.0339451805,-0.5166630745,0.0930591375,0.2839400172,0.2288304567,0.1642683595,-0.3908842802,-0.2956994474,-0.0840063617,0.0476192459,-0.2099985033,0.2810187638,0.0492673144,0.7359642386,0.5960111022,-0.0444244482,-0.304774195,-0.1443673223,-0.2492381781,0.0131039191,-0.1079092845,-0.0197065752,-0.1576778293,0.4348415434,0.2388244569,-0.2917954326,0.1030941457,-0.2023262233,-0.1733230799,-0.1959204376,0.1307047904,-0.1523630619,0.0568117388,0.08723858,0.4376743138,-0.0906561166,0.3069017529,0.0144953104,0.1601238549,-0.3306591213,-0.0659390166,0.2919672728,0.2149828374,-0.3742571175,0.4963657856,0.0697935298,-0.0329746865,-0.0264893714,-0.273147285,0.3278731704,0.093006365,-0.2399257421,-0.0975692719,0.3043729365,-0.050274875,-0.1359650791,0.1112446636,0.2554913163,0.3207032681,-0.0023923717,-0.3764052987,-0.2538140118]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Note that I'm confirming that with the current master branch of dataset, deleting extracted files (without deleting the arrow cache file) lead to **re-extracting** these files when reloading the dataset instead of directly loading the arrow cache file.","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":38,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nNote that I'm confirming that with the current master branch of dataset, deleting extracted files (without deleting the arrow cache file) lead to **re-extracting** these files when reloading the dataset instead of directly loading the arrow cache file.","embeddings":[-0.1638352126,0.0380310751,-0.1447685808,0.1780929118,-0.1592689753,0.2357144654,-0.1746569574,0.287337333,0.1265909523,0.1585224569,0.0674439222,0.5160763264,-0.2275414467,-0.0690875575,-0.2018715143,0.0688017979,-0.2498386949,0.2992703915,-0.1215841621,0.1475564241,-0.1509249061,0.0209516753,0.0688547119,-0.1775142699,-0.0016516814,-0.396625787,0.2204260677,-0.2167356908,-0.1347008646,-0.4136333764,0.0634001121,0.5557292104,0.2655682862,0.1864333749,-0.0001169127,-0.0113526583,0.2978251576,0.1052251905,-0.4620437622,0.1928018481,-0.2715247571,-0.1185481772,0.2721543312,-0.117644541,0.2377959192,-0.4201447964,-0.2090130597,-0.6489426494,0.479297787,-0.1898259521,0.1478467733,0.0329266563,-0.5153906345,0.0615889952,0.2572768927,0.3319445252,-0.0981928185,0.0039390293,0.3674480319,0.0887911618,0.1793729812,0.3715210557,-0.0340011679,-0.2261724919,0.3655919135,-0.0161389373,-0.1732269377,-0.2317009717,0.3097927868,0.0443826169,0.6939400434,-0.4388231933,-0.1589534432,-0.3909534216,0.142165795,-0.5054193735,0.1175634265,0.1817198545,0.0114592006,0.2422340661,-0.1476078779,-0.4430011213,-0.3262475133,-0.1966211945,0.5953676701,-0.4200973511,-0.0878205076,-0.1266939789,0.4859718382,0.0839420781,0.2252024263,-0.1792857945,-0.3020443618,0.3221302927,-0.1004943848,-0.1947907656,-0.3318036199,-0.2837395668,0.1633085012,0.1971438229,0.3921901286,0.0155776553,0.1589901894,-0.00245813,0.3650124967,0.325317502,0.0567488,-0.015445116,0.1955094486,0.1219282299,0.0944667011,-0.0626407117,0.0564416498,-0.0361695439,0.3055577278,-0.0224623028,0.03207618,-0.1643724442,-0.0773370042,0.0661545843,0.1919808686,-0.1818098426,0.0514081493,0.1919044852,0.1945622265,0.108414501,0.0120298024,0.2396143079,0.0325863101,-0.3729590178,0.0624157004,0.0842375457,-0.3152473867,-0.2167695761,0.4251202047,-0.0692520663,-0.0677439123,-0.0776425228,-0.332596451,0.1238681599,0.5096788406,-0.0929375216,0.2749297917,0.3011269271,-0.194448486,-0.2114191949,-0.2887327373,0.0115748607,-0.146205619,0.4718725681,-0.0836265981,-0.2405286133,-0.079288736,0.1582892239,0.09983401,0.2539297342,-0.6136035919,0.1032643318,0.1020422503,0.0629070476,0.0064155911,0.2727427483,-0.0109288385,-0.0711655617,-0.250028491,0.2974756062,-0.7026976347,-0.0151146194,-0.3810774088,-0.2615415752,0.1441979557,0.1360021681,-0.1141004711,0.2153520286,-0.1931096464,0.1585097462,0.2810723484,-0.127763778,-0.4494964182,0.1385644823,0.0316712596,0.221644789,0.2527033687,0.2599445879,0.4064200521,-0.0053484794,-0.0747046247,0.2618270814,0.0072225165,-0.0818888769,-0.31014961,-0.3716854751,0.0091329832,0.2902944982,-0.0464238934,0.0333958231,0.2457732558,-0.0116136139,0.1760638505,0.1377877891,0.1765563488,0.1564887613,0.1129363626,-0.2869826853,-0.2430571169,0.117363438,-0.6472294331,0.0515569635,-0.4302771389,-0.3367313743,-0.2597639859,-0.4017222226,0.1069798246,0.0003136142,-0.2254595608,0.1922654808,0.090428181,-0.0458627604,0.1582460403,-0.0665628016,-0.3913021982,0.0469880179,-0.065038465,0.0207478963,-0.4797946215,0.069530502,0.1714746952,-0.2255218178,0.0300064292,0.0377172492,-0.3066351116,-0.4373158813,-0.1367695034,0.2760284245,0.2677150667,0.2772708535,0.2583813965,0.2232232094,0.1723570675,0.1517799199,0.0993092805,0.1439179927,-0.0193581209,0.1048069149,-0.4482202828,0.3862940371,-0.1830082238,-0.0284164567,-0.1669945568,-0.3487372994,0.2705671489,-0.0806844011,0.0948654562,-0.6014310718,-0.0367365107,0.3301691711,-0.0589249842,0.0794566199,0.0906827226,0.0266449917,0.061290618,-0.0423662998,-0.1162151322,0.3406762481,-0.0697613433,-0.2100413293,0.1170764118,0.4911105633,0.3566942513,0.2214138061,-0.0846716613,-0.1309578121,0.4406019151,0.004432214,0.1452678591,0.0406752266,0.0857581198,0.3085535467,0.0913145542,-0.0110078687,-0.2508101761,-0.0662560016,0.104020901,0.2441475242,0.2336023152,-0.2658806741,-0.3304500878,-0.3818595707,0.4872902632,0.3326804638,0.2534179986,-0.0412212983,0.4535224438,0.1211714,-0.4766841531,0.1249921694,-0.0667654425,0.6247702837,-0.1010287553,-0.2049594671,-0.3821294308,-0.2168403566,0.1266383678,0.0564317144,0.390730232,-0.0646003336,0.4973318279,0.1180510968,0.1314409673,-0.1648124009,-0.0041278186,0.0818762034,0.3110399544,-0.0571077578,0.1339185089,0.1453293711,-0.0676509216,0.1235644221,-0.2136645615,-0.3039454818,-0.1567982435,0.1779378057,0.0637528077,-0.0044931006,-0.2445878983,-0.111591801,-0.1638434529,-0.434940815,0.2674321234,-0.0510960147,0.0239288732,0.1086262837,0.1742657274,-0.0175042432,-0.0008385375,0.0091251042,-0.0748917088,-0.5851693153,0.1650405675,-0.2132187635,0.0230189376,0.2357920557,0.0374633074,0.0251581594,0.3056431711,-0.4097911716,0.0278872866,-0.0611578524,0.1791488081,-0.2906417251,-0.1186671332,0.2030057013,0.2551263869,-0.1212828606,0.0341082998,0.1248948574,0.2755829096,0.3200286627,0.0323269442,0.0342403799,0.2859163582,0.1640781462,0.3512731493,-0.1879602075,0.0512155481,0.2993659973,0.3690423965,0.4838581383,-0.2564122379,0.2488680631,-0.3840307891,0.0199971236,-0.0404896438,0.1524169892,0.0407854207,-0.1463923007,0.2161570936,-0.2363279611,-0.0881399512,-0.1792969555,0.2731652558,-0.2889282107,0.0402077772,0.0800965801,-0.2684657276,-0.0185032208,-0.02338982,0.1533515304,0.0773832723,0.3878053725,-0.0952340364,-0.2750812769,-0.0267894324,-0.2655270994,0.3023177981,-0.2460312247,-0.1332048178,-0.04756587,0.0054382095,0.1472662389,0.2468611896,0.3770068586,-0.1615940481,0.003467219,-0.13444902,0.0493624918,-0.3254891634,0.0385548547,-0.0113813234,0.0659052506,-0.0248180516,0.3186227977,-0.2891429961,-0.2005819529,-0.1354070753,0.565841794,-0.2458693981,-0.0819148868,0.0042757182,-0.2207072675,-0.4849289358,0.0790770501,0.0261178501,-0.1166029125,0.1459041685,-0.2863050997,-0.1278501451,-0.2963278592,0.0831851736,0.0918702781,0.5984148979,-0.2050711215,0.1636439115,-0.061012581,0.1400483847,0.6094427109,0.3160106838,-0.1091067567,-0.0088362759,0.0345186517,-0.3696348369,0.1417111605,0.1127367318,-0.1866817027,0.2563526332,-0.1997178346,0.5040171742,0.0093943654,0.0027100665,-0.2488678247,-0.2839725614,-0.0684408844,-0.2279616147,0.108049944,0.0423949696,-0.1545619071,0.4536227286,-0.1000730097,-0.236952737,0.4782822728,0.1114018857,1.0490735769,-0.0775836781,0.0356169529,-0.16096735,-0.2834871709,0.3261193931,-0.5598512292,0.0843030885,0.0123187294,-0.1201917529,0.1132013127,-0.0343065038,-0.0023790926,0.2772354782,-0.2347500175,0.0795550793,-0.31140095,-0.1200347692,-0.247147277,0.430690378,-0.1301980764,-0.2882220447,-0.0762301609,0.0350224264,0.0873179659,-0.1888197809,-0.218951121,-0.0781412646,0.383589983,-0.0389158651,0.2399488389,-0.0820738599,-0.2556138039,0.3634471893,-0.1974613369,-0.1324067861,-0.1501623392,0.0025843054,-0.1674365699,0.0485405102,0.1268176734,0.0721175671,0.0788132474,0.0102125863,-0.3229548633,0.0176080726,-0.0816233307,-0.1278713793,0.0057415199,0.0871285871,-0.2665798068,-0.3794738352,0.2276147008,0.1466220766,-0.1534769237,0.0173721761,-0.2547750175,-0.1255757213,-0.479109019,0.0489408635,0.1253713071,0.2693444192,-0.12796399,0.2854361832,-0.1708791256,-0.2049617171,-0.2208631337,-0.1151561886,0.2870612741,-0.101997517,0.330894649,-0.1526322067,-0.1944950521,-0.4860873222,0.0062784147,0.089837946,-0.4311336577,0.4288218915,0.0083317179,-0.1597492993,0.1343875825,-0.0785340667,-0.0049223094,-0.0023801187,-0.1550836861,-0.117955409,-0.4771165252,0.2584301531,0.1668663025,0.5414474607,0.0517605096,-0.0477177911,-0.1501578093,0.12489824,-0.2867148817,0.2042428404,0.2524323463,0.1670934558,-0.0562897325,0.3051927984,0.1198346242,-0.3132944107,-0.0517852828,0.0629816353,-0.1310729682,-0.206208989,-0.0797003135,0.0879196972,-0.163065657,-0.0899496004,-0.1748681366,-0.1778810173,0.0288615394,-0.2104351968,0.2703487873,-0.1927797794,-0.1530976743,-0.1547871083,0.4422270656,-0.0978604332,-0.1274126768,0.1500856727,-0.0465313308,0.3000342548,-0.0446943641,0.0621789806,-0.0825495645,-0.0836137608,-0.2250507474,0.1014937311,0.2650090158,0.0771158487,-0.0195489116,-0.0275106225,-0.1250665337,0.1517771631,0.1358239204,0.2690428793,-0.0462784618,-0.1790920943,-0.0391742103,0.2356315255,-0.2341137081,-0.1392617673,-0.0257328991,0.0546709672,-0.170319438,0.2459820658,0.3385914564,0.0634860173,0.0567091256,0.1311381459,0.2675943971,-0.2423135787,0.384311825,0.3634724021,0.0747027546,-0.2442734987,-0.0652984455,0.3188364208,0.1602816582,0.0378592908,-0.1042153835,0.4175528884,0.6066229343,0.3013852537,-0.4863904417,-0.3089416325,-0.2099139988,0.5793690085,-0.0945490077,0.4008535147,0.4404424727,0.1670518368,0.3227195144,-0.356412977,-0.2688576579,0.3678481281,0.0567994677,0.0569508299,0.0324857384,-0.113542743,0.1598705202,0.1240466386,0.1077638045,0.3357599378,-0.0621859916,0.2875965536,0.098116748,0.1508205533,0.5759006739,0.1738013029,0.3641438782,-0.1365813017,-0.159902364,0.1996975243,0.2128989398,0.0841390565,-0.0369849019,-0.0155064054,0.1501622349,-0.3222709,0.0604530089,0.0688138232,0.0356950834,-0.2824096382,0.4310357273,0.2198584527,-0.2790054977,0.0124408482,0.1058833748,-0.0951280147,0.1621370614,0.0995228589,0.1884828508,-0.0682634041,-0.2039425522,0.3901362419,-0.444083035,0.0619216152,-0.1376009732,-0.0918116644,-0.0495721027,0.3588742614,-0.3027020693,0.0067122257,-0.1420395076,0.0905867815,-0.1946097314,0.3490243852,0.1869707257,-0.3395476043,-0.2126796544,-0.3154652715,-0.3378932774,-0.1892919093,-0.2675836682,0.0137952147,0.269831121,0.2674233317,-0.0198091269,0.2280034423,0.2677681446,-0.048060108,-0.1025072783,0.0251173377,-0.501121819,0.1211824864,0.3226194382,0.2113983482,0.1301867515,-0.4562162459,-0.2158824056,-0.0135159045,0.0075380914,-0.1407108754,0.2123311758,0.1369470805,0.6466160417,0.6635488868,-0.0839906931,-0.3119256198,-0.0729778558,-0.1753679216,-0.1198861003,-0.1471249014,-0.0737959594,-0.109639585,0.4317800999,0.294123143,-0.2969919145,0.2574816942,-0.2406549603,-0.2088908851,-0.1787059754,0.2433126718,-0.2513650656,0.05997416,0.0469661802,0.4676736891,-0.0515468679,0.1882142276,0.0739948601,0.1181122735,-0.3579023182,-0.0601623654,0.3388687968,0.2894136906,-0.3498056531,0.4524713159,0.1452544332,0.0975934565,-0.137529999,-0.2947260439,0.2565614879,0.1006885245,-0.2167925686,-0.0259707868,0.2420888394,-0.0488131717,-0.0287118498,0.0757223219,0.3928261101,0.3239247501,-0.0721385255,-0.4051163793,-0.2832627892]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Hi ! That's weird, it doesn't do that on my side (tested on master on my laptop by deleting the `extracted` folder in the download cache directory). You tested with one of the files at https:\/\/huggingface.co\/datasets\/thomwolf\/github-python that you have locally ?","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":41,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nHi ! That's weird, it doesn't do that on my side (tested on master on my laptop by deleting the `extracted` folder in the download cache directory). You tested with one of the files at https:\/\/huggingface.co\/datasets\/thomwolf\/github-python that you have locally ?","embeddings":[-0.0332566053,-0.0617311411,-0.1518054754,0.2321407646,-0.0794417933,0.189954102,-0.1519192904,0.3226890564,0.2679444253,0.1859114766,-0.008661042,0.47859025,-0.2064822614,0.0198211502,-0.1249555647,0.0193799771,-0.2612201273,0.2673110068,-0.0352828801,0.1023952886,-0.1649813056,0.161548242,0.0307119377,-0.2143165022,-0.0014290169,-0.2806631327,0.177387327,-0.0900829583,-0.2368220091,-0.4120254219,0.0911949351,0.4586049616,0.2278508395,0.2842088342,-0.0001189412,0.0202764049,0.3218185008,0.066426903,-0.3416108489,0.0965557545,-0.1090628654,-0.2528342605,0.3365181684,-0.2395073324,0.1118235067,-0.2765804529,-0.2922463417,-0.5912124515,0.5315221548,-0.0952571779,0.1267584264,0.1466713101,-0.355958432,0.2192575932,0.2747636437,0.2166805863,-0.09886498,0.053727679,0.315277338,0.0675988942,0.2849488556,0.3909698725,-0.1213172525,-0.1054870561,0.3051481247,0.0537144914,-0.1651786566,-0.4498274624,0.263142556,-0.0311787631,0.5653121471,-0.4969917536,-0.2195424885,-0.3977091908,0.0205808245,-0.602556169,0.0792899057,0.3545735478,-0.0875537023,0.2304884493,-0.2468704283,-0.3859897554,-0.3075895905,-0.0835531577,0.4763791561,-0.2620851099,-0.1824117601,-0.1066635922,0.4677399695,0.1795823574,0.0424528904,-0.2289459109,-0.2521284521,0.3555209041,-0.1274119616,-0.2224962562,-0.2088644505,-0.3136998713,0.2307114452,0.2699070573,0.2709549963,0.0157219525,0.1045328081,-0.0305523351,0.3759339154,0.4152934551,-0.0611145161,0.1297464371,0.3189991117,0.1017591357,0.0494839922,-0.0757151172,0.109716095,-0.0244888943,0.3162862659,-0.1740889698,0.1749759167,-0.1531502008,-0.0912235156,0.1615382433,0.1361991167,-0.1264770925,0.0433214866,0.2908792198,0.030465981,0.087059319,0.0671079159,0.3209926486,-0.056101162,-0.1605153531,0.0770599991,0.1735166311,-0.2492263913,-0.2693278491,0.421579957,-0.1249038652,0.0887147859,-0.0435720794,-0.216262728,-0.0413144492,0.517576158,-0.1195725799,0.3411772251,0.4210411906,-0.0745880976,-0.114984937,-0.1722786129,-0.1657734364,-0.1743794382,0.460924834,-0.1498116553,-0.2387833297,-0.1282224804,0.1567413956,-0.00324932,0.2452864647,-0.6246505976,0.0005452317,0.1519329697,0.2135063261,0.0848618522,0.2263615876,0.0496609882,-0.1697266549,-0.1571180671,0.3347275853,-0.4668126404,-0.0333851315,-0.2098701894,-0.1474102885,0.1847275645,0.2946440578,-0.091226317,0.2407300323,-0.2020543963,0.1187637821,0.3351454735,-0.2505328953,-0.5275449753,0.2201987356,0.064316839,0.3103695214,0.263060987,0.238543421,0.3052237332,0.0946832523,0.0400316529,0.3342135251,0.0402658992,0.0579853579,-0.403421253,-0.2949002385,0.0445607603,0.2469552606,-0.0413194336,0.1011844873,0.1281719208,-0.1058721989,0.3076839447,0.0411467254,0.1362333149,0.2670237124,0.1770786941,-0.2603784204,-0.1623841375,0.0183591694,-0.4770290256,0.0945220068,-0.4080125093,-0.2634161413,-0.3683611453,-0.3869258463,0.0043123853,0.0810938403,-0.3423129916,0.0619365312,0.0587098934,-0.1631045491,0.3893430233,0.002705645,-0.3460029364,0.151015684,0.020282967,0.102244325,-0.4906042516,0.1922679096,0.1886428148,-0.251817435,-0.0409127697,-0.0152284065,-0.2488425225,-0.4605016112,-0.154968068,0.2493535727,0.2999372184,0.1253156364,0.206798479,0.2235670835,0.1670227051,0.2181142867,0.1772170961,0.222254023,0.0617597066,0.1005986854,-0.2781881988,0.4478914142,-0.1885373592,-0.0291268285,-0.1488328725,-0.3766204417,0.3021794856,-0.0755825639,-0.0541075468,-0.4702609181,-0.0302460603,0.3676839769,-0.0016348548,-0.0202190597,-0.0253918841,-0.047281839,0.1545387059,-0.0499397144,-0.125313133,0.405554086,-0.149950996,-0.1962159425,-0.0140799722,0.415500164,0.4606252313,0.1591747254,-0.0678065792,-0.0173171833,0.373567313,0.0155714555,0.1984344423,0.0710080042,-0.0462578759,0.3129602671,0.2086460143,-0.0325560942,-0.2569279671,0.0208207872,-0.0300837718,0.2917767763,0.1879808456,-0.0984210372,-0.319522351,-0.3765306175,0.4093852937,0.1857545227,0.1845454872,-0.1193320379,0.3986589015,0.1807380468,-0.3674204946,0.0814922825,0.0344223753,0.3880402744,-0.0016941144,-0.3106749356,-0.253988117,-0.1466290206,0.1887052804,-0.0019916501,0.2663029432,-0.1614186019,0.5131768584,-0.0222540926,-0.0057921638,-0.1705052555,0.0209889617,0.0856403336,0.2487094551,0.0692659765,0.1128563136,0.2418886721,0.0192850307,0.16073744,-0.0956477001,-0.2883743644,-0.1349900961,0.1901531368,0.139432922,0.0259200912,-0.2716072798,-0.104124479,-0.2724406123,-0.4422227442,0.3500591815,0.1526203156,0.0999123305,0.1743971407,0.2171635479,-0.0255618915,0.1482309699,0.0802659839,-0.1224801093,-0.6555259824,0.2202030569,-0.2332468033,-0.046239946,0.3175825477,0.0387545489,0.0010782455,0.2596187294,-0.4103809893,-0.1555810571,-0.193201825,0.2453217357,-0.1802003831,0.1367727518,0.1767840534,0.1448154598,-0.0755918175,0.0392446816,0.0460777506,0.1047629714,0.3434068263,-0.004457972,0.1057259217,0.320535779,0.1107369885,0.4776507914,-0.0486312658,0.0883385465,0.4071294069,0.2520943284,0.4920267165,-0.3374051452,0.1258828938,-0.4846446812,-0.093485795,-0.0854883119,0.1300485432,0.155964613,-0.152102977,0.1395096928,-0.2521050572,-0.1673852056,-0.1934487671,0.26682809,-0.2344254851,-0.0259914212,0.0939274281,-0.1695002764,-0.0837181434,-0.0794792324,0.0241836868,0.1382854134,0.2804583907,-0.0628481731,-0.1751363128,-0.1319058537,-0.4479261935,0.315987885,-0.2335048169,-0.2515760958,0.1123895124,-0.0193636231,0.2216553241,0.1352764815,0.4183787107,-0.1490491331,-0.0966638476,-0.075592272,-0.1137547716,-0.3181276321,0.002690048,-0.02132174,0.1850785464,0.0618401095,0.2871908247,-0.2788704038,-0.2535727322,-0.1004381999,0.6766911149,-0.2556499839,-0.0628542006,-0.2358076125,-0.408051759,-0.60403198,0.0970637575,0.1638958752,-0.1048907861,0.0750683546,-0.2510490119,-0.0694815814,-0.3054733872,0.0271753594,0.061598096,0.6772767305,-0.269992888,0.1834549606,-0.1231342927,0.0992158353,0.4116418958,0.482519716,-0.0674106181,-0.0367381647,-0.0192384962,-0.2285833806,0.0878268778,0.1801995039,-0.1216647103,0.2007099241,-0.3159459829,0.6137964129,-0.0255600065,-0.0278348289,-0.0738735795,-0.2462096661,-0.1315940022,-0.3758296371,0.1242166311,0.0124268821,-0.0810214877,0.4736640155,-0.1542472541,-0.2274920195,0.4616504014,-0.0732436851,1.1378904581,-0.0573131926,0.1218130216,-0.0517418943,-0.3184500337,0.3369064331,-0.6598299742,0.1584886461,-0.1112314314,-0.06817922,0.1269742101,-0.0553699955,-0.0240613595,0.1334136426,-0.2422806472,0.2128768712,-0.4090472162,-0.2152429074,-0.3238561749,0.4293532968,-0.1048845574,-0.2767993808,-0.2988873124,0.0471160598,0.121523805,0.0428713635,-0.1723892987,-0.0776229948,0.3175717294,-0.1417823583,0.1199446097,-0.0692977458,-0.1389086396,0.371584028,-0.1808597594,-0.1589767337,-0.3154517114,0.0253913309,-0.040912766,0.1076188684,0.1379154474,-0.0042093545,-0.0192121156,0.0207404774,-0.3300172687,0.0066792201,-0.0624321625,-0.1476429403,-0.0375362113,0.1586971879,-0.3073330522,-0.409011364,0.2464348972,0.2679155469,-0.2021264881,-0.0482521616,-0.2696137726,-0.1242538989,-0.3961208165,0.0285079163,0.1112093255,0.3953776658,-0.2896129191,0.4057909548,-0.2678988278,-0.3382913768,-0.2832980752,-0.0936657339,0.3441705108,-0.0186301209,0.3068976998,-0.1357008219,-0.2476723939,-0.4111408293,-0.0265933573,-0.0254206769,-0.3614841998,0.5023365021,0.0710434914,-0.054200273,0.0357516184,-0.0458611324,-0.0785357952,-0.0049883584,-0.1925136894,-0.0915509239,-0.4491462409,0.1111074388,0.2908155918,0.5239835978,0.1269328594,0.0626064464,-0.0523873009,0.1228004992,-0.2610010207,0.2496437579,0.1228006482,0.101456739,-0.0038102278,0.2337698042,0.1216498241,-0.300934732,-0.0606019683,0.0755942762,-0.1573365033,-0.2063392997,-0.0989713669,0.13120915,-0.1449001431,-0.1276474595,-0.0739609152,-0.1366320401,0.1312752664,-0.252708137,0.4108954668,-0.0777803063,-0.0681053773,-0.1161760092,0.4291162193,-0.1365370899,-0.024872398,0.1099174917,-0.191246897,0.4337045848,0.0322052687,-0.007711886,-0.0308952611,-0.043557182,-0.3207204938,-0.0046776305,0.2004585713,0.0698054805,0.0687473714,-0.1616945714,-0.0815911815,0.139899686,0.2040309459,0.3896818757,-0.1375796795,-0.1116091982,0.0036818793,0.2207110673,-0.2480198294,-0.2681465447,0.0259247962,0.0022775617,-0.1571576148,0.3477062285,0.3385125101,0.156788215,0.1067682207,0.1792643517,0.4243697524,-0.2484146208,0.3661262095,0.1565615535,0.0731896535,-0.1989876479,-0.086117819,0.2434795648,0.1555797607,-0.0020661238,-0.1434388161,0.3063270748,0.5199885368,0.2046111226,-0.577091217,-0.2283822894,-0.2111485451,0.4667412043,-0.2342362106,0.4735678434,0.4336741567,0.1427622437,0.1222108677,-0.2752473652,-0.2910160124,0.5431444645,0.0150197046,-0.0291628689,0.1798154861,-0.0965120941,0.1239811182,0.2251506895,0.1246054322,0.2634052634,-0.0272092558,0.2285267115,0.0037364285,0.091122292,0.5725488067,0.0508422479,0.4281857908,-0.2018005699,-0.1588830501,0.2087461501,0.1749545485,-0.1198238209,-0.0308807362,0.1743926108,0.2122117877,-0.3490408063,0.0587140881,0.0426388755,-0.004761362,-0.2958358228,0.3895008266,0.1521134675,-0.1610446572,-0.0768554285,0.1117371991,-0.1184739023,0.1702170372,0.0884494558,0.1912444383,-0.2271018773,-0.0656104237,0.2860577703,-0.4047800601,-0.0090621449,-0.2457282841,-0.2952024937,-0.0067712418,0.2661706805,-0.2224278301,0.0355192795,-0.1656464934,0.0889626369,-0.1993576586,0.3422857523,0.2350543439,-0.2720923126,-0.3159385026,-0.2059533447,-0.3135770559,-0.1919190139,-0.3323296309,-0.116323851,0.1817055345,0.3304493129,-0.1060380116,0.2373958975,0.1871113926,-0.0952251256,-0.1168179139,0.0020117366,-0.4604878128,0.1378270388,0.178845942,0.2282205969,0.2081568092,-0.5176976323,-0.2647792697,-0.1527632028,0.0007780798,-0.1358692944,0.3029811084,0.1575910896,0.5477007627,0.6887497306,0.0014801199,-0.2566511929,-0.087068662,-0.2421883494,-0.073850818,-0.1291605532,-0.122088179,-0.1594549417,0.3465606868,0.3143826127,-0.3293669224,0.0846820548,-0.2532246709,-0.0861108825,-0.0682175979,0.1455230415,-0.0891815946,-0.0163181294,0.0765483677,0.4157755077,-0.0701462403,0.3245227039,-0.0210703555,0.1589498967,-0.3448210657,-0.1187079176,0.3293447793,0.2186017632,-0.4817661047,0.4003863335,0.1353747249,0.0657074079,-0.1505673528,-0.3618246019,0.3132098019,0.1325524449,-0.2016752958,-0.1304768324,0.3226748705,-0.0866594762,-0.0766901225,0.1098620519,0.3308889568,0.3124461472,-0.0759336278,-0.3197581172,-0.2632790804]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Yes it\u2019s when I load local compressed JSON line files with load_dataset(\u2018json\u2019, data_files=\u2026) ","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":13,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nYes it\u2019s when I load local compressed JSON line files with load_dataset(\u2018json\u2019, data_files=\u2026) ","embeddings":[-0.0635800362,-0.0847523734,-0.1382575929,0.2011316717,-0.0965400264,0.2113160044,-0.196976468,0.3288321793,0.1711597592,0.2009096742,0.1392868757,0.5203614831,-0.2585755289,-0.095453389,-0.2117458284,0.079856731,-0.2573565543,0.2693507373,-0.0222670306,0.1242618114,-0.2414246351,0.1660193652,0.0535003282,-0.1449650675,0.0856962651,-0.4024108052,0.2848064601,-0.2019057572,-0.1029638276,-0.2979504764,0.0280696079,0.4437125027,0.2293639034,0.1763621271,-0.0001169832,0.0102999788,0.2814941704,0.1020329297,-0.4003998935,0.12378411,-0.1906584799,-0.1657997817,0.238501206,-0.1889450401,0.279465735,-0.3739452064,-0.2461220324,-0.649482131,0.4810369015,-0.1473705769,0.1448148787,0.0109573286,-0.5180348754,0.0949772522,0.2730261385,0.3306569159,-0.1542105228,-0.0407175496,0.3553336859,0.1126540676,0.2725184858,0.3803431094,-0.1114516407,-0.2278735191,0.2919649482,0.0121525601,-0.0858529955,-0.1974995434,0.3307258785,0.0150918523,0.5970155597,-0.3527722061,-0.0742606297,-0.344498992,-0.0018316263,-0.5800248384,0.1015805006,0.2347045541,-0.0571065545,0.2165191323,-0.2112868577,-0.3947140574,-0.312858969,-0.1773796827,0.5671154857,-0.4244703352,-0.0826532617,-0.0885035396,0.5740939379,0.1540887356,0.1433882117,-0.0817294866,-0.2929649055,0.3760729134,-0.1425431371,-0.120250456,-0.2880927026,-0.3599862456,0.2681155801,0.0932924002,0.3620294631,0.0055040484,0.219757095,0.048581399,0.3063241541,0.3946848512,0.0918822885,-0.0487503931,0.1659249216,0.0948420018,0.1130609438,-0.0948436558,0.0788979828,-0.0632909387,0.2748019099,-0.0330589749,-0.0287100654,-0.1716949344,-0.0820516497,0.0949533209,0.2081473768,-0.1450390965,0.0903762579,0.2377069443,0.2337224782,0.0674112439,0.0401293375,0.2440683544,-0.0231635869,-0.2795912921,0.0661243647,0.09141922,-0.2972505987,-0.2955617011,0.4095713794,-0.1072728634,-0.0496052951,-0.1214377582,-0.3052442074,0.0387799032,0.5178328753,-0.1238287985,0.2838617265,0.3377973437,-0.1790522039,-0.1495909393,-0.2495439053,-0.0750825182,-0.1833612919,0.4709766507,-0.2032969892,-0.3339942694,-0.1453439444,0.1658548713,0.1310273111,0.2989928424,-0.5026883483,0.081569545,0.1350649744,0.0512877703,-0.0302543771,0.2902136147,-0.0580001846,-0.0936537012,-0.2759104967,0.2684074342,-0.7323592901,-0.0220701229,-0.2772506773,-0.2275446355,0.2773965597,0.228630349,-0.1826620847,0.1682454795,-0.1851234734,0.0969069749,0.3489550948,-0.1746658832,-0.4313072264,0.2473357618,0.0735695586,0.2193296999,0.3583711684,0.1735933572,0.3324270546,0.0125492904,0.0232300796,0.2784374654,0.1063790172,-0.0403750353,-0.4129101932,-0.2761159837,0.0693003833,0.1603296846,-0.0913568661,0.0804016367,0.1787704974,0.0445038378,0.1720892042,0.1066036671,0.1605418175,0.2210215628,0.1384108812,-0.3090108633,-0.1172299534,0.1510982513,-0.59116745,0.0145322653,-0.5845335722,-0.3123106658,-0.2778976262,-0.3892090321,0.0534067042,0.074235037,-0.2267380059,0.1417468488,0.096065633,-0.0950851738,0.2082782835,-0.0333214216,-0.3720234632,0.0020060975,-0.0792518482,-0.0233054291,-0.5341299772,0.0223248824,0.1378320903,-0.2544942498,-0.0557215065,0.0301357321,-0.2559274435,-0.4776660204,-0.1578157246,0.2724581659,0.3247003257,0.2275327444,0.217433542,0.2387457341,0.1792694628,0.2545652092,0.048718933,0.0845372677,-0.0247324239,0.0858078897,-0.370270282,0.3852406442,-0.1330831647,-0.0197265912,-0.1854273528,-0.4231408834,0.2925439179,-0.094213143,0.121211946,-0.518992424,0.1143218949,0.4044379294,-0.1330315024,0.0675328076,0.1132749692,-0.0767724216,0.1019629464,-0.0573651493,-0.1977558434,0.3910286129,-0.0030391708,-0.2319087833,0.046133548,0.4088195264,0.3612288535,0.1625292301,-0.1774450839,-0.146980837,0.4461042583,0.0068955063,0.1660000682,0.0404739641,0.0579954386,0.3246586025,0.1174688935,0.0180919673,-0.2778457999,0.0083286054,0.0858630762,0.3030084968,0.2225315571,-0.1208289862,-0.3327079713,-0.3580002487,0.454020232,0.3447776437,0.1910223663,-0.0919562057,0.4545688629,0.0743405595,-0.396792829,0.0419173986,0.0452621244,0.6345162392,-0.0660132244,-0.2219114453,-0.424015075,-0.2159726918,0.0751345381,0.0251874775,0.3217666745,-0.1082612723,0.4472024739,0.1371442378,0.0753245875,0.0054193311,-0.0464960225,0.0764015242,0.3397600949,0.0341175497,0.1303276271,0.1632763296,-0.0540470146,0.119188197,-0.1434805691,-0.2617385089,-0.1286611408,0.2225173563,0.0013687264,0.0398566648,-0.2956105173,-0.0844877884,-0.1626943946,-0.4200725555,0.2796683013,-0.0371101536,0.0695616454,0.1055243239,0.1892113537,-0.039067369,0.0826264247,0.076404728,-0.101819396,-0.6348325014,0.0827554688,-0.2103749663,-0.0012818597,0.1711402237,0.0554326288,0.0071303239,0.3502431512,-0.396490097,-0.0188177638,0.0186857954,0.1775083095,-0.3602614105,-0.1873993725,0.2488869727,0.2688949704,-0.1270888001,0.0010847113,0.0781250671,0.2552396953,0.325679034,0.1105527207,0.1268286854,0.2913684547,0.226491496,0.3070423007,-0.2636278272,0.0381210707,0.362011075,0.3629217744,0.4459484518,-0.2627597749,0.191279158,-0.3920520842,0.0556997843,-0.1566262394,0.1166174561,0.0137885697,-0.1983213276,0.1687487513,-0.2482326925,-0.1482825875,-0.1055952683,0.342610389,-0.2039889842,0.0377764106,0.0379048884,-0.3149587512,-0.0499389805,-0.0084228795,0.1511137933,0.1273792237,0.3696693182,-0.0651715845,-0.2228645533,-0.039334137,-0.2121529877,0.307115525,-0.2217801213,-0.1349204779,-0.1292728037,0.0492401011,0.157646358,0.163487494,0.5037110448,-0.0791685805,-0.0563689433,-0.1594543457,0.052819673,-0.2899908125,0.0324373357,0.0191640444,0.0829376131,0.0949392617,0.4496697485,-0.2343184203,-0.2162638754,-0.1967165768,0.5488077998,-0.3110198379,-0.157538861,-0.0917828381,-0.2865636349,-0.4673694074,0.0025114836,0.0468707085,-0.0605416894,0.1624680609,-0.3222727478,-0.058884345,-0.3417798579,0.0384316631,0.0353711918,0.6142191887,-0.2584322989,0.0886260793,-0.159195587,0.1720086336,0.4972103238,0.2424914539,-0.0417445004,0.0436437093,0.012087984,-0.3272880018,0.1730201095,0.0706722364,-0.2092171311,0.2919706106,-0.283752799,0.5059887171,0.0516601913,-0.0259855669,-0.1112033725,-0.3278073072,-0.0253595952,-0.2220413238,0.1087233126,0.0539661795,-0.1900668442,0.4209731817,-0.2525550127,-0.2213571966,0.4853894114,0.0631576031,1.0872696638,-0.0786438361,-0.0716803819,-0.182859838,-0.2409425825,0.3603461385,-0.6628372669,0.1060350761,0.0408539362,-0.0952326432,0.1190069616,-0.0351103544,0.0552499443,0.2388339639,-0.1510252357,0.0730217397,-0.3764982522,-0.2307166904,-0.2711305022,0.3491012156,-0.0473499671,-0.2726078629,-0.1125642285,0.0214013457,0.0035254774,-0.1006597281,-0.1841108799,-0.0700005889,0.3896320462,-0.0834498778,0.2893982232,-0.1460356116,-0.2104799449,0.385017693,-0.1758864224,-0.1661659181,-0.3307413757,0.0333319381,-0.251548171,0.1166307479,0.1659657657,0.0274507105,0.0241609812,0.0081873499,-0.3896660209,0.0696830675,-0.0762291029,-0.1722850651,0.0398670174,0.0838400275,-0.3224021494,-0.3170048892,0.2787468731,0.178682059,-0.1638534665,0.0439960659,-0.267300576,-0.1305264533,-0.4703465402,0.1084295362,0.1187369972,0.2714272738,-0.2142945379,0.2887547314,-0.2271916866,-0.2262046486,-0.2250813991,-0.1194844544,0.3463290334,-0.110394299,0.2839418948,-0.2810496986,-0.1979824603,-0.4381260574,-0.0492733531,-0.0164981112,-0.3980126679,0.4417989254,0.1036181673,-0.1894355416,0.1335241497,-0.119073458,0.0214704219,-0.0533580631,-0.1512570083,-0.204745248,-0.4557095766,0.1153083742,0.1973981559,0.5164298415,0.137516886,0.0063431384,-0.1514652073,0.1818770319,-0.2694254518,0.2495364547,0.1736499816,0.1584465653,-0.0758761242,0.2378975898,0.1391951144,-0.2206120193,-0.0551184453,0.1575634778,-0.0962062925,-0.2257269919,-0.1037328318,0.0981703103,-0.1585743576,-0.0587549023,-0.1905527562,-0.166731149,0.0773504674,-0.1729933918,0.2877326906,-0.180464372,-0.1214043945,-0.0593892597,0.4590342641,-0.046130579,0.0171195585,0.1909529269,-0.0611048341,0.3836376369,0.0219791047,0.0176089741,-0.0984174609,-0.0576920249,-0.417044431,0.0036445055,0.2138104439,0.0886856914,-0.0203117747,-0.019900199,-0.1027569324,0.0721066594,0.0828478038,0.2306800187,-0.0788651034,-0.144003123,-0.0770927295,0.2132747024,-0.1612409353,-0.2426515073,-0.0774638802,-0.0106664151,-0.2076933831,0.2267112136,0.4655177593,0.0151988324,0.1472037435,0.0966285914,0.3965957165,-0.2603293061,0.2932195067,0.2389336973,0.0456877425,-0.2071370035,-0.1511296183,0.2748292089,0.1894491613,-0.1020843163,-0.1602698863,0.3451376557,0.6201947331,0.3421747088,-0.5600329041,-0.3081383109,-0.2014006823,0.6582043767,-0.1195684671,0.5161323547,0.4168245494,0.2402068526,0.265820384,-0.2418255359,-0.2287273109,0.460924536,0.075206615,0.0131050935,0.0398535766,-0.1173838377,0.1491580456,0.139987573,0.1885392219,0.4059408903,-0.1243477762,0.2848824263,0.0402185917,0.2685131729,0.5135034323,0.0942838714,0.3942030966,-0.1765801311,-0.1202561781,0.2023445964,0.209495604,0.043447312,-0.0925873891,0.0065447856,0.2171600461,-0.2888061404,-0.0274791606,0.0623845607,0.1011189222,-0.2933245003,0.4945379496,0.2185930014,-0.1706920117,-0.0289129112,0.0775893629,-0.1158454195,0.2149564326,0.1214899048,0.2528750896,-0.0671386123,-0.247498557,0.3243519366,-0.4177721143,-0.0118302805,-0.1405366063,-0.1453604996,-0.032602787,0.2825759947,-0.2619815767,0.0701116547,-0.071001783,0.0898400545,-0.1798762828,0.3586507142,0.1456857622,-0.4000539184,-0.2147163004,-0.3193974793,-0.2431642711,-0.0725082532,-0.2650021315,0.102938354,0.2944024801,0.2489786446,-0.0428319089,0.1848064512,0.107291989,-0.0950152129,-0.1221372485,0.0062834122,-0.4662931263,0.097872071,0.2980542183,0.1921570301,0.1655493826,-0.4090610147,-0.2236194164,-0.08727175,0.0283139665,-0.2389190793,0.2523615062,0.0997848287,0.787630856,0.5745557547,-0.0286482405,-0.3066671789,-0.0792486668,-0.1591223925,0.0304990783,-0.1113064587,0.0105911475,-0.1794501394,0.4483243227,0.1139430255,-0.2500026822,0.2287590504,-0.2351906151,-0.1364353001,-0.1966729313,0.1696378142,-0.1612827927,-0.006680842,0.0953299329,0.3384085,-0.0501802489,0.2275591046,0.0938783884,0.1032033265,-0.3679366112,-0.0514399,0.3150177598,0.304167062,-0.3119878471,0.5644304752,0.1112729535,0.038876541,-0.0516906604,-0.3007228971,0.2683118284,0.158449769,-0.2732674479,-0.0172591954,0.3174834847,-0.0964615047,-0.10701444,0.1297712773,0.2575622201,0.4312021732,-0.0089225071,-0.3286861777,-0.2265087068]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"@thomwolf I'm sorry but I can't reproduce this problem. I'm also using: \r\n```python\r\nds = load_dataset(\"json\", split=\"train\", data_files=data_files, cache_dir=cache_dir)\r\n```\r\nafter having removed the extracted files:\r\n```python\r\nassert sorted((cache_dir \/ \"downloads\" \/ \"extracted\").iterdir()) == []\r\n```\r\n\r\nI get the logging message:\r\n```shell\r\nWARNING datasets.builder:builder.py:531 Reusing dataset json ...\r\n```","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":49,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\n@thomwolf I'm sorry but I can't reproduce this problem. I'm also using: \r\n```python\r\nds = load_dataset(\"json\", split=\"train\", data_files=data_files, cache_dir=cache_dir)\r\n```\r\nafter having removed the extracted files:\r\n```python\r\nassert sorted((cache_dir \/ \"downloads\" \/ \"extracted\").iterdir()) == []\r\n```\r\n\r\nI get the logging message:\r\n```shell\r\nWARNING datasets.builder:builder.py:531 Reusing dataset json ...\r\n```","embeddings":[0.0387398824,0.0011513438,-0.0731728151,0.2888256013,-0.0225971937,0.2442346811,0.0043715644,0.2252011448,0.211116001,0.1112262979,0.1300096661,0.4447046518,-0.3149459064,-0.152233839,-0.1879176497,0.0685452223,-0.1822635829,0.2303551137,0.1179905161,0.1538845748,-0.1485168636,0.0167127326,0.0281603187,-0.1277256161,-0.081199348,-0.3685379624,0.1793391258,-0.2001989782,-0.0709400177,-0.475566864,0.1935171932,0.4872877598,0.2634621859,0.2937750518,-0.0001270321,0.1402311623,0.2824359834,0.1706399322,-0.4946140647,0.073470667,-0.2185482085,-0.131723702,0.278342694,-0.1279088408,0.238957271,-0.5058012605,-0.2874324322,-0.7039566636,0.6001020074,-0.1444820762,0.0226126779,0.0881787464,-0.475264281,0.235341385,0.1795049906,0.3432503641,0.003216923,-0.014561764,0.397413075,0.0410049222,0.2530764639,0.3323472738,-0.1500108838,-0.2315425277,0.2820379734,-0.0412385538,-0.0308066588,-0.303809613,0.1847779602,-0.0036727618,0.6249611378,-0.3946924806,-0.255618304,-0.5423304439,0.0626548827,-0.6009432673,0.0232606772,0.279749155,-0.0825029537,0.2129976153,-0.1887128651,-0.3832755983,-0.3273220956,-0.1123421714,0.5984613895,-0.3998067677,-0.0372391976,-0.0458068214,0.4938412309,0.1461505145,0.2353725284,-0.3142920434,-0.2171897143,0.3846135437,-0.1873886138,-0.0936432704,-0.2239203453,-0.2821353078,0.1414932907,0.1589775831,0.2970098555,-0.1674041748,0.1185295656,0.0908483192,0.3935776651,0.4597049654,-0.0481675826,0.1128554419,0.100851737,0.1619086713,0.0482873619,-0.0979919359,0.1011848226,-0.0617838986,0.4250560403,-0.0321993716,0.0416536331,-0.107187964,-0.2610659897,0.0968772247,-0.0078233099,-0.2051997781,0.1666130871,0.1866741627,0.1723242849,0.0193716679,0.0672903135,0.4301665723,-0.0577309355,-0.2265523523,0.1032965183,0.1011031419,-0.3205889761,-0.3679001331,0.4237632453,-0.1298708916,-0.0259721037,-0.0048201988,-0.25508219,-0.0653327405,0.4840128422,-0.092918694,0.2538093925,0.3099845052,-0.2183120549,-0.0880336314,-0.2122648358,-0.1653785855,-0.1422930807,0.5295100808,-0.1849961281,-0.3411290646,-0.0856946632,0.0599792488,-0.0660016015,0.2800516188,-0.5685242414,0.0611999296,0.1911428124,0.0687458068,-0.041827403,0.2140063345,-0.0375086814,-0.0887151733,-0.2247154415,0.4024045765,-0.6254299879,-0.1333288848,-0.3174979091,-0.1452012062,0.2010465711,0.2420930415,-0.1749440432,0.3494221568,-0.315636009,0.0096579343,0.4006729126,-0.2328708619,-0.4543673396,0.1765417904,0.1190119013,0.4207156897,0.352494508,0.2543875277,0.2594757974,-0.0322538726,-0.0024930341,0.3350212872,0.0773673728,-0.0738491192,-0.2808282971,-0.3792129457,0.2256362885,0.1916751266,-0.0921854079,0.1105522737,0.1678093672,-0.0024405546,0.2609681487,0.1150512472,0.1132845357,0.1940720677,0.1154089645,-0.2010038793,-0.1774228364,0.1313792914,-0.5052556396,0.0986857489,-0.4657924771,-0.3321765363,-0.3756705225,-0.2849501669,0.038604483,0.1174927056,-0.389093101,0.1290554851,-0.0220173094,-0.1279522032,0.3280850351,0.0559472367,-0.3593132496,0.1520723104,-0.0983794555,0.0780223235,-0.5450092554,0.2050356865,0.1843023151,-0.2473080903,-0.1290709525,0.1400124282,-0.2306076288,-0.3627342284,-0.1654000431,0.2036164701,0.3391128778,0.2150984406,0.1560032964,0.2421350926,0.2304073125,0.2009599805,0.0469735563,0.1948219389,0.0115868561,0.0463874564,-0.2821243703,0.3812071681,-0.2219506055,0.0534807071,-0.1775822192,-0.4575609267,0.2535552382,-0.0570940077,0.079079181,-0.4799568653,0.1083344445,0.355214268,-0.0536249131,0.0245037731,0.1122616008,-0.1633778661,0.0036084938,-0.0947122648,-0.1975779533,0.2892430425,-0.1210098863,-0.1792408079,0.0539704822,0.4722078145,0.3872458041,0.0728120878,-0.0692079291,-0.1037085876,0.4217990637,0.0783267841,0.170580253,0.0563217103,-0.0379138552,0.3702068031,0.2599157393,-0.0925854817,-0.2354387939,0.0143223666,0.1043259799,0.2683180273,0.1289220899,-0.0334743373,-0.3978503048,-0.3873176575,0.3365691602,0.1967991292,0.1924797744,-0.146978572,0.3643906415,0.179900229,-0.2070634961,0.0305839833,-0.0449990556,0.4603663683,0.0073027909,-0.3709696531,-0.3444038033,-0.1632081717,0.1228643432,-0.0831061155,0.40229249,-0.2809521854,0.4440789819,0.1275637448,-0.0038780824,-0.0542435609,-0.0177100953,0.0376475342,0.3879231811,0.1876690239,0.1001782194,0.1531307995,0.0379875526,0.096379213,-0.0991242006,-0.144934833,-0.1022407338,0.3233754039,0.0661228448,0.1851321161,-0.1712577045,0.000553999,-0.2757932246,-0.2898995876,0.2013386488,0.0319607258,-0.0162346866,0.2689197361,0.2810165882,-0.0965232104,0.0841756389,0.1488934904,-0.0251834001,-0.7104842663,0.1607454866,-0.0769011527,0.1364143789,0.1890652776,-0.0697535276,-0.0293769576,0.3408835828,-0.4336383343,-0.0120831616,0.0660452917,0.1627597362,-0.2804273069,0.019199362,0.0958866924,0.2754341364,0.0378886424,0.0201795232,0.0985087976,0.1447329521,0.3198245764,0.1731146127,0.1262675673,0.2603633106,0.2043638676,0.4321883619,-0.0622241534,0.0196017493,0.3881896436,0.3163253069,0.4387296736,-0.2470133305,0.0989190862,-0.4738682806,0.0050000832,-0.2182047516,-0.003711178,0.1213676333,-0.0496471711,0.1735690981,-0.1063396111,-0.1391187608,-0.1404744089,0.2118045539,-0.3662151396,0.1025422439,0.0500670895,-0.2705192268,-0.1174218431,-0.1675273478,0.0391082242,0.1657995582,0.5044100285,-0.0209652334,-0.1951973289,-0.1079097316,-0.2236528099,0.2688923478,-0.1963935643,-0.1239866018,0.0256854612,0.0904650465,0.2192270756,0.2238593996,0.4910557568,-0.1035333648,0.0390723012,-0.0954004079,-0.0014426016,-0.2994738817,-0.0398616418,0.0502796806,0.0924927071,-0.0287446491,0.4003672898,-0.2519613802,-0.1651287973,-0.225709185,0.6906113029,-0.2601244152,-0.1884287298,-0.1495396197,-0.2630366087,-0.5465505123,-0.0200443156,-0.0080210408,-0.0236063395,0.0812036246,-0.2832101285,-0.1270304769,-0.2804139853,0.1166423634,-0.0564050488,0.5973092318,-0.2932857573,0.1725893617,-0.0703304857,0.0852039605,0.5068807006,0.3177331686,-0.0923689455,-0.0827543214,-0.1680626422,-0.198306933,0.1099878922,0.0935979113,-0.300325036,0.318071574,-0.2498796731,0.4409916401,-0.0315180682,0.0097209495,0.0271958038,-0.2729805708,-0.1154162362,-0.3969817758,0.0979595706,0.0724428222,-0.14415887,0.4777911007,-0.2132849097,-0.2477706224,0.4918858111,-0.1485553533,1.0456441641,-0.1345221549,-0.0002493626,-0.0578195825,-0.2254691869,0.4228838682,-0.6274007559,0.1990940571,-0.0431004241,-0.0408791862,0.1003762558,-0.1200109944,0.0230390541,0.315231353,-0.0527526215,0.2467504144,-0.4842722118,-0.1601080894,-0.3023614585,0.3316244483,-0.0553002656,-0.1308068782,-0.1981833577,-0.0560085736,0.0174176395,-0.0206172559,-0.2003168911,-0.0525973849,0.4607528746,-0.0518719628,0.2136147022,-0.1244360283,-0.2420250475,0.4143176675,-0.1312363595,-0.2538131475,-0.1891980171,0.1305609792,-0.2244033515,0.0189593863,0.2749981582,-0.0032376095,0.0313760489,-0.0009668538,-0.3601751029,0.105198659,0.0176961515,-0.0305391811,-0.0243017543,0.1501027793,-0.3183411062,-0.4439482689,0.1413226128,0.1588863134,-0.1130427793,-0.0460677147,-0.2940954864,-0.0910576209,-0.3285751641,0.0937715992,0.0375563391,0.3554032445,-0.2288431376,0.2944489121,-0.4424229264,-0.1805368811,-0.138742581,-0.1259227544,0.2781556845,-0.1078396142,0.4058580697,-0.1961036772,-0.1697500497,-0.3114447594,-0.109464474,-0.0591651313,-0.3523538411,0.4959714115,0.1541590989,0.019113889,0.0306142867,-0.1891189218,-0.112822257,-0.0036653695,-0.088388972,-0.2330772579,-0.5126236677,0.1333554834,0.2334330082,0.5337324142,0.0982188135,0.0723924637,-0.2063541263,0.0403027125,-0.1466146111,0.2719170153,0.271477133,0.1760983616,-0.0057530059,0.2576836646,0.1207275614,-0.2755824029,-0.1310796589,-0.0162492059,-0.1007203609,-0.113325499,-0.1385002285,0.1380104721,-0.085807085,-0.0821948871,-0.1708730906,-0.3415997624,0.1951053292,-0.2087672949,0.3060403168,-0.077454865,-0.0922983289,-0.046098277,0.3641610146,-0.1011658609,-0.0318115652,0.1295864731,-0.0777373612,0.3581880331,0.0935714915,0.0184435397,-0.024767952,0.0216320492,-0.5045372248,-0.0154079553,0.2184852809,0.06541159,0.1145337448,-0.1634781212,-0.0516781136,0.0544045828,0.1629258245,0.2467232645,-0.1528826654,-0.0133640971,-0.1064773723,0.1030094922,-0.188770026,-0.2644707561,0.0574011989,0.1221704707,-0.2538683116,0.3539162576,0.3832460046,0.0591589622,0.0459614433,0.0415378772,0.3936091065,-0.2163315415,0.4034718275,0.4159605801,0.0731046423,-0.2110335529,-0.1588383019,0.3227526844,0.1293343604,-0.0114391139,-0.1280555129,0.3922553658,0.588098228,0.1931765229,-0.6003850102,-0.3592983484,-0.2539199591,0.6064805388,-0.0488207899,0.4539389014,0.3565711677,0.2720523477,0.2334132642,-0.1520121396,-0.1246995926,0.4989540279,0.0763918012,-0.0607318878,0.1866460592,-0.1306658089,0.1558776349,0.131953463,0.0924912393,0.282676667,-0.2776646912,0.234921068,-0.0209263489,0.1108352691,0.4764104187,0.1960002184,0.4155327976,-0.1994318515,-0.1705277115,0.2443345189,0.2839078009,0.0255320743,-0.0954852328,0.1415087879,0.1138873771,-0.3440541923,-0.0115645174,0.0395423472,0.0908909142,-0.1983007342,0.4838029444,0.0526313595,-0.1877350956,-0.0884467065,-0.0205582418,-0.0426028818,0.1373265535,0.2074403316,0.2993078828,-0.2762259543,-0.055126749,0.2671678066,-0.3956130147,-0.0130177997,-0.0915183797,-0.1284047812,0.0006955406,0.4648071826,-0.188152045,0.0386157744,-0.1176983863,0.033162225,-0.2790227532,0.1828391105,0.2430692464,-0.325707674,-0.3032860458,-0.1775818616,-0.3610857129,-0.2281069756,-0.3321947753,0.0023377885,0.2257021517,0.1908510476,-0.0866645724,0.1338694692,0.1663001329,-0.0277514402,-0.0381055735,0.0489809141,-0.3745660186,-0.0142569449,0.2151213139,0.1890990287,0.0779751539,-0.4346445799,-0.3108427227,-0.1040524691,-0.07542523,-0.082354635,0.2746676803,0.1471285969,0.7400567532,0.6187962294,-0.0363841057,-0.2140893936,0.0963764414,-0.2081695944,-0.0401764289,-0.1646408141,-0.0168137215,-0.1596775204,0.2693034112,0.2698897421,-0.4261137545,0.2645615935,-0.3312882781,-0.2856910527,-0.1148092672,0.0231930166,-0.2574584782,-0.021233052,-0.0297408085,0.5283959508,0.009937223,0.3121955991,0.0479232296,0.2261252254,-0.3168373108,-0.0664727464,0.3342604339,0.1714877337,-0.3807065487,0.5087202787,0.0780476108,0.0506935678,-0.051804103,-0.3348765075,0.3184893429,0.2205947936,-0.3912208676,0.0043192254,0.2257652581,-0.0556488,-0.0024563391,0.1746977866,0.2327556461,0.4356971979,0.0615052469,-0.2619682252,-0.1745367646]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"Do you confirm the extracted folder stays empty after reloading?","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":10,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\nDo you confirm the extracted folder stays empty after reloading?","embeddings":[-0.0339347236,-0.0321065299,-0.1424028873,0.222521916,-0.0838347748,0.1249039099,-0.1978568435,0.3118233979,0.1508260369,0.1597436368,0.070696786,0.4937373698,-0.1997277737,-0.0565983392,-0.2345223874,0.0499252304,-0.2507733107,0.2931573391,-0.0971169919,0.1164042726,-0.1603369266,0.0832878873,0.0414299928,-0.1676299721,0.0316394679,-0.4089366794,0.1983518451,-0.1657929271,-0.1894752234,-0.30450055,-0.0086474437,0.4840610623,0.3092184067,0.2415572256,-0.0001169057,-0.0542602278,0.3150730729,0.0321099907,-0.4199542105,0.1548909694,-0.2197697312,-0.1091276258,0.2436125278,-0.1722438633,0.2602371573,-0.3404246867,-0.2775099576,-0.6205130219,0.5009018779,-0.2474998236,0.1500027776,-0.0156021276,-0.5308664441,0.1198812425,0.2550978661,0.37518242,-0.1327349991,-0.0232548025,0.4390012026,0.0248183329,0.207804665,0.3292341828,-0.0636956468,-0.2329670936,0.3487396836,0.0023209595,-0.1265799403,-0.2416616827,0.316162616,-0.0095531866,0.6659325361,-0.3607134223,-0.0608702302,-0.2960805297,0.0758794546,-0.5820260048,0.1373921782,0.1668084413,0.0008975066,0.2459333092,-0.1088263243,-0.4359556139,-0.3570625484,-0.1761105955,0.5899226665,-0.4265705347,-0.1314673126,-0.1410305202,0.4829437435,0.1108378544,0.205859378,-0.2231386304,-0.2763829231,0.3679029346,-0.0703587979,-0.1786636114,-0.3136783242,-0.2947987914,0.179433912,0.1920688599,0.3963260651,0.0316002071,0.1785330176,-0.0174679905,0.3233247697,0.44137308,0.1142638624,-0.0125507582,0.1842877716,0.1265815049,0.0552533492,-0.0400088392,0.0086371973,0.0249372814,0.2454721183,0.002941648,-0.0348385535,-0.1688299477,-0.0432852805,0.0686518997,0.229301393,-0.2504942715,0.0209703501,0.213687852,0.2028482407,0.0523477867,0.0751218945,0.2780378759,-0.0020240666,-0.3200554252,0.0788812712,0.1550759971,-0.3619102836,-0.2551314533,0.4356802702,-0.0779297575,-0.0776653513,-0.1424138397,-0.3281650543,0.0423428342,0.4609901607,-0.1098226234,0.276617229,0.2670713365,-0.1292324513,-0.1778230071,-0.2329877615,-0.0704250112,-0.1184445247,0.4533665478,-0.1574326903,-0.2417992204,-0.0483578667,0.1761185974,0.1201338395,0.2642552853,-0.5853342414,0.0771192014,0.0616125129,0.088897258,-0.0177826937,0.3706976771,0.0358544774,-0.100949049,-0.2623008788,0.2925013602,-0.6783759594,-0.012082248,-0.3447231352,-0.2517929077,0.2447189391,0.1862080544,-0.1275531203,0.2607353628,-0.1822628826,0.1797662675,0.3226360381,-0.1090324745,-0.4004465342,0.1686857641,0.0486776121,0.2267651409,0.311960727,0.1936802566,0.3962551355,0.0182836168,-0.0208471045,0.2747553885,0.0875584558,-0.0827371925,-0.3818285167,-0.3359277546,0.0579500496,0.197093904,0.0001516875,0.0662374794,0.209134236,-0.0093571553,0.191854313,0.1080295444,0.1617481858,0.1914378107,0.171369195,-0.2636695504,-0.1998980194,0.1721000969,-0.5970097184,0.0133791389,-0.4932833314,-0.3083331585,-0.1961345971,-0.3287421465,0.0595691316,0.0182997026,-0.2122669518,0.1445482522,0.0974947885,-0.0380274877,0.2254028618,-0.0490453094,-0.3902689219,-0.0074384934,-0.0432441123,0.0044179722,-0.491864413,0.178683579,0.1098586619,-0.2649899721,0.0161313415,-0.0383429676,-0.2490062416,-0.5349908471,-0.1522434354,0.2149124295,0.316139102,0.2839004993,0.2724832892,0.2195220292,0.1518822461,0.2139593661,0.1098077968,0.1306955516,0.0348166041,0.055633463,-0.3905975819,0.415523231,-0.2166260183,-0.0517079793,-0.1640699357,-0.4370445013,0.2481320798,-0.1282289624,0.1633681357,-0.5076729655,0.0639505684,0.3719095886,-0.1636207104,0.0603174828,0.1433140934,-0.0213713367,0.0243366323,-0.0373975076,-0.1714535505,0.3396700621,-0.0488783345,-0.2036289871,0.0009247757,0.4247359633,0.3761192858,0.1886811703,-0.1509467512,-0.178158775,0.4626131952,-0.0126464739,0.1479016244,0.0331653431,0.0371594504,0.3967099488,0.091580689,0.0264063831,-0.3068019748,-0.0451393574,0.1205542088,0.2640251219,0.2750835717,-0.2134003937,-0.2740192711,-0.4003961682,0.5293626189,0.301448971,0.2741215825,-0.0093505941,0.4947927594,0.1254117787,-0.3915703595,0.1068723872,0.0611858331,0.6266986728,-0.0985660106,-0.1802399755,-0.4005054533,-0.1673098207,0.1254660338,0.0381223187,0.3536480665,-0.0948148519,0.4489184916,0.1089851111,0.1347754449,-0.0854839906,-0.0380686261,0.1020324305,0.3293169141,-0.0443261974,0.0848489404,0.1939603835,-0.0018620794,0.0757690743,-0.1620248854,-0.3072976768,-0.1770484,0.2047683001,0.1048506722,-0.0259274561,-0.2306853831,-0.1221353859,-0.1456914246,-0.4128332436,0.2120872885,0.0007541197,0.0018379254,0.0693937466,0.2076062411,-0.0234136805,0.0258013811,0.0348450914,-0.0827021003,-0.62780267,0.1363094598,-0.1785147786,-0.0146830678,0.2013302892,0.0465413071,-0.0104229348,0.2836548686,-0.3474417031,0.0406226404,0.0282029137,0.1582117677,-0.3088590205,-0.07539884,0.246584937,0.2341347039,-0.1350022256,0.0326218233,0.0555384867,0.2230853587,0.3853607178,0.1292105913,0.003032018,0.2501350641,0.1646479964,0.3698381782,-0.2696693838,0.0126559436,0.3677711487,0.3846020103,0.4382919073,-0.2540551126,0.1956698895,-0.4235919118,-0.0161863677,-0.1037967429,0.1453959495,0.052584853,-0.1759960502,0.1383547634,-0.2614876032,-0.1945859492,-0.0930774733,0.3519116044,-0.1752182692,0.0360488221,0.0575147681,-0.2878704667,-0.0548051931,-0.0366332047,0.1220100001,0.1393301785,0.3761164844,-0.0984567404,-0.3003518581,-0.0490374155,-0.1942049265,0.2985492051,-0.228255704,-0.1954660565,-0.0681079403,-0.0030981782,0.1729673743,0.2055794597,0.3765581548,-0.1325200796,0.0145092113,-0.1390942633,-0.0069828918,-0.2180545777,0.0319128484,-0.0563877225,0.0398341566,0.0601739995,0.4271597564,-0.2409655452,-0.1800150275,-0.2330746651,0.5565924048,-0.2762378156,-0.1516360939,-0.0730977282,-0.2022846639,-0.5152980089,0.0120610232,0.0821645409,-0.0658723488,0.1611908674,-0.2780106962,-0.0528528281,-0.3522838354,0.0736717507,0.0603414029,0.5813547969,-0.24583444,0.1322606057,-0.1596652269,0.1895459592,0.5604227781,0.2829778194,-0.1018739492,0.0608400553,-0.007919617,-0.3561514318,0.141359672,0.1467725933,-0.1746397614,0.2231919169,-0.3076801002,0.5163421631,-0.0104743121,-0.0485960618,-0.167415157,-0.3071737587,-0.0143274153,-0.1242671609,0.0868401229,-0.0780904517,-0.1807793826,0.4748015106,-0.116220355,-0.2204426378,0.4675077498,0.0113607515,1.1482275724,-0.053818047,-0.0114424573,-0.1637801677,-0.2676151991,0.2586925328,-0.6262396574,0.1069991887,-0.0095731309,-0.1637341827,0.1230361834,-0.0382339284,0.0021652933,0.2338350117,-0.2251100838,0.0444598421,-0.3625370264,-0.2191787064,-0.2678856254,0.4301961958,-0.0519308783,-0.2422775179,-0.0834073573,0.0181046519,0.0445573889,-0.0359052494,-0.2220378518,-0.0931683332,0.3934630752,-0.0936233327,0.2759169936,-0.0467700362,-0.182510674,0.3509512544,-0.1445247829,-0.1294530034,-0.2664987445,-0.0370047912,-0.2365167141,0.0477952361,0.1510478258,0.0223312527,0.0103517007,-0.0078882845,-0.3724478781,0.0582589656,-0.1029681861,-0.1540122032,0.0165472254,0.1027982309,-0.3006557226,-0.2949490845,0.1755788475,0.1338968277,-0.1199818999,-0.0334898047,-0.2382046431,-0.112195313,-0.5160883665,0.1051663011,0.1275219321,0.3113469183,-0.1856852174,0.3354303539,-0.1698283255,-0.1695794612,-0.2421710938,-0.0310666431,0.3106420934,-0.140077427,0.2755408287,-0.1774081737,-0.228928566,-0.4889847636,0.0403678939,0.026199799,-0.3237668872,0.4313735962,0.0246870238,-0.1679392159,0.151496619,-0.0648375824,0.0741268545,-0.0671295077,-0.1372533143,-0.1496983021,-0.3986370265,0.2477094531,0.212251246,0.494677037,0.0375277624,0.0146422209,-0.0178630557,0.1965650618,-0.2844516337,0.2501190603,0.2395133823,0.1806339771,0.0086932676,0.2311235368,0.0761902258,-0.2496609837,-0.039283406,0.0628472641,-0.1341130286,-0.2104268372,-0.1002223566,0.094151102,-0.1195330322,-0.0364034511,-0.1764139682,-0.2121392936,0.0520071313,-0.1674718261,0.189329192,-0.2362115085,-0.1302536279,-0.1060411409,0.4197223485,-0.0927047655,-0.0445256718,0.1140275002,-0.1142287329,0.3744427264,-0.0224107578,-0.0537372343,-0.0044509131,-0.102406323,-0.3928289115,0.0416296981,0.2210543603,0.1432604492,0.0124871368,-0.025142597,-0.0927220285,0.0445531644,0.0757172629,0.3196428716,-0.0939631611,-0.1378502548,-0.0273172744,0.240393132,-0.2452101856,-0.2192005068,-0.1430390775,0.0222980157,-0.2150977999,0.2858692706,0.2932223976,0.0245836582,0.1109957621,0.0894291252,0.4070224762,-0.2358137369,0.3415766656,0.2267441452,-0.0291797034,-0.2333878875,-0.0855746493,0.2847931683,0.081782937,-0.0736831278,-0.1199501827,0.3711307347,0.5408267975,0.3548240662,-0.5109686852,-0.2650975287,-0.2891122997,0.6113002896,-0.0903850347,0.4628599584,0.4420453906,0.1348528862,0.2384211421,-0.2727707028,-0.2551753521,0.3687810004,0.0546091683,0.0859150141,0.0196470115,-0.1348277777,0.1444601119,0.1185986847,0.1572353989,0.4225819409,-0.0363157578,0.3101007342,0.0163663775,0.1988629103,0.563575387,0.134186551,0.3821549118,-0.1692348123,-0.0848783404,0.1762700826,0.261141777,0.0151888048,-0.05833726,-0.0333093219,0.2449192405,-0.3175620139,-0.0112340068,0.0192963891,0.0231955592,-0.2916857302,0.4890724421,0.2289706916,-0.2178553939,-0.0230298527,0.1015435904,-0.1248281151,0.1776527315,0.149572432,0.153425619,-0.0377111435,-0.2123019397,0.3743781149,-0.3974481225,-0.0033886915,-0.1837846786,-0.1595491916,0.0353924595,0.2818683982,-0.2629832327,0.0362133123,-0.1050778329,0.0889735743,-0.163864091,0.3084605336,0.1806838065,-0.42728284,-0.2450924665,-0.3339603245,-0.2720536292,-0.1136254817,-0.2414874583,-0.0186005421,0.183016628,0.3084487915,-0.0153024513,0.2158743888,0.2325480282,-0.0727788508,-0.1142914072,0.0278166123,-0.5570346713,0.1527901143,0.3619458079,0.2548662126,0.157000035,-0.4414370954,-0.2218935639,-0.0079479879,0.0477314629,-0.1625893414,0.1947262138,0.0612317882,0.7589712739,0.6183800697,-0.0998335183,-0.3007375896,-0.1286147684,-0.1585754305,-0.0296748355,-0.1233728603,0.0270497054,-0.1418568343,0.5546939373,0.1910949498,-0.2811216414,0.262784481,-0.2744608819,-0.2622798383,-0.1670828611,0.2609794438,-0.2399084419,0.0374858789,0.0994686708,0.4447906911,-0.0321784206,0.2319188267,0.0440694578,0.0707565621,-0.4065268338,-0.0261638518,0.312690556,0.3440623879,-0.4073700309,0.5127193332,0.1449337304,0.074493289,-0.046566125,-0.2502864599,0.2750017047,0.1037480906,-0.2694576681,-0.0053661531,0.2735577524,-0.0991428271,-0.0856900215,0.0945449769,0.2164306492,0.4260048866,-0.000826514,-0.4019800425,-0.2154175788]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2604","title":"Add option to delete temporary files (e.g. extracted files) when loading dataset","comments":"> \r\n> \r\n> Do you confirm the extracted folder stays empty after reloading?\r\n\r\nYes, I have the above mentioned assertion on the emptiness of the extracted folder:\r\n```python\r\nassert sorted((cache_dir \/ \"downloads\" \/ \"extracted\").iterdir()) == []\r\n```\r\n","body":"I'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.","comment_length":37,"text":"Add option to delete temporary files (e.g. extracted files) when loading dataset\nI'm loading a dataset constituted of 44 GB of compressed JSON files.\r\n\r\nWhen loading the dataset with the JSON script, extracting the files create about 200 GB of uncompressed files before creating the 180GB of arrow cache tables\r\n\r\nHaving a simple way to delete the extracted files after usage (or even better, to stream extraction\/delete) would be nice to avoid disk cluter.\r\n\r\nI can maybe tackle this one in the JSON script unless you want a more general solution.\n> \r\n> \r\n> Do you confirm the extracted folder stays empty after reloading?\r\n\r\nYes, I have the above mentioned assertion on the emptiness of the extracted folder:\r\n```python\r\nassert sorted((cache_dir \/ \"downloads\" \/ \"extracted\").iterdir()) == []\r\n```\r\n","embeddings":[0.0413070992,-0.0485373661,-0.144972533,0.224829793,-0.0466053225,0.2017893344,-0.1690857112,0.3027119935,0.2245082706,0.1342939734,0.0109500363,0.480185926,-0.268691659,-0.1230428964,-0.1824239641,-0.0093835704,-0.2349088937,0.2767875493,-0.003963063,0.1206632704,-0.1744759232,0.0921203047,0.0378465205,-0.1965695918,0.0525064655,-0.4080628157,0.1274044961,-0.0839395151,-0.1356525421,-0.4438681006,0.0438200198,0.4613906145,0.265135318,0.2756152451,-0.0001182811,-0.0052507161,0.321416229,0.0448974855,-0.4485212564,0.0787613317,-0.1391087621,-0.168904081,0.2623524964,-0.1987499744,0.223350808,-0.3563092947,-0.293863833,-0.6617410779,0.5167124271,-0.1608560979,0.1268815249,0.0683706999,-0.4738253355,0.1933883727,0.2738461792,0.2409303933,-0.0916814432,0.0385858454,0.3927947581,0.0045666494,0.2632922828,0.3668141961,-0.1002369002,-0.2114969343,0.3557851017,0.0218768865,-0.1461001784,-0.32141608,0.2498360574,-0.0287617948,0.6228607297,-0.3944323063,-0.2275240719,-0.3693217933,0.0613993891,-0.6855837107,0.0714127496,0.2646585107,-0.0256918222,0.238250345,-0.0805310756,-0.374007225,-0.3292851746,-0.1420151889,0.5611350536,-0.3349307179,-0.0679379329,-0.1523856223,0.4347500205,0.1455315799,0.2015200257,-0.2469354272,-0.2134846151,0.4398026466,-0.1156603768,-0.1684402525,-0.2826519608,-0.255082339,0.1541438252,0.2384481281,0.3873394728,-0.0318554565,0.1681101024,0.0122853788,0.302649945,0.4701262712,0.0035454736,0.0091398135,0.1823690534,0.0923326388,0.0784320012,-0.0593405664,-0.0071670548,0.0037678275,0.3564743698,-0.0054526506,-0.0003948941,-0.1036438271,-0.0663218647,0.1109705269,0.1465508342,-0.1764888465,0.0581149533,0.2550421357,0.1115225703,0.0797707289,0.1175451726,0.3180947602,-0.006252687,-0.1697294563,0.0901406556,0.1622937173,-0.2889664173,-0.3153970838,0.4541114271,-0.0812455714,0.0166768096,-0.1212240458,-0.2307779342,0.0247761011,0.5310213566,-0.1226880178,0.2855815589,0.3619463146,-0.1461788714,-0.1604757607,-0.1923618317,-0.0846228227,-0.1570649445,0.4940737188,-0.115770638,-0.2477866113,-0.0082167536,0.1721287072,0.0101683578,0.2284962088,-0.5198004842,0.0128179165,0.1061698347,0.0977833718,0.0089794928,0.3370685279,0.0592684001,-0.1662468165,-0.2143195868,0.360255301,-0.5754243135,-0.0783924982,-0.3162104785,-0.1381838024,0.1419519037,0.1934991479,-0.0748170614,0.2696565092,-0.2039485574,0.1503590345,0.3257901073,-0.2086213082,-0.4496496916,0.1334092915,0.0934550762,0.3512125611,0.2929980159,0.2657374442,0.3637495339,0.0370408632,0.0280236807,0.3027046919,0.0446932241,-0.0319463052,-0.3799369037,-0.3065942824,0.1273216903,0.2023749799,0.0401778482,0.0616224483,0.1113649085,-0.0308885518,0.241564095,0.1242302954,0.1132445931,0.2096996307,0.2101258337,-0.2184161395,-0.1411334872,0.0758418366,-0.533705771,0.1161193699,-0.4802550673,-0.2700078785,-0.3258729875,-0.3545676172,-0.0062890369,0.06039932,-0.2571125031,0.0709158108,0.100192897,-0.0983826444,0.3363024294,0.0053298594,-0.4017877579,0.0917589441,-0.0026925183,0.0702562034,-0.442569524,0.1793133169,0.1284513772,-0.2546826303,-0.0913694054,0.0198447704,-0.2252409309,-0.5050381422,-0.1748074442,0.2506622672,0.3291651309,0.233689636,0.2079574913,0.2225165367,0.1781661958,0.2550555766,0.0977902487,0.2198260725,0.0716378614,0.0666633397,-0.3593182564,0.4581740201,-0.226888895,-0.0226653814,-0.1261266172,-0.3892331421,0.2734319866,-0.1001379341,0.0814294145,-0.4512613416,0.0787403956,0.3533690274,-0.0491033792,0.0033766234,0.1296355873,-0.0848263949,-0.0107796937,-0.0596405938,-0.1538949609,0.3206886351,-0.1086348519,-0.1442068368,0.0110897776,0.4614611566,0.4385591745,0.1701882631,-0.1012916416,-0.11482694,0.4571801722,0.0365350693,0.181687206,0.0530165993,-0.048828043,0.3767079115,0.1704260856,0.0017932453,-0.3105232716,-0.0254260805,0.086679019,0.2245713323,0.2444339395,-0.1599863619,-0.2307387292,-0.3647020459,0.4693772793,0.1976751983,0.2197479308,-0.0726308525,0.4689161777,0.1539280713,-0.326340884,0.0959923044,-0.0333921723,0.5423595309,-0.01706874,-0.3114193976,-0.377166003,-0.1681100577,0.2098544985,0.0186219271,0.3155292273,-0.1995816827,0.4949300885,0.08884909,0.0585311837,-0.0464112572,-0.0641171038,0.0497560054,0.2917626202,0.0272737369,0.1036752015,0.2092254162,0.022574693,0.1079359278,-0.094643496,-0.3111015558,-0.1514032483,0.2132693827,0.1755683869,0.0509327352,-0.2745606899,-0.1111386269,-0.2817261815,-0.4231975377,0.2254269272,0.0398009978,0.0331125483,0.1260128021,0.1688347161,-0.061066892,0.1147645488,0.081115447,-0.0450237058,-0.6299257278,0.1831360906,-0.1912486255,-0.0007606788,0.2425493896,-0.0675699636,-0.0057533085,0.3238607943,-0.3674845099,-0.0564071722,-0.0332863927,0.1807260811,-0.3109463751,-0.02231168,0.2386684567,0.2549817562,-0.1006572545,-0.0001963221,0.0948112383,0.1612852961,0.3627622128,0.0858090073,0.0152382841,0.2255530208,0.1552601904,0.4575290084,-0.2319746614,0.0894569457,0.3778058887,0.3623319566,0.483091712,-0.2674523592,0.1350191981,-0.476162523,-0.059716668,-0.1635607779,0.0614998117,0.1268329322,-0.1489725262,0.1271566451,-0.2267463803,-0.2554056048,-0.1087194383,0.34884426,-0.2455918789,0.0747066215,0.0337011926,-0.2601059973,-0.1106662527,-0.0750344023,0.05254611,0.1934334636,0.3555967212,-0.0959618539,-0.2293903232,-0.0704648569,-0.2582458258,0.2436961234,-0.18421942,-0.1575855315,0.050778091,0.0009019715,0.1316079348,0.170419246,0.4188086689,-0.1102035493,-0.0316157676,-0.0432014652,-0.0491015911,-0.2723814547,0.0143882092,-0.043687582,0.051381588,0.0062427637,0.3517798185,-0.2450859547,-0.1610225588,-0.2627820671,0.6113590598,-0.2914118767,-0.146190241,-0.1679039747,-0.2758182287,-0.5205757022,0.0137090394,0.1024279967,-0.0699496046,0.1287714094,-0.2261886299,-0.116565831,-0.3135035038,0.0955886841,0.0261205658,0.6730561852,-0.3106630445,0.1338148117,-0.1499756277,0.1353078932,0.4575920403,0.3512052894,-0.1634143293,0.0324225202,-0.1253552586,-0.266061455,0.1117492691,0.1593424529,-0.2497403324,0.2355032861,-0.3442848921,0.4965210855,-0.001997473,-0.0435360484,-0.1301544309,-0.252844125,-0.0031907249,-0.2554462552,0.1030301452,-0.021974314,-0.1552110314,0.4835703969,-0.1392625868,-0.25539276,0.5033291578,-0.0633165091,1.1462258101,-0.091255486,0.0590803474,-0.0746246874,-0.2671490312,0.294193089,-0.610106647,0.1928513497,-0.0238334946,-0.1474044025,0.1032523587,-0.0321945883,0.0073783016,0.1978934705,-0.2577227056,0.1724757403,-0.4044649005,-0.2458304465,-0.3127624691,0.3859443069,-0.1336411685,-0.2389241755,-0.2229070663,0.0304174349,0.0703630075,0.0096992524,-0.2433690429,-0.0699395463,0.3483138382,-0.1029407606,0.2205169499,-0.1050838009,-0.211561963,0.3625518978,-0.1709127873,-0.1693255603,-0.3747052252,0.0147155654,-0.2277475148,0.0665274113,0.177156806,0.0191218611,-0.0063906428,-0.0273664538,-0.3108091354,0.084057115,-0.0238340534,-0.1177816316,0.0023079407,0.1461178362,-0.3547461629,-0.3214163482,0.2032022327,0.1750044227,-0.1204068288,0.0108268075,-0.3190675378,-0.1161417365,-0.5043124557,0.0812235996,0.114505522,0.3476789594,-0.2278478891,0.3304807544,-0.291590184,-0.2250877023,-0.2206318825,-0.0532794893,0.3131303489,-0.1598387808,0.3334147632,-0.2019819319,-0.2323346287,-0.4242372811,0.0082435263,-0.0975854248,-0.2666522264,0.4921968579,0.0498114564,-0.1244077384,0.0569743961,-0.1461435109,0.085146904,-0.0082857171,-0.1706765592,-0.1901723444,-0.4423626959,0.2361502349,0.3011802733,0.5027703047,0.0333814882,-0.0045292024,-0.0241315272,0.1231794953,-0.2732389867,0.2419102639,0.169009909,0.1864479333,-0.0108399922,0.2098039538,0.1146928519,-0.2905042768,-0.0424861349,0.028029928,-0.1194052994,-0.2052805722,-0.119458206,0.110072948,-0.1223045364,-0.0890904292,-0.1279248744,-0.2548086941,0.1514477581,-0.1967034936,0.2219270021,-0.1408964545,-0.1152803376,-0.0382679813,0.4236804545,-0.0534901656,-0.06674961,0.1244280413,-0.1753025055,0.3281989396,0.0484317504,-0.0580317676,-0.0151760289,-0.0559296682,-0.3425908983,-0.0125505198,0.1476258337,0.0793365985,0.0730557367,-0.0756691396,-0.0917990953,0.0424776338,0.1605764478,0.3133683503,-0.1005902216,-0.0278405175,-0.015204424,0.235552609,-0.3078093231,-0.2028858066,-0.0959709734,0.0422857478,-0.1282796711,0.3526952565,0.2551882863,0.0160210226,0.0785968229,0.1219502091,0.4435407221,-0.262561053,0.3567343056,0.2330165505,0.0335210823,-0.203456223,-0.140412882,0.2835745215,0.0813073963,-0.0411897786,-0.1353509426,0.4125169516,0.5396201611,0.2706173062,-0.5496681929,-0.2321303934,-0.3047351837,0.5648602247,-0.0636367276,0.4833982289,0.4112482965,0.1609768867,0.1994054914,-0.2693638504,-0.2112763375,0.3925042748,0.0376014337,0.0151275843,0.1779447794,-0.1225665063,0.1545842886,0.0890968814,0.1928014904,0.3150663078,-0.0763057321,0.3050920069,0.0403472185,0.1444400996,0.5403456092,0.1726736724,0.4122552276,-0.2008505017,-0.1683285236,0.3069994748,0.2286299467,-0.0467792675,-0.0426460691,0.0607595369,0.2300120741,-0.4002389014,0.0562206097,0.0291171428,0.036165338,-0.2809961736,0.4738404155,0.1764168441,-0.1633598506,-0.0664999783,0.0892370194,-0.1348403841,0.2295845002,0.1274809986,0.1664561629,-0.1332194954,-0.095339939,0.3460355103,-0.4161977768,-0.0393494889,-0.1494923979,-0.2405465394,0.028394578,0.3042207658,-0.1859047115,0.0206843931,-0.0995517522,0.087401256,-0.1506767422,0.2765628397,0.230437845,-0.3117170632,-0.2855139375,-0.2408074737,-0.3313231468,-0.1924170703,-0.3334586918,-0.0793614164,0.203191191,0.2810607255,-0.0578947701,0.2194302082,0.2072461247,-0.0437058285,-0.132068634,0.0096003152,-0.5555422306,0.0932331309,0.2999010086,0.2253646702,0.1715487242,-0.5085917115,-0.290487498,-0.0740842819,0.0490977913,-0.1099371016,0.1973835826,0.1196563244,0.7663525343,0.6555728912,-0.082328923,-0.2813947201,-0.0887070969,-0.2264493555,-0.0294566136,-0.1210761219,-0.0183540117,-0.1915733367,0.4612554312,0.2556870282,-0.3579712212,0.1751772016,-0.2629187405,-0.2018046975,-0.1126138642,0.1124970391,-0.2136930674,0.0236826874,0.0916536823,0.4669551849,-0.0726004466,0.322886765,-0.0376899764,0.1340642869,-0.3692810833,-0.1002738401,0.3550425172,0.253477335,-0.4091673791,0.4840549529,0.1112580597,0.0214007497,-0.0204094499,-0.2937963009,0.3100440204,0.1386470497,-0.2728059888,-0.051329799,0.2545327544,-0.1060625985,-0.0809822828,0.1383152455,0.2301093936,0.349155426,0.0669758096,-0.3895577788,-0.2239033431]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2598","title":"Unable to download omp dataset","comments":"Hi @erikadistefano , thanks for reporting the issue.\r\n\r\nI have created a Pull Request that should fix it. \r\n\r\nOnce merged into master, feel free to update your installed `datasets` library (either by installing it from our GitHub master branch or waiting until our next release) to be able to load omp dataset.","body":"## Describe the bug\r\nThe omp dataset cannot be downloaded because of a DuplicatedKeysError\r\n\r\n## Steps to reproduce the bug\r\nfrom datasets import load_dataset\r\nomp = load_dataset('omp', 'posts_labeled')\r\nprint(omp)\r\n\r\n## Expected results\r\nThis code should download the omp dataset and print the dictionary\r\n\r\n## Actual results\r\nDownloading and preparing dataset omp\/posts_labeled (download: 1.27 MiB, generated: 13.31 MiB, post-processed: Unknown size, total: 14.58 MiB) to \/home\/erika_distefano\/.cache\/huggingface\/datasets\/omp\/posts_labeled\/1.1.0\/2fe5b067be3bff1d4588d5b0cbb9b5b22ae1b9d5b026a8ff572cd389f862735b...\r\n0 examples [00:00, ? examples\/s]2021-07-06 09:43:55.868815: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last): \r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 990, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py\", line 338, in write\r\n self.check_duplicate_keys()\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 3326\r\nKeys should be unique and deterministic in nature\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"hf_datasets.py\", line 32, in \r\n omp = load_dataset('omp', 'posts_labeled')\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 3326\r\nKeys should be unique and deterministic in nature\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Ubuntu 18.04.4 LTS\r\n- Python version: 3.6.9\r\n- PyArrow version: 3.0.0\r\n","comment_length":52,"text":"Unable to download omp dataset\n## Describe the bug\r\nThe omp dataset cannot be downloaded because of a DuplicatedKeysError\r\n\r\n## Steps to reproduce the bug\r\nfrom datasets import load_dataset\r\nomp = load_dataset('omp', 'posts_labeled')\r\nprint(omp)\r\n\r\n## Expected results\r\nThis code should download the omp dataset and print the dictionary\r\n\r\n## Actual results\r\nDownloading and preparing dataset omp\/posts_labeled (download: 1.27 MiB, generated: 13.31 MiB, post-processed: Unknown size, total: 14.58 MiB) to \/home\/erika_distefano\/.cache\/huggingface\/datasets\/omp\/posts_labeled\/1.1.0\/2fe5b067be3bff1d4588d5b0cbb9b5b22ae1b9d5b026a8ff572cd389f862735b...\r\n0 examples [00:00, ? examples\/s]2021-07-06 09:43:55.868815: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last): \r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 990, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py\", line 338, in write\r\n self.check_duplicate_keys()\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 3326\r\nKeys should be unique and deterministic in nature\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"hf_datasets.py\", line 32, in \r\n omp = load_dataset('omp', 'posts_labeled')\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/erika_distefano\/.local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 3326\r\nKeys should be unique and deterministic in nature\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Ubuntu 18.04.4 LTS\r\n- Python version: 3.6.9\r\n- PyArrow version: 3.0.0\r\n\nHi @erikadistefano , thanks for reporting the issue.\r\n\r\nI have created a Pull Request that should fix it. \r\n\r\nOnce merged into master, feel free to update your installed `datasets` library (either by installing it from our GitHub master branch or waiting until our next release) to be able to load omp dataset.","embeddings":[-0.2633759081,-0.1163594574,-0.0825288817,0.100509882,0.2311997861,-0.021742953,0.205230996,0.2830139399,0.0483129919,0.2094642371,-0.2077926248,0.5524355173,-0.2091248482,0.2359020114,0.1329852045,-0.2663642764,-0.1581268013,0.1020685211,-0.3342058063,-0.0926270857,-0.1104445532,0.2932712138,-0.1778841317,0.0904801488,0.0324104577,-0.1805106699,-0.0745370686,-0.0469907336,-0.2778788805,-0.3463790417,0.0518865064,0.0548791774,-0.1014625952,0.3988808095,-0.0001096027,0.117816031,0.0741746128,0.0171126556,-0.1459125131,-0.67196244,-0.1396107972,-0.2014088035,-0.0826201811,-0.363342911,0.2272390574,-0.2570011616,0.0594070926,-0.2267066985,0.435379833,0.3867657781,0.2739731073,0.1229682267,0.5405081511,-0.0700387582,0.13682051,-0.1091347188,-0.1206635237,0.3470193148,0.0586238317,0.1539106667,0.2576222718,0.327263087,0.0221023895,0.205042243,0.0925239623,0.0125784669,0.1395575553,-0.2800306678,0.1792054921,0.2613037527,0.698786974,-0.3968589902,-0.2707764208,-0.1148355007,0.2155599147,-0.1440118998,0.1142741293,0.1241134405,-0.1997843534,0.0727660358,-0.1605892032,-0.0320135504,-0.213559404,0.111754939,0.0657198578,0.2325634211,-0.0038948597,0.2163617611,0.0000728886,-0.0169964246,-0.083187677,-0.1148610041,0.1384497434,0.1905744374,-0.115286611,-0.1532757878,-0.0530042537,-0.643604219,0.2569108903,0.0238949731,0.3082271218,-0.0463234857,-0.213043347,0.0331314951,0.1188271567,0.2266346365,0.1491998136,0.0412930436,0.1060218662,0.1200556606,-0.2173781842,0.0168233905,-0.0801324695,-0.1001487449,0.2647669911,-0.0554638281,0.5122625828,-0.160993576,-0.4319573939,0.2365232557,-0.6948952675,-0.1090141684,-0.1088430882,0.2770724893,-0.0793983415,-0.074644044,0.1808510125,0.0832690075,-0.2342760265,-0.2156969607,-0.1566640288,0.1546242535,-0.0371548496,-0.1352516413,0.1518603414,0.0900973231,0.4250291884,0.1070041656,0.1088750362,-0.5665574074,0.2359183729,0.1804818064,-0.3607868254,0.2312493026,0.1812458336,0.1465978622,-0.0617693551,0.0378138609,-0.1002752408,0.2124563903,-0.1672933847,-0.405683428,-0.0748658478,0.3433825076,0.0531347841,-0.269944638,-0.1625708789,-0.0508225374,0.0389541797,0.0309834313,0.0246787705,0.0224645138,-0.2562174201,-0.2442116439,0.1460748613,0.2647913396,-0.203980118,-0.0658719838,-0.1158687472,0.001087417,0.2173052281,0.2107508481,-0.2382163852,0.492053777,-0.1265227199,-0.0613782816,0.3794507682,-0.2946862876,-0.5540878773,0.1498631239,-0.3093815744,0.0878863856,0.222128287,-0.0451892279,0.1302660108,0.1655368507,0.1286420524,-0.1027142107,-0.1344432682,-0.143959254,-0.3133150041,-0.0950464308,-0.2379143536,0.0358379409,0.1849694997,-0.0782921761,0.1971244067,0.2329875529,0.3517868221,0.0966918841,-0.0691675842,0.2216302603,0.3059597909,-0.3300422132,0.1108259931,-0.2293465883,-0.3739764392,0.3155354261,-0.016556222,-0.1355928779,-0.030045066,-0.1432412714,-0.5316035151,0.1148184314,-0.0472935624,0.071796976,0.1636960208,-0.2225445509,0.119974032,0.1033539325,-0.1992002577,0.3342386186,-0.2297876775,0.1891273111,-0.0412655696,0.0069933338,-0.3651164174,0.0663702562,0.180712536,0.1554024071,0.1503354311,-0.1990832835,-0.1807877421,0.4038277566,-0.1873677075,0.1631094962,-0.1956853271,-0.0584174134,0.2430949211,-0.2473965138,-0.0698491633,-0.045346383,0.1733225137,0.0142511763,-0.0770471171,0.3409189582,0.1212306321,-0.2174375802,0.015748309,0.204194501,0.3598785698,-0.2917144597,0.2594666779,-0.2255247384,0.4635404646,-0.2363403738,0.1157255098,0.0373713225,-0.1989536136,0.2993932068,0.0630537346,0.0063302945,0.1100184321,-0.0354825445,0.0314819813,0.0334412307,-0.0733481124,0.4630728662,0.242342338,0.2329953015,0.0839797929,0.1448700577,-0.2224033475,-0.1442315876,0.2096170336,0.0638770387,0.1202782989,0.3854889274,0.0990730524,0.1486282051,-0.2791571617,-0.2268599272,0.4535354674,0.354529053,-0.0649734363,-0.0871002078,-0.4004331231,-0.2362252027,-0.2478104979,0.0471523367,-0.0626159459,-0.0893814266,0.0793108866,0.3766818941,0.0722202733,0.0774605349,-0.00535336,0.008093534,-0.0871411115,0.0769116729,-0.0113643892,-0.0330483764,-0.2678029537,0.1273901761,0.2779996395,-0.1478267014,0.6885922551,0.0321407653,-0.2163366228,-0.5223962665,-0.2491724938,0.0373586193,0.0187591072,-0.2935934663,0.0439141952,-0.0209343005,-0.136175707,-0.2722957134,0.019805925,-0.2420298159,-0.4843810201,0.0103911264,-0.1160224304,-0.0720853731,-0.2030246556,-0.5646247268,-0.2131234109,-0.2400033623,0.1935483664,0.046699401,0.3808447421,0.2424082458,-0.0673825145,0.1859569699,0.4195982814,0.153919369,-0.1942103952,-0.2394597232,0.1029656082,-0.2137128562,-0.2867950201,0.0474882424,-0.1526742429,0.3554935753,0.0046595586,-0.608931601,-0.1365873367,-0.2141961008,0.0243428126,-0.1862188131,-0.1313263923,0.3098138869,0.0154808182,-0.1365218312,-0.21909073,-0.0629931465,0.2879419029,0.1635590792,0.3820870817,-0.1381602287,0.5215933323,0.2062052786,0.5446686745,0.2092474699,0.0309621841,0.2792491317,-0.0046169646,0.0192993097,-0.0550986156,-0.2558485568,0.0054706396,-0.0831019804,0.0095832767,0.0092906831,-0.0893405303,-0.0721249729,-0.0446569435,0.1175524592,-0.3608146012,-0.3377099037,-0.10467197,-0.4469050765,-0.0484972,0.2157610059,0.0611393824,-0.213088721,-0.0070497207,0.0934162065,0.0202060565,-0.0688990131,-0.0649515688,-0.1495747417,-0.0190238021,-0.8308066726,0.2251755744,0.0450994484,0.288002789,-0.124375321,0.3266197741,0.3935656548,-0.1652689427,0.3911049068,-0.1676344275,0.2698667943,0.0979369879,-0.0790567547,-0.1465240717,-0.1656599343,-0.251033932,0.080031313,0.4006592035,0.3576405346,-0.3497659266,0.1794050485,0.348642081,-0.0631887615,-0.1417799741,-0.2178062201,0.0823130682,-0.4743447006,-0.4810009897,-0.0521650687,0.0439160615,0.353184104,0.0911344066,-0.030413419,-0.0107764928,-0.0366414748,0.1665459275,-0.0943752751,0.1184210852,0.160119921,0.6448461413,0.0918150693,-0.0352791287,0.5881874561,0.6839149594,0.0093940077,-0.5672091842,-0.1577608138,-0.1065871343,-0.1010121331,0.1782255918,0.0908020511,0.0933463126,-0.2088735551,-0.0646629855,0.2587456107,0.2643452287,0.3142026961,0.0723907053,-0.3763007522,-0.2506835163,0.3191961646,-0.0817579851,-0.1552018374,0.5849033594,-0.0678470656,-0.1047824398,0.472150892,-0.0114796655,0.7160589695,-0.3985783458,-0.0927845389,0.0731137022,-0.4017662704,0.6066354513,-0.2347756177,0.1309986413,-0.2987135351,-0.2008704245,0.0399288274,-0.0824519321,0.0222162195,0.0752064139,-0.1967361867,-0.0872491598,0.1580341756,0.0872746408,0.3206403255,0.3674143553,-0.0319647044,-0.0848511159,0.1150053665,0.1373304129,-0.2628635466,0.406024158,-0.0181210414,-0.0668047667,0.1028445289,-0.2992663682,-0.4831861854,0.217360571,-0.1947444975,0.318595767,-0.0621493571,-0.4355201125,0.1259116977,0.4583157003,0.3842694163,0.0739847794,-0.1031519845,0.2048924118,0.2485313416,-0.1178389564,-0.1350045651,-0.1210527197,0.1363639832,-0.1844772547,-0.1647848189,-0.0278721936,0.1377196908,-0.3701806068,0.0673819855,0.0858002454,0.1778977513,0.1031583995,0.101272203,0.1136688441,-0.2541298568,-0.4362241626,0.1229943186,-0.0469448827,-0.2431122959,0.157725364,-0.0330424272,-0.4383397102,0.0451806374,0.4698195755,0.3405234814,-0.0250849947,0.5422228575,-0.1824437827,0.0245187152,-0.2797749341,0.1596972942,0.0143834688,-0.1909028888,0.2175259441,-0.2706026137,0.1917577982,-0.0280452892,0.2737777531,0.2958149016,0.164579168,0.1127592698,-0.4255634248,-0.2509507239,0.0301229004,-0.0641630664,0.134132579,0.0876581445,0.4184315503,-0.0506983399,0.2553928196,-0.3356302381,0.1482075155,-0.1318371445,-0.0594203584,0.1746113151,-0.1717115194,0.0425734259,0.0718937814,0.2041368335,0.2437286526,-0.1099786162,-0.2864663899,-0.2727619112,0.1219254956,-0.0179657936,-0.0088633206,-0.0567950271,-0.2603075504,-0.1260796338,0.1398872882,-0.0748911649,0.4401758611,0.0784294307,-0.1644559354,0.158275485,0.2488705218,-0.2460633665,-0.0389491171,-0.3369632661,0.2750461996,0.1759338826,0.2675978243,0.0292404797,-0.1288139224,-0.2296471894,-0.2728818655,-0.2131318748,-0.146495536,0.2435556054,-0.3018551469,-0.0421684794,0.1519550681,0.2460226268,0.1520501524,-0.2486083061,-0.1632194668,0.1644352525,0.2459692359,-0.3797386289,-0.2860787511,0.2032912821,0.2655281723,0.1033863276,0.3162933886,0.1810362935,-0.127657041,-0.0802385136,-0.0684387758,0.2473837286,0.0403777398,0.0515414365,0.2337288558,-0.0613504536,-0.1316002011,0.2003403157,0.1655679494,0.2464240938,0.2796703577,-0.3739533424,0.1008228883,0.1590705514,0.2548454404,0.1725819707,-0.2700347006,0.0207830369,0.0539291464,-0.2069547176,0.202298969,0.1390527487,0.0993251055,-0.2250609696,-0.0641918778,-0.1498130113,0.1119479835,-0.1164995134,0.0984501541,-0.4847101271,-0.146549955,-0.2188888937,0.1769062132,0.0303366389,-0.1044943929,0.2071796209,0.0987809524,-0.3320992887,-0.4219256043,-0.28494519,0.193777293,0.2210933417,-0.4458440244,0.2861191332,0.1453136057,-0.2279456556,-0.1188646629,0.4128701985,0.2882546186,0.1377207637,-0.0044575576,0.1866989732,0.24326244,0.0614095032,-0.0662980005,0.1467690021,0.3831850588,0.1861084104,0.2336901277,0.0937460959,-0.1420445442,-0.0458601788,0.4072281122,-0.0515641458,0.0260140598,0.2018064857,-0.1265825033,-0.1134503335,-0.0411497653,0.1652198434,-0.2065222263,-0.3550455868,0.0897853971,0.0607543886,0.1378697753,0.0109456433,0.1126450598,-0.1217449084,0.1125570387,0.3411011398,0.2132985294,-0.4258606434,-0.1614370495,-0.6133244634,0.3812159598,-0.3473398685,-0.002553537,0.0322078802,0.2755564749,-0.0445154123,0.0474413894,0.3256109059,0.0206126701,-0.136232689,0.5303824544,-0.3904486001,-0.0794870034,0.1565261483,-0.2918981314,-0.0134287896,-0.2815947235,0.2825532556,-0.272033602,0.0293646194,0.0055840765,-0.0935334414,0.2100780308,-0.07036715,0.3803651929,0.0622596331,0.519521296,-0.0308382399,-0.1754388213,0.0944977477,-0.0112736812,-0.3946163654,0.1818333417,0.3479684591,0.2045311779,-0.0917685628,-0.0474527888,0.0297556352,0.4856366217,0.2023918331,-0.152092725,-0.30988729,0.061570961,-0.225493744,0.1006584615,0.0082628708,0.3684367239,0.0600798763,0.1921826601,-0.3377582729,-0.4812725782,0.5101733208,-0.2088739872,-0.093479909,-0.1610665619,0.375095427,0.0586280078,-0.0781517699,-0.4696164727,-0.1114213467,0.3627079427,-0.4114665389,-0.2139761895,-0.0388944149,-0.0612025857,0.0370035619,-0.061633978,0.5893378258,-0.0260268617,-0.2274932563,0.2875639796,-0.221898526]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"Hi ! Do you have an example in mind that shows how this could be useful ?","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":17,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nHi ! Do you have an example in mind that shows how this could be useful ?","embeddings":[-0.4188199937,-0.1118550226,-0.0211097654,0.164978072,0.4431482255,-0.0737165511,0.5200480819,0.025510639,-0.0000917429,-0.0612897165,0.1383574754,0.3554057777,-0.3664284945,0.1816854477,0.1437233835,-0.126732409,-0.0193027947,0.3366385698,-0.381377995,-0.1455226839,-0.181020543,-0.3653594255,-0.1667024642,0.020922713,-0.18708767,-0.0778727084,-0.0324452557,-0.4812172055,-0.0334892087,-0.0890976936,0.7275010943,0.049632404,-0.0156269856,0.6457778215,-0.0001180408,-0.1917837858,0.0963871926,-0.0768376365,0.0740181282,-0.3926313519,-0.470248878,-0.0931003764,0.0384803675,-0.2704013586,-0.4740886092,-0.0527881496,-0.0390679054,-0.0544612743,0.4018721581,0.1389983594,0.1720724106,-0.4391120672,-0.1384260803,-0.1219020709,-0.0793528929,0.4444640279,-0.2869962156,-0.1327419877,0.1904753894,0.3042643964,0.048442889,-0.0256284904,-0.0494295768,-0.1503827423,0.7289847732,-0.137820065,0.1042367965,-0.4418712556,-0.1786393821,0.4444399178,0.7537171841,-0.3168216944,-0.1675662994,-0.1634083986,0.2200718522,0.2761265039,-0.0231509525,0.2009141445,-0.0280606877,0.2865293026,-0.2226814628,-0.2574950457,-0.422495991,0.0322743244,-0.5201644301,0.1166955978,-0.068190895,0.0087468084,0.0091331927,-0.0526339896,0.2658056319,0.0693863928,0.1179423183,0.1734823734,-0.139836356,-0.5531538129,-0.3461993039,0.2137343585,-0.1130962968,-0.0050097806,-0.1444609761,-0.1590377986,0.1009187475,-0.0037088529,-0.045353055,0.2560715079,0.0444240943,0.2834724188,0.1548118144,-0.3329988122,-0.1865014434,0.1489244401,0.0044968962,-0.1453438699,0.1278155446,0.0848207772,0.0767351985,-0.1519827396,0.059555009,-0.2803952992,-0.0193673726,0.0458842702,-0.1318854988,0.0430733673,0.1812555194,0.0792005211,0.176816076,0.2458331138,0.2344777137,-0.2285812348,0.0617158823,0.2511284649,-0.297173351,-0.1399851143,0.0375425294,0.1639120281,-0.267642051,0.0062306882,-0.1234986782,0.3804902434,-0.0001207096,-0.26225245,0.1476206332,0.2959809303,-0.5519163609,-0.2185193449,0.1582685709,-0.1785986871,-0.2252624333,0.4631130099,-0.1985176504,0.0231161043,0.0901266113,0.1778133661,-0.0336978287,0.0682865679,-0.1685471088,0.3371283114,0.0445051827,-0.1100770682,0.1710585803,-0.3961261213,-0.1047371104,-0.1422396898,-0.0999873653,0.0916873813,-0.6140998602,-0.3623417616,-0.1760552973,-0.0748703852,0.1173579171,0.0689149648,-0.2285491526,0.3040558696,-0.1412665993,-0.1797803938,0.663918972,0.0727627352,-0.2718644738,0.2040672451,-0.5375030637,-0.1342455596,0.1489097625,0.3913097084,0.2420570403,-0.2189547867,-0.5890008211,0.2351159304,-0.0481770113,-0.1363944411,0.0984599069,-0.1721860915,0.5869368315,0.0207448918,0.2949224412,0.4651507437,0.182667166,0.5072157383,-0.1435580254,-0.3711574376,0.2393033206,-0.0810984671,0.4471267462,-0.045060996,0.0037965905,-0.3930869401,-0.0073088813,-0.1223513484,-0.0850009844,0.1331424862,0.3702483177,-0.090662472,0.0221196674,-0.3194077909,-0.1461212337,-0.1193116456,0.1447146386,-0.3999894261,0.280167073,-0.2651736438,-0.1051316038,0.4493666887,0.0060810461,0.0600682721,-0.4010302722,0.2931924462,0.0819307566,0.2227484286,-0.3591738641,-0.0449864492,-0.144190222,-0.1969311386,-0.1015450135,-0.1836638302,-0.0584041364,-0.0512780547,0.0085194856,0.439289093,0.2763770223,-0.4933925867,0.4896831512,0.0793199912,0.0957188308,-0.1939777583,-0.3488175869,0.4720558524,-0.150127992,0.2503143549,-0.3207320869,0.0362431817,0.186323598,-0.433570087,-0.1252280474,-0.0907754898,-0.0459950119,0.0280870683,0.1391773075,0.2028847486,0.0194785055,-0.1024754122,-0.0012694855,-0.0949835256,0.2904396653,0.4638142586,-0.235706374,0.2653409839,-0.2432362437,0.5296252966,0.6370874643,0.1556210518,-0.2533013523,0.2274798006,0.0958712026,-0.1503843367,0.2806729972,-0.0943110287,0.2716952562,0.0350976288,-0.1571458578,0.1369215101,0.070758529,-0.3582987487,-0.0043882094,0.1940810084,-0.1077887714,0.0459180549,-0.1841415167,-0.0973766968,-0.3303296268,0.0659516603,0.072096929,0.1469988227,-0.0945241451,-0.0728322789,0.281208396,0.0661741793,-0.0833279192,0.2607306242,0.236743018,-0.057936959,-0.1471119672,-0.070167996,0.0777691901,0.1580581963,-0.1741780341,0.0169900544,0.2542794049,-0.3422031999,0.1313803941,-0.0990186483,-0.4404471219,-0.0119380821,0.0566025935,0.16734083,0.3422213495,0.0247850493,-0.0896843076,0.0493691154,-0.0207102075,0.0460866168,-0.1860033423,-0.0157513116,-0.2754972279,0.2332953811,0.0840452462,-0.2450775504,-0.2703368366,-0.1052667275,0.1708030254,0.0575782806,-0.0584610067,0.0116345808,0.0825020745,-0.019699581,0.0160593316,0.1571203321,-0.0940387622,0.0360755548,0.2773386836,0.0892527401,-0.1180108115,-0.1655022353,-0.2728357911,-0.0858478546,0.2958499789,-0.3110006452,0.0010259798,-0.1123813093,0.0221872199,-0.0838600472,-0.145645529,0.5014488101,0.050975889,0.0237529818,-0.0218031537,0.0657578409,-0.0743142292,0.1865724623,0.3480320275,0.1766137034,0.4494115412,-0.0166078489,0.2254580557,-0.3424141705,-0.4381712079,-0.26404351,0.0538090803,-0.0042795348,0.1079061702,-0.4174182713,0.2436749786,0.2311609834,-0.3049213886,-0.0121197086,-0.0834345594,-0.2215757072,-0.1662246585,-0.2121593952,0.0478753559,-0.0702751055,0.0171243679,-0.2095749378,-0.0629358962,-0.1967492551,-0.0494220816,0.1260919273,-0.1038436368,0.0292486697,0.5218977928,0.110350661,-0.0452300683,-0.024213545,0.024874121,0.2543198168,-0.0987888202,-0.2955639064,0.0200693775,-0.0834717378,0.05685151,-0.0046996833,0.1922798306,0.7796655297,-0.0551341027,0.1961565167,0.1951418519,-0.3799063265,-0.2024967074,0.0559910163,-0.2858710587,-0.176736176,0.0113886232,0.0822414905,-0.0798510984,-0.1225580871,0.1050653905,0.2761817276,-0.2947768569,-0.3678512275,0.1842408925,0.0087060379,-0.2560683489,0.2149040699,0.0323032513,-0.1509324461,-0.1219079718,-0.0597335212,-0.3734533489,-0.0020950274,0.3103741407,-0.1419566572,-0.1717830896,-0.0993547887,0.2183592021,0.0002128015,0.2302410603,0.096457921,-0.0325966664,0.2010596991,-0.5295643806,0.0463531427,0.0103048673,0.2909169197,-0.2127972692,0.0490581803,-0.0731630772,-0.0993527174,0.2489530742,-0.1713672727,0.0567143448,-0.0679485649,0.0433362685,-0.2862598002,-0.3991357982,0.3170524538,0.1924042553,-0.1370120645,-0.0299176592,-0.0295858625,-0.5042580962,0.1977588087,0.559302032,0.9863254428,0.1159872487,0.1385222822,-0.1127607822,-0.0312075187,0.4583590031,-0.1164896786,-0.0582190119,-0.2965381742,-0.392632395,-0.2365939319,-0.3042209744,0.3992506862,0.5360763073,-0.5129191279,0.2585294545,-0.371717602,-0.0838064775,0.2241666317,-0.1048000306,0.3480706811,-0.3577433527,0.0461030751,0.167915687,-0.0882538781,0.2601305544,0.0308575202,-0.1675019711,-0.0712006614,0.2808563709,-0.0880775005,-0.0689706951,-0.2679661214,-0.1455813944,-0.2007422149,-0.4343781173,0.8832834363,0.3755268753,0.2161356658,0.225745663,-0.1638488919,0.0465083234,-0.0744043067,0.3089933693,0.0002890406,-0.1842873842,0.4555921853,-0.04589637,-0.0858468637,0.0682666227,0.0122593315,-0.1186930016,0.2134862989,-0.1756024659,0.119954288,-0.3734203875,-0.0196654815,0.0330975316,0.0182802081,-0.0086924257,0.1461735219,0.321288377,-0.035990648,0.4378370345,0.359959811,-0.2925547063,-0.0926596597,0.1525407135,0.5887404084,0.0593939833,0.1273666173,-0.3841595948,-0.3116016984,-0.3853751421,0.2694804072,-0.0588144064,-0.2542477846,0.2276105881,0.0699306056,0.0229327809,-0.0068976562,-0.3377266526,0.259275794,0.2940540314,-0.3691020906,-0.4159674942,-0.3563139737,-0.2046509981,0.0908985287,0.0709826574,-0.0747282356,0.1950056106,0.3595223427,0.0223155282,-0.3131049275,-0.0536869019,0.0048860856,0.1169308573,0.1778047085,0.1654969007,-0.3920322955,0.1679166257,0.0183160082,-0.0933931693,-0.1448343247,-0.1975415647,-0.1709586084,0.1506626308,0.1506834179,0.2185853422,-0.1473584175,0.0401099771,-0.0144191757,0.1092646196,-0.0177837498,0.2290980965,-0.0577833503,0.3751970828,0.4489991963,-0.2798541784,-0.4648520648,0.0532281548,0.0720136762,-0.0067777755,-0.1587938517,0.2750158608,0.0991029292,0.0945451036,-0.5071005225,0.0320366733,0.0995945111,-0.0615903102,0.4186533689,-0.0078993058,0.4520020485,0.5059018731,-0.04416354,0.0501449183,-0.1926655322,0.0139052393,0.2509203851,0.1545980126,-0.3792205155,0.1363568902,0.2737792134,0.0160392839,-0.0480938703,0.295191437,-0.0435867496,0.0883008242,-0.0707060248,-0.1038784757,0.4184021056,-0.360319376,0.2763784528,0.6955447197,-0.2741846144,-0.0000256548,0.587677598,0.0294153783,-0.0534790456,0.3487199843,0.1648121029,0.3593232632,0.114098914,-0.0763519928,0.2673564553,0.2705789804,0.1185700744,0.6662791967,-0.0775702447,-0.1824604869,0.1305554211,-0.1851775646,-0.0366713628,-0.2404727638,-0.0460646823,0.1559823304,-0.2395356745,0.3208657503,-0.201394558,-0.056617938,0.1262229532,-0.3402042985,-0.1603171527,-0.2548778653,0.3563571274,0.1266769171,0.0537597537,0.2256072313,0.3504413664,0.0669029355,0.087179102,-0.0889267847,-0.2596403062,0.1178576797,0.1217088699,0.1699084938,0.518145144,0.0359262303,0.1865809113,-0.04960455,0.0585876107,-0.1713820696,-0.2473457456,0.0949910656,0.2725590169,-0.2967213392,-0.1580559909,0.1844722331,0.203292951,-0.1213301197,0.4222786725,0.036692474,0.2091029733,-0.25141114,0.3976300359,0.306756556,-0.1823133081,-0.1009439602,0.1998399049,0.0393758118,-0.0032949816,0.4314260483,-0.4219202697,-0.0242354125,-0.0415411331,0.0387021042,0.2140872627,0.1312822253,0.018133387,0.0149853248,-0.0151186492,0.0497834645,-0.3194951415,0.2027306855,0.1657825559,0.0185094532,-0.1336497813,-0.2075477242,0.0096982699,0.2501333654,0.0554258451,-0.1713771522,-0.1410386413,0.2018368542,-0.0365415439,-0.0278339647,0.1539267749,0.4283744097,0.0223403219,-0.224701941,0.0716225803,0.2997146547,0.0264116917,-0.0329239182,0.3513908982,0.038543161,0.297788322,0.2075157017,0.0420178734,0.6284477711,-0.0537592433,0.3799882531,0.5167906284,-0.0332602002,-0.2373844534,0.2958490849,0.2015533894,-0.2141609639,-0.2449320108,-0.1070013419,0.1380840242,-0.2363910079,0.0062238183,0.0604804233,-0.8246730566,0.0415502228,-0.1709617376,-0.111310266,0.1197766364,0.0274879187,-0.0676506907,-0.0671234727,-0.0524829365,-0.129825443,0.2856090963,0.0781066865,-0.3340091705,-0.1724069118,-0.0688871518,-0.33146438,0.1222174317,-0.0650871173,-0.0006406449,0.4729189277,-0.0696832091,0.3831120431,0.0582936965,-0.0738810375,-0.0992481038,-0.2302218974,0.6369761825,-0.2621823251,0.0422271118,-0.2269104719,-0.0962075368]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"Example:\n\nMerge 2 datasets into one datasets\n\nLabel extraction from dataset\n\ndataset(text, label)\n \u2014> dataset(text, newlabel)\n\nTextCleaning.\n\n\nFor image dataset, \nTransformation are easier (ie linear algebra).\n\n\n\n\n\n\n> On Jul 6, 2021, at 17:39, Quentin Lhoest ***@***.***> wrote:\n> \n> \ufeff\n> Hi ! Do you have an example in mind that shows how this could be useful ?\n> \n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":83,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nExample:\n\nMerge 2 datasets into one datasets\n\nLabel extraction from dataset\n\ndataset(text, label)\n \u2014> dataset(text, newlabel)\n\nTextCleaning.\n\n\nFor image dataset, \nTransformation are easier (ie linear algebra).\n\n\n\n\n\n\n> On Jul 6, 2021, at 17:39, Quentin Lhoest ***@***.***> wrote:\n> \n> \ufeff\n> Hi ! Do you have an example in mind that shows how this could be useful ?\n> \n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n","embeddings":[-0.4854312241,0.1529312581,-0.0886402205,0.1896278262,0.4006462991,0.1068503782,0.507528007,0.1244744882,-0.0116836671,-0.0620446242,0.0533830151,0.3345322311,-0.3507431746,0.1875819862,0.0721608475,-0.1314892918,-0.0283112172,0.3180100322,-0.4435121417,-0.1215898991,-0.2172255516,-0.2452004105,-0.204332307,-0.0350172669,-0.2736256123,-0.1387131512,-0.1118711159,-0.2744330466,-0.1285729706,0.0032409544,0.5669198632,0.2155507505,-0.0712961778,0.5656960607,-0.0001113632,-0.1279024333,0.1100607142,-0.0080487533,-0.1362921596,-0.3996142447,-0.416369915,-0.0603963397,-0.0396468043,-0.2917911112,-0.2664454281,-0.1854079366,0.0237760376,-0.0645149946,0.4134106338,0.0751382485,0.2136539668,-0.4207115173,-0.081319578,-0.1326424181,-0.0725329444,0.4323451221,-0.1573563963,-0.0739583224,0.2559334636,0.2774946094,-0.005224322,0.2231015116,-0.1098077819,-0.2118475437,0.659743607,-0.086872831,-0.0518905669,-0.3680224121,-0.1720120013,0.3020718396,0.6202872396,-0.4165475667,-0.2718209624,-0.2353081554,0.3179525137,0.1380975544,-0.0936644822,0.1536032856,-0.0291529447,0.2862283587,-0.3013944328,-0.3070755005,-0.312664777,0.0769137442,-0.3927911818,0.1838325262,-0.0187025126,0.0787064508,0.0261065941,-0.0120617105,0.2786219716,0.0286741741,0.1734904945,0.1302508563,-0.1575931311,-0.4430968165,-0.347088933,-0.01647182,-0.1724005491,0.082052581,-0.0669375435,-0.1407574117,0.0001506267,0.0678739622,-0.0976173431,0.2801892459,0.0756248534,0.2835099399,0.0927253142,-0.2981130481,-0.1342790425,0.2335567474,-0.1500683874,-0.1828356385,0.0823884085,0.0751636773,0.0881752074,0.0607573465,0.0079319952,-0.3705034256,0.0102080423,0.0104207722,-0.1592562497,0.0114267236,0.2135364115,0.0537415035,0.2268779129,0.1977373362,0.2781538367,-0.2910178304,0.0033119151,0.3380631804,-0.3129458427,-0.1553813368,0.0573053025,0.0989251956,-0.3175148368,-0.0068458575,-0.2444904894,0.386097461,-0.0104768593,-0.2480302602,0.2561534345,0.1999243349,-0.4474495053,-0.1308980584,0.1515335888,-0.2037787586,-0.208618328,0.3421449959,-0.1957412213,0.1629634798,0.0655770227,0.2365572304,0.1148019582,0.0316513292,-0.298053503,0.3904443979,0.0113533875,-0.0630758777,0.102777414,-0.315286696,-0.063845478,-0.0778256282,-0.0714211613,0.1344807297,-0.6597012281,-0.3752602935,-0.0892390683,-0.0400439277,0.0257388875,0.1785104573,-0.1679376811,0.3726211786,-0.1166770533,-0.204284206,0.7212469578,-0.0659181923,-0.1949130446,0.1581442952,-0.5203281045,-0.1914043725,0.1640527993,0.3704347014,0.2596188784,-0.1747305542,-0.4552618265,0.3597725034,-0.1351937354,-0.1334810257,0.1524548978,-0.2831425667,0.4489609897,0.1342541575,0.2727267444,0.2775531411,0.0847700387,0.4158124328,-0.1401902288,-0.4404054284,0.284372896,0.0401456915,0.3044769764,-0.1269488037,-0.0328502059,-0.3859139979,-0.1105206832,-0.126223743,0.0478566401,0.1563088447,0.2692359388,-0.0514915511,-0.0509790517,-0.2854634821,-0.1628869027,-0.2248650044,0.2087797225,-0.3400037587,0.2504788339,-0.3147006035,-0.074569881,0.5503131747,0.0370456353,0.2137278318,-0.5016191602,0.3224230111,0.0449388511,0.2022389323,-0.3285857737,0.0471547246,-0.2616764307,-0.2003247589,0.0024983638,-0.1345227659,-0.0960846022,-0.0462613627,0.0916247442,0.3488053083,0.3109102845,-0.4374358356,0.3571309745,-0.083256118,0.196588859,-0.1748176813,-0.4034092724,0.3144291341,-0.2289667577,0.2899323106,-0.1774757355,0.0508714616,0.1357771158,-0.3550786376,-0.1169614047,-0.1495451182,-0.0605595149,-0.0075774002,0.2507326901,0.2566907704,0.0911047757,-0.0859348252,0.1485472023,-0.0672260299,0.2210759372,0.3868333101,-0.2876425087,0.2263994813,-0.2082078308,0.5587705374,0.4472160637,0.2356890738,-0.2560152709,0.2304827571,0.1021553427,-0.073639825,0.2621409893,0.0267599151,0.4427095056,-0.0132771386,-0.0441405512,0.2116562277,0.0442383848,-0.4906044304,0.0805869475,0.1066009551,-0.0207087882,0.0048884447,-0.1806955338,-0.1815158874,-0.4042205215,0.1540659219,0.0376808681,-0.0611128286,-0.0880035609,-0.0729024857,0.1077074334,0.1231755763,0.0667177662,0.2231042236,0.2378362566,-0.1540003717,-0.0582209639,-0.1108326763,0.057614509,0.1907510757,-0.0666936114,0.0429162681,0.3864461482,-0.3671724498,0.2392293364,-0.1411261559,-0.5004366636,0.1360855401,-0.0246935803,0.0368475728,0.3780713975,-0.0869703665,-0.101020515,-0.0403236784,0.0226899162,0.0577006899,-0.2035497576,0.0045408439,-0.2300645858,0.0799108595,0.0435018539,-0.2263130248,-0.3054023981,-0.2169413716,0.2214498371,-0.0128984582,-0.0439193919,0.0234342162,0.1507755816,0.0317601971,-0.0810158178,0.1676726788,-0.0809066594,0.0390486792,0.1736202836,0.12728104,-0.0919616893,-0.2783718407,-0.3630923927,0.0234737117,0.31731686,-0.2352731079,0.01467063,-0.1455161721,0.0455431454,-0.00765173,-0.0787579641,0.5167153478,0.135990575,-0.0250538439,-0.0448868461,0.0040792534,-0.0832194835,0.0459284186,0.1290344298,0.0709372386,0.3557409942,0.0434626602,0.2524471581,-0.1993198246,-0.3729278743,-0.2999459207,0.1988696754,0.099833928,0.0715788305,-0.4846233428,0.2665576041,0.1256868094,-0.2803901434,-0.0643635392,-0.072648257,-0.1584719718,-0.1652666181,-0.0832464546,-0.1282411218,-0.1157136336,0.1214825809,-0.1738656461,-0.0606970415,-0.1348017156,-0.1101042181,0.0693345591,0.0560910888,0.0124640083,0.4534653723,0.0439972654,-0.0185622647,-0.1429930776,0.053780891,0.2013120353,-0.0198407359,-0.164923057,-0.0192356072,-0.02333064,0.0138601288,-0.1604784578,0.1167049184,0.8743848801,-0.1053460687,0.0392143838,0.1754387319,-0.3215917647,-0.2480315417,0.034238033,-0.3715784252,-0.3055799305,0.0614507943,0.2085668594,-0.113812238,-0.1256839782,0.3286090791,0.3007803261,-0.1876237988,-0.2429513484,0.2264592052,-0.0608755834,-0.3976773024,0.0634020343,0.0986596197,0.0182806011,-0.1287332922,-0.0159141533,-0.3286944926,-0.0786561519,0.3200697005,-0.0887112245,-0.1668912768,-0.0883675888,0.1926545799,-0.0395863876,0.2257262319,0.1378199011,0.1193484366,0.0274984147,-0.6537035704,0.1715267152,0.0505665503,0.4162432849,-0.1999127269,0.0344758257,-0.1658141315,-0.0921526402,0.2252510786,-0.2954198718,0.1215904951,-0.0343767107,0.1489754915,-0.4204351306,-0.3190874755,0.2972900271,0.2758484781,-0.1264448762,0.0362372808,-0.1070564389,-0.4972878397,0.2006771117,0.4948278368,0.8818396926,-0.0173556823,0.236484617,-0.0485710762,-0.2004312873,0.5380058289,-0.0032374978,-0.0201327763,-0.2822466791,-0.3579156995,-0.1545642167,-0.2693386376,0.3293358982,0.5125810504,-0.5189641714,0.2395549119,-0.2172650397,-0.0057562217,0.1921662986,0.0804132149,0.2758338153,-0.5404213071,0.0213345923,0.1845383346,-0.0028104673,0.1811757535,-0.0103377998,-0.1743304431,-0.1444305032,0.2676253021,-0.1415925175,-0.0099307206,-0.2488065511,-0.0027998833,-0.1635055989,-0.3484613895,0.693431139,0.2555440664,0.2147729546,0.2217784673,-0.057856299,-0.0136012919,-0.164458558,0.3608772755,0.0370069146,-0.3177411258,0.5784918666,0.0348818302,0.0611876883,0.0403621495,0.0455642529,-0.0843531489,0.1005487889,-0.1519434452,0.1185914204,-0.3325740099,-0.025045719,0.1056226268,-0.0096455477,-0.112495333,0.1491444707,0.2361220866,0.0416291803,0.2985835671,0.290743649,-0.3393578827,-0.0290111322,0.1557558775,0.5326725245,-0.0060439534,0.2088619024,-0.2699059546,-0.2365808934,-0.4026525915,0.2671622038,0.1195406616,-0.3909738362,0.3062411845,-0.0393647887,-0.0612290278,0.04923179,-0.385186106,0.3085523844,0.2617302835,-0.2858154774,-0.3528090715,-0.3455169499,-0.0381411985,0.1107466519,0.1079318374,0.0000756988,0.3572805226,0.2443053275,0.0375875868,-0.3780190051,-0.0753517747,-0.0252063069,0.0798527747,0.163548857,0.0765361339,-0.3550307751,0.1056521609,0.082699202,-0.0714533329,-0.1629396379,-0.2778495252,-0.1733830273,0.1210464314,0.0908096507,0.2292822301,-0.0004708713,0.1254818588,-0.0314025357,-0.0468612872,0.0633701906,0.0709622279,-0.0459407009,0.2866384983,0.2412418574,-0.1901995391,-0.3151239753,0.073104836,0.035177283,-0.0556340963,-0.1957443953,0.3596335948,0.198205933,0.0916904137,-0.4791559875,0.0773005113,0.1645247638,-0.0897554681,0.3933241367,-0.0731325969,0.3916724026,0.4886705875,0.0866587162,-0.0632252917,-0.1646963358,0.0011307414,0.107295543,0.2124186605,-0.3923549652,0.1034425348,0.276742816,0.0379572809,0.0326937996,0.2100823969,-0.1345962137,0.2381770909,-0.116975747,-0.0781984106,0.479693681,-0.3370963335,0.3642072678,0.6463420391,-0.1479434371,0.0617973991,0.4965008497,0.0711717382,-0.0195746813,0.3492759764,0.0980383679,0.4678318799,0.005631783,-0.0061907321,0.4555330276,0.1775169522,0.1988160908,0.6114429235,-0.0600669645,-0.0282988828,0.0239461083,-0.143137604,-0.0567234233,-0.3246426582,0.0277534146,0.2988834083,-0.2211218327,0.3321267068,-0.1391834617,-0.0811675116,0.0764173865,-0.3467306197,-0.2239057422,-0.3499315083,0.3692118526,0.0660322011,-0.0369878486,0.0641421229,0.1676675826,-0.0217832252,0.0589334518,-0.1439836025,-0.1938712299,0.1554631591,0.091855146,0.1143858731,0.6183263063,0.1413576901,0.2268330157,-0.0266369004,0.0454731062,-0.2194780111,-0.2695859671,-0.0295798052,0.2105143219,-0.2484850287,-0.1852517873,0.2631670237,0.2606516182,-0.1493424624,0.2405787259,0.089551717,0.1632008255,-0.2439084351,0.257274121,0.1971981376,-0.1521504819,-0.0466407277,0.1972215772,-0.1380611211,-0.0838318765,0.5476492643,-0.4606724977,0.0192347206,0.0142844664,0.076179482,0.1886263937,0.2115135491,0.100326851,-0.0400603116,-0.0238133147,0.0371470824,-0.3171057999,0.0832650736,0.1469305605,-0.094707936,-0.1891988814,-0.1004274264,0.1086465642,0.3245387673,0.0671721175,-0.1424287856,-0.1395664662,0.130049631,-0.1766137779,-0.0718234777,0.3381154835,0.2639379799,0.0775396749,-0.3354123831,0.068443574,0.3268901408,0.0650978088,-0.0475229844,0.3619348407,0.1386049092,0.1587195843,0.2067413777,0.0901311561,0.5049049854,-0.0686620176,0.1637928784,0.3704541624,-0.0732799545,-0.2627440989,0.3999932408,0.2048811018,-0.0300378017,-0.2529410124,-0.1391843259,0.1666020751,-0.189974755,0.0365460888,0.0669791177,-0.8383834958,0.0600658581,-0.1159213409,-0.0091716023,0.1660302132,0.0594986193,0.0017780876,-0.0214657765,-0.0139739038,-0.3064361513,0.3038731813,0.0405731276,-0.3939807415,-0.1428613961,-0.163425535,-0.2189141214,0.1842772961,-0.0942626819,0.0096318396,0.4470911622,-0.1043238416,0.335305661,0.2366400957,0.0734264404,0.0055567152,-0.2274941057,0.5061415434,-0.3077508807,-0.1795010865,-0.2111504227,-0.1740984172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"There are already a few transformations that you can apply on a dataset using methods like `dataset.map()`.\r\nYou can find examples in the documentation here:\r\nhttps:\/\/huggingface.co\/docs\/datasets\/processing.html\r\n\r\nYou can merge two datasets with `concatenate_datasets()` or do label extraction with `dataset.map()` for example","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":41,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nThere are already a few transformations that you can apply on a dataset using methods like `dataset.map()`.\r\nYou can find examples in the documentation here:\r\nhttps:\/\/huggingface.co\/docs\/datasets\/processing.html\r\n\r\nYou can merge two datasets with `concatenate_datasets()` or do label extraction with `dataset.map()` for example","embeddings":[-0.4685243368,-0.2306800336,-0.0531573892,0.141038999,0.3790983558,-0.1987152547,0.336720854,0.1695584059,0.1045032665,0.193190679,-0.1746759713,0.2893973589,-0.2316427976,0.5593631268,0.0868366435,-0.2147053182,-0.1576156914,0.3244946003,-0.4510017633,-0.0933915228,-0.2045138329,-0.1517907381,0.0859746113,0.0435048975,-0.2052290738,-0.2564643025,-0.0503218733,-0.1808667779,-0.0929280743,-0.0206905995,0.3956409097,0.2275130004,-0.1231825128,0.461050868,-0.0000994468,-0.107468307,-0.1551022679,-0.0925966576,0.1021683365,-0.168191731,-0.125418961,-0.084633626,0.0220993273,-0.2678519785,-0.5002477169,-0.1091537103,-0.0318488441,0.0631250218,0.5358985066,-0.0103142783,0.2958766818,0.0097887106,-0.0158978384,-0.1169515923,-0.0069825319,0.2734376788,-0.1275932193,-0.018073108,0.1740113497,0.2556479275,-0.1945536286,0.3934059739,-0.0161714107,-0.032958664,0.6259969473,0.0143669872,0.0547637083,-0.4071558416,-0.2678075433,0.4698746502,0.3069075644,-0.2341915965,-0.4158831835,-0.4240956008,0.1417989135,-0.0436883718,0.0182049982,0.0818796903,-0.0191098955,0.2180835307,-0.3345590532,-0.2262868732,-0.2550739348,-0.084840931,-0.4569258094,0.1557009965,-0.1581203043,-0.0320846364,0.208306849,-0.1093471721,-0.057075128,0.2332167178,0.0770113915,0.1242182925,-0.1319326907,-0.4667616785,-0.240433991,0.2420442402,0.0641125143,0.0925941318,-0.101852648,0.0364717245,-0.107458137,0.1789062619,-0.0019420339,0.3197201788,-0.0181512497,0.0871051997,0.2844573259,-0.0984446034,-0.0780216083,0.1426822394,0.1567533314,-0.1871579438,-0.1085611209,-0.047767289,0.0716382936,-0.1214730814,0.0511640087,-0.3087018728,0.267242372,0.1720949113,-0.1121375188,0.2878980935,0.0153888362,0.2037271857,0.3953179419,0.179369539,0.1763352156,-0.1484648734,-0.0825694278,0.1802886128,-0.0494106524,-0.0196711849,-0.0038546352,-0.0378014594,-0.0293676201,0.0261388272,-0.0889280736,0.3310028315,-0.0821674317,-0.1042408645,0.2652435601,0.1671462208,-0.2829819322,-0.2588634789,0.0745518431,-0.1645919681,-0.2382676601,-0.0137226693,-0.1708738953,0.0797658116,0.0186045971,0.2774107158,-0.0035122889,-0.0864806622,-0.21165663,0.4036306143,-0.0128508657,-0.0990921631,0.0406693034,-0.173594445,-0.1943620592,-0.025116479,0.0886336863,-0.0599144511,-0.3292738497,-0.4942750335,0.1610184759,-0.1373925507,-0.0452827141,0.2936730683,-0.0389954075,0.1977722794,-0.062210463,-0.0802091509,0.4939524531,-0.1065551266,-0.1814965606,0.068139568,-0.3399479985,-0.0184119456,0.0083446512,0.3178723752,0.2948327959,-0.0164953284,-0.3483887315,0.3703303933,-0.0752102658,0.1497977972,0.1050369292,-0.2864674032,0.2458959967,0.062972419,0.2355774343,0.2537992299,-0.069267422,0.1117845848,-0.0753493756,-0.4083535373,0.2005532086,0.0748429671,0.2569252849,-0.0320868492,-0.1039824635,-0.4349060059,-0.1174092293,-0.1057571024,-0.1769857854,0.2305080295,0.2835945785,-0.2835356891,-0.0504569672,-0.2405830026,-0.1328191012,-0.1575145721,0.2584118545,-0.2040453404,0.2221158147,-0.3294777274,-0.1069264635,0.4259645343,0.1077912152,0.2475834191,-0.4918275774,0.0894110501,0.0432433337,0.0613956377,0.0378549881,0.0642796755,-0.15450266,-0.2567332387,0.0715664104,0.0709938034,-0.0862856209,0.0828549117,-0.0011877238,0.3227497935,0.2609703243,-0.0784021616,0.143908903,-0.0525297672,0.1788266301,0.0433894731,-0.4305520058,0.6129770875,0.0180025399,0.1722464263,-0.0484185964,0.0495318137,0.300876379,-0.3305992186,-0.014370014,-0.1949433982,-0.1480185986,-0.1301355362,0.2936966717,0.1075424477,-0.2553980947,-0.0235894732,0.2760587335,0.0384417437,0.1950599551,0.3294142187,-0.258631587,0.2395893186,-0.0654359311,0.3712917864,0.2040140629,0.3569667935,-0.1667204201,0.239026159,0.0320154838,-0.0046142754,0.1888573617,-0.0338227116,0.2357216775,0.0783822462,0.0489155911,0.1945629418,-0.2278055698,-0.2059036046,-0.2645703852,0.0031590862,-0.2033651769,0.0034469608,-0.1967256665,-0.1450364143,-0.2393192053,0.0100171873,-0.3171900213,-0.078466624,0.1861803234,0.1125213653,-0.2316172123,0.1188410148,0.2403347492,0.2831951678,0.198699519,-0.0203866176,-0.127515614,-0.1675501466,0.1066242605,0.2194830179,-0.1370417625,0.0977809951,0.4182054698,-0.411868751,0.1238672361,-0.2679352164,-0.4999006391,0.2685489058,-0.1570931524,0.2256887406,0.3385981023,0.1490938365,0.0294232275,-0.0188311692,0.0859243199,-0.1540537477,-0.2891170382,-0.0687830076,-0.2152634859,0.0695159882,-0.1339472681,-0.1737843752,-0.2338173985,-0.35813272,0.3657938838,0.1533703953,0.0451712608,0.0105709638,0.0082780421,-0.0310073849,-0.215448305,0.3004665077,-0.2189868242,-0.178074494,0.2259613872,-0.0719293579,-0.1818494499,-0.2360566854,-0.1790878922,0.0673804134,0.2620018423,-0.1796708554,-0.2952071726,-0.2022519112,0.2044983953,0.0763566718,0.1147262529,0.5687099099,0.1093285531,-0.1792212576,-0.0555118062,-0.0229510516,0.1011578962,0.1597188413,0.0593476593,-0.0260262657,0.368496269,-0.038997788,0.3227413297,-0.3008691072,-0.4758778512,-0.0619685836,-0.0589893572,0.2156079859,0.0416006148,-0.2979401648,0.2278830409,0.04417824,0.0191546455,0.1481875926,-0.0486993492,0.0111457398,-0.2310934514,-0.1589175612,-0.2245370895,-0.3170900047,0.0428312905,0.2425029427,-0.0501595661,-0.0710791945,0.0184232537,-0.0754570961,0.0562100857,0.1994461417,0.4151166975,0.084750317,-0.0235048197,-0.2109864652,0.1167980507,-0.0242337603,0.2419670522,-0.06833224,-0.2246361673,0.0057178801,-0.2587054968,-0.1484929919,-0.135675624,0.6416298747,0.0848639682,0.0767537132,-0.0179064628,-0.4646278322,-0.2970891893,0.2067296356,-0.3552080095,-0.2220585346,0.1504126191,0.3458377123,-0.3562289774,-0.2923684716,0.1644766629,0.3387872577,-0.3072396815,-0.0829121545,-0.0008131768,-0.126312539,-0.3773184121,0.2055283338,0.0385525711,-0.0771327317,0.0005985181,-0.0592508316,-0.3329601884,-0.1186062917,0.2515352368,0.3139459789,0.0628118366,-0.1446579546,0.0171493366,0.1051334813,0.4406676888,0.0418914668,0.2494480908,0.0235106535,-0.570209384,-0.0133866146,0.0422749817,0.1509062201,-0.0021074251,0.1246787757,-0.0105003333,-0.0035681585,0.3054661751,-0.4406086206,0.2906212211,-0.0310881846,0.1782278121,-0.2548517585,-0.104605265,0.4059157372,0.1876244545,-0.2115992606,-0.000697774,0.2190535665,-0.4485887885,0.2102264315,0.3379157782,0.820833683,-0.02207244,0.1324445009,0.2220871747,-0.0561945997,0.4963654876,-0.1502448767,0.0325308591,-0.2691507339,-0.3934499919,-0.1357685626,-0.0652889982,0.2102765888,0.2437312603,-0.3813426793,0.3641525805,-0.0496974662,-0.1586469859,-0.0765664279,0.032895539,0.148718521,-0.5673125386,-0.4194623828,0.25726524,-0.1618788093,0.284717083,-0.0035603181,-0.1318379939,-0.0876574442,0.0457287878,0.09723261,-0.0441240929,-0.325653553,-0.2421275824,0.0073955683,-0.2083853632,0.3350386918,0.2558939159,0.3095935881,0.1981585771,-0.1272188872,0.1192677766,-0.2853118479,0.3346187174,-0.0915669054,-0.1033792421,0.5256097913,-0.030810643,0.0321184471,-0.1150328889,0.0819013938,-0.1203016713,0.4231552482,-0.2049936801,-0.095317252,-0.3909983337,0.0488236323,0.0800536722,0.1095759869,-0.1868070364,0.211766094,0.1607632339,-0.1171950027,0.2752373219,0.4020013809,-0.2543524802,-0.1254382581,0.2494219244,0.2224721462,0.1315016598,0.0198791679,-0.2210659236,-0.2818226218,-0.430657953,0.070774205,0.2052573264,-0.401001513,0.3377908468,0.09767396,-0.3812923133,-0.118126221,0.1145385131,0.3145771623,0.0429113805,-0.1349577308,-0.328235358,-0.5083823204,0.1449072212,0.0421730503,0.1967900544,0.2256726474,0.3546480238,0.049536705,0.2430287153,-0.4819770157,-0.0826899037,-0.2095017433,0.1999917477,0.2634357214,-0.0171152018,-0.0374913327,0.0263990145,0.1727620363,-0.0596017614,-0.2150503099,-0.366091609,-0.3127807081,0.0612799153,-0.1453043818,0.182321474,0.0940944478,0.2562635839,-0.0755431205,-0.2365216315,0.3544233143,0.0906276405,0.0280788615,0.3455359042,0.1791205257,-0.200653553,-0.187082231,-0.0185311232,0.0946881399,-0.1102112308,-0.2036915123,0.1857886016,-0.1922781765,-0.0072237239,0.016615672,0.1514560729,0.3099436462,-0.1421372741,0.3326796889,0.1762284189,0.0830450729,0.5306997299,0.2876535952,-0.0343114808,-0.0275180489,0.123540923,0.231922552,0.3524505198,-0.5596112609,0.0930919051,0.3111482859,-0.1309547871,0.3483591676,0.3157455623,0.0232933685,0.199672237,-0.0048853653,-0.0649216622,0.26689744,-0.3059867322,0.1353481859,0.3643882275,-0.1511551589,0.0086813699,0.4588139653,0.1298285723,0.1662643403,0.1535639167,-0.0163848381,0.4012465477,-0.1160306856,0.024524197,0.4541291893,-0.0029646347,0.0215496197,0.3619544506,0.0846554786,0.0848778263,0.1544637233,0.1645846814,-0.1245513633,-0.452374965,-0.0846994892,0.4605710208,-0.2176191211,0.0525100268,-0.1512880325,-0.1291843057,0.0573182218,-0.1781861782,-0.0756377056,-0.2884871364,0.4399873316,0.0605556965,0.1538340896,-0.0379774906,-0.0289665237,-0.0429917537,0.2498959452,-0.0769526288,0.0208796095,0.2935596406,0.022123808,0.2136040181,0.6038644314,0.1448430866,-0.0849439576,-0.1604552567,0.1932699084,-0.2475610822,-0.2024959326,-0.2125966549,-0.0310603734,-0.2145268172,-0.1687820405,0.377032876,0.3230584562,-0.2493104637,0.1287882626,-0.1560343504,0.2232077271,-0.2034609318,0.195464313,0.2016674578,-0.1328808069,-0.2870441377,-0.053905379,-0.1211345792,0.0998355225,0.1605126858,-0.4364562929,0.1686098427,0.0124979662,0.1488931626,0.2313269526,0.2973573506,0.0701317266,-0.0600894019,0.1038424522,0.0233305618,-0.4550192058,0.0580871776,0.3359411359,-0.1637790948,-0.1979536861,-0.1882534772,0.0778486505,0.3184640408,0.0243486632,-0.2890485525,0.0932431966,-0.0799017921,-0.2769651413,-0.0069261743,-0.0820759535,0.1544284374,0.2809629738,-0.3392101824,-0.0550968163,0.1450465024,0.1700152755,0.0563492179,0.4660274982,-0.2188387215,-0.0072853919,0.2061865479,0.2529653609,0.3293141723,-0.3406133652,0.1847258508,0.3814716637,-0.1514121443,-0.5562723279,0.1668208539,0.1318481863,0.2287925184,-0.067211926,-0.1191268861,0.0181221291,-0.003667041,0.1334282309,0.0585381351,-0.6600281,0.2794378996,0.0282123629,-0.178480342,0.1371931881,0.3119846284,-0.0557687096,-0.1405499429,-0.1189048961,-0.3230000138,0.4701814651,-0.0320267379,-0.2459295243,-0.0723553747,-0.0905000269,0.0480112731,-0.131629169,-0.3248487115,0.1184536219,0.2089709938,-0.064403601,0.0676755682,0.1719876826,-0.1898827553,-0.2444027662,-0.225262925,0.510014236,0.0994271785,-0.2092748284,-0.2593585849,-0.2064970434]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"Ok, sure.\n\nThanks for pointing on functional part.\nMy question is more\n\u201cPhilosophical\u201d\/Design perspective.\n\nThere are 2 perspetive:\n Add transformation methods to \n Dataset Class\n\n\n OR Create a Transformer Class\n which operates on Dataset Class.\n\nT(Dataset) \u2014> Dataset\n\ndatasetnew = MyTransform.transform(dataset)\ndatasetNew.save(path)\n\n\nWhat would be the difficulty\nof implementing a Transformer Class\noperating at dataset level ?\n\n\nthanks\n\n\n\n\n\n\n\n\n\n> On Jul 6, 2021, at 22:00, Quentin Lhoest ***@***.***> wrote:\n> \n> \ufeff\n> There are already a few transformations that you can apply on a dataset using methods like dataset.map().\n> You can find examples in the documentation here:\n> https:\/\/huggingface.co\/docs\/datasets\/processing.html\n> \n> You can merge two datasets with concatenate_datasets() or do label extraction with dataset.map() for example\n> \n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":142,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nOk, sure.\n\nThanks for pointing on functional part.\nMy question is more\n\u201cPhilosophical\u201d\/Design perspective.\n\nThere are 2 perspetive:\n Add transformation methods to \n Dataset Class\n\n\n OR Create a Transformer Class\n which operates on Dataset Class.\n\nT(Dataset) \u2014> Dataset\n\ndatasetnew = MyTransform.transform(dataset)\ndatasetNew.save(path)\n\n\nWhat would be the difficulty\nof implementing a Transformer Class\noperating at dataset level ?\n\n\nthanks\n\n\n\n\n\n\n\n\n\n> On Jul 6, 2021, at 22:00, Quentin Lhoest ***@***.***> wrote:\n> \n> \ufeff\n> There are already a few transformations that you can apply on a dataset using methods like dataset.map().\n> You can find examples in the documentation here:\n> https:\/\/huggingface.co\/docs\/datasets\/processing.html\n> \n> You can merge two datasets with concatenate_datasets() or do label extraction with dataset.map() for example\n> \n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n","embeddings":[-0.1065422595,-0.0538378395,0.0260536186,0.1430307478,0.3365959227,-0.1399590075,0.5725405812,-0.0339156091,-0.0521859936,0.1631585509,0.1168567315,0.2448585033,-0.4620172977,0.4327042997,0.2910209894,-0.3197534084,0.0604292937,0.211038664,-0.5410810113,-0.053649202,-0.3712553084,-0.3355924487,-0.0531556383,0.0580872223,-0.2199225277,-0.1509220451,-0.1271557361,-0.1756850481,-0.1485715061,0.0726781711,0.7325074077,0.0547797345,0.1794544756,0.6730666757,-0.0001188061,-0.0387484692,-0.0175584517,-0.0500610694,-0.0148862395,-0.398799181,-0.2108461708,-0.124951303,-0.0250348169,-0.3303232193,-0.1954077333,-0.0154983262,0.0561694205,-0.0434550978,0.3625746667,-0.0595809221,0.1046545133,-0.2699905038,-0.0658378005,0.0467020571,-0.1154842526,0.3715500236,-0.2796650827,0.0010369624,0.1731648743,0.3063555658,-0.0288388617,0.0741662532,0.0859446824,-0.3187921643,0.912211597,-0.0709178597,-0.0384920985,-0.3968650103,-0.2046209127,0.2444629073,0.4599398673,-0.2721917629,-0.634357214,-0.2657699585,0.0709742159,0.158929944,0.2296385765,-0.0058804196,0.0921110287,0.2785584629,-0.3245720565,-0.2360975146,-0.4755647779,0.093837291,-0.5276674628,0.0506132618,-0.1485310942,0.0155018717,0.0621879064,-0.0744666085,0.1951581836,0.0976155698,0.2587227821,0.2356418073,-0.2632692754,-0.3962749243,-0.1543958336,0.2665106952,-0.0398138203,-0.0534418225,-0.1178392321,-0.1095124632,-0.1519872099,0.1321215183,0.2445401698,0.3347284794,-0.067885384,0.1746526212,0.2386979759,-0.3149813116,0.0606371351,0.0193097442,0.0075841905,-0.1074852049,0.0635666028,0.0290941913,0.2385463417,-0.0591094121,0.2498930842,-0.2912393808,-0.1080684885,-0.0159943588,-0.1475112438,0.0228482094,0.2536226213,0.3971214592,0.3115744293,0.0693930238,0.3559366167,-0.137148425,0.1234385967,0.0896292403,-0.2077120394,-0.0085141491,0.0254325047,-0.0882945061,-0.2700825334,-0.0413119048,-0.0517634265,0.3604640663,0.0132195698,-0.435118556,0.1218436435,0.3219591081,-0.4701114297,-0.2149737626,-0.1325143576,-0.0192559194,-0.1424986422,0.4075756073,-0.1337510794,-0.1032077968,0.2736286521,0.0855005085,0.0136810727,0.0905926451,-0.6206724048,0.3858294487,0.0393181667,-0.1848632097,0.0401023738,-0.3431012332,-0.1524264812,-0.1820155233,-0.1730383933,-0.0275773164,-0.4880181253,-0.3247206807,-0.0589682125,-0.0078160465,0.0338281095,0.2875152826,-0.3181762993,0.5770632625,-0.0649325103,-0.1183059886,0.7609329224,-0.1623450071,-0.3263252676,0.1355405003,-0.3932285011,0.1056018919,-0.0285580903,0.3570363522,0.3934704661,-0.3740145266,-0.5503361821,0.2517984509,-0.050300166,-0.0115070008,0.1889352947,-0.1970737278,0.497328788,0.0133823594,0.2665420771,0.3399263024,0.0465910397,0.3440959454,-0.2913905978,-0.4086551368,0.2489724904,0.0150376074,0.2712851465,-0.007466495,-0.0214922521,-0.4259054065,-0.2000454813,-0.031809289,-0.0618993305,0.326859504,0.3026998341,0.0083405804,0.1577600837,-0.1190226078,-0.2191333771,-0.27046296,-0.0120443031,-0.3185946643,0.3225788772,-0.5568806529,-0.3735568225,0.4142183661,-0.0540543348,0.2334824353,-0.5580812693,0.1156936437,0.0233746469,0.2491554767,-0.3481840193,-0.0079235677,-0.1335957795,-0.2917013466,-0.0953387693,-0.0589290373,-0.0261338111,-0.1575662792,-0.072056435,0.4762631655,0.2876897454,-0.2148327529,0.3182387352,0.0622443222,0.2261413932,-0.1458975971,-0.3460884988,0.2852140963,-0.0535704494,0.3659822345,-0.0880966336,-0.0648378655,0.1404396594,-0.2240234315,-0.2181909084,0.1399024129,-0.1129794866,-0.072604008,0.3441033959,0.1270374954,-0.0325209014,-0.181943059,0.2679167986,-0.1034208536,0.1382667422,0.4796472192,-0.1408270299,0.2142405212,-0.3483157754,0.4952425957,0.621920526,0.1169676855,-0.2442904115,0.3060922623,0.2635382712,-0.0524754114,0.1504600942,0.1145002246,0.236633569,-0.1106892005,0.0115367128,0.355188936,0.1476724297,-0.4427650273,-0.1622827202,-0.0296411291,-0.1099239439,0.0588960797,-0.0287238415,-0.2380894721,-0.3307723105,-0.0129873464,-0.1074386165,0.0961227864,-0.0999431238,-0.1937006116,0.0738501027,0.0243850909,0.1707930714,0.3343965113,0.1304818839,-0.1242378727,0.0231084675,0.0867313594,0.1082578972,0.0292638596,-0.0491125286,0.0383777544,0.2854854465,-0.3548637927,0.101978682,-0.049428083,-0.6258038282,0.2046298385,0.0114048105,0.1709541082,0.3630703688,0.0625498071,0.0274036489,0.0422574393,0.1247981116,-0.0741003156,-0.233915478,0.0146628302,-0.068667911,0.1758811772,-0.0053665605,-0.102647908,-0.1687109917,-0.2311032563,0.4608861506,0.1201040819,0.0761675462,-0.0585082509,-0.2050844431,0.0445598252,0.0041256682,0.2625152469,-0.0093501452,-0.1128629595,0.2273569554,0.0768129528,-0.2224090695,-0.2462584972,-0.2503373921,0.034182623,0.5150563121,-0.2194655985,-0.1157506183,-0.0979348272,0.1770817786,0.0443448499,0.1360614747,0.6754920483,0.1972072721,0.1080714241,-0.0683113262,-0.2257611006,-0.0008916733,0.2898741663,0.124386169,0.4621849954,0.5110164881,-0.0024083038,0.3339212537,-0.1510581374,-0.2582774758,-0.2768117189,0.1763470769,0.0443195552,0.0027491101,-0.4018685818,0.2987800241,-0.071896635,-0.341355294,-0.0681767762,0.0344203748,-0.1198191047,-0.2663063109,-0.2975162268,-0.1594110131,-0.2106073648,0.1818176806,0.0393158086,0.1803296804,-0.0807615295,0.1164886355,-0.0886694267,0.1549246311,0.1667349637,0.6605191827,0.2578846216,-0.0099289091,-0.1072394475,0.0872032642,0.0400460772,0.0366869234,-0.037353266,0.1151957735,0.0354973003,-0.0296959747,-0.1387937963,-0.0774607584,0.673421979,-0.1263867617,-0.0538823642,-0.0925459191,-0.5396211743,-0.2093902677,0.2259621173,-0.2650752664,-0.2224705219,0.1746706665,0.2993132174,-0.1425945461,-0.1577892751,0.1220945418,0.4064098001,-0.3311167061,-0.406367749,0.0549223237,-0.2408884317,-0.1778575778,-0.1317899674,-0.0959737971,-0.2433741838,0.0306980256,0.0425296128,-0.3838761449,0.0047154692,0.1696370393,-0.0302257128,-0.241641596,-0.1622006744,0.1318225712,0.1834803671,0.2295386195,0.1789854914,0.1817311496,0.162709251,-0.6273981929,0.127833426,0.2292990834,0.2480421364,0.015640825,-0.049278453,-0.130185008,0.0483325236,0.1366274655,-0.275672555,-0.0003731506,-0.0455011949,0.3221073747,-0.3522961438,-0.5900357962,0.3108597994,0.2279366404,-0.1028885543,-0.1544230878,0.1743733287,-0.4822404683,0.134019956,0.3647257388,0.9555786252,-0.0171788931,0.0775626078,0.130807519,-0.0602580085,0.4272616208,-0.2183820903,0.0276527964,-0.1985087246,-0.3663853705,-0.3195909262,-0.275773108,0.4778462052,0.4035706818,-0.369599551,0.4447704554,-0.0268549919,0.0830422938,0.3783856034,0.0587674119,0.2714858651,-0.5816074014,-0.3119212985,0.0892602727,-0.1057594642,0.2805681527,-0.1049875021,-0.1866158098,-0.2718273699,0.077004917,0.020539606,-0.0427926518,-0.3292389214,-0.1995063573,-0.0850663036,-0.2173229754,0.8510027528,0.3416753709,0.2968756557,0.2104676664,-0.0236104466,-0.1445991546,-0.1945162863,0.3003870547,0.2033179551,-0.1505187303,0.5049753785,-0.1062807068,0.0509650744,-0.0636290014,-0.0944138691,-0.200209707,0.2499517053,-0.1495440453,0.0199246909,-0.4891405106,0.0657972991,0.0772785991,-0.0438061804,-0.0220829975,0.089304693,0.1573809832,0.0038472957,0.3480000496,0.2357848436,-0.2018477768,-0.0627061576,0.0013767285,0.3610284626,0.1247055307,0.0428968146,-0.3429124057,-0.2802804708,-0.1254491657,0.2406890988,0.0951169804,-0.4211356342,0.2090701908,0.0677030161,-0.0068520899,-0.1850261539,0.0265440904,0.2446223646,-0.0467583016,-0.4049022496,-0.3855032623,-0.50062567,0.1083523855,0.2817618251,-0.0155725712,0.0729669109,0.461671412,0.195802018,0.1672312319,-0.2731506228,-0.1845598519,0.0571261942,0.1818388104,0.3260120153,-0.1249695569,-0.1513395011,0.0886022151,-0.0421320498,-0.082882978,-0.1371410936,-0.1382937878,-0.1004508585,0.1704818606,0.1444212645,0.0493487753,-0.1290219575,0.0181037728,-0.0711755976,-0.1099515483,0.1145004109,0.1153315753,-0.1027354971,0.2702432573,0.1610773057,-0.0335919224,-0.5131677985,0.2962310016,0.0904715434,0.0359370299,-0.2626496851,0.1954117864,-0.0403056778,0.1127259508,-0.4077843428,0.2016228139,0.2528677285,0.076658532,0.4542482793,-0.2142665535,0.5300105214,0.4562421441,0.1906014085,0.0942621008,-0.1394414157,0.0893091187,0.213786602,0.0539295487,-0.3617954254,0.2642563581,0.3354158998,-0.1211386174,0.0393023007,0.4257610738,-0.0247116778,0.1340170652,-0.160720408,-0.1882659942,0.2288958579,-0.1902733892,0.1538883001,0.547901988,-0.1896898299,0.3101469874,0.5182446241,0.2218762636,-0.0831048563,0.2689274848,0.1234965995,0.26371333,0.0905309916,-0.0205284152,0.3966085613,0.0287146065,0.0679489151,0.5042470694,0.0737647116,-0.2143763155,0.1243715957,-0.0610744357,-0.0991160721,-0.3878371716,0.156518802,0.2266200185,-0.1383497864,0.4410702884,-0.0869603083,-0.0259663872,0.1422654092,-0.2406668812,-0.324413985,-0.3866647184,0.3253192902,0.1327977329,0.1506992728,0.041624181,-0.0754017606,-0.051692076,0.2463622093,-0.0627192408,-0.0002610857,0.2278399467,0.2276692241,0.09294945,0.5396186709,0.0761842132,0.2316832691,-0.0407461822,0.1827039421,-0.2275706828,-0.1864681095,0.039115116,0.1910353452,-0.1933148801,0.003873724,0.2546108663,0.0740204081,-0.0374063067,0.2260051966,-0.1073445603,0.4486500919,-0.3857946098,0.3867209554,0.0828126818,-0.1325118095,-0.1683933437,0.1630371362,-0.1195959151,-0.0303969998,0.2465758324,-0.3848781288,0.0494269617,-0.0104873329,0.0293313079,0.2443383634,0.1756816357,0.2190907896,0.1514759064,0.0247954335,0.0265753269,-0.3054023385,-0.0843892172,0.2060105056,-0.1895982474,-0.0302192904,-0.3239158392,0.0980096608,0.461494416,0.0137518765,-0.3440375626,0.0189509559,0.0597928166,-0.1077382788,-0.047847461,0.0716327876,0.3046370149,0.0701382384,-0.2273862809,-0.0252642725,0.2907377183,-0.045555681,0.0754397511,0.4472019374,-0.123066932,0.0623973384,0.138762787,0.2046008706,0.3971917629,-0.06043268,0.1115897149,0.3480598927,-0.0256991871,-0.3430043757,0.4000006318,0.0863488168,-0.0067718397,-0.1952468753,-0.1211528629,-0.0640130118,-0.2995355129,-0.1213253289,0.160443604,-0.7866070271,0.1051194519,-0.1982639879,-0.2150479108,0.1535719782,0.2690836191,-0.0474410504,-0.1647903472,-0.1739923209,0.0095409276,0.4162622392,-0.171244964,-0.2035200298,-0.0616413802,-0.245426178,-0.0794795528,0.0124921743,-0.3885910213,-0.0811210871,0.2913827002,-0.0351548418,0.2398268729,0.1916748583,-0.2044728547,-0.2209127098,-0.1321309656,0.6106513143,-0.3881576359,-0.0434254259,-0.2438924015,-0.0093143983]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"I can imagine that this would be a useful API to implement processing pipelines as transforms. They could be used to perform higher level transforms compared to the atomic transforms allowed by methods like map, filter, etc.\r\n\r\nI guess if you find any transform that could be useful for text dataset processing, image dataset processing etc. we could definitely start having such transforms :)","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":64,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nI can imagine that this would be a useful API to implement processing pipelines as transforms. They could be used to perform higher level transforms compared to the atomic transforms allowed by methods like map, filter, etc.\r\n\r\nI guess if you find any transform that could be useful for text dataset processing, image dataset processing etc. we could definitely start having such transforms :)","embeddings":[-0.490424931,0.0872302279,-0.1906813085,-0.0839954689,0.3473585546,-0.1753206402,0.3381084502,0.2574364245,0.0903777182,0.019790493,0.2580311894,0.3618627489,-0.3850794137,0.4767219126,0.1930702329,-0.2324400842,-0.151134789,0.265183866,-0.4250871539,-0.0168538652,-0.3289776146,-0.1533589661,-0.0330886394,0.1328707039,0.002900715,-0.1466348916,-0.0935134068,-0.3280186951,-0.0702162236,-0.1045169681,0.3167803884,0.1368869245,-0.0556037799,0.4302741885,-0.0001056646,-0.1033560857,0.1528594941,0.0172446407,0.0270981509,-0.3660884202,-0.2878669798,-0.2310528606,-0.0796364918,-0.3515015543,-0.2724731266,-0.1547861993,0.1759620458,-0.3427095115,0.5134334564,0.1626292765,0.2492758334,-0.2664002478,-0.1245223358,-0.0106896963,-0.0541695543,0.4239853621,-0.3033707142,-0.5247583389,0.4609455764,0.23065494,-0.1012291089,0.1187341437,-0.0994046628,-0.1907213777,0.630456388,-0.1156991124,-0.0931705385,-0.5817266107,-0.078451395,0.295892179,0.410266459,-0.2705188394,-0.4758317471,-0.2998233736,-0.0145636881,0.1219541728,-0.2259934396,0.2029168755,-0.0218977723,0.327022016,-0.2893729508,-0.2254891992,-0.3472728729,0.0120361308,-0.2229361981,0.1600910425,-0.097874403,-0.025985254,0.0453669541,0.0403649323,0.2370888144,0.0952521786,0.1654207408,0.1429382712,-0.0912426338,-0.428411901,-0.122181274,0.0104457103,-0.2088413686,0.2001110315,0.0329007506,0.1262459904,-0.0182830803,0.049241744,0.0054219016,0.1322811097,0.0045804046,0.0382177345,0.2880474329,-0.3633047342,-0.0313483216,0.1370051354,0.0354362167,0.0212808382,-0.1402965784,0.1661054641,0.0147123355,0.0149487611,-0.0016047077,-0.2726562321,0.1066770554,0.1336745769,-0.1337660551,0.0749790147,0.1600473225,0.3427148461,0.4094092548,0.0567138046,0.2407179177,-0.1985290796,0.0100685144,0.171739921,-0.2010185272,-0.0314506255,0.0899902508,0.0984608829,-0.1902890354,-0.085748896,-0.1360107511,0.403221637,0.2586643994,-0.0103076641,0.3266676664,0.287661612,-0.3840346634,-0.2247590423,0.0478359126,-0.0214083195,-0.2645701468,0.1571627855,-0.2843391001,-0.0495866835,0.1654486954,0.2562448978,0.0745678619,-0.0825993717,-0.2585091889,0.4359450638,0.0350625217,-0.128864184,0.1879973412,-0.2153193802,-0.2061604261,-0.0929540545,-0.1613592058,-0.085530296,-0.5210264921,-0.4330343306,-0.2928813994,0.0670860708,0.1855181307,0.1258127093,-0.0995190516,0.2549536228,-0.0540064499,0.0176642574,0.6603587866,-0.1747204363,-0.0979027748,0.2688248754,-0.357213378,-0.1598703563,0.1279026419,0.4035031796,0.3632651269,-0.2125305235,-0.4146336615,0.353454262,-0.0909692273,0.0054099564,0.1527169347,-0.2682519853,0.2469784915,0.2128084153,0.3504933119,0.2810118496,0.1294330359,0.1401317567,-0.1105129644,-0.4927369952,0.3531637788,-0.0376637019,0.1912439615,0.0206499044,0.042872902,-0.6312668324,0.188138321,-0.1162087768,-0.270983994,0.1082687154,0.2710978687,-0.1145076528,0.1172646508,-0.155006215,0.0733629018,-0.1630351245,0.2363794148,-0.3905517161,0.1708755344,-0.2011028528,-0.2399889529,0.3937875926,0.0097546885,0.053466253,-0.4360540509,0.1432032436,0.2297930866,0.1855835766,-0.1641368717,0.1989326328,-0.3491078913,-0.2043966353,0.0468196012,-0.228336513,0.0734595135,0.0671370849,0.0497843623,0.5814134479,0.2575738728,-0.0738892183,0.271571666,0.0297425967,0.1512193084,-0.0686642528,-0.1690057069,0.6479952335,-0.0027132689,0.2556301355,-0.1393957436,0.1753908992,0.2669109702,-0.3181121349,-0.265509367,-0.0300369971,-0.03970946,-0.0646846816,0.2421483845,0.0481635667,0.0189805608,0.046282962,0.251547575,0.015413492,0.2425545454,0.3825644851,-0.1510885954,0.2011816353,-0.1937962472,0.578283608,0.4914995432,0.2047153115,-0.2291341573,0.1965982318,0.1805032939,-0.0880012214,0.2533913553,0.159966737,0.305857718,0.0324335583,-0.0279732086,0.1396907419,-0.0631873831,-0.4729793966,-0.1748874038,-0.0551563501,-0.0589844957,0.2467169911,-0.2987349033,-0.3389695585,-0.1829380691,-0.0542922653,0.2892733514,-0.0513308868,0.2056132406,-0.2524846494,-0.1110305563,0.1603519917,-0.0294568706,0.5149572492,0.1022286192,-0.0487784296,-0.2807581127,-0.2141819298,0.0795971602,0.1924239844,0.0171542037,0.0326939672,0.4736204445,-0.2052599788,0.355670929,-0.2228472233,-0.5828427076,0.0999873355,-0.0393820554,0.1567539126,0.3471787572,-0.0640847683,0.1175071076,-0.0812777951,-0.0054268581,-0.1949022859,-0.288954258,-0.0580740124,-0.1537415534,-0.0682870895,-0.1009423882,-0.1854184419,-0.3057166338,-0.3653067052,0.3621490896,0.0168432351,0.1068324968,-0.146415174,-0.0528573878,0.0466547571,0.1201931238,0.1184352413,-0.0254238285,-0.0544987619,0.1331170946,-0.1406106502,-0.1921079904,-0.2421644777,-0.3054256737,0.1779670268,0.3593081534,-0.1374822557,-0.0864887014,-0.179784596,0.0269064307,0.1405414641,-0.0950946137,0.5606482029,0.0580535121,-0.1028714776,-0.1430615485,0.0960488915,-0.1594544053,0.1404044777,0.062264584,0.2014611214,0.4887650311,0.0573834777,0.3966763914,-0.2763507962,-0.5605832934,-0.1552893072,0.0559178889,-0.1592636555,0.0357731134,-0.4223344922,0.1116604879,-0.0961359441,-0.1843737215,0.014401786,-0.0177209787,-0.4235361516,-0.0908083096,0.0125059383,-0.2403864563,-0.1080766618,0.1225372106,-0.05863823,0.127152428,-0.080681771,-0.0054047261,-0.0942281634,0.0064892541,0.250151813,0.3186603487,0.2244582921,-0.0671207383,-0.0895250216,-0.1313562542,0.0592401251,0.1989096701,-0.0286223646,-0.1292306483,-0.0933870748,-0.1276375204,-0.0605837926,-0.0935154334,0.6798669696,-0.2585861981,-0.1833595037,0.1150170863,-0.5211288333,-0.2116234303,0.1572163403,-0.2936236262,-0.1387544572,0.1846016347,0.2109845132,-0.0040839324,-0.096831128,-0.0436809473,0.0557717346,-0.2764881253,-0.1382437497,0.1732911468,-0.1675199717,-0.4046922922,0.0434761532,0.0667044893,-0.1190935969,-0.1624394208,-0.220287025,-0.3782520592,0.0179277323,0.2050927877,-0.1187672466,-0.3053462207,-0.211531803,0.1628768444,-0.0756174698,0.2482209653,0.1217731684,0.0111041255,0.177158162,-0.4412103295,0.2124973685,0.0691914931,0.2468560636,-0.0770299584,0.069558844,0.0136773484,0.0993144438,0.1730927825,-0.4768767357,0.0910902992,-0.0490896776,0.1310815513,-0.1700926721,-0.5385859013,0.4164545238,0.3622052968,-0.1425336152,0.1924139559,-0.1396230459,-0.4583474994,0.3108945489,0.545820117,0.8820564747,0.0202464741,0.1815938056,-0.1626459062,-0.2520898879,0.365999192,-0.3363492191,0.1634489447,-0.3765651882,-0.3504263461,-0.2777534723,-0.1602724642,0.358818084,0.3163309395,-0.44214499,0.2141090184,0.1854150593,-0.2774670422,0.0927125812,0.1222987473,-0.0064123441,-0.5177205801,-0.0617267825,0.2326449454,0.0147148399,0.2346643209,-0.0767123923,-0.2186506391,-0.0635324642,0.2925375998,-0.2410867512,-0.0898591951,-0.1945822239,0.0760053396,-0.1570572853,-0.1579389721,0.5903674364,0.2139989287,0.1131554246,0.1655705571,-0.1522759497,0.0286466982,-0.1678531468,0.2973020077,0.059607666,-0.2248904258,0.5294120312,-0.0713268518,0.0227728896,-0.050038103,0.0985265225,-0.2347023338,0.0566844828,-0.0197727531,0.0146204969,-0.2881886959,0.037441887,0.0680967271,-0.0847849548,-0.0097793834,0.205627799,0.1391351223,0.1021850929,0.4880701005,0.2477913946,-0.2698325813,-0.0187059678,0.048501689,0.2433785796,-0.0282656904,0.107859984,-0.2474265099,-0.3079802394,-0.4038124084,0.1936316341,-0.0800648108,-0.140297696,0.4241030216,0.0560077019,-0.2723399401,0.0227897242,-0.2902472019,0.2731313109,-0.0606276132,-0.3673821688,-0.4418162704,-0.2765259445,0.0984336063,-0.0481786095,0.1748621017,-0.1195066199,0.1579867005,0.2233348787,0.0192567687,-0.4004312158,-0.0206101947,-0.0466725081,0.0773356557,0.1480118334,-0.011920128,-0.3965807557,0.1831103712,0.0164039321,0.0023940662,-0.2839089036,-0.2874855101,-0.1735991985,0.1151174754,-0.051675152,0.2266323566,0.0973276794,0.1461656839,-0.1276879013,-0.0839437991,0.0175205823,0.0877498165,-0.065720275,0.3965675235,0.3393174708,-0.3135195673,-0.1280070692,0.1323916465,0.1197141483,0.0109750889,-0.1404135227,0.4049523473,0.2219867855,0.0319917314,-0.1367147416,0.1751733869,0.2696063817,-0.0433321372,0.5080503225,0.0162172522,0.3049813509,0.5331389308,0.0364385843,0.1318987906,-0.143276006,0.1320046186,0.1662057191,0.2966307104,-0.2822293341,0.0934450999,0.2485661507,0.0942149907,0.055136133,0.2299286872,0.1871549934,0.0846951604,0.1109358147,-0.1304395199,0.3049073815,-0.494229883,0.2468261868,0.3790172637,-0.0694418177,0.0271546505,0.3060771227,0.0716517791,0.0685179979,0.3909011483,0.2271911651,0.5443760753,0.2785642147,0.0679229945,0.3333747089,0.0951083079,-0.0224438682,0.3389113545,-0.1312888563,0.2181161344,0.2518438995,-0.0386553071,0.1300686747,-0.1997902989,-0.1038370952,0.4290379584,-0.2054002881,0.0887924582,-0.2658657432,-0.0644456595,0.1366759539,-0.276855886,-0.0632976666,-0.1097517237,0.4635696113,0.0902050883,0.0708074272,0.1755949855,0.2030889094,-0.1322399974,0.224802658,-0.1567337811,-0.1942661554,0.1774763167,-0.0101630557,0.1475271881,0.4574742913,0.225829348,0.1731732786,-0.4393765628,0.2172538936,-0.3352845311,-0.2082688659,-0.1618868113,0.4071256518,-0.1332312524,0.0322862379,0.1816322207,0.273787111,-0.1549276412,0.1682000756,-0.0191140827,0.2027835697,-0.2083941698,0.5841745734,0.3528769612,-0.1191038191,-0.1195924133,0.1830760986,-0.1006718725,-0.035318207,0.4973428845,-0.5335724354,-0.0101907793,0.0372523628,0.1149991602,0.0839328542,0.2910145819,0.1034574658,0.148253873,0.1419586986,-0.0377888717,-0.1565444767,0.0021484452,0.2513350546,-0.0425239429,-0.1603260636,-0.1410849839,0.0675836802,0.2708384395,-0.0268762633,-0.3874260187,-0.1746884137,0.1546907425,-0.1828951687,0.1026938632,0.111209318,0.2961474061,0.0896124914,-0.2545866966,-0.0773444995,0.2244994491,0.1373787522,-0.0403169617,0.33288306,0.1304400712,0.1238557026,0.2548811138,0.2091063261,0.4130344391,-0.2154668272,0.3468508422,0.2343660891,-0.1899389476,-0.2591225505,0.246782288,0.1187880412,-0.0343256332,-0.2198068351,-0.2305340916,0.0652120635,-0.153087303,0.0055776597,0.1222296879,-0.5944423676,0.0167276971,-0.2975139916,-0.0941293463,-0.0070219971,0.3109409511,-0.000174347,-0.1021811739,-0.0234922022,-0.2513223588,0.2158101052,-0.1368375272,-0.4659645259,-0.0929828808,-0.1370571256,-0.3058809936,0.1579550058,-0.1381411105,-0.1742243767,0.3162984252,-0.2496904284,0.3183437884,0.2175519615,0.0222028624,-0.2063117176,-0.2627237439,0.4805034399,-0.1095903367,-0.0933707878,-0.2571106553,-0.2390016317]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"Thanks for reply.\n\nWhat would be the constraints\nto have\nDataset \u2014> Dataset consistency ?\n\nMain issue would be\nlarger than memory dataset and\nserialization on disk.\n\nTechnically,\none still process at atomic level\nand try to wrap the full results\ninto Dataset\u2026. (!)\n\nWhat would you think ?\n\n\n\n\n\n\n\n\n> On Jul 7, 2021, at 16:51, Quentin Lhoest ***@***.***> wrote:\n> \n> \ufeff\n> I can imagine that this would be a useful API to implement processing pipelines as transforms. They could be used to perform higher level transforms compared to the atomic transforms allowed by methods like map, filter, etc.\n> \n> I guess if you find any transform that could be useful for text dataset processing, image dataset processing etc. we could definitely start having such transforms :)\n> \n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":155,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nThanks for reply.\n\nWhat would be the constraints\nto have\nDataset \u2014> Dataset consistency ?\n\nMain issue would be\nlarger than memory dataset and\nserialization on disk.\n\nTechnically,\none still process at atomic level\nand try to wrap the full results\ninto Dataset\u2026. (!)\n\nWhat would you think ?\n\n\n\n\n\n\n\n\n> On Jul 7, 2021, at 16:51, Quentin Lhoest ***@***.***> wrote:\n> \n> \ufeff\n> I can imagine that this would be a useful API to implement processing pipelines as transforms. They could be used to perform higher level transforms compared to the atomic transforms allowed by methods like map, filter, etc.\n> \n> I guess if you find any transform that could be useful for text dataset processing, image dataset processing etc. we could definitely start having such transforms :)\n> \n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n","embeddings":[-0.4273887277,0.2038237303,-0.0892710388,0.0677168295,0.522382021,-0.1340120435,0.3225390017,0.1911114901,0.013091499,0.0615377389,0.2266187072,0.3506041467,-0.4103445709,0.2984676659,0.0133415768,-0.0542735718,-0.0388857797,0.2965457439,-0.5386686921,-0.0739863887,-0.167423442,-0.2261765152,-0.0746393502,-0.1188812032,-0.0701206326,-0.0560871586,-0.1632028073,-0.4052655995,-0.3135545552,-0.0657363608,0.5253243446,0.2762766778,0.0223159511,0.4755660892,-0.0001143179,-0.1631057411,0.2398658395,0.0084741171,-0.1145883426,-0.443159759,-0.338857919,-0.2092615813,-0.0825239941,-0.3510289192,-0.2636530101,0.0173992235,0.0538823754,-0.3100696504,0.4917915761,-0.0407068171,0.1410671771,-0.3664943576,0.0061739157,0.0395747162,-0.2592406273,0.4160124958,-0.1895077974,-0.265927881,0.2792956829,0.2837553918,-0.107362546,-0.0279848874,-0.0470017865,-0.2350812554,0.6600618958,-0.1943365782,0.0043311757,-0.5277867317,-0.0465322025,0.3039246798,0.6490062475,-0.3981794417,-0.4264175594,-0.1945628077,-0.0383273773,0.1389341503,-0.0498606153,0.244358778,0.017654974,0.3849679828,-0.2907490432,-0.2940432727,-0.4222829044,0.0256696455,-0.3817669153,0.1451612711,-0.0047396049,-0.0499595143,-0.0384320952,0.1308086514,0.3123420775,-0.1127013713,0.2272325754,0.0686238334,-0.1626562476,-0.5328313112,-0.3075406849,0.09869425,0.0344533175,0.1020556912,-0.0561499521,0.0376796611,-0.0647402331,0.0121175535,0.137759611,0.2655591369,-0.0177641474,0.1145247743,0.1995014101,-0.3872465491,0.0304755718,0.2286774367,0.0088771312,-0.0593662374,0.1155702844,0.1483322382,0.0488787293,-0.0583638959,0.0696761757,-0.1357873678,0.0441751406,-0.045509316,-0.2064178139,-0.0037700229,0.210297972,0.4288657904,0.2672246695,0.152155742,0.20641689,-0.2488153875,0.077928558,0.0225015283,-0.3289786279,-0.0778612942,0.0461385325,0.0587496422,-0.310872376,0.1111935154,-0.16365771,0.4183041751,0.1364359558,-0.2913190126,0.1636843681,0.1784058511,-0.3646046817,-0.0293231923,-0.0457364544,0.0250666421,-0.2207814157,0.4000896811,0.0081141181,0.0546802916,0.1735207289,0.1761144698,-0.0178801324,0.136665225,-0.2848555148,0.2927871048,0.162032038,-0.0634124801,0.1134386882,-0.4430219233,-0.1859663129,-0.1998777539,-0.1033181474,0.0172009487,-0.627992034,-0.2453715205,-0.1362820268,0.1093932837,0.036538288,0.2848489285,-0.1213752925,0.225227952,-0.1534097493,-0.2330722362,0.5798242092,-0.0970395505,-0.2454657406,0.2289628536,-0.3728942275,-0.1394857615,0.1957407594,0.2967514098,0.4605460763,-0.2393575311,-0.4908559024,0.2818575203,-0.0648387074,-0.1180446744,0.0126638273,-0.2541800737,0.3696837127,-0.0692067295,0.2602415681,0.199729383,0.0043156072,0.3201088309,-0.0260312054,-0.4473751783,0.3603182435,-0.006686795,0.4549993277,0.0567456298,0.0585832372,-0.3478059173,-0.0341736339,0.0530903563,-0.0473659523,0.246687606,0.3147389293,-0.0091185896,0.1841320395,-0.1566165984,0.033421725,-0.1146244556,0.1116199046,-0.3407637179,0.2966465354,-0.1415416747,-0.2600739002,0.5447855592,0.0473795496,0.1205488965,-0.5010402799,0.3326478302,0.0500790179,0.1671906859,-0.1501583308,0.0597823299,-0.3445490003,-0.192339316,0.0210279152,-0.1629345566,0.131426543,0.0772699341,-0.0494827703,0.6066023707,0.4095857143,-0.1700623482,0.4739151001,0.0510801636,0.1334010363,-0.1194854155,-0.1270043999,0.492796272,-0.0984402373,0.2128577828,-0.0814322904,-0.0044127326,0.2524282634,-0.2985364497,-0.3720683455,-0.0959758386,0.1043319032,-0.0137542076,0.2253137827,0.2128076702,0.015363995,0.0124664828,0.1573317945,-0.0755052119,0.142143786,0.3831191361,-0.1746159941,0.2205764353,-0.2004304379,0.4899602234,0.5510008931,0.1705225408,-0.2451501489,0.1599907428,0.1105057374,-0.065086484,0.2228177488,0.0525744408,0.3805498779,0.2248176634,-0.0298500881,0.1878754646,0.0349261984,-0.3846750557,-0.0897410288,-0.0698584244,-0.0528352521,0.0159650911,-0.2960271239,-0.2227744311,-0.4385722578,-0.0818984509,0.0492287092,-0.0255590435,-0.0404092558,-0.0424437262,0.2382252216,0.1159824878,-0.1045534015,0.3931617141,0.114002265,-0.1347837299,-0.1402972639,-0.0884842128,0.0053355726,0.0395600609,0.0950410217,0.0166571345,0.4365256429,-0.1513846815,0.1293430775,-0.0760705844,-0.5865780711,0.071924217,0.0726535991,0.130654633,0.2849867046,0.0718221292,0.1427114457,0.0545829237,0.034046974,-0.1997928768,-0.3182959855,-0.0681257844,-0.0269995295,-0.0561735779,-0.0062386389,-0.1270177513,-0.3657001257,-0.1590499133,0.2604333758,-0.0370676331,0.0225017499,-0.1918992847,0.0366403386,-0.1675834954,0.0392668284,0.1488780379,-0.1130215228,-0.0118303867,0.2623637021,0.0692795813,-0.2676432431,-0.2812874317,-0.3509054184,0.1491180807,0.2953979671,-0.2414922565,0.0110005792,-0.1852772087,0.0474693291,0.0062265396,-0.0530401058,0.6440244913,0.1352549493,-0.0090843812,-0.1666705012,-0.0123998169,-0.0724096671,0.0961196497,0.1759649664,0.3626632094,0.4776412547,0.1353176981,0.4528478682,-0.0309897773,-0.5186715126,-0.2558555901,0.2427676022,-0.0109792529,0.0309007037,-0.4694468975,0.2096977979,-0.0902997553,-0.406622678,-0.1012916416,-0.0884548277,-0.2011141926,-0.1499274969,-0.0057664569,-0.1017189175,-0.132105127,0.1723469049,-0.1624356806,0.124637492,-0.063758567,-0.0338146761,0.0795984492,-0.0344646461,0.1725523025,0.432340771,0.0813840777,-0.0620546304,-0.0586341433,0.0310685523,-0.0384063981,0.0896900296,-0.1239117086,0.0169133078,0.0803166777,0.0498193055,-0.1409437358,-0.0480195321,0.9176001549,-0.1054358706,0.0751839876,-0.0528622828,-0.4078983665,-0.1606277525,0.0555228069,-0.2564148903,-0.2815268636,0.1257937849,0.1607528478,-0.1736698151,-0.0467674956,-0.1674887985,0.1465177536,-0.3606066108,-0.3681034446,0.1373229474,0.019704394,-0.1851515323,-0.1351667345,0.0603419915,-0.093931675,-0.323459357,-0.1602891237,-0.3461585343,-0.1559464633,0.3014028966,-0.4090832174,-0.1805351675,-0.2664678395,0.3182810247,-0.0156427342,0.2402196079,0.3277978599,0.0834525824,0.096734859,-0.4447584748,0.1128816977,0.0916415825,0.158024013,-0.0036629899,0.1313509196,-0.2520060539,0.1054589748,0.0537693016,-0.5604137182,-0.060542684,-0.0197843276,0.1837174296,-0.4386496544,-0.5073031783,0.281085372,0.3722704053,-0.0437569804,0.1340916008,-0.088326551,-0.538308084,0.3317032456,0.5413069129,1.0115053654,0.035308592,0.0951091945,-0.1618190855,-0.2940162122,0.4121867418,-0.255125165,0.1374658197,-0.4025069773,-0.3175671101,-0.3263966143,-0.3039836287,0.4266259372,0.3661136627,-0.469766587,0.2520799637,0.0660475641,-0.1698084027,0.1787824482,0.3193556666,-0.013073408,-0.4183338284,0.0711157098,0.1410792023,-0.0415542647,0.1963189244,-0.0462236777,-0.272369206,-0.2079685181,0.2208788097,-0.3255499899,-0.1173690856,-0.3221346438,-0.0072343396,-0.1927797049,-0.2832835615,0.5923351049,0.1904491633,0.2429208905,0.2127527744,-0.0905344561,-0.1015583575,-0.0485848263,0.2503943443,0.1308312565,-0.2683895826,0.5263625383,0.0036335667,0.0560063533,-0.1335294992,0.0174610913,-0.1048344523,0.0918322876,-0.3402468562,0.0799732357,-0.4458642304,-0.0323554613,0.0731363967,0.0000034451,0.0337571763,0.1141242906,0.2982451916,-0.005326522,0.5414873362,0.2411539257,-0.1254365891,-0.0880634636,0.2546875775,0.3291701972,0.0857016146,0.0439829156,-0.2423387915,-0.3445209861,-0.2269329578,0.2997029722,-0.0933081806,-0.17675744,0.2913911343,-0.049197726,-0.073546797,-0.0707887337,-0.2497788072,0.3696146309,0.1445183903,-0.3072431982,-0.4929698408,-0.2234714478,0.1468709707,0.0272388328,0.2472435385,-0.0569983013,0.2250389159,0.1466003805,0.051281739,-0.3249670267,-0.1563585699,0.1272403151,0.0378690511,0.0823526904,-0.1206547469,-0.4654295444,0.0243737921,-0.0410165787,-0.0223769676,-0.2515200973,-0.1795809865,0.002502559,0.1760140508,0.0290174522,0.3748070002,-0.0623117872,-0.042919822,0.0252822451,0.0009238463,-0.1133537367,0.1886116117,-0.1212607548,0.3817118108,0.3555004597,-0.2482015938,-0.3381769955,0.217444092,0.1194716319,0.0722379684,-0.024797108,0.4353022575,0.1737557352,0.0868726224,-0.4525137544,0.1167607233,0.1905513108,-0.1261230111,0.402777195,-0.1667664349,0.3428405225,0.5249500871,0.3461026847,0.3309946954,-0.2420707941,-0.0178682655,0.1920123249,0.168471083,-0.224376291,0.1829430163,0.2487548292,0.0333352722,-0.0761647075,0.3599678576,0.1959193498,-0.0136882225,-0.0975944847,-0.0059233671,0.5347859263,-0.5172249675,0.2831372917,0.4529380798,-0.0578202754,0.1300518662,0.395442903,0.1173456088,0.0406698771,0.3292238414,0.1746831834,0.4738540649,0.2645662427,0.024352314,0.3100903928,0.1382420361,-0.0397077091,0.5465513468,-0.0578761473,-0.0461779535,0.1481501162,-0.0553207062,0.1089119837,-0.3581901193,-0.0374864973,0.2535924911,-0.2680290937,0.2417783886,-0.2399579883,-0.1268057525,0.1208409443,-0.2076497823,-0.156116724,-0.1841452122,0.4952987731,0.1778952926,0.0020787844,-0.0086865043,0.2054046243,-0.15188317,0.0813432261,-0.1358140111,-0.0478563718,0.1461289525,0.0562456325,0.0846262127,0.6115331054,0.2377922088,0.3204872906,-0.299259007,0.1391634643,-0.2391991168,-0.2748883665,0.0035142724,0.4170590043,-0.2262089998,-0.1757845283,0.287096858,0.1712303609,-0.0557433031,0.3106923103,0.0455809943,0.240136534,-0.294798255,0.5732987523,0.416983068,-0.0019928885,-0.0044025895,0.1076347902,-0.2917104363,-0.0711558536,0.5056442022,-0.5531600118,-0.0742896944,0.005824938,0.0713660643,0.0575755462,0.1099347696,0.1549496949,0.0550570898,0.0522361174,-0.0616669022,-0.1931787133,0.0594414249,0.1810075045,-0.0620460324,-0.317468524,-0.0042772386,0.1144714877,0.386926353,0.0939161628,-0.290792495,-0.2940655947,0.3518124819,-0.1116025075,0.1389463991,0.1381022781,0.3081413805,0.0559721328,-0.2472725511,0.0599425361,0.2529646754,-0.0038804892,0.0938820317,0.3810404539,0.1513930112,0.1739981472,0.158820793,0.2628678083,0.4578672647,-0.080337368,0.2029063106,0.3254690468,0.0196896177,-0.2989490628,0.4337381423,0.1688085198,-0.0758024231,-0.4323500395,-0.2001509964,0.0462303013,-0.156706214,-0.0028267978,0.0878175125,-0.7870503664,-0.1079171002,-0.2266780585,-0.0185904484,0.0197736062,0.2365136445,0.0901064724,0.1004438922,-0.1907736063,-0.2095700204,0.1808404773,-0.1625893414,-0.3482762575,-0.203956157,-0.2373826355,-0.3200111985,0.0670422316,-0.3271470666,-0.0818442106,0.3845259547,-0.1479876488,0.33480528,0.2760920227,0.0717193857,-0.0876514539,-0.2140066475,0.5781516433,-0.1806405485,-0.0973056927,-0.2706850767,-0.0791067779]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"We can be pretty flexible and not impose any constraints for transforms.\r\n\r\nMoreover, this library is designed to support datasets bigger than memory. The datasets are loaded from the disk via memory mapping, without filling up RAM. Even processing functions like `map` work in a batched fashion to not fill up your RAM. So this shouldn't be an issue","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":59,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nWe can be pretty flexible and not impose any constraints for transforms.\r\n\r\nMoreover, this library is designed to support datasets bigger than memory. The datasets are loaded from the disk via memory mapping, without filling up RAM. Even processing functions like `map` work in a batched fashion to not fill up your RAM. So this shouldn't be an issue","embeddings":[-0.540179193,-0.0189927835,-0.1440081596,0.1232815832,0.4372493029,-0.2170054168,0.245312959,0.1936611384,0.2192274332,0.130491212,0.0615159795,0.187767759,-0.2603258789,0.07238736,0.0481922179,-0.0115978345,-0.0782924369,0.2525763512,-0.5557228327,-0.0573627502,-0.3713938892,-0.1573706418,0.0135659892,-0.1440851539,-0.0328940004,-0.2428776622,0.0715212226,-0.01376657,-0.1378084719,0.0672599301,0.3069247305,0.1395094246,-0.1308079958,0.3437947035,-0.0000978967,-0.0323359221,0.0386952013,-0.1293641925,0.0463968217,0.1244557947,-0.1310040504,-0.1104852259,-0.1095086113,-0.3612022996,-0.2468073219,0.0423793793,0.0283964686,-0.3146081269,0.2295787483,0.1305739284,0.354373306,-0.1699274927,-0.1962252706,-0.053974133,0.2161227465,0.1079380736,-0.0890014172,-0.1663483828,0.2665728331,0.2456260324,-0.312132746,0.2356686741,-0.0461950712,0.0065835495,0.5308229327,-0.1477865726,0.0746076852,-0.1883723289,-0.0439591147,0.2111916542,0.3024063706,-0.3585546911,-0.1495787352,-0.1898489743,-0.0642794892,0.1807121933,0.1255578697,0.0910863951,-0.1162973195,0.1236252263,-0.2932367325,-0.2158868462,-0.185869202,0.0246883072,-0.3864739239,0.0843880698,0.0810551643,-0.0198896565,0.2320849746,-0.1219940484,0.0640956834,0.233398959,0.0315565765,0.0362053812,-0.107584089,-0.428006649,-0.3164810836,-0.0104847215,0.140488416,-0.0769903585,0.0305277966,0.2146950513,0.0684990287,0.0404782556,0.0683705509,0.2132992595,-0.3577131033,0.0629388914,0.3471935987,-0.2959295809,-0.0947281048,0.1421936303,0.1777654141,-0.1919908226,-0.0374680385,-0.187868312,-0.1458259523,-0.1118825898,0.0392557308,-0.1497556716,0.2410176247,0.2816529274,-0.0548955686,0.3797278702,0.0164872911,0.122593388,0.4158708155,-0.0704298615,-0.1016627997,-0.0267893784,-0.1234514937,0.1682244688,-0.2392304242,-0.1178169474,0.1755443364,0.2547565997,-0.0090133147,-0.1393259615,-0.0014683313,0.4502189457,0.1454497874,-0.1632993966,0.3711447716,0.3215205669,-0.3073906898,-0.2312166244,-0.1824297607,0.1355635822,-0.224871546,0.2574270666,-0.0100257434,-0.0601822175,0.2163721919,0.3365978003,-0.0965414643,0.0889814273,-0.0164763294,0.1412249804,0.0578206107,-0.2100098431,0.0223715231,-0.1844257265,-0.1561108828,-0.1307398826,0.0042210277,0.0234843884,-0.3599258363,-0.196546793,-0.0161801074,0.0299439281,0.0393635854,0.2355713397,-0.2584232688,-0.0666704699,-0.1518525928,-0.0959847718,0.5436256528,-0.0873637274,-0.5102764964,0.1716724485,-0.307172209,-0.2066991627,0.0392220244,0.5570349097,0.3030121326,-0.0606620498,-0.2779816985,0.4389851987,-0.0120839952,0.2393550426,-0.1522180885,-0.3807231188,0.2241461128,0.0999747664,0.4358043671,0.069594577,-0.004208317,0.4479361176,0.0191990659,-0.3247076273,0.2287278026,0.0903669521,0.2029692382,-0.1758003682,0.0514797978,-0.3881592751,-0.029957436,-0.0194903668,-0.1231638268,0.1746489406,0.3121495247,-0.0633318052,0.093969591,-0.2406131178,0.0662423149,0.052394297,0.3378172517,-0.0506487973,0.2560933828,-0.3789431751,0.0967584029,0.3695116937,-0.0359232277,0.0900141522,-0.4097872674,0.0930517539,0.0395775288,0.0411274321,0.0174408313,-0.0858047009,-0.159468174,-0.1214016229,0.0888800398,0.0884446204,0.1423340589,-0.0831505358,-0.225153029,0.3935472965,0.2967004776,-0.0864641517,0.4523482323,0.2250223905,0.1164681241,-0.054058414,-0.2326427102,0.3532539606,0.0675226301,0.1176048443,0.0894607902,-0.0354507528,0.3178417683,-0.2105885148,-0.0916877463,-0.2391735613,-0.0803821012,0.1462734789,0.2722612321,0.241806075,-0.1434707642,-0.1761649847,0.2574052215,0.095572263,0.2356152534,0.2567685843,-0.1985844076,0.168779552,-0.0230590459,0.3929041922,0.4536697268,0.3772838116,-0.0321456678,0.0570631735,-0.0334798992,-0.128506422,0.1743486375,-0.0611710101,0.2145086229,-0.0075750812,0.0224812124,-0.0848525614,-0.3140361011,-0.2202801704,0.0788237676,-0.0040893168,-0.1285701245,-0.0825884491,-0.2104412764,-0.2082371861,-0.1002784669,0.1774028689,-0.1065199599,0.1337847114,0.1469562948,0.0870128423,-0.1986066252,0.1812264919,0.0617287904,0.3059969544,0.1493957341,-0.2343177944,-0.010339994,0.0442344993,-0.0209342781,0.2839320004,0.0812600628,0.0301540904,0.3573117256,-0.0325139053,0.0159158316,-0.1075667888,-0.4170719087,0.021991238,0.0508672036,0.1327099055,0.2242882103,0.1550285071,0.1264584213,0.0545450002,0.0223775897,-0.2790865898,-0.3149189949,-0.0249228776,-0.0592014007,0.1274024546,-0.2830105424,-0.2389216274,-0.0929963067,-0.5055266619,0.2309706211,0.1451750249,0.0573420674,0.0028792156,0.2555831075,-0.175446555,0.1253508627,0.1729648709,-0.2668440342,0.023113925,0.4328384399,-0.0556156076,-0.2987732887,-0.1864276677,-0.2419093251,0.216374889,0.5209243298,-0.4316136837,-0.0116054965,-0.2875873744,-0.0104507301,0.0108654983,-0.0455288813,0.5706766844,0.0471077152,-0.2182017267,-0.0778356642,-0.1106931269,0.0597552098,0.150663957,0.016526347,-0.0133856088,0.2726939917,0.0209518801,0.2372093052,-0.4598437548,-0.3791241348,-0.0370535441,0.1754178852,0.132338658,-0.0176740009,-0.1120066792,0.1540552825,0.1146941185,-0.1631689817,0.1865586042,-0.0414883643,-0.4385227859,-0.1449545324,-0.1328935921,0.0028604434,-0.2513259351,0.3997885585,0.0662711263,-0.0650797635,-0.0767925978,0.1190419495,-0.1788247824,-0.0431990102,0.1782182306,0.1142590046,-0.0087259654,-0.073540695,-0.0310928188,0.1208601296,-0.1072629839,0.1333228648,-0.1606767178,0.0035681303,-0.0728182942,-0.2996204793,-0.1183586419,-0.1653675884,0.5815544724,0.0500494838,0.0715659335,0.1510835588,-0.3680042922,-0.3692417443,0.1909379363,-0.2953854501,-0.1915870309,0.0202095527,0.2334669679,-0.1936124563,-0.0312032495,-0.0504270345,0.1912976056,-0.3580103219,-0.1948178262,-0.0154852578,-0.1780780554,-0.3871999383,0.1539223045,0.1001420841,-0.0908321664,0.0161030125,-0.2437742501,-0.325971663,-0.3354439139,0.2459117323,0.0127381794,0.0944812447,-0.1280446649,0.1640539467,-0.1807055175,0.3554745317,0.2075027674,0.066977866,0.1276006103,-0.0331705846,0.1978534311,0.0043810047,0.0737058148,0.1707678586,0.0512657985,-0.0757459402,-0.1046156138,0.2245442122,-0.5025944114,0.2269370258,-0.1176156253,-0.0372035578,-0.1397033632,-0.3046917915,0.3326459229,0.1389250159,-0.2194042355,0.0952933878,-0.1129676402,-0.5864390731,0.2940364778,0.6617208719,0.9825024009,-0.0968456566,0.1830242127,0.0236184876,-0.0951498225,0.2489520758,-0.3341436684,0.1913649589,-0.2713858187,-0.3456095457,0.0392465293,-0.072635524,0.0903847888,0.2218271494,-0.4011258483,0.0230280273,0.0346582979,-0.2270578146,-0.1472181976,0.226057142,0.0816654339,-0.5389626026,-0.1904724985,0.3400667906,0.0100833839,-0.0979929566,-0.0679818466,-0.1252209097,-0.0838793293,0.0722063333,-0.0511192977,-0.1457973421,-0.3910182416,-0.026934173,-0.4349460602,-0.0961698219,0.2528386712,0.1495362073,0.0102971289,0.4106075764,-0.0813753679,0.1445856541,-0.1695801318,0.266882211,-0.0444293544,-0.0862690508,0.4258618653,-0.1075083017,-0.0176085774,-0.0891338438,-0.0558735505,-0.266366899,0.2118800879,-0.133044973,0.1236126274,-0.33392331,-0.0306276605,0.0376899242,-0.1858448386,0.0236257482,0.2972595096,0.0953723863,-0.1094942614,0.4886617661,0.3742894828,-0.267143935,-0.2493671626,0.3068590462,0.4510116279,0.1679682732,0.0681990907,-0.2079339474,-0.3289976418,-0.461337626,-0.0386336632,0.147553429,-0.1566038132,0.3310774863,0.0176412035,-0.4553512931,-0.0362949558,-0.0112304939,0.2609969974,0.1042575762,-0.2485146523,-0.1195608154,-0.57578969,0.116673775,0.0600397885,0.3340652883,-0.0717348307,-0.0291897692,0.0409154072,0.1309844255,-0.5081634521,-0.0618289597,-0.1701747626,0.2392061949,-0.090618819,0.0401869565,-0.199549973,0.1164791882,0.1786226332,0.1987207383,-0.3415727019,-0.3561058342,0.0265030917,0.0729617029,-0.0851012692,0.0381086469,-0.0057898629,-0.1136713251,-0.1005753279,-0.1748030633,0.2322603762,0.0757041201,-0.1358757615,0.161711812,0.3516884148,-0.1904283315,-0.1061378941,0.2564583123,-0.0229883604,0.1287886947,-0.0762278289,0.224842757,0.1289897412,-0.0755504593,-0.0370385014,0.1613633931,0.1092058867,0.0161625799,0.2429385632,0.0872149542,0.0296292845,0.6459552646,0.2213807553,0.1402513832,-0.0733435079,-0.1104496941,0.211389631,0.4208187461,-0.3107750416,-0.1477650106,0.1610415876,-0.1849022806,0.298750937,0.3614836931,0.0597651489,-0.10970781,0.1134000197,0.0597116575,0.2657986283,-0.2369620353,0.151523754,0.291359365,-0.1258794218,-0.1349882931,0.3201216459,0.2068947256,0.2180112898,0.4387917817,0.0376220196,0.3958040178,0.0772475377,0.0723351389,0.0071840049,-0.0593903661,0.084039107,0.5973978043,-0.013415359,0.1384305656,0.1193290129,-0.077409029,0.1103999391,-0.2957307398,-0.1641556025,-0.0149182826,-0.41329965,0.2373102605,0.0107652796,-0.1997172534,-0.1087908074,0.0860128552,-0.0846301094,-0.28409338,0.5777483582,0.0436372459,0.0698707476,0.2187255174,0.2641711235,-0.128750816,-0.095583193,-0.2554258704,0.0436864682,0.1652055383,0.0027947912,0.2931783795,0.3171942234,0.0247330535,0.2477005273,-0.1783778369,0.1587840617,-0.0223436002,-0.1473961174,-0.4861269593,0.1307812184,0.0130430032,0.1222265512,0.3329808712,0.3580688238,-0.2723877728,0.0573449843,0.0822003409,-0.0306330435,-0.098712869,0.1574561894,0.3358223736,-0.1905295104,-0.1132157966,-0.1025379598,-0.1873035729,-0.0711828321,0.2527501881,-0.6251538992,0.173326537,0.1061771438,0.1627587527,0.0676863343,0.524692595,-0.1559083462,-0.0697748959,-0.0372016765,-0.2199507207,-0.3036782146,0.1621911675,0.0176517423,0.0907671973,-0.2603814006,0.1454750448,0.1455657929,0.0833837986,-0.0279174391,-0.2130619884,-0.1423188895,0.0992488042,-0.1822213829,0.0495417938,0.0438819528,0.1255911589,0.2152688056,-0.3644394875,0.0570651069,0.012062422,0.3014870286,-0.0124862101,0.2467904836,-0.0544573851,0.3218396902,0.4679386616,0.1505186707,0.4137197733,-0.1625350416,0.1229261681,0.1877688915,0.0470766798,-0.3208537102,-0.0584467128,0.3024864793,0.0949110612,-0.1568127275,0.0238699615,0.0055291206,0.0495616496,0.0280380193,0.1776486337,-0.3475266993,0.1490118206,-0.0308271851,-0.0213894378,-0.0929821432,0.2472129911,0.0079181036,-0.2108102888,-0.1649531871,-0.4143649936,0.255577445,0.0093731256,-0.2656223476,-0.1167655662,0.0128656691,-0.1665269732,0.0643491894,-0.4184448719,0.0568231791,0.1724010855,-0.1445022076,0.0484218635,0.0856433958,-0.1510891914,-0.2606152594,-0.3129924536,0.4690489769,-0.033630643,-0.151423946,-0.2540898621,-0.2887403071]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"Ok thanks.\n\nBut, Dataset has various flavors.\nIn current design of Dataset,\n how the serialization on disk is done (?)\n\n\nThe main issue is serialization \nof newdataset= Transform(Dataset)\n (ie thats why am referring to Out Of memory dataset\u2026):\n\n Should be part of Transform or part of dataset ?\n\n\n\n\nMaybe, not, since the output is aimed to feed model in memory (?)\n\n\n\n\n\n\n\n\n> On Jul 7, 2021, at 18:04, Quentin Lhoest ***@***.***> wrote:\n> \n> \ufeff\n> We can be pretty flexible and not impose any constraints for transforms.\n> \n> Moreover, this library is designed to support datasets bigger than memory. The datasets are loaded from the disk via memory mapping, without filling up RAM. Even processing functions like map work in a batched fashion to not fill up your RAM. So this shouldn't be an issue\n> \n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":162,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nOk thanks.\n\nBut, Dataset has various flavors.\nIn current design of Dataset,\n how the serialization on disk is done (?)\n\n\nThe main issue is serialization \nof newdataset= Transform(Dataset)\n (ie thats why am referring to Out Of memory dataset\u2026):\n\n Should be part of Transform or part of dataset ?\n\n\n\n\nMaybe, not, since the output is aimed to feed model in memory (?)\n\n\n\n\n\n\n\n\n> On Jul 7, 2021, at 18:04, Quentin Lhoest ***@***.***> wrote:\n> \n> \ufeff\n> We can be pretty flexible and not impose any constraints for transforms.\n> \n> Moreover, this library is designed to support datasets bigger than memory. The datasets are loaded from the disk via memory mapping, without filling up RAM. Even processing functions like map work in a batched fashion to not fill up your RAM. So this shouldn't be an issue\n> \n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n","embeddings":[-0.3028283417,-0.0752469152,-0.05205011,0.2568613887,0.4963202775,0.0333011746,0.3464813828,0.1359570324,0.1084691659,-0.0489086248,0.1433253735,0.1807944179,-0.2497085929,0.0995197967,0.1433494985,0.0294342879,-0.0367588326,0.2082752585,-0.6057528853,-0.110377945,-0.2944284678,-0.1542579234,-0.105689235,-0.1136090979,-0.0622637235,-0.1996254027,-0.1581871808,-0.1262419373,-0.2569142282,-0.075012356,0.420137316,0.0989703834,0.1109477431,0.5628956556,-0.0001049531,-0.1980094761,0.0629667714,-0.1417288631,-0.0416416563,-0.1271471083,-0.110590592,-0.1241359413,-0.0914635956,-0.3788447976,-0.2515878081,-0.012632167,0.042666167,-0.1632636189,0.2673305273,0.1891238987,0.2877258658,-0.2571978569,0.0613154992,-0.1159944683,-0.0906342044,0.4980603456,-0.2132179588,-0.1317691058,0.1639421731,0.2951180041,-0.1768623143,0.1194915324,0.0316155441,-0.0623987168,0.6949888468,-0.1144897491,0.1647530645,-0.3416478038,-0.047490526,0.1770274639,0.6858400702,-0.4208053648,-0.2926061153,-0.1978127658,0.0413535796,0.1125580594,0.0736283138,0.2918520272,-0.1023894027,0.4228276312,-0.290635854,-0.1578572839,-0.212265715,-0.0273612179,-0.3538392782,-0.0272985063,-0.0485696979,0.0704357848,-0.0280704889,-0.0896413997,0.1152463704,0.0687420294,0.085941568,0.0053374721,-0.0471011549,-0.4889363945,-0.2492946684,-0.1057617217,0.1770293117,0.0206169728,-0.1055535376,-0.0825049132,0.0663288385,-0.0880444124,0.17521891,0.3654631376,-0.1550439894,0.1602373272,0.0713833496,-0.3452311158,-0.0499749109,0.0044703535,-0.0236585438,0.0380209535,0.2002761662,-0.0101096481,-0.0094334157,-0.0741236731,0.1005777717,-0.2199292779,0.1885883063,0.1436040848,-0.1644640416,0.256283164,0.1343160272,0.2364140302,0.2878935933,-0.0608223677,0.0728546306,-0.1835631281,-0.0920746624,0.2097067535,-0.3946242034,-0.0316818357,0.0614863746,0.262503773,-0.098427169,-0.1631225944,-0.1027049199,0.293238163,0.0450083837,-0.2885059714,0.37007761,0.2936200202,-0.5175172687,0.0034238908,-0.0082310745,0.090826489,-0.2388359904,0.3887875378,-0.0808293521,-0.0568780564,0.1232487559,0.2654306293,-0.0568023287,0.1304450184,-0.2425501645,0.2334434688,0.2355148196,-0.1140492633,0.0523458868,-0.2796918154,-0.0145570571,-0.2248253822,0.0019177148,0.1308251768,-0.550594151,-0.1371417493,-0.0929638222,0.0054792524,0.2257186919,0.2538051009,-0.2389266789,0.2112432718,-0.2781811357,-0.1598565429,0.5308234692,-0.0717295632,-0.3687420785,0.2272962779,-0.3469464183,-0.2316666543,0.1396826655,0.3911378384,0.3910191357,-0.0739194453,-0.3169713914,0.3255683184,-0.1547599733,0.0172961466,-0.1826123446,-0.2983471453,0.4146893024,0.0043986351,0.3740959764,0.3172221482,-0.0899922773,0.5544752479,0.0361762792,-0.1642745286,0.2736387849,0.0395088308,0.3699331284,-0.2187348157,0.0669548362,-0.3890495896,-0.1323824525,-0.0164715406,0.0716964751,0.2384104282,0.1521542966,0.0425674058,-0.0351381898,-0.1884233207,0.1005936936,-0.1441040337,0.2575828731,-0.0682228804,0.2908599973,-0.3608987927,-0.1201941073,0.3291369975,0.0169595778,0.0997303575,-0.4722500443,0.135290727,-0.1037477329,0.0637206659,-0.1334819943,-0.1195375398,-0.1471011341,-0.2059051692,0.027681306,0.045600161,0.1006959081,-0.072777003,-0.1369668692,0.4407846332,0.3833371401,-0.2213459909,0.4104960263,0.1753512174,0.197182104,-0.1581744701,-0.1563640684,0.3226000369,0.0591880493,0.0539624803,-0.0045891628,-0.0444103889,0.2749353051,-0.1846638918,-0.1495006084,-0.2213106602,-0.093039088,0.1528907418,0.2064580917,0.2672087252,-0.0180213116,-0.0872312263,0.2759074867,-0.0993304923,0.1754228622,0.2399032116,-0.2363844812,0.0967985913,-0.1155927703,0.3368810117,0.4515927434,0.2688371539,-0.0172095709,0.0994065106,0.2248305678,-0.1374188513,0.2106560916,-0.010123007,0.3061108887,0.0928651989,-0.0435660295,0.0148412464,-0.1426321864,-0.2731765807,0.2011459768,0.1052608714,-0.0868094042,-0.0827878341,-0.0791172609,-0.0586189143,-0.3210164011,0.1139508635,0.0012527817,0.1354949921,-0.0451803915,0.0133053074,0.0678712949,0.1548039466,-0.0258904975,0.2285887897,0.1403194815,-0.022201566,-0.0016004929,0.0719159022,-0.012956596,0.1860071272,0.0082577905,-0.0523206517,0.3632893264,0.0237425417,0.0723289773,-0.0993584543,-0.3952199817,0.100798808,-0.0392054431,0.1362577081,0.2487015575,0.1583067477,0.0638481453,0.073958613,0.1003383845,-0.1901111752,-0.2742712498,0.0383124985,-0.0836613625,0.1695398688,-0.1737304032,-0.3793549538,-0.2794141173,-0.3382386863,0.1888036579,0.0238642674,0.0039382828,0.1018079519,0.0851486549,-0.1637770087,0.1167969406,0.135999456,-0.2013310045,0.0089232298,0.3245040178,0.0960316285,-0.3421610594,-0.1381985247,-0.154382363,0.01020806,0.3889779449,-0.3169546425,-0.0313842595,-0.2151265293,-0.028546948,-0.0322704613,-0.1737960577,0.625001967,-0.0079119746,-0.1444297731,-0.0402277,-0.1125065014,-0.0756177604,0.1865358055,0.127004683,0.05392823,0.4150843918,0.1548915505,0.1784219891,-0.2313495427,-0.3481871486,-0.1597625017,0.2163899988,0.0208490193,-0.0015360153,-0.2815902829,0.2986260056,0.086776264,-0.5257769823,0.0635800883,-0.1433752477,-0.3126085997,-0.0938735828,0.0416002832,-0.0591737591,-0.1023656875,0.3085528016,-0.0628911182,-0.001626252,-0.0956037566,0.0832504481,-0.0903696939,-0.0007783305,0.1365067959,0.1792053282,0.0448025167,-0.0823966786,-0.2790713608,0.0333365537,-0.1754755378,-0.0688802972,-0.2427608222,0.1018722355,-0.0368377045,-0.2086670548,-0.0448292084,-0.0471707322,0.8162045479,0.1139855534,0.0306032952,0.0867245495,-0.3628522158,-0.322917968,0.0645702481,-0.3020896912,-0.2120197564,-0.0534544662,0.2099127471,-0.2198083997,-0.0817434117,-0.0645168945,0.3001189232,-0.3440364599,-0.308396697,0.087712869,-0.2214542031,-0.3976357579,0.0095789125,0.184038341,-0.0245554708,0.0119086569,-0.1444052011,-0.2201026529,-0.1155397296,0.2485006005,-0.1722968668,0.0177520104,-0.1120066494,0.1586957574,-0.1353849322,0.3628400862,0.4106985033,0.049474068,0.1806004643,-0.2303600907,0.211571753,-0.0082221534,0.128854394,0.0311821438,0.0427934788,-0.2118957639,0.0750582069,0.1492679268,-0.6959347725,0.0593150817,-0.1067282706,-0.1500359178,-0.2943808436,-0.4452714324,0.2117899805,0.12461707,-0.1842980981,-0.0401221216,-0.0634937733,-0.6139559746,0.3885996938,0.6128232479,1.1169228554,-0.1842900515,0.1669409573,-0.0168989729,-0.0058910968,0.3774906993,-0.2371506095,0.1998218596,-0.3550825715,-0.3183297515,-0.1002992168,-0.1418188512,0.1677480191,0.3499812782,-0.4512499273,0.039248839,-0.0306312628,-0.0302703027,-0.0552794635,0.2415483296,0.1953426301,-0.537481606,-0.1339253932,0.2882777154,-0.1208229437,-0.034762986,-0.0453361683,-0.1534027755,-0.0932938978,0.1907481998,-0.1508702636,-0.067285046,-0.3210911453,0.0556668714,-0.4634639323,-0.1997614503,0.3177556098,0.2684037983,0.1501512975,0.2492229789,-0.022403881,0.0255250689,-0.2658393085,0.2352416515,-0.0859216675,-0.0709661022,0.5328398347,-0.0623401888,-0.0518441796,-0.0348497145,-0.0688450038,-0.1706681252,0.1275119632,-0.2459351569,0.134200871,-0.503185451,-0.1402988732,0.0312630609,-0.1325476468,0.0143527035,0.2422595769,0.274544239,-0.1834605336,0.4280769527,0.2632658482,-0.2569596767,-0.204016,0.2488785684,0.5733271837,0.1230105907,0.2374245971,-0.1156910509,-0.3798985779,-0.4102578163,0.1769221574,0.293238759,-0.2035410106,0.3463568687,0.0076653818,-0.1694399714,-0.0017319446,-0.057373751,0.3999804854,0.0833484977,-0.3539372087,-0.1449834555,-0.5922728181,0.0875967294,0.0081702685,0.1891020387,-0.0583878495,0.0969300941,0.1270013452,0.055208236,-0.4772725105,-0.1786121726,-0.1551102698,0.2088940591,-0.0917118192,-0.0565584451,-0.2293162793,0.0196996816,0.1121013686,0.0307326019,-0.2857632339,-0.2874863446,0.1481781751,0.1052133292,0.1582019329,0.1143800914,-0.1988067776,-0.1584184319,-0.1051569805,-0.1556755751,0.1166793332,0.1672087461,-0.026931392,0.3737908006,0.3921233714,-0.1973208189,-0.3632752895,0.0381701328,0.0736835971,0.133545965,-0.0087955929,0.2220734805,0.2070567161,0.0143724289,-0.5705882907,0.2322947085,0.1467027664,0.0700938925,0.3555316925,-0.0842887089,0.2152724862,0.7550534606,0.2347373515,0.0145714115,-0.075227201,-0.0008534527,0.1678280383,0.3209893405,-0.209294647,0.0889710709,0.1455574483,-0.2011818439,0.0874312222,0.2652133703,-0.0376698896,0.0738559887,-0.099381797,-0.021686729,0.6841118336,-0.3670733571,0.1809158474,0.4650237858,-0.0427682921,0.01620185,0.4952476025,0.2627806962,0.1423285753,0.454531312,0.1572256684,0.379499048,0.2466205955,0.0828615725,0.204173252,0.0437722094,0.0797859505,0.6636738181,0.037964046,-0.0697575733,0.0407792963,-0.1790005267,-0.0601510406,-0.4161751866,-0.1353780627,-0.0135044875,-0.4005985558,0.2915465534,-0.112486735,-0.0684107095,0.0578974187,-0.0972615182,-0.1537034065,-0.2673508227,0.4747747183,0.0462129004,-0.1205843538,0.1951797158,0.1547541469,-0.0755723715,0.0101365773,-0.2884206474,-0.0650607497,0.1735550463,0.0605660267,0.0986792892,0.5237454772,0.0263672229,0.2559469342,-0.0822419152,0.062913388,-0.0508801378,-0.2541757226,-0.2453350723,0.3532659113,-0.1296031326,0.1875945181,0.1301017255,0.2976488769,-0.1622902155,0.165485993,0.2738925517,0.0677082241,-0.1761024445,0.2827882767,0.1829778701,-0.2444881201,0.017565215,0.0409468487,-0.3012639284,0.1206568331,0.461648643,-0.6105459332,0.1043522879,0.1044352725,0.1269515902,0.0615119226,0.2328920066,0.1164094508,-0.1334686726,-0.0879255608,-0.0016164932,-0.2138328999,0.2222235352,0.0722440705,0.0310865492,-0.2169958204,0.2303780913,0.2244080752,0.1264851838,-0.2505669594,-0.3092609048,-0.2979168892,0.2071534693,-0.0954661071,-0.0076386849,0.0572799854,0.3779459298,0.0538814031,-0.2726292014,0.1260622889,0.1939679533,0.1159630269,-0.0036792478,0.2997324169,0.0336135365,0.245690912,0.3860705495,0.181909591,0.4176517129,-0.0400889888,-0.008874136,0.2440115511,-0.0818879232,-0.2122546881,0.2055052668,0.2114602923,0.1955604851,-0.2826955616,0.023086546,-0.065523535,-0.1057370156,-0.1479047537,-0.0403799564,-0.6305205822,0.0302140545,-0.126418218,-0.2030179352,0.0873136446,0.201850161,0.0949390978,-0.071717307,-0.1438317001,-0.3844311535,0.2091503292,-0.0003152927,-0.3548848331,-0.1294166744,-0.1019817814,-0.1853729039,0.0868706778,-0.5133615732,0.1284306794,0.3005841076,-0.0106591349,0.0992452577,0.1067641973,0.0080752466,-0.1457769722,-0.2062027603,0.7358550429,-0.2172154188,-0.1924630255,-0.26944682,-0.1281165034]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2596","title":"Transformer Class on dataset","comments":"I'm not sure I understand, could you elaborate a bit more please ?\r\n\r\nEach dataset is a wrapper of a PyArrow Table that contains all the data. The table is loaded from an arrow file on the disk.\r\nWe have an ArrowWriter and ArrowReader class to write\/read arrow tables on disk or in in-memory buffers.","body":"Just wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":55,"text":"Transformer Class on dataset\nJust wondering if you have intenttion to create\r\n\r\nTransformerClass :\r\n dataset --> dataset\r\n\r\nand make determnistic transformation (ie not fit).\r\n\r\n\r\n\r\n\r\n\r\n\nI'm not sure I understand, could you elaborate a bit more please ?\r\n\r\nEach dataset is a wrapper of a PyArrow Table that contains all the data. The table is loaded from an arrow file on the disk.\r\nWe have an ArrowWriter and ArrowReader class to write\/read arrow tables on disk or in in-memory buffers.","embeddings":[-0.3604480028,-0.056124229,0.0116531681,0.2704983056,0.3946083188,-0.0971958861,0.3508632779,0.0482230633,-0.100028649,-0.1611097008,0.1317931861,0.3815632463,-0.3154996634,-0.0833913833,0.2542794347,-0.2325899005,0.0182949621,0.4430117309,-0.3059166968,-0.1537890881,-0.2024315447,-0.2975797653,-0.1411009431,0.1563541144,-0.108661823,-0.1525422484,-0.1539614797,-0.3630880117,-0.3081549406,-0.1566582322,0.6911965609,0.0511289686,0.0235325266,0.5484084487,-0.0001298558,-0.0235750005,0.0727151707,-0.1433799416,0.1298894733,-0.1923117191,-0.2210621834,-0.033829201,0.1811142266,-0.3892242312,-0.3804796934,-0.2979854345,-0.0040941914,0.051528912,0.5385863185,0.1942201555,0.0866092443,-0.3026257753,0.0551640689,0.1068512052,0.2160902172,0.3861628771,-0.2014035583,0.0175133739,0.1515752673,0.2822321057,-0.15152888,-0.1318753958,-0.1761910468,-0.0272739865,0.8690388799,0.0244412255,0.1509451866,-0.3632725477,-0.2043594569,0.2908260524,0.7759717107,-0.5236781836,-0.1550633758,-0.0679519549,0.1838050336,0.1550900936,0.0083903447,0.2853395939,-0.1065251231,0.2195320576,-0.1110390648,-0.2602996826,-0.5028033257,0.1028025374,-0.521471858,0.1457862109,-0.0008760057,0.1302547306,-0.0523233823,0.0741430968,0.420173645,0.0259791929,0.0769913048,0.2010128349,-0.1059709489,-0.4257588089,-0.4286718965,0.1216747537,-0.0518270582,0.0706445873,-0.0520828366,-0.2472423762,0.1668996215,0.0298987795,-0.0434882902,0.2679726779,-0.0275975652,0.3352452517,0.0658622608,-0.3162097633,-0.1533956379,0.1382795125,-0.0609802641,-0.2558608055,0.2248512506,0.0547069982,0.145677194,-0.0902904868,0.1242999882,-0.2215955853,-0.0831000656,-0.1453952938,-0.2076639533,0.1542184204,0.2172251642,0.0267329048,0.2401559651,0.2879996896,0.1310458928,-0.2053761184,0.080245629,0.2888303101,-0.4228098392,-0.1289700866,0.0684534237,0.169024229,-0.3287172318,0.0584953316,-0.2053339928,0.2638606131,0.0007679323,-0.321113646,0.251339674,0.3396914303,-0.4704187214,-0.1540033817,0.1538810432,-0.1303744316,-0.2196123451,0.6994740963,-0.1940143704,-0.0101558361,0.0722472891,0.0758103952,-0.0002412482,0.0676918253,-0.2430974692,0.166962117,0.2110141367,-0.1898816079,0.3288912773,-0.3618783653,0.1388187855,-0.2993890345,-0.088786222,0.1197607517,-0.7907062769,-0.2019080669,-0.014581684,-0.0888852552,0.184751302,0.2106100917,-0.3140335381,0.3180654943,-0.119111225,-0.0558806583,0.7501407862,0.0790741518,-0.3657938242,0.1410096586,-0.4655134976,-0.3085185885,0.2650138736,0.483253926,0.2425493002,-0.0904032215,-0.5927631259,0.1744460166,0.0494939573,-0.1906802803,0.0165515766,-0.1602030545,0.4672031403,0.0275996495,0.2890078723,0.2277741879,0.1661888808,0.1460767388,-0.0301545355,-0.3529479206,0.2881473005,0.0217043664,0.3800839484,-0.0008921814,0.0429793745,-0.2651746571,-0.0465587564,-0.116854921,-0.214377284,0.1626737565,0.1254605502,-0.085994035,0.0401732624,-0.218502894,-0.2053408772,-0.1495812833,0.0275096651,-0.376270473,0.3453819156,-0.3231988549,-0.204575941,0.3623326123,0.032749325,0.2075287998,-0.4299808443,0.4497601688,-0.0851841941,0.0589445829,-0.2049827874,-0.1223996505,-0.0487961695,-0.1631944478,-0.0185094532,-0.0753683671,0.0226385035,-0.2253183275,-0.0744982287,0.3795029819,0.2443033308,-0.6910700798,0.6359952688,-0.0714266971,0.1448139697,-0.2594786584,-0.1614321768,0.4715094864,-0.0678142235,0.2263817489,-0.2780797184,-0.0126201846,0.1119962484,-0.3335904181,-0.1063085422,0.1066768095,-0.0678891838,0.1117890179,0.2380772233,0.2574709952,0.0649888888,-0.186487183,0.1255374253,-0.1597726792,0.2351240218,0.4662559032,-0.2535711229,0.2055148482,-0.1791690141,0.3958261311,0.6401337385,0.1718729734,-0.3124704659,0.2381339371,0.0387583189,-0.0405567028,0.2972995937,-0.0722031221,0.3564677536,0.0087075802,-0.1942992806,0.1325372458,-0.0467172563,-0.2978574038,0.0913365409,0.1649499834,-0.2284638584,0.0794563442,-0.0708314478,-0.1709196419,-0.4144055843,-0.0693010539,0.0078859618,0.2818384171,-0.0862996578,-0.1729284525,0.3115230799,0.0573587082,-0.1573171616,0.1988903731,0.2325231135,-0.1714354753,-0.1934318691,-0.1323370934,-0.0262800548,0.0499790609,-0.0661576539,0.0992844701,0.2294830531,-0.2976982892,0.2404411435,-0.0680991858,-0.294751972,0.0200556256,-0.0614587069,0.022762306,0.2964047194,0.258754164,-0.1494279504,-0.1004934162,0.1512501091,-0.1010407507,-0.3608795702,0.1753015518,-0.2611690462,0.0857438892,0.179305926,-0.1949099451,-0.2642946243,-0.0162403081,0.1638755649,0.1531158835,0.0431888364,-0.1628793478,0.1258719712,-0.0752598196,-0.0013590266,0.100743182,0.0334941223,0.1717408448,0.3631780148,0.1620592624,-0.1893088967,-0.0503245331,-0.2994683385,0.0299040433,0.3481313288,-0.4328601956,0.0511916466,-0.1269654632,0.126892671,-0.147893399,-0.013433788,0.6234856248,0.1828055978,0.0386355966,-0.0361397006,-0.0435907207,-0.0636737719,0.016723467,0.1787310392,0.132709071,0.4253961444,-0.12833713,0.458335191,-0.3325755298,-0.35143435,-0.1117647588,0.1025599465,-0.0301402267,0.131637916,-0.3031069338,0.0468979031,0.1495187879,-0.3250490725,-0.0367085598,-0.2300450653,-0.3270386159,-0.1254158914,-0.1680469215,-0.0972600728,-0.0746328831,0.245324865,-0.1871758848,-0.0418294333,-0.3286375701,-0.0422058813,0.0985610336,-0.0402651131,0.0708337724,0.4972974658,-0.0194490217,0.0159368776,-0.0106006255,-0.0361704007,0.1508090198,0.011837705,-0.251788497,0.0891629085,0.0207467582,0.0858293995,0.1704929173,-0.0048480234,0.7906422615,0.1083032861,0.1197155491,0.2335706055,-0.1213772744,-0.4022609591,-0.0198285431,-0.3051272631,-0.1082741395,-0.1983741075,-0.1338706017,-0.2571406066,-0.0809813663,0.0834439471,0.2612773478,-0.366565913,-0.4619607627,-0.0363259427,-0.1190256774,-0.2458909899,0.1944992542,0.0799970403,-0.1093364134,-0.0393669009,-0.0230131373,-0.2647075653,-0.1444280297,0.1912236959,-0.1442517489,0.0128758876,-0.1909678876,0.0757486299,-0.1049499735,0.2206996679,0.2764229178,0.1894171387,0.1228580624,-0.4357866347,-0.0072484883,0.0178951491,0.414827764,-0.1693429798,-0.0173246879,0.0359382071,-0.0527621098,0.1164671406,-0.1306697577,-0.0785839781,-0.0155790709,-0.0330440737,-0.2313090116,-0.3976014555,0.3808095455,0.1818301678,-0.140666008,0.1326076239,0.1900203377,-0.4119344056,0.2964876592,0.6284173131,1.2970558405,-0.0131902415,0.186427176,0.0619845353,-0.2102872431,0.3373145759,-0.1518898457,-0.0384922735,-0.4181610942,-0.2754038274,-0.2438250631,-0.4449936152,0.3696979582,0.4542197883,-0.5083640814,0.1258341968,-0.3970276117,-0.1894249618,0.2302428186,-0.0515982285,0.309528023,-0.2607072294,-0.0556334443,0.009317128,-0.1761725396,0.1614353508,-0.0317940265,-0.1821483225,-0.2313051671,0.1159128845,-0.0707366467,-0.1343581676,-0.5251278281,-0.1050064638,-0.257414937,-0.4517552555,0.6619155407,0.3650534451,0.0085170148,0.0933761969,-0.091467239,0.0136749931,0.0320891589,0.1699231267,-0.0420560241,-0.2248436511,0.481806308,-0.1197159961,-0.155088529,0.1149731353,-0.1105771884,-0.1878626645,0.3535494506,-0.1395908445,0.281042546,-0.447185576,-0.0013260759,-0.0551852323,-0.0623601042,0.0393944718,0.0522319935,0.3158757389,-0.0062005525,0.3939512968,0.2616309226,-0.3410563469,0.0032758531,0.1149434745,0.6427344084,0.1686145514,0.2632628977,-0.270757705,-0.3766040206,-0.2814811468,0.1783767194,0.2387970686,-0.5504466295,0.0765404627,0.1007907614,0.1389284134,-0.0252306964,-0.2555324435,0.4737342,0.3242426515,-0.4022254348,-0.4078060687,-0.43561849,-0.0340487286,0.1871985346,0.0944365785,0.0212511793,0.2463050634,0.4592216313,0.1894714683,-0.2077483982,0.0023558873,0.0191597771,0.1419451237,-0.0059982482,0.0766406953,-0.3780401349,0.1673590541,-0.0596420467,-0.0741113871,-0.1974337846,-0.0940064192,0.0293082409,0.229769513,0.1479869932,0.1731090099,-0.2500618696,-0.0883189365,0.0347708054,0.1502580196,-0.0807121918,0.1347415596,-0.1625943035,0.3228533268,0.6502870917,-0.2357778847,-0.7550691962,0.073810719,-0.0346309766,0.0732270852,-0.1203264743,0.217337206,0.1160207987,-0.0137326634,-0.4053081274,0.1018752307,0.1381121278,-0.0314449593,0.3276544511,-0.1083854288,0.3197031021,0.1807759106,0.079679735,0.3073895872,-0.1853256822,-0.0103971502,0.2037962824,0.0623655058,-0.3804176152,0.0889636576,0.2093677819,0.0341319665,-0.1200725734,0.2622236311,-0.0607715286,0.0368853733,-0.1232184097,-0.0107915513,0.377045244,-0.2025353163,0.099825114,0.7080127597,-0.2438093275,0.0200608522,0.6556429863,0.197628215,-0.1139337495,0.1668041199,0.0704852417,0.383641541,0.1070641652,-0.1209114864,0.2517763674,0.1935066134,-0.0298947655,0.5973751545,0.1824341863,-0.3034851253,0.1653964669,0.0709727257,0.2092889547,-0.3164494634,0.058798857,0.0830916017,-0.2280347496,0.4133187532,-0.323368609,-0.1060566157,0.1456353366,-0.2937208116,-0.2135301083,-0.2387198657,0.3417325318,0.2120060325,0.0011814075,0.1860435307,0.3911277652,0.0605411194,0.0301330276,-0.1463622451,0.0761442035,0.1960951239,0.1521429569,0.1496317983,0.3942979276,0.1632653326,0.2761095464,-0.151746273,0.0723585188,-0.0845651105,-0.2777258158,0.0282801539,0.2136629224,-0.0870824009,-0.0966928527,0.2509816885,0.1430818141,-0.0689424425,0.2980152369,0.0975503102,0.1908214241,-0.0413772799,0.3700016439,0.1667381376,-0.4476439357,-0.0104994429,0.0961705223,-0.1885864735,0.0903825983,0.4336329401,-0.5003530979,-0.0619827919,-0.0532988012,-0.0292844232,0.2192305923,0.3402998149,0.1301074326,0.3010827601,-0.230379343,-0.0687949583,-0.3981025517,0.2100150138,0.1048308164,-0.1326296628,-0.0165349524,-0.1178207546,0.111249797,0.2530871332,0.1387646198,-0.0464429408,-0.2062134445,0.3914126456,-0.1201333329,0.0649346709,0.2923530936,0.4736348689,-0.0678983256,-0.2138273716,0.0961941257,0.3186488152,-0.0526873507,-0.0614868067,0.2939700782,0.1828927547,0.43956846,0.3321734965,0.0050552036,0.8048728704,0.0074450681,0.4048197567,0.5635287762,0.0819564909,-0.2770855427,0.3589548171,0.2951975465,-0.1419907361,-0.2297152579,-0.105080381,-0.0272404458,-0.2122631669,-0.0760066137,0.2210277617,-0.7341039777,0.0497249253,-0.13599886,-0.1535840482,0.1058367416,-0.0475432351,0.1144968495,-0.0552567616,-0.2382290959,-0.1923136264,0.3582506478,-0.0686089918,-0.2909313738,0.0617491566,-0.0494968556,-0.1480585635,0.1791734993,-0.256744504,0.0507356524,0.4911362827,-0.0887820274,0.2391620427,-0.0044988059,-0.1032989845,0.011300819,-0.186433658,0.8142256737,-0.329693377,0.0203690436,-0.4875219464,-0.259996593]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2595","title":"ModuleNotFoundError: No module named 'datasets.tasks' while importing common voice datasets","comments":"Hi @profsatwinder.\r\n\r\nIt looks like you are using an old version of `datasets`. Please update it with `pip install -U datasets` and indicate if the problem persists.","body":"Error traceback:\r\n---------------------------------------------------------------------------\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset, load_metric\r\n 2 \r\n----> 3 common_voice_train = load_dataset(\"common_voice\", \"pa-IN\", split=\"train+validation\")\r\n 4 common_voice_test = load_dataset(\"common_voice\", \"pa-IN\", split=\"test\")\r\n\r\n9 frames\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/common_voice\/078d412587e9efeb0ae2e574da99c31e18844c496008d53dc5c60f4159ed639b\/common_voice.py in ()\r\n 19 \r\n 20 import datasets\r\n---> 21 from datasets.tasks import AutomaticSpeechRecognition\r\n 22 \r\n 23 \r\n\r\nModuleNotFoundError: No module named 'datasets.tasks'","comment_length":27,"text":"ModuleNotFoundError: No module named 'datasets.tasks' while importing common voice datasets\nError traceback:\r\n---------------------------------------------------------------------------\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset, load_metric\r\n 2 \r\n----> 3 common_voice_train = load_dataset(\"common_voice\", \"pa-IN\", split=\"train+validation\")\r\n 4 common_voice_test = load_dataset(\"common_voice\", \"pa-IN\", split=\"test\")\r\n\r\n9 frames\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/common_voice\/078d412587e9efeb0ae2e574da99c31e18844c496008d53dc5c60f4159ed639b\/common_voice.py in ()\r\n 19 \r\n 20 import datasets\r\n---> 21 from datasets.tasks import AutomaticSpeechRecognition\r\n 22 \r\n 23 \r\n\r\nModuleNotFoundError: No module named 'datasets.tasks'\nHi @profsatwinder.\r\n\r\nIt looks like you are using an old version of `datasets`. Please update it with `pip install -U datasets` and indicate if the problem persists.","embeddings":[-0.4680482745,-0.2621825933,-0.0422590971,-0.1249682978,0.2041023821,0.1450624764,0.4402432442,0.2543092072,0.2220357209,0.1348177344,-0.2110196054,0.2088879645,-0.279766947,-0.0780611783,0.0812453702,-0.0557086468,0.0423195399,0.2207903266,0.0042473162,-0.1911050677,-0.0319739655,0.1782887131,-0.3306013346,0.2055014223,-0.3548139334,-0.2180955708,0.0779372454,0.0423675068,-0.0422454886,-0.2438998669,0.1500824243,-0.0146506298,0.070500724,0.4818688333,-0.0001044994,0.0464495383,0.3543689251,-0.0503025986,-0.3212524354,-0.4222036004,-0.3318475187,0.0348274596,-0.0461382121,-0.1579176635,0.1033446863,0.1319067329,0.1263466626,-0.1368252337,0.5385984182,0.3482384384,0.2840605378,0.4131548703,0.1567139328,-0.1267151833,0.0197893772,0.2498226613,-0.0403087847,0.2972659767,0.0895575508,-0.1931072623,0.3447549641,0.2893808484,-0.2390359789,0.063667126,0.0798237845,-0.0465243123,-0.0414960198,-0.4952995181,0.0744473934,0.132193625,0.3183714747,-0.4103887081,-0.1930001378,-0.2050786465,0.1101318151,-0.2723012567,0.0004303494,-0.0182032287,-0.132379517,0.1062286645,0.1162843257,-0.1884781122,0.0229160469,0.0268927347,-0.0017303053,-0.0327819213,-0.0579357445,-0.1311632842,0.2440393865,-0.1913584322,-0.1221389621,0.236902833,0.1770546585,0.5055808425,-0.4833761454,-0.1723853052,0.2230124623,-0.2537678778,0.0828389078,0.0121455379,0.0260496307,0.1074604318,-0.1929781884,0.2938083112,0.1571602672,0.2055592388,0.2407598346,0.3080771863,0.2705402672,0.1983297765,-0.0529231802,0.1062482819,-0.2731648088,-0.16814138,0.2600198984,-0.0516615026,0.4945361614,-0.1721232682,-0.5520052314,0.0645157322,-0.1806008816,-0.0973164216,0.0437590554,0.2725822926,-0.0597683676,0.2682129443,0.0992677733,0.0980217084,-0.3125715852,-0.083981581,-0.2083452344,0.1204670966,-0.1339889616,0.0375394039,0.1968412548,-0.1620706022,0.1611405611,0.0262966827,0.1254214644,-0.0994550213,0.0607768632,-0.0804267302,-0.3314159214,0.1653537005,-0.0582629256,-0.0140873948,0.044594083,0.0253963694,-0.1504641473,0.0868899375,-0.3289185762,-0.1916423887,0.0417402312,0.2643729448,-0.0014749963,-0.259262681,-0.1852260083,0.0195644684,-0.0112829041,-0.3709941208,-0.1309114099,-0.3607839644,0.066148676,0.1108472869,0.3037349284,0.2592234015,-0.0803736299,-0.254234314,-0.1639973223,-0.1825810671,0.0133437915,0.3237006068,-0.0084946277,0.1405990422,-0.0637404993,0.4380916059,0.2472978979,-0.452693969,-0.2491369992,0.1340448409,0.0961463302,0.0678056851,-0.1399039924,-0.214760378,0.1372883916,-0.0111433854,0.3468581736,0.1733139008,-0.0133900391,-0.1076784283,-0.0589140691,-0.1439610422,0.1319515407,0.2343855649,0.0049605146,-0.04832251,-0.0427682996,0.1929655075,0.1269651949,-0.1789568216,0.0387045667,0.2070473135,0.4356282353,0.0695726126,-0.089049004,-0.4199829996,-0.1745185405,0.0376326963,0.021216793,0.2585080266,-0.2317058444,-0.0989650488,-0.405416131,-0.1051226109,-0.3773017526,0.1202742383,0.1669739634,0.099082917,-0.0610881634,0.0066819703,-0.2997961044,0.3757269382,-0.1313783824,0.0944234282,-0.3323701024,0.3736220896,-0.237708658,-0.2670727968,0.2409504801,0.411372602,-0.0041607246,0.0625160709,-0.1170939207,0.2125421613,-0.0076911575,0.172198087,-0.2736349404,-0.075165987,0.0657028928,-0.3961906731,0.0338857807,0.1110476479,0.0426315181,-0.0886901394,0.028556956,0.1296461821,0.1537565589,0.1063358858,0.1165112779,0.0868430957,0.1682522446,-0.0683105662,-0.2281047553,-0.0361748189,0.2848306596,-0.1852824241,0.4031330645,-0.0381330252,-0.6622078419,-0.3254399598,0.4939145148,-0.023165077,0.2633921504,0.0882816017,-0.2232540697,0.2004333287,0.1287862211,-0.1824391335,0.4371800423,0.2529809177,-0.1184101105,0.2785458565,-0.1213825792,-0.0148254698,-0.0021355702,0.0376375988,0.0944892615,0.1906543225,0.0539587885,0.1309286952,-0.4432955384,-0.2492175102,-0.2248527259,0.1723359972,-0.3543302417,-0.034156058,-0.2780617774,-0.048387751,-0.1897361577,-0.0976622403,-0.2422444671,-0.0885605216,-0.0185894705,0.1371502578,0.1353910863,0.4319052696,-0.0441789739,-0.0619513579,0.046910055,-0.4555506408,-0.1673716903,-0.1692593843,-0.0982952863,0.1030317545,0.3574747145,0.0153019419,0.2674027383,-0.2192237973,-0.0046595428,0.0495451801,-0.3638139665,-0.0044207326,-0.2471049726,0.1006778479,0.028096972,0.1944648176,-0.087781027,-0.2716493011,0.2568017244,-0.2902956009,-0.3031878173,0.0084943045,-0.0715098754,-0.2179573327,-0.1056061089,-0.6792387962,-0.4916081429,-0.3824001849,0.1891574264,-0.004323863,-0.0504479334,0.5760887861,-0.0038484496,0.1662248075,-0.0396011472,0.2001395673,-0.1855196953,-0.00141783,0.3684045672,-0.1762257516,-0.198447302,0.0650722608,-0.2073300034,0.6107491851,-0.1771792918,-0.1790088862,-0.0992143452,-0.1153991073,0.2072138935,-0.1426563114,0.1987741143,0.2562193871,0.2243711203,-0.0665225461,-0.0385431163,-0.2290917933,-0.0892246291,0.5440238714,0.1236746758,0.0299236327,0.3696120977,-0.2337699234,0.7312172651,0.2542453706,0.0529726259,0.225074172,-0.1787748784,0.2573381066,-0.2329882383,-0.4067516625,0.2250757366,0.0523331612,0.1240250543,0.1890376806,-0.0755172074,-0.1717136353,-0.308570981,-0.1609124243,-0.2336978465,-0.2439306676,-0.1122148558,-0.2118361443,0.3328001499,0.1365736425,0.0779008269,-0.1717391908,-0.0623184666,0.1407050192,0.1314403117,-0.06792932,0.0676836148,-0.3149287105,-0.09493009,-0.2611026168,0.1321661919,-0.0671634376,0.3845616877,-0.3105469048,-0.0053711208,-0.090608865,0.1061191633,-0.0268048961,0.051029183,-0.1677804887,0.1527031213,0.2210835665,-0.1567210555,-0.0859284997,-0.1342551708,0.0122811776,0.1434632242,0.1530689299,-0.2973598838,0.0127240783,0.3386218548,0.3727748394,0.0377482809,-0.0941172987,-0.266810149,-0.3024866879,-0.4897350371,0.1207361445,0.1983530521,0.0963838547,-0.199314937,0.1669947654,0.2083149999,-0.1086240038,0.4872302413,0.1101528257,0.1720112115,-0.0501766764,0.3456845284,0.5056827068,0.1342127621,0.2906112671,0.6657876372,-0.022307463,-0.5903895497,-0.1238747984,-0.0629287139,0.3958955705,0.0872491077,0.1059036329,-0.0479561463,0.3212256432,-0.0264970846,-0.0484167039,0.1525449604,0.5146484375,0.3927659392,-0.3394790888,-0.4118445218,0.0950060934,0.1855013967,0.0875615999,0.0681461543,-0.1043024361,-0.1613092422,-0.0356018841,-0.484215647,0.7983003259,0.3278929591,0.0445062108,0.2252581567,-0.030371327,0.0399945565,-0.1018434018,0.0884786174,-0.132490173,0.0586846545,-0.0059453035,-0.1115389168,0.2056113929,0.0889565349,-0.4066588879,0.4010350704,0.0836590976,-0.0207543559,0.0382337049,0.2802622318,0.0214720927,-0.140996933,-0.1178429127,0.2391809225,-0.1585398465,0.5873627067,0.0903311521,-0.0947030783,-0.0624560639,-0.3076584041,-0.1190492511,0.2365182787,-0.1763168722,0.1899975538,0.3196704686,-0.0780479908,0.3400167823,0.0243677832,0.0111994185,-0.0196507536,-0.2427959442,0.0951937586,0.1217556447,0.0140334871,-0.0462796129,0.0185666215,0.5849072337,-0.1009639651,-0.4325719476,0.1455033123,-0.0550938919,0.0390628092,-0.2691911757,0.0498390198,0.2839344442,-0.1750938743,-0.2218398154,0.122325711,0.2192093432,-0.0271995161,0.2155409157,0.0649078935,-0.1225234345,0.3755879104,0.0395610817,0.0722048804,-0.0408567376,0.2515895963,-0.0187330022,-0.3264916837,0.6129220724,-0.075041689,-0.1617054343,-0.3027546704,0.0731457993,0.1165141538,-0.2352831513,-0.0564119704,0.1171291843,0.0903480351,0.0291938931,0.1955384612,-0.0177138988,-0.248647958,-0.0644360483,-0.1412009597,-0.398311764,0.2289278507,0.1119449958,0.3001005054,-0.1009243131,0.3502114415,0.24706164,-0.0204842873,-0.4086298943,-0.0172807295,-0.1506142467,0.2297730297,0.386777252,-0.2138602734,0.079557769,0.0045471182,0.2706660032,0.0681105554,-0.1541389972,-0.2902034223,-0.1380581707,0.0749273971,0.108120203,-0.0461759642,0.1655641198,0.0306889918,-0.0927981958,-0.1629673839,-0.0240252782,0.0706077889,-0.1004457921,0.261172682,0.0199568924,-0.0094588054,0.246232897,0.1467534155,-0.0240725391,0.1554422826,-0.0228901897,0.0898328051,0.1484561116,0.0780455545,-0.2269878536,0.0927135423,0.3652599454,-0.2079784125,0.091952391,-0.4146076739,-0.1869242787,0.0035451811,0.2747445405,0.2939597964,-0.2162155062,0.0663702041,0.3227999508,0.2919089794,-0.433970809,0.217602551,0.3344878554,0.00392337,0.1013886407,-0.0108274808,0.0384636074,0.0470238067,0.1753734201,0.1658098102,0.1026381701,-0.3446681798,0.3764196336,0.0506742969,0.0758729503,-0.1463964134,0.3820573986,-0.0301786587,0.1045023501,0.5094951987,-0.4414913356,0.1682896614,0.0115654087,0.1542970538,0.2767112851,-0.3233615458,-0.2022305876,-0.1139013395,-0.0536592305,0.0028553321,-0.0626457557,0.6269554496,-0.3885323405,-0.1094072685,-0.1444596052,0.1354536712,-0.3061712086,0.0700819194,0.0340694822,-0.1839968115,-0.1791858226,-0.2247749716,0.1355302334,-0.082457155,0.4152858853,0.0984390751,-0.1780042648,-0.2696000338,-0.1987647414,0.2105157673,0.0625380427,-0.0852094516,0.3057459593,-0.0383055247,-0.1322033703,0.0668961331,0.5210747123,0.5806873441,0.1711712778,-0.0728710815,0.1826114357,-0.0855727717,-0.07937181,0.0943482742,0.1167582944,0.0382122844,0.3208302557,0.1235490367,0.2243593633,-0.226411134,-0.0492910109,-0.1167298704,-0.2444069237,-0.446511656,0.3851093948,-0.2668535411,-0.1144590676,-0.3867430091,-0.0310380254,-0.5898923278,-0.0372401401,0.4932135344,0.2721070647,-0.1028746143,-0.3833261728,0.1120099574,-0.2403333634,0.1437900066,0.3910233676,0.2161833495,0.0266759526,0.0195010677,-0.6945980191,0.1206174046,-0.0724870265,-0.0154803135,-0.0486516356,-0.0889657214,0.0467394441,0.1318782419,0.1182086915,-0.2654316425,-0.2165034413,-0.2908381224,-0.304194361,0.0143067241,0.0956117213,0.0445631035,0.1643180847,-0.2925551236,-0.0871799886,-0.3029634356,0.0716293305,-0.1985855997,0.0138236349,-0.0220372267,0.1548384875,0.3778554201,0.0707522184,0.7247420549,-0.1219094768,-0.0149170402,0.0884628668,-0.3524926603,-0.1980545819,0.1428397596,0.0400935113,0.4865868688,-0.1656808406,-0.1466167867,-0.2866219282,0.0134917637,0.0952576622,-0.3367992342,-0.0814708844,0.1075755432,-0.0959665701,0.2380140573,-0.0568129383,0.0383794494,-0.182372883,-0.000693341,-0.3166334331,-0.4183671176,0.6252818108,-0.0994776711,-0.4117881656,-0.4622044861,0.2439309359,-0.1391502172,-0.2128487974,-0.4571861029,0.1743622124,0.3370921612,-0.0477892943,-0.0868127123,0.2577061057,0.0695325956,-0.0047335685,0.0295038838,0.2987222075,0.049352888,-0.0268669501,0.2987110615,-0.15217188]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2595","title":"ModuleNotFoundError: No module named 'datasets.tasks' while importing common voice datasets","comments":"@albertvillanova Thanks for the information. I updated it to 1.9.0 and the issue is resolved. Thanks again. ","body":"Error traceback:\r\n---------------------------------------------------------------------------\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset, load_metric\r\n 2 \r\n----> 3 common_voice_train = load_dataset(\"common_voice\", \"pa-IN\", split=\"train+validation\")\r\n 4 common_voice_test = load_dataset(\"common_voice\", \"pa-IN\", split=\"test\")\r\n\r\n9 frames\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/common_voice\/078d412587e9efeb0ae2e574da99c31e18844c496008d53dc5c60f4159ed639b\/common_voice.py in ()\r\n 19 \r\n 20 import datasets\r\n---> 21 from datasets.tasks import AutomaticSpeechRecognition\r\n 22 \r\n 23 \r\n\r\nModuleNotFoundError: No module named 'datasets.tasks'","comment_length":17,"text":"ModuleNotFoundError: No module named 'datasets.tasks' while importing common voice datasets\nError traceback:\r\n---------------------------------------------------------------------------\r\nModuleNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset, load_metric\r\n 2 \r\n----> 3 common_voice_train = load_dataset(\"common_voice\", \"pa-IN\", split=\"train+validation\")\r\n 4 common_voice_test = load_dataset(\"common_voice\", \"pa-IN\", split=\"test\")\r\n\r\n9 frames\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/common_voice\/078d412587e9efeb0ae2e574da99c31e18844c496008d53dc5c60f4159ed639b\/common_voice.py in ()\r\n 19 \r\n 20 import datasets\r\n---> 21 from datasets.tasks import AutomaticSpeechRecognition\r\n 22 \r\n 23 \r\n\r\nModuleNotFoundError: No module named 'datasets.tasks'\n@albertvillanova Thanks for the information. I updated it to 1.9.0 and the issue is resolved. Thanks again. ","embeddings":[-0.3636057675,-0.3833416104,-0.0181181207,-0.1324386895,0.1585254967,0.1734296679,0.408313483,0.2774201035,0.1942006648,0.0959576741,-0.1962040365,0.2137221694,-0.2857515514,-0.0307673588,0.1209338233,0.0243292525,0.0525843576,0.1785367876,0.0639957488,-0.2289384007,-0.0830177069,0.179896608,-0.3790596128,0.2335020602,-0.3515877426,-0.1942125112,0.0089904023,0.0472989678,-0.0594284609,-0.2619594634,0.1584800035,-0.0244397465,0.0247928109,0.5354664326,-0.0001072218,0.0130804796,0.3887681067,-0.0951573327,-0.3423600495,-0.4812251627,-0.5180749893,-0.0198785383,-0.0378897265,-0.1568300575,0.1432608515,0.0931807831,0.1367631555,-0.0211385265,0.518473804,0.3234174848,0.2760933042,0.4396746457,0.1758444011,-0.1285552084,-0.0919889957,0.3200479448,-0.0312032439,0.3851700425,0.0718947947,-0.2481864244,0.3491150737,0.3169421256,-0.2483986616,-0.0398039781,0.2099717557,-0.070443131,-0.0864186138,-0.5517331362,0.1612678468,0.1745328158,0.3149979413,-0.4002381265,-0.2445603013,-0.1867613643,0.1450690776,-0.297483623,0.0211215783,0.0581536815,-0.1166366786,0.048545327,0.0884919837,-0.2153846622,0.0324986577,-0.0098934444,0.037946146,0.0695422217,-0.0714510754,-0.1244164407,0.2239089608,-0.1729114354,-0.2198520452,0.1894756556,0.1876290888,0.5041465163,-0.4748544097,-0.1837355942,0.2259812951,-0.2747797668,0.0592493415,-0.1257465929,0.0696833134,0.0972484797,-0.141737625,0.2598145902,0.2328961939,0.2657309473,0.3367702663,0.3275896609,0.3152905703,0.320507586,-0.0319153629,0.0961842313,-0.2667919397,-0.1119739935,0.2743022144,-0.00721126,0.5157472491,-0.1445227414,-0.4783940911,0.0268434417,-0.1586011499,-0.0491169132,0.0459224992,0.3150606751,-0.0719008446,0.2049201578,0.1140737161,0.0707881227,-0.33948946,-0.0207248442,-0.2056521177,0.1439721435,-0.1351425797,0.039093025,0.166922763,-0.2369956672,0.2426794171,-0.0384485722,0.1271665543,-0.1646114588,-0.002661827,0.0350436643,-0.2853616476,0.218885541,-0.0577582195,-0.0250663739,-0.006532751,0.0594038293,-0.1507555693,0.148841694,-0.385599345,-0.2488822043,0.0353923291,0.2451751083,0.0019521569,-0.2012861073,-0.2043664157,0.1103080884,-0.0199954249,-0.3280108869,-0.1137684807,-0.3114471734,-0.0183950011,0.0832226723,0.314004153,0.2145241797,-0.0957234651,-0.2608369887,-0.1386352628,-0.1814469546,-0.0146558946,0.2992719412,-0.0283241291,0.1103681177,-0.0679047406,0.5170898438,0.1305839866,-0.4023426175,-0.2130296677,0.0811728686,0.0615915954,0.0784978718,-0.1494661421,-0.226228103,0.1217366979,-0.0053586592,0.2895030081,0.2189677507,0.0232907105,-0.0666875616,-0.07012631,-0.1601055413,0.073183924,0.1477267593,0.043798741,-0.0165528934,-0.083794862,0.1944784969,0.1807451546,-0.2012950331,0.0373748615,0.2082426846,0.4772288799,0.0610024035,0.0018083723,-0.2899799049,-0.2439673394,0.0906843767,-0.0490243696,0.3566188514,-0.3079186678,-0.1824970543,-0.3866797984,-0.052775111,-0.3918936253,0.0559836365,0.1665537059,0.0353793055,-0.1400140822,0.0161962528,-0.3451236188,0.2648701966,-0.0309727509,0.1227003783,-0.3519035876,0.3853035271,-0.1914645284,-0.3086241186,0.2147639543,0.4524838924,-0.0079535302,0.0620539598,-0.1368801594,0.1645605415,-0.0686163455,0.2083025426,-0.253475666,-0.1135930791,0.1204184741,-0.3686172664,-0.0335526131,0.1560337245,0.1338924617,-0.1023146734,0.0601926297,0.192547828,0.2630729377,0.0872495919,0.1627519578,0.0740887225,0.0583264157,-0.0703037307,-0.1851859391,-0.0571805499,0.2701919079,-0.1746159494,0.3756133914,-0.0258409549,-0.604056716,-0.3535889685,0.4849467874,-0.0782649219,0.2236102521,0.0556300953,-0.2568132579,0.1159279943,0.1705579758,-0.3374951184,0.3594591618,0.1739274263,-0.1083806679,0.3482666612,-0.0841783807,0.0032746568,-0.0126491254,0.1101145372,0.0521428324,0.1774294227,0.0945809409,0.0503328927,-0.4496413469,-0.26560691,-0.1863158792,0.186430946,-0.3769983053,-0.0612788089,-0.2536341846,-0.0062754131,-0.1210898235,-0.1308910251,-0.29107517,-0.1262822747,-0.0285301488,0.1420176327,0.116486527,0.411904335,0.0562704466,-0.0428694561,0.0566488318,-0.3579311669,-0.2157428414,-0.2365891933,-0.054414209,0.06168833,0.3786766231,0.0383745506,0.1699182689,-0.1812996864,-0.0000645514,0.1045465395,-0.4350382388,-0.0063254042,-0.2335228622,0.0633962974,0.0017652768,0.1968349516,-0.0975241214,-0.3132129312,0.2633445561,-0.2720876634,-0.2986532748,0.0646651611,-0.0801601335,-0.2136177123,-0.1200264841,-0.5569186211,-0.4403799176,-0.3566887379,0.1819927096,-0.0772865787,-0.0431814045,0.6514210105,0.0293715503,0.237774834,-0.1074320525,0.161714375,-0.1048827395,0.0539454333,0.3376305997,-0.1701786071,-0.1958740652,0.1166847646,-0.2572161853,0.6778261065,-0.1940079182,-0.2012849599,-0.0728161186,-0.0509069674,0.2408657968,-0.1346973926,0.1593551189,0.2059505135,0.1566209793,-0.0554307662,-0.0127729671,-0.2480054945,-0.0972873941,0.5820992589,0.0944190845,0.0600190833,0.3471995294,-0.2486725897,0.786429286,0.2308702469,0.110745497,0.2190027684,-0.0982874706,0.267440021,-0.2108428031,-0.4376784861,0.2226893604,0.0914874896,0.1118222624,0.2504206598,-0.0326574594,-0.1213685721,-0.3556512892,-0.3254825473,-0.1687034667,-0.2966779172,-0.1167401001,-0.0817803219,0.3221668303,0.1697291285,0.1055976599,-0.2187429816,-0.0654276311,0.1467369348,0.2421664,-0.0970553383,0.065382272,-0.306646347,-0.1483701169,-0.3405385017,0.1368670166,-0.0517593287,0.4558203518,-0.2143550217,0.0210489072,-0.0781608522,0.1187463775,0.0026040876,0.109401457,-0.1714257002,0.1266745627,0.0890690014,-0.1200414002,-0.104627803,-0.0660476312,0.0593859106,0.1450898498,0.2695623934,-0.3026839793,0.0177181996,0.3039089739,0.376061976,0.0805392489,-0.135009557,-0.3770295382,-0.3506780863,-0.4728805721,0.1429031789,0.1724353582,0.0970224142,-0.1271730959,0.1632328928,0.2196709365,-0.1573037654,0.442021966,0.1218705326,0.1329137683,-0.028629059,0.3401709497,0.4768105745,0.066462487,0.237512812,0.6372557878,-0.0236761365,-0.7577488422,-0.1979634911,-0.1201224551,0.3438262045,0.112163417,0.090191707,0.0276927836,0.4193373024,-0.0080799079,-0.0076427571,0.113189131,0.48464185,0.4113264382,-0.4018982947,-0.3383566439,0.0783466622,0.1432867646,0.1592214704,0.1371793151,-0.0960872397,-0.2214962244,-0.0248497315,-0.5586670041,0.7840056419,0.322755456,0.0612405092,0.2629365325,-0.0033959302,0.0799395666,0.045301538,0.1685222387,-0.1381043792,-0.0160070863,0.0119671859,-0.0974415541,0.220621407,0.0779653341,-0.4101417363,0.4231623709,0.0714922249,0.0468714237,0.0075262901,0.2793138921,0.1080201268,-0.1398216784,-0.1461079419,0.1999603063,-0.1303404868,0.632605195,0.1015598178,-0.1258403659,-0.030548444,-0.314910084,-0.1065031141,0.2004971653,-0.2200043201,0.2250460833,0.3635545075,-0.0740767419,0.3070265651,-0.0908043012,-0.0672450513,0.0124379434,-0.2876226008,0.0829657763,0.1172918007,0.0674089417,-0.0960060731,0.0318056829,0.6111606956,-0.1110403463,-0.4229340255,0.167319268,-0.0954173431,0.1045304537,-0.2354660332,0.0267726872,0.2249741256,-0.0980881602,-0.225570783,0.1045550406,0.2335185409,-0.060801629,0.1946658343,0.0483964533,-0.1451740861,0.33759588,0.0521147177,0.0055138986,-0.0695343241,0.2893537879,-0.0004022074,-0.3251859546,0.6093985438,-0.0939882323,-0.1770113409,-0.2521192431,0.0321931802,0.2130113542,-0.2881563008,0.0263941325,0.1457322836,0.0958659947,0.0091585051,0.2715731263,-0.0357916355,-0.233773753,-0.0870273337,-0.1145758629,-0.4128243625,0.2796646059,0.1166093647,0.2642892897,-0.129111141,0.4132292271,0.3312872648,-0.0469710752,-0.376903832,-0.0129263531,-0.1994544268,0.1677209139,0.3669532537,-0.1935517937,0.0903575346,0.0834858045,0.2580676973,0.08665438,-0.1299781054,-0.2424425781,-0.1228833646,0.0876505375,0.1171844825,-0.0328589752,0.1796568185,0.0119785834,-0.1475057751,-0.1580754966,0.0210805703,0.0714745224,-0.1193956956,0.1821638942,0.076413855,-0.0071999617,0.2596047819,0.2125116587,0.0411075726,0.1299154907,0.0073170559,0.0255975872,0.1358450651,0.1108726338,-0.1450424194,0.0927249417,0.2530200779,-0.2015181482,0.0851775035,-0.3775853515,-0.1525233835,0.0016042957,0.20830369,0.3257995546,-0.1462607682,0.0995503142,0.3164529204,0.2548805773,-0.4426314831,0.2402202338,0.3323379159,0.0445717312,0.0907235742,-0.0140095064,0.0340851694,0.0505933501,0.1929453909,0.145035997,0.001135315,-0.2689981163,0.385049969,-0.0095105218,0.0465697497,-0.080815196,0.3681237102,-0.0208935682,0.0974600837,0.382058233,-0.4378886521,0.241949439,0.0315863863,0.1459220052,0.2872114778,-0.3768296242,-0.2674653232,-0.1129546687,-0.0158966929,0.0211884547,-0.0738535449,0.6236789227,-0.4865311682,-0.0626909733,-0.0776228383,0.1505668461,-0.3340906799,0.1247662231,0.0412308127,-0.1665499955,-0.1529593468,-0.2408024073,0.1342776418,-0.0620056055,0.3735053837,0.0824028328,-0.1566254199,-0.2083680034,-0.2844169438,0.21982494,0.0288598239,-0.0974245742,0.2185845971,-0.0091725662,-0.0740261599,0.0873245224,0.5061692595,0.5795962214,0.2384861261,-0.0331308618,0.1049508005,-0.0168664958,-0.0602551997,0.1299707741,0.0700257197,0.1368560791,0.4141387343,0.0976887569,0.1724712402,-0.1997913271,-0.0529920012,-0.0429135673,-0.2157284617,-0.4559637904,0.2976952493,-0.312818557,-0.1389408112,-0.382270366,-0.0345834903,-0.5582609773,0.0663024336,0.4672362208,0.2679565549,-0.1678644419,-0.3979560137,0.0895160437,-0.2667342424,0.1480169147,0.4019308686,0.1818224639,-0.0192097239,0.0471448451,-0.679825604,0.1862460375,-0.0133905532,-0.0708920732,-0.0427570716,-0.1561949402,-0.024506323,0.0166963302,0.0877141356,-0.3457701504,-0.1732069552,-0.3885888159,-0.3407887816,-0.0383819789,0.0890555307,-0.0284403954,0.1961668432,-0.2559446096,-0.0215818286,-0.2670929432,0.0397152714,-0.1974106282,-0.0096197547,-0.0293373019,0.2362259626,0.320160836,-0.0436137579,0.685218513,-0.0979719311,-0.063276425,0.0546017475,-0.2957471013,-0.2335707545,0.1820387542,0.0998794213,0.4836899638,-0.1316227913,-0.0821451694,-0.283893168,-0.003802537,0.0510100164,-0.3323509693,-0.0647727847,0.196378693,-0.0617847256,0.2327576131,0.0000854118,0.0397845395,-0.2082762867,-0.0568772592,-0.3205325603,-0.4250100255,0.6887363195,-0.1174832135,-0.4332188964,-0.4882303476,0.2215625495,-0.1690745801,-0.2508116364,-0.4525445402,0.1826771647,0.3078407049,-0.0280828234,-0.1032514796,0.2085326612,0.0283309985,0.0559333488,-0.0203062631,0.2545564771,0.1018206105,-0.0872147083,0.2926311493,-0.1661280543]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2591","title":"Cached dataset overflowing disk space","comments":"Hi! I'm transferring this issue over to `datasets`","body":"I'm training a Swedish Wav2vec2 model on a Linux GPU and having issues that the huggingface cached dataset folder is completely filling up my disk space (I'm training on a dataset of around 500 gb).\r\n\r\nThe cache folder is 500gb (and now my disk space is full).\r\n\r\nIs there a way to toggle caching or set the caching to be stored on a different device (I have another drive with 4 tb that could hold the caching files).\r\n\r\nThis might not technically be a bug, but I was unsure and I felt that the bug was the closest one.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 186, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 397, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1983, in _map_single\r\n writer.finalize()\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 418, in finalize\r\n self.pa_writer.close()\r\n File \"pyarrow\/ipc.pxi\", line 402, in pyarrow.lib._CRecordBatchWriter.close\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nOSError: [Errno 28] Error writing bytes to file. Detail: [errno 28] No space left on device\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n","comment_length":8,"text":"Cached dataset overflowing disk space\nI'm training a Swedish Wav2vec2 model on a Linux GPU and having issues that the huggingface cached dataset folder is completely filling up my disk space (I'm training on a dataset of around 500 gb).\r\n\r\nThe cache folder is 500gb (and now my disk space is full).\r\n\r\nIs there a way to toggle caching or set the caching to be stored on a different device (I have another drive with 4 tb that could hold the caching files).\r\n\r\nThis might not technically be a bug, but I was unsure and I felt that the bug was the closest one.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 186, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 397, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1983, in _map_single\r\n writer.finalize()\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 418, in finalize\r\n self.pa_writer.close()\r\n File \"pyarrow\/ipc.pxi\", line 402, in pyarrow.lib._CRecordBatchWriter.close\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nOSError: [Errno 28] Error writing bytes to file. Detail: [errno 28] No space left on device\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\nHi! I'm transferring this issue over to `datasets`","embeddings":[0.0375595093,-0.4222126603,0.0985903814,0.4514758587,0.1011737883,0.2050561309,0.0744853392,0.141904518,0.1445289999,0.0556368642,0.4010794461,-0.2280424833,-0.1242998913,0.1127196699,0.1466887593,0.0686893761,0.2591468096,-0.2458593994,-0.0610483401,0.1060899645,-0.1954928488,0.2656945884,0.0236382782,-0.1252904385,-0.2641810179,-0.1312523186,0.1114902273,0.0915549919,0.1169939488,-0.2215207964,0.6455171704,-0.037543118,0.2969840467,0.3342037201,-0.0001271115,-0.2188411504,0.0915412679,-0.181593284,-0.0922594294,0.5504892468,0.1914410889,-0.3445687294,-0.0522951484,0.0128892465,-0.1940163374,0.1353006959,-0.0969077572,-0.4302129745,0.2448510528,0.0724848434,0.0480299257,-0.0077622137,-0.1181686893,0.1708012968,-0.1092367396,-0.0044049905,-0.2483725548,0.1998682469,0.2498452663,-0.1698941886,-0.1137219444,0.5557486415,-0.0590143166,-0.0844287351,0.5253401995,0.0833818987,-0.3562970459,-0.1106624231,0.3813790679,-0.0801356807,0.314463526,-0.4211860299,-0.0359757133,-0.5416691303,0.1242926195,-0.4084812701,0.4917798042,0.1439301074,-0.3081754148,0.2055083513,-0.2643952072,-0.4621683359,-0.2639658749,-0.1211516634,-0.0510151535,-0.1692601889,-0.1779297143,0.0038869162,0.4981890023,-0.1243522316,-0.0180550069,-0.1148280278,-0.1485933661,0.1664890051,-0.4704052806,-0.1601628959,-0.1597560495,0.4838789701,0.1778159291,0.0343691669,0.0580152795,-0.2219776958,0.0135496575,-0.0545078851,0.0384909436,0.6338197589,-0.4302008152,0.1479855031,0.0260615703,-0.1004753411,-0.0664962679,-0.0840651393,-0.0061190045,-0.388866663,0.2431183755,-0.3094467223,0.0366061553,-0.266004473,-0.1329783052,0.1096001938,0.1877570152,0.1407824308,0.2514514029,0.3864438534,0.1821056455,0.3051152825,-0.322846204,0.2218101174,-0.2443430275,-0.1192895547,-0.0440737009,-0.0846645087,0.0494980812,0.3295774758,0.3280156553,-0.1158240363,-0.0438293442,-0.1668728292,0.2368198782,-0.0264211819,0.0309991352,-0.2585927844,0.2868451476,0.1123239547,-0.3598536849,0.225169003,0.2425775528,0.3989651501,-0.2411044687,0.1247824728,-0.3550453782,-0.2518386543,0.1480785757,-0.01426965,-0.1831807494,-0.0969632864,-0.429746151,0.2190823555,0.4013512135,-0.2894096375,0.0460615568,0.0167300683,-0.4555115998,-0.1271671653,0.2183404118,0.390550375,-0.1844455749,-0.0037565609,0.0871429145,0.4355638623,0.0485231653,0.6233537793,-0.1413376033,0.0224382952,-0.3221297562,-0.2130482346,-0.034653984,-0.5695789456,-0.7795040607,0.015956128,0.0507356077,0.2829304338,0.3464263976,0.1854577065,0.0370459333,-0.1173866913,-0.1299520284,0.2160572261,-0.0045058765,-0.0558645725,-0.3231567442,-0.5304920673,0.28534621,-0.1147800237,0.0466500521,0.1006035507,-0.2822220922,0.0011679371,0.1973249465,-0.0316379927,0.2541683018,0.3844375014,0.293066293,-0.3115614951,-0.0663591474,0.0042587491,-0.4575053751,0.1950948834,0.1783739477,-0.3753206432,0.1245837435,-0.1003920138,0.0144068571,-0.2264458388,-0.1339516789,-0.2168265283,-0.1374632716,-0.0339444168,0.4003777504,-0.176329121,-0.2991106212,0.3400560319,0.0025759132,-0.0406318903,-0.3259862363,-0.0154328085,0.1447563022,-0.0170684289,-0.2598402798,0.1888251454,0.0648194179,-0.0355281867,0.0368644558,0.2335698307,0.0226677191,0.3187166452,0.349332124,0.3796356022,0.2128698975,-0.0915223658,0.2820823789,0.2902056575,-0.0620261133,0.0524619706,-0.3751501739,0.2003801614,0.0710650906,0.1950615644,-0.113981016,-0.2368826121,0.251411289,0.1360599846,0.0944542736,-0.1216271147,0.2244897187,0.0017589895,0.392508775,0.0476589166,-0.0416111052,0.2306888103,0.3817815483,0.2312813848,0.0795109794,0.3927400112,-0.1257767677,-0.2583568692,0.1078153104,0.3400515914,0.5201392174,0.1659973711,0.1927121878,-0.08257889,0.0334070846,-0.0248670559,0.1405932903,-0.2922542989,-0.1662439853,0.0999909863,0.1785677075,-0.1703336537,-0.1151010394,0.2065084279,-0.0072104023,0.0374703854,-0.1230290011,0.2915499508,-0.0051883152,0.0669604018,-0.300422281,0.2018046081,-0.4009771645,-0.0063932566,0.1357572228,0.5235956907,0.0959122628,0.0231753308,-0.1968869716,0.7337744236,-0.0781162232,-0.2568981051,0.0806490034,-0.0120087182,-0.1814791262,-0.1487001628,0.3093964458,-0.5017920732,0.097177811,0.1724740267,-0.0851729885,-0.2979379892,0.1057201475,-0.0711285472,0.1724738181,0.0262167249,-0.3023276925,0.3171690702,0.0086356783,-0.0908934399,-0.086091809,-0.1090197414,-0.0805503279,-0.1161442846,0.0500084944,0.4038137794,0.1403172016,0.0388037674,-0.260648638,-0.230347693,0.2758387029,-0.1536940336,-0.0432945602,0.081139639,0.0112992609,-0.1315152347,-0.123419635,-0.2881757617,-0.2525757253,-0.7670419812,0.3175759912,0.0659810528,-0.0736874044,0.148288399,0.2214990407,0.3519730866,0.1975888908,-0.5262952447,-0.1080436409,-0.1927305013,0.2778547108,-0.2042674869,-0.2437518984,0.3360224962,-0.1253866404,0.1910074055,0.0439109616,-0.1972723454,0.2947356105,0.2608200908,0.3492497802,-0.075725235,-0.0599262416,0.0863768682,0.9180575013,0.0366472527,0.1179880798,0.234326914,0.2871668935,0.448636353,-0.1861969531,-0.2464789301,0.0511529483,-0.1345723718,-0.2713452578,0.1046254337,0.3082745969,0.1466263235,0.1631294936,-0.0689219236,0.2187523991,-0.2138775885,0.3311528563,-0.1401932538,0.3264656365,0.0280979425,0.0892387256,-0.3383798301,-0.2278049141,0.0030129068,0.1278017759,0.3850557208,0.2612068057,-0.2789206207,-0.2433248758,-0.4827173948,0.1956553757,-0.0384937972,0.0403399207,-0.3785454333,-0.2253629714,0.1297899038,0.0203341916,0.9234856367,0.0464557074,0.1821256429,-0.2568491399,-0.0968071967,-0.0758301541,0.0022396299,0.2907464802,-0.2174909264,-0.1467928439,0.4098923504,0.0378759801,-0.0872071162,-0.3201644421,0.3639973402,-0.4575732052,0.0065998193,-0.2253907919,-0.0537638292,-0.0625004247,0.1516194493,0.1337892264,0.1230588481,0.0934425071,0.1320100725,-0.1702726483,-0.0427670442,-0.0555934869,0.0200952571,0.5271491408,0.1192736775,0.040874768,0.2363306433,0.2522959709,0.1998598278,0.2740609944,0.0932858884,-0.0258027948,-0.0214497242,0.0511358455,0.118663609,0.4327240884,-0.2288735211,0.2974843085,0.0999590904,0.1264847368,-0.7642996907,0.0768661276,0.2269201428,0.38949579,-0.0471373685,-0.3855696321,0.3176097572,0.0167662092,-0.0140857603,0.2437085509,0.11867439,-0.5224874616,0.0029736734,0.0166976638,1.0220947266,-0.0052735298,0.525090158,0.010605826,0.0838317573,0.4028090239,-0.5466372371,0.358567059,0.057771489,-0.2558864951,-0.0354821868,0.0891322419,-0.0084904106,-0.0728917718,-0.1517244875,0.3895679116,-0.0946481824,0.2141801715,-0.4183894098,-0.0863080621,-0.0117627503,-0.225390628,-0.5132772326,0.0545488559,-0.2742116153,0.1035268828,-0.0508341454,0.0319067053,0.1635920554,-0.0210020393,-0.3381097615,-0.2094533741,-0.2528520525,-0.0258976184,-0.0363922082,-0.3285655081,-0.0165616721,0.2719003558,0.3318611681,0.090514265,-0.3641145229,0.1196023524,-0.1798490584,0.0628120303,-0.2151516676,-0.0983467251,-0.1674999148,-0.0548016615,0.0174991153,0.0707686469,0.0337856375,-0.4108776748,0.0175109897,0.0322574079,-0.133073926,-0.0900963992,-0.2087695301,0.142443493,0.0319382474,-0.0690165684,0.0454461277,0.2176920623,-0.3402993679,0.1204393357,-0.2136463821,-0.2409314513,-0.0380491912,0.4774670005,0.0289599057,-0.2921250165,0.5220627785,-0.0114159333,0.0539456345,-0.1883496344,0.1402017921,0.1701508462,-0.1835626066,-0.0535128005,0.0281142499,-0.2855751216,-0.2057352066,-0.2029029429,0.4924883842,-0.2519830465,-0.3050515354,-0.1449885815,-0.4076443911,-0.0208897814,-0.2762925625,-0.025776986,-0.0961904824,-0.298763752,-0.0198851377,0.191641435,-0.143170774,0.0142275328,-0.0001218162,-0.1673362702,0.1038815528,0.0274556838,0.4854634106,-0.063469477,-0.1909932196,-0.1595243067,-0.0006293667,-0.0392893553,-0.0618141703,0.1540685743,-0.1610926986,-0.0124184377,-0.0185569637,-0.2299521565,0.0444082171,-0.3296545148,0.1573790163,0.2143016011,0.2089236677,-0.0216309913,-0.3265133798,0.0002107329,0.0032439975,0.1196151152,0.2825595737,-0.1058217585,0.0820926502,0.1894634068,-0.0195170864,-0.0718130022,-0.1082033962,0.240028888,0.0303458832,-0.107560128,0.0023846808,-0.0374656357,-0.014393243,0.3485848606,0.1381258965,0.190123722,-0.0317511,-0.0102057382,0.3290903568,0.0324434116,-0.0629224181,-0.3029333055,0.1396771967,-0.1079889536,-0.0974477902,0.3520597816,0.0270169806,-0.2138873786,0.0861029252,0.1473431438,0.4955666959,-0.297668457,0.3628247082,0.1930975467,0.2126588225,0.1036958024,0.2905109823,0.3293461502,0.2367110997,0.5512296557,-0.0336069688,0.1553391069,0.2877417207,0.0097719086,-0.2069424391,-0.3993991315,-0.486281544,0.2589117885,-0.2133095413,0.2273601741,-0.1336095035,0.1467940807,-0.0414488241,-0.6776145101,-0.3064134121,-0.0712345764,-0.243396759,-0.3478887081,0.2802794278,-0.0567444153,0.2844323516,0.2872743011,0.0796922743,-0.2331490219,0.4815829396,-0.0546671972,-0.018017767,0.1753602177,0.0876791477,0.2863029838,0.1103289127,-0.1547991335,0.2614775896,-0.0174217559,-0.1297922432,0.0276955385,0.1919708699,0.2432427853,0.1001448929,-0.0423505977,0.2072948515,0.0817155987,0.1787054688,-0.3429947495,0.1529142559,-0.1525604129,0.0753110126,0.0337519757,-0.0319902077,0.1195744276,0.2038128823,-0.1108585969,0.2348304689,-0.4639111757,-0.049699828,-0.2446763068,0.0765302926,-0.2179709226,0.2800672352,-0.2816495597,0.1942261755,0.5648450851,-0.0442581624,0.3497547209,0.1905335933,-0.0117880618,-0.0112693394,0.7627525926,0.1087859422,-0.3031585813,-0.3050415814,0.0308541004,-0.4886029661,0.080743596,-0.315281868,0.2396721989,-0.2080786824,0.1056509987,0.0669874474,-0.0189331584,-0.0772993714,-0.0975531638,0.0512745306,0.0465316996,-0.3450731933,-0.0946508124,0.1758098304,0.2280649245,0.0866623074,-0.3995473981,0.2115194201,0.1642509699,-0.1586640775,0.4196974635,0.1869817823,-0.4918373823,0.257959336,-0.1285668164,0.1905246526,0.1110507548,0.2345273197,0.0312414672,-0.4896028042,-0.0348943174,-0.2327322215,-0.0441685989,-0.1201228574,0.2055239379,-0.5847114325,0.0387659147,-0.3769586384,0.203678757,0.3197950125,-0.1638128161,0.092655614,-0.0125312367,0.1860568374,0.0903329775,0.4170554578,0.4198202789,-0.0685718283,0.0642155781,-0.3439031839,0.1352642328,0.2230726928,-0.3471212983,-0.3600997329,0.257081151,0.3946217299,0.5234395266,-0.3522129059,-0.3459502459,0.1152452901,0.0333072208,0.0821380839,-0.1448383927,0.2011440843,-0.2863756418,-0.253328681,0.1424975246,0.3100413084,0.3002067804,-0.1675550342,0.2714894712,-0.3209958375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2591","title":"Cached dataset overflowing disk space","comments":"I'm using the datasets concatenate dataset to combine the datasets and then train.\r\ntrain_dataset = concatenate_datasets([dataset1, dataset2, common_voice_train])\r\n\r\n","body":"I'm training a Swedish Wav2vec2 model on a Linux GPU and having issues that the huggingface cached dataset folder is completely filling up my disk space (I'm training on a dataset of around 500 gb).\r\n\r\nThe cache folder is 500gb (and now my disk space is full).\r\n\r\nIs there a way to toggle caching or set the caching to be stored on a different device (I have another drive with 4 tb that could hold the caching files).\r\n\r\nThis might not technically be a bug, but I was unsure and I felt that the bug was the closest one.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 186, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 397, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1983, in _map_single\r\n writer.finalize()\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 418, in finalize\r\n self.pa_writer.close()\r\n File \"pyarrow\/ipc.pxi\", line 402, in pyarrow.lib._CRecordBatchWriter.close\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nOSError: [Errno 28] Error writing bytes to file. Detail: [errno 28] No space left on device\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n","comment_length":18,"text":"Cached dataset overflowing disk space\nI'm training a Swedish Wav2vec2 model on a Linux GPU and having issues that the huggingface cached dataset folder is completely filling up my disk space (I'm training on a dataset of around 500 gb).\r\n\r\nThe cache folder is 500gb (and now my disk space is full).\r\n\r\nIs there a way to toggle caching or set the caching to be stored on a different device (I have another drive with 4 tb that could hold the caching files).\r\n\r\nThis might not technically be a bug, but I was unsure and I felt that the bug was the closest one.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 186, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 397, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1983, in _map_single\r\n writer.finalize()\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 418, in finalize\r\n self.pa_writer.close()\r\n File \"pyarrow\/ipc.pxi\", line 402, in pyarrow.lib._CRecordBatchWriter.close\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nOSError: [Errno 28] Error writing bytes to file. Detail: [errno 28] No space left on device\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\nI'm using the datasets concatenate dataset to combine the datasets and then train.\r\ntrain_dataset = concatenate_datasets([dataset1, dataset2, common_voice_train])\r\n\r\n","embeddings":[0.0375595093,-0.4222126603,0.0985903814,0.4514758587,0.1011737883,0.2050561309,0.0744853392,0.141904518,0.1445289999,0.0556368642,0.4010794461,-0.2280424833,-0.1242998913,0.1127196699,0.1466887593,0.0686893761,0.2591468096,-0.2458593994,-0.0610483401,0.1060899645,-0.1954928488,0.2656945884,0.0236382782,-0.1252904385,-0.2641810179,-0.1312523186,0.1114902273,0.0915549919,0.1169939488,-0.2215207964,0.6455171704,-0.037543118,0.2969840467,0.3342037201,-0.0001271115,-0.2188411504,0.0915412679,-0.181593284,-0.0922594294,0.5504892468,0.1914410889,-0.3445687294,-0.0522951484,0.0128892465,-0.1940163374,0.1353006959,-0.0969077572,-0.4302129745,0.2448510528,0.0724848434,0.0480299257,-0.0077622137,-0.1181686893,0.1708012968,-0.1092367396,-0.0044049905,-0.2483725548,0.1998682469,0.2498452663,-0.1698941886,-0.1137219444,0.5557486415,-0.0590143166,-0.0844287351,0.5253401995,0.0833818987,-0.3562970459,-0.1106624231,0.3813790679,-0.0801356807,0.314463526,-0.4211860299,-0.0359757133,-0.5416691303,0.1242926195,-0.4084812701,0.4917798042,0.1439301074,-0.3081754148,0.2055083513,-0.2643952072,-0.4621683359,-0.2639658749,-0.1211516634,-0.0510151535,-0.1692601889,-0.1779297143,0.0038869162,0.4981890023,-0.1243522316,-0.0180550069,-0.1148280278,-0.1485933661,0.1664890051,-0.4704052806,-0.1601628959,-0.1597560495,0.4838789701,0.1778159291,0.0343691669,0.0580152795,-0.2219776958,0.0135496575,-0.0545078851,0.0384909436,0.6338197589,-0.4302008152,0.1479855031,0.0260615703,-0.1004753411,-0.0664962679,-0.0840651393,-0.0061190045,-0.388866663,0.2431183755,-0.3094467223,0.0366061553,-0.266004473,-0.1329783052,0.1096001938,0.1877570152,0.1407824308,0.2514514029,0.3864438534,0.1821056455,0.3051152825,-0.322846204,0.2218101174,-0.2443430275,-0.1192895547,-0.0440737009,-0.0846645087,0.0494980812,0.3295774758,0.3280156553,-0.1158240363,-0.0438293442,-0.1668728292,0.2368198782,-0.0264211819,0.0309991352,-0.2585927844,0.2868451476,0.1123239547,-0.3598536849,0.225169003,0.2425775528,0.3989651501,-0.2411044687,0.1247824728,-0.3550453782,-0.2518386543,0.1480785757,-0.01426965,-0.1831807494,-0.0969632864,-0.429746151,0.2190823555,0.4013512135,-0.2894096375,0.0460615568,0.0167300683,-0.4555115998,-0.1271671653,0.2183404118,0.390550375,-0.1844455749,-0.0037565609,0.0871429145,0.4355638623,0.0485231653,0.6233537793,-0.1413376033,0.0224382952,-0.3221297562,-0.2130482346,-0.034653984,-0.5695789456,-0.7795040607,0.015956128,0.0507356077,0.2829304338,0.3464263976,0.1854577065,0.0370459333,-0.1173866913,-0.1299520284,0.2160572261,-0.0045058765,-0.0558645725,-0.3231567442,-0.5304920673,0.28534621,-0.1147800237,0.0466500521,0.1006035507,-0.2822220922,0.0011679371,0.1973249465,-0.0316379927,0.2541683018,0.3844375014,0.293066293,-0.3115614951,-0.0663591474,0.0042587491,-0.4575053751,0.1950948834,0.1783739477,-0.3753206432,0.1245837435,-0.1003920138,0.0144068571,-0.2264458388,-0.1339516789,-0.2168265283,-0.1374632716,-0.0339444168,0.4003777504,-0.176329121,-0.2991106212,0.3400560319,0.0025759132,-0.0406318903,-0.3259862363,-0.0154328085,0.1447563022,-0.0170684289,-0.2598402798,0.1888251454,0.0648194179,-0.0355281867,0.0368644558,0.2335698307,0.0226677191,0.3187166452,0.349332124,0.3796356022,0.2128698975,-0.0915223658,0.2820823789,0.2902056575,-0.0620261133,0.0524619706,-0.3751501739,0.2003801614,0.0710650906,0.1950615644,-0.113981016,-0.2368826121,0.251411289,0.1360599846,0.0944542736,-0.1216271147,0.2244897187,0.0017589895,0.392508775,0.0476589166,-0.0416111052,0.2306888103,0.3817815483,0.2312813848,0.0795109794,0.3927400112,-0.1257767677,-0.2583568692,0.1078153104,0.3400515914,0.5201392174,0.1659973711,0.1927121878,-0.08257889,0.0334070846,-0.0248670559,0.1405932903,-0.2922542989,-0.1662439853,0.0999909863,0.1785677075,-0.1703336537,-0.1151010394,0.2065084279,-0.0072104023,0.0374703854,-0.1230290011,0.2915499508,-0.0051883152,0.0669604018,-0.300422281,0.2018046081,-0.4009771645,-0.0063932566,0.1357572228,0.5235956907,0.0959122628,0.0231753308,-0.1968869716,0.7337744236,-0.0781162232,-0.2568981051,0.0806490034,-0.0120087182,-0.1814791262,-0.1487001628,0.3093964458,-0.5017920732,0.097177811,0.1724740267,-0.0851729885,-0.2979379892,0.1057201475,-0.0711285472,0.1724738181,0.0262167249,-0.3023276925,0.3171690702,0.0086356783,-0.0908934399,-0.086091809,-0.1090197414,-0.0805503279,-0.1161442846,0.0500084944,0.4038137794,0.1403172016,0.0388037674,-0.260648638,-0.230347693,0.2758387029,-0.1536940336,-0.0432945602,0.081139639,0.0112992609,-0.1315152347,-0.123419635,-0.2881757617,-0.2525757253,-0.7670419812,0.3175759912,0.0659810528,-0.0736874044,0.148288399,0.2214990407,0.3519730866,0.1975888908,-0.5262952447,-0.1080436409,-0.1927305013,0.2778547108,-0.2042674869,-0.2437518984,0.3360224962,-0.1253866404,0.1910074055,0.0439109616,-0.1972723454,0.2947356105,0.2608200908,0.3492497802,-0.075725235,-0.0599262416,0.0863768682,0.9180575013,0.0366472527,0.1179880798,0.234326914,0.2871668935,0.448636353,-0.1861969531,-0.2464789301,0.0511529483,-0.1345723718,-0.2713452578,0.1046254337,0.3082745969,0.1466263235,0.1631294936,-0.0689219236,0.2187523991,-0.2138775885,0.3311528563,-0.1401932538,0.3264656365,0.0280979425,0.0892387256,-0.3383798301,-0.2278049141,0.0030129068,0.1278017759,0.3850557208,0.2612068057,-0.2789206207,-0.2433248758,-0.4827173948,0.1956553757,-0.0384937972,0.0403399207,-0.3785454333,-0.2253629714,0.1297899038,0.0203341916,0.9234856367,0.0464557074,0.1821256429,-0.2568491399,-0.0968071967,-0.0758301541,0.0022396299,0.2907464802,-0.2174909264,-0.1467928439,0.4098923504,0.0378759801,-0.0872071162,-0.3201644421,0.3639973402,-0.4575732052,0.0065998193,-0.2253907919,-0.0537638292,-0.0625004247,0.1516194493,0.1337892264,0.1230588481,0.0934425071,0.1320100725,-0.1702726483,-0.0427670442,-0.0555934869,0.0200952571,0.5271491408,0.1192736775,0.040874768,0.2363306433,0.2522959709,0.1998598278,0.2740609944,0.0932858884,-0.0258027948,-0.0214497242,0.0511358455,0.118663609,0.4327240884,-0.2288735211,0.2974843085,0.0999590904,0.1264847368,-0.7642996907,0.0768661276,0.2269201428,0.38949579,-0.0471373685,-0.3855696321,0.3176097572,0.0167662092,-0.0140857603,0.2437085509,0.11867439,-0.5224874616,0.0029736734,0.0166976638,1.0220947266,-0.0052735298,0.525090158,0.010605826,0.0838317573,0.4028090239,-0.5466372371,0.358567059,0.057771489,-0.2558864951,-0.0354821868,0.0891322419,-0.0084904106,-0.0728917718,-0.1517244875,0.3895679116,-0.0946481824,0.2141801715,-0.4183894098,-0.0863080621,-0.0117627503,-0.225390628,-0.5132772326,0.0545488559,-0.2742116153,0.1035268828,-0.0508341454,0.0319067053,0.1635920554,-0.0210020393,-0.3381097615,-0.2094533741,-0.2528520525,-0.0258976184,-0.0363922082,-0.3285655081,-0.0165616721,0.2719003558,0.3318611681,0.090514265,-0.3641145229,0.1196023524,-0.1798490584,0.0628120303,-0.2151516676,-0.0983467251,-0.1674999148,-0.0548016615,0.0174991153,0.0707686469,0.0337856375,-0.4108776748,0.0175109897,0.0322574079,-0.133073926,-0.0900963992,-0.2087695301,0.142443493,0.0319382474,-0.0690165684,0.0454461277,0.2176920623,-0.3402993679,0.1204393357,-0.2136463821,-0.2409314513,-0.0380491912,0.4774670005,0.0289599057,-0.2921250165,0.5220627785,-0.0114159333,0.0539456345,-0.1883496344,0.1402017921,0.1701508462,-0.1835626066,-0.0535128005,0.0281142499,-0.2855751216,-0.2057352066,-0.2029029429,0.4924883842,-0.2519830465,-0.3050515354,-0.1449885815,-0.4076443911,-0.0208897814,-0.2762925625,-0.025776986,-0.0961904824,-0.298763752,-0.0198851377,0.191641435,-0.143170774,0.0142275328,-0.0001218162,-0.1673362702,0.1038815528,0.0274556838,0.4854634106,-0.063469477,-0.1909932196,-0.1595243067,-0.0006293667,-0.0392893553,-0.0618141703,0.1540685743,-0.1610926986,-0.0124184377,-0.0185569637,-0.2299521565,0.0444082171,-0.3296545148,0.1573790163,0.2143016011,0.2089236677,-0.0216309913,-0.3265133798,0.0002107329,0.0032439975,0.1196151152,0.2825595737,-0.1058217585,0.0820926502,0.1894634068,-0.0195170864,-0.0718130022,-0.1082033962,0.240028888,0.0303458832,-0.107560128,0.0023846808,-0.0374656357,-0.014393243,0.3485848606,0.1381258965,0.190123722,-0.0317511,-0.0102057382,0.3290903568,0.0324434116,-0.0629224181,-0.3029333055,0.1396771967,-0.1079889536,-0.0974477902,0.3520597816,0.0270169806,-0.2138873786,0.0861029252,0.1473431438,0.4955666959,-0.297668457,0.3628247082,0.1930975467,0.2126588225,0.1036958024,0.2905109823,0.3293461502,0.2367110997,0.5512296557,-0.0336069688,0.1553391069,0.2877417207,0.0097719086,-0.2069424391,-0.3993991315,-0.486281544,0.2589117885,-0.2133095413,0.2273601741,-0.1336095035,0.1467940807,-0.0414488241,-0.6776145101,-0.3064134121,-0.0712345764,-0.243396759,-0.3478887081,0.2802794278,-0.0567444153,0.2844323516,0.2872743011,0.0796922743,-0.2331490219,0.4815829396,-0.0546671972,-0.018017767,0.1753602177,0.0876791477,0.2863029838,0.1103289127,-0.1547991335,0.2614775896,-0.0174217559,-0.1297922432,0.0276955385,0.1919708699,0.2432427853,0.1001448929,-0.0423505977,0.2072948515,0.0817155987,0.1787054688,-0.3429947495,0.1529142559,-0.1525604129,0.0753110126,0.0337519757,-0.0319902077,0.1195744276,0.2038128823,-0.1108585969,0.2348304689,-0.4639111757,-0.049699828,-0.2446763068,0.0765302926,-0.2179709226,0.2800672352,-0.2816495597,0.1942261755,0.5648450851,-0.0442581624,0.3497547209,0.1905335933,-0.0117880618,-0.0112693394,0.7627525926,0.1087859422,-0.3031585813,-0.3050415814,0.0308541004,-0.4886029661,0.080743596,-0.315281868,0.2396721989,-0.2080786824,0.1056509987,0.0669874474,-0.0189331584,-0.0772993714,-0.0975531638,0.0512745306,0.0465316996,-0.3450731933,-0.0946508124,0.1758098304,0.2280649245,0.0866623074,-0.3995473981,0.2115194201,0.1642509699,-0.1586640775,0.4196974635,0.1869817823,-0.4918373823,0.257959336,-0.1285668164,0.1905246526,0.1110507548,0.2345273197,0.0312414672,-0.4896028042,-0.0348943174,-0.2327322215,-0.0441685989,-0.1201228574,0.2055239379,-0.5847114325,0.0387659147,-0.3769586384,0.203678757,0.3197950125,-0.1638128161,0.092655614,-0.0125312367,0.1860568374,0.0903329775,0.4170554578,0.4198202789,-0.0685718283,0.0642155781,-0.3439031839,0.1352642328,0.2230726928,-0.3471212983,-0.3600997329,0.257081151,0.3946217299,0.5234395266,-0.3522129059,-0.3459502459,0.1152452901,0.0333072208,0.0821380839,-0.1448383927,0.2011440843,-0.2863756418,-0.253328681,0.1424975246,0.3100413084,0.3002067804,-0.1675550342,0.2714894712,-0.3209958375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2591","title":"Cached dataset overflowing disk space","comments":"Hi @BirgerMoell.\r\n\r\nYou have several options:\r\n- to set caching to be stored on a different path location, other than the default one (`~\/.cache\/huggingface\/datasets`):\r\n - either setting the environment variable `HF_DATASETS_CACHE` with the path to the new cache location\r\n - or by passing it with the parameter `cache_dir` when loading each of the datasets: `dataset = load_dataset(..., cache_dir=your_new_location)`\r\n\r\n You can get all the information in the docs: https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#cache-directory\r\n- I wouldn't recommend disabling caching, because current implementation generates cache files anyway, although in a temporary directory and they are deleted when the session closes. See details here: https:\/\/huggingface.co\/docs\/datasets\/processing.html#enable-or-disable-caching\r\n- You could alternatively load the datasets in streaming mode. This is a new feature which allows loading the datasets without downloading the entire files. More information here: https:\/\/huggingface.co\/docs\/datasets\/dataset_streaming.html","body":"I'm training a Swedish Wav2vec2 model on a Linux GPU and having issues that the huggingface cached dataset folder is completely filling up my disk space (I'm training on a dataset of around 500 gb).\r\n\r\nThe cache folder is 500gb (and now my disk space is full).\r\n\r\nIs there a way to toggle caching or set the caching to be stored on a different device (I have another drive with 4 tb that could hold the caching files).\r\n\r\nThis might not technically be a bug, but I was unsure and I felt that the bug was the closest one.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 186, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 397, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1983, in _map_single\r\n writer.finalize()\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 418, in finalize\r\n self.pa_writer.close()\r\n File \"pyarrow\/ipc.pxi\", line 402, in pyarrow.lib._CRecordBatchWriter.close\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nOSError: [Errno 28] Error writing bytes to file. Detail: [errno 28] No space left on device\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n","comment_length":127,"text":"Cached dataset overflowing disk space\nI'm training a Swedish Wav2vec2 model on a Linux GPU and having issues that the huggingface cached dataset folder is completely filling up my disk space (I'm training on a dataset of around 500 gb).\r\n\r\nThe cache folder is 500gb (and now my disk space is full).\r\n\r\nIs there a way to toggle caching or set the caching to be stored on a different device (I have another drive with 4 tb that could hold the caching files).\r\n\r\nThis might not technically be a bug, but I was unsure and I felt that the bug was the closest one.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 186, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 397, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1983, in _map_single\r\n writer.finalize()\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 418, in finalize\r\n self.pa_writer.close()\r\n File \"pyarrow\/ipc.pxi\", line 402, in pyarrow.lib._CRecordBatchWriter.close\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nOSError: [Errno 28] Error writing bytes to file. Detail: [errno 28] No space left on device\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\nHi @BirgerMoell.\r\n\r\nYou have several options:\r\n- to set caching to be stored on a different path location, other than the default one (`~\/.cache\/huggingface\/datasets`):\r\n - either setting the environment variable `HF_DATASETS_CACHE` with the path to the new cache location\r\n - or by passing it with the parameter `cache_dir` when loading each of the datasets: `dataset = load_dataset(..., cache_dir=your_new_location)`\r\n\r\n You can get all the information in the docs: https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#cache-directory\r\n- I wouldn't recommend disabling caching, because current implementation generates cache files anyway, although in a temporary directory and they are deleted when the session closes. See details here: https:\/\/huggingface.co\/docs\/datasets\/processing.html#enable-or-disable-caching\r\n- You could alternatively load the datasets in streaming mode. This is a new feature which allows loading the datasets without downloading the entire files. More information here: https:\/\/huggingface.co\/docs\/datasets\/dataset_streaming.html","embeddings":[0.0375595093,-0.4222126603,0.0985903814,0.4514758587,0.1011737883,0.2050561309,0.0744853392,0.141904518,0.1445289999,0.0556368642,0.4010794461,-0.2280424833,-0.1242998913,0.1127196699,0.1466887593,0.0686893761,0.2591468096,-0.2458593994,-0.0610483401,0.1060899645,-0.1954928488,0.2656945884,0.0236382782,-0.1252904385,-0.2641810179,-0.1312523186,0.1114902273,0.0915549919,0.1169939488,-0.2215207964,0.6455171704,-0.037543118,0.2969840467,0.3342037201,-0.0001271115,-0.2188411504,0.0915412679,-0.181593284,-0.0922594294,0.5504892468,0.1914410889,-0.3445687294,-0.0522951484,0.0128892465,-0.1940163374,0.1353006959,-0.0969077572,-0.4302129745,0.2448510528,0.0724848434,0.0480299257,-0.0077622137,-0.1181686893,0.1708012968,-0.1092367396,-0.0044049905,-0.2483725548,0.1998682469,0.2498452663,-0.1698941886,-0.1137219444,0.5557486415,-0.0590143166,-0.0844287351,0.5253401995,0.0833818987,-0.3562970459,-0.1106624231,0.3813790679,-0.0801356807,0.314463526,-0.4211860299,-0.0359757133,-0.5416691303,0.1242926195,-0.4084812701,0.4917798042,0.1439301074,-0.3081754148,0.2055083513,-0.2643952072,-0.4621683359,-0.2639658749,-0.1211516634,-0.0510151535,-0.1692601889,-0.1779297143,0.0038869162,0.4981890023,-0.1243522316,-0.0180550069,-0.1148280278,-0.1485933661,0.1664890051,-0.4704052806,-0.1601628959,-0.1597560495,0.4838789701,0.1778159291,0.0343691669,0.0580152795,-0.2219776958,0.0135496575,-0.0545078851,0.0384909436,0.6338197589,-0.4302008152,0.1479855031,0.0260615703,-0.1004753411,-0.0664962679,-0.0840651393,-0.0061190045,-0.388866663,0.2431183755,-0.3094467223,0.0366061553,-0.266004473,-0.1329783052,0.1096001938,0.1877570152,0.1407824308,0.2514514029,0.3864438534,0.1821056455,0.3051152825,-0.322846204,0.2218101174,-0.2443430275,-0.1192895547,-0.0440737009,-0.0846645087,0.0494980812,0.3295774758,0.3280156553,-0.1158240363,-0.0438293442,-0.1668728292,0.2368198782,-0.0264211819,0.0309991352,-0.2585927844,0.2868451476,0.1123239547,-0.3598536849,0.225169003,0.2425775528,0.3989651501,-0.2411044687,0.1247824728,-0.3550453782,-0.2518386543,0.1480785757,-0.01426965,-0.1831807494,-0.0969632864,-0.429746151,0.2190823555,0.4013512135,-0.2894096375,0.0460615568,0.0167300683,-0.4555115998,-0.1271671653,0.2183404118,0.390550375,-0.1844455749,-0.0037565609,0.0871429145,0.4355638623,0.0485231653,0.6233537793,-0.1413376033,0.0224382952,-0.3221297562,-0.2130482346,-0.034653984,-0.5695789456,-0.7795040607,0.015956128,0.0507356077,0.2829304338,0.3464263976,0.1854577065,0.0370459333,-0.1173866913,-0.1299520284,0.2160572261,-0.0045058765,-0.0558645725,-0.3231567442,-0.5304920673,0.28534621,-0.1147800237,0.0466500521,0.1006035507,-0.2822220922,0.0011679371,0.1973249465,-0.0316379927,0.2541683018,0.3844375014,0.293066293,-0.3115614951,-0.0663591474,0.0042587491,-0.4575053751,0.1950948834,0.1783739477,-0.3753206432,0.1245837435,-0.1003920138,0.0144068571,-0.2264458388,-0.1339516789,-0.2168265283,-0.1374632716,-0.0339444168,0.4003777504,-0.176329121,-0.2991106212,0.3400560319,0.0025759132,-0.0406318903,-0.3259862363,-0.0154328085,0.1447563022,-0.0170684289,-0.2598402798,0.1888251454,0.0648194179,-0.0355281867,0.0368644558,0.2335698307,0.0226677191,0.3187166452,0.349332124,0.3796356022,0.2128698975,-0.0915223658,0.2820823789,0.2902056575,-0.0620261133,0.0524619706,-0.3751501739,0.2003801614,0.0710650906,0.1950615644,-0.113981016,-0.2368826121,0.251411289,0.1360599846,0.0944542736,-0.1216271147,0.2244897187,0.0017589895,0.392508775,0.0476589166,-0.0416111052,0.2306888103,0.3817815483,0.2312813848,0.0795109794,0.3927400112,-0.1257767677,-0.2583568692,0.1078153104,0.3400515914,0.5201392174,0.1659973711,0.1927121878,-0.08257889,0.0334070846,-0.0248670559,0.1405932903,-0.2922542989,-0.1662439853,0.0999909863,0.1785677075,-0.1703336537,-0.1151010394,0.2065084279,-0.0072104023,0.0374703854,-0.1230290011,0.2915499508,-0.0051883152,0.0669604018,-0.300422281,0.2018046081,-0.4009771645,-0.0063932566,0.1357572228,0.5235956907,0.0959122628,0.0231753308,-0.1968869716,0.7337744236,-0.0781162232,-0.2568981051,0.0806490034,-0.0120087182,-0.1814791262,-0.1487001628,0.3093964458,-0.5017920732,0.097177811,0.1724740267,-0.0851729885,-0.2979379892,0.1057201475,-0.0711285472,0.1724738181,0.0262167249,-0.3023276925,0.3171690702,0.0086356783,-0.0908934399,-0.086091809,-0.1090197414,-0.0805503279,-0.1161442846,0.0500084944,0.4038137794,0.1403172016,0.0388037674,-0.260648638,-0.230347693,0.2758387029,-0.1536940336,-0.0432945602,0.081139639,0.0112992609,-0.1315152347,-0.123419635,-0.2881757617,-0.2525757253,-0.7670419812,0.3175759912,0.0659810528,-0.0736874044,0.148288399,0.2214990407,0.3519730866,0.1975888908,-0.5262952447,-0.1080436409,-0.1927305013,0.2778547108,-0.2042674869,-0.2437518984,0.3360224962,-0.1253866404,0.1910074055,0.0439109616,-0.1972723454,0.2947356105,0.2608200908,0.3492497802,-0.075725235,-0.0599262416,0.0863768682,0.9180575013,0.0366472527,0.1179880798,0.234326914,0.2871668935,0.448636353,-0.1861969531,-0.2464789301,0.0511529483,-0.1345723718,-0.2713452578,0.1046254337,0.3082745969,0.1466263235,0.1631294936,-0.0689219236,0.2187523991,-0.2138775885,0.3311528563,-0.1401932538,0.3264656365,0.0280979425,0.0892387256,-0.3383798301,-0.2278049141,0.0030129068,0.1278017759,0.3850557208,0.2612068057,-0.2789206207,-0.2433248758,-0.4827173948,0.1956553757,-0.0384937972,0.0403399207,-0.3785454333,-0.2253629714,0.1297899038,0.0203341916,0.9234856367,0.0464557074,0.1821256429,-0.2568491399,-0.0968071967,-0.0758301541,0.0022396299,0.2907464802,-0.2174909264,-0.1467928439,0.4098923504,0.0378759801,-0.0872071162,-0.3201644421,0.3639973402,-0.4575732052,0.0065998193,-0.2253907919,-0.0537638292,-0.0625004247,0.1516194493,0.1337892264,0.1230588481,0.0934425071,0.1320100725,-0.1702726483,-0.0427670442,-0.0555934869,0.0200952571,0.5271491408,0.1192736775,0.040874768,0.2363306433,0.2522959709,0.1998598278,0.2740609944,0.0932858884,-0.0258027948,-0.0214497242,0.0511358455,0.118663609,0.4327240884,-0.2288735211,0.2974843085,0.0999590904,0.1264847368,-0.7642996907,0.0768661276,0.2269201428,0.38949579,-0.0471373685,-0.3855696321,0.3176097572,0.0167662092,-0.0140857603,0.2437085509,0.11867439,-0.5224874616,0.0029736734,0.0166976638,1.0220947266,-0.0052735298,0.525090158,0.010605826,0.0838317573,0.4028090239,-0.5466372371,0.358567059,0.057771489,-0.2558864951,-0.0354821868,0.0891322419,-0.0084904106,-0.0728917718,-0.1517244875,0.3895679116,-0.0946481824,0.2141801715,-0.4183894098,-0.0863080621,-0.0117627503,-0.225390628,-0.5132772326,0.0545488559,-0.2742116153,0.1035268828,-0.0508341454,0.0319067053,0.1635920554,-0.0210020393,-0.3381097615,-0.2094533741,-0.2528520525,-0.0258976184,-0.0363922082,-0.3285655081,-0.0165616721,0.2719003558,0.3318611681,0.090514265,-0.3641145229,0.1196023524,-0.1798490584,0.0628120303,-0.2151516676,-0.0983467251,-0.1674999148,-0.0548016615,0.0174991153,0.0707686469,0.0337856375,-0.4108776748,0.0175109897,0.0322574079,-0.133073926,-0.0900963992,-0.2087695301,0.142443493,0.0319382474,-0.0690165684,0.0454461277,0.2176920623,-0.3402993679,0.1204393357,-0.2136463821,-0.2409314513,-0.0380491912,0.4774670005,0.0289599057,-0.2921250165,0.5220627785,-0.0114159333,0.0539456345,-0.1883496344,0.1402017921,0.1701508462,-0.1835626066,-0.0535128005,0.0281142499,-0.2855751216,-0.2057352066,-0.2029029429,0.4924883842,-0.2519830465,-0.3050515354,-0.1449885815,-0.4076443911,-0.0208897814,-0.2762925625,-0.025776986,-0.0961904824,-0.298763752,-0.0198851377,0.191641435,-0.143170774,0.0142275328,-0.0001218162,-0.1673362702,0.1038815528,0.0274556838,0.4854634106,-0.063469477,-0.1909932196,-0.1595243067,-0.0006293667,-0.0392893553,-0.0618141703,0.1540685743,-0.1610926986,-0.0124184377,-0.0185569637,-0.2299521565,0.0444082171,-0.3296545148,0.1573790163,0.2143016011,0.2089236677,-0.0216309913,-0.3265133798,0.0002107329,0.0032439975,0.1196151152,0.2825595737,-0.1058217585,0.0820926502,0.1894634068,-0.0195170864,-0.0718130022,-0.1082033962,0.240028888,0.0303458832,-0.107560128,0.0023846808,-0.0374656357,-0.014393243,0.3485848606,0.1381258965,0.190123722,-0.0317511,-0.0102057382,0.3290903568,0.0324434116,-0.0629224181,-0.3029333055,0.1396771967,-0.1079889536,-0.0974477902,0.3520597816,0.0270169806,-0.2138873786,0.0861029252,0.1473431438,0.4955666959,-0.297668457,0.3628247082,0.1930975467,0.2126588225,0.1036958024,0.2905109823,0.3293461502,0.2367110997,0.5512296557,-0.0336069688,0.1553391069,0.2877417207,0.0097719086,-0.2069424391,-0.3993991315,-0.486281544,0.2589117885,-0.2133095413,0.2273601741,-0.1336095035,0.1467940807,-0.0414488241,-0.6776145101,-0.3064134121,-0.0712345764,-0.243396759,-0.3478887081,0.2802794278,-0.0567444153,0.2844323516,0.2872743011,0.0796922743,-0.2331490219,0.4815829396,-0.0546671972,-0.018017767,0.1753602177,0.0876791477,0.2863029838,0.1103289127,-0.1547991335,0.2614775896,-0.0174217559,-0.1297922432,0.0276955385,0.1919708699,0.2432427853,0.1001448929,-0.0423505977,0.2072948515,0.0817155987,0.1787054688,-0.3429947495,0.1529142559,-0.1525604129,0.0753110126,0.0337519757,-0.0319902077,0.1195744276,0.2038128823,-0.1108585969,0.2348304689,-0.4639111757,-0.049699828,-0.2446763068,0.0765302926,-0.2179709226,0.2800672352,-0.2816495597,0.1942261755,0.5648450851,-0.0442581624,0.3497547209,0.1905335933,-0.0117880618,-0.0112693394,0.7627525926,0.1087859422,-0.3031585813,-0.3050415814,0.0308541004,-0.4886029661,0.080743596,-0.315281868,0.2396721989,-0.2080786824,0.1056509987,0.0669874474,-0.0189331584,-0.0772993714,-0.0975531638,0.0512745306,0.0465316996,-0.3450731933,-0.0946508124,0.1758098304,0.2280649245,0.0866623074,-0.3995473981,0.2115194201,0.1642509699,-0.1586640775,0.4196974635,0.1869817823,-0.4918373823,0.257959336,-0.1285668164,0.1905246526,0.1110507548,0.2345273197,0.0312414672,-0.4896028042,-0.0348943174,-0.2327322215,-0.0441685989,-0.1201228574,0.2055239379,-0.5847114325,0.0387659147,-0.3769586384,0.203678757,0.3197950125,-0.1638128161,0.092655614,-0.0125312367,0.1860568374,0.0903329775,0.4170554578,0.4198202789,-0.0685718283,0.0642155781,-0.3439031839,0.1352642328,0.2230726928,-0.3471212983,-0.3600997329,0.257081151,0.3946217299,0.5234395266,-0.3522129059,-0.3459502459,0.1152452901,0.0333072208,0.0821380839,-0.1448383927,0.2011440843,-0.2863756418,-0.253328681,0.1424975246,0.3100413084,0.3002067804,-0.1675550342,0.2714894712,-0.3209958375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2591","title":"Cached dataset overflowing disk space","comments":"Hi @BirgerMoell,\r\n\r\nWe are planning to add a new feature to datasets, which could be interesting in your case: Add the option to delete temporary files (decompressed files) from the cache directory (see: #2481, #2604).\r\n\r\nWe will ping you once this feature is implemented, so that the size of your cache directory will be considerably reduced.","body":"I'm training a Swedish Wav2vec2 model on a Linux GPU and having issues that the huggingface cached dataset folder is completely filling up my disk space (I'm training on a dataset of around 500 gb).\r\n\r\nThe cache folder is 500gb (and now my disk space is full).\r\n\r\nIs there a way to toggle caching or set the caching to be stored on a different device (I have another drive with 4 tb that could hold the caching files).\r\n\r\nThis might not technically be a bug, but I was unsure and I felt that the bug was the closest one.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 186, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 397, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1983, in _map_single\r\n writer.finalize()\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 418, in finalize\r\n self.pa_writer.close()\r\n File \"pyarrow\/ipc.pxi\", line 402, in pyarrow.lib._CRecordBatchWriter.close\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nOSError: [Errno 28] Error writing bytes to file. Detail: [errno 28] No space left on device\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n","comment_length":56,"text":"Cached dataset overflowing disk space\nI'm training a Swedish Wav2vec2 model on a Linux GPU and having issues that the huggingface cached dataset folder is completely filling up my disk space (I'm training on a dataset of around 500 gb).\r\n\r\nThe cache folder is 500gb (and now my disk space is full).\r\n\r\nIs there a way to toggle caching or set the caching to be stored on a different device (I have another drive with 4 tb that could hold the caching files).\r\n\r\nThis might not technically be a bug, but I was unsure and I felt that the bug was the closest one.\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 186, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 397, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1983, in _map_single\r\n writer.finalize()\r\n File \"\/home\/birger\/miniconda3\/envs\/wav2vec2\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 418, in finalize\r\n self.pa_writer.close()\r\n File \"pyarrow\/ipc.pxi\", line 402, in pyarrow.lib._CRecordBatchWriter.close\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nOSError: [Errno 28] Error writing bytes to file. Detail: [errno 28] No space left on device\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\nHi @BirgerMoell,\r\n\r\nWe are planning to add a new feature to datasets, which could be interesting in your case: Add the option to delete temporary files (decompressed files) from the cache directory (see: #2481, #2604).\r\n\r\nWe will ping you once this feature is implemented, so that the size of your cache directory will be considerably reduced.","embeddings":[0.0375595093,-0.4222126603,0.0985903814,0.4514758587,0.1011737883,0.2050561309,0.0744853392,0.141904518,0.1445289999,0.0556368642,0.4010794461,-0.2280424833,-0.1242998913,0.1127196699,0.1466887593,0.0686893761,0.2591468096,-0.2458593994,-0.0610483401,0.1060899645,-0.1954928488,0.2656945884,0.0236382782,-0.1252904385,-0.2641810179,-0.1312523186,0.1114902273,0.0915549919,0.1169939488,-0.2215207964,0.6455171704,-0.037543118,0.2969840467,0.3342037201,-0.0001271115,-0.2188411504,0.0915412679,-0.181593284,-0.0922594294,0.5504892468,0.1914410889,-0.3445687294,-0.0522951484,0.0128892465,-0.1940163374,0.1353006959,-0.0969077572,-0.4302129745,0.2448510528,0.0724848434,0.0480299257,-0.0077622137,-0.1181686893,0.1708012968,-0.1092367396,-0.0044049905,-0.2483725548,0.1998682469,0.2498452663,-0.1698941886,-0.1137219444,0.5557486415,-0.0590143166,-0.0844287351,0.5253401995,0.0833818987,-0.3562970459,-0.1106624231,0.3813790679,-0.0801356807,0.314463526,-0.4211860299,-0.0359757133,-0.5416691303,0.1242926195,-0.4084812701,0.4917798042,0.1439301074,-0.3081754148,0.2055083513,-0.2643952072,-0.4621683359,-0.2639658749,-0.1211516634,-0.0510151535,-0.1692601889,-0.1779297143,0.0038869162,0.4981890023,-0.1243522316,-0.0180550069,-0.1148280278,-0.1485933661,0.1664890051,-0.4704052806,-0.1601628959,-0.1597560495,0.4838789701,0.1778159291,0.0343691669,0.0580152795,-0.2219776958,0.0135496575,-0.0545078851,0.0384909436,0.6338197589,-0.4302008152,0.1479855031,0.0260615703,-0.1004753411,-0.0664962679,-0.0840651393,-0.0061190045,-0.388866663,0.2431183755,-0.3094467223,0.0366061553,-0.266004473,-0.1329783052,0.1096001938,0.1877570152,0.1407824308,0.2514514029,0.3864438534,0.1821056455,0.3051152825,-0.322846204,0.2218101174,-0.2443430275,-0.1192895547,-0.0440737009,-0.0846645087,0.0494980812,0.3295774758,0.3280156553,-0.1158240363,-0.0438293442,-0.1668728292,0.2368198782,-0.0264211819,0.0309991352,-0.2585927844,0.2868451476,0.1123239547,-0.3598536849,0.225169003,0.2425775528,0.3989651501,-0.2411044687,0.1247824728,-0.3550453782,-0.2518386543,0.1480785757,-0.01426965,-0.1831807494,-0.0969632864,-0.429746151,0.2190823555,0.4013512135,-0.2894096375,0.0460615568,0.0167300683,-0.4555115998,-0.1271671653,0.2183404118,0.390550375,-0.1844455749,-0.0037565609,0.0871429145,0.4355638623,0.0485231653,0.6233537793,-0.1413376033,0.0224382952,-0.3221297562,-0.2130482346,-0.034653984,-0.5695789456,-0.7795040607,0.015956128,0.0507356077,0.2829304338,0.3464263976,0.1854577065,0.0370459333,-0.1173866913,-0.1299520284,0.2160572261,-0.0045058765,-0.0558645725,-0.3231567442,-0.5304920673,0.28534621,-0.1147800237,0.0466500521,0.1006035507,-0.2822220922,0.0011679371,0.1973249465,-0.0316379927,0.2541683018,0.3844375014,0.293066293,-0.3115614951,-0.0663591474,0.0042587491,-0.4575053751,0.1950948834,0.1783739477,-0.3753206432,0.1245837435,-0.1003920138,0.0144068571,-0.2264458388,-0.1339516789,-0.2168265283,-0.1374632716,-0.0339444168,0.4003777504,-0.176329121,-0.2991106212,0.3400560319,0.0025759132,-0.0406318903,-0.3259862363,-0.0154328085,0.1447563022,-0.0170684289,-0.2598402798,0.1888251454,0.0648194179,-0.0355281867,0.0368644558,0.2335698307,0.0226677191,0.3187166452,0.349332124,0.3796356022,0.2128698975,-0.0915223658,0.2820823789,0.2902056575,-0.0620261133,0.0524619706,-0.3751501739,0.2003801614,0.0710650906,0.1950615644,-0.113981016,-0.2368826121,0.251411289,0.1360599846,0.0944542736,-0.1216271147,0.2244897187,0.0017589895,0.392508775,0.0476589166,-0.0416111052,0.2306888103,0.3817815483,0.2312813848,0.0795109794,0.3927400112,-0.1257767677,-0.2583568692,0.1078153104,0.3400515914,0.5201392174,0.1659973711,0.1927121878,-0.08257889,0.0334070846,-0.0248670559,0.1405932903,-0.2922542989,-0.1662439853,0.0999909863,0.1785677075,-0.1703336537,-0.1151010394,0.2065084279,-0.0072104023,0.0374703854,-0.1230290011,0.2915499508,-0.0051883152,0.0669604018,-0.300422281,0.2018046081,-0.4009771645,-0.0063932566,0.1357572228,0.5235956907,0.0959122628,0.0231753308,-0.1968869716,0.7337744236,-0.0781162232,-0.2568981051,0.0806490034,-0.0120087182,-0.1814791262,-0.1487001628,0.3093964458,-0.5017920732,0.097177811,0.1724740267,-0.0851729885,-0.2979379892,0.1057201475,-0.0711285472,0.1724738181,0.0262167249,-0.3023276925,0.3171690702,0.0086356783,-0.0908934399,-0.086091809,-0.1090197414,-0.0805503279,-0.1161442846,0.0500084944,0.4038137794,0.1403172016,0.0388037674,-0.260648638,-0.230347693,0.2758387029,-0.1536940336,-0.0432945602,0.081139639,0.0112992609,-0.1315152347,-0.123419635,-0.2881757617,-0.2525757253,-0.7670419812,0.3175759912,0.0659810528,-0.0736874044,0.148288399,0.2214990407,0.3519730866,0.1975888908,-0.5262952447,-0.1080436409,-0.1927305013,0.2778547108,-0.2042674869,-0.2437518984,0.3360224962,-0.1253866404,0.1910074055,0.0439109616,-0.1972723454,0.2947356105,0.2608200908,0.3492497802,-0.075725235,-0.0599262416,0.0863768682,0.9180575013,0.0366472527,0.1179880798,0.234326914,0.2871668935,0.448636353,-0.1861969531,-0.2464789301,0.0511529483,-0.1345723718,-0.2713452578,0.1046254337,0.3082745969,0.1466263235,0.1631294936,-0.0689219236,0.2187523991,-0.2138775885,0.3311528563,-0.1401932538,0.3264656365,0.0280979425,0.0892387256,-0.3383798301,-0.2278049141,0.0030129068,0.1278017759,0.3850557208,0.2612068057,-0.2789206207,-0.2433248758,-0.4827173948,0.1956553757,-0.0384937972,0.0403399207,-0.3785454333,-0.2253629714,0.1297899038,0.0203341916,0.9234856367,0.0464557074,0.1821256429,-0.2568491399,-0.0968071967,-0.0758301541,0.0022396299,0.2907464802,-0.2174909264,-0.1467928439,0.4098923504,0.0378759801,-0.0872071162,-0.3201644421,0.3639973402,-0.4575732052,0.0065998193,-0.2253907919,-0.0537638292,-0.0625004247,0.1516194493,0.1337892264,0.1230588481,0.0934425071,0.1320100725,-0.1702726483,-0.0427670442,-0.0555934869,0.0200952571,0.5271491408,0.1192736775,0.040874768,0.2363306433,0.2522959709,0.1998598278,0.2740609944,0.0932858884,-0.0258027948,-0.0214497242,0.0511358455,0.118663609,0.4327240884,-0.2288735211,0.2974843085,0.0999590904,0.1264847368,-0.7642996907,0.0768661276,0.2269201428,0.38949579,-0.0471373685,-0.3855696321,0.3176097572,0.0167662092,-0.0140857603,0.2437085509,0.11867439,-0.5224874616,0.0029736734,0.0166976638,1.0220947266,-0.0052735298,0.525090158,0.010605826,0.0838317573,0.4028090239,-0.5466372371,0.358567059,0.057771489,-0.2558864951,-0.0354821868,0.0891322419,-0.0084904106,-0.0728917718,-0.1517244875,0.3895679116,-0.0946481824,0.2141801715,-0.4183894098,-0.0863080621,-0.0117627503,-0.225390628,-0.5132772326,0.0545488559,-0.2742116153,0.1035268828,-0.0508341454,0.0319067053,0.1635920554,-0.0210020393,-0.3381097615,-0.2094533741,-0.2528520525,-0.0258976184,-0.0363922082,-0.3285655081,-0.0165616721,0.2719003558,0.3318611681,0.090514265,-0.3641145229,0.1196023524,-0.1798490584,0.0628120303,-0.2151516676,-0.0983467251,-0.1674999148,-0.0548016615,0.0174991153,0.0707686469,0.0337856375,-0.4108776748,0.0175109897,0.0322574079,-0.133073926,-0.0900963992,-0.2087695301,0.142443493,0.0319382474,-0.0690165684,0.0454461277,0.2176920623,-0.3402993679,0.1204393357,-0.2136463821,-0.2409314513,-0.0380491912,0.4774670005,0.0289599057,-0.2921250165,0.5220627785,-0.0114159333,0.0539456345,-0.1883496344,0.1402017921,0.1701508462,-0.1835626066,-0.0535128005,0.0281142499,-0.2855751216,-0.2057352066,-0.2029029429,0.4924883842,-0.2519830465,-0.3050515354,-0.1449885815,-0.4076443911,-0.0208897814,-0.2762925625,-0.025776986,-0.0961904824,-0.298763752,-0.0198851377,0.191641435,-0.143170774,0.0142275328,-0.0001218162,-0.1673362702,0.1038815528,0.0274556838,0.4854634106,-0.063469477,-0.1909932196,-0.1595243067,-0.0006293667,-0.0392893553,-0.0618141703,0.1540685743,-0.1610926986,-0.0124184377,-0.0185569637,-0.2299521565,0.0444082171,-0.3296545148,0.1573790163,0.2143016011,0.2089236677,-0.0216309913,-0.3265133798,0.0002107329,0.0032439975,0.1196151152,0.2825595737,-0.1058217585,0.0820926502,0.1894634068,-0.0195170864,-0.0718130022,-0.1082033962,0.240028888,0.0303458832,-0.107560128,0.0023846808,-0.0374656357,-0.014393243,0.3485848606,0.1381258965,0.190123722,-0.0317511,-0.0102057382,0.3290903568,0.0324434116,-0.0629224181,-0.3029333055,0.1396771967,-0.1079889536,-0.0974477902,0.3520597816,0.0270169806,-0.2138873786,0.0861029252,0.1473431438,0.4955666959,-0.297668457,0.3628247082,0.1930975467,0.2126588225,0.1036958024,0.2905109823,0.3293461502,0.2367110997,0.5512296557,-0.0336069688,0.1553391069,0.2877417207,0.0097719086,-0.2069424391,-0.3993991315,-0.486281544,0.2589117885,-0.2133095413,0.2273601741,-0.1336095035,0.1467940807,-0.0414488241,-0.6776145101,-0.3064134121,-0.0712345764,-0.243396759,-0.3478887081,0.2802794278,-0.0567444153,0.2844323516,0.2872743011,0.0796922743,-0.2331490219,0.4815829396,-0.0546671972,-0.018017767,0.1753602177,0.0876791477,0.2863029838,0.1103289127,-0.1547991335,0.2614775896,-0.0174217559,-0.1297922432,0.0276955385,0.1919708699,0.2432427853,0.1001448929,-0.0423505977,0.2072948515,0.0817155987,0.1787054688,-0.3429947495,0.1529142559,-0.1525604129,0.0753110126,0.0337519757,-0.0319902077,0.1195744276,0.2038128823,-0.1108585969,0.2348304689,-0.4639111757,-0.049699828,-0.2446763068,0.0765302926,-0.2179709226,0.2800672352,-0.2816495597,0.1942261755,0.5648450851,-0.0442581624,0.3497547209,0.1905335933,-0.0117880618,-0.0112693394,0.7627525926,0.1087859422,-0.3031585813,-0.3050415814,0.0308541004,-0.4886029661,0.080743596,-0.315281868,0.2396721989,-0.2080786824,0.1056509987,0.0669874474,-0.0189331584,-0.0772993714,-0.0975531638,0.0512745306,0.0465316996,-0.3450731933,-0.0946508124,0.1758098304,0.2280649245,0.0866623074,-0.3995473981,0.2115194201,0.1642509699,-0.1586640775,0.4196974635,0.1869817823,-0.4918373823,0.257959336,-0.1285668164,0.1905246526,0.1110507548,0.2345273197,0.0312414672,-0.4896028042,-0.0348943174,-0.2327322215,-0.0441685989,-0.1201228574,0.2055239379,-0.5847114325,0.0387659147,-0.3769586384,0.203678757,0.3197950125,-0.1638128161,0.092655614,-0.0125312367,0.1860568374,0.0903329775,0.4170554578,0.4198202789,-0.0685718283,0.0642155781,-0.3439031839,0.1352642328,0.2230726928,-0.3471212983,-0.3600997329,0.257081151,0.3946217299,0.5234395266,-0.3522129059,-0.3459502459,0.1152452901,0.0333072208,0.0821380839,-0.1448383927,0.2011440843,-0.2863756418,-0.253328681,0.1424975246,0.3100413084,0.3002067804,-0.1675550342,0.2714894712,-0.3209958375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2585","title":"sqaud_v2 dataset contains misalignment between the answer text and the context value at the answer index","comments":"Hi @mmajurski, thanks for reporting this issue.\r\n\r\nIndeed this misalignment arises because the source dataset context field contains leading blank spaces (and these are counted within the answer_start), while our datasets loading script removes these leading blank spaces.\r\n\r\nI'm going to fix our script so that all leading blank spaces in the source dataset are kept, and there is no misalignment between the answer text and the answer_start within the context.","body":"## Describe the bug\r\nThe built in huggingface squad_v2 dataset that you can access via datasets.load_dataset contains mis-alignment between the answers['text'] and the characters in the context at the location specified by answers['answer_start'].\r\n\r\nFor example:\r\nid = '56d1f453e7d4791d009025bd'\r\nanswers = {'text': ['Pure Land'], 'answer_start': [146]}\r\nHowever the actual text in context at location 146 is 'ure Land,'\r\nWhich is an off-by-one error from the correct answer.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\ndef check_context_answer_alignment(example):\r\n for a_idx in range(len(example['answers']['text'])):\r\n # check raw dataset for answer consistency between context and answer\r\n answer_text = example['answers']['text'][a_idx]\r\n a_st_idx = example['answers']['answer_start'][a_idx]\r\n a_end_idx = a_st_idx + len(example['answers']['text'][a_idx])\r\n answer_text_from_context = example['context'][a_st_idx:a_end_idx]\r\n if answer_text != answer_text_from_context:\r\n #print(example['id'])\r\n return False\r\n return True\r\n\r\ndataset = datasets.load_dataset('squad_v2', split='train', keep_in_memory=True)\r\n\r\nstart_len = len(dataset)\r\ndataset = dataset.filter(check_context_answer_alignment,\r\n num_proc=1,\r\n keep_in_memory=True)\r\nend_len = len(dataset)\r\nprint('{} instances contain mis-alignment between the answer text and answer index.'.format(start_len - end_len))\r\n```\r\n\r\n## Expected results\r\nThis code should result in 0 rows being filtered out from the dataset.\r\n\r\n## Actual results\r\nThis filter command results in 258 rows being flagged as containing a discrepancy between the text contained within answers['text'] and the text in example['context'] at the answers['answer_start'] location.\r\n\r\nThis code will reproduce the problem and produce the following count:\r\n\"258 instances contain mis-alignment between the answer text and answer index.\"\r\n\r\n## Environment info\r\nSteps to rebuilt the Conda environment:\r\n```\r\n# create a virtual environment to stuff all these packages into\r\nconda create -n round8 python=3.8 -y\r\n\r\n# activate the virtual environment\r\nconda activate round8\r\n\r\n# install pytorch (best done through conda to handle cuda dependencies)\r\nconda install pytorch torchvision torchtext cudatoolkit=11.1 -c pytorch-lts -c nvidia\r\n\r\npip install jsonpickle transformers datasets matplotlib\r\n```\r\n\r\nOS: Ubuntu 20.04\r\nPython 3.8\r\n\r\nResult of `conda env export`:\r\n```\r\nname: round8\r\nchannels:\r\n - pytorch-lts\r\n - nvidia\r\n - defaults\r\ndependencies:\r\n - _libgcc_mutex=0.1=main\r\n - _openmp_mutex=4.5=1_gnu\r\n - blas=1.0=mkl\r\n - brotlipy=0.7.0=py38h27cfd23_1003\r\n - bzip2=1.0.8=h7b6447c_0\r\n - ca-certificates=2021.5.25=h06a4308_1\r\n - certifi=2021.5.30=py38h06a4308_0\r\n - cffi=1.14.5=py38h261ae71_0\r\n - chardet=4.0.0=py38h06a4308_1003\r\n - cryptography=3.4.7=py38hd23ed53_0\r\n - cudatoolkit=11.1.74=h6bb024c_0\r\n - ffmpeg=4.2.2=h20bf706_0\r\n - freetype=2.10.4=h5ab3b9f_0\r\n - gmp=6.2.1=h2531618_2\r\n - gnutls=3.6.15=he1e5248_0\r\n - idna=2.10=pyhd3eb1b0_0\r\n - intel-openmp=2021.2.0=h06a4308_610\r\n - jpeg=9b=h024ee3a_2\r\n - lame=3.100=h7b6447c_0\r\n - lcms2=2.12=h3be6417_0\r\n - ld_impl_linux-64=2.35.1=h7274673_9\r\n - libffi=3.3=he6710b0_2\r\n - libgcc-ng=9.3.0=h5101ec6_17\r\n - libgomp=9.3.0=h5101ec6_17\r\n - libidn2=2.3.1=h27cfd23_0\r\n - libopus=1.3.1=h7b6447c_0\r\n - libpng=1.6.37=hbc83047_0\r\n - libstdcxx-ng=9.3.0=hd4cf53a_17\r\n - libtasn1=4.16.0=h27cfd23_0\r\n - libtiff=4.2.0=h85742a9_0\r\n - libunistring=0.9.10=h27cfd23_0\r\n - libuv=1.40.0=h7b6447c_0\r\n - libvpx=1.7.0=h439df22_0\r\n - libwebp-base=1.2.0=h27cfd23_0\r\n - lz4-c=1.9.3=h2531618_0\r\n - mkl=2021.2.0=h06a4308_296\r\n - mkl-service=2.3.0=py38h27cfd23_1\r\n - mkl_fft=1.3.0=py38h42c9631_2\r\n - mkl_random=1.2.1=py38ha9443f7_2\r\n - ncurses=6.2=he6710b0_1\r\n - nettle=3.7.3=hbbd107a_1\r\n - ninja=1.10.2=hff7bd54_1\r\n - numpy=1.20.2=py38h2d18471_0\r\n - numpy-base=1.20.2=py38hfae3a4d_0\r\n - olefile=0.46=py_0\r\n - openh264=2.1.0=hd408876_0\r\n - openssl=1.1.1k=h27cfd23_0\r\n - pillow=8.2.0=py38he98fc37_0\r\n - pip=21.1.2=py38h06a4308_0\r\n - pycparser=2.20=py_2\r\n - pyopenssl=20.0.1=pyhd3eb1b0_1\r\n - pysocks=1.7.1=py38h06a4308_0\r\n - python=3.8.10=h12debd9_8\r\n - pytorch=1.8.1=py3.8_cuda11.1_cudnn8.0.5_0\r\n - readline=8.1=h27cfd23_0\r\n - requests=2.25.1=pyhd3eb1b0_0\r\n - setuptools=52.0.0=py38h06a4308_0\r\n - six=1.16.0=pyhd3eb1b0_0\r\n - sqlite=3.35.4=hdfb4753_0\r\n - tk=8.6.10=hbc83047_0\r\n - torchtext=0.9.1=py38\r\n - torchvision=0.9.1=py38_cu111\r\n - typing_extensions=3.7.4.3=pyha847dfd_0\r\n - urllib3=1.26.4=pyhd3eb1b0_0\r\n - wheel=0.36.2=pyhd3eb1b0_0\r\n - x264=1!157.20191217=h7b6447c_0\r\n - xz=5.2.5=h7b6447c_0\r\n - zlib=1.2.11=h7b6447c_3\r\n - zstd=1.4.9=haebb681_0\r\n - pip:\r\n - click==8.0.1\r\n - cycler==0.10.0\r\n - datasets==1.8.0\r\n - dill==0.3.4\r\n - filelock==3.0.12\r\n - fsspec==2021.6.0\r\n - huggingface-hub==0.0.8\r\n - joblib==1.0.1\r\n - jsonpickle==2.0.0\r\n - kiwisolver==1.3.1\r\n - matplotlib==3.4.2\r\n - multiprocess==0.70.12.2\r\n - packaging==20.9\r\n - pandas==1.2.4\r\n - pyarrow==3.0.0\r\n - pyparsing==2.4.7\r\n - python-dateutil==2.8.1\r\n - pytz==2021.1\r\n - regex==2021.4.4\r\n - sacremoses==0.0.45\r\n - tokenizers==0.10.3\r\n - tqdm==4.49.0\r\n - transformers==4.6.1\r\n - xxhash==2.0.2\r\nprefix: \/home\/mmajurski\/anaconda3\/envs\/round8\r\n```\r\n","comment_length":71,"text":"sqaud_v2 dataset contains misalignment between the answer text and the context value at the answer index\n## Describe the bug\r\nThe built in huggingface squad_v2 dataset that you can access via datasets.load_dataset contains mis-alignment between the answers['text'] and the characters in the context at the location specified by answers['answer_start'].\r\n\r\nFor example:\r\nid = '56d1f453e7d4791d009025bd'\r\nanswers = {'text': ['Pure Land'], 'answer_start': [146]}\r\nHowever the actual text in context at location 146 is 'ure Land,'\r\nWhich is an off-by-one error from the correct answer.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\ndef check_context_answer_alignment(example):\r\n for a_idx in range(len(example['answers']['text'])):\r\n # check raw dataset for answer consistency between context and answer\r\n answer_text = example['answers']['text'][a_idx]\r\n a_st_idx = example['answers']['answer_start'][a_idx]\r\n a_end_idx = a_st_idx + len(example['answers']['text'][a_idx])\r\n answer_text_from_context = example['context'][a_st_idx:a_end_idx]\r\n if answer_text != answer_text_from_context:\r\n #print(example['id'])\r\n return False\r\n return True\r\n\r\ndataset = datasets.load_dataset('squad_v2', split='train', keep_in_memory=True)\r\n\r\nstart_len = len(dataset)\r\ndataset = dataset.filter(check_context_answer_alignment,\r\n num_proc=1,\r\n keep_in_memory=True)\r\nend_len = len(dataset)\r\nprint('{} instances contain mis-alignment between the answer text and answer index.'.format(start_len - end_len))\r\n```\r\n\r\n## Expected results\r\nThis code should result in 0 rows being filtered out from the dataset.\r\n\r\n## Actual results\r\nThis filter command results in 258 rows being flagged as containing a discrepancy between the text contained within answers['text'] and the text in example['context'] at the answers['answer_start'] location.\r\n\r\nThis code will reproduce the problem and produce the following count:\r\n\"258 instances contain mis-alignment between the answer text and answer index.\"\r\n\r\n## Environment info\r\nSteps to rebuilt the Conda environment:\r\n```\r\n# create a virtual environment to stuff all these packages into\r\nconda create -n round8 python=3.8 -y\r\n\r\n# activate the virtual environment\r\nconda activate round8\r\n\r\n# install pytorch (best done through conda to handle cuda dependencies)\r\nconda install pytorch torchvision torchtext cudatoolkit=11.1 -c pytorch-lts -c nvidia\r\n\r\npip install jsonpickle transformers datasets matplotlib\r\n```\r\n\r\nOS: Ubuntu 20.04\r\nPython 3.8\r\n\r\nResult of `conda env export`:\r\n```\r\nname: round8\r\nchannels:\r\n - pytorch-lts\r\n - nvidia\r\n - defaults\r\ndependencies:\r\n - _libgcc_mutex=0.1=main\r\n - _openmp_mutex=4.5=1_gnu\r\n - blas=1.0=mkl\r\n - brotlipy=0.7.0=py38h27cfd23_1003\r\n - bzip2=1.0.8=h7b6447c_0\r\n - ca-certificates=2021.5.25=h06a4308_1\r\n - certifi=2021.5.30=py38h06a4308_0\r\n - cffi=1.14.5=py38h261ae71_0\r\n - chardet=4.0.0=py38h06a4308_1003\r\n - cryptography=3.4.7=py38hd23ed53_0\r\n - cudatoolkit=11.1.74=h6bb024c_0\r\n - ffmpeg=4.2.2=h20bf706_0\r\n - freetype=2.10.4=h5ab3b9f_0\r\n - gmp=6.2.1=h2531618_2\r\n - gnutls=3.6.15=he1e5248_0\r\n - idna=2.10=pyhd3eb1b0_0\r\n - intel-openmp=2021.2.0=h06a4308_610\r\n - jpeg=9b=h024ee3a_2\r\n - lame=3.100=h7b6447c_0\r\n - lcms2=2.12=h3be6417_0\r\n - ld_impl_linux-64=2.35.1=h7274673_9\r\n - libffi=3.3=he6710b0_2\r\n - libgcc-ng=9.3.0=h5101ec6_17\r\n - libgomp=9.3.0=h5101ec6_17\r\n - libidn2=2.3.1=h27cfd23_0\r\n - libopus=1.3.1=h7b6447c_0\r\n - libpng=1.6.37=hbc83047_0\r\n - libstdcxx-ng=9.3.0=hd4cf53a_17\r\n - libtasn1=4.16.0=h27cfd23_0\r\n - libtiff=4.2.0=h85742a9_0\r\n - libunistring=0.9.10=h27cfd23_0\r\n - libuv=1.40.0=h7b6447c_0\r\n - libvpx=1.7.0=h439df22_0\r\n - libwebp-base=1.2.0=h27cfd23_0\r\n - lz4-c=1.9.3=h2531618_0\r\n - mkl=2021.2.0=h06a4308_296\r\n - mkl-service=2.3.0=py38h27cfd23_1\r\n - mkl_fft=1.3.0=py38h42c9631_2\r\n - mkl_random=1.2.1=py38ha9443f7_2\r\n - ncurses=6.2=he6710b0_1\r\n - nettle=3.7.3=hbbd107a_1\r\n - ninja=1.10.2=hff7bd54_1\r\n - numpy=1.20.2=py38h2d18471_0\r\n - numpy-base=1.20.2=py38hfae3a4d_0\r\n - olefile=0.46=py_0\r\n - openh264=2.1.0=hd408876_0\r\n - openssl=1.1.1k=h27cfd23_0\r\n - pillow=8.2.0=py38he98fc37_0\r\n - pip=21.1.2=py38h06a4308_0\r\n - pycparser=2.20=py_2\r\n - pyopenssl=20.0.1=pyhd3eb1b0_1\r\n - pysocks=1.7.1=py38h06a4308_0\r\n - python=3.8.10=h12debd9_8\r\n - pytorch=1.8.1=py3.8_cuda11.1_cudnn8.0.5_0\r\n - readline=8.1=h27cfd23_0\r\n - requests=2.25.1=pyhd3eb1b0_0\r\n - setuptools=52.0.0=py38h06a4308_0\r\n - six=1.16.0=pyhd3eb1b0_0\r\n - sqlite=3.35.4=hdfb4753_0\r\n - tk=8.6.10=hbc83047_0\r\n - torchtext=0.9.1=py38\r\n - torchvision=0.9.1=py38_cu111\r\n - typing_extensions=3.7.4.3=pyha847dfd_0\r\n - urllib3=1.26.4=pyhd3eb1b0_0\r\n - wheel=0.36.2=pyhd3eb1b0_0\r\n - x264=1!157.20191217=h7b6447c_0\r\n - xz=5.2.5=h7b6447c_0\r\n - zlib=1.2.11=h7b6447c_3\r\n - zstd=1.4.9=haebb681_0\r\n - pip:\r\n - click==8.0.1\r\n - cycler==0.10.0\r\n - datasets==1.8.0\r\n - dill==0.3.4\r\n - filelock==3.0.12\r\n - fsspec==2021.6.0\r\n - huggingface-hub==0.0.8\r\n - joblib==1.0.1\r\n - jsonpickle==2.0.0\r\n - kiwisolver==1.3.1\r\n - matplotlib==3.4.2\r\n - multiprocess==0.70.12.2\r\n - packaging==20.9\r\n - pandas==1.2.4\r\n - pyarrow==3.0.0\r\n - pyparsing==2.4.7\r\n - python-dateutil==2.8.1\r\n - pytz==2021.1\r\n - regex==2021.4.4\r\n - sacremoses==0.0.45\r\n - tokenizers==0.10.3\r\n - tqdm==4.49.0\r\n - transformers==4.6.1\r\n - xxhash==2.0.2\r\nprefix: \/home\/mmajurski\/anaconda3\/envs\/round8\r\n```\r\n\nHi @mmajurski, thanks for reporting this issue.\r\n\r\nIndeed this misalignment arises because the source dataset context field contains leading blank spaces (and these are counted within the answer_start), while our datasets loading script removes these leading blank spaces.\r\n\r\nI'm going to fix our script so that all leading blank spaces in the source dataset are kept, and there is no misalignment between the answer text and the answer_start within the context.","embeddings":[-0.2330299318,-0.3295379579,-0.0441096425,0.3749362826,0.0878231078,-0.0763419643,0.1057570204,0.2279032022,-0.2927472591,0.1577830762,-0.0841630697,0.0450688601,0.3250361681,0.0685452819,-0.0926083177,0.1523635834,0.1756330729,0.0104246661,0.009945604,-0.1725040823,-0.2248985022,0.2991383374,-0.2980325818,0.03003061,-0.342564702,-0.0102489283,0.0364246443,0.1131834313,-0.1175084189,-0.4159388542,0.2125024199,-0.0936790556,0.0077113751,0.2638638318,-0.000116531,-0.1706253737,0.0728382915,-0.0973288417,-0.2682839036,0.0041923714,-0.0435589068,-0.1216914803,-0.1298896968,-0.1058420315,-0.2390315384,0.2717562318,-0.0673231781,-0.2866247594,0.6096699834,0.2443222255,0.1451372653,0.2883639336,0.0109265558,-0.064738296,0.1786869168,-0.1235733703,0.0756943524,0.119126901,0.1456331909,0.1079365462,-0.0346998237,0.4713858068,-0.1343523413,-0.0318910703,-0.0061442321,0.2904502153,0.206617564,-0.1278317124,0.0998568758,0.1311751604,0.2598810494,-0.2484471202,-0.424808681,-0.4569473565,-0.0358279571,-0.1224056184,0.3317508101,0.039453499,-0.0311524291,0.3647889495,-0.1365324855,0.0710276216,-0.0857497603,0.127139762,-0.0876931027,0.2383429706,-0.1219492704,0.1492218971,-0.2677045166,-0.1386347562,-0.0631168187,-0.1375988126,-0.105233863,0.4171854258,-0.5139366984,-0.1013480797,-0.1679380983,0.0907929763,0.2361780256,0.0694944859,-0.0699167773,0.0015248328,0.2659206092,0.0613521114,0.3417307138,0.2773298621,-0.1180535555,0.255628705,0.1020577699,0.1474464983,0.0414646193,0.1034675837,0.1996169537,-0.1841570288,-0.1082466468,-0.1547538042,0.1684667319,-0.1887482852,-0.5222857594,0.4338434935,-0.6319300532,0.2756719887,-0.0214355942,0.3482971787,-0.2216847241,0.2669170201,0.3325076103,0.244284451,-0.0650029629,-0.1851683259,-0.133084923,-0.0969801545,-0.0144069865,-0.0344590582,0.0072183558,-0.1293735057,0.2271303385,0.2961053848,0.2128195316,-0.2205923349,-0.3132048249,-0.1648734659,0.1410585493,0.0306854323,0.0271047894,0.1876549721,0.1873004138,-0.2884297669,0.073195301,0.0930062234,-0.2383058369,0.0768734515,-0.0992043242,0.2245198935,-0.1660035253,-0.056426052,0.0465995781,0.2456641197,0.2966586649,-0.2863129377,0.2683805823,-0.203826189,-0.3300075233,-0.0881249383,0.3650223017,0.2920378745,-0.1985461414,-0.0659254789,-0.0561406538,0.1563176215,0.0265871864,0.237526387,0.2730125487,0.2942620516,-0.4212450981,0.2674057782,0.1037338749,-0.8540112376,-0.5590248704,-0.1742199212,-0.1163506806,0.225110665,-0.1036156043,0.0728597566,0.4342785478,0.1543433517,0.2735286653,0.5391799212,0.0679247305,0.1189373881,-0.3740178347,0.0432915017,-0.0926679522,0.1235257834,-0.2477580458,-0.1743770838,0.0530980192,-0.1834742129,0.1845173985,0.1049054489,0.0700077787,0.242408067,0.3292447031,0.1207566783,0.1795453727,-0.5024951696,-0.4840292037,0.0933753848,-0.0992036015,0.0597215667,-0.0624961443,-0.1360670328,-0.3975332677,-0.0361760221,-0.3520284891,-0.3007717729,0.0909299403,0.0083287833,0.1742563844,0.1125237122,-0.1105441302,0.6274502277,-0.2027090788,0.2574426234,-0.0648542717,0.1838029921,-0.0369916596,0.0197878219,0.0257437229,0.4265458286,-0.0366344415,0.0865879729,-0.2453000844,0.5439850688,0.2354308665,0.1975178421,-0.2265350223,-0.0887751728,0.2265448421,-0.4291588366,-0.2648878694,0.2957112789,0.1549057662,0.1138918698,-0.3117320538,0.59599787,0.0686130524,0.0849598721,-0.2245444506,-0.0018041183,0.2800176144,-0.1242726669,-0.1135764197,-0.2309002727,0.2905323505,0.1432120651,0.5214198828,0.1039449498,-0.3891157508,0.1001307517,0.267216295,-0.1695472002,0.0794455707,0.1168767288,0.0412754938,-0.0822434872,0.024764752,0.2209540159,0.2899647057,0.2044537365,-0.3409023285,0.3199863434,-0.036653012,0.1461652517,0.2144389898,-0.1729850173,-0.0400366671,0.4994864166,0.3725070655,0.2427582741,-0.1843069196,0.1724185944,-0.3970791101,0.152523905,-0.5550450087,-0.021352414,-0.1496844441,-0.0735898316,-0.1993011385,-0.0192141831,-0.2774994969,-0.1124377027,0.248433575,0.1559061408,-0.0537631586,0.2769265175,0.0002657939,0.4594179392,0.0978769809,-0.1988540739,-0.0586240068,-0.0650393665,-0.1624031663,-0.0090979105,0.0156265758,0.1998775154,0.1458058804,-0.4826640189,-0.3008002043,-0.2854555845,-0.3453061879,0.2101686448,-0.0686363429,0.3613208234,0.1664168537,-0.0186989903,-0.3620541394,-0.2030521333,0.164502278,-0.0291347131,-0.4773875177,-0.0524751432,0.0556921959,-0.2053315192,-0.0412881449,-0.4563171864,-0.0989391729,-0.2709364295,0.0162285641,-0.1817271709,0.0823787078,0.2436575741,-0.1389108002,0.0348106995,-0.2918559611,0.1260592341,-0.5039397478,-0.0472881943,0.0939144045,0.0430192649,-0.3436332941,-0.2887289226,-0.232024312,-0.1749058217,-0.0350877568,-0.1375848353,-0.2258420736,-0.1793827713,0.2166627795,-0.2268270254,0.2435190976,0.2265461981,-0.1808743179,0.0880100653,-0.2281240225,-0.4454982877,0.1523829252,0.3971565962,0.2353447974,-0.1117967144,0.3300126195,-0.0246539097,0.6546520591,0.4053243101,0.226618126,0.1022702307,-0.2168853581,0.3332633674,-0.2054528892,-0.2187942415,0.312690258,0.1287728995,-0.3271473944,0.3284796774,0.039798025,0.1749130636,0.015510424,0.1942849457,-0.5804231763,-0.1875118911,-0.2166323811,0.0498929955,0.1647057831,0.0902877003,-0.0723394454,0.1494957656,-0.1061443985,-0.0524449386,0.1069306061,-0.0442453772,0.236469686,-0.8369106054,-0.2997891307,0.0433075465,0.2258636355,0.3995067477,0.381581068,-0.0134445364,-0.093993254,0.1968369186,0.0923821777,0.7246232629,0.0159390848,0.0829525888,-0.2899780869,0.1853271425,-0.4381762743,-0.2349051684,-0.1948949248,0.1377697438,0.2811782062,0.1698484123,-0.2208506167,-0.0408705473,0.1911017746,0.0646320134,-0.0984831601,-0.0093899928,-0.2653487027,-0.1490731835,-0.3438111842,0.2851322889,0.4409757853,0.2931801081,0.478546828,-0.0608856529,0.1911562532,0.1568410844,-0.1183883771,0.2444164157,0.5145237446,-0.1394737959,0.020231897,0.2098539174,-0.0472850315,0.0350847431,0.531881988,-0.0985551476,-0.3209655881,-0.0355942957,-0.0423497148,0.3858833313,0.3409575224,0.0271561947,0.007218238,-0.0423274375,0.1705225259,-0.239471674,-0.1145948768,0.3838074207,0.5305989385,-0.0723313093,-0.2432465702,0.226994887,-0.1459895521,0.0304384511,0.0929574072,0.5290966034,-0.2636906207,0.4520533383,-0.1095875725,1.0020316839,0.328188926,0.1134260595,0.497885257,-0.0472013615,0.3578611016,-0.0754101947,0.2748137712,-0.4139471352,0.3110217452,-0.0377646796,-0.1198738515,-0.1178149134,-0.1698682457,-0.2265703827,0.2860626876,-0.1526471376,0.2311641425,-0.0626616031,-0.0085225478,-0.2034142613,-0.033530008,-0.4369697273,0.0932308212,0.014678021,0.2703944445,-0.148771584,0.1750009209,-0.3390996456,-0.3742335439,-0.6190102696,-0.0299747158,-0.4763568342,0.0047969306,0.4499208033,-0.0897761583,-0.0364160649,0.4331648052,0.3017061949,0.0679342151,-0.0369109735,0.1858628243,-0.0765430257,0.3762842119,0.4295200109,-0.0519072898,0.1574102938,-0.0374897495,-0.3992099166,-0.1864605993,0.0142571572,-0.2147545069,0.0705938786,0.1132239476,0.0116750086,-0.180656597,-0.3797228932,-0.1218563616,0.288859576,-0.2152142823,0.1218225509,-0.1044649407,-0.1833918095,0.1680883616,0.2302089781,-0.3828324676,-0.029355986,0.4591567516,-0.2447421104,0.0446338654,0.4361592531,0.0126489829,-0.1020924821,-0.1210553348,0.1857443005,-0.0898654163,-0.1195793897,0.020980645,-0.1442859769,-0.3595216572,-0.2137382925,0.3384000063,0.38391608,-0.1161921322,-0.0502924994,-0.6957364082,-0.0235519614,0.2530730963,0.0438782871,-0.0451187789,0.362829417,0.0801646635,-0.1754003018,-0.0703575015,-0.3165868223,-0.0696842819,-0.2544342875,0.1378873289,-0.0004854769,-0.054904744,0.2277368754,-0.2508426905,0.0973833054,-0.0025261098,-0.1287060976,-0.2157400548,-0.1059362367,0.151648134,0.0958362594,-0.2368230373,-0.1105897501,-0.0767657012,0.1523080468,-0.190477252,0.128565371,0.4913201928,0.151816532,0.1998494267,-0.518520236,-0.1265322566,0.0309705436,0.0234835055,-0.0831899941,-0.3460215032,0.1145583838,0.2857742608,-0.2541635334,-0.0568224117,0.250439167,0.1971273869,0.156353116,0.1079917178,0.1915987283,-0.0982195735,-0.0190188605,0.183392778,0.2271950245,0.3787548244,-0.4000142813,-0.0452516899,-0.0354742371,0.1983865798,-0.3783816695,-0.0351037718,0.2172053754,-0.188268885,-0.0205085445,0.2156239897,0.1364915371,-0.2648781836,-0.0650718063,0.2325160801,0.580716908,-0.0888280421,0.0113576148,0.1476523876,0.0274071768,0.2149521708,0.3176786602,-0.1152953655,0.2021837384,0.1865382344,-0.1520356238,0.0439828038,-0.2052265704,0.1520625502,0.1444454789,0.0466857925,0.3828479052,0.3497182131,-0.4396210313,0.011120189,0.1317227036,0.4576255679,-0.1160483956,-0.2384144217,-0.2936193049,-0.0883242488,-0.2381866127,-0.0578628816,-0.0807035342,-0.1052349582,-0.1390684694,-0.0989655182,-0.0628812537,-0.2211410105,0.1579808146,0.1681921482,-0.2182729393,-0.3492713273,-0.2746627033,0.2447257638,0.1839176714,-0.1713453531,0.341062516,0.0664092451,0.0094089806,0.3050126731,0.5498544574,0.5111480951,0.2144199312,-0.1568036973,0.1962303221,-0.1739758551,-0.0277864691,0.011248799,0.110214144,-0.2986117303,-0.4947999716,0.0919381231,0.2147872746,-0.1876727939,0.01391479,0.0613702126,0.1129160225,-0.3813641965,-0.2247827202,-0.3311124146,0.3151141405,-0.1817768663,0.1097344384,-0.1495829225,-0.0258891806,0.1415275484,-0.1616253704,0.2025576085,0.0786848292,0.0715653673,0.1909437478,0.04452819,0.6243495941,-0.2013409585,-0.0681901425,-0.0578789152,-0.5805677772,0.1312075257,0.0310452115,-0.1154971719,0.2015237659,0.2146641016,-0.1598927528,0.2209586054,0.4334922135,0.0617087074,-0.1903527081,0.0112976078,-0.2144546956,0.0767382309,-0.2031755149,0.1104440317,-0.0068181152,-0.2531497777,0.2053184509,0.1481136978,-0.0577719957,0.1885320842,0.1527107358,-0.0800265297,-0.4327370226,-0.1294350922,0.3209770322,0.2201793194,-0.2354950458,0.1969352812,-0.0664646477,-0.1612056196,-0.5237137675,0.1405882686,-0.3926342726,0.293876946,-0.0463227294,-0.3050654531,0.0475321859,0.2376043051,0.3541375399,-0.2929815948,-0.4807341993,0.1937257349,-0.1500291824,0.2315478325,-0.0873755366,0.1933842897,-0.080139026,0.457072854,-0.3387340903,-0.3580105603,0.3238231838,-0.3490354419,-0.1189630553,0.0141020631,0.3399590254,0.2818251848,-0.1485277414,-0.505400002,-0.1532177329,0.2660462856,-0.0004354485,-0.2164424956,0.1116041318,-0.0506298356,-0.0587325767,0.0299290884,0.0477241166,0.3681152165,-0.1083867252,0.2889886498,-0.3507986069]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2585","title":"sqaud_v2 dataset contains misalignment between the answer text and the context value at the answer index","comments":"If you are going to be altering the data cleaning from the source Squad dataset, here is one thing to consider.\r\nThere are occasional double spaces separating words which it might be nice to get rid of. \r\n\r\nEither way, thank you.","body":"## Describe the bug\r\nThe built in huggingface squad_v2 dataset that you can access via datasets.load_dataset contains mis-alignment between the answers['text'] and the characters in the context at the location specified by answers['answer_start'].\r\n\r\nFor example:\r\nid = '56d1f453e7d4791d009025bd'\r\nanswers = {'text': ['Pure Land'], 'answer_start': [146]}\r\nHowever the actual text in context at location 146 is 'ure Land,'\r\nWhich is an off-by-one error from the correct answer.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\ndef check_context_answer_alignment(example):\r\n for a_idx in range(len(example['answers']['text'])):\r\n # check raw dataset for answer consistency between context and answer\r\n answer_text = example['answers']['text'][a_idx]\r\n a_st_idx = example['answers']['answer_start'][a_idx]\r\n a_end_idx = a_st_idx + len(example['answers']['text'][a_idx])\r\n answer_text_from_context = example['context'][a_st_idx:a_end_idx]\r\n if answer_text != answer_text_from_context:\r\n #print(example['id'])\r\n return False\r\n return True\r\n\r\ndataset = datasets.load_dataset('squad_v2', split='train', keep_in_memory=True)\r\n\r\nstart_len = len(dataset)\r\ndataset = dataset.filter(check_context_answer_alignment,\r\n num_proc=1,\r\n keep_in_memory=True)\r\nend_len = len(dataset)\r\nprint('{} instances contain mis-alignment between the answer text and answer index.'.format(start_len - end_len))\r\n```\r\n\r\n## Expected results\r\nThis code should result in 0 rows being filtered out from the dataset.\r\n\r\n## Actual results\r\nThis filter command results in 258 rows being flagged as containing a discrepancy between the text contained within answers['text'] and the text in example['context'] at the answers['answer_start'] location.\r\n\r\nThis code will reproduce the problem and produce the following count:\r\n\"258 instances contain mis-alignment between the answer text and answer index.\"\r\n\r\n## Environment info\r\nSteps to rebuilt the Conda environment:\r\n```\r\n# create a virtual environment to stuff all these packages into\r\nconda create -n round8 python=3.8 -y\r\n\r\n# activate the virtual environment\r\nconda activate round8\r\n\r\n# install pytorch (best done through conda to handle cuda dependencies)\r\nconda install pytorch torchvision torchtext cudatoolkit=11.1 -c pytorch-lts -c nvidia\r\n\r\npip install jsonpickle transformers datasets matplotlib\r\n```\r\n\r\nOS: Ubuntu 20.04\r\nPython 3.8\r\n\r\nResult of `conda env export`:\r\n```\r\nname: round8\r\nchannels:\r\n - pytorch-lts\r\n - nvidia\r\n - defaults\r\ndependencies:\r\n - _libgcc_mutex=0.1=main\r\n - _openmp_mutex=4.5=1_gnu\r\n - blas=1.0=mkl\r\n - brotlipy=0.7.0=py38h27cfd23_1003\r\n - bzip2=1.0.8=h7b6447c_0\r\n - ca-certificates=2021.5.25=h06a4308_1\r\n - certifi=2021.5.30=py38h06a4308_0\r\n - cffi=1.14.5=py38h261ae71_0\r\n - chardet=4.0.0=py38h06a4308_1003\r\n - cryptography=3.4.7=py38hd23ed53_0\r\n - cudatoolkit=11.1.74=h6bb024c_0\r\n - ffmpeg=4.2.2=h20bf706_0\r\n - freetype=2.10.4=h5ab3b9f_0\r\n - gmp=6.2.1=h2531618_2\r\n - gnutls=3.6.15=he1e5248_0\r\n - idna=2.10=pyhd3eb1b0_0\r\n - intel-openmp=2021.2.0=h06a4308_610\r\n - jpeg=9b=h024ee3a_2\r\n - lame=3.100=h7b6447c_0\r\n - lcms2=2.12=h3be6417_0\r\n - ld_impl_linux-64=2.35.1=h7274673_9\r\n - libffi=3.3=he6710b0_2\r\n - libgcc-ng=9.3.0=h5101ec6_17\r\n - libgomp=9.3.0=h5101ec6_17\r\n - libidn2=2.3.1=h27cfd23_0\r\n - libopus=1.3.1=h7b6447c_0\r\n - libpng=1.6.37=hbc83047_0\r\n - libstdcxx-ng=9.3.0=hd4cf53a_17\r\n - libtasn1=4.16.0=h27cfd23_0\r\n - libtiff=4.2.0=h85742a9_0\r\n - libunistring=0.9.10=h27cfd23_0\r\n - libuv=1.40.0=h7b6447c_0\r\n - libvpx=1.7.0=h439df22_0\r\n - libwebp-base=1.2.0=h27cfd23_0\r\n - lz4-c=1.9.3=h2531618_0\r\n - mkl=2021.2.0=h06a4308_296\r\n - mkl-service=2.3.0=py38h27cfd23_1\r\n - mkl_fft=1.3.0=py38h42c9631_2\r\n - mkl_random=1.2.1=py38ha9443f7_2\r\n - ncurses=6.2=he6710b0_1\r\n - nettle=3.7.3=hbbd107a_1\r\n - ninja=1.10.2=hff7bd54_1\r\n - numpy=1.20.2=py38h2d18471_0\r\n - numpy-base=1.20.2=py38hfae3a4d_0\r\n - olefile=0.46=py_0\r\n - openh264=2.1.0=hd408876_0\r\n - openssl=1.1.1k=h27cfd23_0\r\n - pillow=8.2.0=py38he98fc37_0\r\n - pip=21.1.2=py38h06a4308_0\r\n - pycparser=2.20=py_2\r\n - pyopenssl=20.0.1=pyhd3eb1b0_1\r\n - pysocks=1.7.1=py38h06a4308_0\r\n - python=3.8.10=h12debd9_8\r\n - pytorch=1.8.1=py3.8_cuda11.1_cudnn8.0.5_0\r\n - readline=8.1=h27cfd23_0\r\n - requests=2.25.1=pyhd3eb1b0_0\r\n - setuptools=52.0.0=py38h06a4308_0\r\n - six=1.16.0=pyhd3eb1b0_0\r\n - sqlite=3.35.4=hdfb4753_0\r\n - tk=8.6.10=hbc83047_0\r\n - torchtext=0.9.1=py38\r\n - torchvision=0.9.1=py38_cu111\r\n - typing_extensions=3.7.4.3=pyha847dfd_0\r\n - urllib3=1.26.4=pyhd3eb1b0_0\r\n - wheel=0.36.2=pyhd3eb1b0_0\r\n - x264=1!157.20191217=h7b6447c_0\r\n - xz=5.2.5=h7b6447c_0\r\n - zlib=1.2.11=h7b6447c_3\r\n - zstd=1.4.9=haebb681_0\r\n - pip:\r\n - click==8.0.1\r\n - cycler==0.10.0\r\n - datasets==1.8.0\r\n - dill==0.3.4\r\n - filelock==3.0.12\r\n - fsspec==2021.6.0\r\n - huggingface-hub==0.0.8\r\n - joblib==1.0.1\r\n - jsonpickle==2.0.0\r\n - kiwisolver==1.3.1\r\n - matplotlib==3.4.2\r\n - multiprocess==0.70.12.2\r\n - packaging==20.9\r\n - pandas==1.2.4\r\n - pyarrow==3.0.0\r\n - pyparsing==2.4.7\r\n - python-dateutil==2.8.1\r\n - pytz==2021.1\r\n - regex==2021.4.4\r\n - sacremoses==0.0.45\r\n - tokenizers==0.10.3\r\n - tqdm==4.49.0\r\n - transformers==4.6.1\r\n - xxhash==2.0.2\r\nprefix: \/home\/mmajurski\/anaconda3\/envs\/round8\r\n```\r\n","comment_length":41,"text":"sqaud_v2 dataset contains misalignment between the answer text and the context value at the answer index\n## Describe the bug\r\nThe built in huggingface squad_v2 dataset that you can access via datasets.load_dataset contains mis-alignment between the answers['text'] and the characters in the context at the location specified by answers['answer_start'].\r\n\r\nFor example:\r\nid = '56d1f453e7d4791d009025bd'\r\nanswers = {'text': ['Pure Land'], 'answer_start': [146]}\r\nHowever the actual text in context at location 146 is 'ure Land,'\r\nWhich is an off-by-one error from the correct answer.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\n\r\ndef check_context_answer_alignment(example):\r\n for a_idx in range(len(example['answers']['text'])):\r\n # check raw dataset for answer consistency between context and answer\r\n answer_text = example['answers']['text'][a_idx]\r\n a_st_idx = example['answers']['answer_start'][a_idx]\r\n a_end_idx = a_st_idx + len(example['answers']['text'][a_idx])\r\n answer_text_from_context = example['context'][a_st_idx:a_end_idx]\r\n if answer_text != answer_text_from_context:\r\n #print(example['id'])\r\n return False\r\n return True\r\n\r\ndataset = datasets.load_dataset('squad_v2', split='train', keep_in_memory=True)\r\n\r\nstart_len = len(dataset)\r\ndataset = dataset.filter(check_context_answer_alignment,\r\n num_proc=1,\r\n keep_in_memory=True)\r\nend_len = len(dataset)\r\nprint('{} instances contain mis-alignment between the answer text and answer index.'.format(start_len - end_len))\r\n```\r\n\r\n## Expected results\r\nThis code should result in 0 rows being filtered out from the dataset.\r\n\r\n## Actual results\r\nThis filter command results in 258 rows being flagged as containing a discrepancy between the text contained within answers['text'] and the text in example['context'] at the answers['answer_start'] location.\r\n\r\nThis code will reproduce the problem and produce the following count:\r\n\"258 instances contain mis-alignment between the answer text and answer index.\"\r\n\r\n## Environment info\r\nSteps to rebuilt the Conda environment:\r\n```\r\n# create a virtual environment to stuff all these packages into\r\nconda create -n round8 python=3.8 -y\r\n\r\n# activate the virtual environment\r\nconda activate round8\r\n\r\n# install pytorch (best done through conda to handle cuda dependencies)\r\nconda install pytorch torchvision torchtext cudatoolkit=11.1 -c pytorch-lts -c nvidia\r\n\r\npip install jsonpickle transformers datasets matplotlib\r\n```\r\n\r\nOS: Ubuntu 20.04\r\nPython 3.8\r\n\r\nResult of `conda env export`:\r\n```\r\nname: round8\r\nchannels:\r\n - pytorch-lts\r\n - nvidia\r\n - defaults\r\ndependencies:\r\n - _libgcc_mutex=0.1=main\r\n - _openmp_mutex=4.5=1_gnu\r\n - blas=1.0=mkl\r\n - brotlipy=0.7.0=py38h27cfd23_1003\r\n - bzip2=1.0.8=h7b6447c_0\r\n - ca-certificates=2021.5.25=h06a4308_1\r\n - certifi=2021.5.30=py38h06a4308_0\r\n - cffi=1.14.5=py38h261ae71_0\r\n - chardet=4.0.0=py38h06a4308_1003\r\n - cryptography=3.4.7=py38hd23ed53_0\r\n - cudatoolkit=11.1.74=h6bb024c_0\r\n - ffmpeg=4.2.2=h20bf706_0\r\n - freetype=2.10.4=h5ab3b9f_0\r\n - gmp=6.2.1=h2531618_2\r\n - gnutls=3.6.15=he1e5248_0\r\n - idna=2.10=pyhd3eb1b0_0\r\n - intel-openmp=2021.2.0=h06a4308_610\r\n - jpeg=9b=h024ee3a_2\r\n - lame=3.100=h7b6447c_0\r\n - lcms2=2.12=h3be6417_0\r\n - ld_impl_linux-64=2.35.1=h7274673_9\r\n - libffi=3.3=he6710b0_2\r\n - libgcc-ng=9.3.0=h5101ec6_17\r\n - libgomp=9.3.0=h5101ec6_17\r\n - libidn2=2.3.1=h27cfd23_0\r\n - libopus=1.3.1=h7b6447c_0\r\n - libpng=1.6.37=hbc83047_0\r\n - libstdcxx-ng=9.3.0=hd4cf53a_17\r\n - libtasn1=4.16.0=h27cfd23_0\r\n - libtiff=4.2.0=h85742a9_0\r\n - libunistring=0.9.10=h27cfd23_0\r\n - libuv=1.40.0=h7b6447c_0\r\n - libvpx=1.7.0=h439df22_0\r\n - libwebp-base=1.2.0=h27cfd23_0\r\n - lz4-c=1.9.3=h2531618_0\r\n - mkl=2021.2.0=h06a4308_296\r\n - mkl-service=2.3.0=py38h27cfd23_1\r\n - mkl_fft=1.3.0=py38h42c9631_2\r\n - mkl_random=1.2.1=py38ha9443f7_2\r\n - ncurses=6.2=he6710b0_1\r\n - nettle=3.7.3=hbbd107a_1\r\n - ninja=1.10.2=hff7bd54_1\r\n - numpy=1.20.2=py38h2d18471_0\r\n - numpy-base=1.20.2=py38hfae3a4d_0\r\n - olefile=0.46=py_0\r\n - openh264=2.1.0=hd408876_0\r\n - openssl=1.1.1k=h27cfd23_0\r\n - pillow=8.2.0=py38he98fc37_0\r\n - pip=21.1.2=py38h06a4308_0\r\n - pycparser=2.20=py_2\r\n - pyopenssl=20.0.1=pyhd3eb1b0_1\r\n - pysocks=1.7.1=py38h06a4308_0\r\n - python=3.8.10=h12debd9_8\r\n - pytorch=1.8.1=py3.8_cuda11.1_cudnn8.0.5_0\r\n - readline=8.1=h27cfd23_0\r\n - requests=2.25.1=pyhd3eb1b0_0\r\n - setuptools=52.0.0=py38h06a4308_0\r\n - six=1.16.0=pyhd3eb1b0_0\r\n - sqlite=3.35.4=hdfb4753_0\r\n - tk=8.6.10=hbc83047_0\r\n - torchtext=0.9.1=py38\r\n - torchvision=0.9.1=py38_cu111\r\n - typing_extensions=3.7.4.3=pyha847dfd_0\r\n - urllib3=1.26.4=pyhd3eb1b0_0\r\n - wheel=0.36.2=pyhd3eb1b0_0\r\n - x264=1!157.20191217=h7b6447c_0\r\n - xz=5.2.5=h7b6447c_0\r\n - zlib=1.2.11=h7b6447c_3\r\n - zstd=1.4.9=haebb681_0\r\n - pip:\r\n - click==8.0.1\r\n - cycler==0.10.0\r\n - datasets==1.8.0\r\n - dill==0.3.4\r\n - filelock==3.0.12\r\n - fsspec==2021.6.0\r\n - huggingface-hub==0.0.8\r\n - joblib==1.0.1\r\n - jsonpickle==2.0.0\r\n - kiwisolver==1.3.1\r\n - matplotlib==3.4.2\r\n - multiprocess==0.70.12.2\r\n - packaging==20.9\r\n - pandas==1.2.4\r\n - pyarrow==3.0.0\r\n - pyparsing==2.4.7\r\n - python-dateutil==2.8.1\r\n - pytz==2021.1\r\n - regex==2021.4.4\r\n - sacremoses==0.0.45\r\n - tokenizers==0.10.3\r\n - tqdm==4.49.0\r\n - transformers==4.6.1\r\n - xxhash==2.0.2\r\nprefix: \/home\/mmajurski\/anaconda3\/envs\/round8\r\n```\r\n\nIf you are going to be altering the data cleaning from the source Squad dataset, here is one thing to consider.\r\nThere are occasional double spaces separating words which it might be nice to get rid of. \r\n\r\nEither way, thank you.","embeddings":[-0.2330299318,-0.3295379579,-0.0441096425,0.3749362826,0.0878231078,-0.0763419643,0.1057570204,0.2279032022,-0.2927472591,0.1577830762,-0.0841630697,0.0450688601,0.3250361681,0.0685452819,-0.0926083177,0.1523635834,0.1756330729,0.0104246661,0.009945604,-0.1725040823,-0.2248985022,0.2991383374,-0.2980325818,0.03003061,-0.342564702,-0.0102489283,0.0364246443,0.1131834313,-0.1175084189,-0.4159388542,0.2125024199,-0.0936790556,0.0077113751,0.2638638318,-0.000116531,-0.1706253737,0.0728382915,-0.0973288417,-0.2682839036,0.0041923714,-0.0435589068,-0.1216914803,-0.1298896968,-0.1058420315,-0.2390315384,0.2717562318,-0.0673231781,-0.2866247594,0.6096699834,0.2443222255,0.1451372653,0.2883639336,0.0109265558,-0.064738296,0.1786869168,-0.1235733703,0.0756943524,0.119126901,0.1456331909,0.1079365462,-0.0346998237,0.4713858068,-0.1343523413,-0.0318910703,-0.0061442321,0.2904502153,0.206617564,-0.1278317124,0.0998568758,0.1311751604,0.2598810494,-0.2484471202,-0.424808681,-0.4569473565,-0.0358279571,-0.1224056184,0.3317508101,0.039453499,-0.0311524291,0.3647889495,-0.1365324855,0.0710276216,-0.0857497603,0.127139762,-0.0876931027,0.2383429706,-0.1219492704,0.1492218971,-0.2677045166,-0.1386347562,-0.0631168187,-0.1375988126,-0.105233863,0.4171854258,-0.5139366984,-0.1013480797,-0.1679380983,0.0907929763,0.2361780256,0.0694944859,-0.0699167773,0.0015248328,0.2659206092,0.0613521114,0.3417307138,0.2773298621,-0.1180535555,0.255628705,0.1020577699,0.1474464983,0.0414646193,0.1034675837,0.1996169537,-0.1841570288,-0.1082466468,-0.1547538042,0.1684667319,-0.1887482852,-0.5222857594,0.4338434935,-0.6319300532,0.2756719887,-0.0214355942,0.3482971787,-0.2216847241,0.2669170201,0.3325076103,0.244284451,-0.0650029629,-0.1851683259,-0.133084923,-0.0969801545,-0.0144069865,-0.0344590582,0.0072183558,-0.1293735057,0.2271303385,0.2961053848,0.2128195316,-0.2205923349,-0.3132048249,-0.1648734659,0.1410585493,0.0306854323,0.0271047894,0.1876549721,0.1873004138,-0.2884297669,0.073195301,0.0930062234,-0.2383058369,0.0768734515,-0.0992043242,0.2245198935,-0.1660035253,-0.056426052,0.0465995781,0.2456641197,0.2966586649,-0.2863129377,0.2683805823,-0.203826189,-0.3300075233,-0.0881249383,0.3650223017,0.2920378745,-0.1985461414,-0.0659254789,-0.0561406538,0.1563176215,0.0265871864,0.237526387,0.2730125487,0.2942620516,-0.4212450981,0.2674057782,0.1037338749,-0.8540112376,-0.5590248704,-0.1742199212,-0.1163506806,0.225110665,-0.1036156043,0.0728597566,0.4342785478,0.1543433517,0.2735286653,0.5391799212,0.0679247305,0.1189373881,-0.3740178347,0.0432915017,-0.0926679522,0.1235257834,-0.2477580458,-0.1743770838,0.0530980192,-0.1834742129,0.1845173985,0.1049054489,0.0700077787,0.242408067,0.3292447031,0.1207566783,0.1795453727,-0.5024951696,-0.4840292037,0.0933753848,-0.0992036015,0.0597215667,-0.0624961443,-0.1360670328,-0.3975332677,-0.0361760221,-0.3520284891,-0.3007717729,0.0909299403,0.0083287833,0.1742563844,0.1125237122,-0.1105441302,0.6274502277,-0.2027090788,0.2574426234,-0.0648542717,0.1838029921,-0.0369916596,0.0197878219,0.0257437229,0.4265458286,-0.0366344415,0.0865879729,-0.2453000844,0.5439850688,0.2354308665,0.1975178421,-0.2265350223,-0.0887751728,0.2265448421,-0.4291588366,-0.2648878694,0.2957112789,0.1549057662,0.1138918698,-0.3117320538,0.59599787,0.0686130524,0.0849598721,-0.2245444506,-0.0018041183,0.2800176144,-0.1242726669,-0.1135764197,-0.2309002727,0.2905323505,0.1432120651,0.5214198828,0.1039449498,-0.3891157508,0.1001307517,0.267216295,-0.1695472002,0.0794455707,0.1168767288,0.0412754938,-0.0822434872,0.024764752,0.2209540159,0.2899647057,0.2044537365,-0.3409023285,0.3199863434,-0.036653012,0.1461652517,0.2144389898,-0.1729850173,-0.0400366671,0.4994864166,0.3725070655,0.2427582741,-0.1843069196,0.1724185944,-0.3970791101,0.152523905,-0.5550450087,-0.021352414,-0.1496844441,-0.0735898316,-0.1993011385,-0.0192141831,-0.2774994969,-0.1124377027,0.248433575,0.1559061408,-0.0537631586,0.2769265175,0.0002657939,0.4594179392,0.0978769809,-0.1988540739,-0.0586240068,-0.0650393665,-0.1624031663,-0.0090979105,0.0156265758,0.1998775154,0.1458058804,-0.4826640189,-0.3008002043,-0.2854555845,-0.3453061879,0.2101686448,-0.0686363429,0.3613208234,0.1664168537,-0.0186989903,-0.3620541394,-0.2030521333,0.164502278,-0.0291347131,-0.4773875177,-0.0524751432,0.0556921959,-0.2053315192,-0.0412881449,-0.4563171864,-0.0989391729,-0.2709364295,0.0162285641,-0.1817271709,0.0823787078,0.2436575741,-0.1389108002,0.0348106995,-0.2918559611,0.1260592341,-0.5039397478,-0.0472881943,0.0939144045,0.0430192649,-0.3436332941,-0.2887289226,-0.232024312,-0.1749058217,-0.0350877568,-0.1375848353,-0.2258420736,-0.1793827713,0.2166627795,-0.2268270254,0.2435190976,0.2265461981,-0.1808743179,0.0880100653,-0.2281240225,-0.4454982877,0.1523829252,0.3971565962,0.2353447974,-0.1117967144,0.3300126195,-0.0246539097,0.6546520591,0.4053243101,0.226618126,0.1022702307,-0.2168853581,0.3332633674,-0.2054528892,-0.2187942415,0.312690258,0.1287728995,-0.3271473944,0.3284796774,0.039798025,0.1749130636,0.015510424,0.1942849457,-0.5804231763,-0.1875118911,-0.2166323811,0.0498929955,0.1647057831,0.0902877003,-0.0723394454,0.1494957656,-0.1061443985,-0.0524449386,0.1069306061,-0.0442453772,0.236469686,-0.8369106054,-0.2997891307,0.0433075465,0.2258636355,0.3995067477,0.381581068,-0.0134445364,-0.093993254,0.1968369186,0.0923821777,0.7246232629,0.0159390848,0.0829525888,-0.2899780869,0.1853271425,-0.4381762743,-0.2349051684,-0.1948949248,0.1377697438,0.2811782062,0.1698484123,-0.2208506167,-0.0408705473,0.1911017746,0.0646320134,-0.0984831601,-0.0093899928,-0.2653487027,-0.1490731835,-0.3438111842,0.2851322889,0.4409757853,0.2931801081,0.478546828,-0.0608856529,0.1911562532,0.1568410844,-0.1183883771,0.2444164157,0.5145237446,-0.1394737959,0.020231897,0.2098539174,-0.0472850315,0.0350847431,0.531881988,-0.0985551476,-0.3209655881,-0.0355942957,-0.0423497148,0.3858833313,0.3409575224,0.0271561947,0.007218238,-0.0423274375,0.1705225259,-0.239471674,-0.1145948768,0.3838074207,0.5305989385,-0.0723313093,-0.2432465702,0.226994887,-0.1459895521,0.0304384511,0.0929574072,0.5290966034,-0.2636906207,0.4520533383,-0.1095875725,1.0020316839,0.328188926,0.1134260595,0.497885257,-0.0472013615,0.3578611016,-0.0754101947,0.2748137712,-0.4139471352,0.3110217452,-0.0377646796,-0.1198738515,-0.1178149134,-0.1698682457,-0.2265703827,0.2860626876,-0.1526471376,0.2311641425,-0.0626616031,-0.0085225478,-0.2034142613,-0.033530008,-0.4369697273,0.0932308212,0.014678021,0.2703944445,-0.148771584,0.1750009209,-0.3390996456,-0.3742335439,-0.6190102696,-0.0299747158,-0.4763568342,0.0047969306,0.4499208033,-0.0897761583,-0.0364160649,0.4331648052,0.3017061949,0.0679342151,-0.0369109735,0.1858628243,-0.0765430257,0.3762842119,0.4295200109,-0.0519072898,0.1574102938,-0.0374897495,-0.3992099166,-0.1864605993,0.0142571572,-0.2147545069,0.0705938786,0.1132239476,0.0116750086,-0.180656597,-0.3797228932,-0.1218563616,0.288859576,-0.2152142823,0.1218225509,-0.1044649407,-0.1833918095,0.1680883616,0.2302089781,-0.3828324676,-0.029355986,0.4591567516,-0.2447421104,0.0446338654,0.4361592531,0.0126489829,-0.1020924821,-0.1210553348,0.1857443005,-0.0898654163,-0.1195793897,0.020980645,-0.1442859769,-0.3595216572,-0.2137382925,0.3384000063,0.38391608,-0.1161921322,-0.0502924994,-0.6957364082,-0.0235519614,0.2530730963,0.0438782871,-0.0451187789,0.362829417,0.0801646635,-0.1754003018,-0.0703575015,-0.3165868223,-0.0696842819,-0.2544342875,0.1378873289,-0.0004854769,-0.054904744,0.2277368754,-0.2508426905,0.0973833054,-0.0025261098,-0.1287060976,-0.2157400548,-0.1059362367,0.151648134,0.0958362594,-0.2368230373,-0.1105897501,-0.0767657012,0.1523080468,-0.190477252,0.128565371,0.4913201928,0.151816532,0.1998494267,-0.518520236,-0.1265322566,0.0309705436,0.0234835055,-0.0831899941,-0.3460215032,0.1145583838,0.2857742608,-0.2541635334,-0.0568224117,0.250439167,0.1971273869,0.156353116,0.1079917178,0.1915987283,-0.0982195735,-0.0190188605,0.183392778,0.2271950245,0.3787548244,-0.4000142813,-0.0452516899,-0.0354742371,0.1983865798,-0.3783816695,-0.0351037718,0.2172053754,-0.188268885,-0.0205085445,0.2156239897,0.1364915371,-0.2648781836,-0.0650718063,0.2325160801,0.580716908,-0.0888280421,0.0113576148,0.1476523876,0.0274071768,0.2149521708,0.3176786602,-0.1152953655,0.2021837384,0.1865382344,-0.1520356238,0.0439828038,-0.2052265704,0.1520625502,0.1444454789,0.0466857925,0.3828479052,0.3497182131,-0.4396210313,0.011120189,0.1317227036,0.4576255679,-0.1160483956,-0.2384144217,-0.2936193049,-0.0883242488,-0.2381866127,-0.0578628816,-0.0807035342,-0.1052349582,-0.1390684694,-0.0989655182,-0.0628812537,-0.2211410105,0.1579808146,0.1681921482,-0.2182729393,-0.3492713273,-0.2746627033,0.2447257638,0.1839176714,-0.1713453531,0.341062516,0.0664092451,0.0094089806,0.3050126731,0.5498544574,0.5111480951,0.2144199312,-0.1568036973,0.1962303221,-0.1739758551,-0.0277864691,0.011248799,0.110214144,-0.2986117303,-0.4947999716,0.0919381231,0.2147872746,-0.1876727939,0.01391479,0.0613702126,0.1129160225,-0.3813641965,-0.2247827202,-0.3311124146,0.3151141405,-0.1817768663,0.1097344384,-0.1495829225,-0.0258891806,0.1415275484,-0.1616253704,0.2025576085,0.0786848292,0.0715653673,0.1909437478,0.04452819,0.6243495941,-0.2013409585,-0.0681901425,-0.0578789152,-0.5805677772,0.1312075257,0.0310452115,-0.1154971719,0.2015237659,0.2146641016,-0.1598927528,0.2209586054,0.4334922135,0.0617087074,-0.1903527081,0.0112976078,-0.2144546956,0.0767382309,-0.2031755149,0.1104440317,-0.0068181152,-0.2531497777,0.2053184509,0.1481136978,-0.0577719957,0.1885320842,0.1527107358,-0.0800265297,-0.4327370226,-0.1294350922,0.3209770322,0.2201793194,-0.2354950458,0.1969352812,-0.0664646477,-0.1612056196,-0.5237137675,0.1405882686,-0.3926342726,0.293876946,-0.0463227294,-0.3050654531,0.0475321859,0.2376043051,0.3541375399,-0.2929815948,-0.4807341993,0.1937257349,-0.1500291824,0.2315478325,-0.0873755366,0.1933842897,-0.080139026,0.457072854,-0.3387340903,-0.3580105603,0.3238231838,-0.3490354419,-0.1189630553,0.0141020631,0.3399590254,0.2818251848,-0.1485277414,-0.505400002,-0.1532177329,0.2660462856,-0.0004354485,-0.2164424956,0.1116041318,-0.0506298356,-0.0587325767,0.0299290884,0.0477241166,0.3681152165,-0.1083867252,0.2889886498,-0.3507986069]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2583","title":"Error iteration over IterableDataset using Torch DataLoader","comments":"Hi ! This is because you first need to format the dataset for pytorch:\r\n\r\n```python\r\n>>> import torch\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('oscar', \"unshuffled_deduplicated_en\", split='train', streaming=True)\r\n>>> torch_iterable_dataset = dataset.with_format(\"torch\")\r\n>>> assert isinstance(torch_iterable_dataset, torch.utils.data.IterableDataset)\r\n>>> dataloader = torch.utils.data.DataLoader(torch_iterable_dataset, batch_size=4)\r\n>>> next(iter(dataloader))\r\n{'id': tensor([0, 1, 2, 3]), 'text': ['Mtendere Village was inspired...]}\r\n```\r\n\r\nThis is because the pytorch dataloader expects a subclass of `torch.utils.data.IterableDataset`. Since you can't pass an arbitrary iterable to a pytorch dataloader, you first need to build an object that inherits from `torch.utils.data.IterableDataset` using `with_format(\"torch\")` for example.\r\n","body":"## Describe the bug\r\nI have an IterableDataset (created using streaming=True) and I am trying to create batches using Torch DataLoader class by passing this IterableDataset to it. This throws error which is pasted below. I can do the same by using Torch IterableDataset. One thing I noticed is that in the former case when I look at the dataloader.sampler class I get torch.utils.data.sampler.SequentialSampler while the latter one gives torch.utils.data.dataloader._InfiniteConstantSampler. \r\n\r\nI am not sure if this is how it is meant to be used, but that's what seemed reasonable to me. \r\n\r\n## Steps to reproduce the bug\r\n\r\n1. Does not work.\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('oscar', \"unshuffled_deduplicated_en\", split='train', streaming=True)\r\n>>> dataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\n>>> dataloader.sampler\r\n\r\n>>> for batch in dataloader:\r\n... print(batch)\r\n```\r\n\r\n2. Works.\r\n```python\r\nimport torch\r\nfrom torch.utils.data import Dataset, IterableDataset, DataLoader\r\nclass CustomIterableDataset(IterableDataset):\r\n 'Characterizes a dataset for PyTorch'\r\n def __init__(self, data):\r\n 'Initialization'\r\n self.data = data\r\n\r\n\r\n def __iter__(self):\r\n return iter(self.data)\r\n\r\n\r\ndata = list(range(12))\r\ndataset = CustomIterableDataset(data)\r\ndataloader = DataLoader(dataset, batch_size=4)\r\nprint(\"dataloader: \", dataloader.sampler)\r\nfor batch in dataloader:\r\n print(batch)\r\n```\r\n\r\n## Expected results\r\nTo get batches of data with the batch size as 4. Output from the latter one (2) though Datasource is different here so actual data is different.\r\ndataloader: \r\ntensor([0, 1, 2, 3])\r\ntensor([4, 5, 6, 7])\r\ntensor([ 8, 9, 10, 11])\r\n\r\n## Actual results\r\n\r\n\r\n...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 435, in __next__\r\n data = self._next_data()\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 474, in _next_data\r\n index = self._next_index() # may raise StopIteration\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 427, in _next_index\r\n return next(self._sampler_iter) # may raise StopIteration\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/sampler.py\", line 227, in __iter__\r\n for idx in self.sampler:\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/sampler.py\", line 67, in __iter__\r\n return iter(range(len(self.data_source)))\r\nTypeError: object of type 'IterableDataset' has no len()\r\n\r\n## Environment info\r\n\r\n- `datasets` version: '1.8.1.dev0'\r\n- Platform: Linux\r\n- Python version: Python 3.6.8\r\n- PyArrow version: '3.0.0'\r\n","comment_length":93,"text":"Error iteration over IterableDataset using Torch DataLoader\n## Describe the bug\r\nI have an IterableDataset (created using streaming=True) and I am trying to create batches using Torch DataLoader class by passing this IterableDataset to it. This throws error which is pasted below. I can do the same by using Torch IterableDataset. One thing I noticed is that in the former case when I look at the dataloader.sampler class I get torch.utils.data.sampler.SequentialSampler while the latter one gives torch.utils.data.dataloader._InfiniteConstantSampler. \r\n\r\nI am not sure if this is how it is meant to be used, but that's what seemed reasonable to me. \r\n\r\n## Steps to reproduce the bug\r\n\r\n1. Does not work.\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('oscar', \"unshuffled_deduplicated_en\", split='train', streaming=True)\r\n>>> dataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\n>>> dataloader.sampler\r\n\r\n>>> for batch in dataloader:\r\n... print(batch)\r\n```\r\n\r\n2. Works.\r\n```python\r\nimport torch\r\nfrom torch.utils.data import Dataset, IterableDataset, DataLoader\r\nclass CustomIterableDataset(IterableDataset):\r\n 'Characterizes a dataset for PyTorch'\r\n def __init__(self, data):\r\n 'Initialization'\r\n self.data = data\r\n\r\n\r\n def __iter__(self):\r\n return iter(self.data)\r\n\r\n\r\ndata = list(range(12))\r\ndataset = CustomIterableDataset(data)\r\ndataloader = DataLoader(dataset, batch_size=4)\r\nprint(\"dataloader: \", dataloader.sampler)\r\nfor batch in dataloader:\r\n print(batch)\r\n```\r\n\r\n## Expected results\r\nTo get batches of data with the batch size as 4. Output from the latter one (2) though Datasource is different here so actual data is different.\r\ndataloader: \r\ntensor([0, 1, 2, 3])\r\ntensor([4, 5, 6, 7])\r\ntensor([ 8, 9, 10, 11])\r\n\r\n## Actual results\r\n\r\n\r\n...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 435, in __next__\r\n data = self._next_data()\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 474, in _next_data\r\n index = self._next_index() # may raise StopIteration\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 427, in _next_index\r\n return next(self._sampler_iter) # may raise StopIteration\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/sampler.py\", line 227, in __iter__\r\n for idx in self.sampler:\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/sampler.py\", line 67, in __iter__\r\n return iter(range(len(self.data_source)))\r\nTypeError: object of type 'IterableDataset' has no len()\r\n\r\n## Environment info\r\n\r\n- `datasets` version: '1.8.1.dev0'\r\n- Platform: Linux\r\n- Python version: Python 3.6.8\r\n- PyArrow version: '3.0.0'\r\n\nHi ! This is because you first need to format the dataset for pytorch:\r\n\r\n```python\r\n>>> import torch\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('oscar', \"unshuffled_deduplicated_en\", split='train', streaming=True)\r\n>>> torch_iterable_dataset = dataset.with_format(\"torch\")\r\n>>> assert isinstance(torch_iterable_dataset, torch.utils.data.IterableDataset)\r\n>>> dataloader = torch.utils.data.DataLoader(torch_iterable_dataset, batch_size=4)\r\n>>> next(iter(dataloader))\r\n{'id': tensor([0, 1, 2, 3]), 'text': ['Mtendere Village was inspired...]}\r\n```\r\n\r\nThis is because the pytorch dataloader expects a subclass of `torch.utils.data.IterableDataset`. Since you can't pass an arbitrary iterable to a pytorch dataloader, you first need to build an object that inherits from `torch.utils.data.IterableDataset` using `with_format(\"torch\")` for example.\r\n","embeddings":[-0.1759884804,-0.3325825632,-0.0121410461,0.2365849912,0.1552752554,0.0204123762,0.4311739802,0.0018048278,-0.1689608544,0.2460096478,0.0963452384,0.2442091852,-0.3354809284,-0.4065969288,-0.2495775521,-0.2147274315,0.0861800462,-0.217220217,-0.2279188633,-0.0843941569,-0.1583958417,-0.0314952843,-0.2381762117,-0.0334465913,-0.0213509388,-0.0244186632,-0.120598048,0.0094267903,0.0826323181,-0.5168218017,0.4132375419,0.0784671381,0.5836470127,0.5928797722,-0.0001213014,0.0804407671,0.5695524812,-0.1488481462,-0.3406754136,0.0339830406,0.1853104383,0.1231119111,0.1892292798,-0.2503037155,0.0085637579,-0.1373459846,0.0413325764,-0.4429116547,0.0277451836,0.0911385715,0.127351135,0.1404467672,-0.2152099013,-0.1551946849,0.0873393342,0.0838435441,-0.0660060793,0.3676383197,0.6034711003,0.1462023854,-0.2805481553,0.1023533046,-0.0515720993,0.1446328014,0.2025340199,-0.2065145671,-0.1518067718,-0.1411313266,0.0165851172,0.137726143,0.4296510816,-0.3647227585,-0.2566857934,-0.2134337723,-0.0908371732,-0.0081759226,-0.0515622795,0.2455942184,-0.2517602444,-0.1019132212,-0.1053520739,0.2945753634,-0.2935186625,0.1506047249,0.1208930612,0.0607558563,-0.0605423823,0.4181459546,-0.0258095749,-0.0496773608,0.4104271531,-0.1466057748,0.1139799654,0.1791273206,-0.4098126292,0.1215738282,-0.074593015,0.0343862325,0.1090590954,0.2285431772,0.191299811,-0.0721809641,0.2148056626,-0.0758831427,0.2809829414,0.2104015499,0.1607042253,0.1309028268,-0.0484363362,0.1294634193,0.1207149625,-0.043157395,-0.0131197236,-0.2782546878,0.3315110207,0.2297443449,0.2427385151,0.0218726397,-0.2630295455,-0.0671552569,-0.377215147,-0.046477098,0.0921518356,0.031209385,-0.0663540661,-0.01823931,0.0260313656,0.2194151133,0.0019598796,-0.1363487989,-0.044967562,-0.0634313747,-0.2369847596,0.1653112769,0.0138125839,-0.5259135365,0.2178764939,0.0743197277,0.1240426004,0.0981908739,0.1407571584,-0.1538307667,0.087651819,-0.1886782199,0.1085830554,0.2937149107,0.0909665525,0.3701990843,0.1703479141,0.3358133435,-0.2139513344,-0.4311694205,-0.0041146064,0.1315057129,-0.1349411905,0.1484013498,-0.1661923081,0.0139729884,0.3592304885,-0.086299561,0.2563602626,-0.3309457302,0.0854897127,-0.3074098527,-0.1566838324,0.3201508522,-0.5023243427,-0.0094216177,0.0944096446,-0.1242311895,0.4520845413,0.1644668728,-0.380759865,0.3596851826,-0.4563367665,-0.0022490702,0.0037128867,-0.4484592378,-0.4353860021,0.2547225654,0.0600690469,0.29165712,0.1757151634,0.0443592407,0.4281608462,-0.3241616189,0.1788843125,0.0181607902,-0.2364029586,-0.1766932756,-0.3059003353,0.06542968,0.3222926259,0.1215381771,0.0787648037,0.2113111764,-0.0018342615,0.0364539027,0.2559021413,-0.162260294,0.1733245999,0.2712677717,-0.0221524425,0.1062509865,0.1877069771,-0.0639051124,-0.3826922774,0.1902539432,-0.0341444239,0.0252886638,0.0248268265,-0.0357971936,0.0746423528,0.1515569687,-0.4089162052,-0.1956710517,0.0207006149,0.3267149329,0.2401629686,-0.3410871029,0.051025603,0.177932322,-0.3679131567,-0.0905418396,-0.6530374289,-0.0157553479,0.1605352908,-0.2572544813,0.0066909054,0.0078652846,0.2089659721,-0.2063672841,-0.0949184895,0.2147931606,-0.1238057837,-0.1281567514,-0.1784879714,-0.0105119431,0.0613333322,-0.2036591768,-0.0033433426,0.3811857998,0.2206561118,-0.1643147618,0.1571412086,0.4330125153,-0.1378979385,0.4838735759,-0.0734657273,-0.0011852734,0.0547133498,0.2082166523,-0.4597417116,0.1447856128,0.2501415312,-0.062214449,0.4288068414,-0.1978160143,-0.3633216321,0.0964463502,-0.0651243404,-0.0165036712,-0.133232519,0.176212579,-0.2831240296,0.0837609991,0.1375372857,-0.0159248263,0.6980891824,0.0048870426,0.0211558826,-0.1067589521,0.1212168932,-0.0663993061,0.0172511786,0.2396637946,0.0811793506,0.1726081222,0.0910499394,-0.1229048222,-0.0504769608,-0.4383977354,-0.1312052011,0.1998901069,-0.0802580416,0.300724566,0.0954560488,-0.1390556544,-0.3806454539,-0.5877685547,0.0185129624,-0.2711995542,-0.2329483777,0.5140383244,0.0763324574,0.1100412235,0.0438049771,0.1240867674,0.3323763013,-0.1041971892,-0.0159296971,-0.0392114073,-0.3659599423,-0.0314605087,-0.0236697625,-0.6017323136,0.3704046309,0.1343916655,0.0404745713,-0.479201287,0.1423618495,0.0473591872,-0.0661919266,0.0706465095,0.4086326361,-0.0504630692,0.2938432693,-0.2023874968,0.3072182834,-0.0948523581,0.0008258498,0.0853473917,-0.093167305,-0.0026095277,0.1463578939,-0.0368866585,-0.2398698926,-0.2074588835,-0.0861339569,0.0731796995,-0.0081962924,0.1651799828,0.0242369957,0.0953033939,0.4552520216,-0.1504246593,-0.0639602095,-0.4166634679,0.1938693374,-0.3335467279,-0.2859089077,-0.2267332226,-0.2294795811,0.2497810721,0.3717745543,-0.3056227267,-0.1961064041,-0.0602116659,0.2252768129,-0.0817707479,-0.0425034091,0.2050268203,-0.1281113476,0.0889525265,-0.1457213461,-0.1033764184,0.0690963343,-0.2425730675,0.3543992937,0.2895650268,0.5595616102,0.1378503442,0.9526199102,0.4625449479,-0.1511872262,0.2041900307,0.0080109807,-0.1384082437,-0.0857749134,-0.3842646182,0.215283066,-0.210038051,-0.312305063,0.1660802066,-0.2149699628,-0.1047037169,0.0827315524,0.097253643,-0.0089916717,-0.1037500724,0.175597474,-0.5570787191,0.1904158592,-0.146860525,0.2426728606,-0.2869228125,-0.1258829087,-0.036966078,0.0424440578,0.2459213287,0.0515738204,-0.1893106252,0.0190867279,-0.6127878428,0.3754479289,0.1767864078,0.6406632066,0.1874254346,-0.1155641079,0.1879219562,0.0049159718,0.7799481153,-0.1074881777,-0.2398386002,-0.0036270251,-0.1669650525,-0.7163993716,-0.4133670926,-0.1469399631,0.7751927376,0.1555383503,-0.0751459971,-0.1394633949,-0.0331797041,0.1941219419,0.0733928084,-0.2068856657,-0.1690134257,-0.118754968,-0.3793283403,-0.0894915834,-0.2840992808,0.0517001189,0.1906178743,-0.0854415298,0.0612732656,-0.2614085376,-0.0143883331,0.058354672,-0.0894985497,0.1643391103,-0.2137307078,0.2124439925,-0.0122111337,0.2840583324,0.4879257679,0.2832309008,-0.0035637938,-0.5100503564,0.2887409925,-0.1830687225,0.037848182,0.2061848491,-0.18861036,0.0803613588,-0.1013224572,-0.0213958174,-0.2109725326,-0.0934891105,0.3020628393,0.0530122146,-0.6308019757,-0.5626425743,0.1728119403,0.2680606544,0.0526801869,0.6681536436,-0.3985095918,-0.2715950012,0.2776996791,0.0838837326,0.7267509699,0.1485255957,-0.0557478853,0.232614845,0.2090160102,0.0186763667,0.3115614057,0.2369673103,-0.3453316987,-0.2272444814,-0.0269917212,-0.1897988021,0.122177735,0.182643488,-0.0632820278,0.331823945,-0.1436598897,0.0459797606,0.1270072609,-0.1268932074,-0.212975353,-0.1338790059,0.0956513137,0.0851191282,0.0057658018,0.256490469,-0.1728526652,-0.0926578641,0.0668793842,-0.1555244178,-0.2618394196,0.044389829,-0.2230003327,0.0985847488,0.2456184924,-0.5232536197,0.0501229167,0.2158941329,0.3522298932,-0.362772733,-0.0026628512,-0.0294200219,0.1182429418,-0.0233424194,-0.2029037625,-0.2077227235,0.4870966673,0.0405509211,0.0995516479,0.3205985725,0.0713343248,-0.4681006372,-0.2282389253,0.2966569662,0.4771983027,-0.539290607,-0.2319696546,-0.2508278489,-0.2199194878,-0.2363430858,0.0104449727,0.0206891093,-0.2876823843,0.194111377,-0.174382478,-0.1259075701,0.0919264257,0.5244768262,0.3316209018,-0.0492083058,0.4610216618,0.1272783279,-0.0780189186,0.0172996707,-0.0337951519,0.2742298543,-0.6039997339,0.332097441,-0.3019652367,0.5657223463,-0.047775317,0.3233405352,0.0145815322,0.0645348728,-0.3575257957,-0.2525319457,-0.261600703,-0.1055721492,0.0183102172,0.2605781853,0.0524430908,0.1213143095,0.0727654248,-0.0829664692,-0.1959934682,-0.1801847517,-0.0205957778,0.1617283523,0.1508175433,0.0620185807,-0.1045728847,0.3312746286,-0.0536824539,0.4063277245,0.0199251119,-0.1504729092,-0.0276327562,0.16993469,0.0605212823,-0.1047080979,-0.1692232341,-0.2060013264,0.0063541098,-0.0602118671,0.2894991636,0.2464603037,0.0399105288,0.1423619092,0.3173556924,0.0431836694,-0.0860947818,0.0958950147,0.1015618071,0.337426573,0.1774801463,0.1356684715,0.1120508984,-0.173380658,-0.3201328516,0.1238116249,0.2855196893,0.220807448,-0.0624331348,-0.5807863474,0.1231827065,0.0236730147,0.1875421256,0.3630838394,-0.3631697893,-0.1094984561,0.0990886614,0.0556390882,0.0512083806,-0.0702484548,-0.1832477003,0.0333972834,0.0226572249,0.2685614526,-0.0871393457,-0.227184236,0.1649125814,-0.0110573228,0.3474280238,-0.1377611458,0.1077495366,0.3673091829,0.1531876475,0.0195036884,0.3287825286,0.0738375485,0.0621762462,0.4536164105,0.0433848351,0.6263985634,0.3133954704,-0.2698846459,0.1048175022,-0.3597240746,-0.1315761507,-0.0644463822,0.1428934187,0.2984360754,0.0719249174,0.4628485143,-0.1762643456,-0.315238148,0.0487584695,-0.2423642129,-0.0856386051,-0.1142396703,-0.3930785358,-0.1982114166,0.2366153449,-0.0277272314,-0.0908587649,-0.0120655186,0.1886201799,-0.1014318094,-0.2059575021,-0.5857432485,-0.1734381318,0.2839123011,-0.0624021553,-0.1222302243,0.1527830362,0.1245499253,0.156281963,0.1507200897,0.3055030107,0.498591125,0.5325033069,-0.1081288159,0.1050312147,-0.1505073458,-0.0687116161,-0.0157819428,0.1756821871,-0.1252213269,-0.1440999061,0.4051362276,0.0661148205,-0.0606726147,0.3941460848,0.3221775889,0.0434570909,-0.2341540456,0.1143153235,-0.127859652,0.0654473901,-0.1382031143,0.015658075,-0.3902926445,-0.0595417954,0.3030416071,-0.3966250122,0.1129759476,-0.0389321037,0.0263055712,-0.4196701646,0.3758586943,0.3221401572,0.1738783121,-0.2213015854,-0.1402682811,-0.1009962857,0.1792589575,-0.2492873818,0.1136275679,-0.0460798666,0.3102738559,-0.1072589383,0.1317005008,0.1980299652,0.0027528044,-0.0639494881,0.776920259,-0.2023666054,-0.4061402977,0.0042584776,0.089118354,0.0659047961,-0.4869763255,-0.0633903146,-0.0967294574,0.0390020907,-0.1039431468,-0.2570984662,0.2701437175,-0.0108226789,0.3218630552,0.1759919375,0.0821825713,0.0472046435,-0.03934687,-0.4764406681,-0.0358141996,-0.0921208933,0.1186222658,0.0973648578,0.2948976755,0.0137291569,-0.1338668913,-0.2598083317,0.3552681804,0.3722296953,0.0636513829,-0.3439753056,0.2925575674,-0.4009210765,0.1828521341,0.016919136,0.325688839,0.2241050303,0.3691232204,-0.2892661691,-0.3230032027,0.5461325645,-0.8041620255,-0.3581658602,0.0187149793,-0.1205215231,-0.0071662795,-0.0818941668,-0.3705402911,0.083790116,0.3755952716,-0.109396413,0.0316317715,0.0740495846,0.0892176032,0.1829013824,-0.0885553285,0.1710429788,0.2066063434,-0.082591936,0.2809377909,-0.3231044412]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2583","title":"Error iteration over IterableDataset using Torch DataLoader","comments":"Thank you for that and the example! \r\n\r\nWhat you said makes total sense; I just somehow missed that and assumed HF IterableDataset was a subclass of Torch IterableDataset. ","body":"## Describe the bug\r\nI have an IterableDataset (created using streaming=True) and I am trying to create batches using Torch DataLoader class by passing this IterableDataset to it. This throws error which is pasted below. I can do the same by using Torch IterableDataset. One thing I noticed is that in the former case when I look at the dataloader.sampler class I get torch.utils.data.sampler.SequentialSampler while the latter one gives torch.utils.data.dataloader._InfiniteConstantSampler. \r\n\r\nI am not sure if this is how it is meant to be used, but that's what seemed reasonable to me. \r\n\r\n## Steps to reproduce the bug\r\n\r\n1. Does not work.\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('oscar', \"unshuffled_deduplicated_en\", split='train', streaming=True)\r\n>>> dataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\n>>> dataloader.sampler\r\n\r\n>>> for batch in dataloader:\r\n... print(batch)\r\n```\r\n\r\n2. Works.\r\n```python\r\nimport torch\r\nfrom torch.utils.data import Dataset, IterableDataset, DataLoader\r\nclass CustomIterableDataset(IterableDataset):\r\n 'Characterizes a dataset for PyTorch'\r\n def __init__(self, data):\r\n 'Initialization'\r\n self.data = data\r\n\r\n\r\n def __iter__(self):\r\n return iter(self.data)\r\n\r\n\r\ndata = list(range(12))\r\ndataset = CustomIterableDataset(data)\r\ndataloader = DataLoader(dataset, batch_size=4)\r\nprint(\"dataloader: \", dataloader.sampler)\r\nfor batch in dataloader:\r\n print(batch)\r\n```\r\n\r\n## Expected results\r\nTo get batches of data with the batch size as 4. Output from the latter one (2) though Datasource is different here so actual data is different.\r\ndataloader: \r\ntensor([0, 1, 2, 3])\r\ntensor([4, 5, 6, 7])\r\ntensor([ 8, 9, 10, 11])\r\n\r\n## Actual results\r\n\r\n\r\n...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 435, in __next__\r\n data = self._next_data()\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 474, in _next_data\r\n index = self._next_index() # may raise StopIteration\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 427, in _next_index\r\n return next(self._sampler_iter) # may raise StopIteration\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/sampler.py\", line 227, in __iter__\r\n for idx in self.sampler:\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/sampler.py\", line 67, in __iter__\r\n return iter(range(len(self.data_source)))\r\nTypeError: object of type 'IterableDataset' has no len()\r\n\r\n## Environment info\r\n\r\n- `datasets` version: '1.8.1.dev0'\r\n- Platform: Linux\r\n- Python version: Python 3.6.8\r\n- PyArrow version: '3.0.0'\r\n","comment_length":28,"text":"Error iteration over IterableDataset using Torch DataLoader\n## Describe the bug\r\nI have an IterableDataset (created using streaming=True) and I am trying to create batches using Torch DataLoader class by passing this IterableDataset to it. This throws error which is pasted below. I can do the same by using Torch IterableDataset. One thing I noticed is that in the former case when I look at the dataloader.sampler class I get torch.utils.data.sampler.SequentialSampler while the latter one gives torch.utils.data.dataloader._InfiniteConstantSampler. \r\n\r\nI am not sure if this is how it is meant to be used, but that's what seemed reasonable to me. \r\n\r\n## Steps to reproduce the bug\r\n\r\n1. Does not work.\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('oscar', \"unshuffled_deduplicated_en\", split='train', streaming=True)\r\n>>> dataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\n>>> dataloader.sampler\r\n\r\n>>> for batch in dataloader:\r\n... print(batch)\r\n```\r\n\r\n2. Works.\r\n```python\r\nimport torch\r\nfrom torch.utils.data import Dataset, IterableDataset, DataLoader\r\nclass CustomIterableDataset(IterableDataset):\r\n 'Characterizes a dataset for PyTorch'\r\n def __init__(self, data):\r\n 'Initialization'\r\n self.data = data\r\n\r\n\r\n def __iter__(self):\r\n return iter(self.data)\r\n\r\n\r\ndata = list(range(12))\r\ndataset = CustomIterableDataset(data)\r\ndataloader = DataLoader(dataset, batch_size=4)\r\nprint(\"dataloader: \", dataloader.sampler)\r\nfor batch in dataloader:\r\n print(batch)\r\n```\r\n\r\n## Expected results\r\nTo get batches of data with the batch size as 4. Output from the latter one (2) though Datasource is different here so actual data is different.\r\ndataloader: \r\ntensor([0, 1, 2, 3])\r\ntensor([4, 5, 6, 7])\r\ntensor([ 8, 9, 10, 11])\r\n\r\n## Actual results\r\n\r\n\r\n...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 435, in __next__\r\n data = self._next_data()\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 474, in _next_data\r\n index = self._next_index() # may raise StopIteration\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py\", line 427, in _next_index\r\n return next(self._sampler_iter) # may raise StopIteration\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/sampler.py\", line 227, in __iter__\r\n for idx in self.sampler:\r\n File \"\/data\/leshekha\/lib\/HFDatasets\/lib\/python3.6\/site-packages\/torch\/utils\/data\/sampler.py\", line 67, in __iter__\r\n return iter(range(len(self.data_source)))\r\nTypeError: object of type 'IterableDataset' has no len()\r\n\r\n## Environment info\r\n\r\n- `datasets` version: '1.8.1.dev0'\r\n- Platform: Linux\r\n- Python version: Python 3.6.8\r\n- PyArrow version: '3.0.0'\r\n\nThank you for that and the example! \r\n\r\nWhat you said makes total sense; I just somehow missed that and assumed HF IterableDataset was a subclass of Torch IterableDataset. ","embeddings":[-0.1759884804,-0.3325825632,-0.0121410461,0.2365849912,0.1552752554,0.0204123762,0.4311739802,0.0018048278,-0.1689608544,0.2460096478,0.0963452384,0.2442091852,-0.3354809284,-0.4065969288,-0.2495775521,-0.2147274315,0.0861800462,-0.217220217,-0.2279188633,-0.0843941569,-0.1583958417,-0.0314952843,-0.2381762117,-0.0334465913,-0.0213509388,-0.0244186632,-0.120598048,0.0094267903,0.0826323181,-0.5168218017,0.4132375419,0.0784671381,0.5836470127,0.5928797722,-0.0001213014,0.0804407671,0.5695524812,-0.1488481462,-0.3406754136,0.0339830406,0.1853104383,0.1231119111,0.1892292798,-0.2503037155,0.0085637579,-0.1373459846,0.0413325764,-0.4429116547,0.0277451836,0.0911385715,0.127351135,0.1404467672,-0.2152099013,-0.1551946849,0.0873393342,0.0838435441,-0.0660060793,0.3676383197,0.6034711003,0.1462023854,-0.2805481553,0.1023533046,-0.0515720993,0.1446328014,0.2025340199,-0.2065145671,-0.1518067718,-0.1411313266,0.0165851172,0.137726143,0.4296510816,-0.3647227585,-0.2566857934,-0.2134337723,-0.0908371732,-0.0081759226,-0.0515622795,0.2455942184,-0.2517602444,-0.1019132212,-0.1053520739,0.2945753634,-0.2935186625,0.1506047249,0.1208930612,0.0607558563,-0.0605423823,0.4181459546,-0.0258095749,-0.0496773608,0.4104271531,-0.1466057748,0.1139799654,0.1791273206,-0.4098126292,0.1215738282,-0.074593015,0.0343862325,0.1090590954,0.2285431772,0.191299811,-0.0721809641,0.2148056626,-0.0758831427,0.2809829414,0.2104015499,0.1607042253,0.1309028268,-0.0484363362,0.1294634193,0.1207149625,-0.043157395,-0.0131197236,-0.2782546878,0.3315110207,0.2297443449,0.2427385151,0.0218726397,-0.2630295455,-0.0671552569,-0.377215147,-0.046477098,0.0921518356,0.031209385,-0.0663540661,-0.01823931,0.0260313656,0.2194151133,0.0019598796,-0.1363487989,-0.044967562,-0.0634313747,-0.2369847596,0.1653112769,0.0138125839,-0.5259135365,0.2178764939,0.0743197277,0.1240426004,0.0981908739,0.1407571584,-0.1538307667,0.087651819,-0.1886782199,0.1085830554,0.2937149107,0.0909665525,0.3701990843,0.1703479141,0.3358133435,-0.2139513344,-0.4311694205,-0.0041146064,0.1315057129,-0.1349411905,0.1484013498,-0.1661923081,0.0139729884,0.3592304885,-0.086299561,0.2563602626,-0.3309457302,0.0854897127,-0.3074098527,-0.1566838324,0.3201508522,-0.5023243427,-0.0094216177,0.0944096446,-0.1242311895,0.4520845413,0.1644668728,-0.380759865,0.3596851826,-0.4563367665,-0.0022490702,0.0037128867,-0.4484592378,-0.4353860021,0.2547225654,0.0600690469,0.29165712,0.1757151634,0.0443592407,0.4281608462,-0.3241616189,0.1788843125,0.0181607902,-0.2364029586,-0.1766932756,-0.3059003353,0.06542968,0.3222926259,0.1215381771,0.0787648037,0.2113111764,-0.0018342615,0.0364539027,0.2559021413,-0.162260294,0.1733245999,0.2712677717,-0.0221524425,0.1062509865,0.1877069771,-0.0639051124,-0.3826922774,0.1902539432,-0.0341444239,0.0252886638,0.0248268265,-0.0357971936,0.0746423528,0.1515569687,-0.4089162052,-0.1956710517,0.0207006149,0.3267149329,0.2401629686,-0.3410871029,0.051025603,0.177932322,-0.3679131567,-0.0905418396,-0.6530374289,-0.0157553479,0.1605352908,-0.2572544813,0.0066909054,0.0078652846,0.2089659721,-0.2063672841,-0.0949184895,0.2147931606,-0.1238057837,-0.1281567514,-0.1784879714,-0.0105119431,0.0613333322,-0.2036591768,-0.0033433426,0.3811857998,0.2206561118,-0.1643147618,0.1571412086,0.4330125153,-0.1378979385,0.4838735759,-0.0734657273,-0.0011852734,0.0547133498,0.2082166523,-0.4597417116,0.1447856128,0.2501415312,-0.062214449,0.4288068414,-0.1978160143,-0.3633216321,0.0964463502,-0.0651243404,-0.0165036712,-0.133232519,0.176212579,-0.2831240296,0.0837609991,0.1375372857,-0.0159248263,0.6980891824,0.0048870426,0.0211558826,-0.1067589521,0.1212168932,-0.0663993061,0.0172511786,0.2396637946,0.0811793506,0.1726081222,0.0910499394,-0.1229048222,-0.0504769608,-0.4383977354,-0.1312052011,0.1998901069,-0.0802580416,0.300724566,0.0954560488,-0.1390556544,-0.3806454539,-0.5877685547,0.0185129624,-0.2711995542,-0.2329483777,0.5140383244,0.0763324574,0.1100412235,0.0438049771,0.1240867674,0.3323763013,-0.1041971892,-0.0159296971,-0.0392114073,-0.3659599423,-0.0314605087,-0.0236697625,-0.6017323136,0.3704046309,0.1343916655,0.0404745713,-0.479201287,0.1423618495,0.0473591872,-0.0661919266,0.0706465095,0.4086326361,-0.0504630692,0.2938432693,-0.2023874968,0.3072182834,-0.0948523581,0.0008258498,0.0853473917,-0.093167305,-0.0026095277,0.1463578939,-0.0368866585,-0.2398698926,-0.2074588835,-0.0861339569,0.0731796995,-0.0081962924,0.1651799828,0.0242369957,0.0953033939,0.4552520216,-0.1504246593,-0.0639602095,-0.4166634679,0.1938693374,-0.3335467279,-0.2859089077,-0.2267332226,-0.2294795811,0.2497810721,0.3717745543,-0.3056227267,-0.1961064041,-0.0602116659,0.2252768129,-0.0817707479,-0.0425034091,0.2050268203,-0.1281113476,0.0889525265,-0.1457213461,-0.1033764184,0.0690963343,-0.2425730675,0.3543992937,0.2895650268,0.5595616102,0.1378503442,0.9526199102,0.4625449479,-0.1511872262,0.2041900307,0.0080109807,-0.1384082437,-0.0857749134,-0.3842646182,0.215283066,-0.210038051,-0.312305063,0.1660802066,-0.2149699628,-0.1047037169,0.0827315524,0.097253643,-0.0089916717,-0.1037500724,0.175597474,-0.5570787191,0.1904158592,-0.146860525,0.2426728606,-0.2869228125,-0.1258829087,-0.036966078,0.0424440578,0.2459213287,0.0515738204,-0.1893106252,0.0190867279,-0.6127878428,0.3754479289,0.1767864078,0.6406632066,0.1874254346,-0.1155641079,0.1879219562,0.0049159718,0.7799481153,-0.1074881777,-0.2398386002,-0.0036270251,-0.1669650525,-0.7163993716,-0.4133670926,-0.1469399631,0.7751927376,0.1555383503,-0.0751459971,-0.1394633949,-0.0331797041,0.1941219419,0.0733928084,-0.2068856657,-0.1690134257,-0.118754968,-0.3793283403,-0.0894915834,-0.2840992808,0.0517001189,0.1906178743,-0.0854415298,0.0612732656,-0.2614085376,-0.0143883331,0.058354672,-0.0894985497,0.1643391103,-0.2137307078,0.2124439925,-0.0122111337,0.2840583324,0.4879257679,0.2832309008,-0.0035637938,-0.5100503564,0.2887409925,-0.1830687225,0.037848182,0.2061848491,-0.18861036,0.0803613588,-0.1013224572,-0.0213958174,-0.2109725326,-0.0934891105,0.3020628393,0.0530122146,-0.6308019757,-0.5626425743,0.1728119403,0.2680606544,0.0526801869,0.6681536436,-0.3985095918,-0.2715950012,0.2776996791,0.0838837326,0.7267509699,0.1485255957,-0.0557478853,0.232614845,0.2090160102,0.0186763667,0.3115614057,0.2369673103,-0.3453316987,-0.2272444814,-0.0269917212,-0.1897988021,0.122177735,0.182643488,-0.0632820278,0.331823945,-0.1436598897,0.0459797606,0.1270072609,-0.1268932074,-0.212975353,-0.1338790059,0.0956513137,0.0851191282,0.0057658018,0.256490469,-0.1728526652,-0.0926578641,0.0668793842,-0.1555244178,-0.2618394196,0.044389829,-0.2230003327,0.0985847488,0.2456184924,-0.5232536197,0.0501229167,0.2158941329,0.3522298932,-0.362772733,-0.0026628512,-0.0294200219,0.1182429418,-0.0233424194,-0.2029037625,-0.2077227235,0.4870966673,0.0405509211,0.0995516479,0.3205985725,0.0713343248,-0.4681006372,-0.2282389253,0.2966569662,0.4771983027,-0.539290607,-0.2319696546,-0.2508278489,-0.2199194878,-0.2363430858,0.0104449727,0.0206891093,-0.2876823843,0.194111377,-0.174382478,-0.1259075701,0.0919264257,0.5244768262,0.3316209018,-0.0492083058,0.4610216618,0.1272783279,-0.0780189186,0.0172996707,-0.0337951519,0.2742298543,-0.6039997339,0.332097441,-0.3019652367,0.5657223463,-0.047775317,0.3233405352,0.0145815322,0.0645348728,-0.3575257957,-0.2525319457,-0.261600703,-0.1055721492,0.0183102172,0.2605781853,0.0524430908,0.1213143095,0.0727654248,-0.0829664692,-0.1959934682,-0.1801847517,-0.0205957778,0.1617283523,0.1508175433,0.0620185807,-0.1045728847,0.3312746286,-0.0536824539,0.4063277245,0.0199251119,-0.1504729092,-0.0276327562,0.16993469,0.0605212823,-0.1047080979,-0.1692232341,-0.2060013264,0.0063541098,-0.0602118671,0.2894991636,0.2464603037,0.0399105288,0.1423619092,0.3173556924,0.0431836694,-0.0860947818,0.0958950147,0.1015618071,0.337426573,0.1774801463,0.1356684715,0.1120508984,-0.173380658,-0.3201328516,0.1238116249,0.2855196893,0.220807448,-0.0624331348,-0.5807863474,0.1231827065,0.0236730147,0.1875421256,0.3630838394,-0.3631697893,-0.1094984561,0.0990886614,0.0556390882,0.0512083806,-0.0702484548,-0.1832477003,0.0333972834,0.0226572249,0.2685614526,-0.0871393457,-0.227184236,0.1649125814,-0.0110573228,0.3474280238,-0.1377611458,0.1077495366,0.3673091829,0.1531876475,0.0195036884,0.3287825286,0.0738375485,0.0621762462,0.4536164105,0.0433848351,0.6263985634,0.3133954704,-0.2698846459,0.1048175022,-0.3597240746,-0.1315761507,-0.0644463822,0.1428934187,0.2984360754,0.0719249174,0.4628485143,-0.1762643456,-0.315238148,0.0487584695,-0.2423642129,-0.0856386051,-0.1142396703,-0.3930785358,-0.1982114166,0.2366153449,-0.0277272314,-0.0908587649,-0.0120655186,0.1886201799,-0.1014318094,-0.2059575021,-0.5857432485,-0.1734381318,0.2839123011,-0.0624021553,-0.1222302243,0.1527830362,0.1245499253,0.156281963,0.1507200897,0.3055030107,0.498591125,0.5325033069,-0.1081288159,0.1050312147,-0.1505073458,-0.0687116161,-0.0157819428,0.1756821871,-0.1252213269,-0.1440999061,0.4051362276,0.0661148205,-0.0606726147,0.3941460848,0.3221775889,0.0434570909,-0.2341540456,0.1143153235,-0.127859652,0.0654473901,-0.1382031143,0.015658075,-0.3902926445,-0.0595417954,0.3030416071,-0.3966250122,0.1129759476,-0.0389321037,0.0263055712,-0.4196701646,0.3758586943,0.3221401572,0.1738783121,-0.2213015854,-0.1402682811,-0.1009962857,0.1792589575,-0.2492873818,0.1136275679,-0.0460798666,0.3102738559,-0.1072589383,0.1317005008,0.1980299652,0.0027528044,-0.0639494881,0.776920259,-0.2023666054,-0.4061402977,0.0042584776,0.089118354,0.0659047961,-0.4869763255,-0.0633903146,-0.0967294574,0.0390020907,-0.1039431468,-0.2570984662,0.2701437175,-0.0108226789,0.3218630552,0.1759919375,0.0821825713,0.0472046435,-0.03934687,-0.4764406681,-0.0358141996,-0.0921208933,0.1186222658,0.0973648578,0.2948976755,0.0137291569,-0.1338668913,-0.2598083317,0.3552681804,0.3722296953,0.0636513829,-0.3439753056,0.2925575674,-0.4009210765,0.1828521341,0.016919136,0.325688839,0.2241050303,0.3691232204,-0.2892661691,-0.3230032027,0.5461325645,-0.8041620255,-0.3581658602,0.0187149793,-0.1205215231,-0.0071662795,-0.0818941668,-0.3705402911,0.083790116,0.3755952716,-0.109396413,0.0316317715,0.0740495846,0.0892176032,0.1829013824,-0.0885553285,0.1710429788,0.2066063434,-0.082591936,0.2809377909,-0.3231044412]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2573","title":"Finding right block-size with JSON loading difficult for user","comments":"This was actually a second error arising from a too small block-size in the json reader.\r\n\r\nFinding the right block size is difficult for the layman user","body":"As reported by @thomwolf, while loading a JSON Lines file with \"json\" loading script, he gets\r\n> json.decoder.JSONDecodeError: Extra data: line 2 column 1 (char 383)\r\n","comment_length":27,"text":"Finding right block-size with JSON loading difficult for user\nAs reported by @thomwolf, while loading a JSON Lines file with \"json\" loading script, he gets\r\n> json.decoder.JSONDecodeError: Extra data: line 2 column 1 (char 383)\r\n\nThis was actually a second error arising from a too small block-size in the json reader.\r\n\r\nFinding the right block size is difficult for the layman user","embeddings":[0.0956068933,0.0020218808,-0.2247490883,0.4068727493,0.1341332197,-0.0870383829,0.3515785933,0.3959694803,0.5053705573,0.4565078318,0.2659341097,-0.0984663814,0.0224605687,-0.063838549,-0.0379775129,-0.2568934858,-0.2618601322,0.2319747657,0.1568873078,0.3039900362,-0.1997620016,0.2665425837,0.1062276065,-0.1609105617,0.2236228734,0.0587515533,0.0914319232,0.0959857106,-0.0711931884,-0.4359497428,-0.0447092541,-0.1151688546,-0.0743276477,0.4498323202,-0.0000995917,0.0249896757,0.3473541737,-0.0728502572,-0.4421042204,-0.120011963,-0.060400039,-0.5030124187,0.20351623,-0.199337095,-0.021843465,-0.1310319453,-0.0668413788,-0.1065114066,0.462911278,0.1220876947,0.3715203702,-0.237700671,0.0456450135,0.0361532383,0.2381219566,0.0637214929,0.0411675163,0.3781586289,0.3516416848,0.5234601498,0.1095333919,0.0675432384,-0.0711688772,-0.0298221298,0.0829154477,-0.3186947703,0.2817250192,-0.0485040992,0.1941674054,0.2429829389,0.4865899086,0.4166166186,-0.0474563912,-0.3242163062,0.0322855152,0.1813371778,0.3077044785,0.0795376003,-0.253692627,-0.022642646,-0.2651890218,-0.0648075789,-0.026416013,-0.036925856,-0.0485683642,-0.0989516228,-0.1725890189,-0.0340950973,0.152447924,-0.1861773729,-0.0281526316,-0.1080389991,-0.3090918362,-0.0638663843,0.0473019257,-0.1721327156,0.1723929793,-0.4616991282,0.0737602338,0.0164623838,0.5297000408,0.1309640706,0.124232538,0.0838950947,0.1776370704,-0.0081726704,-0.0359089375,-0.195377633,0.1494318098,0.2763272822,0.1767874658,-0.0923806652,-0.2731758058,-0.2625056207,-0.2382466495,0.0424292795,-0.0853547454,-0.1323301643,-0.0086915521,0.2200216949,-0.429397583,-0.0551880151,0.0649292693,0.1727652103,0.0687273592,0.2640414536,0.2109944671,-0.0350077674,-0.1069791988,-0.0796667859,-0.1454554647,0.0986789018,0.1120595708,-0.1924844682,0.2174410969,0.1487836242,0.1483711451,-0.1516091973,-0.30999735,-0.3423232734,0.1462390721,0.0635312274,-0.1348585039,0.2650139928,0.186412096,-0.1065845564,-0.2369070202,-0.1228933558,-0.180056408,0.2556236088,-0.2207956612,-0.1268772036,0.0242798924,0.3463029563,0.024233982,0.14293167,-0.1664892137,-0.0688213333,0.0027225125,-0.1732430905,0.0776588246,-0.2282051593,0.044811815,-0.2076376379,-0.1221623197,0.1634411961,-0.3751145601,-0.1120071635,0.1969165355,-0.0536480919,0.0594989881,0.3059161901,-0.2070640475,0.2804148495,-0.0869016349,-0.1196012944,0.2914662957,0.1173439994,-0.2984392643,0.5185750127,-0.1299733669,0.2412544191,0.2093473673,-0.029225586,0.1271212846,0.0056954655,0.388454318,0.2863802016,0.1438802481,0.2596033812,-0.3090475798,-0.1199647635,0.0019169644,0.1428799927,-0.0248802826,-0.27548334,-0.0984636024,-0.0076827239,0.1241575181,-0.2261896878,-0.0808173344,0.2229843736,-0.0063510253,-0.1971470416,0.1822516471,0.2619439662,-0.1677093208,0.0242061354,-0.0506117046,-0.2076396495,0.0557606407,0.0121830124,-0.0724804997,-0.1452801079,0.1470714957,0.2062346935,0.3015975356,0.1180322543,0.123223044,-0.2187989354,0.10078381,0.0669476688,0.0789797679,0.0460177325,-0.2172989994,0.0409305245,-0.0568748116,-0.093319878,0.2606039941,-0.0812589452,-0.0284129344,-0.3079867065,-0.2069393396,0.0034053493,0.1463601738,-0.0012324867,-0.1660547256,-0.1716948301,-0.0142641552,0.452152431,0.1640970111,0.3425682187,0.0488318279,-0.1035275534,-0.0192071386,0.3329329193,-0.1527856439,0.093455568,-0.0145512316,-0.2997948229,0.401919961,0.0413827226,0.122253634,-0.0321630612,0.3996909857,0.554302454,-0.0444615781,-0.026263807,-0.1600777805,-0.0489269942,0.4573387802,-0.2258951366,0.0071738209,0.3564730287,0.1343528479,-0.1963520348,-0.105223082,0.1182256564,0.1299024075,0.1888304204,-0.0610684529,-0.1487724781,0.4033785462,-0.4765149951,0.1759030074,-0.0582790934,0.1923301816,0.2282369733,0.1666168869,-0.1296429485,-0.1703408808,-0.2919543982,-0.1598014683,0.1414598823,-0.2160116732,-0.0383913256,-0.1749043018,0.0171440672,0.1160070002,-0.1941159666,0.0634469017,-0.2764542699,0.186484009,0.0857999846,-0.3864562213,-0.2314625233,-0.1342249066,0.4629586041,-0.1204371005,0.388907522,-0.2953087091,-0.0841377974,-0.2262499034,0.2532871366,0.4552177787,-0.0919125453,0.0816128999,-0.274921149,0.0461798646,0.347112149,-0.0816960931,-0.0107905008,0.022841543,-0.0600496754,0.0560306832,0.4496861696,0.0836112052,0.030221235,-0.1500127167,-0.0744829252,-0.2961722612,0.2650219798,-0.0021039881,-0.1048985273,-0.3383951187,-0.2061034441,0.107868664,-0.3975123763,0.4412335753,0.2321575731,0.212052241,-0.3862463832,0.2198593765,0.1508169919,0.1889527887,0.3159148097,-0.2085716128,-0.4334502518,0.1112232432,-0.2329666913,-0.4228050709,0.1300275177,0.1339902282,0.0498610251,-0.2667621374,-0.4891366065,0.3003756106,-0.0521570146,-0.2767086625,-0.1889620572,-0.1978449821,0.2448020577,0.0081736157,-0.1918315291,0.0152293304,-0.2356634438,0.0061501791,0.1423393786,0.148025915,-0.1886839271,0.6826385856,-0.0103884991,0.0446234606,0.1303070486,0.0706444681,0.0657464042,0.3156469762,0.0240764227,-0.2853400409,0.3214404583,0.1970461458,0.143458724,-0.0817008466,0.2895276546,0.0371530838,-0.2142821252,-0.141053766,-0.4289241433,-0.273473978,-0.2403489202,0.3684609532,-0.0589797646,-0.3968171775,-0.2104893774,0.2031517029,-0.173184678,0.0008149679,-0.115225032,0.1386918426,0.1925499588,-0.0058937613,-0.1709821969,-0.0764891356,-0.2906216085,0.5571376085,0.085389331,0.0533546582,-0.0164594837,-0.0685135201,0.1735541672,-0.3187838495,0.3556574881,0.0575480089,-0.0762582272,0.0424537957,0.0448651947,-0.2691235244,-0.0734236166,0.0714506432,0.0639276579,0.3373168707,0.0944461972,-0.3351978958,0.0434361175,0.0778740197,0.0335208327,-0.0920766592,-0.4355370402,-0.1980988681,-0.3307770491,-0.3583378196,0.0288312007,-0.1152273118,0.1873101592,0.0975570828,-0.4058392048,0.2624620199,-0.1972243935,-0.0597405545,0.2077490836,-0.0900273174,-0.1461898386,0.0606456026,0.0556646995,0.2191551775,0.4232973754,0.5937592983,0.3618925214,0.1425871551,0.3213865757,-0.2036596537,0.2121600956,0.2889624834,0.1237675771,-0.0542605929,-0.02534841,0.3709647954,0.1296097338,-0.2740955949,0.0858497173,-0.0309426691,-0.119336687,-0.1653838754,0.2441494465,-0.2641585469,0.1816567183,0.1978206336,-0.2616374195,-0.1393801868,0.3920158446,0.0121423248,0.9516810179,0.0117814811,-0.0407247245,0.2059564441,-0.6339952946,0.2703417242,-0.5762810707,0.0921203196,-0.2677564025,0.4031209052,0.0943103805,-0.1034552157,0.0441919491,0.0501292124,-0.2925863266,-0.0953405499,-0.2573273182,-0.1921393126,-0.319734633,0.3754507899,-0.3018941879,-0.0932731479,-0.2033963799,0.2914934754,-0.131343022,0.0059480234,-0.1016754061,-0.0961300358,-0.2755258381,-0.3008351922,-0.2541140616,0.1800955534,0.0781224221,0.026045166,0.1926259547,-0.0724613741,-0.0118427603,-0.3353956342,-0.0861087069,0.1424119622,-0.0768290535,0.2569777668,-0.247529164,-0.107786946,0.0033482953,0.2006391436,0.1615879536,-0.3208276927,-0.1810965687,0.0709887892,-0.1186242327,-0.1217617393,-0.1446329653,-0.0039851917,-0.4416051507,-0.3048274517,-0.369040221,-0.1454228908,-0.0676659346,0.0177948456,0.3172253966,0.1601400673,0.0666755587,-0.237467289,0.1774479002,-0.2549625337,-0.2523485422,-0.0809461623,0.2365011573,-0.0279063061,0.3630595803,0.0131120672,-0.2868361771,-0.3017170727,0.0583488978,0.3081487715,-0.0034956569,0.1487828791,-0.0847282931,-0.1603409946,0.0976043865,0.0924249738,0.1385640651,0.061239887,0.2169467509,-0.3212396502,-0.2478015721,0.2087906897,0.0137166111,0.2162072808,0.0073690712,-0.0463709831,-0.0944187939,0.2117325515,-0.4128862619,-0.001519372,-0.1317281723,-0.056888707,-0.3676460087,-0.0979962274,-0.1211700067,0.0854511112,0.2758596838,0.3163309395,-0.1931315958,-0.3425486386,-0.0458514653,0.0235786475,0.0598580316,0.0199138708,-0.0880574062,0.0046932781,-0.0312834382,-0.0007756549,0.1649954617,-0.2054679245,-0.0788373873,-0.0732968822,0.0996497646,0.0162744503,-0.051150158,0.2448523492,-0.1358742416,0.6824114323,-0.0861571506,0.0478593707,-0.0614266284,0.1209924817,-0.2335137874,0.2943442166,0.0573895834,0.2390201241,0.3425367773,-0.086735189,0.2997988164,-0.1794269085,-0.1363956183,0.3500252664,-0.2331953645,0.1883656979,-0.1489792168,0.330804497,0.0368777551,-0.0338601433,0.0600362606,-0.0045372848,-0.1960623115,0.086902447,-0.0463104695,-0.2042616159,-0.0384362303,0.1573482305,0.1158953309,-0.0035343766,-0.2411171049,0.0361595973,-0.386270225,-0.1831169575,0.1398923546,-0.0815116391,0.3740171492,0.0527293719,-0.0417466573,-0.1409883946,0.1649287641,0.3472918868,-0.220756188,-0.3389645219,0.3110887706,0.3915059566,-0.1116965041,0.2904826999,0.3967685103,-0.2645716667,0.0052975854,0.3095879555,-0.4420294464,0.0661105737,-0.2967109978,-0.0467391014,-0.3116157949,-0.122461386,-0.2448616624,0.1646090001,-0.0635072514,-0.0268443432,-0.2367588729,0.2867299318,-0.2356090248,0.0839225203,0.0872336626,-0.3086821735,0.0986694768,-0.3815233409,0.197134763,0.2708278298,-0.1181470528,0.2543821633,-0.187121436,0.3644879758,0.2129338235,-0.0197812356,-0.0783834532,-0.0713374242,-0.1270236373,-0.1773276031,0.4721388221,0.4094125032,0.0881333053,0.1295313537,0.314296633,-0.3490211666,0.1909381002,0.1290359497,0.0187682919,-0.1486460119,0.0549853109,-0.253521651,-0.1547024846,-0.2582303584,-0.3317993283,-0.3616253138,-0.3557009697,-0.2150494456,0.0240763016,0.0756877363,-0.056570448,0.1521946341,-0.0117870383,0.4091024697,0.08521761,-0.1188729852,0.1934007853,-0.0815606937,-0.0450483672,0.1781631112,-0.3505958617,0.0393824019,-0.0146680018,0.1155027673,0.1772553474,-0.0137808947,0.1960994452,0.3159481287,-0.0588477366,-0.0466036052,-0.2327634245,0.0024331703,0.4833305478,-0.1535350233,-0.0513487272,-0.088759318,0.2628226876,-0.2804350555,0.2732723653,-0.0750522912,0.0209093075,0.0570633151,0.2809641361,0.2937183678,0.0545738712,0.3408738971,-0.1181644127,-0.2074314803,0.0148196863,-0.0482553095,0.0912162811,0.3455092311,0.1185510606,0.3530785143,-0.0112940641,0.0972548351,-0.0652899072,0.1041337103,-0.5952878594,0.0762316436,-0.1569371372,-0.1008722261,-0.1479348987,0.2144839764,0.1892669201,0.0780683458,0.0398857445,-0.2250813544,-0.2369148582,0.1293319315,0.2331683189,0.1052929386,-0.1817365587,0.249472931,0.016226463,-0.2458003461,-0.0074239424,-0.4358265102,0.2334847897,0.272474885,-0.0710910112,0.0312001426,0.2523408532,0.0391324237,-0.0248585995,-0.1025632992,0.3251886964,0.1013375744,0.0339279957,-0.1910836399,-0.1868948489]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2569","title":"Weights of model checkpoint not initialized for RobertaModel for Bertscore","comments":"Hi @suzyahyah, thanks for reporting.\r\n\r\nThe message you get is indeed not an error message, but a warning coming from Hugging Face `transformers`. The complete warning message is:\r\n```\r\nSome weights of the model checkpoint at roberta-large were not used when initializing RobertaModel: ['lm_head.decoder.weight', 'lm_head.dense.weight', 'lm_head.dense.bias', 'lm_head.layer_norm.bias', 'lm_head.bias', 'lm_head.layer_norm.weight']\r\n- This IS expected if you are initializing RobertaModel from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).\r\n- This IS NOT expected if you are initializing RobertaModel from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).\r\n```\r\n\r\nIn this case, this behavior IS expected and you can safely ignore the warning message.\r\n\r\nThe reason is that you are just using RoBERTa to get the contextual embeddings of the input sentences\/tokens, thus leaving away its head layer, whose weights are ignored.\r\n\r\nFeel free to reopen this issue if you need further explanations.","body":"When applying bertscore out of the box, \r\n\r\n```Some weights of the model checkpoint at roberta-large were not used when initializing RobertaModel: ['lm_head.decoder.weight', 'lm_head.bias', 'lm_head.dense.bias', 'lm_head.layer_norm.bias', 'lm_head.dense.weight', 'lm_head.layer_norm.weight']```\r\n\r\nFollowing the typical usage from https:\/\/huggingface.co\/docs\/datasets\/loading_metrics.html\r\n\r\n```\r\nfrom datasets import load_metric\r\nmetric = load_metric('bertscore')\r\n\r\n# Example of typical usage\r\nfor batch in dataset:\r\n inputs, references = batch\r\n predictions = model(inputs)\r\n metric.add_batch(predictions=predictions, references=references)\r\nscore = metric.compute(lang=\"en\")\r\n#score = metric.compute(model_type=\"roberta-large\") # gives the same error\r\n```\r\n\r\nI am concerned about this because my usage shouldn't require any further fine-tuning and most people would expect to use BertScore out of the box? I realised the huggingface code is a wrapper around https:\/\/github.com\/Tiiiger\/bert_score, but I think this repo is anyway relying on the model code and weights from huggingface repo.... \r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1041-aws-x86_64-with-glibc2.27\r\n- Python version: 3.9.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":167,"text":"Weights of model checkpoint not initialized for RobertaModel for Bertscore\nWhen applying bertscore out of the box, \r\n\r\n```Some weights of the model checkpoint at roberta-large were not used when initializing RobertaModel: ['lm_head.decoder.weight', 'lm_head.bias', 'lm_head.dense.bias', 'lm_head.layer_norm.bias', 'lm_head.dense.weight', 'lm_head.layer_norm.weight']```\r\n\r\nFollowing the typical usage from https:\/\/huggingface.co\/docs\/datasets\/loading_metrics.html\r\n\r\n```\r\nfrom datasets import load_metric\r\nmetric = load_metric('bertscore')\r\n\r\n# Example of typical usage\r\nfor batch in dataset:\r\n inputs, references = batch\r\n predictions = model(inputs)\r\n metric.add_batch(predictions=predictions, references=references)\r\nscore = metric.compute(lang=\"en\")\r\n#score = metric.compute(model_type=\"roberta-large\") # gives the same error\r\n```\r\n\r\nI am concerned about this because my usage shouldn't require any further fine-tuning and most people would expect to use BertScore out of the box? I realised the huggingface code is a wrapper around https:\/\/github.com\/Tiiiger\/bert_score, but I think this repo is anyway relying on the model code and weights from huggingface repo.... \r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1041-aws-x86_64-with-glibc2.27\r\n- Python version: 3.9.5\r\n- PyArrow version: 3.0.0\r\n\nHi @suzyahyah, thanks for reporting.\r\n\r\nThe message you get is indeed not an error message, but a warning coming from Hugging Face `transformers`. The complete warning message is:\r\n```\r\nSome weights of the model checkpoint at roberta-large were not used when initializing RobertaModel: ['lm_head.decoder.weight', 'lm_head.dense.weight', 'lm_head.dense.bias', 'lm_head.layer_norm.bias', 'lm_head.bias', 'lm_head.layer_norm.weight']\r\n- This IS expected if you are initializing RobertaModel from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).\r\n- This IS NOT expected if you are initializing RobertaModel from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).\r\n```\r\n\r\nIn this case, this behavior IS expected and you can safely ignore the warning message.\r\n\r\nThe reason is that you are just using RoBERTa to get the contextual embeddings of the input sentences\/tokens, thus leaving away its head layer, whose weights are ignored.\r\n\r\nFeel free to reopen this issue if you need further explanations.","embeddings":[-0.0595500432,-0.3645298481,0.0901097059,0.1289699376,0.4273766279,0.0748271346,0.2718206644,0.124234505,0.1251102686,0.1333975643,-0.1491384357,0.217213437,-0.1455290765,-0.0643486157,-0.0716609731,-0.2196490318,0.1432188153,-0.0499439314,-0.2094158232,-0.1764126867,-0.1302692592,0.1864649951,-0.2054327428,0.1243986264,-0.3585760295,0.1707725227,0.0284248386,0.3952339292,0.0566756651,-0.4421780407,0.4252820313,-0.0705891028,0.1829725206,0.487798959,-0.0001228714,0.0064606671,0.2213185281,-0.0540871173,-0.0313102044,-0.1773597002,0.0916281044,-0.2407545894,0.1299378872,-0.0846756846,0.0850726888,0.1098496169,-0.0971314088,-0.0917770565,0.4431142807,-0.0203818697,0.0755341873,0.6024700403,0.0763729885,-0.1494729072,0.0460663699,-0.1314279884,0.0313534476,0.7358220816,-0.153144151,-0.0208545029,-0.0985676721,0.3127400577,0.1029277667,0.3258488476,0.2433318794,-0.0050746533,0.0890485644,-0.0216680765,-0.3173093498,0.347205162,-0.0981717333,-0.1046690792,-0.2188250422,-0.2243639082,0.0887015238,-0.2474700511,0.1257462204,-0.0246987063,0.0381265357,0.1720036417,-0.2667013705,-0.0669068918,0.0676952899,-0.0849525183,0.1294219196,-0.1038843542,-0.3312368989,0.0494496226,0.5595440269,0.0194062963,-0.4946108758,-0.0332812555,0.0093517732,0.1292148232,-0.1981736571,0.0252225641,0.2562126815,0.3080791235,0.2015607953,-0.0441699661,0.0902203843,0.0475660749,-0.1666268557,0.0227408167,0.0656780973,0.3708826303,-0.1095082983,-0.0424451455,0.1951088607,0.4521815777,0.2514283657,-0.1259821802,-0.2047226727,-0.1200857311,0.0782651305,-0.1153678373,0.0622206256,-0.1526997685,-0.4366090596,-0.024324052,-0.0479732677,-0.2746985853,0.4581758082,0.2476866394,-0.0365002938,0.2026105672,-0.1451325715,0.3258635104,-0.1985165775,0.0236692689,-0.1916703731,0.0359917991,-0.5499864817,0.1129402071,0.1525399834,-0.514990449,0.3871522546,0.0121976295,0.4898635447,0.0477317236,-0.1363251656,-0.0647763461,0.0756391883,0.3032776117,-0.0555824786,0.3049114943,0.3483037055,0.248380214,-0.1436179429,-0.1127427444,-0.3222931623,-0.4240387082,0.3303536773,-0.0249223299,-0.2013020664,-0.0021148333,-0.6301630735,0.1443105489,0.0616411753,0.0386339538,-0.0829501227,0.1372282505,-0.1097322479,0.0338978805,0.461948365,0.5912770629,-0.4559357762,-0.3831556439,0.0175581947,-0.0373061523,0.2624933422,0.5404408574,-0.2662966847,0.2490768284,-0.1183156371,-0.1341705322,-0.0909342244,-0.2990626991,-0.1987127215,-0.1149462387,0.1860004514,0.0329382047,0.1031875983,-0.3880258799,-0.5773239136,-0.1748377681,0.077793479,-0.1614882648,0.0580865182,0.0355185121,-0.2055669576,-0.2308661193,0.0268766955,0.2811305523,-0.1402849257,0.0905096754,-0.2136240155,0.1375385672,-0.088408336,-0.1537056565,-0.0406620838,-0.0437063091,0.1233562157,-0.1413982064,0.0559836887,0.1983401328,-0.0795108303,0.2182419151,0.0872282758,0.0330903269,0.1745310426,0.0800559595,-0.2793905139,-0.099970907,-0.050696224,-0.0824540108,-0.0876182467,0.0734473541,0.0435826406,-0.0325288326,-0.4597765803,-0.1559244096,-0.3609873354,0.163539663,-0.1609194726,0.0514067635,-0.0361126959,0.0468638912,0.0299032778,0.3188659847,-0.0681626424,-0.2337530851,-0.0063475822,0.368624717,0.0399781242,-0.0959596038,0.2741053998,-0.0035358192,0.1728624254,-0.1739950478,-0.1187923774,-0.1149655432,0.0476638488,0.0078627998,-0.1623607129,0.3741194606,-0.039869532,0.1298767328,0.1223521829,-0.084031418,0.1379922032,0.0420225747,-0.0950801671,-0.2151756138,0.1294486225,0.1283426434,0.5055886507,-0.1493334323,-0.1057035476,-0.5590142608,-0.0101487963,0.023932267,-0.0259494651,0.222036168,-0.260264039,0.0751149952,-0.1623806953,-0.375482142,0.3970047235,0.0489646494,0.0447649397,-0.0641902685,0.0965383574,-0.1662309468,-0.1341766119,0.4033741653,-0.1383150816,0.0557984523,-0.0470858626,0.0068544578,-0.2336542755,0.0577766262,-0.0103694098,0.0481163561,-0.5006847978,0.2451440841,0.0958232656,0.2131485641,-0.2310560346,-0.6032778025,-0.2174278796,-0.1217876971,0.0403481983,-0.0738132745,0.0724710897,0.2759858668,0.3919827342,0.619857192,-0.0538355932,0.1658966094,-0.3506987393,0.0307397451,-0.0841733068,0.0006196868,0.0281954948,-0.2045597732,-0.1048332751,-0.3925685585,-0.0318240859,0.1536333114,-0.3445519805,0.1895877123,-0.3401435018,0.3981795311,0.3173835278,-0.0608042143,-0.1456557065,-0.2803340852,0.3773913085,-0.1592202485,-0.0017578057,-0.0181132909,-0.1346401572,-0.0219188705,-0.0361186415,0.0024108915,-0.0466626883,-0.1200052127,0.2488232106,0.2780738175,-0.1349141896,-0.0081026675,0.1877327263,0.309073925,-0.2970644832,-0.102199845,-0.1530170888,-0.6735306978,0.4321505129,-0.1006259322,-0.2623509169,-0.1679246724,0.0318279415,0.3910511732,-0.4273377955,-0.2776669264,-0.3598842025,-0.1445216686,-0.067840293,0.1398446113,0.3970988691,-0.0778832212,-0.1432484984,0.0223997086,-0.099999927,-0.1330249608,0.0972602144,-0.1907607168,0.3079788387,-0.3312959969,0.2309684902,-0.325683713,0.9833400846,0.208368659,-0.2381806374,0.5505979061,0.1054871976,0.3776525855,-0.178900972,-0.1699359417,0.0257006567,-0.0011686841,0.2771570683,0.1856427789,0.1174837276,0.2526870072,-0.1581128985,-0.2227057666,-0.0586996265,-0.1845900267,-0.092274949,0.1744085699,0.0404517911,0.0434068218,-0.1081390753,-0.1012126505,-0.0601810999,0.2614684701,0.3391654789,0.2163940668,-0.153073132,-0.1500354111,0.0009446493,-0.535390079,0.2343062758,-0.1653895229,-0.1055973694,0.2081577927,-0.3411153555,0.1482871175,0.1174767166,0.4149771929,0.2935397327,0.0004638947,-0.2337073982,-0.0740333945,-0.2577543259,-0.0245777518,-0.2915293276,-0.1611652076,0.5257940292,0.4730908275,-0.4316529036,-0.0833314657,0.1507504135,-0.0110321958,-0.0640361607,-0.1930378526,-0.5834462047,0.0240985751,-0.1810365021,0.1633041203,-0.0103390105,0.1443923712,-0.0586336777,0.3244631886,-0.0369304493,-0.2063413113,0.2530874908,0.3682602942,0.1876194924,0.0778899863,0.2066581249,0.5539930463,-0.0000884141,0.5046098232,0.6657990217,-0.1601139158,-0.379365623,0.1752096713,0.2613639235,0.1779566258,0.2454217523,-0.0718367696,-0.1452245712,0.2186293602,0.3847096562,-0.4183936119,0.2498674095,0.3549114466,0.150960654,-0.1904462576,-0.1720649898,0.4522171319,0.043394845,-0.0110087013,-0.0559039526,0.3944805264,0.0250470135,0.1934130341,0.1315059066,0.9412798882,-0.1404451728,0.289434284,0.2644696832,0.3819971383,0.4347298145,0.0780259147,-0.0474181361,-0.1110303625,0.0099815205,0.0039277594,-0.0804315805,0.1772286594,0.0330030918,-0.1441045851,0.306850493,-0.1944562495,0.0006029397,-0.2966683507,-0.0225429013,-0.0508292839,-0.2607232332,-0.3046509326,-0.0399433859,0.2054139376,0.4303842485,-0.04038186,-0.0531850345,0.0336150005,-0.3173627555,-0.042304825,0.1069529802,-0.0235018544,-0.1178590357,0.4349119365,0.0037764485,0.2492846102,0.0636048168,0.650043726,0.054842785,-0.2759638727,0.3410854638,-0.0260582063,-0.027199151,-0.1503067762,0.0005937011,0.2195633501,0.0923214927,-0.1672540903,-0.0031068176,-0.0732172206,-0.0034423,-0.2601659,0.0199466012,0.1987615526,-0.3643711507,0.0852134675,0.1690531969,0.1256234497,-0.0979291424,0.0158787183,0.0073378896,-0.0804222524,0.0851542056,0.141345039,-0.2557505965,-0.1532671303,0.4424973428,-0.350553304,0.0770265236,0.4665473402,-0.153902337,-0.1273454428,0.0990454406,-0.2718664408,0.760689795,-0.8098618984,-0.0061646975,0.4467017949,0.1676995307,0.1587513238,0.4819998741,0.4868229628,-0.1703998744,-0.0927318931,-0.1318473667,-0.2867135108,0.1372986287,-0.1310386956,0.0091236811,-0.0741150454,0.1504412591,0.0757746249,0.3769184947,-0.1660311967,0.1314229816,0.0432728715,-0.1004372984,-0.0626352653,-0.0242757332,0.5027931333,0.0748706236,-0.0275889579,0.0237128437,-0.2135196477,-0.0519273058,-0.3896521032,0.1608540714,0.0143394768,-0.2431519926,0.0314339437,0.098700434,-0.0299403984,-0.2585273087,-0.0159452837,-0.0764776096,0.0971184298,-0.1120954528,0.5111289024,0.0089679416,-0.0662010759,-0.1408343762,0.1427725703,0.0031545456,0.233493641,-0.1755473614,-0.1201871634,-0.0454078019,0.1187156662,0.0554708056,0.3773248196,0.0398187041,-0.0046973829,0.0470230542,-0.0108899567,-0.1105004027,0.3558961451,0.3031350076,-0.0842780694,-0.2095178515,0.2264949977,0.044575084,-0.2369077355,-0.0932702869,-0.1619830132,0.2636479437,-0.0085690441,0.2554135919,0.1725535095,0.3013778031,0.0927800387,-0.0370462835,0.0561464839,0.0379341319,-0.065473102,0.2098346204,-0.1669711024,0.2181403786,0.7928684354,0.2075893581,0.290872395,0.1803369671,-0.1650341451,0.421248585,-0.4339540303,0.1494487822,-0.2449096441,-0.560225606,-0.0903785527,0.1666412055,0.1425825655,-0.1524408609,-0.193371281,0.7876696587,-0.0848463178,0.0509207249,-0.1816951782,0.200151816,0.0781923011,-0.5429596305,-0.0651720092,-0.1270938665,0.0230900403,-0.0550476611,0.0252030659,-0.1196330786,0.3264279068,-0.050512895,-0.0569638461,-0.1790418476,-0.6971853375,0.2321909666,0.2278520614,-0.0520918146,0.0736198649,0.3492474556,-0.1455153525,0.2686407566,0.4473618269,0.4319784939,0.1288591623,0.2252587229,-0.0570137836,0.0503857844,0.1868867874,-0.1474188268,0.4391914308,0.2323348373,0.1657163054,0.0475359447,0.0522572398,-0.1726900786,0.144165054,-0.3620407283,0.4175055325,-0.2680935562,0.4726895094,-0.2722771466,0.1367379278,-0.3582394421,0.1517602056,-0.3563919663,0.635563314,0.1526313424,-0.0697697029,-0.0993387327,-0.154582575,-0.0141338045,0.1526906043,0.5158993006,0.3221859932,-0.1246201992,-0.5601328015,-0.0678570867,-0.5862094164,0.1026224941,-0.0393406004,0.3152080476,-0.0407281183,0.0282726418,-0.0719601437,0.152036339,-0.0168855581,-0.3900322914,-0.265737474,-0.1565983146,-0.1269951165,0.0841122195,-0.5294066072,-0.0348447673,-0.0058973958,0.0617950559,0.0942763463,0.0456386507,-0.1566915512,-0.016681809,-0.0648738593,-0.070524402,0.3578889668,0.1826927364,0.0087100761,0.3179999888,0.0148989642,-0.0370213017,-0.1153957173,-0.2476243824,-0.140077576,-0.0810342431,-0.1806669533,0.0882509649,-0.3646003604,-0.2100068629,-0.4426750839,0.0865498707,0.2546904087,-0.2744086385,-0.3757106662,0.0107779689,-0.0295321178,-0.1368322968,0.0929704383,0.3709932864,0.0407098494,0.3060778081,-0.2777658403,0.0099572614,0.7441989779,-0.5834292769,0.0964181423,-0.1124395356,0.0488063432,0.2738863826,-0.4745759964,-0.9858583212,0.2380353659,0.2644089758,0.1279473901,-0.4400335848,0.1439383924,-0.2524216175,0.0990390554,-0.1626115292,0.2319542021,-0.0170271769,-0.2271023542,0.03104355,-0.1964699179]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2569","title":"Weights of model checkpoint not initialized for RobertaModel for Bertscore","comments":"Hi @suzyahyah, I have created a Pull Request to filter out that warning message in this specific case, since the behavior is as expected and the warning message can only cause confusion for users (as in your case).","body":"When applying bertscore out of the box, \r\n\r\n```Some weights of the model checkpoint at roberta-large were not used when initializing RobertaModel: ['lm_head.decoder.weight', 'lm_head.bias', 'lm_head.dense.bias', 'lm_head.layer_norm.bias', 'lm_head.dense.weight', 'lm_head.layer_norm.weight']```\r\n\r\nFollowing the typical usage from https:\/\/huggingface.co\/docs\/datasets\/loading_metrics.html\r\n\r\n```\r\nfrom datasets import load_metric\r\nmetric = load_metric('bertscore')\r\n\r\n# Example of typical usage\r\nfor batch in dataset:\r\n inputs, references = batch\r\n predictions = model(inputs)\r\n metric.add_batch(predictions=predictions, references=references)\r\nscore = metric.compute(lang=\"en\")\r\n#score = metric.compute(model_type=\"roberta-large\") # gives the same error\r\n```\r\n\r\nI am concerned about this because my usage shouldn't require any further fine-tuning and most people would expect to use BertScore out of the box? I realised the huggingface code is a wrapper around https:\/\/github.com\/Tiiiger\/bert_score, but I think this repo is anyway relying on the model code and weights from huggingface repo.... \r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1041-aws-x86_64-with-glibc2.27\r\n- Python version: 3.9.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":38,"text":"Weights of model checkpoint not initialized for RobertaModel for Bertscore\nWhen applying bertscore out of the box, \r\n\r\n```Some weights of the model checkpoint at roberta-large were not used when initializing RobertaModel: ['lm_head.decoder.weight', 'lm_head.bias', 'lm_head.dense.bias', 'lm_head.layer_norm.bias', 'lm_head.dense.weight', 'lm_head.layer_norm.weight']```\r\n\r\nFollowing the typical usage from https:\/\/huggingface.co\/docs\/datasets\/loading_metrics.html\r\n\r\n```\r\nfrom datasets import load_metric\r\nmetric = load_metric('bertscore')\r\n\r\n# Example of typical usage\r\nfor batch in dataset:\r\n inputs, references = batch\r\n predictions = model(inputs)\r\n metric.add_batch(predictions=predictions, references=references)\r\nscore = metric.compute(lang=\"en\")\r\n#score = metric.compute(model_type=\"roberta-large\") # gives the same error\r\n```\r\n\r\nI am concerned about this because my usage shouldn't require any further fine-tuning and most people would expect to use BertScore out of the box? I realised the huggingface code is a wrapper around https:\/\/github.com\/Tiiiger\/bert_score, but I think this repo is anyway relying on the model code and weights from huggingface repo.... \r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1041-aws-x86_64-with-glibc2.27\r\n- Python version: 3.9.5\r\n- PyArrow version: 3.0.0\r\n\nHi @suzyahyah, I have created a Pull Request to filter out that warning message in this specific case, since the behavior is as expected and the warning message can only cause confusion for users (as in your case).","embeddings":[-0.1506308615,-0.297534883,0.0790532753,0.0513515696,0.4647553563,0.0405767187,0.1979359239,0.190786466,0.1385852247,0.2084496319,-0.1303769946,0.2869918644,-0.1275676936,-0.086621359,-0.1238216907,-0.1949991286,0.1129312143,-0.0918190628,-0.2066622376,-0.1772622168,-0.168073222,0.1075783893,-0.2270771414,0.0877471119,-0.2647553086,0.1180240363,0.1158174053,0.3999868035,-0.0279442426,-0.549695909,0.4129525721,-0.1024811044,0.1832319349,0.432197988,-0.000118405,-0.0350618102,0.2312027663,-0.0642960742,-0.0272502974,-0.1799430996,0.0714151114,-0.21532619,0.1907900274,-0.0851066485,0.0962506235,0.110652335,-0.0638516322,-0.0977282301,0.2969880998,0.0147003299,0.1060960814,0.5801304579,0.091524452,-0.2033586353,0.0562141798,-0.2434053868,-0.0107169114,0.7996891737,-0.1194128245,0.0211138446,-0.1755490601,0.3603536487,0.0470004603,0.4131556749,0.2314981222,-0.0450551659,0.1458597034,-0.0140760988,-0.2861908376,0.363350749,-0.0495934747,-0.0992543772,-0.1434774399,-0.1704610288,0.1701110154,-0.3235532343,0.1518207192,-0.0207377486,-0.0049407082,0.1767668426,-0.1856806725,-0.0096387072,0.0392065085,-0.0255855788,0.1122339368,-0.1080647111,-0.3088824153,0.0715796649,0.5233258605,0.0219151527,-0.5161144137,0.0442516692,-0.0158390105,0.1050248444,-0.2215551883,0.0371879861,0.276848644,0.3556411266,0.2921352386,-0.051609721,0.1328827888,0.0495795757,-0.1677485108,0.0735968426,0.0619302392,0.3298578262,-0.0936576203,-0.1504183561,0.1854511499,0.4818689823,0.2026166916,-0.1468577385,-0.1522711217,-0.1516358703,0.1331294626,-0.0134107424,0.0515862331,-0.1503574103,-0.4884830117,0.0293980297,-0.0861961246,-0.2325775027,0.4540870488,0.2400120497,-0.0287968498,0.1914060861,-0.2195283026,0.3667712212,-0.1675243676,-0.0024104565,-0.2121832073,0.0200953875,-0.5994926691,0.0451913923,0.1670409292,-0.521114707,0.4407982528,0.0861022994,0.4616357088,0.0439056158,-0.1134123355,-0.1241905019,0.1050330102,0.3450675309,0.0275230482,0.2916050553,0.3321642876,0.2506280839,-0.1550834179,-0.0865121037,-0.3208598793,-0.4442728162,0.3166167736,0.0127649326,-0.12097352,-0.0538565628,-0.6307356954,0.1684771478,0.0612261593,0.0587103777,-0.0680129528,0.1115413606,-0.1867161691,-0.014588085,0.371406734,0.5238423944,-0.5235716701,-0.2953406274,0.0224866439,-0.0566582307,0.3146739006,0.4856605828,-0.2403754294,0.2253071368,-0.1106553748,-0.1308024079,-0.0051892768,-0.324064523,-0.1994791031,-0.055202812,0.1629431844,0.0029754476,0.1301812232,-0.4134738743,-0.5584423542,-0.1973029375,0.1586217284,-0.1974609792,0.0787228346,0.0256364979,-0.2331430912,-0.2044692934,0.0141701829,0.2407873869,-0.1170875207,0.0346441828,-0.1564632952,0.0461315364,-0.0578327477,-0.1170396432,-0.0120325265,-0.1053555831,0.1301748604,-0.1696342826,0.0152136404,0.1239441633,-0.0621881448,0.2335908562,0.1463427544,0.0846133754,0.2386511117,0.0692473352,-0.2949136198,-0.0895646289,-0.0579354838,-0.0117454203,-0.0491656289,0.0096058138,0.014750029,0.0431583896,-0.4249891639,-0.2600969076,-0.4469791353,0.1492612511,-0.0618759692,-0.0507430844,-0.0595060028,0.0249604285,0.0614042431,0.2770500779,-0.0201047286,-0.1134727448,-0.0162944812,0.3425474763,0.0704258457,-0.1281533688,0.2877139151,-0.0041894238,0.1435242444,-0.0778608173,-0.0890887976,-0.1550526321,0.0574499778,0.0166235436,-0.1131711453,0.3734732568,0.0200649891,0.0996995345,0.0735728964,-0.0686987117,0.1884991378,0.0305553917,0.0090344138,-0.2739720643,0.1552571803,0.1375312805,0.4346508682,-0.1644020975,-0.1481812298,-0.4836677313,0.0090099173,0.0718987063,0.0309538599,0.2066996843,-0.2137859315,0.0322789662,-0.1540057659,-0.3394297361,0.441845119,0.0757387802,0.13170816,-0.0668469593,0.0565903634,-0.2283736914,-0.0916003287,0.3643068075,-0.0837100744,0.1223224849,-0.0144600356,0.0050253039,-0.2743749321,0.1103991047,0.0533314422,0.1116011888,-0.4886286557,0.2111949772,0.0160638969,0.207506299,-0.2277429998,-0.5473071337,-0.1978655905,-0.1674134731,0.1108879074,0.0035506173,0.0155517273,0.245834291,0.3519870043,0.5763667226,-0.0105392141,0.1820797473,-0.3252362013,0.0379519351,-0.0824356228,0.0348154306,0.0212817378,-0.1395526379,-0.0369860195,-0.2842594087,-0.0685697123,0.1353874356,-0.3167825043,0.1629516482,-0.3225497305,0.4349209368,0.3414088786,-0.0739268288,-0.1379303783,-0.241278097,0.3552574217,-0.173766762,-0.0142197516,-0.045967672,-0.1685479283,-0.0038151995,-0.0415971838,-0.0552913882,-0.0487063453,-0.1575495154,0.1859256625,0.2891166508,-0.1052072197,-0.1152042001,0.1889691502,0.2362823486,-0.3005891442,-0.1001724005,-0.2634636164,-0.711925149,0.3826666772,-0.1397305578,-0.2567561269,-0.1022106484,0.0898858234,0.4123462141,-0.4059282541,-0.3205484748,-0.4146136343,-0.0838825405,-0.0104269674,0.1121117994,0.3206905425,-0.1259178519,-0.1657734066,-0.0424434617,-0.1663160324,-0.0597809553,0.0870635062,-0.2837340534,0.3143175542,-0.3372662663,0.2276594788,-0.275947541,0.9847343564,0.2884800732,-0.2326099277,0.6071634293,0.1287951469,0.3356271088,-0.2486676723,-0.1645040512,0.0805723518,-0.0062946249,0.3026959002,0.2106640935,0.0329568088,0.2064095289,-0.1247750819,-0.2830570936,-0.0179487858,-0.2100504786,-0.0514960177,0.1163407415,-0.0407658443,0.0494884364,-0.1192303523,-0.0830015168,-0.094812721,0.2419172227,0.3854697943,0.1510196775,-0.1535047889,-0.1704185009,0.0412966348,-0.6059466004,0.2302510589,-0.146732375,-0.0699241683,0.1800374836,-0.2958262265,0.1627699733,0.1540587097,0.3248283267,0.149426654,0.0431981757,-0.2296824604,0.0439063273,-0.199987486,-0.0090323994,-0.2797452211,-0.0904327407,0.5518627167,0.3548371792,-0.431806922,-0.126186654,0.1528724581,-0.0610856675,-0.0218940247,-0.2179819494,-0.5934867263,0.0278048944,-0.1854323894,0.2032898664,0.0351571925,0.1893443763,-0.0960712284,0.2732439339,-0.0187803283,-0.1408814341,0.2616773844,0.2927852869,0.1352367252,0.1493744701,0.2491376549,0.4505408108,0.0821149126,0.5949651003,0.6329041719,-0.1401255578,-0.3148524463,0.1595425159,0.2627013028,0.1785634309,0.2328179479,-0.0730757192,-0.0564229861,0.1428257376,0.366260469,-0.3818347156,0.2371979505,0.3204720318,0.1241927668,-0.1649454981,-0.1584479809,0.4748922586,0.0454471894,-0.0617037416,0.0083014769,0.3129754663,-0.008951704,0.1839712858,0.1459771544,0.8244406581,-0.0699422732,0.2399156243,0.3293839693,0.4202503264,0.3364492655,0.0560554937,-0.0588701889,-0.1274717748,0.0277729593,0.0196382664,-0.0679951236,0.1738191992,0.0013341047,-0.1454503238,0.3617592454,-0.2176555842,-0.0577369146,-0.2969801128,-0.0312656313,-0.1756336242,-0.273486644,-0.3289920092,-0.0107205724,0.1990984976,0.4005931914,-0.0305472054,-0.0526181161,0.0823788494,-0.2989168465,-0.0793270692,0.0563017242,-0.0462608933,-0.1133680269,0.3503374755,0.0046777143,0.2855458856,0.0208186917,0.6671979427,0.0981374606,-0.2601712048,0.3924080431,0.0511799455,-0.0457669981,-0.1384091079,-0.0449973829,0.2247891575,0.0273371451,-0.1969127357,-0.0282545909,-0.0863884464,0.0070136185,-0.2205931544,-0.0307057798,0.2182715088,-0.4100441337,0.0399590395,0.1802456975,0.2025473565,-0.1137860194,0.0380155556,0.073240146,-0.0505742393,0.1269164979,0.1035780087,-0.244387731,-0.1462875158,0.4675122201,-0.3212060928,0.05447153,0.4901280105,-0.1703803837,-0.1374824196,0.0577415675,-0.2169572413,0.7585762739,-0.7254886031,-0.009794008,0.4498122334,0.1767362058,0.2023490071,0.4736381769,0.4459123313,-0.1422871798,-0.0326176547,-0.2269625664,-0.2666088641,0.1268623322,-0.1752666235,-0.0224275663,-0.1089373827,0.1597594619,0.057791356,0.3241817951,-0.2059388459,0.1709911972,0.0463718437,-0.0893223435,-0.1259393692,-0.0153222047,0.5025229454,0.1386449188,0.0006159012,0.0289893951,-0.2684345543,-0.0892339945,-0.4054860473,0.1571566612,-0.0036493139,-0.2023523301,0.0257345773,0.0387168936,-0.0907378122,-0.2082268894,-0.0270166509,-0.1116173044,0.0785916001,-0.1236197576,0.5146752,0.0623298772,0.0086635286,-0.1430716962,0.0994519666,0.0091000283,0.1324457973,-0.2453328818,-0.1330103278,-0.0814004168,0.1160520837,0.0345563255,0.2790315747,0.0890147388,-0.0414735079,0.0757666454,-0.0428216346,-0.1139972582,0.3317842484,0.2764271498,-0.1133795008,-0.2586683929,0.2536015511,0.0940288305,-0.2695655227,-0.0959036723,-0.1607613266,0.3303249776,0.0078529688,0.183161661,0.177429229,0.2701616287,0.0747521669,-0.0377350412,-0.0785819292,-0.0005341077,-0.0852169245,0.1470084488,-0.1624278426,0.1945862621,0.7594560385,0.2029119879,0.3006829023,0.2762670517,-0.148109138,0.475481987,-0.3361104131,0.075734809,-0.2649821341,-0.6027315259,-0.0962197036,0.2004272938,0.044958286,-0.1512364745,-0.0951129273,0.7941794991,-0.1164684147,0.009147739,-0.2164340764,0.169070974,0.0232783426,-0.5665177703,-0.1921999902,-0.1465446055,0.098167181,-0.0266948044,0.0336825214,-0.0354320295,0.3262622356,-0.0325410664,-0.0700089633,-0.2572225332,-0.685357511,0.2440310717,0.2412668765,-0.041725345,0.0630862862,0.3763276339,-0.1548972279,0.2460984737,0.4070383608,0.4606924951,0.1535179615,0.2155493647,-0.1659933031,0.0658780411,0.1538668275,-0.1654477119,0.4261158407,0.3403889537,0.1815124303,0.1295982897,0.0511618406,-0.1931253225,0.1973882765,-0.2678566873,0.4426292777,-0.2131014466,0.4483055472,-0.2437185496,0.1948981583,-0.368203789,0.167910561,-0.3469511867,0.6887559891,0.1995061785,-0.0317157209,-0.0457602218,-0.2536663115,0.0153081398,0.160255298,0.5332732201,0.2655237317,-0.1902430356,-0.6655281186,-0.0283840429,-0.6237758994,0.1466343999,-0.0807162896,0.3106873631,-0.0688114464,0.0719400719,-0.0749488175,0.2002549917,0.0032365718,-0.4071140587,-0.1927724034,-0.0897803828,-0.204093188,0.0571961328,-0.5257990956,-0.0884627849,0.0108543457,0.007926668,0.1253991425,-0.0568900034,-0.0977352113,-0.0592689663,-0.050639566,-0.0246365685,0.4130625427,0.2529041171,0.0334224664,0.3423547149,0.0045939628,-0.021417357,-0.1209856048,-0.2568950951,-0.1487036198,-0.0593312718,-0.1564587206,0.0970749408,-0.2969249487,-0.1412892938,-0.4015672207,0.105599463,0.1789361238,-0.2523504496,-0.3688322008,0.0523611456,0.0418275408,-0.0956951827,0.0365738682,0.3607118428,0.0672344863,0.3435561061,-0.3030108809,-0.0911935344,0.757011652,-0.5725848079,0.0790633708,-0.1314101517,0.0740841106,0.2822542787,-0.467666328,-1.0831574202,0.2719725072,0.2745968401,0.1263022274,-0.4521111548,0.1626844704,-0.2787555158,0.1043330133,-0.1492418796,0.2096953988,-0.0098473076,-0.352124393,0.0244109835,-0.2266332954]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2561","title":"Existing cache for local dataset builder file updates is ignored with `ignore_verifications=True`","comments":"Hi ! I just tried to reproduce what you said:\r\n- create a local builder class\r\n- use `load_dataset`\r\n- update the builder class code\r\n- use `load_dataset` again (with or without `ignore_verifications=True`)\r\nAnd it creates a new cache, as expected.\r\n\r\nWhat modifications did you do to your builder's code ?","body":"## Describe the bug\r\nIf i have local file defining a dataset builder class and I load it using `load_dataset` functionality, the existing cache is ignored whenever the file is update even with `ignore_verifications=True`. This slows down debugging and cache generator for very large datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\n- Create a local dataset builder class\r\n- load the local builder class file using `load_dataset` and let the cache build\r\n- update the file's content\r\n- The cache should rebuilt.\r\n\r\n## Expected results\r\n\r\nWith `ignore_verifications=True`, `load_dataset` should pick up existing cache.\r\n\r\n## Actual results\r\n\r\nCreates new cache.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\r\n","comment_length":51,"text":"Existing cache for local dataset builder file updates is ignored with `ignore_verifications=True`\n## Describe the bug\r\nIf i have local file defining a dataset builder class and I load it using `load_dataset` functionality, the existing cache is ignored whenever the file is update even with `ignore_verifications=True`. This slows down debugging and cache generator for very large datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\n- Create a local dataset builder class\r\n- load the local builder class file using `load_dataset` and let the cache build\r\n- update the file's content\r\n- The cache should rebuilt.\r\n\r\n## Expected results\r\n\r\nWith `ignore_verifications=True`, `load_dataset` should pick up existing cache.\r\n\r\n## Actual results\r\n\r\nCreates new cache.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\r\n\nHi ! I just tried to reproduce what you said:\r\n- create a local builder class\r\n- use `load_dataset`\r\n- update the builder class code\r\n- use `load_dataset` again (with or without `ignore_verifications=True`)\r\nAnd it creates a new cache, as expected.\r\n\r\nWhat modifications did you do to your builder's code ?","embeddings":[-0.380481869,0.4886573255,0.0326324254,0.1935642362,0.1446633488,0.1507260799,0.3192700744,0.3748850226,0.1290490031,0.2183320373,0.2149606794,0.3630369902,0.1198911443,-0.2817024291,0.0510923229,0.2790060341,0.1163431108,0.143495515,0.21445328,-0.0771563873,-0.028279366,-0.0800596401,-0.0401381962,-0.2102086246,-0.4505546987,0.1976117641,0.0528690033,0.0853829533,-0.0405516736,-0.466229558,0.3908791542,0.2441875637,-0.0562269986,0.2653215528,-0.0001090661,0.1760650277,0.3208662868,0.0526437834,-0.3978691399,0.1389088631,-0.3474245369,-0.4606624544,0.0223906413,-0.0884896964,-0.2164719403,-0.2490846813,-0.1407262981,-0.191591084,0.3035663664,0.0564859174,0.2780434191,0.0750867352,-0.1023399532,-0.0168681145,0.1199553385,-0.0507756919,-0.053615395,0.1609927416,0.2253709286,0.0361661166,-0.430483222,0.3174596727,-0.1896671653,0.2203765512,0.4303051829,-0.0973287672,0.16110304,-0.0093295556,0.377348423,0.0414983593,0.5871434808,-0.3299917579,-0.2986133099,-0.2611524761,-0.1944366395,-0.0155707281,0.2321196795,-0.183269605,-0.2992415428,0.1735045016,-0.4138687849,-0.2707409859,0.067833133,-0.1559703946,0.0528805628,-0.2511184216,0.0992153659,-0.0926114172,-0.089377813,0.1688511074,0.4349325597,-0.468609035,-0.0739733875,0.2702969313,-0.4989158809,0.0219631698,0.3845898807,0.3972643912,0.238409996,0.302289784,-0.1899261922,0.2409279048,0.1860504299,0.0968388021,-0.0826028883,0.3067308366,0.0410077795,-0.0109668709,0.2831624448,0.2031198889,-0.5477035046,-0.0137982024,0.0638632923,-0.2574612796,0.5983974338,-0.0898954123,0.0680672005,-0.328872472,-0.0316492282,0.1889811307,-0.0006035879,-0.1473223269,0.2239474058,0.2570385039,-0.1294848025,0.0850459188,-0.2500111759,-0.2127594501,-0.3465076685,-0.354465872,-0.3988960683,-0.2386138588,0.0201397929,0.3679382503,0.485526979,-0.5491024852,0.3118240833,0.2401132435,-0.061511606,0.0796848014,0.1753485054,-0.1361715347,0.1791896224,0.2202351242,-0.2889445424,0.0660595745,-0.0249982737,-0.1994285434,-0.0721063465,0.3408291936,-0.1144278422,-0.7981068492,0.0093575949,0.1745110154,-0.2340953797,0.0518296994,0.283639878,0.1155551523,0.4194816649,-0.3065300584,-0.1302641928,-0.0479084477,-0.2993509173,-0.1270203888,0.4868237972,0.6713590622,-0.4300766587,0.0946420655,-0.124225527,0.0880809203,0.0045686769,0.1541885585,-0.3215012252,0.4164039791,-0.2850262523,-0.4484735131,0.2019233555,-0.2974612415,-0.829376936,0.2752141654,0.1786656231,0.2141062915,0.1349737346,0.039904274,-0.0856212005,-0.1601605862,0.097990945,0.2460227013,-0.2004423738,0.0236593261,-0.4308620989,-0.1939272434,0.0640484244,0.0740330741,0.0089768032,0.2423089296,0.0550841652,0.1132432595,0.0285036881,0.008789924,0.134250015,0.383466959,0.1400841326,-0.1290516704,0.2051851153,-0.1118596196,-0.4544261098,0.4190022349,-0.2235322893,-0.1305979639,0.0498482175,-0.299634099,-0.1256818622,-0.0823088884,-0.4173101783,-0.2251382321,0.1377799064,0.1733632386,0.4545215666,-0.0684815571,-0.1343213469,0.2744274139,-0.171483323,0.159884885,-0.0228320528,-0.0699828118,-0.078765139,-0.0692961961,-0.3138266504,0.0410866849,0.1141755059,-0.0592819937,-0.2225729078,0.2792674899,0.0170241632,0.0943570212,0.0013861436,0.1983225644,0.0460142158,0.1652133614,0.1288751215,-0.1250752062,0.0781465024,-0.0983776748,-0.0527022667,0.0726588592,0.2021352798,0.1724301577,-0.2249980867,-0.3149552941,0.2511095107,-0.1476520598,0.0871010423,-0.0884525925,-0.0426745564,0.0656267852,0.4767363369,0.2285358906,-0.0164817311,0.168270871,0.4884268045,0.2000714242,-0.2914832234,0.0221241917,-0.1331483424,-0.2133834809,0.1711711138,0.3344317675,0.3634784222,0.1734696627,0.2099324018,0.0661090314,-0.0327125154,-0.1634075195,0.0613576658,-0.1118580773,-0.1488316953,0.4112389982,0.3244159222,0.1536159664,-0.3777528703,0.2912766635,0.1497383416,0.1518630385,-0.0133895529,0.1734291315,-0.4750730991,-0.111807391,-0.1549550593,0.1557514071,-0.3087697923,-0.1850283742,0.1412915736,0.1024342999,-0.0044620861,0.2299690992,-0.0596316159,0.314935118,0.0060207392,-0.5304799676,-0.3073922098,-0.0541435517,-0.4707733989,-0.0123460377,0.4609992206,-0.242698729,0.2919667661,-0.2242351323,-0.0838991329,-0.202625528,-0.2120319456,0.0953063741,0.1066370159,0.4123833776,-0.1788090318,-0.0082731778,-0.0438163728,0.0297020208,0.2912225723,-0.1716905236,-0.1785908192,-0.1113298386,-0.0883523896,0.1494503766,-0.2810606062,-0.2027480602,0.0114694666,-0.378338486,-0.1207697615,0.1415114999,0.0631860867,0.1967853606,-0.0582334176,0.1748851538,0.0205823891,0.1483393461,-0.2669385076,-0.6465246081,0.2283112556,-0.2430549413,-0.0668883547,0.0384100638,-0.3015871942,0.1177911907,-0.0815823972,-0.4913438261,-0.0105528599,-0.4558912516,0.3086244464,-0.0715983137,0.0444752313,0.4206596911,-0.002462599,-0.0336006358,-0.1908629388,-0.3150304258,0.1238770708,-0.1581401676,0.3299719989,0.0541929416,-0.0427867249,0.1322250515,0.5684074759,0.145355016,-0.0803417638,0.3159600496,0.1188294739,0.4034343362,-0.3024224937,-0.2068282366,0.0561797544,-0.0480517596,-0.2586320639,0.0179704074,0.0922228396,0.0153533928,-0.1323777586,-0.024140453,0.0221359171,-0.2735556662,0.00252109,-0.3864329457,0.2015069127,0.4415509403,0.1765926033,-0.1888560504,-0.1417000145,0.1498922408,0.2254766971,0.1601369679,-0.045687966,0.0659559965,0.2351466119,-0.3701852262,0.1026782468,0.0984814614,-0.0682747662,0.0720146969,0.0008163174,0.2745981812,-0.0770791098,0.2790735364,-0.4338414967,0.0234027598,0.2077310383,-0.0449019559,0.0446939655,-0.2526711524,0.2096978873,0.2986282706,-0.150401473,0.27864182,-0.0400711484,-0.2590215504,0.0644097701,0.3787697554,-0.1512272656,-0.4095380306,-0.2608659565,0.0080990838,-0.1622662544,0.0933761075,-0.1808726043,0.2914334834,-0.0097641619,-0.0358071029,-0.1010073423,0.3268053532,-0.1703639477,0.0109592555,0.2811035216,-0.1458037794,0.3419135511,0.0604261272,0.0640219972,0.1709908545,0.4756256044,-0.0876848176,-0.208173722,-0.1658854336,-0.0252961051,-0.0827368051,0.234068498,-0.0805177167,-0.1607118994,0.2102657109,-0.0868855491,-0.0937161371,-0.0870414972,0.0128914639,-0.0236869361,0.0838645697,-0.4713274837,0.1676433533,0.1242035329,-0.0518050119,0.3790417016,-0.2187293023,0.0275238417,0.2006244063,0.1142494157,0.6171649098,0.122116223,-0.0726749599,0.0029372554,-0.2452917248,0.341504544,-0.3462281823,0.2468890101,-0.2816813588,-0.2803459764,0.1727059782,-0.1924578995,0.2364638597,0.0897737443,-0.0701668337,0.1336214691,-0.0885071531,0.2817087173,-0.0107069295,0.2285520732,-0.1176189929,-0.1444789469,-0.0195711702,0.2138471156,0.012485303,0.335318476,-0.058260031,0.0159007609,0.1801537424,-0.0808224902,-0.1948034912,-0.1231329963,-0.1186831817,0.0012879252,0.284040153,-0.2037246376,0.0151566174,0.0649265498,0.2296742946,0.1114740074,-0.1871330291,0.1602970511,0.1367701292,0.3419645429,-0.1230106726,-0.2972232401,0.3527776897,-0.0682600513,-0.1647280753,-0.195394218,-0.5220221877,-0.4138992429,0.260061115,0.1896180958,0.3121992946,0.078032665,0.100407958,-0.0140930349,0.2869430184,-0.2069870383,0.2122658491,0.1076693609,-0.1670370251,0.1476463526,-0.339902997,-0.2350502461,-0.0520666875,0.3169326484,0.224697113,-0.1340264231,0.4436530471,-0.2458906621,-0.0958602205,-0.1339603513,-0.0281456318,0.1562257409,-0.2629675567,-0.0357556567,0.0425796695,0.0624866486,0.1657258272,-0.0400656089,0.0600907579,0.0544878654,-0.3269216716,-0.1526608616,-0.3810037076,-0.2966793478,-0.1081871763,0.0335305221,0.0041417032,-0.0133238481,0.2901389599,0.3709746003,-0.3479486108,-0.0777553841,-0.253752023,0.1006593928,-0.3202537596,0.1699164659,0.073723495,-0.0583094843,-0.016535731,0.0942819566,-0.3026222885,-0.2772940695,-0.0129699344,0.1127234697,0.0548074357,0.2695973516,-0.1076488867,0.0404739082,0.0232927855,-0.0481975004,0.2192638367,-0.1349796653,-0.1380728781,0.2205067724,0.1599507183,0.3212198317,-0.0997262821,-0.0877701938,-0.119157128,0.235064283,0.0335857086,0.0239467323,0.0368626118,-0.0175101422,0.146134451,-0.0298812147,-0.151352942,0.1505284458,0.2623203993,-0.2599433959,0.0070397928,0.3197336793,0.2233694643,0.1086493731,0.000646273,-0.1364182085,0.3405466676,0.2584249973,-0.1966732591,-0.2378261983,0.3885471225,-0.0508065745,0.0853362903,-0.0145698758,0.4374082983,-0.2111658007,0.2243067771,0.1498717368,0.2936809659,-0.4318622947,0.0127436062,0.3747748435,0.1390727907,0.0621830262,0.2875871658,0.0014002188,0.0061589186,0.5597172976,-0.0744695961,0.1425116807,0.0042761005,0.2571490407,-0.1812987924,-0.5544173121,0.4394108951,0.1259934604,-0.1168916076,-0.1993806511,-0.2302345484,0.2791203558,0.1457984596,-0.2319589704,-0.1967093647,0.1830404401,-0.3216516078,0.0736054704,0.0673251748,-0.0609896891,0.1429519206,0.1117721573,0.169110015,0.1647931188,0.2556159198,-0.148861438,-0.1340917498,-0.248791486,0.1285080314,-0.0155018112,0.0161446221,-0.0873767361,0.2900003493,0.001436456,-0.1904775053,0.0177702438,0.4018315971,0.2471944392,0.2909375727,-0.0528565124,0.1229542345,-0.2008673102,0.0174718127,-0.4287215173,0.3778696954,-0.2583732903,-0.2916315198,0.2562733591,0.068200849,-0.1824136227,-0.0073565869,-0.1421477348,0.1414724439,-0.5035196543,0.3143913448,0.3248215616,-0.037974298,-0.0030623905,0.1779228002,-0.1650939286,-0.1860365272,0.4572031796,-0.0709720179,0.0209458601,-0.3190409541,0.119236052,-0.2245951295,0.5049152374,0.2660608292,0.0646054,-0.2951208055,-0.1944384426,-0.7159305811,0.2183600813,-0.0852484107,0.2858008742,0.1396151483,0.2323763669,-0.1937198788,0.1679183692,0.0035759029,-0.0542434417,0.2418112308,0.0408345163,-0.2626483142,0.1586986333,-0.0982364714,-0.2893235385,-0.0120389359,-0.3866194189,0.4166209698,-0.1746986508,0.1220311895,0.238194868,-0.1944958866,0.4156767428,0.2763544917,0.3132993579,0.2699008286,0.0833806023,0.1488393694,0.0722179413,-0.1836619675,0.2596637011,-0.2895092964,0.1402838081,-0.2510017157,0.0963058248,-0.4065718353,-0.1825558841,-0.3155361116,0.3408100009,0.0981943458,-0.0447466858,-0.0821548551,-0.1506191939,-0.2725712359,0.3956179619,0.2849607468,0.0411710963,0.0760045871,0.0999494717,-0.4842992127,-0.0423126258,0.3631964326,-0.3208836615,-0.2423743159,-0.2733334303,0.3320650756,-0.0275917649,-0.3213027716,-0.14974913,0.0142918415,0.3661471903,-0.0442024432,0.0851751789,0.2846250534,-0.1038482115,0.4637131095,-0.0950284302,-0.1761485785,0.0414662734,0.1291504353,-0.0996753126,-0.1462930143]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2561","title":"Existing cache for local dataset builder file updates is ignored with `ignore_verifications=True`","comments":"Hi @lhoestq. Thanks for your reply. I just did minor modifications for which it should not regenerate cache (for e.g. Adding a print statement). Overall, regardless of cache miss, there should be an explicit option to allow reuse of existing cache if author knows cache shouldn't be affected.","body":"## Describe the bug\r\nIf i have local file defining a dataset builder class and I load it using `load_dataset` functionality, the existing cache is ignored whenever the file is update even with `ignore_verifications=True`. This slows down debugging and cache generator for very large datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\n- Create a local dataset builder class\r\n- load the local builder class file using `load_dataset` and let the cache build\r\n- update the file's content\r\n- The cache should rebuilt.\r\n\r\n## Expected results\r\n\r\nWith `ignore_verifications=True`, `load_dataset` should pick up existing cache.\r\n\r\n## Actual results\r\n\r\nCreates new cache.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\r\n","comment_length":48,"text":"Existing cache for local dataset builder file updates is ignored with `ignore_verifications=True`\n## Describe the bug\r\nIf i have local file defining a dataset builder class and I load it using `load_dataset` functionality, the existing cache is ignored whenever the file is update even with `ignore_verifications=True`. This slows down debugging and cache generator for very large datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\n- Create a local dataset builder class\r\n- load the local builder class file using `load_dataset` and let the cache build\r\n- update the file's content\r\n- The cache should rebuilt.\r\n\r\n## Expected results\r\n\r\nWith `ignore_verifications=True`, `load_dataset` should pick up existing cache.\r\n\r\n## Actual results\r\n\r\nCreates new cache.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\r\n\nHi @lhoestq. Thanks for your reply. I just did minor modifications for which it should not regenerate cache (for e.g. Adding a print statement). Overall, regardless of cache miss, there should be an explicit option to allow reuse of existing cache if author knows cache shouldn't be affected.","embeddings":[-0.2966097295,0.4845605791,0.0497074835,0.1848862022,0.080392383,0.2186549455,0.2824616134,0.3745069206,0.1949127316,0.1779461205,0.2753896415,0.3544703126,0.0864098966,-0.2514214218,-0.0044403952,0.2905792594,0.0580530651,0.1453519464,0.1856721044,-0.0063897953,-0.0618961118,-0.059082225,-0.0472507104,-0.2655098736,-0.4217759669,0.1886301935,0.0811306983,-0.0028491432,-0.0397531688,-0.4942772985,0.4173816442,0.2595971823,0.0410107411,0.1719833463,-0.000109742,0.1775733232,0.2939718962,0.0204252359,-0.3915853202,0.1280866712,-0.3183541596,-0.3609194756,-0.015063501,-0.1168086305,-0.1470687985,-0.2058143914,-0.0759644806,-0.3138148487,0.3438763916,0.0339676887,0.2579464912,0.0727289319,-0.1777894199,0.0051544788,0.0632820651,-0.0600984395,-0.0318004042,0.1550073326,0.3051303923,0.046208445,-0.3910071254,0.4045751691,-0.2028275877,0.1022235006,0.4206958413,-0.2342661321,0.1239166856,-0.0326370671,0.3514985144,0.0883087069,0.6280643344,-0.3615947962,-0.2906777561,-0.2804444134,-0.1889223009,0.0089398455,0.1707501113,-0.2470255047,-0.1697717011,0.2825199962,-0.4102947116,-0.3994595408,0.0205434579,-0.1617592722,0.2046763301,-0.2736331522,0.0841637328,-0.0900348127,-0.1739295423,0.2221592069,0.4534497559,-0.4566788077,-0.0571203716,0.2458976954,-0.489816159,0.0414768793,0.3280723691,0.3956191838,0.2704196274,0.3395018578,-0.0873380825,0.2306830585,0.1428929865,0.1323721856,-0.0760943443,0.3461059034,0.044928655,-0.0047850083,0.234240368,0.2124478966,-0.4609407783,-0.0363115892,0.0972765535,-0.2108423263,0.5844724774,-0.0553591624,0.0181646664,-0.3203701377,-0.0643922389,0.2362914383,-0.035235174,-0.1552275121,0.1838333309,0.1808869988,-0.0278257858,0.2197253555,-0.1585133523,-0.187197119,-0.2385390699,-0.4174991846,-0.3815182745,-0.2056736797,0.025520239,0.3334385157,0.4631291926,-0.4986813962,0.3289692998,0.2453932315,-0.0809459016,0.0382729545,0.2030310184,-0.067565456,0.1907198876,0.2264155447,-0.3242384791,0.0466676243,0.0138496961,-0.1422075778,-0.0683883429,0.3774857223,-0.1055401862,-0.772066474,0.0577211417,0.1467430592,-0.230830729,0.0597771481,0.2950683534,0.0941423997,0.4354889095,-0.304712832,-0.0714970902,-0.039928697,-0.2986983955,-0.1711079329,0.4547204077,0.7276176214,-0.4251804352,0.0580194816,-0.0966184065,0.0903699026,-0.0047410023,0.1897814274,-0.2760145962,0.3597519994,-0.3368421793,-0.4499462843,0.1937788427,-0.3304957151,-0.7936922908,0.2364359796,0.2361168861,0.2683826983,0.1507766694,0.0912846029,-0.1303571761,-0.1721943468,0.0760541037,0.2031333297,-0.2156373709,-0.067044355,-0.4511630535,-0.2889775932,0.0535785966,0.0131588541,0.0885536969,0.3096219003,0.0585951172,0.1182934642,0.0537008829,-0.0029381225,0.1418864578,0.292199105,0.1424280852,-0.1862851828,0.1863689572,-0.1150948778,-0.4901274145,0.4268352985,-0.2512301505,-0.2352859676,0.0420520902,-0.3022159338,-0.1098713502,-0.1406947076,-0.3829514086,-0.2034128904,0.1235081479,0.2031829953,0.4170442224,-0.0460982583,-0.1830382198,0.3019635975,-0.2747035325,0.1761454195,-0.0656104535,-0.1344092637,-0.091384545,0.0057708807,-0.3230375946,0.0396357104,0.1474550366,-0.0976060852,-0.1709591001,0.2789711356,0.0086257644,0.1368511319,-0.0129127828,0.2477175295,0.0481240042,0.1332277358,0.0803824067,-0.195008412,0.0764137283,-0.0752483681,-0.0221899841,0.055542957,0.1095236242,0.0519843027,-0.2265179306,-0.3205174208,0.2170547247,-0.2078904808,0.0622809231,-0.1489880681,-0.071980387,0.0899278969,0.4432958663,0.1769195795,-0.0288119819,0.151440084,0.3761355877,0.2306368798,-0.2651818395,-0.0584282055,-0.0324974321,-0.2006670386,0.2052030712,0.4338606894,0.2664705515,0.1773048192,0.237342909,0.057374645,0.0024196065,-0.1575553268,0.0431730747,-0.1791941375,-0.1714890748,0.447245121,0.264668107,0.2256841362,-0.3657458127,0.3604898751,0.1917818487,0.1741558909,0.0068332399,0.1442461759,-0.4783364236,-0.1179984584,-0.1782153994,0.239092797,-0.2455631196,-0.2018035203,0.1303868145,0.1122721359,-0.0641460195,0.1965553313,-0.1686368585,0.3677161932,-0.0814805105,-0.5583733916,-0.3459394872,0.0020714966,-0.4155192673,0.0044738641,0.4792733788,-0.3283120394,0.3057935536,-0.2684009969,-0.1086743176,-0.2305951566,-0.2436314076,0.0833076164,0.1805217117,0.311242938,-0.266936183,-0.0569928139,-0.0704203323,0.1075663865,0.1130451784,-0.1787520796,-0.1501226276,-0.1828025877,-0.0643627346,0.1532535404,-0.3532607257,-0.2649860978,0.0208952688,-0.3403119147,-0.0538648665,0.0546617955,0.0128820492,0.2154653817,-0.1387061477,0.1285962611,0.0274688806,0.1715714484,-0.3066546321,-0.6944122314,0.2179216892,-0.1704295427,-0.0264715925,0.0731141716,-0.2889467478,0.1435308009,0.0199487656,-0.5505864024,-0.0304025579,-0.3796710968,0.2839161754,-0.1302035153,-0.0076424284,0.4433671832,0.0021885843,0.0139110675,-0.2255669087,-0.2300461531,0.1580529362,-0.1709071249,0.3343122602,0.0929654166,-0.1266642362,0.258610189,0.5497055054,0.2341137975,-0.0189248305,0.3620266616,0.2053526938,0.5025876164,-0.2435244918,-0.1163855791,0.0630784854,-0.1085820347,-0.2857517004,-0.0677752644,0.0894800872,0.0560047552,-0.1363673508,0.0104678394,-0.0187594648,-0.2934157848,-0.0642722175,-0.440979749,0.2686774731,0.4563596547,0.1301658452,-0.1031800359,-0.105915159,0.1789331585,0.3160368502,0.2749322653,0.0554623492,0.0449782237,0.2654494643,-0.3704788089,0.1303278208,0.0983075425,0.0600287095,0.0849120468,0.0311317686,0.1852438897,-0.0033048915,0.2500499785,-0.5590522289,0.0235622339,0.1375408918,0.0043953317,0.183459878,-0.2239947617,0.2630492747,0.2345827818,-0.1239518821,0.227033183,0.0529048517,-0.1990065873,0.0187146906,0.3892876208,-0.1164693758,-0.4054345489,-0.1874206513,0.0491074771,-0.1712283641,0.0784710497,-0.0590008274,0.1518256664,-0.0021439369,0.0228617582,-0.1453221142,0.3600177765,-0.1597720385,-0.0199764296,0.2620860338,-0.0905334502,0.3267650306,0.0706418231,0.0305499528,0.2073298246,0.5026637316,-0.2005053759,-0.178902626,-0.1821707487,-0.0325581953,-0.0892440528,0.2249555737,-0.0973756984,-0.1954540908,0.2267183065,-0.1740563214,-0.084635973,-0.1170171797,-0.0445941426,-0.0284842085,0.0423105061,-0.4739639759,0.1151125059,0.1791702062,-0.0562358126,0.3488116562,-0.2945680916,-0.0199167598,0.1836634278,0.0458031222,0.5849218965,0.0938150138,-0.1063082367,-0.1129330024,-0.2617773116,0.3956884742,-0.3350907564,0.2682983279,-0.2448899746,-0.2729746401,0.1506207287,-0.168727994,0.2396327406,0.1026629061,-0.0605538599,0.2022613287,-0.0745358765,0.2639939189,0.038241446,0.2278757691,-0.1351741552,-0.0660475492,0.0956498235,0.1695221066,-0.0046303356,0.3136819899,-0.1343436837,0.0260594245,0.274004221,-0.0983227268,-0.2454523742,-0.1350778788,-0.1414004564,0.0108642662,0.1976471394,-0.1424393207,0.0825926363,0.1155972704,0.273648262,0.1074541807,-0.1614090651,0.1261475235,0.1274575293,0.2821273208,-0.1728496104,-0.2204597145,0.3994212449,-0.0048010172,-0.1251102388,-0.2156900913,-0.5188755989,-0.4346002936,0.2353600413,0.1876213253,0.3133313656,0.0822434276,0.0446385704,0.0506644398,0.2856608629,-0.2207333148,0.1887081861,0.1294856966,-0.1102918312,0.1852196306,-0.4093119204,-0.2512069345,-0.05825077,0.2818199992,0.2550117671,-0.2335409969,0.3360430896,-0.2673621178,-0.0726704001,-0.1185964793,-0.0566904508,0.0648118034,-0.214540571,-0.0365528315,0.0350261144,0.0648187473,0.1371730715,-0.0890341252,0.1072421372,0.062624529,-0.302865386,-0.199423641,-0.3694554269,-0.2732175291,-0.116004467,0.1117703766,0.015584589,-0.0442191549,0.2239628434,0.2931734324,-0.3268520236,-0.0356395729,-0.2011874616,0.0778631642,-0.3165007234,0.211591512,0.0781904757,-0.1265774071,-0.0621048957,0.1234081089,-0.3284885287,-0.2365286201,-0.004544707,0.1296794116,0.1019274816,0.267626971,-0.1630535126,-0.0343169756,0.0506565943,0.0170906726,0.1799688041,-0.0315654688,-0.0869813338,0.3448064029,0.0950809345,0.3468888104,-0.0890010074,-0.0734966099,-0.1203303412,0.1932713836,0.0215042736,0.0928594545,0.0224212501,0.0153669957,0.1091334596,-0.0546688735,-0.1389503777,0.1786765605,0.1755829304,-0.2113979906,-0.0071103298,0.3065178096,0.1843283623,0.075141117,-0.0144774541,-0.1549142897,0.3658693135,0.2336486578,-0.2562911808,-0.2456650734,0.381750524,0.0460430756,0.058105696,0.0672457218,0.5152278543,-0.2925961614,0.3185657859,0.1348695904,0.2733685076,-0.480312109,0.1180782542,0.3190508187,0.1784733832,0.0145691382,0.2404754907,0.0797084793,0.0531543456,0.4855402708,-0.0829697102,0.1410018951,-0.0101917582,0.3159423769,-0.1835172623,-0.5806641579,0.4861351252,0.1387703419,-0.1531383842,-0.0575711802,-0.2067167163,0.2538540661,0.179243505,-0.221708402,-0.1657070965,0.1109394804,-0.338753432,0.0691982657,0.1441845298,-0.1334664673,0.1394822001,0.0860454515,0.193305105,0.1997037083,0.3072156012,-0.1525801569,-0.1720905453,-0.1757915616,0.1346706748,0.0232484564,0.0733004585,-0.1188502312,0.2348504812,-0.030356627,-0.227211237,0.0028518739,0.4154131413,0.1996842325,0.1991998255,0.0122236665,0.1301175207,-0.2028800845,0.1017636955,-0.4308285415,0.4320874214,-0.2502795458,-0.3689311147,0.2801718116,0.0266978648,-0.113344878,0.0911168829,-0.1685042083,0.1127073243,-0.5603573322,0.3034974635,0.3499676883,-0.0123252645,0.048785232,0.2230322659,-0.1427257806,-0.1782526672,0.4750252068,-0.1103386953,0.0264910478,-0.2779117525,0.1259624064,-0.1848772317,0.3783934414,0.3177627027,-0.0603679642,-0.2719249725,-0.1675003022,-0.7562376261,0.2122111022,-0.0488382019,0.3328855336,0.200785026,0.1593889743,-0.195991382,0.1406607181,-0.051837258,-0.0379986465,0.2316822708,0.0111095104,-0.2035105079,0.1085803881,-0.068735607,-0.3280591965,0.0358975865,-0.3965867758,0.3446267843,-0.1365159005,0.0843564942,0.2075682431,-0.1935440004,0.3796270192,0.2734094262,0.337860018,0.359084785,0.0153125534,0.1303505003,0.0188465174,-0.1475566775,0.217538178,-0.2528638542,0.1426426768,-0.2502283156,0.0500825867,-0.5288494825,-0.2059171349,-0.3160488307,0.3023276925,0.1304203868,-0.0567561202,-0.1321529448,-0.1284247339,-0.3266018629,0.4613103271,0.2988040745,0.0799179971,0.0740531236,0.1341237575,-0.4689102173,-0.0401123427,0.3417390585,-0.3154911399,-0.1534037739,-0.306592077,0.331587404,-0.0630788654,-0.328173995,-0.2368665487,0.0108551811,0.3737574518,-0.063425146,0.1217779368,0.3035766184,-0.1761278063,0.4700821638,-0.0523521416,-0.07079757,0.1152305081,0.0913198665,-0.0554489978,-0.1846041828]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2561","title":"Existing cache for local dataset builder file updates is ignored with `ignore_verifications=True`","comments":"The cache is based on the hash of the dataset builder's code, so changing the code makes it recompute the cache.\r\n\r\nYou could still rename the cache directory of your previous computation to the new expected cache directory if you want to avoid having to recompute it and if you're sure that it would generate the exact same result.\r\n\r\nThe verifications are data integrity verifications: it checks the checksums of the downloaded files, as well as the size of the generated splits.","body":"## Describe the bug\r\nIf i have local file defining a dataset builder class and I load it using `load_dataset` functionality, the existing cache is ignored whenever the file is update even with `ignore_verifications=True`. This slows down debugging and cache generator for very large datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\n- Create a local dataset builder class\r\n- load the local builder class file using `load_dataset` and let the cache build\r\n- update the file's content\r\n- The cache should rebuilt.\r\n\r\n## Expected results\r\n\r\nWith `ignore_verifications=True`, `load_dataset` should pick up existing cache.\r\n\r\n## Actual results\r\n\r\nCreates new cache.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\r\n","comment_length":82,"text":"Existing cache for local dataset builder file updates is ignored with `ignore_verifications=True`\n## Describe the bug\r\nIf i have local file defining a dataset builder class and I load it using `load_dataset` functionality, the existing cache is ignored whenever the file is update even with `ignore_verifications=True`. This slows down debugging and cache generator for very large datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\n- Create a local dataset builder class\r\n- load the local builder class file using `load_dataset` and let the cache build\r\n- update the file's content\r\n- The cache should rebuilt.\r\n\r\n## Expected results\r\n\r\nWith `ignore_verifications=True`, `load_dataset` should pick up existing cache.\r\n\r\n## Actual results\r\n\r\nCreates new cache.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\r\n\nThe cache is based on the hash of the dataset builder's code, so changing the code makes it recompute the cache.\r\n\r\nYou could still rename the cache directory of your previous computation to the new expected cache directory if you want to avoid having to recompute it and if you're sure that it would generate the exact same result.\r\n\r\nThe verifications are data integrity verifications: it checks the checksums of the downloaded files, as well as the size of the generated splits.","embeddings":[-0.3809878528,0.4386587143,-0.003196154,0.1750701964,0.0950203612,0.1552421749,0.2317519635,0.4199692309,0.1590616554,0.3092049956,0.1621158272,0.2830117941,0.1100006551,-0.2082076222,-0.032141652,0.2218123078,0.0682655275,0.1848935187,0.1504187733,-0.0182196461,-0.1042756811,-0.0340703353,-0.0095534287,-0.2419549674,-0.4294947982,0.0829604268,0.063556686,0.1601958722,-0.0879584178,-0.4699014127,0.3021024168,0.2475569993,-0.1526563764,0.2725529373,-0.0001049965,0.1780284494,0.3562346697,0.0568509437,-0.4711803198,0.2416318953,-0.3674046397,-0.4435736239,-0.1549166441,-0.1537539214,-0.0183799956,-0.1941011548,-0.1145722046,-0.2818766832,0.3508931994,0.0716103017,0.2827044129,0.1405942589,-0.2049485892,-0.0155738434,0.1999707669,-0.0148219187,-0.0281336438,0.1525065601,0.1852069646,0.0982399881,-0.4011557102,0.3540205956,-0.2052344829,0.166930899,0.3523953557,-0.1947120577,0.0473334119,-0.0936723351,0.3847611547,0.1260740608,0.5368218422,-0.2587720454,-0.3534723818,-0.2496431321,-0.2884817719,-0.0149809429,0.2838677168,-0.1221734807,-0.1945977658,0.1572081745,-0.4552613497,-0.2325942665,0.0808344558,-0.2229066193,0.0629747808,-0.2296004891,0.1435914338,-0.0874812603,0.0441175587,0.223488003,0.3453530073,-0.4724609554,-0.0402999818,0.2510861754,-0.4561106861,-0.0766494721,0.2587413788,0.3883616328,0.2604282498,0.3886110485,-0.1932558566,0.2456337363,0.1801903993,0.098462306,-0.1247398332,0.3354095221,-0.0490097664,0.0047542527,0.3091166914,0.2427242994,-0.3884593546,0.0066368589,0.1257502437,-0.2951217592,0.4921707511,0.0260522179,-0.057826031,-0.3234983683,-0.0804766789,0.1806787252,0.0233980939,-0.1273870617,0.166769281,0.1957849562,-0.0826733932,0.1365427077,-0.3365083337,-0.2401606888,-0.341868192,-0.3335377872,-0.3955495358,-0.1528071612,0.0255934075,0.3271663487,0.4925564229,-0.5260605216,0.3480464518,0.1541673839,-0.0543538705,0.099848561,0.2237308472,-0.2249774635,0.2071833611,0.3538979292,-0.2637627125,-0.0269211829,-0.0084566167,-0.0912471712,-0.1013355702,0.2589486539,-0.1230050176,-0.7289390564,0.118471168,0.212846145,-0.2772568762,0.0854331926,0.2786011398,0.0754414275,0.3407641351,-0.3217763901,-0.0556517057,-0.0820501521,-0.278036207,-0.129054755,0.44631055,0.7053080201,-0.4558100998,0.1331191659,-0.1671929061,-0.0553618968,0.0092204576,0.2276038826,-0.2356608957,0.3070384264,-0.3199964762,-0.2536692619,0.1878542304,-0.2249673903,-0.8255181313,0.2686665952,0.149260968,0.1901006699,0.2113611251,0.0449058972,0.0253551658,-0.2250094712,0.0002093963,0.2408092767,-0.2048729062,0.0776339844,-0.4715503156,-0.230891794,0.0494501889,0.0304169375,0.0650147498,0.1332587898,0.1275087893,0.1479833424,0.002820184,-0.0282347649,0.1279258579,0.2897245884,0.0725994706,-0.1606552303,0.2341433614,-0.1299551874,-0.4772045016,0.4268706739,-0.2430205494,-0.161271885,-0.0033574956,-0.367875576,-0.1315849721,-0.0959358439,-0.3227362037,-0.1761829257,0.1626995504,0.2217373401,0.4972623885,-0.0863663256,-0.092161499,0.2364337891,-0.2325900346,0.1563906223,-0.0437477678,-0.0619531497,-0.0965528041,-0.0037325129,-0.2672517896,0.0637132898,0.0612828881,-0.1438021511,-0.2119125128,0.3813095391,0.182417959,0.1483064294,0.1038815156,0.207360059,0.0454355367,0.0915351883,0.2094063312,-0.1386991888,0.0882934034,-0.0692195669,-0.1016848087,0.0941345692,-0.0026223233,0.1150359735,-0.097411871,-0.3273003995,0.2436901778,-0.0979226083,0.0482175723,-0.1521125287,-0.0224174112,0.1089246124,0.4705246091,0.2494598329,0.0959344208,0.1308138222,0.5666050315,0.0951122195,-0.3064001799,0.008478391,-0.0920882374,-0.2652546465,0.0858353749,0.4567349851,0.3705622852,0.1872846335,0.2971365452,0.0433766544,0.0250684544,-0.1981807351,0.0852613896,-0.0498640835,-0.2039026469,0.3712655306,0.2285541445,0.1648545414,-0.4026487768,0.2453194857,0.1885005534,0.1817783713,0.0076121073,0.210149616,-0.4117991328,-0.1393776834,-0.1930423081,0.155893892,-0.3065185547,-0.2943997979,0.1703382581,-0.0018876696,-0.1134918034,0.2163655162,-0.0438544005,0.3727186322,-0.0144591276,-0.5106043816,-0.3518974185,-0.0204889514,-0.421954304,0.0638473481,0.4755040109,-0.2497112304,0.364374578,-0.3301583529,-0.1467280984,-0.1775905192,-0.2530392408,0.0571762808,0.0023792256,0.2441037297,-0.1194614619,0.0094949184,0.090172559,0.073113054,0.1342111528,-0.1966596693,-0.2162160128,-0.1232861504,0.0012073183,0.1617399752,-0.405503273,-0.2719637454,0.0034306503,-0.3567607701,0.0014948995,0.1071014106,0.0538732335,0.1546714753,-0.1395678371,0.2075217217,0.0358731449,0.0946294367,-0.337197274,-0.5732303858,0.1985998154,-0.1797296554,-0.0840433538,0.0857976228,-0.2623572052,0.0923823118,0.0246700477,-0.5293719172,0.0089809969,-0.5315983295,0.2878866196,-0.1223090738,0.0137544731,0.4414839447,0.0198047347,-0.1030043513,-0.2068895102,-0.2004306614,0.0841571763,-0.2701410949,0.2128931284,-0.0614318289,-0.090494439,0.1922164559,0.6441383958,0.1705699861,-0.0261244494,0.2631525397,0.2355026901,0.5081669688,-0.3046405017,-0.1536633223,0.1751738042,-0.1124335676,-0.2807420194,0.0942912027,0.0032025189,-0.0760571286,-0.1071785092,0.0054230741,0.0717361271,-0.2113446444,0.0174798295,-0.3770815134,0.174826622,0.3805962503,0.216280207,-0.1443027854,-0.1160825044,0.1850569248,0.2957296371,0.174750194,0.0181116108,0.013210549,0.2020646483,-0.3103516698,0.1260381043,0.1875198036,-0.0421074182,0.0342034064,-0.0716048777,0.2343113422,-0.0502523072,0.310957849,-0.4699048102,0.039634943,0.1316846162,0.0003003621,0.1948234588,-0.2603046298,0.096525535,0.2191571742,-0.0178051386,0.3281489611,-0.0614307411,-0.1760919094,0.0945458412,0.3984564543,-0.0820436105,-0.3254150152,-0.3235057592,0.0369306318,-0.1915463656,0.142537266,-0.109192498,0.1812101603,-0.1364021897,-0.0018457026,-0.0889678001,0.2300605774,-0.1454863846,0.0393684134,0.2379764915,-0.1225233302,0.22648485,0.06669157,0.0999516621,0.228730008,0.4857907593,-0.196926102,-0.1524712145,-0.1413664371,-0.059384387,-0.1390206516,0.2855303288,-0.0680386275,-0.2639187276,0.2175980955,-0.004712794,-0.1272703409,-0.0137210786,0.0422125384,-0.0254559759,0.034945678,-0.444201678,0.1618110985,0.1868470311,-0.033510156,0.2781071067,-0.1521117687,0.0795830563,0.1330114305,0.1249538362,0.6822166443,0.1362113506,-0.1145956814,0.0002069581,-0.2710430026,0.3784745038,-0.3629117608,0.2440561652,-0.1973891705,-0.3495337963,0.1336464286,-0.1639500409,0.2852556109,0.0307868179,-0.143657282,0.1867617369,-0.0287960973,0.1945047677,0.0096166022,0.3009592891,-0.1734447032,-0.1199178472,0.1038741022,0.2350699753,0.0205893517,0.2723823786,-0.0777696669,0.0324533693,0.1645316929,-0.0691591799,-0.2880913019,-0.1545039564,-0.0871582404,0.0442418829,0.1170277297,-0.1048893705,-0.0277322251,0.0329482928,0.2732405663,0.1882366985,-0.169549942,0.190285176,0.1060741991,0.3167861104,-0.0597576834,-0.2441647798,0.407333225,-0.080827795,-0.1148024574,-0.2614194453,-0.5250527263,-0.3327232897,0.3132185936,0.1088080332,0.2869955003,0.0436147973,0.1405775398,-0.0438864157,0.1577403098,-0.2174314857,0.2091279775,0.042746298,-0.1072449088,0.1430872232,-0.3217616975,-0.3278847039,-0.1194734201,0.2778670192,0.1766344011,-0.1231905371,0.3894484639,-0.293332994,-0.179352656,-0.1765353382,-0.0402748324,-0.0411626063,-0.269742012,0.0075382008,-0.0401744582,0.0050673829,0.1045181677,-0.0712547824,0.1229317486,0.1439367533,-0.3369544148,-0.2467673868,-0.3537612557,-0.2653901279,-0.1480856687,0.1309174597,-0.0345484838,-0.0122091034,0.2249914259,0.382973969,-0.3870071173,-0.0468675643,-0.2649410367,0.0812324509,-0.3280068636,0.0506658889,0.0909307152,-0.0747460648,0.0090028094,0.1508876681,-0.3463103771,-0.2939866781,0.0540275797,0.1059437543,-0.046906244,0.1963615268,-0.1199758053,0.1190235689,-0.0014981033,-0.049134966,0.2123953551,-0.2160633057,-0.1563726366,0.2548014522,0.1707797647,0.353397876,-0.0799337327,-0.0274657737,-0.183037594,0.1936137974,-0.0437574051,0.132066384,-0.0450856909,-0.0873599201,0.1132305041,-0.031498488,-0.0336392336,0.1384854317,0.2454380542,-0.1636517197,-0.0036880577,0.2253755331,0.2210756689,0.1636663228,0.0717784092,-0.2607776523,0.3172480464,0.2696538568,-0.1820483953,-0.2011013478,0.3371920884,-0.0247246865,0.1318006366,0.0341989323,0.4432098567,-0.1671998054,0.2255140245,0.1544443965,0.3434157968,-0.3536663353,0.0417474844,0.4806295037,0.1429196894,0.0976358503,0.2186352462,-0.0359709859,0.0562150888,0.4986139536,-0.1254051328,0.3106615543,-0.06867861,0.2836295068,-0.2062566429,-0.556445837,0.4018728733,0.1693214178,-0.1316091716,-0.1121614799,-0.1634998173,0.2880296111,0.029241683,-0.2176132649,-0.2586454153,0.15242818,-0.2848618329,0.0762705505,0.0821425989,-0.0851984248,0.055418171,0.1326171011,0.1994865984,0.2265625,0.3473984897,-0.0989744216,-0.1429739743,-0.0877808779,0.1233608946,-0.0985961705,0.0712055117,-0.1479970813,0.1824072152,0.0958579332,-0.1535413712,0.0640589446,0.3616479039,0.2506943643,0.2123777717,-0.0989350379,0.0259598736,-0.1711537242,0.018798301,-0.3725170493,0.4313400686,-0.2985082865,-0.2817282081,0.3423575163,0.0857559443,-0.1949165016,0.0019100879,-0.2204264849,0.0274499729,-0.4845627248,0.3101215959,0.3515853286,-0.030931849,-0.0904952437,0.2066890597,-0.2065312862,-0.2010108382,0.2995293736,-0.1196033359,0.071751371,-0.250905633,0.1303449273,-0.2079915106,0.6261104941,0.351813525,0.0281966031,-0.2392153889,-0.1770207733,-0.6708555222,0.2010459006,-0.0954009071,0.1895970851,0.0860061795,0.2077537328,-0.1447227597,0.1590956748,-0.0037625514,0.0648978502,0.2173777521,-0.0050047245,-0.2826894224,0.1956052035,-0.1473674774,-0.3527964056,0.0055219987,-0.4427983761,0.4373824894,-0.151498884,0.1282616109,0.3331017196,-0.1716022044,0.4262593687,0.2726679146,0.252473712,0.3181019425,0.1200245842,0.0899264142,0.0215407368,-0.1689340919,0.3327594399,-0.2900877595,0.1465853453,-0.2002493441,0.0694028512,-0.3555560708,-0.097299464,-0.2597144842,0.4706433415,0.1652919948,0.0576935746,-0.0976826102,-0.2062508017,-0.2786985934,0.4057398736,0.3590345383,0.0832565203,0.1341121644,0.0653557554,-0.4515414834,-0.1034893095,0.3416598737,-0.2379947901,-0.1830805838,-0.3390711546,0.3901357651,0.0657899454,-0.2984901667,-0.1949422807,0.0399752185,0.3311713636,-0.0626843795,0.0069870073,0.311277926,-0.1424303055,0.3858706653,-0.0999677628,-0.1195295677,0.0070809158,0.0974014327,-0.0177118443,-0.1525213569]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2561","title":"Existing cache for local dataset builder file updates is ignored with `ignore_verifications=True`","comments":"Hi @apsdehal,\r\n\r\nIf you decide to follow @lhoestq's suggestion to rename the cache directory of your previous computation to the new expected cache directory, you can do the following to get the name of the new expected cache directory once #2500 is merged:\r\n```python\r\nfrom datasets import load_dataset_builder\r\ndataset_builder = load_dataset_builder(\"path\/to\/your\/dataset\")\r\nprint(dataset_builder.cache_dir)\r\n```\r\n\r\nThis way, you don't have to recompute the hash of the dataset script yourself each time you modify the script.","body":"## Describe the bug\r\nIf i have local file defining a dataset builder class and I load it using `load_dataset` functionality, the existing cache is ignored whenever the file is update even with `ignore_verifications=True`. This slows down debugging and cache generator for very large datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\n- Create a local dataset builder class\r\n- load the local builder class file using `load_dataset` and let the cache build\r\n- update the file's content\r\n- The cache should rebuilt.\r\n\r\n## Expected results\r\n\r\nWith `ignore_verifications=True`, `load_dataset` should pick up existing cache.\r\n\r\n## Actual results\r\n\r\nCreates new cache.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\r\n","comment_length":73,"text":"Existing cache for local dataset builder file updates is ignored with `ignore_verifications=True`\n## Describe the bug\r\nIf i have local file defining a dataset builder class and I load it using `load_dataset` functionality, the existing cache is ignored whenever the file is update even with `ignore_verifications=True`. This slows down debugging and cache generator for very large datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\n- Create a local dataset builder class\r\n- load the local builder class file using `load_dataset` and let the cache build\r\n- update the file's content\r\n- The cache should rebuilt.\r\n\r\n## Expected results\r\n\r\nWith `ignore_verifications=True`, `load_dataset` should pick up existing cache.\r\n\r\n## Actual results\r\n\r\nCreates new cache.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\r\n\nHi @apsdehal,\r\n\r\nIf you decide to follow @lhoestq's suggestion to rename the cache directory of your previous computation to the new expected cache directory, you can do the following to get the name of the new expected cache directory once #2500 is merged:\r\n```python\r\nfrom datasets import load_dataset_builder\r\ndataset_builder = load_dataset_builder(\"path\/to\/your\/dataset\")\r\nprint(dataset_builder.cache_dir)\r\n```\r\n\r\nThis way, you don't have to recompute the hash of the dataset script yourself each time you modify the script.","embeddings":[-0.4378628135,0.4782138169,-0.0076685501,0.1180528328,0.1569270641,0.1881317943,0.2960320413,0.4827200174,0.1750372797,0.3437815905,0.2252853215,0.3112469912,0.0643949583,-0.1951674372,0.0266977418,0.322676897,0.045707155,0.166235581,0.0950702727,0.005609889,-0.1577256471,-0.0480541214,-0.0432750061,-0.2891950309,-0.3779666424,0.1461428553,0.1288987696,0.1029422581,0.0418603197,-0.5005040765,0.3477988839,0.2545360923,-0.0790350735,0.3221059442,-0.0001039051,0.1450937688,0.3499033153,0.0292263068,-0.4508456588,0.1350848079,-0.3023414612,-0.3922794461,-0.0445121154,-0.1520197988,-0.1497523338,-0.1942988634,-0.0928564519,-0.3335602283,0.2368957698,0.0695878118,0.3009346724,0.1146359071,-0.174826473,-0.0558505058,0.1333578676,-0.0726914853,-0.0339046456,0.1525844485,0.171585083,-0.0266665537,-0.3792466223,0.4052736163,-0.2034496516,0.1334887892,0.399666816,-0.1477893144,0.0812321007,-0.0307958946,0.3842017055,0.0088948812,0.5154066086,-0.3620007336,-0.3648876846,-0.2112908363,-0.2311091572,-0.1044781134,0.2209356427,-0.1983426809,-0.2347516865,0.1736158878,-0.3927207589,-0.2276475728,0.0417950638,-0.180952549,0.0880759656,-0.2250020504,0.1053643301,-0.1372450739,-0.125476405,0.1452204734,0.3674782217,-0.394854635,-0.0215280261,0.228969425,-0.4612696171,-0.0239970554,0.3944355845,0.3140067756,0.2373757213,0.3474124372,-0.1095322147,0.2829091549,0.1750382781,0.0915098861,-0.1357051134,0.3506705761,-0.0340687074,-0.052787438,0.3768588603,0.2751902342,-0.4690761566,0.0056921476,0.106025368,-0.3259635866,0.5800098777,0.0720466673,0.0524250269,-0.3129971921,-0.0475726798,0.2453954369,0.0636419654,-0.0961804464,0.2122593969,0.2610956132,-0.1054353863,0.2311409265,-0.1733355671,-0.215301618,-0.3223363459,-0.2802629173,-0.4096826911,-0.1537312418,0.0119259618,0.3056171834,0.4829461873,-0.4689620733,0.316123575,0.1618570089,0.0118938787,0.1007125154,0.1903785169,-0.1322101802,0.2143111974,0.2632401288,-0.2132689357,0.0102520604,0.0313112512,-0.1175557598,-0.1058414653,0.223121658,-0.0317617171,-0.7341607809,0.0889080092,0.233043924,-0.2381561249,0.0474885292,0.2636314332,0.0336059257,0.3829751313,-0.2503570616,-0.1222589388,-0.0942947567,-0.3573611975,-0.175920859,0.4980244339,0.7052147985,-0.3996891379,0.1528285742,-0.115663141,0.0775155723,-0.066553548,0.1591843665,-0.2795010209,0.3339782953,-0.2976713479,-0.4289454818,0.1798465103,-0.3847131431,-0.788780272,0.2251588106,0.1457596868,0.149137944,0.1818509996,0.0396104641,-0.0169460047,-0.2184168547,0.1615031213,0.2312603593,-0.1634507626,0.0983477458,-0.4349821806,-0.2086125165,0.0089624533,0.0688338131,0.0870128646,0.169584766,0.056815207,0.0829610974,0.026787024,-0.0496538766,0.1402308345,0.3736862838,0.1132957041,-0.1159346849,0.220341593,-0.1655070335,-0.4990051389,0.4116998613,-0.1820967197,-0.1447318047,-0.0531154536,-0.322712332,-0.1338289082,-0.087515153,-0.3375990689,-0.1391991526,0.1884564161,0.2496445775,0.4460933805,-0.0771123692,-0.158167094,0.2838539183,-0.1640733927,0.0899734572,-0.1326985955,-0.1409479529,-0.0700131431,-0.0991641507,-0.1954044551,0.0574336536,0.1353019476,-0.0667103156,-0.26080966,0.3024924695,0.0793655366,0.1147842035,0.0711160004,0.2790294588,-0.0350753628,0.1858082116,0.1745343953,-0.0790572241,0.1426222324,-0.0384484343,-0.0413477905,0.1354292035,0.0463986173,0.1125914901,-0.1388845444,-0.3127244115,0.295370847,-0.1245715842,0.0633455738,-0.1481835991,-0.0057835588,0.0748842359,0.4566465914,0.1771675199,0.0018894294,0.2123999894,0.4804141819,0.1649549454,-0.2486862838,-0.0117339864,-0.09440431,-0.2098719925,0.0730080605,0.397223413,0.3072849214,0.2216350138,0.2192882299,0.0572420284,-0.0068486789,-0.1589208096,0.0911510214,-0.0753751621,-0.1550502926,0.3534370065,0.2896508873,0.1620254964,-0.3402825594,0.1717146039,0.1014906913,0.177421093,-0.0199414697,0.1480083019,-0.437425226,-0.1086441725,-0.174734205,0.205991596,-0.1812389791,-0.2643585503,0.0996612683,0.1013825685,-0.1185423583,0.216680184,-0.1225303784,0.3018139303,0.0229896735,-0.4956262708,-0.2863645554,-0.0362147503,-0.3606921732,0.031658262,0.4571745396,-0.2570876777,0.3414334357,-0.2753068507,-0.1161041185,-0.2035597265,-0.2270614058,0.0009948392,0.0311385747,0.3004502058,-0.1247740686,0.0048386138,0.0545157269,0.0616223998,0.1966441125,-0.2101594061,-0.161397323,-0.0959529504,-0.000732612,0.098420009,-0.3548962176,-0.3140829802,0.016634535,-0.4324187934,-0.0886578187,0.1247779801,0.0167996623,0.1820453554,-0.1176902428,0.1404753327,0.0436921827,0.1650148481,-0.2781701982,-0.6232032776,0.270997107,-0.2506586909,-0.0944469348,0.0615543984,-0.3288154304,0.1137563214,-0.0210148767,-0.4772715867,-0.0800592601,-0.5147484541,0.2753629386,-0.1654993743,0.031567663,0.4313963056,0.0792466849,-0.1137080565,-0.2364391834,-0.2883403301,0.0790760666,-0.2812061012,0.1944599748,0.0056791999,0.0140724061,0.18723315,0.5367400646,0.1950557381,-0.0902496651,0.2941389084,0.1639422178,0.4542673528,-0.3149476051,-0.2042661756,0.0530165359,-0.0757592767,-0.2786774337,0.0489867106,0.0071325153,0.0000148406,-0.1413308531,-0.041352056,-0.0463015735,-0.3057073057,0.0590570681,-0.4534632564,0.1954779029,0.3803459704,0.1741364896,-0.232110858,-0.1006532758,0.1646799147,0.2463067025,0.1682643741,-0.0457250141,0.0351355225,0.2795372009,-0.4036783874,0.162037313,0.1288981438,-0.0083644632,0.0930651724,-0.0572279803,0.1177354828,-0.0911324546,0.2825429142,-0.5072576404,0.0911323354,0.2038347572,-0.1125687286,0.062775813,-0.1989260763,0.1504738182,0.2622158527,-0.0879827738,0.2507500052,-0.0011108541,-0.2351133525,0.1253533661,0.3870460391,-0.1197637916,-0.33446908,-0.2623523176,-0.0277381651,-0.1591908187,0.0565851517,-0.0812543854,0.2400640696,-0.0308113471,-0.0278974604,-0.1352307796,0.2370305955,-0.1084025875,0.0180764943,0.2628670633,-0.084426105,0.3803780973,0.0252024028,0.0302188098,0.2365126908,0.4647636712,-0.125058949,-0.2326643169,-0.1947994232,0.0530399568,-0.0449093133,0.2518130243,-0.1161489189,-0.2082093954,0.1460901201,-0.0102907792,-0.1474388838,-0.0304774381,-0.0205701292,-0.0034657426,0.0006031991,-0.4913100898,0.204158172,0.1895584613,-0.0449840464,0.3198823929,-0.2744066119,-0.0281308349,0.2554374337,0.1006878242,0.6605774164,0.0964937061,-0.03535055,-0.0036210583,-0.2316719592,0.3259824216,-0.3080957234,0.2457036823,-0.2795040607,-0.3074895144,0.1735803038,-0.1925346404,0.1891011,0.0348283648,-0.124057509,0.2253088355,-0.1151042283,0.1912562549,-0.0458404534,0.3150481582,-0.1868624389,-0.1745786667,-0.0274510235,0.2539723217,0.0469782427,0.24615632,-0.0815269873,0.0187766925,0.1505829394,-0.0808518678,-0.2379864156,-0.0665984973,-0.107482031,0.060202606,0.1213902608,-0.1067700237,-0.0436116904,0.057633888,0.3026908338,0.2505254745,-0.1588268578,0.1485432386,0.0560751893,0.3196144104,-0.065843448,-0.2128132582,0.4055592418,-0.0617627129,-0.1589684486,-0.2594929338,-0.4273688793,-0.3726550341,0.2680479586,0.1460202634,0.2779169679,0.0533105284,0.077103816,-0.0018372759,0.1734862626,-0.2306042016,0.2520277798,0.1348027885,-0.1515694559,0.1891325861,-0.3363292515,-0.2991268039,-0.0673695058,0.3114566207,0.1595468223,-0.1517668366,0.4051167071,-0.2084005028,-0.1419317126,-0.185300678,-0.0475523733,0.0767187178,-0.2326750606,0.0687842891,0.0404996388,0.0431034528,0.1582069546,-0.0608673505,0.0395646878,0.0245332643,-0.3133846521,-0.2225445658,-0.428566128,-0.2193321288,-0.0632849261,0.1223854497,-0.0012289485,-0.0267296191,0.1503131837,0.2760211229,-0.4142199457,-0.0566009283,-0.2791680992,0.1024125516,-0.2881475091,0.0984876081,0.1146466061,-0.0766550824,0.0632297769,0.1305019557,-0.3284518719,-0.3228774071,0.0242012627,0.0945781991,-0.0166070927,0.2232036442,-0.1147042066,0.0794668421,0.0184371397,-0.0887312219,0.2097735405,-0.0178425964,-0.1104142889,0.267986685,0.1221219972,0.3223766088,-0.0099017527,-0.0430623591,-0.1550920308,0.2193196267,-0.0433367863,0.1086943671,-0.0544866882,-0.0705574676,0.132929638,-0.0570239946,-0.1508288831,0.192397356,0.2223356664,-0.1864930242,-0.0602302514,0.3218355477,0.2400194556,0.0907398909,-0.0054480275,-0.1957660019,0.3003565967,0.2832587659,-0.2194077522,-0.1706596315,0.2801479697,-0.0132269356,0.1595283896,0.0650127083,0.3555441201,-0.1990950257,0.2337106317,0.1321997046,0.3232167661,-0.4168417752,0.0862707123,0.2853973508,0.1657533795,0.0082563991,0.2053154707,0.0671943873,0.0971907303,0.6143580079,-0.02835197,0.2189553678,-0.0285652671,0.2101496905,-0.2261810601,-0.5105987191,0.4365909398,0.1550196856,-0.1180957183,-0.1057624817,-0.1610130668,0.2295512259,0.107305266,-0.1292933077,-0.194942534,0.1497918367,-0.2843475044,0.0343772471,0.1221878976,-0.024144033,0.1386813819,0.1023364365,0.2187937945,0.1756784022,0.2815102637,-0.0567291044,-0.1032158285,-0.0950154215,0.1916714311,-0.0450918674,0.1194567531,-0.1141229868,0.1823381931,0.1099841371,-0.1482106298,-0.0076163127,0.3853509128,0.3195372522,0.2789935172,-0.0787198171,0.1248615757,-0.1749562025,-0.0083576823,-0.4222159386,0.3339467645,-0.3401620686,-0.2274769098,0.324901402,0.1272558421,-0.2051461637,0.0182348397,-0.1452877223,0.0690739676,-0.4051777124,0.3360981047,0.3596473336,-0.015595274,-0.101364933,0.2376695871,-0.2326172143,-0.2524329424,0.4349987507,-0.0604382083,0.0878379717,-0.3028021455,0.1456753165,-0.1554861516,0.484449923,0.2788093388,-0.0002208715,-0.2452317178,-0.1420565695,-0.6544242501,0.2114124149,-0.0518941283,0.2111930549,0.0496996492,0.2326919436,-0.1671123952,0.1167374849,0.0590450615,-0.0167423338,0.1796305627,0.0328867212,-0.2991632521,0.1567018181,-0.159565717,-0.3394167721,0.089454025,-0.4561432302,0.3197131157,-0.2177821845,0.1850194931,0.2190605253,-0.1568605602,0.3673104048,0.2313755304,0.342959255,0.2560586929,0.1062690914,0.0294540841,-0.0163916778,-0.2015939057,0.2178560942,-0.3097857535,0.1981265992,-0.220191732,0.0975879952,-0.4092788696,-0.1760576814,-0.2828541398,0.3462895453,0.0879388005,-0.0051467461,-0.0652862266,-0.0974264219,-0.2284780741,0.405036658,0.2715637088,0.0712458044,0.0581016764,0.1829338074,-0.4816671014,-0.1742366999,0.4195163846,-0.221542716,-0.2811935544,-0.2682036161,0.3721109331,-0.0936366767,-0.2679523528,-0.2221828401,0.0980302244,0.3129083514,-0.0035396584,0.0003510162,0.3377867937,-0.15323627,0.3676834702,-0.0670661405,-0.1377289891,0.0599398427,0.1395842284,-0.1061303988,-0.155070886]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2559","title":"Memory usage consistently increases when processing a dataset with `.map`","comments":"Hi ! Can you share the function you pass to `map` ?\r\nI know you mentioned it would be hard to share some code but this would really help to understand what happened","body":"## Describe the bug\r\n\r\nI have a HF dataset with image paths stored in it and I am trying to load those image paths using `.map` with `num_proc=80`. I am noticing that the memory usage consistently keeps on increasing with time. I tried using `DEFAULT_WRITER_BATCH_SIZE=10` in the builder to decrease arrow writer's batch size but that doesn't seem to help.\r\n\r\n## Steps to reproduce the bug\r\n\r\nProviding code as it is would be hard. I can provide a MVP if that helps.\r\n\r\n## Expected results\r\n\r\nMemory usage should become consistent after some time following the launch of processing.\r\n\r\n## Actual results\r\n\r\nMemory usage keeps on increasing.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0","comment_length":33,"text":"Memory usage consistently increases when processing a dataset with `.map`\n## Describe the bug\r\n\r\nI have a HF dataset with image paths stored in it and I am trying to load those image paths using `.map` with `num_proc=80`. I am noticing that the memory usage consistently keeps on increasing with time. I tried using `DEFAULT_WRITER_BATCH_SIZE=10` in the builder to decrease arrow writer's batch size but that doesn't seem to help.\r\n\r\n## Steps to reproduce the bug\r\n\r\nProviding code as it is would be hard. I can provide a MVP if that helps.\r\n\r\n## Expected results\r\n\r\nMemory usage should become consistent after some time following the launch of processing.\r\n\r\n## Actual results\r\n\r\nMemory usage keeps on increasing.\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.0-52-generic-x86_64-with-debian-bullseye-sid\r\n- Python version: 3.7.7\r\n- PyArrow version: 3.0.0\nHi ! Can you share the function you pass to `map` ?\r\nI know you mentioned it would be hard to share some code but this would really help to understand what happened","embeddings":[-0.1570354253,-0.1291135401,0.0070893918,0.4211396873,0.1809348166,-0.0064084614,0.0589690395,0.1185609922,0.2151501626,0.1314368397,0.4036096632,0.5179601908,-0.1873755753,-0.0342993326,-0.0775477067,0.073951751,0.1007091478,0.0621853173,-0.002483153,0.0814324021,-0.2891268432,0.1485365182,-0.2638915777,-0.0169580318,-0.3848563731,-0.2375293523,0.0283594988,0.1595099121,0.2712178528,-0.4738775492,-0.1240217835,-0.0056957845,0.1005657837,0.5225373507,-0.0001255529,-0.0305608511,0.401930809,-0.0119665451,-0.0148503724,0.1246175542,-0.3412507474,-0.1495825499,0.0193182342,-0.0983853266,0.2345033735,-0.0282174107,-0.0144428136,-0.4091306925,0.2877306044,0.0745106563,0.1072863117,-0.0405011214,0.0402137116,0.2015887648,0.3487725258,0.3877718747,0.0770824477,0.1127612144,0.0478502326,-0.2627837956,-0.4136327207,0.2883907557,-0.0236985404,0.1897912771,0.5099977255,-0.0134917554,0.3311572373,-0.3816513121,0.4529604316,-0.1067861766,0.1417180002,-0.2396407723,0.230925858,-0.2007993907,-0.3240593374,-0.0723847225,0.257769227,0.1840465218,-0.3346029818,-0.0982146934,-0.2691874802,-0.2370991707,-0.1393745989,0.1706352383,0.1292566508,-0.3141852617,-0.0201004203,0.2519589067,0.3842785954,-0.1438483298,0.045506876,-0.3902648091,-0.0603136346,0.2051589787,-0.2177036256,0.3616625965,0.1598579139,-0.1390434057,0.186015591,-0.4799719453,0.0471167825,0.0493901037,-0.0707272589,0.3078666925,0.5049361587,0.0071358401,0.0339402556,0.3567737043,0.1445606202,0.0739017352,-0.4335692227,-0.09687379,-0.0254753511,-0.1504217684,0.5621683598,-0.4341062903,-0.0089524696,-0.0162021536,0.0838859826,0.1440406144,-0.2474454194,-0.1192666739,-0.1069318354,0.2773066759,0.1608290523,0.212488994,0.0798254684,-0.1200265735,-0.2981453836,0.1034978926,-0.0617954768,-0.2247256786,-0.4726648033,0.2819334269,-0.1137935072,-0.023724284,0.0799849033,0.0839325264,-0.1809536368,-0.4311408103,0.2190316468,-0.3097909391,0.3081334233,0.1337933093,-0.2712439299,0.1661079973,-0.2565067708,0.2414965183,0.0015335188,0.6213239431,-0.4938792288,-0.2463744134,0.0666981637,-0.0554476641,0.1610040069,0.2372991443,-0.289039582,0.0802726299,0.413513422,-0.1228782684,-0.0541271642,-0.1508701444,-0.3637880087,-0.3568014205,0.0744894594,0.5120196939,-0.2421150357,-0.0570950247,-0.0698018894,0.1361998171,0.3940269351,0.3394079208,-0.1452197582,0.2179701626,-0.2710808516,-0.0092691397,0.0535811782,-0.135399431,-0.4603150189,0.1626189798,0.0354248844,0.0441171899,-0.1676273495,0.0917235762,0.1136214286,-0.020611966,0.0810675174,0.2399548143,0.0672847927,0.2178075016,-0.3566509783,-0.3306137621,-0.1602676958,-0.0686361566,0.0161735695,-0.1525756121,0.0998775214,-0.3316319585,0.4156735837,-0.1726185232,0.1996031255,0.3873454928,0.32808882,0.0715711117,0.0820139423,-0.0371198021,0.0218242742,0.0685922205,-0.5524809957,-0.4697282314,-0.089287214,0.05952904,0.1769143194,0.2053637803,-0.095445618,-0.0883561969,-0.0736562163,-0.1179291904,-0.3036301136,-0.0247292574,0.1444119513,0.1526732743,0.0656351298,-0.0442605987,0.027166184,0.2399702519,0.0752367079,-0.2047433108,-0.201455757,0.0656852722,0.0661339164,0.0654207543,-0.261890471,0.1513354033,0.2842341065,0.106396459,-0.0741834193,0.0739029199,0.3849378228,-0.1497890353,0.2134182006,-0.0641799569,0.105747506,-0.0667208433,0.121331118,-0.2330114394,0.3269650042,0.3512523472,-0.0582070462,-0.1533294916,-0.0210708529,-0.0199612882,0.4386912584,-0.1539627463,-0.1473774612,0.087358512,0.2650532126,0.288436383,0.057037428,0.0676861703,0.483543694,0.1863622665,-0.0907853097,0.4215968847,-0.4808891118,-0.1382156312,0.2403400391,-0.0141240573,0.1993006617,0.0344419442,-0.1749187708,0.0550536364,0.0971427187,-0.0127408607,0.1064732373,0.2050014138,0.5473662615,0.1184342504,0.2693426311,-0.0106128324,-0.1087218076,-0.0634394959,0.0204938166,0.0222310498,-0.2989660501,0.2088721395,-0.2458423674,0.1614700109,-0.1152484119,0.1659799516,-0.2664535046,-0.0865114555,-0.1947995275,0.2522456348,-0.0238519162,0.0529238395,0.0373384468,0.1334961206,0.2425183505,-0.2201697379,-0.1335164756,-0.2635029554,-0.3734341562,-0.1866772175,0.3813974261,-0.18527399,0.3777766526,0.1496836841,0.1176155582,-0.2089557648,-0.1244721934,-0.0028930365,0.3083295524,0.2173012644,-0.0969915763,0.1768705547,-0.1049770266,0.0970579609,0.186447382,-0.4787376225,0.0666563064,0.1396715343,-0.119236283,-0.2064532787,-0.0172842238,0.0637788624,0.0498527773,-0.2209955454,0.2391790897,0.0881538615,0.3557169437,0.2330054939,0.4233998358,0.225534454,-0.2216945291,-0.2498881966,-0.1388605386,-0.0541365966,0.0314728022,-0.0852679312,-0.0235585552,0.0020438365,-0.0498107262,0.1298592687,0.3587283492,-0.6982625127,0.0301660262,-0.2623217106,0.2872711122,-0.1540242583,0.2327871621,0.1900490522,0.2522499859,0.0068761553,-0.1199147254,-0.3412647545,0.226945892,0.1021318734,0.2622027993,0.4573037624,0.2340842485,0.1723798066,0.6430386305,0.36811167,-0.1270593405,0.5097683072,-0.1093711257,0.1468427032,-0.1853081882,-0.0802948028,-0.1467505097,-0.3551236391,-0.1013460681,0.0564733148,0.1598134935,-0.3769489825,0.300753355,-0.2385727465,-0.0146850161,-0.3792770207,0.3319390416,-0.088636905,0.4131893814,0.0318720341,-0.0961550027,-0.4214274287,-0.4405423403,0.1351609379,-0.400267303,0.1780541241,-0.039699275,-0.0401163436,-0.0138434954,-0.5272111297,0.3067462444,0.0390462428,0.2621274889,-0.0297251288,0.0491983667,0.2986944914,-0.0760305226,0.5504159331,-0.0347133651,0.0523166209,-0.089277938,0.0212841593,-0.4756469727,0.0552174337,-0.0031104495,0.3808129132,0.0940501913,0.3508795202,0.0697250515,-0.1875783652,0.0132522471,0.0421578586,-0.2354724109,-0.4961806536,-0.1325454116,-0.1841088533,-0.4304438233,0.0109233474,-0.0465680286,0.0336023383,0.3165078759,-0.3026174903,0.1214015037,0.1358111054,0.0017412055,-0.2351055741,0.5520071983,-0.0232513864,0.0332288444,0.3382865489,-0.1980383694,0.283485949,0.4085797369,-0.2060970664,0.1184658632,-0.1302348822,-0.1976927668,0.2686095834,0.2723291218,0.0560342371,-0.0271525364,-0.0440820977,0.240311712,-0.3981290162,0.0236621723,0.4464373291,0.1539865881,-0.1511222571,-0.4074274302,0.3398567736,0.3760689497,-0.1697612852,0.8557540774,-0.319221139,-0.274861902,0.4450088739,0.4598819613,0.7785283327,-0.3535764217,0.3509236574,-0.1129698381,-0.4027541876,-0.3458400369,-0.1995653659,0.0818229169,-0.4845446646,0.0333420895,0.0910126641,-0.2487891316,0.2100740224,0.3734500706,-0.0257142261,0.0497209094,-0.1206620336,0.192477867,0.049335625,0.0023938187,0.2156998217,-0.201844573,0.1236746162,0.0153348418,0.0675122812,-0.1863373667,-0.0363620222,0.0534432381,0.234995991,-0.1699988544,-0.2345427722,-0.1118467003,-0.5601986647,0.2767541111,-0.0396716781,0.2962335944,0.0277380534,0.0559035391,-0.4646374583,0.1299652308,0.0435671844,0.0832821801,0.4669635892,0.2642526329,0.1506468803,0.0136833284,0.0841357559,-0.0185435526,-0.3130716681,0.1003361419,-0.2381833643,-0.236597091,-0.4385222793,0.3223330975,0.2301787287,0.0799803436,-0.2859732509,-0.0877583399,-0.1497965157,0.1472655088,0.0254738908,0.1634956747,0.0104221432,0.7786630988,-0.215199247,-0.3480201662,0.0423698388,0.1496226341,-0.0276839659,0.1380488873,0.2621843517,0.026845783,-0.1049381793,-0.0378299579,0.2424715608,0.0739662722,-0.2794212401,0.1698547155,0.0420329571,-0.1616180241,-0.0963286906,-0.1669205278,-0.1237559691,-0.2514385879,-0.2342485934,-0.1139549538,-0.5196937323,-0.1651578099,-0.2643464506,0.2441367507,0.0532663204,-0.4230463803,0.1131858751,0.4644656479,-0.1100267991,0.1683980823,0.133918941,0.1894161105,-0.0244433992,0.0971354842,0.1988367736,-0.1879514903,-0.1095895469,0.2368977964,-0.2532469928,-0.0861600637,-0.219805792,0.2332850695,-0.1953471303,-0.2172621489,0.0440846942,-0.2931400836,0.1356250048,-0.3491421938,0.2578807175,0.0838899761,-0.2328194678,0.0373020731,0.2388148159,-0.3311472535,0.220101729,0.4545754492,-0.0425617471,0.4388764501,0.1643185765,0.1744069308,0.181005314,-0.083299391,-0.1607128084,-0.0558805615,0.1490997076,0.0030748246,0.2341903597,-0.2745847702,0.0132634528,0.0800812766,0.0003390793,0.1882117093,-0.2667371035,-0.1475630403,0.0492926203,0.0782807767,0.0134438984,-0.2528131306,0.4869853258,0.1127323806,-0.2317196131,0.1365434676,0.2346690148,-0.1223779544,-0.2093344331,0.2005377859,0.0721568912,-0.539724052,0.0281284098,0.1619892865,0.0147951692,0.0210542027,0.2359562069,0.2058278471,0.2101635784,0.5615077615,0.308863163,0.0974673182,0.3279837966,0.0873563439,-0.1292039901,-0.5117123723,-0.0755153075,0.6525540948,-0.2939630449,0.3701236546,0.0529409796,0.3967393637,0.3359198868,-0.5242992043,-0.0725553036,0.319509685,-0.1616052538,-0.164382115,0.2400985062,0.006177125,-0.1096999943,0.1531411707,0.1216986552,-0.1237889454,0.7778648734,0.2835395038,-0.0875753909,-0.1889849901,-0.1279660761,0.2420140952,0.3362987339,-0.2656769753,0.2335510552,-0.2037661225,0.0624339432,-0.1059367508,-0.0529247634,0.4486576915,0.44293648,-0.201753363,0.1336961985,-0.045043882,0.1585483253,-0.1356239766,0.2330148518,-0.0345740765,0.1134563014,-0.0394150876,-0.0856285393,0.0248609614,-0.2527594864,0.2364505529,0.4464548826,-0.1928068697,0.2965643406,0.0010081339,-0.2750468552,0.130255729,-0.2166998684,-0.1702964455,-0.025325723,0.6866876483,-0.6103777885,0.1735483557,-0.243972972,-0.0195740908,-0.2057754397,0.5154132247,0.0212637112,0.1183216572,-0.2750031352,-0.0961365625,-0.4455814362,0.0383725502,-0.4475418329,0.0241199862,0.2209085822,-0.0205838159,-0.4118451178,-0.1685489565,0.4470587373,-0.1555504054,0.1067610234,0.2095529437,-0.4646930695,0.1016210765,-0.1060333326,0.0645982772,0.0489251651,-0.5800876021,0.310274601,-0.2063927054,-0.0724288374,-0.216164887,-0.1279750168,0.0831951872,0.1493741572,0.7480970025,0.1544769108,0.0933578089,-0.1013715044,-0.1662064195,-0.2386661619,-0.0092877988,-0.1082655862,-0.300763011,0.0482407287,0.2131523788,-0.4186708629,-0.3162444532,-0.0958471447,-0.0973260105,-0.0126645286,-0.1017910093,-0.0313046053,-0.0376612134,-0.0718604177,0.2585251033,-0.0472588912,0.4419737458,0.0065752571,0.1275044382,-0.2060889602,-0.1701948643,0.3051473796,-0.5047957301,-0.5666652322,0.2486843616,0.2222003192,-0.291354835,-0.4132839739,-0.4088248014,-0.3057343662,0.3120827973,0.1108765006,-0.3098548055,0.0957601368,-0.2816905677,0.0954844803,-0.0551786721,0.3852954805,0.0510675237,-0.1665999889,0.2219670415,-0.1396966875]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2554","title":"Multilabel metrics not supported","comments":"Hi @GuillemGSubies, thanks for reporting.\r\n\r\nI have made a PR to fix this issue and allow metrics to be computed also for multilabel classification problems.","body":"When I try to use a metric like F1 macro I get the following error:\r\n\r\n```\r\nTypeError: int() argument must be a string, a bytes-like object or a number, not 'list'\r\n```\r\nThere is an explicit casting here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075\/src\/datasets\/features.py#L274\r\n\r\nAnd looks like this is because here\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075\/metrics\/f1\/f1.py#L88\r\n\r\nthe features can only be integers, so we cannot use that F1 for multilabel. Instead, if I create the following F1 (ints replaced with sequence of ints), it will work:\r\n\r\n```python\r\nclass F1(datasets.Metric):\r\n def _info(self):\r\n return datasets.MetricInfo(\r\n description=_DESCRIPTION,\r\n citation=_CITATION,\r\n inputs_description=_KWARGS_DESCRIPTION,\r\n features=datasets.Features(\r\n {\r\n \"predictions\": datasets.Sequence(datasets.Value(\"int32\")),\r\n \"references\": datasets.Sequence(datasets.Value(\"int32\")),\r\n }\r\n ),\r\n reference_urls=[\"https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html\"],\r\n )\r\n\r\n def _compute(self, predictions, references, labels=None, pos_label=1, average=\"binary\", sample_weight=None):\r\n return {\r\n \"f1\": f1_score(\r\n references,\r\n predictions,\r\n labels=labels,\r\n pos_label=pos_label,\r\n average=average,\r\n sample_weight=sample_weight,\r\n ),\r\n }\r\n```\r\n","comment_length":25,"text":"Multilabel metrics not supported\nWhen I try to use a metric like F1 macro I get the following error:\r\n\r\n```\r\nTypeError: int() argument must be a string, a bytes-like object or a number, not 'list'\r\n```\r\nThere is an explicit casting here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075\/src\/datasets\/features.py#L274\r\n\r\nAnd looks like this is because here\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075\/metrics\/f1\/f1.py#L88\r\n\r\nthe features can only be integers, so we cannot use that F1 for multilabel. Instead, if I create the following F1 (ints replaced with sequence of ints), it will work:\r\n\r\n```python\r\nclass F1(datasets.Metric):\r\n def _info(self):\r\n return datasets.MetricInfo(\r\n description=_DESCRIPTION,\r\n citation=_CITATION,\r\n inputs_description=_KWARGS_DESCRIPTION,\r\n features=datasets.Features(\r\n {\r\n \"predictions\": datasets.Sequence(datasets.Value(\"int32\")),\r\n \"references\": datasets.Sequence(datasets.Value(\"int32\")),\r\n }\r\n ),\r\n reference_urls=[\"https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html\"],\r\n )\r\n\r\n def _compute(self, predictions, references, labels=None, pos_label=1, average=\"binary\", sample_weight=None):\r\n return {\r\n \"f1\": f1_score(\r\n references,\r\n predictions,\r\n labels=labels,\r\n pos_label=pos_label,\r\n average=average,\r\n sample_weight=sample_weight,\r\n ),\r\n }\r\n```\r\n\nHi @GuillemGSubies, thanks for reporting.\r\n\r\nI have made a PR to fix this issue and allow metrics to be computed also for multilabel classification problems.","embeddings":[-0.2212646902,-0.1788480431,0.0044772709,0.3008992672,0.6725280881,-0.135990411,0.4806858599,-0.1023353487,0.2283084393,0.3166798353,-0.1240749434,0.2981927693,-0.2808501124,0.4046333432,-0.2508816123,-0.176120922,-0.2169739455,-0.3327276409,-0.0599899776,0.1059186533,-0.42456007,0.0359555557,-0.3928985894,-0.0474174023,-0.2053830773,0.0682160556,0.1161154285,-0.1135555729,-0.2787645161,-0.329521507,0.0360861979,-0.0185579155,-0.0064595272,0.4338566363,-0.0001161838,-0.1123573929,0.2313463986,0.0006338583,-0.1215105578,-0.2388671041,0.1006879508,-0.1241763309,0.0392229259,-0.2491005361,0.0151843484,0.0634786934,-0.3937759399,-0.1104588732,-0.0291239601,0.3288370967,0.1258651316,-0.062961854,0.4294290543,-0.0971220285,-0.1565203965,-0.015151497,-0.2259768844,0.2436192334,0.0067122462,0.1893280596,0.1993249655,0.367898792,0.0551728383,0.0967107713,0.3370931745,-0.0944951847,0.1595984548,-0.1680668592,-0.1602084935,0.2209853083,-0.061623808,-0.0668189675,-0.592457056,-0.0729389489,0.0992546678,-0.8592536449,0.1026922315,-0.1433411539,-0.0054159793,-0.1943722367,-0.4583276808,-0.1244826615,-0.1109633297,0.2019526213,-0.1810643077,0.4406156242,-0.1584268361,0.2374359667,0.3454503417,-0.1518542767,-0.2052839398,0.0081735235,0.05832633,0.205610618,-0.3568544388,0.1093110591,-0.1768293828,0.0013985224,0.1842876226,-0.1633948386,-0.2385578007,0.1422113329,-0.3320725262,0.0641800165,0.122553058,0.3998177052,0.2546080351,0.1608308256,0.072079584,0.1423228383,-0.1401439011,-0.156781882,-0.0698929578,-0.36560449,0.2567168474,0.3489768207,-0.2015082091,0.0135226008,-0.3946036994,-0.0448176153,-0.0712453648,0.0082791513,0.6896387935,0.091719225,-0.0318083167,0.2214342058,0.1153474003,0.4436763823,0.068796277,-0.2091442794,-0.0214124601,0.0479799248,-0.3339827061,0.048268538,0.1806904525,-0.0014864613,0.0539554358,0.1387739629,0.4493706822,-0.2276753336,-0.1483097225,-0.2939902842,-0.2085352242,0.1655087322,-0.0190413315,0.1512502879,0.3160668612,-0.2302957177,0.0357704349,0.0669233426,-0.4116387963,0.0382435285,-0.0666187927,0.0712282583,0.0483757332,0.1025643498,-0.2542997301,0.2213602811,-0.1325968355,-0.0665762275,-0.0036140094,-0.4038878083,-0.2503187954,-0.1125381142,0.2882920504,-0.1103660837,0.2029678673,-0.141621694,0.0057303826,0.1887963414,0.0668105632,0.119579576,-0.0640976131,0.03594134,0.0520436391,-0.0502258129,0.5578497052,-0.77246207,0.0221548453,-0.1110544279,0.1622603387,-0.1910859942,0.1772535294,0.0577876568,-0.0404196531,-0.2197519988,0.4865655899,0.0727670565,0.0616688654,-0.1428972483,-0.089566648,-0.1569128186,-0.0444326475,0.0377983116,0.3760761619,-0.0571965948,0.1041150764,0.380446434,-0.1173127294,-0.2153507024,0.0217813794,0.2393888831,-0.108428888,-0.2037583739,-0.021731142,-0.519529283,-0.2728665769,0.2056166083,0.0130411144,0.1000807732,0.3230245709,-0.1054260582,-0.2257852107,0.1438708901,0.0650469065,0.1493526995,0.0401068889,-0.0796515793,-0.0936166421,-0.0056920196,-0.131843403,-0.0627465919,0.1535237432,0.157718271,0.351640135,-0.0704767704,-0.0087629184,-0.08259812,0.0973036438,0.3517900109,0.434766084,0.1407412291,-0.0306357052,0.3947176039,0.2970744967,-0.5094220042,0.1806082875,0.3879396617,0.3607369959,-0.0730337054,-0.1506708413,-0.0504656769,0.2503890693,-0.2185509503,-0.1461922526,0.3504961729,-0.3115260899,0.3024228811,0.1430209726,0.3508873284,0.2961013317,0.1546096355,-0.2565431297,-0.3138019741,0.1629935205,-0.2991591692,-0.100536041,-0.109675996,-0.3435279727,-0.073240377,0.4491719604,0.2249610424,0.3339295983,0.1307243556,-0.045668751,0.1011583656,-0.2851280272,-0.4623366594,0.3690942228,0.0325323045,-0.0324040838,-0.0058949338,-0.2700992525,-0.1267909259,0.3215012848,0.0240250342,0.0201605055,0.0583690777,0.0777406767,0.012755651,0.0487495102,-0.0738748312,0.1537625641,0.0149455015,-0.3629771173,0.0582905561,-0.2453487366,-0.0463334993,-0.0935079902,-0.7585228682,-0.1761825681,-0.3346891999,-0.0403061807,0.2386613339,0.2290695608,0.2197218984,0.1551398933,0.2061389089,0.1241323799,0.1301218271,-0.0368105024,-0.1414488852,-0.1046102792,-0.0305022672,0.1273281425,-0.0723791197,0.0829007328,0.1874415874,0.2402452976,-0.2738409936,-0.5661293268,-0.0316678844,-0.0450724177,0.6306042075,0.2490514666,-0.359347105,-0.3242051303,0.2313190401,0.5897344947,-0.0427177288,0.2254867256,0.3594716191,-0.0859736204,-0.2206724435,-0.3616174459,0.3093791008,-0.0327621847,-0.2078435868,0.2410562336,0.214681536,-0.1038077176,0.3323223889,0.0637742728,0.1596950591,0.333335042,-0.0583828874,-0.2203381211,0.1170348898,0.1210966036,-0.2581457496,-0.2353103757,-0.0685076118,0.2108470351,0.1027471945,0.0835000724,-0.0877119899,-0.6157526374,0.0433085933,-0.1099518612,0.0962612256,0.353251636,-0.0236948617,-0.0236146413,-0.005031175,-0.295424372,-0.2439953089,0.1375765949,-0.2149140388,0.1514301598,-0.0051483507,0.1247481629,0.1719860137,0.1946258247,0.3539066315,0.0950792506,0.4159371555,0.017634863,0.2028158158,-0.1227415055,-0.496014297,0.2217462361,-0.0935750976,0.0730176196,0.0644242316,0.0248819347,0.2972833514,-0.2300756127,0.405313164,-0.4326803684,-0.2399402261,-0.1216048226,-0.0679211244,-0.0599618703,0.006851099,-0.0686789975,-0.4538890719,-0.0174990222,-0.0197698828,0.0618204623,-0.0652201772,-0.0690514445,-0.1912250221,-0.5020269156,-0.0877895802,0.3013484478,0.1033227816,0.3473795056,-0.0139065851,0.4603227973,-0.0922817588,0.0699033737,0.4486807287,-0.3780311048,-0.164893046,0.4228254855,0.0128902271,-0.3764905334,-0.0492267087,-0.3481476009,0.1504316479,0.217348367,0.5315402746,-0.354280293,-0.1013862118,0.297542721,0.0651202351,-0.0705221593,0.041090358,-0.1671097726,-0.2446780354,-0.0693645701,-0.1013950258,0.1543728858,0.4377833605,0.115724802,0.2029922307,-0.4573968947,-0.1873931289,0.1859023422,-0.1936880648,0.0104943952,0.1093759686,0.0429961383,-0.1465334743,-0.2478698492,-0.0152926072,0.3211365044,-0.0384486169,-0.6306096911,0.0358046107,-0.2560366988,0.3225225508,0.0526942052,-0.2152550817,0.0528170466,-0.3137719333,0.4578759372,-0.0779881179,0.0748886541,0.4235395193,0.0654163659,-0.1905528009,-0.1197462156,0.4245311022,0.2586815655,-0.1111522391,0.5211781859,0.0430513322,-0.3598251939,-0.0759665519,0.5890723467,0.697598815,-0.2226115763,-0.2586009204,0.1672002077,0.082353048,0.5033921003,0.193038106,0.1268890947,-0.0474989377,-0.1426634043,-0.0115385586,0.0176427756,0.0291983299,-0.0812455863,-0.2721316218,0.4355119765,-0.3233117163,-0.1192600951,-0.0509028435,0.0212952197,0.1017358601,-0.3318385482,0.0844058618,0.108577691,-0.0523183942,-0.0902343392,-0.1292123795,-0.044476673,0.0880723819,-0.078350164,-0.3915770054,-0.2225100845,-0.2379131764,-0.0229880251,0.1737508327,0.2766945362,-0.0461669341,0.371722728,0.8833089471,-0.0897018462,-0.2677648962,0.0498953126,-0.4189761281,0.1067224592,-0.1613913924,-0.0561747253,0.3169288039,0.0474147499,0.1192992255,0.1314159036,0.1853279322,-0.092199713,0.0002037796,0.0091381036,-0.1528532654,-0.0397046804,0.1344159693,-0.146582067,0.1696133614,-0.2701961398,0.030994853,-0.0420934483,0.0146137271,0.3356792629,-0.1573169827,0.0131884003,0.1277573407,0.0560097843,-0.013466835,-0.1197617352,-0.0039821025,0.1132163182,-0.0096258847,-0.0553261377,0.0066845352,-0.1056758538,-0.2056487203,0.4466662109,0.2419058383,-0.128834486,0.2151128799,0.730645895,0.017122604,0.1984635293,0.008393513,-0.4410927594,0.0696252659,0.2799544036,-0.3629091382,-0.267287612,-0.3785532713,0.2672263682,0.0635074154,0.0819196776,-0.2287498564,-0.0512527227,-0.3522161245,-0.0120719913,0.0662601143,-0.1235128343,-0.0876618922,0.2427785844,0.0234477073,-0.0590097457,-0.1183175743,-0.0645193234,-0.3597303629,0.1503220797,0.1956708729,0.0865002424,-0.0612365045,-0.185258314,-0.2648679614,-0.295303911,-0.2420809269,0.3174176812,-0.1273352504,0.1342581064,-0.2086674273,0.4146332741,0.2062968761,0.0100608235,0.2734764814,0.2579463124,-0.1087879911,0.1787582487,0.2047462761,0.0324226394,-0.3954749405,-0.0599185899,-0.0124586243,-0.0995980129,0.2261797041,-0.1581040621,0.2550148666,0.3032354712,0.2987625301,0.0989706144,-0.1419446021,0.0377803706,-0.0452909321,0.0828032419,-0.0907506868,-0.342012316,-0.1665186584,0.1157026663,0.1041974351,0.4284951091,0.2192299366,0.1410691589,0.1890291274,-0.2953188419,0.2003716677,-0.1582850963,-0.1535173059,0.2303367406,0.2160973102,0.1102144197,0.513872087,0.0970058441,0.3311564922,0.5222990513,-0.2508504987,0.1869079918,-0.0841608867,0.0725290924,0.030389091,0.0947314575,0.1601162851,0.0957945287,-0.2019709796,-0.0214736722,-0.4733527899,0.0767652169,-0.7301226258,-0.2037562281,-0.3301917017,0.4148744643,0.091410622,-0.2277126163,-0.3952122331,-0.1649126709,-0.1175167263,-0.1400968134,-0.1764771491,0.2861173749,0.3323256671,-0.0380243957,0.1797473431,-0.2377058268,-0.7110541463,0.4802883863,0.2804578841,-0.1638116986,0.2355116308,0.2511024177,0.1285534501,0.1014561281,0.6210411191,0.4934696853,0.3399618864,-0.1255110204,-0.2191705704,-0.2566690147,-0.1811932623,-0.0969048366,0.1962878853,0.4486635923,0.0767921582,0.1382708997,0.0415118299,-0.1604165435,0.3234142661,-0.310670197,0.1860601753,-0.2885332108,0.2526659369,-0.0916091278,0.0181876309,0.0682785809,-0.1238181815,-0.3758455515,-0.0621205494,0.2017790377,-0.0491471514,0.2422328293,-0.061436221,0.0370758586,0.2704456747,0.2974847257,0.0136611946,-0.0163828246,-0.2051001489,0.0752962753,-0.5276011825,-0.1579034179,0.1876159906,-0.0403055288,0.2682581544,0.0356174931,-0.0518562943,0.2787484527,0.2536309659,-0.0934768245,0.3922421634,0.1186252013,-0.3982809782,0.1605590582,0.1794386357,-0.1133188978,0.060497202,-0.2399597466,0.1822281629,-0.3501202762,0.0218922459,-0.2242321074,0.2188231349,0.0837772116,0.2790906131,0.2434481233,0.2892829776,0.4600458741,0.1918538213,-0.027339885,0.0433669277,0.0758345053,0.1163827032,0.0913149416,-0.1747367829,0.300878197,-0.3320218325,0.0073987804,-0.1984241456,0.0825693011,0.0664646104,-0.1721780598,-0.2462133169,0.2692374885,-0.3208706379,-0.0951891989,0.3345551491,0.4876784682,-0.035236761,0.4263415933,0.0300742518,-0.3520091474,0.3873305023,-0.2040195316,-0.3546736836,-0.1390377581,0.2584775984,0.0670106411,-0.1050567925,-0.650236249,-0.1225927025,0.2604496777,0.2896447182,-0.5856113434,0.2949889898,-0.1904971749,-0.1474583894,0.0136125553,-0.3887819052,-0.0953037366,0.031689696,0.1335062832,0.2647954822]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2554","title":"Multilabel metrics not supported","comments":"Looks nice, thank you very much! \ud83d\ude80 ","body":"When I try to use a metric like F1 macro I get the following error:\r\n\r\n```\r\nTypeError: int() argument must be a string, a bytes-like object or a number, not 'list'\r\n```\r\nThere is an explicit casting here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075\/src\/datasets\/features.py#L274\r\n\r\nAnd looks like this is because here\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075\/metrics\/f1\/f1.py#L88\r\n\r\nthe features can only be integers, so we cannot use that F1 for multilabel. Instead, if I create the following F1 (ints replaced with sequence of ints), it will work:\r\n\r\n```python\r\nclass F1(datasets.Metric):\r\n def _info(self):\r\n return datasets.MetricInfo(\r\n description=_DESCRIPTION,\r\n citation=_CITATION,\r\n inputs_description=_KWARGS_DESCRIPTION,\r\n features=datasets.Features(\r\n {\r\n \"predictions\": datasets.Sequence(datasets.Value(\"int32\")),\r\n \"references\": datasets.Sequence(datasets.Value(\"int32\")),\r\n }\r\n ),\r\n reference_urls=[\"https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html\"],\r\n )\r\n\r\n def _compute(self, predictions, references, labels=None, pos_label=1, average=\"binary\", sample_weight=None):\r\n return {\r\n \"f1\": f1_score(\r\n references,\r\n predictions,\r\n labels=labels,\r\n pos_label=pos_label,\r\n average=average,\r\n sample_weight=sample_weight,\r\n ),\r\n }\r\n```\r\n","comment_length":7,"text":"Multilabel metrics not supported\nWhen I try to use a metric like F1 macro I get the following error:\r\n\r\n```\r\nTypeError: int() argument must be a string, a bytes-like object or a number, not 'list'\r\n```\r\nThere is an explicit casting here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075\/src\/datasets\/features.py#L274\r\n\r\nAnd looks like this is because here\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075\/metrics\/f1\/f1.py#L88\r\n\r\nthe features can only be integers, so we cannot use that F1 for multilabel. Instead, if I create the following F1 (ints replaced with sequence of ints), it will work:\r\n\r\n```python\r\nclass F1(datasets.Metric):\r\n def _info(self):\r\n return datasets.MetricInfo(\r\n description=_DESCRIPTION,\r\n citation=_CITATION,\r\n inputs_description=_KWARGS_DESCRIPTION,\r\n features=datasets.Features(\r\n {\r\n \"predictions\": datasets.Sequence(datasets.Value(\"int32\")),\r\n \"references\": datasets.Sequence(datasets.Value(\"int32\")),\r\n }\r\n ),\r\n reference_urls=[\"https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.f1_score.html\"],\r\n )\r\n\r\n def _compute(self, predictions, references, labels=None, pos_label=1, average=\"binary\", sample_weight=None):\r\n return {\r\n \"f1\": f1_score(\r\n references,\r\n predictions,\r\n labels=labels,\r\n pos_label=pos_label,\r\n average=average,\r\n sample_weight=sample_weight,\r\n ),\r\n }\r\n```\r\n\nLooks nice, thank you very much! \ud83d\ude80 ","embeddings":[-0.2124199718,-0.2649220824,-0.0277619977,0.2559404671,0.6675208211,-0.0826078355,0.4885375798,-0.1410621554,0.2341555059,0.2856134772,-0.0567766465,0.289280206,-0.3134280145,0.4336830676,-0.2627758086,-0.1557724029,-0.1786039323,-0.3309242427,-0.0193667393,0.056326557,-0.4603808224,-0.0141714513,-0.4058319032,-0.1103204489,-0.1867831647,0.0540845022,0.0439994186,-0.179564178,-0.3006687164,-0.3042010367,-0.0482368842,0.0099910721,0.0019855225,0.4133024216,-0.0001172044,-0.0961452276,0.1851305068,0.0622617453,-0.0915136039,-0.2571463585,0.0376628451,-0.0685174465,0.0606373288,-0.2913108468,0.0676790997,0.0842296183,-0.3277393579,-0.1105658859,-0.0562231801,0.3314417303,0.1266815513,-0.0303302612,0.3409955502,-0.1471312344,-0.127254203,-0.0489696302,-0.246680975,0.1912675947,0.0643864572,0.2749714255,0.2831326425,0.3785557449,-0.0164635126,0.1418750733,0.3418272138,-0.1445437968,0.1742265522,-0.1804527789,-0.0840655193,0.2936742306,-0.0154681196,-0.0710409656,-0.5178975463,-0.048205208,0.1052433178,-0.8815641999,0.0009349983,-0.1171477437,0.0724968389,-0.1621774137,-0.4928304851,-0.1282107681,-0.0584561639,0.2359929681,-0.2209374458,0.4811474681,-0.1743741781,0.2771565914,0.3042705655,-0.1642657667,-0.2307415456,-0.0007919242,-0.0103189191,0.1154758334,-0.2931984067,0.1554030776,-0.1724403054,0.0016908324,0.2294374108,-0.1145178974,-0.1969256699,0.1014421582,-0.3698745072,0.0230735335,0.1819493473,0.3484312296,0.2628297508,0.1121478826,0.0159408152,0.0865135789,-0.1663219929,-0.1896100491,-0.0237525161,-0.3516687155,0.2535765767,0.2110987902,-0.1148264259,0.0294320714,-0.4340936542,-0.0322211757,-0.0817598999,-0.068486847,0.6535591483,0.1022422239,-0.0262735467,0.2003535032,0.0917576551,0.5340031981,0.121176511,-0.1451220661,-0.0655244365,0.0531786568,-0.3686731458,-0.0240693726,0.0939935073,0.0458186679,0.0950980261,0.1066411808,0.4535258114,-0.2314519286,-0.0762413144,-0.289611727,-0.1970680505,0.163720265,0.0179127641,0.2281707972,0.2503730655,-0.217388019,0.038964279,0.1046793461,-0.3861687481,0.0763562471,-0.1014577597,0.0841044784,0.0269536451,0.1011254936,-0.1764360517,0.1720023155,-0.119184956,-0.0674515963,0.0433142334,-0.4102597833,-0.2807618976,-0.1576472521,0.2251879871,-0.1397388279,0.2327783108,-0.1540888548,-0.0030654941,0.2407552749,0.0110637322,0.1075658202,-0.0451794788,0.0712130442,0.0440585203,-0.0556483045,0.6493510008,-0.7993515134,0.0260259565,-0.0318466611,0.1663932949,-0.2466087341,0.2025268525,0.1195847467,0.0071383989,-0.2354849279,0.4778146744,0.0479400046,0.0934943855,-0.1308830231,-0.0100600151,-0.2094726413,-0.0257139131,0.0615649708,0.3398530185,-0.0596070997,0.0994833037,0.4313283861,-0.1017776132,-0.1870978177,0.0360614322,0.2481694371,-0.0516361594,-0.18591398,-0.095182538,-0.537764132,-0.2647865117,0.1758207828,0.0234382618,0.1065385342,0.4161784649,-0.1272862107,-0.2348756045,0.1630679369,0.0454282649,0.14111498,0.0239961781,0.0069570788,-0.1445300579,0.0290560164,-0.1237955913,-0.1163067296,0.191255793,0.1505110413,0.3247577846,-0.1260708123,0.0161776431,-0.093300797,0.0500006266,0.3134622276,0.4629411697,0.1626097113,0.009443,0.3725735247,0.2446921915,-0.5158541799,0.1558482349,0.4375321269,0.4626836181,-0.1070585996,-0.1068642586,-0.0743039921,0.2218687832,-0.2918804288,-0.1673095077,0.3222904801,-0.3136721253,0.3093440831,0.0693235546,0.4086685479,0.2878317237,0.1247440949,-0.2625716627,-0.3497853875,0.1894638538,-0.2525572479,-0.1000085697,-0.1282973737,-0.3680262268,0.0252064914,0.4290017188,0.1889296919,0.3358874023,0.0430230275,-0.0914986879,0.1228001788,-0.2380401641,-0.4941650629,0.4001939893,0.0803648382,0.0391329266,-0.024516074,-0.2542221546,-0.1830373555,0.2734763324,-0.0153464144,0.010893506,0.0169928148,0.0225922,0.0079474002,0.101067856,-0.0369775183,0.1209874302,0.0879549533,-0.3463976979,0.0943582579,-0.24239555,0.0217054226,-0.0678672045,-0.8001259565,-0.1436436772,-0.403285116,-0.0749144256,0.3388578296,0.1911534667,0.210531354,0.1459226012,0.1641516387,0.158874467,0.0756177753,-0.0206451956,-0.0774553567,0.0101702409,-0.0122669265,0.1309569627,0.0076318267,0.0849996954,0.1464076936,0.2315855324,-0.2913502455,-0.5188516378,-0.0031358537,-0.017644858,0.6203856468,0.2519396245,-0.2975183129,-0.2741418779,0.2768825889,0.5358743668,0.0162217431,0.2645593882,0.3505828977,-0.1863172203,-0.2382915914,-0.3560905457,0.334020108,-0.0466383733,-0.2529991269,0.2238712311,0.240534842,-0.1578063667,0.2627293766,0.176294595,0.1791907102,0.3448338509,-0.028551409,-0.2406422645,0.0978496149,0.122048296,-0.2557986677,-0.207194984,-0.0246640816,0.2649290264,0.0795969144,0.0631021485,-0.1278837621,-0.6162597537,0.0123551385,-0.1196846887,0.1717787236,0.3743750751,-0.0162245724,-0.0433353148,0.0035014767,-0.2936521471,-0.1666945517,0.1382914484,-0.1842327267,0.1477874815,0.0273152273,0.0960973576,0.1149578914,0.1504339874,0.38669011,0.0514217503,0.4117040932,-0.064155221,0.2091182917,-0.1938811243,-0.4914137125,0.2146579772,-0.0512598455,0.0699252114,0.0734620541,-0.0627991632,0.3350553513,-0.1617418975,0.4425060153,-0.4361581206,-0.2379225641,-0.0796519369,-0.0873462707,-0.1390441507,-0.0183441713,-0.1346884221,-0.3797444701,0.0119314706,-0.0404049531,0.0521624312,-0.0998514295,-0.1216749176,-0.214387536,-0.4705291688,-0.0600763373,0.3202597499,0.0529363416,0.2904346585,-0.0524487533,0.3947703242,-0.0844655558,0.0837923363,0.4947989881,-0.4074826241,-0.1686656475,0.363673836,0.0575676113,-0.3777257502,-0.0260665379,-0.3697838485,0.2485998422,0.15937002,0.5363642573,-0.3472309411,-0.088783741,0.32000646,0.0758738369,-0.1197142452,0.0870342106,-0.1816606671,-0.2839707434,-0.15154019,-0.1236584857,0.1818479896,0.4089003205,0.1526658535,0.2374615818,-0.4579610527,-0.1921864897,0.2231288254,-0.1784106344,-0.0079766763,0.1004050896,0.0424514785,-0.1547402591,-0.2790640891,-0.0294381082,0.3090852797,-0.0470658727,-0.5943213701,0.0804763511,-0.241994366,0.3470680118,0.0608575009,-0.1303436458,0.1187158376,-0.2834308147,0.4525172412,-0.0683207363,0.1074945107,0.4983336329,0.0390491076,-0.1911455989,-0.1997217685,0.3875386417,0.2372294515,-0.1467208266,0.5585436821,0.066093199,-0.3383106887,-0.0836797431,0.5380544066,0.705791831,-0.2002166212,-0.2312993854,0.2113631368,0.0565261804,0.4557104409,0.2014187276,0.0394101106,-0.0915660709,-0.160526365,0.0211289525,-0.0094215628,0.0619248711,-0.0751715079,-0.297930181,0.3922971487,-0.3810495436,-0.0554327443,-0.0923303738,-0.038841296,0.1160878986,-0.2740875483,0.1374502331,0.0965259522,-0.0613583513,-0.1843062043,-0.1380311549,-0.0580423698,0.1286965907,-0.095291689,-0.3758384883,-0.2738791108,-0.187144205,-0.0849190503,0.1752566695,0.2964728773,-0.0160606168,0.4056352973,0.9025189877,-0.0880136192,-0.2796676755,0.0960420296,-0.4201096892,0.1374956071,-0.1985163689,-0.0789425448,0.3197676539,0.070866622,0.0469681285,0.1874832809,0.2261170596,-0.0547596626,-0.0100824805,0.0692432448,-0.2189766616,-0.119744204,0.0206462089,-0.0286813304,0.1279129386,-0.2912886441,0.0157362539,-0.0220564269,-0.019832518,0.3424466252,-0.120482035,0.0006601104,0.0727300644,0.03394752,-0.0268984158,-0.0885095894,-0.0346981809,0.0906978324,0.0086601265,-0.0847729445,0.0568686873,-0.0933081284,-0.1726879925,0.428958118,0.2283845097,-0.1406365633,0.173134774,0.6909303665,0.0053258035,0.2540560067,0.0142761609,-0.3888173401,0.1143674925,0.3589311838,-0.336407721,-0.2101251632,-0.4384399354,0.2148440927,0.1217932329,0.04600209,-0.23363325,-0.0080757067,-0.2659704685,-0.0066048964,0.060949631,-0.0911362246,-0.0778691247,0.2413511872,0.0333975889,-0.1031416282,-0.1583179086,-0.0789260715,-0.404712081,0.1600351632,0.2023983002,0.0883708373,-0.032511957,-0.2478602678,-0.2132098824,-0.2659039795,-0.2115649879,0.3085674345,-0.1133212522,0.1264417171,-0.1476258487,0.3681426942,0.1213265136,-0.0055228872,0.2675191164,0.2274594754,-0.0779115483,0.1620307714,0.240172714,0.0179422293,-0.3960998654,-0.0478878506,0.0716410875,-0.0879023299,0.1310120076,-0.1587348431,0.3005702198,0.3109324872,0.3294330537,0.0736039206,-0.1870553792,0.0539639331,-0.0683692843,0.0994333029,-0.0911893398,-0.3655578792,-0.1387530565,0.0886234045,0.1254117042,0.4795015156,0.2135245502,0.1446723342,0.1444963962,-0.2641450167,0.2513200045,-0.1831342429,-0.1646749228,0.2412635535,0.1318867058,0.1003081426,0.5232380629,0.1629894674,0.3496321738,0.5778896213,-0.2151542455,0.2063863426,-0.0367759131,0.0194313414,0.0855761766,0.1383262128,0.16384767,0.0856896192,-0.25901407,-0.0224058609,-0.4371235669,0.0915290415,-0.7155743837,-0.2267733067,-0.3219794035,0.434987843,0.0514351614,-0.2300952971,-0.4359657466,-0.1703725457,-0.1536077708,-0.1560948789,-0.1497558057,0.2554893494,0.3891400099,-0.0960062221,0.215602532,-0.2389438748,-0.632670939,0.3725712299,0.1754340976,-0.2038854361,0.2288094759,0.2496183515,0.1283124536,0.1223018393,0.6421701312,0.4826949835,0.2781298459,-0.1067787483,-0.1766202003,-0.298923105,-0.1976251453,-0.1415518373,0.135800913,0.4508109093,0.0365635008,0.1026513875,0.0584670492,-0.1898612529,0.4722270668,-0.2238818705,0.1745929718,-0.3038282394,0.23047553,-0.0851877406,-0.0414032266,0.1003336832,-0.0946044847,-0.435808599,-0.0617324114,0.277731359,-0.1104777828,0.252467066,-0.0163942501,0.0365639888,0.3115611076,0.2436770648,0.0175290238,0.0525340103,-0.2053918839,0.1017318144,-0.5426405072,-0.1555577666,0.1841942519,-0.0024478058,0.2856030166,0.0248564053,-0.0498502627,0.2749907076,0.2725119889,-0.0734323785,0.3833194673,0.1382140368,-0.3990606368,0.1147438064,0.2230138481,0.0020321535,0.0664459318,-0.3007329404,0.1887245178,-0.3871386647,0.0522104912,-0.2465687245,0.1945902109,0.0857204422,0.2836503983,0.2529189885,0.2705838084,0.4399699569,0.1805644035,-0.0328347236,0.0605230294,0.061339628,0.1640443653,0.0187801495,-0.1255956292,0.2805776894,-0.3616366684,0.0208423994,-0.1719080806,0.002462768,0.0492042825,-0.2043322325,-0.2894050479,0.3535212874,-0.3975231946,-0.067303665,0.2838601768,0.5105024576,-0.0039950632,0.4010973573,0.0831482559,-0.3167667091,0.4016519189,-0.1574972719,-0.3965288699,-0.0818812847,0.2341418564,0.0107412152,-0.0575532354,-0.6545410156,-0.1647710353,0.2822321355,0.2498640716,-0.5913078189,0.2522931397,-0.1604697406,-0.1991998851,0.0408674292,-0.3741008341,-0.0538129769,-0.0171347708,0.1061006635,0.1980424672]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2553","title":"load_dataset(\"web_nlg\") NonMatchingChecksumError","comments":"Hi ! Thanks for reporting. This is due to the WebNLG repository that got updated today.\r\nI just pushed a fix at #2558 - this shouldn't happen anymore in the future.","body":"Hi! It seems the WebNLG dataset gives a NonMatchingChecksumError.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('web_nlg', name=\"release_v3.0_en\", split=\"dev\")\r\n```\r\n\r\nGives\r\n\r\n```\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/gitlab.com\/shimorina\/webnlg-dataset\/-\/archive\/master\/webnlg-dataset-master.zip']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: macOS-11.3.1-x86_64-i386-64bit\r\n- Python version: 3.9.4\r\n- PyArrow version: 3.0.0\r\n\r\nAlso tested on Linux, with python 3.6.8","comment_length":31,"text":"load_dataset(\"web_nlg\") NonMatchingChecksumError\nHi! It seems the WebNLG dataset gives a NonMatchingChecksumError.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('web_nlg', name=\"release_v3.0_en\", split=\"dev\")\r\n```\r\n\r\nGives\r\n\r\n```\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/gitlab.com\/shimorina\/webnlg-dataset\/-\/archive\/master\/webnlg-dataset-master.zip']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: macOS-11.3.1-x86_64-i386-64bit\r\n- Python version: 3.9.4\r\n- PyArrow version: 3.0.0\r\n\r\nAlso tested on Linux, with python 3.6.8\nHi ! Thanks for reporting. This is due to the WebNLG repository that got updated today.\r\nI just pushed a fix at #2558 - this shouldn't happen anymore in the future.","embeddings":[-0.1927948296,0.1298394948,-0.1363762766,0.0406420827,0.2651570141,0.0086788097,0.1319979727,0.4511449337,0.2963729203,0.086196363,-0.0655137524,0.2723962069,0.0220884141,-0.0255221389,-0.2154660523,0.5214452147,0.075551413,0.016244445,-0.1093697101,-0.1461790502,-0.2066713274,0.1184713989,-0.0733739063,-0.1003217548,0.0636151209,0.2236549407,0.249917537,0.2134229541,-0.1248321831,-0.2628319561,0.3232803941,-0.0495467335,-0.0851117894,0.2211042494,-0.0001071248,0.1515761912,0.3256528974,-0.0256441869,-0.4283015132,0.0641530007,-0.5148388743,-0.5065795779,-0.0188274365,-0.2006548047,-0.0035357026,0.2438147515,0.0948674753,-0.1746021509,0.1221942008,0.1583404988,0.2839369178,0.5558196306,-0.005537665,0.1301896572,0.0390270576,-0.1125152558,-0.1401671618,0.2185910642,0.0809826106,-0.1305690259,-0.3386833072,0.172543779,-0.2540836036,0.2696690559,0.1147568449,-0.098241888,0.231546253,-0.0817004517,0.1634391844,0.3882210851,0.1614823788,-0.2111613154,-0.2111985087,-0.1415819675,-0.0977410823,-0.0648079664,0.2398231328,0.3723757565,-0.2639215589,-0.0806604624,-0.3824211955,0.240636006,0.1303239018,0.1618121117,0.0931689367,0.2458066791,0.0526399836,-0.0170343257,-0.0117075425,-0.2015507966,0.1708225906,-0.4176906049,-0.1616995633,0.0688045099,-0.5657450557,-0.1234653592,0.2419984341,0.5164675117,0.3947955966,0.4451564848,0.2517745197,0.244819656,-0.2092456222,0.1271482259,-0.0440329276,0.1494938731,0.2171945274,-0.0837601125,0.325415343,0.002421665,0.026335502,0.2340878099,0.1602009684,-0.1498559713,0.574976027,0.0011791078,0.1858932823,-0.1588828862,-0.2398215681,0.2297268063,0.115285553,-0.1611865759,0.0508991592,0.3672366738,-0.1573462486,0.2044910789,-0.0839353651,0.1732858866,-0.1468550116,-0.2998718023,-0.2878326774,-0.2022191882,-0.0596311949,-0.0296896789,0.2052143216,-0.0602190644,0.3227382302,0.1260931194,0.0152073121,-0.0398379937,0.1705985218,0.1035452336,-0.080589667,0.2938563526,0.0489572063,0.1108821929,0.2325149775,0.0842916146,-0.151392132,0.0721646696,-0.1199170724,-0.2106702328,0.0581686981,0.2775703967,-0.3827539682,-0.1413070709,-0.0195726398,-0.4089507163,0.2865951657,-0.0916863605,0.0641917363,-0.2404451072,-0.1441319138,-0.2984170318,-0.1050308645,0.1696690917,0.2377004325,-0.0296219829,-0.1784476489,-0.1316381991,0.3148423731,0.3782692552,-0.1069779247,-0.1547899544,-0.2393767089,0.1331310719,0.5495225787,-0.1487931311,-0.4355504215,0.1457467973,-0.0122829396,0.2730146945,-0.1001160145,0.246824041,-0.428412199,0.0206717346,-0.2852042317,0.2863252759,0.0123607414,0.2956215441,-0.2442984581,-0.2357428223,0.539798677,0.0765879378,-0.0410942547,-0.0127481418,0.1415346265,-0.2447985858,0.3284392953,-0.2872792482,-0.0617064722,0.073921077,0.2371894121,0.0206666738,0.0100722183,-0.0122238947,-0.5249134898,0.2644129097,-0.1044167131,0.0427522399,-0.0249289013,-0.070720382,-0.1582019478,0.0300226379,-0.1433977038,0.0346075222,0.1706935912,0.5151420832,0.2642631233,0.0445589423,0.0240484457,0.283557117,-0.3699106276,0.0000202275,-0.5306888223,0.2723832428,-0.0884009078,-0.0901313201,0.084347032,-0.0441488251,-0.0575295016,0.1041356251,-0.2234271467,0.3572004437,0.0527738556,-0.0378873199,0.0728344172,0.329175204,-0.1157856584,-0.2637716532,-0.0720948353,0.6472374201,0.1213914752,-0.107889019,-0.1684261411,0.5147510171,-0.1285653263,0.1689170897,-0.0175983291,0.1398556232,0.2636683583,-0.1620629877,-0.2084147483,-0.1159797087,0.2796030045,-0.1819343716,0.2042548656,0.1351904422,-0.2331168056,0.0146668628,0.4084135592,-0.1645733267,0.2219512314,0.1339526176,0.236446023,-0.2454288006,-0.0098468829,0.1951207072,0.366452992,0.1531024873,0.0423474237,0.059913259,-0.2892881334,-0.232820347,0.0300007369,0.0216752701,0.2292783856,0.2969275415,0.2905500531,-0.2966411412,-0.4513899386,-0.106641233,-0.0866571814,0.2381994575,-0.3005838692,-0.0691240951,-0.3129399717,-0.2774086595,-0.2933863997,-0.3166306913,-0.4951165915,-0.5132744908,-0.0091083134,0.0563070513,-0.1829127222,0.3383725286,-0.4450571239,0.0377307571,0.1850384772,0.086943835,-0.0377457477,-0.1209628209,-0.0876403674,0.055346366,0.377540797,0.0115627637,0.3231642842,-0.3149303496,-0.0758119151,-0.3380955756,-0.228727147,0.0264027063,-0.0950912088,0.0999893546,0.213042438,0.2359596342,-0.0065541537,-0.1909570694,0.3453708291,-0.2321027815,-0.4061721563,0.3326815367,-0.0245837048,-0.0256828181,0.0126291486,-0.0888281986,-0.097761564,-0.2715517581,-0.0555886887,0.1748376638,0.2174036652,0.2924055159,-0.0828535631,-0.0749794692,0.1126661077,0.4408074617,-0.3611868024,-0.5222997665,0.4244419038,0.0466344878,-0.3514764011,-0.1904101968,-0.0677130073,0.2386674881,-0.0087470897,-0.4237614572,-0.1579840779,-0.3770035803,0.2211373299,0.1855123341,-0.0317657329,0.1636946499,0.0585967749,-0.1209819615,-0.3027124703,-0.079032056,0.1829794794,-0.2658408284,0.2607697248,-0.1361594796,0.2820146382,-0.0035267258,0.2051795721,0.4078882933,0.0132677527,0.0129668256,-0.0036062836,0.3086514175,-0.0445414297,-0.2472734749,-0.1389975995,0.0274222251,0.0640770644,0.2704855204,-0.0137382634,-0.1399645358,-0.0576543994,0.0702786967,-0.1378274858,-0.2034451514,0.0669690222,-0.0615939796,0.1262950003,0.1868982315,0.1167983338,-0.2275166363,-0.3898288906,0.0272282399,0.2789310813,-0.2099248469,-0.0508675314,-0.3776351213,0.0720062554,-0.0977065489,0.3176906407,0.1439626068,0.483145982,0.2175756246,-0.410409838,-0.009487967,-0.1612326801,0.129508391,-0.4809759259,0.3939151764,0.0182958171,0.1202988997,-0.4662635624,-0.1257723123,0.0240087956,-0.2370339334,0.5430121422,0.0031227709,-0.1567394584,0.0069204159,0.0016436375,0.3630622327,-0.1951063871,-0.2624641955,-0.4572242796,-0.042518191,-0.3031876385,-0.0849622712,-0.291883707,0.1151710823,-0.0952505842,0.1547553092,-0.0363609679,-0.0599315912,0.2802999318,0.1140879691,0.0919066221,0.0681542084,0.2160034776,-0.3673734069,0.1742977202,-0.0950673148,0.629080236,-0.0904105157,-0.4050771892,-0.0200800095,-0.1138672158,-0.2192772329,0.1024190858,0.0712338984,-0.0211342648,0.3180248141,-0.0292249769,-0.1321376413,-0.052964516,-0.0654845387,0.0568437427,-0.3134823143,-0.2322527617,0.2438934743,0.024928052,-0.0923280269,0.3096947968,-0.3169836104,0.0038187143,0.199173376,0.2315113544,0.8551769853,0.0946296602,0.0524023771,0.2603976429,-0.2381093353,0.0461070165,-0.2218386531,0.1222623661,-0.384721607,-0.1620908529,0.0296291765,-0.2397546917,0.1104680151,0.1308330894,0.0456580594,0.3157446086,0.0824448839,0.265435636,0.02678556,0.2626695633,-0.1982169598,-0.0077394708,-0.1783957332,0.2799905539,-0.1096340716,0.2839101255,-0.0245878119,-0.0027103839,-0.0680274516,-0.1203391477,-0.3087733388,0.3578498065,0.2986808419,0.349922508,0.3109762371,-0.0301482398,-0.4435335398,0.2308912873,0.3204562068,0.170921579,-0.2796056271,0.1042745709,0.3298264444,0.321156323,-0.0569303408,0.0360348597,0.2986278832,-0.034251824,-0.1673176438,-0.0113267414,-0.0020908199,-0.4932841063,-0.046410162,-0.0454819091,-0.3058390319,-0.2540874779,0.0605031624,0.0196043216,0.1603304148,-0.2026749253,0.1930448115,0.2711469233,-0.1588141173,0.2549833655,0.0608973913,-0.3740058839,-0.2073373795,0.5218612552,0.1486141384,0.0697739869,0.3076372743,0.0654314458,-0.2238656282,-0.2767473757,0.0103030903,0.0817589685,-0.2700168788,0.0866527557,-0.2723056376,0.1638420969,0.0175199956,0.4179403484,0.1980932504,0.2834116817,0.1301139593,-0.4619596601,-0.1199916974,-0.1962405741,0.2207478881,0.2335412651,-0.3102976978,0.0291037746,-0.0688461959,-0.0642739907,-0.4017803967,-0.1179599762,-0.3063415885,0.0454386547,-0.0378149562,-0.0823216736,0.0140701383,0.0197720118,0.2036683559,-0.1700205505,-0.3156564236,-0.287689209,0.0085890675,0.0834122673,0.0585625321,-0.1869750172,0.061514575,-0.0504394993,-0.0372771136,-0.1653628051,-0.0062807868,0.0559807383,-0.1245548651,0.2315860987,0.0302836727,-0.0545867793,0.1691436917,-0.1704310179,-0.4853298068,0.093788214,0.0261775684,0.0188592542,0.1663453728,0.0416227318,0.2209568322,0.1370637417,-0.0457316712,-0.0730069131,0.3050605059,-0.3060208559,-0.0913374722,0.3514891863,0.2890299857,0.2422015965,-0.3309823573,-0.0974548236,0.1171243861,0.3108786047,-0.3707127273,-0.0434040315,0.2625821531,0.1552303135,0.1360214204,0.0985777453,0.0700103045,-0.3690330684,0.5197038054,0.0253093373,0.1697219014,-0.4504546821,0.1651875675,0.3777585924,-0.0752394125,-0.1158111989,0.1755263507,-0.2276334316,-0.0030607823,-0.0224717893,-0.0467961952,0.3226110041,-0.1982214302,0.1586285979,-0.288612783,-0.4204319119,0.4601888359,0.2164718062,-0.1307115108,0.0211127158,0.300140053,0.2406678349,0.000314089,0.2337664515,-0.5237197876,0.2197773308,-0.1660951227,-0.1239840537,-0.1766879261,-0.1308230013,0.0225972366,0.325063765,-0.0398417972,0.0951620117,0.0081484392,0.1755941063,-0.1504278332,-0.6300577521,0.0377605297,-0.0301664323,-0.0750697181,-0.2521837652,0.1274860948,0.3072917461,-0.1701805741,0.0424116813,0.3488596976,0.3462336659,0.3489651382,0.2465882003,0.012970956,-0.0232694577,-0.0298771132,-0.3037691414,0.7039049864,0.0558110923,0.016437415,0.3652207255,0.1872573644,-0.1858240515,0.0569779016,-0.1207122281,-0.1957026869,-0.0281005148,0.5019013286,0.023771137,0.1474241465,-0.3337136209,0.2036661655,-0.3072110116,0.0872114077,0.1041135862,0.2170729339,0.1662150025,-0.2520652413,0.1290138662,-0.1272193193,0.465726465,0.1348124593,-0.0919346511,-0.0446526147,-0.3844293356,-0.5979169011,0.270507127,-0.2651817501,-0.0466285348,-0.0824073553,0.272816211,-0.0514353514,0.1950716227,0.173467949,-0.0461846851,-0.1869150251,0.062799789,-0.2228792459,0.0221348554,0.0849633291,-0.2045165747,-0.045315627,-0.2967478037,0.2486698776,0.0341974683,0.2029132247,-0.0903885141,-0.0434535965,-0.0509298854,0.2771998942,0.2398995757,0.3135384023,0.524189353,-0.0809628069,-0.1597727388,-0.24264054,-0.1934986264,-0.124801822,0.3427433372,-0.1332729161,0.2896223664,-0.0487711281,-0.1610382795,-0.26639691,0.2788797319,0.1420763284,-0.2243938744,-0.1259956211,0.0790791586,-0.2387904376,0.0993521586,-0.2077052146,-0.0864711031,0.1735779494,0.4444509447,-0.1318268031,-0.2226052582,0.4384515584,-0.0963414237,-0.2203106731,-0.2880280018,0.1975808591,-0.1742270738,-0.2328074574,-0.6331061721,0.0692528635,0.3489447832,0.1409208179,-0.0939427167,0.2052643746,-0.1768370569,0.2215215266,-0.0507070869,0.2084629238,0.1035976335,-0.3344599903,0.0540775098,-0.1685060263]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2553","title":"load_dataset(\"web_nlg\") NonMatchingChecksumError","comments":"This is fixed on `master` now :)\r\nWe'll do a new release soon !","body":"Hi! It seems the WebNLG dataset gives a NonMatchingChecksumError.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('web_nlg', name=\"release_v3.0_en\", split=\"dev\")\r\n```\r\n\r\nGives\r\n\r\n```\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/gitlab.com\/shimorina\/webnlg-dataset\/-\/archive\/master\/webnlg-dataset-master.zip']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: macOS-11.3.1-x86_64-i386-64bit\r\n- Python version: 3.9.4\r\n- PyArrow version: 3.0.0\r\n\r\nAlso tested on Linux, with python 3.6.8","comment_length":14,"text":"load_dataset(\"web_nlg\") NonMatchingChecksumError\nHi! It seems the WebNLG dataset gives a NonMatchingChecksumError.\r\n\r\n## Steps to reproduce the bug\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('web_nlg', name=\"release_v3.0_en\", split=\"dev\")\r\n```\r\n\r\nGives\r\n\r\n```\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/gitlab.com\/shimorina\/webnlg-dataset\/-\/archive\/master\/webnlg-dataset-master.zip']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: macOS-11.3.1-x86_64-i386-64bit\r\n- Python version: 3.9.4\r\n- PyArrow version: 3.0.0\r\n\r\nAlso tested on Linux, with python 3.6.8\nThis is fixed on `master` now :)\r\nWe'll do a new release soon !","embeddings":[-0.2213514149,0.1451930702,-0.1627065092,0.054831665,0.262732625,0.0005360552,0.1060018763,0.4832601845,0.3059208691,0.115994513,-0.0086998483,0.2747579515,-0.0059694932,0.0131163495,-0.2656615973,0.5328830481,0.0758582205,0.0331232883,-0.1119012684,-0.1202461645,-0.197989136,0.1313314438,-0.1261050403,-0.1625899374,0.066963762,0.2043554187,0.2423000187,0.2251994759,-0.1546619087,-0.25867787,0.3184999824,-0.0195882507,-0.0890438482,0.2732493579,-0.0001062257,0.168469131,0.3201108873,-0.0426909067,-0.3961649537,0.0392263085,-0.5058069825,-0.4818948209,-0.0048784083,-0.2138543725,-0.0109464861,0.2540812492,0.0780448988,-0.1685751379,0.1059200987,0.1463116407,0.2905976772,0.5583295226,0.015996702,0.1360504329,0.0916187167,-0.0865557566,-0.1386726946,0.193609342,0.0604513139,-0.1617763788,-0.3319349587,0.2184549868,-0.2171528339,0.2741572559,0.1065650284,-0.0845912918,0.2172036022,-0.1025966331,0.1710924059,0.3330104053,0.1714979559,-0.2114464492,-0.245823279,-0.1235723943,-0.1380158365,-0.0562530085,0.2332320511,0.3630893528,-0.255007863,-0.0771387815,-0.4136483967,0.2636136115,0.1233549789,0.1959425509,0.0515147485,0.2853710651,0.0701634064,-0.0449275188,0.0086996378,-0.2012542486,0.1703356206,-0.3862312436,-0.1421861798,0.0664267465,-0.5192736387,-0.1498858035,0.2591401041,0.5276124477,0.4039345682,0.4392689764,0.1958511025,0.2424522191,-0.1997988373,0.1211751103,-0.0106464066,0.1615688205,0.1949439794,-0.0839853734,0.3756630123,0.0440772437,-0.0146304313,0.2144354433,0.1387906373,-0.1356241256,0.5838980675,0.0139064137,0.1938016415,-0.1715454757,-0.2758587897,0.2424544096,0.0813562274,-0.1759569794,0.0936708301,0.4139363766,-0.1691817194,0.1787906438,-0.0826985389,0.1599846333,-0.178807959,-0.2603595257,-0.3013014495,-0.1948549002,-0.1006649733,-0.0532993115,0.1948911548,-0.0652758628,0.3730781376,0.1309435964,0.0521975048,-0.0292556919,0.1672415882,0.0672233477,-0.0797651708,0.3275880814,0.0650301203,0.0725562871,0.2623938918,0.051401563,-0.1442592293,0.0570121221,-0.1111410111,-0.1955203265,0.0478117727,0.2903845608,-0.4006389678,-0.1621252894,-0.0189837907,-0.3678647578,0.320248574,-0.0982220545,0.0617847629,-0.169508189,-0.0938759893,-0.2932614684,-0.0692824796,0.1751596183,0.2353035957,-0.0282806419,-0.1533160359,-0.1173754036,0.3033266068,0.3283869326,-0.1207039878,-0.1437745541,-0.2056747824,0.1444712877,0.528049171,-0.1462114751,-0.4569272399,0.1110413894,0.0208251979,0.2483598888,-0.106357947,0.2036237121,-0.415436089,0.011831414,-0.2405976951,0.2689469159,-0.0217602588,0.3086409271,-0.2725068033,-0.2499764562,0.5041952729,0.0884087756,-0.0178865604,-0.0473934673,0.1978051215,-0.2654822171,0.3275718987,-0.2987480164,-0.0708770305,0.0747108907,0.236616239,0.037259195,-0.0132464319,-0.0312267989,-0.5213804841,0.2340013385,-0.0987521559,0.0822667405,-0.0782284215,-0.0864103809,-0.1457846314,0.044871822,-0.1321758926,0.0444228537,0.1946576089,0.5132420063,0.2196192443,0.0085928477,0.0084085427,0.2146371901,-0.3017440736,0.0150596732,-0.4622909427,0.2809109986,-0.1104325578,-0.1225858033,0.0918966979,-0.0395374335,-0.0096824262,0.0663842708,-0.2127833664,0.3572992384,0.0566199087,-0.022123944,0.1243640482,0.2896420062,-0.1161517948,-0.2390986979,-0.0540458336,0.6289769411,0.0951028988,-0.08430361,-0.1892063171,0.527900517,-0.1324796379,0.157607168,-0.0149237029,0.1560308486,0.2574345171,-0.1585913599,-0.1889032125,-0.1386745572,0.3034152687,-0.1875821799,0.1542806476,0.1169057488,-0.2296345979,0.0263036713,0.4182861447,-0.1536764652,0.1969830096,0.1149320528,0.220950067,-0.2691132426,0.0047984105,0.1780553013,0.3732897043,0.1775916815,0.0537720434,0.0365419425,-0.2578305304,-0.2073852122,0.0649595112,0.0218336713,0.2422101945,0.2878108621,0.2809993625,-0.3075436056,-0.48106125,-0.1338565052,-0.0663479641,0.2497373968,-0.2675333321,-0.0650018379,-0.2904701233,-0.2531986833,-0.2250424922,-0.3509303331,-0.4928422272,-0.5039510131,0.0066800504,0.0738953128,-0.1908652037,0.3247650564,-0.4282859564,0.0368850827,0.1874407232,0.0657660812,-0.0499028414,-0.1186413467,-0.0775129274,0.0670184419,0.3970268965,-0.0233558044,0.3343366086,-0.3247566521,-0.0832288042,-0.3105060756,-0.2660367191,0.041972544,-0.1165441796,0.1404514313,0.2159852237,0.249858126,-0.014495871,-0.209147945,0.3455652297,-0.2148543894,-0.4010935128,0.302172929,-0.0083120791,-0.0596818626,-0.0119254813,-0.1110825241,-0.1173277497,-0.2770037353,-0.0859008729,0.2545133829,0.2168653607,0.2868855,-0.05115721,-0.1020708755,0.1370263994,0.4096435905,-0.3898355067,-0.576438427,0.4489720464,0.0224821512,-0.3417538404,-0.1682290733,-0.0730731934,0.2707675993,0.0129964016,-0.4342179,-0.1597047001,-0.3655693531,0.2646606266,0.1822001636,0.0081770094,0.2111605704,0.0925381109,-0.1488553286,-0.3051070571,-0.1053634137,0.1433532685,-0.2497019172,0.2870573997,-0.1694842279,0.2709267735,0.0160942469,0.3179975152,0.3712558448,0.0324144736,0.0738546327,-0.0057653524,0.2980764508,-0.0236303601,-0.2479379922,-0.1298051625,0.0220812578,0.0460745953,0.2868149877,-0.008304894,-0.1710108519,-0.0975011736,0.085702762,-0.1919400543,-0.2153317034,0.0455744267,-0.1077846736,0.1153173298,0.1890487671,0.1433933973,-0.247059375,-0.3662156165,0.0392427295,0.287127465,-0.2380149812,-0.056203004,-0.4046994746,0.0644035563,-0.1052755117,0.3195423186,0.1142057255,0.4199598134,0.2235459685,-0.4013907015,-0.0019008897,-0.20271267,0.1489788294,-0.4617826641,0.3531602025,-0.0173760168,0.0968110934,-0.3971095979,-0.134450376,0.0453021079,-0.1848848164,0.5334529281,0.0409584045,-0.2062513679,-0.0058939815,0.0148762092,0.3193853498,-0.1780484617,-0.2774283886,-0.4635246098,-0.0469762534,-0.323254019,-0.0442594253,-0.2452683151,0.1526942849,-0.0601358041,0.1457556188,-0.0485744327,-0.0626742095,0.2843716145,0.1673700809,0.1140093058,0.0326926783,0.2014360875,-0.3140478432,0.161115855,-0.0335639603,0.6571639776,-0.0789663866,-0.3968741894,-0.0594282225,-0.1136820689,-0.1899602711,0.0947522596,0.0630723611,-0.0398875587,0.3449781239,0.0091258641,-0.1479754746,-0.0185157117,-0.0736140311,0.0491680242,-0.3156715333,-0.2358543277,0.205253765,0.0286074392,-0.087707676,0.2927547395,-0.2949785888,-0.026159592,0.2215590179,0.2573698461,0.8447790146,0.0808877945,0.053253077,0.2461447716,-0.1737594754,0.0128081357,-0.2896065712,0.1390639544,-0.3688563704,-0.1840431243,0.027086271,-0.2233599573,0.0876484066,0.0893075094,0.0088732494,0.3345078528,0.0915966704,0.2711271048,0.0123324767,0.2327925414,-0.1573941857,-0.0007190611,-0.1688838899,0.2948543429,-0.0709469542,0.2966869175,-0.0605423413,-0.0058526886,-0.0807993338,-0.1554281116,-0.3041285276,0.3281774521,0.242676869,0.2952567637,0.2857362926,-0.0479134843,-0.4588710964,0.2498746514,0.2947556674,0.209729135,-0.2689047158,0.104209207,0.350551039,0.3174282312,-0.0829775929,0.0281413253,0.3137257993,-0.0467532054,-0.152598843,-0.0166690238,-0.0194943808,-0.5119175911,-0.0293534175,-0.0434403047,-0.280850023,-0.2628359497,0.0637885034,0.0297494549,0.1483173668,-0.2331930548,0.2072930783,0.286318332,-0.1718876511,0.2813049853,0.0962675512,-0.3578621745,-0.2276430279,0.5100736618,0.1462387294,0.0595569462,0.2607145011,0.0683530346,-0.2553168535,-0.3010512292,0.0156904776,0.0780968666,-0.2503837645,0.1001632884,-0.2535720468,0.1003591344,0.0221004263,0.4288523495,0.1977008581,0.2999122143,0.1475726366,-0.5127922893,-0.1690769196,-0.1684523672,0.186232537,0.2321443111,-0.2771873772,0.0338445604,-0.0854864642,-0.083433792,-0.4164291024,-0.1152668297,-0.3219229579,0.0577882752,-0.0735772476,-0.0984115526,0.045394931,0.0237490386,0.2299669087,-0.1661884934,-0.2737678587,-0.3000994325,-0.0238303095,0.0856792927,0.0386790037,-0.1903773695,0.0640197396,-0.0222180169,-0.0230914373,-0.2003702521,-0.0103000905,0.0182903521,-0.1077608317,0.2139082551,0.0741156712,-0.0631567016,0.1222865582,-0.1872828305,-0.4698032141,0.0751174986,0.0022571988,0.0265497267,0.1358854324,0.0007198068,0.2162117958,0.1389553249,-0.0231768955,-0.0353494175,0.3007337153,-0.2976224124,-0.0902649984,0.3910673559,0.3255714774,0.263328433,-0.3182221651,-0.0699483529,0.1326231062,0.328932941,-0.3979270458,-0.0367444009,0.209445104,0.1314918101,0.161520943,0.0928775966,0.0345215797,-0.3324862719,0.4629323781,0.0171230976,0.1609465778,-0.4062530398,0.1807287186,0.3925140798,-0.1083863899,-0.1036391482,0.1704111993,-0.2163061947,-0.0427648276,0.0175729003,-0.0615011603,0.3166530132,-0.1538115889,0.1167634204,-0.3164425492,-0.3681871295,0.4608348012,0.2053994834,-0.0783845782,-0.0043466715,0.3023123145,0.2112044841,0.0039352523,0.2162747532,-0.5057460666,0.2252803892,-0.1712272763,-0.1263666898,-0.1531042159,-0.166779846,0.021186443,0.3041287959,-0.0159998108,0.113070108,0.0289615747,0.1772446483,-0.1336572915,-0.619366467,0.0515482798,-0.0240657534,-0.0792232305,-0.2832977176,0.1691380888,0.2999796867,-0.1581955105,0.0213909112,0.3186531961,0.3698222041,0.3467496634,0.2153699994,-0.0122759687,-0.0485783555,-0.0554547459,-0.3482451141,0.6697833538,0.0664939582,0.0897683129,0.3467674255,0.2042520195,-0.2035214007,0.0894244388,-0.1190764382,-0.2130391896,-0.001450871,0.4948196113,0.0562143847,0.1364334673,-0.2968692183,0.1819120497,-0.2802018225,0.0695863143,0.0510594845,0.1938721687,0.185784623,-0.2667454779,0.1338080019,-0.0869624838,0.4629203379,0.125213936,-0.0601979606,-0.0631924495,-0.3704704642,-0.6255463362,0.2520496547,-0.2688539326,-0.039503742,-0.0776410401,0.2663390934,-0.0243722275,0.2224301249,0.1972408444,-0.0545728467,-0.1499534845,0.0546178818,-0.2404419482,0.0385767594,0.0833958611,-0.2028624713,-0.006862089,-0.3001235425,0.2398748249,-0.000336762,0.2237552702,-0.123647213,-0.0675257146,-0.0457931422,0.2612448633,0.2630409598,0.2758581638,0.5277030468,-0.1312941909,-0.1769468635,-0.2351822853,-0.1852657348,-0.1566816419,0.318138361,-0.1091941521,0.3288838863,-0.0704692602,-0.1761979908,-0.2555302382,0.2730384767,0.1290792972,-0.2011207789,-0.1244075969,0.0718516707,-0.2332198769,0.0586271584,-0.2094827145,-0.0636002347,0.1611866653,0.4360879362,-0.1482816637,-0.2269006521,0.4565494061,-0.1053558588,-0.2178788185,-0.282066524,0.1561088562,-0.1650542915,-0.2419304699,-0.6360208392,0.0911171436,0.3413097262,0.1483868808,-0.1569426209,0.2083608657,-0.1520657837,0.2139817327,-0.0452757776,0.2479707897,0.0994550362,-0.334982425,0.0493275449,-0.1779127568]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2552","title":"Keys should be unique error on code_search_net","comments":"Two questions:\r\n- with `datasets-cli env` we don't have any information on the dataset script version used. Should we give access to this somehow? Either as a note in the Error message or as an argument with the name of the dataset to `datasets-cli env`?\r\n- I don't really understand why the id is duplicated in the code of `code_search_net`, how can I debug this actually?","body":"## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n","comment_length":66,"text":"Keys should be unique error on code_search_net\n## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n\nTwo questions:\r\n- with `datasets-cli env` we don't have any information on the dataset script version used. Should we give access to this somehow? Either as a note in the Error message or as an argument with the name of the dataset to `datasets-cli env`?\r\n- I don't really understand why the id is duplicated in the code of `code_search_net`, how can I debug this actually?","embeddings":[-0.0094925556,0.004909059,-0.1037806347,0.3735046387,0.0469307639,-0.0595812909,0.2355328649,0.2376770377,0.0580709688,0.0420985892,-0.0894128904,0.4072095454,-0.1577007771,0.1043384299,0.1628589183,0.0670307279,0.0143380752,0.1504513174,0.1202021763,-0.1288990974,-0.3279083371,0.2217665762,-0.2497897297,0.0505501106,-0.2336093187,-0.1389334798,-0.2191094309,0.2946093976,-0.1718128622,-0.6023648381,0.2365141064,0.2469079047,-0.304754287,0.348526001,-0.0001129588,0.1454983503,0.3868984282,-0.0112753259,-0.2460678816,-0.1964260042,-0.1086823568,-0.1141097546,0.0937169865,-0.3164001703,0.018459348,-0.1501702219,-0.1131840348,-0.5023266077,0.876992166,0.2101984024,0.2571647763,0.230317384,0.3121251166,-0.2214271277,0.2857811153,-0.1449810416,-0.0408077687,0.4033092856,0.1780676097,-0.050453268,-0.001569794,0.2535579801,0.149145022,-0.0922281891,0.4889210761,0.004278875,0.0523612984,-0.3802721798,0.1833044142,0.0842173025,0.1399863809,-0.2937673628,-0.2534074783,-0.2852416635,-0.0006022927,-0.2383116484,0.3581800461,0.1130280718,-0.1832094043,0.1009692475,-0.0234330706,-0.2443882823,-0.0209400747,-0.0669795573,-0.0024758936,0.000144258,-0.0406801328,0.1741480231,0.1436638087,-0.030949913,0.2942076325,-0.133569628,0.1088615656,0.2196424156,-0.4874308705,0.1088525727,-0.0417375639,-0.0666240379,0.2526279986,0.0501061976,0.1586542875,-0.1383236498,0.0582781099,0.0365461148,0.136211589,0.1176389456,0.024738526,0.2395918816,0.1972186118,-0.2188272476,-0.2711265981,-0.0564664528,-0.17859146,-0.1516632885,0.4401807189,0.0809084475,0.5045513511,0.0516365357,-0.1957731843,0.1798712909,-0.1808783412,-0.1719259471,0.0739718005,0.3315834403,0.1826923192,-0.3608222306,0.0812146813,0.2304950655,-0.1570223868,-0.2011741549,-0.3156377375,-0.1495875716,-0.0772534907,-0.1267662793,0.2154261917,-0.5755309463,0.2917330861,0.2776085734,-0.0764898807,-0.4185130596,-0.0106018735,0.0661275685,-0.21158427,0.1040804014,-0.0396964923,0.0929643214,-0.0853247195,-0.2381432056,-0.0866200179,0.2018012553,-0.21553424,-0.3841583431,-0.0113813402,0.270018816,0.0460587405,0.1273243129,0.0172670651,0.0344303735,0.2237819284,0.1911880821,0.0374225043,-0.0336898118,-0.0244738683,-0.3202982843,0.23071374,0.6258887649,-0.1388694644,-0.0994992405,-0.166160807,-0.0337914713,0.0866958424,0.3992448449,-0.2378602475,0.426987797,-0.3161387146,0.2103259116,0.7068001032,-0.2062661797,-0.3250540495,0.1362059861,-0.2523133457,-0.1497699767,0.1035354063,0.026996633,0.1862990856,0.095646143,0.1519941688,-0.0751006007,0.0733930171,-0.0979357809,-0.384691447,-0.3182408512,0.1399685591,-0.0788120851,0.1862515062,-0.1908337772,0.2699958682,-0.1556688398,0.0582116097,-0.1312351972,-0.1205604523,0.0026754646,0.2714601457,0.0408043377,0.0887514651,-0.0086452868,-0.4823821485,0.374763608,-0.3360308111,-0.2697573602,-0.1959292889,-0.3034653664,-0.3515293598,0.1438819766,-0.5467948318,0.0142956581,0.2060532719,-0.0663145036,0.2171556205,0.1020635888,-0.176940009,0.1594810337,0.0542384125,0.1100345775,-0.2068196833,0.2509876788,-0.473295927,-0.0428255834,0.0588917732,0.2304445505,0.3486605883,-0.126670599,-0.1688676924,0.444019258,-0.2193270922,0.0511581339,-0.0163343493,-0.0888499841,0.1398420334,-0.3815078437,0.2165362686,0.2500270009,0.1734208018,-0.1731620431,-0.0787188336,0.3603438139,-0.111679703,0.0948637575,-0.0413667299,-0.0870466456,0.0792675242,-0.1197792962,0.1631211042,-0.1041850671,0.2782212496,-0.3106930554,0.3890615702,0.1747398674,-0.2443796545,0.1206956282,0.2313815206,-0.040806178,0.1052942649,0.0120361065,0.032384526,-0.191202417,-0.2877425849,0.2358517349,0.3477755487,0.154177025,-0.0584747009,0.2757799625,-0.0113031315,-0.2352381796,0.3930318356,-0.1218933836,-0.1429179609,0.3976445794,0.1085420996,0.0489772335,-0.3879248798,-0.1217015237,0.1714469045,0.0767205209,-0.2829786837,0.0125734387,-0.1297802329,-0.0831947252,-0.456538856,0.5397939086,-0.1158734262,-0.1127381995,0.252034843,0.0170565713,0.0975915939,0.1204392239,-0.2307248563,0.1806775331,-0.0148847988,0.1146529019,-0.2355740368,-0.1124270409,-0.077190429,0.03742823,0.3936970532,-0.0605512783,0.5720286965,-0.1889272928,0.1064837798,-0.1775661409,-0.3476791382,-0.0131592071,-0.0041039702,0.1664582789,0.0505314544,-0.1058803722,-0.035335023,-0.378918618,0.2900122404,-0.3102478683,-0.4068109095,-0.0542965159,0.0398769937,-0.0646373257,-0.2149248123,-0.5273548961,-0.265047133,-0.2430540472,0.3169803917,-0.0962151885,0.1279311031,0.3379567564,-0.1227716357,-0.0101920851,-0.0059263171,0.2016517222,-0.3594234586,0.0480577499,0.0874655396,-0.0509008281,-0.353307426,-0.0688932985,-0.1448617578,0.1514542401,-0.0905499533,-0.3825456202,0.2584376633,-0.231079936,0.4560710192,-0.1607944965,0.3305143416,0.2242964655,0.2599559724,-0.1505051553,-0.185819298,-0.0531722978,0.1141134128,-0.125399828,0.0685934052,0.1404928416,0.2881910503,-0.3000931144,0.8568351269,0.6044374108,0.0024756135,0.2031872422,-0.173413977,0.3541770577,-0.2369024009,-0.4189771414,-0.2487502396,0.0740314499,0.024824189,0.1166035682,-0.074014008,-0.0949571133,0.1337455511,-0.1691589952,-0.6246013641,-0.3106158376,0.0508954749,-0.2232756764,0.1503931284,0.0552002974,0.0079762815,0.0529072955,0.0654388741,-0.0247862693,0.2386342138,0.0278170221,0.0320964567,-0.1855977625,-0.1713015288,-0.3948457539,0.108729355,-0.1418867707,0.1687642634,-0.0074779456,0.1403222829,0.4162751138,-0.1948324144,0.6553739309,-0.082054466,0.118701376,0.113919422,0.1600474268,-0.0918410644,-0.1216057166,-0.0004836839,0.4200724661,0.1214304119,0.0828676298,-0.286349386,0.0328855701,-0.0618459992,0.2935726047,-0.3494208753,-0.418186903,-0.1062088385,-0.3285907507,-0.5123425722,-0.0003708878,-0.0332548544,0.2004807144,-0.0638975054,-0.0373939686,-0.1862738729,-0.2158709317,0.0450355783,0.2135528028,0.2768926919,0.054016538,0.5974946022,0.439026773,-0.1527076066,0.0138948085,0.7316986322,-0.0312716179,0.0088354507,-0.3602286279,-0.214563936,-0.1516319811,0.3003151417,-0.1722510159,0.1454564035,0.064309679,-0.005578598,0.0442820303,-0.2298399806,0.1667325944,-0.2030818909,-0.1967425197,-0.3953807354,0.2921684086,-0.1211480498,-0.3467521667,0.2700574696,0.1420119554,-0.191045925,0.3929702938,-0.2961888909,0.9225983024,-0.2770524025,-0.0354177915,-0.1204794198,-0.3473197818,0.2588879466,-0.0852122381,-0.0664185956,-0.5006073117,0.0952881575,0.1400972009,-0.0975073576,-0.0751616433,-0.0089118378,-0.3079712093,0.1783438027,-0.0532489605,0.0315936655,-0.1495840997,0.3273097873,-0.0030509271,0.2265324146,-0.0421694145,0.1000897214,0.1819884777,0.2940016985,-0.0690353364,-0.0146773607,0.1485506594,-0.578879714,-0.2620584369,0.0455853231,-0.3623944521,0.491085887,0.0204073694,-0.2879943848,0.2039721161,0.0256681852,0.2825469375,0.155127123,-0.0330335163,0.0975018218,0.3119630814,0.1528388113,-0.1907716542,-0.0541854128,-0.0668792501,-0.0191121455,-0.3339544833,-0.0638494417,-0.098152943,-0.385492444,-0.1896460503,0.229513064,-0.2025195658,-0.0292345583,0.1831748486,-0.0261086021,-0.1139790416,-0.2872944772,0.1506655663,0.1937769502,0.0281574912,0.1674936414,-0.1279280186,-0.1594589949,0.0427719019,0.1584786326,0.3779702187,0.0691800192,0.4716702104,-0.2742197216,-0.0576696806,-0.2229479998,0.0907727182,0.0751189739,-0.6839030981,0.2268487066,-0.3302604854,0.2216809392,0.0492409877,0.445679307,0.372816205,0.264169544,0.0447097048,-0.1884460151,-0.1231536418,-0.024940785,0.0557023399,0.2507334054,0.03276366,0.1545563042,-0.0578995384,0.2285488248,-0.3535400927,0.2854165435,-0.0628061518,0.1187647283,-0.0041630487,-0.1193921119,-0.1067295671,-0.0584335364,0.163981691,0.0769250542,-0.2023906559,-0.2468737364,-0.0515812077,0.1313583255,-0.005643406,0.0820989683,-0.2349615246,-0.2209154963,0.2747095823,-0.0153558794,0.2010286897,-0.1215077341,0.1296389848,0.037145555,0.1333367974,0.127461493,-0.2856585681,0.2258189768,-0.2711217105,0.3647975028,-0.0204248093,0.1812969297,-0.0559413582,-0.0842082277,-0.3009376228,0.0659644604,-0.1919424534,0.1232486293,0.1786118001,-0.2007998526,-0.1515879333,0.3735954463,0.1895248443,0.5776078701,-0.3747799993,0.1233211383,0.0016047447,0.2342489511,-0.3270578086,-0.1232975349,0.0141037293,0.0219933055,0.2465510368,0.4026124179,0.3325186074,-0.2041254789,-0.1000590175,0.0773329064,0.4145536423,0.0431953408,0.3692278564,0.4215740561,0.2249730229,-0.1458379626,0.3178175092,-0.008857199,0.0532550141,0.5327891707,-0.5467882752,0.1335881203,0.0073111695,0.248273313,0.2667936981,-0.1006904468,0.4065020382,0.0637443438,-0.0616154112,0.0232217666,0.0591041371,0.1641822904,-0.1535845995,-0.0986169055,-0.0675118864,0.0940413177,-0.1425421685,-0.1154043153,-0.1221371591,-0.2859399915,-0.2399975806,0.1165320724,-0.0442254096,0.0304312427,0.1857380867,-0.0055708699,-0.320953548,-0.395937562,0.2486814857,0.2477921247,0.360360086,-0.3663314879,0.3189526796,0.5262146592,-0.0918747112,0.1858082861,0.5166122317,0.67971766,0.2297296375,-0.1870812625,0.2632651627,0.1030771509,-0.1497942358,-0.1714255959,0.1047997922,-0.0949581712,0.1528459638,0.1419484019,0.0963760242,-0.1254532933,-0.0555638261,0.2238544077,0.0978103653,-0.0999177396,0.2334596664,-0.1785313487,-0.1811484396,0.0959306508,0.1979930699,-0.3587418497,-0.0058027813,0.3178543448,0.2464194894,0.1359600723,-0.1989139616,0.13014099,0.0504276007,0.3355688155,0.2014838904,0.0995328203,-0.2653712034,-0.0299465042,-0.7278295755,0.0578824542,-0.4435489178,-0.0309019033,0.2030548155,0.3628490865,-0.2993817031,0.0497820824,0.371684283,0.0444039591,0.0148728387,0.2185641229,-0.170569852,-0.0004783668,0.4338772297,-0.2329913676,0.0971011743,-0.3571490049,0.3673857749,-0.3747152686,0.0223204643,-0.1711813807,0.0642180368,0.1893087924,-0.2624619603,0.4224396646,0.1264918,0.3507632911,0.0014028184,-0.1700942367,-0.203534171,-0.112419337,-0.2247008085,0.2866600752,0.4753935933,0.1072760671,-0.0141486097,-0.4560535848,-0.1325496882,0.2371395379,0.0571238138,-0.0928513706,-0.0732209608,0.0577744097,-0.1940234005,0.114231497,0.0330923498,0.2063193917,-0.1023555845,0.1780995578,-0.2683041394,-0.6693448424,0.4163756669,-0.1604196578,-0.0692104027,-0.1580542177,0.3268049657,0.2550383508,0.1188758686,-0.6650725603,0.1800191253,0.557053268,-0.2718786597,-0.0380278938,0.1505858749,-0.1034700722,-0.023336485,-0.008343,-0.0389363542,-0.0169709083,-0.2956866324,0.2166404873,-0.3160583079]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2552","title":"Keys should be unique error on code_search_net","comments":"Thanks for reporting. There was indeed an issue with the keys. The key was the addition of the file id and row id, which resulted in collisions. I just opened a PR to fix this at https:\/\/github.com\/huggingface\/datasets\/pull\/2555\r\n\r\nTo help users debug this kind of errors we could try to show a message like this\r\n```python\r\nDuplicateKeysError: both 42th and 1337th examples have the same keys `48`.\r\nPlease fix the dataset script at \r\n```\r\n\r\nThis way users who what to look for if they want to debug this issue. I opened an issue to track this: https:\/\/github.com\/huggingface\/datasets\/issues\/2556","body":"## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n","comment_length":97,"text":"Keys should be unique error on code_search_net\n## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n\nThanks for reporting. There was indeed an issue with the keys. The key was the addition of the file id and row id, which resulted in collisions. I just opened a PR to fix this at https:\/\/github.com\/huggingface\/datasets\/pull\/2555\r\n\r\nTo help users debug this kind of errors we could try to show a message like this\r\n```python\r\nDuplicateKeysError: both 42th and 1337th examples have the same keys `48`.\r\nPlease fix the dataset script at \r\n```\r\n\r\nThis way users who what to look for if they want to debug this issue. I opened an issue to track this: https:\/\/github.com\/huggingface\/datasets\/issues\/2556","embeddings":[-0.039390035,-0.0147919394,-0.092823565,0.3366979957,0.0873194709,-0.0139394747,0.2003903836,0.2712210715,0.0882736444,0.0667391494,-0.0861993954,0.3934495449,-0.1623921543,0.1329388171,0.1347277313,0.0915056542,-0.0428930745,0.1898689717,0.118189007,-0.056120418,-0.3183745444,0.2934969664,-0.222816214,0.0328287929,-0.2074050754,-0.1297896057,-0.1850280464,0.2394023091,-0.169512704,-0.6275754571,0.2506465316,0.2204239964,-0.3045646548,0.3702608347,-0.0001118734,0.1776437461,0.3890072703,-0.0119130826,-0.263148427,-0.1894607991,-0.0932743922,-0.0726644769,0.0986738652,-0.2597336173,-0.0612497963,-0.1425023079,-0.1062145233,-0.5259363651,0.8551874161,0.2249429375,0.261464715,0.1924865991,0.3430315554,-0.2581971586,0.2435848862,-0.1772041172,-0.0567812659,0.3766361773,0.2065399587,-0.0528299324,0.0148361363,0.2883768678,0.1508664638,-0.1099837124,0.4550983906,-0.0461321585,0.0562952235,-0.3445614576,0.202942282,0.0850430951,0.0933940634,-0.2972278595,-0.211820215,-0.2630234361,0.0436524637,-0.2451478392,0.3421130776,0.1394786388,-0.2177512348,0.1649510264,-0.0173711162,-0.258102119,0.004423562,-0.0469903834,-0.0124232657,-0.0174261965,-0.0472720303,0.1445003897,0.1904307753,-0.0294545386,0.2612523139,-0.0770370811,0.0319689102,0.1867050976,-0.5232821107,0.1069345474,-0.0528017282,-0.1080394983,0.2210678756,0.04155799,0.1781447381,-0.0987889767,0.0630774796,0.0046159402,0.1186117828,0.1113321632,0.0510718673,0.2479469478,0.2089866847,-0.1799351871,-0.2521626651,-0.0465332828,-0.2012174428,-0.1685853302,0.370346278,0.0584278665,0.5287957191,0.0775724277,-0.2316820472,0.1241981462,-0.1455936879,-0.1006391644,0.0610974915,0.3620432913,0.1859445423,-0.3567807674,0.1096722409,0.2256824821,-0.1423757821,-0.194125846,-0.3171714544,-0.106359385,-0.0811322033,-0.156812489,0.2419344336,-0.6133795381,0.2899116576,0.2416549325,-0.1092535332,-0.4836315811,0.0230332892,0.1235139221,-0.1988665313,0.0956891775,0.010164707,0.0778541639,-0.0203281436,-0.238665089,-0.0512514226,0.1840789467,-0.2536869943,-0.3633141816,-0.0695421621,0.2863248885,0.0670304596,0.0897065252,0.0178208277,0.036976248,0.2047014832,0.1793242544,0.0022217699,-0.0245714337,-0.0792566091,-0.2655107379,0.2568253875,0.635163486,-0.0922877565,-0.1067955345,-0.1737157553,-0.0022584435,0.0813440755,0.3821578324,-0.2429331839,0.3917122483,-0.3513783813,0.1783625484,0.713447988,-0.1958878785,-0.3543924987,0.1866781712,-0.2593115568,-0.218294844,0.0873140693,0.0251206383,0.1783084273,0.1153895631,0.157443732,-0.0760309249,0.0586169437,-0.0519526079,-0.3551417589,-0.3531219363,0.1054895669,-0.0788374171,0.1729315072,-0.1593489647,0.2709896266,-0.1526414305,0.1180687621,-0.1213864759,-0.0849593729,0.0823330656,0.2390115112,0.0496873073,0.0824484378,-0.0632929653,-0.4461555481,0.3696315587,-0.2972278595,-0.266858995,-0.239024654,-0.2505201697,-0.4130327106,0.1440363824,-0.5862923265,-0.0180854406,0.2047824413,-0.0346153788,0.2092174143,0.1034080163,-0.1269436926,0.1520022005,0.0593991727,0.0834335387,-0.1985911578,0.2522063851,-0.4453632534,-0.0550451502,0.0948074311,0.2252966017,0.325219512,-0.0994668528,-0.1627309918,0.417509675,-0.2410899103,0.0180140231,-0.0007237954,-0.1209466383,0.0787326172,-0.3930574059,0.1947408468,0.2712776065,0.178377375,-0.1665648371,-0.0573369339,0.3320143819,-0.0891046971,0.0955382586,-0.0527458787,-0.0315741859,0.1117780209,-0.1201968044,0.2202439606,-0.1167671606,0.2911606431,-0.3164441288,0.3656363487,0.1319552809,-0.2356605232,0.1314296573,0.3060259521,0.0220883489,0.1882434636,0.0140608279,0.0087792343,-0.155099526,-0.3161092103,0.2443338484,0.3074198365,0.1628214121,-0.1048359126,0.2712975144,0.0289769229,-0.2005146146,0.3751117587,-0.1769654602,-0.1046681628,0.3829626143,0.1048186272,0.0200045686,-0.4195503891,-0.1060770303,0.166590333,0.0904481038,-0.2432655543,0.024589492,-0.1573229134,-0.0550038442,-0.4926802218,0.5171210766,-0.1103608385,-0.1431573182,0.2413244843,0.0723625869,0.1362358183,0.1244721189,-0.2315515727,0.1443307847,0.0402568541,0.0933430716,-0.208307907,-0.1138941646,-0.0725762397,0.0338538177,0.37839064,-0.0770852938,0.5647823811,-0.1624939442,0.1164632887,-0.2422077358,-0.3452756703,-0.017885197,0.0072638318,0.1485368758,0.0805482119,-0.0375982709,-0.0214283541,-0.3726305664,0.2817128599,-0.3179251552,-0.4135041833,-0.0009355024,0.0158012602,-0.0357335135,-0.2129419148,-0.5386603475,-0.2469630688,-0.266546309,0.3176681697,-0.1062420309,0.0821354836,0.3633770347,-0.0593880489,-0.0168305058,-0.0493381135,0.1703391075,-0.3394080698,0.045024801,0.046050258,-0.0631044284,-0.3325830698,-0.0430321284,-0.1231541857,0.1562623382,-0.1182264686,-0.3490812778,0.2830144465,-0.2649712265,0.4167036414,-0.2116276324,0.2905671299,0.2058410645,0.256126225,-0.1482191086,-0.1712091118,-0.02909109,0.1058599502,-0.171593979,0.0603614785,0.0947430432,0.2958513498,-0.2571831346,0.8182141185,0.5736106038,-0.0000046855,0.1971295476,-0.2320547104,0.3561851382,-0.2597362697,-0.4400476217,-0.2620584071,0.1012614444,0.0823969543,0.101261504,-0.089070037,-0.1051109284,0.1480726302,-0.1535298675,-0.5878965855,-0.3327074647,0.0677499026,-0.2154907286,0.1211415082,0.0649843216,0.0236835554,0.0236726534,0.0156503525,-0.0114403795,0.224405542,-0.0429973938,0.0903418586,-0.1426973492,-0.1601071805,-0.3927232921,0.1269591451,-0.1455414593,0.1879532188,-0.0046167173,0.1819086671,0.3899122477,-0.2039209902,0.6486160159,-0.0842430964,0.1471614242,0.116316177,0.1549592018,-0.105467163,-0.1031057835,0.0434398092,0.4334650636,0.1104534641,0.1101810709,-0.2643679082,0.0120887654,-0.0417419448,0.2710542083,-0.35629493,-0.3989876807,-0.1732003987,-0.3362305164,-0.5221696496,0.0070487121,0.0132363122,0.2334473729,-0.0566029213,-0.0964119136,-0.2230377495,-0.2034578025,0.0082272487,0.201933682,0.2648091018,0.1155600846,0.5737646818,0.3952605128,-0.1276341677,-0.0011024169,0.7134286165,0.0046015871,-0.0141345151,-0.301850915,-0.2454432845,-0.1719881892,0.2688542604,-0.1816078871,0.1681399643,0.0785085186,0.021937253,0.0292041432,-0.1947407871,0.1829778254,-0.1791462153,-0.2184189856,-0.3810112774,0.2849347889,-0.1403224021,-0.2812668383,0.2780410945,0.1590031236,-0.1447203308,0.3344264925,-0.3454682231,0.8578016758,-0.3358392119,-0.0549023189,-0.1418085247,-0.3329260349,0.2424285263,-0.0555238351,0.0201000795,-0.5142740607,0.0449112058,0.1704356521,-0.0552185737,-0.0848077536,-0.0177540313,-0.3104019463,0.1720506102,-0.047649432,0.0248287395,-0.1602256745,0.345887363,-0.0029213047,0.2451022416,-0.0235070344,0.1250651926,0.151576966,0.2819944024,-0.0366752967,-0.0448581986,0.1928787827,-0.5627144575,-0.274626106,0.1214069277,-0.3507395089,0.5328578949,0.0126898605,-0.258488059,0.2454972565,0.0261310376,0.2486754656,0.1919755936,-0.0656968355,0.1184825376,0.286691308,0.0737583637,-0.2213355452,-0.0155502837,-0.0890037715,-0.0279262215,-0.3782758117,-0.0527391694,-0.0517929047,-0.4121177197,-0.1450531781,0.2573261857,-0.1829818338,-0.039404165,0.1902216077,-0.0515704639,-0.1065126359,-0.31052953,0.1581815779,0.1754858643,0.0420293063,0.0981867686,-0.0494146273,-0.1876265705,0.0787870809,0.1677099317,0.3655773699,0.0784647241,0.5398337841,-0.2415946871,-0.0110983588,-0.2637151778,0.0708359331,0.0914916396,-0.689343214,0.2305616885,-0.2692235708,0.2263799906,0.1349015087,0.4542607963,0.3118366897,0.2074229121,0.0682095513,-0.1899862587,-0.0834034309,-0.0749873519,0.0079053314,0.2383566052,0.017030362,0.1854347438,-0.0320163853,0.213806361,-0.3605460823,0.2981470823,-0.0902474299,0.0879207402,-0.0235433746,-0.1306426823,-0.035133481,-0.0144165214,0.1685692519,0.100668475,-0.2077368498,-0.2502122223,-0.0640063658,0.124471277,-0.042442333,0.1501917988,-0.2243754566,-0.2086766213,0.2310709804,-0.0412695631,0.2202374637,-0.1125026643,0.1427197754,0.0607993044,0.1483040899,0.1030613482,-0.2411946058,0.191882506,-0.2597800493,0.3232286572,-0.0608832687,0.1569918543,-0.0297607277,-0.0701464266,-0.3087400198,0.0846204162,-0.1682215929,0.09435606,0.154928118,-0.2569451034,-0.1614758819,0.4119190276,0.1723029315,0.5156676173,-0.3425476551,0.1220132783,-0.03145805,0.2493457794,-0.3464598358,-0.1409687847,0.0218816157,0.0230217315,0.2520653903,0.3462911844,0.2895456851,-0.2067204565,-0.0891882926,0.0885485932,0.4030732214,-0.0098054567,0.3582567871,0.3623288274,0.2651084065,-0.2178713977,0.3288599849,-0.0276891086,0.0573129803,0.536087811,-0.5569239259,0.1019916385,-0.0197682995,0.2291837931,0.214448154,-0.1321156323,0.3921252191,0.0536373816,-0.107160978,-0.0169932302,-0.0008155518,0.1937872618,-0.185079962,-0.0418528803,-0.1078157425,0.0945615098,-0.1713007838,-0.0954019725,-0.1070974544,-0.2283350825,-0.2598010898,0.1113855988,0.0088822497,0.065078862,0.1844663173,0.0013827501,-0.3270095289,-0.3965142667,0.2018178254,0.2869683504,0.3742625117,-0.4109542668,0.2661156356,0.4641633332,-0.0921058059,0.1796158105,0.5124383569,0.6277316809,0.2143345624,-0.1546014249,0.2471620142,0.0746892169,-0.1664489806,-0.177601546,0.078996323,-0.096925348,0.1750383377,0.1689578891,0.1014591753,-0.1353826225,-0.1017671078,0.263949424,0.1437013149,-0.105798617,0.1999002099,-0.1909977943,-0.1376964748,-0.0063627157,0.2526304424,-0.374394387,0.0232564043,0.336202234,0.2152958661,0.1036928222,-0.2392004728,0.1387794167,0.046992898,0.3233960271,0.1746049374,0.0572601594,-0.2274860889,-0.0484919809,-0.6852201223,0.1117905155,-0.4422994554,-0.0206722263,0.1845692992,0.3831756711,-0.3131615818,0.0189342294,0.3542641997,0.0676847845,0.0086913286,0.2228708118,-0.215997234,0.0082057081,0.4384151697,-0.2551456392,0.1303380132,-0.402612865,0.3538030684,-0.376016736,0.0528254993,-0.122826077,0.0515918918,0.1246993691,-0.2892438471,0.4157178998,0.097764723,0.3264960349,-0.0135341734,-0.2033665776,-0.2298591286,-0.1841998547,-0.2236932218,0.2773218453,0.5202602744,0.1267429739,-0.0173211899,-0.427680105,-0.1936795115,0.2420183271,0.0435462184,-0.0591895133,-0.0666078553,0.0637033135,-0.1687977463,0.152901724,0.0310584754,0.2029289752,-0.083538942,0.1606072634,-0.2491607368,-0.6767690778,0.394970119,-0.1588925868,-0.1128195971,-0.1074142233,0.3577104509,0.2368585914,0.0830922723,-0.6305253506,0.2515726388,0.5428174734,-0.2058785409,-0.0201533772,0.1489615291,-0.1389390975,0.0323462822,0.0318879709,-0.0470930375,0.0382376686,-0.2773482502,0.2415047884,-0.3176975548]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2552","title":"Keys should be unique error on code_search_net","comments":"and are we sure there are not a lot of datasets which are now broken with this change?","body":"## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n","comment_length":18,"text":"Keys should be unique error on code_search_net\n## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n\nand are we sure there are not a lot of datasets which are now broken with this change?","embeddings":[-0.0469085425,-0.0271531846,-0.1106117964,0.3490996659,0.081276536,-0.0156195574,0.1811806709,0.2770838141,0.0815910995,0.0544187576,-0.0599072017,0.3916929364,-0.1846611351,0.116615966,0.1564046592,0.078515403,-0.0312846079,0.1728187054,0.1106826067,-0.0542455949,-0.3256242573,0.2887380123,-0.2328311056,0.0389522165,-0.1848244071,-0.1620794386,-0.1911866963,0.2588772476,-0.1845095456,-0.6111863256,0.2418148369,0.2137483656,-0.3018578887,0.3558101058,-0.0001112483,0.1782946587,0.3963311613,-0.0183047689,-0.2428982258,-0.1785934716,-0.0976837426,-0.0869019106,0.0782905892,-0.2927312851,-0.0522992276,-0.1311941594,-0.1020006314,-0.5351581573,0.8553102612,0.2324268073,0.2680365145,0.1950211674,0.3377575576,-0.235691309,0.2538096011,-0.1840531081,-0.0478936695,0.3911312521,0.2176291645,-0.0469379611,-0.0018590186,0.3106984198,0.1507784575,-0.1226941347,0.4385913908,-0.0377838165,0.0561736748,-0.3310019672,0.1806346625,0.0725954771,0.1040767357,-0.3062256277,-0.2338497192,-0.2663578689,0.0390654281,-0.2513284683,0.3439498246,0.1335716099,-0.2220121175,0.1639823616,-0.0111537613,-0.2648167908,0.0115648899,-0.0475299433,-0.0154275643,0.0178863611,-0.0589888878,0.1368977427,0.160793066,-0.0466827862,0.2724993527,-0.0896123201,0.0467299111,0.1941236407,-0.5073847175,0.1113422439,-0.0263008662,-0.0870054588,0.2304593474,0.0269335285,0.1648990214,-0.1091128364,0.0596225448,0.0060264375,0.1255843341,0.1220989898,0.0224412084,0.244217068,0.2085755616,-0.1591237634,-0.2433698773,-0.0350925513,-0.1751270741,-0.191812858,0.3740758002,0.0837449282,0.5239519477,0.0817943066,-0.2396693677,0.13707757,-0.1299029738,-0.1012648121,0.0700322166,0.3665414751,0.160808742,-0.3450777531,0.116074048,0.185338378,-0.1395103931,-0.1890879124,-0.3252248168,-0.0816784799,-0.0858611017,-0.1678928435,0.2184514701,-0.5809813142,0.2865787446,0.2305636853,-0.0979512781,-0.475730598,0.0076656276,0.1049675941,-0.2080143392,0.0840548128,0.0011776964,0.059263289,-0.023182882,-0.2230984271,-0.0681596249,0.1733560413,-0.2561249435,-0.3439769447,-0.0812844634,0.2983009517,0.0687769279,0.0940390453,0.0188799016,0.0360048451,0.2115163952,0.1689544171,0.019319294,-0.0201480016,-0.0689401329,-0.2622361481,0.2715479732,0.6408420801,-0.0852002129,-0.1143703461,-0.1630569994,-0.0119723873,0.0820221528,0.3735702038,-0.2195251137,0.4098963439,-0.3701931834,0.1702483743,0.7158523202,-0.2098324448,-0.3514036536,0.1768588573,-0.2566369176,-0.1927757412,0.1062855944,0.017745195,0.171772033,0.1271534413,0.191995427,-0.074500829,0.0496413931,-0.0505873188,-0.3718301654,-0.3530538976,0.1030012369,-0.0865597725,0.1955659837,-0.1717545241,0.2626270652,-0.1302562356,0.1170928478,-0.1145149097,-0.0960906073,0.08498054,0.2491381764,0.0279336032,0.0814496204,-0.0725878999,-0.4511611462,0.3614913523,-0.2994681001,-0.2637138069,-0.2208357751,-0.2707116306,-0.3951714635,0.1444454044,-0.5617762804,-0.021909453,0.2265327275,-0.0335865505,0.2018054277,0.089362435,-0.1509428024,0.1715616882,0.0490317941,0.0780372918,-0.2100775987,0.2596832216,-0.4771119952,-0.0632256791,0.0932601988,0.2274794132,0.3199900389,-0.1160648391,-0.1683141291,0.4271498024,-0.2329930514,0.0158185381,-0.0029284244,-0.1300105751,0.0871503279,-0.3934369981,0.1725026667,0.2751096785,0.1673909575,-0.1346253902,-0.0697996989,0.3326274157,-0.098880671,0.0816605836,-0.0342422314,-0.0534674115,0.1030966714,-0.1120853573,0.2092016786,-0.1217363179,0.280827105,-0.3323273659,0.395688206,0.1194216758,-0.2414901406,0.1092133597,0.2924526334,0.0013845633,0.1975161284,0.0097058034,0.0288969744,-0.1784740537,-0.2953672111,0.2325143069,0.2991614342,0.1731813103,-0.108634688,0.2740420699,0.0366743244,-0.1799518317,0.3811586499,-0.1649849117,-0.1044396386,0.39415434,0.081823498,0.0342088789,-0.4060649276,-0.1442296058,0.1447810829,0.10745617,-0.2713996768,0.0040042065,-0.1440774053,-0.0816541612,-0.4855396152,0.5519646406,-0.0967328995,-0.1340057999,0.2336675078,0.0417370275,0.1260758787,0.1139730886,-0.2359859347,0.1488276422,0.0432489254,0.1060621962,-0.2104985863,-0.104381077,-0.0720650554,0.0447643474,0.3772628605,-0.0930300429,0.5541758537,-0.1986333728,0.0996911302,-0.2164834887,-0.3579094112,-0.0040377402,0.0208136421,0.149884209,0.1100701541,-0.0402450822,-0.0256814659,-0.3755349815,0.2560067475,-0.3044043183,-0.4428147972,-0.0225434508,0.0187568497,-0.0395334139,-0.2074314505,-0.5418492556,-0.2662756741,-0.2766748965,0.292663157,-0.0981774926,0.0814607516,0.3725579083,-0.0679178238,0.0038487245,-0.0269739237,0.1974369437,-0.3267024755,0.0312487204,0.0353545435,-0.0529341437,-0.3246124983,-0.057834886,-0.1475159973,0.173725903,-0.1259884685,-0.3620150685,0.2623095214,-0.2389128059,0.426699996,-0.2036780864,0.3028892279,0.2129474282,0.2573758662,-0.1596046388,-0.1698618531,-0.0314722806,0.0892744213,-0.167788744,0.0556728058,0.1018413827,0.304336071,-0.2619749606,0.8216831684,0.5829095244,-0.0059695388,0.2016521543,-0.2118443847,0.3733597398,-0.2440511435,-0.4355399013,-0.2459855527,0.1285446435,0.0407847725,0.0970505998,-0.0921052098,-0.0974657461,0.1396706849,-0.1615277827,-0.6219164729,-0.3212406933,0.0509340763,-0.2273284495,0.1179317012,0.0508584827,0.008976738,0.033571668,0.0444792248,0.0025212166,0.2345269322,-0.0114630517,0.0759303421,-0.1530279964,-0.1533860415,-0.3725166023,0.1175916344,-0.1397171915,0.193103835,-0.0105235409,0.1913721263,0.3669941723,-0.2281316966,0.6453694105,-0.0840140283,0.1381125599,0.1138134673,0.1591111571,-0.0830507576,-0.1050435305,0.0388189703,0.4268300831,0.1269040108,0.1294588596,-0.272939086,0.0242649168,-0.0232013259,0.2627456188,-0.3573333919,-0.3812642992,-0.1662815511,-0.324618578,-0.4989025891,0.0290859845,0.0064901868,0.2311416268,-0.0217208341,-0.0889583081,-0.2271213084,-0.2112903744,0.0111825662,0.2008690387,0.2838079035,0.0930746794,0.5682291985,0.3745242953,-0.1244232282,-0.0145360259,0.7123065591,-0.0103469836,-0.0092323916,-0.3011119068,-0.2046632171,-0.1842921078,0.2573680878,-0.19497177,0.1513123661,0.0869634598,0.0218750834,0.0457168557,-0.1883669794,0.1670678705,-0.1628591269,-0.2159621567,-0.3804614842,0.28824839,-0.1184282079,-0.2861639559,0.2685067356,0.1581974924,-0.1777515113,0.3684619367,-0.3176839054,0.8326913714,-0.3146328032,-0.052345816,-0.1263722926,-0.3259542584,0.2553507984,-0.0859473869,0.0067332201,-0.4939081669,0.0464636348,0.1598235369,-0.0380564183,-0.0747022554,-0.0262723658,-0.3291626275,0.1905131489,-0.0592301115,0.026452899,-0.1630082428,0.3300626576,-0.000765758,0.2471743077,-0.0477951802,0.1346662641,0.1436000466,0.2798327804,-0.0514159501,-0.030750161,0.1794309765,-0.5663870573,-0.2774447501,0.109191522,-0.3608016372,0.52219522,0.0183589738,-0.2535381317,0.2114667892,0.0314418525,0.2328994274,0.2051720172,-0.0495857224,0.1088980958,0.2894084752,0.0992841274,-0.2045168281,-0.0102028213,-0.0519634746,-0.0296652745,-0.3613168299,-0.0470949225,-0.062256787,-0.4015721083,-0.1501973569,0.2481306344,-0.1729173511,-0.0350113213,0.1649698615,-0.0270376559,-0.1076019481,-0.3131228089,0.1662441641,0.1862301975,0.029859066,0.1124023497,-0.0546169095,-0.169259727,0.0750196949,0.169458583,0.3665231168,0.0696555749,0.5085343719,-0.2613405287,-0.0337564163,-0.26124686,0.080045633,0.1148148999,-0.6623125672,0.2191943377,-0.2941145301,0.2200989276,0.1073093116,0.4256075621,0.3321913183,0.2134764493,0.0559269302,-0.205626905,-0.1126111373,-0.048178874,0.0279103816,0.236250475,0.0393019691,0.1607012153,-0.0426157676,0.1991036534,-0.3712837994,0.2617740333,-0.1110305041,0.0789095759,-0.015682565,-0.1059615612,-0.0660709813,-0.0290733073,0.1858225614,0.0987493768,-0.1861193478,-0.2591012418,-0.0545436591,0.1198962927,-0.0358448029,0.1415406018,-0.2157308608,-0.1956248879,0.2309210896,-0.0465358049,0.1875669509,-0.1173269898,0.1464101672,0.0604631864,0.1159075946,0.1060184315,-0.249331519,0.1973869056,-0.2469549179,0.3111076653,-0.057277672,0.1815192103,-0.0298763048,-0.0798529759,-0.3063936234,0.0847236514,-0.1822786033,0.1095403582,0.1555236876,-0.2479234338,-0.1675855517,0.4212432802,0.1682537049,0.5145013928,-0.3406195343,0.1270991415,-0.0219485331,0.2609172165,-0.3668580651,-0.1234404743,0.0033128208,0.0253490303,0.2622594833,0.3413487673,0.2693054974,-0.191019997,-0.0815276653,0.0918128043,0.400223881,0.0021996188,0.3656622767,0.3726800978,0.2654097676,-0.2072130889,0.2973615229,-0.0409238376,0.0624074191,0.5390818119,-0.5399002433,0.1284631491,-0.0205059703,0.2541928887,0.2257818878,-0.1200048551,0.3996991813,0.0580264181,-0.0998467579,0.0097591048,0.013767126,0.1629028618,-0.1720842123,-0.0554949455,-0.071970351,0.0846283734,-0.1727139503,-0.1010475531,-0.109577246,-0.2272497118,-0.2541548312,0.0966638103,0.0138639556,0.0599566214,0.1649866998,0.0004623942,-0.3333249986,-0.3664300442,0.2015645951,0.2978005707,0.3739057481,-0.408154279,0.3000581563,0.4755453765,-0.0917498469,0.1509276927,0.531871438,0.6291666627,0.2287991345,-0.155759871,0.226746425,0.1105186641,-0.161161378,-0.19540748,0.061420206,-0.11543262,0.1707046926,0.1617058218,0.1147132516,-0.1398622245,-0.0747310445,0.265265882,0.1053485051,-0.0900573283,0.2006717324,-0.1858335286,-0.1396438628,0.0156244487,0.250236094,-0.3926045597,0.0097748563,0.3296543062,0.2164011598,0.1218209565,-0.2242482752,0.1425646544,0.0673690811,0.3181637824,0.1936120689,0.0830997378,-0.2443352491,-0.032158915,-0.6895320415,0.0927471668,-0.4029485583,-0.055442147,0.1859170943,0.3722107112,-0.315990746,0.0117311776,0.3509395421,0.0610805266,0.0317478999,0.2196607143,-0.2321172655,-0.009379602,0.4260524213,-0.2580840588,0.1416169703,-0.4072997272,0.3420375884,-0.3757020533,0.056476485,-0.1472166032,0.0422160588,0.1315653175,-0.2852279842,0.411778897,0.0977901816,0.3219894171,-0.0438527353,-0.2032979131,-0.2257692367,-0.1698814929,-0.2365563661,0.27471596,0.5092764497,0.116309531,-0.0046949629,-0.4373764992,-0.1929321438,0.2320550829,0.0616868325,-0.0785891041,-0.0752977654,0.07226751,-0.1834450215,0.1228627115,0.0518226996,0.2025102824,-0.091045849,0.1757325381,-0.252078712,-0.6924307942,0.3932871819,-0.1511267126,-0.0783274621,-0.1192815453,0.3561945856,0.2467123568,0.0758872554,-0.6449910402,0.2416941226,0.5567663312,-0.1992720366,-0.0071392427,0.1507195234,-0.1236560568,0.0373367332,0.0260232612,-0.0416060314,0.0341908894,-0.274558574,0.2524394095,-0.323622793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2552","title":"Keys should be unique error on code_search_net","comments":"Thanks to the dummy data, we know for sure that most of them work as expected.\r\n`code_search_net` wasn't caught because the dummy data only have one dummy data file while the dataset script can actually load several of them using `os.listdir`. Let me take a look at all the other datasets that use `os.listdir` to see if the keys are alright","body":"## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n","comment_length":61,"text":"Keys should be unique error on code_search_net\n## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n\nThanks to the dummy data, we know for sure that most of them work as expected.\r\n`code_search_net` wasn't caught because the dummy data only have one dummy data file while the dataset script can actually load several of them using `os.listdir`. Let me take a look at all the other datasets that use `os.listdir` to see if the keys are alright","embeddings":[-0.0364270136,-0.0201952066,-0.0964384675,0.3276674449,0.0761770457,-0.013212272,0.2080996335,0.2912021279,0.0984992534,0.071184732,-0.067037195,0.3835592568,-0.1711656004,0.1363995969,0.1288103014,0.0837996826,-0.0438017175,0.1780875623,0.1199660823,-0.0594822355,-0.3179877698,0.2856626809,-0.234905839,0.0165575091,-0.2090522498,-0.137646243,-0.1760066003,0.2507343888,-0.1824110895,-0.5908303857,0.2711940408,0.2093391418,-0.2992361486,0.3508662581,-0.0001107919,0.1900058538,0.3981910348,-0.0256017838,-0.2827827632,-0.1791191101,-0.0728083253,-0.0612183549,0.0831817463,-0.2482592016,-0.0818561614,-0.1224978566,-0.1049753577,-0.5397647619,0.8384168148,0.2460178584,0.2726716101,0.2112665921,0.3223752081,-0.2598754168,0.2420286089,-0.1614331901,-0.0639911294,0.3828731775,0.1833221018,-0.0479908027,0.0057221493,0.2981964052,0.1404444277,-0.0875044614,0.4186807275,-0.0323041193,0.0468773171,-0.3087764084,0.2146259993,0.102823697,0.0884276927,-0.3128267825,-0.2070395052,-0.2669886947,0.0362550654,-0.2407446951,0.3180388212,0.1313280463,-0.202814877,0.1684758216,0.0087359957,-0.2333392948,0.0168732386,-0.035682261,-0.0120590134,0.0075637288,-0.0477772057,0.1442216039,0.1870479286,-0.0301870052,0.2658228576,-0.1093940288,0.019527711,0.1790449172,-0.5136678815,0.1230201051,-0.0518566482,-0.0984727517,0.2277046442,0.028717719,0.1698817313,-0.0687136874,0.0624927841,0.0053730593,0.0905160308,0.1125339046,0.0269704033,0.2809763849,0.2035713792,-0.1502015889,-0.2411468476,-0.0358261578,-0.1814642251,-0.1567694843,0.3382406533,0.0708147213,0.5329917669,0.0759846792,-0.2476930916,0.1277484149,-0.1235774383,-0.0753641129,0.0569920801,0.3604098856,0.1586510837,-0.3382246196,0.1139072999,0.2257108837,-0.1590736508,-0.2025012225,-0.3317087293,-0.119950287,-0.0704322308,-0.1687156856,0.2553776503,-0.5790083408,0.3142931163,0.2457666844,-0.1194907054,-0.4515373409,0.0783602819,0.1040028781,-0.1953116655,0.0890586823,0.0208360329,0.0583035946,0.0076981285,-0.2369559258,-0.0453378856,0.1677516699,-0.2531954348,-0.3557937145,-0.0943062007,0.3007490635,0.0646655783,0.0803303868,0.0169255044,0.055400297,0.1994582266,0.1568813622,-0.0127063803,-0.03325589,-0.1019226387,-0.2483245134,0.2735359669,0.6290325522,-0.1159738824,-0.0979010016,-0.1622074544,-0.0023760367,0.0797294006,0.3462894559,-0.2412421405,0.39027071,-0.3654875755,0.1659935266,0.6817964911,-0.2055112571,-0.3552485406,0.1524267942,-0.2477335483,-0.1710256934,0.0933934301,0.0315536559,0.1626765132,0.1522602588,0.1557519287,-0.066762425,0.057032384,-0.0515627451,-0.3554735482,-0.3471826017,0.1130477786,-0.067095235,0.1661296934,-0.1674634963,0.241747722,-0.1381936371,0.125456661,-0.112406373,-0.0802565515,0.0907338858,0.2145911753,0.0272686835,0.0985970348,-0.0643676594,-0.4726127684,0.3678174019,-0.2958661616,-0.237139076,-0.2157186568,-0.2840786278,-0.4142509103,0.1149873361,-0.5730288029,-0.0325938165,0.220072031,-0.0267939493,0.2193904966,0.0880645588,-0.120731473,0.1589354426,0.0497187562,0.0876840279,-0.2432176024,0.2517299056,-0.4583709538,-0.0632469282,0.1095043197,0.2277159989,0.3038507998,-0.1221485436,-0.1701619029,0.4219498336,-0.2227303982,0.0020727266,-0.0018699454,-0.1487248391,0.0674977824,-0.3830093145,0.1845069975,0.299370259,0.1757295728,-0.1566673368,-0.0986901745,0.3583337963,-0.0740240216,0.0907488316,-0.0510714389,-0.0368811749,0.1225299686,-0.1088033319,0.2126482427,-0.1271819621,0.2830677629,-0.3175019026,0.402438879,0.1329597384,-0.232151404,0.1211523637,0.2946249545,0.0163619891,0.1883885264,-0.0024340528,-0.021498641,-0.1830296367,-0.2807162106,0.2342668921,0.2953345776,0.1781710535,-0.1118414998,0.2570243478,0.0040008421,-0.187197119,0.3797668815,-0.1941545904,-0.1340707093,0.3846864104,0.0899585038,-0.0086621232,-0.4447206259,-0.1110007465,0.1650653332,0.1223477051,-0.2493936718,0.0003657769,-0.163993448,-0.0636476278,-0.4797675014,0.5345814824,-0.0994899422,-0.1488792896,0.2435329407,0.0811946914,0.1193659082,0.1278974265,-0.2404330522,0.149900198,0.0576753877,0.081865415,-0.2238828242,-0.1108079851,-0.0923991054,0.0561681278,0.3637707829,-0.0652652755,0.5341525674,-0.1901935488,0.0905751288,-0.2226313949,-0.3574887812,-0.0129970387,0.002620938,0.1580307782,0.1113686189,-0.0347031057,-0.0118298205,-0.3635375202,0.270929873,-0.3152554929,-0.4279221892,0.0140691744,0.0075032748,-0.0309336297,-0.2294654548,-0.5562415123,-0.2520560026,-0.2643299401,0.2726472616,-0.0757407993,0.0702347681,0.3859768212,-0.0329326019,-0.0093224272,-0.028510984,0.1671642959,-0.3440155685,0.0109188473,0.0526414216,-0.0584129617,-0.3240551949,-0.0440948233,-0.1682155877,0.1607696265,-0.1315598488,-0.3624717593,0.2376160473,-0.2448980063,0.4199641347,-0.2150039226,0.2762877047,0.2194638401,0.2579503059,-0.1546107978,-0.1946175992,-0.043108236,0.1191187724,-0.1772721857,0.0579537489,0.0751885027,0.2850073278,-0.2456805408,0.8134369254,0.5724945068,-0.0077527193,0.1956249475,-0.2384260297,0.3781944811,-0.2580958009,-0.4343471825,-0.2718248665,0.1200620532,0.0793170184,0.1062154546,-0.0887999013,-0.0725793093,0.1693485826,-0.1328802407,-0.5772477388,-0.322095722,0.0516159087,-0.2195011079,0.0935852677,0.0843149647,0.0302751176,0.0016910095,0.0297221392,-0.0056097922,0.1969579905,-0.0401836038,0.0871122703,-0.1645721197,-0.1121440381,-0.3914951086,0.142877847,-0.1441078931,0.1839484721,0.0162565894,0.1846572906,0.3610930145,-0.1915576756,0.6522553563,-0.0972904935,0.1387352198,0.1152109206,0.1727878451,-0.1377714127,-0.1088003144,0.0332587659,0.4142427444,0.1114908978,0.1376121789,-0.2753408253,0.0131415976,-0.0005718403,0.2770029008,-0.3481082618,-0.3533656895,-0.1978776902,-0.3296109438,-0.5215113163,0.0043432377,-0.0022254363,0.2339484692,-0.0310418382,-0.098644264,-0.2354491353,-0.2289232165,0.0203458425,0.2096415162,0.2850694358,0.1258572489,0.5712017417,0.3908025026,-0.1034237444,-0.0132422447,0.7329349518,0.0072246217,-0.0141438823,-0.3090282977,-0.2262702137,-0.1921972334,0.2538832426,-0.1747230142,0.1761264354,0.0652289391,0.0201868508,0.0349428169,-0.1436337531,0.1644269824,-0.1691852659,-0.205705598,-0.3801513314,0.2729162872,-0.1259981692,-0.2761525512,0.2663392723,0.1311991662,-0.1692606062,0.3703281283,-0.3384573162,0.8030844927,-0.3222218454,-0.0323644541,-0.1175398901,-0.3198094964,0.2389767468,-0.0710944012,0.0164572187,-0.4927273691,0.0529790781,0.1650752574,-0.0380417965,-0.0772082284,-0.0188326817,-0.3268424571,0.1943096071,-0.04452179,0.0551317111,-0.1772856414,0.3472311497,-0.0141337141,0.2519333959,0.0079932502,0.1487937272,0.1591982841,0.2623224854,-0.0231434889,-0.0313843489,0.1624129266,-0.5612019897,-0.2426977456,0.1193097308,-0.381688118,0.5089933872,-0.0113082733,-0.2430344224,0.1899625063,0.0247279424,0.2385044545,0.2175934166,-0.0599934123,0.1229249239,0.2727579176,0.0899910331,-0.2118234783,-0.0277948845,-0.0924524739,-0.0050221677,-0.3614406586,-0.0488175899,-0.0599451624,-0.4107629061,-0.1464833766,0.2362815738,-0.1562005877,-0.062380746,0.1588268131,-0.0425824188,-0.1080782637,-0.3281361461,0.1703136265,0.155690968,0.0344356932,0.0936869606,-0.0587181374,-0.2066199481,0.0575105399,0.1759092361,0.3556886911,0.0973457471,0.5280073285,-0.2695705295,-0.0025807566,-0.2692103684,0.065910317,0.1300946176,-0.6779008508,0.2404310554,-0.2633764446,0.1785432547,0.1193555593,0.4354133606,0.3073660135,0.2008407265,0.0557549037,-0.1955749243,-0.1281788647,-0.0845308006,0.0431707017,0.2558164299,0.0304476079,0.1719136983,-0.0592599623,0.2200327516,-0.3742519617,0.2876404226,-0.0920171663,0.1150857583,-0.0360314026,-0.1365343779,-0.0439343601,-0.0217161085,0.1914599687,0.1017006263,-0.2023001164,-0.2717778385,-0.0802880675,0.113682352,-0.0390358046,0.1121336147,-0.2161178589,-0.1972960085,0.19720833,-0.0550850891,0.2046240717,-0.1175462902,0.1393636614,0.0641289279,0.1185412332,0.1053301543,-0.2349288315,0.1644869894,-0.2630435824,0.3048258424,-0.0665520579,0.15409033,-0.0324409455,-0.0691732466,-0.2817690969,0.0859373286,-0.1532095671,0.0907230303,0.1475838572,-0.2426227629,-0.1706993431,0.4171833396,0.1768940091,0.4963460267,-0.3218606412,0.1213969365,-0.0222584736,0.2640508711,-0.3719104528,-0.1372202337,0.0454551689,0.0078017167,0.2777574956,0.3237073421,0.2908854187,-0.1721037924,-0.0758723691,0.0758376569,0.3826352656,0.0070852907,0.3628194332,0.3571462035,0.2445202023,-0.2220552713,0.3068503141,-0.0539518408,0.0639207438,0.5503354669,-0.5318267941,0.136503756,-0.0347368494,0.230537504,0.2037721872,-0.1257787347,0.4079788327,0.0730624497,-0.11642088,0.015899213,0.0352265202,0.1720743924,-0.2178192437,-0.0557479672,-0.1123491526,0.0730849281,-0.175526306,-0.1190880165,-0.1176810265,-0.2050513625,-0.2649655342,0.1112110391,0.0344491079,0.0620379336,0.1656076312,0.0011820153,-0.335629642,-0.3886713386,0.1887097359,0.3114578128,0.3505722582,-0.3815402985,0.2776072025,0.4696751237,-0.1024904922,0.1744853109,0.5233427882,0.6046814322,0.2045619637,-0.161085248,0.230729565,0.0859481096,-0.1587414593,-0.2143244594,0.0859880373,-0.1125596091,0.150861904,0.1863923818,0.1192652807,-0.1523855776,-0.1243504658,0.2570668459,0.0892342255,-0.1157788485,0.2096571773,-0.1934622079,-0.1191295609,-0.0106685292,0.2577622533,-0.3703105152,0.0087721329,0.3382308185,0.1999278069,0.1214424744,-0.25324893,0.1432627738,0.0362461098,0.3206444085,0.1674529761,0.0708228052,-0.2496795058,-0.0344960429,-0.6907975078,0.1137467474,-0.434858501,-0.0502481647,0.1779345423,0.3789178133,-0.295219183,0.024981739,0.3320696056,0.071195662,0.0678387284,0.2166432589,-0.2155784965,0.0088594938,0.4026424587,-0.2635599375,0.1491953582,-0.4507478774,0.3349403739,-0.3472178876,0.0670801103,-0.1321837753,0.0437041409,0.1463814378,-0.2922665775,0.4322301447,0.0993864834,0.324154675,-0.0317808911,-0.180434376,-0.2188673615,-0.1486958563,-0.2408022434,0.258140415,0.5101234913,0.1396809369,-0.0000907326,-0.4157195985,-0.1875611693,0.2462857068,0.0892048255,-0.0388061106,-0.0925739929,0.0967665613,-0.2200208753,0.1512515098,0.0268630013,0.2260272354,-0.0823854953,0.1834840775,-0.2212614715,-0.6726047397,0.3810141087,-0.1573601812,-0.0951677784,-0.1255318671,0.3582267165,0.2455357611,0.0702350214,-0.6323145628,0.2581470609,0.5377963185,-0.2006970346,-0.0005393851,0.1608229727,-0.1072048321,0.0514505394,0.0228129644,0.0017285283,0.0554157384,-0.2744817138,0.2620239556,-0.3209923804]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2552","title":"Keys should be unique error on code_search_net","comments":"I found one issue on `fever` (PR here: https:\/\/github.com\/huggingface\/datasets\/pull\/2557)\r\nAll the other ones seem fine :)","body":"## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n","comment_length":16,"text":"Keys should be unique error on code_search_net\n## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n\nI found one issue on `fever` (PR here: https:\/\/github.com\/huggingface\/datasets\/pull\/2557)\r\nAll the other ones seem fine :)","embeddings":[-0.0202083215,0.0007684655,-0.0907565653,0.3496333361,0.085136801,-0.0248717666,0.1991802007,0.2581068277,0.0932296291,0.0736749247,-0.0927868485,0.3846124113,-0.1742851436,0.1145088226,0.1579982638,0.0978715718,-0.0322047956,0.1860354841,0.1146632507,-0.0663524717,-0.3220596611,0.2716599703,-0.2497789562,0.0057933927,-0.2198529243,-0.1277942508,-0.181460768,0.2327201515,-0.1582820863,-0.6227064729,0.2830570638,0.2367789745,-0.3231881559,0.3545233607,-0.0001138885,0.1849657595,0.3915029764,-0.0082440907,-0.272043705,-0.1919538826,-0.0973862335,-0.0876455605,0.1220966503,-0.2598312497,-0.0403411798,-0.126745373,-0.1003250331,-0.5317751765,0.8246197701,0.1936426014,0.2487318069,0.1887022704,0.3378750384,-0.2372305989,0.2439075857,-0.1666713655,-0.055227574,0.3877444565,0.1985958368,-0.0504906029,0.0121911727,0.2858553231,0.1549561769,-0.1102307737,0.4621378481,-0.052585803,0.0763141215,-0.3710743189,0.2049998045,0.089278467,0.0974366218,-0.2761585116,-0.2324383557,-0.2789843976,0.0560453385,-0.2132116556,0.3593420684,0.1221939251,-0.2076237649,0.1624036431,-0.0172570534,-0.2577531338,0.0126821715,-0.0440757908,0.0079860017,-0.0301998556,-0.0491084233,0.1541401595,0.1662726849,-0.03205375,0.2796732485,-0.0996795073,0.0411549285,0.2203778476,-0.557258606,0.1509339511,-0.0677143708,-0.097979486,0.2128643841,0.0290231053,0.1994989216,-0.0848766863,0.0650746226,0.0161665604,0.1034396291,0.096539937,0.0342402793,0.2327070087,0.2190907598,-0.1835646629,-0.251726985,-0.0357222743,-0.2129186839,-0.1737252921,0.4017063677,0.0611099601,0.5241869688,0.0679923967,-0.2271723598,0.1321724951,-0.1534138471,-0.100855045,0.0860638395,0.3410011828,0.1949203908,-0.3373978138,0.1137649566,0.2290328294,-0.1387767345,-0.2006437629,-0.3102109432,-0.121383667,-0.0552540682,-0.1454878449,0.2526148558,-0.6270567775,0.2642531097,0.2605431676,-0.1020764187,-0.4912848175,0.003810955,0.102369979,-0.2192972749,0.0943130627,0.0159664955,0.079572551,-0.0476267003,-0.2335376292,-0.0544433184,0.1780006588,-0.2537408769,-0.3525573909,-0.0227050148,0.2670414746,0.0590396114,0.1252042651,-0.0038168507,0.0386782661,0.2306306362,0.1851243079,-0.012323305,-0.0314189382,-0.0888888836,-0.2552238405,0.26879704,0.6540945768,-0.0965850651,-0.1106154993,-0.1760225296,-0.0000576481,0.1063128933,0.3705928326,-0.2221122235,0.4160015881,-0.3452946544,0.1597891599,0.7105045319,-0.1933838129,-0.3874007761,0.1839333773,-0.2476288974,-0.1824145019,0.106587939,0.0144339176,0.1848349273,0.1223945469,0.1634444296,-0.0733876675,0.0813080817,-0.0535241216,-0.3427564502,-0.3470592499,0.121237345,-0.0639005452,0.1830687076,-0.1579374671,0.272606492,-0.1373436153,0.0896514803,-0.1177653819,-0.063160114,0.0818895698,0.2199142277,0.0230338331,0.1041528732,-0.0576092042,-0.4630142152,0.381847471,-0.2959958017,-0.2730376422,-0.2498998493,-0.246673286,-0.3670990169,0.1595121622,-0.5803844333,0.0287356377,0.1856029034,-0.0443434604,0.2221629173,0.101912193,-0.1063280851,0.174161166,0.0609682389,0.0783875287,-0.20480676,0.2609229386,-0.4255307615,-0.0561136641,0.1025159657,0.2394187897,0.3366996348,-0.1111951694,-0.1904087812,0.4078956842,-0.2392886132,0.0221072398,-0.0348242894,-0.1301355064,0.0792379975,-0.387062341,0.1635849327,0.2781241834,0.1803104579,-0.1550774127,-0.0412182547,0.3161665499,-0.112228632,0.1176383421,-0.0673278421,-0.0462845862,0.1096145287,-0.1085916609,0.2148876637,-0.1273214817,0.3092455566,-0.322094202,0.4101284146,0.1456072778,-0.22843799,0.1148155704,0.2885634601,0.0129627753,0.1708480716,0.0142394491,0.0231226739,-0.1852747947,-0.3099054098,0.2503823042,0.307439208,0.1368938386,-0.1100538298,0.257715553,0.0272050723,-0.2018003911,0.3694401085,-0.1471651793,-0.1145038083,0.4094091654,0.1236055195,0.0418062434,-0.4395470321,-0.1031289399,0.2066828907,0.0979180336,-0.2481191754,0.0145452265,-0.1656813622,-0.0346917063,-0.4692341983,0.5736931562,-0.1025880724,-0.1179140359,0.2339111418,0.0687985197,0.1535279602,0.1181849316,-0.2238681316,0.1479697227,0.040018931,0.0913883373,-0.2332361639,-0.0914759412,-0.0854198113,0.004380466,0.4014088511,-0.1010481715,0.5399111509,-0.1814332008,0.079408668,-0.2130985856,-0.3361154199,-0.023754539,0.0386971571,0.1665667146,0.06583976,-0.0652712137,-0.0287215542,-0.3721289635,0.2697654366,-0.3284951448,-0.4169460833,0.0101507735,0.0055707325,-0.0413727947,-0.241752103,-0.5230997205,-0.2458373308,-0.2457344681,0.3262642622,-0.1252706051,0.0861840993,0.3265456855,-0.0873571783,-0.0341761373,-0.037580125,0.1786943078,-0.3457049727,0.04664151,0.0505956672,-0.0377124622,-0.3421707153,-0.058241915,-0.1246190444,0.1684557348,-0.0901200026,-0.3649651706,0.3040499389,-0.2496056408,0.4348332584,-0.216963917,0.2921915948,0.1888551712,0.252348423,-0.1235338822,-0.1784010381,-0.0185365994,0.1148375496,-0.1724200696,0.0444516055,0.1156713217,0.2916902006,-0.2708283961,0.8078584671,0.5670900941,0.0105675859,0.2099637091,-0.2143362761,0.3573082387,-0.2475783229,-0.4330516458,-0.265617162,0.1070937589,0.0813839361,0.0956932455,-0.0795380622,-0.0807897821,0.1657532603,-0.1809013635,-0.5866481066,-0.3168035448,0.0558157377,-0.2445258647,0.1324726641,0.0851006135,0.0268228948,0.0293130334,-0.009579719,-0.0265898239,0.2357901782,-0.0056340466,0.0794821903,-0.1579551846,-0.1649989933,-0.3975367844,0.1181500554,-0.1273110956,0.2269959748,0.0055994368,0.2317032814,0.4163600504,-0.2022299618,0.6757409573,-0.0679794475,0.1560729444,0.0936870947,0.1393200755,-0.1231689751,-0.1203206852,0.073318772,0.44538486,0.1178607047,0.1218138635,-0.2651014924,0.0125441635,-0.0674534217,0.2935519516,-0.3799757659,-0.409884721,-0.1448253989,-0.3187561333,-0.5102410316,-0.0192871578,-0.0145017533,0.2496225834,-0.0626726672,-0.0809310451,-0.2437484562,-0.215495348,0.0145714162,0.2108637542,0.2731423676,0.1007199958,0.6041335464,0.4393433034,-0.1413893849,-0.0042285016,0.7062295079,-0.0068030041,-0.0212428924,-0.2818036079,-0.2356341928,-0.1919905841,0.267948091,-0.2039373964,0.1777370721,0.0793219879,-0.0019129969,0.0330989622,-0.1960717738,0.1702937931,-0.1882783473,-0.2366938591,-0.3877386749,0.3027285635,-0.1271834671,-0.2844857275,0.2639060318,0.1304769665,-0.1688389033,0.3598023951,-0.3438882828,0.8287071586,-0.3545681536,-0.0525071472,-0.1525413841,-0.3235087097,0.2501273751,-0.0668521896,0.0228038728,-0.5010052919,0.0630506501,0.164028272,-0.0611053742,-0.1104051992,0.000188096,-0.2942344844,0.1739930362,-0.0544597283,0.020479992,-0.1534423828,0.3549319506,0.0091453847,0.218832463,-0.0245842841,0.1050479934,0.1587536484,0.2887305319,-0.0325891711,-0.0355949923,0.1833779365,-0.5770735741,-0.2792325318,0.1072814316,-0.3708567917,0.5369707346,0.0195550546,-0.2575698197,0.2425426841,0.0130721899,0.2327265888,0.2016302198,-0.0526870303,0.1224905252,0.3005026281,0.1007064953,-0.2253258228,-0.0210914314,-0.1039732173,-0.0016896373,-0.3443885446,-0.077352576,-0.0794806927,-0.4005855918,-0.1658963412,0.263105005,-0.1736078411,-0.0105283447,0.1767033786,-0.0521218181,-0.089037843,-0.3136577904,0.1497512311,0.1800006926,0.0409670696,0.1119866148,-0.0774741098,-0.1747510135,0.085968338,0.1816874146,0.3723405898,0.0833173394,0.5415468216,-0.2741075754,-0.0104319956,-0.2523803711,0.0836683959,0.0918598399,-0.7165339589,0.2361987382,-0.3014400303,0.2387851775,0.1011593491,0.4325396121,0.3051663637,0.2136659324,0.0406877548,-0.2169865519,-0.085824661,-0.0969070047,0.030451268,0.2258669436,0.0072451313,0.1676853895,-0.0845434889,0.2230578512,-0.3288599849,0.2706349194,-0.0811362043,0.0841261372,-0.0047695646,-0.1192708239,-0.0520376004,-0.0461233445,0.1594183147,0.0948787108,-0.1736592799,-0.2326162755,-0.0896770433,0.1331890374,-0.0400187112,0.1329048127,-0.2325379103,-0.2018631548,0.221135363,-0.0618867539,0.1928774565,-0.1243791506,0.1366808414,0.0372685567,0.1535414755,0.099937737,-0.2555323541,0.2024778724,-0.2809892297,0.3246926963,-0.043404337,0.1691111028,-0.0195342433,-0.0407928601,-0.3310754597,0.0457027443,-0.2060308605,0.089342609,0.1524748653,-0.2731484175,-0.1586348414,0.4062788785,0.158953324,0.5390267968,-0.3238693178,0.1248404384,-0.0324668884,0.2207842022,-0.3509299457,-0.15090096,0.0421092883,0.0099140061,0.2370727807,0.3453682363,0.3118670583,-0.2216415405,-0.1029436961,0.0660705641,0.40315786,0.0129868798,0.3764742911,0.3809967339,0.2930631936,-0.2026401162,0.2937021554,-0.0405768491,0.0268594678,0.5647311211,-0.5716761947,0.0970169008,-0.0100530321,0.2584647834,0.2096128464,-0.1318118125,0.3875591457,0.0626457557,-0.1136688963,0.0254143439,-0.0024647517,0.168188557,-0.200914219,-0.0553317815,-0.088798359,0.0949239433,-0.1822154522,-0.1194704026,-0.1094762161,-0.2381825298,-0.2414564937,0.126698643,0.0207602177,0.0668733418,0.1796784401,0.0243960209,-0.3266175091,-0.3924658895,0.2119547278,0.3055470884,0.3821725249,-0.4020872414,0.2971867323,0.4687800109,-0.108518891,0.1810619384,0.4946075976,0.6445351839,0.236722827,-0.1513954401,0.2144293487,0.0908138081,-0.1545377076,-0.1945607066,0.1061461791,-0.1023361087,0.1814478785,0.1494651884,0.0785034746,-0.1149477437,-0.1056778431,0.2771041691,0.1448203474,-0.0929138884,0.2185651064,-0.1976521313,-0.134417221,0.0211391225,0.2508802712,-0.3458345234,0.0088227186,0.3492432237,0.2417682558,0.1199208871,-0.2315484434,0.1283793002,0.034570653,0.3339750171,0.1753886491,0.0514892489,-0.247703895,-0.0324748643,-0.6966251731,0.0962712616,-0.4398213625,-0.0247141588,0.2195525914,0.3855265975,-0.3456201851,-0.009113214,0.372990489,0.0727437288,0.0275592767,0.2396108806,-0.2225517035,-0.0010233882,0.4384168684,-0.2554382086,0.1187904179,-0.4289164245,0.3651635945,-0.3669345379,0.0287134871,-0.1510292888,0.0678249225,0.1020132527,-0.2757465243,0.4158664942,0.0889690444,0.3340982199,-0.0057505616,-0.201738745,-0.2366868705,-0.1630813181,-0.2188815624,0.2959427834,0.5006561279,0.1152568832,-0.0208942667,-0.4221185148,-0.178897962,0.2249132544,0.0389955826,-0.0718408749,-0.0791553631,0.0578772575,-0.201552242,0.1607161015,0.0240388867,0.2076045871,-0.0899779871,0.1652424932,-0.2389786839,-0.6555591822,0.3939450085,-0.1794788241,-0.0907581821,-0.1361942142,0.3525639474,0.2518313229,0.0771782845,-0.6191431284,0.2389184088,0.5750244856,-0.2288495898,-0.0097293826,0.1533532143,-0.1298980564,0.0472925864,0.0175554194,-0.0761202276,0.0305063445,-0.294687748,0.2548615038,-0.308223784]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2552","title":"Keys should be unique error on code_search_net","comments":"Hi! Got same error when loading other dataset:\r\n```python3\r\nload_dataset('wikicorpus', 'raw_en')\r\n```\r\n\r\ntb:\r\n```pytb\r\n---------------------------------------------------------------------------\r\nDuplicatedKeysError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.8\/site-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 1109 example = self.info.features.encode_example(record)\r\n-> 1110 writer.write(example, key)\r\n 1111 finally:\r\n\r\n\/opt\/conda\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py in write(self, example, key, writer_batch_size)\r\n 341 if self._check_duplicates:\r\n--> 342 self.check_duplicate_keys()\r\n 343 # Re-intializing to empty list for next batch\r\n\r\n\/opt\/conda\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py in check_duplicate_keys(self)\r\n 352 if hash in tmp_record:\r\n--> 353 raise DuplicatedKeysError(key)\r\n 354 else:\r\n\r\nDuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 519\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\nVersion: datasets==1.11.0","body":"## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n","comment_length":91,"text":"Keys should be unique error on code_search_net\n## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n\nHi! Got same error when loading other dataset:\r\n```python3\r\nload_dataset('wikicorpus', 'raw_en')\r\n```\r\n\r\ntb:\r\n```pytb\r\n---------------------------------------------------------------------------\r\nDuplicatedKeysError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.8\/site-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 1109 example = self.info.features.encode_example(record)\r\n-> 1110 writer.write(example, key)\r\n 1111 finally:\r\n\r\n\/opt\/conda\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py in write(self, example, key, writer_batch_size)\r\n 341 if self._check_duplicates:\r\n--> 342 self.check_duplicate_keys()\r\n 343 # Re-intializing to empty list for next batch\r\n\r\n\/opt\/conda\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py in check_duplicate_keys(self)\r\n 352 if hash in tmp_record:\r\n--> 353 raise DuplicatedKeysError(key)\r\n 354 else:\r\n\r\nDuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 519\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\nVersion: datasets==1.11.0","embeddings":[-0.05065668,-0.0265310202,-0.0967089161,0.3376669586,0.0938270018,-0.0051850188,0.209637031,0.2723225057,0.0978665054,0.0846779868,-0.0767664909,0.3797584176,-0.1855772287,0.1312500238,0.1376769692,0.0704786927,-0.0451734737,0.1832596064,0.1311877966,-0.0342166089,-0.3247793019,0.277204752,-0.2256615758,0.010591249,-0.2162512988,-0.1412390471,-0.1852134913,0.2377099842,-0.1705391407,-0.6306680441,0.2620755136,0.2049673349,-0.2932028174,0.3494946063,-0.0001131814,0.1847909689,0.3866307437,-0.0220922306,-0.2522614896,-0.1984636784,-0.0995154232,-0.0947171897,0.0844924301,-0.2655925751,-0.0616117455,-0.1218316779,-0.0980843827,-0.563854754,0.8383651972,0.2186895013,0.2488029152,0.1780722141,0.3565167189,-0.2654960454,0.2333432734,-0.1770569235,-0.0573638491,0.3910858035,0.2137886882,-0.046179913,0.0189008489,0.2944657803,0.1549513191,-0.0945589989,0.4434221685,-0.0574329123,0.0748626143,-0.3529263735,0.2176122218,0.0829745308,0.1021128893,-0.2907671332,-0.224451527,-0.2823086083,0.0506774113,-0.2336997837,0.3289036155,0.1434268206,-0.2258563042,0.1656320989,-0.0085747242,-0.251427263,0.0117150955,-0.0579374619,0.0218853708,-0.0263175592,-0.0522210151,0.1410633177,0.1878550053,-0.0514880866,0.2746832967,-0.066602625,0.0352339335,0.1945719272,-0.5258197784,0.1267573237,-0.044437319,-0.1213004068,0.2348861694,0.0344490856,0.1818594038,-0.1160668433,0.0746419951,0.0022266628,0.0957321972,0.1307920367,0.014150287,0.2462746054,0.1825146526,-0.1548223495,-0.2489557713,-0.055968482,-0.1899846941,-0.1946251988,0.3649508655,0.0828335583,0.5289831758,0.086161226,-0.2361945361,0.1337899566,-0.1345017403,-0.0975798815,0.0642859116,0.3334595859,0.1980132908,-0.3291807473,0.1161269173,0.2278708071,-0.1273085922,-0.1955416352,-0.3080868721,-0.0857724398,-0.0730136856,-0.1487104446,0.2354016155,-0.6151371598,0.2906035185,0.2476571947,-0.1106656343,-0.5036480427,0.0178255811,0.103980951,-0.2419441491,0.078496024,0.0232211147,0.0666555315,-0.0222959407,-0.2303782851,-0.0598928668,0.1667026579,-0.2588890791,-0.375463903,-0.0666591451,0.2756346166,0.0991151854,0.1002610549,0.0241976697,0.0435334444,0.2254664153,0.1764665097,-0.0194035489,-0.0314910486,-0.0847402588,-0.2560476065,0.261962533,0.6545096636,-0.0967016146,-0.0910337567,-0.1727345586,0.0162480902,0.0978435725,0.3790159523,-0.2228413671,0.4225227237,-0.3517396152,0.1507118344,0.7040376067,-0.1903835386,-0.3690498769,0.1839616001,-0.25062567,-0.2029010057,0.1125364676,0.011254319,0.166931361,0.1460940689,0.1736392975,-0.0832218081,0.0437686481,-0.0466866717,-0.3396582007,-0.378565073,0.1158137321,-0.0809869766,0.1927709728,-0.1396760941,0.2820678353,-0.1008895338,0.106518738,-0.1311348677,-0.0840880349,0.085260585,0.2358700037,0.0379792601,0.0610436015,-0.0941919982,-0.4492209256,0.375194788,-0.2969653308,-0.2601015866,-0.2236411721,-0.2457150519,-0.4040378928,0.1304817796,-0.5598601699,-0.0123337079,0.1953626424,-0.026377935,0.2334782481,0.094678469,-0.1328272969,0.1824873984,0.0621087104,0.0786608309,-0.2175976783,0.2431551814,-0.4592080414,-0.0629118904,0.0995280519,0.2300848365,0.3229334354,-0.0937632397,-0.1757675558,0.4119703174,-0.2277362198,-0.0013415141,-0.000053395,-0.1339236647,0.0696881488,-0.3773822188,0.1881813556,0.2873413563,0.1758788228,-0.1463004947,-0.0467566364,0.3089581728,-0.0962026119,0.0982418284,-0.0553944409,-0.0379563756,0.1037116274,-0.1151415408,0.2339863181,-0.1282823086,0.3135376573,-0.3360981941,0.3928044736,0.1252299547,-0.2248101681,0.1092213914,0.2842074037,0.0226602666,0.1894483566,0.0273580495,0.0480376966,-0.1708833426,-0.3229432404,0.2553579807,0.3024672568,0.1463909,-0.13264139,0.2653288841,0.0214540493,-0.1830316931,0.3622650206,-0.1740483046,-0.1003461033,0.4041172564,0.0908254161,0.0271128844,-0.4047189355,-0.1392025799,0.1715613306,0.1138504818,-0.2378466725,0.0071048629,-0.1528921574,-0.0564409643,-0.4925183058,0.5608199239,-0.1072300524,-0.1246385574,0.2069615126,0.0719944164,0.1526001841,0.1005075648,-0.2353828847,0.1593100876,0.0551266931,0.1080132574,-0.1970785707,-0.0866711512,-0.0614470541,0.0214707665,0.3701880574,-0.1035158187,0.5540136099,-0.1882575303,0.1039422974,-0.2396972924,-0.3182742596,-0.0383764543,0.0301017147,0.1602738202,0.0871214345,-0.0271787085,-0.0237962883,-0.3870374262,0.2663693428,-0.2896523178,-0.4052238166,-0.0074309725,-0.0020919775,-0.0275115389,-0.2106699198,-0.5379372239,-0.2609393895,-0.2703044415,0.3129994869,-0.1199052706,0.0880830139,0.3569857478,-0.0480923764,-0.0172877945,-0.0212365892,0.1838794798,-0.3261013627,0.0197959747,0.033879146,-0.0585760958,-0.3170383871,-0.074226886,-0.1409898698,0.1756734997,-0.1075962782,-0.3542785645,0.286362946,-0.2775952816,0.4044704735,-0.2260819674,0.3047446311,0.1968455315,0.256400764,-0.1357321888,-0.1624037623,-0.0287843235,0.1043877974,-0.1742865741,0.074448064,0.1234032661,0.3143497109,-0.2439942062,0.8015075326,0.601149261,0.0021506308,0.2192438692,-0.2171153575,0.3666773736,-0.2702762485,-0.4478713572,-0.2633076906,0.1035987139,0.0686427727,0.0786147267,-0.0901965722,-0.0777262673,0.1561603993,-0.1707780063,-0.5983709693,-0.3052642345,0.0630594492,-0.254040271,0.1322244406,0.0592924021,0.0095057152,0.0147971325,0.0199251752,-0.024745604,0.2436976433,-0.0160879139,0.0968082845,-0.1489836276,-0.1618861258,-0.3931615055,0.131275475,-0.1290553808,0.211686343,-0.0039111823,0.2356928587,0.3793331385,-0.2040454149,0.6450525522,-0.0790483505,0.1470449865,0.1092166901,0.1435657293,-0.0940992236,-0.1058793068,0.0680073574,0.4331174195,0.1125947088,0.1072045788,-0.2608776093,0.0433252789,-0.0442133769,0.296264708,-0.3561616242,-0.3749849498,-0.1564511359,-0.3320351839,-0.5376149416,-0.0113805756,0.0059254584,0.2577068806,-0.0122771524,-0.0873329565,-0.2434146106,-0.2121430784,0.0184216369,0.1975051761,0.2777834237,0.1220577806,0.572491765,0.4022454619,-0.1494081616,0.0039839582,0.7032373548,-0.018190328,-0.0268693082,-0.3009820879,-0.2183618098,-0.184100464,0.2520702481,-0.182036981,0.1687922329,0.0891662836,0.0256716758,0.02156692,-0.1831193268,0.1712032408,-0.1753460765,-0.2259693444,-0.4079323113,0.2826685309,-0.12923944,-0.2515246272,0.2798965275,0.130074814,-0.1542232782,0.3632351458,-0.3352123797,0.8218469024,-0.3532815278,-0.0676455796,-0.1762682498,-0.3206603527,0.2579026818,-0.0719608665,0.0293071847,-0.4959406257,0.05418773,0.173623383,-0.0599406287,-0.0763888285,-0.0104474304,-0.3085425198,0.190341711,-0.0599964298,0.0012933286,-0.1672787964,0.3595794141,-0.0123127857,0.2399487942,-0.0135558872,0.1152837798,0.1341265589,0.2973425686,-0.050401859,-0.0284443051,0.1919243932,-0.5713018179,-0.3050013781,0.1107295454,-0.36812675,0.5553808808,-0.0081700524,-0.2508479655,0.2377807796,0.0325038433,0.2473184019,0.2280904502,-0.0505328551,0.1029120609,0.2869988382,0.0714267641,-0.2414245456,-0.0139442487,-0.0804189816,-0.0125871524,-0.3690924644,-0.0620405041,-0.0703915656,-0.4344570637,-0.1671416163,0.2464843988,-0.1694508791,-0.0364787579,0.1722529382,-0.0551458597,-0.0988622084,-0.3119081855,0.1464177966,0.1796561778,0.0623882301,0.1058061197,-0.0682824478,-0.1866519898,0.080572255,0.1547392607,0.3759058714,0.063302137,0.5445219278,-0.2628656328,-0.007223187,-0.2490151823,0.0811503008,0.0862506032,-0.6860795021,0.2478183508,-0.2962574661,0.2240407467,0.1087722033,0.4142891467,0.3356956542,0.2232793272,0.0819596872,-0.2150944322,-0.120950371,-0.0753168315,0.01272697,0.2521207035,0.0051373621,0.1877124608,-0.0497151464,0.2116050124,-0.3393207788,0.2870150506,-0.0936994404,0.0792336762,-0.0118992841,-0.1192263141,-0.0605761632,-0.0137026832,0.1599118263,0.1208096296,-0.1890536547,-0.2317387015,-0.0821919516,0.1298318058,-0.0272765197,0.148358807,-0.2258434147,-0.2204286903,0.2216612995,-0.0574712604,0.1928474307,-0.0927043855,0.1492186189,0.0430969,0.1456860453,0.0988435,-0.2510830164,0.197516948,-0.2540414333,0.3194800317,-0.0578661263,0.1743402034,0.0008012624,-0.0530412421,-0.3475046754,0.061116226,-0.1615915447,0.0822992623,0.1322798878,-0.2498302311,-0.1537819505,0.4204137921,0.1611459702,0.5025945306,-0.3153916001,0.1437795162,-0.023762973,0.2335669696,-0.3553115726,-0.1331441849,0.0018798779,0.0134974504,0.2440273315,0.3504210114,0.2872249186,-0.1826773882,-0.1038487256,0.0704801828,0.4004673958,0.007782924,0.3844712079,0.3722216785,0.278423965,-0.2335223556,0.2874755263,-0.043161355,0.0500269085,0.5659123659,-0.5653172135,0.1077284217,-0.0338053741,0.2446524352,0.207935825,-0.1193798706,0.385962218,0.057640411,-0.1204963326,0.0073298337,-0.0034068655,0.1665775776,-0.2071653455,-0.0567429736,-0.0929066911,0.0841624811,-0.176660642,-0.1067894325,-0.0823224187,-0.2316903919,-0.2621318698,0.1127241775,0.0305349007,0.0736295804,0.1796397716,-0.0074665812,-0.3481133878,-0.3785908222,0.2002186179,0.321432054,0.367128104,-0.4137477875,0.2705955505,0.4744780958,-0.1166892722,0.1705087423,0.5382396579,0.6460700631,0.2355008721,-0.1561446637,0.2338993549,0.0935708657,-0.1518336236,-0.1770099699,0.0792698115,-0.1090334579,0.1749890745,0.1633067578,0.0851404667,-0.1243620738,-0.1141829491,0.262571305,0.138003394,-0.1096436009,0.2217371464,-0.1939567029,-0.1094134673,0.0232875217,0.2653509974,-0.3807305694,0.0055837547,0.343703419,0.2179344296,0.1009056568,-0.2297047973,0.1310104728,0.0447446443,0.3138360083,0.1847361773,0.0450944826,-0.2534657717,-0.039429687,-0.7021536231,0.0885074437,-0.4366108179,-0.0163164642,0.1816154867,0.3796536922,-0.338712126,0.0108205061,0.3618087471,0.0711823478,0.0239166599,0.2403919399,-0.2153799981,-0.0084127076,0.4215864837,-0.2574015558,0.140798822,-0.4249763787,0.3416321278,-0.3546767533,0.040543668,-0.1433149874,0.0585684702,0.1231763735,-0.3012675345,0.3987598121,0.1030943543,0.3309637904,-0.0177954324,-0.231041193,-0.239416346,-0.1723138243,-0.2147094756,0.2892085314,0.5318424702,0.1468508244,-0.0255861953,-0.4306866229,-0.209337309,0.2462720126,0.0495482162,-0.0583305918,-0.0740982443,0.0639474466,-0.1839034408,0.1469751,0.0575052723,0.2097403705,-0.0927490145,0.1925729662,-0.2331942171,-0.676718533,0.3900068402,-0.1394375563,-0.1063907817,-0.1198499426,0.346855402,0.2457020283,0.0856784135,-0.6321530938,0.2525626421,0.5547974706,-0.2041346282,-0.0026269513,0.1426166743,-0.1399857253,0.0241543483,0.0314404629,-0.0570403375,0.0354685038,-0.2795387208,0.272446245,-0.3017102778]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2552","title":"Keys should be unique error on code_search_net","comments":"Fixed by #2555.","body":"## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n","comment_length":3,"text":"Keys should be unique error on code_search_net\n## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n\nFixed by #2555.","embeddings":[-0.0533968322,-0.0143727232,-0.1070379838,0.3222682178,0.0807847753,-0.0239301436,0.1970542669,0.2859337926,0.0986966565,0.0808722898,-0.0773823187,0.3987371922,-0.1725007296,0.1318332255,0.1314204782,0.0896402597,-0.0515288673,0.1915138811,0.1092215478,-0.0468224883,-0.316539228,0.2875036299,-0.2310720831,0.0087489085,-0.1970688999,-0.1292038709,-0.16486229,0.2445614934,-0.1677804738,-0.612480402,0.2543263137,0.22000359,-0.3083204031,0.3416550756,-0.0001108894,0.1955513954,0.387308538,-0.006268532,-0.2742982507,-0.1843951195,-0.0847209841,-0.0831532925,0.0977364555,-0.2522194088,-0.0616638288,-0.1263833195,-0.1068454981,-0.5318850875,0.8346508741,0.2181143761,0.271707058,0.2026619315,0.3351421952,-0.262803793,0.246955961,-0.1953627318,-0.0622251481,0.3748189807,0.206589669,-0.050834693,-0.0060305046,0.3016416728,0.1487404406,-0.1080271378,0.433770597,-0.0445030481,0.0523363873,-0.321533531,0.2175844014,0.0836502463,0.0774272606,-0.2930870652,-0.2062520385,-0.2766878605,0.0427994356,-0.2333026677,0.3327721953,0.1154182628,-0.2088796347,0.1602129638,0.0027357705,-0.2427585125,0.0192203578,-0.0510309227,-0.0091924705,0.0070543136,-0.0450794548,0.1399084479,0.1945799738,-0.0403306633,0.2562682629,-0.0754380599,0.032699462,0.1822511256,-0.5125237107,0.1279069036,-0.032607574,-0.104706414,0.2095108628,0.0395220928,0.180271402,-0.079987973,0.0715635195,0.0041978406,0.0961282924,0.1027244925,0.0378833376,0.2409258187,0.2083101571,-0.1635383815,-0.2565801442,-0.0384946615,-0.1870898902,-0.1824906766,0.3729274869,0.0699900165,0.5246412754,0.0895226076,-0.2392974943,0.1232693866,-0.1234850213,-0.0857701451,0.0720574185,0.3702450097,0.1859916151,-0.3403613865,0.1165724695,0.2051274925,-0.1431888193,-0.1829725802,-0.3314496577,-0.1097373068,-0.0631923452,-0.1655055881,0.2568235695,-0.6022323966,0.2901754677,0.2320697755,-0.104047142,-0.4787913263,0.0359531529,0.1257972717,-0.2027774304,0.0975408852,0.0179266762,0.060191568,-0.0128213558,-0.244178012,-0.0437334664,0.1718090773,-0.2469541579,-0.346116513,-0.0687887743,0.2999408245,0.0847266987,0.0890746266,0.0250012465,0.0341085084,0.2055002302,0.1669212133,-0.0050359406,-0.0313373432,-0.0880494863,-0.2548519969,0.2763558328,0.6298667192,-0.079994753,-0.1094424725,-0.1712303311,-0.0042484938,0.0792527869,0.3580657244,-0.2363410592,0.3886481822,-0.3495095372,0.1838909537,0.6973569393,-0.1961959749,-0.3745010495,0.1674307734,-0.2635224164,-0.1904798895,0.0918365046,0.0257173963,0.1769699752,0.1361354887,0.176670447,-0.0727123842,0.0693897828,-0.0450743288,-0.348092109,-0.3425667882,0.1156722307,-0.0734088048,0.1833988577,-0.1655219495,0.2732811868,-0.1198534891,0.119783856,-0.1168712452,-0.082456775,0.0823877454,0.223249644,0.0279584751,0.0859961957,-0.0747660771,-0.4576280713,0.3750629425,-0.2880455852,-0.2653601766,-0.2408017814,-0.2551548779,-0.3996886909,0.138279289,-0.5677933097,-0.0174775422,0.2221558392,-0.0235242154,0.2057960629,0.0829384699,-0.1275241822,0.1573619992,0.0609475486,0.0781474486,-0.2076239437,0.245385319,-0.4501036108,-0.0626533628,0.1087213755,0.2308883369,0.3151966333,-0.0982680321,-0.1675671786,0.4158402383,-0.2375197113,0.0144378403,-0.0141418278,-0.1483062953,0.0622466505,-0.3815093637,0.1785711795,0.2746314108,0.1735600382,-0.153909415,-0.0621020347,0.3305004537,-0.0955859721,0.0992713645,-0.0376189798,-0.0444837175,0.1111544818,-0.1081904098,0.2102451473,-0.1301473528,0.2881899774,-0.327945292,0.3865830004,0.1341595799,-0.2301338017,0.1248834059,0.3010281622,0.0242341794,0.1928975731,-0.0004765169,0.0053069117,-0.1678134501,-0.3130809069,0.2438852191,0.2847448289,0.1720804423,-0.1023228243,0.2566377819,0.020099422,-0.1971287727,0.3716135919,-0.1795191616,-0.1076992452,0.3959313035,0.1076503322,0.0092673851,-0.4451441169,-0.126077354,0.1587785631,0.1019695252,-0.2394163758,-0.0040552537,-0.1615384072,-0.0540601574,-0.4680299163,0.5390894413,-0.0999609381,-0.1468264461,0.2408241481,0.0808430836,0.1227976531,0.1253394932,-0.2418409884,0.1440522373,0.0573358983,0.0837442651,-0.2209590971,-0.096763514,-0.0733008236,0.0495434105,0.3611040115,-0.0959244221,0.5439587235,-0.166275084,0.1021010131,-0.2262979746,-0.3556706905,-0.029339809,0.002184473,0.1323457211,0.0943808928,-0.0311663784,-0.0181329604,-0.3841351867,0.2713574469,-0.3131960034,-0.4201098382,0.0105116079,0.0190849863,-0.0370074809,-0.2320653498,-0.533623755,-0.2426783293,-0.2830356658,0.3077528477,-0.0909874439,0.0709447786,0.3692486584,-0.0507768616,-0.0068613724,-0.0346290022,0.1844833642,-0.3331341147,0.0398061946,0.0507979356,-0.0577668995,-0.3258475959,-0.0627020523,-0.1336278915,0.1605076343,-0.1120974869,-0.3567942083,0.2776930928,-0.2597378492,0.4277976155,-0.2070044577,0.2878881991,0.2005620152,0.2624753714,-0.1615035385,-0.1764029413,-0.0316534601,0.1085411161,-0.1759968549,0.0398468859,0.0768966302,0.2778222859,-0.242135942,0.7894445062,0.5564331412,0.0068793525,0.2016498595,-0.221604839,0.3635284007,-0.2604294419,-0.4388687015,-0.2594321966,0.1166966036,0.0802225545,0.0993626714,-0.0904172584,-0.0896538123,0.1587798595,-0.1564350277,-0.5858994126,-0.3235034645,0.0644672364,-0.2261741161,0.106585443,0.0809580088,0.0333757699,0.0128276125,0.0246659759,-0.0108781606,0.2094115466,-0.0311140493,0.079485625,-0.1433953941,-0.1512023211,-0.4019238353,0.1308379024,-0.1423098892,0.192145288,0.0098953964,0.1752304286,0.3757168055,-0.2031295151,0.643211782,-0.0757270679,0.1482163221,0.1100512296,0.1494389027,-0.1083078235,-0.1044784635,0.0457977019,0.4213986397,0.1108620241,0.133942619,-0.2678880394,0.0188503377,-0.0329622738,0.2570330501,-0.3487218916,-0.3668176234,-0.1734670103,-0.3336412311,-0.5230059624,-0.0028251943,0.0033894628,0.2519948184,-0.034329962,-0.1003243104,-0.2416806072,-0.2277779281,0.0046701678,0.2114621997,0.2671382427,0.1269395947,0.5715430975,0.3864926696,-0.1117819622,-0.0191210601,0.7062813044,0.0008950132,-0.0177768841,-0.298941344,-0.2241455168,-0.1911696047,0.2594124973,-0.1943227351,0.178103283,0.0905480236,0.0387033187,0.0407104008,-0.168817535,0.1665786654,-0.180982545,-0.2091946304,-0.3766992688,0.2843481898,-0.1441028267,-0.2738368809,0.2588084042,0.1349831522,-0.1615850329,0.3583953977,-0.3398579061,0.8040597439,-0.3355532885,-0.0493056625,-0.1353061944,-0.3193356693,0.2495797724,-0.0624440946,0.027413208,-0.4884461462,0.0388658047,0.1779942364,-0.0363348126,-0.091912657,-0.0178975835,-0.3016380966,0.1731093526,-0.0393079296,0.0295191966,-0.1663223207,0.3395930231,-0.0045519578,0.2305639237,-0.0205249488,0.139823541,0.1506290138,0.2689973414,-0.0310202725,-0.0261450354,0.1887905598,-0.5750886798,-0.2639360428,0.1243832633,-0.3629533648,0.5416007638,0.0033331506,-0.2426807731,0.2085205168,0.0241508596,0.2231005579,0.2118819803,-0.0465240255,0.1267839372,0.2899291217,0.0893023014,-0.2263979614,-0.0138728647,-0.0861660838,-0.0081978505,-0.3483110964,-0.0548679717,-0.0553767495,-0.4253959656,-0.1436448246,0.2697444856,-0.1736813188,-0.0323105641,0.1729264706,-0.0446044244,-0.1069929376,-0.3301971853,0.1695884615,0.1641590446,0.0394321494,0.0943687409,-0.0541383997,-0.1942344755,0.0725209489,0.1783529371,0.3647486269,0.0794348866,0.5327529907,-0.2519989908,-0.0175303593,-0.2679443359,0.0580362305,0.1101654992,-0.6800182462,0.2487094104,-0.2705337107,0.2052155137,0.1257912815,0.432469815,0.2996316552,0.2080721259,0.057952486,-0.1952282339,-0.1033286601,-0.0759744272,0.0186634548,0.2412575632,0.0214397311,0.1878387481,-0.0651526079,0.2113649547,-0.3700384498,0.2825139463,-0.0981565937,0.0932007506,-0.0317372344,-0.1316179782,-0.0413661152,-0.0153935654,0.1966882795,0.1158256978,-0.1923485547,-0.2686248124,-0.0765504017,0.1141903773,-0.0406052247,0.1412652433,-0.2305047661,-0.197676599,0.2151770294,-0.0629517287,0.2182739228,-0.1200023144,0.1405351162,0.0481515452,0.130387947,0.1115672141,-0.2257810235,0.1814168543,-0.2625011802,0.3120441139,-0.0575848073,0.1644658893,-0.0297552049,-0.0560417138,-0.2961214781,0.0803730413,-0.1634289175,0.092863597,0.149157688,-0.2472429276,-0.1845309138,0.4066803753,0.1691397876,0.500020802,-0.3139244318,0.1140939817,-0.0376803763,0.2621702552,-0.3583672941,-0.1372183263,0.0178194735,0.0145977484,0.2742555737,0.3299461007,0.2642634809,-0.1989264339,-0.1037048697,0.0906431973,0.3875153065,0.0018439601,0.3536663949,0.3642461896,0.2744798362,-0.2310163826,0.2994637191,-0.037017338,0.0650061443,0.5564428568,-0.5413181186,0.1201018468,-0.0313466378,0.2433680445,0.201953575,-0.1229586378,0.3917524815,0.0637797639,-0.1149648726,0.0039911871,0.0109215965,0.1570710093,-0.198129192,-0.045021262,-0.0959235728,0.0825832263,-0.1843908876,-0.1251336634,-0.1091137305,-0.2221059501,-0.2628844678,0.1147605553,0.0369972624,0.0733826607,0.1713698655,0.002199153,-0.3206441104,-0.3787291646,0.1820993572,0.2968489826,0.3597643971,-0.4006875157,0.2696298659,0.4791851938,-0.0966947153,0.1685678512,0.5011821985,0.6143705845,0.2107598931,-0.1580123603,0.2359049618,0.0840263665,-0.1646708399,-0.2151739895,0.0764852911,-0.1019977182,0.1857760698,0.1742145568,0.1188110337,-0.1476275176,-0.1119240001,0.2629233301,0.1317374408,-0.0926465318,0.1966638416,-0.1932251751,-0.1300823838,-0.0012055866,0.2552567124,-0.3705136478,0.007609217,0.3208140731,0.215689525,0.1155446544,-0.239186272,0.1444088221,0.0489762351,0.3232978284,0.165722847,0.0628538281,-0.2255798429,-0.0453299023,-0.6783316731,0.1091699824,-0.4244835377,-0.0323454738,0.1800197959,0.3937352002,-0.3103295863,0.0069444296,0.3555320501,0.0707263276,0.035365358,0.2188761979,-0.2089483291,0.0003661659,0.43764624,-0.2704176605,0.1366697103,-0.4220765829,0.3297281265,-0.3720422387,0.0721950531,-0.1420854479,0.0426822118,0.1323527098,-0.2960659564,0.4249586463,0.0904715061,0.3035509288,-0.0378847457,-0.2052284479,-0.2310865521,-0.1704552174,-0.2223263383,0.2685654163,0.5100566149,0.1445935369,-0.0016432472,-0.4153850377,-0.1796501428,0.2437456399,0.060816247,-0.0556702614,-0.0710487217,0.0742131248,-0.1949863136,0.1504758596,0.0274487995,0.2121456712,-0.079520382,0.1658010632,-0.236993745,-0.6941068769,0.3908527493,-0.1587508023,-0.1104017049,-0.1261643469,0.3640736639,0.2386291176,0.0852466896,-0.6239665151,0.2493974119,0.5322613716,-0.2069462091,-0.0212811977,0.1549338996,-0.1419076174,0.0427102,0.0181371663,-0.047616668,0.0404423922,-0.2712614834,0.2481130064,-0.3176469207]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2552","title":"Keys should be unique error on code_search_net","comments":"The wikicorpus issue has been fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/2844\r\n\r\nWe'll do a new release of `datasets` soon :)","body":"## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n","comment_length":17,"text":"Keys should be unique error on code_search_net\n## Describe the bug\r\nLoading `code_search_net` seems not possible at the moment.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n>>> load_dataset('code_search_net')\r\nDownloading: 8.50kB [00:00, 3.09MB\/s] \r\nDownloading: 19.1kB [00:00, 10.1MB\/s] \r\nNo config specified, defaulting to: code_search_net\/all\r\nDownloading and preparing dataset code_search_net\/all (download: 4.77 GiB, generated: 5.99 GiB, post-processed: Unknown size, total: 10.76 GiB) to \/Users\/thomwolf\/.cache\/huggingface\/datasets\/code_search_net\/all\/1.0.0\/b3e8278faf5d67da1d06981efbeac3b76a2900693bd2239bbca7a4a3b0d6e52a...\r\nTraceback (most recent call last): \r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/builder.py\", line 1067, in _prepare_split\r\n writer.write(example, key)\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 343, in write\r\n self.check_duplicate_keys()\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/datasets\/arrow_writer.py\", line 354, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 48\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.1.dev0\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n- Python version: 3.8.5\r\n- PyArrow version: 2.0.0\r\n\nThe wikicorpus issue has been fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/2844\r\n\r\nWe'll do a new release of `datasets` soon :)","embeddings":[-0.0742809549,-0.0351793021,-0.086682789,0.3581603765,0.0623259954,-0.0354929715,0.1684754789,0.2639017701,0.0927660689,0.1236139983,-0.0445139334,0.3784416318,-0.1519159526,0.1347224712,0.1137825847,0.0393825993,-0.0278353021,0.1998654604,0.1008107662,-0.0348957106,-0.348711282,0.3278391957,-0.2847043276,0.0620546378,-0.2731282413,-0.1773471981,-0.1778877974,0.200719893,-0.2140284181,-0.6042189598,0.2684114873,0.2027064562,-0.2295345366,0.379314661,-0.0001112176,0.1849371642,0.3967281282,-0.0099956645,-0.252895385,-0.1352210194,-0.0593605377,-0.0799274743,0.1142255887,-0.3115119338,-0.0552829355,-0.1050350145,-0.0662939399,-0.4870356917,0.8415837288,0.2476684451,0.2632397115,0.2291175723,0.3477338552,-0.2014206797,0.258848995,-0.2904579937,-0.082405217,0.3073558211,0.1880030036,-0.0712194666,-0.0259935986,0.3081966639,0.1307695657,-0.0733631998,0.3915862739,-0.0154980402,0.0570181794,-0.3279979229,0.2298202515,0.1004076973,0.1076005101,-0.2757329941,-0.1830676496,-0.2352993786,0.0456514582,-0.2553208768,0.3790766001,0.1081376746,-0.1849624813,0.1618368924,0.018458413,-0.1499607563,0.0437140539,-0.0016606788,-0.051175721,0.1737132818,-0.0323066302,0.1444085538,0.1939402521,-0.0547873341,0.3041885793,-0.0899312124,0.0066322144,0.1831382215,-0.5206641555,0.1190481037,-0.0558351502,-0.0297692921,0.188243255,-0.0209248234,0.1625713557,-0.0958015844,0.0999090895,0.0181384943,0.1590435803,0.0995517075,0.0560125671,0.262648195,0.2037771493,-0.1440712363,-0.2874125242,-0.0467572995,-0.1567611247,-0.1744732261,0.3424735665,0.0209654979,0.4984278679,0.0877819881,-0.1772675961,0.0851962641,-0.1482088268,-0.1240128204,0.0419483893,0.3618213236,0.1723600775,-0.2316883355,0.1341062337,0.2210865617,-0.1648888886,-0.0792284459,-0.3258473575,-0.0771105886,-0.0448642634,-0.150892809,0.2584308684,-0.5565131903,0.3257267475,0.216427207,-0.0575943179,-0.5052257776,0.066263631,0.1426427811,-0.2543682158,0.1114748195,0.0145561239,0.0953841284,-0.0699025765,-0.2688690126,-0.033451952,0.1893269569,-0.2687191963,-0.3343837559,-0.120107539,0.290027231,0.0964327157,0.0909281671,-0.0038086718,-0.0060241935,0.1928693354,0.1217826903,0.0052761976,-0.0563435927,-0.0965555757,-0.2477224767,0.2557833791,0.5845033526,-0.0346923061,-0.1073500365,-0.1453454643,-0.047122445,0.09786053,0.3257711828,-0.256755501,0.4472723305,-0.3528083265,0.2633313835,0.688100636,-0.244620651,-0.3804998696,0.0933755264,-0.2463257015,-0.1834537238,0.0795573816,0.0523541681,0.1828671992,0.1490884125,0.114357844,-0.0442781225,0.1039808318,-0.0017024446,-0.3555277586,-0.4046012759,0.1209371164,-0.0609739199,0.1574059725,-0.1855020672,0.2923003137,-0.0762466267,0.1631142497,-0.1323100775,-0.1166701168,0.0497569442,0.1757787764,0.0060181851,0.0691934004,-0.0537211634,-0.4638299644,0.3711787164,-0.3174013495,-0.2441738099,-0.2198593169,-0.2543005347,-0.4006926417,0.1687960774,-0.571015954,-0.0822550878,0.2377775311,-0.0140658766,0.1930306405,0.0721662194,-0.1365028918,0.1315037161,0.0369368494,0.0550393388,-0.203536734,0.2733279765,-0.4755397141,-0.0901231393,0.1020118743,0.2486847192,0.3709596694,-0.0731649026,-0.1867815852,0.398981601,-0.1945809573,0.0200651065,0.0209246557,-0.1216245145,0.058594767,-0.3596253693,0.1598382741,0.2624950409,0.1872991323,-0.2088916302,-0.0540271401,0.3151096404,-0.068915613,0.1212315932,-0.0756698549,-0.0245238524,0.1108407378,-0.1342621893,0.2704889178,-0.1226503253,0.2525312006,-0.3048864901,0.2955518365,0.1302850693,-0.1617019475,0.1204012707,0.296366334,-0.016450977,0.1959628314,-0.0138482805,0.0177933909,-0.1833081543,-0.2783831656,0.2625537217,0.2929568589,0.1389880627,-0.0898682252,0.2467435449,0.0580504611,-0.1775527,0.3787803352,-0.1610396206,-0.1312910765,0.3710986972,0.0966158137,-0.0010818283,-0.3934862316,-0.0888896585,0.1221456006,0.1192339733,-0.1838064343,-0.0478310101,-0.1479789168,-0.0141880084,-0.4854323864,0.55943048,-0.1174688265,-0.2091077268,0.2476030886,0.1138995141,0.1475590765,0.1345303059,-0.2055603117,0.1488521695,0.0309208147,0.0777810737,-0.24337551,-0.1421609074,-0.0806783065,0.0434354879,0.3722387254,-0.0059759771,0.5033736825,-0.1726587117,0.1110564321,-0.2709143162,-0.3622726202,-0.011248257,0.0630451441,0.1775150448,0.1069365516,0.0299363397,-0.049996499,-0.4720452726,0.2490351945,-0.323165834,-0.4180160761,-0.0479350127,0.0075530084,-0.0420717336,-0.2085278779,-0.5222279429,-0.2750418484,-0.2573629022,0.2600233853,-0.0206884667,0.0890500769,0.2367381603,-0.024525512,0.0394696817,-0.0720371976,0.1314702928,-0.3177736998,0.0144972801,0.0866977721,-0.0611512884,-0.3245949149,-0.0370558761,-0.1435474306,0.208547473,-0.1518795043,-0.4038965106,0.221521616,-0.2699886858,0.4851068556,-0.1768900305,0.3683383167,0.2399503887,0.2866680324,-0.1356308162,-0.1880393475,-0.0889881849,0.0879557952,-0.2125381529,0.0945468917,0.0932359025,0.2151426226,-0.3133164048,0.8754813671,0.541955471,0.0103153884,0.1690273732,-0.249479875,0.3023969829,-0.2638945282,-0.401509583,-0.1732810736,0.0919193178,0.1000395566,0.1628784686,-0.0817174315,-0.1143440977,0.1291572005,-0.1651426554,-0.6445286274,-0.3470391333,0.0457326546,-0.0895171687,0.0543886237,0.0552177094,0.0347841457,-0.0177383609,-0.0084580705,0.0148496907,0.1984946728,-0.0989724621,0.0816219896,-0.1183469892,-0.0773094893,-0.4334527254,0.1488144249,-0.1107251942,0.2057377696,0.0669469833,0.1918520778,0.3870505989,-0.253490895,0.5979360342,-0.09488336,0.162070483,0.110263899,0.153328836,-0.0196097102,-0.0942645594,0.0241729412,0.3204050064,0.1508574337,0.0656291991,-0.299921453,0.0080905808,-0.0079524228,0.2704363465,-0.3470428884,-0.3830890656,-0.1817530245,-0.3612961471,-0.5679562688,-0.038894847,-0.0180033464,0.290887028,-0.0397793688,-0.0919554904,-0.2838910818,-0.2085041404,-0.0233911294,0.2433039993,0.2719501555,0.1303282678,0.5048880577,0.3754343688,-0.1317442954,-0.1215367615,0.6549913883,0.013013307,0.0737068653,-0.3199548423,-0.2183622867,-0.17773965,0.2366467416,-0.2096500844,0.2116485536,0.0596794188,0.0354308411,0.0525262356,-0.1354892999,0.2041814178,-0.1521792859,-0.181176275,-0.3168835938,0.3276797235,-0.0797863007,-0.2935816646,0.3204473555,0.1830193996,-0.1361523569,0.3454624712,-0.2941662669,0.864649415,-0.3622950613,-0.0112603176,-0.0950295925,-0.3614687324,0.2783358693,-0.0859615877,0.0235061459,-0.4701197147,0.089057751,0.1547519714,-0.0065730098,-0.073673211,-0.0190724749,-0.3405537307,0.1603783369,-0.0125335027,0.0628308952,-0.1861361414,0.3396660089,0.0210426189,0.2069726884,-0.0783225447,0.1233535483,0.106362626,0.2603191435,-0.0611044429,-0.0262385383,0.2197211236,-0.605375886,-0.2562443018,0.1415903121,-0.4317052662,0.5209038258,-0.0354227796,-0.1900562495,0.1860741675,0.1064293161,0.2052595615,0.2176427394,-0.0729290694,0.1493111402,0.2258951217,0.0472839512,-0.2381495684,0.0004402872,-0.1366214454,-0.0497531518,-0.3872920871,-0.0726489872,-0.0593208298,-0.4390932024,-0.094891578,0.323553592,-0.1323817819,-0.0263052806,0.1509141177,0.0208661519,-0.1581373066,-0.3281897902,0.1640160829,0.1439607888,-0.0019016332,0.1028021947,-0.0573713407,-0.1600798666,0.0399427079,0.2016285658,0.3525449336,-0.0204139575,0.5597220659,-0.2238451391,-0.0222035255,-0.2568180859,0.0505393259,0.0784840137,-0.6631227732,0.1979609281,-0.2169672251,0.1372680366,0.0952701122,0.5443470478,0.2849123776,0.2319777161,0.0891342163,-0.2373345941,-0.1445601881,-0.1138235107,0.0212968327,0.2813127637,0.0177842062,0.1847895235,-0.0796057284,0.2442847043,-0.3585905433,0.2909465134,-0.1517928988,0.1126165316,0.0235835034,-0.1157478616,-0.047515735,0.0386179276,0.2042898238,0.1302141547,-0.1892314553,-0.25830549,-0.1250487715,0.1193683594,0.019728424,0.1022297218,-0.1927415282,-0.1939717978,0.2376626581,-0.0976801515,0.2315524518,-0.1262791902,0.1216461807,0.0351070799,0.1617292464,0.1328461766,-0.300421387,0.1358696073,-0.2406536639,0.3165053427,-0.05069004,0.1333663613,-0.0060607661,-0.0459962189,-0.1634325385,0.0433419086,-0.1613444686,0.156269744,0.1328203827,-0.2350576222,-0.2607904971,0.4305790067,0.0915816426,0.5241196752,-0.3124518096,0.0793467835,0.0077804015,0.2575986683,-0.3564535081,-0.1353175789,0.1281305104,-0.0129006999,0.2492410392,0.3392430246,0.2372636497,-0.2636077404,-0.1254513711,0.0786132216,0.2940698266,-0.1136117801,0.388700366,0.3554524481,0.2433124483,-0.2754882872,0.2628868222,-0.0542764589,0.0325674713,0.5613678694,-0.5922800303,0.2081758529,-0.0940593779,0.255430609,0.2358905375,-0.0976408869,0.4575901628,0.0953567326,-0.0775236636,0.0187959578,-0.0119834486,0.0684350058,-0.207780987,-0.015233824,-0.1346884817,0.0715188384,-0.1995581686,-0.1121547595,-0.0987811163,-0.222635448,-0.2354946136,0.1105639935,0.0322937481,0.0883142054,0.1989616454,0.0262333136,-0.3541198373,-0.3746516705,0.1173283085,0.2755414546,0.4009862542,-0.3965549171,0.260845542,0.499607563,-0.1016847044,0.1175333261,0.5371187925,0.6057381034,0.2470074296,-0.1342766434,0.2150330693,0.0629096255,-0.1497749388,-0.1834006608,0.1092464775,-0.1083949804,0.1856255978,0.1824517399,0.1080617011,-0.152909711,-0.1572519839,0.2447272241,0.0984146073,-0.1011576951,0.1219464168,-0.1924872249,-0.135200277,-0.0399501249,0.2947030962,-0.3455330133,0.0732990429,0.2401662469,0.2233254462,0.1486766189,-0.2627126276,0.1377590001,0.0887553692,0.3642181158,0.1823172122,0.0794754326,-0.2567425072,-0.0662510544,-0.7227305174,0.1154727265,-0.4371538162,-0.0587759949,0.1690632403,0.3498442173,-0.3532672822,0.0450927727,0.357853204,0.1252616644,0.1136297882,0.2033278346,-0.2470277548,-0.0464198776,0.4598298371,-0.2959706485,0.1475212276,-0.4268782437,0.3193941116,-0.361877054,0.0644194335,-0.1150156185,0.0719149783,0.0668062195,-0.3483568132,0.3682030737,0.1753947288,0.3102426231,-0.0742492378,-0.1854369342,-0.2399973571,-0.1463766396,-0.2306625545,0.1734178811,0.5155050159,0.1801417321,0.0183062255,-0.3433961272,-0.2121744752,0.196549505,0.068401359,-0.0055206334,-0.0196541045,0.059781421,-0.1644994617,0.1473258436,0.0969853848,0.2417077124,-0.087868154,0.1769556701,-0.2219187915,-0.6492443085,0.396417588,-0.2278424054,-0.1365828365,-0.128093496,0.3268310726,0.2046232969,0.1002063528,-0.6723276377,0.2960059047,0.5039242506,-0.1620292068,-0.0190109964,0.1430853456,-0.123635374,0.0088826027,0.0108755771,-0.0222046953,0.0805404261,-0.2500542402,0.2628282607,-0.3124885261]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2549","title":"Handling unlabeled datasets","comments":"Hi @nelson-liu,\r\n\r\nYou can pass the parameter `features` to `load_dataset`: https:\/\/huggingface.co\/docs\/datasets\/_modules\/datasets\/load.html#load_dataset\r\n\r\nIf you look at the code of the MNLI script you referred in your question (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/multi_nli\/multi_nli.py#L62-L77), you can see how the Features were originally specified. \r\n\r\nFeel free to use it as a template, customize it and pass it to `load_dataset` using the parameter `features`.","body":"Hi!\r\n\r\nIs there a way for datasets to produce unlabeled instances (e.g., the `ClassLabel` can be nullable).\r\n\r\nFor example, I want to use the MNLI dataset reader ( https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/multi_nli\/multi_nli.py ) on a file that doesn't have the `gold_label` field. I tried setting `\"label\": data.get(\"gold_label\")`, but got the following error:\r\n\r\n```\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 989, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 953, in encode_example\r\n return encode_nested_example(self, example)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 848, in encode_nested_example\r\n k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 848, in \r\n k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 875, in encode_nested_example\r\n return schema.encode_example(obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 653, in encode_example\r\n if not -1 <= example_data < self.num_classes:\r\nTypeError: '<=' not supported between instances of 'int' and 'NoneType'\r\n```\r\n\r\nWhat's the proper way to handle reading unlabeled datasets, especially for downstream usage with Transformers?","comment_length":55,"text":"Handling unlabeled datasets\nHi!\r\n\r\nIs there a way for datasets to produce unlabeled instances (e.g., the `ClassLabel` can be nullable).\r\n\r\nFor example, I want to use the MNLI dataset reader ( https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/multi_nli\/multi_nli.py ) on a file that doesn't have the `gold_label` field. I tried setting `\"label\": data.get(\"gold_label\")`, but got the following error:\r\n\r\n```\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 989, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 953, in encode_example\r\n return encode_nested_example(self, example)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 848, in encode_nested_example\r\n k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 848, in \r\n k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 875, in encode_nested_example\r\n return schema.encode_example(obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 653, in encode_example\r\n if not -1 <= example_data < self.num_classes:\r\nTypeError: '<=' not supported between instances of 'int' and 'NoneType'\r\n```\r\n\r\nWhat's the proper way to handle reading unlabeled datasets, especially for downstream usage with Transformers?\nHi @nelson-liu,\r\n\r\nYou can pass the parameter `features` to `load_dataset`: https:\/\/huggingface.co\/docs\/datasets\/_modules\/datasets\/load.html#load_dataset\r\n\r\nIf you look at the code of the MNLI script you referred in your question (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/multi_nli\/multi_nli.py#L62-L77), you can see how the Features were originally specified. \r\n\r\nFeel free to use it as a template, customize it and pass it to `load_dataset` using the parameter `features`.","embeddings":[0.0588133819,0.0781008974,0.1316254437,0.3273362815,0.2503514886,0.1380114406,0.7898842692,-0.1512288302,0.0572320819,0.1499080062,-0.200078249,0.5356727242,-0.2677472532,0.24402228,-0.1644505858,0.4176734388,-0.1134410575,0.2267287523,0.2766293287,-0.2599217296,-0.3965278864,-0.1135896668,-0.325173378,-0.0457837433,-0.4231288731,-0.2310268432,0.1210811734,-0.0139455795,-0.1970751286,-0.4348151386,0.2606198788,0.2651962638,0.0579359941,0.2684318423,-0.000120314,0.0050641368,-0.0486157686,-0.1119301543,-0.4111914337,-0.1604851186,-0.17992194,-0.1166345552,-0.0756473094,-0.3255845606,-0.1712770909,-0.2064837366,-0.0073193642,-0.5275917053,-0.236964345,0.1300067902,0.0282530356,0.129878059,-0.2276054174,0.1804244518,0.0662300587,0.1123861447,-0.0983037725,0.0424691327,-0.0031264429,0.1706990153,0.1761711538,0.456905812,-0.1983845532,0.0251814239,0.2305648625,0.1658895463,-0.2503337264,-0.1352570355,0.0453840084,0.5727470517,0.4895333052,-0.5312574506,-0.8156602979,-0.5682113767,0.5466715693,-0.4145795703,0.0094724102,0.1875493079,0.0612979271,0.2065876871,-0.2948110998,-0.0639906004,-0.1535109729,-0.1351516098,-0.215939641,0.4090186656,0.0057898844,0.083067745,0.0186886974,-0.1346999109,0.1350486726,-0.138262406,-0.0474892966,0.4597667456,-0.1665136963,-0.366247952,-0.1840955913,0.0842086077,0.2371453047,0.1862984598,0.2142308056,-0.1185085326,-0.395050019,0.228852123,0.465693593,-0.1114895418,0.2576128244,0.2515033484,0.4025672972,-0.1563059539,-0.1025376692,-0.0510976352,-0.3554551303,0.1981924921,0.3089798987,0.0072208126,0.0130516663,-0.2379139364,-0.3461011946,0.054074578,-0.2671500444,0.1201147884,0.0662340969,0.1098269001,0.1551239789,0.130673632,0.0043763681,0.0855145678,-0.0975475907,-0.5176511407,0.0880837739,-0.0377777591,0.0948991179,0.0195289087,0.1188480631,0.1252898127,0.2857361734,-0.0127967335,0.0077336165,-0.0752635971,0.1977365017,-0.1479917914,0.1000809968,0.1078649014,0.1397909373,0.2605170012,-0.0957131758,-0.5286998749,-0.0942174643,0.2075319588,-0.0390456878,0.0118883234,-0.1286029965,0.0225009583,-0.4288024902,-0.2167243659,-0.037531402,0.0427329019,-0.024934452,-0.0478496067,0.1455179304,-0.2070304751,0.1521824747,-0.1765888333,0.1779623777,0.3938357532,-0.5988759995,-0.3255690932,-0.0107932687,-0.0681707934,-0.2921360135,-0.0235867389,-0.0219930653,0.1814170927,-0.1385188103,-0.0577694215,0.1128369942,-0.4254541993,-0.2444216013,0.0326566286,0.0170047916,0.0275673475,0.1215592697,0.4517728984,-0.1504065841,0.0214292668,-0.0560304932,0.0455516577,-0.0517212115,-0.362842381,-0.0019084015,-0.1636172831,0.0268481933,0.1302967817,0.0897681341,0.2137241513,0.0989583284,0.2463676631,0.0110974433,-0.3854056299,0.1742380708,0.3612041473,0.3009287417,0.077179864,0.0826647729,-0.2592610717,-0.592498064,0.357568115,-0.3484352529,0.154369399,0.0791183338,-0.5093025565,0.1660781056,-0.1975696087,-0.2076828182,-0.0491877161,-0.0308530536,-0.0516687185,-0.0010149746,-0.0134302462,-0.4805208743,-0.2239594311,-0.0684634596,0.1867066175,-0.2691110373,-0.1177120134,0.4422515333,-0.0155847454,-0.2933373451,0.3634032607,0.0749587417,-0.0475974344,0.2478732616,0.1806419045,0.0792778432,-0.0156236924,0.1070772335,0.2868723571,0.0241860393,0.1035673842,-0.1018624827,0.0068803094,0.0740688369,-0.0478818789,-0.3704842925,0.7585379481,-0.0143596008,0.2423820347,0.0692146122,-0.0360741727,0.2637818754,-0.1088464558,0.0059202481,-0.3193671405,-0.1511460245,-0.0976229236,0.2514329851,0.2600204647,-0.4639050663,0.0723712519,0.1971778274,-0.0341641195,0.1932201684,-0.2416765392,-0.4811052382,0.4233587384,0.0534696691,0.2950209081,0.2030082047,0.2268395871,0.1559764445,0.0888147354,0.1586731672,-0.020764336,0.2215025276,0.0650207102,0.1606153101,0.25936234,-0.0705102384,-0.2032631487,0.0245898329,-0.0898438245,0.1240271926,0.1306619644,-0.3001923859,0.0432797857,-0.4224675596,-0.2937405109,-0.2737027705,0.1452568769,-0.3171314597,0.0007475782,0.145641312,0.1495463252,-0.0781785846,0.0361940861,-0.2981946468,0.2971700132,-0.1557240486,-0.2295411527,0.1275843382,-0.1776760221,-0.1574129462,0.0109988218,0.095933646,-0.2238840908,0.2040944993,0.1271135658,-0.1750091314,-0.0877167657,-0.0940775573,0.1397818178,0.1698424071,0.4188260138,0.3386104107,-0.235151276,-0.013696691,0.4573945701,0.160925433,0.1649491638,0.0267143901,-0.2706840336,-0.0651539937,0.4000677466,-0.1524949223,-0.2732073069,-0.0821761191,-0.103979528,0.1600445658,0.1093073264,0.149368003,0.1068682298,-0.3497598469,0.135965094,0.0897206739,-0.231913358,0.2457119226,-0.3874400854,0.4512012601,-0.1277274936,0.0841623843,0.2434750348,-0.1037124991,0.1409987807,0.1112418845,-0.155796513,0.0534110367,0.4583177269,-0.2154126018,0.1566188037,-0.0838330612,0.0185860302,0.256011039,0.1351924092,-0.2604410052,0.0881984234,0.1130204648,-0.3000995815,0.0162524376,0.232488513,-0.2371097803,0.3170667887,0.2986520231,0.2956102192,-0.0584118478,0.0348767824,-0.1211847961,0.3143197894,0.3201858997,0.0934146494,-0.1946063787,0.296184063,-0.1129747406,0.3518060148,0.1224239841,0.1934314072,-0.0892665833,0.2257009447,-0.3296516538,-0.1333008558,-0.1614063382,-0.0904649943,0.0884622484,-0.009232495,0.1853941381,-0.123632744,-0.1474440843,-0.0547953658,-0.0228386335,0.3190750778,0.0961967707,-0.5007603168,-0.3644496202,0.0005032349,0.1805497706,0.0882219598,0.1723244786,0.1141868234,0.1356215328,0.2444736958,0.231114313,0.3968502879,-0.8019368649,0.0440767556,0.3882579505,0.0623443164,-0.1298643947,0.1688720584,-0.3431433737,0.0576106422,-0.0860664696,0.3440917432,-0.2848518491,-0.0774758533,0.2539089322,0.1106144935,-0.3914368451,-0.2414443046,-0.0032642351,-0.0086060986,-0.0926106498,-0.0796570331,-0.1077553034,0.2246290743,-0.2143976241,-0.0325562246,-0.4636360705,0.3866451979,0.3230821788,0.0573858246,0.0829205886,-0.0989327729,0.2488309443,0.4137302041,0.2953953147,0.5238576531,0.5176036954,-0.1507750303,-0.3284547925,-0.2257456928,-0.0502107069,0.4424321651,-0.0907643139,0.1496494412,0.2319215536,0.1023851782,-0.153636232,-0.0959235206,-0.1006712392,0.0434700549,-0.0950622186,-0.4350956678,-0.5098007917,0.3801496625,0.136225298,-0.0567051917,0.2898118496,0.2649101615,-0.4261862636,0.0274882223,0.3192895651,0.7659423947,-0.1236862317,0.0004804889,-0.2136633396,0.0164637472,0.8572262526,-0.194813326,0.1371896267,-0.5271185637,0.1272506416,-0.083002843,0.0584128127,0.2789562941,0.690320611,0.3886288404,0.3688079119,-0.1227856725,0.1099552289,-0.0724978596,-0.0127376243,-0.0276429076,-0.29114604,-0.0317770094,-0.054518085,0.0289331842,0.071803391,-0.1977015883,-0.0207408294,0.0386900194,0.1972528845,0.1341837645,-0.1853779554,-0.2127626389,-0.236400649,-0.0999077708,-0.2412441224,0.0070270556,0.5078380108,0.4107720256,0.2300405502,-0.102325581,0.1562463641,-0.0384852849,0.2010703832,0.0804535672,0.1842309237,0.4358699322,0.1954389364,-0.2531307936,-0.0107194781,0.0484076664,0.1301898211,0.2571732998,-0.1982814968,0.221161291,-0.0484374277,0.1967833787,-0.26187253,0.2173197865,-0.1928611845,0.003255269,0.1539716274,-0.2409613729,0.0861347988,-0.2422809899,-0.2054116279,-0.1147637293,0.3811560273,0.226390183,0.070886761,0.2734775245,-0.0664198324,-0.0972513482,-0.1872229725,0.470464766,0.2158821672,-0.2648013532,0.3644737005,0.5038281679,0.1376714706,-0.0285635423,-0.0027729755,-0.1975708455,0.0543682948,-0.2784394026,-0.2706344724,-0.1602126211,0.0887570307,-0.0236793272,0.1900788248,0.1855686307,-0.0134274028,-0.1325759888,-0.2978618145,-0.1617359221,-0.0599069335,-0.4192727506,0.0981310755,0.2057622224,0.6343883872,-0.0262719803,0.0019223376,0.0125623532,-0.038064409,0.2053106129,-0.0335737579,-0.3360456526,0.2107453495,0.2821726799,0.1085625514,0.3158654869,-0.2067905962,-0.1548735946,-0.3689013124,0.2481536418,0.2495817095,0.1385821998,0.4728965163,-0.1120224819,0.229837656,-0.6494517326,0.1717785895,0.1408324987,-0.0035494745,0.1042631119,0.2618666887,0.2079256922,0.1574041098,0.0363411754,-0.0172389019,0.0367373452,-0.1150877923,0.2508651614,-0.4141451418,0.2339369208,0.3280203938,0.393759042,-0.0550214574,-0.1032535285,0.0485829078,0.5089664459,0.0508138463,-0.3674024045,-0.1784171611,0.2600318491,-0.1713799983,0.2319699675,0.5777534842,0.2352407426,0.1841718405,-0.2669292092,-0.2425086796,0.108028546,-0.1819589585,0.218917951,0.5804853439,0.0595461279,0.0142012527,0.153760016,0.2075563371,0.3028893769,0.6620522738,0.0744496435,0.5605964661,0.1112168208,-0.237756297,0.0142710172,-0.4676465988,0.102534622,0.0133355828,0.1929287761,0.2917953134,-0.2008613497,0.0400884077,-0.5066401362,-0.1107312217,0.0842201933,0.0371948034,-0.2296021134,-0.1390457302,-0.4124967754,-0.1178517193,0.1028666273,0.0535999984,-0.0389872305,-0.3978669345,-0.2143848836,-0.1525274515,0.0840239599,-0.4460490346,0.1257645935,0.2423612326,0.3388944566,-0.2475620061,-0.1468477696,0.0446668938,0.1613610089,0.3350238204,0.3214452565,-0.1634106338,0.0659049675,0.0908109322,-0.0273048524,-0.3406300247,0.0065458049,-0.0965166762,0.005317606,-0.2625386417,-0.025656838,0.2166795433,-0.0608716495,-0.0266126469,-0.2695067823,0.3386754394,0.0509649701,-0.3194707334,0.32659325,-0.0815227851,-0.3562274575,-0.1618670821,-0.0343674235,-0.0326832868,-0.1692185849,0.2405819893,0.1130356565,0.0635592937,0.1748879105,0.029249806,-0.0438840389,0.1595609784,-0.0691295341,-0.1020608023,-0.0972221121,0.1864404082,-0.7694787979,0.0722595677,-0.040623188,-0.2358599305,-0.0514596403,0.0480815843,-0.0759260282,-0.1725056022,0.031229293,0.1307892352,0.3904410601,0.455370158,-0.1930674314,-0.4410226643,-0.2298396528,-0.1280018985,-0.1229814067,-0.4427229464,-0.1242487207,-0.1068349257,-0.2093497515,0.1240680516,0.2093604803,0.1191808134,-0.0332781561,0.3878779411,0.0790843815,0.4218620658,0.1751775295,0.0334284604,-0.0418238975,0.1080196798,-0.39315328,0.1997748166,0.0557200946,0.1888753474,-0.3027550578,-0.1966567338,-0.197038427,-0.0015008703,-0.1153632849,0.1205095202,-0.4037130475,0.1716375947,-0.5385056138,0.235014528,0.2344378233,0.0084132999,-0.0264050774,0.0564745106,-0.3050249219,-0.1013591811,0.3029195666,-0.4455325902,-0.0240370668,-0.1198956966,0.4738295972,0.3564075828,-0.206001699,-0.5122574568,0.1980585903,0.6573614478,0.0719357282,0.0762585774,-0.1705794781,-0.3097846508,-0.0885075927,-0.1510897279,-0.2408014983,0.0244323816,0.010371129,0.2113201618,-0.1359473169]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2549","title":"Handling unlabeled datasets","comments":"ah got it, thanks!","body":"Hi!\r\n\r\nIs there a way for datasets to produce unlabeled instances (e.g., the `ClassLabel` can be nullable).\r\n\r\nFor example, I want to use the MNLI dataset reader ( https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/multi_nli\/multi_nli.py ) on a file that doesn't have the `gold_label` field. I tried setting `\"label\": data.get(\"gold_label\")`, but got the following error:\r\n\r\n```\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 989, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 953, in encode_example\r\n return encode_nested_example(self, example)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 848, in encode_nested_example\r\n k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 848, in \r\n k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 875, in encode_nested_example\r\n return schema.encode_example(obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 653, in encode_example\r\n if not -1 <= example_data < self.num_classes:\r\nTypeError: '<=' not supported between instances of 'int' and 'NoneType'\r\n```\r\n\r\nWhat's the proper way to handle reading unlabeled datasets, especially for downstream usage with Transformers?","comment_length":4,"text":"Handling unlabeled datasets\nHi!\r\n\r\nIs there a way for datasets to produce unlabeled instances (e.g., the `ClassLabel` can be nullable).\r\n\r\nFor example, I want to use the MNLI dataset reader ( https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/multi_nli\/multi_nli.py ) on a file that doesn't have the `gold_label` field. I tried setting `\"label\": data.get(\"gold_label\")`, but got the following error:\r\n\r\n```\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 989, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 953, in encode_example\r\n return encode_nested_example(self, example)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 848, in encode_nested_example\r\n k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 848, in \r\n k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 875, in encode_nested_example\r\n return schema.encode_example(obj)\r\n File \"\/home\/nfliu\/miniconda3\/envs\/debias\/lib\/python3.7\/site-packages\/datasets\/features.py\", line 653, in encode_example\r\n if not -1 <= example_data < self.num_classes:\r\nTypeError: '<=' not supported between instances of 'int' and 'NoneType'\r\n```\r\n\r\nWhat's the proper way to handle reading unlabeled datasets, especially for downstream usage with Transformers?\nah got it, thanks!","embeddings":[0.0588133819,0.0781008974,0.1316254437,0.3273362815,0.2503514886,0.1380114406,0.7898842692,-0.1512288302,0.0572320819,0.1499080062,-0.200078249,0.5356727242,-0.2677472532,0.24402228,-0.1644505858,0.4176734388,-0.1134410575,0.2267287523,0.2766293287,-0.2599217296,-0.3965278864,-0.1135896668,-0.325173378,-0.0457837433,-0.4231288731,-0.2310268432,0.1210811734,-0.0139455795,-0.1970751286,-0.4348151386,0.2606198788,0.2651962638,0.0579359941,0.2684318423,-0.000120314,0.0050641368,-0.0486157686,-0.1119301543,-0.4111914337,-0.1604851186,-0.17992194,-0.1166345552,-0.0756473094,-0.3255845606,-0.1712770909,-0.2064837366,-0.0073193642,-0.5275917053,-0.236964345,0.1300067902,0.0282530356,0.129878059,-0.2276054174,0.1804244518,0.0662300587,0.1123861447,-0.0983037725,0.0424691327,-0.0031264429,0.1706990153,0.1761711538,0.456905812,-0.1983845532,0.0251814239,0.2305648625,0.1658895463,-0.2503337264,-0.1352570355,0.0453840084,0.5727470517,0.4895333052,-0.5312574506,-0.8156602979,-0.5682113767,0.5466715693,-0.4145795703,0.0094724102,0.1875493079,0.0612979271,0.2065876871,-0.2948110998,-0.0639906004,-0.1535109729,-0.1351516098,-0.215939641,0.4090186656,0.0057898844,0.083067745,0.0186886974,-0.1346999109,0.1350486726,-0.138262406,-0.0474892966,0.4597667456,-0.1665136963,-0.366247952,-0.1840955913,0.0842086077,0.2371453047,0.1862984598,0.2142308056,-0.1185085326,-0.395050019,0.228852123,0.465693593,-0.1114895418,0.2576128244,0.2515033484,0.4025672972,-0.1563059539,-0.1025376692,-0.0510976352,-0.3554551303,0.1981924921,0.3089798987,0.0072208126,0.0130516663,-0.2379139364,-0.3461011946,0.054074578,-0.2671500444,0.1201147884,0.0662340969,0.1098269001,0.1551239789,0.130673632,0.0043763681,0.0855145678,-0.0975475907,-0.5176511407,0.0880837739,-0.0377777591,0.0948991179,0.0195289087,0.1188480631,0.1252898127,0.2857361734,-0.0127967335,0.0077336165,-0.0752635971,0.1977365017,-0.1479917914,0.1000809968,0.1078649014,0.1397909373,0.2605170012,-0.0957131758,-0.5286998749,-0.0942174643,0.2075319588,-0.0390456878,0.0118883234,-0.1286029965,0.0225009583,-0.4288024902,-0.2167243659,-0.037531402,0.0427329019,-0.024934452,-0.0478496067,0.1455179304,-0.2070304751,0.1521824747,-0.1765888333,0.1779623777,0.3938357532,-0.5988759995,-0.3255690932,-0.0107932687,-0.0681707934,-0.2921360135,-0.0235867389,-0.0219930653,0.1814170927,-0.1385188103,-0.0577694215,0.1128369942,-0.4254541993,-0.2444216013,0.0326566286,0.0170047916,0.0275673475,0.1215592697,0.4517728984,-0.1504065841,0.0214292668,-0.0560304932,0.0455516577,-0.0517212115,-0.362842381,-0.0019084015,-0.1636172831,0.0268481933,0.1302967817,0.0897681341,0.2137241513,0.0989583284,0.2463676631,0.0110974433,-0.3854056299,0.1742380708,0.3612041473,0.3009287417,0.077179864,0.0826647729,-0.2592610717,-0.592498064,0.357568115,-0.3484352529,0.154369399,0.0791183338,-0.5093025565,0.1660781056,-0.1975696087,-0.2076828182,-0.0491877161,-0.0308530536,-0.0516687185,-0.0010149746,-0.0134302462,-0.4805208743,-0.2239594311,-0.0684634596,0.1867066175,-0.2691110373,-0.1177120134,0.4422515333,-0.0155847454,-0.2933373451,0.3634032607,0.0749587417,-0.0475974344,0.2478732616,0.1806419045,0.0792778432,-0.0156236924,0.1070772335,0.2868723571,0.0241860393,0.1035673842,-0.1018624827,0.0068803094,0.0740688369,-0.0478818789,-0.3704842925,0.7585379481,-0.0143596008,0.2423820347,0.0692146122,-0.0360741727,0.2637818754,-0.1088464558,0.0059202481,-0.3193671405,-0.1511460245,-0.0976229236,0.2514329851,0.2600204647,-0.4639050663,0.0723712519,0.1971778274,-0.0341641195,0.1932201684,-0.2416765392,-0.4811052382,0.4233587384,0.0534696691,0.2950209081,0.2030082047,0.2268395871,0.1559764445,0.0888147354,0.1586731672,-0.020764336,0.2215025276,0.0650207102,0.1606153101,0.25936234,-0.0705102384,-0.2032631487,0.0245898329,-0.0898438245,0.1240271926,0.1306619644,-0.3001923859,0.0432797857,-0.4224675596,-0.2937405109,-0.2737027705,0.1452568769,-0.3171314597,0.0007475782,0.145641312,0.1495463252,-0.0781785846,0.0361940861,-0.2981946468,0.2971700132,-0.1557240486,-0.2295411527,0.1275843382,-0.1776760221,-0.1574129462,0.0109988218,0.095933646,-0.2238840908,0.2040944993,0.1271135658,-0.1750091314,-0.0877167657,-0.0940775573,0.1397818178,0.1698424071,0.4188260138,0.3386104107,-0.235151276,-0.013696691,0.4573945701,0.160925433,0.1649491638,0.0267143901,-0.2706840336,-0.0651539937,0.4000677466,-0.1524949223,-0.2732073069,-0.0821761191,-0.103979528,0.1600445658,0.1093073264,0.149368003,0.1068682298,-0.3497598469,0.135965094,0.0897206739,-0.231913358,0.2457119226,-0.3874400854,0.4512012601,-0.1277274936,0.0841623843,0.2434750348,-0.1037124991,0.1409987807,0.1112418845,-0.155796513,0.0534110367,0.4583177269,-0.2154126018,0.1566188037,-0.0838330612,0.0185860302,0.256011039,0.1351924092,-0.2604410052,0.0881984234,0.1130204648,-0.3000995815,0.0162524376,0.232488513,-0.2371097803,0.3170667887,0.2986520231,0.2956102192,-0.0584118478,0.0348767824,-0.1211847961,0.3143197894,0.3201858997,0.0934146494,-0.1946063787,0.296184063,-0.1129747406,0.3518060148,0.1224239841,0.1934314072,-0.0892665833,0.2257009447,-0.3296516538,-0.1333008558,-0.1614063382,-0.0904649943,0.0884622484,-0.009232495,0.1853941381,-0.123632744,-0.1474440843,-0.0547953658,-0.0228386335,0.3190750778,0.0961967707,-0.5007603168,-0.3644496202,0.0005032349,0.1805497706,0.0882219598,0.1723244786,0.1141868234,0.1356215328,0.2444736958,0.231114313,0.3968502879,-0.8019368649,0.0440767556,0.3882579505,0.0623443164,-0.1298643947,0.1688720584,-0.3431433737,0.0576106422,-0.0860664696,0.3440917432,-0.2848518491,-0.0774758533,0.2539089322,0.1106144935,-0.3914368451,-0.2414443046,-0.0032642351,-0.0086060986,-0.0926106498,-0.0796570331,-0.1077553034,0.2246290743,-0.2143976241,-0.0325562246,-0.4636360705,0.3866451979,0.3230821788,0.0573858246,0.0829205886,-0.0989327729,0.2488309443,0.4137302041,0.2953953147,0.5238576531,0.5176036954,-0.1507750303,-0.3284547925,-0.2257456928,-0.0502107069,0.4424321651,-0.0907643139,0.1496494412,0.2319215536,0.1023851782,-0.153636232,-0.0959235206,-0.1006712392,0.0434700549,-0.0950622186,-0.4350956678,-0.5098007917,0.3801496625,0.136225298,-0.0567051917,0.2898118496,0.2649101615,-0.4261862636,0.0274882223,0.3192895651,0.7659423947,-0.1236862317,0.0004804889,-0.2136633396,0.0164637472,0.8572262526,-0.194813326,0.1371896267,-0.5271185637,0.1272506416,-0.083002843,0.0584128127,0.2789562941,0.690320611,0.3886288404,0.3688079119,-0.1227856725,0.1099552289,-0.0724978596,-0.0127376243,-0.0276429076,-0.29114604,-0.0317770094,-0.054518085,0.0289331842,0.071803391,-0.1977015883,-0.0207408294,0.0386900194,0.1972528845,0.1341837645,-0.1853779554,-0.2127626389,-0.236400649,-0.0999077708,-0.2412441224,0.0070270556,0.5078380108,0.4107720256,0.2300405502,-0.102325581,0.1562463641,-0.0384852849,0.2010703832,0.0804535672,0.1842309237,0.4358699322,0.1954389364,-0.2531307936,-0.0107194781,0.0484076664,0.1301898211,0.2571732998,-0.1982814968,0.221161291,-0.0484374277,0.1967833787,-0.26187253,0.2173197865,-0.1928611845,0.003255269,0.1539716274,-0.2409613729,0.0861347988,-0.2422809899,-0.2054116279,-0.1147637293,0.3811560273,0.226390183,0.070886761,0.2734775245,-0.0664198324,-0.0972513482,-0.1872229725,0.470464766,0.2158821672,-0.2648013532,0.3644737005,0.5038281679,0.1376714706,-0.0285635423,-0.0027729755,-0.1975708455,0.0543682948,-0.2784394026,-0.2706344724,-0.1602126211,0.0887570307,-0.0236793272,0.1900788248,0.1855686307,-0.0134274028,-0.1325759888,-0.2978618145,-0.1617359221,-0.0599069335,-0.4192727506,0.0981310755,0.2057622224,0.6343883872,-0.0262719803,0.0019223376,0.0125623532,-0.038064409,0.2053106129,-0.0335737579,-0.3360456526,0.2107453495,0.2821726799,0.1085625514,0.3158654869,-0.2067905962,-0.1548735946,-0.3689013124,0.2481536418,0.2495817095,0.1385821998,0.4728965163,-0.1120224819,0.229837656,-0.6494517326,0.1717785895,0.1408324987,-0.0035494745,0.1042631119,0.2618666887,0.2079256922,0.1574041098,0.0363411754,-0.0172389019,0.0367373452,-0.1150877923,0.2508651614,-0.4141451418,0.2339369208,0.3280203938,0.393759042,-0.0550214574,-0.1032535285,0.0485829078,0.5089664459,0.0508138463,-0.3674024045,-0.1784171611,0.2600318491,-0.1713799983,0.2319699675,0.5777534842,0.2352407426,0.1841718405,-0.2669292092,-0.2425086796,0.108028546,-0.1819589585,0.218917951,0.5804853439,0.0595461279,0.0142012527,0.153760016,0.2075563371,0.3028893769,0.6620522738,0.0744496435,0.5605964661,0.1112168208,-0.237756297,0.0142710172,-0.4676465988,0.102534622,0.0133355828,0.1929287761,0.2917953134,-0.2008613497,0.0400884077,-0.5066401362,-0.1107312217,0.0842201933,0.0371948034,-0.2296021134,-0.1390457302,-0.4124967754,-0.1178517193,0.1028666273,0.0535999984,-0.0389872305,-0.3978669345,-0.2143848836,-0.1525274515,0.0840239599,-0.4460490346,0.1257645935,0.2423612326,0.3388944566,-0.2475620061,-0.1468477696,0.0446668938,0.1613610089,0.3350238204,0.3214452565,-0.1634106338,0.0659049675,0.0908109322,-0.0273048524,-0.3406300247,0.0065458049,-0.0965166762,0.005317606,-0.2625386417,-0.025656838,0.2166795433,-0.0608716495,-0.0266126469,-0.2695067823,0.3386754394,0.0509649701,-0.3194707334,0.32659325,-0.0815227851,-0.3562274575,-0.1618670821,-0.0343674235,-0.0326832868,-0.1692185849,0.2405819893,0.1130356565,0.0635592937,0.1748879105,0.029249806,-0.0438840389,0.1595609784,-0.0691295341,-0.1020608023,-0.0972221121,0.1864404082,-0.7694787979,0.0722595677,-0.040623188,-0.2358599305,-0.0514596403,0.0480815843,-0.0759260282,-0.1725056022,0.031229293,0.1307892352,0.3904410601,0.455370158,-0.1930674314,-0.4410226643,-0.2298396528,-0.1280018985,-0.1229814067,-0.4427229464,-0.1242487207,-0.1068349257,-0.2093497515,0.1240680516,0.2093604803,0.1191808134,-0.0332781561,0.3878779411,0.0790843815,0.4218620658,0.1751775295,0.0334284604,-0.0418238975,0.1080196798,-0.39315328,0.1997748166,0.0557200946,0.1888753474,-0.3027550578,-0.1966567338,-0.197038427,-0.0015008703,-0.1153632849,0.1205095202,-0.4037130475,0.1716375947,-0.5385056138,0.235014528,0.2344378233,0.0084132999,-0.0264050774,0.0564745106,-0.3050249219,-0.1013591811,0.3029195666,-0.4455325902,-0.0240370668,-0.1198956966,0.4738295972,0.3564075828,-0.206001699,-0.5122574568,0.1980585903,0.6573614478,0.0719357282,0.0762585774,-0.1705794781,-0.3097846508,-0.0885075927,-0.1510897279,-0.2408014983,0.0244323816,0.010371129,0.2113201618,-0.1359473169]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2548","title":"Field order issue in loading json","comments":"Hi @luyug, thanks for reporting.\r\n\r\nThe good news is that we fixed this issue only 9 days ago: #2507.\r\n\r\nThe patch is already in the master branch of our repository and it will be included in our next `datasets` release version 1.9.0.\r\n\r\nFeel free to reopen the issue if the problem persists.","body":"## Describe the bug\r\nThe `load_dataset` function expects columns in alphabetical order when loading json files.\r\n\r\nSimilar bug was previously reported for csv in #623 and fixed in #684.\r\n## Steps to reproduce the bug\r\n\r\nFor a json file `j.json`,\r\n```\r\n{\"c\":321, \"a\": 1, \"b\": 2}\r\n```\r\nRunning the following,\r\n```\r\nf= datasets.Features({'a': Value('int32'), 'b': Value('int32'), 'c': Value('int32')})\r\njson_data = datasets.load_dataset('json', data_files='j.json', features=f)\r\n```\r\n\r\n\r\n## Expected results\r\nA successful load.\r\n## Actual results\r\n```\r\nFile \"pyarrow\/table.pxi\", line 1409, in pyarrow.lib.Table.cast\r\nValueError: Target schema's field names are not matching the table's field names: ['c', 'a', 'b'], ['a', 'b', 'c']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.10.0-957.1.3.el7.x86_64-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyArrow version: 3.0.0\r\n\r\n","comment_length":52,"text":"Field order issue in loading json\n## Describe the bug\r\nThe `load_dataset` function expects columns in alphabetical order when loading json files.\r\n\r\nSimilar bug was previously reported for csv in #623 and fixed in #684.\r\n## Steps to reproduce the bug\r\n\r\nFor a json file `j.json`,\r\n```\r\n{\"c\":321, \"a\": 1, \"b\": 2}\r\n```\r\nRunning the following,\r\n```\r\nf= datasets.Features({'a': Value('int32'), 'b': Value('int32'), 'c': Value('int32')})\r\njson_data = datasets.load_dataset('json', data_files='j.json', features=f)\r\n```\r\n\r\n\r\n## Expected results\r\nA successful load.\r\n## Actual results\r\n```\r\nFile \"pyarrow\/table.pxi\", line 1409, in pyarrow.lib.Table.cast\r\nValueError: Target schema's field names are not matching the table's field names: ['c', 'a', 'b'], ['a', 'b', 'c']\r\n```\r\n\r\n## Environment info\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-3.10.0-957.1.3.el7.x86_64-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyArrow version: 3.0.0\r\n\r\n\nHi @luyug, thanks for reporting.\r\n\r\nThe good news is that we fixed this issue only 9 days ago: #2507.\r\n\r\nThe patch is already in the master branch of our repository and it will be included in our next `datasets` release version 1.9.0.\r\n\r\nFeel free to reopen the issue if the problem persists.","embeddings":[0.1825091839,0.2316938639,-0.0159273576,0.2105989754,0.329020232,-0.0630036443,0.2235180438,0.4288536906,0.0090349196,0.0226684678,0.0754536316,0.6574710608,0.3253482878,-0.0124667594,-0.0245099887,-0.10777089,-0.0028557745,0.2404144555,-0.0928326845,0.0957933143,-0.3594990075,0.0710984692,-0.169530943,0.1843690276,0.1127448007,0.205004558,0.1421498358,0.1688929945,-0.1974498183,-0.4280341864,-0.1163161322,0.1603664458,0.0409124792,0.1538206637,-0.0001129423,-0.0930031985,0.2057088763,0.044096224,-0.1617198884,-0.4524683654,-0.1871806234,-0.2486564964,0.1546484977,-0.2271186858,-0.0396723188,-0.2239807397,-0.0603111573,0.055517558,0.1781316847,0.0954668894,0.191365093,0.1215380579,0.1931028664,0.1893153936,0.5152660608,-0.0020206263,-0.1111088321,0.2756317854,0.1847001314,-0.0094251689,-0.0758256093,0.0324886329,-0.1814019382,0.0318687186,0.1560178697,0.1483386159,-0.0425271988,0.087413609,-0.099468179,0.0486950763,0.5981535316,-0.4596160352,-0.2934252918,-0.200739786,0.1998674721,-0.2957478464,0.507579565,0.1562162042,0.2286590934,0.0513668843,0.0142116835,0.2856676579,-0.185043633,0.1023253724,-0.1568431407,0.1329457015,0.0515414216,-0.0219508745,-0.1376312226,-0.2529479861,0.3639647365,-0.2273322046,-0.1900892556,0.1349727213,-0.3493430018,0.0503967255,0.135819599,-0.0709700063,0.1523932964,0.1112184078,-0.0516666435,0.0156818945,0.0222560447,0.1730181277,0.2025149465,0.1798478961,0.0475179255,-0.0730152875,0.1967939883,0.0729970485,-0.062958777,-0.0940551385,-0.012366672,-0.195101589,-0.0103183212,0.0408960693,0.0609145984,-0.0840106681,-0.2918557525,0.0429290719,-0.124547407,-0.1908009052,-0.0035287847,0.3132716715,-0.0751944855,0.0949023589,0.0929289013,0.2363107204,0.0139913876,-0.0071011106,-0.0657398924,-0.1489697695,-0.0229369029,0.115148738,0.216383785,0.1122139841,0.1306640655,0.3509099782,0.0715359002,-0.0543669574,-0.047882814,-0.1265518814,-0.1324493438,0.4157491028,0.0547996126,0.1828192919,-0.0814390033,-0.1750667989,-0.2666810155,0.3741872311,-0.4962205291,-0.1093590036,-0.0579978302,0.2085763216,-0.1713588089,-0.1254500598,-0.6543222666,-0.1421352774,0.0294263437,-0.2914305627,-0.0671935529,-0.1449342221,0.1742371023,-0.2461764216,-0.1334021091,0.0312421825,-0.5555409789,0.0891386792,0.090237312,0.0520497598,-0.0067552254,-0.0432028584,-0.1329811662,0.1114545092,-0.0728556961,0.4425647259,0.0826909915,-0.2095859349,-0.3545516431,0.100312449,-0.18328996,0.1564613581,0.0873684362,0.0080047539,0.1473624557,0.0594875179,0.086240083,0.3296462893,0.1209139526,0.3840726018,-0.1493302286,-0.1144720092,0.1772043109,0.2748941481,-0.3706638515,0.0561122,0.1017419025,-0.3683663011,0.1644266546,-0.2329896688,-0.2976112962,0.2651005387,0.138519451,0.1046695411,0.292607367,0.0249064788,-0.4560880363,0.2608802319,0.0239530224,-0.0398318134,-0.5410388112,-0.1111306474,-0.022122426,0.0373545773,-0.2153449506,0.4330096543,0.1082401946,0.0037269872,-0.2064828724,-0.0993746668,-0.0000993488,0.0159663931,-0.1661820263,0.0472490974,-0.3866334856,0.454411447,-0.2604219317,-0.1414031982,0.1588269919,0.0324664339,0.2302994281,-0.2600081265,-0.145140484,0.2956563532,0.2853767574,0.0675241351,-0.2429714054,0.0512437634,-0.0149855288,0.0517478958,-0.1420660764,0.132409662,0.0144065954,-0.0093706083,-0.0859903172,0.7995302677,-0.0237492621,0.2128822505,0.0033282584,-0.3676089942,0.3820781112,0.0352012217,0.0164426453,-0.0694161803,0.3017965853,0.5649797916,-0.1108657792,0.2574886978,-0.360769093,0.1042413488,0.941155076,-0.125324294,-0.1056051776,0.0731294304,0.2090286911,0.0162601694,0.0963968784,-0.1628759503,0.3760441542,0.1892888546,-0.6180878282,-0.0649496317,0.0113626365,-0.1525395066,0.0734227449,-0.0244072117,0.2099510282,0.398982197,0.4225431979,0.0975220725,-0.389383018,-0.1733162105,0.1060144082,0.0876098499,-0.6017731428,0.0633966029,-0.3623118103,0.0237795599,-0.1168357581,-0.2730719745,-0.202742666,-0.0075551504,0.0238268878,0.265352428,-0.1859909743,0.2071102113,0.1384101808,0.0389135219,0.0704559311,-0.3349529803,-0.5142762065,-0.0779049844,-0.177746594,0.0132324751,0.1084139347,0.0567992218,0.0161678959,-0.1116787121,-0.0580410548,0.0624616817,-0.5082748532,-0.0208232459,-0.2801716924,0.0598618537,0.0490928814,0.2172631621,0.0719870105,-0.3592016399,0.2828599215,0.1679742932,-0.4280129969,0.5928493738,-0.1754595935,-0.2120300084,-0.1437628865,-0.4086635113,0.1344608366,-0.1835444719,0.257566154,0.2847779989,0.1752562523,-0.1098787487,-0.1036997512,-0.3340315521,-0.0910992697,0.0293778107,-0.0326463729,-0.0653820187,0.3305379152,-0.1414422691,-0.2782587409,-0.0305189118,-0.0579182319,-0.3105297685,0.0535399392,-0.3014820814,-0.0857230425,-0.0933257565,0.4113368094,0.061827153,-0.1718056351,0.1675982773,0.1254292279,-0.0224112235,-0.0237901509,-0.0857908353,-0.0717317909,0.3714417517,0.1108713374,-0.1400422752,0.6356081367,-0.0889731348,0.2890284061,-0.2067594528,0.0643243119,0.22384049,-0.1482941955,-0.0879818723,-0.2037138343,0.1653652936,-0.0510773435,-0.0864713416,0.2216326296,0.2641670406,-0.0263777953,-0.0538219139,0.1418453157,0.2738384604,-0.4508648515,-0.0728435516,0.1599713117,-0.0060576671,-0.0755466446,0.125087589,0.0617544279,-0.1885136068,-0.2236435264,0.0674155429,-0.0334339887,-0.1557301134,-0.2595062256,-0.3962624371,-0.3487423956,0.0122313164,0.0061358931,0.1084262058,0.1493601948,0.2657660544,0.1275406927,0.251495719,-0.2238464206,0.7247189879,0.1659289896,0.2741288841,0.1798027009,0.2624098659,-0.2966653407,-0.1868926287,-0.1661119759,0.0253556632,0.0108434549,0.5456570387,-0.4508349895,-0.0624197125,0.0905571431,0.1579536647,-0.1082998365,-0.466344893,-0.4540721178,-0.1535979956,-0.0223283637,-0.1013708338,-0.1307668686,0.2807715535,-0.5067601204,-0.1119935066,0.2080164701,-0.0572818294,-0.1946265996,0.2885277569,0.297103256,-0.0922506005,0.1950149536,0.0013473763,-0.0363157615,0.2976644039,0.497046262,0.1947048753,-0.5289745331,-0.1166008636,-0.2787715793,-0.0588415973,-0.0205584057,-0.1162029207,0.1930961907,-0.2420474738,-0.0785324946,0.0164960716,-0.3432072103,0.6345765591,-0.2099701017,-0.1358886659,-0.3663291037,0.6602265239,-0.121286884,-0.0563302897,0.1938370615,0.3561738431,-0.3036479056,0.4781931639,0.0364380181,0.6812087893,-0.1965676546,0.0833839178,0.6457627416,-0.2965186834,0.2771173716,-0.5008105636,0.2874607444,-0.7126541734,-0.2874288559,-0.0656849146,-0.1931163818,0.0727455243,0.5663932562,0.0660878047,0.0732796267,-0.2345211357,0.2931254506,0.0699239224,0.1652540118,0.0377009436,-0.0289714653,-0.5482338667,0.037354596,-0.081094183,-0.2535057068,-0.0766954795,0.0609590337,-0.1117840633,-0.2436929941,0.1308026314,0.3869266808,0.0226173494,0.0508018062,0.2855570912,0.0385333002,0.088800557,0.2166153342,-0.2158967555,-0.0013294951,0.0236454532,0.2722308338,0.3553789854,0.2548782527,0.0919452459,0.1086064354,0.06120006,-0.0214268863,-0.1499882489,-0.1036273167,0.0034807255,-0.2005111277,-0.090992108,0.2189017236,-0.1750911623,-0.2435913086,-0.1304230541,-0.0996214896,0.3487404883,-0.1398784965,0.1851748824,-0.1517706066,0.1083834767,0.3658823073,-0.3383458257,-0.2280916125,-0.0779541731,0.4607666731,0.0233240165,0.2834366262,0.4499661624,0.2891089618,-0.2189537287,-0.0851855949,-0.0400452875,0.3467143178,-0.5800539851,-0.0313578732,0.0554741509,0.1587134302,0.1733552814,0.3075587749,-0.0680432543,-0.2246256769,0.0093908291,-0.3348077238,0.2595437169,0.2816804647,0.3085287511,-0.211473316,0.3378100395,0.2253371477,-0.2173426747,-0.0767680779,-0.2654393017,0.0477251485,-0.0196577292,0.2062471658,-0.0972455442,0.1512371451,0.0935845152,0.0041225306,0.3094446063,0.0137115298,-0.0044119088,-0.2657625675,-0.0569331385,0.0078866715,0.0833401084,0.1785647869,-0.2587290704,-0.414953202,-0.1024875417,-0.0685172826,0.0444543771,0.1058640629,0.0996871963,0.1753333062,0.2332048416,0.1407747716,-0.0602740385,0.1024987996,-0.4341541827,0.0901339576,0.1616143137,0.3027602732,-0.3599027991,0.2037092,-0.4562960267,0.0721853673,-0.1324938685,0.2397702187,0.5103180408,-0.5786815882,0.1536405534,-0.0345996171,0.4610272348,0.3679774106,-0.3090326488,-0.3603273034,-0.0108076269,0.2446335554,-0.2465851605,-0.4968979061,-0.0314646587,0.0577031411,-0.1892639548,-0.1650169641,0.1161840037,0.0026959619,0.0129008563,-0.0198732149,0.4421185255,0.0950346813,0.0347701125,0.4580205977,0.0337548368,0.1267139018,0.3582769036,0.2588685751,-0.1064045206,0.1960481852,0.3231471479,-0.1048637927,-0.0939709246,0.0039251619,-0.1641119868,-0.4030667841,0.0611235276,0.147383973,0.0829796642,-0.0616877526,0.4316123724,0.0788903013,0.1544599235,-0.0432288386,-0.2516341209,-0.0408380553,0.0137428194,0.0829189271,-0.1507077217,-0.1326127052,-0.0189146847,-0.0031861516,-0.3623932004,0.0942898989,0.0717238337,0.1786758751,-0.073544398,-0.6868671179,0.4297246039,-0.0577261075,0.3208509982,-0.0260377023,0.37364012,-0.0392372012,-0.1635448784,0.2002722323,-0.2996743619,0.1698544025,0.3468171656,0.1554803252,-0.331259042,-0.2791924775,-0.122140564,0.0515566804,0.27277565,-0.0942548215,0.119674094,0.1257961392,0.2250459045,-0.1804128289,-0.0053906031,0.4681790769,0.1990907192,0.0463554524,0.1772397012,-0.1658030003,-0.4063176811,0.090867497,-0.2096222788,-0.0249380749,-0.1211010516,0.0041882275,0.3609237969,-0.0596130826,-0.1571272761,0.1015829742,0.1396051645,0.2515595555,0.1618898064,-0.0134948986,-0.203197062,-0.1123998761,-0.3537177145,0.2708966732,-0.0409103073,-0.0233560614,0.5161078572,0.1403433681,0.0534569658,-0.1625390202,0.0083469776,0.0214836635,-0.2109867632,0.290402323,-0.2972913086,-0.1069241762,-0.058204066,0.0654288679,-0.4646750689,-0.1339449137,0.1110644937,0.0387885123,0.0987861156,-0.4376764297,0.1262200773,0.1070031896,-0.0342797227,0.2831784189,-0.1299281269,0.3202366531,-0.0308626462,0.0992102176,0.0209310967,-0.3393149972,-0.1914260983,0.3711230457,0.0069414587,0.2236998677,0.2191451937,0.0719184503,-0.2390631288,0.0160303265,-0.3385501504,-0.1205538362,-0.2732465863,0.1303788573,-0.0079321312,0.0492310226,-0.3172138631,-0.041233398,0.1546374261,0.1962885261,-0.0877081007,-0.2779091597,0.1896759868,-0.1386563629,-0.1206435263,0.2642492056,0.2381527573,-0.2384243459,0.013423522,-0.2246548682,0.0950940028,0.3887329996,-0.0772806779,-0.1293906718,0.2517741024,0.1122353747,0.1122589111,-0.1991973519,0.1574222594,0.1900288016,-0.2249364108,-0.1681822836,-0.0726456493]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2547","title":"Dataset load_from_disk is too slow","comments":"Hi ! It looks like an issue with the virtual disk you are using.\r\n\r\nWe load datasets using memory mapping. In general it makes it possible to load very big files instantaneously since it doesn't have to read the file (it just assigns virtual memory to the file on disk).\r\nHowever there happens to be issues with virtual disks (for example on spot instances), for which memory mapping does a pass over the entire file, and this takes a while. We are discussing about this issue here: #2252 \r\n\r\nMemory mapping is something handled by the OS so we can't do much about it, though we're still trying to figure out what's causing this behavior exactly to see what we can do.","body":"@lhoestq \r\n## Describe the bug\r\nIt's not normal that I have to wait 7-8 hours for a dataset to be loaded from disk, as there are no preprocessing steps, it's only loading it with load_from_disk. I have 96 cpus, however only 1 is used for this, which is inefficient. Moreover, its usage is at 1%... This is happening in the context of a language model training, therefore I'm wasting 100$ each time I have to load the dataset from disk again (because the spot instance was stopped by aws and I need to relaunch it for example). \r\n\r\n## Steps to reproduce the bug\r\nJust get the oscar in spanish (around 150GGB) and try to first save in disk and then load the processed dataset. It's not dependent on the task you're doing, it just depends on the size of the text dataset.\r\n\r\n## Expected results\r\nI expect the dataset to be loaded in a normal time, by using the whole machine for loading it, I mean if you store the dataset in multiple files (.arrow) and then load it from multiple files, you can use multiprocessing for that and therefore don't waste so much time. \r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Ubuntu 18\r\n- Python version: 3.8\r\n\r\n\r\nI've seen you're planning to include a streaming mode for load_dataset, but that only saves the downloading and processing time, that's not being a problem for me, you cannot save the pure loading from disk time, therefore that's not a solution for my use case or for anyone who wants to use your library for training a language model. ","comment_length":121,"text":"Dataset load_from_disk is too slow\n@lhoestq \r\n## Describe the bug\r\nIt's not normal that I have to wait 7-8 hours for a dataset to be loaded from disk, as there are no preprocessing steps, it's only loading it with load_from_disk. I have 96 cpus, however only 1 is used for this, which is inefficient. Moreover, its usage is at 1%... This is happening in the context of a language model training, therefore I'm wasting 100$ each time I have to load the dataset from disk again (because the spot instance was stopped by aws and I need to relaunch it for example). \r\n\r\n## Steps to reproduce the bug\r\nJust get the oscar in spanish (around 150GGB) and try to first save in disk and then load the processed dataset. It's not dependent on the task you're doing, it just depends on the size of the text dataset.\r\n\r\n## Expected results\r\nI expect the dataset to be loaded in a normal time, by using the whole machine for loading it, I mean if you store the dataset in multiple files (.arrow) and then load it from multiple files, you can use multiprocessing for that and therefore don't waste so much time. \r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Ubuntu 18\r\n- Python version: 3.8\r\n\r\n\r\nI've seen you're planning to include a streaming mode for load_dataset, but that only saves the downloading and processing time, that's not being a problem for me, you cannot save the pure loading from disk time, therefore that's not a solution for my use case or for anyone who wants to use your library for training a language model. \nHi ! It looks like an issue with the virtual disk you are using.\r\n\r\nWe load datasets using memory mapping. In general it makes it possible to load very big files instantaneously since it doesn't have to read the file (it just assigns virtual memory to the file on disk).\r\nHowever there happens to be issues with virtual disks (for example on spot instances), for which memory mapping does a pass over the entire file, and this takes a while. We are discussing about this issue here: #2252 \r\n\r\nMemory mapping is something handled by the OS so we can't do much about it, though we're still trying to figure out what's causing this behavior exactly to see what we can do.","embeddings":[-0.4744633436,-0.365211457,-0.0910320878,0.5078071952,0.2949647009,0.0777809173,0.0872955769,0.2941164076,0.5636069775,0.1137160286,0.1030264795,0.3100511134,-0.0364946872,-0.1310184747,-0.0510295592,0.0120529942,0.2169391513,0.0340653881,0.1016836837,-0.1622139961,-0.167964071,0.1942941397,-0.1397533417,-0.2468277961,-0.2608512938,-0.1675216854,0.0360899568,0.1326748133,0.0141112888,-0.23193717,0.082776159,-0.0234671514,0.1747191697,0.4756506383,-0.0001066134,-0.0308672152,0.1014186889,0.0678356439,-0.3980649114,-0.0591615625,0.2369137555,-0.2943418324,0.0769659132,-0.1994225979,-0.0850722864,-0.0845536739,-0.1045388654,-0.4806056023,0.3054249883,0.0995374694,0.2420826107,0.2846265435,-0.2973237038,-0.0019262831,-0.0428477451,0.0577618815,-0.2270344049,0.2517683208,0.4787022769,0.1249555349,-0.1382647008,0.3056685925,-0.0878099948,0.1649827063,0.4531141818,0.0637824163,0.0809570104,-0.2099038512,-0.0163599234,0.1667419672,0.4594858885,0.0003169189,-0.2677719593,-0.275803715,0.0870442018,-0.2303051054,0.3020731807,0.2804101706,-0.1732685119,0.1456238031,-0.0725262165,-0.1091924235,-0.0347356312,0.120602265,0.2601843178,0.0220489316,-0.1033077836,0.0419218689,0.3998857737,0.1031062454,0.1836707592,-0.2756842375,0.1813414097,0.4096225202,-0.4817183316,0.0225810446,-0.0420286059,0.1797642559,-0.0553691722,0.1661653519,0.257504195,0.1878594309,-0.1930667013,0.1321935654,0.1726056337,0.2541685998,-0.2310456932,0.066409573,0.277910471,0.1457004845,-0.2281666994,-0.0095429411,-0.265899986,-0.1400371045,0.3268366456,-0.2910885811,-0.1571619064,-0.1417412162,-0.0625425503,0.0716491267,0.1459768862,-0.1332921833,0.2146691382,0.3002417684,-0.2892740369,0.5329866409,0.0812344179,-0.1594093293,-0.3561357856,-0.0713704526,-0.1798520088,0.0125497803,-0.2574570477,0.182236284,0.2045733333,-0.1747212112,0.1689618081,0.127179116,0.178788498,-0.2528987825,0.2386086136,-0.5147583485,-0.0775189623,0.1764651835,-0.0012884715,0.2229483724,0.0669550672,0.010185034,-0.3426543772,0.1036737561,-0.3803787231,-0.4763144553,-0.1604786217,0.2296297103,0.0005404514,0.0643452406,-0.529604435,0.063398838,-0.0359070562,0.0567388125,-0.2487528175,-0.0802192762,-0.3271106184,0.0043929671,0.3452154696,0.3170844018,-0.4786154628,0.108867012,-0.2045880705,0.1125852689,0.3233202696,0.6517827511,-0.3402391374,0.2586215138,-0.2039432526,-0.0241454113,0.152699098,-0.2088017762,-0.3435871601,0.680685997,0.0049908007,0.0012142297,0.098638393,0.3005006909,0.0809623003,-0.0457398221,0.1633227468,0.5188064575,0.1197433919,0.3266189396,-0.3972658217,-0.2258613855,0.0233361088,0.5041811466,-0.328664422,-0.1150424033,-0.0572323948,0.3690591156,0.3817795217,-0.1068834737,-0.0421333611,0.3159361482,0.0826601237,0.0314597301,0.0057274178,-0.0402595028,-0.213591814,0.369096905,0.2848334908,0.0523241125,0.1105625033,-0.0959375575,-0.1257336289,-0.0045078266,-0.2928821146,0.1405548304,0.0251154341,0.0356315821,0.0581365228,-0.12447384,-0.228387922,0.2344729304,-0.3492327034,-0.1504371166,-0.2534069121,0.0683864951,-0.0063446853,0.0837816447,-0.0515191182,-0.0774432421,-0.0224777665,-0.0123394961,-0.2500486374,0.3335142136,0.1480024606,0.4500734806,0.1707649082,0.0197047815,0.1539163291,0.1502105892,0.2900156975,0.0290655233,0.3547214866,-0.2086054385,-0.1581657529,0.3160486221,-0.0086462023,0.3458160162,0.1937657744,-0.2984397411,0.3042435944,0.1253898293,-0.0200541057,-0.0765722021,0.4769965112,0.1933380216,0.6596895456,0.1480723023,-0.2960383594,-0.150371775,0.4899028838,0.0065293098,-0.0681883842,0.2815354764,-0.0161944125,-0.2959465384,-0.0236283336,0.1372419745,0.4383119643,0.2162736505,0.02002188,-0.0095435074,-0.0210167281,-0.1808339357,0.2527481019,-0.0400845967,0.5054352283,0.1015432924,-0.0506686829,-0.0109918648,-0.4174702168,-0.259871006,-0.0033295986,0.1630130112,-0.1468356848,0.2073659003,-0.0957700312,0.1224526465,-0.3151164353,0.2789246142,-0.0359257609,-0.0728959888,-0.1824884862,0.2659865916,0.107479021,-0.0019792265,-0.089438498,0.2249300331,-0.0581273213,-0.2849130034,-0.2735427022,-0.0603667796,-0.3348503113,0.0103034228,0.3252160549,0.0741825551,0.3003816903,-0.0823547244,-0.2148215026,-0.1328992695,0.0521079823,0.017144816,-0.0165992621,0.2699772716,-0.1210673079,0.1912131608,-0.0505453683,-0.0959266648,0.2041478008,-0.2538374364,-0.0389258564,-0.0102831358,-0.073555328,-0.0037473917,-0.0207954496,-0.4478420317,-0.3565234244,-0.3237344921,0.0292117689,-0.0933991149,0.1269902885,-0.1361833364,0.1047573388,0.0641794801,0.1965306103,0.1443864554,-0.2523111701,-0.5606966615,0.3110788465,0.0203318261,-0.2803843617,0.0262683649,0.2196990252,0.2357210368,0.0345379338,-0.6052536964,-0.0271210689,-0.3552590907,0.021931354,-0.3490780592,0.0371208824,0.0630600676,-0.0853773654,-0.0792649984,0.1527556181,-0.1620277911,-0.0691517517,-0.0767719001,0.077483885,-0.1906053871,0.2672797143,0.0426366776,0.5003497601,0.1409885734,0.1045719534,0.3580861688,0.1680099517,0.1199369058,-0.3736137152,-0.130430609,0.0154236499,-0.3150401711,-0.3742136061,0.2172494233,0.0767825246,-0.3954347968,-0.174583897,-0.08634229,-0.1705093831,-0.1927301586,0.1544846296,-0.2539516389,0.1625977308,-0.0412333049,0.1380117387,-0.0679569021,-0.0923270658,-0.0335166864,-0.0306000207,0.0900248885,-0.1337501854,-0.3750216961,0.0986416414,-0.4460467696,0.1425751299,-0.2722061872,0.186204046,-0.0156177329,0.015772067,0.2318134606,-0.0727579594,0.7050452828,0.102679573,0.1966913342,0.0261168759,-0.2520487309,-0.4085820615,0.0178833548,-0.1041604578,-0.0642227679,0.0086273532,0.5170795321,-0.145253256,-0.2341355234,-0.2064348906,0.4217880964,-0.1872287393,-0.3244764507,-0.2321756333,-0.2543639541,-0.3492271602,0.1607441753,0.0668263286,0.3056792617,0.1028153673,-0.0273499899,-0.0050509972,0.2019695789,0.2199247628,0.0266661197,0.2480990589,-0.0263269562,0.4335161746,-0.0052601323,0.0647027344,-0.0664953291,0.5060542822,-0.1639938951,-0.2544701099,0.0545277819,-0.0748027638,0.1697997302,0.4769681692,0.0424272604,-0.0674119666,0.2645461857,0.2375968397,-0.3703552783,0.0503737219,0.3129148483,0.1920322925,-0.1302802116,-0.6680116057,0.3885315359,0.2552737594,0.0441932529,0.1103970557,-0.2265155464,-0.1272187233,0.5779409409,0.2083392143,0.963752389,-0.1362684816,0.2553078234,-0.0221573133,0.1892080754,0.1544011086,-0.4441037178,-0.0409279987,-0.3033842146,-0.201082781,0.0773726851,-0.0669563934,0.1298740804,0.4100768864,-0.0162325595,0.3677227795,-0.0204636548,-0.3300700486,-0.3168064952,0.5213741064,-0.3694129884,-0.6282692552,-0.6542042494,0.16493617,-0.1258344203,0.0604873374,-0.0272845849,0.2898995876,0.1325596273,-0.157210201,-0.1678628474,0.1088898405,-0.171956107,0.1888804734,-0.307700038,-0.2053186297,0.0581100173,0.1794909239,0.0225232262,0.1846447289,-0.0376792885,0.4627088904,-0.1927570552,0.2538876534,0.1208069772,-0.3102424741,0.3044306636,-0.2151000351,-0.1387359351,-0.2026305199,-0.2663205862,-0.206663534,0.195740059,0.036147397,0.125794366,0.0684804171,-0.6111649275,0.1892088205,-0.0238641296,-0.2318448871,0.1828299612,0.2452365905,-0.077387251,0.3932022452,0.0544845313,-0.3213698268,-0.0695642009,0.3503504395,0.185396418,-0.1453484148,0.4302146733,0.2795015872,-0.2068384439,-0.2503645122,0.1676186919,0.0483137555,-0.4503316581,0.2712937891,0.1372842193,-0.0482425578,0.0251611248,0.0064540645,0.2230455875,-0.1812041104,-0.1374008507,-0.309538424,-0.1729569882,-0.0386961959,0.2079716474,0.0943128094,0.1608486772,-0.2859983444,-0.0956854075,0.0522897728,-0.3249441683,0.0849162042,-0.2763910592,-0.0223495364,-0.1640485525,-0.0449341908,0.3319895864,-0.0107507491,0.1544734985,0.0064231325,-0.1199088991,-0.242011264,-0.2305976152,0.1225053892,-0.1347487122,0.0003669955,0.1364346147,-0.4455792308,-0.1300946921,-0.3052485287,-0.0121419262,0.0698814765,0.0120547377,-0.1074533463,-0.037240807,0.0851489604,-0.1263146549,0.1323620677,-0.1252923757,0.313264966,0.1154140234,0.010453823,0.3787934482,-0.2101465911,-0.4721017182,-0.1458455175,0.0856334269,0.3004267812,0.2204204947,-0.0682259575,0.1085130945,0.3402151465,0.3384627402,0.2614001632,-0.0121019641,-0.1884236336,0.0401678495,0.2403880656,-0.1858049333,-0.3267370164,-0.1238302961,-0.2268266976,-0.0059494623,0.088468425,0.0675780475,0.4614363909,-0.2952952385,0.0183461811,0.3483889401,-0.319653213,0.0307650268,0.2340436876,0.1609738469,0.1844977289,0.1374912262,0.0627418458,0.1188931763,0.7941011786,-0.1557662487,0.3140955269,0.1635064781,0.5316987634,-0.2107441574,-0.3584108949,0.1603096575,0.2819747031,-0.2023050189,0.1567443311,-0.0790773854,0.1474893391,-0.3605521917,-0.2502915859,-0.1841076314,0.0501419231,-0.1614351571,-0.0824110508,-0.034497723,0.0899179354,-0.1224070266,0.3268423975,-0.2461451143,0.0680851787,0.0277387537,0.3039017618,0.0580049418,-0.1363678873,-0.0264327992,-0.1593113989,0.2774821818,-0.2918614149,0.0870017856,-0.0817563459,-0.1345005482,-0.1333439201,0.0755703151,0.2274031043,0.4301581085,0.0833781287,0.1108024046,-0.03221526,0.0555050597,-0.1291928142,0.2067604065,0.1577795297,0.5026854277,0.3084705472,0.0885601267,-0.2033230066,-0.0085542593,-0.1716056466,0.3376588821,-0.3392180502,0.3443558216,0.0564855486,-0.1245845556,-0.4495334029,0.3309647739,-0.2590315938,0.0737504661,0.3909416795,-0.0353820063,-0.0522631668,-0.2203856409,0.0913043395,-0.0858968943,0.7191553712,0.2900618017,0.0001212094,-0.3954747021,-0.0362261645,-0.7739748359,-0.030782992,-0.1820878983,0.2303028852,0.0497662574,0.3654824197,0.1856742948,0.2261641771,-0.091274187,-0.1180469841,-0.1717384309,0.1461450607,-0.1976796091,-0.0900552049,-0.1394363642,0.1651573628,-0.1136929989,-0.5302267075,0.1717645824,0.0819090605,-0.0673016086,-0.1615783274,-0.1335855424,-0.0131775029,0.2143162638,0.4995814562,0.054426685,0.0735202059,-0.1049974635,-0.1087205559,-0.0519076698,-0.0537870303,-0.0982229039,0.1696146131,0.0841461718,0.2834580541,-0.0939562172,0.2206939757,-0.3580946028,0.2006818652,-0.2234624922,-0.3017807901,-0.250353694,-0.1162723228,0.1512673497,0.0865322426,0.1971613467,0.3609008789,0.0298659317,-0.1782892942,-0.3926269114,-0.1595789045,0.2519140542,0.01678898,-0.2076327056,-0.0553050265,0.2613335848,0.1508577913,-0.068516463,-0.5005113482,0.0650299489,0.2177363038,-0.0242247134,-0.197094202,0.3102473915,-0.0516865812,0.0360893048,-0.1081531867,0.3465841413,-0.1472852081,-0.3784306943,0.0138742123,-0.3208793402]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2547","title":"Dataset load_from_disk is too slow","comments":"Okay, that's exactly my case, with spot instances... Therefore this isn't something we can change in any way to be able to load the dataset faster? I mean, what do you do internally at huggingface for being able to use spot instances with datasets efficiently?","body":"@lhoestq \r\n## Describe the bug\r\nIt's not normal that I have to wait 7-8 hours for a dataset to be loaded from disk, as there are no preprocessing steps, it's only loading it with load_from_disk. I have 96 cpus, however only 1 is used for this, which is inefficient. Moreover, its usage is at 1%... This is happening in the context of a language model training, therefore I'm wasting 100$ each time I have to load the dataset from disk again (because the spot instance was stopped by aws and I need to relaunch it for example). \r\n\r\n## Steps to reproduce the bug\r\nJust get the oscar in spanish (around 150GGB) and try to first save in disk and then load the processed dataset. It's not dependent on the task you're doing, it just depends on the size of the text dataset.\r\n\r\n## Expected results\r\nI expect the dataset to be loaded in a normal time, by using the whole machine for loading it, I mean if you store the dataset in multiple files (.arrow) and then load it from multiple files, you can use multiprocessing for that and therefore don't waste so much time. \r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Ubuntu 18\r\n- Python version: 3.8\r\n\r\n\r\nI've seen you're planning to include a streaming mode for load_dataset, but that only saves the downloading and processing time, that's not being a problem for me, you cannot save the pure loading from disk time, therefore that's not a solution for my use case or for anyone who wants to use your library for training a language model. ","comment_length":45,"text":"Dataset load_from_disk is too slow\n@lhoestq \r\n## Describe the bug\r\nIt's not normal that I have to wait 7-8 hours for a dataset to be loaded from disk, as there are no preprocessing steps, it's only loading it with load_from_disk. I have 96 cpus, however only 1 is used for this, which is inefficient. Moreover, its usage is at 1%... This is happening in the context of a language model training, therefore I'm wasting 100$ each time I have to load the dataset from disk again (because the spot instance was stopped by aws and I need to relaunch it for example). \r\n\r\n## Steps to reproduce the bug\r\nJust get the oscar in spanish (around 150GGB) and try to first save in disk and then load the processed dataset. It's not dependent on the task you're doing, it just depends on the size of the text dataset.\r\n\r\n## Expected results\r\nI expect the dataset to be loaded in a normal time, by using the whole machine for loading it, I mean if you store the dataset in multiple files (.arrow) and then load it from multiple files, you can use multiprocessing for that and therefore don't waste so much time. \r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Ubuntu 18\r\n- Python version: 3.8\r\n\r\n\r\nI've seen you're planning to include a streaming mode for load_dataset, but that only saves the downloading and processing time, that's not being a problem for me, you cannot save the pure loading from disk time, therefore that's not a solution for my use case or for anyone who wants to use your library for training a language model. \nOkay, that's exactly my case, with spot instances... Therefore this isn't something we can change in any way to be able to load the dataset faster? I mean, what do you do internally at huggingface for being able to use spot instances with datasets efficiently?","embeddings":[-0.3196368515,-0.5587565303,-0.0620263182,0.516035676,0.232860297,0.0203786287,0.1246295944,0.1379356384,0.5566672683,0.1127059087,-0.0614803024,0.3583331406,-0.0153432656,0.2113305628,0.0768192783,-0.0829387158,0.2018542588,0.0174787268,0.0967767015,-0.0334709622,-0.0602658466,0.1847729385,-0.1777219623,-0.1801541001,-0.3432554603,-0.1481812596,0.0169770177,0.1757047623,-0.091173932,-0.3421683908,0.174323976,0.0779763013,0.1314018667,0.5717477202,-0.0001136536,-0.0459838435,0.1177239493,0.0748291686,-0.3427474797,0.0368377455,0.0366960578,-0.2794430852,0.152512759,-0.1323544234,-0.1332791746,0.1293530762,-0.0407350846,-0.3339373171,0.3590659499,-0.0157278776,0.1833483726,0.5802416205,-0.3182553649,-0.0626381412,-0.0680631995,0.0878693536,-0.2041863352,0.2897658646,0.4282694161,0.137509197,-0.1503990889,0.4081258178,-0.0401742682,0.1290538311,0.554825604,0.1066801548,-0.090983063,-0.1906510293,-0.1276057512,0.274615258,0.4216815233,-0.0359468348,-0.3770565987,-0.4575647116,0.0806800872,-0.3198552728,0.2432198226,0.1389402896,-0.0813618898,0.123814553,-0.1747931242,-0.1454377323,-0.0106413187,0.0115347812,0.3172792792,0.072094053,-0.1995429844,0.0143463109,0.4727319777,0.1806197912,-0.1624889672,-0.1308171749,0.1810790747,0.2825061083,-0.57771492,0.0114664305,0.0010847271,0.412504226,0.051838845,0.2784292996,0.3223172128,0.2358660251,-0.3069178164,0.0911345482,0.233955577,0.3614220321,-0.0564243272,-0.1345004886,0.3064207733,0.3840125203,-0.1135774478,0.0095585249,-0.2387743741,-0.1069194525,0.1238984764,-0.312969178,0.0145296305,-0.1244764552,-0.0202005524,-0.1086966917,0.1678449959,-0.1577461064,0.246166572,0.3766629696,-0.2775833607,0.5177403092,0.1202461943,-0.142570287,-0.4330539405,-0.1303923428,-0.1732846946,0.0804429352,-0.1950098425,0.2662882507,0.0701352134,-0.4665999115,0.2580045164,0.1210588962,0.2060399204,-0.2420329005,0.0148696592,-0.3456922174,-0.2548391819,0.1120524853,-0.0221589487,0.2393349111,0.003117749,0.0291967262,-0.3391521573,-0.1242933497,-0.3038776815,-0.5340563059,-0.1559813768,0.1528094262,-0.0333183035,0.0070495252,-0.6708773375,0.175240472,-0.0309536923,0.1613244861,-0.0681560263,0.0349203907,-0.4282566607,0.0129329655,0.4087439775,0.4829123616,-0.3364719748,-0.0092636868,-0.1738152951,0.0008567188,0.2438302189,0.6055168509,-0.2631828487,0.3598325253,-0.1960412264,0.0458980389,0.0975217521,-0.3626187742,-0.189652279,0.401699096,-0.0491989031,0.0704114214,0.1953025758,0.2154714763,0.0499428287,-0.0069148042,0.0264242832,0.4943832755,0.1099171937,0.3077587187,-0.3275250793,-0.2363768965,-0.0106486408,0.4756228328,-0.4628458619,-0.1358227432,-0.189200148,0.436730504,0.3493830264,-0.0706919506,-0.0973179042,0.2146663219,0.0543624461,0.1839344203,-0.0258474592,-0.0392578691,-0.3798598945,0.3749347329,0.2192986906,0.2074831128,0.0683715269,-0.1953056902,-0.2741436362,0.0775463134,-0.3143353462,0.015247494,-0.0685912594,0.0763565674,0.16917786,-0.0933396667,-0.3474593759,0.2462809086,-0.3357875049,0.0676223934,-0.4804140329,0.0079303021,0.0789154395,0.0794270784,-0.1203793883,0.0181700457,-0.0685081258,-0.0977360755,-0.091565758,0.3952661455,0.0936069041,0.4953532815,0.1984574646,0.1474683583,0.2836602628,0.1705910563,0.1966609508,-0.1005665436,0.332821548,-0.1249453798,-0.2328208238,0.3337990642,-0.1116134524,0.2934541702,0.148508504,-0.2852832675,0.1642889827,0.1510002911,-0.0583695024,-0.0630397275,0.4609430432,0.0204651263,0.6008419991,0.215366289,-0.4440368414,-0.043703191,0.4861674309,-0.0150728589,-0.0587383583,0.2730801105,0.0624162965,-0.3525634706,0.0695389658,-0.1610847116,0.2996737957,0.1361426562,-0.0543624014,0.0600378923,0.087023735,-0.1404526681,0.2030684352,-0.0107664689,0.3830636144,-0.0683272555,0.0322462134,-0.0401494056,-0.4450127482,-0.1755079329,-0.1217239648,0.1107884347,-0.2857523859,0.145692572,-0.0557505004,-0.0033409924,-0.2302478254,0.3559330106,-0.0930296034,-0.0899568051,-0.1406043768,0.3619558513,0.1214620918,0.0018907607,-0.043534968,0.4041738808,-0.034210328,-0.1786132604,-0.3653002083,-0.0047926516,-0.1858634949,-0.0485286415,0.2934470773,0.1590782553,0.2970312238,-0.2010054439,-0.1913919598,-0.2566900253,-0.0117398575,0.004149633,-0.0588130392,0.3551350534,0.0871948153,0.2114967257,-0.2618366182,-0.0664914399,0.2615098953,-0.294922173,-0.1493818164,-0.1307641417,-0.1245448589,0.1220935136,0.0281824488,-0.2825679779,-0.2752526104,-0.3055755496,0.3190359175,-0.0953012258,0.0536931753,-0.144734472,-0.0529215299,0.1388360262,-0.0915215611,0.1111416742,-0.1933532953,-0.6070045829,0.2168903053,0.0247757211,-0.2385846972,-0.1031488627,0.2235171795,0.1149047464,-0.1313179731,-0.5084902644,-0.3393250406,-0.3350628614,0.0989805162,-0.2253898531,0.152825743,0.1881372631,-0.1391521394,-0.0249838643,0.1779857278,-0.2847417593,-0.0286925212,0.0348592848,-0.0209886953,-0.0396118462,0.2691692412,-0.0004037227,0.468415916,0.1950828135,0.1571453214,0.4356394708,0.1224709377,0.1987567246,-0.403903991,-0.1974372566,0.0201035608,-0.2472994179,-0.2876935899,0.2333061546,0.1175003648,0.0322779641,-0.1820494384,-0.0895072967,-0.2380468547,-0.1836570054,0.075268425,0.0269322488,0.0589255132,-0.0642209053,0.0812672526,0.020664556,-0.0583693758,0.0886469111,0.0663988367,0.182816267,-0.1110026091,-0.5808101296,0.1746718287,-0.6069723368,0.2017239034,-0.2025546432,0.0624966472,0.0504118949,0.0110519584,0.4005098641,0.0054570753,0.724716723,0.0903052911,0.1827917695,-0.0531373955,-0.4591615796,-0.4324049056,0.1206627861,-0.094657518,-0.1170938239,0.0451913849,0.5109405518,-0.3123778999,-0.3635507524,-0.1208665967,0.3289903402,-0.1704082638,-0.1602179855,-0.2254456133,-0.3087471724,-0.2820453048,0.1915449053,0.0622891225,0.3389370739,0.2983134985,0.0704508796,0.0111546637,0.2225702852,0.1877049208,0.0558677129,0.2207463831,0.0630624071,0.2699099183,0.1948717386,-0.0403421931,-0.1628281474,0.5641958117,-0.2501838207,-0.4906096756,-0.0146965552,0.1262732297,0.2365377098,0.5669199824,0.0421113931,0.119796209,0.1443805099,0.3374323547,-0.4358737469,0.0840514079,0.2717902064,0.2514378726,-0.1345534772,-0.5810294747,0.4566181004,0.1905065924,0.0488357693,0.0569090843,0.0513386801,-0.0842369571,0.6181397438,0.068698287,1.0370105505,-0.2821862996,0.2254857421,0.1522788703,0.2190827131,0.5381292701,-0.4257055819,0.0011359917,-0.2776856422,-0.1213341504,0.0625816733,-0.046239797,0.0946654081,0.2631706893,0.0226857271,0.4595285952,-0.1524495184,-0.2197600603,-0.3194171786,0.4795187116,-0.2825456262,-0.7124117613,-0.9103118181,0.0863134041,-0.08077503,0.2299457341,-0.0965871736,0.2374042124,0.1809936613,-0.0895968303,-0.1275954992,-0.0315458514,-0.2507607937,0.0424583666,-0.1638723314,-0.1292311549,0.0311028585,0.3478128314,0.1300769895,0.0361164771,-0.0443165153,0.3795782924,-0.3030950129,0.2124039531,0.0925701782,-0.2542066574,0.2649157047,-0.1542871445,-0.0823559836,-0.1755854785,-0.174619168,-0.3735761642,0.2857707739,0.0387679003,0.039281521,-0.0304132663,-0.5706739426,0.1415370703,0.1314602047,-0.2844032645,0.0980652124,0.2616696358,-0.0888563171,0.2768679559,0.1133859754,-0.2748235464,-0.067826435,0.3945206404,-0.0038811376,-0.1927152872,0.3684560657,0.3933858573,-0.2031614929,-0.1406018138,0.1354209185,0.1984264255,-0.6394310594,0.3688584566,0.2252013236,-0.0853408501,-0.032191474,0.2528886795,0.2513149977,-0.1531388313,-0.0675501302,-0.294572562,-0.1395416111,0.1295035928,0.1711527258,-0.0087237079,0.2367130071,-0.2157865465,-0.0892509967,0.0814835578,-0.2553875148,0.0630020425,-0.3963884711,-0.0154437637,0.1368574351,-0.0703490674,0.4796371162,-0.0704574361,0.0749921501,0.0268754419,-0.2657931149,-0.1725008339,-0.1794169694,0.1676754206,-0.182335034,-0.0706502944,0.165552929,-0.1704592258,-0.1135937795,-0.3698256016,0.179892838,0.199099198,0.0540253222,-0.193341136,-0.0070027262,0.0685852095,-0.2307346612,0.1076776311,0.0636108741,0.1913830936,0.1256137341,-0.0902946144,0.1836617142,-0.2086755484,-0.0562755503,-0.042978853,0.2864102721,0.1725776941,0.1507643014,-0.0274282061,-0.0452587865,0.3327915072,0.3738416433,0.2403503209,-0.0165241398,-0.1362867057,0.011544792,0.1501931548,-0.1640739739,-0.245748803,-0.0173645951,-0.1074973941,-0.0516964644,0.1891137362,0.1765632033,0.5343304873,-0.4159370959,0.0900053084,0.278603375,-0.2512699366,0.0952276662,0.1014289409,0.0517053679,0.0592831373,0.2839736342,0.1568907797,0.1252664626,0.4663273394,-0.1248373613,0.4365477562,-0.0518177152,0.6817247272,-0.037854407,-0.4280995429,0.1009420604,0.4241499603,-0.0978153646,0.0548257492,-0.1713283658,0.3906822801,-0.3606950939,-0.3087567687,-0.1585167199,0.1802538037,-0.1367848516,-0.1669901311,0.0572696477,0.0247286242,-0.1000588909,0.339120388,-0.2464726865,-0.0384191573,0.1525425017,0.1303636432,0.0312286876,-0.2246231288,-0.248613432,-0.1329121143,0.4031760693,-0.1926607043,0.1321589351,-0.1036326885,-0.1441441178,-0.1213669926,0.3481108844,0.1797877103,0.2219308913,0.2054885328,0.1745470613,-0.1545777917,0.1652445644,-0.0052414052,0.1515233964,0.0699849129,0.3075494766,0.1711009592,0.0384392142,-0.1586203277,-0.0355841033,-0.2871457636,0.3118085563,-0.4168596864,0.232824564,-0.0707158223,-0.0420554094,-0.4076864123,0.3572513461,-0.1667425185,0.1307487935,0.3537974358,-0.0285261199,-0.1019279808,-0.2110534906,0.0544466823,0.0225339178,0.6195693016,0.3593432903,-0.0996862352,-0.3158842325,-0.0345318951,-0.7715910077,-0.0830790997,0.0270683337,0.1821212023,0.2142477036,0.3033528924,0.0685295984,0.2734255493,-0.0024457525,-0.1144323945,-0.0322492085,-0.0043265494,-0.1468897313,0.0715842843,-0.2752621472,0.0167888515,-0.0520381518,-0.4649930298,0.2568732202,-0.0127402591,-0.1827230603,-0.0955054536,0.0389735475,-0.0595955104,0.0259904359,0.4170560241,0.0598113313,-0.0962143466,-0.0336410813,-0.1641093194,-0.1254540235,-0.1693587452,-0.1114802808,0.1880920678,-0.0811499581,0.2960055172,-0.1214584336,0.2103291601,-0.4020768702,0.3637862802,-0.1327590197,-0.3278025985,-0.2649255991,-0.0511695668,-0.0675088763,-0.0144006675,0.2698023319,0.3055670857,-0.0117319357,-0.219447121,-0.3290022612,-0.1330343187,0.4405803978,-0.0863306001,-0.1167693734,-0.0473910309,0.2936117053,0.31692186,-0.2588176727,-0.6736818552,0.0697593018,0.0982681587,0.1238983497,-0.14124915,0.3487517536,-0.1722726971,-0.0608772077,-0.1790949404,0.2636466324,-0.1334081739,-0.2623075843,0.1287028342,-0.3133035898]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2547","title":"Dataset load_from_disk is too slow","comments":"There are no solutions yet unfortunately.\r\nWe're still trying to figure out a way to make the loading instantaneous on such disks, I'll keep you posted","body":"@lhoestq \r\n## Describe the bug\r\nIt's not normal that I have to wait 7-8 hours for a dataset to be loaded from disk, as there are no preprocessing steps, it's only loading it with load_from_disk. I have 96 cpus, however only 1 is used for this, which is inefficient. Moreover, its usage is at 1%... This is happening in the context of a language model training, therefore I'm wasting 100$ each time I have to load the dataset from disk again (because the spot instance was stopped by aws and I need to relaunch it for example). \r\n\r\n## Steps to reproduce the bug\r\nJust get the oscar in spanish (around 150GGB) and try to first save in disk and then load the processed dataset. It's not dependent on the task you're doing, it just depends on the size of the text dataset.\r\n\r\n## Expected results\r\nI expect the dataset to be loaded in a normal time, by using the whole machine for loading it, I mean if you store the dataset in multiple files (.arrow) and then load it from multiple files, you can use multiprocessing for that and therefore don't waste so much time. \r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Ubuntu 18\r\n- Python version: 3.8\r\n\r\n\r\nI've seen you're planning to include a streaming mode for load_dataset, but that only saves the downloading and processing time, that's not being a problem for me, you cannot save the pure loading from disk time, therefore that's not a solution for my use case or for anyone who wants to use your library for training a language model. ","comment_length":26,"text":"Dataset load_from_disk is too slow\n@lhoestq \r\n## Describe the bug\r\nIt's not normal that I have to wait 7-8 hours for a dataset to be loaded from disk, as there are no preprocessing steps, it's only loading it with load_from_disk. I have 96 cpus, however only 1 is used for this, which is inefficient. Moreover, its usage is at 1%... This is happening in the context of a language model training, therefore I'm wasting 100$ each time I have to load the dataset from disk again (because the spot instance was stopped by aws and I need to relaunch it for example). \r\n\r\n## Steps to reproduce the bug\r\nJust get the oscar in spanish (around 150GGB) and try to first save in disk and then load the processed dataset. It's not dependent on the task you're doing, it just depends on the size of the text dataset.\r\n\r\n## Expected results\r\nI expect the dataset to be loaded in a normal time, by using the whole machine for loading it, I mean if you store the dataset in multiple files (.arrow) and then load it from multiple files, you can use multiprocessing for that and therefore don't waste so much time. \r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Ubuntu 18\r\n- Python version: 3.8\r\n\r\n\r\nI've seen you're planning to include a streaming mode for load_dataset, but that only saves the downloading and processing time, that's not being a problem for me, you cannot save the pure loading from disk time, therefore that's not a solution for my use case or for anyone who wants to use your library for training a language model. \nThere are no solutions yet unfortunately.\r\nWe're still trying to figure out a way to make the loading instantaneous on such disks, I'll keep you posted","embeddings":[-0.469478786,-0.2423017919,-0.1178942844,0.3941159248,0.2421814054,0.0529478155,0.1993055195,0.2657711208,0.4733243883,0.1050993577,0.0533883423,0.4494888484,-0.0244759563,0.0135968365,-0.0266761817,-0.013558601,0.2004590183,0.0515525267,0.2093926072,-0.1210182309,-0.166077584,0.2531619072,-0.1768193841,-0.2215415686,-0.2757839262,-0.1387374848,-0.0035660972,0.0403216183,-0.068663761,-0.3196170032,0.1828830093,0.1244310066,0.2284581363,0.4482245445,-0.00010555,-0.1314772964,0.133610636,0.0169427358,-0.3660498261,-0.1134866551,0.2358428538,-0.2445044667,0.1371787488,-0.2250014544,-0.0972655639,0.0251809154,-0.0460048281,-0.4642716348,0.2052056044,0.1093912497,0.2575610578,0.3051843643,-0.3067190647,-0.0636639223,-0.0051101944,0.0161650516,-0.234357208,0.2038806826,0.4900560379,0.1526582986,-0.1329111159,0.3433520794,-0.0540964566,0.0930550024,0.4101567566,0.007168029,0.0953570008,-0.2674491704,-0.011626713,0.2036076039,0.4939620197,-0.0521098524,-0.3432075679,-0.30284518,0.1475219429,-0.2496155351,0.2738335133,0.2413631976,-0.1156351566,0.1319803894,-0.0537608378,-0.187780425,-0.0633387938,0.0681013688,0.2679540813,0.2253524214,-0.0873984322,0.0108181862,0.4163218439,0.1147833914,0.2195016891,-0.2432810217,0.1509135514,0.374463439,-0.5127294064,-0.0124600101,-0.0313438661,0.2573702335,-0.0977262333,0.229055196,0.1536374092,0.2092105746,-0.1661462784,0.1606682539,0.1446455866,0.2459539473,-0.0865710378,-0.0402124599,0.2442095727,0.1976842284,-0.1823555082,-0.0017857405,-0.2736601233,-0.1554443985,0.2644675672,-0.2677514851,-0.1355288327,-0.1284794062,-0.0555531271,0.0388844013,0.1759007871,-0.1476335824,0.1911531687,0.3013442755,-0.2921019495,0.5728609562,0.0260151401,-0.0990992263,-0.3289968073,-0.0375294648,-0.1681725234,0.0735851973,-0.249655813,0.1538652331,0.1731598526,-0.2611287236,0.2046505958,0.0870450661,0.1814054102,-0.2219474912,0.2221283466,-0.4922517836,-0.115070492,0.1565223634,0.0439555049,0.2417901009,0.0218334626,-0.0940539241,-0.3755871356,0.1060028449,-0.3327483237,-0.4518424273,-0.2468843907,0.2598523498,0.0015027871,-0.040048454,-0.5704126954,0.1138264015,0.0026696075,0.0800517574,-0.1564571112,-0.0796247572,-0.2967770994,-0.0024338544,0.3181245327,0.3346172869,-0.4551331699,0.102175422,-0.2604034841,0.0749707818,0.3448600471,0.5159237981,-0.3132815063,0.2919723392,-0.1433934718,0.0827732757,0.2209132463,-0.2219853252,-0.2678492963,0.6216792464,0.0714554116,-0.0712878704,0.1130911037,0.2571855783,0.177236706,-0.0366671681,0.1337171644,0.5584762692,0.1512366235,0.2853977382,-0.3553659022,-0.2408134788,0.0228472352,0.5096572042,-0.2727359235,-0.1557220221,-0.1301405728,0.3853493333,0.3681561947,-0.1020006388,-0.0429258533,0.24489218,0.1422255635,0.0588854179,-0.0217141639,-0.0522179306,-0.209490642,0.3154984415,0.3402776122,0.105510138,0.0518874526,-0.1499329358,-0.2064654827,0.0269759055,-0.3376523852,0.0885742903,0.0651175827,0.0703162625,0.0686208978,-0.0590329617,-0.2992327809,0.2201518267,-0.3418131769,-0.1120349914,-0.2888448238,0.0647155195,-0.0210965723,0.0440994389,-0.0066615953,-0.1089289188,-0.015061412,-0.0243874881,-0.2034513652,0.3786915243,0.0924975276,0.5323041081,0.1840373278,-0.0338038206,0.1784894913,0.1857829392,0.1977801919,0.0105078947,0.3661213517,-0.1660588086,-0.1404550225,0.3742795587,-0.0605217814,0.2691720724,0.2077660859,-0.2657869756,0.3263538182,0.1359153837,-0.0913011804,-0.1102432013,0.4074093699,0.089459002,0.6007938385,0.1311140507,-0.3572582304,-0.0762512758,0.5789567828,0.0450012088,-0.0418994576,0.3076453805,0.0911244899,-0.3170955777,0.0501960255,0.1875695586,0.4310155809,0.1976147294,-0.0040011485,-0.0328989699,-0.0056909975,-0.2190225869,0.2751500607,-0.0797054172,0.498817414,0.0467041694,-0.0254617687,-0.006002218,-0.4627438784,-0.3295349777,-0.0532215647,0.1768090725,-0.0895202681,0.2188511938,-0.075361684,0.0838565305,-0.2535158098,0.2120685279,-0.0295272581,-0.1037035435,-0.1198300719,0.28434515,0.1204025373,0.0475351065,-0.1557546854,0.1716058105,-0.1020782739,-0.2200122178,-0.3387371898,-0.0934301838,-0.2868705392,0.0351008698,0.2899343073,0.0647917837,0.3139715791,-0.1468579024,-0.1171031594,-0.1601425558,-0.0196714327,-0.0257332213,-0.0159380417,0.1736114323,-0.0721044019,0.1581274718,-0.1802784801,-0.1895720959,0.1762481779,-0.2932653725,0.0095894625,-0.0883764476,-0.0860165283,-0.0133092999,-0.0503963567,-0.4648726583,-0.3632452488,-0.2891048789,0.0190057158,-0.0670994297,0.1013495997,-0.2616087496,0.0097601423,0.0619439259,0.1845126003,0.1032932326,-0.2362322658,-0.585478127,0.322601378,0.0049837912,-0.2711999714,0.0211331621,0.1947817206,0.2655994892,-0.0799174011,-0.538618207,-0.0461725965,-0.3027786314,0.0212117378,-0.3066812456,0.018114157,0.1055008546,-0.0077714929,-0.0813887194,0.0991656929,-0.169482708,-0.0149223125,-0.0797677115,0.083615467,-0.2301028073,0.2255464047,0.010092889,0.4175733924,0.055357825,0.0708941519,0.4242661297,0.243262589,0.1240621135,-0.3561916053,-0.1205060557,0.0345835239,-0.2498476803,-0.2961779535,0.2965631783,0.0390266366,-0.4405575097,-0.2313330472,-0.0357487835,-0.2201548368,-0.1747718602,0.0717260092,-0.2013130486,0.0571289957,-0.0484484546,0.1518857032,-0.0060724313,-0.0490809307,-0.0065013068,0.0212808885,0.0920034423,-0.148169294,-0.3790249825,0.0960948169,-0.5191190243,0.1156713739,-0.2607972026,0.1470319927,0.0420127399,0.0311343335,0.3150844872,-0.1063322201,0.6286598444,0.0305875186,0.1294413358,0.0650368854,-0.2170046717,-0.4151456058,0.0074853864,-0.154278025,-0.0826796666,0.1175875887,0.5377980471,-0.2096422017,-0.2062813938,-0.2187013328,0.2994832397,-0.2043037415,-0.3161745071,-0.2104469985,-0.3497510552,-0.3888547421,0.1937903464,0.0499513522,0.3467878699,0.0920253322,-0.0125946719,-0.0598893501,0.2289756089,0.2043618858,0.0788377225,0.2240424752,0.0071833916,0.3724558651,-0.092207931,0.0635767654,0.0013742587,0.4416355789,-0.1714116186,-0.2881593108,0.070831731,-0.0826338828,0.2258850783,0.4218269885,0.0237919521,-0.0086855767,0.1257955432,0.2253558487,-0.2926580608,0.0639968514,0.2721913457,0.1781928837,-0.1965345293,-0.6573073268,0.3948166966,0.2069479525,0.0046930448,0.0570728332,-0.1848088354,-0.1417870373,0.689068377,0.201516211,1.0291241407,-0.1868833303,0.2222974896,-0.0363855511,0.1555663645,0.2503242195,-0.4073494971,-0.0044394722,-0.3594462276,-0.1366780847,0.087859191,0.0014562157,0.1326104403,0.4059722722,-0.1406882852,0.3619869649,-0.0185154751,-0.3825742006,-0.2977780998,0.5117338896,-0.3943451643,-0.6922773123,-0.8247359991,0.167798996,-0.0810490325,0.1186974868,-0.0903273523,0.2641170323,0.2134920955,-0.1382889301,-0.1650550663,0.1232707947,-0.1168855131,0.1214117929,-0.3188132942,-0.2166332453,0.0729496256,0.2318356335,0.1302154809,0.1892757565,-0.004162618,0.4877101779,-0.2195078731,0.2118656337,0.1033782586,-0.2827644348,0.308196187,-0.2601314187,-0.1553424597,-0.2089232653,-0.1640534103,-0.236153841,0.303773284,0.0139227575,0.1338056028,0.0698402673,-0.5948092937,0.1890896857,0.0155131202,-0.2636429071,0.2033052146,0.2456428558,-0.1535345018,0.3958010674,0.059680637,-0.3271471262,-0.0964698717,0.3515591025,0.1941083968,-0.1444524527,0.3934570253,0.3167483211,-0.2243391722,-0.2559329569,0.1372367442,-0.0305804573,-0.5054119229,0.2752221227,0.1805755347,-0.0334491543,-0.0006365163,0.0063777929,0.1991868764,-0.1418983936,-0.1174499616,-0.3069978654,-0.1299629509,-0.0863467604,0.1642923206,0.0839918405,0.2119122744,-0.3602120876,-0.0737627819,0.0395037606,-0.3430844545,0.1233982816,-0.3643568158,0.0190980211,-0.1297381967,-0.0280149262,0.2609645724,-0.017678041,0.1585829407,0.0504698083,-0.0842273906,-0.2597214878,-0.1886479557,0.1182471588,-0.1186012775,0.0447765775,0.1561884284,-0.3250782192,-0.2172202021,-0.2890641391,0.0348476358,0.0980561227,0.0104703568,-0.0939081386,0.0520677976,0.1179297715,-0.160536617,0.0660485104,-0.0870737135,0.3416157961,0.0650926828,0.0499588326,0.3553997278,-0.1964927316,-0.2897398174,-0.1704972237,0.073591888,0.3466828465,0.1583281457,-0.0846706703,0.0896618143,0.3591156304,0.3362050354,0.257681191,0.0216135904,-0.2313483208,0.0998944342,0.2502575815,-0.2334995121,-0.2964687943,-0.1021904498,-0.1455088109,-0.0343764648,0.0510738902,0.0636559203,0.4499304295,-0.2592739761,0.0849270001,0.2981085777,-0.2808162272,0.1106245294,0.26478374,0.1184213981,0.1860944927,0.1724980474,0.0764960274,0.0882382095,0.7180646658,-0.1674936712,0.4605653882,0.1990515888,0.5430527329,-0.1301974058,-0.2953585386,0.108894147,0.2278901786,-0.163866818,0.121145077,0.0229620337,0.2102442533,-0.2791336775,-0.2531220019,-0.2078100294,0.0798742995,-0.1857957095,-0.1577622443,-0.0551924109,0.068592824,-0.1077869087,0.3235854805,-0.2514153421,0.1503015757,0.0585514456,0.2984164953,0.1284313351,-0.1307538152,-0.0483747199,-0.2393860519,0.3301278651,-0.2393922657,0.1104550585,-0.0595015027,-0.1310568303,-0.1504596919,0.1248567998,0.2495143265,0.3417238891,0.0862351432,0.1263657361,-0.1184461787,0.0892466083,-0.0592333563,0.2225809991,0.2490220666,0.4293945432,0.3248225451,0.1182188541,-0.2089403123,0.0747366175,-0.1469841301,0.292044878,-0.2994968891,0.4099962711,0.1020161882,-0.1169013381,-0.4516202509,0.355976373,-0.2491376549,0.063545607,0.3920060992,-0.0215290654,-0.0770393014,-0.3009452224,0.1012689844,-0.0411057286,0.6939358115,0.2713715136,-0.0689361989,-0.2933212519,-0.051594574,-0.7838363051,-0.0110740345,-0.0357743129,0.292311281,0.0586512834,0.3379094303,0.1290598065,0.2557680905,-0.0540650785,-0.0681863725,-0.1145163178,0.1603102088,-0.2366978973,0.0467103012,-0.1564765871,0.0683117211,-0.1357945651,-0.4821786284,0.1385382116,-0.0724946558,-0.048545558,-0.1844356805,-0.1335465312,0.0081034182,0.1373850107,0.4866031706,0.0385474451,0.0232347474,-0.147089228,-0.0542540029,0.0028939648,-0.0420576222,-0.0661673695,0.1356499642,0.0847152025,0.2798967957,-0.0646764189,0.2580752373,-0.3375319541,0.2599024773,-0.2294941247,-0.3023751974,-0.2581327558,-0.1102368236,0.1125591844,0.006415498,0.1501609832,0.3269652426,0.0328190289,-0.1675225347,-0.3837408721,-0.2113311291,0.3055428267,-0.0138703659,-0.2495283782,-0.1542966068,0.2067962289,0.2052088231,-0.0726262406,-0.5406602621,0.1536536366,0.163872391,0.0282975882,-0.174040094,0.3224799931,-0.0978929102,0.0442292057,-0.1590591669,0.2737866044,-0.1268524081,-0.3126335144,-0.0827788338,-0.3731062412]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2543","title":"switching some low-level log.info's to log.debug?","comments":"Hi @stas00, thanks for pointing out this issue with logging.\r\n\r\nI agree that `datasets` can sometimes be too verbose... I can create a PR and we could discuss there the choice of the log levels for different parts of the code.","body":"In https:\/\/github.com\/huggingface\/transformers\/pull\/12276 we are now changing the examples to have `datasets` on the same log level as `transformers`, so that one setting can do a consistent logging across all involved components.\r\n\r\nThe trouble is that now we get a ton of these:\r\n\r\n```\r\n06\/23\/2021 12:15:31 - INFO - datasets.utils.filelock - Lock 139627640431136 acquired on \/home\/stas\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock\r\n06\/23\/2021 12:15:31 - INFO - datasets.arrow_writer - Done writing 50 examples in 12280 bytes \/home\/stas\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.\r\n06\/23\/2021 12:15:31 - INFO - datasets.arrow_dataset - Set __getitem__(key) output type to python objects for no columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n06\/23\/2021 12:15:31 - INFO - datasets.utils.filelock - Lock 139627640431136 released on \/home\/stas\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock\r\n```\r\n\r\nMay I suggest that these can be `log.debug` as it's no informative to the user.\r\n\r\nMore examples: these are not informative - too much information:\r\n```\r\n06\/23\/2021 12:14:26 - INFO - datasets.load - Checking \/home\/stas\/.cache\/huggingface\/datasets\/downloads\/459933f1fe47711fad2f6ff8110014ff189120b45ad159ef5b8e90ea43a174fa.e23e7d1259a8c6274a82a42a8936dd1b87225302c6dc9b7261beb3bc2daac640.py for additional imports.\r\n06\/23\/2021 12:14:27 - INFO - datasets.builder - Constructing Dataset for split train, validation, test, from \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/0d9fb3e814712c785176ad8cdb9f465fbe6479000ee6546725db30ad8a8b5f8a\r\n\r\n```\r\n\r\nWhile these are:\r\n```\r\n06\/23\/2021 12:14:27 - INFO - datasets.info - Loading Dataset Infos from \/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/0d9fb3e814712c785176ad8cdb9f465fbe6479000ee6546725db30ad8a8b5f8a\r\n06\/23\/2021 12:14:27 - WARNING - datasets.builder - Reusing dataset wmt16 (\/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/0d9fb3e814712c785176ad8cdb9f465fbe6479000ee6546725db30ad8a8b5f8a)\r\n```\r\n\r\nI also realize that `transformers` examples don't have do use `info` for `datasets` to let the default `warning` keep logging to less noisy.\r\n\r\nBut I think currently the log levels are slightly misused and skewed by 1 level. Many `warnings` will better be `info`s and most `info`s be `debug`.\r\n\r\ne.g.:\r\n\r\n```\r\n06\/23\/2021 12:14:27 - WARNING - datasets.builder - Reusing dataset wmt16 (\/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/0d9fb3e814712c785176ad8cdb9f465fbe6479000ee6546725db30ad8a8b5f8a)\r\n```\r\n\r\nwhy is this a warning? it is informing me that the cache is used, there is nothing to be worried about. I'd have it as `info`.\r\n\r\nWarnings are typically something that's bordering error or the first thing to check when things don't work as expected.\r\n\r\ninfrequent info is there to inform of the different stages or important events.\r\n\r\nEverything else is debug.\r\n\r\nAt least the way I understand things. \r\n","comment_length":41,"text":"switching some low-level log.info's to log.debug?\nIn https:\/\/github.com\/huggingface\/transformers\/pull\/12276 we are now changing the examples to have `datasets` on the same log level as `transformers`, so that one setting can do a consistent logging across all involved components.\r\n\r\nThe trouble is that now we get a ton of these:\r\n\r\n```\r\n06\/23\/2021 12:15:31 - INFO - datasets.utils.filelock - Lock 139627640431136 acquired on \/home\/stas\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock\r\n06\/23\/2021 12:15:31 - INFO - datasets.arrow_writer - Done writing 50 examples in 12280 bytes \/home\/stas\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.\r\n06\/23\/2021 12:15:31 - INFO - datasets.arrow_dataset - Set __getitem__(key) output type to python objects for no columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n06\/23\/2021 12:15:31 - INFO - datasets.utils.filelock - Lock 139627640431136 released on \/home\/stas\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock\r\n```\r\n\r\nMay I suggest that these can be `log.debug` as it's no informative to the user.\r\n\r\nMore examples: these are not informative - too much information:\r\n```\r\n06\/23\/2021 12:14:26 - INFO - datasets.load - Checking \/home\/stas\/.cache\/huggingface\/datasets\/downloads\/459933f1fe47711fad2f6ff8110014ff189120b45ad159ef5b8e90ea43a174fa.e23e7d1259a8c6274a82a42a8936dd1b87225302c6dc9b7261beb3bc2daac640.py for additional imports.\r\n06\/23\/2021 12:14:27 - INFO - datasets.builder - Constructing Dataset for split train, validation, test, from \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/0d9fb3e814712c785176ad8cdb9f465fbe6479000ee6546725db30ad8a8b5f8a\r\n\r\n```\r\n\r\nWhile these are:\r\n```\r\n06\/23\/2021 12:14:27 - INFO - datasets.info - Loading Dataset Infos from \/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/0d9fb3e814712c785176ad8cdb9f465fbe6479000ee6546725db30ad8a8b5f8a\r\n06\/23\/2021 12:14:27 - WARNING - datasets.builder - Reusing dataset wmt16 (\/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/0d9fb3e814712c785176ad8cdb9f465fbe6479000ee6546725db30ad8a8b5f8a)\r\n```\r\n\r\nI also realize that `transformers` examples don't have do use `info` for `datasets` to let the default `warning` keep logging to less noisy.\r\n\r\nBut I think currently the log levels are slightly misused and skewed by 1 level. Many `warnings` will better be `info`s and most `info`s be `debug`.\r\n\r\ne.g.:\r\n\r\n```\r\n06\/23\/2021 12:14:27 - WARNING - datasets.builder - Reusing dataset wmt16 (\/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/0d9fb3e814712c785176ad8cdb9f465fbe6479000ee6546725db30ad8a8b5f8a)\r\n```\r\n\r\nwhy is this a warning? it is informing me that the cache is used, there is nothing to be worried about. I'd have it as `info`.\r\n\r\nWarnings are typically something that's bordering error or the first thing to check when things don't work as expected.\r\n\r\ninfrequent info is there to inform of the different stages or important events.\r\n\r\nEverything else is debug.\r\n\r\nAt least the way I understand things. \r\n\nHi @stas00, thanks for pointing out this issue with logging.\r\n\r\nI agree that `datasets` can sometimes be too verbose... I can create a PR and we could discuss there the choice of the log levels for different parts of the code.","embeddings":[0.1725228131,-0.3411969244,0.087156266,0.2024417371,0.2560847104,0.1124867052,0.5520288348,0.3101122975,-0.0718619078,-0.3018711507,-0.0416375324,0.1757286787,-0.2178721428,0.2808289826,0.0503808483,-0.163519308,-0.0827343389,-0.0110472916,-0.2270587385,-0.0930731893,-0.0216768421,0.2245499939,-0.0460353866,0.2822227478,-0.4451558888,-0.2340548187,0.1867844164,0.141461879,0.2126887739,-0.4743229747,0.0037430937,-0.0164442528,0.0430156179,-0.0290570315,-0.000123391,0.1806585491,0.3962735832,-0.0402114503,-0.4393552542,0.0913482085,-0.4999396801,-0.4355217516,0.2987112105,-0.1670996398,0.2413566113,-0.6675630212,0.3559740782,-0.3311716914,0.22117275,0.2089207321,0.102292031,0.1760010123,-0.3780121803,0.0249666851,0.4033831656,0.4810088873,-0.1526447982,0.1070233881,0.2368955016,0.0921347141,-0.1562441736,0.3786007464,0.0120764785,-0.1597690731,0.3421988189,0.1947605759,0.4285323322,0.0483221002,0.2384692132,0.3680716157,0.0961686224,-0.3306325674,-0.4083094895,-0.6607736349,0.0907099992,-0.1062864065,-0.0957999155,-0.0971783623,-0.1848097444,0.2638143003,-0.1493404508,-0.1274099499,0.0364527479,-0.212759614,0.2145827115,-0.0887042731,-0.3330810964,0.0859099627,-0.0800955966,-0.2473695129,0.0023985605,-0.0961467102,-0.1939752251,-0.1913670003,-0.0224671215,-0.1187931299,0.1438100636,0.0872294679,-0.0508307405,0.2879404128,0.0360640846,0.0421623513,0.2165592909,-0.0343364775,0.1342561841,0.1307466626,0.5509587526,0.2441019863,0.2075189054,-0.008192936,0.3790483475,-0.1100816131,-0.07710132,-0.0546142608,0.0875650421,0.1399331987,0.4767226875,-0.2333270907,-0.1709547639,-0.043091476,-0.0543124415,0.0099578304,0.0024802366,0.2184966952,0.4596283436,0.1707849503,0.1712561101,0.1859490573,0.209877342,0.1148033589,-0.0688310787,-0.2183574289,-0.1541655362,0.2162984461,0.0351495035,-0.0420308933,0.2268222868,0.0731881335,0.1887268573,-0.1175669506,0.0699681491,0.2732225955,0.1356358379,0.2352600694,-0.5406377316,0.2473048121,0.155899629,-0.3606781065,-0.3538204432,0.0626212433,-0.100223124,-0.3030183315,-0.0439281613,-0.0210314821,-0.3015890718,0.1510032266,-0.0315897204,0.3249503076,0.0678391531,0.270339936,0.3980059326,0.0139181223,0.2147996873,0.0231182612,0.0114351986,0.4221026003,-0.0385522731,-0.6235333681,-0.4271850288,-0.0567950346,0.0983283967,0.0723430291,-0.0330573395,0.2198322415,-0.1125571579,-0.0502590127,0.1916441917,-0.3836287558,0.2618423998,0.4828167558,-0.2888585627,0.1380439848,0.442145288,0.0850284174,-0.0783957765,-0.07494279,0.1810775101,-0.1197182536,0.0337337144,-0.0326996781,-0.1435083747,-0.3740644157,-0.1517745852,0.0044647134,-0.0677142441,0.2607993186,-0.1733944565,0.3144968748,0.179753378,-0.1150381491,0.0127915889,0.0781221986,0.2459311634,-0.2255862951,0.0905504897,0.0810733587,-0.5876051188,0.0058921529,-0.3214451373,0.0416774079,-0.1873554289,-0.4808943868,-0.1556637734,-0.1422926784,0.1597002745,-0.4150123596,-0.0192603786,0.1258560866,-0.2073339522,-0.1992819011,-0.4069609046,0.010996813,-0.0986962989,0.2772515714,-0.3563072085,0.1548285335,0.3443709314,0.1146947518,-0.1531070918,0.0763476193,0.139760837,-0.0651040599,0.1548536122,0.3248720467,0.0179695636,0.6038194299,0.1306494921,0.4166247249,0.0582289696,0.0767300874,0.1445856839,0.0760668665,-0.1495212764,0.1523808688,-0.0657393932,0.4161950946,0.0636922196,0.2811014354,0.0199383218,0.1184126213,-0.2317800373,-0.171087712,-0.4485034645,-0.3369162679,0.1558888704,-0.0243383646,0.1225605533,0.1830338836,-0.2320762873,0.1382439733,0.4867061973,-0.3173246384,0.0239928346,0.1297966838,-0.2656702101,0.1580067575,0.2322458178,0.3231045604,0.2867370844,0.2394700944,0.068737857,0.1981016099,0.1802074015,0.023949286,0.2909261286,0.2072955519,-0.0971031711,0.1177337766,-0.1308977157,-0.1448405832,0.1288966388,0.0378817469,-0.2107934803,0.0132204499,-0.422955364,0.2986923456,-0.2806160152,-0.2747745216,-0.5732249618,0.1672849655,-0.2866936922,-0.4487184882,0.1492959261,-0.3464496732,-0.3457983136,0.3194854558,-0.0344772078,0.5397189856,-0.1021586061,0.1989433616,-0.4335125685,-0.0686373562,-0.0341499858,-0.0356619656,0.138811931,0.1156102866,0.0614189617,-0.2835664153,-0.1546472311,-0.2253981233,-0.3233311474,0.1690949798,-0.2062082887,0.2027805299,0.2550833523,-0.0227308329,0.0622035787,-0.3148314059,-0.1103437915,-0.0673675835,-0.0574807227,0.0171102509,-0.0726096183,0.3511283994,-0.2880993187,-0.0399906263,0.1052839607,-0.3069507182,0.7045112848,-0.3959383965,0.1219861582,0.6507042646,-0.2035203725,0.4366971552,-0.0826170146,0.3348893225,0.1882422417,-0.0518544838,0.0956461728,-0.0461849608,0.2713853121,-0.5008628368,-0.1019024476,-0.2713529468,-0.1391539276,-0.5069953203,-0.2255197614,-0.4302291274,-0.4604372382,0.2304776609,0.0909764171,-0.0586985834,0.1700603515,-0.042185206,0.1322055012,-0.2719264925,0.155021295,-0.1330126077,-0.0089530069,0.0638806447,0.1364895105,0.1463652104,0.4343576431,0.3973280787,0.0051057199,-0.5073907375,0.0641243458,0.317185998,0.0283218995,-0.2440285832,-0.1385973692,0.0207971558,-0.3648214936,-0.0674673989,0.1058591902,0.176632002,0.100424476,0.2310435176,0.0112018595,-0.4799422622,0.0750439391,-0.2151429802,-0.119996801,-0.226141423,0.3898008168,-0.0957811624,0.054086037,0.1418318003,0.3962545693,-0.0332373157,-0.0589784421,-0.3074739873,-0.189088583,-0.5284596086,0.4252341092,-0.0845706835,0.0548270009,-0.1712548286,-0.6365629435,0.0225231443,0.291370362,-0.0731277838,-0.0352564901,0.0659099147,0.2296977788,0.0327776112,-0.2901079953,-0.0954055488,-0.4920993447,0.4512870908,0.4576242864,0.5969647765,-0.4474855959,-0.395989418,0.1912863702,-0.240059182,-0.007618668,0.2011221796,-0.2185287923,-0.0768996552,-0.1053551137,-0.3052063584,0.1202377975,-0.0721469447,-0.4518235326,0.1842970103,0.0454415493,0.1027014926,0.3930849731,0.1082368866,0.0826000944,-0.0221832152,0.0595437549,0.248027429,0.1985568553,0.027242003,0.5696771741,0.0346864946,-0.3558327854,0.0104209809,0.0990778282,0.2703762949,0.6357893348,-0.162992999,0.3763466477,0.2563649714,0.3062865734,-0.3317093253,0.1575781107,0.3540697694,0.2253126055,-0.3683909774,0.0830763355,0.6284782887,0.1544093639,-0.1889298856,0.5192847252,0.5616207719,-0.4118129015,0.2686728239,-0.0577188097,1.0515260696,-0.0787271485,0.2243228555,-0.1575524807,-0.480335772,0.6950417757,0.2416572571,0.1644827425,-0.3481084406,-0.1683155,0.1125309244,-0.193037048,0.2504825294,0.0173901971,-0.2088644952,-0.1195114851,-0.2981643081,0.2147113085,-0.1959163845,0.1088664755,-0.292840153,-0.2047685385,-0.5198539495,-0.0758502185,-0.0310848448,0.109659262,-0.0427430719,-0.1458023787,-0.0122350156,-0.1332191974,-0.4343958795,-0.4371861517,0.3731702268,0.0117542977,0.4163282216,-0.3817235231,0.0303493682,0.1053375304,0.317835331,0.0770623833,-0.0895017236,-0.0941721201,-0.1707302183,0.1812679917,0.1049099341,-0.0067997887,0.2435864359,0.2293702066,-0.0794018134,0.1551425308,0.0602948368,-0.1993078887,0.0233047474,-0.2982432842,-0.3991319537,-0.2535714507,0.4907600284,0.2858037651,-0.0144483652,0.0263370946,0.0096713658,0.1196761653,-0.0478092916,-0.0264439005,0.1351392716,-0.5101807117,-0.006517204,0.2070388198,-0.0250479579,-0.0228497926,0.2854443192,-0.0648542792,-0.1146365032,-0.0353613496,0.4134577513,0.3352815211,-0.4740405083,0.3915547431,-0.0885794088,-0.2742888331,-0.1122130007,0.019424038,0.2273577303,0.1004941463,-0.3486249745,0.0035253118,-0.5281161666,0.6595492959,-0.018185785,0.3323071897,-0.4016903639,0.0708377808,-0.1490448564,-0.090348579,-0.1915858984,-0.0619834736,-0.2875707746,0.0550510101,0.0313851573,-0.0612781271,0.3700456619,0.1234829947,0.0375284664,-0.2259990424,-0.0690301359,0.0227400046,-0.0792108253,0.1676869839,-0.2555173635,-0.2914051116,0.1479959041,0.0130964415,0.0895102918,-0.1912076473,0.4658153951,0.4780658782,0.1995007843,0.496457696,-0.3006253839,0.2692775428,0.007000342,-0.1602909416,0.2643518448,0.2556018531,-0.0178001206,0.177261889,-0.3485271037,0.0726410449,0.2780468762,0.4325278401,0.0382560715,0.0531348772,0.0897495672,0.1792062819,-0.2310957611,0.1210649163,0.0841101483,0.4048541486,-0.1895457804,0.1114734709,-0.2315522134,0.0273107216,0.0749081671,0.1155580878,-0.163249746,-0.0315433107,0.0055296519,0.4606437683,-0.1514130831,-0.1861306727,0.5105665922,0.0279873256,-0.176178053,-0.0367702208,0.3669159114,0.2180600166,-0.15188618,-0.0931126326,0.3715485632,0.118919462,0.3561661839,-0.1837507337,0.0892959982,0.1574726701,-0.3492941856,0.0968265161,0.164032355,-0.0351879336,-0.0019465553,0.3601388037,-0.3140517771,0.0749963596,0.0771070421,0.2259914726,0.078642495,0.0234907772,-0.064890638,-0.2727424204,-0.1009083167,-0.3651106358,0.351572603,-0.0430485718,-0.2642073929,-0.2607120872,-0.0327246226,-0.3277746439,0.1657454818,-0.1916968077,-0.0087693678,-0.0123801539,0.0968671143,0.0728986859,0.6242207289,-0.226707682,-0.110211134,0.2185063362,0.2157104909,0.4750937223,0.5750283003,0.4337749481,0.0192058384,-0.0654248521,0.1829201579,0.1801286638,0.077289775,0.1416488141,0.4321017861,-0.2179041952,0.0185493752,0.1032306254,0.0475987419,0.0534620658,-0.1279210299,-0.2434788942,-0.0033829249,-0.343465656,0.759490788,-0.1994993836,-0.1174688935,0.1308294982,-0.1452653706,-0.3394357264,-0.5112398267,0.4027397931,0.1448567957,0.0304174647,0.0709595159,0.0536245704,-0.2660658658,0.1413356811,0.3206531405,0.2523708642,-0.0082954057,0.0624714755,-0.0955492929,0.0539703853,0.2519247234,-0.173342824,0.0429269224,-0.1633802205,0.1606699526,0.1417628974,0.1481829733,-0.2656475902,-0.063365601,-0.143798843,-0.2084694952,-0.1232077256,0.0342563838,-0.1047701538,0.1150149256,-0.0754298121,0.140526861,0.1080746129,-0.1116309762,-0.1077926755,-0.0374131016,0.2306878418,0.2661684453,0.2055827528,0.1215523258,-0.1538725346,0.061189048,-0.6028913856,-0.1888510287,-0.3893948197,-0.1632141471,0.1578995883,-0.2710790634,-0.0309375972,-0.2790280879,-0.1647611409,-0.2068668753,0.1150203422,0.1234321594,-0.1767787486,-0.2039229721,0.1764041781,-0.2128604203,-0.1381817311,0.0512238964,0.3176149726,0.0969869643,0.0999709666,-0.1503223777,-0.2495144755,0.5768634677,-0.2938459218,-0.096803613,0.2487762719,0.5207483768,0.1378049552,-0.1285882145,-0.6660390496,-0.1548404247,0.2529763281,-0.3253514469,0.103119202,0.0039318847,-0.3482114971,-0.1231612116,-0.1825647205,0.266315788,0.0733964369,0.0431314744,-0.056429632,-0.3381456435]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2542","title":"`datasets.keyhash.DuplicatedKeysError` for `drop` and `adversarial_qa\/adversarialQA`","comments":"very much related: https:\/\/github.com\/huggingface\/datasets\/pull\/2333","body":"## Describe the bug\r\nFailure to generate the datasets (`drop` and subset `adversarialQA` from `adversarial_qa`) because of duplicate keys.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"drop\")\r\nload_dataset(\"adversarial_qa\", \"adversarialQA\")\r\n```\r\n\r\n## Expected results\r\nThe examples keys should be unique.\r\n\r\n## Actual results\r\n```bash\r\n>>> load_dataset(\"drop\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset drop\/default (download: 7.92 MiB, generated: 111.88 MiB, post-processed: Unknown size, total: 119.80 MiB) to \/home\/hf\/.cache\/huggingface\/datasets\/drop\/default\/0.1.0\/7a94f1e2bb26c4b5c75f89857c06982967d7416e5af935a9374b9bccf5068026...\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 28553293-d719-441b-8f00-ce3dc6df5398\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1044-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":4,"text":"`datasets.keyhash.DuplicatedKeysError` for `drop` and `adversarial_qa\/adversarialQA`\n## Describe the bug\r\nFailure to generate the datasets (`drop` and subset `adversarialQA` from `adversarial_qa`) because of duplicate keys.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"drop\")\r\nload_dataset(\"adversarial_qa\", \"adversarialQA\")\r\n```\r\n\r\n## Expected results\r\nThe examples keys should be unique.\r\n\r\n## Actual results\r\n```bash\r\n>>> load_dataset(\"drop\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset drop\/default (download: 7.92 MiB, generated: 111.88 MiB, post-processed: Unknown size, total: 119.80 MiB) to \/home\/hf\/.cache\/huggingface\/datasets\/drop\/default\/0.1.0\/7a94f1e2bb26c4b5c75f89857c06982967d7416e5af935a9374b9bccf5068026...\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 28553293-d719-441b-8f00-ce3dc6df5398\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1044-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\nvery much related: https:\/\/github.com\/huggingface\/datasets\/pull\/2333","embeddings":[-0.0905093253,-0.1160073131,0.00945515,0.470908016,0.0840874016,0.0286583472,0.3028091192,0.2397654057,0.1277973205,0.1814353019,-0.1439216137,0.5057473183,-0.048295714,0.1198933572,0.0266699288,-0.032785099,0.0012848077,0.0136723919,-0.2982798517,-0.0762103871,-0.2426401526,0.387050271,-0.2559149265,0.10335318,0.0598408356,-0.1042755619,-0.1572466195,0.3131027222,-0.0433670767,-0.2897949517,0.1407245398,0.0112546887,-0.2836559713,0.4620527029,-0.0001087736,-0.0179601684,0.1347592473,0.063810952,-0.4384645522,-0.1568871439,-0.2514858246,-0.0515594594,-0.0878485814,-0.2769972682,0.1303325295,-0.2024025172,-0.3193139136,-0.376553297,0.7059348822,0.0964307413,0.258453995,0.2994432449,0.0578671731,-0.2653636336,0.1903439015,0.03764534,-0.1994792819,0.2062627226,-0.0147559959,-0.0670088753,0.3541854918,0.160368681,-0.0145851849,-0.0506396517,0.2040350139,0.0678948611,0.0129833156,-0.1696084887,0.340441376,0.3239115477,0.2472787201,-0.5006279945,-0.2000300586,-0.2800831497,0.2844666243,-0.0895541832,0.1983643919,0.0411963165,-0.248807773,0.1426331103,0.066547215,0.1313625276,0.0587465577,-0.0792369768,0.1552489251,-0.0006441801,0.0497320481,0.0894895196,0.1573434621,-0.1723117828,-0.0185095444,-0.3647033572,0.0199675113,-0.1255145669,-0.5385659933,0.0129776504,-0.0629277751,0.0256692767,0.4364517331,0.0741026253,0.2214869559,-0.0984062999,-0.0482476018,0.0024116111,0.0399042964,0.3961126208,0.1762890518,0.0912107676,0.1504677236,0.1907208413,-0.2307083607,0.1641299576,0.0568992607,-0.1409792006,0.2863470614,0.1551881731,0.6009869576,-0.0693046823,-0.4248117208,0.2033160925,-0.1676324755,-0.124566488,-0.2327310294,0.1176542267,0.0935675427,-0.0219017286,-0.125678286,0.2016167492,-0.329765439,-0.1748307794,-0.2609870434,-0.0125090387,-0.0085719796,-0.0722294599,0.0838416889,-0.181197688,0.277915746,0.3087557852,-0.0046123383,-0.1379147619,0.2745801508,0.0929096639,-0.2666784823,0.0398336798,0.1705012321,0.0668969154,0.187929526,-0.3602023125,0.0600283071,0.1122215241,-0.0289515667,-0.1286867708,-0.1176909655,0.3158142269,0.0256931968,0.0972319841,-0.1766062826,0.1494122595,0.1119302288,0.0906810015,-0.0044025579,0.039072413,-0.1598896235,-0.4287461638,0.2847690582,0.7104182839,-0.3256116211,-0.1319526285,-0.1431197524,0.0301795825,0.1442250609,0.2293822169,-0.1759794354,0.2083805352,-0.3001039922,0.1173925847,0.0831043348,-0.342885524,-0.4917336702,0.2908804715,-0.3543904424,0.1900268048,0.1736855656,-0.0551680289,0.2153705508,-0.0025115306,0.011651733,-0.0165486876,-0.0907604471,-0.1353102922,-0.3101507425,-0.3366243541,0.1749132574,-0.2637753487,0.0284835584,0.0655985177,0.3573909402,-0.0704982579,0.2868715525,-0.0610246696,0.0467902683,0.0863300711,0.253734529,-0.0603276454,0.2205081582,-0.019394381,-0.5272825956,0.3890169859,-0.2856177986,0.0389644131,-0.1500831693,-0.2936635315,-0.5489760041,0.054138422,-0.4572040737,-0.2501839697,0.248770386,0.1914100796,-0.0284573697,0.1142127141,-0.064734377,0.2220266163,0.0449401326,0.0490150303,-0.4242556393,0.2599344254,-0.3817819953,0.052681651,0.083897166,0.2135799378,0.4154089987,-0.0532197431,-0.1252102405,0.3929850459,-0.0362856239,0.015139333,0.2010873556,-0.1390540749,0.2122402787,-0.1353643239,0.2662816644,0.107059367,0.183084175,-0.1398240626,-0.012119934,0.2761809826,-0.1850383729,0.1524389237,-0.0569083467,-0.0053153718,0.1371273845,-0.2301723808,0.1511038244,-0.2182911783,0.0741774589,-0.106757611,0.175869137,-0.0325058177,-0.1456296444,0.177807942,0.183447957,-0.1542737037,0.1718087494,-0.0479231402,0.0327788517,-0.0946388319,0.0258924644,0.5896530747,0.6832953095,0.1840275824,0.0930984914,0.1825769097,0.0101524424,-0.173652187,0.4018145204,-0.2306769639,-0.1879463941,0.3399558961,0.1338035613,0.0912758708,-0.3779127896,-0.0704192966,0.1394079626,0.2234371454,-0.2576568425,-0.0136110792,-0.4151098728,-0.0433590114,-0.1474149376,0.2553113997,-0.0163310636,-0.3288873732,0.1003978476,-0.1700997353,0.0378062688,0.2827537656,-0.080297403,0.2193176299,-0.0255129263,0.3116757572,-0.1784826368,0.0661221445,-0.2623886168,0.1238634214,0.0868652239,0.131130144,0.4407131076,-0.1406992227,-0.3182706237,-0.3026193976,-0.203552708,0.1379104704,-0.0104992753,0.3564635217,0.3850200176,0.0186122004,-0.1929819733,-0.357203126,0.2555490136,-0.0953965262,-0.4541634917,0.0896259025,-0.0987125114,0.0308767278,-0.1922013015,-0.5104909539,-0.1726352721,-0.2520100176,0.0773612335,-0.1438827962,0.1406255066,0.3311223984,-0.0303112864,0.1407981217,0.0260429103,0.0988141224,-0.4347645342,-0.1841461509,0.0602049045,-0.2762188315,-0.297478646,-0.0169336833,-0.1227035522,0.315713346,-0.1157890856,-0.4564087093,-0.3070779145,-0.2492566258,0.3409292698,-0.1840697676,0.2868571877,0.1675689518,0.1474985778,-0.1426562071,-0.2416717261,-0.0538782403,0.2804362774,-0.2117324024,0.040080253,0.1509509385,0.5182060599,0.0724675581,0.6133498549,0.5002933145,0.0882742405,0.2791177928,-0.0928221941,0.2891702354,-0.1708509326,-0.4779526591,-0.2943904102,0.1661699414,0.2061332017,0.2267427891,-0.1307823807,0.1260145754,0.0918801799,-0.091425918,-0.4983181357,-0.2809929252,-0.2584249675,-0.2079364657,0.0594168864,0.0559903383,0.0944306627,-0.0589040853,0.1520696878,0.18828623,0.1523125619,0.0250050239,0.0236908644,-0.3282147944,0.0291185901,-0.4109385312,0.2250795364,-0.2089009136,0.3213142753,-0.1174951866,0.1202059984,0.0278093852,-0.0108156241,0.6370674372,-0.3670227826,0.1646800786,0.0842312351,-0.1588380337,-0.1211832836,-0.0301569253,-0.0135971233,0.1148917899,-0.0315824524,0.228514716,-0.5454471707,0.1311465502,0.2034050226,0.2716737092,-0.174193725,-0.2394859493,-0.0387507789,-0.4927273989,-0.498588264,-0.1477437764,0.0117497053,0.2282522768,-0.1652650833,-0.0506201908,0.0157096274,-0.1714573205,0.1775358617,0.1479468644,0.2621711791,0.2157793343,0.6348118782,0.1356801093,-0.1521138996,0.1617397368,0.6787909269,-0.1573693752,-0.2183036357,-0.2806072533,-0.1729327589,-0.2247734517,0.2754718661,-0.1484749019,0.3007211089,-0.1415906698,0.0204616878,0.2139416039,-0.3148276806,0.2035317868,-0.1699671,0.0089356564,-0.3534320891,0.0157118123,-0.0586178079,-0.2973249555,0.1694923192,0.1423399448,-0.2212103009,0.3661328554,-0.1916655004,0.8777389526,-0.0657743216,0.0092350086,0.1628193259,-0.2247971445,0.3618406057,0.0900778249,-0.1640419364,-0.4612333179,-0.300560534,0.158187151,0.0517323539,0.1566579789,0.1095851064,-0.299536407,0.2086388171,-0.1181925237,0.0969923064,-0.0475664362,0.2875964344,-0.2880068421,0.0701049641,-0.118870534,0.1461495161,0.0694422424,0.049491249,0.0489635803,-0.1171791106,0.1988665015,-0.4868976176,-0.2058993727,0.0336259902,-0.3923662305,0.3286435008,-0.1180856898,-0.3370572329,0.2058611065,0.4307642281,0.4622982144,0.2482508868,0.0187580381,0.0546287596,-0.0070002414,0.2869389653,-0.1258177012,-0.2343723774,-0.0116663417,-0.1105600446,-0.3548017442,-0.1090689451,0.1537514627,-0.082078144,-0.2488624305,0.1152273417,-0.0675447211,-0.1246659681,0.0234342292,-0.1411093771,-0.177653119,-0.4434713423,0.1303865016,0.131235823,-0.1754398644,0.2219368517,-0.188381508,-0.1558760554,-0.0415972248,0.3397012949,0.1883183569,0.2474416196,0.5482302904,0.0551555045,0.0425936803,-0.2531750798,-0.0076904814,0.0406281911,-0.4769050777,0.1915148795,-0.2640390694,0.1442859322,0.1088689938,0.3577151,0.2171230465,0.1366252601,0.1687989384,-0.3037644625,-0.1269867867,0.0299505889,0.1908327937,0.2174738497,0.1731236726,-0.0191916954,-0.1452844143,0.3274478018,-0.3689516187,0.2378379107,-0.0046652718,-0.0102385459,0.0007049664,-0.1199511737,0.0396756493,0.1199862361,0.2029100507,0.2261535078,-0.0559355654,-0.2679732144,-0.2521988153,0.1316014975,-0.218498081,-0.1682141423,-0.0694025606,-0.2364601046,-0.0586731881,-0.0839180946,0.4266134501,0.137183696,0.0029950093,0.1578784138,0.1529670358,0.1456464529,-0.1631798893,0.1880171001,-0.2690665126,0.016333811,0.1415002346,0.0948353857,-0.0189504791,-0.1715521067,-0.3372675776,0.0826946646,-0.0560484268,0.1779657453,0.2785222828,-0.2052522451,-0.2647082806,0.4296860993,0.297428906,0.0996681973,-0.1986869276,0.0051001068,0.0335961021,0.2705053985,-0.3992035091,-0.4369934499,0.1719170362,-0.1399970651,0.1139143407,0.3059450984,0.225457862,-0.3515104651,-0.2786740959,0.1173193231,0.383474797,-0.0804792121,0.222623691,0.448584646,-0.0854703635,-0.0763477683,0.0265165586,-0.1229587868,-0.0261203535,0.3270574212,-0.2358086705,0.3113099337,-0.1389688998,0.0790472999,0.2696990371,-0.3729182184,0.4485185444,0.1381930858,-0.252594918,0.022935573,0.1536448896,0.1603442281,-0.0875271633,0.0804734528,-0.2175440043,0.1058716252,-0.177334547,-0.0396665819,-0.1882720739,-0.2591831088,-0.1829653382,-0.0938074887,-0.0120979194,-0.1587508172,0.3571461737,-0.0653849542,-0.2649568617,-0.2993311584,0.0447383635,0.295560658,0.1373045295,-0.4245435894,0.0067302636,0.5594311953,-0.0806818008,-0.0790805444,0.2808078527,0.5457740426,0.1270319521,-0.0827852041,0.1383132339,0.1929411739,-0.0637374967,-0.1260762811,0.214398101,-0.0173141062,0.1705465317,0.223416239,0.1514259875,-0.2110050023,-0.2321826369,0.4559077621,-0.1453309059,-0.3635572493,0.192087546,-0.2378398627,-0.0323568098,-0.0969727114,0.3651038408,-0.2949544787,-0.1943055987,0.1872397214,0.2347214222,0.140871346,-0.4275604486,0.1380085349,-0.1543761939,0.1261260808,0.2467401326,-0.1832294464,-0.3851372898,-0.015749732,-0.7335856557,0.2009659857,-0.150406599,-0.070766449,0.0706763044,0.2328766584,-0.0135105625,0.0742938668,0.2322016209,0.2283024788,-0.0936845392,0.4965483248,-0.4009833038,-0.2250744253,0.1137502342,-0.1236983314,0.0905506536,-0.431661576,0.2643917501,-0.0150298644,0.0841797739,-0.055553969,0.0026559797,0.329531461,-0.2390793115,0.2479810417,0.301571548,0.4559423029,-0.1107432842,-0.2801846862,0.0129051777,-0.0368635356,-0.041126512,0.28016904,0.3217188418,0.2556381524,0.1058465466,-0.2885212898,-0.2245197296,0.3387840688,0.2152883261,0.0428710915,-0.3165065348,0.4041502774,-0.0445990339,0.0526788533,0.1392771304,0.3226180077,-0.0196398906,0.2745648026,-0.3327932656,-0.6344330311,0.3417711258,-0.322576791,-0.179686144,-0.274009347,0.3677443564,-0.0831620172,-0.1276747286,-0.2130842358,0.1397543401,0.429546237,-0.2491013557,0.0651856884,0.3638628423,-0.0633863732,0.1249608472,0.2045358419,0.1447397172,0.1241833046,-0.2364504784,0.5228363872,-0.1628024578]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2542","title":"`datasets.keyhash.DuplicatedKeysError` for `drop` and `adversarial_qa\/adversarialQA`","comments":"Hi @VictorSanh, thank you for reporting this issue with duplicated keys.\r\n\r\n- The issue with \"adversarial_qa\" was fixed 23 days ago: #2433. Current version of `datasets` (1.8.0) includes the patch.\r\n- I am investigating the issue with `drop`. I'll ping you to keep you informed.","body":"## Describe the bug\r\nFailure to generate the datasets (`drop` and subset `adversarialQA` from `adversarial_qa`) because of duplicate keys.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"drop\")\r\nload_dataset(\"adversarial_qa\", \"adversarialQA\")\r\n```\r\n\r\n## Expected results\r\nThe examples keys should be unique.\r\n\r\n## Actual results\r\n```bash\r\n>>> load_dataset(\"drop\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset drop\/default (download: 7.92 MiB, generated: 111.88 MiB, post-processed: Unknown size, total: 119.80 MiB) to \/home\/hf\/.cache\/huggingface\/datasets\/drop\/default\/0.1.0\/7a94f1e2bb26c4b5c75f89857c06982967d7416e5af935a9374b9bccf5068026...\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 28553293-d719-441b-8f00-ce3dc6df5398\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1044-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":45,"text":"`datasets.keyhash.DuplicatedKeysError` for `drop` and `adversarial_qa\/adversarialQA`\n## Describe the bug\r\nFailure to generate the datasets (`drop` and subset `adversarialQA` from `adversarial_qa`) because of duplicate keys.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"drop\")\r\nload_dataset(\"adversarial_qa\", \"adversarialQA\")\r\n```\r\n\r\n## Expected results\r\nThe examples keys should be unique.\r\n\r\n## Actual results\r\n```bash\r\n>>> load_dataset(\"drop\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset drop\/default (download: 7.92 MiB, generated: 111.88 MiB, post-processed: Unknown size, total: 119.80 MiB) to \/home\/hf\/.cache\/huggingface\/datasets\/drop\/default\/0.1.0\/7a94f1e2bb26c4b5c75f89857c06982967d7416e5af935a9374b9bccf5068026...\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 28553293-d719-441b-8f00-ce3dc6df5398\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1044-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\nHi @VictorSanh, thank you for reporting this issue with duplicated keys.\r\n\r\n- The issue with \"adversarial_qa\" was fixed 23 days ago: #2433. Current version of `datasets` (1.8.0) includes the patch.\r\n- I am investigating the issue with `drop`. I'll ping you to keep you informed.","embeddings":[-0.0905093253,-0.1160073131,0.00945515,0.470908016,0.0840874016,0.0286583472,0.3028091192,0.2397654057,0.1277973205,0.1814353019,-0.1439216137,0.5057473183,-0.048295714,0.1198933572,0.0266699288,-0.032785099,0.0012848077,0.0136723919,-0.2982798517,-0.0762103871,-0.2426401526,0.387050271,-0.2559149265,0.10335318,0.0598408356,-0.1042755619,-0.1572466195,0.3131027222,-0.0433670767,-0.2897949517,0.1407245398,0.0112546887,-0.2836559713,0.4620527029,-0.0001087736,-0.0179601684,0.1347592473,0.063810952,-0.4384645522,-0.1568871439,-0.2514858246,-0.0515594594,-0.0878485814,-0.2769972682,0.1303325295,-0.2024025172,-0.3193139136,-0.376553297,0.7059348822,0.0964307413,0.258453995,0.2994432449,0.0578671731,-0.2653636336,0.1903439015,0.03764534,-0.1994792819,0.2062627226,-0.0147559959,-0.0670088753,0.3541854918,0.160368681,-0.0145851849,-0.0506396517,0.2040350139,0.0678948611,0.0129833156,-0.1696084887,0.340441376,0.3239115477,0.2472787201,-0.5006279945,-0.2000300586,-0.2800831497,0.2844666243,-0.0895541832,0.1983643919,0.0411963165,-0.248807773,0.1426331103,0.066547215,0.1313625276,0.0587465577,-0.0792369768,0.1552489251,-0.0006441801,0.0497320481,0.0894895196,0.1573434621,-0.1723117828,-0.0185095444,-0.3647033572,0.0199675113,-0.1255145669,-0.5385659933,0.0129776504,-0.0629277751,0.0256692767,0.4364517331,0.0741026253,0.2214869559,-0.0984062999,-0.0482476018,0.0024116111,0.0399042964,0.3961126208,0.1762890518,0.0912107676,0.1504677236,0.1907208413,-0.2307083607,0.1641299576,0.0568992607,-0.1409792006,0.2863470614,0.1551881731,0.6009869576,-0.0693046823,-0.4248117208,0.2033160925,-0.1676324755,-0.124566488,-0.2327310294,0.1176542267,0.0935675427,-0.0219017286,-0.125678286,0.2016167492,-0.329765439,-0.1748307794,-0.2609870434,-0.0125090387,-0.0085719796,-0.0722294599,0.0838416889,-0.181197688,0.277915746,0.3087557852,-0.0046123383,-0.1379147619,0.2745801508,0.0929096639,-0.2666784823,0.0398336798,0.1705012321,0.0668969154,0.187929526,-0.3602023125,0.0600283071,0.1122215241,-0.0289515667,-0.1286867708,-0.1176909655,0.3158142269,0.0256931968,0.0972319841,-0.1766062826,0.1494122595,0.1119302288,0.0906810015,-0.0044025579,0.039072413,-0.1598896235,-0.4287461638,0.2847690582,0.7104182839,-0.3256116211,-0.1319526285,-0.1431197524,0.0301795825,0.1442250609,0.2293822169,-0.1759794354,0.2083805352,-0.3001039922,0.1173925847,0.0831043348,-0.342885524,-0.4917336702,0.2908804715,-0.3543904424,0.1900268048,0.1736855656,-0.0551680289,0.2153705508,-0.0025115306,0.011651733,-0.0165486876,-0.0907604471,-0.1353102922,-0.3101507425,-0.3366243541,0.1749132574,-0.2637753487,0.0284835584,0.0655985177,0.3573909402,-0.0704982579,0.2868715525,-0.0610246696,0.0467902683,0.0863300711,0.253734529,-0.0603276454,0.2205081582,-0.019394381,-0.5272825956,0.3890169859,-0.2856177986,0.0389644131,-0.1500831693,-0.2936635315,-0.5489760041,0.054138422,-0.4572040737,-0.2501839697,0.248770386,0.1914100796,-0.0284573697,0.1142127141,-0.064734377,0.2220266163,0.0449401326,0.0490150303,-0.4242556393,0.2599344254,-0.3817819953,0.052681651,0.083897166,0.2135799378,0.4154089987,-0.0532197431,-0.1252102405,0.3929850459,-0.0362856239,0.015139333,0.2010873556,-0.1390540749,0.2122402787,-0.1353643239,0.2662816644,0.107059367,0.183084175,-0.1398240626,-0.012119934,0.2761809826,-0.1850383729,0.1524389237,-0.0569083467,-0.0053153718,0.1371273845,-0.2301723808,0.1511038244,-0.2182911783,0.0741774589,-0.106757611,0.175869137,-0.0325058177,-0.1456296444,0.177807942,0.183447957,-0.1542737037,0.1718087494,-0.0479231402,0.0327788517,-0.0946388319,0.0258924644,0.5896530747,0.6832953095,0.1840275824,0.0930984914,0.1825769097,0.0101524424,-0.173652187,0.4018145204,-0.2306769639,-0.1879463941,0.3399558961,0.1338035613,0.0912758708,-0.3779127896,-0.0704192966,0.1394079626,0.2234371454,-0.2576568425,-0.0136110792,-0.4151098728,-0.0433590114,-0.1474149376,0.2553113997,-0.0163310636,-0.3288873732,0.1003978476,-0.1700997353,0.0378062688,0.2827537656,-0.080297403,0.2193176299,-0.0255129263,0.3116757572,-0.1784826368,0.0661221445,-0.2623886168,0.1238634214,0.0868652239,0.131130144,0.4407131076,-0.1406992227,-0.3182706237,-0.3026193976,-0.203552708,0.1379104704,-0.0104992753,0.3564635217,0.3850200176,0.0186122004,-0.1929819733,-0.357203126,0.2555490136,-0.0953965262,-0.4541634917,0.0896259025,-0.0987125114,0.0308767278,-0.1922013015,-0.5104909539,-0.1726352721,-0.2520100176,0.0773612335,-0.1438827962,0.1406255066,0.3311223984,-0.0303112864,0.1407981217,0.0260429103,0.0988141224,-0.4347645342,-0.1841461509,0.0602049045,-0.2762188315,-0.297478646,-0.0169336833,-0.1227035522,0.315713346,-0.1157890856,-0.4564087093,-0.3070779145,-0.2492566258,0.3409292698,-0.1840697676,0.2868571877,0.1675689518,0.1474985778,-0.1426562071,-0.2416717261,-0.0538782403,0.2804362774,-0.2117324024,0.040080253,0.1509509385,0.5182060599,0.0724675581,0.6133498549,0.5002933145,0.0882742405,0.2791177928,-0.0928221941,0.2891702354,-0.1708509326,-0.4779526591,-0.2943904102,0.1661699414,0.2061332017,0.2267427891,-0.1307823807,0.1260145754,0.0918801799,-0.091425918,-0.4983181357,-0.2809929252,-0.2584249675,-0.2079364657,0.0594168864,0.0559903383,0.0944306627,-0.0589040853,0.1520696878,0.18828623,0.1523125619,0.0250050239,0.0236908644,-0.3282147944,0.0291185901,-0.4109385312,0.2250795364,-0.2089009136,0.3213142753,-0.1174951866,0.1202059984,0.0278093852,-0.0108156241,0.6370674372,-0.3670227826,0.1646800786,0.0842312351,-0.1588380337,-0.1211832836,-0.0301569253,-0.0135971233,0.1148917899,-0.0315824524,0.228514716,-0.5454471707,0.1311465502,0.2034050226,0.2716737092,-0.174193725,-0.2394859493,-0.0387507789,-0.4927273989,-0.498588264,-0.1477437764,0.0117497053,0.2282522768,-0.1652650833,-0.0506201908,0.0157096274,-0.1714573205,0.1775358617,0.1479468644,0.2621711791,0.2157793343,0.6348118782,0.1356801093,-0.1521138996,0.1617397368,0.6787909269,-0.1573693752,-0.2183036357,-0.2806072533,-0.1729327589,-0.2247734517,0.2754718661,-0.1484749019,0.3007211089,-0.1415906698,0.0204616878,0.2139416039,-0.3148276806,0.2035317868,-0.1699671,0.0089356564,-0.3534320891,0.0157118123,-0.0586178079,-0.2973249555,0.1694923192,0.1423399448,-0.2212103009,0.3661328554,-0.1916655004,0.8777389526,-0.0657743216,0.0092350086,0.1628193259,-0.2247971445,0.3618406057,0.0900778249,-0.1640419364,-0.4612333179,-0.300560534,0.158187151,0.0517323539,0.1566579789,0.1095851064,-0.299536407,0.2086388171,-0.1181925237,0.0969923064,-0.0475664362,0.2875964344,-0.2880068421,0.0701049641,-0.118870534,0.1461495161,0.0694422424,0.049491249,0.0489635803,-0.1171791106,0.1988665015,-0.4868976176,-0.2058993727,0.0336259902,-0.3923662305,0.3286435008,-0.1180856898,-0.3370572329,0.2058611065,0.4307642281,0.4622982144,0.2482508868,0.0187580381,0.0546287596,-0.0070002414,0.2869389653,-0.1258177012,-0.2343723774,-0.0116663417,-0.1105600446,-0.3548017442,-0.1090689451,0.1537514627,-0.082078144,-0.2488624305,0.1152273417,-0.0675447211,-0.1246659681,0.0234342292,-0.1411093771,-0.177653119,-0.4434713423,0.1303865016,0.131235823,-0.1754398644,0.2219368517,-0.188381508,-0.1558760554,-0.0415972248,0.3397012949,0.1883183569,0.2474416196,0.5482302904,0.0551555045,0.0425936803,-0.2531750798,-0.0076904814,0.0406281911,-0.4769050777,0.1915148795,-0.2640390694,0.1442859322,0.1088689938,0.3577151,0.2171230465,0.1366252601,0.1687989384,-0.3037644625,-0.1269867867,0.0299505889,0.1908327937,0.2174738497,0.1731236726,-0.0191916954,-0.1452844143,0.3274478018,-0.3689516187,0.2378379107,-0.0046652718,-0.0102385459,0.0007049664,-0.1199511737,0.0396756493,0.1199862361,0.2029100507,0.2261535078,-0.0559355654,-0.2679732144,-0.2521988153,0.1316014975,-0.218498081,-0.1682141423,-0.0694025606,-0.2364601046,-0.0586731881,-0.0839180946,0.4266134501,0.137183696,0.0029950093,0.1578784138,0.1529670358,0.1456464529,-0.1631798893,0.1880171001,-0.2690665126,0.016333811,0.1415002346,0.0948353857,-0.0189504791,-0.1715521067,-0.3372675776,0.0826946646,-0.0560484268,0.1779657453,0.2785222828,-0.2052522451,-0.2647082806,0.4296860993,0.297428906,0.0996681973,-0.1986869276,0.0051001068,0.0335961021,0.2705053985,-0.3992035091,-0.4369934499,0.1719170362,-0.1399970651,0.1139143407,0.3059450984,0.225457862,-0.3515104651,-0.2786740959,0.1173193231,0.383474797,-0.0804792121,0.222623691,0.448584646,-0.0854703635,-0.0763477683,0.0265165586,-0.1229587868,-0.0261203535,0.3270574212,-0.2358086705,0.3113099337,-0.1389688998,0.0790472999,0.2696990371,-0.3729182184,0.4485185444,0.1381930858,-0.252594918,0.022935573,0.1536448896,0.1603442281,-0.0875271633,0.0804734528,-0.2175440043,0.1058716252,-0.177334547,-0.0396665819,-0.1882720739,-0.2591831088,-0.1829653382,-0.0938074887,-0.0120979194,-0.1587508172,0.3571461737,-0.0653849542,-0.2649568617,-0.2993311584,0.0447383635,0.295560658,0.1373045295,-0.4245435894,0.0067302636,0.5594311953,-0.0806818008,-0.0790805444,0.2808078527,0.5457740426,0.1270319521,-0.0827852041,0.1383132339,0.1929411739,-0.0637374967,-0.1260762811,0.214398101,-0.0173141062,0.1705465317,0.223416239,0.1514259875,-0.2110050023,-0.2321826369,0.4559077621,-0.1453309059,-0.3635572493,0.192087546,-0.2378398627,-0.0323568098,-0.0969727114,0.3651038408,-0.2949544787,-0.1943055987,0.1872397214,0.2347214222,0.140871346,-0.4275604486,0.1380085349,-0.1543761939,0.1261260808,0.2467401326,-0.1832294464,-0.3851372898,-0.015749732,-0.7335856557,0.2009659857,-0.150406599,-0.070766449,0.0706763044,0.2328766584,-0.0135105625,0.0742938668,0.2322016209,0.2283024788,-0.0936845392,0.4965483248,-0.4009833038,-0.2250744253,0.1137502342,-0.1236983314,0.0905506536,-0.431661576,0.2643917501,-0.0150298644,0.0841797739,-0.055553969,0.0026559797,0.329531461,-0.2390793115,0.2479810417,0.301571548,0.4559423029,-0.1107432842,-0.2801846862,0.0129051777,-0.0368635356,-0.041126512,0.28016904,0.3217188418,0.2556381524,0.1058465466,-0.2885212898,-0.2245197296,0.3387840688,0.2152883261,0.0428710915,-0.3165065348,0.4041502774,-0.0445990339,0.0526788533,0.1392771304,0.3226180077,-0.0196398906,0.2745648026,-0.3327932656,-0.6344330311,0.3417711258,-0.322576791,-0.179686144,-0.274009347,0.3677443564,-0.0831620172,-0.1276747286,-0.2130842358,0.1397543401,0.429546237,-0.2491013557,0.0651856884,0.3638628423,-0.0633863732,0.1249608472,0.2045358419,0.1447397172,0.1241833046,-0.2364504784,0.5228363872,-0.1628024578]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2542","title":"`datasets.keyhash.DuplicatedKeysError` for `drop` and `adversarial_qa\/adversarialQA`","comments":"Hi @VictorSanh, the issue is already fixed and merged into master branch and will be included in our next release version 1.9.0.","body":"## Describe the bug\r\nFailure to generate the datasets (`drop` and subset `adversarialQA` from `adversarial_qa`) because of duplicate keys.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"drop\")\r\nload_dataset(\"adversarial_qa\", \"adversarialQA\")\r\n```\r\n\r\n## Expected results\r\nThe examples keys should be unique.\r\n\r\n## Actual results\r\n```bash\r\n>>> load_dataset(\"drop\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset drop\/default (download: 7.92 MiB, generated: 111.88 MiB, post-processed: Unknown size, total: 119.80 MiB) to \/home\/hf\/.cache\/huggingface\/datasets\/drop\/default\/0.1.0\/7a94f1e2bb26c4b5c75f89857c06982967d7416e5af935a9374b9bccf5068026...\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 28553293-d719-441b-8f00-ce3dc6df5398\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1044-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":22,"text":"`datasets.keyhash.DuplicatedKeysError` for `drop` and `adversarial_qa\/adversarialQA`\n## Describe the bug\r\nFailure to generate the datasets (`drop` and subset `adversarialQA` from `adversarial_qa`) because of duplicate keys.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"drop\")\r\nload_dataset(\"adversarial_qa\", \"adversarialQA\")\r\n```\r\n\r\n## Expected results\r\nThe examples keys should be unique.\r\n\r\n## Actual results\r\n```bash\r\n>>> load_dataset(\"drop\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset drop\/default (download: 7.92 MiB, generated: 111.88 MiB, post-processed: Unknown size, total: 119.80 MiB) to \/home\/hf\/.cache\/huggingface\/datasets\/drop\/default\/0.1.0\/7a94f1e2bb26c4b5c75f89857c06982967d7416e5af935a9374b9bccf5068026...\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 28553293-d719-441b-8f00-ce3dc6df5398\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1044-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\nHi @VictorSanh, the issue is already fixed and merged into master branch and will be included in our next release version 1.9.0.","embeddings":[-0.0905093253,-0.1160073131,0.00945515,0.470908016,0.0840874016,0.0286583472,0.3028091192,0.2397654057,0.1277973205,0.1814353019,-0.1439216137,0.5057473183,-0.048295714,0.1198933572,0.0266699288,-0.032785099,0.0012848077,0.0136723919,-0.2982798517,-0.0762103871,-0.2426401526,0.387050271,-0.2559149265,0.10335318,0.0598408356,-0.1042755619,-0.1572466195,0.3131027222,-0.0433670767,-0.2897949517,0.1407245398,0.0112546887,-0.2836559713,0.4620527029,-0.0001087736,-0.0179601684,0.1347592473,0.063810952,-0.4384645522,-0.1568871439,-0.2514858246,-0.0515594594,-0.0878485814,-0.2769972682,0.1303325295,-0.2024025172,-0.3193139136,-0.376553297,0.7059348822,0.0964307413,0.258453995,0.2994432449,0.0578671731,-0.2653636336,0.1903439015,0.03764534,-0.1994792819,0.2062627226,-0.0147559959,-0.0670088753,0.3541854918,0.160368681,-0.0145851849,-0.0506396517,0.2040350139,0.0678948611,0.0129833156,-0.1696084887,0.340441376,0.3239115477,0.2472787201,-0.5006279945,-0.2000300586,-0.2800831497,0.2844666243,-0.0895541832,0.1983643919,0.0411963165,-0.248807773,0.1426331103,0.066547215,0.1313625276,0.0587465577,-0.0792369768,0.1552489251,-0.0006441801,0.0497320481,0.0894895196,0.1573434621,-0.1723117828,-0.0185095444,-0.3647033572,0.0199675113,-0.1255145669,-0.5385659933,0.0129776504,-0.0629277751,0.0256692767,0.4364517331,0.0741026253,0.2214869559,-0.0984062999,-0.0482476018,0.0024116111,0.0399042964,0.3961126208,0.1762890518,0.0912107676,0.1504677236,0.1907208413,-0.2307083607,0.1641299576,0.0568992607,-0.1409792006,0.2863470614,0.1551881731,0.6009869576,-0.0693046823,-0.4248117208,0.2033160925,-0.1676324755,-0.124566488,-0.2327310294,0.1176542267,0.0935675427,-0.0219017286,-0.125678286,0.2016167492,-0.329765439,-0.1748307794,-0.2609870434,-0.0125090387,-0.0085719796,-0.0722294599,0.0838416889,-0.181197688,0.277915746,0.3087557852,-0.0046123383,-0.1379147619,0.2745801508,0.0929096639,-0.2666784823,0.0398336798,0.1705012321,0.0668969154,0.187929526,-0.3602023125,0.0600283071,0.1122215241,-0.0289515667,-0.1286867708,-0.1176909655,0.3158142269,0.0256931968,0.0972319841,-0.1766062826,0.1494122595,0.1119302288,0.0906810015,-0.0044025579,0.039072413,-0.1598896235,-0.4287461638,0.2847690582,0.7104182839,-0.3256116211,-0.1319526285,-0.1431197524,0.0301795825,0.1442250609,0.2293822169,-0.1759794354,0.2083805352,-0.3001039922,0.1173925847,0.0831043348,-0.342885524,-0.4917336702,0.2908804715,-0.3543904424,0.1900268048,0.1736855656,-0.0551680289,0.2153705508,-0.0025115306,0.011651733,-0.0165486876,-0.0907604471,-0.1353102922,-0.3101507425,-0.3366243541,0.1749132574,-0.2637753487,0.0284835584,0.0655985177,0.3573909402,-0.0704982579,0.2868715525,-0.0610246696,0.0467902683,0.0863300711,0.253734529,-0.0603276454,0.2205081582,-0.019394381,-0.5272825956,0.3890169859,-0.2856177986,0.0389644131,-0.1500831693,-0.2936635315,-0.5489760041,0.054138422,-0.4572040737,-0.2501839697,0.248770386,0.1914100796,-0.0284573697,0.1142127141,-0.064734377,0.2220266163,0.0449401326,0.0490150303,-0.4242556393,0.2599344254,-0.3817819953,0.052681651,0.083897166,0.2135799378,0.4154089987,-0.0532197431,-0.1252102405,0.3929850459,-0.0362856239,0.015139333,0.2010873556,-0.1390540749,0.2122402787,-0.1353643239,0.2662816644,0.107059367,0.183084175,-0.1398240626,-0.012119934,0.2761809826,-0.1850383729,0.1524389237,-0.0569083467,-0.0053153718,0.1371273845,-0.2301723808,0.1511038244,-0.2182911783,0.0741774589,-0.106757611,0.175869137,-0.0325058177,-0.1456296444,0.177807942,0.183447957,-0.1542737037,0.1718087494,-0.0479231402,0.0327788517,-0.0946388319,0.0258924644,0.5896530747,0.6832953095,0.1840275824,0.0930984914,0.1825769097,0.0101524424,-0.173652187,0.4018145204,-0.2306769639,-0.1879463941,0.3399558961,0.1338035613,0.0912758708,-0.3779127896,-0.0704192966,0.1394079626,0.2234371454,-0.2576568425,-0.0136110792,-0.4151098728,-0.0433590114,-0.1474149376,0.2553113997,-0.0163310636,-0.3288873732,0.1003978476,-0.1700997353,0.0378062688,0.2827537656,-0.080297403,0.2193176299,-0.0255129263,0.3116757572,-0.1784826368,0.0661221445,-0.2623886168,0.1238634214,0.0868652239,0.131130144,0.4407131076,-0.1406992227,-0.3182706237,-0.3026193976,-0.203552708,0.1379104704,-0.0104992753,0.3564635217,0.3850200176,0.0186122004,-0.1929819733,-0.357203126,0.2555490136,-0.0953965262,-0.4541634917,0.0896259025,-0.0987125114,0.0308767278,-0.1922013015,-0.5104909539,-0.1726352721,-0.2520100176,0.0773612335,-0.1438827962,0.1406255066,0.3311223984,-0.0303112864,0.1407981217,0.0260429103,0.0988141224,-0.4347645342,-0.1841461509,0.0602049045,-0.2762188315,-0.297478646,-0.0169336833,-0.1227035522,0.315713346,-0.1157890856,-0.4564087093,-0.3070779145,-0.2492566258,0.3409292698,-0.1840697676,0.2868571877,0.1675689518,0.1474985778,-0.1426562071,-0.2416717261,-0.0538782403,0.2804362774,-0.2117324024,0.040080253,0.1509509385,0.5182060599,0.0724675581,0.6133498549,0.5002933145,0.0882742405,0.2791177928,-0.0928221941,0.2891702354,-0.1708509326,-0.4779526591,-0.2943904102,0.1661699414,0.2061332017,0.2267427891,-0.1307823807,0.1260145754,0.0918801799,-0.091425918,-0.4983181357,-0.2809929252,-0.2584249675,-0.2079364657,0.0594168864,0.0559903383,0.0944306627,-0.0589040853,0.1520696878,0.18828623,0.1523125619,0.0250050239,0.0236908644,-0.3282147944,0.0291185901,-0.4109385312,0.2250795364,-0.2089009136,0.3213142753,-0.1174951866,0.1202059984,0.0278093852,-0.0108156241,0.6370674372,-0.3670227826,0.1646800786,0.0842312351,-0.1588380337,-0.1211832836,-0.0301569253,-0.0135971233,0.1148917899,-0.0315824524,0.228514716,-0.5454471707,0.1311465502,0.2034050226,0.2716737092,-0.174193725,-0.2394859493,-0.0387507789,-0.4927273989,-0.498588264,-0.1477437764,0.0117497053,0.2282522768,-0.1652650833,-0.0506201908,0.0157096274,-0.1714573205,0.1775358617,0.1479468644,0.2621711791,0.2157793343,0.6348118782,0.1356801093,-0.1521138996,0.1617397368,0.6787909269,-0.1573693752,-0.2183036357,-0.2806072533,-0.1729327589,-0.2247734517,0.2754718661,-0.1484749019,0.3007211089,-0.1415906698,0.0204616878,0.2139416039,-0.3148276806,0.2035317868,-0.1699671,0.0089356564,-0.3534320891,0.0157118123,-0.0586178079,-0.2973249555,0.1694923192,0.1423399448,-0.2212103009,0.3661328554,-0.1916655004,0.8777389526,-0.0657743216,0.0092350086,0.1628193259,-0.2247971445,0.3618406057,0.0900778249,-0.1640419364,-0.4612333179,-0.300560534,0.158187151,0.0517323539,0.1566579789,0.1095851064,-0.299536407,0.2086388171,-0.1181925237,0.0969923064,-0.0475664362,0.2875964344,-0.2880068421,0.0701049641,-0.118870534,0.1461495161,0.0694422424,0.049491249,0.0489635803,-0.1171791106,0.1988665015,-0.4868976176,-0.2058993727,0.0336259902,-0.3923662305,0.3286435008,-0.1180856898,-0.3370572329,0.2058611065,0.4307642281,0.4622982144,0.2482508868,0.0187580381,0.0546287596,-0.0070002414,0.2869389653,-0.1258177012,-0.2343723774,-0.0116663417,-0.1105600446,-0.3548017442,-0.1090689451,0.1537514627,-0.082078144,-0.2488624305,0.1152273417,-0.0675447211,-0.1246659681,0.0234342292,-0.1411093771,-0.177653119,-0.4434713423,0.1303865016,0.131235823,-0.1754398644,0.2219368517,-0.188381508,-0.1558760554,-0.0415972248,0.3397012949,0.1883183569,0.2474416196,0.5482302904,0.0551555045,0.0425936803,-0.2531750798,-0.0076904814,0.0406281911,-0.4769050777,0.1915148795,-0.2640390694,0.1442859322,0.1088689938,0.3577151,0.2171230465,0.1366252601,0.1687989384,-0.3037644625,-0.1269867867,0.0299505889,0.1908327937,0.2174738497,0.1731236726,-0.0191916954,-0.1452844143,0.3274478018,-0.3689516187,0.2378379107,-0.0046652718,-0.0102385459,0.0007049664,-0.1199511737,0.0396756493,0.1199862361,0.2029100507,0.2261535078,-0.0559355654,-0.2679732144,-0.2521988153,0.1316014975,-0.218498081,-0.1682141423,-0.0694025606,-0.2364601046,-0.0586731881,-0.0839180946,0.4266134501,0.137183696,0.0029950093,0.1578784138,0.1529670358,0.1456464529,-0.1631798893,0.1880171001,-0.2690665126,0.016333811,0.1415002346,0.0948353857,-0.0189504791,-0.1715521067,-0.3372675776,0.0826946646,-0.0560484268,0.1779657453,0.2785222828,-0.2052522451,-0.2647082806,0.4296860993,0.297428906,0.0996681973,-0.1986869276,0.0051001068,0.0335961021,0.2705053985,-0.3992035091,-0.4369934499,0.1719170362,-0.1399970651,0.1139143407,0.3059450984,0.225457862,-0.3515104651,-0.2786740959,0.1173193231,0.383474797,-0.0804792121,0.222623691,0.448584646,-0.0854703635,-0.0763477683,0.0265165586,-0.1229587868,-0.0261203535,0.3270574212,-0.2358086705,0.3113099337,-0.1389688998,0.0790472999,0.2696990371,-0.3729182184,0.4485185444,0.1381930858,-0.252594918,0.022935573,0.1536448896,0.1603442281,-0.0875271633,0.0804734528,-0.2175440043,0.1058716252,-0.177334547,-0.0396665819,-0.1882720739,-0.2591831088,-0.1829653382,-0.0938074887,-0.0120979194,-0.1587508172,0.3571461737,-0.0653849542,-0.2649568617,-0.2993311584,0.0447383635,0.295560658,0.1373045295,-0.4245435894,0.0067302636,0.5594311953,-0.0806818008,-0.0790805444,0.2808078527,0.5457740426,0.1270319521,-0.0827852041,0.1383132339,0.1929411739,-0.0637374967,-0.1260762811,0.214398101,-0.0173141062,0.1705465317,0.223416239,0.1514259875,-0.2110050023,-0.2321826369,0.4559077621,-0.1453309059,-0.3635572493,0.192087546,-0.2378398627,-0.0323568098,-0.0969727114,0.3651038408,-0.2949544787,-0.1943055987,0.1872397214,0.2347214222,0.140871346,-0.4275604486,0.1380085349,-0.1543761939,0.1261260808,0.2467401326,-0.1832294464,-0.3851372898,-0.015749732,-0.7335856557,0.2009659857,-0.150406599,-0.070766449,0.0706763044,0.2328766584,-0.0135105625,0.0742938668,0.2322016209,0.2283024788,-0.0936845392,0.4965483248,-0.4009833038,-0.2250744253,0.1137502342,-0.1236983314,0.0905506536,-0.431661576,0.2643917501,-0.0150298644,0.0841797739,-0.055553969,0.0026559797,0.329531461,-0.2390793115,0.2479810417,0.301571548,0.4559423029,-0.1107432842,-0.2801846862,0.0129051777,-0.0368635356,-0.041126512,0.28016904,0.3217188418,0.2556381524,0.1058465466,-0.2885212898,-0.2245197296,0.3387840688,0.2152883261,0.0428710915,-0.3165065348,0.4041502774,-0.0445990339,0.0526788533,0.1392771304,0.3226180077,-0.0196398906,0.2745648026,-0.3327932656,-0.6344330311,0.3417711258,-0.322576791,-0.179686144,-0.274009347,0.3677443564,-0.0831620172,-0.1276747286,-0.2130842358,0.1397543401,0.429546237,-0.2491013557,0.0651856884,0.3638628423,-0.0633863732,0.1249608472,0.2045358419,0.1447397172,0.1241833046,-0.2364504784,0.5228363872,-0.1628024578]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2542","title":"`datasets.keyhash.DuplicatedKeysError` for `drop` and `adversarial_qa\/adversarialQA`","comments":"thank you!","body":"## Describe the bug\r\nFailure to generate the datasets (`drop` and subset `adversarialQA` from `adversarial_qa`) because of duplicate keys.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"drop\")\r\nload_dataset(\"adversarial_qa\", \"adversarialQA\")\r\n```\r\n\r\n## Expected results\r\nThe examples keys should be unique.\r\n\r\n## Actual results\r\n```bash\r\n>>> load_dataset(\"drop\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset drop\/default (download: 7.92 MiB, generated: 111.88 MiB, post-processed: Unknown size, total: 119.80 MiB) to \/home\/hf\/.cache\/huggingface\/datasets\/drop\/default\/0.1.0\/7a94f1e2bb26c4b5c75f89857c06982967d7416e5af935a9374b9bccf5068026...\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 28553293-d719-441b-8f00-ce3dc6df5398\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1044-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":2,"text":"`datasets.keyhash.DuplicatedKeysError` for `drop` and `adversarial_qa\/adversarialQA`\n## Describe the bug\r\nFailure to generate the datasets (`drop` and subset `adversarialQA` from `adversarial_qa`) because of duplicate keys.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"drop\")\r\nload_dataset(\"adversarial_qa\", \"adversarialQA\")\r\n```\r\n\r\n## Expected results\r\nThe examples keys should be unique.\r\n\r\n## Actual results\r\n```bash\r\n>>> load_dataset(\"drop\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset drop\/default (download: 7.92 MiB, generated: 111.88 MiB, post-processed: Unknown size, total: 119.80 MiB) to \/home\/hf\/.cache\/huggingface\/datasets\/drop\/default\/0.1.0\/7a94f1e2bb26c4b5c75f89857c06982967d7416e5af935a9374b9bccf5068026...\r\nTraceback (most recent call last): \r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 992, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 409, in finalize\r\n self.check_duplicate_keys()\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 349, in check_duplicate_keys\r\n raise DuplicatedKeysError(key)\r\ndatasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 28553293-d719-441b-8f00-ce3dc6df5398\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.0-1044-gcp-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\nthank you!","embeddings":[-0.0905093253,-0.1160073131,0.00945515,0.470908016,0.0840874016,0.0286583472,0.3028091192,0.2397654057,0.1277973205,0.1814353019,-0.1439216137,0.5057473183,-0.048295714,0.1198933572,0.0266699288,-0.032785099,0.0012848077,0.0136723919,-0.2982798517,-0.0762103871,-0.2426401526,0.387050271,-0.2559149265,0.10335318,0.0598408356,-0.1042755619,-0.1572466195,0.3131027222,-0.0433670767,-0.2897949517,0.1407245398,0.0112546887,-0.2836559713,0.4620527029,-0.0001087736,-0.0179601684,0.1347592473,0.063810952,-0.4384645522,-0.1568871439,-0.2514858246,-0.0515594594,-0.0878485814,-0.2769972682,0.1303325295,-0.2024025172,-0.3193139136,-0.376553297,0.7059348822,0.0964307413,0.258453995,0.2994432449,0.0578671731,-0.2653636336,0.1903439015,0.03764534,-0.1994792819,0.2062627226,-0.0147559959,-0.0670088753,0.3541854918,0.160368681,-0.0145851849,-0.0506396517,0.2040350139,0.0678948611,0.0129833156,-0.1696084887,0.340441376,0.3239115477,0.2472787201,-0.5006279945,-0.2000300586,-0.2800831497,0.2844666243,-0.0895541832,0.1983643919,0.0411963165,-0.248807773,0.1426331103,0.066547215,0.1313625276,0.0587465577,-0.0792369768,0.1552489251,-0.0006441801,0.0497320481,0.0894895196,0.1573434621,-0.1723117828,-0.0185095444,-0.3647033572,0.0199675113,-0.1255145669,-0.5385659933,0.0129776504,-0.0629277751,0.0256692767,0.4364517331,0.0741026253,0.2214869559,-0.0984062999,-0.0482476018,0.0024116111,0.0399042964,0.3961126208,0.1762890518,0.0912107676,0.1504677236,0.1907208413,-0.2307083607,0.1641299576,0.0568992607,-0.1409792006,0.2863470614,0.1551881731,0.6009869576,-0.0693046823,-0.4248117208,0.2033160925,-0.1676324755,-0.124566488,-0.2327310294,0.1176542267,0.0935675427,-0.0219017286,-0.125678286,0.2016167492,-0.329765439,-0.1748307794,-0.2609870434,-0.0125090387,-0.0085719796,-0.0722294599,0.0838416889,-0.181197688,0.277915746,0.3087557852,-0.0046123383,-0.1379147619,0.2745801508,0.0929096639,-0.2666784823,0.0398336798,0.1705012321,0.0668969154,0.187929526,-0.3602023125,0.0600283071,0.1122215241,-0.0289515667,-0.1286867708,-0.1176909655,0.3158142269,0.0256931968,0.0972319841,-0.1766062826,0.1494122595,0.1119302288,0.0906810015,-0.0044025579,0.039072413,-0.1598896235,-0.4287461638,0.2847690582,0.7104182839,-0.3256116211,-0.1319526285,-0.1431197524,0.0301795825,0.1442250609,0.2293822169,-0.1759794354,0.2083805352,-0.3001039922,0.1173925847,0.0831043348,-0.342885524,-0.4917336702,0.2908804715,-0.3543904424,0.1900268048,0.1736855656,-0.0551680289,0.2153705508,-0.0025115306,0.011651733,-0.0165486876,-0.0907604471,-0.1353102922,-0.3101507425,-0.3366243541,0.1749132574,-0.2637753487,0.0284835584,0.0655985177,0.3573909402,-0.0704982579,0.2868715525,-0.0610246696,0.0467902683,0.0863300711,0.253734529,-0.0603276454,0.2205081582,-0.019394381,-0.5272825956,0.3890169859,-0.2856177986,0.0389644131,-0.1500831693,-0.2936635315,-0.5489760041,0.054138422,-0.4572040737,-0.2501839697,0.248770386,0.1914100796,-0.0284573697,0.1142127141,-0.064734377,0.2220266163,0.0449401326,0.0490150303,-0.4242556393,0.2599344254,-0.3817819953,0.052681651,0.083897166,0.2135799378,0.4154089987,-0.0532197431,-0.1252102405,0.3929850459,-0.0362856239,0.015139333,0.2010873556,-0.1390540749,0.2122402787,-0.1353643239,0.2662816644,0.107059367,0.183084175,-0.1398240626,-0.012119934,0.2761809826,-0.1850383729,0.1524389237,-0.0569083467,-0.0053153718,0.1371273845,-0.2301723808,0.1511038244,-0.2182911783,0.0741774589,-0.106757611,0.175869137,-0.0325058177,-0.1456296444,0.177807942,0.183447957,-0.1542737037,0.1718087494,-0.0479231402,0.0327788517,-0.0946388319,0.0258924644,0.5896530747,0.6832953095,0.1840275824,0.0930984914,0.1825769097,0.0101524424,-0.173652187,0.4018145204,-0.2306769639,-0.1879463941,0.3399558961,0.1338035613,0.0912758708,-0.3779127896,-0.0704192966,0.1394079626,0.2234371454,-0.2576568425,-0.0136110792,-0.4151098728,-0.0433590114,-0.1474149376,0.2553113997,-0.0163310636,-0.3288873732,0.1003978476,-0.1700997353,0.0378062688,0.2827537656,-0.080297403,0.2193176299,-0.0255129263,0.3116757572,-0.1784826368,0.0661221445,-0.2623886168,0.1238634214,0.0868652239,0.131130144,0.4407131076,-0.1406992227,-0.3182706237,-0.3026193976,-0.203552708,0.1379104704,-0.0104992753,0.3564635217,0.3850200176,0.0186122004,-0.1929819733,-0.357203126,0.2555490136,-0.0953965262,-0.4541634917,0.0896259025,-0.0987125114,0.0308767278,-0.1922013015,-0.5104909539,-0.1726352721,-0.2520100176,0.0773612335,-0.1438827962,0.1406255066,0.3311223984,-0.0303112864,0.1407981217,0.0260429103,0.0988141224,-0.4347645342,-0.1841461509,0.0602049045,-0.2762188315,-0.297478646,-0.0169336833,-0.1227035522,0.315713346,-0.1157890856,-0.4564087093,-0.3070779145,-0.2492566258,0.3409292698,-0.1840697676,0.2868571877,0.1675689518,0.1474985778,-0.1426562071,-0.2416717261,-0.0538782403,0.2804362774,-0.2117324024,0.040080253,0.1509509385,0.5182060599,0.0724675581,0.6133498549,0.5002933145,0.0882742405,0.2791177928,-0.0928221941,0.2891702354,-0.1708509326,-0.4779526591,-0.2943904102,0.1661699414,0.2061332017,0.2267427891,-0.1307823807,0.1260145754,0.0918801799,-0.091425918,-0.4983181357,-0.2809929252,-0.2584249675,-0.2079364657,0.0594168864,0.0559903383,0.0944306627,-0.0589040853,0.1520696878,0.18828623,0.1523125619,0.0250050239,0.0236908644,-0.3282147944,0.0291185901,-0.4109385312,0.2250795364,-0.2089009136,0.3213142753,-0.1174951866,0.1202059984,0.0278093852,-0.0108156241,0.6370674372,-0.3670227826,0.1646800786,0.0842312351,-0.1588380337,-0.1211832836,-0.0301569253,-0.0135971233,0.1148917899,-0.0315824524,0.228514716,-0.5454471707,0.1311465502,0.2034050226,0.2716737092,-0.174193725,-0.2394859493,-0.0387507789,-0.4927273989,-0.498588264,-0.1477437764,0.0117497053,0.2282522768,-0.1652650833,-0.0506201908,0.0157096274,-0.1714573205,0.1775358617,0.1479468644,0.2621711791,0.2157793343,0.6348118782,0.1356801093,-0.1521138996,0.1617397368,0.6787909269,-0.1573693752,-0.2183036357,-0.2806072533,-0.1729327589,-0.2247734517,0.2754718661,-0.1484749019,0.3007211089,-0.1415906698,0.0204616878,0.2139416039,-0.3148276806,0.2035317868,-0.1699671,0.0089356564,-0.3534320891,0.0157118123,-0.0586178079,-0.2973249555,0.1694923192,0.1423399448,-0.2212103009,0.3661328554,-0.1916655004,0.8777389526,-0.0657743216,0.0092350086,0.1628193259,-0.2247971445,0.3618406057,0.0900778249,-0.1640419364,-0.4612333179,-0.300560534,0.158187151,0.0517323539,0.1566579789,0.1095851064,-0.299536407,0.2086388171,-0.1181925237,0.0969923064,-0.0475664362,0.2875964344,-0.2880068421,0.0701049641,-0.118870534,0.1461495161,0.0694422424,0.049491249,0.0489635803,-0.1171791106,0.1988665015,-0.4868976176,-0.2058993727,0.0336259902,-0.3923662305,0.3286435008,-0.1180856898,-0.3370572329,0.2058611065,0.4307642281,0.4622982144,0.2482508868,0.0187580381,0.0546287596,-0.0070002414,0.2869389653,-0.1258177012,-0.2343723774,-0.0116663417,-0.1105600446,-0.3548017442,-0.1090689451,0.1537514627,-0.082078144,-0.2488624305,0.1152273417,-0.0675447211,-0.1246659681,0.0234342292,-0.1411093771,-0.177653119,-0.4434713423,0.1303865016,0.131235823,-0.1754398644,0.2219368517,-0.188381508,-0.1558760554,-0.0415972248,0.3397012949,0.1883183569,0.2474416196,0.5482302904,0.0551555045,0.0425936803,-0.2531750798,-0.0076904814,0.0406281911,-0.4769050777,0.1915148795,-0.2640390694,0.1442859322,0.1088689938,0.3577151,0.2171230465,0.1366252601,0.1687989384,-0.3037644625,-0.1269867867,0.0299505889,0.1908327937,0.2174738497,0.1731236726,-0.0191916954,-0.1452844143,0.3274478018,-0.3689516187,0.2378379107,-0.0046652718,-0.0102385459,0.0007049664,-0.1199511737,0.0396756493,0.1199862361,0.2029100507,0.2261535078,-0.0559355654,-0.2679732144,-0.2521988153,0.1316014975,-0.218498081,-0.1682141423,-0.0694025606,-0.2364601046,-0.0586731881,-0.0839180946,0.4266134501,0.137183696,0.0029950093,0.1578784138,0.1529670358,0.1456464529,-0.1631798893,0.1880171001,-0.2690665126,0.016333811,0.1415002346,0.0948353857,-0.0189504791,-0.1715521067,-0.3372675776,0.0826946646,-0.0560484268,0.1779657453,0.2785222828,-0.2052522451,-0.2647082806,0.4296860993,0.297428906,0.0996681973,-0.1986869276,0.0051001068,0.0335961021,0.2705053985,-0.3992035091,-0.4369934499,0.1719170362,-0.1399970651,0.1139143407,0.3059450984,0.225457862,-0.3515104651,-0.2786740959,0.1173193231,0.383474797,-0.0804792121,0.222623691,0.448584646,-0.0854703635,-0.0763477683,0.0265165586,-0.1229587868,-0.0261203535,0.3270574212,-0.2358086705,0.3113099337,-0.1389688998,0.0790472999,0.2696990371,-0.3729182184,0.4485185444,0.1381930858,-0.252594918,0.022935573,0.1536448896,0.1603442281,-0.0875271633,0.0804734528,-0.2175440043,0.1058716252,-0.177334547,-0.0396665819,-0.1882720739,-0.2591831088,-0.1829653382,-0.0938074887,-0.0120979194,-0.1587508172,0.3571461737,-0.0653849542,-0.2649568617,-0.2993311584,0.0447383635,0.295560658,0.1373045295,-0.4245435894,0.0067302636,0.5594311953,-0.0806818008,-0.0790805444,0.2808078527,0.5457740426,0.1270319521,-0.0827852041,0.1383132339,0.1929411739,-0.0637374967,-0.1260762811,0.214398101,-0.0173141062,0.1705465317,0.223416239,0.1514259875,-0.2110050023,-0.2321826369,0.4559077621,-0.1453309059,-0.3635572493,0.192087546,-0.2378398627,-0.0323568098,-0.0969727114,0.3651038408,-0.2949544787,-0.1943055987,0.1872397214,0.2347214222,0.140871346,-0.4275604486,0.1380085349,-0.1543761939,0.1261260808,0.2467401326,-0.1832294464,-0.3851372898,-0.015749732,-0.7335856557,0.2009659857,-0.150406599,-0.070766449,0.0706763044,0.2328766584,-0.0135105625,0.0742938668,0.2322016209,0.2283024788,-0.0936845392,0.4965483248,-0.4009833038,-0.2250744253,0.1137502342,-0.1236983314,0.0905506536,-0.431661576,0.2643917501,-0.0150298644,0.0841797739,-0.055553969,0.0026559797,0.329531461,-0.2390793115,0.2479810417,0.301571548,0.4559423029,-0.1107432842,-0.2801846862,0.0129051777,-0.0368635356,-0.041126512,0.28016904,0.3217188418,0.2556381524,0.1058465466,-0.2885212898,-0.2245197296,0.3387840688,0.2152883261,0.0428710915,-0.3165065348,0.4041502774,-0.0445990339,0.0526788533,0.1392771304,0.3226180077,-0.0196398906,0.2745648026,-0.3327932656,-0.6344330311,0.3417711258,-0.322576791,-0.179686144,-0.274009347,0.3677443564,-0.0831620172,-0.1276747286,-0.2130842358,0.1397543401,0.429546237,-0.2491013557,0.0651856884,0.3638628423,-0.0633863732,0.1249608472,0.2045358419,0.1447397172,0.1241833046,-0.2364504784,0.5228363872,-0.1628024578]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2538","title":"Loading partial dataset when debugging","comments":"Hi ! `load_dataset` downloads the full dataset once and caches it, so that subsequent calls to `load_dataset` just reloads the dataset from your disk.\r\nThen when you specify a `split` in `load_dataset`, it will just load the requested split from the disk. If your specified split is a sliced split (e.g. `\"train[:10]\"`), then it will load the 10 first rows of the train split that you have on disk.\r\n\r\nTherefore, as long as you don't delete your cache, all your calls to `load_dataset` will be very fast. Except the first call that downloads the dataset of course ^^","body":"I am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!","comment_length":98,"text":"Loading partial dataset when debugging\nI am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!\nHi ! `load_dataset` downloads the full dataset once and caches it, so that subsequent calls to `load_dataset` just reloads the dataset from your disk.\r\nThen when you specify a `split` in `load_dataset`, it will just load the requested split from the disk. If your specified split is a sliced split (e.g. `\"train[:10]\"`), then it will load the 10 first rows of the train split that you have on disk.\r\n\r\nTherefore, as long as you don't delete your cache, all your calls to `load_dataset` will be very fast. Except the first call that downloads the dataset of course ^^","embeddings":[-0.3671891391,-0.1431681812,-0.0216082856,0.3622045815,-0.067530632,0.2756697237,0.5740375519,0.4168047011,0.2612759769,-0.0835103467,-0.082844846,0.1403204501,-0.1739207208,0.1820890456,0.0995866209,0.1087071076,-0.1426474005,0.3165223598,0.0734355226,-0.0784858689,0.1707005054,-0.26123932,-0.1748515218,-0.1787908077,-0.4388399422,0.0658321753,0.0724343061,0.1917069405,0.2921259999,-0.1966113895,0.3809176385,0.1237108037,0.0042912429,0.1477989405,-0.0001250374,0.0485777669,0.4530166984,-0.1177272201,-0.4413953125,0.0793153346,-0.237233296,-0.0172563251,0.4449100196,-0.0426382981,-0.0530504771,-0.187539041,0.0052164318,-0.3129993975,0.2829741836,-0.0120449858,0.0667552426,0.0641538724,-0.5506688356,0.1383746117,0.2196933031,-0.0400902294,0.0063698953,0.3839592934,0.4114600122,0.0257234722,-0.3189215958,0.143329367,-0.1303001493,0.2568044364,0.3651657701,-0.0175332408,0.016991362,-0.4844467342,0.277962625,0.3971960843,0.6139980555,0.0177978165,-0.324831605,-0.5615004301,0.0680839494,-0.1019032151,0.1404101104,-0.0329516903,-0.4404362738,-0.0019115366,-0.503777504,-0.1788008213,0.0656843781,0.1815218478,0.0742575899,-0.2963497937,0.0635769144,0.1199378818,0.4592187107,0.2257744819,0.3110461533,-0.0591473207,0.0867825672,0.1040572375,-0.3470100164,0.0552104115,-0.1715562642,-0.0937866569,-0.0360143371,0.5328012109,-0.0335619822,0.462310195,0.1792253256,0.0048949281,0.5818147659,-0.1044395044,0.1357749999,0.1604953557,0.3302040696,-0.3076428771,-0.0562027469,-0.0291394927,-0.4441228807,-0.1446470916,0.2911393344,-0.0331761204,-0.0609830953,-0.3301861286,-0.0984939113,-0.3457719386,0.1044795364,-0.027723629,0.2356433272,0.4508670866,0.0771400854,0.0577649698,-0.1905168742,0.0967730135,-0.2860884368,-0.0646092147,-0.1252984405,-0.179455772,-0.0533976667,0.1343159825,0.345648706,-0.3430443108,0.3036074042,-0.0918700993,0.4442437291,-0.1942007691,0.1545960009,-0.2542306483,-0.0301960651,0.1774717122,-0.0150245437,0.2808523476,0.1144268736,0.2564080954,-0.3054350317,0.4385035038,-0.114683114,-0.3874349892,0.0642130077,-0.0014648666,-0.3882314563,0.2442891449,-0.1981087327,0.2364187539,0.1910140961,-0.2338833213,-0.2300262004,0.0961127058,0.0001006437,-0.1139457971,0.6939179301,0.3801719248,-0.1435455978,-0.2851575911,-0.2555277944,0.0346329436,0.2755737603,0.2422008961,-0.4427700639,0.2324467003,-0.1369287521,-0.0750613958,0.5813103914,-0.0458763093,-0.3899812698,0.4592676461,-0.0660858899,0.2072698623,-0.0639574006,0.5403299332,0.2244986594,0.0629709959,0.1113066897,0.4323061109,-0.1434763521,0.1165145338,-0.2158845663,-0.1530449539,0.1313552409,0.4665034413,0.0372895077,0.1652179956,-0.0386094227,0.1996426433,0.1884550005,0.0003334567,-0.0666797683,0.1926486045,-0.1065419167,0.0639995784,0.0371609405,-0.2812865674,-0.5614965558,0.3774626255,0.1746428758,-0.3102475107,-0.0822079629,-0.0205233563,-0.101674065,-0.1576439291,-0.0180426165,-0.085049808,-0.1891314685,0.0147956302,-0.0793387517,-0.3969837427,-0.4567694068,0.1184231043,0.0288374275,-0.0162682217,-0.3463562131,0.1118151173,0.3637426794,-0.0564941429,-0.0430023111,-0.08845599,0.0026182537,-0.2663026452,-0.0660595968,0.3327568471,0.148797974,0.2904263139,0.045086097,0.1757802665,-0.0158164985,-0.0804262608,0.3634447455,-0.0561743379,0.0023397489,-0.1930642873,-0.19840464,0.3541494906,-0.4528671801,0.4073086679,0.0399317071,-0.2771592736,0.1374360621,-0.1565610468,-0.1314674616,-0.0885141939,0.4653721154,-0.2589767575,0.3783656657,0.2475851625,-0.0831934586,-0.1018943042,0.2595183253,-0.0359505638,0.004566073,0.2960415483,-0.46206671,-0.0278332178,-0.1357196867,-0.1868363023,0.0805441737,0.1411464661,0.1734262109,-0.2135947496,0.2454040945,-0.0822357908,0.1437816322,0.0757913962,0.105934225,0.1493422091,-0.1592341065,-0.1643154919,-0.1599853039,0.0717431307,0.158798337,0.1586150974,-0.2362281531,0.1022941917,-0.4182833135,-0.1048171669,0.0990508944,-0.0547475442,-0.0408968218,-0.122612536,0.0478475541,0.3270416558,0.0698409304,0.2051166296,-0.2745992839,0.1811829209,-0.1321081966,-0.3716071546,-0.2293088287,-0.3260007799,-0.0815040916,-0.0047446671,0.4853346646,0.1031744331,0.3148230612,-0.2068870813,-0.2574830949,0.1653002203,-0.0430504046,-0.0358235128,0.2538663745,0.2143922299,-0.1287669539,0.3431736529,0.1944276243,-0.362919867,0.0824215934,-0.1285372823,-0.0336526223,0.086115025,0.2274838239,0.3637343943,-0.0477724448,-0.134433195,-0.0298836417,-0.3539547622,0.3215943277,0.0293443948,0.1498514712,0.067358695,-0.1317503154,-0.1673803478,-0.0274369903,0.1491861194,0.0864903107,-0.3391231894,0.4030130506,0.1747459769,-0.0408624448,-0.0752274394,-0.0718040094,-0.1546743363,0.2917643487,-0.6640383601,-0.0668251142,-0.128480494,-0.0040828804,-0.1210261285,-0.2857667804,-0.002188307,-0.1174402833,0.0631621554,0.1598545462,-0.1683164537,-0.1388295293,-0.2689239085,0.0819728673,-0.0497700535,0.2759536803,-0.0506890528,0.8458833694,0.041929394,-0.1853019148,-0.1689160317,-0.1192442477,0.0638334975,-0.3244429231,-0.4014126956,-0.0687760115,-0.1970940828,-0.263364315,0.1077177972,-0.0919820219,-0.3551757336,-0.1895462126,0.0282494649,-0.2745781541,-0.3581511676,0.4824312925,-0.5855872631,-0.140605256,-0.0282719806,0.2764353752,-0.0449182056,-0.2472790927,0.0181325395,0.0796073601,0.4708025753,-0.2335610688,-0.0139494129,0.1771229357,-0.1779390275,0.1654905379,-0.1226425096,0.1473449618,-0.02570007,-0.1633833349,0.2757573426,0.1785820723,0.7884396911,0.0170748662,0.1327192634,0.0002017408,-0.5102229118,-0.2963167131,0.0534745976,0.1002969965,0.5788521767,0.1988783032,0.3522874713,-0.3964131773,-0.0815310553,0.1228852272,0.3506296277,-0.2957377732,-0.1107417941,-0.3007184267,0.0792071819,-0.3384441137,0.0602166057,-0.0772344545,-0.1689599752,0.1111793071,-0.1326546967,-0.2855384052,0.0597577319,0.1412393898,0.2928125262,0.2557173371,-0.051807642,0.5995861888,0.2853755653,-0.0873787031,0.2835316956,0.4308952987,0.0039379736,-0.2233617753,0.2127960175,-0.0280541833,0.1954458654,0.4685197473,0.0028096132,-0.1305248886,0.2168253064,0.1986552775,-0.3661758304,0.1299190372,0.0626510009,-0.0761784762,-0.3292380869,-0.5705028772,0.4089976549,-0.0779562369,0.0118412022,0.5679505467,-0.2431020439,-0.0053953566,0.1903681606,0.0203534476,1.0178778172,-0.3330104947,0.0218968485,-0.4650077522,0.0914332718,0.4882529974,-0.2961688936,0.0646971613,-0.3654837906,-0.6286565065,0.078601189,-0.2877317369,-0.2745550871,0.2768291831,0.2540509999,0.2377897203,-0.4282457829,0.0898493528,-0.0995339453,0.2724858522,0.0906432047,-0.3121676147,-0.0334696323,0.0022576079,-0.0345318504,0.1255575716,0.0643879399,0.0665349066,-0.0685932413,0.1550923437,-0.392731905,0.1411187798,-0.1374561042,-0.0428908579,-0.2702481449,-0.3985818624,-0.0254866704,0.1370051354,-0.0683277622,0.0879688188,-0.0561874099,0.3046511412,0.0296229403,0.1737749875,0.0210134387,0.0706022233,-0.0502466857,-0.0093013281,-0.0992267728,-0.1337225288,-0.2553774714,-0.4720454812,0.323376894,0.239531517,0.1415905654,-0.0632365122,-0.2486247122,-0.0001776128,0.1796437055,-0.1793679595,0.0291047823,0.4345216751,-0.0641019046,0.1792690158,0.2324335128,-0.3975849152,0.0245463792,0.3809718192,0.1816916168,-0.0583332106,0.4532392919,-0.0917834863,-0.3143451512,-0.0893868133,0.0159125533,0.329349488,-0.5196863413,-0.0000233216,-0.1610176563,0.0594340339,-0.1526347697,-0.2847933769,-0.0055978498,0.0689764023,-0.3740065694,-0.3733099699,-0.1986030191,0.3582197726,0.025333168,0.3640442491,-0.2689649463,-0.1113041267,-0.1866380274,0.5100815892,-0.1515256912,-0.1166600063,-0.1506282985,0.3793258071,0.0103443367,0.0230710935,0.356502533,-0.1719985157,-0.0177373886,0.0263421852,0.1387483478,-0.0877264738,0.0856961012,0.223717764,-0.2166979462,-0.2318814695,-0.1974543631,-0.128396526,-0.0637037307,-0.2944410443,0.220137611,0.1698251963,0.1967494786,-0.106033884,0.4144618511,0.0202572551,-0.4970488846,0.1724216044,-0.2905471325,0.3164823055,-0.0908004865,0.204643026,0.0721128583,-0.0054178783,-0.0110092489,-0.0986153409,0.365193665,-0.1053293869,0.1401947141,0.0574864149,0.2886664867,0.0617079102,0.3608944416,0.7714855671,0.0302278902,0.2253796905,0.0788227469,0.0353274792,-0.0415503941,-0.1432342529,0.0548329093,-0.3868787289,-0.0227043554,0.2216560245,-0.2289808542,0.2641266584,-0.0151986843,-0.2538966835,0.2530401945,-0.172375828,0.0172838662,0.3965563476,-0.2170441747,-0.1060305685,0.0804063827,0.2679737508,0.0435836315,0.3496535718,-0.1336517036,0.2164198011,0.1409597248,0.1506497115,-0.2677202225,-0.3968942761,0.007602626,0.5739414692,0.1509142965,0.0038488649,-0.1768155992,0.0664811656,0.1489868313,-0.2371134758,-0.0311329719,0.0436324552,0.0971298665,-0.0746427625,0.2067611217,-0.1738481671,-0.2689622939,0.117878899,-0.1730868667,-0.3819600642,-0.0769189522,0.2927039862,0.1724595726,-0.1080509201,0.1192582995,0.0203601345,0.2263833284,-0.363355726,0.1058879867,-0.1151911318,0.2659334838,0.4520157576,-0.0552790202,0.0113519272,0.1452592015,-0.1387480497,0.0534792095,0.3016168475,-0.0364706442,-0.278196156,0.3117523193,0.151029706,0.3166750371,0.0354430377,0.023774283,-0.0851155594,0.1745290309,-0.2254258394,-0.0548546575,-0.5020073056,0.3572721481,0.0979873687,-0.2567735314,0.0218000859,-0.1460699737,0.1641468108,-0.5156952143,0.2081427723,-0.2873867452,-0.0298748165,0.3652812541,0.0081697619,-0.0818294063,0.6307969093,0.0844632909,0.5784643888,-0.2518799901,0.0471645407,-0.4567259252,0.0864483416,-0.0638266951,0.0068628076,0.426787883,0.2150807828,0.3521044254,0.376336664,0.0156499352,-0.0039929007,0.3473134935,0.3020207584,-0.0365598872,-0.0076758917,-0.5346394777,0.2176374644,-0.001232447,-0.3525676727,0.1920086443,-0.0523051359,-0.1589458287,0.4528041184,-0.4606049657,0.4527320266,0.3637913764,0.4324217439,-0.20643574,0.6541488171,0.0310749486,-0.1299526393,-0.0698156431,-0.5255054235,-0.2674126923,0.3101967573,0.1402409971,-0.1344031841,-0.0694592893,0.0419761278,-0.3419904709,0.1872281581,-0.1282745898,0.031474404,-0.0409942605,-0.4278574288,0.1215899885,0.2747301161,0.2673530579,0.1640353948,-0.1089977324,-0.1843394041,-0.176342696,0.0502826124,0.5359644294,-0.1613197029,-0.0261933096,-0.0143750394,0.3824871778,0.1976982653,0.2183241248,-0.5468689799,-0.0375994779,0.321560204,-0.2120874971,-0.0172293,-0.0083579281,0.041081924,-0.0789170638,-0.2472642809,0.1625506282,0.0581363477,-0.0685482472,-0.3469407856,-0.3269047141]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2538","title":"Loading partial dataset when debugging","comments":"That\u2019s a use case for the new streaming feature, no?","body":"I am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!","comment_length":10,"text":"Loading partial dataset when debugging\nI am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!\nThat\u2019s a use case for the new streaming feature, no?","embeddings":[-0.4618416727,-0.102580905,-0.025752956,0.2337713093,-0.051758606,0.2635930181,0.5277156234,0.4005592167,0.2569629252,-0.1201939359,-0.1030896306,0.1262804568,-0.1891560256,0.319886893,0.1285772324,0.1116196364,-0.1549823433,0.3901554942,0.0464991555,-0.0444915928,0.1791653633,-0.2458690554,-0.2178136259,-0.2557118535,-0.3894646764,0.1399041861,0.1941753328,0.1080066487,0.2481357306,-0.2124390453,0.279222846,0.1428738981,0.0199283324,0.0886864662,-0.0001234191,0.0469665714,0.5393361449,-0.1561617851,-0.3813607097,0.0623483472,-0.2308821976,0.0414909385,0.4316727817,0.0639943704,-0.1057052985,-0.2190790623,0.059023954,-0.3659281731,0.2402294278,0.0179461911,0.0680566952,0.0922229514,-0.5196190476,0.1996423602,0.1699413061,-0.0672104433,-0.0379431657,0.3697941601,0.4956247509,0.0282911509,-0.3447718024,0.1807883382,-0.1016988903,0.2968763411,0.4516940415,-0.03757561,-0.0814368427,-0.5123176575,0.2233945429,0.4641260505,0.6338376403,0.0543267913,-0.3536936641,-0.5326236486,0.0780728161,-0.1574284285,0.0852639154,0.0369668081,-0.4755298793,-0.0293945428,-0.5009676814,-0.2024194002,0.0072298413,0.1696307808,0.076727964,-0.1433760524,0.0919718742,0.0884490013,0.3524726331,0.2323951423,0.3257061541,0.0005199163,0.0613918342,0.0630152524,-0.3638934493,0.0484213196,-0.1183654442,-0.1185642928,-0.0526662841,0.5092153549,0.0580150262,0.4422687292,0.1848828197,-0.0186963975,0.5542943478,-0.166710943,0.1844276786,0.1117468849,0.4024398327,-0.2712471187,-0.075239487,-0.0603839196,-0.3251871765,-0.0256637055,0.211069718,-0.0904691666,0.0802553594,-0.3290146291,-0.1621712595,-0.3206707835,0.1117741317,-0.0572155453,0.2240299135,0.4008879364,0.0827954412,0.1118234992,-0.2240075618,0.1618285179,-0.2316546738,-0.1088145897,-0.1044897437,-0.2317078412,0.0212524366,0.142583549,0.3564444482,-0.4767852724,0.3878217638,-0.1212813333,0.496145457,-0.1656162441,0.1180372909,-0.2463481724,-0.0080069909,0.1903519928,0.0457971469,0.2110758424,0.0502380319,0.2300215065,-0.2483243048,0.4224554002,0.0160858203,-0.3958751857,0.06764061,0.0049810465,-0.3173011243,0.1975702643,-0.2565830052,0.3656565249,0.0438528396,-0.2786504328,-0.2825683653,0.1230687797,-0.0718150362,-0.0774129778,0.6567648053,0.3396761417,-0.1159662828,-0.2745162547,-0.1582387537,0.0309632216,0.2606007457,0.1784227043,-0.3953612447,0.1429364085,-0.0829807296,-0.0466924347,0.5551092029,-0.0680239499,-0.3728124201,0.5051888227,-0.0970688909,0.2050602734,0.0079934448,0.55975914,0.2968554497,0.0639114082,0.0781977698,0.5011104345,-0.1713937819,0.1517863721,-0.1580218077,-0.1192599162,0.1811213642,0.4442256391,0.0760874674,0.2311649919,-0.0294977631,0.2078089118,0.192607671,-0.0241392367,-0.0678892136,0.1311438084,-0.0846363977,0.012138837,0.017012896,-0.3080239892,-0.5797471404,0.3251579404,0.2776846588,-0.2878667116,-0.001347727,-0.055020012,-0.0602279156,-0.084368743,0.0366121382,-0.1187169999,-0.1669366062,0.0506960303,-0.1060582995,-0.3865200579,-0.5519036055,0.1676009446,-0.0392771438,0.006214391,-0.3382095993,0.155194968,0.4177765548,-0.0627374649,0.0343081951,-0.1699812412,-0.0197968297,-0.1868831366,-0.0742487684,0.2685055435,0.0870884433,0.3352881372,0.0682213083,0.152760312,0.0482203923,-0.1519392282,0.3698000908,0.0358053036,0.0251946244,-0.1936995834,-0.1829127222,0.3175721169,-0.4332684577,0.404866606,0.0997565314,-0.3253595233,0.1493109465,-0.0668246597,-0.1783299893,-0.0890224651,0.4542120397,-0.2830055356,0.2615095973,0.1651979685,-0.1922606975,-0.0481080376,0.1852844656,0.0025163377,-0.0768503025,0.2905119359,-0.4832884073,-0.0584181473,-0.0754809901,-0.2126501799,0.1257589757,0.1621417701,0.1335122436,-0.1928811818,0.2892286777,-0.1340759695,0.1899056733,0.061240159,-0.0098128887,0.2302778065,-0.0694152191,-0.1790932864,-0.2063154131,0.1162926108,0.1392467171,0.0958648697,-0.2468064874,0.0193928406,-0.4034201503,-0.1673395485,0.1219615638,-0.0532147922,0.0399921723,-0.1152456254,0.0677935258,0.4481601417,0.0882637575,0.1591424644,-0.2369639128,0.1818353832,-0.1231536418,-0.3379866779,-0.2372072488,-0.2695005536,0.0348652676,0.0286593027,0.4643958211,0.1250773966,0.3015447855,-0.1532873213,-0.2289263308,0.0365691483,-0.0827354789,-0.0704389736,0.2252677679,0.2647013664,-0.1556956917,0.3476936221,0.2554940879,-0.3633581698,0.0627185255,-0.187729463,-0.0039560171,0.067152299,0.1994595379,0.3382553756,-0.0015099525,-0.1535556316,-0.0398266688,-0.3905102909,0.3798890412,0.0587484464,0.0980590954,-0.0049225371,-0.1593741924,-0.2267712504,-0.0080364309,0.103411749,0.1243548989,-0.2944485247,0.3739866316,0.1305330545,-0.0782380626,-0.0427151397,0.0202005487,-0.169980824,0.3058058321,-0.6667987704,-0.0614984185,-0.0837989524,0.0254719369,-0.0999502167,-0.3474080265,-0.0521479808,-0.1554697156,0.0788622424,0.151866734,-0.1528112441,-0.210093379,-0.2367157638,-0.0236662719,0.0244893413,0.3652833998,-0.1151266992,0.8401448131,0.0272514857,-0.129733786,-0.0788631439,-0.1857493669,0.0985026807,-0.427608341,-0.3038358092,0.0089387503,-0.1566064507,-0.2213377655,0.1260207444,-0.0645420179,-0.3878088892,-0.2043304294,0.0167914461,-0.2512308955,-0.2771837115,0.460624516,-0.5683704019,-0.1918531358,0.0014471741,0.2608177364,0.0667298064,-0.3761633933,-0.0430654883,0.0534788184,0.5056428313,-0.2046144605,0.1284352988,0.1829364747,-0.2712808549,0.1388496757,-0.1109561101,0.192074731,-0.0614328012,-0.2084713876,0.2566938698,0.1572646499,0.5863550305,-0.0249303468,0.138621062,-0.0354547501,-0.5109820366,-0.269021064,0.0615428425,0.127234444,0.5593417883,0.0993041843,0.3203271627,-0.309086442,-0.0501509346,0.1296133846,0.3176483214,-0.2428078353,-0.0475824401,-0.3079327047,0.1203906387,-0.3814363182,0.0352075025,-0.0082781473,-0.2453715354,0.0280830096,-0.1946384907,-0.2410963476,0.1045727655,0.1399040967,0.3244496584,0.2134123147,-0.0445407107,0.649538219,0.2398111224,-0.1037493572,0.2434858978,0.4407175183,0.1274421215,-0.3001966476,0.2127260715,-0.0035877924,0.2028710246,0.4694598615,-0.0552615896,-0.1039111093,0.228346169,0.1858546585,-0.3672137856,0.1412596405,0.0798806995,-0.0374710746,-0.2828614712,-0.5809779763,0.3910367489,-0.155269444,0.0656153634,0.6884926558,-0.255843401,0.0068565067,0.1231166869,-0.0307950322,0.8409698009,-0.2465376854,-0.0075965938,-0.4377614856,-0.0216929466,0.3462761045,-0.3172473907,0.0921152309,-0.3574266732,-0.670846343,0.0362057611,-0.2611832917,-0.2723329663,0.2379785329,0.2063838243,0.2402263582,-0.3561227918,0.2270268649,-0.0953851491,0.3263635337,-0.031643752,-0.316229403,-0.0760564059,0.0035794876,-0.0746150091,0.0640340373,0.068114683,0.0846310332,0.0191127453,0.1187155098,-0.4208443761,0.1199488118,-0.1674191207,-0.1319732517,-0.3016619682,-0.406619519,-0.0363399871,0.1566478163,-0.2292238921,0.0272928849,-0.0882437006,0.3108077645,0.0714387298,0.2470597178,-0.0051356102,0.0475204363,-0.0913679898,-0.067892164,-0.1081134453,-0.1085763946,-0.2539457679,-0.4697411954,0.2954316735,0.2288636267,0.1547358036,-0.0121515868,-0.2570487261,-0.0190434642,0.157438606,-0.2076603323,0.0335943475,0.4046293497,-0.1016411856,0.095451124,0.3214437366,-0.3873172402,-0.0297238771,0.4137916267,0.1872593313,0.029655803,0.4463548064,-0.1353328526,-0.332863152,-0.0705278814,0.0687077194,0.3111311793,-0.4531735182,-0.0129998084,-0.1198500544,0.0505841114,-0.1199148446,-0.2602984607,-0.138342768,0.1077019572,-0.3955348134,-0.405801326,-0.1800206602,0.3185860217,-0.0367625728,0.2865371108,-0.2760222256,-0.1548084468,-0.1226850227,0.489331305,-0.1480391771,-0.1418845803,-0.0477601923,0.3679990768,0.0376001373,0.0915758312,0.350163132,-0.1405949146,-0.0234594382,0.0316378213,0.1397195011,-0.0906178504,0.0939943939,0.2169879973,-0.1134838313,-0.2207477838,-0.2464702576,-0.1296947002,-0.1638120562,-0.2431360781,0.2836966813,0.1326399148,0.1534757316,-0.1365827769,0.4886804223,-0.0744159669,-0.4228806198,0.2052877396,-0.2833788991,0.3555661142,-0.0974548459,0.195906207,0.1238417327,0.0052460609,0.0051599024,-0.0968445465,0.3412667215,-0.0833126381,0.1674262136,0.0595712885,0.284758389,0.0132993972,0.3948419392,0.7892304659,0.0376321301,0.2534658015,0.033081688,0.0429722145,-0.0701533332,-0.1387659609,0.0218762811,-0.3851896524,0.0295835696,0.1926603168,-0.169814527,0.2166269124,-0.0374457389,-0.280795306,0.1871645153,-0.1849461049,0.0611010492,0.3270016611,-0.2020641714,-0.0647283271,-0.0070895292,0.3134896457,0.0156018669,0.3131305277,-0.1360258311,0.1691908389,0.1632255167,0.1291279495,-0.2764316201,-0.3887904882,-0.0896227062,0.4974320233,0.0948132277,0.0755414367,-0.0708939806,-0.0212442037,0.1872078776,-0.2191982567,-0.0239305981,0.0488882549,0.0979629233,-0.1189466491,0.2429552227,-0.1304811537,-0.2127443105,0.233399272,-0.1534046233,-0.3120705485,-0.1717305481,0.3220469058,0.1697544754,-0.1042305306,0.1932674348,-0.0658255816,0.238501817,-0.3144374788,0.1395092607,-0.0900015682,0.3653866649,0.543977499,-0.0090439189,-0.0112545127,0.1114167273,-0.0845208913,0.1455014944,0.286768645,-0.0343525447,-0.2405533195,0.3433546424,0.2526391149,0.365473479,0.0378909148,0.0270354711,-0.062659204,0.2577911317,-0.1897717565,-0.0977112576,-0.4667253494,0.2651487887,0.134999916,-0.2426283211,-0.0190086346,-0.1830703765,0.1519307196,-0.4795233309,0.2447836101,-0.3326300085,-0.0164611116,0.3441998661,0.0146086318,-0.0277434438,0.5848677754,-0.0039543235,0.5623373389,-0.1745825857,0.0236414466,-0.4510498941,0.094445385,-0.0103296041,0.0339828432,0.4037602544,0.20571886,0.3718590736,0.4044769406,-0.0174823776,-0.0630356595,0.2518298924,0.3131319582,-0.0036647953,-0.0759909227,-0.5298848152,0.262067616,-0.0881872848,-0.27548334,0.2095134258,-0.0268222503,-0.126799196,0.4265170097,-0.4036890566,0.3812583685,0.2878258526,0.3506270349,-0.261703819,0.6447544694,-0.0358278826,-0.1158271953,-0.0429695323,-0.5366233587,-0.3209978938,0.2687728703,0.1165952384,-0.0770887434,0.0012191868,-0.0038539844,-0.2868612111,0.2027552426,-0.1711700708,-0.0629348457,-0.0997593552,-0.3930239081,0.0817360952,0.3122024536,0.213815257,0.1499870718,-0.1524044573,-0.3016652167,-0.1575541049,-0.0113379965,0.5734301209,-0.1860197037,0.0119749084,0.0007315184,0.2990679145,0.2517391741,0.295163393,-0.6067880392,0.0143569,0.2771971226,-0.284145385,0.0734170377,0.0344816521,0.0614946038,-0.0459574014,-0.2838643193,0.0834596157,0.1185156181,-0.125688076,-0.4011085629,-0.3403424621]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2538","title":"Loading partial dataset when debugging","comments":"Hi @reachtarunhere.\r\n\r\nBesides the above insights provided by @lhoestq and @thomwolf, there is also a Dataset feature in progress (I plan to finish it this week): #2249, which will allow you, when calling `load_dataset`, to pass the option to download\/preprocess\/cache only some specific split(s), which will definitely speed up your workflow.\r\n\r\nIf this feature is interesting for you, I can ping you once it will be merged into the master branch.","body":"I am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!","comment_length":71,"text":"Loading partial dataset when debugging\nI am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!\nHi @reachtarunhere.\r\n\r\nBesides the above insights provided by @lhoestq and @thomwolf, there is also a Dataset feature in progress (I plan to finish it this week): #2249, which will allow you, when calling `load_dataset`, to pass the option to download\/preprocess\/cache only some specific split(s), which will definitely speed up your workflow.\r\n\r\nIf this feature is interesting for you, I can ping you once it will be merged into the master branch.","embeddings":[-0.4255480468,-0.1098934039,-0.0348530896,0.3032792509,-0.0704117268,0.282392472,0.5759145617,0.4311378002,0.3011431992,-0.0897593573,-0.0946438685,0.1700280756,-0.1675223559,0.2383288145,0.0475689396,0.0998504236,-0.1617254913,0.3242848217,0.0387713164,-0.0660673827,0.1287385225,-0.2229061127,-0.248351872,-0.2115873992,-0.4659939408,0.0293295104,0.0986263454,0.1705517471,0.3179525137,-0.1268362105,0.3389091194,0.137517035,-0.0084876288,0.1335718185,-0.0001241957,-0.0107220188,0.5006647706,-0.0931077823,-0.4362995028,0.0735286698,-0.237421155,-0.0477150455,0.4577589929,0.0274892692,-0.083421059,-0.1815837771,0.0253198538,-0.3546507955,0.2971426845,-0.0064251428,0.0737070292,0.1067952514,-0.4661458731,0.1442133933,0.1794302315,-0.050980635,-0.0700000003,0.3940604627,0.4568606913,-0.0344275609,-0.2619629502,0.1470095068,-0.0907033384,0.3317944407,0.3868725598,-0.0311696939,0.0120002786,-0.5471946001,0.2819919586,0.4399408996,0.6287299395,0.0329469703,-0.3415046036,-0.5828240514,0.0757654309,-0.0979445726,0.1490002275,0.0070247171,-0.4245784581,0.0065797223,-0.4864515066,-0.1778587997,0.0478371009,0.1568618119,0.0531837158,-0.2893976867,0.036896497,0.0943091363,0.3902567327,0.2198377252,0.3261693418,-0.057307262,0.0780214816,0.0870090798,-0.3868720233,0.0628545433,-0.1920491308,-0.0946867764,-0.0610498674,0.5311383009,-0.0742087141,0.4788445234,0.2084883004,-0.0358149149,0.5376525521,-0.150255248,0.1837615222,0.1738742143,0.3688533306,-0.3586819172,-0.0799363852,-0.0411484465,-0.3891741633,-0.1940121651,0.2670511305,-0.0223625023,-0.0176770724,-0.3543414772,-0.1577949077,-0.367539376,0.1221827939,-0.0502620265,0.2032260895,0.427698493,0.0881245658,0.0653591603,-0.1983948499,0.1510773003,-0.2707566917,-0.1073966995,-0.0971074551,-0.2272587568,-0.0123133911,0.1486579478,0.3650503159,-0.3609235585,0.2793487906,-0.0693353638,0.4252460003,-0.2300672531,0.1399095654,-0.266392529,-0.025323879,0.192095831,0.0380922072,0.2751621008,0.0690244585,0.2701106071,-0.3026752472,0.3971216977,-0.0496055931,-0.3965829909,0.0440241247,-0.0021049343,-0.3572974503,0.2418508828,-0.1939354539,0.2439127415,0.156740576,-0.149572894,-0.2970223129,0.0720491782,-0.0504864827,-0.0922596157,0.6279121041,0.3850626349,-0.1324379742,-0.1911819428,-0.2131745666,0.0428278968,0.2830062509,0.211183399,-0.3966250122,0.2519930005,-0.1373879462,-0.1186437383,0.5024479032,-0.0747006014,-0.3204850852,0.5063077211,-0.0870788023,0.180222556,-0.0622124895,0.4636337459,0.2160332948,0.04201442,0.0510071367,0.4863911867,-0.1479168683,0.1553367078,-0.2005395889,-0.1223475486,0.1038940474,0.4797152281,0.0456256084,0.1825736314,-0.050061103,0.1873490959,0.178916961,-0.0209913347,-0.07795652,0.1962694228,-0.117625609,0.0428444222,0.0597052202,-0.2848039865,-0.6154308319,0.3532468975,0.1503711194,-0.3225485086,-0.0631126836,-0.0078727929,-0.0616652928,-0.1172562689,-0.0140972454,-0.0845228136,-0.1754078418,0.0163250994,-0.0633150935,-0.4030498862,-0.4566090107,0.1725052595,0.0014144225,0.0076232208,-0.3340208232,0.114945747,0.3976435065,-0.0506278835,-0.0025283408,-0.0804589242,-0.0058342782,-0.2696807384,-0.0657005459,0.3011422455,0.1393207759,0.2979264855,0.1136752144,0.1450328827,-0.0313481092,-0.0681549385,0.3927516043,-0.0917418599,-0.0070267702,-0.2175112814,-0.1865833402,0.3615932167,-0.4406475723,0.4426867068,0.0334503911,-0.2793010473,0.1115421802,-0.1444292963,-0.1060400158,-0.0906952992,0.4702703059,-0.2627221346,0.3744698167,0.2515541315,-0.1180124953,-0.0431040414,0.2224398404,-0.0110001676,-0.0264835767,0.3000358641,-0.4567562938,-0.0279169902,-0.1657408774,-0.1882313937,0.1148747429,0.1626736075,0.1160588637,-0.2489050776,0.2345138341,-0.1317109764,0.1706822962,0.1135738492,0.1078627408,0.1801273376,-0.1570752263,-0.1743776351,-0.1415260434,0.1043536142,0.1491667479,0.1552515477,-0.2420474142,0.0409890823,-0.4336850047,-0.095436275,0.1087247506,0.0012945594,0.0000097473,-0.1684014052,0.0455468781,0.3233199418,0.0870681554,0.1428837925,-0.2267100811,0.1525973529,-0.1484452486,-0.3771046698,-0.2346262038,-0.274638176,-0.0331601575,0.0033810793,0.5305057764,0.1417619139,0.332596004,-0.2133327723,-0.2271145135,0.0889055207,-0.0519532859,-0.0518517308,0.2499420792,0.2526989579,-0.126336351,0.3071929514,0.2569523156,-0.3159406781,0.1098044291,-0.1088883951,0.0034073782,0.067696996,0.160555169,0.3011202514,0.0053965701,-0.1698033214,-0.0307083726,-0.3686235547,0.3665091991,0.0164294913,0.1151424572,0.0609630235,-0.1534951329,-0.2031879276,-0.055141151,0.1240792051,0.1300558448,-0.248816967,0.3999468684,0.1499376148,-0.0674016178,-0.0626190975,-0.0296015553,-0.1598857343,0.2536583245,-0.6322861314,-0.0330953076,-0.1697777808,0.0591506548,-0.0926251635,-0.3379983902,-0.0306049101,-0.1077454016,0.0659354776,0.1735351384,-0.1705049574,-0.1234268397,-0.3009801805,0.0332172327,0.0075061289,0.2793803811,-0.1313439757,0.7960772514,0.0527015403,-0.1776231378,-0.1590410769,-0.1315187067,0.0526378602,-0.3936482072,-0.3983528018,-0.0121327424,-0.1709493697,-0.1929161549,0.1234552786,-0.0569065735,-0.360707283,-0.1844074726,-0.024530001,-0.2462880313,-0.3422317207,0.4858388901,-0.533136189,-0.1508584768,-0.0002184699,0.2596715689,-0.0153870108,-0.2984855771,-0.0077806488,0.0723130554,0.5126267076,-0.2478263825,0.095011957,0.2320862412,-0.2354313433,0.1943082213,-0.1584600657,0.1555484682,-0.0471591502,-0.1470233351,0.3114203811,0.1976337135,0.6516824365,0.0131331189,0.1269878894,-0.0256436411,-0.5850504041,-0.2765258253,0.0595206879,0.0903104022,0.5965149999,0.1576708704,0.3528314531,-0.3361721635,-0.1693975627,0.158276543,0.3369893432,-0.2529734075,-0.154201746,-0.2478730828,0.1085687727,-0.3825848401,0.0695344433,-0.0136522846,-0.2019247711,0.0756776035,-0.1720867306,-0.304081589,0.0818557441,0.140473336,0.2908128798,0.2106689066,-0.1111382023,0.6667954922,0.3049616218,-0.105752185,0.2823256552,0.3889487982,-0.0066263895,-0.2671971917,0.1825739145,-0.1143232435,0.2009791583,0.4967256188,-0.0328111835,-0.180592671,0.1169169173,0.2165847272,-0.3674063981,0.0425879769,0.1022801474,0.0049077636,-0.3011725247,-0.5198283195,0.3954170644,-0.0661352128,0.0396847911,0.6187756658,-0.2771736979,-0.0084437495,0.1720659137,0.021292014,0.9266536236,-0.3237703741,-0.0267254375,-0.4711109698,0.0420479737,0.3918367326,-0.2913408577,0.032728456,-0.3474789262,-0.6259019375,0.0939166024,-0.2549416125,-0.2424487919,0.3078632951,0.269395709,0.2444548011,-0.4067323208,0.0754269958,-0.1080506071,0.3130561709,0.0523109473,-0.3572487533,-0.0330365226,0.0150458254,-0.0155570945,0.1669785529,0.076095663,0.0463344716,-0.043139305,0.1699290872,-0.4182583392,0.1273210645,-0.1174611524,-0.1342811435,-0.2234257311,-0.3589178324,0.0806556866,0.16421251,-0.0663387328,0.1277460456,-0.0771031454,0.3243989646,0.0203707851,0.1677104384,0.0537239723,0.0404102057,-0.0689282119,-0.0560780242,-0.0829165131,-0.2383133322,-0.2540110648,-0.4231798649,0.2556218505,0.25390172,0.1100005731,-0.0319813304,-0.2888383865,-0.0155119542,0.1655260026,-0.2077215016,0.0495569967,0.4450703859,-0.0290836543,0.207500264,0.2286260426,-0.4448326826,-0.019374311,0.3128205538,0.1547842324,-0.0245693903,0.4885365069,-0.0238873139,-0.3482508659,-0.0968003497,0.0324060507,0.347817719,-0.5323247313,0.0432259329,-0.0954325274,0.0717842802,-0.0855788812,-0.2574074864,-0.0827288628,0.0252300333,-0.3723294139,-0.3997138441,-0.1243819445,0.2555161119,-0.0016862547,0.3168868124,-0.2882219255,-0.1357832104,-0.1156779826,0.4980579019,-0.1506539583,-0.1228764355,-0.118801631,0.3464044034,0.0529485606,0.0250471886,0.3450154066,-0.1282456517,-0.0013390569,0.0266418401,0.1603636891,-0.1047791466,0.1074304357,0.2032574564,-0.2422377914,-0.232937023,-0.1401116997,-0.135632351,-0.1378342509,-0.3078646064,0.2151729465,0.1667672545,0.1878569722,-0.1424104869,0.4268201888,-0.031564597,-0.4116509855,0.2446386367,-0.2885431349,0.3497884572,-0.1457008868,0.2330198139,0.1168085337,0.0130677279,0.0273902901,-0.1620254219,0.3498270214,-0.0905718282,0.1439084113,0.0498296544,0.2511837184,0.0712215826,0.3635662496,0.7526994944,0.0421289951,0.2309818119,0.0358690247,0.0266573727,-0.0456357226,-0.0797327608,0.0889467075,-0.3545721769,-0.0360713676,0.2208167911,-0.1860036105,0.2577244639,-0.0299299285,-0.2412745506,0.2492341995,-0.2294996828,-0.0234921277,0.3386665285,-0.1834358722,-0.0722392276,0.1141103357,0.2787198722,0.0044050333,0.3821218312,-0.1071387082,0.1787773818,0.1578714401,0.1865142733,-0.205266118,-0.3974644244,-0.0073942491,0.5795645714,0.1509661376,0.0063893846,-0.155051887,0.0297769476,0.1568142325,-0.2301934958,-0.040392939,0.0557419024,0.1063028723,-0.0871282965,0.2420918941,-0.1243695691,-0.2532741725,0.1570651531,-0.1857650429,-0.3671966195,-0.1061075032,0.3231730461,0.2204078287,-0.1215085983,0.1551022232,-0.0787668228,0.2425985932,-0.3389588892,0.1390447319,-0.1475666463,0.3207711279,0.4877935052,-0.0260483865,0.0615058839,0.1670365483,-0.1229516342,0.0573805645,0.2521367073,-0.0455728471,-0.2194967419,0.3887047768,0.1053222641,0.322039932,0.0493761338,0.0182549581,-0.0762273669,0.1485420167,-0.25210163,-0.0411130376,-0.4598382115,0.3122323453,0.1221919358,-0.2666966915,-0.0598165765,-0.1620401889,0.1932631731,-0.5238361955,0.2419538647,-0.326086551,-0.0765336081,0.3290902674,0.0178690553,-0.078553699,0.6179023981,0.0901753306,0.6525301337,-0.2413716167,0.0449945219,-0.4157395363,0.0718611106,-0.0802068636,-0.0179308578,0.4335511923,0.1877977103,0.3594684601,0.3741153181,0.0650736317,-0.0139173875,0.3873788714,0.3575513959,-0.0418281443,0.0121625923,-0.5447233319,0.1933727711,0.0058101304,-0.3408643007,0.2303863913,0.035368409,-0.1486157477,0.38014099,-0.3933700621,0.4711250067,0.318721652,0.3260389864,-0.2265878618,0.6798644066,0.0250504296,-0.1249546036,-0.0659320801,-0.5260887146,-0.2678968012,0.3594198227,0.1478930414,-0.2117683291,-0.0471148938,0.0144740138,-0.286005348,0.1787313968,-0.1955964118,0.0516442321,-0.0779001787,-0.424318701,0.1215834469,0.2915966809,0.2652496397,0.1465907395,-0.100606218,-0.2337410301,-0.1721858978,0.0066255587,0.5101661682,-0.1248635873,-0.047891587,0.0060933731,0.3391577303,0.1818604171,0.2179677933,-0.5644078851,-0.037099313,0.266212374,-0.2286666483,0.0809670314,0.0042563397,0.1076988578,-0.0629155263,-0.2644650638,0.1196230501,0.0584042445,-0.0623077303,-0.3767543137,-0.321226269]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2538","title":"Loading partial dataset when debugging","comments":"Thanks all for responding.\r\n\r\nHey @albertvillanova \r\n\r\nThanks. Yes, I would be interested.\r\n\r\n@lhoestq I think even if a small split is specified it loads up the full dataset from the disk (please correct me if this is not the case). Because it does seem to be slow to me even on subsequent calls. There is no repeated downloading so it seems that the cache is working.\r\n\r\nI am not aware of the streaming feature @thomwolf mentioned. So I might need to read up on it.","body":"I am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!","comment_length":85,"text":"Loading partial dataset when debugging\nI am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!\nThanks all for responding.\r\n\r\nHey @albertvillanova \r\n\r\nThanks. Yes, I would be interested.\r\n\r\n@lhoestq I think even if a small split is specified it loads up the full dataset from the disk (please correct me if this is not the case). Because it does seem to be slow to me even on subsequent calls. There is no repeated downloading so it seems that the cache is working.\r\n\r\nI am not aware of the streaming feature @thomwolf mentioned. So I might need to read up on it.","embeddings":[-0.4159126282,-0.1481463164,-0.0125230812,0.3571513593,-0.0992514342,0.2774944305,0.5004057884,0.4125302136,0.3049237728,-0.1691022813,-0.0966696814,0.1207269281,-0.1135295779,0.2518861294,0.1217293143,0.0623838231,-0.1812782735,0.3625118136,0.0558305643,-0.1207429096,0.1732830107,-0.2011190802,-0.220249027,-0.2168607861,-0.388948828,0.0764960721,0.1006409377,0.1514767259,0.2842651606,-0.1781843603,0.3325754404,0.1009299606,0.0430348925,0.1529105902,-0.0001255008,0.0039876648,0.5047403574,-0.1904088855,-0.4295722842,0.0861378312,-0.2011617571,0.0131383929,0.4568482935,-0.0062859193,-0.08015275,-0.189124316,0.0196712483,-0.3495224118,0.3023769855,0.013711012,0.0576920882,0.068562746,-0.5156866312,0.1896018237,0.16702272,-0.0335522443,-0.0232582372,0.3747786283,0.4744539857,0.0265353732,-0.3703331649,0.1561926156,-0.1206538528,0.3349695802,0.4263782501,-0.0274212547,-0.0863643587,-0.5245525241,0.3088879585,0.4328550994,0.7110130191,0.0691519678,-0.3323349953,-0.5463773012,0.0553159453,-0.1214704737,0.1506922096,0.0559454076,-0.485868603,-0.0160152316,-0.4747502506,-0.1795731783,0.0094091091,0.1596933603,0.0370570645,-0.2802913487,0.0489187017,0.1070276201,0.3559331,0.2472990155,0.3180226088,-0.1036356241,0.0716243088,0.0855664909,-0.3632460535,0.0396836773,-0.187677294,-0.0634623393,-0.0351119936,0.4928974807,0.0537341237,0.4419742823,0.1951717138,-0.0539120585,0.5882905126,-0.123979561,0.1999562681,0.1320292652,0.3957653344,-0.2681331038,-0.042877268,-0.0303604696,-0.3341363668,-0.0585836023,0.2521030307,-0.0839881822,-0.0564097874,-0.3437398374,-0.1681816727,-0.3101805151,0.1414814293,-0.0881745368,0.2062132657,0.4173725247,0.0737162828,0.0797022656,-0.1862524152,0.1231173575,-0.2308132201,-0.1562998891,-0.1004408002,-0.2052611113,-0.0079848804,0.1595641524,0.380074501,-0.3889921904,0.3479875028,-0.1370400488,0.4121008515,-0.1663668156,0.1136634871,-0.2666145265,0.0171438418,0.2017515898,0.0424345732,0.2724153996,0.08878012,0.2402318567,-0.2842931747,0.4575053453,-0.0078219231,-0.3609118462,0.077408962,-0.0084563959,-0.3604274094,0.2339562178,-0.2522034645,0.3074853122,0.1119802073,-0.2342811525,-0.2774858177,0.0819797292,-0.0516638122,-0.0961853564,0.6512773633,0.4351745844,-0.1835158467,-0.2662341297,-0.174118489,0.0488526449,0.2412636876,0.2247572541,-0.4059408009,0.1950815767,-0.1556660384,-0.0307996646,0.5310043097,-0.0549808741,-0.3930068314,0.5534240603,-0.0423559137,0.1968268305,-0.0140453624,0.4981526732,0.2792948484,0.0834638178,0.0859898105,0.4684686959,-0.1412011087,0.1441644728,-0.2153415531,-0.116155602,0.1395133287,0.4743181169,0.0431787632,0.1887046248,-0.0309543032,0.1427442878,0.2508328855,0.0146926558,-0.0430910103,0.1715885252,-0.1290429682,0.0048909266,0.0708856434,-0.2515200078,-0.6060569882,0.3516191542,0.2121593505,-0.3468952477,-0.0178333838,0.0149980923,-0.1039042026,-0.1333016455,-0.0106944712,-0.1397803575,-0.1877261549,0.0580746755,-0.0552602075,-0.3899659216,-0.5041708946,0.1705367863,0.0138530703,0.0111177294,-0.3354501426,0.1781120449,0.4008476436,-0.0462546609,0.014477171,-0.1621901095,0.0068246978,-0.2368638664,-0.0759179145,0.2959756255,0.1486866921,0.2828854918,0.1145104691,0.1139545068,0.0202637967,-0.113908112,0.4275049567,0.0277546421,0.0352048762,-0.2094919235,-0.2280781865,0.3332342803,-0.4468023181,0.3899790943,0.0867231339,-0.2779377699,0.1355456561,-0.0903228298,-0.131014958,-0.0591623373,0.4690316916,-0.265332371,0.3321263194,0.205968067,-0.1701808721,-0.06763044,0.1805542856,-0.0547414944,-0.0281819832,0.3191450536,-0.5030067563,-0.0855560824,-0.071067363,-0.1535506248,0.1706553847,0.1412255168,0.1257363856,-0.2463175058,0.301281184,-0.0973696262,0.1797518879,0.076446712,0.0239915028,0.2401651293,-0.1692893654,-0.167725414,-0.2075762302,0.1335400492,0.1633007377,0.1261619031,-0.2461011857,0.0559538715,-0.3923882544,-0.137128979,0.0984191149,-0.0263084508,-0.0162439886,-0.1145453006,0.0295515209,0.3922593892,0.1370489895,0.1323356777,-0.2432091385,0.2042604983,-0.1511045545,-0.3757819235,-0.2814654708,-0.2621000111,-0.0485493205,0.0039448799,0.4429609478,0.1515236795,0.2691592872,-0.2304836512,-0.2361422777,0.1046250388,-0.0445631482,-0.0410590507,0.2351595908,0.214591369,-0.1406755447,0.3545921743,0.2172018588,-0.287086308,0.0678704306,-0.1821411699,-0.0087206867,0.079246074,0.203364253,0.3169822693,0.0007923888,-0.1470532268,-0.0798825473,-0.3423280418,0.3489235342,0.0219491087,0.1081119552,0.0131323934,-0.203190133,-0.1880278587,-0.0201029629,0.1227916107,0.0990580246,-0.2736402154,0.3888801932,0.1685303003,-0.1319755614,-0.0197492316,-0.0427994132,-0.1725602746,0.2457069606,-0.7163400054,-0.0363676995,-0.1039070711,-0.0275339149,-0.1411459446,-0.3152741194,-0.0093187643,-0.1638632268,0.0879862458,0.1405756325,-0.2466288656,-0.1660842001,-0.2383487821,0.0327039659,-0.0519156642,0.2790387273,-0.1051080972,0.7948543429,0.0669625178,-0.1358752996,-0.120862551,-0.115293853,0.0748092458,-0.4005529284,-0.3552450836,0.0047250567,-0.2175608873,-0.2861966491,0.1386241764,-0.0846051499,-0.3736602664,-0.2103448361,-0.0040322351,-0.2412600666,-0.2968298793,0.4443390071,-0.4901679456,-0.1639560163,0.0084250383,0.2829984426,0.0149871847,-0.308287859,-0.0442695357,0.1167140529,0.5464727879,-0.2055652589,0.0772358775,0.2033463567,-0.2294431925,0.1801040024,-0.138393119,0.1540877074,-0.0372023098,-0.1868480593,0.302375108,0.2027573735,0.6964053512,0.0290533137,0.1523524523,-0.0294492878,-0.5537374616,-0.3339633346,0.0513137393,0.1236577183,0.5427001715,0.1416492015,0.3319582641,-0.3412101865,-0.0857286528,0.0878414661,0.3475191891,-0.2562970817,-0.116563946,-0.2733144164,0.1115521491,-0.3391994834,0.0460334234,-0.0341027789,-0.2568398714,0.0771778673,-0.1358342916,-0.1969122738,0.0703859925,0.108294867,0.3095799088,0.2561495304,-0.0795991272,0.6283694506,0.2473173589,-0.0746685714,0.2623118758,0.4484378994,0.0541109405,-0.2391327024,0.2559788823,-0.0761414766,0.157031551,0.4805578887,-0.0273215398,-0.1740468144,0.1657722443,0.1821801066,-0.3535288572,0.0296650864,0.0437470041,-0.0363747366,-0.2842463553,-0.5381671786,0.4099707901,-0.1583420187,0.0270081926,0.5929956436,-0.2678991556,0.0433022231,0.1711855233,-0.0341253541,0.9001870751,-0.277594924,0.0181045588,-0.4738502502,0.0255156569,0.3826702535,-0.2529367208,0.0576607659,-0.3758229911,-0.6318646073,0.0451944917,-0.2700021565,-0.2560804486,0.2903873324,0.1918895692,0.2626982033,-0.3450230062,0.1452235729,-0.0896292105,0.2789677978,0.0092760315,-0.3133338392,-0.0787832364,-0.0112406258,-0.0686863586,0.182963118,0.1194632575,0.0324169174,-0.0137159182,0.1281000823,-0.4161299467,0.1509047449,-0.1354141384,-0.1083080173,-0.2610022426,-0.4459335208,0.001343939,0.1891569197,-0.169579342,0.0079538086,-0.0962304622,0.3113494813,-0.0028132608,0.2317947596,0.0342327654,0.0917195901,-0.0884342641,-0.0669412538,-0.0867955089,-0.1193685085,-0.2997490466,-0.4180701673,0.3202327192,0.2411245704,0.1727189273,-0.0488426052,-0.2446347773,-0.0208553448,0.155190587,-0.203387782,0.0130446563,0.4422787428,-0.0547042936,0.1766162962,0.2317460775,-0.3899621665,-0.032445278,0.4000256062,0.1774899065,-0.0122099575,0.4729733765,-0.0983900577,-0.3320080042,-0.0905298591,0.052817408,0.2913815677,-0.496655941,0.0077612139,-0.1362752765,0.1066785976,-0.1356606632,-0.255093962,-0.060070239,0.0896170139,-0.3603338599,-0.4292679429,-0.1603853405,0.2925671041,-0.0006636825,0.2813887,-0.2482956052,-0.1884153634,-0.0865625516,0.4756278694,-0.1367110312,-0.082528688,-0.0671442151,0.338463366,0.0354076549,0.0608307384,0.3525756299,-0.1260369867,-0.0385716073,0.0245586019,0.1656212509,-0.0732397214,0.1285141855,0.2295508087,-0.1557930857,-0.2205272913,-0.2334294766,-0.1802588254,-0.1248191595,-0.2638167143,0.2511461973,0.1583968997,0.169455871,-0.1106831506,0.4914953113,-0.0609234758,-0.4792806804,0.2214919627,-0.2967146039,0.350120157,-0.1263444871,0.2038220465,0.0991052315,-0.0105070248,-0.0357677825,-0.1017039791,0.3124388456,-0.1113364175,0.2047915906,0.0470271483,0.2950058281,0.0029082084,0.3840977252,0.8036357164,0.0334466174,0.2618505359,-0.0099868225,0.023774527,-0.0524262749,-0.1327249557,0.0429707989,-0.3547077179,-0.0327943973,0.2194919288,-0.1778031588,0.1971528679,0.0066397786,-0.2571290731,0.2985552847,-0.1595151424,-0.0018198262,0.3739388287,-0.2226196676,-0.0248825606,0.0594379678,0.2785044014,0.0101030739,0.3171236515,-0.1013442054,0.2024615109,0.1621011645,0.1747725755,-0.287936151,-0.4448860288,-0.0459660515,0.508471489,0.140929684,0.0349626429,-0.1158673242,-0.0183835402,0.1152741536,-0.2419969887,-0.058295045,-0.0145589821,0.1222931594,-0.0667558536,0.2357205003,-0.1662474722,-0.1868186146,0.1992184073,-0.2124332041,-0.3387030661,-0.1223848239,0.3224777579,0.136771962,-0.1114935949,0.0598550402,-0.0568305962,0.2279417813,-0.3398154974,0.1281609833,-0.1308227032,0.3174036741,0.5237315297,-0.0250464119,-0.0044355714,0.1036646217,-0.0880646259,0.0602627657,0.3079098463,-0.0386876352,-0.2340183258,0.3626216352,0.1883052438,0.3122766316,0.0332422182,0.0143885445,-0.0594106466,0.2364652306,-0.2298783809,-0.0372657813,-0.4761095345,0.2671964467,0.1054497734,-0.2371656597,-0.0161484946,-0.1595799029,0.1871889532,-0.4613913298,0.2017504722,-0.3219947219,-0.0484767258,0.3282836974,0.0040228739,-0.0981031731,0.6001446247,0.0621676147,0.560213089,-0.2283136547,0.0213623885,-0.4172227979,0.0945704207,-0.0164530221,-0.0068263086,0.4296316206,0.205763042,0.3834560215,0.3887084126,0.0355893299,0.0380241424,0.2914918363,0.3280867636,-0.0167885758,-0.0589557104,-0.5148507953,0.2842602134,-0.0555005819,-0.3336770535,0.2457808554,0.0031645459,-0.1566981524,0.4101821184,-0.4301159978,0.4235067964,0.3595992029,0.3506815732,-0.2170028389,0.6864227057,0.0228636824,-0.1376712322,-0.0296391081,-0.5314250588,-0.2944951653,0.3133070469,0.1260617375,-0.1157500595,-0.0128575014,0.0236607101,-0.2885636687,0.2185403854,-0.1884383857,0.0344264992,-0.0742425621,-0.4067562521,0.0970124528,0.3279888928,0.2838815451,0.1655914634,-0.1393087655,-0.2604670823,-0.1652660817,0.0357345194,0.5490170717,-0.1703853458,-0.0008907582,0.0074438574,0.330096066,0.1613235772,0.2442736179,-0.5656349063,-0.01879506,0.3026265204,-0.2663389444,0.0532588363,-0.0105472673,0.0401480272,-0.0577693805,-0.246151939,0.185306415,0.0904782861,-0.1217581853,-0.368589282,-0.3309917748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2538","title":"Loading partial dataset when debugging","comments":"@reshinthadithyan I use the .select function to have a fraction of indices.","body":"I am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!","comment_length":12,"text":"Loading partial dataset when debugging\nI am using PyTorch Lightning along with datasets (thanks for so many datasets already prepared and the great splits). \r\n\r\nEvery time I execute load_dataset for the imdb dataset it takes some time even if I specify a split involving very few samples. I guess this due to hashing as per the other issues.\r\n\r\nIs there a way to only load part of the dataset on load_dataset? This would really speed up my workflow.\r\nSomething like a debug mode would really help. Thanks!\n@reshinthadithyan I use the .select function to have a fraction of indices.","embeddings":[-0.447386682,-0.1285310686,-0.0407245494,0.361259222,-0.0726014525,0.3406393528,0.5750495791,0.4246875346,0.3035516143,-0.0632946417,-0.1243882328,0.146734044,-0.1679838449,0.2643848062,0.073236458,0.1298993677,-0.2119746059,0.3839837015,0.0369377621,-0.1140678003,0.1036522835,-0.227300182,-0.2530137897,-0.1577304602,-0.4047118425,0.0763561055,0.1515658796,0.1392650008,0.2717470527,-0.1731746048,0.3543200493,0.0715652481,0.0298938658,0.1139042303,-0.0001244871,0.0348428972,0.5232114792,-0.1667564362,-0.3730357289,0.0409000814,-0.2502275407,0.0191153828,0.4595259428,-0.0197098944,-0.0962022915,-0.2555598319,-0.0040840213,-0.3240305483,0.2482997477,-0.0148501489,0.077368699,0.063697733,-0.5058805346,0.1707746089,0.1863405704,-0.0790968016,0.0232793335,0.3963984549,0.4162446558,-0.0168241672,-0.2896716893,0.1443715543,-0.0961545333,0.2583311796,0.3825308084,-0.021011522,0.0513886437,-0.5506599545,0.2650492489,0.4527368844,0.678920567,0.0725387111,-0.3213124573,-0.508703351,0.0736221299,-0.0657505989,0.1334382892,0.0052725188,-0.4283075035,-0.0058336486,-0.484759748,-0.1213999987,0.0485870317,0.1435144544,0.0577906035,-0.2032269239,0.0486687534,0.1429840028,0.3976275623,0.1985020936,0.3011806011,-0.011427016,0.0910431668,0.0837993473,-0.3526317477,0.0472192205,-0.1387039423,-0.1567374766,-0.0304501951,0.501093924,-0.0567359775,0.4659517705,0.2174210846,-0.0240471456,0.5122939944,-0.1437479407,0.2391391993,0.1768762767,0.3303665221,-0.3814075589,-0.0420052558,-0.0333648883,-0.3823722005,-0.1223800704,0.2276696414,-0.0236820076,-0.0563748069,-0.3556757569,-0.186326474,-0.3144237399,0.0981663167,-0.0440734141,0.2113389522,0.4513784051,0.122356832,-0.0081359725,-0.1916611642,0.1329114586,-0.2563944757,-0.0308192894,-0.1148137227,-0.1986619979,-0.0083535854,0.0838827044,0.3111350536,-0.3730425239,0.3519791365,-0.0854737014,0.4622256458,-0.2014769465,0.1569654942,-0.257150352,0.0191340782,0.209146589,0.0322146565,0.2476131618,0.0836621299,0.184937343,-0.2815234959,0.4369626045,-0.0272104889,-0.3249307275,0.0518303812,0.0003277446,-0.2924782038,0.22351183,-0.2111565173,0.3034892678,0.1096146181,-0.18750076,-0.2653904259,0.0719789118,-0.0591758303,-0.1287774593,0.6780409217,0.4159381986,-0.1474419087,-0.240604952,-0.1284250468,0.0003847936,0.2431779951,0.178666383,-0.4257194102,0.2300992161,-0.1208717152,-0.0707549453,0.5156366229,-0.0336716622,-0.3653177619,0.484049052,-0.0961191878,0.1197000369,-0.0582715869,0.5120244622,0.2024565339,0.0579108782,0.1340440959,0.4516094327,-0.149312377,0.1000592634,-0.1713815778,-0.1060001925,0.1843746454,0.4783651531,0.0790555626,0.182115525,-0.0661780089,0.1072014123,0.20396626,-0.0045875558,-0.0951683074,0.2074518204,-0.1423698664,0.0254071299,0.0625900552,-0.3369059265,-0.573158741,0.3624549806,0.1751195937,-0.3117127419,-0.0204113629,0.0110180806,-0.1206527799,-0.1043994203,-0.0007975939,-0.0336725675,-0.1766057909,-0.002968468,-0.0905310065,-0.3952457011,-0.503282845,0.080838941,0.0235317964,0.0061898851,-0.3445114195,0.1209030822,0.4227523208,-0.0604555495,0.0272985231,-0.1320168972,0.001102603,-0.2820909619,-0.066339761,0.2560337484,0.1403706819,0.2201097012,0.1013264731,0.101820007,-0.0257738698,-0.1299439967,0.3763619363,-0.0336989053,0.0225818455,-0.1797884107,-0.1515427232,0.3810150921,-0.4986324906,0.3906364441,0.0185171459,-0.3176044822,0.1262272447,-0.1534607708,-0.1374496818,-0.1160037294,0.4230465293,-0.2301741242,0.3246203661,0.2353230566,-0.1106415689,-0.0765416995,0.1714252084,-0.0347147137,-0.0239821747,0.2861795127,-0.465757072,0.0145947579,-0.1663379371,-0.2573722005,0.0808927119,0.1597539634,0.0929466859,-0.2641517222,0.2672224939,-0.1038162038,0.167868197,0.0719766691,0.0506881922,0.1988302618,-0.1711767167,-0.1603628695,-0.1973939687,0.0962739587,0.1251773089,0.1454047859,-0.2760202289,0.0460730307,-0.4004756212,-0.1121206582,0.167240575,-0.0453683585,0.0499025546,-0.1244248375,0.0237046778,0.269159466,0.0930054784,0.1686986238,-0.2670358717,0.1221245602,-0.0787679702,-0.3659653664,-0.2293305993,-0.2517744005,-0.0493517593,0.0085523678,0.4665497839,0.1616446823,0.3183942735,-0.2038255334,-0.2458742559,0.1351896673,-0.0993510187,-0.0464732721,0.2011560798,0.2636519969,-0.1133373603,0.3230070472,0.1960374564,-0.3134385347,0.0970407426,-0.0784522295,-0.0210787319,0.0887156799,0.2221948951,0.3033585846,0.0434129126,-0.1537567824,-0.0297427494,-0.3395304382,0.332373172,0.0373399891,0.1385136396,0.0451956056,-0.1379437298,-0.2049051672,-0.0557988361,0.0816137567,0.131382376,-0.2460149825,0.3933897614,0.1874917746,-0.0648906901,-0.0742313117,-0.0394640453,-0.2159910202,0.21273233,-0.6363862157,-0.0219800808,-0.1309787482,0.0088978875,-0.1401080042,-0.2720962763,-0.0386422612,-0.1099147871,0.0608247258,0.1905197501,-0.2001960874,-0.133913666,-0.266546458,0.0304525997,-0.0162873883,0.3362723887,-0.1549628228,0.7903823256,0.0411187112,-0.1731878966,-0.1495545655,-0.2018626779,0.0583077297,-0.3980112374,-0.3809620142,-0.0678028092,-0.053794913,-0.2266841382,0.0946594998,-0.1113554984,-0.3336943984,-0.1670865715,-0.0171615835,-0.2656079531,-0.3168419898,0.5021750927,-0.5481266379,-0.1931648701,0.012283775,0.273855418,0.0015495013,-0.2858014107,-0.0113992086,0.0688269138,0.511847198,-0.2614889741,0.0986485779,0.1802844554,-0.1812159866,0.1979950219,-0.1076414809,0.1721325368,-0.0112856571,-0.1828271002,0.2780281007,0.1796883047,0.6776302457,0.0143232867,0.1497751623,0.0204356704,-0.572806716,-0.3478342891,0.0494266823,0.1031564176,0.6130971313,0.1012171656,0.3887572885,-0.3497774601,-0.1270559132,0.1823102087,0.3537808955,-0.2841822505,-0.1132295355,-0.3334885836,0.0744769424,-0.3311693668,0.0834967569,-0.0623944737,-0.2250567824,0.1181978807,-0.2068853676,-0.3096635044,-0.0059824055,0.1465038359,0.3246575594,0.2367628366,-0.0363450758,0.6749070287,0.2587372661,-0.1354241371,0.305347234,0.4033463895,-0.0060702595,-0.2159498185,0.2448267788,-0.0563727841,0.2532421947,0.4493032694,-0.0650620311,-0.1633025557,0.1530885994,0.2045379877,-0.3692820668,0.0382508188,0.0552160628,-0.0107895564,-0.2925263941,-0.519333899,0.4370409846,-0.2015679628,0.0212102365,0.6017226577,-0.1685863882,0.0283260643,0.2049554735,0.0264552571,0.9092683196,-0.3385969698,-0.0587473847,-0.4637214541,0.1262218654,0.3634110689,-0.2128037512,0.0340357088,-0.4131718278,-0.6919583082,0.08339376,-0.2903065979,-0.3036001027,0.2893931866,0.2845687568,0.2182037085,-0.4634791613,0.0865006149,-0.0779621452,0.3162949979,0.0458024628,-0.288652271,-0.0541119948,-0.0015775554,0.017897144,0.106490992,0.0954130366,0.0589656159,-0.0876073614,0.1322075427,-0.3767975867,0.1254599541,-0.1627168953,-0.0521646589,-0.2214471102,-0.4097556174,0.0263661165,0.1379512995,-0.1237177253,0.1154235378,-0.0332169197,0.2928991318,0.0202097762,0.1622357517,0.0003192077,0.0709567815,-0.0759945586,-0.0179370958,-0.1268277764,-0.1780910343,-0.2706538737,-0.4096835256,0.3016135991,0.2849232256,0.0749904811,-0.0830038115,-0.284581244,-0.0674210414,0.1989515573,-0.1553567201,0.0348841436,0.4448956251,-0.0327841938,0.1888445616,0.3392931819,-0.4467830658,0.0118004177,0.3575645685,0.1718283445,0.015415092,0.48432827,-0.1009522304,-0.3009106219,-0.0818130299,0.0901404917,0.3964306712,-0.4315536022,0.0358877666,-0.1114657968,0.1024684161,-0.1210947111,-0.2255038768,-0.0493867882,0.034448076,-0.3786666989,-0.4190530777,-0.1634763479,0.3583892286,0.0019335228,0.3014194667,-0.2658078671,-0.1665570885,-0.1543822885,0.5083736181,-0.1504089087,-0.100149557,-0.1603712589,0.3797670603,0.0784239769,0.050503768,0.3439181447,-0.1566068679,0.0069809463,0.0206836145,0.1808644235,-0.108882308,0.0804140568,0.2091507167,-0.2088889331,-0.165886268,-0.1945036501,-0.1439612806,-0.1471596509,-0.2597184181,0.2163208723,0.1299317628,0.1893774122,-0.1099474058,0.4341238737,-0.0944610834,-0.4457814693,0.2304371595,-0.3292762041,0.3450809717,-0.1134675145,0.2214812338,0.0878206342,-0.0149867665,-0.0349925496,-0.1269187033,0.352868408,-0.1340359598,0.1030413061,0.0702254772,0.3123515546,0.0651106462,0.3824336827,0.8295709491,0.0194585826,0.2710463405,0.0350251272,0.0302219316,-0.0789593533,-0.1003216431,0.0032927534,-0.3895181417,-0.0041898256,0.2287288159,-0.2363290936,0.2510454655,-0.0252089575,-0.2837374806,0.2573925257,-0.1331163049,-0.0671725124,0.3244846165,-0.2504144907,-0.0529482886,0.1134010851,0.285276562,0.0673567653,0.403414458,-0.0986757949,0.1639302671,0.1187493652,0.1689911038,-0.313347429,-0.4320175052,0.0458034985,0.5572580695,0.1556054652,-0.0202267244,-0.0965882242,0.0018765543,0.1617403179,-0.1835684627,-0.00947937,0.0876416638,0.0969868302,-0.1406752318,0.2261236459,-0.1584310234,-0.2125062197,0.1485019475,-0.1555842161,-0.3642870784,-0.1705725789,0.3186227083,0.1933355182,-0.092318587,0.1299974471,-0.0002900891,0.2111216187,-0.3629816771,0.089592278,-0.1197175011,0.3424062729,0.4765996933,-0.0269523803,0.0075267106,0.0924597308,-0.1435025632,0.0450154021,0.2811688781,-0.0455039218,-0.2497178763,0.3247710168,0.1755374372,0.3854746819,0.043670442,0.0389454849,-0.0853974149,0.0920737535,-0.1783699393,-0.027008079,-0.4694225788,0.2343730032,0.1171428859,-0.2405125201,-0.0271336883,-0.1826207936,0.1835137904,-0.4851186574,0.2038838863,-0.3042057157,-0.0156986639,0.4072469473,0.0185205676,-0.073111698,0.6266226768,0.0413480662,0.5901729465,-0.2506599128,0.0733828172,-0.4299322665,0.0622357428,-0.0065097171,-0.005293712,0.4601902664,0.1921355873,0.4238529205,0.3666062057,0.0589230135,0.0716014579,0.3663998246,0.3406557441,0.0131233353,-0.0127049107,-0.5911300778,0.2127960026,-0.0501836464,-0.2984931171,0.2053254396,-0.022070216,-0.1266550571,0.3595016897,-0.4222143888,0.4205596149,0.3250746131,0.4193101823,-0.2373766601,0.7086356878,-0.0096676247,-0.0893202946,0.0006469951,-0.5426053405,-0.2606763244,0.3406434357,0.1833275259,-0.1628098041,-0.0250270814,0.000236332,-0.2701191902,0.1532265842,-0.1649273187,0.0207154602,-0.1041568518,-0.4156496227,0.1158279106,0.265378356,0.238438949,0.1635811478,-0.0982511714,-0.2282428145,-0.1294550747,-0.0099005634,0.5226024389,-0.1604071259,-0.0557767749,0.0147285406,0.3215407431,0.159152776,0.2374351919,-0.5177510977,-0.0128807928,0.3209348023,-0.2451553047,-0.0099429488,-0.0108148726,0.0896197334,-0.0498145148,-0.2753442526,0.1199786291,0.0856049582,-0.0623270124,-0.4467021823,-0.3088541031]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2536","title":"Use `Audio` features for `AutomaticSpeechRecognition` task template","comments":"I'm just retaking and working on #2324. \ud83d\ude09 ","body":"In #2533 we added a task template for speech recognition that relies on the file paths to the audio files. As pointed out by @SBrandeis this is brittle as it doesn't port easily across different OS'. \r\n\r\nThe solution is to use dedicated `Audio` features when casting the dataset. These features are not yet available in `datasets`, but should be included in the `AutomaticSpeechRecognition` template once they are.","comment_length":8,"text":"Use `Audio` features for `AutomaticSpeechRecognition` task template\nIn #2533 we added a task template for speech recognition that relies on the file paths to the audio files. As pointed out by @SBrandeis this is brittle as it doesn't port easily across different OS'. \r\n\r\nThe solution is to use dedicated `Audio` features when casting the dataset. These features are not yet available in `datasets`, but should be included in the `AutomaticSpeechRecognition` template once they are.\nI'm just retaking and working on #2324. \ud83d\ude09 ","embeddings":[-0.4086931646,-0.1050031781,-0.0201717932,-0.059146814,0.1830265522,-0.2864357233,0.3036151528,0.3112901747,0.0281815883,0.2801087201,-0.1647812873,0.4321773052,-0.1427244544,-0.059097372,0.0046525872,-0.1048489511,0.1057824418,0.035588488,0.0927574337,-0.0723030642,-0.3093028069,0.2068701386,-0.2040395737,0.0566584729,-0.2387037873,-0.0808997676,0.1003127545,0.0482092351,0.0035730759,-0.1962217093,-0.255317241,0.1002381071,0.297393918,0.1947508603,-0.0000963321,-0.0411492251,0.3601718843,-0.1581388712,-0.1789464653,-0.0871949419,-0.3767960072,0.2580452263,-0.0995693058,-0.1831744015,-0.1789264977,0.0308268555,0.3169716895,-0.1457207799,0.3007699847,0.4298517704,0.2866197824,0.1993984729,-0.2073910981,-0.0863192677,0.1948210448,0.0420197435,-0.1634683907,-0.1530340761,0.3579793274,0.0885905251,-0.1571439952,0.2450844646,0.0179813411,0.0204808135,0.1513580382,-0.1965816319,-0.0897527561,-0.3769333959,0.0489161015,0.3464509547,0.5198340416,-0.2709258199,-0.1233493984,-0.1951080263,0.0555988587,-0.147610873,0.2158052325,-0.0158634782,-0.2217231989,0.1630329341,-0.1301387846,-0.0640602708,-0.1702833027,-0.1699832082,0.0051031234,0.1107488573,-0.0941171721,-0.2799037397,0.2093928754,-0.2059440166,0.1261190027,0.0716790408,-0.0310270861,0.0013229405,-0.1749815494,-0.1920051724,0.0685875118,0.0464709438,-0.175175935,0.0219066124,0.3185777068,0.3930148482,-0.1309720874,0.2581318915,0.0262891594,0.0449823104,-0.0163258109,-0.1227642894,0.3606704772,0.1184275448,-0.2132749259,-0.147845611,-0.1903038919,-0.1774535626,-0.059472546,0.1236066148,0.1025081351,-0.164781943,-0.3428489864,-0.208362788,0.0757937059,0.1360648572,0.1516780108,0.2997315526,0.0008111633,0.050802961,0.1346632391,0.2947484553,-0.2497188449,-0.158362031,-0.0548873655,-0.0454810783,-0.1308016181,0.0203632265,0.1688898653,0.5101751685,0.3465901017,-0.159585163,0.1566578895,0.002279269,0.1381274313,0.0695623308,-0.0244930815,-0.0519992635,0.0011667087,-0.0982134193,-0.2934915423,0.0167188793,-0.1600058675,0.0388257429,-0.0903314427,-0.0634006783,0.1359282434,0.3139289916,-0.1536630392,-0.2246674597,-0.0574145652,0.3851167262,-0.0838597864,-0.2730384767,0.087355569,-0.2398342192,-0.2534611523,0.0402487777,-0.0100622773,0.0069450103,-0.1596678197,-0.1235274449,-0.1833073497,-0.1389623284,0.1861619502,0.2844409645,-0.0719307065,0.1865472347,-0.1944165975,0.5834636092,0.3377868533,-0.3347233236,-0.1356847137,0.0786628202,-0.1230560318,-0.0268075373,0.1674745828,-0.108656399,0.4572125077,-0.3210780323,-0.1084742323,0.34597224,0.1094340459,0.0409928709,-0.1921253204,-0.1480905712,-0.1425994486,0.1659081131,-0.0088322703,-0.0098174261,0.0423714034,0.2330199927,0.2097470611,-0.2864784896,0.216317907,-0.0194532555,0.4233000576,-0.1072912663,0.0606737509,-0.2867967188,0.1189947575,-0.0150399115,-0.0961974785,0.0846548751,-0.1856307387,-0.2463337481,0.162490949,-0.1424733102,-0.1206437796,0.1494915783,0.2941848338,0.3098862171,-0.1176900938,-0.2090142369,0.0805606022,0.1546125114,-0.0414442793,-0.1798602343,-0.0427375771,0.0793086886,-0.0590477847,-0.1663336456,0.1329140514,0.1804130375,-0.0044005406,0.004467831,0.0061881142,0.3546147645,0.2265910655,0.3611976802,-0.1507905722,0.2201506346,0.0702827573,-0.4674694836,0.2787438035,0.2898763418,0.0865644068,-0.0826253965,-0.076688759,0.0215987545,0.2652099431,0.1270402819,0.3886805475,-0.0470136069,0.3482189775,-0.0447865576,-0.2862187624,-0.0343777761,-0.152124092,-0.0911751837,0.1079891846,0.095588021,-0.7945018411,0.0903532356,0.5879005194,-0.0549497046,0.2874765396,0.1773768365,-0.2114792466,-0.1728336811,0.1503539532,0.0098828515,0.1620692015,0.3717464507,0.3493570685,0.0174667407,-0.2363197058,-0.304600507,0.2277045846,0.1090376303,-0.215942204,0.1638945937,0.0112595186,-0.1863638759,-0.5062639713,-0.1007653028,-0.0145833762,-0.0885102451,-0.3023743629,-0.2195234448,-0.3403604627,-0.3917853832,-0.3816170692,0.2884105444,0.0878639445,-0.0393881947,0.4357614815,0.2359663993,-0.4417285919,0.2441521883,0.0083511295,0.2355541736,-0.2294689119,-0.1162600294,-0.0448290296,0.0136757083,-0.1024614722,0.2344507128,0.3223003447,0.2189682722,0.053747464,-0.038718015,0.1479588896,-0.2570061684,-0.0617503524,0.1097968593,0.1113056242,-0.006030763,-0.0897748396,0.0402872674,-0.1768804044,-0.453376472,-0.066145964,-0.1285496205,-0.2373443842,0.0330547616,-0.0495535582,-0.1353896111,-0.2270616293,-0.3631029427,-0.093350336,-0.5635882616,0.2140187323,-0.1307651252,0.0171323605,0.3450217843,-0.0690367296,0.0428863876,-0.2192558348,0.2802630663,-0.078859359,-0.0953557119,0.2489200532,-0.2555911541,-0.2201833427,-0.1137652546,-0.0778075084,0.3122576475,0.1675541997,-0.0712686852,-0.1939484328,-0.1156159788,0.1885124445,0.1143866479,-0.0243265312,0.4211913049,0.2096885294,-0.1896309555,-0.2936011255,0.00099143,0.2937131524,0.4734397233,0.2362775505,0.149834916,-0.002028795,-0.2104858756,0.6716897488,0.1588516086,0.0129597103,0.422719121,0.0920931175,0.2073975801,-0.2861243784,0.1412866712,0.4234283268,0.043048244,-0.140506193,0.5683684945,-0.007077673,-0.4157935381,0.0631255731,-0.2001614124,-0.2400111705,-0.3546171188,0.0666415915,-0.0147719076,0.0684651807,-0.005298764,0.1817323864,-0.0459335111,-0.0128181782,0.2687338889,0.074036248,0.2587099075,-0.0824049488,-0.1828190982,-0.1358377188,-0.2979820967,0.0396259055,-0.0530112982,0.2623011172,-0.2559121847,-0.461107254,0.0026713745,0.0464606136,0.1357030421,-0.320758611,-0.0247731302,0.1135051847,0.3246741295,0.0579310469,0.1705658883,-0.2652007639,-0.2677805722,0.1005167961,0.0386152901,-0.0220004041,0.0637151897,-0.1562235951,0.1459171027,0.0112860687,0.0627542585,0.1387467086,-0.3152100146,-0.1753987074,0.0969111398,0.323450923,-0.0756859258,-0.1591855288,-0.0220717136,0.0611341819,0.024087837,0.2028177977,0.2267166227,0.257337451,-0.0388885103,0.3998528719,0.0842980519,0.2609118521,0.1205465645,0.2683168352,-0.3110181391,-0.3243673444,-0.0591979772,-0.2281060368,0.311453253,0.3350704908,-0.013946143,-0.0191499442,-0.0272135343,0.1327553689,-0.3574898839,0.3062413931,0.3269279301,0.1200762913,-0.153826654,-0.246895805,0.0709385499,-0.0150164794,-0.3533083498,0.0911409706,-0.191738233,-0.4083380997,0.5319373608,-0.098923251,0.8710600734,0.2300246209,0.0934732631,0.1675094366,-0.1349897236,-0.0792217404,-0.2480356693,0.230116874,0.0995507464,-0.224617824,-0.0189736057,0.1286158264,-0.0353546776,0.408882618,-0.476133883,0.1297675818,0.2587144971,0.0444077402,-0.0988648161,0.3561104536,0.0423224382,-0.3852671087,-0.0006687262,0.2545185387,0.1794464588,-0.0094219698,0.1004771292,-0.2856440246,0.2615228891,-0.1058594659,-0.1319047809,0.1189918518,-0.2125316411,-0.2523253858,-0.112002939,0.0639764369,0.3819412887,0.0755435079,0.3201592565,0.2417000234,-0.3487997949,0.3133288324,0.2726199329,0.3088996708,-0.0365513302,-0.1097034216,0.3781642914,-0.1730969399,-0.1463186741,-0.1535047144,0.0697839335,-0.0327999815,-0.1644587219,-0.1144503877,0.037814077,-0.0978891328,-0.0806450248,0.0306569114,-0.2199447304,-0.2254349887,0.3029114604,-0.049168203,-0.2197429538,0.5200899839,-0.0038990043,-0.0489010662,-0.1395698935,0.1231885403,-0.1906936765,-0.0113372626,0.3652483523,-0.0309449472,-0.2561483681,-0.445826143,-0.1098067909,0.0047426187,-0.1949906945,0.1102661937,0.0977864414,-0.3369669616,-0.0807433203,0.289870441,-0.1704837829,-0.1121769547,-0.4577791989,0.0582098365,-0.3307378888,0.2583136857,0.0011876064,0.3461329043,-0.1693394184,0.1493158937,0.3088995516,-0.2482117862,-0.5040875673,-0.0238392279,-0.0653811544,0.2780072689,0.0811197013,-0.1101868227,0.1053265706,0.1081122458,0.2232851386,0.1077058166,-0.1398743391,-0.3249235153,-0.0734911337,0.0826179385,0.2019304484,-0.0345925614,0.1105433628,-0.1837886572,-0.1806087196,-0.2298028767,0.0250039455,0.1605441123,-0.0593948476,0.3109456897,0.1118803844,-0.1122932136,0.4157781601,0.1004622728,0.0399257615,0.4100133181,-0.1777514368,-0.0034056476,0.0362131856,-0.0666981116,0.2768781185,0.2282360494,0.2264581174,0.0261343922,0.135980323,0.226585418,-0.1098567322,0.2192001045,0.2609254718,0.1284379661,-0.0203901585,-0.2875512242,0.4885424674,0.4400545061,-0.2770483196,-0.0169103369,0.0880407616,-0.3699617684,0.2847819328,0.2227419168,0.1266119778,-0.224667877,0.1178940013,0.0480399765,0.2002303004,-0.3083555698,0.135796845,0.2340245396,0.2088409811,-0.0831121951,0.0643796101,0.0170063619,0.226424858,0.5066189766,-0.2777348161,0.0918323323,0.1908119172,0.0639957115,0.4014004469,-0.1950380355,-0.1011952683,0.0162384659,-0.1308182329,0.1275267601,0.0059765503,0.2419672608,-0.1460411996,-0.0489063561,-0.0907385051,0.0858457163,-0.2741266191,-0.2091843486,-0.1691160202,-0.1637024432,-0.1371312737,-0.1986323446,0.089081578,0.3713915646,0.0590474829,-0.0584666468,0.1436846256,0.0881195664,-0.0604757853,-0.2206505388,0.1444152892,-0.267164737,0.355274111,-0.1108772606,-0.1264697909,0.5917447209,0.3655821085,0.3704828024,0.0976921692,-0.2749675512,0.1668695956,-0.1492282152,-0.0765970498,-0.3827383518,0.2444763482,0.1470358372,0.2518158853,0.4284171462,0.2284623533,-0.2074988186,0.1358778626,-0.0817871094,-0.2379300892,-0.1965880543,0.3660475016,0.134014532,-0.070844382,-0.2911688387,-0.2584937513,-0.5256902575,-0.1413001865,0.3006537855,-0.1700284928,0.1398434788,-0.1137069315,0.1594375372,-0.247569412,0.2202997655,0.1821688414,-0.0380734019,0.0077766064,-0.1761452556,-0.4746288955,0.1130940765,0.1282566637,0.0024786014,-0.1828770489,0.0700487867,0.1161030158,-0.2107350379,0.0047373404,-0.0021813277,-0.0694671273,0.1991161108,-0.1900323182,0.1436292529,0.0776623189,0.2009733021,0.0697507188,-0.2771353126,0.0757044107,0.1694975793,0.2007475495,-0.0252318643,0.0847772658,-0.069606021,-0.1808115244,0.3404690027,-0.076287739,0.282096535,-0.4090268314,0.1196506396,-0.0226122923,-0.0751718581,-0.3510738909,-0.2397661805,-0.0392061844,0.4691056311,-0.0441666134,0.3125301898,-0.1604415327,-0.2603332102,-0.0071973521,0.0669128224,-0.2082976252,0.231682241,-0.4568899274,0.0437148884,0.0405689888,0.0477812998,-0.0683505833,-0.2110679299,-0.2908849716,-0.7309217453,0.4421878457,-0.0341662951,-0.3487739265,-0.3762152195,0.4532550871,-0.4248762727,0.1753982902,-0.7348421812,-0.0292822216,0.1533158123,-0.0218099039,0.1436608285,-0.0083566243,0.0705620795,-0.0325439796,-0.1810919493,0.2310097516,0.2829473317,0.07420367,-0.0719799995,-0.3776144981]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2532","title":"Tokenizer's normalization preprocessor cause misalignment in return_offsets_mapping for tokenizer classification task","comments":"Hi @jerryIsHere, thanks for reporting the issue. But are you sure this is a bug in HuggingFace **Datasets**?","body":"[This colab notebook](https:\/\/colab.research.google.com\/drive\/151gKyo0YIwnlznrOHst23oYH_a3mAe3Z?usp=sharing) implements a token classification input pipeline extending the logic from [this hugging example](https:\/\/huggingface.co\/transformers\/custom_datasets.html#tok-ner).\r\n\r\nThe pipeline works fine with most instance in different languages, but unfortunately, [the Japanese Kana ligature (a form of abbreviation? I don't know Japanese well)](https:\/\/en.wikipedia.org\/wiki\/Kana_ligature) break the alignment of `return_offsets_mapping`:\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/122904371-db192700-d382-11eb-8917-1775db76db69.png)\r\n\r\nWithout the try catch block, it riase `ValueError: NumPy boolean array indexing assignment cannot assign 88 input values to the 87 output values where the mask is true`, example shown here [(another colab notebook)](https:\/\/colab.research.google.com\/drive\/1MmOqf3ppzzdKKyMWkn0bJy6DqzOO0SSm?usp=sharing)\r\n\r\nIt is clear that the normalizer is the process that break the alignment, as it is observed that `tokenizer._tokenizer.normalizer.normalize_str('\u30ff')` return '\u30b3\u30c8'.\r\n\r\nOne workaround is to include `tokenizer._tokenizer.normalizer.normalize_str` before the tokenizer preprocessing pipeline, which is also provided in the [first colab notebook](https:\/\/colab.research.google.com\/drive\/151gKyo0YIwnlznrOHst23oYH_a3mAe3Z?usp=sharing) with the name `udposTestDatasetWorkaround`.\r\n\r\nI guess similar logics should be included inside the tokenizer and the offsets_mapping generation process such that user don't need to include them in their code. But I don't understand the code of tokenizer well that I think I am not able to do this.\r\n\r\np.s.\r\n**I am using my own dataset building script in the provided example, but the script should be equivalent to the changes made by this [update](https:\/\/github.com\/huggingface\/datasets\/pull\/2466)**\r\n`get_dataset `is just a simple wrapping for `load_dataset`\r\nand the `tokenizer` is just `XLMRobertaTokenizerFast.from_pretrained(\"xlm-roberta-large\")`","comment_length":18,"text":"Tokenizer's normalization preprocessor cause misalignment in return_offsets_mapping for tokenizer classification task\n[This colab notebook](https:\/\/colab.research.google.com\/drive\/151gKyo0YIwnlznrOHst23oYH_a3mAe3Z?usp=sharing) implements a token classification input pipeline extending the logic from [this hugging example](https:\/\/huggingface.co\/transformers\/custom_datasets.html#tok-ner).\r\n\r\nThe pipeline works fine with most instance in different languages, but unfortunately, [the Japanese Kana ligature (a form of abbreviation? I don't know Japanese well)](https:\/\/en.wikipedia.org\/wiki\/Kana_ligature) break the alignment of `return_offsets_mapping`:\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/122904371-db192700-d382-11eb-8917-1775db76db69.png)\r\n\r\nWithout the try catch block, it riase `ValueError: NumPy boolean array indexing assignment cannot assign 88 input values to the 87 output values where the mask is true`, example shown here [(another colab notebook)](https:\/\/colab.research.google.com\/drive\/1MmOqf3ppzzdKKyMWkn0bJy6DqzOO0SSm?usp=sharing)\r\n\r\nIt is clear that the normalizer is the process that break the alignment, as it is observed that `tokenizer._tokenizer.normalizer.normalize_str('\u30ff')` return '\u30b3\u30c8'.\r\n\r\nOne workaround is to include `tokenizer._tokenizer.normalizer.normalize_str` before the tokenizer preprocessing pipeline, which is also provided in the [first colab notebook](https:\/\/colab.research.google.com\/drive\/151gKyo0YIwnlznrOHst23oYH_a3mAe3Z?usp=sharing) with the name `udposTestDatasetWorkaround`.\r\n\r\nI guess similar logics should be included inside the tokenizer and the offsets_mapping generation process such that user don't need to include them in their code. But I don't understand the code of tokenizer well that I think I am not able to do this.\r\n\r\np.s.\r\n**I am using my own dataset building script in the provided example, but the script should be equivalent to the changes made by this [update](https:\/\/github.com\/huggingface\/datasets\/pull\/2466)**\r\n`get_dataset `is just a simple wrapping for `load_dataset`\r\nand the `tokenizer` is just `XLMRobertaTokenizerFast.from_pretrained(\"xlm-roberta-large\")`\nHi @jerryIsHere, thanks for reporting the issue. But are you sure this is a bug in HuggingFace **Datasets**?","embeddings":[-0.1536964476,0.0790127367,0.0737818331,-0.0382318683,0.0921575204,-0.3246818185,-0.0899818391,0.1060319841,-0.461761117,0.1986739933,-0.2085373253,0.307420373,0.2766766548,0.0308846552,-0.2191910297,-0.1632778943,0.1479931921,0.2190573663,0.0056235306,0.0611228123,-0.1194189712,0.4571387172,-0.212335363,0.2143603265,-0.2866258919,-0.0706363171,0.125577569,-0.315648973,-0.1683979779,-0.3988147974,0.249714613,0.118910946,-0.0836713836,0.2043143213,-0.0001126969,-0.057501819,-0.0180353783,0.0142885102,0.178675279,-0.1964323819,0.128385514,-0.3769181371,-0.3356096447,0.1073889881,-0.0636701211,0.0396103747,0.2408124506,-0.0554189608,0.6431404352,0.4151496589,0.1455552727,-0.0134888468,0.2545183599,0.0607683174,-0.3109235764,0.2689674497,-0.095067516,0.0190099981,-0.1540722102,0.031468261,-0.0746515244,-0.0096175401,-0.0922699124,-0.4802750349,-0.0148855951,0.1148962528,0.3214075863,-0.4625909626,0.1870723367,0.0185327325,-0.2498886287,-0.0450737998,-0.1830774248,-0.3914948106,-0.1368259192,-0.2679938078,0.2000967115,-0.1542632878,-0.1181238517,0.1168286502,0.0120839812,0.1378460675,0.0297428202,-0.212039277,-0.2942862511,0.564735353,-0.0654119551,0.0426447168,-0.2047559917,-0.0101762712,0.3918432295,-0.042820774,0.0059831715,0.4373289049,-0.2852554321,-0.058525499,0.1970698237,-0.4548270106,-0.0079325698,-0.0356674455,0.1242711842,0.2060957551,0.0719589293,0.0609170049,-0.1630160064,0.4642789066,0.2114510983,0.2766360343,0.1968423277,0.1170869023,-0.0398001708,-0.0404408984,0.0627822056,-0.1251495481,-0.0498765483,0.2508600056,-0.3735534847,0.1039711758,-0.2315127552,0.4121437967,-0.0718666613,0.3000814319,0.0338656791,-0.0993585065,0.0393204316,0.2893413007,-0.3403167427,-0.1019151807,-0.0826219469,0.1086550057,-0.1952837557,0.0189833548,-0.2947455049,0.1247092485,-0.1966221482,0.1081779301,-0.0275508109,0.1280115843,0.3569866419,-0.0326926894,-0.0828055292,-0.0947526321,0.5368512273,0.2782027721,0.0261523295,0.1261869222,0.1154814288,-0.3598402143,-0.3748219907,-0.0433910713,-0.2157930881,0.0899901465,0.147903055,0.2012825161,-0.1073767021,0.1348217279,-0.3785095215,0.2858775556,0.3308013678,-0.2465245724,0.0029821259,-0.3679089248,-0.413536936,-0.1898316592,0.3192932606,-0.1442516297,-0.0017331068,-0.009171092,0.1106587127,0.1891391724,0.484011054,0.1339651644,0.3138591945,0.3631411195,-0.4917096794,0.4855646789,0.0987911001,-0.1336460561,-0.0881660357,-0.1546435505,-0.3981322944,0.257417798,-0.1778470278,-0.2853306532,0.1459316611,0.1648342311,0.1642516553,0.4675336182,-0.1345406771,0.0967916846,-0.0824026763,0.3093529642,-0.0516473278,-0.268320024,0.0125274332,-0.0348947681,-0.2307369262,-0.2382221967,-0.1597029567,-0.0885297135,0.1025192365,-0.0872082114,0.0438992865,-0.129617691,0.0238304473,-0.1528648138,0.042863179,-0.0116005866,-0.3619787693,0.2263039052,0.0064995773,-0.1665717661,-0.2440984845,0.1780437678,-0.1699587405,-0.1374195665,0.2298453599,0.0661637709,0.1778306663,0.3334915042,0.0553164519,0.299461484,0.1876499057,0.2598833144,-0.2379402965,0.1655212343,-0.1539325714,0.032883428,-0.1076619104,0.449572742,0.0576862879,-0.1159918755,-0.1262978613,0.2283831835,0.2532852888,-0.0442685932,-0.2839108407,0.0855725929,0.1471211314,0.1632165909,-0.3755307794,0.2840960324,-0.2077518404,-0.1224473789,0.5265046954,0.893276751,-0.0384244286,0.1267523468,-0.145782873,-0.0221706498,0.445866406,0.1181422994,-0.1546640098,-0.2694558501,0.0372086167,0.2975227833,-0.1064928845,0.0252682399,-0.0940389931,0.142189756,0.219361648,0.0666167215,-0.2200761288,0.1170537993,0.0044210218,-0.2848397195,0.173607558,0.0940917656,0.2308221906,0.0403839238,0.1598118842,0.2757774889,-0.2304855138,-0.162351653,0.1398642212,0.0159907062,0.1191512421,0.5081248879,0.1112573445,0.4917434454,-0.2569987476,-0.2506943941,-0.3280736804,0.0469812714,-0.1642935574,0.2328413278,-0.2001020163,-0.3381249607,-0.418884784,-0.1450555176,-0.1150102839,-0.5577634573,0.1381098479,0.0413245931,-0.2752411664,0.4234624803,0.1134367064,0.253870219,-0.1176146194,0.056621965,-0.2270076871,0.0496982858,-0.4241380692,-0.0032034696,0.1883004457,-0.227151975,0.1867785752,-0.3008562922,-0.0523730293,0.1612110883,-0.7381551266,0.1576875597,0.05999602,0.0300798248,0.26732862,-0.2804103792,-0.4084673822,-0.545712173,0.1107583344,0.0272976961,-0.3636149168,0.1239908785,-0.0817677304,-0.1428301781,-0.1644627154,-0.0236302614,0.0692033842,-0.2058781534,-0.0429567695,-0.4525324106,0.2052507252,0.237983197,-0.4658526182,-0.0950666294,-0.4946010411,0.0330812261,-0.1578765512,-0.0893625394,0.2096683681,-0.2081830204,-0.0607302301,-0.2306274623,-0.3089902699,0.4237656295,0.0377434194,-0.0328559466,-0.0219792239,0.0233350601,0.0698873624,-0.1160932183,-0.0024436321,0.0631492287,0.1390253603,-0.0318449587,0.0003424485,0.0977718607,0.0718512982,-0.2497506291,0.1135581806,-0.2141737193,0.0221480038,0.2703062892,0.2845879197,0.2309903651,-0.2108982056,0.0462130792,0.047887072,0.1470340341,-0.2481189519,-0.1872301251,0.1388610601,0.1633287668,-0.0804205015,0.1776386201,0.004338989,0.0124963084,0.0787412077,0.0671452656,0.1239250526,-0.5669611692,0.1740658432,0.0063350252,0.3078863919,-0.0224398095,0.0088058393,-0.1556963027,-0.1675523371,-0.0230150037,0.0912780836,-0.0685239434,0.0833243802,-0.6093361378,-0.4087055922,-0.5790939927,0.1796693653,0.376925528,0.5158495903,0.0631969199,0.0022391395,-0.0609273612,0.1901024878,0.4397821426,-0.0025200252,-0.3540673256,0.1587360054,-0.2482479662,0.0448089205,-0.0599946789,0.1192463264,0.4059151113,0.4624784291,0.6241876483,-0.1079073399,-0.0444511659,-0.2833611071,0.0380276069,-0.0437422544,-0.0279180296,-0.2320089638,-0.1256358027,0.0214120261,0.2826708257,0.3797214329,0.4987152815,0.1252162158,-0.3248326778,-0.1013535187,-0.0680500865,-0.0375916883,0.2665781081,-0.0106772436,0.047892414,0.0593677983,0.0176351611,0.1956764758,0.4381667376,-0.2038099915,0.2262070328,-0.5027626157,-0.2139523476,0.0629585758,0.1709330976,0.2143757492,0.1790310144,0.1252254993,-0.2666707039,0.0195629317,-0.0269471928,0.5451050401,-0.1103776172,0.2474054545,-0.2115402818,0.1946561486,-0.1120188609,0.131949544,-0.1759557873,0.1986523569,0.0094296671,-0.486890316,0.2963673472,0.3950622976,0.9533655047,0.4690330923,0.2323830426,0.3443706632,-0.3253100514,0.5577519536,0.0121843414,0.2279825211,-0.4324993789,-0.1341838092,-0.0456639938,-0.0061298739,-0.103324756,-0.197679624,-0.0330658928,0.1741434038,0.2883644104,0.3233247399,0.0277676452,-0.0982890055,0.0814577416,-0.3617432415,-0.0013467713,0.0234092809,-0.0649725273,0.0034663754,-0.1622011065,0.1170456633,-0.1443162262,0.1657352448,-0.2635445595,-0.0700074658,-0.2716577351,0.0989561677,0.3264746964,-0.2881453037,0.0499258153,0.0368630439,0.6191903949,0.3315456808,0.0225817244,-0.0339668021,0.1469224095,0.2428238094,0.0445360206,-0.4531978369,0.309209764,0.0624043494,0.2238802761,-0.4029238522,0.0980227515,0.2158093899,-0.2508916855,-0.0615017228,0.0382593833,0.0031565586,0.225056991,-0.1111282036,0.2028565556,-0.2256275862,0.1336387396,-0.0337820351,-0.2314991653,-0.0089016352,-0.036962267,-0.3149179816,0.0417387225,0.2063766718,-0.0394491851,-0.5680710077,0.2585470378,-0.3827702403,-0.2513920069,-0.0074507096,0.228117317,0.1966433674,-0.0670427606,0.1369248927,-0.0981108546,-0.2944406569,-0.1854885817,0.2192816138,0.3219822347,-0.1407373101,0.0456988327,-0.3510436416,-0.1588983238,0.0010028041,-0.1786634177,0.2199769467,0.3252376914,0.5128347278,-0.3801932335,0.1378067285,-0.2897762954,-0.2866574228,-0.1854325086,-0.0114550637,-0.4478394091,-0.1218142956,-0.4638456404,-0.066103138,0.039645724,0.4677852988,-0.0939924866,-0.158350572,-0.1563608199,0.0935142785,0.3157995045,-0.1554117352,0.1816110313,0.4209155142,-0.1054447889,-0.2760101259,-0.0125931902,-0.0062853037,-0.0343766622,0.2151854783,0.2160928696,0.4562657177,0.1022046581,0.0215164255,0.1370663792,0.0469140597,0.1742347628,0.4892759919,-0.3065266013,-0.0602696314,0.0650109947,0.1609166563,0.2588519752,-0.1035243124,0.2140516788,0.1330476552,-0.3981345892,0.2124923468,0.0166543946,0.2489041984,-0.0569155812,-0.4193788767,0.4872605503,0.1549360603,-0.0717952773,0.052345138,0.395901233,-0.1198628917,-0.0587132983,0.0534306541,0.2961940169,-0.0174068511,0.2820734382,0.0970098451,0.2905797362,-0.0340777449,0.160357818,0.0292961467,0.0170205757,0.4946971238,0.4404008985,-0.1957182735,0.1561527699,0.2504727244,0.0438954979,0.3230758011,0.0308163334,0.0219732076,-0.092713438,-0.1868673414,-0.2750905156,0.1224021167,-0.0890771151,0.2107143104,0.3414237797,0.0426953658,-0.0634646267,-0.0621657297,-0.0094098868,0.1725620478,-0.6092427969,-0.1043750271,-0.3937650025,-0.0638765469,-0.313013792,-0.0878591761,0.1186138988,0.1606634408,0.0132578239,-0.2074566782,-0.3273744881,-0.0157746989,0.2165648192,0.008498163,0.2358186543,-0.0823178962,0.2951459885,0.0107118497,0.0943171307,-0.2271029204,0.1180295423,0.5410666466,0.1945247352,-0.304899931,-0.1898750812,-0.1428610384,0.0159299877,-0.2217850387,0.3723816276,-0.0461728871,-0.116817221,0.184931159,0.045645982,-0.1747493446,-0.049691733,0.2591233552,0.0956094787,0.2560723424,0.3022283614,0.0270091482,0.0460871421,0.1359401047,-0.016634753,0.104644835,0.0186180752,-0.1913211793,-0.0224693101,0.0496055521,-0.1372322142,0.0889950991,0.2402991056,0.3574056327,0.541200161,-0.0921806991,-0.2061898112,-0.4149608016,-0.3863128722,0.5211239457,0.4785057008,-0.0168483444,-0.0629614368,-0.1964923292,-0.1075174734,0.0957130119,-0.4149106443,0.164754197,0.0897945538,0.1671873629,-0.1554737836,0.2812334895,-0.0012404745,-0.4019349515,-0.1089807749,-0.0724291578,0.2989850044,0.0287750587,0.0601963028,0.2042093724,-0.0134702558,-0.1505280286,0.3201375306,-0.0521138795,0.6280628443,0.1354064494,-0.1455905885,0.4411860406,-0.0176505372,0.0286772344,-0.5843860507,0.3166669011,-0.2392611206,0.3147197962,0.0373157524,-0.4717943668,-0.0502215736,0.0252538621,0.3487576544,-0.1720810384,-0.3881042302,0.3029598296,-0.2138740271,-0.035779573,0.0506294481,0.3882575333,0.2650303841,0.4494661987,-0.2877138555,-0.6340730786,-0.0330076665,-0.2101486921,-0.2857015431,-0.1999204755,-0.1178367883,0.2962208688,-0.1532807201,-0.1381507218,-0.2883766294,0.159563005,-0.2351206839,-0.2998201549,0.2422930896,-0.1733074635,-0.0321718045,-0.1982436776,0.3979831338,0.1741544902,-0.259208262,0.1010877714,-0.1417121887]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2532","title":"Tokenizer's normalization preprocessor cause misalignment in return_offsets_mapping for tokenizer classification task","comments":"> Hi @jerryIsHere, thanks for reporting the issue. But are you sure this is a bug in HuggingFace **Datasets**?\r\n\r\nOh, I am sorry\r\nI would reopen the post on huggingface\/transformers","body":"[This colab notebook](https:\/\/colab.research.google.com\/drive\/151gKyo0YIwnlznrOHst23oYH_a3mAe3Z?usp=sharing) implements a token classification input pipeline extending the logic from [this hugging example](https:\/\/huggingface.co\/transformers\/custom_datasets.html#tok-ner).\r\n\r\nThe pipeline works fine with most instance in different languages, but unfortunately, [the Japanese Kana ligature (a form of abbreviation? I don't know Japanese well)](https:\/\/en.wikipedia.org\/wiki\/Kana_ligature) break the alignment of `return_offsets_mapping`:\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/122904371-db192700-d382-11eb-8917-1775db76db69.png)\r\n\r\nWithout the try catch block, it riase `ValueError: NumPy boolean array indexing assignment cannot assign 88 input values to the 87 output values where the mask is true`, example shown here [(another colab notebook)](https:\/\/colab.research.google.com\/drive\/1MmOqf3ppzzdKKyMWkn0bJy6DqzOO0SSm?usp=sharing)\r\n\r\nIt is clear that the normalizer is the process that break the alignment, as it is observed that `tokenizer._tokenizer.normalizer.normalize_str('\u30ff')` return '\u30b3\u30c8'.\r\n\r\nOne workaround is to include `tokenizer._tokenizer.normalizer.normalize_str` before the tokenizer preprocessing pipeline, which is also provided in the [first colab notebook](https:\/\/colab.research.google.com\/drive\/151gKyo0YIwnlznrOHst23oYH_a3mAe3Z?usp=sharing) with the name `udposTestDatasetWorkaround`.\r\n\r\nI guess similar logics should be included inside the tokenizer and the offsets_mapping generation process such that user don't need to include them in their code. But I don't understand the code of tokenizer well that I think I am not able to do this.\r\n\r\np.s.\r\n**I am using my own dataset building script in the provided example, but the script should be equivalent to the changes made by this [update](https:\/\/github.com\/huggingface\/datasets\/pull\/2466)**\r\n`get_dataset `is just a simple wrapping for `load_dataset`\r\nand the `tokenizer` is just `XLMRobertaTokenizerFast.from_pretrained(\"xlm-roberta-large\")`","comment_length":30,"text":"Tokenizer's normalization preprocessor cause misalignment in return_offsets_mapping for tokenizer classification task\n[This colab notebook](https:\/\/colab.research.google.com\/drive\/151gKyo0YIwnlznrOHst23oYH_a3mAe3Z?usp=sharing) implements a token classification input pipeline extending the logic from [this hugging example](https:\/\/huggingface.co\/transformers\/custom_datasets.html#tok-ner).\r\n\r\nThe pipeline works fine with most instance in different languages, but unfortunately, [the Japanese Kana ligature (a form of abbreviation? I don't know Japanese well)](https:\/\/en.wikipedia.org\/wiki\/Kana_ligature) break the alignment of `return_offsets_mapping`:\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/122904371-db192700-d382-11eb-8917-1775db76db69.png)\r\n\r\nWithout the try catch block, it riase `ValueError: NumPy boolean array indexing assignment cannot assign 88 input values to the 87 output values where the mask is true`, example shown here [(another colab notebook)](https:\/\/colab.research.google.com\/drive\/1MmOqf3ppzzdKKyMWkn0bJy6DqzOO0SSm?usp=sharing)\r\n\r\nIt is clear that the normalizer is the process that break the alignment, as it is observed that `tokenizer._tokenizer.normalizer.normalize_str('\u30ff')` return '\u30b3\u30c8'.\r\n\r\nOne workaround is to include `tokenizer._tokenizer.normalizer.normalize_str` before the tokenizer preprocessing pipeline, which is also provided in the [first colab notebook](https:\/\/colab.research.google.com\/drive\/151gKyo0YIwnlznrOHst23oYH_a3mAe3Z?usp=sharing) with the name `udposTestDatasetWorkaround`.\r\n\r\nI guess similar logics should be included inside the tokenizer and the offsets_mapping generation process such that user don't need to include them in their code. But I don't understand the code of tokenizer well that I think I am not able to do this.\r\n\r\np.s.\r\n**I am using my own dataset building script in the provided example, but the script should be equivalent to the changes made by this [update](https:\/\/github.com\/huggingface\/datasets\/pull\/2466)**\r\n`get_dataset `is just a simple wrapping for `load_dataset`\r\nand the `tokenizer` is just `XLMRobertaTokenizerFast.from_pretrained(\"xlm-roberta-large\")`\n> Hi @jerryIsHere, thanks for reporting the issue. But are you sure this is a bug in HuggingFace **Datasets**?\r\n\r\nOh, I am sorry\r\nI would reopen the post on huggingface\/transformers","embeddings":[-0.1536964476,0.0790127367,0.0737818331,-0.0382318683,0.0921575204,-0.3246818185,-0.0899818391,0.1060319841,-0.461761117,0.1986739933,-0.2085373253,0.307420373,0.2766766548,0.0308846552,-0.2191910297,-0.1632778943,0.1479931921,0.2190573663,0.0056235306,0.0611228123,-0.1194189712,0.4571387172,-0.212335363,0.2143603265,-0.2866258919,-0.0706363171,0.125577569,-0.315648973,-0.1683979779,-0.3988147974,0.249714613,0.118910946,-0.0836713836,0.2043143213,-0.0001126969,-0.057501819,-0.0180353783,0.0142885102,0.178675279,-0.1964323819,0.128385514,-0.3769181371,-0.3356096447,0.1073889881,-0.0636701211,0.0396103747,0.2408124506,-0.0554189608,0.6431404352,0.4151496589,0.1455552727,-0.0134888468,0.2545183599,0.0607683174,-0.3109235764,0.2689674497,-0.095067516,0.0190099981,-0.1540722102,0.031468261,-0.0746515244,-0.0096175401,-0.0922699124,-0.4802750349,-0.0148855951,0.1148962528,0.3214075863,-0.4625909626,0.1870723367,0.0185327325,-0.2498886287,-0.0450737998,-0.1830774248,-0.3914948106,-0.1368259192,-0.2679938078,0.2000967115,-0.1542632878,-0.1181238517,0.1168286502,0.0120839812,0.1378460675,0.0297428202,-0.212039277,-0.2942862511,0.564735353,-0.0654119551,0.0426447168,-0.2047559917,-0.0101762712,0.3918432295,-0.042820774,0.0059831715,0.4373289049,-0.2852554321,-0.058525499,0.1970698237,-0.4548270106,-0.0079325698,-0.0356674455,0.1242711842,0.2060957551,0.0719589293,0.0609170049,-0.1630160064,0.4642789066,0.2114510983,0.2766360343,0.1968423277,0.1170869023,-0.0398001708,-0.0404408984,0.0627822056,-0.1251495481,-0.0498765483,0.2508600056,-0.3735534847,0.1039711758,-0.2315127552,0.4121437967,-0.0718666613,0.3000814319,0.0338656791,-0.0993585065,0.0393204316,0.2893413007,-0.3403167427,-0.1019151807,-0.0826219469,0.1086550057,-0.1952837557,0.0189833548,-0.2947455049,0.1247092485,-0.1966221482,0.1081779301,-0.0275508109,0.1280115843,0.3569866419,-0.0326926894,-0.0828055292,-0.0947526321,0.5368512273,0.2782027721,0.0261523295,0.1261869222,0.1154814288,-0.3598402143,-0.3748219907,-0.0433910713,-0.2157930881,0.0899901465,0.147903055,0.2012825161,-0.1073767021,0.1348217279,-0.3785095215,0.2858775556,0.3308013678,-0.2465245724,0.0029821259,-0.3679089248,-0.413536936,-0.1898316592,0.3192932606,-0.1442516297,-0.0017331068,-0.009171092,0.1106587127,0.1891391724,0.484011054,0.1339651644,0.3138591945,0.3631411195,-0.4917096794,0.4855646789,0.0987911001,-0.1336460561,-0.0881660357,-0.1546435505,-0.3981322944,0.257417798,-0.1778470278,-0.2853306532,0.1459316611,0.1648342311,0.1642516553,0.4675336182,-0.1345406771,0.0967916846,-0.0824026763,0.3093529642,-0.0516473278,-0.268320024,0.0125274332,-0.0348947681,-0.2307369262,-0.2382221967,-0.1597029567,-0.0885297135,0.1025192365,-0.0872082114,0.0438992865,-0.129617691,0.0238304473,-0.1528648138,0.042863179,-0.0116005866,-0.3619787693,0.2263039052,0.0064995773,-0.1665717661,-0.2440984845,0.1780437678,-0.1699587405,-0.1374195665,0.2298453599,0.0661637709,0.1778306663,0.3334915042,0.0553164519,0.299461484,0.1876499057,0.2598833144,-0.2379402965,0.1655212343,-0.1539325714,0.032883428,-0.1076619104,0.449572742,0.0576862879,-0.1159918755,-0.1262978613,0.2283831835,0.2532852888,-0.0442685932,-0.2839108407,0.0855725929,0.1471211314,0.1632165909,-0.3755307794,0.2840960324,-0.2077518404,-0.1224473789,0.5265046954,0.893276751,-0.0384244286,0.1267523468,-0.145782873,-0.0221706498,0.445866406,0.1181422994,-0.1546640098,-0.2694558501,0.0372086167,0.2975227833,-0.1064928845,0.0252682399,-0.0940389931,0.142189756,0.219361648,0.0666167215,-0.2200761288,0.1170537993,0.0044210218,-0.2848397195,0.173607558,0.0940917656,0.2308221906,0.0403839238,0.1598118842,0.2757774889,-0.2304855138,-0.162351653,0.1398642212,0.0159907062,0.1191512421,0.5081248879,0.1112573445,0.4917434454,-0.2569987476,-0.2506943941,-0.3280736804,0.0469812714,-0.1642935574,0.2328413278,-0.2001020163,-0.3381249607,-0.418884784,-0.1450555176,-0.1150102839,-0.5577634573,0.1381098479,0.0413245931,-0.2752411664,0.4234624803,0.1134367064,0.253870219,-0.1176146194,0.056621965,-0.2270076871,0.0496982858,-0.4241380692,-0.0032034696,0.1883004457,-0.227151975,0.1867785752,-0.3008562922,-0.0523730293,0.1612110883,-0.7381551266,0.1576875597,0.05999602,0.0300798248,0.26732862,-0.2804103792,-0.4084673822,-0.545712173,0.1107583344,0.0272976961,-0.3636149168,0.1239908785,-0.0817677304,-0.1428301781,-0.1644627154,-0.0236302614,0.0692033842,-0.2058781534,-0.0429567695,-0.4525324106,0.2052507252,0.237983197,-0.4658526182,-0.0950666294,-0.4946010411,0.0330812261,-0.1578765512,-0.0893625394,0.2096683681,-0.2081830204,-0.0607302301,-0.2306274623,-0.3089902699,0.4237656295,0.0377434194,-0.0328559466,-0.0219792239,0.0233350601,0.0698873624,-0.1160932183,-0.0024436321,0.0631492287,0.1390253603,-0.0318449587,0.0003424485,0.0977718607,0.0718512982,-0.2497506291,0.1135581806,-0.2141737193,0.0221480038,0.2703062892,0.2845879197,0.2309903651,-0.2108982056,0.0462130792,0.047887072,0.1470340341,-0.2481189519,-0.1872301251,0.1388610601,0.1633287668,-0.0804205015,0.1776386201,0.004338989,0.0124963084,0.0787412077,0.0671452656,0.1239250526,-0.5669611692,0.1740658432,0.0063350252,0.3078863919,-0.0224398095,0.0088058393,-0.1556963027,-0.1675523371,-0.0230150037,0.0912780836,-0.0685239434,0.0833243802,-0.6093361378,-0.4087055922,-0.5790939927,0.1796693653,0.376925528,0.5158495903,0.0631969199,0.0022391395,-0.0609273612,0.1901024878,0.4397821426,-0.0025200252,-0.3540673256,0.1587360054,-0.2482479662,0.0448089205,-0.0599946789,0.1192463264,0.4059151113,0.4624784291,0.6241876483,-0.1079073399,-0.0444511659,-0.2833611071,0.0380276069,-0.0437422544,-0.0279180296,-0.2320089638,-0.1256358027,0.0214120261,0.2826708257,0.3797214329,0.4987152815,0.1252162158,-0.3248326778,-0.1013535187,-0.0680500865,-0.0375916883,0.2665781081,-0.0106772436,0.047892414,0.0593677983,0.0176351611,0.1956764758,0.4381667376,-0.2038099915,0.2262070328,-0.5027626157,-0.2139523476,0.0629585758,0.1709330976,0.2143757492,0.1790310144,0.1252254993,-0.2666707039,0.0195629317,-0.0269471928,0.5451050401,-0.1103776172,0.2474054545,-0.2115402818,0.1946561486,-0.1120188609,0.131949544,-0.1759557873,0.1986523569,0.0094296671,-0.486890316,0.2963673472,0.3950622976,0.9533655047,0.4690330923,0.2323830426,0.3443706632,-0.3253100514,0.5577519536,0.0121843414,0.2279825211,-0.4324993789,-0.1341838092,-0.0456639938,-0.0061298739,-0.103324756,-0.197679624,-0.0330658928,0.1741434038,0.2883644104,0.3233247399,0.0277676452,-0.0982890055,0.0814577416,-0.3617432415,-0.0013467713,0.0234092809,-0.0649725273,0.0034663754,-0.1622011065,0.1170456633,-0.1443162262,0.1657352448,-0.2635445595,-0.0700074658,-0.2716577351,0.0989561677,0.3264746964,-0.2881453037,0.0499258153,0.0368630439,0.6191903949,0.3315456808,0.0225817244,-0.0339668021,0.1469224095,0.2428238094,0.0445360206,-0.4531978369,0.309209764,0.0624043494,0.2238802761,-0.4029238522,0.0980227515,0.2158093899,-0.2508916855,-0.0615017228,0.0382593833,0.0031565586,0.225056991,-0.1111282036,0.2028565556,-0.2256275862,0.1336387396,-0.0337820351,-0.2314991653,-0.0089016352,-0.036962267,-0.3149179816,0.0417387225,0.2063766718,-0.0394491851,-0.5680710077,0.2585470378,-0.3827702403,-0.2513920069,-0.0074507096,0.228117317,0.1966433674,-0.0670427606,0.1369248927,-0.0981108546,-0.2944406569,-0.1854885817,0.2192816138,0.3219822347,-0.1407373101,0.0456988327,-0.3510436416,-0.1588983238,0.0010028041,-0.1786634177,0.2199769467,0.3252376914,0.5128347278,-0.3801932335,0.1378067285,-0.2897762954,-0.2866574228,-0.1854325086,-0.0114550637,-0.4478394091,-0.1218142956,-0.4638456404,-0.066103138,0.039645724,0.4677852988,-0.0939924866,-0.158350572,-0.1563608199,0.0935142785,0.3157995045,-0.1554117352,0.1816110313,0.4209155142,-0.1054447889,-0.2760101259,-0.0125931902,-0.0062853037,-0.0343766622,0.2151854783,0.2160928696,0.4562657177,0.1022046581,0.0215164255,0.1370663792,0.0469140597,0.1742347628,0.4892759919,-0.3065266013,-0.0602696314,0.0650109947,0.1609166563,0.2588519752,-0.1035243124,0.2140516788,0.1330476552,-0.3981345892,0.2124923468,0.0166543946,0.2489041984,-0.0569155812,-0.4193788767,0.4872605503,0.1549360603,-0.0717952773,0.052345138,0.395901233,-0.1198628917,-0.0587132983,0.0534306541,0.2961940169,-0.0174068511,0.2820734382,0.0970098451,0.2905797362,-0.0340777449,0.160357818,0.0292961467,0.0170205757,0.4946971238,0.4404008985,-0.1957182735,0.1561527699,0.2504727244,0.0438954979,0.3230758011,0.0308163334,0.0219732076,-0.092713438,-0.1868673414,-0.2750905156,0.1224021167,-0.0890771151,0.2107143104,0.3414237797,0.0426953658,-0.0634646267,-0.0621657297,-0.0094098868,0.1725620478,-0.6092427969,-0.1043750271,-0.3937650025,-0.0638765469,-0.313013792,-0.0878591761,0.1186138988,0.1606634408,0.0132578239,-0.2074566782,-0.3273744881,-0.0157746989,0.2165648192,0.008498163,0.2358186543,-0.0823178962,0.2951459885,0.0107118497,0.0943171307,-0.2271029204,0.1180295423,0.5410666466,0.1945247352,-0.304899931,-0.1898750812,-0.1428610384,0.0159299877,-0.2217850387,0.3723816276,-0.0461728871,-0.116817221,0.184931159,0.045645982,-0.1747493446,-0.049691733,0.2591233552,0.0956094787,0.2560723424,0.3022283614,0.0270091482,0.0460871421,0.1359401047,-0.016634753,0.104644835,0.0186180752,-0.1913211793,-0.0224693101,0.0496055521,-0.1372322142,0.0889950991,0.2402991056,0.3574056327,0.541200161,-0.0921806991,-0.2061898112,-0.4149608016,-0.3863128722,0.5211239457,0.4785057008,-0.0168483444,-0.0629614368,-0.1964923292,-0.1075174734,0.0957130119,-0.4149106443,0.164754197,0.0897945538,0.1671873629,-0.1554737836,0.2812334895,-0.0012404745,-0.4019349515,-0.1089807749,-0.0724291578,0.2989850044,0.0287750587,0.0601963028,0.2042093724,-0.0134702558,-0.1505280286,0.3201375306,-0.0521138795,0.6280628443,0.1354064494,-0.1455905885,0.4411860406,-0.0176505372,0.0286772344,-0.5843860507,0.3166669011,-0.2392611206,0.3147197962,0.0373157524,-0.4717943668,-0.0502215736,0.0252538621,0.3487576544,-0.1720810384,-0.3881042302,0.3029598296,-0.2138740271,-0.035779573,0.0506294481,0.3882575333,0.2650303841,0.4494661987,-0.2877138555,-0.6340730786,-0.0330076665,-0.2101486921,-0.2857015431,-0.1999204755,-0.1178367883,0.2962208688,-0.1532807201,-0.1381507218,-0.2883766294,0.159563005,-0.2351206839,-0.2998201549,0.2422930896,-0.1733074635,-0.0321718045,-0.1982436776,0.3979831338,0.1741544902,-0.259208262,0.1010877714,-0.1417121887]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2528","title":"Logging cannot be set to NOTSET similar to transformers","comments":"Hi @joshzwiebel, thanks for reporting. We are going to align with `transformers`.","body":"## Describe the bug\r\nIn the transformers library you can set the verbosity level to logging.NOTSET to work around the usage of tqdm and IPywidgets, however in Datasets this is no longer possible. This is because transformers set the verbosity level of tqdm with [this](https:\/\/github.com\/huggingface\/transformers\/blob\/b53bc55ba9bb10d5ee279eab51a2f0acc5af2a6b\/src\/transformers\/file_utils.py#L1449) \r\n`disable=bool(logging.get_verbosity() == logging.NOTSET)`\r\nand datasets accomplishes this like [so](https:\/\/github.com\/huggingface\/datasets\/blob\/83554e410e1ab8c6f705cfbb2df7953638ad3ac1\/src\/datasets\/utils\/file_utils.py#L493)\r\n`not_verbose = bool(logger.getEffectiveLevel() > WARNING)`\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\nimport logging\r\ndatasets.logging.get_verbosity = lambda : logging.NOTSET\r\ndatasets.load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\")\r\n```\r\n\r\n## Expected results\r\nThe code should download and load the dataset as normal without displaying progress bars\r\n\r\n## Actual results\r\n```ImportError Traceback (most recent call last)\r\n in \r\n----> 1 datasets.load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\")\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, **config_kwargs)\r\n 713 dataset=True,\r\n 714 return_resolved_file_path=True,\r\n--> 715 use_auth_token=use_auth_token,\r\n 716 )\r\n 717 # Set the base path for downloads as the parent of the script location\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, dynamic_modules_path, return_resolved_file_path, **download_kwargs)\r\n 350 file_path = hf_bucket_url(path, filename=name, dataset=False)\r\n 351 try:\r\n--> 352 local_path = cached_path(file_path, download_config=download_config)\r\n 353 except FileNotFoundError:\r\n 354 raise FileNotFoundError(\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 289 use_etag=download_config.use_etag,\r\n 290 max_retries=download_config.max_retries,\r\n--> 291 use_auth_token=download_config.use_auth_token,\r\n 292 )\r\n 293 elif os.path.exists(url_or_filename):\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 668 headers=headers,\r\n 669 cookies=cookies,\r\n--> 670 max_retries=max_retries,\r\n 671 )\r\n 672 \r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py in http_get(url, temp_file, proxies, resume_size, headers, cookies, timeout, max_retries)\r\n 493 initial=resume_size,\r\n 494 desc=\"Downloading\",\r\n--> 495 disable=not_verbose,\r\n 496 )\r\n 497 for chunk in response.iter_content(chunk_size=1024):\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __init__(self, *args, **kwargs)\r\n 217 total = self.total * unit_scale if self.total else self.total\r\n 218 self.container = self.status_printer(\r\n--> 219 self.fp, total, self.desc, self.ncols)\r\n 220 self.sp = self.display\r\n 221 \r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in status_printer(_, total, desc, ncols)\r\n 95 if IProgress is None: # #187 #451 #558 #872\r\n 96 raise ImportError(\r\n---> 97 \"IProgress not found. Please update jupyter and ipywidgets.\"\r\n 98 \" See https:\/\/ipywidgets.readthedocs.io\/en\/stable\"\r\n 99 \"\/user_install.html\")\r\n\r\nImportError: IProgress not found. Please update jupyter and ipywidgets. See https:\/\/ipywidgets.readthedocs.io\/en\/stable\/user_install.html\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.95-42.163.amzn2.x86_64-x86_64-with-debian-10.8\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\nI am running this code on Deepnote and which important to this issue **does not** support IPywidgets\r\n\r\n","comment_length":12,"text":"Logging cannot be set to NOTSET similar to transformers\n## Describe the bug\r\nIn the transformers library you can set the verbosity level to logging.NOTSET to work around the usage of tqdm and IPywidgets, however in Datasets this is no longer possible. This is because transformers set the verbosity level of tqdm with [this](https:\/\/github.com\/huggingface\/transformers\/blob\/b53bc55ba9bb10d5ee279eab51a2f0acc5af2a6b\/src\/transformers\/file_utils.py#L1449) \r\n`disable=bool(logging.get_verbosity() == logging.NOTSET)`\r\nand datasets accomplishes this like [so](https:\/\/github.com\/huggingface\/datasets\/blob\/83554e410e1ab8c6f705cfbb2df7953638ad3ac1\/src\/datasets\/utils\/file_utils.py#L493)\r\n`not_verbose = bool(logger.getEffectiveLevel() > WARNING)`\r\n## Steps to reproduce the bug\r\n```python\r\nimport datasets\r\nimport logging\r\ndatasets.logging.get_verbosity = lambda : logging.NOTSET\r\ndatasets.load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\")\r\n```\r\n\r\n## Expected results\r\nThe code should download and load the dataset as normal without displaying progress bars\r\n\r\n## Actual results\r\n```ImportError Traceback (most recent call last)\r\n in \r\n----> 1 datasets.load_dataset(\"patrickvonplaten\/librispeech_asr_dummy\")\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, **config_kwargs)\r\n 713 dataset=True,\r\n 714 return_resolved_file_path=True,\r\n--> 715 use_auth_token=use_auth_token,\r\n 716 )\r\n 717 # Set the base path for downloads as the parent of the script location\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, dynamic_modules_path, return_resolved_file_path, **download_kwargs)\r\n 350 file_path = hf_bucket_url(path, filename=name, dataset=False)\r\n 351 try:\r\n--> 352 local_path = cached_path(file_path, download_config=download_config)\r\n 353 except FileNotFoundError:\r\n 354 raise FileNotFoundError(\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 289 use_etag=download_config.use_etag,\r\n 290 max_retries=download_config.max_retries,\r\n--> 291 use_auth_token=download_config.use_auth_token,\r\n 292 )\r\n 293 elif os.path.exists(url_or_filename):\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 668 headers=headers,\r\n 669 cookies=cookies,\r\n--> 670 max_retries=max_retries,\r\n 671 )\r\n 672 \r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py in http_get(url, temp_file, proxies, resume_size, headers, cookies, timeout, max_retries)\r\n 493 initial=resume_size,\r\n 494 desc=\"Downloading\",\r\n--> 495 disable=not_verbose,\r\n 496 )\r\n 497 for chunk in response.iter_content(chunk_size=1024):\r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __init__(self, *args, **kwargs)\r\n 217 total = self.total * unit_scale if self.total else self.total\r\n 218 self.container = self.status_printer(\r\n--> 219 self.fp, total, self.desc, self.ncols)\r\n 220 self.sp = self.display\r\n 221 \r\n\r\n~\/venv\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in status_printer(_, total, desc, ncols)\r\n 95 if IProgress is None: # #187 #451 #558 #872\r\n 96 raise ImportError(\r\n---> 97 \"IProgress not found. Please update jupyter and ipywidgets.\"\r\n 98 \" See https:\/\/ipywidgets.readthedocs.io\/en\/stable\"\r\n 99 \"\/user_install.html\")\r\n\r\nImportError: IProgress not found. Please update jupyter and ipywidgets. See https:\/\/ipywidgets.readthedocs.io\/en\/stable\/user_install.html\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: 1.8.0\r\n- Platform: Linux-5.4.95-42.163.amzn2.x86_64-x86_64-with-debian-10.8\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\nI am running this code on Deepnote and which important to this issue **does not** support IPywidgets\r\n\r\n\nHi @joshzwiebel, thanks for reporting. We are going to align with `transformers`.","embeddings":[-0.48722139,-0.240187183,0.0768826827,-0.0190668833,0.3719440401,-0.0689783096,0.6162853241,0.2110450268,-0.0893884003,-0.0378360525,-0.2720355392,0.2572050691,-0.3901108801,0.0078309849,-0.1585512459,0.0780935585,-0.1523604393,0.1084450334,-0.6500580311,0.0853273124,-0.0198846161,0.0257744342,-0.3033263087,-0.0028240087,-0.4629936218,-0.1876343489,0.2041845024,0.1107775867,0.0303660706,-0.4412383139,0.1379784942,-0.0555309094,0.184201926,0.5385534763,-0.0001098855,0.0920989886,0.4299394786,0.0696690455,-0.5418926477,-0.0772583038,-0.3333098292,-0.3556215465,0.2541621327,-0.296559155,0.0278052147,-0.0886458457,0.1591480225,-0.2350447327,0.4257537127,0.372875303,0.2581309378,0.6359708905,-0.2188830525,0.021910198,0.0125624612,0.3479246795,-0.3160419464,-0.128743276,0.0895232931,0.1278259605,-0.4082004428,0.2119069099,-0.2304086089,0.0651286915,0.1292444766,0.0887975842,0.2447046638,-0.2036676854,0.0903130099,0.3765877485,0.4439706206,-0.6329837441,-0.3063752651,-0.3242901266,-0.2254669219,-0.1921640486,0.2203552127,-0.0364702046,-0.2598054409,0.2695802152,-0.5321219563,-0.0685853884,-0.1993096769,0.1433254778,-0.0073001394,0.0240585506,-0.1673024446,0.1210326403,-0.3479986191,-0.095880039,0.1577248275,-0.0925209373,0.1008363143,0.1044362411,-0.418222487,-0.1542779356,0.2952682078,0.030275356,0.0130817071,0.321611166,-0.2832523584,-0.2092254609,0.0787242129,0.0354567766,0.2722370327,0.2543895543,0.3063734174,0.4471981227,0.3183549345,0.1133371666,0.0549423583,0.029009698,0.2809048295,-0.129689455,0.5406084061,-0.005186283,0.573125124,-0.0346292928,-0.2512248456,-0.1187204048,0.059977334,0.0532799736,0.0774091482,0.2407835275,0.0137032624,-0.2910085618,-0.0000483196,0.355774045,-0.255962193,0.0520179458,-0.1807847768,0.0351750515,-0.1401919127,0.0768361241,0.3424854577,0.0306794029,0.3125410378,0.1650882661,-0.0716217235,0.2084063441,0.2252460122,-0.1143143922,0.3137560785,0.4384763837,-0.0455839969,0.3192904592,0.0556585304,-0.2542351484,-0.1002488807,0.1750276685,0.1544253528,-0.1277620643,-0.0836197436,0.1987078339,-0.5494337082,-0.1896209717,-0.0438366979,0.1200122088,0.1445987076,-0.0902145952,0.2130984366,-0.1433763355,-0.109292008,-0.1688942164,0.2344229072,0.7190428376,-0.243981272,-0.3475469351,-0.3296743929,-0.3661786914,0.244419381,0.1449767649,-0.0758779496,0.2271275073,-0.2162891775,-0.0021065383,0.1575034559,-0.36010167,-0.3241606653,0.3108734787,-0.0502564572,0.3138056695,0.2139761001,0.1352796704,0.2697777152,-0.1635340452,-0.0747057498,0.0520096347,0.13406232,-0.0303227026,-0.1151241809,-0.1871121079,0.0866843611,0.1621356457,0.0257202443,0.1686272174,0.1645424962,0.1709782034,0.1728233993,-0.0644406229,0.148654297,0.320200026,0.0995772034,-0.3636652231,-0.1847637594,-0.3671681583,-0.5082033277,0.0671478584,-0.0881963745,0.1716508865,-0.2907072604,-0.1213751286,-0.2456423342,-0.0995707065,-0.1333458275,-0.2933743596,0.1341315657,0.0189312845,-0.1710985452,-0.0323217176,-0.2549135685,0.355848372,0.0253766999,0.0542019978,-0.1710238904,0.1148362681,0.1140511557,-0.3562551141,0.0736010894,0.2154676318,0.0048297006,0.1330100745,0.0074865236,0.373195529,-0.0446934253,0.3865067363,0.0793055743,0.1915173382,0.1856602877,-0.2384110689,0.0557563566,0.3371387422,-0.118885614,0.1371994466,-0.4931004643,0.1670438051,0.0684681609,0.0979547873,0.1483547837,0.156210497,0.1312961578,-0.2237747163,-0.3916816413,-0.3026470542,0.3827649951,0.0596654117,0.201207757,-0.0639102384,-0.0629701242,-0.0475946963,0.573082149,0.0823858008,0.305786103,-0.0009629583,-0.5265111327,0.2175006568,0.2340147048,-0.0130907781,0.4029511809,0.1702344567,0.1155779883,0.2413203418,-0.0747502893,0.0272363555,0.2088852376,0.1083665788,-0.1483891308,0.0344677754,0.1562070549,-0.0643635765,-0.2748487592,-0.1857725084,-0.1350777447,0.1450893432,-0.4537131786,0.0024304045,-0.206558153,-0.0829233676,-0.3186276555,-0.0359730832,-0.2034393698,-0.2219528705,-0.0019382413,0.1512763798,-0.1470299661,0.432163924,0.0355893336,-0.157117933,0.2716439068,-0.3008927405,-0.344376564,0.0610840209,-0.0656274408,0.0837574676,0.022611266,-0.1279215664,0.3812164366,-0.0029090384,0.0579035878,-0.1854818463,-0.2910711467,0.1267575771,-0.0193443112,0.2840059102,0.245140627,0.1669774652,0.0852068439,-0.0948919281,0.0339614302,-0.1952685863,-0.010402007,0.164924264,-0.1506554037,0.0272845831,-0.0846583545,-0.3471570015,-0.1756225526,-0.3608453572,0.023314286,0.0519495606,0.0652284771,0.6522889137,0.0546899587,0.152633965,-0.155239284,0.1623249054,0.1587495506,-0.2200263292,0.3970078826,-0.1262869239,-0.1168752015,0.1079385355,0.0076374975,-0.0754245669,0.0479101278,-0.3254781961,0.039210625,-0.3632914424,0.2098378986,0.0477120131,0.161643371,0.2856974602,0.2269754857,-0.0661601052,-0.0690331608,-0.1710989624,-0.027497109,-0.067598708,0.2362048626,-0.2275962681,0.1711643487,0.096263051,0.57407552,0.3996042311,0.0882270634,0.0252389926,-0.2916404307,0.42213431,-0.1050845087,-0.3565305471,-0.1241282448,0.0912003815,-0.0958146527,-0.0111777382,0.0323509574,0.1338358223,-0.1120585203,0.0973092094,0.0274049416,-0.3342528641,-0.0247666631,-0.093518056,0.2778598964,0.0621307082,0.3695232272,-0.1322194785,-0.1924636811,0.0833914652,0.1981212199,0.0102649685,-0.0282053705,-0.0060186242,0.2521658242,-0.6046960354,0.1945882589,-0.0293710995,0.1854954064,-0.2051309645,-0.2919847369,0.0652443394,0.3470413685,0.0985881239,-0.252782613,0.098901324,0.1876722723,0.0175636671,-0.6050220132,0.0822365433,-0.2978385985,0.3377988636,0.4880318046,0.3964844644,-0.3030754328,-0.1209224164,0.1688317209,0.1408665925,-0.0727965385,-0.3714132011,-0.4394181073,-0.0291930921,-0.3428992331,-0.0186857767,0.029614741,-0.0354612023,-0.2152525485,-0.1242232844,-0.2154833525,-0.072670728,0.104678832,0.0609032139,0.2305603325,-0.265542984,0.2550721169,0.1290348917,0.149766311,-0.2374249548,0.3992499411,0.09712594,-0.3989897668,-0.0389684699,-0.0760712549,0.0604149625,0.0846623704,-0.1794312745,0.0422360152,0.2374094129,0.1962188482,-0.2025004476,0.0807808563,0.3158175647,0.0821126699,-0.276915282,0.1466727257,0.4003627002,-0.0267881993,0.1524818689,0.3516911566,0.0314704217,-0.2065581977,-0.0592587143,-0.1465022415,0.6821908355,-0.0181481689,0.1703738421,0.3439227343,-0.2364906818,0.4021262228,0.2109418958,0.1257993132,-0.3333476186,-0.2734325528,0.0155699709,-0.2966027856,0.26998806,-0.0059817848,-0.3544386029,0.1488017291,-0.314224124,0.1867583394,-0.0721871406,0.2476145327,-0.1494712532,-0.2213288397,-0.5328850746,0.131569922,-0.0879109129,0.2167019993,0.005809735,-0.1128674895,0.209017247,0.0035684553,-0.1315754503,0.019628996,0.1895251572,-0.0643389747,0.2824927866,-0.2786331177,0.1782898307,-0.0351063497,0.0990842357,-0.0034139911,-0.2733866274,0.1714365929,-0.2115671039,0.3466464281,0.0896698833,0.0471629351,0.4379053712,0.0321087278,-0.148896113,0.4171281755,-0.0920159146,-0.3290343285,0.2491821051,-0.0307877939,0.218544066,-0.4301470816,0.350771606,0.004544178,0.0118388208,-0.0536313578,0.1844478995,0.2294661701,-0.0672131404,0.420897752,-0.2479861677,-0.3017261624,-0.1354811341,0.5306575298,-0.1053472385,0.0867123008,0.5246639252,0.2055189461,-0.3234333992,-0.2011132687,0.1304932386,0.0842084065,-0.4831882119,0.3385468721,0.0463583581,0.2627357244,0.0283657834,0.2866389453,-0.1478771716,-0.057561215,-0.3928842247,-0.1339157075,-0.6094391346,0.0289315172,-0.0423483662,0.1427823305,0.2057613432,0.2532321811,0.1895774156,-0.1641493738,-0.3571145833,-0.3993304968,-0.2616400421,0.0899529904,0.4146482348,-0.1945373267,0.3370793164,-0.0747870728,0.1288067997,-0.1449508816,-0.1257777959,-0.2656235695,-0.2750573456,0.0721522719,-0.0434004851,-0.1954741031,-0.1029808894,-0.0855740756,-0.0431217216,-0.1293317527,0.2566430569,0.4241147935,-0.0640447885,0.2786155343,-0.3341517746,-0.0541046336,-0.2158391178,-0.3053961098,0.0076708584,-0.1346326619,-0.0574209802,0.2070062757,-0.0924588665,0.0851952806,-0.1405638754,0.0606865138,0.2743837535,0.2509633601,0.2010960728,-0.0831329748,-0.1040941551,-0.0004315999,0.3012122214,0.0993250608,-0.16132541,-0.2725438476,-0.0074287569,0.2430133671,-0.160348177,-0.0131043829,-0.0185005181,0.1120187715,0.2352479845,0.1554494053,-0.3892651498,-0.0269077681,0.5375592709,0.2314753383,0.2002269775,-0.533244133,0.0138194729,0.2498283237,-0.1688591838,-0.0318659917,0.4061585069,0.1052031592,0.2204604447,0.4712255597,-0.0627342835,0.1639965326,-0.1396819949,-0.046683684,-0.0009461538,-0.1465410739,-0.1364584267,-0.2323989272,-0.0383215137,0.0168989487,-0.0151445149,-0.0517628305,-0.0797087848,0.284666419,-0.2431593388,0.1033880562,-0.1597328782,0.0196502395,0.1983198524,0.0152161,0.0979558527,-0.0983752608,-0.0566072874,-0.0161893964,0.3249462843,0.0806827396,0.0266489964,-0.4479720592,0.2656417191,0.191455856,0.1587281078,-0.1424429268,-0.0035640462,0.1979357302,0.149497658,-0.0603529736,0.4500035644,0.4902292192,0.1839171648,-0.0100654056,0.1493505687,0.0449565277,-0.1507998407,-0.0899263173,0.3492850661,-0.2843281329,0.1073567495,0.3259610832,0.165894419,-0.2149915099,0.0914504677,-0.1581545472,-0.2505217791,-0.4009094834,0.6213450432,-0.1550001651,-0.1992400438,-0.0726788715,-0.1703714132,-0.1989317089,0.0118878745,0.5169018507,-0.0658379048,0.1085629687,-0.2609561682,0.0991450772,-0.2602249682,0.2161318362,0.3265387416,0.2682598531,-0.3449492753,0.2543475032,-0.3960289061,0.1736629754,0.1175774038,-0.3632749319,-0.0816447213,0.2282201052,-0.3809297383,0.2601781487,0.2782830894,-0.1762162298,-0.0098027382,-0.0486547314,0.0671043396,-0.0640250668,-0.0032395737,-0.1456312835,-0.1002519876,-0.4419339001,0.296289742,-0.0594607517,0.1269200891,-0.0941189602,-0.1949779689,0.2584537566,0.2016349584,0.1129727066,0.12594603,0.1976179034,-0.0869980752,-0.0936471671,-0.2632791698,-0.1835672706,-0.2624524236,-0.0923092142,0.059122812,0.4232577682,-0.1407154649,-0.0884753838,-0.269553721,0.298769474,0.1105840802,-0.1251612604,0.1304879934,0.1488842517,-0.1423884481,0.1790450215,0.0120669128,0.1881940216,0.0270212628,0.3132683039,-0.3097270131,-0.5265215635,0.3525000215,-0.5566977859,-0.4416193664,-0.080479838,0.4142380357,0.1397955865,-0.1802685559,-0.4359262884,0.000943523,0.4036533237,-0.2260699719,-0.0538289137,0.3003604412,-0.0175400935,0.2804589272,0.1048833057,0.3563004136,0.2252512425,-0.0010577575,-0.0057554771,-0.1745932847]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2522","title":"Documentation Mistakes in Dataset: emotion","comments":"Hi,\r\n\r\nthis issue has been already reported in the dataset repo (https:\/\/github.com\/dair-ai\/emotion_dataset\/issues\/2), so this is a bug on their side.","body":"As per documentation,\r\nDataset: emotion\r\nHomepage: https:\/\/github.com\/dair-ai\/emotion_dataset\r\n\r\nDataset: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/emotion\/emotion.py\r\n\r\nPermalink: https:\/\/huggingface.co\/datasets\/viewer\/?dataset=emotion\r\n\r\nEmotion is a dataset of English Twitter messages with eight basic emotions: anger, anticipation, disgust, fear, joy, sadness, surprise, and trust. For more detailed information please refer to the paper.\r\n\r\nBut when we view the data, there are only 6 emotions, anger, fear, joy, sadness, surprise, and trust.","comment_length":20,"text":"Documentation Mistakes in Dataset: emotion\nAs per documentation,\r\nDataset: emotion\r\nHomepage: https:\/\/github.com\/dair-ai\/emotion_dataset\r\n\r\nDataset: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/emotion\/emotion.py\r\n\r\nPermalink: https:\/\/huggingface.co\/datasets\/viewer\/?dataset=emotion\r\n\r\nEmotion is a dataset of English Twitter messages with eight basic emotions: anger, anticipation, disgust, fear, joy, sadness, surprise, and trust. For more detailed information please refer to the paper.\r\n\r\nBut when we view the data, there are only 6 emotions, anger, fear, joy, sadness, surprise, and trust.\nHi,\r\n\r\nthis issue has been already reported in the dataset repo (https:\/\/github.com\/dair-ai\/emotion_dataset\/issues\/2), so this is a bug on their side.","embeddings":[0.2212711871,-0.3492065668,-0.1087545753,0.5762394071,0.2677908242,0.1737760007,0.2942093909,0.1110398769,-0.1283250451,0.203823626,-0.1274428964,0.0429529399,-0.2023335248,-0.1131508723,-0.0055952733,-0.186274454,0.0835631117,-0.0872550756,0.3198854923,-0.1177928001,0.0335479043,0.1983209848,-0.0980703384,0.2302875668,-0.4554688334,-0.0563974157,-0.2686941922,-0.0515943803,-0.0469822958,-0.5062128901,0.0716762841,-0.0896961838,-0.0074797897,0.2895738184,-0.0001154651,-0.111664705,0.3260878623,0.0778146684,0.0454540886,-0.231089592,0.2143916786,-0.3071580231,0.3784171939,-0.0876959562,-0.0448352508,-0.5062558055,0.059291508,-0.0312779248,0.4184461236,0.4927144647,0.1667767763,0.178087756,0.2175596058,0.2025782615,0.4194192886,0.545627892,-0.0178259239,0.0204210337,-0.1378220469,-0.0215042494,-0.1851763129,0.3451380432,0.165079847,0.0712513551,0.1775846481,-0.091846481,0.199638024,-0.579117775,0.1523029506,0.1900897622,0.4783172607,-0.0171625223,-0.3321194351,-0.1146266535,0.0158168022,-0.1334289908,0.1938817799,0.2820130289,-0.1139772758,0.09297169,0.102289319,0.0965632573,-0.1609798819,0.4246109724,0.0660467595,0.1322462559,-0.3756153882,0.0194855109,0.0079664448,-0.2089993507,0.0384880938,0.143693611,-0.0376610756,0.2057614326,-0.1708817631,0.2282327861,-0.0167870335,0.0003168922,-0.0389231481,-0.2421786934,-0.1124494523,0.1332705915,-0.1147631854,-0.0613619089,0.6197059751,0.0495514646,0.1868034154,0.1237083226,0.293532759,0.090946354,-0.0030079882,-0.0017007591,-0.0367357954,0.0221608207,0.0139622083,-0.4305975437,-0.0058878432,0.0918471292,-0.3832752705,-0.0522363745,-0.2342670411,-0.3501877487,-0.0679655895,0.3387868404,0.0737476423,0.300462991,-0.0802049711,0.1375329792,0.017084064,-0.541508615,-0.1882719845,-0.0249801911,-0.2958142161,-0.0081739416,0.1116004735,-0.0923953354,0.6315552592,0.1559411436,-0.2107320428,-0.3750531375,0.088589035,-0.1992551237,0.3007443845,0.1824375391,-0.0359490551,0.2212837785,0.2664774358,-0.0924429819,0.0486649647,0.0915699974,-0.0351156369,-0.0881066769,-0.4956248403,0.0979886055,-0.2035675198,0.0199237596,-0.0390605964,0.6006979346,0.0803495347,0.1519803256,0.223384887,-0.0730353668,0.0122765219,-0.2965444326,0.3505527377,0.3677574098,-0.3522181213,-0.3982218504,-0.3678133786,0.0090231774,0.1439783573,0.4062836766,-0.3304085135,0.1271507442,-0.3378306329,0.2549834847,0.2418566793,-0.1866464168,-0.2345921695,0.112793766,0.357495904,0.0649622232,-0.0762086213,0.2588240504,0.1550677121,0.1859456897,0.0829654858,-0.2622185051,0.3079393506,-0.0407892577,-0.3400007188,-0.197105661,0.1795888096,0.0728721097,-0.0786690786,-0.0564080104,0.2922448516,-0.2469891012,0.6810661554,-0.0197623167,-0.0141569274,0.2730138302,0.1785178334,0.4840732813,0.1251228899,-0.0556451157,-0.1300527006,-0.1114680693,0.1443299949,0.4373609722,0.2046014816,-0.0555772036,-0.2629448473,-0.0983478725,-0.1430063695,-0.4279916584,0.1366409212,0.2002825737,-0.3659696579,0.2478677183,-0.3380771875,0.2200650275,-0.2423121184,0.0935569555,-0.2778650522,0.2316070795,-0.0960300267,0.017993493,0.1508451104,0.0809294209,0.0463765413,0.0718573779,-0.0797579885,0.2689031959,0.0939958096,0.1416842639,0.015278779,0.0105833048,0.3232316971,-0.4997332692,0.0140256248,0.158009991,-0.1684829742,0.2369734049,-0.1351102889,0.1746459752,0.272500664,0.0226436295,0.0001453864,0.1185611188,0.350269109,0.2929783165,-0.4297550917,-0.439863205,0.1170180514,-0.2998411059,0.1213404536,0.1156431139,-0.2244835347,0.2992221117,0.2973560691,0.0801387802,0.1543859988,0.2432106882,-0.5394952297,-0.3813195825,0.0901098847,0.0315185785,0.2323159128,0.2432172596,-0.1390015483,-0.1369745731,0.0364051759,-0.1281929165,0.2262079716,0.2362590432,0.1060448661,-0.0901912525,0.1861984283,0.0976965129,-0.3965107203,0.1131404415,-0.1362866014,0.2927363515,-0.4832673073,-0.1588353068,-0.1684350669,-0.3459651768,0.1443157345,-0.0566449314,-0.1863461137,-0.3293999732,-0.0384605601,-0.3649948537,-0.1907693446,-0.0067478972,-0.1385474503,0.0470802188,-0.0089543629,0.5758171678,-0.2279251367,-0.0529014245,-0.2607600689,0.0117808422,-0.0994350687,0.1297074854,0.0816469789,-0.259659797,-0.2590094805,-0.2978917956,-0.5375893116,0.2664554119,-0.2632860839,0.358553648,0.0612230599,-0.0123534529,-0.4086563587,-0.0889178887,0.2468182296,-0.2753388286,-0.1839739382,-0.1442348659,-0.0939528048,0.0186870303,0.1265518069,-0.5909531713,-0.2242352962,-0.1425719261,0.3636078537,-0.0673638657,0.1479123086,0.169193849,0.0336715467,0.1353575289,-0.3880838454,-0.0101905605,-0.4052891135,-0.2496473789,0.2451399863,-0.3715068996,-0.4380248785,0.1713705361,0.0093548661,0.146030888,-0.2629668117,-0.6767251492,0.0650976449,-0.2844429016,0.1649732143,-0.0391654707,0.2578712404,-0.1962353885,-0.0335553251,-0.1405586302,0.0131707154,-0.3025437593,-0.0472641811,-0.3038459718,0.4315596819,-0.2831240892,0.1189671978,0.0314503573,0.2606206536,0.4236908257,-0.1661621332,0.1330355704,-0.259968251,0.4982194602,0.1351463795,-0.2544529736,0.2596282661,-0.0836423784,0.2704189718,0.4516679645,0.0968760625,-0.0060576592,0.0286700167,-0.0716327876,-0.3876917064,-0.3234597445,-0.0732192695,-0.3693270385,0.2897177637,0.1122931615,-0.1058364362,-0.0475785881,-0.1648263484,-0.064666383,0.2462563068,0.0442378558,0.1387827843,-0.2342515737,0.1010600403,-0.3769010305,0.1123185158,-0.0255665481,0.3216902614,0.0228450801,0.0710413083,-0.0112935929,0.3466481566,0.475274384,-0.1060132459,-0.1243101731,0.1034455672,0.2500164211,-0.6698015928,0.048292052,-0.2016923428,0.1725427955,0.3898587823,0.4813296795,-0.101772815,-0.0332942493,0.2220117003,0.0038328725,-0.0554085858,-0.1946831197,-0.3725695312,-0.2727359533,-0.0943465903,-0.3107717931,0.2005602866,-0.2032244205,-0.1688935608,-0.0818771794,-0.0048373691,-0.00299721,0.1037380323,0.0469429232,0.3300943077,-0.020065425,0.043591883,0.4950048923,0.3157636225,0.0102140903,0.4064956009,-0.0269221589,-0.6741477251,0.0657280013,-0.2502689064,0.0342188925,0.1419753134,-0.1325029731,-0.0033852537,0.166695103,0.128281787,-0.0065622991,0.1957174242,0.6088898778,0.1220680326,0.0894868821,-0.3547284305,0.2650872469,0.051420033,0.0123914573,-0.1013603657,0.5811461806,-0.1059889942,0.2428254783,0.1652475744,0.8558471203,0.0546254069,0.1828423291,0.2816806436,-0.0538383611,0.5717368722,-0.0866698697,0.104508467,-0.3380014896,0.1762049645,-0.0362825915,0.0243873578,-0.0873433799,0.3483557999,-0.192088604,0.070258677,-0.0287696607,0.1276535541,0.1744466126,0.0807511285,-0.1416060477,0.0913454369,-0.3640456498,0.1393849254,-0.0821591318,0.2686319649,0.1377178431,0.0379359163,-0.0713173524,-0.1258801371,-0.5482355952,-0.2577881813,-0.6138542295,-0.0038741943,-0.184158206,0.0087310337,0.2942084074,0.5220600963,0.5753257871,-0.0724816471,-0.0636657327,0.0764410645,0.0112092765,-0.2106311023,0.0168809444,0.0478556938,0.2465670854,-0.0639683083,-0.4747828543,0.1924731433,-0.0319325104,0.1135670021,-0.4081456959,-0.0220132209,-0.2099512666,-0.2016085237,0.1455543339,-0.1266413182,-0.2544453144,-0.2183949947,0.1006955281,0.1826279014,-0.0076851794,0.0369899459,0.0159806889,-0.1750512272,-0.1224474236,0.0530377366,-0.1395868957,-0.3345024586,0.4910681248,0.2866661549,-0.1466168314,-0.1908087581,0.1806304455,0.2962314487,-0.513875246,0.128376022,-0.0219818912,0.0162779801,-0.1513835639,0.2409599125,0.0548996106,0.1371917278,-0.1302475035,-0.2732833028,-0.2836497128,0.4248706102,0.0570408888,0.004639918,-0.0850583166,-0.2370092869,0.1729513705,0.1067769751,-0.2570379674,0.0735718012,0.0665478483,-0.0176843815,0.1118415743,-0.0777178556,0.2077566087,-0.0933492407,0.0825383812,-0.0391169526,0.2671702206,-0.149182409,-0.0442343429,0.1616088748,0.0723474696,-0.1651754528,-0.0152209606,-0.0501182005,-0.0207911003,-0.166955933,0.1597813815,0.4227869809,0.0663981736,0.4139951766,-0.4394425452,0.0645753443,0.327165693,-0.1110580489,0.2328562438,0.4335312247,0.37139678,0.0603553392,0.2591707408,-0.0801139772,-0.0536843725,0.0043767393,-0.1375538707,0.2534367144,0.0494655706,-0.4649819732,0.0077139228,-0.2350022197,0.3385827541,0.168830663,-0.2895054817,0.0920871198,0.0927089453,0.1928060502,-0.3769546747,-0.0334830284,-0.2554332018,-0.18357867,0.1491666883,0.0393546373,0.1262141317,0.0949019566,-0.1320911795,-0.0742513761,0.2503602207,-0.2206671387,0.2582568228,-0.0035882695,-0.3227566779,-0.0463629439,0.2064570487,0.1990871727,0.1075853482,0.3274728358,0.187772572,0.277530849,0.1415822059,0.1120454445,0.2960928082,-0.0117396964,0.2074179202,0.1707001776,-0.3604422808,0.1573473811,-0.1794190407,0.6270523071,0.1679913104,-0.03098155,0.0112486202,-0.0917842761,-0.0041666892,0.0572589375,-0.0046415306,-0.2117327005,-0.3360927701,-0.17080082,-0.1270973533,-0.3938744068,-0.1881665438,0.2726415992,-0.3617930412,-0.6225930452,0.2395993173,0.3184606433,0.4276423454,0.0079054506,0.1754392385,0.3265024722,-0.1069334,0.0372633934,0.3438470066,0.558218658,0.2726561427,0.1357981116,0.058523871,-0.1089517325,-0.1505998224,0.0711600408,0.0317658782,0.3115769923,-0.3008610308,-0.0867409855,0.0576539934,-0.1387263238,0.1372633874,0.0521847457,0.5654798746,-0.2587907016,0.2997586727,-0.684237361,-0.2530015409,-0.1697015464,-0.1047009602,-0.3903720677,0.0478404723,0.4483489692,0.0521797575,0.0657662898,-0.2489425242,0.0888759717,-0.289729178,0.2565428913,0.3679551184,0.0715657398,-0.025419144,-0.1495766491,-0.6116506457,0.0588440746,0.0604285896,-0.2176892906,0.026487872,0.2333491594,0.094855994,0.0595007241,0.4047589004,-0.1115939096,0.1358876377,0.0189147275,-0.1016515568,-0.0852362663,-0.1031051651,0.5716150999,0.0936461166,0.1650381535,0.1955991685,0.1957147568,-0.1012588963,-0.1404992193,0.1763877273,0.0349838957,-0.1917304099,-0.0249973536,0.337870717,0.2155154049,-0.0544068031,-0.3151955307,-0.3827753365,-0.1919527501,-0.0921650231,0.4343304038,0.1147854179,0.4005175829,-0.1333856583,-0.3621267974,-0.3094433844,-0.1144904792,0.0826683268,-0.3093989789,-0.1664935201,0.0467186458,0.2304986566,-0.0848027021,0.0504702926,0.145922184,-0.1524731219,-0.0363558568,-0.2446848899,-0.3524096608,0.318600297,-0.4616992474,0.0756681934,-0.0987288505,0.3212608099,0.5552653074,-0.2004437298,-0.7379689813,-0.0266476255,0.5044902563,0.1435477734,-0.2154505253,0.0133139053,-0.0447971635,0.1685387194,0.0444825441,0.4601897299,0.3010541797,0.0484001264,0.1357054114,-0.1152162999]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2516","title":"datasets.map pickle issue resulting in invalid mapping function","comments":"Hi ! `map` calls `__getstate__` using `dill` to hash your map function. This is used by the caching mechanism to recover previously computed results. That's why you don't see any `__setstate__` call.\r\n\r\nWhy do you change an attribute of your tokenizer when `__getstate__` is called ?","body":"I trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n","comment_length":46,"text":"datasets.map pickle issue resulting in invalid mapping function\nI trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n\nHi ! `map` calls `__getstate__` using `dill` to hash your map function. This is used by the caching mechanism to recover previously computed results. That's why you don't see any `__setstate__` call.\r\n\r\nWhy do you change an attribute of your tokenizer when `__getstate__` is called ?","embeddings":[-0.2248874158,0.1931205094,0.1127529219,0.0287756212,0.0341960229,-0.3237702549,0.1385361105,0.195084691,0.3468089998,-0.100017868,0.1921808273,0.7253850698,-0.1929601282,0.2086696327,-0.0016190895,0.0595789626,0.0705189332,-0.0354076736,-0.0051377355,-0.064012669,-0.2173004895,0.0509161651,-0.3579756021,0.1536132991,-0.2765950263,-0.1768149883,0.0259390548,0.1186882928,-0.0299846679,-0.2711364627,0.2045458853,-0.0787939131,-0.0277055055,0.2411660254,-0.0001231694,-0.080115065,0.0161940679,-0.0863834247,-0.0041622892,-0.2175996751,-0.2230958939,-0.1239771247,-0.1235440746,-0.2294132262,0.0781571865,0.0024631347,-0.2965726852,-0.3813810945,0.6342011094,0.4318727851,0.1199069619,0.3648691177,-0.0678677708,-0.0346794352,-0.1355490088,0.3100190461,0.0528478138,0.22330302,0.1253270954,-0.3918119371,-0.0012149068,0.095629327,-0.1981257647,-0.216010198,0.1527506709,0.0591067895,0.1838099957,0.0106215868,-0.0086782463,0.0929949582,0.2169977278,-0.3573749661,-0.212029174,-0.3761959076,-0.0575140305,-0.1505009383,0.3402034342,-0.2168394774,0.131648913,0.1427595615,-0.0484839231,-0.1044859737,0.223940596,0.3518484533,-0.0390819162,0.2571777999,-0.016609231,0.3984128535,-0.0536790453,-0.1063232347,-0.1807260513,-0.1033443063,0.1016542092,0.4516169131,-0.0112113589,-0.2384910434,0.2105504125,-0.2123305202,-0.0243081972,0.2115104795,-0.1638644338,0.1225485504,-0.364127934,0.2640176713,0.4708405733,0.2212183774,0.1237124354,0.6363039017,0.1985704154,-0.1602922082,-0.3000450432,0.0745262355,0.3102485538,-0.101135999,-0.0373392925,0.1756477803,-0.1027164608,-0.1286521405,-0.0738535076,0.273021996,-0.7321617603,0.0083053699,0.1820911616,0.1216195971,-0.0009521256,-0.222410351,-0.0831374228,0.0777556822,-0.2158571482,0.1148085818,-0.118575193,0.1242495626,-0.3427777886,0.1104225591,-0.0432601124,0.0569451042,0.3175241649,0.1256694645,-0.1688387245,-0.059651278,0.1512505561,-0.2294484973,0.4016970098,-0.0197041202,-0.2360149175,0.0424856395,0.3069401979,-0.4420320988,-0.255575031,-0.059227135,-0.1526567638,-0.141263321,-0.0386343338,0.034725219,-0.2372962832,-0.0047180429,-0.3046909869,0.1028893217,0.3713427186,-0.2058525831,0.0988660082,-0.5015787482,-0.3563106656,-0.2076636106,0.0088023664,0.1774777621,-0.4386508763,-0.2434036136,-0.0980320275,0.1520785093,0.3357068598,0.1245676726,-0.2613388896,0.6007260084,-0.496250093,0.875687778,0.2869651318,-0.1486267298,-0.4010117352,0.0329538062,-0.1728947461,0.2320756316,-0.6788750887,-0.1038570404,0.1650154293,-0.0797584355,0.1138168424,0.3388640583,0.1549204737,-0.0953329578,-0.1282202452,0.0644310266,0.3268656433,-0.2837596238,0.0703485608,-0.0096958773,0.1624441892,-0.2167634517,0.2174033821,0.0343074575,0.1843425035,-0.1125897691,0.4559538662,0.1764437258,-0.1205102578,-0.3156250119,-0.1837683618,0.3253226578,-0.0280872509,-0.026879672,-0.1192952618,-0.0829102993,-0.1131267026,0.0560219474,-0.3399247825,-0.2433561981,0.0299025644,0.0243141092,0.2335766703,0.1753290445,-0.0835654512,-0.2479457259,-0.0506141856,0.0609713532,-0.055893112,0.0568838157,0.1251781583,-0.2686151862,-0.3178895414,0.167403996,0.4087355435,-0.0979360938,-0.1848382354,0.276186198,0.5985630751,-0.2991654873,-0.3962604403,-0.1816080213,-0.1041159853,0.1290626079,-0.2356887609,0.171751067,-0.002783909,-0.2309010923,-0.0256967079,0.5132235289,0.2476934642,0.3112365901,-0.4027119875,0.2516179383,-0.0123048415,-0.086805433,-0.1482175738,-0.337376684,-0.2627907693,0.0373323113,0.3305847347,-0.1344626099,-0.0364917815,0.0407440625,0.3557883203,-0.0735358521,0.3257069588,-0.0644305125,-0.1981063187,-0.1906806678,0.2980228662,0.3740843832,0.5166624188,-0.1092321575,-0.0572246984,0.0850376934,-0.0458905511,-0.0490012653,0.0168410391,-0.1722277254,-0.0378650054,0.1988481432,0.0485895835,-0.0998987854,-0.1812916398,0.009795065,0.0562791452,0.2922277153,-0.2386673987,0.0962414443,-0.5637153387,0.0709874779,-0.2885975242,-0.1303836107,0.0072157839,-0.4439607561,-0.2119714022,0.0941521823,-0.2360053658,0.218233943,-0.1610499918,-0.1597822011,-0.1504624039,-0.2889367342,0.0986514241,-0.2346133739,-0.2234230042,-0.0864400193,0.1288226098,0.1223812774,0.2256283313,0.3576847017,-0.4277692139,-0.34537521,-0.2861611247,0.3642108738,-0.1671608239,0.1216560602,0.3645741045,-0.210863471,-0.072285004,0.1510339975,0.3195462823,-0.1060249582,-0.1192033067,-0.1503362358,0.2773757279,0.024000911,-0.028316278,0.0144707328,-0.1706431657,-0.057648588,0.3715693355,-0.1595648676,0.0871455371,0.167441681,0.0440905541,0.2306474745,-0.1044093817,-0.260627687,-0.4392026067,-0.0985319316,0.1647433937,-0.196296975,-0.1257399321,-0.0366647914,-0.1095462143,0.1613349617,0.0417040139,-0.2359023392,0.0899801478,-0.0043371278,0.3883194923,-0.2089865357,0.1180841029,0.3752018213,0.1685070395,0.0596246868,-0.1573242694,-0.206185326,0.1814013869,0.3865940869,0.215642795,0.1895819902,0.2443961948,0.1597876698,0.8500338793,0.319329828,-0.4346615076,0.1205978617,-0.0580940507,-0.0644246265,-0.0762369111,-0.3137790561,-0.0662493855,-0.2433529049,-0.1201133803,-0.1252258867,-0.2128140479,0.12244533,0.1718871891,0.0738455802,-0.023222262,-0.4763817489,0.1844983995,-0.2536953986,0.1434853822,0.158539772,0.3790485263,-0.1934465319,0.3132577538,0.2337403893,-0.0990404263,0.4059720933,0.0916798934,-0.5711539984,-0.2006433606,-0.3427852392,0.4440487623,0.2547602654,0.1755263656,-0.0901169777,-0.0857625976,-0.1431151628,0.1859130412,0.3545980752,-0.4438407123,-0.0312311929,0.3080912232,-0.1699861586,-0.1782904714,-0.0858226717,0.203334108,0.6135460138,0.0606842488,0.5777099133,-0.132672444,-0.0017202765,-0.1848405302,0.0997186974,-0.0973077565,-0.1813914031,-0.0416360795,0.0443168096,-0.2571537495,0.1307145357,0.1788236648,-0.0311466176,0.2515218854,0.0375728458,-0.1130554006,-0.2486295402,-0.0670836121,0.0204541292,0.2222068012,0.0019844186,0.1563676298,0.2334225625,-0.0332225002,0.2050068825,0.3170377612,-0.1554178596,0.0191608332,-0.2010471821,0.0237923115,0.357534498,0.0755349174,-0.0567949973,0.3018533289,-0.0134978294,-0.1000862494,-0.292260617,0.2588860095,0.1977367997,0.097498633,-0.1309506297,-0.5797993541,0.0610708259,0.0986432731,-0.1772269309,0.360871315,0.0427088626,-0.2918627262,0.2905969024,0.1774501503,0.8005260229,0.165405795,0.0148347542,0.3365236521,0.1808232963,0.3699220419,0.2355190217,-0.0656936392,-0.3427853882,0.3571147621,-0.1838505417,-0.2929320335,0.0024912651,0.0052597281,-0.1912904531,0.2498566806,-0.0131086651,0.1557759792,-0.1824071407,0.1464309692,0.3141553998,-0.2115342617,-0.0345840752,0.1503129303,-0.0739186034,0.2885682285,-0.0357552096,0.1344928145,-0.1474824548,-0.0078977309,-0.2440619469,0.1444129795,-0.5116114616,0.3974108994,0.0836284384,-0.083952494,0.2956176698,0.0821683779,-0.070096612,0.1653814316,-0.0545432083,0.1242014319,0.2346412688,0.201288566,0.1240207553,-0.3333580196,0.2635716498,0.2254087329,-0.1046201363,-0.003838948,-0.1554680616,-0.2256289423,-0.1773545593,0.2104772925,0.2259200811,-0.3565039933,-0.23085545,-0.0348876342,0.0960318521,-0.1957101226,0.0203741658,0.1206538901,0.1306192428,0.2372598052,-0.3842768967,-0.1575082093,0.0443066582,0.4637593925,-0.2750530839,0.213260293,0.4285078347,-0.005432765,-0.0431298763,-0.120883055,0.1267634928,0.0437580012,0.2106195241,0.1623667181,-0.316500932,-0.062936686,-0.2680115402,0.3825358152,-0.0419599488,-0.207461223,-0.0652518421,-0.2240740657,-0.467910558,-0.0544447564,0.2477304786,0.1402742714,0.0923666134,0.5768774748,-0.358694613,0.1836718321,-0.1821164638,0.0947732478,0.0293321442,0.0852856189,0.0898132324,0.0372846238,-0.022229977,-0.2665495872,-0.0872106999,0.450166434,0.0098861773,-0.0926165059,-0.4276529551,0.1731844991,0.5592157245,-0.1824070662,0.074064523,-0.2170207798,-0.0007076903,-0.3593154848,-0.0251798108,0.2778584063,0.0307262968,0.2733374834,0.1893640608,0.1163053662,0.0041885739,-0.0154999653,-0.0967460424,-0.0643709451,0.0176322181,0.0547905043,0.0343399607,-0.0734430775,-0.0807646438,0.2219371945,0.2970215082,0.2084831595,0.2401298434,-0.0455406457,-0.2281159461,-0.1336230189,0.3515206873,0.3570409119,-0.2013129741,-0.2437855601,0.5111559629,0.0836803168,-0.0942781195,0.1351262927,0.275971204,0.0347155482,0.2394305468,0.1732371151,-0.2326285243,0.2811648548,-0.4663683474,-0.0875423476,0.2715706527,-0.0467737988,0.0408420227,0.0809432641,-0.1805369705,0.2933398783,0.2770445049,0.1332241595,0.3300229013,0.2303710878,0.0932229012,0.1252530068,0.1260301918,0.0945226774,0.1989385337,-0.4628060162,0.321962297,0.2426410913,-0.0055868393,0.1944713295,0.1545814723,0.4853104353,-0.5250402093,-0.2774403691,0.2316378653,0.4182745218,-0.2075717449,-0.0414236709,-0.444889009,0.0641767234,-0.0232779253,-0.0191850625,-0.0846731588,0.0040494045,0.179366678,-0.1410146952,-0.0608337894,-0.2040788233,-0.4670405388,-0.1198210195,0.3051195443,-0.2399477512,-0.1056107581,-0.1089061499,0.0585185848,-0.2150144726,-0.1264312118,0.2644938231,0.2069673091,-0.3826082647,-0.1049547568,0.0128401937,0.1148540601,-0.1220486462,0.230877012,-0.2840906084,-0.3314356804,0.3611928821,-0.0268518962,-0.0976092741,-0.1502250284,0.1964079738,0.6410113573,0.0001512721,0.2310475856,-0.0251508113,-0.0345954895,-0.1389731765,0.0527719781,0.0753154084,0.1661891788,0.1570140868,0.1477213651,0.4005223513,0.0043880064,0.0500766337,0.0997291058,0.257255882,0.3555527627,-0.3640000522,-0.5150117278,0.0507324263,-0.5516366959,0.1780667305,-0.2801930308,-0.3762613833,-0.1620282531,-0.0828687921,0.3111831248,0.0322318822,-0.0938538536,0.1567064822,0.111986272,0.187904194,0.1225926653,-0.2437008619,-0.4694603384,-0.3019477725,0.335523665,-0.2901148796,0.265989691,0.2904400229,-0.0163326152,-0.0716635361,-0.3088710904,-0.0686772689,0.0746306479,0.6968876123,0.3198541999,0.1866268665,-0.094004713,0.0994576886,-0.0232537519,0.2076236606,-0.2597785592,0.1713564843,-0.4841794372,0.390179038,0.000346194,0.263250649,-0.6178389192,0.0062231477,0.1685735583,-0.0620838366,-0.5151698589,0.268465966,-0.171492666,0.1405557096,-0.152260989,0.506000638,-0.1623695344,0.556170404,-0.2420714349,-0.1725637019,0.6069844961,-0.5967271328,-0.2871376276,0.0881195441,-0.0204459559,-0.0478141718,-0.3517234921,-0.407502383,-0.2545377314,0.318304956,-0.1005824059,-0.2755890489,-0.1272731423,0.0381449126,0.1181817204,-0.0827642754,0.7448387146,0.0860634446,-0.0579102077,0.3663490415,-0.0087578502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2516","title":"datasets.map pickle issue resulting in invalid mapping function","comments":"@lhoestq because if I try to pickle my custom tokenizer (it contains a pure python pretokenization step in an otherwise rust backed tokenizer) I get\r\n\r\n> Exception: Error while attempting to pickle Tokenizer: Custom PreTokenizer cannot be serialized\r\n\r\nSo I remove the Custom PreTokenizer in `__getstate__` and then restore it in `__setstate__` (since it doesn't contain any state). This is what my `__getstate__` \/ `__setstate__` looks like:\r\n\r\n def __getstate__(self):\r\n \"\"\"\r\n Removes pre_tokenizer since it cannot be pickled\r\n \"\"\"\r\n logger.debug(\"Copy state dict\")\r\n out = self.__dict__.copy()\r\n logger.debug(\"Detaching pre_tokenizer\")\r\n out['_tokenizer'].pre_tokenizer = tokenizers.pre_tokenizers.Sequence([]) \r\n return out\r\n\r\n def __setstate__(self, d):\r\n \"\"\"\r\n Reinstates pre_tokenizer\r\n \"\"\"\r\n logger.debug(\"Reattaching pre_tokenizer\")\r\n self.__dict__ = d\r\n self.backend_tokenizer.pre_tokenizer = self._pre_tokenizer()\r\n\r\nIf this is the case can you think of another way of avoiding my issue?","body":"I trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n","comment_length":121,"text":"datasets.map pickle issue resulting in invalid mapping function\nI trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n\n@lhoestq because if I try to pickle my custom tokenizer (it contains a pure python pretokenization step in an otherwise rust backed tokenizer) I get\r\n\r\n> Exception: Error while attempting to pickle Tokenizer: Custom PreTokenizer cannot be serialized\r\n\r\nSo I remove the Custom PreTokenizer in `__getstate__` and then restore it in `__setstate__` (since it doesn't contain any state). This is what my `__getstate__` \/ `__setstate__` looks like:\r\n\r\n def __getstate__(self):\r\n \"\"\"\r\n Removes pre_tokenizer since it cannot be pickled\r\n \"\"\"\r\n logger.debug(\"Copy state dict\")\r\n out = self.__dict__.copy()\r\n logger.debug(\"Detaching pre_tokenizer\")\r\n out['_tokenizer'].pre_tokenizer = tokenizers.pre_tokenizers.Sequence([]) \r\n return out\r\n\r\n def __setstate__(self, d):\r\n \"\"\"\r\n Reinstates pre_tokenizer\r\n \"\"\"\r\n logger.debug(\"Reattaching pre_tokenizer\")\r\n self.__dict__ = d\r\n self.backend_tokenizer.pre_tokenizer = self._pre_tokenizer()\r\n\r\nIf this is the case can you think of another way of avoiding my issue?","embeddings":[-0.2248874158,0.1931205094,0.1127529219,0.0287756212,0.0341960229,-0.3237702549,0.1385361105,0.195084691,0.3468089998,-0.100017868,0.1921808273,0.7253850698,-0.1929601282,0.2086696327,-0.0016190895,0.0595789626,0.0705189332,-0.0354076736,-0.0051377355,-0.064012669,-0.2173004895,0.0509161651,-0.3579756021,0.1536132991,-0.2765950263,-0.1768149883,0.0259390548,0.1186882928,-0.0299846679,-0.2711364627,0.2045458853,-0.0787939131,-0.0277055055,0.2411660254,-0.0001231694,-0.080115065,0.0161940679,-0.0863834247,-0.0041622892,-0.2175996751,-0.2230958939,-0.1239771247,-0.1235440746,-0.2294132262,0.0781571865,0.0024631347,-0.2965726852,-0.3813810945,0.6342011094,0.4318727851,0.1199069619,0.3648691177,-0.0678677708,-0.0346794352,-0.1355490088,0.3100190461,0.0528478138,0.22330302,0.1253270954,-0.3918119371,-0.0012149068,0.095629327,-0.1981257647,-0.216010198,0.1527506709,0.0591067895,0.1838099957,0.0106215868,-0.0086782463,0.0929949582,0.2169977278,-0.3573749661,-0.212029174,-0.3761959076,-0.0575140305,-0.1505009383,0.3402034342,-0.2168394774,0.131648913,0.1427595615,-0.0484839231,-0.1044859737,0.223940596,0.3518484533,-0.0390819162,0.2571777999,-0.016609231,0.3984128535,-0.0536790453,-0.1063232347,-0.1807260513,-0.1033443063,0.1016542092,0.4516169131,-0.0112113589,-0.2384910434,0.2105504125,-0.2123305202,-0.0243081972,0.2115104795,-0.1638644338,0.1225485504,-0.364127934,0.2640176713,0.4708405733,0.2212183774,0.1237124354,0.6363039017,0.1985704154,-0.1602922082,-0.3000450432,0.0745262355,0.3102485538,-0.101135999,-0.0373392925,0.1756477803,-0.1027164608,-0.1286521405,-0.0738535076,0.273021996,-0.7321617603,0.0083053699,0.1820911616,0.1216195971,-0.0009521256,-0.222410351,-0.0831374228,0.0777556822,-0.2158571482,0.1148085818,-0.118575193,0.1242495626,-0.3427777886,0.1104225591,-0.0432601124,0.0569451042,0.3175241649,0.1256694645,-0.1688387245,-0.059651278,0.1512505561,-0.2294484973,0.4016970098,-0.0197041202,-0.2360149175,0.0424856395,0.3069401979,-0.4420320988,-0.255575031,-0.059227135,-0.1526567638,-0.141263321,-0.0386343338,0.034725219,-0.2372962832,-0.0047180429,-0.3046909869,0.1028893217,0.3713427186,-0.2058525831,0.0988660082,-0.5015787482,-0.3563106656,-0.2076636106,0.0088023664,0.1774777621,-0.4386508763,-0.2434036136,-0.0980320275,0.1520785093,0.3357068598,0.1245676726,-0.2613388896,0.6007260084,-0.496250093,0.875687778,0.2869651318,-0.1486267298,-0.4010117352,0.0329538062,-0.1728947461,0.2320756316,-0.6788750887,-0.1038570404,0.1650154293,-0.0797584355,0.1138168424,0.3388640583,0.1549204737,-0.0953329578,-0.1282202452,0.0644310266,0.3268656433,-0.2837596238,0.0703485608,-0.0096958773,0.1624441892,-0.2167634517,0.2174033821,0.0343074575,0.1843425035,-0.1125897691,0.4559538662,0.1764437258,-0.1205102578,-0.3156250119,-0.1837683618,0.3253226578,-0.0280872509,-0.026879672,-0.1192952618,-0.0829102993,-0.1131267026,0.0560219474,-0.3399247825,-0.2433561981,0.0299025644,0.0243141092,0.2335766703,0.1753290445,-0.0835654512,-0.2479457259,-0.0506141856,0.0609713532,-0.055893112,0.0568838157,0.1251781583,-0.2686151862,-0.3178895414,0.167403996,0.4087355435,-0.0979360938,-0.1848382354,0.276186198,0.5985630751,-0.2991654873,-0.3962604403,-0.1816080213,-0.1041159853,0.1290626079,-0.2356887609,0.171751067,-0.002783909,-0.2309010923,-0.0256967079,0.5132235289,0.2476934642,0.3112365901,-0.4027119875,0.2516179383,-0.0123048415,-0.086805433,-0.1482175738,-0.337376684,-0.2627907693,0.0373323113,0.3305847347,-0.1344626099,-0.0364917815,0.0407440625,0.3557883203,-0.0735358521,0.3257069588,-0.0644305125,-0.1981063187,-0.1906806678,0.2980228662,0.3740843832,0.5166624188,-0.1092321575,-0.0572246984,0.0850376934,-0.0458905511,-0.0490012653,0.0168410391,-0.1722277254,-0.0378650054,0.1988481432,0.0485895835,-0.0998987854,-0.1812916398,0.009795065,0.0562791452,0.2922277153,-0.2386673987,0.0962414443,-0.5637153387,0.0709874779,-0.2885975242,-0.1303836107,0.0072157839,-0.4439607561,-0.2119714022,0.0941521823,-0.2360053658,0.218233943,-0.1610499918,-0.1597822011,-0.1504624039,-0.2889367342,0.0986514241,-0.2346133739,-0.2234230042,-0.0864400193,0.1288226098,0.1223812774,0.2256283313,0.3576847017,-0.4277692139,-0.34537521,-0.2861611247,0.3642108738,-0.1671608239,0.1216560602,0.3645741045,-0.210863471,-0.072285004,0.1510339975,0.3195462823,-0.1060249582,-0.1192033067,-0.1503362358,0.2773757279,0.024000911,-0.028316278,0.0144707328,-0.1706431657,-0.057648588,0.3715693355,-0.1595648676,0.0871455371,0.167441681,0.0440905541,0.2306474745,-0.1044093817,-0.260627687,-0.4392026067,-0.0985319316,0.1647433937,-0.196296975,-0.1257399321,-0.0366647914,-0.1095462143,0.1613349617,0.0417040139,-0.2359023392,0.0899801478,-0.0043371278,0.3883194923,-0.2089865357,0.1180841029,0.3752018213,0.1685070395,0.0596246868,-0.1573242694,-0.206185326,0.1814013869,0.3865940869,0.215642795,0.1895819902,0.2443961948,0.1597876698,0.8500338793,0.319329828,-0.4346615076,0.1205978617,-0.0580940507,-0.0644246265,-0.0762369111,-0.3137790561,-0.0662493855,-0.2433529049,-0.1201133803,-0.1252258867,-0.2128140479,0.12244533,0.1718871891,0.0738455802,-0.023222262,-0.4763817489,0.1844983995,-0.2536953986,0.1434853822,0.158539772,0.3790485263,-0.1934465319,0.3132577538,0.2337403893,-0.0990404263,0.4059720933,0.0916798934,-0.5711539984,-0.2006433606,-0.3427852392,0.4440487623,0.2547602654,0.1755263656,-0.0901169777,-0.0857625976,-0.1431151628,0.1859130412,0.3545980752,-0.4438407123,-0.0312311929,0.3080912232,-0.1699861586,-0.1782904714,-0.0858226717,0.203334108,0.6135460138,0.0606842488,0.5777099133,-0.132672444,-0.0017202765,-0.1848405302,0.0997186974,-0.0973077565,-0.1813914031,-0.0416360795,0.0443168096,-0.2571537495,0.1307145357,0.1788236648,-0.0311466176,0.2515218854,0.0375728458,-0.1130554006,-0.2486295402,-0.0670836121,0.0204541292,0.2222068012,0.0019844186,0.1563676298,0.2334225625,-0.0332225002,0.2050068825,0.3170377612,-0.1554178596,0.0191608332,-0.2010471821,0.0237923115,0.357534498,0.0755349174,-0.0567949973,0.3018533289,-0.0134978294,-0.1000862494,-0.292260617,0.2588860095,0.1977367997,0.097498633,-0.1309506297,-0.5797993541,0.0610708259,0.0986432731,-0.1772269309,0.360871315,0.0427088626,-0.2918627262,0.2905969024,0.1774501503,0.8005260229,0.165405795,0.0148347542,0.3365236521,0.1808232963,0.3699220419,0.2355190217,-0.0656936392,-0.3427853882,0.3571147621,-0.1838505417,-0.2929320335,0.0024912651,0.0052597281,-0.1912904531,0.2498566806,-0.0131086651,0.1557759792,-0.1824071407,0.1464309692,0.3141553998,-0.2115342617,-0.0345840752,0.1503129303,-0.0739186034,0.2885682285,-0.0357552096,0.1344928145,-0.1474824548,-0.0078977309,-0.2440619469,0.1444129795,-0.5116114616,0.3974108994,0.0836284384,-0.083952494,0.2956176698,0.0821683779,-0.070096612,0.1653814316,-0.0545432083,0.1242014319,0.2346412688,0.201288566,0.1240207553,-0.3333580196,0.2635716498,0.2254087329,-0.1046201363,-0.003838948,-0.1554680616,-0.2256289423,-0.1773545593,0.2104772925,0.2259200811,-0.3565039933,-0.23085545,-0.0348876342,0.0960318521,-0.1957101226,0.0203741658,0.1206538901,0.1306192428,0.2372598052,-0.3842768967,-0.1575082093,0.0443066582,0.4637593925,-0.2750530839,0.213260293,0.4285078347,-0.005432765,-0.0431298763,-0.120883055,0.1267634928,0.0437580012,0.2106195241,0.1623667181,-0.316500932,-0.062936686,-0.2680115402,0.3825358152,-0.0419599488,-0.207461223,-0.0652518421,-0.2240740657,-0.467910558,-0.0544447564,0.2477304786,0.1402742714,0.0923666134,0.5768774748,-0.358694613,0.1836718321,-0.1821164638,0.0947732478,0.0293321442,0.0852856189,0.0898132324,0.0372846238,-0.022229977,-0.2665495872,-0.0872106999,0.450166434,0.0098861773,-0.0926165059,-0.4276529551,0.1731844991,0.5592157245,-0.1824070662,0.074064523,-0.2170207798,-0.0007076903,-0.3593154848,-0.0251798108,0.2778584063,0.0307262968,0.2733374834,0.1893640608,0.1163053662,0.0041885739,-0.0154999653,-0.0967460424,-0.0643709451,0.0176322181,0.0547905043,0.0343399607,-0.0734430775,-0.0807646438,0.2219371945,0.2970215082,0.2084831595,0.2401298434,-0.0455406457,-0.2281159461,-0.1336230189,0.3515206873,0.3570409119,-0.2013129741,-0.2437855601,0.5111559629,0.0836803168,-0.0942781195,0.1351262927,0.275971204,0.0347155482,0.2394305468,0.1732371151,-0.2326285243,0.2811648548,-0.4663683474,-0.0875423476,0.2715706527,-0.0467737988,0.0408420227,0.0809432641,-0.1805369705,0.2933398783,0.2770445049,0.1332241595,0.3300229013,0.2303710878,0.0932229012,0.1252530068,0.1260301918,0.0945226774,0.1989385337,-0.4628060162,0.321962297,0.2426410913,-0.0055868393,0.1944713295,0.1545814723,0.4853104353,-0.5250402093,-0.2774403691,0.2316378653,0.4182745218,-0.2075717449,-0.0414236709,-0.444889009,0.0641767234,-0.0232779253,-0.0191850625,-0.0846731588,0.0040494045,0.179366678,-0.1410146952,-0.0608337894,-0.2040788233,-0.4670405388,-0.1198210195,0.3051195443,-0.2399477512,-0.1056107581,-0.1089061499,0.0585185848,-0.2150144726,-0.1264312118,0.2644938231,0.2069673091,-0.3826082647,-0.1049547568,0.0128401937,0.1148540601,-0.1220486462,0.230877012,-0.2840906084,-0.3314356804,0.3611928821,-0.0268518962,-0.0976092741,-0.1502250284,0.1964079738,0.6410113573,0.0001512721,0.2310475856,-0.0251508113,-0.0345954895,-0.1389731765,0.0527719781,0.0753154084,0.1661891788,0.1570140868,0.1477213651,0.4005223513,0.0043880064,0.0500766337,0.0997291058,0.257255882,0.3555527627,-0.3640000522,-0.5150117278,0.0507324263,-0.5516366959,0.1780667305,-0.2801930308,-0.3762613833,-0.1620282531,-0.0828687921,0.3111831248,0.0322318822,-0.0938538536,0.1567064822,0.111986272,0.187904194,0.1225926653,-0.2437008619,-0.4694603384,-0.3019477725,0.335523665,-0.2901148796,0.265989691,0.2904400229,-0.0163326152,-0.0716635361,-0.3088710904,-0.0686772689,0.0746306479,0.6968876123,0.3198541999,0.1866268665,-0.094004713,0.0994576886,-0.0232537519,0.2076236606,-0.2597785592,0.1713564843,-0.4841794372,0.390179038,0.000346194,0.263250649,-0.6178389192,0.0062231477,0.1685735583,-0.0620838366,-0.5151698589,0.268465966,-0.171492666,0.1405557096,-0.152260989,0.506000638,-0.1623695344,0.556170404,-0.2420714349,-0.1725637019,0.6069844961,-0.5967271328,-0.2871376276,0.0881195441,-0.0204459559,-0.0478141718,-0.3517234921,-0.407502383,-0.2545377314,0.318304956,-0.1005824059,-0.2755890489,-0.1272731423,0.0381449126,0.1181817204,-0.0827642754,0.7448387146,0.0860634446,-0.0579102077,0.3663490415,-0.0087578502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2516","title":"datasets.map pickle issue resulting in invalid mapping function","comments":"Actually, maybe I need to deep copy `self.__dict__`? That way `self` isn't modified. That was my intention and I thought it was working - I'll double-check after the weekend.","body":"I trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n","comment_length":29,"text":"datasets.map pickle issue resulting in invalid mapping function\nI trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n\nActually, maybe I need to deep copy `self.__dict__`? That way `self` isn't modified. That was my intention and I thought it was working - I'll double-check after the weekend.","embeddings":[-0.2248874158,0.1931205094,0.1127529219,0.0287756212,0.0341960229,-0.3237702549,0.1385361105,0.195084691,0.3468089998,-0.100017868,0.1921808273,0.7253850698,-0.1929601282,0.2086696327,-0.0016190895,0.0595789626,0.0705189332,-0.0354076736,-0.0051377355,-0.064012669,-0.2173004895,0.0509161651,-0.3579756021,0.1536132991,-0.2765950263,-0.1768149883,0.0259390548,0.1186882928,-0.0299846679,-0.2711364627,0.2045458853,-0.0787939131,-0.0277055055,0.2411660254,-0.0001231694,-0.080115065,0.0161940679,-0.0863834247,-0.0041622892,-0.2175996751,-0.2230958939,-0.1239771247,-0.1235440746,-0.2294132262,0.0781571865,0.0024631347,-0.2965726852,-0.3813810945,0.6342011094,0.4318727851,0.1199069619,0.3648691177,-0.0678677708,-0.0346794352,-0.1355490088,0.3100190461,0.0528478138,0.22330302,0.1253270954,-0.3918119371,-0.0012149068,0.095629327,-0.1981257647,-0.216010198,0.1527506709,0.0591067895,0.1838099957,0.0106215868,-0.0086782463,0.0929949582,0.2169977278,-0.3573749661,-0.212029174,-0.3761959076,-0.0575140305,-0.1505009383,0.3402034342,-0.2168394774,0.131648913,0.1427595615,-0.0484839231,-0.1044859737,0.223940596,0.3518484533,-0.0390819162,0.2571777999,-0.016609231,0.3984128535,-0.0536790453,-0.1063232347,-0.1807260513,-0.1033443063,0.1016542092,0.4516169131,-0.0112113589,-0.2384910434,0.2105504125,-0.2123305202,-0.0243081972,0.2115104795,-0.1638644338,0.1225485504,-0.364127934,0.2640176713,0.4708405733,0.2212183774,0.1237124354,0.6363039017,0.1985704154,-0.1602922082,-0.3000450432,0.0745262355,0.3102485538,-0.101135999,-0.0373392925,0.1756477803,-0.1027164608,-0.1286521405,-0.0738535076,0.273021996,-0.7321617603,0.0083053699,0.1820911616,0.1216195971,-0.0009521256,-0.222410351,-0.0831374228,0.0777556822,-0.2158571482,0.1148085818,-0.118575193,0.1242495626,-0.3427777886,0.1104225591,-0.0432601124,0.0569451042,0.3175241649,0.1256694645,-0.1688387245,-0.059651278,0.1512505561,-0.2294484973,0.4016970098,-0.0197041202,-0.2360149175,0.0424856395,0.3069401979,-0.4420320988,-0.255575031,-0.059227135,-0.1526567638,-0.141263321,-0.0386343338,0.034725219,-0.2372962832,-0.0047180429,-0.3046909869,0.1028893217,0.3713427186,-0.2058525831,0.0988660082,-0.5015787482,-0.3563106656,-0.2076636106,0.0088023664,0.1774777621,-0.4386508763,-0.2434036136,-0.0980320275,0.1520785093,0.3357068598,0.1245676726,-0.2613388896,0.6007260084,-0.496250093,0.875687778,0.2869651318,-0.1486267298,-0.4010117352,0.0329538062,-0.1728947461,0.2320756316,-0.6788750887,-0.1038570404,0.1650154293,-0.0797584355,0.1138168424,0.3388640583,0.1549204737,-0.0953329578,-0.1282202452,0.0644310266,0.3268656433,-0.2837596238,0.0703485608,-0.0096958773,0.1624441892,-0.2167634517,0.2174033821,0.0343074575,0.1843425035,-0.1125897691,0.4559538662,0.1764437258,-0.1205102578,-0.3156250119,-0.1837683618,0.3253226578,-0.0280872509,-0.026879672,-0.1192952618,-0.0829102993,-0.1131267026,0.0560219474,-0.3399247825,-0.2433561981,0.0299025644,0.0243141092,0.2335766703,0.1753290445,-0.0835654512,-0.2479457259,-0.0506141856,0.0609713532,-0.055893112,0.0568838157,0.1251781583,-0.2686151862,-0.3178895414,0.167403996,0.4087355435,-0.0979360938,-0.1848382354,0.276186198,0.5985630751,-0.2991654873,-0.3962604403,-0.1816080213,-0.1041159853,0.1290626079,-0.2356887609,0.171751067,-0.002783909,-0.2309010923,-0.0256967079,0.5132235289,0.2476934642,0.3112365901,-0.4027119875,0.2516179383,-0.0123048415,-0.086805433,-0.1482175738,-0.337376684,-0.2627907693,0.0373323113,0.3305847347,-0.1344626099,-0.0364917815,0.0407440625,0.3557883203,-0.0735358521,0.3257069588,-0.0644305125,-0.1981063187,-0.1906806678,0.2980228662,0.3740843832,0.5166624188,-0.1092321575,-0.0572246984,0.0850376934,-0.0458905511,-0.0490012653,0.0168410391,-0.1722277254,-0.0378650054,0.1988481432,0.0485895835,-0.0998987854,-0.1812916398,0.009795065,0.0562791452,0.2922277153,-0.2386673987,0.0962414443,-0.5637153387,0.0709874779,-0.2885975242,-0.1303836107,0.0072157839,-0.4439607561,-0.2119714022,0.0941521823,-0.2360053658,0.218233943,-0.1610499918,-0.1597822011,-0.1504624039,-0.2889367342,0.0986514241,-0.2346133739,-0.2234230042,-0.0864400193,0.1288226098,0.1223812774,0.2256283313,0.3576847017,-0.4277692139,-0.34537521,-0.2861611247,0.3642108738,-0.1671608239,0.1216560602,0.3645741045,-0.210863471,-0.072285004,0.1510339975,0.3195462823,-0.1060249582,-0.1192033067,-0.1503362358,0.2773757279,0.024000911,-0.028316278,0.0144707328,-0.1706431657,-0.057648588,0.3715693355,-0.1595648676,0.0871455371,0.167441681,0.0440905541,0.2306474745,-0.1044093817,-0.260627687,-0.4392026067,-0.0985319316,0.1647433937,-0.196296975,-0.1257399321,-0.0366647914,-0.1095462143,0.1613349617,0.0417040139,-0.2359023392,0.0899801478,-0.0043371278,0.3883194923,-0.2089865357,0.1180841029,0.3752018213,0.1685070395,0.0596246868,-0.1573242694,-0.206185326,0.1814013869,0.3865940869,0.215642795,0.1895819902,0.2443961948,0.1597876698,0.8500338793,0.319329828,-0.4346615076,0.1205978617,-0.0580940507,-0.0644246265,-0.0762369111,-0.3137790561,-0.0662493855,-0.2433529049,-0.1201133803,-0.1252258867,-0.2128140479,0.12244533,0.1718871891,0.0738455802,-0.023222262,-0.4763817489,0.1844983995,-0.2536953986,0.1434853822,0.158539772,0.3790485263,-0.1934465319,0.3132577538,0.2337403893,-0.0990404263,0.4059720933,0.0916798934,-0.5711539984,-0.2006433606,-0.3427852392,0.4440487623,0.2547602654,0.1755263656,-0.0901169777,-0.0857625976,-0.1431151628,0.1859130412,0.3545980752,-0.4438407123,-0.0312311929,0.3080912232,-0.1699861586,-0.1782904714,-0.0858226717,0.203334108,0.6135460138,0.0606842488,0.5777099133,-0.132672444,-0.0017202765,-0.1848405302,0.0997186974,-0.0973077565,-0.1813914031,-0.0416360795,0.0443168096,-0.2571537495,0.1307145357,0.1788236648,-0.0311466176,0.2515218854,0.0375728458,-0.1130554006,-0.2486295402,-0.0670836121,0.0204541292,0.2222068012,0.0019844186,0.1563676298,0.2334225625,-0.0332225002,0.2050068825,0.3170377612,-0.1554178596,0.0191608332,-0.2010471821,0.0237923115,0.357534498,0.0755349174,-0.0567949973,0.3018533289,-0.0134978294,-0.1000862494,-0.292260617,0.2588860095,0.1977367997,0.097498633,-0.1309506297,-0.5797993541,0.0610708259,0.0986432731,-0.1772269309,0.360871315,0.0427088626,-0.2918627262,0.2905969024,0.1774501503,0.8005260229,0.165405795,0.0148347542,0.3365236521,0.1808232963,0.3699220419,0.2355190217,-0.0656936392,-0.3427853882,0.3571147621,-0.1838505417,-0.2929320335,0.0024912651,0.0052597281,-0.1912904531,0.2498566806,-0.0131086651,0.1557759792,-0.1824071407,0.1464309692,0.3141553998,-0.2115342617,-0.0345840752,0.1503129303,-0.0739186034,0.2885682285,-0.0357552096,0.1344928145,-0.1474824548,-0.0078977309,-0.2440619469,0.1444129795,-0.5116114616,0.3974108994,0.0836284384,-0.083952494,0.2956176698,0.0821683779,-0.070096612,0.1653814316,-0.0545432083,0.1242014319,0.2346412688,0.201288566,0.1240207553,-0.3333580196,0.2635716498,0.2254087329,-0.1046201363,-0.003838948,-0.1554680616,-0.2256289423,-0.1773545593,0.2104772925,0.2259200811,-0.3565039933,-0.23085545,-0.0348876342,0.0960318521,-0.1957101226,0.0203741658,0.1206538901,0.1306192428,0.2372598052,-0.3842768967,-0.1575082093,0.0443066582,0.4637593925,-0.2750530839,0.213260293,0.4285078347,-0.005432765,-0.0431298763,-0.120883055,0.1267634928,0.0437580012,0.2106195241,0.1623667181,-0.316500932,-0.062936686,-0.2680115402,0.3825358152,-0.0419599488,-0.207461223,-0.0652518421,-0.2240740657,-0.467910558,-0.0544447564,0.2477304786,0.1402742714,0.0923666134,0.5768774748,-0.358694613,0.1836718321,-0.1821164638,0.0947732478,0.0293321442,0.0852856189,0.0898132324,0.0372846238,-0.022229977,-0.2665495872,-0.0872106999,0.450166434,0.0098861773,-0.0926165059,-0.4276529551,0.1731844991,0.5592157245,-0.1824070662,0.074064523,-0.2170207798,-0.0007076903,-0.3593154848,-0.0251798108,0.2778584063,0.0307262968,0.2733374834,0.1893640608,0.1163053662,0.0041885739,-0.0154999653,-0.0967460424,-0.0643709451,0.0176322181,0.0547905043,0.0343399607,-0.0734430775,-0.0807646438,0.2219371945,0.2970215082,0.2084831595,0.2401298434,-0.0455406457,-0.2281159461,-0.1336230189,0.3515206873,0.3570409119,-0.2013129741,-0.2437855601,0.5111559629,0.0836803168,-0.0942781195,0.1351262927,0.275971204,0.0347155482,0.2394305468,0.1732371151,-0.2326285243,0.2811648548,-0.4663683474,-0.0875423476,0.2715706527,-0.0467737988,0.0408420227,0.0809432641,-0.1805369705,0.2933398783,0.2770445049,0.1332241595,0.3300229013,0.2303710878,0.0932229012,0.1252530068,0.1260301918,0.0945226774,0.1989385337,-0.4628060162,0.321962297,0.2426410913,-0.0055868393,0.1944713295,0.1545814723,0.4853104353,-0.5250402093,-0.2774403691,0.2316378653,0.4182745218,-0.2075717449,-0.0414236709,-0.444889009,0.0641767234,-0.0232779253,-0.0191850625,-0.0846731588,0.0040494045,0.179366678,-0.1410146952,-0.0608337894,-0.2040788233,-0.4670405388,-0.1198210195,0.3051195443,-0.2399477512,-0.1056107581,-0.1089061499,0.0585185848,-0.2150144726,-0.1264312118,0.2644938231,0.2069673091,-0.3826082647,-0.1049547568,0.0128401937,0.1148540601,-0.1220486462,0.230877012,-0.2840906084,-0.3314356804,0.3611928821,-0.0268518962,-0.0976092741,-0.1502250284,0.1964079738,0.6410113573,0.0001512721,0.2310475856,-0.0251508113,-0.0345954895,-0.1389731765,0.0527719781,0.0753154084,0.1661891788,0.1570140868,0.1477213651,0.4005223513,0.0043880064,0.0500766337,0.0997291058,0.257255882,0.3555527627,-0.3640000522,-0.5150117278,0.0507324263,-0.5516366959,0.1780667305,-0.2801930308,-0.3762613833,-0.1620282531,-0.0828687921,0.3111831248,0.0322318822,-0.0938538536,0.1567064822,0.111986272,0.187904194,0.1225926653,-0.2437008619,-0.4694603384,-0.3019477725,0.335523665,-0.2901148796,0.265989691,0.2904400229,-0.0163326152,-0.0716635361,-0.3088710904,-0.0686772689,0.0746306479,0.6968876123,0.3198541999,0.1866268665,-0.094004713,0.0994576886,-0.0232537519,0.2076236606,-0.2597785592,0.1713564843,-0.4841794372,0.390179038,0.000346194,0.263250649,-0.6178389192,0.0062231477,0.1685735583,-0.0620838366,-0.5151698589,0.268465966,-0.171492666,0.1405557096,-0.152260989,0.506000638,-0.1623695344,0.556170404,-0.2420714349,-0.1725637019,0.6069844961,-0.5967271328,-0.2871376276,0.0881195441,-0.0204459559,-0.0478141718,-0.3517234921,-0.407502383,-0.2545377314,0.318304956,-0.1005824059,-0.2755890489,-0.1272731423,0.0381449126,0.1181817204,-0.0827642754,0.7448387146,0.0860634446,-0.0579102077,0.3663490415,-0.0087578502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2516","title":"datasets.map pickle issue resulting in invalid mapping function","comments":"Doing a deep copy results in the warning:\r\n\r\n> 06\/20\/2021 16:02:15 - WARNING - datasets.fingerprint - Parameter 'function'= of the transform datasets.arrow_dataset.Dataset._map_single couldn't be hashed properly, a random hash was used instead. Make sure your transforms and parameters are serializable with pickle or dill for the dataset fingerprinting and caching to work. If you reuse this transform, the caching mechanism will consider it to be different from the previous calls and recompute everything. This warning is only showed once. Subsequent hashing failures won't be showed.\r\n\r\n\r\n```\r\ndef __getstate__(self):\r\n \"\"\"\r\n Removes pre_tokenizer since it cannot be pickled\r\n \"\"\"\r\n logger.debug(\"Copy state dict\")\r\n out = copy.deepcopy(self.__dict__)\r\n logger.debug(\"Detaching pre_tokenizer\")\r\n out['_tokenizer'].pre_tokenizer = tokenizers.pre_tokenizers.Sequence([]) \r\n return out\r\n```","body":"I trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n","comment_length":114,"text":"datasets.map pickle issue resulting in invalid mapping function\nI trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n\nDoing a deep copy results in the warning:\r\n\r\n> 06\/20\/2021 16:02:15 - WARNING - datasets.fingerprint - Parameter 'function'= of the transform datasets.arrow_dataset.Dataset._map_single couldn't be hashed properly, a random hash was used instead. Make sure your transforms and parameters are serializable with pickle or dill for the dataset fingerprinting and caching to work. If you reuse this transform, the caching mechanism will consider it to be different from the previous calls and recompute everything. This warning is only showed once. Subsequent hashing failures won't be showed.\r\n\r\n\r\n```\r\ndef __getstate__(self):\r\n \"\"\"\r\n Removes pre_tokenizer since it cannot be pickled\r\n \"\"\"\r\n logger.debug(\"Copy state dict\")\r\n out = copy.deepcopy(self.__dict__)\r\n logger.debug(\"Detaching pre_tokenizer\")\r\n out['_tokenizer'].pre_tokenizer = tokenizers.pre_tokenizers.Sequence([]) \r\n return out\r\n```","embeddings":[-0.2248874158,0.1931205094,0.1127529219,0.0287756212,0.0341960229,-0.3237702549,0.1385361105,0.195084691,0.3468089998,-0.100017868,0.1921808273,0.7253850698,-0.1929601282,0.2086696327,-0.0016190895,0.0595789626,0.0705189332,-0.0354076736,-0.0051377355,-0.064012669,-0.2173004895,0.0509161651,-0.3579756021,0.1536132991,-0.2765950263,-0.1768149883,0.0259390548,0.1186882928,-0.0299846679,-0.2711364627,0.2045458853,-0.0787939131,-0.0277055055,0.2411660254,-0.0001231694,-0.080115065,0.0161940679,-0.0863834247,-0.0041622892,-0.2175996751,-0.2230958939,-0.1239771247,-0.1235440746,-0.2294132262,0.0781571865,0.0024631347,-0.2965726852,-0.3813810945,0.6342011094,0.4318727851,0.1199069619,0.3648691177,-0.0678677708,-0.0346794352,-0.1355490088,0.3100190461,0.0528478138,0.22330302,0.1253270954,-0.3918119371,-0.0012149068,0.095629327,-0.1981257647,-0.216010198,0.1527506709,0.0591067895,0.1838099957,0.0106215868,-0.0086782463,0.0929949582,0.2169977278,-0.3573749661,-0.212029174,-0.3761959076,-0.0575140305,-0.1505009383,0.3402034342,-0.2168394774,0.131648913,0.1427595615,-0.0484839231,-0.1044859737,0.223940596,0.3518484533,-0.0390819162,0.2571777999,-0.016609231,0.3984128535,-0.0536790453,-0.1063232347,-0.1807260513,-0.1033443063,0.1016542092,0.4516169131,-0.0112113589,-0.2384910434,0.2105504125,-0.2123305202,-0.0243081972,0.2115104795,-0.1638644338,0.1225485504,-0.364127934,0.2640176713,0.4708405733,0.2212183774,0.1237124354,0.6363039017,0.1985704154,-0.1602922082,-0.3000450432,0.0745262355,0.3102485538,-0.101135999,-0.0373392925,0.1756477803,-0.1027164608,-0.1286521405,-0.0738535076,0.273021996,-0.7321617603,0.0083053699,0.1820911616,0.1216195971,-0.0009521256,-0.222410351,-0.0831374228,0.0777556822,-0.2158571482,0.1148085818,-0.118575193,0.1242495626,-0.3427777886,0.1104225591,-0.0432601124,0.0569451042,0.3175241649,0.1256694645,-0.1688387245,-0.059651278,0.1512505561,-0.2294484973,0.4016970098,-0.0197041202,-0.2360149175,0.0424856395,0.3069401979,-0.4420320988,-0.255575031,-0.059227135,-0.1526567638,-0.141263321,-0.0386343338,0.034725219,-0.2372962832,-0.0047180429,-0.3046909869,0.1028893217,0.3713427186,-0.2058525831,0.0988660082,-0.5015787482,-0.3563106656,-0.2076636106,0.0088023664,0.1774777621,-0.4386508763,-0.2434036136,-0.0980320275,0.1520785093,0.3357068598,0.1245676726,-0.2613388896,0.6007260084,-0.496250093,0.875687778,0.2869651318,-0.1486267298,-0.4010117352,0.0329538062,-0.1728947461,0.2320756316,-0.6788750887,-0.1038570404,0.1650154293,-0.0797584355,0.1138168424,0.3388640583,0.1549204737,-0.0953329578,-0.1282202452,0.0644310266,0.3268656433,-0.2837596238,0.0703485608,-0.0096958773,0.1624441892,-0.2167634517,0.2174033821,0.0343074575,0.1843425035,-0.1125897691,0.4559538662,0.1764437258,-0.1205102578,-0.3156250119,-0.1837683618,0.3253226578,-0.0280872509,-0.026879672,-0.1192952618,-0.0829102993,-0.1131267026,0.0560219474,-0.3399247825,-0.2433561981,0.0299025644,0.0243141092,0.2335766703,0.1753290445,-0.0835654512,-0.2479457259,-0.0506141856,0.0609713532,-0.055893112,0.0568838157,0.1251781583,-0.2686151862,-0.3178895414,0.167403996,0.4087355435,-0.0979360938,-0.1848382354,0.276186198,0.5985630751,-0.2991654873,-0.3962604403,-0.1816080213,-0.1041159853,0.1290626079,-0.2356887609,0.171751067,-0.002783909,-0.2309010923,-0.0256967079,0.5132235289,0.2476934642,0.3112365901,-0.4027119875,0.2516179383,-0.0123048415,-0.086805433,-0.1482175738,-0.337376684,-0.2627907693,0.0373323113,0.3305847347,-0.1344626099,-0.0364917815,0.0407440625,0.3557883203,-0.0735358521,0.3257069588,-0.0644305125,-0.1981063187,-0.1906806678,0.2980228662,0.3740843832,0.5166624188,-0.1092321575,-0.0572246984,0.0850376934,-0.0458905511,-0.0490012653,0.0168410391,-0.1722277254,-0.0378650054,0.1988481432,0.0485895835,-0.0998987854,-0.1812916398,0.009795065,0.0562791452,0.2922277153,-0.2386673987,0.0962414443,-0.5637153387,0.0709874779,-0.2885975242,-0.1303836107,0.0072157839,-0.4439607561,-0.2119714022,0.0941521823,-0.2360053658,0.218233943,-0.1610499918,-0.1597822011,-0.1504624039,-0.2889367342,0.0986514241,-0.2346133739,-0.2234230042,-0.0864400193,0.1288226098,0.1223812774,0.2256283313,0.3576847017,-0.4277692139,-0.34537521,-0.2861611247,0.3642108738,-0.1671608239,0.1216560602,0.3645741045,-0.210863471,-0.072285004,0.1510339975,0.3195462823,-0.1060249582,-0.1192033067,-0.1503362358,0.2773757279,0.024000911,-0.028316278,0.0144707328,-0.1706431657,-0.057648588,0.3715693355,-0.1595648676,0.0871455371,0.167441681,0.0440905541,0.2306474745,-0.1044093817,-0.260627687,-0.4392026067,-0.0985319316,0.1647433937,-0.196296975,-0.1257399321,-0.0366647914,-0.1095462143,0.1613349617,0.0417040139,-0.2359023392,0.0899801478,-0.0043371278,0.3883194923,-0.2089865357,0.1180841029,0.3752018213,0.1685070395,0.0596246868,-0.1573242694,-0.206185326,0.1814013869,0.3865940869,0.215642795,0.1895819902,0.2443961948,0.1597876698,0.8500338793,0.319329828,-0.4346615076,0.1205978617,-0.0580940507,-0.0644246265,-0.0762369111,-0.3137790561,-0.0662493855,-0.2433529049,-0.1201133803,-0.1252258867,-0.2128140479,0.12244533,0.1718871891,0.0738455802,-0.023222262,-0.4763817489,0.1844983995,-0.2536953986,0.1434853822,0.158539772,0.3790485263,-0.1934465319,0.3132577538,0.2337403893,-0.0990404263,0.4059720933,0.0916798934,-0.5711539984,-0.2006433606,-0.3427852392,0.4440487623,0.2547602654,0.1755263656,-0.0901169777,-0.0857625976,-0.1431151628,0.1859130412,0.3545980752,-0.4438407123,-0.0312311929,0.3080912232,-0.1699861586,-0.1782904714,-0.0858226717,0.203334108,0.6135460138,0.0606842488,0.5777099133,-0.132672444,-0.0017202765,-0.1848405302,0.0997186974,-0.0973077565,-0.1813914031,-0.0416360795,0.0443168096,-0.2571537495,0.1307145357,0.1788236648,-0.0311466176,0.2515218854,0.0375728458,-0.1130554006,-0.2486295402,-0.0670836121,0.0204541292,0.2222068012,0.0019844186,0.1563676298,0.2334225625,-0.0332225002,0.2050068825,0.3170377612,-0.1554178596,0.0191608332,-0.2010471821,0.0237923115,0.357534498,0.0755349174,-0.0567949973,0.3018533289,-0.0134978294,-0.1000862494,-0.292260617,0.2588860095,0.1977367997,0.097498633,-0.1309506297,-0.5797993541,0.0610708259,0.0986432731,-0.1772269309,0.360871315,0.0427088626,-0.2918627262,0.2905969024,0.1774501503,0.8005260229,0.165405795,0.0148347542,0.3365236521,0.1808232963,0.3699220419,0.2355190217,-0.0656936392,-0.3427853882,0.3571147621,-0.1838505417,-0.2929320335,0.0024912651,0.0052597281,-0.1912904531,0.2498566806,-0.0131086651,0.1557759792,-0.1824071407,0.1464309692,0.3141553998,-0.2115342617,-0.0345840752,0.1503129303,-0.0739186034,0.2885682285,-0.0357552096,0.1344928145,-0.1474824548,-0.0078977309,-0.2440619469,0.1444129795,-0.5116114616,0.3974108994,0.0836284384,-0.083952494,0.2956176698,0.0821683779,-0.070096612,0.1653814316,-0.0545432083,0.1242014319,0.2346412688,0.201288566,0.1240207553,-0.3333580196,0.2635716498,0.2254087329,-0.1046201363,-0.003838948,-0.1554680616,-0.2256289423,-0.1773545593,0.2104772925,0.2259200811,-0.3565039933,-0.23085545,-0.0348876342,0.0960318521,-0.1957101226,0.0203741658,0.1206538901,0.1306192428,0.2372598052,-0.3842768967,-0.1575082093,0.0443066582,0.4637593925,-0.2750530839,0.213260293,0.4285078347,-0.005432765,-0.0431298763,-0.120883055,0.1267634928,0.0437580012,0.2106195241,0.1623667181,-0.316500932,-0.062936686,-0.2680115402,0.3825358152,-0.0419599488,-0.207461223,-0.0652518421,-0.2240740657,-0.467910558,-0.0544447564,0.2477304786,0.1402742714,0.0923666134,0.5768774748,-0.358694613,0.1836718321,-0.1821164638,0.0947732478,0.0293321442,0.0852856189,0.0898132324,0.0372846238,-0.022229977,-0.2665495872,-0.0872106999,0.450166434,0.0098861773,-0.0926165059,-0.4276529551,0.1731844991,0.5592157245,-0.1824070662,0.074064523,-0.2170207798,-0.0007076903,-0.3593154848,-0.0251798108,0.2778584063,0.0307262968,0.2733374834,0.1893640608,0.1163053662,0.0041885739,-0.0154999653,-0.0967460424,-0.0643709451,0.0176322181,0.0547905043,0.0343399607,-0.0734430775,-0.0807646438,0.2219371945,0.2970215082,0.2084831595,0.2401298434,-0.0455406457,-0.2281159461,-0.1336230189,0.3515206873,0.3570409119,-0.2013129741,-0.2437855601,0.5111559629,0.0836803168,-0.0942781195,0.1351262927,0.275971204,0.0347155482,0.2394305468,0.1732371151,-0.2326285243,0.2811648548,-0.4663683474,-0.0875423476,0.2715706527,-0.0467737988,0.0408420227,0.0809432641,-0.1805369705,0.2933398783,0.2770445049,0.1332241595,0.3300229013,0.2303710878,0.0932229012,0.1252530068,0.1260301918,0.0945226774,0.1989385337,-0.4628060162,0.321962297,0.2426410913,-0.0055868393,0.1944713295,0.1545814723,0.4853104353,-0.5250402093,-0.2774403691,0.2316378653,0.4182745218,-0.2075717449,-0.0414236709,-0.444889009,0.0641767234,-0.0232779253,-0.0191850625,-0.0846731588,0.0040494045,0.179366678,-0.1410146952,-0.0608337894,-0.2040788233,-0.4670405388,-0.1198210195,0.3051195443,-0.2399477512,-0.1056107581,-0.1089061499,0.0585185848,-0.2150144726,-0.1264312118,0.2644938231,0.2069673091,-0.3826082647,-0.1049547568,0.0128401937,0.1148540601,-0.1220486462,0.230877012,-0.2840906084,-0.3314356804,0.3611928821,-0.0268518962,-0.0976092741,-0.1502250284,0.1964079738,0.6410113573,0.0001512721,0.2310475856,-0.0251508113,-0.0345954895,-0.1389731765,0.0527719781,0.0753154084,0.1661891788,0.1570140868,0.1477213651,0.4005223513,0.0043880064,0.0500766337,0.0997291058,0.257255882,0.3555527627,-0.3640000522,-0.5150117278,0.0507324263,-0.5516366959,0.1780667305,-0.2801930308,-0.3762613833,-0.1620282531,-0.0828687921,0.3111831248,0.0322318822,-0.0938538536,0.1567064822,0.111986272,0.187904194,0.1225926653,-0.2437008619,-0.4694603384,-0.3019477725,0.335523665,-0.2901148796,0.265989691,0.2904400229,-0.0163326152,-0.0716635361,-0.3088710904,-0.0686772689,0.0746306479,0.6968876123,0.3198541999,0.1866268665,-0.094004713,0.0994576886,-0.0232537519,0.2076236606,-0.2597785592,0.1713564843,-0.4841794372,0.390179038,0.000346194,0.263250649,-0.6178389192,0.0062231477,0.1685735583,-0.0620838366,-0.5151698589,0.268465966,-0.171492666,0.1405557096,-0.152260989,0.506000638,-0.1623695344,0.556170404,-0.2420714349,-0.1725637019,0.6069844961,-0.5967271328,-0.2871376276,0.0881195441,-0.0204459559,-0.0478141718,-0.3517234921,-0.407502383,-0.2545377314,0.318304956,-0.1005824059,-0.2755890489,-0.1272731423,0.0381449126,0.1181817204,-0.0827642754,0.7448387146,0.0860634446,-0.0579102077,0.3663490415,-0.0087578502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2516","title":"datasets.map pickle issue resulting in invalid mapping function","comments":"Looks like there is still an object that is not pickable in your `tokenize_function` function.\r\n\r\nYou can test if an object can be pickled and hashed by using \r\n```python\r\nfrom datasets.fingerprint import Hasher\r\n\r\nHasher.hash(my_object)\r\n```\r\n\r\nUnder the hood it pickles the object to compute its hash, so it calls `__getstate__` when applicable.","body":"I trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n","comment_length":52,"text":"datasets.map pickle issue resulting in invalid mapping function\nI trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n\nLooks like there is still an object that is not pickable in your `tokenize_function` function.\r\n\r\nYou can test if an object can be pickled and hashed by using \r\n```python\r\nfrom datasets.fingerprint import Hasher\r\n\r\nHasher.hash(my_object)\r\n```\r\n\r\nUnder the hood it pickles the object to compute its hash, so it calls `__getstate__` when applicable.","embeddings":[-0.2248874158,0.1931205094,0.1127529219,0.0287756212,0.0341960229,-0.3237702549,0.1385361105,0.195084691,0.3468089998,-0.100017868,0.1921808273,0.7253850698,-0.1929601282,0.2086696327,-0.0016190895,0.0595789626,0.0705189332,-0.0354076736,-0.0051377355,-0.064012669,-0.2173004895,0.0509161651,-0.3579756021,0.1536132991,-0.2765950263,-0.1768149883,0.0259390548,0.1186882928,-0.0299846679,-0.2711364627,0.2045458853,-0.0787939131,-0.0277055055,0.2411660254,-0.0001231694,-0.080115065,0.0161940679,-0.0863834247,-0.0041622892,-0.2175996751,-0.2230958939,-0.1239771247,-0.1235440746,-0.2294132262,0.0781571865,0.0024631347,-0.2965726852,-0.3813810945,0.6342011094,0.4318727851,0.1199069619,0.3648691177,-0.0678677708,-0.0346794352,-0.1355490088,0.3100190461,0.0528478138,0.22330302,0.1253270954,-0.3918119371,-0.0012149068,0.095629327,-0.1981257647,-0.216010198,0.1527506709,0.0591067895,0.1838099957,0.0106215868,-0.0086782463,0.0929949582,0.2169977278,-0.3573749661,-0.212029174,-0.3761959076,-0.0575140305,-0.1505009383,0.3402034342,-0.2168394774,0.131648913,0.1427595615,-0.0484839231,-0.1044859737,0.223940596,0.3518484533,-0.0390819162,0.2571777999,-0.016609231,0.3984128535,-0.0536790453,-0.1063232347,-0.1807260513,-0.1033443063,0.1016542092,0.4516169131,-0.0112113589,-0.2384910434,0.2105504125,-0.2123305202,-0.0243081972,0.2115104795,-0.1638644338,0.1225485504,-0.364127934,0.2640176713,0.4708405733,0.2212183774,0.1237124354,0.6363039017,0.1985704154,-0.1602922082,-0.3000450432,0.0745262355,0.3102485538,-0.101135999,-0.0373392925,0.1756477803,-0.1027164608,-0.1286521405,-0.0738535076,0.273021996,-0.7321617603,0.0083053699,0.1820911616,0.1216195971,-0.0009521256,-0.222410351,-0.0831374228,0.0777556822,-0.2158571482,0.1148085818,-0.118575193,0.1242495626,-0.3427777886,0.1104225591,-0.0432601124,0.0569451042,0.3175241649,0.1256694645,-0.1688387245,-0.059651278,0.1512505561,-0.2294484973,0.4016970098,-0.0197041202,-0.2360149175,0.0424856395,0.3069401979,-0.4420320988,-0.255575031,-0.059227135,-0.1526567638,-0.141263321,-0.0386343338,0.034725219,-0.2372962832,-0.0047180429,-0.3046909869,0.1028893217,0.3713427186,-0.2058525831,0.0988660082,-0.5015787482,-0.3563106656,-0.2076636106,0.0088023664,0.1774777621,-0.4386508763,-0.2434036136,-0.0980320275,0.1520785093,0.3357068598,0.1245676726,-0.2613388896,0.6007260084,-0.496250093,0.875687778,0.2869651318,-0.1486267298,-0.4010117352,0.0329538062,-0.1728947461,0.2320756316,-0.6788750887,-0.1038570404,0.1650154293,-0.0797584355,0.1138168424,0.3388640583,0.1549204737,-0.0953329578,-0.1282202452,0.0644310266,0.3268656433,-0.2837596238,0.0703485608,-0.0096958773,0.1624441892,-0.2167634517,0.2174033821,0.0343074575,0.1843425035,-0.1125897691,0.4559538662,0.1764437258,-0.1205102578,-0.3156250119,-0.1837683618,0.3253226578,-0.0280872509,-0.026879672,-0.1192952618,-0.0829102993,-0.1131267026,0.0560219474,-0.3399247825,-0.2433561981,0.0299025644,0.0243141092,0.2335766703,0.1753290445,-0.0835654512,-0.2479457259,-0.0506141856,0.0609713532,-0.055893112,0.0568838157,0.1251781583,-0.2686151862,-0.3178895414,0.167403996,0.4087355435,-0.0979360938,-0.1848382354,0.276186198,0.5985630751,-0.2991654873,-0.3962604403,-0.1816080213,-0.1041159853,0.1290626079,-0.2356887609,0.171751067,-0.002783909,-0.2309010923,-0.0256967079,0.5132235289,0.2476934642,0.3112365901,-0.4027119875,0.2516179383,-0.0123048415,-0.086805433,-0.1482175738,-0.337376684,-0.2627907693,0.0373323113,0.3305847347,-0.1344626099,-0.0364917815,0.0407440625,0.3557883203,-0.0735358521,0.3257069588,-0.0644305125,-0.1981063187,-0.1906806678,0.2980228662,0.3740843832,0.5166624188,-0.1092321575,-0.0572246984,0.0850376934,-0.0458905511,-0.0490012653,0.0168410391,-0.1722277254,-0.0378650054,0.1988481432,0.0485895835,-0.0998987854,-0.1812916398,0.009795065,0.0562791452,0.2922277153,-0.2386673987,0.0962414443,-0.5637153387,0.0709874779,-0.2885975242,-0.1303836107,0.0072157839,-0.4439607561,-0.2119714022,0.0941521823,-0.2360053658,0.218233943,-0.1610499918,-0.1597822011,-0.1504624039,-0.2889367342,0.0986514241,-0.2346133739,-0.2234230042,-0.0864400193,0.1288226098,0.1223812774,0.2256283313,0.3576847017,-0.4277692139,-0.34537521,-0.2861611247,0.3642108738,-0.1671608239,0.1216560602,0.3645741045,-0.210863471,-0.072285004,0.1510339975,0.3195462823,-0.1060249582,-0.1192033067,-0.1503362358,0.2773757279,0.024000911,-0.028316278,0.0144707328,-0.1706431657,-0.057648588,0.3715693355,-0.1595648676,0.0871455371,0.167441681,0.0440905541,0.2306474745,-0.1044093817,-0.260627687,-0.4392026067,-0.0985319316,0.1647433937,-0.196296975,-0.1257399321,-0.0366647914,-0.1095462143,0.1613349617,0.0417040139,-0.2359023392,0.0899801478,-0.0043371278,0.3883194923,-0.2089865357,0.1180841029,0.3752018213,0.1685070395,0.0596246868,-0.1573242694,-0.206185326,0.1814013869,0.3865940869,0.215642795,0.1895819902,0.2443961948,0.1597876698,0.8500338793,0.319329828,-0.4346615076,0.1205978617,-0.0580940507,-0.0644246265,-0.0762369111,-0.3137790561,-0.0662493855,-0.2433529049,-0.1201133803,-0.1252258867,-0.2128140479,0.12244533,0.1718871891,0.0738455802,-0.023222262,-0.4763817489,0.1844983995,-0.2536953986,0.1434853822,0.158539772,0.3790485263,-0.1934465319,0.3132577538,0.2337403893,-0.0990404263,0.4059720933,0.0916798934,-0.5711539984,-0.2006433606,-0.3427852392,0.4440487623,0.2547602654,0.1755263656,-0.0901169777,-0.0857625976,-0.1431151628,0.1859130412,0.3545980752,-0.4438407123,-0.0312311929,0.3080912232,-0.1699861586,-0.1782904714,-0.0858226717,0.203334108,0.6135460138,0.0606842488,0.5777099133,-0.132672444,-0.0017202765,-0.1848405302,0.0997186974,-0.0973077565,-0.1813914031,-0.0416360795,0.0443168096,-0.2571537495,0.1307145357,0.1788236648,-0.0311466176,0.2515218854,0.0375728458,-0.1130554006,-0.2486295402,-0.0670836121,0.0204541292,0.2222068012,0.0019844186,0.1563676298,0.2334225625,-0.0332225002,0.2050068825,0.3170377612,-0.1554178596,0.0191608332,-0.2010471821,0.0237923115,0.357534498,0.0755349174,-0.0567949973,0.3018533289,-0.0134978294,-0.1000862494,-0.292260617,0.2588860095,0.1977367997,0.097498633,-0.1309506297,-0.5797993541,0.0610708259,0.0986432731,-0.1772269309,0.360871315,0.0427088626,-0.2918627262,0.2905969024,0.1774501503,0.8005260229,0.165405795,0.0148347542,0.3365236521,0.1808232963,0.3699220419,0.2355190217,-0.0656936392,-0.3427853882,0.3571147621,-0.1838505417,-0.2929320335,0.0024912651,0.0052597281,-0.1912904531,0.2498566806,-0.0131086651,0.1557759792,-0.1824071407,0.1464309692,0.3141553998,-0.2115342617,-0.0345840752,0.1503129303,-0.0739186034,0.2885682285,-0.0357552096,0.1344928145,-0.1474824548,-0.0078977309,-0.2440619469,0.1444129795,-0.5116114616,0.3974108994,0.0836284384,-0.083952494,0.2956176698,0.0821683779,-0.070096612,0.1653814316,-0.0545432083,0.1242014319,0.2346412688,0.201288566,0.1240207553,-0.3333580196,0.2635716498,0.2254087329,-0.1046201363,-0.003838948,-0.1554680616,-0.2256289423,-0.1773545593,0.2104772925,0.2259200811,-0.3565039933,-0.23085545,-0.0348876342,0.0960318521,-0.1957101226,0.0203741658,0.1206538901,0.1306192428,0.2372598052,-0.3842768967,-0.1575082093,0.0443066582,0.4637593925,-0.2750530839,0.213260293,0.4285078347,-0.005432765,-0.0431298763,-0.120883055,0.1267634928,0.0437580012,0.2106195241,0.1623667181,-0.316500932,-0.062936686,-0.2680115402,0.3825358152,-0.0419599488,-0.207461223,-0.0652518421,-0.2240740657,-0.467910558,-0.0544447564,0.2477304786,0.1402742714,0.0923666134,0.5768774748,-0.358694613,0.1836718321,-0.1821164638,0.0947732478,0.0293321442,0.0852856189,0.0898132324,0.0372846238,-0.022229977,-0.2665495872,-0.0872106999,0.450166434,0.0098861773,-0.0926165059,-0.4276529551,0.1731844991,0.5592157245,-0.1824070662,0.074064523,-0.2170207798,-0.0007076903,-0.3593154848,-0.0251798108,0.2778584063,0.0307262968,0.2733374834,0.1893640608,0.1163053662,0.0041885739,-0.0154999653,-0.0967460424,-0.0643709451,0.0176322181,0.0547905043,0.0343399607,-0.0734430775,-0.0807646438,0.2219371945,0.2970215082,0.2084831595,0.2401298434,-0.0455406457,-0.2281159461,-0.1336230189,0.3515206873,0.3570409119,-0.2013129741,-0.2437855601,0.5111559629,0.0836803168,-0.0942781195,0.1351262927,0.275971204,0.0347155482,0.2394305468,0.1732371151,-0.2326285243,0.2811648548,-0.4663683474,-0.0875423476,0.2715706527,-0.0467737988,0.0408420227,0.0809432641,-0.1805369705,0.2933398783,0.2770445049,0.1332241595,0.3300229013,0.2303710878,0.0932229012,0.1252530068,0.1260301918,0.0945226774,0.1989385337,-0.4628060162,0.321962297,0.2426410913,-0.0055868393,0.1944713295,0.1545814723,0.4853104353,-0.5250402093,-0.2774403691,0.2316378653,0.4182745218,-0.2075717449,-0.0414236709,-0.444889009,0.0641767234,-0.0232779253,-0.0191850625,-0.0846731588,0.0040494045,0.179366678,-0.1410146952,-0.0608337894,-0.2040788233,-0.4670405388,-0.1198210195,0.3051195443,-0.2399477512,-0.1056107581,-0.1089061499,0.0585185848,-0.2150144726,-0.1264312118,0.2644938231,0.2069673091,-0.3826082647,-0.1049547568,0.0128401937,0.1148540601,-0.1220486462,0.230877012,-0.2840906084,-0.3314356804,0.3611928821,-0.0268518962,-0.0976092741,-0.1502250284,0.1964079738,0.6410113573,0.0001512721,0.2310475856,-0.0251508113,-0.0345954895,-0.1389731765,0.0527719781,0.0753154084,0.1661891788,0.1570140868,0.1477213651,0.4005223513,0.0043880064,0.0500766337,0.0997291058,0.257255882,0.3555527627,-0.3640000522,-0.5150117278,0.0507324263,-0.5516366959,0.1780667305,-0.2801930308,-0.3762613833,-0.1620282531,-0.0828687921,0.3111831248,0.0322318822,-0.0938538536,0.1567064822,0.111986272,0.187904194,0.1225926653,-0.2437008619,-0.4694603384,-0.3019477725,0.335523665,-0.2901148796,0.265989691,0.2904400229,-0.0163326152,-0.0716635361,-0.3088710904,-0.0686772689,0.0746306479,0.6968876123,0.3198541999,0.1866268665,-0.094004713,0.0994576886,-0.0232537519,0.2076236606,-0.2597785592,0.1713564843,-0.4841794372,0.390179038,0.000346194,0.263250649,-0.6178389192,0.0062231477,0.1685735583,-0.0620838366,-0.5151698589,0.268465966,-0.171492666,0.1405557096,-0.152260989,0.506000638,-0.1623695344,0.556170404,-0.2420714349,-0.1725637019,0.6069844961,-0.5967271328,-0.2871376276,0.0881195441,-0.0204459559,-0.0478141718,-0.3517234921,-0.407502383,-0.2545377314,0.318304956,-0.1005824059,-0.2755890489,-0.1272731423,0.0381449126,0.1181817204,-0.0827642754,0.7448387146,0.0860634446,-0.0579102077,0.3663490415,-0.0087578502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2516","title":"datasets.map pickle issue resulting in invalid mapping function","comments":"I figured it out, the problem is deep copy itself uses pickle (unless you implement `__deepcopy__`). So when I changed `__getstate__` it started throwing an error.\r\n\r\nI'm sure there's a better way of doing this, but in order to return the `__dict__` without the non-pikelable pre-tokeniser and without modifying self I removed the pre-tokenizers, did a deep copy and then re-generated it.\r\n\r\nIt does work - although I noticed Hasher doesn't call `__hash__` if the object being hashed implements it which I feel it should? If it did I could return a hash of the tokenizers.json file instead.\r\n\r\n```\r\n def __getstate__(self):\r\n \"\"\"\r\n Removes pre_tokenizer since it cannot be pickled\r\n \"\"\"\r\n logger.debug(\"Copy state dict\")\r\n self.backend_tokenizer.pre_tokenizer = tokenizers.pre_tokenizers.Sequence([])\r\n out = copy.deepcopy(self.__dict__) #self.__dict__.copy()\r\n self.backend_tokenizer.pre_tokenizer = self._pre_tokenizer()\r\n\r\n return out\r\n```\r\n","body":"I trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n","comment_length":126,"text":"datasets.map pickle issue resulting in invalid mapping function\nI trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n\nI figured it out, the problem is deep copy itself uses pickle (unless you implement `__deepcopy__`). So when I changed `__getstate__` it started throwing an error.\r\n\r\nI'm sure there's a better way of doing this, but in order to return the `__dict__` without the non-pikelable pre-tokeniser and without modifying self I removed the pre-tokenizers, did a deep copy and then re-generated it.\r\n\r\nIt does work - although I noticed Hasher doesn't call `__hash__` if the object being hashed implements it which I feel it should? If it did I could return a hash of the tokenizers.json file instead.\r\n\r\n```\r\n def __getstate__(self):\r\n \"\"\"\r\n Removes pre_tokenizer since it cannot be pickled\r\n \"\"\"\r\n logger.debug(\"Copy state dict\")\r\n self.backend_tokenizer.pre_tokenizer = tokenizers.pre_tokenizers.Sequence([])\r\n out = copy.deepcopy(self.__dict__) #self.__dict__.copy()\r\n self.backend_tokenizer.pre_tokenizer = self._pre_tokenizer()\r\n\r\n return out\r\n```\r\n","embeddings":[-0.2248874158,0.1931205094,0.1127529219,0.0287756212,0.0341960229,-0.3237702549,0.1385361105,0.195084691,0.3468089998,-0.100017868,0.1921808273,0.7253850698,-0.1929601282,0.2086696327,-0.0016190895,0.0595789626,0.0705189332,-0.0354076736,-0.0051377355,-0.064012669,-0.2173004895,0.0509161651,-0.3579756021,0.1536132991,-0.2765950263,-0.1768149883,0.0259390548,0.1186882928,-0.0299846679,-0.2711364627,0.2045458853,-0.0787939131,-0.0277055055,0.2411660254,-0.0001231694,-0.080115065,0.0161940679,-0.0863834247,-0.0041622892,-0.2175996751,-0.2230958939,-0.1239771247,-0.1235440746,-0.2294132262,0.0781571865,0.0024631347,-0.2965726852,-0.3813810945,0.6342011094,0.4318727851,0.1199069619,0.3648691177,-0.0678677708,-0.0346794352,-0.1355490088,0.3100190461,0.0528478138,0.22330302,0.1253270954,-0.3918119371,-0.0012149068,0.095629327,-0.1981257647,-0.216010198,0.1527506709,0.0591067895,0.1838099957,0.0106215868,-0.0086782463,0.0929949582,0.2169977278,-0.3573749661,-0.212029174,-0.3761959076,-0.0575140305,-0.1505009383,0.3402034342,-0.2168394774,0.131648913,0.1427595615,-0.0484839231,-0.1044859737,0.223940596,0.3518484533,-0.0390819162,0.2571777999,-0.016609231,0.3984128535,-0.0536790453,-0.1063232347,-0.1807260513,-0.1033443063,0.1016542092,0.4516169131,-0.0112113589,-0.2384910434,0.2105504125,-0.2123305202,-0.0243081972,0.2115104795,-0.1638644338,0.1225485504,-0.364127934,0.2640176713,0.4708405733,0.2212183774,0.1237124354,0.6363039017,0.1985704154,-0.1602922082,-0.3000450432,0.0745262355,0.3102485538,-0.101135999,-0.0373392925,0.1756477803,-0.1027164608,-0.1286521405,-0.0738535076,0.273021996,-0.7321617603,0.0083053699,0.1820911616,0.1216195971,-0.0009521256,-0.222410351,-0.0831374228,0.0777556822,-0.2158571482,0.1148085818,-0.118575193,0.1242495626,-0.3427777886,0.1104225591,-0.0432601124,0.0569451042,0.3175241649,0.1256694645,-0.1688387245,-0.059651278,0.1512505561,-0.2294484973,0.4016970098,-0.0197041202,-0.2360149175,0.0424856395,0.3069401979,-0.4420320988,-0.255575031,-0.059227135,-0.1526567638,-0.141263321,-0.0386343338,0.034725219,-0.2372962832,-0.0047180429,-0.3046909869,0.1028893217,0.3713427186,-0.2058525831,0.0988660082,-0.5015787482,-0.3563106656,-0.2076636106,0.0088023664,0.1774777621,-0.4386508763,-0.2434036136,-0.0980320275,0.1520785093,0.3357068598,0.1245676726,-0.2613388896,0.6007260084,-0.496250093,0.875687778,0.2869651318,-0.1486267298,-0.4010117352,0.0329538062,-0.1728947461,0.2320756316,-0.6788750887,-0.1038570404,0.1650154293,-0.0797584355,0.1138168424,0.3388640583,0.1549204737,-0.0953329578,-0.1282202452,0.0644310266,0.3268656433,-0.2837596238,0.0703485608,-0.0096958773,0.1624441892,-0.2167634517,0.2174033821,0.0343074575,0.1843425035,-0.1125897691,0.4559538662,0.1764437258,-0.1205102578,-0.3156250119,-0.1837683618,0.3253226578,-0.0280872509,-0.026879672,-0.1192952618,-0.0829102993,-0.1131267026,0.0560219474,-0.3399247825,-0.2433561981,0.0299025644,0.0243141092,0.2335766703,0.1753290445,-0.0835654512,-0.2479457259,-0.0506141856,0.0609713532,-0.055893112,0.0568838157,0.1251781583,-0.2686151862,-0.3178895414,0.167403996,0.4087355435,-0.0979360938,-0.1848382354,0.276186198,0.5985630751,-0.2991654873,-0.3962604403,-0.1816080213,-0.1041159853,0.1290626079,-0.2356887609,0.171751067,-0.002783909,-0.2309010923,-0.0256967079,0.5132235289,0.2476934642,0.3112365901,-0.4027119875,0.2516179383,-0.0123048415,-0.086805433,-0.1482175738,-0.337376684,-0.2627907693,0.0373323113,0.3305847347,-0.1344626099,-0.0364917815,0.0407440625,0.3557883203,-0.0735358521,0.3257069588,-0.0644305125,-0.1981063187,-0.1906806678,0.2980228662,0.3740843832,0.5166624188,-0.1092321575,-0.0572246984,0.0850376934,-0.0458905511,-0.0490012653,0.0168410391,-0.1722277254,-0.0378650054,0.1988481432,0.0485895835,-0.0998987854,-0.1812916398,0.009795065,0.0562791452,0.2922277153,-0.2386673987,0.0962414443,-0.5637153387,0.0709874779,-0.2885975242,-0.1303836107,0.0072157839,-0.4439607561,-0.2119714022,0.0941521823,-0.2360053658,0.218233943,-0.1610499918,-0.1597822011,-0.1504624039,-0.2889367342,0.0986514241,-0.2346133739,-0.2234230042,-0.0864400193,0.1288226098,0.1223812774,0.2256283313,0.3576847017,-0.4277692139,-0.34537521,-0.2861611247,0.3642108738,-0.1671608239,0.1216560602,0.3645741045,-0.210863471,-0.072285004,0.1510339975,0.3195462823,-0.1060249582,-0.1192033067,-0.1503362358,0.2773757279,0.024000911,-0.028316278,0.0144707328,-0.1706431657,-0.057648588,0.3715693355,-0.1595648676,0.0871455371,0.167441681,0.0440905541,0.2306474745,-0.1044093817,-0.260627687,-0.4392026067,-0.0985319316,0.1647433937,-0.196296975,-0.1257399321,-0.0366647914,-0.1095462143,0.1613349617,0.0417040139,-0.2359023392,0.0899801478,-0.0043371278,0.3883194923,-0.2089865357,0.1180841029,0.3752018213,0.1685070395,0.0596246868,-0.1573242694,-0.206185326,0.1814013869,0.3865940869,0.215642795,0.1895819902,0.2443961948,0.1597876698,0.8500338793,0.319329828,-0.4346615076,0.1205978617,-0.0580940507,-0.0644246265,-0.0762369111,-0.3137790561,-0.0662493855,-0.2433529049,-0.1201133803,-0.1252258867,-0.2128140479,0.12244533,0.1718871891,0.0738455802,-0.023222262,-0.4763817489,0.1844983995,-0.2536953986,0.1434853822,0.158539772,0.3790485263,-0.1934465319,0.3132577538,0.2337403893,-0.0990404263,0.4059720933,0.0916798934,-0.5711539984,-0.2006433606,-0.3427852392,0.4440487623,0.2547602654,0.1755263656,-0.0901169777,-0.0857625976,-0.1431151628,0.1859130412,0.3545980752,-0.4438407123,-0.0312311929,0.3080912232,-0.1699861586,-0.1782904714,-0.0858226717,0.203334108,0.6135460138,0.0606842488,0.5777099133,-0.132672444,-0.0017202765,-0.1848405302,0.0997186974,-0.0973077565,-0.1813914031,-0.0416360795,0.0443168096,-0.2571537495,0.1307145357,0.1788236648,-0.0311466176,0.2515218854,0.0375728458,-0.1130554006,-0.2486295402,-0.0670836121,0.0204541292,0.2222068012,0.0019844186,0.1563676298,0.2334225625,-0.0332225002,0.2050068825,0.3170377612,-0.1554178596,0.0191608332,-0.2010471821,0.0237923115,0.357534498,0.0755349174,-0.0567949973,0.3018533289,-0.0134978294,-0.1000862494,-0.292260617,0.2588860095,0.1977367997,0.097498633,-0.1309506297,-0.5797993541,0.0610708259,0.0986432731,-0.1772269309,0.360871315,0.0427088626,-0.2918627262,0.2905969024,0.1774501503,0.8005260229,0.165405795,0.0148347542,0.3365236521,0.1808232963,0.3699220419,0.2355190217,-0.0656936392,-0.3427853882,0.3571147621,-0.1838505417,-0.2929320335,0.0024912651,0.0052597281,-0.1912904531,0.2498566806,-0.0131086651,0.1557759792,-0.1824071407,0.1464309692,0.3141553998,-0.2115342617,-0.0345840752,0.1503129303,-0.0739186034,0.2885682285,-0.0357552096,0.1344928145,-0.1474824548,-0.0078977309,-0.2440619469,0.1444129795,-0.5116114616,0.3974108994,0.0836284384,-0.083952494,0.2956176698,0.0821683779,-0.070096612,0.1653814316,-0.0545432083,0.1242014319,0.2346412688,0.201288566,0.1240207553,-0.3333580196,0.2635716498,0.2254087329,-0.1046201363,-0.003838948,-0.1554680616,-0.2256289423,-0.1773545593,0.2104772925,0.2259200811,-0.3565039933,-0.23085545,-0.0348876342,0.0960318521,-0.1957101226,0.0203741658,0.1206538901,0.1306192428,0.2372598052,-0.3842768967,-0.1575082093,0.0443066582,0.4637593925,-0.2750530839,0.213260293,0.4285078347,-0.005432765,-0.0431298763,-0.120883055,0.1267634928,0.0437580012,0.2106195241,0.1623667181,-0.316500932,-0.062936686,-0.2680115402,0.3825358152,-0.0419599488,-0.207461223,-0.0652518421,-0.2240740657,-0.467910558,-0.0544447564,0.2477304786,0.1402742714,0.0923666134,0.5768774748,-0.358694613,0.1836718321,-0.1821164638,0.0947732478,0.0293321442,0.0852856189,0.0898132324,0.0372846238,-0.022229977,-0.2665495872,-0.0872106999,0.450166434,0.0098861773,-0.0926165059,-0.4276529551,0.1731844991,0.5592157245,-0.1824070662,0.074064523,-0.2170207798,-0.0007076903,-0.3593154848,-0.0251798108,0.2778584063,0.0307262968,0.2733374834,0.1893640608,0.1163053662,0.0041885739,-0.0154999653,-0.0967460424,-0.0643709451,0.0176322181,0.0547905043,0.0343399607,-0.0734430775,-0.0807646438,0.2219371945,0.2970215082,0.2084831595,0.2401298434,-0.0455406457,-0.2281159461,-0.1336230189,0.3515206873,0.3570409119,-0.2013129741,-0.2437855601,0.5111559629,0.0836803168,-0.0942781195,0.1351262927,0.275971204,0.0347155482,0.2394305468,0.1732371151,-0.2326285243,0.2811648548,-0.4663683474,-0.0875423476,0.2715706527,-0.0467737988,0.0408420227,0.0809432641,-0.1805369705,0.2933398783,0.2770445049,0.1332241595,0.3300229013,0.2303710878,0.0932229012,0.1252530068,0.1260301918,0.0945226774,0.1989385337,-0.4628060162,0.321962297,0.2426410913,-0.0055868393,0.1944713295,0.1545814723,0.4853104353,-0.5250402093,-0.2774403691,0.2316378653,0.4182745218,-0.2075717449,-0.0414236709,-0.444889009,0.0641767234,-0.0232779253,-0.0191850625,-0.0846731588,0.0040494045,0.179366678,-0.1410146952,-0.0608337894,-0.2040788233,-0.4670405388,-0.1198210195,0.3051195443,-0.2399477512,-0.1056107581,-0.1089061499,0.0585185848,-0.2150144726,-0.1264312118,0.2644938231,0.2069673091,-0.3826082647,-0.1049547568,0.0128401937,0.1148540601,-0.1220486462,0.230877012,-0.2840906084,-0.3314356804,0.3611928821,-0.0268518962,-0.0976092741,-0.1502250284,0.1964079738,0.6410113573,0.0001512721,0.2310475856,-0.0251508113,-0.0345954895,-0.1389731765,0.0527719781,0.0753154084,0.1661891788,0.1570140868,0.1477213651,0.4005223513,0.0043880064,0.0500766337,0.0997291058,0.257255882,0.3555527627,-0.3640000522,-0.5150117278,0.0507324263,-0.5516366959,0.1780667305,-0.2801930308,-0.3762613833,-0.1620282531,-0.0828687921,0.3111831248,0.0322318822,-0.0938538536,0.1567064822,0.111986272,0.187904194,0.1225926653,-0.2437008619,-0.4694603384,-0.3019477725,0.335523665,-0.2901148796,0.265989691,0.2904400229,-0.0163326152,-0.0716635361,-0.3088710904,-0.0686772689,0.0746306479,0.6968876123,0.3198541999,0.1866268665,-0.094004713,0.0994576886,-0.0232537519,0.2076236606,-0.2597785592,0.1713564843,-0.4841794372,0.390179038,0.000346194,0.263250649,-0.6178389192,0.0062231477,0.1685735583,-0.0620838366,-0.5151698589,0.268465966,-0.171492666,0.1405557096,-0.152260989,0.506000638,-0.1623695344,0.556170404,-0.2420714349,-0.1725637019,0.6069844961,-0.5967271328,-0.2871376276,0.0881195441,-0.0204459559,-0.0478141718,-0.3517234921,-0.407502383,-0.2545377314,0.318304956,-0.1005824059,-0.2755890489,-0.1272731423,0.0381449126,0.1181817204,-0.0827642754,0.7448387146,0.0860634446,-0.0579102077,0.3663490415,-0.0087578502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2516","title":"datasets.map pickle issue resulting in invalid mapping function","comments":"I'm glad you figured something out :)\r\n\r\nRegarding hashing: we're not using hashing for the same purpose as the python `__hash__` purpose (which is in general for dictionary lookups). For example it is allowed for python hashing to not return the same hash across sessions, while our hashing must return the same hashes across sessions for the caching to work properly.","body":"I trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n","comment_length":61,"text":"datasets.map pickle issue resulting in invalid mapping function\nI trained my own tokenizer, and I needed to use a python custom class. Because of this I have to detach the custom step before saving and reattach after restore. I did this using the standard pickle `__get_state__` \/ `__set_state__` mechanism. I think it's correct but it fails when I use it inside a function which is mapped to a dataset, i.e. in the manner of run_mlm.py and other huggingface scripts.\r\n\r\nThe following reproduces the issue - most likely I'm missing something\r\n\r\nA simulated tokeniser which can be pickled\r\n\r\n```\r\nclass CustomTokenizer:\r\n def __init__(self):\r\n self.state = \"init\"\r\n\r\n def __getstate__(self):\r\n print(\"__getstate__ called\")\r\n out = self.__dict__.copy()\r\n self.state = \"pickled\"\r\n return out\r\n \r\n def __setstate__(self, d):\r\n print(\"__setstate__ called\")\r\n self.__dict__ = d\r\n self.state = \"restored\"\r\n\r\ntokenizer = CustomTokenizer()\r\n```\r\n\r\nTest that it actually works - prints \"__getstate__ called\" and \"__setstate__ called\"\r\n```\r\nimport pickle\r\nserialized = pickle.dumps(tokenizer)\r\nrestored = pickle.loads(serialized)\r\nassert restored.state == \"restored\"\r\n```\r\n\r\nSimulate a function that tokenises examples, when dataset.map is called, this function \r\n```\r\ndef tokenize_function(examples):\r\n assert tokenizer.state == \"restored\" # this shouldn't fail but it does\r\n output = tokenizer(examples) # this will fail as tokenizer isn't really a tokenizer\r\n return output\r\n```\r\n\r\nUse map to simulate tokenization\r\n```\r\nimport glob\r\nfrom datasets import load_dataset\r\n\r\nassert tokenizer.state == \"restored\"\r\ntrain_files = glob.glob('train*.csv')\r\nvalidation_files = glob.glob('validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n)\r\n```\r\n\r\nWhat's happening is I can see that __getstate__ is called but not __setstate__, so the state of `tokenize_function` is invalid at the point that it's actually executed. This doesn't matter as far as I can see for the standard tokenizers as they don't use __getstate__ \/ __setstate__. I'm not sure if there's another hook I'm supposed to implement as well?\r\n\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n in \r\n 8 tokenized_datasets = datasets.map(\r\n 9 tokenize_function,\r\n---> 10 batched=True,\r\n 11 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in (.0)\r\n 487 desc=desc,\r\n 488 )\r\n--> 489 for k, dataset in self.items()\r\n 490 }\r\n 491 )\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1633 fn_kwargs=fn_kwargs,\r\n 1634 new_fingerprint=new_fingerprint,\r\n-> 1635 desc=desc,\r\n 1636 )\r\n 1637 else:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 184 }\r\n 185 # apply actual function\r\n--> 186 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 187 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 188 # re-apply format to the output\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1961 indices,\r\n 1962 check_same_num_examples=len(input_dataset.list_indexes()) > 0,\r\n-> 1963 offset=offset,\r\n 1964 )\r\n 1965 except NumExamplesMismatch:\r\n\r\n~\/.pyenv\/versions\/3.7.6\/envs\/xxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1853 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1854 processed_inputs = (\r\n-> 1855 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1856 )\r\n 1857 if update_data is None:\r\n\r\n in tokenize_function(examples)\r\n 1 def tokenize_function(examples):\r\n----> 2 assert tokenizer.state == \"restored\"\r\n 3 tokenizer(examples)\r\n 4 return examples\r\n\r\n\r\n\nI'm glad you figured something out :)\r\n\r\nRegarding hashing: we're not using hashing for the same purpose as the python `__hash__` purpose (which is in general for dictionary lookups). For example it is allowed for python hashing to not return the same hash across sessions, while our hashing must return the same hashes across sessions for the caching to work properly.","embeddings":[-0.2248874158,0.1931205094,0.1127529219,0.0287756212,0.0341960229,-0.3237702549,0.1385361105,0.195084691,0.3468089998,-0.100017868,0.1921808273,0.7253850698,-0.1929601282,0.2086696327,-0.0016190895,0.0595789626,0.0705189332,-0.0354076736,-0.0051377355,-0.064012669,-0.2173004895,0.0509161651,-0.3579756021,0.1536132991,-0.2765950263,-0.1768149883,0.0259390548,0.1186882928,-0.0299846679,-0.2711364627,0.2045458853,-0.0787939131,-0.0277055055,0.2411660254,-0.0001231694,-0.080115065,0.0161940679,-0.0863834247,-0.0041622892,-0.2175996751,-0.2230958939,-0.1239771247,-0.1235440746,-0.2294132262,0.0781571865,0.0024631347,-0.2965726852,-0.3813810945,0.6342011094,0.4318727851,0.1199069619,0.3648691177,-0.0678677708,-0.0346794352,-0.1355490088,0.3100190461,0.0528478138,0.22330302,0.1253270954,-0.3918119371,-0.0012149068,0.095629327,-0.1981257647,-0.216010198,0.1527506709,0.0591067895,0.1838099957,0.0106215868,-0.0086782463,0.0929949582,0.2169977278,-0.3573749661,-0.212029174,-0.3761959076,-0.0575140305,-0.1505009383,0.3402034342,-0.2168394774,0.131648913,0.1427595615,-0.0484839231,-0.1044859737,0.223940596,0.3518484533,-0.0390819162,0.2571777999,-0.016609231,0.3984128535,-0.0536790453,-0.1063232347,-0.1807260513,-0.1033443063,0.1016542092,0.4516169131,-0.0112113589,-0.2384910434,0.2105504125,-0.2123305202,-0.0243081972,0.2115104795,-0.1638644338,0.1225485504,-0.364127934,0.2640176713,0.4708405733,0.2212183774,0.1237124354,0.6363039017,0.1985704154,-0.1602922082,-0.3000450432,0.0745262355,0.3102485538,-0.101135999,-0.0373392925,0.1756477803,-0.1027164608,-0.1286521405,-0.0738535076,0.273021996,-0.7321617603,0.0083053699,0.1820911616,0.1216195971,-0.0009521256,-0.222410351,-0.0831374228,0.0777556822,-0.2158571482,0.1148085818,-0.118575193,0.1242495626,-0.3427777886,0.1104225591,-0.0432601124,0.0569451042,0.3175241649,0.1256694645,-0.1688387245,-0.059651278,0.1512505561,-0.2294484973,0.4016970098,-0.0197041202,-0.2360149175,0.0424856395,0.3069401979,-0.4420320988,-0.255575031,-0.059227135,-0.1526567638,-0.141263321,-0.0386343338,0.034725219,-0.2372962832,-0.0047180429,-0.3046909869,0.1028893217,0.3713427186,-0.2058525831,0.0988660082,-0.5015787482,-0.3563106656,-0.2076636106,0.0088023664,0.1774777621,-0.4386508763,-0.2434036136,-0.0980320275,0.1520785093,0.3357068598,0.1245676726,-0.2613388896,0.6007260084,-0.496250093,0.875687778,0.2869651318,-0.1486267298,-0.4010117352,0.0329538062,-0.1728947461,0.2320756316,-0.6788750887,-0.1038570404,0.1650154293,-0.0797584355,0.1138168424,0.3388640583,0.1549204737,-0.0953329578,-0.1282202452,0.0644310266,0.3268656433,-0.2837596238,0.0703485608,-0.0096958773,0.1624441892,-0.2167634517,0.2174033821,0.0343074575,0.1843425035,-0.1125897691,0.4559538662,0.1764437258,-0.1205102578,-0.3156250119,-0.1837683618,0.3253226578,-0.0280872509,-0.026879672,-0.1192952618,-0.0829102993,-0.1131267026,0.0560219474,-0.3399247825,-0.2433561981,0.0299025644,0.0243141092,0.2335766703,0.1753290445,-0.0835654512,-0.2479457259,-0.0506141856,0.0609713532,-0.055893112,0.0568838157,0.1251781583,-0.2686151862,-0.3178895414,0.167403996,0.4087355435,-0.0979360938,-0.1848382354,0.276186198,0.5985630751,-0.2991654873,-0.3962604403,-0.1816080213,-0.1041159853,0.1290626079,-0.2356887609,0.171751067,-0.002783909,-0.2309010923,-0.0256967079,0.5132235289,0.2476934642,0.3112365901,-0.4027119875,0.2516179383,-0.0123048415,-0.086805433,-0.1482175738,-0.337376684,-0.2627907693,0.0373323113,0.3305847347,-0.1344626099,-0.0364917815,0.0407440625,0.3557883203,-0.0735358521,0.3257069588,-0.0644305125,-0.1981063187,-0.1906806678,0.2980228662,0.3740843832,0.5166624188,-0.1092321575,-0.0572246984,0.0850376934,-0.0458905511,-0.0490012653,0.0168410391,-0.1722277254,-0.0378650054,0.1988481432,0.0485895835,-0.0998987854,-0.1812916398,0.009795065,0.0562791452,0.2922277153,-0.2386673987,0.0962414443,-0.5637153387,0.0709874779,-0.2885975242,-0.1303836107,0.0072157839,-0.4439607561,-0.2119714022,0.0941521823,-0.2360053658,0.218233943,-0.1610499918,-0.1597822011,-0.1504624039,-0.2889367342,0.0986514241,-0.2346133739,-0.2234230042,-0.0864400193,0.1288226098,0.1223812774,0.2256283313,0.3576847017,-0.4277692139,-0.34537521,-0.2861611247,0.3642108738,-0.1671608239,0.1216560602,0.3645741045,-0.210863471,-0.072285004,0.1510339975,0.3195462823,-0.1060249582,-0.1192033067,-0.1503362358,0.2773757279,0.024000911,-0.028316278,0.0144707328,-0.1706431657,-0.057648588,0.3715693355,-0.1595648676,0.0871455371,0.167441681,0.0440905541,0.2306474745,-0.1044093817,-0.260627687,-0.4392026067,-0.0985319316,0.1647433937,-0.196296975,-0.1257399321,-0.0366647914,-0.1095462143,0.1613349617,0.0417040139,-0.2359023392,0.0899801478,-0.0043371278,0.3883194923,-0.2089865357,0.1180841029,0.3752018213,0.1685070395,0.0596246868,-0.1573242694,-0.206185326,0.1814013869,0.3865940869,0.215642795,0.1895819902,0.2443961948,0.1597876698,0.8500338793,0.319329828,-0.4346615076,0.1205978617,-0.0580940507,-0.0644246265,-0.0762369111,-0.3137790561,-0.0662493855,-0.2433529049,-0.1201133803,-0.1252258867,-0.2128140479,0.12244533,0.1718871891,0.0738455802,-0.023222262,-0.4763817489,0.1844983995,-0.2536953986,0.1434853822,0.158539772,0.3790485263,-0.1934465319,0.3132577538,0.2337403893,-0.0990404263,0.4059720933,0.0916798934,-0.5711539984,-0.2006433606,-0.3427852392,0.4440487623,0.2547602654,0.1755263656,-0.0901169777,-0.0857625976,-0.1431151628,0.1859130412,0.3545980752,-0.4438407123,-0.0312311929,0.3080912232,-0.1699861586,-0.1782904714,-0.0858226717,0.203334108,0.6135460138,0.0606842488,0.5777099133,-0.132672444,-0.0017202765,-0.1848405302,0.0997186974,-0.0973077565,-0.1813914031,-0.0416360795,0.0443168096,-0.2571537495,0.1307145357,0.1788236648,-0.0311466176,0.2515218854,0.0375728458,-0.1130554006,-0.2486295402,-0.0670836121,0.0204541292,0.2222068012,0.0019844186,0.1563676298,0.2334225625,-0.0332225002,0.2050068825,0.3170377612,-0.1554178596,0.0191608332,-0.2010471821,0.0237923115,0.357534498,0.0755349174,-0.0567949973,0.3018533289,-0.0134978294,-0.1000862494,-0.292260617,0.2588860095,0.1977367997,0.097498633,-0.1309506297,-0.5797993541,0.0610708259,0.0986432731,-0.1772269309,0.360871315,0.0427088626,-0.2918627262,0.2905969024,0.1774501503,0.8005260229,0.165405795,0.0148347542,0.3365236521,0.1808232963,0.3699220419,0.2355190217,-0.0656936392,-0.3427853882,0.3571147621,-0.1838505417,-0.2929320335,0.0024912651,0.0052597281,-0.1912904531,0.2498566806,-0.0131086651,0.1557759792,-0.1824071407,0.1464309692,0.3141553998,-0.2115342617,-0.0345840752,0.1503129303,-0.0739186034,0.2885682285,-0.0357552096,0.1344928145,-0.1474824548,-0.0078977309,-0.2440619469,0.1444129795,-0.5116114616,0.3974108994,0.0836284384,-0.083952494,0.2956176698,0.0821683779,-0.070096612,0.1653814316,-0.0545432083,0.1242014319,0.2346412688,0.201288566,0.1240207553,-0.3333580196,0.2635716498,0.2254087329,-0.1046201363,-0.003838948,-0.1554680616,-0.2256289423,-0.1773545593,0.2104772925,0.2259200811,-0.3565039933,-0.23085545,-0.0348876342,0.0960318521,-0.1957101226,0.0203741658,0.1206538901,0.1306192428,0.2372598052,-0.3842768967,-0.1575082093,0.0443066582,0.4637593925,-0.2750530839,0.213260293,0.4285078347,-0.005432765,-0.0431298763,-0.120883055,0.1267634928,0.0437580012,0.2106195241,0.1623667181,-0.316500932,-0.062936686,-0.2680115402,0.3825358152,-0.0419599488,-0.207461223,-0.0652518421,-0.2240740657,-0.467910558,-0.0544447564,0.2477304786,0.1402742714,0.0923666134,0.5768774748,-0.358694613,0.1836718321,-0.1821164638,0.0947732478,0.0293321442,0.0852856189,0.0898132324,0.0372846238,-0.022229977,-0.2665495872,-0.0872106999,0.450166434,0.0098861773,-0.0926165059,-0.4276529551,0.1731844991,0.5592157245,-0.1824070662,0.074064523,-0.2170207798,-0.0007076903,-0.3593154848,-0.0251798108,0.2778584063,0.0307262968,0.2733374834,0.1893640608,0.1163053662,0.0041885739,-0.0154999653,-0.0967460424,-0.0643709451,0.0176322181,0.0547905043,0.0343399607,-0.0734430775,-0.0807646438,0.2219371945,0.2970215082,0.2084831595,0.2401298434,-0.0455406457,-0.2281159461,-0.1336230189,0.3515206873,0.3570409119,-0.2013129741,-0.2437855601,0.5111559629,0.0836803168,-0.0942781195,0.1351262927,0.275971204,0.0347155482,0.2394305468,0.1732371151,-0.2326285243,0.2811648548,-0.4663683474,-0.0875423476,0.2715706527,-0.0467737988,0.0408420227,0.0809432641,-0.1805369705,0.2933398783,0.2770445049,0.1332241595,0.3300229013,0.2303710878,0.0932229012,0.1252530068,0.1260301918,0.0945226774,0.1989385337,-0.4628060162,0.321962297,0.2426410913,-0.0055868393,0.1944713295,0.1545814723,0.4853104353,-0.5250402093,-0.2774403691,0.2316378653,0.4182745218,-0.2075717449,-0.0414236709,-0.444889009,0.0641767234,-0.0232779253,-0.0191850625,-0.0846731588,0.0040494045,0.179366678,-0.1410146952,-0.0608337894,-0.2040788233,-0.4670405388,-0.1198210195,0.3051195443,-0.2399477512,-0.1056107581,-0.1089061499,0.0585185848,-0.2150144726,-0.1264312118,0.2644938231,0.2069673091,-0.3826082647,-0.1049547568,0.0128401937,0.1148540601,-0.1220486462,0.230877012,-0.2840906084,-0.3314356804,0.3611928821,-0.0268518962,-0.0976092741,-0.1502250284,0.1964079738,0.6410113573,0.0001512721,0.2310475856,-0.0251508113,-0.0345954895,-0.1389731765,0.0527719781,0.0753154084,0.1661891788,0.1570140868,0.1477213651,0.4005223513,0.0043880064,0.0500766337,0.0997291058,0.257255882,0.3555527627,-0.3640000522,-0.5150117278,0.0507324263,-0.5516366959,0.1780667305,-0.2801930308,-0.3762613833,-0.1620282531,-0.0828687921,0.3111831248,0.0322318822,-0.0938538536,0.1567064822,0.111986272,0.187904194,0.1225926653,-0.2437008619,-0.4694603384,-0.3019477725,0.335523665,-0.2901148796,0.265989691,0.2904400229,-0.0163326152,-0.0716635361,-0.3088710904,-0.0686772689,0.0746306479,0.6968876123,0.3198541999,0.1866268665,-0.094004713,0.0994576886,-0.0232537519,0.2076236606,-0.2597785592,0.1713564843,-0.4841794372,0.390179038,0.000346194,0.263250649,-0.6178389192,0.0062231477,0.1685735583,-0.0620838366,-0.5151698589,0.268465966,-0.171492666,0.1405557096,-0.152260989,0.506000638,-0.1623695344,0.556170404,-0.2420714349,-0.1725637019,0.6069844961,-0.5967271328,-0.2871376276,0.0881195441,-0.0204459559,-0.0478141718,-0.3517234921,-0.407502383,-0.2545377314,0.318304956,-0.1005824059,-0.2755890489,-0.1272731423,0.0381449126,0.1181817204,-0.0827642754,0.7448387146,0.0860634446,-0.0579102077,0.3663490415,-0.0087578502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2514","title":"Can datasets remove duplicated rows?","comments":"Hi ! For now this is probably the best option.\r\nWe might add a feature like this in the feature as well.\r\n\r\nDo you know any deduplication method that works on arbitrary big datasets without filling up RAM ?\r\nOtherwise we can have do the deduplication in memory like pandas but I feel like this is going to be limiting for some cases","body":"**Is your feature request related to a problem? Please describe.**\r\ni find myself more and more relying on datasets just to do all the preprocessing. One thing however, for removing duplicated rows, I couldn't find out how and am always converting datasets to pandas to do that..\r\n\r\n\r\n**Describe the solution you'd like**\r\nhave a functionality of \" remove duplicated rows\"\r\n\r\n**Describe alternatives you've considered**\r\nconvert dataset to pandas, remove duplicate, and convert back...\r\n\r\n\r\n**Additional context**\r\nno","comment_length":63,"text":"Can datasets remove duplicated rows?\n**Is your feature request related to a problem? Please describe.**\r\ni find myself more and more relying on datasets just to do all the preprocessing. One thing however, for removing duplicated rows, I couldn't find out how and am always converting datasets to pandas to do that..\r\n\r\n\r\n**Describe the solution you'd like**\r\nhave a functionality of \" remove duplicated rows\"\r\n\r\n**Describe alternatives you've considered**\r\nconvert dataset to pandas, remove duplicate, and convert back...\r\n\r\n\r\n**Additional context**\r\nno\nHi ! For now this is probably the best option.\r\nWe might add a feature like this in the feature as well.\r\n\r\nDo you know any deduplication method that works on arbitrary big datasets without filling up RAM ?\r\nOtherwise we can have do the deduplication in memory like pandas but I feel like this is going to be limiting for some cases","embeddings":[0.1406852901,-0.1654824466,-0.1650884748,0.1540047675,0.0489430018,0.2394908667,0.2769089341,0.0634394139,-0.337651521,0.0192047395,-0.0341103598,0.2346976399,0.1190901324,0.2062148005,-0.0816838667,0.0102524916,0.0978407934,0.1727186441,-0.0850968063,0.0356911197,-0.074069798,-0.2448657155,-0.2372234464,-0.1401374787,-0.4883829653,-0.070132032,-0.3793184757,-0.0523847528,-0.1877827346,-0.056332361,0.1769872159,0.5257619619,-0.015922714,0.3108595312,-0.0001015379,-0.3572315574,0.0228314269,0.1596541554,-0.3186165392,0.1642242223,-0.3713808358,-0.0732265413,0.0418946855,-0.0257244334,0.0256408472,-0.2802483737,-0.2648801506,-0.2234140188,0.1587850899,0.0940889865,0.2888639867,-0.0277582351,-0.2481730878,-0.0161672868,0.0297317747,0.4935165644,-0.1404056847,0.2517938316,0.452596426,-0.0940500572,0.159520492,0.1083615348,-0.0224840548,0.2351039499,0.1676221341,-0.1702308059,-0.2057350129,-0.1909149438,0.2019746602,0.3055927455,0.7380333543,-0.4114798307,-0.3087230623,-0.2445228994,0.4246825278,-0.0125919525,-0.0807776153,0.0383970067,-0.018863,0.0967520401,-0.3541407883,-0.5405525565,-0.1078156456,-0.2075586021,-0.028380191,-0.5066338778,0.0746085048,-0.1145889312,0.078236565,-0.0259599183,0.4261381924,-0.149689123,0.2013802528,0.3944005072,-0.3874807656,-0.3127180636,-0.1966116428,-0.0839563534,0.2826720476,0.1205257624,0.3587394059,-0.0393952876,-0.105467774,0.2516821921,0.3106621206,0.0013534994,-0.1374455988,0.0365054049,0.052742321,-0.2087919265,-0.2718610466,-0.1024351642,0.0155878998,0.2337319106,0.3121547103,-0.0734114721,0.2197850198,-0.3145796657,-0.3696149886,0.0395031124,-0.2582018971,-0.1338284761,-0.155918017,-0.1158696488,0.1043353975,-0.0742390081,-0.1398553848,-0.0562564954,-0.2685828209,-0.3536014557,-0.0524646491,0.0923020691,-0.1924515069,0.0273811817,0.3380941153,-0.1689910293,-0.0459467173,0.2378150672,0.1735194027,-0.1429393291,0.3021446764,-0.2683894038,0.058059983,-0.1268226653,-0.1946309805,-0.2599046826,-0.1769257486,0.1686663032,-0.0677968115,0.4203227162,0.1222819611,-0.0890091881,-0.02607527,0.3075685799,0.0380233452,-0.1720773876,-0.2657401562,0.2884062529,-0.1524928212,0.133170411,0.0765210241,0.0340102538,-0.0129356328,-0.2317591459,-0.067217961,0.1737954319,-0.4373749197,-0.0078025162,-0.1961129606,-0.3269720972,-0.1153033823,0.389149636,-0.0075870343,0.1604729742,0.0140760802,-0.0453709252,0.046333164,0.0908388793,-0.1502966583,0.2502066195,-0.0282454118,0.2478695214,-0.1319449395,0.3640062213,0.3663207293,-0.0124331871,0.1415612698,0.0014437262,-0.277880013,-0.376496464,-0.2161200643,-0.416598767,0.1194738597,0.1059599742,0.1478085071,0.0597624816,0.1666277796,-0.1710539609,0.1553811878,-0.0631804764,-0.1182341799,-0.056508176,0.4424176216,0.1002992094,0.1001491174,-0.3835816383,-0.4197447896,0.1171433181,-0.1840033531,-0.117498748,0.2460185587,-0.6625521183,0.0903072953,-0.0896113589,-0.0794647709,0.1624142379,0.237485528,-0.3466388285,0.0311691761,-0.2786119282,-0.2357189506,0.3210316598,-0.2857351005,0.0295973048,-0.1031141207,0.3592186272,0.1591356546,-0.0108295381,0.1967078745,0.3669462502,-0.0154448543,0.1201516166,0.0040446571,0.1093618497,0.1113897711,0.0468276031,0.2053043842,0.1967709661,0.1637277007,-0.0249819849,-0.0328928381,0.3001321554,0.0941672623,0.0090635037,-0.4436185956,0.1890527904,-0.3968798518,-0.1433710903,0.0715351328,-0.0711398199,0.0350454971,-0.1869291216,0.1368222237,-0.2567847967,0.1241660938,0.1711545736,-0.1705221832,0.3658866882,-0.4551768899,0.1549885869,-0.1446327567,-0.1276321411,0.1036438197,-0.027899174,-0.0213380922,-0.133742094,0.2878712416,0.3261114955,0.3545413613,0.2429133207,0.2274253368,0.2056377381,0.2314518541,0.1031549647,0.0854174942,-0.1205092072,0.023635624,-0.1802590787,-0.09628544,0.1669261754,-0.3265921474,0.1067498773,0.2158196568,0.1735782027,-0.0868751258,-0.3450119793,-0.5039238334,-0.5576465726,-0.0695749,0.2795039713,-0.0923288912,0.0339048468,0.4795560837,0.1193517372,0.0990646407,-0.0422910415,-0.2414945513,0.6234964728,-0.2915302813,0.0142950574,-0.0168413874,-0.1014138609,-0.0281202346,0.3210752308,0.1391530782,0.3040886223,0.7592910528,0.1577477008,0.0796603933,-0.6362704635,-0.1932237595,-0.0785812587,0.0907432958,0.1842636764,0.0044846823,0.0530583374,-0.1311346442,0.1233926937,-0.1617035866,-0.0167220328,-0.1975779235,-0.2556275129,0.2330762595,0.175649941,-0.0681055039,-0.2742835283,0.0457902439,-0.1980942041,0.2695210278,-0.262224853,0.0951305628,-0.3755032718,-0.0681842864,-0.1175014302,0.1213092804,0.0528002046,-0.3482720852,-0.2078989148,-0.0830827206,0.0397475362,-0.1691691726,-0.1180493534,0.2818202674,-0.0622634403,0.390932709,-0.1327950656,-0.2212595642,0.027614817,0.1558007002,0.0163267218,-0.0522644147,0.2613501251,0.283583194,-0.1511324644,0.0405747779,-0.0436748639,-0.1215490848,0.0741653964,-0.0268685445,0.1740274727,0.2242369503,0.2291259766,0.391993165,0.4572185278,0.0120550031,0.1044704765,0.1064218581,0.1943872869,0.0151882833,-0.393497169,-0.0631571263,-0.252823323,0.1516997069,0.4256210625,0.1379581839,-0.2058085203,0.0294687245,-0.3817224801,-0.2387852073,-0.2324654758,-0.0845031589,-0.4879080355,0.3477216363,-0.001710524,-0.2861219049,0.044184342,-0.0023837227,-0.0917697251,-0.0810923129,0.2863119841,-0.2717479765,-0.2790020704,-0.2390105575,-0.1783727854,0.2477947176,0.1335837245,0.1735503525,0.0309215542,0.2605943382,0.1744631827,0.1475320458,0.4469807744,-0.2209324688,-0.2941132486,-0.0769939646,-0.1500929743,0.1528449655,0.1882310361,-0.059576802,0.1963181049,-0.007738167,-0.0078367442,-0.1292000264,-0.232116878,0.3821360469,0.3925136626,-0.2230862826,-0.3086245954,0.2006906718,-0.150943473,-0.3083112836,0.1182738841,0.0833298713,-0.2034464628,-0.0352648795,-0.2010384947,-0.1889076382,-0.057460729,0.5008252263,-0.1931931973,-0.0677116588,-0.058531858,0.3980497122,0.0129431169,-0.065827243,0.5238155723,0.4703602493,-0.2744763196,-0.2046439052,-0.4385216236,-0.476215899,0.4185963571,0.1051482335,0.0575382523,0.4906478226,-0.1769179553,-0.096455127,-0.260318011,-0.3278205097,0.1610073298,-0.0771318227,-0.2866964936,-0.4126425087,0.156919986,0.2747707665,-0.1696583331,0.3188519478,0.2740939856,-0.2063532472,0.0082202917,0.0602805503,0.7813103199,-0.057261534,-0.0152090192,-0.4095437229,-0.2641257048,0.2027219832,0.1019192412,0.0460514352,0.0212360136,-0.3808659613,-0.0866684988,0.1076633856,0.2176419944,0.3884306252,0.0392843075,0.2841058075,-0.0924634412,-0.0005435649,0.2943615913,0.375017643,0.0731351599,0.1393528432,0.0992627665,0.1993832141,0.1266742051,-0.4207122922,-0.1413902938,0.1719731838,0.478608489,0.0558302701,-0.0638378263,-0.352922529,-0.1126641929,0.2513096631,-0.2462433726,-0.1611236185,0.2035205513,-0.1979615837,0.1111254692,0.0654316917,-0.043432042,-0.0374819599,0.4285085201,0.3341982663,-0.1032542959,-0.104805924,0.2820320129,-0.0139261661,-0.1866949648,0.017701447,0.2604140639,-0.3278184235,-0.2937097549,0.1115783155,-0.2877513468,-0.1948918104,-0.0711988211,-0.0361999162,-0.1690892577,-0.2639679313,0.2227433026,0.2321110219,-0.0886576772,0.5801397562,0.096572943,0.1145547405,-0.284545958,0.0514660887,0.6038421392,-0.2942626178,0.0203121509,-0.0450427085,-0.3466790915,-0.3113832474,0.0495288298,-0.0439983681,-0.1276925057,0.268538326,-0.2381167263,0.2652869225,0.2355503291,-0.2999506593,-0.1005669311,0.0460508876,-0.0921841711,0.0475017801,-0.2061013728,0.1462436318,0.2292242646,0.4124716222,0.0708965287,-0.0468711182,-0.1285553575,0.2888346314,-0.4000198841,0.3993601203,0.0945593044,0.1767641157,0.0824467242,0.1948525757,-0.2577297091,-0.216028437,0.1403772235,-0.0178153664,0.3704651594,-0.3278236985,-0.0225527659,0.0600105189,-0.147668913,0.2092429698,0.100826174,-0.0661532357,-0.0986372307,-0.0651670992,0.1538115144,0.0807398111,0.072188057,0.0130378306,0.39847067,-0.4098277986,-0.3508966863,0.212117821,0.1171258315,0.0153381256,0.2029845268,-0.0538278073,0.0407218002,0.0347627252,-0.2113345414,0.0822405964,0.2176874727,-0.3976355791,0.1352212727,0.0903974399,-0.0065130061,0.0280604772,0.3131864667,0.0237548687,-0.0954862088,-0.0735831931,0.1023028791,0.2557310462,-0.4735104144,-0.0518357232,0.3816914558,-0.2694062889,0.1285037845,0.3236992359,0.1077451929,0.3082872927,0.075988397,-0.148661539,0.0952469632,-0.307705462,0.2454701364,0.1432372332,0.370496273,-0.2405254245,0.1277182251,0.1459761709,-0.0762236789,0.2074323595,0.0497043431,0.2443204522,0.2103810757,0.0827849805,0.3224075139,0.000585471,0.1528162658,0.4286256135,-0.1078520715,0.0441781804,0.2797578275,-0.1247082725,0.1938396692,-0.1352660209,-0.1123760492,0.1474460214,-0.089339599,0.1584052593,0.1259023398,-0.2418941259,-0.1280555427,-0.0236354787,-0.0601814762,-0.0971800834,0.2889662981,-0.0741620883,0.0531001426,-0.1978666484,0.5029247999,0.0863771811,0.2710227072,-0.1500816792,-0.3748132288,0.0151136471,0.0625861511,0.1371517628,0.5981580615,0.0585292913,0.0340734161,-0.0815232098,0.0144750271,-0.2162017822,0.02858462,0.0467494354,0.2041559219,-0.1509693861,-0.105228439,-0.0115822712,0.1715223938,-0.0447368696,0.0386967734,0.1964152455,-0.143628329,-0.7924929261,0.3506558537,0.2331953794,-0.3567838967,0.1086480021,-0.0057643466,-0.0246921629,-0.0635941327,0.5445169806,-0.0859044716,-0.1150617078,-0.018531153,0.1519895196,-0.1441807598,0.1815050393,0.3420636356,0.1265690029,-0.1513135135,-0.0463804044,-0.5696834326,-0.1018697992,-0.2204755247,0.1026959047,0.4844156206,0.0441273861,-0.0674322546,0.1358661801,0.0070429756,-0.2546031475,-0.0132011808,0.2133856118,-0.3235017359,-0.0227958076,-0.006837951,-0.1383009702,-0.1080934778,-0.3611944318,0.1588810086,-0.1554007232,-0.0089335162,-0.0340313613,0.0469430313,0.3325405717,0.1585762352,0.0870804191,0.0702105612,0.2880370617,0.1809135824,0.071149528,0.1565506607,0.2594001889,-0.2120538652,0.3363725841,0.3618619144,-0.1224530339,-0.1858343333,-0.1522149593,0.2615294456,0.2256315947,-0.1041049287,0.0013002714,-0.2095726877,-0.1042670161,0.0674142987,0.0819423422,0.0823079124,0.2622485161,-0.2664477527,0.1350756586,-0.210916549,-0.135509789,0.1331353784,-0.0722148493,-0.0630252063,-0.3058522642,0.26320526,0.0985742509,-0.2345578372,-0.3781541288,-0.1496791393,0.4653114974,-0.169310689,0.1591654122,0.5180373788,0.076300174,-0.0000087746,-0.1389848441,-0.0060039107,0.0172109772,0.0925397351,0.2196024507,-0.2579988539]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2514","title":"Can datasets remove duplicated rows?","comments":"Yes, I'd like to work on this feature once I'm done with #2500, but first I have to do some research, and see if the implementation wouldn't be too complex.\r\n\r\nIn the meantime, maybe [this lib](https:\/\/github.com\/TomScheffers\/pyarrow_ops) can help. However, note that this lib operates directly on pyarrow tables and relies only on `hash` to find duplicates (e.g. `-1` and `-2` have the same hash in Python 3, so this lib will treat them as duplicates), which doesn't make much sense.","body":"**Is your feature request related to a problem? Please describe.**\r\ni find myself more and more relying on datasets just to do all the preprocessing. One thing however, for removing duplicated rows, I couldn't find out how and am always converting datasets to pandas to do that..\r\n\r\n\r\n**Describe the solution you'd like**\r\nhave a functionality of \" remove duplicated rows\"\r\n\r\n**Describe alternatives you've considered**\r\nconvert dataset to pandas, remove duplicate, and convert back...\r\n\r\n\r\n**Additional context**\r\nno","comment_length":80,"text":"Can datasets remove duplicated rows?\n**Is your feature request related to a problem? Please describe.**\r\ni find myself more and more relying on datasets just to do all the preprocessing. One thing however, for removing duplicated rows, I couldn't find out how and am always converting datasets to pandas to do that..\r\n\r\n\r\n**Describe the solution you'd like**\r\nhave a functionality of \" remove duplicated rows\"\r\n\r\n**Describe alternatives you've considered**\r\nconvert dataset to pandas, remove duplicate, and convert back...\r\n\r\n\r\n**Additional context**\r\nno\nYes, I'd like to work on this feature once I'm done with #2500, but first I have to do some research, and see if the implementation wouldn't be too complex.\r\n\r\nIn the meantime, maybe [this lib](https:\/\/github.com\/TomScheffers\/pyarrow_ops) can help. However, note that this lib operates directly on pyarrow tables and relies only on `hash` to find duplicates (e.g. `-1` and `-2` have the same hash in Python 3, so this lib will treat them as duplicates), which doesn't make much sense.","embeddings":[0.104411073,0.0183886345,-0.1353909671,0.094954513,0.0331351422,0.1407132298,0.3450481594,0.2536354959,-0.6958328485,0.0018973832,-0.0936520845,0.3721427321,0.0341928788,0.1650433242,0.1942412257,0.0286316872,0.1286832243,0.2465856373,-0.1124954522,0.116292797,-0.1895906627,-0.2231096476,-0.3207571507,0.0351020806,-0.3403733075,0.0134709924,-0.3700193763,-0.0473617837,-0.2080128342,-0.2208465785,0.2427178174,0.5297170877,0.0183263551,0.2103604227,-0.000106261,-0.284894079,0.0180516597,0.1628575921,-0.3009908497,0.0122177238,-0.0974296927,-0.1236533374,0.1503576785,-0.0711361691,0.0399264731,-0.488966018,-0.2355798781,0.1284597963,0.1914304793,0.1222232431,0.2542186677,0.2170831412,-0.0721717402,0.1295604408,0.281085819,0.3251201808,-0.1984840035,0.4407833815,0.4495327473,-0.1497725546,0.3157269955,0.0626316667,-0.1624632478,0.145337224,0.1373353004,0.0009987081,-0.1338930428,-0.0835928544,0.0399319045,0.2834547162,0.5527540445,-0.6672785282,-0.3417749703,-0.3500338495,0.4291492999,-0.2890269756,-0.0532083958,0.0705096722,-0.0623901561,0.1691336781,-0.4275589287,-0.2888962924,-0.1135681942,-0.0553033948,-0.0859845281,-0.2437136024,0.2126117796,0.0046139881,0.1512090862,-0.0223211814,0.4632273018,-0.0744324103,0.2205946892,0.4448759556,-0.4444244206,-0.2212748528,-0.1027680039,0.1373488903,0.3310642242,0.1832380444,0.2757816017,0.0348050818,0.0536748171,0.2114180028,0.2253226489,-0.0003008748,-0.0716260001,0.1072478965,-0.0484820157,-0.171483025,-0.1903428733,0.1094561592,0.068816565,0.1245209798,0.422583878,-0.0973042846,0.2560529709,-0.1723722667,-0.3565198481,0.0375062786,-0.519122541,-0.2063179165,-0.0685030967,-0.1372722834,0.1091361493,-0.2522799373,-0.1279276907,0.0959540755,-0.3302261531,-0.2666727901,-0.0282258205,0.0720546618,-0.1864466667,-0.0160293803,0.3545638025,-0.2653107345,-0.1158275455,0.2977137864,0.1664303243,0.0528772399,0.2804608941,-0.0771169588,-0.0925080478,0.002487794,-0.1493972689,-0.2366427034,-0.1870187968,0.0621966645,-0.1377648115,0.5223926902,0.1237040982,-0.0923645198,-0.12273027,0.2812554836,-0.1491717845,-0.1853643209,-0.1511612087,0.1568148285,-0.0947597846,0.0086555341,0.1222714037,-0.1431024671,0.0737311468,-0.3464638591,-0.1126491502,0.06481307,-0.4670024514,-0.0562972575,-0.0684908032,-0.279932797,-0.1742832512,0.3257155716,-0.0579803474,0.1904354542,0.1309428364,0.097457625,0.0092130182,0.0978478566,-0.2693904638,0.030833723,0.0046320348,0.2481786311,-0.1455245912,0.2779869139,0.3345446885,-0.0485477149,0.0061093536,0.0208810754,-0.22293441,-0.401534915,-0.1754588336,-0.312045902,0.24020724,0.232738018,0.0118307928,-0.1296452135,0.1328344941,-0.3675982654,0.3119074404,-0.1483109146,-0.1385402828,-0.1938741207,0.4634333253,0.0824154094,0.1329257041,-0.3347445428,-0.4417125583,0.1908194274,-0.3296414018,-0.0624991357,-0.0673098788,-0.6913440228,0.0400846787,0.0445210971,-0.1122875512,0.2088786364,0.1710991412,-0.465197593,0.1041989923,-0.2003914863,-0.3963654041,0.1534633636,-0.22867468,0.214286834,-0.0842430815,0.489299655,0.0366029888,-0.1764849275,0.0690182224,0.4056182504,0.0476470105,0.0649266541,0.199491933,0.2082214355,-0.1004643589,-0.09535373,-0.0311935097,0.2300293446,0.1409668475,0.0056995028,-0.0579594113,0.1789648384,0.0268166959,0.0631047487,-0.3733065426,0.4113070071,-0.3832029104,-0.1537827253,0.0186023284,-0.1228374243,-0.0743722618,-0.150757432,0.1481102556,-0.1434396803,0.1585226655,0.3084122539,-0.1017345488,0.3747660816,-0.3540122211,0.2607422471,-0.1475556642,-0.277892679,0.1260588467,0.007866648,0.0648228303,-0.0456085727,0.2205125839,0.1937395185,0.2955671549,0.233207047,0.1198237985,0.2788223624,0.0400637686,0.0992880315,0.0858443528,-0.1198733002,0.0092498967,-0.1813764274,0.0222154818,0.2317917347,-0.362688899,0.143701449,0.2026194632,0.2773590982,-0.2206283063,-0.191067934,-0.5655443072,-0.4531567693,-0.0762024745,0.0453591384,-0.1780557483,0.0915171355,0.5150851011,0.1038604528,0.013792702,0.1103627086,-0.0746416897,0.3941479921,-0.2792817056,-0.0394747742,-0.1317638457,-0.1765998751,0.0203454494,0.2772711813,0.1195707992,0.1093560979,0.5857251883,0.0744352937,0.0695963129,-0.6109545827,-0.2983681858,-0.0311367474,0.0372670479,0.1495423168,0.085422568,-0.027060302,-0.3043860197,0.0325498171,-0.0417546108,-0.0985049233,-0.1936568767,-0.1042864472,0.1357042491,0.2669232488,-0.1313287169,-0.2455844283,-0.1122041568,-0.005250806,0.3367560804,-0.1136648431,0.162663877,-0.3023939133,0.016721664,-0.1321528405,-0.0836809278,-0.1281010956,-0.2123140395,-0.2869140804,0.0213606544,0.0162554774,-0.1465837061,-0.1533123404,0.1851413399,-0.200535804,0.4018159509,0.0114754131,-0.311875701,0.153409034,0.5076434016,0.0489151403,0.0520673953,0.35602355,0.3535315394,-0.0741010532,0.0246507879,0.0310232937,-0.1591763198,0.1633944958,0.0025987814,0.0769772753,0.138202548,0.2860575914,0.4721869826,0.2066885829,-0.095202677,0.1265106648,-0.0693728775,0.2181539088,0.0455579385,-0.3102408946,-0.1662701964,-0.3074214756,0.1874792874,0.4478248358,0.1748810112,-0.2337745726,0.1441046,-0.233210966,-0.5569915771,-0.2196409851,-0.0079382285,-0.1678031981,0.3260240257,0.0846034884,-0.3632075191,0.0090639535,-0.1497919112,-0.0940444171,-0.1068408415,0.261523217,-0.2522053421,-0.151019901,-0.4240190685,-0.1247038543,0.2552506924,0.1949155182,-0.0751071423,0.2008883655,0.1826677471,0.2519548237,0.0984337553,0.6468613744,-0.1492335051,-0.2107220441,-0.0603857301,0.0109381489,-0.0572135001,0.0634213388,-0.0763694122,0.0660847425,-0.0949915275,-0.1192447692,-0.199976936,-0.0821864083,0.5210831165,0.391253829,-0.1593157351,-0.3596802056,0.03580359,-0.3132453561,-0.2835783362,0.1139167398,0.0217585713,-0.0557770841,-0.0976584256,0.0525503531,-0.0164303463,0.0300415102,0.3727796972,-0.0867574289,-0.0055881594,-0.0923462883,0.2851485908,0.0504658334,-0.2362451255,0.5697203875,0.4987100959,-0.1559572369,-0.2272206396,-0.5975430012,-0.4074410796,0.4083680511,0.0544857644,0.0111147603,0.48163414,-0.2928210199,-0.0611238368,0.0185560063,-0.491501689,0.1985459626,-0.148378849,-0.2121211141,-0.2890119851,0.2951412499,0.2490932494,-0.1999496222,0.0702200532,0.2622963488,-0.1759204715,0.1031376794,0.0268730782,0.8997114301,-0.1537866145,0.0391393453,-0.1291898638,-0.2764852047,0.3032194972,-0.0598264672,0.0454575121,-0.1890165806,-0.2673981488,-0.0778649524,0.0925763622,0.2458872199,0.4782503545,0.112997815,0.1543112844,-0.131753996,0.1419727355,0.3370570242,0.2129842788,0.0325279348,0.1623698175,-0.169706732,0.1479340941,0.1216152161,-0.4452348948,-0.1877203882,0.1676511914,0.3555828035,-0.0666628256,0.0445275307,-0.3174158633,-0.0347608253,0.2446353883,-0.0071866554,-0.3055079579,-0.0170846581,-0.087399438,0.0232436359,-0.1431647539,0.0386561826,-0.0224848762,0.4205991924,0.4071325362,0.029815957,-0.084846355,0.2570121884,0.0094526829,-0.2883405983,0.0997741818,0.3497155011,-0.3559153974,-0.2147188783,0.159134537,-0.2184224725,-0.1804961413,0.1551721096,-0.0712405369,-0.0580587015,-0.3316576481,0.1704318374,0.1180439964,-0.0996863768,0.6222676635,-0.145157218,0.0775386021,-0.2983397543,0.0324796848,0.511844039,-0.2031547129,-0.0017203199,0.1107382998,-0.3159990609,-0.1827511489,0.206024915,-0.0038199304,-0.344048351,0.1151198,-0.0865317956,0.4295632243,0.239229545,-0.2274767011,-0.0982738212,0.0113092475,-0.0716353804,0.0187306181,-0.1199152023,0.1719378978,0.492649436,0.2791413665,0.1893620491,0.2254906446,-0.0105365561,0.1215113029,-0.3703795671,0.4481422007,0.1095800698,0.3079470992,-0.0832728371,0.1583911628,-0.295414567,-0.3457027376,0.1549717486,0.0201544333,0.3334080577,-0.3354876637,-0.1186097935,0.0778183341,-0.1121973693,0.1648175865,0.1211891845,0.0708630309,0.0310243648,-0.00644805,0.1450162232,0.1954202205,0.1516328007,0.0219904762,0.6954905391,-0.3684903979,-0.3162457943,0.0467092656,-0.1004607379,-0.0085438928,0.2305114716,0.007272928,-0.0876678899,0.1252840608,0.0593227483,-0.0296628308,0.1654496491,-0.4229683578,0.2478306144,-0.0688032806,-0.1274454296,-0.1567022651,0.5310279727,-0.0192132425,-0.1377566308,-0.1361900568,0.1865355819,0.2113353163,-0.5376392007,-0.0772432983,0.3420599103,-0.1656554341,0.04846121,0.3049323559,0.0320758745,0.2197640687,-0.0184644759,-0.0835448727,0.2788001299,-0.1496606469,0.1379196197,0.2400130779,0.2825613618,-0.2172234654,0.2884019315,0.0495562032,-0.053860765,0.1368466616,-0.1109046265,0.1620570123,0.0565388203,0.0129835131,0.2799027264,0.0169338323,0.0616242699,0.2446558923,0.0434483886,-0.0971763059,0.3055136204,0.1565429419,0.3416389823,-0.0264828745,0.0217285044,0.2720403969,0.0625353754,0.1540866941,-0.0740801841,-0.2136196792,-0.0782180801,-0.0646037534,-0.1890469193,-0.3106637299,0.3977768123,0.106549643,0.1259816438,-0.3896438479,0.3826603293,-0.0735544413,0.2179976851,-0.0860806778,-0.3578412831,0.1943039745,-0.0324578919,0.0880389586,0.4262210727,0.2488001138,-0.0439271852,-0.2069272399,0.0492283069,-0.4595752656,0.1220458895,0.0760370716,0.131810993,-0.0749094486,-0.0174437072,-0.1294480562,0.1688018143,-0.0743670166,0.106747739,0.229067564,-0.1014801115,-0.7163110375,0.1947499216,0.2168847919,-0.437205106,0.2104262561,-0.0665780529,-0.0675297007,-0.0037529124,0.4338000417,0.1405693442,-0.1521219164,-0.0515961163,0.1071064398,-0.0179619696,0.2052426785,0.3779183924,0.1835570037,-0.2155329287,0.0132900858,-0.5778458118,-0.0691288114,-0.4158000648,0.0295880865,0.5652524233,-0.0484474339,-0.0484683327,0.3098572493,0.0799080729,0.0138322171,-0.1534745693,0.2666604221,-0.3007995188,-0.0290298127,-0.065478079,-0.1326021701,-0.210774824,-0.2510816455,0.1813970804,-0.3836818635,-0.0175374001,-0.241615817,-0.056613598,0.4530106783,0.1486594677,0.2132157087,0.0342444181,0.405618906,0.2609458864,0.0636656806,0.185246855,0.1711768359,-0.2430460006,0.4706923962,0.1666073948,-0.1785307378,-0.0617745034,-0.3375510573,0.2445050031,0.2603080869,-0.1194974333,0.0688638166,-0.1194403246,-0.1358754337,0.1132173911,0.0888071805,-0.1549826115,0.1614748389,-0.248197332,0.3061462939,-0.226302743,-0.3050387204,0.2326824665,-0.2734937072,-0.034901049,-0.2624465227,0.1766578704,-0.064300783,-0.2301067859,-0.365023464,-0.1018805504,0.4784152508,-0.0784607977,0.1206112206,0.4608331025,0.033114057,-0.0399206579,-0.0932518691,-0.0537031777,-0.0345110744,0.1175374389,0.1016300395,-0.3563200235]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2514","title":"Can datasets remove duplicated rows?","comments":"> Hi ! For now this is probably the best option.\r\n> We might add a feature like this in the feature as well.\r\n> \r\n> Do you know any deduplication method that works on arbitrary big datasets without filling up RAM ?\r\n> Otherwise we can have do the deduplication in memory like pandas but I feel like this is going to be limiting for some cases\r\n\r\nGreat if this is can be done. Thanks!!\r\n\r\nNot sure if you are asking me. In any case I don't know of any unfortunately :( in practice if data is really large we normally do it with spark (only for info. I understand this is not useful in developing this library..)","body":"**Is your feature request related to a problem? Please describe.**\r\ni find myself more and more relying on datasets just to do all the preprocessing. One thing however, for removing duplicated rows, I couldn't find out how and am always converting datasets to pandas to do that..\r\n\r\n\r\n**Describe the solution you'd like**\r\nhave a functionality of \" remove duplicated rows\"\r\n\r\n**Describe alternatives you've considered**\r\nconvert dataset to pandas, remove duplicate, and convert back...\r\n\r\n\r\n**Additional context**\r\nno","comment_length":119,"text":"Can datasets remove duplicated rows?\n**Is your feature request related to a problem? Please describe.**\r\ni find myself more and more relying on datasets just to do all the preprocessing. One thing however, for removing duplicated rows, I couldn't find out how and am always converting datasets to pandas to do that..\r\n\r\n\r\n**Describe the solution you'd like**\r\nhave a functionality of \" remove duplicated rows\"\r\n\r\n**Describe alternatives you've considered**\r\nconvert dataset to pandas, remove duplicate, and convert back...\r\n\r\n\r\n**Additional context**\r\nno\n> Hi ! For now this is probably the best option.\r\n> We might add a feature like this in the feature as well.\r\n> \r\n> Do you know any deduplication method that works on arbitrary big datasets without filling up RAM ?\r\n> Otherwise we can have do the deduplication in memory like pandas but I feel like this is going to be limiting for some cases\r\n\r\nGreat if this is can be done. Thanks!!\r\n\r\nNot sure if you are asking me. In any case I don't know of any unfortunately :( in practice if data is really large we normally do it with spark (only for info. I understand this is not useful in developing this library..)","embeddings":[0.1012984961,-0.1910532117,-0.1392705292,0.2864695787,0.0751885027,0.2036610991,0.2715580463,0.0986695886,-0.3335005939,-0.0163339805,-0.00081,0.2826299071,0.0401865058,0.1977713555,-0.0249465518,-0.0297985617,0.1026341692,0.13065283,-0.1036721319,0.0501254201,-0.0256591607,-0.1465695798,-0.270499438,-0.1782128066,-0.4471103847,-0.061712347,-0.4829657078,-0.0588315502,-0.1587790251,0.0186243448,0.2496173531,0.5069915652,0.0081406729,0.3449687064,-0.0001065691,-0.4029147923,-0.0071389847,0.1132893115,-0.1967679709,0.0577062555,-0.2064548135,-0.0378361829,0.0811134204,-0.0923915878,0.0455432497,-0.2710401416,-0.309235394,-0.2610732615,0.1535585672,0.0755504593,0.2574928105,-0.0642401427,-0.2274297774,0.0427543931,0.1963112056,0.5021438599,-0.140786618,0.2988500893,0.5304868817,-0.0445952229,0.1433535367,0.0672994182,-0.1113487929,0.1386645436,0.290704757,-0.1952319741,-0.2926008701,-0.1928344518,0.20410797,0.3246283233,0.8173855543,-0.491063118,-0.2823204696,-0.1588136703,0.4048905373,0.0392115638,-0.0883467644,0.0993426442,-0.0085499687,0.0695747882,-0.3531998694,-0.5799795389,-0.1640008539,-0.1828123778,0.0733937472,-0.4417580664,0.0995519087,-0.0448228754,0.1841219217,-0.0781926736,0.3813770711,-0.1632514894,0.1593037695,0.4425401986,-0.4141801298,-0.3620016873,-0.2003977299,-0.0943413004,0.3296689987,0.1987353414,0.3466756046,-0.0675841346,-0.1132863685,0.2721816897,0.3223301768,0.0195970554,-0.1558397412,0.011317079,0.1128345132,-0.1993379742,-0.2604490221,-0.0415460132,-0.0300454013,0.2280072421,0.3002118468,-0.0754928663,0.0775921121,-0.3900251091,-0.2964450717,0.0931489766,-0.2986606956,-0.1495645493,-0.0994157419,-0.1295012385,0.086814411,-0.1445708573,-0.0661547258,-0.1551336497,-0.2972076833,-0.227149263,-0.0256302543,0.1597307473,-0.223503679,-0.0066864784,0.3262166381,-0.2521714568,-0.0575285591,0.2420903593,0.1999292225,-0.1301860511,0.341953516,-0.284827441,0.0834751949,-0.0906566679,-0.2567923665,-0.2523036003,-0.2064595371,0.1417122781,-0.0916771069,0.4854812622,0.1296812594,-0.0932272449,-0.1475782394,0.2590628862,-0.0467090011,-0.1065523401,-0.27303496,0.2472009212,-0.1362901628,0.0149164218,0.1113040894,-0.0092505803,0.113448061,-0.2967493236,-0.0482333116,0.2049471736,-0.5000967979,0.0429928601,-0.1776574552,-0.294242382,-0.0986089185,0.4649655819,-0.0501374379,0.0564437509,0.0138209406,-0.0734155551,0.0800505951,0.0723868832,-0.164642632,0.3385246396,0.0167544726,0.1923589408,-0.140732497,0.4406460524,0.3836031556,-0.0312123038,0.1031772122,0.0156770442,-0.2498232871,-0.3883444071,-0.234484151,-0.4616602659,0.2157368213,0.0939397588,0.1487682909,-0.0323660076,0.1957757473,-0.0963867307,0.1839366704,-0.0687402487,-0.0946717411,-0.0008339873,0.448630482,0.1039017588,0.1751072258,-0.3972981274,-0.3710746765,0.0808521435,-0.1925429106,-0.1053730324,0.2508815825,-0.65753901,0.1710532457,-0.1175081357,-0.0220814086,0.1245761886,0.1292590648,-0.4144191146,0.0347934216,-0.3078157008,-0.2813328803,0.2298516035,-0.3113639057,0.0324754938,-0.2253491729,0.4157766104,0.1745212674,-0.0179054551,0.1903279573,0.3661206067,-0.0345994048,0.0753782913,-0.0093812048,0.1028031856,0.1476318687,-0.0036964882,0.187470898,0.2373096645,0.2151127607,-0.073116295,0.0110310037,0.3312472701,0.1400051415,-0.0228340421,-0.4782250226,0.1685675234,-0.5081922412,-0.1461952925,0.0462404415,-0.0886555314,0.0201870482,-0.1640032381,0.1715242714,-0.1282509118,0.1904507577,0.1697271615,-0.1296025068,0.3815738857,-0.4475439489,0.0641821921,-0.1406338811,-0.1185539067,0.03778027,0.0418273583,-0.018564444,-0.1150999144,0.3880048096,0.3194039762,0.3859788179,0.2076179981,0.162846908,0.2141689062,0.3135331571,0.0956354737,0.1016093567,-0.1003510654,0.0071907956,-0.1843217164,-0.1023382246,0.2157649696,-0.3473798633,0.0111862263,0.2527470589,0.2058586031,-0.0462606624,-0.2968865633,-0.5272918344,-0.6229697466,-0.10584905,0.2558296025,-0.17891334,0.1530619711,0.425755769,0.1220294461,0.1268965155,-0.0463796481,-0.1135757491,0.6114587784,-0.2947301567,-0.1177484468,-0.0578323454,-0.0651275218,-0.0194145516,0.2763374448,0.189846918,0.3241818845,0.7310884595,0.2817598581,0.1069163233,-0.6828422546,-0.3100990951,-0.1548416018,0.0220710784,0.2122166306,0.1046079472,0.0073591014,-0.1370043904,0.0865820646,-0.1891391426,-0.0357156247,-0.1473559737,-0.2665337026,0.2026568502,0.1801410615,-0.1981701106,-0.276049763,-0.00624335,-0.2233166397,0.345307678,-0.2187124044,0.1618303508,-0.4644885957,-0.0877307281,-0.070093058,0.179972589,0.0134258084,-0.2804414332,-0.2759269178,-0.0374962837,0.0590817742,-0.2052407712,-0.0942788497,0.2906415164,-0.0165990181,0.3591819108,-0.2501959801,-0.2154522091,0.0508950837,0.1491317302,0.0574434325,-0.0064027566,0.2411807626,0.2258652598,-0.0641691908,0.0476743691,-0.1351560056,-0.1079881191,0.0796849057,0.0920585915,0.2104487419,0.2346976846,0.1868864745,0.3986666799,0.4363937378,-0.0304672923,0.1008327827,0.1341678947,0.131381765,-0.0042360174,-0.3950912058,-0.0240605231,-0.2811253667,0.0778106824,0.4160493612,0.172676459,-0.2761066556,0.0108274464,-0.4355359972,-0.1743421406,-0.2066577375,-0.083026208,-0.5281805396,0.2518496215,-0.0584769994,-0.2659347057,0.0218992438,-0.0675447285,-0.0927630588,0.0153090917,0.3267201781,-0.2318802327,-0.1968442798,-0.3175489306,-0.1894106269,0.2490177304,0.1273405403,0.1260680407,0.0645548478,0.2862904668,0.2849059701,0.0944534987,0.5072616935,-0.1550367326,-0.3747442365,-0.0297642574,-0.1918331087,0.1423603445,0.1936911643,-0.0345425867,0.1590632349,0.0462310202,0.0104218153,-0.2269629091,-0.2246122211,0.3740518093,0.4223040044,-0.204880625,-0.3847651482,0.2303045094,-0.2099769861,-0.3856249452,0.1169318855,-0.0015919227,-0.1818374246,-0.0545131713,-0.2308936566,-0.0752632841,-0.0534932725,0.5534107685,-0.2494471371,-0.045238141,-0.161875993,0.4126777053,-0.0461908579,-0.0647723302,0.5135988593,0.4399630427,-0.2340794951,-0.0955068022,-0.3327074051,-0.4687212408,0.4154412448,0.0632594153,-0.0105449017,0.6419352293,-0.2069550306,-0.1050460115,-0.1951548755,-0.3630766571,0.1463473737,-0.0616196133,-0.2987641096,-0.4916356206,0.1742487103,0.2544904053,-0.1544753611,0.40540573,0.2026758939,-0.1384014487,0.0683059171,0.0626354069,0.8880200982,-0.1315444261,-0.0210646931,-0.5048201084,-0.2182020396,0.3203296065,0.1101237908,0.0369021781,-0.153300181,-0.340264976,-0.0973493084,0.0867110714,0.2498960346,0.3294863999,0.0319349021,0.27884534,-0.0634566993,-0.0170542151,0.2026633769,0.4478130639,0.0420329534,0.1184312031,0.0272976272,0.1520118564,0.0843287557,-0.3738181591,-0.1554313004,0.1178789884,0.4924117327,-0.0143024689,-0.1616538912,-0.3362328708,-0.0661118478,0.3038235307,-0.1971484721,-0.1846030056,0.2987505794,-0.175671488,0.0282932371,0.1860212982,-0.0497364402,-0.0338161141,0.3742715418,0.3428649306,-0.0448880307,-0.0347160101,0.3155068457,-0.0240450297,-0.1775733829,0.0328864679,0.2792313397,-0.430300653,-0.202095598,0.106995374,-0.1454862654,-0.1505751908,-0.081698969,0.0012094292,-0.2756042182,-0.2018074393,0.1682783067,0.3091473579,-0.0842484459,0.6608366966,0.0679012313,0.0843944922,-0.3240249753,0.0801819712,0.6168262362,-0.3480920792,-0.0158459675,0.0764677674,-0.3589728773,-0.2794746459,0.0417131223,-0.1507000029,-0.2133918256,0.2705877721,-0.2612151802,0.2985774875,0.2992025912,-0.304817766,-0.068266809,0.119234778,-0.0829762444,0.0452655554,-0.3097729981,0.1152445152,0.1627223194,0.4415406883,0.1304177046,-0.0959677175,-0.1138584837,0.3005700111,-0.3438500762,0.4714667201,0.0688479394,0.190493539,0.078298457,0.2368115485,-0.3214196563,-0.1811047196,0.0666387901,0.0794918463,0.3810458183,-0.2554201186,-0.0792583376,0.0959276482,-0.1776941419,0.1848970503,0.0697865933,-0.1565337777,-0.1252879798,-0.049668882,0.2326223999,0.0428486392,0.0793864951,-0.063546814,0.4134469032,-0.4536335766,-0.4365443289,0.258805722,0.1193438098,0.0316154845,0.207085982,-0.0543186367,0.0982479006,-0.0273258388,-0.2776968777,0.072271578,0.2129624039,-0.4306273758,0.2142936289,0.154484421,-0.0256421454,-0.0029884323,0.3490703106,0.0557062812,-0.0163676068,-0.0679793954,0.1378094256,0.1761341095,-0.4282858074,-0.0741737783,0.3166078329,-0.1787236333,0.012357071,0.2967842817,0.0964596793,0.3377549946,0.004239331,-0.1560533047,0.1886490136,-0.2589979768,0.2207680494,0.1989772767,0.2766292989,-0.2398596555,0.174119845,0.1463068575,-0.0420258865,0.2355798334,0.0415929519,0.2634004354,0.3026799262,0.0452104099,0.322791487,-0.0349332429,0.0739854127,0.44282493,0.0682911947,0.0580403134,0.2874363661,-0.0683395714,0.2384699583,-0.1309801936,-0.0871750265,0.2352374196,-0.0565091632,0.1907279342,0.1485865414,-0.2084195763,-0.1504638642,-0.0080811903,-0.1395138055,-0.0559516326,0.3512403369,-0.0156345516,0.109600313,-0.2197772712,0.5052939057,0.0266908891,0.2342055291,-0.1877643913,-0.3179320097,0.0719372928,0.0776499137,0.1111991853,0.4749700129,0.0632930025,0.0271709431,-0.0570736043,-0.0331966132,-0.2760207057,0.127746284,0.1048715711,0.3314171433,-0.1079455391,-0.0018262631,0.0300680362,0.0967409238,-0.0053437995,0.0893901959,0.1831608266,-0.0931034759,-0.7865310907,0.3493581414,0.2298716605,-0.3472646773,0.0225217491,0.0056214142,-0.0517576709,-0.1072738022,0.5509375334,-0.0814046562,-0.1437779814,-0.0513463877,0.1184336692,-0.2014709711,0.2428555787,0.3888446093,0.1531118453,-0.1822668761,-0.077073656,-0.607568264,-0.1002845839,-0.148525551,0.206498161,0.4242288768,0.0686706752,-0.0160482656,0.1010089889,-0.0390285663,-0.2488709837,0.0020462228,0.215596661,-0.3344157338,0.021590773,-0.1229387671,-0.1379566789,-0.1229776964,-0.3416821063,0.1283521503,-0.2221683562,-0.0732397884,-0.1079543903,-0.0276713353,0.3466143906,0.1421359926,0.1150746569,0.0356480405,0.3848189414,0.2009039223,0.0643572211,0.2148556411,0.2523736656,-0.1405177861,0.3137957156,0.3568800092,-0.0743014887,-0.1540361494,-0.0668986738,0.2339364737,0.3083657622,-0.2212506533,0.0382493064,-0.2708306313,-0.1847289354,0.0084305834,0.0594215579,0.0863314718,0.314329654,-0.2490262091,0.1305454671,-0.2227336913,-0.0613357723,0.2162169367,-0.14438501,-0.0399395339,-0.3423178792,0.2046395689,0.120177947,-0.2263674438,-0.4386417568,-0.1585230678,0.4443906546,-0.132495001,0.1383078396,0.493834585,0.0098453788,0.0560976341,-0.161819607,-0.1014936864,0.037148729,0.091886431,0.2818921804,-0.3022620976]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2513","title":"Corelation should be Correlation","comments":"Hi @colbym-MM, thanks for reporting. We are fixing it.","body":"https:\/\/github.com\/huggingface\/datasets\/blob\/0e87e1d053220e8ecddfa679bcd89a4c7bc5af62\/metrics\/matthews_correlation\/matthews_correlation.py#L66","comment_length":9,"text":"Corelation should be Correlation\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0e87e1d053220e8ecddfa679bcd89a4c7bc5af62\/metrics\/matthews_correlation\/matthews_correlation.py#L66\nHi @colbym-MM, thanks for reporting. We are fixing it.","embeddings":[-0.3292961717,-0.3376269042,-0.1706661433,-0.0578410365,0.1505800188,-0.1545663327,0.0919021443,0.1828321964,-0.1506545395,0.4139116406,-0.1254370213,0.2337972075,0.0946260616,0.2705773115,-0.3405186236,0.1704522967,-0.1515125185,0.4011961818,0.0886283144,0.0473572239,0.0919471532,0.3022273779,-0.0730139539,0.056887798,-0.2156841308,0.1171516851,0.2122066915,0.2140278816,-0.216630131,-0.0258198958,0.1656026244,0.1976089478,-0.4575228393,0.1343396157,-0.0000995074,-0.0885374695,0.4149936438,-0.057559859,0.0992352366,-0.0282627065,0.1173281297,-0.0253510214,0.2471728176,-0.1561224014,-0.2937161028,0.1506008357,-0.0436217189,0.1734566242,0.3515178263,0.0780397132,0.3917894661,0.5046311021,0.0543131977,-0.3481095731,0.1051909775,0.1573972702,-0.0022632552,0.1459787488,0.125581339,-0.107263796,-0.2919933498,0.1619914323,0.1300520748,0.2704963386,-0.0300154164,0.0279464126,0.0587813109,0.0690853596,0.1945026219,0.0129027581,0.2787200809,-0.0436203927,0.0986911878,0.068477802,-0.0021477512,-0.1722410768,0.1230184138,-0.0311761722,0.0279103909,-0.0129122995,0.0458019413,0.2705483139,-0.0754250363,-0.0200849995,-0.1587543786,0.3855138123,-0.3432259262,-0.1088962257,0.2216067761,0.0626837835,0.0474347882,0.3360876143,-0.1586731374,-0.2468543798,-0.0191912223,-0.2007490993,0.1437393725,0.1538912058,0.1234539002,0.075396724,0.0069377953,0.2382209599,0.2275977731,0.1746492237,0.1262753755,0.1092825681,-0.0173057076,0.0642350987,0.2356437594,0.538192749,-0.2131287158,-0.0517908596,0.2232370228,-0.3461720049,-0.0858395323,-0.0386758149,-0.0390928611,-0.2595512867,-0.4685721993,-0.1001967713,0.0726637095,-0.1037671417,-0.0028789393,0.2887077332,0.067470938,0.182682693,0.0376847945,0.1202108115,-0.1514618397,0.0544113256,-0.4064358771,0.1418107152,-0.2057699412,-0.0357072949,0.0706351995,-0.0381695405,0.4480188191,0.1491836458,0.4351722896,-0.0993616804,-0.0671776161,0.1580410153,0.1327570081,0.1822745949,-0.1197331697,-0.0085326191,0.0978068784,-0.2824820876,-0.1116479337,-0.1342002451,-0.3405846953,0.052019354,-0.0540718585,0.347823292,-0.0812616795,-0.0960405171,0.1378563493,0.1732455492,-0.0527976155,-0.0936228409,-0.0532270148,0.1271814704,-0.1324804127,-0.1557102352,0.2845958173,0.1044652089,-0.0945260525,-0.1604371816,0.2312751412,-0.1206151471,0.2370034754,0.1082669869,-0.1991478503,-0.0976468846,-0.1488047093,0.0089313481,-0.037537761,-0.3122423887,-0.1501235217,-0.2181313336,-0.2530890405,-0.0837352201,0.0563131198,-0.0255295243,0.0111075863,0.1281507462,0.1145193353,0.0112573178,0.0627909973,0.0380163305,-0.3519393504,-0.1651260853,-0.0293643363,0.0581582636,0.0121125001,0.0438074321,0.060369201,-0.0974661335,0.2201001197,-0.0745313689,-0.0115952129,0.1539877057,0.2172376662,-0.1046473384,0.039966803,0.0751061589,-0.0683691725,-0.0471652895,-0.245041579,0.4066057503,-0.0982310176,-0.1225503236,-0.4364647269,-0.2185674161,-0.0854689926,-0.1322464347,0.3992758393,-0.0855840966,-0.159053877,0.0550332963,0.0417085998,0.1062579677,-0.0165411495,-0.116516307,-0.0349055678,0.2500118017,-0.2497588247,-0.0776266605,0.4842419624,0.1465557069,0.1796839088,0.1505589336,-0.0587278977,0.1814627349,-0.1211158782,0.4120377898,0.3830243349,0.2302466333,-0.0789792836,-0.0964952037,-0.0365124755,-0.0195470434,0.0734725296,-0.1123859659,0.0880983621,0.4540244937,0.4042424858,-0.0188852269,-0.1181676462,-0.0892203078,0.1285244077,-0.1300586015,-0.021244863,-0.1861355603,0.1231796667,-0.0122515988,-0.1415732354,-0.0681419596,-0.104731068,0.2517956495,0.2845808864,0.0974572301,0.0777964592,0.0289025642,-0.1113610938,-0.3006308079,-0.1694089919,0.052379705,0.2986202836,0.3072963059,0.2396383137,0.017730521,-0.1412918568,-0.3497086167,0.1216825917,0.0238134339,0.0077010482,0.1201123744,0.3387916386,-0.2197735757,-0.580650866,0.2742229402,-0.2454506159,0.0573504865,-0.3054170012,-0.0504222214,-0.0743548051,0.2259108722,0.0835723132,0.0303082671,-0.2684071958,-0.426451087,0.1874294132,0.1456164718,-0.1623486429,0.0850396752,0.1057943106,0.3301292062,0.0053568427,0.0347080566,-0.3025407493,-0.3208661973,-0.1462107301,0.3086653948,-0.2024663836,0.1216423437,0.293099463,-0.2348075211,0.2353079319,-0.0438902974,-0.5915470123,0.0426348411,-0.0965304747,0.0788728893,0.2207683474,0.1497862488,-0.189637661,-0.3359104991,0.4096127748,-0.3585763574,-0.3123655021,-0.035835173,-0.0607865453,-0.0818791687,-0.2135239094,-0.1209826767,0.109921895,-0.2854331732,-0.0259187575,0.1455184072,-0.0387127548,0.1005999669,-0.1743835509,-0.1051727012,-0.4697806537,0.2045826167,-0.424823314,-0.3966568708,0.2957311273,-0.3227920532,-0.3549940884,-0.1006547883,0.2243953943,0.2269629538,-0.3671001792,0.0116610974,-0.5271276832,-0.0515589677,-0.0006136649,0.0660579726,0.1371136308,0.0990972668,0.0504495837,-0.4511801004,-0.24300237,-0.2186338753,0.3414193094,-0.1698427945,-0.0341988057,-0.3194920719,-0.022998644,-0.0010178142,0.2441064268,0.1062623039,-0.2657999098,0.2590441108,-0.2013694793,0.3848338723,-0.0315922201,-0.0053354879,0.0973603055,0.1885153204,0.167131424,0.5195607543,-0.0302164629,-0.1746045202,-0.3547121584,-0.0273829829,-0.2776458263,-0.2319584489,-0.2808958888,0.4092261791,-0.0312311444,0.1853342503,-0.0601319633,-0.0677493662,-0.1868161857,0.3654835224,0.208095938,-0.2893808484,-0.2290108055,-0.0570726357,0.1480111182,-0.4473717511,0.1799962819,-0.052799385,-0.0697319061,-0.1542254984,-0.2575937808,-0.0169250518,-0.115402393,0.3021960855,-0.0329326503,-0.0444001779,0.0994304121,0.065353781,-0.0470414944,0.0572773144,-0.3326043785,0.119954668,0.1322121024,0.4404166341,-0.346960336,-0.0597215742,0.1235483959,-0.0177620556,-0.0728026107,-0.1578844041,-0.2792785466,-0.3091052771,0.0549464859,-0.0714872405,0.1841746122,0.3038558066,-0.0650834888,-0.251660496,0.1364496797,-0.0810872093,0.1725251526,0.1035827547,0.3920295537,0.0384186879,0.0596650429,0.0478632674,0.2885206044,-0.2334754616,0.2419253588,0.0941255167,0.1935555637,-0.0936773866,-0.0957618132,0.1561101675,0.1605150104,-0.166220054,0.0603087582,-0.1293652356,0.3090915382,-0.2191896439,0.0487695336,0.3417219818,0.172318399,-0.0028428438,0.205143407,0.247327894,-0.0538804159,-0.0464797206,0.0625735372,0.3627257347,-0.1150885001,-0.1125370488,0.0057555023,0.8594093919,0.1122133508,0.0067850333,0.3753140569,-0.4103160501,0.2538357675,-0.0068897107,0.0683083683,-0.265263021,-0.4028229415,0.0476059541,-0.0810667127,0.0284373928,0.0346700773,-0.087436907,-0.0018179332,-0.1844551712,0.1456766427,0.0049741426,-0.1991991997,0.1868997961,0.0715739205,-0.1690132916,0.3258064091,0.0838793069,0.1275985837,-0.0000371781,-0.1028216854,-0.1527245492,-0.1837382615,0.0776218176,0.0607676208,-0.2124541104,0.0869275481,0.1126416847,-0.0258899108,-0.0212340038,-0.3442100585,0.3338688016,0.0639080852,-0.2232446671,0.349217236,-0.0726376623,0.042816177,0.1143795252,0.2152214646,-0.0990495086,-0.1779188216,-0.0397919379,-0.1529227793,0.0742282048,-0.1946935058,-0.0358738713,0.0237037726,-0.39396891,-0.2428716421,-0.0572103374,0.1517391652,0.0332170166,-0.2013222724,0.317415297,0.1014175415,-0.1403504163,0.1199078783,0.1175604686,-0.2633980513,-0.0762024894,0.5079191327,-0.1574812382,-0.1178847775,-0.0653247237,-0.1121834666,-0.3992550075,-0.2423877567,-0.311889112,0.1856356114,-0.302554816,0.0242527034,0.1119805574,-0.2537340224,0.1640127152,0.5327690244,0.0737736523,0.1512362063,-0.1010119095,-0.177132979,-0.28137061,0.0588484257,-0.100992769,0.1467886567,0.0942841992,0.1531427652,0.1203580946,-0.0105966572,-0.5328756571,-0.0549291112,-0.183627218,0.0773438662,0.0778465122,-0.4712958038,0.4865084291,0.1177954152,0.3442094326,0.0474275723,-0.2069716603,-0.4148698747,0.0914836675,0.058844693,-0.0794564039,0.261926055,0.0007671161,0.1297603548,-0.2733228803,-0.000263001,0.1727825701,0.0877765864,0.0964398906,0.0489532873,0.0305385999,-0.10144151,0.1751309037,-0.1429836452,0.1111987159,0.0768501982,-0.1007031426,-0.0748492926,-0.384224683,-0.1256182641,0.2406425923,0.3963770568,0.0134246824,0.1727063358,0.1377826482,0.0278805252,-0.4600376487,0.1864420921,0.631102562,0.0746934265,-0.0189875886,-0.1952060014,-0.0504480191,0.3721679747,-0.4618184566,-0.0447195545,0.0025689029,0.0673060194,0.2774090767,0.1117003337,0.2193528712,-0.183880344,0.0546791963,0.1881185472,0.4270968437,-0.3829347789,-0.0912683606,-0.2605382204,0.2489413917,-0.026019793,0.4396728575,0.0646709874,0.1503736526,-0.0807751715,0.0849535689,0.3871847689,0.1246459857,0.190002203,0.1306690872,0.1926693022,0.2154446542,0.2621257603,0.2197921723,-0.0906427354,0.1919233352,0.6540631652,-0.0739868507,-0.1945181936,-0.2902942002,0.0840907171,-0.2096895725,-0.201164037,0.0585786775,-0.2879781723,-0.2399896979,-0.2556109428,0.1858216971,0.4955319762,0.3170305192,-0.090657793,0.1757735908,-0.0129520549,-0.272259742,0.0889862254,0.4889948964,-0.080175668,-0.0192921441,0.1728345752,-0.068012625,0.1321111619,0.3719402552,0.373624146,-0.0335738324,0.1133715659,-0.0853638574,-0.0448617749,-0.1925208718,-0.1273449212,0.090228267,0.2171347737,-0.199460417,0.1968182772,0.3584393263,-0.389480412,-0.0331745744,0.0160434805,0.1634040922,0.0246772394,0.0300272368,-0.165926069,-0.074292019,-0.1798451245,-0.1448262185,-0.3577137291,-0.0081477808,-0.3336323798,-0.0816000849,0.044065915,-0.2588652968,0.2034748942,0.1353015751,0.3167354167,-0.1336336136,-0.0041867523,-0.0589890182,-0.2509634495,-0.3226000369,0.3704240322,0.2472877651,-0.1326949596,-0.1601875722,0.0784285888,0.1173668802,-0.0673658177,0.2513984442,0.130354926,-0.0191891585,-0.5142102838,-0.3321831822,0.17715168,-0.1053546146,-0.43567276,0.30919379,-0.0376268029,0.0469107665,-0.0667481273,0.3371299505,-0.1512961239,0.0418199897,-0.1065250561,-0.0222628023,-0.0026338107,0.0070706876,0.2498872578,-0.3134697676,-0.0079212291,-0.0317238085,-0.2907348275,-0.0563873947,0.111057207,-0.0246203542,0.1548642665,0.0490972511,0.0254107416,-0.046819523,0.1706926674,0.2125174254,-0.1365689188,-0.2830999792,0.2566475868,0.1062416732,0.0132618267,-0.0415478721,0.2145567685,0.0428060144,0.0077657616,-0.1306594759,-0.5940511823,0.4838270545,-0.2066214532,-0.0371202901,0.129665792,0.3020774126,0.1794313937,-0.0384643488,-0.3939996958,0.1185912937,0.0965014398,0.0798151642,-0.170629099,0.419403553,-0.2080658674,-0.0229191557,-0.0223028604,0.0044325055,0.4834254682,-0.1347420216,-0.1366482675,-0.072511062]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2512","title":"seqeval metric does not work with a recent version of sklearn: classification_report() got an unexpected keyword argument 'output_dict'","comments":"Sorry, I was using an old version of sequeval","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\nseqeval = load_metric(\"seqeval\")\r\nseqeval.compute(predictions=[['A']], references=[['A']])\r\n```\r\n\r\n## Expected results\r\nThe function computes a dict with metrics\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset, load_metric\r\n 2 seqeval = load_metric(\"seqeval\")\r\n----> 3 seqeval.compute(predictions=[['A']], references=[['A']])\r\n\r\n~\/p3\/lib\/python3.7\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/seqeval\/81eda1ff004361d4fa48754a446ec69bb7aa9cf4d14c7215f407d1475941c5ff\/seqeval.py in _compute(self, predictions, references, suffix)\r\n 95 \r\n 96 def _compute(self, predictions, references, suffix=False):\r\n---> 97 report = classification_report(y_true=references, y_pred=predictions, suffix=suffix, output_dict=True)\r\n 98 report.pop(\"macro avg\")\r\n 99 report.pop(\"weighted avg\")\r\n\r\nTypeError: classification_report() got an unexpected keyword argument 'output_dict'\r\n```\r\n\r\n## Environment info\r\nsklearn=0.24\r\ndatasets=1.1.3\r\n\r\n","comment_length":9,"text":"seqeval metric does not work with a recent version of sklearn: classification_report() got an unexpected keyword argument 'output_dict'\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, load_metric\r\nseqeval = load_metric(\"seqeval\")\r\nseqeval.compute(predictions=[['A']], references=[['A']])\r\n```\r\n\r\n## Expected results\r\nThe function computes a dict with metrics\r\n\r\n## Actual results\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset, load_metric\r\n 2 seqeval = load_metric(\"seqeval\")\r\n----> 3 seqeval.compute(predictions=[['A']], references=[['A']])\r\n\r\n~\/p3\/lib\/python3.7\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/seqeval\/81eda1ff004361d4fa48754a446ec69bb7aa9cf4d14c7215f407d1475941c5ff\/seqeval.py in _compute(self, predictions, references, suffix)\r\n 95 \r\n 96 def _compute(self, predictions, references, suffix=False):\r\n---> 97 report = classification_report(y_true=references, y_pred=predictions, suffix=suffix, output_dict=True)\r\n 98 report.pop(\"macro avg\")\r\n 99 report.pop(\"weighted avg\")\r\n\r\nTypeError: classification_report() got an unexpected keyword argument 'output_dict'\r\n```\r\n\r\n## Environment info\r\nsklearn=0.24\r\ndatasets=1.1.3\r\n\r\n\nSorry, I was using an old version of sequeval","embeddings":[-0.1199598908,-0.2244085222,0.0414448939,-0.0210395586,0.3008895516,-0.018107865,0.2096486986,0.272505343,0.2254312932,0.0090003349,-0.0891351923,0.6743949056,-0.1125626713,0.1897184104,0.0011524823,-0.0774690285,0.0738754794,-0.2512515485,-0.2874985635,-0.0117382417,-0.5155801177,0.2597082853,-0.261501044,0.2013555914,-0.1008395255,0.0041860887,0.0819644108,-0.2173564583,-0.1931759268,-0.3006435633,0.4675993025,0.0435256958,-0.027547976,0.3225607574,-0.0001143389,-0.1199739352,0.41058743,-0.1721962243,-0.3210989833,-0.2762172818,-0.4928049743,-0.1638211608,0.2805795372,-0.3551905751,0.1354936361,0.3459747136,-0.2022318989,-0.3604930043,0.2919411063,0.3537459373,0.2321998328,-0.0671875179,0.0811053216,-0.3575169146,-0.0471938923,0.202943027,-0.1858641058,0.6882337332,-0.3144969046,-0.108023569,-0.1342169642,0.3469319642,-0.0982129872,-0.1225802153,0.44681409,0.1983401179,0.3948453367,-0.2110654861,0.192485258,-0.1311587244,0.1994934827,-0.3334084153,-0.276391685,0.2208823413,-0.0495300479,-0.7980106473,0.1730573624,-0.0391287804,-0.106230706,-0.1801242232,-0.0287079494,-0.1543680727,-0.3006970286,0.0921578258,-0.2557089627,0.3476885855,-0.1333398521,0.1371735483,0.1580542177,0.0020462656,-0.0274058376,0.1732136011,-0.0167934895,0.0022677591,-0.4321724772,-0.0912935957,0.0777353942,-0.4126805961,0.0322047472,0.0445728041,-0.1278344393,-0.0855602771,0.4967467189,0.0321867131,0.0691018775,0.5994374752,0.3918285668,0.4976956844,0.0512913801,0.2842762768,-0.0988006294,-0.2111562043,0.0790367648,-0.1612646282,0.312191695,0.1304565519,0.3228036761,-0.2084973902,-0.5595058203,0.0938391536,-0.266682148,0.0163439885,-0.1471492201,0.1566410214,0.083661899,0.0791335627,0.407515198,0.4086063504,-0.1138298512,0.0859849378,-0.1538369805,0.3957912326,-0.2415429205,0.0624700598,-0.1019224599,0.0328995734,0.0050392416,0.0377854146,0.2566554546,-0.3578881621,-0.1814467609,0.0490449481,-0.1342267543,0.0024248697,-0.3974169493,0.1270609349,0.2331995368,-0.2914038599,-0.1244011074,-0.0475209616,-0.3739718497,-0.3082201183,-0.0138058271,0.2229057997,-0.1240622625,-0.0630577579,-0.0897705555,0.3628126383,0.0010541568,0.2599656284,0.0389467441,-0.2395419925,-0.5059835315,-0.1115854904,0.4242027402,0.1545285583,-0.4962057769,-0.2048662603,-0.2846322656,0.1264961213,0.0189071335,-0.0595998429,0.0760237128,0.3813437819,-0.0143646859,-0.1314715594,0.4109148681,-0.7786234617,-0.3335829973,-0.0450068042,0.0066421609,-0.2336003333,-0.0122737102,-0.1552322507,0.4917007983,0.1835813373,0.3566707373,0.188194707,0.0507529974,-0.2916214168,-0.3301068842,-0.1005038098,-0.098929584,-0.0448439494,0.4267694652,0.1067989767,0.0913034528,0.5078602433,-0.0150712794,0.1893948615,0.012478373,-0.1076579317,0.5086243749,-0.1667145044,0.0898858756,-0.1704851538,-0.0983734652,0.0394038297,-0.0747826621,0.1526193917,-0.2134158313,-0.0830920264,-0.7450507879,0.3173702359,-0.1238220185,-0.2590775788,0.2143210918,-0.2744357288,0.2286585718,0.3341432214,0.0964171067,-0.1750795394,0.0666370988,0.1215863079,0.039376121,0.1218235269,-0.075449042,-0.2003024369,-0.0046824133,0.2765694857,0.2461287081,0.2204276025,-0.1417785436,0.3119633496,0.0571223423,-0.1051256657,0.1412898749,0.0717190877,0.0225499962,-0.3556120098,0.116540499,0.1339532286,0.1533854604,0.0249734242,-0.0239719357,0.2949794233,0.107202515,0.2733209133,-0.1340610087,0.3818987012,-0.098966904,-0.0219692495,-0.1136927828,-0.163464129,0.1418019235,-0.1713226885,0.0239698775,-0.3419027627,0.2798279822,-0.1856464893,0.4627176225,-0.0849637613,0.2158216536,0.1342319697,-0.2826850116,0.0372908264,-0.1747515798,-0.0359456502,0.296372205,-0.0552583747,-0.1464033276,0.1597023159,-0.354581207,-0.0168268718,0.1866926998,0.208003059,-0.1206212416,0.1781004518,0.2640199661,-0.0769933164,-0.0638479441,0.0164928921,-0.1154930666,0.1587709486,-0.2206350863,0.3378016949,-0.0641868934,0.0416488349,-0.1735430658,0.1118142903,-0.1053266376,-0.5381045341,0.0039727595,-0.0119665368,0.0413140953,0.2563252747,0.032028608,0.0873490497,0.298686415,-0.1172850057,0.0102851791,-0.5094793439,-0.0406946354,0.054145921,-0.132036075,-0.1365925372,0.3365445733,-0.1747053266,0.1662852168,-0.1739983112,-0.4542059898,-0.0424694307,0.0421891548,0.6025862098,0.3151755929,-0.0668452904,-0.214071691,-0.2003279477,0.3947308362,-0.1458831429,-0.0356448367,0.3843100965,-0.1381612271,0.3708941042,-0.2651763558,-0.3710213006,0.0665883124,-0.3385000825,0.1411649883,-0.0057856818,-0.0859414488,0.3685806096,0.1107913032,0.2525849342,-0.0919366553,0.2174213231,-0.2371063083,0.3414668739,0.247801736,0.0020847339,-0.2837006152,0.0362863764,-0.1897434741,0.5101376176,-0.0436262451,-0.4228924215,-0.3919383585,0.0279339738,0.2604332268,0.0333855525,-0.1587833315,0.0857704878,-0.1365304142,-0.0210874099,-0.1484159976,-0.286231935,0.2784942389,0.0958003104,0.0214907788,-0.1448363811,0.6193232536,0.1862675846,0.5361302495,0.0165116396,0.180534333,-0.1455521137,0.1262241453,0.3097314835,-0.0656371564,-0.2948701978,-0.0254808385,-0.2603194714,-0.147184968,-0.0131903179,-0.0998831242,0.0712668449,-0.2242472023,0.0622938536,0.002420228,-0.3222862482,0.0292442665,-0.2747407854,0.1404457688,-0.0790386125,0.3364512026,-0.064847894,-0.139446497,0.1802809834,0.097216621,0.0172440968,0.0461419597,0.0240412392,-0.4767488837,-0.4174202979,0.2766228616,-0.1368886679,0.0136236073,-0.0833611786,-0.0706968978,0.0314275101,0.1593890339,0.2666400969,0.0109094176,-0.0163598452,0.3633792996,-0.179320097,-0.5540354848,-0.2128618211,-0.4287369847,-0.0478383936,0.1911234558,0.1756914556,-0.4851525426,0.1478601843,0.3641083241,0.021418184,-0.1499134004,-0.0083194152,-0.2122128755,-0.1758195609,-0.0820418671,0.1862450242,0.0494053364,0.1921239495,-0.0997816771,0.3237235546,0.0500117429,-0.2185509801,-0.0824417695,0.2117331326,0.2117700279,-0.2941465974,0.0129622286,0.0521280095,-0.1116219163,0.0089953132,0.1963147819,0.0445953161,-0.6646763086,0.1719569564,-0.2183655053,0.2885702252,0.0706637353,-0.2482901067,0.3202180564,-0.3153911233,-0.0645769387,-0.1737233251,0.0023091915,0.5551623106,0.1061853915,-0.1078872979,-0.129189238,0.3883296549,0.1036252156,-0.1476723254,-0.1173868105,-0.1508109421,-0.3060376048,0.262109369,0.2516682148,0.7765686512,0.0337273926,-0.1525421292,0.3891771734,-0.1211719587,0.7539755106,0.1251740605,0.1089027971,-0.1908221394,-0.0646778792,0.1617492139,-0.0222675949,0.0017578864,-0.0986767113,-0.4400956631,0.2965111434,-0.4147157073,0.1477283686,0.0048074555,-0.0511901043,0.2376457751,0.0928132311,-0.1946566701,0.1416923553,-0.2145565301,0.1369942725,-0.1467974037,-0.0032292011,-0.0211502966,-0.2307695597,-0.2551130652,0.0020744249,0.2640746832,0.2162388712,-0.0193479136,-0.4372326434,0.0564463176,0.0783843473,0.2675249577,0.1664368659,-0.1246011779,0.0787400082,-0.0885903761,0.1088358611,0.190386191,0.0172302332,0.1406732351,-0.0721005127,-0.111067757,0.2841851115,0.1086899489,-0.1075236127,0.0554817282,0.0727156401,0.0440467112,-0.3074341118,0.1338057071,-0.2373275012,0.0883553922,-0.2523770332,0.1378428191,0.0588278063,0.0197057705,0.1850012988,-0.0437927917,-0.2395344675,0.0208054036,0.0590178818,0.0831114054,0.0447984114,0.3323073983,-0.1679142118,-0.1958746612,-0.1399493963,-0.2654064,-0.2547517419,-0.5830376744,0.5187286735,0.0963176563,0.0982810259,-0.1397025734,0.3428414464,0.2692143917,-0.0586730503,-0.0461935923,-0.3135467768,0.0738030672,0.3446766436,-0.2497631758,-0.1634405553,-0.2997582257,0.1881363839,-0.0538357794,0.1101043522,-0.2928805053,-0.1423357427,-0.1525247991,0.0530216843,0.2297840416,0.0002586033,0.2047298402,0.1884487271,0.0858950019,0.0806569755,-0.0129922582,-0.1726670265,-0.3248640299,0.1010305658,-0.0336799733,0.0923588052,0.1562631726,0.1023021266,0.0005307294,-0.1299574822,-0.0901828334,0.2525510788,-0.1045581624,0.500813067,-0.1967810541,0.4951876104,0.112060748,0.0178671218,0.0320244357,0.2279212922,-0.1135250032,0.3275372088,0.1678346694,0.1013617143,-0.3311156332,0.0835368931,-0.3298193812,-0.0753332973,0.3485963047,-0.1109153405,-0.1289255619,0.0517022908,0.2715263665,0.2995965481,-0.2194067091,-0.1741759479,-0.2277933359,0.1926083565,-0.1441814899,0.0621454716,-0.0277540497,0.0689166859,0.0563400574,0.2010013759,0.0365624167,-0.1425543129,0.3359173536,0.0424343012,0.1257328093,-0.1308773756,0.4169005156,0.4317820072,0.2173967808,0.2984763086,0.4969815016,-0.0052258368,0.0950319767,0.3617352545,-0.1230399385,-0.0056558186,-0.1571082473,0.1468076557,0.2274854183,0.1908045411,0.0842786282,0.0452137664,-0.4120704532,-0.0794824213,-0.3670049608,0.128272444,-0.2398834378,-0.0136592006,-0.4595514834,-0.1641209573,-0.2643000185,-0.0255628284,-0.0869163126,-0.2533309758,0.0683541968,-0.0673565641,0.0462267324,0.0506881103,0.2180024236,-0.1121399552,-0.0008925105,0.0719655603,-0.0486420095,0.2541557848,0.2816670537,-0.1467849612,-0.0329768993,0.0827528685,0.093235366,0.3355743289,0.4587855637,0.4184202552,0.1990620196,-0.2211630493,-0.09807197,-0.0541128889,-0.1821768731,0.0240321271,0.3827438056,0.2471114099,-0.0047523649,0.1169658825,0.2090123892,-0.197349906,0.3443657458,0.0608236939,0.3915441632,-0.3751223385,0.074995257,-0.0480515026,0.0876155347,-0.0209018681,0.1245962977,-0.1066247374,-0.0535663031,-0.0355080664,0.3448269367,0.0905748755,-0.2809974849,0.1084682718,0.0890708417,0.1620452702,0.1937519163,0.1264421493,-0.2804147601,0.1279836446,-0.3345086277,0.2561509311,0.1225090399,0.0921450928,-0.1003932655,0.1724293828,-0.0617524572,0.3806799054,0.0458317026,-0.0871022418,-0.1620690972,-0.1810978055,-0.3676353991,-0.2423916906,0.1011776701,-0.2980795205,0.0345576257,-0.086666286,-0.0830673352,-0.1629511714,0.0487065688,-0.0088092433,-0.2808691561,0.2787574232,0.8215997219,0.0961959362,0.3475220501,0.6704810262,0.0217950344,-0.1139181703,-0.0712416768,-0.3274981081,-0.1206214651,0.1852816492,-0.1950809658,0.2957595289,-0.0669247136,-0.2532565594,-0.1121036634,0.6487157941,0.1377529502,-0.657815218,-0.30730021,0.144904688,-0.0248758364,-0.2015416324,-0.0854520649,0.2895033062,-0.0342447609,0.2562033832,-0.1051741987,-0.4393104017,0.0909999683,-0.4574047327,-0.4693599641,-0.075863488,0.2827580273,-0.1791931689,0.0152350748,-0.3602010906,-0.0260597654,0.3281439245,0.1909056753,-0.0694261193,0.248964414,-0.1455693394,0.1393265426,-0.0452160649,0.3736135364,0.0142718805,-0.0274027344,0.1887041628,-0.2804619372]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2511","title":"Add C4","comments":"Update on this: I'm computing the checksums of the data files. It will be available soon","body":"## Adding a Dataset\r\n- **Name:** *C4*\r\n- **Description:** *https:\/\/github.com\/allenai\/allennlp\/discussions\/5056*\r\n- **Paper:** *https:\/\/arxiv.org\/abs\/1910.10683*\r\n- **Data:** *https:\/\/huggingface.co\/datasets\/allenai\/c4*\r\n- **Motivation:** *Used a lot for pretraining*\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nShould fix https:\/\/github.com\/huggingface\/datasets\/issues\/1710","comment_length":16,"text":"Add C4\n## Adding a Dataset\r\n- **Name:** *C4*\r\n- **Description:** *https:\/\/github.com\/allenai\/allennlp\/discussions\/5056*\r\n- **Paper:** *https:\/\/arxiv.org\/abs\/1910.10683*\r\n- **Data:** *https:\/\/huggingface.co\/datasets\/allenai\/c4*\r\n- **Motivation:** *Used a lot for pretraining*\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nShould fix https:\/\/github.com\/huggingface\/datasets\/issues\/1710\nUpdate on this: I'm computing the checksums of the data files. It will be available soon","embeddings":[-0.2586114407,-0.2368931323,-0.2156255245,0.0744192898,0.2085535973,-0.0874803811,0.1077219546,0.3031297326,0.0643011406,0.3231354058,-0.2132708132,0.0666778013,-0.1542895585,0.4332842529,-0.0734491944,0.0308050662,0.0055632726,0.2454853207,-0.2516648471,-0.061238613,-0.1237060875,0.1600955725,0.1151486561,-0.3427098691,-0.2370680273,-0.0282374881,-0.0126363691,0.2399484366,-0.2678383291,-0.2220925242,0.1105905995,0.2579703927,-0.1301324666,0.3192816377,-0.0001002748,-0.0244540684,-0.0022872835,-0.1104863659,-0.0692491382,0.2499464452,-0.0230546407,-0.0328948461,-0.0620339401,-0.1048000455,-0.3629558384,0.0126546863,0.0622265562,-0.1381947398,0.3689251542,0.227187708,0.3632343411,0.3869170845,-0.0194074344,-0.1629914492,0.1113213673,0.0167693719,-0.0871038735,0.2593570054,0.2584525049,0.1209056899,-0.2002436519,0.2841012478,0.1847804785,0.0739731714,0.1756887287,0.0254423451,-0.0550467111,0.0629208535,0.0187196117,0.0853793174,0.1215298846,-0.2831244171,-0.3110687137,-0.1756326556,0.0741557702,-0.3809302449,0.1468225867,0.0236500967,0.008267899,-0.0215166695,-0.1706751734,0.0240884051,-0.1968894452,-0.0516155809,0.1005559191,0.0680278316,-0.2644309998,-0.2472828627,0.2497013658,0.0335759297,-0.3803047836,0.1121621057,-0.0670740008,-0.0589751825,-0.0907467976,-0.2901783288,0.0382392816,0.2546216547,0.3078463078,0.1262143254,0.1025282964,0.2484940588,-0.0234189685,0.1000332087,-0.1276970506,-0.078260839,0.035893105,-0.2585762441,0.3095819652,0.1944138557,-0.1259980649,-0.1424466521,0.0289675854,-0.1575796157,-0.0986094177,0.0520872958,-0.1677925587,-0.2475500852,-0.2214194685,0.0771839321,0.0888821557,0.0077095767,0.1176060364,0.5735186338,-0.0618299097,-0.1944669485,0.0046390933,0.2758075297,-0.1213817969,-0.1446335316,-0.3123088777,-0.0546745025,-0.2895224392,0.0227369461,0.3508206904,-0.0656067133,0.3327207863,0.0534064099,0.2864612639,0.1202825606,0.1082647294,0.0188669022,0.1733444482,0.0356118828,-0.0004558177,-0.2364005297,0.1209374517,-0.0725719705,-0.0417091288,-0.4723997414,0.0708002076,0.0765203163,-0.2188246995,0.3166131079,0.002436931,-0.0092682065,-0.0422471538,0.0561761521,-0.0712293535,-0.1864610016,0.0144609539,0.1856624931,-0.0902405903,-0.1852321327,0.1604566276,0.2504974306,0.0348879062,-0.1383182108,0.2225313634,-0.1801272184,0.0053034388,0.1121449545,0.0546575896,-0.1318762898,-0.206851691,0.0868494809,-0.0757228285,-0.1791615039,-0.2858584821,0.0691651702,-0.1840796471,-0.011854006,0.1742110401,0.1349700242,0.0025919797,-0.1063934565,-0.0471261404,0.0593735427,0.0134822046,0.1410415769,-0.2126968503,-0.2363345474,-0.1547973603,0.1216252521,0.1423418671,-0.1273562461,0.2024533153,-0.1252297759,0.3256566823,-0.3942752182,0.0615581498,0.1828734279,0.4015356302,0.021906985,-0.1932616085,-0.1827530116,-0.2574688494,0.0062637813,-0.0590291247,0.1451519877,0.0297113135,-0.3581786156,-0.0883436203,-0.2195280492,-0.1305889934,-0.0849122107,0.3844880164,0.2899780571,0.0879077539,-0.0343079455,0.0201949701,0.3061675727,-0.190151751,0.0976372734,-0.0789303705,0.1960108727,-0.1100703776,-0.1417631805,0.3407817483,0.1319140643,-0.0608002953,-0.1108487546,0.1443618685,0.3013299406,-0.0395784751,0.2133542299,0.2270402163,0.4914596379,0.0338227712,-0.158936128,0.0062979907,0.1531943828,-0.010462366,0.0646910369,-0.1888082325,0.5184321404,0.0301831048,-0.1605615616,0.2123947144,-0.1512403339,0.2263773531,-0.1737162322,-0.0606891066,-0.3493960798,0.1980843395,-0.0980962515,0.2594591081,-0.1316393167,-0.4100782275,0.0487894341,0.2338222265,0.1663601696,0.170386672,-0.0405135304,-0.0363341756,0.0802968368,0.1401967555,0.1761394441,0.213407591,0.5047037005,0.1331339777,0.0588973314,-0.1254735142,-0.1863913238,0.1383117884,-0.2588725984,-0.0642817765,0.2351440638,0.0219158251,-0.1701902002,-0.6001666188,0.0372480676,-0.249778524,-0.0394537561,-0.2543345094,-0.0752418265,-0.1240677088,-0.2492334992,-0.1242171004,0.0269252509,-0.1130862609,-0.1818078905,0.199516654,0.1633022428,-0.3851896822,0.0591588542,-0.2787411213,0.2953506112,0.0269858334,0.0085841501,0.0370016024,0.085481979,0.0408057943,0.2772705853,0.2907311916,0.0665358976,0.628795445,-0.3146470189,0.1830504537,-0.2187590152,-0.5036549568,0.1665034592,-0.1542676091,0.1319329143,0.2156217396,0.1723758429,-0.0999767333,-0.3396267891,0.1567200869,-0.3860035539,-0.3319838941,-0.046469871,-0.0542919151,0.0138935391,-0.3315369487,-0.1770873517,-0.0326982588,-0.2762327194,0.410189718,0.1199416295,0.013964761,0.1550645232,0.0994469449,0.1597021371,-0.3217945695,0.2671185434,-0.3387693465,-0.4841817319,0.2318953276,-0.3925537467,-0.4123530984,-0.192890808,-0.1043146774,0.3161990345,-0.0530861989,-0.2953405976,-0.2839132845,-0.1251228899,0.2562994063,0.0690907761,0.0794284791,0.2865327001,-0.0360680148,-0.3148214519,-0.2869904041,0.009124686,0.3432369232,-0.2602678835,0.0930616409,-0.3781878054,0.0193489045,0.0591104999,0.3250522316,0.180475384,-0.2283665389,0.3548689485,-0.1520234644,0.5535254478,0.0695398077,-0.1883889437,0.2312797308,-0.004539534,0.1252145916,0.4402047396,0.1030635685,0.0941187143,-0.0408257954,0.0296800658,-0.3676127195,-0.3207439482,0.0620300099,0.1089212894,0.1617344767,-0.1250067204,0.0586354956,-0.1335464418,-0.1395232081,0.1339299679,0.2203106433,-0.0487965383,-0.2848830819,-0.3441628516,0.0100541776,-0.4193620086,0.2949059904,0.1592249721,-0.0592814796,-0.1573583633,-0.1976354569,-0.1458630115,-0.1930574775,0.3951094747,0.023937203,0.1395500302,-0.0657826886,-0.1350118518,-0.1307117641,0.2956539094,-0.390679121,-0.1058883518,0.2813071609,0.3926399052,-0.4702376425,-0.0706677735,0.0776190013,0.1879817098,-0.0025737663,0.0136998529,-0.1803116649,-0.0984489769,-0.1680195332,0.2098437548,-0.12098746,0.134672612,-0.0937305912,-0.1281519979,0.0898434445,-0.0872648284,0.3720995486,0.2908614576,0.2714125514,0.1856375486,0.1081400514,0.0484058596,0.1658587009,0.3566215634,0.3592710793,-0.1186856776,0.0882533714,-0.1735043228,-0.2214381844,0.3014198542,0.2030114979,0.1486733109,0.1415748745,0.0163742043,0.1771242917,-0.3032541573,0.3258421123,0.1457597762,0.0752536952,-0.0524989478,-0.0206209347,0.1941458732,-0.1415146887,-0.037385378,0.261151731,0.4078726172,-0.3011979461,0.0937329903,0.2443811297,0.8777470589,-0.0046672262,0.217697069,0.2962688208,-0.3105140924,0.105028756,-0.1164755672,0.0513804294,-0.1365356147,-0.3590020537,0.0525877737,0.0482212231,0.0691625178,-0.0005929003,-0.1677506566,-0.0689116716,0.0936186835,0.1628494859,-0.1722352803,0.0950144753,-0.1619414389,-0.1869973242,-0.2368478477,0.3349007964,0.1559036225,-0.1940421462,-0.158341676,-0.2405658066,-0.0901219621,-0.1532129794,-0.0317376517,-0.0857944638,-0.2066100836,-0.1015003622,0.0833995789,-0.0622988269,-0.072038576,0.0582491718,0.3871597648,0.3838898838,-0.2678079903,0.1037069783,0.0349543244,0.0192259308,-0.1153027192,0.1006241441,-0.0004591541,-0.153299436,-0.0048643863,-0.0906594023,0.1995953918,-0.3959479332,0.073967047,-0.0694265738,-0.3752855659,-0.2290606797,0.0389383025,0.1117535681,-0.1367266178,-0.2934674025,0.2787504494,0.1051097959,-0.1143933088,0.0157205928,0.3547438979,-0.1251776069,-0.2527896166,0.383408457,-0.152192384,-0.1380854398,0.2241638303,0.0676642135,-0.2761642933,-0.4785531461,-0.204191342,0.1426137984,-0.2338983268,0.0328520276,0.0920856893,-0.2548292577,0.0880373418,0.2719169259,0.3056760728,-0.0205095038,0.2051071227,-0.0958158299,-0.619484961,0.2811788321,-0.0765736625,0.4294385016,-0.0263865944,0.441427201,-0.2109744698,0.0565094762,-0.5384542346,-0.0210269634,-0.3181409538,-0.0798597336,0.1433393806,-0.1306783259,0.3084122837,-0.0928664804,0.3257278502,0.0914620459,-0.3117181957,-0.3644774854,-0.1320673227,0.0511494614,-0.0996304676,0.0106854746,0.3214328289,0.1851658821,-0.1380699128,-0.1665500253,0.1431972384,0.0987851396,0.0346628278,0.224277541,0.2374432683,0.0527482368,0.1088702679,-0.2880391777,0.0840386003,-0.0015571039,-0.0754436255,-0.1213696003,-0.329036206,-0.1117813215,0.5154303908,0.317478478,0.1826005578,0.0761368498,0.1818925589,0.0261965748,-0.2194978595,0.3510786891,0.4336693585,-0.0402932875,-0.0891636983,0.1107378528,0.1840313226,0.4534437954,-0.3457513154,0.068099685,0.0552136116,-0.0613621026,0.374915868,0.257320255,0.0839184374,-0.2434906214,0.1740943789,0.1419276446,-0.1442545801,-0.0210443567,0.0100245103,0.1906672269,0.0687269121,-0.1367884576,0.2694217861,0.026739927,0.1927177757,0.1791886389,-0.1262068599,0.2254700214,-0.3400603831,0.0581016466,0.0377326049,-0.0817449838,0.0968952104,0.2272134125,0.0492978133,0.1836009175,0.1171772555,0.5696305037,-0.1565781236,-0.2025785446,-0.359615624,0.1760598421,-0.2035920173,-0.0836501047,-0.1376357824,-0.2224752009,-0.170541063,0.0411342233,0.0798767731,-0.0019958045,0.1915551573,-0.035468027,0.0658656284,-0.2086837143,-0.0682744011,0.1018467322,0.2579044104,-0.1790467054,0.1622300446,0.3193747997,0.02818303,0.2234219164,0.3514073193,0.2412113845,-0.0871994421,0.0893329382,0.1626980752,-0.1762306541,-0.0311367381,-0.3607547581,0.2241190374,0.3361491263,0.1975808591,0.3792403936,0.2971457839,-0.3550035655,-0.0402267165,-0.0842021927,-0.0679760575,-0.070557408,0.1620290875,0.0212596711,-0.084530212,-0.1829674542,-0.162642628,-0.4618337154,0.0602067746,-0.0069878353,-0.1328731626,0.2082100511,-0.0472464338,0.1634314656,0.1856045425,0.5502731204,0.0255942736,0.1202687472,-0.1698035747,-0.4889599085,-0.5210318565,0.2235606313,0.0629495531,0.0278372616,-0.3625488579,0.0703040287,0.4442741871,-0.0635202974,0.2373723239,-0.035265889,0.1218710393,-0.0982064977,-0.4115131795,0.1512268782,-0.0487998314,-0.1151851043,0.2564331591,-0.1479368657,0.0145483455,0.0333208963,0.3110305667,0.005245544,0.1741110981,-0.2488556504,-0.123388201,0.3204123974,0.008249322,0.4328651428,-0.5732752085,-0.2378932536,-0.1087393239,-0.0136973178,-0.1755241901,0.0337197818,0.1392369419,0.5417926908,-0.0298419446,-0.1966596395,-0.0242918413,0.3196481466,0.3039909005,0.0287045538,-0.1446598768,0.3647102416,-0.0003912729,-0.2065259218,-0.0516169444,0.4915200174,-0.0658000857,-0.0153523041,-0.1639591306,-0.3880117834,0.5056558251,-0.3207930326,-0.0266803019,-0.0344345495,0.3556982279,0.1834290475,-0.2193897665,-0.5304563046,0.2442419976,0.0832718015,0.1025891528,-0.1924135685,0.3654026389,-0.1686755866,-0.1641025543,-0.0059415218,0.4586779177,0.2416504472,-0.2849210799,-0.0465817712,-0.3411107957]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2511","title":"Add C4","comments":"Added in #2575 :)","body":"## Adding a Dataset\r\n- **Name:** *C4*\r\n- **Description:** *https:\/\/github.com\/allenai\/allennlp\/discussions\/5056*\r\n- **Paper:** *https:\/\/arxiv.org\/abs\/1910.10683*\r\n- **Data:** *https:\/\/huggingface.co\/datasets\/allenai\/c4*\r\n- **Motivation:** *Used a lot for pretraining*\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nShould fix https:\/\/github.com\/huggingface\/datasets\/issues\/1710","comment_length":4,"text":"Add C4\n## Adding a Dataset\r\n- **Name:** *C4*\r\n- **Description:** *https:\/\/github.com\/allenai\/allennlp\/discussions\/5056*\r\n- **Paper:** *https:\/\/arxiv.org\/abs\/1910.10683*\r\n- **Data:** *https:\/\/huggingface.co\/datasets\/allenai\/c4*\r\n- **Motivation:** *Used a lot for pretraining*\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nShould fix https:\/\/github.com\/huggingface\/datasets\/issues\/1710\nAdded in #2575 :)","embeddings":[-0.2344433367,-0.3906821609,-0.1943427324,0.0365951471,0.260853231,-0.0699108839,0.1942980438,0.2204871923,-0.0224221796,0.3215046525,-0.1395500451,0.135219574,-0.1403120458,0.412068814,0.0247789696,-0.0698187947,-0.0553392954,0.2685842812,-0.1971850395,0.0043837796,-0.1201309487,0.1770246327,0.0639820397,-0.2327889949,-0.3242893219,-0.0952162072,-0.0030869816,0.1822504848,-0.2013797462,-0.2951210141,0.1486285627,0.1537279189,-0.0480645336,0.2340649962,-0.000099735,-0.019610798,-0.0189707801,-0.0715517402,-0.0994714051,0.2024277896,0.0377671011,-0.0110399155,0.0525897965,-0.0594155565,-0.4323686063,0.006170704,0.0860170051,-0.1048312858,0.3951166272,0.2404221594,0.3758192062,0.3852788806,-0.0671650544,-0.2324172109,0.0868422464,0.011395786,-0.1464603543,0.1999960095,0.2307098657,0.0745369941,-0.2138725221,0.3292371035,0.2598564327,0.1822265983,0.1986621022,0.0957833007,-0.0557409041,-0.021023728,0.0779224038,0.0257832743,0.1392601877,-0.1937361211,-0.2446497381,-0.2401965708,0.1267556697,-0.4167433679,0.0410414189,0.0019691151,0.0327267535,0.0244100019,-0.1663618088,-0.0517191477,-0.2158903927,-0.0316852815,0.0903222188,0.0921222568,-0.3035559058,-0.1753581911,0.2182398289,-0.0181631166,-0.321408987,0.2396737188,-0.0755710229,-0.0484239273,-0.071132645,-0.2299449444,0.0898965076,0.2217904925,0.1780834496,0.0309524164,0.0967658162,0.1953027546,0.1000130475,0.093747817,-0.0689985827,-0.1269021928,0.0303257462,-0.2261117846,0.289804846,0.1659276932,-0.137440607,-0.2306738198,0.0550743341,-0.1424314678,-0.1735295206,-0.1012332365,-0.0628956482,-0.1496122926,-0.2296491563,0.0305978339,-0.0161238965,0.0579657219,0.0310306475,0.5734707713,-0.0140834413,-0.1409287602,0.0736849234,0.2848813832,-0.1210069507,-0.1886797994,-0.3421661258,-0.0557121933,-0.2479030937,0.003606667,0.2968689501,-0.0509543568,0.3760034144,0.1318581849,0.2128509283,0.1096834168,0.0544708967,0.1055217907,0.1028580964,0.0024623349,0.0508919433,-0.1751512289,0.1023069695,-0.0854191855,0.0156091284,-0.4179218411,0.0604162328,0.0094343647,-0.2751068175,0.3247367144,0.0984877273,-0.0671711937,-0.0317455791,0.2549746037,-0.0404331833,-0.0783001259,0.0338681601,0.2129720747,-0.1201903522,-0.1339084506,0.1572327316,0.2271873951,0.0289911851,-0.2779015601,0.2520412803,-0.1317164451,0.0777954534,0.0799978077,0.0002347801,-0.0465283915,-0.1701281518,0.0475399122,-0.0541833751,-0.0806211829,-0.2444818318,0.0046582711,-0.2656595111,0.0037702085,0.1509025693,0.0902954489,-0.0218908507,0.0812395364,0.0568003021,0.070149906,-0.0640587807,0.139835909,-0.2337794453,-0.214593038,-0.1413364261,0.1702506244,0.1186616421,-0.0396530218,0.1420298815,-0.0376842059,0.3090627789,-0.3216448128,0.0866875798,0.2092159092,0.3242675662,0.0089461468,-0.1836088896,-0.1528851688,-0.2853154242,-0.029235838,0.0359025225,0.2083282471,0.1154955104,-0.3303009272,-0.1558167487,-0.1922239661,-0.1541130096,-0.1970716417,0.3648617864,0.1732511371,0.009415498,-0.0254284181,-0.0163653623,0.3926531076,-0.165505901,0.166424796,-0.1201985925,0.1577541828,-0.1702105254,-0.1580424905,0.3138777912,0.1038700268,0.0549468547,-0.080930464,0.1194930375,0.2432479411,-0.1677697599,0.3003966212,0.1702382267,0.2947887182,0.0658737198,-0.1010658741,-0.0612781048,0.0702186003,-0.0328161642,0.0590912774,-0.089099586,0.4168736041,0.1803853065,-0.0688817129,0.1885308027,-0.0947139636,0.1563051194,-0.1501801759,-0.0120694842,-0.3425281942,0.166008696,-0.1143769547,0.2718318105,-0.1833514422,-0.5105952621,0.149700895,0.2582325637,0.2820986211,0.164034754,-0.0028264418,-0.1501956284,0.0422093794,0.1503812075,0.1037307605,0.2407307029,0.5099434257,0.1094598174,0.0653790459,-0.0995779783,-0.2738417387,0.1527263224,-0.1961590499,-0.1267359704,0.0893356279,0.1167041361,-0.2023904771,-0.6548163295,0.0400626063,-0.2611787021,0.017373208,-0.2837429643,-0.0989788026,-0.1188535914,-0.2407790869,-0.0296717882,-0.0287268162,-0.0863250196,-0.1765737534,0.2232756466,0.241784513,-0.3360280991,0.1290718913,-0.1889470816,0.3311981857,-0.063713707,0.1345336139,-0.061513342,0.0488411933,-0.0331628323,0.3136166036,0.0382985845,-0.0211997274,0.4916908145,-0.2773482502,0.1852730066,-0.1467280686,-0.5441322327,0.1577872485,-0.0858022347,0.1268968582,0.1932612658,0.2204558402,-0.1938165128,-0.3260070086,0.1832210273,-0.3937160373,-0.2774964869,-0.2058346421,-0.0673979223,0.076486811,-0.3641819656,-0.2514015734,0.0047645052,-0.3476854265,0.3983205259,0.0803255513,-0.0423477478,0.2490770668,0.1167000383,0.1509243995,-0.3584468663,0.2211385369,-0.3931612074,-0.3766806424,0.1289016306,-0.4357405305,-0.3753900528,-0.1991757154,-0.1082544774,0.3652496636,-0.0695038512,-0.2768137157,-0.2748773396,-0.1046139821,0.2093257606,0.0063232658,0.113631174,0.2782217264,-0.0482716598,-0.3148500323,-0.2422292382,-0.0897121876,0.3169451058,-0.1809300333,0.0720232055,-0.391962707,0.080948092,0.0453964286,0.3726854622,0.0538321957,-0.2455335557,0.447317183,-0.2543344796,0.5190086961,0.0164172538,-0.2126834542,0.2253325433,0.0152811771,0.1581837684,0.4282591939,0.1037226394,0.0544313714,-0.0459536649,-0.0004861334,-0.2473383695,-0.3113833666,-0.0301664397,0.2328010201,0.1380427033,-0.0334498808,0.0947986022,-0.1328778565,-0.0284823459,0.0362609476,0.1492090225,-0.0034859276,-0.2496963292,-0.3671901226,0.0453670658,-0.5676614046,0.2498455942,0.0490562133,-0.0714492276,-0.178967312,-0.2127940059,-0.1275651008,-0.0843061432,0.4376881123,0.0498822071,0.0169767551,-0.0847874954,-0.1595825106,-0.2627815604,0.3040933311,-0.3363463879,-0.0377892628,0.2908061445,0.3889855146,-0.4010408819,-0.0854307935,0.1471197605,0.0787090585,0.0190942939,0.0971864015,-0.1957533211,-0.1616914272,-0.2172822207,0.212027356,-0.0418320149,0.1290567815,0.0410113819,-0.1609348208,0.0459767133,-0.1702471673,0.2235934585,0.3728913963,0.2689760029,0.2287289053,0.1379177719,0.1124308184,0.1574763805,0.2794084847,0.3287061155,-0.1076338962,0.0581780672,-0.1318325698,-0.1220538542,0.3923823237,0.2603607476,0.1227228642,0.2968573272,-0.0803986713,0.2561702728,-0.2190978825,0.402680546,0.2079286277,0.0753193125,0.0315775946,-0.0525079928,0.233975634,-0.132142216,0.0173908304,0.2585658729,0.3658845127,-0.3065921068,0.0116778612,0.1251391917,0.8706358671,0.1638902575,0.2155781984,0.32018888,-0.3158908486,0.1423148662,-0.0210631266,0.118104279,-0.1503908038,-0.2230451256,0.1033703759,0.0922752991,0.0555695891,0.0235658959,-0.1292231083,-0.0529488735,0.0874345601,0.215187639,-0.1661209017,0.0372053087,-0.1446150541,-0.2558126152,-0.338742733,0.3025230467,0.0918407366,-0.0882363617,-0.1651778966,-0.2201616019,-0.0224153548,-0.2279692441,0.0845657885,-0.0621999875,-0.2908402383,-0.101875484,0.0407016799,-0.1807431877,-0.003752931,0.0295104459,0.3540422022,0.1998990923,-0.2533261478,0.2224539369,-0.0467376374,-0.0617452972,-0.1633657962,0.1787913591,-0.0904491842,-0.1599682271,-0.0494824983,-0.0934037268,0.1720372736,-0.3102512956,-0.017954275,-0.0202601776,-0.3416411281,-0.2493533045,0.0060019395,0.1569802761,-0.029047003,-0.339045465,0.2825603485,0.0223386511,-0.0732549056,-0.013294192,0.3939886093,-0.1194318235,-0.3077161312,0.378641665,-0.1316069961,-0.0451316461,0.2466545701,0.0762296617,-0.249895528,-0.4240954816,-0.1625063568,0.2643860877,-0.3310705721,0.0875343084,0.1799612492,-0.2712115347,0.1032984555,0.2762470841,0.1313917786,0.0008097842,0.1441404223,0.020765774,-0.6334751844,0.2932320535,-0.1978968382,0.3516800702,0.0547097363,0.4703651965,-0.1233265325,0.0501354039,-0.5301721692,-0.0145775732,-0.261782378,-0.0305768754,0.1536672413,-0.2609249055,0.3702170551,-0.0474862047,0.3372846246,0.1946006715,-0.3156359494,-0.3791424334,-0.1801032126,0.044058539,-0.0458877571,0.0136305429,0.2319969237,0.2421181798,-0.2043872476,-0.1134126857,0.2090076804,0.1583095491,0.1357143074,0.1977695376,0.178833738,0.0587519668,0.164677754,-0.3051436543,0.1414149255,0.0584577918,-0.0459109321,-0.1529089957,-0.3444571793,-0.0683570281,0.4542391896,0.2139399499,0.2135746032,0.0989457667,0.0862984434,0.011925743,-0.2495516837,0.2787104547,0.3647052944,-0.0439545438,-0.0967848897,0.1007360891,0.1884394437,0.4684045315,-0.4216820598,0.1055960134,0.0298482254,-0.1067405194,0.3985698223,0.2672733366,0.0307435468,-0.1672433019,0.1396106929,0.2278959006,-0.0184835028,-0.0815756693,0.0022122737,0.1031240821,0.0164912436,-0.1361324191,0.3969686329,0.1707354188,0.2703040242,0.1861242056,-0.1948961616,0.1780884564,-0.2012618333,0.0387670584,0.1999378502,0.0466831811,-0.0447732173,0.2925358713,0.0580352359,0.1681791395,0.1109297723,0.5485624075,-0.1757249236,-0.2970117629,-0.2540398836,0.1223636046,-0.2329696119,-0.0827624202,-0.0549092069,-0.196206063,-0.2294317484,0.0231994633,0.0317624174,-0.0547924936,0.224004373,-0.1294074208,0.0819624737,-0.2938680947,-0.0917474553,0.1469578892,0.3143924177,-0.0969735086,0.1637091488,0.2131730318,-0.0312776268,0.2220522314,0.4186244607,0.2451396137,-0.1407750994,0.1147199646,0.1356685162,-0.0703529045,-0.0773850381,-0.3427345753,0.0456366315,0.3602008522,0.1662038118,0.2938136458,0.2989548445,-0.3946097493,0.0200323742,-0.0337892175,0.0689972118,-0.1923976839,0.1357389987,-0.0969207957,-0.1170511246,-0.1663878858,-0.1740650088,-0.4302501678,0.1320469677,0.0130676301,-0.187091738,0.1094773859,-0.1336766928,0.1583885849,0.1368099004,0.5094417334,-0.0112745222,0.0819683224,-0.1720140427,-0.4892557561,-0.5047972202,0.2177073359,0.0534733236,0.0570028014,-0.297267437,0.1011878029,0.3718442023,-0.0296259969,0.2696983516,0.0074495855,0.121064052,-0.154440105,-0.3885409534,0.1438043416,-0.0754034892,-0.0747010857,0.3037541509,-0.0868727416,0.0088903671,-0.0214073304,0.2714061141,-0.0707278773,0.1798810214,-0.2232897729,-0.1272991747,0.3452400267,-0.0166034847,0.2549382448,-0.5629046559,-0.2168508917,-0.1134535521,-0.2293744683,-0.2637234032,0.0017269755,0.1459636986,0.5841070414,0.0052856337,-0.2345767468,-0.0233155508,0.2357273549,0.3529059887,0.129469052,-0.0562659204,0.4194053411,0.0928900167,-0.1696400046,-0.0861487165,0.5203219056,-0.0491874591,-0.1113428473,-0.1989598721,-0.5387395024,0.5652921796,-0.3702072799,-0.0738666356,0.0090218978,0.3253996372,0.287076354,-0.3082360327,-0.5058665276,0.190735966,0.0995966718,0.0566689335,-0.1651223749,0.3893477321,-0.1845410913,-0.2024880201,0.0002469565,0.4747718573,0.3225940466,-0.2759100795,0.0468181819,-0.2912892997]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2508","title":"Load Image Classification Dataset from Local ","comments":"Hi ! Is this folder structure a standard, a bit like imagenet ?\r\nIn this case maybe we can consider having a dataset loader for cifar-like, imagenet-like, squad-like, conll-like etc. datasets ?\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nmy_custom_cifar = load_dataset(\"cifar_like\", data_dir=\"path\/to\/data\/dir\")\r\n```\r\n\r\nLet me know what you think","body":"**Is your feature request related to a problem? Please describe.**\r\nYes - we would like to load an image classification dataset with datasets without having to write a custom data loader.\r\n\r\n**Describe the solution you'd like**\r\n\r\nGiven a folder structure with images of each class in each folder, the ability to load these folders into a HuggingFace dataset like \"cifar10\".\r\n\r\n**Describe alternatives you've considered**\r\n\r\nImplement ViT training outside of the HuggingFace Trainer and without datasets (we did this but prefer to stay on the main path)\r\n\r\nWrite custom data loader logic\r\n\r\n**Additional context**\r\n\r\nWe're training ViT on custom dataset\r\n","comment_length":48,"text":"Load Image Classification Dataset from Local \n**Is your feature request related to a problem? Please describe.**\r\nYes - we would like to load an image classification dataset with datasets without having to write a custom data loader.\r\n\r\n**Describe the solution you'd like**\r\n\r\nGiven a folder structure with images of each class in each folder, the ability to load these folders into a HuggingFace dataset like \"cifar10\".\r\n\r\n**Describe alternatives you've considered**\r\n\r\nImplement ViT training outside of the HuggingFace Trainer and without datasets (we did this but prefer to stay on the main path)\r\n\r\nWrite custom data loader logic\r\n\r\n**Additional context**\r\n\r\nWe're training ViT on custom dataset\r\n\nHi ! Is this folder structure a standard, a bit like imagenet ?\r\nIn this case maybe we can consider having a dataset loader for cifar-like, imagenet-like, squad-like, conll-like etc. datasets ?\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nmy_custom_cifar = load_dataset(\"cifar_like\", data_dir=\"path\/to\/data\/dir\")\r\n```\r\n\r\nLet me know what you think","embeddings":[-0.1711518764,-0.1667470187,0.0313529335,0.4532008469,0.2632547319,-0.0811185464,0.2285084426,0.0517842583,0.3194644153,0.2681328654,-0.1980350912,0.0934309736,-0.3347459733,0.3943925798,0.03784661,-0.2031163573,-0.2279976457,0.2942554355,-0.1917820275,-0.2173044384,-0.0529894456,-0.1112344861,0.1236169711,0.0195183475,-0.2421073765,0.1486360133,-0.2809579074,0.4919865131,-0.1451515108,-0.10265477,0.4892801344,0.0337604471,0.368265003,0.6073811054,-0.0001115616,0.1082242206,-0.1246429682,-0.2912990451,-0.122024022,-0.3184724748,-0.0028073406,-0.1293970346,0.2840099931,-0.2369065583,-0.4590673447,-0.0310311709,0.0454267785,-0.3836547732,0.4092215002,0.2136450112,0.1300623864,0.0714408457,-0.4090717733,0.0486208014,-0.2243045121,0.5681055188,-0.1489504427,0.4393532276,0.3459990025,-0.2117604613,-0.0951595157,0.1754909754,0.0813865513,-0.0121847438,0.6244930029,0.326700002,-0.2012632489,-0.2279163748,-0.179352209,0.1679523289,0.2916376889,-0.1329025179,-0.1520915627,-0.1528731138,0.1039553285,-0.1191771477,0.0571648628,0.1517658681,-0.4100264013,0.261583209,-0.2354246527,-0.148532331,-0.4728707969,0.1664527059,-0.0157123823,-0.3108100295,-0.1096825376,-0.0335084014,0.1107050404,-0.0302021056,-0.0637880191,-0.0225583259,0.1726934314,0.521335721,-0.1483541429,-0.2448643744,-0.0624012277,0.1635794342,0.0854707882,-0.1567988545,0.1323319077,-0.0657559037,-0.4273361564,0.0766008571,0.1742494255,0.0852662995,-0.069465898,0.3597370088,-0.1167033762,0.0099350614,-0.4018887877,-0.1022835374,-0.3766005635,-0.0059971726,-0.1649569869,-0.0880271718,-0.0829650909,-0.0514034219,0.2310631126,-0.212762624,0.1315826029,0.1782817245,0.2259933352,0.2244734764,0.0003828079,-0.0224533584,0.1043003276,-0.0969310105,-0.1879051626,-0.0112456102,-0.1539769769,0.1492160112,-0.284630537,0.1611146033,0.0593347289,-0.2404901832,0.0239400174,-0.2232363075,0.4601967931,0.1644294262,-0.0332620516,-0.194562301,0.2785152197,0.2013729513,-0.3364462256,-0.1555517167,0.3225756884,-0.1244988218,-0.2927462161,0.0910836905,-0.4155611396,-0.3023661375,0.1320977956,0.127516076,0.054183308,-0.1185607016,-0.4537299573,0.0471442714,-0.3906651735,-0.0376894549,-0.2005945891,0.0031112428,-0.3844126463,-0.1122067049,0.325560987,0.4995104074,-0.3537164629,-0.1935031563,-0.2759791613,-0.1226335168,-0.0405496061,0.3582922518,-0.2825185061,0.3609904647,-0.231622234,-0.0004718885,0.3596025109,-0.6408893466,-0.1602463424,0.3611235023,-0.1468535662,0.083478868,-0.1106667072,0.4675813615,0.0573210604,-0.0437320732,-0.1097148508,0.7292591333,-0.2461506426,0.1003957465,0.0783915371,-0.3444947898,-0.0595198572,0.3862071633,-0.0255865082,0.261233598,0.0877624601,-0.1202732623,0.1105117723,-0.240093112,-0.01672481,-0.0660725981,0.2149087489,0.2525713146,-0.0356158614,0.0856861174,-0.3853397667,0.3809303939,0.0137658063,0.2332108468,-0.0894564837,-0.0775726736,-0.1049795672,-0.1797708422,-0.2935766876,-0.0740475804,-0.019134203,0.0974868983,-0.0879434943,-0.1982378811,-0.3646779954,0.3607048988,-0.4032907486,0.1460895687,0.155663386,0.1872362494,0.1924535781,0.0663262084,-0.2258826047,0.0304024052,-0.1710557491,-0.2163349837,-0.0901054889,0.1741504818,0.1317851096,0.2255440205,0.2103808373,0.1911410242,0.2752165496,-0.2569105625,0.0800213367,-0.0953982174,0.303080827,0.050475657,-0.2693997324,0.4733867943,-0.0346195288,0.265086323,0.1230627075,-0.1807813197,0.0388387926,-0.117783621,0.007415894,-0.069532901,0.1119439751,0.4586138427,0.3327568173,0.2433035374,-0.1677209586,-0.2217969,-0.0880103931,0.0076388605,0.0139185833,0.295162797,0.0542380363,0.3093998432,-0.0272909105,-0.1097018942,0.5220649838,0.0971358791,-0.1485693455,0.1968320012,0.1740426421,0.0449788608,0.0066723232,-0.00370813,0.1470155716,0.0109731294,-0.0192599483,0.1900916845,-0.3261588812,-0.200176388,-0.2250062525,-0.0429200009,-0.0188866779,0.3271223903,-0.0284519643,-0.435213387,-0.2229123563,-0.1485522985,-0.0564320385,0.4121331275,-0.026557697,-0.1437772214,-0.0114058815,0.0061254282,0.0188556705,0.748334825,-0.0273529217,-0.552380383,0.1248901337,0.0192498825,-0.1453247219,0.0816013664,0.2255043387,-0.0599971749,0.3380980492,-0.0557878092,0.0297195092,-0.2399781793,-0.2019121945,0.2341752946,-0.0794292018,0.329430759,0.0342180133,0.2260845006,0.0151083842,-0.2302158177,0.2885739207,-0.0619158521,-0.2701322734,-0.3402596116,0.1675587147,0.3559232652,-0.1829326898,-0.2913974524,-0.6088712811,-0.3206725121,0.2787880003,0.4164873958,0.2521072626,0.2307675481,0.1515019536,-0.0546728149,-0.1698648781,0.0696597397,0.1167342812,-0.5177243352,0.035861887,-0.2100634724,-0.220499143,-0.2387831956,-0.2466602772,0.1457769126,0.0720841736,-0.38451612,-0.2613418102,-0.0408322029,0.1592840701,-0.0618202612,-0.0013889262,0.1672932506,-0.0954543948,-0.0235284828,0.0389559865,0.2476973534,0.0794435889,0.0094007812,-0.1864801645,0.1917174309,0.1555677801,-0.1275321245,0.4102587998,-0.0690623969,-0.2088337988,0.4698214531,-0.2183326185,0.2523116767,-0.501657486,-0.5047842264,0.2008981258,0.2034907192,-0.3791694045,0.2592870593,0.3521589041,-0.0071075479,-0.3364824355,-0.1287143826,-0.0697979107,-0.2419340163,0.3536166847,0.2839049101,0.3354141712,-0.4179280996,-0.1929782629,-0.194346562,-0.0567977056,-0.0406867936,0.6717273593,0.1609262377,0.0094619198,-0.3218970895,0.2331465185,-0.0091690999,-0.0332697965,0.1991981566,-0.1552209854,-0.1863545328,-0.050049711,0.0616948754,0.1646850556,0.3063277602,0.074585937,0.101086542,0.1251390874,-0.4726458192,-0.1690327227,0.0957720131,0.1085049808,-0.1605380923,-0.2302184403,0.6902582049,0.0756456181,-0.3655419946,0.175186947,0.1750392616,-0.3738020957,-0.101386562,-0.1717061251,0.0359934866,-0.2495566756,-0.0382312201,0.0375485718,0.2471781373,0.3213495016,0.1070403606,-0.1904520243,-0.0473848246,0.1418724805,-0.0774636418,0.0356822833,0.2747861445,0.2058420628,0.2150744796,0.3199415803,-0.0552344099,0.5318437219,-0.2067218721,-0.448445797,-0.0082834158,0.0992641747,0.3700360656,0.3551442921,0.1360489726,-0.00771019,0.0238100607,-0.0558497682,-0.7978026271,0.5541272163,0.0784005821,-0.1719647348,0.0020041654,-0.4529390037,0.5468375087,-0.0532994941,-0.0842841491,0.1839754283,0.2237266749,-0.3484931588,0.1305000186,0.3636848927,1.0036213398,-0.2000190318,0.5979268551,0.1882564723,0.316577673,0.1871587485,0.2571648657,-0.1189418063,-0.1989306211,-0.3950670063,-0.1318368763,-0.1330111921,0.148886621,0.119044058,0.1053663865,0.1067568734,-0.1086650863,-0.0392305814,0.0464435816,0.2183549553,0.0208843723,-0.5506279469,-0.2668254077,0.139769569,0.1630829424,0.3775499165,-0.0552518815,-0.0429413579,-0.077299498,0.0099428939,-0.004394541,-0.139194876,-0.1740027368,0.0335314013,0.0126943821,-0.2328245044,-0.0502990708,0.3883973956,0.2344049066,0.0964228287,-0.0383393057,0.1355278492,-0.3317182958,0.1921803057,-0.07056842,-0.2202104032,0.399959594,0.0428435095,0.0266042035,-0.1041100919,-0.0429281555,-0.4345785677,-0.030589588,-0.0844630897,0.6611809731,-0.2210626602,-0.1806448102,-0.1703483462,0.014989418,0.2041854113,0.1152872071,0.2015032023,0.0900535062,-0.0656384826,0.2074520141,-0.2563242018,0.0653493106,0.1979349703,0.0860574767,-0.4300831258,0.2644721568,0.0282172505,0.1034595519,-0.2177565396,0.5871706605,0.6104701757,-0.0557040647,0.099685207,0.2635746002,-0.147919789,-0.1274323314,0.7271055579,0.3650838137,-0.3910202682,-0.2315898389,-0.2075193524,-0.5043410063,0.187787652,-0.2077351809,0.2692264318,-0.0096119763,0.0223331526,0.194976598,-0.0040621082,-0.3121486306,0.148597464,-0.1957099438,0.0460952073,0.4459797144,0.2056479752,0.3248225152,0.002192715,0.0678553283,-0.0865037367,-0.1329452246,-0.113889806,-0.2136784345,0.1644244492,0.0822074041,0.1318346262,0.0325372778,-0.2356372476,-0.1330015063,-0.1140092835,0.1053071395,-0.0979826525,-0.1127383485,0.1572509706,-0.2052237391,0.0462915674,-0.160426259,0.2952700257,0.0089542465,0.3124412894,0.2140686065,-0.3142935634,-0.1873663664,-0.277660042,0.2408557087,0.1674015075,0.1792428344,-0.0307552218,0.2105464637,0.1559767127,0.231523484,0.0817586929,0.3380711675,-0.1454683095,-0.2125446647,0.154746294,0.4487452507,0.1892141551,-0.1692368537,-0.2411860526,-0.1099906787,-0.1559248269,0.1897260547,0.3677541912,-0.2745054364,0.3570895791,-0.2224761695,-0.1835071445,0.2273589075,0.1385715157,-0.0709574148,0.1136435717,-0.260101527,0.0836565346,0.1571115851,0.2050438821,0.148743704,0.4492584169,-0.4987841845,0.2104026377,-0.4103265107,0.053292077,0.3693914711,-0.370634228,0.0909091309,0.0052357232,0.1692736745,0.0678852499,-0.3774766922,0.2427134067,-0.2050593346,-0.2998977602,0.0458185077,0.0883684084,-0.0854751095,0.0337928534,-0.4702097178,-0.0412395895,-0.1352412403,0.1471666694,-0.1011154279,-0.2601190507,0.0451041162,0.1402022392,0.2343097031,0.2619287074,-0.1562951207,0.249656111,0.072859861,-0.0985257998,0.170527637,0.0140766911,0.1973771006,0.2325736582,0.3065911233,-0.256477952,0.4207333028,0.0621999949,0.2470007688,-0.1252539605,-0.1160264239,-0.0663903952,-0.1322397888,-0.2821481824,0.2717025578,0.1526796967,0.1508155167,-0.1591506153,0.1688978076,-0.1398805082,0.133980453,-0.5048725009,0.3854015768,0.3395011425,0.1274992526,-0.2040971071,0.2874664366,-0.3802792132,0.1660722047,0.2698231339,0.0666692182,0.1398975104,0.2488579005,0.0307346527,0.1064167246,0.6256422997,0.1565334946,0.0580626056,-0.402463764,-0.2386450469,-0.5080851316,0.0719949603,0.0885119215,-0.3129577339,-0.0073909881,0.0532940924,0.2365839928,0.3540362418,-0.1684101671,-0.2577153742,0.1331724375,-0.0545119718,-0.2838410437,-0.1130408421,0.2317336053,0.2961946726,0.0627538413,-0.3775100112,0.1264815181,-0.2398353219,0.066814661,0.2509346008,-0.1073643193,-0.206617415,-0.1168395653,0.3551855981,-0.1814097166,0.1029653996,0.0332704969,-0.0983114615,-0.111162588,-0.0830923915,-0.2492653131,0.0040516709,0.0883747861,0.4344706535,-0.4324679673,0.126624167,-0.5154191256,-0.1808874756,0.1386324763,-0.1103626564,-0.2050598711,0.037108127,0.0302581545,-0.0312015917,0.0093016159,0.4774017632,-0.2749958336,-0.261623174,-0.265099436,-0.1257013977,0.5387348533,-0.3954185247,-0.1003002226,-0.0282263905,-0.0324403867,-0.0426292121,0.1789043546,-0.0998637006,0.2147548944,0.2795602679,0.0839397013,-0.1748629063,0.2535271049,0.149110049,-0.0945383385,-0.0824439079,0.1646547616,-0.0835357085,0.10624028,-0.0935260281,-0.3627170324]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2508","title":"Load Image Classification Dataset from Local ","comments":"Yep that would be sweet - closing for now as we found a workaround. ","body":"**Is your feature request related to a problem? Please describe.**\r\nYes - we would like to load an image classification dataset with datasets without having to write a custom data loader.\r\n\r\n**Describe the solution you'd like**\r\n\r\nGiven a folder structure with images of each class in each folder, the ability to load these folders into a HuggingFace dataset like \"cifar10\".\r\n\r\n**Describe alternatives you've considered**\r\n\r\nImplement ViT training outside of the HuggingFace Trainer and without datasets (we did this but prefer to stay on the main path)\r\n\r\nWrite custom data loader logic\r\n\r\n**Additional context**\r\n\r\nWe're training ViT on custom dataset\r\n","comment_length":14,"text":"Load Image Classification Dataset from Local \n**Is your feature request related to a problem? Please describe.**\r\nYes - we would like to load an image classification dataset with datasets without having to write a custom data loader.\r\n\r\n**Describe the solution you'd like**\r\n\r\nGiven a folder structure with images of each class in each folder, the ability to load these folders into a HuggingFace dataset like \"cifar10\".\r\n\r\n**Describe alternatives you've considered**\r\n\r\nImplement ViT training outside of the HuggingFace Trainer and without datasets (we did this but prefer to stay on the main path)\r\n\r\nWrite custom data loader logic\r\n\r\n**Additional context**\r\n\r\nWe're training ViT on custom dataset\r\n\nYep that would be sweet - closing for now as we found a workaround. ","embeddings":[-0.1886762679,-0.2111069709,0.0259518363,0.4207182229,0.2512441278,-0.081174098,0.1948154122,0.1049177945,0.2327822,0.3297686577,-0.1712318212,0.1252407879,-0.2363784611,0.3856712282,0.0491452478,-0.2130761445,-0.2529838085,0.3919166327,-0.1945611835,-0.1751778275,-0.036593724,-0.0908078328,0.1545467973,0.0123157827,-0.3192312121,0.1354777813,-0.24732171,0.5201258063,-0.1240555719,-0.1091669202,0.4604955912,0.0274884477,0.3040930331,0.4868181646,-0.0001068472,0.0720115677,-0.0899762958,-0.2837551832,-0.0680224523,-0.1697876453,-0.0618719421,-0.0788135082,0.2045942545,-0.1830083877,-0.5179725289,0.0054793865,0.0360626429,-0.4050836563,0.4306099117,0.1914859414,0.1694279313,0.1369747519,-0.3749600053,-0.0619755276,-0.1921081841,0.5134592056,-0.1910932511,0.3487407267,0.4570524991,-0.1313595176,-0.1428183764,0.2557406127,0.1963772774,-0.0298613142,0.5546407104,0.2590894103,-0.1847518981,-0.1755548269,-0.1541946232,0.2354571372,0.3180779517,0.0547957383,-0.043065507,-0.1167956963,0.1160133481,-0.0628652424,0.118917264,0.032766439,-0.4303638339,0.2771244645,-0.275503546,-0.2302494794,-0.5004920363,0.072638236,0.0717259347,-0.4859114885,-0.0955550745,-0.0243719444,0.1358907968,-0.0519038588,-0.202296108,0.1686824858,0.0598122701,0.4052997231,-0.1859391332,-0.2544473708,-0.0306581911,0.0468857065,0.0968627408,-0.1613052636,0.1522735506,0.0271884557,-0.4556382596,0.0831126422,0.1840142757,0.123074986,-0.0728213638,0.2156217694,-0.0761707947,0.020621622,-0.4176223874,-0.156705752,-0.2805726528,0.0293707643,-0.264844209,-0.0737600848,-0.0346610285,-0.1694330126,0.2074032277,-0.1815536022,0.1530923545,0.2163883746,0.167221874,0.2990125716,-0.0204191934,-0.086250931,0.083900325,-0.1788126677,-0.2248739898,-0.1573250741,-0.1704981476,0.0834947526,-0.3529962897,0.1539865583,0.0876948386,-0.2489308864,0.0744474381,-0.200294435,0.4810881913,0.2002457082,-0.0297470354,-0.1332981735,0.3405076563,0.1555591375,-0.3364633918,-0.2068179101,0.2967258394,-0.0243551247,-0.1948940903,0.1142061651,-0.3980308771,-0.2876571417,0.1048900187,0.1378747523,0.0861115009,-0.100459002,-0.3218684196,0.1513062567,-0.4894645512,-0.0354535952,-0.2414630502,0.0209809747,-0.3560899794,-0.0236280076,0.323242873,0.4308150113,-0.3840063214,-0.1856369972,-0.1955722421,-0.1828946918,-0.066821374,0.3184571564,-0.3097456694,0.2534507513,-0.2341014892,0.0362450331,0.267285347,-0.4901867807,-0.1140361875,0.2814023197,-0.2302235812,0.1148289368,-0.1403031498,0.4464541972,0.0510438606,-0.0228721779,-0.1673348397,0.6676790118,-0.2692114413,0.085491851,0.0125718508,-0.357696265,-0.0721535236,0.3716855049,-0.0263674408,0.303170234,0.1524233967,-0.1078434438,0.0135808447,-0.2610098124,0.0274742823,-0.03094754,0.2001755685,0.2589576542,0.0084528197,0.1386582106,-0.3657492995,0.4044002295,-0.0850106776,0.2478879392,0.0571117662,-0.0750035197,-0.1184800938,-0.2347795218,-0.2010245323,-0.2141424865,0.038004268,0.1959475875,-0.1098103449,-0.1741458327,-0.2714770734,0.3022443056,-0.4138531685,0.1242265627,0.1542907059,0.0806485564,0.1807566732,-0.0079021016,-0.0984393954,-0.0138807157,-0.1468479186,-0.1943733394,-0.1448952556,0.201001972,0.0808688477,0.2900575697,0.2213886082,0.0992184505,0.2037042826,-0.3145465851,0.025728615,-0.1150033474,0.2168543488,0.0753654763,-0.3089154363,0.4189842641,-0.0176287685,0.2039635777,0.1366123855,-0.0969603807,0.0412735417,-0.1847309321,0.0481159501,-0.0777521804,0.0984493718,0.4382323325,0.3251095414,0.2041404098,-0.2440632135,-0.1623482406,0.0286384542,0.1426482648,-0.049754627,0.2693118453,0.0367684066,0.2275295407,0.0826457217,-0.20442608,0.4000048935,0.1647106707,-0.1313389093,0.1279948652,0.1336467862,-0.0274786707,-0.0492144115,-0.0243373066,0.1123504043,-0.0507093929,-0.0427080207,0.0963129401,-0.3918791115,-0.0592534095,-0.1463610828,-0.0253697969,0.0079914741,0.213837266,-0.1473637223,-0.4472160935,-0.2030930072,-0.0730712786,-0.0915465504,0.3939895034,0.1113134325,-0.1128744856,-0.1333403885,0.0838942081,0.0971707106,0.8364970684,-0.084485434,-0.3744972646,0.102167502,0.0082857376,-0.2548112273,0.1501130015,0.1946913451,-0.0443976335,0.3770352304,0.0116089433,0.1229370683,-0.2772678733,-0.243027404,0.2331575602,-0.0162586048,0.3008849323,0.0091334227,0.1748001724,0.0508300252,-0.1765118837,0.2870744467,-0.0933634266,-0.2921841443,-0.3772734106,0.197113812,0.3548741639,-0.2070067376,-0.3806007504,-0.4039874077,-0.381519109,0.3130661547,0.3168607354,0.2121033967,0.2155041248,0.1812317967,0.0356069691,-0.2485938966,0.1082441062,0.0212990586,-0.525978148,-0.0442334749,-0.2578098178,-0.2370121926,-0.1707788855,-0.1015114486,0.1129675433,0.0609078854,-0.4162604213,-0.2152824998,-0.0432725623,0.1510328501,-0.0983734652,-0.0977986455,0.2016440928,-0.1030391306,-0.10765443,-0.0466415435,0.2366515249,0.169855088,-0.020706933,-0.1491356194,0.1399475038,0.0922459289,-0.0838694498,0.3661858737,-0.1385658681,-0.1786674559,0.4900537133,-0.1863448769,0.3239941299,-0.5196755528,-0.4126216769,0.23341842,0.2231251299,-0.3081181943,0.3087413609,0.3701803982,-0.0061039869,-0.3232180178,-0.1084538549,0.0864061192,-0.2573907077,0.3719649315,0.3204576671,0.3113158047,-0.3446988761,-0.1278562546,-0.1769230068,-0.0544801168,-0.012496707,0.5363087654,0.2383389026,0.0158008859,-0.3489562869,0.2840011716,-0.052297879,0.0391970463,0.1792689711,-0.0976695716,-0.3398566544,-0.1166902632,0.0867073163,0.1956863105,0.332465142,0.0758615285,0.0625564232,0.1324739605,-0.5369899273,-0.0462294631,0.1135937944,0.1304973215,-0.1589531749,-0.2376618236,0.6659223437,0.1125360653,-0.3397164941,0.1510480046,0.0692621693,-0.3529451489,-0.0382648297,-0.093948938,0.1429973245,-0.2641768157,0.041362118,0.0098300781,0.1945202053,0.2784324884,0.0788051859,-0.1705152839,-0.0501493961,0.0751522556,-0.0165253989,-0.0027183876,0.3210842609,0.2628196478,0.2650375664,0.3896450996,-0.011195221,0.4680131674,-0.2052229941,-0.4266871512,0.0510747507,0.0175106321,0.3602543771,0.397424221,0.1669209599,0.0650434494,0.037151739,-0.0455659293,-0.7420886159,0.5788740516,0.0994194523,-0.1949786693,0.0854951143,-0.297183305,0.3778253496,-0.1893396676,-0.068158716,0.1823446453,0.1883853525,-0.2875302434,0.0256934632,0.3705444336,0.9427644014,-0.1571028829,0.5236108899,0.1488693506,0.2530037761,0.2335362881,0.2440981865,-0.0898443907,-0.1815682352,-0.4510159194,-0.0747257099,-0.1075656712,0.1243050471,0.0175847095,0.0555790067,-0.037424989,0.0472627319,0.0197377633,0.0523444898,0.2100940347,0.1437539607,-0.6261909604,-0.217015326,0.2043070644,0.1420521736,0.3734578192,-0.0588302128,-0.0991917849,-0.0168156568,0.0095277457,0.0288431048,-0.102549769,-0.160813272,0.0025743863,-0.0119074425,-0.1988095492,-0.0237489138,0.2767211795,0.2692287862,0.1334885657,-0.1686823219,0.2248770744,-0.280862391,0.1743305176,-0.0966649801,-0.2264118642,0.4135711789,-0.0023452251,0.0572974607,-0.0622005276,-0.0598381348,-0.4739028811,-0.0451767966,-0.1543251276,0.5844867229,-0.1830486357,-0.1145310774,-0.1397281885,0.0632798597,0.0979997963,0.154435128,0.1759676486,0.0867640898,-0.103600204,0.3364907205,-0.1952341944,0.0790412053,0.2312387228,0.0383483395,-0.4692230225,0.2331270278,0.1021107882,0.1515903622,-0.2425744087,0.4904971719,0.6527702212,-0.1369202584,0.1322085112,0.2864371836,-0.1616901159,-0.1218145117,0.7785048485,0.3183022738,-0.375407666,-0.2007907927,-0.1027071476,-0.5522806048,0.158525452,-0.2521063685,0.3436013162,-0.0455510579,0.0419172458,0.2051225901,0.0561944693,-0.3653967977,0.1072389111,-0.3045700788,0.1057962999,0.461204648,0.188941583,0.3184305131,-0.0116769904,0.0491147339,-0.0248842463,-0.1901056021,-0.1368205547,-0.1964838505,0.1436837465,0.0700952634,0.1653320789,0.0132180713,-0.1935999095,-0.1760112196,-0.0517959669,0.1525475383,-0.1505454928,-0.154184863,0.1229643375,-0.205832243,0.0162480809,-0.1611086875,0.2404272556,0.0500516929,0.2822113335,0.0872723758,-0.4003174305,-0.3019441962,-0.291031003,0.2674355507,0.2314512879,0.1421723515,0.0287216362,0.1238674074,0.2088850141,0.1736758202,0.0286901947,0.2841961086,-0.2751663923,-0.1734536886,0.1666865498,0.5255389214,0.2431488931,-0.1379144043,-0.2056378275,-0.0298969205,-0.2649646699,0.1679307371,0.2660008967,-0.1864689291,0.3659419119,-0.0028816978,-0.1636115909,0.0595598109,0.1019721255,-0.0360138565,0.1206549779,-0.3696935177,0.0777585506,0.1605892479,0.1517809927,0.1647694111,0.3835868239,-0.5289012194,0.167697832,-0.3943482339,0.1047057956,0.3723470867,-0.3754980266,0.0677507967,0.0136040449,0.2016780823,0.1288521439,-0.385760963,0.373013258,-0.2089536488,-0.3937515616,-0.0443030633,0.0968052894,-0.2326326817,0.1015165225,-0.5035973191,-0.1113986373,-0.1923737973,0.1994361728,-0.1202485263,-0.0935886651,0.0401419327,0.0709852353,0.1243532225,0.2939528227,-0.2323836684,0.199571684,0.0606051125,-0.0644198805,0.2141497135,-0.0416806825,0.1890438497,0.3697752953,0.3727453649,-0.3047144115,0.3642728329,0.1633383334,0.144950971,0.0021390987,-0.1041252613,-0.1012499258,-0.0923695415,-0.1800383776,0.1914107054,0.2088695914,0.151544705,-0.1956516057,0.1984682232,-0.2054088712,0.2399984449,-0.532220006,0.255684346,0.4123972058,0.0832966343,-0.2408373058,0.2311385721,-0.3130533099,0.163782686,0.1242028922,0.0397597365,0.0749649405,0.1685434431,0.0610238686,0.0425303169,0.7164620757,0.0678377002,-0.0622111969,-0.4263786972,-0.3025923371,-0.5433431864,0.1142451242,0.1481647938,-0.2015092373,0.0381500758,0.0321704,0.2882297039,0.351501137,-0.0543548241,-0.3175209165,0.2110997289,-0.108780995,-0.3162995875,-0.0860574841,0.3458501101,0.1808620542,0.084256798,-0.2712035775,0.2391748428,-0.1183575019,0.0961945653,0.2811901867,-0.1172475517,-0.2156408429,-0.0834076628,0.387796253,-0.1100727096,0.0393735841,0.0223446861,-0.1182418391,-0.1589830816,-0.1766626239,-0.2595826089,-0.1310537159,0.1386948377,0.5243890285,-0.3780395091,0.1152591929,-0.5571206808,-0.148279205,0.1783438027,0.0008647145,-0.2203398347,0.0276066232,0.0337712206,-0.0751646981,0.0919342637,0.4980617762,-0.2210524976,-0.3598721623,-0.3902246952,-0.1463922709,0.532938838,-0.3308367133,-0.1287924349,-0.0358020552,-0.0102960868,0.0447896495,0.0890759528,-0.1289822012,0.1658098102,0.2047533244,0.1484396607,-0.1726453602,0.2763156891,0.0601439811,-0.0200511962,-0.0815344527,0.2234620601,0.0551929213,0.1248029023,-0.0889974684,-0.3690285981]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2508","title":"Load Image Classification Dataset from Local ","comments":"@lhoestq I think we'll want a generic `image-folder` dataset (same as 'imagenet-like'). This is like `torchvision.datasets.ImageFolder`, and is something vision folks are used to seeing.","body":"**Is your feature request related to a problem? Please describe.**\r\nYes - we would like to load an image classification dataset with datasets without having to write a custom data loader.\r\n\r\n**Describe the solution you'd like**\r\n\r\nGiven a folder structure with images of each class in each folder, the ability to load these folders into a HuggingFace dataset like \"cifar10\".\r\n\r\n**Describe alternatives you've considered**\r\n\r\nImplement ViT training outside of the HuggingFace Trainer and without datasets (we did this but prefer to stay on the main path)\r\n\r\nWrite custom data loader logic\r\n\r\n**Additional context**\r\n\r\nWe're training ViT on custom dataset\r\n","comment_length":25,"text":"Load Image Classification Dataset from Local \n**Is your feature request related to a problem? Please describe.**\r\nYes - we would like to load an image classification dataset with datasets without having to write a custom data loader.\r\n\r\n**Describe the solution you'd like**\r\n\r\nGiven a folder structure with images of each class in each folder, the ability to load these folders into a HuggingFace dataset like \"cifar10\".\r\n\r\n**Describe alternatives you've considered**\r\n\r\nImplement ViT training outside of the HuggingFace Trainer and without datasets (we did this but prefer to stay on the main path)\r\n\r\nWrite custom data loader logic\r\n\r\n**Additional context**\r\n\r\nWe're training ViT on custom dataset\r\n\n@lhoestq I think we'll want a generic `image-folder` dataset (same as 'imagenet-like'). This is like `torchvision.datasets.ImageFolder`, and is something vision folks are used to seeing.","embeddings":[-0.22038351,-0.1735643446,0.0379923396,0.4659894109,0.1979663372,-0.0509574749,0.1801866144,0.0725658014,0.3113462329,0.2405235469,-0.1453477442,0.1418926269,-0.2927550673,0.2843796909,0.0848555639,-0.2747689784,-0.2471704185,0.3313145638,-0.2056817263,-0.1957011819,-0.0449201725,-0.1257203519,0.1448772848,0.0041875867,-0.2547867894,0.1277288049,-0.2468128949,0.4981716275,-0.0797515661,-0.084591344,0.4612981081,-0.0093378248,0.3770538867,0.4620998502,-0.0001086435,0.0818984658,-0.0611452647,-0.2880570889,-0.0763202757,-0.1762279719,-0.014972847,-0.0790724382,0.2707419991,-0.1627689153,-0.4990906715,-0.0147909168,0.0881997123,-0.4173053205,0.4179804921,0.1468567997,0.143450588,0.1118765622,-0.373375535,0.0350770019,-0.1778990924,0.5875381827,-0.1875459254,0.3944593072,0.4111532271,-0.1405942887,-0.1894525886,0.2388451099,0.0990793332,-0.0389614515,0.6692720652,0.2804870307,-0.2617646754,-0.1706671715,-0.2399690151,0.2066524029,0.3181587458,0.0006301517,-0.0447476208,-0.1067001671,0.0801901817,-0.0609744899,0.1026630104,0.1048561856,-0.4013625681,0.2154264152,-0.3154778183,-0.2255514413,-0.5069895387,0.1197830141,0.0038086374,-0.4213098586,-0.0634872317,-0.0050006481,0.1666987985,-0.0544408113,-0.0993967354,0.1049668789,0.0811220929,0.422881633,-0.1536212564,-0.2187401056,-0.0559955835,0.1574492306,0.0980690345,-0.2005217075,0.2193757445,-0.0241949931,-0.4958957136,0.1277993172,0.1326784939,0.098014541,-0.0123289563,0.3163379729,-0.1417284906,-0.0399906933,-0.3384733498,-0.0842937827,-0.3632296324,0.0447499007,-0.2670950294,-0.0891438723,-0.1156796888,-0.1065340936,0.15916273,-0.2201787084,0.1231757551,0.2107804269,0.1692010462,0.2219041735,-0.0250553768,-0.077038534,0.1714544445,-0.1723672003,-0.1978271902,-0.1410319507,-0.1429929435,0.1034082323,-0.3244977593,0.1915985048,0.0878314152,-0.2593480647,0.0646570921,-0.2324134707,0.3678420186,0.2796503007,-0.0336758606,-0.1119669452,0.325874418,0.1366643757,-0.3914694488,-0.2433126122,0.2597413957,0.0465238951,-0.2552483678,0.1332139969,-0.3724262118,-0.2556294799,0.054665979,0.1216500849,0.1061291769,-0.1501431316,-0.3313512504,0.1637575626,-0.4754866064,-0.0514059588,-0.1621988565,-0.014872604,-0.3337172866,-0.0651122704,0.3065491915,0.4737496972,-0.4165000916,-0.1793348193,-0.2316373587,-0.1584114134,-0.106385015,0.3455013633,-0.2947668731,0.2779908478,-0.2246076465,-0.0050772247,0.2705891132,-0.5349384546,-0.0601748042,0.3036560714,-0.2119434923,0.0305139478,-0.1640352458,0.5342690945,0.050146699,-0.0419990346,-0.1745987982,0.6611357927,-0.3526691198,0.0466846749,0.0780655146,-0.3422687352,-0.0644625053,0.4122166336,-0.0797401443,0.2406919003,0.1318224669,-0.1193682402,-0.0236289036,-0.2165368199,-0.0034816517,-0.0660730749,0.0811402947,0.2458647192,-0.0730326772,0.1529961377,-0.3992540836,0.3711704612,0.0097054522,0.190908879,0.0310494136,-0.0621119663,-0.0510725304,-0.2172994465,-0.1721381545,-0.1460330039,-0.0090472726,0.1679480225,-0.1434604377,-0.2179095447,-0.2863901258,0.3025023043,-0.4228336513,0.1167783141,0.1282292902,0.1148817763,0.1717139035,0.0182773247,-0.1940162629,-0.0566648021,-0.1676882505,-0.2380437553,-0.0803513601,0.1865678281,0.0075435643,0.227512762,0.2719519734,0.1917624623,0.2524004281,-0.3821576238,0.1531851739,-0.150438562,0.21512492,0.0628465265,-0.3206401765,0.3980116248,-0.0454428792,0.2160473764,0.0668012053,-0.1179174036,-0.0077410075,-0.1440976858,0.0915236101,-0.0523050018,0.1271104515,0.4711931944,0.2664765716,0.2243493348,-0.218883127,-0.1921231449,-0.0595937409,0.0769026354,-0.0112240668,0.2901335359,-0.0028836692,0.2898470163,0.103418,-0.2469832301,0.4035187662,0.16002886,-0.171587497,0.1276348531,0.1330762953,0.0537488051,-0.0414684862,-0.0240612309,0.100619033,-0.0882860795,-0.1025520265,0.1620601118,-0.3434333503,-0.1346260458,-0.1863878667,-0.0674704611,0.0368096866,0.2644503415,-0.0366185941,-0.4857848883,-0.2420701385,-0.0306953862,-0.1192214787,0.4429273903,0.0953442305,-0.125559181,-0.059894044,0.0385999344,0.0327109471,0.8698264956,-0.0177567732,-0.4855663776,0.1011931449,0.0124969902,-0.2378597111,0.1273885071,0.1968029737,-0.1020843759,0.400246948,-0.0251910165,0.1554798633,-0.2948789299,-0.2264708877,0.2518216372,-0.0938105732,0.2748894691,0.0254154988,0.1629122645,0.0268967357,-0.2239453495,0.242792964,-0.0728503987,-0.2957368493,-0.4095212519,0.1939206272,0.3621399999,-0.2087424397,-0.3324574828,-0.4163191915,-0.3412550688,0.3048316538,0.3539651632,0.2217598855,0.1651599556,0.1471845955,0.0040501459,-0.2702941,0.0946021155,0.114448905,-0.4609877467,-0.0590890162,-0.1896750182,-0.2067407817,-0.2203400135,-0.1727891713,0.1196982041,0.1568202525,-0.45949471,-0.1874851137,0.0145020094,0.2022879869,-0.0684157014,-0.1215318069,0.269607842,-0.0686749667,-0.0675727054,-0.0113547156,0.1930972934,0.1305381954,0.0150990132,-0.0766622573,0.2141602486,0.1200874522,-0.1289938092,0.4839248359,-0.1341385692,-0.2982962132,0.458688587,-0.1482966542,0.2997557223,-0.4732166827,-0.3920402825,0.2545677722,0.2472583205,-0.2930373847,0.2879585922,0.3745384812,-0.0634125993,-0.3145258725,-0.160968408,-0.0095566204,-0.1770733297,0.4041850865,0.2089851052,0.3410283029,-0.4403284192,-0.1249426007,-0.1967133582,-0.0869488418,0.0050696991,0.5884594917,0.2764629126,0.0240621753,-0.2165913582,0.3037583232,0.0321437344,0.0399330966,0.1948237568,-0.1237633154,-0.2956554294,-0.1122803614,0.0854419097,0.1938694417,0.284878701,0.0979407355,0.048267737,0.0827790052,-0.4667034745,-0.0958475918,0.1218658313,0.1544507295,-0.1674496084,-0.289426446,0.5744000077,0.0712845996,-0.3223757446,0.1751597971,0.1193183661,-0.3754455149,-0.0278297551,-0.0399079397,0.1829688996,-0.2103995383,0.017481856,-0.0102238515,0.148694098,0.3407164216,0.0679077655,-0.155469805,-0.0782486424,0.0654698163,-0.0717564523,-0.029060822,0.2698785365,0.2421383411,0.3197583258,0.4253334999,-0.0151428794,0.5078399181,-0.1578489095,-0.3838399947,0.0121014351,0.1267116368,0.3334471881,0.3710128963,0.1919393241,-0.038432125,0.0186787024,-0.137051031,-0.815870285,0.5896148682,0.0729693398,-0.2231980413,0.0463674292,-0.2987622321,0.4696885347,-0.1042820737,-0.1003978997,0.2121422142,0.1468672007,-0.3020868003,0.0595265701,0.4889581501,0.9659129381,-0.2565289736,0.5424615145,0.0919431224,0.2645860016,0.2019232363,0.2472116351,-0.0507729352,-0.2140336633,-0.385464251,-0.1326998025,-0.1157030985,0.0899960697,0.0280095357,0.1046664268,-0.0004791431,0.0318760686,0.0459690653,0.0555038154,0.1865988672,0.1125454605,-0.5767173767,-0.0885066688,0.1656726748,0.097872287,0.3691044152,-0.0501510352,-0.0938527212,-0.0038341619,-0.0055170078,0.0119551048,-0.1070672274,-0.1335557401,0.0020826652,0.006706689,-0.2064534873,-0.0368454494,0.3239462972,0.1903465837,0.0462059416,-0.146029219,0.1698487252,-0.2683291435,0.1435433477,-0.0371730067,-0.257273823,0.427156508,0.0603845678,0.1232574433,-0.0871603414,-0.0342846401,-0.4487549663,-0.1154086441,-0.1401174068,0.6795423031,-0.2016599029,-0.0441879183,-0.1713184416,0.0454779156,0.1947475821,0.1356179863,0.1881379783,0.1464863718,-0.1470835954,0.3069872856,-0.1632525176,0.0422835685,0.1818506867,0.0206909031,-0.483143121,0.1389982551,0.0571578145,0.1379521042,-0.2186934948,0.5137709379,0.6890600324,-0.1856106222,0.0950149223,0.255015552,-0.1769702137,-0.080960691,0.7742312551,0.3607247472,-0.4379698336,-0.2155485451,-0.0806499422,-0.5159037709,0.1880480796,-0.3054280877,0.3426116705,-0.067124106,-0.0205329508,0.1677675098,0.0446032472,-0.3304334283,0.0615924485,-0.2406257689,0.1268831491,0.4618703723,0.2617571056,0.2810432017,0.0487006269,0.047938291,-0.0934771746,-0.1809422523,-0.0897052586,-0.1819967479,0.156942606,0.0965894088,0.1295959502,-0.0247125179,-0.2508901358,-0.1646745205,-0.076592952,0.1185127869,-0.2117545307,-0.1617012024,0.1377967745,-0.1535950601,0.0359751694,-0.1333457977,0.2676193118,0.0069106747,0.3529626429,0.0898396745,-0.3657232821,-0.2456702888,-0.3413461447,0.2444539517,0.241866976,0.1333364397,0.0059953001,0.1923787743,0.260199368,0.161518693,0.0354914702,0.3301908672,-0.2450646013,-0.1639598012,0.1343403757,0.479151547,0.2119674236,-0.130901441,-0.2089427114,-0.0889953598,-0.2090006471,0.1344104111,0.2951006889,-0.2345120162,0.3950603306,-0.0475567393,-0.190367192,0.1842797697,0.1106932238,-0.062436305,0.14889431,-0.315552026,0.06094658,0.2417805195,0.1907817274,0.173851341,0.4474363625,-0.5088698268,0.2313016951,-0.3525955081,0.1327976882,0.4371777475,-0.3301294148,0.088963598,-0.0813509673,0.2755255699,0.1303029209,-0.4082705975,0.260866791,-0.1979530901,-0.3885847628,-0.0071525485,0.0998762771,-0.1769783199,0.1098461598,-0.5392649174,-0.1054767817,-0.0967821926,0.2101845741,-0.0788780302,-0.1732392311,0.1305887252,0.1152082756,0.1759532392,0.2512292564,-0.1543103755,0.2087507397,0.0609660074,-0.0560695678,0.2293683738,-0.0510763973,0.1742599905,0.3724264205,0.3731933534,-0.3360878825,0.3672221601,0.1079408303,0.197318092,-0.0604234524,-0.1145716906,-0.0751977637,-0.0853597596,-0.2778794765,0.219888553,0.132054165,0.1360963583,-0.13982445,0.2305925786,-0.2897707522,0.1470107734,-0.441347152,0.2855165005,0.3836601079,0.1116193086,-0.2781771123,0.2666491866,-0.3031942248,0.155796349,0.1269330978,0.0470831804,0.0562157594,0.2201367617,0.0491623953,0.0408185497,0.6757802367,0.1230631694,-0.0002568653,-0.4220190942,-0.2528709471,-0.4494131207,0.1086915955,0.1498080343,-0.2966712415,0.0146448426,0.0192303658,0.2893818021,0.3351863623,-0.1327323318,-0.2538819909,0.274379909,-0.0767371729,-0.2786461413,-0.1292854995,0.3408537507,0.220634982,0.056782037,-0.285138756,0.2080623806,-0.147838518,0.0750451311,0.2263497859,-0.1668165475,-0.1618995219,-0.0848925933,0.4018754661,-0.1649273634,0.088662684,0.0323633589,-0.11470972,-0.1409032643,-0.1237448454,-0.2987914383,-0.1430527121,0.1465636194,0.4959648252,-0.4086167812,0.1019576937,-0.5701278448,-0.1637083739,0.1276887953,0.0118020885,-0.2440306991,0.0364947468,-0.0801195353,-0.0949007645,0.0979750231,0.4745466709,-0.1897490919,-0.3523243368,-0.360465467,-0.150929302,0.5336057544,-0.3622514606,-0.0179837607,-0.000638146,-0.1082384959,0.0579233579,0.1308809519,-0.1344611645,0.1527675241,0.2543122768,0.1060827747,-0.1274393052,0.2331526577,0.1409254074,-0.0177376736,-0.1020818353,0.1804926842,-0.0122538032,0.1106826514,-0.0972425118,-0.340736866]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2508","title":"Load Image Classification Dataset from Local ","comments":"Opening this back up, since I'm planning on tackling this. Already posted a quick version of it on my account on the hub.\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset('nateraw\/image-folder', data_files='PetImages\/')\r\n```","body":"**Is your feature request related to a problem? Please describe.**\r\nYes - we would like to load an image classification dataset with datasets without having to write a custom data loader.\r\n\r\n**Describe the solution you'd like**\r\n\r\nGiven a folder structure with images of each class in each folder, the ability to load these folders into a HuggingFace dataset like \"cifar10\".\r\n\r\n**Describe alternatives you've considered**\r\n\r\nImplement ViT training outside of the HuggingFace Trainer and without datasets (we did this but prefer to stay on the main path)\r\n\r\nWrite custom data loader logic\r\n\r\n**Additional context**\r\n\r\nWe're training ViT on custom dataset\r\n","comment_length":33,"text":"Load Image Classification Dataset from Local \n**Is your feature request related to a problem? Please describe.**\r\nYes - we would like to load an image classification dataset with datasets without having to write a custom data loader.\r\n\r\n**Describe the solution you'd like**\r\n\r\nGiven a folder structure with images of each class in each folder, the ability to load these folders into a HuggingFace dataset like \"cifar10\".\r\n\r\n**Describe alternatives you've considered**\r\n\r\nImplement ViT training outside of the HuggingFace Trainer and without datasets (we did this but prefer to stay on the main path)\r\n\r\nWrite custom data loader logic\r\n\r\n**Additional context**\r\n\r\nWe're training ViT on custom dataset\r\n\nOpening this back up, since I'm planning on tackling this. Already posted a quick version of it on my account on the hub.\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset('nateraw\/image-folder', data_files='PetImages\/')\r\n```","embeddings":[-0.2027493119,-0.188690275,0.0131782666,0.4011635482,0.2593050301,-0.0279740077,0.1975970417,0.1451179683,0.3119898438,0.2857654095,-0.1935351789,0.1235475317,-0.3047415912,0.3270210922,0.0650774762,-0.2292608172,-0.2278257012,0.3378814161,-0.1562044919,-0.2475795001,-0.0562427118,-0.0963324755,0.1660833657,-0.0088663884,-0.3098389208,0.1793593615,-0.2453823984,0.5447191,-0.1185016185,-0.1016760245,0.4887914062,0.0333637074,0.3943757117,0.5124262571,-0.0001068552,0.087266162,-0.1294454187,-0.265117079,-0.0608895496,-0.2519886792,-0.0206751078,-0.1543317139,0.2418370992,-0.2121085674,-0.4997153878,-0.0551615655,0.0461996533,-0.3964985013,0.4658918977,0.1928687394,0.174645111,0.126873374,-0.342350632,0.02665288,-0.2240594476,0.4555782974,-0.1381314993,0.4050805867,0.4020667374,-0.2422726005,-0.0563127808,0.2717686594,0.0958084315,-0.0036633222,0.594912529,0.3227715492,-0.1611328721,-0.20971632,-0.185771212,0.1864373982,0.1650446355,-0.0789068192,-0.1229935661,-0.1892203987,0.1058700681,-0.1091682389,0.0645306259,0.1194690764,-0.4392844141,0.2350859642,-0.2582420409,-0.1784317493,-0.4480182528,0.0994268954,0.0262657702,-0.3858397901,-0.0895731077,-0.0566947088,0.128129527,-0.0233988855,-0.1673009396,0.0837891847,0.1429525465,0.4985156953,-0.1698465496,-0.2242092639,0.0119308187,0.0944930315,0.0747527033,-0.1614055783,0.1179353595,-0.0337839611,-0.453022927,0.0725452825,0.1935688406,0.1195869669,-0.0463378094,0.3049668074,-0.1034985706,0.0002569417,-0.4169119,-0.1191155016,-0.3012642562,-0.0344271734,-0.1744806916,-0.0468035527,-0.0193034932,-0.138956368,0.1612506807,-0.1624351442,0.1029184312,0.1908903569,0.1504067928,0.2424588352,-0.0482725091,-0.0031934674,0.1031282693,-0.1202344671,-0.2283467352,-0.0636385828,-0.1917521209,0.1401709765,-0.2598928511,0.1119391173,0.075955689,-0.2991689742,0.0413861088,-0.2150568664,0.472856164,0.2309122831,-0.0587947406,-0.156492576,0.3177806139,0.1721372455,-0.3297017217,-0.2200269848,0.3412006199,-0.1647662222,-0.2446691692,0.0908580944,-0.365121156,-0.2542339861,0.09029001,0.1549029052,0.0580705926,-0.1307823509,-0.3527176082,0.1504351497,-0.4175845683,0.0191522017,-0.1944582164,-0.007577274,-0.3421984911,-0.0516983457,0.347889632,0.4690683186,-0.3932281435,-0.1652914137,-0.1618795693,-0.1397892386,-0.0641402453,0.3136607707,-0.2297160625,0.3035108447,-0.1629567295,-0.0052741012,0.2236001641,-0.5785595775,-0.1155698746,0.2601819336,-0.1376008689,0.1651307046,-0.1817654222,0.4197821617,0.0482666679,-0.0234768186,-0.1105472967,0.7269153595,-0.2688494325,0.0874660611,0.0078729391,-0.3116492331,-0.0915126875,0.4174849391,0.018892169,0.2793212235,0.0595956147,-0.0981967524,0.0894353911,-0.258109659,0.0199274793,-0.0467378534,0.2255551368,0.2770264447,-0.0013278293,0.0582767464,-0.3241295815,0.4105852544,-0.0123326601,0.2917645574,-0.1126455665,-0.0683656409,-0.1458888203,-0.1702474952,-0.2722268105,-0.134783566,0.0470264331,0.0864298418,-0.0660315603,-0.158166945,-0.3154458106,0.3485401869,-0.4256828725,0.1370666921,0.1380571127,0.1380340308,0.1871838123,-0.0216990467,-0.1657732278,-0.0027624688,-0.1812498868,-0.1727612019,-0.1171171218,0.2002925724,0.0799686238,0.2456362247,0.2190037668,0.1401597708,0.2687151134,-0.2863899171,-0.0079384455,-0.0720431656,0.2825771272,0.0857939273,-0.2918455005,0.4547752142,-0.0203835387,0.222137019,0.1012874395,-0.1100744158,0.0155491764,-0.1211861894,0.0509902425,-0.0794550925,0.1232160181,0.4209697247,0.3136687577,0.129948929,-0.1777851433,-0.1941128075,-0.0812529027,0.0585995428,-0.0019886738,0.2639885545,0.0237173978,0.2941877246,0.02845034,-0.1595571935,0.42019099,0.1286082864,-0.1448091418,0.158753857,0.0780899003,0.0449383408,-0.0498137437,-0.0107468572,0.0640473589,-0.0586436056,-0.007914192,0.1536595225,-0.3867430389,-0.1319253445,-0.2363811433,0.0018710952,-0.0079262378,0.285233885,-0.1009811908,-0.3875078857,-0.1490825564,-0.1553316861,-0.1084499732,0.3473501205,0.0478672944,-0.1340536028,-0.0920879841,0.0323764384,0.064541012,0.7125998139,-0.02485203,-0.46919626,0.1037002727,-0.0467529036,-0.1814677566,0.1466980875,0.2156877518,-0.0644049793,0.360367775,-0.0669210032,0.0659147054,-0.2202038914,-0.2478675693,0.192679435,-0.0845230892,0.339830488,0.0223369244,0.2182110846,-0.0282246172,-0.1670151502,0.313624084,-0.1251924932,-0.2248562425,-0.3721477985,0.1539634317,0.3486576378,-0.2092300206,-0.3167017102,-0.5308420062,-0.3142337203,0.3089395761,0.4514374435,0.2063345164,0.2160308957,0.1925136894,0.0203400366,-0.2060065269,0.1167829633,0.0709908679,-0.522577107,0.0260534324,-0.2402145565,-0.2574028671,-0.2038243264,-0.1880463362,0.0893688425,0.0179092679,-0.3726289272,-0.3170896471,-0.0948065892,0.1927018315,-0.0982363448,-0.0089151198,0.1456410438,-0.076976262,-0.0927020982,-0.038080804,0.2784571946,0.1374171525,0.0226473361,-0.2067890167,0.1048831344,0.0421145819,-0.0949455202,0.3632555008,-0.1189904734,-0.1858229786,0.4987353683,-0.2188987881,0.2780492604,-0.5007695556,-0.4705104232,0.1554922909,0.2024284899,-0.315797627,0.2361053526,0.3836195469,0.0417031795,-0.3428866863,-0.0829944238,-0.0110559603,-0.2468995154,0.3223659694,0.3423321545,0.2638841569,-0.3573635519,-0.1419983357,-0.2169504315,-0.0788719282,-0.0044166981,0.588270247,0.1983955204,0.005052716,-0.3221976757,0.232148245,-0.0820269212,0.009445061,0.2002619654,-0.1162261665,-0.2406812757,-0.1060224771,0.0583754405,0.1489955038,0.3460451663,0.1016142443,0.0671902522,0.1230659783,-0.5372578502,-0.1438310146,0.1007355303,0.0597226508,-0.2057904601,-0.2554807663,0.6490074992,0.0681871697,-0.3536732495,0.158528313,0.0991672128,-0.3491899967,-0.0328163318,-0.1618230045,0.0673727542,-0.2876484692,-0.0037825634,0.037604548,0.2875193059,0.2628844082,0.1315567642,-0.1734480411,-0.0176330004,0.1536099762,-0.0293117575,0.027830977,0.2857740819,0.2290541083,0.2562815547,0.3251539767,-0.0794727057,0.5093402863,-0.2159670591,-0.4752892554,-0.0389044546,0.1117705032,0.3309125006,0.3534556925,0.1149043888,0.0006218351,-0.0314127281,-0.0187642854,-0.8070137501,0.5762056112,0.0688988715,-0.1418067515,0.0866758525,-0.4121037424,0.4884509742,-0.0991752818,-0.0336984582,0.1793906242,0.2129326314,-0.2754082084,0.1061649323,0.3879108727,0.9502143264,-0.2089948654,0.5667915344,0.1677491814,0.266476661,0.2431260645,0.2348926216,-0.1047671288,-0.1616081446,-0.4155467451,-0.1051981002,-0.076686278,0.1650699973,0.0588499345,0.0765501857,0.0380688682,-0.0239418503,-0.0512513667,0.0152958874,0.1511951536,0.008729564,-0.5815856457,-0.3145025969,0.2094788998,0.1535281837,0.399163872,-0.0469239019,-0.0668687224,-0.0617358908,0.0073630312,0.026992524,-0.1028417274,-0.1125143096,0.0362899564,0.0071028783,-0.234101519,-0.1018063128,0.3608926833,0.2642953694,0.0759973153,-0.129774034,0.193892464,-0.3031798303,0.1749074608,-0.0844344646,-0.1884146929,0.403364867,0.0306339357,0.0402399302,-0.0778917223,-0.0464994684,-0.4463988841,-0.0515013263,-0.0914785415,0.594982028,-0.143631041,-0.1442608088,-0.1526285559,0.0557628572,0.1164382324,0.1565444916,0.2211786509,0.0589644089,-0.0566403717,0.2580007017,-0.2473710477,0.079993546,0.2593111694,0.0732704103,-0.4243011773,0.2710446715,0.1312050968,0.1236436144,-0.2333193272,0.4793847799,0.5929043889,-0.0943112001,0.1389596015,0.262760371,-0.1387142986,-0.0982642546,0.7520093322,0.3514798582,-0.4276151955,-0.1781028062,-0.1849619746,-0.5113028884,0.1392212808,-0.1715734601,0.3388847113,0.0048545864,0.071958892,0.1959533989,0.0211515427,-0.3629047573,0.1149833947,-0.2795414925,0.078199558,0.3643583655,0.2243508697,0.3502086699,-0.0084178178,0.085307166,-0.0262588244,-0.1596098244,-0.1616480649,-0.2193170637,0.1357162446,0.0767210424,0.1213212535,0.0447624028,-0.1669976115,-0.1648409516,-0.0850634053,0.1592330784,-0.0806108192,-0.1487305462,0.1932409257,-0.1417417824,0.0104057882,-0.1049188301,0.2175507396,0.0159711465,0.2904707193,0.1388822347,-0.340390563,-0.2393259406,-0.266735673,0.3110299706,0.1864569187,0.1296159476,-0.017090911,0.198824361,0.1754839569,0.1664516479,0.0432939976,0.3626082838,-0.2004906982,-0.2212094814,0.2235446274,0.4667614698,0.2406203002,-0.1922971755,-0.2187258005,-0.0807728469,-0.1801434457,0.2173093855,0.3076249361,-0.255292207,0.383681953,-0.1784029156,-0.128002584,0.1589686275,0.1056009009,-0.0670827702,0.1668912917,-0.292024374,0.0753034353,0.1400852948,0.1639073193,0.1505865008,0.424130708,-0.5359811783,0.2032942325,-0.4283611476,0.1026886404,0.3870523274,-0.3839820623,0.0944157392,0.0223557148,0.1899942607,0.1633550972,-0.3974065185,0.3631198704,-0.2455971986,-0.3343325555,-0.0152436523,0.1304395646,-0.1191877201,0.0357596464,-0.4416671991,-0.0613719039,-0.1514236927,0.1702166498,-0.0652250126,-0.2150897235,0.0124940854,0.0824307725,0.1849017441,0.2537665367,-0.256547749,0.2063064426,0.1319912076,-0.0657894909,0.1126946509,0.0414291471,0.1676035523,0.2787658572,0.3783047199,-0.2263225764,0.4136285186,0.0632912591,0.2188953161,-0.0334759466,-0.1237981766,-0.0826207176,-0.0444800556,-0.258687079,0.2596135736,0.2082896382,0.1723278463,-0.2008010447,0.134854123,-0.1885299534,0.2333204597,-0.5136280656,0.3722877204,0.3565174937,0.1105501503,-0.220444262,0.2379423529,-0.3508530557,0.1727730036,0.191684261,0.0668124855,0.101879552,0.1869109124,0.060096167,0.0617358163,0.6641269326,0.1494378299,0.021016093,-0.4050767124,-0.2557225227,-0.5360119939,0.0816494599,0.1062443182,-0.2939186394,0.0084801707,0.0306253023,0.2576007247,0.3980213404,-0.1194432825,-0.3155707121,0.1635521501,-0.0418917015,-0.2796112001,-0.1119437292,0.2454476058,0.2100671232,0.0810230747,-0.3084847331,0.183435455,-0.2136031538,0.1177768111,0.2491154522,-0.1012826785,-0.1914178282,-0.1190898046,0.3701430261,-0.1382643133,0.0733894706,-0.0025668193,-0.1067171842,-0.122595571,-0.1300555319,-0.2726760805,-0.0744350329,0.0721400827,0.5224917531,-0.4053119719,0.087604478,-0.488648355,-0.0944117457,0.1875325888,-0.0686932728,-0.2719266713,0.0453876182,0.0730381384,-0.0700621754,0.0103010805,0.5261645913,-0.2377746999,-0.2638458908,-0.327953279,-0.1958243698,0.558826983,-0.413307637,-0.1375681758,-0.0633119121,-0.0475287102,-0.0437315777,0.0730093196,-0.1622506231,0.2003907114,0.2113155574,0.1249167249,-0.212006703,0.3038729727,0.1336133927,-0.0588356145,-0.0838026255,0.1437307149,-0.0280817281,0.1466644108,-0.060199637,-0.3650993705]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2503","title":"SubjQA wrong boolean values in entries","comments":"Hi @arnaudstiegler, thanks for reporting. I'm investigating it.","body":"## Describe the bug\r\nSubjQA seems to have a boolean that's consistently wrong.\r\n\r\nIt defines:\r\n- question_subj_level: The subjectiviy level of the question (on a 1 to 5 scale with 1 being the most subjective).\r\n- is_ques_subjective: A boolean subjectivity label derived from question_subj_level (i.e., scores below 4 are considered as subjective)\r\n\r\nHowever, `is_ques_subjective` seems to have wrong values in the entire dataset.\r\n\r\nFor instance, in the example in the dataset card, we have:\r\n- \"question_subj_level\": 2\r\n- \"is_ques_subjective\": false\r\n\r\nHowever, according to the description, the question should be subjective since the `question_subj_level` is below 4\r\n","comment_length":8,"text":"SubjQA wrong boolean values in entries\n## Describe the bug\r\nSubjQA seems to have a boolean that's consistently wrong.\r\n\r\nIt defines:\r\n- question_subj_level: The subjectiviy level of the question (on a 1 to 5 scale with 1 being the most subjective).\r\n- is_ques_subjective: A boolean subjectivity label derived from question_subj_level (i.e., scores below 4 are considered as subjective)\r\n\r\nHowever, `is_ques_subjective` seems to have wrong values in the entire dataset.\r\n\r\nFor instance, in the example in the dataset card, we have:\r\n- \"question_subj_level\": 2\r\n- \"is_ques_subjective\": false\r\n\r\nHowever, according to the description, the question should be subjective since the `question_subj_level` is below 4\r\n\nHi @arnaudstiegler, thanks for reporting. I'm investigating it.","embeddings":[0.1086944267,0.0404847935,0.0092389937,0.1550252736,-0.3427656591,-0.0266071428,0.0926538333,0.0519785732,-0.0441118591,0.1578255594,0.0165721904,0.4199479818,0.2598315775,0.285935998,-0.2933038473,0.1767487973,0.2668679357,0.1742724329,-0.0549485721,-0.1394340396,-0.4269313216,0.2538077831,-0.5594368577,-0.0361939557,-0.1106423885,-0.22542575,-0.1312589198,-0.2302967906,-0.2374795973,-0.1925224513,0.0934379846,-0.0037480993,-0.0932141617,0.079935886,-0.0001078679,-0.2069956958,0.0723753348,0.0265175495,-0.2101933211,0.032851126,-0.2585700452,-0.0624376684,-0.1321136355,0.0476373173,-0.3968239725,-0.3248187006,-0.1967480928,-0.2597842216,0.3476752639,0.0467392504,0.1953799427,0.0513683818,-0.2066460997,0.2634503543,0.3638375103,-0.1859468222,-0.0904756635,0.1725229472,-0.1388443559,-0.0981842279,0.0976689756,0.3409767449,0.1613233984,0.0693444163,0.1005131379,-0.1086117476,-0.0711431056,-0.1612356454,0.3148467541,0.1732052416,0.3546339571,-0.0774747059,-0.3807255924,0.0419412479,-0.0937728435,-0.0971023962,0.2521139979,0.0387910232,0.0892409012,-0.0387661271,0.0827189833,0.4317024946,0.0422152504,0.0124803623,-0.0861078501,0.089768067,-0.165511474,-0.0216761716,-0.1706858277,-0.1076443344,0.0344790705,-0.0022639195,-0.5591996908,0.0556460209,-0.4330788553,0.015316953,0.0065043792,0.2419092804,-0.2224964648,-0.3837036788,0.3447864056,-0.0303069651,0.1998203993,0.1621707678,0.5325686336,0.1896781474,0.370405525,-0.1838124841,-0.1117741615,-0.2674105465,0.0201293975,-0.0054298239,0.1946291476,-0.2149305344,0.0141406162,0.0393391661,0.121081233,-0.3379256427,-0.5752221346,0.1005345508,-0.4127647579,-0.0774246454,0.1717004031,0.1102424785,0.0467311181,-0.2316418141,-0.1232305169,-0.0244657788,-0.1351023614,-0.3918811381,-0.129091382,0.025706673,-0.2309709042,0.0109899743,0.0253754351,-0.108946532,0.0048837457,0.3396919072,-0.2583560348,-0.2203959674,-0.0283312462,-0.2709927559,0.2741025984,0.1416445971,-0.2923513353,0.1809704006,0.0285246801,-0.1256355047,0.0038451659,0.3668010831,0.0527750663,0.2010226846,0.0306558125,0.1908655912,-0.0614121519,-0.0853592679,0.1896118671,0.3661432266,0.2248979956,0.0245468114,0.1447257847,-0.1714270562,-0.1704276353,-0.1050753519,-0.095365867,0.0798688382,-0.7099152803,-0.0959369466,0.0123920059,0.1499717236,0.0018589956,0.1346984208,0.2285702974,0.2587414682,0.0735446736,-0.0690113679,-0.2563324571,0.1502362788,-0.3472135365,0.1750538647,0.4281513691,-0.1076105163,-0.2047308832,0.2052242756,0.1462742537,0.0059728143,0.12221241,0.1082347482,0.0986734107,-0.1669773459,-0.4362824559,0.0312878899,0.0211695712,0.1986448467,-0.2344741821,-0.0875620097,0.0208606962,-0.4653842449,0.185058862,0.2744430602,-0.2829159796,0.0083542354,0.4928469658,-0.3822821677,0.2702969313,0.0618754998,0.161095351,0.151358366,-0.2949240208,-0.0403018706,0.6218830943,-0.0875844732,-0.2224607021,0.1010837704,0.0422165394,-0.1722308546,0.1404689997,0.1179523915,-0.3721057475,-0.0634600893,-0.2716040313,-0.1298617423,0.1127668694,-0.1156084538,0.2226841897,-0.1070027426,-0.0368037708,0.0902262032,0.4547881484,0.3771903813,0.4752746224,0.1020636037,-0.0787252337,0.2028440237,0.1082331687,-0.1111342609,-0.2018052638,-0.2284781039,-0.1544958055,-0.0177851897,-0.04722986,0.5918661952,-0.0832392946,0.3316909969,0.1294324994,0.4274124503,0.078454718,-0.1012776345,-0.3155736625,0.2165187001,-0.0382006988,-0.1974434406,-0.1862216741,-0.4508140683,0.2089960426,0.0140479067,-0.1022829711,0.2189901471,0.1831719875,0.2945272326,0.315551728,-0.2082607001,0.2113685459,-0.1448933482,0.0957125649,-0.0862429738,0.2207739204,0.4056730568,0.0887244865,0.1711553931,-0.0840777457,-0.1254949123,-0.1189034209,-0.1050963402,0.1902138591,0.178278774,-0.2558560967,-0.0534355938,0.0727953315,0.0938397944,-0.1990550905,0.7170019746,-0.055332575,0.0190006346,-0.3394153714,0.4095416069,-0.3576120138,-0.3678606749,0.0109787444,-0.4763740897,0.1028970554,-0.1900486946,0.4552436769,-0.3954171836,-0.3940078318,0.3679149747,0.4028204381,0.6836955547,-0.1646242291,0.2363882512,0.0510028712,-0.0651171952,-0.5891812444,0.112617895,0.0265494082,-0.1446547955,0.1083635539,-0.0901532918,-0.0310394969,-0.0019047419,-0.2361223549,0.1243400052,-0.0258589908,0.9967814088,-0.276990056,-0.1475044191,-0.1891061664,0.1299015284,-0.1453691721,0.1715866625,-0.0383400396,0.1450566798,-0.1421274394,-0.1709903926,-0.111174196,-0.2888025641,0.0692278892,0.0013702753,-0.113242276,-0.2049562931,0.2201299965,0.4428061843,-0.1418827623,-0.4042373598,0.1063322797,0.0046781143,-0.3059552908,-0.1919153035,0.2700252235,0.0176516231,-0.5485237241,-0.1003788635,-0.1324885488,0.3947390616,-0.2551551163,-0.3398009837,0.2674721479,0.1156701446,-0.2104316354,0.3506485224,-0.4314863682,-0.1012230143,-0.3137769699,-0.1444976777,-0.1705008,-0.6408392191,0.1712003946,0.0828491598,0.4176999927,0.0143066626,0.1683300734,0.1576804817,-0.1679109335,0.3301302493,0.175411582,0.3265022933,0.2780916393,0.3378966451,0.0830716863,-0.1935725063,0.3416213989,-0.0372334421,-0.2525604069,0.4358322024,0.0923087969,-0.2702057958,-0.1565955281,-0.3128049076,-0.292870611,-0.0590361021,-0.1837670654,-0.2335654497,-0.0260720849,0.2174154818,0.44809255,0.0705435649,-0.159980014,0.0138378572,0.0526516922,0.1386069059,-0.0909279212,-0.1020492613,-0.2387267351,0.130207032,0.1935138106,-0.2099672109,0.4386670291,0.0697386339,0.0069293529,0.0480184853,-0.0485860072,0.3132438958,0.1500179768,0.1879092306,0.1593741924,0.2349615544,-0.3872258067,-0.2669921517,-0.0066930121,0.0535586402,0.353803128,-0.2553151846,0.0355976187,0.1523177773,0.1650445461,-0.3105690181,-0.4019656181,-0.4674668014,-0.5130928159,-0.0434752256,0.1322655827,-0.2382610291,-0.0778658688,-0.0501130559,0.0378740989,-0.213486135,-0.055824399,-0.2608462572,0.122231327,0.2898702621,0.1534912586,-0.0731518194,0.3987520039,0.0084083881,0.3870435357,-0.0645662993,0.1885569245,-0.0584623553,0.2860870063,-0.020120224,-0.1384843588,0.0368805416,-0.0247644149,-0.1597116441,0.2168865949,0.0608393066,0.0003910742,-0.2538980544,-0.3209424913,0.1822465062,0.2561471462,0.1904805899,-0.2242921293,0.5012977123,-0.2793630064,-0.3461000919,0.0665202811,0.2444519401,0.0563757233,0.5275489092,0.444391042,0.6823357344,0.0832832754,-0.2015973628,0.0617376566,-0.5795075893,0.2008139789,0.1897365302,0.0327905379,-0.29577595,0.0057109469,0.0328074992,0.1081589684,0.0226492472,-0.0463681594,-0.2439537048,0.2286650687,-0.0432855859,0.2475499064,0.1999292225,-0.4328411818,0.0463740714,0.1934909374,0.0516449325,0.1709925532,0.0794212967,0.1979757249,-0.1561201811,0.0532299057,0.0809718445,-0.0682456419,-0.1365009993,-0.1101310551,0.0294672977,-0.1359082013,0.6114507914,0.0081860544,-0.1737750769,0.3049525917,0.7233242989,0.0312232506,-0.1650012434,0.2243765444,0.0814069733,0.0830178782,-0.0759287551,-0.0355520435,0.4017161429,-0.2073373199,-0.2424644381,-0.1354519874,0.176605925,0.160509333,-0.4235402048,0.1967919469,0.1920951456,-0.0910506546,0.2102546394,-0.3475649655,0.4535573125,-0.2994946837,0.1874474436,-0.0172397494,0.2214671522,0.2587785721,-0.1475878954,-0.0494102761,-0.2666564286,0.0410024114,0.0732267201,0.3569479883,0.2591821551,-0.1972291023,-0.0705196112,-0.1935994178,0.1360177249,0.203421101,0.2293919176,-0.121611692,-0.1258852184,-0.3454179466,-0.096635893,0.4754003882,0.3274008036,0.1240091398,-0.0845170543,-0.4530286789,0.0535154492,0.151571393,-0.2187912017,0.0071665337,-0.0404683314,-0.0320185199,0.1582821757,0.4407413304,-0.3040696084,-0.3075132072,-0.125198558,-0.0749649107,-0.0778145269,-0.0921387747,-0.0547414385,0.084533833,0.1099209338,-0.2102391273,-0.0672074929,-0.1531794667,-0.2270549685,0.0666400269,0.2220398188,0.1254561096,-0.230194971,-0.1300734133,-0.1319652796,-0.0067337593,-0.1409840435,-0.0825301111,0.1432181299,0.5018556714,-0.3013475239,-0.0163627006,-0.0226428639,0.0934189111,0.1913681477,-0.0262033045,0.0151679059,-0.0753220543,-0.0637996271,-0.0593572296,-0.0763601959,0.1175910756,-0.2685797811,0.0336751044,0.1046142951,-0.1771104485,0.2714645267,-0.0215783231,0.0716571137,-0.080995962,-0.0499347448,-0.1160095185,-0.0627261326,0.2081087977,-0.3372123241,-0.1686745584,-0.5594901443,0.070388101,0.0053708437,0.2068969011,0.0117177367,-0.1725110859,-0.1129655093,-0.0344042294,0.0460039005,-0.0764939114,-0.2481325269,0.0310474709,-0.1698401719,0.0715106204,0.3183740079,-0.0484232269,-0.029505739,0.5796294212,-0.0191933718,0.2769597769,0.4304647744,0.1961362809,-0.0370761231,0.1297838986,0.8218683004,0.7138661146,0.0567609295,-0.1451452672,0.161777854,-0.1085897684,0.1470682323,-0.2733016312,-0.4234045446,-0.2159789056,-0.1224903837,-0.2648460567,-0.6692809463,-0.1275978833,0.0218151435,-0.1569614559,-0.1950144321,-0.1166308373,-0.1777846962,-0.0074604754,-0.2130559683,0.0100233713,-0.2707313299,0.2728992105,0.0089492956,-0.0531129055,0.4224569798,-0.0815848187,0.3257421553,0.4410845041,0.1867759228,0.1326977611,0.3957195878,-0.2886687517,0.05456689,0.0071900249,-0.2777746916,0.0549891181,0.2530419827,-0.1008969992,0.575985074,0.2529301941,0.0766595453,-0.1060372964,0.3418752551,0.0910576135,0.7309059501,-0.3629243374,0.0688363016,-0.2589505315,0.0163564775,0.0450370722,-0.0271519627,-0.2573101521,-0.2173153609,-0.1452644914,0.0030241201,0.1407193393,0.1125506237,0.1253523529,-0.0733579472,0.3330389559,-0.0868569613,0.4070806503,-0.2416121811,-0.0280423407,-0.564930737,0.016729746,0.3131297529,0.0697611347,0.1622437984,-0.0284912344,0.0954741091,0.3451727033,0.191942811,0.3001376987,-0.038785208,-0.3129364252,-0.4545789063,-0.1578798443,0.2355280519,0.1526146233,0.0018976778,-0.073671706,0.2914087474,0.0686793178,0.0574670993,0.2785779238,-0.3577061594,0.2636860609,0.0553176068,0.0398983434,0.2203609496,0.0362682529,0.0329764895,-0.0409149863,0.0823313072,0.1189452037,-0.2695453167,-0.1359946281,-0.1256125122,0.4190439582,0.1911336333,-0.0887533426,0.1670397669,0.1044956744,0.1678709388,-0.2785592973,-0.2384754866,0.2490670979,0.2718354762,0.0775364563,-0.0689320564,0.0112844696,-0.0142210647,0.2185996026,-0.2731289268,-0.2237673104,0.4472947717,-0.6256158948,0.2771773934,0.036767561,0.5421524048,0.1418080777,-0.2140448838,-0.1366855204,-0.1059715077,0.3657519519,-0.1967063844,-0.1121848747,0.0648569018,-0.0950532183,0.2117450535,-0.0112746162,0.1887583435,0.3107919991,-0.1589509249,0.3389444649,-0.1313868314]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2503","title":"SubjQA wrong boolean values in entries","comments":"@arnaudstiegler I have just checked that these mismatches are already present in the original dataset: https:\/\/github.com\/megagonlabs\/SubjQA\r\n\r\nWe are going to contact the dataset owners to report this.","body":"## Describe the bug\r\nSubjQA seems to have a boolean that's consistently wrong.\r\n\r\nIt defines:\r\n- question_subj_level: The subjectiviy level of the question (on a 1 to 5 scale with 1 being the most subjective).\r\n- is_ques_subjective: A boolean subjectivity label derived from question_subj_level (i.e., scores below 4 are considered as subjective)\r\n\r\nHowever, `is_ques_subjective` seems to have wrong values in the entire dataset.\r\n\r\nFor instance, in the example in the dataset card, we have:\r\n- \"question_subj_level\": 2\r\n- \"is_ques_subjective\": false\r\n\r\nHowever, according to the description, the question should be subjective since the `question_subj_level` is below 4\r\n","comment_length":27,"text":"SubjQA wrong boolean values in entries\n## Describe the bug\r\nSubjQA seems to have a boolean that's consistently wrong.\r\n\r\nIt defines:\r\n- question_subj_level: The subjectiviy level of the question (on a 1 to 5 scale with 1 being the most subjective).\r\n- is_ques_subjective: A boolean subjectivity label derived from question_subj_level (i.e., scores below 4 are considered as subjective)\r\n\r\nHowever, `is_ques_subjective` seems to have wrong values in the entire dataset.\r\n\r\nFor instance, in the example in the dataset card, we have:\r\n- \"question_subj_level\": 2\r\n- \"is_ques_subjective\": false\r\n\r\nHowever, according to the description, the question should be subjective since the `question_subj_level` is below 4\r\n\n@arnaudstiegler I have just checked that these mismatches are already present in the original dataset: https:\/\/github.com\/megagonlabs\/SubjQA\r\n\r\nWe are going to contact the dataset owners to report this.","embeddings":[0.1235658228,0.0693046823,0.0270314571,0.1645201296,-0.3365960717,-0.0068990644,0.0865071937,0.045239836,-0.0927979127,0.1864392012,-0.0919599161,0.4202955663,0.2149700969,0.254910022,-0.218426764,0.1229733899,0.2286490202,0.1842408031,-0.1363601834,-0.1547209173,-0.4317789078,0.260576278,-0.5335964561,-0.0119052576,-0.1275137514,-0.2023657113,-0.1333579421,-0.1422060132,-0.2238854617,-0.2932470143,0.0730003268,-0.0335393548,-0.0811768249,0.1234094575,-0.0001093275,-0.2034583986,0.0526367277,0.0162325818,-0.2176798731,0.0769172758,-0.2617804706,-0.0417561233,-0.1978150308,0.0486691743,-0.3357685804,-0.4470198154,-0.2089022696,-0.268366456,0.3899082541,0.0091003245,0.1720302105,0.0610974617,-0.1662039608,0.2474562973,0.3539647162,-0.1808389723,-0.0492369607,0.2432310581,-0.0437768176,-0.046365045,0.1507534981,0.3816999495,0.0878882408,0.0792881027,0.1195269004,-0.1314198822,-0.0643833727,-0.2449347526,0.2853113115,0.2433769554,0.340031594,-0.1474963725,-0.4372366667,-0.0865620747,-0.0688515455,-0.1194220558,0.2729384899,0.0267802067,0.1324341595,-0.0465580709,0.1046249047,0.3983652294,0.1085541248,-0.0410750099,-0.1135641858,0.1411972493,-0.1243974566,-0.0282294825,-0.1896745563,-0.1182829961,0.0167922322,0.0029046785,-0.6352514625,0.1034262553,-0.4720294774,-0.0552292466,-0.0651546121,0.1520368755,-0.15406923,-0.3375006616,0.2958487868,-0.0503651798,0.2031734139,0.1381501257,0.5345699787,0.2554335892,0.3524950147,-0.2006354481,-0.0916452259,-0.2544087172,0.0428180024,0.0638490841,0.1985457242,-0.2788588107,-0.0028028965,0.0887786821,0.1084235609,-0.3041123748,-0.5812082291,0.1572570056,-0.4001148641,-0.045035664,0.134021163,0.0927854776,0.047653798,-0.2227297574,-0.0990667716,0.0209519267,-0.134592101,-0.4782003462,-0.1239519566,-0.0238467362,-0.1999012083,0.054493811,0.0194708202,-0.0180341434,-0.0020733154,0.3719302118,-0.1979906112,-0.2239182591,0.0091523537,-0.2884495854,0.2214837223,0.2032552809,-0.2496664226,0.2008807957,0.0493718758,-0.0521150827,-0.0147676887,0.3387775421,0.089169614,0.1819593757,-0.0369847305,0.1804388463,-0.0891753063,-0.0578743257,0.1410854757,0.3233660161,0.2153365165,0.0382418036,0.0631323233,-0.2067233622,-0.1559418738,-0.0916230232,-0.0364298038,0.1502397656,-0.6753096581,-0.1031776294,-0.0081572644,0.2379122674,0.0118778152,0.1806159019,0.191116184,0.2914659977,0.0767869577,-0.0861867368,-0.2743889093,0.1927633435,-0.4252072573,0.0999394506,0.422033906,-0.0812632442,-0.2345722467,0.2056286782,0.1705133617,0.0348280966,0.1862520278,0.0711834431,0.0980645046,-0.1092921942,-0.4336036146,-0.0682207122,-0.0198702402,0.2016846389,-0.1839958429,-0.070262529,0.0519549549,-0.4857279956,0.1369781941,0.1761033088,-0.2915922403,0.0750327855,0.457909286,-0.3810139298,0.2863347232,0.0377027057,0.1168815345,0.1942813396,-0.4058647156,0.0025864381,0.5612463355,-0.0935289487,-0.1990175396,0.1016951129,0.0479906648,-0.1811901033,0.1231813356,0.1443934739,-0.3442240059,-0.0312309042,-0.2829026282,-0.0618521273,0.1143131778,-0.0696996599,0.1398046911,-0.1327401847,-0.0771326274,0.0567477867,0.4595968127,0.3403603435,0.447440505,0.0692550167,-0.0982890725,0.1780465841,0.0816662908,-0.1003291011,-0.1397878826,-0.2158813775,-0.0932421163,-0.100940682,0.048600655,0.5878672004,-0.0976712927,0.345982492,0.1544420719,0.463313669,0.0811572522,-0.1465827227,-0.3019917905,0.2043401748,-0.0304879397,-0.1946274787,-0.172913,-0.4056322277,0.2356466651,0.0524048954,-0.089546673,0.247719571,0.1411375999,0.2954147458,0.3167260289,-0.1854985505,0.2649505734,-0.2104398161,-0.0202971678,-0.0652850494,0.2469478399,0.4416828156,0.1151769087,0.1954725832,-0.0981773287,-0.0625020713,-0.1693164855,-0.0454245768,0.2141800672,0.1679679155,-0.2550610304,-0.0083731012,0.0559946373,0.1583368033,-0.21402888,0.7436457276,-0.0561722852,0.0429213084,-0.3219944835,0.3840662241,-0.3165824413,-0.3022707999,0.0065395301,-0.526925981,0.0282171946,-0.2270455062,0.4689020813,-0.3912793398,-0.4067310393,0.3903204501,0.4209657609,0.6782056689,-0.1489887238,0.1956743896,-0.0283058733,-0.0563011095,-0.5697382689,0.0757160559,0.0366113819,-0.2367685884,0.0595968105,-0.1379311979,-0.1368680298,-0.0072096889,-0.2768760026,0.1192589402,0.0006518387,1.0724867582,-0.2695175707,-0.1467303783,-0.1190344021,0.1193541065,-0.1574742347,0.1590285003,-0.0539910309,0.0906694978,-0.1739587635,-0.1675078869,-0.1246683374,-0.35159567,0.0382580049,0.004620159,-0.1437669843,-0.2055474371,0.2674574852,0.4465821087,-0.1614125818,-0.3993704021,0.0713938922,-0.0173409935,-0.3348317742,-0.2631746233,0.3340443969,0.0522255749,-0.5183662176,-0.0925903469,-0.1297246367,0.4031534195,-0.2657470703,-0.3338995874,0.3003411591,0.0676516593,-0.2214169204,0.3181125522,-0.3968447447,-0.1473349184,-0.2791083753,-0.1210676059,-0.1548233926,-0.6240259409,0.1361133307,0.1735401303,0.4568222761,-0.0487644188,0.1495238245,0.1556363106,-0.0425291322,0.3412079811,0.1651685983,0.3119772077,0.3186435401,0.3356915116,0.083191663,-0.2547818124,0.3756683171,-0.0273993947,-0.2127391845,0.3895865381,0.0599051118,-0.2288595736,-0.1277025938,-0.1800574213,-0.3049177229,-0.0435703844,-0.2200166136,-0.1913350374,-0.0832531899,0.2246245444,0.3975691795,0.0006997898,-0.1159170121,0.0042764763,0.0272913203,0.112770915,0.0074557448,-0.1231147125,-0.2056857646,0.1232669502,0.2521838546,-0.2121995986,0.4180151522,0.0451586321,-0.0097377701,0.0432471149,-0.0618417822,0.3363803327,0.1082416177,0.18739447,0.1663555205,0.292517066,-0.3721614182,-0.2799001038,0.0299684405,0.0526573732,0.3983995616,-0.2495489419,0.0345630199,0.1566703171,0.1528951079,-0.2442190349,-0.3942863047,-0.39825809,-0.525051415,-0.0674085319,0.0731197149,-0.2132783681,-0.0859828442,-0.0457683429,0.036919754,-0.2266895324,-0.0952135772,-0.205331862,0.1177323386,0.3615660667,0.1725515127,-0.132208854,0.3831833005,0.0572224259,0.4139149785,0.0133756362,0.2656077147,-0.0019590529,0.1644227356,0.0157556757,-0.1311364323,-0.0216493271,0.0119559774,-0.1118065715,0.1419661194,0.0780653805,-0.0004626242,-0.2904466689,-0.2381997705,0.1909382045,0.2920924425,0.138316229,-0.2242974788,0.5193617344,-0.2695476115,-0.3641129136,0.1406291723,0.1554797441,0.0487962216,0.4307951629,0.5219877362,0.7752364278,0.0249507483,-0.1994744241,0.0232310928,-0.5342606306,0.227693215,0.1881096661,0.0658886507,-0.3278450668,-0.0222094543,0.0415822119,0.0512277931,0.0864018649,-0.0782105103,-0.2384384125,0.2436134964,-0.0614316538,0.2418659776,0.1969997734,-0.4235354066,0.0483170561,0.2041701376,0.0795643479,0.166733548,0.0505106412,0.2443148643,-0.1379183084,0.0434099026,0.0136464369,-0.0338525586,-0.2221969217,-0.0691595078,-0.0649968684,-0.1792666465,0.5730493069,0.044191841,-0.1402351558,0.3558791578,0.7642988563,0.033544261,-0.130191192,0.2423322201,0.0350238234,0.1353755146,-0.0243632887,-0.0333927833,0.3863116205,-0.163699314,-0.2512901425,-0.0570655726,0.2584724128,0.1336396784,-0.5338660479,0.1852084845,0.1505409181,-0.0688359886,0.2695310414,-0.3762758076,0.4878803492,-0.3313367963,0.1649713218,-0.0474717803,0.153219521,0.2749245465,-0.2633228004,-0.1157359853,-0.2474367023,0.0408621244,0.1218427122,0.4315551817,0.2313307673,-0.1939185113,0.0098416982,-0.1758002192,0.1753427982,0.1273627281,0.1977410167,-0.10554225,-0.1823627204,-0.3872100711,-0.1193847656,0.4147989154,0.3592251539,0.1023105532,-0.0635276586,-0.4572013617,-0.0118480548,0.170110032,-0.2586512268,0.0249619875,0.0414023362,-0.0099096922,0.0454233103,0.4727853239,-0.2946563065,-0.2848927081,-0.143734619,-0.1145408228,-0.079201892,-0.1556377262,-0.0609723404,0.0131851314,0.0809936896,-0.2546264827,-0.0911783576,-0.1349463463,-0.237850666,0.0759965554,0.185825333,0.1357615292,-0.241028592,-0.0968572572,-0.0528470464,-0.0384106077,-0.0851724818,-0.1401199102,0.1708898544,0.5099928975,-0.2567969859,-0.0665274784,-0.068767786,0.1167855784,0.1696681231,-0.0922320411,0.0603907816,-0.0241418891,-0.0374816284,-0.0915850475,-0.0986884832,0.1159597859,-0.3026841581,0.0257621966,0.1222919002,-0.1805089414,0.2352313846,0.0132279787,0.0667850748,-0.0411056876,-0.0670949295,-0.1620511115,-0.0609442778,0.1795124263,-0.3609016538,-0.285972178,-0.5163059831,0.079366155,-0.0257652737,0.1678267866,0.09591607,-0.1292792112,-0.0863103271,-0.0490188487,0.1414652616,-0.0860994905,-0.0971278176,0.1656467915,-0.1560973078,0.0525442958,0.3499757051,0.0512841865,0.0088000903,0.6680774093,-0.0368153676,0.285869956,0.3955477178,0.1617474407,-0.0236239973,0.0540882573,0.7669574022,0.7463683486,0.0661245957,-0.07790301,0.1320261508,-0.1819618195,0.0664941594,-0.329118669,-0.3671015799,-0.2042683959,-0.1462715119,-0.2567375004,-0.7162510753,-0.1572035253,0.0245053023,-0.1584972143,-0.2033134103,-0.1467431039,-0.1335705817,0.0106220245,-0.2728757858,-0.0699335635,-0.338830024,0.3155473769,-0.0242836028,-0.0999641344,0.4305458367,-0.0234449208,0.2384318262,0.4121883214,0.2002364844,0.1885689944,0.3959139884,-0.3454794884,0.0806802213,-0.0005042401,-0.2542779446,0.0388810784,0.1794080734,-0.1321676672,0.4886135161,0.2490169108,0.0578088313,-0.0644316748,0.246882081,0.0444876626,0.735316813,-0.3102240264,0.0769420192,-0.2547957599,0.0198194552,0.1025856063,0.0089667533,-0.2493952662,-0.2108242512,-0.1092764288,0.0135479979,0.1533758938,0.141567409,0.111755684,-0.1167592853,0.3674061298,0.0092888288,0.3763061762,-0.2092628628,0.0337285623,-0.6373873353,0.0088730799,0.3327725828,0.0958693102,0.2232337892,-0.0396269336,0.1314703077,0.3173561692,0.2348477989,0.4174261391,-0.0000524035,-0.3034644723,-0.5666862726,-0.1190749258,0.244263947,0.1538658589,-0.0076226266,-0.065914236,0.3218994439,0.0929826051,0.0182211492,0.3179986179,-0.3438113332,0.2634426057,0.075060308,0.1002883241,0.2569550872,0.0346494019,0.0482903235,-0.0914406925,0.0513267517,0.0839146748,-0.3272604644,-0.1908607185,-0.1339072734,0.4577326775,0.1618555933,-0.1028999984,0.078097038,0.1582153738,0.1992510557,-0.3002577126,-0.307849139,0.2193503827,0.2622329295,0.0981667936,-0.0249890722,0.0775294006,0.0179367065,0.3035348952,-0.2655321956,-0.142457664,0.4916682541,-0.5384700894,0.2442444563,-0.0164526254,0.4995483756,0.1762417555,-0.1370936334,-0.1033865362,-0.1187577322,0.368295908,-0.1611113995,-0.1232278049,0.0827054828,-0.0310241301,0.1907952428,-0.0175492428,0.1547932029,0.3163712323,-0.163660571,0.3270504475,-0.1499864608]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2503","title":"SubjQA wrong boolean values in entries","comments":"I have:\r\n- opened an issue in their repo: https:\/\/github.com\/megagonlabs\/SubjQA\/issues\/3\r\n- written an email to all the paper authors","body":"## Describe the bug\r\nSubjQA seems to have a boolean that's consistently wrong.\r\n\r\nIt defines:\r\n- question_subj_level: The subjectiviy level of the question (on a 1 to 5 scale with 1 being the most subjective).\r\n- is_ques_subjective: A boolean subjectivity label derived from question_subj_level (i.e., scores below 4 are considered as subjective)\r\n\r\nHowever, `is_ques_subjective` seems to have wrong values in the entire dataset.\r\n\r\nFor instance, in the example in the dataset card, we have:\r\n- \"question_subj_level\": 2\r\n- \"is_ques_subjective\": false\r\n\r\nHowever, according to the description, the question should be subjective since the `question_subj_level` is below 4\r\n","comment_length":19,"text":"SubjQA wrong boolean values in entries\n## Describe the bug\r\nSubjQA seems to have a boolean that's consistently wrong.\r\n\r\nIt defines:\r\n- question_subj_level: The subjectiviy level of the question (on a 1 to 5 scale with 1 being the most subjective).\r\n- is_ques_subjective: A boolean subjectivity label derived from question_subj_level (i.e., scores below 4 are considered as subjective)\r\n\r\nHowever, `is_ques_subjective` seems to have wrong values in the entire dataset.\r\n\r\nFor instance, in the example in the dataset card, we have:\r\n- \"question_subj_level\": 2\r\n- \"is_ques_subjective\": false\r\n\r\nHowever, according to the description, the question should be subjective since the `question_subj_level` is below 4\r\n\nI have:\r\n- opened an issue in their repo: https:\/\/github.com\/megagonlabs\/SubjQA\/issues\/3\r\n- written an email to all the paper authors","embeddings":[0.1710738689,0.0587353781,0.0442122258,0.0991450697,-0.3185442686,-0.1043600217,0.0993457064,0.0244082045,-0.1466466784,0.2234863788,-0.0614027008,0.3840125203,0.2275162339,0.2279736251,-0.2728487849,0.1101417691,0.2479696721,0.166661188,-0.0700219274,-0.1800619364,-0.4520644546,0.2599630952,-0.6040433645,-0.0405316427,-0.1244438365,-0.2389922142,-0.1235726103,-0.1611758471,-0.3197330534,-0.2684791386,0.0228940286,-0.0024612597,-0.0706962943,0.1173533425,-0.0001112474,-0.1952677071,0.0751518682,0.0176677648,-0.2246204019,0.085610643,-0.1955979615,-0.0928517729,-0.2246859223,0.0400888659,-0.3261657059,-0.4419411123,-0.2020022124,-0.2190367281,0.3537075222,0.0255618338,0.1417027116,0.0158051699,-0.1125524268,0.2566840649,0.3405130208,-0.2426410466,-0.1025819257,0.2494196445,-0.0194552206,-0.0457550175,0.1129644662,0.3713188767,0.1074483916,0.0999766067,0.0808564126,-0.0649048761,-0.0014376957,-0.2529067099,0.2693755925,0.2289806306,0.3289312124,-0.1468628198,-0.4584599137,-0.0965828001,-0.0855324939,-0.0976805016,0.2790672183,0.0382844768,0.1127962172,-0.0264036488,0.1741399169,0.4021877646,0.0861006007,-0.0118232798,-0.11401885,0.117589727,-0.179119885,-0.0428106003,-0.1124674007,-0.1337602437,-0.0495318323,0.0108837923,-0.6033577323,0.1203535423,-0.4970486462,-0.0238896552,0.0368807875,0.114126496,-0.1481908709,-0.3702002168,0.3033159971,-0.053209506,0.2016269565,0.2120987773,0.5167773962,0.2139938921,0.37682423,-0.2153893858,-0.0588614717,-0.2195380032,0.0617492199,0.0186376739,0.1607973725,-0.2832615674,0.036087621,0.04941817,0.1375556737,-0.2750588357,-0.5347527266,0.1172761023,-0.4085183144,-0.0356437787,0.1520291716,0.0824728534,0.0522991531,-0.2182780951,-0.1242639795,-0.0142753702,-0.1809726954,-0.4663915634,-0.1026122496,-0.0126753598,-0.2421897352,0.0704667196,0.0095849838,-0.0051735188,-0.0326608047,0.3610464036,-0.1792062074,-0.2311148047,-0.045544751,-0.3192038238,0.2102465034,0.1566302478,-0.257568866,0.1769367307,0.0184607133,-0.0885849893,0.0113783451,0.3068144321,0.1314315349,0.2160139829,-0.0264996942,0.1529012471,-0.0488313958,-0.0194252003,0.1305797249,0.3427820802,0.2295735478,0.0766574964,0.097899422,-0.1246005967,-0.108446613,-0.120166406,-0.0074979821,0.0930765569,-0.5938064456,-0.0956555009,0.0028287093,0.2563048899,0.0060078651,0.1614622623,0.2352576256,0.3217531443,0.1038671359,-0.0696827844,-0.2798698545,0.1783294678,-0.4563191533,0.1051230207,0.4366212189,-0.1220444813,-0.221826598,0.2033574879,0.1472888291,0.0721242428,0.1484610587,0.060657151,0.091976963,-0.0978632793,-0.4295600951,-0.0188788548,-0.0461560562,0.1946666688,-0.1881421953,-0.0667494759,0.0751229078,-0.4497666657,0.1388593316,0.2089561969,-0.2969644368,0.0651323795,0.4925538003,-0.4521814287,0.2907610536,0.0284890067,0.1424675137,0.1890472323,-0.4601029158,0.0187790208,0.5386309624,-0.0874888152,-0.2166935951,0.0964654461,0.0762858465,-0.1828972846,0.1264407784,0.0942379162,-0.3206942976,0.0440112501,-0.2463474274,-0.0723955408,0.11738538,-0.053882312,0.1654940397,-0.1231289431,-0.0828999206,0.0102148503,0.5004791021,0.3310031891,0.4476000965,0.082400769,-0.0840768516,0.1700825393,0.0918709338,-0.0996949896,-0.1469714195,-0.2126724571,-0.1487464905,-0.045117598,0.0176735595,0.5685980916,-0.096134901,0.355956316,0.1158235297,0.4205283821,0.0681669116,-0.1510509849,-0.2987359464,0.1734165847,-0.0158575624,-0.2166254073,-0.2364974767,-0.3514859378,0.1823423803,0.0784165338,-0.1122974604,0.2293712199,0.214940846,0.2402786762,0.3250105083,-0.1460169107,0.307464391,-0.1975390613,0.0135957124,-0.0615344234,0.2345063239,0.3891462684,0.0415773876,0.1550029069,-0.1160536259,-0.0861490667,-0.1397906095,-0.031043807,0.1661272645,0.2036787719,-0.2768189907,-0.0534087159,0.0403561257,0.1367254257,-0.1804144233,0.7738114595,-0.0661959127,0.0629098415,-0.2841646075,0.4080078304,-0.3262366354,-0.3555271327,0.065151833,-0.5793977976,0.0429905877,-0.1876525283,0.4197408855,-0.3421962857,-0.371132195,0.3845454156,0.432146281,0.7168484926,-0.0835766047,0.197931245,0.0355790965,-0.0012880722,-0.5488743782,0.0704206377,-0.0014025839,-0.2150136977,0.0272914078,-0.1356779337,-0.0908275098,0.0397351198,-0.2662622929,0.1043313891,0.0013372216,1.0558440685,-0.2724154592,-0.1538788676,-0.101854004,0.1566333324,-0.0917654857,0.1692700982,-0.008325927,0.1373540461,-0.1299726665,-0.1773164272,-0.1217777207,-0.3256254196,0.1215997115,-0.0073376219,-0.1755313873,-0.2254958302,0.2760755122,0.4957951605,-0.099076882,-0.3769023418,0.0817377269,-0.0282826014,-0.2469940633,-0.2265727818,0.3488762975,0.0640418157,-0.501604259,-0.0953377411,-0.1688266695,0.4532333314,-0.2941656709,-0.2607148588,0.3030191064,0.1040576398,-0.2190982103,0.3839105368,-0.3634410799,-0.1291661412,-0.3180090487,-0.1227653474,-0.1539869308,-0.6412751079,0.1094340682,0.1728317142,0.494954139,0.0312018413,0.1230886132,0.215901345,-0.1157743186,0.4156973958,0.1808377504,0.296356976,0.3292664886,0.3150509894,0.1459970623,-0.3184501231,0.4076510072,-0.013140765,-0.2225530893,0.3915548027,0.1070998535,-0.2039831877,-0.1251978725,-0.2320007235,-0.293009311,-0.0953899398,-0.2225006819,-0.1737557501,-0.0804586932,0.1478117108,0.4191001952,0.0185867194,-0.0496437289,0.0072559142,0.0653144643,0.1636179984,-0.0373219587,-0.0649501607,-0.1452895552,0.0865935534,0.2224617451,-0.1967711449,0.4460463226,0.0981758535,0.0612534396,0.0429764576,-0.0720524788,0.2515118122,0.0819990709,0.190233618,0.1813407689,0.2219103277,-0.3612627685,-0.2554587126,0.0454083197,0.0373501256,0.4311312437,-0.3206877112,-0.0464969352,0.1578715593,0.1075587496,-0.3169372976,-0.3542120159,-0.4349246323,-0.5650666952,-0.0700684935,0.0910840854,-0.25302881,-0.0687799826,-0.0058950582,0.044294443,-0.2659388185,-0.0459987,-0.2791157365,0.0803436488,0.3241610229,0.1375290453,-0.1815164387,0.3898333609,0.083271943,0.3620143235,-0.0191580653,0.2222392261,-0.004175181,0.1461882591,0.0003617232,-0.1001675501,-0.0257404614,-0.0471040383,-0.145759657,0.1756811142,0.1380997598,0.0051616007,-0.2981782854,-0.2280026227,0.1883990914,0.2918301821,0.1923448592,-0.1986018866,0.5775856972,-0.2535889149,-0.3238762617,0.1643600166,0.1683275998,0.0528544672,0.3784892857,0.4539310634,0.6912428141,0.0340017043,-0.2650211751,0.0035184082,-0.4962542951,0.1727015227,0.1954564005,0.1080077216,-0.3301402926,0.0022254344,0.0319641829,0.0554594845,0.1138453484,-0.1201774403,-0.249357298,0.2281406075,-0.0764818341,0.2382534891,0.2021413147,-0.4176892638,0.0296194311,0.1498800814,0.0613117777,0.1391269714,0.079340592,0.2710895836,-0.137752071,0.0098856976,0.0082470691,-0.012412414,-0.2050175518,-0.1198811829,-0.0636533275,-0.2108406425,0.6007065177,0.0496907532,-0.162156269,0.344355166,0.7612618208,0.0414605513,-0.1690537483,0.1809292436,-0.0063544801,0.0539402775,-0.0523575656,-0.0242304262,0.3899326921,-0.1600636244,-0.2672144473,-0.0354235023,0.2512818873,0.120862104,-0.4943597317,0.1666124463,0.1627020538,-0.0504370965,0.3389598131,-0.3811135292,0.441352129,-0.3024438024,0.1571769416,0.0596395545,0.186114043,0.2916510701,-0.2486475408,-0.0310851224,-0.2250554115,0.0086328117,0.0705894455,0.447460711,0.1469641626,-0.1963317245,0.0011979877,-0.1352941692,0.1603173763,0.102155067,0.2937581241,-0.0895491913,-0.1734993458,-0.3739548028,-0.113450475,0.4546183646,0.3324969709,0.0349066332,-0.0810138211,-0.4279148579,0.0097935023,0.1897893399,-0.3191972375,-0.0186345074,0.0535007082,-0.0060306517,0.1207752228,0.4614461064,-0.2734963596,-0.3302182555,-0.184092626,-0.1364317685,-0.0317588858,-0.1606333554,-0.0649157017,0.0195151474,0.0582752302,-0.236616835,-0.0580759943,-0.1101901084,-0.2381670326,0.0738930926,0.2222915441,0.1310154051,-0.208708778,-0.1066550389,-0.0951599032,-0.0650417432,-0.1191650704,-0.1671589017,0.1784075201,0.5427194834,-0.3180454075,-0.0727732927,-0.0946562588,0.1349090636,0.1783765405,-0.1064403132,0.04572374,-0.0195693597,-0.0604567491,-0.0733916014,-0.142905727,0.101495482,-0.3249171376,-0.0324073248,0.1791353822,-0.179275468,0.2557197213,-0.0481328256,0.0687513724,-0.1087821946,-0.0806981921,-0.1058293805,-0.0837047771,0.1577275246,-0.2939201593,-0.2751470804,-0.5780192614,0.0810625926,0.0259179287,0.2421536446,0.0658203289,-0.1607862562,-0.1154477149,-0.0842853561,0.1261678338,-0.074329108,-0.1699790061,0.1035673916,-0.1424228251,0.0622132458,0.4021680951,0.0162798092,0.0114184516,0.6619752049,0.0064644124,0.2706698477,0.3829728961,0.199198395,-0.0744822621,0.1442096978,0.7489561439,0.7794600725,0.0877803192,-0.154695183,0.0922623426,-0.2461014837,0.0689404458,-0.2604273558,-0.3266122341,-0.1921123564,-0.1099601835,-0.235773623,-0.7090877295,-0.1479443908,0.0614780858,-0.2298949808,-0.1654640436,-0.1609176397,-0.130974412,-0.0166164525,-0.2493011504,-0.034295883,-0.3071187735,0.3857826889,0.0257287025,-0.0913101882,0.4063305259,-0.078096278,0.2709381878,0.3368436098,0.2180171907,0.2085646242,0.4107733071,-0.3620149493,0.0342453197,0.000755243,-0.2595655024,0.0585566424,0.2109633833,-0.1715664864,0.5173904896,0.1891725212,0.0402592905,-0.0571143739,0.2397412211,0.057312306,0.7155686021,-0.3011633754,0.1196403131,-0.2480365336,0.0413655229,0.084101826,0.0089145582,-0.2025250793,-0.1839364767,-0.1601080298,-0.025551267,0.139183253,0.0968239382,0.0995877385,-0.0579678267,0.3223220408,-0.043169599,0.4640344381,-0.2075226456,0.0518894233,-0.5853188038,-0.0900231972,0.3276987672,0.0366398953,0.2794473171,0.0311095603,0.1066708639,0.4072140157,0.2393845618,0.4312269688,-0.0055292952,-0.3166004419,-0.5569550395,-0.1463944763,0.2237804681,0.1129154563,-0.0277986005,-0.100446485,0.3361706138,0.0746722296,0.0399963744,0.3453142643,-0.2967600226,0.2541286349,0.0569518954,0.1100652441,0.2701626718,-0.0224243179,0.0659559071,-0.1521827132,0.0002197733,0.1862972826,-0.3042448163,-0.2631472051,-0.1038389057,0.4582390785,0.1468835771,-0.134257108,0.0556059964,0.1960705966,0.2331261486,-0.3619260192,-0.2265972048,0.2175063789,0.3137883842,0.0897706896,-0.0526370443,0.0236258525,-0.0488439724,0.2905326188,-0.3144273162,-0.1801204532,0.4388245046,-0.52937603,0.2906626463,-0.0097948536,0.5312720537,0.1658749133,-0.1308614761,-0.1109351814,-0.0946589112,0.3302759528,-0.1105706692,-0.1525578797,0.0244374331,0.0119888857,0.1898100972,0.0036702091,0.1343384534,0.2634167671,-0.0979352966,0.3415637612,-0.1324795634]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2503","title":"SubjQA wrong boolean values in entries","comments":"Please [see my response](https:\/\/github.com\/megagonlabs\/SubjQA\/issues\/3#issuecomment-905160010). There will be a fix in a couple of days.","body":"## Describe the bug\r\nSubjQA seems to have a boolean that's consistently wrong.\r\n\r\nIt defines:\r\n- question_subj_level: The subjectiviy level of the question (on a 1 to 5 scale with 1 being the most subjective).\r\n- is_ques_subjective: A boolean subjectivity label derived from question_subj_level (i.e., scores below 4 are considered as subjective)\r\n\r\nHowever, `is_ques_subjective` seems to have wrong values in the entire dataset.\r\n\r\nFor instance, in the example in the dataset card, we have:\r\n- \"question_subj_level\": 2\r\n- \"is_ques_subjective\": false\r\n\r\nHowever, according to the description, the question should be subjective since the `question_subj_level` is below 4\r\n","comment_length":14,"text":"SubjQA wrong boolean values in entries\n## Describe the bug\r\nSubjQA seems to have a boolean that's consistently wrong.\r\n\r\nIt defines:\r\n- question_subj_level: The subjectiviy level of the question (on a 1 to 5 scale with 1 being the most subjective).\r\n- is_ques_subjective: A boolean subjectivity label derived from question_subj_level (i.e., scores below 4 are considered as subjective)\r\n\r\nHowever, `is_ques_subjective` seems to have wrong values in the entire dataset.\r\n\r\nFor instance, in the example in the dataset card, we have:\r\n- \"question_subj_level\": 2\r\n- \"is_ques_subjective\": false\r\n\r\nHowever, according to the description, the question should be subjective since the `question_subj_level` is below 4\r\n\nPlease [see my response](https:\/\/github.com\/megagonlabs\/SubjQA\/issues\/3#issuecomment-905160010). There will be a fix in a couple of days.","embeddings":[0.1078841239,0.0156454202,0.0252958909,0.1298000067,-0.2156247199,-0.1369084865,0.0789557919,0.0792114735,-0.121195659,0.2209163457,-0.1261701137,0.4446637332,0.2015048265,0.2101600617,-0.2392442673,0.0243513733,0.2125155777,0.1294924468,-0.156169951,-0.1451063454,-0.4437947273,0.2678593397,-0.5591130853,-0.0343240276,-0.171608299,-0.1762482971,-0.1232099831,-0.1076801792,-0.3289825916,-0.3217784166,0.0076412116,0.0098621724,-0.0445199236,0.1426259875,-0.0001050527,-0.1938910782,0.1047164723,0.0202372428,-0.2186999172,0.02379179,-0.2130345106,-0.0177318584,-0.1838343292,0.1085871011,-0.3510945141,-0.4068053067,-0.1953574419,-0.2121617198,0.3728216887,0.030214807,0.2069548965,0.0731095895,-0.1665284634,0.1264034957,0.3284571767,-0.2110949457,-0.145341903,0.2226789743,-0.0061084926,-0.0522830412,0.0797757208,0.3633543551,0.1334154904,0.1183657125,0.1207933128,-0.0924511552,-0.037695203,-0.2533324361,0.2884595394,0.2718881071,0.3816457391,-0.102404058,-0.4347176254,-0.0564079657,-0.079952538,-0.1436015964,0.2506157458,0.0122629507,0.1037202775,-0.079169549,0.124134101,0.4422753453,0.0309008099,-0.0398909301,-0.1293095499,0.1092372909,-0.2011896819,-0.0101448018,-0.0890257582,-0.1437237412,-0.1051873192,0.0382035039,-0.6164880991,0.0896843746,-0.5005204082,-0.0581618957,0.031914819,0.1636416912,-0.1211132556,-0.3252467513,0.2689596713,0.0045965416,0.1843355894,0.1559428126,0.5638609529,0.2239138633,0.3876720667,-0.240400821,-0.0435531884,-0.2489138097,-0.0043956772,0.0323095471,0.205439657,-0.2637923956,0.0247810613,0.0653884113,0.1067539826,-0.3335361481,-0.5306913853,0.0447513945,-0.3258023262,-0.0365552604,0.162762776,0.1874296814,0.049243331,-0.1780377626,-0.1389155388,0.0290630832,-0.1394177675,-0.4701973796,-0.1302654743,-0.0279068369,-0.2514798641,0.0480002612,0.0805167183,0.0371726044,-0.0147057343,0.316675961,-0.146690309,-0.2116389722,0.0303529482,-0.337413758,0.2290263921,0.1952270418,-0.2081382424,0.1602177918,0.0529025197,-0.0269854777,0.0129567087,0.3299602866,0.1502215117,0.2081973255,-0.0221722908,0.20190911,-0.0030561343,-0.0333307274,0.1041437909,0.3153418601,0.185134396,0.1232642531,0.072126843,-0.1433778405,-0.1496178359,-0.1003512293,-0.0502962433,0.0847514719,-0.5601146817,-0.0764279366,-0.0352466516,0.1555561423,-0.0244265106,0.1905686706,0.1863573641,0.272380203,0.0855212882,-0.0332408883,-0.2608580589,0.1775908023,-0.3822654188,0.1188811064,0.3639727533,-0.1182511374,-0.244916603,0.205911234,0.1168062687,0.0990830958,0.1585886478,0.0778981894,0.0759563148,-0.0894242153,-0.4623380899,-0.0423552915,-0.0175817702,0.2353250831,-0.182051897,-0.068527475,0.0390432663,-0.4683903456,0.1705978215,0.1901507229,-0.2734436691,0.0676822811,0.5333811045,-0.4080266654,0.2827703655,0.0107922535,0.1316680312,0.1808864772,-0.4180734158,0.0776943937,0.5860155821,-0.0953002647,-0.2689681649,0.1065695584,0.0783531293,-0.1784012914,0.1818079501,0.0968714803,-0.314580977,-0.0352716781,-0.2252127677,-0.0026632939,0.1741293818,-0.0801820457,0.1730214953,-0.1529103369,-0.10740491,-0.0335847288,0.484310329,0.2997553051,0.4983261824,0.0867985189,-0.0893943682,0.2151368707,0.1279271394,-0.0823510885,-0.1044034734,-0.2382663339,-0.1810604483,-0.0628499612,0.0107257832,0.5453843474,-0.0689737499,0.3212510645,0.0795222744,0.4526009858,0.0827824771,-0.1582430899,-0.2327337414,0.2047129273,-0.0155974245,-0.1867107153,-0.230164811,-0.3718472719,0.1537745446,0.0469394997,-0.1074849814,0.2120721042,0.1410797387,0.2905456424,0.3607884347,-0.1225848421,0.2849938869,-0.1921898127,-0.0257339589,-0.0809656829,0.2435415089,0.326233983,0.084455885,0.2091195583,-0.0837976485,-0.0728600994,-0.1378699243,-0.0671971962,0.186629191,0.1368008107,-0.2311040163,-0.0982552096,0.0744894668,0.0967023,-0.2640463412,0.7692214847,-0.0595285408,0.0239598602,-0.3110316992,0.3185155392,-0.2833787501,-0.2752931118,0.0854061395,-0.5461099148,0.0195280295,-0.2208147794,0.4698496759,-0.3125492632,-0.3880244792,0.378064245,0.3936958015,0.6616831422,-0.1058019251,0.2285186648,0.0402448662,-0.0495512635,-0.5159404278,0.1395889223,0.0083683245,-0.204518795,0.0781889111,-0.0829382986,-0.0678761601,0.0354415551,-0.2837559581,0.1077972874,-0.0687569305,1.0491286516,-0.2394216061,-0.1365912408,-0.1309762448,0.131928876,-0.0518807136,0.0938581824,-0.0464565754,0.1039718166,-0.1310689449,-0.1739444882,-0.1715656668,-0.3940061927,0.0960453078,-0.0466709062,-0.0972711295,-0.2111295015,0.2599931061,0.5046925545,-0.1170601994,-0.357193768,0.042250596,-0.0589483567,-0.3305425346,-0.2346704453,0.3367640972,-0.0243675839,-0.5406258702,-0.0598177575,-0.1033514589,0.4902529418,-0.2907180488,-0.2391142845,0.264402926,0.0985087454,-0.1334737986,0.3457525969,-0.3871741295,-0.0687726736,-0.3274451792,-0.2053570002,-0.178157106,-0.5563392043,0.2077001482,0.217939347,0.4722729027,-0.0265835319,0.1018110886,0.2164962143,-0.008327798,0.317196846,0.2194681466,0.3106540442,0.3153970838,0.265087992,0.0506344438,-0.2692220211,0.4026501775,-0.0347031951,-0.1895988286,0.4249106646,0.1096412539,-0.2834822536,-0.1662667394,-0.1825755388,-0.2632664144,-0.0982692242,-0.2017259151,-0.0998434797,-0.0955644026,0.1610024869,0.372579813,-0.0214932784,-0.0154800806,-0.017948512,0.0524507537,0.1742388755,-0.0866842866,-0.1069167554,-0.1101591587,0.0763605461,0.2488822639,-0.2102680355,0.4049270451,0.0124372495,0.040290691,0.0373990834,-0.0827137008,0.279918313,0.0878548026,0.1663915813,0.2080863863,0.1845299751,-0.3854837716,-0.2265732735,-0.0073562898,0.0034428318,0.4128106833,-0.2838337719,-0.0367008634,0.1073252708,0.105301857,-0.3341072798,-0.3146873116,-0.3984829783,-0.5422887802,-0.0462575182,0.0941668674,-0.1859787554,-0.0265785083,-0.0266988743,-0.0450295843,-0.2808746099,-0.0295262169,-0.2933042943,0.0767029151,0.3623654246,0.1453165114,-0.1536247283,0.3892409205,0.0530667789,0.3490187526,-0.0021113639,0.1700937599,-0.0093389712,0.1469714195,0.0131881526,-0.162760675,-0.0046576066,0.0081787175,-0.1679845303,0.1843100935,0.0791243762,0.0034831488,-0.3126168251,-0.2270733416,0.1840350628,0.2306521833,0.2594152987,-0.1533780396,0.5802628994,-0.2575391829,-0.3947588801,0.1737313718,0.1379369646,0.052668117,0.3628042638,0.4564633369,0.7187488079,0.0311748963,-0.2819401622,0.0413098,-0.51452142,0.1140026599,0.189113304,0.1449657232,-0.3229416907,0.0082439706,0.0614341348,0.0766414404,0.1501955092,-0.0489879921,-0.2675614953,0.1717067361,-0.0789731517,0.2087402493,0.1791489422,-0.3609255254,-0.0128051033,0.1378910244,0.0290504545,0.2185210735,0.0584137999,0.1873106211,-0.0727211386,-0.0095098922,-0.0218447521,-0.0332304835,-0.1531891227,-0.0687229186,-0.1186951846,-0.2165898383,0.5413071513,0.0438259728,-0.1325315833,0.3093574345,0.6902180314,0.0007581609,-0.1699671596,0.2800174356,-0.0298520811,0.124141708,-0.0973239169,-0.0164406393,0.3743915558,-0.1857570261,-0.3134050369,-0.0423005521,0.2072261274,0.1064539552,-0.4617937505,0.1705457121,0.1554320902,-0.0807098523,0.2868440151,-0.3752064109,0.4050652683,-0.3073514998,0.2059989125,0.0123027572,0.1421100497,0.2774687707,-0.1887096614,-0.0542957373,-0.2139101774,0.1062781587,0.0925619602,0.455901742,0.1758491397,-0.2018762678,-0.0153572876,-0.223347649,0.1256690472,0.1125431433,0.2423885465,-0.0565082729,-0.0950918272,-0.3736506999,-0.0508184507,0.4535966814,0.2255276442,0.056494385,-0.1029301733,-0.3656066954,-0.073225677,0.1258585751,-0.3612861335,-0.0033514104,0.0246561542,0.0145930108,0.1365249902,0.4795310795,-0.3566233814,-0.3215708137,-0.1910586357,-0.1010519266,-0.0467890576,-0.2436276376,-0.0644729733,0.0296674874,0.1096922234,-0.2070846707,-0.1735155284,-0.1844582856,-0.2107448727,0.0547923557,0.2060494274,0.1098678634,-0.2236348391,-0.0662462413,-0.1040379405,-0.0479368567,-0.024941247,-0.197478354,0.116817601,0.4568612874,-0.3068595529,-0.0005457038,-0.0125741335,0.130270049,0.1598881632,-0.0779121965,0.0364825651,-0.0833674967,-0.0716839284,-0.1070487797,-0.0986887738,0.0936210603,-0.2261183858,0.0079764454,0.1434172988,-0.1781702936,0.1308995038,-0.0296085272,0.0917970613,-0.1381935179,-0.0352733694,-0.1387369037,-0.1055568084,0.228967756,-0.3008519411,-0.2021581233,-0.5758332014,0.0776181892,0.072062768,0.2196634263,0.107943736,-0.1131732091,-0.0044888225,-0.031396009,0.1584740877,-0.0816691816,-0.1365462542,0.1137649789,-0.1724721491,0.0461781509,0.3820456266,-0.0105383843,0.0121971713,0.6491846442,-0.0384012684,0.3165826201,0.3499768972,0.1997150481,-0.0201673508,0.1599977016,0.7009029388,0.7578041553,0.0614890605,-0.1245578751,0.045205947,-0.1923233122,0.0190007109,-0.2935475111,-0.380114615,-0.1871431619,-0.1286732703,-0.2488827705,-0.6894242167,-0.1609437615,-0.0036977599,-0.2322810739,-0.2159403116,-0.119130671,-0.1177180037,-0.0233480502,-0.2115475684,-0.0282761455,-0.3620203137,0.3541433215,0.0339795724,-0.1035919338,0.4498360455,-0.0433008373,0.2705354989,0.3618825674,0.2316312939,0.1712512672,0.4133895636,-0.3754761815,0.0382299498,0.0134705994,-0.2692512572,0.0443045795,0.2056946456,-0.0908258334,0.4918970466,0.2253949344,0.1106724665,-0.1005335301,0.1952467561,0.0866474435,0.7187842727,-0.2881050706,0.0432560332,-0.214932695,0.0099301953,0.078611888,-0.0805394948,-0.2154162973,-0.1744560152,-0.1637699157,-0.0498710647,0.1193574741,0.02545329,0.1429724246,-0.107928507,0.3923468888,-0.1283168942,0.3971066177,-0.2525192201,-0.0085225794,-0.6135665178,-0.0628590286,0.3399309814,0.1016351506,0.2081202269,0.0242326669,0.1190306693,0.3776516914,0.2008865774,0.374281466,0.0656084269,-0.2396687716,-0.5903016329,-0.0760593936,0.2660719752,0.1337560266,0.0175196789,-0.0821204334,0.3042609096,0.1094170883,0.0679458231,0.2985953987,-0.3129928112,0.262614727,0.1415283531,0.1556359828,0.2645022273,0.0387490503,0.0397642776,-0.1440609843,0.0288873799,0.1145183071,-0.2595925629,-0.294380486,-0.014036946,0.5275291204,0.2039332688,-0.088011086,0.0539391972,0.2366225719,0.205160588,-0.3053720593,-0.1925453544,0.2172718346,0.3083986342,0.0602690652,-0.1087512448,0.0126001257,-0.00501777,0.2415713519,-0.3445965648,-0.2118752152,0.4881680608,-0.4782317877,0.2080924362,-0.0283025987,0.5134817958,0.135270521,-0.1670482159,-0.141918689,-0.0933434963,0.2906729579,-0.0832345262,-0.0966840088,0.0413641818,0.0166038815,0.192694962,-0.0406478122,0.1490509659,0.2806019485,-0.0605726317,0.2938956022,-0.1157161742]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2499","title":" Python Programming Puzzles","comments":"\ud83d\udc40 @TalSchuster","body":"## Adding a Dataset\r\n- **Name:** Python Programming Puzzles\r\n- **Description:** Programming challenge called programming puzzles, as an objective and comprehensive evaluation of program synthesis\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2106.05784.pdf\r\n- **Data:** https:\/\/github.com\/microsoft\/PythonProgrammingPuzzles ([Scrolling through the data](https:\/\/github.com\/microsoft\/PythonProgrammingPuzzles\/blob\/main\/problems\/README.md))\r\n- **Motivation:** Spans a large range of difficulty, problems, and domains. A useful resource for evaluation as we don't have a clear understanding of the abilities and skills of extremely large LMs.\r\n\r\nNote: it's a growing dataset (contributions are welcome), so we'll need careful versioning for this dataset.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":2,"text":" Python Programming Puzzles\n## Adding a Dataset\r\n- **Name:** Python Programming Puzzles\r\n- **Description:** Programming challenge called programming puzzles, as an objective and comprehensive evaluation of program synthesis\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2106.05784.pdf\r\n- **Data:** https:\/\/github.com\/microsoft\/PythonProgrammingPuzzles ([Scrolling through the data](https:\/\/github.com\/microsoft\/PythonProgrammingPuzzles\/blob\/main\/problems\/README.md))\r\n- **Motivation:** Spans a large range of difficulty, problems, and domains. A useful resource for evaluation as we don't have a clear understanding of the abilities and skills of extremely large LMs.\r\n\r\nNote: it's a growing dataset (contributions are welcome), so we'll need careful versioning for this dataset.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n\ud83d\udc40 @TalSchuster","embeddings":[-0.1130172983,-0.1368338317,-0.2620641291,-0.0835469961,0.0358389355,0.0504642762,0.0034132623,0.2657264173,0.0189120006,0.1132145673,0.0565157831,0.2820882797,-0.415989846,0.3511270881,0.2383207679,-0.3106215596,-0.042996034,-0.0099194534,-0.1516429633,0.0325130522,-0.21212928,0.2472493649,-0.0816592127,-0.1468262672,0.0469824038,-0.2560288906,-0.1659639925,0.217937544,-0.0694615319,-0.3740359247,-0.3979611695,0.0396701954,-0.1897542924,0.319380343,-0.0000953417,-0.4529803395,0.1464274377,-0.0271556526,-0.5330190063,-0.0440043621,0.1155821159,-0.249318853,-0.2218259871,-0.1487129629,-0.1123274788,0.0289854035,-0.0693441629,-0.4078004956,0.4747589827,0.24019593,0.3734829128,0.2950743437,0.1882399917,-0.0484627336,0.1520218402,-0.0697945431,-0.1328537017,0.1764703989,0.479308933,-0.1994533241,-0.0156673733,0.0763710365,0.0965293869,-0.1362722367,0.1025375947,-0.2898264825,0.083574079,-0.5018382668,-0.135106504,0.0983733535,0.1450345814,-0.1308940202,-0.2803508937,-0.3875156939,0.0110252108,-0.1367087513,-0.2942920625,0.1817873418,-0.2155560106,0.019134203,0.0351377241,-0.1605034471,0.0022605078,-0.1401302665,0.11111404,0.1748879701,-0.0006833806,-0.1605917215,0.0670457184,-0.095545724,-0.1763480902,0.0502293333,0.0966259688,0.2590123713,-0.0991964117,-0.2498794198,0.2717814744,0.1405649334,0.032812342,0.3089974523,0.1460988075,-0.0196261462,0.3677435219,-0.0929140896,0.0949281678,0.0821012259,-0.1323208958,-0.0586258322,-0.0430718809,0.0612497181,-0.3035412133,0.1061330065,-0.0223845318,-0.1019858643,0.0541567467,0.141129449,-0.0896542966,0.2141446322,-0.1951935142,0.211777702,-0.2121958286,0.0098297056,-0.0095888963,0.0335004143,-0.1108221635,0.0327473991,-0.0126272021,0.050068818,-0.2794600725,-0.0015845497,-0.1702347249,0.0600313991,-0.3445099294,-0.1430073828,0.3644028008,0.1130493879,0.0475043729,0.1675461233,-0.0166860633,-0.281787008,0.3132359087,-0.000245643,0.18106848,-0.0102279317,-0.0818348601,-0.292817682,-0.006964094,-0.2580139041,-0.2648698092,0.1193898916,0.2889271975,0.0474345572,-0.1115078256,0.3787299693,-0.1390280724,0.1403591335,0.2786775231,0.3925020993,-0.1350934952,0.3035817146,-0.0099453311,-0.0996786207,0.2598269284,-0.191699937,0.2147495747,0.3130674064,-0.0700473115,0.0769152194,-0.0170190781,-0.0208810102,-0.3639110923,0.1832103729,0.1125266925,0.4617922902,-0.1711460203,-0.0090091666,0.3931082189,-0.1628927141,0.1382349133,0.0341057144,-0.1847982705,0.0155973323,0.0409635268,0.2908545136,0.0532662012,-0.0121409111,0.1958528608,0.3144518137,-0.0430374555,-0.0864306614,-0.2833215892,0.0249392185,-0.1651664972,0.1257258058,0.1925224364,-0.131222412,0.0468592681,-0.0646603033,0.182095468,-0.1623491943,-0.2575896084,0.100458771,0.2039763778,0.1677746326,0.0937685072,-0.5322014689,-0.1138562635,-0.0883557275,0.2051547021,0.1905958354,0.1354433447,-0.0872560367,-0.1167168394,0.2331330329,0.0999159813,-0.2078662664,0.3412594795,-0.1366814226,0.0918709487,-0.1697859168,-0.1373381168,0.3343254328,0.0885701254,-0.0330696627,-0.2791720033,0.0563989282,-0.1234606281,-0.1313773096,0.1390373111,0.5089416504,-0.0589487106,0.1890802383,0.1932872087,-0.0055757412,0.128437683,-0.2179922611,0.1305809319,0.3248330057,-0.1574540585,-0.2763596475,0.1172187179,0.28045398,-0.0127454372,0.1275875866,0.3104065955,0.3611837029,-0.0455163568,0.1029435992,0.2974846065,0.0584765896,-0.033421915,-0.0262903515,-0.3264105022,-0.2111174762,0.4473000169,0.0556484684,0.2634080946,-0.0596531667,-0.3262686431,0.2028658688,0.4614658952,-0.0460320339,0.2509970069,-0.1629679054,-0.0859655365,0.0346716344,-0.2329474837,0.2189608663,0.0341938622,0.3254306316,0.1897781938,0.2195358574,-0.2768262923,-0.219467476,0.1946178377,0.0435685292,-0.2321813852,0.2288419902,0.1336560845,0.1455089599,-0.0592726246,-0.2978995442,-0.2799660563,-0.0776688606,-0.244922176,0.2276792973,-0.074148044,-0.1418951601,-0.2535203099,0.0209591128,-0.1039745808,-0.4511721134,0.165075928,0.0960043222,-0.2812975347,-0.0166358296,0.0572550781,0.2937250435,-0.2418540567,0.1137345657,0.2068874389,-0.2365594208,0.0092574116,0.3167509735,0.08808662,0.022523975,0.7686954141,-0.4589861035,0.2996346056,-0.1382564753,-0.5123679042,0.0578949191,-0.1210451275,0.6592643857,0.3184371889,-0.0336850211,-0.0028857377,-0.1857105792,0.0377786756,-0.1628763825,-0.0368088931,-0.200826183,0.1737704426,-0.160857752,-0.2219827473,-0.186249122,-0.0467147753,-0.3656191826,0.38010934,0.1562811136,0.2217354476,0.0764976144,-0.0372320861,0.1718139052,0.0620451048,0.14133057,-0.0372312814,-0.134635359,0.0646560863,-0.3423058689,-0.2257515043,-0.2048908323,-0.4728269875,0.1169181615,0.0573044755,-0.0233899243,0.0305613838,-0.0341908261,0.2282916158,-0.0506172292,0.2994266748,-0.2082714587,0.2331350595,-0.2528220415,-0.2164483666,-0.232386753,0.1913551092,0.1109113619,0.0982312858,-0.1937645376,0.0358293951,-0.0831600428,0.5785011053,0.1512454748,0.1086236238,0.1698127985,-0.0227491874,0.2672560811,-0.0267195925,-0.3138284087,0.046919506,0.1697056741,-0.1577292234,0.2471685112,0.2632473111,-0.0175390914,-0.0088439202,0.0446345657,-0.2285562009,-0.0267238989,0.3655669689,-0.1606414765,0.1310187131,-0.0583880544,0.0371369533,-0.3915614784,0.0212304071,-0.1843534857,0.0421238318,0.038078662,-0.0465606861,0.0181043744,-0.2623780966,-0.3243571818,0.1160031408,0.1765453517,-0.0062632966,0.0755018815,-0.3191971779,-0.1897554547,-0.0394448861,0.1306208819,-0.0313728154,-0.0869639888,0.3423403203,0.0411274284,-0.0242578369,-0.0167773478,-0.3552360237,0.0085834991,0.2207346708,0.3590525687,-0.0162589923,0.0095117101,0.0220357291,0.3314071298,-0.0667902082,-0.1074398458,-0.2589554191,-0.4334926605,-0.3034576476,0.1099442914,0.2494662702,0.2289446741,0.0073721232,-0.0216505453,0.0379262306,-0.1136843488,0.3480602503,0.0915684104,-0.0254149009,-0.0549985543,0.136430636,-0.2686590254,-0.0811214969,-0.0953939408,0.1883782446,-0.3609694242,0.2143847197,0.0204192195,-0.0738314912,0.3978024125,0.2276903093,0.0245481376,-0.0606903359,-0.2769490778,0.3180642128,-0.384618938,-0.0583270043,0.0963773578,0.2389092594,-0.2866673768,-0.3703949451,0.0439401083,-0.1413425803,-0.1366539001,0.1394264102,-0.1646774113,-0.1007149071,0.1612150967,0.1610717177,0.7913138866,-0.2208357751,0.1603285074,0.2502910793,-0.3139734566,0.4829625785,0.0978196189,0.1960036159,-0.3161452115,0.0013105014,0.0891214162,-0.3105449378,0.1200226396,0.0698209107,-0.1118763983,0.1968564987,-0.1527346373,-0.1190608963,0.0889392942,0.4514316022,0.0075778244,0.0470386222,-0.4437475204,0.3082951009,0.0252195727,0.0138027277,-0.2757230997,-0.0198972821,-0.1422770321,-0.0159947407,-0.4113853574,0.0799347535,-0.1180117726,-0.0179911796,-0.0003185128,-0.2077369094,0.2137969732,-0.1584607363,0.3163831234,0.486438334,0.0981707498,-0.019015912,0.1760311276,-0.0770082399,0.2526631951,0.1382562518,0.4730743766,-0.2599129975,-0.0525067933,-0.098507911,0.2295685858,-0.143809393,-0.2894064784,-0.0914379656,0.0632623434,-0.0469530858,0.0573679432,0.0801587179,-0.127849102,-0.0820511058,0.2778334618,0.1169834062,-0.1311348528,0.1995596886,-0.0825324431,-0.0994408429,-0.0283635501,-0.3042224348,0.0739464387,-0.0893257111,0.2241064757,0.2654109597,-0.2040149868,-0.5156850815,-0.0463996977,-0.0315184779,0.0174604245,0.2494443655,-0.1060082093,-0.2152883112,0.0288376622,-0.015739385,0.2971253395,0.0632727966,-0.044034183,0.0049604937,-0.302177608,0.0423750095,0.121339336,0.1788960397,-0.3287988603,0.2789233625,0.1062274054,0.2636781335,-0.5504189134,-0.0277832989,-0.347779572,0.034888234,0.070184201,-0.3422400355,-0.4065733552,0.0610883273,0.3732415736,-0.0961578786,-0.4572246075,-0.4206407368,-0.3188762963,-0.0203563496,-0.2169270217,0.0352120586,0.1950778514,0.0038469199,0.0340013541,-0.2092923373,-0.077233091,-0.040694993,0.1003795117,0.0917219594,0.1085135862,0.1476854384,0.1000427827,0.0907354131,0.1108191907,0.168746084,0.0518576466,0.1430417001,0.0737833977,0.0352511294,-0.1518009156,0.055167608,-0.0163662788,0.1834301054,-0.0546524189,0.0286051314,-0.246855557,0.1593728513,0.1270123124,0.2611429393,-0.0014165784,-0.0782701075,0.195936963,0.3931059837,-0.184213683,0.2340249568,0.1818279773,-0.237306416,0.2643768787,0.352034837,-0.0023062148,-0.0099394061,-0.1449270844,-0.1250330657,0.1352436543,0.1617528647,0.282838136,0.180106461,0.1698342115,-0.0253223237,0.0635208338,-0.0623417161,0.0710190535,0.0985834152,-0.1034217402,0.1792205274,0.2320150137,0.0701676533,0.1924393624,0.2607126236,0.2828267813,0.1088452786,0.1078824773,0.2685012817,0.0210468639,0.0001076762,-0.2619569004,0.0530209951,-0.0460070968,0.3099086285,-0.2554865181,-0.0944074839,0.2873756886,-0.2974130213,-0.192853421,-0.1095960662,0.0535530411,-0.2853682935,0.0045474689,0.0103714177,0.2621122599,-0.0858357996,-0.1463328898,-0.0111450981,0.3736934662,-0.2322827876,0.1597076058,0.1619792581,0.1314662695,0.063757576,0.1720757037,0.4383860826,0.1311459839,-0.397904247,0.0848858431,-0.2091064751,-0.062909618,-0.265424639,0.1746910065,0.1066527441,0.2082899064,0.2641356587,0.2678040564,-0.2942237258,-0.2659398913,0.0544703752,-0.3729004562,-0.2548047304,0.388394624,-0.0880572274,-0.1534846723,-0.0868066922,0.1371442825,-0.495487839,-0.175563544,0.3857802749,-0.1092927381,-0.0525201485,-0.0836233422,0.1855084002,-0.0579339303,0.490986526,0.341062814,0.6164212823,-0.1057007983,-0.0773023292,-0.5067762733,0.0461301953,-0.0421451144,-0.2094443589,-0.0579220727,0.1723262668,0.0456611812,0.1901417971,-0.0167836044,0.2426739782,-0.1196846962,-0.0754790306,-0.4187950194,0.2009527981,0.0954052359,-0.0665484667,0.0195329785,-0.0364423953,-0.0083545027,-0.2529987395,0.2868320048,-0.096764259,0.0416146927,0.2972953022,0.0607958473,0.1218619719,0.1195612624,0.341131568,-0.0529318638,-0.586830318,-0.0859753639,-0.0186380912,-0.1245163754,0.1752615571,0.2411126941,0.3146766424,-0.0618853159,0.0172884222,0.0763710588,0.1399593055,0.2357483953,0.1428719759,-0.0467968285,-0.0382724144,0.1751233786,-0.0169421695,0.0605718121,0.4746253192,-0.1489088088,0.1677561402,-0.1270209849,-0.2663491666,0.2664871514,-0.186423108,-0.4249921441,-0.2556322515,0.2020238191,0.0413606986,0.1082691848,-0.5018566847,-0.0839004815,0.2567215562,-0.0964833647,-0.0214342475,0.4202648401,-0.2124410421,0.0445978679,0.0043647476,0.0166114178,-0.1760028899,-0.0296952408,-0.0236543156,-0.2535151243]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2499","title":" Python Programming Puzzles","comments":"Thanks @VictorSanh!\r\nThere's also a [notebook](https:\/\/aka.ms\/python_puzzles) and [demo](https:\/\/aka.ms\/python_puzzles_study) available now to try out some of the puzzles","body":"## Adding a Dataset\r\n- **Name:** Python Programming Puzzles\r\n- **Description:** Programming challenge called programming puzzles, as an objective and comprehensive evaluation of program synthesis\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2106.05784.pdf\r\n- **Data:** https:\/\/github.com\/microsoft\/PythonProgrammingPuzzles ([Scrolling through the data](https:\/\/github.com\/microsoft\/PythonProgrammingPuzzles\/blob\/main\/problems\/README.md))\r\n- **Motivation:** Spans a large range of difficulty, problems, and domains. A useful resource for evaluation as we don't have a clear understanding of the abilities and skills of extremely large LMs.\r\n\r\nNote: it's a growing dataset (contributions are welcome), so we'll need careful versioning for this dataset.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":17,"text":" Python Programming Puzzles\n## Adding a Dataset\r\n- **Name:** Python Programming Puzzles\r\n- **Description:** Programming challenge called programming puzzles, as an objective and comprehensive evaluation of program synthesis\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2106.05784.pdf\r\n- **Data:** https:\/\/github.com\/microsoft\/PythonProgrammingPuzzles ([Scrolling through the data](https:\/\/github.com\/microsoft\/PythonProgrammingPuzzles\/blob\/main\/problems\/README.md))\r\n- **Motivation:** Spans a large range of difficulty, problems, and domains. A useful resource for evaluation as we don't have a clear understanding of the abilities and skills of extremely large LMs.\r\n\r\nNote: it's a growing dataset (contributions are welcome), so we'll need careful versioning for this dataset.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nThanks @VictorSanh!\r\nThere's also a [notebook](https:\/\/aka.ms\/python_puzzles) and [demo](https:\/\/aka.ms\/python_puzzles_study) available now to try out some of the puzzles","embeddings":[-0.0690371692,-0.1561278403,-0.2759868503,-0.0527508967,-0.000647664,0.0701648518,0.0195636209,0.2562186122,0.0034628008,0.1276417226,0.0386161953,0.3095811307,-0.3876287043,0.3575746715,0.2331800461,-0.3480175138,-0.0425296463,-0.0044569364,-0.0990918577,0.0490014628,-0.1924676746,0.2509897351,-0.1473506838,-0.0799463317,0.0459989198,-0.2655206621,-0.1827941686,0.2118450552,-0.0962313265,-0.3995168209,-0.3597231209,0.0380528793,-0.1506151259,0.356187731,-0.0000946857,-0.4483577013,0.1580719501,-0.0328321941,-0.4988930523,-0.0523063503,0.1360356808,-0.2378710955,-0.1941977143,-0.1723545045,-0.1147182286,-0.0105924988,-0.0551272333,-0.3561468124,0.4866496921,0.2586369812,0.3744764626,0.2904421091,0.2213054895,-0.0379726924,0.1492575407,-0.1286346465,-0.157435745,0.2081690282,0.5029332042,-0.1821868271,-0.0104112653,0.0505567081,0.096982792,-0.1274922788,0.0570479035,-0.278098613,0.0993154868,-0.548532486,-0.1464281976,0.1003871411,0.1350398809,-0.1140143126,-0.254365921,-0.3878379762,-0.0214350503,-0.1134718135,-0.3032558262,0.2030613124,-0.2227848172,0.0181849226,0.0416184328,-0.1341142803,-0.0169906076,-0.0963678882,0.0912826732,0.2131258398,-0.0226303414,-0.1626652628,0.1081698835,-0.119375214,-0.1527066082,0.0314427391,0.1135204062,0.2347504348,-0.0868087187,-0.2609752119,0.2708579004,0.1239233166,0.0177956652,0.33556512,0.1480636001,-0.0060744844,0.3773305416,-0.1028960347,0.109333694,0.0928000659,-0.1637730598,-0.019672811,-0.0344462842,0.1125673279,-0.2922190726,0.0885627493,-0.0866945982,-0.1023688614,0.0563460588,0.1442159563,-0.0840299055,0.1820851117,-0.1716403663,0.2033602297,-0.2273284793,0.0299166795,-0.0007513291,0.0107504744,-0.1066906676,0.0260178782,-0.0303215552,0.0574970394,-0.2994653583,-0.0056301914,-0.1562254131,0.0581405722,-0.3145743012,-0.1795540005,0.3554589152,0.0870754495,0.041605141,0.1689746529,0.0125689367,-0.2826557755,0.306252867,-0.0045679612,0.1723783463,-0.017137466,-0.0784272552,-0.3076550066,0.0100729801,-0.2725691199,-0.2808988988,0.1333914995,0.2779684365,0.0449200533,-0.0949021652,0.3816742897,-0.1550905257,0.1056775898,0.2967669964,0.3964276612,-0.1516933888,0.3133289516,-0.0206085835,-0.1080102101,0.2293014824,-0.1770500839,0.2192471027,0.2873530388,-0.0541666485,0.0985069051,-0.0198064931,-0.0220546126,-0.3745416403,0.1586958319,0.1215590015,0.4654797316,-0.1846594065,0.0024684274,0.3773067296,-0.2076592296,0.1459478885,0.0320006981,-0.1530904472,0.0365185849,0.0253290199,0.2588229179,0.055617664,-0.0200537276,0.1840392202,0.314040482,-0.0406618565,-0.0831532478,-0.2736202776,0.0288454704,-0.1419500262,0.1271437854,0.1529227048,-0.1347588003,0.0323846154,-0.1064786389,0.1668428928,-0.1540444195,-0.2847520113,0.101653561,0.1968060136,0.1704759002,0.1061803475,-0.4948402345,-0.0954572186,-0.0782223791,0.2154520601,0.2092569023,0.1178890169,-0.067855604,-0.163056761,0.1954728961,0.0771314278,-0.2147099376,0.3580568433,-0.1830913872,0.1555978656,-0.1312407404,-0.1321453005,0.3327614069,0.0983534679,-0.0328270309,-0.293934077,0.0650842786,-0.1366574764,-0.132299751,0.1677355617,0.48086676,-0.0348925479,0.1765156537,0.1527649909,0.0091347564,0.168367222,-0.2324342728,0.081219174,0.3387724459,-0.156469658,-0.2968168855,0.1251401007,0.3188964725,-0.0220045634,0.1281387359,0.3337660134,0.3432987928,-0.0628278852,0.1198533103,0.3273600638,0.0607918687,-0.0467175134,-0.0157497115,-0.3352227509,-0.1780977398,0.4403584898,-0.0038231546,0.2370430529,-0.0717473552,-0.3317889571,0.1878476292,0.4561134279,-0.0291133486,0.2473886609,-0.1612132192,-0.1126268283,0.058572378,-0.2232492566,0.2160288095,0.0277732573,0.3220371604,0.2059086114,0.223083362,-0.2732074857,-0.2321784943,0.1780881286,0.0639737621,-0.2333123982,0.1899661273,0.1242356747,0.1144082993,-0.0533996299,-0.2826054096,-0.2626927197,-0.0810928941,-0.2464055866,0.2295690775,-0.1131106168,-0.1971317381,-0.2495608032,0.0127566662,-0.0888338089,-0.4756056666,0.1830380708,0.1165233776,-0.2463806719,-0.0390421152,0.041554641,0.2850763798,-0.2686403692,0.1312721819,0.1759013087,-0.2556347549,-0.0451754555,0.3335237205,0.1024491563,0.0451790653,0.7545125484,-0.4607816041,0.301704675,-0.1164506897,-0.5321436524,0.0517706983,-0.0727219954,0.6607948542,0.30450207,-0.0265441313,-0.0186404642,-0.170327276,0.0631332621,-0.1959264278,-0.0507183559,-0.1848328859,0.1850781292,-0.1605685204,-0.2132178396,-0.1860857308,-0.046531029,-0.3705588281,0.3902661502,0.2063408941,0.2195364237,0.0811281949,-0.0710448548,0.1906784475,0.0697576776,0.1549745053,-0.0232916307,-0.1519992501,0.0968977213,-0.3474819064,-0.2306865901,-0.1915353239,-0.4439184368,0.1147794649,0.0001647464,-0.0150546953,0.0378849432,-0.0818679109,0.215500921,-0.0263535175,0.3032610714,-0.2600440979,0.1974380761,-0.2345322967,-0.2070155442,-0.1931428015,0.1899802387,0.1369209886,0.0737718642,-0.175689131,0.0039855745,-0.0646066219,0.5416694283,0.1444898993,0.0965864584,0.1775265783,-0.0395968221,0.2368361354,-0.0356596075,-0.3107915521,0.0096796863,0.1647159904,-0.1436948031,0.2343010455,0.2779715359,-0.0106764538,-0.0441383049,0.0526839197,-0.2258945107,-0.0388458855,0.306301862,-0.1587509215,0.1283335686,-0.07370729,0.0262012407,-0.3611786962,-0.0159424935,-0.187070936,0.0389973037,0.0505211167,-0.0409504995,0.0347645544,-0.249402836,-0.3102258742,0.1455376744,0.1685859561,-0.0082251662,0.0812311247,-0.329603672,-0.1815980971,-0.0434763394,0.1425545663,-0.0119510526,-0.078571029,0.3668577075,0.0096085966,-0.0375765562,-0.030977048,-0.3742429018,-0.0176465679,0.19275482,0.3419071734,0.0150986472,0.0027152593,0.0409251899,0.3311522603,-0.0868282691,-0.0899551511,-0.2745040357,-0.4378645718,-0.3091596067,0.1035918742,0.2543663979,0.1961680204,-0.0251273066,-0.0307610855,0.0362063348,-0.1105725095,0.2938117087,0.1053560972,-0.0059825899,-0.0822005197,0.1436872333,-0.2420022488,-0.0984779075,-0.0850634128,0.1693283319,-0.3131100535,0.1889680326,0.0329419784,-0.048394952,0.3921136558,0.2157172412,0.0349283591,-0.0350554474,-0.2584016919,0.3342838585,-0.3643996716,-0.0816661045,0.0917204171,0.22048226,-0.2457787991,-0.3512534499,0.0310870036,-0.1219968051,-0.11327333,0.1257783324,-0.2043555528,-0.0937829316,0.1382654905,0.1394010633,0.7695464492,-0.2423823625,0.1440494657,0.2465070635,-0.3516128957,0.4781216681,0.0819467306,0.1643395275,-0.3154322803,-0.0308698323,0.0982630104,-0.3076474965,0.1232834831,0.0670363232,-0.1559761912,0.2217409462,-0.1696431637,-0.1114576384,0.0752782077,0.4630977511,0.0173103791,0.0861430392,-0.442892015,0.3158777952,0.0528279096,0.0239630211,-0.2839011252,0.0062365336,-0.1343144178,-0.0093881264,-0.3785085976,0.0911129937,-0.1066075563,0.0213018488,-0.0155666573,-0.1867976636,0.2088782936,-0.1206164286,0.3245629668,0.4761957526,0.0971525759,0.001804048,0.2171012759,-0.1232538894,0.253808707,0.1527249366,0.460475713,-0.2643861771,-0.0798586681,-0.0852969885,0.2536119223,-0.123215802,-0.3077405393,-0.0837399215,0.0708822161,-0.0119851064,0.0479115061,0.0312017854,-0.1383460313,-0.0562494136,0.2917148173,0.1155022308,-0.1229963824,0.1712290347,-0.0439925231,-0.0851932764,-0.0120807644,-0.3101010323,0.0772052482,-0.0763779059,0.1873510033,0.2989393473,-0.1959855855,-0.4983998239,-0.0272515025,-0.0620665401,0.027731277,0.2527610064,-0.1006314456,-0.145980373,-0.0119407661,0.0088391304,0.2668992281,0.0403582975,-0.0414745659,0.0136557799,-0.2834407985,0.038430512,0.1314191669,0.1851040274,-0.3328623772,0.2243736684,0.1172311231,0.2479749769,-0.5505171418,0.0041109133,-0.3728105724,-0.0001516921,0.0744423047,-0.3020956814,-0.4391433895,0.0567219704,0.3743461967,-0.0688629746,-0.4298588634,-0.4343464673,-0.2867398858,-0.0223497227,-0.2107211053,-0.0016585325,0.1870347559,0.0086107384,0.041539114,-0.1880383939,-0.1111691594,-0.0676059276,0.1136836261,0.1113967374,0.0959635153,0.1404606551,0.0950424373,0.0987855271,0.0849588886,0.1962415129,0.0901591852,0.1185632423,0.079428196,0.0507436506,-0.1370887309,0.0502441712,-0.0123636434,0.1506475359,-0.0515547357,0.0267640799,-0.2211907506,0.1216377243,0.1003751978,0.2775790095,-0.0283702482,-0.0733614117,0.2402865589,0.3897968233,-0.1823669523,0.227712214,0.1928217411,-0.2348291427,0.2318038195,0.3255149424,-0.0016309151,0.0244119707,-0.1715511084,-0.1144249067,0.147441566,0.1806930155,0.3262030184,0.181906119,0.1565909684,-0.037912868,0.040394064,-0.0785381123,0.0512471572,0.0784506649,-0.1034296751,0.1659890264,0.2315859795,0.074034147,0.2153995484,0.2852212191,0.254247129,0.1275483519,0.1080038622,0.2308660597,0.0185126513,0.0074525783,-0.2453173548,0.0821883976,-0.0752850398,0.2779008448,-0.2056855261,-0.0719239339,0.3132878244,-0.2913527489,-0.2342718989,-0.1108089462,0.0524084494,-0.2813006639,-0.0089149037,-0.0062080384,0.2553602159,-0.1142995432,-0.1499308795,0.0054702363,0.3869015276,-0.215578258,0.1432788968,0.1367815286,0.1342476159,0.0253645461,0.1762517244,0.4440714419,0.1253836006,-0.4179982841,0.0775766745,-0.2172596604,-0.0695187896,-0.2213422805,0.1724793911,0.1115686223,0.199608475,0.2587201595,0.2700965106,-0.3019435704,-0.2827014029,0.058990553,-0.3771444261,-0.2978942692,0.3868867159,-0.1045128033,-0.1454066336,-0.0439538993,0.101201795,-0.4862314463,-0.180212006,0.3434894681,-0.0849164128,-0.0505051948,-0.0594291687,0.1933020055,-0.0597739443,0.508376956,0.33082816,0.6604347825,-0.0987407714,-0.1256650388,-0.5089107752,0.0274068322,-0.0679127723,-0.2085276097,-0.0288377013,0.155577153,0.0242198538,0.2306130975,-0.020453617,0.2452186793,-0.1169552878,-0.0553173944,-0.4203046858,0.2021094114,0.0966740474,-0.0789750144,-0.0023820894,-0.0157948025,-0.042228166,-0.2822077572,0.2868263125,-0.0480558015,0.0356837064,0.2877689898,0.0443807542,0.141506955,0.1029954776,0.3527123034,-0.0143934898,-0.5512963533,-0.0410051458,-0.0339036472,-0.1100924611,0.188782379,0.2182328552,0.259593457,-0.0630077645,-0.0017134048,0.081893377,0.1197579652,0.2425085157,0.1526297927,-0.0231061615,-0.0692383423,0.1991019994,-0.0270715654,0.0771510154,0.4572689831,-0.1705478579,0.1309229285,-0.0883627757,-0.2311698943,0.251657337,-0.1641234607,-0.400752008,-0.2591993511,0.1849483401,0.0476105064,0.1128682494,-0.4909984171,-0.0792073831,0.2659678161,-0.1039646789,-0.0158232823,0.4436080158,-0.1985335052,0.0322358496,0.0246869028,0.0352488756,-0.1705404967,0.0034696499,-0.0374611765,-0.2289957851]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"That\u2019s interesting thanks, let\u2019s see what we can do. Can you detail your last sentence? I\u2019m not sure I understand it well.","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":22,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\nThat\u2019s interesting thanks, let\u2019s see what we can do. Can you detail your last sentence? I\u2019m not sure I understand it well.","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"Hi ! I just re-ran a quick benchmark and using `to_numpy()` seems to be faster now:\r\n\r\n```python\r\nimport pyarrow as pa # I used pyarrow 3.0.0\r\nimport numpy as np\r\n\r\nn, max_length = 1_000, 512\r\nlow, high, size = 0, 2 << 16, (n, max_length)\r\n\r\ntable = pa.Table.from_pydict({\r\n \"input_ids\": np.random.default_rng(42).integers(low=low, high=high, size=size).tolist()\r\n})\r\n\r\n\r\n%%timeit\r\n_ = table.to_pandas()[\"input_ids\"].to_numpy()\r\n# 1.44 ms \u00b1 80.1 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 1000 loops each)\r\n\r\n%%timeit\r\n_ = table[\"input_ids\"].to_pandas().to_numpy()\r\n# 461 \u00b5s \u00b1 14.2 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 1000 loops each)\r\n\r\n%%timeit\r\n_ = table[\"input_ids\"].to_numpy()\r\n# 317 \u00b5s \u00b1 5.06 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 1000 loops each)\r\n```\r\n\r\nCurrently the conversion from arrow to numpy is done in the NumpyArrowExtractor here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/d6d0ede9486ffad7944642ca9a326e058b676788\/src\/datasets\/formatting\/formatting.py#L143-L166\r\n\r\nLet's update the NumpyArrowExtractor to call `to_numpy` directly and see how our github benchmarks evolve ?__","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":150,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\nHi ! I just re-ran a quick benchmark and using `to_numpy()` seems to be faster now:\r\n\r\n```python\r\nimport pyarrow as pa # I used pyarrow 3.0.0\r\nimport numpy as np\r\n\r\nn, max_length = 1_000, 512\r\nlow, high, size = 0, 2 << 16, (n, max_length)\r\n\r\ntable = pa.Table.from_pydict({\r\n \"input_ids\": np.random.default_rng(42).integers(low=low, high=high, size=size).tolist()\r\n})\r\n\r\n\r\n%%timeit\r\n_ = table.to_pandas()[\"input_ids\"].to_numpy()\r\n# 1.44 ms \u00b1 80.1 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 1000 loops each)\r\n\r\n%%timeit\r\n_ = table[\"input_ids\"].to_pandas().to_numpy()\r\n# 461 \u00b5s \u00b1 14.2 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 1000 loops each)\r\n\r\n%%timeit\r\n_ = table[\"input_ids\"].to_numpy()\r\n# 317 \u00b5s \u00b1 5.06 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 1000 loops each)\r\n```\r\n\r\nCurrently the conversion from arrow to numpy is done in the NumpyArrowExtractor here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/d6d0ede9486ffad7944642ca9a326e058b676788\/src\/datasets\/formatting\/formatting.py#L143-L166\r\n\r\nLet's update the NumpyArrowExtractor to call `to_numpy` directly and see how our github benchmarks evolve ?__","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"Sounds like a plan @lhoestq If you create a PR I'll pick it up and try it out right away! ","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":20,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\nSounds like a plan @lhoestq If you create a PR I'll pick it up and try it out right away! ","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"@lhoestq I can also prepare the PR, just lmk. ","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":9,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\n@lhoestq I can also prepare the PR, just lmk. ","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"I\u2019m not exactly sure how to read the graph but it seems that to_categorical take a lot of time here. Could you share more informations on the features\/stats of your datasets so we could maybe design a synthetic datasets that looks more similar for debugging testing?","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":46,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\nI\u2019m not exactly sure how to read the graph but it seems that to_categorical take a lot of time here. Could you share more informations on the features\/stats of your datasets so we could maybe design a synthetic datasets that looks more similar for debugging testing?","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"I created https:\/\/github.com\/huggingface\/datasets\/pull\/2505 if you want to play with it @vblagoje ","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":11,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\nI created https:\/\/github.com\/huggingface\/datasets\/pull\/2505 if you want to play with it @vblagoje ","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"> I\u2019m not exactly sure how to read the graph but it seems that to_categorical take a lot of time here. Could you share more informations on the features\/stats of your datasets so we could maybe design a synthetic datasets that looks more similar for debugging testing?\r\n\r\n@thomwolf starting from the top, each rectangle represents the cumulative amount of it takes to execute the method call. Therefore, format_batch in torch_formatter.py takes ~20 sec, and the largest portion of that call is taken by to_pandas call and the smaller portion (grey rectangle) by the other method invocation(s) in format_batch (series_to_numpy etc). \r\n\r\nFeatures of the dataset are BERT pre-training model input columns i.e:\r\n```\r\nf = Features({ \r\n \"input_ids\": Sequence(feature=Value(dtype=\"int32\")), \r\n \"attention_mask\": Sequence(feature=Value(dtype=\"int8\")), \r\n \"token_type_ids\": Sequence(feature=Value(dtype=\"int8\")), \r\n \"labels\": Sequence(feature=Value(dtype=\"int32\")), \r\n \"next_sentence_label\": Value(dtype=\"int8\")\r\n})\r\n```\r\n\r\nI'll work with @lhoestq till we get to the bottom of this one. \r\n ","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":140,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\n> I\u2019m not exactly sure how to read the graph but it seems that to_categorical take a lot of time here. Could you share more informations on the features\/stats of your datasets so we could maybe design a synthetic datasets that looks more similar for debugging testing?\r\n\r\n@thomwolf starting from the top, each rectangle represents the cumulative amount of it takes to execute the method call. Therefore, format_batch in torch_formatter.py takes ~20 sec, and the largest portion of that call is taken by to_pandas call and the smaller portion (grey rectangle) by the other method invocation(s) in format_batch (series_to_numpy etc). \r\n\r\nFeatures of the dataset are BERT pre-training model input columns i.e:\r\n```\r\nf = Features({ \r\n \"input_ids\": Sequence(feature=Value(dtype=\"int32\")), \r\n \"attention_mask\": Sequence(feature=Value(dtype=\"int8\")), \r\n \"token_type_ids\": Sequence(feature=Value(dtype=\"int8\")), \r\n \"labels\": Sequence(feature=Value(dtype=\"int32\")), \r\n \"next_sentence_label\": Value(dtype=\"int8\")\r\n})\r\n```\r\n\r\nI'll work with @lhoestq till we get to the bottom of this one. \r\n ","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"@lhoestq the proposed branch is faster, but overall training speedup is a few percentage points. I couldn't figure out how to include the GitHub branch into setup.py, so I couldn't start NVidia optimized Docker-based pre-training run. But on bare metal, there is a slight improvement. I'll do some more performance traces. ","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":51,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\n@lhoestq the proposed branch is faster, but overall training speedup is a few percentage points. I couldn't figure out how to include the GitHub branch into setup.py, so I couldn't start NVidia optimized Docker-based pre-training run. But on bare metal, there is a slight improvement. I'll do some more performance traces. ","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"Hi @vblagoje, to install Datasets from @lhoestq PR reference #2505, you can use:\r\n```shell\r\npip install git+ssh:\/\/git@github.com\/huggingface\/datasets.git@refs\/pull\/2505\/head#egg=datasets\r\n```","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":18,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\nHi @vblagoje, to install Datasets from @lhoestq PR reference #2505, you can use:\r\n```shell\r\npip install git+ssh:\/\/git@github.com\/huggingface\/datasets.git@refs\/pull\/2505\/head#egg=datasets\r\n```","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"Hey @albertvillanova yes thank you, I am aware, I can easily pull it from a terminal command line but then I can't automate docker image builds as dependencies are picked up from setup.py and for some reason setup.py doesn't accept this string format.","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":43,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\nHey @albertvillanova yes thank you, I am aware, I can easily pull it from a terminal command line but then I can't automate docker image builds as dependencies are picked up from setup.py and for some reason setup.py doesn't accept this string format.","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"@vblagoje in that case, you can add this to your `setup.py`:\r\n```python\r\n install_requires=[\r\n \"datasets @ git+ssh:\/\/git@github.com\/huggingface\/datasets.git@refs\/pull\/2505\/head\",\r\n```","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":17,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\n@vblagoje in that case, you can add this to your `setup.py`:\r\n```python\r\n install_requires=[\r\n \"datasets @ git+ssh:\/\/git@github.com\/huggingface\/datasets.git@refs\/pull\/2505\/head\",\r\n```","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"@lhoestq @thomwolf @albertvillanova The new approach is definitely faster, dataloader now takes less than 3% cumulative time (pink rectangle two rectangles to the right of tensor.py backward invocation)\r\n\r\n![Screen Shot 2021-06-16 at 3 05 06 PM](https:\/\/user-images.githubusercontent.com\/458335\/122224432-19de4700-ce82-11eb-982f-d45d4bcc1e41.png)\r\n\r\nWhen we drill down into dataloader next invocation we get:\r\n\r\n![Screen Shot 2021-06-16 at 3 09 56 PM](https:\/\/user-images.githubusercontent.com\/458335\/122224976-a1c45100-ce82-11eb-8d40-59194740d616.png)\r\n\r\nAnd finally format_batch:\r\n\r\n![Screen Shot 2021-06-16 at 3 11 07 PM](https:\/\/user-images.githubusercontent.com\/458335\/122225132-cae4e180-ce82-11eb-8a16-967ab7c1c2aa.png)\r\n\r\n\r\nNot sure this could be further improved but this is definitely a decent step forward.\r\n\r\n","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":80,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\n@lhoestq @thomwolf @albertvillanova The new approach is definitely faster, dataloader now takes less than 3% cumulative time (pink rectangle two rectangles to the right of tensor.py backward invocation)\r\n\r\n![Screen Shot 2021-06-16 at 3 05 06 PM](https:\/\/user-images.githubusercontent.com\/458335\/122224432-19de4700-ce82-11eb-982f-d45d4bcc1e41.png)\r\n\r\nWhen we drill down into dataloader next invocation we get:\r\n\r\n![Screen Shot 2021-06-16 at 3 09 56 PM](https:\/\/user-images.githubusercontent.com\/458335\/122224976-a1c45100-ce82-11eb-8d40-59194740d616.png)\r\n\r\nAnd finally format_batch:\r\n\r\n![Screen Shot 2021-06-16 at 3 11 07 PM](https:\/\/user-images.githubusercontent.com\/458335\/122225132-cae4e180-ce82-11eb-8a16-967ab7c1c2aa.png)\r\n\r\n\r\nNot sure this could be further improved but this is definitely a decent step forward.\r\n\r\n","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"> ```python\r\n> datasets @ git+ssh:\/\/git@github.com\/huggingface\/datasets.git@refs\/pull\/2505\/head\r\n> ```\r\n\r\n@albertvillanova how would I replace datasets dependency in https:\/\/github.com\/huggingface\/transformers\/blob\/master\/setup.py as the above approach is not working. ","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":24,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\n> ```python\r\n> datasets @ git+ssh:\/\/git@github.com\/huggingface\/datasets.git@refs\/pull\/2505\/head\r\n> ```\r\n\r\n@albertvillanova how would I replace datasets dependency in https:\/\/github.com\/huggingface\/transformers\/blob\/master\/setup.py as the above approach is not working. ","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"@vblagoje I tested my proposed approach before posting it here and it worked for me. \r\n\r\nIs it not working in your case because of the SSH protocol? In that case you could try the same approach but using HTTPS:\r\n```\r\n\"datasets @ git+https:\/\/github.com\/huggingface\/datasets.git@refs\/pull\/2505\/head\",\r\n``` ","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":44,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\n@vblagoje I tested my proposed approach before posting it here and it worked for me. \r\n\r\nIs it not working in your case because of the SSH protocol? In that case you could try the same approach but using HTTPS:\r\n```\r\n\"datasets @ git+https:\/\/github.com\/huggingface\/datasets.git@refs\/pull\/2505\/head\",\r\n``` ","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"Also note the blanks before and after the `@`.","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":9,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\nAlso note the blanks before and after the `@`.","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2498","title":"Improve torch formatting performance","comments":"@albertvillanova of course it works. Apologies. I needed to change datasets in all deps references , like [here](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/setup.py#L235) for example. ","body":"**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n","comment_length":20,"text":"Improve torch formatting performance\n**Is your feature request related to a problem? Please describe.**\r\nIt would be great, if possible, to further improve read performance of raw encoded datasets and their subsequent conversion to torch tensors. \r\n\r\nA bit more background. I am working on LM pre-training using HF ecosystem. We use encoded HF Wikipedia and BookCorpus datasets. The training machines are similar to DGX-1 workstations. We use HF trainer torch.distributed training approach on a single machine with 8 GPUs.\r\n\r\nThe current performance is about 30% slower than NVidia optimized BERT [examples](https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/tree\/master\/PyTorch\/LanguageModeling) baseline. Quite a bit of customized code and training loop tricks were used to achieve the baseline performance. It would be great to achieve the same performance while using nothing more than off the shelf HF ecosystem. Perhaps, in the future, with @stas00 work on deepspeed integration, it could even be exceeded. \r\n\r\n**Describe the solution you'd like**\r\nUsing profiling tools we've observed that appx. 25% of cumulative run time is spent on data loader next call.\r\n![dataloader_next](https:\/\/user-images.githubusercontent.com\/458335\/121895543-59742a00-ccee-11eb-85fb-f07715e3f1f6.png)\r\n\r\nAs you can observe most of the data loader next call is spent in HF datasets torch_formatter.py format_batch call. \r\n\r\nDigging a bit deeper into format_batch we can see the following profiler data:\r\n![torch_formatter](https:\/\/user-images.githubusercontent.com\/458335\/121895944-c7b8ec80-ccee-11eb-95d5-5875c5716c30.png)\r\n\r\nOnce again, a lot of time is spent in pyarrow table conversion to pandas which seems like an intermediary step. Offline @lhoestq told me that this approach was, for some unknown reason, faster than direct to numpy conversion. \r\n\r\n**Describe alternatives you've considered**\r\nI am not familiar with pyarrow and have not yet considered the alternatives to the current approach. \r\n\r\nMost of the online advice around data loader performance improvements revolve around increasing number of workers, using pin memory for copying tensors from host device to gpus but we've already tried these avenues without much performance improvement. Weights & Biases dashboard for the pre-training task reports CPU utilization of ~ 10%, GPUs are completely saturated (GPU utilization is above 95% on all GPUs), while disk utilization is above 90%. \r\n\r\n\n@albertvillanova of course it works. Apologies. I needed to change datasets in all deps references , like [here](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/setup.py#L235) for example. ","embeddings":[-0.3501352668,-0.2402974069,-0.0736680552,0.190357089,0.1027540565,0.147706598,0.2599887252,0.6868174076,-0.121130459,-0.0641201735,-0.356641978,0.2431406975,-0.1145989373,-0.0332216397,0.0663354844,-0.3101109862,0.0007658997,0.1003077179,-0.0678961575,-0.1308621615,-0.1019239053,-0.4241634309,0.0556780435,-0.2308925092,-0.1764650196,0.1433885545,-0.0729704648,0.0049904119,-0.0147783048,-0.3914083242,-0.016802527,0.2041123956,0.1916381568,0.2366560847,-0.0001087394,-0.0219625924,0.105563499,0.0298551396,-0.3094554543,0.2616169453,0.3262714446,-0.3157902956,0.0545597821,-0.2053410858,-0.4622154236,-0.2044653594,-0.0115542626,-0.30791834,0.043709524,0.0324950144,0.1543279141,0.2567405403,-0.3000166416,0.1878047138,0.3291626871,0.4852654338,-0.1752856523,0.0986196473,0.3001522422,0.0291677509,-0.5153069496,0.6179615855,-0.1722578853,-0.0325031802,0.177194804,0.0603692159,-0.2424562275,-0.1164502576,-0.0824211463,0.4060195386,0.032524813,-0.2445900589,-0.500759542,-0.4453507662,-0.0309528373,-0.5959249139,-0.100147076,0.1113878042,-0.2834424973,0.1466880888,-0.0680339932,0.0368809551,-0.1203642115,-0.09613619,-0.0547588803,0.2587930262,0.2410730571,-0.2225960046,0.0905139223,0.1961138248,0.4557999671,-0.1157797202,0.2760972977,-0.0045529366,-0.3522320092,-0.1666412801,-0.0236723572,-0.1867082566,-0.1529744416,-0.0668888763,0.1264795959,0.5531223416,-0.0911921859,-0.0197006557,-0.2372612804,0.0375605449,-0.1032606214,0.210592702,0.1966996491,-0.1186811998,0.0677641109,0.1936755478,-0.0780171901,-0.0493074879,-0.3120107949,0.142736569,-0.4815460443,-0.0842902958,-0.1862215847,0.0252111387,-0.1306553185,-0.1333502084,0.0409616493,0.3111973107,-0.2863078713,0.1948336661,0.1680843681,-0.1474616975,-0.2475613654,-0.2679266036,0.0511941016,0.1703957319,-0.2893526852,0.0069734068,0.2033197582,-0.2458306253,-0.0934309065,-0.070954971,0.074785918,0.388759762,0.020032078,-0.2280967683,0.4766450822,-0.0271022301,-0.0679627433,0.0136563703,0.181370452,0.5773720741,-0.279609412,0.1573994309,-0.1916829199,-0.2062162906,-0.1592074186,0.1019587144,0.0700449198,0.007335064,0.0668838844,0.3424280584,0.1331920326,0.0767204538,0.126472339,-0.249081552,-0.3356597424,-0.1178224832,0.3001018167,0.3032240868,-0.1989410818,-0.0473861061,0.4758732319,0.3441790342,0.4220882058,0.4858781695,-0.3000439703,0.1179692149,-0.0290246345,0.0072558247,0.3716585338,-0.369782567,-0.1297076046,0.3484440446,0.074912928,0.2708399892,0.1608398706,0.1948831677,0.4046298862,-0.0059370184,0.0729750544,0.395489186,0.0082202358,-0.0499474593,-0.4824648201,-0.1739059389,0.4433811307,0.3957262933,0.0574343726,-0.3865258992,-0.1188643202,-0.1047391221,0.4031055868,-0.3227175474,0.1457478553,0.108986631,-0.1720594317,0.153413415,0.0151736438,0.2190547287,-0.1121948287,0.2242095768,-0.3645497561,0.3796460927,0.0479122065,-0.1950641721,0.3231682479,-0.0361455195,-0.0039837821,-0.2554191351,0.0407075249,-0.0446182266,-0.2046026289,0.0253762789,0.0424456075,0.3719467223,-0.2666930556,0.1153119951,-0.2319298089,0.1106730998,0.052906137,-0.0528733134,-0.0534533113,0.2816495895,0.0650102496,-0.0430517569,-0.0791779757,0.1298173219,-0.1343417764,0.2028240263,-0.2158796489,0.1773772687,0.4185040295,-0.1259617507,0.2937958241,0.1701381952,-0.0248790234,-0.1208947748,-0.422383219,0.7661342025,0.4835484028,0.1557845175,-0.1161572859,-0.2186772972,-0.2221040428,0.148417592,0.0560706258,0.0844040662,0.5234335065,0.1368191689,0.1734043807,0.0809791461,-0.4177882671,0.3712826371,0.341157198,0.005134359,-0.1374358982,0.4363709092,-0.0785387158,-0.3662787378,0.3380372822,-0.5506458282,-0.005098918,0.233039096,0.2333679199,0.0235326774,-0.0177598242,-0.0314640813,0.0983767435,0.01055016,-0.0542775206,0.0126836607,0.2190360725,0.2533146739,-0.0732507408,-0.0944208428,-0.1324299425,0.0236582067,-0.1572207958,0.1237859577,-0.4127211571,-0.103166461,-0.1553760469,-0.3362688124,-0.0705448091,-0.1799101084,0.0300737731,0.1377863437,-0.0285042822,0.1710874438,0.0217946917,0.0469736978,0.057029631,-0.2233769894,-0.1739785969,-0.1969986111,-0.0307846274,0.0338535979,0.147559166,-0.032628864,0.3699895144,-0.006935156,0.1059629098,0.0763379633,-0.2765926421,0.0485948883,-0.1336117834,-0.2551215291,0.0496093929,-0.1258975714,-0.2280783504,-0.1208018288,0.0972726569,0.0017792792,0.0587237999,-0.1553657651,-0.0902638882,0.1817816049,-0.0192360058,-0.2255808413,-0.245959267,-0.3073492646,-0.0496062972,-0.1677562743,0.156537056,-0.2041864097,-0.0283644665,0.1308830678,0.2082248777,0.0456591472,0.0236809626,-0.4292769134,0.4179665148,-0.2306931019,-0.1688781232,-0.401109308,-0.0202502627,0.0000506992,0.3673450351,-0.4472309649,0.2142145932,-0.0647428632,0.1862570643,-0.2460256517,-0.1165185198,0.1445492953,0.2486154586,-0.0935690776,0.0080999294,0.0720040649,-0.1541757882,-0.2329793125,0.102216132,-0.0745993704,0.3041974008,0.3246068954,0.6840358973,-0.0617227219,-0.3457929492,0.1029266343,0.1235023215,0.1042312533,-0.0933369845,-0.2323143929,0.1221762151,-0.207471773,0.3181696534,0.3299957216,-0.0535863489,-0.2173918486,-0.0138574708,-0.2203790396,0.1459464729,-0.2429201305,0.6411700249,-0.4819494784,0.2805368006,-0.0943444371,0.4774987102,-0.2672187686,-0.238350302,0.2915062308,-0.0068223383,0.2728028595,-0.1776013672,0.1038238779,-0.1798351556,-0.4914883375,0.2444355637,0.384765029,0.1953480095,0.0330887474,-0.0129563333,0.1963345706,-0.0567443818,0.2656693757,0.0066430154,-0.179469645,0.1652773321,-0.5235353708,-0.1951198727,0.0093970569,-0.0036706482,-0.0709712729,-0.0767353922,0.276170522,-0.118723616,-0.3965091407,0.2941373587,-0.053548649,-0.0598531142,-0.0008518329,0.0515360907,-0.0114567308,-0.1247614175,0.1667720228,0.0765728354,-0.0107124522,-0.2117273808,-0.0574950501,-0.1102331653,0.2859824002,0.349495858,0.0255461223,-0.1026383787,0.0217125639,0.1175977513,0.0775162876,0.3361064196,0.4761406183,0.0730550066,-0.3765414357,0.046519354,-0.0345419161,-0.0940706655,0.2185604572,0.5431827903,0.0941531733,0.357247442,-0.295946449,0.0031491355,-0.4404723942,0.1241221875,0.0799106061,0.0164612141,-0.5305218101,-0.5308513045,0.4449137151,0.0952895135,-0.0073607354,0.3080025017,-0.1986234337,-0.1722658277,0.4340583086,0.4619412422,1.0132721663,0.0184626319,0.4371165335,0.0783796534,-0.3031276464,-0.0109508978,-0.6706625819,0.2412650585,-0.3993288279,-0.2291055471,-0.0166533217,0.0206779297,0.1523991525,-0.0672575682,0.0205901675,0.0318231508,0.1562055051,0.0996146426,-0.0527221076,0.1821584404,0.2133357227,-0.7044994235,-0.2040035874,0.113750264,0.0907499269,0.1175582483,-0.0489601791,-0.257995218,-0.166382134,0.1905659139,-0.1383577734,-0.0450620987,-0.2100950629,0.0955520049,-0.2639068961,-0.8112205267,0.3461897075,0.1180452779,0.1777116954,0.1102318391,0.0089759324,0.1276231855,0.1128093377,0.337587446,0.1916223317,-0.3779072165,0.3423671126,-0.0188876651,0.0039915917,0.2039135396,0.045379106,-0.1885495782,-0.0305914376,-0.0427062325,0.283344835,-0.1624303013,0.089905642,-0.1114211306,-0.0500165597,-0.2463736236,0.1121333688,0.1338964254,-0.0950406268,0.2468922734,-0.0048645511,-0.2627278864,-0.2601156235,0.5095308423,0.1535368562,-0.1177702993,0.1691270471,0.3207577169,-0.5200848579,-0.1520256251,0.2127375603,0.452996701,-0.3142904937,-0.1394716501,-0.1035796702,-0.0857295394,-0.0949095935,-0.1106523201,-0.0858022645,-0.1794987768,-0.1017007902,0.1357043386,-0.4016144574,-0.0884814933,-0.2072799355,0.2621747851,0.0059843264,0.4504777491,-0.2052363902,0.2569230497,-0.3582395315,-0.11573679,0.1890013963,0.2031719834,-0.1223324388,0.1105043069,0.1015334576,0.2365470082,0.0470126942,0.3806158304,-0.0133108431,-0.1437841505,-0.2111733556,0.1359448433,-0.1630240828,-0.2040040195,0.1853450537,0.0713820904,-0.2378403097,-0.3986274004,0.2083637863,0.1215956658,0.0794300288,0.265021503,0.1749062985,0.1456349492,-0.0116831949,0.1614361554,0.0436111465,0.3500569761,-0.1312401593,-0.0331345238,-0.154723689,-0.181676507,0.0194093138,0.1629772931,0.428150773,0.0954229012,0.3320984244,0.0279594101,0.0267650392,-0.1860631406,0.3409809768,-0.0844898298,-0.062081337,-0.2638731301,0.3456560373,0.1365155727,-0.0907138139,-0.0626953393,-0.1355545074,-0.4172329307,0.1579431444,0.1829023063,0.5511474013,0.4552224576,0.1613767296,-0.0611742996,0.0412166752,-0.2157762349,-0.1608438045,0.4734087288,0.2522922456,-0.0158942416,0.2946068943,0.0607482418,0.1728645712,0.1979298741,0.1669101864,0.5242637396,0.4437017143,0.2266332954,-0.1603258699,-0.2803914845,-0.2760701478,0.07497596,0.0950399265,0.4341541231,-0.1632954031,0.2683142424,-0.0126119805,-0.4638894796,0.1848493516,0.5748338103,-0.242324993,-0.0748647302,0.2102371454,-0.0191637687,-0.3495026529,0.0930140615,-0.1050724611,-0.0751006529,-0.0156533103,0.3138209581,0.0348405987,0.0191513095,0.1164430305,-0.1667735428,0.4328149855,-0.2130880803,-0.0587464496,0.0340810455,0.005026822,-0.0648341998,-0.0555495434,0.2003889233,-0.09766379,-0.3771409392,-0.0379999056,-0.2076569647,0.1153810099,0.1750670522,0.0863945261,-0.0471863709,-0.2225128561,0.3377093375,0.0271601975,-0.1881805062,0.0681808591,0.0210095141,-0.0788613409,-0.0644232854,0.196948424,0.367418319,-0.1713654846,-0.2173270434,-0.3684085608,-0.244444862,-0.1997157335,0.5486779809,-0.0873466954,-0.0335329995,0.0955941156,0.1014815122,0.0862175971,0.6552398801,0.0800397992,0.1004414037,-0.3057168424,-0.1001805663,-0.0198476128,-0.0347769931,-0.0303325001,-0.1363439262,0.0663212463,-0.0353639536,0.2537302375,-0.01605542,-0.0654062405,-0.1982389241,0.2522170842,0.3421797156,-0.3151411414,0.1931947619,-0.5118898153,0.1501817256,-0.1294438392,-0.2208767533,0.2189540416,0.2018658668,0.0827887207,-0.283098191,-0.1504084021,0.2201898545,0.1271348149,0.4697135091,-0.0993740931,0.0384238362,-0.1637731194,-0.2426306456,-0.1156155318,0.1785310656,-0.1564409435,0.1788856536,0.1588065773,0.3668152392,0.1120522469,-0.5914254189,-0.2831096053,0.6362105012,0.1218942702,-0.1333908141,-0.1297865063,0.3011381626,-0.1914127171,-0.0089398492,0.1173386425,0.3214927614,0.4039704204,0.115647696,-0.6210435629,-0.2591513991,0.3328736424,-0.2074284554,-0.0719588473,-0.1597021669,-0.0799209476,0.269005388,0.5017656684,-0.7551468015,-0.123101294,0.1283411533,0.0959575325,-0.0990820378,0.3250200152,0.3227867484,0.0367257968,-0.3005054593,-0.0376769938,-0.1567607373,-0.0169451032,-0.3125458956,-0.2917817533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2484","title":"Implement loading a dataset builder","comments":"#self-assign","body":"As discussed with @stas00 and @lhoestq, this would allow things like:\r\n\r\n```python\r\nfrom datasets import load_dataset_builder\r\ndataset_name = \"openwebtext\"\r\nbuilder = load_dataset_builder(dataset_name)\r\nprint(builder.cache_dir)\r\n```","comment_length":1,"text":"Implement loading a dataset builder\nAs discussed with @stas00 and @lhoestq, this would allow things like:\r\n\r\n```python\r\nfrom datasets import load_dataset_builder\r\ndataset_name = \"openwebtext\"\r\nbuilder = load_dataset_builder(dataset_name)\r\nprint(builder.cache_dir)\r\n```\n#self-assign","embeddings":[-0.4386495352,0.1124225557,-0.1308667064,-0.0511096641,0.2150122821,0.1549260765,0.4422236681,0.5050137639,0.3474539518,0.2123945653,-0.2762616575,0.1705545038,-0.2276424021,0.2027029395,0.0419683196,0.3161544502,-0.1666620523,0.2632686198,-0.061166998,-0.1672290266,-0.3925251961,-0.1684753746,0.0735683143,-0.0296521802,-0.1065600514,0.0690301731,0.0736255944,0.1788545698,-0.0215610042,-0.2129750401,0.0496805832,0.0659999773,-0.0411656424,0.1969635934,-0.0001012373,0.1056393087,-0.0531033613,-0.0859771743,-0.4466030002,0.0643105805,-0.4056337178,-0.2063390464,0.1509783119,-0.3141410649,-0.1296840608,-0.1943142116,0.3031316996,-0.6111240983,0.2372525781,0.1936213225,0.3256171644,0.4264576435,-0.2099469006,-0.0189473946,0.1897360384,-0.1700785905,-0.1700597554,-0.2352287471,0.1854345053,-0.1043259576,-0.1550463438,0.277177453,-0.0333351903,0.288008064,0.0537235327,0.1306989044,0.1634563655,0.0486091599,0.1281135678,0.1102707088,0.6685625911,-0.4677182734,-0.3939364851,-0.1913031638,-0.2338885367,-0.161824584,0.0183749013,0.3360499144,0.1137421727,0.3574599624,-0.0309891645,-0.1338428706,-0.1585027575,0.0480413511,-0.2062451392,0.4421059787,0.0337585658,-0.1454127133,-0.1308233291,0.0285658874,0.4241783619,-0.0490444414,0.0141654033,0.2452636212,-0.1406351924,0.0102995411,0.3138629198,-0.00038475,0.0547911227,0.4004395604,0.4399606287,0.190397352,-0.2443340272,0.1677496433,0.0985160694,0.1995106786,-0.0278406423,-0.0911072344,0.3389090002,-0.2727329135,-0.225570187,-0.1149049923,-0.1546345502,-0.235239923,0.355963707,0.0904898345,0.3579250574,-0.1051378772,-0.2195517868,-0.056275852,0.2126763016,0.0160956383,-0.0100988308,0.4566091001,-0.161546886,-0.0321089178,0.3500881493,-0.103756994,-0.2063322663,-0.0521410145,-0.2976083755,-0.0833698437,-0.0530152209,-0.0301064234,0.1712602675,0.1493582726,0.2725954056,-0.30571419,0.0505933091,0.0250642281,0.1995487362,0.1242268458,0.3074888885,0.1307092607,0.1807388216,-0.1081985384,0.0917324722,-0.1924542487,-0.2957572937,-0.0470533073,0.0242268331,-0.3520313501,-0.2091060132,0.2963680327,-0.0010704944,-0.1318250149,0.1470997185,-0.1271548271,0.1476410478,0.042111434,-0.2003593445,0.3332744241,-0.0416150093,-0.1178591177,0.1675406247,0.4017371535,-0.0587884337,-0.1409438252,0.1810996532,-0.0155925294,-0.2348888218,-0.1579338312,-0.1212389544,0.2283561826,-0.121345982,-0.1013968438,0.4440607131,-0.4061499536,-0.3013504744,-0.0803037509,-0.086935319,-0.1035479158,0.0653389022,0.1668080539,0.1624435037,-0.1295315921,0.110896036,0.5430915356,0.0353543423,0.4065743387,-0.1065712571,-0.4038910568,-0.1015137061,0.1202948168,0.2124274522,0.1498224139,0.1431852579,0.0094895856,0.2163370252,-0.2182621509,0.1401317865,-0.0747600421,0.2412544042,0.0561892763,-0.2152118981,-0.2376358509,-0.3274627924,0.1901503354,0.0650887415,-0.171262458,0.132986322,-0.3308269978,-0.1918208301,0.2196937352,-0.1479758322,0.1467757821,0.3244475424,0.2962808907,0.1804625243,-0.0346784443,-0.1832453161,0.0890198201,0.1215728372,0.039898511,-0.0267071519,-0.1204898283,0.1964701563,-0.017538568,-0.0142993294,0.0934156477,-0.1569193602,0.1127253324,-0.1130300239,0.4207019806,-0.1611137092,0.1171525121,0.0432764105,0.3380193114,-0.0970739573,-0.0710122213,0.1465927809,0.0625834242,0.0048527229,-0.0184281003,0.0401380323,0.2778857052,0.2363986075,0.0524495877,0.0033314079,0.0003931512,0.0132752387,-0.1938538253,0.0940695554,-0.2320939749,-0.1382420361,0.1423946321,0.2542408109,-0.0413410105,-0.0749150589,0.0553639345,0.209418565,0.2019416839,0.3336173594,0.0798189342,-0.2894494236,0.0187135208,-0.1101952121,0.1492774338,0.2686703503,0.3510842919,0.2594451308,0.0031739709,-0.0244546849,-0.0522725135,0.2841052711,0.0614449009,-0.0793226957,0.1166185513,0.1635621339,-0.342223376,-0.2835030258,0.1211458147,0.0024479022,0.0438244157,-0.2574369013,0.0039559617,-0.190841645,-0.0608423986,0.0696659535,-0.0973089486,0.0990485549,-0.2782455087,0.1833358258,0.165690884,-0.2709335685,0.2768436074,-0.3441629708,0.3614455163,0.0500627011,-0.2153750807,0.0997991785,-0.3508519232,-0.0396318473,0.163386941,0.2029948086,-0.1612579525,0.5418078899,-0.1201263145,0.2365521491,-0.3537326753,0.0639478192,0.0101879537,0.2814016938,0.1614924967,0.0538186394,0.1830991656,0.1145747751,-0.1453064382,0.1119763777,-0.2351751029,-0.0158462152,-0.1853912622,0.15025419,0.1934743971,-0.1871703267,-0.2627907097,-0.3387160301,-0.5517283678,0.1003320962,0.195158571,0.2733900249,0.1209616587,-0.2121967971,-0.2019782364,0.1373336762,0.335024178,-0.2317690998,-0.4820139408,0.4451868236,-0.4399350882,-0.1463478506,-0.1165543795,-0.1675826013,0.0311417468,0.1288097203,-0.2987568676,-0.2944813371,-0.1676894128,-0.0272902194,-0.0596628189,0.2814137042,0.3208370209,0.3565166891,-0.2799209058,-0.1904506981,0.0349465311,0.192748487,-0.1642620414,-0.1252728701,0.0151075255,-0.1170239821,0.1163839549,0.2185883373,-0.148509115,-0.0669908598,0.4996663332,-0.1172322258,0.3154894412,-0.0565003678,-0.0570483617,-0.2735773623,-0.0524680354,-0.0579526611,0.4576520622,0.2869012654,-0.0491081104,-0.2233772427,0.1763802767,-0.3566323221,-0.4512857795,0.054207053,-0.1736109257,0.1063915938,0.1620369852,0.0622943752,-0.5818600655,-0.0605558604,0.1302194595,-0.0542932712,-0.0342787318,-0.2498277575,0.1152573973,0.4033688605,-0.2971266508,0.2720114589,0.0583029315,-0.2140874714,0.1925277263,-0.3345677257,0.0123750791,0.0026693223,0.3101823926,-0.1827374697,0.261898309,0.0748236477,-0.0537579916,-0.1135275885,0.1120743528,0.0263731703,0.0890267119,-0.0333027579,-0.0586301424,-0.3282371461,0.0251564346,0.153061077,0.1508702487,-0.3418646455,-0.011314516,0.006069534,-0.2312919497,-0.2954324484,0.0269122012,0.1168818623,0.1195727736,0.1328362525,0.0911230147,-0.5343552828,0.025518382,0.2235839516,0.1725773513,0.2648279071,0.0820506364,0.1247861683,0.0263447613,0.2657604218,0.0035536154,0.5341163874,-0.4113599956,-0.14844127,-0.1760944575,0.070154272,0.2012531459,0.2475168854,0.0839022547,-0.0975767002,-0.281617105,0.0208395179,-0.2958224416,0.3242608011,-0.1498579234,-0.156412974,-0.1862725317,-0.5517911911,0.4969351888,-0.0009233321,-0.2472214252,0.4861564338,-0.0257061049,-0.3974500299,0.3103632331,0.1083892211,0.8201705217,-0.2211490571,0.2912841141,0.2229505181,-0.1339534372,0.016097879,-0.2770312726,-0.0067988373,-0.2628659904,-0.3930805624,-0.0000544299,-0.0545694083,-0.1812525839,0.1185019463,-0.1659198552,0.1140419319,-0.137041375,0.0750713199,0.0852900296,-0.0312815197,-0.2590627372,-0.4709836841,-0.3592149913,0.304101944,0.011638578,-0.0041519818,-0.2124154717,-0.0197139531,0.1722748876,-0.0464223959,-0.0991120487,0.0080711907,-0.269230932,0.2423478961,-0.274784863,0.0577507429,-0.3058199883,0.2088020295,0.3539170623,0.2202454209,-0.1058524475,0.1249475256,0.1826902926,0.2132245898,-0.2110082507,0.1238427907,0.1727542728,0.080808863,-0.1223377064,0.0538523681,-0.0804773346,-0.4797339439,0.1501040757,0.0944898278,0.3296847045,-0.3257558048,-0.0745317936,0.076229088,0.0200786591,-0.2593949735,0.237236172,0.1547381729,-0.213799417,0.3150458038,0.1948484629,-0.2297902256,-0.0140388738,0.2903612852,-0.0677714571,0.1633123755,0.2805896699,-0.0534919798,-0.1873059869,-0.3221575022,0.048829671,0.0849418864,-0.1204283237,0.4324747026,-0.0058867526,-0.2542565167,-0.0493928604,0.2999030352,0.127870962,0.2996290624,-0.0418444313,-0.0727993622,-0.1567871571,0.0942892954,0.0532976128,0.5002894998,0.1618194282,0.1841808558,-0.2741753757,-0.0711282492,-0.4835540354,0.0913777575,-0.2396837622,0.1363035291,-0.0274768285,0.0117066819,0.2643021047,-0.0141743496,0.2157174498,-0.0434773527,-0.1753415316,-0.4168917239,-0.2314713746,0.104765512,-0.0624010898,0.0633039027,0.0598617345,-0.1284381151,0.0587096997,-0.4123634398,0.1405312121,0.0892063975,-0.0799280778,0.0235708039,0.4021606445,-0.0094765257,0.037111219,0.1160247326,-0.1744594127,0.2777879834,-0.1884236485,0.0744806454,-0.2552344501,-0.1228495613,0.2264015079,0.151152432,0.3040261865,0.0689066947,0.2093619406,-0.0099857086,-0.2478378564,0.3350842595,0.2927630246,0.0251244064,-0.069343321,0.1083991677,0.1710270047,0.3785525262,-0.219536975,-0.0173391253,0.1867946833,-0.3926481009,0.3878831267,0.26678285,0.0371399708,-0.1445843577,-0.2399749905,-0.0110542038,0.1776201874,-0.1699102968,0.1006622985,0.3335637152,0.0883249193,-0.0520056635,-0.085307464,0.2089117914,0.2837933004,0.2062918693,-0.0029272079,0.4163761735,-0.0150622157,-0.0827475488,-0.3613338172,-0.4804900289,0.4866734445,-0.0608475395,0.1670568734,0.3889097273,0.002391398,-0.0071483077,-0.1262432784,0.0704998225,-0.3571039736,0.4469643533,0.0432002135,-0.3685560524,0.1946020424,-0.0184117015,0.0646442845,0.0458734259,0.0249534603,-0.2651970685,-0.1799059659,-0.0704846233,0.4087218046,-0.0861140415,0.2507521808,0.0023451375,0.1956114769,-0.3705331683,-0.0715803951,0.1795305014,-0.0332586765,0.0541009828,0.5107300878,0.1423247159,0.1784802973,-0.2455127835,0.1596587896,-0.1680573374,-0.0266911443,-0.5964573622,-0.2120258212,-0.0311098993,-0.0419322439,0.372300595,0.3206718862,-0.2190718651,-0.0038258964,-0.1791327745,-0.1250610203,-0.0094658975,0.0902764052,0.3568990529,-0.1705809832,-0.1158213839,-0.0337724686,-0.2834278941,0.0291003101,0.1228373423,-0.2093199492,0.2823329568,0.1220153198,0.1457153559,0.2255584002,0.4714052677,0.1308293492,0.1607329398,-0.0656624734,-0.1603191346,-0.5151643157,0.1653640866,0.0916734412,-0.3174725473,-0.1655201614,0.0582624413,0.0453527831,-0.009424109,0.0161937438,0.0046920935,0.0218018144,0.1939229071,-0.3886311948,-0.2627260983,0.2340573072,0.0701092854,0.2360215634,-0.4280852079,-0.248187393,-0.0523750633,0.2937097847,-0.112416856,-0.0340686403,-0.0097925551,-0.1174720004,0.5315550566,-0.0993793532,0.0446466208,-0.2778207958,0.008903523,-0.157702744,-0.2208385766,-0.1426613778,-0.1170150936,0.0257521644,0.0700602531,-0.3827408552,-0.1369270086,-0.0372716263,0.1116005853,0.1648299396,0.2690051198,-0.2047214955,-0.0213779062,-0.0715662166,-0.0228370149,0.0633072779,0.1553797424,-0.2073812038,0.3147789836,0.0575184003,-0.4012447298,0.3633438647,-0.1625498384,-0.3417702317,-0.0608827248,0.4008001387,-0.0642096251,-0.1346590221,-0.3289476335,0.2549346089,0.3340220153,-0.0644124374,-0.1211607158,0.3138403594,0.0501456484,-0.0991485789,-0.1359989792,0.215388298,-0.0306824315,-0.1824738979,-0.3053699136,-0.3229510486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2481","title":"Delete extracted files to save disk space","comments":"My suggestion for this would be to have this enabled by default.\r\n\r\nPlus I don't know if there should be a dedicated issue to that is another functionality. But I propose layered building rather than all at once. That is:\r\n\r\n1. uncompress a handful of files via a generator enough to generate one arrow file\r\n2. process arrow file 1\r\n3. delete all the files that went in and aren't needed anymore.\r\n\r\nrinse and repeat.\r\n\r\n1. This way much less disc space will be required - e.g. on JZ we won't be running into inode limitation, also it'd help with the collaborative hub training project\r\n2. The user doesn't need to go and manually clean up all the huge files that were left after pre-processing\r\n3. It would already include deleting temp files this issue is talking about\r\n\r\nI wonder if the new streaming API would be of help, except here the streaming would be into arrow files as the destination, rather than dataloaders.","body":"As discussed with @stas00 and @lhoestq, allowing the deletion of extracted files would save a great amount of disk space to typical user.","comment_length":164,"text":"Delete extracted files to save disk space\nAs discussed with @stas00 and @lhoestq, allowing the deletion of extracted files would save a great amount of disk space to typical user.\nMy suggestion for this would be to have this enabled by default.\r\n\r\nPlus I don't know if there should be a dedicated issue to that is another functionality. But I propose layered building rather than all at once. That is:\r\n\r\n1. uncompress a handful of files via a generator enough to generate one arrow file\r\n2. process arrow file 1\r\n3. delete all the files that went in and aren't needed anymore.\r\n\r\nrinse and repeat.\r\n\r\n1. This way much less disc space will be required - e.g. on JZ we won't be running into inode limitation, also it'd help with the collaborative hub training project\r\n2. The user doesn't need to go and manually clean up all the huge files that were left after pre-processing\r\n3. It would already include deleting temp files this issue is talking about\r\n\r\nI wonder if the new streaming API would be of help, except here the streaming would be into arrow files as the destination, rather than dataloaders.","embeddings":[-0.1356213838,-0.3015790582,-0.1714526117,0.1553417295,-0.0760238245,-0.0403156914,-0.1220020875,0.5078596473,-0.0759534761,0.5109491944,0.1839041412,0.3984468281,-0.3721122444,0.2519610524,-0.1107841879,-0.1918504983,-0.0977600813,0.3879247606,0.0852558166,0.1664171815,-0.0113373147,-0.1601634026,0.0160193667,-0.1735659987,-0.1610468775,-0.2036674172,0.1742886603,0.0150632374,-0.3164533079,-0.3642852008,-0.0983547419,0.1170425713,0.1618454903,0.2395251989,-0.0001068214,-0.2105756253,0.1879487187,-0.0319870822,-0.3838619888,0.2532771826,-0.0644358918,-0.2182497084,0.020984346,0.0299099069,0.1138881817,-0.3589502871,-0.0091724787,-0.2793225348,0.6590781808,-0.1522957087,0.2111323476,0.1615836918,-0.3104902208,0.1768416911,0.0161841139,0.3898807466,-0.3480933607,0.1210870221,0.4206851125,0.1398861706,0.0320917815,0.3903601766,-0.061599277,-0.0733970255,0.5608956218,-0.2504017949,-0.0018344686,-0.6491248608,0.1009111628,0.0591418147,0.2640399337,-0.0302475076,-0.1682366282,-0.303260237,0.1826171428,-0.4667096436,0.1477618665,0.1348556429,-0.2945114374,0.0794928893,-0.115130432,-0.3061996996,-0.3989314437,-0.059657719,0.4848071337,-0.1565108895,0.045629777,-0.1460523605,0.3003849685,0.184118405,0.3356783092,-0.0832955539,0.0373022445,0.0891576409,-0.0273945481,-0.2479068786,-0.0708846524,-0.2019142658,-0.1583535224,0.2543638349,0.5093041062,0.2072617859,0.2530672252,0.0360715203,0.0824607238,0.1407410502,0.0225181095,-0.3135814369,0.2208501697,0.0810909495,0.1876424104,-0.035811048,-0.1597423255,0.0385523811,-0.1225860119,-0.0592423566,-0.0099593624,0.0264747832,-0.0054677497,-0.0316957384,0.0815773606,0.1349273622,0.0508202836,-0.0802242905,0.2607171237,0.583869338,-0.0523020402,0.2284559458,0.0538581572,-0.165893808,0.0019470691,0.0466800295,-0.1367949098,0.1242415383,0.2767364383,-0.1985018849,-0.0007487249,-0.1014563367,0.0445221774,-0.140229255,0.4040881395,0.1330535859,0.1663027108,0.1791575402,-0.1452494115,-0.2906928658,-0.1634166539,0.2401972413,-0.1942516714,0.317889899,-0.1339348406,-0.4168533087,-0.1584446579,0.2433993071,0.1234101728,0.3315380812,-0.3740736246,-0.0029831263,-0.185118258,-0.0062794401,0.1100404188,0.2283481508,0.1229907721,-0.1229147911,0.0011528463,0.099202618,-0.2082644999,0.0019111834,-0.3521500528,0.1476605386,0.2103511691,0.2112832814,-0.2174930274,0.1508672088,-0.0513635911,0.3056920767,0.3844228089,-0.0396411419,-0.3758777678,0.1250844002,-0.1171268672,0.0635380968,0.3967376947,0.1820128113,0.3838993907,-0.1808836758,-0.1668737382,0.3437715173,-0.0905222297,0.0229771491,-0.5567653179,-0.5417342782,-0.2692141235,0.0540842414,0.2500276864,-0.0277789179,0.1159667894,-0.0277544763,0.1581622213,-0.1157450154,0.2834402025,-0.0659279525,0.2879229784,-0.3269377947,-0.2581423521,0.1229908764,-0.2178051919,0.0017613814,-0.2269024849,-0.4157865942,0.0213951487,-0.3032647073,0.1413082033,-0.0749386176,-0.033028014,0.1762614995,0.1311375797,0.0012698655,0.0004972822,-0.3802503347,-0.517528832,-0.1215186268,-0.0330006853,0.0996584594,-0.2109696716,0.2710273564,-0.0107829068,-0.1132738292,0.098958835,0.0885714367,-0.2441337854,-0.2142184526,-0.0785041526,0.1750764996,-0.2076222003,0.4073339105,0.3072755039,0.3474683464,0.2859517038,-0.3495841622,0.2610292733,0.4026177227,0.0197046753,0.2463868707,-0.2008056343,0.1402627081,-0.1120508313,0.0014032872,-0.1213431656,0.0507838801,-0.145862475,-0.0369989164,-0.0701316074,-0.0479005501,-0.0976387262,0.1465107501,-0.1034493521,0.3378802836,0.2001923323,0.2742253542,0.4219549,0.1870205104,-0.1777384132,0.415522784,-0.0429303795,-0.3651118577,0.1717949361,0.2880438566,0.1948765218,0.1730131954,0.0711779967,-0.180332467,0.5252888799,-0.1838226616,0.4256459773,0.2466791272,0.0225341916,0.508084476,-0.1023883373,0.311640054,-0.2079672515,0.0337461568,0.0400211439,-0.0517089143,0.2694743574,-0.295173645,-0.1319106519,-0.0715440214,0.2251098454,-0.0056277905,0.0820553005,-0.0203735773,0.5460215211,0.2144937515,-0.6292132735,0.2050195485,0.0284267422,0.6209682226,-0.2706537545,0.1086053923,-0.4188214242,-0.0603151433,0.125739947,0.1639572978,0.3364751339,-0.025830403,0.3237527013,0.070252426,0.217561692,-0.3380711079,-0.1101963744,0.1077900827,0.1536597759,-0.2036513537,-0.0970041454,0.2288162112,-0.0879902467,-0.489382416,-0.2472160161,-0.4151543677,-0.0292705372,0.1077464595,0.1544803828,0.0633975565,-0.3697407842,-0.131983757,0.0087766601,-0.733098805,0.4366351664,-0.2078465968,0.1679955572,-0.0826826692,0.2305323035,-0.0202275682,0.1514763832,-0.0668018088,0.0914699063,-0.5894623399,0.0071841232,-0.1340795606,-0.2107977569,0.0536962077,-0.0734702572,0.0184512995,0.3070475459,-0.2912283838,-0.0223293565,-0.0389113612,0.1044672951,-0.1724076271,-0.0754538849,0.334551394,0.2662546337,-0.2806762755,0.1334557384,0.2682353854,0.0744224787,0.4424127936,0.1211573929,0.0608153977,-0.1510605812,0.3017262816,0.440004319,-0.1242219955,0.325374037,0.268820703,0.619294405,0.286211729,-0.2466650307,0.2488624454,0.1597568542,-0.0142295929,0.030119285,0.5266540051,0.1751338691,-0.1330192983,0.1773607433,-0.2286223322,0.0067670424,-0.1373528987,0.4198543131,-0.3682390153,0.0538393967,-0.036829643,-0.1332172155,0.0338170975,0.0297933649,0.1565911472,-0.0338706747,0.6322969198,-0.0506581031,0.0316304751,-0.019824123,-0.4064336121,0.2881213427,-0.0217886977,-0.3238893151,-0.0288006552,-0.3460864127,0.1331807673,0.1648305357,-0.1160241142,-0.0527964272,-0.2456238717,-0.2521002591,0.1071441993,-0.1838885099,0.0594569817,0.0782581642,-0.1445228457,-0.1094828174,0.3484134674,-0.147945255,-0.3527017236,-0.1920200139,0.3462498188,0.0017347308,0.23008053,0.2537962198,-0.2485167533,-0.470205307,0.1756508052,0.0466784425,-0.0842256397,0.0475145206,-0.3103623986,-0.0997417346,-0.0827753544,0.055635836,-0.0160805192,0.5145267248,-0.2448665053,-0.0280132312,-0.0558372587,0.328068167,0.4145462215,-0.0850938112,-0.0283863787,0.0317276902,0.279515177,-0.362119168,0.1528606266,0.4592937827,-0.092892237,0.0700968355,-0.1618494242,0.6410700679,-0.2481253892,-0.107886903,-0.0350547694,-0.2344015837,0.1601708233,0.0726491511,0.0842033997,-0.0746392235,-0.123702772,0.5312130451,0.0928425267,-0.1183322147,0.481025517,0.2648243904,1.1928399801,0.0984791815,0.2629880011,0.0237388462,-0.2545757294,0.1177985743,-0.5239387751,0.2042855173,0.1412040591,-0.1463821232,0.0214509312,0.1541945338,-0.0491921343,0.1785126925,-0.3350701034,-0.1128890812,-0.2717015147,-0.0900179446,-0.1906404793,0.239844352,-0.2895218134,-0.309799701,-0.1892155409,0.098218143,0.0206427276,-0.0075757061,-0.1533195078,-0.1998576969,0.0410634093,-0.1214550808,0.0852410048,-0.0540282093,-0.025707908,0.1632813513,-0.2693378925,-0.2498387396,-0.0495136157,0.0655817538,-0.2643714845,-0.1169710159,-0.1355528831,0.099499099,0.129672572,0.0827893168,-0.1416053772,-0.0721998215,-0.0112586329,-0.3059492409,0.1785344481,-0.0054919072,-0.201695919,-0.2403163463,0.1244636104,0.0450479947,0.1175175533,0.2096674442,-0.056658648,-0.1120105088,-0.5392418504,-0.1017621458,0.2102258354,-0.0896099135,-0.3118331134,0.215178892,0.0900764614,-0.197697565,-0.0840923414,-0.3686060905,0.26865834,-0.0519965179,0.2886566818,-0.4474610686,-0.1929422617,-0.5038599372,0.0540383905,-0.0020484377,-0.2841661274,0.3334969878,-0.0761460736,-0.4090226889,0.0741344765,-0.0396530367,0.2295855284,0.013162287,-0.2982761264,0.0957481414,-0.3086766899,0.3106725216,-0.1109745577,0.3808490634,-0.4966371655,-0.1239670888,0.103709802,0.2151007354,-0.3594698012,0.2248786241,0.1255707741,0.1538799852,-0.1385266185,0.0405912474,0.0187125672,-0.1323861629,-0.0316552706,0.0830751136,-0.2785665393,-0.2416483909,-0.1215564758,0.1332311183,-0.0590285361,-0.1028664783,0.1873571128,0.0986839607,-0.1260303408,-0.1760175675,0.083536759,-0.401843667,-0.0156525914,-0.2817939818,0.3741918206,0.071851626,-0.1541146338,0.1807672679,0.2047514319,0.396823287,-0.2970113456,-0.059815377,0.0040722126,-0.3454163373,-0.0427052192,0.2129657567,0.1374453455,0.103576228,0.2391906679,0.1692904085,-0.0652638525,-0.0482421517,0.2074065804,0.2450065762,0.162361905,-0.0944888815,0.0310066342,0.3144351244,-0.1226928234,-0.0987313539,-0.0921271145,-0.1310138851,-0.1981376857,0.0474094078,0.3405256271,-0.0968255699,-0.1183700711,0.0556953624,-0.1452519596,-0.2564465404,0.3973234892,0.3020839095,-0.0022492043,-0.1539879143,-0.1657870561,-0.1669609994,0.153188765,-0.1523235738,-0.4397920072,0.2478108257,0.4387491345,0.1649576128,-0.1390998811,-0.129975453,-0.2733868062,0.1077059433,-0.0892620161,0.2902970612,0.4061613381,0.4156827033,0.1732331961,-0.60458529,-0.1843392849,0.3101634979,0.0199213102,0.1079857126,-0.0446805023,-0.1847449839,-0.0763441324,0.0646372736,0.1822233349,0.3154657781,0.2762486041,0.4885822535,0.1737255007,0.210249275,0.330847472,0.1891872734,0.5358947515,-0.2661395967,-0.0681468993,0.0712614506,-0.0269336216,-0.0844229758,0.1401829571,0.1318357438,-0.0837927088,-0.2902522385,0.2506684661,0.0318590142,0.1053667367,-0.3437458277,0.1499020606,0.4966713786,0.1513185501,-0.1313322634,0.1357514113,-0.246019274,-0.0192422401,-0.207683593,0.0725888684,-0.1424960941,-0.1264312565,0.3554879427,-0.4864268005,-0.0173536595,-0.1077545509,-0.2301465869,0.0352147259,0.3305037618,-0.1148456559,0.0085049588,-0.1938133091,0.0956974775,-0.1864747107,0.4524026513,0.080753468,-0.2967921197,-0.1621742249,-0.2621740401,-0.414724499,0.0042784335,0.0679866448,0.2935778201,0.1867893636,0.164428547,-0.0221703053,0.1953957528,0.0234600417,-0.1550227404,-0.016104864,0.0711743906,-0.4328565001,0.4441393316,0.5078606009,0.2515916228,0.0945651531,-0.5529877543,0.1167289093,0.0347465836,0.0498265214,0.282808274,0.0953077078,0.1203657612,0.2672490478,0.4411725104,-0.0180258583,-0.3104726076,-0.1752664,-0.317533493,-0.1937985867,-0.0050584753,-0.1977740973,-0.1752252728,0.3306725621,0.4791045487,-0.3103908002,0.2205577046,-0.3596090078,-0.0526754633,-0.1067492366,-0.0309280492,-0.2534036636,-0.1696967185,-0.1470216513,0.3872558177,0.1568235457,0.293831557,-0.0571972802,-0.3464207351,-0.4373600781,-0.126547873,0.3629042506,0.12415196,-0.3802589774,0.2163328379,0.3051617742,0.4904469848,0.0896073878,-0.1151883528,-0.0336240456,0.0037739321,-0.132611528,0.0227950327,0.2835887372,-0.1408280134,-0.1451745927,-0.0007892739,0.3146114647,0.0007042213,-0.161728099,-0.1770175546,-0.4092116058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2480","title":"Set download\/extracted paths configurable","comments":"For example to be able to send uncompressed and temp build files to another volume\/partition, so that the user gets the minimal disk usage on their primary setup - and ends up with just the downloaded compressed data + arrow files, but outsourcing the huge files and building to another partition. e.g. on JZ there is a special partition for fast data, but it's also volatile, so only temp files should go there.\r\n\r\nThink of it as `TMPDIR` so we need the equivalent for `datasets`.","body":"As discussed with @stas00 and @lhoestq, setting these paths configurable may allow to overcome disk space limitation on different partitions\/drives.\r\n\r\nTODO:\r\n- [x] Set configurable extracted datasets path: #2487\r\n- [x] Set configurable downloaded datasets path: #2488\r\n- [ ] Set configurable \"incomplete\" datasets path?","comment_length":85,"text":"Set download\/extracted paths configurable\nAs discussed with @stas00 and @lhoestq, setting these paths configurable may allow to overcome disk space limitation on different partitions\/drives.\r\n\r\nTODO:\r\n- [x] Set configurable extracted datasets path: #2487\r\n- [x] Set configurable downloaded datasets path: #2488\r\n- [ ] Set configurable \"incomplete\" datasets path?\nFor example to be able to send uncompressed and temp build files to another volume\/partition, so that the user gets the minimal disk usage on their primary setup - and ends up with just the downloaded compressed data + arrow files, but outsourcing the huge files and building to another partition. e.g. on JZ there is a special partition for fast data, but it's also volatile, so only temp files should go there.\r\n\r\nThink of it as `TMPDIR` so we need the equivalent for `datasets`.","embeddings":[-0.2962553203,-0.1055004969,-0.1821807027,0.189564541,0.1187052503,-0.0878328681,-0.0424844585,0.2916618586,0.0462339111,0.347949326,-0.1803242266,0.1907633692,0.0141955828,0.3997286856,0.0212996267,-0.0809282586,-0.2167209387,0.2052152604,-0.2652953565,0.180523932,-0.2859375775,0.0219762586,0.0442879908,-0.065782547,-0.0264658909,-0.1949851513,-0.02131119,0.0202082526,-0.1970957816,-0.2983090281,0.1801560968,0.2298398167,0.0550775453,0.3512578309,-0.0000983048,-0.0042582424,0.0631370991,-0.1732392907,-0.6294649839,-0.2513601482,-0.019107502,-0.5226925015,-0.0321825072,-0.2287234068,0.0223126914,-0.245350942,0.1677339673,-0.3536904454,0.1197797135,-0.0993938595,0.3681629896,0.2698152363,-0.043047931,-0.1813036501,0.2348441333,0.2331928313,-0.3291924894,0.0507573597,0.3433953226,0.2788313329,0.0131397964,0.0068487185,0.0295037478,0.3137277365,0.2971015275,-0.030737495,0.1122198477,-0.4712475836,0.0997477323,0.2625061274,0.5066245198,-0.3221046329,-0.3209204972,-0.4933105111,-0.0251851361,-0.1129073128,0.2811666727,0.1690110713,-0.304670006,0.1274890453,-0.3215426207,-0.133631438,-0.2272516489,0.2419533581,-0.0089452351,0.023959199,-0.1358814538,-0.0463341437,0.2854825556,0.1762559414,0.2132878006,-0.4044156969,0.1854973286,-0.0745382607,0.2001808286,-0.2847127914,-0.0808054656,-0.1120696664,0.0643399954,0.0092399204,0.0790888667,-0.0684436858,0.0642204136,0.0518128462,0.2564743459,0.0517005175,-0.2714928985,-0.0614267848,0.2406294793,-0.0381081663,0.217418924,0.048367843,-0.1444502622,-0.3023051322,-0.3724725246,-0.1259241402,-0.0751999244,0.0211896189,-0.3299500048,-0.3805194497,0.4152680039,0.0354148671,0.0559008196,-0.0144029129,0.134567529,-0.1691818684,-0.237179175,0.2818698287,-0.0938867629,0.0925750136,-0.0458540209,-0.0457006469,-0.2854702771,0.0869987905,0.2056019455,0.0078324061,0.1775724739,-0.1700742841,-0.0977415442,0.2256355882,0.4130826294,-0.0440710448,0.0689796582,0.3716663122,0.1851373017,0.1394669265,-0.2464462221,0.3634608686,-0.3581136763,0.2483163476,-0.2918246686,-0.440557152,0.0944593325,0.2806108296,-0.1643900573,0.000140954,-0.1923115402,0.0523219816,-0.0897785947,-0.0549220294,-0.0231637824,0.1035751551,-0.0189472828,-0.2156531066,0.1324109286,0.1738235056,-0.3945197165,-0.0616732948,-0.3099749684,-0.0832765102,0.1494045705,0.2036529183,-0.2687669396,0.1125361696,-0.356792897,0.1980066746,0.3827776611,-0.0618517846,-0.3516402543,0.1939809769,-0.117787607,-0.0740234852,0.2361405343,0.152056098,0.4000906348,-0.0394077897,-0.0989351124,0.5456370115,0.0877942368,0.0515716597,-0.2195086926,-0.6255843639,-0.0960066319,-0.0157759879,0.0799059793,-0.1631885767,0.2333318889,0.0486646704,0.0973784328,-0.0917536095,0.2353197038,0.0877411142,0.1764996648,-0.0881695747,-0.2954939604,0.0713954493,-0.3603202999,0.2684758008,-0.1432165056,-0.2140787393,0.0800394565,-0.0703865215,-0.0012989908,-0.2523118556,0.0203486439,-0.1723855585,0.25114429,0.1763208508,0.1582447588,-0.2860212326,-0.1124312952,-0.1914063394,0.082382448,0.0651015192,0.083324112,0.4587310553,-0.1591846347,-0.0013214529,-0.0579477064,0.1388753951,-0.2309408635,-0.3591414094,0.07074745,0.2254110128,0.0796133652,0.2203642428,0.2830663621,0.2405697554,0.2718525231,0.0907001793,0.2883929014,0.1396323442,-0.2050853968,0.1913954318,-0.3032184541,0.3291305304,-0.1139847264,-0.0407863706,0.1621512622,0.0199148115,0.1656925529,-0.0579242557,0.1096840054,-0.3439187407,-0.0135338139,0.0395466387,0.2599923611,0.2794912159,0.218991667,0.2041254342,0.3102349043,0.2133680433,0.0524346605,0.1946962029,0.010002315,-0.2498704046,-0.0461732782,0.5935592055,0.4426033497,0.4874846041,0.3771624863,-0.2000807077,0.0602378435,-0.3213655353,0.4303214848,-0.0018881229,-0.0261024106,0.1658359319,-0.2078157514,-0.0010186218,-0.1154893413,-0.1544919163,0.2019416541,-0.0175744742,0.0483082868,-0.389996171,-0.3329841197,-0.2143124044,-0.0565149374,0.0724727437,0.0167133603,-0.0889772624,0.348167181,0.1882576644,-0.3835696876,0.1427307427,-0.0516349748,0.3345633149,-0.2055127323,-0.2572612762,-0.0680400208,-0.0189592559,-0.001406703,0.240559727,0.2626760602,-0.0290705059,0.318559587,-0.0115838209,0.0144249005,-0.514113605,-0.1823756844,0.1400062144,0.1323257685,0.1149186715,0.0611002035,0.2755704522,0.2675017118,-0.18912597,0.0262238756,0.1039331481,-0.0953405648,-0.0886662304,0.0824930891,-0.099150762,-0.306763947,-0.2084642351,-0.2901556194,-0.6281812787,0.2831303179,-0.1043651327,0.2514548898,0.1861862391,0.4520158172,-0.0152569767,0.0524816662,-0.0490883254,-0.0354688093,-0.4996625781,0.2473182529,-0.4030629098,0.078699477,-0.1015755534,0.1316581219,0.1765712649,0.4386541843,-0.515838325,-0.0015856476,-0.2677957416,0.3502958119,0.2252785563,0.0487867184,0.4365338087,0.0645734668,-0.2161233872,-0.0506754592,-0.0251302719,0.1632097811,0.3058544993,0.1377157718,0.0263393484,-0.1690971702,0.1141312569,0.3385214806,-0.145069927,0.2844698727,0.3558307588,0.2150807381,0.5061594844,-0.2726740241,0.1611344665,0.280484587,0.0531855971,-0.1758492887,0.3591487706,0.0730059519,-0.0294820424,0.1004183516,-0.2248380035,0.0020400947,-0.2578144073,0.3602555692,-0.0735457763,0.1644790322,-0.0234714095,-0.0519988351,-0.2224878669,0.099679783,-0.1062944904,0.0971950144,0.2565867603,0.2397233993,-0.1812212616,0.1004308835,-0.1038938165,0.1167722642,-0.0764035434,-0.3136499524,-0.1662271321,-0.0961618498,0.0451200716,0.2097720653,0.1315774769,-0.2113986015,-0.1321148127,-0.0274165701,-0.1807540655,-0.4094070792,-0.0728057921,0.0874167755,-0.0580557771,0.0789031535,0.5266671181,-0.1968787909,-0.2153991461,-0.1217999086,0.3659332693,-0.0654598549,0.0785347223,0.1812670529,-0.2583677471,-0.4736134112,-0.0506499,0.0821811855,-0.0766173825,-0.0295024235,-0.2210759372,-0.2395141572,-0.0184686016,-0.0091566751,0.1260850132,0.3321581185,-0.2643472254,0.1873908937,0.1375816166,0.1935504973,0.57376647,0.2259551734,-0.404071331,0.1423267126,0.3827581406,-0.2888180912,0.0072536571,0.3047213256,-0.0270465277,-0.069520995,-0.2594329715,0.2425066531,-0.3263025582,0.0453180075,0.1171565726,-0.0337289535,0.1162496284,-0.2784149945,-0.0462170541,-0.1442334503,-0.0496926792,0.3877956271,0.1219698265,-0.4149861634,0.0624525547,0.2417076975,0.8535844684,0.0596354529,0.2840288281,-0.0271896068,-0.2870078981,0.1652580947,-0.3779338896,-0.0063582258,0.1090159416,-0.1220232695,0.0640725195,-0.0663448498,0.048565805,0.3011220694,-0.2952788174,0.0256754551,-0.0339053422,-0.0607207566,-0.3045534194,0.0530405417,-0.1731689274,-0.3400924206,-0.1264634281,0.26016137,0.0835288018,0.1092492193,-0.0729474127,-0.1383631676,0.0851603374,0.2667833269,-0.0093866037,-0.3033587933,-0.2119657993,0.135273695,-0.2492734641,-0.1983186156,0.068477951,0.1015303135,0.125380069,0.2845492959,-0.3281392455,0.0609660707,-0.1740519404,0.0178485401,-0.0457571261,-0.2124135792,0.0729297176,-0.1777646989,-0.0396775007,-0.1771306247,-0.0823265016,-0.1294359416,0.0814886913,-0.1745873988,-0.0134932231,-0.0381381325,-0.0767596737,-0.1831595898,-0.0874967501,-0.2126350701,0.2323913872,-0.1197541952,0.0281713679,0.2539649904,-0.1687567234,-0.2555294633,-0.3182780743,-0.0453005247,0.0031734481,-0.1112609953,0.1956821084,-0.4225511253,-0.1719763577,-0.4166550338,0.3600195646,-0.1063459441,-0.1055905744,0.1877847612,-0.1937761903,-0.2437065393,-0.2756361365,0.1186489016,0.4634706378,0.146036908,-0.1659792662,-0.1332042515,-0.1763457507,0.2984718978,-0.1030163839,0.4938419759,-0.2298555225,-0.006388444,0.3236541748,0.0469660461,-0.4705732167,0.2156821638,0.0209626295,0.0801619515,0.2978313565,0.0259255022,-0.0845197961,-0.1299291998,0.1000135317,-0.1460113227,-0.3717274964,-0.2606518567,-0.2561435699,0.0818950981,-0.0667125061,-0.0635349974,0.2903601527,-0.3504109383,-0.1260032207,-0.2786107063,0.0276008565,0.0203402136,-0.0193352234,-0.1616066098,0.1106605455,-0.1023945659,-0.3127557337,0.3415350318,0.2284643054,0.3963299394,-0.0050034155,0.2641149461,-0.152800262,-0.1785360873,-0.1952912807,0.2308612168,0.2098824531,0.2751529217,-0.1250120848,0.2627412379,0.1357966065,0.0835193917,0.2154494673,0.2932986915,0.1641919464,-0.1230342314,0.3854133487,0.3824098408,-0.0723155886,-0.0085141268,-0.0328316726,0.0173369478,0.1773392558,0.271579504,0.3052975535,-0.0188170001,0.0958164185,-0.0291585438,0.1579670906,0.0897261947,0.1839332879,0.5496280789,0.0454632901,-0.0588179976,-0.0501451418,-0.0826155171,0.3256779909,0.3156008124,-0.4169416726,0.3879027069,0.2968867719,-0.09835352,-0.0058437693,-0.2980697155,0.1633033752,-0.1229990274,0.0538106672,0.0835566148,-0.0361458808,0.1923865527,-0.2933646142,-0.1780525744,-0.1463740468,0.1173957065,-0.201806426,-0.2556242049,-0.1328032315,-0.4109455943,0.0578253902,0.0312959552,-0.0727863908,-0.1329727322,0.4485038519,-0.0657494888,0.1454711407,-0.0972801149,0.5200176239,0.1480005533,-0.0879779756,-0.1540926248,0.2658529878,0.0016303831,-0.0782704353,0.0029765358,0.2568162382,0.33275792,0.0681134388,-0.044131726,-0.1560366899,0.1860788614,-0.0133511359,-0.4736658633,-0.1506228745,0.1344253719,0.0375676043,0.3874790967,0.3374322951,-0.2408827245,0.0464148931,0.0127061307,-0.2789759338,-0.13714616,0.0342933387,0.3196047843,-0.2384693325,0.3096679449,-0.1275579035,-0.1416871399,-0.2346616834,0.3317603767,-0.2541353106,0.1927250028,0.1957082003,0.1663005054,0.0739166066,0.5967606902,-0.1065526605,-0.2863192856,-0.2140313387,-0.0822027996,-0.4470919371,0.0459093563,-0.1815075576,-0.117420204,-0.2055389732,0.2257558107,0.002251911,0.0561159588,0.2056533545,0.2857435942,0.1163190901,0.288862884,-0.3156137168,-0.0352347493,0.33701244,0.0967785642,0.0349236019,-0.4380877614,0.01212133,0.0611366145,0.1633024514,0.0401520059,0.2764396071,0.3812283874,0.0807804912,0.3357610703,-0.0835609362,0.0662950203,-0.1477476209,-0.0354165211,-0.3296090662,-0.0463789068,-0.143149823,-0.1811527014,0.1976809055,0.1099560037,-0.4670315087,-0.0205278974,-0.3134146631,0.0510444157,0.17039828,0.258392036,0.2977572083,0.0304791406,-0.1053458154,0.1648833752,0.3832371235,0.2409545481,-0.02439899,-0.081954807,-0.2712875307,-0.2433438748,0.0291072205,0.3065900207,-0.2554537058,-0.0098424274,0.0677875653,0.3064528406,0.2222564965,-0.1710164994,0.1331201494,0.3338829279,-0.1364162415,-0.0711197108,0.1565680653,-0.0535921417,-0.0989545658,0.0490135849,0.363745898,0.023652656,-0.3637844026,-0.2438346297,-0.2443176508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2475","title":"Issue in timit_asr database","comments":"This bug was fixed in #1995. Upgrading datasets to version 1.6 fixes the issue!","body":"## Describe the bug\r\nI am trying to load the timit_asr dataset however only the first record is shown (duplicated over all the rows).\r\nI am using the next code line\r\ndataset = load_dataset(\u201ctimit_asr\u201d, split=\u201ctest\u201d).shuffle().select(range(10))\r\n\r\nThe above code result with the same sentence duplicated ten times.\r\nIt also happens when I use the dataset viewer at Streamlit .\r\n\r\n## Steps to reproduce the bug\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\u201ctimit_asr\u201d, split=\u201ctest\u201d).shuffle().select(range(10))\r\ndata = dataset.to_pandas()\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\ntable with different row information\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.4.1 (also occur in the latest version)\r\n- Platform: Linux-4.15.0-143-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.8.1+cu102 (False)\r\n- Tensorflow version (GPU?): 1.15.3 (False)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\r\n- `datasets` version:\r\n- Platform:\r\n- Python version:\r\n- PyArrow version:\r\n","comment_length":14,"text":"Issue in timit_asr database\n## Describe the bug\r\nI am trying to load the timit_asr dataset however only the first record is shown (duplicated over all the rows).\r\nI am using the next code line\r\ndataset = load_dataset(\u201ctimit_asr\u201d, split=\u201ctest\u201d).shuffle().select(range(10))\r\n\r\nThe above code result with the same sentence duplicated ten times.\r\nIt also happens when I use the dataset viewer at Streamlit .\r\n\r\n## Steps to reproduce the bug\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\u201ctimit_asr\u201d, split=\u201ctest\u201d).shuffle().select(range(10))\r\ndata = dataset.to_pandas()\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\ntable with different row information\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.4.1 (also occur in the latest version)\r\n- Platform: Linux-4.15.0-143-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.8.1+cu102 (False)\r\n- Tensorflow version (GPU?): 1.15.3 (False)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\r\n- `datasets` version:\r\n- Platform:\r\n- Python version:\r\n- PyArrow version:\r\n\nThis bug was fixed in #1995. Upgrading datasets to version 1.6 fixes the issue!","embeddings":[-0.0035169786,-0.3718388081,0.0401346385,0.5538208485,0.2874104381,0.1348073632,0.3841583133,0.2454013526,-0.3271685243,0.1167894155,-0.1194758415,0.5017601848,0.0366128646,0.2132893205,0.3299592733,0.1174107939,0.0804404095,0.2233543247,-0.2730519176,-0.1913636327,-0.0837137699,0.0187306833,-0.3005526066,-0.0541513413,-0.3123556674,0.356143713,-0.1421096027,0.0773136839,-0.0027764938,-0.2180716395,0.0472317748,0.2179221213,0.1011598185,0.4397813678,-0.0001224917,-0.0437433049,0.1536538899,0.0376670025,-0.3571340144,-0.4788444638,-0.2536831796,0.1657094359,0.4298107028,-0.0168080609,-0.1919686794,-0.2204069495,0.0153245693,-0.0032241109,0.0580783524,-0.0587263927,0.1279315948,-0.0700518563,-0.309948951,-0.008073437,0.0979072675,0.0198308602,0.1419012845,0.3659571409,-0.0061307163,0.1408261657,0.0504534505,0.1574636251,0.0190609489,0.4090979695,0.1579133719,0.0722733736,0.0735200793,-0.1263838112,0.199909091,0.3509020805,0.6390126944,-0.2943965197,-0.1062136069,-0.1762697697,0.2817424834,-0.08698643,0.1077077538,0.133087039,-0.1625984311,0.0371367522,-0.1377103329,0.1535277665,-0.0191015434,0.2034159452,-0.1244539544,-0.1964819878,-0.0487429053,0.2289688289,0.1935901493,-0.0528363958,0.5600512624,0.0873591453,0.0777280033,0.130802229,-0.6907063723,0.1696691066,0.0721739009,-0.0325104296,0.094234556,0.2098889053,0.7568898201,0.2299492508,0.0220174622,0.0542643517,0.4541719556,0.0384740271,-0.0719310269,0.1269022822,0.0248846076,0.0864315778,-0.2165447474,0.0293507408,-0.139364481,0.2273674905,0.2358614504,-0.2979688644,0.2498427927,-0.2742166519,-0.3862969279,-0.0973418429,-0.2942026258,-0.2052494735,0.0058446289,0.2198114097,0.143686533,0.3313932419,0.0135661382,0.1407025903,-0.1352301091,-0.1151422262,-0.2645073235,-0.1136845648,0.0038655815,0.0393517688,0.174080953,-0.7548032999,0.119043462,0.2988196313,0.0704429522,-0.3749945164,0.1200077534,-0.148879841,-0.0531569682,0.2885536849,0.2717417181,0.2477754503,0.0247594696,-0.1103553548,0.1554389,0.4112883508,0.0332736336,-0.075457871,0.1279896945,0.0739616528,-0.106872879,0.0948502198,-0.5075110197,0.282317996,-0.0402662717,-0.1628939062,-0.2071588933,-0.0856745988,-0.3177864254,-0.237057656,0.205155611,0.4185120165,-0.317825228,-0.0580350906,0.0877428874,0.0224855673,0.2892636657,0.0578873977,-0.2492666245,0.2181712836,-0.3331076205,-0.1878444403,0.2407715023,0.0288726799,-0.4438280463,0.0824931338,-0.2400319129,0.2394569516,0.1515150815,0.2172440588,0.3593176603,0.229555428,0.2278988957,0.136348784,0.0951143131,-0.0320127569,-0.3715804219,0.108915858,0.3166851699,0.3442661166,-0.1897285134,0.4779950082,0.1436803788,0.1430115998,0.480586648,0.0868988112,-0.1557339281,-0.0679573119,-0.141367808,0.0873393267,0.2371721268,-0.2569782734,-0.104639329,0.114471361,0.0702198818,-0.1638260037,0.1512080878,-0.0381399617,-0.1412693262,-0.1575072557,-0.4172098935,0.2337392122,-0.0013295427,0.4847946167,-0.2759517431,0.0077961739,0.018921202,0.3798327148,-0.2186736763,-0.0998165458,-0.4190430343,0.2982143164,0.0183487143,0.0733887106,0.199136138,0.0242233165,0.2635972798,0.001062081,-0.0978280529,0.3068014681,-0.0522252731,0.0980819315,-0.2294414192,-0.3256539404,0.2034775764,-0.1790530682,-0.1523263454,0.1329062134,0.2259669155,-0.2348288149,-0.1908884794,0.0888515636,-0.4352977872,-0.059339609,0.0559671745,-0.0203923807,0.2167896777,0.2009063512,0.054088071,-0.2700942457,0.2669419944,0.0969452336,-0.2118966579,0.0695681423,-0.3587127328,0.3053561151,0.0435449928,-0.1185381263,-0.0010358487,0.0557230972,-0.1536088139,0.0327988304,-0.0142302187,0.1491773725,0.3201248944,0.140716821,0.1753589213,-0.0296908394,-0.0065775411,-0.0287113544,0.0275578797,0.0794805884,-0.2302872539,0.3513810635,0.1529009491,-0.1802951843,-0.4084851742,0.1318781525,0.4431325793,0.4914958477,-0.3171830177,-0.1029348448,-0.5063694119,-0.1688468903,-0.1197777912,-0.1918087751,0.150651142,-0.1039073989,0.0424144231,0.582588613,0.1701629758,0.1234191805,-0.059378624,0.146754384,-0.0415806286,0.4056347609,-0.0827774405,0.2911361158,-0.3537473977,0.0198738687,0.1480403841,0.4129242301,0.345783323,-0.1272243559,0.0866375789,-0.3266282678,-0.1138819382,0.0348282494,-0.0026255788,0.341607362,-0.0980547443,0.2032835484,-0.4332409501,-0.1082395613,0.1651758254,-0.06180517,-0.0860461891,0.262493968,-0.0178274047,0.1571684629,-0.0706277788,-0.3673730195,-0.028507011,-0.2026819885,0.1661893576,0.0136311371,0.1407264918,-0.0618070327,-0.0628626347,-0.0676675811,0.1209830865,0.0624878518,-0.5289276242,-0.3424660861,0.0759001747,0.1124719977,-0.2627945244,-0.2293483317,-0.0050927531,-0.0159993358,0.158577919,-0.4302548766,0.1413328797,0.1013187766,-0.0909470916,-0.288977474,0.0094875479,0.3224421442,-0.0118670072,0.0226783939,-0.0764338747,-0.0925165787,-0.1031485647,-0.0584525242,0.1427298188,0.1165368631,0.4198417664,0.0344364233,0.2283519059,0.3895139992,0.162799269,0.2522086799,-0.2143780142,0.2165403664,-0.1666733176,-0.3694883287,-0.2346042097,-0.1239013001,0.13815476,0.39603585,-0.0687943846,0.0730805323,-0.1683626026,0.0966980383,-0.3277001679,-0.1975039691,-0.0878393427,-0.3895818293,0.0878070071,0.2924848199,0.1140475795,0.1794026047,-0.1967069507,-0.2590691447,0.1011512354,0.0998888761,-0.0855076909,-0.2796019316,-0.3747205138,-0.244121328,0.0374122038,0.2548140883,0.5431174636,0.0872618183,-0.0723344907,0.4424591362,0.2266438156,0.9034943581,-0.2364865839,0.1118959412,0.1163688824,-0.0560200699,-0.307687223,-0.0252503511,-0.1297630221,0.1261994988,-0.0923706591,-0.0735935196,-0.1352334172,0.0983444452,0.1874624789,0.3085967302,-0.3238923848,-0.3346426487,-0.3275571764,-0.1353564262,-0.1294291764,-0.1198240668,-0.1626115292,0.0817146152,-0.2781119347,-0.1312412769,-0.2557206154,-0.0252144784,0.2389314622,0.2292592824,0.5405646563,0.2320352346,0.6274511814,0.6013333797,-0.2336233705,0.1861926317,0.7264572382,-0.0644944087,-0.1752109677,0.2515898347,-0.4403795004,-0.0625117496,0.3564408123,-0.2024867535,0.1901327372,-0.0959645808,-0.1289777458,0.0266820788,-0.0920629725,0.1403669715,-0.127286911,-0.2795651853,-0.6341582537,0.3452590108,-0.0461691767,-0.2197108716,0.4685609937,-0.312957406,0.0617322512,0.219299078,-0.5134537816,0.9313744307,-0.2697829306,0.2258369625,-0.1068179309,-0.164093852,-0.0025164271,-0.0024716307,0.1238741279,-0.4160215855,-0.6455469131,-0.0732767284,-0.089289926,-0.3054291308,0.2855671644,-0.0724272281,0.1596588641,-0.1064738408,0.1483987868,0.2026425004,0.1122846529,0.100789316,0.0971990377,-0.0278849881,-0.0593248494,-0.1661498845,-0.0353697091,0.278950125,0.1174491569,0.3159681261,-0.2417311966,-0.2599372268,0.0473024659,-0.2113705724,0.1991027594,-0.2948182821,-0.471603632,-0.1733734757,-0.0371930003,-0.3053154647,0.0402245075,0.0003335544,-0.0776367411,0.4851885736,0.441753298,-0.3377182186,0.0368821882,0.2285597175,0.1567443311,-0.367464453,0.1212372929,0.3146136701,-0.1519770175,-0.4787367582,0.3721342385,0.0737086758,-0.3079115748,0.0169834234,0.0809656233,0.0353382863,-0.421662271,0.0005130195,0.0699741542,-0.2821937501,0.1218045875,0.0930851251,-0.1535139382,0.0065672644,0.7602151632,0.5507843494,0.1212182343,0.2375285178,-0.0034825553,-0.1811911464,-0.068433024,0.1376174688,-0.0937062427,-0.3521999717,0.220836997,-0.2030648142,0.3683298528,0.060559649,0.1236680076,-0.0481191352,0.0247149989,-0.2025212795,-0.3216696382,-0.0490883924,-0.0197930913,0.220911473,0.3211824298,0.1006778404,0.2827619612,0.0158554148,0.240595296,-0.1552976221,0.2196280956,0.1346051991,0.2111643404,-0.2921218574,-0.0464425012,0.053991951,0.061860498,0.1196215898,0.3494888544,0.2542255819,-0.1298364699,-0.1527456194,0.1997391433,0.1476189047,0.1348201782,-0.3525344729,-0.1841500551,-0.0244389568,0.0541868396,0.1615141332,0.1796976477,0.1138684973,-0.0614135377,0.5707734823,-0.3047700524,0.0621442534,-0.0682387874,-0.2497998029,-0.0991846845,0.1489132345,0.0242047887,-0.0765939429,0.0515944809,-0.5694958568,-0.218003884,-0.0996783674,-0.2897157669,-0.0104425345,-0.3556214869,0.1584205478,0.0435612649,0.4095528722,0.553640306,-0.1880052388,-0.1793334335,-0.1727851927,0.0221511368,-0.3302715123,-0.4186816514,-0.3138356507,0.0359060541,-0.0420401879,0.1558498144,-0.1006062999,-0.6175259948,0.1054020524,-0.1401911676,0.1677859873,-0.4618127942,0.0602735132,0.3355374336,-0.1266040653,-0.2156083286,0.0406896733,0.1781651229,-0.380148381,0.5589037538,-0.0080079352,-0.012836419,-0.3958186805,-0.1521437168,-0.0595990755,-0.4355050921,-0.0179957151,0.4624483585,-0.452793777,0.1983210295,0.1278136224,0.3341597617,0.0192638151,0.127006337,-0.1278090477,-0.316888392,-0.2479605973,-0.0944275409,0.0487670116,-0.3186523318,-0.1704467088,0.0205285735,-0.0872933269,0.0470420569,0.1814609468,0.2262983918,-0.213588357,-0.7137227058,-0.1699905097,0.2464029491,0.0098077953,-0.1441752315,0.259793967,0.3376735151,0.0757399127,0.3888047636,0.1608972102,0.1979900002,0.0123933069,0.079266645,0.0564038865,-0.2088776231,-0.0358273722,0.1481038332,0.1363360137,0.3704130352,-0.1180000231,0.1081165224,0.0058085541,-0.0186300855,0.3974691331,0.4048883021,0.0921251774,-0.5521720648,0.3745788038,-0.1504852772,0.0728532597,0.0423581079,0.0088187531,-0.3597369492,0.1449587196,0.1574673504,0.1463382542,-0.0299891308,-0.1593993753,0.018570086,0.04124064,0.0461306348,0.0567121767,0.1425176114,-0.2140637338,-0.1779855043,-0.5941957831,0.2989788353,-0.0782699138,0.0659886003,0.3159153461,0.3078939915,-0.189925015,0.015150195,0.2904504836,0.2114409506,-0.4122418761,0.6659331918,-0.1820702106,-0.4695115983,-0.1968198121,0.1743108183,-0.1990711987,-0.2279466689,0.3029984832,-0.1717747003,-0.0002793286,-0.1049668938,-0.0252821427,-0.0720260292,0.2797107399,0.0113895871,-0.0666734874,0.3483418524,0.1175403148,0.2047601491,-0.1360943764,-0.2600983381,-0.2059318721,0.4848166108,0.4665034711,0.1646368206,0.1064429134,-0.4742978811,0.0097053945,0.0539317578,-0.0274770968,-0.0341002457,-0.2168218493,-0.0779031515,0.01067919,0.4529963732,-0.1002113298,0.3377586305,-0.0457667895,0.2182195187,-0.1701074839,-0.5406842232,0.4682012796,-0.5862604976,-0.2056438476,-0.2914238572,0.3180342615,-0.1106252298,0.0745078176,-0.4245930016,0.0193992648,0.4931824505,-0.1373592168,0.0267345905,0.4661881924,-0.1348495781,0.256216228,-0.218927145,0.2520153224,0.3302054107,-0.1999833435,0.2350977659,-0.3925742507]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2475","title":"Issue in timit_asr database","comments":"Indeed was a fixed bug.\r\nWorks on version 1.8\r\nThanks ","body":"## Describe the bug\r\nI am trying to load the timit_asr dataset however only the first record is shown (duplicated over all the rows).\r\nI am using the next code line\r\ndataset = load_dataset(\u201ctimit_asr\u201d, split=\u201ctest\u201d).shuffle().select(range(10))\r\n\r\nThe above code result with the same sentence duplicated ten times.\r\nIt also happens when I use the dataset viewer at Streamlit .\r\n\r\n## Steps to reproduce the bug\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\u201ctimit_asr\u201d, split=\u201ctest\u201d).shuffle().select(range(10))\r\ndata = dataset.to_pandas()\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\ntable with different row information\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.4.1 (also occur in the latest version)\r\n- Platform: Linux-4.15.0-143-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.8.1+cu102 (False)\r\n- Tensorflow version (GPU?): 1.15.3 (False)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\r\n- `datasets` version:\r\n- Platform:\r\n- Python version:\r\n- PyArrow version:\r\n","comment_length":10,"text":"Issue in timit_asr database\n## Describe the bug\r\nI am trying to load the timit_asr dataset however only the first record is shown (duplicated over all the rows).\r\nI am using the next code line\r\ndataset = load_dataset(\u201ctimit_asr\u201d, split=\u201ctest\u201d).shuffle().select(range(10))\r\n\r\nThe above code result with the same sentence duplicated ten times.\r\nIt also happens when I use the dataset viewer at Streamlit .\r\n\r\n## Steps to reproduce the bug\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\u201ctimit_asr\u201d, split=\u201ctest\u201d).shuffle().select(range(10))\r\ndata = dataset.to_pandas()\r\n\r\n# Sample code to reproduce the bug\r\n```\r\n\r\n## Expected results\r\ntable with different row information\r\n\r\n## Actual results\r\nSpecify the actual results or traceback.\r\n\r\n## Environment info\r\n\r\n\r\n- `datasets` version: 1.4.1 (also occur in the latest version)\r\n- Platform: Linux-4.15.0-143-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.8.1+cu102 (False)\r\n- Tensorflow version (GPU?): 1.15.3 (False)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\r\n- `datasets` version:\r\n- Platform:\r\n- Python version:\r\n- PyArrow version:\r\n\nIndeed was a fixed bug.\r\nWorks on version 1.8\r\nThanks ","embeddings":[0.0056248805,-0.3979411721,0.0337808244,0.5559312105,0.2933141589,0.1347243488,0.3898530602,0.2464324534,-0.3128981292,0.1001804546,-0.1041535065,0.4860834777,0.0373350121,0.2365354896,0.3352548182,0.1193738654,0.0827093422,0.2177102715,-0.2566957772,-0.2044742554,-0.0797199532,0.0188646987,-0.297102958,-0.0470760055,-0.3123938441,0.3580350578,-0.1445847452,0.0711581782,0.0002215193,-0.214114517,0.0407694057,0.2142827511,0.1101519018,0.4459582567,-0.0001224793,-0.0474875905,0.1529335529,0.0295090601,-0.3462762833,-0.4951920509,-0.246620506,0.1636790335,0.4268999398,-0.0129712448,-0.2035852969,-0.2229595482,0.003545946,0.0058256909,0.0597612001,-0.0599806458,0.129264459,-0.0803964585,-0.2996327281,0.0033442567,0.1008916274,0.0247674137,0.1526795477,0.357227236,-0.0027115087,0.1435573995,0.0702766031,0.1611882001,0.0175218619,0.4145493507,0.1682370007,0.0819296315,0.0866658613,-0.13381432,0.1990442872,0.3332818151,0.6418062449,-0.2891809046,-0.1146525294,-0.1680236012,0.2774809897,-0.117734313,0.1097050682,0.142147705,-0.1723692715,0.0439909324,-0.1222973615,0.15301615,-0.030113399,0.2070792168,-0.1209242046,-0.185602501,-0.0517291091,0.2268612236,0.1994020939,-0.0515241921,0.558006525,0.0837549716,0.0768629387,0.1400007755,-0.6882770061,0.1654391885,0.06778384,-0.0292578712,0.0879018977,0.2199456841,0.7684150934,0.2363758236,0.0202989113,0.0555991642,0.4605643153,0.0540962443,-0.0620750263,0.1312758923,0.0379954353,0.0840284303,-0.2297661453,0.0244856142,-0.1405501813,0.2263147235,0.2301293164,-0.3008561432,0.2543634474,-0.2841982245,-0.3891957402,-0.1016779616,-0.283615768,-0.208096534,0.002211187,0.2331704348,0.1544072181,0.3274967968,0.028295394,0.1405595392,-0.1271810383,-0.1198238358,-0.2591748536,-0.1094121858,-0.0000325923,0.0290385503,0.1760293394,-0.740131557,0.1181929931,0.2749448121,0.0749574527,-0.381275028,0.1251406819,-0.1424267739,-0.042401664,0.2826627195,0.2753904164,0.2366274744,0.022529982,-0.1216866374,0.1555004567,0.422919631,0.0302761141,-0.0699343234,0.1188010275,0.073583506,-0.104587853,0.0908431634,-0.4983101189,0.2783350945,-0.0346191712,-0.1572910547,-0.2161005437,-0.0704079419,-0.3181914687,-0.2330867052,0.2046703696,0.4110208452,-0.3068484962,-0.0658723116,0.0907928124,0.0328801721,0.2894310951,0.0634195283,-0.2533632219,0.2217700928,-0.3355354667,-0.1963275224,0.2616679072,0.0153773092,-0.4468181431,0.0915927514,-0.2389507294,0.2330971509,0.1542991847,0.2254262567,0.35830158,0.2396834195,0.2301186025,0.1379088759,0.0978865102,-0.0293608457,-0.3745519519,0.120479241,0.3184940219,0.3362240195,-0.1927578896,0.477773726,0.137348935,0.145187214,0.4923008978,0.0871254429,-0.1551188976,-0.0668878406,-0.12127316,0.0826407969,0.2386507541,-0.2514390945,-0.0993490741,0.114706181,0.0681049377,-0.170716688,0.1511778831,-0.0404325053,-0.1382754594,-0.1682498455,-0.4172431529,0.2265386432,-0.0035324872,0.473484993,-0.2720677555,-0.0014211884,0.0130565176,0.3815121055,-0.2139223069,-0.0981519371,-0.4101258218,0.307096988,0.0192479081,0.0714671686,0.1961863637,0.0272502564,0.2588761449,-0.0087540317,-0.0929293633,0.2967071533,-0.054640308,0.0913163647,-0.226056546,-0.3229327202,0.2042343318,-0.1737438589,-0.1436575949,0.1335594803,0.2320598215,-0.2432814986,-0.1841963828,0.0994659737,-0.4357467592,-0.0665915534,0.0565569438,-0.0264220089,0.2216915637,0.2030881792,0.0536614992,-0.2630222738,0.2832905352,0.0822819918,-0.2153942585,0.058404237,-0.3620329201,0.3088451922,0.0341832452,-0.1205177084,0.0081544016,0.056733463,-0.144717887,0.0390819423,-0.0215909984,0.1334864348,0.3313490748,0.1416812837,0.1554867476,-0.0271240417,0.0102011645,-0.0335363299,0.0336997136,0.0647844449,-0.2325026244,0.3498497605,0.154527992,-0.1829378903,-0.4008693099,0.1428057402,0.448492527,0.4870320857,-0.3134055734,-0.1067330912,-0.4975292683,-0.1816777885,-0.1235709414,-0.1957681924,0.1507009268,-0.0991004407,0.0422927476,0.5853568912,0.1762648821,0.1341422498,-0.0573368967,0.1550188512,-0.0353208669,0.4142862856,-0.0839059427,0.2835778296,-0.3509893119,0.0144113461,0.152474001,0.3987771273,0.3484049141,-0.1282776892,0.087584354,-0.3280181885,-0.1300551593,0.0358773507,-0.0117633268,0.3443444669,-0.109234646,0.2113010436,-0.4252711236,-0.0963448957,0.1727606952,-0.0637626052,-0.0849625394,0.2630160153,-0.0053264373,0.1532988548,-0.0644939691,-0.3552166522,-0.0440656468,-0.2014307678,0.1771762818,0.0260813572,0.1413502842,-0.0722728372,-0.0699294209,-0.0755577311,0.1194212511,0.0522054136,-0.5218003392,-0.3417723477,0.072943233,0.1080179736,-0.2779786587,-0.2234183699,0.0092066657,-0.0063446574,0.1333943456,-0.4221113026,0.1474864632,0.1085519493,-0.095304966,-0.2865752876,0.0001677668,0.3072421849,-0.0224984773,0.0242249835,-0.0769679248,-0.0889206305,-0.1076639816,-0.0596360564,0.1462442875,0.1253357828,0.420163542,0.0264854208,0.2295158654,0.3870595098,0.1501708627,0.2523775697,-0.2190306634,0.1960753351,-0.1671267599,-0.3774962127,-0.2324671298,-0.1250198036,0.1386948824,0.3840281069,-0.0723783597,0.0642071217,-0.1786606461,0.0967954397,-0.3309627771,-0.2039111108,-0.0877790749,-0.392636627,0.0768379718,0.2955745161,0.1079307795,0.1721527278,-0.207291767,-0.2598460913,0.1028024256,0.0940812901,-0.0929654688,-0.2749293447,-0.3916316926,-0.2400295585,0.0475861058,0.2419138104,0.5383493304,0.0906794816,-0.0694246143,0.445898056,0.2227627635,0.9132775664,-0.234008953,0.1169735715,0.1083235964,-0.0686630458,-0.3096210361,-0.0269935802,-0.125052318,0.1390567422,-0.1083654389,-0.0693501383,-0.1315711588,0.0905326754,0.1867608875,0.2976565361,-0.329041332,-0.340821296,-0.3363567591,-0.1296082288,-0.1427940428,-0.1129725948,-0.1646885425,0.0812120885,-0.2870581746,-0.1326008886,-0.2494586557,-0.0302412566,0.2414226532,0.2248560041,0.543974936,0.2204795629,0.6248975992,0.5942409039,-0.2139112651,0.176674515,0.723412931,-0.0645542145,-0.1724704653,0.2533642054,-0.4609920979,-0.0515498295,0.365365237,-0.2037724257,0.204062134,-0.0877030119,-0.1338185519,0.0212760568,-0.1053291708,0.1483434141,-0.1132354066,-0.2689570785,-0.6262937784,0.3378186822,-0.0512641147,-0.2142157853,0.4733335376,-0.3206619918,0.0620097369,0.2215447724,-0.5245487094,0.9491752386,-0.2505227923,0.2283992916,-0.1212490276,-0.164080292,-0.0237533487,-0.0131789837,0.1366136372,-0.4178075194,-0.6421229839,-0.0757755861,-0.0906335339,-0.2977395356,0.2902068496,-0.0793177709,0.1573922783,-0.1070460156,0.1466048956,0.191525057,0.112691775,0.105482012,0.1157002077,-0.0420036539,-0.0620303713,-0.1623238474,-0.0222715847,0.2684696615,0.1056092903,0.3089279234,-0.2498238087,-0.2534947693,0.0438385643,-0.2113337815,0.1936832219,-0.3087241948,-0.4784846008,-0.179499507,-0.0396261886,-0.3079868257,0.0326886438,-0.0032182692,-0.0863492861,0.4925431013,0.4375332594,-0.3467575312,0.0352037027,0.2302591205,0.1424430311,-0.3850836754,0.133814767,0.321721077,-0.1499120295,-0.4800188243,0.3704206347,0.0694934949,-0.308418721,0.0049526473,0.0905151591,0.0206801556,-0.4158825576,0.0030740395,0.0716441721,-0.3006033003,0.1185243428,0.1156229749,-0.1557547152,0.0105239125,0.7731684446,0.5564835668,0.1333249658,0.2439105213,-0.0032624588,-0.179963246,-0.0702166185,0.134768784,-0.093496114,-0.3545819521,0.222753197,-0.1999768913,0.3705841005,0.0504613258,0.1280688792,-0.0443621501,0.0202720053,-0.2004776299,-0.3241287172,-0.0462814979,-0.0047239205,0.224131912,0.3120579123,0.0859718472,0.2944168448,0.0173116289,0.2398969978,-0.1565557718,0.222439751,0.1397283673,0.2110426873,-0.2878567278,-0.0403467901,0.0541873425,0.0631238595,0.1221518517,0.3560796082,0.2563143671,-0.1314398199,-0.14200899,0.2000613213,0.1501682699,0.1345308125,-0.3552037776,-0.1879475862,-0.0285666473,0.0575053096,0.1622294188,0.1676051021,0.119517386,-0.0632236004,0.5764709711,-0.3063498139,0.0658529475,-0.0651747286,-0.2414977252,-0.0901727602,0.157620132,0.0192030165,-0.0792006701,0.0529109649,-0.5621141791,-0.2215974182,-0.1088602841,-0.2874333262,-0.0018456341,-0.3643865585,0.1651740372,0.034900099,0.4065042138,0.5556183457,-0.1953155994,-0.1729483157,-0.1821032763,0.0235401187,-0.3373126388,-0.4109774232,-0.3154740334,0.0299954731,-0.0427036881,0.166012764,-0.0994182527,-0.6110933423,0.1038360298,-0.1351600736,0.1702670455,-0.468490541,0.0492678098,0.3149185479,-0.1307144761,-0.2090574801,0.0342115685,0.1828796566,-0.3888220489,0.5513429046,-0.0107448259,-0.0167364571,-0.397616148,-0.1456192732,-0.0676324517,-0.4244382381,-0.0162771288,0.4634147882,-0.4529018998,0.1920525879,0.1221838519,0.3261225522,0.0294750761,0.1237580255,-0.1260399967,-0.3066849709,-0.2453119457,-0.0878130123,0.0457898267,-0.3259349465,-0.1767864972,0.0119527383,-0.0865270644,0.0637005195,0.173801139,0.2206592262,-0.209751308,-0.7064387202,-0.1655074358,0.2415878177,0.0156680979,-0.1420388222,0.2794003189,0.3213681281,0.0841900334,0.398635745,0.1638099402,0.1864478588,0.0252208356,0.0809300914,0.0569250211,-0.2127871662,-0.0426319689,0.1548784077,0.1462453157,0.3827766776,-0.114030771,0.0991511717,0.0084686773,-0.0168145299,0.399700433,0.4221745431,0.0932059214,-0.5563765764,0.3720404506,-0.15409863,0.06814605,0.0260499064,0.0002239503,-0.3679634631,0.1464949995,0.1525691897,0.146621123,-0.0231285449,-0.1575547755,0.0190517623,0.0410829373,0.0415199175,0.0407718457,0.1390342265,-0.2147184759,-0.1753223389,-0.5892380476,0.3083989322,-0.0758737773,0.0566560067,0.3058187962,0.3119958639,-0.1753121018,0.0086712018,0.3102050126,0.1992397159,-0.4232991338,0.6652490497,-0.1851636767,-0.4708698392,-0.2072743177,0.2004029602,-0.1980255097,-0.223119393,0.3047703505,-0.16233325,0.0016129619,-0.1069568843,-0.030095581,-0.0782811716,0.2886677384,0.0168594792,-0.0747106299,0.3533467948,0.1192290336,0.2086034566,-0.1297353059,-0.2707828581,-0.2026435584,0.4907976985,0.478630662,0.1456923783,0.109708488,-0.486944437,-0.0005943596,0.0499240831,-0.0317292586,-0.0604245998,-0.2081461847,-0.0890185535,0.0254474264,0.4460380375,-0.1064979732,0.3449551165,-0.0572518706,0.2174514234,-0.1696603745,-0.5218992233,0.470168978,-0.5788564086,-0.1983889043,-0.2767700851,0.3205152452,-0.1122590527,0.0674582794,-0.4219201803,0.0168235358,0.4931388497,-0.13176018,0.0359037966,0.4647358954,-0.1348753124,0.2561511993,-0.2153066844,0.2500782609,0.3206747472,-0.1920317411,0.2244355977,-0.3855502903]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2474","title":"cache_dir parameter for load_from_disk ?","comments":"Hi ! `load_from_disk` doesn't move the data. If you specify a local path to your mounted drive, then the dataset is going to be loaded directly from the arrow file in this directory. The cache files that result from `map` operations are also stored in the same directory by default.\r\n\r\nHowever note than writing data to your google drive actually fills the VM's disk (see https:\/\/github.com\/huggingface\/datasets\/issues\/643)\r\n\r\nGiven that, I don't think that changing the cache directory changes anything.\r\n\r\nLet me know what you think","body":"**Is your feature request related to a problem? Please describe.**\r\nWhen using Google Colab big datasets can be an issue, as they won't fit on the VM's disk. Therefore mounting google drive could be a possible solution. Unfortunatly when loading my own dataset by using the _load_from_disk_ function, the data gets cached to the VM's disk:\r\n\r\n`\r\nfrom datasets import load_from_disk\r\n\r\nmyPreprocessedData = load_from_disk(\"\/content\/gdrive\/MyDrive\/ASR_data\/myPreprocessedData\")\r\n\r\n`\r\nI know that chaching on google drive could slow down learning. But at least it would run.\r\n\r\n**Describe the solution you'd like**\r\nAdd cache_Dir parameter to the load_from_disk function.\r\n\r\n**Describe alternatives you've considered**\r\nIt looks like you could write a custom loading script for the load_dataset function. But this seems to be much too complex for my use case. Is there perhaps a template here that uses the load_from_disk function?\r\n","comment_length":84,"text":"cache_dir parameter for load_from_disk ?\n**Is your feature request related to a problem? Please describe.**\r\nWhen using Google Colab big datasets can be an issue, as they won't fit on the VM's disk. Therefore mounting google drive could be a possible solution. Unfortunatly when loading my own dataset by using the _load_from_disk_ function, the data gets cached to the VM's disk:\r\n\r\n`\r\nfrom datasets import load_from_disk\r\n\r\nmyPreprocessedData = load_from_disk(\"\/content\/gdrive\/MyDrive\/ASR_data\/myPreprocessedData\")\r\n\r\n`\r\nI know that chaching on google drive could slow down learning. But at least it would run.\r\n\r\n**Describe the solution you'd like**\r\nAdd cache_Dir parameter to the load_from_disk function.\r\n\r\n**Describe alternatives you've considered**\r\nIt looks like you could write a custom loading script for the load_dataset function. But this seems to be much too complex for my use case. Is there perhaps a template here that uses the load_from_disk function?\r\n\nHi ! `load_from_disk` doesn't move the data. If you specify a local path to your mounted drive, then the dataset is going to be loaded directly from the arrow file in this directory. The cache files that result from `map` operations are also stored in the same directory by default.\r\n\r\nHowever note than writing data to your google drive actually fills the VM's disk (see https:\/\/github.com\/huggingface\/datasets\/issues\/643)\r\n\r\nGiven that, I don't think that changing the cache directory changes anything.\r\n\r\nLet me know what you think","embeddings":[-0.2717247307,-0.12313582,-0.0702048838,0.1266331822,0.2155744135,-0.0214382932,0.2847643197,0.084847644,0.4296782315,0.3610470295,-0.1760807931,0.2474943846,-0.2383664548,0.2422399968,0.270316273,0.0099474564,0.0024952546,-0.0644657016,-0.0686381236,-0.1333422959,-0.1486033052,0.1585291773,0.0208627786,-0.2857365906,-0.0282743629,-0.1897913963,-0.0265019555,0.2757652402,0.3105606735,-0.3239333332,0.1548090726,0.1240978912,0.157219097,0.5559691191,-0.0001047829,-0.0235921051,-0.0222179685,-0.2570990622,-0.3680811822,-0.1126931608,-0.0655640736,-0.1430470794,0.0789605454,-0.0596486479,-0.4904256761,0.0890217945,-0.0152680352,-0.2515120208,0.1891818643,0.1155601069,0.2483702898,-0.0900903121,-0.3790552914,0.0660773069,0.0514307208,0.0461631231,-0.1904249787,0.5480613112,0.0846348852,0.1086061001,-0.145517841,0.4331534207,-0.0582385734,0.1626028866,0.530264616,0.3962347507,-0.0634314194,-0.1762126684,0.0431517214,-0.1202843711,0.5056242943,-0.2109804749,-0.2322986871,-0.2513057888,0.0228340011,-0.3922423124,0.384424001,0.1623055339,-0.1985076517,0.3777803779,-0.2268848866,-0.3963228464,-0.1039247438,0.0135895964,-0.0000064121,-0.1683011651,-0.2351841927,-0.0759985968,0.1762790829,0.1449361444,0.206283614,-0.0536264814,0.0520501025,0.3046586514,-0.0917593911,-0.1389120519,-0.0464858897,0.3314088881,0.0078661777,0.0147841228,0.4713613689,0.0797277391,-0.2129879445,0.1816708297,-0.067667596,0.3930065036,-0.2098642737,-0.0274156965,0.4752400815,0.1682695001,-0.2131383717,-0.0970623493,0.1519378722,-0.0439626947,0.2183887661,0.140013501,0.0364623815,-0.1857560575,0.2886313796,-0.2089163065,0.1319718659,0.0694612339,0.28249681,0.4558331966,-0.0666587055,-0.1313296705,0.1208478436,-0.2122956663,-0.1652946323,0.1883939356,-0.109368667,0.0455158725,-0.2290371507,0.489289701,0.2463846803,0.0902864486,0.2731122077,-0.2993309796,0.2071622461,0.0754927173,0.0755357817,-0.1780983359,0.3490899801,0.3921796679,-0.0062888535,0.0368461236,-0.0510139763,-0.0927096233,-0.363440901,0.2184134424,-0.4228341281,-0.4356271327,0.4194563925,0.1934923977,-0.185685724,0.0723155886,-0.5352765322,0.3024528325,-0.1795238107,-0.1689608097,-0.0451574512,0.0595604405,-0.5309442282,-0.3391389847,0.1563928425,0.2845223844,-0.1855960786,-0.1120542958,0.1516869664,0.1550040096,-0.137589395,0.3683568239,-0.2482610047,0.0201292392,-0.2403443307,-0.1525033116,0.200971067,-0.3625818491,-0.605230093,0.1620093286,-0.1262727827,-0.0310018901,-0.0377450697,0.1869264394,0.1384861767,-0.1977033615,0.3591814935,0.6201862693,0.0559349991,0.3013579547,-0.3125096262,-0.2830684781,-0.3325062394,0.2417350858,-0.0909443945,0.1030955315,0.1306126416,-0.2827852666,0.0508498624,-0.1940552145,0.0488463007,0.3934384882,0.3601915538,-0.2557527125,-0.0082358364,0.1270047426,-0.652000308,0.3312801719,-0.0341888703,-0.0623960271,-0.0443273447,-0.1631168574,0.1890245229,-0.1728112698,-0.0432419144,-0.0107518909,0.0655686334,0.0757776126,0.2253200263,-0.1849762201,-0.0131545244,0.328877598,0.0081908507,-0.067329213,0.1577627063,0.2264792174,0.0856376439,-0.1127577201,-0.1247871369,-0.0879629105,0.0816989914,-0.1688310802,-0.2173811197,0.3881791234,0.1821926832,0.4562583864,-0.0831858665,0.255866617,0.30075109,-0.0425577536,0.4923293293,0.0063057984,0.0130656706,0.1124818847,-0.3570446372,0.3859908283,-0.1907472908,0.1745758206,0.2347824723,-0.1141711324,0.19006975,-0.1285559833,0.1645834446,-0.1481104046,0.1359440982,0.3211123645,0.3957827389,0.1460084617,-0.1996143311,0.2674739063,0.3323667645,0.0968802571,0.1031925455,0.246134609,-0.3691287637,-0.2252021283,0.1397533119,0.197507143,0.5396437049,0.3424683511,0.3131107688,0.1506482363,0.1247832999,-0.0937287584,0.2106864601,-0.1117323339,0.1853049695,0.0762357116,0.190817982,-0.1487636566,-0.2165201157,0.0760200843,-0.0008258178,0.0549027845,-0.1907076687,0.143323943,0.0854379758,0.0519407131,-0.165815711,-0.0339375921,-0.076560393,0.0708651394,0.0929717124,0.4913678169,-0.1250120252,0.071845524,0.2910555005,0.393671155,-0.1227563918,-0.3675518036,-0.0487354584,-0.0331536904,-0.0719597563,0.0459643975,0.1463933885,-0.0009631031,0.2091947794,0.1547740102,-0.0044358699,-0.3165817857,-0.1744364053,0.1876914352,-0.0269155707,0.524433732,-0.057649862,0.0138232242,-0.0310174618,0.0415321067,-0.0137339924,-0.455465436,-0.1072462574,-0.3994989097,0.1093487367,0.2656004727,-0.3197873831,-0.0110392747,-0.3589963615,-0.5685352683,0.1637336612,0.2817828059,0.2383924723,0.265165031,0.0898809657,-0.0618730783,0.572666347,-0.0546644367,-0.3195027709,-0.3880082369,0.285887599,-0.3585863113,-0.0461908393,0.0663327053,-0.0486336984,0.0415387303,0.2487963736,-0.3152450025,-0.5569884777,-0.0941396281,0.0763135478,-0.1239234954,-0.0483976789,0.3200336695,-0.0705713704,-0.1355658621,-0.0175185837,-0.4911369383,0.0440961309,0.4789563715,0.0262698568,0.1413140446,0.0510880016,-0.0411006957,0.6494165659,-0.3143565357,0.0779204816,0.2143071592,0.2396313101,0.3123824894,-0.266377151,-0.1168241724,0.0004895034,0.0033940761,-0.5504609346,0.3372140825,0.1995887458,-0.3394126296,-0.119218193,-0.381487757,0.0521340519,-0.2729561329,0.3701463938,0.2252449989,0.2053884268,-0.0108913109,0.0168045107,-0.3317088187,-0.1100648195,0.1290931255,0.20593068,0.1658010781,-0.0393399149,-0.106875658,0.0265217032,-0.6030474901,0.3651073575,0.0749172792,-0.4362511635,-0.1904774904,-0.0978045985,0.102555424,-0.0237314217,0.4069846869,-0.3171295524,-0.0407178774,0.0126508465,-0.2084210813,0.0999234319,-0.0199754126,0.0917777196,0.1346763819,-0.1953123659,0.2986092865,-0.0235233381,-0.3415671885,0.0834692344,0.185412541,-0.3344799578,-0.0131441904,-0.1658171117,-0.3321719766,-0.24766846,0.0475294776,-0.0208778884,-0.174914211,0.1017635763,0.0514636561,0.0490510352,-0.0931758955,0.0320487693,0.0503638163,0.4364527762,0.3477067947,-0.0359022059,0.3418847322,0.1838711053,0.1447160244,0.3590496778,-0.313193053,0.1504933238,0.0009914561,-0.0626287535,0.1716903746,0.5609202981,-0.0305366889,0.0166082811,-0.0469164439,0.0513989069,-0.3236756921,0.3021384776,-0.275093019,0.1146375835,-0.0554490127,-0.4245933294,0.5308994055,0.0290660243,-0.2997180223,0.1367155612,-0.0001854716,-0.7025890946,0.4885977209,0.3361812234,0.9242456555,-0.2402123511,0.3218453228,0.0788633451,0.1488070041,0.1762288958,-0.1550417244,0.2432227731,-0.182414636,-0.4360659719,-0.1167536527,-0.1882802397,-0.1224959418,0.010844009,-0.3719363809,0.336663276,0.0038930504,0.0098652337,-0.1484054625,0.3056336641,0.0565599091,-0.5729461312,-0.486371845,0.1915958077,-0.1588524729,0.2162870765,-0.0208585765,-0.052224461,0.0078266393,-0.0142132761,-0.0103387926,-0.1447440386,-0.6259279251,0.182462424,-0.4253158569,-0.399818629,-0.1184005961,0.0155104436,0.2847126424,0.008999981,-0.1155999824,0.2833639681,-0.1056761593,0.2330959886,-0.2701417208,-0.2482078373,0.3209016621,0.0013770269,0.1363827288,-0.048031088,-0.1237723976,-0.5901224613,0.060321521,-0.0143850734,0.03996934,-0.0196123794,-0.2681331038,0.007938616,-0.1973916739,-0.0584130399,0.1618923247,0.136601463,0.0394565649,0.1673317105,0.2108816355,-0.2371620983,-0.1011994407,0.1155039668,0.2318777591,-0.1932785511,0.4154466391,-0.0743582398,-0.0843025893,-0.2334057987,-0.2085594535,0.0307474043,0.0738214031,0.4039310515,0.0249863304,-0.0552656874,-0.0081918985,0.2261240482,0.1802875549,0.0972425193,-0.4014541209,-0.1373057067,-0.3850552738,0.2895958722,-0.1665962785,0.2592766285,0.0285679922,-0.1688112617,0.0164901875,0.0509675592,-0.4256032705,0.1830048114,-0.1100650281,-0.0803719088,0.3154861927,-0.0769215897,0.4419022202,-0.0866327584,0.0802168697,-0.198989585,-0.2595523894,-0.2575899661,-0.2302679121,0.0908226967,-0.1220994443,0.1583171785,-0.1519140601,-0.1913686842,0.1481453329,-0.3160799742,0.2773513496,-0.0007620081,-0.0254210904,-0.2146071941,-0.2145930678,-0.0026448742,-0.0418053381,0.1807190329,-0.0977541357,0.1765023321,0.0674680769,0.0306346789,-0.2144226581,-0.0808250085,0.1843414605,0.0756340921,-0.0544143394,0.0382650979,0.1736978292,0.1568242013,-0.0733184218,0.1325683892,0.0315196216,-0.2181302607,-0.0888698772,-0.0244052391,0.088026844,0.2289022803,0.0511965156,-0.1481765062,-0.000033754,-0.5024091601,0.2911372781,0.4773529172,0.2948723733,0.4668112993,0.1188886762,0.1677166075,0.4517068565,0.0219305716,-0.1184431911,-0.1356632262,0.1526888907,0.2881663442,0.2948812842,0.2481931001,0.336692065,0.2198079079,-0.2805719674,0.0388726778,-0.0968112573,-0.0160589237,-0.0543689951,-0.479113549,0.1115250215,0.1625468135,-0.0872814134,0.3022709787,-0.0425254107,0.0294629093,-0.4516474307,-0.241702646,-0.126077041,0.1258672476,-0.2133701444,-0.1083880365,0.1261664033,0.0069003273,-0.0262637511,0.3184457719,0.0180552974,-0.2803692222,0.2218081951,0.0409236886,0.1581664681,0.2806837559,-0.0110171195,0.2183282971,0.0533361249,-0.0750525296,0.1689574569,-0.3555186391,-0.0037468616,-0.0661052689,-0.064914234,0.0138751967,0.2872537076,-0.2753952146,-0.0253908429,0.0737068802,-0.0202164017,-0.2466812134,-0.1995472461,0.0274628624,0.0292459559,0.059801694,0.1995851398,-0.163457185,0.1789294928,-0.0769739971,0.1154756472,-0.048441153,-0.0661407933,0.3410575092,-0.0693211406,-0.268614769,-0.0373322032,-0.1331006587,0.0084883645,0.1698346734,-0.2074604332,0.0692835599,0.24103342,0.1047196016,0.2239606977,0.6690722704,0.0915638134,0.1458970308,-0.5181546211,0.0084504616,-0.4568884373,-0.0725777522,0.1638171077,0.166103363,-0.2737773359,0.3370563388,0.1787356734,0.1622143537,-0.0216417015,-0.008584627,0.0547087081,-0.1880150735,-0.2519069016,-0.000444837,-0.0424700901,0.0905952752,0.1160649285,-0.4988151789,-0.0402761512,0.0875591785,0.0460802987,-0.0843941197,0.1787142605,-0.0984602198,0.1627934873,0.6224085689,-0.2340026051,0.0567222238,-0.0371387638,-0.1572415829,-0.0879911408,-0.1784507781,-0.2318412215,-0.1272525489,0.0841510072,0.511174798,-0.1256454736,-0.0874004662,-0.3893437684,0.1874754131,0.2763238251,-0.0712516457,0.1412660331,0.07366772,-0.0142611433,0.1381676495,0.2249012738,0.3117011786,-0.0824032202,-0.0772755891,-0.4371891916,0.0175977591,0.4397958219,-0.163270101,-0.1224297509,0.0889563262,0.3794546723,0.1667649448,-0.1413537264,-0.4729729593,0.0586285926,0.1840690076,-0.0497039966,-0.343053937,0.4234716892,-0.0261529088,-0.0265620202,-0.2129845023,0.1733535528,-0.0263956375,-0.3420034647,-0.5036293268,-0.2204976082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2474","title":"cache_dir parameter for load_from_disk ?","comments":"Thanks for your answer! I am a little surprised since I just want to read the dataset.\r\n\r\nAfter debugging a bit, I noticed that the VM\u2019s disk fills up when the tables (generator) are converted to a list:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/5ba149773d23369617563d752aca922081277ec2\/src\/datasets\/table.py#L850\r\n\r\nIf I try to iterate through the table\u2019s generator e.g.: \r\n\r\n`length = sum(1 for x in tables)`\r\n\r\nthe VM\u2019s disk fills up as well.\r\n\r\nI\u2019m running out of Ideas \ud83d\ude04 ","body":"**Is your feature request related to a problem? Please describe.**\r\nWhen using Google Colab big datasets can be an issue, as they won't fit on the VM's disk. Therefore mounting google drive could be a possible solution. Unfortunatly when loading my own dataset by using the _load_from_disk_ function, the data gets cached to the VM's disk:\r\n\r\n`\r\nfrom datasets import load_from_disk\r\n\r\nmyPreprocessedData = load_from_disk(\"\/content\/gdrive\/MyDrive\/ASR_data\/myPreprocessedData\")\r\n\r\n`\r\nI know that chaching on google drive could slow down learning. But at least it would run.\r\n\r\n**Describe the solution you'd like**\r\nAdd cache_Dir parameter to the load_from_disk function.\r\n\r\n**Describe alternatives you've considered**\r\nIt looks like you could write a custom loading script for the load_dataset function. But this seems to be much too complex for my use case. Is there perhaps a template here that uses the load_from_disk function?\r\n","comment_length":69,"text":"cache_dir parameter for load_from_disk ?\n**Is your feature request related to a problem? Please describe.**\r\nWhen using Google Colab big datasets can be an issue, as they won't fit on the VM's disk. Therefore mounting google drive could be a possible solution. Unfortunatly when loading my own dataset by using the _load_from_disk_ function, the data gets cached to the VM's disk:\r\n\r\n`\r\nfrom datasets import load_from_disk\r\n\r\nmyPreprocessedData = load_from_disk(\"\/content\/gdrive\/MyDrive\/ASR_data\/myPreprocessedData\")\r\n\r\n`\r\nI know that chaching on google drive could slow down learning. But at least it would run.\r\n\r\n**Describe the solution you'd like**\r\nAdd cache_Dir parameter to the load_from_disk function.\r\n\r\n**Describe alternatives you've considered**\r\nIt looks like you could write a custom loading script for the load_dataset function. But this seems to be much too complex for my use case. Is there perhaps a template here that uses the load_from_disk function?\r\n\nThanks for your answer! I am a little surprised since I just want to read the dataset.\r\n\r\nAfter debugging a bit, I noticed that the VM\u2019s disk fills up when the tables (generator) are converted to a list:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/5ba149773d23369617563d752aca922081277ec2\/src\/datasets\/table.py#L850\r\n\r\nIf I try to iterate through the table\u2019s generator e.g.: \r\n\r\n`length = sum(1 for x in tables)`\r\n\r\nthe VM\u2019s disk fills up as well.\r\n\r\nI\u2019m running out of Ideas \ud83d\ude04 ","embeddings":[-0.3292525113,-0.1360467821,-0.0741689503,0.3395154774,0.305447638,0.0242107809,0.1128007472,0.1102485061,0.3817079663,0.4735019803,-0.1144267693,0.330824703,-0.3873733878,0.2628678679,0.3451523781,0.1399837583,0.0198713578,0.0477789566,-0.0935734883,-0.1791060865,-0.2441439778,0.1887172014,0.0361785218,-0.3975585699,-0.01812369,-0.1753518283,0.0200118907,0.2108757496,0.1762578338,-0.253587693,0.0706496388,0.1289643645,0.209305346,0.5231045485,-0.0001096625,-0.0347109325,-0.0630265176,-0.2385136336,-0.4836073816,-0.1148203462,0.0183786023,-0.2147289366,0.1308309883,-0.0166224986,-0.4695326984,-0.012113004,-0.1294527352,-0.1720090359,0.2342174202,0.0874240249,0.2182984948,-0.1099984646,-0.3384350836,0.0545466691,0.2123763114,-0.0360593498,-0.2365245521,0.5095515251,0.2890924811,-0.0178813078,-0.1884749979,0.4902845025,-0.15122886,0.1314134002,0.6123000979,0.3670589626,-0.0228238888,-0.2394089997,-0.0000887081,-0.1050684378,0.5449563265,-0.1928210706,-0.1841847003,-0.29294312,0.0818782747,-0.4848721921,0.2964898646,0.2394314259,-0.3697342873,0.4065176249,-0.2246524096,-0.3107135594,-0.1870348901,-0.0529639572,0.0257946067,-0.2531559169,-0.2438654453,-0.0982120782,0.3537666798,0.2249075472,0.4196824729,-0.0605257526,0.030505633,0.2356748879,-0.2678915262,-0.1673245132,0.161732018,0.2779461145,0.1160210669,0.137620464,0.4320089221,0.0113153523,-0.1668656319,0.1139179766,0.0027673913,0.2854044139,-0.2825758159,0.0938473791,0.3335908651,0.1877832562,-0.1780281961,-0.0356745869,-0.0809276775,0.0728077814,0.3274492025,0.0160023961,-0.0252428614,-0.1408070028,0.0222679134,-0.0140229436,0.2142018527,0.2151568383,0.3565920591,0.5565623641,-0.1116352379,-0.025187511,0.1629080474,-0.232388854,-0.204443872,-0.0039903214,-0.1049953625,0.0050656698,-0.1695918739,0.3698783517,0.3263739645,0.0448296852,0.1683118641,-0.2533761859,0.2748964727,-0.0289751757,0.1201196685,-0.3316960037,0.3535512984,0.5145769119,0.1240806654,-0.0336503498,0.0273189042,-0.2320530117,-0.2865387499,0.4115089476,-0.2364871949,-0.4154908359,0.3976668715,0.1789437532,-0.1540034711,0.1263399124,-0.5711731911,0.2753769159,-0.1312900186,-0.0143278968,-0.2039010823,-0.005609049,-0.5529276729,-0.3339086473,0.1536078751,0.3464924395,-0.3733966053,0.0774056837,0.1239692867,0.1717839837,0.0171532035,0.4231011868,-0.2158162594,0.1241510138,-0.2311008275,-0.3277535141,0.0503481403,-0.0883581489,-0.7001388073,0.2320940942,-0.0248876773,0.1038093045,0.1604832411,0.3248530626,0.2106384039,-0.2432838827,0.5372142196,0.5482687354,-0.0456038006,0.2187206,-0.2700901926,-0.2495957762,-0.2289041877,0.2286724001,-0.1061992198,0.1680550426,-0.01130882,-0.2265481204,0.183961764,-0.2329793721,-0.0076500583,0.3606111109,0.3138047755,-0.2296572626,-0.0069621163,0.1488804668,-0.5796355009,0.2576674223,0.0531867109,-0.0545332134,-0.1754681617,-0.2799572349,0.0859792158,0.1150501817,-0.1081493348,0.119952701,0.0452320538,0.0314098634,0.1639125347,-0.362708807,-0.0575693771,0.192623049,-0.2284770906,-0.1036640778,0.0555240102,0.2995518148,0.1773662418,-0.1041514501,0.0284906961,-0.0537629016,0.0820358098,-0.0955642015,-0.0820117593,0.2479249984,0.351405412,0.3084276319,-0.1722155362,0.3389832675,0.225528121,0.0864558443,0.3739484847,-0.0398888402,0.1321095079,0.0006421394,-0.2703698575,0.3023077846,-0.2428352386,0.1529139578,0.1404814273,-0.1356091201,0.2908037603,-0.0365934595,0.1280058026,-0.1456817687,0.1292473525,0.4699782729,0.3004358113,0.1481821686,-0.2916773558,0.2725982666,0.3274208605,-0.001786654,0.1043403745,0.2034421712,-0.3379884958,-0.2950523794,0.1544416845,0.2103461772,0.4295648634,0.287474364,0.2392275631,0.0463792384,-0.0203874465,-0.1091998518,0.267097652,-0.0447409526,0.2852107584,0.1625962853,0.0705031008,-0.3323691785,-0.2747305036,0.1028756499,0.0722262785,0.1728269011,-0.2719561756,0.1694491357,0.0887609646,0.0731124356,-0.0902353004,-0.022102274,0.0151520558,0.0252637956,0.0940146223,0.6236653924,-0.0686962232,-0.0496789776,0.381652832,0.3992088139,-0.0191562548,-0.3055442572,-0.0406587161,-0.031010868,-0.0964486375,0.0532517843,0.2259763181,-0.1503489912,0.3060438037,0.0221153293,0.0237563495,-0.2317753434,-0.1140305549,0.2062560767,-0.0948144719,0.588311255,0.0014552168,0.1454019845,-0.0026164884,0.1788854599,-0.0646845996,-0.3798911273,-0.0150706535,-0.1384127438,0.0829962045,0.1341272593,-0.2145696133,-0.0644524843,-0.2786408663,-0.5284282565,0.0822836161,0.3449890316,0.3361671567,0.2942187786,0.2245372385,-0.0406703651,0.731526792,-0.0306201205,-0.3309214115,-0.2992551625,0.3457358778,-0.2213965207,-0.0686014816,0.2369613498,-0.0908062011,0.0824192837,0.2516969442,-0.5148746371,-0.4952053428,-0.0686283186,0.0294781271,-0.2048715353,-0.0371070243,0.1333857179,0.0059907623,-0.0629636869,-0.0594419353,-0.3281789422,-0.0978182778,0.2368490994,0.0939500704,0.1322534531,0.2417259365,-0.0481892824,0.5023215413,-0.3002935052,0.1572213918,0.1453654617,0.2208264023,0.3550994694,-0.2288267463,-0.1286663264,-0.2484922111,-0.0635367855,-0.6394159794,0.3165562749,0.1906860024,-0.2048769593,-0.0953036025,-0.2165760696,0.1227096543,-0.2248542756,0.4522868395,0.1798549742,0.1870837063,0.0748902857,0.0889941305,-0.3011831641,-0.1907238066,-0.0636866763,0.1096690595,0.2505927682,-0.1272366345,-0.0322670937,-0.0487916693,-0.6629843116,0.2970024943,0.0329043567,-0.3559454679,-0.0226086043,-0.0922589004,0.156696856,-0.043703679,0.6123079658,-0.249463588,-0.0468483157,0.0614045374,-0.2094807923,0.1058541536,0.1069702059,-0.041413907,0.108899653,-0.2605800927,0.2324753851,-0.0353755467,-0.2694305778,0.0681123808,0.4016906619,-0.3747314811,0.0555819757,-0.2616027892,-0.4519450068,-0.2791503072,0.0277672745,0.0763357952,-0.0479643792,-0.0766487569,-0.0470954068,0.0747934505,-0.1755212843,0.070823282,0.1097227857,0.5314676166,0.3468258381,0.1014156789,0.1709948927,0.0772883371,0.181295678,0.5112214684,-0.3182958663,0.0811011642,0.1411788613,-0.1457720399,0.074938044,0.5205647945,-0.1423425823,-0.0133470632,-0.2382354885,0.0977477357,-0.3676801622,0.1682332158,-0.2828285396,0.098656334,-0.0111881671,-0.7148752809,0.4752884209,-0.0168161765,-0.2711822093,0.1365141869,-0.113386713,-0.5768023133,0.5007026792,0.3116943836,1.0014719963,-0.2632242739,0.3655272722,0.1476274729,0.2187669277,0.1361572295,-0.3522646427,0.2528359592,-0.2399026304,-0.5551368594,-0.1111186892,-0.2290709913,-0.1229607612,0.1384823322,-0.3338915706,0.3431206346,-0.0643807054,-0.0451507084,-0.0736979321,0.2739697397,-0.0591371916,-0.464330107,-0.6345859766,0.1490139961,-0.1026690975,-0.0148170125,-0.1019532904,0.1241418123,0.0666725859,-0.0883292258,-0.0709175169,-0.114406608,-0.5100637674,0.1424904913,-0.3498207927,-0.4442533553,-0.1210781038,0.0020120549,0.0633082986,0.159178108,-0.1061998531,0.1952677816,-0.1250753403,0.2468688339,-0.3537704647,-0.1649007648,0.2606595755,0.0050330916,0.0952560976,0.1133154556,-0.11397513,-0.5635504723,-0.0026652517,0.0339859091,0.0527821966,-0.0883004293,-0.1789954454,-0.0133932987,-0.2288237661,-0.0748172104,0.1304322779,0.1742902696,-0.1199737489,0.2225718796,0.1605199277,-0.2823365629,-0.0438031517,0.0644406229,0.4907645881,-0.1109144166,0.4364704192,0.1345934868,-0.0280117076,-0.2728205621,-0.1875159591,-0.0571815856,-0.0791179985,0.4986955523,-0.0469055921,-0.0471794978,0.0717753321,0.2687982917,0.2098950148,0.062321201,-0.4079914987,-0.2259815931,-0.289521724,0.10091611,-0.0116054565,0.2131756544,0.144389838,-0.2463236898,-0.0482417084,0.1008623391,-0.3798359036,0.117768459,-0.1841448098,-0.0030396534,0.0865656435,-0.0323160402,0.4071114659,-0.1312693805,0.034685526,-0.1083798483,-0.2190884501,-0.233046934,-0.2197315395,0.1323817968,-0.175150305,0.2051775753,-0.2381195575,-0.2093167752,0.0741353482,-0.3226132095,0.3540242612,-0.1431896836,-0.039670825,-0.3114109635,-0.0773422644,0.0740278438,-0.0191833917,0.1167486683,-0.1692478508,0.2673142552,0.0588508956,0.0185744073,-0.0110085821,-0.1184472889,0.0162733197,0.1511173993,-0.0811748207,-0.126766324,0.3653866947,0.0866404697,-0.0308534894,0.1455337852,0.0639187768,-0.00261831,-0.0284300502,-0.0874154568,0.103236571,0.1949340999,-0.0153399585,-0.2444605231,-0.2322175503,-0.3353002667,0.2043920308,0.5522078276,0.2231103182,0.396238029,0.1489768177,0.1430341452,0.4616467059,0.0097332727,-0.1177480295,0.0230497532,0.1163854077,0.2166339606,0.1924875826,0.1981835663,0.2370375693,0.3087208569,-0.2465239316,-0.0080163861,-0.0191361941,-0.0263119545,-0.0773027837,-0.5415144563,0.0916395262,0.2027920187,-0.1640401632,0.339810282,-0.1043200865,-0.1156578809,-0.2313120365,-0.0226758197,-0.2210330665,0.2519416809,-0.2052824646,-0.1138441265,-0.0237651654,-0.0254415032,-0.138057515,0.2771424353,0.0455915369,-0.3564942777,0.0397714004,0.0447122492,0.1855156869,0.2119827121,0.1706032455,0.2630199492,0.0052259127,-0.2124300897,0.115431726,-0.15538086,0.1690758914,-0.0190374572,-0.229765296,0.1186696813,0.3068601191,-0.2667123079,-0.0471957214,0.1087817997,-0.0350512266,-0.4230211973,-0.079174526,-0.0261724815,0.2590159476,0.1034063101,0.1192831546,-0.0810740292,0.166545108,0.0135789085,0.0845873132,-0.1595273614,-0.0043845563,0.3106912673,-0.1107708663,-0.294113338,-0.0408080295,-0.2450644523,-0.0385170504,0.319406122,-0.2234350145,0.1000694335,0.2046152502,0.1049986184,0.0640411377,0.7118225694,0.1589998156,0.0762882233,-0.4750406146,-0.0213600546,-0.4419043362,-0.002064877,-0.0311682653,-0.0305289216,-0.1654070467,0.3960045576,0.2405217886,0.2552636564,-0.2573157847,-0.0068065915,-0.1121133044,-0.0385031737,-0.2530663013,-0.1821521968,0.0086789485,0.1265823692,-0.0399237126,-0.5482534766,-0.0145720029,0.1609401107,0.0496111065,-0.1970479041,0.1832741499,-0.0271979328,0.3204393089,0.5041556954,-0.1090563312,0.186358735,-0.0318924449,-0.0949272588,-0.0550479628,-0.1787011325,-0.1692059785,-0.0730096102,0.2044576705,0.4093383253,-0.0356407501,-0.0685094073,-0.28843835,0.1108473539,0.2608984113,0.1009282246,0.0189706217,0.0953329057,-0.0478143021,0.1805865616,0.0938623399,0.2747203112,-0.0633185729,-0.0451847762,-0.3445540071,-0.0310562868,0.2752750516,-0.1956723034,-0.283483237,0.1274483353,0.3155278265,0.0162927527,-0.1920623481,-0.3868098259,0.0132813649,0.1763200164,-0.1257052124,-0.3535123765,0.4556244314,0.1166872606,0.0575452,-0.2295944691,0.1711477786,0.0449680872,-0.3850894272,-0.5124903917,-0.3491408825]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2474","title":"cache_dir parameter for load_from_disk ?","comments":"Indeed reading the data shouldn't increase the VM's disk. Not sure what google colab does under the hood for that to happen","body":"**Is your feature request related to a problem? Please describe.**\r\nWhen using Google Colab big datasets can be an issue, as they won't fit on the VM's disk. Therefore mounting google drive could be a possible solution. Unfortunatly when loading my own dataset by using the _load_from_disk_ function, the data gets cached to the VM's disk:\r\n\r\n`\r\nfrom datasets import load_from_disk\r\n\r\nmyPreprocessedData = load_from_disk(\"\/content\/gdrive\/MyDrive\/ASR_data\/myPreprocessedData\")\r\n\r\n`\r\nI know that chaching on google drive could slow down learning. But at least it would run.\r\n\r\n**Describe the solution you'd like**\r\nAdd cache_Dir parameter to the load_from_disk function.\r\n\r\n**Describe alternatives you've considered**\r\nIt looks like you could write a custom loading script for the load_dataset function. But this seems to be much too complex for my use case. Is there perhaps a template here that uses the load_from_disk function?\r\n","comment_length":22,"text":"cache_dir parameter for load_from_disk ?\n**Is your feature request related to a problem? Please describe.**\r\nWhen using Google Colab big datasets can be an issue, as they won't fit on the VM's disk. Therefore mounting google drive could be a possible solution. Unfortunatly when loading my own dataset by using the _load_from_disk_ function, the data gets cached to the VM's disk:\r\n\r\n`\r\nfrom datasets import load_from_disk\r\n\r\nmyPreprocessedData = load_from_disk(\"\/content\/gdrive\/MyDrive\/ASR_data\/myPreprocessedData\")\r\n\r\n`\r\nI know that chaching on google drive could slow down learning. But at least it would run.\r\n\r\n**Describe the solution you'd like**\r\nAdd cache_Dir parameter to the load_from_disk function.\r\n\r\n**Describe alternatives you've considered**\r\nIt looks like you could write a custom loading script for the load_dataset function. But this seems to be much too complex for my use case. Is there perhaps a template here that uses the load_from_disk function?\r\n\nIndeed reading the data shouldn't increase the VM's disk. Not sure what google colab does under the hood for that to happen","embeddings":[-0.2947950661,-0.2046133727,-0.1245886311,0.2407285869,0.1985381395,0.0009984197,0.2094194442,0.0836256593,0.3559877276,0.4815080464,-0.1554521322,0.2485090196,-0.3452034295,0.282816112,0.2444066256,0.0840865374,-0.0638254359,-0.1021243334,-0.052796308,-0.1500464827,-0.1746011078,0.1620825529,0.0336108543,-0.4488104582,-0.0757848918,-0.151332736,-0.0464410521,0.200165078,0.2447034866,-0.2812951207,0.0641704872,0.0317722261,0.2542558908,0.473564446,-0.0001055217,-0.1294770986,-0.0200287718,-0.2458789945,-0.4441110194,-0.0372917876,-0.095758684,-0.1686767638,0.1495139152,-0.0932383463,-0.5178107619,0.0790239424,-0.0394624323,-0.1595536619,0.1908255666,0.1364302188,0.2488568574,-0.2450764775,-0.3598359525,0.0546795018,0.0662649423,0.0147160226,-0.1717963368,0.6884990931,0.2043837756,0.1545289755,-0.1795848459,0.4665690958,-0.0468552113,0.1521947384,0.4581644237,0.3910410702,0.1475776434,-0.2053609788,0.076880075,-0.1911458671,0.5348823667,-0.1255856901,-0.1675493717,-0.2609573007,0.165681392,-0.3373607397,0.3232167959,0.1641472131,-0.2515545785,0.4267630875,-0.1887676865,-0.404666692,-0.1755983084,-0.012893579,-0.0105824694,-0.0563667193,-0.2141340822,-0.1140272692,0.2221086323,0.1944479197,0.4096280336,0.0553650744,0.134954378,0.2564835548,-0.0272533335,-0.2071182579,-0.0789539963,0.2429348081,-0.0325585008,-0.021466665,0.4337691367,0.0182380453,-0.0849026591,0.1709044427,-0.1430236995,0.281547457,-0.2321080565,-0.0094442796,0.3639002442,0.137361154,-0.2417913526,-0.0757044777,-0.0214492213,0.0920444131,0.258441925,0.1874013245,-0.0370550863,-0.1761892289,0.2375370115,-0.1626896858,0.0557400659,0.1327322274,0.3196319044,0.3923742771,-0.1288207173,-0.1100473627,0.0997797027,-0.3074592054,-0.1705522537,0.047048036,-0.0905656219,0.0611919574,-0.1742025614,0.4017079473,0.3291392624,0.1654947102,0.2862094045,-0.3184047639,0.2037617713,0.033626914,0.1248870492,-0.2953373492,0.303219229,0.3845986426,0.0816503614,-0.13502267,-0.1155547649,-0.1106197909,-0.358395189,0.3294315636,-0.4407780468,-0.4950112104,0.377317667,0.2242396176,-0.2043097317,0.050191924,-0.4798679054,0.3666493595,-0.2225823402,-0.109098196,-0.1160730124,0.0506499112,-0.5522587299,-0.3293835819,0.133204475,0.2489878833,-0.262055248,-0.0554220714,0.0437133648,0.1354732364,-0.060600847,0.3321146071,-0.1957058907,-0.0220701862,-0.1836250722,-0.2804225981,0.1054391488,-0.2935200036,-0.6031992435,0.1867584735,-0.1149791777,0.0187048372,0.0413778946,0.2783218622,0.1554038972,-0.1983818114,0.4627945125,0.6393683553,-0.0005452702,0.1835105866,-0.3206068873,-0.2408221215,-0.3541868329,0.2877818048,0.0215086471,0.1196718588,0.0659591407,-0.2054599077,0.1027824655,-0.1397984624,0.0603698902,0.2875833213,0.4107733071,-0.2825403512,0.0150418077,0.1753421277,-0.538623631,0.2681299448,-0.0175532773,-0.1096555218,0.0511842817,-0.2234460562,0.1211263537,-0.0706112161,0.1311998963,0.0717237443,0.0809118152,0.0541534647,0.1319201142,-0.2969985604,0.0562821142,0.3559165597,-0.0787584335,-0.0990396515,0.3432376087,0.210073635,0.1246256083,-0.085683845,-0.0238235649,-0.1315010339,0.1183646098,-0.1236345097,-0.1380410194,0.2186228931,0.3182613552,0.3651823103,-0.1120717973,0.1964773536,0.25447613,0.0606508963,0.456736207,0.0402255766,0.0466406271,0.0790525451,-0.2858815491,0.3103999794,-0.1372111291,0.1091961265,0.2012239993,-0.1087443829,0.1503808349,-0.1756733656,0.150062874,-0.1859767437,0.1219744384,0.3175137043,0.3086785078,0.1633751541,-0.2031810284,0.2309563011,0.2223933786,0.0800496265,0.0466064289,0.205344975,-0.3644627035,-0.2673252225,0.1972737908,0.2737582922,0.4447880685,0.3685199618,0.3432639241,0.1414576918,0.1480698586,-0.0980152115,0.2551247478,-0.0680994615,0.2847212851,0.1146997213,0.2074634731,-0.1807905734,-0.0850653276,0.0455066711,0.1101966277,0.0680633783,-0.2206230909,0.0879644006,0.1208761483,0.0284159649,-0.1671191305,-0.0309394598,0.0207425188,0.0505535752,0.1200193986,0.5694577098,-0.1172007099,-0.0893695205,0.283852458,0.4693687856,-0.1382218301,-0.3756337464,0.0255945176,0.0326816514,-0.08782994,0.1110660136,0.069385007,-0.0297101587,0.260778904,0.138328746,0.0307778735,-0.1739283204,-0.2027786821,0.1722753346,-0.05329597,0.5861184597,-0.1246021315,-0.1088589728,-0.0743268728,0.1032784507,-0.1498512179,-0.3892001212,0.0036853191,-0.4489950836,0.119580254,0.2415837198,-0.3005292416,-0.0329489298,-0.3920276761,-0.4765082896,0.0372339562,0.3572834134,0.3248090744,0.2059687972,0.1230168492,-0.048452571,0.7896348834,-0.0669557527,-0.3142929077,-0.2918387055,0.3334436417,-0.3300228715,0.0642815158,0.1681377143,-0.1227298751,0.0367093943,0.2973199189,-0.4173502326,-0.6012536883,0.0159900449,0.0203157235,-0.2475111634,-0.1919259429,0.1435024291,-0.0271020513,-0.1219345927,-0.0096824309,-0.3961215913,-0.0704023391,0.4241247773,0.1346845925,0.042803932,0.0413444042,-0.0517120436,0.4943959117,-0.4250248373,0.1386836469,0.1263699085,0.2134533674,0.1903630644,-0.1662686318,-0.0875446945,-0.0142716197,0.0378782302,-0.6092913151,0.3105731308,0.199304536,-0.3015579283,-0.1535800099,-0.4781707525,0.0781090632,-0.1998575181,0.4040936232,0.2191236168,0.2348673344,-0.0079501132,0.0621132217,-0.2967654765,-0.0921781361,0.0147639411,0.1050126031,0.2014401257,-0.0729470029,0.1876026839,0.0099096233,-0.6256811023,0.3484437466,0.1110424772,-0.517683804,-0.0639814436,-0.0886627138,0.1346350759,-0.0923871323,0.3977906704,-0.2652189136,-0.0328225866,0.0341143161,-0.3106330931,0.0737866685,0.017013846,-0.0291328579,0.1078792587,-0.1755040288,0.2598281205,0.0781011581,-0.1723684222,0.1852948964,0.2388052493,-0.3523338437,-0.0049131783,-0.1606566012,-0.3351812363,-0.1652618349,0.0979278982,-0.0110504124,-0.0917559713,0.046596881,0.0504699349,0.0464182384,-0.1190763339,0.0926396176,0.1465318203,0.428119421,0.3504363894,0.0468888059,0.1736234576,0.1927523166,0.2371891141,0.3415151834,-0.2283036113,0.1304275244,0.0443159901,-0.1272239387,0.1842017472,0.4878558815,-0.0533719957,-0.0231594592,-0.1934852153,0.0504534058,-0.2650805414,0.3137112558,-0.3810216784,0.0676027685,-0.0759999231,-0.4482183158,0.5460740924,0.0200342368,-0.2757229209,-0.0217543934,-0.0325504728,-0.6783691049,0.4378556311,0.4553413391,0.9008393288,-0.1119629368,0.2154699415,0.0810915232,0.1416947544,0.1125417948,-0.2148104757,0.2796598375,-0.1595528573,-0.4095725715,-0.0838612616,-0.149556309,-0.1985899955,0.0108049577,-0.430949688,0.2519760728,-0.0740197673,-0.1373814046,-0.1570628732,0.3362627625,-0.0731524378,-0.4427512288,-0.3969441652,0.1862338632,-0.1674734354,0.1053347141,-0.0523008145,0.0524667464,0.0258860793,-0.0553074293,-0.0306449234,-0.1815196276,-0.5493505001,0.1628824323,-0.5084130168,-0.4618141949,-0.0653172955,-0.1032964438,0.2319495529,0.1113220304,-0.1506150514,0.275945276,-0.1374866813,0.1880232543,-0.3067370057,-0.2430710495,0.2787221372,-0.0475118384,0.0689160824,-0.0290537868,-0.045919586,-0.5067521334,0.0380848497,-0.0062371367,0.0113953045,-0.0427772701,-0.1738128066,0.010951655,-0.2437485009,-0.0375044644,0.179800272,0.0767667592,-0.0357060395,0.1759880632,0.304620862,-0.3061667383,-0.112817131,0.0676243156,0.4273447096,-0.1107056513,0.3846515715,-0.1064650118,-0.055689685,-0.2659322321,-0.2632015646,-0.1373116374,0.1585359573,0.4549857974,0.0388788022,-0.0419905633,0.0435741171,0.1575360596,0.2213467509,0.2007952482,-0.4092757404,-0.0901667774,-0.4182666838,0.2230786085,-0.1407077461,0.2151385397,0.0253864881,-0.0874385834,-0.0145794842,0.0859341621,-0.4361751974,0.197999984,-0.0815576464,-0.0296392124,0.1926058233,-0.0759423301,0.3838183582,-0.0198423322,0.0985495523,-0.1513777673,-0.1562871784,-0.2419055998,-0.2562246025,0.0936705619,-0.1467557102,0.2433602959,-0.2257207632,-0.1952906102,0.0614009462,-0.2626337707,0.2993644476,-0.0196474269,0.0245645549,-0.1629642844,-0.1659304947,0.1087572649,-0.0686340183,0.2435048074,-0.0496982299,0.2810969651,0.0465848483,0.084738262,-0.086791113,-0.0344279893,0.1746941656,0.058005061,-0.0447081625,0.0032733511,0.189741388,0.2614059448,0.0164558161,0.0645648763,-0.0556256063,-0.121304132,0.0116985114,-0.0609997846,0.1138401031,0.2637290955,0.013418328,-0.031248454,-0.142574504,-0.4244298041,0.2472470552,0.5229259133,0.2370681167,0.5254070163,0.1175635159,0.1424951404,0.4704644084,0.1671971083,-0.1948324591,-0.1361905038,0.0845340639,0.2150060683,0.1856606752,0.1765724272,0.3422061801,0.1557603925,-0.3438128531,0.0408968106,-0.0635236427,-0.0690376759,-0.0696564764,-0.3990437388,0.0988334119,0.159718886,-0.1371411383,0.3598382175,-0.0741064847,-0.0913309306,-0.3188643157,-0.1197412834,-0.2186481804,0.1206926256,-0.2310610414,-0.0476081222,0.0791314915,0.0099716038,-0.1583374441,0.3416454494,0.065360114,-0.2994521856,0.1420313567,0.0793011338,0.1010844707,0.3124506474,0.0226453505,0.321198076,0.0340062976,-0.1777881086,0.0434946604,-0.3089444935,0.0828939229,-0.0309938658,-0.1719606966,0.0373984464,0.2310183644,-0.2869732678,-0.1562151015,0.0883711874,0.0084795272,-0.3210494816,-0.2012784034,0.1315424442,0.1785123795,0.1131487712,0.1841855645,-0.1465115696,0.2762310803,0.0050307084,0.0224095657,-0.1316520423,-0.125081718,0.413731277,-0.1089558825,-0.3091594279,-0.0926850811,-0.1254784018,-0.0750104785,0.1403380036,-0.2083006054,0.0307833534,0.3496785164,0.1149618849,0.0427882969,0.6843208075,0.0350485407,0.0589936897,-0.4648109972,0.0278107319,-0.3333445191,-0.0074052648,0.0970940217,-0.0000055024,-0.3546730578,0.3303089142,0.1476748735,0.1639567614,-0.1917969137,0.00623101,0.0475088172,-0.159349665,-0.2527301908,-0.1447193772,-0.0132536879,0.0298754796,0.1062250361,-0.4717811942,-0.064739503,0.0717284381,0.0781089589,-0.1220117733,0.0354955643,0.0134362727,0.2344299853,0.6783956885,-0.20293881,0.153533265,-0.0773942918,-0.0472399965,-0.1047920063,-0.2000088394,-0.1764747202,-0.106599316,0.2350355685,0.4294146001,-0.0749500692,-0.0025272432,-0.2665859461,0.247726351,0.2344866246,0.0288892779,0.174554646,0.008323689,-0.0526773781,0.1104175895,0.0687191114,0.2957520187,-0.0955033526,-0.1706161201,-0.3986957073,0.0279796198,0.3589602113,-0.0761857703,-0.117517814,-0.0644374788,0.3950451016,0.0712588876,-0.1321643144,-0.3863277137,0.0009051815,0.2215428352,-0.0425214358,-0.3301816285,0.5074206591,0.098276034,-0.004662605,-0.233047545,0.131780535,-0.0935492218,-0.297511816,-0.5842607617,-0.2332978398]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2472","title":"Fix automatic generation of Zenodo DOI","comments":"I have received a reply from Zenodo support:\r\n> We are currently investigating and fixing this issue related to GitHub releases. As soon as we have solved it we will reach back to you.","body":"After the last release of Datasets (1.8.0), the automatic generation of the Zenodo DOI failed: it appears in yellow as \"Received\", instead of in green as \"Published\".\r\n\r\nI have contacted Zenodo support to fix this issue.\r\n\r\nTODO:\r\n- [x] Check with Zenodo to fix the issue\r\n- [x] Check BibTeX entry is right","comment_length":34,"text":"Fix automatic generation of Zenodo DOI\nAfter the last release of Datasets (1.8.0), the automatic generation of the Zenodo DOI failed: it appears in yellow as \"Received\", instead of in green as \"Published\".\r\n\r\nI have contacted Zenodo support to fix this issue.\r\n\r\nTODO:\r\n- [x] Check with Zenodo to fix the issue\r\n- [x] Check BibTeX entry is right\nI have received a reply from Zenodo support:\r\n> We are currently investigating and fixing this issue related to GitHub releases. As soon as we have solved it we will reach back to you.","embeddings":[-0.0967250019,0.3705003858,-0.0376106165,0.0805334821,0.1489466131,-0.1063510776,0.3863855004,0.4031184614,-0.061123658,0.2544853091,-0.0367732123,0.1469548494,0.2090731561,-0.0825828612,-0.1057561636,0.0061597624,-0.0564027242,0.347769171,-0.0588133708,-0.3109584451,-0.202006802,0.0611617602,-0.0399869643,0.1083547771,0.1429421604,-0.0178717673,-0.0885883048,0.0401961766,-0.6289506555,-0.294123888,0.2304198295,0.3536779583,0.0241569616,0.0977585763,-0.0001149367,-0.193783164,0.4348081052,-0.0376478955,0.1791505516,-0.3938967884,-0.4930344522,-0.3317319751,-0.1378981471,-0.2664651573,0.1761809289,-0.3223012388,0.1303076148,0.1256043911,0.0584312789,0.3948599398,0.1907943189,-0.0837685615,0.4714028239,-0.2440700382,0.1568886638,0.089191094,-0.0636926368,0.0680648983,-0.1046001986,0.2829131484,0.0303520281,0.2908119857,0.0494917929,0.1739325374,0.0865884349,0.1175321266,0.2991642058,-0.2475599647,0.1543980837,0.0806598887,0.7945279479,-0.1233866215,-0.1636312157,0.0783248544,0.2249083668,-0.2218073308,0.081119597,0.1100925356,0.1318532377,0.130947873,0.1229463443,-0.2656325102,-0.2755658031,0.0650090501,0.1293116957,-0.1051816493,-0.0930702239,0.0275286362,-0.0683570355,-0.1167767048,0.2643761635,-0.0598064698,-0.2564129531,0.2311538905,0.2473509312,-0.0809894577,-0.0505684763,0.0341484658,0.3572542667,0.0973298773,0.2150624245,-0.029652616,-0.0738931298,-0.1037192419,0.1122526824,0.0419797897,0.3595068455,-0.0530412607,0.199470371,0.0872092471,-0.0922934487,0.0737133175,0.0192253105,-0.0975429565,0.5840197206,-0.168579638,0.3791415691,-0.4698496461,-0.031587109,0.121334441,-0.0889860243,-0.1356124729,-0.4675282836,0.0189321321,-0.0387980044,0.2661347985,-0.1196072474,0.0638765544,-0.2146764696,-0.1119567603,-0.1042337269,-0.4499019384,0.0859547108,0.2072413117,-0.0272003449,-0.3636471331,0.220649913,0.2520163357,-0.2041303664,-0.5929424167,0.0520067625,0.2018798143,0.1980849802,0.2398324311,0.0729131177,0.2277602255,-0.071303986,-0.1430610567,-0.003468036,0.3119443953,-0.1970106065,-0.4158118367,-0.5545141697,0.1263848543,-0.0154378656,-0.0302750524,0.3378929794,0.172124058,0.2121184468,-0.1259371489,-0.1239725873,-0.0948559791,0.24657318,0.1058486253,0.1277978569,0.2760051489,0.1801678389,-0.059993919,-0.4540840387,0.07300511,0.3593138754,-0.028415475,-0.0059811133,0.3127602637,-0.0990879759,-0.3946625292,0.2607989311,-0.5165682435,-0.3449460864,-0.0686951727,-0.0866520852,0.0106102945,-0.0730016083,-0.1784451753,0.3146195412,0.264207691,-0.3067135215,-0.2928422987,-0.0682026073,-0.0304838419,-0.2615099847,-0.2276733965,-0.3169073462,-0.1217575297,0.6132693291,0.4346250892,0.1965690851,0.2121606916,0.2273948044,0.186016798,-0.055820398,0.105781585,0.5157321095,-0.2429505736,0.2426191866,-0.0186445955,-0.0445015281,0.1058372781,0.364911139,0.1284542531,-0.1423499137,-0.034027677,-0.3487661481,-0.1293665618,-0.1910822093,0.0236552507,0.1080503762,0.1056260392,-0.0852857828,0.1599137634,0.0961170495,-0.0619293377,-0.1601580381,0.1793097854,-0.0670498088,0.1715123653,0.0375265516,-0.0785583481,-0.0061084623,0.0678291842,0.082690306,0.1868734509,-0.212024644,0.2070410699,0.0062331068,0.0665713102,0.1753115058,0.3192522824,0.1791736037,-0.2357140481,-0.0851681307,0.1161766574,-0.2221634984,0.178564027,0.1205098927,0.2113172263,-0.2766640484,-0.2890143692,-0.3767161667,0.1957046986,0.5364695191,0.0458592996,-0.3070167005,-0.3863206208,-0.1415654272,-0.0293307249,0.1237045154,0.1486452818,-0.300234437,0.2740983069,0.5304563642,-0.0463285968,0.0184068996,-0.136000827,0.1112140492,-0.0769464672,0.1097098663,0.3504441679,0.1392980069,0.1322077364,-0.1204899102,0.1311260909,-0.2398717552,-0.1490615904,0.1420025676,0.0199940521,0.5988178253,0.2531802654,0.0380570255,-0.1625417471,-0.4330268204,0.2709622979,0.0133442227,0.4834909141,-0.1430612206,-0.5351665616,-0.0234868526,0.2057470381,-0.3697085083,-0.3503602445,-0.4329450727,-0.1239149496,0.0663367212,0.2066048384,0.1248129681,0.1780059934,0.057008408,0.0615401864,-0.0046262424,0.2375089973,-0.1140942425,0.07026954,0.0646879449,-0.0235295445,-0.1753820926,-0.1083829999,0.2719162405,-0.5216576457,0.1199310943,-0.5896314979,-0.0733040869,0.1976411194,0.1813601404,0.1604100168,-0.2260288149,0.1183457077,0.1115448996,-0.0442179404,0.0312312841,-0.2721320987,-0.1584836841,0.0709126964,-0.4817188978,0.2275580168,-0.0358135141,-0.4244445562,-0.0655133948,-0.0550049245,0.0293306299,0.1248127148,0.31581074,0.3227924407,0.0852842256,-0.1649546772,-0.1046718583,-0.2954024971,-0.102194421,-0.6679029465,0.2470011711,-0.3297753036,-0.3566850722,0.1468521804,0.1116567403,0.3824832737,-0.059308745,-0.2216844708,0.2715043426,0.0712309256,0.5275432467,0.1116093248,-0.225499317,0.5041925907,0.1181064844,-0.0133410562,-0.2628585994,0.2124114633,0.3426111042,0.01155819,0.2765508592,-0.0946914703,0.0140320119,-0.1651981324,0.8457219005,0.1296900958,-0.016075179,-0.0834944397,0.1961854994,0.4897732139,0.0004719502,-0.1514770389,0.0353651904,-0.0071054026,0.0440894179,0.4767307341,0.0500956587,-0.2161652893,-0.0282669365,0.0931128487,-0.570309639,-0.2282783389,-0.3788039982,-0.4062751532,0.2478974015,0.3844501078,0.2499268651,-0.2203362584,-0.209267512,-0.0211007744,0.0414455794,0.2817259431,0.0732870698,-0.1079245433,0.3157145977,-0.4676323235,0.0744109526,0.1037520245,0.225378111,-0.0193955582,0.3193941116,0.1153438464,-0.0286819991,0.0752714649,-0.2320332825,0.2020751685,-0.1445951611,0.1543175876,-0.1736942828,-0.1093853191,-0.098053284,0.1274947077,0.2200846821,0.4640784562,-0.0552610196,-0.0556427315,0.2325370908,-0.2038918287,-0.1424517483,-0.3197479844,-0.1269571185,-0.1669629216,0.0747228935,0.0526803881,-0.1765682101,0.10465049,-0.745044589,-0.4398395419,0.2958354354,0.3356731832,-0.1973886788,-0.0409764647,0.1445718408,0.0108093899,0.3822500706,0.6417276263,-0.1232715622,0.5436773896,0.0476812944,0.0276159029,-0.5039852262,-0.1938749105,-0.3857544959,-0.3228616118,-0.0463007838,0.1902396977,-0.0814385712,0.032138627,-0.2670419812,0.0485086441,-0.2286581099,0.4517758191,0.0935662091,-0.3955610693,-0.240521118,0.0785963386,0.0804695934,-0.1259428263,0.7364740968,0.2169435322,-0.159440726,-0.1123099029,0.3402057588,0.5876361132,-0.344402343,-0.2561013699,-0.1041015014,-0.5278944373,0.3125472963,0.2465445399,-0.0010466605,-0.6754637361,-0.4289260209,0.0101618124,0.0746379346,0.4377726018,0.0988086835,-0.3488592803,-0.162266776,-0.0992376283,0.2801806927,0.298220098,-0.080154404,-0.2423162758,-0.1729039699,-0.0664295107,0.0598763749,-0.3673050404,0.0053494507,-0.2066658288,-0.0192054715,0.0614517666,0.0371587463,-0.2663782835,0.1245199293,0.1992771626,0.0585918017,-0.0870544761,-0.0602893829,-0.6240654588,0.4838114977,0.3043715656,0.0751142427,-0.3853136897,0.3092935681,0.1534983516,-0.0655556247,0.0244935025,0.1299606264,0.1941206455,-0.0426277444,-0.4403536916,-0.2009083927,-0.1321410239,-0.2104883939,-0.4892983139,-0.3531471491,-0.2041273117,0.2366013825,0.0571920499,0.0256891716,0.1611796916,-0.3184297979,0.0937966779,0.1432299018,-0.0172177944,0.173685655,-0.0976154208,-0.2676525712,0.0150883626,0.339376539,0.2690903246,0.5175557733,-0.0262349322,-0.3436092734,-0.0060005104,-0.0978043303,0.3593732119,-0.0548263825,-0.2926119268,-0.0020855975,-0.3965001106,0.2255900949,0.0006402381,0.2546984851,0.0983994827,-0.3241000772,-0.2721592188,-0.2226507962,-0.1150930747,0.0633042455,-0.0822059438,0.0008344767,0.2460262477,-0.0737517849,0.0565664582,0.0160901863,-0.2135611922,-0.0610002093,0.0404814892,0.022131877,-0.0266830269,-0.0488008559,-0.0498564281,-0.2507319748,-0.0487087816,0.0645243675,-0.1613749564,-0.099605225,-0.1827903688,0.1642116457,0.2026184797,0.1977992356,0.2497017831,-0.2101872116,-0.4068311751,-0.0708102062,-0.1022661254,0.4876911342,0.0147070689,0.1685093045,0.1323976815,0.3828552961,-0.1328426749,0.0028584795,0.0281108767,0.2305226922,-0.1195189133,0.3854368627,-0.0718183219,-0.0742235333,-0.2405994833,0.0278490372,-0.1232714504,0.2926117182,0.4203765988,-0.4343615472,0.2582091987,0.2444010675,0.4816324711,-0.1824465841,0.0471990407,-0.2001367211,-0.0078965323,0.2093535066,-0.2823455036,-0.1439035982,-0.006837076,0.0855942294,-0.1265205294,0.1708847731,0.1535381824,-0.5379576683,-0.090320304,0.1077005118,0.3146013021,-0.2242764682,0.1277176291,0.5658208132,-0.0655923486,0.0721996799,0.3955343664,-0.0889114663,-0.0875122026,0.2043167949,0.2443236262,0.2287298441,0.5702473521,0.3067271709,0.1086935177,-0.0521495864,0.02942729,-0.1481008828,0.0401951484,0.0470428951,0.3019080758,0.0064733159,0.0262699034,-0.3086571693,-0.2823632956,-0.0682359338,-0.0162676815,0.1108099818,-0.1570953578,-0.3861691356,0.2526269257,0.0733975098,0.0037788621,0.1801875532,0.1687810123,0.0084437924,-0.373250246,-0.8148137331,0.0297498014,0.0382262431,0.1774600744,-0.2293674052,0.2444329709,-0.3747792244,-0.1698663682,-0.0174339991,0.6133742929,-0.0014461952,0.1128969714,0.4109487236,-0.0525882132,0.1922064275,-0.1282741874,0.1452539116,0.2382663637,0.3600050807,0.1735500097,0.0289491545,0.1182527691,-0.0549949408,0.0728613436,0.6321055889,0.2950083315,0.125405997,-0.107587114,0.0866271406,-0.1764473468,0.0965500996,0.1644863188,-0.1549878865,-0.0991473347,0.100754559,0.0849228352,-0.0510789566,-0.2872258127,0.055120822,-0.0023822258,-0.0983055457,0.1630260795,0.0832355991,0.2478386015,-0.3213146031,-0.4689475298,0.4513774514,0.0749723986,0.3187841773,0.17532067,0.0775498822,-0.0907016098,-0.0964825079,0.3948313892,-0.3154283464,0.1677636802,0.3029379845,-0.3918507695,0.0016643973,-0.1767329872,-0.1617289037,-0.2946323156,-0.3336303532,0.3862124681,0.0862187445,0.0532519333,0.0913769081,-0.0561347641,0.1138782576,0.0822550654,0.3240066171,0.1581222862,-0.0281349402,-0.1749519259,-0.1914308369,-0.2141385078,-0.0954476073,-0.3301589787,0.4264927208,0.0526306555,0.3511557579,-0.0239791926,-0.0246435404,-0.0855717435,0.2809417248,0.0544385649,0.0992323607,-0.6209915876,-0.0439067781,0.0433199368,-0.5385665894,-0.0834996626,0.1761343181,0.0424825996,0.2947028875,-0.0027918713,0.0749356449,0.2606540918,-0.2971814275,-0.156008929,-0.1408725083,0.2689493895,0.3504348099,-0.4279558957,-0.2316553146,0.0006437611,0.3700453043,-0.156818375,-0.0027837721,-0.1221082434,0.047349181,0.2214327604,-0.032808315,0.1912916452,0.0517688952,-0.1450219005,0.1849898547,-0.049189575]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2472","title":"Fix automatic generation of Zenodo DOI","comments":"Other repo maintainers had the same problem with Zenodo. \r\n\r\nThere is an open issue on their GitHub repo: zenodo\/zenodo#2181","body":"After the last release of Datasets (1.8.0), the automatic generation of the Zenodo DOI failed: it appears in yellow as \"Received\", instead of in green as \"Published\".\r\n\r\nI have contacted Zenodo support to fix this issue.\r\n\r\nTODO:\r\n- [x] Check with Zenodo to fix the issue\r\n- [x] Check BibTeX entry is right","comment_length":19,"text":"Fix automatic generation of Zenodo DOI\nAfter the last release of Datasets (1.8.0), the automatic generation of the Zenodo DOI failed: it appears in yellow as \"Received\", instead of in green as \"Published\".\r\n\r\nI have contacted Zenodo support to fix this issue.\r\n\r\nTODO:\r\n- [x] Check with Zenodo to fix the issue\r\n- [x] Check BibTeX entry is right\nOther repo maintainers had the same problem with Zenodo. \r\n\r\nThere is an open issue on their GitHub repo: zenodo\/zenodo#2181","embeddings":[-0.0932828709,0.394243449,-0.0347407646,0.099130474,0.1498733163,-0.1066873893,0.3806301951,0.3882210553,-0.0758683681,0.253555268,-0.064562887,0.2056420445,0.2068252861,-0.0944909677,-0.1203863472,0.016830558,-0.0314425714,0.3863615692,-0.057352256,-0.3248518109,-0.2133872807,0.0766483694,-0.0383451693,0.0802262947,0.1236057132,0.0087414347,-0.1001702398,0.0040567052,-0.6152876019,-0.2550566494,0.2330326736,0.3883582652,0.0097615132,0.1403077543,-0.0001156442,-0.2252505422,0.4270015359,-0.0163295865,0.1792460829,-0.3721206188,-0.5172886252,-0.3386047482,-0.1104393452,-0.2647010088,0.2235631049,-0.3678027391,0.115695253,0.1140099242,0.0215495154,0.3697971702,0.1875544637,-0.1051705927,0.4895671606,-0.217632398,0.1373074055,0.1001364216,-0.0828814432,0.0587363206,-0.1178573444,0.3063135147,0.0154910637,0.2852521241,0.0233926568,0.172807157,0.0930840448,0.1147537529,0.3039465249,-0.1714944541,0.1280236542,0.1230859533,0.832862258,-0.1432768703,-0.2120329738,0.0880180672,0.1983065903,-0.236638248,0.0740668997,0.078993395,0.1280183941,0.1463824213,0.1279241443,-0.2579166591,-0.253556639,0.0778847486,0.0788915157,-0.0996273234,-0.084779948,0.009448722,-0.0460835807,-0.0916435122,0.3670062423,-0.0872830674,-0.2454650104,0.1976322532,0.2415435463,-0.0673406124,-0.0352681242,0.1007865593,0.3323948979,0.1322206557,0.2674002647,-0.0176760405,-0.0796178356,-0.07322184,0.1397774816,-0.0086306017,0.3600791097,-0.1042005643,0.1855026037,0.0518748052,-0.0983477458,0.067417413,0.054464329,-0.1070985794,0.6239064932,-0.1588163525,0.3452750742,-0.4820899963,-0.0601069592,0.1509780437,-0.0923017487,-0.1525580734,-0.4643554091,-0.0169897508,-0.0030420616,0.3226550221,-0.072182335,0.0545920432,-0.2229958028,-0.168489635,-0.1057487652,-0.4463167489,0.0522426479,0.2065855414,-0.0011709336,-0.3521764874,0.197790578,0.2260470688,-0.1673873365,-0.5947030783,0.0487626828,0.1647951901,0.165528819,0.2422035933,0.0377056971,0.2358960509,-0.0561090186,-0.0592608117,-0.0253798738,0.2959363759,-0.1597511321,-0.3853194118,-0.5159863234,0.1229494661,-0.0173366182,-0.0714498237,0.3614366949,0.1926749647,0.2250159681,-0.1650751978,-0.1569584608,-0.0901965573,0.2544915378,0.124928683,0.1172455326,0.3184442818,0.0776741281,0.0068548811,-0.4972431064,0.0783776864,0.3269890547,-0.0679561794,0.0046435008,0.2849656045,-0.0645963103,-0.3945794106,0.2350950241,-0.4871272147,-0.3446820676,-0.0375059582,-0.0600487292,-0.0081504267,-0.0700566843,-0.1746236682,0.2466297299,0.2377924174,-0.3222362101,-0.2767680585,-0.1004822776,-0.0193046629,-0.2484045774,-0.2057971507,-0.2596994638,-0.1154619083,0.6102955341,0.4047065973,0.1696137041,0.2542264163,0.1998075843,0.2200543135,-0.0433983132,0.0964039788,0.4835833311,-0.2071894407,0.2367183715,-0.0505857952,-0.0197005216,0.0789463073,0.3896897435,0.0955225974,-0.0952427834,-0.0432568491,-0.3474522829,-0.1109543815,-0.1589937657,0.05873546,0.1061796546,0.078716293,-0.1146688089,0.1632086188,0.1062244549,-0.1158350781,-0.1601466984,0.1873285919,-0.0361526906,0.142281875,0.0558788218,-0.0835240185,0.0014843171,0.0762866288,0.0873074904,0.1970050782,-0.2366698086,0.2509304583,0.0188194551,0.1150829196,0.101737313,0.3152865767,0.1941678375,-0.224107787,-0.1002208665,0.1239513755,-0.2105144113,0.1670019776,0.0604422688,0.1879071295,-0.2931379676,-0.2764587998,-0.3872968554,0.2040327042,0.5019196868,0.0381810851,-0.3221302629,-0.4127164483,-0.1184822842,-0.0647520795,0.1155248284,0.1912376881,-0.2757062316,0.2671723962,0.4886540473,-0.0309058446,0.0030141713,-0.1926344633,0.0880844742,-0.0898323655,0.1228223294,0.3807592392,0.1672833413,0.149175033,-0.1024721861,0.0924102962,-0.2890713513,-0.1720481217,0.1176505461,0.0428197943,0.6119812131,0.2980555594,0.0219496656,-0.1812718213,-0.4736090899,0.2799232006,0.0439986289,0.4817019701,-0.1301573962,-0.5318512321,-0.048868563,0.2293630987,-0.3871013522,-0.3776044548,-0.3723509908,-0.1700176895,0.0392706692,0.2129530013,0.1064998358,0.1337770522,-0.007893987,0.0321994461,-0.010892381,0.239023149,-0.1423677057,0.0784383789,0.0735010058,-0.0318246074,-0.1976611763,-0.083522208,0.2792720199,-0.5059190392,0.1397278756,-0.5817276239,-0.0640512854,0.1704770923,0.2002034932,0.1675930768,-0.2041617483,0.1052956507,0.1108049899,-0.0587826036,0.0500783063,-0.2695665658,-0.1343619972,0.076053448,-0.4455569685,0.2318740338,-0.0432307497,-0.4657226205,-0.0397085659,-0.0698588639,0.0087665478,0.0806312487,0.2930549383,0.2758255303,0.1201167256,-0.1805644333,-0.098605305,-0.2380491346,-0.1330815405,-0.6524446011,0.2465555221,-0.3133287132,-0.3543634713,0.1500936747,0.1037111729,0.3814441562,-0.0061269663,-0.2047825754,0.2546283305,0.0792150795,0.4864209592,0.1491881609,-0.2074271888,0.5132312179,0.1229927093,-0.0085548116,-0.2754908502,0.2137739062,0.3333303034,-0.0036007545,0.290617913,-0.1034906805,0.0191383585,-0.1552757919,0.7816070318,0.0748125538,-0.0199385565,-0.0556436963,0.2394745946,0.5023627281,0.016449308,-0.1836204976,0.0851341262,-0.0156056676,0.0503268316,0.4427652657,0.0535645112,-0.2275871634,-0.0261368658,0.1506719291,-0.5740029216,-0.2443083227,-0.3904034197,-0.4486213326,0.2665712535,0.4034620523,0.2810754478,-0.1891522706,-0.1890504509,-0.0234121941,-0.0031506147,0.2987802625,0.0547349676,-0.1094797179,0.335160464,-0.4646450579,0.0601857156,0.0745490864,0.2616271973,-0.0187298767,0.318944335,0.1133931279,-0.0025565277,0.1504484266,-0.2476605177,0.1963242888,-0.1526253521,0.1243000031,-0.2300027013,-0.1102542505,-0.1079181135,0.0619094111,0.1958957314,0.4469104409,-0.1038466319,-0.0177887809,0.2152649164,-0.1738678813,-0.1664157361,-0.3462203443,-0.079379119,-0.1432831734,0.0917596519,0.0484491512,-0.2176761627,0.1544992924,-0.7347431779,-0.4275983274,0.2579804063,0.3688342571,-0.131356284,-0.0569331795,0.1297714859,-0.0298403241,0.3940135539,0.6949183941,-0.088680312,0.5424674749,0.016871836,-0.013956924,-0.5385689735,-0.1900947243,-0.380784899,-0.3176026046,-0.0570349097,0.1817416847,-0.0906869844,0.0109105241,-0.3648436069,0.0432275645,-0.1916186661,0.4432834089,0.1211550608,-0.3691505492,-0.2751365304,0.0868400633,0.1485039145,-0.1373633146,0.6784926057,0.2168352008,-0.1530548185,-0.0953433886,0.3730174601,0.5726534128,-0.3677917421,-0.23237288,-0.0831186548,-0.4831365645,0.3436771035,0.196104005,-0.0092897117,-0.6566013098,-0.436960578,-0.0029363527,0.0846448541,0.4113505185,0.1586982906,-0.3403171301,-0.173806861,-0.1196738109,0.2188957483,0.2896239758,-0.1011893451,-0.2579543293,-0.2130929083,-0.0207100976,0.0635609627,-0.3572360277,-0.0194681063,-0.2139989585,-0.0077065784,0.0936864167,0.047674641,-0.2661893666,0.1054154113,0.2157576978,0.0407060757,-0.1121953949,-0.0809914395,-0.5836494565,0.4978783429,0.2646733224,0.1008882672,-0.4010162055,0.2870317996,0.1576528102,-0.0822621807,0.0131520648,0.1258675903,0.1904937774,0.0073670475,-0.4133663774,-0.1846573949,-0.1244508326,-0.1799500734,-0.4760700464,-0.3439109325,-0.1513126343,0.1970789433,0.0551084839,0.0292189177,0.1710063815,-0.3160808682,0.0926830918,0.1718724519,-0.0023675803,0.1961985826,-0.0872699171,-0.2533201575,0.0067745876,0.379340142,0.2609356642,0.4856829643,-0.0324352384,-0.3512033522,-0.0024749173,-0.1070749611,0.3448709846,-0.0937592089,-0.2616343796,0.004919441,-0.3712386489,0.1906501651,0.0390994623,0.1902031004,0.1496659517,-0.3142699897,-0.2813903689,-0.2328201681,-0.089943029,0.0482821129,-0.0701134354,0.0287259277,0.2584739029,-0.0588118248,0.021082364,0.0455815941,-0.2064445615,-0.0414251499,0.0606878437,0.0095794564,-0.0148042655,-0.044113569,-0.0938066244,-0.2502225935,-0.0534851253,0.0864143446,-0.1508473903,-0.0932835266,-0.2262032926,0.1624041945,0.2445539683,0.2056560367,0.2881823182,-0.2260029167,-0.39443928,-0.1152761951,-0.1636573821,0.5030627251,-0.0102194604,0.1572260857,0.1103710756,0.4229063988,-0.1316490024,-0.0161335561,0.0620596856,0.1838669628,-0.1328743845,0.3991327882,-0.0566050969,-0.0718699396,-0.2178449482,0.0268855691,-0.1278129816,0.236422658,0.435515821,-0.4055805802,0.2881438732,0.2377658039,0.5250592232,-0.1516385376,0.0689501911,-0.2050172389,0.0056181513,0.2084183395,-0.2678335309,-0.1486811936,0.0103677055,0.1443114877,-0.1289648861,0.1540709436,0.1336095631,-0.5700016618,-0.1002751291,0.0998232886,0.2820520401,-0.2140442282,0.1105136946,0.6199074388,-0.0803866386,0.1038863882,0.4857153296,-0.0811597407,-0.0776410252,0.2727069855,0.262406677,0.2202563435,0.4952891767,0.2957501709,0.0440323725,-0.0632760152,0.0249672458,-0.1340910047,0.0584232025,0.0841429606,0.2926394343,0.047963433,-0.0447158664,-0.29869017,-0.3202171028,-0.0708666667,-0.025526721,0.0852861032,-0.1081807762,-0.3852229714,0.307762146,0.0741830468,0.0068924162,0.1515877545,0.2006689608,-0.0220511537,-0.3321388662,-0.8933185339,0.0290020276,0.0012891702,0.1753227711,-0.2215000838,0.1813257933,-0.4121181667,-0.1685069352,-0.0005258863,0.6146162152,-0.0506746657,0.1478739679,0.431175977,-0.0753808841,0.1523250788,-0.0951571465,0.1130063236,0.2606200576,0.3594924808,0.2033472508,0.0782859027,0.1092896909,-0.0546959825,0.0638804883,0.6495121121,0.3192133904,0.1096374467,-0.1245618016,0.1133415848,-0.1635334492,0.1063707694,0.1936610043,-0.129844442,-0.1156115308,0.1088026762,0.0930192322,-0.0579093695,-0.2761663496,0.0510126241,0.0342684239,-0.1229973212,0.191200465,0.0921436399,0.2127639949,-0.3060308695,-0.4880693257,0.4472506642,0.0753389224,0.3521840274,0.1332858652,0.0810542405,-0.0617395304,-0.0776737034,0.3390155435,-0.2810379565,0.1368282139,0.3183822632,-0.3833419681,-0.0361452922,-0.1471694708,-0.1681506783,-0.2840764225,-0.3117672205,0.3715990186,0.1087422669,0.0544137247,0.0514032692,-0.0601132661,0.1082706004,0.1798319072,0.3098835051,0.1812488884,0.0078992313,-0.1639168859,-0.1630961597,-0.2410724759,-0.0728934631,-0.3530278206,0.3977640569,0.0407788083,0.3121425211,-0.0480068401,-0.0000995106,-0.0642088503,0.3095231056,0.0588220656,0.0698166713,-0.662165463,-0.0234545134,0.0115175601,-0.5694556832,-0.0872990936,0.2147097737,0.0452801995,0.2947057784,0.012344433,0.0684527829,0.276481092,-0.3221191764,-0.1434512734,-0.1486726701,0.2394924462,0.3593899012,-0.4365219474,-0.1958890706,-0.0282594562,0.3507600129,-0.1879641563,-0.0248397365,-0.1295251846,0.0290364735,0.1974965632,-0.0537030548,0.2217454761,0.0648875013,-0.1482273638,0.1938784719,-0.0448475555]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2472","title":"Fix automatic generation of Zenodo DOI","comments":"I have received the following request from Zenodo support:\r\n> Could you send us the link to the repository as well as the release tag?\r\n\r\nMy reply:\r\n> Sure, here it is:\r\n> - Link to the repository: https:\/\/github.com\/huggingface\/datasets\r\n> - Link to the repository at the release tag: https:\/\/github.com\/huggingface\/datasets\/releases\/tag\/1.8.0\r\n> - Release tag: 1.8.0","body":"After the last release of Datasets (1.8.0), the automatic generation of the Zenodo DOI failed: it appears in yellow as \"Received\", instead of in green as \"Published\".\r\n\r\nI have contacted Zenodo support to fix this issue.\r\n\r\nTODO:\r\n- [x] Check with Zenodo to fix the issue\r\n- [x] Check BibTeX entry is right","comment_length":55,"text":"Fix automatic generation of Zenodo DOI\nAfter the last release of Datasets (1.8.0), the automatic generation of the Zenodo DOI failed: it appears in yellow as \"Received\", instead of in green as \"Published\".\r\n\r\nI have contacted Zenodo support to fix this issue.\r\n\r\nTODO:\r\n- [x] Check with Zenodo to fix the issue\r\n- [x] Check BibTeX entry is right\nI have received the following request from Zenodo support:\r\n> Could you send us the link to the repository as well as the release tag?\r\n\r\nMy reply:\r\n> Sure, here it is:\r\n> - Link to the repository: https:\/\/github.com\/huggingface\/datasets\r\n> - Link to the repository at the release tag: https:\/\/github.com\/huggingface\/datasets\/releases\/tag\/1.8.0\r\n> - Release tag: 1.8.0","embeddings":[0.0188328642,0.2643171251,-0.0537575781,0.1654242426,0.1444075853,-0.0486195646,0.2956543267,0.3008949459,0.0125504285,0.242185697,-0.0125956545,0.1527004838,0.1531535685,-0.0105606355,-0.0150587875,-0.0113817109,-0.1902420074,0.3833129406,0.0192255285,-0.2728261352,-0.0802597106,0.0352201611,0.0213998426,0.075741455,0.0815024301,-0.0937683135,-0.0888004005,0.143186003,-0.6058160067,-0.3880801797,0.3033675551,0.3281611204,-0.0811473876,-0.0234927498,-0.0001181506,-0.1728259623,0.4213643074,-0.0058291624,0.0931776389,-0.4274674356,-0.5783349276,-0.2995273471,-0.2516589463,-0.2421192974,0.1319315135,-0.2870878577,0.1257436424,0.0686942786,0.2115385234,0.3670431376,0.132983759,0.0080191335,0.5302379131,-0.2428427786,0.0904612616,0.1740521342,-0.0342529826,0.1918554306,0.0288992412,0.3487540483,-0.0097329868,0.4010199904,0.025682969,-0.0115816593,0.1483985931,0.0837114826,0.1936552525,-0.3168489039,0.2095567584,0.0580773614,0.7970228195,-0.211556226,-0.3363280296,-0.0487540215,0.2337999642,-0.2339422703,0.0584118553,-0.0532489717,0.1060293615,0.2367158532,0.0488145798,-0.452583462,-0.2751454413,-0.0164492205,0.0710524321,-0.0525216199,-0.1526584178,-0.026970692,0.0195290465,-0.0795237571,-0.0462922491,0.0442449786,-0.176940456,0.2272758484,0.210976243,-0.1259251088,-0.0560807399,0.0083235959,0.5040109754,0.1538022906,0.2533904612,-0.1297270358,-0.1405759007,-0.1524105817,0.1476975381,0.1218779236,0.314168632,-0.1610112339,0.1694394797,0.2534525692,0.0824104697,0.0321332701,-0.0234065708,-0.0403025672,0.3857743144,-0.1556466371,0.4334898889,-0.437556833,-0.1630749702,0.1929126531,-0.181387797,-0.1911931336,-0.3931791782,-0.0156549178,-0.1364433169,0.3304491341,-0.1980864704,-0.014288716,-0.0921550989,-0.1597937495,-0.0890810937,-0.3371344805,0.0141227366,0.3546697199,-0.0559159443,-0.5388818979,0.2696074545,0.3608840704,-0.101627022,-0.5210308433,-0.1006259844,0.1846012324,0.1480939239,0.2446360737,0.0973092839,0.2966916859,-0.1031371355,-0.0750422478,-0.0908156335,0.2668316066,-0.2066263556,-0.4705869853,-0.478235513,0.0741063505,-0.0533963591,-0.0676374063,0.2885314226,0.1652777046,0.0928109363,-0.1225229278,-0.0665579289,0.1615645289,0.1698337644,0.0248986073,0.2625796497,0.3304292858,0.1081319004,-0.1468625516,-0.3137835264,0.0973710865,0.2828282714,0.0033237885,-0.0448510237,0.3146296442,-0.1357925236,-0.3073585033,0.2696188092,-0.5879855156,-0.364151597,-0.0572758131,-0.2582696676,0.0593365468,-0.008058643,-0.1976899505,0.3686468005,0.2593597472,-0.3146285117,-0.3157481849,-0.125650093,-0.0444278605,-0.1898540705,-0.2747274935,-0.3628685772,-0.0656624958,0.4524851441,0.3284434676,0.1630731225,0.2639475167,0.1869410723,0.1493179947,0.0041005574,0.0452151857,0.4665614069,-0.1666944623,0.1884704083,-0.0460650586,-0.1447493285,0.1752282828,0.4449854791,0.170723021,-0.1821965277,-0.1048287451,-0.4984592199,-0.1512636244,-0.123564288,0.0029100718,0.0209837891,0.1246435642,0.0562109686,0.1935500056,-0.0028669671,0.0172534175,-0.1043000892,0.3185937703,-0.2101012766,0.1065792367,-0.02846005,-0.0456147678,-0.0227371119,0.1432827562,0.0684764981,0.1075025499,-0.0491037443,0.1979271173,-0.0298827756,0.2410429865,0.1510243118,0.3712355196,0.3792065978,-0.2344623059,-0.044760447,0.0559506714,-0.2126634121,0.1049376875,0.1136080027,0.2080778778,-0.2715879977,-0.2774095833,-0.35139516,0.1562840641,0.4229256809,0.0323060416,-0.3406136334,-0.3359737396,-0.19703044,-0.2026632279,0.2110149413,0.1531803757,-0.3375046849,0.2807776034,0.5243439078,-0.0864707828,-0.0126692522,-0.1467856318,0.0731581524,-0.1405577064,0.0798447728,0.3469860554,0.1621224284,0.1023155525,-0.0670552179,0.1023797169,-0.1716163605,-0.0997635424,0.1845201999,-0.0271852408,0.5872587562,0.3576720059,-0.0427256338,-0.058404658,-0.2886099517,0.1547620744,0.0236792229,0.3564015031,-0.1751269102,-0.6635761261,0.0064438586,0.0676270798,-0.3993249238,-0.4240987003,-0.5256232619,-0.0824077353,0.0585298873,0.1291621923,0.1624428183,0.1735312939,-0.0767111406,0.223111853,-0.0968481675,0.2269190252,-0.1752964854,0.1168274209,0.0629978627,-0.0443286039,-0.2599350214,-0.1368361413,0.2961815298,-0.5483995676,0.1319540888,-0.5378554463,-0.0621335991,0.2043301016,0.1647994667,0.0299274754,-0.0076469467,0.1783820093,0.1878955066,-0.0386595689,0.0143071227,-0.2300728112,-0.2202730328,-0.1764260679,-0.3641079068,0.3547821045,-0.0295607224,-0.4267731309,-0.0369149931,-0.0573267229,0.1431887001,0.0223086029,0.3342297971,0.4058888853,0.0438097678,-0.1033032015,-0.1444998682,-0.1450365186,-0.0960467011,-0.7143990397,0.251899004,-0.2774967253,-0.2620482147,0.1847258508,0.1000009775,0.5091559291,-0.0343831629,-0.2110319436,0.0629937127,0.0245486014,0.4867337346,-0.0485542752,-0.082927309,0.5620526671,0.150778845,-0.010695219,-0.224782303,0.0754697993,0.2946532071,-0.0628259778,0.2438559681,0.0527600832,0.0985681787,-0.1168366894,1.0462340117,0.2199911028,-0.0246647038,-0.017103808,0.1474039704,0.5294182301,0.1580959111,-0.3583490849,0.0158541966,-0.0138802566,-0.010371536,0.3552688062,0.0933368653,0.0616519749,-0.0682541281,0.005089866,-0.5711330175,-0.2430211157,-0.3952088058,-0.2392047495,0.3242174387,0.3030349612,0.3546321988,-0.1379886568,-0.0496442392,0.1588335633,0.160614863,0.4409466982,0.0526122898,-0.0696622357,0.3064370155,-0.5714339614,0.0200432427,0.2249084115,0.1337825209,-0.1148483008,0.337947309,0.0607433617,-0.0145724183,0.3146234453,-0.1851236373,0.2493982762,-0.2188224047,0.1371629536,0.0563672595,0.0568645075,0.0189098902,0.1150910333,0.3641639948,0.5256968737,-0.1791980714,-0.079407908,0.2453015447,-0.191664353,-0.1595643461,-0.2109298408,-0.0154957417,-0.1973309219,0.0537550561,0.0443392843,-0.0606972612,0.219106093,-0.5079624653,-0.3269940317,0.3720759153,0.2891472876,-0.1736972183,-0.0655642524,0.1174465418,0.1101504192,0.3324817121,0.7585226297,-0.0604214743,0.6344255805,0.1879083216,-0.0318491012,-0.5295669436,-0.3288809657,-0.2241621763,-0.1451194137,0.0828752443,0.1613499373,-0.1453902721,0.0907882601,-0.3072914779,-0.0827849656,-0.1839219481,0.4555115104,0.1427202076,-0.4234263003,-0.2619732022,0.001380153,0.1943770647,-0.1145445332,0.6239343882,0.501647234,-0.2544493377,-0.2650600672,0.3341426551,0.7262828946,-0.27815938,-0.1264921427,-0.1478276253,-0.5781803727,0.5295581222,0.3364044726,-0.0349686816,-0.5804929137,-0.4340507686,-0.0510880984,0.0124363229,0.402118504,0.0846430957,-0.381848067,-0.0412028991,0.0665477067,0.2690662742,0.2535394728,0.0430560745,-0.3380262256,-0.2376369536,-0.1506433636,-0.0272215884,-0.3198177516,0.1685468853,-0.190520525,-0.1722947508,0.0657750443,-0.0054038693,-0.4023111463,0.0775434822,-0.0577157289,-0.1284455657,-0.0067519085,-0.054662101,-0.4198946655,0.4081673026,0.2419962138,0.0624142215,-0.3137251139,0.2111720592,-0.0085237464,-0.073293142,0.011095169,0.0891371742,0.2499184012,-0.0027726328,-0.3028775156,-0.1581942588,-0.209266454,-0.1547766328,-0.4880754948,-0.3939775527,-0.2473040819,0.1327327341,0.0370710492,0.0457210839,0.2262032628,-0.2979454994,0.0354091562,0.1603600383,0.0270033441,0.0302371383,-0.135875687,-0.1697162539,0.0447121039,0.3609546125,0.0797510594,0.3192098439,-0.0592463352,-0.3930160403,-0.0248697475,-0.0414361991,0.2886120975,-0.0445295759,-0.4130453169,-0.0649765953,-0.3906111419,0.1708176434,-0.1037158296,0.2379667461,0.2758376598,-0.2019821256,-0.1730237901,-0.3063979447,-0.157796219,0.157265678,-0.1063341126,0.0400743335,0.2192181051,-0.0518492348,0.1225078106,-0.0750644207,-0.1868622154,-0.1019912735,0.080644317,-0.1005298942,0.0792645961,-0.0451880433,0.0503879748,-0.2418504357,-0.0996202454,0.0135644935,-0.1612738371,-0.0191908944,-0.1587375551,0.205520913,0.1004211009,0.2183869928,0.2726885378,-0.2261933982,-0.3207331896,0.0214371961,-0.0412582196,0.5195782781,0.1747218221,0.0571518131,0.2004870325,0.5155271292,-0.2896468639,0.0241129529,0.0911205187,0.3099288344,-0.1200852245,0.3679488301,-0.1996873766,-0.1086001024,-0.252487272,0.0549876504,0.0658718348,0.0925075486,0.3490975499,-0.3229043782,0.0494440459,0.4477138817,0.6423042417,-0.2155703306,0.0504370555,-0.0551803522,0.0253970977,0.1495009959,-0.2573877275,-0.1280942708,-0.0500715412,0.0550650917,-0.140488416,0.2167467177,0.1621718407,-0.3948625624,-0.0427187122,0.0567936189,0.3184597492,-0.1251234859,0.0826749578,0.5675780773,-0.0376206152,0.0553092323,0.4317943752,0.0260594748,0.0141377794,-0.030945994,0.0548487231,0.2732615769,0.6133189797,0.4035389125,0.2087519467,-0.0298635941,-0.0704298392,-0.1106907576,0.0568301193,0.050235983,0.3190507889,0.1692847162,0.0138792247,-0.3573317826,-0.1901688874,0.1133165285,-0.0108573511,0.0475690961,-0.0060642166,-0.3562538028,0.0854572356,0.0601430163,0.0256296489,0.0280811228,0.2432371825,-0.1658055782,-0.3811866641,-0.8119903207,-0.1659934074,0.0658104718,0.1767832786,-0.1928747147,0.2468496412,-0.3664066195,-0.222312957,-0.0509409942,0.8055132627,0.0757440552,0.042016618,0.4883258939,0.0039262827,0.1536893994,-0.0532319844,0.1195724085,0.2071630955,0.3157704771,-0.0568188354,-0.0037859636,0.0788448229,-0.0448380262,0.0302524809,0.5185053945,0.3099989891,0.0290192254,0.0625400022,-0.0657797828,-0.1486612707,0.0338542126,0.3108851612,-0.2482420057,-0.0159011018,0.041778978,0.0991105288,-0.0912514031,-0.2822519839,0.0414645746,0.0347700566,-0.1620167643,0.3209408522,0.0801597014,0.2211478204,-0.2779092491,-0.5337411165,0.3785619736,0.2400203794,0.1645997018,0.1464401782,0.0233428907,-0.065316014,-0.0679498687,0.3363083899,-0.2585842609,0.2947355509,0.248321712,-0.308590591,0.0111937914,-0.1491391659,-0.2218487859,-0.2559011281,-0.1564924866,0.3803358972,0.1333687603,-0.0102952849,-0.0034805692,0.0294197705,-0.0112862084,0.1360284835,0.2952550948,0.1193813533,-0.0872371495,-0.2152346075,-0.2024948001,-0.2422470301,0.0215569101,-0.4104663134,0.4260915518,0.0285298415,0.2173588723,-0.052106563,-0.1398037672,-0.1788336635,0.2281909883,0.146904856,0.0619119219,-0.5954897404,0.0515131429,0.014974799,-0.6247376204,0.0247276425,0.2606160343,-0.0195604078,0.2545497715,0.0318338647,-0.0379194617,0.3640733957,-0.41009444,-0.0422835425,-0.0495893583,0.309441328,0.405380249,-0.5434498787,-0.4110452235,-0.0000897353,0.3572213948,-0.0869887918,-0.0231014062,-0.0967688039,0.0322583541,0.1769705415,-0.0780812427,0.1460570842,0.1366180032,-0.228455171,0.18690148,-0.0517965592]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2472","title":"Fix automatic generation of Zenodo DOI","comments":"Zenodo issue has been fixed. The 1.8.0 release DOI can be found here: https:\/\/zenodo.org\/record\/4946100#.YMd6vKj7RPY","body":"After the last release of Datasets (1.8.0), the automatic generation of the Zenodo DOI failed: it appears in yellow as \"Received\", instead of in green as \"Published\".\r\n\r\nI have contacted Zenodo support to fix this issue.\r\n\r\nTODO:\r\n- [x] Check with Zenodo to fix the issue\r\n- [x] Check BibTeX entry is right","comment_length":14,"text":"Fix automatic generation of Zenodo DOI\nAfter the last release of Datasets (1.8.0), the automatic generation of the Zenodo DOI failed: it appears in yellow as \"Received\", instead of in green as \"Published\".\r\n\r\nI have contacted Zenodo support to fix this issue.\r\n\r\nTODO:\r\n- [x] Check with Zenodo to fix the issue\r\n- [x] Check BibTeX entry is right\nZenodo issue has been fixed. The 1.8.0 release DOI can be found here: https:\/\/zenodo.org\/record\/4946100#.YMd6vKj7RPY","embeddings":[-0.1276472062,0.3648432195,-0.0427521355,0.086929135,0.1287777722,-0.1165491268,0.350559622,0.4325045347,-0.1157745495,0.2650321722,-0.0857932866,0.1701208949,0.156162262,-0.0318087302,-0.1042906716,0.0138034197,-0.0821277276,0.3472718596,-0.0333306268,-0.3059482574,-0.2411757261,0.1041502431,-0.0310633592,0.0592511371,0.0967376083,-0.0133123342,-0.1155754551,-0.002625301,-0.6373178959,-0.2898760736,0.2117006928,0.3862249851,-0.0122202951,0.091927126,-0.0001115753,-0.2179564983,0.3990101814,-0.0613058619,0.1894159913,-0.3993172944,-0.4966446161,-0.3368517458,-0.1118820235,-0.2210412771,0.1687878668,-0.344227463,0.1168469936,0.1707206368,0.0785860419,0.4088201523,0.2324048579,-0.0964877233,0.462238133,-0.2651804388,0.1193558872,0.0738097057,-0.0664754435,0.0427458249,-0.1232345402,0.3147010207,-0.057273332,0.2952130139,0.0615880117,0.1639275551,0.0338265076,0.0791369379,0.2854732573,-0.2386750281,0.1627262831,0.1508211493,0.8272175193,-0.1040101573,-0.2112257928,0.1218621209,0.1969872415,-0.2398520559,0.0999539196,0.0821903273,0.1315313578,0.1152198911,0.1441225708,-0.2169701457,-0.2635707855,0.0223813057,0.072737433,-0.0621657036,-0.0755716786,-0.0292691458,-0.0162154492,-0.132941708,0.3126554489,-0.0463040359,-0.2404833138,0.2261698246,0.2291472852,-0.056270808,-0.0226141866,0.1099060252,0.3078130484,0.175941214,0.2989215255,0.0299286619,-0.0712770671,-0.0983816832,0.1395368725,0.0053117326,0.376881063,-0.0843499005,0.2142859548,0.0713581368,-0.0579156578,0.0687342286,0.0742444992,-0.1007361114,0.5671770573,-0.163455829,0.3139571846,-0.4937621653,-0.0736599043,0.127096355,0.0081508672,-0.1122738048,-0.4699214399,0.0171350203,-0.0131240822,0.333891809,-0.0442627221,0.0036914533,-0.2457430065,-0.1644554585,-0.137726143,-0.4284023046,0.0954869688,0.1792787015,0.0250738505,-0.3561386168,0.2302707881,0.1848700643,-0.1593869627,-0.510985136,0.0842346326,0.1982409209,0.2018585354,0.2626516819,0.0948756039,0.1765399873,-0.0657955334,-0.1015712991,0.0033277888,0.286929667,-0.1481172293,-0.430593133,-0.5306661129,0.1735018343,-0.0043724887,-0.0513049178,0.4339270294,0.220164299,0.1558127105,-0.1584343314,-0.1603408158,-0.0570722781,0.1968234628,0.1039189845,0.1392633021,0.2979480326,0.1394583136,-0.0364029557,-0.4480935037,0.0200784337,0.3247521222,-0.0593138188,-0.0139837293,0.2366739959,-0.084498696,-0.3633298576,0.2564738393,-0.4945763052,-0.3621280193,-0.0594122522,-0.0666056052,-0.0418133661,-0.0438909009,-0.1598453522,0.3066423833,0.2594448924,-0.307653904,-0.2543961704,-0.0933055729,-0.0350184962,-0.2658015788,-0.1762033105,-0.2635964751,-0.0920261145,0.6148753762,0.3617994189,0.1732919067,0.2241190374,0.2010513246,0.23364757,-0.014715868,0.0965362564,0.5175153017,-0.2195931524,0.2518333793,0.0172665454,-0.0379066318,0.0607143231,0.3674999475,0.1227196157,-0.0883978531,-0.0770754069,-0.3192293346,-0.1353597492,-0.1683202833,0.0033035504,0.1431188285,0.0637131035,-0.0676442981,0.1566046327,0.0798380971,-0.0825694129,-0.0608584099,0.1678389013,-0.0745825842,0.1074951962,0.0233037975,-0.119926706,0.0831123218,0.0723050237,0.0739945024,0.1868750602,-0.2001238614,0.2543473244,0.0713327304,0.0758610964,0.1153724566,0.3495374024,0.2024537325,-0.2067794353,-0.0913805887,0.1121269315,-0.2016371787,0.1503857821,0.0193529651,0.3028654158,-0.2581292689,-0.3044906259,-0.3119139373,0.2759441733,0.5370331407,0.0420377739,-0.2956937551,-0.4251458943,-0.1063094139,-0.1361320764,0.1804952919,0.1569513381,-0.3323892951,0.2972870171,0.4421213269,-0.0406733565,0.0071021589,-0.1665102094,0.0967135653,-0.1172703281,0.1063665226,0.3659424782,0.1049696803,0.2018118501,-0.0775375143,0.1291189045,-0.242711097,-0.2093186378,0.1573490053,0.0089320084,0.5699402094,0.3031816185,-0.0175450332,-0.200857833,-0.5125389099,0.2591793835,0.03345377,0.4638356566,-0.1276204288,-0.5851067901,-0.0330619514,0.2312393934,-0.3544933498,-0.3642880917,-0.3784945905,-0.2268044651,0.1120157763,0.2224253565,0.060370218,0.1534920782,-0.0513778254,0.105823651,0.0270165429,0.320928514,-0.1337502301,0.0215769708,0.0615992658,0.0120056635,-0.1958386153,-0.1157170534,0.3026033342,-0.4803004861,0.1888469011,-0.5613209605,-0.149362877,0.1899141818,0.1756080389,0.1328714937,-0.1659779251,0.0643382072,0.1064553186,-0.1003607437,0.0286189802,-0.3231865764,-0.1802284122,0.0990355462,-0.4877848923,0.2138941288,-0.0612735786,-0.4875466228,-0.0362592079,-0.0642827675,0.0240284,0.0769697428,0.3351006806,0.3139775097,0.0667185336,-0.1480328143,-0.1310467124,-0.2920634151,-0.1410773247,-0.6531655788,0.2963385582,-0.3593808711,-0.4324019253,0.172202751,0.0744459257,0.4255275726,-0.0696390867,-0.1786551178,0.2332183123,0.1021687388,0.5111603737,0.116311118,-0.2305387408,0.4898826778,0.1118389145,-0.0729694292,-0.2673332989,0.2693287432,0.3997789919,-0.0066460101,0.3154290318,-0.1744367182,-0.0175566506,-0.13910155,0.7804270983,0.0574707761,0.0086837523,-0.075068444,0.2255037576,0.50721699,0.0230322834,-0.1232248321,0.0879861116,-0.0129749076,0.0649964064,0.4969039559,0.0448021926,-0.2233751714,-0.0695924461,0.1194937527,-0.5448397398,-0.2431103885,-0.4036025405,-0.4276815951,0.2482360899,0.3956670165,0.2334257811,-0.2130814493,-0.1863588691,-0.0078709489,0.0228246469,0.277697593,0.0364426523,-0.0817918107,0.3758344054,-0.4816722274,0.063466318,0.0330315903,0.2390162349,-0.0362182781,0.2812831104,0.0978709683,-0.0557036884,0.2008822113,-0.2632823288,0.2095757574,-0.1298655421,0.1259958148,-0.2558363676,-0.0908586979,-0.1436152309,0.0720295385,0.2136502713,0.4778341651,-0.0479973853,-0.028459996,0.1839702725,-0.2804489732,-0.1949248463,-0.3254898489,-0.0907830074,-0.1857114434,0.1075779796,0.0581268258,-0.2146456838,0.1260564476,-0.6979616284,-0.3720459938,0.2242470831,0.3119070828,-0.1482236534,-0.0498416051,0.1460080296,0.0376822203,0.3962549865,0.6873075962,-0.0542165935,0.4746967554,0.0132583492,-0.0485315286,-0.5039690137,-0.1862262785,-0.3990829587,-0.3170873821,-0.0492913835,0.1738660783,-0.1319037378,0.0373215899,-0.2970101237,0.0449730828,-0.1989735365,0.3888956308,0.0960109979,-0.3479038179,-0.214418292,0.1044373512,0.124587357,-0.1559471041,0.6445992589,0.1790319085,-0.1340746284,-0.0063656732,0.3804314733,0.5535769463,-0.3415917158,-0.2419618815,-0.0329730883,-0.4711416662,0.2819353938,0.2047358006,0.0031721077,-0.6027874351,-0.4457513988,0.0191165674,0.1090603247,0.3984047771,0.1705501378,-0.3323077559,-0.1714204699,-0.0903653502,0.1975288987,0.2747181952,-0.043683812,-0.255666852,-0.1554715484,-0.0288094338,0.1028524488,-0.3360681832,-0.0407053195,-0.1792011112,-0.0547282286,0.096031785,0.0064624948,-0.2027204484,0.1097486317,0.1656971723,0.0970922038,-0.1256292611,-0.0499522761,-0.6244404316,0.4581741095,0.2410417944,0.061817646,-0.4071315229,0.2993488908,0.170957908,-0.1185935885,0.0205721874,0.1498745531,0.2394733429,-0.0647610277,-0.3887230158,-0.1777350903,-0.105949752,-0.1578865498,-0.3984579742,-0.3985873163,-0.1918846816,0.2266168296,0.073307775,0.0417781323,0.1314663887,-0.2976039052,0.1292819232,0.1350285411,-0.0436460935,0.2026649863,-0.0362062529,-0.2262427062,0.0000524638,0.4080985188,0.2826412022,0.4951074421,-0.0488550104,-0.3336983919,-0.0281085391,-0.1504890174,0.3042360842,-0.0586090274,-0.2590529323,0.0364844501,-0.3464644551,0.209828198,0.0433660895,0.2139803916,0.1353705823,-0.3471018672,-0.280957669,-0.2155943215,-0.1246607453,0.0563825183,-0.1102892682,0.0467658751,0.1846394986,-0.0474021435,0.0704074204,0.078160964,-0.268416822,-0.0441057608,0.0733289644,0.0042425105,-0.0351321921,-0.0920088068,-0.0806460977,-0.2092384398,-0.0177244265,0.0462034382,-0.1873684824,-0.1392490417,-0.2018140703,0.1317555755,0.2358716279,0.1725281775,0.258739531,-0.1984753907,-0.4427166581,-0.0447735302,-0.1312068254,0.4628528953,-0.0233721789,0.1427282393,0.0910133421,0.3948186338,-0.0375929587,-0.0058050109,0.0555652902,0.2120227814,-0.0996396691,0.3358250558,-0.1322316229,-0.0855701938,-0.1518844068,0.0321695805,-0.1317867041,0.1728297621,0.4285942316,-0.42364344,0.2226333767,0.272124052,0.4805395901,-0.1615103334,0.0856857076,-0.1848362088,0.008851367,0.2487794906,-0.348949343,-0.1194930449,-0.0544702336,0.1272089183,-0.1152218804,0.1110468134,0.2049230784,-0.574103713,-0.1160142347,0.1097180024,0.2623594999,-0.2014755905,0.1227218583,0.5477483869,-0.0354107991,0.0969160795,0.4338223636,-0.073665157,-0.0659021288,0.1852622032,0.2381290197,0.2949497998,0.4836460054,0.3265836537,0.0581210181,-0.0607255585,0.026793465,-0.1633801609,0.034418121,0.0674159825,0.2874881029,0.0709676445,-0.0146143446,-0.2789024711,-0.3274447024,-0.0035672009,-0.0053841723,0.0726771131,-0.1098152474,-0.3980038464,0.2318462282,0.0902956799,0.0472747087,0.1393273771,0.1983374059,-0.021959519,-0.3211120963,-0.8208305836,-0.0377119407,0.0473385192,0.1769637316,-0.2228484154,0.1969902068,-0.3605412841,-0.1977749765,0.0153390402,0.6312586665,-0.0204497781,0.0984817594,0.3875378966,-0.0577998534,0.1479715109,-0.1240315363,0.1016149223,0.2667052448,0.3934873343,0.158857733,0.1036819965,0.1431794018,-0.0728524476,0.0356105044,0.6348097324,0.3096057177,0.1589226425,-0.1455379128,0.1285209209,-0.1357479841,0.0712423176,0.1545196623,-0.2094681114,-0.1215065718,0.0602411889,0.0655084103,-0.0326608084,-0.2892557979,0.0790040046,0.0230956171,-0.0666370317,0.1140737459,0.068483904,0.2202108204,-0.3401423395,-0.5243164301,0.4213830829,0.1081629395,0.329716742,0.1140104905,0.1106463298,-0.0590967089,-0.1140896156,0.3103164434,-0.1914676279,0.1593452245,0.2541580498,-0.4248420298,0.0057239062,-0.1420126557,-0.1397618949,-0.2959055603,-0.2360798419,0.3261331022,0.1200800985,0.0870226473,0.0590475164,-0.061032299,0.0888387859,0.1842132956,0.3237660229,0.1706321836,-0.0187726393,-0.1915982366,-0.1493485421,-0.2527066469,-0.0536822118,-0.3470336795,0.3674574494,0.1027021557,0.282192111,-0.0223733168,-0.0556328334,-0.0258126203,0.2936146855,0.041676566,0.0720971748,-0.6444922686,0.0156506579,0.0374180786,-0.511838913,-0.0982788056,0.2026281953,0.0300405119,0.2619768679,0.0259410404,0.0269795414,0.2792870998,-0.3196246028,-0.0680820793,-0.1405108571,0.2971508503,0.3638842702,-0.4150417447,-0.2244248241,-0.0067104367,0.3242398202,-0.1983854324,-0.0049404949,-0.1073397771,0.0074155806,0.1920620054,-0.0906027257,0.1941300929,0.0777634531,-0.1618403047,0.1436613351,-0.04040787]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2470","title":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.","comments":"Hi ! It looks like the issue comes from pyarrow. What version of pyarrow are you using ? How did you install it ?","body":"## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n","comment_length":24,"text":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.\n## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n\nHi ! It looks like the issue comes from pyarrow. What version of pyarrow are you using ? How did you install it ?","embeddings":[-0.4569717646,-0.2173907459,0.0228779204,0.4219461381,0.3113077581,-0.2264823467,0.2361344099,0.1525053084,0.0431262851,0.3643293679,0.5303300619,0.2979793251,-0.3346508443,-0.0465700552,0.047729928,-0.0265675541,0.3659133613,-0.2182355374,0.2511509061,0.3584368825,-0.2708789408,0.1627700925,-0.4191713631,0.2937121987,-0.2652947009,0.082243599,-0.1050226912,0.2865630984,-0.0974157006,-0.5258919597,-0.0305760354,-0.295525372,0.1912617087,0.4583674669,-0.0001182336,-0.0913553163,0.2139718682,0.0120125022,0.0930700898,-0.4430084825,-0.2259867936,-0.2528820932,-0.1637293845,-0.0408718698,0.3133925498,0.0046791183,-0.0663942918,-0.150278613,0.0759297609,0.7205310464,0.1591092348,0.214832738,0.2288711667,-0.1120013967,0.016339317,0.1341472566,-0.0305977836,0.061437469,0.4003654718,-0.2003950328,0.0249149706,-0.1673947722,0.092808485,-0.2669810057,0.305939585,-0.1227102131,0.1998042613,-0.6616144776,0.2974792719,-0.1724279821,-0.1635968983,-0.2878350317,-0.0509196632,-0.1632319391,-0.1430846751,-0.2230149657,0.177413553,-0.1671895236,-0.1571241021,0.0832860544,-0.2861917019,0.1699507236,0.0335625187,0.3086440563,-0.3192946315,0.4002259374,-0.0029774297,0.389618665,0.1262956411,-0.2813790739,-0.3420287967,0.1340495944,-0.0415284634,0.2674230039,-0.2257158607,-0.1663678735,0.1507905126,-0.1738581359,0.2904581726,-0.2406045794,-0.1656422466,0.2874025404,0.048612088,0.2133101672,0.2980216742,-0.0209573247,0.3021420836,0.3840674162,0.0499104597,-0.056064833,-0.102803044,-0.0400411636,0.2264666557,-0.1167617217,0.0892518535,0.1257736236,0.0740240142,-0.1534793824,-0.1198020428,0.4294593036,-0.2137423903,0.2773050666,0.3186097741,0.2963832617,0.3739162087,0.227028951,-0.2862907052,0.0338729285,-0.2570652664,-0.0871776044,-0.0442779921,-0.2006773353,-0.3708454072,0.0992647409,-0.005736412,0.239634648,0.1527641416,0.0450480506,0.2212321013,-0.539868176,0.1125220954,-0.2573070824,0.1339917481,0.0350833498,-0.2941997945,0.1408175528,0.1156675741,-0.1236832365,-0.1660993546,0.1671806574,-0.2124153674,-0.2457581609,-0.1474966705,0.1560774744,-0.0041650292,0.4375189543,-0.0617898293,0.3228711784,0.3464185596,-0.3285045028,-0.0258732624,-0.3166899085,-0.4711762369,-0.2692376971,-0.0478229038,0.2325534225,-0.1191353053,0.0298826676,0.2000778913,0.0202775728,0.0941302255,0.1629669219,-0.1979359239,-0.0177588277,-0.1911362857,0.5243681073,0.2797337174,-0.0643637404,-0.4906041324,0.1102769822,-0.2673051357,-0.1981518269,-0.4909549952,-0.1560129821,0.4023430943,-0.2073682845,0.4260299802,-0.0168935806,-0.1412145048,0.118385382,-0.306717515,0.094998017,0.1623565853,-0.1462418735,0.0806366652,-0.166501984,0.1831175238,-0.0708336085,0.0203282014,0.004628906,-0.0426518135,0.2887397707,0.2136349082,-0.0781941861,0.0351790115,-0.3048202991,-0.0683804154,0.2268817574,0.0347599015,-0.0888614431,-0.0282383524,0.0105143087,-0.0411652811,0.3073398471,0.1391114891,-0.1058630794,0.0216380805,-0.1246979013,0.0030890247,-0.0661532804,-0.0376062728,-0.0197845213,-0.1885098517,0.0246681049,-0.0632356256,-0.1272604764,-0.2010300606,-0.1646624357,-0.3887704015,0.1416743398,0.4239653945,-0.0064662574,-0.2309417278,0.5506316423,0.3362869918,-0.2484891564,-0.6358670592,-0.0847648457,0.3994875848,-0.1538935304,0.1634524167,0.0869446471,-0.0126662534,-0.1323807985,0.1911210567,0.2216383666,0.2824165821,0.1572643816,-0.0959971696,0.0425383486,0.1659151316,0.1229683161,0.0325946435,-0.199791953,-0.0025342163,0.1199669093,0.0561050922,-0.047854811,-0.0479735769,0.2361027151,0.2793250382,-0.0761121586,0.1238081977,0.0580437221,-0.1685209721,-0.0567469634,0.2826824784,-0.0136068668,0.6230338812,0.0613569953,-0.0287729017,0.0040426482,-0.3801935911,-0.0336282663,0.1335722506,0.0782550871,-0.1228369474,0.0503613725,0.2556922138,0.159722358,-0.1924543381,-0.3795136213,0.0199823845,0.3146329224,-0.4612223506,0.2040783465,-0.1188221872,0.2136378586,-0.0778881684,-0.1706882864,-0.1753237247,-0.3915620148,-0.0386055,0.1213032827,0.0050780186,0.1582536548,0.0887682661,-0.1417624056,0.04338919,0.3126336932,0.2620718777,-0.0042142225,-0.4652619362,-0.1256633401,0.3646590412,-0.1426314861,0.1807572693,0.1596735418,-0.4231134653,-0.1377694905,-0.554107368,0.0362168029,0.0458122902,0.2179297954,0.0650527328,0.0002729307,-0.1417898685,-0.0640055388,0.1627510339,-0.0710154772,-0.0511523709,0.1167618632,0.1763634235,-0.3484143615,-0.4004407525,0.0810818374,-0.2404452562,-0.2278887779,0.0398980491,0.0562176853,0.2899165154,0.033601027,0.3173662722,0.2378733456,0.0640233308,-0.2699205875,-0.2033462822,0.2120094001,-0.0509562194,-0.1475808471,-0.2852531075,-0.0752577856,0.1772031188,-0.022968391,0.2650331259,-0.2499954402,-0.0535250045,-0.0659198835,0.3023351431,-0.1044890657,-0.2844492197,0.3918755949,0.1078824848,0.0479716472,-0.0043216893,-0.3812083304,-0.0652099252,-0.2201593518,0.1879525185,-0.0040869415,0.5207585096,0.1579419822,0.5066372156,0.3318142593,-0.5369669795,0.1717587262,-0.2706901133,0.1213353053,-0.1566914618,-0.1280220896,0.3013108969,0.0409448743,-0.0624374636,0.136187464,-0.0629640371,-0.4967781901,0.0518860146,0.0803500637,-0.0168010574,-0.2850890756,0.1508056819,-0.0602970831,0.2637024224,-0.1313890517,0.3329266608,-0.2499023676,-0.0794402286,-0.1635815203,-0.0973349959,-0.1765634567,-0.0466502458,-0.534514606,-0.1086725891,-0.6624104381,0.3221202195,0.1939269751,0.7835565805,0.0214179475,-0.0824445039,-0.2468029261,-0.1681991518,0.6353745461,-0.300203532,0.0723909363,-0.0045123673,0.098818697,-0.6122601032,-0.1256724149,-0.0143848732,0.6673847437,0.2643587291,0.9284402132,0.0463600494,0.045960132,0.0434808359,0.0178261716,0.1433940679,-0.0652095973,-0.0105038164,-0.0599536635,-0.1763828248,0.1956971437,0.2371133864,0.207475841,-0.0157400426,-0.4159704745,0.0754318386,-0.3766656518,0.2011009455,0.1319937855,0.208466664,-0.2982677221,0.1140266657,-0.0784145966,-0.1573798358,0.2803728878,-0.1881632805,0.0264791045,-0.3439128697,0.2833440304,0.1530024409,0.1995321661,0.426977247,-0.0416696221,0.1061635017,-0.3421507776,0.3123049438,-0.2047104388,0.2691136599,0.2535805106,0.2167814076,-0.5693432689,0.0799898058,0.1540472656,0.0905904621,-0.0760773271,0.3940853775,-0.368170321,-0.4619640112,0.3434622586,0.1300132424,0.848302424,0.0364213064,-0.081603311,-0.19214347,0.1661395729,0.1686700284,-0.1858166009,-0.0507305525,-0.3904622197,-0.1557223946,-0.0677676797,-0.1420495659,0.0983293951,0.0727663487,-0.3925172985,0.1129100174,0.1099763438,0.4424754381,-0.0006535093,0.094351247,-0.1428740174,-0.1273342818,-0.0686014518,0.1570535004,0.1238408089,-0.2321622223,-0.14562352,0.3469045758,0.1431711614,-0.4613877833,-0.62011379,-0.0647350997,-0.5778717399,0.1768615544,0.1812040657,0.2300380021,0.1644968092,0.301273495,0.0709967613,0.506118238,-0.2116412073,-0.0245072097,0.0392306633,0.0512521602,0.1469227374,-0.1345072687,-0.0176817924,0.0910632312,-0.0327797458,0.0640930757,0.001057488,0.1164323166,-0.501894176,0.265417248,0.4029581249,-0.196479097,-0.2089856416,-0.2772096992,-0.1896944642,-0.2230817527,0.0718460232,0.0842439681,-0.1488311887,0.1394552439,-0.1531652212,-0.2392156124,0.0301355962,0.4575230777,-0.2256522924,0.0981041044,-0.0154942367,0.0785034224,-0.0677308589,-0.1901301593,0.2326962501,-0.1237555444,-0.0325517394,-0.2231186181,-0.2007786334,-0.0984243006,0.1080042869,0.0606934577,0.1129072234,0.0262527335,-0.016009355,-0.5448374152,-0.4919656813,-0.0073139649,-0.0454394147,0.3145165741,-0.0165020507,0.0936865062,-0.2128308862,0.1364151239,-0.2179399729,0.1666069627,-0.0632879138,0.2086213976,-0.0276215877,0.0983446389,-0.1632549614,-0.1979133934,0.0605035052,0.7531200051,-0.2359837741,-0.0942862928,-0.0525906458,0.1379344016,0.2398593724,-0.1094310135,-0.1033161655,-0.0272614323,0.1039245501,-0.2421072721,0.1028843969,0.2268958539,-0.2034132034,-0.131596297,0.6177958846,0.3815800548,0.1931364983,0.357540071,0.0191768613,0.2638332546,0.1465423852,0.3621217012,0.1001960859,-0.0010375903,-0.0883050188,0.2816980779,0.0660891756,0.1587927341,0.2086260915,-0.0950449184,0.1765199453,0.0912955999,-0.0393662527,0.3638717234,-0.4002837837,-0.0639449432,0.191659078,0.1843820661,0.0324680954,-0.2394108027,0.3049276471,0.1877780557,0.0064044776,0.2658642828,0.0362890139,-0.3785454035,0.079316467,0.2469057888,0.3529776633,-0.0388915651,-0.0370597355,0.2587002814,-0.2121544927,0.1043223813,0.2033436149,-0.1946918219,0.3131908774,0.6266698241,-0.0740214512,0.4274179339,0.1257343292,0.0630890504,-0.0054944996,-0.2483475804,0.0808186531,0.4405095875,-0.1837271601,0.1500857174,0.0620873384,0.257960856,-0.3250874877,-0.1628504544,0.0823815241,0.0784967542,-0.3235331476,-0.0777549371,-0.0641655922,-0.1524268389,-0.1851560324,0.0298068896,-0.1491112709,-0.134274587,0.3058646917,0.105733268,-0.3869328499,-0.2465470433,-0.0842965916,0.0462458469,-0.0254337527,-0.2964484394,0.5547583699,-0.165271312,0.2051642239,0.1515772492,-0.1810009629,0.5931897759,0.6459059119,-0.2287395895,0.1645237505,0.1871111542,0.0369282775,-0.0236070659,0.1389949471,0.0064930003,0.2090339661,0.2872698605,0.0319978818,-0.1442494094,-0.1472728252,0.3358566463,0.2155430019,-0.0512180105,0.1997416914,-0.2628280818,-0.1870484501,0.0019922892,-0.0307528339,-0.1691651791,-0.0553508326,0.3815113008,-0.0056802407,0.4061158597,-0.1410945505,0.0556682125,0.1124868616,0.165840745,0.2160716504,-0.1539394557,-0.4025766253,-0.3473495245,-0.3931940496,0.3725878298,-0.2325892448,0.1685742289,0.0681374967,-0.2374645174,-0.0817601308,0.003409601,0.1784852594,0.2687878013,-0.0693608075,0.5016827583,-0.3206634223,-0.0434216745,-0.0729406029,-0.1339516491,0.2401047647,-0.4375864863,0.4043382704,-0.3041529357,-0.0143265808,-0.1915714145,0.0053744041,-0.227931276,-0.0372267179,0.6843011975,0.2150361538,0.4069308043,-0.0329385772,0.1105626151,-0.2053319067,0.1099141538,-0.2095181644,0.2924145162,0.0320701674,0.1326195002,-0.1456784308,-0.1236873269,-0.072567299,-0.1102667898,0.2603496611,-0.0558728129,-0.2727081776,0.1199415252,-0.2002013028,-0.0307991542,-0.1286313683,0.3280451596,0.1533139497,0.2899936736,0.0323429331,-0.2277695686,0.3998816311,-0.3358049691,-0.3362597823,-0.1846985966,0.1810190976,-0.2116550505,-0.3600912392,-0.5163121819,-0.3579574823,0.3601654172,0.0267328806,-0.5006997585,-0.0882551745,-0.2791074812,-0.0996797755,-0.2518528104,0.4753126502,0.1529987752,-0.1394924521,0.6449896693,-0.3258785605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2470","title":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.","comments":"Thank you for the quick reply! I have `pyarrow==4.0.0`, and I am installing with `pip`. It's not one of my explicit dependencies, so I assume it came along with something else.","body":"## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n","comment_length":31,"text":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.\n## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n\nThank you for the quick reply! I have `pyarrow==4.0.0`, and I am installing with `pip`. It's not one of my explicit dependencies, so I assume it came along with something else.","embeddings":[-0.4569717646,-0.2173907459,0.0228779204,0.4219461381,0.3113077581,-0.2264823467,0.2361344099,0.1525053084,0.0431262851,0.3643293679,0.5303300619,0.2979793251,-0.3346508443,-0.0465700552,0.047729928,-0.0265675541,0.3659133613,-0.2182355374,0.2511509061,0.3584368825,-0.2708789408,0.1627700925,-0.4191713631,0.2937121987,-0.2652947009,0.082243599,-0.1050226912,0.2865630984,-0.0974157006,-0.5258919597,-0.0305760354,-0.295525372,0.1912617087,0.4583674669,-0.0001182336,-0.0913553163,0.2139718682,0.0120125022,0.0930700898,-0.4430084825,-0.2259867936,-0.2528820932,-0.1637293845,-0.0408718698,0.3133925498,0.0046791183,-0.0663942918,-0.150278613,0.0759297609,0.7205310464,0.1591092348,0.214832738,0.2288711667,-0.1120013967,0.016339317,0.1341472566,-0.0305977836,0.061437469,0.4003654718,-0.2003950328,0.0249149706,-0.1673947722,0.092808485,-0.2669810057,0.305939585,-0.1227102131,0.1998042613,-0.6616144776,0.2974792719,-0.1724279821,-0.1635968983,-0.2878350317,-0.0509196632,-0.1632319391,-0.1430846751,-0.2230149657,0.177413553,-0.1671895236,-0.1571241021,0.0832860544,-0.2861917019,0.1699507236,0.0335625187,0.3086440563,-0.3192946315,0.4002259374,-0.0029774297,0.389618665,0.1262956411,-0.2813790739,-0.3420287967,0.1340495944,-0.0415284634,0.2674230039,-0.2257158607,-0.1663678735,0.1507905126,-0.1738581359,0.2904581726,-0.2406045794,-0.1656422466,0.2874025404,0.048612088,0.2133101672,0.2980216742,-0.0209573247,0.3021420836,0.3840674162,0.0499104597,-0.056064833,-0.102803044,-0.0400411636,0.2264666557,-0.1167617217,0.0892518535,0.1257736236,0.0740240142,-0.1534793824,-0.1198020428,0.4294593036,-0.2137423903,0.2773050666,0.3186097741,0.2963832617,0.3739162087,0.227028951,-0.2862907052,0.0338729285,-0.2570652664,-0.0871776044,-0.0442779921,-0.2006773353,-0.3708454072,0.0992647409,-0.005736412,0.239634648,0.1527641416,0.0450480506,0.2212321013,-0.539868176,0.1125220954,-0.2573070824,0.1339917481,0.0350833498,-0.2941997945,0.1408175528,0.1156675741,-0.1236832365,-0.1660993546,0.1671806574,-0.2124153674,-0.2457581609,-0.1474966705,0.1560774744,-0.0041650292,0.4375189543,-0.0617898293,0.3228711784,0.3464185596,-0.3285045028,-0.0258732624,-0.3166899085,-0.4711762369,-0.2692376971,-0.0478229038,0.2325534225,-0.1191353053,0.0298826676,0.2000778913,0.0202775728,0.0941302255,0.1629669219,-0.1979359239,-0.0177588277,-0.1911362857,0.5243681073,0.2797337174,-0.0643637404,-0.4906041324,0.1102769822,-0.2673051357,-0.1981518269,-0.4909549952,-0.1560129821,0.4023430943,-0.2073682845,0.4260299802,-0.0168935806,-0.1412145048,0.118385382,-0.306717515,0.094998017,0.1623565853,-0.1462418735,0.0806366652,-0.166501984,0.1831175238,-0.0708336085,0.0203282014,0.004628906,-0.0426518135,0.2887397707,0.2136349082,-0.0781941861,0.0351790115,-0.3048202991,-0.0683804154,0.2268817574,0.0347599015,-0.0888614431,-0.0282383524,0.0105143087,-0.0411652811,0.3073398471,0.1391114891,-0.1058630794,0.0216380805,-0.1246979013,0.0030890247,-0.0661532804,-0.0376062728,-0.0197845213,-0.1885098517,0.0246681049,-0.0632356256,-0.1272604764,-0.2010300606,-0.1646624357,-0.3887704015,0.1416743398,0.4239653945,-0.0064662574,-0.2309417278,0.5506316423,0.3362869918,-0.2484891564,-0.6358670592,-0.0847648457,0.3994875848,-0.1538935304,0.1634524167,0.0869446471,-0.0126662534,-0.1323807985,0.1911210567,0.2216383666,0.2824165821,0.1572643816,-0.0959971696,0.0425383486,0.1659151316,0.1229683161,0.0325946435,-0.199791953,-0.0025342163,0.1199669093,0.0561050922,-0.047854811,-0.0479735769,0.2361027151,0.2793250382,-0.0761121586,0.1238081977,0.0580437221,-0.1685209721,-0.0567469634,0.2826824784,-0.0136068668,0.6230338812,0.0613569953,-0.0287729017,0.0040426482,-0.3801935911,-0.0336282663,0.1335722506,0.0782550871,-0.1228369474,0.0503613725,0.2556922138,0.159722358,-0.1924543381,-0.3795136213,0.0199823845,0.3146329224,-0.4612223506,0.2040783465,-0.1188221872,0.2136378586,-0.0778881684,-0.1706882864,-0.1753237247,-0.3915620148,-0.0386055,0.1213032827,0.0050780186,0.1582536548,0.0887682661,-0.1417624056,0.04338919,0.3126336932,0.2620718777,-0.0042142225,-0.4652619362,-0.1256633401,0.3646590412,-0.1426314861,0.1807572693,0.1596735418,-0.4231134653,-0.1377694905,-0.554107368,0.0362168029,0.0458122902,0.2179297954,0.0650527328,0.0002729307,-0.1417898685,-0.0640055388,0.1627510339,-0.0710154772,-0.0511523709,0.1167618632,0.1763634235,-0.3484143615,-0.4004407525,0.0810818374,-0.2404452562,-0.2278887779,0.0398980491,0.0562176853,0.2899165154,0.033601027,0.3173662722,0.2378733456,0.0640233308,-0.2699205875,-0.2033462822,0.2120094001,-0.0509562194,-0.1475808471,-0.2852531075,-0.0752577856,0.1772031188,-0.022968391,0.2650331259,-0.2499954402,-0.0535250045,-0.0659198835,0.3023351431,-0.1044890657,-0.2844492197,0.3918755949,0.1078824848,0.0479716472,-0.0043216893,-0.3812083304,-0.0652099252,-0.2201593518,0.1879525185,-0.0040869415,0.5207585096,0.1579419822,0.5066372156,0.3318142593,-0.5369669795,0.1717587262,-0.2706901133,0.1213353053,-0.1566914618,-0.1280220896,0.3013108969,0.0409448743,-0.0624374636,0.136187464,-0.0629640371,-0.4967781901,0.0518860146,0.0803500637,-0.0168010574,-0.2850890756,0.1508056819,-0.0602970831,0.2637024224,-0.1313890517,0.3329266608,-0.2499023676,-0.0794402286,-0.1635815203,-0.0973349959,-0.1765634567,-0.0466502458,-0.534514606,-0.1086725891,-0.6624104381,0.3221202195,0.1939269751,0.7835565805,0.0214179475,-0.0824445039,-0.2468029261,-0.1681991518,0.6353745461,-0.300203532,0.0723909363,-0.0045123673,0.098818697,-0.6122601032,-0.1256724149,-0.0143848732,0.6673847437,0.2643587291,0.9284402132,0.0463600494,0.045960132,0.0434808359,0.0178261716,0.1433940679,-0.0652095973,-0.0105038164,-0.0599536635,-0.1763828248,0.1956971437,0.2371133864,0.207475841,-0.0157400426,-0.4159704745,0.0754318386,-0.3766656518,0.2011009455,0.1319937855,0.208466664,-0.2982677221,0.1140266657,-0.0784145966,-0.1573798358,0.2803728878,-0.1881632805,0.0264791045,-0.3439128697,0.2833440304,0.1530024409,0.1995321661,0.426977247,-0.0416696221,0.1061635017,-0.3421507776,0.3123049438,-0.2047104388,0.2691136599,0.2535805106,0.2167814076,-0.5693432689,0.0799898058,0.1540472656,0.0905904621,-0.0760773271,0.3940853775,-0.368170321,-0.4619640112,0.3434622586,0.1300132424,0.848302424,0.0364213064,-0.081603311,-0.19214347,0.1661395729,0.1686700284,-0.1858166009,-0.0507305525,-0.3904622197,-0.1557223946,-0.0677676797,-0.1420495659,0.0983293951,0.0727663487,-0.3925172985,0.1129100174,0.1099763438,0.4424754381,-0.0006535093,0.094351247,-0.1428740174,-0.1273342818,-0.0686014518,0.1570535004,0.1238408089,-0.2321622223,-0.14562352,0.3469045758,0.1431711614,-0.4613877833,-0.62011379,-0.0647350997,-0.5778717399,0.1768615544,0.1812040657,0.2300380021,0.1644968092,0.301273495,0.0709967613,0.506118238,-0.2116412073,-0.0245072097,0.0392306633,0.0512521602,0.1469227374,-0.1345072687,-0.0176817924,0.0910632312,-0.0327797458,0.0640930757,0.001057488,0.1164323166,-0.501894176,0.265417248,0.4029581249,-0.196479097,-0.2089856416,-0.2772096992,-0.1896944642,-0.2230817527,0.0718460232,0.0842439681,-0.1488311887,0.1394552439,-0.1531652212,-0.2392156124,0.0301355962,0.4575230777,-0.2256522924,0.0981041044,-0.0154942367,0.0785034224,-0.0677308589,-0.1901301593,0.2326962501,-0.1237555444,-0.0325517394,-0.2231186181,-0.2007786334,-0.0984243006,0.1080042869,0.0606934577,0.1129072234,0.0262527335,-0.016009355,-0.5448374152,-0.4919656813,-0.0073139649,-0.0454394147,0.3145165741,-0.0165020507,0.0936865062,-0.2128308862,0.1364151239,-0.2179399729,0.1666069627,-0.0632879138,0.2086213976,-0.0276215877,0.0983446389,-0.1632549614,-0.1979133934,0.0605035052,0.7531200051,-0.2359837741,-0.0942862928,-0.0525906458,0.1379344016,0.2398593724,-0.1094310135,-0.1033161655,-0.0272614323,0.1039245501,-0.2421072721,0.1028843969,0.2268958539,-0.2034132034,-0.131596297,0.6177958846,0.3815800548,0.1931364983,0.357540071,0.0191768613,0.2638332546,0.1465423852,0.3621217012,0.1001960859,-0.0010375903,-0.0883050188,0.2816980779,0.0660891756,0.1587927341,0.2086260915,-0.0950449184,0.1765199453,0.0912955999,-0.0393662527,0.3638717234,-0.4002837837,-0.0639449432,0.191659078,0.1843820661,0.0324680954,-0.2394108027,0.3049276471,0.1877780557,0.0064044776,0.2658642828,0.0362890139,-0.3785454035,0.079316467,0.2469057888,0.3529776633,-0.0388915651,-0.0370597355,0.2587002814,-0.2121544927,0.1043223813,0.2033436149,-0.1946918219,0.3131908774,0.6266698241,-0.0740214512,0.4274179339,0.1257343292,0.0630890504,-0.0054944996,-0.2483475804,0.0808186531,0.4405095875,-0.1837271601,0.1500857174,0.0620873384,0.257960856,-0.3250874877,-0.1628504544,0.0823815241,0.0784967542,-0.3235331476,-0.0777549371,-0.0641655922,-0.1524268389,-0.1851560324,0.0298068896,-0.1491112709,-0.134274587,0.3058646917,0.105733268,-0.3869328499,-0.2465470433,-0.0842965916,0.0462458469,-0.0254337527,-0.2964484394,0.5547583699,-0.165271312,0.2051642239,0.1515772492,-0.1810009629,0.5931897759,0.6459059119,-0.2287395895,0.1645237505,0.1871111542,0.0369282775,-0.0236070659,0.1389949471,0.0064930003,0.2090339661,0.2872698605,0.0319978818,-0.1442494094,-0.1472728252,0.3358566463,0.2155430019,-0.0512180105,0.1997416914,-0.2628280818,-0.1870484501,0.0019922892,-0.0307528339,-0.1691651791,-0.0553508326,0.3815113008,-0.0056802407,0.4061158597,-0.1410945505,0.0556682125,0.1124868616,0.165840745,0.2160716504,-0.1539394557,-0.4025766253,-0.3473495245,-0.3931940496,0.3725878298,-0.2325892448,0.1685742289,0.0681374967,-0.2374645174,-0.0817601308,0.003409601,0.1784852594,0.2687878013,-0.0693608075,0.5016827583,-0.3206634223,-0.0434216745,-0.0729406029,-0.1339516491,0.2401047647,-0.4375864863,0.4043382704,-0.3041529357,-0.0143265808,-0.1915714145,0.0053744041,-0.227931276,-0.0372267179,0.6843011975,0.2150361538,0.4069308043,-0.0329385772,0.1105626151,-0.2053319067,0.1099141538,-0.2095181644,0.2924145162,0.0320701674,0.1326195002,-0.1456784308,-0.1236873269,-0.072567299,-0.1102667898,0.2603496611,-0.0558728129,-0.2727081776,0.1199415252,-0.2002013028,-0.0307991542,-0.1286313683,0.3280451596,0.1533139497,0.2899936736,0.0323429331,-0.2277695686,0.3998816311,-0.3358049691,-0.3362597823,-0.1846985966,0.1810190976,-0.2116550505,-0.3600912392,-0.5163121819,-0.3579574823,0.3601654172,0.0267328806,-0.5006997585,-0.0882551745,-0.2791074812,-0.0996797755,-0.2518528104,0.4753126502,0.1529987752,-0.1394924521,0.6449896693,-0.3258785605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2470","title":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.","comments":"Could you trying reinstalling pyarrow with pip ?\r\nI'm not sure why it would check in your multicurtural-sc directory for source files.","body":"## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n","comment_length":22,"text":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.\n## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n\nCould you trying reinstalling pyarrow with pip ?\r\nI'm not sure why it would check in your multicurtural-sc directory for source files.","embeddings":[-0.4569717646,-0.2173907459,0.0228779204,0.4219461381,0.3113077581,-0.2264823467,0.2361344099,0.1525053084,0.0431262851,0.3643293679,0.5303300619,0.2979793251,-0.3346508443,-0.0465700552,0.047729928,-0.0265675541,0.3659133613,-0.2182355374,0.2511509061,0.3584368825,-0.2708789408,0.1627700925,-0.4191713631,0.2937121987,-0.2652947009,0.082243599,-0.1050226912,0.2865630984,-0.0974157006,-0.5258919597,-0.0305760354,-0.295525372,0.1912617087,0.4583674669,-0.0001182336,-0.0913553163,0.2139718682,0.0120125022,0.0930700898,-0.4430084825,-0.2259867936,-0.2528820932,-0.1637293845,-0.0408718698,0.3133925498,0.0046791183,-0.0663942918,-0.150278613,0.0759297609,0.7205310464,0.1591092348,0.214832738,0.2288711667,-0.1120013967,0.016339317,0.1341472566,-0.0305977836,0.061437469,0.4003654718,-0.2003950328,0.0249149706,-0.1673947722,0.092808485,-0.2669810057,0.305939585,-0.1227102131,0.1998042613,-0.6616144776,0.2974792719,-0.1724279821,-0.1635968983,-0.2878350317,-0.0509196632,-0.1632319391,-0.1430846751,-0.2230149657,0.177413553,-0.1671895236,-0.1571241021,0.0832860544,-0.2861917019,0.1699507236,0.0335625187,0.3086440563,-0.3192946315,0.4002259374,-0.0029774297,0.389618665,0.1262956411,-0.2813790739,-0.3420287967,0.1340495944,-0.0415284634,0.2674230039,-0.2257158607,-0.1663678735,0.1507905126,-0.1738581359,0.2904581726,-0.2406045794,-0.1656422466,0.2874025404,0.048612088,0.2133101672,0.2980216742,-0.0209573247,0.3021420836,0.3840674162,0.0499104597,-0.056064833,-0.102803044,-0.0400411636,0.2264666557,-0.1167617217,0.0892518535,0.1257736236,0.0740240142,-0.1534793824,-0.1198020428,0.4294593036,-0.2137423903,0.2773050666,0.3186097741,0.2963832617,0.3739162087,0.227028951,-0.2862907052,0.0338729285,-0.2570652664,-0.0871776044,-0.0442779921,-0.2006773353,-0.3708454072,0.0992647409,-0.005736412,0.239634648,0.1527641416,0.0450480506,0.2212321013,-0.539868176,0.1125220954,-0.2573070824,0.1339917481,0.0350833498,-0.2941997945,0.1408175528,0.1156675741,-0.1236832365,-0.1660993546,0.1671806574,-0.2124153674,-0.2457581609,-0.1474966705,0.1560774744,-0.0041650292,0.4375189543,-0.0617898293,0.3228711784,0.3464185596,-0.3285045028,-0.0258732624,-0.3166899085,-0.4711762369,-0.2692376971,-0.0478229038,0.2325534225,-0.1191353053,0.0298826676,0.2000778913,0.0202775728,0.0941302255,0.1629669219,-0.1979359239,-0.0177588277,-0.1911362857,0.5243681073,0.2797337174,-0.0643637404,-0.4906041324,0.1102769822,-0.2673051357,-0.1981518269,-0.4909549952,-0.1560129821,0.4023430943,-0.2073682845,0.4260299802,-0.0168935806,-0.1412145048,0.118385382,-0.306717515,0.094998017,0.1623565853,-0.1462418735,0.0806366652,-0.166501984,0.1831175238,-0.0708336085,0.0203282014,0.004628906,-0.0426518135,0.2887397707,0.2136349082,-0.0781941861,0.0351790115,-0.3048202991,-0.0683804154,0.2268817574,0.0347599015,-0.0888614431,-0.0282383524,0.0105143087,-0.0411652811,0.3073398471,0.1391114891,-0.1058630794,0.0216380805,-0.1246979013,0.0030890247,-0.0661532804,-0.0376062728,-0.0197845213,-0.1885098517,0.0246681049,-0.0632356256,-0.1272604764,-0.2010300606,-0.1646624357,-0.3887704015,0.1416743398,0.4239653945,-0.0064662574,-0.2309417278,0.5506316423,0.3362869918,-0.2484891564,-0.6358670592,-0.0847648457,0.3994875848,-0.1538935304,0.1634524167,0.0869446471,-0.0126662534,-0.1323807985,0.1911210567,0.2216383666,0.2824165821,0.1572643816,-0.0959971696,0.0425383486,0.1659151316,0.1229683161,0.0325946435,-0.199791953,-0.0025342163,0.1199669093,0.0561050922,-0.047854811,-0.0479735769,0.2361027151,0.2793250382,-0.0761121586,0.1238081977,0.0580437221,-0.1685209721,-0.0567469634,0.2826824784,-0.0136068668,0.6230338812,0.0613569953,-0.0287729017,0.0040426482,-0.3801935911,-0.0336282663,0.1335722506,0.0782550871,-0.1228369474,0.0503613725,0.2556922138,0.159722358,-0.1924543381,-0.3795136213,0.0199823845,0.3146329224,-0.4612223506,0.2040783465,-0.1188221872,0.2136378586,-0.0778881684,-0.1706882864,-0.1753237247,-0.3915620148,-0.0386055,0.1213032827,0.0050780186,0.1582536548,0.0887682661,-0.1417624056,0.04338919,0.3126336932,0.2620718777,-0.0042142225,-0.4652619362,-0.1256633401,0.3646590412,-0.1426314861,0.1807572693,0.1596735418,-0.4231134653,-0.1377694905,-0.554107368,0.0362168029,0.0458122902,0.2179297954,0.0650527328,0.0002729307,-0.1417898685,-0.0640055388,0.1627510339,-0.0710154772,-0.0511523709,0.1167618632,0.1763634235,-0.3484143615,-0.4004407525,0.0810818374,-0.2404452562,-0.2278887779,0.0398980491,0.0562176853,0.2899165154,0.033601027,0.3173662722,0.2378733456,0.0640233308,-0.2699205875,-0.2033462822,0.2120094001,-0.0509562194,-0.1475808471,-0.2852531075,-0.0752577856,0.1772031188,-0.022968391,0.2650331259,-0.2499954402,-0.0535250045,-0.0659198835,0.3023351431,-0.1044890657,-0.2844492197,0.3918755949,0.1078824848,0.0479716472,-0.0043216893,-0.3812083304,-0.0652099252,-0.2201593518,0.1879525185,-0.0040869415,0.5207585096,0.1579419822,0.5066372156,0.3318142593,-0.5369669795,0.1717587262,-0.2706901133,0.1213353053,-0.1566914618,-0.1280220896,0.3013108969,0.0409448743,-0.0624374636,0.136187464,-0.0629640371,-0.4967781901,0.0518860146,0.0803500637,-0.0168010574,-0.2850890756,0.1508056819,-0.0602970831,0.2637024224,-0.1313890517,0.3329266608,-0.2499023676,-0.0794402286,-0.1635815203,-0.0973349959,-0.1765634567,-0.0466502458,-0.534514606,-0.1086725891,-0.6624104381,0.3221202195,0.1939269751,0.7835565805,0.0214179475,-0.0824445039,-0.2468029261,-0.1681991518,0.6353745461,-0.300203532,0.0723909363,-0.0045123673,0.098818697,-0.6122601032,-0.1256724149,-0.0143848732,0.6673847437,0.2643587291,0.9284402132,0.0463600494,0.045960132,0.0434808359,0.0178261716,0.1433940679,-0.0652095973,-0.0105038164,-0.0599536635,-0.1763828248,0.1956971437,0.2371133864,0.207475841,-0.0157400426,-0.4159704745,0.0754318386,-0.3766656518,0.2011009455,0.1319937855,0.208466664,-0.2982677221,0.1140266657,-0.0784145966,-0.1573798358,0.2803728878,-0.1881632805,0.0264791045,-0.3439128697,0.2833440304,0.1530024409,0.1995321661,0.426977247,-0.0416696221,0.1061635017,-0.3421507776,0.3123049438,-0.2047104388,0.2691136599,0.2535805106,0.2167814076,-0.5693432689,0.0799898058,0.1540472656,0.0905904621,-0.0760773271,0.3940853775,-0.368170321,-0.4619640112,0.3434622586,0.1300132424,0.848302424,0.0364213064,-0.081603311,-0.19214347,0.1661395729,0.1686700284,-0.1858166009,-0.0507305525,-0.3904622197,-0.1557223946,-0.0677676797,-0.1420495659,0.0983293951,0.0727663487,-0.3925172985,0.1129100174,0.1099763438,0.4424754381,-0.0006535093,0.094351247,-0.1428740174,-0.1273342818,-0.0686014518,0.1570535004,0.1238408089,-0.2321622223,-0.14562352,0.3469045758,0.1431711614,-0.4613877833,-0.62011379,-0.0647350997,-0.5778717399,0.1768615544,0.1812040657,0.2300380021,0.1644968092,0.301273495,0.0709967613,0.506118238,-0.2116412073,-0.0245072097,0.0392306633,0.0512521602,0.1469227374,-0.1345072687,-0.0176817924,0.0910632312,-0.0327797458,0.0640930757,0.001057488,0.1164323166,-0.501894176,0.265417248,0.4029581249,-0.196479097,-0.2089856416,-0.2772096992,-0.1896944642,-0.2230817527,0.0718460232,0.0842439681,-0.1488311887,0.1394552439,-0.1531652212,-0.2392156124,0.0301355962,0.4575230777,-0.2256522924,0.0981041044,-0.0154942367,0.0785034224,-0.0677308589,-0.1901301593,0.2326962501,-0.1237555444,-0.0325517394,-0.2231186181,-0.2007786334,-0.0984243006,0.1080042869,0.0606934577,0.1129072234,0.0262527335,-0.016009355,-0.5448374152,-0.4919656813,-0.0073139649,-0.0454394147,0.3145165741,-0.0165020507,0.0936865062,-0.2128308862,0.1364151239,-0.2179399729,0.1666069627,-0.0632879138,0.2086213976,-0.0276215877,0.0983446389,-0.1632549614,-0.1979133934,0.0605035052,0.7531200051,-0.2359837741,-0.0942862928,-0.0525906458,0.1379344016,0.2398593724,-0.1094310135,-0.1033161655,-0.0272614323,0.1039245501,-0.2421072721,0.1028843969,0.2268958539,-0.2034132034,-0.131596297,0.6177958846,0.3815800548,0.1931364983,0.357540071,0.0191768613,0.2638332546,0.1465423852,0.3621217012,0.1001960859,-0.0010375903,-0.0883050188,0.2816980779,0.0660891756,0.1587927341,0.2086260915,-0.0950449184,0.1765199453,0.0912955999,-0.0393662527,0.3638717234,-0.4002837837,-0.0639449432,0.191659078,0.1843820661,0.0324680954,-0.2394108027,0.3049276471,0.1877780557,0.0064044776,0.2658642828,0.0362890139,-0.3785454035,0.079316467,0.2469057888,0.3529776633,-0.0388915651,-0.0370597355,0.2587002814,-0.2121544927,0.1043223813,0.2033436149,-0.1946918219,0.3131908774,0.6266698241,-0.0740214512,0.4274179339,0.1257343292,0.0630890504,-0.0054944996,-0.2483475804,0.0808186531,0.4405095875,-0.1837271601,0.1500857174,0.0620873384,0.257960856,-0.3250874877,-0.1628504544,0.0823815241,0.0784967542,-0.3235331476,-0.0777549371,-0.0641655922,-0.1524268389,-0.1851560324,0.0298068896,-0.1491112709,-0.134274587,0.3058646917,0.105733268,-0.3869328499,-0.2465470433,-0.0842965916,0.0462458469,-0.0254337527,-0.2964484394,0.5547583699,-0.165271312,0.2051642239,0.1515772492,-0.1810009629,0.5931897759,0.6459059119,-0.2287395895,0.1645237505,0.1871111542,0.0369282775,-0.0236070659,0.1389949471,0.0064930003,0.2090339661,0.2872698605,0.0319978818,-0.1442494094,-0.1472728252,0.3358566463,0.2155430019,-0.0512180105,0.1997416914,-0.2628280818,-0.1870484501,0.0019922892,-0.0307528339,-0.1691651791,-0.0553508326,0.3815113008,-0.0056802407,0.4061158597,-0.1410945505,0.0556682125,0.1124868616,0.165840745,0.2160716504,-0.1539394557,-0.4025766253,-0.3473495245,-0.3931940496,0.3725878298,-0.2325892448,0.1685742289,0.0681374967,-0.2374645174,-0.0817601308,0.003409601,0.1784852594,0.2687878013,-0.0693608075,0.5016827583,-0.3206634223,-0.0434216745,-0.0729406029,-0.1339516491,0.2401047647,-0.4375864863,0.4043382704,-0.3041529357,-0.0143265808,-0.1915714145,0.0053744041,-0.227931276,-0.0372267179,0.6843011975,0.2150361538,0.4069308043,-0.0329385772,0.1105626151,-0.2053319067,0.1099141538,-0.2095181644,0.2924145162,0.0320701674,0.1326195002,-0.1456784308,-0.1236873269,-0.072567299,-0.1102667898,0.2603496611,-0.0558728129,-0.2727081776,0.1199415252,-0.2002013028,-0.0307991542,-0.1286313683,0.3280451596,0.1533139497,0.2899936736,0.0323429331,-0.2277695686,0.3998816311,-0.3358049691,-0.3362597823,-0.1846985966,0.1810190976,-0.2116550505,-0.3600912392,-0.5163121819,-0.3579574823,0.3601654172,0.0267328806,-0.5006997585,-0.0882551745,-0.2791074812,-0.0996797755,-0.2518528104,0.4753126502,0.1529987752,-0.1394924521,0.6449896693,-0.3258785605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2470","title":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.","comments":"Sure! I tried reinstalling to get latest. pip was mad because it looks like Datasets currently wants <4.0.0 (which is interesting, because apparently I ended up with 4.0.0 already?), but I gave it a shot anyway:\r\n\r\n```bash\r\n$ pip install --upgrade --force-reinstall pyarrow\r\nCollecting pyarrow\r\n Downloading pyarrow-4.0.1-cp39-cp39-manylinux2014_x86_64.whl (21.9 MB)\r\n |\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 21.9 MB 23.8 MB\/s\r\nCollecting numpy>=1.16.6\r\n Using cached numpy-1.20.3-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (15.4 MB)\r\nInstalling collected packages: numpy, pyarrow\r\n Attempting uninstall: numpy\r\n Found existing installation: numpy 1.20.3\r\n Uninstalling numpy-1.20.3:\r\n Successfully uninstalled numpy-1.20.3\r\n Attempting uninstall: pyarrow\r\n Found existing installation: pyarrow 3.0.0\r\n Uninstalling pyarrow-3.0.0:\r\n Successfully uninstalled pyarrow-3.0.0\r\nERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.\r\ndatasets 1.8.0 requires pyarrow<4.0.0,>=1.0.0, but you have pyarrow 4.0.1 which is incompatible.\r\nSuccessfully installed numpy-1.20.3 pyarrow-4.0.1\r\n```\r\n\r\nTrying it, the same issue:\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121730226-3f470b80-caa4-11eb-85a5-684c44c816da.png)\r\n\r\nI tried installing `\"pyarrow<4.0.0\"`, which gave me 3.0.0. Running, still, same issue.\r\n\r\nI agree it's weird that pyarrow is checking the source code directory for its files. (There is no `pyarrow\/` directory there.) To me, that makes it seem like an issue with how pyarrow is called.\r\n\r\nOut of curiosity, I tried running this with fewer workers to see when the error arises:\r\n\r\n- 1: \u2705\r\n- 2: \u2705\r\n- 4: \u2705\r\n- 8: \u2705\r\n- 10: \u2705\r\n- 11: \u274c \ud83e\udd14\r\n- 12: \u274c\r\n- 16: \u274c\r\n- 32: \u274c\r\n\r\nchecking my datasets:\r\n\r\n```python\r\n>>> datasets\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text'],\r\n num_rows: 389290\r\n })\r\n validation.sc: Dataset({\r\n features: ['text'],\r\n num_rows: 10 # \ud83e\udd14\r\n })\r\n validation.wvs: Dataset({\r\n features: ['text'],\r\n num_rows: 93928\r\n })\r\n})\r\n```\r\n\r\nNew hypothesis: crash if `num_proc` > length of a dataset? \ud83d\ude05\r\n\r\nIf so, this might be totally my fault, as the caller. Could be a docs fix, or maybe this library could do a check to limit `num_proc` for this case?","body":"## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n","comment_length":305,"text":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.\n## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n\nSure! I tried reinstalling to get latest. pip was mad because it looks like Datasets currently wants <4.0.0 (which is interesting, because apparently I ended up with 4.0.0 already?), but I gave it a shot anyway:\r\n\r\n```bash\r\n$ pip install --upgrade --force-reinstall pyarrow\r\nCollecting pyarrow\r\n Downloading pyarrow-4.0.1-cp39-cp39-manylinux2014_x86_64.whl (21.9 MB)\r\n |\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 21.9 MB 23.8 MB\/s\r\nCollecting numpy>=1.16.6\r\n Using cached numpy-1.20.3-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (15.4 MB)\r\nInstalling collected packages: numpy, pyarrow\r\n Attempting uninstall: numpy\r\n Found existing installation: numpy 1.20.3\r\n Uninstalling numpy-1.20.3:\r\n Successfully uninstalled numpy-1.20.3\r\n Attempting uninstall: pyarrow\r\n Found existing installation: pyarrow 3.0.0\r\n Uninstalling pyarrow-3.0.0:\r\n Successfully uninstalled pyarrow-3.0.0\r\nERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.\r\ndatasets 1.8.0 requires pyarrow<4.0.0,>=1.0.0, but you have pyarrow 4.0.1 which is incompatible.\r\nSuccessfully installed numpy-1.20.3 pyarrow-4.0.1\r\n```\r\n\r\nTrying it, the same issue:\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121730226-3f470b80-caa4-11eb-85a5-684c44c816da.png)\r\n\r\nI tried installing `\"pyarrow<4.0.0\"`, which gave me 3.0.0. Running, still, same issue.\r\n\r\nI agree it's weird that pyarrow is checking the source code directory for its files. (There is no `pyarrow\/` directory there.) To me, that makes it seem like an issue with how pyarrow is called.\r\n\r\nOut of curiosity, I tried running this with fewer workers to see when the error arises:\r\n\r\n- 1: \u2705\r\n- 2: \u2705\r\n- 4: \u2705\r\n- 8: \u2705\r\n- 10: \u2705\r\n- 11: \u274c \ud83e\udd14\r\n- 12: \u274c\r\n- 16: \u274c\r\n- 32: \u274c\r\n\r\nchecking my datasets:\r\n\r\n```python\r\n>>> datasets\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text'],\r\n num_rows: 389290\r\n })\r\n validation.sc: Dataset({\r\n features: ['text'],\r\n num_rows: 10 # \ud83e\udd14\r\n })\r\n validation.wvs: Dataset({\r\n features: ['text'],\r\n num_rows: 93928\r\n })\r\n})\r\n```\r\n\r\nNew hypothesis: crash if `num_proc` > length of a dataset? \ud83d\ude05\r\n\r\nIf so, this might be totally my fault, as the caller. Could be a docs fix, or maybe this library could do a check to limit `num_proc` for this case?","embeddings":[-0.4569717646,-0.2173907459,0.0228779204,0.4219461381,0.3113077581,-0.2264823467,0.2361344099,0.1525053084,0.0431262851,0.3643293679,0.5303300619,0.2979793251,-0.3346508443,-0.0465700552,0.047729928,-0.0265675541,0.3659133613,-0.2182355374,0.2511509061,0.3584368825,-0.2708789408,0.1627700925,-0.4191713631,0.2937121987,-0.2652947009,0.082243599,-0.1050226912,0.2865630984,-0.0974157006,-0.5258919597,-0.0305760354,-0.295525372,0.1912617087,0.4583674669,-0.0001182336,-0.0913553163,0.2139718682,0.0120125022,0.0930700898,-0.4430084825,-0.2259867936,-0.2528820932,-0.1637293845,-0.0408718698,0.3133925498,0.0046791183,-0.0663942918,-0.150278613,0.0759297609,0.7205310464,0.1591092348,0.214832738,0.2288711667,-0.1120013967,0.016339317,0.1341472566,-0.0305977836,0.061437469,0.4003654718,-0.2003950328,0.0249149706,-0.1673947722,0.092808485,-0.2669810057,0.305939585,-0.1227102131,0.1998042613,-0.6616144776,0.2974792719,-0.1724279821,-0.1635968983,-0.2878350317,-0.0509196632,-0.1632319391,-0.1430846751,-0.2230149657,0.177413553,-0.1671895236,-0.1571241021,0.0832860544,-0.2861917019,0.1699507236,0.0335625187,0.3086440563,-0.3192946315,0.4002259374,-0.0029774297,0.389618665,0.1262956411,-0.2813790739,-0.3420287967,0.1340495944,-0.0415284634,0.2674230039,-0.2257158607,-0.1663678735,0.1507905126,-0.1738581359,0.2904581726,-0.2406045794,-0.1656422466,0.2874025404,0.048612088,0.2133101672,0.2980216742,-0.0209573247,0.3021420836,0.3840674162,0.0499104597,-0.056064833,-0.102803044,-0.0400411636,0.2264666557,-0.1167617217,0.0892518535,0.1257736236,0.0740240142,-0.1534793824,-0.1198020428,0.4294593036,-0.2137423903,0.2773050666,0.3186097741,0.2963832617,0.3739162087,0.227028951,-0.2862907052,0.0338729285,-0.2570652664,-0.0871776044,-0.0442779921,-0.2006773353,-0.3708454072,0.0992647409,-0.005736412,0.239634648,0.1527641416,0.0450480506,0.2212321013,-0.539868176,0.1125220954,-0.2573070824,0.1339917481,0.0350833498,-0.2941997945,0.1408175528,0.1156675741,-0.1236832365,-0.1660993546,0.1671806574,-0.2124153674,-0.2457581609,-0.1474966705,0.1560774744,-0.0041650292,0.4375189543,-0.0617898293,0.3228711784,0.3464185596,-0.3285045028,-0.0258732624,-0.3166899085,-0.4711762369,-0.2692376971,-0.0478229038,0.2325534225,-0.1191353053,0.0298826676,0.2000778913,0.0202775728,0.0941302255,0.1629669219,-0.1979359239,-0.0177588277,-0.1911362857,0.5243681073,0.2797337174,-0.0643637404,-0.4906041324,0.1102769822,-0.2673051357,-0.1981518269,-0.4909549952,-0.1560129821,0.4023430943,-0.2073682845,0.4260299802,-0.0168935806,-0.1412145048,0.118385382,-0.306717515,0.094998017,0.1623565853,-0.1462418735,0.0806366652,-0.166501984,0.1831175238,-0.0708336085,0.0203282014,0.004628906,-0.0426518135,0.2887397707,0.2136349082,-0.0781941861,0.0351790115,-0.3048202991,-0.0683804154,0.2268817574,0.0347599015,-0.0888614431,-0.0282383524,0.0105143087,-0.0411652811,0.3073398471,0.1391114891,-0.1058630794,0.0216380805,-0.1246979013,0.0030890247,-0.0661532804,-0.0376062728,-0.0197845213,-0.1885098517,0.0246681049,-0.0632356256,-0.1272604764,-0.2010300606,-0.1646624357,-0.3887704015,0.1416743398,0.4239653945,-0.0064662574,-0.2309417278,0.5506316423,0.3362869918,-0.2484891564,-0.6358670592,-0.0847648457,0.3994875848,-0.1538935304,0.1634524167,0.0869446471,-0.0126662534,-0.1323807985,0.1911210567,0.2216383666,0.2824165821,0.1572643816,-0.0959971696,0.0425383486,0.1659151316,0.1229683161,0.0325946435,-0.199791953,-0.0025342163,0.1199669093,0.0561050922,-0.047854811,-0.0479735769,0.2361027151,0.2793250382,-0.0761121586,0.1238081977,0.0580437221,-0.1685209721,-0.0567469634,0.2826824784,-0.0136068668,0.6230338812,0.0613569953,-0.0287729017,0.0040426482,-0.3801935911,-0.0336282663,0.1335722506,0.0782550871,-0.1228369474,0.0503613725,0.2556922138,0.159722358,-0.1924543381,-0.3795136213,0.0199823845,0.3146329224,-0.4612223506,0.2040783465,-0.1188221872,0.2136378586,-0.0778881684,-0.1706882864,-0.1753237247,-0.3915620148,-0.0386055,0.1213032827,0.0050780186,0.1582536548,0.0887682661,-0.1417624056,0.04338919,0.3126336932,0.2620718777,-0.0042142225,-0.4652619362,-0.1256633401,0.3646590412,-0.1426314861,0.1807572693,0.1596735418,-0.4231134653,-0.1377694905,-0.554107368,0.0362168029,0.0458122902,0.2179297954,0.0650527328,0.0002729307,-0.1417898685,-0.0640055388,0.1627510339,-0.0710154772,-0.0511523709,0.1167618632,0.1763634235,-0.3484143615,-0.4004407525,0.0810818374,-0.2404452562,-0.2278887779,0.0398980491,0.0562176853,0.2899165154,0.033601027,0.3173662722,0.2378733456,0.0640233308,-0.2699205875,-0.2033462822,0.2120094001,-0.0509562194,-0.1475808471,-0.2852531075,-0.0752577856,0.1772031188,-0.022968391,0.2650331259,-0.2499954402,-0.0535250045,-0.0659198835,0.3023351431,-0.1044890657,-0.2844492197,0.3918755949,0.1078824848,0.0479716472,-0.0043216893,-0.3812083304,-0.0652099252,-0.2201593518,0.1879525185,-0.0040869415,0.5207585096,0.1579419822,0.5066372156,0.3318142593,-0.5369669795,0.1717587262,-0.2706901133,0.1213353053,-0.1566914618,-0.1280220896,0.3013108969,0.0409448743,-0.0624374636,0.136187464,-0.0629640371,-0.4967781901,0.0518860146,0.0803500637,-0.0168010574,-0.2850890756,0.1508056819,-0.0602970831,0.2637024224,-0.1313890517,0.3329266608,-0.2499023676,-0.0794402286,-0.1635815203,-0.0973349959,-0.1765634567,-0.0466502458,-0.534514606,-0.1086725891,-0.6624104381,0.3221202195,0.1939269751,0.7835565805,0.0214179475,-0.0824445039,-0.2468029261,-0.1681991518,0.6353745461,-0.300203532,0.0723909363,-0.0045123673,0.098818697,-0.6122601032,-0.1256724149,-0.0143848732,0.6673847437,0.2643587291,0.9284402132,0.0463600494,0.045960132,0.0434808359,0.0178261716,0.1433940679,-0.0652095973,-0.0105038164,-0.0599536635,-0.1763828248,0.1956971437,0.2371133864,0.207475841,-0.0157400426,-0.4159704745,0.0754318386,-0.3766656518,0.2011009455,0.1319937855,0.208466664,-0.2982677221,0.1140266657,-0.0784145966,-0.1573798358,0.2803728878,-0.1881632805,0.0264791045,-0.3439128697,0.2833440304,0.1530024409,0.1995321661,0.426977247,-0.0416696221,0.1061635017,-0.3421507776,0.3123049438,-0.2047104388,0.2691136599,0.2535805106,0.2167814076,-0.5693432689,0.0799898058,0.1540472656,0.0905904621,-0.0760773271,0.3940853775,-0.368170321,-0.4619640112,0.3434622586,0.1300132424,0.848302424,0.0364213064,-0.081603311,-0.19214347,0.1661395729,0.1686700284,-0.1858166009,-0.0507305525,-0.3904622197,-0.1557223946,-0.0677676797,-0.1420495659,0.0983293951,0.0727663487,-0.3925172985,0.1129100174,0.1099763438,0.4424754381,-0.0006535093,0.094351247,-0.1428740174,-0.1273342818,-0.0686014518,0.1570535004,0.1238408089,-0.2321622223,-0.14562352,0.3469045758,0.1431711614,-0.4613877833,-0.62011379,-0.0647350997,-0.5778717399,0.1768615544,0.1812040657,0.2300380021,0.1644968092,0.301273495,0.0709967613,0.506118238,-0.2116412073,-0.0245072097,0.0392306633,0.0512521602,0.1469227374,-0.1345072687,-0.0176817924,0.0910632312,-0.0327797458,0.0640930757,0.001057488,0.1164323166,-0.501894176,0.265417248,0.4029581249,-0.196479097,-0.2089856416,-0.2772096992,-0.1896944642,-0.2230817527,0.0718460232,0.0842439681,-0.1488311887,0.1394552439,-0.1531652212,-0.2392156124,0.0301355962,0.4575230777,-0.2256522924,0.0981041044,-0.0154942367,0.0785034224,-0.0677308589,-0.1901301593,0.2326962501,-0.1237555444,-0.0325517394,-0.2231186181,-0.2007786334,-0.0984243006,0.1080042869,0.0606934577,0.1129072234,0.0262527335,-0.016009355,-0.5448374152,-0.4919656813,-0.0073139649,-0.0454394147,0.3145165741,-0.0165020507,0.0936865062,-0.2128308862,0.1364151239,-0.2179399729,0.1666069627,-0.0632879138,0.2086213976,-0.0276215877,0.0983446389,-0.1632549614,-0.1979133934,0.0605035052,0.7531200051,-0.2359837741,-0.0942862928,-0.0525906458,0.1379344016,0.2398593724,-0.1094310135,-0.1033161655,-0.0272614323,0.1039245501,-0.2421072721,0.1028843969,0.2268958539,-0.2034132034,-0.131596297,0.6177958846,0.3815800548,0.1931364983,0.357540071,0.0191768613,0.2638332546,0.1465423852,0.3621217012,0.1001960859,-0.0010375903,-0.0883050188,0.2816980779,0.0660891756,0.1587927341,0.2086260915,-0.0950449184,0.1765199453,0.0912955999,-0.0393662527,0.3638717234,-0.4002837837,-0.0639449432,0.191659078,0.1843820661,0.0324680954,-0.2394108027,0.3049276471,0.1877780557,0.0064044776,0.2658642828,0.0362890139,-0.3785454035,0.079316467,0.2469057888,0.3529776633,-0.0388915651,-0.0370597355,0.2587002814,-0.2121544927,0.1043223813,0.2033436149,-0.1946918219,0.3131908774,0.6266698241,-0.0740214512,0.4274179339,0.1257343292,0.0630890504,-0.0054944996,-0.2483475804,0.0808186531,0.4405095875,-0.1837271601,0.1500857174,0.0620873384,0.257960856,-0.3250874877,-0.1628504544,0.0823815241,0.0784967542,-0.3235331476,-0.0777549371,-0.0641655922,-0.1524268389,-0.1851560324,0.0298068896,-0.1491112709,-0.134274587,0.3058646917,0.105733268,-0.3869328499,-0.2465470433,-0.0842965916,0.0462458469,-0.0254337527,-0.2964484394,0.5547583699,-0.165271312,0.2051642239,0.1515772492,-0.1810009629,0.5931897759,0.6459059119,-0.2287395895,0.1645237505,0.1871111542,0.0369282775,-0.0236070659,0.1389949471,0.0064930003,0.2090339661,0.2872698605,0.0319978818,-0.1442494094,-0.1472728252,0.3358566463,0.2155430019,-0.0512180105,0.1997416914,-0.2628280818,-0.1870484501,0.0019922892,-0.0307528339,-0.1691651791,-0.0553508326,0.3815113008,-0.0056802407,0.4061158597,-0.1410945505,0.0556682125,0.1124868616,0.165840745,0.2160716504,-0.1539394557,-0.4025766253,-0.3473495245,-0.3931940496,0.3725878298,-0.2325892448,0.1685742289,0.0681374967,-0.2374645174,-0.0817601308,0.003409601,0.1784852594,0.2687878013,-0.0693608075,0.5016827583,-0.3206634223,-0.0434216745,-0.0729406029,-0.1339516491,0.2401047647,-0.4375864863,0.4043382704,-0.3041529357,-0.0143265808,-0.1915714145,0.0053744041,-0.227931276,-0.0372267179,0.6843011975,0.2150361538,0.4069308043,-0.0329385772,0.1105626151,-0.2053319067,0.1099141538,-0.2095181644,0.2924145162,0.0320701674,0.1326195002,-0.1456784308,-0.1236873269,-0.072567299,-0.1102667898,0.2603496611,-0.0558728129,-0.2727081776,0.1199415252,-0.2002013028,-0.0307991542,-0.1286313683,0.3280451596,0.1533139497,0.2899936736,0.0323429331,-0.2277695686,0.3998816311,-0.3358049691,-0.3362597823,-0.1846985966,0.1810190976,-0.2116550505,-0.3600912392,-0.5163121819,-0.3579574823,0.3601654172,0.0267328806,-0.5006997585,-0.0882551745,-0.2791074812,-0.0996797755,-0.2518528104,0.4753126502,0.1529987752,-0.1394924521,0.6449896693,-0.3258785605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2470","title":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.","comments":"Good catch ! Not sure why it could raise such a weird issue from pyarrow though\r\nWe should definitely reduce num_proc to the length of the dataset if needed and log a warning.","body":"## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n","comment_length":33,"text":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.\n## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n\nGood catch ! Not sure why it could raise such a weird issue from pyarrow though\r\nWe should definitely reduce num_proc to the length of the dataset if needed and log a warning.","embeddings":[-0.4569717646,-0.2173907459,0.0228779204,0.4219461381,0.3113077581,-0.2264823467,0.2361344099,0.1525053084,0.0431262851,0.3643293679,0.5303300619,0.2979793251,-0.3346508443,-0.0465700552,0.047729928,-0.0265675541,0.3659133613,-0.2182355374,0.2511509061,0.3584368825,-0.2708789408,0.1627700925,-0.4191713631,0.2937121987,-0.2652947009,0.082243599,-0.1050226912,0.2865630984,-0.0974157006,-0.5258919597,-0.0305760354,-0.295525372,0.1912617087,0.4583674669,-0.0001182336,-0.0913553163,0.2139718682,0.0120125022,0.0930700898,-0.4430084825,-0.2259867936,-0.2528820932,-0.1637293845,-0.0408718698,0.3133925498,0.0046791183,-0.0663942918,-0.150278613,0.0759297609,0.7205310464,0.1591092348,0.214832738,0.2288711667,-0.1120013967,0.016339317,0.1341472566,-0.0305977836,0.061437469,0.4003654718,-0.2003950328,0.0249149706,-0.1673947722,0.092808485,-0.2669810057,0.305939585,-0.1227102131,0.1998042613,-0.6616144776,0.2974792719,-0.1724279821,-0.1635968983,-0.2878350317,-0.0509196632,-0.1632319391,-0.1430846751,-0.2230149657,0.177413553,-0.1671895236,-0.1571241021,0.0832860544,-0.2861917019,0.1699507236,0.0335625187,0.3086440563,-0.3192946315,0.4002259374,-0.0029774297,0.389618665,0.1262956411,-0.2813790739,-0.3420287967,0.1340495944,-0.0415284634,0.2674230039,-0.2257158607,-0.1663678735,0.1507905126,-0.1738581359,0.2904581726,-0.2406045794,-0.1656422466,0.2874025404,0.048612088,0.2133101672,0.2980216742,-0.0209573247,0.3021420836,0.3840674162,0.0499104597,-0.056064833,-0.102803044,-0.0400411636,0.2264666557,-0.1167617217,0.0892518535,0.1257736236,0.0740240142,-0.1534793824,-0.1198020428,0.4294593036,-0.2137423903,0.2773050666,0.3186097741,0.2963832617,0.3739162087,0.227028951,-0.2862907052,0.0338729285,-0.2570652664,-0.0871776044,-0.0442779921,-0.2006773353,-0.3708454072,0.0992647409,-0.005736412,0.239634648,0.1527641416,0.0450480506,0.2212321013,-0.539868176,0.1125220954,-0.2573070824,0.1339917481,0.0350833498,-0.2941997945,0.1408175528,0.1156675741,-0.1236832365,-0.1660993546,0.1671806574,-0.2124153674,-0.2457581609,-0.1474966705,0.1560774744,-0.0041650292,0.4375189543,-0.0617898293,0.3228711784,0.3464185596,-0.3285045028,-0.0258732624,-0.3166899085,-0.4711762369,-0.2692376971,-0.0478229038,0.2325534225,-0.1191353053,0.0298826676,0.2000778913,0.0202775728,0.0941302255,0.1629669219,-0.1979359239,-0.0177588277,-0.1911362857,0.5243681073,0.2797337174,-0.0643637404,-0.4906041324,0.1102769822,-0.2673051357,-0.1981518269,-0.4909549952,-0.1560129821,0.4023430943,-0.2073682845,0.4260299802,-0.0168935806,-0.1412145048,0.118385382,-0.306717515,0.094998017,0.1623565853,-0.1462418735,0.0806366652,-0.166501984,0.1831175238,-0.0708336085,0.0203282014,0.004628906,-0.0426518135,0.2887397707,0.2136349082,-0.0781941861,0.0351790115,-0.3048202991,-0.0683804154,0.2268817574,0.0347599015,-0.0888614431,-0.0282383524,0.0105143087,-0.0411652811,0.3073398471,0.1391114891,-0.1058630794,0.0216380805,-0.1246979013,0.0030890247,-0.0661532804,-0.0376062728,-0.0197845213,-0.1885098517,0.0246681049,-0.0632356256,-0.1272604764,-0.2010300606,-0.1646624357,-0.3887704015,0.1416743398,0.4239653945,-0.0064662574,-0.2309417278,0.5506316423,0.3362869918,-0.2484891564,-0.6358670592,-0.0847648457,0.3994875848,-0.1538935304,0.1634524167,0.0869446471,-0.0126662534,-0.1323807985,0.1911210567,0.2216383666,0.2824165821,0.1572643816,-0.0959971696,0.0425383486,0.1659151316,0.1229683161,0.0325946435,-0.199791953,-0.0025342163,0.1199669093,0.0561050922,-0.047854811,-0.0479735769,0.2361027151,0.2793250382,-0.0761121586,0.1238081977,0.0580437221,-0.1685209721,-0.0567469634,0.2826824784,-0.0136068668,0.6230338812,0.0613569953,-0.0287729017,0.0040426482,-0.3801935911,-0.0336282663,0.1335722506,0.0782550871,-0.1228369474,0.0503613725,0.2556922138,0.159722358,-0.1924543381,-0.3795136213,0.0199823845,0.3146329224,-0.4612223506,0.2040783465,-0.1188221872,0.2136378586,-0.0778881684,-0.1706882864,-0.1753237247,-0.3915620148,-0.0386055,0.1213032827,0.0050780186,0.1582536548,0.0887682661,-0.1417624056,0.04338919,0.3126336932,0.2620718777,-0.0042142225,-0.4652619362,-0.1256633401,0.3646590412,-0.1426314861,0.1807572693,0.1596735418,-0.4231134653,-0.1377694905,-0.554107368,0.0362168029,0.0458122902,0.2179297954,0.0650527328,0.0002729307,-0.1417898685,-0.0640055388,0.1627510339,-0.0710154772,-0.0511523709,0.1167618632,0.1763634235,-0.3484143615,-0.4004407525,0.0810818374,-0.2404452562,-0.2278887779,0.0398980491,0.0562176853,0.2899165154,0.033601027,0.3173662722,0.2378733456,0.0640233308,-0.2699205875,-0.2033462822,0.2120094001,-0.0509562194,-0.1475808471,-0.2852531075,-0.0752577856,0.1772031188,-0.022968391,0.2650331259,-0.2499954402,-0.0535250045,-0.0659198835,0.3023351431,-0.1044890657,-0.2844492197,0.3918755949,0.1078824848,0.0479716472,-0.0043216893,-0.3812083304,-0.0652099252,-0.2201593518,0.1879525185,-0.0040869415,0.5207585096,0.1579419822,0.5066372156,0.3318142593,-0.5369669795,0.1717587262,-0.2706901133,0.1213353053,-0.1566914618,-0.1280220896,0.3013108969,0.0409448743,-0.0624374636,0.136187464,-0.0629640371,-0.4967781901,0.0518860146,0.0803500637,-0.0168010574,-0.2850890756,0.1508056819,-0.0602970831,0.2637024224,-0.1313890517,0.3329266608,-0.2499023676,-0.0794402286,-0.1635815203,-0.0973349959,-0.1765634567,-0.0466502458,-0.534514606,-0.1086725891,-0.6624104381,0.3221202195,0.1939269751,0.7835565805,0.0214179475,-0.0824445039,-0.2468029261,-0.1681991518,0.6353745461,-0.300203532,0.0723909363,-0.0045123673,0.098818697,-0.6122601032,-0.1256724149,-0.0143848732,0.6673847437,0.2643587291,0.9284402132,0.0463600494,0.045960132,0.0434808359,0.0178261716,0.1433940679,-0.0652095973,-0.0105038164,-0.0599536635,-0.1763828248,0.1956971437,0.2371133864,0.207475841,-0.0157400426,-0.4159704745,0.0754318386,-0.3766656518,0.2011009455,0.1319937855,0.208466664,-0.2982677221,0.1140266657,-0.0784145966,-0.1573798358,0.2803728878,-0.1881632805,0.0264791045,-0.3439128697,0.2833440304,0.1530024409,0.1995321661,0.426977247,-0.0416696221,0.1061635017,-0.3421507776,0.3123049438,-0.2047104388,0.2691136599,0.2535805106,0.2167814076,-0.5693432689,0.0799898058,0.1540472656,0.0905904621,-0.0760773271,0.3940853775,-0.368170321,-0.4619640112,0.3434622586,0.1300132424,0.848302424,0.0364213064,-0.081603311,-0.19214347,0.1661395729,0.1686700284,-0.1858166009,-0.0507305525,-0.3904622197,-0.1557223946,-0.0677676797,-0.1420495659,0.0983293951,0.0727663487,-0.3925172985,0.1129100174,0.1099763438,0.4424754381,-0.0006535093,0.094351247,-0.1428740174,-0.1273342818,-0.0686014518,0.1570535004,0.1238408089,-0.2321622223,-0.14562352,0.3469045758,0.1431711614,-0.4613877833,-0.62011379,-0.0647350997,-0.5778717399,0.1768615544,0.1812040657,0.2300380021,0.1644968092,0.301273495,0.0709967613,0.506118238,-0.2116412073,-0.0245072097,0.0392306633,0.0512521602,0.1469227374,-0.1345072687,-0.0176817924,0.0910632312,-0.0327797458,0.0640930757,0.001057488,0.1164323166,-0.501894176,0.265417248,0.4029581249,-0.196479097,-0.2089856416,-0.2772096992,-0.1896944642,-0.2230817527,0.0718460232,0.0842439681,-0.1488311887,0.1394552439,-0.1531652212,-0.2392156124,0.0301355962,0.4575230777,-0.2256522924,0.0981041044,-0.0154942367,0.0785034224,-0.0677308589,-0.1901301593,0.2326962501,-0.1237555444,-0.0325517394,-0.2231186181,-0.2007786334,-0.0984243006,0.1080042869,0.0606934577,0.1129072234,0.0262527335,-0.016009355,-0.5448374152,-0.4919656813,-0.0073139649,-0.0454394147,0.3145165741,-0.0165020507,0.0936865062,-0.2128308862,0.1364151239,-0.2179399729,0.1666069627,-0.0632879138,0.2086213976,-0.0276215877,0.0983446389,-0.1632549614,-0.1979133934,0.0605035052,0.7531200051,-0.2359837741,-0.0942862928,-0.0525906458,0.1379344016,0.2398593724,-0.1094310135,-0.1033161655,-0.0272614323,0.1039245501,-0.2421072721,0.1028843969,0.2268958539,-0.2034132034,-0.131596297,0.6177958846,0.3815800548,0.1931364983,0.357540071,0.0191768613,0.2638332546,0.1465423852,0.3621217012,0.1001960859,-0.0010375903,-0.0883050188,0.2816980779,0.0660891756,0.1587927341,0.2086260915,-0.0950449184,0.1765199453,0.0912955999,-0.0393662527,0.3638717234,-0.4002837837,-0.0639449432,0.191659078,0.1843820661,0.0324680954,-0.2394108027,0.3049276471,0.1877780557,0.0064044776,0.2658642828,0.0362890139,-0.3785454035,0.079316467,0.2469057888,0.3529776633,-0.0388915651,-0.0370597355,0.2587002814,-0.2121544927,0.1043223813,0.2033436149,-0.1946918219,0.3131908774,0.6266698241,-0.0740214512,0.4274179339,0.1257343292,0.0630890504,-0.0054944996,-0.2483475804,0.0808186531,0.4405095875,-0.1837271601,0.1500857174,0.0620873384,0.257960856,-0.3250874877,-0.1628504544,0.0823815241,0.0784967542,-0.3235331476,-0.0777549371,-0.0641655922,-0.1524268389,-0.1851560324,0.0298068896,-0.1491112709,-0.134274587,0.3058646917,0.105733268,-0.3869328499,-0.2465470433,-0.0842965916,0.0462458469,-0.0254337527,-0.2964484394,0.5547583699,-0.165271312,0.2051642239,0.1515772492,-0.1810009629,0.5931897759,0.6459059119,-0.2287395895,0.1645237505,0.1871111542,0.0369282775,-0.0236070659,0.1389949471,0.0064930003,0.2090339661,0.2872698605,0.0319978818,-0.1442494094,-0.1472728252,0.3358566463,0.2155430019,-0.0512180105,0.1997416914,-0.2628280818,-0.1870484501,0.0019922892,-0.0307528339,-0.1691651791,-0.0553508326,0.3815113008,-0.0056802407,0.4061158597,-0.1410945505,0.0556682125,0.1124868616,0.165840745,0.2160716504,-0.1539394557,-0.4025766253,-0.3473495245,-0.3931940496,0.3725878298,-0.2325892448,0.1685742289,0.0681374967,-0.2374645174,-0.0817601308,0.003409601,0.1784852594,0.2687878013,-0.0693608075,0.5016827583,-0.3206634223,-0.0434216745,-0.0729406029,-0.1339516491,0.2401047647,-0.4375864863,0.4043382704,-0.3041529357,-0.0143265808,-0.1915714145,0.0053744041,-0.227931276,-0.0372267179,0.6843011975,0.2150361538,0.4069308043,-0.0329385772,0.1105626151,-0.2053319067,0.1099141538,-0.2095181644,0.2924145162,0.0320701674,0.1326195002,-0.1456784308,-0.1236873269,-0.072567299,-0.1102667898,0.2603496611,-0.0558728129,-0.2727081776,0.1199415252,-0.2002013028,-0.0307991542,-0.1286313683,0.3280451596,0.1533139497,0.2899936736,0.0323429331,-0.2277695686,0.3998816311,-0.3358049691,-0.3362597823,-0.1846985966,0.1810190976,-0.2116550505,-0.3600912392,-0.5163121819,-0.3579574823,0.3601654172,0.0267328806,-0.5006997585,-0.0882551745,-0.2791074812,-0.0996797755,-0.2518528104,0.4753126502,0.1529987752,-0.1394924521,0.6449896693,-0.3258785605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2470","title":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.","comments":"This has been fixed in #2566, thanks @connor-mccarthy !\r\nWe'll make a new release soon that includes the fix ;)","body":"## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n","comment_length":20,"text":"Crash when `num_proc` > dataset length for `map()` on a `datasets.Dataset`.\n## Describe the bug\r\nCrash if when using `num_proc` > 1 (I used 16) for `map()` on a `datasets.Dataset`.\r\n\r\nI believe I've had cases where `num_proc` > 1 works before, but now it seems either inconsistent, or depends on my data. I'm not sure whether the issue is on my end, because it's difficult for me to debug! Any tips greatly appreciated, I'm happy to provide more info if it would helps us diagnose.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n# this function will be applied with map()\r\ndef tokenize_function(examples):\r\n return tokenizer(\r\n examples[\"text\"],\r\n padding=PaddingStrategy.DO_NOT_PAD,\r\n truncation=True,\r\n )\r\n\r\n# data_files is a Dict[str, str] mapping name -> path\r\ndatasets = load_dataset(\"text\", data_files={...}) \r\n\r\n# this is where the error happens if num_proc = 16,\r\n# but is fine if num_proc = 1\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=num_workers,\r\n)\r\n```\r\n\r\n## Expected results\r\nThe `map()` function succeeds with `num_proc` > 1.\r\n\r\n## Actual results\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404271-a6cc5200-c910-11eb-8e27-5c893bd04042.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/1170062\/121404362-be0b3f80-c910-11eb-9117-658943029aef.png)\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-glibc2.31\r\n- Python version: 3.9.5\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes, but I think N\/A for this issue\r\n- Using distributed or parallel set-up in script?: Multi-GPU on one machine, but I think also N\/A for this issue\r\n\nThis has been fixed in #2566, thanks @connor-mccarthy !\r\nWe'll make a new release soon that includes the fix ;)","embeddings":[-0.4569717646,-0.2173907459,0.0228779204,0.4219461381,0.3113077581,-0.2264823467,0.2361344099,0.1525053084,0.0431262851,0.3643293679,0.5303300619,0.2979793251,-0.3346508443,-0.0465700552,0.047729928,-0.0265675541,0.3659133613,-0.2182355374,0.2511509061,0.3584368825,-0.2708789408,0.1627700925,-0.4191713631,0.2937121987,-0.2652947009,0.082243599,-0.1050226912,0.2865630984,-0.0974157006,-0.5258919597,-0.0305760354,-0.295525372,0.1912617087,0.4583674669,-0.0001182336,-0.0913553163,0.2139718682,0.0120125022,0.0930700898,-0.4430084825,-0.2259867936,-0.2528820932,-0.1637293845,-0.0408718698,0.3133925498,0.0046791183,-0.0663942918,-0.150278613,0.0759297609,0.7205310464,0.1591092348,0.214832738,0.2288711667,-0.1120013967,0.016339317,0.1341472566,-0.0305977836,0.061437469,0.4003654718,-0.2003950328,0.0249149706,-0.1673947722,0.092808485,-0.2669810057,0.305939585,-0.1227102131,0.1998042613,-0.6616144776,0.2974792719,-0.1724279821,-0.1635968983,-0.2878350317,-0.0509196632,-0.1632319391,-0.1430846751,-0.2230149657,0.177413553,-0.1671895236,-0.1571241021,0.0832860544,-0.2861917019,0.1699507236,0.0335625187,0.3086440563,-0.3192946315,0.4002259374,-0.0029774297,0.389618665,0.1262956411,-0.2813790739,-0.3420287967,0.1340495944,-0.0415284634,0.2674230039,-0.2257158607,-0.1663678735,0.1507905126,-0.1738581359,0.2904581726,-0.2406045794,-0.1656422466,0.2874025404,0.048612088,0.2133101672,0.2980216742,-0.0209573247,0.3021420836,0.3840674162,0.0499104597,-0.056064833,-0.102803044,-0.0400411636,0.2264666557,-0.1167617217,0.0892518535,0.1257736236,0.0740240142,-0.1534793824,-0.1198020428,0.4294593036,-0.2137423903,0.2773050666,0.3186097741,0.2963832617,0.3739162087,0.227028951,-0.2862907052,0.0338729285,-0.2570652664,-0.0871776044,-0.0442779921,-0.2006773353,-0.3708454072,0.0992647409,-0.005736412,0.239634648,0.1527641416,0.0450480506,0.2212321013,-0.539868176,0.1125220954,-0.2573070824,0.1339917481,0.0350833498,-0.2941997945,0.1408175528,0.1156675741,-0.1236832365,-0.1660993546,0.1671806574,-0.2124153674,-0.2457581609,-0.1474966705,0.1560774744,-0.0041650292,0.4375189543,-0.0617898293,0.3228711784,0.3464185596,-0.3285045028,-0.0258732624,-0.3166899085,-0.4711762369,-0.2692376971,-0.0478229038,0.2325534225,-0.1191353053,0.0298826676,0.2000778913,0.0202775728,0.0941302255,0.1629669219,-0.1979359239,-0.0177588277,-0.1911362857,0.5243681073,0.2797337174,-0.0643637404,-0.4906041324,0.1102769822,-0.2673051357,-0.1981518269,-0.4909549952,-0.1560129821,0.4023430943,-0.2073682845,0.4260299802,-0.0168935806,-0.1412145048,0.118385382,-0.306717515,0.094998017,0.1623565853,-0.1462418735,0.0806366652,-0.166501984,0.1831175238,-0.0708336085,0.0203282014,0.004628906,-0.0426518135,0.2887397707,0.2136349082,-0.0781941861,0.0351790115,-0.3048202991,-0.0683804154,0.2268817574,0.0347599015,-0.0888614431,-0.0282383524,0.0105143087,-0.0411652811,0.3073398471,0.1391114891,-0.1058630794,0.0216380805,-0.1246979013,0.0030890247,-0.0661532804,-0.0376062728,-0.0197845213,-0.1885098517,0.0246681049,-0.0632356256,-0.1272604764,-0.2010300606,-0.1646624357,-0.3887704015,0.1416743398,0.4239653945,-0.0064662574,-0.2309417278,0.5506316423,0.3362869918,-0.2484891564,-0.6358670592,-0.0847648457,0.3994875848,-0.1538935304,0.1634524167,0.0869446471,-0.0126662534,-0.1323807985,0.1911210567,0.2216383666,0.2824165821,0.1572643816,-0.0959971696,0.0425383486,0.1659151316,0.1229683161,0.0325946435,-0.199791953,-0.0025342163,0.1199669093,0.0561050922,-0.047854811,-0.0479735769,0.2361027151,0.2793250382,-0.0761121586,0.1238081977,0.0580437221,-0.1685209721,-0.0567469634,0.2826824784,-0.0136068668,0.6230338812,0.0613569953,-0.0287729017,0.0040426482,-0.3801935911,-0.0336282663,0.1335722506,0.0782550871,-0.1228369474,0.0503613725,0.2556922138,0.159722358,-0.1924543381,-0.3795136213,0.0199823845,0.3146329224,-0.4612223506,0.2040783465,-0.1188221872,0.2136378586,-0.0778881684,-0.1706882864,-0.1753237247,-0.3915620148,-0.0386055,0.1213032827,0.0050780186,0.1582536548,0.0887682661,-0.1417624056,0.04338919,0.3126336932,0.2620718777,-0.0042142225,-0.4652619362,-0.1256633401,0.3646590412,-0.1426314861,0.1807572693,0.1596735418,-0.4231134653,-0.1377694905,-0.554107368,0.0362168029,0.0458122902,0.2179297954,0.0650527328,0.0002729307,-0.1417898685,-0.0640055388,0.1627510339,-0.0710154772,-0.0511523709,0.1167618632,0.1763634235,-0.3484143615,-0.4004407525,0.0810818374,-0.2404452562,-0.2278887779,0.0398980491,0.0562176853,0.2899165154,0.033601027,0.3173662722,0.2378733456,0.0640233308,-0.2699205875,-0.2033462822,0.2120094001,-0.0509562194,-0.1475808471,-0.2852531075,-0.0752577856,0.1772031188,-0.022968391,0.2650331259,-0.2499954402,-0.0535250045,-0.0659198835,0.3023351431,-0.1044890657,-0.2844492197,0.3918755949,0.1078824848,0.0479716472,-0.0043216893,-0.3812083304,-0.0652099252,-0.2201593518,0.1879525185,-0.0040869415,0.5207585096,0.1579419822,0.5066372156,0.3318142593,-0.5369669795,0.1717587262,-0.2706901133,0.1213353053,-0.1566914618,-0.1280220896,0.3013108969,0.0409448743,-0.0624374636,0.136187464,-0.0629640371,-0.4967781901,0.0518860146,0.0803500637,-0.0168010574,-0.2850890756,0.1508056819,-0.0602970831,0.2637024224,-0.1313890517,0.3329266608,-0.2499023676,-0.0794402286,-0.1635815203,-0.0973349959,-0.1765634567,-0.0466502458,-0.534514606,-0.1086725891,-0.6624104381,0.3221202195,0.1939269751,0.7835565805,0.0214179475,-0.0824445039,-0.2468029261,-0.1681991518,0.6353745461,-0.300203532,0.0723909363,-0.0045123673,0.098818697,-0.6122601032,-0.1256724149,-0.0143848732,0.6673847437,0.2643587291,0.9284402132,0.0463600494,0.045960132,0.0434808359,0.0178261716,0.1433940679,-0.0652095973,-0.0105038164,-0.0599536635,-0.1763828248,0.1956971437,0.2371133864,0.207475841,-0.0157400426,-0.4159704745,0.0754318386,-0.3766656518,0.2011009455,0.1319937855,0.208466664,-0.2982677221,0.1140266657,-0.0784145966,-0.1573798358,0.2803728878,-0.1881632805,0.0264791045,-0.3439128697,0.2833440304,0.1530024409,0.1995321661,0.426977247,-0.0416696221,0.1061635017,-0.3421507776,0.3123049438,-0.2047104388,0.2691136599,0.2535805106,0.2167814076,-0.5693432689,0.0799898058,0.1540472656,0.0905904621,-0.0760773271,0.3940853775,-0.368170321,-0.4619640112,0.3434622586,0.1300132424,0.848302424,0.0364213064,-0.081603311,-0.19214347,0.1661395729,0.1686700284,-0.1858166009,-0.0507305525,-0.3904622197,-0.1557223946,-0.0677676797,-0.1420495659,0.0983293951,0.0727663487,-0.3925172985,0.1129100174,0.1099763438,0.4424754381,-0.0006535093,0.094351247,-0.1428740174,-0.1273342818,-0.0686014518,0.1570535004,0.1238408089,-0.2321622223,-0.14562352,0.3469045758,0.1431711614,-0.4613877833,-0.62011379,-0.0647350997,-0.5778717399,0.1768615544,0.1812040657,0.2300380021,0.1644968092,0.301273495,0.0709967613,0.506118238,-0.2116412073,-0.0245072097,0.0392306633,0.0512521602,0.1469227374,-0.1345072687,-0.0176817924,0.0910632312,-0.0327797458,0.0640930757,0.001057488,0.1164323166,-0.501894176,0.265417248,0.4029581249,-0.196479097,-0.2089856416,-0.2772096992,-0.1896944642,-0.2230817527,0.0718460232,0.0842439681,-0.1488311887,0.1394552439,-0.1531652212,-0.2392156124,0.0301355962,0.4575230777,-0.2256522924,0.0981041044,-0.0154942367,0.0785034224,-0.0677308589,-0.1901301593,0.2326962501,-0.1237555444,-0.0325517394,-0.2231186181,-0.2007786334,-0.0984243006,0.1080042869,0.0606934577,0.1129072234,0.0262527335,-0.016009355,-0.5448374152,-0.4919656813,-0.0073139649,-0.0454394147,0.3145165741,-0.0165020507,0.0936865062,-0.2128308862,0.1364151239,-0.2179399729,0.1666069627,-0.0632879138,0.2086213976,-0.0276215877,0.0983446389,-0.1632549614,-0.1979133934,0.0605035052,0.7531200051,-0.2359837741,-0.0942862928,-0.0525906458,0.1379344016,0.2398593724,-0.1094310135,-0.1033161655,-0.0272614323,0.1039245501,-0.2421072721,0.1028843969,0.2268958539,-0.2034132034,-0.131596297,0.6177958846,0.3815800548,0.1931364983,0.357540071,0.0191768613,0.2638332546,0.1465423852,0.3621217012,0.1001960859,-0.0010375903,-0.0883050188,0.2816980779,0.0660891756,0.1587927341,0.2086260915,-0.0950449184,0.1765199453,0.0912955999,-0.0393662527,0.3638717234,-0.4002837837,-0.0639449432,0.191659078,0.1843820661,0.0324680954,-0.2394108027,0.3049276471,0.1877780557,0.0064044776,0.2658642828,0.0362890139,-0.3785454035,0.079316467,0.2469057888,0.3529776633,-0.0388915651,-0.0370597355,0.2587002814,-0.2121544927,0.1043223813,0.2033436149,-0.1946918219,0.3131908774,0.6266698241,-0.0740214512,0.4274179339,0.1257343292,0.0630890504,-0.0054944996,-0.2483475804,0.0808186531,0.4405095875,-0.1837271601,0.1500857174,0.0620873384,0.257960856,-0.3250874877,-0.1628504544,0.0823815241,0.0784967542,-0.3235331476,-0.0777549371,-0.0641655922,-0.1524268389,-0.1851560324,0.0298068896,-0.1491112709,-0.134274587,0.3058646917,0.105733268,-0.3869328499,-0.2465470433,-0.0842965916,0.0462458469,-0.0254337527,-0.2964484394,0.5547583699,-0.165271312,0.2051642239,0.1515772492,-0.1810009629,0.5931897759,0.6459059119,-0.2287395895,0.1645237505,0.1871111542,0.0369282775,-0.0236070659,0.1389949471,0.0064930003,0.2090339661,0.2872698605,0.0319978818,-0.1442494094,-0.1472728252,0.3358566463,0.2155430019,-0.0512180105,0.1997416914,-0.2628280818,-0.1870484501,0.0019922892,-0.0307528339,-0.1691651791,-0.0553508326,0.3815113008,-0.0056802407,0.4061158597,-0.1410945505,0.0556682125,0.1124868616,0.165840745,0.2160716504,-0.1539394557,-0.4025766253,-0.3473495245,-0.3931940496,0.3725878298,-0.2325892448,0.1685742289,0.0681374967,-0.2374645174,-0.0817601308,0.003409601,0.1784852594,0.2687878013,-0.0693608075,0.5016827583,-0.3206634223,-0.0434216745,-0.0729406029,-0.1339516491,0.2401047647,-0.4375864863,0.4043382704,-0.3041529357,-0.0143265808,-0.1915714145,0.0053744041,-0.227931276,-0.0372267179,0.6843011975,0.2150361538,0.4069308043,-0.0329385772,0.1105626151,-0.2053319067,0.1099141538,-0.2095181644,0.2924145162,0.0320701674,0.1326195002,-0.1456784308,-0.1236873269,-0.072567299,-0.1102667898,0.2603496611,-0.0558728129,-0.2727081776,0.1199415252,-0.2002013028,-0.0307991542,-0.1286313683,0.3280451596,0.1533139497,0.2899936736,0.0323429331,-0.2277695686,0.3998816311,-0.3358049691,-0.3362597823,-0.1846985966,0.1810190976,-0.2116550505,-0.3600912392,-0.5163121819,-0.3579574823,0.3601654172,0.0267328806,-0.5006997585,-0.0882551745,-0.2791074812,-0.0996797755,-0.2518528104,0.4753126502,0.1529987752,-0.1394924521,0.6449896693,-0.3258785605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2459","title":"`Proto_qa` hosting seems to be broken","comments":"@VictorSanh , I think @mariosasko is already working on it. ","body":"## Describe the bug\r\nThe hosting (on Github) of the `proto_qa` dataset seems broken. I haven't investigated more yet, just flagging it for now. \r\n\r\n@zaidalyafeai if you want to dive into it, I think it's just a matter of changing the links in `proto_qa.py`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"proto_qa\")\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 630, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/hf\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/proto_qa\/445346efaad5c5f200ecda4aa7f0fb50ff1b55edde3003be424a2112c3e8102e\/proto_qa.py\", line 131, in _split_generators\r\n train_fpath = dl_manager.download(_URLs[self.config.name][\"train\"])\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 199, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 195, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 218, in _download\r\n return cached_path(url_or_filename, download_config=download_config)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/iesl\/protoqa-data\/master\/data\/train\/protoqa_train.jsonl\r\n```","comment_length":10,"text":"`Proto_qa` hosting seems to be broken\n## Describe the bug\r\nThe hosting (on Github) of the `proto_qa` dataset seems broken. I haven't investigated more yet, just flagging it for now. \r\n\r\n@zaidalyafeai if you want to dive into it, I think it's just a matter of changing the links in `proto_qa.py`\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"proto_qa\")\r\n```\r\n\r\n## Actual results\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 751, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 630, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/hf\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/proto_qa\/445346efaad5c5f200ecda4aa7f0fb50ff1b55edde3003be424a2112c3e8102e\/proto_qa.py\", line 131, in _split_generators\r\n train_fpath = dl_manager.download(_URLs[self.config.name][\"train\"])\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 199, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 195, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 218, in _download\r\n return cached_path(url_or_filename, download_config=download_config)\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/hf\/dev\/promptsource\/.venv\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/iesl\/protoqa-data\/master\/data\/train\/protoqa_train.jsonl\r\n```\n@VictorSanh , I think @mariosasko is already working on it. ","embeddings":[-0.2379052341,-0.1363887638,-0.1059827581,0.2375226766,0.3108732104,-0.0629653111,0.1478208154,0.3436966836,-0.2986535132,0.176171422,-0.1861490607,-0.1167649403,0.1058836803,0.4656740427,0.2429773957,-0.1706756949,0.0337843932,0.0838806182,-0.1692854911,0.1520941257,-0.016307164,0.0634398162,-0.0931490064,0.2198267877,0.1650674045,0.1234978512,-0.0810085237,0.0410941094,-0.0079999166,-0.3936422467,0.0567568466,-0.0287063029,-0.0193583556,0.5810063481,-0.0000991802,0.1633045673,0.2867327929,-0.0088414419,-0.1739728451,0.2184001058,-0.3235583007,0.0265823714,-0.2811763883,0.1126966998,-0.1597103775,-0.4038235247,0.091311805,-0.0835802853,0.4007878602,0.1863644272,0.3341585398,0.3828071952,0.1708706021,-0.233302623,-0.0297626108,-0.0368175693,-0.1694797426,0.0747401491,0.1489881724,0.0057666595,0.2261247337,0.0826742873,0.008426128,-0.1188435704,-0.1088934019,-0.2838004529,0.0102369273,-0.4537743926,0.2315013409,0.287866801,0.3075620234,-0.1827702075,-0.1664185077,0.0745612308,-0.0047275429,-0.4977577925,0.0735178292,0.3936243653,-0.0378586724,0.2590829134,0.2056573182,-0.1066472307,-0.1351154149,0.0078298235,-0.1148480177,0.3386599422,-0.0320594311,-0.0874797702,-0.2843167186,0.021725012,0.1161626354,0.0827374384,-0.2362588495,-0.165853411,-0.1754939109,-0.0638196096,0.2094714344,0.1432018876,0.187255159,0.2741155028,-0.0584228262,0.1576451063,-0.3430628181,-0.0393534824,0.0413359515,0.0108624473,0.0675265193,-0.0860355049,0.3215428293,0.5422900915,0.1050070599,0.0908572376,-0.1619945019,-0.1993871927,-0.0989580229,-0.0012015839,0.2965599298,-0.0795397833,-0.0896079615,0.1232589409,-0.3004104793,0.0723948404,-0.233921662,0.3159494698,-0.0625911579,0.1020462513,-0.1685367525,0.2210589498,-0.2446062416,-0.4127288461,-0.2823371589,-0.2642573118,0.0368815437,0.0690320134,-0.0792538673,-0.0678116679,0.2291734517,0.0696935356,0.2047234625,-0.1204469204,-0.0000767881,0.039944414,-0.1570401043,0.3807753026,0.1314163804,0.1121471301,0.0085202176,0.015389123,0.020351056,0.0236650612,-0.2174049914,-0.3757149875,-0.210259974,0.3165300488,-0.0479158945,0.1074235439,-0.2266753465,-0.0242353976,0.0298005864,-0.1748613119,-0.0867517591,-0.0327262357,-0.0726491883,-0.105627127,0.4456382394,0.2859564126,-0.1526963711,-0.0077756532,0.0998897478,0.0410335958,0.1350185573,0.0829578787,-0.1225597858,0.1582617015,-0.288018465,0.1121271402,0.0076100277,-0.1597019732,-0.2040525675,0.0148129258,-0.0974145159,-0.2838821709,-0.0843638852,-0.378709197,0.0371232256,-0.005614555,0.0860170797,0.3713031709,0.1164556369,0.0671391562,-0.3918760419,-0.3896547556,-0.2551313341,0.0611317642,-0.0686672032,-0.0464640558,-0.0952262059,-0.1296622157,0.413117379,-0.0287680347,-0.0565860383,0.1632500887,0.2470013797,0.0758249834,-0.0477086492,-0.0091556758,-0.0971810073,0.1064389721,0.0085161421,0.3159912527,-0.0350077376,0.0168817267,-0.3450975716,0.113963671,-0.2389431,-0.3516870737,0.3491958976,0.2255755216,-0.2835783958,0.1713144332,-0.0219909474,-0.1337836534,0.1841474026,0.08519274,-0.3547540307,0.3096849024,-0.2474432439,0.155391857,0.0874311253,0.2418213338,0.1831498742,-0.2009785771,0.1161594316,0.4202190042,0.1324905604,0.2107793242,0.1448425949,-0.0515472218,0.3153675199,-0.2011332959,-0.0094874613,0.0052903397,-0.0459901579,0.0667848289,0.2105385512,0.0301911589,-0.0966025591,0.3223456442,0.0241448656,0.2306976616,0.3043802381,0.0498368628,-0.1869328022,-0.0156777631,0.2744500339,0.3699849844,-0.0022554724,-0.1649072766,-0.2922696173,0.4163262248,0.1036101133,-0.1583666205,0.0441981666,-0.1805638373,-0.3340841532,-0.0096770646,0.2401475161,0.3947191536,0.163410455,0.0574864373,-0.0028029943,0.1044513732,0.101074785,-0.3407148719,0.3349455297,0.069751218,-0.01898445,0.0218939036,0.3180417717,0.0259419195,-0.2255016416,-0.104128696,-0.157069996,-0.0165163446,-0.3663857281,0.286691606,-0.374522984,0.1897575557,-0.2494781166,-0.2513636947,-0.3060948849,-0.3937913179,0.0605279468,-0.0826343149,-0.2188115418,0.288351655,0.0662028491,0.108763583,-0.014548867,-0.1933072507,0.0890485123,0.0733586401,-0.30545035,0.1706156284,0.1282383651,-0.0866463929,0.4270478189,-0.2240703106,-0.3345175087,-0.1096382588,-0.2223081887,0.2351237833,-0.0616551116,0.4050990343,0.1746190488,0.3247613311,0.3603941202,-0.1402509958,0.3394928277,0.0840431303,0.068611823,0.12327905,0.1333150417,0.1227300689,-0.2584322691,-0.0995959342,-0.0345869027,-0.2921948731,-0.0568870679,0.1294692606,0.2554420233,0.0355724245,0.0359341465,-0.0567868315,-0.2359442711,-0.1357828975,-0.3341325521,-0.2289307564,0.2243503034,-0.3328276277,-0.3328516781,-0.1110699326,0.1875952631,0.5986089706,-0.0853982419,-0.5405715704,-0.2938596308,0.0005169924,0.2016407847,0.0260554496,0.0838964954,0.1563915759,0.1820695549,-0.226485908,0.035505373,-0.0395731404,0.0052556866,0.0569705926,0.1621282399,-0.0502945483,0.3163736165,-0.0788374096,0.7790765762,0.0989878327,0.1596416682,0.3119193912,-0.1659963578,0.4360764921,-0.1468851566,-0.1365059018,0.1822456867,0.1197723597,-0.116231069,0.2311961949,0.2345326245,-0.0255398117,-0.2846832871,0.1262333244,-0.1488203257,-0.3882617354,-0.0916604251,-0.1511442512,0.2439389378,-0.0169075206,0.2845713794,0.1095016524,-0.0949033797,0.1961275488,0.119895108,-0.0048303185,-0.2198240757,-0.2745086551,-0.2935863733,-0.5645080805,0.1875353307,-0.4273892343,0.2308639139,-0.0911335498,-0.2385017276,0.1820375323,-0.0306847561,0.2843280137,-0.0431542024,0.2583745718,0.0819342658,0.1323289126,-0.232459873,-0.1942332834,-0.11673785,-0.0966765806,-0.0771452412,-0.0325109065,-0.4118260145,0.0342628658,0.1612755805,0.0229102895,-0.1778585613,0.1775072366,-0.2367499471,-0.4258963466,-0.1518913507,-0.1381720901,-0.0010059935,0.1245913953,-0.2717770636,0.0872168317,0.076463826,0.0445787944,0.0330114104,0.196016416,0.1163989231,0.0436647646,0.018499393,0.2299005538,-0.0044991234,0.2490671873,0.1886816025,-0.016108131,-0.2458720952,0.036886055,-0.026305208,0.1349474341,0.3711208999,-0.0972258672,-0.0743455812,-0.0964121968,0.349948734,-0.4175291955,0.0490532964,0.5251778364,-0.1900066435,-0.2276615351,-0.3218893707,0.4132740498,-0.1635806113,-0.0722097382,0.2187487483,0.1396527886,-0.2659014761,0.564987421,-0.0976491272,0.8777089715,0.1769711226,0.1712427735,0.4936151505,-0.0632367209,0.253185451,-0.2444020212,0.1190773323,-0.2936998904,-0.2367860377,-0.0621664152,0.0078682024,-0.1141776666,0.0539765693,-0.2565304637,0.2667337358,-0.2327837795,0.0900523886,0.1556145251,-0.200735867,0.1046291739,-0.0151282093,-0.6044342518,0.2938360572,-0.0532032698,0.2777733803,-0.1493840516,-0.0536443442,0.1411349326,0.0468634218,-0.4989023805,0.1563367397,-0.2579948008,-0.3110890388,-0.203380242,-0.0469373725,0.0780885145,0.1145275012,0.0491839424,0.5085526109,-0.4008268714,0.3667818904,0.030594267,-0.0047982824,0.1267133653,-0.1222322285,-0.0534644648,-0.1878515333,-0.2836860716,0.1747339368,0.07940837,0.0256478433,-0.1559932679,-0.0218634643,0.2307053953,-0.2820183933,0.1054166779,0.0465062112,0.3338932991,-0.2876547277,0.2579128444,-0.0732366815,-0.0539885238,-0.1556790918,0.162071541,-0.1363334507,-0.1573515683,0.1793420315,-0.0464822724,0.269978404,0.476791054,0.2784003615,0.0451764762,-0.3182638586,0.105186969,-0.2159634084,-0.3155460358,0.0665367693,-0.0368154384,-0.1426376104,-0.1725613326,-0.0875330567,0.0264194552,-0.3243028224,-0.0205735024,-0.2999615967,0.0516892225,0.1542527825,-0.0218412317,0.0549885072,0.1606357992,0.2591747344,-0.0632514134,0.0426593237,-0.4400584698,0.1290011108,-0.0216118898,-0.0688736811,-0.1312078089,-0.1184381321,0.1688165516,-0.0365940444,0.2252024412,-0.1905544698,-0.2858352661,-0.3253731132,-0.1967855692,0.0647978336,-0.1137295663,-0.13741377,-0.0145297507,-0.0978542939,-0.1364192069,-0.0518773198,0.1670124978,-0.3108180463,0.0584008023,-0.0507012457,0.6031697989,0.1621308774,-0.1115588248,0.209192723,0.02589879,0.35926193,0.1388672441,0.0381010734,-0.1134434342,-0.2231873572,-0.0836026073,0.197732985,0.0756033957,0.3848229051,0.4043203294,-0.1562736779,-0.1194169149,0.0211278126,0.3780199885,0.3719846308,-0.2595867515,-0.0940753967,0.1317213178,0.3849290907,-0.1545469314,-0.0864655375,0.079724662,0.0944697484,0.0420704521,0.031888105,0.0938057229,-0.0177677535,-0.0722091347,0.3899414539,0.0387967825,-0.2457260191,0.3045012653,0.3916954994,-0.0883964971,0.0915950164,0.1612267643,-0.343773216,0.0303653516,0.3487805128,0.2343050539,0.3782423139,0.3454763889,-0.1950470358,0.2454549521,-0.3048836887,-0.0952888355,0.1244551241,0.1326957047,-0.0463906936,0.0438142121,0.4446778297,-0.3664544225,-0.1538828313,-0.4007712007,0.2283639461,0.0265298542,-0.1576373577,-0.1185586005,-0.1045647487,-0.159697026,-0.0590527952,-0.0854323655,-0.1885849237,0.1943442672,-0.0063108592,-0.2724882662,-0.3698686659,-0.1513001621,0.0331107676,0.2608832419,0.0627672747,0.2494049519,0.1527308226,-0.117408663,-0.0560474172,0.4864452183,0.4589995742,-0.0315459594,-0.2299812585,-0.1433864981,-0.0576763786,0.0670613497,-0.148945868,0.2082095593,-0.0058168452,0.039401032,0.2283656448,0.2847266197,-0.2700331211,-0.2212295383,0.1580992639,0.1149126217,-0.0897103101,0.4332891703,-0.2159721255,-0.1642546207,-0.2012971789,-0.1377545595,-0.3023885787,0.4225520492,0.4049167335,-0.059008196,0.0526108183,-0.3431285322,0.1636934131,0.0672898144,0.7482516766,0.2475824803,-0.1975529194,-0.0528740697,-0.2334896326,-0.4435038567,0.2637144327,0.0405788384,0.2390314043,-0.1302861124,-0.0501234494,0.0510524698,0.1321127117,-0.0862904862,0.0997540578,-0.10530518,0.2124385387,-0.4610484838,0.124619402,-0.1905298978,0.1961841583,-0.123857677,-0.3027668893,0.4009636641,-0.1307563335,0.131031096,0.3045720458,-0.2147110552,0.1879058033,-0.2575601935,0.4886960387,0.2320825756,-0.2090862691,-0.1006136015,-0.3417688608,-0.5164040327,-0.3741610348,-0.1594553888,0.3071830273,-0.0016410142,0.0120794065,0.1144720092,-0.0803177953,-0.2212474495,0.3589507043,0.1984108388,0.0704772547,-0.3132013083,0.0230560508,0.062855944,-0.1389336139,0.1340041608,0.0975983739,-0.1566055119,-0.0134242903,-0.3114956617,-0.503806293,0.2778944969,-0.0711474046,-0.1194042712,-0.1359512359,0.356230855,-0.0999076292,-0.0358042717,-0.2968714833,0.2774446309,0.1749936193,-0.0046842461,-0.2442577928,0.448746562,-0.1983329505,-0.0202634148,-0.106439285,0.8370672464,0.0111923805,-0.4435793161,0.1709634513,0.0592583567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2458","title":"Revert default in-memory for small datasets","comments":"cc: @krandiash (pinged in reverted PR).","body":"Users are reporting issues and confusion about setting default in-memory to True for small datasets.\r\n\r\nWe see 2 clear use cases of Datasets:\r\n- the \"canonical\" way, where you can work with very large datasets, as they are memory-mapped and cached (after every transformation)\r\n- some edge cases (speed benchmarks, interactive\/exploratory analysis,...), where default in-memory can explicitly be enabled, and no caching will be done\r\n\r\nAfter discussing with @lhoestq we have agreed to:\r\n- revert this feature (implemented in #2182)\r\n- explain in the docs how to optimize speed\/performance by setting default in-memory\r\n\r\ncc: @stas00 https:\/\/github.com\/huggingface\/datasets\/pull\/2409#issuecomment-856210552","comment_length":6,"text":"Revert default in-memory for small datasets\nUsers are reporting issues and confusion about setting default in-memory to True for small datasets.\r\n\r\nWe see 2 clear use cases of Datasets:\r\n- the \"canonical\" way, where you can work with very large datasets, as they are memory-mapped and cached (after every transformation)\r\n- some edge cases (speed benchmarks, interactive\/exploratory analysis,...), where default in-memory can explicitly be enabled, and no caching will be done\r\n\r\nAfter discussing with @lhoestq we have agreed to:\r\n- revert this feature (implemented in #2182)\r\n- explain in the docs how to optimize speed\/performance by setting default in-memory\r\n\r\ncc: @stas00 https:\/\/github.com\/huggingface\/datasets\/pull\/2409#issuecomment-856210552\ncc: @krandiash (pinged in reverted PR).","embeddings":[-0.2187991738,-0.5008435249,0.038315773,0.0978243351,0.1504630148,-0.1621025354,0.0189160705,0.3139829338,0.0372855626,0.3439128995,0.0853734761,0.0300496612,-0.0116538294,0.1190378889,0.0542465262,0.179468587,0.0787219852,0.0944986939,-0.36097458,-0.0633965582,-0.2260774374,-0.2083972692,-0.2288212478,-0.2621673346,-0.2838351429,-0.0180836283,0.1684918255,0.0987582356,-0.2535508275,-0.3965057433,0.3366418779,0.0423141196,-0.051139012,0.0532257147,-0.0001038748,-0.0089625744,0.4153715074,-0.0051870761,-0.0316593461,0.2055862248,-0.1945186257,-0.2797249258,-0.0352295265,-0.1192184091,0.1576732546,0.1043663025,0.2574567497,-0.5586274266,0.2473321408,0.2402334958,0.2925077379,0.0204121247,-0.3377669454,-0.1195778698,-0.0435287319,0.2666583061,-0.1501745284,-0.0550282076,0.2429473251,-0.1313346624,0.0164472442,0.1635841727,-0.1357681304,0.3388287425,0.4272738397,-0.0745303184,-0.1531906873,-0.2241694033,0.0289171059,0.3394555151,0.4031113982,-0.3054738343,-0.3166726828,-0.336563617,-0.0127380872,-0.2298821807,0.1346246451,0.1250378639,-0.0193970855,0.1834094524,-0.5192624331,-0.1441462338,0.049592644,0.0237567443,-0.2660960555,0.005445655,-0.152847901,0.0497653559,0.0752468929,-0.0054889969,-0.0936995074,0.0192247201,-0.2397973984,-0.0265446268,-0.123363927,-0.166468218,0.2905315161,0.2861329317,0.497743845,-0.0229932964,0.1576466858,0.1888439953,0.0645636767,-0.0443753898,-0.0344133638,0.3708163798,-0.0406256579,-0.0846233219,0.6153429747,-0.02001369,0.032588575,0.0775510594,0.3121501207,-0.1623281091,0.2879938483,-0.1401370913,0.2666312754,-0.2415783256,-0.1122572646,0.1538859904,-0.1657716036,0.0753609985,-0.0275297798,0.2790061533,-0.0117702242,0.2919765413,0.0169481635,0.0045900061,-0.3660779297,-0.1108155921,-0.2292646617,-0.0875369459,-0.3304286599,0.2586780488,0.3561860323,-0.0468150228,0.2806923985,0.0771557763,0.178475067,0.1071834713,0.2821258605,-0.3107624054,0.1694002897,0.1530413777,-0.3773081601,0.234989509,0.0972611234,0.1201708317,-0.3419399261,0.3187859654,-0.2635608017,-0.5300819874,0.3335213661,0.2195550799,-0.1049155816,0.0869130716,-0.2520645261,0.3525859714,-0.0861539543,0.0905354545,0.0379256196,-0.2003478557,-0.2640528381,-0.1589144766,0.2109175175,0.2933341265,-0.3656395674,-0.0272886716,-0.0742472634,-0.1162402555,0.1355763674,0.4752572179,-0.4695717692,-0.0307420716,-0.083156161,-0.0572801456,-0.1003074124,-0.0422323309,-0.5617979169,0.0234913975,0.1153202504,-0.2130565494,-0.111141488,0.2588202953,0.193094492,0.114731662,0.0961397812,0.2192376554,0.1643733531,0.5013756752,-0.5176610947,-0.4049455225,-0.019195376,0.1822059155,0.2393774539,0.164800033,-0.0856705308,-0.0233060643,0.013348422,0.0592914447,0.0671228766,0.2907777429,0.3948366344,-0.1980810761,0.1101625487,-0.0218950249,-0.4909823239,0.2632967532,0.1339683235,-0.078604877,0.1569459438,-0.3188015521,-0.0040375004,0.0023089373,0.3200862408,-0.0169770941,0.1807946563,0.0178735461,0.102681078,-0.1612992585,-0.407512784,0.2792132199,-0.197409451,0.1029198915,-0.0709739625,0.0732711703,0.1266524345,-0.0890725553,-0.0976292044,-0.1623296738,-0.1977823824,-0.0141749056,-0.0193781462,0.4939626455,0.0866722688,0.0783568919,-0.0345104337,0.294057548,0.2893883586,-0.1948402524,0.307926178,-0.1502967626,0.0721064433,-0.0000731688,-0.4095214009,0.2281537056,0.0660443455,0.0197042376,0.0654478222,-0.3284133971,-0.1339641809,-0.1908281744,-0.1750675589,-0.1708096713,0.0542207547,0.1190228537,0.1896179765,0.385215044,-0.5036543608,-0.2073132843,0.126617834,0.0539586842,-0.016151458,0.233786881,-0.049372144,-0.2090923041,0.2222340703,0.0691305995,0.4694780409,0.4068779945,0.1208579168,0.0883494616,0.0259011146,0.0092116632,0.1484451443,-0.0714296699,0.1900357157,-0.0566708408,0.4208894074,-0.0461499952,-0.346485436,0.1345022172,0.0401266292,0.3207698464,-0.1678870171,-0.241426751,-0.3049805164,-0.3166844249,0.0095691243,0.318757236,-0.1706967652,-0.1175443009,0.0543438606,-0.107757166,-0.1832354218,0.2067876905,-0.0919625536,0.6479174495,-0.1595799178,0.0993080661,-0.2574399412,0.1049550772,-0.1754477322,0.0572270043,-0.0741915926,-0.0970684513,0.3152087927,0.2318614274,-0.3069363236,-0.2792355716,-0.3482031822,0.1446471065,0.120751068,0.4473277628,-0.0046773814,0.3845615387,-0.0230327286,0.0647039711,0.1038242579,-0.2266568244,0.0213144999,-0.1796644032,0.0048805885,0.22665824,-0.3578588068,-0.2300420552,-0.0344586521,-0.4237993956,0.1857672781,0.0955522433,-0.005766992,0.1077113077,0.1229364052,-0.246156767,-0.011212186,0.2487351596,-0.2128057629,-0.5337743163,0.3445532024,0.0834102258,-0.2076478302,-0.2552721798,0.0597054735,-0.0830281302,0.3750703335,-0.6215053797,-0.3345666528,-0.4536195099,0.0863480866,0.0903064832,0.1497584432,0.3318160772,-0.0444064327,-0.1834748685,0.0436025523,-0.3203803301,0.0367129669,0.005715475,0.1289768815,0.1293694228,0.4308817685,0.3028620481,0.6074613333,0.0640972704,-0.0032882744,0.0975514054,0.2012017369,0.322142452,-0.3475748301,0.1936393976,0.0672604591,-0.0471771695,-0.2952454984,0.14979361,0.1185408607,-0.4228839278,-0.0413677879,0.2674019337,0.0188530628,-0.1295238137,0.020882817,0.0069387141,0.1460891366,0.231441915,0.0065629496,-0.290324986,-0.0525488406,0.0537088327,0.0025712452,0.1048772559,0.0268391371,-0.4852460623,0.0934043229,-0.5561278462,0.1451216042,-0.1924943328,0.2606886029,0.0254478306,-0.3673080504,0.1004319862,-0.0004155005,0.394788444,0.0044884868,-0.2718369365,0.0234411433,-0.1950940192,-0.0674062073,-0.1098162457,-0.0616356097,0.0976044163,0.1996342689,0.2387029082,-0.2371465564,-0.1989568025,0.1297176927,-0.0235046968,0.0643830374,-0.1379919797,-0.1424683779,-0.1139198467,-0.1574848145,0.0627483577,0.0139852948,-0.2245469391,0.0824401602,0.2212925553,-0.3144432306,0.0370475762,0.1807736903,-0.0068506892,0.2829914391,-0.1149414852,0.2080898136,0.0898526534,0.1464464664,0.2414969057,0.4276295006,-0.1189181656,0.1101896986,0.1011361405,0.1791167557,0.2804281116,0.4780022502,-0.0934867859,-0.0483980328,0.0472078659,0.1978943199,-0.7037300467,0.2643836141,0.0070218155,-0.1743259728,-0.4329405129,-0.3057400584,0.5853257775,0.277338326,-0.17805475,0.4260760546,0.283651799,-0.2500902116,0.0859111398,0.2275706232,1.0039902925,-0.0367950499,0.293402791,-0.0351946168,0.1630490571,0.3141222596,-0.1350865215,0.1461464763,-0.2149699926,-0.2642908692,0.1335115433,-0.1900534183,0.078113623,-0.1816644371,0.0953761488,0.0175270867,0.2954483628,-0.0398650095,-0.2394230217,0.0571388453,-0.3389733732,-0.3028466702,0.1191907451,0.2474336028,-0.1621086746,0.0102834841,-0.0696115047,0.2636107206,-0.1066704541,0.0836558938,-0.0717509165,-0.3402318656,-0.2764901221,0.2149440795,-0.0150994323,-0.0378209203,0.1039626375,0.1525458246,0.148348406,0.4872969985,-0.2737690508,0.1031404063,-0.0238853339,0.3031525314,0.2847664058,0.0286053717,0.251368016,0.0815332606,-0.0211816821,-0.114235796,-0.1679257154,0.0728838071,-0.2797622979,-0.2430844605,-0.0994175598,-0.2236745507,-0.2379263937,0.2160255909,0.1669181436,-0.0928575918,0.2482772619,0.1782383919,0.1451325268,0.2919334471,-0.0215121936,-0.3788104057,-0.2249490768,0.5255734324,0.2917609215,-0.0606689714,0.157504335,-0.0234843101,-0.3464677036,-0.1200752109,0.3370844424,0.1170398369,-0.0526929013,0.2844401598,-0.2274679542,-0.1273143142,0.0570580848,0.1740360409,0.1279938519,-0.1877499819,-0.060110528,0.0337942019,-0.5955784917,0.3795018494,-0.2795139253,0.5569556355,-0.0354218334,-0.1593278497,-0.1856450737,-0.1786653548,-0.4331299067,0.0135485316,-0.0888045579,0.0816254318,-0.1195707321,0.373918891,-0.2518975437,-0.213607654,0.0406580903,0.1195770651,-0.3159406781,-0.2957638204,0.0396905728,0.1488591284,-0.0501258485,-0.3152567446,0.2680826783,-0.344876498,0.0115149524,-0.2461018562,0.328492105,0.377697736,-0.0479554348,0.162390545,0.1325922459,-0.1814634353,0.156818822,0.0919766054,0.1599471122,0.0133817149,-0.108908616,0.3041533232,-0.2605000734,-0.1088178828,-0.0856744051,0.1230723485,0.026668271,0.2430939674,0.1445355713,-0.2569819987,0.0195058286,0.2928348184,0.3991469443,0.0888295844,-0.3240458369,0.0524785072,0.2400458753,0.3007773161,0.0025267401,-0.0687353015,-0.03550791,-0.0842300281,0.2511711121,0.2521036565,0.3044990003,0.1978364885,-0.0322840251,0.0327101909,0.3509848416,-0.5152756572,0.0322654508,-0.128785193,-0.0471817106,-0.151030004,0.4487401545,0.2937787473,0.3369753957,0.5905492902,0.4089589715,0.5424235463,0.1343595237,0.2355729192,-0.319732517,-0.3584733009,0.3128856421,0.3870537579,-0.3433884382,0.0200859476,0.1504815072,-0.08860939,-0.0667210519,-0.1508194655,-0.0251812357,-0.2536854744,-0.3205809891,-0.2058205009,-0.0523490719,-0.1922961324,-0.0422640815,0.2913803756,-0.3203575015,-0.1058796495,0.4989079535,0.0421398245,-0.0748407617,-0.1559641361,0.2634370923,-0.0298068207,0.3915459216,-0.3091391623,0.0443177298,-0.3322966695,-0.107100144,-0.0753349289,0.6554329395,0.2665661275,0.1978498697,0.109224841,-0.0028951496,0.0696383491,0.0344365239,-0.2155053765,0.1760673076,0.277551353,0.1791046858,0.1556771994,0.2397782803,-0.1616642326,0.0791222677,-0.3231351078,-0.0564176776,-0.4519142807,-0.2095728666,0.1546356082,-0.210694164,0.2570950389,-0.1091148704,-0.2717449069,0.2141822278,0.2234256566,-0.0960335061,0.1687483639,-0.2350176424,0.1527178437,0.0098314555,0.3907490075,0.1923637837,-0.1102028117,-0.167197004,-0.2357330024,-0.4914355874,-0.06458278,-0.0635159165,0.280166924,0.1386426091,0.189457342,0.1401303709,0.0001410271,-0.1785236746,-0.0534383096,-0.0571589842,-0.0852273777,-0.0540345646,0.0739318505,-0.0443832576,-0.0673189387,0.0559047237,-0.3471015096,0.2162062377,-0.0085152937,0.1012286693,-0.2356850654,0.2030036896,0.2313687205,0.1989689916,0.3967600465,0.2721073031,-0.0841688514,0.0597079508,-0.1216958687,-0.3478772342,-0.2663733661,-0.158235535,-0.1882212758,-0.2899598181,0.3253398538,-0.2435728014,-0.3294013143,-0.0607005358,-0.0369214565,-0.0631317124,0.0891665891,-0.4405935109,-0.1733836979,-0.0191600379,0.0069559719,0.0356071666,0.2278299481,0.1290514171,-0.0595633313,-0.3646673262,-0.2365457714,0.2639270425,0.1838391125,-0.0536169894,-0.3645186126,0.236553967,0.3013576865,-0.009702595,-0.7267893553,-0.0019851297,0.288590461,-0.2552908361,-0.3379472792,0.3077589571,0.0231635161,-0.2093590349,-0.1936673969,0.3601523638,-0.0349941999,-0.1108690202,-0.0253084265,-0.1453530639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2452","title":"MRPC test set differences between torch and tensorflow datasets","comments":"Realized that `tensorflow_datasets` is not provided by Huggingface and should therefore raise the issue there.","body":"## Describe the bug\r\nWhen using `load_dataset(\"glue\", \"mrpc\")` to load the MRPC dataset, the test set includes the labels. When using `tensorflow_datasets.load('glue\/{}'.format('mrpc'))` to load the dataset the test set does not contain the labels. There should be consistency between torch and tensorflow ways of importing the GLUE datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\nMinimal working code \r\n```python\r\nfrom datasets import load_dataset\r\nimport tensorflow as tf\r\nimport tensorflow_datasets\r\n\r\n# torch\r\ndataset = load_dataset(\"glue\", \"mrpc\")\r\n# tf\r\ndata = tensorflow_datasets.load('glue\/{}'.format('mrpc'))\r\ndata = list(data['test'].as_numpy_iterator())\r\nfor i in range(40,50):\r\n tf_sentence1 = data[i]['sentence1'].decode(\"utf-8\") \r\n tf_sentence2 = data[i]['sentence2'].decode(\"utf-8\") \r\n\r\n tf_label = data[i]['label']\r\n \r\n index = data[i]['idx']\r\n print('Index {}'.format(index))\r\n torch_sentence1 = dataset['test']['sentence1'][index]\r\n torch_sentence2 = dataset['test']['sentence2'][index]\r\n\r\n torch_label = dataset['test']['label'][index]\r\n print('Tensorflow: \\n\\tSentence1 {}\\n\\tSentence2 {}\\n\\tLabel {}'.format(tf_sentence1, tf_sentence2, tf_label))\r\n print('Torch: \\n\\tSentence1 {}\\n\\tSentence2 {}\\n\\tLabel {}'.format(torch_sentence1, torch_sentence2, torch_label))\r\n```\r\n\r\nSample output \r\n```\r\nIndex 954\r\nTensorflow: \r\n\tSentence1 Sabri Yakou , an Iraqi native who is a legal U.S. resident , appeared before a federal magistrate yesterday on charges of violating U.S. arms-control laws .\r\n\tSentence2 The elder Yakou , an Iraqi native who is a legal U.S. resident , appeared before a federal magistrate Wednesday on charges of violating U.S. arms control laws .\r\n\tLabel -1\r\nTorch: \r\n\tSentence1 Sabri Yakou , an Iraqi native who is a legal U.S. resident , appeared before a federal magistrate yesterday on charges of violating U.S. arms-control laws .\r\n\tSentence2 The elder Yakou , an Iraqi native who is a legal U.S. resident , appeared before a federal magistrate Wednesday on charges of violating U.S. arms control laws .\r\n\tLabel 1\r\nIndex 711\r\nTensorflow: \r\n\tSentence1 Others keep records sealed for as little as five years or as much as 30 .\r\n\tSentence2 Some states make them available immediately ; others keep them sealed for as much as 30 years .\r\n\tLabel -1\r\nTorch: \r\n\tSentence1 Others keep records sealed for as little as five years or as much as 30 .\r\n\tSentence2 Some states make them available immediately ; others keep them sealed for as much as 30 years .\r\n\tLabel 0\r\n```\r\n\r\n## Expected results\r\nI would expect the datasets to be independent of whether I am working with torch or tensorflow.\r\n\r\n## Actual results\r\nTest set labels are provided in the `datasets.load_datasets()` for MRPC. However MRPC is the only task where the test set labels are not -1.\r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.109+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":15,"text":"MRPC test set differences between torch and tensorflow datasets\n## Describe the bug\r\nWhen using `load_dataset(\"glue\", \"mrpc\")` to load the MRPC dataset, the test set includes the labels. When using `tensorflow_datasets.load('glue\/{}'.format('mrpc'))` to load the dataset the test set does not contain the labels. There should be consistency between torch and tensorflow ways of importing the GLUE datasets.\r\n\r\n## Steps to reproduce the bug\r\n\r\nMinimal working code \r\n```python\r\nfrom datasets import load_dataset\r\nimport tensorflow as tf\r\nimport tensorflow_datasets\r\n\r\n# torch\r\ndataset = load_dataset(\"glue\", \"mrpc\")\r\n# tf\r\ndata = tensorflow_datasets.load('glue\/{}'.format('mrpc'))\r\ndata = list(data['test'].as_numpy_iterator())\r\nfor i in range(40,50):\r\n tf_sentence1 = data[i]['sentence1'].decode(\"utf-8\") \r\n tf_sentence2 = data[i]['sentence2'].decode(\"utf-8\") \r\n\r\n tf_label = data[i]['label']\r\n \r\n index = data[i]['idx']\r\n print('Index {}'.format(index))\r\n torch_sentence1 = dataset['test']['sentence1'][index]\r\n torch_sentence2 = dataset['test']['sentence2'][index]\r\n\r\n torch_label = dataset['test']['label'][index]\r\n print('Tensorflow: \\n\\tSentence1 {}\\n\\tSentence2 {}\\n\\tLabel {}'.format(tf_sentence1, tf_sentence2, tf_label))\r\n print('Torch: \\n\\tSentence1 {}\\n\\tSentence2 {}\\n\\tLabel {}'.format(torch_sentence1, torch_sentence2, torch_label))\r\n```\r\n\r\nSample output \r\n```\r\nIndex 954\r\nTensorflow: \r\n\tSentence1 Sabri Yakou , an Iraqi native who is a legal U.S. resident , appeared before a federal magistrate yesterday on charges of violating U.S. arms-control laws .\r\n\tSentence2 The elder Yakou , an Iraqi native who is a legal U.S. resident , appeared before a federal magistrate Wednesday on charges of violating U.S. arms control laws .\r\n\tLabel -1\r\nTorch: \r\n\tSentence1 Sabri Yakou , an Iraqi native who is a legal U.S. resident , appeared before a federal magistrate yesterday on charges of violating U.S. arms-control laws .\r\n\tSentence2 The elder Yakou , an Iraqi native who is a legal U.S. resident , appeared before a federal magistrate Wednesday on charges of violating U.S. arms control laws .\r\n\tLabel 1\r\nIndex 711\r\nTensorflow: \r\n\tSentence1 Others keep records sealed for as little as five years or as much as 30 .\r\n\tSentence2 Some states make them available immediately ; others keep them sealed for as much as 30 years .\r\n\tLabel -1\r\nTorch: \r\n\tSentence1 Others keep records sealed for as little as five years or as much as 30 .\r\n\tSentence2 Some states make them available immediately ; others keep them sealed for as much as 30 years .\r\n\tLabel 0\r\n```\r\n\r\n## Expected results\r\nI would expect the datasets to be independent of whether I am working with torch or tensorflow.\r\n\r\n## Actual results\r\nTest set labels are provided in the `datasets.load_datasets()` for MRPC. However MRPC is the only task where the test set labels are not -1.\r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.109+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\nRealized that `tensorflow_datasets` is not provided by Huggingface and should therefore raise the issue there.","embeddings":[0.0261485744,-0.4852635562,-0.0406564698,0.4093670249,0.1271916032,0.09912882,0.5130628943,0.230747208,0.0737166852,0.3825484514,-0.1057029217,0.4088272154,-0.0558496229,-0.4865846038,0.1695501059,-0.3017296493,0.1464218497,-0.0920704827,-0.1237552986,-0.2203543633,-0.0687525049,0.1828691363,-0.2936635017,-0.1844654828,-0.156720683,-0.0232121777,-0.0786587745,0.0440008044,0.1698827147,-0.4878647625,0.1352048665,0.0039574974,-0.0758825094,0.6296335459,-0.0001189356,0.1458906531,0.4568511248,-0.0476903319,0.0133067202,-0.3868901432,-0.0054133064,0.1641970128,0.2406085879,-0.0807504505,-0.1679168344,0.2380318791,-0.1524688303,-0.2349459976,0.1612937152,0.4040102959,0.133264631,0.6312502623,-0.0760309771,-0.1164818257,0.4525410831,-0.1993582696,-0.0935989246,-0.0306460615,0.2879213393,0.1278103888,-0.0768549368,0.3218511641,-0.1801728904,0.1891780943,0.1300185323,0.0046263514,-0.0839199722,-0.4597879946,-0.2064427882,0.4059527516,0.3886668682,-0.2746468782,-0.1795115918,-0.3170023561,-0.035715092,-0.2481783479,0.1790881157,0.3536750078,-0.2183131725,0.1314068139,-0.2730304301,0.2830864787,-0.0537571311,0.3323119879,-0.1957848519,0.225935325,0.0312797911,0.0999972448,0.1313652098,0.0053545805,0.3146856427,0.0339371413,0.0620538667,0.008161528,-0.261228621,0.0196104478,0.0636192188,-0.0553953089,0.1330862343,-0.2808221877,0.2946055233,0.1996212602,0.0916738883,0.1940813512,-0.2211866975,0.3942585588,0.1337268949,0.2836015522,-0.0379245728,-0.3140030503,-0.2698280811,0.1454364359,-0.0445575118,0.0287070833,0.0584632494,0.1344109327,0.0514179543,-0.0484617278,-0.4138657749,0.0773653984,-0.3263053,-0.1696856022,-0.2202602476,0.5934234858,-0.2372552902,0.2949207723,-0.1201264486,0.2392455637,-0.1729836315,-0.189748466,-0.1912541986,0.1766444296,-0.1897297353,-0.116504766,0.2497688532,-0.2277820408,0.3021504283,-0.0364691541,-0.0965437442,-0.0238527562,0.1509497166,-0.0550075993,0.7358696461,0.0225617774,0.1828421205,0.140667513,0.0513226576,-0.0169458594,-0.0562738217,0.0521189235,-0.2948814929,0.0802291334,0.1529688388,0.12040589,-0.1925696582,-0.1619033813,0.0754925236,0.1264446974,-0.0435537845,0.1901890486,0.26821661,-0.0774894059,-0.1642974764,-0.2255094349,-0.1302791387,0.344622165,-0.2681716979,-0.3065844774,0.293628782,0.2347957194,-0.0268686786,0.1929280013,0.0135615021,0.2826031744,-0.1652211398,-0.232605651,0.2195355892,-0.8616648316,-0.2695984542,0.1949015409,0.0065686381,0.2051821649,0.2177992612,-0.0519418418,-0.0724701956,0.0933177471,0.2777657211,0.0914550275,0.0103686759,0.1112370193,-0.2181405723,0.0495595895,0.2911790311,0.0014045913,0.1172998026,-0.0416461974,0.0752110705,-0.2959313393,0.3078765273,0.0031297922,-0.2602918148,0.4254790246,-0.2870187163,-0.1407147199,-0.047386568,-0.1010833234,-0.2631437778,0.2191304266,-0.1039094031,0.3021523356,0.112700671,-0.0315030068,-0.1614627689,-0.0917108878,-0.220363304,-0.3311410546,0.057352595,0.2291719764,0.1316349804,-0.0263840966,0.0296224058,0.3999807835,0.2018100619,0.2311398387,-0.0989278182,0.2352641821,0.1024239287,-0.0957424119,-0.0565804206,0.3975553811,0.1845360994,-0.2048081011,-0.0754873082,0.4399977624,-0.0254469197,0.1216707006,-0.4201394618,-0.1171443164,0.4344422519,-0.4096219838,-0.0988099501,0.2031838,0.1255825758,-0.0318556204,-0.1936947107,0.222259745,0.2248143852,0.2307391614,0.1110636145,-0.1285502762,-0.1522250324,-0.0802835599,-0.1840081513,-0.3531493545,0.5630275011,0.0976499096,0.3836700916,0.2488897145,-0.4424460232,0.0375956036,0.4983786047,0.026095869,0.2035005242,-0.1867844909,-0.4458573461,0.1614033431,0.105065152,-0.1024869457,0.5798225403,0.2963430583,-0.1030573249,0.0409140103,-0.1228296235,-0.100265868,0.112481609,0.2082177699,0.0188880917,0.000153512,0.3692679703,0.1109226868,-0.1828511804,0.1988733709,-0.0253415182,-0.1920788884,-0.5446025133,0.0110713346,-0.4301724732,-0.4148855507,-0.2450619638,-0.2872181535,-0.2379797101,-0.3870528936,0.0009965107,-0.0786514878,0.0367396548,0.2030114233,0.3443807065,-0.1007799506,0.0340608917,-0.2861700058,0.0872789994,-0.1952978671,-0.2093060017,-0.069784984,0.07713487,0.0857914761,0.3604156673,-0.3037988245,-0.0860263407,-0.1592538208,-0.3240958154,0.5268493295,-0.2385825515,0.1935058683,0.2617387176,-0.3003004193,-0.0033720059,0.0280535705,0.031810388,-0.2877400219,-0.0593402945,0.0935971811,0.0518820025,-0.2390560359,0.0306804851,-0.6296645403,-0.2213192135,-0.2936131656,0.1027421281,-0.0329730399,0.0066429102,0.1782484204,-0.136363253,-0.0344947837,-0.1142251268,0.2638046145,-0.1993944496,-0.4606876373,0.1238719076,-0.4417025447,-0.5407736301,-0.2513517141,-0.2575095892,0.0077014421,0.1778318882,-0.1692204028,-0.2368616909,-0.2342531234,0.1535665542,0.0514444076,0.3187043965,0.1556939483,-0.1288915575,0.1132317558,-0.4747688472,-0.120506987,0.153414607,0.0590810217,0.1915936768,0.0039260616,0.1071707755,0.061250966,0.4445212781,-0.1030288935,-0.5058110952,0.3583200276,-0.2459377199,0.4808038175,-0.0152407968,-0.267504096,0.0908953995,0.1069890559,0.2713555098,0.2215180695,-0.1758103967,-0.2347477525,-0.2094546705,0.3248157501,-0.447827816,-0.1090401039,-0.0456512049,-0.1603236794,0.0181960575,0.1657184958,0.3096438944,-0.2159806043,-0.3260965049,0.1154042706,-0.0363689288,0.1256204098,0.0888835192,-0.4212674797,0.0416310728,-0.5271978974,0.2519057393,0.3857143819,0.3348063827,-0.0273557827,-0.1982361525,-0.1415959448,0.1894200146,0.5182942748,-0.3986012638,0.0667465478,0.1938921213,0.1717714965,-0.7626214027,-0.1268062592,-0.4225605428,-0.0555558242,0.4223057032,0.1446826011,-0.0863947272,-0.2914789021,-0.0721174851,0.0491872765,-0.0423614345,-0.2509915829,-0.1612650007,0.1055295244,-0.0776395649,-0.0196556337,0.2048522234,0.1878491342,-0.1682690829,-0.1457278877,-0.1287943125,-0.0590154119,0.2605468631,0.1739228666,0.2881631255,0.0615336224,0.1905277967,0.1036713198,0.5682646036,0.0522206314,0.2521030605,-0.4342468083,0.0578103699,0.0766292885,-0.2648721933,0.3054163158,0.2998768985,-0.0126541974,-0.0960637033,-0.6493073702,0.0786402524,-0.2375623435,0.3297789395,0.0691422224,0.2048294246,-0.4163253307,-0.0632149726,0.2043107003,0.1374180913,-0.1201537922,0.2713946104,-0.1305900365,-0.4574914575,0.5324043036,-0.030350225,0.7031251788,-0.057831239,0.2459033877,0.1927008927,0.0246449169,0.5451081991,-0.2435890883,0.1984702796,-0.2252380848,-0.5061349273,-0.0947213098,-0.2841522396,-0.1502871662,0.2378441393,-0.3137894571,0.6216282248,0.0667094216,0.1143938154,0.1837727726,0.2670100331,0.0929721892,-0.344144702,-0.1160385758,0.0057153441,0.2483813018,0.4108079374,-0.1654258072,0.05158953,-0.154468134,-0.1849396378,-0.2239165157,0.0544463545,-0.8971481919,-0.0873610079,0.3152316809,-0.1536289454,-0.0696995407,0.4063764513,0.3651089668,0.0183506273,0.044949092,0.2344439626,0.4681715667,0.1398103833,0.3395697176,-0.066802904,0.6410596967,-0.010494546,-0.0350675806,0.1543358862,-0.1718003154,-0.0279965606,-0.2236998975,0.0400440134,0.2533499897,-0.1420948058,-0.2080612779,0.0559361316,-0.0560282357,-0.0599899255,0.0677218363,-0.0392806903,-0.143091768,0.1871360093,0.0111882342,-0.5029118061,-0.1013019234,0.4887892604,-0.0856809765,-0.3984007239,0.0194320418,0.2082631141,0.0084084645,-0.2477831244,-0.0270492099,0.3757218122,-0.3810605407,-0.0469850898,-0.2838191986,-0.1649595499,0.0074268612,0.2035780102,0.0732887015,0.1287187487,-0.1060063243,-0.1675921828,-0.1157035157,-0.0180189591,-0.0025482343,0.0329141803,0.1127638817,0.1971694231,0.1927319318,0.0124041028,-0.2522570193,0.0422297344,0.0182939246,0.0517058559,0.369271636,0.028529752,0.291448921,-0.1684616953,0.1055060998,0.216408357,-0.0323024243,-0.1908490062,-0.2669506669,0.1268221587,-0.1265551001,0.0483267158,0.0186192952,0.0130958892,0.1995587051,-0.2796386778,0.3475923836,0.2750243545,-0.0618946739,0.3123222888,0.1005775034,0.2225857377,-0.0394548364,0.0839852169,0.1022237986,0.1146096885,0.0186111256,0.1657169759,-0.0150674526,0.0566080064,0.0459227636,-0.0709856302,-0.0331197865,0.0811861381,-0.011181375,-0.3356150091,-0.0382746011,0.1470419914,0.0520125404,0.0601787232,-0.3827112913,0.0538503751,-0.0274264328,0.2136344016,-0.0287975222,-0.2715962231,-0.0212846473,0.2179252356,0.2058618963,0.285921067,0.3748160005,0.1710000783,-0.1631913036,0.2002656758,0.3545741439,-0.1063456908,0.2357365191,0.2015610337,0.0780016109,0.003702047,0.0139350593,-0.2552009225,-0.0480212644,-0.0936487541,0.432280153,0.609767437,0.2216233909,-0.052927617,0.3795582056,0.0532928631,0.038123589,0.0704010874,-0.0008060421,0.3937423825,-0.0849768519,0.7027029395,0.1229556501,-0.4091215432,-0.0550967641,0.0764363483,-0.0342284292,-0.2700658441,-0.0880366191,-0.0298293326,-0.2634902298,-0.0229858048,0.0215562582,-0.2900961637,0.1546148658,0.2581143081,-0.0864197388,-0.2978470325,-0.3602617681,0.0533818752,0.1972672343,-0.2041769028,0.3390128613,-0.121220015,-0.2024329156,0.2504366636,0.0120416731,0.4645029902,0.1406089664,-0.3251254857,-0.1031660289,-0.0637895018,-0.0436435044,-0.071339637,0.0467670262,-0.0821751505,-0.0418519415,0.2714115083,0.13606067,-0.1286405325,-0.0654725954,0.0413731895,0.207407102,-0.2755357325,0.107650131,-0.2716192007,0.0169959851,-0.1972705275,0.1689263135,-0.572490871,0.3044317067,0.585653007,-0.0141738085,0.2190096527,-0.0823730975,0.0733112767,-0.1027477905,0.3238554895,0.1918896735,0.2639640868,-0.2637925148,-0.1038483977,-0.3774582744,0.1888024211,0.1409155875,-0.2214824408,0.1619472355,0.1524030268,0.0237263329,0.1909632683,0.0915486291,0.2092489451,0.0247695539,0.1936853379,-0.5353742242,-0.0970440134,-0.1275385022,0.0535546988,0.180910632,-0.3217891157,-0.2524047792,0.0123878894,0.0058039948,-0.356987685,0.0729943663,0.1320171952,0.2190764099,0.3762047291,0.0051442636,0.7146678567,-0.0234269723,-0.032939069,-0.1327984929,-0.3116542697,-0.1099420339,0.1135383844,0.0040358575,0.2622762024,-0.0556943156,-0.1913021207,-0.1204274297,0.2540628612,0.3894256949,0.0174141228,-0.2802178264,0.1768088192,-0.0515043028,0.19842498,0.2788590193,0.4941806793,-0.0558047555,0.2504893839,0.0057154866,-0.6125565767,0.2101188153,-0.6313806176,-0.0745611712,-0.1029293239,0.171128124,0.1883383244,-0.0063946317,-0.9091857076,-0.1575627625,0.3860874772,-0.2401844114,-0.3059315383,0.05899176,0.140670687,0.1829777062,-0.0591248125,0.0654752254,0.336247474,-0.2855674326,-0.0498244874,-0.1186591089]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2450","title":"BLUE file not found","comments":"Hi ! The `blue` metric doesn't exist, but the `bleu` metric does.\r\nYou can get the full list of metrics [here](https:\/\/github.com\/huggingface\/datasets\/tree\/master\/metrics) or by running\r\n```python\r\nfrom datasets import list_metrics\r\n\r\nprint(list_metrics())\r\n```","body":"Hi, I'm having the following issue when I try to load the `blue` metric.\r\n\r\n```shell\r\nimport datasets\r\nmetric = datasets.load_metric('blue')\r\nTraceback (most recent call last):\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/metrics\/blue\/blue.py\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 332, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/metrics\/blue\/blue.py\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 605, in load_metric\r\n dataset=False,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 343, in prepare_module\r\n combined_path, github_file_path\r\nFileNotFoundError: Couldn't find file locally at blue\/blue.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/metrics\/blue\/blue.py.\r\nThe file is also not present on the master branch on github.\r\n```\r\nHere is dataset installed version info\r\n```shell\r\npip freeze | grep datasets\r\ndatasets==1.7.0\r\n```\r\n","comment_length":31,"text":"BLUE file not found\nHi, I'm having the following issue when I try to load the `blue` metric.\r\n\r\n```shell\r\nimport datasets\r\nmetric = datasets.load_metric('blue')\r\nTraceback (most recent call last):\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/metrics\/blue\/blue.py\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 332, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/metrics\/blue\/blue.py\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 605, in load_metric\r\n dataset=False,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 343, in prepare_module\r\n combined_path, github_file_path\r\nFileNotFoundError: Couldn't find file locally at blue\/blue.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/metrics\/blue\/blue.py.\r\nThe file is also not present on the master branch on github.\r\n```\r\nHere is dataset installed version info\r\n```shell\r\npip freeze | grep datasets\r\ndatasets==1.7.0\r\n```\r\n\nHi ! The `blue` metric doesn't exist, but the `bleu` metric does.\r\nYou can get the full list of metrics [here](https:\/\/github.com\/huggingface\/datasets\/tree\/master\/metrics) or by running\r\n```python\r\nfrom datasets import list_metrics\r\n\r\nprint(list_metrics())\r\n```","embeddings":[-0.2038624436,-0.286285907,-0.0685010329,0.4027132094,0.3037265539,0.0655114725,0.2102440298,0.3386648595,0.0885978788,0.1891393214,-0.2280900776,-0.1502068639,0.0800003409,-0.3868101239,0.2086268514,-0.0838348791,-0.0976857319,0.2933502793,0.1404932737,-0.0119333351,-0.2906016111,0.279425174,-0.0793287903,0.1401927322,-0.2483473867,0.2007489204,0.0558202937,0.4576255977,-0.066158548,-0.4756872058,0.262820214,0.0613073483,0.0844941288,0.6237055659,-0.0001181479,0.1353184283,0.4002448022,-0.026030736,-0.1755044907,-0.7946474552,-0.2668886185,-0.3052861094,-0.1206793189,-0.1363381892,-0.0330706872,-0.0962109268,0.0631731302,-0.1688701361,0.0880968049,0.2695310712,0.1627939492,0.3915831447,0.0068830149,-0.0806269795,0.1766516715,-0.1143770814,0.1108325496,0.5380025506,0.1324937791,0.0339101963,0.1965762675,0.2750691473,0.0797883868,0.098599039,0.3210598528,-0.0133122467,0.5190857649,0.093614094,-0.0156915635,0.2679638863,0.2782544196,-0.0065161688,-0.5074590445,0.0587431863,-0.2388174832,-0.4378614128,0.3182871044,-0.0005305419,0.0924746692,0.040464472,-0.3045379519,-0.3758830428,-0.2266742438,0.008756551,-0.3165692091,0.271229744,-0.1296576858,-0.0393383913,0.131154418,0.035779655,-0.1490480006,-0.0241202097,-0.2086912543,0.1841868162,-0.2370862961,0.1590325236,-0.0566026047,-0.1240671426,0.1691560149,0.1326967925,0.2836335599,-0.0671531931,-0.1154061556,0.2826260924,0.1740510762,0.2684218884,0.0162829645,0.1422417462,0.3623938262,0.4090431035,-0.2026886195,-0.2172914594,-0.0469186157,-0.5095567107,0.2717924714,0.1654091477,0.2333503067,-0.3669013679,-0.546892941,-0.1752153039,0.2014846504,-0.138439551,0.2428230643,0.3890919685,-0.1542295218,0.1355956793,-0.0241192281,0.2843506038,-0.3726007044,-0.0479731485,-0.2461787909,0.270036608,-0.2791169882,0.1852990091,0.134738192,-0.1474324316,0.4780178666,-0.2440681159,0.1274051219,-0.2823776603,0.1358564347,-0.1307548732,-0.0632690564,0.3866440654,-0.0663409978,0.0235976931,0.0756518319,-0.1533665955,-0.1150935367,0.1922052354,-0.3756907582,-0.3689212203,0.0063163307,0.1795061976,-0.6024382114,-0.013512793,-0.1706515402,0.0045132828,0.0897604227,-0.5304581523,-0.1209513769,-0.0536715128,-0.3520642519,-0.1883587986,0.2899969816,0.5583953261,-0.2963094413,-0.3161924183,-0.2241495848,-0.0718351007,0.0118514877,0.1162731722,-0.1957620084,0.2485929281,-0.5058684945,-0.2216970921,0.6039164662,-0.9293052554,-0.3994351625,0.0354591422,-0.314542383,0.1001128778,0.1549662948,-0.0467824712,0.115826495,-0.2321648002,0.302493304,0.0525196567,0.074928388,-0.1573142558,-0.2148272693,-0.1992250383,0.1408916861,0.1859884113,0.3698249459,-0.0691337138,0.2987953722,-0.1159083545,-0.0255059134,0.0242809486,-0.0979231074,0.3807157874,0.0568256937,0.0048563476,0.0850338936,-0.4343167543,0.0962968469,0.2571773529,-0.1961857229,-0.2399356514,-0.0186847337,0.1108483896,-0.4768013954,-0.3033174574,-0.3182031512,-0.0944714919,0.0308126733,0.0620128363,0.0148465876,0.0749977529,-0.17083399,0.3170694709,-0.0074223857,0.0792751908,-0.0620083287,0.2012156248,0.0121677862,-0.0265175514,-0.0333781242,0.2669577599,0.3650774956,-0.1662282646,-0.1953881383,0.4600685835,0.0996977836,0.3213559985,0.2970384955,0.0997050256,0.3341647089,-0.215291068,0.0784406811,-0.0297695249,0.1145220175,-0.0374824256,0.0993940607,0.3468848467,0.0831461996,0.1389874369,0.1891477257,-0.2690497637,0.2665727139,-0.1116631404,0.0468978509,-0.120234102,0.4082636237,-0.0945784077,0.0844262168,-0.0353472419,0.0425563082,-0.1108863801,0.2558621466,-0.1055775434,0.0156317856,0.0747177899,-0.1970437914,-0.0989656746,-0.2834335864,0.2893654704,0.7424980998,0.2036165893,0.0498852096,-0.0030591253,0.0600026026,-0.135204643,0.1827728152,0.1364149898,0.0558149219,0.4532756209,0.1284675151,-0.1120860875,-0.1604457796,-0.2615599632,0.0380720012,-0.0798378214,-0.3087829351,0.2265219539,-0.1820993274,0.181409508,-0.0977322757,-0.0975359753,-0.0847241133,-0.2716193497,0.0139922788,0.1867242455,0.301818192,0.0267870892,-0.1049008816,0.1172481105,-0.0092865555,-0.2614940405,-0.0797163844,-0.226645276,-0.138090089,-0.0651665181,-0.0165860113,-0.0185929723,0.3383127153,-0.4342367053,0.2017265111,-0.2717138529,-0.3112030327,-0.0327798836,0.2021766156,0.5519148111,0.0779538527,0.0549889728,0.0839210004,0.2093093991,0.4240537882,-0.247879833,-0.1166461185,0.1270682067,-0.113442339,0.0935031623,-0.0398880206,-0.4275738299,-0.3291085064,-0.405400306,-0.1222823784,-0.1058298498,0.030580502,0.5178664327,0.0676998943,0.1980971694,0.1595279574,0.151336953,-0.1719416231,-0.5531238914,0.4830413461,-0.3444617689,-0.2105953395,0.3475134671,0.109351635,0.4056742191,0.047029946,-0.3599723279,-0.3414432108,0.1025691852,0.0229943469,-0.1240343228,0.1032207161,0.068411313,-0.1385262012,-0.0798841193,-0.168666631,-0.1995458752,0.0632746145,0.0649987981,0.6346903443,-0.1762986779,0.1323319823,-0.1719237566,0.6171426177,0.0703139454,0.0202168189,0.326261282,0.194279328,0.4952047467,-0.0125613613,-0.2130267769,0.0433266871,0.246070385,-0.0176599026,0.0808744058,0.3187314272,0.0635616258,-0.2410158068,-0.3034925163,-0.2976821661,-0.043066036,-0.2033643425,-0.171039179,0.2434120178,0.1458584815,-0.0820808262,-0.3605437577,-0.0717066005,0.1613418609,0.8195825219,0.1423374712,0.1108169258,0.0290332977,-0.2295909673,-0.015448926,0.2639464438,0.3717529774,0.2841869891,-0.2373904288,0.2865346074,0.0564559139,-0.0776723102,0.5926782489,-0.2880095243,0.1730368584,0.1834766567,-0.0934189856,-0.0887018293,-0.1145091653,0.2175034434,0.2971912622,0.117755577,0.5743358731,-0.160858348,-0.196233809,0.0543240011,0.3400988281,-0.3454618454,-0.2840363681,-0.4877050817,-0.5236839056,-0.0042205197,-0.3721770942,-0.0787026361,0.2904574275,0.0152567113,-0.1196865737,-0.0763870552,0.0144908652,0.1442987323,0.0109875519,0.4414971173,-0.3486977816,0.0901101008,0.6808419228,-0.3170379102,0.2050163895,0.3126320541,0.0865959823,-0.4804007709,0.1092877239,0.1043396369,0.1207940802,0.3142724931,-0.0896870568,-0.2236964256,0.0016913622,-0.0164651796,0.0017646016,0.1162373573,0.1674048901,0.3277085721,-0.3291717172,-0.5208202004,0.0903667361,0.2208288163,0.0332244635,0.4332658947,0.0146515677,-0.2760310173,0.0304517318,-0.1376806051,0.9293695092,-0.1882178187,0.0197137874,0.1726040542,-0.383990407,0.0584003963,-0.2677530944,0.2663508654,-0.1903845221,-0.3532457352,-0.0340050906,-0.3336504996,0.2430554032,-0.0606812984,0.0111931264,0.5956478119,-0.0813547373,0.0939347297,0.051942464,0.0904619917,-0.1423704922,0.0577240251,0.0329636186,0.1051893458,-0.2272537053,0.2990388572,-0.1936215311,-0.230888918,-0.1308116019,-0.0655563995,-0.0890985131,-0.0723471269,-0.0746775791,0.3519776165,-0.0094720293,-0.2856935263,0.0322626196,0.2786594331,0.2132272273,-0.1411055177,-0.2179094106,0.4878760278,-0.3885922134,-0.001628659,-0.0234856009,0.0653723627,0.0743611306,-0.0114005879,-0.421651125,0.1994488835,-0.1593191028,-0.18825683,0.0412396677,0.0750640035,-0.1103762239,0.0909299999,0.014743384,-0.1974089146,-0.0220826846,-0.1685828865,0.1055674553,0.1960796565,-0.2721881568,-0.1259486526,0.1073704734,-0.3957341611,0.0294664074,0.4185599387,-0.0717896819,-0.2697695494,0.2888673842,-0.252653718,-0.186411351,-0.1533373743,0.0639422089,-0.4145129621,-0.155424282,0.2565377951,-0.0062887417,0.2278640419,0.1213084534,0.2209199369,0.2727257907,-0.4424714148,-0.0002790965,-0.6986165047,-0.1593088955,0.1652499437,-0.3136617839,0.0674873069,-0.0574709848,-0.1149328426,-0.1241893992,-0.0054318155,-0.2733653784,0.1323696226,-0.2543610334,-0.0331063978,-0.0485642329,-0.1197383627,0.3125316501,-0.0795154199,0.0412274487,-0.112618424,-0.28047207,-0.1652464718,-0.1427247375,0.1883925647,-0.1919293702,-0.0162004046,0.0087100472,-0.0256241411,0.2526685596,0.027589485,0.0102170827,0.1508691609,0.0549973883,-0.0908617899,-0.258667022,0.0971573293,0.0889883116,0.2601917386,0.1108610407,0.5788546801,0.0635835752,-0.0576905198,-0.0915762559,0.0325405672,-0.072038807,-0.1396462619,0.104491733,0.0304917637,0.3099398017,-0.3526627123,0.1116649806,0.1382086873,0.2125661969,0.1743493378,-0.208732456,0.2819304764,-0.0355677009,0.161100477,-0.0784536451,-0.1347819716,0.2081986517,-0.0087054772,0.0300102066,0.2454165518,0.2398883998,-0.0686846823,0.1356220394,0.1805455834,0.716170311,-0.1656839252,0.1646829247,0.0211391971,0.0997987464,0.0909186453,0.5014144778,0.3404444456,0.1566219628,0.4747625887,-0.09747684,0.2164920121,-0.0298287626,0.3604225814,-0.4585135579,-0.4043613672,-0.1977346838,-0.080923833,-0.0976931304,-0.1050588936,-0.1571647674,0.4915153682,-0.4113043547,0.1484201401,0.0108117824,-0.0091366079,-0.0967927203,0.0585560836,0.1124368757,-0.2258551121,-0.1174156964,-0.0631916597,0.303794235,0.070056349,-0.0320829786,0.0684157014,-0.0301989187,-0.2110674083,0.1346748322,0.3773548901,0.098997049,-0.230566144,0.2642075717,0.347799778,0.1171856821,0.3213607073,0.2767164111,0.3790933788,0.3753465414,-0.1545854211,0.0001059834,-0.0183589514,-0.1806027442,0.041069001,0.2578658164,0.0083794445,-0.1360823214,0.0483013578,0.2008868605,-0.1168294251,0.5617827177,0.0570292212,-0.0070071458,-0.2716127336,0.4583091438,-0.333306402,0.0492208712,-0.3129720986,-0.2718108594,-0.5526043773,-0.0267052893,0.2775914967,0.1155038774,0.1228216141,0.0837406665,0.0487815887,-0.0070327721,-0.0819081143,0.4508321285,0.0906260461,-0.5584523082,0.0252812188,-0.6570866108,0.1153029948,0.1005192846,0.1148271933,0.0499007329,0.0832690373,-0.4690294862,-0.0776049122,0.511138916,0.2343607396,0.5092459917,0.0947927982,-0.3792040348,-0.2515353262,0.1353820413,-0.0111658219,-0.1206037402,-0.3279915452,-0.1026712656,-0.1878145784,0.0213371553,-0.1614062786,-0.0440401174,0.0943923444,0.1164212897,0.3828518391,-0.123639904,0.6549368501,-0.0110115558,0.0350222737,-0.1622268111,-0.2287862003,0.0851554796,-0.1250772774,-0.0343030579,0.2623528838,-0.2591590881,-0.0701133311,-0.2872568667,0.4354524612,-0.0614144355,-0.4957471192,-0.1876620501,-0.0507446192,-0.2257562727,-0.008248684,0.1245377585,0.2674946785,-0.0836365893,0.146944344,-0.1605594158,-0.0961769596,0.6528552175,-0.1847657859,-0.2792511582,0.1726897806,0.3615811765,0.3517997563,0.0711644068,-0.3607063591,0.1369380355,0.3650059104,0.2070905715,-0.20161587,0.2284931093,0.0129004307,-0.0097609796,0.1081143469,-0.3145316243,0.076554276,0.0150702503,-0.0079064807,-0.0129493475]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2450","title":"BLUE file not found","comments":"Ah, my mistake. Thanks for correcting","body":"Hi, I'm having the following issue when I try to load the `blue` metric.\r\n\r\n```shell\r\nimport datasets\r\nmetric = datasets.load_metric('blue')\r\nTraceback (most recent call last):\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/metrics\/blue\/blue.py\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 332, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/metrics\/blue\/blue.py\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 605, in load_metric\r\n dataset=False,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 343, in prepare_module\r\n combined_path, github_file_path\r\nFileNotFoundError: Couldn't find file locally at blue\/blue.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/metrics\/blue\/blue.py.\r\nThe file is also not present on the master branch on github.\r\n```\r\nHere is dataset installed version info\r\n```shell\r\npip freeze | grep datasets\r\ndatasets==1.7.0\r\n```\r\n","comment_length":6,"text":"BLUE file not found\nHi, I'm having the following issue when I try to load the `blue` metric.\r\n\r\n```shell\r\nimport datasets\r\nmetric = datasets.load_metric('blue')\r\nTraceback (most recent call last):\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 320, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/metrics\/blue\/blue.py\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 332, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/metrics\/blue\/blue.py\r\nDuring handling of the above exception, another exception occurred:\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 605, in load_metric\r\n dataset=False,\r\n File \"\/home\/irfan\/environments\/Perplexity_Transformers\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 343, in prepare_module\r\n combined_path, github_file_path\r\nFileNotFoundError: Couldn't find file locally at blue\/blue.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.7.0\/metrics\/blue\/blue.py.\r\nThe file is also not present on the master branch on github.\r\n```\r\nHere is dataset installed version info\r\n```shell\r\npip freeze | grep datasets\r\ndatasets==1.7.0\r\n```\r\n\nAh, my mistake. Thanks for correcting","embeddings":[-0.2038624436,-0.286285907,-0.0685010329,0.4027132094,0.3037265539,0.0655114725,0.2102440298,0.3386648595,0.0885978788,0.1891393214,-0.2280900776,-0.1502068639,0.0800003409,-0.3868101239,0.2086268514,-0.0838348791,-0.0976857319,0.2933502793,0.1404932737,-0.0119333351,-0.2906016111,0.279425174,-0.0793287903,0.1401927322,-0.2483473867,0.2007489204,0.0558202937,0.4576255977,-0.066158548,-0.4756872058,0.262820214,0.0613073483,0.0844941288,0.6237055659,-0.0001181479,0.1353184283,0.4002448022,-0.026030736,-0.1755044907,-0.7946474552,-0.2668886185,-0.3052861094,-0.1206793189,-0.1363381892,-0.0330706872,-0.0962109268,0.0631731302,-0.1688701361,0.0880968049,0.2695310712,0.1627939492,0.3915831447,0.0068830149,-0.0806269795,0.1766516715,-0.1143770814,0.1108325496,0.5380025506,0.1324937791,0.0339101963,0.1965762675,0.2750691473,0.0797883868,0.098599039,0.3210598528,-0.0133122467,0.5190857649,0.093614094,-0.0156915635,0.2679638863,0.2782544196,-0.0065161688,-0.5074590445,0.0587431863,-0.2388174832,-0.4378614128,0.3182871044,-0.0005305419,0.0924746692,0.040464472,-0.3045379519,-0.3758830428,-0.2266742438,0.008756551,-0.3165692091,0.271229744,-0.1296576858,-0.0393383913,0.131154418,0.035779655,-0.1490480006,-0.0241202097,-0.2086912543,0.1841868162,-0.2370862961,0.1590325236,-0.0566026047,-0.1240671426,0.1691560149,0.1326967925,0.2836335599,-0.0671531931,-0.1154061556,0.2826260924,0.1740510762,0.2684218884,0.0162829645,0.1422417462,0.3623938262,0.4090431035,-0.2026886195,-0.2172914594,-0.0469186157,-0.5095567107,0.2717924714,0.1654091477,0.2333503067,-0.3669013679,-0.546892941,-0.1752153039,0.2014846504,-0.138439551,0.2428230643,0.3890919685,-0.1542295218,0.1355956793,-0.0241192281,0.2843506038,-0.3726007044,-0.0479731485,-0.2461787909,0.270036608,-0.2791169882,0.1852990091,0.134738192,-0.1474324316,0.4780178666,-0.2440681159,0.1274051219,-0.2823776603,0.1358564347,-0.1307548732,-0.0632690564,0.3866440654,-0.0663409978,0.0235976931,0.0756518319,-0.1533665955,-0.1150935367,0.1922052354,-0.3756907582,-0.3689212203,0.0063163307,0.1795061976,-0.6024382114,-0.013512793,-0.1706515402,0.0045132828,0.0897604227,-0.5304581523,-0.1209513769,-0.0536715128,-0.3520642519,-0.1883587986,0.2899969816,0.5583953261,-0.2963094413,-0.3161924183,-0.2241495848,-0.0718351007,0.0118514877,0.1162731722,-0.1957620084,0.2485929281,-0.5058684945,-0.2216970921,0.6039164662,-0.9293052554,-0.3994351625,0.0354591422,-0.314542383,0.1001128778,0.1549662948,-0.0467824712,0.115826495,-0.2321648002,0.302493304,0.0525196567,0.074928388,-0.1573142558,-0.2148272693,-0.1992250383,0.1408916861,0.1859884113,0.3698249459,-0.0691337138,0.2987953722,-0.1159083545,-0.0255059134,0.0242809486,-0.0979231074,0.3807157874,0.0568256937,0.0048563476,0.0850338936,-0.4343167543,0.0962968469,0.2571773529,-0.1961857229,-0.2399356514,-0.0186847337,0.1108483896,-0.4768013954,-0.3033174574,-0.3182031512,-0.0944714919,0.0308126733,0.0620128363,0.0148465876,0.0749977529,-0.17083399,0.3170694709,-0.0074223857,0.0792751908,-0.0620083287,0.2012156248,0.0121677862,-0.0265175514,-0.0333781242,0.2669577599,0.3650774956,-0.1662282646,-0.1953881383,0.4600685835,0.0996977836,0.3213559985,0.2970384955,0.0997050256,0.3341647089,-0.215291068,0.0784406811,-0.0297695249,0.1145220175,-0.0374824256,0.0993940607,0.3468848467,0.0831461996,0.1389874369,0.1891477257,-0.2690497637,0.2665727139,-0.1116631404,0.0468978509,-0.120234102,0.4082636237,-0.0945784077,0.0844262168,-0.0353472419,0.0425563082,-0.1108863801,0.2558621466,-0.1055775434,0.0156317856,0.0747177899,-0.1970437914,-0.0989656746,-0.2834335864,0.2893654704,0.7424980998,0.2036165893,0.0498852096,-0.0030591253,0.0600026026,-0.135204643,0.1827728152,0.1364149898,0.0558149219,0.4532756209,0.1284675151,-0.1120860875,-0.1604457796,-0.2615599632,0.0380720012,-0.0798378214,-0.3087829351,0.2265219539,-0.1820993274,0.181409508,-0.0977322757,-0.0975359753,-0.0847241133,-0.2716193497,0.0139922788,0.1867242455,0.301818192,0.0267870892,-0.1049008816,0.1172481105,-0.0092865555,-0.2614940405,-0.0797163844,-0.226645276,-0.138090089,-0.0651665181,-0.0165860113,-0.0185929723,0.3383127153,-0.4342367053,0.2017265111,-0.2717138529,-0.3112030327,-0.0327798836,0.2021766156,0.5519148111,0.0779538527,0.0549889728,0.0839210004,0.2093093991,0.4240537882,-0.247879833,-0.1166461185,0.1270682067,-0.113442339,0.0935031623,-0.0398880206,-0.4275738299,-0.3291085064,-0.405400306,-0.1222823784,-0.1058298498,0.030580502,0.5178664327,0.0676998943,0.1980971694,0.1595279574,0.151336953,-0.1719416231,-0.5531238914,0.4830413461,-0.3444617689,-0.2105953395,0.3475134671,0.109351635,0.4056742191,0.047029946,-0.3599723279,-0.3414432108,0.1025691852,0.0229943469,-0.1240343228,0.1032207161,0.068411313,-0.1385262012,-0.0798841193,-0.168666631,-0.1995458752,0.0632746145,0.0649987981,0.6346903443,-0.1762986779,0.1323319823,-0.1719237566,0.6171426177,0.0703139454,0.0202168189,0.326261282,0.194279328,0.4952047467,-0.0125613613,-0.2130267769,0.0433266871,0.246070385,-0.0176599026,0.0808744058,0.3187314272,0.0635616258,-0.2410158068,-0.3034925163,-0.2976821661,-0.043066036,-0.2033643425,-0.171039179,0.2434120178,0.1458584815,-0.0820808262,-0.3605437577,-0.0717066005,0.1613418609,0.8195825219,0.1423374712,0.1108169258,0.0290332977,-0.2295909673,-0.015448926,0.2639464438,0.3717529774,0.2841869891,-0.2373904288,0.2865346074,0.0564559139,-0.0776723102,0.5926782489,-0.2880095243,0.1730368584,0.1834766567,-0.0934189856,-0.0887018293,-0.1145091653,0.2175034434,0.2971912622,0.117755577,0.5743358731,-0.160858348,-0.196233809,0.0543240011,0.3400988281,-0.3454618454,-0.2840363681,-0.4877050817,-0.5236839056,-0.0042205197,-0.3721770942,-0.0787026361,0.2904574275,0.0152567113,-0.1196865737,-0.0763870552,0.0144908652,0.1442987323,0.0109875519,0.4414971173,-0.3486977816,0.0901101008,0.6808419228,-0.3170379102,0.2050163895,0.3126320541,0.0865959823,-0.4804007709,0.1092877239,0.1043396369,0.1207940802,0.3142724931,-0.0896870568,-0.2236964256,0.0016913622,-0.0164651796,0.0017646016,0.1162373573,0.1674048901,0.3277085721,-0.3291717172,-0.5208202004,0.0903667361,0.2208288163,0.0332244635,0.4332658947,0.0146515677,-0.2760310173,0.0304517318,-0.1376806051,0.9293695092,-0.1882178187,0.0197137874,0.1726040542,-0.383990407,0.0584003963,-0.2677530944,0.2663508654,-0.1903845221,-0.3532457352,-0.0340050906,-0.3336504996,0.2430554032,-0.0606812984,0.0111931264,0.5956478119,-0.0813547373,0.0939347297,0.051942464,0.0904619917,-0.1423704922,0.0577240251,0.0329636186,0.1051893458,-0.2272537053,0.2990388572,-0.1936215311,-0.230888918,-0.1308116019,-0.0655563995,-0.0890985131,-0.0723471269,-0.0746775791,0.3519776165,-0.0094720293,-0.2856935263,0.0322626196,0.2786594331,0.2132272273,-0.1411055177,-0.2179094106,0.4878760278,-0.3885922134,-0.001628659,-0.0234856009,0.0653723627,0.0743611306,-0.0114005879,-0.421651125,0.1994488835,-0.1593191028,-0.18825683,0.0412396677,0.0750640035,-0.1103762239,0.0909299999,0.014743384,-0.1974089146,-0.0220826846,-0.1685828865,0.1055674553,0.1960796565,-0.2721881568,-0.1259486526,0.1073704734,-0.3957341611,0.0294664074,0.4185599387,-0.0717896819,-0.2697695494,0.2888673842,-0.252653718,-0.186411351,-0.1533373743,0.0639422089,-0.4145129621,-0.155424282,0.2565377951,-0.0062887417,0.2278640419,0.1213084534,0.2209199369,0.2727257907,-0.4424714148,-0.0002790965,-0.6986165047,-0.1593088955,0.1652499437,-0.3136617839,0.0674873069,-0.0574709848,-0.1149328426,-0.1241893992,-0.0054318155,-0.2733653784,0.1323696226,-0.2543610334,-0.0331063978,-0.0485642329,-0.1197383627,0.3125316501,-0.0795154199,0.0412274487,-0.112618424,-0.28047207,-0.1652464718,-0.1427247375,0.1883925647,-0.1919293702,-0.0162004046,0.0087100472,-0.0256241411,0.2526685596,0.027589485,0.0102170827,0.1508691609,0.0549973883,-0.0908617899,-0.258667022,0.0971573293,0.0889883116,0.2601917386,0.1108610407,0.5788546801,0.0635835752,-0.0576905198,-0.0915762559,0.0325405672,-0.072038807,-0.1396462619,0.104491733,0.0304917637,0.3099398017,-0.3526627123,0.1116649806,0.1382086873,0.2125661969,0.1743493378,-0.208732456,0.2819304764,-0.0355677009,0.161100477,-0.0784536451,-0.1347819716,0.2081986517,-0.0087054772,0.0300102066,0.2454165518,0.2398883998,-0.0686846823,0.1356220394,0.1805455834,0.716170311,-0.1656839252,0.1646829247,0.0211391971,0.0997987464,0.0909186453,0.5014144778,0.3404444456,0.1566219628,0.4747625887,-0.09747684,0.2164920121,-0.0298287626,0.3604225814,-0.4585135579,-0.4043613672,-0.1977346838,-0.080923833,-0.0976931304,-0.1050588936,-0.1571647674,0.4915153682,-0.4113043547,0.1484201401,0.0108117824,-0.0091366079,-0.0967927203,0.0585560836,0.1124368757,-0.2258551121,-0.1174156964,-0.0631916597,0.303794235,0.070056349,-0.0320829786,0.0684157014,-0.0301989187,-0.2110674083,0.1346748322,0.3773548901,0.098997049,-0.230566144,0.2642075717,0.347799778,0.1171856821,0.3213607073,0.2767164111,0.3790933788,0.3753465414,-0.1545854211,0.0001059834,-0.0183589514,-0.1806027442,0.041069001,0.2578658164,0.0083794445,-0.1360823214,0.0483013578,0.2008868605,-0.1168294251,0.5617827177,0.0570292212,-0.0070071458,-0.2716127336,0.4583091438,-0.333306402,0.0492208712,-0.3129720986,-0.2718108594,-0.5526043773,-0.0267052893,0.2775914967,0.1155038774,0.1228216141,0.0837406665,0.0487815887,-0.0070327721,-0.0819081143,0.4508321285,0.0906260461,-0.5584523082,0.0252812188,-0.6570866108,0.1153029948,0.1005192846,0.1148271933,0.0499007329,0.0832690373,-0.4690294862,-0.0776049122,0.511138916,0.2343607396,0.5092459917,0.0947927982,-0.3792040348,-0.2515353262,0.1353820413,-0.0111658219,-0.1206037402,-0.3279915452,-0.1026712656,-0.1878145784,0.0213371553,-0.1614062786,-0.0440401174,0.0943923444,0.1164212897,0.3828518391,-0.123639904,0.6549368501,-0.0110115558,0.0350222737,-0.1622268111,-0.2287862003,0.0851554796,-0.1250772774,-0.0343030579,0.2623528838,-0.2591590881,-0.0701133311,-0.2872568667,0.4354524612,-0.0614144355,-0.4957471192,-0.1876620501,-0.0507446192,-0.2257562727,-0.008248684,0.1245377585,0.2674946785,-0.0836365893,0.146944344,-0.1605594158,-0.0961769596,0.6528552175,-0.1847657859,-0.2792511582,0.1726897806,0.3615811765,0.3517997563,0.0711644068,-0.3607063591,0.1369380355,0.3650059104,0.2070905715,-0.20161587,0.2284931093,0.0129004307,-0.0097609796,0.1081143469,-0.3145316243,0.076554276,0.0150702503,-0.0079064807,-0.0129493475]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2447","title":"dataset adversarial_qa has no answers in the \"test\" set","comments":"Hi ! I'm pretty sure that the answers are not made available for the test set on purpose because it is part of the DynaBench benchmark, for which you can submit your predictions on the website.\r\nIn any case we should mention this in the dataset card of this dataset.","body":"## Describe the bug\r\nWhen loading the adversarial_qa dataset the 'test' portion has no answers. Only the 'train' and 'validation' portions do. This occurs with all four of the configs ('adversarialQA', 'dbidaf', 'dbert', 'droberta')\r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nexamples = load_dataset('adversarial_qa', 'adversarialQA', script_version=\"master\")['test']\r\nprint('Loaded {:,} examples'.format(len(examples)))\r\nhas_answers = 0\r\nfor e in examples:\r\n if e['answers']['text']:\r\n has_answers += 1\r\nprint('{:,} have answers'.format(has_answers))\r\n>>> Loaded 3,000 examples\r\n>>> 0 have answers\r\n\r\nexamples = load_dataset('adversarial_qa', 'adversarialQA', script_version=\"master\")['validation']\r\n<...code above...>\r\n>>> Loaded 3,000 examples\r\n>>> 3,000 have answers\r\n```\r\n\r\n## Expected results\r\nIf 'test' is a valid dataset, it should have answers. Also note that all of the 'train' and 'validation' sets have answers, there are no \"no answer\" questions with this set (not sure if this is correct or not).\r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.8.0-53-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyArrow version: 1.0.0\r\n\r\n","comment_length":50,"text":"dataset adversarial_qa has no answers in the \"test\" set\n## Describe the bug\r\nWhen loading the adversarial_qa dataset the 'test' portion has no answers. Only the 'train' and 'validation' portions do. This occurs with all four of the configs ('adversarialQA', 'dbidaf', 'dbert', 'droberta')\r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nexamples = load_dataset('adversarial_qa', 'adversarialQA', script_version=\"master\")['test']\r\nprint('Loaded {:,} examples'.format(len(examples)))\r\nhas_answers = 0\r\nfor e in examples:\r\n if e['answers']['text']:\r\n has_answers += 1\r\nprint('{:,} have answers'.format(has_answers))\r\n>>> Loaded 3,000 examples\r\n>>> 0 have answers\r\n\r\nexamples = load_dataset('adversarial_qa', 'adversarialQA', script_version=\"master\")['validation']\r\n<...code above...>\r\n>>> Loaded 3,000 examples\r\n>>> 3,000 have answers\r\n```\r\n\r\n## Expected results\r\nIf 'test' is a valid dataset, it should have answers. Also note that all of the 'train' and 'validation' sets have answers, there are no \"no answer\" questions with this set (not sure if this is correct or not).\r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.8.0-53-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyArrow version: 1.0.0\r\n\r\n\nHi ! I'm pretty sure that the answers are not made available for the test set on purpose because it is part of the DynaBench benchmark, for which you can submit your predictions on the website.\r\nIn any case we should mention this in the dataset card of this dataset.","embeddings":[-0.1097810715,0.1781133115,-0.1390384734,0.3103133142,0.163477093,-0.0822487101,0.3094846606,0.4499256909,-0.0519624054,0.1181879789,0.1611241698,0.3690258563,-0.1847231686,-0.2479670346,0.0857571959,0.1090602502,-0.1185330972,0.0875692368,-0.072045058,-0.3029209673,-0.1661467552,0.0697546378,-0.1738792509,-0.1399212331,-0.2601901889,0.1874664277,-0.1425664574,0.0713289008,0.0379400924,-0.5859718919,0.284163028,-0.1142051965,0.0199329518,0.333371073,-0.0001112121,-0.0196807217,0.2017178386,-0.062374223,-0.4061582983,-0.1288897246,-0.5137016773,-0.258731097,0.1562203616,-0.2437217832,-0.2343089134,-0.2186056525,-0.1217001155,-0.2548487484,0.2813834548,0.5967983007,0.2503398359,0.1645575166,-0.2127765715,-0.181281805,0.0325633399,-0.1947725713,-0.1360811144,0.3340584934,0.0669581518,-0.2101397663,0.2344431132,0.1632981002,0.0818397924,-0.0216818526,-0.0547687039,-0.0577248409,0.2928401232,-0.2534742355,0.0799158737,0.2677954137,0.3713645339,-0.1823160201,-0.2502909899,-0.3610525429,0.2017528862,-0.0994583741,0.2094523758,0.2093106061,-0.2533089519,0.0622792356,-0.0272087287,0.2265873104,-0.0623358525,-0.0309756249,-0.252605617,0.4209857583,0.071873188,-0.1054385453,-0.057288602,0.155698061,0.0889472514,-0.3112669885,-0.0931109264,0.1401284486,-0.437082231,-0.0814287961,-0.0774305537,-0.0911607966,0.2894621193,0.1992064863,0.3563617766,-0.1585877538,0.1394703835,-0.1587534994,0.1482955068,0.4554486871,0.2018250078,0.2326283455,0.0324285403,0.1187070757,-0.0405059978,0.1136837304,-0.2118980885,-0.2287192792,0.130190894,0.2099892199,0.385589391,-0.2329800129,-0.3287391067,0.3438715339,-0.3541021347,-0.1151092127,-0.0645849332,0.3388816714,-0.3086433709,0.1680484265,-0.1507396251,0.2771680653,-0.2411506772,-0.2918148637,-0.2530305386,-0.1557367593,-0.3436105549,-0.1896779835,0.3385716081,0.2474781722,0.2008586377,0.0807726979,0.1476811916,-0.0245352928,0.2117058486,-0.3234573901,0.1355656236,0.3594354391,-0.1031559482,0.1419624388,0.155846864,-0.4854317009,0.078876555,0.5665237904,0.1898899525,-0.1312184483,0.0712206215,0.2513558865,-0.3031832278,0.0074214432,0.0121108666,0.3110601008,0.0177013874,-0.0367622524,0.3407376409,-0.3106961846,0.0375176482,-0.2748856246,0.6375951171,0.4450051785,-0.3981798887,0.1778471321,0.0053822226,-0.0573090203,-0.1676468551,-0.2092047036,0.0114109581,0.1053504422,-0.1846325696,-0.0219071824,0.2112338394,-0.3385699987,-0.4776214957,-0.1273833364,0.1589169949,0.2947568297,0.0232288837,-0.0901725516,0.1367347389,0.250972122,0.0836565495,0.1850374043,-0.0195128527,-0.2015656829,-0.4516374171,0.022746345,0.3778617084,0.369409591,0.2642970085,-0.0260876324,-0.1160885617,-0.3355936706,0.258883208,0.2518937588,-0.0911189467,0.1675364822,0.1315605342,-0.1300174147,0.0336127728,-0.0891024619,-0.4266718328,0.3443760872,0.0254590958,0.1381693631,0.1315917969,0.0282772034,-0.4881201982,0.0094749797,-0.3255443871,-0.5843749046,0.1927192807,0.2720829248,-0.0129175643,0.3838590384,-0.1999856383,0.2762494385,-0.0239834562,-0.0731474161,-0.180801928,0.2311245054,-0.0168623086,-0.1607399285,-0.0748773739,0.3288769126,-0.0815293714,0.0815682784,-0.1294564605,0.1215953976,0.0694558918,-0.3059769869,0.1249368936,0.4802087843,0.0908802971,-0.3299107254,-0.1332969666,0.2975678146,0.2645037472,-0.0471613258,-0.1427665651,0.3639020622,-0.0844002962,0.2512452602,-0.1064912006,0.1490966529,0.2662833631,-0.2230786979,-0.1498378217,0.0792866573,0.101217553,0.0388417765,0.3362998664,0.1632045805,-0.30365628,0.0747909546,0.3500984013,-0.0923981294,0.1034012362,0.0270148441,-0.1550431103,0.0407557152,0.2616890967,0.3052296937,0.663975358,0.2440157086,0.0889746249,0.0574707016,0.0073665157,-0.2085778862,0.1379159093,-0.2342082411,0.0052957698,0.4090867341,0.2028958201,0.0527467206,-0.0890481845,0.1290135384,-0.0421295464,0.1984057575,-0.2945136428,-0.0023095177,-0.175065726,0.1735696644,0.1295403391,-0.2093019634,-0.2406195253,-0.1649426967,0.2955542505,-0.3521289825,0.0030031323,0.3188256621,-0.2050446272,0.2272210419,-0.1827200055,0.0839221925,-0.2140140831,-0.1988140792,-0.4165837169,0.1731057316,-0.0749871582,-0.0569408685,0.2524851561,-0.2461756319,-0.1882236302,0.0628474876,-0.0934975147,0.0965294465,-0.0814576745,0.6559173465,-0.0215197261,0.07231085,-0.3762665391,-0.0415688604,0.3522254825,-0.0448904447,-0.0101421401,0.1523817182,0.240395084,0.0174525771,-0.0148715144,-0.7003239393,-0.3180599511,-0.209393099,0.0168716703,-0.0373114683,0.1344526559,0.1834020615,0.267659843,0.1186418608,-0.2419462055,-0.0092015984,-0.1599935293,-0.0506427437,0.3487632871,-0.1489490867,-0.5024710298,-0.1494236588,-0.2436115593,0.2617567778,-0.1332828999,-0.5106116533,-0.0782305673,-0.2623902559,0.3865682781,-0.189012453,-0.0559495389,0.3177070916,-0.1946865171,-0.0775459409,-0.2881661057,-0.0538055412,0.3896780312,-0.0496959127,0.1622001827,-0.2436951101,0.2100416571,0.0601756051,0.5309695005,0.1187317669,0.1414560527,0.2202576697,-0.2097058743,0.3259542584,-0.2176638842,-0.3961415589,0.028869614,0.2446224838,0.1261309236,0.2928257287,0.0213247128,0.0901763365,-0.1461826861,0.0045509222,-0.4204443395,-0.2203031331,0.1375111192,-0.6682155132,-0.0042713466,0.027372418,0.120307371,-0.0047598104,-0.0708991215,-0.1687237173,0.0749271661,-0.0112583991,-0.071102418,-0.3702082634,-0.0972271413,-0.0766305625,0.1608388424,-0.1226447299,0.4465252161,-0.0860416666,-0.1159292907,0.1933716238,0.1177214533,0.7862274647,-0.043729797,-0.1693386137,0.3484087288,-0.0574469119,-0.4323269427,-0.2542470694,-0.2168700099,-0.0288234726,-0.1057343781,-0.1255831867,-0.3221476674,-0.0818284899,-0.0219380576,0.1284460127,-0.3099874556,-0.1010548621,-0.3357022107,-0.197364375,-0.1284754872,0.2869135439,0.0632433593,0.2137039602,-0.2322818637,-0.2141743451,-0.1322432905,-0.1597712934,0.2611672878,0.2644218504,0.2692783177,-0.046671018,0.2410868555,0.0192893948,-0.1497373134,0.0243906081,0.3924176991,-0.0886380747,0.0058868336,0.1942727864,-0.307692796,0.5265529156,0.2358260602,-0.2248591781,-0.2387011945,-0.4207187295,0.1047620848,-0.1507255137,-0.0318781883,0.3642633855,0.004929001,0.0762219876,-0.4885349572,0.0843964666,-0.2333078682,-0.2787137628,0.1871903241,0.1780787855,-0.1186535954,0.2873227,0.0047131954,0.9321403503,-0.0109090796,0.1728473306,0.4116937816,-0.0254122112,0.1783474386,-0.2065520287,0.1283474267,-0.2668017447,-0.5641277432,0.2046461701,-0.1305363923,0.0850248784,0.256545186,-0.063496381,0.2525484562,-0.2689885199,0.0201526806,0.2733145654,-0.0011872649,-0.1138403714,0.3495713472,-0.5553538203,0.1829525083,0.1221838892,0.0446792878,-0.1100507379,-0.0654555708,-0.2123873532,-0.1899006963,-0.3045633435,0.1847120076,-0.4208047092,-0.1004084572,0.4122463167,-0.1661561579,-0.089938499,0.6880053282,0.3051171899,0.2804513872,-0.0177321136,0.2763749957,0.0597614422,0.2724177241,0.2788799703,0.0651513115,0.1753791124,-0.2156325728,-0.5686722994,0.2594637573,-0.0479034036,-0.1648970991,-0.1486432701,0.1658141613,0.1582077146,-0.0979382843,-0.1528484374,-0.0464154631,0.2232116014,-0.0733547434,0.2347389311,-0.0565918125,-0.1036368534,0.3340063393,0.2225227803,-0.3776361346,-0.0384170786,0.2935939431,0.0073563973,0.3854387701,0.6925075054,0.0737850443,0.0034340923,-0.353564322,-0.1314850748,0.4564798474,-0.5438268185,0.1469975263,-0.0994199887,-0.0860918835,-0.1598635465,0.0676042587,0.2713715732,-0.2666700482,0.0945769176,-0.4723008573,0.003577638,0.2808353901,0.0397526957,0.0787774622,0.0395650603,-0.1446521878,-0.0091663133,0.2214588374,-0.4289362729,-0.0268759262,-0.4231907129,0.0726351663,-0.5557461381,-0.1046495289,0.234384656,-0.078605324,0.2053902149,-0.242931962,-0.1031935737,-0.3527093232,-0.1048281118,0.1370175183,-0.025814224,-0.0174612757,-0.2341402918,-0.2490492463,-0.134771049,-0.079195708,0.1172598079,0.0486217067,-0.0304800831,0.2638103664,0.2025879323,0.287247926,-0.1684889942,0.1114346758,-0.3002915382,-0.2510043085,0.1055681035,0.031623695,0.1123343855,-0.0497318096,-0.012790909,0.0725676492,-0.2300700843,0.0176178943,0.2510806322,-0.1145708561,0.2062974274,0.0430524088,0.0313124098,0.1156378239,-0.3033095598,0.1346184909,0.0145561816,0.2969893515,-0.4703717232,0.0845935643,0.0236771684,0.2388776988,0.1069848463,0.1374826133,0.2607826889,-0.3211278617,0.1386264712,0.232689485,-0.0135323852,-0.0238461979,0.1704480797,0.536796689,-0.2039663792,0.0089846905,-0.1139198467,-0.1224127784,-0.1746314466,-0.010902836,-0.188633725,0.3858889937,0.0586990714,-0.2206844985,-0.0185337681,-0.0937949196,0.1625819206,0.0596138053,-0.0633414686,-0.0821070895,-0.0366841704,0.2787972689,-0.0097260168,0.1430375427,-0.2206464857,-0.1821599305,0.0156309195,-0.0064553651,-0.1274029016,-0.1969024092,-0.327578336,-0.3353548944,-0.043258559,-0.1840711385,0.1011587158,-0.0019415362,-0.2903952599,-0.1264003217,0.3109066188,0.3682549894,0.1263703853,-0.0104895653,0.2845536768,0.2754612863,0.1594679505,0.1338799,0.5755476356,0.0708153024,0.2835338116,-0.3751259446,0.0089568933,0.1179051995,-0.0702083185,-0.2208065391,0.5092964172,0.110339798,0.1771219522,0.0724059343,0.2322898954,-0.2277542502,0.2104780972,0.0882982165,0.0542932339,-0.1977096796,0.1949516386,0.0193385091,0.0532162227,0.0095316954,0.1372045279,-0.429400295,0.0619726889,0.1822433025,-0.1039695889,0.0761234164,-0.3123710155,0.1164202765,-0.246245876,0.2577493191,0.1471637338,0.349937737,-0.2469520271,-0.0355953239,-0.57020998,0.1726404279,0.118252866,0.1476763189,0.2577409148,-0.0239245389,0.0133776572,-0.067011863,0.0889681354,0.2568540275,-0.2957231998,0.1841738522,-0.5471511483,-0.3395519853,-0.2542598248,0.2178045064,-0.0658861026,-0.1400207579,0.0318742283,-0.0573087633,0.071266681,0.2236217111,-0.0883623809,0.2697107792,0.2985754311,0.1739414036,0.1848462075,0.4591228664,0.1161993369,-0.0168215614,-0.1099833697,0.0506256595,0.0350617096,0.2203918546,0.0867137909,0.1476565301,0.0235554148,-0.1127965748,0.1414250433,0.1362455934,0.200488627,-0.0160135832,-0.2218059003,0.0102881258,0.1975569576,0.2081076801,-0.0014722819,0.1964023411,-0.1311181039,0.2788808048,-0.2168022841,-0.2409404963,0.3765819371,-0.2864558697,-0.0280715358,-0.1468076259,0.1772669405,0.1771667749,-0.0874620378,-0.413985163,-0.2478870749,0.4810094237,-0.0441986024,-0.0525893793,0.2417891175,-0.0084548518,0.147353664,0.0200292096,0.3900100887,0.0255171228,0.0300645344,-0.0610901043,-0.2780912519]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2447","title":"dataset adversarial_qa has no answers in the \"test\" set","comments":"Makes sense, but not intuitive for someone searching through the datasets. Thanks for adding the note to clarify.","body":"## Describe the bug\r\nWhen loading the adversarial_qa dataset the 'test' portion has no answers. Only the 'train' and 'validation' portions do. This occurs with all four of the configs ('adversarialQA', 'dbidaf', 'dbert', 'droberta')\r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nexamples = load_dataset('adversarial_qa', 'adversarialQA', script_version=\"master\")['test']\r\nprint('Loaded {:,} examples'.format(len(examples)))\r\nhas_answers = 0\r\nfor e in examples:\r\n if e['answers']['text']:\r\n has_answers += 1\r\nprint('{:,} have answers'.format(has_answers))\r\n>>> Loaded 3,000 examples\r\n>>> 0 have answers\r\n\r\nexamples = load_dataset('adversarial_qa', 'adversarialQA', script_version=\"master\")['validation']\r\n<...code above...>\r\n>>> Loaded 3,000 examples\r\n>>> 3,000 have answers\r\n```\r\n\r\n## Expected results\r\nIf 'test' is a valid dataset, it should have answers. Also note that all of the 'train' and 'validation' sets have answers, there are no \"no answer\" questions with this set (not sure if this is correct or not).\r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.8.0-53-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyArrow version: 1.0.0\r\n\r\n","comment_length":18,"text":"dataset adversarial_qa has no answers in the \"test\" set\n## Describe the bug\r\nWhen loading the adversarial_qa dataset the 'test' portion has no answers. Only the 'train' and 'validation' portions do. This occurs with all four of the configs ('adversarialQA', 'dbidaf', 'dbert', 'droberta')\r\n\r\n## Steps to reproduce the bug\r\n```\r\nfrom datasets import load_dataset\r\nexamples = load_dataset('adversarial_qa', 'adversarialQA', script_version=\"master\")['test']\r\nprint('Loaded {:,} examples'.format(len(examples)))\r\nhas_answers = 0\r\nfor e in examples:\r\n if e['answers']['text']:\r\n has_answers += 1\r\nprint('{:,} have answers'.format(has_answers))\r\n>>> Loaded 3,000 examples\r\n>>> 0 have answers\r\n\r\nexamples = load_dataset('adversarial_qa', 'adversarialQA', script_version=\"master\")['validation']\r\n<...code above...>\r\n>>> Loaded 3,000 examples\r\n>>> 3,000 have answers\r\n```\r\n\r\n## Expected results\r\nIf 'test' is a valid dataset, it should have answers. Also note that all of the 'train' and 'validation' sets have answers, there are no \"no answer\" questions with this set (not sure if this is correct or not).\r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.8.0-53-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyArrow version: 1.0.0\r\n\r\n\nMakes sense, but not intuitive for someone searching through the datasets. Thanks for adding the note to clarify.","embeddings":[-0.0380188413,0.1775163263,-0.1093787625,0.3759938478,0.1982209831,-0.0062384722,0.3613942266,0.3695282042,-0.064131923,0.0352284424,0.1037120596,0.3792202771,-0.1990861148,-0.1972342879,0.0432114191,0.1114880666,-0.0155535378,0.1072037444,-0.0602785125,-0.3192011118,-0.2311211377,0.0949319154,-0.2330937833,-0.0949132293,-0.2921154499,0.1194490716,-0.2043208033,0.1561518759,0.0097583635,-0.5535634756,0.2813077569,-0.1492449194,-0.0167530496,0.4255777895,-0.0001118015,-0.0490550287,0.2256864905,-0.0351710208,-0.3814617097,-0.2443181276,-0.4445735216,-0.2766770124,0.1798815578,-0.2216132879,-0.1694054008,-0.2560254335,-0.1121231094,-0.1137276664,0.2730475664,0.5292611718,0.2409110069,0.0941973403,-0.1843172163,-0.133368358,0.1004744396,-0.094907403,-0.1246037483,0.3207915425,0.1239120662,-0.2727337778,0.2450588793,0.161215663,0.0185226779,-0.1245784909,0.0030624783,0.0022467026,0.2287279367,-0.2950288951,0.0905454755,0.2719071209,0.5253135562,-0.1579341441,-0.2285391539,-0.3373923302,0.1966067106,-0.0812611207,0.2187606841,0.2738299966,-0.1987095326,0.0751456246,-0.0344927274,0.2293820381,-0.0945535004,0.0515003391,-0.3328437805,0.3991541564,0.0483285598,-0.0972458422,-0.1221484467,0.1052148938,-0.0369532965,-0.2576110065,-0.0345442742,0.2256599665,-0.5387450457,-0.0272687823,0.0190763511,0.0055163186,0.2636825144,0.3198458254,0.2951920927,-0.2196024954,-0.0240470916,-0.1002113745,0.2141585946,0.3907952011,0.1931944937,0.2982254028,0.0379006118,0.0519817732,-0.0990941077,0.1056296602,-0.2550399899,-0.2089608759,0.2388080806,0.1443412602,0.4825346768,-0.2045682222,-0.2926563323,0.33776173,-0.4190512002,-0.14759776,-0.0480976924,0.4000511169,-0.2404103428,0.0922677219,-0.1325743198,0.2541822791,-0.2532403469,-0.2583611012,-0.2444072962,-0.1555507779,-0.3043295741,-0.2328317314,0.2940024734,0.0891265571,0.2044060081,0.1051169336,0.0744469836,0.0001031693,0.1703083217,-0.3443287909,0.1115709841,0.3897327185,-0.1513686329,0.2131369114,0.2146556675,-0.383913815,0.02497885,0.5199177265,0.0757908151,-0.0811048597,0.0184673406,0.2310703397,-0.2154954672,-0.0448467247,0.0854428932,0.3162041008,0.0786264464,-0.1287725717,0.2978664339,-0.3184589744,0.0218626391,-0.3275707662,0.5358551145,0.5741321445,-0.4068905711,0.1616363525,-0.0375148766,-0.054950811,-0.1443333924,-0.1478255391,0.0347904004,0.1809509993,-0.2410242856,0.0704634935,0.1228723824,-0.3233373165,-0.3895311654,-0.0886804685,0.189490363,0.3247493207,0.0354473218,-0.1182397828,-0.0009166035,0.1604258269,0.1369958073,0.1970087141,-0.0356904455,-0.1906371862,-0.382917881,0.0364628918,0.4099152982,0.3123600483,0.2516899705,-0.0237091016,-0.0720952228,-0.43276003,0.2436808944,0.2088194191,-0.1370542198,0.1104608774,0.2225861549,-0.0562018119,0.1033200324,-0.122828342,-0.4905353189,0.3253698051,0.1002145559,0.2054318339,0.0529075786,-0.0096935313,-0.4930993319,-0.0230460372,-0.3354252577,-0.5171920061,0.1825162619,0.251416862,-0.0514378697,0.3741511106,-0.2586688697,0.2646823525,0.0258375742,0.0281636659,-0.1718530804,0.2696374357,-0.0179110821,-0.0689149946,-0.1050148308,0.2524372339,0.0164250154,0.0803575516,-0.0995145738,0.1832994074,0.0995594263,-0.253220886,-0.0125661502,0.4138720632,0.107801415,-0.2240742296,-0.101709418,0.2977459133,0.3071589768,-0.1076306626,-0.1429254711,0.2557155192,-0.0801076889,0.3540757895,-0.0574544966,0.1349477023,0.2534440458,-0.2289259285,-0.1601988971,0.0503910258,0.1779779941,0.0509596504,0.4108891189,0.2816082239,-0.3163895905,0.1226554215,0.2962827384,-0.1855596006,0.0877043977,0.0665116683,-0.1073962599,0.1110722125,0.2065813094,0.3290757835,0.6542757154,0.2463446856,0.0147651481,0.0547763333,-0.0812356845,-0.2099500895,0.1697426289,-0.1730263829,0.096653007,0.4493786395,0.1986962557,0.1220147237,-0.1478346735,0.0132789714,-0.0728756264,0.2129718065,-0.4133994281,0.0900946483,-0.2010109127,0.1117902175,0.0624766946,-0.2461852133,-0.2787481248,-0.2188784778,0.2188399732,-0.440474391,-0.0339393541,0.2548559308,-0.2764028907,0.2695633173,-0.1428872794,0.0761199817,-0.2370153219,-0.1646420807,-0.4196374416,0.1350830942,0.018620966,-0.1070711389,0.2462532967,-0.2813128531,-0.2374907285,0.1074723601,-0.1896662116,0.1197578311,-0.1743261963,0.6320790052,0.0549536794,0.0948213637,-0.2963761389,-0.0584234372,0.3697791398,-0.0066816909,-0.0874721557,0.1727496982,0.2285975516,0.0186050907,-0.0421031266,-0.7604624033,-0.3359158635,-0.2818158567,-0.0193221159,0.0256294515,0.1912407279,0.3360476792,0.247378543,0.1308222115,-0.2239892334,0.0551474728,-0.1709394157,-0.0125009604,0.3971096873,-0.1424566507,-0.4635104537,-0.0660351589,-0.2137093693,0.2111829817,-0.0478711091,-0.4919604957,-0.0667051822,-0.2664311826,0.2749800086,-0.1996053755,0.0907574371,0.2340377569,-0.1649851054,-0.0545590706,-0.2723251283,-0.1440219879,0.3495979905,-0.09067256,0.1211499199,-0.1907510161,0.2533167601,-0.0005464607,0.4581643641,0.0876599401,0.1873094589,0.2734171748,-0.1752267182,0.4184911251,-0.232049495,-0.4264899194,0.0626349822,0.3019270599,-0.0071080551,0.3204470873,0.0907081515,0.094932422,-0.1675326675,-0.0532751828,-0.5038219094,-0.2043175399,0.136664331,-0.5980351567,-0.0114598554,-0.0190638714,0.1001692563,0.0272329524,-0.0467808992,-0.2239553183,0.1314517856,0.0312177129,-0.0720691085,-0.4294918478,-0.0658830479,-0.0928294584,0.1315648705,-0.1544346809,0.4282396734,-0.08382947,-0.1761914939,0.1934204549,0.1368543357,0.9014118314,-0.0768921822,-0.1909402311,0.3454948962,-0.0574556962,-0.4061219692,-0.2854189575,-0.2000975758,0.0180401299,-0.096505858,-0.0459479094,-0.3890874386,-0.0728922635,0.0727900788,0.1111529917,-0.2667551935,-0.1379171908,-0.3741106391,-0.2309468985,-0.1302396059,0.2148495317,0.056260623,0.1579615772,-0.2149170339,-0.1812154949,-0.0974434987,-0.1467895061,0.3230846524,0.2487399131,0.2922794521,-0.0289353505,0.1983170062,-0.0189653449,-0.1342496425,0.067194894,0.5639817715,-0.1479319483,-0.1124052927,0.1748312712,-0.2797798514,0.5428379774,0.2414059043,-0.187085107,-0.2742554843,-0.3951379955,0.0226076059,-0.135164991,-0.1171409413,0.3611705601,-0.0359364226,-0.0520918183,-0.5506644249,0.1386537701,-0.1921025664,-0.2379793525,0.1963515878,0.2204592079,-0.1451257765,0.4353651404,-0.0496724285,0.923674345,-0.0572254471,0.1938799769,0.4561693668,0.0061992947,0.3132887185,-0.1844435185,0.0707451776,-0.2815214694,-0.5043560266,0.1649122834,-0.1241306737,0.0716559738,0.3066304326,-0.1143863052,0.2874799073,-0.3841590881,0.1655444056,0.302968353,-0.1174090877,-0.1647827476,0.335822612,-0.6336981654,0.1677838266,0.144730106,-0.0005199336,-0.1407932937,0.0240201559,-0.2332928926,-0.2857978344,-0.3050089478,0.165056631,-0.3812279701,-0.1564989835,0.4907232225,-0.126595661,-0.0396447293,0.6422729492,0.1531253457,0.240921244,-0.0450376943,0.2599819899,0.0654424801,0.2937280238,0.2277170718,-0.0405051559,0.3724682927,-0.0904825181,-0.5008720756,0.2400508225,-0.0874692425,-0.1317161322,-0.1945071816,0.2537884414,0.1201157421,-0.1207486242,-0.1940658391,-0.0599333085,0.2955269217,-0.0697546974,0.2095025331,0.0012138132,-0.0050029284,0.3027437627,0.221565485,-0.3633124232,-0.0551403202,0.2310792357,0.0297792405,0.3197757602,0.7269604802,0.1023237482,-0.0020689485,-0.3283357918,-0.1111943573,0.4242294431,-0.4804447591,0.1672976613,-0.0230816491,-0.0255545229,-0.0829312056,0.0299031977,0.1886943579,-0.2846041322,0.0424859971,-0.5493791699,0.0757252127,0.2848809958,0.0453587174,0.1237667799,0.1141305789,-0.1974124312,0.0480159223,0.2211819738,-0.4114348888,-0.01587479,-0.4132530987,0.0946036503,-0.5187206864,0.0080775311,0.1861641854,-0.1222391874,0.2089359313,-0.2573083937,-0.0663395077,-0.3209450841,-0.1183250844,0.1647927165,0.0406224728,-0.0084510427,-0.2012792528,-0.2471342236,-0.1016666219,-0.1423341483,0.0885669664,0.130761072,-0.0714398623,0.2590895593,0.154194504,0.24488464,-0.1563756019,0.1397523433,-0.330016464,-0.3168635368,0.2431919277,0.0163534563,0.0441778675,-0.00576352,0.0132150296,0.1420336366,-0.2390473187,0.000405962,0.3030936122,-0.1098769009,0.2360560596,-0.035262268,0.0341412686,0.2104491442,-0.3356778026,0.1342542022,0.1254933327,0.2870281637,-0.4436323047,0.1100875884,0.0421583094,0.2080154419,0.0710081682,0.1533443183,0.1693098098,-0.3946666718,0.0037216162,0.2072191834,0.1631776392,0.0951807722,0.2361098975,0.5443534255,-0.2297906727,0.0586381331,-0.0552058741,-0.1819313467,-0.1783489883,0.0119293462,-0.1506992131,0.2770531178,-0.0804102123,-0.1646328121,-0.0688692555,-0.1995176822,0.2449262142,0.1648693085,-0.0788615197,-0.0950233713,-0.0478952155,0.267169863,-0.009718935,0.1053664312,-0.2423919737,-0.1233429909,0.0048694997,0.0407748297,-0.2498860359,-0.2379471213,-0.2651355863,-0.3146309257,-0.1096802652,-0.2998905778,0.1182250232,0.0206025671,-0.2783414423,-0.2631770372,0.3560770154,0.2780511379,0.1248760298,-0.0630218014,0.2743388712,0.2478071004,0.1056769267,0.2066354752,0.5216087699,0.2522513866,0.4706481695,-0.2216774672,-0.0317643173,0.0302272048,-0.0459728763,-0.2372834831,0.5032395124,0.0284663271,0.2390486747,0.0533830784,0.2061213553,-0.1913520396,0.2023074925,0.1169336587,0.0357270129,-0.2932557464,0.1734133363,-0.0888204351,0.0540347621,0.0014768103,0.1422104537,-0.5204792619,0.018835457,0.2046675235,-0.0279563926,0.1461400688,-0.2721902728,0.11039491,-0.1316324323,0.235606581,0.2926119864,0.3262927234,-0.2702353895,0.0057397955,-0.6387885809,0.1406664401,0.0854919329,0.1015134081,0.2069918513,-0.0526742525,0.0517801531,-0.0708198175,0.0629346222,0.2157251835,-0.2645030022,0.1946753114,-0.55256778,-0.3293536007,-0.2111410201,0.2391730398,-0.1025165915,-0.1405192167,0.0238683112,-0.1510090828,0.0693714917,0.1784316301,-0.0718040243,0.2928219438,0.260678947,0.186637938,0.2264152765,0.4868241847,0.1238885298,-0.006214804,-0.2220587283,0.0404415391,-0.0468368419,0.2719148099,0.0386465117,0.1407654285,0.0147916088,-0.1896041483,0.1678962559,0.1222179756,0.1433686316,-0.0380159505,-0.2138062119,0.0654852688,0.1460559815,0.1398108751,-0.0475838967,0.1428248584,-0.1131304875,0.2552953959,-0.2337553054,-0.339769721,0.3369910419,-0.3220007122,-0.0184333455,-0.2039360106,0.1805533767,0.1736849844,0.0004843457,-0.4455561042,-0.2057563066,0.5004960299,-0.0006246518,-0.0927117839,0.2928690016,-0.0615588501,0.0878623798,-0.035013441,0.3499043882,0.0436590314,0.0033152625,-0.0028539025,-0.2417056113]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2446","title":"`yelp_polarity` is broken","comments":"```\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/script_runner.py\", line 332, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 233, in \r\n configs = get_confs(option)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 604, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 588, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 148, in get_confs\r\n builder_cls = nlp.load.import_main_class(module_path[0], dataset=True)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 85, in import_main_class\r\n module = importlib.import_module(module_path)\r\nFile \"\/usr\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\nFile \"\", line 1006, in _gcd_import\r\nFile \"\", line 983, in _find_and_load\r\nFile \"\", line 967, in _find_and_load_unlocked\r\nFile \"\", line 677, in _load_unlocked\r\nFile \"\", line 728, in exec_module\r\nFile \"\", line 219, in _call_with_frames_removed\r\nFile \"\/home\/sasha\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/yelp_polarity\/a770787b2526bdcbfc29ac2d9beb8e820fbc15a03afd3ebc4fb9d8529de57544\/yelp_polarity.py\", line 36, in \r\n from datasets.tasks import TextClassification\r\n```","body":"![image](https:\/\/user-images.githubusercontent.com\/22514219\/120828150-c4a35b00-c58e-11eb-8083-a537cee4dbb3.png)\r\n","comment_length":118,"text":"`yelp_polarity` is broken\n![image](https:\/\/user-images.githubusercontent.com\/22514219\/120828150-c4a35b00-c58e-11eb-8083-a537cee4dbb3.png)\r\n\n```\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/script_runner.py\", line 332, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 233, in \r\n configs = get_confs(option)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 604, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 588, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 148, in get_confs\r\n builder_cls = nlp.load.import_main_class(module_path[0], dataset=True)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 85, in import_main_class\r\n module = importlib.import_module(module_path)\r\nFile \"\/usr\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\nFile \"\", line 1006, in _gcd_import\r\nFile \"\", line 983, in _find_and_load\r\nFile \"\", line 967, in _find_and_load_unlocked\r\nFile \"\", line 677, in _load_unlocked\r\nFile \"\", line 728, in exec_module\r\nFile \"\", line 219, in _call_with_frames_removed\r\nFile \"\/home\/sasha\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/yelp_polarity\/a770787b2526bdcbfc29ac2d9beb8e820fbc15a03afd3ebc4fb9d8529de57544\/yelp_polarity.py\", line 36, in \r\n from datasets.tasks import TextClassification\r\n```","embeddings":[-0.1024581045,-0.4164357781,-0.1597157121,0.1106378362,0.2700803578,-0.1342438161,0.0620417446,0.324896127,0.115024209,0.0208610073,-0.0733032897,0.0981361493,-0.0775282383,0.088160865,-0.0272907391,-0.0547005758,-0.0670888275,0.5007214546,-0.279864639,-0.1024006978,-0.187374115,0.3047941029,0.0469926558,0.1529836655,-0.0215161424,-0.0998178199,0.075682044,0.2807609439,-0.2198141068,-0.3285527229,0.0454352498,0.0923940316,-0.2174301147,0.1122374758,-0.0001101976,-0.0683468431,0.3514304161,-0.0494548529,0.0267474912,-0.3109245896,0.0556759872,-0.1162542105,0.0716132447,-0.175342679,-0.130285576,-0.2852269709,0.0069500348,-0.0522034727,0.5125470757,0.217505917,0.2951674759,0.0969505981,0.07746052,0.1119322181,0.0685760379,0.2308505774,-0.0519353822,0.0755407736,-0.1282691211,-0.4185512364,-0.0851896256,0.3796780705,0.0417252518,0.0093526682,0.1816565841,-0.2398160696,0.64311409,0.0964671522,0.0561751276,0.0915924609,-0.2606853843,0.0230086315,0.117477037,0.2456567436,-0.1648456007,-0.4188649952,0.0975314304,0.3891115189,-0.2444213927,-0.1872090697,-0.0527212434,-0.096131444,-0.2303685993,0.2770906687,-0.0000369342,0.7835253477,-0.001544368,0.1822980791,0.1811497062,0.0369461775,0.195466727,0.3975942433,-0.1270466596,0.0670382679,-0.0557206124,0.0359687619,0.2267304361,0.1506973207,0.0570467897,0.2722670436,0.3639073372,0.0971900225,0.2167173773,0.1387993097,0.0001440624,0.1953643113,0.0682745352,0.4745216668,0.2713787854,0.5279783607,0.1639027745,0.1432444304,0.1657558382,-0.4216600358,-0.0301527083,0.1145276353,0.0603313223,0.1303584427,-0.2250370532,0.0096477158,-0.0831861794,0.2637470663,0.202659294,0.2387269735,0.0568376854,0.2667765915,0.2185367197,0.0769397616,-0.0639017597,0.0330666006,-0.1990626603,0.3460360169,-0.3091924787,-0.0692949891,0.078894414,0.3844735622,0.5319616199,-0.1714345962,-0.2256480902,-0.0870869234,-0.02714861,-0.0683751851,0.1519236267,0.250631094,-0.0791389421,-0.0542153344,0.2363597304,0.0717259124,-0.0663518533,-0.1300756931,-0.0314719155,0.0940547511,-0.1985687315,0.3028070033,0.0391189456,0.0301341806,0.0850731954,0.0400215872,0.2629771531,-0.2812221646,0.1904451996,-0.2715656161,-0.2210510522,-0.1143209264,0.074341163,0.3664713204,-0.0416049883,-0.4093360305,-0.293621242,-0.1419834346,0.223440975,0.1826951951,0.0799330994,-0.0352958068,-0.2192550302,0.0191173293,0.3005329072,-0.3774944544,-0.1865869164,-0.1834850907,-0.1630088538,-0.0756063834,-0.0483941808,0.1819626689,-0.0024486179,0.1328393668,0.155842334,0.650113523,0.1210432053,0.1192088872,-0.4229591787,-0.0155736674,0.2117554098,-0.052793894,0.2383706421,-0.0355702266,0.0632933974,-0.1881220937,0.1184785962,-0.0154105201,0.2185810059,0.2496159971,0.5116145015,0.0674153715,0.086702086,0.0805798769,0.0377302803,-0.0798633546,-0.9023961425,0.1307978183,-0.3075226247,0.1495981961,-0.2495387644,-0.1359956414,-0.4057085216,-0.4762876928,0.2574538291,0.1651846766,0.1010157689,0.0820708126,0.1044291332,0.3060425222,-0.3019443452,0.0823413953,0.0181695465,0.2572535276,-0.2900827527,-0.0522765554,0.0459960885,0.2013465911,0.3136645854,-0.1686371714,-0.1792932898,0.3776599765,-0.5558848977,0.4288749993,-0.1506473124,-0.3552914262,0.22080037,-0.5869683027,-0.0280094035,0.1380987912,-0.0364759825,-0.0319529921,0.1685712636,0.2660727799,-0.1766497344,0.1643361598,0.1421674043,0.2842082977,0.1625176817,-0.0230881367,-0.1731597185,-0.3042883575,0.4437310696,-0.0175074041,0.1051840931,-0.3311535716,0.017173795,0.0191193298,0.4013838172,0.3117001057,0.1925226599,-0.0715919659,0.048675023,-0.1927369386,0.1532043964,0.2270621359,0.3311017752,0.1304068714,-0.2781845629,0.0966533795,-0.1665710211,-0.137836948,0.2107174098,0.0464989357,0.3183383942,0.3407025933,0.3252840936,-0.0894416347,-0.3340715766,-0.1943020225,-0.2050064951,0.228572458,0.0228807777,0.0516976751,-0.2642994523,-0.0561268441,0.0060938667,0.1965384483,-0.0209378339,-0.2515963614,0.1971224248,-0.0967067629,-0.2391358465,0.4282850623,0.1994766295,0.1254674941,0.2627290189,-0.1705452055,-0.3780843914,-0.2816615403,-0.0331926383,0.0982737243,-0.0743730143,-0.0480498262,0.4588577449,-0.1241153181,-0.0208083335,-0.2733244598,-0.466245681,0.0604558438,-0.1246377155,-0.0323550403,0.1386374533,0.2650431991,0.0291096382,0.0644889772,0.4058018327,-0.5469167829,-0.129997313,0.0779849812,-0.0843900666,0.0141160209,-0.3546925485,0.0162592232,-0.151968345,-0.4868755639,0.1228972673,0.4273002148,0.1107620597,0.1560126692,0.1129691377,-0.118295446,-0.1105833575,0.2593542337,-0.1041664407,0.3105356097,0.2750077248,-0.3885119557,-0.2793481052,-0.2813337445,-0.041482117,0.5379199982,0.0729843602,-0.2579219043,-0.0823838785,0.0530899949,-0.1429593712,-0.2245258987,-0.1798430383,0.295620054,0.0567408986,-0.1748363674,0.0133334519,0.1366025358,-0.0788140669,0.0510627814,0.2558814585,-0.0252816975,0.0437959507,0.0180152804,0.3831216991,-0.0396491699,-0.059311904,0.2597792149,0.0844619572,0.3718746603,0.0490069576,-0.3023099899,-0.0909235179,0.2394493073,0.2573047578,0.0541873984,-0.0231465995,-0.1395311654,-0.2728272378,0.177087307,-0.2595816851,-0.0192697681,-0.2645175755,0.0544520989,0.2426661104,0.0951151997,-0.0836053193,0.0078834444,-0.1807267964,0.0618781038,0.2265647203,-0.3716441691,0.1437590271,-0.0756446421,-0.5033358335,-0.0518128835,0.245409444,-0.0410103239,0.0489411242,-0.1886306554,0.003123669,0.1159720048,-0.2220709324,0.3767777979,0.237682119,0.0985258073,0.0503795743,0.2114095837,-0.3266764879,-0.2567704916,-0.1279067248,0.1398043483,0.0037041297,0.303419143,-0.2448831201,-0.0859331712,0.4820078015,-0.138535738,-0.1042616293,0.0710156038,-0.3813587427,-0.3186557293,-0.0304114372,-0.0683777332,0.0783219263,0.1801552773,0.0039214725,0.0600615181,0.3741314411,-0.0926764458,-0.2901074886,0.0783734769,0.2200597525,-0.3125345707,-0.3215502799,0.0070638168,-0.1349464506,-0.2527891994,0.2348996252,0.1567543298,0.049360849,0.0944043919,0.0677431971,-0.2000508606,0.2899401784,-0.0645483956,-0.0421376675,-0.0564921536,0.3262818754,0.2556849122,0.0156845786,0.2430730611,-0.0254160538,-0.0338974558,-0.117784448,0.2917452753,0.0228211842,0.1835059375,0.2056350559,0.270979166,-0.2117458135,0.0800266042,0.2155640721,0.8698325157,0.1203985214,0.2494456619,0.4566153288,-0.1443168074,0.397970438,-0.309132576,0.173383832,-0.2170266658,0.7918404937,0.0458682366,-0.0601841174,0.2042423934,-0.3551328182,-0.1539207548,0.2807689607,0.055980023,-0.1374189109,-0.0944259837,-0.0407578535,-0.2088607997,-0.050047908,0.1101367772,0.2160217464,-0.3534607589,0.3168384433,-0.2780526876,-0.0105568254,0.0044396012,-0.2202251852,-0.3149413168,0.1178950146,-0.3058873713,0.3330855072,0.1586165279,-0.3753379583,-0.2043130398,0.1201803535,0.0023953649,0.1498411,-0.1004087403,0.3681905568,0.0705657229,-0.1395627707,-0.2048745602,0.391443789,0.1122763231,-0.177280426,-0.1457835436,0.1083117574,-0.1231200993,-0.1705342978,-0.4195892513,0.1231823713,0.2489798218,0.2318018973,0.0314113684,-0.1299745739,-0.2163278013,-0.341688782,0.2031344324,-0.1083763838,-0.0590417683,-0.1964464933,0.3437514603,-0.1848868728,-0.0158504806,0.1488685161,-0.0897963867,-0.0017373548,0.2791011035,0.0897134915,-0.2856801748,-0.3556283712,-0.2173357606,-0.1536955833,-0.3202960491,0.1819196045,-0.0684527233,-0.3601391017,-0.121163331,0.0910346434,-0.0356679223,0.2908215821,0.1632186472,-0.3125679195,-0.1938587427,0.1337806731,-0.2705552578,0.030572731,0.1145085692,0.17810978,-0.157926634,-0.0553048365,-0.3672716916,0.0183307473,-0.2803925574,-0.019764388,-0.0945991948,-0.309276849,0.0475358069,0.2122885585,0.2952652276,0.3387914598,-0.1974691898,-0.2868537903,-0.1151243076,0.0832903758,0.0768210441,-0.2846932411,0.0162924938,-0.1160239428,-0.0209358558,0.0953287408,-0.0849208087,0.0878762081,0.1053588688,0.1731076688,-0.4861341417,-0.0337018296,0.3043298721,-0.0317983814,-0.265417248,0.0546754412,-0.188723281,0.0560606569,0.0769909322,-0.0855597481,-0.3580903113,0.0578422248,0.0130077396,0.1760048717,0.2488035411,-0.1545947194,0.014413476,-0.4138321579,0.4470154345,0.1585949063,-0.3002401888,-0.0020471334,0.0883621052,0.2202453017,-0.3378839493,0.3136426508,-0.2345174402,-0.035339158,0.2800394297,0.0551910587,-0.2639222443,-0.3170891404,0.1199295446,0.3092920184,0.4165694714,-0.5629336834,0.0328441747,-0.2319966704,-0.109667547,0.1346082985,0.3232007921,0.1284903288,0.1752014905,0.1240878329,-0.2571130693,-0.040596731,-0.3657486737,0.019105928,-0.0160914548,-0.0366423316,-0.091012843,0.1378771067,-0.2509682178,0.3228325248,0.0818235725,0.7464497089,-0.2874426246,-0.0339757092,-0.1460583061,0.1454366744,-0.1406720877,0.1058819667,-0.4547056854,-0.180870235,-0.3080899417,0.0974007696,0.2200728506,0.1771291345,-0.238386035,0.2249790281,-0.3436040878,-0.2806710005,0.3494089544,0.0303444006,0.2118942291,-0.1712238193,0.0802692175,0.3841170371,0.062484581,-0.2722819448,0.1710564345,0.2106820792,0.234440878,-0.0908481479,-0.1804371029,-0.120175533,-0.3061048388,-0.0151191493,0.1298440993,0.5514697433,-0.209016189,-0.1516843587,0.2590768635,-0.3603489399,0.4619835615,0.3421060443,0.3487669528,0.270468533,0.171418488,-0.3577379286,-0.0914184302,-0.1237706915,-0.1109347045,-0.5832605958,-0.247385323,-0.274343878,0.1002849936,0.1159112975,-0.3735369742,0.1060738787,0.0299115106,0.1652316898,0.3347088397,-0.3359488845,-0.1154945791,-0.0728588477,-0.2500897944,0.2025323212,-0.0433936194,-0.0494602323,-0.3409715891,0.2301681787,-0.2987864316,0.0475336947,0.3991130292,0.2826303244,0.0443495512,-0.3554227352,-0.2299550921,0.1981904805,0.1117231101,-0.0601395145,0.2495933622,-0.3151139319,-0.0492017344,-0.2633419931,0.2330156565,-0.0152768325,-0.1871239543,0.0017422886,0.1091874838,0.0368754715,0.1457736939,0.0754961371,-0.1785765886,-0.2145087719,-0.2647871375,-0.3179630637,-0.118497692,-0.0743314326,-0.0863893926,0.4793865085,0.0589315481,-0.5347625017,-0.3082464039,0.3783529699,-0.1460174173,-0.2836263776,-0.1616503149,0.3121758997,0.129075706,0.0019300682,-0.191252172,0.0003066979,0.0616069622,-0.1369276643,-0.579182744,-0.4299157858,0.48207587,-0.1373570561,-0.2608008981,0.3844519258,0.2948332131,0.2673252225,0.1125455201,-0.1652177125,0.1308746785,-0.0490655974,0.1708981544,-0.1524116546,0.3338855207,-0.3131613433,-0.0769903734,0.0495462157,-0.2198229134,0.3098995984,0.0017402216,0.1381971687,-0.0770448446]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2446","title":"`yelp_polarity` is broken","comments":"Solved by updating the `nlpviewer`","body":"![image](https:\/\/user-images.githubusercontent.com\/22514219\/120828150-c4a35b00-c58e-11eb-8083-a537cee4dbb3.png)\r\n","comment_length":5,"text":"`yelp_polarity` is broken\n![image](https:\/\/user-images.githubusercontent.com\/22514219\/120828150-c4a35b00-c58e-11eb-8083-a537cee4dbb3.png)\r\n\nSolved by updating the `nlpviewer`","embeddings":[-0.0200818982,-0.3092260063,-0.0927101597,-0.0165109169,0.1731459647,-0.2567180395,0.2033396363,0.1944490969,0.188550204,0.0823514089,-0.1388387978,0.2763346732,-0.0382111184,0.0415191799,-0.1053624302,-0.0491201505,-0.0072587072,0.4544561207,-0.2256203443,-0.1107593998,-0.21920982,0.3271355331,-0.0736514702,0.1060536057,-0.0355644636,0.0649530888,0.1932220757,0.15152587,-0.1450338811,-0.2897987366,0.0579925478,0.067933403,-0.2745875716,-0.0100898845,-0.0001070358,0.0232599992,0.3882442713,0.017305389,0.109052591,-0.4431198537,0.037238989,-0.0582427904,0.258646071,0.0051352815,-0.2671557367,-0.2403930128,0.0414380431,0.1970580816,0.3147628903,0.0871280432,0.3253395557,0.0346961319,0.0119829094,0.0570048019,0.1987484992,0.2725264132,0.0726500973,-0.1167755127,-0.0811701044,-0.3973815441,-0.178090632,0.2237029523,0.2871476412,-0.0582279041,0.0049044858,-0.129035309,0.7550250888,0.1408989578,0.0614678897,0.0642508194,-0.27306813,0.0998626649,0.1676183343,0.1488410383,-0.1109473482,-0.2310122252,0.1897274703,0.2872374356,-0.2031494081,-0.2342985868,-0.115369767,-0.0783918202,-0.1941447407,0.2688858807,-0.1025843546,0.778917551,0.0719313994,0.1676601917,0.1010914594,-0.0464041308,0.2021995783,0.4712458253,-0.0873986259,0.1085754111,-0.0746714994,0.001990343,0.2885327637,0.2024378628,-0.0905223489,0.2338658124,0.26006639,0.257422179,0.1163431033,0.1456009448,0.1903345436,0.0999985486,0.1739029288,0.4170952439,0.3002719283,0.4679611921,0.2501412928,0.1404837817,0.1760880649,-0.3430702686,-0.0997207463,-0.007316127,0.1020072326,0.0086457068,-0.2074843198,-0.05621152,-0.1812451631,0.2166467905,0.2033608854,0.1748642772,0.0977701843,0.3714202344,0.2258160859,0.0097504677,-0.1164817214,-0.0368492976,-0.1897538155,0.2553676367,-0.3868358731,-0.2213166207,0.1604045928,0.4252347648,0.4636264145,-0.0275675766,-0.1850759685,-0.0262593701,-0.1781490445,0.0726601258,0.0693980828,0.0791009143,0.0437751897,-0.1058819965,0.0265937243,-0.1485952139,0.0259931572,-0.0102660442,0.0091785137,0.2061223388,-0.1442278773,0.3222893476,-0.0389256105,0.0586477444,0.1924082786,0.2003264576,0.1342159212,-0.1940461248,0.2262342423,-0.247060895,-0.0357045233,0.038808696,-0.0308365449,0.1178062931,-0.1101895645,-0.4893554449,-0.2270245403,-0.3536872864,0.1657245755,0.2059611529,0.0378041863,-0.1349884719,-0.0917157754,0.0787403136,0.2347389311,-0.2136191279,-0.2049319595,-0.2767326534,-0.2984164357,-0.1122439429,-0.1894598007,0.1737445444,-0.0420446657,0.1007316262,0.1157315895,0.5945643783,0.2130706906,0.1902569532,-0.4638769031,0.0407055728,0.2497034073,0.0705691352,0.224012062,-0.0460263938,0.0618045032,0.0683620349,0.2615686655,-0.0133235864,0.2353276163,0.0625434369,0.677929759,0.021526495,0.091592893,0.0574843958,0.0287610944,-0.1729325205,-0.8883360028,0.2984176278,-0.0512850806,0.1443578154,-0.1169694141,-0.0677409098,-0.3330611289,-0.2815824151,0.314378649,0.0875622705,0.0527269132,0.2313656658,0.1575228721,0.1249063686,-0.2888310254,0.0463768095,0.144494921,0.2108756751,-0.1540300995,0.0567599721,0.0457330272,0.1816065013,0.2367818803,-0.072748661,-0.0736611113,0.2421265244,-0.4045485854,0.3722610474,-0.1951992065,-0.3755057752,0.2626803219,-0.5422523618,-0.2465536743,0.0897424147,-0.0577510968,-0.0375571586,-0.0677623749,0.2763516605,0.0253054537,0.212168321,0.095820941,0.2843980789,0.249944225,-0.0623834655,-0.1570227742,-0.3592318594,0.3760054111,-0.0804922432,-0.0223022774,-0.2255315781,-0.144862324,0.1679353416,0.1984559149,0.3187920153,0.198335126,0.0090646436,0.0709791556,-0.2147693485,0.1156595349,0.2192702591,0.0622571446,0.1234699488,-0.2508196235,0.1286441386,-0.2909008265,-0.2301950306,0.1856511682,0.1329302341,0.2708942592,0.2456829697,0.381659776,-0.1427160352,-0.3955886066,0.0326165669,-0.1752175391,0.1688720286,-0.0555596203,-0.1384469718,-0.2835014462,-0.1190363318,0.0388611071,0.078191936,-0.0460603312,-0.283151567,0.3188844025,-0.0819779187,-0.2211127132,0.4174658954,0.2395003736,0.2702545226,0.1054818556,0.0244916454,-0.1846250892,-0.2566592097,-0.0224624593,0.1920325011,-0.1158085093,-0.0128004905,0.321295023,-0.1067401171,0.1326783597,-0.2953078449,-0.5893121958,0.0916894674,-0.1029200107,-0.0791422725,0.1535940468,0.0623051897,-0.036481373,0.1444422752,0.3220663667,-0.4520685673,-0.1642165333,0.1549222469,-0.0910736695,-0.0623714291,-0.3400526643,0.0223829951,0.0367963538,-0.3645442128,0.3099125028,0.4184516668,0.0638239235,-0.0068374318,0.0409347378,-0.126821965,-0.2619682848,0.0365581587,0.0037262817,0.3498035371,0.2677513957,-0.4467646182,-0.3529115617,-0.188878879,0.0340122059,0.4241081178,0.0502626933,-0.2277788967,-0.1123092324,0.103130877,-0.2093066871,-0.0188954603,-0.2831341624,0.2530283928,0.1051916704,-0.219196409,-0.012552876,-0.0211472344,0.0971134976,0.05914554,0.3043213785,-0.0697439089,-0.0331053138,-0.0393274166,0.3158397079,-0.0860571787,-0.0120360777,0.2388045192,0.2003787905,0.3195843101,0.0738105029,-0.1181000695,-0.1296387166,0.3068482578,0.2798942924,0.2602995336,-0.0483552068,-0.3579597771,-0.2607403696,0.041555956,-0.2103601843,-0.1106554791,-0.2085867524,0.0812724829,0.2035516053,0.136609897,-0.0709268451,0.0742768943,-0.2543475032,0.0337662548,-0.0133475149,-0.342677325,0.029824432,-0.098922886,-0.521451056,-0.0075365119,0.0473340005,-0.0687209591,0.1554630846,-0.1789398789,-0.2665852308,0.1628742665,-0.2347321212,0.2277695686,0.3524974883,0.1054366007,-0.0192443151,0.0768569559,-0.4516458213,-0.2077907324,-0.1312174052,0.2473808974,-0.0425070375,0.2324923426,-0.1755375117,0.0141453659,0.2323556095,-0.3426977992,-0.1371633112,-0.0206899121,-0.1941460371,-0.1271475255,0.0283274259,-0.0431186445,0.0158810448,0.1240056008,-0.0396592468,0.0381974578,0.4081473351,-0.0823418498,-0.3001587391,0.1972868294,0.2335483432,-0.2599794567,-0.228782773,-0.0707847551,-0.0888416395,-0.2895107865,0.1573660076,0.3053196073,0.0030012431,0.1080954075,-0.0789904743,-0.1052453741,0.337523371,0.0363539085,0.0781832635,-0.1280150712,0.3300984502,0.2345588803,0.0359132215,0.1444364339,-0.0906651393,-0.0222569183,0.0697062835,0.3057299852,-0.1483672559,0.0444242135,0.0578223392,0.4494871497,-0.2509330213,0.1029473469,0.2669774294,0.7822646499,0.1600007564,0.0926848575,0.3260193169,-0.1732465774,0.5052621961,-0.2997729182,0.1832494289,-0.3620665967,0.9499100447,0.0569016822,0.0320004709,0.1335609257,-0.225377351,-0.0924967453,0.0094168922,0.1593202949,-0.2640455961,0.0546731241,-0.0374100506,-0.0501525663,-0.0293093715,0.169143647,0.2675110102,-0.3634817898,0.1325007528,-0.2678315341,0.0628447756,-0.0955693498,-0.1131007075,-0.252676934,0.1321297586,-0.1209516525,0.2468245476,0.2633874118,-0.2131027281,-0.0690877214,-0.0849412531,-0.0031243914,0.0816496238,-0.2247779816,0.3561171293,0.115698196,-0.1760397702,-0.1527584791,0.3807429075,0.0228380729,-0.1816585362,-0.1655492485,0.0210999958,-0.0836729705,-0.0963075012,-0.5465927124,0.1378605515,0.0690792799,0.2713038027,0.0388841294,-0.0780771449,-0.1424950361,-0.3346582651,0.2710796893,-0.0664152056,-0.05739443,-0.2225224376,0.5471709371,-0.1647915095,-0.0331501476,0.2144970745,-0.1301083714,0.027893506,0.264736712,0.1652011573,-0.2885868549,-0.3430823088,-0.2837084234,-0.0956492126,-0.408922106,0.1014949381,-0.0390748866,-0.3212599158,-0.0720475093,0.2143044174,-0.034248177,0.2311059386,0.0391447172,-0.2264294028,-0.0825291425,0.057040181,-0.2443787009,0.0108022941,0.0803782567,0.1283199638,0.0262418799,-0.1152898446,-0.3984130919,-0.0830629617,-0.3307058513,0.0685805604,-0.1936058402,-0.2884078026,-0.1079270467,0.102888599,0.2792747021,0.2811260223,-0.0980037674,-0.3550280631,-0.1024267524,0.0551861711,0.2759084702,-0.219064638,-0.108997494,0.1278333515,0.0371065475,0.2619271874,-0.1481955051,-0.016224578,0.0175633039,0.1670730561,-0.3427129388,-0.225209415,0.3167023361,-0.0172980186,-0.0822359994,-0.0176398158,-0.1812302172,0.0043900944,0.1167252138,-0.0324148312,-0.1234590113,0.0926194638,-0.0461049452,0.2257238775,0.1325053573,-0.1844209135,0.1092274413,-0.2747400701,0.3620736897,0.1522572786,-0.2161072791,-0.1568992883,0.2469806969,0.2622681856,-0.3504869044,0.3680068552,-0.0324307196,0.0234573465,0.2466838956,-0.0222439915,-0.1736797094,-0.3240403831,0.0669545755,0.3464282453,0.3817280531,-0.5726109147,0.0305570345,-0.2217250019,-0.1883001029,0.1725296974,0.2170787454,0.1374454647,0.1261212081,0.1478134245,-0.082858257,-0.0359599367,-0.4394110441,0.0840622112,0.1757773161,0.0383011997,-0.0352536924,0.1078011766,-0.2800416946,0.2165806293,0.1913434565,0.8253935575,-0.1877599806,-0.0048049274,-0.2134243399,0.2347618341,-0.1986558735,0.058105737,-0.4606626332,-0.2585514486,-0.2801048458,0.1136169657,0.1704025716,0.2026474327,-0.3002609015,0.1952019632,-0.3547443151,-0.3089993,0.2250457406,-0.0703439191,0.2338573039,-0.0808070898,-0.0180875473,0.218139872,-0.0481739827,-0.0647863075,0.1013072059,-0.0073062903,0.2569058537,-0.1707378179,-0.2057699114,-0.2066798806,-0.2118935287,-0.0027709482,0.171556592,0.6098131537,-0.214419663,-0.0459670052,0.2654424906,-0.3973271549,0.5642440319,0.3053239584,0.3309663832,0.318589896,0.1330789775,-0.2780371606,-0.0141000226,-0.1505184025,-0.2713238001,-0.495998174,-0.3452085555,-0.3117900193,0.0106366407,0.0946223512,-0.5409673452,0.1311484575,0.2353310585,0.2524316311,0.2251801044,-0.3391823471,0.0822603554,-0.2245000452,-0.2443832308,0.2280924618,-0.0837521031,0.024975894,-0.2951916158,0.2647166252,-0.2166949958,0.0746132955,0.2797321081,0.4450488985,-0.1457368433,-0.3859178126,-0.205615133,0.3305384219,0.1010753363,-0.1430072784,0.2130230814,-0.0945286378,0.1163291857,-0.1304947436,0.2461233139,-0.0221777596,-0.0350047424,-0.0714496449,0.1733003259,-0.0027398248,0.1250875145,-0.0172777511,-0.0830609426,-0.1294529885,-0.2506172359,-0.3536204994,-0.1870965362,-0.111001797,-0.1663085967,0.407169193,0.1586851925,-0.5291237831,-0.1185790598,0.1325595677,-0.2040678114,-0.1222427934,-0.1220230609,0.2514171004,0.0857350454,-0.0913749784,-0.1502857655,-0.1771609485,0.0970258936,-0.1843605042,-0.4401524663,-0.4739480317,0.2574661076,-0.2076841593,-0.2380717099,0.3738728166,0.1370832771,0.0966346934,-0.0568771213,-0.2498506606,0.0022528486,-0.0422136113,0.1426769048,-0.0898225084,0.1735610068,-0.3458985984,-0.1576354951,-0.0890249014,-0.1205528751,0.3444818854,0.0575490929,0.1410998702,-0.0877125487]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2444","title":"Sentence Boundaries missing in Dataset: xtreme \/ udpos","comments":"Hi,\r\n\r\nThis is a known issue. More info on this issue can be found in #2061. If you are looking for an open-source contribution, there are step-by-step instructions in the linked issue that you can follow to fix it.","body":"I was browsing through annotation guidelines, as suggested by the datasets introduction.\r\n\r\nThe guidlines saids \"There must be exactly one blank line after every sentence, including the last sentence in the file. Empty sentences are not allowed.\" in the [Sentence Boundaries and Comments section](https:\/\/universaldependencies.org\/format.html#sentence-boundaries-and-comments)\r\n\r\nBut the sentence boundaries seems not to be represented by huggingface datasets features well. I found out that multiple sentence are concatenated together as a 1D array, without any delimiter.\r\n\r\nPAN-x, which is another token classification subset from xtreme do represent the sentence boundary using a 2D array.\r\n\r\nYou may compare in PAN-x.en and udpos.English in the explorer:\r\n https:\/\/huggingface.co\/datasets\/viewer\/?dataset=xtreme","comment_length":39,"text":"Sentence Boundaries missing in Dataset: xtreme \/ udpos\nI was browsing through annotation guidelines, as suggested by the datasets introduction.\r\n\r\nThe guidlines saids \"There must be exactly one blank line after every sentence, including the last sentence in the file. Empty sentences are not allowed.\" in the [Sentence Boundaries and Comments section](https:\/\/universaldependencies.org\/format.html#sentence-boundaries-and-comments)\r\n\r\nBut the sentence boundaries seems not to be represented by huggingface datasets features well. I found out that multiple sentence are concatenated together as a 1D array, without any delimiter.\r\n\r\nPAN-x, which is another token classification subset from xtreme do represent the sentence boundary using a 2D array.\r\n\r\nYou may compare in PAN-x.en and udpos.English in the explorer:\r\n https:\/\/huggingface.co\/datasets\/viewer\/?dataset=xtreme\nHi,\r\n\r\nThis is a known issue. More info on this issue can be found in #2061. If you are looking for an open-source contribution, there are step-by-step instructions in the linked issue that you can follow to fix it.","embeddings":[0.2943785489,-0.3165698946,-0.0003395347,-0.0429260321,0.2085690349,-0.2181834131,0.2861591578,-0.2059640586,-0.0860845149,0.0388182215,-0.1333658993,-0.032144051,0.0991284773,0.0123276971,-0.0314271636,-0.37335971,0.0662141219,-0.0248676892,-0.0890788957,-0.2008087337,0.0172808338,0.19551301,-0.2496371567,-0.1431765407,-0.9076256156,0.082934469,-0.441617012,-0.1027689502,-0.2490838319,-0.2852181792,0.4068395495,0.2977807522,0.0146640893,0.1980400383,-0.000119285,-0.2438721806,0.0082747815,-0.2250849456,0.1475875229,-0.116915077,-0.0043435404,-0.3055847287,0.1942625791,-0.0991955325,-0.1677924693,0.3073951304,-0.0375575386,-0.2729660273,0.3280765414,0.2558254004,0.047795143,0.0436938405,0.1891580224,-0.1726178229,0.2665705681,0.109638311,0.0460626557,-0.1800417304,0.3944066167,-0.0963980556,-0.1886882335,0.0528421216,-0.0747718737,-0.05295467,0.0326096527,-0.0139410626,0.0553775951,-0.3024042547,-0.0813537911,0.351790309,0.5863409638,-0.1634722799,-0.4324625731,-0.2785465419,0.1534543782,-0.0422806628,0.0232623424,0.1487395912,0.0191482678,0.1376584768,-0.2143916488,-0.007037167,-0.2035733759,0.1972466856,-0.4043771923,0.5016332865,-0.1200178415,-0.2557636201,-0.1704626083,0.0702378452,0.3050934374,0.1517990977,-0.0597095229,-0.0428074785,-0.1932374686,0.0560365841,-0.0593727157,-0.2394413948,0.270050317,0.1130454764,-0.0215370879,-0.0920686871,-0.028517304,0.1502853334,0.3564156592,0.0459969155,0.4407730997,-0.1441407949,0.1073602065,0.1852706373,0.0624562651,-0.1162273884,-0.204515636,0.0379617959,0.1711208224,0.027792763,0.0768402442,-0.2226639986,-0.1109056324,0.0910214558,-0.1715718657,0.2357388437,-0.0843475163,0.1162339896,0.1526813209,0.3613802791,-0.1246704087,0.1369122714,-0.0593390316,-0.5313051343,-0.1652615368,-0.0679921582,-0.1691628546,0.1295256913,0.0916913971,0.2039807737,0.4520845711,-0.0667562038,0.2060504407,-0.3830917478,-0.2101044953,-0.1992096305,0.2979212701,0.2983342707,0.090955697,-0.0935404822,0.0057702451,0.0271281283,0.1800231785,0.115302287,-0.0377768166,0.1452331543,-0.217824921,0.0280968975,-0.4752112627,-0.1052292064,0.0204412527,0.1591146439,0.195896253,0.2108123153,0.3726465404,0.0334666967,-0.1673221141,0.0539823137,0.2322040051,0.5055274963,-0.4218670726,-0.2860972881,0.0067024492,0.1054777727,0.1910517067,0.1601894349,0.2430484444,0.1763904244,-0.2068833113,0.4748584032,-0.1279444695,-0.1247969791,-0.1877070367,0.0819592029,0.1985340565,0.1063800827,0.0142898057,-0.3189462125,0.7664393783,0.2003558278,0.1362506151,-0.2044346035,0.3295777738,-0.0468569435,-0.0303291138,0.055160325,-0.2202139497,-0.0789166316,-0.0859380215,-0.1518075317,0.1531657428,0.2992606163,0.2501058877,0.0424009971,0.2810055614,0.451996088,0.2274077386,0.3054513335,-0.052371487,-0.3569984436,-0.3652436733,-0.1456579268,0.3876273036,0.202541858,0.2256770581,0.0326027051,-0.0716571584,-0.0734834895,-0.1509144157,-0.4594607651,0.0537529774,0.1306964904,-0.1917301565,0.0219784603,-0.2497675717,0.1522014141,-0.0449553281,0.3213044107,0.0248551648,0.3513791859,0.2706263661,-0.4771026671,0.1784388721,0.587245822,-0.079562895,0.1270792335,-0.0334462412,0.4789627492,0.4988962412,-0.1589893252,-0.1316779256,-0.106226854,0.3002455831,-0.5324885845,-0.0362059735,-0.1594479084,-0.1272919029,-0.065529421,-0.3021337092,0.2023843378,0.3503062427,-0.0740159974,-0.0771419629,0.2796509266,0.0286558308,-0.0842890665,-0.4774984419,-0.566264689,0.2487124801,-0.1395751983,0.1832011938,0.2994828522,-0.2072235942,-0.0993355066,0.2111136764,-0.0500906594,0.1915109754,0.1477712691,-0.1585731953,0.049129989,0.2867007852,-0.0171468724,0.1938892156,0.4177370369,0.1222360879,-0.0983990878,-0.1228454411,-0.148239091,-0.0460715294,0.2629246414,-0.1783530861,0.0441677347,0.1996251792,0.152828455,-0.3149123788,0.0574433394,-0.1866769195,-0.0235043392,-0.3852814734,-0.4310305119,-0.2788581848,-0.6325992942,0.2539135814,0.0348162912,-0.4936919212,-0.2842388749,-0.0321608782,-0.3742286563,0.0344480351,0.0170961041,0.0484232344,0.2826184928,0.1043350399,0.483754009,-0.0180591848,-0.2675775588,-0.0042579747,-0.0576888248,-0.1463287473,0.3795585036,-0.0502777845,0.0338556208,0.1338170469,-0.2877043784,-0.2033989429,0.1954875141,0.0709628016,0.3235013187,0.4326769412,-0.0626671612,-0.4241930544,0.3033116758,0.1547109038,0.1255280524,-0.2496196479,-0.0930201635,0.0321451537,-0.0865748003,-0.1464926749,0.0536402538,0.0980486721,-0.2061930895,0.1678492129,-0.1835826933,-0.0553531796,-0.1754871458,-0.188967526,0.2563677728,-0.4381485879,-0.0193597339,0.0108618336,-0.8254088759,0.3866718709,-0.1991353929,-0.3030286729,-0.0154322237,0.0977002978,0.2604309618,-0.1815440208,-0.2502306402,-0.2417726517,-0.3236402273,0.2169124633,0.0544666387,-0.1011909172,0.1095847338,-0.0515855327,0.0892269015,-0.2553340793,-0.1623624414,0.4873225391,0.1539411992,0.1689654887,-0.2814294398,0.1701750904,0.2800267637,-0.3305484354,0.4690202475,0.2252591848,0.4293317497,0.1059248075,0.3980954289,-0.1127362475,-0.0663272068,0.1267487556,-0.1707401723,0.2599265277,0.3350878656,0.2443295866,0.3813293576,0.0683961362,-0.1298306435,-0.2731616795,-0.21265091,-0.1859507114,-0.5198609829,0.1279313862,-0.2289843112,-0.1323072761,0.048058413,-0.046577055,-0.1859839708,0.223239705,0.5266064405,0.1856376827,-0.3416755497,0.6310095191,-0.1177590042,-0.0751292109,0.2515766919,0.1422967911,-0.1750202924,0.0762990788,-0.0102010798,0.4444431663,0.4584877491,-0.1506924927,-0.3880316317,0.0155880209,-0.071754314,-0.0491618253,-0.0222768541,-0.1248715371,-0.4263614416,0.6463209391,0.2460529804,-0.1870411634,-0.0923993438,-0.0211071968,0.0353135206,0.0441335775,-0.2755171061,-0.0964139253,0.1326760203,-0.1947399676,-0.0805144086,0.3271882832,-0.0477222316,0.1682492644,-0.4310177267,0.1979884803,-0.4687083662,0.0631610081,-0.0286277104,0.286227107,-0.5002985597,-0.0274898298,0.3533815742,0.0308714695,0.4632567763,0.0392195508,-0.3126153648,-0.3794095516,0.3603869081,-0.6718878746,0.9012407064,0.0327450372,0.3511559367,0.1237362921,-0.1261217296,0.0960148051,0.1673152894,0.1320742518,0.1645965874,0.2219867259,-0.0047335755,-0.1285553575,-0.2284502685,0.0097481422,0.2290443182,-0.1419848651,0.3926765323,-0.209992066,0.1256396174,-0.0053798333,0.8337066174,-0.1103570089,0.2992193103,-0.2163316607,0.0626733527,0.5165963769,0.2686865032,-0.0392163545,0.2749159932,0.0224694293,-0.2080705911,-0.0493425168,0.274428159,0.1604605764,-0.3019710183,0.3058300316,-0.0237956382,-0.0782111436,0.1890056729,-0.3220506608,0.1228764132,-0.2821014225,-0.1451452076,-0.0302128568,-0.0742968395,0.1339921355,0.0108509269,-0.0462072454,0.0139514916,0.0954719186,-0.4640967548,-0.0705145895,-0.331846267,-0.1265804917,0.3696522713,-0.0086720483,-0.0765327662,0.4826587737,0.2945119739,0.2009195238,-0.0621028952,0.2080149353,0.537925601,-0.1614275724,-0.0231218785,0.1985255331,0.2764620483,-0.2877350748,-0.2914534211,0.0316628516,-0.1044870615,-0.123370178,-0.1003988683,-0.194570154,0.0574098751,-0.0272033438,0.0505013987,-0.1697612852,0.2405719608,-0.3615009189,0.0475933552,0.2042313814,-0.0285423994,-0.3973739445,0.0763703138,-0.145333156,-0.2536848485,0.2905171812,-0.3669510484,-0.0524050742,0.2908208072,0.4180589914,-0.0232693683,-0.2249288857,0.2397970855,0.2150541395,-0.4241418242,-0.0069755577,-0.2416140139,-0.0111032575,0.0034772325,0.356651634,0.3360308409,-0.0411199145,-0.0921993256,-0.2644581199,-0.2211116552,-0.0745719224,-0.2075251937,0.0130849769,-0.06447009,0.5586349964,0.3201162815,-0.036625877,-0.1439365,0.062555775,-0.2379827946,-0.1734874547,0.1556695551,-0.2263437212,-0.2235224992,-0.1376832724,-0.0068528987,0.3970900476,0.4999190867,-0.0784099475,-0.1922360957,0.1174255386,0.0035970572,-0.0362477265,0.1459848881,0.5382344723,-0.2701543868,-0.1230417415,0.1378374398,0.2036233693,0.0756100714,0.1761165112,0.1619105488,0.0741148368,-0.2448042482,-0.121069476,0.2436768711,-0.197334215,-0.0061692381,-0.0227317419,0.0637619346,0.1073124856,0.1678414345,-0.1087199524,0.033780586,-0.045387622,-0.0774209425,0.0730628595,-0.1050178334,0.0388193354,0.168424651,-0.0884607807,-0.038513869,-0.1398370266,0.2101245224,0.1170870736,-0.2189162225,0.1035358086,0.0641263649,0.3692781925,0.0166753884,0.0967761278,0.1667140126,-0.0169994459,-0.149147138,0.0196671765,-0.0195958894,0.0630777329,0.2611026764,0.4098836482,0.2130204439,0.0644922554,0.0811727047,0.0015834764,0.2016781569,0.2348062098,0.3369478583,0.4751454294,0.1061124653,0.3609163463,0.2039249539,-0.25652349,-0.1137657911,0.0190850534,-0.1151513681,-0.0528184436,-0.4103647768,0.5589283109,0.1277994961,-0.1880593151,-0.2263303846,0.3180320859,-0.191934064,-0.0837650299,-0.1091188192,0.0113220261,-0.079680182,-0.0715601742,0.0832062066,0.0945593938,-0.0004828028,-0.2478149831,0.0285852086,-0.417579025,-0.2683475316,-0.207250461,0.2011404634,-0.174280867,0.4245044887,0.3192593455,-0.40639925,0.0081334952,0.4427731335,0.302146554,0.0902664661,0.4132663608,-0.0982320458,0.047594294,-0.1608252823,0.1377708912,0.2472910583,0.2786760926,0.1676155776,0.2435364574,0.0586561561,-0.0141701875,0.2170735598,-0.1283196211,0.0881886706,-0.3246140182,0.1093244255,-0.0298856385,-0.1438878179,0.0332684517,-0.081102401,-0.4142846167,-0.0020252597,0.1221646369,-0.1779777408,-0.2098808289,-0.1173431128,-0.0173731837,-0.0454543605,0.162369132,0.5658299327,-0.005157189,0.2236793637,-0.4870315194,-0.4129174948,-0.0991768911,0.019333981,-0.0554684624,0.2044575959,-0.1164033189,-0.0141880903,-0.0705078766,-0.243666634,0.2037537247,-0.3597389162,-0.195864439,-0.6218894124,-0.0008837344,0.39712587,0.0003161494,-0.0462668724,0.4576946497,-0.064816907,0.248202756,-0.118897818,0.3025880754,0.2950904369,-0.0113145309,0.4382409155,-0.4488315582,0.0804611444,0.3396882713,0.0221409332,0.0857643634,-0.2124494016,-0.0740529895,0.0467416234,0.3410715759,0.1519690901,0.3718175292,-0.1551256329,-0.0752488822,0.1088762432,0.1650443077,0.2037392557,-0.2129629999,-0.225045085,0.082672134,-0.0612428673,-0.2613786161,0.1654132009,-0.2865150571,-0.1310978085,-0.1278981417,0.0536397584,-0.2311439216,-0.0159513075,-0.1312959045,0.0444015935,-0.4656938314,-0.0499694794,0.6577653885,-0.1784222573,-0.898114562,-0.1978029013,0.4490499198,0.1846192926,-0.4968597591,0.3874936998,-0.0522783734,-0.0359413698,-0.042872481,0.6013662815,0.1980889142,-0.1977722496,0.1258393973,-0.0992822647]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2444","title":"Sentence Boundaries missing in Dataset: xtreme \/ udpos","comments":"Closed by #2466.","body":"I was browsing through annotation guidelines, as suggested by the datasets introduction.\r\n\r\nThe guidlines saids \"There must be exactly one blank line after every sentence, including the last sentence in the file. Empty sentences are not allowed.\" in the [Sentence Boundaries and Comments section](https:\/\/universaldependencies.org\/format.html#sentence-boundaries-and-comments)\r\n\r\nBut the sentence boundaries seems not to be represented by huggingface datasets features well. I found out that multiple sentence are concatenated together as a 1D array, without any delimiter.\r\n\r\nPAN-x, which is another token classification subset from xtreme do represent the sentence boundary using a 2D array.\r\n\r\nYou may compare in PAN-x.en and udpos.English in the explorer:\r\n https:\/\/huggingface.co\/datasets\/viewer\/?dataset=xtreme","comment_length":3,"text":"Sentence Boundaries missing in Dataset: xtreme \/ udpos\nI was browsing through annotation guidelines, as suggested by the datasets introduction.\r\n\r\nThe guidlines saids \"There must be exactly one blank line after every sentence, including the last sentence in the file. Empty sentences are not allowed.\" in the [Sentence Boundaries and Comments section](https:\/\/universaldependencies.org\/format.html#sentence-boundaries-and-comments)\r\n\r\nBut the sentence boundaries seems not to be represented by huggingface datasets features well. I found out that multiple sentence are concatenated together as a 1D array, without any delimiter.\r\n\r\nPAN-x, which is another token classification subset from xtreme do represent the sentence boundary using a 2D array.\r\n\r\nYou may compare in PAN-x.en and udpos.English in the explorer:\r\n https:\/\/huggingface.co\/datasets\/viewer\/?dataset=xtreme\nClosed by #2466.","embeddings":[0.2949484587,-0.3351140022,-0.0134666478,-0.0679348484,0.1731439829,-0.1927722394,0.3030518591,-0.226330474,-0.0803332105,0.0512380786,-0.1084665507,-0.0391737781,0.1022449806,-0.0219379924,0.0005638007,-0.3351674676,0.0731719732,-0.0372602977,-0.0477517098,-0.2020673603,0.0344883651,0.17796579,-0.2859562635,-0.1534460187,-0.9007300735,0.0959525853,-0.4002174735,-0.1333491653,-0.2582620084,-0.295188427,0.398598671,0.3141582608,-0.0170702562,0.2185044289,-0.0001224476,-0.2265435606,0.0145918215,-0.2281181067,0.1109113619,-0.1040380821,-0.0174459461,-0.2937981486,0.1903034896,-0.1461071074,-0.1512830108,0.3339186609,0.0091063511,-0.2913050354,0.3058179915,0.2599942684,0.0229477976,0.0592507869,0.1262426525,-0.1703512073,0.2642913461,0.101960279,0.0327060632,-0.1890524626,0.3610964119,-0.0827611163,-0.209859252,0.0287117027,-0.0829826146,-0.0130860768,0.0537852198,0.0088099279,0.0472520962,-0.2880236804,-0.0750967115,0.3958187997,0.5908740759,-0.1768332422,-0.4533177018,-0.2913461626,0.1220475957,-0.015474217,0.0151322596,0.1583005637,0.0085145915,0.1556455642,-0.2019555569,0.0109278122,-0.1993221492,0.163648203,-0.4600059986,0.5282253623,-0.1202925891,-0.2238467932,-0.1669298857,0.0928838998,0.2762417197,0.1289450228,-0.0383535735,-0.0398723446,-0.1794800162,0.0415616408,-0.0318336897,-0.2355515808,0.2450718433,0.116126515,0.0296103638,-0.0895684808,-0.0589380488,0.1809388399,0.3392975628,0.007879803,0.4522628784,-0.1467561722,0.1152355149,0.1737014651,0.0611209981,-0.0989195257,-0.2122358978,0.0589465834,0.1882738769,-0.0070216418,0.1191392317,-0.2255941331,-0.1789055169,0.1126533076,-0.1812348217,0.1906188726,-0.0838791505,0.1430996358,0.1698950678,0.3244161606,-0.1540812701,0.1313964725,-0.0441910848,-0.5134134293,-0.1446290314,-0.0683530644,-0.174174726,0.1261108071,0.094884485,0.2156309783,0.4674039483,-0.0525438823,0.2115533203,-0.3536052108,-0.132035315,-0.2284722179,0.374673903,0.3108576536,0.0832986012,-0.076024875,0.0167401973,0.0226558316,0.1304598004,0.1217616871,0.0052758893,0.1329207718,-0.1833054274,0.0175353419,-0.4835573733,-0.1085759774,0.036382582,0.1491630375,0.163135469,0.2021386027,0.4457437992,0.047601942,-0.1371382177,0.0493335761,0.2027596086,0.4924277663,-0.4465982914,-0.3137145936,0.0162282381,0.0982916877,0.1764440686,0.1589629948,0.2555944622,0.1723021418,-0.2307984233,0.4900501668,-0.0650669783,-0.1523386091,-0.1730719507,0.086023435,0.1804276854,0.0953027457,0.0593715161,-0.2618142962,0.736934185,0.1855226755,0.1176813021,-0.2649475336,0.3506861925,-0.0663577095,-0.0217641108,0.0795650631,-0.1824647635,-0.0833772421,-0.088702485,-0.1811772287,0.1596080065,0.3610343039,0.2759050727,0.0734633058,0.2351198345,0.4387180209,0.2117446661,0.2715566456,-0.0658807009,-0.3674932718,-0.3757318556,-0.1161547452,0.4687045813,0.1929935664,0.3049728274,-0.0220520198,-0.0819522515,-0.0796199962,-0.164127633,-0.4644187987,0.0400416404,0.1349746883,-0.2204728872,0.0077829701,-0.3027048111,0.1133687198,-0.0693737566,0.3264012337,0.0194281861,0.3171497583,0.2832654715,-0.4360127151,0.128639251,0.5544006824,-0.1191734895,0.1277284473,0.0330133438,0.4666340351,0.5460426211,-0.1823410839,-0.1313551515,-0.0590378121,0.275970459,-0.5171937943,-0.0266112369,-0.1573470682,-0.1377404928,-0.0421284288,-0.3745609522,0.2157206833,0.3822157681,-0.0775442272,-0.0522261709,0.2184159607,0.0064443522,-0.0860703513,-0.5235785842,-0.5783962607,0.2430794537,-0.0986931995,0.2105854899,0.3344167769,-0.2402667701,-0.0969232395,0.2293501496,-0.0740955845,0.2077848613,0.1200115532,-0.1365498006,0.0143128624,0.2665316164,-0.0306115709,0.1988758445,0.4013659656,0.1923032254,-0.0974015519,-0.170914501,-0.1667033732,-0.0404013209,0.2519716322,-0.2160396278,0.0096167699,0.2180336863,0.1873250902,-0.3134056926,0.0085278898,-0.1966118962,-0.0441438071,-0.3809345365,-0.4305824041,-0.2682775259,-0.6533597708,0.2658167183,0.0551551431,-0.480692625,-0.3008632362,-0.0418163165,-0.3488638997,0.0421679914,0.0111443745,-0.0608764738,0.3066629767,0.1124837324,0.4369810522,-0.0010278196,-0.2302103788,0.0127260517,-0.0682212114,-0.1947155595,0.4006639421,-0.0220092386,0.0343717299,0.113664344,-0.3178385794,-0.2207548618,0.2266160101,0.0967836156,0.2931870818,0.4458592236,-0.0568883158,-0.4022474289,0.3216414154,0.1290322095,0.1202807128,-0.2705097795,-0.0961288363,0.0187055301,-0.0669107512,-0.1449723095,0.0997761413,0.1338218898,-0.2000702769,0.1449309886,-0.1607627422,-0.0384248272,-0.1751274168,-0.2261041105,0.2575850487,-0.4379251599,-0.0408570804,0.0126226889,-0.8649773598,0.3815352619,-0.1666413993,-0.2851656675,0.0137610128,0.0952552333,0.2641241848,-0.1397159696,-0.249098435,-0.289457947,-0.3380870521,0.245077163,0.0985400826,-0.1195473224,0.1496760547,-0.1019697487,0.0999504477,-0.2420303375,-0.1930312663,0.512329042,0.1579627842,0.1272254139,-0.3088991642,0.1511734426,0.3163472116,-0.3553407788,0.4600084722,0.2168492973,0.426407516,0.0647357553,0.4434846044,-0.1216337159,-0.0248620883,0.128687039,-0.1464240849,0.2223377228,0.3041374385,0.1974419653,0.3998407423,0.0681650862,-0.0921177566,-0.2636969984,-0.181638211,-0.1979334205,-0.4967590868,0.1573645473,-0.2189117223,-0.0864854082,0.0379589908,-0.0321280099,-0.2403016984,0.2626798451,0.5323818922,0.1662474424,-0.3670920134,0.6611530185,-0.1292329282,-0.104982689,0.2424109429,0.157207191,-0.1853829622,0.0530598834,-0.0259715579,0.4200502634,0.4202244282,-0.1910874844,-0.3233763278,0.0485233553,-0.0556238852,-0.1015676335,-0.0159733314,-0.1647796482,-0.3749451637,0.6857441068,0.2511520386,-0.1794800609,-0.097354047,-0.0203646515,0.028426839,0.0514519438,-0.262291491,-0.1283293664,0.1299456358,-0.1431708932,-0.0584535338,0.3740888834,-0.0841403306,0.2283667922,-0.422023505,0.1641007066,-0.4671693742,0.085990876,0.0238511246,0.2488094568,-0.4898030758,-0.05850612,0.3728099465,0.0165721085,0.4403966069,0.0288833398,-0.3212684393,-0.3592952192,0.4034202099,-0.6743367314,0.8701492548,0.0592523962,0.3544526994,0.1591870338,-0.1253082752,0.0928011388,0.1635180563,0.1694906056,0.1608440578,0.2136859745,-0.0427441299,-0.1365336627,-0.207666114,0.0125834038,0.2093409896,-0.1559251994,0.4119896889,-0.2075376809,0.1080910414,0.0042382837,0.8450747728,-0.084182106,0.3358730078,-0.2017342746,0.1100057065,0.5272215009,0.274504602,-0.0200810768,0.2804138958,0.0070395512,-0.214809671,-0.0949205384,0.2649002373,0.1696041971,-0.310215801,0.3110075593,0.0027549579,-0.0279173441,0.1940527856,-0.305313915,0.0498739667,-0.2130707651,-0.1580016613,-0.0534158312,-0.0252677705,0.1374196708,-0.0007372499,-0.0105392952,0.002995651,0.0913239941,-0.4850497544,-0.1202642024,-0.2985579967,-0.1236405075,0.3384289443,-0.0271000322,-0.0976619869,0.4779925942,0.3175880313,0.1716344953,-0.0474314466,0.2161839604,0.4855120778,-0.1456459761,0.0160024967,0.1439149529,0.2710727155,-0.2838165164,-0.2601913214,0.0701885745,-0.1221962273,-0.1384642273,-0.042828396,-0.1759338081,0.0424866788,-0.0412500873,0.0352036655,-0.1745621264,0.2088452876,-0.3453226388,0.0246668309,0.2122803926,0.0198984072,-0.4231006801,0.0839667395,-0.1660550237,-0.2560262084,0.2840535939,-0.3832184076,-0.0081498064,0.2846981585,0.367600888,-0.0081050238,-0.2222242206,0.2707359493,0.2168028653,-0.4060319364,0.006143976,-0.2490618527,-0.0323485509,0.0151076643,0.3775051236,0.3403463662,-0.0424712338,-0.1401213706,-0.2726335227,-0.1930283755,-0.0419456735,-0.2388759702,0.0599643588,-0.0812942088,0.5570396781,0.3487193286,-0.0785155892,-0.1256166697,0.0562057681,-0.2202661783,-0.1651371866,0.151191175,-0.2035998106,-0.2146244049,-0.1689092666,-0.0230948403,0.3553434908,0.4696233273,-0.0615144111,-0.2161220163,0.1374688894,0.009275889,-0.0219067298,0.1494520307,0.5255020857,-0.2655134201,-0.1467996091,0.1363945603,0.2240489423,0.1003237218,0.1901465803,0.1617901772,0.0469199531,-0.2776301801,-0.139443934,0.2277263105,-0.1923836321,0.0272140242,0.0130479038,0.0476257242,0.0951702297,0.2131075561,-0.0719112232,0.0404739119,-0.0768711194,-0.139479503,0.052392751,-0.0825911686,-0.0013423121,0.1778900027,-0.0488502011,-0.0414169505,-0.1596706212,0.2037329823,0.1084711179,-0.2450300455,0.0881039649,0.0881145447,0.4030985832,0.048177857,0.1260286272,0.1520148516,-0.0010533268,-0.1547638476,0.0440157726,-0.0327437669,0.0636266395,0.2353346795,0.4330115914,0.2309398353,0.0950264782,0.1045528352,0.0080252495,0.231381312,0.2484096885,0.3937909603,0.5262463093,0.1331878006,0.3672992289,0.2061893791,-0.2560003698,-0.1364704669,0.0473044626,-0.1210114211,-0.043092709,-0.3516905606,0.4767515659,0.1649041176,-0.1247643605,-0.2297073454,0.2538859844,-0.1842840612,-0.1200203747,-0.0997581035,0.0054743388,-0.0500699021,-0.064647831,0.051017344,0.0828431249,-0.0129512437,-0.2733073533,0.0872878656,-0.4021320939,-0.1928845793,-0.2275559604,0.2035225183,-0.1907956153,0.4024657607,0.3099822998,-0.3884496987,0.0381318592,0.4578489959,0.2952003777,0.0532072373,0.4370012283,-0.0673219562,0.0305750594,-0.1920578927,0.0904178768,0.2080383003,0.2877852321,0.1638086736,0.2774435282,0.0529159456,0.0109754009,0.262994349,-0.1415746659,0.0655962378,-0.3947277963,0.0975910127,-0.0216612127,-0.1501387358,0.0653591752,-0.0285975765,-0.4013192058,-0.011343061,0.0974163339,-0.1872508377,-0.1904862821,-0.1078211665,-0.0329202674,-0.0189087968,0.1512490362,0.5576463342,0.0483137928,0.2664676905,-0.5254781842,-0.4554860592,-0.082938686,-0.0208982155,-0.10602431,0.2168312073,-0.0720264763,-0.0076387855,-0.0387089998,-0.198524788,0.2651242614,-0.392677933,-0.1527600139,-0.5893936157,-0.0294168368,0.3821199536,0.0262569841,-0.1025057659,0.4024057388,-0.0742120072,0.2647199035,-0.1309686452,0.3175792396,0.2961915731,-0.0121343117,0.4494165778,-0.4036705196,0.0770488828,0.3041946292,0.0112261083,0.0700656548,-0.2057007551,-0.0030244589,0.0318597853,0.331225872,0.1476236284,0.3841544688,-0.158008635,-0.0387701578,0.1640706509,0.1860900372,0.2060954124,-0.2314427644,-0.2006061971,0.0794249848,-0.0831413642,-0.2648019195,0.1411605328,-0.2778375745,-0.1275728047,-0.1497097611,0.0562532693,-0.2145192027,-0.0455981158,-0.1670219004,0.0586627088,-0.5022032261,-0.0633877888,0.6659346223,-0.1974996775,-0.9030703306,-0.2063631862,0.4862928987,0.1802442372,-0.5378523469,0.3778819144,-0.0558725968,-0.0812057182,-0.0459639244,0.6167016029,0.1588868499,-0.2229042053,0.0714154169,-0.1554821432]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2443","title":"Some tests hang on Windows","comments":"Hi ! That would be nice indeed to at least have a warning, since we don't handle the max path length limit.\r\nAlso if we could have an error instead of an infinite loop I'm sure windows users will appreciate that","body":"Currently, several tests hang on Windows if the max path limit of 260 characters is not disabled. This happens due to the changes introduced by #2223 that cause an infinite loop in `WindowsFileLock` described in #2220. This can be very tricky to debug, so I think now is a good time to address these issues\/PRs. IMO throwing an error is too harsh, but maybe we can emit a warning in the top-level `__init__.py ` on startup if long paths are not enabled.\r\n","comment_length":41,"text":"Some tests hang on Windows\nCurrently, several tests hang on Windows if the max path limit of 260 characters is not disabled. This happens due to the changes introduced by #2223 that cause an infinite loop in `WindowsFileLock` described in #2220. This can be very tricky to debug, so I think now is a good time to address these issues\/PRs. IMO throwing an error is too harsh, but maybe we can emit a warning in the top-level `__init__.py ` on startup if long paths are not enabled.\r\n\nHi ! That would be nice indeed to at least have a warning, since we don't handle the max path length limit.\r\nAlso if we could have an error instead of an infinite loop I'm sure windows users will appreciate that","embeddings":[-0.2391643524,-0.0217614975,-0.1179873645,0.0013239856,0.3143847287,-0.0502426848,0.1832035333,0.1566540301,0.164340362,0.3078767061,0.7059500217,-0.0965389833,-0.4796687365,0.042514395,-0.1620287299,-0.1105980501,0.1247819215,-0.122185722,-0.083477959,0.2735829353,-0.5415411592,0.4021037221,-0.1029828191,0.2443026751,-0.4127779603,-0.0245216042,-0.1770639867,0.3733247817,0.0386725366,-0.6603713632,0.0152102541,0.0478582531,0.141298905,0.2531118095,-0.0001164809,-0.1899032295,0.5350030661,0.2290475368,0.012908604,0.1523456424,0.0528207049,-0.1746689975,0.2479258776,-0.1053414717,0.2352506965,0.359431982,-0.1607311517,-0.0304966606,0.0932784081,0.3737002909,0.1050343737,0.3648024499,-0.0494908802,-0.0883277059,0.448241502,0.0189046152,-0.2878830731,0.2482592911,0.4959424734,-0.4592775702,-0.1171995699,-0.0880855694,0.0908664688,0.1507358998,0.022930041,-0.0589563549,0.5751327276,-0.4512700737,-0.2003280818,0.4209567606,-0.2349319309,-0.0156885237,-0.2226243019,-0.28502509,0.1935049295,-0.4098216891,0.5957505107,0.0356167369,-0.1813475937,0.0468960479,-0.022190107,-0.1439809948,-0.1132972986,0.0460565984,-0.2228390276,0.0445665568,0.2310000807,0.0520929731,0.3748267591,-0.0120594194,0.1558532715,0.4347129762,-0.1441374719,-0.2462709397,0.0365051292,-0.1012238041,0.1483522207,0.3499832749,0.2254705727,0.1086855456,-0.3143654466,-0.0978057683,0.5534831285,0.0053729136,-0.3763042688,-0.1428903341,0.0891837329,-0.0676920786,0.4931468368,0.2424603254,-0.0430940166,0.2192297727,-0.0567031801,-0.8042609096,0.0714339167,0.5170311332,0.3584701717,-0.3616822958,-0.0329385325,0.1662298441,-0.1827291697,0.0982009694,0.2688486576,0.2686883509,0.2209935486,0.0942668989,0.0555387214,0.1434252113,-0.2021653652,0.3693977594,0.1390004754,0.0366625823,-0.3773708344,0.1078154892,0.2827774286,0.0050585396,0.2136135101,-0.1209830791,0.3965454996,-0.042351421,0.1081949323,-0.0675143376,0.0929299369,0.2159458846,-0.2991596162,-0.0727473348,-0.012955891,0.165175274,-0.262547493,0.297672689,0.0235384591,-0.4552656412,0.1728268713,0.1255333126,-0.2673514783,0.1717526317,0.4315384626,-0.1196921542,-0.198164016,0.0798500106,0.3290472329,-0.3684563935,-0.0752470195,-0.1185630411,0.1537753791,0.4675206542,-0.0121820001,-0.0108969482,0.2409825325,-0.1799976379,-0.0743229166,0.3938121796,-0.1296010911,0.0681994259,-0.4522504807,-0.0216788501,-0.0568144731,-0.0295012016,-0.009467205,0.0381269567,-0.2809455097,-0.1634589881,0.2682085931,-0.1446130574,0.048056446,-0.2025567293,0.0440236591,0.2531568706,0.0291315578,0.1608382165,-0.4407097697,-0.4872566164,0.4007733166,0.1522914469,0.2565484345,-0.3678263724,0.0031310327,0.2906616628,0.2672812343,-0.0543467067,-0.394531697,0.2366376966,0.152739495,0.1699732393,-0.1435275376,0.2348627746,0.0475034416,0.0943569541,-0.3476723433,-0.0704795197,0.0128654009,-0.2680282295,0.0518457443,0.160308674,0.1790085733,0.0569023974,0.0398340598,-0.0999593213,0.3378398716,0.1612319052,0.0489723682,-0.1757089943,-0.6226357222,0.0756283551,0.1384879649,0.0773188993,-0.1261323392,-0.1271213293,-0.0843830556,0.195040077,0.1032838598,-0.1099059507,0.0767368078,0.2410556972,0.145432964,-0.0404212698,-0.0846027583,0.4348882139,0.1316210777,0.0110705979,0.0123361936,0.421738863,-0.4075356722,0.0736928061,0.0555152893,0.2390256226,-0.3201412261,0.4193177223,0.0374493822,-0.2509611249,-0.0963294208,-0.3603075445,-0.431891948,0.1608762145,0.4325932562,0.2577133179,-0.1193564534,-0.0644837841,-0.0555407405,-0.3684889078,0.4165738523,-0.0080654258,0.0075694863,0.4114581347,0.2304674685,0.1487060487,-0.0904229656,-0.0724782348,0.264903903,0.1598542482,-0.0322241485,0.1685835421,-0.3454689682,-0.2781630158,0.345310688,0.2235974073,-0.1453368664,0.2507613599,0.0226723533,-0.1234521344,-0.1451883167,-0.4776329696,-0.0312441513,0.0449361317,-0.4879174232,-0.1537807584,0.2533180416,-0.3396518528,-0.1639818698,-0.3365969062,-0.137875393,-0.2301132381,0.2494189292,0.1771198064,-0.3231560588,0.1255191863,-0.0189247839,0.3794437647,0.1063991562,0.004055337,-0.12838085,0.1367351711,-0.0805394873,-0.0059561692,0.3297410905,-0.1653635353,0.2761060596,-0.0437184162,-0.1772616953,-0.2435234338,-0.4369324446,-0.0262845885,0.1350604892,0.3849754035,0.2874762118,0.2002984583,0.4496801794,-0.3664644063,0.1500198543,-0.1894405931,0.3117998838,-0.1405071616,0.1857137084,0.0739964917,-0.3200155795,-0.0531776994,0.0043495414,-0.4196608961,0.3991326094,-0.0794469267,0.0366798043,0.0747519657,0.1206176728,0.2620082796,-0.1830413342,-0.0178466681,0.3086880147,-0.2026559561,0.1383386701,0.0618734881,-0.1948967874,0.0705917999,-0.2418643087,-0.4351091683,-0.0551089756,-0.5333495736,-0.3359261453,-0.3845426738,0.5380304456,-0.062015336,-0.2795823812,0.3068859875,0.0603766404,-0.1483675539,0.2410379946,0.0354514122,0.0023783932,-0.0745716244,0.3487825096,-0.0238350239,0.2383545637,-0.1507552713,0.1219179034,0.0534962863,0.2662055492,-0.0433165617,-0.1001688093,0.2181071788,-0.1563321501,0.1521645635,-0.0021526071,-0.1536776572,-0.5739147067,0.1440783292,0.2271134555,0.1518520117,0.3060595989,0.0829183385,0.2084761709,-0.3937461972,0.2117389143,-0.2044584006,0.0711211115,-0.434615314,0.067497097,0.2741535008,0.0690521151,0.3096498549,0.4431302547,0.1157631874,-0.0226891302,0.2122389972,-0.2494978309,-0.3500971794,-0.1736768633,0.1631728113,0.3219476044,-0.3620350063,-0.3022389114,-0.0163498521,-0.3897597194,0.7135555148,0.059542641,-0.0114424946,0.3372002542,0.0989015922,0.0539373495,-0.4292014837,-0.3051165044,-0.0239910968,0.1910817921,0.2757386565,0.1417052746,0.065801546,0.0688887015,0.3272838891,-0.0564779937,-0.0212374628,-0.2575510144,-0.2793650031,-0.3662132323,0.5595241189,0.0650473759,-0.1544142962,-0.4675737023,-0.0880890936,0.0760632753,-0.3324145377,-0.1271100044,0.2750613987,0.149870947,-0.276196748,0.0794333816,-0.0951214209,-0.0618822686,-0.1473118663,-0.0181595311,-0.0943748504,0.3219315708,0.1056081504,-0.2941696942,0.1336150616,0.6162634492,-0.2420330346,0.3068232536,-0.3942895532,0.6207204461,-0.3629673421,0.2847329676,0.526358366,0.0913804099,-0.0375170298,0.2233712971,0.1246892437,0.1991850585,-0.1372984499,0.4933137298,0.1511910707,-0.367644608,0.1610738933,-0.0475364961,0.8037680984,-0.1301330179,0.2636250556,0.2768618762,-0.2134628147,0.0316328257,0.2631060779,0.0049588326,-0.0212567542,-0.0527893677,0.1332623065,-0.1213088408,0.4038451314,-0.2893591821,-0.4949132502,-0.0097368378,0.1173050776,-0.368260175,-0.3090281487,0.3184364736,-0.4009123743,-0.0595816672,-0.2672671974,0.0754588395,0.2497920543,0.137883991,-0.2718999088,0.2237083614,0.3766262829,-0.3399772346,-0.2614260912,-0.1627697349,-0.3189672828,-0.0399445668,-0.1736756563,-0.0212010667,-0.0268190168,-0.2331460267,0.0863535926,0.4869732261,-0.1737555414,0.0750488713,0.1375372112,-0.2199283093,0.1454793662,-0.1300818324,0.0407856666,-0.0873631686,-0.0885955244,-0.1629542559,-0.3459845185,-0.0665066019,-0.3142248392,-0.1424838156,0.2073128819,-0.1719293743,0.0763893723,-0.1434756815,-0.2541716993,-0.0238429941,0.139724791,-0.076994665,-0.1207297146,0.219414264,-0.1506750733,-0.2820222676,-0.1134567484,-0.1849845797,-0.4514195025,0.1019988358,0.3560183644,0.3655843437,0.0873413533,-0.2292153388,0.3393521905,-0.1949692816,-0.2206104249,0.2283402979,0.4495655298,-0.0937003791,0.0368118845,0.2841462493,0.3606710136,-0.0033702597,0.002098666,-0.2205459177,-0.4499720931,0.0822678432,-0.0524765328,0.508962512,-0.149676308,0.0384734236,-0.0635653585,-0.5542202592,-0.2688800991,0.0762819573,-0.317466408,-0.0567343049,-0.1012338102,-0.1626291424,0.2217428386,0.1146094128,0.0061608101,0.0648665652,0.0491265282,-0.1822008193,-0.1530569494,0.0839402452,-0.0784139708,0.0538298152,0.3450561166,-0.2897572219,0.4671118855,0.0912741944,0.049350597,0.1538402736,0.0329513364,0.0941421166,0.2365308851,0.0854239687,-0.0492996015,0.2130630463,-0.1033444554,0.0968944132,-0.1043924615,0.2595861256,-0.1601207405,0.1574349254,-0.0625765324,0.2868470848,-0.1371771544,-0.0394677147,-0.0627598614,0.1878602356,-0.1910609901,0.0976790786,0.1249362081,0.2845531404,-0.3436902165,0.0611680895,0.2983241975,0.2478049397,-0.105340369,0.0937683508,-0.4325433671,-0.2716394067,0.0753200725,0.2076967955,-0.0161249619,-0.3263406456,0.2761287093,0.2176982462,0.1730788052,-0.0593965799,-0.2027761936,0.5297368169,-0.0908666477,-0.0821958333,0.0526892319,0.0511147529,0.1483463645,-0.1169774085,0.1564046443,0.4308781326,-0.0206558183,0.117517136,-0.0875058249,-0.0925823376,0.0252845176,-0.2110583633,0.1873038411,-0.2725271583,-0.0403416567,0.4988441467,-0.2770440578,-0.1210123375,-0.2796237469,-0.0848678574,-0.0767670274,-0.3070852458,0.1241739541,0.1249072701,0.017194394,0.1148021221,0.1018812731,0.0415189303,0.1317458004,0.2958168983,0.046754878,-0.1289890558,0.1441718638,-0.0461649559,0.1282079071,-0.4147498608,-0.0191635098,0.2371945679,0.0628780574,0.1751341671,0.1436931789,0.7410426736,0.188440606,-0.2957277596,-0.1581428647,-0.1831716746,0.1981003135,0.0384298787,0.3525544703,0.1490245461,0.4684921503,0.1160620376,0.1216357127,-0.2220141143,0.4912021756,-0.2774022818,-0.2572142184,-0.0872855783,0.1711492836,0.4528829157,-0.0994067043,-0.3272868097,0.022356391,-0.3697116375,-0.1242754757,0.2412133366,-0.1216954887,0.0919313878,0.1452404708,0.04165712,0.1441595107,0.4421734512,0.4410912693,0.4239453971,-0.2546844482,0.0249417406,-0.4240611494,-0.1157246605,-0.2682241797,-0.0529700741,0.1830562949,-0.2452221066,-0.3702529967,0.0285025258,-0.1508502215,0.0853592306,-0.1534638852,-0.0207221489,-0.3501726389,0.1093712151,0.3746509254,-0.0772379264,-0.0913721174,-0.2115367055,0.2984956801,-0.0262157805,-0.0554387011,0.067315042,-0.0303801075,0.0693889558,0.4362842441,0.0004089497,0.0930492952,0.1103085801,-0.0463981517,-0.1083779559,-0.5494907498,-0.3835366964,0.1194440722,-0.1858762801,0.0514288247,0.0423034392,-0.188593328,-0.2767034769,-0.273067683,-0.0042147716,-0.0806570575,0.0321926028,0.0828681216,0.0449162051,0.1737250686,0.1800950319,-0.0167122744,0.4024492502,0.1928697079,-0.0116305612,0.011189647,-0.2728140354,0.4084324241,-0.2736480832,-0.2728582025,-0.0494354554,0.3432900906,-0.2039267123,0.0199955106,-0.4867305458,-0.1714116484,-0.0184334163,0.1484360546,-0.3963748515,0.2862764001,0.0007690586,-0.0395036787,-0.0259568878,0.0696401075,-0.1661198884,-0.0036587184,-0.1213396788,-0.2500241399]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2443","title":"Some tests hang on Windows","comments":"Unfortunately, I know this problem very well... \ud83d\ude05 \r\n\r\nI remember having proposed to throw an error instead of hanging in an infinite loop #2220: 60c7d1b6b71469599a27147a08100f594e7a3f84, 8c8ab60018b00463edf1eca500e434ff061546fc \r\nbut @lhoestq told me:\r\n> Note that the filelock module comes from this project that hasn't changed in years - while still being used by ten of thousands of projects:\r\nhttps:\/\/github.com\/benediktschmitt\/py-filelock\r\n> \r\n> Unless we have proper tests for this, I wouldn't recommend to change it\r\n\r\nI opened an Issue requesting a warning\/error at startup for that case: #2224","body":"Currently, several tests hang on Windows if the max path limit of 260 characters is not disabled. This happens due to the changes introduced by #2223 that cause an infinite loop in `WindowsFileLock` described in #2220. This can be very tricky to debug, so I think now is a good time to address these issues\/PRs. IMO throwing an error is too harsh, but maybe we can emit a warning in the top-level `__init__.py ` on startup if long paths are not enabled.\r\n","comment_length":85,"text":"Some tests hang on Windows\nCurrently, several tests hang on Windows if the max path limit of 260 characters is not disabled. This happens due to the changes introduced by #2223 that cause an infinite loop in `WindowsFileLock` described in #2220. This can be very tricky to debug, so I think now is a good time to address these issues\/PRs. IMO throwing an error is too harsh, but maybe we can emit a warning in the top-level `__init__.py ` on startup if long paths are not enabled.\r\n\nUnfortunately, I know this problem very well... \ud83d\ude05 \r\n\r\nI remember having proposed to throw an error instead of hanging in an infinite loop #2220: 60c7d1b6b71469599a27147a08100f594e7a3f84, 8c8ab60018b00463edf1eca500e434ff061546fc \r\nbut @lhoestq told me:\r\n> Note that the filelock module comes from this project that hasn't changed in years - while still being used by ten of thousands of projects:\r\nhttps:\/\/github.com\/benediktschmitt\/py-filelock\r\n> \r\n> Unless we have proper tests for this, I wouldn't recommend to change it\r\n\r\nI opened an Issue requesting a warning\/error at startup for that case: #2224","embeddings":[-0.1316355467,-0.0140703134,-0.1183958724,0.0093432153,0.216333136,0.009574512,0.2583823204,0.1399549693,0.2193913609,0.1932305545,0.6185334921,-0.2044840306,-0.456207931,-0.0727645084,-0.1960126758,-0.0045183972,0.1505007446,-0.1235165298,-0.0943990722,0.235921964,-0.5354189277,0.3176811337,-0.1172528714,0.197901547,-0.5318851471,0.0011337473,-0.230655551,0.3467150927,0.1202580854,-0.5881578326,-0.0131818801,0.1594189256,0.0917500779,0.2956953645,-0.0001211409,-0.1967275292,0.4859236777,0.2037378699,0.0005869582,0.1154859439,0.0509205498,-0.1330162883,0.1301416457,-0.0721323341,0.3160827458,0.1963718981,-0.1202354878,-0.0964280218,0.1415560693,0.3219186962,0.0639845505,0.2413581312,-0.0692552328,-0.045142699,0.5415896177,0.0615072176,-0.1869721115,0.3898575306,0.611108005,-0.5240539908,-0.2011203617,-0.0953899026,0.102151759,0.0478345342,0.0524179973,-0.105921939,0.6723552942,-0.3971517086,-0.2064733058,0.3575099111,-0.2766179442,-0.0687448755,-0.2651711106,-0.3719166517,0.2571754456,-0.4026556611,0.5238242745,0.0008922243,-0.1837704629,0.1181459203,0.0411817655,-0.1604121327,-0.0762065351,0.0516591072,-0.2029982507,-0.0085944729,0.2746643424,0.0897675678,0.32466802,-0.0021188331,0.2055932879,0.318670392,-0.066254437,-0.2283949852,0.0005716895,-0.0741439983,0.1557633579,0.4239689112,0.1852652282,0.0662041605,-0.3897069395,-0.0668179989,0.6030475497,-0.0184411891,-0.3685918748,-0.1789285839,0.1245021299,0.0080424519,0.4500328302,0.2327784747,-0.118313089,0.2438619733,-0.0260402206,-0.8544873595,0.031572856,0.502163887,0.3412607908,-0.3400120735,-0.0478162915,0.2321031243,-0.2983105779,0.0759851411,0.4083694816,0.2198483944,0.2574112117,0.1645141244,0.1485760808,0.1445606053,-0.2195202857,0.4130954146,0.1404797435,0.0966344103,-0.3439186215,0.2156903595,0.2637860179,0.0620283112,0.1262094975,-0.1099425182,0.4001162648,-0.0769519806,0.1251265407,-0.1827649921,0.1465218514,0.1761445403,-0.3164168,-0.0967078134,-0.0023512586,0.1427814215,-0.1257780343,0.30697155,0.0105354311,-0.3850651085,0.2843335867,0.0792791545,-0.3350974023,0.2820642292,0.4276290536,-0.1684337705,-0.1730345041,-0.0340239182,0.361295104,-0.4424392283,-0.0297962986,-0.0627548099,0.0382816941,0.4527541399,0.007491949,0.0220764205,0.3214856386,-0.14285703,-0.1722471863,0.3537982404,-0.0991480872,0.1600213498,-0.42561692,0.0160495043,-0.0326019414,-0.0677002221,-0.1579433233,-0.01255247,-0.2833658457,-0.1163136736,0.2613494694,-0.1442889422,0.0733992308,-0.2898311019,0.1417184174,0.1491006166,0.0669078603,0.0689505488,-0.3761093616,-0.4131203294,0.3417448997,0.1601039469,0.2079893053,-0.4094273746,0.0172305983,0.405678153,0.3045521379,-0.054792311,-0.3384215832,0.256454587,0.2818982005,0.1988409162,-0.1141442433,0.2293933779,-0.0392691307,0.1177417338,-0.5058175325,-0.1871633828,-0.0469944924,-0.2523780465,0.0760947168,0.1335120052,0.1980001181,0.1414019167,0.0146605158,-0.0860019401,0.2619712651,0.11723838,0.1211637482,-0.1346013099,-0.6055215001,0.0955851302,0.1001957655,0.1685338169,-0.0450489558,-0.2398446947,-0.1322713643,0.1978309005,0.1731105149,-0.1637741327,0.0901984274,0.1834789962,0.1919383705,0.0020781409,-0.1233244389,0.4388432205,0.1687589884,0.0375101492,-0.01679652,0.3242324889,-0.3502953649,0.0391615257,-0.0130549595,0.2228693515,-0.3726812899,0.3832468987,0.0069089103,-0.300624758,-0.1137449592,-0.3819709718,-0.4186284244,0.108128354,0.4916052818,0.3734047115,-0.0281401817,-0.0706891492,0.090347074,-0.4116041362,0.3713473976,-0.0994368717,0.0052002044,0.3837336302,0.2857146561,0.1169907525,-0.0796300769,-0.1088966951,0.3620381951,0.1165082827,-0.0148175731,0.1597653627,-0.195213899,-0.2953649759,0.2977644801,0.2104063034,-0.1678679287,0.3470380604,0.0618427135,-0.1153237969,-0.0405805446,-0.5296561122,0.0951711684,0.0495553166,-0.4471700788,-0.1198968887,0.1955694407,-0.2618548274,-0.1542730629,-0.320756048,-0.1069817096,-0.2502382398,0.2079909444,0.2245192528,-0.3119969964,0.0664316937,-0.0508479849,0.3877587318,0.1660600752,0.0168033969,-0.094253011,0.1470428854,0.0232992526,-0.0521201491,0.3705758452,-0.168691501,0.2437377125,-0.0290287286,-0.2418398112,-0.2861432135,-0.4756320715,0.0531958416,0.1204546094,0.3724136353,0.3183584511,0.2138862014,0.4626654983,-0.3509402573,0.1200631931,-0.2073254585,0.352994293,-0.1723328084,0.1432347894,0.169257462,-0.3495153487,0.0519581251,0.0000856496,-0.3818429112,0.4376588464,-0.0873671994,0.0153942006,0.0734839588,0.027739536,0.3518814743,-0.1488550305,0.0140369358,0.3381622732,-0.2036547363,0.0736833513,0.1348721981,-0.1714611799,-0.0310961585,-0.2577048838,-0.5798020363,-0.0206013136,-0.5306005478,-0.4399010837,-0.3137562573,0.5588341951,-0.0460235141,-0.2840919495,0.2859584093,-0.0491278321,-0.0965188593,0.1969574541,0.08932776,0.0327275768,-0.0323025845,0.3258376122,0.0719382837,0.2525540292,-0.1229371652,0.056706097,0.0907779485,0.2744390666,-0.0498083979,-0.0268117804,0.2113294899,-0.1781023592,0.1561137289,-0.0355601646,-0.1675130576,-0.5556102991,0.173182413,0.1922280192,0.19232288,0.2859148979,-0.0293291602,0.1788475811,-0.4784809053,0.232171163,-0.2071212232,0.0086901505,-0.4147836864,0.0375189036,0.1523621678,0.0048920773,0.2627697289,0.4239975214,0.2038238049,-0.0822512656,0.3217107654,-0.1980769932,-0.3215710223,-0.1190275326,0.1892000735,0.370375216,-0.3403119743,-0.3152082562,-0.0356046669,-0.4756583869,0.7116439342,0.2410149127,0.02690861,0.308771342,0.1259396374,0.1196006089,-0.4440324306,-0.3544288576,0.0016207689,0.2721108496,0.3886832595,0.0662790611,0.0055716448,-0.0373762548,0.375303179,-0.0752808601,-0.0729089826,-0.1245009452,-0.1841607839,-0.3895639479,0.46912238,0.1008502021,-0.1185636967,-0.4459836483,-0.0183147565,0.1641580462,-0.3795089424,-0.1110716909,0.2170251459,0.0882701203,-0.272993803,0.0277935751,-0.083251588,-0.030651046,-0.0870450884,-0.0524896123,-0.0597247146,0.3577674627,0.2049935311,-0.3324190974,0.0660900548,0.7001744509,-0.2707610428,0.3558590114,-0.3035083413,0.546035409,-0.3487128317,0.2272804379,0.5174638629,0.0943469554,-0.037448667,0.306904614,0.0911040828,0.2328852117,-0.1738639623,0.5167903304,0.1529824436,-0.3960298896,0.1211365461,-0.0166369975,0.8352264166,-0.0260070488,0.2238602936,0.2544744611,-0.1888336092,0.0354264267,0.2474341094,-0.0558233485,-0.0309395771,-0.1968034059,0.1435757279,-0.1611329764,0.4754770696,-0.2515172064,-0.3428009748,-0.0000899531,0.156944558,-0.4665779173,-0.2782803178,0.3115173876,-0.4264758527,-0.018658651,-0.3455867171,0.0174774639,0.2856178284,0.0943987668,-0.3185611069,0.2241392881,0.3486357629,-0.3245661557,-0.2068626732,-0.2578181922,-0.2988286316,-0.0194199607,-0.1755230576,-0.0056736642,-0.0470779128,-0.2395443171,0.0531325452,0.4995245337,-0.1278753728,0.0119272703,0.124128364,-0.1470466703,0.162072435,-0.1486047059,0.0711086169,-0.0737256631,-0.065780662,-0.1510278285,-0.3049004376,-0.1187240183,-0.2922258079,-0.1670825779,0.2822095752,-0.1781995296,0.0037347723,-0.2081586719,-0.3232519329,-0.0005246587,0.0970786959,0.021199815,-0.1766433865,0.2180000842,-0.2662189603,-0.2962591946,-0.1030668318,-0.1590947509,-0.5041978955,0.0084366109,0.4400086701,0.4161369205,0.0701132938,-0.1844922155,0.4453377724,-0.1864585429,-0.2794839442,0.1728159934,0.4562320411,-0.0831811056,0.0385242403,0.240349263,0.4058396816,0.0994956195,0.0002570224,-0.2946366668,-0.4661720693,0.0758159235,-0.0643081814,0.4723452032,-0.09569332,0.0600741915,-0.0809220523,-0.5390076637,-0.2347770482,0.1358180791,-0.3285686374,-0.0168094523,-0.1387875527,-0.2169166505,0.1695279628,0.1015311182,-0.0084708808,0.1124909818,0.0623668209,-0.1454924047,-0.1496883929,0.1069569066,-0.0222423486,0.0489728078,0.317234844,-0.3200664818,0.5094142556,0.105234921,0.0390116498,0.1349602491,0.0971847102,0.0359213203,0.242534101,0.1091914549,-0.0556532256,0.2172622681,-0.0748106018,-0.0327353328,-0.1311033368,0.2655623555,-0.206149295,0.1868949682,-0.2202154249,0.2822438776,-0.1155879572,-0.0418300852,0.0345661454,0.2016751319,-0.1292195022,0.2081947774,0.1083213985,0.2710419297,-0.3482831419,0.0315377563,0.3321911395,0.1888281405,-0.0979036018,0.109931156,-0.3305422068,-0.284910351,0.048243124,0.2159027457,-0.1243167147,-0.3476738036,0.1930438429,0.2405394316,0.1450353861,-0.0219821949,-0.2696686089,0.6022278666,-0.0187443681,-0.0526357517,0.0863804817,0.0695993006,0.1505975127,-0.1465857327,0.0184847917,0.3359057605,-0.1084495634,0.1143249869,-0.0243179388,-0.1211229563,0.0491300598,-0.1018947065,0.1775196344,-0.26161623,-0.1102492362,0.4556134343,-0.2607677579,-0.0235032327,-0.1751489937,0.0115548298,-0.0845474899,-0.3239192963,0.1808032244,0.1568061709,-0.0000404252,0.0279425792,0.0359420553,0.0430746675,0.0123825772,0.3713068068,0.0484474599,-0.0324998759,0.0719863325,-0.0767456442,0.1368587315,-0.3781597316,0.0038229602,0.2791757286,0.1289520413,0.2566325665,0.0438771658,0.6962896585,0.1485913098,-0.3574948013,-0.1588429362,-0.1733488142,0.1942709982,0.0344551653,0.4090585709,0.1812072992,0.360802412,0.0517147146,0.102548562,-0.2340460122,0.4883556068,-0.1853107959,-0.2448047251,-0.1238558516,0.180403471,0.4797127843,-0.0693587586,-0.2663089335,0.0615455881,-0.332978189,-0.1397316456,0.1575901061,-0.1344189346,0.1206689253,0.2623581588,0.0250269324,0.0435602553,0.3081722558,0.3810090125,0.3903263509,-0.2383050323,0.1200440899,-0.3823440969,-0.0943542272,-0.1524959207,-0.0872327685,0.1905202717,-0.3165167272,-0.281804055,-0.0003023286,-0.1478676647,0.0490879305,-0.1348014325,-0.0256959684,-0.3480947018,0.1525303423,0.4213390052,0.0395939611,-0.1225897148,-0.1860152036,0.2646531463,-0.0097409645,-0.0786394998,0.0731905103,-0.1151801795,0.0466732495,0.4070211351,-0.1093025282,0.1019969806,0.1223400086,-0.0606301017,-0.1871463805,-0.5519605279,-0.3645993173,0.1471647024,-0.1969409883,-0.0041503143,0.0791535154,-0.2016838044,-0.3115654886,-0.1854510456,0.0419493355,-0.0735170394,-0.0577250272,0.04423308,0.0926527381,0.1334033757,0.1736555248,0.0220984183,0.4676498473,0.1801199913,-0.0346239433,-0.0110538388,-0.2337912172,0.4812085927,-0.3890375793,-0.2401884198,-0.0704149008,0.388758719,-0.2648902535,-0.0252298284,-0.437930882,-0.1141225994,-0.0028231288,0.1101779342,-0.4098558128,0.193004027,0.0465502553,-0.0217689537,-0.0551146567,0.0229187477,-0.1459445804,0.0514838025,-0.1566633582,-0.2133200914]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2443","title":"Some tests hang on Windows","comments":"@albertvillanova Thanks for additional info on this issue.\r\n\r\nYes, I think the best option is to throw an error instead of suppressing it in a loop. I've considered 2 more options, but I don't really like them:\r\n1. create a temporary file with a filename longer than 255 characters on import; if this fails, long paths are not enabled and raise a warning. I'm not sure about this approach because I don't like the idea of creating a temporary file on import for this purpose.\r\n2. check if long paths are enabled with [this code](https:\/\/stackoverflow.com\/a\/46546731\/14095927). As mentioned in the comment, this code relies on an undocumented function and Win10-specific.","body":"Currently, several tests hang on Windows if the max path limit of 260 characters is not disabled. This happens due to the changes introduced by #2223 that cause an infinite loop in `WindowsFileLock` described in #2220. This can be very tricky to debug, so I think now is a good time to address these issues\/PRs. IMO throwing an error is too harsh, but maybe we can emit a warning in the top-level `__init__.py ` on startup if long paths are not enabled.\r\n","comment_length":109,"text":"Some tests hang on Windows\nCurrently, several tests hang on Windows if the max path limit of 260 characters is not disabled. This happens due to the changes introduced by #2223 that cause an infinite loop in `WindowsFileLock` described in #2220. This can be very tricky to debug, so I think now is a good time to address these issues\/PRs. IMO throwing an error is too harsh, but maybe we can emit a warning in the top-level `__init__.py ` on startup if long paths are not enabled.\r\n\n@albertvillanova Thanks for additional info on this issue.\r\n\r\nYes, I think the best option is to throw an error instead of suppressing it in a loop. I've considered 2 more options, but I don't really like them:\r\n1. create a temporary file with a filename longer than 255 characters on import; if this fails, long paths are not enabled and raise a warning. I'm not sure about this approach because I don't like the idea of creating a temporary file on import for this purpose.\r\n2. check if long paths are enabled with [this code](https:\/\/stackoverflow.com\/a\/46546731\/14095927). As mentioned in the comment, this code relies on an undocumented function and Win10-specific.","embeddings":[-0.2172692865,0.0115094697,-0.1549207866,0.096899502,0.2777678668,-0.1149821952,0.2952309549,0.1591924578,0.2075246274,0.3304062784,0.5141476393,-0.1415558904,-0.3705517948,0.0012734199,-0.1112247929,0.0127639482,0.1575827301,-0.0090894178,-0.1111329198,0.2419539392,-0.5817576647,0.2973512411,-0.1208924428,0.0697888806,-0.4657229781,0.0211458933,-0.1338057667,0.2217046022,0.0226238631,-0.5973180532,-0.0495577939,0.0873737857,0.1996967643,0.3299791515,-0.000115835,-0.1732414663,0.4302377701,0.1597503871,-0.0091918902,0.1272987425,0.0230976343,-0.2545146644,0.127109006,-0.0732763708,0.2885791957,0.3217958808,-0.1892525107,-0.0902725309,0.035655614,0.3643991351,0.0982453078,0.2990233004,-0.0031338937,-0.0298449732,0.5178439021,-0.0710393488,-0.2678998113,0.2770674527,0.6322718263,-0.4585854709,-0.0864377171,-0.0931096077,0.1384407431,0.0669378117,0.047749646,-0.086216785,0.5759895444,-0.4731736779,-0.0358525813,0.4336645007,-0.2455301434,0.1136651784,-0.2645104825,-0.3538419008,0.1901870668,-0.4038680196,0.6520926356,-0.0036579394,-0.1152871102,-0.0088345902,-0.0514102653,-0.1113078669,-0.1150888428,0.0267864317,-0.3807653189,-0.0167559907,0.2789621055,0.0224454012,0.2998565733,-0.1439074427,0.1176183745,0.2302717268,-0.0288799331,-0.1442059577,-0.0252502151,-0.0608080775,0.153076902,0.3980491459,0.1461873353,0.0952547565,-0.2470443994,-0.0520447493,0.504622519,-0.0110126752,-0.4035009444,0.001145718,0.0288733244,-0.0997332931,0.4837907851,0.2388659567,-0.1527222544,0.2103301138,-0.1078674048,-0.8629803658,0.0475985967,0.4625174105,0.2471917123,-0.3369776607,-0.0723709315,0.2264099568,-0.1641430259,0.1272981614,0.3300949633,0.3593279719,0.2021065354,0.0927483588,0.0393642299,0.09883596,-0.2775814831,0.4377490878,0.1801630855,0.0555022359,-0.280264765,-0.0411236696,0.3397875726,-0.0915286392,0.1394988745,-0.1317785829,0.2847776413,-0.1143499315,0.0431971699,-0.1867536306,0.3103039265,0.1996109188,-0.1265117824,-0.1007327959,-0.0554307103,-0.0333479494,-0.1700852513,0.2480184734,0.0035550778,-0.3172843754,0.2403085679,0.1077100709,-0.3247683644,0.2129339427,0.4897372425,-0.191144824,-0.2119082063,0.0256810207,0.2849576473,-0.4452630877,-0.0673917085,-0.0775783285,0.1355928928,0.4370955527,-0.164684996,0.1038484424,0.2707796693,-0.1279554516,-0.0830543712,0.3319161534,-0.1123683006,0.0614607148,-0.4950167239,0.1328362375,-0.0266108532,-0.0144911064,0.0068828682,-0.0183742233,-0.3378442824,-0.0011196708,0.2841236591,-0.2005919814,0.1528516114,-0.2663381398,0.1396324635,0.3033238649,0.101021722,0.0706548616,-0.299346447,-0.3317329288,0.3780610859,0.1408145577,0.1832021326,-0.4835068583,0.0631887093,0.352191031,0.2053332776,-0.101463832,-0.34100613,0.3710107803,0.2650710642,0.1796691269,-0.1401556283,0.2892664075,-0.0001360326,0.1813503355,-0.4352968633,-0.1165557504,0.0808815435,-0.2388078123,0.0949820429,0.1937977523,0.1762120575,0.1112867966,0.0619298629,-0.0714723095,0.3773116171,0.1645885408,0.1133344099,-0.1022687629,-0.5447921157,0.087233685,0.0233133771,0.1140279546,-0.1472142041,-0.1905942559,-0.0353814811,0.1501614451,0.2323101312,-0.1448000968,0.1112753227,0.1464542896,0.1687177569,0.0312835239,-0.2211984843,0.4166812301,0.1917732656,0.1476076096,-0.0157336127,0.3653412461,-0.2827352285,0.0291609131,0.0455632098,0.2608527243,-0.3534026146,0.4727430344,0.1586640924,-0.2761830986,0.0587769374,-0.3594389856,-0.4150963128,0.0549218133,0.6074181199,0.2353583574,-0.000370414,-0.0002103605,0.1212193891,-0.2808741033,0.5422959924,0.0568547174,-0.0394089483,0.3032362461,0.1459266394,0.1916733086,-0.0751290396,0.0067373938,0.2901608348,0.1829991937,0.0901913419,0.0914913565,-0.3158392906,-0.2774303854,0.4461654425,0.232043758,-0.1507961303,0.3037210107,0.0946591645,-0.1597191095,-0.0675068349,-0.5304828882,0.1028740928,-0.0298284534,-0.4843880236,-0.1231740639,0.1148471609,-0.3094687462,-0.3127938509,-0.3310083151,-0.1815257818,-0.2108958811,0.2179022729,0.2613316476,-0.3223773539,0.1804905236,0.0376184992,0.3314249814,0.0883596614,-0.1428709328,-0.0527734235,0.2226195931,-0.0607940629,-0.0157803297,0.4539106488,-0.2740629613,0.1814881265,0.0092139021,-0.1629222333,-0.2455580831,-0.4740955532,-0.0063958294,0.0298700947,0.2510574758,0.2609376311,0.2094195038,0.4769947529,-0.2721766531,0.093031235,-0.1760141253,0.3480132818,0.0100299781,0.1739299148,0.0580613799,-0.345510751,-0.039953623,0.0382979959,-0.3544498682,0.3413948119,-0.0103595871,0.0404326841,0.1787366122,0.161766693,0.330088228,-0.0868528783,-0.0767032132,0.2947015464,-0.1412647814,0.2555148005,0.0037729361,-0.2388873398,0.1578398645,-0.1839338243,-0.4813085794,0.0166409221,-0.4579631984,-0.2532841861,-0.4618512094,0.5632853508,-0.0207420327,-0.1893278956,0.4016420543,-0.0025791954,-0.1776579171,0.1646380424,0.1219462752,0.0253792126,-0.0352754742,0.3957588077,-0.0465228967,0.1563889682,-0.1581442207,-0.051425755,0.0910616294,0.2764056921,0.0424139984,-0.0738487616,0.2500141561,-0.1753449589,0.1355897486,-0.0705150887,-0.2777818739,-0.56074512,0.2470707446,0.1676240563,0.1236436069,0.1883511543,-0.0788276419,0.3090281785,-0.4608138502,0.3542456031,-0.1459750533,0.0333634987,-0.328438133,0.0127495453,0.0957138538,-0.0203414168,0.1347704679,0.2582072914,0.1035258025,-0.0879031643,0.1836244911,-0.2839953303,-0.4292935431,-0.1621987075,0.2014072239,0.4037257433,-0.3110534251,-0.2601674199,-0.0667255595,-0.4968472123,0.707783401,0.0704316422,0.1334697604,0.3319799006,0.0088189719,0.0172392037,-0.3556056023,-0.375123322,-0.1942510903,0.2851825655,0.3584681451,0.0886965692,0.0828644633,-0.0194454305,0.3784443438,-0.0700273886,0.0862086564,-0.3121836185,-0.2375123203,-0.3458348513,0.4420398176,0.1834160239,-0.037881691,-0.4988885224,-0.0989255756,0.0018958112,-0.4247414768,-0.0299913269,0.2790254653,0.023910448,-0.1926021129,0.1185649037,-0.1690094322,-0.0550748669,-0.0781916305,0.1308693737,-0.1462377459,0.1646686643,0.0714078099,-0.3979042172,0.2017241567,0.6667283177,-0.2257810384,0.267336458,-0.3896172047,0.4993730485,-0.4118640423,0.2541935742,0.3653788567,0.2016883492,-0.0261402577,0.3741235137,0.1126386523,0.2195997536,-0.0678744316,0.4350931644,0.1655338556,-0.4541857541,0.1107249111,-0.0357790813,0.7684594989,-0.1025959179,0.2361129671,0.1936135292,-0.272657752,0.0635888055,0.180794552,-0.0152714914,0.0121654086,-0.1770032793,0.1316200495,-0.1737526357,0.4840862453,-0.2300782651,-0.3916828334,-0.0171940699,0.1450698823,-0.2043371052,-0.2742081881,0.3656720817,-0.4138526917,-0.1832092553,-0.326726675,0.0717030168,0.2668474615,0.1039491594,-0.2877571285,0.2532528043,0.2862624824,-0.3419090211,-0.1302772909,-0.1696576923,-0.3729836643,-0.0192673076,-0.3036741018,-0.1988729984,-0.0507909507,-0.1214193106,0.0458581261,0.4932070971,-0.1412708908,0.092087172,0.1528895348,-0.1999787539,0.1736022532,-0.1752399057,0.0730566829,-0.131025359,-0.0874770805,-0.1194305867,-0.235084936,0.031761542,-0.1612134427,-0.0762142688,0.1734034419,-0.1093053147,0.045687817,-0.1932756156,-0.217442289,-0.091774337,0.134964928,-0.2188099921,-0.1472105831,0.2437861264,-0.1576461792,-0.2920387089,-0.1590313166,-0.0159774441,-0.5558786392,0.0937789902,0.4450945258,0.416370362,-0.027555421,-0.2198133916,0.4310459197,-0.2096080333,-0.2238228619,0.1515854001,0.453561306,-0.1138733774,-0.0563181601,0.354550004,0.2941829264,-0.046444159,-0.0360759571,-0.3193818927,-0.3624599278,-0.0140301762,0.0657080263,0.4530293345,-0.1570481658,0.1419938207,-0.0550155528,-0.4760751426,-0.2775469124,0.0929115042,-0.4631843865,-0.0248706602,-0.2316605896,-0.1815753132,0.0835804343,0.0925404727,0.0119066527,0.1723335236,0.0727681145,-0.2090094239,-0.1925768405,0.0740879327,-0.0589477532,0.0469003133,0.3545170724,-0.3356933296,0.3852955997,0.0282090306,0.043265067,0.1046054587,0.0090737557,0.0372019634,0.2115103602,0.0809888616,-0.053393621,0.2145834267,-0.1268549114,0.1809292734,-0.0225600023,0.2321632504,-0.2318402976,0.217737034,-0.1435003132,0.2415046692,-0.1232784241,-0.0051571294,-0.0066484543,0.147330299,-0.145492956,0.0873111337,0.0597389042,0.3479053676,-0.3795172572,0.0672825128,0.4205959439,0.2158334404,-0.1561333388,0.0957379118,-0.3546185195,-0.3607589006,0.141210705,0.2810517848,-0.0189430118,-0.3506754935,0.293282032,0.1566913426,0.3039092124,0.0890587494,-0.2764709592,0.5668206811,0.0337315835,-0.0283818766,0.0461749099,0.1355449706,0.135297671,-0.0225715525,0.0025074638,0.3725320995,-0.1431926787,0.2135623991,-0.0902713463,-0.1606612802,0.0194220617,-0.1444691867,0.1291565597,-0.1779028177,-0.0277129114,0.3827035427,-0.3580883443,-0.0696325228,-0.2955851853,0.0538028814,-0.0824809745,-0.23889108,0.0795329586,0.1937398612,0.0749653354,0.1684906036,0.1426730603,0.0473220311,0.06776838,0.4221164584,0.1413723677,-0.0595813766,0.1098185778,-0.1941904128,0.1477009207,-0.463653326,-0.0268103909,0.2835348547,0.1742487997,0.1700770259,0.0491864718,0.6987282634,0.0957893282,-0.2471297234,-0.1762970984,-0.2665339112,0.2089702338,-0.0835699588,0.3567848802,0.0949687511,0.3964253664,0.1978457123,0.1072713211,-0.2538060248,0.2967761755,-0.1387280673,-0.3420832455,-0.1190982684,0.093678318,0.498916924,-0.0352753513,-0.3549276292,-0.0653397813,-0.2709409595,0.0372456387,0.2163382769,-0.1412046403,0.1156894043,0.2288043499,0.0481812283,0.0198706537,0.3926139474,0.2761422992,0.4154982865,-0.2311154455,-0.0514356084,-0.4386861622,-0.1502940357,-0.3161567748,-0.1545128077,0.1450296193,-0.2663190663,-0.3011598587,0.0207017139,-0.0948858857,0.0588882305,-0.0972879231,0.0100234812,-0.369794786,-0.0255805515,0.3494021893,-0.0572845861,-0.1743710637,-0.1982958466,0.3089749515,-0.1150209755,-0.0442843065,0.0825243294,-0.0622957349,0.0444777012,0.3745492399,0.0711262971,0.0429100394,0.1308486015,-0.018859949,-0.0639654025,-0.5000753999,-0.1529311687,0.1654265821,-0.1978626996,0.1462386996,-0.0194315873,-0.2850457132,-0.2424501777,-0.1718780994,0.0413159579,-0.0024932153,0.1153156608,0.0667586252,0.0396885984,0.2512597144,0.2332679629,0.0352139138,0.498524785,0.0693052188,-0.0566730797,-0.0659211949,-0.303914696,0.3496402502,-0.2846904993,-0.3333358467,-0.0133777801,0.4205602407,-0.250023067,-0.0112793585,-0.3974562287,-0.0966489315,-0.0782213658,0.1323837787,-0.4807889462,0.2450626343,-0.0051570809,0.0007845252,-0.1002022848,-0.0114858067,-0.0952915028,0.1334641129,-0.1964176893,-0.1109139174]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2441","title":"DuplicatedKeysError on personal dataset","comments":"Hi ! In your dataset script you must be yielding examples like\r\n```python\r\nfor line in file:\r\n ...\r\n yield key, {...}\r\n```\r\n\r\nSince `datasets` 1.7.0 we enforce the keys to be unique.\r\nHowever it looks like your examples generator creates duplicate keys: at least two examples have key 0.\r\n\r\nYou can fix that by making sure that your keys are unique.\r\n\r\nFor example if you use a counter to define the key of each example, make sure that your counter is not reset to 0 in during examples generation (between two open files for examples).\r\n\r\nLet me know if you have other questions :)","body":"## Describe the bug\r\nEver since today, I have been getting a DuplicatedKeysError while trying to load my dataset from my own script.\r\nError returned when running this line: `dataset = load_dataset('\/content\/drive\/MyDrive\/Thesis\/Datasets\/book_preprocessing\/goodreads_maharjan_trimmed_and_nered\/goodreadsnered.py')`\r\nNote that my script was working fine with earlier versions of the Datasets library. Cannot say with 100% certainty if I have been doing something wrong with my dataset script this whole time or if this is simply a bug with the new version of datasets.\r\n\r\n## Steps to reproduce the bug\r\nI cannot provide code to reproduce the error as I am working with my own dataset. I can however provide my script if requested.\r\n\r\n## Expected results\r\nFor my data to be loaded.\r\n\r\n## Actual results\r\n**DuplicatedKeysError** exception is raised\r\n```\r\nDownloading and preparing dataset good_reads_practice_dataset\/main_domain (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/good_reads_practice_dataset\/main_domain\/1.1.0\/64ff7c3fee2693afdddea75002eb6887d4fedc3d812ae3622128c8504ab21655...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nDuplicatedKeysError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 dataset = load_dataset('\/content\/drive\/MyDrive\/Thesis\/Datasets\/book_preprocessing\/goodreads_maharjan_trimmed_and_nered\/goodreadsnered.py')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, **config_kwargs)\r\n 749 try_from_hf_gcs=try_from_hf_gcs,\r\n 750 base_path=base_path,\r\n--> 751 use_auth_token=use_auth_token,\r\n 752 )\r\n 753 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 573 if not downloaded_from_gcs:\r\n 574 self._download_and_prepare(\r\n--> 575 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 576 )\r\n 577 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 650 try:\r\n 651 # Prepare split will record examples associated to the split\r\n--> 652 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 653 except OSError as e:\r\n 654 raise OSError(\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 990 writer.write(example, key)\r\n 991 finally:\r\n--> 992 num_examples, num_bytes = writer.finalize()\r\n 993 \r\n 994 split_generator.split_info.num_examples = num_examples\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in finalize(self, close_stream)\r\n 407 # In case current_examples < writer_batch_size, but user uses finalize()\r\n 408 if self._check_duplicates:\r\n--> 409 self.check_duplicate_keys()\r\n 410 # Re-intializing to empty list for next batch\r\n 411 self.hkey_record = []\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in check_duplicate_keys(self)\r\n 347 for hash, key in self.hkey_record:\r\n 348 if hash in tmp_record:\r\n--> 349 raise DuplicatedKeysError(key)\r\n 350 else:\r\n 351 tmp_record.add(hash)\r\n\r\nDuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 0\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.7.9\r\n- PyArrow version: 3.0.0\r\n","comment_length":104,"text":"DuplicatedKeysError on personal dataset\n## Describe the bug\r\nEver since today, I have been getting a DuplicatedKeysError while trying to load my dataset from my own script.\r\nError returned when running this line: `dataset = load_dataset('\/content\/drive\/MyDrive\/Thesis\/Datasets\/book_preprocessing\/goodreads_maharjan_trimmed_and_nered\/goodreadsnered.py')`\r\nNote that my script was working fine with earlier versions of the Datasets library. Cannot say with 100% certainty if I have been doing something wrong with my dataset script this whole time or if this is simply a bug with the new version of datasets.\r\n\r\n## Steps to reproduce the bug\r\nI cannot provide code to reproduce the error as I am working with my own dataset. I can however provide my script if requested.\r\n\r\n## Expected results\r\nFor my data to be loaded.\r\n\r\n## Actual results\r\n**DuplicatedKeysError** exception is raised\r\n```\r\nDownloading and preparing dataset good_reads_practice_dataset\/main_domain (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/good_reads_practice_dataset\/main_domain\/1.1.0\/64ff7c3fee2693afdddea75002eb6887d4fedc3d812ae3622128c8504ab21655...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nDuplicatedKeysError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 dataset = load_dataset('\/content\/drive\/MyDrive\/Thesis\/Datasets\/book_preprocessing\/goodreads_maharjan_trimmed_and_nered\/goodreadsnered.py')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, **config_kwargs)\r\n 749 try_from_hf_gcs=try_from_hf_gcs,\r\n 750 base_path=base_path,\r\n--> 751 use_auth_token=use_auth_token,\r\n 752 )\r\n 753 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 573 if not downloaded_from_gcs:\r\n 574 self._download_and_prepare(\r\n--> 575 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 576 )\r\n 577 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 650 try:\r\n 651 # Prepare split will record examples associated to the split\r\n--> 652 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 653 except OSError as e:\r\n 654 raise OSError(\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 990 writer.write(example, key)\r\n 991 finally:\r\n--> 992 num_examples, num_bytes = writer.finalize()\r\n 993 \r\n 994 split_generator.split_info.num_examples = num_examples\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in finalize(self, close_stream)\r\n 407 # In case current_examples < writer_batch_size, but user uses finalize()\r\n 408 if self._check_duplicates:\r\n--> 409 self.check_duplicate_keys()\r\n 410 # Re-intializing to empty list for next batch\r\n 411 self.hkey_record = []\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in check_duplicate_keys(self)\r\n 347 for hash, key in self.hkey_record:\r\n 348 if hash in tmp_record:\r\n--> 349 raise DuplicatedKeysError(key)\r\n 350 else:\r\n 351 tmp_record.add(hash)\r\n\r\nDuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 0\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.7.9\r\n- PyArrow version: 3.0.0\r\n\nHi ! In your dataset script you must be yielding examples like\r\n```python\r\nfor line in file:\r\n ...\r\n yield key, {...}\r\n```\r\n\r\nSince `datasets` 1.7.0 we enforce the keys to be unique.\r\nHowever it looks like your examples generator creates duplicate keys: at least two examples have key 0.\r\n\r\nYou can fix that by making sure that your keys are unique.\r\n\r\nFor example if you use a counter to define the key of each example, make sure that your counter is not reset to 0 in during examples generation (between two open files for examples).\r\n\r\nLet me know if you have other questions :)","embeddings":[-0.1864345074,-0.0565234087,0.0734844357,0.2193849683,0.1425481141,0.0292863455,0.5267676711,0.2207098007,0.0588233471,0.0872558504,-0.1790747494,0.2924934328,-0.1353885084,-0.1258903295,0.3225165606,0.1644579619,0.0292240959,-0.0048958994,-0.1378866285,0.0562093109,-0.1669512391,0.2724630833,-0.0154245673,0.4006582499,-0.1122406051,0.0180170108,-0.1549707204,0.3373055458,0.1574178636,-0.5524030328,0.3723440766,0.115193449,-0.0585309304,0.6388737559,-0.000114377,0.089326188,0.3678568304,0.0726804435,-0.3368042707,-0.2682757378,0.0349326618,-0.0251707342,0.1887518764,-0.3304275572,0.0402530059,-0.229095608,-0.0011710926,-0.3105666935,0.3904740512,0.2292785496,0.2183040529,0.0367066115,-0.0755213723,-0.1794561446,0.1490657926,0.3316625059,0.1221522838,0.2524907887,0.2090585381,-0.1218607128,0.093560718,0.0922568589,-0.2326767594,0.2715856433,0.4658198953,0.1381968111,-0.2277673036,-0.0389199071,0.2810682356,0.4387474358,0.4561787844,-0.5366321206,-0.2995787263,-0.4711150825,0.1770800203,-0.2411703467,0.1227191463,-0.0223870464,-0.1665583104,0.2151415795,-0.1649234295,-0.0753314644,-0.1069493145,0.0315973312,0.0011806944,-0.1223172769,-0.049361892,0.271197468,0.1206141412,-0.1859199107,-0.0420606621,-0.0807006657,-0.062629737,0.0479095578,-0.1217092127,-0.0121265231,0.0719313547,-0.0356171615,0.4212903976,0.1260746419,0.1623898894,0.0511264056,0.0721501485,0.0158360321,0.4798832834,0.0257830061,0.0838383511,0.253839016,0.2858423293,0.1228547394,-0.1708331406,-0.0057329764,-0.037885841,0.0616610274,0.5174378753,-0.0436327457,0.700039506,-0.2715612054,-0.2010089159,0.0989135578,0.0076636784,-0.1383434385,-0.0781496763,0.1666328758,0.0921997353,0.0152029637,-0.180369392,0.0483593643,-0.2964209914,-0.1243328527,-0.1058553755,0.090281494,0.0710753873,-0.030139178,0.32089746,-0.4846819341,0.1354746073,0.0996275842,0.022484228,-0.3461916447,0.1496575475,0.0179346725,-0.1550012827,0.0956206843,0.1365298182,0.372443825,0.0551547557,-0.2492426038,-0.1250035167,0.1580288708,-0.3686243892,-0.4084016979,-0.1084923893,0.223093614,-0.2670642436,-0.1035706699,-0.1620784849,0.055188071,0.3332821131,-0.1818750203,-0.0662663952,-0.2544150054,-0.4359380305,-0.2930590808,-0.0488225035,0.627838254,-0.3001243174,-0.0828407332,-0.5352852345,-0.1706262529,-0.2267044485,0.268750906,-0.3589932323,0.5027720928,-0.3484795094,0.2010123581,0.0668412521,-0.3456726968,-0.6580873132,0.5311324,-0.2746317685,0.327347517,0.0416119695,0.1276919842,0.4651711285,-0.1233073696,0.1120899841,-0.1775349528,0.0862104595,-0.1243482232,-0.0574253239,-0.1677972376,0.050897643,0.0288599618,0.0928659365,-0.2194238007,0.2983483076,0.0872548148,0.2812464535,-0.2383116633,-0.0349764228,0.1662206501,0.1789264679,0.0728690475,0.0283241384,-0.1941446811,-0.4341951609,0.3175955117,0.0040579163,-0.3278971016,-0.2090773284,-0.1817663461,-0.4485370219,0.0098392963,-0.5775942206,0.0009188339,0.0714324936,0.2051878422,-0.0812789798,0.0258612875,-0.1853131205,0.4073454738,0.1822107881,0.0808900222,-0.2352645993,0.0039496641,-0.0226734448,0.0041180314,0.0736214295,0.0747316927,0.3385287821,-0.0815312937,-0.1256201267,0.4060436487,0.064183481,0.1542711258,-0.187399596,-0.1152945533,0.1179354936,-0.154614687,0.3159890771,0.1982626766,0.2679871917,-0.1654774994,-0.3924897015,0.3851958513,0.0091077089,-0.1224951595,-0.1656302661,-0.0794710889,0.0632036924,-0.2009999901,-0.0268606432,-0.2052624524,0.2908895612,-0.2018519938,0.0799277723,0.1402326524,-0.1887188554,0.1398288161,0.1679102331,-0.1818193793,0.2790426016,0.1638421267,-0.0441890433,0.0209840108,-0.0386851579,0.2999103069,0.5103562474,0.153732866,-0.0037189524,0.2641640007,-0.1228415072,-0.0646803752,0.0696571469,-0.0417983532,0.0226826686,0.2188365459,0.0694565699,0.0686904863,-0.2538977861,0.0455449224,0.0083703725,0.3561425507,-0.4474920034,0.1444788277,-0.5289639831,-0.0410224162,-0.2909207344,0.1154117361,0.068720296,-0.1183312982,-0.0059037218,0.3284442425,0.0763666332,-0.0614555366,-0.2195770591,0.0737695545,-0.1509956568,-0.0838600397,0.0374967493,-0.1066306606,-0.457036078,0.0393652692,0.4354344904,-0.0755674616,0.4688960612,0.1013907269,-0.0867710486,-0.5481936932,-0.1828355342,-0.1221205518,-0.0135880848,0.1786353141,0.2338028997,0.0347757079,0.1136032194,-0.0659089908,0.0344041139,-0.1430894136,-0.2460508496,0.1748118997,-0.1032154411,-0.0170347877,-0.0606005527,-0.4004840553,-0.2146552801,0.0589309596,0.2180157304,-0.0278810523,-0.0025197237,0.1291808933,-0.0794784054,-0.0344787128,0.0746449903,-0.1346056461,-0.3363359571,-0.1720222831,0.2224801928,-0.14923127,-0.345107913,-0.0564711317,-0.0509004332,0.0318320617,-0.1824121773,-0.7195866704,-0.0192839764,-0.1149610728,0.2436561137,-0.2316603661,-0.1231254414,0.1163996905,0.0450868905,0.0219756514,-0.1702771932,-0.4202033281,-0.1165173352,-0.1416843086,0.0535265766,0.1716160625,0.5161118507,0.0215465147,0.3910152316,0.4786762893,-0.0518728308,0.1198170781,-0.0967197791,0.391258955,-0.1630542278,-0.5665510893,-0.0886632353,-0.1595169753,-0.0959168449,0.109044984,-0.141204074,-0.0564710759,0.0449788719,-0.1053584889,-0.4014692903,-0.2534804642,-0.1894110739,-0.0761737451,-0.1574556231,0.1937532574,0.2676413059,-0.1156040877,-0.0787375718,0.1312841326,0.0939033106,0.0195747502,0.0200156886,-0.3347143829,-0.0318388268,-0.5773205757,0.2624218762,0.0507724211,0.2812111676,-0.1805754006,0.2767280638,0.1589816064,-0.1883494258,0.3186620474,-0.1713173091,0.1666193902,0.2084120363,-0.3217483163,-0.3334866166,-0.1418097913,0.0553301051,0.3321937621,-0.1960080713,0.3390961885,-0.301961571,0.2847577631,0.4102033675,0.2261133343,-0.300919205,-0.457372576,0.0091470117,-0.6291176677,-0.4878902733,-0.0378224589,-0.0553398281,0.1563607454,-0.0442626253,0.0646786913,0.0115400823,0.0188315008,0.2603390217,0.1147316098,0.2303770781,0.0658604577,0.5187873244,0.1961691827,-0.0277723148,0.279868871,0.7565236688,0.001823833,-0.2737170458,-0.1594572216,-0.2390248626,-0.0713587999,0.2079110742,-0.1034442112,0.097773239,0.2249877602,-0.2202137411,-0.0234991498,-0.0852021053,0.1944959313,0.026329156,-0.4011048675,-0.3828482926,0.431596905,0.1299850792,-0.1661482453,0.3216854036,-0.1509178132,-0.0649240687,0.0862527639,-0.3026022315,0.5584703088,-0.2048691213,0.1103623211,0.1195008904,-0.4390955269,0.2995179892,-0.0239378847,-0.1661559939,-0.5218623281,-0.1000197008,0.0326341316,0.0657745972,0.096465975,0.0393105373,-0.0514230654,0.3023513556,0.0951553211,0.0253832638,0.0172300767,0.5617966652,0.0428265594,-0.0991237089,-0.0670165047,0.1566982865,-0.1801558584,0.2203921676,-0.0533136018,0.0197015628,-0.0324234664,-0.3704388142,-0.5147485733,0.0243518166,-0.0554232188,0.4853644371,-0.0828788206,-0.6532334685,0.1719492525,0.1229546815,0.238294512,0.0626022145,-0.2026261091,0.1107624397,-0.0813313201,0.2891862094,-0.2048490793,-0.244732067,0.1065135151,-0.0338678658,-0.411108166,-0.0314404815,0.1727112681,-0.1459559053,-0.1441899985,0.1145526171,0.292016238,-0.297840625,-0.1786839515,-0.1561232209,0.0387242883,-0.2815293372,0.053031791,0.1902587116,0.0553611815,0.1686558872,-0.1481540203,-0.3456920981,-0.0585029162,0.5251168013,0.1614068002,-0.2164437175,0.5713971257,0.0784837157,-0.0110600414,-0.1443631798,0.3910468519,0.2197487801,-0.2600576878,0.075141646,-0.1700660735,0.4170729518,0.1211233214,-0.1060315445,0.090955168,-0.0373879746,0.01344804,-0.1337712705,-0.2923740745,-0.0398552343,0.0645568669,0.2733722925,0.3497540355,0.0700478256,0.021049032,0.0084096128,-0.2461220622,0.2840166092,0.3811727464,-0.1247937754,0.3689233959,0.0878285021,-0.1213931888,0.0827203244,0.1207188293,0.3267295659,-0.0182491578,-0.1973792762,-0.1536381543,0.1597957611,-0.0764238834,0.0101838158,-0.1052649394,-0.1975092739,0.042086117,-0.0829911605,0.2619949877,0.4193260372,0.1065906882,0.2137805521,0.2205275595,-0.0996406227,-0.183651045,0.2850447893,-0.2902155817,0.1308636069,0.1731794924,0.248304382,0.1261893064,0.0332860611,-0.3144943118,0.0719708949,0.148118481,-0.088149257,0.1994998455,-0.4802075922,0.0026393796,0.2147420198,0.1991143972,0.3138101995,-0.3521753848,-0.2660816312,0.0199558288,0.1040719971,-0.3379886448,-0.3375861943,0.3838552833,0.1304217577,-0.1627376974,0.3674614727,0.0118581876,-0.2065409273,-0.1925642937,0.0000249554,0.5280101299,-0.1580040455,0.1635504067,0.3557379842,-0.1377314925,-0.2449371964,0.1990891546,0.0457413979,0.2326608747,0.7291664481,-0.3192643821,0.0695519745,0.0388671346,0.2995016873,0.119696416,-0.5269600749,0.2326116413,0.0199155286,-0.1420402825,0.0734836832,0.3483227789,0.2425382435,-0.0950403214,0.0766014457,-0.3192046881,-0.0517552495,-0.1597227901,-0.1258331239,0.0142505532,-0.1257435232,0.026150303,0.095063895,-0.2226903141,-0.1025366336,0.2904983461,0.0164554939,-0.4882953167,-0.4008246064,0.1408111304,0.0700542629,0.077807039,-0.2178088725,0.134912163,0.4335056245,0.0613660887,0.258336544,0.2638992965,0.6102486849,0.1389959455,-0.0220020711,0.1325817704,-0.1553770155,0.0196851864,0.1498385221,0.1509149075,-0.1279217154,0.362713784,0.3473048508,0.1291383505,0.0603024252,0.2542678416,0.3459598124,-0.0217789169,-0.4354406297,0.1667446941,-0.2433710396,-0.1651654243,-0.1628522426,0.381146878,-0.2777805626,-0.2351034731,0.5976246595,0.3726883233,0.0739829391,-0.2082563788,0.0601051003,-0.1204657182,0.1670228392,0.2927010655,-0.1981873661,-0.2354176641,0.1743615419,-0.9540211558,0.3124175966,-0.5523523092,-0.0202428233,0.0663271919,0.2064768076,-0.1396538168,0.0777139068,0.3370631933,-0.0765122175,-0.142531693,0.2806809247,-0.1841120422,-0.2737287283,0.2567863464,-0.0319788978,0.0418113098,-0.2001496702,0.2216446996,-0.2371189445,0.0011960795,-0.1737746745,-0.1126349494,0.1364224702,-0.1621296108,0.4371235967,0.1686959118,0.509370625,0.1869760454,-0.107384339,-0.3273393512,0.0053863325,-0.1892150342,0.5253096819,0.166409418,0.4103934765,0.0189346354,-0.1680552065,-0.1164797172,0.1793960184,-0.016118167,0.0582556091,-0.0924997479,0.1548611373,-0.0696645081,0.0682693645,-0.0576005727,0.2311602533,-0.0504845865,0.4076499641,-0.1823980957,-0.5174458623,0.5359555483,-0.3246198297,-0.1127242371,-0.2242871821,0.319124043,-0.2730090618,-0.0922168195,-0.439629972,0.1477435529,0.5595074296,-0.0903044045,0.0617006533,0.1447773129,-0.2487705648,0.1322369277,0.0044943751,0.2902173102,-0.00843082,-0.1849803627,0.4908718169,-0.1793061048]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2441","title":"DuplicatedKeysError on personal dataset","comments":"Yup, I indeed was generating duplicate keys. Fixed it and now it's working.","body":"## Describe the bug\r\nEver since today, I have been getting a DuplicatedKeysError while trying to load my dataset from my own script.\r\nError returned when running this line: `dataset = load_dataset('\/content\/drive\/MyDrive\/Thesis\/Datasets\/book_preprocessing\/goodreads_maharjan_trimmed_and_nered\/goodreadsnered.py')`\r\nNote that my script was working fine with earlier versions of the Datasets library. Cannot say with 100% certainty if I have been doing something wrong with my dataset script this whole time or if this is simply a bug with the new version of datasets.\r\n\r\n## Steps to reproduce the bug\r\nI cannot provide code to reproduce the error as I am working with my own dataset. I can however provide my script if requested.\r\n\r\n## Expected results\r\nFor my data to be loaded.\r\n\r\n## Actual results\r\n**DuplicatedKeysError** exception is raised\r\n```\r\nDownloading and preparing dataset good_reads_practice_dataset\/main_domain (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/good_reads_practice_dataset\/main_domain\/1.1.0\/64ff7c3fee2693afdddea75002eb6887d4fedc3d812ae3622128c8504ab21655...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nDuplicatedKeysError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 dataset = load_dataset('\/content\/drive\/MyDrive\/Thesis\/Datasets\/book_preprocessing\/goodreads_maharjan_trimmed_and_nered\/goodreadsnered.py')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, **config_kwargs)\r\n 749 try_from_hf_gcs=try_from_hf_gcs,\r\n 750 base_path=base_path,\r\n--> 751 use_auth_token=use_auth_token,\r\n 752 )\r\n 753 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 573 if not downloaded_from_gcs:\r\n 574 self._download_and_prepare(\r\n--> 575 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 576 )\r\n 577 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 650 try:\r\n 651 # Prepare split will record examples associated to the split\r\n--> 652 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 653 except OSError as e:\r\n 654 raise OSError(\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 990 writer.write(example, key)\r\n 991 finally:\r\n--> 992 num_examples, num_bytes = writer.finalize()\r\n 993 \r\n 994 split_generator.split_info.num_examples = num_examples\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in finalize(self, close_stream)\r\n 407 # In case current_examples < writer_batch_size, but user uses finalize()\r\n 408 if self._check_duplicates:\r\n--> 409 self.check_duplicate_keys()\r\n 410 # Re-intializing to empty list for next batch\r\n 411 self.hkey_record = []\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in check_duplicate_keys(self)\r\n 347 for hash, key in self.hkey_record:\r\n 348 if hash in tmp_record:\r\n--> 349 raise DuplicatedKeysError(key)\r\n 350 else:\r\n 351 tmp_record.add(hash)\r\n\r\nDuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 0\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.7.9\r\n- PyArrow version: 3.0.0\r\n","comment_length":13,"text":"DuplicatedKeysError on personal dataset\n## Describe the bug\r\nEver since today, I have been getting a DuplicatedKeysError while trying to load my dataset from my own script.\r\nError returned when running this line: `dataset = load_dataset('\/content\/drive\/MyDrive\/Thesis\/Datasets\/book_preprocessing\/goodreads_maharjan_trimmed_and_nered\/goodreadsnered.py')`\r\nNote that my script was working fine with earlier versions of the Datasets library. Cannot say with 100% certainty if I have been doing something wrong with my dataset script this whole time or if this is simply a bug with the new version of datasets.\r\n\r\n## Steps to reproduce the bug\r\nI cannot provide code to reproduce the error as I am working with my own dataset. I can however provide my script if requested.\r\n\r\n## Expected results\r\nFor my data to be loaded.\r\n\r\n## Actual results\r\n**DuplicatedKeysError** exception is raised\r\n```\r\nDownloading and preparing dataset good_reads_practice_dataset\/main_domain (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/good_reads_practice_dataset\/main_domain\/1.1.0\/64ff7c3fee2693afdddea75002eb6887d4fedc3d812ae3622128c8504ab21655...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nDuplicatedKeysError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 dataset = load_dataset('\/content\/drive\/MyDrive\/Thesis\/Datasets\/book_preprocessing\/goodreads_maharjan_trimmed_and_nered\/goodreadsnered.py')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, **config_kwargs)\r\n 749 try_from_hf_gcs=try_from_hf_gcs,\r\n 750 base_path=base_path,\r\n--> 751 use_auth_token=use_auth_token,\r\n 752 )\r\n 753 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 573 if not downloaded_from_gcs:\r\n 574 self._download_and_prepare(\r\n--> 575 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 576 )\r\n 577 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 650 try:\r\n 651 # Prepare split will record examples associated to the split\r\n--> 652 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 653 except OSError as e:\r\n 654 raise OSError(\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 990 writer.write(example, key)\r\n 991 finally:\r\n--> 992 num_examples, num_bytes = writer.finalize()\r\n 993 \r\n 994 split_generator.split_info.num_examples = num_examples\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in finalize(self, close_stream)\r\n 407 # In case current_examples < writer_batch_size, but user uses finalize()\r\n 408 if self._check_duplicates:\r\n--> 409 self.check_duplicate_keys()\r\n 410 # Re-intializing to empty list for next batch\r\n 411 self.hkey_record = []\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in check_duplicate_keys(self)\r\n 347 for hash, key in self.hkey_record:\r\n 348 if hash in tmp_record:\r\n--> 349 raise DuplicatedKeysError(key)\r\n 350 else:\r\n 351 tmp_record.add(hash)\r\n\r\nDuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\nFound duplicate Key: 0\r\nKeys should be unique and deterministic in nature\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.7.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.7.9\r\n- PyArrow version: 3.0.0\r\n\nYup, I indeed was generating duplicate keys. Fixed it and now it's working.","embeddings":[-0.1864345074,-0.0565234087,0.0734844357,0.2193849683,0.1425481141,0.0292863455,0.5267676711,0.2207098007,0.0588233471,0.0872558504,-0.1790747494,0.2924934328,-0.1353885084,-0.1258903295,0.3225165606,0.1644579619,0.0292240959,-0.0048958994,-0.1378866285,0.0562093109,-0.1669512391,0.2724630833,-0.0154245673,0.4006582499,-0.1122406051,0.0180170108,-0.1549707204,0.3373055458,0.1574178636,-0.5524030328,0.3723440766,0.115193449,-0.0585309304,0.6388737559,-0.000114377,0.089326188,0.3678568304,0.0726804435,-0.3368042707,-0.2682757378,0.0349326618,-0.0251707342,0.1887518764,-0.3304275572,0.0402530059,-0.229095608,-0.0011710926,-0.3105666935,0.3904740512,0.2292785496,0.2183040529,0.0367066115,-0.0755213723,-0.1794561446,0.1490657926,0.3316625059,0.1221522838,0.2524907887,0.2090585381,-0.1218607128,0.093560718,0.0922568589,-0.2326767594,0.2715856433,0.4658198953,0.1381968111,-0.2277673036,-0.0389199071,0.2810682356,0.4387474358,0.4561787844,-0.5366321206,-0.2995787263,-0.4711150825,0.1770800203,-0.2411703467,0.1227191463,-0.0223870464,-0.1665583104,0.2151415795,-0.1649234295,-0.0753314644,-0.1069493145,0.0315973312,0.0011806944,-0.1223172769,-0.049361892,0.271197468,0.1206141412,-0.1859199107,-0.0420606621,-0.0807006657,-0.062629737,0.0479095578,-0.1217092127,-0.0121265231,0.0719313547,-0.0356171615,0.4212903976,0.1260746419,0.1623898894,0.0511264056,0.0721501485,0.0158360321,0.4798832834,0.0257830061,0.0838383511,0.253839016,0.2858423293,0.1228547394,-0.1708331406,-0.0057329764,-0.037885841,0.0616610274,0.5174378753,-0.0436327457,0.700039506,-0.2715612054,-0.2010089159,0.0989135578,0.0076636784,-0.1383434385,-0.0781496763,0.1666328758,0.0921997353,0.0152029637,-0.180369392,0.0483593643,-0.2964209914,-0.1243328527,-0.1058553755,0.090281494,0.0710753873,-0.030139178,0.32089746,-0.4846819341,0.1354746073,0.0996275842,0.022484228,-0.3461916447,0.1496575475,0.0179346725,-0.1550012827,0.0956206843,0.1365298182,0.372443825,0.0551547557,-0.2492426038,-0.1250035167,0.1580288708,-0.3686243892,-0.4084016979,-0.1084923893,0.223093614,-0.2670642436,-0.1035706699,-0.1620784849,0.055188071,0.3332821131,-0.1818750203,-0.0662663952,-0.2544150054,-0.4359380305,-0.2930590808,-0.0488225035,0.627838254,-0.3001243174,-0.0828407332,-0.5352852345,-0.1706262529,-0.2267044485,0.268750906,-0.3589932323,0.5027720928,-0.3484795094,0.2010123581,0.0668412521,-0.3456726968,-0.6580873132,0.5311324,-0.2746317685,0.327347517,0.0416119695,0.1276919842,0.4651711285,-0.1233073696,0.1120899841,-0.1775349528,0.0862104595,-0.1243482232,-0.0574253239,-0.1677972376,0.050897643,0.0288599618,0.0928659365,-0.2194238007,0.2983483076,0.0872548148,0.2812464535,-0.2383116633,-0.0349764228,0.1662206501,0.1789264679,0.0728690475,0.0283241384,-0.1941446811,-0.4341951609,0.3175955117,0.0040579163,-0.3278971016,-0.2090773284,-0.1817663461,-0.4485370219,0.0098392963,-0.5775942206,0.0009188339,0.0714324936,0.2051878422,-0.0812789798,0.0258612875,-0.1853131205,0.4073454738,0.1822107881,0.0808900222,-0.2352645993,0.0039496641,-0.0226734448,0.0041180314,0.0736214295,0.0747316927,0.3385287821,-0.0815312937,-0.1256201267,0.4060436487,0.064183481,0.1542711258,-0.187399596,-0.1152945533,0.1179354936,-0.154614687,0.3159890771,0.1982626766,0.2679871917,-0.1654774994,-0.3924897015,0.3851958513,0.0091077089,-0.1224951595,-0.1656302661,-0.0794710889,0.0632036924,-0.2009999901,-0.0268606432,-0.2052624524,0.2908895612,-0.2018519938,0.0799277723,0.1402326524,-0.1887188554,0.1398288161,0.1679102331,-0.1818193793,0.2790426016,0.1638421267,-0.0441890433,0.0209840108,-0.0386851579,0.2999103069,0.5103562474,0.153732866,-0.0037189524,0.2641640007,-0.1228415072,-0.0646803752,0.0696571469,-0.0417983532,0.0226826686,0.2188365459,0.0694565699,0.0686904863,-0.2538977861,0.0455449224,0.0083703725,0.3561425507,-0.4474920034,0.1444788277,-0.5289639831,-0.0410224162,-0.2909207344,0.1154117361,0.068720296,-0.1183312982,-0.0059037218,0.3284442425,0.0763666332,-0.0614555366,-0.2195770591,0.0737695545,-0.1509956568,-0.0838600397,0.0374967493,-0.1066306606,-0.457036078,0.0393652692,0.4354344904,-0.0755674616,0.4688960612,0.1013907269,-0.0867710486,-0.5481936932,-0.1828355342,-0.1221205518,-0.0135880848,0.1786353141,0.2338028997,0.0347757079,0.1136032194,-0.0659089908,0.0344041139,-0.1430894136,-0.2460508496,0.1748118997,-0.1032154411,-0.0170347877,-0.0606005527,-0.4004840553,-0.2146552801,0.0589309596,0.2180157304,-0.0278810523,-0.0025197237,0.1291808933,-0.0794784054,-0.0344787128,0.0746449903,-0.1346056461,-0.3363359571,-0.1720222831,0.2224801928,-0.14923127,-0.345107913,-0.0564711317,-0.0509004332,0.0318320617,-0.1824121773,-0.7195866704,-0.0192839764,-0.1149610728,0.2436561137,-0.2316603661,-0.1231254414,0.1163996905,0.0450868905,0.0219756514,-0.1702771932,-0.4202033281,-0.1165173352,-0.1416843086,0.0535265766,0.1716160625,0.5161118507,0.0215465147,0.3910152316,0.4786762893,-0.0518728308,0.1198170781,-0.0967197791,0.391258955,-0.1630542278,-0.5665510893,-0.0886632353,-0.1595169753,-0.0959168449,0.109044984,-0.141204074,-0.0564710759,0.0449788719,-0.1053584889,-0.4014692903,-0.2534804642,-0.1894110739,-0.0761737451,-0.1574556231,0.1937532574,0.2676413059,-0.1156040877,-0.0787375718,0.1312841326,0.0939033106,0.0195747502,0.0200156886,-0.3347143829,-0.0318388268,-0.5773205757,0.2624218762,0.0507724211,0.2812111676,-0.1805754006,0.2767280638,0.1589816064,-0.1883494258,0.3186620474,-0.1713173091,0.1666193902,0.2084120363,-0.3217483163,-0.3334866166,-0.1418097913,0.0553301051,0.3321937621,-0.1960080713,0.3390961885,-0.301961571,0.2847577631,0.4102033675,0.2261133343,-0.300919205,-0.457372576,0.0091470117,-0.6291176677,-0.4878902733,-0.0378224589,-0.0553398281,0.1563607454,-0.0442626253,0.0646786913,0.0115400823,0.0188315008,0.2603390217,0.1147316098,0.2303770781,0.0658604577,0.5187873244,0.1961691827,-0.0277723148,0.279868871,0.7565236688,0.001823833,-0.2737170458,-0.1594572216,-0.2390248626,-0.0713587999,0.2079110742,-0.1034442112,0.097773239,0.2249877602,-0.2202137411,-0.0234991498,-0.0852021053,0.1944959313,0.026329156,-0.4011048675,-0.3828482926,0.431596905,0.1299850792,-0.1661482453,0.3216854036,-0.1509178132,-0.0649240687,0.0862527639,-0.3026022315,0.5584703088,-0.2048691213,0.1103623211,0.1195008904,-0.4390955269,0.2995179892,-0.0239378847,-0.1661559939,-0.5218623281,-0.1000197008,0.0326341316,0.0657745972,0.096465975,0.0393105373,-0.0514230654,0.3023513556,0.0951553211,0.0253832638,0.0172300767,0.5617966652,0.0428265594,-0.0991237089,-0.0670165047,0.1566982865,-0.1801558584,0.2203921676,-0.0533136018,0.0197015628,-0.0324234664,-0.3704388142,-0.5147485733,0.0243518166,-0.0554232188,0.4853644371,-0.0828788206,-0.6532334685,0.1719492525,0.1229546815,0.238294512,0.0626022145,-0.2026261091,0.1107624397,-0.0813313201,0.2891862094,-0.2048490793,-0.244732067,0.1065135151,-0.0338678658,-0.411108166,-0.0314404815,0.1727112681,-0.1459559053,-0.1441899985,0.1145526171,0.292016238,-0.297840625,-0.1786839515,-0.1561232209,0.0387242883,-0.2815293372,0.053031791,0.1902587116,0.0553611815,0.1686558872,-0.1481540203,-0.3456920981,-0.0585029162,0.5251168013,0.1614068002,-0.2164437175,0.5713971257,0.0784837157,-0.0110600414,-0.1443631798,0.3910468519,0.2197487801,-0.2600576878,0.075141646,-0.1700660735,0.4170729518,0.1211233214,-0.1060315445,0.090955168,-0.0373879746,0.01344804,-0.1337712705,-0.2923740745,-0.0398552343,0.0645568669,0.2733722925,0.3497540355,0.0700478256,0.021049032,0.0084096128,-0.2461220622,0.2840166092,0.3811727464,-0.1247937754,0.3689233959,0.0878285021,-0.1213931888,0.0827203244,0.1207188293,0.3267295659,-0.0182491578,-0.1973792762,-0.1536381543,0.1597957611,-0.0764238834,0.0101838158,-0.1052649394,-0.1975092739,0.042086117,-0.0829911605,0.2619949877,0.4193260372,0.1065906882,0.2137805521,0.2205275595,-0.0996406227,-0.183651045,0.2850447893,-0.2902155817,0.1308636069,0.1731794924,0.248304382,0.1261893064,0.0332860611,-0.3144943118,0.0719708949,0.148118481,-0.088149257,0.1994998455,-0.4802075922,0.0026393796,0.2147420198,0.1991143972,0.3138101995,-0.3521753848,-0.2660816312,0.0199558288,0.1040719971,-0.3379886448,-0.3375861943,0.3838552833,0.1304217577,-0.1627376974,0.3674614727,0.0118581876,-0.2065409273,-0.1925642937,0.0000249554,0.5280101299,-0.1580040455,0.1635504067,0.3557379842,-0.1377314925,-0.2449371964,0.1990891546,0.0457413979,0.2326608747,0.7291664481,-0.3192643821,0.0695519745,0.0388671346,0.2995016873,0.119696416,-0.5269600749,0.2326116413,0.0199155286,-0.1420402825,0.0734836832,0.3483227789,0.2425382435,-0.0950403214,0.0766014457,-0.3192046881,-0.0517552495,-0.1597227901,-0.1258331239,0.0142505532,-0.1257435232,0.026150303,0.095063895,-0.2226903141,-0.1025366336,0.2904983461,0.0164554939,-0.4882953167,-0.4008246064,0.1408111304,0.0700542629,0.077807039,-0.2178088725,0.134912163,0.4335056245,0.0613660887,0.258336544,0.2638992965,0.6102486849,0.1389959455,-0.0220020711,0.1325817704,-0.1553770155,0.0196851864,0.1498385221,0.1509149075,-0.1279217154,0.362713784,0.3473048508,0.1291383505,0.0603024252,0.2542678416,0.3459598124,-0.0217789169,-0.4354406297,0.1667446941,-0.2433710396,-0.1651654243,-0.1628522426,0.381146878,-0.2777805626,-0.2351034731,0.5976246595,0.3726883233,0.0739829391,-0.2082563788,0.0601051003,-0.1204657182,0.1670228392,0.2927010655,-0.1981873661,-0.2354176641,0.1743615419,-0.9540211558,0.3124175966,-0.5523523092,-0.0202428233,0.0663271919,0.2064768076,-0.1396538168,0.0777139068,0.3370631933,-0.0765122175,-0.142531693,0.2806809247,-0.1841120422,-0.2737287283,0.2567863464,-0.0319788978,0.0418113098,-0.2001496702,0.2216446996,-0.2371189445,0.0011960795,-0.1737746745,-0.1126349494,0.1364224702,-0.1621296108,0.4371235967,0.1686959118,0.509370625,0.1869760454,-0.107384339,-0.3273393512,0.0053863325,-0.1892150342,0.5253096819,0.166409418,0.4103934765,0.0189346354,-0.1680552065,-0.1164797172,0.1793960184,-0.016118167,0.0582556091,-0.0924997479,0.1548611373,-0.0696645081,0.0682693645,-0.0576005727,0.2311602533,-0.0504845865,0.4076499641,-0.1823980957,-0.5174458623,0.5359555483,-0.3246198297,-0.1127242371,-0.2242871821,0.319124043,-0.2730090618,-0.0922168195,-0.439629972,0.1477435529,0.5595074296,-0.0903044045,0.0617006533,0.1447773129,-0.2487705648,0.1322369277,0.0044943751,0.2902173102,-0.00843082,-0.1849803627,0.4908718169,-0.1793061048]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2440","title":"Remove `extended` field from dataset tagger","comments":"The tagger also doesn't insert the value for the `size_categories` field automatically, so this should be fixed too","body":"## Describe the bug\r\nWhile working on #2435 I used the [dataset tagger](https:\/\/huggingface.co\/datasets\/tagging\/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:\r\n\r\n```\r\ndataset_name = 'arcd'\r\n\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\n def test_changed_dataset_card(dataset_name):\r\n card_path = repo_path \/ \"datasets\" \/ dataset_name \/ \"README.md\"\r\n assert card_path.exists()\r\n error_messages = []\r\n try:\r\n ReadMe.from_readme(card_path)\r\n except Exception as readme_error:\r\n error_messages.append(f\"The following issues have been found in the dataset cards:\\nREADME:\\n{readme_error}\")\r\n try:\r\n DatasetMetadata.from_readme(card_path)\r\n except Exception as metadata_error:\r\n error_messages.append(\r\n f\"The following issues have been found in the dataset cards:\\nYAML tags:\\n{metadata_error}\"\r\n )\r\n \r\n if error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset cards:\r\nE YAML tags:\r\nE __init__() got an unexpected keyword argument 'extended'\r\n\r\ntests\/test_dataset_cards.py:70: ValueError\r\n```\r\n\r\nConsider either removing this tag from the tagger or including it as part of the validation step in the CI.\r\n\r\ncc @yjernite ","comment_length":18,"text":"Remove `extended` field from dataset tagger\n## Describe the bug\r\nWhile working on #2435 I used the [dataset tagger](https:\/\/huggingface.co\/datasets\/tagging\/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:\r\n\r\n```\r\ndataset_name = 'arcd'\r\n\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\n def test_changed_dataset_card(dataset_name):\r\n card_path = repo_path \/ \"datasets\" \/ dataset_name \/ \"README.md\"\r\n assert card_path.exists()\r\n error_messages = []\r\n try:\r\n ReadMe.from_readme(card_path)\r\n except Exception as readme_error:\r\n error_messages.append(f\"The following issues have been found in the dataset cards:\\nREADME:\\n{readme_error}\")\r\n try:\r\n DatasetMetadata.from_readme(card_path)\r\n except Exception as metadata_error:\r\n error_messages.append(\r\n f\"The following issues have been found in the dataset cards:\\nYAML tags:\\n{metadata_error}\"\r\n )\r\n \r\n if error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset cards:\r\nE YAML tags:\r\nE __init__() got an unexpected keyword argument 'extended'\r\n\r\ntests\/test_dataset_cards.py:70: ValueError\r\n```\r\n\r\nConsider either removing this tag from the tagger or including it as part of the validation step in the CI.\r\n\r\ncc @yjernite \nThe tagger also doesn't insert the value for the `size_categories` field automatically, so this should be fixed too","embeddings":[-0.0470579714,-0.0278825946,-0.0063311323,0.2230904996,0.3980685771,0.3846163154,0.3813194633,0.203803122,0.133192867,0.3070672452,0.3498389125,0.3464969993,-0.3376166523,0.0101224994,-0.1240958795,0.0616643392,0.1870365888,-0.0848196596,0.2828492224,0.1796520501,-0.2078076452,-0.1280446649,-0.015804451,-0.2396082878,-0.2903832495,0.0013581972,-0.1328539848,-0.0206578374,-0.2535718977,-0.6862009168,0.4132631421,0.0345202722,-0.0556321964,0.1361313462,-0.0001189538,-0.0852048844,0.1941103041,-0.0051044491,-0.1610985994,-0.294326067,0.0116299735,-0.0132540064,0.0771376863,-0.1406927407,0.2041279078,-0.2700171769,-0.2972790003,-0.2101826072,-0.0064122593,0.3029443622,0.1134225354,-0.1724954247,-0.2265181392,-0.2223748863,0.3100630343,0.1275651306,-0.1088298261,0.132264182,0.4971141517,0.3124465644,-0.0372097306,0.4784123003,-0.019173285,-0.0323127694,0.3561680019,-0.1500074863,0.3728245795,-0.3121429384,0.2381992489,-0.1095877811,0.3060843945,-0.4363087118,-0.1398069859,-0.6247736812,-0.0318108983,-0.483725518,0.3642436564,-0.1988351792,-0.0016445724,-0.1294550151,-0.2243828475,-0.2147236317,-0.0542467237,0.0448021181,-0.3654717207,0.4700638652,0.1729986668,0.0621889867,0.1544353217,-0.0747496262,-0.0489725061,0.085245356,-0.2398613393,0.0857923329,-0.0082028601,-0.0928957537,-0.0311395042,-0.0814366192,0.1322938055,0.3225845695,-0.2078666985,-0.0327617377,0.2405249625,0.1316328347,0.310002923,0.03076474,0.0436907746,0.3443269432,0.0164351966,0.1438977867,-0.0169156455,-0.1115257591,0.0808673874,-0.0741375759,0.4703764319,0.1669950783,0.2705390751,-0.2235886604,-0.2313901931,0.1506878585,-0.3144266307,0.1037450284,0.1048623845,0.1004370973,0.0077369781,-0.0336846113,-0.3246629536,0.3663532734,0.1444291621,0.0521876365,-0.0436124839,0.1409448832,-0.0900342688,-0.1279349923,0.3507840335,-0.0277270097,0.2396265715,0.1880385727,0.1474274248,0.0099657355,-0.1440269798,-0.1226906851,0.2804158926,0.3793877363,-0.0323032402,0.1165248752,0.2214988768,-0.2776072323,-0.1801417917,0.1413510591,-0.2764790952,-0.5645540357,-0.1676930487,0.2017687261,-0.0390372798,-0.0194359347,-0.2403202951,-0.0538733937,0.3568716943,-0.0461056493,0.1374583244,-0.3215439618,-0.0683542639,-0.0430002883,0.0962239876,0.300490737,-0.0930942744,-0.1085047349,-0.5604650378,0.2998743951,-0.0664228797,-0.2599172592,-0.3118373752,0.1593887359,-0.3636178672,0.2771801054,0.362359345,-0.5000998378,-0.2899017036,-0.0787229538,-0.1293275058,0.4150394201,0.0940617323,-0.1839195937,0.2706879973,-0.2614488006,-0.1139896661,-0.2425147444,-0.4982540011,-0.1073555425,-0.1978641599,-0.0011880208,0.3833704889,0.064986445,0.3685127497,-0.2640117407,0.1988302022,0.2539690435,0.0429284386,-0.1136466414,0.049541939,0.189852342,0.2255969793,-0.0586890019,0.1237284616,-0.2138512284,-0.4337544143,0.1700278819,-0.1862152517,0.0550416559,-0.4036817253,-0.4727565646,-0.337982893,0.0708444566,-0.2964731157,-0.2775361538,0.0362860039,-0.0333929248,-0.0529800169,-0.0850006193,0.0008828238,-0.1241720989,-0.3564129174,0.3598924279,0.0580932833,-0.1835439503,-0.1832258999,-0.0699560642,-0.2722539604,0.158375293,0.2830740213,-0.1735773236,-0.213320002,0.2277534455,0.3332911432,-0.0209884644,-0.2749373615,0.539573729,0.3640347123,0.0174277853,-0.1698105782,0.1185639724,-0.1406784505,-0.1380033791,-0.2306803763,0.0682132393,-0.2076488435,0.330758363,-0.0102807237,-0.1152118817,0.1514413655,0.0575256161,-0.1673299521,-0.3952556849,-0.1256975383,-0.2747526765,-0.0283359587,0.0621759482,0.1700967103,0.3082009554,0.4491210282,0.4563256502,-0.1045935601,0.1277947277,-0.1929861009,-0.0935542211,0.1554531008,0.3781324625,0.2799788415,0.1411229819,0.0164398514,0.3384657502,0.1400482208,-0.1087927446,0.4774755239,0.3182540536,0.1658418924,0.4758277833,0.3180497587,0.1766327769,-0.1152629033,-0.0490471795,0.100695841,0.2709631026,-0.294903785,0.2232638896,-0.4055165648,-0.3722479343,-0.0678367019,-0.4162577689,-0.3396762908,-0.5190808177,0.3021348715,0.1177894697,-0.3278955519,0.1806784123,-0.247670874,-0.0960876867,0.0339501053,-0.1522390842,-0.121889703,0.0645907149,-0.2581368983,-0.0085340953,0.1556768566,-0.4164269865,0.4039628506,-0.1176032275,0.1496205479,-0.3149619102,-0.7408567667,0.2484392524,-0.0770198628,0.1834637523,0.4292975664,-0.3148034513,0.0215790421,-0.2592755854,0.2269996107,0.0908662155,0.1376240551,0.0316740908,0.0850266889,0.1913524568,-0.3398470879,-0.539755106,-0.2005249709,-0.1603257805,-0.1500421762,0.1650024354,0.0840091854,0.4994533658,-0.064620696,-0.2609850466,-0.2369151264,0.0092040608,-0.0938684493,-0.0743312463,0.1962911189,-0.0398308374,-0.0746402442,0.3503736556,-0.139124766,0.5066404939,0.12604177,-0.2408871949,-0.1568250954,0.3411068916,0.3540205956,-0.0566255637,-0.3169088364,0.3661273122,0.2036691457,0.0364861041,-0.0601680465,0.1416758746,0.1864183247,-0.1816585511,0.2106060982,-0.3667785823,0.0976002589,0.3018325865,0.3059273362,-0.1678301394,-0.0497624427,0.2520911098,-0.2006797791,0.0213903189,-0.1348640472,-0.6103516817,-0.0721231923,-0.1914785653,-0.2737713158,0.1863114685,-0.1272969693,-0.0307635535,0.1952723414,0.1689474285,-0.0118310098,-0.3345033228,0.1618983001,-0.2411805093,0.2018144131,0.1747648418,0.2280922681,-0.2410244644,-0.1832772642,0.110542044,0.103289932,0.3081328273,0.0265364032,0.2159497589,-0.3218404055,-0.048848737,0.2912314236,0.3632239699,0.037826661,-0.0877951384,0.063919872,-0.0299534164,-0.0749516785,0.4396287799,-0.2327893525,-0.0144696254,0.3920862377,-0.1407586336,0.0403800867,-0.2063895166,-0.1154251993,0.1810028702,0.1622503102,0.4688838124,-0.3142159879,-0.0319625363,0.3033488393,0.2184576094,-0.1087774336,0.1308354139,-0.1025861204,-0.1923557967,-0.2805983126,0.1967998892,0.0857549682,0.3269478381,-0.0196066629,-0.0167713072,0.0317768343,-0.010498587,0.1638726741,0.264871031,0.313511461,-0.2611885369,0.2118148655,-0.2683932185,0.4072530568,0.7266278267,0.4322222173,0.0651960745,-0.1843228489,-0.0189701486,-0.3502708077,0.3898514807,0.1154389903,-0.1276160479,0.2575131059,-0.2796077132,0.0542521551,-0.1976775229,0.1481023431,0.3795391023,-0.173825413,-0.2594434023,-0.3678873479,0.0777111202,0.2639745772,-0.0066361502,0.3888726532,0.0390351824,-0.5234270096,0.1186384559,0.1450538784,0.8001855612,0.4121473432,-0.1915742755,0.194761157,0.2244767547,0.0911051035,-0.4601960182,0.1581682414,-0.271705538,-0.15344356,0.0277516507,0.0556967482,0.4720816612,0.110867314,-0.2156368196,0.0796130002,-0.2481335551,0.4691539705,-0.1593142599,0.185276553,-0.5069157481,-0.268512547,-0.1197809353,0.049115859,-0.0728325918,-0.146759674,-0.1670562178,-0.0262888651,0.0460174903,-0.4585576952,-0.1864122897,-0.2002387345,-0.1889805347,-0.1539574564,-0.0787985623,0.0619412027,0.3307195306,-0.2198083103,0.2294680327,0.2972361743,-0.1191086322,0.2214436084,-0.045634646,0.0829476342,0.2213918716,-0.1759573072,0.2480489165,-0.0622777194,0.1254355609,-0.0810699016,-0.1246331036,-0.1629860103,-0.1854172945,-0.1845380068,0.6051138043,-0.4712047577,0.0146682821,-0.2269564271,0.0826344416,-0.444041729,0.0536613576,-0.1614522189,-0.1165186912,0.1619338691,-0.2414563745,-0.2038701326,-0.1600064039,0.0554606579,0.047819607,-0.0963207856,0.3979774117,-0.0989366919,-0.1335577369,-0.2292313576,-0.0289920606,0.1888291091,-0.5756084919,0.10933844,0.0664856434,-0.0455095097,-0.0235738717,0.4011648595,0.086347349,0.1172612309,-0.0820661262,-0.3288559616,-0.2480509281,0.0923303813,0.0430359133,0.0985753536,-0.1008963585,0.0736204237,-0.143351838,0.1773617715,-0.2430761158,-0.0637702346,-0.0391498618,0.197244063,-0.1992156506,0.1110503301,-0.0232235081,-0.1965449452,-0.0029943788,0.1898087561,0.0057758596,-0.0791421682,-0.1569434106,0.2112522125,0.0204794556,0.1821600348,0.2190771401,0.1740813553,-0.0587721542,-0.1220609024,0.2485522628,-0.094421193,0.1258091033,-0.0343819484,0.2373393923,0.2272668928,-0.1735194325,-0.0632438958,-0.0681630448,0.3235395253,-0.0665058866,0.2477603555,-0.160002619,-0.0853480026,-0.0305449609,0.1990442723,-0.0561818182,-0.1165543273,0.3627637327,-0.2627104819,0.1160384342,0.3600439727,0.2566112876,0.16750063,-0.2431188971,-0.0144634657,-0.0252244081,0.1866099536,-0.2930409908,-0.147227034,0.1264133602,-0.1274361759,0.0823215246,0.522434175,0.6653687358,-0.0732866451,0.1759086996,0.0678957552,0.8685905933,-0.1464315206,0.3343615532,0.7550765872,-0.2798928022,0.1776983887,0.2474984676,0.05291152,0.137559548,0.3602353632,-0.2074296623,0.6691162586,0.5490683913,-0.3221603632,0.1135336235,-0.3197205663,0.2606502771,0.1739158183,0.2225123197,0.0861207172,0.1814705431,0.1143550053,0.0886542201,-0.4692407548,-0.1042803451,0.1961992532,-0.2581466436,0.1572608054,0.0136737647,-0.3153924644,-0.5290760398,0.0869102702,-0.0707683787,-0.1452488601,0.1395495087,-0.1513897479,-0.365554899,-0.1286621541,-0.063243039,0.0082281213,0.2547105849,-0.38957268,0.2708716094,0.3742643893,-0.2127828747,0.1764421761,0.1181961447,0.4805864394,0.1340564191,0.1719903052,-0.1956600398,-0.2507948577,-0.0836449564,0.0430110544,0.2725761235,-0.1612664014,0.1940100193,0.1253467798,0.0611021444,-0.1453901231,-0.0215725694,0.1518563479,0.1295015663,-0.075859122,0.2887374461,0.2826043963,0.0710288361,-0.1200045794,0.1136223599,-0.231961742,-0.2884580791,0.5546866655,0.3244502246,0.196227774,-0.3049543202,0.0690613016,0.1938880235,0.1592255682,0.2849515378,-0.3145399988,-0.2621354759,-0.0424961857,-0.6376393437,0.1383606642,-0.0609902851,-0.1454482824,0.4037163258,-0.0723091289,0.1364809126,0.0812711567,0.3085353673,0.2047010213,0.4365625083,0.1080796644,-0.0242800191,0.0164654534,0.1539180726,-0.1898455769,-0.0292333663,-0.2114117593,0.21178478,-0.4118198752,0.0192795917,-0.1006442085,-0.2203858942,0.3989229798,0.0903164148,0.49344033,0.0468414538,0.252085954,-0.0962051898,-0.1233733445,-0.3604350984,-0.060385175,-0.293746084,0.3651178181,-0.035745725,0.3513460159,-0.0554305092,-0.247226283,-0.369751513,0.0400843024,0.0363651104,-0.1157740727,-0.2941093445,0.2927540541,-0.1815777272,0.2877059877,0.0763626397,0.0394820943,0.1525593698,0.210079208,-0.1876696944,-0.5920581222,0.573780477,-0.1900604367,-0.3546431959,-0.4177166224,0.32766065,0.1319286823,-0.1886220425,-0.6175740957,-0.2775549889,0.4226591587,0.0249398835,-0.04034549,0.1850890964,0.0946728364,-0.0841679573,-0.0873373002,0.0166871864,0.0217028614,0.128986299,0.0918277949,-0.1965800375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2440","title":"Remove `extended` field from dataset tagger","comments":"Thanks for reporting. Indeed the `extended` tag doesn't exist. Not sure why we had that in the tagger.\r\nThe repo of the tagger is here if someone wants to give this a try: https:\/\/github.com\/huggingface\/datasets-tagging\r\nOtherwise I can probably fix it next week","body":"## Describe the bug\r\nWhile working on #2435 I used the [dataset tagger](https:\/\/huggingface.co\/datasets\/tagging\/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:\r\n\r\n```\r\ndataset_name = 'arcd'\r\n\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\n def test_changed_dataset_card(dataset_name):\r\n card_path = repo_path \/ \"datasets\" \/ dataset_name \/ \"README.md\"\r\n assert card_path.exists()\r\n error_messages = []\r\n try:\r\n ReadMe.from_readme(card_path)\r\n except Exception as readme_error:\r\n error_messages.append(f\"The following issues have been found in the dataset cards:\\nREADME:\\n{readme_error}\")\r\n try:\r\n DatasetMetadata.from_readme(card_path)\r\n except Exception as metadata_error:\r\n error_messages.append(\r\n f\"The following issues have been found in the dataset cards:\\nYAML tags:\\n{metadata_error}\"\r\n )\r\n \r\n if error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset cards:\r\nE YAML tags:\r\nE __init__() got an unexpected keyword argument 'extended'\r\n\r\ntests\/test_dataset_cards.py:70: ValueError\r\n```\r\n\r\nConsider either removing this tag from the tagger or including it as part of the validation step in the CI.\r\n\r\ncc @yjernite ","comment_length":42,"text":"Remove `extended` field from dataset tagger\n## Describe the bug\r\nWhile working on #2435 I used the [dataset tagger](https:\/\/huggingface.co\/datasets\/tagging\/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:\r\n\r\n```\r\ndataset_name = 'arcd'\r\n\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\n def test_changed_dataset_card(dataset_name):\r\n card_path = repo_path \/ \"datasets\" \/ dataset_name \/ \"README.md\"\r\n assert card_path.exists()\r\n error_messages = []\r\n try:\r\n ReadMe.from_readme(card_path)\r\n except Exception as readme_error:\r\n error_messages.append(f\"The following issues have been found in the dataset cards:\\nREADME:\\n{readme_error}\")\r\n try:\r\n DatasetMetadata.from_readme(card_path)\r\n except Exception as metadata_error:\r\n error_messages.append(\r\n f\"The following issues have been found in the dataset cards:\\nYAML tags:\\n{metadata_error}\"\r\n )\r\n \r\n if error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset cards:\r\nE YAML tags:\r\nE __init__() got an unexpected keyword argument 'extended'\r\n\r\ntests\/test_dataset_cards.py:70: ValueError\r\n```\r\n\r\nConsider either removing this tag from the tagger or including it as part of the validation step in the CI.\r\n\r\ncc @yjernite \nThanks for reporting. Indeed the `extended` tag doesn't exist. Not sure why we had that in the tagger.\r\nThe repo of the tagger is here if someone wants to give this a try: https:\/\/github.com\/huggingface\/datasets-tagging\r\nOtherwise I can probably fix it next week","embeddings":[0.0003140183,-0.0987946317,-0.0033113675,0.1912126392,0.4241888523,0.4057804346,0.3773061633,0.2321246117,0.1182753444,0.2808997929,0.3130059838,0.4294629097,-0.2918011546,0.0003099469,-0.0962812677,0.0519277453,0.2192724496,-0.0698820129,0.2905523479,0.1498899311,-0.186719507,-0.1042956784,-0.0217016246,-0.2240479439,-0.2687342763,-0.0041395938,-0.1522147208,-0.0004327623,-0.1924079955,-0.6609166265,0.3868137896,0.071211189,-0.0397313572,0.1408297569,-0.0001184678,-0.1125504524,0.2438776344,-0.0025301739,-0.1776151061,-0.2817538679,-0.0076940292,0.0383164957,0.1209352612,-0.1555958986,0.1453635097,-0.1953763962,-0.3327715695,-0.2186891437,0.1178630143,0.3451340199,0.1246328354,-0.0914979726,-0.2098382413,-0.1890344769,0.3394220471,0.0909312591,-0.0966884047,0.0902480483,0.5893580914,0.3097759187,0.0236048829,0.4658851326,-0.0466955267,-0.0535481274,0.3191026449,-0.1811038852,0.3747120798,-0.242455855,0.2049732953,-0.091123715,0.3223930895,-0.4056682885,-0.1556642652,-0.5504049063,-0.007573023,-0.5235106349,0.3982792199,-0.2334059775,-0.0243907496,-0.0912935212,-0.2531798184,-0.2212180197,-0.0475980751,0.0559128895,-0.3510925472,0.4075273275,0.1735143363,0.0653679967,0.1431018114,-0.0861212239,0.020751128,0.0846145377,-0.1744771153,0.0337960012,0.0334306881,-0.1315354258,-0.1024269983,-0.1374259442,0.1256537884,0.37399894,-0.2233776599,-0.0094907843,0.2022276223,0.145347923,0.343242377,-0.0394410305,0.0646753237,0.3392271698,0.0410912186,0.1974247396,-0.0468314588,-0.1100905985,0.0627237186,-0.0449791104,0.4439435005,0.1949736178,0.2984240651,-0.2926565707,-0.1970297247,0.0969917178,-0.3704855144,0.1451703757,0.1344015449,0.1267314255,0.05404374,-0.1025555655,-0.3116002679,0.3116007745,0.1424890906,-0.0113603882,-0.0340749249,0.153189078,-0.1016091183,-0.1060060188,0.4092148542,-0.0865495801,0.2645270228,0.1723595411,0.139774099,-0.0269341618,-0.0855332464,-0.0696397275,0.319128871,0.376521796,-0.0374684446,0.1472157985,0.2145492285,-0.2550094724,-0.1550869346,0.1358287483,-0.3036946654,-0.5740443468,-0.1560348868,0.1887000501,-0.0966336057,-0.0061226613,-0.194587782,-0.1367241293,0.3593397141,-0.0559064709,0.1324734092,-0.302239567,-0.1135979816,0.0077784802,0.1533000767,0.3194821179,-0.0298087504,-0.0656731501,-0.5532761216,0.2635629177,-0.1193991527,-0.2619855702,-0.3319779634,0.1100359559,-0.361199379,0.2333570719,0.278512001,-0.5548009872,-0.2303901762,-0.0377747454,-0.1724076718,0.4368152022,0.0398087837,-0.2358095646,0.2606691122,-0.3076125681,-0.1031483337,-0.2270416617,-0.472763896,-0.1627990901,-0.2732887864,0.0489266776,0.3657885492,0.0566968694,0.2494950742,-0.2655074,0.1501193792,0.2188585997,0.0700702593,-0.1165327206,0.066332899,0.2436310649,0.2496330291,-0.0021935413,0.1035938188,-0.1965005398,-0.4023182392,0.1305842251,-0.1783620119,0.0230578389,-0.48124066,-0.45917207,-0.3062449992,0.0376051329,-0.3534753621,-0.3354027569,0.0344671011,-0.0630825013,-0.0725845546,-0.0820656568,0.023466358,-0.1426420212,-0.3872588873,0.3070746958,0.0084980102,-0.1334138066,-0.1711049527,-0.04227373,-0.2389017195,0.1958137006,0.3363584578,-0.1897206604,-0.2188038081,0.2598940432,0.3583898842,-0.0879120901,-0.1984903067,0.5692500472,0.3428311348,-0.1001196802,-0.1322922409,0.138531357,-0.1558125615,-0.1278757155,-0.286218226,0.0895515457,-0.2382875681,0.3804147542,-0.0131929172,-0.0287683997,0.2370084226,0.0562688224,-0.2296647578,-0.429304719,-0.1230942458,-0.2335540205,-0.0564448833,0.0019826295,0.1389630288,0.3440558314,0.3959482014,0.4019793868,-0.1673779041,0.1014385596,-0.1910217106,-0.1044660136,0.1607940793,0.3864328563,0.2456977218,0.128486529,-0.0229314622,0.3320823908,0.1404837966,-0.110407874,0.4423740804,0.2740832269,0.1690316647,0.4438692629,0.2920733094,0.1619816571,-0.0570952408,-0.0997478217,0.1282672137,0.2701994479,-0.304061234,0.2250166982,-0.4346056879,-0.3808312714,-0.1063380837,-0.402180016,-0.3194605112,-0.5558412075,0.2902688682,0.0936841071,-0.3065008819,0.280662179,-0.1872084737,-0.1169866472,0.006309628,-0.1769814044,-0.1444192827,0.0145446789,-0.207355693,-0.0171891246,0.1901349276,-0.3280123472,0.3918944001,-0.0801410377,0.1511206329,-0.4546266198,-0.6928055286,0.2421106696,-0.1202883944,0.2134212404,0.4348098338,-0.3150678277,0.050928276,-0.2641372681,0.2383198291,-0.0089948224,0.1518588662,0.0307016578,0.0889758766,0.1576651633,-0.2906535566,-0.5436253548,-0.1608187854,-0.113581486,-0.1207598075,0.1297326535,0.0558198169,0.4889589548,-0.0584840141,-0.2023570389,-0.3087671399,-0.0078364406,-0.1023400053,-0.1442334056,0.1379031986,-0.0920272544,-0.1064266562,0.4192366302,-0.0954574198,0.5402063727,0.1076406017,-0.2199103832,-0.1565547138,0.3927101493,0.4241115451,-0.0569099598,-0.3841508031,0.3486684561,0.2105719,0.0443726331,-0.0648899153,0.1158821732,0.2074986249,-0.2082594186,0.2325932384,-0.3750937879,0.08650098,0.2408459336,0.3512165546,-0.1425565034,-0.0145412823,0.186694175,-0.2168249637,-0.0001610692,-0.1665429771,-0.5890359282,-0.1397545785,-0.1733062118,-0.2427230179,0.1731244773,-0.1461893469,-0.0375954434,0.177425012,0.1735090166,-0.0029557324,-0.3457013667,0.1309201121,-0.2558939159,0.2480551302,0.1915761977,0.2138014883,-0.2268821448,-0.194649443,0.1226844043,0.1072449386,0.3399111927,0.0026734485,0.198987335,-0.3184266686,-0.0341056399,0.303574264,0.3770940304,0.1147945225,-0.0863249898,0.0864196867,-0.0071301088,-0.0962148458,0.4647549391,-0.2693887949,0.0421697944,0.4391360879,-0.1976352185,0.0112480838,-0.2105627805,-0.1383235157,0.1764829755,0.0870646462,0.5664535761,-0.2283640951,-0.1016432047,0.3886175156,0.2248515487,-0.099574618,0.0920536667,-0.092640318,-0.2424954623,-0.2701031864,0.1623062342,0.0810568184,0.295846343,-0.0756679699,-0.0434921198,0.0718372241,0.0424280129,0.1050648093,0.2345249355,0.3128808141,-0.2771994472,0.2501500547,-0.2137081474,0.4078693986,0.6896817684,0.4112485647,-0.072214663,-0.1640152931,-0.0656770691,-0.4268177748,0.3582837284,0.1291314811,-0.1379517615,0.310828805,-0.2521000206,0.0059105675,-0.1566771716,0.095368363,0.3929443359,-0.099306725,-0.2628518343,-0.4332625568,0.0545970835,0.2735936344,0.008202713,0.3860192299,-0.0064823101,-0.4396151602,0.1277210116,0.1599190235,0.8890958428,0.3560686409,-0.1739722192,0.1735225767,0.2280184925,0.1016049087,-0.5014361739,0.1365305334,-0.2667913437,-0.2394099534,0.0321224332,0.0506548807,0.4719640911,0.0500566848,-0.2175472975,0.0857044607,-0.260022074,0.4788054228,-0.1253934801,0.222303763,-0.4344605207,-0.2390309423,-0.1349336952,0.0585309342,-0.0794541612,-0.1676386148,-0.1614966244,-0.0375652649,0.0528181493,-0.4639972746,-0.1672094762,-0.1268185526,-0.135450393,-0.0952777192,-0.0986088663,0.0675148219,0.3694373071,-0.2576726079,0.2425047606,0.2825701535,-0.16631338,0.2520093024,0.0206613969,0.0080710268,0.1592754275,-0.1857250482,0.2694397569,-0.1088746935,0.0810611248,-0.0553093217,-0.1282543987,-0.1588982344,-0.1465991437,-0.1983085424,0.6032127142,-0.4415479302,0.0166257247,-0.1621663868,0.0522065647,-0.4735493958,0.0747748166,-0.1701308638,-0.179458335,0.1120443642,-0.2498127669,-0.2068417519,-0.1144918278,0.0962016433,0.0242827348,-0.1313823611,0.4699946642,-0.0377028435,-0.1012934819,-0.2787187397,-0.0280088298,0.1924147457,-0.616658628,0.1606228799,0.1137121543,0.051764708,-0.0011957788,0.4517729282,0.0332446247,0.1049059182,-0.0914536342,-0.3305730224,-0.2469259948,0.0511205196,0.0127127031,0.0842094645,-0.0904360488,0.1768889427,-0.1297350675,0.1352158487,-0.265419215,-0.0056980164,-0.0873777568,0.2000610381,-0.1278535128,0.0612977892,-0.0018428336,-0.1975076646,0.018512208,0.1883536577,-0.0046709627,-0.0946779102,-0.1496195197,0.2090851963,0.0016977144,0.1834045798,0.2538028061,0.1821084768,-0.079822965,-0.1073185429,0.2055536658,-0.0453776047,0.1551739872,0.0124377767,0.2630967796,0.2316503525,-0.2443834543,-0.0706602931,-0.0822778866,0.2896994352,-0.0646962747,0.206053555,-0.2416512966,-0.0996076837,-0.0065178215,0.2077780813,-0.1156988516,-0.0670349449,0.3926126361,-0.2521973848,0.1110124961,0.3472044468,0.2288534641,0.2076355964,-0.3032678664,-0.0264956914,-0.0016696894,0.1850781143,-0.2708517313,-0.1050333455,0.1775525212,-0.1248033643,0.058413133,0.4639509916,0.6760467887,-0.1183538213,0.1980180144,0.0746130198,0.8444798589,-0.2484196872,0.3639233708,0.7368881106,-0.2064301819,0.1531536877,0.1691211164,0.0838205293,0.09748815,0.3269822896,-0.2433440536,0.6369713545,0.5483292937,-0.280674696,0.1314940602,-0.3452028632,0.3214640021,0.1574708372,0.2007051408,0.1241414845,0.1072484404,0.2155112624,0.0706408694,-0.4382077456,-0.1155315489,0.2446926534,-0.218302384,0.1294466853,0.0293901097,-0.2767006457,-0.550752759,0.1026754826,-0.0600919612,-0.1193011031,0.1065624952,-0.1250668019,-0.3660973907,-0.1628534645,-0.0241947174,0.0017169947,0.2560476065,-0.3403729498,0.250041455,0.3928289711,-0.1883889139,0.1698669046,0.0363670513,0.4413490295,0.0943279117,0.2049486786,-0.1747897714,-0.2660327256,-0.0579445064,0.1427544951,0.3077872694,-0.1079170555,0.2048006654,0.1520262212,0.0630788654,-0.138006568,-0.0514989123,0.1939499378,0.2117455006,-0.0383965336,0.2215190679,0.3344192803,0.0279595759,-0.1896013021,0.1348512322,-0.258495003,-0.26113832,0.5472016335,0.3617829382,0.1623189449,-0.2242135704,0.0737819746,0.1545419693,0.1596466601,0.2870761752,-0.2660982311,-0.3241258562,-0.086848639,-0.5948204994,0.1610307246,-0.0657766834,-0.1399641931,0.3689259291,-0.0742445365,0.0975769535,0.0477713495,0.2893304527,0.1838389337,0.3929618597,0.0817233175,-0.0025130555,0.0164841339,0.2030582726,-0.1696239263,-0.0084994333,-0.1675106734,0.1412529647,-0.3266453445,0.0127197895,-0.0809665844,-0.2827871144,0.3916544318,0.0767508,0.4173844457,0.0474376231,0.2677125037,-0.0613669194,-0.1450713277,-0.3603590727,-0.0893696919,-0.2728504539,0.3768959641,0.0206657834,0.2532949746,-0.0403604098,-0.1949863732,-0.405608654,0.0496129058,0.0149269355,-0.1289158762,-0.3502374887,0.2659265995,-0.1478919536,0.3055540621,0.0678815022,0.0693785399,0.158433944,0.1978629082,-0.2647052705,-0.505902648,0.647875607,-0.1815772355,-0.3230546713,-0.3615927994,0.3531632721,0.0831717104,-0.1900714934,-0.6228551269,-0.3255912066,0.4235510826,0.0198500827,-0.0901093557,0.2049212456,0.0926014483,-0.0225007422,-0.079733476,0.0000323311,0.0759062469,0.1838484854,0.0889564455,-0.1694913656]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2440","title":"Remove `extended` field from dataset tagger","comments":"I've opened a PR on `datasets-tagging` to fix the issue \ud83d\ude80 ","body":"## Describe the bug\r\nWhile working on #2435 I used the [dataset tagger](https:\/\/huggingface.co\/datasets\/tagging\/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:\r\n\r\n```\r\ndataset_name = 'arcd'\r\n\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\n def test_changed_dataset_card(dataset_name):\r\n card_path = repo_path \/ \"datasets\" \/ dataset_name \/ \"README.md\"\r\n assert card_path.exists()\r\n error_messages = []\r\n try:\r\n ReadMe.from_readme(card_path)\r\n except Exception as readme_error:\r\n error_messages.append(f\"The following issues have been found in the dataset cards:\\nREADME:\\n{readme_error}\")\r\n try:\r\n DatasetMetadata.from_readme(card_path)\r\n except Exception as metadata_error:\r\n error_messages.append(\r\n f\"The following issues have been found in the dataset cards:\\nYAML tags:\\n{metadata_error}\"\r\n )\r\n \r\n if error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset cards:\r\nE YAML tags:\r\nE __init__() got an unexpected keyword argument 'extended'\r\n\r\ntests\/test_dataset_cards.py:70: ValueError\r\n```\r\n\r\nConsider either removing this tag from the tagger or including it as part of the validation step in the CI.\r\n\r\ncc @yjernite ","comment_length":11,"text":"Remove `extended` field from dataset tagger\n## Describe the bug\r\nWhile working on #2435 I used the [dataset tagger](https:\/\/huggingface.co\/datasets\/tagging\/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:\r\n\r\n```\r\ndataset_name = 'arcd'\r\n\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\n def test_changed_dataset_card(dataset_name):\r\n card_path = repo_path \/ \"datasets\" \/ dataset_name \/ \"README.md\"\r\n assert card_path.exists()\r\n error_messages = []\r\n try:\r\n ReadMe.from_readme(card_path)\r\n except Exception as readme_error:\r\n error_messages.append(f\"The following issues have been found in the dataset cards:\\nREADME:\\n{readme_error}\")\r\n try:\r\n DatasetMetadata.from_readme(card_path)\r\n except Exception as metadata_error:\r\n error_messages.append(\r\n f\"The following issues have been found in the dataset cards:\\nYAML tags:\\n{metadata_error}\"\r\n )\r\n \r\n if error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset cards:\r\nE YAML tags:\r\nE __init__() got an unexpected keyword argument 'extended'\r\n\r\ntests\/test_dataset_cards.py:70: ValueError\r\n```\r\n\r\nConsider either removing this tag from the tagger or including it as part of the validation step in the CI.\r\n\r\ncc @yjernite \nI've opened a PR on `datasets-tagging` to fix the issue \ud83d\ude80 ","embeddings":[-0.0320098661,-0.0937643871,-0.0110269468,0.19322294,0.4230774045,0.3770270348,0.3622706831,0.1902802736,0.1291031688,0.2867848873,0.3256170452,0.4400387704,-0.3483436108,0.007049731,-0.1155106798,0.0442021862,0.2011447847,-0.0742028505,0.27200827,0.1735339314,-0.2071188837,-0.1316568553,-0.0475848392,-0.2151355892,-0.2691774368,-0.0407511517,-0.0967888907,-0.0263973102,-0.1682283431,-0.6403113604,0.377333343,0.0350419134,-0.0522250086,0.1123930663,-0.0001182011,-0.0747788921,0.20648247,0.0014024153,-0.2296306342,-0.2901522815,-0.0194802787,0.0727964863,0.083451204,-0.149478808,0.14069134,-0.2327216119,-0.3207336068,-0.2701170146,0.0572776534,0.3396079242,0.1286982596,-0.069569014,-0.2590894699,-0.2201712728,0.2914168537,0.0780351758,-0.086377345,0.0863597095,0.5600705743,0.3074058294,-0.0191553123,0.4982021153,-0.0498550646,-0.0701460615,0.3103295565,-0.180755645,0.366032213,-0.2491345853,0.2572181225,-0.1329858899,0.2995718122,-0.3957487047,-0.1483563036,-0.5467221141,-0.0117409462,-0.4626344144,0.3749768138,-0.2789793611,-0.0492298044,-0.1145552993,-0.2554495931,-0.2553431988,-0.0299690478,0.0495613068,-0.3395854831,0.4150592387,0.1821377426,0.0835222751,0.1143762097,-0.0600792505,-0.0116304411,0.0697044581,-0.1638213396,0.0606055111,0.0255110878,-0.0799348056,-0.0403110161,-0.114714019,0.1158881038,0.3459371924,-0.2217330039,-0.0123612741,0.2420717627,0.1459457725,0.2996522188,-0.0361040831,0.0793564767,0.3696037233,0.0368547179,0.1771431565,-0.0720298812,-0.1178867444,0.0616023131,-0.0365053974,0.4414254129,0.223126173,0.3199301362,-0.2423403561,-0.2510874569,0.1252715886,-0.3584168255,0.1180083305,0.161966145,0.072210446,0.0639805123,-0.0952699631,-0.3154793382,0.3288344443,0.1268927604,0.0000795771,-0.0350162983,0.1344118714,-0.1103807837,-0.1260101497,0.3652181923,-0.0959346369,0.2694652081,0.1808016151,0.1519821733,-0.013807632,-0.104527548,-0.0909447223,0.3029901087,0.3666310608,-0.0500145592,0.1625397503,0.2221072912,-0.2666473687,-0.1571682096,0.1237380877,-0.2765294313,-0.5819149613,-0.141705066,0.202510342,-0.0360600613,-0.0044433256,-0.2196554542,-0.1100438684,0.3617346883,-0.0546459667,0.1380179822,-0.3108513951,-0.0583009571,-0.0136126522,0.1555989832,0.2820190191,-0.032139983,-0.1080522314,-0.5058831573,0.2444759011,-0.0916973427,-0.2325035632,-0.358289063,0.1503871977,-0.3794141114,0.2267109603,0.2776651382,-0.5416305661,-0.2368653864,-0.0472650416,-0.1780878901,0.424099952,0.0482414179,-0.2168157995,0.2884165645,-0.2665423453,-0.0994042233,-0.1953285486,-0.4932041466,-0.1366241574,-0.2401223183,0.048920285,0.4064281583,0.0409246609,0.3124667406,-0.2493998557,0.1414117068,0.2439652979,0.0491696298,-0.1284169257,0.0837283731,0.180858776,0.2482501566,-0.0286910925,0.104937695,-0.2174370736,-0.4376920462,0.1303619891,-0.1424686462,0.0466147438,-0.4977133274,-0.472730577,-0.2948199809,0.0556252562,-0.3475441635,-0.3413008749,0.0507688709,-0.0499410927,-0.0677777454,-0.1185079142,0.018837316,-0.1784024835,-0.3634283841,0.3370849788,0.0049464675,-0.1501518786,-0.1683347076,-0.0494851246,-0.2717081308,0.1706851423,0.311000824,-0.1825943887,-0.2070895582,0.2146307379,0.3356451094,-0.0289794598,-0.2008900195,0.5149354935,0.3177017868,-0.1016400531,-0.16792202,0.1287610829,-0.1516993195,-0.1305071414,-0.2382172942,0.0778784677,-0.2249399871,0.3994395435,-0.0079522952,-0.0455342717,0.2107138783,0.0447723754,-0.1938901693,-0.4042080641,-0.1145448908,-0.2650676966,-0.0137771778,0.0336015783,0.1337542236,0.3095012009,0.3590165675,0.4295859337,-0.1792279929,0.104778789,-0.1892232001,-0.093173191,0.1511626989,0.3702284992,0.2388580739,0.1312263608,0.0123878988,0.3343346119,0.1477113515,-0.0878835246,0.453586787,0.3005019426,0.1714978218,0.4850888848,0.2569748461,0.1725996882,-0.0994604379,-0.0977629796,0.1425180733,0.2894711792,-0.2897137105,0.2342165858,-0.4268131852,-0.3682851195,-0.0790216997,-0.4025772512,-0.3061148226,-0.532530427,0.2797587514,0.0894918516,-0.329638809,0.2667110264,-0.213867113,-0.1255255938,0.0223198961,-0.2212658972,-0.1316394955,0.0321168788,-0.1843409538,-0.0013521706,0.1675419956,-0.3621970713,0.422965318,-0.0881995037,0.1190701872,-0.4410029948,-0.677088201,0.2451714575,-0.1175453737,0.2005646378,0.469116658,-0.3100031912,0.0788096637,-0.2881054878,0.2457691431,0.0527252108,0.146230042,0.0172326155,0.0868152156,0.1990579367,-0.2958247662,-0.5580770969,-0.1557831764,-0.1196798012,-0.175099656,0.089340657,0.0668268874,0.4803521931,-0.0671280995,-0.2321808934,-0.2774071395,0.0092710331,-0.1052990332,-0.0972178951,0.1232296377,-0.0296869334,-0.0609119907,0.3611297607,-0.1126603931,0.5347832441,0.1851719469,-0.1922099441,-0.1545419693,0.3898555636,0.4409189522,-0.0649071783,-0.3786566556,0.3523936272,0.2157933265,0.039013125,-0.0542271733,0.1085305959,0.1656885594,-0.2449733615,0.2290465087,-0.362138778,0.1338315457,0.2665470839,0.3398310244,-0.1069592908,-0.0275909007,0.1969534159,-0.2092139423,-0.0186661277,-0.1466377378,-0.6128413081,-0.1463774443,-0.2135242373,-0.31770733,0.1444468498,-0.1416952759,-0.0028891722,0.1840176284,0.1629431695,-0.0300007537,-0.3352801204,0.1621931493,-0.2777743042,0.2791376114,0.1961392313,0.2394621074,-0.2211411595,-0.134958744,0.1408403516,0.1108187437,0.3649988174,0.0001000636,0.1832244396,-0.3261752129,-0.0484689996,0.2791549861,0.3597432673,0.1182011887,-0.0933854729,0.0798087716,0.0080473153,-0.0890379995,0.4644637704,-0.2546453178,0.0191494785,0.3981511295,-0.1905373931,0.0822159424,-0.2051461339,-0.1430037916,0.2013113499,0.104025729,0.5999001265,-0.2841422856,-0.0675628632,0.3493292034,0.2176429182,-0.0783234462,0.1188864484,-0.0733054429,-0.2363784909,-0.2695176005,0.1824834198,0.0858407617,0.3010356724,0.0097735645,-0.004760399,0.0839342698,0.0473065823,0.1255203336,0.2291854471,0.2837549746,-0.2789385319,0.2580251992,-0.2297008932,0.3700134456,0.7422603369,0.4000979662,-0.0285873115,-0.1785421073,-0.0875912383,-0.3678397238,0.3708818555,0.0804524496,-0.1493668109,0.3331658244,-0.2330316752,0.0214250647,-0.1824967265,0.1053785309,0.3917914331,-0.0998148024,-0.289211899,-0.3875777125,0.0473058596,0.2857320309,-0.0230440944,0.3781681657,0.00083971,-0.5162772536,0.1284729689,0.1597513407,0.7705391049,0.3863981366,-0.1855426282,0.2050935775,0.1972079575,0.1104643717,-0.4401886165,0.1150053814,-0.2539369762,-0.2053036094,0.0159089174,0.0430039093,0.4937824309,0.0928783789,-0.1783533394,0.0736949816,-0.2672769427,0.4987086058,-0.1321155876,0.1991834939,-0.4336214364,-0.2389552891,-0.1270631403,0.0482878983,-0.1155458838,-0.146817416,-0.163087219,-0.0229732059,0.0547058173,-0.466593653,-0.1804111004,-0.1476681083,-0.1464713812,-0.1106525734,-0.0728618875,0.0458340161,0.3858575225,-0.2659149468,0.2236855775,0.2590919733,-0.1163565665,0.2448017448,-0.0065143993,0.0532293208,0.1860630214,-0.2198518217,0.248524949,-0.077340126,0.1218382195,-0.0466373116,-0.1447132677,-0.1941456497,-0.1593937874,-0.1929103732,0.6028686762,-0.4449506402,0.0085030431,-0.181870237,0.071036987,-0.4804348648,0.0605488867,-0.1846135706,-0.1585884839,0.1554136276,-0.2380546629,-0.2017221302,-0.1194382682,0.0836062133,0.0394908935,-0.1324040294,0.4583229125,-0.0754016414,-0.1538486928,-0.2550146878,-0.0377536975,0.1872317493,-0.5791506767,0.1442901939,0.1076668799,0.0215744544,-0.026669642,0.401689291,0.0692809746,0.1119343564,-0.1006020531,-0.311965555,-0.2653410435,0.0817633644,0.0173049401,0.0790659264,-0.0846888721,0.1661981195,-0.1378832608,0.123114638,-0.2523404062,-0.024547087,-0.0619166754,0.2396284938,-0.1370595098,0.0753155202,-0.0162456501,-0.2204684168,0.0200932622,0.1575829685,0.0309096053,-0.0955521539,-0.1381864697,0.2041949481,0.036003381,0.1450525671,0.2322570831,0.1886535585,-0.0373491906,-0.0894572362,0.230984807,-0.0272437166,0.1587688476,0.0115979398,0.2110564411,0.2475294024,-0.2357632518,-0.0742401034,-0.1002502441,0.2860633433,-0.0533093847,0.2232506275,-0.2110876739,-0.1046605483,-0.0778455287,0.2271034867,-0.0804627165,-0.0676537231,0.3759816885,-0.2518021464,0.0950861499,0.3574811816,0.2441662848,0.1870732605,-0.3089560568,-0.0559827425,-0.0019504882,0.1949564219,-0.2646116912,-0.1269689351,0.1731017977,-0.1789910793,0.0807771161,0.4619521797,0.6363974214,-0.0897296518,0.1896642745,0.0296624061,0.8626216054,-0.1971339285,0.345892489,0.7383452654,-0.2502454817,0.1841810048,0.1887702644,0.0745090917,0.0775646418,0.370562613,-0.2420366853,0.646247983,0.5885434151,-0.2799963951,0.1149151176,-0.319666177,0.2739922702,0.1182508618,0.1751985401,0.0987980813,0.1423454881,0.1888429672,0.0668274313,-0.4418269992,-0.0798788145,0.2540167868,-0.2606357634,0.1267412901,-0.0219981279,-0.2824585438,-0.5228692889,0.1136839092,-0.0689855739,-0.1342749745,0.1042650044,-0.1514091045,-0.3841322958,-0.1322554946,-0.0286236145,-0.015579192,0.2375309765,-0.3725696206,0.2465431392,0.3660732508,-0.1815975606,0.1918540746,0.0483258925,0.43745628,0.0980342403,0.2083120048,-0.1957828701,-0.2718233764,-0.0485977158,0.0833563879,0.3087087572,-0.1459353864,0.1833215803,0.1369089186,0.0744469166,-0.1481654793,-0.0483752303,0.1919300258,0.1745472699,-0.0282274541,0.1795263439,0.3036722243,0.0614366457,-0.1333840489,0.1591781974,-0.1795026213,-0.2685680985,0.5571242571,0.3380035162,0.1735129803,-0.2133608311,0.0767950192,0.1824379414,0.1512424648,0.3064234257,-0.261279434,-0.3386270404,-0.0777239725,-0.6321675181,0.1432313621,-0.0562368743,-0.1568603367,0.3735543787,-0.0976385102,0.1404124647,0.045702666,0.2682737708,0.2316977084,0.4431173503,0.1063270941,0.0316517353,0.0433480181,0.1736647338,-0.1688483208,-0.0223289058,-0.1970438361,0.1939309686,-0.308938086,0.0277123321,-0.1191413626,-0.2667638659,0.3924006224,0.1075056717,0.4319445789,0.0583993755,0.2930966616,-0.1036525816,-0.1393924356,-0.3521636128,-0.0900604427,-0.2822354138,0.3842729926,-0.0054420475,0.2585756779,0.0049408493,-0.2114175111,-0.3754516244,0.0852789804,0.0144334827,-0.1255456209,-0.3520838022,0.3189050555,-0.1653857082,0.3043383658,0.0559064522,0.0430632383,0.1771782935,0.1918556541,-0.2409588844,-0.5613312721,0.6039035916,-0.1921795011,-0.3458337188,-0.3816736937,0.3418906927,0.1194456965,-0.195979014,-0.5978175402,-0.326726079,0.452060014,-0.0045039565,-0.0681277961,0.189874962,0.1181299388,-0.0492893532,-0.0793451816,0.0230095591,0.033820834,0.1604052186,0.1045582965,-0.1852162331]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2440","title":"Remove `extended` field from dataset tagger","comments":"thanks ! this is fixed now","body":"## Describe the bug\r\nWhile working on #2435 I used the [dataset tagger](https:\/\/huggingface.co\/datasets\/tagging\/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:\r\n\r\n```\r\ndataset_name = 'arcd'\r\n\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\n def test_changed_dataset_card(dataset_name):\r\n card_path = repo_path \/ \"datasets\" \/ dataset_name \/ \"README.md\"\r\n assert card_path.exists()\r\n error_messages = []\r\n try:\r\n ReadMe.from_readme(card_path)\r\n except Exception as readme_error:\r\n error_messages.append(f\"The following issues have been found in the dataset cards:\\nREADME:\\n{readme_error}\")\r\n try:\r\n DatasetMetadata.from_readme(card_path)\r\n except Exception as metadata_error:\r\n error_messages.append(\r\n f\"The following issues have been found in the dataset cards:\\nYAML tags:\\n{metadata_error}\"\r\n )\r\n \r\n if error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset cards:\r\nE YAML tags:\r\nE __init__() got an unexpected keyword argument 'extended'\r\n\r\ntests\/test_dataset_cards.py:70: ValueError\r\n```\r\n\r\nConsider either removing this tag from the tagger or including it as part of the validation step in the CI.\r\n\r\ncc @yjernite ","comment_length":6,"text":"Remove `extended` field from dataset tagger\n## Describe the bug\r\nWhile working on #2435 I used the [dataset tagger](https:\/\/huggingface.co\/datasets\/tagging\/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:\r\n\r\n```\r\ndataset_name = 'arcd'\r\n\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\n def test_changed_dataset_card(dataset_name):\r\n card_path = repo_path \/ \"datasets\" \/ dataset_name \/ \"README.md\"\r\n assert card_path.exists()\r\n error_messages = []\r\n try:\r\n ReadMe.from_readme(card_path)\r\n except Exception as readme_error:\r\n error_messages.append(f\"The following issues have been found in the dataset cards:\\nREADME:\\n{readme_error}\")\r\n try:\r\n DatasetMetadata.from_readme(card_path)\r\n except Exception as metadata_error:\r\n error_messages.append(\r\n f\"The following issues have been found in the dataset cards:\\nYAML tags:\\n{metadata_error}\"\r\n )\r\n \r\n if error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset cards:\r\nE YAML tags:\r\nE __init__() got an unexpected keyword argument 'extended'\r\n\r\ntests\/test_dataset_cards.py:70: ValueError\r\n```\r\n\r\nConsider either removing this tag from the tagger or including it as part of the validation step in the CI.\r\n\r\ncc @yjernite \nthanks ! this is fixed now","embeddings":[-0.0462113172,-0.0768266544,-0.0064129154,0.1970952451,0.4162620902,0.3510395288,0.3961618841,0.2223574519,0.0990221426,0.270236671,0.3219407499,0.4198800325,-0.33541435,0.0101606306,-0.1084655747,0.0410654843,0.1885299236,-0.0782069564,0.2728714049,0.1705937237,-0.2169271111,-0.1167993844,-0.0283868201,-0.2122808844,-0.27659747,-0.0125878882,-0.1147346199,0.0142809357,-0.1798166186,-0.6478677988,0.3944429755,0.0409708023,-0.063722454,0.1354271024,-0.0001185429,-0.0698691905,0.2243359238,-0.0024084155,-0.2134162635,-0.278563112,0.0196863487,0.0530185103,0.115748778,-0.157409668,0.1339526623,-0.1959283352,-0.3080334365,-0.243189007,0.0794527084,0.3410100043,0.1256148517,-0.0707778931,-0.2217970043,-0.2267844826,0.3395689428,0.1074107662,-0.0962771848,0.0867756754,0.5839314461,0.3088136017,-0.0191380233,0.485578537,-0.0430630334,-0.0495966189,0.3058120608,-0.1533795148,0.389829874,-0.2507185936,0.2189472765,-0.1223787665,0.3149010539,-0.38846457,-0.1537944824,-0.5451709628,-0.0302891079,-0.4939097464,0.3950106502,-0.2648709714,-0.0556533039,-0.122346364,-0.240243569,-0.2127438486,-0.0544397198,0.0489542522,-0.3670960665,0.4150324762,0.1905630827,0.0626522526,0.1069174632,-0.0787484497,0.0144505892,0.1130781546,-0.177213341,0.0409196801,0.0363634713,-0.1124949455,-0.0657275766,-0.1612489522,0.1218479276,0.3649869263,-0.2219605148,0.0079832906,0.2170217782,0.1405786574,0.3028357625,-0.0307402909,0.0555499904,0.3601344228,0.0725807399,0.1745892316,-0.0741098449,-0.1123137847,0.050138142,-0.0480894558,0.4435277879,0.2302498966,0.3146905303,-0.2855033576,-0.2174763829,0.1254793704,-0.3605772257,0.117136851,0.1331933141,0.1105858609,0.0480825566,-0.0936186463,-0.2851441503,0.3182325959,0.1237713248,0.0119940387,-0.0296305995,0.1283523887,-0.1236450896,-0.1250665635,0.3814832568,-0.0508620292,0.2682861686,0.1685968786,0.151822418,-0.0094796317,-0.0737340674,-0.0996450186,0.3050988317,0.3659884334,-0.0300566722,0.1552374363,0.2130547315,-0.2564800084,-0.1561023742,0.1063994467,-0.2949412167,-0.5918496251,-0.1387456208,0.2065220773,-0.0334179513,-0.0006560928,-0.1766006202,-0.1388459504,0.374944061,-0.0552312881,0.1263389587,-0.324493736,-0.0785211772,-0.0146065075,0.1369053274,0.2880402207,-0.0366274081,-0.0800124332,-0.5086553693,0.2800020874,-0.0853074566,-0.2411043942,-0.3470096588,0.1081900075,-0.3836238384,0.2224304229,0.3002963364,-0.5620270371,-0.235311538,-0.0415155739,-0.1748321205,0.4121417701,0.063278839,-0.2127539963,0.2772896886,-0.2991476357,-0.0866026133,-0.1899987608,-0.4843269289,-0.1344492137,-0.2575746179,0.050413385,0.3582370579,0.0601625964,0.3219889402,-0.2620067596,0.1519873738,0.2293834686,0.0756829083,-0.1429022849,0.0784100294,0.1997951269,0.2894491553,-0.0357674249,0.1141686887,-0.1817525625,-0.4283948541,0.1401445568,-0.1817204952,0.0554073304,-0.4622297883,-0.449739933,-0.3020059168,0.0372620597,-0.3457895815,-0.3504852653,0.0527531244,-0.0250278637,-0.0712613687,-0.0976184607,0.0124706272,-0.1626376659,-0.3748069108,0.3281826675,0.0270247366,-0.1147811711,-0.1708924472,-0.0693489984,-0.2571965158,0.1735017449,0.3265558779,-0.2019084245,-0.213666752,0.2125941813,0.355129391,-0.0508289225,-0.2171712965,0.5169973373,0.3442586958,-0.0980874076,-0.1575529426,0.1415550858,-0.1379870027,-0.1513208896,-0.2316264361,0.0945222527,-0.2365680486,0.371137917,0.0158189908,-0.0413984731,0.2318272591,0.0431259014,-0.1787792891,-0.4130811691,-0.1036863178,-0.2552954257,-0.0175741166,0.0127652623,0.1325827837,0.3341290653,0.3762396872,0.4367774725,-0.1748934537,0.0993461311,-0.1775492728,-0.1059266031,0.1412207931,0.3532337844,0.2486837506,0.117782101,-0.0127263414,0.3396542966,0.1192497611,-0.1062043086,0.445471406,0.2484806329,0.1888837069,0.4824754298,0.2603002787,0.1690918058,-0.0823542252,-0.0995711312,0.1413826346,0.2849093378,-0.2919005752,0.2150722295,-0.389575839,-0.3772101402,-0.101147376,-0.4087667763,-0.3029141426,-0.538539052,0.2825340629,0.1277633607,-0.333381772,0.2526756525,-0.2101630419,-0.129680872,0.0427706912,-0.1742612123,-0.1313567609,0.0175991077,-0.1997762322,-0.005083228,0.1865524352,-0.3542138636,0.393158555,-0.1047494113,0.117423065,-0.4593425691,-0.6823964119,0.2540036142,-0.1064707562,0.2167126238,0.4542515278,-0.3102813363,0.0710545853,-0.287393719,0.2632094622,0.0283847488,0.1383710355,0.0403513797,0.085065417,0.1689662635,-0.2944970727,-0.5482293963,-0.1813831925,-0.1107177362,-0.1892957091,0.1308873892,0.0761193857,0.4946201146,-0.0935571194,-0.2367350459,-0.2716738582,0.0179661438,-0.0909313411,-0.1045179218,0.1077850759,-0.0672242343,-0.104539603,0.3566236496,-0.0956845507,0.518566072,0.1654370129,-0.2120472342,-0.1788322479,0.3612322807,0.4307524562,-0.0530284308,-0.3768746257,0.3619385958,0.2214019597,0.0332433879,-0.0637262836,0.119305633,0.1937485933,-0.2215959132,0.2185125649,-0.3767977357,0.0975748897,0.2391467243,0.352383405,-0.1427764297,-0.0426770896,0.2103664875,-0.2011178881,-0.0333241932,-0.1518466026,-0.6007657647,-0.1218887493,-0.2096941322,-0.3028373718,0.1385840327,-0.1564135402,-0.039240893,0.1636978388,0.1801854968,-0.0304766037,-0.3475955129,0.1442922354,-0.2649933398,0.2648338974,0.1984336525,0.2262377292,-0.2280855179,-0.1495907158,0.1498630345,0.1015605405,0.3522155285,-0.0222996417,0.1771226227,-0.32665658,-0.0240104198,0.2815905809,0.3814673722,0.1261553317,-0.1013399214,0.0646872669,-0.0171753969,-0.1078366414,0.4502841532,-0.2529578805,0.0450865962,0.4356574416,-0.2005776465,0.0477056727,-0.2262141407,-0.161124751,0.185057953,0.087391302,0.5924451351,-0.2567925751,-0.06254448,0.3545739353,0.2144550383,-0.1005252823,0.1415825039,-0.0761358365,-0.2398456633,-0.2477613091,0.1921687126,0.0597070083,0.2991336584,-0.0441986956,-0.0316812508,0.0579718947,0.0392749645,0.1009857953,0.2512366474,0.3023381531,-0.3052487373,0.2778814137,-0.2456796318,0.3937590718,0.7258188128,0.3945235014,-0.0326370224,-0.1563652307,-0.064299047,-0.3916947842,0.3442964554,0.1237754524,-0.1404122114,0.3191731274,-0.2491399795,0.0116509581,-0.1621254086,0.1063333377,0.3952849805,-0.1003417373,-0.2918831408,-0.3863886595,0.0321529023,0.2784665227,-0.0148248831,0.3592590094,-0.0116771664,-0.4866648912,0.1436099261,0.1759919077,0.8011077046,0.3823474944,-0.1719627976,0.2111870944,0.2398453355,0.0741974041,-0.4748704731,0.1287387908,-0.2940504551,-0.2361823618,0.0171618946,0.0262859464,0.472294867,0.0368018486,-0.1968370974,0.0553081073,-0.2571065128,0.5028215647,-0.1378272623,0.1976374984,-0.4355250001,-0.2098915279,-0.1390814632,0.0505192429,-0.085852161,-0.1429692358,-0.1778292954,-0.0374711864,0.0402961187,-0.4741259515,-0.1772120148,-0.1377987117,-0.1393326819,-0.1068647727,-0.077772446,0.039111428,0.3457908928,-0.2903071642,0.2705644667,0.2901952565,-0.1204118505,0.2276877016,0.0199122578,0.0498606153,0.1983307004,-0.2175799757,0.2517167628,-0.111251913,0.1083702371,-0.0280163176,-0.1463258564,-0.1774200499,-0.1583220214,-0.2002956867,0.5863072276,-0.4421055317,0.0195702426,-0.1735641062,0.0502710976,-0.4784705639,0.0644178689,-0.1552637666,-0.1774474531,0.1376023293,-0.2386980951,-0.2039878815,-0.1201643497,0.0759421363,0.0504263267,-0.1121023819,0.4542832673,-0.0559863783,-0.1230918095,-0.2534426451,-0.036921382,0.1961049438,-0.6105554104,0.1413625479,0.1014959663,0.0113893356,-0.0171954669,0.4512276351,0.076390937,0.1462314278,-0.0810016766,-0.3487627804,-0.2624402344,0.0846297964,0.0080770692,0.083752133,-0.0965045393,0.1409650892,-0.1507002562,0.1200335249,-0.2567129135,-0.0121416505,-0.0811488852,0.2473770678,-0.156775102,0.0365413949,-0.0292805731,-0.1994656473,0.0275560282,0.1827025414,0.0124404216,-0.0914104655,-0.1412364095,0.2031496763,0.0234610718,0.1413698792,0.2361170352,0.1948635578,-0.0498455651,-0.0861367509,0.2289134711,-0.0362479538,0.1445638984,0.0047795223,0.2619268894,0.2533825636,-0.2463396341,-0.0347234607,-0.0958832353,0.2670418322,-0.0828657672,0.2237543911,-0.2288952023,-0.1162637323,-0.0165315606,0.2180735916,-0.1116804406,-0.0524961278,0.3777392209,-0.267377466,0.1199024245,0.3429394066,0.251157701,0.2295672297,-0.2969734967,-0.0327852368,0.0016466775,0.1976564229,-0.2747769952,-0.1354241073,0.1470186412,-0.1583087146,0.0785114691,0.4500414431,0.6559212208,-0.107269913,0.2049577981,0.0648151338,0.8469398022,-0.1884472519,0.3461844623,0.7431451082,-0.2624563873,0.1756053716,0.1512257308,0.0635330155,0.0928581282,0.3343069255,-0.2475540936,0.6398672462,0.5777912736,-0.3004900217,0.1267537922,-0.3176538646,0.3000890017,0.1291022599,0.1962518394,0.088102378,0.1188545153,0.1969531924,0.0691865683,-0.4480983317,-0.0981715396,0.2520669401,-0.2584923208,0.1354370415,-0.0022484639,-0.2974014878,-0.5245420337,0.115410991,-0.0800461918,-0.0963653475,0.0927196369,-0.118907474,-0.372076124,-0.1356552243,-0.0070305513,0.0083532268,0.2270049155,-0.3585032821,0.2855601311,0.3843245208,-0.1870392412,0.1828489602,0.0498850383,0.4491161108,0.1273240745,0.1821558326,-0.1730732173,-0.2683154941,-0.0526320189,0.0917801186,0.2783191204,-0.1260081977,0.1761749089,0.1561253369,0.0768370554,-0.1444422603,-0.0335673466,0.1709514409,0.1756077409,-0.0224840753,0.2150027454,0.3005295992,0.0527935103,-0.1401572973,0.1382834315,-0.2138616294,-0.284026742,0.5582799315,0.3528470099,0.1898805499,-0.1978656501,0.0783447996,0.1737494767,0.1521022469,0.2952041328,-0.2642081976,-0.3219383657,-0.0771070868,-0.6138168573,0.1785257459,-0.0547026545,-0.1580157578,0.3402022719,-0.0982999429,0.129017368,0.0445053689,0.3014612794,0.2182102799,0.4378831089,0.11198318,-0.0198727045,0.0335856453,0.1914449185,-0.1842509359,-0.0388259403,-0.2096883059,0.1826887131,-0.3032147586,0.0374278128,-0.0884572342,-0.30649665,0.3922858238,0.1237302944,0.4069455862,0.0599160716,0.3075306714,-0.0834075361,-0.1405850798,-0.3655076027,-0.1063112989,-0.2531937063,0.3644946516,0.0139062619,0.27498281,0.0067550456,-0.1841181964,-0.4159359932,0.095780015,0.0143594826,-0.1253898293,-0.3448574245,0.3184100091,-0.1527571678,0.2758260667,0.0716578439,0.0467913635,0.1778230518,0.1935364753,-0.2533567846,-0.5619792938,0.6317414641,-0.1839078069,-0.3355757296,-0.3740915358,0.3441366851,0.092389822,-0.1795834154,-0.5915859342,-0.3286912739,0.4473665357,-0.0018052678,-0.0823574141,0.1801314652,0.1402060837,-0.0615430847,-0.0944894403,0.0370961577,0.0551455878,0.1590408683,0.083669126,-0.1766809225]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2434","title":"Extend QuestionAnsweringExtractive template to handle nested columns","comments":"this is also the case for the following datasets and configurations:\r\n\r\n* `mlqa` with config `mlqa-translate-train.ar`\r\n\r\n","body":"Currently the `QuestionAnsweringExtractive` task template and `preprare_for_task` only support \"flat\" features. We should extend the functionality to cover QA datasets like:\r\n\r\n* `iapp_wiki_qa_squad`\r\n* `parsinlu_reading_comprehension`\r\n\r\nwhere the nested features differ with those from `squad` and trigger an `ArrowNotImplementedError`:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in \r\n----> 1 ds.prepare_for_task(\"question-answering-extractive\")[0]\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in prepare_for_task(self, task)\r\n 1436 # We found a template so now flush `DatasetInfo` to skip the template update in `DatasetInfo.__post_init__`\r\n 1437 dataset.info.task_templates = None\r\n-> 1438 dataset = dataset.cast(features=template.features)\r\n 1439 return dataset\r\n 1440 \r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in cast(self, features, batch_size, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, num_proc)\r\n 977 format = self.format\r\n 978 dataset = self.with_format(\"arrow\")\r\n--> 979 dataset = dataset.map(\r\n 980 lambda t: t.cast(schema),\r\n 981 batched=True,\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1600 \r\n 1601 if num_proc is None or num_proc == 1:\r\n-> 1602 return self._map_single(\r\n 1603 function=function,\r\n 1604 with_indices=with_indices,\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 176 }\r\n 177 # apply actual function\r\n--> 178 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 179 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 180 # re-apply format to the output\r\n\r\n~\/git\/datasets\/src\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1940 ) # Something simpler?\r\n 1941 try:\r\n-> 1942 batch = apply_function_on_filtered_inputs(\r\n 1943 batch,\r\n 1944 indices,\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1836 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1837 processed_inputs = (\r\n-> 1838 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1839 )\r\n 1840 if update_data is None:\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in (t)\r\n 978 dataset = self.with_format(\"arrow\")\r\n 979 dataset = dataset.map(\r\n--> 980 lambda t: t.cast(schema),\r\n 981 batched=True,\r\n 982 batch_size=batch_size,\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.cast()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.ChunkedArray.cast()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/compute.py in cast(arr, target_type, safe)\r\n 241 else:\r\n 242 options = CastOptions.unsafe(target_type)\r\n--> 243 return call_function(\"cast\", [arr], options)\r\n 244 \r\n 245 \r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.call_function()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.Function.call()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: Unsupported cast from struct, answer_start: list, text: list> to struct using function cast_struct\r\n```","comment_length":16,"text":"Extend QuestionAnsweringExtractive template to handle nested columns\nCurrently the `QuestionAnsweringExtractive` task template and `preprare_for_task` only support \"flat\" features. We should extend the functionality to cover QA datasets like:\r\n\r\n* `iapp_wiki_qa_squad`\r\n* `parsinlu_reading_comprehension`\r\n\r\nwhere the nested features differ with those from `squad` and trigger an `ArrowNotImplementedError`:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in \r\n----> 1 ds.prepare_for_task(\"question-answering-extractive\")[0]\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in prepare_for_task(self, task)\r\n 1436 # We found a template so now flush `DatasetInfo` to skip the template update in `DatasetInfo.__post_init__`\r\n 1437 dataset.info.task_templates = None\r\n-> 1438 dataset = dataset.cast(features=template.features)\r\n 1439 return dataset\r\n 1440 \r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in cast(self, features, batch_size, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, num_proc)\r\n 977 format = self.format\r\n 978 dataset = self.with_format(\"arrow\")\r\n--> 979 dataset = dataset.map(\r\n 980 lambda t: t.cast(schema),\r\n 981 batched=True,\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc)\r\n 1600 \r\n 1601 if num_proc is None or num_proc == 1:\r\n-> 1602 return self._map_single(\r\n 1603 function=function,\r\n 1604 with_indices=with_indices,\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 176 }\r\n 177 # apply actual function\r\n--> 178 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 179 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 180 # re-apply format to the output\r\n\r\n~\/git\/datasets\/src\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 395 # Call actual function\r\n 396 \r\n--> 397 out = func(self, *args, **kwargs)\r\n 398 \r\n 399 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, desc)\r\n 1940 ) # Something simpler?\r\n 1941 try:\r\n-> 1942 batch = apply_function_on_filtered_inputs(\r\n 1943 batch,\r\n 1944 indices,\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in apply_function_on_filtered_inputs(inputs, indices, check_same_num_examples, offset)\r\n 1836 effective_indices = [i + offset for i in indices] if isinstance(indices, list) else indices + offset\r\n 1837 processed_inputs = (\r\n-> 1838 function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1839 )\r\n 1840 if update_data is None:\r\n\r\n~\/git\/datasets\/src\/datasets\/arrow_dataset.py in (t)\r\n 978 dataset = self.with_format(\"arrow\")\r\n 979 dataset = dataset.map(\r\n--> 980 lambda t: t.cast(schema),\r\n 981 batched=True,\r\n 982 batch_size=batch_size,\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.cast()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.ChunkedArray.cast()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/compute.py in cast(arr, target_type, safe)\r\n 241 else:\r\n 242 options = CastOptions.unsafe(target_type)\r\n--> 243 return call_function(\"cast\", [arr], options)\r\n 244 \r\n 245 \r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.call_function()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.Function.call()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/miniconda3\/envs\/datasets\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: Unsupported cast from struct, answer_start: list, text: list> to struct using function cast_struct\r\n```\nthis is also the case for the following datasets and configurations:\r\n\r\n* `mlqa` with config `mlqa-translate-train.ar`\r\n\r\n","embeddings":[-0.4004985392,-0.3374337554,-0.0784154385,0.2124934494,0.0623010099,-0.0605397411,0.342718184,0.6537569761,0.2787890732,0.1644310951,-0.3850086629,0.8013219237,0.0938052833,0.1116523817,-0.22402969,-0.1465635002,0.0285621081,0.1905269176,-0.0690934658,0.1011490896,-0.317885071,0.193488434,-0.2012185156,0.1028063819,0.2194111347,-0.228253141,-0.0353386253,0.088482976,0.2185596377,-0.1091627553,0.353946507,0.2710413635,0.0640781745,-0.0168413743,-0.0001118251,-0.0238936115,-0.0758058652,-0.0531460196,-0.212666899,0.1347667426,-0.2212988734,-0.2578913867,0.0375700369,-0.3121807277,-0.1900653839,-0.2097728848,-0.0742548779,-0.0294896681,0.4723312259,-0.0135347294,0.1707757711,0.353399694,0.226946041,-0.1244799048,0.1267433614,-0.0933816358,-0.4431982636,-0.0782305524,0.5466852188,0.0300399885,0.0169388112,0.3652411997,-0.0343711115,0.107914716,0.023896955,-0.0528597124,0.3286340833,-0.1229999363,0.0111094778,0.4290266335,-0.0217081048,-0.1874396801,-0.2082417905,-0.3809464574,0.0123564163,-0.2188023478,-0.0093456274,-0.2322489768,-0.3328902423,0.1813619733,0.3615427911,-0.4655230343,-0.3431660831,-0.092115812,0.0644313693,0.3115049005,0.0014113832,-0.0799374655,-0.1401880085,-0.1510992944,0.1287179738,-0.0096265841,-0.0296482965,0.1576338708,-0.509965241,-0.4060970247,0.1071711034,-0.5747662783,0.1126487926,0.3861805201,0.4989239275,0.1735469252,0.0279558115,0.206945017,0.1508289874,0.1981992871,0.0859325454,0.1923662871,-0.0598652773,0.5967301726,-0.1092450023,-0.1933967024,-0.1167256311,-0.0851828158,0.1492389739,0.0838888288,0.5520023108,-0.0021331876,-0.3467601538,0.0427193008,0.0344929621,-0.0730466098,-0.2322002053,0.1289039552,0.0269288309,0.2737087011,-0.4161029756,0.3354622126,-0.107742779,0.0315742232,-0.0187874679,-0.1710878909,0.2539903522,0.3146185577,0.1120186672,0.1311676055,0.0525480881,-0.0430616252,0.4304729104,-0.0478942022,0.2082478702,0.0399069563,0.1974537373,0.1108006239,-0.2531925738,-0.285309881,0.222916618,-0.0126007618,-0.1032246649,0.3345327377,-0.1871690303,-0.2716550529,-0.2834213972,0.1354265809,-0.1475449204,-0.0186512191,-0.0152721088,0.3522073328,0.1309848577,-0.1189485714,0.2892568409,-0.1386577785,-0.1265509725,-0.2178006023,0.2230325192,-0.0989434645,-0.2481732666,0.0340819955,-0.2819128931,0.0077523724,-0.1541363746,0.1380896121,0.1155997887,0.1912062764,-0.0904866233,0.2903517187,0.1303349733,-0.4194324911,-0.1716729999,0.1422038227,-0.1727183312,0.4925249219,0.0901335031,-0.217970863,0.3718209267,-0.1345768869,-0.0932563394,0.1010200903,-0.0755134597,-0.262001574,-0.2281343192,-0.2768455446,0.0552995168,0.2245056331,-0.0404817089,-0.0239651147,-0.2805787623,-0.240359664,0.2465333492,-0.2161897421,0.0740369186,-0.272703737,0.6029760838,-0.2477521151,0.1103423685,-0.5407899022,-0.77259022,0.0796709806,-0.3421511948,0.1253489554,-0.2214034945,-0.1351489127,-0.2753249705,0.1095229983,-0.3861001432,-0.3508324027,0.1197643429,-0.1778397113,-0.1192595065,-0.333622247,-0.1127010956,-0.1888127774,0.1344316602,0.1812633574,-0.1693035066,0.0854024887,0.1900447011,0.0097132232,-0.0249612387,0.4596466124,-0.0569810867,0.0291838497,0.0498901382,0.098061271,0.0428169928,-0.0526704006,0.0180584732,-0.1237667128,-0.1701565981,-0.2258189619,0.0442458019,0.1731623262,0.0419866703,-0.0607020482,-0.1348133832,0.6354981065,0.2211730629,0.6569632888,-0.1559611559,-0.1290232688,0.0943490937,-0.049836684,-0.0629422441,-0.2428585291,-0.2287387401,0.2319350243,-0.0882754624,-0.4189498127,-0.3073641956,0.5817826986,0.0810655281,0.1374222338,0.0011099088,-0.1110069007,-0.0673702136,-0.0404732414,0.254290849,-0.2399120033,0.3141377866,0.3139423728,-0.065449588,0.0390544459,-0.0922980085,-0.0944457129,0.2009768635,-0.0393718854,-0.1034151614,0.4231419563,0.4579702318,-0.2617941499,-0.1843265146,-0.0396275111,-0.1588330716,-0.0572177731,-0.1092867255,0.0183848664,-0.0835384056,0.1203019843,-0.0117673762,0.1021931469,0.1570769697,-0.277908504,0.4337159395,0.216811195,-0.2702797651,0.3933640122,-0.0181811377,0.4547273219,0.0914212391,-0.2880818844,-0.3084332943,-0.3010666966,0.1893123835,0.2279930115,-0.1081851795,-0.1813700348,0.1264741421,0.130594328,0.202128455,-0.3754136562,-0.3695494533,0.0802297369,-0.1746841818,0.4358965456,-0.0005782454,0.2324522436,-0.0349891372,-0.4227176905,0.3027733266,0.3596452773,-0.1245285347,0.007156651,-0.0550786592,0.1573766172,-0.0598692596,-0.5525564551,0.1279469281,-0.459766537,0.300028652,-0.0259809047,0.2484274805,0.0255915448,0.1529374421,-0.1002791896,-0.3239284456,-0.1522279233,-0.1322710067,0.144474566,0.4026347399,-0.3067209721,0.0858300477,-0.3118074834,-0.2724570632,0.4823209047,0.1247967258,-0.0917277262,-0.1788227856,0.0560312644,0.5228121877,0.1135839745,-0.1982823759,0.2239517868,0.1521131694,-0.0610609092,0.0546885505,0.2073439658,0.3603225052,0.0601630099,-0.0624172501,-0.0270571057,0.2810737789,-0.1344600469,0.6102373004,0.1061382443,0.6012282968,0.3398057222,-0.0528019369,0.1166456267,-0.1236595362,-0.2203693241,0.1173630059,-0.0170339923,0.0119720353,0.2393427938,0.1865437925,-0.2366127372,0.102124311,-0.0993667096,0.1055637151,-0.4779638052,0.2101210952,-0.0725417212,0.3333581388,0.052525565,0.0694065392,-0.3075441122,-0.1093289033,0.0998681933,-0.0191154741,0.1872664243,-0.3970310986,-0.0782968178,-0.2667006552,-0.1847986877,0.3976645172,0.2287739664,0.1634255201,0.0757724941,-0.4814979434,0.0494246632,-0.0165100824,0.2341278791,-0.0981100053,-0.1952940524,0.313722074,-0.09294983,-0.343793422,-0.2501413822,-0.2566255927,-0.1055371836,-0.231404379,0.748519063,-0.5161657929,-0.0789523572,0.1844935715,0.0071829394,-0.31185323,-0.1599471122,0.0913125724,-0.2091603577,-0.2450037897,0.2980203032,0.1331293285,0.3195592165,0.0549395941,-0.1745123565,-0.2765749991,-0.040504396,0.077967383,0.4612746835,0.0308449902,-0.3243045211,0.1789200753,-0.0541225858,-0.2305426449,-0.0411806442,0.2416390926,-0.3083823025,-0.3228373528,-0.2223910689,-0.2464657575,0.3405139446,0.2358868867,-0.2208076566,0.4975652993,-0.2715390623,0.1151025221,-0.0058118836,-0.1423479617,0.1611718684,0.2125761062,0.06848526,-0.2608027756,0.1550548375,0.1196022555,-0.328846246,0.0244952142,-0.0617087372,-0.1090536863,0.1171464249,0.070243448,0.697475791,0.3031575978,0.0221206676,0.6793621778,-0.162607342,0.1202450916,-0.1493446529,0.0532497019,-0.2023036927,-0.2083169818,0.0938203409,0.0594180264,0.15321064,0.2220788002,-0.1507140696,0.2946811914,-0.0077931727,0.1605764776,0.1185396612,-0.070102483,-0.1148159653,-0.1214280352,-0.2685714066,0.0190218631,-0.106689781,-0.3126922846,0.1520840228,-0.1784346849,0.1498322785,-0.1442405134,-0.2309501767,0.0950113013,-0.2048009038,-0.13690193,0.2875278294,-0.503208518,0.4567984939,-0.2757572234,-0.1087519005,-0.1135817841,-0.1341153383,-0.0185430795,0.2259441614,0.0906741172,-0.1244182512,-0.2219094485,0.3769589067,-0.0928608477,-0.0782057121,0.0012886013,0.0785741732,0.0406579785,-0.2615446448,0.2578839362,0.1981559843,-0.2064759433,-0.2651970983,-0.1959351152,0.1422434896,-0.4721253216,0.114191182,-0.2125702351,-0.0638121292,0.2251409292,0.2377054989,-0.1281977594,0.1455713212,-0.0293277446,0.3228374124,0.0243811272,0.557051301,0.1077824235,-0.08799918,-0.3893527389,-0.0960673988,0.3653009534,-0.3925564587,0.4129045606,0.1317708343,-0.1598312557,-0.1478015929,0.2787010372,0.0747213066,-0.132560581,-0.2192814499,-0.2287515551,-0.0814961642,0.385379225,-0.1765208244,0.1850264519,0.0824656039,-0.0714645684,-0.3542011678,0.4459698796,-0.3408142328,-0.0097484104,0.0725125298,-0.0375281088,-0.0621653497,0.0381144881,0.2486619651,0.2622764111,0.096264571,0.0499067605,-0.1349864751,-0.2726428807,-0.182158798,0.1333673298,-0.0570102781,0.074983798,-0.0720611215,0.1433748454,0.0168694388,-0.1345733106,0.1496858001,-0.4678490162,-0.0120261526,0.2440705895,0.1179509014,0.0236427058,-0.3257214129,0.3457039595,0.0626994744,0.0807158127,-0.0221131388,-0.0019439438,-0.2662564218,-0.0457217246,0.213912338,0.3572964668,0.3749022484,0.1953272521,0.2280796468,0.2977215648,-0.3576919436,-0.415135026,-0.1824419498,0.2406379879,0.1808640957,-0.1292002201,-0.120100297,0.2277158797,-0.3833834231,0.0965554565,0.3412307799,-0.2544451356,0.3618658185,0.3307238817,-0.0065313717,-0.1882486492,0.1555595398,0.1999706626,-0.0098780748,-0.1116261035,-0.0519808009,0.5616613626,0.2358347625,-0.0543840639,0.5195079446,-0.121410951,0.2474182248,0.0503482148,-0.0618534647,0.145121038,0.4857702255,-0.0191891473,0.4051707685,0.0345225707,-0.199208647,-0.0564112701,0.3465835452,0.0792785585,0.2575397193,0.4537781477,0.3087075651,0.1058323458,-0.0425818078,0.192065686,-0.1651949883,-0.3434915543,-0.1943694204,-0.2801077366,-0.2088404,-0.3808355927,-0.1824438274,-0.1929715276,0.4127450883,-0.196121648,0.184238404,0.0878352001,0.3296379149,0.2366622686,0.1715803891,-0.014072516,0.3426659405,0.4159706533,0.0234647263,0.1276071668,0.4066373408,0.2009385526,-0.0599648729,-0.3517586887,-0.1842295527,-0.3521150053,-0.1823284924,0.0067697135,0.2215096951,-0.1202450469,0.1417792588,0.3804834187,0.1290813237,-0.1590753496,0.126243934,0.1178934276,0.0061521041,-0.3411141336,0.3959719241,0.0082837474,-0.2934152782,0.1177167967,-0.2684540153,-0.1586904824,-0.0155188376,0.1246442422,-0.141715169,-0.0374875292,-0.0108927023,0.0675142705,-0.1479430646,0.2800037861,0.1133942157,-0.1268400401,-0.1358676851,0.0824644268,-0.405144006,0.3420463204,0.0683750734,-0.1710830182,0.2279713601,0.0964195132,0.1493554264,0.4303812683,-0.0037878847,-0.0471724272,-0.1360999942,0.3810090423,-0.832749486,0.0307475422,0.5487805009,0.0535662882,0.1971005797,-0.2480857819,0.0202133209,0.4855688214,-0.0476235636,0.1241229773,-0.0867305845,0.2254624069,-0.3691396415,0.1861017197,0.173130542,0.1237802729,-0.0506397486,-0.1343679577,-0.0293767117,-0.0333654769,-0.2019147277,0.1373996884,0.1758999228,0.3460509181,0.0347754508,-0.2025481611,0.1435955316,0.2318672538,-0.0811656117,-0.2745198905,-0.183077544,0.3552872539,0.1164726615,-0.1305212975,-0.237274915,0.1530246884,0.1424942017,0.0470586717,-0.153107062,-0.5564714074,0.4793992639,0.0170533042,-0.601374507,0.100932762,0.1521765143,0.0249979012,-0.1862440854,-0.463799268,-0.0391586795,0.3138693571,-0.0124705946,0.1759585142,0.3320384324,0.0188714899,-0.0121068032,-0.3110931218,0.373627156,0.186700955,0.0191646274,-0.5798617601,-0.2618544698]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2431","title":"DuplicatedKeysError when trying to load adversarial_qa","comments":"Thanks for reporting !\r\n#2433 fixed the issue, thanks @mariosasko :)\r\n\r\nWe'll do a patch release soon of the library.\r\nIn the meantime, you can use the fixed version of adversarial_qa by adding `script_version=\"master\"` in `load_dataset`","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndataset = load_dataset('adversarial_qa', 'adversarialQA')\r\n```\r\n\r\n## Expected results\r\nThe dataset should be loaded into memory\r\n\r\n## Actual results\r\n\r\n>DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\n>Found duplicate Key: 4d3cb5677211ee32895ca9c66dad04d7152254d4\r\n>Keys should be unique and deterministic in nature\r\n>\r\n>\r\n>During handling of the above exception, another exception occurred:\r\n>\r\n>DuplicatedKeysError Traceback (most recent call last)\r\n>\r\n>\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in check_duplicate_keys(self)\r\n> 347 for hash, key in self.hkey_record:\r\n> 348 if hash in tmp_record:\r\n>--> 349 raise DuplicatedKeysError(key)\r\n> 350 else:\r\n> 351 tmp_record.add(hash)\r\n>\r\n>DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\n>Found duplicate Key: 4d3cb5677211ee32895ca9c66dad04d7152254d4\r\n>Keys should be unique and deterministic in nature\r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.109+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n","comment_length":36,"text":"DuplicatedKeysError when trying to load adversarial_qa\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndataset = load_dataset('adversarial_qa', 'adversarialQA')\r\n```\r\n\r\n## Expected results\r\nThe dataset should be loaded into memory\r\n\r\n## Actual results\r\n\r\n>DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\n>Found duplicate Key: 4d3cb5677211ee32895ca9c66dad04d7152254d4\r\n>Keys should be unique and deterministic in nature\r\n>\r\n>\r\n>During handling of the above exception, another exception occurred:\r\n>\r\n>DuplicatedKeysError Traceback (most recent call last)\r\n>\r\n>\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in check_duplicate_keys(self)\r\n> 347 for hash, key in self.hkey_record:\r\n> 348 if hash in tmp_record:\r\n>--> 349 raise DuplicatedKeysError(key)\r\n> 350 else:\r\n> 351 tmp_record.add(hash)\r\n>\r\n>DuplicatedKeysError: FAILURE TO GENERATE DATASET !\r\n>Found duplicate Key: 4d3cb5677211ee32895ca9c66dad04d7152254d4\r\n>Keys should be unique and deterministic in nature\r\n\r\n## Environment info\r\n- `datasets` version: 1.7.0\r\n- Platform: Linux-5.4.109+-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyArrow version: 3.0.0\r\n\nThanks for reporting !\r\n#2433 fixed the issue, thanks @mariosasko :)\r\n\r\nWe'll do a patch release soon of the library.\r\nIn the meantime, you can use the fixed version of adversarial_qa by adding `script_version=\"master\"` in `load_dataset`","embeddings":[-0.1690554023,0.0245385617,0.0173124857,0.2362533659,0.2499980032,-0.2299038023,0.3607746065,0.223856926,0.0596654527,0.1077191457,0.1362492144,0.5518656969,-0.1071600392,-0.1683701426,0.1626174599,0.0517192595,-0.0379656665,0.1883038431,0.0392815843,0.0287615117,-0.3011764586,0.3779449761,-0.1012482494,0.3241074383,0.0359348692,-0.0679111406,-0.1033208892,0.1514307857,0.2137738913,-0.4177423418,0.0700682923,-0.0187095534,-0.1166656017,0.4811602533,-0.0001052441,0.144102633,0.2889031172,0.1035589874,-0.0913373306,-0.1600612849,0.0185088813,-0.1071981043,0.1912548542,-0.1924386919,0.1262396574,-0.4752852619,0.0131603889,-0.3006721735,0.5054063201,0.1201551408,0.3179467022,0.0903188139,0.2629330754,-0.171065256,0.1195176542,-0.1868845075,-0.1742420495,0.1196182147,-0.0303588919,0.0372036919,0.2526727319,0.2869953215,-0.0103847869,0.0183801148,0.4412295222,-0.0151753277,0.2555160224,0.1374571323,0.3210968971,0.3423620164,0.2900976837,-0.4888226092,-0.214271605,-0.0686696917,0.3486289978,-0.2374397665,0.236017555,0.027807869,-0.2709132135,-0.0538720042,0.3976077139,-0.1472533643,-0.0434323885,-0.1063056663,0.1602600515,0.1248760298,-0.0217962377,0.1693464369,0.3264658153,-0.1023193002,0.0459521711,0.0315127932,-0.2117156833,-0.0338133611,-0.4873888493,0.1138649359,-0.0784596279,-0.0931205899,0.2700706124,0.0484963879,0.512791276,-0.0115022063,0.0825756639,0.1235229596,0.1281139255,0.2494140565,0.1686249971,0.2489651293,0.2112814188,0.0933455601,-0.2349935323,0.024731908,-0.0834695995,-0.3852000535,0.6409913301,0.0559605286,0.3691559136,-0.0636514947,-0.248886019,0.1071039736,-0.3669061065,-0.0592118315,-0.2935501337,0.339161396,0.1958414912,0.0174902473,-0.1358401626,0.2525438666,-0.2060322464,0.2294732928,-0.2119791061,-0.1100280657,-0.061905086,-0.1172626242,0.0400002785,-0.1178540066,0.2964857817,-0.0017159585,-0.0026419838,-0.2703484595,0.3662492335,0.0210249461,-0.2416959405,0.2553783357,0.1975595951,0.1291869432,0.0458439291,-0.379288882,-0.1068110019,0.0993222892,-0.0052411021,-0.0512386039,-0.0127534429,0.2953703105,0.1497000754,0.0862279758,-0.0123280352,-0.0367017016,0.1766476929,-0.0752214566,0.0585838184,-0.0121802548,-0.1647233218,-0.442214638,0.0968423411,0.4755463004,-0.3256644011,-0.1278250664,-0.3280483186,-0.0849097446,-0.2783800066,0.2302599549,-0.2746446133,0.0413547792,-0.2059623748,0.0704587251,0.254255712,-0.4608035088,-0.4768517017,0.318764925,-0.3031131029,0.18343243,0.1363978684,0.1553700417,0.1448499858,0.0580859296,-0.0169934295,0.0008209471,0.1661281884,-0.1249247044,-0.2674152553,-0.3802102506,0.1170748994,-0.1609794647,0.0419294089,-0.2055484802,0.177867353,-0.4174796939,0.3148236275,-0.0983029604,0.0229328666,0.0728653371,0.2474905103,-0.0617426001,0.2550160289,0.024097519,-0.3038271964,0.251642257,-0.2327656895,-0.1694363505,-0.2557512224,-0.2266550213,-0.2787438333,0.1975636929,-0.5176190734,-0.166650936,0.259850353,0.2048997879,-0.1156930104,0.2234192938,-0.0729155019,0.0727873445,0.1837258041,0.0310812965,-0.2157280594,-0.1280266047,-0.2033466995,-0.00254299,0.1591006666,0.1641196609,0.105101198,-0.0211301092,-0.0315848105,0.3745127916,-0.2419859618,0.1823240966,-0.0071171322,-0.0134901097,0.0761555731,-0.3286634088,0.2699786425,0.1486452967,0.29554528,-0.1044862643,0.2927990854,0.2971185744,-0.0134968944,0.2014833391,-0.0366398916,0.0408822075,0.2538094521,-0.1656628549,0.0626064092,-0.0536196008,0.1130050868,0.0100470856,0.0348612666,-0.0021447081,-0.0330811813,0.2147905082,0.230107531,-0.0221389886,0.3043051064,0.0318967178,-0.2252117097,0.0949227586,-0.1188314036,0.5473522544,0.4136870205,0.1460587382,-0.3134827316,-0.0018698144,-0.1162209734,-0.2087380886,0.394338131,-0.1712438613,0.2710690796,0.2361287773,0.0873066187,0.021922851,-0.3811853826,0.0386912338,0.0817479938,0.2271606475,-0.1483618915,0.0967334434,-0.2366157621,0.2362318933,-0.1268054247,0.0108367437,-0.0239075106,-0.2116208225,0.0958054364,-0.1182431355,-0.0491660722,0.2551221848,-0.0901474953,0.1280748099,-0.1118554324,0.1302352399,0.0685867667,-0.2182480097,-0.3099152446,0.1367270499,0.2610578239,-0.0785213709,0.4030058086,0.0804523006,-0.0342302173,-0.1852927208,-0.191251263,-0.0399555415,-0.0497093648,0.2836412191,0.1351198554,-0.0747143477,-0.1941614002,-0.4077005386,0.2535795569,-0.2552897036,-0.176333636,0.3357872963,-0.0151262116,0.0049349866,-0.1734302789,-0.4460121393,-0.1923153996,-0.1484011561,0.2947089672,0.0645741895,0.1293051243,0.022704903,0.2246602774,0.0428250432,-0.1031477824,0.1216512248,-0.2890095711,0.1607942581,0.113841258,-0.1703122407,-0.3239609897,0.0529669784,-0.1017126292,0.360444665,-0.1304558516,-0.601433754,-0.2104379833,-0.2563964725,0.1294313669,-0.051540833,0.0978013575,0.2213895619,0.2180900574,-0.2045319378,0.0228927303,-0.1176154837,0.4647151232,-0.5940796137,-0.0129853617,0.0150222387,0.5313708782,-0.0693257302,0.5452752709,0.2173437327,0.0534125008,0.1459403783,-0.0911503434,0.3452010155,-0.1222963855,-0.2979429364,-0.295717895,-0.04432166,0.2269752175,0.0703490525,-0.0713013113,-0.1328836679,0.0996658579,0.0637233853,-0.5250399709,-0.4105526507,-0.0504653379,-0.3686192632,-0.0499646217,0.125163883,0.1396446079,-0.119626984,-0.0425439142,0.2148649246,0.1139081568,-0.0498430207,-0.2372042984,-0.1185947806,-0.1230585724,-0.449226141,0.2901737094,-0.1970184743,0.3927428424,-0.0283109937,-0.0395989344,0.2069393992,-0.3311928213,0.5487363338,-0.1006397307,-0.0437834412,0.1423230916,0.004482789,-0.3103791177,-0.0675509349,-0.005320421,0.0821350366,-0.3869895339,0.1475908458,-0.4015746415,0.325081557,-0.0271142963,0.0716884062,-0.4544391632,-0.2566442192,-0.3444777131,-0.3723651767,-0.5866625309,-0.0118524553,-0.246175915,0.2684402168,-0.0486089773,-0.0742864162,-0.1318245083,-0.2240801007,0.1418439746,0.1856937855,0.3297496736,-0.0783314034,0.4084164202,0.1614056379,-0.2644542754,0.066890426,0.4830089509,-0.04351772,-0.2060808092,-0.2193892896,-0.3509432971,0.0361798257,0.2154800892,-0.2175709456,0.3857224584,-0.2800894678,-0.1288882047,0.0625397936,-0.3895275891,0.3317033052,-0.0128449062,0.1893234402,-0.2071923763,0.1652026325,-0.0126962224,-0.3141309023,0.2452116758,0.1695786417,-0.0903836563,0.5747087598,-0.1488593221,0.8625911474,-0.1967768669,0.1831266433,0.2162335813,-0.4381442368,0.2829995751,0.0253959578,-0.067696318,-0.5362507105,-0.2563013434,0.2176161855,0.024692446,-0.04977547,-0.0239700489,-0.2305036932,0.0604666807,-0.2146956027,-0.0899975076,0.0369567648,0.1902624369,-0.2748835385,0.0282420218,-0.1978779733,0.1951878816,0.0725922436,-0.0656854585,0.0461747833,-0.0735720992,0.2139962614,-0.421246469,-0.3048992753,0.1753005087,-0.252887696,0.2598187327,0.0801970661,-0.307921201,0.1604757607,0.2478317469,0.2872768342,0.1719996631,0.0351867042,0.2488935292,0.2083690912,0.1343954504,-0.1962566972,-0.1868822873,-0.0186943039,-0.1554929912,-0.340211153,0.3303759396,0.1148404703,-0.3770139515,-0.0781607032,0.1308063418,0.0611938797,0.2026581466,0.0681278855,-0.0313435756,-0.0867437944,-0.2570920885,0.2101154327,0.1591988951,-0.0914342999,0.1320965737,-0.0596611798,-0.2535030246,0.1193293035,0.2277128249,0.2914981842,0.1522347033,0.7089361548,0.0403515287,0.0634063482,-0.3153806031,-0.1432933807,0.2455528975,-0.6300361753,0.2436055243,0.0223504864,0.0298903808,0.0513813309,0.1311262399,0.0982888192,0.0846286938,0.0221147276,-0.3444291353,-0.1053890139,0.0957377329,-0.0054350384,0.1040807664,0.033886902,-0.1115774438,-0.2146911919,0.426320672,-0.3956038654,0.2180331945,0.0735007077,-0.0979469568,-0.4599565268,-0.1142995507,0.155190438,0.2482208312,0.2072069943,0.3262910247,-0.1163023189,-0.304641217,-0.1620006263,0.1191147491,-0.1236592606,0.0134607395,-0.0771646723,-0.3239645958,0.1320306808,0.0880462304,0.2979100943,-0.0093370872,-0.1027643085,-0.0421517976,0.2517244816,0.1904056221,-0.1696843952,0.1836373508,-0.5097129345,0.1367590874,0.1345285177,0.1311323196,-0.1236748993,-0.0908113495,-0.0826575756,0.1135312915,-0.0396270081,0.146299094,0.2333976328,-0.2341088653,-0.2200934589,0.0182027891,0.2221576422,0.1798067689,-0.2681423724,-0.1640091389,-0.0306903962,0.301974833,-0.3929407597,-0.2168429494,-0.1873518676,-0.0635854527,-0.0176597666,0.1850602925,0.1885098219,-0.4881368279,-0.0743756965,0.2150141299,0.134907946,-0.3536396623,0.0407510884,0.2481795996,-0.1442987025,-0.1378642321,0.2240298092,-0.1203108579,0.215792045,0.3417678177,-0.4048341811,0.1180556491,-0.0068648714,0.0424533859,0.1101279706,-0.3249895871,0.3864814341,0.1536853462,-0.1335543096,0.0375901386,0.3025678992,0.6008815765,0.0319885015,0.1380994022,-0.3952950239,0.033575885,-0.2369766384,-0.2100066245,-0.194609493,-0.1895808876,-0.3405833542,-0.0322144367,-0.0924285352,0.0909312367,0.2315273732,-0.0269406941,-0.3150287569,-0.2569419742,0.2479465306,0.1572834402,0.2248806655,-0.2477217168,0.0930370018,0.4507555664,-0.0367642455,-0.0188956633,0.3361511528,0.3789549768,0.1391470283,-0.231650874,0.236138165,0.0226496458,0.0811645687,0.0013859137,0.2505359352,0.1345093697,0.2800769806,0.1820329875,0.192076385,-0.2005641013,0.0931581259,0.3853110373,0.2345333248,-0.0048910026,0.1036343426,-0.3479995728,-0.0272679366,-0.161302492,0.4776095748,-0.2901833057,-0.024458373,0.2981216908,0.0215832219,-0.0813905299,-0.7242833376,0.1511066407,-0.1152197048,0.3489242792,0.294498086,-0.0292275008,-0.3234564364,0.0851724669,-0.7443155646,0.5754794478,-0.3601246774,0.2356722355,-0.0357267596,0.204594329,-0.0286514182,-0.07412716,0.2015185505,0.4417279661,-0.331448704,0.2870698571,-0.3599570096,-0.1302859783,0.2769171894,-0.2032385617,0.103304565,-0.4686330855,0.0324178971,-0.5130493641,0.1535775363,-0.0300121233,-0.2659941912,0.2487077266,-0.0412469283,0.4548517168,0.166301325,0.1476598382,-0.0329004228,-0.3738212883,-0.1941197962,-0.0635113716,-0.2092291862,0.1470029205,0.4681989551,0.1502508223,0.0012092597,0.0563175939,-0.0963239372,0.3003609776,-0.069293499,0.0320126414,-0.2587635815,0.279517442,-0.004311583,-0.0246220436,0.0537909083,0.1890216917,0.012757916,0.234276697,-0.2320858687,-0.4410297871,0.3400584459,-0.2176022232,-0.006342791,0.0182968769,0.3659908175,-0.0719234943,-0.0738373771,-0.4508941174,0.1418466866,0.334736377,0.0527791195,-0.045541849,0.056456279,-0.0670094267,0.0091567906,-0.0320915245,0.2730198801,0.0710982382,-0.0779992566,0.3131793439,-0.2689362168]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2426","title":"Saving Graph\/Structured Data in Datasets","comments":"It should probably work out of the box to save structured data. If you want to show an example we can help you.","body":"Thanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.","comment_length":23,"text":"Saving Graph\/Structured Data in Datasets\nThanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.\nIt should probably work out of the box to save structured data. If you want to show an example we can help you.","embeddings":[-0.2739242315,0.0389155634,-0.0391023234,0.2492293268,0.1765464395,-0.1106182709,-0.0422585346,0.0588869825,0.0385465659,0.157548368,-0.1896000952,0.4191503823,-0.4267219007,0.6438033581,-0.1479446143,-0.2593741715,0.2712560892,0.0387915932,0.2738686502,0.0248014051,-0.1155969352,0.0541642085,0.0334124379,-0.0629185215,-0.2414860725,-0.483335495,-0.155989781,0.0525340214,-0.3321605921,-0.4602074623,0.193873778,-0.1048882827,0.3438137174,0.2096619457,-0.0001199866,-0.0553125031,0.1516453773,-0.1994683594,-0.0566304997,-0.2451012731,-0.0566121265,-0.2753434479,0.2017225623,-0.4001507759,-0.0335036851,-0.4654900134,0.0522395968,-0.2573404014,0.3798193038,0.1249340549,0.1100876033,-0.0983889103,0.357000947,0.1881575286,0.1299807578,0.5715721846,-0.1208103076,0.3607281148,0.013202833,0.1853478253,0.4580911994,-0.005127124,-0.1430881321,-0.1577080637,0.2900474668,-0.0249025803,-0.0308635589,-0.4098470807,-0.1116895378,0.1381860822,0.723748982,-0.4365120232,-0.233100906,-0.1412198097,-0.1959114671,-0.1156859398,-0.0175559595,0.152929157,-0.2070350796,0.2562131286,0.2139635235,-0.3798537552,-0.358709693,0.4734244347,-0.0559350103,0.2613530755,-0.0914311931,0.1273645312,-0.1421518177,-0.1118723825,0.0549218766,-0.1248368919,0.1320752054,0.0886373594,-0.0203796141,-0.0758737028,-0.4801715612,0.2287084609,0.0287253894,0.3036389649,0.0697164163,0.0168967787,0.3490336835,0.2395004481,0.475723654,-0.1442130953,0.0471447818,0.3832100034,0.4966682792,-0.3662723601,0.0406668074,-0.1784888506,-0.2933403552,0.1213971376,0.4435192943,-0.3187876046,0.5578758121,0.0079512931,-0.1359265149,0.0446229801,-0.1969697922,0.0967279375,-0.0216956679,0.3609926403,-0.2775350809,0.0732495338,-0.1412934065,0.5445889831,0.0454772003,0.2662028372,0.1382092535,0.0149851786,0.0218704604,0.0009210279,0.1089767292,-0.1986514181,0.0825165883,0.0825564787,-0.1388048232,-0.2052012384,0.0733292103,-0.3362773061,0.2886628807,0.1905325502,-0.0180045962,0.1952151805,-0.1086379811,-0.2577902675,-0.0838287324,0.2124571204,-0.2254009247,-0.5140501261,-0.5806310773,0.052874241,-0.2283154726,-0.2235742062,-0.323785007,0.1371755898,0.445918262,-0.0290160812,-0.0331984125,-0.1683663428,-0.1464661211,-0.5850507021,0.0567371398,-0.2350818515,-0.4784719348,0.0991271138,-0.3251682818,-0.0744251832,0.1291402578,0.4273463786,-0.0596012808,0.3404157162,0.0357297994,0.4138521254,0.698769629,-0.2777142227,-0.4148166776,-0.0819008127,0.4841585159,0.0040979031,-0.1289055198,0.3573928773,0.5263537169,-0.1022598296,-0.3043392599,0.0996801779,-0.0032445143,-0.0020724984,-0.1499859542,0.0159728248,0.3745208383,0.0442034192,-0.1373299807,-0.2539263368,0.2561991215,-0.2004583329,0.1385451257,0.0192221981,0.0950797871,0.1007085964,0.5383986831,0.5469770432,0.0907589272,-0.3336781263,-0.3127003908,0.0721707642,-0.2295341194,-0.308480233,-0.5126116872,-0.3048506677,-0.2286320329,0.2266468406,-0.1705741882,0.027937334,-0.0004071034,-0.1398181915,-0.0901462585,0.2064345628,-0.3856292069,0.0508762822,-0.0188888069,0.0089701163,0.1010847688,0.2701417804,-0.1949193776,-0.5788186789,0.0939568654,0.2997945845,-0.010896395,0.0467017516,-0.0175425373,0.3709216416,0.1792910844,-0.1432299167,-0.3749114573,0.1712652594,0.0279154982,-0.3871057034,0.0973690972,0.139295429,0.2226231247,0.0337360725,-0.13293989,0.2601615787,0.3219934702,-0.1201745346,-0.1562555879,0.1346736252,0.1626520753,0.119449079,-0.0708395168,-0.0142743047,-0.1685674936,-0.0778061152,0.0543080159,0.1200311333,-0.7086869478,0.0711341724,0.4323029518,-0.0024226049,0.5254663825,-0.059575066,-0.0401118435,-0.1502237916,0.2823343575,-0.03111214,0.2100045532,0.0562598929,-0.1707645208,-0.0801717415,-0.0489392206,-0.0759968981,0.0649271458,0.1203808561,0.3846667707,0.1639330387,0.1507395059,-0.034177877,-0.0859737024,0.2178903818,-0.1813958287,0.0015845292,-0.3497137129,0.1840836704,-0.1545985192,-0.4507990181,0.091239199,-0.2079921067,-0.1476707608,-0.231444791,-0.0228111409,0.0237650219,0.0121921031,0.1620066315,0.4222955406,0.2165821344,-0.1318423897,-0.3234273493,-0.2239544094,-0.3585410118,-0.1564082354,-0.011779882,0.0598242693,0.1675311774,0.4270932078,-0.039924264,0.2172190696,-0.3474119306,-0.2564814687,-0.0247166138,-0.063722536,0.0088332221,0.0550150499,0.3775037527,-0.1045713127,-0.1879948527,-0.0885133892,-0.1572542936,-0.1157544479,0.0610345379,-0.1081991792,-0.0642103553,0.109876208,-0.1394550204,-0.8664664626,-0.1614281684,0.3969891965,0.3282330036,0.061320819,-0.0835260227,-0.0021313094,-0.119987458,-0.1332860291,0.0077423435,-0.031291943,-0.3910387158,0.6375262141,-0.3082252741,-0.4324339628,0.2933846712,-0.0717979148,0.1257753819,-0.0882094726,0.1132861972,0.3267559707,-0.0515328199,0.4475869238,0.0574893318,-0.0259936173,0.2197533995,0.0480363369,0.1845579892,-0.1009557396,0.189058587,0.2827368379,0.1272770911,0.321865499,-0.0799065977,0.1303382814,0.0115075987,0.2465844899,0.0448854938,-0.0933634266,0.4398342669,-0.1336839944,0.3183667064,0.0108390311,0.0069366996,-0.0040597073,-0.1631752402,0.0090190563,0.0922248736,0.088914305,-0.2952551544,-0.0644934773,-0.2197118998,0.0109106973,-0.153057009,0.133110866,-0.5054294467,0.2884161174,-0.1327303946,-0.1492357254,0.0692283586,-0.1835989654,0.0185348392,0.2247162014,0.0847114474,0.2152620405,0.024094265,0.0453175567,-0.4616928697,0.1995841563,0.2359960079,-0.001528295,-0.1771086752,-0.0131538603,0.2584995031,0.136202842,0.004395837,-0.0446180254,-0.3111343384,-0.1692455411,0.0505853184,-0.4030066729,0.1320203841,-0.0676533207,0.0516197495,0.0647099391,0.319065392,-0.3750118017,-0.0330993384,-0.1412836611,0.2624394596,-0.3222723007,-0.1376175433,-0.1584223807,-0.2395705134,-0.4591308236,-0.2636276186,0.3954102099,-0.1299705058,0.1575474739,-0.1445972472,-0.3792842925,0.041169133,-0.1711989194,-0.0569994114,0.4456487596,-0.4513582885,-0.203185156,0.2764333189,0.2749667466,0.2852410078,0.3952458501,0.0585568845,-0.3404387534,-0.1505059153,0.0688276738,0.393848747,-0.2941744626,-0.2827282548,-0.0331750661,-0.1187790483,-0.0880132467,-0.2690441608,-0.1278584749,0.2524332106,0.0330475383,-0.730305016,-0.5065047741,0.3701466918,0.1977503747,-0.0441390313,-0.0819270313,0.2267474085,-0.1044174656,0.5230234265,0.336945802,0.9222271442,-0.1236640215,0.3628592193,0.3219562173,-0.4654832184,0.3435896933,-0.0497577526,-0.2487834096,-0.3438928425,0.3321546912,-0.170214504,0.0735957846,0.2791543007,0.1188886762,-0.3840119243,0.0435865223,-0.2547366321,-0.2646394074,-0.055368606,0.415532738,-0.0775093958,0.0823907778,-0.3506851792,0.0696490332,0.0457903445,0.0893840343,-0.1200310439,-0.0315538682,-0.2052202076,0.0014496854,-0.1673033684,-0.192152977,-0.0001375319,0.3817147315,0.1153983101,-0.3777297139,0.2035635859,0.5662435293,0.243087858,-0.0402707569,-0.2280299366,0.0273313075,0.09482041,-0.0905546471,0.1944992691,-0.0795727074,0.4053023458,-0.1497245729,-0.1920763999,0.0298310444,-0.0605867654,-0.2365449667,-0.31731534,-0.0388990082,-0.0681997538,-0.2013991028,0.0308904853,-0.4597535133,-0.0601130724,0.0414439701,0.0602422953,0.375154525,0.1638067663,0.4845051765,-0.1038536206,0.052689258,-0.0722562298,0.090582259,0.0046548177,0.2955243587,0.3351482451,-0.0535709597,-0.2657906413,-0.1244489998,0.4560087919,0.0366503261,-0.106848754,0.1845739484,0.1027294025,0.1321051121,-0.287302345,0.5487669706,0.1577096581,-0.4202432334,-0.1482123137,-0.3057685792,-0.4151207209,0.1710181236,0.192871958,0.2847711742,0.0346566141,0.1072654426,0.1212044731,0.0023185434,-0.1816729754,-0.0584642477,-0.1107701436,0.1396639645,0.1899677366,0.0981935114,-0.0309617221,-0.204658553,-0.0433750413,-0.1228898913,0.1459270716,-0.0717408061,-0.1759452522,0.196218431,0.0762231499,-0.135706827,0.0826891363,-0.3776551187,-0.089908272,0.2189729959,-0.1723040044,0.370501101,-0.0560695864,0.1345966607,0.0978346616,0.0006529429,0.0382282361,0.1371964812,-0.1900976002,0.1095947996,-0.2090770453,0.1986453235,-0.0120244762,-0.198012501,0.2365083098,0.1115507185,0.2743534744,-0.1883937418,0.3860310316,-0.2259959579,0.054920733,-0.2953667641,0.3457432985,0.2153708041,-0.0679592341,0.1284864843,0.3097656071,0.0269792862,-0.2391414344,0.0055603483,0.0193402786,-0.1614257097,-0.1401199102,0.1201207489,0.2192446142,0.4591211081,-0.7182412744,-0.0712944567,0.151619181,0.059244018,0.1226512715,0.0577367768,-0.142590031,0.0160026848,0.1898609698,0.3188076913,0.1759219766,0.4872881472,-0.0977185667,0.1729775965,0.4726815522,-0.1042809263,0.5654816031,-0.0388983302,-0.1138923839,-0.4172440469,0.0201839209,-0.1344977617,0.3396229744,0.4090042412,0.1617929786,-0.2944789231,0.1116717085,0.5377633572,0.0106216194,-0.0998328626,-0.1528786272,-0.0495046228,0.0271226931,-0.3353942931,-0.3285278678,-0.449331671,0.116735898,0.3489614427,0.1912849545,-0.3651035726,0.4992555678,-0.0378665179,0.2248036712,-0.0209787469,0.2331258208,-0.2217001021,-0.0811195299,-0.1414305121,0.1262820065,0.500813365,0.2805133462,-0.1380147189,0.3987898231,-0.1735538542,-0.2531891465,0.0304629076,0.0083875209,0.1520697623,0.188951984,0.4188032746,-0.0069607836,0.0702834055,0.1499118954,0.1195798591,0.2691441774,-0.0219239239,0.4289731681,-0.2066787779,-0.2854379117,-0.0260286257,-0.0637116954,-0.1970925629,-0.218703717,0.3100643456,0.0867256001,0.2249694914,-0.2830573916,0.012339375,0.0635156408,0.564782083,0.1533934772,0.6376802921,0.0431213155,-0.0837929398,-0.4523647726,-0.1175952479,-0.0585143827,-0.166807726,0.0635967553,0.0668436736,-0.024544606,0.4297702909,0.1752510369,0.455604583,-0.0212628283,0.1491663456,-0.1350543499,-0.0380245596,-0.1125688925,0.1605095416,0.011380408,-0.0399842076,0.3054420948,-0.0912986323,0.0355631895,-0.1888400614,-0.0639921725,-0.0251224823,-0.3501341641,0.3043224514,-0.0399399027,0.2214451283,0.0427275598,0.1490404159,0.3592251837,-0.0212914478,-0.2375466675,0.3152908087,0.0587146766,0.2484672517,-0.3070229292,0.2642485499,-0.0496712103,-0.1571848392,-0.1861047298,-0.0388189219,-0.1620766222,0.0006382578,0.1151321903,0.0066862721,-0.0655350462,0.0716569722,-0.2507295609,0.0382839739,-0.3548092246,-0.1365326792,0.5200733542,-0.0038205732,-0.1697152406,0.0650631562,-0.0101881186,0.3891577125,0.1384936571,-0.4002582133,0.0299438015,0.3884984255,-0.131535694,0.0317554474,0.0854040831,0.19755584,-0.0195277967,-0.0831729695,0.5987548828,-0.0818712711,0.201020211,0.0144745065,-0.3707670867]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2426","title":"Saving Graph\/Structured Data in Datasets","comments":"An example of a toy dataset is like:\r\n```json\r\n[\r\n {\r\n \"name\": \"mike\",\r\n \"friends\": [\r\n \"tom\",\r\n \"lily\"\r\n ],\r\n \"articles\": [\r\n {\r\n \"title\": \"aaaaa\",\r\n \"reader\": [\r\n \"tom\",\r\n \"lucy\"\r\n ]\r\n }\r\n ]\r\n },\r\n {\r\n \"name\": \"tom\",\r\n \"friends\": [\r\n \"mike\",\r\n \"bbb\"\r\n ],\r\n \"articles\": [\r\n {\r\n \"title\": \"xxxxx\",\r\n \"reader\": [\r\n \"tom\",\r\n \"qqqq\"\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n```\r\nWe can use the friendship relation to build a directional graph, and a user node can be represented using the articles written by himself. And the relationship between articles can be built when the article has read by the same user.\r\nThis dataset can be used to model the heterogeneous relationship between users and articles, and this graph can be used to build recommendation systems to recommend articles to the user, or potential friends to the user.","body":"Thanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.","comment_length":131,"text":"Saving Graph\/Structured Data in Datasets\nThanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.\nAn example of a toy dataset is like:\r\n```json\r\n[\r\n {\r\n \"name\": \"mike\",\r\n \"friends\": [\r\n \"tom\",\r\n \"lily\"\r\n ],\r\n \"articles\": [\r\n {\r\n \"title\": \"aaaaa\",\r\n \"reader\": [\r\n \"tom\",\r\n \"lucy\"\r\n ]\r\n }\r\n ]\r\n },\r\n {\r\n \"name\": \"tom\",\r\n \"friends\": [\r\n \"mike\",\r\n \"bbb\"\r\n ],\r\n \"articles\": [\r\n {\r\n \"title\": \"xxxxx\",\r\n \"reader\": [\r\n \"tom\",\r\n \"qqqq\"\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n```\r\nWe can use the friendship relation to build a directional graph, and a user node can be represented using the articles written by himself. And the relationship between articles can be built when the article has read by the same user.\r\nThis dataset can be used to model the heterogeneous relationship between users and articles, and this graph can be used to build recommendation systems to recommend articles to the user, or potential friends to the user.","embeddings":[-0.2233352512,0.1035750657,-0.0550303981,0.2553001046,0.1664751619,-0.0883487985,-0.0420417003,0.0489242636,0.1316880137,0.0172995813,-0.1817377806,0.341371417,-0.4227148592,0.5882750154,-0.2004154623,-0.25589028,0.2201137543,-0.0427169427,0.3299941719,0.0772028118,-0.0972134098,-0.0045653326,0.1638395786,-0.0592396893,-0.245912984,-0.5715228319,-0.1594649255,0.0477374308,-0.3542378545,-0.4808264375,0.2615370452,-0.0621038005,0.2797881067,0.1629059613,-0.000119489,0.0557818934,0.1224136502,-0.1544818282,-0.0868340507,-0.3309611082,0.0056771506,-0.2981207371,0.2152486444,-0.3602895141,0.0141386576,-0.52231884,0.0453781597,-0.4032321274,0.4603859782,0.0864814296,0.1160716712,-0.1106244102,0.3321138024,0.2047654837,0.1417979449,0.6720956564,-0.1498085856,0.3692043722,0.0153676271,0.1761931777,0.4121541381,-0.0536443405,-0.1396135986,-0.1696149707,0.3264270425,0.0092136934,-0.025214877,-0.3577064574,-0.1132700592,0.142080605,0.5881964564,-0.3921891153,-0.2603586912,-0.2230053246,-0.2509318292,-0.0681216493,-0.0151373716,0.1142061576,-0.2154325545,0.2317511141,0.1675646752,-0.3765946925,-0.2405542284,0.5173786283,0.0040855436,0.224382177,-0.0817498192,0.1435154378,-0.1506551653,-0.2028892338,0.05542266,-0.239014551,0.1657639444,0.019779142,0.0126152048,-0.0969205722,-0.4118796587,0.1422938555,0.0904703215,0.2531909347,0.1131650805,0.0034745373,0.2498451173,0.212865442,0.5089201331,-0.1791360974,0.0563799664,0.3806176484,0.3717682362,-0.3154613674,0.0485287569,-0.1778391749,-0.2879604399,0.1798106581,0.4373398721,-0.3225421011,0.5189981461,0.0792660937,-0.1570601612,0.0459627546,-0.1656923741,0.1057720855,0.0260034595,0.3263950646,-0.2364593893,0.1145993993,-0.127290681,0.5873839259,0.0855612308,0.3219810128,0.1119706407,-0.0072313198,0.0196744576,0.0170335937,0.1154987887,-0.0898103714,0.0913378447,0.0445006564,-0.2310366333,-0.2699974775,0.0820463523,-0.3541987836,0.2971456945,0.1638869345,-0.0560777448,0.1963002831,-0.0513299592,-0.3526574969,-0.0780925602,0.123974517,-0.214168191,-0.4256947041,-0.5040130615,0.0581177212,-0.315015614,-0.2132231295,-0.3674656451,0.1815354526,0.4864902794,-0.0217237826,-0.1316982061,-0.1553574353,-0.212444514,-0.5415158868,0.0038440039,-0.1813927591,-0.4119336307,0.0449710228,-0.4088256359,0.0243230797,0.0672061667,0.4411496818,-0.082571201,0.3443979025,0.0977900997,0.406766504,0.6312229633,-0.2161991,-0.4147624373,-0.0524231456,0.3830607533,0.1183177084,-0.1405139416,0.3945300579,0.509745419,-0.1003322005,-0.2299243808,0.1021637768,-0.0321283638,0.0126858987,-0.1281439066,-0.0780200809,0.3725044131,0.0612571314,-0.1228097677,-0.2798149288,0.2493323237,-0.1987602264,0.0786968321,0.0687674433,0.1506728977,0.1583855301,0.5065504313,0.4946050644,-0.0140913855,-0.3567608297,-0.4202125072,0.0701819435,-0.1202671453,-0.3514761627,-0.5534553528,-0.3246555328,-0.2232934237,0.2171337903,-0.1879822016,0.1065857261,0.0359379686,-0.0916216373,-0.1437990367,0.1442363709,-0.2992870212,0.0262181088,0.0541749597,0.0288671777,0.0494713858,0.3267568052,-0.1388345212,-0.5361049771,0.1419084817,0.2938562334,0.0160471667,0.0654351041,0.020452043,0.3442263305,0.2059183568,-0.1587535143,-0.2881811559,0.2673157752,0.0207438711,-0.3377420306,0.0994936377,0.0964006558,0.2454944104,0.0149286073,-0.1673009694,0.3099408746,0.3228692114,-0.0099839615,-0.1581581235,0.1404611617,0.1189101934,0.1425235868,-0.0679236501,-0.0623697378,-0.2031425983,-0.0518230461,0.0549689941,0.1553440541,-0.6136972904,0.124884963,0.4467571676,0.1083694026,0.5621824265,0.0328808688,0.003314625,-0.1703310758,0.2184522599,0.0379648097,0.216180563,0.0516937338,-0.1855855584,-0.110531278,-0.0116996532,-0.0349002294,0.028251458,0.1544921845,0.4188004732,0.1511193663,0.2366502136,0.0117272949,-0.0266444236,0.1544203609,-0.1849400699,-0.0042223101,-0.3672721386,0.2306480855,-0.1960738152,-0.2586520016,0.0719429404,-0.2118722647,-0.1705953479,-0.2689159214,-0.0383747257,0.0383530669,-0.0827082321,0.1073316187,0.4270685017,0.1616069824,-0.1902111322,-0.3846328557,-0.1311557889,-0.4501526058,-0.1093700379,-0.0138796307,0.1302648187,0.1611512005,0.4358736575,-0.0209491886,0.2071597129,-0.2987848222,-0.2861085832,-0.1119458824,-0.0864021927,-0.0002219397,0.0333612598,0.3284339905,-0.1627060026,-0.2232958674,-0.0354165696,-0.0906159729,-0.1039472148,0.0554788858,-0.0491634831,-0.0825529769,0.1610196978,-0.1338107437,-0.8484687209,-0.165742889,0.4261323214,0.2923759818,0.0407998636,-0.0893886313,0.0410691537,-0.1693545878,-0.1445688754,-0.0015500413,-0.005745261,-0.3910816014,0.5569652915,-0.297108084,-0.3576735854,0.2122366726,-0.1462715864,0.0288633723,-0.0205753241,0.0949492902,0.4115853906,-0.0594883896,0.4241309762,0.1043307558,-0.1644164026,0.2474941015,0.1327968836,0.1801097244,-0.064238511,0.1386542022,0.3372526467,0.0933458805,0.2491908222,-0.1027858853,0.0881969109,0.0299030598,0.2436656952,0.055164326,-0.2277391851,0.4077163041,-0.0759844184,0.2586165071,-0.0077718794,-0.0284326859,-0.0658639818,-0.1733735502,0.0518307574,0.0568675064,0.0884222314,-0.2805773318,-0.0161420051,-0.1658566892,-0.0078087579,-0.1966060847,0.1610701233,-0.5294158459,0.3078149557,-0.0815821066,-0.1512292922,0.000750009,-0.1890054643,-0.0780671611,0.2192106098,0.0345448852,0.2282939255,-0.0131349359,0.0290684476,-0.4362006783,0.1897961199,0.268563211,0.0482558012,-0.1776027828,0.0300147794,0.175556466,0.1354052126,-0.0394631587,-0.081386067,-0.2729462683,-0.1763349324,0.0311280414,-0.4062754512,0.1925645769,-0.0398803838,0.065726541,0.0311273038,0.4377597272,-0.4519427717,-0.027209809,-0.1801562756,0.2542631626,-0.3058665991,-0.1896768659,-0.1122757271,-0.2221027762,-0.4085019827,-0.2553930283,0.4043835998,-0.0842117369,0.1050650552,-0.1419720799,-0.399253577,0.1159557253,-0.1226835027,-0.1181738749,0.4155303836,-0.4326835275,-0.1347812414,0.265237391,0.2992002666,0.3488849998,0.3828358948,0.1606918871,-0.2972449362,-0.174863264,0.0539425202,0.310757041,-0.2941873372,-0.1882867515,-0.0540661067,-0.0418896489,-0.0992156416,-0.2898101509,-0.0834393129,0.2686804235,-0.0104876244,-0.6994720101,-0.4712933302,0.3735628426,0.1548803747,-0.0439873151,-0.1446902454,0.1700032055,-0.149175629,0.5449377298,0.3925794661,0.8639636636,-0.137297526,0.3636466265,0.2922455966,-0.489328742,0.4083568454,-0.0765318871,-0.3226093352,-0.3465977311,0.3702889681,-0.1670933813,0.0797833279,0.2839083076,0.1858793795,-0.3257012367,0.034268789,-0.2680805922,-0.2359673679,-0.03075625,0.431848973,-0.0221588854,0.0607449226,-0.2921316028,0.0805914253,-0.0678455159,-0.0000774691,-0.1068580896,-0.0424481295,-0.3113137782,0.017347591,-0.2258692682,-0.2482300103,0.0091149006,0.3994217217,0.1247969717,-0.3601287007,0.2805285752,0.4910919964,0.2668821514,0.0105500771,-0.2003247142,0.0108856335,0.0244670324,-0.1112179831,0.241052255,-0.1784822792,0.33447963,-0.1055420861,-0.2018777579,-0.060170453,-0.0311829988,-0.1187306195,-0.2716255784,-0.0526388623,-0.1225861013,-0.2589848042,-0.01063905,-0.4220399559,-0.0879995674,-0.0030057083,0.0585306399,0.3459264934,0.194560647,0.3843897283,-0.1221578121,0.07462091,-0.0483520143,0.1350472271,-0.0138579728,0.3325591087,0.3632879853,-0.0549686737,-0.2740263939,-0.1561952978,0.4622799754,0.0958908647,-0.0743362382,0.1519630402,0.0807852447,0.0521466471,-0.3380199075,0.426867485,0.1066436917,-0.3751498163,-0.0491338857,-0.3442888856,-0.3983802497,0.2167146206,0.159757778,0.2631758451,-0.0089030191,0.1021148786,0.0464648083,0.0376727059,-0.1907521337,-0.0286247358,-0.1103163511,0.130376488,0.1121982187,0.0385999307,-0.0410784855,-0.2520422637,-0.0201742984,-0.1345951408,0.1123709455,-0.0853066072,-0.1513262838,0.1684648097,0.1194574162,-0.1256007999,0.0726765096,-0.3188901544,-0.1045239717,0.1903571039,-0.2234043181,0.3806156516,-0.0345957875,0.1050267369,0.0083016781,0.0254154745,0.0691225603,0.1472540796,-0.2052019387,0.099452123,-0.2112653852,0.1887049079,0.0214152504,-0.226731807,0.2309467047,0.1460653692,0.2607263923,-0.1214080229,0.4373712838,-0.2486296147,0.0087226843,-0.197149232,0.3064255118,0.2360520065,-0.0393494777,0.1593473405,0.1976500154,0.0569279939,-0.1567937732,0.0267651081,0.1081993654,-0.1126913279,-0.0967968628,0.1712249368,0.1399676949,0.5176048279,-0.719239831,-0.0999459699,0.1777595282,0.0955881625,0.1979864836,-0.000753797,-0.0449617617,-0.0323677845,0.1952230036,0.3257002234,0.2375146002,0.4222401679,-0.0629470795,0.1541092545,0.4920126796,-0.0420317054,0.5321179628,-0.0603517964,-0.0898596272,-0.3307671249,0.0903912559,-0.1168531105,0.301903218,0.3768695891,0.1933685839,-0.3365649581,0.1037804261,0.5163173079,0.0221250504,-0.1533877701,-0.2020446509,-0.0388360433,-0.1057575196,-0.3559310138,-0.3709657192,-0.5535500646,0.0485522971,0.2532098889,0.2285641283,-0.4509907961,0.5552634001,0.0018845377,0.2011065334,-0.0249459352,0.2227440029,-0.1253732592,-0.1210839897,-0.0836583301,0.0906806588,0.4870869517,0.228807047,-0.2157652676,0.4297358692,-0.1882042289,-0.3056252301,-0.0229485091,0.0028567116,0.0876649693,0.210530594,0.4193348587,0.0305864979,0.0448114164,0.1424073577,0.1101067811,0.3276667893,-0.0487611853,0.4861747324,-0.2393774539,-0.2932775021,0.0763849914,-0.0813082829,-0.1369047314,-0.2732716501,0.3981994987,0.0815889165,0.2730681598,-0.3055604994,0.0256105158,0.0092547843,0.6038179398,0.0910225883,0.5298618674,-0.0217415802,-0.0480633788,-0.3787655234,-0.1793809682,-0.1186905429,-0.1883232594,0.1057669893,0.0931359306,-0.0406337827,0.4481936693,0.0297932867,0.5166373253,-0.0900732055,0.1324812621,-0.1779606193,-0.0411554202,-0.1122468188,0.1276655495,0.1043585017,-0.076900579,0.2495915741,-0.0289029106,0.0344782136,-0.1968002915,-0.0419660136,-0.0598189831,-0.4242297113,0.2686575651,0.0139557254,0.2215149403,0.1019033715,0.1035749987,0.2701401711,-0.0107207708,-0.2577411234,0.432690084,0.0173458327,0.4144567549,-0.3793774843,0.3594712615,-0.0173909571,-0.2136908472,-0.1642682999,-0.0564805418,-0.1417870671,0.0088394359,0.1383752972,0.0567426607,-0.030434655,0.1238712892,-0.2212103754,0.1201573387,-0.2631855607,-0.1337825507,0.4790452123,0.1374792457,-0.1566676497,0.1865205765,-0.0100894021,0.465260148,0.0785822868,-0.3664717674,0.0568115823,0.4066559374,-0.1128983498,0.0361753404,0.0049631768,0.1560529917,-0.061094027,-0.0728053749,0.5850744247,-0.1229251623,0.2136714756,0.0621661991,-0.339114517]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2426","title":"Saving Graph\/Structured Data in Datasets","comments":"Hi,\r\n\r\nyou can do the following to load this data into a `Dataset`:\r\n```python\r\nfrom datasets import Dataset\r\nexamples = [\r\n {\r\n \"name\": \"mike\",\r\n \"friends\": [\r\n \"tom\",\r\n \"lily\"\r\n ],\r\n \"articles\": [\r\n {\r\n \"title\": \"aaaaa\",\r\n \"reader\": [\r\n \"tom\",\r\n \"lucy\"\r\n ]\r\n }\r\n ]\r\n },\r\n {\r\n \"name\": \"tom\",\r\n \"friends\": [\r\n \"mike\",\r\n \"bbb\"\r\n ],\r\n \"articles\": [\r\n {\r\n \"title\": \"xxxxx\",\r\n \"reader\": [\r\n \"tom\",\r\n \"qqqq\"\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n\r\nkeys = examples[0].keys()\r\nvalues = [ex.values() for ex in examples]\r\ndataset = Dataset.from_dict({k: list(v) for k, v in zip(keys, zip(*values))})\r\n```\r\n\r\nLet us know if this works for you.","body":"Thanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.","comment_length":93,"text":"Saving Graph\/Structured Data in Datasets\nThanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.\nHi,\r\n\r\nyou can do the following to load this data into a `Dataset`:\r\n```python\r\nfrom datasets import Dataset\r\nexamples = [\r\n {\r\n \"name\": \"mike\",\r\n \"friends\": [\r\n \"tom\",\r\n \"lily\"\r\n ],\r\n \"articles\": [\r\n {\r\n \"title\": \"aaaaa\",\r\n \"reader\": [\r\n \"tom\",\r\n \"lucy\"\r\n ]\r\n }\r\n ]\r\n },\r\n {\r\n \"name\": \"tom\",\r\n \"friends\": [\r\n \"mike\",\r\n \"bbb\"\r\n ],\r\n \"articles\": [\r\n {\r\n \"title\": \"xxxxx\",\r\n \"reader\": [\r\n \"tom\",\r\n \"qqqq\"\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n\r\nkeys = examples[0].keys()\r\nvalues = [ex.values() for ex in examples]\r\ndataset = Dataset.from_dict({k: list(v) for k, v in zip(keys, zip(*values))})\r\n```\r\n\r\nLet us know if this works for you.","embeddings":[-0.2856687903,0.0489118472,-0.0369235054,0.2954734564,0.1761306524,-0.0925177932,0.0076426757,0.094109267,0.1752270609,0.0359997712,-0.2266489267,0.414647758,-0.4315076768,0.6326939464,-0.1680036932,-0.1835950762,0.2540252805,-0.0057012681,0.2815619409,0.067555666,-0.1143528074,0.0210009422,0.0942106321,-0.1012063399,-0.2402513772,-0.4843244255,-0.1383965611,0.1061426699,-0.2759398222,-0.5049470663,0.2628543377,-0.0757849887,0.3589752614,0.2296425849,-0.0001212197,0.0293496773,0.1541253775,-0.1582915485,-0.1681702733,-0.2760806382,-0.0338869095,-0.3862594068,0.2224015743,-0.352077812,0.0265035257,-0.5047537088,-0.0631149709,-0.438708365,0.4396774471,0.1020215526,0.1048100814,-0.0953225791,0.2954246104,0.189921692,0.0589852892,0.6107455492,-0.137282595,0.3834092319,0.0041535716,0.1410946846,0.4431878626,-0.0102145523,-0.2097631842,-0.1612381041,0.3201059997,0.0171319246,-0.0291725714,-0.3698993623,-0.1352078766,0.1660224795,0.5845979452,-0.3652295768,-0.2983254194,-0.3096181154,-0.228136152,-0.1148663908,0.0039704191,0.110199824,-0.198173821,0.2807420194,0.2201821208,-0.3308597505,-0.1807452291,0.4849873483,-0.0611000247,0.146174714,-0.1192958355,0.1419625729,-0.1227738187,-0.1803835034,0.0996391997,-0.2622883916,0.2115786076,0.05403658,-0.0296069365,-0.0414730161,-0.3642190099,0.124395512,0.0635046139,0.2624091208,0.1064799502,-0.013124737,0.2416505665,0.2026436925,0.4457499087,-0.2023298144,0.0707388073,0.4032110572,0.2888768613,-0.261646986,0.0439677425,-0.1640574038,-0.3500635922,0.0812721327,0.4830957651,-0.340732038,0.5467669368,0.0318036415,-0.1989638209,0.0567570068,-0.2073794156,0.0930052772,-0.0059705758,0.3341024816,-0.2197184861,0.1738167852,-0.1130866855,0.6057108045,0.0968959257,0.2678085268,0.0694801584,-0.0594241396,0.0333949141,0.0329499394,0.1419242173,-0.1907092333,0.045240175,0.0679358542,-0.1966766864,-0.3019955158,0.0097883483,-0.3076047003,0.3215050101,0.1786852032,-0.0007942825,0.2441357523,0.0554694794,-0.333433032,-0.0856918544,0.1601164341,-0.2456672341,-0.4013805389,-0.470400691,0.0464392267,-0.270311147,-0.2009264976,-0.423334837,0.1407182515,0.4591409564,-0.0270312373,-0.1629072279,-0.1851441264,-0.2272674441,-0.5391336679,0.0616365634,-0.130131498,-0.491052717,0.0238610953,-0.4131832123,0.033468686,0.0880380198,0.4129672348,-0.0644395277,0.3483416736,0.0359830782,0.3879745305,0.6230746508,-0.2450202405,-0.4455579817,-0.0123277232,0.4037216306,0.1489109099,-0.1765155643,0.3048385382,0.5396374464,-0.0855405629,-0.2140922397,0.1027064994,0.0024445693,0.0117912684,-0.1402405351,-0.045989152,0.4148258567,0.0751411691,-0.1731809676,-0.2115935981,0.1960125417,-0.2831936777,0.1302322745,0.0918998718,0.1341879368,0.1718764603,0.4094119668,0.5433160663,0.0042917985,-0.3294644058,-0.3844675124,0.1148533821,-0.1275493354,-0.3256943524,-0.5862291455,-0.2956007421,-0.2968480289,0.2654924691,-0.2253479362,0.1046174839,0.0233859792,-0.1460328102,-0.1592219323,0.1599626839,-0.3197219372,0.0339863636,0.0065495037,0.0329063013,0.0861336514,0.3561074734,-0.0928437412,-0.518142581,0.1240737587,0.2805194259,0.0706996918,0.1026900634,-0.0418438092,0.2740919292,0.2330722362,-0.2299735099,-0.3369737864,0.2312617749,-0.022167882,-0.3557398021,0.126083374,0.1292925179,0.2492378354,-0.0251101404,-0.1084318608,0.3343934715,0.27631706,0.0176386684,-0.2407128066,0.1901579052,0.1251969486,0.1463831514,-0.0831875801,-0.0966418684,-0.2315422595,-0.0299728718,0.0397298895,0.1298664361,-0.5606173277,0.07129246,0.3628565073,0.0552307628,0.4966156483,0.0064899344,-0.0281091835,-0.1260257065,0.234453842,0.0343133286,0.2439900935,0.0448081456,-0.1889482588,-0.1231997237,0.0325147659,-0.0325880311,0.0398155078,0.1706948131,0.4787935615,0.1793331206,0.1353459209,0.0705786571,-0.0077467966,0.1529106349,-0.1868605614,0.0215745121,-0.3932467997,0.2597412765,-0.1891775876,-0.3411652446,0.0717251301,-0.1798442006,-0.1124335229,-0.2484556437,-0.1268178672,0.0519933403,-0.0120929889,0.180280745,0.4102582932,0.1516159475,-0.1177892238,-0.4625420272,-0.1461651623,-0.4068709016,-0.108137019,-0.029054001,0.1182833463,0.129303202,0.3737321794,-0.0375159755,0.2248944044,-0.2988058329,-0.1950057596,-0.1388469785,-0.0749744102,-0.0545152053,-0.0024956635,0.3713651001,-0.1401944757,-0.1446201652,-0.0166386869,-0.1091725528,-0.069999665,0.0943960547,-0.0415677279,-0.0403397083,0.2053181529,-0.0593520068,-0.8583173156,-0.1564684808,0.3328521252,0.3435877562,0.0337750241,-0.0736743361,0.1039049029,-0.1537514925,-0.1172204241,-0.0354967192,-0.0155316768,-0.2680869699,0.6005049348,-0.2475970387,-0.3712285459,0.2166536897,-0.1438214034,0.0600721203,-0.0142243141,0.0875076354,0.4726675749,-0.0464403443,0.4293411076,0.0283014551,-0.1236563548,0.2986907065,0.1369947642,0.1880536079,-0.0151023027,0.176908493,0.2579629719,0.0381058529,0.163757652,-0.0598191656,0.1514721066,0.018788632,0.2478972524,0.1197220013,-0.1546406895,0.4252412021,-0.1102481782,0.2781702876,-0.0261385851,-0.1024712473,-0.1288133711,-0.150689438,0.0448802188,0.0461362042,0.0839887112,-0.2124664485,-0.0288285781,-0.1592271924,-0.0319373347,-0.1974444538,0.1420326978,-0.5685598254,0.3177016079,-0.0778088421,-0.1205713153,-0.01379955,-0.1891199797,-0.0773552135,0.1908205003,0.0961592421,0.2022231966,-0.0088536739,0.1580521017,-0.3908388317,0.2159328759,0.2422118038,0.0526715033,-0.1613968313,-0.0098298267,0.1805404127,0.2001651078,-0.0332736075,-0.0504457615,-0.2560588121,-0.133578971,0.032940831,-0.5067497492,0.118435286,-0.0100114876,0.0254855268,0.0300400667,0.3673708141,-0.4127818942,-0.0161992349,-0.1694826484,0.3126722872,-0.3179882467,-0.2013420463,-0.1354960799,-0.192778036,-0.5086798072,-0.2729519904,0.4607399106,-0.0016771553,0.0428382009,-0.1104805768,-0.4287120402,0.1029512882,-0.1134523377,-0.1434390396,0.4799451828,-0.4377681613,-0.127734527,0.2868560255,0.2002214342,0.3274329305,0.4442690909,0.0914469138,-0.401863426,-0.186556071,0.0629839972,0.2902643681,-0.3098469079,-0.2368327677,-0.0595165528,-0.0432763211,-0.1370960772,-0.3614167571,-0.103391327,0.2624729872,-0.0037975826,-0.7144719958,-0.539575994,0.4125770926,0.197072506,-0.027719019,-0.0266004279,0.1530395299,-0.1341238469,0.5577990413,0.2633269131,0.8369047046,-0.1708225608,0.3860082626,0.3709448874,-0.3662647903,0.3103200197,0.0348481461,-0.3661974072,-0.3733038604,0.3256160319,-0.1508176625,0.0435672626,0.2262845188,0.2361872047,-0.2926365733,0.0447802767,-0.3811260164,-0.2370942384,0.016040843,0.3597187102,-0.0095487004,0.0325434543,-0.3372530043,0.078279309,-0.0309544206,0.0969293863,-0.0989189371,0.0048985584,-0.3105246723,0.0184317715,-0.2310585827,-0.1915613711,0.0278338771,0.3982832134,0.1025097668,-0.3914256394,0.2562240958,0.5818545222,0.1734419763,-0.0048548859,-0.1794057488,0.0245774072,0.053972248,-0.1081004217,0.2703412771,-0.1511366218,0.3229625523,-0.0809687823,-0.1821543276,0.0223522149,-0.0114421444,-0.1420909911,-0.1729144156,-0.0079508061,-0.0497271791,-0.3564708531,-0.0662157759,-0.4418888986,-0.0266374405,0.0121281622,0.0438930616,0.3539246619,0.1604397744,0.3738444746,-0.1117231175,0.0817152113,-0.0045218114,0.152437225,-0.0543767922,0.2796642184,0.377810061,-0.0395720713,-0.2724963427,-0.1358537227,0.5030568838,0.0316752233,-0.0970069394,0.2096060961,0.1103025377,0.1278747171,-0.3132586181,0.4684441686,0.1165468246,-0.3360753357,-0.0291698389,-0.3738365173,-0.3828911781,0.197391659,0.1345099062,0.2569147646,0.0327695385,0.0586008281,0.0130570913,-0.0410902798,-0.1750811338,0.0078683253,-0.0504735187,0.0552771725,0.1755063832,0.0495285653,-0.0223344788,-0.2421898395,-0.0060716677,-0.1527203172,0.1510724574,-0.0859100968,-0.1604235619,0.1865711808,0.0894698724,-0.1686234176,0.0824473426,-0.3796802759,-0.1631352752,0.1391598135,-0.2217632234,0.4110220075,-0.0309099928,0.1000932679,0.1153017879,0.0248635877,-0.0379967764,0.224134773,-0.3059077859,0.1076153666,-0.2326215059,0.1671913117,0.0619522855,-0.1668602675,0.1264882833,0.181488812,0.2913072109,-0.1705123931,0.4657720029,-0.3192029595,0.0873787031,-0.1996007264,0.3123028874,0.2466650903,-0.1209041551,0.1089831889,0.2264735103,0.0490045212,-0.182328701,0.0479788221,0.0929598659,-0.0954233631,-0.1176793128,0.2002972662,0.1202904508,0.4316860139,-0.7079563141,-0.0772754401,0.1330684721,0.0625941455,0.1939899772,0.0485740788,0.0312315729,-0.0131815635,0.2086907625,0.3057984412,0.1872418225,0.4361662567,-0.0657551363,0.1167403013,0.4263691008,-0.0823037624,0.4778571129,-0.1740698814,-0.0776170939,-0.3145121634,-0.0231841449,-0.1125490442,0.3007282317,0.382478714,0.1244913414,-0.3221501112,0.1214069575,0.484528929,-0.0043229284,-0.0835251063,-0.1307428181,-0.0245633293,-0.0537144542,-0.3768719435,-0.3602025211,-0.5454347134,0.0062992941,0.2937610745,0.2227457911,-0.4700936675,0.4952847064,0.056502454,0.2026818693,-0.0080148671,0.1636492163,-0.1567015946,-0.0396030806,-0.0184499435,0.0724216029,0.5308017731,0.2071003318,-0.1774478257,0.4039839804,-0.1485021114,-0.2613715231,0.0007312517,-0.0535338894,0.0601243936,0.2574847043,0.4366300106,0.0371917002,0.0439021438,0.0237559173,0.0899638534,0.2652653456,-0.0644070432,0.5160330534,-0.2475353628,-0.2820348144,-0.0111838467,-0.0204453301,-0.0623748004,-0.2469958514,0.4320987165,0.1445878595,0.2569583952,-0.2672492266,0.020395996,0.0289813932,0.5063236356,0.1182213426,0.6070608497,-0.0375145525,-0.0313390829,-0.4656193852,-0.1290886402,-0.1342174411,-0.224941045,0.1277496666,0.0706052929,-0.0366853029,0.5012503862,0.0455979519,0.4711269438,-0.0299162772,0.1629820615,-0.1218772307,-0.1208420247,-0.1137338877,0.1352229714,0.0601317808,-0.100780569,0.2258527577,-0.0911611915,0.0519866236,-0.2024913281,-0.083968401,-0.0687900707,-0.3774838448,0.2490080148,-0.0397431962,0.2931400239,0.1143603846,0.1087013111,0.2703839242,-0.0232464895,-0.2724152207,0.4581118524,-0.0268511176,0.3115202785,-0.3828103542,0.3603460789,-0.0382879302,-0.2142516673,-0.1353540868,-0.0859589726,-0.1741978824,0.0308947265,0.1518740505,0.1291862875,-0.0300155077,0.0815060437,-0.2136297822,0.1677756608,-0.2586874068,-0.1427021623,0.5019131303,0.1070627794,-0.1688990146,0.214993313,0.0102932975,0.4040285051,0.1307591796,-0.3111591935,0.0307093412,0.4374919832,-0.1101523861,0.0601362437,0.057638552,0.1347296536,-0.0329809859,-0.1160571873,0.5957214832,-0.1341553032,0.2520398498,0.0268845689,-0.3416142166]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2426","title":"Saving Graph\/Structured Data in Datasets","comments":"Thank you so much, and that works! I also have a question that if the dataset is very large, that cannot be loaded into the memory. How to create the Dataset?","body":"Thanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.","comment_length":31,"text":"Saving Graph\/Structured Data in Datasets\nThanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.\nThank you so much, and that works! I also have a question that if the dataset is very large, that cannot be loaded into the memory. How to create the Dataset?","embeddings":[-0.2723234296,0.0077559124,-0.0335844606,0.331335783,0.1474175602,-0.0770562068,-0.0660822317,0.0120472154,0.0752707794,0.1229880601,-0.173162654,0.313380152,-0.4319543242,0.5779833198,-0.0695673078,-0.2389570773,0.276214838,0.0760909691,0.2164649516,-0.0062684393,-0.1167420521,-0.0144838151,0.0624190047,-0.1413793266,-0.2844819725,-0.4980396032,-0.1976485997,0.1711293012,-0.4135592282,-0.498990953,0.2010753453,-0.0681336671,0.3138073087,0.2875531018,-0.0001202716,-0.0988923907,0.1517480463,-0.213938266,-0.0875948146,-0.2176559567,-0.0626532882,-0.3494538665,0.1856263429,-0.3744631708,0.0488128364,-0.461779207,0.0895982981,-0.2455245554,0.3802903295,0.2123028189,0.1083455458,-0.1020672917,0.3337271512,0.1387737095,0.2244872898,0.5536146164,-0.1602343321,0.3449639678,0.0203347094,0.184986949,0.3843227625,-0.1057037115,-0.1318714768,-0.1821804941,0.397329092,-0.0415188149,-0.0887694955,-0.4944135845,-0.0841982812,0.1948035955,0.7593106031,-0.4477229416,-0.2266944349,-0.1608174443,-0.2214411497,-0.0948323086,0.0108969165,0.2465351522,-0.2418276668,0.2171609253,0.1256919056,-0.4189335406,-0.3556168973,0.4478442967,-0.0639713556,0.2398026288,-0.0703294426,0.1305978894,-0.1145715863,-0.0866908953,0.0702921525,-0.1638026536,0.1005724594,0.1569358408,-0.0077922223,-0.0993697867,-0.4392795265,0.2444678545,0.131171748,0.2988360226,0.0962082222,-0.0442811921,0.3254297972,0.2030017674,0.4563657939,-0.1295352578,-0.0343310535,0.3509793282,0.4708310664,-0.4111912251,0.0793208852,-0.1461351365,-0.32575804,0.1213463098,0.4596614242,-0.3857610226,0.4507372975,0.018676294,-0.0674884617,0.0762663037,-0.1422474682,0.0838796049,-0.020079799,0.3746221066,-0.2849635184,0.0356439836,-0.1820646822,0.5128691792,-0.0016849629,0.2907336354,0.1076042131,0.0358643532,-0.0395977162,0.0730571374,0.1775100082,-0.15936023,0.1261921823,0.1004850268,-0.1680474877,-0.1745099723,0.1169634685,-0.3805246055,0.2957401872,0.2588338852,0.0160455797,0.1812125444,-0.150212273,-0.2076259255,-0.0921968967,0.2719935477,-0.1546025276,-0.539816916,-0.5330708027,0.0546012931,-0.2730056643,-0.1832983792,-0.3185444772,0.1531112045,0.4208410382,-0.0368215814,-0.0124009838,-0.1630356759,-0.1278664768,-0.6147217751,0.0856520161,-0.1255210489,-0.4851516187,0.1229011342,-0.3247629702,-0.0987713113,0.1090563759,0.4521107972,-0.07078702,0.3401065469,0.0563110188,0.3587877154,0.7113128901,-0.218078509,-0.4243231714,-0.070299536,0.4794392884,-0.0058620842,-0.057980042,0.3966946304,0.5118689537,-0.0480294339,-0.3228681684,0.130073607,-0.0515295602,-0.0404157937,-0.1537346542,-0.0078461068,0.3993906081,0.0581793077,-0.0624420084,-0.2450851649,0.2789312899,-0.1595264077,0.0747456029,-0.0010716429,0.1093669608,0.1044967324,0.4632634521,0.5171387196,0.0917439237,-0.2976774275,-0.3671291471,0.0860744566,-0.2413827032,-0.2384937704,-0.4375576377,-0.2809787095,-0.2324994802,0.184506461,-0.1535614282,0.0133439358,-0.0315263346,-0.1356746554,0.0234829839,0.159497261,-0.4104891121,0.0924570039,-0.0393071696,0.0162757635,0.0546815991,0.2408659607,-0.1573641151,-0.5781391859,0.1061069295,0.301720202,-0.0689623728,0.0416807383,0.0111880675,0.3606559932,0.2074419707,-0.1765874177,-0.3877708912,0.2146490812,0.0640983433,-0.3676768243,0.1983052939,0.0748228133,0.2533234954,-0.02154967,-0.1256420463,0.200494796,0.2751399875,-0.1084681824,-0.0767166615,0.0466155261,0.1586217582,0.1110260859,-0.0413988233,0.0030366348,-0.1252223104,-0.0301832072,0.1659463048,0.1786604971,-0.687571764,0.025208259,0.4318841398,-0.0037641011,0.4825694561,-0.0363643691,-0.1058223397,-0.1832434237,0.2555556893,-0.0549234636,0.2452308983,0.0651387647,-0.1634278744,-0.0654099584,-0.057419017,-0.1317062527,0.0382650271,0.1137054339,0.4122775495,0.183090955,0.1618249565,-0.0466056988,-0.1662721485,0.1843385845,-0.1508415788,0.0023883434,-0.3080722094,0.143203631,-0.0645182356,-0.4157723784,0.0653207153,-0.1595942676,-0.2017191797,-0.1290533543,0.0060436875,0.0526842847,0.0520017371,0.1046508104,0.411814183,0.2988615334,-0.1228376329,-0.2645308971,-0.2101124376,-0.3297555447,-0.1263043582,-0.0112475362,0.0675989836,0.1480524838,0.4714926481,-0.0406836756,0.2307069451,-0.3062478006,-0.1926319599,-0.0408775695,-0.1002945676,0.0242455732,0.0078032231,0.3974918425,-0.0643022507,-0.1873313934,-0.0652493984,-0.2046804577,-0.1442505866,0.0302599259,-0.1169730499,-0.0828620344,0.121373646,-0.1776868105,-0.8541564941,-0.1905872524,0.4202074707,0.3776107728,0.0883553177,-0.0926016644,0.044257585,-0.0691745207,-0.110158816,0.0476266779,-0.0483815484,-0.4029603899,0.6650779843,-0.3638497889,-0.4688949585,0.2904838324,-0.1085298881,0.1245299429,-0.0135353534,-0.0370056778,0.4082781076,-0.1297804862,0.402040571,0.0598957911,0.0274004787,0.2436969727,0.0090727108,0.1797289401,-0.0924161375,0.1197548881,0.2141741216,0.134723261,0.345688045,-0.0561615191,0.1580355465,-0.0328666791,0.3319851756,0.0610183068,-0.0342557207,0.4159286618,-0.1219007596,0.3614265025,-0.0465202145,0.0066815489,0.0552164651,-0.1459611803,-0.0111309374,0.1058525518,0.1118346229,-0.3828292489,-0.0498012304,-0.1883652061,0.0158545524,-0.0781419873,0.1724596471,-0.4791879952,0.2654161751,-0.1513158083,-0.1574131697,0.0224548336,-0.1981697828,-0.0630537271,0.1586485952,0.0935719311,0.2189332098,-0.0522055291,-0.005063124,-0.493866384,0.1305492073,0.2328785509,0.0222804286,-0.144769907,-0.0269361511,0.2809049487,0.134988308,0.1031431034,-0.0445408225,-0.331562072,-0.1095427498,0.0517794751,-0.4385114014,0.1658964455,-0.0629877001,0.0676268116,0.0451608524,0.2955589592,-0.3560398817,-0.0140203545,-0.1216108575,0.3484960198,-0.3610788882,-0.1794099212,-0.1570308357,-0.2636582255,-0.4634796679,-0.22464405,0.3272734582,-0.0851483867,0.2009483129,-0.1607329398,-0.3252104819,0.0007557994,-0.1558558643,-0.0221047848,0.4108929634,-0.4057790041,-0.1393278539,0.3009976745,0.254001528,0.3001542389,0.4777843952,0.0116620241,-0.271230191,-0.0929112881,0.087177068,0.4172769487,-0.2592752874,-0.2191576511,-0.0518106706,-0.1136209592,-0.0823026448,-0.344452858,-0.0657454208,0.2050578296,-0.0459362604,-0.7719864249,-0.5602840185,0.3869167268,0.2202914953,0.0203528889,-0.043257799,0.2119410783,-0.1232543737,0.5424807668,0.3151690364,0.9842785001,-0.2049352229,0.4104818702,0.312466979,-0.3715561032,0.3692463636,-0.1294334233,-0.2402048409,-0.3524491787,0.3045667708,-0.1524689198,-0.0004674518,0.2512142062,0.1252006143,-0.3326058984,0.0366881378,-0.2355013043,-0.2196179032,-0.07064116,0.5262088776,-0.0947105214,0.0324643999,-0.317235589,0.0677489787,0.0741962641,0.0060955305,-0.1172755137,0.0133996485,-0.1907144487,0.0263571832,-0.2165495455,-0.2170370668,-0.0563148148,0.329503119,0.0704746097,-0.4314472675,0.1900761127,0.5449610353,0.2234722078,0.030778816,-0.2126306444,0.0420063883,0.0630682036,-0.16717273,0.1841310412,-0.0465420522,0.4224024117,-0.179425478,-0.1933525205,0.0541544743,-0.1087846681,-0.2336307317,-0.3441565335,-0.0660136268,-0.0137215145,-0.1989692301,0.0267663319,-0.4472454488,-0.0417771786,0.0314723998,0.0567563325,0.3994976878,0.1630465984,0.4689699113,-0.113103658,0.0200668462,-0.1455467194,0.0673469231,0.0898702592,0.2726691067,0.2975699604,0.0182563663,-0.2582915723,-0.1258302331,0.4669548869,0.1385335922,-0.1383716762,0.1922924221,0.1118066013,0.0984766856,-0.3236134052,0.5095869303,0.2069763094,-0.39881742,-0.1580053121,-0.2553685308,-0.4718832672,0.2013965994,0.1836304665,0.3284221292,0.0582956634,0.0409693681,0.1304242462,0.1056583896,-0.1835722923,-0.0439128876,-0.0952205434,0.1519567668,0.1714344919,0.0804858059,-0.0600729659,-0.2175820023,-0.0571822375,-0.1220174134,0.0932596698,-0.0788313672,-0.1427465379,0.2128281146,0.1266997606,-0.1882382333,0.0565320626,-0.438497752,-0.0879148096,0.1547568142,-0.0857401937,0.3494836986,-0.0872718319,0.0425749011,0.1263460815,-0.0152615858,-0.0318190344,0.201622203,-0.1714976877,0.1312598735,-0.1559431106,0.1554667503,0.0338128209,-0.2313486487,0.2518280745,0.1107576415,0.280708462,-0.2182866782,0.3831951916,-0.2193217278,0.0598401986,-0.2706255615,0.3563504517,0.3107167482,-0.0594900176,0.169294849,0.3296419978,0.0168050956,-0.1965285987,-0.0234868359,-0.0028800655,-0.14995116,-0.1149393097,0.1819030643,0.2069322765,0.3982496858,-0.7131308317,-0.0477532707,0.1602548361,0.0673815534,0.0853356197,0.0466799811,-0.1553267837,-0.0143554574,0.2013260722,0.3755819798,0.1560039222,0.5233569145,-0.0484213419,0.2199649811,0.4708999693,-0.086634405,0.5086176395,-0.060634274,-0.128753081,-0.3366001844,0.0302630514,-0.1679256409,0.2574034333,0.3914016187,0.1623525918,-0.2951770127,0.0879093856,0.4683952928,-0.0019766451,-0.0939678699,-0.1216472089,-0.1184937581,0.042773135,-0.2274584025,-0.3452318311,-0.5257751346,0.1253990084,0.2845238745,0.1502968967,-0.3566370308,0.574862659,-0.0152891818,0.1335635036,-0.0234063771,0.2750990987,-0.1628663242,-0.1051651239,-0.1668434441,0.1493135989,0.4687004983,0.3185720146,-0.1878052056,0.4099747837,-0.1455452144,-0.2621717751,-0.0087765548,0.0808148608,0.1458844543,0.1913474649,0.3824696541,-0.0117541216,0.0504341647,0.1233558059,0.1379035115,0.178033039,-0.0310949013,0.4093840718,-0.2234399319,-0.288064599,0.0133323688,-0.003194476,-0.2659605443,-0.195313409,0.3789481521,0.0326267891,0.2002425045,-0.3258498311,0.0094099836,-0.0000613016,0.6366890073,0.1890742481,0.6842777133,0.0008040303,-0.1100557372,-0.4394081831,-0.0787859336,-0.1487989575,-0.1403146982,0.0441944785,0.1181310266,-0.0763417855,0.4245951772,0.1432764977,0.4282594323,-0.0480072647,0.1388755143,-0.0844316259,-0.0526339933,-0.1193915606,0.1963618398,-0.0268599745,-0.0602936335,0.3545543551,-0.1976439804,0.0021772138,-0.1862405986,-0.033030767,0.0260430276,-0.3248854578,0.3500730097,-0.0369290486,0.2633863688,0.0643498972,0.1154605821,0.3198368847,-0.0296797976,-0.2034392655,0.3049371541,0.1073152795,0.2416838855,-0.3868366778,0.2544582784,-0.0254418086,-0.1201010123,-0.2326817513,0.0141728269,-0.1213488504,-0.0236025564,0.0822708011,-0.0074760742,0.0068533635,0.0909060314,-0.2743730843,-0.0275851637,-0.2920644879,-0.104169257,0.4838922322,0.022296913,-0.1679777503,0.0180601459,-0.0405054353,0.4224330187,0.1209572777,-0.4229340255,0.0966996178,0.3826317489,-0.1275714487,-0.0449397862,0.104165487,0.1581151634,-0.0396687947,-0.0948925465,0.6039980054,-0.1448592246,0.1285214275,0.0327844135,-0.4027280211]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2426","title":"Saving Graph\/Structured Data in Datasets","comments":"If your dataset doesn't fit in memory, store it in a local file and load it from there. Check out [this chapter](https:\/\/huggingface.co\/docs\/datasets\/master\/loading_datasets.html#from-local-files) in the docs for more info.","body":"Thanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.","comment_length":28,"text":"Saving Graph\/Structured Data in Datasets\nThanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.\nIf your dataset doesn't fit in memory, store it in a local file and load it from there. Check out [this chapter](https:\/\/huggingface.co\/docs\/datasets\/master\/loading_datasets.html#from-local-files) in the docs for more info.","embeddings":[-0.237810269,-0.1662292928,0.0330213755,0.3780060709,0.2134668827,-0.0944955647,-0.0819416642,0.0058013732,0.2306800038,0.1115748957,-0.3286620677,0.2727145553,-0.4143139124,0.7541984916,0.0344124287,-0.2124170363,0.2468240112,-0.0094740018,0.2130971253,0.0727865249,-0.1470599622,0.0766597688,0.122483775,-0.0655740127,-0.3424080908,-0.463511318,-0.144516468,0.2216348797,-0.2589560449,-0.4306832254,0.2377919406,-0.0232453737,0.3134662509,0.2755083144,-0.0001271783,0.0192323532,0.1439924836,-0.198069483,-0.1501704305,-0.2263995558,-0.0105654718,-0.2625780106,0.2062537819,-0.2558062971,-0.0249745864,-0.4207357168,0.0195486937,-0.2141057849,0.4344231188,0.0629910976,0.0510321781,0.0888584852,0.2963438928,0.1291370839,0.0012458344,0.7133287191,-0.0915173143,0.4353260696,-0.0811251178,0.1337190717,0.4062704146,0.0324031152,-0.1193139926,-0.1696986854,0.4198621213,0.0327906273,-0.1076439098,-0.2930718362,-0.0799694434,0.1958370507,0.5325228572,-0.3814711571,-0.2554703355,-0.4010131359,-0.1833397299,-0.1577882916,0.1092292666,0.101676859,-0.2182467729,0.2751451135,0.0796673223,-0.4316377342,-0.2243959159,0.462623328,-0.0305591356,0.1390717477,-0.218229875,0.1348426789,-0.009732022,-0.1697131693,-0.1591190845,-0.2166929692,0.1530637443,0.1046103239,0.0219988041,-0.0494470745,-0.3984912932,0.3054518104,0.1760906577,0.2698923647,0.0397170559,0.0487186201,0.1419565082,0.177365616,0.4741599858,-0.1392013729,0.0289493147,0.2929439843,0.4270748198,-0.1479544044,0.0825609341,-0.1716121733,-0.2451376319,0.1686770618,0.2907402217,-0.4451834857,0.5732555985,-0.0094405822,-0.1498821974,0.0264444333,-0.0496975146,0.0983386114,0.0392809436,0.4119528532,-0.2485650033,0.0763239861,-0.0822553113,0.567754209,0.0806292444,0.1816281974,0.0675828531,-0.1055876538,0.0586666875,0.1883953214,0.119540222,-0.304302156,0.1431476474,0.0493285544,-0.117508173,-0.2298802137,-0.1051741913,-0.2983853817,0.2503584325,0.1874824315,-0.034736421,0.2177912146,0.0264807232,-0.251060605,-0.0614526682,0.0711186901,-0.1817571968,-0.5472614169,-0.4872515202,-0.0172648542,-0.296346128,-0.1825623661,-0.494740665,0.2153961957,0.4036822021,-0.0304517038,-0.0772516876,-0.0381743051,-0.361159116,-0.4873213172,0.1917564422,-0.0130669074,-0.3687559068,-0.070466958,-0.2764284015,-0.06379278,0.0365636498,0.5089984536,-0.0615215823,0.3808539212,-0.0218580235,0.3172659874,0.6162870526,-0.2900633514,-0.3520003855,-0.0438315198,0.3447287977,0.1639666408,-0.1431290656,0.255654633,0.4974556565,-0.0588796884,-0.2075304687,0.0913296044,-0.0154418899,-0.0106452741,-0.1777820438,-0.0888601243,0.3596826494,0.0504076965,-0.1555024683,-0.1489335746,0.2017619014,-0.2562267482,0.1623859257,0.0518335551,0.1262307614,0.1539584398,0.4118930101,0.674677372,-0.0025748683,-0.380671829,-0.4766902924,0.1435331255,-0.0869005248,-0.2942626774,-0.4281311631,-0.3349537849,-0.2720730305,0.2078000903,-0.1555913836,0.0490857139,-0.0890747085,-0.1226063222,-0.0303374156,0.24661614,-0.4448167086,0.1691298634,-0.0641332716,0.1097605452,-0.0555861406,0.3028223515,-0.0643483549,-0.5284366012,0.1100298166,0.3356063664,-0.0236818772,0.0538898818,-0.0074759074,0.3950042427,0.1024578959,-0.1147257537,-0.2665483654,0.2249416262,0.0773112848,-0.3018085063,0.0808286816,0.0188892446,0.2364879698,0.0676601678,-0.1550954729,0.2655336857,0.2925465107,-0.034726575,-0.1800360978,0.0837549642,0.0568516552,0.17951864,-0.0535778515,-0.1041107699,-0.0837018713,-0.1460365802,0.1154142171,0.2012958825,-0.7476040721,0.0817903206,0.4663318098,-0.0182795506,0.4762844443,0.0598481223,-0.0826509669,-0.1684434116,0.2419523001,-0.1110014096,0.2505150735,0.0205011591,-0.1729886234,-0.0328332148,-0.0232802629,-0.0379643291,-0.0422997959,0.1455172747,0.3761143684,0.1378004253,0.1253286153,0.0087273847,-0.1385634691,0.2837544084,-0.2259045243,0.0164922141,-0.4621778131,0.1632553637,-0.1930512041,-0.4235275388,-0.006028852,-0.1084075347,-0.3135083914,-0.1574716419,-0.1068442985,0.1374977231,0.0040074326,0.1223513037,0.3831165135,0.3042136133,-0.1437426805,-0.2974486053,-0.239982143,-0.2823421657,-0.0940711275,-0.1037602574,0.1134150475,0.1871296316,0.4331095517,-0.1567296386,0.1698223501,-0.3087498844,-0.1628099084,-0.0224445332,-0.1605901271,0.0554261617,0.0662362576,0.3839928508,-0.1775969267,-0.1332483888,0.0893267021,-0.1924712062,-0.1230789125,0.0382372439,-0.0354033336,-0.0072129765,0.1521919072,0.0253635533,-0.7563423514,-0.1886426359,0.5028291941,0.2522119284,0.0238973498,-0.0296341702,-0.0192307364,-0.0403144583,-0.3316256702,0.0194334313,-0.0847092867,-0.4716016948,0.5566013455,-0.2912874818,-0.4345383644,0.2250980288,-0.0294178072,-0.0016179786,-0.1036294177,0.0356378816,0.2269741744,-0.0987163037,0.3692942262,0.0368262716,-0.0142815448,0.2845596075,-0.000279107,0.2400384396,-0.0237688515,0.0509027764,0.2991060019,0.1358183473,0.2495039254,-0.0381164216,0.1825693548,-0.029887015,0.3044150174,0.1194671169,-0.0898281857,0.3953668177,-0.1584452987,0.4451116621,-0.0311501641,-0.0932234302,0.0465212129,-0.1361174285,0.148824051,0.1300960332,0.1521754265,0.0301486608,-0.0227373354,-0.1833896786,0.0052987388,-0.2169131935,0.1409316063,-0.3431179523,0.297162056,-0.0625432953,-0.1378450245,0.024536157,-0.2161645144,-0.0243809111,0.3769986331,0.1443920583,0.2630928755,-0.0981368721,0.0922479481,-0.4891065955,0.2030301839,0.2482052594,-0.0379975475,-0.2532474995,-0.0251315068,0.2564848065,0.2246561199,0.1220309064,0.0435848944,-0.2990808487,-0.297909826,-0.1101707667,-0.5257943869,0.1900992543,0.0069583007,0.0007102028,0.0075407554,0.4951515198,-0.4599431753,-0.1785256118,-0.1128562167,0.3327127993,-0.3560933769,-0.1817073822,-0.2515665889,-0.2221289724,-0.5160810947,-0.2160278857,0.4248581827,-0.0517174751,0.1393969506,-0.1036180556,-0.3414806724,0.0714399889,-0.1353787035,-0.0908671394,0.4022269547,-0.3095941246,-0.1547002792,0.456852138,0.2821012735,0.2365688682,0.4839627147,0.0233904179,-0.4414555132,-0.2041065693,0.1150660291,0.3530122936,-0.1686894,-0.1933995783,-0.0165236089,0.0535567775,-0.0436288603,-0.4761084914,-0.0063218516,0.2209944576,0.11451976,-0.7278291583,-0.4517309368,0.4075703621,0.1989436001,0.0232937951,-0.0869522393,0.3862119317,-0.1088402644,0.4511083961,0.171370551,0.920620203,-0.3042573035,0.4588991404,0.3918348253,-0.3748668432,0.4236673117,-0.0161824394,-0.2834759951,-0.2926483154,0.2958174348,-0.2013873607,0.124474071,0.2746739984,0.0297483206,-0.2378523052,0.1377848387,-0.194422543,-0.1548203975,-0.0723189935,0.3833324909,-0.0824032947,-0.0499591716,-0.4727933407,-0.0082902964,-0.0444188192,0.1927435547,-0.0833240822,0.0354112014,-0.1781819165,-0.0354208797,-0.2267344892,-0.3245777488,-0.0947827473,0.2185376287,0.2355640978,-0.4162791967,0.1968208104,0.5587382913,0.253203392,-0.0242749676,-0.2452732027,-0.0615352131,-0.0694263279,-0.1300839782,0.1039223447,-0.1281896234,0.3510268033,-0.0988306105,-0.1566266567,-0.0491092131,-0.0027278292,-0.151480481,-0.247232914,-0.0694772378,-0.191577211,-0.3497976959,-0.0484246798,-0.4830571413,0.0785324797,-0.0209787413,-0.004174537,0.3805873692,0.1568733901,0.2481210381,-0.0552266948,0.0544712842,-0.068919085,0.201989308,-0.2097505927,0.2892911136,0.3349259794,-0.0249455497,-0.2534463406,-0.064378731,0.4565857947,0.1960139126,-0.2084733546,0.1738604158,0.1887452751,0.0873214677,-0.371257931,0.5464473367,0.1676831096,-0.4892342091,0.0068765045,-0.2661756575,-0.4534219205,0.2466026843,0.1706192493,0.2125742733,0.143290922,0.157999143,0.0639650077,0.0403115377,-0.1287721992,-0.018794328,-0.1039855778,0.0510511585,0.3722291291,0.0420410447,0.1012656465,-0.2688124776,-0.0840556547,-0.1287344694,0.1210713908,-0.0133125316,-0.1868758649,0.2314411849,0.0717604831,-0.1721500456,0.0660553277,-0.3535119593,-0.0502739325,0.0726570114,-0.0781745985,0.4331773221,0.0200851485,0.0593629889,0.0098101934,0.0293241553,0.0489457771,0.2316126078,-0.1596681029,0.1100056842,-0.2222000808,0.0564764664,-0.0708735138,-0.1899221539,0.2187069952,0.185351789,0.383146286,-0.1810681224,0.398471117,-0.2720332444,-0.0031760137,-0.2061009407,0.3805424571,0.2312155813,-0.0632025823,0.1723066419,0.325319171,-0.0598840937,-0.1709196419,0.0686494857,0.1401774734,-0.1394279897,-0.1276977807,0.2266844511,0.1291902065,0.4667309225,-0.7082306743,-0.0597292632,0.1112584695,0.1040728018,0.1130272374,-0.0223206524,-0.0193721466,-0.0005893737,0.2922588289,0.4302027225,0.2038756013,0.4065985084,-0.0584568977,0.1299313158,0.2378509641,0.051853653,0.546364665,-0.2018978149,-0.1400672197,-0.2965394258,0.0133124338,-0.1272659153,0.1368792802,0.5901451111,0.0090803821,-0.39947474,0.0755189583,0.5644210577,0.0026345116,-0.1513636559,-0.0362434387,-0.0394277759,0.003465987,-0.2722338438,-0.2876990736,-0.5580126047,0.1271226555,0.242232725,0.2279909253,-0.448431313,0.3282668293,0.0688506961,0.2519052327,0.0404617935,0.2493932247,-0.1919648051,-0.071862936,-0.0172176417,0.2116292417,0.4467218816,0.1513146162,-0.0992312655,0.428517282,-0.1903179139,-0.1783988178,0.0021036691,0.0057383003,0.1007185802,0.1863934696,0.4147551358,-0.0348259658,0.0907186642,0.0266912375,-0.0193428993,0.2751912475,-0.1650656164,0.5213029385,-0.2983398438,-0.202845633,-0.066774413,-0.0671221986,-0.2131318748,-0.0435939357,0.3483567834,0.063923046,0.2537431121,-0.2583738863,-0.0237737335,0.1026847884,0.5535253882,0.2162878066,0.5189980268,0.0687717795,-0.057701461,-0.49127689,-0.1256464571,-0.0572120063,-0.183850944,0.0308066234,0.0029729118,-0.0727822706,0.3946724236,0.0961569101,0.4237008393,-0.0105359638,0.0309104435,-0.0358678214,-0.1061554328,-0.2830891013,0.105676055,0.0646324158,-0.0537189133,0.3075629175,-0.0429795943,-0.0801165104,-0.1211295277,0.0881644487,-0.2088938355,-0.4123039246,0.2223679721,0.0092395814,0.2638677061,0.1034504399,0.045279026,0.2686206996,-0.0465898253,-0.2727016509,0.382232368,-0.0694519654,0.2768940628,-0.4674035311,0.2228346467,-0.1484358609,-0.0775288492,-0.0480352379,-0.0436978824,-0.1375497133,0.0348941833,0.0297295246,-0.0244258009,0.1754740477,0.1926557273,-0.2406093329,0.0332073756,-0.252490133,-0.0686557144,0.5929351449,0.022446176,-0.1151450276,0.1116814464,0.043940708,0.560339272,-0.0473893695,-0.5693662763,0.0072652544,0.3408538997,-0.0636249483,0.0488683656,0.0711791068,0.0867063031,-0.0900590569,-0.1279159039,0.6649675965,-0.083202742,0.2170121819,0.1679089516,-0.2830289602]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2426","title":"Saving Graph\/Structured Data in Datasets","comments":"Nice! Thanks for your help.","body":"Thanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.","comment_length":5,"text":"Saving Graph\/Structured Data in Datasets\nThanks for this amazing library! And my question is I have structured data that is organized with a graph. For example, a dataset with users' friendship relations and user's articles. When I try to save a python dict in the dataset, an error occurred ``did not recognize Python value type when inferring an Arrow data type''.\r\nAlthough I also know that storing a python dict in pyarrow datasets is not the best practice, but I have no idea about how to save structured data in the Datasets. \r\n\r\nThank you very much for your help.\nNice! Thanks for your help.","embeddings":[-0.2841216624,0.0266983286,-0.0496116839,0.2592530847,0.1681542397,-0.1244001165,-0.0174353383,0.0479331389,0.0541954637,0.1503934264,-0.2165412158,0.4113406539,-0.4343291521,0.6310096383,-0.1379674077,-0.2617000341,0.2769474685,0.0372624062,0.2771196067,0.0114676813,-0.1179588959,0.0482694991,0.0485160798,-0.0686145201,-0.2485456765,-0.5032104254,-0.1472541392,0.068224296,-0.3383066356,-0.4679697752,0.1917086095,-0.0970357209,0.305381,0.2369099706,-0.0001210282,-0.0562430918,0.1604664326,-0.1899514943,-0.0872493088,-0.2424679548,-0.029226521,-0.2869589925,0.2097710818,-0.3833188117,-0.0044435696,-0.4766096473,0.0479367971,-0.2548909485,0.3660090566,0.1242099777,0.1080285013,-0.0860133171,0.3315427601,0.1790641695,0.1547197551,0.588534832,-0.1323008686,0.3661423922,0.0416585803,0.1821733266,0.4353068471,0.0034135152,-0.1510414183,-0.173096329,0.2937896848,-0.0408372283,-0.0387933478,-0.4126718044,-0.1015877277,0.1593149453,0.7135501504,-0.4165401161,-0.2383114845,-0.1703110188,-0.2181325853,-0.1009319201,-0.0310281776,0.1470523179,-0.1848394573,0.2477009147,0.1991881579,-0.3879391551,-0.3278194368,0.4786539078,-0.0796560794,0.2878671288,-0.0833580941,0.1239681765,-0.1539126486,-0.1244710386,0.0591384843,-0.1360456645,0.1301975399,0.0838861391,0.0019253648,-0.0863485858,-0.4713477492,0.2273230851,0.0411024541,0.3066672981,0.1055564582,0.0159695614,0.3213527799,0.2260525972,0.4952710867,-0.1595544964,0.0259024836,0.3967303336,0.4859671593,-0.356570363,0.0394766144,-0.1707198322,-0.2782934606,0.1246410236,0.4304837584,-0.3397375941,0.5514327288,0.0056842975,-0.139626354,0.0537976958,-0.1749348491,0.0813301429,0.0027892168,0.3566479981,-0.2515427768,0.0862334743,-0.145427227,0.5512216091,0.0249548666,0.2714252472,0.1232312247,0.0114962235,0.0132228844,-0.0064821779,0.1102076992,-0.1893864423,0.0976420641,0.0717823654,-0.1293924153,-0.2248810381,0.0838540345,-0.3347125053,0.3123422861,0.1920189857,-0.0389790311,0.2077109665,-0.0955785587,-0.2691794038,-0.0658239648,0.2175432146,-0.2327611297,-0.5069137216,-0.5937488079,0.0485212654,-0.2319661677,-0.2397316992,-0.3102378845,0.13584961,0.4642160535,-0.0335586928,-0.016889492,-0.1633044481,-0.1371935308,-0.5706154108,0.0241618883,-0.2182810009,-0.4665491581,0.0868670493,-0.337066561,-0.0482161529,0.1360135823,0.456110388,-0.0810950547,0.3392061591,0.0613524169,0.3989382386,0.7109857202,-0.2810422182,-0.4344409406,-0.0786120594,0.4597835243,-0.0129767526,-0.1200763136,0.3644968271,0.5232423544,-0.0890418366,-0.3044044375,0.0743968487,-0.003841952,-0.0004082654,-0.1656683087,-0.0121907452,0.4052241743,0.0457141995,-0.111083433,-0.2634261847,0.2684825361,-0.223661691,0.1339295805,0.0210237354,0.0913813412,0.0915135294,0.5121109486,0.5301193595,0.0883933976,-0.3427695036,-0.3199818432,0.0571642183,-0.2196976095,-0.3135938048,-0.50762254,-0.3200414777,-0.2272316813,0.2335603386,-0.1739004254,0.029458249,-0.0092446683,-0.1409551948,-0.0968217552,0.2200828344,-0.3944184184,0.0383464694,-0.001918109,0.0385975502,0.0955543742,0.2951526046,-0.1748662889,-0.5797656775,0.0979821756,0.3278578818,-0.0146706915,0.0423655994,-0.0160814449,0.3460858762,0.1809360981,-0.1434504688,-0.3534111083,0.179299444,0.0467723981,-0.3682452738,0.1073432341,0.1279942989,0.2103717029,0.033899527,-0.1199027449,0.2677151859,0.3105708957,-0.115898408,-0.1658835709,0.1397317201,0.1465287358,0.1352618635,-0.0625336096,-0.0416738354,-0.1372070611,-0.0592733249,0.0731503367,0.1440672874,-0.7061895728,0.077058576,0.4367786944,0.0270884633,0.5205628872,-0.0668705329,-0.0325142145,-0.1538984179,0.2958709598,-0.0266648494,0.2053173333,0.0548971593,-0.1652005166,-0.0948483571,-0.0343751907,-0.0926699266,0.0498094037,0.1291927695,0.3859781921,0.163282901,0.1575922072,-0.0273873024,-0.1223383397,0.2176821679,-0.1742672622,0.0025888789,-0.3510479331,0.1947001815,-0.1646115333,-0.4331569076,0.0856512338,-0.2088753432,-0.1645665169,-0.2367683351,-0.0040504821,0.0308340006,-0.0090210233,0.1559923142,0.413297087,0.2328831553,-0.1264953911,-0.3499773145,-0.2382426113,-0.3510257304,-0.1498568207,-0.0200585704,0.0612706058,0.1346156448,0.426366955,-0.022975402,0.2318389118,-0.3560204208,-0.2614273429,-0.0589313842,-0.0870553404,-0.0122567518,0.0449244268,0.3619051874,-0.1130526438,-0.1848883182,-0.073438853,-0.1528510004,-0.1102313176,0.0523401871,-0.0910529569,-0.0631368309,0.1202831566,-0.123589538,-0.8660712242,-0.1593050212,0.4051977694,0.3309882283,0.0584078319,-0.0806771144,-0.0035444782,-0.1540699601,-0.1424028724,-0.0067969938,-0.0434636027,-0.4086027145,0.6192677021,-0.3015931845,-0.4213538468,0.2497227192,-0.0790835395,0.1218820438,-0.0628010184,0.0946581811,0.3439460993,-0.0709532872,0.4456189573,0.0651979297,-0.0363210738,0.2191872299,0.063645117,0.1822438687,-0.0953482091,0.1691356897,0.2736745179,0.1447312981,0.3038574159,-0.0989243388,0.1102454215,0.0146523258,0.2644544244,0.065755114,-0.1000515819,0.4447125494,-0.1048211679,0.3259735405,-0.0145026017,-0.0187916458,-0.0050007557,-0.1352424324,0.0207416564,0.1201295778,0.095038183,-0.3222895563,-0.0464365371,-0.2418021113,0.0132942069,-0.1476363391,0.134015277,-0.5244495273,0.2803181708,-0.1185977459,-0.1340767592,0.0506491028,-0.1792117804,0.0019615304,0.2161165625,0.057800889,0.2175221294,0.0146164242,0.0171747077,-0.4540729821,0.2044790089,0.2387604266,0.0119634783,-0.1724223495,-0.0026381451,0.2731368542,0.1483608335,0.0120003745,-0.0663042963,-0.3229424357,-0.1528742015,0.0603577532,-0.3818270862,0.1252575666,-0.0428564586,0.0667193756,0.0556711219,0.3173848689,-0.3635873795,-0.0289634354,-0.1335092783,0.2494827211,-0.3107518554,-0.1240490079,-0.1492804438,-0.2513454258,-0.452506274,-0.2425408363,0.3984553218,-0.1104324535,0.1683814824,-0.1492936313,-0.3681343198,0.0317806005,-0.1788109988,-0.0667683631,0.4276967645,-0.4314914346,-0.2061053216,0.3085004091,0.2865608335,0.2890154421,0.4032464027,0.0461586826,-0.3243086934,-0.1714724898,0.0421539359,0.4028622806,-0.2759926319,-0.2559232712,-0.0315315984,-0.1050217375,-0.0858861953,-0.2660092711,-0.1357110143,0.2340765893,0.026414454,-0.7436004877,-0.5021514297,0.363727212,0.1991765946,-0.0408622287,-0.0471396111,0.2178131044,-0.1144923344,0.5535894036,0.3522467315,0.9528594613,-0.1480493397,0.3756441474,0.3192255199,-0.4501427412,0.3512215316,-0.0601966679,-0.2651688457,-0.3414159417,0.3303072751,-0.1732361913,0.0516402423,0.271792978,0.1120338589,-0.3655573726,0.0402636379,-0.2885537446,-0.2566679418,-0.0618283041,0.4239861071,-0.053476043,0.0896265656,-0.3430037498,0.0637798831,0.0470033884,0.0775193647,-0.1325614303,-0.0250138678,-0.2283075899,-0.0013281667,-0.1819839776,-0.2292514294,-0.0003284101,0.3887287378,0.1208996475,-0.4101457596,0.2078385204,0.561380744,0.2419106662,-0.0271689445,-0.228816539,0.0235120151,0.0797628835,-0.1089358851,0.1868199855,-0.0930187553,0.4044603109,-0.140601024,-0.185838595,0.0134275742,-0.0311677158,-0.2307653278,-0.3247739375,-0.0281728338,-0.0568388104,-0.2426908016,0.0490746051,-0.4675457478,-0.0655806661,0.0262196492,0.0505637676,0.3686614633,0.1873699725,0.4653540254,-0.0979471579,0.0441833772,-0.0798148215,0.0921662375,0.0151226083,0.284193337,0.3312571347,-0.0705738515,-0.2774217427,-0.1393246949,0.4736006558,0.0616722032,-0.1129214242,0.201341033,0.1030427516,0.1083725467,-0.3092980385,0.5350552201,0.1630360931,-0.4120645523,-0.1590043753,-0.2777443528,-0.4121989608,0.1872730553,0.1605760157,0.3066107333,0.0179783292,0.0978630185,0.1253119558,-0.0132037792,-0.1768645048,-0.0446455777,-0.1038016304,0.1428653151,0.1964751929,0.0779784843,-0.0522719696,-0.2041474432,-0.0500906296,-0.1482426524,0.1295485944,-0.0727825612,-0.1689884067,0.2017162293,0.0919424295,-0.1404850185,0.0886624753,-0.3650427163,-0.0752870142,0.2086860836,-0.1580018401,0.3556466103,-0.0415586382,0.1248644814,0.0896185264,-0.0085411659,0.0314319432,0.1758132577,-0.202881813,0.1221606731,-0.1993849576,0.1843562424,0.0155860996,-0.2141652852,0.2574011385,0.1295868307,0.2741523385,-0.1748604327,0.3764035404,-0.2413075417,0.0517622642,-0.2666993141,0.3347614706,0.2342572063,-0.0855310634,0.1181913763,0.3075523078,0.0217745285,-0.2194829732,-0.0261109266,0.0284446161,-0.1401365697,-0.1308552176,0.1413942128,0.2123751342,0.4253672063,-0.7093172669,-0.0690081865,0.1396739334,0.05800182,0.1518891156,0.0575475357,-0.1201289371,-0.0085611092,0.2071281224,0.3403806388,0.1612823457,0.4989518225,-0.0736874714,0.1838923544,0.4943055511,-0.0997152254,0.5593063235,-0.0311794467,-0.1089319959,-0.3896370828,0.0321778171,-0.1449196488,0.3243095577,0.4041045904,0.172501862,-0.3057445586,0.1246306226,0.5459190607,-0.0147057073,-0.1045813262,-0.1598334312,-0.0505966656,0.0123628844,-0.3287473023,-0.3465027809,-0.4625299871,0.108880505,0.3329596817,0.191283673,-0.3969893456,0.5271252394,-0.0338282399,0.2193278372,-0.0186604857,0.2441181839,-0.2057516873,-0.0705343038,-0.1102791801,0.1335024983,0.4997254014,0.2674865127,-0.1726030558,0.3909478784,-0.1849000007,-0.2632442713,0.0030157515,0.0195180755,0.1247343048,0.1883065999,0.4394977391,-0.0127039356,0.0678972453,0.1405875683,0.1291375458,0.2408783734,-0.0252324771,0.4444268644,-0.215532437,-0.2974543869,0.0012676738,-0.0627623051,-0.2026788443,-0.2096818089,0.3399343193,0.0830626264,0.2323112935,-0.2821787596,0.0073595205,0.0738260001,0.5606043339,0.1552883238,0.6646101475,0.0370366648,-0.0969469324,-0.4408203363,-0.1049630642,-0.0683525503,-0.1591607332,0.0647787973,0.0789176971,-0.027098326,0.4226264954,0.1832793057,0.437764436,-0.0153542273,0.1525796354,-0.1372033209,-0.0303705484,-0.1080871299,0.1675405204,0.0018987632,-0.0156170148,0.3025383353,-0.1045776606,0.0269559491,-0.2052236199,-0.0551944599,-0.0426292345,-0.371296078,0.2955659032,-0.0280489177,0.2211886346,0.0667313263,0.1516574472,0.3440728784,-0.0286210813,-0.2324383557,0.3236463666,0.0587510839,0.2675921023,-0.3327796459,0.2883047163,-0.0300495271,-0.1523626596,-0.1668872833,0.0046555414,-0.1544424891,0.0112546626,0.1345131099,-0.0119809294,-0.0656271055,0.092756696,-0.2301007658,0.0377657861,-0.3420924544,-0.1543703228,0.512570262,0.0094515383,-0.1749468893,0.0674499348,0.0001615482,0.420153439,0.1291292906,-0.403175801,0.0351672843,0.4053023756,-0.1398120672,0.0222347118,0.0863811597,0.2055108398,-0.0230304282,-0.086260885,0.6118322611,-0.0927111804,0.208769083,0.0379401296,-0.3804062307]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2424","title":"load_from_disk and save_to_disk are not compatible with each other","comments":"Hi,\r\n\r\n`load_dataset` returns an instance of `DatasetDict` if `split` is not specified, so instead of `Dataset.load_from_disk`, use `DatasetDict.load_from_disk` to load the dataset from disk.","body":"## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n","comment_length":24,"text":"load_from_disk and save_to_disk are not compatible with each other\n## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n\nHi,\r\n\r\n`load_dataset` returns an instance of `DatasetDict` if `split` is not specified, so instead of `Dataset.load_from_disk`, use `DatasetDict.load_from_disk` to load the dataset from disk.","embeddings":[-0.1792232394,-0.225251615,-0.1031944528,0.443765372,0.1700460911,-0.0056649549,0.288071841,0.3563308418,0.3407506943,0.0764616281,0.0730309263,0.3753902614,0.0848360732,0.1554043442,-0.1350735724,-0.0391581059,0.2747603059,0.048006542,0.2421990782,-0.115025036,-0.2098629475,0.400549978,-0.203281939,-0.0734000653,-0.1446835399,-0.0062051648,-0.2334032953,0.3358605206,0.3330038488,0.0827944577,0.2947779894,-0.2153790295,0.0816993341,0.5532495975,-0.0001212123,0.0112388376,0.160898447,-0.229187578,-0.2692857683,-0.6934784651,-0.1458991766,-0.254847616,0.0668642372,-0.3095525503,0.1461135596,-0.0618361346,-0.0437009335,-0.2818159163,0.519669354,0.1374070048,0.1133457944,0.1389506608,0.1779476702,-0.2703320086,-0.0786266327,0.4976059198,0.0025216881,0.3422203362,0.1431116164,-0.139275983,0.1311556548,-0.0501510464,-0.3340208232,-0.1618638635,0.2924522758,0.2686187327,0.2871638238,-0.3989811838,-0.0131455446,0.1011754498,0.4063547254,-0.1671675295,-0.331186831,-0.3169541061,-0.3207164407,0.0708923116,0.4782576263,0.0573517419,-0.3660798967,0.1695609838,-0.1249224544,-0.3215165734,0.0082832025,0.3736847341,-0.04334566,-0.0029226099,-0.078749001,0.0695815757,0.2542456388,0.0742168278,0.3182521462,-0.4341986179,0.2093426138,0.0751498416,-0.1967533976,-0.0305962283,-0.1692512184,-0.4386776388,-0.1910796613,0.3895240724,0.2439085394,-0.1618518233,-0.4091942608,0.3818172216,0.4627968669,0.2550275624,0.1295374036,0.2542569339,0.2904804349,0.0670523569,-0.0025622717,0.0827606767,0.2443393469,-0.3209077418,0.3263378143,-0.168192938,0.3287460506,-0.1333357096,-0.3826584816,-0.0483474135,-0.2059504539,-0.096618399,-0.0288152806,0.1742832214,0.1753856093,0.3235180378,0.2330272198,0.3827495873,-0.1726453304,0.4203166366,-0.1407727301,-0.0053146444,0.1534060091,0.1977638453,0.137517795,-0.1364088207,0.1692527831,0.1127422452,-0.0249592569,-0.1497893631,-0.0999917462,-0.3721024394,0.1490969062,0.1833569258,0.3355745971,0.1456549317,0.1002760902,-0.361836344,-0.237632066,0.2797680795,-0.4703740776,-0.1679827422,0.084968619,0.1425002515,-0.2271607667,0.0532756262,-0.6181437969,-0.0054427949,0.2329247296,-0.3323367536,0.0348131061,-0.1326627731,-0.165564388,-0.4953819811,0.3575298786,0.2533137798,-0.3116635978,-0.0661654621,-0.1282549948,-0.1001225859,0.0401784927,0.3860417008,-0.0586670674,0.3588559926,-0.392334938,0.1302446872,0.551763773,-0.5543698668,-0.329724282,0.4475794137,-0.1710844934,-0.0719404668,0.0096968254,0.1479272544,0.2011791319,-0.1595356911,0.2984156311,0.4097066224,0.0290239155,-0.0377856418,-0.3559076488,-0.1730239838,0.1189952567,0.1487328559,-0.0316416919,0.1294300258,0.1749072373,0.0735625997,0.2986240983,0.0119523965,0.1110604629,0.37243855,0.3932012618,-0.0431029461,-0.1116136461,-0.139479652,-0.7380242348,0.215712145,0.2114938349,-0.3154179752,-0.0078514479,0.1417946965,-0.0605179667,-0.1606418788,-0.1623939276,0.1112134829,-0.0190429799,-0.0295296758,0.1480240673,0.0633510649,-0.2396693975,0.2358975261,0.2464085519,0.0769229084,-0.2499374896,0.6901504993,-0.0967807099,-0.3731459975,-0.091365397,-0.0424572788,0.3784249127,-0.0463225879,-0.1258752048,0.4283832014,0.0684175417,0.3490032256,-0.3699417412,-0.2295208722,0.1426036358,0.1124600768,0.0141563369,0.336281687,0.1857881546,-0.2280341089,-0.3808369637,0.1877964288,-0.2250871807,0.3004726171,0.1914522946,0.0634959638,0.1003570184,-0.1119061038,0.1852492094,-0.235955596,0.1871626377,-0.1902701706,0.2068027407,-0.2088867724,-0.1850789487,0.0922372937,0.3652824461,-0.053681381,0.2342032194,-0.0175956674,0.0251882598,-0.0746917427,-0.0590968728,0.4634986818,0.7871444225,0.0422297232,0.0795527399,-0.0054729604,0.0074960729,-0.2254068106,0.1815459877,0.0908007845,0.3721469045,0.413138926,0.0500583239,0.0782987699,0.0164352059,-0.1771486253,-0.1368957013,0.0040486008,-0.3926257789,0.0614943951,-0.1727619022,0.1391112208,-0.41207394,-0.0962932557,-0.1234484836,-0.4050314724,-0.2254335433,0.634055078,-0.0669438317,-0.1725351065,0.1766725183,0.0252850801,-0.1996213794,-0.2302349508,-0.2482907921,0.1143594161,-0.2080485374,-0.0911775306,0.2168107033,0.0659446642,0.2270480096,-0.184716627,-0.2007342279,-0.2244721353,0.0086451713,0.0563396327,-0.0481998771,0.174309954,0.0051670028,0.1690931022,0.1144624725,-0.3755486012,0.1989772469,-0.1584615707,-0.2648602128,0.1604645997,0.2826429904,-0.1355348974,-0.2202664614,-0.3173787594,-0.4011374414,-0.4135324955,0.1916930228,-0.0158063285,0.112751171,-0.3547134399,0.1864437312,0.0553877577,0.2779679298,-0.0414725766,-0.1246274933,-0.129309088,0.4934219718,-0.1611725539,-0.2432639301,0.0910674185,0.1499760598,0.0022744983,0.0416839868,-0.3326735795,-0.0748596862,-0.0855374411,0.1137662902,-0.2176433802,0.0636477023,0.2987585366,-0.0172817074,0.0534693897,-0.0840159729,-0.3136729002,0.1490674019,0.21666421,0.1762974709,-0.0757616907,0.3940104842,0.0701805875,0.3671786189,0.2821439207,0.0682801232,0.2314012498,0.1976292133,0.2039630115,-0.2118751109,-0.3651670516,-0.217256397,-0.1256068647,-0.3346657157,-0.1396237314,-0.0191876106,-0.0728396773,-0.2661356628,-0.2252121866,-0.4899787009,-0.3013902307,-0.109698981,0.0758392662,0.0644719526,-0.145582661,0.2430927008,-0.0647757351,0.0884331316,0.0669178218,0.4528947175,0.2329869419,0.0132697737,-0.3872072697,-0.2801979482,0.035410434,0.3589225411,-0.0231552571,0.0943201035,-0.2866734862,-0.1208518296,-0.1231397018,-0.1881711781,0.8430709839,-0.0435816683,0.1572844982,0.2460634559,-0.1343022138,-0.328399986,0.0381864831,0.188536182,0.0769943148,0.1844556034,0.8155303001,-0.0272243805,-0.1024333388,-0.1037097052,0.2324947566,-0.2788468897,-0.4167487621,-0.0561375991,-0.1679396182,-0.4362307191,-0.2866224349,-0.0066366466,0.0994212553,-0.1910349429,-0.203799516,0.0356108248,-0.197009325,0.1286070198,-0.1980958581,0.4845178425,-0.4603369534,0.1581513584,0.1250541955,-0.1178248599,0.5534567833,0.4204623699,-0.071782954,-0.2405165881,0.1173941344,-0.2397986054,0.1263392568,0.2820703983,-0.0190952942,-0.2404928505,0.0300330054,0.0292028133,-0.2740194201,-0.3262087107,0.1233977303,-0.2438352406,-0.2024731189,-0.4568466842,0.3333617151,0.20426853,-0.0587805286,-0.0277162176,-0.0599198602,-0.2325790823,0.5505395532,-0.2199411243,0.9964282513,0.1399216652,0.1335690767,0.2671432197,-0.2895307541,0.2396452725,-0.1239444166,0.0249922387,-0.3002525568,0.07632,-0.1777347922,-0.3403819203,0.0162170529,0.3958284259,-0.2803139687,0.35866189,-0.2020865977,-0.0247780848,-0.233007133,0.0909928307,-0.1795624942,-0.1530536711,-0.4331517518,0.0824333876,0.1317228526,0.3334020674,0.0991494209,-0.2072178125,-0.4210858941,-0.0186660383,-0.1602161825,0.2978706062,0.0030853509,0.4448386729,-0.1195057631,-0.4097126126,-0.0725059509,0.5926669836,0.3466898203,-0.1597685665,-0.172764957,-0.1731955856,0.036187537,0.3085708618,-0.1741181165,-0.314591378,0.5150313377,-0.044453714,-0.0975140259,-0.0807083398,-0.0946884528,-0.2956655324,-0.2916203737,0.1140673608,-0.1707693338,-0.1510406882,-0.3946333528,0.0568458289,0.1067782789,-0.3121574521,0.0623828582,0.1775386781,0.1490916759,0.0456080064,-0.3520313799,-0.0600606091,0.025652511,0.620067358,-0.0398476645,0.0294857807,0.2415608466,-0.0791287646,-0.0901850685,-0.2299123704,0.2913432419,0.0566617176,-0.3113962114,0.4832256138,-0.4003125429,0.0823214874,-0.1740467101,0.161111325,0.2215655446,-0.116993852,0.0083233137,-0.4607629776,-0.2677718103,0.2494452447,0.3027375937,0.095439516,0.1861708164,-0.128805846,-0.0506031252,-0.0834197104,-0.1854325533,0.1311274469,-0.0117555978,-0.0833499283,-0.0141417729,-0.0124184452,0.0749378651,-0.309016794,0.0161395166,-0.1502273381,-0.0030724574,-0.0983645022,-0.0645914748,0.1623411328,0.1286141425,0.0638500974,-0.2278719246,-0.3625853062,0.2352804542,-0.2045985609,0.2090546638,0.2792466283,0.0417335443,0.1091272011,0.0875456631,0.3675711453,-0.2056933343,0.1747206599,-0.1276892722,0.2338372171,-0.0179812908,0.0752438828,-0.0176666565,-0.0139823649,-0.6035295725,-0.0825048313,0.1156985313,0.3541322947,0.2102701068,-0.2770418525,-0.0456900373,-0.0975321233,0.398101449,0.3070780337,-0.2893286347,0.2781081498,-0.0179389361,0.1350535303,0.1429455429,-0.1580067128,0.0837234929,-0.090536207,-0.1928894222,0.0621973872,-0.1712001413,0.0006845302,-0.475207895,0.1193490177,0.5830206275,-0.2806250453,0.1496831328,0.4546429515,0.0754818246,0.5349249244,-0.1644011885,-0.1342391223,0.194725588,0.3073861003,0.0979937091,0.2507735193,0.2416252345,0.1135457829,0.2560833395,-0.2981274426,-0.0953677967,-0.1129898429,-0.1657360047,0.0248742253,0.1350264102,0.0902404934,-0.5479127169,-0.0568016805,-0.1020678878,-0.0223808754,-0.112239562,0.0154866548,0.0265096016,-0.2203860134,0.1093628556,-0.0453573279,-0.024555143,-0.1533993632,0.1082820222,0.2340093404,-0.0837568343,0.2301015556,0.0436769389,0.1465765685,-0.0433394797,-0.1569602937,-0.0072988798,0.1156498119,-0.2302282155,-0.1400012672,0.1382829696,0.5224794149,0.4858752191,-0.121873945,-0.0937869027,0.1671836525,-0.2302484959,-0.0242689401,0.0959388167,-0.2098716497,0.2996823788,0.2823518515,0.2254709154,-0.0106327878,0.2151260823,0.2488845736,0.4378120303,-0.2219176739,0.7030802965,-0.2753863037,-0.2516451776,0.0655336678,-0.193869859,-0.2221294343,-0.2310145944,0.0551737547,0.0968652442,0.2961736321,0.1833065897,0.0521609783,0.2360132784,0.0285923183,0.1093116179,-0.2365082353,-0.2159183174,0.1889209598,-0.4864019156,0.0884934291,0.156170696,0.0870368704,-0.2600170374,0.1971454024,0.0686659589,0.0330048949,0.1119749248,0.0239815693,0.1465542167,0.3653161824,-0.1289297938,-0.2046670616,0.1195952818,-0.0333623588,0.1753516048,-0.4723130465,0.3102141321,0.0657490045,-0.0150607508,-0.022406958,0.0842882916,0.0771067366,0.331301719,0.3999200463,0.1438800246,0.2604223192,-0.1196507439,-0.2254442424,-0.0617230088,0.0908673108,-0.0903902873,0.3285110891,-0.2051671892,0.6097778082,-0.4582568705,0.1216178238,-0.3751807511,0.0197446141,-0.2229548544,-0.130483374,-0.2956731319,0.1677443385,0.0091389017,0.1232137904,0.543220818,0.1687327921,-0.1220479459,0.2851135731,-0.3222294748,-0.079083994,0.3124597371,-0.0419963636,-0.1428691149,-0.0274195131,0.2287321538,0.2062350661,0.26954633,-0.4439637065,-0.3900857568,0.3673894703,-0.1451245993,-0.1367970705,0.1274524331,-0.1071220562,-0.0976591632,0.0441746153,0.282545507,0.2658179104,-0.4070364535,0.2763123214,-0.2208061069]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2424","title":"load_from_disk and save_to_disk are not compatible with each other","comments":"Thanks it worked!","body":"## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n","comment_length":3,"text":"load_from_disk and save_to_disk are not compatible with each other\n## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n\nThanks it worked!","embeddings":[-0.1595656872,-0.2487840652,-0.1072165668,0.4770942628,0.1850848198,-0.0196556132,0.2860177457,0.3149394691,0.3124919534,0.0835074484,0.0650347918,0.3543157279,0.126731813,0.1374331862,-0.1022723168,-0.0356756151,0.2724686265,0.0335400067,0.1832104325,-0.1090334579,-0.1747179627,0.4066781998,-0.1909293532,-0.0496043377,-0.1460332721,-0.0029834183,-0.2315905392,0.289855957,0.3541850746,0.111123696,0.2647699714,-0.2563732266,0.0877904147,0.5903381705,-0.0001227107,0.001515971,0.1593749821,-0.2241782397,-0.2198133767,-0.704529047,-0.1381594837,-0.2435390651,0.0674528331,-0.3141705096,0.1478727758,-0.0318320803,-0.0246844236,-0.2582590878,0.470615685,0.1172057241,0.1112174392,0.1270015091,0.1604432464,-0.251085341,-0.0763741881,0.4869512916,0.0073117996,0.3368804157,0.1364031285,-0.1862502247,0.151224345,-0.052957803,-0.3191910386,-0.1687724888,0.3105096817,0.3161045909,0.3003470898,-0.4083753526,0.0024574813,0.0670385808,0.4233000875,-0.1732080281,-0.3333834112,-0.2843082845,-0.308190465,0.102085501,0.5042901039,0.0836972073,-0.3945860565,0.1581367701,-0.1534678936,-0.3392028809,-0.0059720399,0.4016949534,0.0048802374,-0.0097245527,-0.0839798674,0.0747777894,0.2512534857,0.0990440547,0.2953085005,-0.4594422877,0.2207243145,0.0674191639,-0.1900134683,-0.0332234502,-0.1949251443,-0.4186321199,-0.2288553566,0.4153904021,0.2649355829,-0.1904456466,-0.3884089589,0.388825357,0.4547083378,0.2629271448,0.0989019796,0.2313123792,0.3263717592,0.061869394,-0.0340188667,0.0596172921,0.2456667125,-0.3161002696,0.2987661064,-0.2097992152,0.3172332048,-0.1505419463,-0.3655763268,-0.0653347075,-0.1994051486,-0.0877898112,-0.026009528,0.1749265939,0.2043834925,0.3032310307,0.2323543727,0.3965183496,-0.1574741602,0.4741445184,-0.1204879731,0.0328433216,0.1288115531,0.1933674812,0.1150395572,-0.1457891911,0.1796792895,0.130478397,0.0058707097,-0.1580342948,-0.0961435884,-0.3848310411,0.1339249164,0.1899985969,0.3493919075,0.1767187417,0.078060247,-0.3450113535,-0.2323467135,0.2821398973,-0.4777747989,-0.1342684776,0.1106451899,0.128205955,-0.194697991,0.0268598348,-0.6501268744,-0.0360539034,0.2340123206,-0.3362546265,0.0756384656,-0.0882385075,-0.1477496475,-0.496388346,0.3191135228,0.2266878039,-0.3046130538,-0.0372097604,-0.087749064,-0.0748175308,0.0445746556,0.3926349282,-0.0692032799,0.3462982178,-0.4320230782,0.1431025565,0.5570830703,-0.5935671926,-0.3035718799,0.4841957688,-0.1630117744,-0.1155269444,0.0504550785,0.1818388999,0.1703788638,-0.1578356028,0.2428192347,0.3929957747,0.0413987339,-0.0326055028,-0.3310392201,-0.1486743987,0.1518514305,0.1307430565,-0.0332674421,0.1798928231,0.1631192118,0.0419692248,0.2938974798,0.0187329166,0.1040213481,0.3925706148,0.4037343562,-0.0857312977,-0.1471502185,-0.1381686777,-0.7201148868,0.188556388,0.2083150893,-0.3183430433,0.0332239345,0.1627714783,-0.019143628,-0.1793095618,-0.145960629,0.108061783,-0.0400352366,-0.0286839623,0.1699956059,0.0338665396,-0.2660995126,0.2471517175,0.2612700462,0.0822639912,-0.1972579062,0.739774406,-0.091376327,-0.4061838984,-0.0970645845,-0.0943134725,0.3797573745,-0.0472855084,-0.1226754263,0.4271793067,0.0547959283,0.3652001023,-0.3209685087,-0.1855662465,0.1543432027,0.1313878447,0.019213086,0.3637965322,0.2039880008,-0.2391972393,-0.3542276919,0.1560796648,-0.1982220411,0.3136927485,0.2173661143,0.0768385604,0.074058421,-0.1199532598,0.1947493255,-0.2082377076,0.227824077,-0.1874376088,0.217818439,-0.2075045556,-0.2071570903,0.0768731683,0.3283700049,-0.0319051147,0.2245039642,-0.0191460773,0.028791273,-0.0410433449,-0.0407005064,0.4313800335,0.8115184903,0.0366873704,0.0886332914,0.0016591682,0.0220497381,-0.2417170256,0.184048906,0.1089527681,0.3671648502,0.3982813954,0.0398074016,0.0747788176,0.0279956274,-0.1627234817,-0.1428728253,-0.0060803415,-0.384285152,0.0723917112,-0.1575641334,0.1052181423,-0.4154405892,-0.0876143202,-0.1457629651,-0.3786204457,-0.2327490896,0.6626579762,-0.0481921583,-0.170508936,0.1716439426,0.0215009581,-0.1961759031,-0.2065805048,-0.304432869,0.1448419541,-0.1905741096,-0.1063733846,0.2446895689,0.045127973,0.2173737586,-0.127027005,-0.1912667751,-0.2704422474,0.0261345208,0.0447319523,-0.0371415429,0.1957577914,-0.0324773304,0.16990453,0.1151474193,-0.3525065184,0.2180260718,-0.1712355167,-0.2643699646,0.1535909921,0.2804917991,-0.1527781188,-0.2180304825,-0.2865968943,-0.390506804,-0.3859126866,0.1687770486,-0.0043767993,0.108151488,-0.3876335025,0.1903870702,0.016262617,0.2796458304,-0.0612094365,-0.1233197376,-0.1599696577,0.5070112944,-0.169705227,-0.2612731159,0.0830278546,0.1991830319,0.021381462,0.0366944894,-0.328542918,-0.1395133287,-0.0976009443,0.0838351548,-0.2179671228,0.1172204316,0.301386416,-0.0599898584,0.0480828434,-0.0874895453,-0.2976807356,0.1706503481,0.2329417765,0.1880690604,-0.0518954247,0.402951628,0.0624062791,0.3283655047,0.2941242158,0.0674550384,0.2754198909,0.1638332009,0.2132244557,-0.2129698396,-0.3431883454,-0.2254795283,-0.1101827994,-0.3350201547,-0.1647328436,-0.0333539769,-0.0707755759,-0.2955695093,-0.2317157388,-0.5043907762,-0.2524388134,-0.1263995916,0.0700904503,0.0375894979,-0.14671354,0.2306548953,-0.0550427437,0.0831736252,0.089134492,0.4317876101,0.246277675,0.0048889965,-0.4074720144,-0.2888074219,0.0507117137,0.346193254,-0.0426117443,0.0982037187,-0.2952648103,-0.1146567464,-0.0907456204,-0.1912423372,0.8715537786,-0.0662276074,0.1341736913,0.2167951465,-0.1737675518,-0.3339246511,0.0237346496,0.1881411374,0.0601056255,0.1695568562,0.8116832376,0.0174716748,-0.1093191504,-0.102391541,0.1877800375,-0.2784635127,-0.4538881183,-0.0774828941,-0.1557758451,-0.4280588925,-0.2892130911,-0.0058272546,0.0757523179,-0.1792573631,-0.206071645,0.0390600711,-0.2124982774,0.1514513344,-0.2228820324,0.4663800895,-0.4871139824,0.1305273324,0.1166117936,-0.0889570639,0.5388104916,0.4070642292,-0.0466861464,-0.217029646,0.143678084,-0.2550883591,0.1716179103,0.2963429093,-0.0169478878,-0.2279179841,-0.0190175064,0.0253425427,-0.2907688916,-0.3602236211,0.1421951652,-0.2178894877,-0.1619148403,-0.3986265659,0.3536889255,0.1958695203,-0.0565602221,-0.0301484372,-0.0639551729,-0.2460217327,0.4808594286,-0.2389088571,1.0419052839,0.1527377367,0.1683439314,0.2496782839,-0.3072264493,0.2062925994,-0.1811119765,0.034091536,-0.3200466037,0.0759802908,-0.209895879,-0.370919317,0.0236417931,0.4209342599,-0.2596467137,0.3910067081,-0.2230376005,0.0425109975,-0.2487469912,0.0717652962,-0.1818908453,-0.1241207793,-0.4598342776,0.0677552745,0.1574733853,0.3618867695,0.0599955693,-0.1943176985,-0.4519231021,-0.0057173707,-0.1580233723,0.3133456707,-0.0037278049,0.4612222314,-0.1122145727,-0.4451151192,-0.0982863754,0.64280653,0.3418216407,-0.1841230839,-0.1538851559,-0.1752275378,-0.0121024437,0.3082693815,-0.1878391504,-0.3142383397,0.4799483716,-0.0777719468,-0.112646699,-0.0608613342,-0.0840075091,-0.3230049908,-0.3041831553,0.1080386788,-0.2127009928,-0.1277087182,-0.4113620818,0.0633343533,0.1083676219,-0.2985093892,0.0478022397,0.1776783168,0.1266821772,0.0535404831,-0.3249801993,-0.0428134687,0.0340874307,0.6477483511,-0.05183677,0.0112725375,0.2238583565,-0.1013279781,-0.1060369387,-0.2200156152,0.2713661492,0.0748135746,-0.279330194,0.4690675139,-0.4121708572,0.1013782695,-0.1914082319,0.1425457448,0.2134708613,-0.0815304741,-0.0053488761,-0.4717307091,-0.2600123882,0.246042043,0.3038271368,0.0858951733,0.1440369487,-0.1822425872,-0.0036482471,-0.0876636729,-0.1683358848,0.1466923058,-0.0045454572,-0.0743013248,0.0480004065,-0.0100321611,0.0786938816,-0.2978360653,0.0055743414,-0.154003188,0.0257809442,-0.0830930769,-0.0385039784,0.168500036,0.1376339495,0.0678984076,-0.2152459919,-0.3629514277,0.2402343601,-0.1830365658,0.2020526379,0.2544336617,0.0461283289,0.1273222119,0.0893549919,0.3335489035,-0.2084372789,0.1926219761,-0.1359673291,0.2643968761,0.0103344135,0.0408923663,0.0178206638,-0.0069088824,-0.618465662,-0.0851199403,0.1212383136,0.3569440842,0.1959403604,-0.2924271226,-0.0065078093,-0.1001695916,0.3876193166,0.326431334,-0.2759440541,0.2804960608,-0.0192505699,0.1258967668,0.1424836516,-0.1592239588,0.0962671489,-0.0851382613,-0.232189104,0.0563241318,-0.1598895639,0.0085676368,-0.4640292823,0.0860014856,0.5629779696,-0.2901818156,0.0992163718,0.3750498295,0.0714287013,0.5369858146,-0.1613673419,-0.1101678535,0.1651523858,0.3240866065,0.0842393637,0.2433509231,0.2388918549,0.1020734161,0.2560626268,-0.2802579105,-0.1272277534,-0.1226292849,-0.1499768645,-0.019856045,0.1216476709,0.0512998663,-0.4810649157,-0.0488514155,-0.1156407669,0.0022943716,-0.1096465588,0.0258716624,0.0237378348,-0.2313769609,0.1645299047,-0.0079364628,-0.0257796925,-0.1528257579,0.1364082396,0.261289835,-0.0790912434,0.2542348206,0.0895272195,0.1597646326,-0.0464878753,-0.145381555,-0.002382874,0.0896249712,-0.2094263881,-0.1683007479,0.1582950503,0.5300831795,0.4781470299,-0.1119417548,-0.0959315524,0.1614122987,-0.2065228224,-0.005049652,0.1067054048,-0.1786062419,0.3350985944,0.2557011247,0.2118682861,0.0177860763,0.238518104,0.2124417573,0.4132989645,-0.239986524,0.6598535776,-0.2796023786,-0.2736819685,0.0341404304,-0.211424306,-0.207692951,-0.2436102629,0.0650256798,0.0942773446,0.2750737369,0.1874266267,0.0422007218,0.2386183888,0.0101690013,0.1421139389,-0.2085793018,-0.2479724884,0.161601156,-0.4652258754,0.1095929816,0.1689287871,0.0947668999,-0.31607306,0.2029401958,0.0616937727,0.0192684364,0.1731039584,0.0218644198,0.1463047415,0.3816424012,-0.1519901454,-0.1788972914,0.0894198045,-0.0157359224,0.1650359184,-0.46314165,0.2906401455,0.0592658333,-0.0247980729,-0.0359175801,0.1331382841,0.0159402322,0.3331526518,0.4398376346,0.1215740219,0.311619401,-0.0980700329,-0.2183061987,-0.0474477932,0.0712434798,-0.0828848109,0.3152991533,-0.2111347318,0.5689573288,-0.4723352194,0.1223061383,-0.4043778777,0.019831378,-0.2494304031,-0.1205270886,-0.2540346682,0.1367166191,0.0165239498,0.1174787879,0.5744800568,0.1445678771,-0.1202992946,0.2867481709,-0.3384572864,-0.0416926183,0.2935964465,-0.0216832329,-0.1451454759,-0.0248441882,0.2221045792,0.1856491268,0.2783804834,-0.4522714913,-0.4267606437,0.3633867204,-0.1452108324,-0.1196956635,0.1337875426,-0.1052086651,-0.0871677995,0.052165553,0.2752436996,0.2671508491,-0.3887984157,0.2417088598,-0.2403727323]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2424","title":"load_from_disk and save_to_disk are not compatible with each other","comments":"Though I see a stream of issues open by people lost between datasets and datasets dicts so maybe there is here something that could be better in terms of UX. Could be better error handling or something else smarter to even avoid said errors but maybe we should think about this. Reopening to use this issue as a discussion place but feel free to open a new open if you prefer @lhoestq @albertvillanova ","body":"## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n","comment_length":73,"text":"load_from_disk and save_to_disk are not compatible with each other\n## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n\nThough I see a stream of issues open by people lost between datasets and datasets dicts so maybe there is here something that could be better in terms of UX. Could be better error handling or something else smarter to even avoid said errors but maybe we should think about this. Reopening to use this issue as a discussion place but feel free to open a new open if you prefer @lhoestq @albertvillanova ","embeddings":[-0.1793285012,-0.1415978968,-0.1055235788,0.5006904006,0.232578963,-0.0382465087,0.2578227222,0.2903450429,0.3386763036,0.0991070271,0.1047393084,0.3404804468,0.0760164931,0.1851234734,-0.2164248675,-0.0467502438,0.3020486534,-0.0183789227,0.1567269117,-0.0449476875,-0.1527918279,0.2874240279,-0.1635298729,-0.05758439,-0.1719533801,-0.0479435101,-0.196713388,0.3407809138,0.3019942343,0.0461686142,0.2732267976,-0.2321476042,0.1589491069,0.5605933666,-0.0001191241,-0.0277949944,0.1933393627,-0.1634420305,-0.3407094181,-0.6506851912,-0.1944092214,-0.2385559678,0.0649952441,-0.2768592536,0.1821877509,-0.0138578583,-0.0538221039,-0.310264051,0.5078191161,0.0963830948,0.1465382278,0.1002467722,0.192025125,-0.2128233761,-0.0376710705,0.5029975176,-0.0007013583,0.3070380688,0.2560757995,-0.2046595514,0.1296961159,-0.0047982051,-0.3322282434,-0.189792186,0.2666126192,0.2611280084,0.3266581595,-0.4210820198,-0.0327870436,0.1346203238,0.4311968982,-0.1612837315,-0.3976997435,-0.3701413572,-0.2544538677,0.145589307,0.4366273284,0.0056382585,-0.3736253381,0.2122143656,-0.1339206547,-0.385587126,-0.0302912518,0.3730505705,0.0058063013,0.0287204664,-0.0525110476,0.0366714634,0.2215721309,0.0582590476,0.3583062887,-0.4107931554,0.1827104837,0.0393775925,-0.249452889,-0.0691118389,-0.1533340663,-0.4124203026,-0.1539743692,0.3790569305,0.2561164498,-0.1636759341,-0.3973768055,0.3728151321,0.5320214629,0.2313528955,0.1168094054,0.2314741015,0.3084314764,0.1047696173,-0.0210200436,0.0529793687,0.2307306677,-0.3375527859,0.3026453853,-0.104715243,0.3373525143,-0.196644783,-0.40621382,-0.0097613307,-0.1994801462,-0.0894898251,-0.0438000187,0.1508357078,0.1162971035,0.3933481276,0.2099098563,0.3266170621,-0.122673586,0.3789060116,-0.107279852,-0.018295221,0.1081808656,0.2428033799,0.0142367547,-0.0631554574,0.1086314246,0.1245500222,-0.0098791681,-0.1332287937,-0.0954652503,-0.3437676728,0.1277273595,0.1409400105,0.2713589966,0.1469076723,0.0382456034,-0.4243357182,-0.1878847331,0.2419470251,-0.4898731112,-0.1394298077,-0.0174274407,0.1673590839,-0.1537126601,0.0566571765,-0.6717962027,-0.0232001469,0.1584375352,-0.3381961584,0.0564060397,-0.0369184949,-0.0973192081,-0.5008745193,0.2721493542,0.277169168,-0.3709772527,-0.0047246674,-0.1214739978,-0.0433652066,0.0253739264,0.427850455,-0.0759251118,0.3326679468,-0.4297975898,0.0854031965,0.5042351484,-0.6016662121,-0.3212787211,0.5045099854,-0.1311884373,-0.0434036553,0.0758243874,0.1683423966,0.2189977765,-0.2220080495,0.2906537652,0.4203119278,-0.0469291471,-0.0481425636,-0.4034248888,-0.2235569358,0.1317161918,0.1483624876,-0.0357850529,0.1120731458,0.1524415165,-0.0020958986,0.3085451424,-0.0010440451,0.0577642731,0.3298602402,0.4671284258,-0.0828381777,-0.1864544898,-0.1380204707,-0.7963288426,0.208411023,0.2326538265,-0.2737752795,0.0200373456,0.09049505,0.006242624,-0.1295400262,-0.1372879744,0.081004329,0.0064008664,-0.087876454,0.12609604,0.0337573476,-0.220084697,0.2324874699,0.155164808,0.0982590988,-0.2315646857,0.6826292872,-0.1195002496,-0.3252774477,-0.0909527689,-0.0521271825,0.2896011472,-0.0309513304,-0.1150024235,0.380851686,0.0248580091,0.4010757506,-0.2816483974,-0.1012844443,0.1768703461,0.1265096068,-0.013036415,0.3360197544,0.1940553635,-0.1694559157,-0.3583262861,0.1533226967,-0.2350907028,0.3569944501,0.1710189134,0.0887068361,0.0445472039,-0.152260825,0.1597272009,-0.2137804031,0.188740477,-0.1867073923,0.2605465055,-0.1405669898,-0.1402639151,0.1336381584,0.3338999152,-0.0303892437,0.2459747493,0.0107367234,0.0479866527,-0.0385259576,-0.0095809922,0.4764660001,0.7985130548,0.0833330154,0.0470817275,0.0200402141,0.0087478077,-0.2281494737,0.1811587662,0.1527590603,0.4110870957,0.4088664651,0.0798020884,0.1150384247,0.1026185676,-0.245300889,-0.24656564,0.0196930971,-0.4122724235,0.062038321,-0.1948617846,0.0563521162,-0.3862085044,-0.0460762531,-0.1631375253,-0.4521231055,-0.1730680466,0.5514714122,-0.1993460506,-0.1709204167,0.1264359653,0.066756241,-0.2848277688,-0.1417156607,-0.2706845999,0.0450143516,-0.2285294682,-0.0703290775,0.3768827617,0.0135416761,0.2489432693,-0.1580121815,-0.1588632911,-0.2698636055,-0.0948809162,0.0372722447,0.0466151386,0.2869644463,-0.0291697308,0.1743344218,0.1645389646,-0.391613096,0.1812032014,-0.1359606981,-0.2781122923,0.0960861817,0.2482705414,-0.101070419,-0.1235447899,-0.3715144098,-0.4118355215,-0.4568195939,0.1000407562,-0.0465258956,0.0883387774,-0.3574300408,0.2374668568,-0.039750237,0.2538008094,0.0114846425,-0.1109165773,-0.2015268207,0.3960470855,-0.0928211287,-0.2273955941,0.0414545499,0.1561560184,0.0511109531,0.0232173596,-0.3708794415,-0.1539649963,-0.0623584278,0.0670587122,-0.1709103435,0.0792751983,0.3046562374,0.0384576432,0.0212896951,-0.0239468683,-0.2883480489,0.0651211813,0.2094322592,0.2247606665,-0.0381383076,0.3764009476,0.0836886168,0.3308734298,0.2706696391,0.0717619061,0.2684164047,0.1924279183,0.2364748865,-0.1833929569,-0.3487490118,-0.1726327538,-0.1326691508,-0.3924682736,-0.1421627998,-0.0119236689,-0.0297626983,-0.3161026835,-0.2475093901,-0.5281929374,-0.3219799995,-0.0321004763,0.0037490416,-0.0355911665,-0.1123595536,0.2373548746,0.0101932455,0.1177192852,0.1573983878,0.4070904851,0.2495688796,0.0149291661,-0.381252557,-0.3094475269,0.070604451,0.3708429635,0.0020471143,0.1234323233,-0.2986415923,-0.1773284376,-0.1252613962,-0.1186028942,0.7820263505,-0.1833035648,0.025450917,0.1466884762,-0.1405288875,-0.2576371729,0.0375256129,0.1946367174,0.0166628864,0.1348632574,0.8020758629,-0.0442900881,-0.1353100836,-0.0912219062,0.2376186103,-0.2858724892,-0.395129621,-0.0272366367,-0.0948163122,-0.45247823,-0.3063827157,-0.0201471895,0.0110055879,-0.1826762855,-0.1882516742,-0.0693840161,-0.1406568736,0.119962208,-0.1822683662,0.4162624478,-0.5548484325,0.1575388312,0.1434378773,-0.0321305208,0.6181356907,0.4264189601,-0.1206234768,-0.2985330224,0.0696001425,-0.1932495683,0.1675164849,0.3233307898,-0.0035517933,-0.314917028,-0.0210111812,0.0555360131,-0.3314693272,-0.3960347474,0.0845650733,-0.1123113036,-0.2535326779,-0.3851904273,0.3009378016,0.2938147783,-0.1184904054,-0.0505024418,-0.0279624034,-0.2261317372,0.6079070568,-0.1274576038,1.1012152433,0.1730005443,0.2065189034,0.2037585825,-0.3399909735,0.3108735383,-0.1348894686,-0.0448052175,-0.3215373755,0.1035078764,-0.1803060472,-0.3237329423,0.0219975673,0.3883426785,-0.2020188421,0.4719235897,-0.2464172244,0.026270641,-0.2103811353,0.1127803475,-0.2025840133,-0.1193396673,-0.4821546376,0.0872996077,0.1393262893,0.2534198761,0.0168541931,-0.1910112798,-0.4468170702,0.0347218513,-0.1573345065,0.2845717072,0.0294333417,0.4148929417,-0.0706743822,-0.4161209464,-0.0417974219,0.6158403754,0.3683529794,-0.1017177626,-0.1701795012,-0.1532874107,-0.0333521813,0.3466129303,-0.1187566891,-0.3605044484,0.6040200591,-0.0108607961,-0.0963197798,-0.0474192277,-0.0903425291,-0.247945562,-0.403690666,0.081242308,-0.1405258775,-0.1467219144,-0.3641982675,-0.0188263524,0.1204581484,-0.3135337532,0.0866106823,0.180217132,0.207186833,0.1089263409,-0.2220783979,-0.0461017452,0.0343565345,0.5484598875,-0.0815237537,0.0112922257,0.1419294029,-0.0924002752,-0.1385920942,-0.24987863,0.3250659108,0.1075172797,-0.3133608699,0.4331635535,-0.3652237654,0.0389792882,-0.2060312927,0.0727408528,0.2630292475,-0.1315069348,-0.0264270902,-0.4607176185,-0.2394621223,0.3285416067,0.3012675643,0.13964656,0.0597363152,-0.241564855,-0.1102965698,-0.1080041304,-0.2241877764,0.0902845412,-0.0659574866,-0.1217610165,0.0137540223,-0.0132506648,0.0217709839,-0.3241366744,0.0364238918,-0.160280019,0.029897118,-0.1078219488,-0.0112192817,0.1491944939,0.1544567496,0.1018150225,-0.1634311974,-0.3246350586,0.2782616317,-0.217679441,0.1702616066,0.2939698696,0.0696022809,0.0768551901,0.0691796169,0.3163084388,-0.2097793669,0.3629968166,-0.0915644243,0.3446460664,-0.0537187159,0.0638063699,0.02729146,-0.0174489059,-0.4779081643,-0.0171687063,0.1812154949,0.402502656,0.21777156,-0.2360191345,-0.0166793298,0.0349960886,0.3909659684,0.3330205977,-0.2517447472,0.2638549805,0.0510596484,0.1676001251,0.1398449391,-0.1539597064,0.0919143409,-0.1504649371,-0.2148330212,0.0673750266,-0.1821635216,-0.0023587963,-0.4155852199,0.0969722271,0.5428466201,-0.2998737097,0.0530369282,0.3338719904,0.0535738692,0.4518009424,-0.1316324323,-0.1856984645,0.2075605541,0.2608276904,0.0370783582,0.2347220331,0.2281683832,0.1164507419,0.2893737257,-0.2608206272,-0.0247359611,-0.0742181242,-0.2026892006,-0.0764251351,0.1436886638,0.0332627818,-0.3727567196,-0.0760634318,0.0362281762,-0.0156933274,-0.1121500209,0.0482911989,0.0265506953,-0.2150345147,0.0969182402,-0.0428931527,-0.0720705017,-0.2111688852,0.1460961848,0.2383223921,-0.0736743584,0.1633494496,0.0815587491,0.1595100313,-0.0003990717,-0.1480988115,-0.0380109847,0.1211210042,-0.1941271573,-0.171332255,0.1959544569,0.61284554,0.4658444226,-0.1990788132,-0.0843776911,0.1803434938,-0.1835072041,-0.0640548691,0.1193507686,-0.2507123947,0.2812378407,0.2919749618,0.2541736662,-0.0342527591,0.207743898,0.1241123155,0.3720479608,-0.2656431198,0.6275001168,-0.2644356489,-0.2241107523,0.0326905213,-0.1488490254,-0.2340363115,-0.312730372,0.0963899046,0.0324013419,0.2786929905,0.1412012875,0.0748659223,0.2165148258,0.0332985073,0.2245589793,-0.1520264,-0.2459196746,0.2690982819,-0.5443849564,0.0925261676,0.1871257722,0.0994299129,-0.2845580578,0.1964010149,0.0975480527,0.1387923807,0.132019788,0.0729864463,0.1194150597,0.3543981314,-0.2389385849,-0.2182529867,0.1716094166,-0.1318479925,0.1520500481,-0.4748724103,0.2885645032,0.0872716904,-0.0113048451,-0.0492980443,0.1728170067,0.0974347591,0.2943975329,0.4146996439,0.1715382636,0.270788461,-0.1150451526,-0.2479918599,0.0120855477,0.0166091044,-0.1123506948,0.3540955782,-0.1954910904,0.5083062649,-0.4584724903,0.0578743257,-0.3943744302,-0.0188650675,-0.2488162965,-0.0672474504,-0.3928897977,0.1923331618,0.062591143,0.1629087478,0.4411374331,0.104962118,-0.1042922288,0.2898521721,-0.2736344635,-0.1244100332,0.2987194657,0.0868920237,-0.0897098035,-0.0402321033,0.2144563794,0.2227745801,0.22268866,-0.3981020153,-0.4122733772,0.3479643166,-0.1655627787,-0.0955845043,0.1723352224,-0.056851618,-0.0364663005,0.0312630981,0.2184519172,0.1830181479,-0.3039124906,0.3530835509,-0.1925851256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2424","title":"load_from_disk and save_to_disk are not compatible with each other","comments":"We should probably improve the error message indeed.\r\n\r\nAlso note that there exists a function `load_from_disk` that can load a Dataset or a DatasetDict. Under the hood it calls either `Dataset.load_from_disk` or `DatasetDict.load_from_disk`:\r\n\r\n\r\n```python\r\nfrom datasets import load_from_disk\r\n\r\ndataset_dict = load_from_disk(\"path\/to\/dataset\/dict\")\r\nsingle_dataset = load_from_disk(\"path\/to\/single\/dataset\")\r\n```","body":"## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n","comment_length":45,"text":"load_from_disk and save_to_disk are not compatible with each other\n## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n\nWe should probably improve the error message indeed.\r\n\r\nAlso note that there exists a function `load_from_disk` that can load a Dataset or a DatasetDict. Under the hood it calls either `Dataset.load_from_disk` or `DatasetDict.load_from_disk`:\r\n\r\n\r\n```python\r\nfrom datasets import load_from_disk\r\n\r\ndataset_dict = load_from_disk(\"path\/to\/dataset\/dict\")\r\nsingle_dataset = load_from_disk(\"path\/to\/single\/dataset\")\r\n```","embeddings":[-0.2114283144,-0.1918451488,-0.1080043539,0.4175554216,0.1788415313,0.0094063906,0.2517890036,0.3598573506,0.3484734297,0.1281708628,0.0906293541,0.3500769436,0.0675619841,0.1234149337,-0.1204770133,-0.0371128097,0.2396690845,0.0431480221,0.2499337792,-0.1075347662,-0.1955868453,0.3563712239,-0.1399710327,-0.0368440039,-0.1406249255,-0.0493298694,-0.1832460165,0.3529135883,0.329156965,0.0321098045,0.3073035479,-0.1563008726,0.0693301931,0.5561497808,-0.0001211894,0.0506671369,0.1738408804,-0.2001836896,-0.3015095294,-0.6945695281,-0.1187138036,-0.2993363142,0.1153070852,-0.3185425103,0.1460939348,-0.0933302566,-0.0410898775,-0.2919644415,0.5449647903,0.1292127073,0.1088067889,0.1360334903,0.2161679566,-0.2247788757,-0.0733537972,0.5043619871,0.0144879296,0.3683722615,0.1602687091,-0.122659348,0.1436018348,-0.007756067,-0.3498789668,-0.1579461545,0.2968854904,0.2928346992,0.3333088756,-0.4449427724,-0.0770189464,0.1268185824,0.3759043515,-0.1877381653,-0.3891636133,-0.3483689725,-0.2338569313,0.0552521795,0.4390164316,0.0421449281,-0.3749604523,0.1507019699,-0.0870305896,-0.3912684023,-0.0318068601,0.3521456718,-0.013760414,-0.0051348791,-0.0147847869,0.0822793916,0.2828969061,0.0774772167,0.3303999901,-0.4270054698,0.2116922587,0.0744733959,-0.241938442,-0.0240220483,-0.1862122864,-0.4937504232,-0.1886006147,0.3319095671,0.2741460502,-0.1271965653,-0.4449871182,0.3601480424,0.4342702627,0.2644770145,0.1294641793,0.2789503634,0.2962292135,0.0728928521,0.0235371199,0.0931862965,0.2209709436,-0.3597979248,0.2701231837,-0.0994902179,0.329267174,-0.1243443787,-0.4007611275,-0.0372810289,-0.1868371516,-0.1043993011,-0.0128229316,0.1663793325,0.1677876413,0.3757869601,0.2494017929,0.3480872214,-0.1479360014,0.4127962887,-0.109518148,0.0165517218,0.2048794925,0.2110491097,0.1036957279,-0.142040506,0.1754673123,0.0609239563,-0.0050596925,-0.1145134419,-0.0856750086,-0.3981600106,0.1377871335,0.1861119121,0.3335360587,0.166163668,0.0975963995,-0.3431741297,-0.2430256754,0.2610028684,-0.4517631233,-0.1967012286,0.0310486481,0.1310895234,-0.185141772,0.0582766943,-0.6198506355,-0.0169688426,0.1802481711,-0.3511140049,-0.0158939008,-0.1464461535,-0.1892220676,-0.4825013876,0.3719311953,0.2832886875,-0.3165066242,-0.0348567627,-0.0943608731,-0.1288660616,0.0539679341,0.4559873343,-0.0792085081,0.4309543073,-0.3708837926,0.0749579817,0.58112818,-0.633163929,-0.3446838856,0.4848427474,-0.1842655241,-0.0113264732,0.0165183768,0.15371494,0.2192757875,-0.156722486,0.3526002169,0.4051528871,0.0413565971,-0.0281782337,-0.3694812953,-0.1913498342,0.1292472333,0.1685738564,-0.0730328038,0.098166585,0.1295511127,0.0467545502,0.3483135402,0.001223473,0.0949436575,0.3625749946,0.4391660094,-0.0557367876,-0.1810932308,-0.0853820518,-0.7521965504,0.2291172594,0.1643960625,-0.263703227,-0.0664817318,0.1166827083,-0.0023818291,-0.1435466707,-0.1489107907,0.1193036884,-0.0247613564,-0.0315305926,0.1420332044,0.0465006046,-0.2370294929,0.2715092003,0.1500969678,0.0716424659,-0.1933504045,0.6443470716,-0.0932986811,-0.3494719565,-0.054322958,-0.0531428568,0.3440527916,-0.0708367154,-0.1461755186,0.4169158936,0.0370488241,0.3551664948,-0.3490378261,-0.2183392048,0.1753199697,0.1694666445,-0.0041925162,0.3294335604,0.1812501401,-0.2228418291,-0.4059947729,0.1439863294,-0.2043489814,0.3077221215,0.1737815589,0.0624669492,0.1120544747,-0.0566764772,0.1743007004,-0.2174956352,0.1451800615,-0.1665930003,0.2363595068,-0.1829375327,-0.1936642975,0.0557036027,0.3306387365,-0.0622398816,0.2126478255,0.0016045637,0.0850016102,-0.0840710625,-0.0744769275,0.4744531214,0.7585263252,0.0646188185,0.0331504419,-0.0072399406,-0.0344476923,-0.2042892575,0.199331075,0.0764113888,0.4182401597,0.4030751884,0.0133546088,0.114487879,0.0406402461,-0.2641814649,-0.1368008554,0.0475820154,-0.415517807,0.010881952,-0.1339762807,0.1105847731,-0.4369588196,-0.0678196177,-0.0970941931,-0.4311989844,-0.2040608376,0.6053654552,-0.1286749095,-0.1061858907,0.2125469893,0.020337332,-0.2196084112,-0.2715381086,-0.2055036873,0.0993707925,-0.1967610568,-0.1058103889,0.2419972718,0.0936895758,0.1949482709,-0.1839586198,-0.2100103498,-0.2731385827,-0.0082734302,0.0595859364,-0.0474168584,0.1841149777,0.0165117271,0.2255676687,0.1444560289,-0.4229511023,0.2088562548,-0.190148443,-0.2385510355,0.0926063582,0.2182631195,-0.1516928077,-0.2054022998,-0.3321664035,-0.4588548541,-0.4398188591,0.1718239784,-0.0543600917,0.1566059887,-0.3143665195,0.2442070991,0.0351690687,0.2324008942,0.0136316558,-0.0756383464,-0.1940948218,0.4806297421,-0.1420886219,-0.2508003414,0.0512710363,0.1974127591,0.0401312299,0.0472098663,-0.3657919466,-0.1209231243,-0.0864043906,0.0940967426,-0.2144783735,0.0691624358,0.2659473717,0.0285011437,0.0436960645,-0.0268319305,-0.3267620802,0.133175686,0.2165144533,0.1586675644,-0.0398017727,0.4220996201,0.0351590104,0.3567742705,0.2694926858,0.0391378067,0.1881526411,0.1792059094,0.2134340107,-0.1884242594,-0.3604229987,-0.2073598653,-0.1219459549,-0.3263211548,-0.1609517783,-0.049916748,-0.0379022881,-0.2592145205,-0.2296525985,-0.5250450969,-0.3166351318,-0.0507385321,0.0321842469,0.059485577,-0.1449345797,0.223460421,-0.0556595288,0.1530026644,0.1052327901,0.4250106514,0.2477372289,-0.0111325532,-0.3744657636,-0.2794954479,0.0033856444,0.3654209077,-0.0124469725,0.1025326774,-0.2650876641,-0.084719032,-0.1380921304,-0.1689258814,0.8337205648,-0.0480090417,0.1163205951,0.205423668,-0.1429502368,-0.3742010593,0.0357749276,0.1966569275,0.0813589469,0.1771714091,0.8088899255,-0.0656376705,-0.1427797079,-0.0996179581,0.2629836202,-0.3095370531,-0.3826812804,-0.0486196168,-0.1773068309,-0.4840148687,-0.2591538727,-0.0424859747,0.1025680155,-0.1666193604,-0.2044738531,0.0305095688,-0.1813812405,0.120746024,-0.1673353463,0.5097058415,-0.438762337,0.1563759148,0.1626300067,-0.1062507257,0.596285224,0.423392117,-0.1189095825,-0.2646346986,0.0914943516,-0.2110733539,0.1422107071,0.3218801916,0.0081472145,-0.2362228036,0.054218661,0.0670404658,-0.2457410693,-0.302038759,0.118814595,-0.1924898326,-0.2191018909,-0.5185208917,0.3012195528,0.2540456951,-0.0829499811,-0.0501749031,-0.0375279672,-0.2199164778,0.5497977138,-0.1713287532,1.0208890438,0.1000314429,0.1488337517,0.264228642,-0.3185460567,0.3059230149,-0.1494653672,0.0128649231,-0.2513437867,0.0981441662,-0.1380978972,-0.3005196452,-0.000525544,0.4005233347,-0.2648542225,0.3912701607,-0.1632882506,-0.0496528745,-0.2358871847,0.2159726173,-0.1243713573,-0.1938482523,-0.4818901122,0.0606299788,0.1260810345,0.3104171157,0.0920327529,-0.2086247802,-0.3821691275,-0.0258883722,-0.1622567773,0.271166414,-0.0465014316,0.4824004471,-0.1542370468,-0.4039051235,-0.0476117656,0.5758296251,0.2975664437,-0.1253306568,-0.1778574437,-0.1835592538,0.0205672868,0.3025239408,-0.2065059245,-0.3314888179,0.5029808283,-0.0643265843,-0.15009664,-0.0460401624,-0.1378159076,-0.3255543113,-0.2770249248,0.0802121833,-0.1919632852,-0.1068411544,-0.3981924057,0.0614352226,0.0949863195,-0.3091369569,0.0496560372,0.2182285637,0.1166982055,0.0542917512,-0.2879574895,-0.0937966928,0.0547178015,0.5448535085,-0.0171784591,-0.0035188526,0.2803351879,-0.0981831253,-0.0658634901,-0.2270571738,0.2269714922,0.0664357767,-0.2888086736,0.4507120252,-0.3576386869,0.0879854038,-0.201642096,0.1661579907,0.2737417519,-0.0611139461,-0.0055420063,-0.4276255369,-0.2893223166,0.2825799584,0.2939669192,0.1758793145,0.1567588747,-0.1529950947,-0.0603668652,-0.0623954684,-0.1899172813,0.0835017115,-0.0053910888,-0.1260296553,0.0172252133,0.0114718117,0.0790745616,-0.309907198,0.0263556503,-0.1711457819,0.0255522653,-0.0779644102,-0.0253904201,0.1646276414,0.0881364122,0.0896511823,-0.2134756297,-0.3668814301,0.261945188,-0.2076133937,0.222959578,0.2996290326,0.0483129546,0.090627566,0.0874326751,0.3761312366,-0.2290219665,0.2095327973,-0.1292053759,0.2233264595,0.0278263129,0.0835723579,-0.0008777375,-0.0324948616,-0.541876018,-0.0831164047,0.1482139528,0.3846343756,0.1514449567,-0.2208705544,-0.067636013,-0.0453106351,0.3868575096,0.3088821173,-0.2721372247,0.2473815978,-0.0292655993,0.1323946118,0.117071487,-0.1798063368,0.0262991842,-0.1083295345,-0.1815361977,0.0645572543,-0.1594798118,0.0272420496,-0.455383122,0.1263843626,0.5783295631,-0.2782410681,0.1346314847,0.3411013484,0.0592298172,0.474124819,-0.2006497085,-0.1741560251,0.2148996741,0.3066390753,0.0227488391,0.2576081753,0.2274410278,0.1206659824,0.2622335553,-0.2492772788,-0.0630299523,-0.0845090374,-0.2448536009,-0.01928542,0.1495051682,0.0278854277,-0.528380692,-0.0836867765,-0.0605075508,-0.0044037034,-0.1037560478,-0.0038001765,-0.0128486725,-0.2140301615,0.1286591142,-0.0605460852,0.0047688633,-0.176268205,0.1050168052,0.2381134629,-0.0978631675,0.2465442419,0.0645329282,0.1737645119,-0.0015649138,-0.1358410269,-0.0082426947,0.1637688577,-0.2731030881,-0.1656356305,0.1954284757,0.5575493574,0.5093078613,-0.0790171176,-0.0652617887,0.1576014012,-0.1975927502,-0.0450353958,0.0495883152,-0.2058167905,0.2743483484,0.3095663786,0.2349815965,-0.0162885766,0.2123918384,0.2283772975,0.4528605342,-0.2124085128,0.700260222,-0.2316213399,-0.2523343563,0.0432469696,-0.1375972629,-0.2326693088,-0.2715756893,0.0729434341,0.1044164523,0.263813585,0.1911268681,0.0524812452,0.2173949331,0.0901614949,0.1593888849,-0.1863623857,-0.2668263316,0.2266171873,-0.5496716499,0.0657139421,0.1354391873,0.0748809949,-0.249620378,0.2157055587,0.0673128739,0.0721877366,0.1537304521,0.0226837378,0.0963698924,0.3004963398,-0.1831515282,-0.1540468931,0.1283848882,-0.0283377972,0.1851824075,-0.4644731283,0.2657773495,0.0602811389,-0.0176665075,-0.0461481549,0.1165552288,0.0930098742,0.3153388798,0.4169470072,0.1484839171,0.2761135101,-0.1564550102,-0.2638331652,-0.0681927055,0.0650025755,-0.0887399539,0.3637222648,-0.1543534994,0.6169583201,-0.4431063831,0.0800405741,-0.4158985317,0.0172056723,-0.237482816,-0.1623784155,-0.2831349671,0.1799577326,0.0546908453,0.1158025414,0.510155797,0.119252637,-0.087879926,0.2665167153,-0.3581798375,-0.0754754618,0.3443349302,-0.0105969841,-0.0800801218,-0.0681003705,0.237819925,0.1639595181,0.3511958122,-0.4744574726,-0.3576359451,0.3663258255,-0.1155626327,-0.1743567288,0.1275445819,-0.0925646946,-0.1026215553,0.0377682,0.2736036777,0.1886525899,-0.4187124074,0.1985134631,-0.2418085486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2424","title":"load_from_disk and save_to_disk are not compatible with each other","comments":"I just opened #2437 to improve the error message","body":"## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n","comment_length":9,"text":"load_from_disk and save_to_disk are not compatible with each other\n## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n\nI just opened #2437 to improve the error message","embeddings":[-0.1681076586,-0.2181345373,-0.1153695211,0.4917304516,0.1969197094,0.0046861297,0.2464758456,0.3459077179,0.3205581903,0.100092113,0.0605783239,0.3524062634,0.120111756,0.1083519608,-0.1472454667,-0.0516677275,0.2791848481,0.0488681383,0.1801396608,-0.0774676427,-0.1710391641,0.4125433564,-0.1592954248,-0.0264890026,-0.1447586268,-0.0380912386,-0.2147710621,0.3203584254,0.3362852335,0.0826837122,0.2582362294,-0.2613320053,0.064913094,0.5925084949,-0.0001221407,0.0139371483,0.1574796736,-0.206108138,-0.2432180643,-0.6777523756,-0.1485946774,-0.2478170097,0.072476469,-0.2876604199,0.1617719233,-0.045493558,-0.03758898,-0.2565718293,0.4926871955,0.1315734237,0.1094972789,0.1354215443,0.1756507754,-0.2687014937,-0.0551934242,0.4812216163,-0.0120862918,0.328632921,0.1463250816,-0.1795566827,0.0982658565,-0.0520140901,-0.3284205198,-0.165022254,0.2781501412,0.3072389662,0.3302423954,-0.4115948975,-0.0071139378,0.0604404323,0.4102462232,-0.1800015271,-0.3227064013,-0.291913569,-0.302272886,0.0930885747,0.5114881992,0.0282057058,-0.4021246135,0.1714725345,-0.1323171258,-0.3451258838,0.0083597135,0.3800449967,0.001526663,-0.0207191054,-0.0711020753,0.0654491782,0.2787291408,0.0850345269,0.2673446536,-0.4254047871,0.2211666107,0.0623981878,-0.1951057166,-0.0331187546,-0.1750410795,-0.4435802996,-0.2203290015,0.4073023498,0.2666565776,-0.1875852793,-0.3890313804,0.379765451,0.4512857199,0.2616424859,0.0806065127,0.2174141258,0.3133448958,0.0666088015,-0.0073043797,0.0613994971,0.2424916029,-0.3516310453,0.2873622775,-0.1758417785,0.3219048083,-0.1603074819,-0.4064446092,-0.0217962246,-0.2054226846,-0.0751561522,-0.0323566049,0.17398341,0.1808242053,0.3021276593,0.2187263519,0.3852776587,-0.1451257318,0.4757513106,-0.1187289953,0.0473655388,0.1122779995,0.2014897913,0.1136166528,-0.1388703883,0.1731721014,0.1313720942,-0.0032757423,-0.1365222037,-0.1053300723,-0.3891339004,0.1210224777,0.1929780692,0.3499800265,0.1815025955,0.1003985256,-0.353298068,-0.2375360429,0.2514547706,-0.46472615,-0.1655843407,0.074187465,0.1312120557,-0.1848405153,0.0432234742,-0.6336085796,-0.05608299,0.2237087786,-0.336417973,0.0538036413,-0.0808429793,-0.1097015962,-0.4844127893,0.3189862669,0.243085295,-0.2992188334,-0.0014028449,-0.1030079797,-0.066504553,0.0386970304,0.4155364931,-0.0625758991,0.3524427116,-0.4352921844,0.1447637379,0.5695642829,-0.5937177539,-0.304023236,0.4891279936,-0.1743187457,-0.0930987969,0.0828193203,0.1616436988,0.1391991228,-0.161436528,0.296833992,0.4005681574,0.0298066847,-0.0404148437,-0.3487366438,-0.1759303063,0.1461810917,0.15119192,-0.0441960245,0.1488188505,0.1641560942,0.0429228656,0.280397445,0.0175034795,0.0822796673,0.3875748515,0.4135050178,-0.0876996517,-0.182226181,-0.1343616545,-0.7202801108,0.1720478088,0.2006560862,-0.2879221141,0.0164850447,0.1553368866,-0.0534991659,-0.1811895519,-0.15030168,0.0980397463,-0.024071712,-0.0152521711,0.1922767758,0.0716798976,-0.2339811772,0.2577826679,0.2102522701,0.0735481903,-0.2167375237,0.7177053094,-0.1179541051,-0.4030925035,-0.078644231,-0.071577549,0.3650529981,-0.0156266987,-0.1269289851,0.4093128443,0.0484787188,0.3706368208,-0.3049777746,-0.1789018512,0.1439112127,0.1402574331,-0.0033492905,0.3717390895,0.1897397339,-0.2216052264,-0.3530495167,0.1508677453,-0.1988351643,0.3073973954,0.2216867656,0.0657739043,0.0907767713,-0.1037429199,0.1823021919,-0.2113704979,0.2228257209,-0.1776246578,0.2325557023,-0.2000538409,-0.2013054341,0.0638012514,0.3221771717,-0.0196614582,0.227024287,-0.0112389503,0.0396187343,-0.0571785197,-0.0396197699,0.4296669066,0.8001823425,0.0427514762,0.07894665,0.0001031519,-0.0132325795,-0.2327139378,0.1877426356,0.1032246128,0.3849906325,0.4424247742,0.0426854901,0.1018510908,0.0192886796,-0.2288484126,-0.1542618573,0.0287192892,-0.4017923474,0.0608347729,-0.1544523835,0.0989850983,-0.4164668024,-0.0932406858,-0.1306797117,-0.4050268829,-0.2060406208,0.6740753651,-0.0753210336,-0.1659659892,0.1665352285,0.0155087123,-0.173973307,-0.2037667185,-0.2820259333,0.157212019,-0.1822092384,-0.0967486128,0.2623652518,0.0339623019,0.2298693359,-0.159076497,-0.1843864471,-0.2865065634,0.0085833166,0.047441069,-0.0424786583,0.2190158218,0.0061728195,0.184199214,0.1415010393,-0.4138883054,0.2306030989,-0.1534198523,-0.2696866393,0.1534806341,0.2420785427,-0.1632349938,-0.1972946525,-0.3272341192,-0.3907184899,-0.4309913516,0.1575457901,-0.0291907676,0.1153655052,-0.3603465855,0.2149811834,0.0524358191,0.2726559043,-0.0424094684,-0.1198060811,-0.1524910331,0.4949252605,-0.1646557599,-0.2357034534,0.0774480402,0.1924850941,0.0148697011,0.0334796198,-0.3435988128,-0.1371316165,-0.1055366993,0.1019806489,-0.2360516638,0.0973130241,0.2969856858,-0.0161304381,0.0343816616,-0.0496855006,-0.3006058931,0.1463090777,0.2443744391,0.1894333959,-0.0727317184,0.4041881859,0.0615110919,0.3414723575,0.2757255733,0.0686196983,0.2690562606,0.1771882921,0.2081651241,-0.2319686413,-0.3407472074,-0.2366323918,-0.112148568,-0.3678083122,-0.1531658024,-0.0395940691,-0.0666682497,-0.2742435634,-0.2144827694,-0.5046338439,-0.2423088998,-0.0818685666,0.070224002,0.0238798428,-0.157987535,0.2208390981,-0.0395012237,0.0974809825,0.1029806882,0.4299116433,0.2445394695,0.0022246256,-0.4030502141,-0.3132345974,0.0426497236,0.3428678513,-0.020211827,0.1102962941,-0.3220983148,-0.1102884039,-0.1069926247,-0.1912423223,0.8496170044,-0.1010018587,0.1206636876,0.2289735973,-0.144491151,-0.3213548362,0.0279253349,0.1981248707,0.0702427104,0.170038715,0.829059124,0.0263381079,-0.1218597218,-0.0907756761,0.2002664506,-0.2680677176,-0.4190010726,-0.0671719015,-0.1590631008,-0.4474275708,-0.2539502084,0.0028049755,0.0721462667,-0.1678455025,-0.2035073489,0.045406919,-0.1956140846,0.1486805975,-0.19415766,0.4682685137,-0.4977220595,0.1268540472,0.0861467794,-0.091901429,0.5544959903,0.4019116759,-0.0486627258,-0.2147897035,0.1577795297,-0.2418036014,0.1508675218,0.3062355816,-0.004511572,-0.2335560322,-0.0323236808,0.0334373787,-0.2834764123,-0.337682575,0.1505901664,-0.2281668335,-0.1704630554,-0.3786385059,0.3434433639,0.2219545245,-0.0639181808,-0.0311050061,-0.0542742126,-0.2441559434,0.4927356243,-0.2300070673,1.017547965,0.1686084569,0.1556392461,0.2808715403,-0.2906357646,0.2072784454,-0.2032857388,0.0351046696,-0.2937127054,0.0798091218,-0.1998883486,-0.3617560267,0.0512316227,0.3818473518,-0.2669062018,0.3821332157,-0.2240542471,0.033227805,-0.2588190436,0.0792668462,-0.197872743,-0.122322984,-0.4968791604,0.0613832735,0.1626824886,0.3321671188,0.0666371062,-0.1987428516,-0.4520527422,0.001752971,-0.177780211,0.2927499115,0.0077797403,0.4690177441,-0.1218079329,-0.4343040884,-0.0882845223,0.6285147071,0.3361135721,-0.1638363749,-0.1784063131,-0.1855338514,-0.0224822834,0.3073958457,-0.1806085408,-0.3306483328,0.4966150522,-0.0787229761,-0.1012059301,-0.0397893935,-0.0975532532,-0.3287011683,-0.3009430766,0.0798832104,-0.2010727525,-0.1375395656,-0.4131033719,0.0577495471,0.1038301513,-0.3059010506,0.050138548,0.170348227,0.1361612678,0.0269280169,-0.3247143328,-0.0408072621,0.0286559779,0.6101368666,-0.0618074462,0.005965041,0.2080488503,-0.0875976756,-0.1017532796,-0.2302686423,0.2586213946,0.1115736142,-0.295904696,0.4376428425,-0.4072172642,0.0715248138,-0.1995759904,0.160214141,0.2563301027,-0.0968805552,0.016502941,-0.4600673318,-0.2476039082,0.2655860484,0.3078342378,0.1108050048,0.1215878725,-0.1981823891,-0.0357036553,-0.1041994095,-0.1725202799,0.1346328557,-0.025627248,-0.0812274367,0.0488666669,-0.0054875147,0.0921454802,-0.3015746474,0.0103543652,-0.1510090679,0.0087954169,-0.078673631,-0.0195539873,0.1662075967,0.1238947138,0.0685310066,-0.1932589561,-0.3482718766,0.2474868894,-0.1780784428,0.2154512405,0.2484663725,0.0504480004,0.1020037606,0.0642257631,0.3402286172,-0.1970232427,0.1901216656,-0.1404139102,0.2589696348,0.0254010595,0.0437342823,0.0079999305,-0.0069859885,-0.5984080434,-0.075346835,0.1278583705,0.3667227328,0.1660803109,-0.2780553997,-0.0458589643,-0.0671062022,0.3789433837,0.3226670325,-0.2714571953,0.2736759484,-0.0141682038,0.1367654651,0.1546488851,-0.1643156707,0.0832922831,-0.1215425208,-0.2290084511,0.0415943302,-0.1418477744,-0.004806865,-0.4417809248,0.1144258901,0.5740604401,-0.3051282167,0.1060018986,0.3886294067,0.0555611402,0.5100387931,-0.1622303128,-0.1494645774,0.1944114119,0.3098216653,0.0502313077,0.2493999302,0.2396464795,0.1096768007,0.2278083414,-0.2601919472,-0.081879124,-0.0888715386,-0.1501507163,-0.0319200195,0.1504244059,0.0736699402,-0.491401583,-0.0393665619,-0.1229659915,-0.0232871696,-0.1199936196,0.0270919893,0.049964115,-0.2554581761,0.1470973194,0.0010821526,-0.0338502191,-0.1405462921,0.1482603252,0.2401221097,-0.090352267,0.2368424535,0.0866849795,0.1662480384,-0.0359359495,-0.1589110345,0.0077050868,0.0968554318,-0.2458583117,-0.1830453426,0.1719953716,0.5635792017,0.4967333078,-0.136543721,-0.1247806773,0.1610536724,-0.2035741359,-0.0159357954,0.1188536286,-0.2030373663,0.3163867891,0.264298141,0.2215055078,-0.0066992757,0.242980212,0.1884783357,0.4206314683,-0.2239786834,0.6464163661,-0.2771496475,-0.2561892271,0.0847750828,-0.2057349831,-0.2246133089,-0.2484834194,0.0643294603,0.094038114,0.2793364227,0.1921066046,0.049111601,0.2419462651,0.0153489523,0.1396918297,-0.2213706523,-0.25222826,0.1817067564,-0.4862005711,0.1119265407,0.1848092675,0.0915237442,-0.3148504496,0.2107546329,0.060674455,0.0591742657,0.1621070951,0.0016317694,0.1503932923,0.3700338006,-0.1825974286,-0.1661577523,0.1116635352,-0.0427446105,0.1498661041,-0.4698968828,0.2739081681,0.0684273168,-0.0145280147,-0.0516377278,0.125639841,0.0316874124,0.3504182398,0.4354067147,0.1439888179,0.3071696162,-0.1101925001,-0.2642188966,-0.0585841946,0.0411954224,-0.0755819231,0.2946276665,-0.1979242861,0.5968320966,-0.4701016545,0.1207490861,-0.4092777967,0.026999414,-0.2325184047,-0.0760272741,-0.2734827995,0.1630407125,0.025326049,0.1283552796,0.5528568625,0.1284504086,-0.0941740423,0.2982805669,-0.336194694,-0.0659695417,0.2789391875,-0.0118248211,-0.1435843706,-0.0045504514,0.2297326326,0.1956397146,0.303992033,-0.4641840756,-0.4126655757,0.3504126668,-0.1335388571,-0.1364109218,0.1472329795,-0.1015168726,-0.101750508,0.0650068372,0.2746614814,0.2564640343,-0.3799902499,0.2520468831,-0.2555972636]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2424","title":"load_from_disk and save_to_disk are not compatible with each other","comments":"Superseded by #2462 ","body":"## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n","comment_length":3,"text":"load_from_disk and save_to_disk are not compatible with each other\n## Describe the bug\r\nload_from_disk and save_to_disk are not compatible. When I use save_to_disk to save a dataset to disk it works perfectly but given the same directory load_from_disk throws an error that it can't find state.json. looks like the load_from_disk only works on one split\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"art\")\r\ndataset.save_to_disk(\"mydir\")\r\nd = Dataset.load_from_disk(\"mydir\")\r\n```\r\n\r\n## Expected results\r\nIt is expected that these two functions be the reverse of each other without more manipulation\r\n\r\n## Actual results\r\nFileNotFoundError: [Errno 2] No such file or directory: 'mydir\/art\/state.json'\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.0-73-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.7.10\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: \r\n- Using distributed or parallel set-up in script?: \r\n\r\n\nSuperseded by #2462 ","embeddings":[-0.1741850376,-0.2175427824,-0.1080487669,0.4549724758,0.1950396746,-0.0049299239,0.2797517478,0.3457045555,0.3133310974,0.0889768675,0.0539686345,0.3709700704,0.1151137576,0.1141666248,-0.1080199927,-0.0385595188,0.2935668528,0.0366213173,0.1656907648,-0.0801072717,-0.1850747019,0.3960942924,-0.1790975481,-0.0474063791,-0.1371829361,-0.0177341197,-0.2328820229,0.3070189059,0.3266166449,0.0799820051,0.266621083,-0.2427215278,0.0644829571,0.5731480718,-0.0001207328,-0.0023862247,0.1753364652,-0.2129469216,-0.2404965758,-0.6553148627,-0.1368522495,-0.2628899217,0.0761246383,-0.3093193471,0.1531127542,-0.0590892285,-0.0378882252,-0.257784158,0.4702774286,0.1388538331,0.1295469552,0.1565347016,0.1883912235,-0.2807461619,-0.0569723807,0.4619095027,-0.0107013676,0.3501386046,0.1577370763,-0.1821838915,0.090612635,-0.0413173176,-0.3210597336,-0.1778950989,0.3031707406,0.3026190102,0.3090175986,-0.3831441104,-0.0029779936,0.0592823289,0.4123271406,-0.1801104993,-0.3320390582,-0.2979134023,-0.3130901754,0.0821294114,0.5067589283,0.0338608287,-0.3873398304,0.1694743186,-0.1336309463,-0.3335934579,0.0200518146,0.3749040663,-0.0131652215,-0.004759734,-0.0873841494,0.0647590831,0.2659870684,0.0794325173,0.2965069413,-0.4374370277,0.2262961268,0.0571193099,-0.1841377765,-0.0282017402,-0.1590981632,-0.4328099787,-0.2227683514,0.4121014476,0.2615071833,-0.1750761718,-0.370072484,0.389716655,0.4258052707,0.2676742375,0.0766051039,0.2089511454,0.3083907068,0.0780285001,-0.0086487494,0.0549238101,0.2698770463,-0.3326566219,0.302005887,-0.2005934864,0.3374833465,-0.1524735987,-0.4039333165,-0.0387604833,-0.179682225,-0.0768074691,-0.0395287983,0.1770813763,0.1785613447,0.3173444569,0.2240362912,0.3809550107,-0.1436279565,0.4768669307,-0.1327084005,0.0291987285,0.1252922416,0.1794851571,0.1223834381,-0.1469604075,0.171829164,0.1330770254,0.0043957895,-0.1334266961,-0.1071257219,-0.3840948641,0.1573795974,0.1899792403,0.3244677484,0.1522443891,0.0793783963,-0.3377861679,-0.2238088548,0.2747085094,-0.4578325152,-0.1552637666,0.0696945637,0.1509216577,-0.1869102716,0.0181139577,-0.6262210608,-0.0293959323,0.2122650594,-0.3289220631,0.0712535903,-0.0950983241,-0.1200670302,-0.5036562085,0.3249674737,0.2296145558,-0.2967191637,-0.0297500677,-0.1166439429,-0.0795292705,0.0401144624,0.3828337193,-0.067750223,0.3416960537,-0.416181922,0.1386785805,0.5395057201,-0.5907687545,-0.3202836215,0.4617722034,-0.160029918,-0.0720641539,0.0360519029,0.1694364846,0.1761552244,-0.1557423919,0.3026770949,0.3973477483,0.0374396071,-0.0389236659,-0.349655956,-0.1652843356,0.1448416859,0.1477775425,-0.0291067604,0.170980081,0.1572743803,0.0460217334,0.2836943567,0.0260278732,0.0969678909,0.3936819434,0.4079156816,-0.089186959,-0.1752797961,-0.1382092685,-0.7377340198,0.180347234,0.2155535519,-0.291621834,0.0200978853,0.1495955586,-0.0513606034,-0.1568711102,-0.1353937238,0.0878796577,-0.0126533406,-0.0181339402,0.177340582,0.0548683256,-0.2440370321,0.2422351092,0.2197212875,0.0840933174,-0.2013452351,0.708096087,-0.111263521,-0.4014307261,-0.0785604566,-0.0670916736,0.3803290725,-0.030093994,-0.132810384,0.4391628802,0.0431517214,0.3606216609,-0.3284679949,-0.1975552738,0.1478406638,0.151746586,-0.0044674734,0.3687996268,0.1904227287,-0.2202705741,-0.3646808863,0.1705083996,-0.188096568,0.2699759305,0.2146614939,0.0690702796,0.0913046151,-0.1123899892,0.1743608266,-0.2026274204,0.2067235857,-0.1707656235,0.2278927267,-0.210497722,-0.1929557025,0.0896043107,0.352747649,-0.0527392849,0.2286965102,0.0000292414,0.0336607248,-0.0686584339,-0.0418202206,0.4375598133,0.8096470237,0.0420154631,0.104570359,0.0047306828,0.0005254145,-0.2354211658,0.2001147121,0.1121289283,0.368101716,0.4121638834,0.0638748854,0.0911850259,0.0100965397,-0.1801524907,-0.1468693465,0.0189258363,-0.3808590174,0.0710718259,-0.159812361,0.1180791482,-0.4205246866,-0.0882191584,-0.134395346,-0.4101064801,-0.2210122496,0.6655914783,-0.0855315477,-0.1720484048,0.1647708863,0.0220322404,-0.1741300374,-0.2039710283,-0.2965391874,0.1563245654,-0.1970203072,-0.0818116367,0.2310630679,0.0189663786,0.2236329019,-0.128901571,-0.1780097336,-0.2815443873,-0.0044004428,0.0485556163,-0.054037977,0.2166307122,-0.0096397065,0.1441196799,0.1407511532,-0.3934119642,0.2086977214,-0.1831230074,-0.2654940486,0.1455908865,0.2697132826,-0.1600453854,-0.2131361961,-0.307263881,-0.4002050161,-0.432318598,0.16489169,-0.0142886061,0.1011734605,-0.352956742,0.1884857714,0.0497646779,0.2820631862,-0.056831751,-0.1381598264,-0.1580025107,0.4889185727,-0.1787905246,-0.2611337304,0.0643889531,0.1965977401,0.0092602652,0.0476113856,-0.3205881715,-0.1230755448,-0.102270104,0.10839127,-0.2159924805,0.0926870853,0.3044285476,-0.0204092283,0.0234332345,-0.0667332262,-0.2982936203,0.1279776692,0.2367207408,0.1669728011,-0.05856473,0.3831931353,0.0725770667,0.3273150325,0.2649621367,0.0553882718,0.271140784,0.1843617857,0.2193586826,-0.2430397272,-0.3511774838,-0.2322369367,-0.1355687529,-0.3366701603,-0.1519722342,-0.0492396094,-0.0792031363,-0.2739039361,-0.2010303736,-0.5077374578,-0.2647830546,-0.0977065116,0.0601314418,0.0555660613,-0.1404201686,0.2530914545,-0.0588107407,0.0916051269,0.097285673,0.4274416864,0.2471559346,-0.0103408089,-0.4146606028,-0.2602306008,0.0161137711,0.3595975935,-0.0270818751,0.1069888845,-0.3083598614,-0.1445273906,-0.1019494012,-0.2024014145,0.8335738182,-0.0719461367,0.1202824041,0.2423020452,-0.1393557638,-0.3338867128,0.0319127291,0.1846304685,0.0673711598,0.1916098744,0.8335217834,0.0164853446,-0.1284486204,-0.0762011334,0.1787870377,-0.2634699345,-0.4084690511,-0.0771087036,-0.1675077975,-0.4340544939,-0.2657866478,0.0005173006,0.0734298527,-0.195518136,-0.1995940357,0.0234586783,-0.1964735836,0.1317658722,-0.210076347,0.4697755575,-0.4801239371,0.1290376633,0.0925836712,-0.0894241035,0.5371629,0.3915741444,-0.0618874542,-0.2098420709,0.1597415209,-0.2422109246,0.1448732167,0.3070548475,-0.0092940032,-0.2117106169,-0.0224603303,0.0146687217,-0.3057922721,-0.3296527267,0.149797678,-0.2261209488,-0.1653711498,-0.3951313794,0.3515120447,0.201805234,-0.0690681413,-0.0389370695,-0.0604854003,-0.242228955,0.4926623106,-0.232982412,1.011516571,0.1551804096,0.1683779955,0.298137635,-0.2979191542,0.20187217,-0.181317538,0.0427345335,-0.2935013175,0.0727456883,-0.1981283277,-0.3603641093,0.0406589396,0.3813097179,-0.2840274274,0.3681771457,-0.2059717178,0.0551169664,-0.2273862064,0.080427289,-0.198308602,-0.1451503783,-0.4947015941,0.0852470249,0.1555424482,0.3395808935,0.0596735962,-0.2046436816,-0.4431525171,-0.0046470766,-0.1458463073,0.3135176897,0.0267383158,0.4741326571,-0.1069147885,-0.4402856231,-0.10879419,0.5982846022,0.3252527416,-0.1624962091,-0.1572424769,-0.1681185514,-0.0247350261,0.3246048391,-0.1675503999,-0.3310705125,0.5112570524,-0.0672809929,-0.1019872874,-0.0523897223,-0.0864626244,-0.3549807668,-0.301412046,0.0960793123,-0.1994906813,-0.1413325518,-0.4102976918,0.0694327727,0.0997650474,-0.3283160925,0.0647770017,0.1554223299,0.1417983174,0.041538544,-0.3381700814,-0.0322354324,0.0318711549,0.6338948011,-0.0632138476,0.003613356,0.2152079344,-0.0881011784,-0.1048128083,-0.2349233627,0.249082461,0.0923441201,-0.2820670903,0.4679807723,-0.4216115773,0.0836714581,-0.1931373477,0.1515527219,0.2180527151,-0.1097215712,0.0062096105,-0.4527426958,-0.2724584639,0.2592954934,0.319224447,0.1115489826,0.1625521332,-0.1731567681,-0.0350025967,-0.0966922268,-0.1937218904,0.1576375365,-0.0097846007,-0.0722496808,0.0596815012,-0.042585928,0.1050514504,-0.2939847708,0.0178318322,-0.1390362382,0.0092274956,-0.0997530669,-0.0457246751,0.1622449905,0.1276208907,0.0480925217,-0.2203915715,-0.358972162,0.2340079546,-0.1833221316,0.2182694823,0.2429759204,0.036278218,0.1227678359,0.0624634586,0.3598231673,-0.161417231,0.1746925861,-0.1456300318,0.2639669776,0.0069482327,0.0590112396,0.0148985535,-0.0112004336,-0.6159088016,-0.0592935383,0.1346312463,0.3673656285,0.1984713674,-0.2746374905,-0.0487985611,-0.0846737698,0.3978588283,0.3077802062,-0.2704910338,0.2545284033,-0.0257278569,0.1500354409,0.1468463689,-0.1740762591,0.0787556916,-0.1045076922,-0.2047307789,0.0310791265,-0.1416525841,0.0085821608,-0.4585697055,0.1147698835,0.5566176772,-0.2922749221,0.1082894132,0.403530091,0.0707889125,0.5296090245,-0.134408474,-0.1294270158,0.194217518,0.3386872113,0.0935606807,0.2758493125,0.2410784215,0.1001741067,0.2482503057,-0.2913377881,-0.0902839825,-0.1153521389,-0.1471332461,0.0035682505,0.1595460922,0.0774582028,-0.4891764522,-0.0473516732,-0.1188406199,-0.0045102164,-0.1281019449,0.0030977849,0.0528532825,-0.2461189181,0.1562211812,0.0012201039,-0.0344943814,-0.1437168568,0.1547335684,0.2462415248,-0.0753972828,0.2397570461,0.0884191543,0.1345272958,-0.0230617281,-0.1505259573,-0.0200029034,0.1098239422,-0.2235678732,-0.1769298911,0.1741528362,0.5485336185,0.4762207568,-0.1445257962,-0.1101788655,0.1535407603,-0.2013629824,-0.0241847355,0.1205772236,-0.1917116344,0.3214882314,0.2714609206,0.2333847731,-0.0161363631,0.2277991027,0.2036608607,0.4182049036,-0.2545616329,0.6589789391,-0.2746686935,-0.2712714374,0.0553288758,-0.1882340461,-0.2282122672,-0.2550166249,0.0803884491,0.0941591933,0.289435029,0.1793227047,0.0594320484,0.2557249665,0.0254196767,0.1344584227,-0.2323471308,-0.2434663326,0.1721504629,-0.4794439971,0.1147091761,0.200221315,0.1137315929,-0.3045061529,0.2004140764,0.0708703697,0.0373187102,0.123494789,-0.0057117715,0.1468186527,0.3458431363,-0.1695055217,-0.1611426473,0.1065401062,-0.0553547367,0.1592568308,-0.4846781492,0.2827045321,0.051592581,-0.0024237102,-0.0639998987,0.1160035655,0.0359393135,0.3272209466,0.448566705,0.1511261165,0.2832087278,-0.1110362932,-0.2636474371,-0.081228897,0.0411652401,-0.0691884756,0.2973334491,-0.2241839021,0.5822125077,-0.4309474528,0.1228323653,-0.403334856,0.035134837,-0.2413585186,-0.0975302681,-0.2868119478,0.1488729417,0.0084481109,0.11702618,0.5552584529,0.1625957936,-0.1036910117,0.3009862602,-0.3470260203,-0.0658441931,0.2851970494,-0.0280648936,-0.158643648,-0.0292342603,0.218660146,0.1948000491,0.2704411149,-0.4688139856,-0.4140500128,0.3445144594,-0.1536240429,-0.1378882527,0.1637196541,-0.1055007428,-0.1004948542,0.0542978384,0.2887746692,0.2711272836,-0.3789721727,0.2430325449,-0.2436443716]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2415","title":"Cached dataset not loaded","comments":"It actually seems to happen all the time in above configuration:\r\n* the function `filter_by_duration` correctly loads cached processed dataset\r\n* the function `prepare_dataset` is always reexecuted\r\n\r\nI end up solving the issue by saving to disk my dataset at the end but I'm still wondering if it's a bug or limitation here.","body":"## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No","comment_length":53,"text":"Cached dataset not loaded\n## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\nIt actually seems to happen all the time in above configuration:\r\n* the function `filter_by_duration` correctly loads cached processed dataset\r\n* the function `prepare_dataset` is always reexecuted\r\n\r\nI end up solving the issue by saving to disk my dataset at the end but I'm still wondering if it's a bug or limitation here.","embeddings":[-0.2978544831,-0.0423550606,-0.0758303478,0.4236162901,-0.0719461665,-0.0276887827,0.2326412499,0.3540042639,0.2089607567,-0.0697676763,0.0410520583,0.1886659861,-0.2849694192,-0.2252903432,0.1128703952,0.2975529432,0.0528036244,0.1180419326,-0.1090572625,-0.1540792584,-0.0150639126,0.2576181591,-0.0469234958,-0.093652606,-0.2029935271,0.1055824384,0.0262820609,0.1340813786,0.0921806395,-0.2740397155,0.5368282199,0.0920564905,0.1350365132,0.5165358186,-0.0001210673,0.1466437578,0.2859062552,-0.215739131,-0.2181520611,-0.1280438304,-0.2788822949,-0.0832798928,-0.1238182411,-0.2300582379,-0.0120474277,0.0122144446,-0.0481124632,-0.8536034226,0.459944576,0.0661035106,0.1565443277,-0.1820604801,-0.3796681762,0.1259405762,0.038267713,-0.07413809,0.0205731262,0.1491375864,0.1322319508,-0.0945808291,-0.0412483253,0.5044329762,-0.0829141736,-0.0581240319,0.0016792879,0.0538822822,0.2330061644,-0.1952100992,0.1459397674,0.1084957868,0.1963786185,-0.2124014497,-0.3501954079,-0.2727812529,-0.3075908422,-0.2135212123,0.3252165914,-0.1555833817,0.1468669474,0.1651790589,-0.6091327071,-0.0123855583,0.1913139969,-0.0265445951,-0.0197416823,-0.1129705459,-0.168192789,0.1520161331,0.1083762646,-0.0320031121,0.2745790482,-0.1657436788,-0.0053228564,0.4561800063,-0.0696609691,0.2084786743,0.3096645474,0.0638183728,-0.0398250967,0.2631065547,0.1751727313,0.1068111286,-0.0345857143,-0.0644117445,0.1306186318,0.2133267969,0.2404883653,0.2522965968,-0.049673751,-0.1146898791,-0.5343170762,0.0586863309,-0.1254806519,-0.2830507755,0.7806978822,0.1787222922,-0.0620118529,-0.401501447,0.064640604,0.041482646,-0.42515257,-0.3526496589,0.1482111812,0.4227842987,-0.22410357,0.2358968258,-0.0378687978,-0.1386456192,-0.2654689252,0.1546202749,-0.1668414921,-0.2038098127,-0.3884489834,0.1796002686,0.37313658,-0.4005204141,0.2701569796,0.0872643217,0.0604942627,-0.2748971879,0.165197596,-0.1340657026,0.5168511868,0.0657960027,0.0053642825,0.2182988822,0.1590651125,0.0378639959,-0.2242561281,0.1018682271,-0.5072760582,-0.1304601878,0.318675667,0.0764246434,-0.2396333665,0.3016883731,-0.209837392,0.2644234896,0.3779603243,-0.3355059624,-0.1166514009,-0.2069757432,-0.2666144669,-0.4785521328,0.4195537567,0.6090718508,-0.2946233451,-0.3102104068,-0.1382859498,-0.0281040333,-0.0599078014,0.0682887435,-0.2763515711,0.0992670506,-0.3678852618,0.0261640139,0.4927723706,-0.3572143912,-0.4876314104,0.0474483036,0.250859648,0.5972858667,-0.0802121386,-0.1463817358,0.1930467933,-0.2266284823,0.2826514542,0.1132326648,0.0213674568,-0.1429440379,-0.1665978581,-0.1039288044,0.1472658366,-0.1259729266,0.3338222206,0.1621620655,0.1865675449,-0.1245006174,0.1304099411,-0.0470252037,-0.0186559297,0.3194216788,-0.0883403495,-0.0465562865,0.1233384088,-0.2534306645,-0.3030246496,0.4161659777,-0.277041018,-0.5969229937,0.0289568231,0.0484691896,-0.207630232,-0.0645195767,-0.170820117,-0.2434182167,0.0316517092,0.2297227979,0.2284024656,-0.0210998245,-0.0545175895,0.5442660451,0.0102770561,-0.2740502059,-0.2433663309,0.1692317128,0.2576344907,-0.1716133356,-0.3055517972,0.0876919404,0.3064357936,-0.1491903067,-0.1570721865,0.1695891917,0.0594948232,0.2013689876,-0.2249587923,0.2929288149,-0.1600542963,-0.1222959012,0.1584830284,0.0284264144,-0.0416131653,-0.1636959612,-0.0239242911,0.3278394341,0.1689307094,0.0020596988,0.0516499132,-0.18602705,0.1895095557,-0.1054123193,-0.117535077,-0.2751552165,0.1547856033,-0.1886118948,0.4881473184,0.2394631356,0.0860364437,0.1719860733,0.4886368513,0.2141722739,0.1340205371,-0.0534790717,-0.3166265786,-0.1791449189,-0.1831711382,0.0051268237,0.162058875,-0.000577304,-0.0210616924,-0.023351742,0.0179936346,0.2533574998,0.0779890493,-0.2341278642,0.3470124602,0.1444575787,0.119849056,0.0359204039,-0.1744266897,0.0458373725,0.1211205721,-0.0410864949,-0.466869086,0.1364646703,-0.3292939067,0.2719795704,-0.2452030629,-0.0638307258,0.0286573898,-0.2246178985,-0.1919744313,0.3194694519,0.2880186141,0.2144105434,-0.3410937786,0.1146627292,-0.1894728988,-0.1603946239,-0.0100804102,-0.1131348908,-0.3489112854,-0.0086926008,0.1438347846,-0.2055016458,0.2075018138,-0.0255988836,-0.1401118487,-0.2677106261,-0.3595719039,0.0935123265,0.2070462853,0.3455573022,-0.2412908524,-0.0900674835,-0.20699431,-0.0637390614,0.1708731949,-0.2313913256,-0.1428841799,-0.0684183836,-0.0057264441,0.0865047798,0.0938635096,-0.4032764435,-0.3004500568,0.0697977245,-0.1072492748,-0.0847632661,0.0376866087,0.0733904243,-0.1441195458,-0.2160439491,-0.0327869318,-0.0495372526,-0.4220378101,-0.4178791642,0.2509204149,-0.2321704179,0.0369442664,-0.1274863183,-0.1213700399,-0.0014036332,0.5802029371,-0.3909652531,0.1654570103,-0.0096159568,0.0502640828,0.000515733,-0.1375783086,0.444417119,0.1967123896,0.1541717947,-0.2419928163,-0.411700666,0.0349578336,-0.0705297217,0.2873363793,0.1238329336,0.1701889038,-0.0201322362,1.0965393782,0.3184102774,-0.2103504837,0.0661234036,-0.0407362431,0.2642996013,-0.0109162964,-0.3952404857,-0.1654396057,-0.1520673931,-0.1927796453,0.0578009188,0.063286595,-0.308313489,-0.0400371552,0.2140312791,-0.7177140713,-0.208044365,0.2859528661,-0.6640021205,0.2785124779,0.1591979861,0.1727760732,-0.5495753884,0.068691723,0.082135193,-0.1490960866,0.5069204569,0.0073018586,-0.1675949842,0.1631207913,0.0367319956,0.1653803885,0.2624444664,0.2216524333,-0.123574324,-0.0967764705,-0.2273300737,-0.0536234789,0.4456322789,-0.602388978,0.2966638803,0.3834201992,-0.3278230131,-0.1631358862,-0.2292153388,0.1275748312,0.3774194717,0.1738161296,0.2964749932,-0.0295395795,0.162295267,-0.0784908682,0.1458637416,-0.2814039886,-0.0492803566,-0.0455858894,0.135243699,0.1419732124,-0.0365296789,-0.2272629142,-0.0217074193,-0.0517015569,-0.053078834,-0.4242401421,0.0739685223,0.2555040121,0.3246227205,0.5262849927,0.0624686703,0.4390742183,0.2741711736,0.0351012796,-0.0851440579,0.5787379742,-0.2023314238,-0.1998823136,0.009701441,0.2453024536,-0.1176062599,0.0950476453,-0.385099113,-0.1949116141,0.1685321629,-0.1146743968,-0.1057470292,-0.1660600156,0.055524677,0.2818845212,-0.5762344599,-0.6871487498,0.4059185982,0.4733220041,-0.4103278518,0.7267813087,-0.5869186521,-0.2629710436,0.2095643282,-0.1846629679,0.7170375586,-0.2186813653,-0.0995574594,-0.0923304334,0.3079008162,-0.0393021181,-0.172490254,-0.0400164574,-0.2709791064,-0.1853233725,-0.1254699975,-0.2476443499,0.084847413,0.2775765359,0.0025101651,0.3366933763,-0.2414913028,0.4807322919,0.0184305776,-0.0930507928,0.1714968532,-0.060281191,-0.3047181368,0.1155695915,0.0777915493,0.1758089811,-0.0594418831,-0.0584447756,-0.1147260442,-0.0674225315,-0.3334651887,0.2548058629,-0.6193088293,0.4070167542,-0.1185960695,-0.1381233037,-0.1180677861,0.1156828254,0.1650652289,0.0837796777,-0.0401754305,0.2410151958,0.326459676,0.1791501194,0.0417668931,-0.1866993457,0.1389885992,0.0957636237,-0.1071276888,0.1587900221,-0.1544303447,-0.375262022,-0.176004231,0.1844714731,0.0091225058,-0.014729579,-0.2240322977,0.2788906991,0.2617226243,0.115472585,0.0510925129,0.2583168447,0.1049676538,0.3374639452,-0.1267133802,-0.2926823497,0.0319456309,0.546869874,0.1948584914,-0.1740589887,0.2723882496,-0.1755636781,-0.0421947949,-0.0335020795,-0.3243037164,0.104490377,-0.2906941175,0.0863463804,-0.3525417745,0.1369680464,0.2956292927,-0.151681602,0.1772856414,0.2408417314,-0.1096671373,-0.3795477748,-0.0640935153,-0.0311764106,-0.0599763021,0.0721737891,0.1183805019,-0.297593236,-0.1280510724,0.331507951,-0.2037172467,0.1426204294,0.138836503,0.0612695515,-0.0742919669,0.2138422877,0.2881692052,-0.2710391283,-0.0206499267,0.0642175302,-0.3246044219,-0.0957717597,-0.1054293364,0.1195218414,-0.2374317795,0.1265368909,-0.0961164013,-0.0452616774,0.2535043955,-0.2583850324,0.1769426316,0.4429425895,0.2472510934,-0.1213730574,0.0362662934,-0.2371957153,-0.1984448284,0.2644020021,-0.1955887228,-0.0001742452,0.0041380827,0.3197752535,0.0020018611,-0.0661249831,-0.1795742959,-0.1680130959,-0.0040729488,0.1874286681,0.3716533184,-0.1524595618,0.1358595341,0.135865286,0.229644686,0.2164588422,-0.1311828345,0.0259596594,0.00648139,0.1072962359,-0.0830922425,0.0035993913,0.3172124028,0.1602545381,-0.1353239864,0.278639257,0.1040155143,-0.0837675557,0.2344512939,0.0386339016,0.3939882517,-0.3727237284,0.3957624435,0.2192043662,0.2516993582,0.1452663988,0.1796540171,0.1162650362,0.0013625554,0.256772697,0.178279236,0.3827572465,0.2070094794,0.0789957568,-0.1574572772,-0.5870573521,-0.0352019109,0.3819710314,-0.3661758602,0.501277566,0.0281729121,0.058995869,0.1634183675,-0.3162510097,-0.3057076931,0.3864521682,-0.0145207196,-0.2368498594,0.1871301383,-0.2024393827,0.0762404501,0.0389943346,-0.007976192,0.0427843966,0.5030100942,0.1696486175,-0.221239239,-0.0135474857,-0.18846865,0.1144961566,0.4267623127,-0.1168560162,0.1467285603,-0.0639424175,-0.1349025369,0.0471755788,0.3376615345,0.4609377384,0.2616270781,0.1134784073,0.0790454522,0.4846385121,0.1015991122,-0.2699878216,0.3003548682,-0.0987127572,-0.2325969636,0.151421845,0.1502105594,-0.0925896019,0.2429991663,-0.2007202357,0.3491633534,-0.2665272057,0.0634320751,-0.1339989007,0.2627901137,-0.0019667032,0.1811152548,-0.049006585,-0.1975012273,0.7122183442,0.1442073137,0.2382027507,0.0097113065,0.0515268631,0.2391056567,0.1620854586,0.3663572967,0.058446452,-0.1798638701,-0.0593992993,-0.6679871082,0.0174868777,-0.3761491776,0.2178829014,-0.0890114754,0.2133032382,0.2604070902,-0.0863466859,0.2404281944,0.001278261,0.1334682405,0.2444926947,-0.4332608581,-0.2156454623,-0.1834638268,0.0501266681,0.0466087759,-0.430588603,0.0228872839,-0.056356173,0.0701468289,0.0467086211,-0.2373895347,0.1958326846,-0.112088196,0.644041419,0.1405759156,0.3196020126,-0.0674552545,0.0807409436,-0.3065145314,-0.1040730104,0.1997594982,0.2339827865,-0.0820058882,0.2814269662,-0.3691291809,0.0220728666,-0.2654205859,-0.0618536472,0.1581526697,0.0049349377,-0.0418930389,-0.1768803149,0.0234032944,0.1227749065,0.1063580886,0.5370633602,-0.1948029548,0.3811496198,0.0126683507,-0.0843542144,0.2367663532,-0.5975607634,-0.0592407882,-0.1238263398,0.2851480246,-0.1292723119,-0.0741126239,-0.2619338632,-0.0332840607,0.4934832156,-0.200522542,0.0032077192,0.0730269998,-0.0920789316,0.2536216378,-0.1244288683,0.0757653117,0.103560783,-0.0784405246,0.0703940243,-0.2011801451]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2415","title":"Cached dataset not loaded","comments":"Hi ! The hash used for caching `map` results is the fingerprint of the resulting dataset. It is computed using three things:\r\n- the old fingerprint of the dataset\r\n- the hash of the function\r\n- the hash of the other parameters passed to `map`\r\n\r\nYou can compute the hash of your function (or any python object) with\r\n```python\r\nfrom datasets.fingerprint import Hasher\r\n\r\nmy_func = lambda x: x + 1\r\nprint(Hasher.hash(my_func))\r\n```\r\n\r\nIf `prepare_dataset` is always executed, maybe this is because your `processor` has a different hash each time you want to execute it.","body":"## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No","comment_length":94,"text":"Cached dataset not loaded\n## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\nHi ! The hash used for caching `map` results is the fingerprint of the resulting dataset. It is computed using three things:\r\n- the old fingerprint of the dataset\r\n- the hash of the function\r\n- the hash of the other parameters passed to `map`\r\n\r\nYou can compute the hash of your function (or any python object) with\r\n```python\r\nfrom datasets.fingerprint import Hasher\r\n\r\nmy_func = lambda x: x + 1\r\nprint(Hasher.hash(my_func))\r\n```\r\n\r\nIf `prepare_dataset` is always executed, maybe this is because your `processor` has a different hash each time you want to execute it.","embeddings":[-0.2978544831,-0.0423550606,-0.0758303478,0.4236162901,-0.0719461665,-0.0276887827,0.2326412499,0.3540042639,0.2089607567,-0.0697676763,0.0410520583,0.1886659861,-0.2849694192,-0.2252903432,0.1128703952,0.2975529432,0.0528036244,0.1180419326,-0.1090572625,-0.1540792584,-0.0150639126,0.2576181591,-0.0469234958,-0.093652606,-0.2029935271,0.1055824384,0.0262820609,0.1340813786,0.0921806395,-0.2740397155,0.5368282199,0.0920564905,0.1350365132,0.5165358186,-0.0001210673,0.1466437578,0.2859062552,-0.215739131,-0.2181520611,-0.1280438304,-0.2788822949,-0.0832798928,-0.1238182411,-0.2300582379,-0.0120474277,0.0122144446,-0.0481124632,-0.8536034226,0.459944576,0.0661035106,0.1565443277,-0.1820604801,-0.3796681762,0.1259405762,0.038267713,-0.07413809,0.0205731262,0.1491375864,0.1322319508,-0.0945808291,-0.0412483253,0.5044329762,-0.0829141736,-0.0581240319,0.0016792879,0.0538822822,0.2330061644,-0.1952100992,0.1459397674,0.1084957868,0.1963786185,-0.2124014497,-0.3501954079,-0.2727812529,-0.3075908422,-0.2135212123,0.3252165914,-0.1555833817,0.1468669474,0.1651790589,-0.6091327071,-0.0123855583,0.1913139969,-0.0265445951,-0.0197416823,-0.1129705459,-0.168192789,0.1520161331,0.1083762646,-0.0320031121,0.2745790482,-0.1657436788,-0.0053228564,0.4561800063,-0.0696609691,0.2084786743,0.3096645474,0.0638183728,-0.0398250967,0.2631065547,0.1751727313,0.1068111286,-0.0345857143,-0.0644117445,0.1306186318,0.2133267969,0.2404883653,0.2522965968,-0.049673751,-0.1146898791,-0.5343170762,0.0586863309,-0.1254806519,-0.2830507755,0.7806978822,0.1787222922,-0.0620118529,-0.401501447,0.064640604,0.041482646,-0.42515257,-0.3526496589,0.1482111812,0.4227842987,-0.22410357,0.2358968258,-0.0378687978,-0.1386456192,-0.2654689252,0.1546202749,-0.1668414921,-0.2038098127,-0.3884489834,0.1796002686,0.37313658,-0.4005204141,0.2701569796,0.0872643217,0.0604942627,-0.2748971879,0.165197596,-0.1340657026,0.5168511868,0.0657960027,0.0053642825,0.2182988822,0.1590651125,0.0378639959,-0.2242561281,0.1018682271,-0.5072760582,-0.1304601878,0.318675667,0.0764246434,-0.2396333665,0.3016883731,-0.209837392,0.2644234896,0.3779603243,-0.3355059624,-0.1166514009,-0.2069757432,-0.2666144669,-0.4785521328,0.4195537567,0.6090718508,-0.2946233451,-0.3102104068,-0.1382859498,-0.0281040333,-0.0599078014,0.0682887435,-0.2763515711,0.0992670506,-0.3678852618,0.0261640139,0.4927723706,-0.3572143912,-0.4876314104,0.0474483036,0.250859648,0.5972858667,-0.0802121386,-0.1463817358,0.1930467933,-0.2266284823,0.2826514542,0.1132326648,0.0213674568,-0.1429440379,-0.1665978581,-0.1039288044,0.1472658366,-0.1259729266,0.3338222206,0.1621620655,0.1865675449,-0.1245006174,0.1304099411,-0.0470252037,-0.0186559297,0.3194216788,-0.0883403495,-0.0465562865,0.1233384088,-0.2534306645,-0.3030246496,0.4161659777,-0.277041018,-0.5969229937,0.0289568231,0.0484691896,-0.207630232,-0.0645195767,-0.170820117,-0.2434182167,0.0316517092,0.2297227979,0.2284024656,-0.0210998245,-0.0545175895,0.5442660451,0.0102770561,-0.2740502059,-0.2433663309,0.1692317128,0.2576344907,-0.1716133356,-0.3055517972,0.0876919404,0.3064357936,-0.1491903067,-0.1570721865,0.1695891917,0.0594948232,0.2013689876,-0.2249587923,0.2929288149,-0.1600542963,-0.1222959012,0.1584830284,0.0284264144,-0.0416131653,-0.1636959612,-0.0239242911,0.3278394341,0.1689307094,0.0020596988,0.0516499132,-0.18602705,0.1895095557,-0.1054123193,-0.117535077,-0.2751552165,0.1547856033,-0.1886118948,0.4881473184,0.2394631356,0.0860364437,0.1719860733,0.4886368513,0.2141722739,0.1340205371,-0.0534790717,-0.3166265786,-0.1791449189,-0.1831711382,0.0051268237,0.162058875,-0.000577304,-0.0210616924,-0.023351742,0.0179936346,0.2533574998,0.0779890493,-0.2341278642,0.3470124602,0.1444575787,0.119849056,0.0359204039,-0.1744266897,0.0458373725,0.1211205721,-0.0410864949,-0.466869086,0.1364646703,-0.3292939067,0.2719795704,-0.2452030629,-0.0638307258,0.0286573898,-0.2246178985,-0.1919744313,0.3194694519,0.2880186141,0.2144105434,-0.3410937786,0.1146627292,-0.1894728988,-0.1603946239,-0.0100804102,-0.1131348908,-0.3489112854,-0.0086926008,0.1438347846,-0.2055016458,0.2075018138,-0.0255988836,-0.1401118487,-0.2677106261,-0.3595719039,0.0935123265,0.2070462853,0.3455573022,-0.2412908524,-0.0900674835,-0.20699431,-0.0637390614,0.1708731949,-0.2313913256,-0.1428841799,-0.0684183836,-0.0057264441,0.0865047798,0.0938635096,-0.4032764435,-0.3004500568,0.0697977245,-0.1072492748,-0.0847632661,0.0376866087,0.0733904243,-0.1441195458,-0.2160439491,-0.0327869318,-0.0495372526,-0.4220378101,-0.4178791642,0.2509204149,-0.2321704179,0.0369442664,-0.1274863183,-0.1213700399,-0.0014036332,0.5802029371,-0.3909652531,0.1654570103,-0.0096159568,0.0502640828,0.000515733,-0.1375783086,0.444417119,0.1967123896,0.1541717947,-0.2419928163,-0.411700666,0.0349578336,-0.0705297217,0.2873363793,0.1238329336,0.1701889038,-0.0201322362,1.0965393782,0.3184102774,-0.2103504837,0.0661234036,-0.0407362431,0.2642996013,-0.0109162964,-0.3952404857,-0.1654396057,-0.1520673931,-0.1927796453,0.0578009188,0.063286595,-0.308313489,-0.0400371552,0.2140312791,-0.7177140713,-0.208044365,0.2859528661,-0.6640021205,0.2785124779,0.1591979861,0.1727760732,-0.5495753884,0.068691723,0.082135193,-0.1490960866,0.5069204569,0.0073018586,-0.1675949842,0.1631207913,0.0367319956,0.1653803885,0.2624444664,0.2216524333,-0.123574324,-0.0967764705,-0.2273300737,-0.0536234789,0.4456322789,-0.602388978,0.2966638803,0.3834201992,-0.3278230131,-0.1631358862,-0.2292153388,0.1275748312,0.3774194717,0.1738161296,0.2964749932,-0.0295395795,0.162295267,-0.0784908682,0.1458637416,-0.2814039886,-0.0492803566,-0.0455858894,0.135243699,0.1419732124,-0.0365296789,-0.2272629142,-0.0217074193,-0.0517015569,-0.053078834,-0.4242401421,0.0739685223,0.2555040121,0.3246227205,0.5262849927,0.0624686703,0.4390742183,0.2741711736,0.0351012796,-0.0851440579,0.5787379742,-0.2023314238,-0.1998823136,0.009701441,0.2453024536,-0.1176062599,0.0950476453,-0.385099113,-0.1949116141,0.1685321629,-0.1146743968,-0.1057470292,-0.1660600156,0.055524677,0.2818845212,-0.5762344599,-0.6871487498,0.4059185982,0.4733220041,-0.4103278518,0.7267813087,-0.5869186521,-0.2629710436,0.2095643282,-0.1846629679,0.7170375586,-0.2186813653,-0.0995574594,-0.0923304334,0.3079008162,-0.0393021181,-0.172490254,-0.0400164574,-0.2709791064,-0.1853233725,-0.1254699975,-0.2476443499,0.084847413,0.2775765359,0.0025101651,0.3366933763,-0.2414913028,0.4807322919,0.0184305776,-0.0930507928,0.1714968532,-0.060281191,-0.3047181368,0.1155695915,0.0777915493,0.1758089811,-0.0594418831,-0.0584447756,-0.1147260442,-0.0674225315,-0.3334651887,0.2548058629,-0.6193088293,0.4070167542,-0.1185960695,-0.1381233037,-0.1180677861,0.1156828254,0.1650652289,0.0837796777,-0.0401754305,0.2410151958,0.326459676,0.1791501194,0.0417668931,-0.1866993457,0.1389885992,0.0957636237,-0.1071276888,0.1587900221,-0.1544303447,-0.375262022,-0.176004231,0.1844714731,0.0091225058,-0.014729579,-0.2240322977,0.2788906991,0.2617226243,0.115472585,0.0510925129,0.2583168447,0.1049676538,0.3374639452,-0.1267133802,-0.2926823497,0.0319456309,0.546869874,0.1948584914,-0.1740589887,0.2723882496,-0.1755636781,-0.0421947949,-0.0335020795,-0.3243037164,0.104490377,-0.2906941175,0.0863463804,-0.3525417745,0.1369680464,0.2956292927,-0.151681602,0.1772856414,0.2408417314,-0.1096671373,-0.3795477748,-0.0640935153,-0.0311764106,-0.0599763021,0.0721737891,0.1183805019,-0.297593236,-0.1280510724,0.331507951,-0.2037172467,0.1426204294,0.138836503,0.0612695515,-0.0742919669,0.2138422877,0.2881692052,-0.2710391283,-0.0206499267,0.0642175302,-0.3246044219,-0.0957717597,-0.1054293364,0.1195218414,-0.2374317795,0.1265368909,-0.0961164013,-0.0452616774,0.2535043955,-0.2583850324,0.1769426316,0.4429425895,0.2472510934,-0.1213730574,0.0362662934,-0.2371957153,-0.1984448284,0.2644020021,-0.1955887228,-0.0001742452,0.0041380827,0.3197752535,0.0020018611,-0.0661249831,-0.1795742959,-0.1680130959,-0.0040729488,0.1874286681,0.3716533184,-0.1524595618,0.1358595341,0.135865286,0.229644686,0.2164588422,-0.1311828345,0.0259596594,0.00648139,0.1072962359,-0.0830922425,0.0035993913,0.3172124028,0.1602545381,-0.1353239864,0.278639257,0.1040155143,-0.0837675557,0.2344512939,0.0386339016,0.3939882517,-0.3727237284,0.3957624435,0.2192043662,0.2516993582,0.1452663988,0.1796540171,0.1162650362,0.0013625554,0.256772697,0.178279236,0.3827572465,0.2070094794,0.0789957568,-0.1574572772,-0.5870573521,-0.0352019109,0.3819710314,-0.3661758602,0.501277566,0.0281729121,0.058995869,0.1634183675,-0.3162510097,-0.3057076931,0.3864521682,-0.0145207196,-0.2368498594,0.1871301383,-0.2024393827,0.0762404501,0.0389943346,-0.007976192,0.0427843966,0.5030100942,0.1696486175,-0.221239239,-0.0135474857,-0.18846865,0.1144961566,0.4267623127,-0.1168560162,0.1467285603,-0.0639424175,-0.1349025369,0.0471755788,0.3376615345,0.4609377384,0.2616270781,0.1134784073,0.0790454522,0.4846385121,0.1015991122,-0.2699878216,0.3003548682,-0.0987127572,-0.2325969636,0.151421845,0.1502105594,-0.0925896019,0.2429991663,-0.2007202357,0.3491633534,-0.2665272057,0.0634320751,-0.1339989007,0.2627901137,-0.0019667032,0.1811152548,-0.049006585,-0.1975012273,0.7122183442,0.1442073137,0.2382027507,0.0097113065,0.0515268631,0.2391056567,0.1620854586,0.3663572967,0.058446452,-0.1798638701,-0.0593992993,-0.6679871082,0.0174868777,-0.3761491776,0.2178829014,-0.0890114754,0.2133032382,0.2604070902,-0.0863466859,0.2404281944,0.001278261,0.1334682405,0.2444926947,-0.4332608581,-0.2156454623,-0.1834638268,0.0501266681,0.0466087759,-0.430588603,0.0228872839,-0.056356173,0.0701468289,0.0467086211,-0.2373895347,0.1958326846,-0.112088196,0.644041419,0.1405759156,0.3196020126,-0.0674552545,0.0807409436,-0.3065145314,-0.1040730104,0.1997594982,0.2339827865,-0.0820058882,0.2814269662,-0.3691291809,0.0220728666,-0.2654205859,-0.0618536472,0.1581526697,0.0049349377,-0.0418930389,-0.1768803149,0.0234032944,0.1227749065,0.1063580886,0.5370633602,-0.1948029548,0.3811496198,0.0126683507,-0.0843542144,0.2367663532,-0.5975607634,-0.0592407882,-0.1238263398,0.2851480246,-0.1292723119,-0.0741126239,-0.2619338632,-0.0332840607,0.4934832156,-0.200522542,0.0032077192,0.0730269998,-0.0920789316,0.2536216378,-0.1244288683,0.0757653117,0.103560783,-0.0784405246,0.0703940243,-0.2011801451]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2415","title":"Cached dataset not loaded","comments":"> If `prepare_dataset` is always executed, maybe this is because your `processor` has a different hash each time you want to execute it.\r\n\r\nYes I\u00a0think that was the issue.\r\n\r\nFor the hash of the function:\r\n* does it consider just the name or the actual code of the function\r\n* does it consider variables that are not passed explicitly as parameters to the functions (such as the processor here)","body":"## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No","comment_length":70,"text":"Cached dataset not loaded\n## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\n> If `prepare_dataset` is always executed, maybe this is because your `processor` has a different hash each time you want to execute it.\r\n\r\nYes I\u00a0think that was the issue.\r\n\r\nFor the hash of the function:\r\n* does it consider just the name or the actual code of the function\r\n* does it consider variables that are not passed explicitly as parameters to the functions (such as the processor here)","embeddings":[-0.2978544831,-0.0423550606,-0.0758303478,0.4236162901,-0.0719461665,-0.0276887827,0.2326412499,0.3540042639,0.2089607567,-0.0697676763,0.0410520583,0.1886659861,-0.2849694192,-0.2252903432,0.1128703952,0.2975529432,0.0528036244,0.1180419326,-0.1090572625,-0.1540792584,-0.0150639126,0.2576181591,-0.0469234958,-0.093652606,-0.2029935271,0.1055824384,0.0262820609,0.1340813786,0.0921806395,-0.2740397155,0.5368282199,0.0920564905,0.1350365132,0.5165358186,-0.0001210673,0.1466437578,0.2859062552,-0.215739131,-0.2181520611,-0.1280438304,-0.2788822949,-0.0832798928,-0.1238182411,-0.2300582379,-0.0120474277,0.0122144446,-0.0481124632,-0.8536034226,0.459944576,0.0661035106,0.1565443277,-0.1820604801,-0.3796681762,0.1259405762,0.038267713,-0.07413809,0.0205731262,0.1491375864,0.1322319508,-0.0945808291,-0.0412483253,0.5044329762,-0.0829141736,-0.0581240319,0.0016792879,0.0538822822,0.2330061644,-0.1952100992,0.1459397674,0.1084957868,0.1963786185,-0.2124014497,-0.3501954079,-0.2727812529,-0.3075908422,-0.2135212123,0.3252165914,-0.1555833817,0.1468669474,0.1651790589,-0.6091327071,-0.0123855583,0.1913139969,-0.0265445951,-0.0197416823,-0.1129705459,-0.168192789,0.1520161331,0.1083762646,-0.0320031121,0.2745790482,-0.1657436788,-0.0053228564,0.4561800063,-0.0696609691,0.2084786743,0.3096645474,0.0638183728,-0.0398250967,0.2631065547,0.1751727313,0.1068111286,-0.0345857143,-0.0644117445,0.1306186318,0.2133267969,0.2404883653,0.2522965968,-0.049673751,-0.1146898791,-0.5343170762,0.0586863309,-0.1254806519,-0.2830507755,0.7806978822,0.1787222922,-0.0620118529,-0.401501447,0.064640604,0.041482646,-0.42515257,-0.3526496589,0.1482111812,0.4227842987,-0.22410357,0.2358968258,-0.0378687978,-0.1386456192,-0.2654689252,0.1546202749,-0.1668414921,-0.2038098127,-0.3884489834,0.1796002686,0.37313658,-0.4005204141,0.2701569796,0.0872643217,0.0604942627,-0.2748971879,0.165197596,-0.1340657026,0.5168511868,0.0657960027,0.0053642825,0.2182988822,0.1590651125,0.0378639959,-0.2242561281,0.1018682271,-0.5072760582,-0.1304601878,0.318675667,0.0764246434,-0.2396333665,0.3016883731,-0.209837392,0.2644234896,0.3779603243,-0.3355059624,-0.1166514009,-0.2069757432,-0.2666144669,-0.4785521328,0.4195537567,0.6090718508,-0.2946233451,-0.3102104068,-0.1382859498,-0.0281040333,-0.0599078014,0.0682887435,-0.2763515711,0.0992670506,-0.3678852618,0.0261640139,0.4927723706,-0.3572143912,-0.4876314104,0.0474483036,0.250859648,0.5972858667,-0.0802121386,-0.1463817358,0.1930467933,-0.2266284823,0.2826514542,0.1132326648,0.0213674568,-0.1429440379,-0.1665978581,-0.1039288044,0.1472658366,-0.1259729266,0.3338222206,0.1621620655,0.1865675449,-0.1245006174,0.1304099411,-0.0470252037,-0.0186559297,0.3194216788,-0.0883403495,-0.0465562865,0.1233384088,-0.2534306645,-0.3030246496,0.4161659777,-0.277041018,-0.5969229937,0.0289568231,0.0484691896,-0.207630232,-0.0645195767,-0.170820117,-0.2434182167,0.0316517092,0.2297227979,0.2284024656,-0.0210998245,-0.0545175895,0.5442660451,0.0102770561,-0.2740502059,-0.2433663309,0.1692317128,0.2576344907,-0.1716133356,-0.3055517972,0.0876919404,0.3064357936,-0.1491903067,-0.1570721865,0.1695891917,0.0594948232,0.2013689876,-0.2249587923,0.2929288149,-0.1600542963,-0.1222959012,0.1584830284,0.0284264144,-0.0416131653,-0.1636959612,-0.0239242911,0.3278394341,0.1689307094,0.0020596988,0.0516499132,-0.18602705,0.1895095557,-0.1054123193,-0.117535077,-0.2751552165,0.1547856033,-0.1886118948,0.4881473184,0.2394631356,0.0860364437,0.1719860733,0.4886368513,0.2141722739,0.1340205371,-0.0534790717,-0.3166265786,-0.1791449189,-0.1831711382,0.0051268237,0.162058875,-0.000577304,-0.0210616924,-0.023351742,0.0179936346,0.2533574998,0.0779890493,-0.2341278642,0.3470124602,0.1444575787,0.119849056,0.0359204039,-0.1744266897,0.0458373725,0.1211205721,-0.0410864949,-0.466869086,0.1364646703,-0.3292939067,0.2719795704,-0.2452030629,-0.0638307258,0.0286573898,-0.2246178985,-0.1919744313,0.3194694519,0.2880186141,0.2144105434,-0.3410937786,0.1146627292,-0.1894728988,-0.1603946239,-0.0100804102,-0.1131348908,-0.3489112854,-0.0086926008,0.1438347846,-0.2055016458,0.2075018138,-0.0255988836,-0.1401118487,-0.2677106261,-0.3595719039,0.0935123265,0.2070462853,0.3455573022,-0.2412908524,-0.0900674835,-0.20699431,-0.0637390614,0.1708731949,-0.2313913256,-0.1428841799,-0.0684183836,-0.0057264441,0.0865047798,0.0938635096,-0.4032764435,-0.3004500568,0.0697977245,-0.1072492748,-0.0847632661,0.0376866087,0.0733904243,-0.1441195458,-0.2160439491,-0.0327869318,-0.0495372526,-0.4220378101,-0.4178791642,0.2509204149,-0.2321704179,0.0369442664,-0.1274863183,-0.1213700399,-0.0014036332,0.5802029371,-0.3909652531,0.1654570103,-0.0096159568,0.0502640828,0.000515733,-0.1375783086,0.444417119,0.1967123896,0.1541717947,-0.2419928163,-0.411700666,0.0349578336,-0.0705297217,0.2873363793,0.1238329336,0.1701889038,-0.0201322362,1.0965393782,0.3184102774,-0.2103504837,0.0661234036,-0.0407362431,0.2642996013,-0.0109162964,-0.3952404857,-0.1654396057,-0.1520673931,-0.1927796453,0.0578009188,0.063286595,-0.308313489,-0.0400371552,0.2140312791,-0.7177140713,-0.208044365,0.2859528661,-0.6640021205,0.2785124779,0.1591979861,0.1727760732,-0.5495753884,0.068691723,0.082135193,-0.1490960866,0.5069204569,0.0073018586,-0.1675949842,0.1631207913,0.0367319956,0.1653803885,0.2624444664,0.2216524333,-0.123574324,-0.0967764705,-0.2273300737,-0.0536234789,0.4456322789,-0.602388978,0.2966638803,0.3834201992,-0.3278230131,-0.1631358862,-0.2292153388,0.1275748312,0.3774194717,0.1738161296,0.2964749932,-0.0295395795,0.162295267,-0.0784908682,0.1458637416,-0.2814039886,-0.0492803566,-0.0455858894,0.135243699,0.1419732124,-0.0365296789,-0.2272629142,-0.0217074193,-0.0517015569,-0.053078834,-0.4242401421,0.0739685223,0.2555040121,0.3246227205,0.5262849927,0.0624686703,0.4390742183,0.2741711736,0.0351012796,-0.0851440579,0.5787379742,-0.2023314238,-0.1998823136,0.009701441,0.2453024536,-0.1176062599,0.0950476453,-0.385099113,-0.1949116141,0.1685321629,-0.1146743968,-0.1057470292,-0.1660600156,0.055524677,0.2818845212,-0.5762344599,-0.6871487498,0.4059185982,0.4733220041,-0.4103278518,0.7267813087,-0.5869186521,-0.2629710436,0.2095643282,-0.1846629679,0.7170375586,-0.2186813653,-0.0995574594,-0.0923304334,0.3079008162,-0.0393021181,-0.172490254,-0.0400164574,-0.2709791064,-0.1853233725,-0.1254699975,-0.2476443499,0.084847413,0.2775765359,0.0025101651,0.3366933763,-0.2414913028,0.4807322919,0.0184305776,-0.0930507928,0.1714968532,-0.060281191,-0.3047181368,0.1155695915,0.0777915493,0.1758089811,-0.0594418831,-0.0584447756,-0.1147260442,-0.0674225315,-0.3334651887,0.2548058629,-0.6193088293,0.4070167542,-0.1185960695,-0.1381233037,-0.1180677861,0.1156828254,0.1650652289,0.0837796777,-0.0401754305,0.2410151958,0.326459676,0.1791501194,0.0417668931,-0.1866993457,0.1389885992,0.0957636237,-0.1071276888,0.1587900221,-0.1544303447,-0.375262022,-0.176004231,0.1844714731,0.0091225058,-0.014729579,-0.2240322977,0.2788906991,0.2617226243,0.115472585,0.0510925129,0.2583168447,0.1049676538,0.3374639452,-0.1267133802,-0.2926823497,0.0319456309,0.546869874,0.1948584914,-0.1740589887,0.2723882496,-0.1755636781,-0.0421947949,-0.0335020795,-0.3243037164,0.104490377,-0.2906941175,0.0863463804,-0.3525417745,0.1369680464,0.2956292927,-0.151681602,0.1772856414,0.2408417314,-0.1096671373,-0.3795477748,-0.0640935153,-0.0311764106,-0.0599763021,0.0721737891,0.1183805019,-0.297593236,-0.1280510724,0.331507951,-0.2037172467,0.1426204294,0.138836503,0.0612695515,-0.0742919669,0.2138422877,0.2881692052,-0.2710391283,-0.0206499267,0.0642175302,-0.3246044219,-0.0957717597,-0.1054293364,0.1195218414,-0.2374317795,0.1265368909,-0.0961164013,-0.0452616774,0.2535043955,-0.2583850324,0.1769426316,0.4429425895,0.2472510934,-0.1213730574,0.0362662934,-0.2371957153,-0.1984448284,0.2644020021,-0.1955887228,-0.0001742452,0.0041380827,0.3197752535,0.0020018611,-0.0661249831,-0.1795742959,-0.1680130959,-0.0040729488,0.1874286681,0.3716533184,-0.1524595618,0.1358595341,0.135865286,0.229644686,0.2164588422,-0.1311828345,0.0259596594,0.00648139,0.1072962359,-0.0830922425,0.0035993913,0.3172124028,0.1602545381,-0.1353239864,0.278639257,0.1040155143,-0.0837675557,0.2344512939,0.0386339016,0.3939882517,-0.3727237284,0.3957624435,0.2192043662,0.2516993582,0.1452663988,0.1796540171,0.1162650362,0.0013625554,0.256772697,0.178279236,0.3827572465,0.2070094794,0.0789957568,-0.1574572772,-0.5870573521,-0.0352019109,0.3819710314,-0.3661758602,0.501277566,0.0281729121,0.058995869,0.1634183675,-0.3162510097,-0.3057076931,0.3864521682,-0.0145207196,-0.2368498594,0.1871301383,-0.2024393827,0.0762404501,0.0389943346,-0.007976192,0.0427843966,0.5030100942,0.1696486175,-0.221239239,-0.0135474857,-0.18846865,0.1144961566,0.4267623127,-0.1168560162,0.1467285603,-0.0639424175,-0.1349025369,0.0471755788,0.3376615345,0.4609377384,0.2616270781,0.1134784073,0.0790454522,0.4846385121,0.1015991122,-0.2699878216,0.3003548682,-0.0987127572,-0.2325969636,0.151421845,0.1502105594,-0.0925896019,0.2429991663,-0.2007202357,0.3491633534,-0.2665272057,0.0634320751,-0.1339989007,0.2627901137,-0.0019667032,0.1811152548,-0.049006585,-0.1975012273,0.7122183442,0.1442073137,0.2382027507,0.0097113065,0.0515268631,0.2391056567,0.1620854586,0.3663572967,0.058446452,-0.1798638701,-0.0593992993,-0.6679871082,0.0174868777,-0.3761491776,0.2178829014,-0.0890114754,0.2133032382,0.2604070902,-0.0863466859,0.2404281944,0.001278261,0.1334682405,0.2444926947,-0.4332608581,-0.2156454623,-0.1834638268,0.0501266681,0.0466087759,-0.430588603,0.0228872839,-0.056356173,0.0701468289,0.0467086211,-0.2373895347,0.1958326846,-0.112088196,0.644041419,0.1405759156,0.3196020126,-0.0674552545,0.0807409436,-0.3065145314,-0.1040730104,0.1997594982,0.2339827865,-0.0820058882,0.2814269662,-0.3691291809,0.0220728666,-0.2654205859,-0.0618536472,0.1581526697,0.0049349377,-0.0418930389,-0.1768803149,0.0234032944,0.1227749065,0.1063580886,0.5370633602,-0.1948029548,0.3811496198,0.0126683507,-0.0843542144,0.2367663532,-0.5975607634,-0.0592407882,-0.1238263398,0.2851480246,-0.1292723119,-0.0741126239,-0.2619338632,-0.0332840607,0.4934832156,-0.200522542,0.0032077192,0.0730269998,-0.0920789316,0.2536216378,-0.1244288683,0.0757653117,0.103560783,-0.0784405246,0.0703940243,-0.2011801451]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2415","title":"Cached dataset not loaded","comments":"> does it consider just the name or the actual code of the function\r\n\r\nIt looks at the name and the actual code and all variables such as recursively. It uses `dill` to do so, which is based on `pickle`.\r\nBasically the hash is computed using the pickle bytes of your function (computed using `dill` to support most python objects).\r\n\r\n> does it consider variables that are not passed explicitly as parameters to the functions (such as the processor here)\r\n\r\nYes it does thanks to recursive pickling.","body":"## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No","comment_length":87,"text":"Cached dataset not loaded\n## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\n> does it consider just the name or the actual code of the function\r\n\r\nIt looks at the name and the actual code and all variables such as recursively. It uses `dill` to do so, which is based on `pickle`.\r\nBasically the hash is computed using the pickle bytes of your function (computed using `dill` to support most python objects).\r\n\r\n> does it consider variables that are not passed explicitly as parameters to the functions (such as the processor here)\r\n\r\nYes it does thanks to recursive pickling.","embeddings":[-0.2978544831,-0.0423550606,-0.0758303478,0.4236162901,-0.0719461665,-0.0276887827,0.2326412499,0.3540042639,0.2089607567,-0.0697676763,0.0410520583,0.1886659861,-0.2849694192,-0.2252903432,0.1128703952,0.2975529432,0.0528036244,0.1180419326,-0.1090572625,-0.1540792584,-0.0150639126,0.2576181591,-0.0469234958,-0.093652606,-0.2029935271,0.1055824384,0.0262820609,0.1340813786,0.0921806395,-0.2740397155,0.5368282199,0.0920564905,0.1350365132,0.5165358186,-0.0001210673,0.1466437578,0.2859062552,-0.215739131,-0.2181520611,-0.1280438304,-0.2788822949,-0.0832798928,-0.1238182411,-0.2300582379,-0.0120474277,0.0122144446,-0.0481124632,-0.8536034226,0.459944576,0.0661035106,0.1565443277,-0.1820604801,-0.3796681762,0.1259405762,0.038267713,-0.07413809,0.0205731262,0.1491375864,0.1322319508,-0.0945808291,-0.0412483253,0.5044329762,-0.0829141736,-0.0581240319,0.0016792879,0.0538822822,0.2330061644,-0.1952100992,0.1459397674,0.1084957868,0.1963786185,-0.2124014497,-0.3501954079,-0.2727812529,-0.3075908422,-0.2135212123,0.3252165914,-0.1555833817,0.1468669474,0.1651790589,-0.6091327071,-0.0123855583,0.1913139969,-0.0265445951,-0.0197416823,-0.1129705459,-0.168192789,0.1520161331,0.1083762646,-0.0320031121,0.2745790482,-0.1657436788,-0.0053228564,0.4561800063,-0.0696609691,0.2084786743,0.3096645474,0.0638183728,-0.0398250967,0.2631065547,0.1751727313,0.1068111286,-0.0345857143,-0.0644117445,0.1306186318,0.2133267969,0.2404883653,0.2522965968,-0.049673751,-0.1146898791,-0.5343170762,0.0586863309,-0.1254806519,-0.2830507755,0.7806978822,0.1787222922,-0.0620118529,-0.401501447,0.064640604,0.041482646,-0.42515257,-0.3526496589,0.1482111812,0.4227842987,-0.22410357,0.2358968258,-0.0378687978,-0.1386456192,-0.2654689252,0.1546202749,-0.1668414921,-0.2038098127,-0.3884489834,0.1796002686,0.37313658,-0.4005204141,0.2701569796,0.0872643217,0.0604942627,-0.2748971879,0.165197596,-0.1340657026,0.5168511868,0.0657960027,0.0053642825,0.2182988822,0.1590651125,0.0378639959,-0.2242561281,0.1018682271,-0.5072760582,-0.1304601878,0.318675667,0.0764246434,-0.2396333665,0.3016883731,-0.209837392,0.2644234896,0.3779603243,-0.3355059624,-0.1166514009,-0.2069757432,-0.2666144669,-0.4785521328,0.4195537567,0.6090718508,-0.2946233451,-0.3102104068,-0.1382859498,-0.0281040333,-0.0599078014,0.0682887435,-0.2763515711,0.0992670506,-0.3678852618,0.0261640139,0.4927723706,-0.3572143912,-0.4876314104,0.0474483036,0.250859648,0.5972858667,-0.0802121386,-0.1463817358,0.1930467933,-0.2266284823,0.2826514542,0.1132326648,0.0213674568,-0.1429440379,-0.1665978581,-0.1039288044,0.1472658366,-0.1259729266,0.3338222206,0.1621620655,0.1865675449,-0.1245006174,0.1304099411,-0.0470252037,-0.0186559297,0.3194216788,-0.0883403495,-0.0465562865,0.1233384088,-0.2534306645,-0.3030246496,0.4161659777,-0.277041018,-0.5969229937,0.0289568231,0.0484691896,-0.207630232,-0.0645195767,-0.170820117,-0.2434182167,0.0316517092,0.2297227979,0.2284024656,-0.0210998245,-0.0545175895,0.5442660451,0.0102770561,-0.2740502059,-0.2433663309,0.1692317128,0.2576344907,-0.1716133356,-0.3055517972,0.0876919404,0.3064357936,-0.1491903067,-0.1570721865,0.1695891917,0.0594948232,0.2013689876,-0.2249587923,0.2929288149,-0.1600542963,-0.1222959012,0.1584830284,0.0284264144,-0.0416131653,-0.1636959612,-0.0239242911,0.3278394341,0.1689307094,0.0020596988,0.0516499132,-0.18602705,0.1895095557,-0.1054123193,-0.117535077,-0.2751552165,0.1547856033,-0.1886118948,0.4881473184,0.2394631356,0.0860364437,0.1719860733,0.4886368513,0.2141722739,0.1340205371,-0.0534790717,-0.3166265786,-0.1791449189,-0.1831711382,0.0051268237,0.162058875,-0.000577304,-0.0210616924,-0.023351742,0.0179936346,0.2533574998,0.0779890493,-0.2341278642,0.3470124602,0.1444575787,0.119849056,0.0359204039,-0.1744266897,0.0458373725,0.1211205721,-0.0410864949,-0.466869086,0.1364646703,-0.3292939067,0.2719795704,-0.2452030629,-0.0638307258,0.0286573898,-0.2246178985,-0.1919744313,0.3194694519,0.2880186141,0.2144105434,-0.3410937786,0.1146627292,-0.1894728988,-0.1603946239,-0.0100804102,-0.1131348908,-0.3489112854,-0.0086926008,0.1438347846,-0.2055016458,0.2075018138,-0.0255988836,-0.1401118487,-0.2677106261,-0.3595719039,0.0935123265,0.2070462853,0.3455573022,-0.2412908524,-0.0900674835,-0.20699431,-0.0637390614,0.1708731949,-0.2313913256,-0.1428841799,-0.0684183836,-0.0057264441,0.0865047798,0.0938635096,-0.4032764435,-0.3004500568,0.0697977245,-0.1072492748,-0.0847632661,0.0376866087,0.0733904243,-0.1441195458,-0.2160439491,-0.0327869318,-0.0495372526,-0.4220378101,-0.4178791642,0.2509204149,-0.2321704179,0.0369442664,-0.1274863183,-0.1213700399,-0.0014036332,0.5802029371,-0.3909652531,0.1654570103,-0.0096159568,0.0502640828,0.000515733,-0.1375783086,0.444417119,0.1967123896,0.1541717947,-0.2419928163,-0.411700666,0.0349578336,-0.0705297217,0.2873363793,0.1238329336,0.1701889038,-0.0201322362,1.0965393782,0.3184102774,-0.2103504837,0.0661234036,-0.0407362431,0.2642996013,-0.0109162964,-0.3952404857,-0.1654396057,-0.1520673931,-0.1927796453,0.0578009188,0.063286595,-0.308313489,-0.0400371552,0.2140312791,-0.7177140713,-0.208044365,0.2859528661,-0.6640021205,0.2785124779,0.1591979861,0.1727760732,-0.5495753884,0.068691723,0.082135193,-0.1490960866,0.5069204569,0.0073018586,-0.1675949842,0.1631207913,0.0367319956,0.1653803885,0.2624444664,0.2216524333,-0.123574324,-0.0967764705,-0.2273300737,-0.0536234789,0.4456322789,-0.602388978,0.2966638803,0.3834201992,-0.3278230131,-0.1631358862,-0.2292153388,0.1275748312,0.3774194717,0.1738161296,0.2964749932,-0.0295395795,0.162295267,-0.0784908682,0.1458637416,-0.2814039886,-0.0492803566,-0.0455858894,0.135243699,0.1419732124,-0.0365296789,-0.2272629142,-0.0217074193,-0.0517015569,-0.053078834,-0.4242401421,0.0739685223,0.2555040121,0.3246227205,0.5262849927,0.0624686703,0.4390742183,0.2741711736,0.0351012796,-0.0851440579,0.5787379742,-0.2023314238,-0.1998823136,0.009701441,0.2453024536,-0.1176062599,0.0950476453,-0.385099113,-0.1949116141,0.1685321629,-0.1146743968,-0.1057470292,-0.1660600156,0.055524677,0.2818845212,-0.5762344599,-0.6871487498,0.4059185982,0.4733220041,-0.4103278518,0.7267813087,-0.5869186521,-0.2629710436,0.2095643282,-0.1846629679,0.7170375586,-0.2186813653,-0.0995574594,-0.0923304334,0.3079008162,-0.0393021181,-0.172490254,-0.0400164574,-0.2709791064,-0.1853233725,-0.1254699975,-0.2476443499,0.084847413,0.2775765359,0.0025101651,0.3366933763,-0.2414913028,0.4807322919,0.0184305776,-0.0930507928,0.1714968532,-0.060281191,-0.3047181368,0.1155695915,0.0777915493,0.1758089811,-0.0594418831,-0.0584447756,-0.1147260442,-0.0674225315,-0.3334651887,0.2548058629,-0.6193088293,0.4070167542,-0.1185960695,-0.1381233037,-0.1180677861,0.1156828254,0.1650652289,0.0837796777,-0.0401754305,0.2410151958,0.326459676,0.1791501194,0.0417668931,-0.1866993457,0.1389885992,0.0957636237,-0.1071276888,0.1587900221,-0.1544303447,-0.375262022,-0.176004231,0.1844714731,0.0091225058,-0.014729579,-0.2240322977,0.2788906991,0.2617226243,0.115472585,0.0510925129,0.2583168447,0.1049676538,0.3374639452,-0.1267133802,-0.2926823497,0.0319456309,0.546869874,0.1948584914,-0.1740589887,0.2723882496,-0.1755636781,-0.0421947949,-0.0335020795,-0.3243037164,0.104490377,-0.2906941175,0.0863463804,-0.3525417745,0.1369680464,0.2956292927,-0.151681602,0.1772856414,0.2408417314,-0.1096671373,-0.3795477748,-0.0640935153,-0.0311764106,-0.0599763021,0.0721737891,0.1183805019,-0.297593236,-0.1280510724,0.331507951,-0.2037172467,0.1426204294,0.138836503,0.0612695515,-0.0742919669,0.2138422877,0.2881692052,-0.2710391283,-0.0206499267,0.0642175302,-0.3246044219,-0.0957717597,-0.1054293364,0.1195218414,-0.2374317795,0.1265368909,-0.0961164013,-0.0452616774,0.2535043955,-0.2583850324,0.1769426316,0.4429425895,0.2472510934,-0.1213730574,0.0362662934,-0.2371957153,-0.1984448284,0.2644020021,-0.1955887228,-0.0001742452,0.0041380827,0.3197752535,0.0020018611,-0.0661249831,-0.1795742959,-0.1680130959,-0.0040729488,0.1874286681,0.3716533184,-0.1524595618,0.1358595341,0.135865286,0.229644686,0.2164588422,-0.1311828345,0.0259596594,0.00648139,0.1072962359,-0.0830922425,0.0035993913,0.3172124028,0.1602545381,-0.1353239864,0.278639257,0.1040155143,-0.0837675557,0.2344512939,0.0386339016,0.3939882517,-0.3727237284,0.3957624435,0.2192043662,0.2516993582,0.1452663988,0.1796540171,0.1162650362,0.0013625554,0.256772697,0.178279236,0.3827572465,0.2070094794,0.0789957568,-0.1574572772,-0.5870573521,-0.0352019109,0.3819710314,-0.3661758602,0.501277566,0.0281729121,0.058995869,0.1634183675,-0.3162510097,-0.3057076931,0.3864521682,-0.0145207196,-0.2368498594,0.1871301383,-0.2024393827,0.0762404501,0.0389943346,-0.007976192,0.0427843966,0.5030100942,0.1696486175,-0.221239239,-0.0135474857,-0.18846865,0.1144961566,0.4267623127,-0.1168560162,0.1467285603,-0.0639424175,-0.1349025369,0.0471755788,0.3376615345,0.4609377384,0.2616270781,0.1134784073,0.0790454522,0.4846385121,0.1015991122,-0.2699878216,0.3003548682,-0.0987127572,-0.2325969636,0.151421845,0.1502105594,-0.0925896019,0.2429991663,-0.2007202357,0.3491633534,-0.2665272057,0.0634320751,-0.1339989007,0.2627901137,-0.0019667032,0.1811152548,-0.049006585,-0.1975012273,0.7122183442,0.1442073137,0.2382027507,0.0097113065,0.0515268631,0.2391056567,0.1620854586,0.3663572967,0.058446452,-0.1798638701,-0.0593992993,-0.6679871082,0.0174868777,-0.3761491776,0.2178829014,-0.0890114754,0.2133032382,0.2604070902,-0.0863466859,0.2404281944,0.001278261,0.1334682405,0.2444926947,-0.4332608581,-0.2156454623,-0.1834638268,0.0501266681,0.0466087759,-0.430588603,0.0228872839,-0.056356173,0.0701468289,0.0467086211,-0.2373895347,0.1958326846,-0.112088196,0.644041419,0.1405759156,0.3196020126,-0.0674552545,0.0807409436,-0.3065145314,-0.1040730104,0.1997594982,0.2339827865,-0.0820058882,0.2814269662,-0.3691291809,0.0220728666,-0.2654205859,-0.0618536472,0.1581526697,0.0049349377,-0.0418930389,-0.1768803149,0.0234032944,0.1227749065,0.1063580886,0.5370633602,-0.1948029548,0.3811496198,0.0126683507,-0.0843542144,0.2367663532,-0.5975607634,-0.0592407882,-0.1238263398,0.2851480246,-0.1292723119,-0.0741126239,-0.2619338632,-0.0332840607,0.4934832156,-0.200522542,0.0032077192,0.0730269998,-0.0920789316,0.2536216378,-0.1244288683,0.0757653117,0.103560783,-0.0784405246,0.0703940243,-0.2011801451]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2415","title":"Cached dataset not loaded","comments":"Thanks for these explanations. I'm closing the issue.","body":"## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No","comment_length":8,"text":"Cached dataset not loaded\n## Describe the bug\r\nI have a large dataset (common_voice, english) where I use several map and filter functions.\r\nSometimes my cached datasets after specific functions are not loaded.\r\nI always use the same arguments, same functions, no seed\u2026\r\n\r\n## Steps to reproduce the bug\r\n```python\r\ndef filter_by_duration(batch):\r\n return (\r\n batch[\"duration\"] <= 10\r\n and batch[\"duration\"] >= 1\r\n and len(batch[\"target_text\"]) > 5\r\n )\r\n\r\ndef prepare_dataset(batch):\r\n batch[\"input_values\"] = processor(\r\n batch[\"speech\"], sampling_rate=batch[\"sampling_rate\"][0]\r\n ).input_values\r\n with processor.as_target_processor():\r\n batch[\"labels\"] = processor(batch[\"target_text\"]).input_ids\r\n return batch\r\n\r\ntrain_dataset = train_dataset.filter(\r\n filter_by_duration,\r\n remove_columns=[\"duration\"],\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# PROBLEM HERE -> below function is reexecuted and cache is not loaded\r\ntrain_dataset = train_dataset.map(\r\n prepare_dataset,\r\n remove_columns=train_dataset.column_names,\r\n batch_size=training_args.per_device_train_batch_size,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n)\r\n\r\n# Later in script\r\nset_caching_enabled(False)\r\n# apply map on trained model to eval\/test sets\r\n\r\n```\r\n\r\n## Expected results\r\nThe cached dataset should always be reloaded.\r\n\r\n## Actual results\r\nThe function is reexecuted.\r\n\r\nI have access to cached files `cache-xxxxx.arrow`.\r\nIs there a way I can somehow load manually 2 versions and see how the hash was created for debug purposes (to know if it's an issue with dataset or function)?\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.5\r\n- PyTorch version (GPU?): 1.8.1+cu102 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\nThanks for these explanations. I'm closing the issue.","embeddings":[-0.2978544831,-0.0423550606,-0.0758303478,0.4236162901,-0.0719461665,-0.0276887827,0.2326412499,0.3540042639,0.2089607567,-0.0697676763,0.0410520583,0.1886659861,-0.2849694192,-0.2252903432,0.1128703952,0.2975529432,0.0528036244,0.1180419326,-0.1090572625,-0.1540792584,-0.0150639126,0.2576181591,-0.0469234958,-0.093652606,-0.2029935271,0.1055824384,0.0262820609,0.1340813786,0.0921806395,-0.2740397155,0.5368282199,0.0920564905,0.1350365132,0.5165358186,-0.0001210673,0.1466437578,0.2859062552,-0.215739131,-0.2181520611,-0.1280438304,-0.2788822949,-0.0832798928,-0.1238182411,-0.2300582379,-0.0120474277,0.0122144446,-0.0481124632,-0.8536034226,0.459944576,0.0661035106,0.1565443277,-0.1820604801,-0.3796681762,0.1259405762,0.038267713,-0.07413809,0.0205731262,0.1491375864,0.1322319508,-0.0945808291,-0.0412483253,0.5044329762,-0.0829141736,-0.0581240319,0.0016792879,0.0538822822,0.2330061644,-0.1952100992,0.1459397674,0.1084957868,0.1963786185,-0.2124014497,-0.3501954079,-0.2727812529,-0.3075908422,-0.2135212123,0.3252165914,-0.1555833817,0.1468669474,0.1651790589,-0.6091327071,-0.0123855583,0.1913139969,-0.0265445951,-0.0197416823,-0.1129705459,-0.168192789,0.1520161331,0.1083762646,-0.0320031121,0.2745790482,-0.1657436788,-0.0053228564,0.4561800063,-0.0696609691,0.2084786743,0.3096645474,0.0638183728,-0.0398250967,0.2631065547,0.1751727313,0.1068111286,-0.0345857143,-0.0644117445,0.1306186318,0.2133267969,0.2404883653,0.2522965968,-0.049673751,-0.1146898791,-0.5343170762,0.0586863309,-0.1254806519,-0.2830507755,0.7806978822,0.1787222922,-0.0620118529,-0.401501447,0.064640604,0.041482646,-0.42515257,-0.3526496589,0.1482111812,0.4227842987,-0.22410357,0.2358968258,-0.0378687978,-0.1386456192,-0.2654689252,0.1546202749,-0.1668414921,-0.2038098127,-0.3884489834,0.1796002686,0.37313658,-0.4005204141,0.2701569796,0.0872643217,0.0604942627,-0.2748971879,0.165197596,-0.1340657026,0.5168511868,0.0657960027,0.0053642825,0.2182988822,0.1590651125,0.0378639959,-0.2242561281,0.1018682271,-0.5072760582,-0.1304601878,0.318675667,0.0764246434,-0.2396333665,0.3016883731,-0.209837392,0.2644234896,0.3779603243,-0.3355059624,-0.1166514009,-0.2069757432,-0.2666144669,-0.4785521328,0.4195537567,0.6090718508,-0.2946233451,-0.3102104068,-0.1382859498,-0.0281040333,-0.0599078014,0.0682887435,-0.2763515711,0.0992670506,-0.3678852618,0.0261640139,0.4927723706,-0.3572143912,-0.4876314104,0.0474483036,0.250859648,0.5972858667,-0.0802121386,-0.1463817358,0.1930467933,-0.2266284823,0.2826514542,0.1132326648,0.0213674568,-0.1429440379,-0.1665978581,-0.1039288044,0.1472658366,-0.1259729266,0.3338222206,0.1621620655,0.1865675449,-0.1245006174,0.1304099411,-0.0470252037,-0.0186559297,0.3194216788,-0.0883403495,-0.0465562865,0.1233384088,-0.2534306645,-0.3030246496,0.4161659777,-0.277041018,-0.5969229937,0.0289568231,0.0484691896,-0.207630232,-0.0645195767,-0.170820117,-0.2434182167,0.0316517092,0.2297227979,0.2284024656,-0.0210998245,-0.0545175895,0.5442660451,0.0102770561,-0.2740502059,-0.2433663309,0.1692317128,0.2576344907,-0.1716133356,-0.3055517972,0.0876919404,0.3064357936,-0.1491903067,-0.1570721865,0.1695891917,0.0594948232,0.2013689876,-0.2249587923,0.2929288149,-0.1600542963,-0.1222959012,0.1584830284,0.0284264144,-0.0416131653,-0.1636959612,-0.0239242911,0.3278394341,0.1689307094,0.0020596988,0.0516499132,-0.18602705,0.1895095557,-0.1054123193,-0.117535077,-0.2751552165,0.1547856033,-0.1886118948,0.4881473184,0.2394631356,0.0860364437,0.1719860733,0.4886368513,0.2141722739,0.1340205371,-0.0534790717,-0.3166265786,-0.1791449189,-0.1831711382,0.0051268237,0.162058875,-0.000577304,-0.0210616924,-0.023351742,0.0179936346,0.2533574998,0.0779890493,-0.2341278642,0.3470124602,0.1444575787,0.119849056,0.0359204039,-0.1744266897,0.0458373725,0.1211205721,-0.0410864949,-0.466869086,0.1364646703,-0.3292939067,0.2719795704,-0.2452030629,-0.0638307258,0.0286573898,-0.2246178985,-0.1919744313,0.3194694519,0.2880186141,0.2144105434,-0.3410937786,0.1146627292,-0.1894728988,-0.1603946239,-0.0100804102,-0.1131348908,-0.3489112854,-0.0086926008,0.1438347846,-0.2055016458,0.2075018138,-0.0255988836,-0.1401118487,-0.2677106261,-0.3595719039,0.0935123265,0.2070462853,0.3455573022,-0.2412908524,-0.0900674835,-0.20699431,-0.0637390614,0.1708731949,-0.2313913256,-0.1428841799,-0.0684183836,-0.0057264441,0.0865047798,0.0938635096,-0.4032764435,-0.3004500568,0.0697977245,-0.1072492748,-0.0847632661,0.0376866087,0.0733904243,-0.1441195458,-0.2160439491,-0.0327869318,-0.0495372526,-0.4220378101,-0.4178791642,0.2509204149,-0.2321704179,0.0369442664,-0.1274863183,-0.1213700399,-0.0014036332,0.5802029371,-0.3909652531,0.1654570103,-0.0096159568,0.0502640828,0.000515733,-0.1375783086,0.444417119,0.1967123896,0.1541717947,-0.2419928163,-0.411700666,0.0349578336,-0.0705297217,0.2873363793,0.1238329336,0.1701889038,-0.0201322362,1.0965393782,0.3184102774,-0.2103504837,0.0661234036,-0.0407362431,0.2642996013,-0.0109162964,-0.3952404857,-0.1654396057,-0.1520673931,-0.1927796453,0.0578009188,0.063286595,-0.308313489,-0.0400371552,0.2140312791,-0.7177140713,-0.208044365,0.2859528661,-0.6640021205,0.2785124779,0.1591979861,0.1727760732,-0.5495753884,0.068691723,0.082135193,-0.1490960866,0.5069204569,0.0073018586,-0.1675949842,0.1631207913,0.0367319956,0.1653803885,0.2624444664,0.2216524333,-0.123574324,-0.0967764705,-0.2273300737,-0.0536234789,0.4456322789,-0.602388978,0.2966638803,0.3834201992,-0.3278230131,-0.1631358862,-0.2292153388,0.1275748312,0.3774194717,0.1738161296,0.2964749932,-0.0295395795,0.162295267,-0.0784908682,0.1458637416,-0.2814039886,-0.0492803566,-0.0455858894,0.135243699,0.1419732124,-0.0365296789,-0.2272629142,-0.0217074193,-0.0517015569,-0.053078834,-0.4242401421,0.0739685223,0.2555040121,0.3246227205,0.5262849927,0.0624686703,0.4390742183,0.2741711736,0.0351012796,-0.0851440579,0.5787379742,-0.2023314238,-0.1998823136,0.009701441,0.2453024536,-0.1176062599,0.0950476453,-0.385099113,-0.1949116141,0.1685321629,-0.1146743968,-0.1057470292,-0.1660600156,0.055524677,0.2818845212,-0.5762344599,-0.6871487498,0.4059185982,0.4733220041,-0.4103278518,0.7267813087,-0.5869186521,-0.2629710436,0.2095643282,-0.1846629679,0.7170375586,-0.2186813653,-0.0995574594,-0.0923304334,0.3079008162,-0.0393021181,-0.172490254,-0.0400164574,-0.2709791064,-0.1853233725,-0.1254699975,-0.2476443499,0.084847413,0.2775765359,0.0025101651,0.3366933763,-0.2414913028,0.4807322919,0.0184305776,-0.0930507928,0.1714968532,-0.060281191,-0.3047181368,0.1155695915,0.0777915493,0.1758089811,-0.0594418831,-0.0584447756,-0.1147260442,-0.0674225315,-0.3334651887,0.2548058629,-0.6193088293,0.4070167542,-0.1185960695,-0.1381233037,-0.1180677861,0.1156828254,0.1650652289,0.0837796777,-0.0401754305,0.2410151958,0.326459676,0.1791501194,0.0417668931,-0.1866993457,0.1389885992,0.0957636237,-0.1071276888,0.1587900221,-0.1544303447,-0.375262022,-0.176004231,0.1844714731,0.0091225058,-0.014729579,-0.2240322977,0.2788906991,0.2617226243,0.115472585,0.0510925129,0.2583168447,0.1049676538,0.3374639452,-0.1267133802,-0.2926823497,0.0319456309,0.546869874,0.1948584914,-0.1740589887,0.2723882496,-0.1755636781,-0.0421947949,-0.0335020795,-0.3243037164,0.104490377,-0.2906941175,0.0863463804,-0.3525417745,0.1369680464,0.2956292927,-0.151681602,0.1772856414,0.2408417314,-0.1096671373,-0.3795477748,-0.0640935153,-0.0311764106,-0.0599763021,0.0721737891,0.1183805019,-0.297593236,-0.1280510724,0.331507951,-0.2037172467,0.1426204294,0.138836503,0.0612695515,-0.0742919669,0.2138422877,0.2881692052,-0.2710391283,-0.0206499267,0.0642175302,-0.3246044219,-0.0957717597,-0.1054293364,0.1195218414,-0.2374317795,0.1265368909,-0.0961164013,-0.0452616774,0.2535043955,-0.2583850324,0.1769426316,0.4429425895,0.2472510934,-0.1213730574,0.0362662934,-0.2371957153,-0.1984448284,0.2644020021,-0.1955887228,-0.0001742452,0.0041380827,0.3197752535,0.0020018611,-0.0661249831,-0.1795742959,-0.1680130959,-0.0040729488,0.1874286681,0.3716533184,-0.1524595618,0.1358595341,0.135865286,0.229644686,0.2164588422,-0.1311828345,0.0259596594,0.00648139,0.1072962359,-0.0830922425,0.0035993913,0.3172124028,0.1602545381,-0.1353239864,0.278639257,0.1040155143,-0.0837675557,0.2344512939,0.0386339016,0.3939882517,-0.3727237284,0.3957624435,0.2192043662,0.2516993582,0.1452663988,0.1796540171,0.1162650362,0.0013625554,0.256772697,0.178279236,0.3827572465,0.2070094794,0.0789957568,-0.1574572772,-0.5870573521,-0.0352019109,0.3819710314,-0.3661758602,0.501277566,0.0281729121,0.058995869,0.1634183675,-0.3162510097,-0.3057076931,0.3864521682,-0.0145207196,-0.2368498594,0.1871301383,-0.2024393827,0.0762404501,0.0389943346,-0.007976192,0.0427843966,0.5030100942,0.1696486175,-0.221239239,-0.0135474857,-0.18846865,0.1144961566,0.4267623127,-0.1168560162,0.1467285603,-0.0639424175,-0.1349025369,0.0471755788,0.3376615345,0.4609377384,0.2616270781,0.1134784073,0.0790454522,0.4846385121,0.1015991122,-0.2699878216,0.3003548682,-0.0987127572,-0.2325969636,0.151421845,0.1502105594,-0.0925896019,0.2429991663,-0.2007202357,0.3491633534,-0.2665272057,0.0634320751,-0.1339989007,0.2627901137,-0.0019667032,0.1811152548,-0.049006585,-0.1975012273,0.7122183442,0.1442073137,0.2382027507,0.0097113065,0.0515268631,0.2391056567,0.1620854586,0.3663572967,0.058446452,-0.1798638701,-0.0593992993,-0.6679871082,0.0174868777,-0.3761491776,0.2178829014,-0.0890114754,0.2133032382,0.2604070902,-0.0863466859,0.2404281944,0.001278261,0.1334682405,0.2444926947,-0.4332608581,-0.2156454623,-0.1834638268,0.0501266681,0.0466087759,-0.430588603,0.0228872839,-0.056356173,0.0701468289,0.0467086211,-0.2373895347,0.1958326846,-0.112088196,0.644041419,0.1405759156,0.3196020126,-0.0674552545,0.0807409436,-0.3065145314,-0.1040730104,0.1997594982,0.2339827865,-0.0820058882,0.2814269662,-0.3691291809,0.0220728666,-0.2654205859,-0.0618536472,0.1581526697,0.0049349377,-0.0418930389,-0.1768803149,0.0234032944,0.1227749065,0.1063580886,0.5370633602,-0.1948029548,0.3811496198,0.0126683507,-0.0843542144,0.2367663532,-0.5975607634,-0.0592407882,-0.1238263398,0.2851480246,-0.1292723119,-0.0741126239,-0.2619338632,-0.0332840607,0.4934832156,-0.200522542,0.0032077192,0.0730269998,-0.0920789316,0.2536216378,-0.1244288683,0.0757653117,0.103560783,-0.0784405246,0.0703940243,-0.2011801451]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2413","title":"AttributeError: 'DatasetInfo' object has no attribute 'task_templates'","comments":"Hi ! Can you try using a more up-to-date version ? We added the task_templates in `datasets` 1.7.0.\r\n\r\nIdeally when you're working on new datasets, you should install and use the local version of your fork of `datasets`. Here I think you tried to run the 1.7.0 tests with the 1.6.2 code","body":"## Describe the bug\r\nHello, \r\nI'm trying to add dataset and contribute, but test keep fail with below cli.\r\n` RUN_SLOW=1 pytest tests\/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_`\r\n\r\n## Steps to reproduce the bug\r\nIt seems like a bug when I see an error with the existing dataset, not the dataset I'm trying to add.\r\n\r\n` RUN_SLOW=1 pytest tests\/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_`\r\n\r\n\r\n## Expected results\r\nAll test passed\r\n\r\n## Actual results\r\n```\r\n # check that dataset is not empty\r\n self.parent.assertListEqual(sorted(dataset_builder.info.splits.keys()), sorted(dataset))\r\n for split in dataset_builder.info.splits.keys():\r\n # check that loaded datset is not empty\r\n self.parent.assertTrue(len(dataset[split]) > 0)\r\n \r\n # check that we can cast features for each task template\r\n> task_templates = dataset_builder.info.task_templates\r\nE AttributeError: 'DatasetInfo' object has no attribute 'task_templates'\r\n\r\ntests\/test_dataset_common.py:175: AttributeError\r\n```\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Darwin-20.4.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyTorch version (GPU?): 1.7.0 (False)\r\n- Tensorflow version (GPU?): 2.3.0 (False)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n","comment_length":52,"text":"AttributeError: 'DatasetInfo' object has no attribute 'task_templates'\n## Describe the bug\r\nHello, \r\nI'm trying to add dataset and contribute, but test keep fail with below cli.\r\n` RUN_SLOW=1 pytest tests\/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_`\r\n\r\n## Steps to reproduce the bug\r\nIt seems like a bug when I see an error with the existing dataset, not the dataset I'm trying to add.\r\n\r\n` RUN_SLOW=1 pytest tests\/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_`\r\n\r\n\r\n## Expected results\r\nAll test passed\r\n\r\n## Actual results\r\n```\r\n # check that dataset is not empty\r\n self.parent.assertListEqual(sorted(dataset_builder.info.splits.keys()), sorted(dataset))\r\n for split in dataset_builder.info.splits.keys():\r\n # check that loaded datset is not empty\r\n self.parent.assertTrue(len(dataset[split]) > 0)\r\n \r\n # check that we can cast features for each task template\r\n> task_templates = dataset_builder.info.task_templates\r\nE AttributeError: 'DatasetInfo' object has no attribute 'task_templates'\r\n\r\ntests\/test_dataset_common.py:175: AttributeError\r\n```\r\n\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Darwin-20.4.0-x86_64-i386-64bit\r\n- Python version: 3.7.7\r\n- PyTorch version (GPU?): 1.7.0 (False)\r\n- Tensorflow version (GPU?): 2.3.0 (False)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\nHi ! Can you try using a more up-to-date version ? We added the task_templates in `datasets` 1.7.0.\r\n\r\nIdeally when you're working on new datasets, you should install and use the local version of your fork of `datasets`. Here I think you tried to run the 1.7.0 tests with the 1.6.2 code","embeddings":[-0.4735528529,-0.2719739676,-0.0121506173,0.0597860627,0.2671819627,0.2102233022,0.2545935214,0.3851537704,0.0319131128,0.1861087233,0.0951910689,0.4420559108,-0.1509742588,0.0039889752,0.0563462004,-0.0786740929,-0.0314524733,0.1539635062,0.3197634816,0.0171360672,-0.2879475057,-0.1133656055,-0.0043047555,-0.1182009429,-0.1542956233,-0.171497196,-0.1579617858,0.0511678495,0.1129547581,-0.3455854356,0.4298336804,-0.1083598733,-0.0093474761,0.4918485284,-0.00012102,0.178600207,0.3251259029,-0.0006494033,-0.4968328774,-0.065669544,-0.391399622,-0.3287818432,0.0636085197,-0.1634951979,-0.2025725842,-0.3342350125,-0.0632523447,-0.3298051059,-0.086768806,0.4178372025,0.1655729413,0.5218763351,-0.1509704888,-0.4315092266,0.121765703,0.1786002666,-0.2803600729,0.170839712,0.6216207147,-0.1034466848,0.3462252319,0.0053155129,-0.0010401886,0.1662091613,0.1322756559,-0.1352190524,0.223420009,-0.3972377777,-0.0180807989,0.0762008876,0.5973716378,-0.2986184955,-0.4307505786,-0.1419692785,0.2230953574,0.0322043635,-0.0421730131,-0.1669620126,-0.2400552779,-0.0351164341,-0.0903396234,-0.1531327963,-0.1619226635,0.0752020255,0.1189873144,0.022189403,0.1850389838,0.111264661,-0.2130625248,0.1433154047,-0.09717796,-0.0581392869,-0.2552188039,0.1132180989,-0.2312996686,-0.2929350138,0.1680669487,-0.3805031776,0.2416609228,0.3471602798,-0.0348081961,-0.1617386937,0.1267192066,0.2000541091,0.030334333,0.2343725264,0.1062364429,0.2329974473,0.0663586408,-0.1103372946,0.0122648925,-0.1069707945,0.0525073558,-0.2436966449,0.2287586629,0.1440772712,0.6398686767,-0.0755982921,-0.560692668,-0.0253961179,-0.0251773056,0.0050571836,0.2622101605,0.4388638139,0.0128706601,-0.097531423,-0.1945374012,0.3225837648,-0.2628580034,-0.0823602229,-0.2578427196,0.2040020376,0.0043739043,-0.1395382732,0.322098285,-0.1710180491,0.2165559679,0.0552464873,0.1847632378,-0.0080519458,0.2094966918,-0.3163890839,-0.2044803649,0.3319312632,0.0183881335,0.3858396113,0.0564305186,-0.293910861,-0.1657142639,0.2179040313,-0.2973391414,-0.5190382004,-0.0173203126,0.2060691118,-0.0601036549,-0.1723547131,-0.35112831,-0.0557960756,0.1549577266,-0.1310038269,-0.1286191493,-0.3681014478,-0.1906939,-0.1142805442,0.2976214886,0.48236835,-0.3817064464,-0.0535848923,-0.0898585692,-0.1456292123,0.0348261669,-0.0019266937,-0.0912468433,0.3778104186,-0.2277224213,-0.0918507427,0.3079971969,-0.4803474844,-0.3461225927,0.3316580057,-0.0846224725,-0.0342730843,0.1829599142,0.0866437927,0.2815199792,-0.2343328148,0.000161327,0.1537230611,-0.2427243888,-0.0347398706,-0.2711488903,-0.0072603929,0.4636033475,0.2265042663,0.176244691,0.0114101162,0.2185559273,0.0250426698,0.064493753,-0.1237813979,0.1112319976,0.1950040609,0.7014116049,-0.1778082103,0.2300510108,-0.2532387674,-0.686537683,0.344271034,0.298545599,0.0529374108,-0.0783262998,0.0465380736,-0.4575315416,0.2122500539,-0.4423654974,-0.3045431077,-0.023945054,0.1752167642,-0.0368821435,-0.2862261534,-0.1815100312,0.1214267612,-0.3243822455,0.0647803023,0.1674654931,0.3153729439,-0.0987430438,-0.2384933978,-0.1168052033,0.0454538949,0.1586704254,-0.1916502714,-0.2262795568,0.1804823279,0.071852535,0.0765519291,0.2165350169,-0.0281479061,0.2404047698,-0.2461782694,-0.1484141648,0.1785614192,0.2710611224,-0.176605463,-0.0259797741,0.0694446266,0.0350021124,0.3100695014,-0.0453376472,-0.0502872467,0.3370392323,-0.0335725881,0.101717338,-0.0198310129,0.258048296,0.0984007344,0.4189310968,-0.2681364119,-0.301053375,0.1251331717,0.0404604636,0.0795261264,-0.0336546488,0.0976856053,-0.1843661219,0.210137248,0.1349643618,0.1800945401,0.6353419423,0.173872456,-0.1109515801,-0.1063899249,-0.1677987128,-0.0614228733,0.0778772384,-0.0424953401,0.2484363019,0.3933539987,0.2878506482,0.0008509708,-0.1763962656,-0.4252843559,0.3472664952,0.4965883493,-0.3807483613,-0.0114441672,0.042327635,-0.1184769869,-0.1318303198,-0.1406994015,0.0334356315,-0.3064957261,0.1057032496,0.1973102242,-0.014866285,0.2224334031,0.0229597576,0.0128619829,0.0781135038,-0.5534873605,-0.3724315464,-0.1585111469,-0.1578884572,0.0962258652,0.2058898956,-0.2709441483,0.4508914053,-0.0215346403,-0.0256652683,-0.2439593822,-0.2913550436,-0.0090347659,-0.072403416,0.6468446851,-0.0969461352,0.301179409,0.2462478578,-0.2688610554,0.4366424978,-0.0899263397,-0.2616486251,0.1155974641,-0.0554235242,-0.2192442715,-0.1361955702,-0.6507858634,-0.276322335,-0.2776764035,-0.1019040868,0.1167071089,0.0516929589,0.4311493337,0.2770000398,0.125761807,0.2852220535,0.2057918161,0.1320800483,0.0203058086,0.2052997798,-0.041547291,-0.3083122969,-0.0945478156,-0.1930867136,0.1220797598,0.2034581453,-0.1758288741,-0.3868483305,-0.2175997049,0.523040235,-0.1018581465,-0.1444369555,0.394279778,0.0856026635,0.0116369911,-0.1326035112,-0.1141730323,-0.0093503641,0.1062882841,0.1252449304,-0.1373667717,0.2916113138,-0.1365160644,0.6200549006,0.1993594617,-0.1321611255,0.2084064335,-0.3504512608,0.2625141442,-0.1612603813,-0.7129328251,0.2227699161,0.064211905,-0.330975771,-0.0510928258,-0.1329248548,-0.0910040289,0.0487357974,-0.0341808721,0.003298552,0.0740836114,-0.0482872985,-0.1781901717,0.2059230059,-0.0206801333,0.082637459,-0.3545071483,-0.15849334,0.1571548134,0.6275927424,-0.1528115124,-0.0835008398,-0.2704843879,-0.1670535803,-0.3188592792,0.0824508294,0.1177517995,0.4406627119,-0.1966100484,0.0235199016,0.1314789355,-0.0246565398,0.5827756524,-0.2723338306,0.0095703527,0.3472291827,-0.0413709432,-0.2590356469,-0.1730453521,-0.3693192601,0.4203239977,0.1536146551,0.6140229106,-0.0226108171,0.0086647682,0.0132775418,0.2609716952,-0.1683845073,-0.0599217117,-0.1185138822,-0.4029223621,-0.2713449597,0.2448353618,0.0387072824,0.3778189719,0.1394174397,-0.0750323236,-0.4280233383,-0.1475997567,0.2268198878,0.1832056642,0.1736440808,0.1870418787,0.5258866549,-0.4211353362,0.1038896516,0.5118353367,0.3237636983,-0.2278761715,-0.1201091632,-0.0314740725,-0.2060822397,0.601096034,0.019678425,-0.1302602291,0.0310511738,-0.3375561237,0.0540330596,-0.3637283444,0.1568115354,0.3803380728,0.0833119974,-0.3442886174,-0.5051572919,0.128378123,0.1814952493,0.0012116757,0.4686942995,-0.1212277189,-0.1734326184,0.1350249648,-0.0839678198,0.7279384732,0.0567631088,-0.1743060499,0.4782586992,0.089844279,0.2299923152,-0.1944004595,0.1369218081,-0.3040011823,-0.6132839322,0.175073415,-0.2351094186,0.2782959938,-0.0064080544,-0.0599623248,0.1307994574,0.0532904416,0.240463227,0.0838885829,0.4036744237,0.0649515912,-0.090415962,-0.0953709781,0.1300752908,-0.1313066632,0.2946180701,-0.0342864469,0.0626519993,-0.1721254587,-0.1849189252,-0.293775022,0.1219400689,-0.1266431063,0.1483822912,0.1917473078,-0.5770078897,0.3892250955,-0.0981841534,0.1473284662,0.1135730892,-0.1717559546,0.1886997968,-0.0298448,0.2157132626,0.0590928756,-0.166265294,0.5821647048,-0.067997627,-0.1611411273,0.2086188793,-0.217986986,-0.136062935,-0.0340337679,-0.0556782335,0.1239263341,-0.4858964682,-0.2893350422,0.0965529904,0.1161946505,-0.2459219694,0.0801224038,0.1002980545,-0.2555170655,0.2285135686,-0.0605742559,-0.2476042509,-0.0166405719,0.0132460082,0.2224865705,0.1310912371,0.7664225101,0.0360793062,0.063909106,-0.2532014549,0.029471809,0.3147097528,-0.4400612414,0.0714858472,0.0573586822,0.1125411615,0.0217952039,0.233272016,-0.0062285657,0.2217055857,-0.0227417462,-0.4234823883,-0.3272491395,-0.0912705287,-0.0148762912,0.2613730431,-0.1000276208,0.1467792988,0.1819497943,0.1697796583,-0.2714860737,-0.1207595617,-0.1202977449,0.2443989515,0.1946798414,0.2275515944,0.3158340156,-0.1297940761,0.0698057488,-0.2630220354,-0.0597519986,-0.1642076075,-0.1447724849,0.1487659216,0.0992453694,0.1020178944,-0.1359678209,-0.266330868,0.1779774874,-0.1022468805,0.0681575164,-0.0923413113,-0.192423135,0.320966512,0.4243399501,0.0326330438,-0.1276932359,0.2927679121,-0.0328306779,0.2154638618,-0.0123054693,0.1906448156,0.195874989,0.1034165248,-0.4052976668,0.0003747585,-0.115518935,0.0756030008,0.0816716403,-0.3190637827,0.2422188371,0.0013434709,0.1588743031,0.2017826289,-0.2113505453,0.2005332857,0.2323740721,0.1727867573,-0.3805335462,0.0132233407,0.3443752825,-0.0909360796,0.220723629,0.1588815451,0.0000560348,-0.083429642,0.1408260316,0.1894121617,0.13268058,-0.2120161653,0.0879701748,0.1616252214,0.0620883517,-0.0248757433,0.1053942144,0.044382222,-0.0480980724,0.1944150031,-0.123354502,0.2816064954,0.3173201978,-0.1410178244,0.1363497972,-0.1346052438,0.3186090589,-0.2182856947,0.0391699113,-0.2141794115,0.1011769027,0.660155654,-0.1364597082,-0.0901879296,0.0490123853,0.0140312035,-0.298976928,0.047108721,-0.1300397217,-0.0423310846,-0.1368878484,-0.1899407357,-0.0318264738,-0.2115705311,0.1176530421,-0.0210621487,-0.2188140303,-0.1381505579,0.3286408186,0.5079281926,-0.0606167503,-0.2960720956,0.3181618154,-0.0162804965,-0.1382190734,-0.0445127599,0.4015032351,0.5818588734,0.3397773504,0.0334290303,0.0961989388,-0.0238559321,-0.2488906533,0.0481393822,0.1956442744,-0.1965762526,-0.043489825,0.2811099291,0.1509518623,-0.0905115232,0.0437691472,0.061918091,-0.2591669559,-0.3974075615,0.7871944308,0.0563673005,-0.3766653836,-0.1480691135,0.0879432708,-0.2859017253,-0.0843191892,0.5475353003,0.0342186168,0.026958663,-0.1053189412,0.0475581884,-0.1306767166,0.2556800544,0.3318831325,-0.1418061703,-0.4705780149,0.0799428225,-0.5435574651,0.2905920744,-0.2085742652,-0.2139266431,0.0180650987,0.1728942245,-0.1140333861,-0.0173730608,0.2625356615,0.125136748,0.175987184,0.1286887378,-0.3384450078,-0.2251588851,-0.3229908347,0.1170925647,0.1361964941,-0.1330691129,0.0005556246,-0.1072094887,0.0337517224,-0.141115889,-0.2769261003,0.4984536767,-0.1127067506,0.2805220783,-0.0547093265,0.6785972714,0.0421353467,0.0487816855,-0.0744155049,-0.4086259007,-0.1927952617,0.3152897358,0.0242178924,0.1720992327,0.0264170691,-0.1725986451,-0.3324017227,0.0225762725,-0.0961749256,-0.3547200263,-0.0100093801,0.2915281951,-0.2276990563,0.1695405245,0.1270207316,-0.0010680574,0.0956637561,0.1680359691,-0.2658647895,-0.4773276448,0.3552927077,-0.2708622515,-0.2547750473,-0.205025509,0.2875132859,-0.0579046682,0.1274867356,-0.3261030018,-0.1663334072,0.5073917508,-0.1839230508,-0.0795116201,0.3061133921,0.2008480132,0.1990653276,-0.0001379169,0.1951451153,-0.0786879212,-0.0451828204,0.1501549482,-0.1302929223]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2412","title":"Docstring mistake: dataset vs. metric","comments":"> I can provide a PR l8er...\r\n\r\nSee #2425 ","body":"This:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/d95b95f8cf3cb0cff5f77a675139b584dcfcf719\/src\/datasets\/load.py#L582\r\n\r\nShould better be something like:\r\n\r\n`a metric identifier on HuggingFace AWS bucket (list all available metrics and ids with ``datasets.list_metrics()``)`\r\n\r\nI can provide a PR l8er...","comment_length":9,"text":"Docstring mistake: dataset vs. metric\nThis:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/d95b95f8cf3cb0cff5f77a675139b584dcfcf719\/src\/datasets\/load.py#L582\r\n\r\nShould better be something like:\r\n\r\n`a metric identifier on HuggingFace AWS bucket (list all available metrics and ids with ``datasets.list_metrics()``)`\r\n\r\nI can provide a PR l8er...\n> I can provide a PR l8er...\r\n\r\nSee #2425 ","embeddings":[-0.1369899213,-0.3554062247,-0.1418960094,0.1769212484,0.3456723392,-0.0362680741,0.1891716868,0.0313960351,0.2447891086,0.1154253036,-0.1037692726,0.0663549006,0.0386143476,0.2515497208,0.0644848943,-0.1598270535,-0.060389746,-0.0500844643,0.0560807362,-0.0268647224,-0.3186773062,0.3177779615,-0.0659913272,-0.1314190626,-0.1436258703,0.043055281,0.032090079,0.0983669683,-0.5195533037,-0.331392765,0.2579959631,0.0861897618,0.1546093822,0.3069421053,-0.0001050712,-0.1087716967,0.1915326715,-0.0599697679,-0.2200823128,-0.3071030676,-0.2904322445,-0.1485239863,0.2120488584,-0.3109486699,-0.1549810618,0.0720195994,-0.0808921009,0.029760899,0.2454552203,0.3583383262,0.2799392641,0.5363788009,0.0921334028,-0.2711108625,0.0571195893,-0.0530673787,-0.202066198,0.3962208927,0.118728362,0.0467618294,-0.1966291666,0.4474366605,0.0615308322,0.1575723141,0.4222788811,0.1879176348,0.0741885751,0.0565194562,0.1093464568,-0.0919505358,0.2674985826,-0.4487399459,-0.3527985215,-0.1271832883,0.0011156964,-0.4809899926,0.052798219,0.0004971843,0.093171373,0.1827649772,-0.3862278461,-0.2993210554,-0.1326019317,0.0836352855,-0.1610203087,0.1290115863,-0.3498753011,-0.0834533423,0.0333554931,-0.0858433023,-0.2921169698,0.0036322763,-0.0682930052,0.0484668612,-0.2137848735,-0.0139876353,0.2306144983,0.2300419956,0.1982400566,0.4018967152,0.0650253072,-0.0323007219,0.0477913283,0.2009908408,0.0093352152,0.3021989167,0.3787873089,0.2349911779,-0.069038868,0.4383250773,0.20238249,-0.3352252245,0.0694883764,-0.1755695939,-0.05399701,-0.2653626502,0.242716074,-0.2027421445,-0.4189256132,0.1856262535,0.1137957051,-0.1678974628,0.151866734,0.3555713594,0.0950100347,-0.2225698382,0.2009442002,0.3324963748,-0.1837731898,-0.1345189065,-0.3583266437,0.1926366389,-0.3552437723,0.1034065783,-0.0590468086,-0.2541776001,0.464791894,-0.0658695772,0.1949488521,-0.0996931046,-0.1546958983,0.2519117892,-0.1882584989,0.1824600697,0.1252838969,-0.0300128013,0.243930608,-0.2319082916,-0.0045897341,-0.1483695656,-0.2762667537,-0.245356679,-0.2011993676,0.2877593935,-0.220705241,-0.0557298623,-0.1880077571,0.2574056983,0.0920785442,-0.0632984787,0.0731765479,0.1999274194,-0.2260135561,-0.1814037114,0.339555949,0.3590626419,0.0856240317,-0.3611961901,0.1067929119,0.030494133,0.1843943,0.321977973,0.0352829918,0.0474180989,-0.0320811123,0.2384090275,0.2683448493,-0.5044909716,-0.1561730653,0.1970742941,-0.1183884069,-0.1477110535,0.0380154774,-0.1881818771,0.1694646031,-0.057656385,0.2632719576,-0.1485155821,-0.029650405,-0.0238047615,-0.2631935477,-0.1137202159,-0.309352845,-0.0260108653,0.0128852865,0.2307423204,0.2109606862,0.149345845,0.0853260309,0.0275432505,-0.022330638,0.3140347898,0.1519524008,-0.0974829718,-0.0639349744,-0.202567935,-0.4446407855,0.0240929872,-0.0159100071,0.2852667272,0.0030830582,-0.1229389384,-0.4785580635,0.1487239897,0.07856071,-0.3214308918,0.1161264256,-0.0919924676,0.0646213293,0.0633342192,-0.1362060606,-0.1956682056,-0.2705943584,0.2777217925,-0.4926015139,0.088793464,-0.0324331485,0.0369925685,0.3154793084,0.0510451943,0.1747072637,0.096375674,-0.0036497817,0.3432334065,-0.1163238734,0.3498062789,0.2825202346,0.4732847512,0.0738541707,-0.066111058,-0.0265969019,-0.2232058793,0.1133751348,-0.1339005381,-0.4386962354,0.3667370379,0.0270361025,0.1630885601,-0.0375543088,0.0303048175,0.0406359881,0.0090676313,-0.2104444653,-0.2644634247,0.3192302585,-0.1385011077,0.1470571756,0.0001870279,-0.086833179,0.0641851202,0.4066406488,0.0224233214,-0.0555753559,0.241960004,-0.0190714598,-0.0052903788,0.0623210333,0.1164784133,0.4129032195,0.3990406096,0.041805841,0.0993139371,0.1508330852,-0.3591250479,0.2619267702,0.0257120505,-0.1794894189,0.0518752337,0.0047923103,-0.1132465675,-0.2733709216,0.0418188646,-0.1042840928,0.3272950947,-0.404050976,-0.0122396685,-0.0433475226,-0.118610546,0.06264171,-0.1929108799,-0.2162053138,-0.3116529882,0.0368850082,0.2138305455,0.0224902704,0.1545723677,-0.0033315152,0.2807436585,0.0189248547,0.0161116011,-0.3090783954,-0.0481459275,-0.0264787506,0.1356045753,0.0620812885,-0.0365818962,0.4629909694,-0.2658271492,0.2248556465,-0.4801389575,-0.3179039955,0.1279691458,-0.1989565343,0.3403461277,0.4238622189,0.138843447,-0.2542389929,0.0339762829,0.4636236727,-0.1336700618,-0.2277037054,-0.1697055548,0.0053293239,-0.1735257655,-0.1122630686,-0.2758335173,0.1408159137,-0.4156904221,0.2241674811,0.2733454108,-0.0170990676,0.0118817491,0.1364119649,0.074738048,0.0104510421,-0.0934438035,-0.3147366643,-0.722050786,0.1745068878,-0.1922535151,-0.315864563,-0.0149708176,0.2682374716,0.4019905627,-0.0371823385,-0.2776469886,-0.3580754399,-0.0506959707,0.0642230362,0.0606010631,0.3071913421,0.1114287153,-0.038732484,-0.1571645588,-0.2510859668,-0.2721949816,0.0341949575,-0.2212587595,0.3095272481,-0.0424134322,0.1849252433,-0.0022746506,0.7403101325,0.3918928504,-0.0648320094,0.3681847155,-0.1944408566,0.2418417335,-0.0669907182,-0.3487725258,0.1367516369,0.0983963981,0.1068468913,0.417109549,0.2800422013,0.3285652101,-0.0217709299,-0.2272139639,-0.3553292155,-0.2392626256,0.0098033622,0.0126936901,0.0388155095,0.0645616651,0.0354937613,-0.2253553271,0.058030773,-0.017732868,0.2504944801,0.1231630966,0.1482173055,-0.0500381291,0.2490478605,-0.6522081494,0.3657720089,-0.1737143993,-0.1292887926,-0.2234284878,0.1091892272,-0.0027773362,0.0732938573,0.4496645033,-0.2410998344,-0.0808971599,-0.1258058995,-0.1303829104,-0.213474676,0.2245084643,-0.1920987368,0.2988505363,0.0657960325,0.4204018414,-0.5267184973,-0.0729183033,0.1252190471,-0.3640127778,-0.0930684432,-0.3379521966,-0.2427843064,-0.3275200129,-0.3989728093,-0.0915278941,0.2453257293,0.0639190003,0.0913825929,0.2016075104,0.08342316,0.0786496401,0.2675435841,0.0504257679,0.1138610393,-0.0515007898,-0.0297341309,0.1931416243,0.0557656288,0.2625733614,0.6614326835,-0.1269600838,-0.4558272958,0.0178274494,0.1848732084,0.3116720915,0.2424278557,0.0415918119,0.2087954283,0.2063431591,0.3352116346,-0.1896569133,-0.0076292041,0.2925043106,0.2948120534,-0.1746258438,-0.1924529225,0.5163763762,0.0019369818,-0.0093733557,0.3041927516,0.4812480807,-0.2712072432,0.0235795937,0.0075888499,0.7988295555,-0.0389039293,-0.0577277914,0.3728939593,-0.0284100529,0.678589344,-0.3476472199,-0.1038290858,-0.2412153929,-0.066797398,0.0079538208,0.0570213087,0.0835176632,-0.0845739245,-0.1417501122,0.313285321,-0.160070926,0.2052977085,-0.0647693798,0.2526805997,-0.331559062,-0.2837646306,-0.5145003796,0.2589155138,-0.0922073945,0.3155817389,-0.2347978354,-0.157935366,0.1073612124,-0.4173962474,-0.2697360218,-0.1812197417,0.0284068435,-0.18705827,0.1670511514,-0.0073994678,0.0881913975,0.1940266788,0.2094298005,0.0101282699,-0.373224169,0.1711439788,-0.3692679703,-0.0436121933,-0.0303626806,-0.0366580784,0.0757039711,-0.164852798,-0.0921674222,-0.1131407917,0.0464831367,-0.0911273956,-0.0532128438,0.0055582486,-0.1432148218,-0.4522740245,-0.1368871331,-0.1419882625,-0.2189254463,-0.23429811,0.189926818,0.1062907726,-0.0547167733,-0.0816884637,0.3186552227,-0.1905155629,-0.162533164,0.4209215939,-0.2001669258,-0.1919793338,0.0895518064,0.1338779181,-0.1641080081,-0.2814767659,-0.1603309214,0.4161812961,-0.2777767181,0.1968358606,0.1532770842,0.1406347454,0.0607285686,0.2444701344,0.3793388009,-0.0941904783,0.1825460494,-0.4524051249,-0.2017320096,0.4335436821,-0.2015569657,0.1291364133,-0.1049900502,0.1924929023,0.1148664728,0.1526827216,-0.4196881652,-0.1271089762,-0.3219664097,0.0147429872,0.3220666647,-0.0352193378,0.4099052846,-0.0169896223,0.1682608426,0.1422094703,-0.3291974962,-0.2896979451,-0.1535224915,0.1186939031,0.0568514429,0.064225629,-0.0334656164,0.1369307488,-0.2491377145,-0.1764353067,0.0216498133,0.2790236771,0.0284868684,-0.0479959808,0.0763979405,0.1287650913,0.1531406045,-0.1395307779,0.2309925854,0.1097539663,-0.2613290548,-0.0812841356,-0.122559987,-0.0032588274,0.1834439486,0.2243005335,0.4293866158,0.2297611386,0.2005444765,-0.0420603789,-0.0092059132,-0.0760614648,0.4785506427,0.0638398454,-0.1789747775,0.1326900423,-0.1338484138,0.3346493542,-0.3642168045,-0.1124296188,-0.0548033528,0.0775173157,0.1144924164,0.268091023,-0.0601689667,0.0111532332,0.0177564267,0.0097447159,0.5043485165,-0.375780791,-0.0256867893,-0.1590376347,-0.1332166791,0.1455655396,0.5047094822,0.2245737165,0.1105153337,0.1437439173,-0.2101964504,0.3814542294,-0.0620530508,0.2657042742,-0.0038223043,-0.231562987,-0.1940076202,0.1695038229,-0.0238626413,0.0959322974,-0.0129431356,0.51743716,0.0416020714,-0.0533491112,-0.2396288067,0.2376961708,-0.1370607167,-0.1290767789,-0.3128732741,-0.1920586228,0.138947174,-0.1845131367,0.0256241541,-0.2094101608,0.3694067001,-0.1531731188,0.1432806551,-0.4811302423,-0.3411111236,0.1696869731,0.1949613541,-0.2470092475,0.2888572216,0.1385754794,-0.0675915107,0.2106428593,0.3617844582,0.6626459956,0.0390651748,0.4829784036,-0.1039739698,-0.2207463831,-0.1287294477,-0.0729157776,0.2277341634,0.1234151274,-0.2191890627,0.1569187641,0.2750017643,-0.2820508778,0.2874896228,-0.0906155407,-0.0002927266,-0.1699617058,0.1880122423,-0.3504303694,-0.0369727872,-0.2686794698,0.116624929,-0.535158217,0.0707357228,0.2183248401,-0.0394619219,0.2629826665,-0.0918968171,0.1340940595,0.1892806143,0.2581159472,0.3220217824,0.1750477105,-0.4155546725,-0.0530743338,-0.4669246674,-0.0143772205,0.0794879273,0.0605571419,0.0185987297,0.1483553797,0.0671079531,0.1662390977,0.0182244051,0.0799623206,0.0008959466,0.0789148808,-0.3299058974,-0.0092801843,-0.1563940197,-0.0718959868,0.2179852128,0.1090154052,0.1708552092,-0.155763045,0.0616527982,-0.3938594162,0.4140229523,0.0477939434,0.2129092216,0.0367512107,0.2294604629,0.3348456323,-0.3326281309,0.0864575505,-0.1204344481,-0.3173261285,-0.1907368898,0.2844718397,-0.1135386676,0.1604587436,-0.1259060353,-0.3361112475,-0.2603803575,0.2113907784,-0.0086425962,-0.150267899,-0.2489551455,0.2523683012,-0.1561409384,-0.0005607987,-0.0552718155,0.414534986,0.100842908,0.2208876759,-0.1407539696,-0.4408956468,0.7458025217,-0.3814136386,-0.1240806282,0.355358392,0.2076824456,0.5019419193,-0.3132382035,-0.9022820592,0.0935367048,0.271158725,-0.0512546822,-0.2528682649,0.2630099058,-0.1616929621,0.075741075,-0.0775683448,0.2614367306,0.1240926534,-0.2855709791,0.140740782,-0.2329535037]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2407","title":".map() function got an unexpected keyword argument 'cache_file_name'","comments":"Hi @cindyxinyiwang,\r\nDid you try adding `.arrow` after `cache_file_name` argument? Here I think they're expecting something like that only for a cache file:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e08362256fb157c0b3038437fc0d7a0bbb50de5c\/src\/datasets\/arrow_dataset.py#L1556-L1558","body":"## Describe the bug\r\n\r\nI'm trying to save the result of datasets.map() to a specific file, so that I can easily share it among multiple computers without reprocessing the dataset. However, when I try to pass an argument 'cache_file_name' to the .map() function, it throws an error that \".map() function got an unexpected keyword argument 'cache_file_name'\". \r\n\r\nI believe I'm using the latest dataset 1.6.2. Also seems like the document and the actual code indicates there is an argument 'cache_file_name' for the .map() function.\r\n\r\nHere is the code I use\r\n## Steps to reproduce the bug\r\n```datasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n cache_file_name=\"my_tokenized_file\"\r\n)\r\n```\r\n\r\n## Actual results\r\n tokenized_datasets = datasets.map(\r\nTypeError: map() got an unexpected keyword argument 'cache_file_name'\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.6.2\r\n- Platform:Linux-4.18.0-193.28.1.el8_2.x86_64-x86_64-with-glibc2.10\r\n- Python version:3.8.5\r\n- PyArrow version:3.0.0\r\n","comment_length":24,"text":".map() function got an unexpected keyword argument 'cache_file_name'\n## Describe the bug\r\n\r\nI'm trying to save the result of datasets.map() to a specific file, so that I can easily share it among multiple computers without reprocessing the dataset. However, when I try to pass an argument 'cache_file_name' to the .map() function, it throws an error that \".map() function got an unexpected keyword argument 'cache_file_name'\". \r\n\r\nI believe I'm using the latest dataset 1.6.2. Also seems like the document and the actual code indicates there is an argument 'cache_file_name' for the .map() function.\r\n\r\nHere is the code I use\r\n## Steps to reproduce the bug\r\n```datasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n cache_file_name=\"my_tokenized_file\"\r\n)\r\n```\r\n\r\n## Actual results\r\n tokenized_datasets = datasets.map(\r\nTypeError: map() got an unexpected keyword argument 'cache_file_name'\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.6.2\r\n- Platform:Linux-4.18.0-193.28.1.el8_2.x86_64-x86_64-with-glibc2.10\r\n- Python version:3.8.5\r\n- PyArrow version:3.0.0\r\n\nHi @cindyxinyiwang,\r\nDid you try adding `.arrow` after `cache_file_name` argument? Here I think they're expecting something like that only for a cache file:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e08362256fb157c0b3038437fc0d7a0bbb50de5c\/src\/datasets\/arrow_dataset.py#L1556-L1558","embeddings":[0.1323000193,0.0689882264,0.0382361263,0.0000564463,0.0520310178,0.2963714898,0.1708019078,0.3389790654,-0.0635626316,-0.0203007851,0.1289214194,0.610024035,-0.2410923839,-0.5483763218,0.0488028489,0.090259932,0.3929358423,-0.0865352526,0.2103406191,0.0160696357,-0.3766417205,0.3897094429,-0.0564443506,0.2779166102,-0.4511603117,-0.2072949708,-0.0894433111,-0.0664880276,-0.0896062851,-0.391181618,0.2553226352,0.0901256204,-0.0756616443,0.489871949,-0.0001236526,-0.0626008809,0.0621398427,-0.164627403,-0.2050175667,-0.3376171291,-0.108075425,0.0543018244,-0.0680716634,-0.497097224,0.2129514813,-0.0833946764,0.1886314303,-0.6428586841,0.1455971301,0.172502622,0.1301087439,-0.1011054739,-0.2413280308,-0.1786019355,0.1032098681,-0.1168403178,-0.0114621865,0.2041410208,0.035358049,-0.2602057159,0.1032075211,0.3366818726,-0.3803628087,-0.0966832191,0.4977585375,0.0177516527,-0.2116977423,-0.3391501904,0.2500306368,-0.1566718221,0.1236489341,-0.4110817015,-0.3522394896,-0.1329432279,-0.3212546408,-0.1088350564,0.1613844782,-0.0994668975,-0.1431579143,-0.0172122009,0.0187122691,-0.2126890421,0.0077328901,0.1374987811,0.2252310216,0.1087041497,-0.1958861947,0.3618341386,0.3491643369,-0.0833931491,-0.2166385353,-0.3524973094,-0.0401295535,0.3894569576,-0.1926616281,-0.0108110849,0.0461867489,0.1170019284,0.0980422422,-0.3093523681,-0.0926846117,-0.145090729,0.0813978985,0.3073478937,0.2356880158,0.2836824954,0.3462310135,0.4869079292,-0.1531836241,-0.3088080585,-0.2916905582,-0.2059519738,0.0763019547,-0.1290071458,0.5090178847,-0.0678104758,0.2955312729,-0.1016577184,0.0983532891,0.0917264596,-0.3417958319,-0.0036549708,-0.0990243405,0.341468215,0.1448268741,-0.0943260938,-0.0087400302,0.2082662582,0.0484691709,-0.1415781528,-0.1375662535,0.0520832837,-0.253726095,0.2101465315,-0.1983944476,-0.0464085862,0.1395723373,0.0137630878,0.170556888,-0.1162838638,0.160295859,-0.0472937115,0.2373288572,-0.1563026309,-0.1153677031,0.4976298213,0.2821669877,-0.2278868705,-0.1183428764,0.0254225265,-0.5410766602,-0.3162823617,0.000062654,0.022127213,-0.0346530601,0.1262811273,-0.426648587,0.0120566403,0.5163833499,-0.2676182985,0.0308113955,-0.0304689668,-0.2301573753,-0.4813354611,0.020797763,0.3271371126,-0.0734873638,-0.0302486084,-0.0271823667,0.1944456547,-0.0033592696,0.1681511849,-0.0935864821,0.1899894774,-0.2116828561,0.7205870152,0.5101905465,-0.2484844923,-0.5775424838,0.273391515,-0.2147663087,0.013901134,-0.0861738399,0.0286193695,0.4520129263,-0.1365366876,0.437209785,-0.0731702372,-0.0409542918,0.1518031508,-0.2699029744,-0.2399378121,0.1261731833,-0.2583991587,-0.0506732762,0.1763263047,0.0805187747,0.0621099956,0.1262601614,0.097930178,0.234285295,0.0838978663,0.2890298367,0.0728417113,0.0257608369,-0.008435946,-0.2922082543,0.0839115307,0.196461454,-0.4200127721,-0.3298364878,-0.4760627747,-0.0561076403,0.0424852632,-0.1884646863,-0.0734005347,0.0064832517,0.0784215108,0.0639972761,0.1206236333,-0.0759832785,-0.0023852827,0.0435928106,-0.0542773716,0.0612000935,-0.2633388042,-0.0796586573,-0.1117172316,-0.2368734181,-0.1921242774,0.4008702934,-0.0590945631,-0.3225570917,0.4013506174,-0.0077051329,0.117150262,-0.0612250417,0.2972923815,-0.2990592122,-0.0420838594,0.0486955419,0.0069589787,0.0879930481,-0.0257945675,0.1187412217,0.1925860494,0.2606340051,0.1826837212,-0.2336699963,0.2960456014,-0.0235924106,0.0437992625,0.0012439389,-0.2449383885,-0.0402426831,-0.1408083737,0.3976156712,-0.1182780489,0.0518170111,0.0109181646,0.4879014492,0.125406608,0.1525567174,-0.0630090386,-0.1764896959,0.025393324,0.089999795,0.3629103899,0.5742313266,-0.0069505544,0.0451245978,-0.0071809585,0.1451582313,0.0854675472,0.0900083482,0.0456476398,-0.1075735614,-0.2344181985,-0.0852103606,0.240983218,0.0570673868,0.0389851555,0.1391420662,0.101846315,-0.1943262368,0.2980097234,-0.092138119,-0.0072416645,-0.0684707388,-0.0650317073,-0.2314721495,-0.4921033382,-0.1023306847,0.1390251219,0.0433198437,0.2613464892,-0.20452106,0.194872275,0.0464905873,-0.4844405353,-0.0153643424,-0.2388944328,-0.0431445278,-0.204358235,0.0273899734,-0.3609799147,0.421936214,-0.0387925506,-0.0085765989,-0.3726302087,-0.1466664672,0.1330472082,-0.1224432066,0.1531076729,0.1415273398,0.1062693521,-0.3267288804,-0.194094494,0.1713119298,-0.2614380121,0.0331206471,-0.078739509,0.1121828184,-0.2484862953,-0.1560818255,-0.0955816433,-0.0614487641,-0.1265228242,0.3240824342,-0.3173361421,0.0880131945,0.1432165205,-0.027240362,0.2221166342,-0.3109336495,-0.10956911,-0.3583908081,-0.4785100818,0.4354128838,-0.2844851911,-0.2151241452,0.0019907448,0.057266511,0.1640390307,0.2437956184,-0.2821912467,-0.2356435806,0.1157440618,0.2570863962,0.0183189958,-0.0090194829,0.2191235572,-0.0111932848,0.0872837752,-0.3030955493,-0.4302428663,0.1351031661,0.1022798121,0.3065890372,0.4003380239,0.2808996737,0.3629715443,0.7278517485,0.3356896043,-0.0266913902,0.4212534726,-0.1764850616,0.3586527407,-0.0308426954,-0.4958994091,-0.0759457126,-0.096144706,-0.2645400763,-0.0491271913,-0.0876551569,-0.1360626966,0.1183231547,-0.1151531041,-0.2305315733,-0.4181848466,0.0533832833,-0.3917154968,0.1326438487,0.0615811758,0.1490504593,-0.5956489444,-0.0849507302,0.1158467755,0.2924676836,0.1579387635,0.1532180309,-0.3381687105,-0.0135259628,-0.4353027046,0.5255327821,0.0376698077,0.3008788228,-0.1652258188,0.1278519928,-0.0330789238,0.2386569977,0.7517746687,-0.2644421756,0.031440001,0.0712054521,0.2308716476,-0.2254027724,0.0461838506,0.2163041979,0.8413039446,0.1609923691,0.4745667875,-0.1487785876,-0.1196168587,-0.2819727659,0.0935418904,-0.1004221663,-0.0052041779,-0.0415228941,-0.0565000251,-0.4008994997,-0.0090705799,0.235584259,-0.0150325317,0.3674220145,0.1424491554,0.1331077665,0.0558584183,0.1616948247,0.191403836,0.3781094551,-0.0754214451,-0.2005077004,0.2696893811,-0.0140879313,0.1037214994,0.2063851655,-0.2506126463,-0.1245501563,-0.1312492937,-0.0998670459,0.3893234432,0.1580215096,-0.2658316791,0.080271028,-0.0858954266,-0.0175331943,0.0140579883,0.2113144249,0.4594945312,-0.0593054257,-0.5275682807,-0.4904386997,0.2581529319,0.2997872531,-0.2697074711,0.8450321555,0.0675557256,-0.3200567365,0.4222558141,-0.1042792946,0.8706273437,0.0174507443,0.233898133,-0.1201487109,0.1141838059,0.3580418825,-0.026596101,-0.0544714406,-0.1759698838,0.073018454,0.0455647334,0.0271887723,0.0868418962,0.4136997461,-0.3510037363,0.2176757902,-0.0524417832,0.4112336934,-0.0447034687,0.0122163193,-0.0464386493,-0.1867203414,-0.0871264189,0.10723719,-0.0259740278,0.1294765621,-0.0183796231,-0.1637247801,-0.0068756598,-0.4977631569,-0.1725837588,-0.0030542703,-0.5952806473,0.3011161089,-0.1135485545,0.0605969504,0.0023503704,0.3404524326,0.0794319287,0.2227647454,-0.0841487274,-0.018820988,0.1150722355,0.006619012,-0.1201634184,-0.2954678237,0.1425801665,0.1661803871,0.0112981601,0.1389872432,-0.0704510435,-0.5288639069,-0.1938417107,0.0623061135,0.2274815738,-0.3000990152,-0.4811905622,-0.2001834065,-0.1463021636,-0.053829357,0.0190693196,0.0449602865,-0.0798003599,0.3188449442,-0.1389811635,-0.243743822,0.2529884577,0.3342953324,-0.3944354951,-0.2294706702,0.4249681234,-0.340541929,-0.0756213889,-0.0775876716,0.0927992016,-0.0627633929,-0.0084368149,0.1716315448,-0.2090996802,0.1777826101,-0.0807370916,0.117148228,0.217057839,0.2274258286,0.0067639537,-0.3460564315,-0.2396903783,0.067724511,-0.1967144608,0.1427447051,0.1849365532,-0.0298067853,0.0721746311,0.1650880873,-0.1944781691,0.1778861731,0.01750925,0.0619778186,0.3951482475,0.3479487002,0.2850729823,-0.1957460195,-0.0405347161,0.2143423259,-0.1704339236,-0.0689629689,-0.3461568654,0.2059270591,0.2186740935,0.025395995,0.1010091007,-0.3307479918,0.1104192138,-0.3575364947,0.0508436225,0.3860601485,-0.0687766746,0.3203295171,0.3072118163,0.1299646646,0.0060877791,0.0349603705,-0.0458491743,0.1752848327,0.0297457334,0.2047571391,0.0185490381,-0.0132541293,-0.1200756505,0.1309159845,-0.0850157216,0.0991977975,0.1645425409,-0.1579203159,0.1042553112,0.4336757958,0.0309245512,0.1451113373,-0.4857906699,-0.0307353865,0.2163935006,0.0697331578,0.1154649705,-0.2508092225,0.7882000208,0.1106002107,-0.073002018,0.234962061,0.0217755362,-0.0178702008,-0.1021969095,-0.170132935,0.4406094849,-0.4514471292,0.3755295277,0.2748088241,0.2161708474,0.3323142529,0.2810053229,0.3359888494,0.1987060905,0.5423124433,-0.2920403779,0.2530552149,0.2998801768,-0.1417287588,0.1987141818,-0.2295819819,0.16809237,0.1797321737,-0.2209814191,0.4865629375,-0.0482227057,0.4005236924,-0.3815602362,-0.0810335055,0.1354394853,0.1693060249,-0.2599409521,-0.0303710923,0.0784511864,-0.0027461094,-0.0521514118,-0.1279712766,-0.1171713993,-0.1499399543,0.6191293001,0.002300336,0.0840351358,0.1402328759,-0.1033389047,0.1035468653,0.4382250905,-0.4216110706,-0.066862613,-0.3315870762,0.0253687873,0.0993204936,0.1796460599,0.5354409814,0.5416946411,-0.0827153474,0.1162161604,0.0940868855,-0.0946573764,0.0472855307,0.0480317101,-0.102966316,-0.2641995251,0.2053313255,0.037208572,0.0143864509,0.1431319863,0.2195749283,0.4603921175,-0.0320070833,0.1295123398,-0.1524962038,-0.2574126124,-0.2486703396,0.2730790377,-0.3320175409,-0.0996738821,0.368712008,0.1682631969,0.2206092924,-0.1890610456,0.0401911512,0.2439327985,0.0987141579,0.207496956,-0.0657489896,-0.5534555316,-0.1550088972,-0.6261844039,0.1647972763,-0.0627425462,0.137770161,0.0240122508,-0.0295651946,-0.0035253125,0.1277083009,0.1839986593,-0.1906885058,-0.2176295817,0.338257432,-0.1625491679,-0.158642903,0.0577393472,0.1596433818,0.4403271973,-0.2399888784,-0.0172734242,0.0005746132,-0.0213475823,-0.3502492607,-0.1870134473,0.0824535787,-0.0299031101,0.6764065027,0.0250769332,0.2400174737,-0.0070618764,-0.2346413285,-0.0556816161,0.1813787818,-0.2350433022,-0.1356525123,-0.1792160124,0.2597793043,-0.414819032,-0.2854936123,-0.5130299926,-0.1274336427,-0.0590394065,-0.2208183408,-0.1585983187,0.2819311321,-0.0363105126,0.2087019682,0.073071003,0.4152940512,-0.2337919176,0.629226625,-0.1960882843,-0.0745547637,0.5311341882,-0.6217187643,-0.4185054898,0.1602221727,0.2648943365,-0.0957688093,-0.4335862696,-0.4637288451,-0.1815104485,0.4318005741,0.0812447593,-0.3228625953,-0.0238837972,-0.1468506008,0.1304251552,0.0383933894,0.5380346179,0.0021399576,-0.1710185111,0.3173922002,-0.3794923127]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2407","title":".map() function got an unexpected keyword argument 'cache_file_name'","comments":"Hi ! `cache_file_name` is an argument of the `Dataset.map` method. Can you check that your `dataset` is indeed a `Dataset` object ?\r\n\r\nIf you loaded several splits, then it would actually be a `DatasetDict` (one dataset per split, in a dictionary).\r\nIn this case, since there are several datasets in the dict, the `DatasetDict.map` method requires a `cache_file_names` argument (with an 's'), so that you can provide one file name per split.","body":"## Describe the bug\r\n\r\nI'm trying to save the result of datasets.map() to a specific file, so that I can easily share it among multiple computers without reprocessing the dataset. However, when I try to pass an argument 'cache_file_name' to the .map() function, it throws an error that \".map() function got an unexpected keyword argument 'cache_file_name'\". \r\n\r\nI believe I'm using the latest dataset 1.6.2. Also seems like the document and the actual code indicates there is an argument 'cache_file_name' for the .map() function.\r\n\r\nHere is the code I use\r\n## Steps to reproduce the bug\r\n```datasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n cache_file_name=\"my_tokenized_file\"\r\n)\r\n```\r\n\r\n## Actual results\r\n tokenized_datasets = datasets.map(\r\nTypeError: map() got an unexpected keyword argument 'cache_file_name'\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.6.2\r\n- Platform:Linux-4.18.0-193.28.1.el8_2.x86_64-x86_64-with-glibc2.10\r\n- Python version:3.8.5\r\n- PyArrow version:3.0.0\r\n","comment_length":72,"text":".map() function got an unexpected keyword argument 'cache_file_name'\n## Describe the bug\r\n\r\nI'm trying to save the result of datasets.map() to a specific file, so that I can easily share it among multiple computers without reprocessing the dataset. However, when I try to pass an argument 'cache_file_name' to the .map() function, it throws an error that \".map() function got an unexpected keyword argument 'cache_file_name'\". \r\n\r\nI believe I'm using the latest dataset 1.6.2. Also seems like the document and the actual code indicates there is an argument 'cache_file_name' for the .map() function.\r\n\r\nHere is the code I use\r\n## Steps to reproduce the bug\r\n```datasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n cache_file_name=\"my_tokenized_file\"\r\n)\r\n```\r\n\r\n## Actual results\r\n tokenized_datasets = datasets.map(\r\nTypeError: map() got an unexpected keyword argument 'cache_file_name'\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.6.2\r\n- Platform:Linux-4.18.0-193.28.1.el8_2.x86_64-x86_64-with-glibc2.10\r\n- Python version:3.8.5\r\n- PyArrow version:3.0.0\r\n\nHi ! `cache_file_name` is an argument of the `Dataset.map` method. Can you check that your `dataset` is indeed a `Dataset` object ?\r\n\r\nIf you loaded several splits, then it would actually be a `DatasetDict` (one dataset per split, in a dictionary).\r\nIn this case, since there are several datasets in the dict, the `DatasetDict.map` method requires a `cache_file_names` argument (with an 's'), so that you can provide one file name per split.","embeddings":[0.0843966529,0.0935957208,0.0456615947,0.0183026306,0.0415576138,0.2793120444,0.2214866579,0.3731047213,-0.0277668294,-0.0000239611,0.1310784668,0.6276783943,-0.2474791855,-0.4753703177,0.0740652829,0.1246322691,0.3925755322,-0.0903926194,0.2432297915,-0.0068806205,-0.3810337186,0.373701781,-0.0541739129,0.255556196,-0.4504035413,-0.1992647797,-0.1135985106,-0.0584842935,-0.0518890321,-0.346601218,0.182830736,0.1161076277,-0.0857539549,0.4750781953,-0.000123576,-0.0917597637,0.0360114127,-0.1631390154,-0.255508095,-0.2864112854,-0.144255355,0.0374925956,-0.107219696,-0.5347388983,0.2405566126,-0.0832321048,0.1901536733,-0.6212079525,0.154182896,0.1646748781,0.1337841898,-0.0938229486,-0.2403854579,-0.168412298,0.0872015953,-0.0968493819,-0.0081237555,0.1451229453,-0.0240640473,-0.2789250612,0.1103762835,0.3297784328,-0.3978562057,-0.1006857231,0.4735762477,0.0346832797,-0.2390235364,-0.3423895538,0.2542084157,-0.164136529,0.1279025376,-0.4114086926,-0.3495681286,-0.1497306377,-0.3343961835,-0.1376321614,0.1707211137,-0.0676344037,-0.1131243929,-0.000665666,0.0142389946,-0.2325133979,0.0578134768,0.1314535737,0.1871845275,0.1065116525,-0.1610174924,0.3629954159,0.32640028,-0.069491446,-0.2115776837,-0.3912232816,-0.0488409474,0.3552266657,-0.168315053,-0.0316751972,0.0834232643,0.0786998421,0.0730248466,-0.3154939115,-0.129665643,-0.1537213624,0.1106229201,0.3209197223,0.2893005908,0.2396935821,0.3267398179,0.5005446672,-0.1488441229,-0.288900733,-0.2772119343,-0.1784284413,0.0600887984,-0.0740306005,0.5499634743,-0.0462909229,0.283651799,-0.0727455467,0.05186373,0.0815827176,-0.3601100445,-0.0161578674,-0.078132242,0.3521354496,0.1568032801,-0.074087061,-0.0127624804,0.171187073,0.0065890537,-0.1879358441,-0.1625734568,0.0591677204,-0.2378071845,0.2446793616,-0.1718951613,-0.0271145198,0.0770326331,0.0156629775,0.1433115304,-0.1758872271,0.1657017171,-0.0921750665,0.2263487726,-0.1334418356,-0.08519499,0.5168968439,0.25256446,-0.2299636155,-0.1541354805,0.0480262861,-0.5041528344,-0.3401092589,0.0116798328,0.0334643573,-0.0309841894,0.1496296972,-0.3882625997,0.0301382802,0.5299628377,-0.2702380717,-0.0235784054,-0.0733261108,-0.2511709332,-0.4917123318,0.0383105204,0.349750191,-0.109250024,-0.0363359489,-0.0776748732,0.155304119,0.0279318616,0.1516963094,-0.1027360708,0.2707727253,-0.2081891149,0.7426680923,0.4991734326,-0.1963742375,-0.6254301667,0.2615728974,-0.1825743765,-0.0140554942,-0.0932946056,0.03504676,0.5258513689,-0.1505950391,0.4445669651,-0.0091094961,-0.0476919077,0.2018995434,-0.2865401208,-0.2564961016,0.1329108328,-0.2272552699,-0.0024048383,0.1572242826,0.0922307596,0.1445804387,0.1027358994,0.0925472528,0.225297153,0.0467780381,0.2800409794,0.0934413821,0.0233280342,-0.0101456726,-0.2804198265,0.0546406284,0.2277635783,-0.4543118477,-0.2949874103,-0.5119615197,-0.0737078711,0.0596263073,-0.1761666536,-0.0526076779,0.0141654415,0.1137335747,0.0338054113,0.0811935514,-0.0729893744,-0.0186709911,0.0781614259,-0.0853207186,0.0549881682,-0.2141624987,-0.1000522599,-0.1230626479,-0.2338081449,-0.1642668247,0.412866652,-0.035357222,-0.3176954389,0.4300271571,0.0383871421,0.0824360773,-0.1083932891,0.2794752419,-0.3045786321,-0.0324729867,0.0536520779,0.0173651762,0.1332926899,-0.0131594846,0.0782387033,0.162786901,0.2941243649,0.152560398,-0.2231174558,0.3086067736,-0.0300187785,0.0038742777,0.0457966402,-0.2656097114,-0.0447885618,-0.1367946267,0.3153089583,-0.0962775946,0.0436739475,0.0585999489,0.5465804935,0.0896509588,0.2021795958,-0.0539966114,-0.1480334103,-0.0148798497,0.0500316955,0.370580405,0.5701339245,-0.0074127931,0.0329305939,0.0049894121,0.1401070505,0.1061861664,0.0774008334,0.0539289638,-0.1377099901,-0.2030059993,-0.0961868837,0.2135051191,0.0581591241,0.0547279902,0.1970949173,0.1184690967,-0.2186033428,0.3140494227,-0.1162796542,0.0529644303,-0.0950969607,-0.0377407558,-0.2213388383,-0.5031988621,-0.086987704,0.1824336052,0.0225831866,0.2560198605,-0.1359323263,0.2338398546,-0.0110110706,-0.4088107944,0.0171660986,-0.2472446859,-0.08292377,-0.2146368325,0.0458928868,-0.2960109711,0.433589071,-0.0514927767,-0.065821439,-0.3763533831,-0.0999039561,0.1506199837,-0.128230989,0.16671817,0.1436538994,0.1126276553,-0.2614251077,-0.2257205695,0.1647644639,-0.27018857,0.0349736214,-0.0669245496,0.1397510022,-0.2369216233,-0.142332226,-0.1207683608,-0.0906235054,-0.116723381,0.258715868,-0.3342749476,0.1245620847,0.0998176336,-0.0638102144,0.2011466622,-0.2483415306,-0.1572299004,-0.4022387862,-0.3883705735,0.3892168105,-0.270865798,-0.2293523401,0.0011085076,0.0592357889,0.105509229,0.2628535032,-0.2950545549,-0.2078715563,0.1155484989,0.2173290551,0.0064052753,-0.0176845882,0.27014485,-0.0041283728,0.0952356234,-0.2970133722,-0.4936104417,0.1092556641,0.1201159284,0.3229607642,0.3394294083,0.3143760562,0.3534232378,0.710734129,0.3199200928,-0.0045737433,0.391482234,-0.1586185992,0.3285063207,-0.0312450491,-0.515438199,-0.0435898602,-0.1356969029,-0.3268224895,-0.0196899828,-0.1299832165,-0.1846941859,0.1277697384,-0.1052170172,-0.2755036056,-0.4394861758,0.079189688,-0.3375879824,0.1755042523,0.0630628765,0.1905632466,-0.6088643074,-0.0744217187,0.0866465271,0.2727277875,0.1386161894,0.1183940545,-0.3839710951,0.0072866376,-0.3965055645,0.5132543445,0.031205954,0.3311236203,-0.1937033087,0.107031092,-0.0153562184,0.1930265129,0.7922773361,-0.2679607868,0.045488704,0.1210801452,0.2072666287,-0.1734303981,0.0165106878,0.176727429,0.8107902408,0.1846639663,0.4598003626,-0.1755235195,-0.1126585454,-0.2083562315,0.0926601216,-0.1504731774,-0.0065759905,-0.0410153344,-0.0956522897,-0.4337189496,-0.017076008,0.2169876844,0.0194635149,0.3484913409,0.1774245948,0.1002261788,0.0034343104,0.1716186851,0.195819065,0.3997894228,-0.0367517099,-0.1422632188,0.2697028518,-0.0535964817,0.1151066348,0.1766400635,-0.2531785071,-0.1062384024,-0.1047325432,-0.1075570136,0.3750821352,0.1819928437,-0.2561524808,0.0344507918,-0.1116351411,-0.0082290675,-0.059259776,0.2258591205,0.4638691247,-0.0926958099,-0.5957253575,-0.5398175716,0.2328742743,0.3057694137,-0.2729085684,0.784509182,0.0274978969,-0.3102461398,0.4755221605,-0.1116868779,0.9144572616,-0.0222541261,0.2302893698,-0.135551542,0.1126970872,0.3551304936,-0.0616424493,-0.0126772719,-0.1980343163,0.0328161865,0.0430368595,0.0327878445,0.0927533507,0.4047366977,-0.3663408756,0.2410905361,-0.1058428138,0.4469736814,-0.0407314524,-0.0070345374,-0.0390015393,-0.2001373023,-0.0524476022,0.1204576045,-0.0571443141,0.121959053,-0.0035046898,-0.1379832327,0.0385709926,-0.493968308,-0.2175468802,0.0263958778,-0.5055758953,0.3241205812,-0.1637339145,0.0933815315,-0.0147192795,0.3070189357,0.0865072459,0.2390553206,-0.0346438028,-0.0008355866,0.0770904496,0.0066745649,-0.1286074668,-0.2843939364,0.1596182287,0.1853232086,-0.0073416471,0.1485629976,-0.0257387795,-0.5060883164,-0.2224616408,0.0730461627,0.2858823836,-0.3665453196,-0.438590169,-0.1720102131,-0.1113936231,-0.1033290774,0.0238906834,0.010730477,-0.1097253561,0.3575002849,-0.2395234257,-0.2372450978,0.2461716831,0.3517597914,-0.3285679221,-0.1857031137,0.4542012215,-0.34026438,-0.0653129891,-0.1118479446,0.111395061,-0.0910291597,0.0006981475,0.2068597078,-0.2725410461,0.1638411433,-0.1023332775,0.1258582175,0.1827770323,0.2274897993,-0.0206734333,-0.3733623922,-0.1848761439,0.0356907956,-0.1771189421,0.1585329026,0.2361739576,0.002523022,0.024298016,0.2208945453,-0.2025887817,0.1812765896,-0.0320364311,0.0858319849,0.3761409223,0.3112924695,0.2769512534,-0.2245524377,-0.0314897038,0.1976859868,-0.1631717831,-0.0721167177,-0.3499280214,0.2080765069,0.2200913429,0.0250810757,0.0658119991,-0.3174197376,0.1206403151,-0.3502321243,0.0448617376,0.3928798735,-0.0669546276,0.2886337936,0.354804486,0.135775283,0.01507458,0.0360131264,-0.049216602,0.1835970581,0.0180043727,0.2534424961,0.0322392508,-0.0113042602,-0.1823342443,0.1354929358,-0.0019151209,0.1167614162,0.1840479076,-0.1613006741,0.136233598,0.4229466319,0.0161658339,0.1729230285,-0.5007998943,-0.0365425982,0.2302373648,0.0679261833,0.1230835468,-0.2761168778,0.8008127213,0.0760899112,-0.1169250458,0.2435141057,0.0345423855,0.0245291702,-0.1695734859,-0.1669195443,0.4612924159,-0.4105742872,0.3882528543,0.3578399122,0.2091784626,0.3332790732,0.2539384663,0.3218890131,0.1807016283,0.5784000754,-0.258454293,0.2908930182,0.3010332286,-0.1214266345,0.1967751831,-0.2728750706,0.193494916,0.1676933169,-0.1804718524,0.4687674046,-0.0906146243,0.3237780929,-0.4149036705,-0.0803274587,0.0942230225,0.1906607151,-0.2526555359,-0.0348542482,0.0795268565,0.0199235715,-0.0897658691,-0.0753986984,-0.1135911867,-0.1683868319,0.6030051112,0.0085332366,0.0657512322,0.1192067564,-0.1418523192,0.0852058902,0.4310850203,-0.4683701992,-0.0725180283,-0.3837422729,0.0691385418,0.1228722557,0.1599069983,0.554443121,0.5351935625,-0.1123419926,0.0913888812,0.1016571522,-0.0620642416,0.0256304182,0.0167524107,-0.1384332329,-0.1652918905,0.2106099725,0.0366320275,0.0164089836,0.0945241526,0.2368489057,0.4203828871,-0.0320598595,0.1690512598,-0.1136381403,-0.22543253,-0.2443024218,0.2671291232,-0.3228886127,-0.1418663114,0.381688714,0.180131346,0.2626311481,-0.1505586654,0.0469924398,0.2338539362,0.1337417364,0.1655443311,-0.1106802151,-0.5532167554,-0.1350879818,-0.6108707786,0.1502811611,-0.0991240591,0.1134090722,0.0417402051,0.0030684597,-0.0390887856,0.1086745113,0.1402222961,-0.1363838166,-0.2191719711,0.3542179167,-0.1631863117,-0.1612084359,0.041701097,0.1446032226,0.4254318178,-0.2717137039,0.0298630688,0.0203138236,-0.0150781032,-0.3144214451,-0.2270759046,0.0902834088,-0.0580685586,0.648899734,0.0289296936,0.2742853761,0.0063963928,-0.2149114907,-0.0472533479,0.1743014306,-0.2483456433,-0.1251496524,-0.1784656346,0.2291238606,-0.3856540322,-0.1919009835,-0.5190024972,-0.1139736325,-0.059141092,-0.1711239517,-0.1509468108,0.2457058579,-0.0137111228,0.2305443287,0.1110587716,0.437844336,-0.2536054552,0.6320556402,-0.1998130083,-0.0673571005,0.5248292089,-0.6005620956,-0.4457896948,0.1041662022,0.2797317207,-0.1180543154,-0.4135190845,-0.4719399214,-0.1909065992,0.438302964,0.0340473615,-0.324711293,-0.015911486,-0.0834916979,0.109438926,0.0421535559,0.4812168777,0.0117099779,-0.1895609647,0.3745303452,-0.3594847918]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2407","title":".map() function got an unexpected keyword argument 'cache_file_name'","comments":"I think you are right. I used cache_file_names={data1: name1, data2: name2} and it works. Thank you!","body":"## Describe the bug\r\n\r\nI'm trying to save the result of datasets.map() to a specific file, so that I can easily share it among multiple computers without reprocessing the dataset. However, when I try to pass an argument 'cache_file_name' to the .map() function, it throws an error that \".map() function got an unexpected keyword argument 'cache_file_name'\". \r\n\r\nI believe I'm using the latest dataset 1.6.2. Also seems like the document and the actual code indicates there is an argument 'cache_file_name' for the .map() function.\r\n\r\nHere is the code I use\r\n## Steps to reproduce the bug\r\n```datasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n cache_file_name=\"my_tokenized_file\"\r\n)\r\n```\r\n\r\n## Actual results\r\n tokenized_datasets = datasets.map(\r\nTypeError: map() got an unexpected keyword argument 'cache_file_name'\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.6.2\r\n- Platform:Linux-4.18.0-193.28.1.el8_2.x86_64-x86_64-with-glibc2.10\r\n- Python version:3.8.5\r\n- PyArrow version:3.0.0\r\n","comment_length":16,"text":".map() function got an unexpected keyword argument 'cache_file_name'\n## Describe the bug\r\n\r\nI'm trying to save the result of datasets.map() to a specific file, so that I can easily share it among multiple computers without reprocessing the dataset. However, when I try to pass an argument 'cache_file_name' to the .map() function, it throws an error that \".map() function got an unexpected keyword argument 'cache_file_name'\". \r\n\r\nI believe I'm using the latest dataset 1.6.2. Also seems like the document and the actual code indicates there is an argument 'cache_file_name' for the .map() function.\r\n\r\nHere is the code I use\r\n## Steps to reproduce the bug\r\n```datasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n cache_file_name=\"my_tokenized_file\"\r\n)\r\n```\r\n\r\n## Actual results\r\n tokenized_datasets = datasets.map(\r\nTypeError: map() got an unexpected keyword argument 'cache_file_name'\r\n\r\n## Environment info\r\n\r\n- `datasets` version:1.6.2\r\n- Platform:Linux-4.18.0-193.28.1.el8_2.x86_64-x86_64-with-glibc2.10\r\n- Python version:3.8.5\r\n- PyArrow version:3.0.0\r\n\nI think you are right. I used cache_file_names={data1: name1, data2: name2} and it works. Thank you!","embeddings":[0.1506065279,0.0931876749,0.046234984,-0.0068933293,0.049635455,0.2938355803,0.2111461759,0.3586550951,-0.06443993,0.0176359806,0.1194801331,0.6066251397,-0.2475739419,-0.4879621267,0.0684003457,0.1206603795,0.414899081,-0.0868131071,0.2050106525,0.0021265608,-0.3779777288,0.3605404794,-0.0720000491,0.2415003031,-0.4130324721,-0.1476259381,-0.1109034866,-0.0780590028,-0.0512373336,-0.3494436741,0.1783299595,0.1331271678,-0.089754872,0.4608037174,-0.0001248522,-0.0772017464,0.0131908813,-0.189556241,-0.1999052316,-0.3421218097,-0.1010760218,0.0225374438,-0.0683813095,-0.5165248513,0.2275047153,-0.1730200946,0.2001903653,-0.5802561641,0.1264273971,0.1558328122,0.1211594194,-0.1118961796,-0.2370792776,-0.1584720612,0.0761795342,-0.1248546541,-0.0106951464,0.1948423535,0.0108601814,-0.2778089643,0.1566221267,0.3239407837,-0.4211137295,-0.0880382657,0.4996004999,0.0565097071,-0.220529899,-0.3682149351,0.2468452156,-0.1557586044,0.160047695,-0.4097859859,-0.3636485338,-0.1235275194,-0.3060948849,-0.1057340354,0.1689180583,-0.0969173983,-0.1287402064,-0.0269522741,0.0185058676,-0.2359417677,0.0517394058,0.1558367908,0.2515682876,0.0878491476,-0.1667111367,0.3883554041,0.3064814508,-0.0860620886,-0.2235984504,-0.4085984528,-0.030748941,0.3572908044,-0.184774816,-0.0417935438,0.0571031123,0.0732770637,0.0752723292,-0.3570677042,-0.1044029444,-0.1761839241,0.1044635549,0.3414004147,0.239831984,0.2543738782,0.3329280317,0.4893091917,-0.1597922146,-0.3360438347,-0.311604172,-0.2010548264,0.0660401359,-0.096471779,0.5460054874,-0.0766859651,0.2803613842,-0.102068536,0.1100974455,0.0750942156,-0.3782733083,-0.012523056,-0.0800797418,0.3749354184,0.1374799162,-0.0799544752,-0.0131762614,0.1978539377,0.0574048348,-0.1682079285,-0.1408262104,0.088381663,-0.2228275836,0.2366519272,-0.1865341961,-0.0189995598,0.1354494244,0.0265366361,0.1726419181,-0.1520012319,0.2027388215,-0.0868520737,0.2472155839,-0.1467598379,-0.0666591749,0.5224140882,0.2376119941,-0.2446204424,-0.1467581391,0.0654154345,-0.519585371,-0.288936466,-0.000584065,0.0114896651,-0.0296990406,0.1388429105,-0.4297335446,0.0088157123,0.5321179032,-0.3010300696,-0.0069852872,-0.0502438992,-0.2498446405,-0.5160012841,0.0298702642,0.3006653488,-0.0805004016,-0.032352604,0.0187519081,0.2007806152,0.0075065377,0.1544819474,-0.0984018594,0.2479175478,-0.1984270662,0.6995722055,0.5317043662,-0.2337763011,-0.5732627511,0.2688948214,-0.1798194349,-0.0256475136,-0.0756236315,0.0339843743,0.4914753139,-0.1528477967,0.4469078183,-0.0536132418,-0.0290826,0.192588523,-0.2574965954,-0.2234037519,0.1693952978,-0.2530673146,-0.0420028828,0.1683845967,0.073596552,0.0956488624,0.100527972,0.0775981769,0.2457409054,0.0606872328,0.2730278373,0.0878141597,0.0292353295,0.0117293689,-0.2583968937,0.0582039654,0.2064500004,-0.449622184,-0.2803353667,-0.4923896492,-0.0269086044,0.0401753038,-0.1573172957,-0.0334996209,-0.003660334,0.0983634219,0.0660050213,0.1121909395,-0.0879581571,-0.0005152253,0.0636178255,-0.0778424814,0.0891707167,-0.2292373031,-0.0712287501,-0.128863588,-0.2181703746,-0.2020585835,0.4292916954,-0.0318069942,-0.2988354564,0.3882876039,-0.0271624662,0.0665292367,-0.0784718245,0.3237133622,-0.3152004182,-0.0091996025,0.0507583246,-0.0057599894,0.1366644204,-0.0144883664,0.0982632786,0.1785289943,0.3157641292,0.1690732986,-0.2228194475,0.3009205759,-0.0301989783,0.0125484364,0.0300274026,-0.2449702471,-0.0398891121,-0.1226134375,0.3339978755,-0.1022685394,-0.0005504786,0.0352188759,0.5029191375,0.1220408231,0.1732726097,-0.0695688948,-0.1503664851,-0.0012800422,0.0664687455,0.3170949221,0.5943683386,-0.0066189603,0.0173457023,-0.0088958219,0.1392169744,0.1031204611,0.0691981465,0.0365273133,-0.1297702789,-0.2433543503,-0.0627132431,0.2150774449,0.0854618847,0.0491771139,0.2067358196,0.1132831797,-0.2130065709,0.2950794399,-0.0789687857,0.0191499013,-0.075493522,-0.0520404354,-0.2320698649,-0.4821262956,-0.0861163065,0.2031510174,0.0416536294,0.2774496377,-0.1584031582,0.1630256176,-0.0084109399,-0.4518019557,0.0133335823,-0.2092440873,-0.043516513,-0.2183418721,0.0205727592,-0.3122446537,0.4293563962,-0.0221570786,-0.0626373515,-0.3932145834,-0.1321213841,0.1364067942,-0.132469818,0.1697255224,0.1360776573,0.1069907397,-0.284819752,-0.1963796169,0.1869968027,-0.2661243677,0.0500925705,-0.0728318393,0.1140383333,-0.2556231916,-0.1344592273,-0.0887681097,-0.0870846584,-0.0854689404,0.268052429,-0.3216689527,0.1308672875,0.0859511495,-0.0522363074,0.1807772517,-0.2963650227,-0.1283027381,-0.3490850627,-0.4505206943,0.4080706537,-0.2937375605,-0.2143178582,0.0160292219,0.0618532076,0.1160241738,0.237045154,-0.2521747947,-0.2652908862,0.1330054551,0.2280884087,0.0131181097,-0.0057230312,0.2380248308,0.0031640935,0.0866204724,-0.3071552813,-0.4675841033,0.1193784252,0.1409864128,0.3189295232,0.3793570101,0.2845683098,0.3596894145,0.6702024341,0.3310705125,-0.005253714,0.425998807,-0.1899456233,0.3477729857,-0.0295606814,-0.5100730658,-0.0894977972,-0.0939289406,-0.2986631989,-0.0397361927,-0.11994721,-0.155311361,0.0920087099,-0.1217898726,-0.2851125002,-0.4131676555,0.0624878258,-0.3208054304,0.1612466872,0.0793834403,0.1429513097,-0.6122845411,-0.1216693074,0.0658832714,0.2694426179,0.1453983486,0.1211967319,-0.3798371553,-0.0284326915,-0.4286347628,0.5219662786,0.0320536159,0.3255200088,-0.1689075381,0.1328068972,-0.0162343141,0.2232998759,0.7702389956,-0.300186038,0.0226608943,0.1039687395,0.2140932977,-0.2086614072,0.0142739872,0.192670092,0.8246279359,0.164061442,0.4790226519,-0.1594995856,-0.1409149319,-0.2343809158,0.1006201953,-0.1453937143,-0.0247448552,-0.0478684939,-0.0961720422,-0.4246598482,-0.050325796,0.1972025037,-0.0002180763,0.357405901,0.1722485572,0.1264429539,0.0527610183,0.2019991726,0.1873435974,0.3689807355,-0.0163777322,-0.12771523,0.2373881787,-0.0350011066,0.1215638593,0.1370646507,-0.2305275947,-0.1081721634,-0.1359265596,-0.1148120761,0.4389282167,0.1496143639,-0.2844192684,0.0829154924,-0.1337013692,-0.0319950879,-0.0172156803,0.212102145,0.4726811349,-0.078292422,-0.5838581324,-0.5335868597,0.2385367006,0.307476759,-0.2767955661,0.8010517359,0.0006075576,-0.319924742,0.4546947777,-0.1077264845,0.8967576623,-0.0293937698,0.2561295033,-0.148526758,0.1116107479,0.365603894,-0.0295310244,-0.0222608857,-0.1998390406,0.0409981646,0.0428980179,0.0122949658,0.0913725793,0.4195532501,-0.3640224934,0.267534405,-0.0755035207,0.4595959485,-0.0435271487,-0.0055507822,-0.0019312092,-0.2061171681,-0.0843581557,0.1025355384,-0.0380674414,0.0937514752,0.0043151961,-0.1688732058,0.0125750815,-0.4977371097,-0.155701369,0.0034702518,-0.535148263,0.3101593554,-0.1394639909,0.0956991315,-0.0343221501,0.3570439219,0.0853720233,0.2437949181,-0.0573577732,-0.0332692899,0.0785985962,0.0092070987,-0.1305179447,-0.2771018744,0.1249171123,0.1810337901,-0.0051080766,0.1753000021,-0.025132779,-0.519482851,-0.2173883319,0.0601097234,0.2489191145,-0.3463416994,-0.4690988362,-0.176880613,-0.1415276825,-0.0865632966,0.0110344673,0.0197935179,-0.1361038089,0.3602216244,-0.220791012,-0.2241202891,0.2638535202,0.3675530553,-0.3189499676,-0.2010809332,0.4495681226,-0.3525358438,-0.0513463691,-0.0735214874,0.0972588509,-0.1046672836,0.0384297706,0.203956604,-0.1828582883,0.197828576,-0.1194085106,0.1648614854,0.1596924216,0.2241194844,0.0317611732,-0.348198384,-0.2153420299,0.0479376987,-0.1564866006,0.1336771101,0.2249795645,-0.0118741291,0.0472120531,0.1785405278,-0.1908364594,0.1848515719,-0.0267697815,0.0882775262,0.3773245811,0.3435542583,0.2894652188,-0.201167658,-0.040561799,0.1854470372,-0.1447353959,-0.0630849749,-0.3401402235,0.2215380967,0.2339004874,0.0338977166,0.0786913559,-0.349947989,0.0929036066,-0.3637593091,0.0423919596,0.4307113588,-0.0725799948,0.3073854148,0.3182022274,0.1126613989,0.0097082723,0.0585851409,-0.0421373136,0.1638426632,0.023089271,0.2250066102,0.0366747081,0.0061762128,-0.1736589372,0.1346369833,-0.0644849166,0.0658219829,0.158786267,-0.1427582353,0.1276531816,0.4364624023,0.0293166284,0.1625669897,-0.5086873174,-0.0070617567,0.2376177609,0.050873287,0.1090688109,-0.260699451,0.8278591037,0.1090266258,-0.1083182395,0.23992531,0.0589891076,0.0016378475,-0.1321201921,-0.1859840751,0.4091364145,-0.4331194758,0.3383302093,0.288038671,0.2341212332,0.3518940508,0.2614619434,0.370103389,0.1764348149,0.5423188806,-0.2519457936,0.2570311725,0.2938836217,-0.1519248039,0.1742676944,-0.2455426157,0.2053670585,0.1600109339,-0.1853313148,0.4770709276,-0.1064114496,0.3385760784,-0.3598348796,-0.0561264195,0.1143939123,0.1950637996,-0.2536484599,-0.0618436001,0.0725082234,0.0098166512,-0.0620443895,-0.0928367451,-0.1013494432,-0.188324362,0.6053315997,0.0192473494,0.0836117566,0.1208087876,-0.1201506779,0.1298997849,0.4147449434,-0.4476347864,-0.0883442909,-0.3582007289,0.0668256879,0.1162706688,0.1713826507,0.5463880897,0.5371773243,-0.0913759246,0.1108284742,0.0789530054,-0.075906381,0.0481178313,0.0418879278,-0.1036809608,-0.181725949,0.1953211278,0.024169229,0.0287163518,0.116941072,0.2279562801,0.4599262476,-0.0382427946,0.1489447355,-0.1141954362,-0.2389262468,-0.2414095253,0.2638718188,-0.342925638,-0.1206549257,0.3887702227,0.2108164728,0.2338464856,-0.1618371159,0.0352787226,0.2122348696,0.0910511687,0.1747748852,-0.0927336514,-0.5577592254,-0.1271519214,-0.5940537453,0.1552420855,-0.0787497535,0.1418318152,0.029293336,0.0033742401,-0.0001279065,0.1125844046,0.1446472555,-0.1640222073,-0.242003113,0.3722458482,-0.1334304512,-0.1698625833,0.0288319439,0.1734134406,0.4382997155,-0.245284602,0.0051156734,-0.0184462313,-0.0242949408,-0.3747481704,-0.1871188432,0.0896143541,-0.0573669039,0.6901994348,0.0102396216,0.2794249654,0.0101226661,-0.2189331353,-0.0119489152,0.1673456877,-0.2361444682,-0.1449659914,-0.1862489432,0.2186086476,-0.4118637145,-0.2017904222,-0.5065118074,-0.1514182687,-0.0831607804,-0.2064054608,-0.1174833477,0.2421707511,0.0191765167,0.2319876552,0.0684673041,0.4040102065,-0.2606873512,0.6287738085,-0.2165254951,-0.0439625531,0.5280613899,-0.6110010147,-0.4437100291,0.1085877717,0.2843797803,-0.1579755396,-0.4147579968,-0.4719444513,-0.166752845,0.448336184,0.0761113688,-0.3220528364,0.0097972872,-0.1244867072,0.1342135519,0.0622292235,0.4502387643,-0.0207829978,-0.1780711114,0.2883855999,-0.3513783216]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2401","title":"load_dataset('natural_questions') fails with \"ValueError: External features info don't match the dataset\"","comments":"I faced the similar problem. Downgrading datasets to 1.5.0 fixed it.","body":"## Describe the bug\r\nload_dataset('natural_questions') throws ValueError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndatasets = load_dataset('natural_questions', split='validation[:10]')\r\n```\r\n\r\n## Expected results\r\nCall to load_dataset returns data.\r\n\r\n## Actual results\r\n```\r\nUsing custom data configuration default\r\nReusing dataset natural_questions (\/mnt\/d\/huggingface\/datasets\/natural_questions\/default\/0.0.2\/19bc04755018a3ad02ee74f7045cde4ba9b4162cb64450a87030ab786b123b76)\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in \r\n----> 1 datasets = load_dataset('natural_questions', split='validation[:10]', cache_dir='\/mnt\/d\/huggingface\/datasets')\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 756 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\r\n 757 )\r\n--> 758 ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n 759 if save_infos:\r\n 760 builder_instance._save_infos()\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in as_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 735 \r\n 736 # Create a dataset for each of the given splits\r\n--> 737 datasets = utils.map_nested(\r\n 738 partial(\r\n 739 self._build_single_dataset,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 193 # Singleton\r\n 194 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 195 return function(data_struct)\r\n 196 \r\n 197 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _build_single_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 762 \r\n 763 # Build base dataset\r\n--> 764 ds = self._as_dataset(\r\n 765 split=split,\r\n 766 in_memory=in_memory,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _as_dataset(self, split, in_memory)\r\n 838 in_memory=in_memory,\r\n 839 )\r\n--> 840 return Dataset(**dataset_kwargs)\r\n 841 \r\n 842 def _post_process(self, dataset: Dataset, resources_paths: Dict[str, str]) -> Optional[Dataset]:\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint)\r\n 271 assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n 272 if self.info.features.type != inferred_features.type:\r\n--> 273 raise ValueError(\r\n 274 \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n 275 self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'id': Value(dtype='string', id=None), 'document': {'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None), 'html': Value(dtype='string', id=None), 'tokens': Sequence(feature={'token': Value(dtype='string', id=None), 'is_html': Value(dtype='bool', id=None)}, length=-1, id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': Sequence(feature={'id': Value(dtype='string', id=None), 'long_answer': {'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None)}, 'short_answers': Sequence(feature={'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}, length=-1, id=None), 'yes_no_answer': ClassLabel(num_classes=2, names=['NO', 'YES'], names_file=None, id=None)}, length=-1, id=None)}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>>, id: string, question: struct>>\r\n\r\nbut expected something like\r\n{'id': Value(dtype='string', id=None), 'document': {'html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'tokens': {'is_html': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'token': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'url': Value(dtype='string', id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': {'id': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'long_answer': [{'end_byte': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'start_token': Value(dtype='int64', id=None)}], 'short_answers': [{'end_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'end_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'text': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}], 'yes_no_answer': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None)}}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>, url: string>, id: string, question: struct>>\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n- Python version: 3.8.3\r\n- PyTorch version (GPU?): 1.6.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n","comment_length":11,"text":"load_dataset('natural_questions') fails with \"ValueError: External features info don't match the dataset\"\n## Describe the bug\r\nload_dataset('natural_questions') throws ValueError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndatasets = load_dataset('natural_questions', split='validation[:10]')\r\n```\r\n\r\n## Expected results\r\nCall to load_dataset returns data.\r\n\r\n## Actual results\r\n```\r\nUsing custom data configuration default\r\nReusing dataset natural_questions (\/mnt\/d\/huggingface\/datasets\/natural_questions\/default\/0.0.2\/19bc04755018a3ad02ee74f7045cde4ba9b4162cb64450a87030ab786b123b76)\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in \r\n----> 1 datasets = load_dataset('natural_questions', split='validation[:10]', cache_dir='\/mnt\/d\/huggingface\/datasets')\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 756 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\r\n 757 )\r\n--> 758 ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n 759 if save_infos:\r\n 760 builder_instance._save_infos()\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in as_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 735 \r\n 736 # Create a dataset for each of the given splits\r\n--> 737 datasets = utils.map_nested(\r\n 738 partial(\r\n 739 self._build_single_dataset,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 193 # Singleton\r\n 194 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 195 return function(data_struct)\r\n 196 \r\n 197 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _build_single_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 762 \r\n 763 # Build base dataset\r\n--> 764 ds = self._as_dataset(\r\n 765 split=split,\r\n 766 in_memory=in_memory,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _as_dataset(self, split, in_memory)\r\n 838 in_memory=in_memory,\r\n 839 )\r\n--> 840 return Dataset(**dataset_kwargs)\r\n 841 \r\n 842 def _post_process(self, dataset: Dataset, resources_paths: Dict[str, str]) -> Optional[Dataset]:\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint)\r\n 271 assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n 272 if self.info.features.type != inferred_features.type:\r\n--> 273 raise ValueError(\r\n 274 \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n 275 self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'id': Value(dtype='string', id=None), 'document': {'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None), 'html': Value(dtype='string', id=None), 'tokens': Sequence(feature={'token': Value(dtype='string', id=None), 'is_html': Value(dtype='bool', id=None)}, length=-1, id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': Sequence(feature={'id': Value(dtype='string', id=None), 'long_answer': {'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None)}, 'short_answers': Sequence(feature={'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}, length=-1, id=None), 'yes_no_answer': ClassLabel(num_classes=2, names=['NO', 'YES'], names_file=None, id=None)}, length=-1, id=None)}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>>, id: string, question: struct>>\r\n\r\nbut expected something like\r\n{'id': Value(dtype='string', id=None), 'document': {'html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'tokens': {'is_html': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'token': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'url': Value(dtype='string', id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': {'id': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'long_answer': [{'end_byte': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'start_token': Value(dtype='int64', id=None)}], 'short_answers': [{'end_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'end_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'text': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}], 'yes_no_answer': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None)}}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>, url: string>, id: string, question: struct>>\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n- Python version: 3.8.3\r\n- PyTorch version (GPU?): 1.6.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\nI faced the similar problem. Downgrading datasets to 1.5.0 fixed it.","embeddings":[-0.3010365963,-0.1931253821,-0.0652654469,0.525786221,0.2114789933,-0.1320342124,0.4025248289,0.3253356814,0.1363790929,0.0655856878,-0.1739462912,0.3462786973,-0.1743816137,0.3402386606,0.0053835479,-0.0349213295,-0.033921551,0.0606687069,-0.0714647621,0.0316084325,-0.1666446477,0.1117764711,-0.1739413589,0.054349374,-0.1699436158,0.0612640306,0.0389056914,0.368008852,0.0395361893,-0.4640865624,0.1402531266,-0.2512812614,0.200597018,0.3641215563,-0.0001097937,0.1101842001,0.291016221,-0.1340754777,-0.4646355808,-0.3495002687,-0.4326409698,-0.2932563126,0.1844670177,-0.1025997847,-0.0750925839,-0.0901743248,-0.1942299753,-0.1198888496,0.2101251334,0.5170678496,0.2509104908,0.437240392,0.1481525749,-0.1356033385,0.4015932977,0.2591476738,-0.1198510081,0.3900898993,-0.1157002077,-0.1915698498,0.1298338175,0.1801397651,-0.0604048744,0.3196884692,0.3086515367,0.0979741514,0.0387078598,-0.2461425811,0.0445514359,0.2624997497,0.5269556642,-0.2871050835,-0.2662248313,-0.3345457017,-0.1267289966,-0.3140507936,0.1942808628,0.1228377894,-0.2235041857,0.1248339564,-0.068113409,0.0506800823,-0.1982334703,0.1594713777,-0.1901111454,0.0679849163,-0.1613807976,0.0686992332,-0.2411332875,-0.2050860971,0.2681346238,-0.1306806505,-0.040640682,-0.0284468103,-0.158993572,-0.1046421081,0.0473441184,-0.2293193638,0.1360967755,0.3597458601,-0.0197180305,-0.0360403508,0.0957899392,0.2053233534,0.2986126244,0.177685976,0.1130070016,0.3782274723,0.1557691097,0.3454990983,-0.1647492349,-0.0209359992,0.0980151296,-0.0155570917,0.3737146258,0.0471422449,0.6195658445,-0.3021340966,-0.3641215265,0.2766881585,0.0009493929,-0.0309526678,0.0115113985,0.34006688,-0.2023818344,0.1892527938,-0.0823120698,0.1425789297,0.0140412003,-0.2875742018,-0.2640197575,-0.0711601377,-0.014275156,0.1304302961,0.1474407762,-0.0116192913,0.5592864156,0.0787607655,0.0984397903,-0.0673231632,0.1605854183,-0.1946494579,-0.1344540566,0.2918545604,-0.209122315,0.0362084657,0.4381088614,-0.0508765057,-0.0775992349,-0.0562448762,-0.1957972497,-0.4581328928,-0.1711842418,0.2804891765,-0.3117066324,0.0094732922,0.2535792291,-0.1273874938,0.3626094162,-0.2142152637,-0.1140045077,-0.2887541056,-0.2575203478,-0.2485756278,0.2768083215,0.4994139075,-0.3477646112,0.0705651119,0.1204925999,-0.1663465947,0.1035433859,0.0806002319,-0.1228351295,0.3711713552,-0.1696024537,-0.1638680995,0.1695781052,-0.2986466289,-0.4003855586,0.3155426979,-0.0153261796,0.2871543467,0.0063026901,0.0271411482,0.1573456824,0.1157109886,0.4620014727,0.4014619291,-0.066910401,-0.0074357758,-0.2804535925,-0.2457351685,0.1595118046,0.2452587187,0.0332712159,0.06809894,0.2984835804,0.0001923513,0.0488624536,-0.0455991328,-0.2236338407,0.1471460015,0.3994964957,0.0760490149,-0.1120455712,-0.404699862,-0.5526606441,0.3127443492,-0.3151362538,0.0898777246,-0.0624681339,0.0992171317,-0.6718125939,0.1659998745,-0.2465763986,-0.3998812437,0.1665603518,0.3555616736,-0.168122068,-0.0031244948,-0.1956553459,0.4240124524,-0.2365444303,0.1162518114,-0.3416495919,0.2825235426,-0.0393327996,-0.149338156,0.0628314167,0.0276369713,0.167953372,0.0517680161,-0.2381964028,0.4361684024,0.2144528329,0.0494904742,-0.2391491681,-0.1940413862,0.038697008,-0.2645381391,0.163651675,0.2008208781,0.1864885539,-0.0701437369,-0.1489441991,0.3302778304,0.0553814918,0.1013679281,-0.0695879534,-0.0139205735,0.0889150873,-0.0030233478,-0.0507578999,-0.2198005319,0.2232966572,-0.0017620587,0.2004920393,0.0682676956,-0.3529761434,0.0907458141,0.2752231658,0.0731488168,0.1277240962,0.1467070132,-0.2322896719,-0.1234295219,0.2753238678,0.0469856486,0.6293234229,0.1825696975,-0.165564239,0.2986156046,-0.15912202,-0.0204912219,0.2052815408,-0.0421392247,0.1960897297,0.2865037024,0.1824352592,0.1393212676,-0.178505227,-0.0943207219,-0.1112810969,0.1985169053,-0.4880636632,0.079517372,-0.1686079204,0.0649334341,-0.0422706679,-0.1734807342,-0.2261099219,-0.1904407442,-0.1937994361,0.2278338224,-0.1691826135,0.2074276656,-0.2505809069,-0.0847058743,0.1458775997,-0.526855588,0.089645423,-0.0621372499,-0.2492348999,0.0662923008,0.2406881154,0.1643317789,0.267888993,-0.20216088,-0.151760608,-0.2189625651,-0.1686671227,-0.0300829411,0.074105829,0.770889461,0.1303961128,0.2613179088,0.379313767,-0.2804147601,0.2815556228,0.0498569459,-0.1869068891,0.0964192227,0.064125143,0.0366347134,-0.0271369424,-0.6000498533,-0.36951527,-0.5040061474,-0.0255676452,0.2252414972,0.1044660807,0.5914885998,0.2259821892,0.098211661,0.0685926825,0.2290950865,-0.2079316676,-0.1246068254,0.2001141459,-0.1017325744,-0.1273368746,-0.316460073,-0.0996490195,0.1105185002,0.0181393623,-0.5311613679,-0.0326023139,-0.3451296091,0.4823538363,-0.1425576508,0.0308493003,0.3642105162,-0.063392967,-0.0323227346,-0.0954335257,-0.2647781968,0.0323960744,-0.0000882497,0.2433171421,0.0356643125,0.4075246155,-0.0353547633,0.6303832531,0.0092987791,-0.0344562195,0.3832820654,-0.1496753991,0.416369319,-0.3017952442,-0.5337983966,0.0131714232,-0.2167703509,-0.1959250569,0.1570640951,-0.0784802437,-0.1687743813,-0.1831905693,-0.0932203308,-0.3046920598,-0.0606943481,0.0384095423,-0.0187649019,0.1651714593,0.1132292226,0.1195074767,-0.2817521989,-0.0889264867,-0.1094715968,0.0146534182,-0.0797828808,-0.0825566798,-0.4356510639,0.2557926178,-0.1198370233,0.2290960848,0.1021284238,0.3126807511,-0.116187349,-0.3825466335,-0.1960823387,-0.0726532489,0.2147463858,0.0184873715,0.0753258839,0.1536622196,-0.1412212104,-0.3610154688,-0.2004318684,-0.1549527943,0.113098599,0.2339238226,0.6427520514,-0.1891706288,0.0071767429,0.3524083495,0.4035503864,-0.1031610444,-0.2549925745,-0.2080092579,-0.1045188084,-0.3992632627,-0.0318549238,-0.1161562353,0.2455658913,-0.0078822803,0.0437904298,-0.1692227125,-0.0803853348,0.165555656,0.2271063626,0.1503321826,-0.1269869804,0.284586221,-0.1758479029,0.1441669017,0.0596954003,0.6913921237,0.0657192096,-0.5916351676,-0.0717137083,0.0396200567,0.1844923943,0.2397723347,0.0448651798,-0.0272457544,-0.0495887175,0.0868238807,-0.0873017907,0.3527728617,0.0881813541,-0.121158205,0.1119030043,-0.2668753564,0.3709553778,-0.1393034011,-0.0502592623,0.2567106485,-0.1017212197,-0.1936982721,-0.0222703051,0.1220172495,0.6600776315,0.3586020172,0.1172699258,0.4765803516,-0.1983785629,0.3195495307,-0.0597341172,-0.1360106617,-0.4920931756,-0.3510355353,0.1049575582,-0.1960784346,0.1987957358,-0.1476992816,-0.0226633139,0.3582038581,-0.0114521012,0.4813610911,0.181717664,0.1715059429,-0.0177972503,-0.024297215,-0.2199698389,0.187038362,-0.0949707925,0.1076645777,-0.0729456171,-0.039432548,-0.0519838631,-0.1821260899,-0.2247526646,0.1746825129,-0.3567138612,0.092994228,0.2788364887,-0.3486048877,-0.1215511039,0.2856612504,0.1725815684,0.1285102963,-0.1277089417,-0.124090977,0.0180381462,0.0834357515,0.0509496666,-0.0223008767,0.3129682243,-0.1497761309,-0.3154302835,0.17894198,-0.1141532063,-0.1367426813,0.0276512485,0.0460858345,0.0429792181,-0.6083951592,-0.3587333858,-0.2729334235,0.2343661785,-0.2964299023,0.1671108007,-0.0850273147,-0.159299016,-0.0390192419,-0.0390178151,-0.3504946232,-0.188681826,0.2096121609,0.0100413337,0.0955855548,0.4750776589,-0.0024073629,0.085909158,-0.281863451,0.5697032809,-0.1319766641,-0.2841341496,0.1106292233,0.0628202036,-0.1130802184,-0.2623446584,0.1433285475,0.2304039598,-0.0656383932,0.2568435073,-0.4319812059,-0.3417362571,0.1113909185,-0.3017697036,0.2283263505,0.0737322196,-0.0314689614,-0.1762896925,0.1253466308,-0.371548444,-0.0490871407,-0.2450396568,0.1156439707,0.1753619909,-0.0509934947,0.0549703278,0.0774088129,0.2249481529,-0.0035852455,-0.1168234944,-0.3020421863,-0.0865272358,0.1254264265,0.0515467748,-0.0695332363,-0.1507494003,-0.0611952282,-0.1140329838,-0.1327302009,0.0390418172,0.1042219698,-0.0347810686,0.3725389242,-0.1999056935,0.2814030349,-0.1046492755,0.2537442446,-0.1507167965,0.1896493137,-0.2035243213,0.1280552447,-0.1644871533,-0.000014853,-0.2832544148,0.0390093923,-0.0349154696,0.0817560554,0.2742252648,-0.2003487498,-0.2600533068,0.0305856019,0.0199263673,0.3129118681,-0.2150289416,0.1065102741,0.1815644205,0.2663259804,-0.3734115958,-0.010485691,0.1419816762,-0.1060630083,0.0743484572,0.2163880616,0.2126561552,-0.0611200705,0.2493149489,0.0930695385,0.2510064244,0.0173221454,0.0926623568,0.3244889081,-0.2003924251,-0.043881692,0.2114525735,0.1855243593,0.2891336381,0.4851191044,-0.079485178,0.0525735989,-0.2255954295,0.0710886717,0.0361781418,-0.3354212046,0.0672081336,0.0693156421,-0.0936621055,-0.00470267,-0.1472409517,0.0748917535,-0.2594290078,0.0290211085,-0.2763265669,-0.0886882246,-0.0123605458,-0.017395312,-0.2431390733,-0.2479863018,-0.1207344383,-0.053788323,-0.0487490408,-0.1577880383,0.1042184457,-0.1172417551,-0.3449355066,-0.4382255077,0.1962499619,0.4863179624,-0.187797606,-0.3739557862,0.3553467691,0.1638759375,0.1163184866,-0.0050221337,0.2935783565,0.5776576996,0.4492668509,-0.2183886468,-0.0133697409,-0.0903499722,-0.2572478652,-0.08058808,0.1788306087,-0.2036101818,0.2041838914,0.3896346092,0.2425412089,-0.2860693038,0.3087479472,0.0671433136,-0.014772282,-0.6452242136,0.3085484207,0.0373875014,0.0155810714,0.1040226147,-0.0377722569,-0.2956029773,-0.0928971395,0.4650352895,0.2191892415,0.2604660094,0.0643747449,0.1007259563,-0.2586954534,0.2847928107,0.3491569161,0.2334001511,-0.302959919,0.1163394004,-0.6907718182,0.1602863669,-0.0385213606,-0.1857215613,0.145208776,0.1284459084,0.189406991,0.3235723674,-0.2789386511,0.1875156462,0.1158161089,-0.102913402,-0.3438047171,-0.2960200012,-0.3301110864,0.1387993097,0.0364331864,-0.4284615219,0.0283448994,-0.2654493451,0.0856036842,-0.0148275346,-0.1632930338,0.2235354483,0.1154315844,0.4646860659,0.1625410169,0.4427508414,-0.0643838346,-0.0877063125,-0.1593246162,-0.326819092,-0.2241960913,0.0881990641,0.0083235903,0.6604252458,-0.0688739419,-0.186289534,-0.2489366531,0.547285974,0.1704099625,-0.1335785091,0.1080821976,0.0587932207,0.005058486,0.016521344,0.0720723271,0.086332269,-0.1636992246,0.2174746543,-0.0827195421,-0.2598025203,0.4770698547,-0.0727573186,-0.3654350936,-0.0709003955,0.2040002495,-0.0339087285,-0.0137331747,-0.373477608,0.1011847407,0.5264702439,-0.0827589035,-0.2293682843,0.1287205517,0.0001831334,0.119935222,-0.0393447988,0.4087265432,-0.0403168648,-0.1360573173,0.0768616349,-0.1536669582]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2401","title":"load_dataset('natural_questions') fails with \"ValueError: External features info don't match the dataset\"","comments":"Thanks for reporting, I'm looking into it","body":"## Describe the bug\r\nload_dataset('natural_questions') throws ValueError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndatasets = load_dataset('natural_questions', split='validation[:10]')\r\n```\r\n\r\n## Expected results\r\nCall to load_dataset returns data.\r\n\r\n## Actual results\r\n```\r\nUsing custom data configuration default\r\nReusing dataset natural_questions (\/mnt\/d\/huggingface\/datasets\/natural_questions\/default\/0.0.2\/19bc04755018a3ad02ee74f7045cde4ba9b4162cb64450a87030ab786b123b76)\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in \r\n----> 1 datasets = load_dataset('natural_questions', split='validation[:10]', cache_dir='\/mnt\/d\/huggingface\/datasets')\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 756 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\r\n 757 )\r\n--> 758 ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n 759 if save_infos:\r\n 760 builder_instance._save_infos()\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in as_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 735 \r\n 736 # Create a dataset for each of the given splits\r\n--> 737 datasets = utils.map_nested(\r\n 738 partial(\r\n 739 self._build_single_dataset,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 193 # Singleton\r\n 194 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 195 return function(data_struct)\r\n 196 \r\n 197 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _build_single_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 762 \r\n 763 # Build base dataset\r\n--> 764 ds = self._as_dataset(\r\n 765 split=split,\r\n 766 in_memory=in_memory,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _as_dataset(self, split, in_memory)\r\n 838 in_memory=in_memory,\r\n 839 )\r\n--> 840 return Dataset(**dataset_kwargs)\r\n 841 \r\n 842 def _post_process(self, dataset: Dataset, resources_paths: Dict[str, str]) -> Optional[Dataset]:\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint)\r\n 271 assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n 272 if self.info.features.type != inferred_features.type:\r\n--> 273 raise ValueError(\r\n 274 \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n 275 self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'id': Value(dtype='string', id=None), 'document': {'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None), 'html': Value(dtype='string', id=None), 'tokens': Sequence(feature={'token': Value(dtype='string', id=None), 'is_html': Value(dtype='bool', id=None)}, length=-1, id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': Sequence(feature={'id': Value(dtype='string', id=None), 'long_answer': {'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None)}, 'short_answers': Sequence(feature={'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}, length=-1, id=None), 'yes_no_answer': ClassLabel(num_classes=2, names=['NO', 'YES'], names_file=None, id=None)}, length=-1, id=None)}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>>, id: string, question: struct>>\r\n\r\nbut expected something like\r\n{'id': Value(dtype='string', id=None), 'document': {'html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'tokens': {'is_html': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'token': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'url': Value(dtype='string', id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': {'id': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'long_answer': [{'end_byte': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'start_token': Value(dtype='int64', id=None)}], 'short_answers': [{'end_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'end_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'text': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}], 'yes_no_answer': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None)}}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>, url: string>, id: string, question: struct>>\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n- Python version: 3.8.3\r\n- PyTorch version (GPU?): 1.6.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n","comment_length":7,"text":"load_dataset('natural_questions') fails with \"ValueError: External features info don't match the dataset\"\n## Describe the bug\r\nload_dataset('natural_questions') throws ValueError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndatasets = load_dataset('natural_questions', split='validation[:10]')\r\n```\r\n\r\n## Expected results\r\nCall to load_dataset returns data.\r\n\r\n## Actual results\r\n```\r\nUsing custom data configuration default\r\nReusing dataset natural_questions (\/mnt\/d\/huggingface\/datasets\/natural_questions\/default\/0.0.2\/19bc04755018a3ad02ee74f7045cde4ba9b4162cb64450a87030ab786b123b76)\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in \r\n----> 1 datasets = load_dataset('natural_questions', split='validation[:10]', cache_dir='\/mnt\/d\/huggingface\/datasets')\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 756 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\r\n 757 )\r\n--> 758 ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n 759 if save_infos:\r\n 760 builder_instance._save_infos()\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in as_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 735 \r\n 736 # Create a dataset for each of the given splits\r\n--> 737 datasets = utils.map_nested(\r\n 738 partial(\r\n 739 self._build_single_dataset,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 193 # Singleton\r\n 194 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 195 return function(data_struct)\r\n 196 \r\n 197 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _build_single_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 762 \r\n 763 # Build base dataset\r\n--> 764 ds = self._as_dataset(\r\n 765 split=split,\r\n 766 in_memory=in_memory,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _as_dataset(self, split, in_memory)\r\n 838 in_memory=in_memory,\r\n 839 )\r\n--> 840 return Dataset(**dataset_kwargs)\r\n 841 \r\n 842 def _post_process(self, dataset: Dataset, resources_paths: Dict[str, str]) -> Optional[Dataset]:\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint)\r\n 271 assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n 272 if self.info.features.type != inferred_features.type:\r\n--> 273 raise ValueError(\r\n 274 \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n 275 self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'id': Value(dtype='string', id=None), 'document': {'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None), 'html': Value(dtype='string', id=None), 'tokens': Sequence(feature={'token': Value(dtype='string', id=None), 'is_html': Value(dtype='bool', id=None)}, length=-1, id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': Sequence(feature={'id': Value(dtype='string', id=None), 'long_answer': {'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None)}, 'short_answers': Sequence(feature={'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}, length=-1, id=None), 'yes_no_answer': ClassLabel(num_classes=2, names=['NO', 'YES'], names_file=None, id=None)}, length=-1, id=None)}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>>, id: string, question: struct>>\r\n\r\nbut expected something like\r\n{'id': Value(dtype='string', id=None), 'document': {'html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'tokens': {'is_html': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'token': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'url': Value(dtype='string', id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': {'id': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'long_answer': [{'end_byte': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'start_token': Value(dtype='int64', id=None)}], 'short_answers': [{'end_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'end_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'text': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}], 'yes_no_answer': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None)}}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>, url: string>, id: string, question: struct>>\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n- Python version: 3.8.3\r\n- PyTorch version (GPU?): 1.6.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\nThanks for reporting, I'm looking into it","embeddings":[-0.3010365963,-0.1931253821,-0.0652654469,0.525786221,0.2114789933,-0.1320342124,0.4025248289,0.3253356814,0.1363790929,0.0655856878,-0.1739462912,0.3462786973,-0.1743816137,0.3402386606,0.0053835479,-0.0349213295,-0.033921551,0.0606687069,-0.0714647621,0.0316084325,-0.1666446477,0.1117764711,-0.1739413589,0.054349374,-0.1699436158,0.0612640306,0.0389056914,0.368008852,0.0395361893,-0.4640865624,0.1402531266,-0.2512812614,0.200597018,0.3641215563,-0.0001097937,0.1101842001,0.291016221,-0.1340754777,-0.4646355808,-0.3495002687,-0.4326409698,-0.2932563126,0.1844670177,-0.1025997847,-0.0750925839,-0.0901743248,-0.1942299753,-0.1198888496,0.2101251334,0.5170678496,0.2509104908,0.437240392,0.1481525749,-0.1356033385,0.4015932977,0.2591476738,-0.1198510081,0.3900898993,-0.1157002077,-0.1915698498,0.1298338175,0.1801397651,-0.0604048744,0.3196884692,0.3086515367,0.0979741514,0.0387078598,-0.2461425811,0.0445514359,0.2624997497,0.5269556642,-0.2871050835,-0.2662248313,-0.3345457017,-0.1267289966,-0.3140507936,0.1942808628,0.1228377894,-0.2235041857,0.1248339564,-0.068113409,0.0506800823,-0.1982334703,0.1594713777,-0.1901111454,0.0679849163,-0.1613807976,0.0686992332,-0.2411332875,-0.2050860971,0.2681346238,-0.1306806505,-0.040640682,-0.0284468103,-0.158993572,-0.1046421081,0.0473441184,-0.2293193638,0.1360967755,0.3597458601,-0.0197180305,-0.0360403508,0.0957899392,0.2053233534,0.2986126244,0.177685976,0.1130070016,0.3782274723,0.1557691097,0.3454990983,-0.1647492349,-0.0209359992,0.0980151296,-0.0155570917,0.3737146258,0.0471422449,0.6195658445,-0.3021340966,-0.3641215265,0.2766881585,0.0009493929,-0.0309526678,0.0115113985,0.34006688,-0.2023818344,0.1892527938,-0.0823120698,0.1425789297,0.0140412003,-0.2875742018,-0.2640197575,-0.0711601377,-0.014275156,0.1304302961,0.1474407762,-0.0116192913,0.5592864156,0.0787607655,0.0984397903,-0.0673231632,0.1605854183,-0.1946494579,-0.1344540566,0.2918545604,-0.209122315,0.0362084657,0.4381088614,-0.0508765057,-0.0775992349,-0.0562448762,-0.1957972497,-0.4581328928,-0.1711842418,0.2804891765,-0.3117066324,0.0094732922,0.2535792291,-0.1273874938,0.3626094162,-0.2142152637,-0.1140045077,-0.2887541056,-0.2575203478,-0.2485756278,0.2768083215,0.4994139075,-0.3477646112,0.0705651119,0.1204925999,-0.1663465947,0.1035433859,0.0806002319,-0.1228351295,0.3711713552,-0.1696024537,-0.1638680995,0.1695781052,-0.2986466289,-0.4003855586,0.3155426979,-0.0153261796,0.2871543467,0.0063026901,0.0271411482,0.1573456824,0.1157109886,0.4620014727,0.4014619291,-0.066910401,-0.0074357758,-0.2804535925,-0.2457351685,0.1595118046,0.2452587187,0.0332712159,0.06809894,0.2984835804,0.0001923513,0.0488624536,-0.0455991328,-0.2236338407,0.1471460015,0.3994964957,0.0760490149,-0.1120455712,-0.404699862,-0.5526606441,0.3127443492,-0.3151362538,0.0898777246,-0.0624681339,0.0992171317,-0.6718125939,0.1659998745,-0.2465763986,-0.3998812437,0.1665603518,0.3555616736,-0.168122068,-0.0031244948,-0.1956553459,0.4240124524,-0.2365444303,0.1162518114,-0.3416495919,0.2825235426,-0.0393327996,-0.149338156,0.0628314167,0.0276369713,0.167953372,0.0517680161,-0.2381964028,0.4361684024,0.2144528329,0.0494904742,-0.2391491681,-0.1940413862,0.038697008,-0.2645381391,0.163651675,0.2008208781,0.1864885539,-0.0701437369,-0.1489441991,0.3302778304,0.0553814918,0.1013679281,-0.0695879534,-0.0139205735,0.0889150873,-0.0030233478,-0.0507578999,-0.2198005319,0.2232966572,-0.0017620587,0.2004920393,0.0682676956,-0.3529761434,0.0907458141,0.2752231658,0.0731488168,0.1277240962,0.1467070132,-0.2322896719,-0.1234295219,0.2753238678,0.0469856486,0.6293234229,0.1825696975,-0.165564239,0.2986156046,-0.15912202,-0.0204912219,0.2052815408,-0.0421392247,0.1960897297,0.2865037024,0.1824352592,0.1393212676,-0.178505227,-0.0943207219,-0.1112810969,0.1985169053,-0.4880636632,0.079517372,-0.1686079204,0.0649334341,-0.0422706679,-0.1734807342,-0.2261099219,-0.1904407442,-0.1937994361,0.2278338224,-0.1691826135,0.2074276656,-0.2505809069,-0.0847058743,0.1458775997,-0.526855588,0.089645423,-0.0621372499,-0.2492348999,0.0662923008,0.2406881154,0.1643317789,0.267888993,-0.20216088,-0.151760608,-0.2189625651,-0.1686671227,-0.0300829411,0.074105829,0.770889461,0.1303961128,0.2613179088,0.379313767,-0.2804147601,0.2815556228,0.0498569459,-0.1869068891,0.0964192227,0.064125143,0.0366347134,-0.0271369424,-0.6000498533,-0.36951527,-0.5040061474,-0.0255676452,0.2252414972,0.1044660807,0.5914885998,0.2259821892,0.098211661,0.0685926825,0.2290950865,-0.2079316676,-0.1246068254,0.2001141459,-0.1017325744,-0.1273368746,-0.316460073,-0.0996490195,0.1105185002,0.0181393623,-0.5311613679,-0.0326023139,-0.3451296091,0.4823538363,-0.1425576508,0.0308493003,0.3642105162,-0.063392967,-0.0323227346,-0.0954335257,-0.2647781968,0.0323960744,-0.0000882497,0.2433171421,0.0356643125,0.4075246155,-0.0353547633,0.6303832531,0.0092987791,-0.0344562195,0.3832820654,-0.1496753991,0.416369319,-0.3017952442,-0.5337983966,0.0131714232,-0.2167703509,-0.1959250569,0.1570640951,-0.0784802437,-0.1687743813,-0.1831905693,-0.0932203308,-0.3046920598,-0.0606943481,0.0384095423,-0.0187649019,0.1651714593,0.1132292226,0.1195074767,-0.2817521989,-0.0889264867,-0.1094715968,0.0146534182,-0.0797828808,-0.0825566798,-0.4356510639,0.2557926178,-0.1198370233,0.2290960848,0.1021284238,0.3126807511,-0.116187349,-0.3825466335,-0.1960823387,-0.0726532489,0.2147463858,0.0184873715,0.0753258839,0.1536622196,-0.1412212104,-0.3610154688,-0.2004318684,-0.1549527943,0.113098599,0.2339238226,0.6427520514,-0.1891706288,0.0071767429,0.3524083495,0.4035503864,-0.1031610444,-0.2549925745,-0.2080092579,-0.1045188084,-0.3992632627,-0.0318549238,-0.1161562353,0.2455658913,-0.0078822803,0.0437904298,-0.1692227125,-0.0803853348,0.165555656,0.2271063626,0.1503321826,-0.1269869804,0.284586221,-0.1758479029,0.1441669017,0.0596954003,0.6913921237,0.0657192096,-0.5916351676,-0.0717137083,0.0396200567,0.1844923943,0.2397723347,0.0448651798,-0.0272457544,-0.0495887175,0.0868238807,-0.0873017907,0.3527728617,0.0881813541,-0.121158205,0.1119030043,-0.2668753564,0.3709553778,-0.1393034011,-0.0502592623,0.2567106485,-0.1017212197,-0.1936982721,-0.0222703051,0.1220172495,0.6600776315,0.3586020172,0.1172699258,0.4765803516,-0.1983785629,0.3195495307,-0.0597341172,-0.1360106617,-0.4920931756,-0.3510355353,0.1049575582,-0.1960784346,0.1987957358,-0.1476992816,-0.0226633139,0.3582038581,-0.0114521012,0.4813610911,0.181717664,0.1715059429,-0.0177972503,-0.024297215,-0.2199698389,0.187038362,-0.0949707925,0.1076645777,-0.0729456171,-0.039432548,-0.0519838631,-0.1821260899,-0.2247526646,0.1746825129,-0.3567138612,0.092994228,0.2788364887,-0.3486048877,-0.1215511039,0.2856612504,0.1725815684,0.1285102963,-0.1277089417,-0.124090977,0.0180381462,0.0834357515,0.0509496666,-0.0223008767,0.3129682243,-0.1497761309,-0.3154302835,0.17894198,-0.1141532063,-0.1367426813,0.0276512485,0.0460858345,0.0429792181,-0.6083951592,-0.3587333858,-0.2729334235,0.2343661785,-0.2964299023,0.1671108007,-0.0850273147,-0.159299016,-0.0390192419,-0.0390178151,-0.3504946232,-0.188681826,0.2096121609,0.0100413337,0.0955855548,0.4750776589,-0.0024073629,0.085909158,-0.281863451,0.5697032809,-0.1319766641,-0.2841341496,0.1106292233,0.0628202036,-0.1130802184,-0.2623446584,0.1433285475,0.2304039598,-0.0656383932,0.2568435073,-0.4319812059,-0.3417362571,0.1113909185,-0.3017697036,0.2283263505,0.0737322196,-0.0314689614,-0.1762896925,0.1253466308,-0.371548444,-0.0490871407,-0.2450396568,0.1156439707,0.1753619909,-0.0509934947,0.0549703278,0.0774088129,0.2249481529,-0.0035852455,-0.1168234944,-0.3020421863,-0.0865272358,0.1254264265,0.0515467748,-0.0695332363,-0.1507494003,-0.0611952282,-0.1140329838,-0.1327302009,0.0390418172,0.1042219698,-0.0347810686,0.3725389242,-0.1999056935,0.2814030349,-0.1046492755,0.2537442446,-0.1507167965,0.1896493137,-0.2035243213,0.1280552447,-0.1644871533,-0.000014853,-0.2832544148,0.0390093923,-0.0349154696,0.0817560554,0.2742252648,-0.2003487498,-0.2600533068,0.0305856019,0.0199263673,0.3129118681,-0.2150289416,0.1065102741,0.1815644205,0.2663259804,-0.3734115958,-0.010485691,0.1419816762,-0.1060630083,0.0743484572,0.2163880616,0.2126561552,-0.0611200705,0.2493149489,0.0930695385,0.2510064244,0.0173221454,0.0926623568,0.3244889081,-0.2003924251,-0.043881692,0.2114525735,0.1855243593,0.2891336381,0.4851191044,-0.079485178,0.0525735989,-0.2255954295,0.0710886717,0.0361781418,-0.3354212046,0.0672081336,0.0693156421,-0.0936621055,-0.00470267,-0.1472409517,0.0748917535,-0.2594290078,0.0290211085,-0.2763265669,-0.0886882246,-0.0123605458,-0.017395312,-0.2431390733,-0.2479863018,-0.1207344383,-0.053788323,-0.0487490408,-0.1577880383,0.1042184457,-0.1172417551,-0.3449355066,-0.4382255077,0.1962499619,0.4863179624,-0.187797606,-0.3739557862,0.3553467691,0.1638759375,0.1163184866,-0.0050221337,0.2935783565,0.5776576996,0.4492668509,-0.2183886468,-0.0133697409,-0.0903499722,-0.2572478652,-0.08058808,0.1788306087,-0.2036101818,0.2041838914,0.3896346092,0.2425412089,-0.2860693038,0.3087479472,0.0671433136,-0.014772282,-0.6452242136,0.3085484207,0.0373875014,0.0155810714,0.1040226147,-0.0377722569,-0.2956029773,-0.0928971395,0.4650352895,0.2191892415,0.2604660094,0.0643747449,0.1007259563,-0.2586954534,0.2847928107,0.3491569161,0.2334001511,-0.302959919,0.1163394004,-0.6907718182,0.1602863669,-0.0385213606,-0.1857215613,0.145208776,0.1284459084,0.189406991,0.3235723674,-0.2789386511,0.1875156462,0.1158161089,-0.102913402,-0.3438047171,-0.2960200012,-0.3301110864,0.1387993097,0.0364331864,-0.4284615219,0.0283448994,-0.2654493451,0.0856036842,-0.0148275346,-0.1632930338,0.2235354483,0.1154315844,0.4646860659,0.1625410169,0.4427508414,-0.0643838346,-0.0877063125,-0.1593246162,-0.326819092,-0.2241960913,0.0881990641,0.0083235903,0.6604252458,-0.0688739419,-0.186289534,-0.2489366531,0.547285974,0.1704099625,-0.1335785091,0.1080821976,0.0587932207,0.005058486,0.016521344,0.0720723271,0.086332269,-0.1636992246,0.2174746543,-0.0827195421,-0.2598025203,0.4770698547,-0.0727573186,-0.3654350936,-0.0709003955,0.2040002495,-0.0339087285,-0.0137331747,-0.373477608,0.1011847407,0.5264702439,-0.0827589035,-0.2293682843,0.1287205517,0.0001831334,0.119935222,-0.0393447988,0.4087265432,-0.0403168648,-0.1360573173,0.0768616349,-0.1536669582]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2401","title":"load_dataset('natural_questions') fails with \"ValueError: External features info don't match the dataset\"","comments":"I just opened #2438 to fix this :)","body":"## Describe the bug\r\nload_dataset('natural_questions') throws ValueError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndatasets = load_dataset('natural_questions', split='validation[:10]')\r\n```\r\n\r\n## Expected results\r\nCall to load_dataset returns data.\r\n\r\n## Actual results\r\n```\r\nUsing custom data configuration default\r\nReusing dataset natural_questions (\/mnt\/d\/huggingface\/datasets\/natural_questions\/default\/0.0.2\/19bc04755018a3ad02ee74f7045cde4ba9b4162cb64450a87030ab786b123b76)\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in \r\n----> 1 datasets = load_dataset('natural_questions', split='validation[:10]', cache_dir='\/mnt\/d\/huggingface\/datasets')\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 756 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\r\n 757 )\r\n--> 758 ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n 759 if save_infos:\r\n 760 builder_instance._save_infos()\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in as_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 735 \r\n 736 # Create a dataset for each of the given splits\r\n--> 737 datasets = utils.map_nested(\r\n 738 partial(\r\n 739 self._build_single_dataset,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 193 # Singleton\r\n 194 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 195 return function(data_struct)\r\n 196 \r\n 197 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _build_single_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 762 \r\n 763 # Build base dataset\r\n--> 764 ds = self._as_dataset(\r\n 765 split=split,\r\n 766 in_memory=in_memory,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _as_dataset(self, split, in_memory)\r\n 838 in_memory=in_memory,\r\n 839 )\r\n--> 840 return Dataset(**dataset_kwargs)\r\n 841 \r\n 842 def _post_process(self, dataset: Dataset, resources_paths: Dict[str, str]) -> Optional[Dataset]:\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint)\r\n 271 assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n 272 if self.info.features.type != inferred_features.type:\r\n--> 273 raise ValueError(\r\n 274 \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n 275 self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'id': Value(dtype='string', id=None), 'document': {'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None), 'html': Value(dtype='string', id=None), 'tokens': Sequence(feature={'token': Value(dtype='string', id=None), 'is_html': Value(dtype='bool', id=None)}, length=-1, id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': Sequence(feature={'id': Value(dtype='string', id=None), 'long_answer': {'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None)}, 'short_answers': Sequence(feature={'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}, length=-1, id=None), 'yes_no_answer': ClassLabel(num_classes=2, names=['NO', 'YES'], names_file=None, id=None)}, length=-1, id=None)}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>>, id: string, question: struct>>\r\n\r\nbut expected something like\r\n{'id': Value(dtype='string', id=None), 'document': {'html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'tokens': {'is_html': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'token': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'url': Value(dtype='string', id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': {'id': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'long_answer': [{'end_byte': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'start_token': Value(dtype='int64', id=None)}], 'short_answers': [{'end_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'end_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'text': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}], 'yes_no_answer': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None)}}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>, url: string>, id: string, question: struct>>\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n- Python version: 3.8.3\r\n- PyTorch version (GPU?): 1.6.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n","comment_length":8,"text":"load_dataset('natural_questions') fails with \"ValueError: External features info don't match the dataset\"\n## Describe the bug\r\nload_dataset('natural_questions') throws ValueError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndatasets = load_dataset('natural_questions', split='validation[:10]')\r\n```\r\n\r\n## Expected results\r\nCall to load_dataset returns data.\r\n\r\n## Actual results\r\n```\r\nUsing custom data configuration default\r\nReusing dataset natural_questions (\/mnt\/d\/huggingface\/datasets\/natural_questions\/default\/0.0.2\/19bc04755018a3ad02ee74f7045cde4ba9b4162cb64450a87030ab786b123b76)\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in \r\n----> 1 datasets = load_dataset('natural_questions', split='validation[:10]', cache_dir='\/mnt\/d\/huggingface\/datasets')\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 756 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\r\n 757 )\r\n--> 758 ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n 759 if save_infos:\r\n 760 builder_instance._save_infos()\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in as_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 735 \r\n 736 # Create a dataset for each of the given splits\r\n--> 737 datasets = utils.map_nested(\r\n 738 partial(\r\n 739 self._build_single_dataset,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 193 # Singleton\r\n 194 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 195 return function(data_struct)\r\n 196 \r\n 197 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _build_single_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 762 \r\n 763 # Build base dataset\r\n--> 764 ds = self._as_dataset(\r\n 765 split=split,\r\n 766 in_memory=in_memory,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _as_dataset(self, split, in_memory)\r\n 838 in_memory=in_memory,\r\n 839 )\r\n--> 840 return Dataset(**dataset_kwargs)\r\n 841 \r\n 842 def _post_process(self, dataset: Dataset, resources_paths: Dict[str, str]) -> Optional[Dataset]:\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint)\r\n 271 assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n 272 if self.info.features.type != inferred_features.type:\r\n--> 273 raise ValueError(\r\n 274 \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n 275 self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'id': Value(dtype='string', id=None), 'document': {'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None), 'html': Value(dtype='string', id=None), 'tokens': Sequence(feature={'token': Value(dtype='string', id=None), 'is_html': Value(dtype='bool', id=None)}, length=-1, id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': Sequence(feature={'id': Value(dtype='string', id=None), 'long_answer': {'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None)}, 'short_answers': Sequence(feature={'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}, length=-1, id=None), 'yes_no_answer': ClassLabel(num_classes=2, names=['NO', 'YES'], names_file=None, id=None)}, length=-1, id=None)}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>>, id: string, question: struct>>\r\n\r\nbut expected something like\r\n{'id': Value(dtype='string', id=None), 'document': {'html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'tokens': {'is_html': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'token': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'url': Value(dtype='string', id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': {'id': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'long_answer': [{'end_byte': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'start_token': Value(dtype='int64', id=None)}], 'short_answers': [{'end_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'end_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'text': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}], 'yes_no_answer': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None)}}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>, url: string>, id: string, question: struct>>\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n- Python version: 3.8.3\r\n- PyTorch version (GPU?): 1.6.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\nI just opened #2438 to fix this :)","embeddings":[-0.3010365963,-0.1931253821,-0.0652654469,0.525786221,0.2114789933,-0.1320342124,0.4025248289,0.3253356814,0.1363790929,0.0655856878,-0.1739462912,0.3462786973,-0.1743816137,0.3402386606,0.0053835479,-0.0349213295,-0.033921551,0.0606687069,-0.0714647621,0.0316084325,-0.1666446477,0.1117764711,-0.1739413589,0.054349374,-0.1699436158,0.0612640306,0.0389056914,0.368008852,0.0395361893,-0.4640865624,0.1402531266,-0.2512812614,0.200597018,0.3641215563,-0.0001097937,0.1101842001,0.291016221,-0.1340754777,-0.4646355808,-0.3495002687,-0.4326409698,-0.2932563126,0.1844670177,-0.1025997847,-0.0750925839,-0.0901743248,-0.1942299753,-0.1198888496,0.2101251334,0.5170678496,0.2509104908,0.437240392,0.1481525749,-0.1356033385,0.4015932977,0.2591476738,-0.1198510081,0.3900898993,-0.1157002077,-0.1915698498,0.1298338175,0.1801397651,-0.0604048744,0.3196884692,0.3086515367,0.0979741514,0.0387078598,-0.2461425811,0.0445514359,0.2624997497,0.5269556642,-0.2871050835,-0.2662248313,-0.3345457017,-0.1267289966,-0.3140507936,0.1942808628,0.1228377894,-0.2235041857,0.1248339564,-0.068113409,0.0506800823,-0.1982334703,0.1594713777,-0.1901111454,0.0679849163,-0.1613807976,0.0686992332,-0.2411332875,-0.2050860971,0.2681346238,-0.1306806505,-0.040640682,-0.0284468103,-0.158993572,-0.1046421081,0.0473441184,-0.2293193638,0.1360967755,0.3597458601,-0.0197180305,-0.0360403508,0.0957899392,0.2053233534,0.2986126244,0.177685976,0.1130070016,0.3782274723,0.1557691097,0.3454990983,-0.1647492349,-0.0209359992,0.0980151296,-0.0155570917,0.3737146258,0.0471422449,0.6195658445,-0.3021340966,-0.3641215265,0.2766881585,0.0009493929,-0.0309526678,0.0115113985,0.34006688,-0.2023818344,0.1892527938,-0.0823120698,0.1425789297,0.0140412003,-0.2875742018,-0.2640197575,-0.0711601377,-0.014275156,0.1304302961,0.1474407762,-0.0116192913,0.5592864156,0.0787607655,0.0984397903,-0.0673231632,0.1605854183,-0.1946494579,-0.1344540566,0.2918545604,-0.209122315,0.0362084657,0.4381088614,-0.0508765057,-0.0775992349,-0.0562448762,-0.1957972497,-0.4581328928,-0.1711842418,0.2804891765,-0.3117066324,0.0094732922,0.2535792291,-0.1273874938,0.3626094162,-0.2142152637,-0.1140045077,-0.2887541056,-0.2575203478,-0.2485756278,0.2768083215,0.4994139075,-0.3477646112,0.0705651119,0.1204925999,-0.1663465947,0.1035433859,0.0806002319,-0.1228351295,0.3711713552,-0.1696024537,-0.1638680995,0.1695781052,-0.2986466289,-0.4003855586,0.3155426979,-0.0153261796,0.2871543467,0.0063026901,0.0271411482,0.1573456824,0.1157109886,0.4620014727,0.4014619291,-0.066910401,-0.0074357758,-0.2804535925,-0.2457351685,0.1595118046,0.2452587187,0.0332712159,0.06809894,0.2984835804,0.0001923513,0.0488624536,-0.0455991328,-0.2236338407,0.1471460015,0.3994964957,0.0760490149,-0.1120455712,-0.404699862,-0.5526606441,0.3127443492,-0.3151362538,0.0898777246,-0.0624681339,0.0992171317,-0.6718125939,0.1659998745,-0.2465763986,-0.3998812437,0.1665603518,0.3555616736,-0.168122068,-0.0031244948,-0.1956553459,0.4240124524,-0.2365444303,0.1162518114,-0.3416495919,0.2825235426,-0.0393327996,-0.149338156,0.0628314167,0.0276369713,0.167953372,0.0517680161,-0.2381964028,0.4361684024,0.2144528329,0.0494904742,-0.2391491681,-0.1940413862,0.038697008,-0.2645381391,0.163651675,0.2008208781,0.1864885539,-0.0701437369,-0.1489441991,0.3302778304,0.0553814918,0.1013679281,-0.0695879534,-0.0139205735,0.0889150873,-0.0030233478,-0.0507578999,-0.2198005319,0.2232966572,-0.0017620587,0.2004920393,0.0682676956,-0.3529761434,0.0907458141,0.2752231658,0.0731488168,0.1277240962,0.1467070132,-0.2322896719,-0.1234295219,0.2753238678,0.0469856486,0.6293234229,0.1825696975,-0.165564239,0.2986156046,-0.15912202,-0.0204912219,0.2052815408,-0.0421392247,0.1960897297,0.2865037024,0.1824352592,0.1393212676,-0.178505227,-0.0943207219,-0.1112810969,0.1985169053,-0.4880636632,0.079517372,-0.1686079204,0.0649334341,-0.0422706679,-0.1734807342,-0.2261099219,-0.1904407442,-0.1937994361,0.2278338224,-0.1691826135,0.2074276656,-0.2505809069,-0.0847058743,0.1458775997,-0.526855588,0.089645423,-0.0621372499,-0.2492348999,0.0662923008,0.2406881154,0.1643317789,0.267888993,-0.20216088,-0.151760608,-0.2189625651,-0.1686671227,-0.0300829411,0.074105829,0.770889461,0.1303961128,0.2613179088,0.379313767,-0.2804147601,0.2815556228,0.0498569459,-0.1869068891,0.0964192227,0.064125143,0.0366347134,-0.0271369424,-0.6000498533,-0.36951527,-0.5040061474,-0.0255676452,0.2252414972,0.1044660807,0.5914885998,0.2259821892,0.098211661,0.0685926825,0.2290950865,-0.2079316676,-0.1246068254,0.2001141459,-0.1017325744,-0.1273368746,-0.316460073,-0.0996490195,0.1105185002,0.0181393623,-0.5311613679,-0.0326023139,-0.3451296091,0.4823538363,-0.1425576508,0.0308493003,0.3642105162,-0.063392967,-0.0323227346,-0.0954335257,-0.2647781968,0.0323960744,-0.0000882497,0.2433171421,0.0356643125,0.4075246155,-0.0353547633,0.6303832531,0.0092987791,-0.0344562195,0.3832820654,-0.1496753991,0.416369319,-0.3017952442,-0.5337983966,0.0131714232,-0.2167703509,-0.1959250569,0.1570640951,-0.0784802437,-0.1687743813,-0.1831905693,-0.0932203308,-0.3046920598,-0.0606943481,0.0384095423,-0.0187649019,0.1651714593,0.1132292226,0.1195074767,-0.2817521989,-0.0889264867,-0.1094715968,0.0146534182,-0.0797828808,-0.0825566798,-0.4356510639,0.2557926178,-0.1198370233,0.2290960848,0.1021284238,0.3126807511,-0.116187349,-0.3825466335,-0.1960823387,-0.0726532489,0.2147463858,0.0184873715,0.0753258839,0.1536622196,-0.1412212104,-0.3610154688,-0.2004318684,-0.1549527943,0.113098599,0.2339238226,0.6427520514,-0.1891706288,0.0071767429,0.3524083495,0.4035503864,-0.1031610444,-0.2549925745,-0.2080092579,-0.1045188084,-0.3992632627,-0.0318549238,-0.1161562353,0.2455658913,-0.0078822803,0.0437904298,-0.1692227125,-0.0803853348,0.165555656,0.2271063626,0.1503321826,-0.1269869804,0.284586221,-0.1758479029,0.1441669017,0.0596954003,0.6913921237,0.0657192096,-0.5916351676,-0.0717137083,0.0396200567,0.1844923943,0.2397723347,0.0448651798,-0.0272457544,-0.0495887175,0.0868238807,-0.0873017907,0.3527728617,0.0881813541,-0.121158205,0.1119030043,-0.2668753564,0.3709553778,-0.1393034011,-0.0502592623,0.2567106485,-0.1017212197,-0.1936982721,-0.0222703051,0.1220172495,0.6600776315,0.3586020172,0.1172699258,0.4765803516,-0.1983785629,0.3195495307,-0.0597341172,-0.1360106617,-0.4920931756,-0.3510355353,0.1049575582,-0.1960784346,0.1987957358,-0.1476992816,-0.0226633139,0.3582038581,-0.0114521012,0.4813610911,0.181717664,0.1715059429,-0.0177972503,-0.024297215,-0.2199698389,0.187038362,-0.0949707925,0.1076645777,-0.0729456171,-0.039432548,-0.0519838631,-0.1821260899,-0.2247526646,0.1746825129,-0.3567138612,0.092994228,0.2788364887,-0.3486048877,-0.1215511039,0.2856612504,0.1725815684,0.1285102963,-0.1277089417,-0.124090977,0.0180381462,0.0834357515,0.0509496666,-0.0223008767,0.3129682243,-0.1497761309,-0.3154302835,0.17894198,-0.1141532063,-0.1367426813,0.0276512485,0.0460858345,0.0429792181,-0.6083951592,-0.3587333858,-0.2729334235,0.2343661785,-0.2964299023,0.1671108007,-0.0850273147,-0.159299016,-0.0390192419,-0.0390178151,-0.3504946232,-0.188681826,0.2096121609,0.0100413337,0.0955855548,0.4750776589,-0.0024073629,0.085909158,-0.281863451,0.5697032809,-0.1319766641,-0.2841341496,0.1106292233,0.0628202036,-0.1130802184,-0.2623446584,0.1433285475,0.2304039598,-0.0656383932,0.2568435073,-0.4319812059,-0.3417362571,0.1113909185,-0.3017697036,0.2283263505,0.0737322196,-0.0314689614,-0.1762896925,0.1253466308,-0.371548444,-0.0490871407,-0.2450396568,0.1156439707,0.1753619909,-0.0509934947,0.0549703278,0.0774088129,0.2249481529,-0.0035852455,-0.1168234944,-0.3020421863,-0.0865272358,0.1254264265,0.0515467748,-0.0695332363,-0.1507494003,-0.0611952282,-0.1140329838,-0.1327302009,0.0390418172,0.1042219698,-0.0347810686,0.3725389242,-0.1999056935,0.2814030349,-0.1046492755,0.2537442446,-0.1507167965,0.1896493137,-0.2035243213,0.1280552447,-0.1644871533,-0.000014853,-0.2832544148,0.0390093923,-0.0349154696,0.0817560554,0.2742252648,-0.2003487498,-0.2600533068,0.0305856019,0.0199263673,0.3129118681,-0.2150289416,0.1065102741,0.1815644205,0.2663259804,-0.3734115958,-0.010485691,0.1419816762,-0.1060630083,0.0743484572,0.2163880616,0.2126561552,-0.0611200705,0.2493149489,0.0930695385,0.2510064244,0.0173221454,0.0926623568,0.3244889081,-0.2003924251,-0.043881692,0.2114525735,0.1855243593,0.2891336381,0.4851191044,-0.079485178,0.0525735989,-0.2255954295,0.0710886717,0.0361781418,-0.3354212046,0.0672081336,0.0693156421,-0.0936621055,-0.00470267,-0.1472409517,0.0748917535,-0.2594290078,0.0290211085,-0.2763265669,-0.0886882246,-0.0123605458,-0.017395312,-0.2431390733,-0.2479863018,-0.1207344383,-0.053788323,-0.0487490408,-0.1577880383,0.1042184457,-0.1172417551,-0.3449355066,-0.4382255077,0.1962499619,0.4863179624,-0.187797606,-0.3739557862,0.3553467691,0.1638759375,0.1163184866,-0.0050221337,0.2935783565,0.5776576996,0.4492668509,-0.2183886468,-0.0133697409,-0.0903499722,-0.2572478652,-0.08058808,0.1788306087,-0.2036101818,0.2041838914,0.3896346092,0.2425412089,-0.2860693038,0.3087479472,0.0671433136,-0.014772282,-0.6452242136,0.3085484207,0.0373875014,0.0155810714,0.1040226147,-0.0377722569,-0.2956029773,-0.0928971395,0.4650352895,0.2191892415,0.2604660094,0.0643747449,0.1007259563,-0.2586954534,0.2847928107,0.3491569161,0.2334001511,-0.302959919,0.1163394004,-0.6907718182,0.1602863669,-0.0385213606,-0.1857215613,0.145208776,0.1284459084,0.189406991,0.3235723674,-0.2789386511,0.1875156462,0.1158161089,-0.102913402,-0.3438047171,-0.2960200012,-0.3301110864,0.1387993097,0.0364331864,-0.4284615219,0.0283448994,-0.2654493451,0.0856036842,-0.0148275346,-0.1632930338,0.2235354483,0.1154315844,0.4646860659,0.1625410169,0.4427508414,-0.0643838346,-0.0877063125,-0.1593246162,-0.326819092,-0.2241960913,0.0881990641,0.0083235903,0.6604252458,-0.0688739419,-0.186289534,-0.2489366531,0.547285974,0.1704099625,-0.1335785091,0.1080821976,0.0587932207,0.005058486,0.016521344,0.0720723271,0.086332269,-0.1636992246,0.2174746543,-0.0827195421,-0.2598025203,0.4770698547,-0.0727573186,-0.3654350936,-0.0709003955,0.2040002495,-0.0339087285,-0.0137331747,-0.373477608,0.1011847407,0.5264702439,-0.0827589035,-0.2293682843,0.1287205517,0.0001831334,0.119935222,-0.0393447988,0.4087265432,-0.0403168648,-0.1360573173,0.0768616349,-0.1536669582]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2401","title":"load_dataset('natural_questions') fails with \"ValueError: External features info don't match the dataset\"","comments":"Hi ! This has been fixed in the 1.8.0 release of `datasets`","body":"## Describe the bug\r\nload_dataset('natural_questions') throws ValueError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndatasets = load_dataset('natural_questions', split='validation[:10]')\r\n```\r\n\r\n## Expected results\r\nCall to load_dataset returns data.\r\n\r\n## Actual results\r\n```\r\nUsing custom data configuration default\r\nReusing dataset natural_questions (\/mnt\/d\/huggingface\/datasets\/natural_questions\/default\/0.0.2\/19bc04755018a3ad02ee74f7045cde4ba9b4162cb64450a87030ab786b123b76)\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in \r\n----> 1 datasets = load_dataset('natural_questions', split='validation[:10]', cache_dir='\/mnt\/d\/huggingface\/datasets')\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 756 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\r\n 757 )\r\n--> 758 ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n 759 if save_infos:\r\n 760 builder_instance._save_infos()\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in as_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 735 \r\n 736 # Create a dataset for each of the given splits\r\n--> 737 datasets = utils.map_nested(\r\n 738 partial(\r\n 739 self._build_single_dataset,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 193 # Singleton\r\n 194 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 195 return function(data_struct)\r\n 196 \r\n 197 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _build_single_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 762 \r\n 763 # Build base dataset\r\n--> 764 ds = self._as_dataset(\r\n 765 split=split,\r\n 766 in_memory=in_memory,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _as_dataset(self, split, in_memory)\r\n 838 in_memory=in_memory,\r\n 839 )\r\n--> 840 return Dataset(**dataset_kwargs)\r\n 841 \r\n 842 def _post_process(self, dataset: Dataset, resources_paths: Dict[str, str]) -> Optional[Dataset]:\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint)\r\n 271 assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n 272 if self.info.features.type != inferred_features.type:\r\n--> 273 raise ValueError(\r\n 274 \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n 275 self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'id': Value(dtype='string', id=None), 'document': {'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None), 'html': Value(dtype='string', id=None), 'tokens': Sequence(feature={'token': Value(dtype='string', id=None), 'is_html': Value(dtype='bool', id=None)}, length=-1, id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': Sequence(feature={'id': Value(dtype='string', id=None), 'long_answer': {'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None)}, 'short_answers': Sequence(feature={'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}, length=-1, id=None), 'yes_no_answer': ClassLabel(num_classes=2, names=['NO', 'YES'], names_file=None, id=None)}, length=-1, id=None)}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>>, id: string, question: struct>>\r\n\r\nbut expected something like\r\n{'id': Value(dtype='string', id=None), 'document': {'html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'tokens': {'is_html': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'token': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'url': Value(dtype='string', id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': {'id': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'long_answer': [{'end_byte': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'start_token': Value(dtype='int64', id=None)}], 'short_answers': [{'end_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'end_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'text': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}], 'yes_no_answer': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None)}}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>, url: string>, id: string, question: struct>>\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n- Python version: 3.8.3\r\n- PyTorch version (GPU?): 1.6.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n","comment_length":12,"text":"load_dataset('natural_questions') fails with \"ValueError: External features info don't match the dataset\"\n## Describe the bug\r\nload_dataset('natural_questions') throws ValueError\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndatasets = load_dataset('natural_questions', split='validation[:10]')\r\n```\r\n\r\n## Expected results\r\nCall to load_dataset returns data.\r\n\r\n## Actual results\r\n```\r\nUsing custom data configuration default\r\nReusing dataset natural_questions (\/mnt\/d\/huggingface\/datasets\/natural_questions\/default\/0.0.2\/19bc04755018a3ad02ee74f7045cde4ba9b4162cb64450a87030ab786b123b76)\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in \r\n----> 1 datasets = load_dataset('natural_questions', split='validation[:10]', cache_dir='\/mnt\/d\/huggingface\/datasets')\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 756 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\r\n 757 )\r\n--> 758 ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n 759 if save_infos:\r\n 760 builder_instance._save_infos()\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in as_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 735 \r\n 736 # Create a dataset for each of the given splits\r\n--> 737 datasets = utils.map_nested(\r\n 738 partial(\r\n 739 self._build_single_dataset,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 193 # Singleton\r\n 194 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 195 return function(data_struct)\r\n 196 \r\n 197 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _build_single_dataset(self, split, run_post_process, ignore_verifications, in_memory)\r\n 762 \r\n 763 # Build base dataset\r\n--> 764 ds = self._as_dataset(\r\n 765 split=split,\r\n 766 in_memory=in_memory,\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/builder.py in _as_dataset(self, split, in_memory)\r\n 838 in_memory=in_memory,\r\n 839 )\r\n--> 840 return Dataset(**dataset_kwargs)\r\n 841 \r\n 842 def _post_process(self, dataset: Dataset, resources_paths: Dict[str, str]) -> Optional[Dataset]:\r\n\r\n~\/miniconda3\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint)\r\n 271 assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n 272 if self.info.features.type != inferred_features.type:\r\n--> 273 raise ValueError(\r\n 274 \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n 275 self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'id': Value(dtype='string', id=None), 'document': {'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None), 'html': Value(dtype='string', id=None), 'tokens': Sequence(feature={'token': Value(dtype='string', id=None), 'is_html': Value(dtype='bool', id=None)}, length=-1, id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': Sequence(feature={'id': Value(dtype='string', id=None), 'long_answer': {'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None)}, 'short_answers': Sequence(feature={'start_token': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'end_byte': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)}, length=-1, id=None), 'yes_no_answer': ClassLabel(num_classes=2, names=['NO', 'YES'], names_file=None, id=None)}, length=-1, id=None)}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>>, id: string, question: struct>>\r\n\r\nbut expected something like\r\n{'id': Value(dtype='string', id=None), 'document': {'html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'tokens': {'is_html': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'token': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'url': Value(dtype='string', id=None)}, 'question': {'text': Value(dtype='string', id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}, 'annotations': {'id': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'long_answer': [{'end_byte': Value(dtype='int64', id=None), 'end_token': Value(dtype='int64', id=None), 'start_byte': Value(dtype='int64', id=None), 'start_token': Value(dtype='int64', id=None)}], 'short_answers': [{'end_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'end_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_byte': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'start_token': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'text': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}], 'yes_no_answer': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None)}}\r\nwith type\r\nstruct, long_answer: list>, short_answers: list, end_token: list, start_byte: list, start_token: list, text: list>>, yes_no_answer: list>, document: struct, token: list>, url: string>, id: string, question: struct>>\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n- Python version: 3.8.3\r\n- PyTorch version (GPU?): 1.6.0 (False)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Using GPU in script?: No\r\n- Using distributed or parallel set-up in script?: No\r\n\nHi ! This has been fixed in the 1.8.0 release of `datasets`","embeddings":[-0.3010365963,-0.1931253821,-0.0652654469,0.525786221,0.2114789933,-0.1320342124,0.4025248289,0.3253356814,0.1363790929,0.0655856878,-0.1739462912,0.3462786973,-0.1743816137,0.3402386606,0.0053835479,-0.0349213295,-0.033921551,0.0606687069,-0.0714647621,0.0316084325,-0.1666446477,0.1117764711,-0.1739413589,0.054349374,-0.1699436158,0.0612640306,0.0389056914,0.368008852,0.0395361893,-0.4640865624,0.1402531266,-0.2512812614,0.200597018,0.3641215563,-0.0001097937,0.1101842001,0.291016221,-0.1340754777,-0.4646355808,-0.3495002687,-0.4326409698,-0.2932563126,0.1844670177,-0.1025997847,-0.0750925839,-0.0901743248,-0.1942299753,-0.1198888496,0.2101251334,0.5170678496,0.2509104908,0.437240392,0.1481525749,-0.1356033385,0.4015932977,0.2591476738,-0.1198510081,0.3900898993,-0.1157002077,-0.1915698498,0.1298338175,0.1801397651,-0.0604048744,0.3196884692,0.3086515367,0.0979741514,0.0387078598,-0.2461425811,0.0445514359,0.2624997497,0.5269556642,-0.2871050835,-0.2662248313,-0.3345457017,-0.1267289966,-0.3140507936,0.1942808628,0.1228377894,-0.2235041857,0.1248339564,-0.068113409,0.0506800823,-0.1982334703,0.1594713777,-0.1901111454,0.0679849163,-0.1613807976,0.0686992332,-0.2411332875,-0.2050860971,0.2681346238,-0.1306806505,-0.040640682,-0.0284468103,-0.158993572,-0.1046421081,0.0473441184,-0.2293193638,0.1360967755,0.3597458601,-0.0197180305,-0.0360403508,0.0957899392,0.2053233534,0.2986126244,0.177685976,0.1130070016,0.3782274723,0.1557691097,0.3454990983,-0.1647492349,-0.0209359992,0.0980151296,-0.0155570917,0.3737146258,0.0471422449,0.6195658445,-0.3021340966,-0.3641215265,0.2766881585,0.0009493929,-0.0309526678,0.0115113985,0.34006688,-0.2023818344,0.1892527938,-0.0823120698,0.1425789297,0.0140412003,-0.2875742018,-0.2640197575,-0.0711601377,-0.014275156,0.1304302961,0.1474407762,-0.0116192913,0.5592864156,0.0787607655,0.0984397903,-0.0673231632,0.1605854183,-0.1946494579,-0.1344540566,0.2918545604,-0.209122315,0.0362084657,0.4381088614,-0.0508765057,-0.0775992349,-0.0562448762,-0.1957972497,-0.4581328928,-0.1711842418,0.2804891765,-0.3117066324,0.0094732922,0.2535792291,-0.1273874938,0.3626094162,-0.2142152637,-0.1140045077,-0.2887541056,-0.2575203478,-0.2485756278,0.2768083215,0.4994139075,-0.3477646112,0.0705651119,0.1204925999,-0.1663465947,0.1035433859,0.0806002319,-0.1228351295,0.3711713552,-0.1696024537,-0.1638680995,0.1695781052,-0.2986466289,-0.4003855586,0.3155426979,-0.0153261796,0.2871543467,0.0063026901,0.0271411482,0.1573456824,0.1157109886,0.4620014727,0.4014619291,-0.066910401,-0.0074357758,-0.2804535925,-0.2457351685,0.1595118046,0.2452587187,0.0332712159,0.06809894,0.2984835804,0.0001923513,0.0488624536,-0.0455991328,-0.2236338407,0.1471460015,0.3994964957,0.0760490149,-0.1120455712,-0.404699862,-0.5526606441,0.3127443492,-0.3151362538,0.0898777246,-0.0624681339,0.0992171317,-0.6718125939,0.1659998745,-0.2465763986,-0.3998812437,0.1665603518,0.3555616736,-0.168122068,-0.0031244948,-0.1956553459,0.4240124524,-0.2365444303,0.1162518114,-0.3416495919,0.2825235426,-0.0393327996,-0.149338156,0.0628314167,0.0276369713,0.167953372,0.0517680161,-0.2381964028,0.4361684024,0.2144528329,0.0494904742,-0.2391491681,-0.1940413862,0.038697008,-0.2645381391,0.163651675,0.2008208781,0.1864885539,-0.0701437369,-0.1489441991,0.3302778304,0.0553814918,0.1013679281,-0.0695879534,-0.0139205735,0.0889150873,-0.0030233478,-0.0507578999,-0.2198005319,0.2232966572,-0.0017620587,0.2004920393,0.0682676956,-0.3529761434,0.0907458141,0.2752231658,0.0731488168,0.1277240962,0.1467070132,-0.2322896719,-0.1234295219,0.2753238678,0.0469856486,0.6293234229,0.1825696975,-0.165564239,0.2986156046,-0.15912202,-0.0204912219,0.2052815408,-0.0421392247,0.1960897297,0.2865037024,0.1824352592,0.1393212676,-0.178505227,-0.0943207219,-0.1112810969,0.1985169053,-0.4880636632,0.079517372,-0.1686079204,0.0649334341,-0.0422706679,-0.1734807342,-0.2261099219,-0.1904407442,-0.1937994361,0.2278338224,-0.1691826135,0.2074276656,-0.2505809069,-0.0847058743,0.1458775997,-0.526855588,0.089645423,-0.0621372499,-0.2492348999,0.0662923008,0.2406881154,0.1643317789,0.267888993,-0.20216088,-0.151760608,-0.2189625651,-0.1686671227,-0.0300829411,0.074105829,0.770889461,0.1303961128,0.2613179088,0.379313767,-0.2804147601,0.2815556228,0.0498569459,-0.1869068891,0.0964192227,0.064125143,0.0366347134,-0.0271369424,-0.6000498533,-0.36951527,-0.5040061474,-0.0255676452,0.2252414972,0.1044660807,0.5914885998,0.2259821892,0.098211661,0.0685926825,0.2290950865,-0.2079316676,-0.1246068254,0.2001141459,-0.1017325744,-0.1273368746,-0.316460073,-0.0996490195,0.1105185002,0.0181393623,-0.5311613679,-0.0326023139,-0.3451296091,0.4823538363,-0.1425576508,0.0308493003,0.3642105162,-0.063392967,-0.0323227346,-0.0954335257,-0.2647781968,0.0323960744,-0.0000882497,0.2433171421,0.0356643125,0.4075246155,-0.0353547633,0.6303832531,0.0092987791,-0.0344562195,0.3832820654,-0.1496753991,0.416369319,-0.3017952442,-0.5337983966,0.0131714232,-0.2167703509,-0.1959250569,0.1570640951,-0.0784802437,-0.1687743813,-0.1831905693,-0.0932203308,-0.3046920598,-0.0606943481,0.0384095423,-0.0187649019,0.1651714593,0.1132292226,0.1195074767,-0.2817521989,-0.0889264867,-0.1094715968,0.0146534182,-0.0797828808,-0.0825566798,-0.4356510639,0.2557926178,-0.1198370233,0.2290960848,0.1021284238,0.3126807511,-0.116187349,-0.3825466335,-0.1960823387,-0.0726532489,0.2147463858,0.0184873715,0.0753258839,0.1536622196,-0.1412212104,-0.3610154688,-0.2004318684,-0.1549527943,0.113098599,0.2339238226,0.6427520514,-0.1891706288,0.0071767429,0.3524083495,0.4035503864,-0.1031610444,-0.2549925745,-0.2080092579,-0.1045188084,-0.3992632627,-0.0318549238,-0.1161562353,0.2455658913,-0.0078822803,0.0437904298,-0.1692227125,-0.0803853348,0.165555656,0.2271063626,0.1503321826,-0.1269869804,0.284586221,-0.1758479029,0.1441669017,0.0596954003,0.6913921237,0.0657192096,-0.5916351676,-0.0717137083,0.0396200567,0.1844923943,0.2397723347,0.0448651798,-0.0272457544,-0.0495887175,0.0868238807,-0.0873017907,0.3527728617,0.0881813541,-0.121158205,0.1119030043,-0.2668753564,0.3709553778,-0.1393034011,-0.0502592623,0.2567106485,-0.1017212197,-0.1936982721,-0.0222703051,0.1220172495,0.6600776315,0.3586020172,0.1172699258,0.4765803516,-0.1983785629,0.3195495307,-0.0597341172,-0.1360106617,-0.4920931756,-0.3510355353,0.1049575582,-0.1960784346,0.1987957358,-0.1476992816,-0.0226633139,0.3582038581,-0.0114521012,0.4813610911,0.181717664,0.1715059429,-0.0177972503,-0.024297215,-0.2199698389,0.187038362,-0.0949707925,0.1076645777,-0.0729456171,-0.039432548,-0.0519838631,-0.1821260899,-0.2247526646,0.1746825129,-0.3567138612,0.092994228,0.2788364887,-0.3486048877,-0.1215511039,0.2856612504,0.1725815684,0.1285102963,-0.1277089417,-0.124090977,0.0180381462,0.0834357515,0.0509496666,-0.0223008767,0.3129682243,-0.1497761309,-0.3154302835,0.17894198,-0.1141532063,-0.1367426813,0.0276512485,0.0460858345,0.0429792181,-0.6083951592,-0.3587333858,-0.2729334235,0.2343661785,-0.2964299023,0.1671108007,-0.0850273147,-0.159299016,-0.0390192419,-0.0390178151,-0.3504946232,-0.188681826,0.2096121609,0.0100413337,0.0955855548,0.4750776589,-0.0024073629,0.085909158,-0.281863451,0.5697032809,-0.1319766641,-0.2841341496,0.1106292233,0.0628202036,-0.1130802184,-0.2623446584,0.1433285475,0.2304039598,-0.0656383932,0.2568435073,-0.4319812059,-0.3417362571,0.1113909185,-0.3017697036,0.2283263505,0.0737322196,-0.0314689614,-0.1762896925,0.1253466308,-0.371548444,-0.0490871407,-0.2450396568,0.1156439707,0.1753619909,-0.0509934947,0.0549703278,0.0774088129,0.2249481529,-0.0035852455,-0.1168234944,-0.3020421863,-0.0865272358,0.1254264265,0.0515467748,-0.0695332363,-0.1507494003,-0.0611952282,-0.1140329838,-0.1327302009,0.0390418172,0.1042219698,-0.0347810686,0.3725389242,-0.1999056935,0.2814030349,-0.1046492755,0.2537442446,-0.1507167965,0.1896493137,-0.2035243213,0.1280552447,-0.1644871533,-0.000014853,-0.2832544148,0.0390093923,-0.0349154696,0.0817560554,0.2742252648,-0.2003487498,-0.2600533068,0.0305856019,0.0199263673,0.3129118681,-0.2150289416,0.1065102741,0.1815644205,0.2663259804,-0.3734115958,-0.010485691,0.1419816762,-0.1060630083,0.0743484572,0.2163880616,0.2126561552,-0.0611200705,0.2493149489,0.0930695385,0.2510064244,0.0173221454,0.0926623568,0.3244889081,-0.2003924251,-0.043881692,0.2114525735,0.1855243593,0.2891336381,0.4851191044,-0.079485178,0.0525735989,-0.2255954295,0.0710886717,0.0361781418,-0.3354212046,0.0672081336,0.0693156421,-0.0936621055,-0.00470267,-0.1472409517,0.0748917535,-0.2594290078,0.0290211085,-0.2763265669,-0.0886882246,-0.0123605458,-0.017395312,-0.2431390733,-0.2479863018,-0.1207344383,-0.053788323,-0.0487490408,-0.1577880383,0.1042184457,-0.1172417551,-0.3449355066,-0.4382255077,0.1962499619,0.4863179624,-0.187797606,-0.3739557862,0.3553467691,0.1638759375,0.1163184866,-0.0050221337,0.2935783565,0.5776576996,0.4492668509,-0.2183886468,-0.0133697409,-0.0903499722,-0.2572478652,-0.08058808,0.1788306087,-0.2036101818,0.2041838914,0.3896346092,0.2425412089,-0.2860693038,0.3087479472,0.0671433136,-0.014772282,-0.6452242136,0.3085484207,0.0373875014,0.0155810714,0.1040226147,-0.0377722569,-0.2956029773,-0.0928971395,0.4650352895,0.2191892415,0.2604660094,0.0643747449,0.1007259563,-0.2586954534,0.2847928107,0.3491569161,0.2334001511,-0.302959919,0.1163394004,-0.6907718182,0.1602863669,-0.0385213606,-0.1857215613,0.145208776,0.1284459084,0.189406991,0.3235723674,-0.2789386511,0.1875156462,0.1158161089,-0.102913402,-0.3438047171,-0.2960200012,-0.3301110864,0.1387993097,0.0364331864,-0.4284615219,0.0283448994,-0.2654493451,0.0856036842,-0.0148275346,-0.1632930338,0.2235354483,0.1154315844,0.4646860659,0.1625410169,0.4427508414,-0.0643838346,-0.0877063125,-0.1593246162,-0.326819092,-0.2241960913,0.0881990641,0.0083235903,0.6604252458,-0.0688739419,-0.186289534,-0.2489366531,0.547285974,0.1704099625,-0.1335785091,0.1080821976,0.0587932207,0.005058486,0.016521344,0.0720723271,0.086332269,-0.1636992246,0.2174746543,-0.0827195421,-0.2598025203,0.4770698547,-0.0727573186,-0.3654350936,-0.0709003955,0.2040002495,-0.0339087285,-0.0137331747,-0.373477608,0.1011847407,0.5264702439,-0.0827589035,-0.2293682843,0.1287205517,0.0001831334,0.119935222,-0.0393447988,0.4087265432,-0.0403168648,-0.1360573173,0.0768616349,-0.1536669582]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2400","title":"Concatenate several datasets with removed columns is not working.","comments":"Hi,\r\n\r\ndid you fill out the env info section manually or by copy-pasting the output of the `datasets-cli env` command?\r\n\r\nThis code should work without issues on 1.6.2 version (I'm working on master (1.6.2.dev0 version) and can't reproduce this error).","body":"## Describe the bug\r\n\r\nYou can't concatenate datasets when you removed columns before.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\nwikiann= load_dataset(\"wikiann\",\"en\")\r\n\r\nwikiann[\"train\"] = wikiann[\"train\"].remove_columns([\"langs\",\"spans\"])\r\nwikiann[\"test\"] = wikiann[\"test\"].remove_columns([\"langs\",\"spans\"])\r\n\r\nassert wikiann[\"train\"].features.type == wikiann[\"test\"].features.type\r\n\r\nconcate = concatenate_datasets([wikiann[\"train\"],wikiann[\"test\"]])\r\n```\r\n\r\n## Expected results\r\nMerged dataset \r\n\r\n\r\n## Actual results\r\n```python\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'ner_tags': Sequence(feature=ClassLabel(num_classes=7, names=['O', 'B-PER', 'I-PER', 'B-ORG', 'I-ORG', 'B-LOC', 'I-LOC'], names_file=None, id=None), length=-1, id=None), 'langs': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'spans': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}\r\nwith type\r\nstruct, ner_tags: list, spans: list, tokens: list>\r\n\r\nbut expected something like\r\n{'ner_tags': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}\r\nwith type\r\nstruct, tokens: list>\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: ~1.6.2~ 1.5.0\r\n- Platform: macos\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":40,"text":"Concatenate several datasets with removed columns is not working.\n## Describe the bug\r\n\r\nYou can't concatenate datasets when you removed columns before.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\nwikiann= load_dataset(\"wikiann\",\"en\")\r\n\r\nwikiann[\"train\"] = wikiann[\"train\"].remove_columns([\"langs\",\"spans\"])\r\nwikiann[\"test\"] = wikiann[\"test\"].remove_columns([\"langs\",\"spans\"])\r\n\r\nassert wikiann[\"train\"].features.type == wikiann[\"test\"].features.type\r\n\r\nconcate = concatenate_datasets([wikiann[\"train\"],wikiann[\"test\"]])\r\n```\r\n\r\n## Expected results\r\nMerged dataset \r\n\r\n\r\n## Actual results\r\n```python\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'ner_tags': Sequence(feature=ClassLabel(num_classes=7, names=['O', 'B-PER', 'I-PER', 'B-ORG', 'I-ORG', 'B-LOC', 'I-LOC'], names_file=None, id=None), length=-1, id=None), 'langs': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'spans': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}\r\nwith type\r\nstruct, ner_tags: list, spans: list, tokens: list>\r\n\r\nbut expected something like\r\n{'ner_tags': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}\r\nwith type\r\nstruct, tokens: list>\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: ~1.6.2~ 1.5.0\r\n- Platform: macos\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n\nHi,\r\n\r\ndid you fill out the env info section manually or by copy-pasting the output of the `datasets-cli env` command?\r\n\r\nThis code should work without issues on 1.6.2 version (I'm working on master (1.6.2.dev0 version) and can't reproduce this error).","embeddings":[-0.0503957234,-0.1630990207,-0.0372494049,0.29215765,0.0712865591,0.1608157754,0.4275573492,0.3972988129,-0.128023386,0.0971976742,-0.2116857469,0.2830114663,0.1164146364,0.2401473969,-0.3315172791,-0.2897301912,0.3033925593,-0.0497661643,-0.1116106212,0.0089246044,-0.2896185219,0.1585718542,-0.5898987651,-0.0812414214,-0.2130750418,0.3191348612,-0.4302580953,0.0101843085,0.0158585962,-0.0107666077,0.2489322722,0.0348747857,0.0911828578,0.2840434313,-0.0001121243,-0.1066203564,0.2169155329,-0.098146677,-0.0510174669,-0.0768572837,-0.3660033047,-0.3289982677,0.0968496427,-0.1380754113,0.0926101133,-0.3934742808,-0.5519912243,-0.3150902092,-0.0737218633,0.3357467353,0.1743545085,0.1843100339,0.1966743767,0.0580199733,0.098164551,-0.1065447256,-0.1400445551,-0.3029312193,-0.0800745711,-0.2223571837,0.4281584024,0.2023564875,-0.1775318682,-0.2188138068,-0.1170001701,0.2738669217,-0.1349679083,-0.3314731121,-0.0674612671,0.1771031767,0.5298755169,-0.1704706401,-0.0786020607,-0.1111541465,0.1492639929,-0.3533147871,0.1071003452,0.2344943136,0.0967376977,0.0004615577,-0.0760485828,0.1861793846,-0.0712400973,0.1990367621,0.0029855049,0.3789429069,0.1802271307,0.0926468223,-0.000508545,-0.2979050875,0.2016520947,-0.451084435,-0.1082627848,0.100753814,-0.6059805155,-0.0505556278,-0.0591483302,-0.220883444,0.1962039024,0.0668276921,-0.3245195746,-0.084399648,0.1913082153,0.1033208817,0.1777590513,0.4177729487,-0.1163810045,0.2942094803,-0.2089640349,0.1199076697,-0.0970709324,0.083201699,0.0753655881,-0.0561545268,0.070139192,0.101112254,0.3909551799,-0.0969191641,-0.1147804409,0.1432804912,-0.3422627747,0.0044983653,-0.1969391853,-0.1263696402,-0.1667243391,0.5428573489,0.0377590321,0.3465742469,0.1597060263,-0.0662202612,-0.2639191151,0.0068230163,-0.0230933372,-0.349783808,-0.0136387125,-0.0729780421,0.0252169706,0.3798228204,0.0718706697,-0.0370393842,-0.2197403163,-0.1956977546,0.2782518864,-0.040623337,0.0026141929,-0.2228974551,0.4038158059,-0.1640163362,0.0380972587,-0.043440748,-0.5051544309,-0.1425156891,-0.6818566918,0.2404864877,0.0318961516,0.0104360897,-0.141025722,0.1947817057,0.1299948692,0.2286318243,-0.1718499511,-0.2043316811,0.0844891518,-0.1165866628,0.3006387353,0.1032018512,-0.2258979231,0.0896844864,0.2380647808,0.024186071,0.4148750305,0.1680238545,0.0080546904,0.0937932432,-0.3190907836,-0.0370072946,0.4073163569,-0.3568253219,-0.0753332525,0.0344850644,0.0019476183,0.396525234,0.0630723238,-0.1725126654,-0.0213235896,0.1155074462,0.3902072012,0.377129674,0.008525488,-0.0484941602,-0.0952870846,-0.3098079562,0.3817209899,-0.0649060011,0.0163189266,0.297275424,0.1458047479,-0.3483148217,0.375438869,-0.2316856831,-0.0468890034,0.3996683955,0.1350973696,0.0140624885,-0.0787636489,-0.0277048089,-0.622395575,0.0030916058,0.0060159098,0.1521741599,0.0274602231,-0.1527986079,-0.0295592137,-0.1535816789,-0.2728900611,0.0649764165,0.1627149731,0.1823699772,0.0888013095,-0.1181537658,-0.222929731,0.4120450318,-0.0862699673,0.0703940541,-0.3407647014,0.3083175719,-0.1723789126,0.032928057,0.1073514447,0.3806806207,0.3077317476,0.031149758,-0.1154230684,0.3030853868,-0.1770817786,-0.1594836861,-0.1415891051,-0.1729774475,0.031404905,-0.1056798846,-0.2074016482,-0.0851076618,-0.0073306849,-0.0969136134,-0.0933882594,0.1859098375,-0.0348238647,0.3872608542,0.0237940811,0.0162364934,0.2036125213,-0.1143120527,-0.0444548801,-0.1632085294,0.1956177205,0.0571594164,-0.2243569642,0.0947116464,-0.3509671688,0.101021491,0.4427131414,0.3745536506,-0.2229179591,-0.0197401363,-0.2566661835,-0.05037193,0.0802748725,-0.0437158383,0.5458464026,0.2856977582,-0.139158994,0.223729834,0.1402375251,-0.0241754893,0.1097457185,0.1719204485,0.0535752065,0.4272288382,0.3137719035,0.1490446478,-0.0310781747,-0.1394143701,0.0814510658,0.1154050753,-0.2603968382,0.0682633147,-0.4254438877,-0.3282284439,-0.0643066242,-0.4015084207,-0.1965873688,-0.5253537297,-0.2588239014,-0.0019736115,-0.2272165418,0.3588680029,-0.2451846302,-0.1407823861,0.1112647876,-0.3164879084,-0.1617892981,-0.0060373219,-0.0109855114,0.1099305451,-0.0284194648,-0.0724929869,-0.1141834706,-0.1195501536,-0.0290251151,-0.2793536782,-0.1802812666,0.0783889368,-0.410856694,-0.0059508011,0.0833566934,-0.1461045146,0.0428666398,-0.5514668822,0.2130687535,0.3447151184,-0.0989946499,0.353197813,0.1951990426,-0.0460549332,-0.0398075543,-0.4187425971,-0.1764497608,-0.3359154463,0.0367100723,0.0865780115,0.0953492224,0.2800819874,0.1913312525,-0.1775089651,0.087542586,0.0280540325,-0.1698313802,-0.172742635,0.1965539306,0.0289864335,-0.1954082847,0.037702404,-0.0466100387,0.1134588271,0.1409336329,-0.1158427373,-0.0908907801,-0.127644971,0.3326511085,-0.1209760234,0.2143722773,0.1329697669,0.1540289819,0.0231206939,0.078750819,0.028014645,0.0431594811,0.1299306899,0.3429623842,-0.2213283777,0.137257725,-0.1495714039,0.1375365853,0.3419882953,-0.0257243589,0.2826379538,0.0222580042,0.3406074941,-0.1352028996,-0.5328831673,-0.1271975487,-0.0814224854,0.3290216029,-0.0329630785,-0.2476583272,-0.1753962487,-0.1729934663,0.2439470887,-0.0694781989,-0.0917678475,-0.0333740674,-0.2583687305,0.2206797302,0.1147601455,-0.26897645,-0.1551436186,-0.1641207337,-0.0245630164,0.0801429674,0.0375051685,-0.1497065872,-0.6010102034,0.0100803887,-0.240525201,0.3405512273,0.2059509307,0.4148744047,0.2138051391,-0.3147104383,-0.3111159503,0.0981573388,0.3010884523,-0.2733204067,-0.5399464965,0.585916996,0.2233323902,-0.1713488251,-0.4675942361,-0.1997636408,-0.1970880181,0.2386059314,0.5842803121,-0.1334958971,-0.1447021216,0.5269443989,0.6180104017,0.1199457422,0.0410640053,-0.3213901818,-0.0627560541,-0.4176593423,-0.2730881572,-0.0228929017,0.5833283067,-0.4107410014,-0.1739822477,-0.1997663677,-0.4016899168,0.2125867307,-0.2919649482,0.3892847598,0.1077715531,0.2085872144,-0.4233523011,0.485882163,0.1032780185,0.4562618136,0.1394287348,-0.5673007369,0.016768517,-0.1674880385,0.7560812235,-0.1930605769,-0.1519735754,0.4494765699,-0.5763666034,0.3119747639,0.0449160822,0.2468320429,0.3549008667,0.0151377106,-0.1996210217,-0.0914015323,0.028181918,0.1854792088,0.0522529595,0.5647614002,-0.0058152722,-0.2855694294,-0.0244538523,-0.2005772591,0.7556119561,0.3028379977,0.0208409261,0.0965687037,0.1147253737,0.3575703502,-0.020463476,0.3085598648,-0.3534913957,-0.1734673232,-0.0389287546,-0.1595134884,0.1411314458,-0.3761194944,-0.355876714,0.2786339819,-0.4931744933,0.7269926667,-0.0347013585,0.188721776,0.10594403,-0.200825125,-0.0680047423,0.0378720872,0.0580184348,-0.1867534369,-0.1229830384,0.0072531677,-0.0799759701,-0.1067744121,0.0135239307,0.4419242442,0.004734118,0.4068088531,-0.0896738097,-0.3423575461,-0.0609219149,0.4606698453,0.2999409139,0.0102734501,0.1020503715,-0.2231286764,0.1134877577,-0.0160720237,0.0467890725,-0.3310846686,-0.0681069046,0.0927124545,-0.3878518939,0.3156060874,0.1516726762,-0.3560869396,-0.2975388467,0.1084222123,0.0818375722,-0.48824507,-0.1700423211,0.1675438434,-0.0017363734,-0.2665569484,0.1225984544,-0.0974287987,-0.2891640663,0.2967380583,-0.0802771449,-0.28687042,-0.1964721233,0.3217976987,0.2862678468,0.0131937442,0.4645836651,0.128436029,0.1328782141,-0.0833978131,0.0675331876,-0.1073498949,-0.1771719009,-0.0076978682,-0.174520269,0.0185005032,-0.0949167833,0.5044787526,-0.2347294986,0.2245818973,0.1139484197,-0.179313913,-0.3401869833,0.1036676019,-0.1466750652,0.0892515704,-0.0959016681,0.3043541014,-0.1747474819,0.1204196364,-0.2747637033,-0.0956493467,-0.2462108284,0.1867581457,0.1189923361,-0.0417108014,0.0259538908,-0.1496899128,0.146457538,0.1957099289,0.253636241,-0.2474799901,-0.0826303363,0.0871705711,0.2038493156,-0.0752913207,-0.090073131,-0.0664062202,-0.2871436179,-0.1995355636,-0.1168175787,-0.0653047189,-0.0417941324,0.2797829211,-0.1537990421,-0.0764093995,-0.036126975,0.0254090372,-0.040943075,0.0646216497,-0.2511429787,-0.063955836,-0.2318916172,-0.0310029108,-0.2864593863,0.025183104,-0.1243507564,0.3092530668,0.2894902825,-0.1083313823,-0.2965265512,-0.1483740807,0.2037281692,0.2329983413,-0.1955141425,0.0803614259,0.0674550384,0.1903170347,-0.4417676926,-0.2918830216,0.0822314247,0.1420915276,0.0977668762,0.1850869507,0.2802190185,-0.155501157,-0.0577224083,-0.0908955112,0.3061224818,0.1604911089,0.4371010959,0.4279775321,0.0571260601,0.1286177039,0.1029726118,0.2378238738,-0.0385002717,0.1931988299,0.1888796687,-0.0699635446,0.2364192307,-0.1155474409,0.0791444108,-0.1605754942,0.3062432408,0.0459764041,0.0083453776,0.109252207,-0.0170332715,-0.1481685191,-0.0365285836,-0.3210131228,-0.0650919452,0.3818079531,-0.1106385514,0.1032072529,-0.0424409807,-0.1661801636,0.1270096898,-0.0202532299,0.1957321912,0.0077174278,0.3294883072,-0.1101324782,-0.0789016336,-0.0990316048,0.2679001093,0.5671166778,0.3819376528,-0.2040128708,-0.0562765338,0.2738361657,0.1424807459,-0.0493477918,0.2027297467,0.5455584526,0.267303437,0.0571308322,-0.0507809147,0.0271811392,-0.2287236303,0.0995417908,0.275002867,0.0248401985,-0.339669466,0.2354063988,0.1432966739,-0.1791848838,0.067043364,0.0911583155,0.1455981582,-0.5377765298,0.4148109257,0.1168230847,-0.2847920656,-0.0198490545,0.0530587211,-0.0716053545,0.0064345393,0.6549490094,0.4100564718,0.2773191929,0.078654699,0.1211827546,0.1426071674,0.2290956527,0.4957567751,0.0465815179,-0.4582920372,-0.2425277382,-0.4170460999,-0.0589077175,-0.0299219433,-0.1018600687,0.1288499981,0.2419574112,0.2251465023,0.1913026273,-0.0598401204,0.2551972866,0.2330652028,-0.2104756683,-0.6309728026,-0.2250903994,0.025339447,0.0410836749,0.0213108491,-0.3583005965,0.394884795,0.1316540837,0.078984946,-0.1651486307,0.0224083401,0.1879859865,0.0349592231,-0.0051151114,0.2527990639,0.3567374945,-0.0489396676,-0.0415803865,-0.0485593788,-0.1991138756,-0.1130388528,0.1264909059,-0.1278738976,0.1283636093,-0.1019973382,0.1006888598,-0.1587141156,0.1035266444,0.4020667374,0.1055741161,-0.3391427696,0.0887743235,0.2914126217,0.0935687795,0.189646095,0.5305503607,-0.0892121941,0.4117039442,-0.1116233096,-0.4936121702,0.2872141898,0.0231086779,-0.6537188888,-0.1910220236,0.1563529372,0.2566809654,0.1443287879,-0.3373329043,-0.0105282618,0.4899207652,-0.1523715109,-0.1201294288,0.3732355535,0.0071295714,0.2497646511,0.024418354,-0.1045162901,0.1468240023,-0.1472362578,0.0055601969,0.0045180423]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2400","title":"Concatenate several datasets with removed columns is not working.","comments":"@mariosasko you are right I was still on `1.5.0`. ","body":"## Describe the bug\r\n\r\nYou can't concatenate datasets when you removed columns before.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\nwikiann= load_dataset(\"wikiann\",\"en\")\r\n\r\nwikiann[\"train\"] = wikiann[\"train\"].remove_columns([\"langs\",\"spans\"])\r\nwikiann[\"test\"] = wikiann[\"test\"].remove_columns([\"langs\",\"spans\"])\r\n\r\nassert wikiann[\"train\"].features.type == wikiann[\"test\"].features.type\r\n\r\nconcate = concatenate_datasets([wikiann[\"train\"],wikiann[\"test\"]])\r\n```\r\n\r\n## Expected results\r\nMerged dataset \r\n\r\n\r\n## Actual results\r\n```python\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'ner_tags': Sequence(feature=ClassLabel(num_classes=7, names=['O', 'B-PER', 'I-PER', 'B-ORG', 'I-ORG', 'B-LOC', 'I-LOC'], names_file=None, id=None), length=-1, id=None), 'langs': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'spans': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}\r\nwith type\r\nstruct, ner_tags: list, spans: list, tokens: list>\r\n\r\nbut expected something like\r\n{'ner_tags': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}\r\nwith type\r\nstruct, tokens: list>\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: ~1.6.2~ 1.5.0\r\n- Platform: macos\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n","comment_length":9,"text":"Concatenate several datasets with removed columns is not working.\n## Describe the bug\r\n\r\nYou can't concatenate datasets when you removed columns before.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset, concatenate_datasets\r\n\r\nwikiann= load_dataset(\"wikiann\",\"en\")\r\n\r\nwikiann[\"train\"] = wikiann[\"train\"].remove_columns([\"langs\",\"spans\"])\r\nwikiann[\"test\"] = wikiann[\"test\"].remove_columns([\"langs\",\"spans\"])\r\n\r\nassert wikiann[\"train\"].features.type == wikiann[\"test\"].features.type\r\n\r\nconcate = concatenate_datasets([wikiann[\"train\"],wikiann[\"test\"]])\r\n```\r\n\r\n## Expected results\r\nMerged dataset \r\n\r\n\r\n## Actual results\r\n```python\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'ner_tags': Sequence(feature=ClassLabel(num_classes=7, names=['O', 'B-PER', 'I-PER', 'B-ORG', 'I-ORG', 'B-LOC', 'I-LOC'], names_file=None, id=None), length=-1, id=None), 'langs': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'spans': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}\r\nwith type\r\nstruct, ner_tags: list, spans: list, tokens: list>\r\n\r\nbut expected something like\r\n{'ner_tags': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)}\r\nwith type\r\nstruct, tokens: list>\r\n```\r\n## Environment info\r\n\r\n- `datasets` version: ~1.6.2~ 1.5.0\r\n- Platform: macos\r\n- Python version: 3.8.5\r\n- PyArrow version: 3.0.0\r\n\n@mariosasko you are right I was still on `1.5.0`. ","embeddings":[-0.0503957234,-0.1630990207,-0.0372494049,0.29215765,0.0712865591,0.1608157754,0.4275573492,0.3972988129,-0.128023386,0.0971976742,-0.2116857469,0.2830114663,0.1164146364,0.2401473969,-0.3315172791,-0.2897301912,0.3033925593,-0.0497661643,-0.1116106212,0.0089246044,-0.2896185219,0.1585718542,-0.5898987651,-0.0812414214,-0.2130750418,0.3191348612,-0.4302580953,0.0101843085,0.0158585962,-0.0107666077,0.2489322722,0.0348747857,0.0911828578,0.2840434313,-0.0001121243,-0.1066203564,0.2169155329,-0.098146677,-0.0510174669,-0.0768572837,-0.3660033047,-0.3289982677,0.0968496427,-0.1380754113,0.0926101133,-0.3934742808,-0.5519912243,-0.3150902092,-0.0737218633,0.3357467353,0.1743545085,0.1843100339,0.1966743767,0.0580199733,0.098164551,-0.1065447256,-0.1400445551,-0.3029312193,-0.0800745711,-0.2223571837,0.4281584024,0.2023564875,-0.1775318682,-0.2188138068,-0.1170001701,0.2738669217,-0.1349679083,-0.3314731121,-0.0674612671,0.1771031767,0.5298755169,-0.1704706401,-0.0786020607,-0.1111541465,0.1492639929,-0.3533147871,0.1071003452,0.2344943136,0.0967376977,0.0004615577,-0.0760485828,0.1861793846,-0.0712400973,0.1990367621,0.0029855049,0.3789429069,0.1802271307,0.0926468223,-0.000508545,-0.2979050875,0.2016520947,-0.451084435,-0.1082627848,0.100753814,-0.6059805155,-0.0505556278,-0.0591483302,-0.220883444,0.1962039024,0.0668276921,-0.3245195746,-0.084399648,0.1913082153,0.1033208817,0.1777590513,0.4177729487,-0.1163810045,0.2942094803,-0.2089640349,0.1199076697,-0.0970709324,0.083201699,0.0753655881,-0.0561545268,0.070139192,0.101112254,0.3909551799,-0.0969191641,-0.1147804409,0.1432804912,-0.3422627747,0.0044983653,-0.1969391853,-0.1263696402,-0.1667243391,0.5428573489,0.0377590321,0.3465742469,0.1597060263,-0.0662202612,-0.2639191151,0.0068230163,-0.0230933372,-0.349783808,-0.0136387125,-0.0729780421,0.0252169706,0.3798228204,0.0718706697,-0.0370393842,-0.2197403163,-0.1956977546,0.2782518864,-0.040623337,0.0026141929,-0.2228974551,0.4038158059,-0.1640163362,0.0380972587,-0.043440748,-0.5051544309,-0.1425156891,-0.6818566918,0.2404864877,0.0318961516,0.0104360897,-0.141025722,0.1947817057,0.1299948692,0.2286318243,-0.1718499511,-0.2043316811,0.0844891518,-0.1165866628,0.3006387353,0.1032018512,-0.2258979231,0.0896844864,0.2380647808,0.024186071,0.4148750305,0.1680238545,0.0080546904,0.0937932432,-0.3190907836,-0.0370072946,0.4073163569,-0.3568253219,-0.0753332525,0.0344850644,0.0019476183,0.396525234,0.0630723238,-0.1725126654,-0.0213235896,0.1155074462,0.3902072012,0.377129674,0.008525488,-0.0484941602,-0.0952870846,-0.3098079562,0.3817209899,-0.0649060011,0.0163189266,0.297275424,0.1458047479,-0.3483148217,0.375438869,-0.2316856831,-0.0468890034,0.3996683955,0.1350973696,0.0140624885,-0.0787636489,-0.0277048089,-0.622395575,0.0030916058,0.0060159098,0.1521741599,0.0274602231,-0.1527986079,-0.0295592137,-0.1535816789,-0.2728900611,0.0649764165,0.1627149731,0.1823699772,0.0888013095,-0.1181537658,-0.222929731,0.4120450318,-0.0862699673,0.0703940541,-0.3407647014,0.3083175719,-0.1723789126,0.032928057,0.1073514447,0.3806806207,0.3077317476,0.031149758,-0.1154230684,0.3030853868,-0.1770817786,-0.1594836861,-0.1415891051,-0.1729774475,0.031404905,-0.1056798846,-0.2074016482,-0.0851076618,-0.0073306849,-0.0969136134,-0.0933882594,0.1859098375,-0.0348238647,0.3872608542,0.0237940811,0.0162364934,0.2036125213,-0.1143120527,-0.0444548801,-0.1632085294,0.1956177205,0.0571594164,-0.2243569642,0.0947116464,-0.3509671688,0.101021491,0.4427131414,0.3745536506,-0.2229179591,-0.0197401363,-0.2566661835,-0.05037193,0.0802748725,-0.0437158383,0.5458464026,0.2856977582,-0.139158994,0.223729834,0.1402375251,-0.0241754893,0.1097457185,0.1719204485,0.0535752065,0.4272288382,0.3137719035,0.1490446478,-0.0310781747,-0.1394143701,0.0814510658,0.1154050753,-0.2603968382,0.0682633147,-0.4254438877,-0.3282284439,-0.0643066242,-0.4015084207,-0.1965873688,-0.5253537297,-0.2588239014,-0.0019736115,-0.2272165418,0.3588680029,-0.2451846302,-0.1407823861,0.1112647876,-0.3164879084,-0.1617892981,-0.0060373219,-0.0109855114,0.1099305451,-0.0284194648,-0.0724929869,-0.1141834706,-0.1195501536,-0.0290251151,-0.2793536782,-0.1802812666,0.0783889368,-0.410856694,-0.0059508011,0.0833566934,-0.1461045146,0.0428666398,-0.5514668822,0.2130687535,0.3447151184,-0.0989946499,0.353197813,0.1951990426,-0.0460549332,-0.0398075543,-0.4187425971,-0.1764497608,-0.3359154463,0.0367100723,0.0865780115,0.0953492224,0.2800819874,0.1913312525,-0.1775089651,0.087542586,0.0280540325,-0.1698313802,-0.172742635,0.1965539306,0.0289864335,-0.1954082847,0.037702404,-0.0466100387,0.1134588271,0.1409336329,-0.1158427373,-0.0908907801,-0.127644971,0.3326511085,-0.1209760234,0.2143722773,0.1329697669,0.1540289819,0.0231206939,0.078750819,0.028014645,0.0431594811,0.1299306899,0.3429623842,-0.2213283777,0.137257725,-0.1495714039,0.1375365853,0.3419882953,-0.0257243589,0.2826379538,0.0222580042,0.3406074941,-0.1352028996,-0.5328831673,-0.1271975487,-0.0814224854,0.3290216029,-0.0329630785,-0.2476583272,-0.1753962487,-0.1729934663,0.2439470887,-0.0694781989,-0.0917678475,-0.0333740674,-0.2583687305,0.2206797302,0.1147601455,-0.26897645,-0.1551436186,-0.1641207337,-0.0245630164,0.0801429674,0.0375051685,-0.1497065872,-0.6010102034,0.0100803887,-0.240525201,0.3405512273,0.2059509307,0.4148744047,0.2138051391,-0.3147104383,-0.3111159503,0.0981573388,0.3010884523,-0.2733204067,-0.5399464965,0.585916996,0.2233323902,-0.1713488251,-0.4675942361,-0.1997636408,-0.1970880181,0.2386059314,0.5842803121,-0.1334958971,-0.1447021216,0.5269443989,0.6180104017,0.1199457422,0.0410640053,-0.3213901818,-0.0627560541,-0.4176593423,-0.2730881572,-0.0228929017,0.5833283067,-0.4107410014,-0.1739822477,-0.1997663677,-0.4016899168,0.2125867307,-0.2919649482,0.3892847598,0.1077715531,0.2085872144,-0.4233523011,0.485882163,0.1032780185,0.4562618136,0.1394287348,-0.5673007369,0.016768517,-0.1674880385,0.7560812235,-0.1930605769,-0.1519735754,0.4494765699,-0.5763666034,0.3119747639,0.0449160822,0.2468320429,0.3549008667,0.0151377106,-0.1996210217,-0.0914015323,0.028181918,0.1854792088,0.0522529595,0.5647614002,-0.0058152722,-0.2855694294,-0.0244538523,-0.2005772591,0.7556119561,0.3028379977,0.0208409261,0.0965687037,0.1147253737,0.3575703502,-0.020463476,0.3085598648,-0.3534913957,-0.1734673232,-0.0389287546,-0.1595134884,0.1411314458,-0.3761194944,-0.355876714,0.2786339819,-0.4931744933,0.7269926667,-0.0347013585,0.188721776,0.10594403,-0.200825125,-0.0680047423,0.0378720872,0.0580184348,-0.1867534369,-0.1229830384,0.0072531677,-0.0799759701,-0.1067744121,0.0135239307,0.4419242442,0.004734118,0.4068088531,-0.0896738097,-0.3423575461,-0.0609219149,0.4606698453,0.2999409139,0.0102734501,0.1020503715,-0.2231286764,0.1134877577,-0.0160720237,0.0467890725,-0.3310846686,-0.0681069046,0.0927124545,-0.3878518939,0.3156060874,0.1516726762,-0.3560869396,-0.2975388467,0.1084222123,0.0818375722,-0.48824507,-0.1700423211,0.1675438434,-0.0017363734,-0.2665569484,0.1225984544,-0.0974287987,-0.2891640663,0.2967380583,-0.0802771449,-0.28687042,-0.1964721233,0.3217976987,0.2862678468,0.0131937442,0.4645836651,0.128436029,0.1328782141,-0.0833978131,0.0675331876,-0.1073498949,-0.1771719009,-0.0076978682,-0.174520269,0.0185005032,-0.0949167833,0.5044787526,-0.2347294986,0.2245818973,0.1139484197,-0.179313913,-0.3401869833,0.1036676019,-0.1466750652,0.0892515704,-0.0959016681,0.3043541014,-0.1747474819,0.1204196364,-0.2747637033,-0.0956493467,-0.2462108284,0.1867581457,0.1189923361,-0.0417108014,0.0259538908,-0.1496899128,0.146457538,0.1957099289,0.253636241,-0.2474799901,-0.0826303363,0.0871705711,0.2038493156,-0.0752913207,-0.090073131,-0.0664062202,-0.2871436179,-0.1995355636,-0.1168175787,-0.0653047189,-0.0417941324,0.2797829211,-0.1537990421,-0.0764093995,-0.036126975,0.0254090372,-0.040943075,0.0646216497,-0.2511429787,-0.063955836,-0.2318916172,-0.0310029108,-0.2864593863,0.025183104,-0.1243507564,0.3092530668,0.2894902825,-0.1083313823,-0.2965265512,-0.1483740807,0.2037281692,0.2329983413,-0.1955141425,0.0803614259,0.0674550384,0.1903170347,-0.4417676926,-0.2918830216,0.0822314247,0.1420915276,0.0977668762,0.1850869507,0.2802190185,-0.155501157,-0.0577224083,-0.0908955112,0.3061224818,0.1604911089,0.4371010959,0.4279775321,0.0571260601,0.1286177039,0.1029726118,0.2378238738,-0.0385002717,0.1931988299,0.1888796687,-0.0699635446,0.2364192307,-0.1155474409,0.0791444108,-0.1605754942,0.3062432408,0.0459764041,0.0083453776,0.109252207,-0.0170332715,-0.1481685191,-0.0365285836,-0.3210131228,-0.0650919452,0.3818079531,-0.1106385514,0.1032072529,-0.0424409807,-0.1661801636,0.1270096898,-0.0202532299,0.1957321912,0.0077174278,0.3294883072,-0.1101324782,-0.0789016336,-0.0990316048,0.2679001093,0.5671166778,0.3819376528,-0.2040128708,-0.0562765338,0.2738361657,0.1424807459,-0.0493477918,0.2027297467,0.5455584526,0.267303437,0.0571308322,-0.0507809147,0.0271811392,-0.2287236303,0.0995417908,0.275002867,0.0248401985,-0.339669466,0.2354063988,0.1432966739,-0.1791848838,0.067043364,0.0911583155,0.1455981582,-0.5377765298,0.4148109257,0.1168230847,-0.2847920656,-0.0198490545,0.0530587211,-0.0716053545,0.0064345393,0.6549490094,0.4100564718,0.2773191929,0.078654699,0.1211827546,0.1426071674,0.2290956527,0.4957567751,0.0465815179,-0.4582920372,-0.2425277382,-0.4170460999,-0.0589077175,-0.0299219433,-0.1018600687,0.1288499981,0.2419574112,0.2251465023,0.1913026273,-0.0598401204,0.2551972866,0.2330652028,-0.2104756683,-0.6309728026,-0.2250903994,0.025339447,0.0410836749,0.0213108491,-0.3583005965,0.394884795,0.1316540837,0.078984946,-0.1651486307,0.0224083401,0.1879859865,0.0349592231,-0.0051151114,0.2527990639,0.3567374945,-0.0489396676,-0.0415803865,-0.0485593788,-0.1991138756,-0.1130388528,0.1264909059,-0.1278738976,0.1283636093,-0.1019973382,0.1006888598,-0.1587141156,0.1035266444,0.4020667374,0.1055741161,-0.3391427696,0.0887743235,0.2914126217,0.0935687795,0.189646095,0.5305503607,-0.0892121941,0.4117039442,-0.1116233096,-0.4936121702,0.2872141898,0.0231086779,-0.6537188888,-0.1910220236,0.1563529372,0.2566809654,0.1443287879,-0.3373329043,-0.0105282618,0.4899207652,-0.1523715109,-0.1201294288,0.3732355535,0.0071295714,0.2497646511,0.024418354,-0.1045162901,0.1468240023,-0.1472362578,0.0055601969,0.0045180423]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2396","title":"strange datasets from OSCAR corpus","comments":"Hi ! Thanks for reporting\r\ncc @pjox is this an issue from the data ?\r\n\r\nAnyway we should at least mention that OSCAR could contain such contents in the dataset card, you're totally right @jerryIsHere ","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/119260850-4f876b80-bc07-11eb-8894-124302600643.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/119260875-675eef80-bc07-11eb-9da4-ee27567054ac.png)\r\nFrom the [official site ](https:\/\/oscar-corpus.com\/), the Yue Chinese dataset should have 2.2KB data.\r\n7 training instances is obviously not a right number.\r\nAs I can read Yue Chinese, I call tell the last instance is definitely not something that would appear on Common Crawl.\r\nAnd even if you don't read Yue Chinese, you can tell the first six instance are problematic.\r\n(It is embarrassing, as the 7 training instances look exactly like something from a pornographic novel or flitting messages in a chat of a dating app)\r\nIt might not be the problem of the huggingface\/datasets implementation, because when I tried to download the dataset from the official site, I found out that the zip file is corrupted.\r\nI will try to inform the host of OSCAR corpus later.\r\nAwy a remake about this dataset in huggingface\/datasets is needed, perhaps after the host of the dataset fixes the issue.\r\n\r\n> Hi @jerryIsHere , sorry for the late response! Sadly this is normal, the problem comes form fasttext's classifier which we used to create the original corpus. In general the classifier is not really capable of properly recognizing Yue Chineese so the file ends un being just noise from Common Crawl. Some of these problems with OSCAR were already discussed [here](https:\/\/arxiv.org\/pdf\/2103.12028.pdf) but we are working on explicitly documenting the problems by language on our website. In fact, could please you open an issue on [our repo](https:\/\/github.com\/oscar-corpus\/oscar-website\/issues) as well so that we can track it?\r\n\r\nThanks a lot, the new post is here:\r\nhttps:\/\/github.com\/oscar-corpus\/oscar-website\/issues\/11","comment_length":35,"text":"strange datasets from OSCAR corpus\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/119260850-4f876b80-bc07-11eb-8894-124302600643.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/119260875-675eef80-bc07-11eb-9da4-ee27567054ac.png)\r\nFrom the [official site ](https:\/\/oscar-corpus.com\/), the Yue Chinese dataset should have 2.2KB data.\r\n7 training instances is obviously not a right number.\r\nAs I can read Yue Chinese, I call tell the last instance is definitely not something that would appear on Common Crawl.\r\nAnd even if you don't read Yue Chinese, you can tell the first six instance are problematic.\r\n(It is embarrassing, as the 7 training instances look exactly like something from a pornographic novel or flitting messages in a chat of a dating app)\r\nIt might not be the problem of the huggingface\/datasets implementation, because when I tried to download the dataset from the official site, I found out that the zip file is corrupted.\r\nI will try to inform the host of OSCAR corpus later.\r\nAwy a remake about this dataset in huggingface\/datasets is needed, perhaps after the host of the dataset fixes the issue.\r\n\r\n> Hi @jerryIsHere , sorry for the late response! Sadly this is normal, the problem comes form fasttext's classifier which we used to create the original corpus. In general the classifier is not really capable of properly recognizing Yue Chineese so the file ends un being just noise from Common Crawl. Some of these problems with OSCAR were already discussed [here](https:\/\/arxiv.org\/pdf\/2103.12028.pdf) but we are working on explicitly documenting the problems by language on our website. In fact, could please you open an issue on [our repo](https:\/\/github.com\/oscar-corpus\/oscar-website\/issues) as well so that we can track it?\r\n\r\nThanks a lot, the new post is here:\r\nhttps:\/\/github.com\/oscar-corpus\/oscar-website\/issues\/11\nHi ! Thanks for reporting\r\ncc @pjox is this an issue from the data ?\r\n\r\nAnyway we should at least mention that OSCAR could contain such contents in the dataset card, you're totally right @jerryIsHere ","embeddings":[0.2676993608,0.057094764,-0.0023013265,0.5770170689,0.1565007865,0.0609702878,-0.0124235256,0.2492677867,-0.2645931542,0.0486133806,-0.579642415,-0.1079981476,0.0431049913,-0.051175423,-0.0094940485,-0.301930815,0.110553138,0.0599582233,0.0344043113,-0.369253844,-0.1291408092,0.2438743412,-0.0991414487,0.0684380084,-0.0104771517,-0.1464854181,-0.2549023926,-0.0072285933,-0.3976078331,0.0282757394,0.1691304445,0.1257212758,0.0447246879,0.378584832,-0.0001098784,-0.1675397009,0.2241498679,-0.032616768,0.1237274259,-0.0597777031,0.1777801663,0.0427329913,-0.0099441046,0.0341181159,-0.2434695065,-0.2776968777,0.0040472257,0.1497946978,0.2936551273,0.410572499,0.2379986495,0.0630983114,-0.1171702296,-0.1336971223,0.2072796822,-0.0323365964,-0.2481697649,0.1810347885,0.466270566,0.0761901438,-0.1121683344,0.0576442331,0.0283341352,-0.101253666,0.2170051336,-0.1111785099,-0.2650502622,-0.4842260778,0.2932800949,0.5078215599,0.307980001,-0.0811569616,-0.2589426637,-0.2738285959,0.0988934413,-0.2739084661,0.2932356894,0.3606124222,0.0745738298,0.2289150208,-0.2088944763,0.1954556257,-0.1180684865,0.4085975289,-0.2678520679,-0.1681371033,-0.223840341,0.0197889302,0.0670833662,-0.0670693293,0.0395493396,-0.2790008485,0.0216593854,0.2727837563,-0.1698675603,-0.1353237629,0.0378287397,0.4480695426,0.1993468702,-0.1269877404,-0.0557646081,0.1898595542,-0.3269468844,-0.0471593998,0.3354088068,0.1453732997,-0.1572799683,-0.1377139837,0.1423854232,0.1614499837,0.0849004611,0.2207645029,-0.0777042508,-0.0842675045,-0.2245463133,-0.3745216727,0.0991557464,-0.3770408928,-0.2749187946,0.3003405631,-0.6244720817,0.2917198241,0.0357688703,0.1298946887,-0.1998814493,0.0667819455,-0.016715385,0.2029186487,-0.1194482073,-0.0397366621,-0.1921678632,0.2545002103,-0.217064783,0.2372613102,0.0401417799,-0.2705197036,0.2493536174,0.1856439114,0.3698456585,-0.1369044632,-0.0799858868,-0.4004633129,-0.0195614137,0.0929225683,-0.1745463461,0.1127903014,0.0614734329,-0.3261088133,-0.1222709864,-0.1741286218,-0.3078811765,-0.1431825906,0.0786640644,0.1830638647,-0.4335425794,0.0327018574,-0.3430885971,0.3751347363,0.0194555651,0.3101088107,0.3069826961,-0.0848392099,-0.2614415288,0.110479854,-0.1510209292,0.3771541417,-0.0065252073,-0.0835282505,-0.1661023796,0.0815068334,0.3079614937,0.2400632799,-0.0228678118,-0.0866745338,-0.2789590657,0.1016659811,-0.0555678383,-0.2338879108,-0.3346047997,0.2996109128,0.1844694763,0.2973280251,-0.1442996413,-0.1274283528,0.2322862893,-0.2734360099,0.2217368335,0.1740197837,0.0513248742,-0.027263226,-0.3290061653,-0.1187138036,-0.358600378,0.0662762225,0.130272001,0.0879968554,0.0359005406,0.0622565188,0.3574254215,0.026874952,0.0745497197,0.13262631,-0.1339467168,0.4023646414,0.1660578102,0.0027954392,0.0745985582,0.2480973452,-0.3026051223,0.5085480213,0.0180819575,-0.2257502228,-0.1997607797,-0.2386162728,-0.2820656598,-0.245531112,0.1467761993,-0.184961915,0.0019979251,0.3076104522,0.0463864841,0.0455984473,-0.2320450544,0.022844201,-0.5470091105,0.1783732176,0.0564264841,0.1313604116,0.1921941787,0.2265807688,0.0022634515,-0.1288478822,0.0188172571,0.0711189955,-0.3784909248,0.5216875076,0.0105420277,0.3390645385,0.4842693806,-0.3351532221,0.186759457,0.0962930098,0.205712527,-0.0595745817,-0.1192198545,0.4944604635,0.1703856438,-0.1687858999,0.0706616193,0.0164971575,0.283459574,-0.1336004138,-0.2227505445,-0.2548031509,0.424082756,0.1479568034,0.4548326135,0.1943281591,-0.4106912315,0.2208998203,0.0080121411,-0.0442878269,0.2007290721,0.3062259853,-0.1937873214,-0.0257208925,0.0227153599,0.2468966693,0.1487173736,0.0625743568,-0.0749768093,0.3186788559,-0.069458954,-0.3558135331,0.111846745,0.1962320656,0.4446726143,0.0999088213,0.3605704606,0.220251292,-0.7133963704,0.126008302,-0.0119821373,0.3289736509,-0.1542913318,0.2327584177,-0.1093545407,-0.3838168383,-0.425598979,-0.0526062325,-0.1197228208,-0.0765793249,0.1917120218,-0.0079616001,-0.0710900426,-0.0606718212,-0.2533044517,0.2126707733,-0.2307349592,0.3920391798,-0.097000666,-0.0001325482,-0.1199717075,0.1299186945,0.1205610186,0.0156219909,0.426179111,-0.2234457135,-0.144994095,-0.2055267841,-0.5277351737,0.3550101519,-0.037032783,-0.0828411877,-0.0709609762,-0.1230314299,-0.2948217392,-0.0460335873,0.0550760813,-0.1752235442,-0.1702314615,0.056853313,-0.1404699534,-0.0086305682,0.0384075195,-0.4563492537,-0.2169178724,-0.0600899011,0.2215566635,0.1342093796,0.1609433293,0.4600666165,-0.2288878113,0.140054822,0.1669894904,0.0807909295,-0.5246050954,-0.3567255139,0.3421804607,-0.3204010725,-0.4378070235,-0.20665133,-0.2681742311,-0.1136468872,0.0419772081,-0.3773085177,-0.5895934105,-0.2654289603,-0.1950018555,0.390293777,-0.3890113235,0.0661180988,-0.3487103283,-0.1124316305,0.0002544484,0.3013922572,-0.0172216762,0.0075142249,0.2562912703,-0.2389841676,0.5512993932,0.0627559274,0.3488258719,0.2001673132,0.0785247684,0.4007724524,0.1514521688,0.2562434077,-0.0319219492,-0.1340349764,0.3567915261,-0.1609951556,-0.0110346237,0.4110300839,0.1014012247,-0.2588830292,-0.0115151526,0.1257117689,-0.270486176,-0.2327979058,0.0137277488,-0.0752975866,-0.0444323197,0.0619645007,-0.0168288313,0.2466303259,-0.4781698287,-0.0459046029,0.0849940479,-0.0483802557,-0.0480614416,-0.1555660069,0.0361824594,-0.6853789091,0.1299283803,0.0055808201,0.2157566249,-0.0981483981,0.3015866578,0.2229267806,0.1142501011,0.6046202183,-0.0360771455,0.0197583586,0.0640874878,-0.1593836099,-0.2227322459,-0.1761957109,-0.4290166199,0.2636863589,0.2155922055,0.2350151688,-0.1400407255,-0.0076076165,0.2108382881,0.0438339673,-0.1988004595,0.0452601276,-0.0140930125,-0.5783287287,0.020486258,-0.0566003509,0.257024318,0.2461138368,-0.2465652525,-0.1783549041,0.0273977965,0.3921889961,0.522521019,0.0822655186,0.0778018087,0.5562979579,0.1265050322,0.0485975221,0.2794081867,0.2785663009,0.7305167317,-0.1383529007,-0.4596145451,-0.078700617,-0.3730379641,-0.0539411902,0.5012969971,0.0733194873,-0.046170041,-0.3145380318,-0.0212493297,-0.2173012048,0.1982852817,0.367354095,0.0678473562,-0.3099002242,-0.338165164,0.6486569643,0.097301513,-0.2117157727,-0.0075569572,0.265858382,-0.2391379178,0.3573741317,0.012272316,0.9542484283,-0.1912836432,-0.0442245007,0.2648425996,0.2007787973,0.4890199304,0.0048596938,0.0470102727,-0.2552776039,-0.4762350917,-0.0891338885,0.1125535965,0.1524897963,0.2524732053,0.0504174754,0.169568792,0.1507980525,-0.3307522833,-0.00462418,-0.0930545852,-0.1097794473,-0.1424431354,-0.3858094513,0.2335760146,0.0878824666,0.4907717109,-0.025132807,0.0482202098,-0.3262656331,0.0106008183,-0.5220407248,0.0588058867,0.1309580505,-0.0944531858,0.3296045959,-0.5177643895,0.2315041274,0.0937837586,0.3234200776,-0.212633267,-0.2467211932,0.3399108648,-0.0834050551,-0.3022833765,0.3012394905,-0.0905864313,0.34712556,-0.1234992743,-0.1297503412,0.1222688556,0.2607758045,0.1445859969,-0.2115234286,-0.0985134766,-0.0890083164,-0.1729493141,-0.4047572017,-0.1194863096,-0.1400160342,-0.1010028943,0.1534444988,-0.015741393,0.0184017122,-0.0176607948,0.0893049836,-0.3160211146,-0.0297373142,0.4177772403,0.2908807993,-0.0866720974,0.3345596194,0.5006690025,-0.0924457535,-0.0569555759,-0.135657534,0.1240203381,-0.3993449807,0.0166614223,-0.1370998621,-0.1621250659,-0.1353639364,0.0718333498,-0.0321010873,-0.169067204,0.0348090865,-0.4083505571,0.048429247,0.0493490361,0.309458226,0.0582229756,-0.0381126143,0.0904550254,0.2750759423,0.1498869061,-0.3072505891,0.2171390355,-0.2439662814,0.033765275,-0.1528746337,0.0050278138,0.0345686302,-0.2459265292,-0.0381523594,0.0866533071,-0.1406113356,-0.1784694791,-0.0995471925,0.1121714562,-0.0125547107,-0.3512524366,0.2863205373,0.1491298378,-0.0890470892,-0.2008609176,0.1019810736,0.5271870494,0.205179289,-0.0790420622,0.0798615143,-0.1667254865,0.0778205767,-0.0986037776,0.0689814985,0.3401399851,0.1839011163,-0.0672482625,0.1160735115,-0.2392149568,-0.0607522205,-0.0328315385,-0.3738975823,-0.2426214814,0.4297305942,-0.0269519109,0.251141727,0.695482254,0.0882526487,0.3521155119,-0.3342868686,0.0167474337,-0.0096740825,0.1583522409,-0.3926720321,-0.0477947257,-0.0476361886,0.2072638869,-0.0001071624,0.0749138817,0.0460642464,-0.022994088,-0.2722858787,0.008051252,0.5071793199,0.0163915567,0.208509326,0.0689454377,0.1197313145,0.0585184842,0.2436468154,-0.0928816646,0.2050593495,0.2713142633,-0.0106765134,0.3332244456,-0.281311661,0.3489289582,0.3693645298,-0.3195890188,0.013630989,0.5468193889,-0.2157413661,0.2831245661,0.1977006644,0.2938272357,0.1392150819,-0.1310464293,-0.3772868216,0.1210148484,-0.0039990265,-0.1015880853,-0.0990433469,-0.3409675062,-0.1125148088,0.1773215979,-0.330996722,-0.1919046789,0.2041882277,-0.0504714921,-0.4578523338,-0.3864658773,0.0886508971,-0.0775413811,0.3063214719,0.0372155793,0.3571665287,0.2162098289,-0.1343719065,0.1002368554,0.5791752934,0.129008323,0.1282622367,0.3204135299,0.2209004164,-0.1628935188,0.0165252779,0.0122791892,0.1575547904,0.1201130524,-0.099198252,0.0856005996,0.1103232279,-0.1372419596,0.2281878144,0.0672517046,0.1438181996,-0.1229054779,0.4894866645,-0.1464986503,0.0709644929,-0.2475923002,0.0226247218,-0.3075389564,0.2617521882,0.2810641229,-0.2151386142,0.1608298123,-0.1471676677,0.0899276584,-0.0447107702,0.3532955348,0.581543088,-0.0098432153,-0.0143454662,-0.0663926601,-0.5804973245,0.4264128506,-0.2836047113,0.1714906693,0.0004597301,0.1692210883,0.0245569777,0.1825418174,-0.3031076193,0.3341662586,-0.1446931362,-0.128556788,-0.3643918633,0.1117894948,-0.2892955244,0.1427312642,-0.0977642462,-0.0098533118,0.3696789145,-0.3764726818,0.0249701776,-0.1294029355,0.0040037222,-0.4783974886,-0.0495420694,0.2290337682,0.2319302708,0.0133247739,-0.1163659021,-0.0666055903,-0.2614999413,-0.2494395971,-0.0704268068,0.4824530482,-0.0602832101,-0.131870538,0.1580566913,-0.0932491198,0.240411669,0.3770562708,0.0144608496,-0.173925817,-0.4429989457,0.0399757475,-0.1874047965,-0.0352264531,-0.0608854108,0.55699265,-0.1725472063,-0.1392384917,0.0391079038,-0.5822050571,0.3299417496,-0.5559088588,0.1658035368,-0.2232458889,0.1830234975,0.0688657761,-0.2526189685,-0.4450642467,0.1104285643,0.1024456322,-0.0768706501,-0.1648370922,-0.0664818808,-0.0641322657,-0.1986686438,-0.1243117973,0.2940542698,0.2435918748,-0.2361119092,0.3124830723,-0.0227929354]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2396","title":"strange datasets from OSCAR corpus","comments":"Hi @jerryIsHere , sorry for the late response! Sadly this is normal, the problem comes form fasttext's classifier which we used to create the original corpus. In general the classifier is not really capable of properly recognizing Yue Chineese so the file ends un being just noise from Common Crawl. Some of these problems with OSCAR were already discussed [here](https:\/\/arxiv.org\/pdf\/2103.12028.pdf) but we are working on explicitly documenting the problems by language on our website. In fact, could please you open an issue on [our repo](https:\/\/github.com\/oscar-corpus\/oscar-website\/issues) as well so that we can track it?","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/119260850-4f876b80-bc07-11eb-8894-124302600643.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/119260875-675eef80-bc07-11eb-9da4-ee27567054ac.png)\r\nFrom the [official site ](https:\/\/oscar-corpus.com\/), the Yue Chinese dataset should have 2.2KB data.\r\n7 training instances is obviously not a right number.\r\nAs I can read Yue Chinese, I call tell the last instance is definitely not something that would appear on Common Crawl.\r\nAnd even if you don't read Yue Chinese, you can tell the first six instance are problematic.\r\n(It is embarrassing, as the 7 training instances look exactly like something from a pornographic novel or flitting messages in a chat of a dating app)\r\nIt might not be the problem of the huggingface\/datasets implementation, because when I tried to download the dataset from the official site, I found out that the zip file is corrupted.\r\nI will try to inform the host of OSCAR corpus later.\r\nAwy a remake about this dataset in huggingface\/datasets is needed, perhaps after the host of the dataset fixes the issue.\r\n\r\n> Hi @jerryIsHere , sorry for the late response! Sadly this is normal, the problem comes form fasttext's classifier which we used to create the original corpus. In general the classifier is not really capable of properly recognizing Yue Chineese so the file ends un being just noise from Common Crawl. Some of these problems with OSCAR were already discussed [here](https:\/\/arxiv.org\/pdf\/2103.12028.pdf) but we are working on explicitly documenting the problems by language on our website. In fact, could please you open an issue on [our repo](https:\/\/github.com\/oscar-corpus\/oscar-website\/issues) as well so that we can track it?\r\n\r\nThanks a lot, the new post is here:\r\nhttps:\/\/github.com\/oscar-corpus\/oscar-website\/issues\/11","comment_length":93,"text":"strange datasets from OSCAR corpus\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/119260850-4f876b80-bc07-11eb-8894-124302600643.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/119260875-675eef80-bc07-11eb-9da4-ee27567054ac.png)\r\nFrom the [official site ](https:\/\/oscar-corpus.com\/), the Yue Chinese dataset should have 2.2KB data.\r\n7 training instances is obviously not a right number.\r\nAs I can read Yue Chinese, I call tell the last instance is definitely not something that would appear on Common Crawl.\r\nAnd even if you don't read Yue Chinese, you can tell the first six instance are problematic.\r\n(It is embarrassing, as the 7 training instances look exactly like something from a pornographic novel or flitting messages in a chat of a dating app)\r\nIt might not be the problem of the huggingface\/datasets implementation, because when I tried to download the dataset from the official site, I found out that the zip file is corrupted.\r\nI will try to inform the host of OSCAR corpus later.\r\nAwy a remake about this dataset in huggingface\/datasets is needed, perhaps after the host of the dataset fixes the issue.\r\n\r\n> Hi @jerryIsHere , sorry for the late response! Sadly this is normal, the problem comes form fasttext's classifier which we used to create the original corpus. In general the classifier is not really capable of properly recognizing Yue Chineese so the file ends un being just noise from Common Crawl. Some of these problems with OSCAR were already discussed [here](https:\/\/arxiv.org\/pdf\/2103.12028.pdf) but we are working on explicitly documenting the problems by language on our website. In fact, could please you open an issue on [our repo](https:\/\/github.com\/oscar-corpus\/oscar-website\/issues) as well so that we can track it?\r\n\r\nThanks a lot, the new post is here:\r\nhttps:\/\/github.com\/oscar-corpus\/oscar-website\/issues\/11\nHi @jerryIsHere , sorry for the late response! Sadly this is normal, the problem comes form fasttext's classifier which we used to create the original corpus. In general the classifier is not really capable of properly recognizing Yue Chineese so the file ends un being just noise from Common Crawl. Some of these problems with OSCAR were already discussed [here](https:\/\/arxiv.org\/pdf\/2103.12028.pdf) but we are working on explicitly documenting the problems by language on our website. In fact, could please you open an issue on [our repo](https:\/\/github.com\/oscar-corpus\/oscar-website\/issues) as well so that we can track it?","embeddings":[0.2676993608,0.057094764,-0.0023013265,0.5770170689,0.1565007865,0.0609702878,-0.0124235256,0.2492677867,-0.2645931542,0.0486133806,-0.579642415,-0.1079981476,0.0431049913,-0.051175423,-0.0094940485,-0.301930815,0.110553138,0.0599582233,0.0344043113,-0.369253844,-0.1291408092,0.2438743412,-0.0991414487,0.0684380084,-0.0104771517,-0.1464854181,-0.2549023926,-0.0072285933,-0.3976078331,0.0282757394,0.1691304445,0.1257212758,0.0447246879,0.378584832,-0.0001098784,-0.1675397009,0.2241498679,-0.032616768,0.1237274259,-0.0597777031,0.1777801663,0.0427329913,-0.0099441046,0.0341181159,-0.2434695065,-0.2776968777,0.0040472257,0.1497946978,0.2936551273,0.410572499,0.2379986495,0.0630983114,-0.1171702296,-0.1336971223,0.2072796822,-0.0323365964,-0.2481697649,0.1810347885,0.466270566,0.0761901438,-0.1121683344,0.0576442331,0.0283341352,-0.101253666,0.2170051336,-0.1111785099,-0.2650502622,-0.4842260778,0.2932800949,0.5078215599,0.307980001,-0.0811569616,-0.2589426637,-0.2738285959,0.0988934413,-0.2739084661,0.2932356894,0.3606124222,0.0745738298,0.2289150208,-0.2088944763,0.1954556257,-0.1180684865,0.4085975289,-0.2678520679,-0.1681371033,-0.223840341,0.0197889302,0.0670833662,-0.0670693293,0.0395493396,-0.2790008485,0.0216593854,0.2727837563,-0.1698675603,-0.1353237629,0.0378287397,0.4480695426,0.1993468702,-0.1269877404,-0.0557646081,0.1898595542,-0.3269468844,-0.0471593998,0.3354088068,0.1453732997,-0.1572799683,-0.1377139837,0.1423854232,0.1614499837,0.0849004611,0.2207645029,-0.0777042508,-0.0842675045,-0.2245463133,-0.3745216727,0.0991557464,-0.3770408928,-0.2749187946,0.3003405631,-0.6244720817,0.2917198241,0.0357688703,0.1298946887,-0.1998814493,0.0667819455,-0.016715385,0.2029186487,-0.1194482073,-0.0397366621,-0.1921678632,0.2545002103,-0.217064783,0.2372613102,0.0401417799,-0.2705197036,0.2493536174,0.1856439114,0.3698456585,-0.1369044632,-0.0799858868,-0.4004633129,-0.0195614137,0.0929225683,-0.1745463461,0.1127903014,0.0614734329,-0.3261088133,-0.1222709864,-0.1741286218,-0.3078811765,-0.1431825906,0.0786640644,0.1830638647,-0.4335425794,0.0327018574,-0.3430885971,0.3751347363,0.0194555651,0.3101088107,0.3069826961,-0.0848392099,-0.2614415288,0.110479854,-0.1510209292,0.3771541417,-0.0065252073,-0.0835282505,-0.1661023796,0.0815068334,0.3079614937,0.2400632799,-0.0228678118,-0.0866745338,-0.2789590657,0.1016659811,-0.0555678383,-0.2338879108,-0.3346047997,0.2996109128,0.1844694763,0.2973280251,-0.1442996413,-0.1274283528,0.2322862893,-0.2734360099,0.2217368335,0.1740197837,0.0513248742,-0.027263226,-0.3290061653,-0.1187138036,-0.358600378,0.0662762225,0.130272001,0.0879968554,0.0359005406,0.0622565188,0.3574254215,0.026874952,0.0745497197,0.13262631,-0.1339467168,0.4023646414,0.1660578102,0.0027954392,0.0745985582,0.2480973452,-0.3026051223,0.5085480213,0.0180819575,-0.2257502228,-0.1997607797,-0.2386162728,-0.2820656598,-0.245531112,0.1467761993,-0.184961915,0.0019979251,0.3076104522,0.0463864841,0.0455984473,-0.2320450544,0.022844201,-0.5470091105,0.1783732176,0.0564264841,0.1313604116,0.1921941787,0.2265807688,0.0022634515,-0.1288478822,0.0188172571,0.0711189955,-0.3784909248,0.5216875076,0.0105420277,0.3390645385,0.4842693806,-0.3351532221,0.186759457,0.0962930098,0.205712527,-0.0595745817,-0.1192198545,0.4944604635,0.1703856438,-0.1687858999,0.0706616193,0.0164971575,0.283459574,-0.1336004138,-0.2227505445,-0.2548031509,0.424082756,0.1479568034,0.4548326135,0.1943281591,-0.4106912315,0.2208998203,0.0080121411,-0.0442878269,0.2007290721,0.3062259853,-0.1937873214,-0.0257208925,0.0227153599,0.2468966693,0.1487173736,0.0625743568,-0.0749768093,0.3186788559,-0.069458954,-0.3558135331,0.111846745,0.1962320656,0.4446726143,0.0999088213,0.3605704606,0.220251292,-0.7133963704,0.126008302,-0.0119821373,0.3289736509,-0.1542913318,0.2327584177,-0.1093545407,-0.3838168383,-0.425598979,-0.0526062325,-0.1197228208,-0.0765793249,0.1917120218,-0.0079616001,-0.0710900426,-0.0606718212,-0.2533044517,0.2126707733,-0.2307349592,0.3920391798,-0.097000666,-0.0001325482,-0.1199717075,0.1299186945,0.1205610186,0.0156219909,0.426179111,-0.2234457135,-0.144994095,-0.2055267841,-0.5277351737,0.3550101519,-0.037032783,-0.0828411877,-0.0709609762,-0.1230314299,-0.2948217392,-0.0460335873,0.0550760813,-0.1752235442,-0.1702314615,0.056853313,-0.1404699534,-0.0086305682,0.0384075195,-0.4563492537,-0.2169178724,-0.0600899011,0.2215566635,0.1342093796,0.1609433293,0.4600666165,-0.2288878113,0.140054822,0.1669894904,0.0807909295,-0.5246050954,-0.3567255139,0.3421804607,-0.3204010725,-0.4378070235,-0.20665133,-0.2681742311,-0.1136468872,0.0419772081,-0.3773085177,-0.5895934105,-0.2654289603,-0.1950018555,0.390293777,-0.3890113235,0.0661180988,-0.3487103283,-0.1124316305,0.0002544484,0.3013922572,-0.0172216762,0.0075142249,0.2562912703,-0.2389841676,0.5512993932,0.0627559274,0.3488258719,0.2001673132,0.0785247684,0.4007724524,0.1514521688,0.2562434077,-0.0319219492,-0.1340349764,0.3567915261,-0.1609951556,-0.0110346237,0.4110300839,0.1014012247,-0.2588830292,-0.0115151526,0.1257117689,-0.270486176,-0.2327979058,0.0137277488,-0.0752975866,-0.0444323197,0.0619645007,-0.0168288313,0.2466303259,-0.4781698287,-0.0459046029,0.0849940479,-0.0483802557,-0.0480614416,-0.1555660069,0.0361824594,-0.6853789091,0.1299283803,0.0055808201,0.2157566249,-0.0981483981,0.3015866578,0.2229267806,0.1142501011,0.6046202183,-0.0360771455,0.0197583586,0.0640874878,-0.1593836099,-0.2227322459,-0.1761957109,-0.4290166199,0.2636863589,0.2155922055,0.2350151688,-0.1400407255,-0.0076076165,0.2108382881,0.0438339673,-0.1988004595,0.0452601276,-0.0140930125,-0.5783287287,0.020486258,-0.0566003509,0.257024318,0.2461138368,-0.2465652525,-0.1783549041,0.0273977965,0.3921889961,0.522521019,0.0822655186,0.0778018087,0.5562979579,0.1265050322,0.0485975221,0.2794081867,0.2785663009,0.7305167317,-0.1383529007,-0.4596145451,-0.078700617,-0.3730379641,-0.0539411902,0.5012969971,0.0733194873,-0.046170041,-0.3145380318,-0.0212493297,-0.2173012048,0.1982852817,0.367354095,0.0678473562,-0.3099002242,-0.338165164,0.6486569643,0.097301513,-0.2117157727,-0.0075569572,0.265858382,-0.2391379178,0.3573741317,0.012272316,0.9542484283,-0.1912836432,-0.0442245007,0.2648425996,0.2007787973,0.4890199304,0.0048596938,0.0470102727,-0.2552776039,-0.4762350917,-0.0891338885,0.1125535965,0.1524897963,0.2524732053,0.0504174754,0.169568792,0.1507980525,-0.3307522833,-0.00462418,-0.0930545852,-0.1097794473,-0.1424431354,-0.3858094513,0.2335760146,0.0878824666,0.4907717109,-0.025132807,0.0482202098,-0.3262656331,0.0106008183,-0.5220407248,0.0588058867,0.1309580505,-0.0944531858,0.3296045959,-0.5177643895,0.2315041274,0.0937837586,0.3234200776,-0.212633267,-0.2467211932,0.3399108648,-0.0834050551,-0.3022833765,0.3012394905,-0.0905864313,0.34712556,-0.1234992743,-0.1297503412,0.1222688556,0.2607758045,0.1445859969,-0.2115234286,-0.0985134766,-0.0890083164,-0.1729493141,-0.4047572017,-0.1194863096,-0.1400160342,-0.1010028943,0.1534444988,-0.015741393,0.0184017122,-0.0176607948,0.0893049836,-0.3160211146,-0.0297373142,0.4177772403,0.2908807993,-0.0866720974,0.3345596194,0.5006690025,-0.0924457535,-0.0569555759,-0.135657534,0.1240203381,-0.3993449807,0.0166614223,-0.1370998621,-0.1621250659,-0.1353639364,0.0718333498,-0.0321010873,-0.169067204,0.0348090865,-0.4083505571,0.048429247,0.0493490361,0.309458226,0.0582229756,-0.0381126143,0.0904550254,0.2750759423,0.1498869061,-0.3072505891,0.2171390355,-0.2439662814,0.033765275,-0.1528746337,0.0050278138,0.0345686302,-0.2459265292,-0.0381523594,0.0866533071,-0.1406113356,-0.1784694791,-0.0995471925,0.1121714562,-0.0125547107,-0.3512524366,0.2863205373,0.1491298378,-0.0890470892,-0.2008609176,0.1019810736,0.5271870494,0.205179289,-0.0790420622,0.0798615143,-0.1667254865,0.0778205767,-0.0986037776,0.0689814985,0.3401399851,0.1839011163,-0.0672482625,0.1160735115,-0.2392149568,-0.0607522205,-0.0328315385,-0.3738975823,-0.2426214814,0.4297305942,-0.0269519109,0.251141727,0.695482254,0.0882526487,0.3521155119,-0.3342868686,0.0167474337,-0.0096740825,0.1583522409,-0.3926720321,-0.0477947257,-0.0476361886,0.2072638869,-0.0001071624,0.0749138817,0.0460642464,-0.022994088,-0.2722858787,0.008051252,0.5071793199,0.0163915567,0.208509326,0.0689454377,0.1197313145,0.0585184842,0.2436468154,-0.0928816646,0.2050593495,0.2713142633,-0.0106765134,0.3332244456,-0.281311661,0.3489289582,0.3693645298,-0.3195890188,0.013630989,0.5468193889,-0.2157413661,0.2831245661,0.1977006644,0.2938272357,0.1392150819,-0.1310464293,-0.3772868216,0.1210148484,-0.0039990265,-0.1015880853,-0.0990433469,-0.3409675062,-0.1125148088,0.1773215979,-0.330996722,-0.1919046789,0.2041882277,-0.0504714921,-0.4578523338,-0.3864658773,0.0886508971,-0.0775413811,0.3063214719,0.0372155793,0.3571665287,0.2162098289,-0.1343719065,0.1002368554,0.5791752934,0.129008323,0.1282622367,0.3204135299,0.2209004164,-0.1628935188,0.0165252779,0.0122791892,0.1575547904,0.1201130524,-0.099198252,0.0856005996,0.1103232279,-0.1372419596,0.2281878144,0.0672517046,0.1438181996,-0.1229054779,0.4894866645,-0.1464986503,0.0709644929,-0.2475923002,0.0226247218,-0.3075389564,0.2617521882,0.2810641229,-0.2151386142,0.1608298123,-0.1471676677,0.0899276584,-0.0447107702,0.3532955348,0.581543088,-0.0098432153,-0.0143454662,-0.0663926601,-0.5804973245,0.4264128506,-0.2836047113,0.1714906693,0.0004597301,0.1692210883,0.0245569777,0.1825418174,-0.3031076193,0.3341662586,-0.1446931362,-0.128556788,-0.3643918633,0.1117894948,-0.2892955244,0.1427312642,-0.0977642462,-0.0098533118,0.3696789145,-0.3764726818,0.0249701776,-0.1294029355,0.0040037222,-0.4783974886,-0.0495420694,0.2290337682,0.2319302708,0.0133247739,-0.1163659021,-0.0666055903,-0.2614999413,-0.2494395971,-0.0704268068,0.4824530482,-0.0602832101,-0.131870538,0.1580566913,-0.0932491198,0.240411669,0.3770562708,0.0144608496,-0.173925817,-0.4429989457,0.0399757475,-0.1874047965,-0.0352264531,-0.0608854108,0.55699265,-0.1725472063,-0.1392384917,0.0391079038,-0.5822050571,0.3299417496,-0.5559088588,0.1658035368,-0.2232458889,0.1830234975,0.0688657761,-0.2526189685,-0.4450642467,0.1104285643,0.1024456322,-0.0768706501,-0.1648370922,-0.0664818808,-0.0641322657,-0.1986686438,-0.1243117973,0.2940542698,0.2435918748,-0.2361119092,0.3124830723,-0.0227929354]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2391","title":"Missing original answers in kilt-TriviaQA","comments":"That could be useful indeed! Feel free to open a PR on the dataset card if you already have some code that runs, otherwise we'll take care of it soon :) ","body":"I previously opened an issue at https:\/\/github.com\/facebookresearch\/KILT\/issues\/42 but from the answer of @fabiopetroni it seems that the problem comes from HF-datasets\r\n\r\n## Describe the bug\r\nThe `answer` field in kilt-TriviaQA, e.g. `kilt_tasks['train_triviaqa'][0]['output']['answer']` contains a list of alternative answer which are accepted for the question. \r\nHowever it'd be nice to know the original answer to the question (the only fields in `output` are `'answer', 'meta', 'provenance'`)\r\n\r\n## How to fix\r\nIt can be fixed by retrieving the original answer from the original TriviaQA (e.g. `trivia_qa['train'][0]['answer']['value']`), perhaps at the same place as here where one retrieves the questions https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md#loading-the-kilt-knowledge-source-and-task-data\r\n\r\ncc @yjernite who previously answered to an issue about KILT and TriviaQA :)\r\n","comment_length":31,"text":"Missing original answers in kilt-TriviaQA\nI previously opened an issue at https:\/\/github.com\/facebookresearch\/KILT\/issues\/42 but from the answer of @fabiopetroni it seems that the problem comes from HF-datasets\r\n\r\n## Describe the bug\r\nThe `answer` field in kilt-TriviaQA, e.g. `kilt_tasks['train_triviaqa'][0]['output']['answer']` contains a list of alternative answer which are accepted for the question. \r\nHowever it'd be nice to know the original answer to the question (the only fields in `output` are `'answer', 'meta', 'provenance'`)\r\n\r\n## How to fix\r\nIt can be fixed by retrieving the original answer from the original TriviaQA (e.g. `trivia_qa['train'][0]['answer']['value']`), perhaps at the same place as here where one retrieves the questions https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md#loading-the-kilt-knowledge-source-and-task-data\r\n\r\ncc @yjernite who previously answered to an issue about KILT and TriviaQA :)\r\n\nThat could be useful indeed! Feel free to open a PR on the dataset card if you already have some code that runs, otherwise we'll take care of it soon :) ","embeddings":[0.5064159632,-0.3483679891,-0.0351329036,0.0628515109,-0.1213426515,-0.1011349186,0.034442164,0.3133405745,0.1268104464,0.1838567704,0.1640157402,0.4927667677,0.1122973189,0.4052688479,-0.2509842813,0.3889521956,-0.0829049274,-0.0019121934,-0.0266753733,-0.3187006116,-0.2322140038,0.2161713094,-0.1054576263,-0.0127488822,-0.4187035561,0.1789601743,-0.1225842908,0.0886476785,0.1447677314,-0.3618231416,0.4138162434,-0.0295042358,-0.0571988709,-0.0511866622,-0.0001163996,-0.2085256577,0.0355243646,-0.141021058,-0.4676496387,-0.0007738866,-0.5212789774,0.0164583251,-0.0597766973,-0.1613026559,-0.0574857295,0.2864459753,-0.0024450046,-0.0210050289,0.3094422221,0.2542654574,0.1770555824,-0.0932435393,-0.063484855,-0.1800219119,0.0356934071,0.2446768582,-0.1647231281,0.1539984494,-0.1439371705,-0.0503535867,0.0279170666,0.3840641975,0.1803435981,0.1095787734,-0.2104799598,-0.1323056668,0.1236590222,-0.2465537935,0.4007653296,0.2841539085,0.2348448187,0.1206990033,-0.3248027265,-0.1977267116,0.0932543203,0.288092196,-0.0109733539,0.1244026572,-0.1363507509,-0.2952623367,-0.2086613029,-0.02564626,-0.0678828508,-0.0072240941,-0.1617991924,0.2719786465,0.1087460071,0.041688405,-0.1180529445,-0.064709954,0.2020545453,-0.0071485466,-0.2675997019,0.2320914716,-0.4236135185,-0.201063782,0.1482287794,0.0433905087,0.2846264243,-0.2984972894,0.0577431023,0.1936496496,-0.2878252864,-0.1157380044,0.2814102471,0.2249249816,0.5015596151,-0.4662697911,-0.0686959401,0.1793825924,-0.1848450154,-0.1055990234,-0.0697763488,0.0285260156,-0.1266579926,-0.1790535748,0.0918540284,-0.3960404992,-0.220307827,0.0945565924,-0.3354539275,0.0928730145,-0.0802250504,0.0914254412,-0.3331416547,0.4611908793,-0.1675022542,0.2119443864,-0.3096483052,-0.2978350818,-0.1636331826,-0.1647340506,0.0661306605,0.0875089914,0.2834359705,-0.2787396312,0.4228960276,0.2885068953,0.0934127346,-0.1077329889,-0.0583772771,-0.1930044442,0.0104149515,0.1265088022,0.0996813327,0.1406677067,0.170772627,-0.0101910122,0.1328339279,0.3715587854,-0.4798253477,-0.3926911652,-0.1743836254,0.2275592387,-0.0689651892,-0.1529335082,-0.098171398,0.2556852996,-0.1350388378,-0.133900553,-0.0644677281,-0.3494915366,-0.0867917091,-0.2528090179,0.4844788313,0.1443109065,-0.2544125319,0.2270882875,-0.2028786838,0.0399902239,-0.2231209129,0.1565094143,-0.0392991826,0.3329613209,-0.1608744562,-0.0407970771,-0.105517827,-0.4473587275,-0.2160711437,-0.364700228,0.4545496404,0.3787327707,-0.240792349,0.1538425535,0.0745173544,0.1185933501,-0.0344050825,0.0907148793,-0.1299792081,-0.1017844528,-0.3159190416,-0.3424827456,0.1604639441,0.2800517678,0.0927965119,-0.0462395847,0.2145369053,0.1177990288,0.2015682906,0.1085855216,-0.2322513312,0.2903354466,0.4023075998,0.1829581261,0.1544663459,0.1252215505,-0.295643419,0.1648569554,-0.5367688537,0.0007433234,0.375613749,-0.275592804,-0.554350853,0.1562990099,-0.1554647237,-0.5120695233,0.1259919703,-0.1458701342,0.2678463161,0.0769694597,0.0319223143,0.2687189877,-0.290915519,-0.1769288331,-0.159297958,0.2640773654,-0.019435944,-0.1374590546,-0.0875332206,0.636172533,0.269698292,0.3276682198,-0.0062105502,0.2320399284,0.1752087325,0.1612212062,0.4861148596,-0.0401581004,0.2118989676,-0.0589986183,-0.2136869878,0.2125492543,-0.0642785951,-0.0561990663,-0.3204424679,0.2087258101,0.1455164999,0.1726152003,-0.137897402,0.1327901781,-0.215197295,-0.2184007019,0.2137721628,0.0241881274,-0.0226678252,-0.2658396661,-0.535695672,-0.03125415,-0.3318933249,0.3433724344,0.2486389577,0.2554905713,-0.0049298019,0.0634896159,-0.1412683576,-0.0893693864,0.4690452218,-0.2126735896,0.1278372258,0.166465044,-0.128238827,-0.0613431446,0.0182388797,-0.0580957755,-0.0592058152,-0.0200524386,0.2152731419,0.2801452875,0.6623122692,0.1440820098,-0.103091754,0.4766434431,-0.0511283539,-0.0771931186,-0.0545729361,-0.1799441725,-0.477512002,-0.1386460364,0.0692698434,0.0777663663,-0.1331335306,-0.4419929683,0.4497723579,0.3484996855,-0.1111549437,0.1979820728,0.001284398,0.1198353022,-0.2222640663,0.049632296,-0.3405997157,-0.4228145182,-0.4497844875,0.1223601997,-0.3684006333,-0.0903138518,0.2718051076,-0.0971284881,-0.3501998782,-0.5005816817,-0.5326234102,0.0099324714,0.1049496457,0.4307731986,-0.0346926339,0.3675716221,-0.6405918002,-0.2677495778,0.3842469156,-0.035001751,-0.1797939241,0.0280367061,0.0755793452,0.1794508398,-0.1399714947,-0.506370604,-0.0688341036,-0.0853633359,0.212743938,0.2330769449,0.1755038351,0.298068136,0.0535352863,-0.1522396058,-0.0593041852,-0.0652659982,-0.4701882303,-0.1478026062,0.3093015254,-0.1650629789,-0.3541282117,0.1435091197,-0.0879173577,0.2715161443,-0.0638272166,-0.5980352163,-0.6656985879,0.0528208502,0.4026936293,-0.2056811154,-0.094350636,0.134178102,-0.1361105442,-0.0195530448,-0.0448328629,-0.025064338,0.4879376292,0.426646769,0.5124357939,0.1280546635,0.0136329252,0.3042178452,0.7666119337,0.1936040968,0.5757521987,0.1775548905,-0.2589359879,0.1625192612,-0.0556806438,-0.1149135381,0.5098050833,-0.0782857612,0.1934153587,0.2823203504,0.0999527797,0.567021966,0.0473700203,0.3638904095,-0.3983494341,-0.4199623764,-0.2484639585,-0.1325901151,0.3250523508,0.2249948233,0.1674158871,-0.0775937662,-0.4030824304,0.2956237495,-0.1099240705,0.1770304441,0.0153236231,-0.2180565596,-0.2417607605,-0.1092478707,0.1904704422,0.0765500218,0.1864427626,-0.0629480705,0.1542728394,0.2369973063,-0.0557325259,0.7960611582,0.0846259966,-0.101981394,-0.1867235005,0.065167509,-0.2904493809,-0.3532454073,-0.0397939309,-0.3415675759,0.017859675,0.2351996005,-0.249916777,-0.0148185585,0.1009588167,0.1128629446,-0.2615755796,0.1712518334,-0.098867394,-0.3827979267,-0.2567311227,-0.096759364,-0.095036827,0.2531138062,-0.0929266736,0.0238668174,-0.0120588038,0.1606690288,0.0480755083,0.24130252,0.2478301972,0.0851692781,0.1172995046,-0.0155271245,-0.1903509796,0.066843003,0.3774348795,0.2302328646,0.183760941,0.0732130334,-0.2065120935,0.393271029,0.2801283598,0.0807282105,0.2001470029,-0.1146983579,-0.0163936522,-0.2932943404,0.1788850427,0.1789172292,-0.2708420157,0.0600631684,-0.0104098879,0.2207981348,-0.1354476213,-0.0936426371,0.0685280189,0.2277286798,-0.0387290344,0.0613758527,0.298455894,0.8620378375,0.2510136664,0.0091869365,0.2357157767,-0.188904658,0.1107184812,-0.2218530029,0.2238613367,-0.3860090971,-0.0746997669,0.1703315675,0.0738084987,-0.2233560085,0.0923312828,0.101116389,0.3213037252,0.1428895146,0.2470908761,-0.0429474711,0.1438771784,-0.1422431469,-0.0621830747,0.0630836338,0.2091884911,-0.1976959556,0.3709493876,-0.1852802038,-0.1061220989,-0.0573601685,0.1074712202,-0.1681906581,0.2297087163,-0.4546374083,0.2138388306,0.6036265492,-0.2568200231,-0.2494297624,0.4904197454,0.3904775977,0.2837065756,-0.1608925313,0.2527780831,0.3619234264,0.2905690968,0.1596663147,0.3011656404,-0.0441927128,-0.2004941255,-0.2744786739,0.2391584218,-0.2606873214,0.1532793492,-0.5899584889,-0.0484356172,0.1494723409,-0.0534892045,-0.1584602743,-0.0726192072,0.1473394334,-0.1602448672,0.1208741963,-0.2276438177,0.1610534191,0.2469837219,-0.0065854131,-0.1457173824,-0.1351134032,0.3775758445,0.074324131,0.0356282331,0.6234513521,-0.1352857053,-0.076297231,-0.0436789058,0.0039996523,-0.4881559908,-0.0785829723,0.0283741057,0.0598429553,-0.2092085183,0.0018560152,0.6113401055,0.0977057293,-0.0517054498,0.1108690128,-0.4042427242,-0.0231331382,0.5424612164,-0.1096951589,-0.0116479751,0.0896503106,-0.0129199922,-0.082152985,0.0850663036,-0.3287646472,-0.1148810536,-0.3234301507,0.2987617552,-0.2868504524,0.0571608692,0.1276669055,-0.1636975855,0.0520794876,-0.0788236558,-0.0122041712,-0.2355417907,-0.3841152787,0.0905056223,0.2504940629,0.1839337796,-0.0198661983,-0.3337651193,-0.0573528595,0.0127571514,0.0325047486,0.238720268,-0.0132551044,0.18521218,-0.0736936778,0.3311601579,-0.2495123893,0.1607386619,0.0807543397,0.0236901697,0.1116962358,-0.2004224807,-0.1917161793,0.0619911812,-0.0235579293,-0.0134965852,-0.0685669407,-0.222032696,0.1084036902,-0.0838566571,-0.0439926535,0.0526218675,-0.0861187503,-0.1279671788,-0.2177357674,0.028618291,0.2521513402,0.1769888997,-0.2502527237,0.2626273036,-0.1607589126,-0.1608959287,-0.1667327583,0.1170612425,0.3037760258,-0.2903404832,0.5298941731,0.2549230754,-0.2279793471,-0.6408807039,0.2101260126,0.0719323382,-0.0914486572,-0.0221449286,0.0974088907,-0.3312944174,-0.2036069334,0.0768251792,0.0181046072,0.1511261612,0.2797660232,-0.0919452012,0.4834863544,0.1185492575,0.0429401174,0.0517752357,-0.1283738464,0.2372884005,0.0968291312,0.3769885302,-0.1475054026,-0.3596107066,-0.1484083533,0.1534838229,-0.2363644987,-0.0405377597,-0.0969620496,-0.1626701653,-0.1764695197,0.1278991997,0.0158062391,-0.2733189464,0.0212886985,0.2561477125,-0.3647534847,-0.3657858372,-0.2174108922,0.229673475,0.4608760178,0.1824729592,0.4139089584,-0.0393182449,0.1923700124,0.2209993601,0.3500199616,-0.072174795,0.1619295925,-0.1938936263,-0.0745090172,0.1242268756,0.0568597242,0.2854766548,0.2052726746,0.3876189291,-0.1374131143,0.2548476458,0.1557367444,-0.1746132374,-0.1642604917,-0.0253732782,-0.0942310169,-0.2613272667,0.2215280682,0.2376087755,-0.0374144986,0.2884159982,0.0523326807,-0.0962535739,-0.1368668377,-0.3797746003,0.3199318349,0.0954286978,-0.0156355724,0.0930591077,-0.140992254,-0.0770708621,0.3124872446,-0.1272165328,-0.2280492038,-0.1602298319,-0.4929161668,-0.0255816672,0.223879531,0.0563351028,0.2680434883,-0.0726590455,-0.1488180012,-0.0042564124,-0.1019177884,-0.5042887926,0.2965972424,-0.0493113287,-0.2549828589,0.0658000857,-0.0294590965,-0.206483081,0.0119885383,-0.0348472223,0.2386120707,0.0460577235,-0.067853108,0.2699091434,-0.2649444342,0.1679950804,-0.0620010868,-0.1175932959,0.0600352883,0.0688100308,0.0579304658,-0.0877125338,0.2034641206,-0.2216168344,0.0733709559,-0.2331159115,-0.4398472309,0.3352454305,0.305906117,-0.2100493014,0.0481630974,0.3895558119,0.3441247642,0.0108140009,-0.2174219638,0.2059721202,0.0136938049,0.0295113437,0.3768521845,0.1416673213,-0.019489361,0.2824307382,-0.2910391986,-0.2156056017,0.3001195788,-0.2305030674,-0.1259000748,-0.2456979454,0.393378526,-0.2405364513,-0.1775010228,-0.5544116497,-0.3222292066,0.3366212547,0.0032441858,0.418428719,0.0111279404,-0.1815105379,0.0367986485,0.0118322633,0.0938411877,0.0879660845,0.0302498732,0.1788499951,-0.0241483301]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2391","title":"Missing original answers in kilt-TriviaQA","comments":"I can open a PR but there is 2 details to fix:\r\n- the name for the corresponding key (e.g. `original_answer`)\r\n- how to implement it: I\u2019m not sure what happens when you map `lambda x: {'input': ...}`\u00a0as it keeps the other keys (e.g. `output`) intact but here since we want to set a nested value (e.g. `x['output']['original_answer']`) I implemented it with a regular function (not lambda), see below\r\n\r\n```py\r\ndef add_original_answer(x, trivia_qa, triviaqa_map):\r\n i = triviaqa_map[x['id']]\r\n x['output']['original_answer'] = trivia_qa['validation'][i]['answer']['value']\r\n return x\r\n```","body":"I previously opened an issue at https:\/\/github.com\/facebookresearch\/KILT\/issues\/42 but from the answer of @fabiopetroni it seems that the problem comes from HF-datasets\r\n\r\n## Describe the bug\r\nThe `answer` field in kilt-TriviaQA, e.g. `kilt_tasks['train_triviaqa'][0]['output']['answer']` contains a list of alternative answer which are accepted for the question. \r\nHowever it'd be nice to know the original answer to the question (the only fields in `output` are `'answer', 'meta', 'provenance'`)\r\n\r\n## How to fix\r\nIt can be fixed by retrieving the original answer from the original TriviaQA (e.g. `trivia_qa['train'][0]['answer']['value']`), perhaps at the same place as here where one retrieves the questions https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md#loading-the-kilt-knowledge-source-and-task-data\r\n\r\ncc @yjernite who previously answered to an issue about KILT and TriviaQA :)\r\n","comment_length":84,"text":"Missing original answers in kilt-TriviaQA\nI previously opened an issue at https:\/\/github.com\/facebookresearch\/KILT\/issues\/42 but from the answer of @fabiopetroni it seems that the problem comes from HF-datasets\r\n\r\n## Describe the bug\r\nThe `answer` field in kilt-TriviaQA, e.g. `kilt_tasks['train_triviaqa'][0]['output']['answer']` contains a list of alternative answer which are accepted for the question. \r\nHowever it'd be nice to know the original answer to the question (the only fields in `output` are `'answer', 'meta', 'provenance'`)\r\n\r\n## How to fix\r\nIt can be fixed by retrieving the original answer from the original TriviaQA (e.g. `trivia_qa['train'][0]['answer']['value']`), perhaps at the same place as here where one retrieves the questions https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md#loading-the-kilt-knowledge-source-and-task-data\r\n\r\ncc @yjernite who previously answered to an issue about KILT and TriviaQA :)\r\n\nI can open a PR but there is 2 details to fix:\r\n- the name for the corresponding key (e.g. `original_answer`)\r\n- how to implement it: I\u2019m not sure what happens when you map `lambda x: {'input': ...}`\u00a0as it keeps the other keys (e.g. `output`) intact but here since we want to set a nested value (e.g. `x['output']['original_answer']`) I implemented it with a regular function (not lambda), see below\r\n\r\n```py\r\ndef add_original_answer(x, trivia_qa, triviaqa_map):\r\n i = triviaqa_map[x['id']]\r\n x['output']['original_answer'] = trivia_qa['validation'][i]['answer']['value']\r\n return x\r\n```","embeddings":[0.5155783296,-0.3537389636,0.0327610411,0.0395886675,-0.105386138,-0.0799774006,-0.0295903832,0.2565177381,0.085573785,0.1321024299,0.0442913473,0.5437101126,0.1936984211,0.4238483906,-0.1813199818,0.3287062645,0.0641255006,0.1050164923,0.1597800553,-0.2960401475,-0.3145171106,0.2330386192,-0.1170250624,0.1646067649,-0.1306224465,0.0120377354,-0.2015926093,0.21834369,0.2492164373,-0.3835995197,0.3044094443,-0.0027334217,-0.1335724294,-0.0683760867,-0.0001235717,-0.2483267486,0.0100177377,-0.2631385028,-0.4476844966,0.0135608371,-0.5491896272,0.0013693378,-0.2267566025,-0.2673223615,0.0246863365,0.0820418149,-0.0023390166,-0.181188643,0.2089129686,0.1528659314,0.0594027154,-0.1457809657,0.1589805633,-0.0494781211,0.0174915977,0.0762605891,-0.1484265029,0.1695866734,-0.1361159682,-0.0244756676,0.1385612339,0.2619105279,0.105464071,0.0110791102,-0.1540766358,-0.0243223608,0.10231556,-0.244240582,0.4118763208,0.1882626712,0.1229106709,-0.2038338333,-0.4319516122,-0.2443827987,0.1036714762,-0.018734375,-0.0171048008,0.0892703384,-0.2156305611,-0.2821560502,-0.00090461,-0.2265105546,-0.0220113378,0.1320800334,-0.055072777,0.3479349017,0.2165877968,0.1588949263,-0.1401011348,-0.0632673502,0.3329878449,-0.1604586393,-0.1640904397,0.3152612746,-0.3240547776,-0.1331782192,0.1900509894,0.0019526612,0.2189777344,-0.4318755567,-0.0086097261,0.1560154557,-0.2777463198,-0.0269539673,0.2008716017,0.2189187109,0.5501311421,-0.2920511663,-0.0707457289,0.2034399956,-0.1683296859,-0.1551692039,0.0187576879,0.1357354522,-0.0358962528,-0.0002326736,0.1213153899,-0.2009133399,-0.3044298589,0.3365494013,-0.5707329512,0.1031196564,-0.2290992439,0.1706618965,-0.1619618982,0.4370584786,-0.119920671,0.2700347304,-0.2904263437,-0.1402242482,-0.088785097,-0.0078456737,0.0948832482,0.2193794847,0.1299312264,-0.355532378,0.4627596736,0.3198022246,0.1934595704,-0.0461257696,-0.0313192494,-0.1281021386,0.0169504937,0.1240085885,0.1192691326,0.0273681041,0.2130442262,-0.2978673875,0.0743862987,0.3816873133,-0.3751229048,-0.3407919407,-0.1022927612,0.0808336139,-0.1557133645,-0.0852613896,-0.1102893203,0.3784490824,-0.0152195701,-0.1495497972,-0.0393542536,-0.3106988668,-0.1329434663,-0.3905701339,0.49024266,0.1699148417,-0.3541273773,0.3064566851,-0.2360414863,0.0928541571,-0.3066458702,-0.0370559171,-0.076302357,0.4838697016,-0.1619303674,0.0490621962,-0.2314925045,-0.4575146437,-0.3383847177,-0.3009569645,0.4830967486,0.4284089208,-0.0870445669,0.0214253701,0.0319995806,0.0131465476,0.0061029866,-0.1040413007,-0.1317987144,-0.214095667,-0.1762363464,-0.1944406778,-0.049631577,0.0566412881,0.0829337388,0.1037125587,0.283506453,0.2077239007,0.1905883551,0.0045289639,-0.1778160781,0.3082348406,0.3007472456,0.1162420735,0.1203731447,0.0259790253,-0.3274600208,0.1223377064,-0.6618936062,-0.0308636203,0.1369944066,-0.3331814706,-0.5493906736,0.3154106736,-0.1641771197,-0.6335079074,0.0482662804,-0.0674963593,0.2100677788,-0.0569680668,-0.0661991537,0.2212019414,-0.1994565427,-0.185448423,-0.2481945306,0.1694604307,-0.0260955114,-0.0751713812,-0.1915444434,0.5666322708,0.355817914,0.3527083099,0.0031313398,0.3100954294,0.2255421579,0.0509434752,0.4634577036,0.0723986775,0.225365907,0.082031548,-0.1245186478,0.2991172373,0.0801994801,0.0168047603,-0.3568072915,0.421233654,0.2299394757,0.0837266445,-0.1474848092,-0.0554867722,-0.2449639142,-0.2046156079,0.1883871853,-0.0484557226,-0.1712999642,-0.2022147179,-0.5047810674,-0.1734983027,-0.3660624325,0.4657456279,0.1779556125,0.0015858189,0.0422763675,-0.0474720486,-0.140778631,-0.0348974094,0.4081193805,-0.0663329139,0.3968331814,0.1243698671,-0.1609617919,-0.0058463388,-0.0208608415,0.036013227,0.1278193295,-0.0164298099,0.0743787959,0.4520234168,0.6341477036,0.0820695981,-0.0746798068,0.4535393715,-0.0603403263,0.0384703092,-0.1819293648,-0.1642486453,-0.4214442372,0.0934764221,0.0367738456,0.19537884,0.0722384825,-0.5293933153,0.3032277822,0.4172869325,-0.0395120159,0.2344897091,0.045224268,0.0601693951,-0.0857849643,-0.0692597851,-0.2800633609,-0.4454108179,-0.3175086677,0.0101327766,-0.4948932827,-0.211940378,0.1414636821,0.1508740634,-0.3360295892,-0.6223651767,-0.6542856097,-0.0059413896,-0.0045577791,0.5484629273,0.0428466164,0.3344559669,-0.6263720393,-0.4282755256,0.3463867903,0.1127221212,-0.3175826967,0.022467725,-0.0236435961,0.1038113162,0.0179196876,-0.5704700351,-0.0489381142,-0.0977127925,0.1817075014,0.0674074367,0.2584240735,0.102878049,0.0457065217,-0.2438624799,0.1191970482,-0.1969721168,-0.5478495359,0.0836739764,0.3846223652,-0.2477894425,-0.2864775956,-0.0700969845,-0.3900149167,0.2580229342,-0.0927009881,-0.4573445916,-0.6468146443,0.1500904858,0.5276089907,-0.1340635866,0.0861618742,0.1564029455,0.0242378414,0.0743472874,-0.1985616088,-0.1410773993,0.4161976576,0.3931500316,0.5974636674,0.3126388192,0.0723790601,0.3485159576,0.7757428288,0.2300544679,0.5763300061,0.1780343205,-0.3186384737,0.1277631819,-0.034983214,-0.2192827463,0.2881546319,-0.1039438844,0.2534276843,0.2045451403,0.0285413619,0.373711139,0.1092816368,0.2637449503,-0.3957763612,-0.4265753925,-0.2401214242,-0.1653005183,0.4256432652,0.1469210535,0.1787017137,-0.2362689227,-0.303822875,0.283318162,-0.2621351182,0.151965633,0.0429622754,-0.1555477381,-0.2987037599,-0.2114507854,0.3432161212,0.1276800632,0.1706939042,0.0484018326,0.1696101129,0.2524091005,-0.1208019853,0.8238663077,-0.0635682568,-0.1314235777,-0.1152850837,-0.0395146459,-0.2656165659,-0.1850191057,0.0102546951,-0.1166962087,0.0241238102,0.381629765,-0.3781962991,0.0880152583,-0.0164086949,0.0268286597,-0.3905110061,0.0672976822,-0.1110441834,-0.470772624,-0.430236727,-0.0862661749,-0.1416791677,0.4008589685,0.1594100147,0.2181693912,-0.0276574921,0.2795979679,0.2141216397,0.2251775861,0.2607110441,0.1595428437,0.0613127463,-0.0761819705,-0.2943114936,0.0192233305,0.5156428218,0.127186656,0.2365971357,0.0179174673,-0.1256866157,0.3058919311,0.353326261,0.0412076525,0.224942565,-0.1922955662,0.0153246075,-0.2425510883,0.0591746122,0.1155349985,-0.2436773479,0.0901327953,-0.1724962741,0.1682245433,-0.2947555184,-0.3521316946,0.086889118,0.2420440316,-0.1013927683,0.3036596477,0.3035930097,0.9488306046,0.3353246152,0.1135626137,0.3608902097,-0.2075495124,0.0258863401,-0.2734702229,0.1519681364,-0.3740267158,-0.0625596866,0.0572385713,0.0583629422,-0.2141710371,0.1685255021,0.0524644889,0.4384443164,0.2437238246,0.1848827004,0.0702228323,0.1409515142,-0.0924054608,-0.0747848079,0.0029515878,0.0531415418,-0.1154729947,0.3401060104,-0.090874806,-0.1692330986,-0.2210327536,0.0344555862,-0.226444453,0.1442914605,-0.5238325596,0.3133117557,0.4568421543,-0.2757675648,-0.254119724,0.7030171752,0.4300964773,0.2325909287,-0.0726566464,0.2624039054,0.5008444786,0.2284643203,0.2381435335,0.2171832323,-0.0182109065,-0.0843087882,-0.294089973,0.3372737467,-0.2506558597,0.1089558601,-0.6670827866,-0.0575196557,-0.0427640863,-0.1426392049,-0.1850868911,-0.0336046107,0.1458840668,-0.2189784944,-0.0095642693,-0.2536067069,0.0922193527,0.2272176892,-0.2408324778,-0.1084252819,-0.0668555722,0.4478659034,0.1054668128,0.1571369916,0.6415705085,-0.2098605186,-0.0364837237,0.0930473581,0.064663671,-0.4503222704,-0.0296362881,0.1232792586,0.0214888696,-0.3334921896,-0.1061205491,0.4281898439,0.1537292004,0.0174186472,0.1043468714,-0.4397270679,-0.1445714235,0.4621887207,-0.117621772,0.0034605518,0.214208141,0.0082865031,-0.2010956556,0.2436837256,-0.2319589555,-0.1701847464,-0.1865469515,0.1614428759,0.0275459141,0.1224576905,0.2052406371,-0.0890582204,0.0160150956,-0.1305370331,0.0444641933,-0.1165537536,-0.4527635574,0.1614403278,0.11642351,0.2653885484,-0.0771372542,-0.3488806188,0.0651357621,-0.0611055791,0.124157913,0.3242108226,0.039947208,0.3424029052,0.0166315995,0.2856262624,-0.1570260823,-0.0011665333,0.0997801274,0.1727511734,0.2072710693,-0.0285778791,-0.212638855,0.0460995138,0.060050942,0.0381188355,-0.0444414094,-0.2118056118,0.3314485252,0.1655603051,-0.1196496785,-0.0943982601,-0.0308662429,-0.1426915824,-0.2555013895,0.1220716238,0.0399624482,0.0314756818,-0.3359080553,0.1770796478,-0.0440032519,-0.2765420675,-0.1121823117,0.2542101741,0.3468826413,-0.2869996428,0.3937596977,0.1090268418,-0.1169777736,-0.6292848587,-0.0057365252,0.2432087511,0.0002796413,-0.0183627009,0.2823640704,-0.1612988263,-0.1265545636,0.2200452685,0.100147061,0.0030182737,0.3094759583,-0.0766100287,0.4954618812,0.1501493007,0.1288388968,0.0092948619,-0.0002637517,0.1924198568,0.230856806,0.2963753641,0.1318616569,-0.3133875132,0.0684031695,0.3214664161,-0.1737070829,-0.2978072464,-0.1678691953,-0.1606396586,-0.1957285404,0.013651086,0.0772793069,-0.407787472,0.1681638807,0.182603091,-0.2792726457,-0.2892889678,-0.1758623868,0.2151628882,0.5069206953,0.077009581,0.3692753017,0.1024635658,0.1428624541,0.1320142746,0.2746454477,0.067481108,0.0623275861,-0.3842316866,0.1307674944,0.2049889565,-0.027688738,0.3567188382,0.1330915987,0.3253226876,0.0434273183,0.3325113654,0.0128083481,-0.1045854241,-0.1991184801,0.1404850334,-0.0553094745,-0.1745198518,0.2698664963,0.2881715298,-0.0493475832,0.4204522967,0.1491864622,-0.1316367239,-0.2629206479,-0.2125591636,0.2651269138,0.2261155099,0.0008914815,0.0369848646,-0.0291613173,-0.1942655891,0.3717502356,-0.1683085263,-0.2988691926,0.0201078765,-0.4621519744,-0.0688242763,0.2794607282,-0.2333631366,0.2594431937,-0.1044077501,0.0804863796,-0.0147828236,-0.2231643349,-0.6184481978,0.1799286157,0.1711007506,-0.1629858464,-0.0195142347,0.081234172,-0.1083704829,0.0270613618,-0.0725059211,0.0882162899,0.2514555752,-0.111946933,0.2420817614,-0.3079433441,0.2020721734,-0.2945342064,-0.1013778746,0.1801868975,0.0555455536,-0.1379777193,-0.0538951568,0.3241066933,-0.0261430182,-0.0984105989,-0.1921048462,-0.3853358328,0.3587110639,0.2550293803,-0.4032894373,0.0664922893,0.287165612,0.3047101796,-0.1547925323,-0.2566091716,0.4110637605,-0.090285711,0.0615222156,0.2689650357,0.267934978,-0.0887555107,0.5461474657,-0.1659614295,-0.3539319336,0.4041138887,-0.3229782283,-0.0972538367,0.0051640994,0.3167745173,-0.302054584,-0.0745038614,-0.4578879178,-0.2353713959,0.337751627,-0.0885355622,0.1365435869,0.0259771775,-0.3285058141,-0.0027856636,-0.0035060418,0.1100809574,0.1873667091,-0.0189367421,0.0626486987,-0.1144459099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"Looks like there are multiple issues regarding this (#2386, #2322) and it's a WIP #2329. Currently these datasets are being loaded in-memory which is causing this issue. Quoting @mariosasko here for a quick fix:\r\n\r\n> set `keep_in_memory` to `False` when loading a dataset (`sst = load_dataset(\"sst\", keep_in_memory=False)`) to prevent it from loading in-memory. Currently, in-memory datasets fail to find cached files due to this check (always False for them)\r\n\r\n","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":69,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nLooks like there are multiple issues regarding this (#2386, #2322) and it's a WIP #2329. Currently these datasets are being loaded in-memory which is causing this issue. Quoting @mariosasko here for a quick fix:\r\n\r\n> set `keep_in_memory` to `False` when loading a dataset (`sst = load_dataset(\"sst\", keep_in_memory=False)`) to prevent it from loading in-memory. Currently, in-memory datasets fail to find cached files due to this check (always False for them)\r\n\r\n","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"Hi ! Since `datasets` 1.6.0 we no longer keep small datasets (<250MB) on disk and load them in RAM instead by default. This makes data processing and iterating on data faster. However datasets in RAM currently have no way to reload previous results from the cache (since nothing is written on disk). We are working on making the caching work for datasets in RAM.\r\n\r\nUntil then, I'd recommend passing `keep_in_memory=False` to the calls to `load_dataset` like here:\r\n\r\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/223943872e8c9c3fc11db3c6e93da07f5177423f\/examples\/pytorch\/language-modeling\/run_clm.py#L233\r\n\r\nThis way you say explicitly that you want your dataset to stay on the disk, and it will be able to recover previously computed results from the cache.","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":106,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nHi ! Since `datasets` 1.6.0 we no longer keep small datasets (<250MB) on disk and load them in RAM instead by default. This makes data processing and iterating on data faster. However datasets in RAM currently have no way to reload previous results from the cache (since nothing is written on disk). We are working on making the caching work for datasets in RAM.\r\n\r\nUntil then, I'd recommend passing `keep_in_memory=False` to the calls to `load_dataset` like here:\r\n\r\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/223943872e8c9c3fc11db3c6e93da07f5177423f\/examples\/pytorch\/language-modeling\/run_clm.py#L233\r\n\r\nThis way you say explicitly that you want your dataset to stay on the disk, and it will be able to recover previously computed results from the cache.","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"gotcha! thanks Quentin","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":3,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\ngotcha! thanks Quentin","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"OK, It doesn't look like we can use the proposed workaround - see https:\/\/github.com\/huggingface\/transformers\/issues\/11801\r\n\r\nCould you please add an env var for us to be able to turn off this unwanted in our situation behavior? It is really problematic for dev work, when one needs to restart the training very often and needs a quick startup time. Manual editing of standard scripts is not a practical option when one uses examples.\r\n\r\nThis could also be a problem for tests, which will be slower because of lack of cache, albeit usually we use tiny datasets there. I think we want caching for tests.\r\n\r\nThank you.","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":104,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nOK, It doesn't look like we can use the proposed workaround - see https:\/\/github.com\/huggingface\/transformers\/issues\/11801\r\n\r\nCould you please add an env var for us to be able to turn off this unwanted in our situation behavior? It is really problematic for dev work, when one needs to restart the training very often and needs a quick startup time. Manual editing of standard scripts is not a practical option when one uses examples.\r\n\r\nThis could also be a problem for tests, which will be slower because of lack of cache, albeit usually we use tiny datasets there. I think we want caching for tests.\r\n\r\nThank you.","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"Hi @stas00, \r\n\r\nYou are right: an env variable is needed to turn off this behavior. I am adding it.\r\n\r\nFor the moment there is a config parameter to turn off this behavior: `datasets.config.MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES = None`\r\n\r\nYou can find this info in the docs:\r\n- in the docstring of the parameter `keep_in_memory` of the function [`load_datasets`](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/loading_methods.html#datasets.load_dataset):\r\n- in a Note in the docs about [Loading a Dataset](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#from-the-huggingface-hub)\r\n\r\n> The default in \ud83e\udd17Datasets is to memory-map the dataset on drive if its size is larger than datasets.config.MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES (default 250 MiB); otherwise, the dataset is copied in-memory. This behavior can be disabled by setting datasets.config.MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES = None, and in this case the dataset is not loaded in memory.","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":115,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nHi @stas00, \r\n\r\nYou are right: an env variable is needed to turn off this behavior. I am adding it.\r\n\r\nFor the moment there is a config parameter to turn off this behavior: `datasets.config.MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES = None`\r\n\r\nYou can find this info in the docs:\r\n- in the docstring of the parameter `keep_in_memory` of the function [`load_datasets`](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/loading_methods.html#datasets.load_dataset):\r\n- in a Note in the docs about [Loading a Dataset](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#from-the-huggingface-hub)\r\n\r\n> The default in \ud83e\udd17Datasets is to memory-map the dataset on drive if its size is larger than datasets.config.MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES (default 250 MiB); otherwise, the dataset is copied in-memory. This behavior can be disabled by setting datasets.config.MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES = None, and in this case the dataset is not loaded in memory.","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"Yes, but this still requires one to edit the standard example scripts, so if I'm doing that already I just as well can add `keep_in_memory=False`.\r\n\r\nMay be the low hanging fruit is to add `MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES` env var to match the config, and if the user sets it to 0, then it'll be the same as `keep_in_memory=False` or `datasets.config.MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=0`?","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":58,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nYes, but this still requires one to edit the standard example scripts, so if I'm doing that already I just as well can add `keep_in_memory=False`.\r\n\r\nMay be the low hanging fruit is to add `MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES` env var to match the config, and if the user sets it to 0, then it'll be the same as `keep_in_memory=False` or `datasets.config.MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=0`?","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"@stas00, however, for the moment, setting the value to `0` is equivalent to the opposite, i.e. `keep_in_memory=True`. This means the max size until which I load in memory is 0 bytes.\r\n\r\nTell me if this is logical\/convenient, or I should change it.","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":42,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\n@stas00, however, for the moment, setting the value to `0` is equivalent to the opposite, i.e. `keep_in_memory=True`. This means the max size until which I load in memory is 0 bytes.\r\n\r\nTell me if this is logical\/convenient, or I should change it.","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"In my PR, to turn off current default bahavior, you should set env variable to one of: `{\"\", \"OFF\", \"NO\", \"FALSE\"}`.\r\n\r\nFor example:\r\n```\r\nMAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=\r\n```","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":26,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nIn my PR, to turn off current default bahavior, you should set env variable to one of: `{\"\", \"OFF\", \"NO\", \"FALSE\"}`.\r\n\r\nFor example:\r\n```\r\nMAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=\r\n```","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"IMHO, this behaviour is not very intuitive, as 0 is a normal quantity of bytes. So `MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=0` to me reads as don't cache ever.\r\n\r\nAlso \"SIZE_IN_BYTES\" that can take one of `{\"\", \"OFF\", \"NO\", \"FALSE\"}` is also quite odd.\r\n\r\nI think supporting a very simple `MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES` that can accept any numerical value to match the name of the variable, requires minimal logic and is very straightforward. \r\n\r\nSo if you could adjust this logic - then `MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=0` is all that's needed to not do in-memory datasets.\r\n\r\nDoes it make sense?","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":89,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nIMHO, this behaviour is not very intuitive, as 0 is a normal quantity of bytes. So `MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=0` to me reads as don't cache ever.\r\n\r\nAlso \"SIZE_IN_BYTES\" that can take one of `{\"\", \"OFF\", \"NO\", \"FALSE\"}` is also quite odd.\r\n\r\nI think supporting a very simple `MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES` that can accept any numerical value to match the name of the variable, requires minimal logic and is very straightforward. \r\n\r\nSo if you could adjust this logic - then `MAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=0` is all that's needed to not do in-memory datasets.\r\n\r\nDoes it make sense?","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"I understand your point @stas00, as I am not very convinced with current implementation.\r\n\r\nMy concern is: which numerical value should then pass a user who wants `keep_in_memory=True` by default, independently of dataset size? Currently it is `0` for this case.","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":41,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nI understand your point @stas00, as I am not very convinced with current implementation.\r\n\r\nMy concern is: which numerical value should then pass a user who wants `keep_in_memory=True` by default, independently of dataset size? Currently it is `0` for this case.","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"That's a good question, and again the normal bytes can be used for that:\r\n```\r\nMAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=1e12 # (~2**40)\r\n```\r\nSince it's unlikely that anybody will have more than 1TB RAM.\r\n\r\nIt's also silly that it uses BYTES and not MBYTES - that level of refinement doesn't seem to be of a practical use in this context.\r\n\r\nNot sure when it was added and if there are back-compat issues here, but perhaps it could be renamed `MAX_IN_MEMORY_DATASET_SIZE` and support 1M, 1G, 1T, etc. \r\n\r\nBut scientific notation is quite intuitive too, as each 000 zeros is the next M, G, T multiplier. Minus the discrepancy of 1024 vs 1000, which adds up. And it is easy to write down `1e12`, as compared to `1099511627776` (2**40). (`1.1e12` is more exact).\r\n","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":127,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nThat's a good question, and again the normal bytes can be used for that:\r\n```\r\nMAX_IN_MEMORY_DATASET_SIZE_IN_BYTES=1e12 # (~2**40)\r\n```\r\nSince it's unlikely that anybody will have more than 1TB RAM.\r\n\r\nIt's also silly that it uses BYTES and not MBYTES - that level of refinement doesn't seem to be of a practical use in this context.\r\n\r\nNot sure when it was added and if there are back-compat issues here, but perhaps it could be renamed `MAX_IN_MEMORY_DATASET_SIZE` and support 1M, 1G, 1T, etc. \r\n\r\nBut scientific notation is quite intuitive too, as each 000 zeros is the next M, G, T multiplier. Minus the discrepancy of 1024 vs 1000, which adds up. And it is easy to write down `1e12`, as compared to `1099511627776` (2**40). (`1.1e12` is more exact).\r\n","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"Great! Thanks, @stas00.\r\n\r\nI am implementing your suggestion to turn off default value when set to `0`.\r\n\r\nFor the other suggestion (allowing different metric prefixes), I will discuss with @lhoestq to agree on its implementation.","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":35,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nGreat! Thanks, @stas00.\r\n\r\nI am implementing your suggestion to turn off default value when set to `0`.\r\n\r\nFor the other suggestion (allowing different metric prefixes), I will discuss with @lhoestq to agree on its implementation.","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2387","title":"datasets 1.6 ignores cache","comments":"Awesome! Thank you, @albertvillanova!!!\r\n\r\n","body":"Moving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n","comment_length":4,"text":"datasets 1.6 ignores cache\nMoving from https:\/\/github.com\/huggingface\/transformers\/issues\/11801#issuecomment-845546612 \r\n\r\nQuoting @VictorSanh:\r\n\r\n> \r\n> I downgraded datasets to `1.5.0` and printed `tokenized_datasets.cache_files` (L335):\r\n> \r\n> > `{'train': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-c6aefe81ca4e5152.arrow'}], 'validation': [{'filename': '\/home\/victor\/.cache\/huggingface\/datasets\/openwebtext10k\/plain_text\/1.0.0\/3a8df094c671b4cb63ed0b41f40fb3bd855e9ce2e3765e5df50abcdfb5ec144b\/cache-97cf4c813e6469c6.arrow'}]}`\r\n> \r\n> while the same command with the latest version of datasets (actually starting at `1.6.0`) gives:\r\n> > `{'train': [], 'validation': []}`\r\n> \r\n\r\nI also confirm that downgrading to `datasets==1.5.0` makes things fast again - i.e. cache is used.\r\n\r\nto reproduce:\r\n```\r\nUSE_TF=0 python examples\/pytorch\/language-modeling\/run_clm.py \\\r\n --model_name_or_path gpt2 \\\r\n --dataset_name \"stas\/openwebtext-10k\" \\\r\n --output_dir output_dir \\\r\n --overwrite_output_dir \\\r\n --do_train \\\r\n --do_eval \\\r\n --max_train_samples 1000 \\\r\n --max_eval_samples 200 \\\r\n --per_device_train_batch_size 4 \\\r\n --per_device_eval_batch_size 4 \\\r\n --num_train_epochs 1 \\\r\n --warmup_steps 8 \\\r\n --block_size 64 \\\r\n --fp16 \\\r\n --report_to none\r\n```\r\n\r\nthe first time the startup is slow and some 5 tqdm bars. It shouldn't do it on consequent runs. but with `datasets>1.5.0` it rebuilds on every run.\r\n\r\n@lhoestq \r\n\nAwesome! Thank you, @albertvillanova!!!\r\n\r\n","embeddings":[-0.2158487439,0.0203107707,0.0422426425,0.1404154301,0.1821744591,0.0462308712,0.1409803778,0.3139572144,0.0551589616,-0.0644533038,-0.2359380573,-0.0579473414,-0.0247005746,-0.3315773904,-0.1081899628,0.0198047999,0.2144653797,0.0230753701,-0.1678417474,-0.0409687683,-0.1443585306,0.2673944831,-0.128291443,-0.228637591,-0.6212518811,0.1536585689,-0.0539875105,-0.0578289442,-0.1436334997,-0.3942032158,0.3681198657,0.2826853693,0.1136951819,0.4599291086,-0.00012043,0.1595689058,0.2031688243,0.0211555008,-0.4104935527,-0.0573553517,0.0354547277,-0.0839594007,0.0214983989,0.1640308499,-0.2790547013,-0.0306378007,0.1911177784,-0.5746431947,0.452419132,0.3432961404,0.160297215,0.5254045725,-0.1487574875,-0.0378798321,-0.0608341657,-0.0301156137,-0.332855463,0.0926357806,0.1314037144,0.0842170715,-0.2770285308,0.3524879217,0.097985439,-0.0218208209,0.16133672,0.0107811904,-0.1585707217,0.0202849843,0.2987552285,0.10706193,0.3787642717,-0.2141118944,-0.5825324059,-0.3364526331,-0.3538715541,0.0064531881,0.3440105021,-0.0655077994,0.1379141808,0.3956792057,-0.4331495464,-0.0684470385,0.0699945837,-0.0044340761,0.0611454509,0.2066850215,-0.1162460893,-0.0994941965,0.1231028289,0.149908945,0.1937531978,-0.1635281295,-0.109124884,0.2992784679,-0.3559890985,-0.0108128795,0.1666994542,0.0396621898,0.0110875499,0.1188242361,0.0640139356,0.1364050955,0.0925285518,-0.2549798191,-0.0728779137,0.5833615065,0.1947101653,0.2663045526,0.0184179097,0.0475953147,-0.1567291319,-0.096233882,-0.0447993651,-0.2194101065,0.2228741646,-0.0724640489,0.1600194871,-0.177988708,-0.0083685135,0.1688308865,0.0522615016,-0.1071201414,0.1425199807,0.297921747,-0.1997744441,0.2688975036,-0.0717916638,-0.1660918742,-0.2827946544,0.0425893366,-0.3636681736,-0.0625563711,-0.2846522033,0.2555699944,0.2676339149,-0.372418493,0.4367183745,0.1324850917,0.2268454731,0.0182676762,0.0358226374,-0.1124993712,0.3517661691,0.1498368829,-0.4831211865,0.4104645848,0.0979266092,0.1543205231,-0.2801340222,-0.034504652,-0.1624662131,-0.2589809299,0.1788692027,0.0967067927,-0.4759327471,-0.2851259112,0.1238623857,0.2492318451,0.3530565798,-0.1826398224,0.0576875657,-0.054647509,-0.3108603656,-0.0424191542,0.2072167397,0.6110782027,-0.0987188816,-0.2701899707,0.3160687387,0.0783183947,0.1078270078,0.2469794154,-0.1151805148,-0.2041630894,-0.0944092944,-0.0451543592,0.2127512693,-0.2799159884,-0.6564025283,0.0895091146,0.0316258781,0.2786871493,0.0627193525,-0.030414138,0.1169167235,-0.2188137472,-0.169888556,0.1760187596,0.0873951912,0.1103237867,-0.4886095524,-0.3197338879,0.0049685156,0.0149404025,0.1607397348,0.1425743401,-0.2066916376,0.1983790696,0.0647579879,0.1064668968,0.2382353395,0.4176746905,-0.0577855743,-0.2508044243,0.2393510789,-0.2066840678,-0.4679914415,0.3813743293,-0.0582001358,0.1504364312,0.0513900258,-0.0919354334,-0.314740777,-0.1900457591,-0.3224794865,-0.4129860103,0.0291994661,0.083561644,0.4664493203,-0.0627890825,-0.2708913088,0.4161113501,0.1776850522,0.2026192099,-0.1979149729,-0.0367069468,-0.0714992806,-0.1292344034,-0.3720748723,-0.0548154041,0.1507507265,-0.2703870535,-0.1646059304,0.4357369542,-0.0549746081,0.1264852285,-0.0163788572,0.2787240446,0.1310151666,0.2094838619,0.1261975765,0.1861446947,-0.1718526632,-0.0207077246,-0.0400916971,0.3744085133,0.1988709271,0.212100476,0.1096512601,-0.0578966551,-0.0523809791,0.020036336,-0.0271951109,-0.3291665018,0.2592810392,-0.0970329493,0.6289586425,0.1915283948,0.0682512671,0.2737976611,0.7349142432,0.1994312406,-0.2323375195,0.1495774388,-0.2541774213,-0.3552419245,0.1600754559,0.280841589,0.3655433357,0.0967631564,0.4016495049,0.1217315421,-0.0592810474,-0.0925822258,0.1318967044,-0.103768073,0.0365564413,-0.0743083283,0.0565556027,0.2631292343,-0.3260219693,0.136027202,0.0117203426,0.1084349826,-0.1194188669,0.2409161031,-0.3367220759,0.2784646749,-0.1354539096,-0.1432013363,-0.3787076771,-0.2035416663,-0.048936978,0.1152366698,-0.0025185787,0.3248536587,-0.2567089796,0.1713041961,-0.0361298993,-0.3038983345,-0.4968311787,-0.1464488059,-0.2653330266,-0.0522758998,-0.0077604922,-0.3781483173,-0.041796647,-0.1704156846,-0.2427740842,-0.1978735626,-0.329761982,0.1543189883,-0.072553426,-0.0289234985,-0.0718471333,0.0568940639,-0.1304771602,0.0907302946,0.1732173264,-0.3506852686,-0.1693695635,-0.0369205885,0.0841066614,0.1431051046,-0.2130721211,-0.2104904354,0.0077060619,-0.2791153491,0.1315926015,-0.1691376567,-0.0363107882,0.5730288625,0.0180480685,0.0493544452,-0.154198274,0.1935802996,-0.2817358673,-0.5460473299,0.4480652511,-0.0068342546,-0.3691204488,-0.1439013928,0.0137629816,0.1789387465,0.3921634257,-0.1815630645,-0.2465928942,-0.3960894048,0.0011066424,0.2109925449,0.0658638254,0.6789283156,-0.012012572,-0.0564476885,-0.2156410962,-0.1154806092,0.1403704733,-0.1744911224,0.321870178,-0.2863911688,0.1133488566,0.261436671,0.6822909117,0.2117644399,-0.0542893298,0.2297441661,0.2226077169,0.4835837185,-0.3701800406,-0.2817359567,-0.0303613804,-0.2276449054,0.0423181877,0.1051360816,0.2854079008,0.1345894635,-0.1757140905,0.0838291422,0.0781061649,-0.2813549638,0.039714478,-0.3440360427,0.3322945535,0.4198822677,0.3526895046,-0.2939573824,-0.0616108328,0.0850312188,0.1677763462,0.241633445,0.161315158,-0.1308396757,0.6268690228,-0.7827771306,0.2176955342,-0.0210799146,0.2715777457,0.117168352,-0.2589393854,0.0737981275,-0.0793794692,0.4684254229,0.0345295779,-0.0915621519,0.0898290426,-0.2584157288,-0.3263619244,-0.067605108,0.0506654344,0.0655268729,0.5230583549,0.305990994,-0.261277616,-0.572140038,-0.0025242311,0.0265639573,0.0670713112,-0.0700507909,-0.3095417619,-0.010731726,-0.0510093756,-0.0454554223,0.2116691768,0.0512173213,0.0097731221,0.2060348988,-0.2357091606,-0.0749712214,0.1123635992,0.0876082331,0.2482483387,-0.1353745311,0.2141057402,0.3516706228,0.0993538499,-0.1307245195,0.3770356774,-0.1581927985,-0.5206752419,-0.0382089131,0.2025392801,-0.1003960818,0.3610553443,-0.1493229568,-0.0494242087,0.1187873781,0.0308994744,-0.0949648097,0.3457081616,0.2662645578,0.1549057513,-0.3207552135,-0.4492103755,0.4268127084,0.0861395448,-0.0303659104,0.3022578955,0.3320183754,-0.1887545437,0.138139382,-0.0381244868,0.718544066,0.1204460189,0.3983494639,0.1804788113,-0.1022650301,0.4165359735,0.1965270489,0.2679464221,-0.1517039686,-0.3694930077,0.0400474258,-0.0396894589,0.0779861584,0.1464396715,-0.0772888511,0.3025547564,0.165171504,0.2353433967,0.152645573,0.1048942953,-0.1742328852,-0.0978281572,-0.0854356661,0.1495160609,-0.0977488309,0.3520143628,-0.0446067974,-0.0288661867,0.0228391122,-0.3122123778,-0.1936957687,0.0801447183,-0.4066699147,0.1365056038,0.3283130527,-0.3234783411,0.1001867205,0.0218602195,0.4756090939,-0.1211263761,-0.1954516768,0.2320108861,-0.1895305365,0.1707837135,0.1086139381,-0.177397266,0.385437876,0.1122103482,0.0376974754,-0.3645128906,-0.250379473,-0.4268319011,0.1346522123,0.0056340098,-0.1293735504,-0.2019222379,0.0396191776,0.0577243529,-0.129512012,-0.2113541365,0.1154869497,0.0673520714,0.0494736768,0.0662550107,-0.1544148326,-0.2119868696,-0.1764064431,0.6264654398,-0.1077430397,-0.5085910559,0.4606703818,-0.0396037921,-0.2720078826,-0.0072721401,-0.2138532549,0.29101336,-0.507191956,0.266651541,-0.0812650472,-0.314702183,0.0867408812,0.1454060525,0.0332268812,-0.0632966235,-0.2446782738,0.0686218515,-0.511470139,0.0086054411,-0.200459972,0.2193791419,-0.0729593039,0.0024338742,-0.0114564318,0.1509165168,-0.2475846857,0.0313376114,0.065550223,-0.0200345833,0.0715788826,-0.4475181401,0.519418776,-0.2106151283,-0.0788263828,0.0893336907,-0.3692188859,-0.139516905,-0.2798086107,0.1558230817,0.0201896243,-0.0632973909,-0.1017561853,0.2242700607,-0.1317821145,-0.3850981593,0.2663278282,0.4267380536,-0.0859412178,0.2177353948,0.1543671489,0.2008510083,0.1570768356,-0.1120454744,-0.1153177992,0.0805329904,0.0798636377,0.232343927,0.0232997742,0.0737446398,-0.0667184517,0.0598417781,0.2692273557,-0.0168123096,0.2843044698,-0.2442565113,-0.1633828282,0.1765942276,0.2694580555,0.2560473084,0.2352306098,-0.3531941772,0.0514986143,0.1367585212,-0.0665064231,-0.0913145691,0.2504135668,0.1206495985,0.1605099887,0.0472538956,0.0817921236,0.1953168958,0.519335866,0.2348034829,0.5798423886,-0.2966858745,0.3942886889,0.145529002,0.3598196507,0.1383571327,0.6228572726,0.0592763014,0.0088254651,0.0637837723,-0.184096694,0.4451974928,-0.1497716755,0.2619179785,-0.1469971836,-0.2622156441,-0.1380573064,0.0083395969,-0.1496467739,0.2342556119,-0.1989998668,0.3088906407,-0.2092136741,-0.1399448514,-0.1402635276,0.0096942354,-0.1827800423,-0.3270062506,0.2731042504,0.0293133538,-0.0254419725,0.0349011049,-0.0406330414,0.0132086398,0.3504340351,-0.1268636882,0.0577689707,-0.2136361152,-0.1655310988,-0.086990118,0.3347528875,-0.0947559327,0.3914239109,-0.1091882661,-0.1141861081,-0.0627325848,0.5630706549,0.244389087,0.2078852952,0.033292599,0.1316380203,-0.1075236574,-0.0312001854,-0.102912955,0.3895367682,-0.0637682229,-0.1537944525,0.0288896244,0.0281649679,-0.0744896382,0.0483428687,-0.4329925478,0.013191212,-0.2983934879,0.5955242515,-0.1327138692,-0.2584170997,-0.0577879697,0.1808206439,-0.1852998137,0.1427333653,0.1971939206,-0.1472716779,0.0389881171,-0.2511451542,0.0460631549,0.2118889093,0.5874791145,0.4427193999,0.0371648632,-0.2049155086,-0.0314137004,-0.5466330051,0.0109709362,0.1660831124,0.4006892443,-0.0006393692,0.0278087873,-0.1901566386,0.0617383979,0.0248270892,0.1584170908,-0.0657321364,-0.12598373,-0.3469351828,0.2608814836,0.0132860746,0.0004390477,0.1867113113,-0.2607448399,0.0867998451,0.0810841992,-0.0324986205,-0.1374961734,0.0821656659,0.0162176862,-0.1607312411,0.2406508178,0.3036295176,-0.0884025693,0.1702005863,-0.1911836714,-0.6736296415,-0.0930993557,-0.189581275,-0.0692995265,-0.1484887898,0.3655905724,-0.3894742131,-0.4164470434,-0.2334029973,0.2534016967,0.1525858492,-0.028603889,-0.1299511045,-0.0457870029,-0.1260604113,0.2234048843,0.2347216904,0.4405471087,-0.0294994097,0.1446450502,-0.6241056919,-0.3679643273,0.4854192138,-0.3664082885,-0.2516687512,-0.1198854595,0.1506163031,0.3509034216,-0.4559537172,-0.7183164358,0.0642748773,0.1543937624,0.0468179733,-0.2576728165,0.2499382198,0.0383849852,0.1265490204,-0.1445231736,0.5700888634,0.2609195113,-0.1883661598,-0.0265075509,-0.1023060232]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2386","title":"Accessing Arrow dataset cache_files","comments":"Thanks @bhavitvyamalik for referencing the workaround. Setting `keep_in_memory=False` is working.","body":"## Describe the bug\r\nIn datasets 1.5.0 the following code snippet would have printed the cache_files:\r\n\r\n```\r\ntrain_data = load_dataset('conll2003', split='train', cache_dir='data')\r\nprint(train_data.cache_files[0]['filename'])\r\n\r\n```\r\n\r\nHowever, in the newest release (1.6.1), it prints an empty list.\r\n\r\nI also tried loading the dataset with `keep_in_memory=True` argument but still `cache_files` is empty.\r\n\r\nWas wondering if this is a bug or I need to pass additional arguments so I can access the cache_files.\r\n","comment_length":10,"text":"Accessing Arrow dataset cache_files\n## Describe the bug\r\nIn datasets 1.5.0 the following code snippet would have printed the cache_files:\r\n\r\n```\r\ntrain_data = load_dataset('conll2003', split='train', cache_dir='data')\r\nprint(train_data.cache_files[0]['filename'])\r\n\r\n```\r\n\r\nHowever, in the newest release (1.6.1), it prints an empty list.\r\n\r\nI also tried loading the dataset with `keep_in_memory=True` argument but still `cache_files` is empty.\r\n\r\nWas wondering if this is a bug or I need to pass additional arguments so I can access the cache_files.\r\n\nThanks @bhavitvyamalik for referencing the workaround. Setting `keep_in_memory=False` is working.","embeddings":[0.0086587435,0.1484433562,-0.0586204529,0.5825809836,0.014361524,0.2871320844,0.3393753767,0.3756525815,0.1552539617,0.1141940132,0.0963894278,0.3016970456,-0.2102267593,-0.5456681252,-0.0608220473,0.1094387621,0.1178415641,0.0517380498,-0.103199251,-0.1534640193,-0.1906240135,0.1310354471,-0.258679986,-0.0604599379,-0.3278146684,0.0888491496,-0.0627641007,0.1011972353,-0.0205198918,-0.5282450914,0.352294445,-0.0570568964,0.1428263038,0.1720666438,-0.0001263243,0.2992685139,0.5814350843,-0.0007952896,-0.4517445564,-0.1649363339,-0.4208176732,-0.1645340323,0.375937283,-0.0457995906,0.0850858614,-0.1399679333,0.1954983175,-0.8507196903,0.1330686063,0.3136349022,0.1056241393,0.0721357614,-0.2231870145,-0.0739162862,0.4686887562,0.0180107728,-0.1748925894,0.0241117571,0.0943951383,0.1462513655,-0.0681407228,0.4670679569,0.1520577967,-0.0541826151,0.1155000106,0.3008659482,-0.141857326,-0.2561483681,0.3042883277,0.1247970164,0.6636638045,-0.1021313816,-0.4949543476,-0.2492912561,-0.0989101604,-0.248441726,0.3422538936,0.1396725923,-0.1056610644,0.1982167959,-0.0847189277,-0.2601931691,-0.1995744705,0.1832585484,0.3425370455,-0.0334722176,0.0027912389,0.1365611851,0.1273950785,0.1450673789,0.6272753477,-0.2561600506,-0.4113564789,0.2873600423,-0.1532011777,0.3220556974,0.0637374893,0.0133709423,-0.0464088172,0.0888248086,0.1724103987,0.030413188,-0.127825141,-0.1089667454,0.223803699,0.268392235,0.2681394517,0.3390173018,-0.1513076127,-0.1213689297,-0.1738257259,-0.3415277302,-0.2074166834,-0.1350316554,0.4055297077,-0.1584483534,0.1130053699,-0.2951798439,0.1730503589,0.0794473663,-0.1831752658,-0.1228943989,0.086268045,0.3214913607,0.0829796419,0.4810509384,0.127933979,0.0835562646,-0.0200376678,-0.0498616286,-0.2547412515,-0.1034734845,-0.2702747285,0.0460588373,0.1482286155,-0.1758184582,0.1818381101,-0.2282372862,-0.0115227345,-0.0856870338,0.1918987781,-0.1257788092,0.567653954,0.1497793347,-0.1158944517,0.2721174657,0.1325927228,0.3631063402,0.0700255036,0.4299562573,-0.2127260417,-0.3171436191,-0.0112274494,0.0106315399,-0.079392679,0.1497696489,-0.098808229,0.1408194453,0.1590952128,-0.291775614,0.2035699636,0.0507252663,-0.1950235963,-0.127655372,-0.0362629555,0.4726661146,-0.3713498712,-0.0149878534,-0.1334593147,0.0049027996,0.0986405984,-0.3131472766,-0.2770813406,-0.2741310894,-0.4712885618,0.0479241572,0.7344582081,-0.4827947021,-0.8122273684,0.0097561916,0.1709824502,-0.0483296961,0.2135448754,0.2141939253,0.0522542074,0.2042657137,-0.0313903466,-0.1660847217,0.0546763912,-0.2511238158,-0.4041962028,-0.3138143122,-0.0736442208,-0.0876412764,0.2214115262,0.1611496508,0.0526479855,-0.2722453773,0.3443700969,0.1270004511,0.2154869288,0.3752092719,0.0404780209,-0.2582918108,0.1166333705,0.0322583131,-0.5155984759,0.4081654549,0.0385778472,-0.4469814599,-0.1077129766,-0.100883007,0.0144798504,0.1909885257,-0.1519772559,0.0752921626,-0.0534781218,0.3185630441,0.1834270805,-0.0739492327,-0.2905937731,0.3302761912,-0.0229285657,-0.0825063512,-0.176288113,0.1309015304,0.1947007179,-0.2225819081,-0.3791825175,0.002701323,0.0690419525,0.0239199474,-0.1951388419,0.3188824058,0.2432767302,-0.1265917718,0.0902723372,0.0307057686,0.0528054573,-0.0498722605,0.153416425,0.051751785,0.0865192264,0.0944843814,-0.2828694284,0.3017774522,-0.0532146432,0.3028884828,0.1095866933,0.1716531366,0.0292319208,0.2574460208,-0.0044608074,-0.3053285182,0.0639042258,-0.115476504,0.5305544138,0.0208692867,-0.2788176835,0.3812891543,0.4863036871,0.2384699434,-0.2873437405,-0.0009183827,-0.2804931402,-0.2516019344,0.3040845692,0.2880062759,0.3502087295,0.1908795536,0.2654625773,-0.2523513436,0.2879526317,-0.0186319854,0.1779497713,0.0685132518,0.1391948313,-0.1554175764,0.208966583,-0.1432071626,-0.0216792114,0.0106252255,0.0328710414,0.1285139769,-0.1750454009,0.1075765789,-0.1017452553,-0.0235283449,-0.0414828509,-0.1798665076,-0.0777829662,-0.4012187421,-0.1064021066,0.000874267,0.0145827876,0.1055185199,-0.6484495401,0.2991164923,0.2183873802,-0.0592365414,-0.3168590665,-0.0341891721,-0.3496877849,-0.1038995534,-0.0214460772,-0.3102265,-0.0601383187,0.1995315254,0.1586916447,-0.238146469,0.2124938071,0.1980479956,0.1234354675,-0.0044616438,-0.0050789202,-0.0258639194,-0.2013290077,0.194137156,-0.0277927183,-0.1111192033,-0.2147837728,0.0470559075,0.2491459399,0.090708673,-0.2080153823,-0.1322383583,-0.2748843133,-0.1493902206,0.0579008274,0.0161566548,0.0914268941,0.2227537632,0.1756011546,0.3468943238,-0.0360372886,-0.1640057564,-0.1476368159,-0.3944389224,0.3308269978,-0.2032197118,-0.2568877935,0.0785796717,-0.1508683115,0.2093843073,0.474488467,-0.4510912597,-0.0634741932,-0.0991527364,0.0934528038,-0.0725364536,-0.1138728857,0.4701070189,0.0903246924,-0.0099004358,-0.3014013767,0.0044999947,0.3002731204,-0.1212807,0.1628671139,0.0934897512,0.0207599476,-0.0193384103,0.8765259385,0.2412121743,0.2116504461,0.1132475883,-0.2280811369,0.5593954325,-0.0811173171,-0.1498134881,-0.1027145758,0.1136083379,-0.2922467291,-0.0803041458,0.2483395189,-0.088776283,0.0279261358,0.2586578727,-0.126980722,-0.1698693037,-0.0670122728,-0.6727398038,0.0833240002,0.1362859309,0.4376981556,-0.2855377197,-0.3053559065,-0.100876756,0.043339707,0.338565737,-0.1040980369,0.0741843134,0.2048360705,-0.3630968332,0.4238590598,0.0506163575,0.3643123806,0.1376971751,-0.398579061,0.075584054,0.094665058,0.6393511295,-0.1488383412,-0.151247561,0.3506399095,0.0348771848,-0.3291810453,-0.0740456954,-0.0323478803,0.3307231069,-0.147788763,0.0509788841,0.1001650542,-0.2690404654,-0.0626025796,0.0007723961,-0.138958171,0.1369044781,-0.1931449026,0.2139216512,0.1568505019,-0.0955152139,-0.0010446856,0.028802691,-0.0867957398,-0.0895199105,-0.4511216581,-0.0215952899,0.0054187067,0.0192460604,0.5926301479,-0.1284525543,0.1233923733,0.557498157,0.1454776973,0.3370680511,0.4932541549,-0.2455871701,-0.2683755457,0.1343157291,-0.3856577277,0.175740853,0.342512846,-0.4271121621,-0.0752919167,0.0485731326,-0.322630167,0.1478862911,0.0857872516,0.0822701082,-0.1502696425,-0.1762333661,-0.5304474235,0.3096626103,0.2138592899,-0.1700600088,0.94642061,-0.0908570364,-0.2770588696,0.1572758555,-0.1397411376,0.7183568478,-0.0598551519,0.3775262237,0.281811744,-0.2295470834,0.2828819156,0.233212769,0.2644119561,-0.2091979086,-0.1778427213,-0.0554671623,-0.1414114237,-0.1862270534,0.2838456333,0.0936555117,0.0308614522,-0.1609484404,0.0806705728,-0.1390294731,-0.1893261969,0.0542798601,0.0246673469,-0.0231302045,0.0350053422,0.0708455518,0.0480273068,-0.1158078685,-0.1556974798,-0.2204706669,-0.2171077877,-0.1917044669,0.2204909027,-0.3305668831,0.1196434125,0.3158498108,-0.5341467261,-0.2013469189,0.3490875065,-0.1075580791,-0.160080269,-0.2186230868,0.1364399195,0.1999458075,0.0417969339,-0.1611132771,-0.0733822435,0.2364817858,0.1214688122,-0.1398881674,0.1821077466,-0.4131307304,-0.4865494967,0.0554604642,0.2321104407,0.1520544589,-0.0267263371,-0.4183078706,-0.0219803303,-0.2869098186,0.0090408297,0.0201289356,0.2196043283,0.0359769352,-0.2081708759,0.0939713269,-0.2855896652,0.1591140777,0.4515765607,-0.0378478356,-0.2862846553,0.528652072,-0.0602921583,-0.0262560323,-0.0573223531,-0.1543737203,0.426196754,-0.4404309094,0.304777503,-0.149119705,0.2291273028,0.1148793399,0.0605639741,0.0354314372,0.1940597296,-0.2255056798,-0.185987711,-0.3933669925,0.3161686361,-0.3777481616,0.3321886957,-0.295440495,-0.6421486139,-0.0450790189,-0.3069440722,-0.1461935937,0.0965793878,0.0467512645,-0.107100144,-0.0371385515,0.165316388,0.4207697511,-0.2527632713,-0.0608038157,-0.1065040305,-0.1657715589,-0.0544176549,-0.014858623,0.1862252802,-0.0287833773,-0.0863691419,-0.1617950052,-0.3100075424,-0.015639754,-0.1361569613,-0.0003640213,0.1476743072,-0.1924207658,0.2413215339,0.1468044221,0.1865966767,-0.2037779838,0.2682129443,-0.0390308797,0.3286945224,0.070951201,0.1832629293,0.1746859103,0.0866331309,0.0595187247,-0.0029407996,-0.2138644606,0.0524809211,0.1314964294,-0.3910873532,0.0661246032,0.2843994796,0.1779811531,0.2959285975,-0.0613031872,0.0834191069,-0.1142036691,0.098962456,-0.0474899262,-0.0391699634,0.0634363741,0.3626732826,-0.065959096,0.1230736747,0.0987326279,-0.322113812,0.1874577701,0.1662789434,0.1525571644,-0.3932755888,0.2196412832,0.4273037016,0.3567899168,0.1886700243,0.2715615928,0.2820469737,0.0874503106,0.4459552765,-0.0289268773,0.3359936476,0.5023398399,0.1340837628,-0.3409581482,-0.5579535365,-0.1491236091,-0.0043444163,-0.1635191888,0.199373886,-0.1183273271,0.3072599769,0.1716505438,-0.11757157,0.0251214262,-0.2461457998,-0.0717066228,-0.2310865223,0.1974198818,-0.2202671915,-0.0801818892,0.0478912368,0.0485899523,0.0357942097,0.3153189719,0.0132622663,0.1874216199,-0.1415961534,0.2562049329,0.5658798814,0.1668455303,-0.1749982238,0.2494397163,-0.1950921863,0.0750241205,0.3231140971,0.1571608037,0.2034118176,0.1972947717,-0.0882718787,0.0020066996,0.1002648845,-0.1781275421,-0.1882523745,0.3402649462,0.2912041843,-0.0430682637,-0.0316013098,0.024401322,0.0581114702,0.5427252054,-0.1515842676,0.3370585442,-0.386810571,0.0773757249,-0.1112306863,-0.3298600018,-0.1760322154,-0.0507654548,-0.173246935,-0.1092505902,0.5164635777,0.1002367884,0.0176755656,-0.0179830492,-0.0209620278,-0.0750138164,0.1869273782,0.4633506238,0.3729032576,-0.1557870805,-0.1013801694,-0.2454878241,0.1531476229,-0.0933990851,0.2288911045,0.0448203422,0.3053553402,-0.1142381951,-0.224408105,0.4505221248,-0.0161233079,0.0908384323,0.1279372871,-0.5454428792,-0.24189879,0.2605953515,0.0059336303,0.0411922745,-0.4107004106,-0.1584772915,-0.0038958394,-0.1569649726,0.1845959872,-0.397035569,0.0546930134,0.3001666069,0.4898158908,-0.0127002159,-0.1039635763,0.2541372776,-0.2060459703,-0.7876588106,-0.2488608658,0.0990297422,0.0087975888,-0.1319762766,0.3823498487,-0.5604308248,-0.0382443741,-0.210777238,-0.09530586,0.1714805961,0.1005784273,0.0126295937,0.1303395331,-0.3942715824,0.2801752687,0.3067973256,0.1814804375,-0.0501066819,-0.0532519221,-0.2928853035,0.1332689226,0.2281432599,-0.4740186334,-0.2031558007,0.0175216664,0.1480816752,0.2902068794,-0.2829039693,-0.7554504871,-0.1915969849,0.4218648374,-0.2905318737,-0.0580878109,-0.2483447194,0.1447306722,0.3765780628,-0.0686614588,0.5989040136,0.0672025755,0.0557456873,-0.1233315542,-0.3069169819]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2377","title":"ArrowDataset.save_to_disk produces files that cannot be read using pyarrow.feather","comments":"Hi ! This is because we are actually using the arrow streaming format. We plan to switch to the arrow IPC format.\r\nMore info at #1933 ","body":"## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom pyarrow import feather\r\n\r\ndataset = load_dataset('imdb', split='train')\r\ndataset.save_to_disk('dataset_dir')\r\ntable = feather.read_table('dataset_dir\/dataset.arrow')\r\n```\r\n\r\n## Expected results\r\nI expect that the saved dataset can be read by the official Apache Arrow methods.\r\n\r\n## Actual results\r\n```\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/pyarrow\/feather.py\", line 236, in read_table\r\n reader.open(source, use_memory_map=memory_map)\r\n File \"pyarrow\/feather.pxi\", line 67, in pyarrow.lib.FeatherReader.open\r\n File \"pyarrow\/error.pxi\", line 123, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Not a Feather V1 or Arrow IPC file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets-1.6.2\r\n- Platform: Linux\r\n- Python version: 3.7\r\n- PyArrow version: 0.17.1, also 2.0.0\r\n","comment_length":26,"text":"ArrowDataset.save_to_disk produces files that cannot be read using pyarrow.feather\n## Describe the bug\r\nA clear and concise description of what the bug is.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\nfrom pyarrow import feather\r\n\r\ndataset = load_dataset('imdb', split='train')\r\ndataset.save_to_disk('dataset_dir')\r\ntable = feather.read_table('dataset_dir\/dataset.arrow')\r\n```\r\n\r\n## Expected results\r\nI expect that the saved dataset can be read by the official Apache Arrow methods.\r\n\r\n## Actual results\r\n```\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/pyarrow\/feather.py\", line 236, in read_table\r\n reader.open(source, use_memory_map=memory_map)\r\n File \"pyarrow\/feather.pxi\", line 67, in pyarrow.lib.FeatherReader.open\r\n File \"pyarrow\/error.pxi\", line 123, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Not a Feather V1 or Arrow IPC file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: datasets-1.6.2\r\n- Platform: Linux\r\n- Python version: 3.7\r\n- PyArrow version: 0.17.1, also 2.0.0\r\n\nHi ! This is because we are actually using the arrow streaming format. We plan to switch to the arrow IPC format.\r\nMore info at #1933 ","embeddings":[-0.2040329725,0.1602186114,-0.0240918901,0.3715007603,0.1489743888,0.2544102669,-0.0069093485,0.5975209475,-0.516123116,-0.1188536435,-0.3667038381,0.6461250186,-0.0722652897,-0.7478836775,0.0741805732,0.2550881207,0.1921531856,0.1028440222,-0.082889922,-0.0657771602,-0.0897659138,0.2092618346,-0.0571896695,0.2752649486,-0.3663752675,-0.0884740204,0.0842268914,0.2411667407,0.0667258576,-0.3410418332,0.3246599734,-0.1335393041,0.0821229294,0.1735127866,-0.0001142378,0.1800326556,0.3546250463,-0.114441365,-0.2745331228,0.0962590724,0.3013077676,0.0216187648,0.2277007848,-0.3599592745,0.2159604132,-0.4508233666,0.0920102224,-0.1863130331,0.4654636681,0.5436754823,0.1880560517,0.1298346967,0.247970745,0.1184635311,0.6443868876,0.0946033746,-0.1738111973,0.6543809175,-0.1060598642,-0.1333584189,-0.2428010702,-0.0375029743,-0.2614388764,-0.0618286058,-0.0463792607,0.1491945684,-0.1361512244,-0.2698401511,-0.0836420432,-0.0086214123,0.2848907113,-0.5713864565,-0.3896680772,0.0149825756,0.1416034847,-0.1129797548,0.203139618,0.2324097902,-0.2008633614,0.2662531137,0.3600381315,0.105133526,-0.3550495505,0.1775662303,0.1963576525,-0.3350893855,-0.1383097768,0.0752857625,0.1055062041,0.2164954394,0.4860531986,0.0233407393,-0.1866559684,-0.0319987424,-0.0597093664,0.1197859868,-0.0345686451,0.0261297375,0.1266890764,0.2536793947,0.3568298519,0.0311271641,-0.0248768926,0.2956557572,-0.2425471246,-0.2075383067,-0.0120289112,-0.2447337359,0.4849910438,-0.0072272983,-0.1094048023,-0.129802078,0.0873428732,-0.1138179153,0.3794026375,-0.2756100297,0.3993114531,-0.1765921563,-0.43215096,0.0621131286,-0.1248621419,0.1126479954,-0.0811195746,0.2452442944,0.0682345256,-0.1348800212,0.1660973281,0.4020653069,-0.1224156693,-0.3004590869,-0.0893309787,-0.0793376938,0.1035904214,0.0683884621,0.1880318969,-0.2568871081,0.2237656564,0.1502276063,0.0028199984,0.0867314413,-0.0153814107,-0.3452115655,0.0290859472,0.1992779672,0.1487944126,0.206622541,0.0915319398,-0.2093018442,-0.1230401173,0.1477116942,-0.2002523094,-0.0906479284,-0.2770582438,0.1281217784,0.004672695,-0.0147553897,-0.6079197526,-0.0893438086,0.079785265,-0.368920207,0.0663040206,-0.2139708102,0.2391676009,-0.5171179771,0.196696654,0.198701188,-0.4703057408,0.3383235037,-0.187883988,-0.1323550642,0.1699742377,0.236460939,-0.1800809205,-0.0397697911,-0.3184486628,0.3659442067,0.628264606,-0.2115264535,-0.5290198326,0.0265543051,0.0094446475,-0.0924774632,0.1212489232,0.1006608605,0.192357555,-0.0783827305,0.0730787665,0.1113942042,0.1752781272,-0.0964541286,-0.4493743479,0.0013989729,-0.1251825392,0.0139608746,-0.3602592349,-0.2302970439,0.4638059437,-0.3822005391,0.3907346129,0.0159339271,0.2205717117,0.1957494169,0.319806993,-0.051801037,-0.0862396061,-0.1210420877,-0.3550294936,0.0492109507,0.0426891483,-0.2842832506,-0.3007982969,-0.3553681672,-0.0504073538,0.0846613348,-0.1455820948,0.0046237539,0.1144097373,-0.0577399954,0.1321231425,0.426502049,-0.1304387003,-0.0774213746,-0.1839461625,0.1863158643,0.0773608461,0.3832542002,-0.1710474342,-0.4252557755,0.0546784848,-0.0013697831,-0.0663407519,-0.0181767717,-0.1635926217,0.4444546998,-0.104848206,-0.0027795583,0.1065694913,0.0027234366,0.0258759726,-0.4440297186,0.0702937171,0.2675184608,0.2114117295,0.114753738,-0.1618662775,0.4025797546,0.2345505208,-0.0526555367,-0.0595331341,-0.076381743,0.2920064926,0.1740585566,0.0049437806,0.0926574841,-0.0624662712,0.0698854998,0.4033080935,-0.2243360728,-0.2936953604,0.1701296717,0.4582487941,0.1070467383,0.2035182863,0.1900099069,-0.458039999,-0.0727926493,0.2644941509,0.0998542607,0.3613568246,0.2305187732,0.1219336167,-0.1843561232,-0.3392295241,-0.2428066432,0.1225639209,0.103658475,0.3005486727,0.2048089504,0.3009166718,-0.0272119697,-0.5037742853,-0.1916365772,-0.0973271206,0.02382995,-0.2101465315,0.2568069994,-0.2381262481,-0.3353670239,-0.2172453701,-0.124833174,-0.1586039513,-0.3103268743,-0.0365534388,0.5780177116,0.0232730042,-0.11160402,-0.1507937908,-0.0620743558,0.2144187242,-0.1583542824,-0.3896921873,0.0800497457,-0.5274578929,0.0443884544,0.392510891,-0.0837984458,0.3802197576,0.1537794024,0.1395398676,-0.4493108988,0.0021023147,-0.1122934893,-0.2809176743,0.4190345705,-0.3382554054,0.1093484536,-0.2637506425,-0.0891233534,0.1717892587,-0.0154298469,-0.1937982142,0.1257688403,0.0371739492,-0.007128987,-0.0095254062,-0.1934448183,-0.1541962028,-0.3427513242,0.279450804,-0.0502474569,0.0649999604,-0.0202814043,0.254016012,0.4780289233,0.0682370141,0.0917002708,0.0518788509,-0.1818199456,0.7161714435,-0.1722089499,-0.3640959263,0.1633569449,-0.0550005399,0.1100773588,0.0610596612,-0.5219791532,-0.2839545906,-0.006811006,0.6309642196,-0.066388838,-0.0105872657,0.1917037666,0.0353512838,-0.0799474493,-0.0474486426,0.0840427801,-0.1530009359,0.0323280655,0.0995260775,-0.1042967141,0.3168242574,-0.0642553568,0.6338437796,0.146045655,0.0935227126,0.5382612944,-0.3165524304,0.3479831517,-0.1960766315,-0.0964326486,-0.3640474081,-0.3192086816,0.0621135011,0.0054774084,0.0595304929,-0.1623480767,0.1042286158,0.018826386,-0.3889211416,-0.331084609,0.1735133231,-0.1545094997,-0.101887174,-0.0340695642,0.1319405884,0.3461717665,-0.2330643982,0.149590686,0.3263342977,0.1793850362,0.1115914658,0.1782157719,-0.1987205148,-0.3324721754,0.2440050393,-0.0607047528,0.1120349392,0.0934569016,-0.1182398871,0.2939030528,-0.116898261,0.5011329651,0.1405365616,-0.0242126565,0.2702575326,0.267038554,-0.6215153337,-0.0747949928,-0.0314125307,0.0143584982,-0.060167212,0.2260790169,-0.0339015238,0.0088196509,-0.1876796931,0.2643366754,-0.0638044178,0.1802326292,-0.2145405561,-0.2797142565,-0.281208992,-0.0233741421,0.151554659,-0.0318394043,-0.2828219533,-0.2102072686,-0.2041368634,-0.2065012306,0.0623885393,0.121378839,0.3808788955,-0.2450421005,0.1007575542,-0.0931303725,-0.0102142319,0.5208442211,0.5118024349,0.2257418633,-0.4611839652,-0.1422623694,-0.1319176853,0.0384697169,0.1068285257,-0.2773246765,-0.0265366025,-0.2617263198,-0.1741290241,0.064876914,0.0540832654,0.2572917044,-0.3083182275,-0.070675835,-0.0182440802,0.731711328,0.1117015183,0.0836656541,0.2863133252,0.3211808503,-0.1072591692,0.4641636908,0.2122834772,0.7479281425,0.043137949,0.4643588662,0.5308567882,-0.2122963518,0.0160339009,-0.1649824828,0.1785253137,-0.3695942461,-0.0326738842,-0.0299425181,0.0134265795,-0.1000159234,0.0142353373,0.0010392972,-0.0607739463,0.0199613739,0.001128585,-0.0676589534,-0.1116179898,-0.218418479,-0.0588077679,-0.1353928149,0.1330664754,0.0535563044,-0.0894671679,0.1921913922,-0.3127528727,-0.5645931959,0.0088156508,-0.1491820514,0.157237798,-0.1951975375,0.4169195294,-0.1762772202,-0.5911453962,-0.1303219944,0.2761573493,-0.2100264877,-0.2087070644,-0.1128743961,-0.1383062303,0.325248301,-0.1204835549,-0.141692698,-0.2638709545,0.3116105497,-0.0208358988,0.090375796,0.1997550875,-0.0059994315,-0.222235769,0.2424612194,0.1891068369,0.2453873307,-0.102924034,-0.10395246,-0.3850191832,-0.3170557022,-0.2708206177,0.1112982482,0.060735561,0.1559600979,-0.0029874155,-0.0495336279,-0.077205345,0.0608417988,0.525767982,-0.1084994599,0.1239104271,0.348838836,-0.2154362053,-0.2444496453,-0.1800950617,0.3286757171,0.0323287249,-0.7142599821,0.3077320158,-0.0378189087,0.3314643502,-0.209483996,-0.1696855277,0.1757101715,-0.2275860161,-0.2607493699,-0.3319696784,-0.3052562177,0.2826490402,0.1260707974,0.0277666375,0.1368633062,-0.2062249184,0.1197325811,-0.2621000111,-0.2328123748,0.1825926751,0.2293160856,-0.0286845583,-0.0013062964,0.1293844283,0.2112018317,0.0492569506,0.1066417918,0.2611012161,0.0131677166,-0.2077660561,-0.0308493413,0.1410416663,0.0917202234,-0.0365271494,-0.0383390673,-0.2797231078,-0.0027393363,-0.1004419923,-0.147117272,-0.1317083389,-0.0407161377,-0.0580725297,0.4979653358,0.2660637498,-0.005335527,0.0013513896,-0.1619920582,0.1802113354,-0.1033188626,0.0690186769,-0.1313317716,-0.0996306688,-0.3911681175,-0.0428951681,-0.1632254273,0.0117556304,0.24430646,-0.2647142112,-0.1916566938,0.038172327,0.2037824243,0.4649304152,-0.1592673957,-0.1972120553,0.309466362,0.1419485658,-0.0694742948,-0.0542083681,-0.0278011877,-0.1513531655,-0.1106505617,-0.0502440818,-0.0706550404,-0.0053181467,-0.1682131439,-0.064365536,0.0870016143,-0.1888093054,-0.0426288471,0.2759276628,0.142425403,0.1502589285,0.2690136731,-0.0499714315,0.342037946,0.5846748948,-0.0860492289,0.210579589,0.2341168672,-0.199513644,-0.119570151,-0.2119227499,-0.2484053671,-0.3364697993,0.0029922517,0.213861227,0.2108577192,0.650398612,0.106139496,-0.1071249694,0.0391622297,0.1595211029,-0.098877117,0.1020458341,0.0903363004,-0.0435119048,0.1965125203,0.0737741143,-0.3059620261,-0.2051436454,0.1116173491,0.3435764611,-0.1187390611,-0.0592509918,0.1965910643,0.1987667084,0.2077500075,-0.0424804799,0.0357110426,0.2661932111,0.0226387288,-0.2092641592,-0.0500584841,0.4795774817,0.0533769242,-0.0390313305,0.1192437112,-0.0269431043,-0.0728559494,-0.0308473874,0.0892176405,0.2304474711,0.2329432368,0.297541678,0.1788152605,-0.1063542888,-0.0122928899,-0.1355466843,0.1412945986,0.0330212638,0.2003348023,-0.2876665294,-0.3980258703,-0.2061076909,-0.2261141539,-0.0600929074,0.1351624429,0.3732072413,-0.0087251971,-0.0343861617,-0.0231072363,0.0742298588,-0.084338218,0.2900454402,0.4074524343,0.0998025686,-0.1530338824,-0.0733737126,-0.4634938836,0.1758205742,-0.1239106953,0.1178748757,-0.0990302786,0.2496943325,0.0420528837,0.0901108012,0.1534992009,-0.2341082841,0.2216779888,0.5094465017,-0.2163200974,-0.0703231469,-0.0387993604,-0.0673303902,-0.0695792586,-0.4108839333,0.2714143395,-0.1225953847,0.1092187017,-0.0908555239,-0.1614263803,0.1238614768,0.2390544862,0.6692219377,-0.1242606416,0.2127544731,-0.0750222206,-0.0866603479,-0.300047189,0.0526146851,-0.0731733814,0.1781781167,0.0002240012,0.5266051888,-0.0404526964,-0.3038596213,-0.1741482764,0.307214886,-0.0574831553,-0.0235131104,-0.0527946055,0.0384699665,-0.1573580056,0.1157290339,-0.057801038,0.027186662,0.0116744153,0.2189774066,-0.4629120529,-0.6108031273,0.412887305,-0.3112291098,-0.0361100473,0.2989273369,0.4274262786,0.2502659559,0.0931094438,-0.3529934883,-0.1512593031,0.4758540094,-0.2447625846,-0.2351056486,0.2724817097,0.0116761932,0.1664245278,-0.0315091684,0.5087145567,0.0213007573,-0.2097076625,0.0494028442,-0.4608906209]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2373","title":"Loading dataset from local path","comments":"Version below works, checked again in the docs, and data_files should be a path.\r\n```\r\nds = datasets.load_dataset('my_script.py', \r\n data_files='\/data\/dir\/corpus.txt', \r\n cache_dir='.')\r\n```","body":"I'm trying to load a local dataset with the code below\r\n\r\n```\r\nds = datasets.load_dataset('my_script.py', \r\n data_files='corpus.txt', \r\n data_dir='\/data\/dir', \r\n cache_dir='.')\r\n```\r\nBut internally a BuilderConfig is created, which tries to use getmtime on the data_files string, without using data_dir. Is this a bug or am I not using the load_dataset correctly?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/bc61954083f74e6460688202e9f77dde2475319c\/src\/datasets\/builder.py#L153","comment_length":21,"text":"Loading dataset from local path\nI'm trying to load a local dataset with the code below\r\n\r\n```\r\nds = datasets.load_dataset('my_script.py', \r\n data_files='corpus.txt', \r\n data_dir='\/data\/dir', \r\n cache_dir='.')\r\n```\r\nBut internally a BuilderConfig is created, which tries to use getmtime on the data_files string, without using data_dir. Is this a bug or am I not using the load_dataset correctly?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/bc61954083f74e6460688202e9f77dde2475319c\/src\/datasets\/builder.py#L153\nVersion below works, checked again in the docs, and data_files should be a path.\r\n```\r\nds = datasets.load_dataset('my_script.py', \r\n data_files='\/data\/dir\/corpus.txt', \r\n cache_dir='.')\r\n```","embeddings":[-0.3634896576,0.2240853161,0.053444095,0.4514975846,0.1415300369,-0.1207066849,0.498796463,0.0649361387,0.1139449701,0.1279033422,0.1670618504,0.1164789125,-0.000026259,0.1618969291,0.2033519,0.2234757394,-0.000567297,0.0768432692,-0.1561119705,-0.0686451644,-0.2324312925,0.1196476519,-0.0156468209,0.1871809959,-0.3966724575,0.3161482811,0.0041334904,0.6595543027,0.020150587,-0.1594536304,0.4635256231,-0.0194137301,0.2626406848,0.6896225214,-0.0001233974,0.3195054829,0.4197114408,-0.1370545328,-0.4451933503,-0.3630022407,-0.3213688731,-0.1205844432,0.2730474472,-0.11777132,-0.3139307797,0.1684602052,0.1296423227,-0.4796307087,0.1251054704,0.3512626588,0.0864334106,0.1955178678,-0.4008831084,-0.0074883299,0.0875306055,0.4546239674,0.0039398121,0.2868144512,0.2027198225,-0.0509557985,0.0403102152,-0.1094479337,-0.114551954,0.1670668274,0.7290754914,0.2044696212,0.246295616,0.1195372492,0.1271799505,0.2097671032,0.8114512563,-0.2189270705,-0.3567315042,-0.3475878537,-0.295979619,-0.1321891546,0.2913793921,0.1959009916,-0.328309536,0.1550272554,-0.2826429009,-0.1759109199,-0.0385543704,0.3217872381,0.0216448698,0.2445595264,-0.1323623061,0.2320985943,-0.1640591025,0.0989941284,0.3219019771,-0.3972786367,-0.0250569135,0.2357768565,0.1988216937,0.3457135856,0.0492765792,0.4159120023,-0.040215753,0.0188087318,-0.048678197,0.0319599025,-0.0594865642,0.1328590661,0.4287555814,0.0450413488,0.1661290675,0.145797208,0.1568096727,0.0159146264,-0.1326692402,-0.1107086539,-0.0979724526,-0.4574611783,-0.1004288048,-0.1933715194,0.2295648158,-0.1493961513,0.0464507863,-0.2724032402,0.252602905,-0.0867495164,0.3000946343,0.3881671131,-0.1646872759,-0.1338105947,-0.0842716098,0.2251116484,-0.1790882349,0.0483147316,-0.358327806,-0.2650708258,-0.2043486834,0.027656192,-0.0473335683,-0.2438429892,0.283828944,-0.0513345227,0.1549018025,-0.1537651718,-0.1994173676,0.136416167,0.1271741986,0.3635897934,-0.2581698596,0.1701279432,0.2217472941,-0.3667539656,-0.2471632659,0.2983249128,-0.3065575659,-0.542529583,0.0149518829,0.0559000187,-0.212438792,0.0018625653,-0.0638151765,-0.1449575126,0.4413236678,0.011180127,-0.1396732777,-0.0297437217,-0.5766786337,-0.0687704608,0.2972009778,0.6269584894,-0.4744592309,-0.229979068,-0.0354942158,-0.1582820714,0.1261746436,0.0954372138,-0.5179555416,0.5660901666,-0.2751978636,-0.2663215995,0.5033833981,-0.449216783,-0.109479934,0.5727268457,-0.246774137,-0.1474408507,0.0393458232,0.1660744101,-0.1855107844,-0.1198162809,0.2940329611,0.4177986681,-0.0429040566,0.1091134325,-0.0920208097,-0.1166869253,0.1799337864,0.1157259643,-0.0208533984,0.2262751162,0.1564219147,-0.1143688932,0.160383895,-0.0202934071,0.093998827,0.2106742412,-0.2537521124,0.0658293813,0.1278431267,0.1546729654,-0.528696537,0.3985966146,-0.0480761528,-0.1569336355,0.0695543811,-0.0701723471,-0.1584976017,0.17531991,-0.4332988858,0.0715495795,-0.1169660091,0.2941356897,-0.0104163317,-0.1862038374,-0.262555927,0.0907969996,-0.0827613473,0.1009166092,-0.2276257128,-0.0032427341,0.0723079368,-0.013983231,-0.2208838165,0.1398984194,0.0464977883,-0.1580699682,-0.1367745697,0.3328722417,0.1175295264,0.4391644299,-0.1691031307,0.0723802596,0.0857858881,0.2635891438,0.1544679999,0.0564928278,0.1003271565,-0.1937914342,-0.129720971,0.2046453506,-0.1600632668,0.3911741376,-0.0227396917,-0.0961444601,0.177777648,0.0437285453,-0.1670854241,-0.0535982102,-0.0064443788,0.0601862855,0.4769275486,0.2809217572,-0.1788144112,0.0292527415,0.0615935959,-0.0224738009,-0.0784141272,0.0449133664,-0.1692564338,-0.1153314561,-0.0713187829,0.0733580515,0.3055044413,0.0455869548,0.1843348593,0.0199747384,0.071401298,-0.3972021639,0.1588449776,0.0964899808,0.0015545648,0.1022952124,-0.2346717864,-0.2873325646,-0.1957287788,0.2077186555,0.0872740969,-0.0018091075,-0.5557534695,0.2986184061,-0.3373382688,-0.2322006524,-0.3124395609,-0.5424563885,-0.1791621,-0.1586726308,-0.1613505185,0.1112186983,-0.0387292504,-0.0187358502,-0.0045743803,0.1716792434,-0.0236294903,-0.6744971871,-0.1274794489,-0.2660986781,-0.3435392082,-0.1125141382,0.4159903526,0.0857210457,-0.0353805013,-0.3524414897,-0.0573636666,-0.1474471837,0.2198313922,0.3596020639,0.0932917297,0.753444612,0.1883265227,0.0202108771,0.2376392633,0.0874288008,0.2951365113,0.1826603562,0.018356679,-0.0442107543,0.0366412774,-0.0859246105,-0.3417372704,-0.0414708368,-0.2050759196,-0.1807593256,0.1391971409,0.2741288543,0.1545012593,0.6428108811,0.1556607336,0.0804975107,-0.1419402212,0.1903131604,-0.122363098,-0.8258020878,0.0887810662,-0.145987466,-0.2825871706,-0.1199327782,-0.036819838,0.175004527,-0.1200867295,-0.3841398954,0.016501788,0.0389106274,0.1934177876,0.3129930496,-0.0268756375,0.1477226615,-0.1665030718,0.1391680688,-0.1151883155,-0.2201762944,0.304744184,-0.0463766195,0.0681138188,0.3982470036,0.2701755762,-0.349062413,0.6581444144,-0.0419468917,-0.135484308,0.2687523961,-0.3474255204,0.5696567297,-0.3677194118,-0.1921464354,0.1278744489,0.2734452486,-0.5633006692,0.3523603678,0.3173483908,0.0421457104,-0.2780067921,-0.1425402164,-0.0093342932,-0.2635152936,-0.0356827825,-0.0357737392,-0.0839617401,0.2509799898,0.2607079744,-0.1420855969,-0.1288410872,-0.0528032482,0.5240070224,0.1827690005,0.0712557435,-0.0331554189,0.0538785122,-0.3607849777,0.0930320472,-0.0018490237,0.0711949542,-0.0294784587,-0.0747773871,0.1303329021,0.0075901668,0.5740774274,-0.0426572338,0.2700049579,0.1973776221,-0.1476932913,-0.4956389368,-0.0050180582,0.3244312108,0.3547219038,-0.2119250149,0.5544860959,0.0248028394,-0.3523011208,-0.1425226629,0.288828671,-0.3024058342,-0.2755249143,-0.2828812003,-0.09085875,-0.1693260074,-0.1359177679,-0.2644427717,0.1604723334,0.3173460662,-0.0542619824,0.0199967362,0.2740503252,-0.1012237146,-0.2405223846,0.2603538632,-0.0512351431,0.1918953061,0.3695375919,0.2724651992,0.1330745071,0.4891783595,-0.1525875926,0.0589672066,0.1066776514,-0.113842994,0.2623552084,0.3542262018,-0.2661116123,-0.2850041986,0.435588181,-0.0714087635,-0.4431219697,0.1451996118,0.0650886223,-0.0585033596,-0.0124943126,-0.7041835189,0.3211218417,-0.0287692212,-0.0916713178,0.3414559066,-0.276104629,-0.3676846325,0.0366566293,0.051457379,0.5803357363,-0.1109289154,0.3393071294,0.4912239909,0.1037682444,-0.1205616519,-0.0198097471,-0.116348587,-0.3502036333,0.0732993707,-0.0753939599,-0.2070186734,0.0714948401,0.4098730981,0.0838324577,0.3290243149,-0.2101148963,0.4944488406,0.0073868916,0.2272144705,0.0601578616,-0.353441447,-0.0565701425,0.0522741452,0.064260304,0.284347266,-0.1775835007,0.096294865,0.0419204757,-0.1211866885,0.1079063714,0.0236521065,-0.1989729255,0.0379890725,0.1669697911,-0.3038426936,-0.3285149634,0.3571901321,0.185563758,-0.084210217,-0.2812748551,0.1942753941,-0.145205304,0.1037394926,-0.161561802,-0.0587676428,0.1501354575,-0.0364120118,0.008747844,-0.3542821407,-0.3601642847,-0.3469100893,0.2884579599,0.0409501418,0.0974462181,-0.1604705304,-0.3576007485,0.0155936955,0.315689981,-0.0022195692,0.0440252461,0.1933984458,-0.0006950215,0.0494024679,-0.0000182195,-0.522116065,-0.1139214337,0.2639822066,-0.059195362,0.139902547,0.4883067906,0.1480217427,-0.1267182678,-0.0911748558,0.2864241302,0.2808643579,-0.0819671527,0.081863746,0.2499215305,0.0719754174,0.0344536975,0.4461497962,0.1257199347,-0.1293550432,-0.2481327951,-0.3615415394,-0.3472002447,0.059997879,-0.1547261178,0.1391308159,0.2028879076,-0.0808751956,0.1917208284,-0.2964784801,-0.1524124444,-0.0502970777,0.0777829066,0.0693687126,0.2164888531,0.2236094326,0.3147010505,-0.0005695213,-0.0675306544,-0.3426747024,-0.2485060394,-0.0901537165,-0.0261395015,0.2068739235,0.012220202,-0.0691921636,0.0507295877,-0.3420846164,-0.3368499279,-0.2921670377,0.2289035469,0.2773245573,-0.0991031229,-0.045830138,0.2431878746,0.132675916,-0.0690466017,0.1624124795,-0.0383186452,0.3669086695,-0.0795844197,0.1248300299,-0.3500242531,-0.0815616697,0.1967009157,0.2148936838,0.0490618981,0.1604729295,0.3099260628,-0.4675811827,0.180360496,0.14351888,0.4164133072,0.2039120644,-0.1885019988,0.0596927404,0.2156945616,-0.0126175927,-0.0489641428,-0.137092188,0.1080824882,0.0304838996,-0.0306821875,0.3173310459,0.1685622633,-0.08036948,-0.0899000317,0.0471223369,0.4480670393,-0.0706882551,-0.0507538356,0.3608314395,-0.1141717434,0.359721601,0.2721737325,0.3458629549,0.1139599159,0.5598076582,-0.0526599064,0.0000243527,0.2040197104,0.1723182052,-0.1634253114,-0.5923697352,0.3253557086,0.2390280366,-0.1098720208,-0.1611026824,-0.1157057136,0.5343716741,-0.2219631076,0.2505008578,-0.2162912637,0.1709721684,-0.1029276997,-0.4251091182,-0.3079388738,-0.1412219554,0.2282500267,-0.0186241493,-0.1614607424,-0.064449735,-0.1494418085,-0.1645069271,0.1885359585,0.0629336536,0.3134792447,-0.0528451093,-0.4776817858,-0.2475821078,0.3105486929,-0.2090513706,0.1323187202,0.3261492848,0.0168294068,0.2841062248,0.3334079981,0.0473643169,-0.096279487,-0.0710704029,-0.1093626693,-0.3242408931,0.1764926612,-0.3072253168,0.0820762217,0.4704864919,0.0694290102,-0.0697874948,0.0879534706,0.1533005238,0.3244493306,-0.2235340774,0.313650012,0.1243327856,-0.0068724081,-0.0762718245,0.1700989455,-0.2068653256,-0.0677663982,0.610519886,0.0770614594,0.3127712607,-0.0215629991,-0.0132977683,-0.0179821271,0.6097301245,0.1546905041,0.264940232,-0.207294479,0.0087386556,-0.7007184625,0.0271119792,-0.0970675871,0.3181787133,0.1544747651,0.1189372689,0.1033614054,0.0525310747,0.0455731302,-0.0127336476,0.106608443,-0.1811340749,0.2178354412,-0.3707609773,-0.2538194954,0.1991679072,0.0606617779,-0.4984818697,0.080563575,-0.3372354209,-0.0870117471,0.0640451312,0.0036411204,0.2089632899,-0.0651070476,0.252884686,-0.1077311411,0.1323127151,0.061413303,0.066924803,-0.4829733372,-0.2013020813,-0.3909189403,0.0929685459,-0.5008547306,0.0383653156,-0.6236851215,0.1119696125,-0.6095497608,-0.0392553546,0.1357252747,-0.1387305111,-0.0422170348,-0.0555080175,-0.415022999,0.1858195513,0.1746155471,0.0111049954,-0.0177153181,0.1192523837,0.0035476787,-0.0809878036,0.6704128981,-0.2766206861,-0.2386036366,-0.0440811776,-0.072656922,0.0300052352,-0.3232818544,-0.227857396,0.1609710306,0.4065617621,-0.0712540522,0.0153429443,-0.2961613238,-0.0806938484,0.2529427409,-0.1226013079,0.1222223565,0.016147783,-0.2187944651,-0.2271164507,-0.3131069243]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2363","title":"Trying to use metric.compute but get OSError","comments":"also, I test the function on some little data , get the same message:\r\n\r\n```\r\nPython 3.8.5 (default, Jan 27 2021, 15:41:15)\r\n[GCC 9.3.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric('accuracy')\r\n>>> metric.add_batch(predictions=[1, 1, 1, 1], references=[1, 1, 0, 0])\r\n2021-05-15 16:39:17.240991: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\n>>> metric.compute()\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 112, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```","body":"I want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n","comment_length":113,"text":"Trying to use metric.compute but get OSError\nI want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n\nalso, I test the function on some little data , get the same message:\r\n\r\n```\r\nPython 3.8.5 (default, Jan 27 2021, 15:41:15)\r\n[GCC 9.3.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric('accuracy')\r\n>>> metric.add_batch(predictions=[1, 1, 1, 1], references=[1, 1, 0, 0])\r\n2021-05-15 16:39:17.240991: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\n>>> metric.compute()\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 112, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```","embeddings":[-0.3944184184,-0.2529711425,-0.0330115929,0.3631941974,0.253131628,-0.1011049151,0.2328529209,0.139267087,0.2106887102,0.6745797396,-0.0271198936,0.0674137399,-0.0561968796,0.0007951714,-0.1410534978,-0.1666642725,-0.1239938736,0.0129452525,0.0000074239,0.0780162588,-0.3373332918,0.0424939394,-0.2203928828,0.0718222335,-0.0541991033,-0.2079253197,0.206342876,0.042613849,-0.5113474727,-0.2387506664,0.2351529449,-0.3208845258,0.2170303017,0.8475168347,-0.0001149249,0.1326398402,0.0087183854,-0.1004110426,-0.0945497677,-0.276073128,0.2104712427,-0.2102968097,0.0614107549,-0.2894953787,0.1416248083,-0.025001768,-0.2049153298,-0.3347912133,0.2580504715,0.5588136911,0.1094384342,0.4805767834,0.0571895204,-0.1703307778,-0.374795258,-0.344244957,0.0392587148,0.670753479,-0.0942461342,-0.1250064075,-0.0927235708,-0.0397556424,0.2341285497,0.0877322108,0.206516698,-0.1326481998,0.8326205611,-0.2251918614,-0.1489875764,-0.0572964847,0.1004799753,0.0542257167,-0.3662271798,0.3212551177,-0.0640987307,-0.7123773694,-0.0083443522,-0.0554771312,-0.329859823,-0.1663537025,-0.3520748615,-0.0161977299,-0.4670949578,0.0650909245,-0.2273944765,0.2691727877,-0.1618948281,0.0761636347,0.314235419,0.3356452584,0.0720360652,-0.0285175629,0.1678747386,0.2304376811,-0.5341535807,-0.0079725357,-0.1267386079,-0.0935950652,-0.2341202796,-0.2478045374,0.3044139445,0.0229335669,0.2392277718,0.27803123,-0.1958615184,0.8154187799,0.0828209221,0.1414985061,0.0289000422,0.3671554923,-0.0761498585,-0.0763362721,-0.0666468665,-0.3928147554,0.3305636346,0.1671797335,-0.1789496094,-0.0470937341,-0.5103244781,-0.0825716555,-0.3074188232,-0.1921125054,0.3589552343,0.0782146156,-0.1096397415,0.1300759017,0.2878340185,0.190229252,-0.2090027928,-0.1163052544,-0.2376890033,0.325322479,-0.2651518881,0.1427433044,-0.2088134736,-0.2825702131,0.2108990848,-0.0339325406,0.2929007113,-0.372744292,0.1489905268,-0.1023257896,-0.0621559471,0.0755864009,0.1042392701,-0.2266150713,0.4715836942,-0.188901484,0.1012283415,0.0352442786,-0.6336891055,-0.4329087138,0.0577113368,0.2142496556,-0.1043527424,0.11758966,-0.2650551498,0.0452635139,0.3239228129,0.1826907694,-0.0352026895,-0.1126695797,-0.417783916,-0.147210747,0.2840802968,-0.2580549717,-0.245167166,0.0119533483,0.0527974665,-0.091761142,0.2312343568,0.238289997,-0.1830744445,0.6743611097,-0.0939318314,-0.2390212417,0.5787590742,-0.5383852124,-0.0357063636,-0.2172711045,-0.193100974,-0.1919911057,0.1511501074,-0.107356593,0.0182099994,0.0467908122,0.4387013316,0.0395708159,-0.0889608786,-0.3561971188,-0.1995178908,0.2851071954,0.2279101908,0.0419763997,0.4570365548,0.2853614986,0.1695158035,-0.0684723854,0.1887974292,-0.0180304926,-0.2879210114,-0.0030238016,0.0400323123,-0.3892459273,0.2894599736,-0.0779185146,0.2788540721,0.2235236019,-0.246622175,0.0826151818,0.3609803319,0.2236855477,-0.4406271875,0.2417325824,-0.1057087854,-0.0480109528,0.0733944699,-0.2125060558,0.2519195378,-0.0479268841,-0.0860952213,0.0504066125,-0.4933129251,-0.0235673264,-0.1894487143,0.0653727204,-0.1262327582,-0.0448605902,-0.1458986551,-0.0153140696,0.1543092728,0.0556869097,-0.2693291008,0.4298380017,0.2029634118,-0.1456071436,-0.0144350883,0.1559263766,-0.1474841088,0.1328604072,-0.0093299672,-0.1915466785,0.2004588246,-0.1141059697,0.0535784177,0.3843781948,0.1535725743,0.0171225872,-0.1076225862,0.0651368126,-0.0502192862,0.2021970153,0.0288014244,0.0270202309,0.4426712394,-0.3593238294,0.0580615848,-0.3788809478,0.3278771341,-0.4405206144,0.3899686038,-0.0393587872,0.0601105914,-0.0102301594,0.1284069866,-0.0726182908,-0.3743269444,-0.2551704347,0.4715960026,-0.0478855297,0.0405350626,0.0953617096,-0.0711544156,-0.0679310337,0.0304413363,0.3493910134,0.2070938945,0.0212720446,0.0449011773,-0.0521538779,-0.0465460382,-0.1306073666,-0.0280886311,0.3493729532,-0.225588873,0.1348799318,0.075287506,0.2183084935,-0.1993987262,-0.1545158476,-0.2003663927,-0.3204384446,0.0115698399,0.1411631107,0.3208896518,-0.1376408786,0.1372115463,0.1250700653,0.4607310593,0.2262209803,0.2846042216,-0.1636532545,0.0487440601,0.0333901569,-0.0979563445,-0.215823099,0.2466624081,-0.3523681462,0.0774699152,-0.0886008814,-0.1845653355,-0.1599787623,0.0074018487,0.2293597162,0.1024663746,-0.0420796461,-0.255674541,0.0428051874,0.2096294463,-0.2503049076,-0.3007627726,0.374556154,-0.058750622,-0.0117408587,-0.1317933202,-0.1412438005,-0.1016193479,-0.2883880734,-0.0360812098,0.2483190298,0.2062901258,0.2738169432,-0.005697303,0.3814315796,0.4019247591,-0.0543019064,-0.1986670345,-0.547450006,0.2148372084,-0.0925638452,-0.2276468575,0.0701643974,-0.0187872667,0.1677819192,0.1912883967,-0.2886455953,-0.673584044,0.1889113188,0.2871622443,-0.2609939575,0.1604726017,0.2804620564,-0.3490063846,-0.0287137385,-0.2814407945,-0.103891097,0.2245209068,0.0848220885,0.2112163156,-0.221072346,0.4646460712,0.0941406488,0.6489886642,0.0296453517,-0.1281051636,0.2985831499,0.1537393332,0.2447576225,-0.0038530349,-0.1338460147,0.0489935353,0.1064159945,-0.2038707584,0.1017027572,-0.0557483211,0.065122366,-0.3501721025,-0.2313645333,0.0471409708,-0.0168843288,-0.0369388424,-0.4160637259,0.3232347369,-0.0301211886,0.2697743475,-0.2884373367,0.1355518103,0.135763377,0.1578714401,0.1288013011,-0.0568930842,-0.0139695946,-0.170657292,-0.2526807189,0.393801868,0.0498392619,0.2946966588,-0.0771579444,0.3258003592,-0.1086524948,0.1709336042,0.6087872386,-0.1312526315,0.0350221097,0.1956077218,-0.2016815692,-0.3555338085,-0.0019054203,-0.3041402996,0.1678823978,0.0728453323,0.7506990433,-0.0244571082,0.0154515868,-0.0228771269,0.3795055747,-0.222904548,-0.386636287,-0.259060055,-0.239826262,0.2020212114,0.0542650297,0.0234459173,0.3708866239,0.1194965094,0.2322513759,-0.0566958673,0.1982782334,0.0937500745,-0.0649097115,0.0116167786,-0.1418979466,-0.1390072256,0.054092776,-0.0561870784,0.2651714385,0.0636063591,-0.0457775891,-0.5454137325,0.1777900755,0.1208341345,0.2388094217,-0.0866866931,-0.2268003374,-0.1198794544,-0.0450631194,0.2277987301,0.1475597471,0.1014586389,0.2041599751,0.3210230768,-0.2983633578,0.0091705713,0.4024381936,0.2755218446,0.1890118122,0.1654900014,-0.2372071296,-0.1963230819,0.0678781047,0.3185476363,0.4146768153,0.0382899977,-0.3658927679,0.524125874,0.2566727698,0.1425125748,-0.2115787119,0.1814755797,-0.1015443057,-0.4452896118,-0.1305099875,-0.2162659913,0.1773560345,-0.1759329587,-0.1006761268,0.5212775469,-0.2374328077,-0.0595539846,0.2230506241,0.0124302059,-0.2171092629,-0.1357540488,-0.173538208,0.094317548,-0.1034941152,0.1680279076,-0.2548354268,-0.0296272207,-0.0258480571,-0.0543110035,0.0665848777,-0.1761960685,-0.3191459775,0.2255784273,0.1246949434,-0.307084173,0.023898663,0.2432399392,0.5752543211,-0.0661970228,0.0134257153,-0.012469504,-0.1434542537,-0.0496817417,0.1397010684,-0.1298236251,0.2827926874,-0.2351443321,0.0005081835,0.106917724,-0.2279693782,-0.3003264666,0.4230240583,0.0332099907,0.1173307002,0.0298740864,0.0688449815,-0.3049264252,-0.1818187386,-0.1941626072,0.1032953188,0.1751461476,-0.4289224446,0.3889814317,-0.0574683212,-0.37502563,0.2138925493,0.1949261874,0.2153660208,-0.2504068017,0.2916764319,-0.0620119981,-0.1055545658,-0.058654286,-0.166161254,0.0028955101,-0.0192075521,0.1990184635,0.018567238,0.352901727,0.0213241875,0.2117160261,0.3017182946,0.0511455461,-0.1592131406,-0.7366719246,-0.2020980716,0.0577510595,-0.0001070497,-0.1427347213,-0.1370648146,0.2167479694,-0.1322854012,0.1760246009,-0.2509111762,-0.0105598839,-0.2350025028,0.1568703502,0.0424210578,-0.148717314,0.2262828201,0.1627236456,0.0156815946,0.2249858826,-0.2116453201,-0.1787355393,0.1025301293,0.1621220708,-0.059812367,-0.0652413741,0.3917291164,0.0371191427,-0.3599075377,0.0558811985,-0.2302165478,0.3071429133,-0.1411307454,-0.1288602799,0.1621950716,0.6288257837,0.090284504,0.0673449486,0.005225874,0.2742750943,-0.2251279801,0.1829318851,-0.2780632973,-0.1428376585,-0.3569362462,-0.1561434567,0.0151984952,-0.1730259657,0.2919380367,-0.2861347795,0.0282903668,-0.2088741958,0.3628519475,0.0899800509,-0.1975496709,0.0473193005,0.0589776151,0.1172167733,-0.0800810456,0.1394037902,0.172870785,0.0388543792,0.0148108676,0.332151413,-0.01107537,0.1098471284,-0.0826835781,-0.0821306184,0.0725666881,0.0972760096,-0.042032022,-0.1037478149,-0.1657619178,0.1477657557,0.4743666351,0.0668724999,0.2182660401,0.0428017601,-0.1522810459,0.3543897271,0.127581805,-0.0173842087,-0.2526034713,-0.0494091697,-0.1836319864,0.0211601369,-0.0532549359,-0.1062147692,-0.2489045709,0.0879066959,-0.4121108651,0.1509920657,-0.2156967521,0.1265384853,-0.0160518922,0.3691642284,-0.3033030033,-0.2289079726,0.0783943161,-0.2160382569,0.2938553691,0.1115008816,-0.2649542689,-0.0027564408,-0.2185145319,0.1251319349,-0.4348650575,0.5451754332,0.2303571552,-0.3177498579,0.1950152814,0.2211612761,0.2615064979,-0.0830582082,0.3307033479,0.4944124222,0.4311096668,-0.0547136292,0.102584891,0.1507220715,-0.1535886228,0.1190046594,0.1592704803,0.1158066168,-0.2425831556,0.5202123523,0.1174396276,-0.2335672528,0.3855023086,-0.2376741767,0.4925028384,-0.389862299,0.0585529953,-0.0203453898,0.0341270305,-0.3185689747,-0.0753840879,-0.155657202,0.1755642593,-0.1999762207,0.0327101126,0.0760909542,-0.078606151,0.0525699221,-0.298938632,0.3373302817,0.3318254352,0.297365725,-0.3848649859,-0.0060853418,-0.4286868274,0.1391412169,0.0209710095,0.0802477226,-0.0116070267,0.0408270136,0.0421428196,0.2475109547,0.4816519022,0.0979122743,0.1235919222,0.2023987174,-0.4470593035,-0.1758026332,-0.1573079526,-0.3478933573,0.0999621078,-0.2129716724,-0.0290640537,-0.3214386106,0.0842481554,0.0336777754,-0.0423010625,0.232040301,0.6045469642,0.2200201899,0.067227751,0.4585684538,0.0525996536,0.1894745231,0.3232182264,-0.0218697302,-0.2543055415,0.0623143725,-0.0039302143,0.2870970368,-0.0663463175,-0.287137717,-0.1954174787,0.6739629507,0.0616975725,-0.5848380923,-0.0630172789,0.1214392781,0.0211719479,-0.1331175119,0.4323017597,0.066798456,0.1964479834,0.2162305713,-0.2929868996,0.0216373373,0.4165513813,-0.1725697666,-0.382404089,-0.0623855852,0.1373431087,0.1988154203,-0.2434106916,-0.4655195475,0.0824157745,0.4198485613,-0.1423687786,-0.1497416049,0.1252377182,-0.3407678008,0.2748536766,0.2406638563,0.0039691809,-0.2093468755,-0.2314013988,-0.0217551906,-0.1318709552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2363","title":"Trying to use metric.compute but get OSError","comments":"Hi @hyusterr,\r\nIf you look at the example provided in `metrics\/accuracy.py`, it only does `metric.compute()` to calculate the accuracy. Here's an example:\r\n```\r\nfrom datasets import load_metric\r\nmetric = load_metric('accuracy')\r\noutput = metric.compute(predictions=[1, 1, 1, 1], references=[1, 1, 0, 0])\r\nprint(output['accuracy']) # 0.5\r\n```\r\n","body":"I want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n","comment_length":44,"text":"Trying to use metric.compute but get OSError\nI want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n\nHi @hyusterr,\r\nIf you look at the example provided in `metrics\/accuracy.py`, it only does `metric.compute()` to calculate the accuracy. Here's an example:\r\n```\r\nfrom datasets import load_metric\r\nmetric = load_metric('accuracy')\r\noutput = metric.compute(predictions=[1, 1, 1, 1], references=[1, 1, 0, 0])\r\nprint(output['accuracy']) # 0.5\r\n```\r\n","embeddings":[-0.3944184184,-0.2529711425,-0.0330115929,0.3631941974,0.253131628,-0.1011049151,0.2328529209,0.139267087,0.2106887102,0.6745797396,-0.0271198936,0.0674137399,-0.0561968796,0.0007951714,-0.1410534978,-0.1666642725,-0.1239938736,0.0129452525,0.0000074239,0.0780162588,-0.3373332918,0.0424939394,-0.2203928828,0.0718222335,-0.0541991033,-0.2079253197,0.206342876,0.042613849,-0.5113474727,-0.2387506664,0.2351529449,-0.3208845258,0.2170303017,0.8475168347,-0.0001149249,0.1326398402,0.0087183854,-0.1004110426,-0.0945497677,-0.276073128,0.2104712427,-0.2102968097,0.0614107549,-0.2894953787,0.1416248083,-0.025001768,-0.2049153298,-0.3347912133,0.2580504715,0.5588136911,0.1094384342,0.4805767834,0.0571895204,-0.1703307778,-0.374795258,-0.344244957,0.0392587148,0.670753479,-0.0942461342,-0.1250064075,-0.0927235708,-0.0397556424,0.2341285497,0.0877322108,0.206516698,-0.1326481998,0.8326205611,-0.2251918614,-0.1489875764,-0.0572964847,0.1004799753,0.0542257167,-0.3662271798,0.3212551177,-0.0640987307,-0.7123773694,-0.0083443522,-0.0554771312,-0.329859823,-0.1663537025,-0.3520748615,-0.0161977299,-0.4670949578,0.0650909245,-0.2273944765,0.2691727877,-0.1618948281,0.0761636347,0.314235419,0.3356452584,0.0720360652,-0.0285175629,0.1678747386,0.2304376811,-0.5341535807,-0.0079725357,-0.1267386079,-0.0935950652,-0.2341202796,-0.2478045374,0.3044139445,0.0229335669,0.2392277718,0.27803123,-0.1958615184,0.8154187799,0.0828209221,0.1414985061,0.0289000422,0.3671554923,-0.0761498585,-0.0763362721,-0.0666468665,-0.3928147554,0.3305636346,0.1671797335,-0.1789496094,-0.0470937341,-0.5103244781,-0.0825716555,-0.3074188232,-0.1921125054,0.3589552343,0.0782146156,-0.1096397415,0.1300759017,0.2878340185,0.190229252,-0.2090027928,-0.1163052544,-0.2376890033,0.325322479,-0.2651518881,0.1427433044,-0.2088134736,-0.2825702131,0.2108990848,-0.0339325406,0.2929007113,-0.372744292,0.1489905268,-0.1023257896,-0.0621559471,0.0755864009,0.1042392701,-0.2266150713,0.4715836942,-0.188901484,0.1012283415,0.0352442786,-0.6336891055,-0.4329087138,0.0577113368,0.2142496556,-0.1043527424,0.11758966,-0.2650551498,0.0452635139,0.3239228129,0.1826907694,-0.0352026895,-0.1126695797,-0.417783916,-0.147210747,0.2840802968,-0.2580549717,-0.245167166,0.0119533483,0.0527974665,-0.091761142,0.2312343568,0.238289997,-0.1830744445,0.6743611097,-0.0939318314,-0.2390212417,0.5787590742,-0.5383852124,-0.0357063636,-0.2172711045,-0.193100974,-0.1919911057,0.1511501074,-0.107356593,0.0182099994,0.0467908122,0.4387013316,0.0395708159,-0.0889608786,-0.3561971188,-0.1995178908,0.2851071954,0.2279101908,0.0419763997,0.4570365548,0.2853614986,0.1695158035,-0.0684723854,0.1887974292,-0.0180304926,-0.2879210114,-0.0030238016,0.0400323123,-0.3892459273,0.2894599736,-0.0779185146,0.2788540721,0.2235236019,-0.246622175,0.0826151818,0.3609803319,0.2236855477,-0.4406271875,0.2417325824,-0.1057087854,-0.0480109528,0.0733944699,-0.2125060558,0.2519195378,-0.0479268841,-0.0860952213,0.0504066125,-0.4933129251,-0.0235673264,-0.1894487143,0.0653727204,-0.1262327582,-0.0448605902,-0.1458986551,-0.0153140696,0.1543092728,0.0556869097,-0.2693291008,0.4298380017,0.2029634118,-0.1456071436,-0.0144350883,0.1559263766,-0.1474841088,0.1328604072,-0.0093299672,-0.1915466785,0.2004588246,-0.1141059697,0.0535784177,0.3843781948,0.1535725743,0.0171225872,-0.1076225862,0.0651368126,-0.0502192862,0.2021970153,0.0288014244,0.0270202309,0.4426712394,-0.3593238294,0.0580615848,-0.3788809478,0.3278771341,-0.4405206144,0.3899686038,-0.0393587872,0.0601105914,-0.0102301594,0.1284069866,-0.0726182908,-0.3743269444,-0.2551704347,0.4715960026,-0.0478855297,0.0405350626,0.0953617096,-0.0711544156,-0.0679310337,0.0304413363,0.3493910134,0.2070938945,0.0212720446,0.0449011773,-0.0521538779,-0.0465460382,-0.1306073666,-0.0280886311,0.3493729532,-0.225588873,0.1348799318,0.075287506,0.2183084935,-0.1993987262,-0.1545158476,-0.2003663927,-0.3204384446,0.0115698399,0.1411631107,0.3208896518,-0.1376408786,0.1372115463,0.1250700653,0.4607310593,0.2262209803,0.2846042216,-0.1636532545,0.0487440601,0.0333901569,-0.0979563445,-0.215823099,0.2466624081,-0.3523681462,0.0774699152,-0.0886008814,-0.1845653355,-0.1599787623,0.0074018487,0.2293597162,0.1024663746,-0.0420796461,-0.255674541,0.0428051874,0.2096294463,-0.2503049076,-0.3007627726,0.374556154,-0.058750622,-0.0117408587,-0.1317933202,-0.1412438005,-0.1016193479,-0.2883880734,-0.0360812098,0.2483190298,0.2062901258,0.2738169432,-0.005697303,0.3814315796,0.4019247591,-0.0543019064,-0.1986670345,-0.547450006,0.2148372084,-0.0925638452,-0.2276468575,0.0701643974,-0.0187872667,0.1677819192,0.1912883967,-0.2886455953,-0.673584044,0.1889113188,0.2871622443,-0.2609939575,0.1604726017,0.2804620564,-0.3490063846,-0.0287137385,-0.2814407945,-0.103891097,0.2245209068,0.0848220885,0.2112163156,-0.221072346,0.4646460712,0.0941406488,0.6489886642,0.0296453517,-0.1281051636,0.2985831499,0.1537393332,0.2447576225,-0.0038530349,-0.1338460147,0.0489935353,0.1064159945,-0.2038707584,0.1017027572,-0.0557483211,0.065122366,-0.3501721025,-0.2313645333,0.0471409708,-0.0168843288,-0.0369388424,-0.4160637259,0.3232347369,-0.0301211886,0.2697743475,-0.2884373367,0.1355518103,0.135763377,0.1578714401,0.1288013011,-0.0568930842,-0.0139695946,-0.170657292,-0.2526807189,0.393801868,0.0498392619,0.2946966588,-0.0771579444,0.3258003592,-0.1086524948,0.1709336042,0.6087872386,-0.1312526315,0.0350221097,0.1956077218,-0.2016815692,-0.3555338085,-0.0019054203,-0.3041402996,0.1678823978,0.0728453323,0.7506990433,-0.0244571082,0.0154515868,-0.0228771269,0.3795055747,-0.222904548,-0.386636287,-0.259060055,-0.239826262,0.2020212114,0.0542650297,0.0234459173,0.3708866239,0.1194965094,0.2322513759,-0.0566958673,0.1982782334,0.0937500745,-0.0649097115,0.0116167786,-0.1418979466,-0.1390072256,0.054092776,-0.0561870784,0.2651714385,0.0636063591,-0.0457775891,-0.5454137325,0.1777900755,0.1208341345,0.2388094217,-0.0866866931,-0.2268003374,-0.1198794544,-0.0450631194,0.2277987301,0.1475597471,0.1014586389,0.2041599751,0.3210230768,-0.2983633578,0.0091705713,0.4024381936,0.2755218446,0.1890118122,0.1654900014,-0.2372071296,-0.1963230819,0.0678781047,0.3185476363,0.4146768153,0.0382899977,-0.3658927679,0.524125874,0.2566727698,0.1425125748,-0.2115787119,0.1814755797,-0.1015443057,-0.4452896118,-0.1305099875,-0.2162659913,0.1773560345,-0.1759329587,-0.1006761268,0.5212775469,-0.2374328077,-0.0595539846,0.2230506241,0.0124302059,-0.2171092629,-0.1357540488,-0.173538208,0.094317548,-0.1034941152,0.1680279076,-0.2548354268,-0.0296272207,-0.0258480571,-0.0543110035,0.0665848777,-0.1761960685,-0.3191459775,0.2255784273,0.1246949434,-0.307084173,0.023898663,0.2432399392,0.5752543211,-0.0661970228,0.0134257153,-0.012469504,-0.1434542537,-0.0496817417,0.1397010684,-0.1298236251,0.2827926874,-0.2351443321,0.0005081835,0.106917724,-0.2279693782,-0.3003264666,0.4230240583,0.0332099907,0.1173307002,0.0298740864,0.0688449815,-0.3049264252,-0.1818187386,-0.1941626072,0.1032953188,0.1751461476,-0.4289224446,0.3889814317,-0.0574683212,-0.37502563,0.2138925493,0.1949261874,0.2153660208,-0.2504068017,0.2916764319,-0.0620119981,-0.1055545658,-0.058654286,-0.166161254,0.0028955101,-0.0192075521,0.1990184635,0.018567238,0.352901727,0.0213241875,0.2117160261,0.3017182946,0.0511455461,-0.1592131406,-0.7366719246,-0.2020980716,0.0577510595,-0.0001070497,-0.1427347213,-0.1370648146,0.2167479694,-0.1322854012,0.1760246009,-0.2509111762,-0.0105598839,-0.2350025028,0.1568703502,0.0424210578,-0.148717314,0.2262828201,0.1627236456,0.0156815946,0.2249858826,-0.2116453201,-0.1787355393,0.1025301293,0.1621220708,-0.059812367,-0.0652413741,0.3917291164,0.0371191427,-0.3599075377,0.0558811985,-0.2302165478,0.3071429133,-0.1411307454,-0.1288602799,0.1621950716,0.6288257837,0.090284504,0.0673449486,0.005225874,0.2742750943,-0.2251279801,0.1829318851,-0.2780632973,-0.1428376585,-0.3569362462,-0.1561434567,0.0151984952,-0.1730259657,0.2919380367,-0.2861347795,0.0282903668,-0.2088741958,0.3628519475,0.0899800509,-0.1975496709,0.0473193005,0.0589776151,0.1172167733,-0.0800810456,0.1394037902,0.172870785,0.0388543792,0.0148108676,0.332151413,-0.01107537,0.1098471284,-0.0826835781,-0.0821306184,0.0725666881,0.0972760096,-0.042032022,-0.1037478149,-0.1657619178,0.1477657557,0.4743666351,0.0668724999,0.2182660401,0.0428017601,-0.1522810459,0.3543897271,0.127581805,-0.0173842087,-0.2526034713,-0.0494091697,-0.1836319864,0.0211601369,-0.0532549359,-0.1062147692,-0.2489045709,0.0879066959,-0.4121108651,0.1509920657,-0.2156967521,0.1265384853,-0.0160518922,0.3691642284,-0.3033030033,-0.2289079726,0.0783943161,-0.2160382569,0.2938553691,0.1115008816,-0.2649542689,-0.0027564408,-0.2185145319,0.1251319349,-0.4348650575,0.5451754332,0.2303571552,-0.3177498579,0.1950152814,0.2211612761,0.2615064979,-0.0830582082,0.3307033479,0.4944124222,0.4311096668,-0.0547136292,0.102584891,0.1507220715,-0.1535886228,0.1190046594,0.1592704803,0.1158066168,-0.2425831556,0.5202123523,0.1174396276,-0.2335672528,0.3855023086,-0.2376741767,0.4925028384,-0.389862299,0.0585529953,-0.0203453898,0.0341270305,-0.3185689747,-0.0753840879,-0.155657202,0.1755642593,-0.1999762207,0.0327101126,0.0760909542,-0.078606151,0.0525699221,-0.298938632,0.3373302817,0.3318254352,0.297365725,-0.3848649859,-0.0060853418,-0.4286868274,0.1391412169,0.0209710095,0.0802477226,-0.0116070267,0.0408270136,0.0421428196,0.2475109547,0.4816519022,0.0979122743,0.1235919222,0.2023987174,-0.4470593035,-0.1758026332,-0.1573079526,-0.3478933573,0.0999621078,-0.2129716724,-0.0290640537,-0.3214386106,0.0842481554,0.0336777754,-0.0423010625,0.232040301,0.6045469642,0.2200201899,0.067227751,0.4585684538,0.0525996536,0.1894745231,0.3232182264,-0.0218697302,-0.2543055415,0.0623143725,-0.0039302143,0.2870970368,-0.0663463175,-0.287137717,-0.1954174787,0.6739629507,0.0616975725,-0.5848380923,-0.0630172789,0.1214392781,0.0211719479,-0.1331175119,0.4323017597,0.066798456,0.1964479834,0.2162305713,-0.2929868996,0.0216373373,0.4165513813,-0.1725697666,-0.382404089,-0.0623855852,0.1373431087,0.1988154203,-0.2434106916,-0.4655195475,0.0824157745,0.4198485613,-0.1423687786,-0.1497416049,0.1252377182,-0.3407678008,0.2748536766,0.2406638563,0.0039691809,-0.2093468755,-0.2314013988,-0.0217551906,-0.1318709552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2363","title":"Trying to use metric.compute but get OSError","comments":"I thought I can use Metric to collect predictions and references, this follows the step from huggingface's sample colab.\r\nBTW, I fix the problem by setting other cache_dir in load_metric, but I'm still wondering about the mechanism.","body":"I want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n","comment_length":37,"text":"Trying to use metric.compute but get OSError\nI want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n\nI thought I can use Metric to collect predictions and references, this follows the step from huggingface's sample colab.\r\nBTW, I fix the problem by setting other cache_dir in load_metric, but I'm still wondering about the mechanism.","embeddings":[-0.3944184184,-0.2529711425,-0.0330115929,0.3631941974,0.253131628,-0.1011049151,0.2328529209,0.139267087,0.2106887102,0.6745797396,-0.0271198936,0.0674137399,-0.0561968796,0.0007951714,-0.1410534978,-0.1666642725,-0.1239938736,0.0129452525,0.0000074239,0.0780162588,-0.3373332918,0.0424939394,-0.2203928828,0.0718222335,-0.0541991033,-0.2079253197,0.206342876,0.042613849,-0.5113474727,-0.2387506664,0.2351529449,-0.3208845258,0.2170303017,0.8475168347,-0.0001149249,0.1326398402,0.0087183854,-0.1004110426,-0.0945497677,-0.276073128,0.2104712427,-0.2102968097,0.0614107549,-0.2894953787,0.1416248083,-0.025001768,-0.2049153298,-0.3347912133,0.2580504715,0.5588136911,0.1094384342,0.4805767834,0.0571895204,-0.1703307778,-0.374795258,-0.344244957,0.0392587148,0.670753479,-0.0942461342,-0.1250064075,-0.0927235708,-0.0397556424,0.2341285497,0.0877322108,0.206516698,-0.1326481998,0.8326205611,-0.2251918614,-0.1489875764,-0.0572964847,0.1004799753,0.0542257167,-0.3662271798,0.3212551177,-0.0640987307,-0.7123773694,-0.0083443522,-0.0554771312,-0.329859823,-0.1663537025,-0.3520748615,-0.0161977299,-0.4670949578,0.0650909245,-0.2273944765,0.2691727877,-0.1618948281,0.0761636347,0.314235419,0.3356452584,0.0720360652,-0.0285175629,0.1678747386,0.2304376811,-0.5341535807,-0.0079725357,-0.1267386079,-0.0935950652,-0.2341202796,-0.2478045374,0.3044139445,0.0229335669,0.2392277718,0.27803123,-0.1958615184,0.8154187799,0.0828209221,0.1414985061,0.0289000422,0.3671554923,-0.0761498585,-0.0763362721,-0.0666468665,-0.3928147554,0.3305636346,0.1671797335,-0.1789496094,-0.0470937341,-0.5103244781,-0.0825716555,-0.3074188232,-0.1921125054,0.3589552343,0.0782146156,-0.1096397415,0.1300759017,0.2878340185,0.190229252,-0.2090027928,-0.1163052544,-0.2376890033,0.325322479,-0.2651518881,0.1427433044,-0.2088134736,-0.2825702131,0.2108990848,-0.0339325406,0.2929007113,-0.372744292,0.1489905268,-0.1023257896,-0.0621559471,0.0755864009,0.1042392701,-0.2266150713,0.4715836942,-0.188901484,0.1012283415,0.0352442786,-0.6336891055,-0.4329087138,0.0577113368,0.2142496556,-0.1043527424,0.11758966,-0.2650551498,0.0452635139,0.3239228129,0.1826907694,-0.0352026895,-0.1126695797,-0.417783916,-0.147210747,0.2840802968,-0.2580549717,-0.245167166,0.0119533483,0.0527974665,-0.091761142,0.2312343568,0.238289997,-0.1830744445,0.6743611097,-0.0939318314,-0.2390212417,0.5787590742,-0.5383852124,-0.0357063636,-0.2172711045,-0.193100974,-0.1919911057,0.1511501074,-0.107356593,0.0182099994,0.0467908122,0.4387013316,0.0395708159,-0.0889608786,-0.3561971188,-0.1995178908,0.2851071954,0.2279101908,0.0419763997,0.4570365548,0.2853614986,0.1695158035,-0.0684723854,0.1887974292,-0.0180304926,-0.2879210114,-0.0030238016,0.0400323123,-0.3892459273,0.2894599736,-0.0779185146,0.2788540721,0.2235236019,-0.246622175,0.0826151818,0.3609803319,0.2236855477,-0.4406271875,0.2417325824,-0.1057087854,-0.0480109528,0.0733944699,-0.2125060558,0.2519195378,-0.0479268841,-0.0860952213,0.0504066125,-0.4933129251,-0.0235673264,-0.1894487143,0.0653727204,-0.1262327582,-0.0448605902,-0.1458986551,-0.0153140696,0.1543092728,0.0556869097,-0.2693291008,0.4298380017,0.2029634118,-0.1456071436,-0.0144350883,0.1559263766,-0.1474841088,0.1328604072,-0.0093299672,-0.1915466785,0.2004588246,-0.1141059697,0.0535784177,0.3843781948,0.1535725743,0.0171225872,-0.1076225862,0.0651368126,-0.0502192862,0.2021970153,0.0288014244,0.0270202309,0.4426712394,-0.3593238294,0.0580615848,-0.3788809478,0.3278771341,-0.4405206144,0.3899686038,-0.0393587872,0.0601105914,-0.0102301594,0.1284069866,-0.0726182908,-0.3743269444,-0.2551704347,0.4715960026,-0.0478855297,0.0405350626,0.0953617096,-0.0711544156,-0.0679310337,0.0304413363,0.3493910134,0.2070938945,0.0212720446,0.0449011773,-0.0521538779,-0.0465460382,-0.1306073666,-0.0280886311,0.3493729532,-0.225588873,0.1348799318,0.075287506,0.2183084935,-0.1993987262,-0.1545158476,-0.2003663927,-0.3204384446,0.0115698399,0.1411631107,0.3208896518,-0.1376408786,0.1372115463,0.1250700653,0.4607310593,0.2262209803,0.2846042216,-0.1636532545,0.0487440601,0.0333901569,-0.0979563445,-0.215823099,0.2466624081,-0.3523681462,0.0774699152,-0.0886008814,-0.1845653355,-0.1599787623,0.0074018487,0.2293597162,0.1024663746,-0.0420796461,-0.255674541,0.0428051874,0.2096294463,-0.2503049076,-0.3007627726,0.374556154,-0.058750622,-0.0117408587,-0.1317933202,-0.1412438005,-0.1016193479,-0.2883880734,-0.0360812098,0.2483190298,0.2062901258,0.2738169432,-0.005697303,0.3814315796,0.4019247591,-0.0543019064,-0.1986670345,-0.547450006,0.2148372084,-0.0925638452,-0.2276468575,0.0701643974,-0.0187872667,0.1677819192,0.1912883967,-0.2886455953,-0.673584044,0.1889113188,0.2871622443,-0.2609939575,0.1604726017,0.2804620564,-0.3490063846,-0.0287137385,-0.2814407945,-0.103891097,0.2245209068,0.0848220885,0.2112163156,-0.221072346,0.4646460712,0.0941406488,0.6489886642,0.0296453517,-0.1281051636,0.2985831499,0.1537393332,0.2447576225,-0.0038530349,-0.1338460147,0.0489935353,0.1064159945,-0.2038707584,0.1017027572,-0.0557483211,0.065122366,-0.3501721025,-0.2313645333,0.0471409708,-0.0168843288,-0.0369388424,-0.4160637259,0.3232347369,-0.0301211886,0.2697743475,-0.2884373367,0.1355518103,0.135763377,0.1578714401,0.1288013011,-0.0568930842,-0.0139695946,-0.170657292,-0.2526807189,0.393801868,0.0498392619,0.2946966588,-0.0771579444,0.3258003592,-0.1086524948,0.1709336042,0.6087872386,-0.1312526315,0.0350221097,0.1956077218,-0.2016815692,-0.3555338085,-0.0019054203,-0.3041402996,0.1678823978,0.0728453323,0.7506990433,-0.0244571082,0.0154515868,-0.0228771269,0.3795055747,-0.222904548,-0.386636287,-0.259060055,-0.239826262,0.2020212114,0.0542650297,0.0234459173,0.3708866239,0.1194965094,0.2322513759,-0.0566958673,0.1982782334,0.0937500745,-0.0649097115,0.0116167786,-0.1418979466,-0.1390072256,0.054092776,-0.0561870784,0.2651714385,0.0636063591,-0.0457775891,-0.5454137325,0.1777900755,0.1208341345,0.2388094217,-0.0866866931,-0.2268003374,-0.1198794544,-0.0450631194,0.2277987301,0.1475597471,0.1014586389,0.2041599751,0.3210230768,-0.2983633578,0.0091705713,0.4024381936,0.2755218446,0.1890118122,0.1654900014,-0.2372071296,-0.1963230819,0.0678781047,0.3185476363,0.4146768153,0.0382899977,-0.3658927679,0.524125874,0.2566727698,0.1425125748,-0.2115787119,0.1814755797,-0.1015443057,-0.4452896118,-0.1305099875,-0.2162659913,0.1773560345,-0.1759329587,-0.1006761268,0.5212775469,-0.2374328077,-0.0595539846,0.2230506241,0.0124302059,-0.2171092629,-0.1357540488,-0.173538208,0.094317548,-0.1034941152,0.1680279076,-0.2548354268,-0.0296272207,-0.0258480571,-0.0543110035,0.0665848777,-0.1761960685,-0.3191459775,0.2255784273,0.1246949434,-0.307084173,0.023898663,0.2432399392,0.5752543211,-0.0661970228,0.0134257153,-0.012469504,-0.1434542537,-0.0496817417,0.1397010684,-0.1298236251,0.2827926874,-0.2351443321,0.0005081835,0.106917724,-0.2279693782,-0.3003264666,0.4230240583,0.0332099907,0.1173307002,0.0298740864,0.0688449815,-0.3049264252,-0.1818187386,-0.1941626072,0.1032953188,0.1751461476,-0.4289224446,0.3889814317,-0.0574683212,-0.37502563,0.2138925493,0.1949261874,0.2153660208,-0.2504068017,0.2916764319,-0.0620119981,-0.1055545658,-0.058654286,-0.166161254,0.0028955101,-0.0192075521,0.1990184635,0.018567238,0.352901727,0.0213241875,0.2117160261,0.3017182946,0.0511455461,-0.1592131406,-0.7366719246,-0.2020980716,0.0577510595,-0.0001070497,-0.1427347213,-0.1370648146,0.2167479694,-0.1322854012,0.1760246009,-0.2509111762,-0.0105598839,-0.2350025028,0.1568703502,0.0424210578,-0.148717314,0.2262828201,0.1627236456,0.0156815946,0.2249858826,-0.2116453201,-0.1787355393,0.1025301293,0.1621220708,-0.059812367,-0.0652413741,0.3917291164,0.0371191427,-0.3599075377,0.0558811985,-0.2302165478,0.3071429133,-0.1411307454,-0.1288602799,0.1621950716,0.6288257837,0.090284504,0.0673449486,0.005225874,0.2742750943,-0.2251279801,0.1829318851,-0.2780632973,-0.1428376585,-0.3569362462,-0.1561434567,0.0151984952,-0.1730259657,0.2919380367,-0.2861347795,0.0282903668,-0.2088741958,0.3628519475,0.0899800509,-0.1975496709,0.0473193005,0.0589776151,0.1172167733,-0.0800810456,0.1394037902,0.172870785,0.0388543792,0.0148108676,0.332151413,-0.01107537,0.1098471284,-0.0826835781,-0.0821306184,0.0725666881,0.0972760096,-0.042032022,-0.1037478149,-0.1657619178,0.1477657557,0.4743666351,0.0668724999,0.2182660401,0.0428017601,-0.1522810459,0.3543897271,0.127581805,-0.0173842087,-0.2526034713,-0.0494091697,-0.1836319864,0.0211601369,-0.0532549359,-0.1062147692,-0.2489045709,0.0879066959,-0.4121108651,0.1509920657,-0.2156967521,0.1265384853,-0.0160518922,0.3691642284,-0.3033030033,-0.2289079726,0.0783943161,-0.2160382569,0.2938553691,0.1115008816,-0.2649542689,-0.0027564408,-0.2185145319,0.1251319349,-0.4348650575,0.5451754332,0.2303571552,-0.3177498579,0.1950152814,0.2211612761,0.2615064979,-0.0830582082,0.3307033479,0.4944124222,0.4311096668,-0.0547136292,0.102584891,0.1507220715,-0.1535886228,0.1190046594,0.1592704803,0.1158066168,-0.2425831556,0.5202123523,0.1174396276,-0.2335672528,0.3855023086,-0.2376741767,0.4925028384,-0.389862299,0.0585529953,-0.0203453898,0.0341270305,-0.3185689747,-0.0753840879,-0.155657202,0.1755642593,-0.1999762207,0.0327101126,0.0760909542,-0.078606151,0.0525699221,-0.298938632,0.3373302817,0.3318254352,0.297365725,-0.3848649859,-0.0060853418,-0.4286868274,0.1391412169,0.0209710095,0.0802477226,-0.0116070267,0.0408270136,0.0421428196,0.2475109547,0.4816519022,0.0979122743,0.1235919222,0.2023987174,-0.4470593035,-0.1758026332,-0.1573079526,-0.3478933573,0.0999621078,-0.2129716724,-0.0290640537,-0.3214386106,0.0842481554,0.0336777754,-0.0423010625,0.232040301,0.6045469642,0.2200201899,0.067227751,0.4585684538,0.0525996536,0.1894745231,0.3232182264,-0.0218697302,-0.2543055415,0.0623143725,-0.0039302143,0.2870970368,-0.0663463175,-0.287137717,-0.1954174787,0.6739629507,0.0616975725,-0.5848380923,-0.0630172789,0.1214392781,0.0211719479,-0.1331175119,0.4323017597,0.066798456,0.1964479834,0.2162305713,-0.2929868996,0.0216373373,0.4165513813,-0.1725697666,-0.382404089,-0.0623855852,0.1373431087,0.1988154203,-0.2434106916,-0.4655195475,0.0824157745,0.4198485613,-0.1423687786,-0.1497416049,0.1252377182,-0.3407678008,0.2748536766,0.2406638563,0.0039691809,-0.2093468755,-0.2314013988,-0.0217551906,-0.1318709552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2363","title":"Trying to use metric.compute but get OSError","comments":"I tried this code on a colab notebook and it worked fine (with gpu enabled):\r\n```\r\nfrom datasets import load_metric\r\nmetric = load_metric('accuracy')\r\noutput = metric.add_batch(predictions=[1, 1, 1, 1], references=[1, 1, 0, 0])\r\nfinal_score = metric.compute()\r\nprint(final_score) # 0.5\r\n```\r\nAlso, in `load_metric`, I saw `cache_dir` is optional and it defaults to `~\/.datasets\/`","body":"I want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n","comment_length":53,"text":"Trying to use metric.compute but get OSError\nI want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n\nI tried this code on a colab notebook and it worked fine (with gpu enabled):\r\n```\r\nfrom datasets import load_metric\r\nmetric = load_metric('accuracy')\r\noutput = metric.add_batch(predictions=[1, 1, 1, 1], references=[1, 1, 0, 0])\r\nfinal_score = metric.compute()\r\nprint(final_score) # 0.5\r\n```\r\nAlso, in `load_metric`, I saw `cache_dir` is optional and it defaults to `~\/.datasets\/`","embeddings":[-0.3944184184,-0.2529711425,-0.0330115929,0.3631941974,0.253131628,-0.1011049151,0.2328529209,0.139267087,0.2106887102,0.6745797396,-0.0271198936,0.0674137399,-0.0561968796,0.0007951714,-0.1410534978,-0.1666642725,-0.1239938736,0.0129452525,0.0000074239,0.0780162588,-0.3373332918,0.0424939394,-0.2203928828,0.0718222335,-0.0541991033,-0.2079253197,0.206342876,0.042613849,-0.5113474727,-0.2387506664,0.2351529449,-0.3208845258,0.2170303017,0.8475168347,-0.0001149249,0.1326398402,0.0087183854,-0.1004110426,-0.0945497677,-0.276073128,0.2104712427,-0.2102968097,0.0614107549,-0.2894953787,0.1416248083,-0.025001768,-0.2049153298,-0.3347912133,0.2580504715,0.5588136911,0.1094384342,0.4805767834,0.0571895204,-0.1703307778,-0.374795258,-0.344244957,0.0392587148,0.670753479,-0.0942461342,-0.1250064075,-0.0927235708,-0.0397556424,0.2341285497,0.0877322108,0.206516698,-0.1326481998,0.8326205611,-0.2251918614,-0.1489875764,-0.0572964847,0.1004799753,0.0542257167,-0.3662271798,0.3212551177,-0.0640987307,-0.7123773694,-0.0083443522,-0.0554771312,-0.329859823,-0.1663537025,-0.3520748615,-0.0161977299,-0.4670949578,0.0650909245,-0.2273944765,0.2691727877,-0.1618948281,0.0761636347,0.314235419,0.3356452584,0.0720360652,-0.0285175629,0.1678747386,0.2304376811,-0.5341535807,-0.0079725357,-0.1267386079,-0.0935950652,-0.2341202796,-0.2478045374,0.3044139445,0.0229335669,0.2392277718,0.27803123,-0.1958615184,0.8154187799,0.0828209221,0.1414985061,0.0289000422,0.3671554923,-0.0761498585,-0.0763362721,-0.0666468665,-0.3928147554,0.3305636346,0.1671797335,-0.1789496094,-0.0470937341,-0.5103244781,-0.0825716555,-0.3074188232,-0.1921125054,0.3589552343,0.0782146156,-0.1096397415,0.1300759017,0.2878340185,0.190229252,-0.2090027928,-0.1163052544,-0.2376890033,0.325322479,-0.2651518881,0.1427433044,-0.2088134736,-0.2825702131,0.2108990848,-0.0339325406,0.2929007113,-0.372744292,0.1489905268,-0.1023257896,-0.0621559471,0.0755864009,0.1042392701,-0.2266150713,0.4715836942,-0.188901484,0.1012283415,0.0352442786,-0.6336891055,-0.4329087138,0.0577113368,0.2142496556,-0.1043527424,0.11758966,-0.2650551498,0.0452635139,0.3239228129,0.1826907694,-0.0352026895,-0.1126695797,-0.417783916,-0.147210747,0.2840802968,-0.2580549717,-0.245167166,0.0119533483,0.0527974665,-0.091761142,0.2312343568,0.238289997,-0.1830744445,0.6743611097,-0.0939318314,-0.2390212417,0.5787590742,-0.5383852124,-0.0357063636,-0.2172711045,-0.193100974,-0.1919911057,0.1511501074,-0.107356593,0.0182099994,0.0467908122,0.4387013316,0.0395708159,-0.0889608786,-0.3561971188,-0.1995178908,0.2851071954,0.2279101908,0.0419763997,0.4570365548,0.2853614986,0.1695158035,-0.0684723854,0.1887974292,-0.0180304926,-0.2879210114,-0.0030238016,0.0400323123,-0.3892459273,0.2894599736,-0.0779185146,0.2788540721,0.2235236019,-0.246622175,0.0826151818,0.3609803319,0.2236855477,-0.4406271875,0.2417325824,-0.1057087854,-0.0480109528,0.0733944699,-0.2125060558,0.2519195378,-0.0479268841,-0.0860952213,0.0504066125,-0.4933129251,-0.0235673264,-0.1894487143,0.0653727204,-0.1262327582,-0.0448605902,-0.1458986551,-0.0153140696,0.1543092728,0.0556869097,-0.2693291008,0.4298380017,0.2029634118,-0.1456071436,-0.0144350883,0.1559263766,-0.1474841088,0.1328604072,-0.0093299672,-0.1915466785,0.2004588246,-0.1141059697,0.0535784177,0.3843781948,0.1535725743,0.0171225872,-0.1076225862,0.0651368126,-0.0502192862,0.2021970153,0.0288014244,0.0270202309,0.4426712394,-0.3593238294,0.0580615848,-0.3788809478,0.3278771341,-0.4405206144,0.3899686038,-0.0393587872,0.0601105914,-0.0102301594,0.1284069866,-0.0726182908,-0.3743269444,-0.2551704347,0.4715960026,-0.0478855297,0.0405350626,0.0953617096,-0.0711544156,-0.0679310337,0.0304413363,0.3493910134,0.2070938945,0.0212720446,0.0449011773,-0.0521538779,-0.0465460382,-0.1306073666,-0.0280886311,0.3493729532,-0.225588873,0.1348799318,0.075287506,0.2183084935,-0.1993987262,-0.1545158476,-0.2003663927,-0.3204384446,0.0115698399,0.1411631107,0.3208896518,-0.1376408786,0.1372115463,0.1250700653,0.4607310593,0.2262209803,0.2846042216,-0.1636532545,0.0487440601,0.0333901569,-0.0979563445,-0.215823099,0.2466624081,-0.3523681462,0.0774699152,-0.0886008814,-0.1845653355,-0.1599787623,0.0074018487,0.2293597162,0.1024663746,-0.0420796461,-0.255674541,0.0428051874,0.2096294463,-0.2503049076,-0.3007627726,0.374556154,-0.058750622,-0.0117408587,-0.1317933202,-0.1412438005,-0.1016193479,-0.2883880734,-0.0360812098,0.2483190298,0.2062901258,0.2738169432,-0.005697303,0.3814315796,0.4019247591,-0.0543019064,-0.1986670345,-0.547450006,0.2148372084,-0.0925638452,-0.2276468575,0.0701643974,-0.0187872667,0.1677819192,0.1912883967,-0.2886455953,-0.673584044,0.1889113188,0.2871622443,-0.2609939575,0.1604726017,0.2804620564,-0.3490063846,-0.0287137385,-0.2814407945,-0.103891097,0.2245209068,0.0848220885,0.2112163156,-0.221072346,0.4646460712,0.0941406488,0.6489886642,0.0296453517,-0.1281051636,0.2985831499,0.1537393332,0.2447576225,-0.0038530349,-0.1338460147,0.0489935353,0.1064159945,-0.2038707584,0.1017027572,-0.0557483211,0.065122366,-0.3501721025,-0.2313645333,0.0471409708,-0.0168843288,-0.0369388424,-0.4160637259,0.3232347369,-0.0301211886,0.2697743475,-0.2884373367,0.1355518103,0.135763377,0.1578714401,0.1288013011,-0.0568930842,-0.0139695946,-0.170657292,-0.2526807189,0.393801868,0.0498392619,0.2946966588,-0.0771579444,0.3258003592,-0.1086524948,0.1709336042,0.6087872386,-0.1312526315,0.0350221097,0.1956077218,-0.2016815692,-0.3555338085,-0.0019054203,-0.3041402996,0.1678823978,0.0728453323,0.7506990433,-0.0244571082,0.0154515868,-0.0228771269,0.3795055747,-0.222904548,-0.386636287,-0.259060055,-0.239826262,0.2020212114,0.0542650297,0.0234459173,0.3708866239,0.1194965094,0.2322513759,-0.0566958673,0.1982782334,0.0937500745,-0.0649097115,0.0116167786,-0.1418979466,-0.1390072256,0.054092776,-0.0561870784,0.2651714385,0.0636063591,-0.0457775891,-0.5454137325,0.1777900755,0.1208341345,0.2388094217,-0.0866866931,-0.2268003374,-0.1198794544,-0.0450631194,0.2277987301,0.1475597471,0.1014586389,0.2041599751,0.3210230768,-0.2983633578,0.0091705713,0.4024381936,0.2755218446,0.1890118122,0.1654900014,-0.2372071296,-0.1963230819,0.0678781047,0.3185476363,0.4146768153,0.0382899977,-0.3658927679,0.524125874,0.2566727698,0.1425125748,-0.2115787119,0.1814755797,-0.1015443057,-0.4452896118,-0.1305099875,-0.2162659913,0.1773560345,-0.1759329587,-0.1006761268,0.5212775469,-0.2374328077,-0.0595539846,0.2230506241,0.0124302059,-0.2171092629,-0.1357540488,-0.173538208,0.094317548,-0.1034941152,0.1680279076,-0.2548354268,-0.0296272207,-0.0258480571,-0.0543110035,0.0665848777,-0.1761960685,-0.3191459775,0.2255784273,0.1246949434,-0.307084173,0.023898663,0.2432399392,0.5752543211,-0.0661970228,0.0134257153,-0.012469504,-0.1434542537,-0.0496817417,0.1397010684,-0.1298236251,0.2827926874,-0.2351443321,0.0005081835,0.106917724,-0.2279693782,-0.3003264666,0.4230240583,0.0332099907,0.1173307002,0.0298740864,0.0688449815,-0.3049264252,-0.1818187386,-0.1941626072,0.1032953188,0.1751461476,-0.4289224446,0.3889814317,-0.0574683212,-0.37502563,0.2138925493,0.1949261874,0.2153660208,-0.2504068017,0.2916764319,-0.0620119981,-0.1055545658,-0.058654286,-0.166161254,0.0028955101,-0.0192075521,0.1990184635,0.018567238,0.352901727,0.0213241875,0.2117160261,0.3017182946,0.0511455461,-0.1592131406,-0.7366719246,-0.2020980716,0.0577510595,-0.0001070497,-0.1427347213,-0.1370648146,0.2167479694,-0.1322854012,0.1760246009,-0.2509111762,-0.0105598839,-0.2350025028,0.1568703502,0.0424210578,-0.148717314,0.2262828201,0.1627236456,0.0156815946,0.2249858826,-0.2116453201,-0.1787355393,0.1025301293,0.1621220708,-0.059812367,-0.0652413741,0.3917291164,0.0371191427,-0.3599075377,0.0558811985,-0.2302165478,0.3071429133,-0.1411307454,-0.1288602799,0.1621950716,0.6288257837,0.090284504,0.0673449486,0.005225874,0.2742750943,-0.2251279801,0.1829318851,-0.2780632973,-0.1428376585,-0.3569362462,-0.1561434567,0.0151984952,-0.1730259657,0.2919380367,-0.2861347795,0.0282903668,-0.2088741958,0.3628519475,0.0899800509,-0.1975496709,0.0473193005,0.0589776151,0.1172167733,-0.0800810456,0.1394037902,0.172870785,0.0388543792,0.0148108676,0.332151413,-0.01107537,0.1098471284,-0.0826835781,-0.0821306184,0.0725666881,0.0972760096,-0.042032022,-0.1037478149,-0.1657619178,0.1477657557,0.4743666351,0.0668724999,0.2182660401,0.0428017601,-0.1522810459,0.3543897271,0.127581805,-0.0173842087,-0.2526034713,-0.0494091697,-0.1836319864,0.0211601369,-0.0532549359,-0.1062147692,-0.2489045709,0.0879066959,-0.4121108651,0.1509920657,-0.2156967521,0.1265384853,-0.0160518922,0.3691642284,-0.3033030033,-0.2289079726,0.0783943161,-0.2160382569,0.2938553691,0.1115008816,-0.2649542689,-0.0027564408,-0.2185145319,0.1251319349,-0.4348650575,0.5451754332,0.2303571552,-0.3177498579,0.1950152814,0.2211612761,0.2615064979,-0.0830582082,0.3307033479,0.4944124222,0.4311096668,-0.0547136292,0.102584891,0.1507220715,-0.1535886228,0.1190046594,0.1592704803,0.1158066168,-0.2425831556,0.5202123523,0.1174396276,-0.2335672528,0.3855023086,-0.2376741767,0.4925028384,-0.389862299,0.0585529953,-0.0203453898,0.0341270305,-0.3185689747,-0.0753840879,-0.155657202,0.1755642593,-0.1999762207,0.0327101126,0.0760909542,-0.078606151,0.0525699221,-0.298938632,0.3373302817,0.3318254352,0.297365725,-0.3848649859,-0.0060853418,-0.4286868274,0.1391412169,0.0209710095,0.0802477226,-0.0116070267,0.0408270136,0.0421428196,0.2475109547,0.4816519022,0.0979122743,0.1235919222,0.2023987174,-0.4470593035,-0.1758026332,-0.1573079526,-0.3478933573,0.0999621078,-0.2129716724,-0.0290640537,-0.3214386106,0.0842481554,0.0336777754,-0.0423010625,0.232040301,0.6045469642,0.2200201899,0.067227751,0.4585684538,0.0525996536,0.1894745231,0.3232182264,-0.0218697302,-0.2543055415,0.0623143725,-0.0039302143,0.2870970368,-0.0663463175,-0.287137717,-0.1954174787,0.6739629507,0.0616975725,-0.5848380923,-0.0630172789,0.1214392781,0.0211719479,-0.1331175119,0.4323017597,0.066798456,0.1964479834,0.2162305713,-0.2929868996,0.0216373373,0.4165513813,-0.1725697666,-0.382404089,-0.0623855852,0.1373431087,0.1988154203,-0.2434106916,-0.4655195475,0.0824157745,0.4198485613,-0.1423687786,-0.1497416049,0.1252377182,-0.3407678008,0.2748536766,0.2406638563,0.0039691809,-0.2093468755,-0.2314013988,-0.0217551906,-0.1318709552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2363","title":"Trying to use metric.compute but get OSError","comments":"Hi ! By default it caches the predictions and references used to compute the metric in `~\/.cache\/huggingface\/datasets\/metrics` (not `~\/.datasets\/`). Let me update the documentation @bhavitvyamalik .\r\n\r\nThe cache is used to store all the predictions and references passed to `add_batch` for example in order to compute the metric later when `compute` is called.\r\n\r\nI think the issue might come from the cache directory that is used by default. Can you check that you have the right permissions ? Otherwise feel free to set `cache_dir` to another location.","body":"I want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n","comment_length":87,"text":"Trying to use metric.compute but get OSError\nI want to use metric.compute from load_metric('accuracy') to get training accuracy, but receive OSError. I am wondering what is the mechanism behind the metric calculation, why would it report an OSError?\r\n\r\n```python\r\n195 for epoch in range(num_train_epochs):\r\n196 model.train()\r\n197 for step, batch in enumerate(train_loader):\r\n198 # print(batch['input_ids'].shape)\r\n199 outputs = model(**batch)\r\n200\r\n201 loss = outputs.loss\r\n202 loss \/= gradient_accumulation_steps\r\n203 accelerator.backward(loss)\r\n204\r\n205 predictions = outputs.logits.argmax(dim=-1)\r\n206 metric.add_batch(\r\n207 predictions=accelerator.gather(predictions),\r\n208 references=accelerator.gather(batch['labels'])\r\n209 )\r\n210 progress_bar.set_postfix({'loss': loss.item(), 'train batch acc.': train_metrics})\r\n211\r\n212 if (step + 1) % 50 == 0 or step == len(train_loader) - 1:\r\n213 train_metrics = metric.compute()\r\n```\r\n\r\nthe error message is as below:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"run_multi.py\", line 273, in \r\n main()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"run_multi.py\", line 213, in main\r\n train_metrics = metric.compute()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 391, in compute\r\n self._finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/metric.py\", line 342, in _finalize\r\n self.writer.finalize()\r\n File \"\/home\/yshuang\/.local\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 370, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n```\r\n\r\n## Environment info\r\n\r\n- `datasets` version: 1.6.1\r\n- Platform: Linux NAME=\"Ubuntu\" VERSION=\"20.04.1 LTS (Focal Fossa)\"\r\n- Python version: python3.8.5\r\n- PyArrow version: 4.0.0\r\n\nHi ! By default it caches the predictions and references used to compute the metric in `~\/.cache\/huggingface\/datasets\/metrics` (not `~\/.datasets\/`). Let me update the documentation @bhavitvyamalik .\r\n\r\nThe cache is used to store all the predictions and references passed to `add_batch` for example in order to compute the metric later when `compute` is called.\r\n\r\nI think the issue might come from the cache directory that is used by default. Can you check that you have the right permissions ? Otherwise feel free to set `cache_dir` to another location.","embeddings":[-0.3944184184,-0.2529711425,-0.0330115929,0.3631941974,0.253131628,-0.1011049151,0.2328529209,0.139267087,0.2106887102,0.6745797396,-0.0271198936,0.0674137399,-0.0561968796,0.0007951714,-0.1410534978,-0.1666642725,-0.1239938736,0.0129452525,0.0000074239,0.0780162588,-0.3373332918,0.0424939394,-0.2203928828,0.0718222335,-0.0541991033,-0.2079253197,0.206342876,0.042613849,-0.5113474727,-0.2387506664,0.2351529449,-0.3208845258,0.2170303017,0.8475168347,-0.0001149249,0.1326398402,0.0087183854,-0.1004110426,-0.0945497677,-0.276073128,0.2104712427,-0.2102968097,0.0614107549,-0.2894953787,0.1416248083,-0.025001768,-0.2049153298,-0.3347912133,0.2580504715,0.5588136911,0.1094384342,0.4805767834,0.0571895204,-0.1703307778,-0.374795258,-0.344244957,0.0392587148,0.670753479,-0.0942461342,-0.1250064075,-0.0927235708,-0.0397556424,0.2341285497,0.0877322108,0.206516698,-0.1326481998,0.8326205611,-0.2251918614,-0.1489875764,-0.0572964847,0.1004799753,0.0542257167,-0.3662271798,0.3212551177,-0.0640987307,-0.7123773694,-0.0083443522,-0.0554771312,-0.329859823,-0.1663537025,-0.3520748615,-0.0161977299,-0.4670949578,0.0650909245,-0.2273944765,0.2691727877,-0.1618948281,0.0761636347,0.314235419,0.3356452584,0.0720360652,-0.0285175629,0.1678747386,0.2304376811,-0.5341535807,-0.0079725357,-0.1267386079,-0.0935950652,-0.2341202796,-0.2478045374,0.3044139445,0.0229335669,0.2392277718,0.27803123,-0.1958615184,0.8154187799,0.0828209221,0.1414985061,0.0289000422,0.3671554923,-0.0761498585,-0.0763362721,-0.0666468665,-0.3928147554,0.3305636346,0.1671797335,-0.1789496094,-0.0470937341,-0.5103244781,-0.0825716555,-0.3074188232,-0.1921125054,0.3589552343,0.0782146156,-0.1096397415,0.1300759017,0.2878340185,0.190229252,-0.2090027928,-0.1163052544,-0.2376890033,0.325322479,-0.2651518881,0.1427433044,-0.2088134736,-0.2825702131,0.2108990848,-0.0339325406,0.2929007113,-0.372744292,0.1489905268,-0.1023257896,-0.0621559471,0.0755864009,0.1042392701,-0.2266150713,0.4715836942,-0.188901484,0.1012283415,0.0352442786,-0.6336891055,-0.4329087138,0.0577113368,0.2142496556,-0.1043527424,0.11758966,-0.2650551498,0.0452635139,0.3239228129,0.1826907694,-0.0352026895,-0.1126695797,-0.417783916,-0.147210747,0.2840802968,-0.2580549717,-0.245167166,0.0119533483,0.0527974665,-0.091761142,0.2312343568,0.238289997,-0.1830744445,0.6743611097,-0.0939318314,-0.2390212417,0.5787590742,-0.5383852124,-0.0357063636,-0.2172711045,-0.193100974,-0.1919911057,0.1511501074,-0.107356593,0.0182099994,0.0467908122,0.4387013316,0.0395708159,-0.0889608786,-0.3561971188,-0.1995178908,0.2851071954,0.2279101908,0.0419763997,0.4570365548,0.2853614986,0.1695158035,-0.0684723854,0.1887974292,-0.0180304926,-0.2879210114,-0.0030238016,0.0400323123,-0.3892459273,0.2894599736,-0.0779185146,0.2788540721,0.2235236019,-0.246622175,0.0826151818,0.3609803319,0.2236855477,-0.4406271875,0.2417325824,-0.1057087854,-0.0480109528,0.0733944699,-0.2125060558,0.2519195378,-0.0479268841,-0.0860952213,0.0504066125,-0.4933129251,-0.0235673264,-0.1894487143,0.0653727204,-0.1262327582,-0.0448605902,-0.1458986551,-0.0153140696,0.1543092728,0.0556869097,-0.2693291008,0.4298380017,0.2029634118,-0.1456071436,-0.0144350883,0.1559263766,-0.1474841088,0.1328604072,-0.0093299672,-0.1915466785,0.2004588246,-0.1141059697,0.0535784177,0.3843781948,0.1535725743,0.0171225872,-0.1076225862,0.0651368126,-0.0502192862,0.2021970153,0.0288014244,0.0270202309,0.4426712394,-0.3593238294,0.0580615848,-0.3788809478,0.3278771341,-0.4405206144,0.3899686038,-0.0393587872,0.0601105914,-0.0102301594,0.1284069866,-0.0726182908,-0.3743269444,-0.2551704347,0.4715960026,-0.0478855297,0.0405350626,0.0953617096,-0.0711544156,-0.0679310337,0.0304413363,0.3493910134,0.2070938945,0.0212720446,0.0449011773,-0.0521538779,-0.0465460382,-0.1306073666,-0.0280886311,0.3493729532,-0.225588873,0.1348799318,0.075287506,0.2183084935,-0.1993987262,-0.1545158476,-0.2003663927,-0.3204384446,0.0115698399,0.1411631107,0.3208896518,-0.1376408786,0.1372115463,0.1250700653,0.4607310593,0.2262209803,0.2846042216,-0.1636532545,0.0487440601,0.0333901569,-0.0979563445,-0.215823099,0.2466624081,-0.3523681462,0.0774699152,-0.0886008814,-0.1845653355,-0.1599787623,0.0074018487,0.2293597162,0.1024663746,-0.0420796461,-0.255674541,0.0428051874,0.2096294463,-0.2503049076,-0.3007627726,0.374556154,-0.058750622,-0.0117408587,-0.1317933202,-0.1412438005,-0.1016193479,-0.2883880734,-0.0360812098,0.2483190298,0.2062901258,0.2738169432,-0.005697303,0.3814315796,0.4019247591,-0.0543019064,-0.1986670345,-0.547450006,0.2148372084,-0.0925638452,-0.2276468575,0.0701643974,-0.0187872667,0.1677819192,0.1912883967,-0.2886455953,-0.673584044,0.1889113188,0.2871622443,-0.2609939575,0.1604726017,0.2804620564,-0.3490063846,-0.0287137385,-0.2814407945,-0.103891097,0.2245209068,0.0848220885,0.2112163156,-0.221072346,0.4646460712,0.0941406488,0.6489886642,0.0296453517,-0.1281051636,0.2985831499,0.1537393332,0.2447576225,-0.0038530349,-0.1338460147,0.0489935353,0.1064159945,-0.2038707584,0.1017027572,-0.0557483211,0.065122366,-0.3501721025,-0.2313645333,0.0471409708,-0.0168843288,-0.0369388424,-0.4160637259,0.3232347369,-0.0301211886,0.2697743475,-0.2884373367,0.1355518103,0.135763377,0.1578714401,0.1288013011,-0.0568930842,-0.0139695946,-0.170657292,-0.2526807189,0.393801868,0.0498392619,0.2946966588,-0.0771579444,0.3258003592,-0.1086524948,0.1709336042,0.6087872386,-0.1312526315,0.0350221097,0.1956077218,-0.2016815692,-0.3555338085,-0.0019054203,-0.3041402996,0.1678823978,0.0728453323,0.7506990433,-0.0244571082,0.0154515868,-0.0228771269,0.3795055747,-0.222904548,-0.386636287,-0.259060055,-0.239826262,0.2020212114,0.0542650297,0.0234459173,0.3708866239,0.1194965094,0.2322513759,-0.0566958673,0.1982782334,0.0937500745,-0.0649097115,0.0116167786,-0.1418979466,-0.1390072256,0.054092776,-0.0561870784,0.2651714385,0.0636063591,-0.0457775891,-0.5454137325,0.1777900755,0.1208341345,0.2388094217,-0.0866866931,-0.2268003374,-0.1198794544,-0.0450631194,0.2277987301,0.1475597471,0.1014586389,0.2041599751,0.3210230768,-0.2983633578,0.0091705713,0.4024381936,0.2755218446,0.1890118122,0.1654900014,-0.2372071296,-0.1963230819,0.0678781047,0.3185476363,0.4146768153,0.0382899977,-0.3658927679,0.524125874,0.2566727698,0.1425125748,-0.2115787119,0.1814755797,-0.1015443057,-0.4452896118,-0.1305099875,-0.2162659913,0.1773560345,-0.1759329587,-0.1006761268,0.5212775469,-0.2374328077,-0.0595539846,0.2230506241,0.0124302059,-0.2171092629,-0.1357540488,-0.173538208,0.094317548,-0.1034941152,0.1680279076,-0.2548354268,-0.0296272207,-0.0258480571,-0.0543110035,0.0665848777,-0.1761960685,-0.3191459775,0.2255784273,0.1246949434,-0.307084173,0.023898663,0.2432399392,0.5752543211,-0.0661970228,0.0134257153,-0.012469504,-0.1434542537,-0.0496817417,0.1397010684,-0.1298236251,0.2827926874,-0.2351443321,0.0005081835,0.106917724,-0.2279693782,-0.3003264666,0.4230240583,0.0332099907,0.1173307002,0.0298740864,0.0688449815,-0.3049264252,-0.1818187386,-0.1941626072,0.1032953188,0.1751461476,-0.4289224446,0.3889814317,-0.0574683212,-0.37502563,0.2138925493,0.1949261874,0.2153660208,-0.2504068017,0.2916764319,-0.0620119981,-0.1055545658,-0.058654286,-0.166161254,0.0028955101,-0.0192075521,0.1990184635,0.018567238,0.352901727,0.0213241875,0.2117160261,0.3017182946,0.0511455461,-0.1592131406,-0.7366719246,-0.2020980716,0.0577510595,-0.0001070497,-0.1427347213,-0.1370648146,0.2167479694,-0.1322854012,0.1760246009,-0.2509111762,-0.0105598839,-0.2350025028,0.1568703502,0.0424210578,-0.148717314,0.2262828201,0.1627236456,0.0156815946,0.2249858826,-0.2116453201,-0.1787355393,0.1025301293,0.1621220708,-0.059812367,-0.0652413741,0.3917291164,0.0371191427,-0.3599075377,0.0558811985,-0.2302165478,0.3071429133,-0.1411307454,-0.1288602799,0.1621950716,0.6288257837,0.090284504,0.0673449486,0.005225874,0.2742750943,-0.2251279801,0.1829318851,-0.2780632973,-0.1428376585,-0.3569362462,-0.1561434567,0.0151984952,-0.1730259657,0.2919380367,-0.2861347795,0.0282903668,-0.2088741958,0.3628519475,0.0899800509,-0.1975496709,0.0473193005,0.0589776151,0.1172167733,-0.0800810456,0.1394037902,0.172870785,0.0388543792,0.0148108676,0.332151413,-0.01107537,0.1098471284,-0.0826835781,-0.0821306184,0.0725666881,0.0972760096,-0.042032022,-0.1037478149,-0.1657619178,0.1477657557,0.4743666351,0.0668724999,0.2182660401,0.0428017601,-0.1522810459,0.3543897271,0.127581805,-0.0173842087,-0.2526034713,-0.0494091697,-0.1836319864,0.0211601369,-0.0532549359,-0.1062147692,-0.2489045709,0.0879066959,-0.4121108651,0.1509920657,-0.2156967521,0.1265384853,-0.0160518922,0.3691642284,-0.3033030033,-0.2289079726,0.0783943161,-0.2160382569,0.2938553691,0.1115008816,-0.2649542689,-0.0027564408,-0.2185145319,0.1251319349,-0.4348650575,0.5451754332,0.2303571552,-0.3177498579,0.1950152814,0.2211612761,0.2615064979,-0.0830582082,0.3307033479,0.4944124222,0.4311096668,-0.0547136292,0.102584891,0.1507220715,-0.1535886228,0.1190046594,0.1592704803,0.1158066168,-0.2425831556,0.5202123523,0.1174396276,-0.2335672528,0.3855023086,-0.2376741767,0.4925028384,-0.389862299,0.0585529953,-0.0203453898,0.0341270305,-0.3185689747,-0.0753840879,-0.155657202,0.1755642593,-0.1999762207,0.0327101126,0.0760909542,-0.078606151,0.0525699221,-0.298938632,0.3373302817,0.3318254352,0.297365725,-0.3848649859,-0.0060853418,-0.4286868274,0.1391412169,0.0209710095,0.0802477226,-0.0116070267,0.0408270136,0.0421428196,0.2475109547,0.4816519022,0.0979122743,0.1235919222,0.2023987174,-0.4470593035,-0.1758026332,-0.1573079526,-0.3478933573,0.0999621078,-0.2129716724,-0.0290640537,-0.3214386106,0.0842481554,0.0336777754,-0.0423010625,0.232040301,0.6045469642,0.2200201899,0.067227751,0.4585684538,0.0525996536,0.1894745231,0.3232182264,-0.0218697302,-0.2543055415,0.0623143725,-0.0039302143,0.2870970368,-0.0663463175,-0.287137717,-0.1954174787,0.6739629507,0.0616975725,-0.5848380923,-0.0630172789,0.1214392781,0.0211719479,-0.1331175119,0.4323017597,0.066798456,0.1964479834,0.2162305713,-0.2929868996,0.0216373373,0.4165513813,-0.1725697666,-0.382404089,-0.0623855852,0.1373431087,0.1988154203,-0.2434106916,-0.4655195475,0.0824157745,0.4198485613,-0.1423687786,-0.1497416049,0.1252377182,-0.3407678008,0.2748536766,0.2406638563,0.0039691809,-0.2093468755,-0.2314013988,-0.0217551906,-0.1318709552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2356","title":"How to Add New Metrics Guide","comments":"Hi ! sorry for the late response \r\n\r\nIt would be fantastic to have a guide for adding metrics as well ! Currently we only have this template here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/templates\/new_metric_script.py\r\n\r\nWe can also include test utilities for metrics in the guide.\r\n\r\nWe have a pytest suite with commands that you can use to make sure your metric works as expected.\r\nIt has two useful commands:\r\n\r\n1. This commands tests the code in the `Examples:` desction of the docstring of the metric:\r\n```\r\npytest tests\/test_metric_common.py::LocalMetricTest::test_load_metric_\r\n```\r\nThis will run this code for example:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e0787aa2a781cc15a80f7597f56d1f12e23df4c9\/metrics\/accuracy\/accuracy.py#L40-L45\r\n\r\nMoreover this test is meant to be fast so users are free to add patches to the metric to avoid intensive computations.\r\nAnd example of intensive call patch can be found here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e0787aa2a781cc15a80f7597f56d1f12e23df4c9\/tests\/test_metric_common.py#L138-L151\r\n\r\n2. This test runs the same thing as 1. except that it doesn't use patches (the real metric is used):\r\n```\r\nRUN_SLOW=1 pytest tests\/test_metric_common.py::LocalMetricTest::test_load_metric_\r\n```\r\n\r\nFinally additional metric-specific tests can be added to `test_metric_common.py`.\r\n\r\nVoila :) Feel free to ping me if you have any question or if I can help\r\n","body":"**Is your feature request related to a problem? Please describe.**\r\nCurrently there is an absolutely fantastic guide for how to contribute a new dataset to the library. However, there isn't one for adding new metrics.\r\n\r\n**Describe the solution you'd like**\r\nI'd like for a guide in a similar style to the dataset guide for adding metrics. I believe many of the content in the dataset guide such as setup can be easily copied over with minimal changes. Also, from what I've seen with existing metrics, it shouldn't be as complicated, especially in documentation of the metric, mainly just citation and usage. The most complicated part I see would be in automated tests that run the new metrics, but y'all's test suite seem pretty comprehensive, so it might not be that hard.\r\n\r\n**Describe alternatives you've considered**\r\nOne alternative would be just not having the metrics be community generated and so would not need a step by step guide. New metrics would just be proposed as issues and the internal team would take care of them. However, I think it makes more sense to have a step by step guide for contributors to follow.\r\n\r\n**Additional context**\r\nI'd be happy to help with creating this guide as I am very interested in adding software engineering metrics to the library :nerd_face:, the part I would need guidance on would be testing.\r\n\r\nP.S. Love the library and community y'all have built! :hugs: \r\n","comment_length":176,"text":"How to Add New Metrics Guide\n**Is your feature request related to a problem? Please describe.**\r\nCurrently there is an absolutely fantastic guide for how to contribute a new dataset to the library. However, there isn't one for adding new metrics.\r\n\r\n**Describe the solution you'd like**\r\nI'd like for a guide in a similar style to the dataset guide for adding metrics. I believe many of the content in the dataset guide such as setup can be easily copied over with minimal changes. Also, from what I've seen with existing metrics, it shouldn't be as complicated, especially in documentation of the metric, mainly just citation and usage. The most complicated part I see would be in automated tests that run the new metrics, but y'all's test suite seem pretty comprehensive, so it might not be that hard.\r\n\r\n**Describe alternatives you've considered**\r\nOne alternative would be just not having the metrics be community generated and so would not need a step by step guide. New metrics would just be proposed as issues and the internal team would take care of them. However, I think it makes more sense to have a step by step guide for contributors to follow.\r\n\r\n**Additional context**\r\nI'd be happy to help with creating this guide as I am very interested in adding software engineering metrics to the library :nerd_face:, the part I would need guidance on would be testing.\r\n\r\nP.S. Love the library and community y'all have built! :hugs: \r\n\nHi ! sorry for the late response \r\n\r\nIt would be fantastic to have a guide for adding metrics as well ! Currently we only have this template here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/templates\/new_metric_script.py\r\n\r\nWe can also include test utilities for metrics in the guide.\r\n\r\nWe have a pytest suite with commands that you can use to make sure your metric works as expected.\r\nIt has two useful commands:\r\n\r\n1. This commands tests the code in the `Examples:` desction of the docstring of the metric:\r\n```\r\npytest tests\/test_metric_common.py::LocalMetricTest::test_load_metric_\r\n```\r\nThis will run this code for example:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e0787aa2a781cc15a80f7597f56d1f12e23df4c9\/metrics\/accuracy\/accuracy.py#L40-L45\r\n\r\nMoreover this test is meant to be fast so users are free to add patches to the metric to avoid intensive computations.\r\nAnd example of intensive call patch can be found here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e0787aa2a781cc15a80f7597f56d1f12e23df4c9\/tests\/test_metric_common.py#L138-L151\r\n\r\n2. This test runs the same thing as 1. except that it doesn't use patches (the real metric is used):\r\n```\r\nRUN_SLOW=1 pytest tests\/test_metric_common.py::LocalMetricTest::test_load_metric_\r\n```\r\n\r\nFinally additional metric-specific tests can be added to `test_metric_common.py`.\r\n\r\nVoila :) Feel free to ping me if you have any question or if I can help\r\n","embeddings":[0.0184292234,0.009165708,-0.0002697915,-0.1558319926,0.0394446068,0.0882947743,-0.0663276687,0.0305395015,0.0264215395,0.0817607865,0.0160229094,0.2692242265,-0.1336126328,0.2791265249,0.252602607,-0.1617622674,-0.1102725938,-0.0691408813,0.1617088318,0.0204251409,-0.002773423,-0.021874046,0.329408586,0.0545105003,-0.1164637133,-0.0480870865,-0.1330723464,0.0665191337,-0.4833083153,-0.6584537029,0.1586213112,0.1352889538,-0.1188627034,0.4845243394,-0.0001103193,-0.3142381608,0.2092059851,-0.0670046434,-0.1198505983,-0.441141665,0.0299328081,-0.7234885097,0.1960804909,-0.0932002142,0.0312845409,-0.1347152293,-0.1554330438,0.1224784926,-0.0282175038,0.1885768771,0.213010639,0.0976783112,0.0443958864,-0.3083938956,-0.260330379,0.3114709854,-0.4239771366,0.5435327291,0.3079588115,0.1700689942,0.0951610059,-0.0022227019,0.0336410254,-0.0354855694,0.6798270941,-0.0897223577,0.0804402381,0.1071784347,-0.3254208863,0.3772416711,0.4795588553,-0.3143450618,-0.3093032837,-0.283180058,0.1683874726,-0.4429858625,-0.2639279366,-0.1353823096,-0.0871687531,-0.2493702322,-0.266956389,-0.4700167477,-0.2433303297,0.1533593982,0.1787814051,0.0073762131,0.0057962895,-0.1015962586,0.2997991443,0.1384744048,-0.4045351446,0.0882186815,-0.0804798678,0.3847275078,0.0626204833,-0.4114014506,0.2671383619,0.0210768189,0.4693697691,0.1305872798,0.3220776618,0.0416282825,-0.3315683007,0.1019887552,0.069723919,0.0302171819,0.2064802796,0.0259620585,0.0907073021,0.3396171331,0.4559905827,-0.2520832121,0.3968576193,-0.0956557021,0.0499762073,0.0685843676,0.1471230835,-0.3993322253,-0.0377423763,-0.3121259809,0.432828635,-0.4076386392,0.4414565265,0.0897256657,-0.0520035923,-0.2915807068,-0.3402695656,0.3253027797,-0.1487537324,0.2075919509,-0.0739264563,0.1344438344,-0.3010628521,0.5152187943,0.1827646196,-0.2941655815,0.2420096695,0.1666758358,0.1522644162,-0.199954167,0.0022773335,0.2404260784,0.0478422493,-0.0570169687,-0.123834759,0.1695913672,-0.1199385375,-0.2851220071,-0.3717520535,-0.0453540944,-0.3488799632,-0.4204467237,-0.2678246498,0.1413258314,-0.3109639585,-0.3429507613,-0.1060872748,0.4933302999,-0.3894296587,-0.1969149858,0.0484503955,0.4269002676,-0.2820628285,-0.3100942969,0.0637532398,0.357925266,-0.3396798968,-0.2324168235,-0.2341893762,0.0253998954,-0.1374936253,-0.0389815457,-0.0297508687,0.1589485407,0.0738943592,-0.1113869697,0.5824443102,-0.2802065611,0.2215571702,0.0327623673,-0.085318692,-0.2565297186,-0.1304361671,0.4572606385,0.0265267976,-0.2065899521,-0.2832887471,0.0571033023,-0.1942892671,-0.2472899556,-0.1541683972,-0.1237749606,0.2240421027,-0.1152999178,-0.0695195496,-0.0026222249,0.4326041341,-0.0192571636,-0.1784527302,-0.102321595,-0.2335809767,0.2795448005,0.52559793,-0.0219726469,0.2193109542,-0.2187476605,-0.1563498527,-0.0179751441,0.1284736246,-0.1166479439,0.3591010869,-0.1786890775,-0.4749299586,-0.0524333864,0.0522365682,0.066976361,0.0221097227,-0.3339243829,-0.042014949,0.0546156801,-0.1838302016,-0.0239482522,-0.3259169459,0.443764776,-0.0501364954,0.0161871444,0.0387043133,0.0317185409,0.0827146769,0.5306402445,0.2860185206,0.0981434435,-0.0662294701,0.1483388841,0.0292118378,0.0674835891,0.2121868581,0.6915304065,0.546990037,-0.0805040002,0.203546226,-0.2614835501,-0.0956279859,-0.1592144519,-0.2688032389,0.2466148138,-0.3072992265,0.1097352952,-0.1203326136,0.0513707474,-0.0616922006,-0.0289074238,-0.1241117194,-0.1857165992,0.1790335476,-0.0951611921,-0.0663760304,-0.0675027668,0.2252746373,0.2073013633,0.1898086518,-0.1084432602,0.0301645417,0.3718456924,0.1367536336,-0.1395901293,-0.1679075956,0.0820075125,0.1165526733,0.2425193489,0.0056011179,0.0916838497,-0.3783510327,-0.1168581396,-0.089568764,0.1801922768,-0.3604492545,-0.0228362791,0.0145989731,0.1239486039,0.0442001298,-0.0018293137,-0.1631917953,0.1606388688,-0.1625336409,-0.1156409532,0.0826383904,0.0563561097,-0.0095752887,-0.2589283288,-0.3223095238,-0.2484735399,0.2069886923,-0.0176496804,0.1311489791,-0.0120332576,0.1771413535,0.313678056,-0.3538408875,-0.0062791016,-0.1025248691,-0.2769832015,-0.0146990437,0.1558507681,0.0837280005,-0.0107058007,0.542370975,0.0823427513,0.3953062594,-0.4368837178,-0.8153743744,-0.0724791139,-0.1126619577,0.6125946641,0.0842719227,0.0896968171,0.1166645661,-0.2269663066,0.0211907309,-0.1636058241,0.0361229591,0.04617057,-0.3235256672,0.0530880876,-0.1868775785,-0.2285001278,-0.0836725086,-0.3801620603,0.1529833227,0.1903473437,0.0673877522,0.2270582765,0.0454426929,0.0681231618,-0.0045258016,0.2067427188,-0.0990071222,0.0303437915,0.2220388055,-0.3982443213,-0.1208576784,0.252918601,0.0172525626,0.2168854773,-0.1977648586,-0.2195447087,-0.7192444205,0.0633444116,0.3662299514,0.1567400992,0.3119517267,0.1860899627,0.031219136,0.0116891274,-0.0507835075,-0.5496695638,-0.0624050535,0.3355000317,0.2113073468,-0.1976547241,0.2261877358,0.1639116406,0.5142987967,0.1251865923,-0.1498404741,0.2382318825,0.0807622224,0.552067697,0.0971253142,-0.2344640195,0.083427839,0.1130742952,0.1381238103,0.0112292729,0.4134224653,-0.047334034,-0.059586443,-0.176943019,-0.1702030152,0.0639380887,0.0918143988,0.0583691113,0.240531221,-0.2475563288,-0.009952819,-0.5171266198,-0.4916653037,0.1898654103,0.6618602872,0.1130466238,0.1411307901,-0.300295949,-0.2005688995,-0.3676937222,0.0851902291,0.2254043967,-0.3414384723,-0.3552610874,0.1465350538,-0.0313886143,0.1897523254,-0.1730224192,-0.2952411473,-0.5991567969,0.244756937,0.0904881582,-0.0445317999,0.1522836238,0.1154043004,0.3179458082,0.1051432863,0.2786827087,-0.3807907701,-0.1559417844,0.2681200504,0.4588249028,-0.0700059682,-0.044475846,0.1528838724,-0.4423783422,0.0944515988,0.0025821307,-0.3348359466,0.1386279613,0.3370295465,-0.1488547027,0.253372252,0.0784650221,0.2578750849,0.026288148,0.0526403524,0.3170262575,0.0884940773,0.1409422606,-0.2362565249,-0.0629438013,0.0061264173,-0.1569485962,0.1550159454,-0.0296910498,-0.2291724682,0.5229886174,0.0898672715,0.2265159041,0.507243216,-0.1621755809,0.0462648384,-0.3842125237,0.11181885,0.0300918538,0.2098210156,-0.3015302122,-0.3861687481,0.5291603208,0.0627260879,-0.1092025861,0.6401628256,0.2059105784,0.1134096682,0.0134447329,0.1225927323,0.8073385954,-0.1980905235,-0.2586926222,0.0910703093,-0.4162579775,0.4987086654,-0.1907399446,-0.255646348,-0.2205879241,0.0898179263,0.0656922832,-0.0381285809,0.0204967838,-0.2923638821,-0.0121048111,0.2856832147,-0.1299082637,-0.070424445,-0.022637926,0.4557738006,0.0465857498,-0.3004553914,0.2458121479,0.2028039098,-0.1255786419,-0.0508205742,-0.0916766226,-0.1765006781,-0.0268977582,0.0362506546,-0.3564399481,-0.5494884849,0.2038478553,-0.1903162003,0.2237992287,-0.4408729076,0.2794647813,-0.0170745123,0.5803315043,0.0122538786,-0.2243372351,0.1015141308,-0.6351746917,0.0541653968,-0.0711620227,0.0436093025,0.2332556546,-0.1384500265,0.0471984595,0.2041129768,0.0631218255,-0.2692430317,-0.1460914016,-0.1066579446,0.0170494393,-0.3227482736,0.2849548459,0.1865736246,0.1908513308,-0.2029922903,0.1163508296,0.4865889847,-0.0669853911,0.1861115843,0.1797464192,0.0426099524,-0.2171354443,-0.1957912594,-0.0896973833,-0.4844108522,-0.1337010562,-0.1219653934,-0.1105271652,-0.092899017,-0.149239704,0.1604579687,-0.0780040696,-0.1232781261,0.2798514664,0.0764325187,0.1370571256,0.2057883441,0.1991370916,-0.0557145923,-0.1154419333,-0.1009295508,-0.1944975257,0.1754546762,-0.216511637,0.1490339935,-0.5839186907,-0.0183276646,0.0264895223,0.3320849538,-0.3175542355,-0.0264538545,0.1045346335,0.1118404269,0.2753580809,0.029427262,-0.1566524655,-0.1903887242,-0.0623635724,-0.2546935678,-0.1563497484,-0.1091449484,-0.3250694871,0.1531180292,-0.2212014794,0.0897999257,-0.0839981288,0.041127257,0.0471489877,0.1006320864,0.010900734,0.0874092281,-0.1541685611,0.0948510095,-0.2594510317,0.0788015202,0.0737436339,0.0457845666,0.2143075615,0.4922334552,0.0893116668,0.1163202226,-0.1913698614,0.0406657122,0.125886634,0.2075711042,0.1541228741,-0.2356119454,0.0355450511,0.2782092094,0.0968456715,0.2868942916,0.0373785868,-0.1993086487,0.0351724252,0.3389795423,0.0153883407,0.1513997018,0.0328002162,-0.0844230279,0.0770939812,0.1764703393,0.2765749097,0.074194409,0.5186918974,0.2016321719,0.3859261572,-0.1801867634,0.0759760961,-0.0941981003,0.1787389815,-0.1183591783,0.0894717723,-0.0716501325,0.5985187888,0.1002664864,0.0165919214,-0.2216980755,-0.0850746781,0.3150957823,0.0589004457,0.0701206699,0.0637234822,0.0152083607,0.1097335294,-0.147644341,0.3266406059,-0.2443140149,0.1892072409,0.5347024798,-0.0361418277,-0.2718058527,-0.0539482534,-0.1050115004,0.128790915,0.0659514293,-0.2847744226,-0.1338751316,-0.196233958,-0.2545784116,0.2274820805,-0.3350029886,0.1266346127,-0.2393223941,0.1538042277,-0.0168619137,0.0430134013,0.2615385652,0.2364612222,0.0756507218,-0.0086820954,-0.100796409,0.0309826806,-0.0299991015,0.3788925707,0.1901307106,-0.1459893137,0.0140393535,0.4152963758,-0.1666008681,-0.0902854651,0.3696526289,0.1923995465,-0.1116637737,-0.1447280049,-0.147663191,0.1399005651,-0.1301928163,0.3774015307,-0.4194701016,-0.1162066758,-0.3492373526,0.3389995694,0.3243453503,-0.1370758414,-0.1120394021,-0.2174183577,-0.06917537,-0.0427202694,0.5133609176,-0.0530813336,-0.174845621,-0.2077672333,0.1112169623,0.2073675394,-0.1023726612,0.3140542507,0.3296172321,-0.2229017764,-0.1180862635,-0.4054396749,-0.2519232929,0.153862685,0.1275785714,0.021071054,-0.0239094105,-0.0661350489,0.3210529387,0.3211809397,0.002587721,0.1724500954,-0.182783857,-0.3393128812,-0.1350646317,0.0585565828,-0.2547019422,-0.1227249131,0.3190796077,-0.0818207636,-0.4071906507,-0.0169707742,-0.1636160463,0.1709468514,0.2222639322,0.5071320534,0.264021486,0.0503472537,0.4483633041,0.1060717478,-0.0998588949,-0.1260732859,-0.033013247,0.1640865654,0.0844278634,-0.1180986762,0.299447149,0.0042779385,-0.2160635889,-0.2486462891,0.5709847808,0.087832585,-0.1702438891,-0.0867074504,0.051918529,-0.0879796669,-0.1736577749,0.3191544116,0.3881107569,0.076636903,0.4116769731,0.077970311,-0.0692616031,0.2546755373,-0.0881049782,0.1379313022,0.0783845112,0.249388501,0.474599123,-0.0439440273,-0.5338285565,-0.0624817088,0.2976032197,-0.0779588073,-0.1717201024,0.5125023127,-0.1398001611,-0.0533151813,-0.1188735813,-0.0235224236,0.0585027933,-0.1079894677,0.1956657171,-0.0109942816]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2350","title":"`FaissIndex.save` throws error on GPU","comments":"Just in case, this is a workaround that I use in my code and it seems to do the job.\r\n\r\n```python\r\nif use_gpu_index:\r\n data[\"train\"]._indexes[\"text_emb\"].faiss_index = faiss.index_gpu_to_cpu(data[\"train\"]._indexes[\"text_emb\"].faiss_index)\r\n```","body":"## Describe the bug\r\n\r\nAfter training an index with a factory string `OPQ16_128,IVF512,PQ32` on GPU, `.save_faiss_index` throws this error.\r\n\r\n```\r\n File \"index_wikipedia.py\", line 119, in \r\n data[\"train\"].save_faiss_index(\"text_emb\", index_save_path)\r\n File \"\/home\/vlialin\/miniconda3\/envs\/cat\/lib\/python3.8\/site-packages\/datasets\/search.py\", line 470, in save_faiss_index\r\n index.save(file)\r\n File \"\/home\/vlialin\/miniconda3\/envs\/cat\/lib\/python3.8\/site-packages\/datasets\/search.py\", line 334, in save\r\n faiss.write_index(index, str(file))\r\n File \"\/home\/vlialin\/miniconda3\/envs\/cat\/lib\/python3.8\/site-packages\/faiss\/swigfaiss_avx2.py\", line 5654, in write_index\r\n return _swigfaiss.write_index(*args)\r\nRuntimeError: Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at \/root\/miniconda3\/conda-bld\/faiss-pkg_1613235005464\/work\/faiss\/impl\/index_write.cpp:453: don't know how to serialize this type of index\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nAny dataset will do, I just selected a familiar one.\r\n\r\n```python\r\nimport numpy as np\r\nimport datasets\r\nINDEX_STR = \"OPQ16_128,IVF512,PQ32\"\r\nINDEX_SAVE_PATH = \"will_not_save.faiss\"\r\n\r\ndata = datasets.load_dataset(\"Fraser\/news-category-dataset\", split=f\"train[:10000]\")\r\n\r\ndef encode(item):\r\n return {\"text_emb\": np.random.randn(768).astype(np.float32)}\r\n\r\ndata = data.map(encode)\r\n\r\ndata.add_faiss_index(column=\"text_emb\", string_factory=INDEX_STR, train_size=10_000, device=0)\r\ndata.save_faiss_index(\"text_emb\", INDEX_SAVE_PATH)\r\n```\r\n\r\n## Expected results\r\nSaving the index\r\n\r\n## Actual results\r\nError in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) ... don't know how to serialize this type of index\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-4.15.0-142-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\r\n\r\n\r\nI will be proposing a fix in a couple of minutes","comment_length":27,"text":"`FaissIndex.save` throws error on GPU\n## Describe the bug\r\n\r\nAfter training an index with a factory string `OPQ16_128,IVF512,PQ32` on GPU, `.save_faiss_index` throws this error.\r\n\r\n```\r\n File \"index_wikipedia.py\", line 119, in \r\n data[\"train\"].save_faiss_index(\"text_emb\", index_save_path)\r\n File \"\/home\/vlialin\/miniconda3\/envs\/cat\/lib\/python3.8\/site-packages\/datasets\/search.py\", line 470, in save_faiss_index\r\n index.save(file)\r\n File \"\/home\/vlialin\/miniconda3\/envs\/cat\/lib\/python3.8\/site-packages\/datasets\/search.py\", line 334, in save\r\n faiss.write_index(index, str(file))\r\n File \"\/home\/vlialin\/miniconda3\/envs\/cat\/lib\/python3.8\/site-packages\/faiss\/swigfaiss_avx2.py\", line 5654, in write_index\r\n return _swigfaiss.write_index(*args)\r\nRuntimeError: Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at \/root\/miniconda3\/conda-bld\/faiss-pkg_1613235005464\/work\/faiss\/impl\/index_write.cpp:453: don't know how to serialize this type of index\r\n```\r\n\r\n## Steps to reproduce the bug\r\n\r\nAny dataset will do, I just selected a familiar one.\r\n\r\n```python\r\nimport numpy as np\r\nimport datasets\r\nINDEX_STR = \"OPQ16_128,IVF512,PQ32\"\r\nINDEX_SAVE_PATH = \"will_not_save.faiss\"\r\n\r\ndata = datasets.load_dataset(\"Fraser\/news-category-dataset\", split=f\"train[:10000]\")\r\n\r\ndef encode(item):\r\n return {\"text_emb\": np.random.randn(768).astype(np.float32)}\r\n\r\ndata = data.map(encode)\r\n\r\ndata.add_faiss_index(column=\"text_emb\", string_factory=INDEX_STR, train_size=10_000, device=0)\r\ndata.save_faiss_index(\"text_emb\", INDEX_SAVE_PATH)\r\n```\r\n\r\n## Expected results\r\nSaving the index\r\n\r\n## Actual results\r\nError in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) ... don't know how to serialize this type of index\r\n\r\n## Environment info\r\n- `datasets` version: 1.6.2\r\n- Platform: Linux-4.15.0-142-generic-x86_64-with-glibc2.10\r\n- Python version: 3.8.8\r\n- PyTorch version (GPU?): 1.8.1+cu111 (True)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\r\n\r\n\r\nI will be proposing a fix in a couple of minutes\nJust in case, this is a workaround that I use in my code and it seems to do the job.\r\n\r\n```python\r\nif use_gpu_index:\r\n data[\"train\"]._indexes[\"text_emb\"].faiss_index = faiss.index_gpu_to_cpu(data[\"train\"]._indexes[\"text_emb\"].faiss_index)\r\n```","embeddings":[-0.1138690859,0.2311160564,0.0493216403,0.2159428895,0.3592372239,0.1624383777,0.5158483982,0.5090635419,0.2508896887,0.2249036729,0.0669301525,0.2369603813,0.0509322956,-0.0680036098,-0.1380849034,-0.0389072336,0.2700771689,0.2628393769,0.157356739,-0.1693143994,-0.4407014251,0.0116727334,-0.0017445267,0.2576278448,-0.2485831231,-0.044677835,-0.0931610987,0.0168835968,-0.1002748013,-0.4397268891,0.0679610968,-0.2787279487,0.4836601615,0.3083126843,-0.0001119511,0.16159226,0.2866710722,-0.1724367291,-0.2263427526,0.1483052224,-0.0678200722,0.1392590106,0.0382678434,-0.2912577093,-0.1710164845,-0.2987514436,0.0940966308,-0.2719932199,-0.123602353,0.3052643836,0.2272423357,-0.1925358325,0.3092999756,-0.1131985635,0.1893387586,-0.0202897564,-0.1342300624,0.0498123802,-0.3920942843,0.3410638869,0.2851158679,0.3660792112,-0.0126446206,-0.4190097153,-0.2800433636,0.1759340912,0.60136801,-0.2439565659,0.116647087,-0.0416836329,0.1341054738,-0.2888173461,-0.3871852458,0.1896890849,0.2492049485,-0.493991971,-0.0477072038,0.0656199455,0.0472116917,0.2803214788,0.5671783686,-0.1080207601,-0.1454774588,-0.0482442752,0.1810242683,0.076971449,-0.045570109,0.2025454789,-0.090509668,-0.2889374197,-0.1879069954,-0.0037431747,0.0984284058,-0.0547393262,-0.3431196213,0.0062898733,-0.175707534,-0.4342462718,-0.0055380403,-0.0066519608,-0.4312520623,-0.008348275,0.2703360617,0.118852973,-0.3501147032,0.1605631411,0.009781952,-0.1619330496,0.1525248736,0.0218443051,-0.2132831514,-0.0922401622,-0.1168507561,0.1236494854,-0.1891160011,-0.0733328015,0.2162267417,-0.2217499167,-0.5828648806,0.3758404553,-0.3595431447,0.1377798021,0.0111374157,0.5140858293,0.0791313127,0.050104931,0.1652794629,0.2339881808,-0.1110544726,-0.1419420689,-0.2111119181,-0.1108440608,0.2596277893,0.2357220501,-0.1644803286,-0.3422324657,-0.0027447089,0.0669749305,-0.3403433263,-0.0236196872,0.1093347296,-0.2806960344,0.002856513,0.303132683,0.1009419486,0.0751547217,0.191299662,-0.3268490136,-0.1066069156,0.0402038284,0.0674366206,-0.1408487111,-0.3724992275,0.2187590897,0.2860943973,-0.1972838342,0.6523479223,0.0753616542,0.3555603027,0.1765325069,0.0953457132,-0.067618072,-0.0979846045,-0.3917908072,0.3624824286,0.1430670321,-0.1860331595,-0.0113125201,0.0957461894,0.0662762523,0.3802833259,0.5683287382,0.3160372674,0.2390820831,-0.2121520191,0.4203566611,0.272642225,-0.3779995739,-0.280284673,0.1330714524,-0.1357293874,-0.2128633261,0.2011710256,0.0591533929,0.1582609862,-0.0584350452,0.4598637223,0.1182971597,0.06503205,-0.108853817,-0.4900937974,-0.5541249514,0.2036344111,0.1098531485,0.0084758783,-0.1918828636,-0.1755236685,-0.5428344011,0.2811937928,-0.2432644367,-0.0737854317,0.0249400605,0.7233158946,0.1537608802,0.2423792332,-0.0366046019,0.1383715272,-0.0622231737,-0.1854722351,-0.0290439725,-0.1147089824,-0.2371757776,-0.0223276298,0.0183746554,0.0365044102,0.0494049825,0.1141124964,0.0131477229,-0.1199359447,0.3434326053,-0.2094563991,0.0398113243,-0.3076533973,0.0626216009,-0.2081642598,0.0722263381,-0.3703285158,-0.4066974223,-0.1425864846,0.1701209843,0.3474281132,-0.1863129437,0.0125181749,0.3988081813,0.0344950929,-0.1558338702,0.1012302041,0.0262103546,-0.0489243567,-0.2536347806,0.0307387523,0.5145238042,0.1880791783,0.1472197622,-0.0048773056,0.3355096877,0.1963059157,-0.0242392085,-0.0846971646,0.1306782216,0.4605296254,-0.0087657906,-0.0339583382,-0.2989700735,0.0284292139,0.0268369317,0.0555951558,-0.0418341942,-0.6681858897,0.2340365648,0.2718895376,0.00093151,0.5030539632,0.0890595317,-0.3418864012,0.0547037199,-0.0895656869,-0.5434659719,0.2345535606,0.1372830421,-0.1467340738,-0.1237027422,-0.33535707,-0.1771036536,0.4169137776,-0.2255945355,0.0904394537,0.1040292457,0.1014716923,0.0593896024,-0.1880995184,0.0337482095,-0.1696738899,0.0177561492,-0.3170367479,0.1562788039,-0.1312709153,0.013449572,-0.2444274873,-0.1842757612,-0.0321661569,0.0443149991,0.0538874753,0.1305157989,-0.1142224073,0.284424752,-0.3202932477,0.3281571269,0.4209784865,-0.1238969639,-0.0517931059,-0.0182623453,-0.3266088963,-0.0397244319,0.0600185879,-0.0189724937,0.1622944325,-0.082202144,-0.0453144275,-0.2157865018,-0.0934125483,0.0835173726,-0.1467777193,0.3859787583,-0.0344992802,0.087016359,0.0051547461,-0.03441412,0.1084697768,0.0458991528,-0.0604748949,-0.0019139997,-0.1769495606,0.031945616,-0.1569327712,-0.208674103,-0.1914224327,-0.2342451215,0.2827096581,-0.0373622552,0.2085962892,0.1510173678,0.0411782004,0.2148143202,0.1120457947,-0.0829849541,-0.3446111381,-0.1167994365,0.3281378448,-0.0233734585,-0.487514019,0.0882834867,0.0634188727,-0.1923591197,-0.0627999678,-0.0927179381,0.3474021852,-0.0381610282,0.0832146183,-0.1545877904,0.1555236429,0.2991626561,-0.0797351897,-0.0500682704,-0.1272290051,-0.1483065933,0.2376608849,0.0784068853,0.3545718789,-0.2086637467,0.3585048616,-0.0277999267,0.4486905336,0.2896327674,-0.3432827294,0.2418365777,-0.0633315369,0.2357196063,-0.0512965359,-0.1933173388,0.2188630253,0.0113077005,-0.2435846031,0.1216401458,-0.3592109084,0.0010470334,0.2625235617,0.1317616999,-0.0125195011,-0.2619675398,0.1138253286,-0.0019274667,0.3059886396,-0.163909182,0.3812544346,0.0767232627,0.140803948,0.2754314244,0.4634262025,0.2528236806,-0.0780757964,0.1676052511,-0.1933817863,-0.3587157428,0.2600285411,0.0077132154,0.7269444466,-0.0758038089,-0.171739161,0.2119948417,-0.045222275,0.7033804655,-0.1883667707,-0.3067083359,0.3672285974,0.3515645266,-0.5859757662,-0.1705937386,-0.1365602463,0.0983960852,0.0748953745,0.4549067914,-0.096212782,-0.1658404171,0.0827238634,0.0176609922,-0.3054210842,-0.2055086046,-0.3597078919,-0.293201983,-0.1851569414,0.2824061811,-0.2168985456,0.2141839266,-0.0700112358,0.266512692,-0.5732181072,-0.1213622242,-0.0349181332,-0.0092258155,0.328904897,-0.1224837005,0.1129456162,0.0889962837,0.088052325,0.215499267,-0.0249006432,0.1573171914,0.07492432,0.1120043099,-0.08666832,0.0969557762,0.2478033006,-0.1018243283,0.0557654612,-0.1926347762,-0.0416368358,-0.3843403757,-0.0372369699,0.2731849253,-0.0681490824,-0.3149110079,-0.0548362508,0.5327314138,-0.0808583573,-0.0924986303,0.0570368767,0.2378219962,-0.2765051126,0.7876927853,0.5123694539,0.8745913506,-0.0249332618,0.2387755513,0.3136335015,-0.4298110604,0.4504750371,-0.0873857886,0.3265615106,-0.6909171343,-0.2089759707,0.0071772272,0.0387976617,0.1789012104,-0.3683890402,-0.4958117306,-0.0103025893,-0.2931323647,-0.0426613316,0.0684217364,0.1311561465,-0.1014822572,-0.3177869916,-0.1130020842,0.1260932833,0.1939209998,-0.0398441255,-0.0478853397,-0.1176901609,-0.120188117,0.0173483938,-0.23809053,0.1713317633,-0.0718619749,0.4490128756,-0.2436015755,-0.1858198792,-0.233213827,-0.3181995153,0.5306907892,-0.1426222622,-0.1438910961,-0.1689338833,-0.0461881347,-0.0520061515,-0.0457245223,-0.2577606738,0.4293529093,0.1032442078,-0.1271422505,0.003966351,0.1561392546,-0.0028365764,-0.2947388589,-0.136777401,-0.416726768,-0.1705379635,-0.2591062188,-0.2123648226,0.1281217039,-0.1432858855,0.1266583353,0.1254921257,-0.2148133367,0.3663400412,-0.1937049329,-0.1992466897,-0.0968680754,0.3377802372,0.1716618091,0.22840105,0.4187211692,-0.0312773027,-0.1312450618,-0.2868302166,-0.0087426109,0.0588750094,0.1039057672,0.0444741175,-0.3001897633,-0.1903582811,-0.0765734017,-0.0567488223,0.0982520357,-0.4079138935,-0.3621655405,-0.198867172,-0.3984547555,0.0485116988,-0.0694913715,0.1906043887,-0.0342383571,0.2257736772,-0.1961165071,-0.2054575086,-0.3127098978,0.0522203743,-0.2639787495,0.1025281325,-0.1675952673,-0.0767601058,0.1980795413,-0.0422759317,0.1842722446,0.2328882068,0.0275049768,-0.1886847317,-0.3180929124,0.094292514,0.0779786259,0.0868215486,-0.1164679453,-0.0805213749,-0.1669326276,-0.0123439189,-0.1684740335,0.1831952929,-0.0591792092,0.1752375364,-0.0839023069,0.136741817,-0.0152720539,-0.004783351,0.1312592328,0.3450068533,0.0749549493,0.1724459231,-0.4182553589,0.0649813116,-0.1920062453,0.0097517576,0.2220517546,0.1787428707,0.1739415824,-0.2661895752,-0.338511765,0.0960966125,-0.0441372842,0.2958493829,-0.3334267139,-0.1905685812,0.1959242374,0.220736444,-0.0675949603,-0.034242481,0.2121056467,-0.3065629601,-0.1031314209,0.2337240279,0.3573590815,0.2369272262,-0.3551806211,0.1168408319,0.2302570641,0.306124717,-0.017234629,0.0000812426,0.1974563301,0.1980648786,0.1406739205,0.1616263092,0.5372236371,0.7073373795,0.0098228343,0.1949501336,0.2589001358,-0.0168180279,-0.0678319559,0.0341058038,0.2200885862,0.0367187299,-0.0308126137,0.0902553573,-0.005316765,0.1024517417,-0.297632426,-0.3377819657,-0.0103417039,0.2341649383,-0.2971342206,-0.1746603549,-0.015743956,0.1265753061,-0.0371046849,-0.1824046373,0.0277615152,0.1088751256,-0.0475750975,0.1855154186,-0.1767652631,-0.1655455679,-0.2276676595,0.5172172785,0.3538103104,-0.2013293356,0.0686090812,0.0342892334,-0.1833887994,-0.2204377502,0.2049185783,0.4763285518,0.4229625762,-0.4053070843,0.3371579945,-0.0608097948,0.0121298004,0.0333166681,-0.1948703527,0.0502511188,0.0437494814,0.2616239786,0.1441911757,-0.0435963944,-0.1528134197,0.2445415109,0.4285230041,-0.0031446205,-0.0318151638,-0.2115763873,0.0656905696,-0.0291512981,-0.0721508712,-0.3379934132,0.3102944195,0.4012337327,-0.0870110542,0.2544293106,0.1473809332,0.0780670494,-0.1207248718,0.2295898646,0.1237019375,-0.0928331241,-0.2431087047,0.027510941,-0.132309705,0.2722803652,-0.0667425543,0.1401506364,0.1952130198,0.5284093618,0.0032887326,-0.2131545544,0.2024678141,0.1564042419,0.3337155581,0.3043870926,-0.2011167854,-0.0649317726,-0.3702446818,-0.1155204475,-0.001927143,-0.2382215858,0.3254854679,0.010603522,0.1508497894,-0.1065704003,-0.2740846276,-0.0899966881,-0.2683733702,0.8571715951,0.1699066013,-0.1377744973,-0.2573096752,-0.339368999,-0.066035457,0.0179766305,-0.3312304616,0.0966607705,-0.0171484575,0.2599122226,-0.2001356781,-0.2401793599,-0.3278620541,0.6025990844,0.1020162329,-0.1882013232,-0.2061789781,0.3153871,-0.2367208302,-0.1087487563,0.0324265473,0.2104197741,0.0471045822,0.5866944194,-0.1895516813,-0.2843569219,0.5402361751,-0.322381556,-0.2281233221,-0.1142283231,0.4193398654,0.3192804158,0.307844609,-0.5473868251,-0.0813798606,0.2216365486,0.0549654774,-0.1970957965,-0.1857766509,0.2253252417,-0.0525562018,-0.1044452414,0.0102511458,0.0774492323,0.1365794688,0.0969914272,-0.2957920432]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2347","title":"Add an API to access the language and pretty name of a dataset","comments":"Hi ! With @bhavitvyamalik we discussed about having something like\r\n```python\r\nfrom datasets import load_dataset_card\r\n\r\ndataset_card = load_dataset_card(\"squad\")\r\nprint(dataset_card.metadata.pretty_name)\r\n# Stanford Question Answering Dataset (SQuAD)\r\nprint(dataset_card.metadata.languages)\r\n# [\"en\"]\r\n\r\n```\r\nWhat do you think ?\r\n\r\nI don't know if you already have a way to load the model tags in `transformers` but we can agree on the API to have something consistent.\r\n\r\nAlso note that the pretty name would only be used to show users something prettier than a dataset id, but in the end the source of truth will stay the dataset id (here `squad`).","body":"It would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.","comment_length":95,"text":"Add an API to access the language and pretty name of a dataset\nIt would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.\nHi ! With @bhavitvyamalik we discussed about having something like\r\n```python\r\nfrom datasets import load_dataset_card\r\n\r\ndataset_card = load_dataset_card(\"squad\")\r\nprint(dataset_card.metadata.pretty_name)\r\n# Stanford Question Answering Dataset (SQuAD)\r\nprint(dataset_card.metadata.languages)\r\n# [\"en\"]\r\n\r\n```\r\nWhat do you think ?\r\n\r\nI don't know if you already have a way to load the model tags in `transformers` but we can agree on the API to have something consistent.\r\n\r\nAlso note that the pretty name would only be used to show users something prettier than a dataset id, but in the end the source of truth will stay the dataset id (here `squad`).","embeddings":[-0.188386336,0.1110786572,-0.1189651936,0.3318739533,0.3078963459,0.1275812835,0.2714886963,0.2705741525,-0.1630472392,0.2370295674,0.182367295,0.5307036042,-0.2200144529,0.3426440358,0.2076429725,-0.1973491013,-0.0799138099,-0.0694945529,0.0691232532,-0.068991445,-0.1761096567,-0.090749979,0.1633303612,0.1583644748,-0.0301870462,0.0261960234,-0.0522684716,-0.3547734916,-0.0490883105,-0.3833298385,0.0265071467,0.2685779333,0.2704098821,0.0535490401,-0.0001118512,-0.1514771879,0.428049475,-0.0854010656,-0.3574964106,-0.3493791521,-0.3073886037,-0.1295091808,0.1739201546,-0.3760478795,-0.1767501384,-0.3267531395,0.2022295594,-0.3540710807,0.2644023895,-0.0619541481,0.176770106,0.1459209621,-0.1946241856,-0.0098479623,0.1031239405,0.6372214556,-0.1374300569,0.0719209313,0.3668981791,-0.0315231904,0.0537887476,0.4254769981,0.1457024813,-0.099273853,0.1142932922,0.2550070584,0.0585154332,-0.2441239804,0.1967636049,0.308009088,0.6004468203,-0.3262732625,-0.5997585058,-0.3656826019,-0.022137465,0.0269224588,-0.1059647277,0.1007210612,0.0055736718,0.325719744,0.0648374185,-0.1178197637,-0.3247561455,0.1506285816,0.2871362567,0.673175633,0.0636098012,0.22268866,-0.2268857211,-0.05360654,0.2841743827,-0.2307621688,-0.0192712378,0.3466577232,0.0307829492,-0.2591355145,0.0764825791,-0.0677650273,0.0404178016,0.2686677575,0.0246033706,0.2520871162,-0.1839669943,0.057052467,0.1863696575,-0.0583458804,0.3656691611,0.3474362493,0.1395958662,-0.2905345857,0.3369954526,-0.1128106937,-0.4009593725,0.3998141885,0.1455465257,0.236240387,0.1147171482,0.0293195825,0.1392043382,0.0995041132,-0.0031157157,-0.193773061,0.0437826328,0.2090449929,-0.1029655933,0.4838320613,-0.0575311519,-0.0558884256,0.261975795,-0.2849515378,-0.0396051444,-0.0428765304,-0.1556304097,0.2016425729,0.1335633397,0.4709702134,0.1575273275,-0.2021083236,0.0211380981,0.0888663605,0.5216773152,-0.1072577387,0.1699405164,-0.1429651678,-0.1553065479,0.2478838414,0.0340855382,-0.331638068,-0.3818343282,0.0909754857,-0.1794744432,-0.0662017539,-0.2235243022,0.2019513249,-0.12367782,-0.0743874162,-0.3969196677,0.546120584,0.0621863715,-0.0336483493,0.094790183,0.0439010486,-0.080586791,-0.0177111849,-0.1008541659,0.4807066321,-0.3225237131,-0.3198370337,0.1414936781,-0.1929112971,-0.1637776345,-0.1636966318,-0.4127165377,0.278162241,0.0034137366,-0.0515932515,0.4448745847,-0.433339715,-0.0318218097,0.1604062766,-0.0531949773,-0.0830703527,0.2868572176,0.206754297,0.1231195852,0.0766081512,-0.0214541331,0.4316636026,-0.1471989006,0.0135139525,0.0924810171,-0.578383863,0.1879703403,0.4853284061,0.1501314491,-0.0893118456,0.0238716174,0.2388889939,0.4389698803,-0.1998252571,0.0331594422,-0.1284463406,0.2546330988,-0.0727186948,0.1066728234,-0.2864490449,-0.4631745815,0.0916081369,-0.1845386773,0.3253059089,-0.2072031498,-0.3501712084,0.1826668829,0.069185175,-0.1576730162,0.2429236025,0.1133722365,-0.0418601297,-0.1135338247,-0.1330611408,-0.1494903415,0.0675101578,-0.0991156325,-0.0889893919,-0.5047960877,-0.0932677686,0.0735899955,0.1328448653,-0.3385683298,0.0157211944,-0.1293935925,-0.154828459,-0.0818671733,0.0013924312,0.3518503904,0.1150767803,0.1976159066,0.6180114746,0.290583998,-0.2617429793,0.0504497923,0.3226174414,0.2168333828,0.0121784154,-0.3475636244,0.3583140075,-0.0680054799,0.2737100422,0.0967417136,-0.1490737051,0.4255134761,0.1746971607,-0.3774278164,-0.3048305213,-0.0601780042,0.1465276033,0.147524789,-0.0568139926,0.000165446,0.1434458047,0.5072767138,-0.08790344,0.0534009859,0.0310843028,-0.0818457156,-0.0111306952,0.0422716625,0.2109892517,0.2227065563,0.2407680154,-0.2791877389,-0.0600433983,0.283783406,-0.0428526103,0.2694918215,0.408303231,0.0491667688,-0.2313789427,-0.0485485569,0.1354128569,0.0925141424,-0.5166596174,0.1762725413,0.0856549963,-0.1000048593,0.270493418,0.1103067249,-0.3297911286,-0.3040924072,-0.356790185,-0.1466909945,-0.2314637899,0.0276809353,-0.4174096286,-0.2768644392,0.1285554767,-0.1056592539,0.2818367183,-0.2166183144,-0.2801747918,-0.080882065,-0.6389325857,-0.182870999,0.0282986108,0.1939239651,-0.1233168095,0.4811219871,-0.1316085905,0.2442051917,0.0948468819,-0.5168853998,0.2221330851,0.1477455497,0.1503617316,0.0017751112,-0.0010787621,0.3242390156,-0.0951399058,0.0896564275,-0.1282002032,0.2495244443,-0.0675789341,-0.0633373111,-0.247516349,-0.1216358691,-0.3174870312,-0.6008809209,-0.1647494882,0.0497582965,-0.0585563146,0.1887986958,0.3861448467,-0.131436184,0.2721607983,-0.166587919,-0.0442869253,-0.1468530446,-0.0880572498,0.1331970394,-0.10978356,-0.1724701673,-0.1961832345,-0.2997507751,0.2696593106,-0.2651053965,0.0918115377,-0.1628443152,-0.0597254261,0.0677622482,0.0137443012,0.0391947739,0.3031950891,0.1896317601,0.0316593982,0.138198033,-0.1563609689,0.1373777241,0.0796183944,0.0916441157,-0.0959923789,0.1144304425,-0.233294487,0.4291267693,-0.057969071,-0.2486724555,0.0941292346,0.1176648661,-0.0398776568,0.1980214566,-0.2120458484,-0.1962487549,-0.0698767677,-0.1064824536,0.3461346924,-0.0721686259,-0.4198964238,-0.210419178,-0.0121268565,-0.2709896266,-0.2303487659,0.4262935221,-0.2471023202,0.029464718,0.1658520848,0.1617808044,-0.3989671767,-0.0047365571,0.0586278141,0.0179061387,0.0122409137,-0.1063569114,-0.4593088329,-0.2083370984,0.1658546478,0.2547630072,0.1404826194,0.1163140461,-0.1070870459,-0.059194319,-0.0427075736,0.0485874526,0.3832095265,-0.2628858387,-0.2734737992,-0.1513476521,-0.0940894112,-0.2443834245,-0.0047431728,-0.4492405951,-0.2510789335,0.0672510713,0.5958282351,-0.3425453901,-0.1368212402,-0.0067864908,0.1104082018,-0.3247512579,-0.1207775623,-0.0612855069,0.0893330798,-0.3224770129,0.0090525886,0.0288016889,-0.0605014339,0.1205050647,0.0186339058,0.0029254323,0.212298587,0.2857385576,0.5010877252,0.0051692072,0.10105443,0.101924412,0.1425134689,0.065304473,-0.2637544274,0.2954247892,-0.1171863675,-0.4510862529,0.1569934934,-0.1934814751,0.410979569,0.2477842122,0.1050267294,0.2390011251,0.3234278858,-0.0419203565,-0.5307347178,0.2066841274,-0.1548741013,0.1049121916,-0.5274234414,-0.9195456505,0.6403459907,-0.035996493,-0.2454494536,0.1327768564,0.2789333463,-0.3952602446,0.9277620316,0.2805678844,1.1838779449,-0.1109055579,0.1133580953,-0.2915828526,-0.2011342496,0.6841204762,-0.2174706161,-0.0981718749,-0.3596976697,-0.1068282649,-0.1066991761,-0.0206151437,0.170184955,0.4003402889,-0.2707736492,0.0070978296,0.0679802895,0.2797981203,0.0900011733,0.696133554,-0.1064019725,-0.5078454614,-0.1485538483,0.1092171595,-0.2880716026,-0.0006130933,-0.1443094909,-0.2325149924,-0.2226059437,-0.0717897117,-0.3035658002,0.1049666852,-0.2098395675,0.1903997213,-0.0677833632,0.0253501236,0.4092522264,-0.0756430104,0.2082840502,0.3011295497,-0.3239886165,0.1833632886,-0.2004252374,0.1588977128,0.1242030263,-0.097476393,0.3460097313,0.1220505759,-0.3134865761,0.354521215,0.2316941917,0.0650300086,-0.5912466049,-0.0323423296,0.2109403163,-0.1195839047,-0.2272166014,-0.1336141974,-0.1775825471,0.008897298,0.1382501721,0.2032518387,-0.2360197902,0.1328653544,0.2284072638,-0.0286654308,-0.0831754729,-0.0430631414,0.0103243329,-0.1383846551,0.2229240537,0.073172003,-0.0880995095,-0.2573052347,0.158940658,0.294539541,-0.5770974755,-0.344764024,0.2088222504,-0.3290333152,-0.1639159769,-0.2163973749,0.1821503937,-0.0085369926,-0.0631075948,-0.2203497738,-0.3973639905,0.2896134555,0.1392595917,0.2403062135,0.0678501278,-0.3488306999,-0.2570100427,0.2480110824,-0.3264061511,-0.3125216067,0.0208861846,0.2037793249,0.1241406873,0.0463331752,0.036868766,-0.2900848091,0.0525608212,-0.0961519629,0.0332685113,-0.2268172652,-0.4099669456,0.1431797296,0.1258359551,0.0287392028,-0.1444209665,-0.0890864879,0.1120790318,-0.3430493772,0.0900224745,-0.0147025399,0.1279264838,0.4479000866,0.140070647,0.0852236673,-0.2326353192,0.2551234961,0.0884528756,0.2350688726,0.3074111938,0.2131789327,-0.0286646988,-0.0068740994,-0.0918991119,0.2367048413,0.323235929,0.2591973841,0.0199374054,0.0331576876,0.3183542788,0.3111644089,-0.0162219647,0.100843288,0.0148461638,-0.0186626911,0.1883951128,0.2323776484,-0.2617147565,0.0040873387,-0.0288032796,-0.1045663878,0.1420136839,0.4495282173,-0.0260925684,0.4359008372,0.0027754528,-0.0543318987,0.512665391,0.0542792343,0.2052687407,0.3901941776,0.0100830747,0.1212394312,0.2000378072,0.1871794313,0.2621740699,0.2354494333,-0.1585978717,0.5301888585,0.0568320639,0.1388507336,0.2717408538,0.1085636243,-0.0960551724,-0.1925899088,0.036962606,0.4369643331,0.1960623562,-0.0556247868,0.0479136854,0.0679945424,-0.2876179218,-0.1570555568,0.0580421053,-0.2667608857,0.0368300565,-0.0722604543,-0.0605612658,-0.1712153852,0.0636689961,-0.3644526601,0.0271636415,0.1770180464,-0.0277990866,0.0680298358,0.0085922843,0.2444881499,0.1561803371,-0.2653382123,-0.1434295624,0.0931202173,-0.068806462,-0.2415265441,0.1719186902,0.0235701557,0.1138699204,-0.1319382191,0.3289744258,-0.1361106485,-0.1000301465,-0.1604032815,0.1897663623,0.1998181343,0.0366170369,0.0214318782,0.1722930074,-0.1096117496,-0.3068856299,0.1988538057,0.0478212573,-0.0392372459,0.6089425683,0.195598051,-0.1415049881,-0.1341377795,0.1517565697,-0.5333811045,0.2032198608,0.5581694841,0.1650266796,0.0316838957,-0.0624929741,0.0436933525,-0.0113292942,0.3779026866,0.1115214899,0.2741173506,-0.3451054394,0.3276253641,-0.410155952,-0.1148513183,0.1545855701,0.1877103895,0.037359301,-0.0740534216,0.3624297976,0.1715920717,-0.2454886734,0.1477734149,-0.0864703208,0.1268274486,-0.2789252102,-0.2352340072,-0.1752721071,0.2359231412,0.0659240186,-0.183932215,-0.0510216579,0.1628499925,-0.0599640571,-0.6815500855,-0.1216642931,0.2556994259,-0.0708698407,0.3999885917,-0.0054890164,0.3638622463,-0.0179951265,0.0898324996,-0.1918762624,-0.1961183548,-0.0333201773,0.4306816161,0.1151777208,0.1724217981,-0.0141095212,-0.1996969581,-0.2273341417,-0.3305166662,-0.2913440168,-0.369266361,-0.4320269227,-0.0503303595,-0.1914570034,0.0151615273,-0.1938505471,0.1640466154,0.064658165,0.0513589792,0.0739249811,0.0380448997,0.3093402088,-0.3006005585,-0.3384078741,-0.0731964335,0.088076368,-0.3810874522,0.1482792497,-0.6339802742,0.1331757605,0.3060176671,-0.1305230558,-0.0326129422,0.284879595,0.4409700036,0.1968102604,-0.306433022,0.2921519876,-0.0798793435,0.0057390383,-0.2772657573,-0.184896037]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2347","title":"Add an API to access the language and pretty name of a dataset","comments":"That works for me!","body":"It would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.","comment_length":4,"text":"Add an API to access the language and pretty name of a dataset\nIt would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.\nThat works for me!","embeddings":[-0.3502853215,0.0495913401,-0.1775059849,0.2301256657,0.3177360892,0.164116323,0.3228104711,0.3208495677,0.0157352686,0.2803491652,0.1759234518,0.5126197934,-0.1418594569,0.437076807,0.2498487085,-0.2016717792,-0.0521459691,0.0593625121,-0.0460758545,-0.1050447673,-0.3137137294,0.0223157145,0.1363231391,0.0568493865,-0.0872413442,-0.0047346586,-0.0278062075,-0.218979612,-0.1850874126,-0.3700324595,0.0040160716,0.3026528358,0.2828627229,0.0047072764,-0.000105022,-0.0820352361,0.3481831849,-0.0933142528,-0.358210355,-0.2037404925,-0.3364557326,-0.1231404841,0.2374293208,-0.3038119972,-0.090959698,-0.3380981088,0.1973797679,-0.3010746837,0.2803879976,0.0913372859,0.2451131493,0.2186708748,-0.2827739716,-0.1032993793,0.041213993,0.5932502747,-0.1691324711,-0.0584159642,0.2925573885,0.0089819916,-0.1575833559,0.3041018844,0.2174441069,-0.201265648,0.1312519759,0.0953644812,-0.0236835089,-0.2210141271,0.1110852286,0.1934835315,0.6563660502,-0.229124248,-0.5306963921,-0.2462700009,-0.0573833175,-0.0000148497,-0.0265471246,0.1455394477,0.0320522748,0.3205779493,0.0877848938,-0.1604950726,-0.4117412567,0.0246803984,0.3270285428,0.5738529563,0.0884325579,0.1475249678,-0.1164591014,0.0464229323,0.3325067461,-0.2122336179,-0.0741344318,0.2803394198,-0.0031893149,-0.2666256726,0.1766436249,-0.008972927,-0.0816363469,0.1481118202,0.1669183224,0.3597914279,-0.1179801673,0.1007161885,0.1618638337,-0.0622021593,0.2824542522,0.1774471998,0.2934402823,-0.1575571001,0.3308173418,-0.1783311814,-0.3749627173,0.4540731311,0.0369345285,0.1794706732,-0.0119289923,0.0097173229,0.2109908462,0.0387989767,0.268304497,-0.255613178,0.0459817797,0.2590984106,0.0490740649,0.4851492047,-0.0796926543,-0.0193878971,0.1131202951,-0.3784913719,-0.1101419181,-0.0445373766,-0.2079063058,0.2154807597,0.1015275493,0.4976240098,0.1912083775,-0.1691650599,-0.1019909903,0.1785482466,0.6041755676,0.055481974,0.1064164266,0.0332368575,-0.1475004107,0.0526219495,0.0397732519,-0.089981854,-0.3440059423,0.0736240223,-0.1403566599,-0.265725702,-0.1995448172,0.3029581308,-0.1101839691,-0.0528028309,-0.2632117271,0.5745816827,-0.0570780486,-0.1201522723,0.1889041215,0.0408783928,-0.1137892604,0.0020749571,-0.2369404733,0.3243205547,-0.3611980081,-0.3412049413,-0.0217220299,-0.3638513386,-0.1127201095,-0.2186565399,-0.5122693777,0.1852211654,-0.0193516966,0.0722446367,0.6039071679,-0.2448027432,0.0371318348,0.1789212674,-0.0190262739,-0.1772219241,0.2319673002,0.2188230455,0.1496622115,0.0177906919,-0.0186440218,0.3611595035,-0.1499570757,0.028592376,0.0204741955,-0.4643070102,0.0957165882,0.4961187243,0.1423460245,0.0435371064,0.110700056,0.2175605595,0.2672693729,-0.2181571424,0.0123155685,-0.0600889288,0.2513706684,-0.1872873902,0.1553940773,-0.1945987642,-0.5429002047,0.029860938,-0.3054523766,0.3567158282,-0.0342100039,-0.262740761,0.183981508,0.1283997148,-0.1578704417,0.0974001363,0.190051496,0.0864034817,-0.1687295437,-0.1532251388,-0.288631767,0.07213673,-0.1307543218,-0.0745190755,-0.477301091,-0.1784743071,0.0220527798,0.0891442373,-0.1957420558,-0.0270764474,-0.1845665276,-0.1967969388,-0.199268803,0.082233794,0.4154603183,0.1460976005,0.0775599927,0.594943285,0.3038707674,-0.1350553185,0.0778357089,0.1188931242,0.13697429,-0.017200904,-0.3609302938,0.3951181471,0.0491595231,0.1597483307,0.0818152502,-0.0890428126,0.3511003256,0.182405293,-0.2881779969,-0.2858135104,-0.036678765,0.1958215982,0.229323104,0.0303749721,-0.0721056238,0.1711070389,0.5291377306,-0.101560019,-0.0799956098,0.1815533489,0.0322926044,-0.1644975394,0.1376699507,0.3227258027,0.2521500885,0.2604472637,-0.2471482754,0.0001228164,0.2196914703,-0.1143733114,0.2618025243,0.2752870917,0.0898934975,-0.196024552,-0.084078446,0.0804397538,-0.1352247,-0.53808707,0.1698540002,0.1771606207,-0.0240960587,0.1678847224,0.0284425765,-0.170105204,-0.1178930774,-0.1495739818,-0.0745385289,-0.315857321,0.1330419928,-0.3276257217,-0.4173078537,0.16819489,-0.2520864308,0.4436039329,-0.2020245939,-0.0618132614,-0.2754162252,-0.5820621848,-0.2273128331,0.1093734056,0.1749397069,-0.0498565324,0.5028125048,-0.0604990683,0.2424672395,0.1387492269,-0.4640847743,0.2977093756,0.1029239967,0.0883198231,0.1097376496,0.0871210694,0.3324848711,-0.1931634098,0.1016655713,-0.1869425923,0.0684472397,-0.0702064335,-0.0692188516,-0.2810708582,-0.1501561254,-0.34378317,-0.3140520751,-0.3931384683,0.0230133999,-0.1311063766,0.199306041,0.2840713859,-0.0892761052,0.2960542738,-0.0417933874,-0.0106075145,-0.1418711692,-0.1628510207,0.1296890229,-0.2293480933,-0.2680482566,-0.2585278451,-0.2692846656,0.3795395792,-0.0765688047,0.0221331194,-0.0341239944,-0.1667846888,0.0180299804,-0.0167476274,-0.0374106243,0.3663125336,0.3298105597,-0.1086948365,0.0266891848,-0.1859556288,0.0947545692,0.1873093694,-0.0379850827,-0.1440155059,0.1242351905,-0.3088566959,0.4195920527,-0.1741080284,-0.2057474554,0.2360073328,0.0915873498,0.0670209303,0.0210121777,-0.1431158781,-0.2163945585,-0.0538789555,-0.0440160744,0.3755688965,-0.157106027,-0.4871304035,-0.1770736128,-0.0778405741,-0.1369544268,-0.1414766759,0.4658180177,-0.3028299212,0.1483553797,0.1614520997,0.112734139,-0.3045597374,0.152327314,0.0705470666,-0.0334560648,0.0475637987,-0.1947263777,-0.4566761255,-0.1834852993,0.006837179,0.2205854356,0.0597326458,0.1407147199,-0.1661607027,-0.240264684,-0.1315689832,0.0159731302,0.3934558332,-0.3746570349,-0.5076360703,-0.1830558479,-0.1393463612,-0.2476564199,0.0243289843,-0.3903344274,-0.1614426225,-0.0219838209,0.5740508437,-0.2528949976,-0.2716288269,0.0194749366,0.0893154219,-0.320472151,-0.1798454076,-0.0681939349,0.0647286847,-0.4000614285,-0.0773709491,-0.0544069372,-0.1334854215,0.0885966569,-0.0305349622,0.1328147203,0.1702814102,0.1900164485,0.5404166579,-0.064354457,0.0172245316,0.0523762405,0.0383152589,0.0021495798,-0.1446664035,0.2340634018,-0.0815491751,-0.293277353,0.2824063599,-0.2295643538,0.366379559,0.1730614007,0.1175469458,0.2223206609,0.3088794351,-0.0436380059,-0.4194453955,0.2200592756,-0.115033254,0.0259508658,-0.5116327405,-0.7796267867,0.5268663168,-0.2366655171,-0.1871548891,0.3022610843,0.2399468273,-0.3070839047,0.9652080536,0.2879996002,1.0420871973,-0.0368284397,0.1153606474,-0.1860263497,-0.1088872775,0.5330466032,-0.2417856455,-0.0582749322,-0.2939398885,-0.0283906106,-0.1768308133,-0.0230487287,0.1609066874,0.3204155266,-0.3210467994,-0.042169109,0.2319729626,0.0940586329,-0.0339390151,0.7472791672,-0.1495342553,-0.4917582273,-0.0563861728,0.1782340407,-0.220859468,-0.029676307,-0.172586292,-0.2806499004,-0.0065051313,-0.1343878359,-0.2719815671,0.0836559907,-0.1893547028,0.1448550224,-0.0859775394,0.1110875458,0.4488755167,-0.0658603311,0.099176906,0.1780147403,-0.334672451,0.1831165105,-0.243066594,0.160959363,0.0471713766,-0.1602298021,0.3300664425,-0.0061336271,-0.2738345265,0.1501136869,0.0984179005,0.1169010028,-0.3561852872,-0.0701902211,0.143724516,-0.2473666072,-0.187155962,-0.0827369541,-0.3574693203,-0.1275292635,0.2088584006,-0.0297143217,-0.1078812182,0.1074275076,0.2971345782,-0.0025594593,-0.1677225828,-0.0636990592,-0.0733199939,-0.2026983798,0.2309023291,0.0403234623,-0.1168681532,-0.4061762393,0.0921739638,0.4648104608,-0.5979719758,-0.1245609224,0.0754336417,-0.3889811337,-0.0919485167,-0.0421141721,0.1858377606,-0.024954956,-0.0752549171,-0.1307805628,-0.4842279851,0.2325913608,0.0647199154,0.4048976898,-0.0415765606,-0.3539332449,-0.1002324522,0.2611479163,-0.4148260653,-0.384272933,-0.0404307768,0.1837529093,0.1813320071,0.066886507,0.0534570217,-0.1556785554,0.1046464667,-0.0149829444,-0.0865059495,-0.2605804801,-0.3772441745,0.1242399514,0.0979581475,0.0425453819,-0.0585454926,0.0351991765,-0.054712832,-0.2839814126,0.0956799239,-0.1495872736,-0.0392264165,0.4464496672,0.204905808,0.0888658091,-0.1657830626,0.1240682527,0.0732691884,0.2275289446,0.2540443838,0.1481739283,-0.0318934731,-0.0773013458,0.0260488782,0.336258918,0.2707450092,0.2418014854,0.0101879332,0.1705252975,0.2663040757,0.1873089522,0.0290882401,0.0660428628,-0.0152295977,0.0083105937,0.1828339547,0.302895844,-0.2531895638,0.0363909639,-0.0325107761,-0.1884499192,0.1454906017,0.3276316822,0.1544748694,0.3267995715,0.2779004872,0.0128524788,0.3999137878,-0.0796907395,0.2873780429,0.3595487475,-0.0787759125,0.1612105072,0.2858903408,0.1505380273,0.2155576944,0.252771616,-0.1137245372,0.490539968,0.096702747,0.1669734269,0.2135974169,-0.0539368503,0.0547907501,-0.1521181315,0.001974337,0.4300383627,0.3131800294,-0.0117240483,0.0949692875,0.0998370573,-0.3478962481,-0.0759031177,-0.0833484232,-0.1416791528,-0.0246912576,-0.1020157561,-0.0924947187,-0.0929386467,0.1467873454,-0.0648030713,0.0513094291,0.168238163,0.0809711516,0.1105282009,0.12516132,0.2162425518,0.0263027195,-0.2132523805,-0.0623501912,0.2642304897,-0.2062572241,-0.1537845135,0.1060859188,0.0207514074,0.1568329334,-0.1161991954,0.3835520744,-0.0656320006,-0.1395930946,-0.1880833209,0.2528643608,0.2533898354,0.2515535653,0.001387429,0.2197742164,-0.1517357677,-0.2567914426,0.1435356885,-0.0742891729,0.0963314623,0.54572469,0.4107125401,-0.2559317946,-0.1704484373,0.1838677824,-0.4129803777,0.2071996033,0.4369111061,0.0156527702,0.0364673622,-0.1050668955,0.0854003057,0.0056218393,0.5967510939,0.1174326167,0.2117950171,-0.2688120008,0.0826737285,-0.2323383093,-0.0823716447,0.1235492751,0.2964871228,0.0910956189,-0.0905394331,0.3698862791,0.0627399907,-0.0672686547,0.0894702151,-0.1246023998,0.0717184991,-0.2718918324,-0.1151234433,-0.0713769719,0.1809432805,-0.039010141,-0.1806721836,0.054855153,0.2262910455,0.0211782139,-0.6979060769,-0.1020198688,0.2269871831,-0.0190783441,0.3922932148,0.1135794744,0.309338659,-0.2318971008,0.1690592766,-0.1715503186,-0.2231942415,-0.0045522889,0.2192415744,0.198089391,0.1350924373,0.0184539855,-0.2109641731,-0.3719199002,-0.2666019201,-0.2134855092,-0.1445928663,-0.3611726463,0.0859314129,-0.1895067096,-0.0688693225,-0.0744379535,0.0984994471,0.1696610749,-0.1148534641,-0.0552482605,0.0527004935,0.3233285546,-0.1557818949,-0.4293305874,0.0458808281,-0.0193116143,-0.3372173905,0.1913745999,-0.5748399496,0.1452222914,0.2445679605,-0.0897105634,-0.1065029651,0.465631783,0.3442919552,0.0969081149,-0.2697210312,0.3814453781,0.0208088644,-0.095573023,-0.3887976408,-0.2115291655]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2347","title":"Add an API to access the language and pretty name of a dataset","comments":"maybe use the hub-backed dataset_info method? (so there's only one parser of README.md metadata)?","body":"It would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.","comment_length":14,"text":"Add an API to access the language and pretty name of a dataset\nIt would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.\nmaybe use the hub-backed dataset_info method? (so there's only one parser of README.md metadata)?","embeddings":[-0.382519871,0.0132843601,-0.1555714309,0.3113779426,0.3000566959,0.2545586228,0.3027870953,0.3353427052,-0.0138798403,0.3614401817,0.051078029,0.4932179451,-0.2035108805,0.3929581344,0.3005828261,-0.1664565951,-0.0400700867,-0.0375355147,0.0795915872,-0.0435439199,-0.2573100328,-0.0304007493,0.143102929,0.089222692,-0.0866039991,0.0475868396,-0.0231739208,-0.2801446915,-0.2100512534,-0.3556900918,0.0326040909,0.2761154771,0.2883509994,0.0284637492,-0.000111529,-0.1112070233,0.4743473232,-0.1261722147,-0.3554672897,-0.2672204971,-0.3943985999,-0.1439613551,0.2066294551,-0.3411194384,-0.0257762205,-0.2639379799,0.2164467275,-0.3559041321,0.1725357026,0.0626279488,0.1912742406,0.1313304752,-0.2323884517,-0.1017961726,0.1056906506,0.6263754368,-0.144379735,-0.0231474005,0.3764878511,0.0552008934,-0.1615622491,0.3400904834,0.0746036544,-0.137576744,0.1550125331,0.2139600217,0.0207617711,-0.1308389902,0.1921700537,0.2548314631,0.8016534448,-0.3804926276,-0.5545297861,-0.3470294178,-0.0895756781,0.000549968,-0.0743359923,0.2041956335,0.0208753143,0.3687666059,0.0549212955,-0.2355682254,-0.3561957479,0.0231102668,0.3485353291,0.6599277258,0.0948710218,0.1736351848,-0.138735339,-0.0019289218,0.5073009133,-0.251806289,-0.0953684449,0.3278715312,0.1157637462,-0.2555829585,0.1286350042,-0.0184291061,-0.031544257,0.2447680384,0.0488362201,0.2842275202,-0.2811157405,0.097469084,0.1191880032,-0.0914768428,0.3690194488,0.2230026573,0.2695890069,-0.2423205227,0.4464545548,-0.1780708581,-0.4069765806,0.5265490413,0.0307816844,0.199094519,-0.0695647076,-0.0190331023,0.1810664982,0.0239023156,0.2267699093,-0.2630895078,0.0233818423,0.2009836137,0.0379873998,0.4408716261,-0.0910511687,-0.002234681,0.2224881649,-0.3663612902,-0.0850903541,0.0363831744,-0.1836175919,0.2663980722,0.0837678835,0.515724659,0.2024513632,-0.1691033542,-0.0375328995,0.1080490276,0.4824996591,-0.0065419008,0.116141513,0.0487967767,-0.0362266675,0.1640733331,-0.0194177143,-0.1410223693,-0.4276416898,0.0155095616,-0.2663888633,-0.25898844,-0.1859139353,0.1919544488,-0.1180998906,0.0327487402,-0.388687551,0.5425735712,-0.0513567887,-0.040957544,0.1596074998,0.1490183771,-0.1359752566,-0.0729370341,-0.2246603221,0.3244206905,-0.2734502256,-0.2662237883,0.0113274725,-0.2298122644,-0.1351867169,-0.2447400838,-0.4690416157,0.2624190152,-0.0373392962,0.0931429937,0.5576632023,-0.5127329826,-0.0088941325,0.1425562799,0.0489226505,-0.2458119094,0.2287059128,0.1792592257,0.1185104027,-0.0261673741,0.0974640623,0.3858812153,-0.1910532117,0.0223704614,0.192329824,-0.5699858665,0.0100323819,0.4592341781,0.1663509756,-0.0837202817,0.2037396729,0.3062742949,0.3540766239,-0.1438972503,0.0948609412,-0.0039308029,0.1672941148,-0.1847397536,0.1527121812,-0.2712331116,-0.4977675676,0.0717808828,-0.112314485,0.3537005782,-0.2318066508,-0.2829921246,0.2255385518,0.1219486818,-0.1157411411,0.156272158,0.1232983172,0.0082985237,-0.2365090549,-0.1434396505,-0.2550841272,0.1369340122,-0.0850533321,-0.0531746782,-0.5448283553,-0.0864705816,0.1394861639,0.0991121978,-0.2355987877,0.0254585072,-0.1526152939,-0.1169924438,-0.1096147746,0.0343311094,0.3937690854,0.2774084508,0.1349646598,0.698913157,0.2996772528,-0.1463650763,0.169770211,-0.0126307849,0.1328804195,-0.0970049128,-0.3589525223,0.328364253,-0.0470566265,0.2075079829,0.0211817361,-0.0790644959,0.3100651801,0.2122882307,-0.3273974657,-0.3172622919,-0.0217201691,0.1549547017,0.2145112902,-0.08478719,0.1557432264,0.122679837,0.5731419325,-0.1241306365,-0.0248400643,0.0490899459,-0.0492575094,-0.1181121022,0.1767469794,0.1987424344,0.1878615767,0.1950035542,-0.1955697834,0.0530859791,0.2239174992,-0.0549269356,0.2263594717,0.340411365,0.1749640405,-0.2310612351,-0.1177041233,0.1239614859,-0.0467152782,-0.5198426843,0.1980891973,0.2032971084,-0.0531556197,0.1968861818,0.0428790599,-0.0753786415,-0.1284825504,-0.3832616508,-0.1838586777,-0.318559736,-0.0777195767,-0.2470966429,-0.2786354721,0.074968107,-0.2303389013,0.4132597148,-0.2005615085,-0.0441117957,-0.2279595286,-0.5827357769,-0.118052341,0.0277491529,0.1978812218,-0.004313861,0.5062886477,-0.1334144175,0.316726476,0.1148122922,-0.3814626932,0.2815646231,0.0940135941,0.0821937099,0.1088660881,0.0156015446,0.4287016392,-0.1381285638,0.0241017304,-0.1182787046,0.1206660792,-0.1136659011,-0.0545083322,-0.2837180793,-0.0953601524,-0.191307053,-0.4482689798,-0.3104404807,0.0156516116,-0.1775952131,0.2284779549,0.2870112658,-0.2695159912,0.2960008085,-0.1404755265,-0.0400333703,-0.138589263,-0.2160336375,0.0618612468,-0.1857069582,-0.1422276944,-0.188936606,-0.2502924502,0.4162764549,-0.1059164628,0.0160748232,-0.0791459158,-0.0857845247,0.0484033376,-0.0154457148,0.0028956775,0.3300645947,0.3177496791,0.0522625521,0.0510101132,-0.2254760861,0.0482391976,0.1712310165,-0.047595121,-0.0547528304,0.0671824142,-0.2109037936,0.2497275621,-0.1583255678,-0.1090565249,0.1084005758,0.0682086051,0.024877755,0.1149933711,-0.1472264826,-0.1775392592,-0.0431997217,-0.1193582416,0.4011523724,-0.0825181529,-0.3941972852,-0.2258717269,-0.0971387625,-0.2110519409,-0.2165944725,0.3612154424,-0.3785186708,0.0462567247,0.1453927159,0.1834389865,-0.3600265682,0.059144292,0.1387261003,0.0240974091,-0.0424129963,-0.1504551917,-0.4631061256,-0.1951566786,-0.0517478064,0.198721543,0.0929715782,0.1890624315,-0.1264425963,-0.0416370332,-0.0743095875,0.0658165962,0.414519906,-0.4606897533,-0.3246561587,-0.1970992237,-0.0563521236,-0.2238284945,-0.0119615952,-0.4146484733,-0.0935769901,-0.0926216766,0.724352777,-0.226440534,-0.2320034504,-0.0160716362,0.0238381736,-0.2543182075,-0.2290788591,0.0169259515,0.0548014529,-0.3053891659,-0.109762907,0.0045976941,-0.0936541632,0.1399969459,0.0246749967,0.114468962,0.190186277,0.2980965078,0.4336780608,-0.0841347948,0.0968005285,0.0619528554,0.1629604399,0.050567992,-0.100571245,0.2845266759,-0.1551850438,-0.4038881958,0.1885102838,-0.2898671031,0.4289567173,0.2216413915,0.1618853509,0.282668978,0.2553931773,-0.1552065015,-0.5864426494,0.2438239902,-0.1122415066,0.0419226922,-0.7167118788,-0.8676812053,0.6783362031,-0.1185826957,-0.1949784905,0.3085178435,0.2437497377,-0.392591387,0.8725911379,0.3021670282,1.0867829323,-0.0595860705,0.1156732216,-0.2351612002,-0.1452880949,0.5709779263,-0.3308143318,-0.0455745161,-0.3060330749,-0.0262307003,-0.2218384296,-0.0204175618,0.1903684139,0.3359439671,-0.1727037132,-0.0494896881,0.3274557292,0.2788982093,0.0061152363,0.6649950147,-0.226886645,-0.5201141834,-0.036334116,0.1010254249,-0.2963991165,-0.0155749219,-0.1501872391,-0.2149117291,-0.1265394241,-0.0933461785,-0.4119506776,0.0215704683,-0.0986618251,0.1290515065,-0.1357485205,0.0133359516,0.4020500779,-0.1917871088,0.2309774011,0.2535200119,-0.3715169132,0.2342334688,-0.2795985639,0.1197355092,0.0444908254,-0.1328150332,0.2868949473,0.0504314229,-0.2823980451,0.2203228772,0.15207991,0.073009111,-0.3427943885,-0.0461761244,0.1971013844,-0.1608904004,-0.2490368634,-0.0501356907,-0.1760364473,-0.0332085118,0.1139744073,0.0989602283,-0.1846347898,0.1576601863,0.1888440847,-0.0788817257,-0.1221880466,0.0030631362,-0.0402868986,-0.3135068119,0.1186334193,0.0736710578,-0.1119331717,-0.2531694174,0.1199168786,0.4659533799,-0.5444821119,-0.3060614467,0.0613782518,-0.3274764717,-0.0739740133,-0.1839659959,0.2017278373,-0.0748894885,-0.0525118001,-0.2536860704,-0.4518549144,0.3175568879,-0.0403866284,0.3316699862,0.0561204888,-0.4216320217,-0.3096024692,0.2220255733,-0.3124023378,-0.289208591,0.0730077922,0.1892004758,0.1039353013,0.1293196976,0.0590515137,-0.2674793899,0.0378901586,-0.0447059274,-0.0663506165,-0.1580506563,-0.4090871215,0.1509952694,0.0792523623,0.0648069084,-0.0257361233,-0.1340882629,-0.0713403374,-0.3585807383,0.1209126934,0.0309843812,0.0401415192,0.4076748192,0.3948901296,0.1239317656,-0.2358321995,0.2776529491,0.0524910055,0.2577156126,0.1555420905,0.2383891344,-0.0514002889,0.0074200067,0.1084148884,0.3047497272,0.3230584562,0.1960056275,0.0511228032,0.1569581628,0.3583663404,0.2375690639,0.0321676098,0.0781168491,0.0026557266,-0.0093091922,0.2011524886,0.2149125487,-0.1647000313,0.006964846,-0.0446486585,-0.0247125346,0.0519944653,0.3314432204,0.2045866996,0.4511499405,0.0779803693,-0.0186585505,0.5682343245,0.0386715643,0.2201537341,0.4977555871,-0.0212705936,0.2002587914,0.2808569074,0.2378834337,0.2652047575,0.4048501849,-0.1260121912,0.5046952367,0.2107446939,0.1700340062,0.2183267772,-0.1453456283,-0.0928882957,-0.1969637573,0.0332620703,0.4342416525,0.2646166086,-0.0819685683,0.1058478281,0.1069512293,-0.308403492,-0.1003798842,-0.0533567518,-0.1788995266,0.1193139181,-0.1091176346,-0.0850610361,-0.0931362063,0.0827830583,-0.2223093957,-0.0446697064,0.1180314496,0.0882511362,0.1596375257,0.0331230909,0.25953269,0.0425266847,-0.2797541618,-0.0979192853,0.116063185,-0.1958253682,-0.1240559965,0.0431170389,0.0306374431,0.0451051816,-0.1291425079,0.3846647441,-0.1120031327,-0.0914893895,-0.1456621587,0.1750522703,0.2802906036,0.1978929043,0.0188897308,0.1568670124,-0.0474002734,-0.3200163543,0.118593581,-0.0892769769,0.0831352845,0.5165130496,0.3158290088,-0.2402651012,-0.140886873,0.2320820093,-0.3646382391,0.1261700243,0.5776978731,0.0428974517,0.0144046126,-0.023764886,0.0217930153,-0.0066338754,0.4919044673,0.2791668475,0.2031033486,-0.2209550142,0.0476768352,-0.2774269581,-0.0697021335,0.0869194195,0.2521314025,0.0826768354,-0.1852786392,0.4106392562,0.0526027456,-0.1143485531,0.2113519162,-0.059919633,0.1116982698,-0.2882396877,-0.2637892365,-0.0851863697,0.1223352551,-0.0717254356,-0.134153828,0.0446305983,0.268879056,-0.0952054933,-0.7371988297,0.0122702289,0.2557095587,0.0195285305,0.4136569202,0.0326624848,0.3648235798,-0.1785209626,0.1327504516,-0.2633847296,-0.1412115097,0.0188808199,0.2682233155,0.0689026341,0.0588288978,-0.0690946877,-0.1509847194,-0.3108577132,-0.2595832646,-0.339584738,-0.121415697,-0.4853745997,0.0161523502,-0.3182705343,-0.0515386723,-0.0509188324,0.0337888114,0.1769850254,-0.0059504383,0.1084819958,0.0881265178,0.3382861018,-0.1268095225,-0.4318947196,0.0017507498,-0.0258029606,-0.4314211309,0.1465003788,-0.6659142971,0.099572517,0.278773129,-0.1434538513,-0.1094182804,0.3935502768,0.401426971,0.0878676623,-0.2980800867,0.3209183216,0.046121344,-0.1199628636,-0.3783833981,-0.1940392852]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2347","title":"Add an API to access the language and pretty name of a dataset","comments":"What dataset_info method are you talking about @julien-c ? In `huggingface_hub` I can only see `model_info`.","body":"It would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.","comment_length":16,"text":"Add an API to access the language and pretty name of a dataset\nIt would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.\nWhat dataset_info method are you talking about @julien-c ? In `huggingface_hub` I can only see `model_info`.","embeddings":[-0.2759426534,-0.1491751969,-0.0873624384,0.4150775969,0.3421386778,0.1189722791,0.3897387385,0.3709770143,-0.0223603975,0.3746368587,0.0779034197,0.5090548396,-0.0601509586,0.3841827214,0.3461498022,-0.2023470253,-0.1172480658,0.0129679348,-0.0803340152,-0.1211486757,-0.1962136924,0.0521035045,0.1108366102,0.1241701245,-0.2669221163,0.0454564132,-0.0540004075,-0.1388418525,-0.0790893584,-0.3375242352,0.0865435675,0.2850354612,0.2838588655,0.0294624493,-0.0001076396,-0.0245080348,0.4776881635,-0.1515951008,-0.3618721366,-0.2217716128,-0.3296572268,-0.1733139902,0.2418285757,-0.4034040272,-0.1884512007,-0.2138314545,0.224421382,-0.1580801308,0.3432016373,0.0513024628,0.2172973156,0.3994908929,-0.102676332,-0.1373630464,0.04761355,0.6418023705,-0.2143192291,-0.0689423978,0.2645703852,0.1362203658,-0.1340632737,0.486395061,0.2200396359,-0.1540040821,0.0501422063,0.1936683506,0.0280212779,-0.1995519847,0.1938809752,0.298142463,0.525097549,-0.2898886502,-0.4724097848,-0.3455622494,-0.0765761286,0.0059020873,-0.0632784516,0.1038327143,-0.078157559,0.3310469985,-0.0015482671,-0.182741642,-0.3698796928,-0.0884884149,0.3872871697,0.6222913861,-0.0667792186,0.2156384438,-0.1351896524,-0.0614886768,0.2825260162,-0.1690822244,-0.1277253777,0.2868182361,-0.0393238179,-0.3108914495,0.2393340319,0.1192906275,0.0675930381,0.1376781464,0.1428592801,0.3573163748,-0.1809933037,0.0888535008,0.1385019571,-0.0307538137,0.3093574345,0.2264119238,0.2360432446,-0.021527959,0.3387371898,-0.2190620899,-0.3292751908,0.4511210024,0.0249057077,0.2304005325,0.1044783145,-0.0164831672,0.1443421096,-0.0354278199,0.2760775387,-0.2522013187,0.178229019,0.2969044745,-0.0513974875,0.3812257051,-0.1183512956,-0.1153667048,0.0954156965,-0.3546554446,-0.1189889982,-0.0383488499,-0.1936372966,0.2752449512,0.0766616315,0.3325634599,0.1666308194,-0.1525792032,0.0196104068,0.2120028734,0.3784129322,0.0297197718,0.1751953512,0.0140853906,-0.1419986784,0.1474352181,0.0596461184,-0.0874045566,-0.4299757481,-0.0136393746,-0.382239908,-0.3169518113,-0.2568475008,0.1982749104,-0.0872393548,0.0501367599,-0.3038825989,0.6148662567,-0.0784490779,-0.106677115,0.1258229613,0.1114142835,-0.1736579835,-0.0588228069,-0.0618404336,0.4257553816,-0.2679534853,-0.3829319477,-0.0307526831,-0.3184658289,-0.047787156,-0.10053844,-0.5002002716,0.2181773335,-0.0935139284,0.0518310741,0.4971801341,-0.3273154199,-0.0425875671,0.1693952829,-0.01667629,-0.0804448128,0.2692882121,0.1571595073,0.053384535,0.0325234272,-0.004432918,0.2725387812,-0.1532925218,0.0294047911,0.0630578995,-0.5404416919,0.054611031,0.3931621611,0.0242334809,0.0829235017,0.1160224527,0.2529343665,0.4412276149,-0.1962568462,0.0433011539,0.0485788845,0.242580533,-0.1425728947,0.1200500354,-0.2552195489,-0.5282456875,0.134233579,-0.2967631817,0.3904132843,-0.2090803534,-0.3232308626,0.1848082244,0.1415601373,-0.1126807183,0.0030877152,0.1554501951,0.0829647332,-0.0735130161,-0.2124800831,-0.2460973412,0.1851912141,-0.0752023384,-0.0761938393,-0.6570014358,-0.103659898,0.0326333903,0.0452943072,-0.2922037244,-0.0061010495,-0.0503631309,-0.2164345533,-0.1234729812,0.0844233334,0.295961678,0.1233161539,0.0977291688,0.5998771787,0.2218786627,-0.1326135248,-0.0450715944,0.0476418138,0.2004657686,0.029669065,-0.3212014735,0.4076057076,0.0848569646,0.1917535216,0.1073310822,-0.0558085777,0.3379788995,0.2143078893,-0.2994888425,-0.317826122,-0.0228854213,0.2004351169,0.2086059302,-0.0387231223,-0.0429479368,0.12272688,0.6142543554,-0.0771296024,-0.0394029319,0.1776136607,0.0620752759,-0.0580253229,0.2085754126,0.1728255749,0.2795028389,0.2470294684,-0.2898241282,0.0175849069,0.2250059992,-0.042623952,0.2176297903,0.2541165948,0.1007624194,-0.2975418866,-0.0693697482,0.0942350775,-0.0805080384,-0.5052796602,0.1470464617,0.1081088036,-0.1146179438,0.2134939879,0.0048204483,-0.1865454167,-0.2260833979,-0.2593690157,-0.2423830926,-0.2321575135,0.0639260337,-0.3443917036,-0.2731078863,0.167504862,-0.1685518026,0.4209560454,-0.1939249188,-0.0794449151,-0.327447623,-0.4993188977,-0.1878132522,0.0564041473,0.111486353,-0.0859057009,0.4333528876,-0.1752043664,0.2551008463,0.0451659784,-0.4706531763,0.3559008241,-0.0337875858,0.109446831,0.2159413844,0.1232699975,0.3101499677,-0.0756680965,0.1584525108,-0.209682107,0.118352443,-0.0982742012,-0.0985588655,-0.2147697955,-0.1783460081,-0.2231878638,-0.357454896,-0.3391286731,0.1538420618,-0.1468399763,0.2661872208,0.3840101659,-0.1879553944,0.3901928663,-0.1318159252,0.0069877296,-0.1764029115,-0.281493336,0.0871337429,-0.2605356872,-0.2236509025,-0.2518216968,-0.2001709491,0.3316057324,-0.1917027384,-0.0048782853,-0.1415728331,-0.3200156689,0.0396235846,0.0485229306,0.1005735472,0.5203320384,0.1714390367,-0.0038117054,0.0036428815,-0.3489682674,0.1289702058,0.185622558,0.0381870158,-0.0448597036,0.1224044189,-0.2703688741,0.3509935141,-0.1466285139,-0.1390264928,0.2506451309,-0.022420099,0.1232724488,-0.020252917,-0.175989911,-0.1975342631,-0.0261259209,-0.1281015426,0.2923987508,-0.0998625979,-0.2114877254,-0.2865624428,-0.1411036402,-0.1324397624,-0.2634537816,0.4139840901,-0.1196353734,0.0964192003,0.1225680709,0.1635970771,-0.3575393558,0.098944746,0.0790403336,0.0718184412,0.0586111359,-0.1438412517,-0.4632593095,-0.20924595,-0.0405965596,0.3120363951,0.0798461065,0.0989732519,-0.2552241981,-0.1933975071,-0.0437651686,0.0099864053,0.3893160224,-0.297493279,-0.4550481737,-0.1842501611,-0.2364064753,-0.3232527375,0.0257844366,-0.4025108814,-0.2186630964,-0.0348210782,0.6307066083,-0.3656851053,-0.3308067322,0.0633178428,0.0039728582,-0.3565931618,-0.1163438484,-0.1900060475,0.0624508969,-0.3366025984,-0.0849280059,0.0200806428,-0.0210251436,0.1621296853,0.0979931876,0.1038385108,0.1954334676,0.1701419652,0.5530716181,-0.0327245556,0.0767157376,0.0087067289,0.2072020471,0.1496104151,-0.1514511257,0.365521431,-0.1082000136,-0.3777086437,0.2084170431,-0.0988678485,0.4041383862,0.3018050492,0.121410735,0.3060575128,0.3197751641,0.0132703576,-0.5712265372,0.2327616662,0.0038012664,0.0152422963,-0.5912832022,-0.7295432091,0.6037890911,-0.1806968302,-0.1662048399,0.2758647203,0.4441757798,-0.3743599057,0.8279211521,0.2387399375,1.1571503878,0.0049295896,0.1082107201,-0.2190335095,-0.1763715744,0.6383492947,-0.1340519488,-0.0354390591,-0.3163116276,-0.1272507459,-0.1556326151,0.045439519,0.2552592158,0.2740697265,-0.2787871957,0.0618556589,0.1432453543,0.203065455,-0.0726199746,0.5821612477,-0.2201788723,-0.5712998509,-0.2741285563,0.1341062039,-0.2298972756,0.1336826086,-0.1209980026,-0.3156326711,-0.0603222921,-0.2010242194,-0.364135921,0.0349948406,-0.1870691776,0.0755529553,0.0728012845,0.1158524454,0.34650594,-0.0590805933,0.1576420069,0.2652761638,-0.4127757549,0.2081745863,-0.3386718035,0.1951896548,-0.0009970294,-0.1399055421,0.3260091543,0.0263080429,-0.4010033906,0.2274025828,0.1553508043,0.0811642483,-0.4271915257,-0.0589178316,0.1183510572,-0.135100916,-0.1909099668,-0.1581843495,-0.2536539435,-0.1162066534,0.1612607241,0.0736358017,-0.20750615,0.0647428855,0.2558290958,-0.0246776789,-0.1540743411,-0.0130586643,-0.1613878012,-0.1768429577,0.1780763417,0.0551510975,-0.057113912,-0.2974298,0.1033768281,0.55762887,-0.6720743775,-0.2136091888,0.1543538868,-0.3262690604,-0.1795612276,-0.0015338825,0.148967281,-0.089567773,-0.0993056148,-0.1393262893,-0.5680269003,0.2761219442,0.022277277,0.3341533542,0.022573892,-0.2911904752,-0.1335535049,0.3174354732,-0.366687119,-0.3855105639,-0.0382177494,0.1187954843,0.2027061582,0.0907611772,0.2256063968,-0.1537210345,0.0361881815,0.0366791114,-0.0618629418,-0.1893775165,-0.4390673041,0.129111737,0.0757048726,0.1053548381,-0.1097318456,-0.0448511317,0.0111414865,-0.3443035781,0.2489362359,0.0468478128,0.1269065887,0.543554306,0.2113695741,0.0683460385,-0.2815256417,0.1473039985,0.1519031078,0.3099833727,0.2464330941,0.0900694802,-0.2311067879,-0.0207084026,-0.011920332,0.395129472,0.4228619933,0.1954633892,-0.0516204312,0.1905959696,0.2433227748,0.1549430639,0.1229506508,-0.0251097549,0.0028265747,0.0069449353,0.2330272794,0.2194883525,-0.1700420678,0.0364912078,-0.0324669331,-0.1786352545,0.1470065862,0.4730040431,0.0629318729,0.4394761026,0.1933121681,-0.0391317643,0.3884008825,-0.0180234518,0.1633432508,0.4092825055,-0.0543864928,0.2335223556,0.4315808415,0.1384379417,0.290599525,0.3131816089,-0.141236186,0.5300723314,0.0180794485,0.1420877874,0.3447794616,-0.0635802671,-0.0474326722,-0.2558323741,0.0265271682,0.4020937383,0.1237141788,0.0359302647,0.0616868325,0.0532746315,-0.3657339215,-0.0872216672,-0.1554146111,-0.2191476226,0.0888074115,-0.0220284462,-0.0962613076,-0.1248002127,0.1701055616,-0.0733878687,0.0472960733,0.0660973042,0.0869513005,0.1095827073,-0.0615011938,0.325879395,0.1874179542,-0.2220154256,-0.0190381836,0.1866034865,-0.2159181088,-0.1455266774,0.2004791945,0.0651551634,0.06976033,0.0586056672,0.3083477616,-0.064299725,-0.0909113288,-0.146963194,0.2606359124,0.1961841583,0.1402208805,0.0009805966,0.1745384634,-0.131776914,-0.2461850941,0.0526569374,0.1050315648,-0.0519278534,0.5253123641,0.1650741398,-0.2007721514,-0.2089076191,0.1433084011,-0.434869647,0.1977425367,0.4484319687,0.0286457762,0.0409162343,-0.1360032707,0.0782456696,-0.0245528612,0.6048408747,0.1333010644,0.0883971676,-0.2785167694,0.1097171754,-0.3128061891,-0.0625596046,0.1769056767,0.1858273447,0.1176982373,-0.1474124491,0.3364085257,0.0009133685,-0.1275829822,-0.025956016,-0.042730324,0.0922177508,-0.2509528399,-0.154651165,-0.2082225978,0.0799684003,0.0380616263,-0.1675682813,0.0793036148,0.2785617709,-0.0724616945,-0.686901927,0.0119714011,0.1952445358,-0.0633666068,0.2783082128,0.125090763,0.4131275117,-0.2456312776,0.1188131198,-0.2510003448,-0.1941519827,-0.0302020516,0.2311796397,0.1099976003,0.0896770731,0.0272130277,-0.3131050169,-0.4556752145,-0.1158591211,-0.1330257803,-0.1714403331,-0.3130781054,0.0191309098,-0.2408618629,-0.1093681604,0.0500591211,0.1687130928,0.1611702442,-0.1018982977,0.043099232,0.0226521622,0.3717010319,-0.1529971361,-0.471778512,0.0047616973,0.0835169926,-0.4209587872,0.0268653836,-0.7214283347,0.1415534019,0.1885473281,-0.1172914505,-0.1767595112,0.4244628847,0.3451029956,0.1086066589,-0.2637030482,0.4320712984,0.1148653552,-0.0227838121,-0.223690033,-0.1537883282]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2347","title":"Add an API to access the language and pretty name of a dataset","comments":"hmm the equivalent method in `datasets` (which could go into `huggingface_hub` at some point)","body":"It would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.","comment_length":14,"text":"Add an API to access the language and pretty name of a dataset\nIt would be super nice to have an API to get some metadata of the dataset from the name and args passed to `load_dataset`. This way we could programmatically infer the language and the name of a dataset when creating model cards automatically in the Transformers examples scripts.\nhmm the equivalent method in `datasets` (which could go into `huggingface_hub` at some point)","embeddings":[-0.3680251837,-0.035552416,-0.1518368125,0.2967908084,0.3060852289,0.1962565929,0.3240097761,0.3253296316,0.087283209,0.3536399305,0.1184640452,0.4867266119,-0.1382242441,0.4719510674,0.3167955577,-0.2252021879,-0.0660580769,0.0260168929,-0.0123070562,-0.0528531,-0.2736876607,0.0310785528,0.1902249306,0.0862634778,-0.1520760059,0.0071770218,-0.0066345958,-0.2180203795,-0.1954889745,-0.3322659135,0.0439915247,0.3177286983,0.3046394885,-0.0019296911,-0.0001068616,-0.0696629509,0.3831649721,-0.0986470282,-0.3452882469,-0.2134877741,-0.2982555926,-0.1532998085,0.2293319404,-0.3321766257,-0.1240756139,-0.2666774988,0.2220786959,-0.2615324259,0.2814106047,0.0167006645,0.2252594382,0.3396882415,-0.275313586,-0.1248062253,0.0269638225,0.633112669,-0.199988842,-0.0174285341,0.252566725,-0.0088960193,-0.1500615478,0.3573870063,0.2171346843,-0.2105234563,0.1585272402,0.1879028976,-0.0590561107,-0.1842231303,0.0889706314,0.2734551132,0.5650496483,-0.3309367299,-0.5644130707,-0.3185558915,-0.0753830969,-0.0539302714,-0.0194823761,0.1382933855,-0.0080589177,0.3583204746,0.1026079804,-0.2196923643,-0.3568755984,0.0016661769,0.3241778612,0.5333986282,0.0662472248,0.1499340385,-0.1072284505,0.0201892238,0.2803978622,-0.1854182631,-0.0573717281,0.2712131739,-0.0297683012,-0.2591710389,0.1622039378,0.078534618,0.0021666407,0.2023764849,0.1699711233,0.3230954409,-0.2065614015,0.1319925189,0.1250026971,-0.0685240626,0.3235705197,0.1477289349,0.3172385395,-0.0980507284,0.4315104485,-0.1832234561,-0.3647798002,0.4938082695,0.0280048624,0.175017193,-0.0070224204,0.0299963634,0.1697029173,0.0163015351,0.3119880259,-0.2369177043,0.084548682,0.2463130951,0.0714767501,0.4490157962,-0.0907731652,-0.0070609632,0.1614465564,-0.3447469771,-0.1196521223,-0.0325230286,-0.2009731233,0.2717501223,0.0854225829,0.3866281807,0.1827501655,-0.1852335781,-0.0533221439,0.1727667153,0.4882667959,0.0798892975,0.0947406143,0.0343732871,-0.1242163405,0.0556065813,-0.0096812025,-0.0614245087,-0.3933309317,-0.007127549,-0.1642850786,-0.2493080199,-0.2373835891,0.2440943569,-0.0994494855,-0.020009907,-0.3665913343,0.5723448992,-0.0796224102,-0.061179474,0.192571491,0.0972823799,-0.1677733809,0.0137874922,-0.1762911379,0.3450471759,-0.285681963,-0.3903023303,0.0109745404,-0.3977089226,-0.1601867825,-0.1402354389,-0.4790296555,0.1778436899,-0.0494677797,0.0817715004,0.5297881365,-0.2967269719,-0.0017605101,0.105602622,-0.0418994427,-0.1322196871,0.2235865742,0.2228494734,0.1158136204,0.0325738937,0.0390398018,0.3577947021,-0.1366805583,0.029193284,0.1140810698,-0.530636549,0.093121171,0.4786497951,0.1130470783,0.0567130521,0.0897560194,0.2256623805,0.3201318383,-0.2247365713,0.0352878198,-0.0502256714,0.2166538537,-0.1710066497,0.1145795807,-0.2528070211,-0.5673567653,0.0771520138,-0.283911109,0.3759111464,-0.1256531626,-0.3268595636,0.2285159826,0.1517625004,-0.1373387277,0.1435526609,0.1493249834,0.0619705133,-0.1947816461,-0.1896431446,-0.3053133488,0.1335832775,-0.1311443597,0.0013650913,-0.5493671298,-0.1441169232,0.0674896538,0.1183326542,-0.2485745549,0.0182711706,-0.1951661855,-0.1701793224,-0.1033648625,0.0313952528,0.3702283502,0.1957342476,0.1383496821,0.6464543343,0.3155127466,-0.109535329,0.0258641466,0.0636382699,0.1373928636,-0.0073815789,-0.3357206285,0.4083741307,0.050063353,0.1801996529,0.0272855982,-0.0988618582,0.2947947383,0.2184789479,-0.3287475705,-0.2778521478,-0.0166214705,0.1795458943,0.228635639,-0.0034521173,-0.024443198,0.1105321571,0.5334147811,-0.1428393722,-0.0068150228,0.19717215,0.0577787235,-0.1258722246,0.1510443538,0.1931411922,0.210497424,0.2303768843,-0.2380103022,0.0501608364,0.2176095843,-0.0730912089,0.2241094261,0.3024942577,0.0829804242,-0.2374754548,-0.0523425601,0.1053141132,-0.1641715169,-0.5079719424,0.126243785,0.1236367226,-0.0383033566,0.15601331,0.0654195473,-0.1302739382,-0.1332456023,-0.199933812,-0.2029634565,-0.2872855365,0.0568242185,-0.3045387864,-0.3660069704,0.1280496418,-0.1622783095,0.5079782009,-0.212865293,-0.074907802,-0.2986730039,-0.6184287667,-0.1499622613,0.0701080561,0.1223047972,-0.0684886947,0.5011821389,-0.109010525,0.2616670132,0.0702737495,-0.4411019683,0.286698699,0.0160910022,0.0880644098,0.158023268,0.1444804221,0.3501608372,-0.1447490156,0.0855710506,-0.1813066304,0.0719918087,-0.0902647898,-0.1001979932,-0.2445299327,-0.1407452524,-0.2203695774,-0.3846755922,-0.3935340345,0.1136918962,-0.1702356189,0.2049326003,0.2982764542,-0.1946869195,0.2927332222,-0.1368301213,-0.0238556862,-0.1583251208,-0.2252918035,0.0621358976,-0.2183223665,-0.2107007653,-0.2737297416,-0.2505477965,0.3809813559,-0.1373119205,-0.0004025562,-0.1076471955,-0.1751169413,0.0948550701,0.0236622915,0.0125211636,0.3688835502,0.2971470356,-0.0325562321,0.0563375391,-0.2598154545,0.1156632826,0.2472844273,-0.0542724691,-0.0648236871,0.0834747404,-0.2805164456,0.343282789,-0.1477194726,-0.1717998385,0.2754822373,0.0471777655,0.0906752273,0.0165043492,-0.1515355259,-0.2000186592,-0.0171331614,-0.0457236655,0.3483969271,-0.1341115981,-0.3336562216,-0.2158038765,-0.1384945065,-0.1423364282,-0.2033449262,0.4105173945,-0.2624240518,0.1075735241,0.1589523405,0.1162145734,-0.3104801476,0.1439140886,0.14428702,0.0174697675,0.0449387394,-0.1610216945,-0.501794517,-0.1875893325,-0.0635669678,0.245858565,0.0854938924,0.1466901302,-0.1857710183,-0.1823737472,-0.0915700048,0.0446881615,0.4494254291,-0.3508166969,-0.4701811075,-0.2350138575,-0.1669451445,-0.2837841511,0.0526609197,-0.3574932218,-0.2052765787,-0.0733999237,0.6108876467,-0.2656170428,-0.3010943532,0.0326369703,0.1143676713,-0.3196981847,-0.1658059359,-0.115610227,0.0375717431,-0.3636305034,-0.0463627018,-0.0436286703,-0.0621681064,0.1639201641,0.01257258,0.0983244032,0.179638043,0.2133537382,0.5282828212,-0.0536407828,0.0363731645,0.0456325598,0.1562178731,0.0496574789,-0.1797400564,0.3147396445,-0.1476307362,-0.3837729096,0.2408501804,-0.1626692265,0.3979377151,0.2299193591,0.1368302107,0.2969046831,0.2942723036,-0.0371340401,-0.5465805531,0.2344856113,-0.0919105783,0.0910681486,-0.5677520037,-0.7539510131,0.6067982912,-0.1610238999,-0.1764906794,0.2845115662,0.3725021183,-0.3317753077,0.9597309232,0.272493422,1.0739361048,-0.0848384723,0.1396097988,-0.1723479629,-0.096190989,0.6217882037,-0.2569561303,-0.0578274615,-0.2863363326,-0.0817717686,-0.1951461732,0.0125887366,0.1735803634,0.2861514091,-0.2212015688,0.0121116694,0.2382904142,0.2042289823,-0.0249509942,0.6904378533,-0.2295961976,-0.5797513723,-0.1509587467,0.1486731321,-0.241521433,0.0791051537,-0.1766870767,-0.2255636305,-0.0206753872,-0.1295952797,-0.3293623328,0.0227844156,-0.200044781,0.0915677398,-0.0456728525,0.1076989919,0.3874123693,-0.0608497523,0.1641775072,0.212797299,-0.3460482955,0.2270107567,-0.3329410255,0.1793019176,0.0129147274,-0.1485571265,0.3282196224,0.039296519,-0.3293558061,0.184805125,0.1632613093,0.1101750061,-0.3533958495,-0.0505334809,0.0941840857,-0.1908586472,-0.2123963237,-0.114140816,-0.2285723835,-0.1323099881,0.1649833322,0.0344831944,-0.1680629104,0.0545408092,0.30472821,-0.0577421114,-0.1657471061,-0.0006666267,-0.139812544,-0.252104193,0.1255599409,0.0941034481,-0.0763560012,-0.3552132845,0.0855173916,0.5314562917,-0.6359408498,-0.1862015128,0.1318677217,-0.4156267941,-0.1823753715,-0.0325161032,0.2169919759,-0.0916247591,-0.0485157222,-0.1876969934,-0.4991033673,0.296446979,0.0706433803,0.3821381927,0.0716775581,-0.3681830466,-0.2093350589,0.245501101,-0.3735681474,-0.3550162613,-0.0221779365,0.1496612579,0.2193430662,0.0557120629,0.1655662805,-0.1966049373,0.0771777406,0.0189793017,-0.0751258656,-0.2227002531,-0.4356565773,0.1331580579,0.0609120317,0.1079930514,-0.0414886437,0.0151416557,-0.0265874751,-0.3518209159,0.197747618,-0.0666080341,0.0339313485,0.4414327741,0.2275703698,0.0732498467,-0.1984057426,0.1584715992,0.1050488874,0.2506375015,0.2254651785,0.1440544873,-0.1091434807,-0.0219266936,0.0481481291,0.385599345,0.3897276223,0.195670113,-0.032641273,0.213138774,0.2462243289,0.1858932376,0.0978042409,0.0315072611,0.0179076456,0.0477484986,0.1679240316,0.2662694752,-0.232397303,0.028913781,0.0091274269,-0.1585342288,0.1634850502,0.3877548575,0.1611148715,0.445625037,0.1759949476,-0.0695051923,0.4177449346,-0.0311470926,0.2396049052,0.3668141067,-0.026498219,0.1803301126,0.3357956409,0.2088882625,0.2622889876,0.2981678843,-0.1451648325,0.5021366477,0.0476546139,0.2092818171,0.2789627612,-0.0992056429,-0.0316151157,-0.2134975493,0.010543271,0.4210825264,0.2479662895,0.0237332936,0.0599106587,0.0438610874,-0.3583818078,-0.0258244071,-0.086730808,-0.2247735858,0.0449679457,-0.0828998908,-0.0754020363,-0.0902183056,0.1827233136,-0.1393024772,0.0371807814,0.1355399042,0.1294522882,0.0783875063,0.0184035506,0.2691107094,0.0728795454,-0.2044600844,-0.0659626201,0.2457884401,-0.2111194432,-0.1652362198,0.1463949233,-0.0032451842,-0.0033172579,-0.076191999,0.4241475165,-0.1282701045,-0.1022258773,-0.1910683066,0.2044961452,0.1922947168,0.2714376152,0.0446709059,0.1913745403,-0.1323662847,-0.3367794454,0.095231913,0.0072948467,0.0341398008,0.5695911646,0.2894803882,-0.2388479263,-0.2365601659,0.1838847846,-0.4037019312,0.1803047359,0.4729957879,0.0050785695,0.0423502028,-0.1004720479,0.0699505284,-0.0204775725,0.5861886144,0.2237776965,0.1793128401,-0.2617933154,0.11813052,-0.2588198185,-0.0715953112,0.1413943619,0.2715402842,0.1081627533,-0.1823727787,0.3663530946,0.0555097423,-0.1057191566,0.12723355,-0.1030505672,0.0393907949,-0.2703351676,-0.1543967277,-0.0989169702,0.1099215075,-0.0112798084,-0.1567997485,0.0500373654,0.2179257125,-0.0623150803,-0.7347015738,-0.0037037632,0.1749933809,-0.121363543,0.3861846924,0.0719247013,0.2960927188,-0.2171062976,0.112337783,-0.1695034653,-0.2164759487,-0.0347989984,0.2291823179,0.0912461281,0.1428077966,0.004374336,-0.2434690297,-0.3581710458,-0.2253588438,-0.2177964896,-0.1662863195,-0.4074644446,0.0773547739,-0.2452638894,-0.0903428122,0.0234287325,0.1331742704,0.1880894005,-0.0823205337,0.050726369,0.0503492206,0.327755332,-0.1737489253,-0.4083665609,0.0615468882,-0.036791116,-0.3223434985,0.0888489857,-0.6640225053,0.1371552348,0.1973090172,-0.1092551649,-0.1683815271,0.4782003462,0.3514437377,0.1177451983,-0.2872995436,0.3952622712,0.0411661156,-0.1040241569,-0.338345468,-0.2194025666]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2345","title":"[Question] How to move and reuse preprocessed dataset? ","comments":"@lhoestq @LysandreJik","body":"Hi, I am training a gpt-2 from scratch using run_clm.py.\r\n\r\nI want to move and reuse the preprocessed dataset (It take 2 hour to preprocess),\r\n\r\nI tried to :\r\n\r\ncopy path_to_cache_dir\/datasets to new_cache_dir\/datasets\r\nset export HF_DATASETS_CACHE=\"new_cache_dir\/\"\r\nbut the program still re-preprocess the whole dataset without loading cache.\r\n\r\nI also tried to torch.save(lm_datasets, fw), but the saved file is only 14M.\r\n\r\nWhat is the proper way to do this?","comment_length":2,"text":"[Question] How to move and reuse preprocessed dataset? \nHi, I am training a gpt-2 from scratch using run_clm.py.\r\n\r\nI want to move and reuse the preprocessed dataset (It take 2 hour to preprocess),\r\n\r\nI tried to :\r\n\r\ncopy path_to_cache_dir\/datasets to new_cache_dir\/datasets\r\nset export HF_DATASETS_CACHE=\"new_cache_dir\/\"\r\nbut the program still re-preprocess the whole dataset without loading cache.\r\n\r\nI also tried to torch.save(lm_datasets, fw), but the saved file is only 14M.\r\n\r\nWhat is the proper way to do this?\n@lhoestq @LysandreJik","embeddings":[-0.2037422657,-0.1724879295,0.0268734302,0.3251445293,0.2044835836,0.2692605257,-0.031548433,0.2629524767,-0.1478934884,-0.1774816811,-0.0785922483,0.1326426715,-0.1410595328,0.0142101115,0.3060074747,-0.1791174859,0.1607154459,-0.0127882138,-0.2236266434,0.0363891013,-0.0876350999,-0.0407485589,-0.0578159653,-0.1901448071,-0.3203300834,-0.1126613244,-0.26538077,0.1856277585,0.0888046324,-0.2042430043,0.1529036909,0.2086126506,0.371784687,0.5639864802,-0.000124839,0.0764158517,-0.3017306626,-0.1859615892,-0.2845976949,0.0063411905,0.3748137057,-0.1054427922,-0.145716846,-0.1478052139,-0.4197368324,-0.3773116767,-0.1443356425,-0.5287983418,0.5374683738,0.245860368,0.0382170901,-0.4078690112,-0.2978642285,-0.0193990115,0.0391579233,-0.0714992508,0.0461683758,0.1882416606,0.3121959567,0.0798146352,0.2954444289,-0.0017390234,-0.1582686454,0.1989332139,0.1065374613,-0.2691984773,-0.2758403718,0.0165232532,0.2386970222,-0.0432179309,0.428688854,-0.550098598,-0.017777741,-0.2704801857,0.0573882982,-0.5959832072,-0.3207006752,0.1671169698,0.0948769227,0.4144234657,-0.3603258431,-0.5639224052,-0.025134379,-0.3872517943,0.2103292942,0.1642148644,-0.0116572268,0.2137106657,-0.005654858,0.3518165648,0.2906157374,0.0445327573,-0.2772253752,0.1719689667,0.0326810591,0.1142993644,-0.1864017397,0.266726464,-0.0464955084,0.4047196209,0.0266900603,-0.2525605857,0.3430196941,-0.197506085,-0.040205773,0.2442400604,-0.2051400244,0.2982929051,0.0139638912,-0.3265722096,-0.4057730138,-0.2980324328,-0.3589153588,-0.1293609887,0.3829704821,0.3398736119,0.0428095125,-0.1984188408,0.0318860151,0.072004661,-0.0532349572,-0.3289200664,-0.1549503356,0.1246782616,0.0008079647,0.0842313468,0.2572757006,0.0833331719,-0.0951300859,-0.4230962396,-0.1609107554,0.1740496308,-0.3163066506,0.3192405105,-0.0386611335,-0.086428076,0.2452008426,0.0591847673,0.0546834357,-0.1092542335,0.4917868078,-0.1168196276,0.1755405217,0.1816837341,-0.0915065259,0.1738700867,0.2889578938,0.4091693759,-0.2392396033,0.5512588024,-0.2484468818,-0.2810080051,0.4954839945,-0.057399217,0.0836624503,0.1513405889,-0.2178062052,0.1132414266,0.3672728837,0.048496794,-0.0329513997,-0.2560317218,-0.170861274,-0.2792309821,0.2227985561,0.0445028171,-0.3925894201,-0.0627989098,0.3297199309,0.2338337153,0.3675372899,0.2388240397,-0.2528591752,0.474999398,-0.2176841199,-0.3385119438,0.4861726165,-0.2923808396,-0.0215509646,0.0613918416,-0.0583870038,-0.0313797444,0.0246582683,0.1168971583,-0.2042223513,-0.266448468,-0.3556271195,0.496357739,0.0213241428,-0.0101616411,-0.280069232,-0.2093247622,0.2209414095,-0.2495112121,0.0489906929,0.1227238625,-0.0255096145,0.0094047943,0.2300142497,-0.132817179,0.1104627401,0.1681937277,0.1763862222,0.1822406501,-0.089075774,0.0263703167,-0.2529900372,0.0775227174,-0.1778028458,-0.4124350846,0.3950369358,-0.1360621005,-0.244375512,-0.1842935383,-0.187776953,-0.1884699166,-0.0733592361,0.1853184849,0.5064260364,-0.2330239713,-0.1911128163,0.4028528035,-0.1136337593,0.0489118136,0.0420634374,0.1884287745,0.0141248973,-0.1507340521,-0.1047518924,-0.0963275433,0.3396182358,-0.2363124043,-0.0564035475,0.3245693743,0.0947458446,0.0577574447,0.1166906208,0.2510468364,-0.1127326414,-0.1965842098,0.2850088775,-0.0936760604,-0.0514469892,-0.076889582,-0.2179377526,0.3493608832,0.127235055,0.1676337123,0.1506398618,-0.2498599738,-0.1014707834,-0.1440264583,0.2415553629,0.2031805515,0.0413916521,0.1590043306,0.3775345385,0.1738455594,-0.1158208996,0.3023030162,0.2799594402,0.1016128585,-0.0386351831,-0.1699267328,-0.1958602816,-0.3057859242,0.1205069721,0.0359109938,0.0231777038,0.064110443,0.3708545566,0.2742257118,0.031645041,-0.0623750724,0.0131334402,-0.3560841382,0.4349400997,0.0090112714,-0.0285059605,0.0435290076,0.2908142507,0.4834903777,0.2608244419,0.4479074776,-0.1814387441,0.2282052934,-0.3077180386,0.1190449521,-0.0779858232,0.1779007167,0.1194066554,0.0711013526,-0.19525446,0.366496563,0.4076517224,0.0395413786,-0.1278954893,0.065717198,-0.1572267562,-0.2336248308,0.0454107225,-0.3153561056,-0.1168715283,-0.0510903336,0.3141408861,-0.1052696109,0.3447365165,0.1098658815,-0.0518683158,-0.3686341941,0.2937695384,0.1703063995,0.0507639907,0.1293572932,-0.171444878,0.2300544828,0.0533351116,0.3051926792,-0.033642441,-0.0675596371,-0.0813741907,-0.1840611994,-0.2088778019,0.1758762747,-0.0345539972,-0.2672415674,-0.1675367206,-0.0158603583,0.0902827755,0.1921525002,0.0178319179,-0.2154280245,0.0515539758,-0.2210651189,0.3124235272,0.2273920476,-0.198780939,-0.7538731098,0.0953265801,-0.1479717195,-0.043594718,0.1005572826,0.17168957,0.2611336708,0.2950744629,-0.416205734,-0.1209918633,-0.0671724528,0.1790215224,-0.071904026,0.2136950493,0.3615890443,0.0339532793,0.0429486223,0.0033962044,-0.0288676713,0.3880075812,0.0443103164,0.0969483554,0.2219842523,0.2752407789,0.2128897309,0.7723720074,0.163275212,-0.1174795851,0.2269292474,0.0490236431,0.2644726634,-0.3078288436,0.0495082401,0.1314727217,0.070180729,-0.3050029576,0.1452895999,0.0025596127,-0.0946520269,-0.1295571774,0.0933206826,-0.0394243076,-0.1361430734,0.1506855041,-0.1275052577,0.4249049425,0.1299781203,-0.1360868514,-0.2358926237,-0.0968796983,0.0283524748,-0.1863287836,0.5199896097,0.4029473662,-0.5126817822,-0.0654617622,-0.3681846857,0.3157458305,-0.1118135825,0.0177553035,0.0128629003,-0.1116603166,0.1684498042,-0.238444522,0.1012098193,-0.1835324317,0.092434451,0.3786740601,-0.0278597455,0.0476296246,-0.1456466317,0.028141005,0.2856521308,0.1656172872,-0.1855030805,-0.1865389347,-0.3018854856,0.0470550358,0.281098932,-0.5025234222,0.1166764647,-0.154782638,-0.5233402252,-0.2441644818,0.15462102,0.2958908975,0.0701764449,-0.1344407052,0.3634575009,-0.2022899985,0.0347230956,0.3862490058,-0.091094099,0.3662982285,0.1582044363,0.1547720432,0.0132170841,0.3087052405,0.1560252458,0.3203488588,-0.3949987292,0.4435811341,-0.1788884997,0.0473347567,0.4202805459,0.1437384486,-0.1332038343,0.1353102475,-0.3002368212,0.0509689786,-0.4809866548,0.0347059257,0.0185532346,-0.0781791732,-0.5431815386,-0.7855383158,-0.0365551338,0.4058017433,-0.133058086,0.3685834408,-0.2587459385,-0.3118244708,-0.0158723276,0.2281906903,1.1096663475,-0.3849042654,0.3514353037,-0.0025822367,-0.0989537463,0.150844872,-0.3531893492,-0.0938404351,-0.2853488326,-0.1579939872,0.0114270169,-0.1049266383,0.2980239391,0.2393813729,-0.0857728124,0.4363353848,-0.3984698355,-0.2795701623,-0.072871767,0.2209583819,0.2505064905,0.1623320878,0.355594635,0.0600997806,-0.2601300478,-0.1203506291,-0.1497873962,0.2636909783,0.1426990777,0.2464292198,-0.3293325305,-0.1097229868,-0.6422778964,0.0095779821,-0.5371730328,-0.5418801904,-0.0584469028,0.0277245604,0.6591243148,-0.019553829,0.0593473651,0.219306156,-0.0870590508,-0.0227576699,-0.1203799099,0.1719948947,0.2975162268,0.0344529413,-0.1119437218,0.3145791888,0.0150017031,-0.3532521725,-0.1846017241,-0.3647325337,0.2997435033,-0.2022811472,-0.2930469811,0.5546284318,0.0663830265,-0.1531872302,0.0372969732,0.3012940884,-0.2650881708,0.3291013241,-0.3083041012,-0.2711604834,-0.0612544231,0.0665621385,0.3336153924,-0.3272016943,0.3351078033,0.2273415327,-0.0519020632,-0.0854356959,-0.0041752085,0.2194867283,-0.1630469412,0.261269927,0.0724441186,0.16127792,0.0301478989,-0.1457449794,-0.0148556205,0.0847776756,-0.0529853888,0.1232206449,-0.3189747036,-0.1851715893,0.2105150074,0.334517777,0.2113431245,0.0180148166,-0.0801691636,0.3630075455,-0.1871197522,0.2131416351,0.1167738587,0.2090552896,0.1731694937,-0.0149762649,-0.1457211822,-0.101484783,-0.0340913422,-0.1155992821,-0.0280044563,-0.1342687607,-0.3054619431,0.221489042,-0.1279334128,-0.0007056993,0.0571934879,-0.5830870271,0.0283490214,-0.2783137858,0.0596866012,0.2825347781,0.0895804986,0.0960265994,0.3960859179,-0.3068305254,-0.0959295779,-0.0587040298,0.1168596298,-0.0694138631,0.0783706307,0.161950618,0.191434741,-0.1132181659,-0.1865689605,-0.0690296069,0.4618935883,0.07844726,0.073121883,-0.151887536,0.0142666474,-0.0987955332,0.1756359786,0.0752671286,-0.0327069461,0.2617372572,0.4847383201,-0.0543433875,-0.3114573061,0.233027488,0.2590301335,-0.1613175273,-0.1862479895,0.4517156184,0.3919871747,0.4477859437,-0.2221024781,-0.0819137022,0.4238718748,-0.1886423379,-0.0040112576,0.0646130741,0.3197864294,0.0000191297,0.2516067028,0.4281054437,0.0541470274,0.0130510265,-0.0470898002,0.3980367184,0.3786913455,-0.0635832176,-0.1610403657,-0.4522519708,-0.1009056941,0.1210686043,0.1296036541,0.2532998323,-0.2896716595,0.2196651846,0.3302277923,-0.6784617901,0.1657744199,0.5863934159,-0.0934614763,-0.2720733285,-0.2799411118,-0.0597388446,-0.126886189,0.1779456139,-0.075611487,-0.3054000139,0.0904622525,0.0139110275,0.1328616291,-0.0604757182,0.1270985007,0.0136205712,-0.0089429785,0.0709026754,0.4314080179,0.0672563985,-0.1901019812,0.0773522556,-0.0143186282,0.1632647067,-0.0600573383,-0.2245257646,0.2516721785,0.0487945005,0.1767394692,-0.1776888967,0.0284280088,-0.2321080416,-0.1489962488,-0.1164279059,-0.053503532,0.0577366911,0.0203472618,-0.227629438,-0.0197945349,-0.1001541615,0.4848842323,0.3583626747,-0.3412519097,-0.0041220351,0.4337994754,-0.295822978,0.1552665532,0.7841237783,-0.0435909964,0.0482570566,-0.0948641896,-0.0017825268,0.3293181956,0.2038672119,0.2150726765,0.0269077644,-0.2487292588,-0.3539150357,-0.505223155,-0.0300479513,-0.258543402,0.0935067981,-0.1221058816,0.025329005,0.2965248823,-0.1230019033,0.305164814,0.0759470612,0.2798596025,0.2413270026,-0.2273377627,-0.3116932213,-0.1137630716,0.4301926792,0.0185147319,-0.2269172668,-0.0510482118,0.414873898,-0.117249608,-0.3136877418,0.0039053052,0.0861420259,0.197144419,0.3696238399,-0.0797255635,-0.0073159635,0.1522541195,0.1385730356,-0.0773031637,0.261254251,-0.0415900238,-0.1998658329,-0.3113097847,-0.0026868235,-0.6894416809,-0.2891818285,-0.3405935168,-0.2973170877,0.1273684353,-0.0173066035,-0.3378043473,-0.3073319793,-0.1306822747,0.3457747102,0.3317645788,0.4195410013,-0.2065376341,0.0163962692,-0.4942605495,0.1641459912,0.6596114635,-0.0249125287,-0.3328962922,0.0188382864,0.0237659458,0.0954892114,0.0257299077,-0.1929258704,0.1273186207,0.3199720085,-0.0614460707,-0.1536313146,0.405164957,-0.3205523789,0.1034705415,0.0581033714,0.2845223546,-0.020578241,-0.2213194817,-0.4385995567,-0.066765815]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2345","title":"[Question] How to move and reuse preprocessed dataset? ","comments":"Hi :) Can you share with us the code you used ?<\/s>\r\n\r\nEDIT: from https:\/\/github.com\/huggingface\/transformers\/issues\/11665#issuecomment-838348291 I understand you're using the run_clm.py script. Can you share your logs ?\r\n","body":"Hi, I am training a gpt-2 from scratch using run_clm.py.\r\n\r\nI want to move and reuse the preprocessed dataset (It take 2 hour to preprocess),\r\n\r\nI tried to :\r\n\r\ncopy path_to_cache_dir\/datasets to new_cache_dir\/datasets\r\nset export HF_DATASETS_CACHE=\"new_cache_dir\/\"\r\nbut the program still re-preprocess the whole dataset without loading cache.\r\n\r\nI also tried to torch.save(lm_datasets, fw), but the saved file is only 14M.\r\n\r\nWhat is the proper way to do this?","comment_length":28,"text":"[Question] How to move and reuse preprocessed dataset? \nHi, I am training a gpt-2 from scratch using run_clm.py.\r\n\r\nI want to move and reuse the preprocessed dataset (It take 2 hour to preprocess),\r\n\r\nI tried to :\r\n\r\ncopy path_to_cache_dir\/datasets to new_cache_dir\/datasets\r\nset export HF_DATASETS_CACHE=\"new_cache_dir\/\"\r\nbut the program still re-preprocess the whole dataset without loading cache.\r\n\r\nI also tried to torch.save(lm_datasets, fw), but the saved file is only 14M.\r\n\r\nWhat is the proper way to do this?\nHi :) Can you share with us the code you used ?<\/s>\r\n\r\nEDIT: from https:\/\/github.com\/huggingface\/transformers\/issues\/11665#issuecomment-838348291 I understand you're using the run_clm.py script. Can you share your logs ?\r\n","embeddings":[-0.1601244807,-0.326133579,0.0653828979,0.3396508694,0.2864599824,0.1715729535,0.1354581118,0.1812243313,-0.0809112191,-0.1558374763,-0.1377898008,0.0398773029,-0.168959111,0.0587198995,0.2985677719,-0.2673764825,0.1552250385,0.0188227296,-0.3201619685,0.0475619994,-0.0599813052,0.0041698474,-0.0590553656,-0.2007913291,-0.4416160882,-0.0313769318,-0.248014912,0.1666658074,0.1828492731,-0.1862577051,0.25329265,0.2420495898,0.3576463163,0.6828821898,-0.0001243633,0.1335567683,-0.2728155851,-0.2317272723,-0.2766866386,0.033128757,0.3927209675,-0.1214862019,-0.1350308955,-0.1829705387,-0.4399235547,-0.3136433661,-0.1345021874,-0.4278991222,0.7499090433,0.1695947349,0.0427683331,-0.2132157981,-0.2778227925,0.008503221,0.0278573241,0.0635983124,-0.0142037226,0.130135268,0.2992685139,0.0561226234,0.2704514265,0.0200520474,-0.1339302659,0.1400889754,0.1526925415,-0.2002475262,-0.2576467991,-0.0023741957,0.2037418187,-0.0177668817,0.3946386278,-0.5167760849,-0.0736816972,-0.3228442371,0.0029749095,-0.5202156305,-0.2885060012,0.1219025403,0.0930978507,0.4745025337,-0.42115587,-0.4792868793,-0.1270094216,-0.423194319,0.2090489119,0.1758548915,-0.0334332548,0.2210970074,-0.0002789216,0.3428712487,0.246133998,0.0327330157,-0.2068687677,0.2302808762,-0.0386438072,0.0416807681,-0.185777992,0.1895509958,-0.0354911052,0.3463308215,-0.0219855066,-0.2044128329,0.3041285872,-0.2163733989,-0.0310898088,0.3699341118,-0.238204971,0.369102031,0.0616072305,-0.2548559606,-0.3396961093,-0.2820080221,-0.3452650607,-0.0796073601,0.3917407095,0.3445879817,0.1377707124,-0.1579306573,0.0967462808,0.0325179249,0.030502595,-0.3261357546,-0.1352859437,0.19699651,0.0150934337,0.1261731684,0.2938515246,0.0826364085,-0.1128118411,-0.3416502774,-0.1811539978,0.1455423385,-0.3331011534,0.3481995463,0.0287016276,-0.014869987,0.2117708921,0.0153675238,0.1050676927,-0.1169733703,0.4738585651,-0.079364568,0.1513180733,0.2509043515,-0.1366546899,0.2494835854,0.2912724912,0.3461928368,-0.237950623,0.5287489295,-0.2154992372,-0.3498134613,0.4890762269,-0.0687368363,0.028009437,0.1787005514,-0.2850988507,0.0660907924,0.4531784952,0.0426065885,-0.0363986418,-0.2588700056,-0.1559437811,-0.2766834199,0.2264140844,0.154808268,-0.3293088377,-0.1583404541,0.3445459604,0.228521198,0.3519313633,0.386734128,-0.299557656,0.4961073995,-0.2172493935,-0.3410748541,0.4171503484,-0.3627423346,-0.0831839815,0.0610206202,-0.0586275943,0.0295937695,0.0395397469,0.0728784502,-0.1641704142,-0.3265634775,-0.3916639984,0.525267899,0.0657280236,0.0954927728,-0.3008169234,-0.2864235342,0.2404656112,-0.1850553602,0.0782072246,0.164751485,-0.1351279318,0.069657661,0.265425086,-0.1466569453,0.1503960043,0.211293295,0.1797007769,0.1057514101,-0.0425681919,0.0228968654,-0.4002119899,0.1264652163,-0.2417495549,-0.2840818465,0.3042552471,-0.1635408103,-0.2996501029,-0.1877946258,-0.2450942993,-0.2375672013,-0.0899938345,0.1676106304,0.4852058291,-0.2807754874,-0.2159505934,0.4022931755,-0.0248985793,0.0942259729,-0.0614655539,0.2730467618,-0.0532378256,-0.1499190778,-0.1567530632,-0.1414903849,0.3739397228,-0.3398053944,-0.1057071835,0.3445952237,0.0315981284,0.1384757161,0.0972635448,0.3569538295,-0.0082065947,-0.2292221934,0.2719107866,-0.0102203768,-0.0274838805,-0.0911909416,-0.3177952766,0.3725720644,0.0262950454,0.1716701537,0.0985738561,-0.2230465859,-0.1204782575,-0.2000549436,0.1363731623,0.2361083478,0.0083200103,0.160623163,0.3446005285,0.2063460797,-0.0403262451,0.2403444201,0.3612610102,0.0553271919,-0.0649843514,-0.1507966071,-0.228865236,-0.3063418269,0.0669332594,0.0615004003,0.043263942,0.0668413118,0.2860628664,0.2623815238,0.0553233512,-0.0461747982,0.0767881572,-0.2649209499,0.4027986526,0.0500800721,0.0191394705,0.0634204149,0.2220074236,0.4308409095,0.2273504883,0.4133346081,-0.1975356042,0.2897772193,-0.2299524993,0.1958617717,-0.1645824909,0.1251373291,0.0235758014,0.106869705,-0.2166393995,0.3630876243,0.3512009382,0.0673116222,-0.00158157,0.069393672,-0.0673962459,-0.2549570799,-0.0874466449,-0.2556751966,-0.0744591281,-0.0861398056,0.2902083993,-0.1208374947,0.2638791203,0.1115595102,-0.0741840675,-0.2830797732,0.2449823618,0.2341542393,0.0082205543,0.1375828236,-0.1480499357,0.3238554001,0.1407762319,0.25285092,-0.0139525691,-0.1429397762,-0.0660809055,-0.1644589752,-0.1860044897,0.1651820093,-0.0793921351,-0.2665188909,-0.1722411066,-0.0440107808,0.1693967134,0.1464495957,0.0157912131,-0.0370964408,0.04608858,-0.228190586,0.3308419287,0.1997729689,-0.1972014308,-0.7081502676,0.1870312095,-0.0855588317,-0.1360951811,0.0187754184,0.1196089908,0.2030193955,0.2234844118,-0.3637008369,-0.1337093264,-0.1479481161,0.0738007575,-0.0395294949,0.2716372609,0.4804903269,0.0577302165,0.0834795535,0.0219519809,-0.0803537294,0.3806947172,0.1315556765,0.1364147067,0.19788526,0.2309724987,0.1304398179,0.8657284975,0.1493235081,-0.0853927806,0.1692263782,0.0596273579,0.2112223059,-0.3018945754,-0.0521552004,0.0147803249,0.0333520807,-0.3499486446,0.1703591347,0.0205531754,-0.1342290193,-0.1104608029,0.0983526185,-0.0135286013,-0.2050616294,0.1740505248,0.0571834594,0.416400522,0.1700700521,-0.0592019819,-0.2642399967,-0.0906216055,0.0867497101,-0.1253446639,0.5051723719,0.3344731033,-0.4985047281,-0.0593768433,-0.4051256776,0.2841690779,-0.1172834933,0.0056697009,0.0332439281,-0.1058059633,0.2011013031,-0.2289655954,0.2696990967,-0.1669823974,0.0831260383,0.3571746051,-0.1333175153,-0.0036285482,-0.1654502898,0.001246225,0.2042503357,0.2525528371,-0.0918646753,-0.2378341556,-0.3394630253,0.0686186925,0.2582443655,-0.5208461881,0.0637169853,-0.2125194669,-0.4796688557,-0.3496029675,0.13028723,0.3499252796,0.0401254334,-0.1907970607,0.3724321127,-0.1583038718,0.0489155985,0.3282653391,-0.0834723487,0.3856498599,0.1348904669,0.1237285957,0.1112386584,0.1829808056,0.0165107902,0.2914616168,-0.2944533229,0.3888118565,-0.1182244271,0.0383458249,0.3628846109,0.2960481644,-0.0848832875,0.1007577553,-0.1963428706,0.1107583568,-0.5120970011,0.0160590261,0.075787276,-0.061205741,-0.5460215807,-0.7416172624,0.0027431922,0.3119477928,-0.1413746774,0.3256017268,-0.1791395396,-0.3482351899,0.1249273792,0.1337278336,1.2701575756,-0.3682627976,0.3932271302,0.0407306813,-0.0616525337,0.2148334682,-0.3995699883,-0.0484813452,-0.3540832698,-0.1910413057,0.041421473,-0.1402025074,0.27013129,0.1966874301,-0.1686205417,0.4666837752,-0.3877736032,-0.1828290373,-0.0821986571,0.2741085291,0.3067518473,0.0062169395,0.2844866812,0.0441115238,-0.2560626268,0.0312994197,-0.1887556165,0.1766507924,0.149213016,0.1573364735,-0.3251625597,-0.0547240563,-0.5776564479,0.1004327312,-0.3510814309,-0.454256773,0.025712125,0.0467069633,0.6525635123,0.0427269116,-0.0445942879,0.2635794878,-0.220193103,0.0209863409,-0.2038311809,0.1770691276,0.3226625919,0.0799309909,-0.1761853844,0.232417196,-0.0096874069,-0.3173390627,-0.1525359899,-0.4009660184,0.2816938758,-0.288767606,-0.2285507321,0.5524759293,-0.0141875409,-0.1786886901,0.0377793796,0.3203521073,-0.3139811456,0.3353334069,-0.2060263157,-0.2728105485,-0.052052632,0.1235836595,0.3762610257,-0.3565504253,0.4092620909,0.2852985561,-0.0503522381,-0.0718017966,-0.0397052653,0.1862610877,-0.4198502302,0.2662210166,0.0312094241,0.0564384423,0.0017469855,0.0117727416,-0.0020350455,0.0972583368,-0.1192916557,0.1641535312,-0.4360486865,-0.1915821433,0.2404498309,0.4087908864,0.2104061097,0.053677734,-0.0588932671,0.3609116077,-0.2057773918,0.145569399,0.0549994744,0.1449221373,0.2417033762,-0.2105190605,-0.0113021815,-0.1670223027,-0.0560569502,-0.0815702304,-0.0990906358,-0.1287226975,-0.3700258434,0.2282631695,-0.0982034057,-0.062337365,-0.0429419689,-0.5152864456,0.0373097695,-0.3559055328,0.1376357526,0.2939298451,0.1237008274,0.1071033701,0.3744119406,-0.306448251,-0.1304825991,-0.0919971913,0.1106681824,-0.1535521001,0.1488151401,0.1349716932,0.1269028932,-0.1172866821,-0.2406548858,-0.0245024953,0.5484562516,0.1106055677,0.208498463,-0.1628067344,-0.0436511599,-0.1176139116,0.1929164976,0.0388988256,-0.1015439183,0.2350192666,0.3835322261,-0.0437623784,-0.3034848869,0.2242760509,0.2838372588,-0.1523146182,-0.1874622703,0.4942164123,0.3747985065,0.4605586529,-0.1184173822,0.02667832,0.5552828312,-0.2129226625,0.0199746303,0.0100018717,0.2258826792,0.0183153749,0.3231091201,0.364128828,0.0960739702,-0.0058233831,-0.0262358207,0.2425688654,0.216564253,-0.026168583,-0.165641278,-0.3985560536,-0.1074305326,0.1001915336,0.0502805896,0.244231984,-0.2618632317,0.2348312736,0.2859803736,-0.5793049335,0.0439931005,0.6533359289,-0.0878018886,-0.2688229382,-0.1439164579,-0.0232388917,-0.1741877049,0.145249024,-0.0670853034,-0.2645215392,0.1441343725,0.0096609918,0.1244768128,-0.0039135306,0.1286540329,0.0638973415,0.073581852,0.0244560037,0.3757623136,0.1357917041,-0.183997333,-0.0558844656,0.0121547636,0.1858714074,-0.0888787806,-0.2200862467,0.3250590563,0.056581445,0.1778753102,-0.108105652,0.0461839698,-0.2707103193,-0.0779381767,-0.1502161324,-0.0563367195,0.0050877286,-0.0975059569,-0.2703382969,-0.0184379239,-0.0828581676,0.5797555447,0.2944806814,-0.3680410385,-0.081416063,0.3954348862,-0.3513241112,0.1719429791,0.723590672,-0.0404162519,0.1165964752,-0.1081167161,-0.0002356541,0.3333337605,0.249860689,0.2139374167,0.065950647,-0.2839468718,-0.2710180581,-0.5490266681,-0.0766463727,-0.1191195995,0.1520472169,-0.0725880489,0.0408806726,0.2885599732,-0.1487170011,0.2767313421,-0.0182299335,0.2915890217,0.1825948954,-0.2095009238,-0.197172612,-0.1070296317,0.4269379377,0.0258569028,-0.2944353521,-0.0081186136,0.4932320714,-0.1112406179,-0.3396423459,0.0601266734,0.0122495918,0.1071740836,0.2979592085,-0.0336137973,0.0103997337,0.1751953959,0.099164404,-0.1063952595,0.1840147525,-0.0318379439,-0.2652281523,-0.2596701682,0.1223709956,-0.6407398582,-0.3083273172,-0.4538687468,-0.3385769129,0.143699497,-0.0420784242,-0.441083461,-0.2471866012,-0.255881846,0.3739536107,0.3258477151,0.398901552,-0.1653431803,0.0093378481,-0.5237102509,0.1157125533,0.786352098,-0.0277510472,-0.4555926919,0.0094989929,0.0352793075,0.0927550495,0.0146480529,-0.2981964946,0.1291433573,0.2376129329,-0.0424108021,-0.2237713039,0.4407900572,-0.321593076,0.0996613726,0.0451737009,0.3537795842,0.0059905606,-0.2401881814,-0.4362029433,-0.0664886311]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2345","title":"[Question] How to move and reuse preprocessed dataset? ","comments":"Also note that for the caching to work, you must reuse the exact same parameters as in the first run. Did you change any parameter ? The `preprocessing_num_workers` should also stay the same","body":"Hi, I am training a gpt-2 from scratch using run_clm.py.\r\n\r\nI want to move and reuse the preprocessed dataset (It take 2 hour to preprocess),\r\n\r\nI tried to :\r\n\r\ncopy path_to_cache_dir\/datasets to new_cache_dir\/datasets\r\nset export HF_DATASETS_CACHE=\"new_cache_dir\/\"\r\nbut the program still re-preprocess the whole dataset without loading cache.\r\n\r\nI also tried to torch.save(lm_datasets, fw), but the saved file is only 14M.\r\n\r\nWhat is the proper way to do this?","comment_length":33,"text":"[Question] How to move and reuse preprocessed dataset? \nHi, I am training a gpt-2 from scratch using run_clm.py.\r\n\r\nI want to move and reuse the preprocessed dataset (It take 2 hour to preprocess),\r\n\r\nI tried to :\r\n\r\ncopy path_to_cache_dir\/datasets to new_cache_dir\/datasets\r\nset export HF_DATASETS_CACHE=\"new_cache_dir\/\"\r\nbut the program still re-preprocess the whole dataset without loading cache.\r\n\r\nI also tried to torch.save(lm_datasets, fw), but the saved file is only 14M.\r\n\r\nWhat is the proper way to do this?\nAlso note that for the caching to work, you must reuse the exact same parameters as in the first run. Did you change any parameter ? The `preprocessing_num_workers` should also stay the same","embeddings":[-0.2378590852,-0.1530073881,0.0416558683,0.3084846735,0.1833169609,0.2134883404,-0.0576088615,0.2500776649,-0.1095818356,-0.2252917588,-0.0464636534,0.1580044329,-0.1260224581,-0.035717044,0.2702538073,-0.1903045774,0.1926396787,-0.0515302345,-0.2087463439,-0.0057137888,-0.0511988364,-0.0350198485,-0.0222177655,-0.2187884301,-0.3487526178,-0.1432226002,-0.2869339883,0.1577489823,0.1409121901,-0.2185288221,0.1496663988,0.1988651156,0.3916561306,0.5650043488,-0.0001259738,0.0803208277,-0.2605284452,-0.1995313019,-0.2637192011,0.0004304107,0.4051909149,-0.1469468921,-0.1489050686,-0.1150196642,-0.4202831984,-0.3849588931,-0.1287567019,-0.5107620955,0.5873053074,0.2024842203,0.0258474611,-0.4192654788,-0.3031598032,-0.0005872123,0.0771552026,-0.1091331542,0.0281405747,0.1772205532,0.2982748449,0.0634795725,0.2317209989,-0.0059604747,-0.1695144475,0.2283945233,0.0896402299,-0.2463833988,-0.2650186718,0.010324467,0.213329792,-0.078211531,0.3787163794,-0.5245104432,-0.0408925042,-0.3021622896,0.0440787487,-0.629689455,-0.2901814878,0.1405390203,0.0898738578,0.332952261,-0.3665982187,-0.5225884914,-0.0271649975,-0.3522252738,0.2370804697,0.1714153886,-0.0160608422,0.2046978027,0.0173268523,0.3591575623,0.3000130951,0.0683678314,-0.2626776993,0.17619057,0.0096514747,0.1386127174,-0.1080907062,0.3031721413,-0.0575353019,0.3915197551,0.0235153437,-0.2391524464,0.3578472435,-0.1992240697,-0.0719343126,0.2576841712,-0.168346256,0.2766534686,0.0102211814,-0.3261593282,-0.4254625738,-0.3058959246,-0.3417910039,-0.1531976014,0.3950912058,0.3623047769,0.0495441109,-0.1994163543,0.0779158622,0.1261309236,-0.0675558746,-0.360421598,-0.1975093037,0.1143200621,-0.0364066064,0.0673463419,0.2185407728,0.1009308472,-0.1091736108,-0.4340987504,-0.1762934029,0.1403093934,-0.3342283368,0.4014991522,0.0100634415,-0.1399188787,0.3073939383,0.1028969064,0.045238059,-0.1445742249,0.4853774011,-0.1454342008,0.2267767638,0.1687609255,-0.0933346897,0.1907972693,0.3086638451,0.4347533882,-0.2161506414,0.50892663,-0.2469209731,-0.3078389168,0.5280898213,-0.0791080892,0.0558473766,0.1770818979,-0.1443105489,0.103744328,0.4253030717,0.0528267846,-0.0352413356,-0.2709804475,-0.1938498914,-0.2852814496,0.2383836806,0.0972619951,-0.3305194676,-0.0629702955,0.3851653934,0.3416414261,0.3617277741,0.2083150893,-0.2422458082,0.4805541635,-0.208283484,-0.3495693207,0.4447630346,-0.2732563913,-0.0126191489,0.0592036396,0.000230851,-0.0096721193,0.0283690915,0.0542139523,-0.1592286974,-0.2620946169,-0.365652889,0.421566844,0.006927344,0.0015507707,-0.3235527277,-0.1850039065,0.2035008967,-0.2905150056,0.0656410754,0.1388631463,-0.034485165,-0.030191265,0.20496957,-0.0819884315,0.0843489841,0.1955619156,0.1572903991,0.1549064219,-0.1008209363,0.0708129257,-0.23960042,0.087539427,-0.1613678187,-0.3961811364,0.4401755631,-0.0977196917,-0.2322435528,-0.2417341769,-0.1962097287,-0.190013513,-0.0740765259,0.2564918697,0.4803428054,-0.2477550954,-0.1338165253,0.4130370319,-0.1129544452,0.0086515453,0.0035676907,0.1617397815,-0.0026037612,-0.1892520338,-0.053903766,-0.1240687817,0.415946871,-0.2461105585,-0.0509943403,0.3192646503,0.0949898288,0.0605702847,0.0752692968,0.2710303068,-0.1237048209,-0.1199661046,0.3089157045,-0.1230340078,-0.0858797804,-0.133708626,-0.2027669996,0.3510611653,0.1217094883,0.1586760283,0.1118556857,-0.29674685,-0.1390333027,-0.122890234,0.2126417309,0.222205013,0.08911933,0.1400203109,0.3727264106,0.1329216361,-0.1442529112,0.3077135682,0.2793326676,0.1369310617,-0.0644071251,-0.162967667,-0.1781316102,-0.342042625,0.1087360233,0.0169055182,0.0321849734,0.0411658995,0.394356817,0.3003914356,-0.0095940884,-0.0722705796,0.0120829102,-0.3623243272,0.4106180668,0.0669660494,-0.0287395343,0.0599024408,0.2683700323,0.4951936603,0.2788164616,0.4521010816,-0.1961208582,0.2168153077,-0.3017246127,0.129743591,-0.1224794164,0.1606664211,0.0639332682,0.0758642182,-0.2171610594,0.3982006907,0.3844769001,0.0536038615,-0.1365617514,0.076228559,-0.14812316,-0.2067567408,0.0052702059,-0.3171885014,-0.1434646398,-0.0825677887,0.2953956127,-0.1293855458,0.3584155738,0.0957578644,-0.076369077,-0.299849689,0.3141582608,0.1587602496,0.0824835002,0.1322542727,-0.1929991841,0.2659243941,0.1027571559,0.2607206702,-0.0364822485,-0.1100326553,-0.0644034594,-0.199487403,-0.2438248992,0.156023398,-0.0177233648,-0.2850745022,-0.0929065049,0.0174159743,0.0426195227,0.2385849059,0.0228475221,-0.1819270849,0.0555911511,-0.261844337,0.304982394,0.2563683093,-0.2394011319,-0.7709587216,0.1479658186,-0.1281965524,-0.0332674049,0.0822986066,0.1963883191,0.2900821269,0.2994524837,-0.4189986587,-0.1111751124,-0.0321101211,0.1402683705,-0.0752459392,0.2125627398,0.3601725399,0.049723994,0.0564321429,0.0076198885,-0.0885587707,0.3789902329,0.06662862,0.1641147584,0.2058381885,0.3059243262,0.1700585186,0.7813290358,0.2020617276,-0.1195316613,0.2247864157,0.070247978,0.2524743378,-0.3247464597,0.0542767346,0.1614908278,0.0351382233,-0.3057978749,0.0849258974,-0.0191720612,-0.1112907231,-0.1250434816,0.1055364683,-0.0307058748,-0.1223670766,0.1600783467,-0.1368754357,0.4417608678,0.1152697057,-0.0597147755,-0.3022227883,-0.1132885665,0.0453083888,-0.213859275,0.528631568,0.4135552347,-0.4654914439,-0.0517196767,-0.3803641498,0.311193496,-0.1160935983,0.0681876987,0.0343908221,-0.1186765954,0.1350325644,-0.2068353593,0.096149303,-0.2219334692,0.0872861966,0.3639413416,-0.0061689895,0.1004083008,-0.2459686995,0.0381249823,0.3307689428,0.2019870728,-0.1875589639,-0.1658796966,-0.3008567095,0.0466714837,0.2752130032,-0.4799245894,0.1115132868,-0.2050670534,-0.4821567237,-0.2187338471,0.2105711102,0.2567538321,0.0695932955,-0.173583433,0.3810070455,-0.1715442687,0.0167105589,0.3543286026,-0.0866811201,0.3993421197,0.1708689481,0.1461695731,0.0603262,0.2983466685,0.1328887492,0.2768137157,-0.4191559255,0.4938594103,-0.1247186586,0.0020691173,0.3709501624,0.194118768,-0.1325176209,0.1068121195,-0.3122163117,0.0476856008,-0.4304374158,0.0203566439,0.0458398163,-0.1062617004,-0.5128148198,-0.8027752638,-0.0159528758,0.3981413245,-0.1304418892,0.4157746136,-0.3875638843,-0.2967139482,-0.0329834484,0.2522015274,1.1055426598,-0.2789765,0.3033004999,-0.0125613641,-0.108507745,0.1800975949,-0.3405998349,-0.0375915729,-0.2607918084,-0.156931743,0.0234983638,-0.1165351272,0.2680090368,0.2212571502,-0.0987249836,0.4781957865,-0.3981011212,-0.2551437914,-0.0945032611,0.2080539018,0.2884462178,0.1694542468,0.4030929208,0.0486780033,-0.2707054019,-0.0789211318,-0.1453622133,0.2496760786,0.1875510365,0.2118333727,-0.3258017898,-0.0970606655,-0.6363190413,-0.0228213202,-0.4688453078,-0.5695709586,-0.0640964955,0.0270000044,0.6920267344,-0.0783740804,0.0521787405,0.1922741532,-0.1054191366,-0.0001253343,-0.1312661916,0.1047302186,0.3144969642,0.0321064331,-0.0976962373,0.2907635272,0.0027100369,-0.3139575422,-0.2378308773,-0.3568241894,0.264089942,-0.2140872926,-0.2281088382,0.5874043107,0.0975055993,-0.1491780281,0.0260644834,0.2976787984,-0.238839373,0.358235091,-0.3314952552,-0.2371261567,-0.0552787408,0.09359359,0.3266553283,-0.3253431916,0.3527347744,0.2297740281,-0.068124935,-0.0492398888,-0.0766159147,0.1966187358,-0.1945143789,0.2431261688,0.0282321703,0.1477132142,0.0749343634,-0.1244105548,-0.020303607,0.1662440896,-0.0029203929,0.1760596782,-0.2919256985,-0.2285665572,0.2335788161,0.3254612684,0.1455575824,0.0357898474,-0.0362041518,0.3847229779,-0.1643302441,0.1703977287,0.1472100466,0.1958282143,0.0967095718,-0.0278392751,-0.1505879164,-0.079514578,-0.0613503754,-0.0719887763,-0.0424630977,-0.1017549261,-0.3052811027,0.2270800918,-0.1314198971,0.0105124088,0.0254078116,-0.5276867747,0.0600914098,-0.2723101377,0.0410042927,0.2522368431,0.0761446878,0.1077260301,0.3695380092,-0.2398174703,-0.0173993874,-0.1128975451,0.0950242653,-0.0766147152,0.1053006649,0.204301089,0.2123811543,-0.1024310291,-0.140551582,-0.0946181715,0.3850333095,0.0994777903,0.0607647263,-0.1666939259,-0.0035701101,-0.0791089982,0.151100412,0.1283086985,-0.0213948451,0.1841116548,0.4707873762,-0.0509520434,-0.3081831634,0.1896889955,0.2716931403,-0.11237894,-0.1997253448,0.4546198547,0.3958207965,0.3837063909,-0.1981132179,-0.0502294116,0.4299533069,-0.2597358823,0.0139861321,0.0014535153,0.3135933876,0.0052336855,0.2828738093,0.3944245577,-0.0108018927,-0.0037894333,-0.0097479485,0.3508373499,0.3881848752,-0.0930477381,-0.2104658633,-0.4430541396,-0.0984277353,0.143086493,0.1050590798,0.2753296793,-0.32741189,0.2162556499,0.3400505483,-0.6805121303,0.1070686951,0.5620388985,-0.1046513468,-0.2644446194,-0.3083596826,-0.0716602355,-0.1136858761,0.211553961,-0.0649633035,-0.2648225427,0.0952799544,0.0017644105,0.0734731704,-0.0537666455,0.0690992549,-0.0216052569,0.0395682007,0.0684899911,0.4542489648,0.0731869414,-0.2653938234,0.0276849829,-0.0323735513,0.1969394833,-0.0131794987,-0.1961666942,0.2902059853,0.1038330644,0.1731046289,-0.1679555178,0.0450798683,-0.2011702508,-0.1919150054,-0.1337240785,-0.0789508671,0.064927198,0.0693725273,-0.2745789886,-0.0363626033,-0.0463298038,0.4864162803,0.341753751,-0.3559111953,-0.0175144337,0.4189008176,-0.280547142,0.1513672471,0.7690737844,-0.0326195136,0.045784086,-0.1132402197,-0.0052768751,0.2971598208,0.2374255359,0.1925991327,0.0036205885,-0.2640182078,-0.3714710772,-0.4844547808,-0.0427697338,-0.2741440237,0.1524680257,-0.1331397891,0.0744177103,0.2540328801,-0.1676154882,0.2727189064,0.0335945562,0.2813007534,0.2858588994,-0.2592865229,-0.2775993049,-0.1373998225,0.4601693749,0.0557294227,-0.145520851,-0.0602986924,0.3954575956,-0.1013860777,-0.3069753647,-0.0261680633,0.0952987745,0.2149173766,0.3513469696,-0.0576541312,-0.0238311291,0.1455117911,0.1376879066,-0.1406626999,0.2595166862,-0.0416241661,-0.1788038164,-0.2974762619,-0.0157369878,-0.6356416941,-0.286619097,-0.4114281535,-0.249758631,0.1543263495,0.0200927891,-0.2996421754,-0.2943605781,-0.1123159379,0.3462572992,0.3360644877,0.4151576459,-0.1947688162,0.0636702254,-0.5413557291,0.1431347281,0.6640558839,-0.0283078998,-0.3156377673,-0.0669046566,0.0613720715,0.112382777,0.081952475,-0.2363597006,0.1080253944,0.2914659977,-0.0750484243,-0.1893652231,0.4150597155,-0.3093854487,0.1259085387,0.0467565134,0.2780730128,0.0212976206,-0.2434836775,-0.4002308249,-0.0611645542]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2345","title":"[Question] How to move and reuse preprocessed dataset? ","comments":"> Also note that for the caching to work, you must reuse the exact same parameters as in the first run. Did you change any parameter ? The `preprocessing_num_workers` should also stay the same\r\n\r\nI only changed the `preprocessing_num_workers` maybe it is the problem~ I will try again~","body":"Hi, I am training a gpt-2 from scratch using run_clm.py.\r\n\r\nI want to move and reuse the preprocessed dataset (It take 2 hour to preprocess),\r\n\r\nI tried to :\r\n\r\ncopy path_to_cache_dir\/datasets to new_cache_dir\/datasets\r\nset export HF_DATASETS_CACHE=\"new_cache_dir\/\"\r\nbut the program still re-preprocess the whole dataset without loading cache.\r\n\r\nI also tried to torch.save(lm_datasets, fw), but the saved file is only 14M.\r\n\r\nWhat is the proper way to do this?","comment_length":48,"text":"[Question] How to move and reuse preprocessed dataset? \nHi, I am training a gpt-2 from scratch using run_clm.py.\r\n\r\nI want to move and reuse the preprocessed dataset (It take 2 hour to preprocess),\r\n\r\nI tried to :\r\n\r\ncopy path_to_cache_dir\/datasets to new_cache_dir\/datasets\r\nset export HF_DATASETS_CACHE=\"new_cache_dir\/\"\r\nbut the program still re-preprocess the whole dataset without loading cache.\r\n\r\nI also tried to torch.save(lm_datasets, fw), but the saved file is only 14M.\r\n\r\nWhat is the proper way to do this?\n> Also note that for the caching to work, you must reuse the exact same parameters as in the first run. Did you change any parameter ? The `preprocessing_num_workers` should also stay the same\r\n\r\nI only changed the `preprocessing_num_workers` maybe it is the problem~ I will try again~","embeddings":[-0.2302794009,-0.1823711097,0.0412170552,0.3404656649,0.1980711371,0.2086258978,-0.0375071019,0.2514610887,-0.1037038416,-0.2087676227,-0.0314362459,0.163025707,-0.1305736154,-0.0440304205,0.2582703829,-0.191184625,0.1755548716,-0.0452623554,-0.1915269196,-0.0165702756,-0.053558208,-0.0251102448,-0.0321363769,-0.2102221549,-0.3548283279,-0.1217458174,-0.2918241918,0.1653580219,0.1479510814,-0.2046285123,0.1572978795,0.1787934154,0.4092198908,0.5932112932,-0.0001262194,0.0605907179,-0.2513768375,-0.1923409402,-0.2729692161,-0.0186120551,0.3915597498,-0.1380922049,-0.1416184306,-0.1292648911,-0.416855067,-0.3540060818,-0.1318838298,-0.5027413964,0.6082751155,0.2023899108,0.0339937508,-0.4183064401,-0.3058912456,-0.0001165216,0.0776552707,-0.1023506001,0.0107616307,0.1995076388,0.3009852469,0.0655802563,0.2533354461,-0.0054526818,-0.191324845,0.2161302567,0.0771678165,-0.2489583045,-0.2419671565,-0.0063001793,0.2172265649,-0.0644972771,0.3854005635,-0.4955865145,-0.0214141812,-0.2836808264,0.0525718816,-0.6237027049,-0.2847483754,0.1535848528,0.0742787942,0.3391981721,-0.3694743216,-0.5197529793,-0.0286423266,-0.3435761034,0.2452744395,0.1527326554,-0.0363954119,0.2196631432,-0.0064444202,0.3711723387,0.3126406968,0.067528598,-0.2623577118,0.173351571,-0.0181058608,0.1303376704,-0.1184555069,0.3414288163,-0.0596794747,0.3688027859,0.0189884286,-0.2177861929,0.3481891453,-0.1931370795,-0.0610701963,0.2535793185,-0.1575440913,0.2674420476,0.0153020546,-0.30593732,-0.4410021901,-0.306471169,-0.3355246484,-0.1766227037,0.4219074249,0.3731341362,0.0536548346,-0.226765126,0.0651227906,0.1151025519,-0.0963107273,-0.3628315628,-0.1881166846,0.1191907451,-0.0267978217,0.061994154,0.2298413664,0.1135953516,-0.1300108284,-0.4259298742,-0.1784644574,0.1209823266,-0.3620635569,0.4047139883,0.0170489047,-0.1473718137,0.2812189162,0.1347008198,0.0576779209,-0.157856673,0.4618590176,-0.1600634158,0.2225682586,0.1569197327,-0.0888811126,0.2168241441,0.3071734011,0.4267371297,-0.2022321969,0.5157366395,-0.2954542935,-0.2999769151,0.5050640106,-0.0835787505,0.0763936043,0.2000036687,-0.141600132,0.0793812647,0.4280877113,0.0587629639,-0.0315423943,-0.2942147851,-0.1964375675,-0.2926189303,0.2295431942,0.0953661278,-0.327676177,-0.0593125261,0.3742058873,0.3352854848,0.3759862483,0.2220870554,-0.2464765012,0.4664205015,-0.2396627218,-0.3466240168,0.4473837018,-0.2679067552,-0.0157277305,0.0745586678,-0.0023964485,-0.0045465534,0.0178567097,0.0399054177,-0.1705620289,-0.2612144649,-0.3734619617,0.4028814733,0.0182685181,-0.0181437545,-0.3461932838,-0.1834283769,0.2144181877,-0.281565994,0.0642997101,0.1546642631,-0.0236602668,-0.0397349149,0.2138565183,-0.07063438,0.0812874883,0.2107388824,0.1581497192,0.1610248387,-0.08663515,0.0636222064,-0.2418338954,0.0950836539,-0.1858502328,-0.3736918569,0.426628679,-0.079285875,-0.2309802771,-0.2579778731,-0.2068706006,-0.1903482378,-0.0772554427,0.2529401481,0.4489644766,-0.2513560653,-0.103517361,0.4261999726,-0.1011625826,-0.006149414,0.0129419537,0.1810733825,-0.0196843203,-0.1941667199,-0.0644103885,-0.1135629416,0.4528129399,-0.2504135966,-0.0720244199,0.3093142807,0.0809951723,0.0377836116,0.0612075031,0.2628931403,-0.124869898,-0.1418714672,0.2970556319,-0.1096230149,-0.0812778026,-0.1403728276,-0.1864896864,0.3085307479,0.1251228601,0.1671189964,0.0939229652,-0.2757715583,-0.1386353374,-0.1613757759,0.2355558127,0.2098315358,0.0903167352,0.1427326202,0.3833736777,0.1253003031,-0.1511943042,0.3049738705,0.2918567061,0.1568562388,-0.065999724,-0.1570335031,-0.181794703,-0.3211259544,0.0839108229,0.0257231202,0.0624967143,0.0498829,0.3670087755,0.2906754017,0.0059052287,-0.084963195,0.0275960341,-0.3611502349,0.4245193601,0.0833941028,-0.0329412408,0.0688961223,0.2673463821,0.4997732639,0.2703295946,0.4532830119,-0.2201969177,0.2037452012,-0.3384391069,0.1429186314,-0.1421448737,0.1770968139,0.0565393642,0.0676854327,-0.2187446505,0.3874480724,0.3927195668,0.0559774712,-0.1153547391,0.0760849863,-0.1575406641,-0.1825377643,-0.0212741699,-0.3034434617,-0.1745171398,-0.0893428773,0.3128996789,-0.1144714206,0.3428845704,0.0956346393,-0.0854871348,-0.3158885539,0.3085270226,0.148235172,0.0799029842,0.1736142486,-0.1999173611,0.2825099528,0.0987995341,0.2641468346,-0.0171636902,-0.1196132898,-0.0731900632,-0.1901337057,-0.2272614092,0.1431858093,-0.02286979,-0.2841074765,-0.1119892299,0.0000451936,0.0224696305,0.2144094259,0.0258827228,-0.1833224446,0.0513681285,-0.2649686038,0.2935570776,0.2573852837,-0.257855773,-0.7665913701,0.1563871354,-0.1181642041,-0.0654659048,0.0767871588,0.2234165817,0.2946231067,0.2762686312,-0.4194144309,-0.0827117041,-0.0344308205,0.1499554515,-0.0958808959,0.2032488734,0.3651020229,0.0533620566,0.0602739044,-0.0092965188,-0.1047998443,0.3742393553,0.0774437636,0.206361562,0.215963617,0.2940153778,0.145108521,0.7868188024,0.195152089,-0.1160394549,0.2585864961,0.0802676603,0.2512440085,-0.3090752661,-0.0128970165,0.171986267,0.0376320556,-0.3038145304,0.0874180794,-0.0287990868,-0.1278553456,-0.130604133,0.0801125988,-0.0449876636,-0.130328387,0.1563013494,-0.1270851344,0.4329031706,0.1149054766,-0.0622673668,-0.2800683677,-0.1100105569,0.0520335101,-0.1749269962,0.5241789818,0.4021216333,-0.4852683842,-0.063607268,-0.3512471616,0.3219734728,-0.1157842427,0.1028265581,0.0073922537,-0.0797646269,0.1573293209,-0.182096377,0.1464356035,-0.2514882982,0.081182681,0.3863942921,-0.0053782617,0.0993621573,-0.2538025975,0.053465981,0.3308380544,0.2037239224,-0.181319505,-0.1752060354,-0.3217464089,0.0434069782,0.2783793211,-0.4798920453,0.0575096011,-0.2167820632,-0.4627769589,-0.214927569,0.1868020892,0.2578516304,0.0962295532,-0.1862252504,0.3677944839,-0.1546339989,0.0286895838,0.3485646844,-0.1175281107,0.407396853,0.1658473611,0.147004351,0.0444776416,0.3160403073,0.1348305047,0.262426585,-0.4285261035,0.4940423369,-0.1070592925,-0.001958438,0.3777117431,0.196923852,-0.1340242475,0.1191707477,-0.3015602231,0.070352599,-0.4203509688,-0.0222119391,0.0624336824,-0.0916472003,-0.5123933554,-0.7761762142,-0.014444028,0.3869948983,-0.1343574226,0.4250030518,-0.4090653956,-0.3004296422,-0.0322500244,0.2574757636,1.1520472765,-0.249940142,0.2839636505,-0.0163691621,-0.1121745035,0.1752405763,-0.320735693,-0.050877627,-0.2835557163,-0.1752741933,0.0249645896,-0.133909151,0.2799001634,0.2194498479,-0.1174451485,0.4820213616,-0.427449584,-0.2495140582,-0.0907144845,0.2142428756,0.2973355949,0.172482118,0.4000270665,0.0470794961,-0.2730310857,-0.0555628836,-0.145810023,0.2094357163,0.1826082617,0.2097636759,-0.3370374441,-0.090216428,-0.6582045555,-0.026400296,-0.4553880692,-0.5626409054,-0.0016358134,0.0187767614,0.6962363124,-0.0597636066,0.0341452174,0.1885809898,-0.1038896665,0.0123058083,-0.1183058321,0.1131057888,0.3138386607,0.0337910727,-0.1306759119,0.2751539946,0.0025538146,-0.3120924234,-0.2595795989,-0.3384279013,0.2628728151,-0.2288309634,-0.2692021728,0.5659242272,0.1044118851,-0.1746212542,0.0316332616,0.2832493782,-0.2325007617,0.3902266324,-0.3128139973,-0.2564024329,-0.0495731197,0.1170377657,0.3268792331,-0.3238463998,0.3862734437,0.2198897004,-0.0780946091,-0.0562833324,-0.0784553513,0.1917869896,-0.2133915871,0.2346998751,0.0292929858,0.1681252271,0.0750394389,-0.1442066431,-0.0251833834,0.1845805198,-0.0068635028,0.148393169,-0.2823743224,-0.2391581535,0.2126634121,0.3108687997,0.1389663965,0.0435547344,-0.0113830119,0.377479285,-0.169541195,0.1738160402,0.1361582279,0.1900034398,0.1055747196,-0.0430118293,-0.1715748906,-0.0924955681,-0.0621798746,-0.0726877525,-0.0386429876,-0.1044059768,-0.3125816286,0.2272194475,-0.1248397082,0.0186505858,0.0106055634,-0.5460450649,0.0660820827,-0.2595504522,0.0099601364,0.2569547594,0.078603901,0.1129866466,0.3769864142,-0.2383505702,-0.0419708118,-0.0839817673,0.0883619785,-0.0725463629,0.114932254,0.176842764,0.2323459983,-0.0996143073,-0.1672747582,-0.122951448,0.3646799624,0.115470551,0.0624608509,-0.1464221776,0.019738853,-0.0620910861,0.1360793263,0.1175241098,-0.0306683742,0.1765670627,0.485673219,-0.0533962287,-0.3026413023,0.2002395689,0.2910361886,-0.0990786776,-0.2204585373,0.455316782,0.3883110285,0.3621027768,-0.1899762601,-0.0446704775,0.4329750836,-0.2688747942,0.0050528646,0.0129922954,0.3136689663,0.0137779666,0.2958070338,0.3763059676,-0.0367890522,-0.0068096952,-0.0109448032,0.3377029598,0.3985646665,-0.0833692998,-0.1828801036,-0.4347554445,-0.1114668101,0.1532503814,0.0929496959,0.2628058791,-0.3464435935,0.23017928,0.323969394,-0.6916007996,0.1316185445,0.5701608658,-0.1210775226,-0.2360218763,-0.3140922785,-0.0584085695,-0.1190509498,0.1784202158,-0.0688439757,-0.2500265241,0.1047569737,0.0315688737,0.0387094505,-0.0470056906,0.0387098826,0.0005981295,0.0377257988,0.0561169796,0.4386214614,0.0568428785,-0.2471431941,0.0157621894,-0.0118671162,0.1960346401,0.0346037261,-0.181761682,0.2370755672,0.1254266948,0.1816650033,-0.154965356,0.0577238463,-0.1978987753,-0.1778492928,-0.1292394996,-0.0755382776,0.0559551753,0.0750966817,-0.2512834966,0.0002850342,-0.0540964045,0.4733882844,0.342115283,-0.3579433858,-0.0299593173,0.4078529179,-0.2729263008,0.1792832464,0.7854659557,-0.0429317616,0.0463991575,-0.1331852674,-0.0052664289,0.2972895503,0.2528789043,0.2097212225,-0.0023796589,-0.292339623,-0.3610755205,-0.5111787915,-0.0479325168,-0.2610587478,0.1552469134,-0.1150251031,0.058611352,0.2591908574,-0.1830439866,0.3036310673,0.0425810963,0.3051353693,0.2993865907,-0.2836439013,-0.2719522119,-0.1196966544,0.4597982466,0.0529472977,-0.1639258564,-0.0264039375,0.4131957889,-0.1012801602,-0.3161117733,-0.0223771576,0.0793436319,0.2279126644,0.3597213924,-0.0542268567,-0.020811053,0.1456582695,0.1436702907,-0.1295814067,0.2380605489,-0.0272845346,-0.1591938734,-0.27623564,-0.0207706746,-0.6245563626,-0.2775925398,-0.4387020469,-0.2465454936,0.1248994023,0.0334265828,-0.3082304001,-0.3033782244,-0.0933986455,0.355297327,0.334115237,0.417835325,-0.1940508932,0.0660084337,-0.5593668222,0.1297997385,0.6791755557,-0.0191565324,-0.3333543241,-0.0575463623,0.0880826414,0.0938506275,0.0655806735,-0.2357253134,0.0716258734,0.3024847507,-0.0642756671,-0.159628436,0.4308495522,-0.3177461028,0.1543847024,0.065936543,0.2563073337,0.0453030653,-0.2220456451,-0.3671284616,-0.0503213555]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2344","title":"Is there a way to join multiple datasets in one?","comments":"Hi ! We don't have `join`\/`merge` on a certain column as in pandas.\r\nMaybe you can just use the [concatenate_datasets](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?highlight=concatenate#datasets.concatenate_datasets) function.\r\n","body":"**Is your feature request related to a problem? Please describe.**\nI need to join 2 datasets, one that is in the hub and another I've created from my files. Is there an easy way to join these 2? \n\n**Describe the solution you'd like**\nId like to join them with a merge or join method, just like pandas dataframes. \n\n**Additional context**\nIf you want to extend an existing dataset with more data, for example for training a language model, you need that functionality. I've not found it in the documentation.","comment_length":21,"text":"Is there a way to join multiple datasets in one?\n**Is your feature request related to a problem? Please describe.**\nI need to join 2 datasets, one that is in the hub and another I've created from my files. Is there an easy way to join these 2? \n\n**Describe the solution you'd like**\nId like to join them with a merge or join method, just like pandas dataframes. \n\n**Additional context**\nIf you want to extend an existing dataset with more data, for example for training a language model, you need that functionality. I've not found it in the documentation.\nHi ! We don't have `join`\/`merge` on a certain column as in pandas.\r\nMaybe you can just use the [concatenate_datasets](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?highlight=concatenate#datasets.concatenate_datasets) function.\r\n","embeddings":[-0.4773362577,-0.6935091019,-0.0979438573,0.1795870066,0.1399442703,0.324791193,-0.2019497603,0.0913561285,0.0232805461,0.0314714238,-0.5830703974,-0.0147510488,0.2058378458,0.4802639186,0.0973033533,-0.3352275193,0.2061462551,0.0533265024,-0.0682545528,0.1907300651,-0.0057576806,0.1205627769,-0.1383992136,-0.0527564175,-0.4373386502,0.1436782777,-0.5212435722,-0.1436893493,-0.2237620801,0.1499155164,0.3690565825,0.3095445633,0.1582986861,0.4412491322,-0.0001130937,-0.2398640066,0.0831307396,0.091894336,-0.1158550158,-0.2447850555,-0.2960631847,-0.2401031852,0.1332097501,0.1241770536,-0.0557513945,-0.0315558836,0.0795208886,0.0654474273,0.0969701633,-0.1098946631,0.1444479525,0.0797101557,0.2500294149,-0.1734854579,-0.1637144983,0.511146009,0.0108181126,0.1524355412,0.4946920872,-0.0485214777,0.6609489918,-0.0064986385,0.0665581748,-0.0683701783,0.2285611778,0.100775443,-0.0508412197,-0.4745711088,-0.3607732952,0.3082880974,0.5458722711,-0.1255922467,-0.4932384789,-0.2368196696,0.2477696389,0.0213851891,-0.191336289,0.3288097382,0.2086051255,0.0240966771,-0.2064361274,-0.3917130828,0.0532743707,0.1583522111,-0.0697357133,0.1957956254,-0.0513200685,0.0908239409,0.0695555285,-0.2681914568,0.0352073982,-0.1660829335,-0.1516615003,0.1595116407,-0.6084676981,-0.3992604911,-0.2260915339,0.3628484607,0.4802445173,0.2024169266,-0.1985312402,0.0058766617,-0.4315240085,0.2300775647,0.2825003564,-0.2399666607,0.0240399688,-0.2077750862,0.3054992557,-0.1043953821,-0.1197053269,-0.0667931288,0.1038361266,-0.1081261933,-0.456207931,-0.2621081769,0.1290219277,-0.0245840736,0.0039546536,-0.319190383,0.1243011728,-0.070965834,-0.0413121134,0.0606067218,-0.1655995399,0.1167970821,-0.1396435946,0.2302725613,-0.1907328963,-0.2304217815,-0.0756720826,0.1119525135,0.0165523402,0.2893992066,-0.1712617874,-0.494278729,-0.2040009201,0.4133867621,0.4074558914,0.0469176136,-0.0839378759,-0.2842968106,-0.1569360793,0.0950027779,0.2700360417,-0.0443907417,0.0842060596,0.0237631667,-0.1463833451,-0.1203252599,-0.4403831959,-0.2885366082,-0.4382292032,0.0907670334,0.0465172119,-0.0341365039,-0.4737575352,0.4752351642,-0.2497937083,-0.0918000713,-0.0570735969,0.1923232675,-0.1775273681,-0.1139970347,0.1159601584,0.0173621904,-0.0776893198,-0.1454377472,0.1351715624,-0.0703127682,-0.3292372525,0.1861366332,-0.2383997291,0.1550084203,-0.1052752882,0.0049520712,0.3695624471,-0.4302903414,-0.0287413299,0.0589243174,0.1559230536,-0.1439162195,-0.0754961893,0.2700874507,0.043855112,0.1177537665,0.1722823381,0.4885490835,-0.1066983491,-0.1704289168,0.1633426249,-0.2834759057,0.208545506,0.0972488075,-0.031383235,-0.2148171514,0.0863160118,-0.166401282,0.2248512506,-0.0607077293,0.0528883524,0.2243683487,0.0630205795,0.2102553099,-0.0084778657,-0.4905330539,-0.4413351715,-0.0259323455,0.2855248153,0.084995456,0.2845337391,-0.2495342046,0.1780862808,-0.2793402374,-0.0916923732,-0.0079673845,0.0058748634,-0.0218319129,0.1856889576,-0.2826339304,-0.5492193699,0.685990572,0.1156111211,0.0149121145,-0.5255337358,0.7264195681,0.0302156247,0.0821208283,0.0499669015,0.2766754329,0.0578535758,-0.0626276135,0.2358025312,0.1307605207,-0.3658678532,0.2062020451,0.4405407906,0.1315921247,0.3730511367,-0.2377302498,-0.260142833,-0.3811036348,0.0679416433,-0.195644021,-0.360506475,0.2850862443,-0.0207605176,0.3033422232,0.0646377504,0.0476941578,0.0972117633,-0.0369289704,-0.1704229712,0.0372312292,-0.0251717437,-0.0196590051,-0.0002518992,0.1699335426,-0.397913456,0.0302122831,0.1181701124,-0.0063374308,-0.0161388051,-0.0998308435,-0.2815756202,0.0955167264,-0.0223459806,0.0471738093,0.4209526479,0.2107751369,-0.1005109251,0.0912071019,-0.0222610291,0.0151997237,0.0410538353,0.1977774501,-0.0018745521,0.1576871872,0.2379131317,0.0629280582,-0.0465944596,-0.0710174143,0.2590674758,-0.0271979924,-0.1693974882,-0.2094568163,-0.128433004,-0.3241541982,-0.0318349861,-0.3003270328,-0.4572021365,-0.2940026522,-0.1320739985,0.1283115149,0.1783221513,-0.0108237285,0.1779537946,0.5747698545,-0.2819282413,0.2220961452,-0.1566817611,0.2854074836,0.2318284512,0.13720119,0.1110639572,0.1105616465,0.3009609878,-0.0158740133,0.1589298397,-0.4940417409,-0.1534124017,0.0295919552,-0.2387657017,0.2145949751,0.0540904552,0.2275940627,-0.0427714847,-0.3297334909,0.305698663,0.175301984,-0.1237514392,-0.1903065741,0.0014500894,0.1199492887,-0.0652679875,0.0478746928,-0.3397681713,-0.2575618625,0.1969037205,-0.1347544789,0.0366297662,-0.2529940605,-0.1064912751,0.0301943403,-0.3341850042,-0.0353308916,-0.046087902,-0.4317243695,0.2534382939,-0.2205061316,-0.1305094659,0.1344530284,0.0573150516,0.0983774513,-0.3787458837,-0.0478381813,-0.3517192006,-0.2464513332,0.0267123915,0.0777891278,0.3056207299,0.0439632945,-0.0687061101,0.0510463901,0.0427715704,-0.1938399971,-0.1942002624,0.4216268957,0.2125367075,-0.1220898703,0.2178722471,-0.2056950927,0.1473842412,0.644236207,0.2356084585,-0.0413697995,0.0447314195,0.1844355911,-0.1091811806,-0.4813727736,-0.0742047429,-0.1004672498,0.4089356363,0.1773378998,0.3692170978,0.4297561347,-0.3885721266,-0.1469445229,0.0480184332,-0.0226585194,-0.1660974324,-0.1242579296,0.0802994147,-0.282866478,-0.6389855146,-0.0855932906,-0.0426258966,-0.1398236901,0.4228808284,0.083070606,-0.0825103968,-0.3596101403,-0.1980522424,-0.2996658087,0.2538965642,0.2277431041,-0.1285482943,0.037304949,0.3393589854,-0.0337129496,0.1487991065,0.5826120973,-0.1726087779,-0.6704808474,0.0616284274,-0.1711139977,-0.2093910277,-0.0673206747,0.1376402527,0.1414289325,-0.0379880331,0.606541872,-0.192656219,-0.2045238316,0.4182096124,0.260318011,-0.1448494345,-0.137181133,-0.0255755838,-0.0489086248,-0.3120493293,-0.2289447933,0.0596590564,0.2473796457,0.061299514,-0.1433029175,-0.0324929729,-0.1259240955,0.4951268733,-0.2331517637,0.061114572,0.3976516426,0.1118561104,0.218894586,0.0608471669,-0.0251233485,0.5964621305,-0.0073032416,-0.5920336843,-0.0474391542,0.0537922047,0.8099610806,-0.0254008733,0.1866259128,0.6380625963,-0.2588059902,0.124218598,-0.2145996392,0.0865785778,0.2060464323,0.2696506381,-0.7012066245,-0.2455543429,0.5334430337,0.1279528737,0.072125487,0.250633359,0.3512210846,-0.0461604744,-0.1374610066,-0.0891829804,0.9792351723,-0.0136232581,0.0646688864,-0.0848006681,-0.0947196633,0.6272912025,-0.1484618038,0.0552020334,0.0178433955,-0.2724736631,-0.1328746229,-0.0650325939,0.1201246381,-0.1120243967,-0.1766736209,0.1030926779,-0.0423504375,0.326620698,-0.0839351043,0.259812206,0.0555086508,-0.1796266437,-0.3709902763,0.0588785596,-0.0756691769,-0.1263667792,-0.0926461294,-0.0362317786,0.1469322741,0.193890661,-0.2606265843,-0.3029327095,0.3281199038,0.165167138,0.1333445907,-0.1908025444,0.2359549999,0.3386783898,0.3834403157,0.01565657,-0.2581949532,0.0790784359,-0.4021692872,-0.0843715742,-0.3797151744,-0.2739932239,0.1269820035,0.0440067761,-0.2601495683,0.3154331744,0.3332213163,-0.3063253462,-0.3582176566,0.3029268086,-0.1516794413,-0.2119670659,-0.2580427229,0.3263221383,0.3057923019,-0.1407382637,0.0533414558,0.2744378746,-0.2274742424,0.1149603277,0.0177614354,-0.0982254371,-0.0811064392,0.2735075653,0.1508398503,-0.2412650734,0.1571580917,0.2811252177,0.025403982,0.0433968119,0.1141719148,0.4336085916,-0.1087922901,0.0388734713,-0.0965190604,0.3796593547,0.0731243193,0.1997227073,0.1604219079,-0.0799412802,-0.0885858908,-0.2232183963,-0.1712340266,0.3421502709,-0.0258371513,0.2213172615,0.0329446644,0.1406092048,0.1405458003,0.5654829741,-0.2208259553,0.0864147618,-0.1965106428,0.0217312388,0.1606549919,0.115095377,-0.258548826,-0.0973560587,-0.0471022725,-0.0568879507,0.104708083,-0.1868646741,0.0041204751,0.1633819044,0.2580339909,-0.0123333503,-0.0900284052,-0.0745777115,0.0345583446,-0.1989393979,0.1217484698,-0.0164493993,0.083584033,-0.197871089,0.3910062611,-0.3001542985,-0.2309037894,0.1493074745,0.3866788447,0.1070403606,0.0033815498,-0.0618900172,-0.1058141664,-0.0933873132,-0.0233909991,0.1882751137,0.361615032,-0.1570625305,0.0098911393,0.1190665588,-0.0215581004,0.0221303497,0.4316721261,-0.0803105757,-0.0918823928,0.6225711703,0.0248998497,0.0419758894,-0.1596770287,-0.212771222,0.2212358117,0.1080699638,-0.0306184497,0.0542711392,0.092487216,0.2969582081,-0.1719502658,-0.0238207076,0.1018379405,-0.0110292854,0.2880906165,0.3324660361,0.2076010853,-0.1269050688,0.4156457186,0.2111039609,0.0739036202,0.1609977484,-0.2167823315,0.0563257784,0.0735724121,-0.0141430208,0.7791044116,-0.0039288872,0.1016911715,-0.174867034,0.0833350867,-0.4081781805,-0.0040062051,0.2450895309,-0.2864967287,-0.3804768622,-0.097484827,0.5531991124,0.0412749313,-0.1027697176,0.342895329,-0.1211675107,0.1395777464,0.1233708411,0.122298494,-0.4584869146,0.2442854941,-0.0641886294,0.2247371823,0.0617756546,0.3395593762,0.6070730686,0.2601694167,-0.0309476573,-0.1975222677,0.2071586847,0.0227769185,-0.1020897776,0.5970635414,0.2625040114,0.0122281313,0.0513061881,0.3758941591,-0.0599457808,0.058875218,0.2854728699,-0.0849190727,-0.0516267978,-0.3078393936,0.2585432231,0.1601927578,0.0167522766,-0.057140436,-0.4110811651,0.1542838067,-0.8882024884,0.2430042326,-0.2121292502,-0.1556624919,-0.0812814906,0.1495216638,-0.2142646313,-0.1669244468,0.5107842088,-0.3032633066,-0.1526392996,-0.0496548451,0.0925268009,0.1310433894,0.3922297657,0.4841098785,0.3431787193,-0.1779951006,-0.3353415132,-0.2935390472,-0.2749537826,0.1368604749,0.1734771729,-0.1628403366,0.0611597598,-0.0481851362,0.0391289964,0.1245373711,-0.1670336425,0.1734158844,0.0174610335,-0.2168399394,-0.1275696009,0.0965036973,-0.0845418572,-0.0809870809,0.246857509,0.4023828804,0.0646210089,-0.1149134114,-0.0456651561,0.2484323978,0.2490338236,-0.3340333402,-0.0683509558,0.1755551994,0.4713202715,-0.0781218186,0.2838208079,-0.0082992585,-0.3150534928,-0.0818116441,0.1970591396,-0.2204046696,0.1018044874,-0.3778835535,-0.0571265705,0.1697275639,-0.1676938087,-0.0246204846,0.1312611699,-0.2086883187,-0.1202334762,-0.1719345152,-0.1216088831,0.2517898381,0.3607232571,-0.0874531269,-0.067593433,0.1652060747,-0.1206565201,0.0616155379,0.0914516523,-0.308652252,-0.3141691983,-0.0062717209,0.2522942722,0.1439670175,-0.5772434473,-0.3370821476,0.2370826006,0.0840223953,-0.0770445764,0.5987442732,0.2726123631,-0.0307158045,-0.2170369625,0.0288372487,0.1116119623,0.0265543461,-0.0034458775,-0.1477662176]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2337","title":"NonMatchingChecksumError for web_of_science dataset","comments":"I've raised a PR for this. Should work with `dataset = load_dataset(\"web_of_science\", \"WOS11967\", ignore_verifications=True)`once it gets merged into the main branch. Thanks for reporting this! ","body":"NonMatchingChecksumError when trying to download the web_of_science dataset. \r\n\r\n>NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/data.mendeley.com\/datasets\/9rw3vkcfy4\/6\/files\/c9ea673d-5542-44c0-ab7b-f1311f7d61df\/WebOfScience.zip?dl=1']\r\n\r\nSetting `ignore_verfications=True` results in OSError.\r\n\r\n>OSError: Cannot find data file. \r\nOriginal error:\r\n[Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/37ab2c42f50d553c1d0ea432baca3e9e11fedea4aeec63a81e6b7e25dd10d4e7\/WOS5736\/X.txt'\r\n\r\n```python\r\ndataset = load_dataset('web_of_science', 'WOS5736')\r\n```\r\nThere are 3 data instances and they all don't work. 'WOS5736', 'WOS11967', 'WOS46985'\r\n\r\ndatasets 1.6.2\r\npython 3.7.10\r\nUbuntu 18.04.5 LTS","comment_length":25,"text":"NonMatchingChecksumError for web_of_science dataset\nNonMatchingChecksumError when trying to download the web_of_science dataset. \r\n\r\n>NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/data.mendeley.com\/datasets\/9rw3vkcfy4\/6\/files\/c9ea673d-5542-44c0-ab7b-f1311f7d61df\/WebOfScience.zip?dl=1']\r\n\r\nSetting `ignore_verfications=True` results in OSError.\r\n\r\n>OSError: Cannot find data file. \r\nOriginal error:\r\n[Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/37ab2c42f50d553c1d0ea432baca3e9e11fedea4aeec63a81e6b7e25dd10d4e7\/WOS5736\/X.txt'\r\n\r\n```python\r\ndataset = load_dataset('web_of_science', 'WOS5736')\r\n```\r\nThere are 3 data instances and they all don't work. 'WOS5736', 'WOS11967', 'WOS46985'\r\n\r\ndatasets 1.6.2\r\npython 3.7.10\r\nUbuntu 18.04.5 LTS\nI've raised a PR for this. Should work with `dataset = load_dataset(\"web_of_science\", \"WOS11967\", ignore_verifications=True)`once it gets merged into the main branch. Thanks for reporting this! ","embeddings":[-0.085158661,-0.0791746303,-0.0687736422,0.2122479528,0.1446293741,0.1670502424,-0.0674462169,0.3561598957,0.335711509,0.1703765094,-0.184128806,-0.088294372,0.0320244357,-0.049100548,-0.1792019904,0.3910678029,0.1115438193,0.0704055727,0.1489616632,-0.0231386833,-0.1831050366,0.3623233736,-0.0717558861,-0.1976115257,-0.1742103994,0.2103101313,0.0770297796,0.2272589356,-0.1206633896,-0.2419532239,0.5031519532,0.1476858705,0.2182934433,0.5435404181,-0.0001259768,0.1516777873,0.2565944791,-0.034308359,-0.2887716889,-0.29092893,-0.2795560658,-0.475911051,-0.058786463,0.0108694946,0.1746040583,0.1657092422,0.0882877856,-0.1844337583,0.0431266464,0.3276795447,0.0915033445,0.5787203312,0.2556365728,0.0954884216,0.2089751959,0.1225741282,-0.0354046486,0.3028368354,0.3427023888,0.2210127562,0.1164619476,0.060722854,-0.2214396,0.3517247438,0.1020267382,-0.0368372649,0.173568368,-0.5107841492,0.2583927214,0.5678054094,0.4979865849,-0.0043832846,-0.419526428,-0.1657646149,-0.3733310699,0.0157464128,0.3665995002,0.2591349781,-0.3134577572,0.0133686997,-0.3301946521,0.0845313892,-0.0772109628,0.3896190524,0.0180250891,0.2835505307,-0.0295883548,-0.0571471676,0.11614912,-0.1018674225,0.5316739082,-0.4382450879,0.2005492896,0.144161582,-0.4212915897,-0.1766135991,-0.1564654708,0.5054700375,0.3802501559,0.0973420963,0.1199492291,0.1752368659,-0.3893419206,0.1672763377,0.0753296688,0.1147176474,-0.0474486351,0.0135330446,0.312235117,0.2439682484,0.0297439005,0.1369536817,0.1577558666,-0.3242730796,0.3243071139,-0.2313977182,0.3200098574,-0.4231874049,-0.3306280077,0.0906516463,-0.3126273453,-0.0955943987,0.1481835991,0.3303675652,-0.1278497875,0.4152513444,0.0071039302,0.1833061278,-0.0802410394,-0.1089758426,-0.2528930306,-0.0673437864,0.0680782795,0.2231645435,0.3387229145,-0.0464915372,0.5633104444,0.0471912175,0.2553281784,-0.024210358,0.064512372,0.119226411,-0.06533131,0.5076720119,0.0377697907,0.1763928682,0.19255732,-0.0206301846,-0.1292277277,0.1545487642,-0.2998304069,-0.3466961086,-0.0361652412,0.0729087889,-0.5573632717,-0.2970697284,0.0667590573,-0.3345190883,0.1852492541,-0.3027693033,-0.0240202639,-0.1686641872,-0.0076714717,-0.1457055658,0.0374204963,0.2380313128,-0.1447969675,-0.0407205299,-0.0382016487,-0.0833038241,0.2342764288,0.1772887707,-0.0524787046,-0.0786318034,-0.4035301208,-0.0802104995,0.2724952996,-0.1574369669,-0.6148737073,0.0999209955,0.1808791608,0.2084606886,0.2243495136,0.1730824858,-0.3209434748,-0.0814131498,-0.1034110412,-0.1201060042,-0.0156137291,0.0100259213,-0.3111905158,-0.1605876833,0.1613960117,0.1905451119,0.1421376765,0.0388945043,0.0776122063,-0.29169029,0.4403509796,-0.2310882211,-0.0102114165,0.1556785256,0.2587715685,0.2319766581,0.0506166928,-0.1816351861,-0.5557805896,0.3231894672,-0.1367996335,-0.0096906079,-0.0300934799,-0.0773777515,-0.1430769116,-0.2569597661,-0.2845996916,-0.2503167689,-0.0902674571,0.2949232161,0.4509862959,-0.0489331745,-0.1575189382,0.0379108451,-0.1419367343,0.1865469217,-0.356816113,0.3374642432,-0.0452543609,0.0145251537,-0.0272784624,0.0376796983,0.1270903647,-0.0440379828,-0.1227161363,0.4065564871,0.0182057507,-0.0507407933,0.1054895893,0.4436122179,0.135043934,-0.1297820359,-0.0214370992,0.3785462379,0.0495000184,0.0335715413,-0.2013786584,0.2474854439,-0.1390593946,0.3544166684,0.0880438387,0.2606643736,0.2448540181,-0.1513050795,-0.1840907633,-0.0671158582,0.4213168025,-0.3561789691,0.0980976373,-0.0824201182,-0.1569029391,0.0316677988,0.0791498795,-0.2353097349,0.0087870592,0.2170219123,0.0629356205,-0.0883164853,-0.0595988743,0.5852566957,0.5754346848,0.1997919679,0.1184806228,0.0807274505,-0.0301167313,-0.1202846467,0.1156509146,0.2258462012,0.0875288695,0.3959544301,0.0566203743,-0.0406732559,-0.4490393698,-0.2895910144,-0.0693226755,0.2049111873,-0.4617797732,-0.0714839995,-0.2673533559,-0.0268006809,-0.4079286754,-0.1423026025,-0.6074799299,-0.5206985474,0.0005134312,0.3735798597,-0.0295475926,0.3360081911,-0.4761890173,-0.0309328362,-0.1708243489,-0.0682737678,-0.1113139614,0.2401206195,-0.0735438541,-0.1028328463,0.456233263,-0.0082454281,0.2831138968,-0.3246206641,-0.0525074378,-0.7112343907,-0.3908772171,0.0480473489,0.0272902772,-0.0165778957,0.3302422464,0.2918686569,0.0696731359,-0.2094524503,0.2388788015,-0.3470227718,-0.275816381,0.1425112337,-0.0270260032,-0.0324203856,0.1690679789,-0.1787805706,-0.2738606334,-0.3106218874,-0.1619167924,0.0988685191,0.1175975427,0.3009375632,-0.0066987467,0.0593076795,0.0564735122,0.2426354736,-0.3133750558,-0.8532565236,0.4654781818,0.0037950312,-0.1866527945,0.2090541869,-0.0538757145,0.2935708463,0.2297284901,-0.4464188218,-0.3201853335,-0.3100651503,-0.1366550326,0.2622306049,0.171727851,0.4347871244,-0.1502236873,-0.0250732712,-0.0867382511,-0.1759123802,0.0649075434,-0.1306042522,0.6317172647,-0.0790114552,0.0577214025,0.0454547815,0.7125914097,0.3863265514,0.0674954578,0.304278791,0.2701839805,0.3162385821,0.0691310689,-0.2132716924,-0.2078008056,-0.1137950867,-0.0180374142,0.2296580672,0.0889313444,0.0873777047,-0.1719105095,-0.0079711471,-0.3711526096,-0.1685080677,-0.1288425475,-0.085193105,0.2630025446,0.1696459949,-0.0437430739,-0.138473779,-0.2854967713,0.0224270578,0.389308393,-0.0161675867,0.0505798645,-0.4088653624,0.1129736304,-0.2739293277,0.4751008451,-0.037259981,0.3057077229,0.0980384201,-0.1773195565,0.0461823121,-0.0003063008,0.6603267789,-0.3992723227,0.2456143796,-0.2270676345,-0.0527215116,-0.2520557046,-0.2740662694,0.0396930166,0.1629039943,0.3987030685,0.3397414982,-0.2540696263,-0.0806216598,0.2021558434,0.0874188766,-0.0450020172,-0.4370992482,-0.2332210392,-0.2644630373,-0.3989752531,-0.2029905915,0.0157417804,0.1682739705,-0.018936513,-0.0000235067,-0.0035435243,-0.0968364775,0.4360724986,-0.1660076529,0.1386368573,0.1040053517,0.1729806662,0.2166641653,0.1723202467,0.4092180729,0.7318953872,-0.2874979079,-0.6002776027,-0.0113575049,-0.0527178571,0.059787672,0.3057090342,0.0964490771,-0.3372845352,0.563247323,0.198870942,0.0585830286,-0.0768780857,0.0675514266,0.0707513466,-0.2937156856,-0.2277707756,0.2731252313,0.3429972529,0.0426031388,0.1635281593,0.1138819233,-0.0248718951,0.1626589,0.1482758373,1.0862958431,0.0878437236,0.1860835552,-0.0522381179,-0.1104113311,0.2627029419,-0.1908109635,0.1054059044,-0.3536967635,-0.30446136,-0.2195750028,-0.2797238231,0.1659423709,0.1664831936,-0.1162376553,0.5488876104,0.1168221459,0.2160012126,0.2047855705,0.1635703892,-0.5091705918,-0.0332469717,-0.0590834543,0.0884188637,-0.0928000957,0.5007269382,-0.0778229311,-0.1452210248,-0.2754146755,-0.1927773803,-0.2730580866,0.159544155,-0.1317400485,0.2402003407,0.3220721185,-0.2370830625,-0.113027975,0.488699168,0.3835788369,-0.1671031266,-0.3553119898,0.1355219334,0.2764635682,-0.1244989261,-0.0171230976,-0.3096616566,0.2277709842,-0.0881357118,-0.1494827271,-0.0479171388,-0.0436859652,-0.5708991885,-0.0020887859,-0.2316403538,-0.1237522289,-0.4168468416,-0.126259014,-0.1965942234,0.0217029881,-0.270331651,0.0317287035,0.3245995343,-0.3015615344,0.1005060747,0.0035000972,-0.2388787568,-0.2121371627,0.4402431548,-0.0634018332,-0.0947238654,0.3009137809,0.1090557203,-0.1930950135,-0.0311588403,0.0378629006,-0.189049989,-0.3966252208,0.1735886186,-0.0929014534,0.0876044109,-0.0711564645,0.2744777501,0.2370478958,0.258066386,-0.0138267539,-0.6462165117,-0.1248562038,-0.0952459201,0.0612847246,0.5304857492,-0.4701808393,-0.0797349215,0.0009785234,-0.2288040221,-0.1774576902,0.1099165156,-0.2394400686,-0.2627349198,-0.1782147139,0.0223824568,0.1300973743,-0.1634641886,-0.0548021831,0.0018498197,-0.2210846692,-0.0403416231,-0.0434737764,0.2257995158,0.2479028255,-0.121025525,0.3199219704,0.0110338116,-0.0919963792,-0.0054701297,-0.0953246579,0.3137368858,0.0040579708,-0.0022810341,0.1096825972,-0.0783089921,-0.1045235768,-0.0253287479,-0.206926018,0.1630871445,0.1618339866,0.1151995286,-0.0566391461,-0.0262813866,0.1852611452,0.1343863755,-0.0019727156,-0.1917341202,0.2688359618,-0.3860968053,-0.2062043846,0.5657047033,0.4447371066,0.4240089953,-0.3189670742,0.3215558231,0.1060511842,0.1201568618,-0.2795187235,-0.1221104041,0.4327147305,0.0603626817,-0.0285354052,0.2022353262,0.0442619547,-0.3004891574,0.4922033846,0.2466418296,0.2334192991,-0.3281928301,0.3674349785,0.4295713007,-0.1627134532,0.1419421136,0.2974653244,-0.0661256686,0.0168983322,-0.0369440801,-0.0829525888,0.2433850616,-0.4784404933,0.1266919822,-0.0710530728,-0.4233189225,0.1946549565,0.0133204274,-0.0027761983,-0.0304527748,0.0699214786,0.3076066375,-0.263022244,0.112767756,-0.5174113512,0.0381692089,-0.035842333,0.0914097577,0.0797534063,-0.3225444257,0.0346795395,0.1557123959,0.0230995063,-0.2504737973,-0.0276709702,-0.0621574745,-0.3479332626,-0.4791832566,-0.0341070183,0.0095143802,-0.02146082,-0.0762251616,0.0561429188,0.2304467708,-0.3766308725,0.078048341,0.4358033836,0.6305704117,0.2479379624,-0.0078430558,0.0214254968,-0.2103945911,-0.0309742484,-0.047282964,0.3579619527,0.0084608681,-0.0292126331,0.4339888692,0.0156590845,-0.091590412,0.0162390526,-0.3148887455,0.0372813232,-0.2002287358,0.5952569246,-0.0213359781,0.1963330358,-0.0331073254,0.0542357974,-0.3736696839,-0.3054956794,-0.0647713467,0.0438076109,0.1210549474,-0.0202753674,0.0060150879,-0.195438236,0.2912979126,0.4983139038,0.0990587696,-0.3213220537,-0.0941101387,-0.8508207202,0.1570828855,-0.1376277357,0.1217195839,0.0024772175,0.14142555,-0.0202258267,0.1938099712,0.4343990386,0.0081655281,-0.1585168093,0.0657148212,0.0636699945,-0.1766451597,-0.0116655948,-0.02283445,0.1525709182,-0.2536917925,0.1695951074,0.0036754156,-0.0415648371,0.0542045683,0.1688704342,-0.0785414875,-0.1620916724,0.1832969487,0.1747029722,0.5801798105,0.0081683192,-0.1735315174,-0.1242438331,-0.063071996,-0.3637900949,0.3822866678,-0.0729066879,0.3957030773,-0.334664762,-0.2453686595,-0.3081839383,0.1894775778,0.1044971198,-0.1284387559,-0.0745332986,0.1596523374,-0.2124974579,0.2788161635,0.3082185388,-0.0399155095,-0.1191357747,0.2621364892,-0.2827535272,-0.2975854278,0.6069638133,0.03731516,0.0027404572,-0.2981251776,0.1216122359,-0.1206108183,-0.0510753207,-0.5369361043,0.0036438771,0.3522105813,0.0643135533,-0.3270994723,0.132646203,-0.1859229654,-0.0038302029,0.0317367725,0.4268348515,0.2133131474,-0.3390214443,0.1384729147,-0.1304610521]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2330","title":"Allow passing `desc` to `tqdm` in `Dataset.map()`","comments":"Hi @lhoestq,\r\nShould we change `desc` in [pbar](https:\/\/github.com\/huggingface\/datasets\/blob\/81fcf88172ed5e3026ef68aed4c0ec6980372333\/src\/datasets\/arrow_dataset.py#L1860) to something meaningful?","body":"It's normal to have many `map()` calls, and some of them can take a few minutes,\r\nit would be nice to have a description on the progress bar.\r\n\r\nAlternative solution:\r\nPrint the description before\/after the `map()` call.","comment_length":11,"text":"Allow passing `desc` to `tqdm` in `Dataset.map()`\nIt's normal to have many `map()` calls, and some of them can take a few minutes,\r\nit would be nice to have a description on the progress bar.\r\n\r\nAlternative solution:\r\nPrint the description before\/after the `map()` call.\nHi @lhoestq,\r\nShould we change `desc` in [pbar](https:\/\/github.com\/huggingface\/datasets\/blob\/81fcf88172ed5e3026ef68aed4c0ec6980372333\/src\/datasets\/arrow_dataset.py#L1860) to something meaningful?","embeddings":[-0.3072710037,-0.0466327779,-0.1089392528,-0.1826233566,0.3455191255,-0.0337330885,0.0899253562,0.1168782711,-0.3769188821,0.1873242259,0.1908902377,0.6768376231,0.0713502988,0.003169067,-0.0518927909,-0.0042376984,-0.1745890379,0.3236307502,-0.1186980307,0.1601096094,-0.3103112876,0.0022689218,-0.2588056326,0.4401148558,-0.0576699078,-0.2495460361,-0.034188915,0.00190501,-0.0848528221,-0.8410236239,-0.0787710324,0.4994030595,-0.2055618465,0.1110525206,-0.0001215994,0.0029145184,0.4593873918,0.0408072695,-0.1329794824,0.2496578097,-0.2398094982,-0.4344353974,0.551186502,-0.5093436241,0.2677410245,-0.1887652874,0.0759569705,-0.2976067662,0.0851228759,-0.1769500524,0.1757276505,0.2615217865,-0.5217784047,-0.0450647362,0.1423230916,-0.0314468518,-0.1901314408,0.0443354845,0.5492420197,-0.2565240264,-0.3967480063,0.4426681101,-0.0321578607,0.1810791343,0.2126697004,-0.0494978093,0.7965030074,-0.2568858862,0.1686064303,0.4457052052,0.5620917678,-0.4755859077,-0.3325846195,-0.2114982605,-0.1058290228,-0.1745497733,-0.2128906101,-0.0804342031,-0.3764392436,0.2044846565,-0.4933661819,0.0933172852,0.0392252505,-0.1457511187,0.0430186577,-0.1293406487,0.0860719606,0.2024744451,-0.0520003475,-0.1431442499,0.0738009959,-0.0940795317,-0.0987012237,0.140619278,-0.2670850456,-0.270945549,0.3573226035,-0.0592694469,0.0030517429,0.2379104048,-0.1351323724,0.1036520228,-0.2139101923,0.254622668,0.3084129691,-0.0480647124,0.4439511299,0.0759405345,-0.1012311056,-0.3503714204,0.0169175658,-0.0498899408,0.249105826,-0.1674872339,0.4004657269,0.2560196817,0.105828464,-0.0888704583,-0.0179858338,0.1331774443,0.2402548492,0.1474353224,-0.0671393797,0.6026400924,0.2194808573,0.2251881659,-0.0108078606,-0.2273169756,0.190542385,-0.0732044727,-0.1797383279,-0.1927359551,-0.2795526683,0.031425409,-0.4915224612,0.3545736969,0.0965812504,-0.2645706832,0.2802160978,0.1016683206,-0.110752143,-0.0135478731,0.3312871754,0.443598479,-0.3514434695,0.3003787994,0.1156251058,0.2027931064,-0.3349898458,0.4062410593,-0.189402923,-0.2118563056,-0.0261345301,0.0450393036,0.2013643533,0.040031109,-0.0917568654,0.0870028362,0.0885727257,-0.1317393482,0.2953703105,0.0082915258,-0.331502825,-0.3361302614,-0.0851317868,0.14392142,-0.4127627015,-0.1525770426,-0.1879814118,-0.0686161742,0.46899575,0.131669715,-0.343061775,0.3049196899,0.0413502157,0.2516263723,0.6775072813,-0.2305899262,-0.3023616076,0.5443847179,-0.2071109563,-0.2756797075,0.0145497574,0.0551021807,0.3899468184,-0.061633341,0.1977811009,-0.1046165153,-0.0942086801,0.3620114923,-0.0428860001,-0.1592458338,0.1674716771,0.0658023879,-0.0859584659,0.0309840348,0.139613077,-0.4066047072,0.4817200899,0.1161017194,-0.0121252099,0.05295977,0.3084961176,-0.3186561167,-0.2289988846,0.0499279201,-0.2609936893,0.1519788355,0.0802391246,0.0818757936,-0.1902191192,-0.3986062109,-0.1781596988,0.4615989327,0.1075879186,-0.0400308631,0.0281589758,-0.2775934935,-0.1536908597,0.0543195195,-0.0810145885,-0.3951455355,-0.2834453583,0.023315765,0.2867541313,-0.0593489408,0.2591720819,-0.1201344728,-0.0373364761,-0.007019083,0.0287231766,0.1588220298,-0.2995089591,0.4061307907,0.0430141389,0.3881074786,0.1531360298,0.2640830576,-0.1134597436,-0.192898646,0.0987504721,-0.0803662241,-0.1457854956,-0.0703062192,0.0604966395,0.3114670813,0.1652771831,0.5148589611,0.0360511057,0.2178665251,0.0352383442,0.1227186918,-0.2736406326,-0.0424827375,-0.0863583311,0.417655766,0.3628048301,-0.0644334182,-0.0912964866,0.2977759838,0.5989565253,-0.0584964193,-0.1227983758,0.1076183021,-0.1905966848,0.079464972,0.3552560806,0.020057952,0.028757995,0.2129210085,-0.0933251902,0.0450039245,-0.2113839388,-0.0361410789,0.1632660031,0.1856961548,0.112935856,-0.170593977,-0.0058884611,-0.3485253453,-0.0993824378,-0.3496234417,0.0585720874,0.3103129268,-0.1382211447,-0.0904071927,0.0596311018,-0.2052439749,-0.0353769735,-0.2906074226,0.0265384354,-0.6030112505,-0.0166394711,-0.0740305111,-0.2435645163,0.4365400374,-0.2881198525,-0.1143134683,0.5323266387,-0.0683567598,-0.2052150369,-0.404369086,0.2695590854,-0.0684492961,0.0314100683,-0.2258301079,0.3889926672,-0.1434429437,0.2305974364,-0.2582543492,-0.046942804,0.2209360152,-0.0389317535,-0.1885608286,0.2259573787,0.2952449322,-0.1880596429,0.123553291,0.1861379296,-0.2096957564,-0.2099896967,-0.0508149192,0.0579681545,0.0193243846,-0.0937866867,0.1336030215,-0.1489562243,-0.3866201341,0.3697223365,-0.4466917813,0.3708949983,0.2791883945,-0.1703437269,0.1503603905,-0.3276536465,-0.2689218521,0.0953211114,-0.3909830451,0.0422779433,0.0149786286,0.1479041427,-0.0357056782,0.0186138339,0.1082307771,-0.0979803652,-0.2466458678,-0.4173072875,-0.2290323526,0.4694497585,-0.1056448147,-0.1143243164,0.4141492844,0.220314905,-0.1677263081,0.0457869954,-0.3092603087,0.0020521472,0.391805917,-0.0254416484,-0.0722826645,0.3030149639,-0.095482178,0.6081329584,0.2357957959,-0.0683078319,-0.1732721925,-0.1007460877,0.1018380225,0.1546055824,-0.0225582216,-0.3580590785,-0.0283435471,-0.3491327465,0.0628366768,-0.3110090792,-0.17959176,0.2184303701,0.3978835642,-0.2516092062,-0.2520862222,0.0772436857,-0.4256778657,0.2163166702,-0.2195917815,0.2034888864,-0.0434182473,0.1429280043,0.0882387087,0.1622564644,0.2648357451,-0.2827656567,-0.1943306476,-0.0948676318,-0.6611031294,0.1265069693,0.0856343061,0.156081304,-0.2277347594,0.0628640205,0.187470302,0.3181170821,0.2722467184,-0.5783202052,-0.0657238364,0.1639238298,0.0684204176,-0.358964026,0.1880295575,-0.039567817,0.4599541128,0.1769430786,0.1857537329,-0.3063776791,-0.1715427786,0.0217305645,-0.2876066267,-0.2856378257,-0.2168765813,-0.2452751547,-0.0324293524,-0.3162260652,0.1737104654,0.2033004016,-0.1427773684,-0.0072469874,-0.1791038811,0.1451260149,0.0316230394,0.0401565209,0.2240277678,0.2859920263,0.1577167362,0.0020923675,0.1207668111,0.0377400778,-0.1580850333,0.4212833941,-0.1565069556,-0.1828219295,0.2516844571,0.0422880091,0.5727679729,0.5166627765,-0.0057944325,0.4574667811,-0.0509534515,0.3048254251,-0.203649357,0.337218225,0.170529902,-0.1515514255,-0.257758528,-0.3016842008,0.2801211476,-0.1188799813,-0.2501350343,0.422868818,0.0435765013,-0.0114650084,0.016504677,0.1098674759,1.1463087797,-0.4182510078,0.0694864988,-0.1635508388,-0.2303236276,-0.1379033774,0.0869337842,-0.2026983202,-0.551201582,0.1404602677,-0.0523540899,-0.2750806808,0.1423650831,0.2595633864,-0.3630469739,0.0532370657,0.1888811588,0.1226075292,0.0544092543,0.0383559912,0.2581376433,-0.5784074068,-0.3684664965,-0.0214350093,-0.0707429424,-0.370439291,-0.0654928237,-0.056405507,0.1195936725,-0.0707748607,-0.1817066371,-0.2800461054,-0.4279144406,-0.0545717441,0.0862331986,-0.3214488924,0.0017669545,0.0324350968,-0.1698949188,0.2067947984,-0.0811549053,0.2371582091,0.0093690194,0.3928011656,-0.0688716248,-0.0984151736,0.2633372843,0.1563239992,-0.0414996557,0.5083097219,-0.2331703007,-0.3960368335,-0.1274160892,0.1148259938,-0.014121633,-0.0419608653,0.1520823538,0.0611069351,-0.1303218305,-0.0733059347,0.0501675569,0.4284384251,-0.2042183578,0.3699139357,0.1682662815,-0.365144521,0.1674433053,-0.007170537,-0.0325437002,0.2095936388,0.3339779973,-0.1572727561,-0.2968187332,-0.1734149009,0.1824237108,0.244134903,-0.2956413329,0.1378153265,-0.0342117585,0.056133084,-0.1010484397,0.3718962371,-0.2467257082,0.2415933311,-0.4561505616,0.0254475474,-0.5924559832,0.1963911653,-0.2529989779,0.352824688,-0.1060981676,0.3185599744,0.0734239668,0.1617462188,-0.2528199553,-0.1576077193,-0.1554459333,0.2978178859,0.0722156987,0.128270179,0.1433280855,-0.1425302625,-0.0518781357,0.4143431485,-0.3873563707,-0.1484775096,0.08266332,0.2285252213,0.131143719,0.1370667219,-0.0222776346,0.0198533442,0.006104392,-0.1478289217,0.0826528221,-0.1017867476,-0.0608244352,0.0679787546,0.1139191315,0.1492170841,0.0474393368,0.1352811903,0.083153598,0.5757147074,-0.0728435516,0.0123411678,-0.0069888174,0.0148030017,0.5606937408,0.1707991958,0.2374644279,0.0206232127,0.165443778,0.0380118378,0.0306129549,-0.0199984871,0.3704387546,-0.0791879967,-0.3550455868,-0.2031693757,0.1123690605,0.1206800267,-0.079914771,-0.0123302918,0.2756312788,-0.1167592332,0.2071902007,0.2739208341,-0.0213921703,0.2737828791,0.2268152833,0.0014208172,-0.1449127644,-0.2695441544,-0.1305820197,0.323802799,0.0736638531,0.1578816473,0.6479834318,0.4076951742,0.2172995061,0.3777754009,0.2644347548,0.5589678288,0.1722508222,-0.1800815761,-0.0919614658,-0.0341864116,0.1475350261,-0.0478322841,-0.1012943164,0.2968015075,0.2229041308,0.2139360905,-0.1273013055,-0.0965303928,-0.1671859771,0.0624662079,-0.2160784304,-0.1748737395,0.3564842343,0.043707341,0.2026521713,-0.0516503863,0.0469068959,-0.0329069644,0.5420635939,-0.2210308462,0.4852814972,-0.1462919861,-0.1921389401,0.0574793741,0.4696072638,-0.3269206285,-0.1596680284,-0.2118863016,0.1957513988,-0.0119611984,0.2591776848,0.2310431153,0.1715106219,-0.2901810408,-0.0595741384,-0.0231592264,0.0685800985,0.1602127403,-0.015750058,-0.0443686768,-0.1762479097,0.3066939116,0.0192213394,-0.0533375069,-0.0713458657,0.0441449285,0.2587499619,-0.0158730391,0.3462969959,0.0945443362,-0.2089578211,0.0182363354,0.1684522778,-0.2411034107,-0.1758974344,0.3886463046,-0.3514864743,0.0216893684,-0.2935937941,0.0363319851,-0.1105771363,0.4864827096,0.333114922,0.3494307101,-0.381146878,-0.3620235026,-0.5225533843,-0.0489429012,-0.0514124446,-0.2406903356,0.2740222514,0.1793921888,0.141865775,0.4638444483,0.4708786607,-0.3044942021,-0.1186609119,0.2215524763,-0.2251726091,-0.1456356496,0.065423131,-0.1142001078,0.030695647,-0.1015980318,0.1782718599,0.3443032503,0.0087356325,-0.2053551376,0.1450529695,0.0823126286,0.1587791145,0.4628461599,-0.0868258104,0.1750554591,0.0400757864,-0.0155834304,0.413502574,-0.3070274889,0.2044427246,0.1386117935,-0.1268128306,0.2547102273,0.0880572349,-0.0515589975,-0.2640472651,-0.2790506184,0.1556993276,-0.2398375124,-0.1983914971,-0.0236046966,0.0243072454,0.0579300858,-0.3988565207,-0.0135476282,0.3835868835,0.3672170043,0.0257459842,-0.6022506356,0.3368788064,-0.0433191955,-0.5273472667,0.3428891897,0.1499558985,-0.2435362637,0.1638964117,-0.8755454421,-0.1355119348,0.1957211196,-0.028244799,-0.1089924946,0.2201502472,-0.3667953908,0.1497602612,-0.1935126334,0.3791491389,-0.1043318361,0.0295014717,-0.3907364607,-0.3946561813]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2330","title":"Allow passing `desc` to `tqdm` in `Dataset.map()`","comments":"I think the user could pass the `desc` parameter to `map` so that it can be displayed in the tqdm progress bar, as suggested by @cccntu.\r\n\r\nWhen there's no multiprocessing, the `desc` of the progress bar could be the `desc` passed by the user.\r\nIn multiprocessing, we were already using a `desc` equal to `\"#\" + str(rank)`.\r\nWe can change it to be `(desc or \"\") + \"#\" + str(rank)` instead.\r\n\r\nIn the end, since both `desc` and `rank` could be None, we can have:\r\n```python\r\npbar_desc = (desc or \"\") + \"#\" + str(rank) if rank is not None else desc\r\n```\r\n\r\nFinally let's remember that if we add `desc` as a new parameter to `map`, we should add it to the `ignore_kwargs` list of the `@fingerprint_transform` decorator of `Dataset._map_single` since we don't want this parameter to affect the fingerprint of the resulting dataset.","body":"It's normal to have many `map()` calls, and some of them can take a few minutes,\r\nit would be nice to have a description on the progress bar.\r\n\r\nAlternative solution:\r\nPrint the description before\/after the `map()` call.","comment_length":145,"text":"Allow passing `desc` to `tqdm` in `Dataset.map()`\nIt's normal to have many `map()` calls, and some of them can take a few minutes,\r\nit would be nice to have a description on the progress bar.\r\n\r\nAlternative solution:\r\nPrint the description before\/after the `map()` call.\nI think the user could pass the `desc` parameter to `map` so that it can be displayed in the tqdm progress bar, as suggested by @cccntu.\r\n\r\nWhen there's no multiprocessing, the `desc` of the progress bar could be the `desc` passed by the user.\r\nIn multiprocessing, we were already using a `desc` equal to `\"#\" + str(rank)`.\r\nWe can change it to be `(desc or \"\") + \"#\" + str(rank)` instead.\r\n\r\nIn the end, since both `desc` and `rank` could be None, we can have:\r\n```python\r\npbar_desc = (desc or \"\") + \"#\" + str(rank) if rank is not None else desc\r\n```\r\n\r\nFinally let's remember that if we add `desc` as a new parameter to `map`, we should add it to the `ignore_kwargs` list of the `@fingerprint_transform` decorator of `Dataset._map_single` since we don't want this parameter to affect the fingerprint of the resulting dataset.","embeddings":[-0.3543031216,0.0359196849,-0.0605332516,-0.169031918,0.3744505346,-0.1667570919,0.2543700635,0.2177440226,-0.3104316592,0.4162537158,0.2674728036,0.6729348302,-0.0182513837,0.0839984864,-0.1400039196,0.0010156956,-0.2245403677,0.2880353928,-0.1542477608,0.1733601391,-0.3427264392,-0.0123510053,-0.1521170139,0.333489269,-0.1336866617,-0.2490088791,0.0189019199,0.0414581671,-0.0060942615,-0.6088420153,-0.2313154191,0.4572477639,-0.1192472279,0.2423422784,-0.0001132663,0.0449169613,0.3025875092,-0.0090298932,0.0343071297,-0.0876949951,-0.2307193577,-0.2086026073,0.205450058,-0.5455300808,0.0810967162,0.0083245039,0.0779942423,-0.2727372646,0.2467460781,-0.2837902308,0.2267532796,0.3224823475,-0.4963641465,-0.1730308235,0.0051274141,-0.1487742811,-0.2355037779,-0.0538833328,0.4061657488,-0.0787890255,-0.3605757654,0.5466031432,0.0331227109,0.1048731655,0.0424411297,0.0684083477,0.9416398406,-0.3111875951,0.2162596583,0.5596245527,-0.0025465908,-0.403375566,-0.2936964333,-0.1252213269,0.1010691598,-0.2364679128,-0.1962261945,-0.1318239421,-0.4667769074,0.0572201423,-0.6661195159,0.2062962949,0.0636753291,-0.1756588966,-0.1616928428,0.1324810833,0.1472689211,0.175353691,0.048448775,-0.0012471844,-0.0412459597,0.0121267904,0.1333086044,0.0419924408,-0.2462373972,-0.2159461379,0.2609503269,-0.1506172419,0.0285712127,0.0838271677,-0.1659707427,0.2715797424,-0.1430589855,0.2127112597,0.1533040553,-0.0197043065,0.2840502262,-0.0085473321,0.0391802825,-0.3217777312,-0.1140473261,0.063157618,0.3379700184,-0.2410391718,0.5201619864,0.4561315179,-0.0073194918,0.0579646379,-0.0625929534,0.1070722118,0.2622315288,0.1972302049,-0.050163772,0.5713624954,0.2755984664,0.2373299748,0.0270602927,-0.1368020475,0.0655463561,-0.1194635928,-0.1515819579,-0.2082107812,-0.3167417049,0.1212709546,-0.3725419641,0.2893479168,0.0622703917,-0.0132815307,0.3908943534,0.0847775042,-0.0046318881,-0.1438905895,0.3979234993,0.3534492552,-0.0868191645,0.1720015407,0.142945081,-0.0032617503,-0.2947628498,0.4820659459,0.0177822802,-0.0298867673,-0.0325824879,0.1280913502,0.0457737334,0.2533397079,-0.1417144686,0.1408379674,0.0359459631,0.0297054537,0.258611083,-0.177757591,-0.4316973984,-0.3306638598,0.0106278807,0.1676989198,-0.1842898875,-0.2401033938,-0.1703087538,-0.0373635143,0.305379957,-0.0256110393,-0.2176831216,0.3224095404,-0.0214518029,0.3570397496,0.6694242954,-0.2494210154,-0.1231464222,0.469635278,-0.3237856627,-0.1044310108,0.1116032004,0.0702892765,0.3541651666,0.0041367575,-0.0440221578,-0.09036991,-0.1212184802,0.3055946529,-0.0965264514,-0.1311973631,0.225217849,0.0808431059,0.1800924391,0.043225605,-0.1003475562,-0.3944578767,0.4690576494,-0.0687007979,0.1408855915,0.105878599,0.036459282,-0.274223119,-0.2740654945,-0.1592959166,-0.1313504279,0.3375793397,-0.1488400847,0.0257428996,-0.237604931,-0.3887973726,-0.1905229688,0.2381239533,-0.0400184989,0.000591166,0.1616580188,-0.1433953643,-0.2247552127,-0.1466299444,-0.0102205034,-0.1418243498,-0.0116472915,-0.1284280717,0.2948251069,0.0363100208,0.1882728189,-0.1888583899,-0.0742232352,0.0934998617,0.1896006167,0.0957732946,-0.1555179656,0.4198977351,0.0185335502,0.1653473228,0.0407476798,0.126327157,-0.0886523649,-0.1433205456,-0.0579831526,0.1885076016,-0.0610125884,-0.06293156,0.0237915125,0.5794562101,0.0664438531,0.3983498812,-0.0122988941,0.1748764813,0.1880546212,-0.1832468957,-0.0905973166,-0.1429748237,-0.0115729617,0.4477988183,0.383444041,-0.1673359573,-0.2730872035,0.3041702211,0.5253994465,0.138740316,-0.0068682539,-0.0814467221,-0.1799041778,0.2821110785,0.1481162906,0.0180238858,0.1794286221,0.3364721835,0.1017204076,0.1056354865,-0.1593882143,-0.0237740036,0.0749181658,0.0301051624,-0.0055219666,0.016417196,-0.0589955263,-0.1441398263,-0.0818932801,-0.2558988333,0.0643959418,0.0688643083,-0.2571947575,-0.0724465474,-0.0504352711,-0.0811872184,0.0147793312,-0.192697987,-0.0432978719,-0.4102104008,0.16133596,0.0397835337,-0.2512671947,0.4082573056,-0.0590766892,-0.2016568184,0.1945892125,-0.0901442692,-0.0886031464,-0.3783821762,0.1800910234,0.0163422469,0.1017844379,0.0101125734,0.3469286561,0.0168993864,0.1499071568,-0.2127230465,0.0196143445,0.0517327711,-0.1476740092,-0.0338900983,0.0720474049,0.2288357466,-0.275609225,0.1529309005,0.245251745,-0.2540294528,-0.1178651676,0.0562493429,0.1432034671,-0.0320925787,-0.2721735835,0.1490704566,-0.1415087581,-0.3524087071,0.079158619,-0.3988301456,0.3791470826,-0.0064547686,-0.1370365322,0.0351493247,0.0223249495,-0.2152631879,-0.0719897747,-0.4446079731,0.0634605587,-0.1008327901,-0.073219806,-0.0186567698,0.0207427554,0.115023829,0.0150276115,-0.1133948267,-0.5253043175,-0.1396808326,0.4327143431,-0.2014661133,-0.1334905773,0.5469970703,0.1849723756,-0.26857391,-0.0150798149,-0.2984207869,0.0338325128,0.1508781612,0.0023235672,-0.2157335579,0.1579723954,-0.0173316039,0.7264468074,0.2910222113,-0.126720041,-0.114783816,-0.1962477267,0.1106468216,0.1317901909,-0.0290173069,-0.4755138755,-0.0374588482,-0.0940501839,0.1027610675,-0.2709338963,-0.0640768334,0.2485486418,0.4160204828,-0.4120849371,-0.357522428,0.1175739318,-0.3366911709,0.2249913663,-0.3050579727,0.2969436646,-0.1149432287,0.1209469587,0.012109369,0.2440431565,0.1088889167,-0.2996542752,0.0664762035,0.0148181487,-0.3753647208,0.2600144446,0.1935792416,-0.0139995227,-0.128584981,-0.0522100665,0.2372803241,0.0844521075,0.3091477752,-0.5974338055,0.1905173659,0.4091605246,-0.290332973,-0.2686609924,0.0744363889,-0.0934717804,0.2835863233,0.1928517967,0.2323409021,-0.2917527854,-0.0816627443,-0.204918474,-0.2462274879,-0.3318457603,-0.2268570513,-0.2825606465,-0.0673750341,-0.105678387,0.3289418817,0.1057234108,-0.0931392014,-0.1121614277,-0.1835068166,0.085079886,-0.1289615929,0.0089083361,0.4399094284,0.2581612766,0.0957551226,0.0231382456,0.0018137683,0.0294709727,-0.2969562411,0.2772450149,-0.0737154856,-0.054808408,0.11601758,-0.1062045842,0.3790581226,0.4680798352,0.0209911894,0.4216799736,-0.0877086967,0.3533377945,-0.2584062517,0.256025821,0.0746825188,-0.2012855709,-0.1703725159,-0.1626459509,0.3817444742,-0.0297057722,-0.2717422247,0.1986742765,-0.2033969313,-0.1684937328,0.0089621544,0.2284095436,1.098941803,-0.6049200296,-0.0631915331,-0.3199733198,-0.1295970827,-0.1932378709,0.1619354188,-0.1454752833,-0.3798640668,-0.0958779603,-0.0446434505,-0.1203744262,0.1903120279,0.2228494585,-0.4763143957,0.0453110263,0.0826769322,-0.0158165041,-0.0537970401,0.0857195333,0.2267693728,-0.5300146937,-0.2672763169,0.0878039524,-0.003131591,-0.4076291323,-0.1108337417,-0.1381043196,0.1012069806,0.0013530379,-0.2048480213,-0.1415641755,-0.5126553178,-0.0310581625,0.0223970991,-0.2806537151,-0.0010208186,-0.0021850474,0.0012861395,0.134808138,-0.0737180337,0.2051578909,0.0328309461,0.4995693862,-0.0395827293,-0.0346125029,0.3274907172,0.0430510007,-0.1543291211,0.4600406885,-0.0885946751,-0.3782998919,-0.0880598947,0.1505749077,0.0895118788,0.104546912,0.4316311777,0.0705661774,0.1072730422,-0.1773658991,0.1075254604,0.3297344148,-0.301353246,0.466485858,0.1509518027,-0.3971974254,0.1685273796,0.0400902405,0.0715516061,0.3789876401,0.2130925208,-0.3200155199,-0.1801005155,-0.3070732355,0.0981554016,-0.005128236,-0.0047616535,0.2817222774,0.0922124907,0.1112033576,0.0824301615,0.3829148114,-0.1607064009,0.1288014501,-0.4710475206,-0.0861841515,-0.3992663324,-0.0023446076,-0.2232076079,0.2558085322,-0.1603619903,0.700314641,0.0832813457,0.2738613188,-0.3663798273,-0.2554730177,-0.4581249952,0.39759323,0.1494739354,0.1711195409,0.0417525582,-0.0895034596,0.0275863688,0.3001737595,-0.5012282133,-0.2796919048,-0.146778509,0.1485368311,0.0718939975,0.3211847842,0.0104425065,0.1530555487,0.1007179096,-0.1485731453,0.2995369136,0.1371569782,-0.1262771189,0.2434355468,0.1613571197,0.2583529055,-0.1614622772,-0.0658649281,0.0488376841,0.3448450863,-0.1857880056,0.0928952098,-0.0177288707,0.0814499259,0.4266801178,0.0163479336,0.2751743495,0.0750815868,0.0246076174,0.2015183568,0.1230052114,-0.0335370488,0.2313364148,-0.0955634862,-0.280313164,-0.2249711305,0.2154681534,0.2163048536,-0.1700076014,0.1758911759,0.1357088536,-0.1021308079,0.356472671,0.3878494203,0.0229428988,0.2096785456,0.266277343,-0.0907113552,-0.1220262945,-0.4007737339,-0.0565473437,0.2515042126,0.168618843,0.2333994508,0.5188537836,0.2669459879,0.0628480688,0.312174201,0.0969325155,0.7731795311,0.0355854295,-0.2280782908,-0.1063074693,0.0778793544,0.1772416979,0.0026592955,-0.0090826927,0.2526454031,0.1282432824,0.1609683037,-0.24937208,-0.1211344153,-0.1930092275,0.1743644327,-0.1143894047,-0.2444416881,0.2082290798,0.0702758208,-0.1169953197,-0.1188067049,0.1732619703,0.0206393115,0.7649599314,-0.0720501244,0.323946327,-0.0148437936,-0.439348042,-0.0521437488,0.2582207322,-0.4002414346,-0.1658609957,-0.149653554,0.277395308,0.2187279463,0.3526684642,0.1083387434,0.146221295,-0.3900358677,-0.0591477491,-0.0979830846,0.1233994067,-0.0057189302,-0.1346590966,-0.0186478067,-0.0586037785,0.3146674037,0.1123372689,-0.1652235985,0.0274717994,0.1435278654,0.3673883975,-0.0919075087,0.4869936407,0.1941068619,-0.1355677992,0.0483947471,-0.0333840735,-0.2675362527,-0.142897889,0.3206524551,-0.4000179768,0.0588194244,-0.226305455,0.1019982994,0.0596870445,0.362126112,0.1174310222,0.0723797157,-0.4723439217,-0.2349482328,-0.6552355289,0.0114019373,-0.0769951716,-0.2930529118,0.1641541272,0.2058352679,0.005148117,0.3640120029,0.4453575909,-0.243460089,-0.1943772286,0.2442207634,-0.3748534918,0.0006777131,0.0834980011,-0.2616188824,0.1132396311,-0.2640103996,0.0854104087,0.2979796827,0.1450808644,0.0914872959,-0.0240645055,0.19539693,0.3445814848,0.4568000138,-0.1306682229,0.3079547584,-0.1257483959,0.2529365718,0.3577656448,0.0457691438,-0.075084649,0.0522866733,0.1410209686,0.3408601582,0.0279194396,-0.0452304669,-0.1750472188,-0.1457654983,0.057588093,-0.1890040636,-0.1555407047,0.0548096001,-0.0198324993,0.1372141987,-0.3962384164,0.0775620267,0.2672230303,0.4194971323,-0.1872834414,-0.6303867698,0.2855833471,-0.2099762559,-0.6549606919,0.0069323503,0.2808082998,-0.5046144128,0.1485663503,-0.7743097544,-0.0654677972,0.1463049501,0.1020645276,-0.1546311677,0.1827256382,-0.50043571,-0.0918131694,-0.1753419191,0.1103482619,-0.0111007588,0.0091225756,-0.3428884149,-0.3361288607]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2327","title":"A syntax error in example","comments":"cc @beurkinger but I think this has been fixed internally and will soon be updated right ?","body":"![image](https:\/\/user-images.githubusercontent.com\/6883957\/117315905-b47a5c00-aeba-11eb-91eb-b2a4a0212a56.png)\r\n\r\nSorry to report with an image, I can't find the template source code of this snippet.","comment_length":17,"text":"A syntax error in example\n![image](https:\/\/user-images.githubusercontent.com\/6883957\/117315905-b47a5c00-aeba-11eb-91eb-b2a4a0212a56.png)\r\n\r\nSorry to report with an image, I can't find the template source code of this snippet.\ncc @beurkinger but I think this has been fixed internally and will soon be updated right ?","embeddings":[0.0715632811,-0.4951067865,-0.1797281951,-0.1540903896,0.0774730146,-0.2150441259,0.1975012869,0.2825195789,-0.3641221225,0.1687918156,0.2558269501,0.3997045755,0.0529384203,0.0039085206,0.0124356346,-0.1999744922,0.0728849694,0.2567012608,0.0948015749,0.0897849798,-0.1415306926,0.4616273344,-0.2060655206,0.1898974925,-0.1933685988,-0.0379654169,0.0317138806,0.3373529017,-0.2273010314,-0.0470030457,0.0010915148,0.064178206,-0.1115034297,0.0319319181,-0.0000933617,0.0573273972,0.4510402381,-0.0477175824,-0.0617727712,-0.0962015763,-0.3083204925,0.1501756608,-0.0974037796,-0.2114107609,-0.2634438872,-0.0932573527,0.1575714946,0.225198999,0.6356444359,0.1518195719,0.4007506669,-0.0871496126,0.0372064859,-0.2963175476,0.1952328831,0.1279038936,-0.0493087508,-0.0032211577,0.3966778517,-0.0520781614,0.3011013269,0.7066950202,0.1491575241,-0.0617366955,0.1567130387,-0.0642567426,0.0901056379,-0.1462881416,0.1106479466,0.2388547957,0.1958262771,-0.004171296,-0.0273603406,0.1946362555,-0.2383274287,0.0582046025,0.104520224,0.1131511033,-0.0325616226,0.1628715843,-0.1777629256,-0.3388653398,-0.4302212298,0.0763217136,0.1298819184,0.1238824949,-0.2709282637,-0.0848228782,-0.0750402808,-0.0733413398,0.0690799356,-0.0947975218,-0.1755657643,0.032806363,0.3898117542,0.0336384363,0.2636396885,-0.1664782315,0.0907298997,0.2093675584,0.3645391166,0.1617857963,0.044568494,0.4841555059,-0.059862718,-0.2177202255,0.1092422828,0.0952872112,0.2170198262,0.2489142716,0.1988035738,-0.1722972095,-0.2023472637,-0.4475755095,-0.1721684784,-0.0195734091,0.1949792057,0.0756070614,-0.0213516075,-0.1340056956,0.0416855104,0.0311296862,-0.0521577597,0.2935717702,0.2862028182,-0.2618910074,0.2229367793,-0.0731547251,-0.0839980394,-0.2374912649,-0.3668646514,0.4344537556,-0.3073208928,-0.3818004131,-0.2056149691,-0.099942334,0.1797950417,0.0794072598,0.3844331205,-0.1954474896,0.2693597972,0.1235044524,0.1823216081,0.0405629426,-0.0613383763,-0.0378034748,0.1874403358,-0.140776962,0.0633979663,0.2403682172,-0.3574039936,0.0198715869,-0.3501662314,0.3257005811,0.0071030757,0.0871215984,-0.0015992568,0.1873692721,0.2495847344,-0.017472405,0.0723603144,-0.0797396749,0.0049851905,-0.1812770367,0.0313532576,0.1445374489,-0.1001026705,-0.0282105524,0.0006581482,-0.1491847038,0.1289750934,0.1001662761,-0.085537374,0.2190530449,-0.3449908793,0.3650205135,0.1883977503,-0.2218052745,0.2569124997,0.3305473626,-0.0898411795,-0.1762397736,-0.1385308057,-0.1838117838,-0.0027715589,-0.0424129516,-0.076133959,0.2300001532,0.0800398514,0.1793266833,-0.2126228958,0.0947306901,-0.1867510974,-0.0717876181,-0.0061205765,-0.0782548711,0.1166824177,0.3036964536,0.1652324796,-0.318279922,-0.0693618283,0.1438332051,0.4180040359,-0.1671893597,0.010069483,0.0409813747,0.1267256588,-0.1673319787,-0.2309223115,0.0167770237,-0.0212326106,0.1463430226,-0.2395848334,-0.0800207183,-0.1342269331,-0.3119918108,0.4072885513,0.0106123425,-0.016910756,-0.1834714264,0.0215902235,0.0133522497,-0.2049185336,-0.0322175324,0.025399331,0.049333401,-0.3071591258,-0.0733259171,0.0361094326,-0.2548871338,0.0245670862,-0.1094739214,-0.1592314988,0.4442049861,0.1413282901,0.0414529964,-0.0152209299,0.0102455113,-0.0901464671,-0.5474354029,-0.0892877877,0.1133851111,0.063894704,-0.0388867743,-0.2313878387,0.1792845875,0.1370069832,-0.0850085989,0.104477331,0.0620643087,0.0435256027,0.0470072515,-0.081009686,-0.1532190442,-0.0685969591,0.194492057,-0.0653938949,-0.3109587431,-0.3392188847,-0.088443853,0.203523308,0.0368342586,0.2094865739,0.0864068642,0.1118897274,-0.0480871908,0.1330356151,-0.1928911656,0.0090983538,0.1427801847,-0.0994561166,-0.0307848882,0.0792718157,-0.3007695675,0.3233331442,0.1179037541,0.1392183453,-0.0968128815,0.0771728456,-0.1826923937,-0.2567153573,-0.1586351693,-0.1180126369,-0.0927415043,-0.170383811,0.2097693831,-0.1907847524,-0.2285812646,0.0054413294,0.121594049,-0.0043110652,-0.2447490394,0.3674565554,-0.0179614741,-0.174718678,0.2753228247,0.1864567846,0.121570386,0.0565675907,0.2379210442,-0.2804912925,0.0258785095,-0.1376214027,0.3204863369,0.0677374303,-0.2399541438,0.1588524431,-0.1785077453,0.2150992751,-0.1911414862,-0.6027473807,0.2668872476,0.0793695301,0.1450598538,0.3221103549,0.2325113714,-0.0992604196,0.1294579506,0.3499355018,-0.0739874393,-0.2206541598,0.0096322596,0.0324231088,-0.0643796548,-0.4315524399,0.2253299654,0.1619851887,-0.2519227266,-0.1134442315,0.2032539546,0.2085518837,0.499602437,0.5264030695,0.1237835735,-0.1164513379,0.0399327874,0.0033893934,-0.315153867,0.1725258529,-0.3455632627,-0.2890867591,-0.1712598205,0.0984721258,0.1768655628,-0.1460026205,0.0523329452,-0.3894007802,0.0924828947,-0.0556530654,-0.0745255426,-0.0114935422,-0.0547323339,-0.0129131963,-0.4187718332,-0.5239934325,0.1694461256,0.1538664848,-0.3789800406,0.1989573538,0.0611085333,-0.087557666,0.2441907227,0.4768975377,0.0424322933,-0.0711891651,0.2790886164,0.0041550957,0.3523954153,-0.0233964939,0.1190907583,0.2066285461,0.363634944,-0.018282719,0.1448671222,-0.0236815549,-0.0788129196,-0.0408137813,-0.0138904303,-0.1867888719,0.0739570558,-0.00707176,0.3263540566,0.0798627213,0.0956488103,0.0213641506,0.1200428605,0.2129473835,0.18820934,0.387815088,0.2118499726,-0.0233112089,-0.0469512045,-0.1843382418,0.0923286602,0.1610891968,0.10995204,-0.1553365737,-0.1177560315,-0.2147241384,0.0611135848,-0.3169390261,0.4080032408,-0.2792223096,-0.1637157351,0.1321358085,-0.0995102972,0.0626944527,-0.058375638,-0.2980416417,0.398947686,0.1735858619,0.0962038562,-0.1565978825,-0.0632659793,0.0652255267,-0.5374460816,-0.1647794545,0.0501794219,-0.0044471668,-0.2646267414,-0.3756535351,-0.207976386,0.2388739586,-0.0106988652,0.2318128496,-0.1614081115,0.0356576629,-0.0671873316,-0.175816834,0.2079935223,0.1705128849,-0.2478565127,-0.2505509853,-0.1706192791,-0.0673833489,0.1498774439,-0.0424335413,-0.033712592,0.0404799171,0.0624480508,-0.0881765857,0.1336309165,0.1885130107,-0.3140097558,-0.2016717792,-0.0443115458,0.1782079786,0.164380163,0.0213042572,0.1333990395,0.1246834844,0.0323382579,-0.24602139,0.2191867381,0.0945932344,0.0395983532,0.2605484724,-0.0328815058,-0.2100171745,-0.0165730584,0.2635546923,0.8582571745,0.2074985355,0.0427850299,0.2182802111,-0.2088026851,0.1898964047,-0.0901072994,-0.0922126099,-0.3557022214,-0.0908879265,0.0697551668,-0.1454473585,0.1278451383,0.1065113693,-0.0965304822,-0.1428257525,-0.2703927159,0.1838329136,0.0104385968,-0.0242331848,0.1529514492,0.0700927526,0.0235543624,0.3410838544,0.0684477761,0.0254692007,-0.2867955565,-0.3670866787,-0.1807395816,-0.3988815248,-0.2417755127,0.3232107759,-0.1162724718,0.1979614496,0.3024292588,-0.16817002,0.0923651904,-0.1475239247,0.023821665,0.0503206179,-0.2876914442,0.3686482608,0.2929595709,-0.1519884765,-0.3081815243,0.348005116,0.2104075551,-0.3964599669,0.1499019563,0.1032108441,0.0711078048,0.0543967709,-0.3257026672,-0.0879754275,-0.0482918657,0.046583686,-0.2456073761,-0.0358539894,-0.2566569746,-0.3148053586,0.3563577533,0.1371207088,-0.1007086784,-0.1744980961,0.313700676,0.0182477366,0.1877688915,-0.0360825472,-0.0464049093,-0.0668418407,0.1393823922,0.2138672173,-0.2586386502,-0.3567465544,-0.0580908693,-0.1454874873,-0.3680625558,0.3555624187,-0.1243837029,-0.1583322734,-0.1762834042,0.3298704624,0.0042307782,0.1877345294,0.0301690847,-0.0621547922,0.0345112197,-0.0330368169,-0.5213058591,0.2617722154,-0.1892851293,0.1511415094,0.3945093155,-0.0811316222,-0.557228744,0.0285614729,-0.0901461467,-0.0746213943,0.0081122303,0.0527405776,0.0504554845,0.0249890964,0.2616671026,0.3452384472,-0.2784825265,-0.4186275303,-0.1332170814,0.0851489827,0.0555993468,-0.1165198013,0.3011631668,0.3788214028,-0.2120733559,0.054980211,-0.0322543755,-0.2884773612,0.0911688805,0.1308811903,0.0294637401,-0.1145010218,0.3699089885,-0.245359078,0.255815357,0.3970482349,-0.1166401356,-0.0859459639,-0.013187753,-0.3108522892,0.2506210804,0.1698597074,0.1149440333,0.2961706817,0.040193744,-0.144576937,0.1179864109,-0.2378507107,0.1796547174,-0.0705138668,-0.4775745571,0.0499942638,-0.1605377942,0.3734444976,-0.2396960407,0.0875542983,0.0259527098,0.0550787188,0.0277316123,0.1427137405,-0.1049511135,0.0520016216,-0.0014738799,0.1725534648,0.0775345117,-0.2465997636,-0.0430273972,0.1976915747,-0.2594724894,0.1988372803,0.0102924304,0.1281149089,0.1632475555,0.3542839587,-0.0841073245,0.0624637976,0.1840746701,0.0511635952,0.3145600259,-0.0667686909,-0.0732243359,0.1928116381,0.4372864068,0.3121578693,0.0745019764,0.0863800198,0.4601459503,0.1916143447,0.0246332455,0.316878438,0.0409519598,-0.1502854526,-0.3962423205,-0.0534962192,-0.3450787365,-0.2443112284,0.0385401808,0.0660979599,0.2023185045,-0.0611171424,0.0533979386,-0.0326912031,-0.0049795606,-0.0337296873,0.0840256214,-0.3494595289,0.2607451379,0.1190831736,-0.0260091536,0.1218363717,0.2161661386,0.5713266134,0.2512052059,0.1727866381,-0.207377255,-0.164415583,-0.1553098857,0.1315274686,0.0071028308,0.0951862782,-0.090344198,0.2418675125,0.3023291528,-0.3576982021,0.2153081,0.18498528,-0.0085020065,0.0095105013,-0.0305110198,-0.1985211968,-0.0481644832,-0.3636624515,0.0193186495,-0.4257338345,-0.1030887142,0.097545512,0.1103422269,-0.055046279,0.0044380128,0.179121837,0.1831070185,0.3706059456,0.0624001436,0.1879453808,-0.0618245266,-0.373886317,-0.1120659411,0.2231788337,-0.2363315374,-0.1832208633,0.0711996034,0.2132915109,0.0434757285,-0.0104120784,0.3475263417,0.2639898062,-0.0352098458,0.2855528295,-0.2770672441,0.1265669763,0.1273532212,-0.176458165,0.2080287784,-0.1252673566,0.0252976324,-0.0786890984,0.3584910333,-0.1522973478,-0.1829757839,-0.1008876562,-0.0752158165,0.2810131609,0.0117124869,-0.2090331763,-0.3099663258,-0.2741809189,0.1419515759,-0.1518728733,-0.2074362636,0.4317160547,0.1817796677,0.1159687266,0.0878034756,-0.3782545924,-0.2854495347,0.0264465325,-0.1224921644,-0.2001462877,-0.2672688663,0.1795648336,-0.1553334743,0.033318568,-0.1667780727,-0.0431229621,-0.1626894474,-0.2106925398,-0.4100949764,-0.6633718014,0.5129126906,-0.3869130313,-0.3491482139,0.1915044636,0.0613825023,-0.0574717261,-0.1136663184,-0.2272025794,0.1705521494,0.195565924,0.0586176664,-0.1175833866,0.1158121228,0.1681635529,0.0610022396,-0.0177463181,0.0851602182,0.2769685984,-0.1020570174,-0.1372485757,-0.3400894105]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2327","title":"A syntax error in example","comments":"This issue has been fixed.","body":"![image](https:\/\/user-images.githubusercontent.com\/6883957\/117315905-b47a5c00-aeba-11eb-91eb-b2a4a0212a56.png)\r\n\r\nSorry to report with an image, I can't find the template source code of this snippet.","comment_length":5,"text":"A syntax error in example\n![image](https:\/\/user-images.githubusercontent.com\/6883957\/117315905-b47a5c00-aeba-11eb-91eb-b2a4a0212a56.png)\r\n\r\nSorry to report with an image, I can't find the template source code of this snippet.\nThis issue has been fixed.","embeddings":[-0.0118735088,-0.4756166637,-0.1859151572,-0.0986319557,0.0613417216,-0.238526687,0.2298802882,0.2826513052,-0.3375259936,0.2428027093,0.1962114573,0.4457080066,0.0382778496,0.0966375992,0.0458242446,-0.1407996863,0.0370828472,0.3030714691,0.0513857938,0.0391901992,-0.1129494533,0.4401147068,-0.21532996,0.2181288153,-0.1851880699,0.0171554685,0.0521536209,0.2690093517,-0.2384880483,-0.0024255915,0.0017096375,0.0385675468,-0.1389970928,0.0572607517,-0.0000911048,0.0489375703,0.4487837851,-0.0476382971,-0.1015279666,-0.0264369268,-0.2779799402,0.140307337,-0.1659363657,-0.2238977849,-0.2873155475,-0.1644967347,0.1532058567,0.2659866214,0.6600322723,0.1470054537,0.4188614488,-0.0440656394,0.0461575277,-0.2770752609,0.199544549,0.0956225395,-0.0665788651,-0.0370206237,0.3342565298,-0.0652861446,0.2149709612,0.6437575221,0.105946511,-0.1021540314,0.1283149719,-0.0719493181,0.0504253544,-0.0767917186,0.149156183,0.1728185266,0.2160967737,-0.0678712428,0.0330826081,0.214944303,-0.1757199317,0.0753652826,0.0837428644,0.0775495693,0.0081048738,0.1680189073,-0.1371504217,-0.3096079826,-0.3956130743,0.0569132268,0.1497615427,0.0634156168,-0.3012185991,-0.1133846864,-0.0975861847,-0.0369499177,0.1208696067,-0.0730855316,-0.1593846232,0.0080190254,0.3419110775,0.0234312024,0.2852382958,-0.1295316964,0.1128700376,0.2427001745,0.3477541506,0.1722848564,0.0942681655,0.445274055,-0.0799641684,-0.1900733262,0.0682659224,0.0840843543,0.1531047672,0.258662641,0.1412904412,-0.1685145497,-0.1986907721,-0.4010641575,-0.1651077121,-0.0590783618,0.1589517742,0.1053389907,-0.02514356,-0.1209298372,0.0227359571,0.0496426634,-0.0184312053,0.3085201979,0.2068647742,-0.2300750762,0.228004232,-0.0509530678,-0.070471473,-0.1910655051,-0.427408278,0.3994250894,-0.3194102645,-0.3580946624,-0.153330937,-0.0592665151,0.2196889669,0.1419698894,0.3506931365,-0.1709810495,0.2568647861,0.1868283302,0.1508562863,0.0378292166,-0.0011356192,-0.0310296901,0.1778866947,-0.0833840594,0.0614405572,0.1591362506,-0.3388735354,-0.0444243774,-0.3049573898,0.3518997729,0.0171826817,0.1318827271,0.0140864505,0.144548431,0.1997556239,-0.0069012237,-0.0309162494,-0.0215710886,-0.0140896309,-0.2013436109,0.1197718978,0.1807678491,-0.0774667785,-0.0217394102,0.0397043265,-0.1710309386,0.1538698822,0.0714838728,-0.1065080315,0.1797307283,-0.3849446476,0.3694549501,0.2014430016,-0.2174983919,0.1586993039,0.3259782493,-0.1465036273,-0.2208780944,-0.1116874889,-0.1583962739,0.0392137393,-0.0301770922,-0.0864853859,0.1970350444,0.0904103667,0.1601937413,-0.1909549385,0.0669389665,-0.243189469,-0.0145559991,-0.0193894375,-0.0344510525,0.07652089,0.2466625869,0.144576475,-0.2588056028,-0.028578274,0.156402573,0.4285593033,-0.1295853853,0.0240366161,0.0289857127,0.1514994949,-0.1761446446,-0.2805260718,0.0410140231,0.0147129465,0.1389840394,-0.3106948435,-0.0238345284,-0.1409646273,-0.2755191922,0.4462563097,0.0716675222,-0.0135540105,-0.1375795901,0.0376910716,0.0157217998,-0.184227109,-0.0851499364,0.0003319906,0.0841758251,-0.3456303477,-0.0600308813,0.1064865068,-0.2507141829,0.0242480375,-0.0535018481,-0.0943570808,0.3823904097,0.0868310034,0.0311326049,0.0100856461,0.0432339944,-0.1087483093,-0.5943900347,-0.0350195803,0.0872334093,0.0497554243,-0.0432701707,-0.0877576172,0.1858593822,0.1008690745,-0.1118587479,0.0972467512,0.053293407,0.0982277468,-0.0198877975,-0.0688891858,-0.1459852159,-0.027413737,0.2089174241,-0.0462878607,-0.2401484549,-0.2974525094,-0.0386204422,0.2865251005,0.0217563417,0.1833830774,0.0665158033,0.0495774671,-0.0618683621,0.1191912144,-0.1884271801,0.030412754,0.1681758612,-0.066722326,-0.0584033281,0.0078471182,-0.2769371271,0.2903131545,0.0476122759,0.1091128588,-0.0333741829,0.0360653102,-0.1859167665,-0.2619710267,-0.1375338584,-0.1496951431,-0.086035274,-0.1292456239,0.1795493066,-0.1785801798,-0.1939334124,-0.0254204161,0.1177024841,0.0318057761,-0.1906790584,0.3797313869,0.0281380676,-0.1207373515,0.3412282169,0.2189115435,0.15360111,0.084371075,0.1785039902,-0.2613633275,0.0675873533,-0.1763108224,0.3513890803,-0.0063586817,-0.1324241161,0.1278689802,-0.2071157247,0.2576618791,-0.181637913,-0.4966984689,0.2716535926,0.1200555041,0.1814741194,0.3359618783,0.2755220532,-0.1518499702,0.0855754763,0.3804481328,-0.0797175765,-0.2430215627,0.04363757,0.0031873859,-0.1226344407,-0.3872664869,0.2184130251,0.1752150059,-0.2789428234,-0.1340549141,0.141140759,0.1539719552,0.3966163397,0.5041601658,0.1624166667,-0.0739545971,0.106215246,-0.0448817387,-0.3024614751,0.1741007417,-0.362698704,-0.3095275462,-0.153504923,0.0225593951,0.1695298404,-0.1606042534,0.0130910473,-0.3478903174,0.141198203,-0.0967196524,-0.0479642525,0.0440582559,-0.0139261754,-0.0150281023,-0.4043776691,-0.546507895,0.1407913864,0.1921581626,-0.456159234,0.235895589,0.0251873489,-0.0938252285,0.1936469376,0.4945642054,0.0848060995,-0.1145045012,0.2549591959,-0.013876833,0.3474779725,0.0166823566,0.0105367471,0.1328596473,0.3930579424,0.0113363862,0.1869530976,-0.0382954516,-0.0905663967,-0.0588897653,-0.0444092192,-0.2099647075,0.0580569766,-0.0249857213,0.3559287488,0.1430826336,0.0610508099,-0.053201206,0.0903323144,0.2320184857,0.2355131358,0.3929509223,0.2172281891,-0.0222721156,-0.0515003093,-0.1759861261,0.0734780058,0.1053051203,0.0672559738,-0.1925124526,-0.1360677183,-0.2199809849,0.0695726499,-0.3006555438,0.4702368379,-0.233793512,-0.1446689814,0.1714849472,-0.0967562199,0.0681465939,-0.0170142036,-0.2629466355,0.3398911953,0.2267022729,0.0610519648,-0.2379565537,-0.0491060503,0.0433988757,-0.4809137583,-0.1612307131,0.0210759696,-0.0735974461,-0.2680262327,-0.3099285364,-0.173480168,0.2365290821,0.0655084923,0.2161346823,-0.2169105262,-0.0007297086,-0.1451392174,-0.1173834428,0.2193903774,0.1770615429,-0.1872045845,-0.2067737281,-0.1737493128,-0.0166719649,0.1236905381,-0.0556234866,0.0081686405,0.0285372268,0.0883491859,-0.0616723411,0.0872127488,0.1240079701,-0.3175060749,-0.1877035499,-0.0772553161,0.1528622955,0.1347176135,0.0158923287,0.1105653569,0.1152037233,0.0686257258,-0.254288733,0.202769354,0.0082229301,0.0048476704,0.2451908439,-0.0653726086,-0.1300888509,-0.0313946456,0.2423349172,0.9280289412,0.1525127441,-0.0143706659,0.3044354916,-0.1352744102,0.1288121641,-0.0616883263,-0.0877853036,-0.3887127042,-0.0807461813,0.1142654717,-0.1794138104,0.1104053557,0.1382091045,-0.0661727861,-0.2329590172,-0.1809568554,0.1779400408,-0.009058686,-0.0208545458,0.1184176356,0.1050031632,0.0162174832,0.3563632667,0.080527477,0.0443560407,-0.2586807907,-0.3839924335,-0.1645181775,-0.3315911889,-0.1955628842,0.3054246902,-0.1351272911,0.1945282221,0.274358362,-0.2230660319,0.0728967488,-0.1011383384,0.0058207801,0.0777702853,-0.2349387258,0.3144372702,0.28932482,-0.274161756,-0.3331644237,0.3498936892,0.1158920899,-0.3798339963,0.0816859007,0.0912798196,0.1225518361,0.0238727462,-0.2456092685,-0.1109407246,-0.011000067,-0.0149011649,-0.2309147865,-0.087234877,-0.2012457103,-0.2860348225,0.3764710426,0.1447373629,-0.0815802068,-0.2083890438,0.3181856573,0.0662888512,0.1311181933,0.0099091856,-0.020081846,-0.0364569128,0.154012531,0.2153555304,-0.2754331529,-0.3943071663,-0.0844020098,-0.1401327252,-0.3308058381,0.3105459213,-0.0752886161,-0.1842616647,-0.1209640205,0.2698129416,0.0397490263,0.1870062947,0.0603011921,-0.0817918926,-0.0002977088,0.0275868867,-0.4221720099,0.2758831382,-0.1440282911,0.1909467131,0.3073821068,-0.0465305597,-0.5807148218,-0.0153328795,-0.1475233585,-0.1240901053,0.0056560868,-0.020457359,0.0461036488,0.0809650645,0.3145197928,0.2973503172,-0.3198586404,-0.4533140063,-0.0967136174,0.0660653785,0.0315237604,-0.0838940069,0.2313832343,0.2938205302,-0.2756278813,0.0286722034,-0.0084861498,-0.3068834841,0.1043148562,0.0906223357,-0.0191168413,-0.1320683211,0.241669178,-0.2630804777,0.2172098011,0.3661402464,-0.0968275666,-0.1235052645,0.0031151401,-0.3037320375,0.1837209016,0.1450288445,0.0937537849,0.2774573267,0.0913138613,-0.0891552791,0.0657541603,-0.1630048603,0.1344282925,-0.084848538,-0.454559654,0.0899854079,-0.1354420185,0.4138968289,-0.2556310296,0.0510538183,-0.0348963216,-0.0136576984,0.0359036587,0.1451810896,-0.130389899,0.0649866462,0.0337808132,0.1860633045,0.0588778518,-0.183022052,-0.0639667213,0.2342568487,-0.2067485452,0.1969847977,0.0631553978,0.1834882498,0.1423412561,0.3240292966,-0.0630464032,0.097760193,0.0645396188,0.1180393621,0.229320839,-0.099866569,-0.0990235731,0.2133854926,0.3539803624,0.273658216,-0.0093258917,0.1021121442,0.3707510829,0.2241242826,-0.0494319759,0.3683395982,-0.0185502302,-0.1177773327,-0.4142472446,-0.0554157682,-0.2833877802,-0.2338245213,0.1282663345,0.1223165914,0.1245310679,-0.0741674006,0.0251099784,0.0198555682,0.0080067543,0.0155800385,0.0797508284,-0.3011894822,0.2260739058,0.0404818356,-0.001582237,0.102598533,0.1859733015,0.5189894438,0.2380378246,0.1878685951,-0.2005041689,-0.1146204919,-0.2035016418,0.0171502773,-0.0239701904,0.1229626536,-0.0906273276,0.2218311727,0.3215356171,-0.4031616449,0.1100145727,0.1826948375,0.0054884888,0.0837630779,-0.0242447406,-0.231485188,-0.035292238,-0.4187523425,0.0094800005,-0.4549061656,-0.0416707695,0.1120540127,0.1139818355,-0.0293285679,-0.0577524565,0.1926025748,0.084190622,0.3568504751,0.0557613894,0.2002831548,-0.0152188893,-0.4316188991,-0.1256766021,0.2646879554,-0.200295791,-0.2117464244,0.0279254988,0.2018200755,0.0242557917,-0.0616906695,0.2695121169,0.2525903881,-0.0037238246,0.2595821321,-0.2765399516,0.1013205871,0.0642145276,-0.1428593695,0.2367432266,-0.1159876138,-0.0369600654,-0.0134008061,0.396093905,-0.1495370716,-0.1511077285,-0.1170059741,-0.1297816187,0.3263142407,-0.0389712788,-0.1682530791,-0.3299039006,-0.2207044214,0.1530582309,-0.1525356472,-0.1607644856,0.4000603259,0.2518279552,0.1326845884,0.1169493496,-0.329197228,-0.2463496476,0.0218351409,-0.024335254,-0.1746138036,-0.2534829378,0.2153192461,-0.176276505,0.0309328549,-0.1420364976,0.0354089029,-0.133632496,-0.2205446362,-0.3692039549,-0.662027657,0.4939661324,-0.407941848,-0.3204495311,0.1834515184,0.0535863414,-0.128516674,-0.0915778801,-0.2438147515,0.2013573796,0.1480086297,0.0382662751,-0.0794220269,0.1438975185,0.1678116471,0.091512166,-0.0269079823,0.159956336,0.2577880919,-0.0804404616,-0.1150569469,-0.3235681951]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2323","title":"load_dataset(\"timit_asr\") gives back duplicates of just one sample text","comments":"Upgrading datasets to version 1.6 fixes the issue","body":"## Describe the bug\r\nWhen you look up on key [\"train\"] and then ['text'], you get back a list with just one sentence duplicated 4620 times. Namely, the sentence \"Would such an act of refusal be useful?\". Similarly when you look up ['test'] and then ['text'], the list is one sentence repeated \"The bungalow was pleasantly situated near the shore.\" 1680 times. \r\n\r\nI tried to work around the issue by downgrading to datasets version 1.3.0, inspired by [this post](https:\/\/www.gitmemory.com\/issue\/huggingface\/datasets\/2052\/798904836) and removing the entire huggingface directory from ~\/.cache, but I still get the same issue. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(timit['train']['text'])\r\nprint(timit['test']['text'])\r\n```\r\n\r\n## Expected Result\r\nRows of diverse text, like how it is shown in the [wav2vec2.0 tutorial](https:\/\/colab.research.google.com\/github\/patrickvonplaten\/notebooks\/blob\/master\/Fine_tuning_Wav2Vec2_for_English_ASR.ipynb)\r\n\"Screen\r\n\r\n\r\n## Actual results\r\nRows of repeated text.\r\n\"Screen\r\n\r\n\r\n## Versions\r\n- Datasets: 1.3.0\r\n- Python: 3.9.1\r\n- Platform: macOS-11.2.1-x86_64-i386-64bit}\r\n","comment_length":8,"text":"load_dataset(\"timit_asr\") gives back duplicates of just one sample text\n## Describe the bug\r\nWhen you look up on key [\"train\"] and then ['text'], you get back a list with just one sentence duplicated 4620 times. Namely, the sentence \"Would such an act of refusal be useful?\". Similarly when you look up ['test'] and then ['text'], the list is one sentence repeated \"The bungalow was pleasantly situated near the shore.\" 1680 times. \r\n\r\nI tried to work around the issue by downgrading to datasets version 1.3.0, inspired by [this post](https:\/\/www.gitmemory.com\/issue\/huggingface\/datasets\/2052\/798904836) and removing the entire huggingface directory from ~\/.cache, but I still get the same issue. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(timit['train']['text'])\r\nprint(timit['test']['text'])\r\n```\r\n\r\n## Expected Result\r\nRows of diverse text, like how it is shown in the [wav2vec2.0 tutorial](https:\/\/colab.research.google.com\/github\/patrickvonplaten\/notebooks\/blob\/master\/Fine_tuning_Wav2Vec2_for_English_ASR.ipynb)\r\n\"Screen\r\n\r\n\r\n## Actual results\r\nRows of repeated text.\r\n\"Screen\r\n\r\n\r\n## Versions\r\n- Datasets: 1.3.0\r\n- Python: 3.9.1\r\n- Platform: macOS-11.2.1-x86_64-i386-64bit}\r\n\nUpgrading datasets to version 1.6 fixes the issue","embeddings":[0.2821879089,-0.4890296757,0.0636294037,0.3835485876,0.1550767869,0.0093223024,0.2025510371,0.1971771866,-0.1960430145,0.0225637201,-0.0537860543,0.2588163614,-0.0607607588,-0.0019165596,0.1897579134,0.0341104642,0.1301131546,-0.0039515593,-0.0855849162,-0.2164427489,-0.0349717848,0.5018290877,-0.1255064309,0.039624732,-0.3864379525,0.0896656886,-0.2256180942,0.0954400674,0.0274894368,-0.4388808608,0.2015984505,0.2562795579,0.0076610618,0.1743332148,-0.0001294062,-0.1238207743,0.103335917,0.1966973394,-0.2948189676,-0.337613523,0.0371453874,-0.0082015544,0.476550132,0.0588161834,-0.1446308494,-0.1229952723,0.0461377911,-0.0911030397,0.5953748822,0.0530859642,0.0554759242,-0.0213040039,-0.2731775641,-0.1502211094,0.30012393,0.1408754438,0.0361843631,0.0638974905,0.2574379146,0.0740013942,0.0937062427,0.4073674381,-0.2614638805,0.0803650841,0.0201924555,0.2505540252,-0.3521604538,-0.2842356563,0.1567363143,0.3818914294,0.4144081473,-0.1949874014,-0.2438047528,-0.5332146287,0.161943242,0.0539523214,0.1597354859,-0.0827361718,-0.239992395,0.132328406,-0.272169441,-0.0274029933,-0.0044496772,0.0038483818,0.1031661034,-0.2842705548,-0.1285128742,0.1618483514,0.3656548262,-0.0383118913,-0.0753812864,-0.0201201346,-0.0413320735,0.1814562082,-0.4986139834,0.0300714653,0.0270303711,-0.0612340719,0.2511675954,0.0161169358,0.4327557385,-0.0579363778,-0.1301839948,-0.0773007572,0.41262725,0.3775301576,0.0727243051,-0.0954676718,0.0092438171,-0.1186076403,-0.2098524123,0.0403310657,-0.0111010782,0.1610854268,0.0967437252,-0.2585274279,0.3425212801,-0.3690470159,-0.4510843158,0.0706897229,-0.5093212724,-0.1209715977,-0.1602241993,0.093509011,0.1373853534,0.3813998401,0.1867179424,0.1462110281,-0.1727949232,-0.2743343115,-0.134309411,-0.1454474628,-0.0307364892,0.1082837433,0.1738090068,-0.5300574899,0.2907099128,0.3179664016,0.341807574,-0.4182767868,-0.1233289689,-0.0930299759,-0.1239289194,0.0292858966,-0.0926786661,0.4591899216,0.0831208751,-0.0351972915,0.0112058921,-0.0125043103,-0.2134313434,-0.1196761578,0.2937436998,-0.0058470955,-0.2225892097,-0.0508112013,0.0654349998,0.3292100132,0.0748777837,0.1093474776,-0.1791653037,-0.0059522255,-0.4267409444,-0.1410837471,0.0414114632,0.3480868042,-0.2709081471,-0.0908948332,0.2106882185,0.2735222876,-0.0548251383,0.5467896461,-0.0427984148,0.4276206791,-0.2642585337,0.0796886086,0.0182303637,-0.4098272324,-0.2936970294,0.2606933713,-0.1333964318,0.3572215736,0.2087152153,0.1109746322,0.2958289087,0.1419264823,0.1747303605,0.0112375775,0.3328603804,-0.1269817948,-0.2913201153,-0.2165751457,0.0697632954,0.0462378487,-0.3213202953,0.213571474,0.0712145939,0.0611044355,0.4559496045,0.0401258767,-0.0278414022,0.0761946365,0.1083892062,0.0714350268,0.3267050087,-0.057660155,-0.2763386369,0.0627625287,0.1535568237,0.0899405926,0.1892700791,-0.0697184727,-0.2410974801,-0.0870902613,-0.3644341528,-0.1309642196,-0.0653520897,0.3186425269,0.0336826257,0.0968320817,-0.070033446,0.487472713,-0.1397870034,0.0868068263,-0.3534024656,0.0399518311,0.1423211247,0.078276664,0.2095784694,0.3197647631,0.2262900472,0.1240757406,0.0834152624,0.2634825408,0.029740002,0.1141453683,-0.1574550569,-0.212614432,0.1984330416,-0.2226008475,-0.0833782852,0.1785469055,0.2592630982,-0.1271862537,-0.2556278706,0.195391506,0.1241352111,0.2335562855,-0.0585896485,0.021532936,0.0859504044,-0.0684443489,-0.1082612723,-0.2807025909,0.5420552492,-0.1498948187,0.218012318,0.1538831294,-0.3125824928,0.2360616326,0.2239007503,-0.0717561767,-0.0559593439,0.3721970022,-0.1574176699,-0.0650917515,-0.1354452372,0.0964777842,0.3051019311,0.1097943187,0.0447403416,0.027225472,-0.1308660805,0.045597259,0.2455358952,-0.0921415836,-0.0883161649,0.3082364798,0.3380037546,-0.0945692807,-0.598446548,0.1292411536,-0.0442171581,0.2760109007,-0.3640033603,0.012894542,-0.3676787019,0.0385611914,-0.3294149637,-0.026490096,-0.2005130351,-0.1136622652,0.188616544,0.3473803699,0.2270390242,0.0175411087,0.2276711464,0.3561970294,-0.0489334911,0.2410355657,-0.08919736,0.1284423023,-0.3455183506,-0.0424879417,-0.0230716765,0.0373932831,0.2573551238,-0.1349247098,-0.3074090779,-0.4002928138,-0.2227882594,0.253131032,-0.2650930285,0.410592556,0.3220893443,-0.0436666459,-0.4077951908,-0.2074154466,0.0193402227,-0.0297960863,-0.401899308,0.2117161304,-0.0361040346,0.0585928671,-0.0396243893,-0.5401884317,0.0536182858,0.050110504,0.5322161913,0.0152634261,0.0834391713,0.2894309461,0.0147456396,0.0821889639,-0.188985303,0.1407686025,-0.5706542134,-0.2644888163,0.1145198941,0.155229792,-0.2780033648,-0.2397743165,0.0120847588,-0.0123424381,-0.1511860639,-0.5954256654,-0.262006253,-0.1077622399,0.0324840359,-0.1540211886,0.3147455752,0.22699669,0.1243949309,0.0203050952,-0.216123566,0.0191078782,0.5199985504,0.1948654354,0.1195437759,0.107226871,0.1179140136,-0.0223576836,0.3257386088,0.8683472276,0.1731048524,0.2302144766,-0.1726848483,0.1653012335,-0.2504996061,-0.3475958109,-0.2155773193,-0.2688345611,0.0127172098,0.431297034,0.1373870373,0.140871346,-0.1646451652,0.1525147706,-0.1869924366,-0.2055777758,-0.1615321636,-0.1367037445,0.11357636,0.1412164569,-0.028997859,0.2722263336,-0.2952245176,-0.1493304074,-0.0974014029,0.1342126131,-0.1416782141,-0.3199513853,-0.074440755,-0.6836324334,0.0644329041,0.2459241897,0.2961187065,-0.0062791235,-0.0122515298,0.4832975864,0.2160003483,1.0575666428,-0.0956710875,0.0187031496,-0.2254541218,-0.0384578519,-0.2621954978,0.0948655829,-0.2956033945,-0.1604373455,-0.0091202706,0.1500043571,-0.1343819499,0.0799587816,0.053867653,0.4325528741,-0.2853339016,-0.1934048533,-0.3532069623,-0.2623336911,-0.3100244105,-0.0245487355,-0.018167289,0.1022895724,-0.2388502806,-0.0308564622,0.0258367434,-0.1209612638,0.4737134278,0.1313102841,0.6074252725,-0.0171290226,0.4328970909,0.5953530073,-0.1411606222,-0.3634350896,0.8624765873,-0.1211226508,-0.2612604797,-0.1437160224,-0.3327286541,0.2774755657,0.6624307036,-0.2186656594,0.3310763538,-0.1124000996,-0.0247950163,-0.2145534605,-0.1021718904,0.3796004653,0.2069099247,0.0043728543,-0.3903522491,0.3599166572,-0.0825879648,-0.2580901682,0.1440051347,0.0782555342,-0.2731482983,0.325918138,-0.2066819668,1.1999918222,-0.1144596189,0.4627687335,0.0744244978,0.1892920434,0.4935740829,0.0038518589,-0.0439088121,-0.3514220417,-0.076208517,-0.0938368961,0.0494186021,-0.1900263727,0.0923126638,-0.3971046805,0.2927147448,0.3565191031,0.1551822871,0.0880177766,0.0852866322,-0.0418357998,-0.0442516059,-0.1079174951,-0.1045988426,-0.0658707991,0.1889068186,0.257589817,0.1798462123,0.2313103527,-0.5975795984,-0.476439774,0.0072849481,-0.3215946853,-0.0100113079,0.1055299491,-0.3966953158,0.0478975177,0.3964281082,0.1667467803,0.1279726624,-0.1238549054,-0.1648803502,0.2242880017,0.428678751,-0.1358835846,-0.2301557213,-0.0153529942,0.0764355883,-0.441925317,-0.0202909205,0.2607076168,-0.0377194807,-0.25432688,0.1783215702,-0.2497574389,-0.3596180975,0.0049403296,0.2692559063,0.1709797382,-0.3426191211,-0.0310657229,0.1471382678,-0.3102154732,0.088294372,0.0200827327,-0.3249722123,0.0588383414,0.7058208585,0.3741140962,-0.1189956516,0.5002724528,0.2675578594,0.0116620082,-0.0957498252,0.226036936,-0.0397436507,-0.4815365672,0.2151355743,-0.002908526,0.09179423,-0.2900525331,0.426771909,-0.021726232,0.1830295175,-0.1333239675,-0.2961260676,-0.0423954651,0.2187686116,0.13949655,0.1619938612,-0.1724784672,-0.1619065553,0.0097095631,0.1109667495,-0.1637848616,0.1157054976,0.0443555899,0.1690419018,0.0517503396,-0.0303461272,0.2929798961,0.0675780475,0.071134679,0.2659066021,0.0648534074,-0.09327811,-0.2981212139,0.2246005237,-0.0966291726,0.1079116836,-0.0679994002,-0.0075088846,0.2013832331,-0.199802503,0.2344643027,0.3591980934,0.1373309791,0.129059419,-0.1139102578,-0.1763348281,0.1196943,-0.0580043867,-0.3185044229,-0.0318246596,0.3222025931,-0.000190768,0.0447842069,-0.0174140316,0.1147963405,0.1557329744,0.147576198,-0.1878322363,-0.2165790349,-0.1440662146,-0.1786869168,0.0539104007,0.3492898941,0.3345215321,-0.3116487861,-0.1201452538,0.0710493997,0.024994174,-0.3629961014,-0.3286014795,-0.0276242942,-0.0674707219,-0.110719882,0.4090248346,0.1963606477,-0.3866886199,0.2300868034,-0.1666325033,0.267883569,-0.2556222081,0.1114893332,0.4208318591,0.1170658842,-0.2309045345,-0.0539012924,-0.102771759,-0.0972770974,0.5784748197,-0.365550518,0.0661406144,-0.4556359351,0.3279430866,0.2335660458,-0.1663713306,0.3428981304,0.2844083607,-0.3349226415,0.2329865247,0.0152564663,0.5815436244,-0.0264838561,-0.0638352633,-0.2792891562,0.1324142665,-0.1662414074,-0.3447313905,-0.0006581515,-0.3484948874,-0.3311074674,0.2563079,-0.0774773061,-0.2266082764,0.5246610641,0.1639321595,-0.0409048982,-0.7086568475,-0.3736552596,-0.0045372285,0.2766523361,-0.2027623653,0.4182367325,0.3229885697,0.0039526508,0.3944139779,0.4252387285,0.3170160949,0.0883705392,-0.0516745709,0.3155106008,-0.4504492581,-0.0059706555,0.0368653499,0.3054410815,-0.1242536828,-0.1066462398,0.0966200233,-0.102817066,0.0515899286,0.0424061045,0.2062606364,0.0443429165,-0.5337705016,0.2679801881,-0.1764916331,0.1740657091,-0.0651561692,0.1449666619,-0.5101766586,0.2612885237,0.3920439184,0.0583924651,0.0105528086,-0.0828596056,0.0329217464,-0.1383634061,0.2497992963,0.2675231993,0.1021079421,-0.2600739002,-0.1450134516,-0.5910230875,0.0332322828,-0.1908808649,-0.0054462044,0.1442645788,-0.0413945653,-0.0358709693,0.2769308388,-0.0004990975,0.0354808979,-0.4129131138,0.2628389001,-0.2723473907,-0.2296372205,0.2392846048,0.2164620608,-0.1023936942,-0.1503431052,0.1333624274,-0.0975973085,-0.1930176318,-0.0528880693,0.1203178987,-0.2028674036,-0.0992030203,-0.025235869,0.2476500273,0.2405091673,0.0045695081,-0.1746524721,-0.1247410849,-0.2587486804,-0.3179563582,0.3191432953,0.2317633778,0.3002185524,-0.1924441308,-0.2965009809,0.0314552374,0.1041969284,0.282802254,0.0189734064,-0.4451439083,0.1156836525,-0.0879513696,0.3841794133,-0.1302316636,0.465686053,-0.0369808264,0.183060497,-0.1826772392,-0.568195045,0.4401472509,-0.6994359493,-0.1401675195,-0.0263723247,0.3481683433,0.0437882319,-0.2374210656,-0.9038203955,-0.1497897804,0.3655125797,0.0880114585,-0.1035750508,0.3861141205,-0.0456485115,-0.0585038513,-0.148834005,0.3935339749,0.4965204895,-0.1668690443,0.3587186635,-0.5926187634]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2323","title":"load_dataset(\"timit_asr\") gives back duplicates of just one sample text","comments":"This bug was fixed in #1995. Upgrading the `datasets` should work! ","body":"## Describe the bug\r\nWhen you look up on key [\"train\"] and then ['text'], you get back a list with just one sentence duplicated 4620 times. Namely, the sentence \"Would such an act of refusal be useful?\". Similarly when you look up ['test'] and then ['text'], the list is one sentence repeated \"The bungalow was pleasantly situated near the shore.\" 1680 times. \r\n\r\nI tried to work around the issue by downgrading to datasets version 1.3.0, inspired by [this post](https:\/\/www.gitmemory.com\/issue\/huggingface\/datasets\/2052\/798904836) and removing the entire huggingface directory from ~\/.cache, but I still get the same issue. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(timit['train']['text'])\r\nprint(timit['test']['text'])\r\n```\r\n\r\n## Expected Result\r\nRows of diverse text, like how it is shown in the [wav2vec2.0 tutorial](https:\/\/colab.research.google.com\/github\/patrickvonplaten\/notebooks\/blob\/master\/Fine_tuning_Wav2Vec2_for_English_ASR.ipynb)\r\n\"Screen\r\n\r\n\r\n## Actual results\r\nRows of repeated text.\r\n\"Screen\r\n\r\n\r\n## Versions\r\n- Datasets: 1.3.0\r\n- Python: 3.9.1\r\n- Platform: macOS-11.2.1-x86_64-i386-64bit}\r\n","comment_length":11,"text":"load_dataset(\"timit_asr\") gives back duplicates of just one sample text\n## Describe the bug\r\nWhen you look up on key [\"train\"] and then ['text'], you get back a list with just one sentence duplicated 4620 times. Namely, the sentence \"Would such an act of refusal be useful?\". Similarly when you look up ['test'] and then ['text'], the list is one sentence repeated \"The bungalow was pleasantly situated near the shore.\" 1680 times. \r\n\r\nI tried to work around the issue by downgrading to datasets version 1.3.0, inspired by [this post](https:\/\/www.gitmemory.com\/issue\/huggingface\/datasets\/2052\/798904836) and removing the entire huggingface directory from ~\/.cache, but I still get the same issue. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(timit['train']['text'])\r\nprint(timit['test']['text'])\r\n```\r\n\r\n## Expected Result\r\nRows of diverse text, like how it is shown in the [wav2vec2.0 tutorial](https:\/\/colab.research.google.com\/github\/patrickvonplaten\/notebooks\/blob\/master\/Fine_tuning_Wav2Vec2_for_English_ASR.ipynb)\r\n\"Screen\r\n\r\n\r\n## Actual results\r\nRows of repeated text.\r\n\"Screen\r\n\r\n\r\n## Versions\r\n- Datasets: 1.3.0\r\n- Python: 3.9.1\r\n- Platform: macOS-11.2.1-x86_64-i386-64bit}\r\n\nThis bug was fixed in #1995. Upgrading the `datasets` should work! ","embeddings":[0.2821879089,-0.4890296757,0.0636294037,0.3835485876,0.1550767869,0.0093223024,0.2025510371,0.1971771866,-0.1960430145,0.0225637201,-0.0537860543,0.2588163614,-0.0607607588,-0.0019165596,0.1897579134,0.0341104642,0.1301131546,-0.0039515593,-0.0855849162,-0.2164427489,-0.0349717848,0.5018290877,-0.1255064309,0.039624732,-0.3864379525,0.0896656886,-0.2256180942,0.0954400674,0.0274894368,-0.4388808608,0.2015984505,0.2562795579,0.0076610618,0.1743332148,-0.0001294062,-0.1238207743,0.103335917,0.1966973394,-0.2948189676,-0.337613523,0.0371453874,-0.0082015544,0.476550132,0.0588161834,-0.1446308494,-0.1229952723,0.0461377911,-0.0911030397,0.5953748822,0.0530859642,0.0554759242,-0.0213040039,-0.2731775641,-0.1502211094,0.30012393,0.1408754438,0.0361843631,0.0638974905,0.2574379146,0.0740013942,0.0937062427,0.4073674381,-0.2614638805,0.0803650841,0.0201924555,0.2505540252,-0.3521604538,-0.2842356563,0.1567363143,0.3818914294,0.4144081473,-0.1949874014,-0.2438047528,-0.5332146287,0.161943242,0.0539523214,0.1597354859,-0.0827361718,-0.239992395,0.132328406,-0.272169441,-0.0274029933,-0.0044496772,0.0038483818,0.1031661034,-0.2842705548,-0.1285128742,0.1618483514,0.3656548262,-0.0383118913,-0.0753812864,-0.0201201346,-0.0413320735,0.1814562082,-0.4986139834,0.0300714653,0.0270303711,-0.0612340719,0.2511675954,0.0161169358,0.4327557385,-0.0579363778,-0.1301839948,-0.0773007572,0.41262725,0.3775301576,0.0727243051,-0.0954676718,0.0092438171,-0.1186076403,-0.2098524123,0.0403310657,-0.0111010782,0.1610854268,0.0967437252,-0.2585274279,0.3425212801,-0.3690470159,-0.4510843158,0.0706897229,-0.5093212724,-0.1209715977,-0.1602241993,0.093509011,0.1373853534,0.3813998401,0.1867179424,0.1462110281,-0.1727949232,-0.2743343115,-0.134309411,-0.1454474628,-0.0307364892,0.1082837433,0.1738090068,-0.5300574899,0.2907099128,0.3179664016,0.341807574,-0.4182767868,-0.1233289689,-0.0930299759,-0.1239289194,0.0292858966,-0.0926786661,0.4591899216,0.0831208751,-0.0351972915,0.0112058921,-0.0125043103,-0.2134313434,-0.1196761578,0.2937436998,-0.0058470955,-0.2225892097,-0.0508112013,0.0654349998,0.3292100132,0.0748777837,0.1093474776,-0.1791653037,-0.0059522255,-0.4267409444,-0.1410837471,0.0414114632,0.3480868042,-0.2709081471,-0.0908948332,0.2106882185,0.2735222876,-0.0548251383,0.5467896461,-0.0427984148,0.4276206791,-0.2642585337,0.0796886086,0.0182303637,-0.4098272324,-0.2936970294,0.2606933713,-0.1333964318,0.3572215736,0.2087152153,0.1109746322,0.2958289087,0.1419264823,0.1747303605,0.0112375775,0.3328603804,-0.1269817948,-0.2913201153,-0.2165751457,0.0697632954,0.0462378487,-0.3213202953,0.213571474,0.0712145939,0.0611044355,0.4559496045,0.0401258767,-0.0278414022,0.0761946365,0.1083892062,0.0714350268,0.3267050087,-0.057660155,-0.2763386369,0.0627625287,0.1535568237,0.0899405926,0.1892700791,-0.0697184727,-0.2410974801,-0.0870902613,-0.3644341528,-0.1309642196,-0.0653520897,0.3186425269,0.0336826257,0.0968320817,-0.070033446,0.487472713,-0.1397870034,0.0868068263,-0.3534024656,0.0399518311,0.1423211247,0.078276664,0.2095784694,0.3197647631,0.2262900472,0.1240757406,0.0834152624,0.2634825408,0.029740002,0.1141453683,-0.1574550569,-0.212614432,0.1984330416,-0.2226008475,-0.0833782852,0.1785469055,0.2592630982,-0.1271862537,-0.2556278706,0.195391506,0.1241352111,0.2335562855,-0.0585896485,0.021532936,0.0859504044,-0.0684443489,-0.1082612723,-0.2807025909,0.5420552492,-0.1498948187,0.218012318,0.1538831294,-0.3125824928,0.2360616326,0.2239007503,-0.0717561767,-0.0559593439,0.3721970022,-0.1574176699,-0.0650917515,-0.1354452372,0.0964777842,0.3051019311,0.1097943187,0.0447403416,0.027225472,-0.1308660805,0.045597259,0.2455358952,-0.0921415836,-0.0883161649,0.3082364798,0.3380037546,-0.0945692807,-0.598446548,0.1292411536,-0.0442171581,0.2760109007,-0.3640033603,0.012894542,-0.3676787019,0.0385611914,-0.3294149637,-0.026490096,-0.2005130351,-0.1136622652,0.188616544,0.3473803699,0.2270390242,0.0175411087,0.2276711464,0.3561970294,-0.0489334911,0.2410355657,-0.08919736,0.1284423023,-0.3455183506,-0.0424879417,-0.0230716765,0.0373932831,0.2573551238,-0.1349247098,-0.3074090779,-0.4002928138,-0.2227882594,0.253131032,-0.2650930285,0.410592556,0.3220893443,-0.0436666459,-0.4077951908,-0.2074154466,0.0193402227,-0.0297960863,-0.401899308,0.2117161304,-0.0361040346,0.0585928671,-0.0396243893,-0.5401884317,0.0536182858,0.050110504,0.5322161913,0.0152634261,0.0834391713,0.2894309461,0.0147456396,0.0821889639,-0.188985303,0.1407686025,-0.5706542134,-0.2644888163,0.1145198941,0.155229792,-0.2780033648,-0.2397743165,0.0120847588,-0.0123424381,-0.1511860639,-0.5954256654,-0.262006253,-0.1077622399,0.0324840359,-0.1540211886,0.3147455752,0.22699669,0.1243949309,0.0203050952,-0.216123566,0.0191078782,0.5199985504,0.1948654354,0.1195437759,0.107226871,0.1179140136,-0.0223576836,0.3257386088,0.8683472276,0.1731048524,0.2302144766,-0.1726848483,0.1653012335,-0.2504996061,-0.3475958109,-0.2155773193,-0.2688345611,0.0127172098,0.431297034,0.1373870373,0.140871346,-0.1646451652,0.1525147706,-0.1869924366,-0.2055777758,-0.1615321636,-0.1367037445,0.11357636,0.1412164569,-0.028997859,0.2722263336,-0.2952245176,-0.1493304074,-0.0974014029,0.1342126131,-0.1416782141,-0.3199513853,-0.074440755,-0.6836324334,0.0644329041,0.2459241897,0.2961187065,-0.0062791235,-0.0122515298,0.4832975864,0.2160003483,1.0575666428,-0.0956710875,0.0187031496,-0.2254541218,-0.0384578519,-0.2621954978,0.0948655829,-0.2956033945,-0.1604373455,-0.0091202706,0.1500043571,-0.1343819499,0.0799587816,0.053867653,0.4325528741,-0.2853339016,-0.1934048533,-0.3532069623,-0.2623336911,-0.3100244105,-0.0245487355,-0.018167289,0.1022895724,-0.2388502806,-0.0308564622,0.0258367434,-0.1209612638,0.4737134278,0.1313102841,0.6074252725,-0.0171290226,0.4328970909,0.5953530073,-0.1411606222,-0.3634350896,0.8624765873,-0.1211226508,-0.2612604797,-0.1437160224,-0.3327286541,0.2774755657,0.6624307036,-0.2186656594,0.3310763538,-0.1124000996,-0.0247950163,-0.2145534605,-0.1021718904,0.3796004653,0.2069099247,0.0043728543,-0.3903522491,0.3599166572,-0.0825879648,-0.2580901682,0.1440051347,0.0782555342,-0.2731482983,0.325918138,-0.2066819668,1.1999918222,-0.1144596189,0.4627687335,0.0744244978,0.1892920434,0.4935740829,0.0038518589,-0.0439088121,-0.3514220417,-0.076208517,-0.0938368961,0.0494186021,-0.1900263727,0.0923126638,-0.3971046805,0.2927147448,0.3565191031,0.1551822871,0.0880177766,0.0852866322,-0.0418357998,-0.0442516059,-0.1079174951,-0.1045988426,-0.0658707991,0.1889068186,0.257589817,0.1798462123,0.2313103527,-0.5975795984,-0.476439774,0.0072849481,-0.3215946853,-0.0100113079,0.1055299491,-0.3966953158,0.0478975177,0.3964281082,0.1667467803,0.1279726624,-0.1238549054,-0.1648803502,0.2242880017,0.428678751,-0.1358835846,-0.2301557213,-0.0153529942,0.0764355883,-0.441925317,-0.0202909205,0.2607076168,-0.0377194807,-0.25432688,0.1783215702,-0.2497574389,-0.3596180975,0.0049403296,0.2692559063,0.1709797382,-0.3426191211,-0.0310657229,0.1471382678,-0.3102154732,0.088294372,0.0200827327,-0.3249722123,0.0588383414,0.7058208585,0.3741140962,-0.1189956516,0.5002724528,0.2675578594,0.0116620082,-0.0957498252,0.226036936,-0.0397436507,-0.4815365672,0.2151355743,-0.002908526,0.09179423,-0.2900525331,0.426771909,-0.021726232,0.1830295175,-0.1333239675,-0.2961260676,-0.0423954651,0.2187686116,0.13949655,0.1619938612,-0.1724784672,-0.1619065553,0.0097095631,0.1109667495,-0.1637848616,0.1157054976,0.0443555899,0.1690419018,0.0517503396,-0.0303461272,0.2929798961,0.0675780475,0.071134679,0.2659066021,0.0648534074,-0.09327811,-0.2981212139,0.2246005237,-0.0966291726,0.1079116836,-0.0679994002,-0.0075088846,0.2013832331,-0.199802503,0.2344643027,0.3591980934,0.1373309791,0.129059419,-0.1139102578,-0.1763348281,0.1196943,-0.0580043867,-0.3185044229,-0.0318246596,0.3222025931,-0.000190768,0.0447842069,-0.0174140316,0.1147963405,0.1557329744,0.147576198,-0.1878322363,-0.2165790349,-0.1440662146,-0.1786869168,0.0539104007,0.3492898941,0.3345215321,-0.3116487861,-0.1201452538,0.0710493997,0.024994174,-0.3629961014,-0.3286014795,-0.0276242942,-0.0674707219,-0.110719882,0.4090248346,0.1963606477,-0.3866886199,0.2300868034,-0.1666325033,0.267883569,-0.2556222081,0.1114893332,0.4208318591,0.1170658842,-0.2309045345,-0.0539012924,-0.102771759,-0.0972770974,0.5784748197,-0.365550518,0.0661406144,-0.4556359351,0.3279430866,0.2335660458,-0.1663713306,0.3428981304,0.2844083607,-0.3349226415,0.2329865247,0.0152564663,0.5815436244,-0.0264838561,-0.0638352633,-0.2792891562,0.1324142665,-0.1662414074,-0.3447313905,-0.0006581515,-0.3484948874,-0.3311074674,0.2563079,-0.0774773061,-0.2266082764,0.5246610641,0.1639321595,-0.0409048982,-0.7086568475,-0.3736552596,-0.0045372285,0.2766523361,-0.2027623653,0.4182367325,0.3229885697,0.0039526508,0.3944139779,0.4252387285,0.3170160949,0.0883705392,-0.0516745709,0.3155106008,-0.4504492581,-0.0059706555,0.0368653499,0.3054410815,-0.1242536828,-0.1066462398,0.0966200233,-0.102817066,0.0515899286,0.0424061045,0.2062606364,0.0443429165,-0.5337705016,0.2679801881,-0.1764916331,0.1740657091,-0.0651561692,0.1449666619,-0.5101766586,0.2612885237,0.3920439184,0.0583924651,0.0105528086,-0.0828596056,0.0329217464,-0.1383634061,0.2497992963,0.2675231993,0.1021079421,-0.2600739002,-0.1450134516,-0.5910230875,0.0332322828,-0.1908808649,-0.0054462044,0.1442645788,-0.0413945653,-0.0358709693,0.2769308388,-0.0004990975,0.0354808979,-0.4129131138,0.2628389001,-0.2723473907,-0.2296372205,0.2392846048,0.2164620608,-0.1023936942,-0.1503431052,0.1333624274,-0.0975973085,-0.1930176318,-0.0528880693,0.1203178987,-0.2028674036,-0.0992030203,-0.025235869,0.2476500273,0.2405091673,0.0045695081,-0.1746524721,-0.1247410849,-0.2587486804,-0.3179563582,0.3191432953,0.2317633778,0.3002185524,-0.1924441308,-0.2965009809,0.0314552374,0.1041969284,0.282802254,0.0189734064,-0.4451439083,0.1156836525,-0.0879513696,0.3841794133,-0.1302316636,0.465686053,-0.0369808264,0.183060497,-0.1826772392,-0.568195045,0.4401472509,-0.6994359493,-0.1401675195,-0.0263723247,0.3481683433,0.0437882319,-0.2374210656,-0.9038203955,-0.1497897804,0.3655125797,0.0880114585,-0.1035750508,0.3861141205,-0.0456485115,-0.0585038513,-0.148834005,0.3935339749,0.4965204895,-0.1668690443,0.3587186635,-0.5926187634]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2323","title":"load_dataset(\"timit_asr\") gives back duplicates of just one sample text","comments":"Thanks @ekeleshian for having reported.\r\n\r\nI am closing this issue once that you updated `datasets`. Feel free to reopen it if the problem persists.","body":"## Describe the bug\r\nWhen you look up on key [\"train\"] and then ['text'], you get back a list with just one sentence duplicated 4620 times. Namely, the sentence \"Would such an act of refusal be useful?\". Similarly when you look up ['test'] and then ['text'], the list is one sentence repeated \"The bungalow was pleasantly situated near the shore.\" 1680 times. \r\n\r\nI tried to work around the issue by downgrading to datasets version 1.3.0, inspired by [this post](https:\/\/www.gitmemory.com\/issue\/huggingface\/datasets\/2052\/798904836) and removing the entire huggingface directory from ~\/.cache, but I still get the same issue. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(timit['train']['text'])\r\nprint(timit['test']['text'])\r\n```\r\n\r\n## Expected Result\r\nRows of diverse text, like how it is shown in the [wav2vec2.0 tutorial](https:\/\/colab.research.google.com\/github\/patrickvonplaten\/notebooks\/blob\/master\/Fine_tuning_Wav2Vec2_for_English_ASR.ipynb)\r\n\"Screen\r\n\r\n\r\n## Actual results\r\nRows of repeated text.\r\n\"Screen\r\n\r\n\r\n## Versions\r\n- Datasets: 1.3.0\r\n- Python: 3.9.1\r\n- Platform: macOS-11.2.1-x86_64-i386-64bit}\r\n","comment_length":24,"text":"load_dataset(\"timit_asr\") gives back duplicates of just one sample text\n## Describe the bug\r\nWhen you look up on key [\"train\"] and then ['text'], you get back a list with just one sentence duplicated 4620 times. Namely, the sentence \"Would such an act of refusal be useful?\". Similarly when you look up ['test'] and then ['text'], the list is one sentence repeated \"The bungalow was pleasantly situated near the shore.\" 1680 times. \r\n\r\nI tried to work around the issue by downgrading to datasets version 1.3.0, inspired by [this post](https:\/\/www.gitmemory.com\/issue\/huggingface\/datasets\/2052\/798904836) and removing the entire huggingface directory from ~\/.cache, but I still get the same issue. \r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(timit['train']['text'])\r\nprint(timit['test']['text'])\r\n```\r\n\r\n## Expected Result\r\nRows of diverse text, like how it is shown in the [wav2vec2.0 tutorial](https:\/\/colab.research.google.com\/github\/patrickvonplaten\/notebooks\/blob\/master\/Fine_tuning_Wav2Vec2_for_English_ASR.ipynb)\r\n\"Screen\r\n\r\n\r\n## Actual results\r\nRows of repeated text.\r\n\"Screen\r\n\r\n\r\n## Versions\r\n- Datasets: 1.3.0\r\n- Python: 3.9.1\r\n- Platform: macOS-11.2.1-x86_64-i386-64bit}\r\n\nThanks @ekeleshian for having reported.\r\n\r\nI am closing this issue once that you updated `datasets`. Feel free to reopen it if the problem persists.","embeddings":[0.2821879089,-0.4890296757,0.0636294037,0.3835485876,0.1550767869,0.0093223024,0.2025510371,0.1971771866,-0.1960430145,0.0225637201,-0.0537860543,0.2588163614,-0.0607607588,-0.0019165596,0.1897579134,0.0341104642,0.1301131546,-0.0039515593,-0.0855849162,-0.2164427489,-0.0349717848,0.5018290877,-0.1255064309,0.039624732,-0.3864379525,0.0896656886,-0.2256180942,0.0954400674,0.0274894368,-0.4388808608,0.2015984505,0.2562795579,0.0076610618,0.1743332148,-0.0001294062,-0.1238207743,0.103335917,0.1966973394,-0.2948189676,-0.337613523,0.0371453874,-0.0082015544,0.476550132,0.0588161834,-0.1446308494,-0.1229952723,0.0461377911,-0.0911030397,0.5953748822,0.0530859642,0.0554759242,-0.0213040039,-0.2731775641,-0.1502211094,0.30012393,0.1408754438,0.0361843631,0.0638974905,0.2574379146,0.0740013942,0.0937062427,0.4073674381,-0.2614638805,0.0803650841,0.0201924555,0.2505540252,-0.3521604538,-0.2842356563,0.1567363143,0.3818914294,0.4144081473,-0.1949874014,-0.2438047528,-0.5332146287,0.161943242,0.0539523214,0.1597354859,-0.0827361718,-0.239992395,0.132328406,-0.272169441,-0.0274029933,-0.0044496772,0.0038483818,0.1031661034,-0.2842705548,-0.1285128742,0.1618483514,0.3656548262,-0.0383118913,-0.0753812864,-0.0201201346,-0.0413320735,0.1814562082,-0.4986139834,0.0300714653,0.0270303711,-0.0612340719,0.2511675954,0.0161169358,0.4327557385,-0.0579363778,-0.1301839948,-0.0773007572,0.41262725,0.3775301576,0.0727243051,-0.0954676718,0.0092438171,-0.1186076403,-0.2098524123,0.0403310657,-0.0111010782,0.1610854268,0.0967437252,-0.2585274279,0.3425212801,-0.3690470159,-0.4510843158,0.0706897229,-0.5093212724,-0.1209715977,-0.1602241993,0.093509011,0.1373853534,0.3813998401,0.1867179424,0.1462110281,-0.1727949232,-0.2743343115,-0.134309411,-0.1454474628,-0.0307364892,0.1082837433,0.1738090068,-0.5300574899,0.2907099128,0.3179664016,0.341807574,-0.4182767868,-0.1233289689,-0.0930299759,-0.1239289194,0.0292858966,-0.0926786661,0.4591899216,0.0831208751,-0.0351972915,0.0112058921,-0.0125043103,-0.2134313434,-0.1196761578,0.2937436998,-0.0058470955,-0.2225892097,-0.0508112013,0.0654349998,0.3292100132,0.0748777837,0.1093474776,-0.1791653037,-0.0059522255,-0.4267409444,-0.1410837471,0.0414114632,0.3480868042,-0.2709081471,-0.0908948332,0.2106882185,0.2735222876,-0.0548251383,0.5467896461,-0.0427984148,0.4276206791,-0.2642585337,0.0796886086,0.0182303637,-0.4098272324,-0.2936970294,0.2606933713,-0.1333964318,0.3572215736,0.2087152153,0.1109746322,0.2958289087,0.1419264823,0.1747303605,0.0112375775,0.3328603804,-0.1269817948,-0.2913201153,-0.2165751457,0.0697632954,0.0462378487,-0.3213202953,0.213571474,0.0712145939,0.0611044355,0.4559496045,0.0401258767,-0.0278414022,0.0761946365,0.1083892062,0.0714350268,0.3267050087,-0.057660155,-0.2763386369,0.0627625287,0.1535568237,0.0899405926,0.1892700791,-0.0697184727,-0.2410974801,-0.0870902613,-0.3644341528,-0.1309642196,-0.0653520897,0.3186425269,0.0336826257,0.0968320817,-0.070033446,0.487472713,-0.1397870034,0.0868068263,-0.3534024656,0.0399518311,0.1423211247,0.078276664,0.2095784694,0.3197647631,0.2262900472,0.1240757406,0.0834152624,0.2634825408,0.029740002,0.1141453683,-0.1574550569,-0.212614432,0.1984330416,-0.2226008475,-0.0833782852,0.1785469055,0.2592630982,-0.1271862537,-0.2556278706,0.195391506,0.1241352111,0.2335562855,-0.0585896485,0.021532936,0.0859504044,-0.0684443489,-0.1082612723,-0.2807025909,0.5420552492,-0.1498948187,0.218012318,0.1538831294,-0.3125824928,0.2360616326,0.2239007503,-0.0717561767,-0.0559593439,0.3721970022,-0.1574176699,-0.0650917515,-0.1354452372,0.0964777842,0.3051019311,0.1097943187,0.0447403416,0.027225472,-0.1308660805,0.045597259,0.2455358952,-0.0921415836,-0.0883161649,0.3082364798,0.3380037546,-0.0945692807,-0.598446548,0.1292411536,-0.0442171581,0.2760109007,-0.3640033603,0.012894542,-0.3676787019,0.0385611914,-0.3294149637,-0.026490096,-0.2005130351,-0.1136622652,0.188616544,0.3473803699,0.2270390242,0.0175411087,0.2276711464,0.3561970294,-0.0489334911,0.2410355657,-0.08919736,0.1284423023,-0.3455183506,-0.0424879417,-0.0230716765,0.0373932831,0.2573551238,-0.1349247098,-0.3074090779,-0.4002928138,-0.2227882594,0.253131032,-0.2650930285,0.410592556,0.3220893443,-0.0436666459,-0.4077951908,-0.2074154466,0.0193402227,-0.0297960863,-0.401899308,0.2117161304,-0.0361040346,0.0585928671,-0.0396243893,-0.5401884317,0.0536182858,0.050110504,0.5322161913,0.0152634261,0.0834391713,0.2894309461,0.0147456396,0.0821889639,-0.188985303,0.1407686025,-0.5706542134,-0.2644888163,0.1145198941,0.155229792,-0.2780033648,-0.2397743165,0.0120847588,-0.0123424381,-0.1511860639,-0.5954256654,-0.262006253,-0.1077622399,0.0324840359,-0.1540211886,0.3147455752,0.22699669,0.1243949309,0.0203050952,-0.216123566,0.0191078782,0.5199985504,0.1948654354,0.1195437759,0.107226871,0.1179140136,-0.0223576836,0.3257386088,0.8683472276,0.1731048524,0.2302144766,-0.1726848483,0.1653012335,-0.2504996061,-0.3475958109,-0.2155773193,-0.2688345611,0.0127172098,0.431297034,0.1373870373,0.140871346,-0.1646451652,0.1525147706,-0.1869924366,-0.2055777758,-0.1615321636,-0.1367037445,0.11357636,0.1412164569,-0.028997859,0.2722263336,-0.2952245176,-0.1493304074,-0.0974014029,0.1342126131,-0.1416782141,-0.3199513853,-0.074440755,-0.6836324334,0.0644329041,0.2459241897,0.2961187065,-0.0062791235,-0.0122515298,0.4832975864,0.2160003483,1.0575666428,-0.0956710875,0.0187031496,-0.2254541218,-0.0384578519,-0.2621954978,0.0948655829,-0.2956033945,-0.1604373455,-0.0091202706,0.1500043571,-0.1343819499,0.0799587816,0.053867653,0.4325528741,-0.2853339016,-0.1934048533,-0.3532069623,-0.2623336911,-0.3100244105,-0.0245487355,-0.018167289,0.1022895724,-0.2388502806,-0.0308564622,0.0258367434,-0.1209612638,0.4737134278,0.1313102841,0.6074252725,-0.0171290226,0.4328970909,0.5953530073,-0.1411606222,-0.3634350896,0.8624765873,-0.1211226508,-0.2612604797,-0.1437160224,-0.3327286541,0.2774755657,0.6624307036,-0.2186656594,0.3310763538,-0.1124000996,-0.0247950163,-0.2145534605,-0.1021718904,0.3796004653,0.2069099247,0.0043728543,-0.3903522491,0.3599166572,-0.0825879648,-0.2580901682,0.1440051347,0.0782555342,-0.2731482983,0.325918138,-0.2066819668,1.1999918222,-0.1144596189,0.4627687335,0.0744244978,0.1892920434,0.4935740829,0.0038518589,-0.0439088121,-0.3514220417,-0.076208517,-0.0938368961,0.0494186021,-0.1900263727,0.0923126638,-0.3971046805,0.2927147448,0.3565191031,0.1551822871,0.0880177766,0.0852866322,-0.0418357998,-0.0442516059,-0.1079174951,-0.1045988426,-0.0658707991,0.1889068186,0.257589817,0.1798462123,0.2313103527,-0.5975795984,-0.476439774,0.0072849481,-0.3215946853,-0.0100113079,0.1055299491,-0.3966953158,0.0478975177,0.3964281082,0.1667467803,0.1279726624,-0.1238549054,-0.1648803502,0.2242880017,0.428678751,-0.1358835846,-0.2301557213,-0.0153529942,0.0764355883,-0.441925317,-0.0202909205,0.2607076168,-0.0377194807,-0.25432688,0.1783215702,-0.2497574389,-0.3596180975,0.0049403296,0.2692559063,0.1709797382,-0.3426191211,-0.0310657229,0.1471382678,-0.3102154732,0.088294372,0.0200827327,-0.3249722123,0.0588383414,0.7058208585,0.3741140962,-0.1189956516,0.5002724528,0.2675578594,0.0116620082,-0.0957498252,0.226036936,-0.0397436507,-0.4815365672,0.2151355743,-0.002908526,0.09179423,-0.2900525331,0.426771909,-0.021726232,0.1830295175,-0.1333239675,-0.2961260676,-0.0423954651,0.2187686116,0.13949655,0.1619938612,-0.1724784672,-0.1619065553,0.0097095631,0.1109667495,-0.1637848616,0.1157054976,0.0443555899,0.1690419018,0.0517503396,-0.0303461272,0.2929798961,0.0675780475,0.071134679,0.2659066021,0.0648534074,-0.09327811,-0.2981212139,0.2246005237,-0.0966291726,0.1079116836,-0.0679994002,-0.0075088846,0.2013832331,-0.199802503,0.2344643027,0.3591980934,0.1373309791,0.129059419,-0.1139102578,-0.1763348281,0.1196943,-0.0580043867,-0.3185044229,-0.0318246596,0.3222025931,-0.000190768,0.0447842069,-0.0174140316,0.1147963405,0.1557329744,0.147576198,-0.1878322363,-0.2165790349,-0.1440662146,-0.1786869168,0.0539104007,0.3492898941,0.3345215321,-0.3116487861,-0.1201452538,0.0710493997,0.024994174,-0.3629961014,-0.3286014795,-0.0276242942,-0.0674707219,-0.110719882,0.4090248346,0.1963606477,-0.3866886199,0.2300868034,-0.1666325033,0.267883569,-0.2556222081,0.1114893332,0.4208318591,0.1170658842,-0.2309045345,-0.0539012924,-0.102771759,-0.0972770974,0.5784748197,-0.365550518,0.0661406144,-0.4556359351,0.3279430866,0.2335660458,-0.1663713306,0.3428981304,0.2844083607,-0.3349226415,0.2329865247,0.0152564663,0.5815436244,-0.0264838561,-0.0638352633,-0.2792891562,0.1324142665,-0.1662414074,-0.3447313905,-0.0006581515,-0.3484948874,-0.3311074674,0.2563079,-0.0774773061,-0.2266082764,0.5246610641,0.1639321595,-0.0409048982,-0.7086568475,-0.3736552596,-0.0045372285,0.2766523361,-0.2027623653,0.4182367325,0.3229885697,0.0039526508,0.3944139779,0.4252387285,0.3170160949,0.0883705392,-0.0516745709,0.3155106008,-0.4504492581,-0.0059706555,0.0368653499,0.3054410815,-0.1242536828,-0.1066462398,0.0966200233,-0.102817066,0.0515899286,0.0424061045,0.2062606364,0.0443429165,-0.5337705016,0.2679801881,-0.1764916331,0.1740657091,-0.0651561692,0.1449666619,-0.5101766586,0.2612885237,0.3920439184,0.0583924651,0.0105528086,-0.0828596056,0.0329217464,-0.1383634061,0.2497992963,0.2675231993,0.1021079421,-0.2600739002,-0.1450134516,-0.5910230875,0.0332322828,-0.1908808649,-0.0054462044,0.1442645788,-0.0413945653,-0.0358709693,0.2769308388,-0.0004990975,0.0354808979,-0.4129131138,0.2628389001,-0.2723473907,-0.2296372205,0.2392846048,0.2164620608,-0.1023936942,-0.1503431052,0.1333624274,-0.0975973085,-0.1930176318,-0.0528880693,0.1203178987,-0.2028674036,-0.0992030203,-0.025235869,0.2476500273,0.2405091673,0.0045695081,-0.1746524721,-0.1247410849,-0.2587486804,-0.3179563582,0.3191432953,0.2317633778,0.3002185524,-0.1924441308,-0.2965009809,0.0314552374,0.1041969284,0.282802254,0.0189734064,-0.4451439083,0.1156836525,-0.0879513696,0.3841794133,-0.1302316636,0.465686053,-0.0369808264,0.183060497,-0.1826772392,-0.568195045,0.4401472509,-0.6994359493,-0.1401675195,-0.0263723247,0.3481683433,0.0437882319,-0.2374210656,-0.9038203955,-0.1497897804,0.3655125797,0.0880114585,-0.1035750508,0.3861141205,-0.0456485115,-0.0585038513,-0.148834005,0.3935339749,0.4965204895,-0.1668690443,0.3587186635,-0.5926187634]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2322","title":"Calls to map are not cached.","comments":"I tried upgrading to `datasets==1.6.2` and downgrading to `1.6.0`. Both versions produce the same output.\r\n\r\nDowngrading to `1.5.0` works and produces the following output for me:\r\n\r\n```bash\r\nDownloading: 9.20kB [00:00, 3.94MB\/s] \r\nDownloading: 5.99kB [00:00, 3.29MB\/s] \r\nNo config specified, defaulting to: sst\/default\r\nDownloading and preparing dataset sst\/default (download: 6.83 MiB, generated: 3.73 MiB, post-processed: Unknown size, total: 10.56 MiB) to \/home\/johannes\/.cache\/huggingface\/datasets\/sst\/default\/1.0.0\/a16a45566b63b2c3179e6c1d0f8edadde56e45570ee8cf99394fbb738491d34b...\r\n Dataset sst downloaded and prepared to \/home\/johannes\/.cache\/huggingface\/datasets\/sst\/default\/1.0.0\/a16a45566b63b2c3179e6c1d0f8edadde56e45570ee8cf99394fbb738491d34b. Subsequent calls will reuse this data.\r\nexecuted [0, 1]\r\n#0: 0%| | 0\/5 [00:00>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\nDownloading: 14.7kB [00:00, 4.91MB\/s]\r\nDownloading: 3.07MB [00:00, 32.6MB\/s]\r\nDownloading and preparing dataset oscar\/unshuffled_deduplicated_af (download: 62.93 MiB, generated: 163.38 MiB, post-processed: Unknown size, total: 226.32 MiB) to C:\\Users\\sgraaf\\.cache\\huggingface\\datasets\\oscar\\unshuffled_deduplicated_af\\1.0.0\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 81.0\/81.0 [00:00<00:00, 40.5kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 66.0M\/66.0M [00:18<00:00, 3.50MB\/s]\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\load.py\", line 745, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 979, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\tqdm\\std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"C:\\Users\\sgraaf\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\oscar\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464\\oscar.py\", line 359, in _generate_examples\r\n for line in f:\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\encodings\\cp1252.py\", line 23, in decode\r\n return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 7454: character maps to \r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n- Datasets: 1.6.2\r\n- Python: 3.9.4 (tags\/v3.9.4:1f2e308, Apr 6 2021, 13:40:21) [MSC v.1928 64 bit (AMD64)]\r\n- Platform: Windows-10-10.0.19041-SP0","comment_length":57,"text":"UnicodeDecodeError for OSCAR (Afrikaans)\n## Describe the bug\r\nWhen loading the [OSCAR dataset](https:\/\/huggingface.co\/datasets\/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\n```\r\n\r\n## Expected results\r\nAnything but an error, really.\r\n\r\n## Actual results\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\nDownloading: 14.7kB [00:00, 4.91MB\/s]\r\nDownloading: 3.07MB [00:00, 32.6MB\/s]\r\nDownloading and preparing dataset oscar\/unshuffled_deduplicated_af (download: 62.93 MiB, generated: 163.38 MiB, post-processed: Unknown size, total: 226.32 MiB) to C:\\Users\\sgraaf\\.cache\\huggingface\\datasets\\oscar\\unshuffled_deduplicated_af\\1.0.0\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 81.0\/81.0 [00:00<00:00, 40.5kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 66.0M\/66.0M [00:18<00:00, 3.50MB\/s]\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\load.py\", line 745, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 979, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\tqdm\\std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"C:\\Users\\sgraaf\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\oscar\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464\\oscar.py\", line 359, in _generate_examples\r\n for line in f:\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\encodings\\cp1252.py\", line 23, in decode\r\n return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 7454: character maps to \r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n- Datasets: 1.6.2\r\n- Python: 3.9.4 (tags\/v3.9.4:1f2e308, Apr 6 2021, 13:40:21) [MSC v.1928 64 bit (AMD64)]\r\n- Platform: Windows-10-10.0.19041-SP0\nThanks for reporting, @sgraaf.\r\n\r\nI am going to have a look at it. \r\n\r\nI guess the expected codec is \"UTF-8\". Normally, when no explicitly codec is passed, Python uses one which is platform-dependent. For Linux machines, the default codec is `utf_8`, which is OK. However for Windows machine, the default codec is `cp1252`, which causes the problem.","embeddings":[-0.0694424137,-0.0689756423,-0.0686594769,0.5462196469,0.4998692274,0.0235804226,0.1604425609,0.1600644588,-0.2532077134,0.2616168559,0.108140856,0.0745659769,-0.1329873204,-0.1621266156,0.0694556683,-0.2411268204,0.0238732174,-0.0095882108,0.1490591168,-0.1138725877,-0.3042188883,0.3047537208,-0.146704644,0.1465251744,-0.338201195,0.1170730814,0.0047265538,0.2575016916,-0.1414587945,-0.3523659706,0.1280923933,0.0163177848,0.1069401726,0.2742300928,-0.0001185049,-0.0547297038,0.4611512125,-0.1315254271,-0.3736711144,0.0470520742,-0.2316299528,0.1640679091,-0.1074656174,-0.0671967864,0.1775648743,-0.2891407609,-0.0651192144,-0.5603585839,0.018415628,0.322922349,0.1736242771,0.1185969636,0.1850381345,0.0575322583,-0.1859926581,-0.1070656851,-0.0142301479,0.0994632468,0.2410674989,0.1616299301,-0.0901438296,0.1463269442,0.1150426418,0.0891946778,0.1299639195,0.0778747648,-0.1804778725,-0.2484130412,0.3013782501,0.1579844356,0.3460531831,-0.4334869385,-0.386044234,-0.3108896613,0.1961866468,-0.5193542242,0.540989697,-0.0237485357,-0.1159027442,0.2356258333,-0.0887801498,0.078484565,-0.1790719777,0.2531147897,-0.2060515881,0.0477679521,-0.2439979166,-0.00667615,-0.1567632258,-0.109207958,0.0756576806,-0.3706617951,-0.0559559315,0.3437899053,-0.179793492,0.1441420764,-0.0641237795,0.1609281451,0.2896220088,0.2038017958,0.1995994896,-0.0058476198,-0.3779214919,-0.0436835438,0.1444817781,0.3555548787,-0.0572909042,-0.0332700424,0.2502485216,0.3896280825,0.0751844347,-0.0033730227,-0.1319143623,-0.2046168447,0.2316856533,0.0210273936,0.5934262872,-0.2065752447,-0.2736486793,0.2963079214,-0.1085711718,-0.1333846599,-0.1332821995,0.1001211107,-0.1807520241,0.208304435,-0.1411847919,0.263507992,-0.0920619071,-0.3318838179,-0.1641348898,-0.2244988978,-0.0947454721,0.104287982,0.0092695877,0.0026700087,0.1869610399,0.0432528481,0.3022830784,-0.0885123089,-0.0725461096,-0.1705320179,-0.0138035817,0.0101132309,0.0544214398,0.2440924793,0.0441389903,-0.4594571888,0.0363978371,0.0950224996,-0.0368120633,-0.2857238352,-0.190805763,0.1487388909,-0.1962005198,-0.05188182,-0.2405536324,0.2764452696,0.1800576001,0.0541132875,0.2551351488,-0.0168271754,-0.0834420025,-0.1390443891,0.2472237051,0.5666908622,-0.3448715806,-0.0487917103,-0.1240096763,-0.2066462934,0.3429841101,0.2263717651,0.0529710837,-0.0220640656,-0.3649465442,0.1118094996,0.1237515584,-0.246365875,-0.543892324,0.1850887686,-0.0684071407,0.5069641471,0.1817364693,-0.1587031037,0.2452930361,-0.0106923943,0.2900948226,0.4672313631,-0.2506344914,-0.0112695359,-0.253282845,-0.2451152056,0.0314227976,0.2176154852,0.0890725926,0.0002360298,0.1373472661,0.2070339918,0.3857545555,-0.104228802,0.1944541037,0.1247579083,0.0118005583,0.3576553762,0.0679639205,-0.2294073999,-0.282086283,0.2640694678,-0.0024920979,0.1133578196,-0.5008912683,-0.1440782249,-0.3841666877,-0.133248657,-0.4603827298,-0.407979995,0.0958776325,0.2784141898,0.1699373871,0.5170188546,-0.355160147,0.5020518303,-0.0425743796,-0.0007125136,-0.5245204568,0.2013834864,0.1213725284,-0.2475564778,-0.1367651522,0.2605232596,-0.0419438966,-0.0601239018,-0.3732912242,0.4005674124,-0.0718277916,0.2017915398,0.0296432581,-0.0306066349,0.3568613827,-0.2730566859,-0.0263988581,0.2786277831,0.2903375924,-0.0451177098,0.1648098528,0.2245022357,0.0779714659,0.1424601823,0.0697221607,-0.0256782267,0.2891086042,0.1915865093,-0.0097916322,-0.216820851,0.2216823101,-0.2710436583,0.5030974746,0.0332432501,-0.2335601598,-0.0266738199,0.6612030268,-0.1373115629,0.126795128,0.26213485,-0.1895713508,0.1265211254,-0.1143813953,0.1996651441,0.383133918,-0.0028643475,-0.0537522845,0.154264167,0.2330895215,-0.1416677535,0.1646430194,-0.0629854426,0.2722617984,0.1404127479,-0.1023301706,0.0954846963,-0.4391815364,-0.2598067224,-0.1511851847,0.2690092325,-0.3030655384,0.0640029982,-0.3796649575,-0.0206939261,-0.1904437095,0.0154795432,-0.3886276186,-0.1819051802,-0.1342800707,-0.0216699317,0.0310603883,0.1142271087,-0.3544104993,-0.0076761735,0.0488115102,-0.1199690923,-0.2093363106,-0.1015667543,-0.3103099763,0.0438975021,0.3231027424,-0.0926567391,0.2462631166,-0.2765359282,-0.0418532155,-0.1076047346,-0.221381247,0.0839273483,-0.0488724895,0.0420714654,0.0048807901,-0.1037534922,-0.1238439083,-0.219120428,0.2420269549,-0.0446107425,-0.2791256309,-0.0322687142,0.2220712453,0.0287394859,0.0866416246,-0.136137262,-0.2833299041,-0.2109611034,0.1948475689,-0.0700791627,0.0065472955,-0.004164679,-0.1305136085,0.206185624,0.0590830483,0.0376083516,-0.3174690902,-0.157781139,0.4085605145,-0.310995847,-0.3512374461,-0.1355366409,-0.0830545723,-0.0184940137,-0.0540398359,-0.2928423882,-0.0980516225,-0.510127306,0.0637507737,-0.1237606332,-0.0457393974,0.119070597,-0.157507062,0.0576680489,-0.1093934625,-0.1019479036,-0.099363856,-0.2305915356,0.2241044194,0.0622040257,0.4960558414,0.0121934908,0.21265845,0.3057525158,0.2123214155,0.526016891,-0.0799332038,0.1849859506,-0.128806293,-0.1884320527,0.2252397984,-0.0842235684,0.0127589097,0.2701328099,0.0234232247,0.0515855663,0.0052422374,-0.0492020361,-0.4067462981,-0.2655234635,0.0225820895,-0.2999838293,-0.0312157609,0.0302822553,-0.0431811661,-0.0429149941,-0.004856878,0.0555360615,0.2275538892,0.117038995,0.0395337269,-0.0138052711,0.463462323,-0.4846142828,0.1904788017,0.2164639831,0.368820399,-0.1864619851,0.1061565429,0.1600918472,-0.0152382031,0.5245283246,-0.2009750903,0.0450648032,-0.0242345091,0.0215711053,-0.1524719298,0.0056638322,-0.1911643595,0.0840869546,0.0751026943,-0.0573343486,-0.3224104047,0.1299594939,0.2902525961,0.1279769987,-0.2466364503,-0.0584840328,-0.2767928541,-0.6488556266,-0.4599291682,0.0440945327,0.1775838882,0.6439508796,0.0380225666,0.0368811823,-0.3710877895,-0.1193650439,0.2180927098,0.2218856663,0.2045163959,0.0522636361,0.1377334595,-0.0309110936,-0.1569290459,0.1311826259,0.7023574114,0.1565992832,-0.8256710172,0.0120185725,-0.3925683796,0.2404544502,0.303863287,-0.3360858262,-0.0839434117,-0.0898070112,0.045434434,-0.2287013829,-0.047865551,0.3065976501,0.1627944857,-0.3890811503,-0.3552235067,0.3005340397,0.1599294096,0.1147744283,0.3221263587,0.6737329364,-0.3928653002,0.3598769903,-0.1468215585,0.8503291011,-0.0660849437,0.083355166,0.2286052257,-0.1141342074,0.1319810301,0.0775085613,-0.0339698717,-0.2698893845,-0.5620769858,-0.0572103225,0.0915215015,0.3920048773,0.1631708443,0.0972269326,0.2071475536,0.0075961538,-0.0747890621,0.2738659978,-0.0034667561,-0.6421940923,-0.2249639481,-0.7355553508,0.0970179886,-0.0224779081,0.3151266575,-0.0714314878,-0.1402932703,-0.1581398696,-0.3351461887,-0.3856018484,-0.0154363299,-0.3776572943,-0.0970994532,-0.1545861065,-0.1117138788,0.1473340243,0.4229134917,0.6014969945,-0.0652299672,0.0985844508,0.1983693242,0.0026572002,0.0442011021,0.0897218809,-0.0528167076,0.2270966172,-0.1585269272,-0.2550147474,0.1195955351,0.1011184603,-0.198221311,0.0599999763,-0.009447475,-0.0429216847,0.0028320961,-0.2840370834,-0.3613981009,0.0897815526,-0.1192422062,0.0948574245,0.1246794686,-0.1993409842,-0.0894752145,-0.0859324187,-0.2890739739,-0.1181159914,0.3048832119,-0.0217162948,-0.2639842331,0.5028188229,0.1184295416,-0.0716893226,0.0056202356,0.1334239095,0.383113116,-0.7260894775,0.1097734272,-0.1273098141,0.1592686474,-0.4053401351,-0.0723259225,0.1595667601,-0.0180266481,0.0894450694,-0.4488123357,-0.4160528481,0.0744157881,0.2130593359,0.1293924749,0.0954491049,-0.192347616,-0.0250340644,-0.1422930509,-0.2089497745,0.1566259563,-0.0382911749,-0.0264956728,0.1362144202,0.0213510599,0.3478306234,-0.0249611195,0.1240819469,0.3377991021,0.1442821771,-0.1547534764,-0.2414449006,0.1742484123,0.0036252374,-0.2690984607,0.1239994541,0.0635249689,0.0573854223,-0.2646684647,0.3544514179,0.2616269886,0.0747516975,0.0032804376,0.4272879064,0.2713807225,-0.2036113441,0.1037187353,-0.097784467,0.652607739,0.1991455406,0.1095703617,0.0275043286,0.0897115842,0.1037354395,-0.1280574352,-0.153515175,0.2041809559,0.4038520753,-0.4951635897,0.1244053468,0.1553735286,0.1547781527,0.3628264368,-0.3015831709,-0.1846852601,0.0290584918,0.1114222705,-0.1472486705,-0.0250142701,0.1978164464,-0.050646849,-0.1373553872,0.3244960308,0.013393743,0.0463707894,-0.273154974,0.0127834706,0.15477781,0.1434918642,0.1568887532,0.5182115436,0.0003666666,0.1642681807,-0.0344317444,-0.0143855475,0.0959053114,0.6655558944,-0.4488252401,0.3969167471,-0.3500706851,0.1745541245,0.0767898262,-0.4070474803,0.0888987407,0.1055201143,-0.1830558032,0.3550574183,0.3304263651,-0.0402164049,0.0390585847,-0.024165839,-0.1799833775,-0.0039237412,0.0731732547,-0.0629029199,0.2421871424,-0.0489679277,-0.4993311465,0.0469325408,-0.2214452326,-0.0606461577,0.1470314562,0.1422177255,-0.2054488808,-0.4592171609,-0.0614412017,0.0940991193,0.3104107976,-0.1470007896,0.2697252929,0.5597763658,-0.050930541,0.083506979,0.3272659183,0.3617138267,0.3501305878,-0.0930441543,0.2183519602,-0.1705189943,0.0922725797,-0.041103676,0.3157975078,0.0335394889,0.2923209667,0.1796877682,0.1172872931,-0.0610099547,-0.113932617,0.1808080226,0.0667183772,-0.1626317203,0.5590126514,-0.2413318604,0.3378422856,-0.4312125742,0.1158481315,-0.4262698889,0.0569704808,0.1062830687,0.0436158553,0.152222082,-0.1087149158,0.0243047178,0.0584188998,0.3454767466,0.4733144045,-0.0609361231,-0.0635546148,-0.2280711979,-0.8209503889,0.2867543101,0.0702626333,-0.0507209338,0.0075879861,0.0676316097,-0.157578826,0.2200596482,0.1838781387,0.3849220276,0.0557090715,0.421577543,-0.2175575048,-0.3669783473,-0.0342484079,0.0892867371,-0.1560131758,-0.2319588661,0.1906741709,-0.1902059317,0.0089190593,-0.1362985522,0.0176735688,-0.3930063248,-0.3073526025,0.2711995244,0.4918371439,0.3157919347,-0.200627178,-0.1065057889,-0.3567888439,0.0511441678,-0.0569167994,0.7189384103,0.0713302493,0.3113724589,-0.0486296155,-0.3071857691,-0.3909928203,0.6503095627,0.0137485759,0.0019374921,-0.3354512155,0.208679229,0.006966962,0.160742268,0.1457736343,0.2864105701,-0.1862343848,0.3455361128,-0.0165793058,-0.5363335013,0.4404785633,-0.3810058534,-0.2944444418,0.0402828828,0.1934369504,0.38284114,-0.3138970733,-0.5248433948,0.3539697826,0.2991941273,-0.125530079,-0.1020757779,0.1371118724,-0.0490201786,0.033830002,-0.1445921808,0.3745250404,-0.0207860004,-0.0856956244,0.5031970739,-0.3108055294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2319","title":"UnicodeDecodeError for OSCAR (Afrikaans)","comments":"Awesome, thank you. \ud83d\ude03 ","body":"## Describe the bug\r\nWhen loading the [OSCAR dataset](https:\/\/huggingface.co\/datasets\/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\n```\r\n\r\n## Expected results\r\nAnything but an error, really.\r\n\r\n## Actual results\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\nDownloading: 14.7kB [00:00, 4.91MB\/s]\r\nDownloading: 3.07MB [00:00, 32.6MB\/s]\r\nDownloading and preparing dataset oscar\/unshuffled_deduplicated_af (download: 62.93 MiB, generated: 163.38 MiB, post-processed: Unknown size, total: 226.32 MiB) to C:\\Users\\sgraaf\\.cache\\huggingface\\datasets\\oscar\\unshuffled_deduplicated_af\\1.0.0\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 81.0\/81.0 [00:00<00:00, 40.5kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 66.0M\/66.0M [00:18<00:00, 3.50MB\/s]\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\load.py\", line 745, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 979, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\tqdm\\std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"C:\\Users\\sgraaf\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\oscar\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464\\oscar.py\", line 359, in _generate_examples\r\n for line in f:\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\encodings\\cp1252.py\", line 23, in decode\r\n return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 7454: character maps to \r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n- Datasets: 1.6.2\r\n- Python: 3.9.4 (tags\/v3.9.4:1f2e308, Apr 6 2021, 13:40:21) [MSC v.1928 64 bit (AMD64)]\r\n- Platform: Windows-10-10.0.19041-SP0","comment_length":4,"text":"UnicodeDecodeError for OSCAR (Afrikaans)\n## Describe the bug\r\nWhen loading the [OSCAR dataset](https:\/\/huggingface.co\/datasets\/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\n```\r\n\r\n## Expected results\r\nAnything but an error, really.\r\n\r\n## Actual results\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\nDownloading: 14.7kB [00:00, 4.91MB\/s]\r\nDownloading: 3.07MB [00:00, 32.6MB\/s]\r\nDownloading and preparing dataset oscar\/unshuffled_deduplicated_af (download: 62.93 MiB, generated: 163.38 MiB, post-processed: Unknown size, total: 226.32 MiB) to C:\\Users\\sgraaf\\.cache\\huggingface\\datasets\\oscar\\unshuffled_deduplicated_af\\1.0.0\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 81.0\/81.0 [00:00<00:00, 40.5kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 66.0M\/66.0M [00:18<00:00, 3.50MB\/s]\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\load.py\", line 745, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 979, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\tqdm\\std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"C:\\Users\\sgraaf\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\oscar\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464\\oscar.py\", line 359, in _generate_examples\r\n for line in f:\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\encodings\\cp1252.py\", line 23, in decode\r\n return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 7454: character maps to \r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n- Datasets: 1.6.2\r\n- Python: 3.9.4 (tags\/v3.9.4:1f2e308, Apr 6 2021, 13:40:21) [MSC v.1928 64 bit (AMD64)]\r\n- Platform: Windows-10-10.0.19041-SP0\nAwesome, thank you. \ud83d\ude03 ","embeddings":[-0.0694424137,-0.0689756423,-0.0686594769,0.5462196469,0.4998692274,0.0235804226,0.1604425609,0.1600644588,-0.2532077134,0.2616168559,0.108140856,0.0745659769,-0.1329873204,-0.1621266156,0.0694556683,-0.2411268204,0.0238732174,-0.0095882108,0.1490591168,-0.1138725877,-0.3042188883,0.3047537208,-0.146704644,0.1465251744,-0.338201195,0.1170730814,0.0047265538,0.2575016916,-0.1414587945,-0.3523659706,0.1280923933,0.0163177848,0.1069401726,0.2742300928,-0.0001185049,-0.0547297038,0.4611512125,-0.1315254271,-0.3736711144,0.0470520742,-0.2316299528,0.1640679091,-0.1074656174,-0.0671967864,0.1775648743,-0.2891407609,-0.0651192144,-0.5603585839,0.018415628,0.322922349,0.1736242771,0.1185969636,0.1850381345,0.0575322583,-0.1859926581,-0.1070656851,-0.0142301479,0.0994632468,0.2410674989,0.1616299301,-0.0901438296,0.1463269442,0.1150426418,0.0891946778,0.1299639195,0.0778747648,-0.1804778725,-0.2484130412,0.3013782501,0.1579844356,0.3460531831,-0.4334869385,-0.386044234,-0.3108896613,0.1961866468,-0.5193542242,0.540989697,-0.0237485357,-0.1159027442,0.2356258333,-0.0887801498,0.078484565,-0.1790719777,0.2531147897,-0.2060515881,0.0477679521,-0.2439979166,-0.00667615,-0.1567632258,-0.109207958,0.0756576806,-0.3706617951,-0.0559559315,0.3437899053,-0.179793492,0.1441420764,-0.0641237795,0.1609281451,0.2896220088,0.2038017958,0.1995994896,-0.0058476198,-0.3779214919,-0.0436835438,0.1444817781,0.3555548787,-0.0572909042,-0.0332700424,0.2502485216,0.3896280825,0.0751844347,-0.0033730227,-0.1319143623,-0.2046168447,0.2316856533,0.0210273936,0.5934262872,-0.2065752447,-0.2736486793,0.2963079214,-0.1085711718,-0.1333846599,-0.1332821995,0.1001211107,-0.1807520241,0.208304435,-0.1411847919,0.263507992,-0.0920619071,-0.3318838179,-0.1641348898,-0.2244988978,-0.0947454721,0.104287982,0.0092695877,0.0026700087,0.1869610399,0.0432528481,0.3022830784,-0.0885123089,-0.0725461096,-0.1705320179,-0.0138035817,0.0101132309,0.0544214398,0.2440924793,0.0441389903,-0.4594571888,0.0363978371,0.0950224996,-0.0368120633,-0.2857238352,-0.190805763,0.1487388909,-0.1962005198,-0.05188182,-0.2405536324,0.2764452696,0.1800576001,0.0541132875,0.2551351488,-0.0168271754,-0.0834420025,-0.1390443891,0.2472237051,0.5666908622,-0.3448715806,-0.0487917103,-0.1240096763,-0.2066462934,0.3429841101,0.2263717651,0.0529710837,-0.0220640656,-0.3649465442,0.1118094996,0.1237515584,-0.246365875,-0.543892324,0.1850887686,-0.0684071407,0.5069641471,0.1817364693,-0.1587031037,0.2452930361,-0.0106923943,0.2900948226,0.4672313631,-0.2506344914,-0.0112695359,-0.253282845,-0.2451152056,0.0314227976,0.2176154852,0.0890725926,0.0002360298,0.1373472661,0.2070339918,0.3857545555,-0.104228802,0.1944541037,0.1247579083,0.0118005583,0.3576553762,0.0679639205,-0.2294073999,-0.282086283,0.2640694678,-0.0024920979,0.1133578196,-0.5008912683,-0.1440782249,-0.3841666877,-0.133248657,-0.4603827298,-0.407979995,0.0958776325,0.2784141898,0.1699373871,0.5170188546,-0.355160147,0.5020518303,-0.0425743796,-0.0007125136,-0.5245204568,0.2013834864,0.1213725284,-0.2475564778,-0.1367651522,0.2605232596,-0.0419438966,-0.0601239018,-0.3732912242,0.4005674124,-0.0718277916,0.2017915398,0.0296432581,-0.0306066349,0.3568613827,-0.2730566859,-0.0263988581,0.2786277831,0.2903375924,-0.0451177098,0.1648098528,0.2245022357,0.0779714659,0.1424601823,0.0697221607,-0.0256782267,0.2891086042,0.1915865093,-0.0097916322,-0.216820851,0.2216823101,-0.2710436583,0.5030974746,0.0332432501,-0.2335601598,-0.0266738199,0.6612030268,-0.1373115629,0.126795128,0.26213485,-0.1895713508,0.1265211254,-0.1143813953,0.1996651441,0.383133918,-0.0028643475,-0.0537522845,0.154264167,0.2330895215,-0.1416677535,0.1646430194,-0.0629854426,0.2722617984,0.1404127479,-0.1023301706,0.0954846963,-0.4391815364,-0.2598067224,-0.1511851847,0.2690092325,-0.3030655384,0.0640029982,-0.3796649575,-0.0206939261,-0.1904437095,0.0154795432,-0.3886276186,-0.1819051802,-0.1342800707,-0.0216699317,0.0310603883,0.1142271087,-0.3544104993,-0.0076761735,0.0488115102,-0.1199690923,-0.2093363106,-0.1015667543,-0.3103099763,0.0438975021,0.3231027424,-0.0926567391,0.2462631166,-0.2765359282,-0.0418532155,-0.1076047346,-0.221381247,0.0839273483,-0.0488724895,0.0420714654,0.0048807901,-0.1037534922,-0.1238439083,-0.219120428,0.2420269549,-0.0446107425,-0.2791256309,-0.0322687142,0.2220712453,0.0287394859,0.0866416246,-0.136137262,-0.2833299041,-0.2109611034,0.1948475689,-0.0700791627,0.0065472955,-0.004164679,-0.1305136085,0.206185624,0.0590830483,0.0376083516,-0.3174690902,-0.157781139,0.4085605145,-0.310995847,-0.3512374461,-0.1355366409,-0.0830545723,-0.0184940137,-0.0540398359,-0.2928423882,-0.0980516225,-0.510127306,0.0637507737,-0.1237606332,-0.0457393974,0.119070597,-0.157507062,0.0576680489,-0.1093934625,-0.1019479036,-0.099363856,-0.2305915356,0.2241044194,0.0622040257,0.4960558414,0.0121934908,0.21265845,0.3057525158,0.2123214155,0.526016891,-0.0799332038,0.1849859506,-0.128806293,-0.1884320527,0.2252397984,-0.0842235684,0.0127589097,0.2701328099,0.0234232247,0.0515855663,0.0052422374,-0.0492020361,-0.4067462981,-0.2655234635,0.0225820895,-0.2999838293,-0.0312157609,0.0302822553,-0.0431811661,-0.0429149941,-0.004856878,0.0555360615,0.2275538892,0.117038995,0.0395337269,-0.0138052711,0.463462323,-0.4846142828,0.1904788017,0.2164639831,0.368820399,-0.1864619851,0.1061565429,0.1600918472,-0.0152382031,0.5245283246,-0.2009750903,0.0450648032,-0.0242345091,0.0215711053,-0.1524719298,0.0056638322,-0.1911643595,0.0840869546,0.0751026943,-0.0573343486,-0.3224104047,0.1299594939,0.2902525961,0.1279769987,-0.2466364503,-0.0584840328,-0.2767928541,-0.6488556266,-0.4599291682,0.0440945327,0.1775838882,0.6439508796,0.0380225666,0.0368811823,-0.3710877895,-0.1193650439,0.2180927098,0.2218856663,0.2045163959,0.0522636361,0.1377334595,-0.0309110936,-0.1569290459,0.1311826259,0.7023574114,0.1565992832,-0.8256710172,0.0120185725,-0.3925683796,0.2404544502,0.303863287,-0.3360858262,-0.0839434117,-0.0898070112,0.045434434,-0.2287013829,-0.047865551,0.3065976501,0.1627944857,-0.3890811503,-0.3552235067,0.3005340397,0.1599294096,0.1147744283,0.3221263587,0.6737329364,-0.3928653002,0.3598769903,-0.1468215585,0.8503291011,-0.0660849437,0.083355166,0.2286052257,-0.1141342074,0.1319810301,0.0775085613,-0.0339698717,-0.2698893845,-0.5620769858,-0.0572103225,0.0915215015,0.3920048773,0.1631708443,0.0972269326,0.2071475536,0.0075961538,-0.0747890621,0.2738659978,-0.0034667561,-0.6421940923,-0.2249639481,-0.7355553508,0.0970179886,-0.0224779081,0.3151266575,-0.0714314878,-0.1402932703,-0.1581398696,-0.3351461887,-0.3856018484,-0.0154363299,-0.3776572943,-0.0970994532,-0.1545861065,-0.1117138788,0.1473340243,0.4229134917,0.6014969945,-0.0652299672,0.0985844508,0.1983693242,0.0026572002,0.0442011021,0.0897218809,-0.0528167076,0.2270966172,-0.1585269272,-0.2550147474,0.1195955351,0.1011184603,-0.198221311,0.0599999763,-0.009447475,-0.0429216847,0.0028320961,-0.2840370834,-0.3613981009,0.0897815526,-0.1192422062,0.0948574245,0.1246794686,-0.1993409842,-0.0894752145,-0.0859324187,-0.2890739739,-0.1181159914,0.3048832119,-0.0217162948,-0.2639842331,0.5028188229,0.1184295416,-0.0716893226,0.0056202356,0.1334239095,0.383113116,-0.7260894775,0.1097734272,-0.1273098141,0.1592686474,-0.4053401351,-0.0723259225,0.1595667601,-0.0180266481,0.0894450694,-0.4488123357,-0.4160528481,0.0744157881,0.2130593359,0.1293924749,0.0954491049,-0.192347616,-0.0250340644,-0.1422930509,-0.2089497745,0.1566259563,-0.0382911749,-0.0264956728,0.1362144202,0.0213510599,0.3478306234,-0.0249611195,0.1240819469,0.3377991021,0.1442821771,-0.1547534764,-0.2414449006,0.1742484123,0.0036252374,-0.2690984607,0.1239994541,0.0635249689,0.0573854223,-0.2646684647,0.3544514179,0.2616269886,0.0747516975,0.0032804376,0.4272879064,0.2713807225,-0.2036113441,0.1037187353,-0.097784467,0.652607739,0.1991455406,0.1095703617,0.0275043286,0.0897115842,0.1037354395,-0.1280574352,-0.153515175,0.2041809559,0.4038520753,-0.4951635897,0.1244053468,0.1553735286,0.1547781527,0.3628264368,-0.3015831709,-0.1846852601,0.0290584918,0.1114222705,-0.1472486705,-0.0250142701,0.1978164464,-0.050646849,-0.1373553872,0.3244960308,0.013393743,0.0463707894,-0.273154974,0.0127834706,0.15477781,0.1434918642,0.1568887532,0.5182115436,0.0003666666,0.1642681807,-0.0344317444,-0.0143855475,0.0959053114,0.6655558944,-0.4488252401,0.3969167471,-0.3500706851,0.1745541245,0.0767898262,-0.4070474803,0.0888987407,0.1055201143,-0.1830558032,0.3550574183,0.3304263651,-0.0402164049,0.0390585847,-0.024165839,-0.1799833775,-0.0039237412,0.0731732547,-0.0629029199,0.2421871424,-0.0489679277,-0.4993311465,0.0469325408,-0.2214452326,-0.0606461577,0.1470314562,0.1422177255,-0.2054488808,-0.4592171609,-0.0614412017,0.0940991193,0.3104107976,-0.1470007896,0.2697252929,0.5597763658,-0.050930541,0.083506979,0.3272659183,0.3617138267,0.3501305878,-0.0930441543,0.2183519602,-0.1705189943,0.0922725797,-0.041103676,0.3157975078,0.0335394889,0.2923209667,0.1796877682,0.1172872931,-0.0610099547,-0.113932617,0.1808080226,0.0667183772,-0.1626317203,0.5590126514,-0.2413318604,0.3378422856,-0.4312125742,0.1158481315,-0.4262698889,0.0569704808,0.1062830687,0.0436158553,0.152222082,-0.1087149158,0.0243047178,0.0584188998,0.3454767466,0.4733144045,-0.0609361231,-0.0635546148,-0.2280711979,-0.8209503889,0.2867543101,0.0702626333,-0.0507209338,0.0075879861,0.0676316097,-0.157578826,0.2200596482,0.1838781387,0.3849220276,0.0557090715,0.421577543,-0.2175575048,-0.3669783473,-0.0342484079,0.0892867371,-0.1560131758,-0.2319588661,0.1906741709,-0.1902059317,0.0089190593,-0.1362985522,0.0176735688,-0.3930063248,-0.3073526025,0.2711995244,0.4918371439,0.3157919347,-0.200627178,-0.1065057889,-0.3567888439,0.0511441678,-0.0569167994,0.7189384103,0.0713302493,0.3113724589,-0.0486296155,-0.3071857691,-0.3909928203,0.6503095627,0.0137485759,0.0019374921,-0.3354512155,0.208679229,0.006966962,0.160742268,0.1457736343,0.2864105701,-0.1862343848,0.3455361128,-0.0165793058,-0.5363335013,0.4404785633,-0.3810058534,-0.2944444418,0.0402828828,0.1934369504,0.38284114,-0.3138970733,-0.5248433948,0.3539697826,0.2991941273,-0.125530079,-0.1020757779,0.1371118724,-0.0490201786,0.033830002,-0.1445921808,0.3745250404,-0.0207860004,-0.0856956244,0.5031970739,-0.3108055294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2319","title":"UnicodeDecodeError for OSCAR (Afrikaans)","comments":"@sgraaf, I have just merged the fix in the master branch.\r\n\r\nYou can either:\r\n- install `datasets` from source code\r\n- wait until we make the next release of `datasets`\r\n- set the `utf-8` codec as your default instead of `cp1252`. This can be done by activating the Python [UTF-8 mode](https:\/\/www.python.org\/dev\/peps\/pep-0540) either by passing the command-line option `-X utf8` or by setting the environment variable `PYTHONUTF8=1`.","body":"## Describe the bug\r\nWhen loading the [OSCAR dataset](https:\/\/huggingface.co\/datasets\/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\n```\r\n\r\n## Expected results\r\nAnything but an error, really.\r\n\r\n## Actual results\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\nDownloading: 14.7kB [00:00, 4.91MB\/s]\r\nDownloading: 3.07MB [00:00, 32.6MB\/s]\r\nDownloading and preparing dataset oscar\/unshuffled_deduplicated_af (download: 62.93 MiB, generated: 163.38 MiB, post-processed: Unknown size, total: 226.32 MiB) to C:\\Users\\sgraaf\\.cache\\huggingface\\datasets\\oscar\\unshuffled_deduplicated_af\\1.0.0\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 81.0\/81.0 [00:00<00:00, 40.5kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 66.0M\/66.0M [00:18<00:00, 3.50MB\/s]\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\load.py\", line 745, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 979, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\tqdm\\std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"C:\\Users\\sgraaf\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\oscar\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464\\oscar.py\", line 359, in _generate_examples\r\n for line in f:\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\encodings\\cp1252.py\", line 23, in decode\r\n return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 7454: character maps to \r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n- Datasets: 1.6.2\r\n- Python: 3.9.4 (tags\/v3.9.4:1f2e308, Apr 6 2021, 13:40:21) [MSC v.1928 64 bit (AMD64)]\r\n- Platform: Windows-10-10.0.19041-SP0","comment_length":66,"text":"UnicodeDecodeError for OSCAR (Afrikaans)\n## Describe the bug\r\nWhen loading the [OSCAR dataset](https:\/\/huggingface.co\/datasets\/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\n```\r\n\r\n## Expected results\r\nAnything but an error, really.\r\n\r\n## Actual results\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"oscar\", \"unshuffled_deduplicated_af\")\r\nDownloading: 14.7kB [00:00, 4.91MB\/s]\r\nDownloading: 3.07MB [00:00, 32.6MB\/s]\r\nDownloading and preparing dataset oscar\/unshuffled_deduplicated_af (download: 62.93 MiB, generated: 163.38 MiB, post-processed: Unknown size, total: 226.32 MiB) to C:\\Users\\sgraaf\\.cache\\huggingface\\datasets\\oscar\\unshuffled_deduplicated_af\\1.0.0\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 81.0\/81.0 [00:00<00:00, 40.5kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 66.0M\/66.0M [00:18<00:00, 3.50MB\/s]\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\load.py\", line 745, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 652, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\datasets\\builder.py\", line 979, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\site-packages\\tqdm\\std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"C:\\Users\\sgraaf\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\oscar\\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464\\oscar.py\", line 359, in _generate_examples\r\n for line in f:\r\n File \"C:\\Users\\sgraaf\\AppData\\Local\\Programs\\Python\\Python39\\lib\\encodings\\cp1252.py\", line 23, in decode\r\n return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 7454: character maps to \r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n- Datasets: 1.6.2\r\n- Python: 3.9.4 (tags\/v3.9.4:1f2e308, Apr 6 2021, 13:40:21) [MSC v.1928 64 bit (AMD64)]\r\n- Platform: Windows-10-10.0.19041-SP0\n@sgraaf, I have just merged the fix in the master branch.\r\n\r\nYou can either:\r\n- install `datasets` from source code\r\n- wait until we make the next release of `datasets`\r\n- set the `utf-8` codec as your default instead of `cp1252`. This can be done by activating the Python [UTF-8 mode](https:\/\/www.python.org\/dev\/peps\/pep-0540) either by passing the command-line option `-X utf8` or by setting the environment variable `PYTHONUTF8=1`.","embeddings":[-0.0694424137,-0.0689756423,-0.0686594769,0.5462196469,0.4998692274,0.0235804226,0.1604425609,0.1600644588,-0.2532077134,0.2616168559,0.108140856,0.0745659769,-0.1329873204,-0.1621266156,0.0694556683,-0.2411268204,0.0238732174,-0.0095882108,0.1490591168,-0.1138725877,-0.3042188883,0.3047537208,-0.146704644,0.1465251744,-0.338201195,0.1170730814,0.0047265538,0.2575016916,-0.1414587945,-0.3523659706,0.1280923933,0.0163177848,0.1069401726,0.2742300928,-0.0001185049,-0.0547297038,0.4611512125,-0.1315254271,-0.3736711144,0.0470520742,-0.2316299528,0.1640679091,-0.1074656174,-0.0671967864,0.1775648743,-0.2891407609,-0.0651192144,-0.5603585839,0.018415628,0.322922349,0.1736242771,0.1185969636,0.1850381345,0.0575322583,-0.1859926581,-0.1070656851,-0.0142301479,0.0994632468,0.2410674989,0.1616299301,-0.0901438296,0.1463269442,0.1150426418,0.0891946778,0.1299639195,0.0778747648,-0.1804778725,-0.2484130412,0.3013782501,0.1579844356,0.3460531831,-0.4334869385,-0.386044234,-0.3108896613,0.1961866468,-0.5193542242,0.540989697,-0.0237485357,-0.1159027442,0.2356258333,-0.0887801498,0.078484565,-0.1790719777,0.2531147897,-0.2060515881,0.0477679521,-0.2439979166,-0.00667615,-0.1567632258,-0.109207958,0.0756576806,-0.3706617951,-0.0559559315,0.3437899053,-0.179793492,0.1441420764,-0.0641237795,0.1609281451,0.2896220088,0.2038017958,0.1995994896,-0.0058476198,-0.3779214919,-0.0436835438,0.1444817781,0.3555548787,-0.0572909042,-0.0332700424,0.2502485216,0.3896280825,0.0751844347,-0.0033730227,-0.1319143623,-0.2046168447,0.2316856533,0.0210273936,0.5934262872,-0.2065752447,-0.2736486793,0.2963079214,-0.1085711718,-0.1333846599,-0.1332821995,0.1001211107,-0.1807520241,0.208304435,-0.1411847919,0.263507992,-0.0920619071,-0.3318838179,-0.1641348898,-0.2244988978,-0.0947454721,0.104287982,0.0092695877,0.0026700087,0.1869610399,0.0432528481,0.3022830784,-0.0885123089,-0.0725461096,-0.1705320179,-0.0138035817,0.0101132309,0.0544214398,0.2440924793,0.0441389903,-0.4594571888,0.0363978371,0.0950224996,-0.0368120633,-0.2857238352,-0.190805763,0.1487388909,-0.1962005198,-0.05188182,-0.2405536324,0.2764452696,0.1800576001,0.0541132875,0.2551351488,-0.0168271754,-0.0834420025,-0.1390443891,0.2472237051,0.5666908622,-0.3448715806,-0.0487917103,-0.1240096763,-0.2066462934,0.3429841101,0.2263717651,0.0529710837,-0.0220640656,-0.3649465442,0.1118094996,0.1237515584,-0.246365875,-0.543892324,0.1850887686,-0.0684071407,0.5069641471,0.1817364693,-0.1587031037,0.2452930361,-0.0106923943,0.2900948226,0.4672313631,-0.2506344914,-0.0112695359,-0.253282845,-0.2451152056,0.0314227976,0.2176154852,0.0890725926,0.0002360298,0.1373472661,0.2070339918,0.3857545555,-0.104228802,0.1944541037,0.1247579083,0.0118005583,0.3576553762,0.0679639205,-0.2294073999,-0.282086283,0.2640694678,-0.0024920979,0.1133578196,-0.5008912683,-0.1440782249,-0.3841666877,-0.133248657,-0.4603827298,-0.407979995,0.0958776325,0.2784141898,0.1699373871,0.5170188546,-0.355160147,0.5020518303,-0.0425743796,-0.0007125136,-0.5245204568,0.2013834864,0.1213725284,-0.2475564778,-0.1367651522,0.2605232596,-0.0419438966,-0.0601239018,-0.3732912242,0.4005674124,-0.0718277916,0.2017915398,0.0296432581,-0.0306066349,0.3568613827,-0.2730566859,-0.0263988581,0.2786277831,0.2903375924,-0.0451177098,0.1648098528,0.2245022357,0.0779714659,0.1424601823,0.0697221607,-0.0256782267,0.2891086042,0.1915865093,-0.0097916322,-0.216820851,0.2216823101,-0.2710436583,0.5030974746,0.0332432501,-0.2335601598,-0.0266738199,0.6612030268,-0.1373115629,0.126795128,0.26213485,-0.1895713508,0.1265211254,-0.1143813953,0.1996651441,0.383133918,-0.0028643475,-0.0537522845,0.154264167,0.2330895215,-0.1416677535,0.1646430194,-0.0629854426,0.2722617984,0.1404127479,-0.1023301706,0.0954846963,-0.4391815364,-0.2598067224,-0.1511851847,0.2690092325,-0.3030655384,0.0640029982,-0.3796649575,-0.0206939261,-0.1904437095,0.0154795432,-0.3886276186,-0.1819051802,-0.1342800707,-0.0216699317,0.0310603883,0.1142271087,-0.3544104993,-0.0076761735,0.0488115102,-0.1199690923,-0.2093363106,-0.1015667543,-0.3103099763,0.0438975021,0.3231027424,-0.0926567391,0.2462631166,-0.2765359282,-0.0418532155,-0.1076047346,-0.221381247,0.0839273483,-0.0488724895,0.0420714654,0.0048807901,-0.1037534922,-0.1238439083,-0.219120428,0.2420269549,-0.0446107425,-0.2791256309,-0.0322687142,0.2220712453,0.0287394859,0.0866416246,-0.136137262,-0.2833299041,-0.2109611034,0.1948475689,-0.0700791627,0.0065472955,-0.004164679,-0.1305136085,0.206185624,0.0590830483,0.0376083516,-0.3174690902,-0.157781139,0.4085605145,-0.310995847,-0.3512374461,-0.1355366409,-0.0830545723,-0.0184940137,-0.0540398359,-0.2928423882,-0.0980516225,-0.510127306,0.0637507737,-0.1237606332,-0.0457393974,0.119070597,-0.157507062,0.0576680489,-0.1093934625,-0.1019479036,-0.099363856,-0.2305915356,0.2241044194,0.0622040257,0.4960558414,0.0121934908,0.21265845,0.3057525158,0.2123214155,0.526016891,-0.0799332038,0.1849859506,-0.128806293,-0.1884320527,0.2252397984,-0.0842235684,0.0127589097,0.2701328099,0.0234232247,0.0515855663,0.0052422374,-0.0492020361,-0.4067462981,-0.2655234635,0.0225820895,-0.2999838293,-0.0312157609,0.0302822553,-0.0431811661,-0.0429149941,-0.004856878,0.0555360615,0.2275538892,0.117038995,0.0395337269,-0.0138052711,0.463462323,-0.4846142828,0.1904788017,0.2164639831,0.368820399,-0.1864619851,0.1061565429,0.1600918472,-0.0152382031,0.5245283246,-0.2009750903,0.0450648032,-0.0242345091,0.0215711053,-0.1524719298,0.0056638322,-0.1911643595,0.0840869546,0.0751026943,-0.0573343486,-0.3224104047,0.1299594939,0.2902525961,0.1279769987,-0.2466364503,-0.0584840328,-0.2767928541,-0.6488556266,-0.4599291682,0.0440945327,0.1775838882,0.6439508796,0.0380225666,0.0368811823,-0.3710877895,-0.1193650439,0.2180927098,0.2218856663,0.2045163959,0.0522636361,0.1377334595,-0.0309110936,-0.1569290459,0.1311826259,0.7023574114,0.1565992832,-0.8256710172,0.0120185725,-0.3925683796,0.2404544502,0.303863287,-0.3360858262,-0.0839434117,-0.0898070112,0.045434434,-0.2287013829,-0.047865551,0.3065976501,0.1627944857,-0.3890811503,-0.3552235067,0.3005340397,0.1599294096,0.1147744283,0.3221263587,0.6737329364,-0.3928653002,0.3598769903,-0.1468215585,0.8503291011,-0.0660849437,0.083355166,0.2286052257,-0.1141342074,0.1319810301,0.0775085613,-0.0339698717,-0.2698893845,-0.5620769858,-0.0572103225,0.0915215015,0.3920048773,0.1631708443,0.0972269326,0.2071475536,0.0075961538,-0.0747890621,0.2738659978,-0.0034667561,-0.6421940923,-0.2249639481,-0.7355553508,0.0970179886,-0.0224779081,0.3151266575,-0.0714314878,-0.1402932703,-0.1581398696,-0.3351461887,-0.3856018484,-0.0154363299,-0.3776572943,-0.0970994532,-0.1545861065,-0.1117138788,0.1473340243,0.4229134917,0.6014969945,-0.0652299672,0.0985844508,0.1983693242,0.0026572002,0.0442011021,0.0897218809,-0.0528167076,0.2270966172,-0.1585269272,-0.2550147474,0.1195955351,0.1011184603,-0.198221311,0.0599999763,-0.009447475,-0.0429216847,0.0028320961,-0.2840370834,-0.3613981009,0.0897815526,-0.1192422062,0.0948574245,0.1246794686,-0.1993409842,-0.0894752145,-0.0859324187,-0.2890739739,-0.1181159914,0.3048832119,-0.0217162948,-0.2639842331,0.5028188229,0.1184295416,-0.0716893226,0.0056202356,0.1334239095,0.383113116,-0.7260894775,0.1097734272,-0.1273098141,0.1592686474,-0.4053401351,-0.0723259225,0.1595667601,-0.0180266481,0.0894450694,-0.4488123357,-0.4160528481,0.0744157881,0.2130593359,0.1293924749,0.0954491049,-0.192347616,-0.0250340644,-0.1422930509,-0.2089497745,0.1566259563,-0.0382911749,-0.0264956728,0.1362144202,0.0213510599,0.3478306234,-0.0249611195,0.1240819469,0.3377991021,0.1442821771,-0.1547534764,-0.2414449006,0.1742484123,0.0036252374,-0.2690984607,0.1239994541,0.0635249689,0.0573854223,-0.2646684647,0.3544514179,0.2616269886,0.0747516975,0.0032804376,0.4272879064,0.2713807225,-0.2036113441,0.1037187353,-0.097784467,0.652607739,0.1991455406,0.1095703617,0.0275043286,0.0897115842,0.1037354395,-0.1280574352,-0.153515175,0.2041809559,0.4038520753,-0.4951635897,0.1244053468,0.1553735286,0.1547781527,0.3628264368,-0.3015831709,-0.1846852601,0.0290584918,0.1114222705,-0.1472486705,-0.0250142701,0.1978164464,-0.050646849,-0.1373553872,0.3244960308,0.013393743,0.0463707894,-0.273154974,0.0127834706,0.15477781,0.1434918642,0.1568887532,0.5182115436,0.0003666666,0.1642681807,-0.0344317444,-0.0143855475,0.0959053114,0.6655558944,-0.4488252401,0.3969167471,-0.3500706851,0.1745541245,0.0767898262,-0.4070474803,0.0888987407,0.1055201143,-0.1830558032,0.3550574183,0.3304263651,-0.0402164049,0.0390585847,-0.024165839,-0.1799833775,-0.0039237412,0.0731732547,-0.0629029199,0.2421871424,-0.0489679277,-0.4993311465,0.0469325408,-0.2214452326,-0.0606461577,0.1470314562,0.1422177255,-0.2054488808,-0.4592171609,-0.0614412017,0.0940991193,0.3104107976,-0.1470007896,0.2697252929,0.5597763658,-0.050930541,0.083506979,0.3272659183,0.3617138267,0.3501305878,-0.0930441543,0.2183519602,-0.1705189943,0.0922725797,-0.041103676,0.3157975078,0.0335394889,0.2923209667,0.1796877682,0.1172872931,-0.0610099547,-0.113932617,0.1808080226,0.0667183772,-0.1626317203,0.5590126514,-0.2413318604,0.3378422856,-0.4312125742,0.1158481315,-0.4262698889,0.0569704808,0.1062830687,0.0436158553,0.152222082,-0.1087149158,0.0243047178,0.0584188998,0.3454767466,0.4733144045,-0.0609361231,-0.0635546148,-0.2280711979,-0.8209503889,0.2867543101,0.0702626333,-0.0507209338,0.0075879861,0.0676316097,-0.157578826,0.2200596482,0.1838781387,0.3849220276,0.0557090715,0.421577543,-0.2175575048,-0.3669783473,-0.0342484079,0.0892867371,-0.1560131758,-0.2319588661,0.1906741709,-0.1902059317,0.0089190593,-0.1362985522,0.0176735688,-0.3930063248,-0.3073526025,0.2711995244,0.4918371439,0.3157919347,-0.200627178,-0.1065057889,-0.3567888439,0.0511441678,-0.0569167994,0.7189384103,0.0713302493,0.3113724589,-0.0486296155,-0.3071857691,-0.3909928203,0.6503095627,0.0137485759,0.0019374921,-0.3354512155,0.208679229,0.006966962,0.160742268,0.1457736343,0.2864105701,-0.1862343848,0.3455361128,-0.0165793058,-0.5363335013,0.4404785633,-0.3810058534,-0.2944444418,0.0402828828,0.1934369504,0.38284114,-0.3138970733,-0.5248433948,0.3539697826,0.2991941273,-0.125530079,-0.1020757779,0.1371118724,-0.0490201786,0.033830002,-0.1445921808,0.3745250404,-0.0207860004,-0.0856956244,0.5031970739,-0.3108055294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2318","title":"[api request] API to obtain \"dataset_module\" dynamic path?","comments":"Hi @richardliaw, \r\n\r\nFirst, thanks for the compliments.\r\n\r\nIn relation with your request, currently, the dynamic modules path is obtained this way:\r\n```python\r\nfrom datasets.load import init_dynamic_modules, MODULE_NAME_FOR_DYNAMIC_MODULES\r\n\r\ndynamic_modules_path = init_dynamic_modules(MODULE_NAME_FOR_DYNAMIC_MODULES)\r\n```\r\n\r\nLet me know if it is OK for you this way. \r\n\r\nI could set `MODULE_NAME_FOR_DYNAMIC_MODULES` as default value, so that you could instead obtain the path with:\r\n```\r\ndynamic_modules_path = datasets.load.init_dynamic_modules()\r\n```","body":"**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n","comment_length":63,"text":"[api request] API to obtain \"dataset_module\" dynamic path?\n**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n\nHi @richardliaw, \r\n\r\nFirst, thanks for the compliments.\r\n\r\nIn relation with your request, currently, the dynamic modules path is obtained this way:\r\n```python\r\nfrom datasets.load import init_dynamic_modules, MODULE_NAME_FOR_DYNAMIC_MODULES\r\n\r\ndynamic_modules_path = init_dynamic_modules(MODULE_NAME_FOR_DYNAMIC_MODULES)\r\n```\r\n\r\nLet me know if it is OK for you this way. \r\n\r\nI could set `MODULE_NAME_FOR_DYNAMIC_MODULES` as default value, so that you could instead obtain the path with:\r\n```\r\ndynamic_modules_path = datasets.load.init_dynamic_modules()\r\n```","embeddings":[-0.1038029417,-0.2795985937,-0.1310288161,0.0527429543,0.3254469931,-0.2634874582,-0.0613957755,0.1563359052,-0.2127547264,0.3393896222,-0.0749454126,0.7778300047,-0.4725230038,0.3166542649,0.2836638987,-0.4223529398,-0.1075101271,0.0503048524,-0.2626203299,0.0036111402,-0.1428721696,0.2316277176,-0.084783338,0.2267727405,0.0757528618,-0.2109445781,-0.0228330232,0.0175177902,-0.138037622,-0.3746137917,0.2695262134,0.1497001052,0.2535261214,0.2199819386,-0.0001094077,0.0686421096,0.0977332592,-0.0621917918,-0.385055393,-0.5734552741,0.1486319751,-0.1373895407,0.1928843558,-0.2899366617,-0.0638167486,-0.274437964,-0.0536290333,-0.7661019564,0.291287601,0.063306205,0.2035726011,0.3836931884,-0.2115834653,0.0141160162,0.0111685703,0.2845671475,-0.1351316273,-0.1789131463,0.1712046117,0.1474252492,-0.0069576534,0.2718309462,-0.0836598128,0.1184935868,0.5701004267,0.092134282,-0.4181138277,-0.0629471838,0.1164077893,0.3929319978,0.4341738224,-0.4605739713,-0.3385897577,-0.3974283934,-0.2387160212,-0.1298523992,0.1684211493,0.1541239619,-0.2948129475,0.1381622702,0.0280635785,-0.1709858924,-0.2388433367,-0.0365590118,0.1353721321,0.0721849278,0.0338096544,0.0492584147,0.2138145566,-0.2562360764,0.2530487478,0.0001356122,0.1630683988,0.3501803577,-0.1332852989,0.0437952392,0.2593631446,0.0081867296,0.0949191153,0.1528411061,0.0803343207,0.07105726,-0.1307509691,-0.0033430359,0.1743541211,-0.0688316077,0.1667058021,0.266246289,0.0422718786,0.2100567371,0.0943904743,-0.0139658703,-0.0658002421,0.1338001788,-0.0473947525,-0.0295378156,0.0953877941,0.1512423158,-0.1286311597,-0.2674999237,-0.3199114501,-0.2002335787,0.3884600103,0.5135529637,-0.0828726888,0.0863124654,0.2151854038,-0.2800034285,0.1129065976,0.0486184992,-0.1234717444,0.1900316179,-0.1880220175,0.1826998442,0.1336837262,-0.1558955312,0.3612341881,-0.2448184192,0.2501977682,-0.0375415795,0.1097567528,0.211521998,0.2910254896,0.0866850987,-0.0058065658,-0.0145474803,-0.2215718031,0.1193017662,-0.4832944274,0.0427645445,-0.2613914013,-0.4978818297,0.0284053702,0.1506852806,-0.1773408055,-0.1017165482,-0.265321821,0.1563615948,-0.352812171,0.0960061997,-0.0530509278,0.0474222191,-0.2700793743,-0.1650848687,0.273958385,0.3499602973,-0.0629201978,-0.2511388361,0.0323060714,0.0200772565,-0.1705696881,-0.0503633469,-0.480982095,0.217200622,-0.1995638609,0.111890547,0.4243351817,-0.3662731946,0.1924604923,-0.0459742136,-0.059218064,-0.2813163102,0.0999210626,-0.0642255843,0.0895033553,0.0366246998,0.1906103939,0.4679450095,0.2107554525,-0.0519933738,0.0623168536,-0.4383768439,-0.1321052164,0.0569139458,0.0535727851,-0.098087281,0.042050872,-0.0495754778,0.5005158186,-0.0934973136,-0.0479240566,0.2800538242,0.2826190293,0.2624945641,0.2190205753,-0.4658496976,-0.1698426157,0.2003215104,-0.1806832254,0.004847005,-0.2005132735,-0.1883307993,0.0107190842,0.2374990433,0.0430157296,0.2332790941,0.1127445623,-0.0729937777,-0.0062701162,-0.0852662623,-0.4607096016,0.0859270841,-0.2912900746,0.1009320691,-0.2687898874,0.2666110992,0.1665558666,0.0113884602,-0.0941922814,0.306388706,0.0804970264,-0.1595148742,0.066316694,0.1853282154,0.1383262128,0.1945545971,-0.0541788638,0.4117592573,0.1305964142,-0.0497927703,0.1247769743,-0.1700545996,0.0635828003,0.3343411386,0.2617124915,0.4313472509,-0.3116960824,0.3364180326,0.2746247947,0.1848143488,0.2339006513,-0.0022998799,-0.3090560138,-0.0668614656,0.105562605,0.1788025945,0.023215035,-0.2177972645,-0.2678339779,0.1311531663,0.2675953507,0.0370154195,-0.0143411215,-0.0825007781,-0.1222535223,0.0992908403,0.1431242675,0.4640809298,0.2016625553,0.372367382,-0.0311848782,0.103514336,-0.1654464155,0.0827460364,0.508587718,0.5266926289,0.0420886427,-0.3552269638,-0.0315018892,0.0765445754,-0.1907278746,-0.364616096,-0.1555599123,-0.058780808,-0.3679342866,0.2199699581,-0.4546753764,-0.0473184288,-0.6218673587,-0.5643889308,0.0178714413,0.1321030408,-0.060395278,-0.3999856114,-0.1962894052,0.1978435069,0.0772585496,0.2225201875,-0.2102829069,-0.5583546162,-0.0730556175,-0.1039896309,-0.1112963036,-0.0039003547,0.1009964943,0.1162631139,0.4589595199,-0.0660966411,0.1545366645,-0.2830373645,-0.2296126336,0.420738101,0.0460110791,0.073922433,0.3897780776,0.0098355543,0.2122607678,-0.1580740362,0.076707907,-0.0547339618,-0.1132117584,-0.3992922306,0.0876093879,0.0643574595,-0.0876141563,-0.3364678025,-0.2850174308,-0.2404596359,0.1316914856,0.0220055655,0.3412692249,0.3866489828,0.0774681643,0.1976983249,-0.3972137272,0.3130909801,-0.0743631944,-0.0252564643,0.1658196896,-0.0637148693,-0.2479181886,0.1867426187,-0.1397674084,0.3825788796,0.2226330936,-0.174667418,-0.55777812,0.0760991797,0.3783728182,-0.115717873,0.1571434289,0.4352731705,0.3465403318,-0.0645105317,-0.0938565657,-0.6340901256,0.3788377643,-0.0431331247,0.1673584878,0.2510476112,0.3467082083,0.1777441055,0.7528575063,-0.181751579,-0.4236459732,0.0311472118,-0.1504568309,0.1973709315,0.2868965268,-0.2159771472,-0.333032161,-0.238171652,-0.0834506452,0.0249342099,0.179941237,-0.0035557584,-0.2561033368,-0.0242580641,-0.0254070107,-0.2009700835,-0.061016541,0.1494755149,0.4692341387,0.2079645842,0.2638307512,-0.4698414505,0.0571561344,0.2276513875,0.1129219085,0.0671468452,0.0289807711,-0.0694196522,-0.0837766379,-0.030762516,0.2128607184,0.2331228852,0.0638222024,0.0785243213,-0.0474786386,0.357979238,0.0862004682,0.3167372048,0.4252171516,-0.1599026173,-0.1514633745,-0.2411319464,-0.0701697022,0.0445044339,-0.026758315,0.0390932932,0.1133594215,0.3360709846,-0.2208693027,-0.4445349276,0.0832481682,0.1801422387,-0.1273140162,0.0432912298,-0.1254707128,-0.6150271297,-0.1043439656,-0.1816945821,0.1410344988,0.2752096355,0.3697339296,-0.132362783,0.0107322196,-0.1800209135,0.1455291957,-0.0918153599,0.4292846024,0.0462661423,0.2616246343,0.3307581246,0.4209097028,-0.2178494334,0.2163706869,-0.3547985852,-0.0497568697,-0.1411769837,0.235957101,0.0529505536,0.4503337443,-0.0923175216,0.5791481137,0.3868659735,-0.0814831406,-0.3159941733,0.0603873506,0.1594022959,-0.0859970227,0.11713662,-0.8254059553,0.7735911608,-0.0561267138,-0.2654657066,0.4140822589,0.3090521395,-0.4338614345,0.063164711,0.5055969357,0.8474004865,0.0751848444,-0.075317733,0.4453955293,-0.467010051,0.2378028929,0.2643934786,0.1760012805,-0.2317067832,0.2607176006,-0.1877785325,-0.1911853105,0.0461387075,-0.2341165692,0.2179815471,0.1462669075,0.4515358806,0.1433019787,0.1041421518,0.2537241578,-0.0849172622,-0.3625350893,-0.4506554902,0.0646405369,-0.1980610341,0.0823966712,0.0796023533,-0.3622604012,0.1989402324,-0.2618861496,-0.0911210626,-0.3854855001,-0.6159552932,0.3089271486,0.4366497099,-0.2593919933,0.2312399149,0.0309975147,-0.0144267594,0.0217801407,-0.3404830992,0.3322176933,-0.1413490325,-0.2605072558,-0.1024760082,0.0596327744,0.2745764256,-0.0274847001,-0.0314291529,0.098097682,-0.1036355495,-0.1628209651,-0.1983742267,0.235234648,0.4340149462,0.0471857861,0.233590737,0.2397925258,0.4019862115,0.0107304454,0.1145023406,0.1893706024,-0.1899371445,0.2291092873,0.2565895915,0.0498802029,-0.0817737132,0.3258640766,-0.0360977538,-0.3304715455,0.4503824711,0.0344177857,-0.1876458526,-0.0733856559,0.0844617411,-0.1419008821,-0.3722993731,0.1286992729,0.133062005,0.0027370409,0.0227725711,0.2405536771,0.1204586029,-0.2028244585,-0.4959562123,0.0500296317,-0.0993115753,0.2616818845,-0.487293601,0.1782850474,0.046684172,0.1094487384,-0.1473510563,-0.2423119247,-0.3322330117,0.1697492748,0.0756281242,-0.2347113788,-0.0099389264,-0.0483514927,0.1035885513,0.0196105782,0.0206688736,-0.1674184054,-0.3417086005,-0.1054176837,-0.2636520565,0.1290740371,-0.2934762836,0.2898746133,-0.2584205568,-0.4429605007,-0.3115537465,-0.0917408764,-0.1657723188,0.2544405758,0.016520068,-0.290993005,-0.1005763784,-0.182067886,-0.1897600442,0.1860294342,0.0641118288,0.1085283831,0.3473661244,0.3174054921,-0.1919296533,-0.139494732,0.0671388209,0.2283728123,0.0205008294,-0.2499941289,0.2506266832,-0.01825854,-0.2987321019,0.1155869141,0.2533141971,0.0698419586,-0.1862195134,0.024839513,0.3518086076,0.1476494372,-0.2163857669,-0.1372430772,-0.1596531123,0.027729148,0.1589305401,0.4751015306,0.3294625282,0.044688452,0.2192899287,0.1590889096,0.4575735033,-0.084018141,-0.0202972535,0.2867109478,0.1116621569,0.0987731889,0.3216600716,0.459540844,0.4684455097,0.3703983426,-0.4155923724,0.6599217057,0.0969216749,-0.1093707755,0.0021117507,-0.243208468,0.0989543423,-0.0209329333,0.1255952865,-0.0371264927,-0.1475565732,0.601926744,-0.3544938862,0.1671015918,-0.2633250058,0.1947573572,0.0336240642,-0.1714271158,-0.0108015481,-0.2624391913,-0.3131428361,-0.1437803209,0.3506458998,-0.2271131128,0.0382734314,-0.2672570646,0.1363361925,-0.0091721788,-0.2429690063,0.2316109389,0.0077917194,-0.2223390788,-0.0314069502,0.0286649559,-0.1828943491,0.0107149109,-0.0349058583,0.041091118,0.1311830431,-0.1510450542,-0.070830442,0.0141315665,0.0545675196,0.0197320972,-0.0278931595,-0.1538022906,0.3528762758,0.3085981011,0.097533986,-0.1188476682,-0.0774694607,0.0332265049,0.3475908637,-0.308332026,0.6529901028,0.0617268011,0.1379269958,-0.1629935056,0.0113351978,-0.3320845962,0.065508537,0.2414770722,-0.1651806086,-0.0960701481,0.0734680742,0.0696211755,-0.0173026379,0.234587878,0.131628558,0.2944741547,0.0163609032,-0.0926444381,-0.4293479919,-0.2535050213,0.2615555227,-0.1690464467,0.1274643838,0.0533078127,0.2237377912,0.0018432395,0.150822401,-0.2255041897,0.1424327642,-0.0156660937,-0.2573007047,-0.1917879432,0.0597271845,0.0831735134,0.0479983613,-0.0525408685,-0.2635933459,-0.2478509098,-0.0997065753,-0.3285038173,0.1265524626,0.1037196517,-0.3290543258,0.2502278984,-0.1427303851,-0.0076587955,0.0810544118,0.1488341242,-0.2445700467,-0.0470976382,0.0523998514,-0.29871279,0.1247968674,0.1630404592,-0.121184893,-0.2223839164,-0.4487820864,0.0580842718,0.2074413896,0.0270482544,-0.1167502105,0.1439340115,-0.4063348174,0.028887311,0.1521342844,0.1841321886,0.0181778055,0.1298269033,-0.270092696,-0.2129979283,0.7706225514,-0.6101862788,-0.1864351034,-0.0291615427,0.338709116,0.1310017705,-0.0895153135,-0.6234183311,-0.0783440322,0.2426125258,-0.1379645914,0.002424388,0.1277170628,-0.1117190272,0.1759688109,-0.0884257331,-0.225172475,0.0719399229,-0.0747233629,-0.1060916409,-0.1518386602]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2318","title":"[api request] API to obtain \"dataset_module\" dynamic path?","comments":"Hi @albertvillanova, the default value proposal seems great :) Looking forward to this!","body":"**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n","comment_length":13,"text":"[api request] API to obtain \"dataset_module\" dynamic path?\n**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n\nHi @albertvillanova, the default value proposal seems great :) Looking forward to this!","embeddings":[-0.1192443147,-0.2359804809,-0.1255786419,0.0934086367,0.2970490158,-0.2991087437,-0.0950807631,0.1645036191,-0.2220382839,0.4028836787,-0.0430308506,0.780302763,-0.4451965392,0.3963929713,0.2424199879,-0.4457994699,-0.0987388194,0.0189938862,-0.282120645,-0.008350689,-0.107388638,0.2385216355,-0.0602002405,0.2233847529,-0.0048816293,-0.2270368636,-0.0022966983,-0.0629020706,-0.2275694907,-0.3151089549,0.299893111,0.0989250541,0.27499488,0.2342559397,-0.0001114414,0.016716212,0.1283120066,-0.0144618042,-0.3508321643,-0.6066398025,0.0599938668,-0.0309799928,0.2399614155,-0.2156696916,-0.0483233668,-0.2462839335,-0.0348588228,-0.7317524552,0.3012373745,0.0846928731,0.1744774282,0.2852261662,-0.2718065977,-0.0304426122,0.0195239596,0.3676588237,-0.1538048387,-0.1849719584,0.2404304147,0.1661438346,-0.0986476019,0.2649643421,-0.0067253844,0.0739354044,0.60059762,0.0952091068,-0.4916483164,-0.0388300642,0.1459251642,0.4695992172,0.512434721,-0.4513013363,-0.2661778629,-0.3193818331,-0.1964663118,-0.0979674309,0.2388626933,0.2052019835,-0.3035822511,0.1686713696,0.0072838077,-0.2028544843,-0.3409159482,-0.0251484532,0.1404250711,0.0453839861,0.003523801,0.0269564744,0.2472968996,-0.2482444048,0.2477356046,-0.0204502065,0.136771813,0.3299212456,-0.1496309787,-0.0424894877,0.248051554,0.0382701978,0.1105846986,0.1910864115,0.0922004804,0.1052888185,-0.2125988305,0.0079988996,0.1947668493,-0.0576602072,0.2017603368,0.129346922,0.0588142276,0.1640901864,0.1307228953,0.0140254861,-0.0003119953,0.2690988481,-0.1431022584,-0.076581791,-0.0264656469,0.0963847339,-0.0922370031,-0.2389263511,-0.2369943261,-0.2370826453,0.3945921361,0.5261980891,-0.0241173096,0.0368802026,0.1272231936,-0.2781774998,0.1634347439,-0.0139866741,-0.1240984127,0.1559949964,-0.1791002452,0.2213028669,0.1386582255,-0.1782446057,0.3648886085,-0.1870120913,0.2481601685,-0.0244265292,0.177593872,0.2132495195,0.3006585836,0.0462700389,0.0000731435,-0.053180825,-0.1626324952,0.1358424127,-0.4828062356,0.0833732635,-0.2933544815,-0.5377111435,-0.0534183308,0.1192512065,-0.1438257545,-0.124606967,-0.250341177,0.2495575994,-0.4220460355,0.0857928172,-0.0426141582,0.111116983,-0.2591056228,-0.1555458754,0.1704513133,0.2911574841,-0.0486130118,-0.262878865,0.0174783263,0.0074337316,-0.2108134627,-0.0182446446,-0.5295177698,0.1558253616,-0.2032384425,0.2204259485,0.4310370982,-0.338108629,0.1717428118,-0.0559447967,-0.0995883569,-0.404773891,0.1144824922,0.0048457342,0.0711646229,0.0238389559,0.1988725662,0.4034148753,0.195938915,-0.0504458435,0.0785586685,-0.5779812932,-0.174423635,0.0673464388,-0.0427297652,-0.1407471597,0.1062152833,-0.1014302522,0.4560680687,-0.1020417362,-0.0817037672,0.2598120868,0.3111793697,0.2131940275,0.1566463113,-0.4622723162,-0.190103665,0.1301058978,-0.2127268314,0.0068839923,-0.1220950261,-0.2320679426,0.0581379719,0.2307007909,0.1055637449,0.2477626652,0.0781188384,-0.0582163036,-0.0631932244,-0.0906854048,-0.5204542279,-0.0079999464,-0.3347383142,0.1228476316,-0.2543461621,0.2557995617,0.1842712909,0.0509452298,-0.0322412178,0.2582678795,0.0556427464,-0.143029958,0.0626136437,0.1636192054,0.1379079074,0.1994696558,0.033380162,0.497864604,0.1465552449,-0.0639306381,0.1600272357,-0.196131438,0.0156443045,0.2978382409,0.1401095837,0.3943191171,-0.3211243451,0.2639146149,0.2379556894,0.1564858705,0.2034573406,0.0328663662,-0.3246210515,-0.0730893016,0.1433756948,0.1279814839,0.0174280927,-0.1748274565,-0.2556751966,0.1400428414,0.2995645702,0.0511810072,0.0010056719,0.0085415887,-0.1273634434,0.0428741872,0.2158556879,0.4490804672,0.2274695784,0.3647868633,-0.0619657189,0.0651794001,-0.1389253139,0.0491272211,0.4699911475,0.4599807858,0.0529877059,-0.355704993,-0.0430448838,0.0966607407,-0.2734756768,-0.3000676334,-0.142752558,-0.0305837747,-0.2911926508,0.1451640278,-0.4359740019,-0.0662338361,-0.6336832047,-0.4837952256,-0.0102860611,0.1251722723,-0.0256414767,-0.4063815773,-0.224179104,0.1821217239,0.0568338893,0.3980732858,-0.2544726729,-0.5546968579,-0.1048081666,-0.0688798353,-0.1704572737,-0.0177037027,0.1318677962,0.1132078543,0.4859386086,0.008566604,0.2435871959,-0.3255787492,-0.3347913027,0.4399992824,0.1154871061,0.1039993167,0.382452637,0.0086310385,0.1521567553,-0.1645860225,0.0689032301,-0.0280201714,-0.1026293784,-0.415440619,0.1168051288,0.0830482021,-0.0539463982,-0.3830569386,-0.2808697224,-0.1914702952,0.1518782228,-0.0416553319,0.336820811,0.3303649724,0.0721026883,0.2437145263,-0.4835163951,0.2784747183,-0.0910281762,-0.0551644973,0.1220732033,-0.0317443945,-0.2670062184,0.1913972795,-0.1195077151,0.3806470037,0.1811300963,-0.2702911496,-0.5449126363,0.0861135498,0.4407222271,-0.0921331719,0.1026195362,0.5017553568,0.3343057632,-0.0363957845,-0.0711407289,-0.6541674137,0.3871555626,-0.0687244833,0.1942713261,0.2685588598,0.3242745697,0.1890627742,0.7807953954,-0.1177470013,-0.4486914277,0.0651046634,-0.1582908481,0.2228229046,0.2523452044,-0.2232629508,-0.2594988346,-0.22644642,-0.0959166437,0.0819191709,0.2119965851,-0.0055796667,-0.2571833432,-0.0411472805,0.009764934,-0.1871319562,-0.0182562638,0.1872059703,0.4395620227,0.2162239403,0.263664633,-0.4420650005,0.04292145,0.2321644574,0.1242113858,0.10933806,0.0239284169,-0.1404200941,-0.1070540771,-0.0203626249,0.2224467844,0.2137450278,0.0539741479,0.0242538396,0.0018948637,0.3642623425,0.0875115991,0.3523096442,0.4002794325,-0.2053508162,-0.2647417188,-0.245720312,-0.0033112234,0.1270685345,-0.0003209422,0.0706205219,0.1109461784,0.4252747893,-0.2035249323,-0.4369705915,0.0955582559,0.0882873237,-0.0702227354,0.0263909027,-0.0353248566,-0.5628284812,-0.0512137897,-0.1706589609,0.1051520929,0.2384044379,0.3485447168,-0.1460548937,0.0358627662,-0.1671164632,0.1093012691,-0.1892296672,0.3966543078,0.0386239588,0.2394258529,0.2920612693,0.4887080193,-0.208644554,0.1512367576,-0.3634923398,-0.0044928198,-0.1049746946,0.1779812574,0.0833556652,0.519090414,-0.0502860323,0.5679901242,0.4834351242,-0.0706375167,-0.2944825292,0.0423331521,0.1434528381,-0.1117569134,0.087810643,-0.7443140745,0.8651666045,-0.0276839919,-0.2762111425,0.4507704079,0.3465531468,-0.3800905347,0.0247731358,0.5472705364,0.9188605547,0.1607439518,-0.0009737652,0.4028138816,-0.5161280632,0.3020170629,0.2022553384,0.177196756,-0.2765772343,0.2032737136,-0.2360307574,-0.2281658053,0.0443609916,-0.2173782885,0.2090111971,0.0473402739,0.4860451221,0.2853049934,0.1423449218,0.2782467306,-0.1923161894,-0.346087873,-0.3599821627,0.0507354848,-0.1956144422,-0.0035406235,0.0708899945,-0.3982607722,0.1966228336,-0.2598725259,-0.1333376318,-0.4370934963,-0.6310657859,0.2642351985,0.4431126118,-0.2624378204,0.2428935766,0.0520295277,-0.0202704873,0.0282778516,-0.4402227998,0.3375801444,-0.2047325224,-0.2143470943,-0.0446908027,0.0264847316,0.3103302121,-0.0752327293,-0.0237552673,0.1051776186,-0.0951531529,-0.2016964704,-0.2975940406,0.2202626765,0.3985711038,0.0069660321,0.175359562,0.2288755924,0.376178205,0.0289408416,0.0967334509,0.1783102006,-0.1377399862,0.2392453551,0.3259305954,0.0540608764,-0.1247813478,0.373854816,-0.0421436541,-0.4010379612,0.3864420652,0.0092615178,-0.168719396,-0.0785808265,0.1343327612,-0.0895690918,-0.394572407,0.0792535618,0.0495532528,-0.0794320479,0.0329679102,0.2272607386,0.1546521634,-0.1906757653,-0.4660451114,0.0630464032,-0.1008894965,0.3001457453,-0.5050112605,0.2017272264,0.0505319387,-0.0084174583,-0.1568949819,-0.2422372699,-0.3123594522,0.1727910042,0.1342415512,-0.2394591421,0.0160574727,0.0486208238,0.0876503512,0.0290379953,-0.0329189338,-0.1045398861,-0.3793998957,-0.0843138695,-0.255728513,0.1422023326,-0.2615115047,0.2950270176,-0.2652973235,-0.4497604072,-0.3879786432,-0.0635347962,-0.2148142755,0.1714127362,0.0065171891,-0.3124706745,-0.120706521,-0.2051804215,-0.1293306053,0.2085736692,0.1449133456,0.1511185914,0.3459276259,0.2651391327,-0.2113369107,-0.1476480514,0.1080112383,0.2639444172,0.008056134,-0.2359960824,0.2195504606,-0.037921939,-0.3153920472,0.1045504287,0.2803036273,-0.0087559791,-0.2047688514,0.0502917767,0.398728013,0.1259504557,-0.1600995809,-0.1655356586,-0.1473776996,0.0875243992,0.090873003,0.4086368382,0.3111740947,0.0944907069,0.2572927475,0.1505096704,0.4488633573,-0.1161592007,0.0361034684,0.3004515767,0.0291332677,0.0794080049,0.3317962289,0.4673808813,0.469466567,0.4072720706,-0.3842874467,0.7063808441,0.1151896641,-0.1229231507,-0.0108707985,-0.2341440022,0.0936705545,0.0247204676,0.130295217,-0.0472965129,-0.1143210903,0.6202298999,-0.3160804808,0.0859462544,-0.3119387031,0.18803747,-0.0124346055,-0.1176342815,-0.0798678771,-0.3124755025,-0.2268970013,-0.0655113831,0.3174627423,-0.1448322535,0.0434429459,-0.2163871527,0.1098164767,-0.0309147947,-0.193275094,0.274897486,-0.0149936024,-0.1418790519,0.0439141989,-0.0087193046,-0.1567621231,0.0679851696,-0.0221836008,0.0009261243,0.1633030623,-0.0859925523,-0.0719492137,-0.0146590183,0.0329703577,0.0419292264,-0.0196571257,-0.0431335904,0.3101606965,0.3439054191,0.0739045516,-0.0692464709,-0.0107555846,0.0305215437,0.4017837048,-0.3936827779,0.5672363043,0.0680576861,0.0877921581,-0.0858224556,-0.03064874,-0.2883415818,0.0979951993,0.18905209,-0.1677280217,-0.1107391864,0.0279490445,0.0606351756,-0.0153569328,0.2750056684,0.1182803139,0.2239967138,0.0070110685,-0.1456069499,-0.4248638749,-0.2719306946,0.3512255847,-0.1223770007,0.1000487953,-0.0397796258,0.2192400098,0.0166768916,0.1739378422,-0.1994983852,0.2642294765,-0.0268058572,-0.2759322524,-0.1839023679,0.1292188913,0.0799989253,0.0225003604,-0.0367465131,-0.2201511711,-0.0977721587,-0.1348531097,-0.3652352989,0.1127883792,0.0222769473,-0.3438945413,0.2925704718,-0.0822254196,-0.0468608774,0.0917457491,0.2224715203,-0.2173406929,-0.0008136557,0.0359518491,-0.3316890895,0.1213293299,0.1889057457,-0.1526789814,-0.2419308424,-0.4498143494,0.0343057215,0.2056037635,0.0696602836,-0.1652209759,0.1396449506,-0.4324088097,-0.0229439382,0.1883947998,0.2015006542,0.0419347286,0.1393878907,-0.2385141402,-0.133436054,0.7548966408,-0.6378833055,-0.146620065,-0.0172348507,0.2961886227,0.1702851206,-0.1387294531,-0.6024070382,-0.1083646268,0.2030390203,-0.1396337152,0.019982582,0.1760910451,-0.1361820102,0.1917113364,-0.1044927686,-0.1401063949,0.1311807781,-0.1035191119,-0.0717357546,-0.155414179]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2318","title":"[api request] API to obtain \"dataset_module\" dynamic path?","comments":"I like the idea as well ! thanks @albertvillanova ","body":"**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n","comment_length":9,"text":"[api request] API to obtain \"dataset_module\" dynamic path?\n**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n\nI like the idea as well ! thanks @albertvillanova ","embeddings":[-0.0715709329,-0.2262378633,-0.1513926685,0.1596531868,0.2925485969,-0.2739342451,-0.1207628474,0.1491294205,-0.2353176922,0.3420878351,-0.0302424543,0.7563670874,-0.4744618535,0.3760714233,0.281167537,-0.4820159376,-0.109488517,-0.0033817105,-0.2524858713,0.0120169166,-0.0957384109,0.2206683308,-0.0734882504,0.2258647978,0.0079630967,-0.2546795905,-0.0356744342,-0.0634931549,-0.2515848577,-0.2921498716,0.3152516484,0.0871357098,0.2588400543,0.2571384311,-0.0001125954,0.004969595,0.090837799,-0.0150505695,-0.3594183624,-0.5971360803,0.0594909005,-0.0925777853,0.2382089049,-0.2151919752,-0.0022104611,-0.2780035436,-0.040508531,-0.7394971251,0.3323914111,0.0607341304,0.1650945693,0.2808467746,-0.2516329885,0.0082772039,0.025817506,0.3935748935,-0.1761958301,-0.2006755322,0.295698911,0.1604554504,-0.0758963823,0.2667439282,-0.0367462747,0.0615741126,0.5828230381,0.0497203507,-0.5414122939,-0.0385879017,0.1276909858,0.4285574853,0.5348728299,-0.4504088163,-0.296035707,-0.3133638501,-0.2100010216,-0.0663065314,0.19749327,0.199285835,-0.3210361302,0.1537103951,-0.0154080065,-0.2557679713,-0.3433293402,-0.0517514497,0.201032415,0.0322980471,0.0301440116,0.0186986886,0.2041103989,-0.2355963141,0.2858401239,0.0049677151,0.14617154,0.3492870927,-0.1562351435,-0.04607464,0.2432495058,0.0232164301,0.1283154339,0.2065716237,0.1001705006,0.0845767409,-0.2042628825,-0.0122768916,0.2219229192,-0.0685281828,0.1889935583,0.1752980649,0.0394567698,0.1589556336,0.0910396725,0.0103948405,-0.0404775962,0.2850301266,-0.1476002485,-0.0442427546,-0.0412372835,0.0836659372,-0.0764949694,-0.2452889383,-0.2848002017,-0.2372208685,0.3892763257,0.4723510444,-0.0049760225,0.0151317827,0.1578300446,-0.2983418107,0.1827229112,0.0231753923,-0.1068718657,0.1691015214,-0.1931519657,0.250725776,0.1211967319,-0.1388201714,0.3422932327,-0.1839130372,0.2232414335,-0.0525148772,0.1567001045,0.2341081351,0.28389588,0.0240607504,-0.0002942037,-0.0533986278,-0.1752424836,0.1670508683,-0.473095715,0.0734137222,-0.2884525061,-0.5198731422,-0.0563308075,0.0934075639,-0.145841375,-0.1512646228,-0.2438306212,0.2197324336,-0.4169390202,0.0860321596,-0.0213152077,0.1332706213,-0.2540419102,-0.1530595124,0.1281196624,0.3148926198,-0.0525904894,-0.2368785888,0.016068114,0.0488406643,-0.1903237849,-0.0038042984,-0.4953263998,0.1567065865,-0.2235729247,0.1983281374,0.4840585887,-0.3624569476,0.1698249876,-0.0316582397,-0.0957047716,-0.3844743967,0.1449628472,0.0049770819,0.0358829498,0.0469794609,0.1755563319,0.4307869971,0.1895030439,-0.090360105,0.0964076594,-0.5630605817,-0.1552044451,0.0537548624,-0.0316385776,-0.1246224642,0.1365335137,-0.0989810601,0.4867522418,-0.1022997424,-0.06934084,0.2462878227,0.2892418504,0.2683608532,0.1794160455,-0.4992589951,-0.1602211297,0.1219828352,-0.2344980091,0.0176467672,-0.1518499553,-0.2182080299,0.0823772773,0.2206687033,0.11377386,0.230423823,0.0639063939,-0.1100528538,-0.0433410704,-0.1096116453,-0.4932194352,-0.0248245467,-0.3215841055,0.1224106625,-0.2648107111,0.2604577243,0.2115884125,0.0703785717,-0.0352031961,0.2778664529,0.0526896492,-0.1519785672,0.0593977645,0.0744207874,0.1469548941,0.184335053,0.0751591176,0.5323821306,0.1353834569,-0.0667533502,0.1945181787,-0.2259695679,0.0209826846,0.2765145004,0.2229263633,0.3700436652,-0.3378068209,0.3178957403,0.259296298,0.1926462948,0.2136887163,0.0461604781,-0.2812060416,-0.0396230444,0.1722409576,0.143887639,0.0221076552,-0.2271492779,-0.1987493336,0.1524049938,0.3175561428,0.0785224661,0.010522265,-0.022995336,-0.1219253764,0.059095446,0.1909933686,0.4412010908,0.1915376931,0.357542485,-0.0920841247,0.0759723186,-0.072846204,0.0640159249,0.4859339893,0.5009200573,0.0476383716,-0.3722443283,-0.070635587,0.1177125946,-0.2420990318,-0.3559581935,-0.1279858947,-0.03981499,-0.2770503461,0.1532433331,-0.4592067301,-0.1149933487,-0.6361504793,-0.5350412726,-0.0068779159,0.1521460414,-0.0415756144,-0.4301598072,-0.1730868071,0.1507072151,0.0573836118,0.3607397676,-0.2596814334,-0.5457544923,-0.1075374633,-0.0874734968,-0.1345088333,-0.0198201407,0.1802768111,0.1357111633,0.4953314662,0.01176864,0.266693294,-0.3546221554,-0.3064699769,0.4399485886,0.1013118252,0.0736469701,0.3733629882,-0.0168200489,0.1552999318,-0.1434012651,0.0467413813,-0.0134852827,-0.1099841073,-0.4161855876,0.1019306406,0.073048681,-0.0248958487,-0.3331699371,-0.2991986573,-0.1828419715,0.1880793124,-0.0560364611,0.3576302528,0.3447981179,0.0479699038,0.2368636876,-0.4663241804,0.264336586,-0.0744856074,-0.0528311022,0.0977041349,-0.030867368,-0.2535563707,0.2071957439,-0.0878424719,0.3844105303,0.2020772099,-0.2797503769,-0.5342267156,0.1015204713,0.4397081733,-0.0901888758,0.0797019303,0.5040370226,0.3253481388,-0.0120288674,-0.0503931977,-0.6377399564,0.3959643841,-0.0680103898,0.1986982375,0.2658723593,0.3279469311,0.1755133718,0.7688278556,-0.1155988723,-0.4027715325,0.0604580119,-0.1440828145,0.1693125367,0.2883001566,-0.2650906742,-0.2647750676,-0.2616938055,-0.129887566,0.0604530387,0.2356179208,0.0088570286,-0.2563450336,-0.0762848258,0.0109624462,-0.1535546184,-0.0412260033,0.1672307253,0.4519492984,0.218801707,0.2552594841,-0.4393710196,0.059957426,0.252084434,0.1477391124,0.1498459876,0.0316446982,-0.1080778241,-0.155882284,0.0044406629,0.2117329985,0.2444834709,0.0695731565,-0.003676354,0.0567412935,0.3774356544,0.1157394722,0.3304950595,0.3889866769,-0.2172111422,-0.2562645972,-0.2289268672,0.027646374,0.1249274239,0.0113684228,0.0763891712,0.0902891979,0.4204506278,-0.1666353643,-0.4254294038,0.095123522,0.1066044867,-0.0953660309,0.0142263761,-0.0270686727,-0.5661179423,-0.0458875597,-0.1835055351,0.0992510021,0.2047625482,0.3708977401,-0.1818471849,0.0463498794,-0.1409290433,0.1172326058,-0.224024266,0.3593381345,0.0168222617,0.2393189371,0.2853223383,0.4651551247,-0.212771371,0.1727165729,-0.3444378674,-0.0219970904,-0.1229914501,0.1917362958,0.1011572778,0.4704846442,-0.0259352289,0.6190360188,0.4835693836,-0.1036292017,-0.2900442779,-0.0348135419,0.1614524871,-0.0809980333,0.0776027888,-0.7981911898,0.8521542549,-0.0080210781,-0.2405977845,0.4786370993,0.3154906631,-0.3955683708,0.026053926,0.5296068788,0.9231008887,0.1477058083,-0.056239713,0.3918225467,-0.5130006075,0.3227371573,0.2009095103,0.1705505401,-0.2606891692,0.1936377436,-0.2539634109,-0.2299778461,0.0371501669,-0.2520067394,0.2286065966,0.0620852262,0.4469465911,0.263094306,0.1255008727,0.278909564,-0.1579746455,-0.3334781826,-0.4113105536,0.0396113247,-0.1939235628,0.0128561752,0.0667111725,-0.4172227681,0.1737458557,-0.2322037965,-0.1728783697,-0.414814204,-0.6353833675,0.295322597,0.41248703,-0.2920282781,0.2850149274,0.0198659748,-0.0242610108,0.0108945584,-0.4168024361,0.3280274868,-0.1785102636,-0.250004828,-0.0537900366,0.0307765454,0.2872059643,-0.0685378611,-0.0351907574,0.0873222873,-0.1119816676,-0.2221680433,-0.2938449085,0.2362501025,0.4347923398,0.0297075938,0.2002646476,0.2416098565,0.3798702061,0.040974848,0.0787320659,0.2250270694,-0.1432593465,0.2511987984,0.3163281083,0.0934816748,-0.1110531166,0.3114649951,-0.0354225188,-0.4304850996,0.4005537629,0.0473905653,-0.1779955178,-0.0711910278,0.1111238748,-0.1035529524,-0.3938211501,0.0756265596,0.0673045516,-0.0421942621,0.0115949456,0.2065100223,0.1503222287,-0.1683192849,-0.4835576713,0.004646502,-0.0963150188,0.2801903188,-0.5498678684,0.1587651521,0.0191205144,-0.0018293647,-0.1598447561,-0.2013882399,-0.2839499116,0.1628508419,0.1431470811,-0.2605548501,0.0229084753,0.0297863483,0.0736310408,0.0467233174,-0.0548646897,-0.0768637583,-0.3278090656,-0.0589446165,-0.2479655892,0.1479586065,-0.2688071132,0.3106965423,-0.2542723715,-0.4377108216,-0.3863190413,-0.0467090271,-0.2399065495,0.1902445853,0.0341813192,-0.3199730814,-0.1186763346,-0.2228709906,-0.1941912919,0.2524873316,0.1282140166,0.1769001335,0.3718853891,0.2770297527,-0.1520201564,-0.1425380558,0.0944765732,0.2508153319,0.0297547448,-0.2445514053,0.2493702471,-0.0148770735,-0.2938300073,0.1417720169,0.2502235174,0.0238544606,-0.1566014141,0.0575945787,0.4005158544,0.1091093197,-0.1718084663,-0.1675898582,-0.1400907636,0.0771132186,0.0441411473,0.405872494,0.2937821746,0.073852174,0.2307672948,0.1385568529,0.4522203803,-0.0814950019,0.0300061498,0.2902031541,0.0642748997,0.0615383461,0.3179042637,0.4576174319,0.4516461194,0.3868235946,-0.4042824805,0.7036194205,0.1393147111,-0.1226203963,-0.0045483718,-0.2043928206,0.0706474632,0.0071851611,0.1542053372,-0.0275540072,-0.1269075274,0.6159351468,-0.2911730707,0.1290331632,-0.3046154082,0.2177102566,0.0140577909,-0.0595650487,-0.044104334,-0.3048033118,-0.2222346067,-0.0820649117,0.333661288,-0.1270152181,0.0322106704,-0.2303172946,0.1142927557,0.0121909557,-0.2177217454,0.3160595,-0.0185351614,-0.1616608053,-0.0005604536,0.008978189,-0.1450826973,0.0423344299,-0.033945214,-0.0057612956,0.1455549598,-0.083632566,-0.0650179833,-0.0410969481,0.0454203673,0.0488949902,0.0203191843,-0.0859280676,0.302105695,0.2990092635,0.0430934504,-0.0527017377,-0.0245905537,0.035627827,0.4039624333,-0.3784259856,0.5710351467,0.0384067073,0.0714472532,-0.1447587609,-0.0165708587,-0.2650619745,0.0684634373,0.2181402296,-0.1814795732,-0.1432467997,0.0499139763,0.0477016531,-0.0691485479,0.2825455368,0.1682593673,0.2423474938,0.0257301461,-0.1611098647,-0.4147757292,-0.2849624753,0.3607400656,-0.1499712169,0.1012367681,-0.0398398899,0.199895367,0.0146554671,0.1758114398,-0.2040684521,0.2794881165,0.025278572,-0.3216680884,-0.1745391786,0.1317727864,0.0878221616,0.0178296026,-0.0147397881,-0.2210965753,-0.1122673005,-0.160448432,-0.376701802,0.087348111,0.0309507605,-0.3266299665,0.3035985827,-0.061718028,-0.0260188282,0.1003752127,0.1862243414,-0.2419344038,-0.0144354003,0.0524668135,-0.301999867,0.1310646683,0.161187917,-0.1590801477,-0.2105103731,-0.4507226944,0.0619893186,0.1676827371,0.0370543376,-0.1672323942,0.1189489737,-0.4366020262,-0.0039181043,0.2037552297,0.1683233082,0.0583820529,0.1166526228,-0.2196156979,-0.0866042972,0.7436992526,-0.6192292571,-0.1626989096,0.0090176286,0.3121676147,0.1226929948,-0.1742892861,-0.6150499582,-0.1534845382,0.2068738192,-0.1163315475,0.1063900217,0.1825445443,-0.137316972,0.2070714533,-0.0748620182,-0.1884391308,0.1064032465,-0.0998461545,-0.0528154448,-0.1345544606]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2318","title":"[api request] API to obtain \"dataset_module\" dynamic path?","comments":"Hi @richardliaw, the feature is on the master branch and will be included in the next release in a couple of weeks.","body":"**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n","comment_length":22,"text":"[api request] API to obtain \"dataset_module\" dynamic path?\n**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n\nHi @richardliaw, the feature is on the master branch and will be included in the next release in a couple of weeks.","embeddings":[-0.1270397156,-0.225305289,-0.1368040591,0.1205858663,0.2940106094,-0.3005486429,-0.0833771154,0.1747500896,-0.2136734873,0.3675940335,-0.0689021051,0.7768819332,-0.4621844292,0.3776905835,0.2999127507,-0.4458760321,-0.0975358039,0.0067398096,-0.322686404,0.0176408105,-0.1013591662,0.2357002795,-0.0494804457,0.2150381953,-0.0331281647,-0.2271060199,-0.0215991996,-0.0324077271,-0.2154179066,-0.3063216805,0.2938253284,0.0977493972,0.241946429,0.2549487054,-0.0001097169,0.0399271362,0.0869922489,-0.0253267381,-0.3892126083,-0.5852143764,0.086991176,-0.0527439713,0.2232045084,-0.2114178687,-0.0571547747,-0.2286361903,-0.0388008952,-0.7245185375,0.3127938807,0.0970972478,0.1863785684,0.3525506854,-0.2461271733,-0.0608689785,0.0263192523,0.3660175502,-0.1790616959,-0.22310628,0.2439689934,0.1758071482,-0.0902159587,0.2793928981,-0.0147101153,0.0854094401,0.5754789114,0.0873526558,-0.5218421817,-0.0078821545,0.1398223191,0.4570405781,0.4978515506,-0.4548654556,-0.3124949336,-0.3435848653,-0.2102581561,-0.061194744,0.2432989478,0.1618670225,-0.3192092776,0.1748104542,0.0064415107,-0.2058244348,-0.296787858,-0.0477677807,0.1724842042,0.0452364311,0.0128453113,0.0101400046,0.2376514375,-0.2335121781,0.2568751872,-0.0087721646,0.1329406351,0.3248983026,-0.1687449962,-0.0509651899,0.2447003424,0.0347744301,0.1226374879,0.1836777031,0.1095809937,0.1045258865,-0.2002750337,0.0051323455,0.2168016136,-0.0542629547,0.168360725,0.1642332524,0.043715205,0.1569670737,0.0752833635,0.0036233903,-0.0262984224,0.2485636324,-0.1683986038,-0.068460539,-0.0536364317,0.0930320621,-0.0980267897,-0.2727898359,-0.2284814715,-0.2464374006,0.402005285,0.5189792514,-0.0103534022,0.0032723539,0.1346749365,-0.3061783016,0.1540817767,0.0342514217,-0.132878527,0.1700687557,-0.1964595616,0.2172746956,0.1251200438,-0.1617827266,0.3757390678,-0.2147464454,0.2080125064,-0.0146880941,0.1403799355,0.2258154005,0.302125901,0.056196183,0.0005969711,-0.0795303434,-0.196061492,0.1319363117,-0.4790400267,0.0572812036,-0.2756034732,-0.5086496472,-0.0531340875,0.1233089715,-0.1517894715,-0.1602447033,-0.2275237292,0.250295788,-0.4091596305,0.0668224245,-0.0462653302,0.1205807403,-0.251566112,-0.1417647004,0.1624892354,0.319162637,-0.0573723167,-0.2667974234,0.0106131323,0.0205455888,-0.2142793238,0.0149703249,-0.4976260364,0.1595586538,-0.2514204681,0.2069936395,0.4655227959,-0.3496899009,0.1979121715,-0.0496314839,-0.120559372,-0.3511764109,0.1423155665,-0.0178619046,0.0505459346,0.0221116059,0.1919242144,0.4152484834,0.1833771169,-0.0652413368,0.1088487878,-0.5048277974,-0.191604346,0.0347432494,-0.0281982888,-0.111048907,0.1313598752,-0.0748344138,0.4719240963,-0.128629446,-0.0647891089,0.2573931813,0.2756946981,0.2501274049,0.1638350338,-0.5089278221,-0.1832372248,0.1345116049,-0.223482579,0.0207616147,-0.1253190935,-0.2170520872,0.0546782538,0.2268900424,0.1063441262,0.2143793106,0.0887199268,-0.0532281138,-0.019846499,-0.0994883627,-0.5061228871,-0.0065902886,-0.3069564402,0.1298153549,-0.2599502504,0.263589561,0.1946105659,0.0440865792,-0.044573117,0.2768558264,0.0326316915,-0.1502081454,0.0656056702,0.1314016879,0.1356819868,0.1755419225,0.0802517757,0.4977859557,0.1447063684,-0.050052885,0.1689353287,-0.222633332,0.016557388,0.2989408672,0.1698878706,0.3969588578,-0.2993741035,0.3082689941,0.274392724,0.171094507,0.2389397472,0.0241720192,-0.2681674063,-0.0689967498,0.1494846195,0.1469724178,0.0204165131,-0.1865819544,-0.2366634011,0.1510697305,0.3428978026,0.0796978697,0.022950815,-0.0155319609,-0.1506819576,0.0635261834,0.1780986041,0.4427559078,0.2015550584,0.3840312958,-0.0342352688,0.0657879636,-0.1435747892,0.0426152982,0.480145961,0.4844023287,0.047703024,-0.3410339952,-0.067403391,0.0669592395,-0.3018414676,-0.3176433742,-0.1428513527,-0.0518546775,-0.2832365334,0.1495362222,-0.4663021863,-0.0819149539,-0.657181561,-0.4793883264,-0.0096388683,0.1438778043,-0.0076334723,-0.4043323994,-0.2307163328,0.1738907099,0.0714429244,0.3288937211,-0.2518242002,-0.5264746547,-0.114265956,-0.0889327675,-0.1420903355,0.0081492029,0.1766569018,0.1147655919,0.4805981219,0.0131506929,0.2469646335,-0.3617047668,-0.3256879449,0.4542023242,0.0781988725,0.0940469131,0.3639710248,-0.0134570859,0.1510278285,-0.1486555189,0.0746757463,-0.0454936512,-0.123011969,-0.4061354697,0.1090995073,0.0852028355,-0.0579174198,-0.3914550841,-0.2709724903,-0.2069480121,0.1916268766,-0.028917376,0.3386306465,0.3339799345,0.0931618363,0.2643855214,-0.4949843884,0.2537980676,-0.1022889838,-0.0478170477,0.1019314602,-0.0608179532,-0.2543387413,0.1836875081,-0.1126869917,0.3660377264,0.1741805226,-0.2480429113,-0.5372393727,0.0946132615,0.4121119082,-0.0586607344,0.1167090386,0.4918664992,0.3178735375,-0.0698463693,-0.0744871646,-0.6162253022,0.3979557157,-0.0852422044,0.2072379142,0.2233785987,0.3112289906,0.1901506782,0.7600682974,-0.1596223712,-0.42568627,0.0601219386,-0.1740730554,0.2207672447,0.2339052558,-0.2157909572,-0.2629825175,-0.2709943056,-0.095469363,0.0814522803,0.2104849219,0.0039520613,-0.2497093976,-0.0238252878,0.0264837593,-0.1686851531,-0.0031427913,0.198259294,0.4025060833,0.2280314863,0.2561766207,-0.4469816685,0.0559056066,0.2370326668,0.1111933291,0.1208506227,0.0199223664,-0.1564542502,-0.1482174098,-0.005031995,0.2213250548,0.2344180346,0.0478564911,-0.008420364,-0.0040333117,0.3566393852,0.1168896332,0.3550269902,0.3904077709,-0.2318372726,-0.2277101129,-0.2749612033,-0.0112798307,0.1291375756,-0.0172639303,0.0635054335,0.0876656622,0.4194371104,-0.1899559647,-0.4452036321,0.110592626,0.0996984839,-0.0870775655,0.0365874991,-0.0436143503,-0.5624928474,-0.0576422028,-0.1527360529,0.1215315387,0.2176288813,0.3544310033,-0.192473501,0.0319263265,-0.150006026,0.1050284728,-0.1843225956,0.4025805295,0.0508517399,0.2276389003,0.3149351776,0.491992712,-0.1902744472,0.1715232283,-0.3597531319,0.0065760948,-0.0965789184,0.2052886486,0.1015694737,0.4806386232,-0.032595288,0.5874004364,0.4865578413,-0.0947553217,-0.2769834101,0.0323534347,0.158147037,-0.0775581822,0.1104384586,-0.7509758472,0.8524295688,-0.0180823337,-0.2312948853,0.4544404745,0.2854153514,-0.3981997073,0.030143749,0.5078800917,0.8721266985,0.1242512837,-0.037690945,0.4104700983,-0.4999958277,0.3050886691,0.1763010323,0.1813927442,-0.2525940835,0.2059394568,-0.2253310531,-0.2310686558,0.0554068722,-0.2109219134,0.2230186611,0.0626591891,0.4559387565,0.2968809009,0.1157773808,0.2567669153,-0.149341926,-0.3624578714,-0.4080199897,0.0696040764,-0.1638591141,0.0116393734,0.069564648,-0.4029490352,0.1844186336,-0.2217804641,-0.12941508,-0.4136715531,-0.6379282475,0.2743671536,0.4318449497,-0.2715861797,0.2260098904,0.0433892347,-0.0286401641,0.0123944581,-0.3978767395,0.3720235527,-0.2066404372,-0.2087387294,-0.0821841061,0.022354776,0.3002604544,-0.0787798241,-0.0408403315,0.0990634337,-0.1257716566,-0.2412292361,-0.2255236208,0.2224192917,0.3901077807,0.0331859812,0.2292697579,0.2431899309,0.3690357804,-0.0099196695,0.102027148,0.1914518327,-0.1580658406,0.2423637062,0.3357068002,0.072722286,-0.1312860996,0.3530756533,-0.0332530029,-0.3652002513,0.4034724832,0.0571376272,-0.1780078411,-0.1003581285,0.0797298551,-0.0782631338,-0.3799103498,0.1269384623,0.0823453143,-0.0951909721,0.0366906375,0.2173463851,0.1455174983,-0.1780812442,-0.5291553736,0.0446872897,-0.104207173,0.2746506035,-0.5229539871,0.2043560892,0.0394080952,0.0369306579,-0.1375302523,-0.1906044334,-0.3217546046,0.1310565025,0.1074657068,-0.2314845026,0.0065797223,0.0149172042,0.1272598356,0.0509818569,-0.0200027805,-0.0790565833,-0.3563222587,-0.0998895466,-0.2363280505,0.1352294534,-0.2705956101,0.3342074752,-0.2743694186,-0.3926012814,-0.3475574553,-0.0800920352,-0.1927576065,0.1803085804,0.0111980969,-0.3118948042,-0.119164668,-0.254396677,-0.1364374608,0.2073961347,0.1417258829,0.1169023663,0.3541044891,0.2619065344,-0.2186802179,-0.1539203078,0.1065283418,0.2579851449,0.0185770057,-0.2296202034,0.2262381911,-0.0180714205,-0.3451118469,0.1336897463,0.2566279471,0.0240584332,-0.145364821,0.0515603311,0.3952967525,0.1313833594,-0.1798523813,-0.1597663462,-0.1133295819,0.0490257367,0.1101578698,0.4035378098,0.3108295798,0.081598103,0.2650838494,0.1428501308,0.4174602032,-0.087879926,0.0325314403,0.2982436419,0.0801320076,0.081343174,0.3302181959,0.4191918373,0.4734615088,0.4140080214,-0.4155042768,0.6876769662,0.0832934082,-0.1216451749,0.0012814695,-0.2178991586,0.0898938552,0.0082712434,0.1344248503,-0.0401978828,-0.1357735246,0.6071438789,-0.3177061379,0.1087908372,-0.3577255607,0.2070801705,-0.0203853697,-0.1085221693,-0.0752280727,-0.3105896711,-0.2112378031,-0.0754498392,0.3396895826,-0.1315981895,0.0607940294,-0.272448808,0.1377929002,-0.0065004742,-0.1932078451,0.2700701654,-0.0457599275,-0.1384177357,0.0202916246,0.002197823,-0.1543736458,0.0738035068,-0.0091056665,-0.0112820519,0.1025681719,-0.0627627596,-0.0828380957,-0.0517350659,0.0369274467,0.0181666203,0.0122521017,-0.1111308262,0.279679805,0.3424684107,0.0743121654,-0.0894782469,-0.0241070129,-0.0020486552,0.3875116408,-0.362311393,0.5390936136,0.0565106161,0.0931208432,-0.1340156198,-0.0465817265,-0.2787457705,0.0764466897,0.2130423933,-0.206896171,-0.1263848394,0.0543924645,0.0693988279,-0.037272431,0.3193175495,0.1211507469,0.2074967623,0.0297434609,-0.1358012259,-0.4270625412,-0.2781203389,0.3304317296,-0.1402631402,0.0962634534,-0.0102873249,0.2111393958,0.0173726678,0.1982536614,-0.2382508218,0.2866846025,-0.0140271857,-0.3081609011,-0.1660570502,0.0965938494,0.0895441696,0.0552214161,-0.0357670225,-0.220484063,-0.1086996198,-0.1243984029,-0.3356556594,0.1046070233,0.027272651,-0.3417004645,0.3023058772,-0.0926134214,-0.0048267469,0.0784266517,0.2078556865,-0.2367387563,-0.0439939611,0.0392095074,-0.3017890751,0.1240683347,0.1988803446,-0.1457273066,-0.2062171102,-0.428042382,0.0671990067,0.1955474168,0.0422182977,-0.1168776453,0.1359466016,-0.4302164316,-0.0137939258,0.2242737859,0.1927515715,0.0687335506,0.1144163683,-0.2228600234,-0.1458034068,0.7569946647,-0.628046751,-0.1650455892,-0.0156983994,0.291521579,0.1642668247,-0.1685465723,-0.6236018538,-0.1161717549,0.1995682269,-0.1070710272,0.0781649426,0.1628123373,-0.1534256339,0.1962367296,-0.0869594738,-0.1758803725,0.1021919176,-0.1038907021,-0.0632885471,-0.1314676255]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2318","title":"[api request] API to obtain \"dataset_module\" dynamic path?","comments":"awesome work @albertvillanova !","body":"**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n","comment_length":4,"text":"[api request] API to obtain \"dataset_module\" dynamic path?\n**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nThis is an awesome library. \r\n\r\nIt seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\n\r\nThis is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. \r\n\r\nI'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case.\r\n\r\nBy offering this API, we will be able to address the following issues (by patching the ray integration sufficiently):\r\n\r\nhttps:\/\/github.com\/huggingface\/blog\/issues\/106\r\nhttps:\/\/github.com\/huggingface\/transformers\/issues\/11565\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/34\r\nhttps:\/\/discuss.huggingface.co\/t\/using-hyperparameter-search-in-trainer\/785\/35\r\n\r\n\nawesome work @albertvillanova !","embeddings":[-0.1235195994,-0.2129467726,-0.1454212964,0.1316784769,0.2975060344,-0.2882803679,-0.1093318313,0.1675436944,-0.2321175933,0.3611184061,-0.0613422617,0.7779220939,-0.4597415626,0.3714591265,0.2889253497,-0.4567815065,-0.1055705771,0.008220938,-0.2979366481,0.011237124,-0.1017431617,0.2270932794,-0.0599272847,0.2292527407,-0.0150762442,-0.2438205034,-0.0263518505,-0.0364981182,-0.2267149836,-0.3021511436,0.3008244932,0.092894949,0.2565479577,0.2530418038,-0.0001097595,0.0302356426,0.0839397758,-0.0181854274,-0.3759013414,-0.5799815059,0.0720690414,-0.0689032227,0.2171185315,-0.2307339609,-0.0289479755,-0.2544797361,-0.0382940173,-0.7315433025,0.3256861866,0.098178938,0.1890392601,0.3301767111,-0.2411682457,-0.035413865,0.0308030695,0.3716481328,-0.1764228195,-0.2218154967,0.2680900097,0.1705570221,-0.0901674926,0.2833645642,-0.0264832769,0.0795286074,0.5768831372,0.0678419769,-0.5200684071,-0.0220175926,0.1428516805,0.447545886,0.5037162304,-0.4579201341,-0.3037011325,-0.3232639432,-0.2070348859,-0.0666099563,0.2223795801,0.1783015579,-0.308994323,0.1628851295,-0.0114134476,-0.2161559165,-0.3109675944,-0.0565504245,0.1756594479,0.0519175231,0.0206788182,0.014722948,0.2274508029,-0.245881319,0.2732420862,0.001260209,0.1373042464,0.3271966279,-0.1731330752,-0.0453802049,0.2340688705,0.0166646466,0.1157825366,0.1877549738,0.1055251732,0.1008433923,-0.1917366385,-0.0050071245,0.2099799216,-0.0491865464,0.1697733104,0.1808643937,0.0405807495,0.152415514,0.0772749111,0.0077800611,-0.0397094004,0.2585529089,-0.1516104937,-0.0675424039,-0.0564775802,0.0967974365,-0.1046222448,-0.2639604807,-0.2434509546,-0.2360841036,0.3941462934,0.5158168077,-0.00995964,0.0021620675,0.1383769959,-0.3086359501,0.1622195989,0.0260991007,-0.1287235171,0.1883672774,-0.1929554045,0.2164583057,0.1220329106,-0.1568613797,0.3627625108,-0.2190739214,0.221993506,-0.0203846227,0.1528937668,0.2193863839,0.2931146324,0.0483277999,0.0059941714,-0.0768813118,-0.2022196352,0.1398389041,-0.4742925465,0.0573124886,-0.2900441289,-0.5152980685,-0.0492533967,0.1282336414,-0.1408796906,-0.1493531615,-0.2270478457,0.2249857932,-0.4121249914,0.0687299371,-0.0382150188,0.1057880595,-0.2466267347,-0.1457213759,0.1600653678,0.3075908124,-0.0609425716,-0.2508646846,0.0003785124,0.0320375413,-0.1965206712,0.0065217395,-0.4949878752,0.1584880352,-0.2345175445,0.2152707428,0.4839906693,-0.3609963059,0.1825468838,-0.053247273,-0.1086631268,-0.3760794401,0.1334020346,-0.0129046068,0.0686126724,0.0339317173,0.1770154834,0.4179819524,0.189646855,-0.0776028559,0.1088229865,-0.5219867826,-0.1771243215,0.058146432,-0.0163488965,-0.1314736903,0.1323249191,-0.0681199357,0.4681271017,-0.1190659627,-0.0702559724,0.2610254586,0.2670190334,0.2439154536,0.1721882969,-0.5049237609,-0.1641213298,0.1335295588,-0.2264221609,0.0174754634,-0.138375476,-0.2020910531,0.0590308681,0.2268193811,0.0977479145,0.2121064961,0.0937867165,-0.0683980584,-0.0250807833,-0.1102238595,-0.4998072386,-0.0036447532,-0.3121667504,0.1178364977,-0.2572193444,0.2635036707,0.1885767281,0.045493979,-0.0393308997,0.2784275115,0.0412752591,-0.1564606428,0.0580539927,0.1200410873,0.1285791993,0.1827969104,0.0695614219,0.5018494129,0.1360618025,-0.06232417,0.1747602075,-0.2153770626,0.0236706454,0.3002019525,0.1885359436,0.3824970126,-0.3175560236,0.3075825274,0.2762338519,0.1816346198,0.23698318,0.0309340972,-0.266163528,-0.0620889738,0.1603273004,0.1518375427,0.0218448713,-0.20147039,-0.2209951878,0.1438581198,0.3369916379,0.0646541119,0.0187542699,-0.0183163974,-0.1388073117,0.0660874248,0.1876347363,0.4441204369,0.202424705,0.3858090937,-0.0549058579,0.0671587065,-0.1260344982,0.0471274145,0.481741637,0.4879460633,0.0502936766,-0.3586165905,-0.065025501,0.0798940063,-0.2813826203,-0.3513381481,-0.1445528716,-0.0465051644,-0.2778961957,0.1567336917,-0.4632063806,-0.0942371935,-0.6389959455,-0.4855386913,0.0004112917,0.14201653,-0.0191555265,-0.4196766913,-0.2162741721,0.1651802361,0.0624544248,0.3347507715,-0.2452793121,-0.5293625593,-0.1110782474,-0.0929849073,-0.1553521454,0.009219015,0.1701626629,0.1264100522,0.4897284508,-0.0032089176,0.2628183365,-0.3540847301,-0.3020901978,0.4527813494,0.0854158178,0.0772527158,0.3571250439,-0.0072837346,0.1463849097,-0.1522010267,0.0646870807,-0.0343740769,-0.1154635847,-0.4052948654,0.1012460887,0.0678163245,-0.040928863,-0.3848119974,-0.2942708731,-0.2006362528,0.1633167416,-0.0377856344,0.3474760652,0.3295829296,0.0769839063,0.2679837346,-0.4769464135,0.2720761597,-0.0910995901,-0.0481851026,0.1106543839,-0.0478020012,-0.261929363,0.1979567111,-0.1008032262,0.3765464127,0.1913295835,-0.2490435392,-0.5310202241,0.0950435475,0.4171668589,-0.0741218477,0.1039132252,0.4974088073,0.3218806386,-0.0581721924,-0.0656398311,-0.6264390349,0.3956491053,-0.0711627826,0.1999528408,0.2290048748,0.3185458779,0.1800083071,0.7651033998,-0.1430972517,-0.4281297624,0.0659140274,-0.1566709578,0.1976844519,0.2571000159,-0.2290777862,-0.2549378276,-0.2675397992,-0.1034706011,0.0733744949,0.2149969488,0.0015149051,-0.2493592203,-0.0388682038,0.0155445598,-0.1667148173,-0.0311193764,0.1903120279,0.4279967844,0.2288799882,0.2653007209,-0.4430680573,0.0572748519,0.2435655296,0.1190731749,0.1204751655,0.0224878453,-0.1430853903,-0.1525773257,0.0093451133,0.2305473685,0.2403454483,0.0562480092,-0.0025713819,0.012651273,0.3610892296,0.1065056249,0.34610641,0.3804276288,-0.2187417746,-0.233607918,-0.2533991933,-0.0045086392,0.1211583167,-0.019260468,0.0604617223,0.0964101404,0.4115911722,-0.1777203679,-0.4282768369,0.10484799,0.1137032434,-0.0928164124,0.0320362151,-0.0402143337,-0.5746632814,-0.0575913377,-0.1594376564,0.1056602821,0.2271474004,0.3589656949,-0.1772580445,0.0194962937,-0.1432567686,0.1095982119,-0.1811476052,0.3865906894,0.0377578773,0.2251571417,0.2992192507,0.4852485061,-0.2039067894,0.1626007855,-0.3590512574,-0.0129240779,-0.1072495729,0.2010064423,0.0840502977,0.4739887118,-0.0279347245,0.5851618648,0.4860494733,-0.094111003,-0.275598824,0.0264668111,0.1545367986,-0.0776021183,0.0952638909,-0.7681438327,0.8332893848,-0.0020385091,-0.2328877896,0.4483564198,0.2982200086,-0.3976546228,0.036130324,0.5277879834,0.8905420899,0.1269765198,-0.0472048335,0.4033127129,-0.5028147697,0.3033471406,0.1722150296,0.1802416891,-0.2479023486,0.1912981719,-0.2333951741,-0.2229727954,0.052127663,-0.2304918766,0.2254910171,0.0721341521,0.4702086449,0.281067729,0.1255818605,0.2685324252,-0.1452265531,-0.349614054,-0.3963912725,0.0703806505,-0.1837145388,-0.0012118113,0.0736522749,-0.4045619071,0.1861966252,-0.2250355184,-0.1398214996,-0.4126962721,-0.6360215545,0.2944432199,0.4200670421,-0.2806042731,0.2462684214,0.0378165133,-0.0290013291,0.0128112109,-0.4015739262,0.3555457294,-0.1908158958,-0.2174788862,-0.0686787069,0.0224416126,0.2942309678,-0.0834321901,-0.0293133259,0.1073824689,-0.1146005765,-0.2156724036,-0.2437959015,0.2361137122,0.4085608125,0.04071467,0.2233376801,0.2400912195,0.3691821396,0.0097975759,0.1025284827,0.1883865893,-0.1441885978,0.2483382076,0.3264454305,0.0767828226,-0.1234833151,0.3358946145,-0.0382364243,-0.3731149733,0.4016260505,0.0411933362,-0.182248801,-0.101059638,0.0945196301,-0.0939629376,-0.3864131272,0.1047559306,0.0607095398,-0.0813983381,0.0318483561,0.210593313,0.1479676068,-0.1794154793,-0.505048275,0.0323646329,-0.1038588285,0.2819686532,-0.5254949331,0.176445812,0.0391433388,0.0194407981,-0.1514148414,-0.2003892213,-0.3206597567,0.1388090998,0.105799906,-0.2390641272,0.0064020334,0.0209077876,0.0983133018,0.0503931828,-0.0213209148,-0.0772660375,-0.3517033756,-0.097509861,-0.2411739379,0.1330244243,-0.2766582072,0.3133024871,-0.2704691589,-0.4174567759,-0.3623639047,-0.0671116859,-0.2102098167,0.1636414081,0.0123657892,-0.3173542619,-0.136455074,-0.2462094128,-0.1636691242,0.2190497369,0.1314278692,0.1351769567,0.3537027538,0.2821862698,-0.1771048158,-0.1531466842,0.097943753,0.2494394034,0.0315592363,-0.2441201657,0.2276178449,-0.0198333357,-0.3199503124,0.1288930625,0.2590270042,0.0277256146,-0.1569231302,0.0460064933,0.3989006877,0.1357211024,-0.1802713126,-0.1676594913,-0.136313796,0.0602561571,0.0978849977,0.4047309458,0.3025794327,0.0777039006,0.2379756421,0.1461731791,0.4372665286,-0.0857618749,0.0280483384,0.3044452667,0.0751446038,0.075803034,0.3219586611,0.4336282611,0.457463026,0.4145367742,-0.4089034796,0.7172420025,0.1134332195,-0.1123367324,0.0042442754,-0.2085939795,0.0899351686,0.0084280483,0.1405514479,-0.0392522179,-0.1292493939,0.5968949795,-0.302041173,0.1166353226,-0.3412135839,0.2032292932,-0.0083738007,-0.0945436954,-0.0622063875,-0.3106940389,-0.2207392603,-0.0767056942,0.3374914229,-0.1359111071,0.050795868,-0.2535589635,0.1267877072,0.0082887672,-0.1935077161,0.2789848447,-0.0505364388,-0.1523425579,0.0093677454,0.0181664191,-0.1581915021,0.0539879464,-0.0196640491,-0.0008604829,0.1143902838,-0.0886440352,-0.078864187,-0.035216596,0.0331205167,0.0249044783,0.0109622274,-0.1063539982,0.2856220007,0.3378874063,0.0764028728,-0.0857575759,-0.0329716057,0.0164459981,0.3971990645,-0.3694297671,0.5578833818,0.0530893244,0.0879878923,-0.1328178495,-0.0312937833,-0.2781651616,0.0875936896,0.2235903293,-0.2002781183,-0.1245881468,0.0539740324,0.0667039007,-0.0466759503,0.315269351,0.1300387084,0.2197506279,0.027327884,-0.1397500783,-0.4227002859,-0.2712610662,0.3342705667,-0.1539742351,0.0944451466,-0.0207823291,0.2021907866,0.0175552107,0.1915638149,-0.2037034035,0.2801629007,0.0046908557,-0.3100990653,-0.1667499095,0.1211380884,0.0886008814,0.0426503234,-0.0447238646,-0.2237106413,-0.1181129739,-0.1249560788,-0.3439107239,0.0933697522,0.0365253761,-0.3440477848,0.3037234843,-0.0899616703,-0.0133908652,0.0831494406,0.2090787441,-0.2336117178,-0.0195588674,0.0470437892,-0.3091340363,0.1356414407,0.183193624,-0.1479846239,-0.2002520114,-0.4356529415,0.0588429719,0.1864165366,0.046225097,-0.1401934177,0.1325750947,-0.4271811545,-0.0114650279,0.2095905244,0.179301694,0.0679821819,0.1108358726,-0.2301347852,-0.131327644,0.7448346615,-0.6288006306,-0.1708487719,-0.0077181901,0.2932465076,0.1553701609,-0.1628079265,-0.6073155403,-0.1271734536,0.1987855285,-0.1169762835,0.0755172744,0.1684044749,-0.1390177161,0.2043369114,-0.0804120898,-0.1817895919,0.1069691405,-0.1172403693,-0.07000193,-0.1270733923]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2316","title":"Incorrect version specification for pyarrow","comments":"Fixed by #2317.","body":"## Describe the bug\r\nThe pyarrow dependency is incorrectly specified in setup.py file, in [this line](https:\/\/github.com\/huggingface\/datasets\/blob\/3a3e5a4da20bfcd75f8b6a6869b240af8feccc12\/setup.py#L77).\r\nAlso as a snippet:\r\n```python\r\n \"pyarrow>=1.0.0<4.0.0\",\r\n```\r\n\r\n## Steps to reproduce the bug\r\n```bash\r\n pip install \"pyarrow>=1.0.0<4.0.0\"\r\n```\r\n\r\n## Expected results\r\nIt is expected to get a pyarrow version between 1.0.0 (inclusive) and 4.0.0 (exclusive).\r\n\r\n## Actual results\r\npip ignores the specified versions since there is a missing comma between the lower and upper limits. Therefore, pip installs the latest pyarrow version from PYPI, which is 4.0.0.\r\nThis is especially problematic since \"conda env export\" fails due to incorrect version specification. Here is the conda error as well:\r\n```bash\r\nconda env export\r\nInvalidVersionSpec: Invalid version '1.0.0<4.0.0': invalid character(s)\r\n```\r\n\r\n\r\n## Fix suggestion\r\nPut a comma between the version limits which means replacing the line in setup.py file with the following:\r\n```python\r\n \"pyarrow>=1.0.0,<4.0.0\",\r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.2\r\n- Python: 3.7.10 (default, Feb 26 2021, 18:47:35) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.0-42-generic-x86_64-with-debian-buster-sid\r\n```\r\n","comment_length":3,"text":"Incorrect version specification for pyarrow\n## Describe the bug\r\nThe pyarrow dependency is incorrectly specified in setup.py file, in [this line](https:\/\/github.com\/huggingface\/datasets\/blob\/3a3e5a4da20bfcd75f8b6a6869b240af8feccc12\/setup.py#L77).\r\nAlso as a snippet:\r\n```python\r\n \"pyarrow>=1.0.0<4.0.0\",\r\n```\r\n\r\n## Steps to reproduce the bug\r\n```bash\r\n pip install \"pyarrow>=1.0.0<4.0.0\"\r\n```\r\n\r\n## Expected results\r\nIt is expected to get a pyarrow version between 1.0.0 (inclusive) and 4.0.0 (exclusive).\r\n\r\n## Actual results\r\npip ignores the specified versions since there is a missing comma between the lower and upper limits. Therefore, pip installs the latest pyarrow version from PYPI, which is 4.0.0.\r\nThis is especially problematic since \"conda env export\" fails due to incorrect version specification. Here is the conda error as well:\r\n```bash\r\nconda env export\r\nInvalidVersionSpec: Invalid version '1.0.0<4.0.0': invalid character(s)\r\n```\r\n\r\n\r\n## Fix suggestion\r\nPut a comma between the version limits which means replacing the line in setup.py file with the following:\r\n```python\r\n \"pyarrow>=1.0.0,<4.0.0\",\r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.2\r\n- Python: 3.7.10 (default, Feb 26 2021, 18:47:35) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.0-42-generic-x86_64-with-debian-buster-sid\r\n```\r\n\nFixed by #2317.","embeddings":[-0.3293253481,0.0613468364,-0.1033367887,0.0941155851,0.1090466529,-0.1797627658,0.2043182254,0.4323440194,-0.4242305458,0.0479667597,0.1352877915,0.4154728353,0.1616445631,0.1031846479,0.0247078035,-0.1516472697,0.1754688323,0.2537163198,-0.2592993379,-0.0661627054,-0.046094656,0.3526418507,-0.189809978,0.2315452695,-0.2163686305,0.0538548119,-0.0411416963,0.0746229663,-0.2986971736,-0.4668654799,0.3813257813,-0.0361553431,0.1174036041,0.3492774069,-0.0001052237,0.0277146436,0.1867129654,-0.0540586486,-0.1895418912,-0.320959717,0.1390068531,-0.3200803995,0.1888888031,-0.132141009,0.1311011761,-0.3027854264,0.038142655,0.2845487297,0.281165272,0.1622896194,0.3061073422,0.335278511,0.4839234054,-0.1481698304,0.2130458802,-0.0130237993,-0.188703835,0.488160044,0.0443516411,-0.1712583452,-0.0440881588,0.0696140826,-0.1064724699,-0.1078136712,-0.0029544374,0.1725744605,0.22390154,-0.1449862421,-0.121186398,0.1718390435,0.2160875499,-0.5827276707,-0.4012673795,0.0241212491,-0.0008576339,-0.3473529518,0.1414670348,0.2095406204,-0.333848834,0.3731095791,-0.17317155,0.1086833775,-0.2264311761,-0.0772836879,-0.2200175226,0.2009439617,0.0589540005,0.0952578634,0.0760676786,-0.058706522,0.1703604013,0.1857180297,-0.0988129303,0.046869915,-0.1297956109,0.0574927554,-0.0306528956,0.172236383,0.3219581544,0.062497586,-0.0737257004,-0.0239403136,0.2240965068,0.2091137171,-0.0510442145,0.1456472427,-0.0136546018,0.0382199548,0.2326868623,0.0353379585,0.2764451504,0.0109081389,0.1045693755,-0.3013595343,-0.0381445102,-0.0307929311,0.2257788181,-0.1361337453,-0.4114389122,0.0781747997,-0.1830581427,-0.0860646367,-0.1615540683,0.3412765861,-0.1650919467,-0.051098574,0.0832005814,0.3095273972,-0.2042611986,0.0075881174,-0.1478618234,-0.1037297696,-0.1207627654,-0.0771150514,0.0945199654,-0.1798327267,0.3564105928,0.1775455624,0.118987821,0.2616992295,-0.1589668244,0.0141710956,-0.1544570774,0.3122213781,-0.0442310274,0.0632874817,0.1040203199,-0.1985595673,-0.2106844187,0.1250106841,-0.0205734391,-0.199022606,-0.5355875492,0.2504501045,-0.1581709534,0.0293771271,0.0034543583,-0.14783445,0.0300420746,-0.2212673575,-0.0211083572,-0.2772291005,0.2893983424,-0.3916004896,0.3669196963,0.1085338965,-0.4669374526,-0.0420990735,0.372186631,-0.402235359,0.0657996982,0.1088156924,-0.1545504779,0.0713914782,0.0014730332,0.0216971021,0.1729075164,-0.3737107217,-0.4191465974,-0.121221967,-0.0169613697,0.0537693799,-0.0118430611,-0.1711357981,0.1623157412,0.0189831294,-0.0410342701,0.1893713474,0.0756572783,-0.0790025145,-0.2709807456,-0.4034540355,-0.0419960581,0.1258122623,-0.0791862607,-0.2095832825,0.0670630112,-0.4043299556,0.3473365307,-0.0944849849,0.0252832342,0.0492690764,0.3415067196,0.0838256627,0.0381655693,-0.0390198976,-0.392875433,0.1174721867,-0.2547550797,0.080548048,-0.4295913875,-0.1674927175,-0.1604496837,0.3074555993,-0.0418503694,-0.1806700975,0.2436359227,-0.0074701165,0.2683659792,0.1743633598,-0.0777883381,0.3098963499,0.1421180367,0.2211226225,0.034110669,0.2832371593,-0.4890272617,-0.3314470947,-0.0138839511,0.285951972,0.2144787461,-0.1515024453,0.0610691607,0.3274956346,-0.1227918565,-0.101743564,-0.239798665,-0.1654363424,0.0491140857,-0.2266006917,0.1774207205,0.0003486252,-0.101185225,0.1426268816,0.1894907206,0.1896850318,0.2209903896,-0.01608878,-0.02134756,0.1157084331,0.3608475626,-0.0167750958,0.1227014586,-0.1656402946,0.1205751151,0.2075038105,0.2094353288,-0.0532915257,-0.3605071902,0.0773165897,0.3440695405,-0.1500602216,-0.0024427299,0.1351140141,-0.1480791122,0.166579023,0.266492188,-0.2073448151,0.1913185418,0.2652365863,-0.1097983047,0.1217498109,-0.1287274361,-0.1556119323,0.2028847188,0.0023200568,0.1581461132,0.0936209634,0.158900395,-0.0623839386,-0.2236666679,-0.0923313126,-0.21738787,0.135978356,-0.1987077147,0.1780671775,-0.3118232787,0.0673016235,-0.1731317192,-0.6719259024,-0.3010737002,-0.0502465926,0.2024639994,0.1333040744,0.0256907362,0.4673159719,0.1424296349,-0.2037860453,-0.0630650297,-0.4036450088,-0.0976305753,-0.1236376613,-0.4228557646,0.1430127472,0.0813175887,-0.0155181652,0.0724253729,-0.2368448079,-0.0250510741,-0.2270698696,-0.4384095669,0.1378343552,-0.4806168973,0.4755058289,0.344029218,0.1300124675,-0.0463802889,-0.248724997,0.2872928083,-0.1248899996,-0.2263915539,0.0435730964,-0.0778800547,-0.2906534076,-0.2573845387,-0.218208164,-0.4303585291,-0.2946029902,0.3301987648,0.218099162,0.098419778,0.1522650123,0.4002890885,0.0085725309,-0.1158535182,0.1882134825,-0.0120231137,-0.0879467055,0.2370005101,-0.1813811064,-0.3930126727,0.009854109,0.0455041789,0.2319517881,0.0633845404,-0.4236885607,-0.236452058,0.0823147371,0.7113948464,-0.0548868142,0.0664705858,0.4633465111,0.2060410827,-0.055312112,-0.1593297869,-0.2722122669,-0.049432341,-0.1940241754,0.114145115,-0.0162031204,0.2518673837,-0.058787901,0.5857924223,0.2168657929,-0.154530555,0.2992658615,-0.1270862967,0.4248113334,-0.1714911759,-0.0907654017,0.020985892,-0.0412753001,0.0452066101,0.1764709353,-0.0370213278,-0.2535754144,-0.2158640325,0.0456515923,-0.3392736018,-0.2188977003,-0.0055488949,0.3993224502,0.1759165376,0.1165815815,0.3013985753,-0.0686802864,0.0012773873,0.0607332475,-0.0396783799,-0.2266084254,0.0639575273,-0.07315927,-0.1869666427,-0.4118630588,0.3873225749,0.111157611,-0.0122864079,0.1472923756,-0.2160772234,0.1190051511,-0.1082766652,0.5949860215,0.1399854571,-0.186591655,0.1525640637,0.2598967254,-0.5327020288,-0.2362198234,-0.0058069127,0.064704299,0.4111000896,0.0248274654,-0.503857851,-0.2122707367,0.4767184258,0.3166432977,0.0481977388,-0.0495377705,-0.4007003605,-0.4447710216,-0.182782799,0.0476032645,-0.0854070112,0.194715485,-0.0983999297,-0.04115358,0.0216735248,0.0959654823,-0.1554980576,0.4328534305,0.2383036762,-0.398876369,0.1149061024,-0.0508430861,0.0226443391,0.4592907429,0.3905208707,-0.0181481875,-0.2505517006,-0.1318744272,-0.038825497,0.1390227526,0.1336060762,-0.1228039041,-0.0401397571,-0.1681434214,0.120873265,0.0321735516,-0.1178957522,0.4463046491,0.0252871756,-0.3146312535,0.0403658301,0.3002142012,0.1123904809,-0.0972419381,0.1197596192,0.2501278222,-0.1535490155,0.2365332544,0.0663883016,0.8796625137,0.0524958298,0.146524474,0.5330759883,-0.3249711692,0.0512233414,-0.0824112892,0.0325842611,-0.3899542689,-0.1889344454,0.0051186346,-0.065890871,0.1733802557,-0.1188552529,-0.3746958971,0.1945249587,-0.0403671488,0.1530838162,0.1817508191,-0.0982728228,-0.004683787,-0.1071982682,-0.2389144897,0.2295481116,0.1135735735,0.0466433652,-0.1790203303,-0.1219273359,-0.2541183829,-0.4695954323,-0.264923811,0.0976735577,-0.3670772016,0.0509488396,0.2838154733,-0.4009480476,-0.1341637671,0.1047598645,0.0032647736,0.1159105599,-0.1210362539,0.1247271523,0.019114038,-0.1292452067,0.0832043961,0.0944281518,0.1535068601,-0.1368537843,-0.3145421147,0.1585733593,-0.1132577583,-0.0354715399,0.1456578374,0.0980617702,0.1635437906,-0.196000576,0.1814534962,-0.1310325563,0.1861944497,-0.2179530859,0.2099562734,-0.0461457148,0.114586845,0.3664984405,-0.2001594603,-0.1493896246,-0.224314332,0.1128523946,0.000319701,0.0173492059,0.1889068037,0.0618466735,-0.2567341328,-0.1768965274,-0.2231314331,-0.252230078,-0.5580479503,-0.1639229357,0.1225187182,-0.0022524991,-0.0126863811,0.3384084404,0.1924357712,-0.0212451592,0.0118161971,-0.1916674674,-0.2103589028,0.36499542,-0.0371272899,0.0065912725,0.3337614238,0.3662502766,0.1420865804,-0.0800987333,-0.3967453241,0.0617961697,-0.0848206729,0.1311184019,0.0639746487,-0.0862734839,0.3302310109,-0.0459184796,0.2952847779,0.1084055603,-0.2434734553,-0.2890139222,-0.099945493,0.0397814661,0.0088201053,-0.1723639667,0.1582697779,0.0434240624,-0.0600672513,-0.0973254517,0.2177072465,0.1783843488,-0.0232946165,0.0544806868,0.1313370466,0.3154245913,0.0430142619,-0.1311328709,0.0316684805,0.3279381096,0.0505274199,0.2542048097,-0.2126435637,0.0010273649,0.0794543326,0.1228700206,0.0741543472,0.0475755259,0.1854095608,-0.2598520815,-0.2091854811,-0.0874058381,0.5106782317,0.2949961424,-0.1765620559,-0.0205495059,0.5580257177,0.3270153701,-0.3597737551,0.021111628,0.1520129144,0.0680902228,-0.02101494,0.2339702398,0.1657903492,-0.3489792347,0.2064490318,0.2095534652,0.3763013482,0.0670302063,0.0699480698,0.3206043541,0.0728342384,0.0998959467,0.4084194899,0.1044874787,0.3153886199,0.223275125,-0.4326778054,0.0168834962,-0.2191358805,-0.0473701805,-0.0987266228,-0.0962391868,-0.0989297405,-0.0838767588,0.219338879,-0.2212996036,0.1237929985,0.5065464377,0.0314564034,-0.1168002188,-0.343757391,0.2599305511,-0.1198728681,0.0327203274,0.0128668156,-0.0211701933,0.183626771,-0.0413369089,-0.1070670635,-0.2952034771,0.5606056452,0.2222759724,-0.1070805714,-0.4725451171,0.0968666673,0.1217810288,-0.0088638887,-0.034368109,0.4105711877,0.3222196996,-0.2995681167,-0.3258165419,0.420039922,0.7677022219,0.2794877589,0.0502342358,-0.1065219194,-0.2805844247,0.0992316753,-0.1560940593,0.1957740635,0.0568641387,0.2759606242,0.251834631,0.3021621108,-0.2460110635,-0.1331487149,0.2334879786,0.3410659134,-0.2025342435,0.1807749271,-0.1899020523,-0.2077006847,-0.0467238612,-0.0854666531,-0.3218032718,0.4074016213,0.110907048,0.2778560519,-0.0352341346,-0.0953776464,0.1318977177,0.1153359264,0.3660244048,0.2457199842,0.4075174034,-0.144238919,-0.1997267753,-0.3286935687,0.2906815708,-0.4103168547,0.0428001434,-0.0439622402,0.0650557652,0.0973325297,0.2076601237,0.300968051,0.0974480808,-0.0175859127,0.0856944025,-0.3471741974,0.0478885062,0.0164919905,-0.2352905869,0.1410652846,-0.2627708912,0.1050517187,-0.4169732332,0.1712429523,-0.2676775157,0.0953667685,0.1639345139,-0.0832864493,0.5615702868,0.0679681674,0.2375492007,-0.0096863676,-0.125141263,-0.2957378626,-0.2580848932,-0.2334179729,0.1816101372,0.00929577,0.1356420815,-0.162859112,-0.3730594814,-0.2566678822,0.4734206796,0.1324440092,0.0122010699,0.0797727853,0.2107094228,-0.0507361069,-0.1312570423,-0.0914758816,0.1109829322,-0.0968874022,0.1835145354,-0.5022741556,-0.4769975543,0.4085148573,-0.4291891158,-0.1012323573,0.0166277681,0.0838664845,0.022117598,-0.2742372155,-0.4858311117,0.2620688379,0.2436240166,0.0419571362,-0.3534299433,0.2466799617,-0.2294899821,-0.193641454,-0.0530626513,0.5304144621,0.024812948,-0.3870817721,0.1323703378,-0.2105240375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2308","title":"Add COCO evaluation metrics","comments":"Hi @NielsRogge, \r\nI'd like to contribute these metrics to datasets. Let's start with `CocoEvaluator` first? Currently how are are you sending the ground truths and predictions in coco_evaluator?\r\n","body":"I'm currently working on adding Facebook AI's DETR model (end-to-end object detection with Transformers) to HuggingFace Transformers. The model is working fine, but regarding evaluation, I'm currently relying on external `CocoEvaluator` and `PanopticEvaluator` objects which are defined in the original repository ([here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py#L22) and [here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/panoptic_eval.py#L13) respectively). \r\n\r\nRunning these in a notebook gives you nice summaries like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/48327001\/116878842-326f0680-ac20-11eb-9061-d6da02193694.png)\r\n\r\nIt would be great if we could import these metrics from the Datasets library, something like this:\r\n\r\n```\r\nimport datasets\r\n\r\nmetric = datasets.load_metric('coco')\r\n\r\nfor model_input, gold_references in evaluation_dataset:\r\n model_predictions = model(model_inputs)\r\n metric.add_batch(predictions=model_predictions, references=gold_references)\r\n\r\nfinal_score = metric.compute()\r\n```\r\n\r\nI think this would be great for object detection and semantic\/panoptic segmentation in general, not just for DETR. Reproducing results of object detection papers would be way easier.\r\n\r\nHowever, object detection and panoptic segmentation evaluation is a bit more complex than accuracy (it's more like a summary of metrics at different thresholds rather than a single one). I'm not sure how to proceed here, but happy to help making this possible.\r\n\r\n\r\n\r\n","comment_length":28,"text":"Add COCO evaluation metrics\nI'm currently working on adding Facebook AI's DETR model (end-to-end object detection with Transformers) to HuggingFace Transformers. The model is working fine, but regarding evaluation, I'm currently relying on external `CocoEvaluator` and `PanopticEvaluator` objects which are defined in the original repository ([here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py#L22) and [here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/panoptic_eval.py#L13) respectively). \r\n\r\nRunning these in a notebook gives you nice summaries like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/48327001\/116878842-326f0680-ac20-11eb-9061-d6da02193694.png)\r\n\r\nIt would be great if we could import these metrics from the Datasets library, something like this:\r\n\r\n```\r\nimport datasets\r\n\r\nmetric = datasets.load_metric('coco')\r\n\r\nfor model_input, gold_references in evaluation_dataset:\r\n model_predictions = model(model_inputs)\r\n metric.add_batch(predictions=model_predictions, references=gold_references)\r\n\r\nfinal_score = metric.compute()\r\n```\r\n\r\nI think this would be great for object detection and semantic\/panoptic segmentation in general, not just for DETR. Reproducing results of object detection papers would be way easier.\r\n\r\nHowever, object detection and panoptic segmentation evaluation is a bit more complex than accuracy (it's more like a summary of metrics at different thresholds rather than a single one). I'm not sure how to proceed here, but happy to help making this possible.\r\n\r\n\r\n\r\n\nHi @NielsRogge, \r\nI'd like to contribute these metrics to datasets. Let's start with `CocoEvaluator` first? Currently how are are you sending the ground truths and predictions in coco_evaluator?\r\n","embeddings":[-0.3099649251,-0.210986048,-0.0590808243,0.0364998765,0.1357953399,-0.1280122548,0.0327846818,-0.1446955502,-0.2019299865,0.1269806623,-0.6827290654,0.1239695549,-0.1395809948,0.1238021106,-0.2454466075,-0.0975500122,-0.0049184803,-0.0187490061,-0.217513442,0.1276486516,0.0637156069,0.1212180033,0.2470147163,0.1320974678,-0.2228360325,-0.1528722197,-0.2780018449,-0.3175072372,-0.1612353325,-0.3200707436,0.3088329732,0.4587270916,0.2178368121,0.3828972876,-0.0001260293,0.1676148772,0.0120117078,0.0352617018,-0.0486765914,-0.0712401196,0.0626677647,-0.3141658008,0.2967291176,-0.1566159576,0.2011528611,-0.0271209609,-0.1151306778,0.0279608723,0.3960064352,0.0947105587,0.0307992771,0.1865343302,-0.0129996138,-0.0139087019,-0.3567220569,0.3218263984,-0.3141763806,0.3093022108,0.0804275125,-0.1705961674,-0.1404986829,0.1835096776,0.3468079865,-0.1027476788,0.5674605966,0.0961440355,0.4028461576,-0.3214431107,-0.4800122976,0.1687449068,0.1818929464,-0.2496030927,-0.5749029517,-0.3565101624,-0.1467401981,-0.3485468328,-0.4368962348,-0.0716112927,-0.108432062,0.2087444514,-0.3374734819,-0.0290727485,-0.2903195024,-0.0355794653,-0.0136548253,0.3278526366,0.1791389734,-0.1927564144,0.388684839,-0.0898672342,0.3580983579,-0.2984927595,0.1656743586,0.0580153205,-0.2435355037,-0.2330769151,0.2841905653,0.1397185475,-0.0865409076,0.0263522509,0.291955173,0.0951316431,-0.2298988104,0.1196050867,0.1383743584,0.5930799842,0.078654632,0.6068129539,0.170524165,0.3567244709,0.0604417101,0.009807582,0.2361737043,0.1867909133,0.08795771,0.1499005407,-0.1823438704,-0.2842317522,-0.1777065545,0.1698988229,0.0837589502,-0.074615106,0.2970551252,0.4641896486,-0.1873854101,0.4703545272,-0.0868343562,-0.0211216658,0.0057172333,0.0514900386,-0.0179805923,0.3525495529,-0.2764952779,0.0843371525,0.1849103272,0.1202506199,-0.0409353636,-0.5112404823,0.7779881954,-0.3419728875,-0.1739482284,-0.3053329885,0.3040142059,0.2109246552,-0.0706052631,-0.2278837711,0.028767949,-0.2351464182,-0.2879236042,-0.2090793699,-0.0509464219,-0.2891754508,0.1410698146,0.0335556567,-0.5960389972,0.1051091179,-0.237392962,0.5540138483,-0.3729892075,0.1248563081,-0.08957766,-0.2057229578,-0.1545006633,-0.0026389603,0.282823503,0.2545711398,-0.1710279435,-0.2035298645,-0.096007891,-0.2274876982,-0.1596027762,0.2408809215,-0.0294255428,0.2013362944,0.1114363521,0.0861898661,0.1921635121,-0.7362812161,-0.1907059103,0.039125368,0.2540967762,0.2841348052,-0.0689087138,0.3182002902,0.1190395057,-0.1914613396,-0.1278437227,0.4185301661,-0.0923452452,-0.2677813768,-0.1036554724,0.0313858092,0.1740298271,0.1357136071,0.4682197869,-0.1663037837,0.101213932,0.0368285961,-0.207832247,-0.3178376257,-0.2059400678,-0.0450444967,0.3274961114,-0.0367590077,0.0985477045,-0.2202026844,-0.5051130652,0.1482058764,-0.1031731889,0.1890636832,0.1133558303,-0.3304752707,-0.0975143537,0.4347307086,-0.0014060313,-0.1480617523,0.02245887,-0.1518128365,0.2043616027,-0.026348481,-0.1034101248,0.1999194324,0.0833177567,0.2868762016,0.4424619973,0.3334579468,-0.057746239,0.1427787244,0.1259192526,0.4548034668,-0.0623297058,-0.0239386689,0.0574869961,0.3577261269,-0.0979781523,0.2656302154,0.2493121773,0.5448932052,0.1440551728,-0.3306070268,0.1368448138,-0.0180839933,0.0213611238,-0.0567232221,-0.013299495,0.7422036529,0.0734468102,0.4684187174,0.1045975983,-0.1195817664,0.042006027,0.0761806145,-0.2893473208,-0.3267185688,0.1725258678,-0.2551752925,0.1510015428,-0.118014209,-0.0232432783,-0.0939322412,-0.1533479691,-0.1365054548,0.1462494731,0.2674663067,0.0887184739,-0.092035003,-0.0518967509,-0.234629631,0.2595597506,0.0696014464,0.0040451973,0.2768526971,-0.1759034842,0.2066102177,0.1795294583,0.4463194609,-0.1659672856,0.2943893373,0.2906155288,-0.0403938219,-0.2838439047,-0.4341475666,-0.3130204678,-0.0064297346,-0.1635503918,0.3149732947,0.1163866445,-0.3425177336,0.1712343991,-0.4348503947,-0.2230145335,-0.5114440322,0.1025417224,0.0085555222,0.084088102,0.0869053304,0.0160094947,0.5051114559,-0.2759878039,0.1010564044,-0.0918443426,-0.3128788173,0.1251651943,-0.0227441937,0.2510174811,-0.0505835451,0.2167745084,0.06350407,0.3050380349,0.0786861703,-0.5782645941,0.161975652,0.0128238061,0.1595941633,-0.009425465,-0.5442686677,0.1438509077,-0.3647788167,0.1384721994,-0.1547429115,-0.068238616,-0.0656606033,-0.1959404051,-0.1478056759,-0.1425560862,0.0340810306,-0.4164380133,-0.1596177816,0.3661914468,0.1965493858,-0.0222530216,0.1161043867,0.0819729641,-0.0072670886,0.015481689,-0.1338333189,0.0023299209,0.0536471866,0.3982647359,-0.1423727721,-0.020068042,-0.242856279,-0.4220930934,-0.0924972445,-0.2059858441,-0.3302642405,-0.7189273834,-0.1237746179,0.1407007575,0.0311928671,-0.0348916277,0.0768194944,-0.2370531559,0.1992862374,-0.0613395236,-0.4014445245,0.0908249021,-0.0623569861,-0.083991617,-0.4966912866,0.4393984675,0.2411254495,0.380070895,-0.0167775396,-0.3829095066,0.0695053786,0.0900348425,0.3596771359,0.0528986715,-0.1561409533,0.1884684861,-0.1911197901,-0.2710874379,0.2511743903,0.0950738192,-0.1881806254,-0.3442677855,0.04526788,-0.3981234729,-0.2436230481,-0.0015464936,0.2201658785,0.0994086787,0.0086950669,-0.0166082047,-0.5513292551,-0.0858264193,0.3991059065,0.2545210421,0.0767552853,-0.0948708206,-0.2385220826,-0.1900904328,-0.4015961587,0.3356547356,0.1114157438,0.0072918958,0.1262076199,-0.0031765508,0.0109875686,-0.1006437317,0.1530740708,-0.0662385151,-0.0091539836,0.0727168098,-0.5084453225,-0.0441977195,-0.2963524163,-0.2866689563,-0.0147866383,0.2140396386,0.5403144956,-0.3491865993,0.1352050304,0.214856267,0.0536219329,-0.2273655385,-0.0271938182,-0.1333905607,0.1154649928,0.0138833057,-0.0546660572,0.1017285213,0.1657826602,0.0941172391,0.1694841385,0.0033784793,-0.035575524,0.5560325384,0.5890778303,-0.0228788257,0.0049228589,0.0503083132,-0.0736029744,-0.4726648927,0.2202365398,0.2540068924,-0.0232181512,-0.846832335,-0.0817957819,0.1925516278,0.3330128193,-0.0464383401,0.0763191432,-0.2225629836,-0.2167150676,0.2232943028,-0.3131755888,0.5295398235,0.0977689177,0.2532327771,-0.2018689662,0.0246959012,0.3580567539,0.2071653157,-0.1761367619,0.0569431707,0.0899651796,-0.153889209,0.3100932539,0.4626712799,0.7989937663,-0.1242219806,-0.1367195398,-0.0001975645,-0.1432891339,0.2762157619,-0.4615858793,-0.0108004445,-0.1875384301,-0.2533771694,-0.0440357924,-0.1522351652,0.0488850996,-0.184582904,-0.0008906316,0.1672566235,0.2464726567,0.1864145845,0.1270349473,0.1973675787,0.0732199699,-0.3102345467,0.0292290747,0.0533295199,-0.2029115856,0.1816488802,-0.1132239923,0.0155738266,-0.313431859,0.1782194972,-0.2240444869,-0.047443036,-0.0893832147,0.2175898254,0.1645731032,-0.149340719,0.0840918496,0.3634406626,0.4752680659,-0.1508083194,0.0376059376,0.0593169332,-0.3069787621,0.1144080833,0.0553652346,-0.230626449,0.2586888969,-0.073738955,0.3419593275,0.1915057003,0.4165429473,-0.2845354676,0.1757487357,-0.1674331874,0.176913619,-0.0472232811,0.3336514235,0.1380684674,-0.1131135449,-0.0418058671,-0.0265045203,0.1103858352,-0.216594249,0.3255285025,-0.0434482619,-0.2242950797,0.0244910717,0.087303035,-0.065869838,-0.1756337434,-0.0384289399,-0.148648262,-0.2053458393,-0.01017607,-0.2198366076,-0.1991187036,-0.2939013839,0.3008575439,-0.2469031215,-0.4895503521,-0.2054042369,0.3679880798,0.4144978821,0.2888382077,-0.2223988175,-0.3008726835,-0.3259771168,0.2588813901,-0.2112446576,0.1525604576,-0.2292189449,0.607791543,-0.1435199082,0.2965234518,-0.1856720597,-0.2812434137,-0.2371106893,-0.0588165,-0.4407038093,-0.3586856723,-0.0555451214,-0.0189077575,-0.1065518707,0.2003398687,-0.2286507934,-0.0167121291,-0.2466875613,0.1922568679,0.0466640666,0.4406832755,-0.1525684893,0.3062048554,0.0616215207,-0.380271107,0.0805014968,0.1096490771,-0.0679831356,0.2895841002,0.2054458857,0.1984710246,0.0933200195,-0.0518018603,0.3235302866,0.353048116,0.4071328342,0.3210071027,-0.3119117022,-0.0700864717,0.1335509717,0.2069118619,0.1380882114,-0.1105321869,0.4389410317,0.2935257554,-0.0526559055,0.1060616821,0.3521795273,0.1685093343,0.09604837,0.1776010394,-0.164395079,0.047221683,-0.0686861873,-0.0459914431,0.2625201046,-0.1437592506,0.2043302655,0.4219452441,0.320017606,0.262047857,0.163194567,0.0780930445,0.2289646864,0.5206672549,0.2069575489,-0.0092780637,-0.0111818621,0.1314433217,0.4788704515,-0.2426757663,0.4709958732,-0.0607688203,-0.2056522518,0.1885233223,-0.4297228456,0.2462845445,0.3417514563,0.0191679392,-0.1782657653,-0.2274445146,0.3287647963,0.3067310154,-0.0350281186,-0.0914730951,-0.0242021065,-0.1226980239,-0.3589450121,0.2241210192,0.1636246741,-0.1052354053,-0.1486037076,-0.0799847543,-0.1974096745,0.091148518,0.038368728,-0.4937095046,-0.1336367428,0.0289222896,0.0179260243,0.336360693,-0.1937909275,0.3053176403,0.7154999375,-0.14009884,-0.113074474,0.4139331579,0.0292661116,-0.1019619107,0.3565452099,0.2964403033,-0.1349702179,-0.4047288597,0.2447700799,-0.0138267148,0.0196942464,0.006577143,0.1163655892,0.0413885228,0.2430705428,0.2479826808,0.0298962276,-0.1251018494,0.0728876218,-0.3502909541,0.3562813997,-0.4247967005,0.1356851906,0.3046932817,0.0273336917,-0.0798519775,0.0281216539,-0.0415624939,-0.4593973458,0.0430443026,-0.3000650406,-0.1201367453,-0.0400581509,-0.0092174187,-0.1653909236,0.2681224942,0.3173994124,0.2378945202,-0.3654517829,-0.1649194062,-0.4199304581,0.0316959396,0.1823453456,-0.3640840352,-0.0365194529,0.2397165,-0.0328222476,0.5731806755,-0.0296350848,0.3089938462,0.1334939003,0.0964872763,-0.4332215786,0.192015186,-0.4717691839,-0.0964951292,0.1334383488,0.055418577,-0.2559971511,0.0372667722,-0.1144233122,0.1294708848,-0.1716092825,-0.0825998336,0.5387237668,0.3669650257,0.1494873166,0.4115780294,0.1393589824,0.1398129612,0.094468981,-0.2344833761,0.1178086102,0.0759973601,-0.1502671689,0.3561973274,-0.2179719508,0.0830599219,-0.0056158514,0.2488784194,0.4969787002,-0.5032250285,-0.0846447721,-0.0199216437,-0.2355301529,0.08058016,0.1499191672,0.5444469452,0.1967546344,0.2469244301,0.1093314663,-0.3478524387,0.4284908473,-0.2423878908,0.0558310933,-0.2117149383,0.0536825098,0.3772603273,-0.0949614346,-0.5903131962,-0.1176030561,0.1731557548,-0.1132120788,-0.1280548871,0.2338012755,-0.1917490661,0.0629500374,-0.1018352285,0.124675177,-0.04514689,0.052667208,-0.1917762905,-0.2176475525]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2308","title":"Add COCO evaluation metrics","comments":"Great!\r\n\r\nHere's a notebook that illustrates how I'm using `CocoEvaluator`: https:\/\/drive.google.com\/file\/d\/1VV92IlaUiuPOORXULIuAdtNbBWCTCnaj\/view?usp=sharing\r\n\r\nThe evaluation is near the end of the notebook.\r\n\r\n","body":"I'm currently working on adding Facebook AI's DETR model (end-to-end object detection with Transformers) to HuggingFace Transformers. The model is working fine, but regarding evaluation, I'm currently relying on external `CocoEvaluator` and `PanopticEvaluator` objects which are defined in the original repository ([here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py#L22) and [here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/panoptic_eval.py#L13) respectively). \r\n\r\nRunning these in a notebook gives you nice summaries like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/48327001\/116878842-326f0680-ac20-11eb-9061-d6da02193694.png)\r\n\r\nIt would be great if we could import these metrics from the Datasets library, something like this:\r\n\r\n```\r\nimport datasets\r\n\r\nmetric = datasets.load_metric('coco')\r\n\r\nfor model_input, gold_references in evaluation_dataset:\r\n model_predictions = model(model_inputs)\r\n metric.add_batch(predictions=model_predictions, references=gold_references)\r\n\r\nfinal_score = metric.compute()\r\n```\r\n\r\nI think this would be great for object detection and semantic\/panoptic segmentation in general, not just for DETR. Reproducing results of object detection papers would be way easier.\r\n\r\nHowever, object detection and panoptic segmentation evaluation is a bit more complex than accuracy (it's more like a summary of metrics at different thresholds rather than a single one). I'm not sure how to proceed here, but happy to help making this possible.\r\n\r\n\r\n\r\n","comment_length":20,"text":"Add COCO evaluation metrics\nI'm currently working on adding Facebook AI's DETR model (end-to-end object detection with Transformers) to HuggingFace Transformers. The model is working fine, but regarding evaluation, I'm currently relying on external `CocoEvaluator` and `PanopticEvaluator` objects which are defined in the original repository ([here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py#L22) and [here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/panoptic_eval.py#L13) respectively). \r\n\r\nRunning these in a notebook gives you nice summaries like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/48327001\/116878842-326f0680-ac20-11eb-9061-d6da02193694.png)\r\n\r\nIt would be great if we could import these metrics from the Datasets library, something like this:\r\n\r\n```\r\nimport datasets\r\n\r\nmetric = datasets.load_metric('coco')\r\n\r\nfor model_input, gold_references in evaluation_dataset:\r\n model_predictions = model(model_inputs)\r\n metric.add_batch(predictions=model_predictions, references=gold_references)\r\n\r\nfinal_score = metric.compute()\r\n```\r\n\r\nI think this would be great for object detection and semantic\/panoptic segmentation in general, not just for DETR. Reproducing results of object detection papers would be way easier.\r\n\r\nHowever, object detection and panoptic segmentation evaluation is a bit more complex than accuracy (it's more like a summary of metrics at different thresholds rather than a single one). I'm not sure how to proceed here, but happy to help making this possible.\r\n\r\n\r\n\r\n\nGreat!\r\n\r\nHere's a notebook that illustrates how I'm using `CocoEvaluator`: https:\/\/drive.google.com\/file\/d\/1VV92IlaUiuPOORXULIuAdtNbBWCTCnaj\/view?usp=sharing\r\n\r\nThe evaluation is near the end of the notebook.\r\n\r\n","embeddings":[-0.2997457683,-0.2173083425,-0.0662911907,0.0201683957,0.1273829192,-0.1438953131,0.0299044643,-0.1443723142,-0.2041875869,0.134513706,-0.6726332307,0.1221146658,-0.1353891045,0.1291405559,-0.2434386015,-0.0863759667,-0.0150743863,-0.0097319195,-0.2172834873,0.1266033947,0.0592169836,0.1436478943,0.2437028587,0.1383200288,-0.2407882363,-0.1455112398,-0.2821148336,-0.3128210008,-0.152200982,-0.3235957623,0.2970020175,0.4498107135,0.195288673,0.3987880647,-0.000123575,0.1599732488,0.0171702933,0.0391538106,-0.0614747815,-0.0754960254,0.0772497281,-0.3173386157,0.2860774696,-0.1607625932,0.1821161509,-0.0290837213,-0.1013191044,0.0469154418,0.4056439698,0.100031741,0.050513804,0.1995001435,0.0012993686,-0.0145343486,-0.3424551785,0.3313701451,-0.3202508688,0.2981693149,0.083472468,-0.1853355765,-0.1613387167,0.199203074,0.3513002992,-0.093131505,0.568644166,0.0972017869,0.395811975,-0.3362832069,-0.4863196313,0.1939113289,0.1688250899,-0.2461128086,-0.5697387457,-0.3576303124,-0.16592668,-0.3212452233,-0.4314124286,-0.0633493662,-0.1420812011,0.2109694183,-0.3318120539,-0.0063184197,-0.2902599871,-0.0367846377,-0.0175289921,0.3369067013,0.1609748751,-0.2007407993,0.3905267715,-0.0829602629,0.3388564885,-0.2830789089,0.1836016178,0.0522666834,-0.2357368022,-0.239590764,0.2833957672,0.1345157325,-0.0850757062,0.0227248017,0.2996721864,0.1121045128,-0.2344082743,0.1268322021,0.1345121264,0.5766767263,0.060616944,0.5718221068,0.1867628545,0.3516945243,0.0615282729,0.0060395375,0.24470523,0.1921391636,0.0925656334,0.1304136068,-0.1720277667,-0.2655124962,-0.1750139445,0.1668262184,0.0971483886,-0.0691358447,0.2887948751,0.4693868756,-0.1852558553,0.4720817208,-0.0911772996,-0.0504620373,0.0083446885,0.0523504913,-0.0340195857,0.3280818462,-0.2931898832,0.0744849443,0.1807041764,0.1405318379,-0.0318055563,-0.5192122459,0.7789002657,-0.313687712,-0.1726381481,-0.2899074256,0.3224669695,0.2182406932,-0.0977120996,-0.2309614271,0.0274781007,-0.2426817864,-0.283683002,-0.2099500149,-0.0465671197,-0.2794578969,0.1206553802,0.0554584488,-0.5932693481,0.1008231863,-0.2202059776,0.572905004,-0.3765848875,0.1203671843,-0.0749170631,-0.1959534883,-0.1592183113,-0.012403016,0.2833305895,0.2463438511,-0.1743961424,-0.2132157981,-0.0936428979,-0.2362224162,-0.1588311791,0.2393621355,-0.0196399577,0.1804813296,0.1095646843,0.0924026668,0.185248211,-0.7222111821,-0.2145638615,0.048907049,0.2471457124,0.2894208729,-0.0734357759,0.3351033926,0.1267104,-0.189346835,-0.1360853314,0.4152053893,-0.0811438933,-0.2573272288,-0.1198692918,0.0354173258,0.1681862622,0.1451547891,0.433103621,-0.1665911525,0.1054945812,0.0371051319,-0.196060434,-0.320304662,-0.2076268047,-0.0496512391,0.3412632644,-0.0403319113,0.1081841141,-0.2308375984,-0.5144692063,0.1474981904,-0.1135462821,0.2028590292,0.122453101,-0.3373811841,-0.0957704484,0.4329608381,-0.0011986231,-0.1656338274,0.0466326438,-0.1520049572,0.2136487514,-0.0242753141,-0.1066238359,0.2016839087,0.0871169791,0.2898570895,0.4406019151,0.3309725523,-0.0673665702,0.1346359849,0.1199563444,0.4655914009,-0.0769188702,-0.0163238421,0.0534669049,0.3662233055,-0.0929570124,0.2508169115,0.2465990037,0.5548986197,0.1488536447,-0.3216001391,0.1336142272,-0.0017458077,0.0225114916,-0.0474049821,-0.0396965444,0.730409801,0.0812966004,0.463642776,0.1052093357,-0.1008718163,0.0464760251,0.0716416091,-0.3060565889,-0.3219819963,0.1666344404,-0.2491999418,0.1512070149,-0.0932580084,-0.0341611989,-0.0734885633,-0.1332815289,-0.1414821446,0.1587101221,0.2704225779,0.0616405271,-0.108017832,-0.0460405312,-0.2232921869,0.2471750826,0.0910435095,-0.0043952297,0.2587950528,-0.1916436404,0.1918939501,0.1756489724,0.4494325817,-0.1697312295,0.2736298144,0.2883634567,-0.0343804993,-0.2886705995,-0.4417698383,-0.328854233,-0.0141829066,-0.1479980499,0.3167216778,0.1225950569,-0.3476931453,0.1796490699,-0.436488688,-0.2220805287,-0.5205181837,0.1152007803,0.0132634863,0.0673807859,0.0899344906,0.0288928337,0.5286703706,-0.2699851096,0.1107042506,-0.1122965142,-0.3123542666,0.0993663445,0.0015263066,0.2481557727,-0.0478702746,0.2108082622,0.064555347,0.3195941746,0.0707352087,-0.5984122753,0.163905561,0.0002433452,0.16894871,-0.0078940392,-0.543690145,0.1415248364,-0.359223485,0.133017078,-0.180943653,-0.0755390897,-0.0621161759,-0.1881648302,-0.1498838812,-0.1533188075,0.0218634848,-0.407507062,-0.1856070459,0.3884068429,0.2024872005,-0.0216110572,0.1255498827,0.0721987411,0.0083553931,0.0168718901,-0.1339852065,-0.0112587996,0.0598986857,0.3984030187,-0.1504701227,-0.0178765357,-0.2467008978,-0.4240752459,-0.0942104906,-0.2257972658,-0.3218148947,-0.7311675549,-0.1432559639,0.1486822814,0.038619373,-0.0254496988,0.0705624372,-0.2284892648,0.1728915125,-0.0559927337,-0.3789964318,0.0764611885,-0.0758435726,-0.0935095847,-0.4966576099,0.4219006598,0.2515825033,0.3898653388,-0.0341858119,-0.392049551,0.0780420676,0.0776520148,0.3641143739,0.0493736491,-0.1420731992,0.1943820715,-0.1970599145,-0.2544766665,0.2512594759,0.0943183079,-0.2052671015,-0.325204432,0.037259791,-0.3865357339,-0.2463575751,0.004864688,0.2265776396,0.0744298995,0.0035523891,-0.016765926,-0.5261825919,-0.0947540104,0.4008866549,0.2383138388,0.0825293213,-0.0988602266,-0.2404682636,-0.1687589437,-0.4156934023,0.3411152363,0.0932508484,-0.0129253846,0.1346775144,-0.0368649177,-0.0030849124,-0.0987497121,0.1665324271,-0.0610169396,-0.0342471227,0.0618021749,-0.5149249434,-0.0635028109,-0.2830658257,-0.2945513129,-0.0157704242,0.211601451,0.5451014042,-0.3397910595,0.125939101,0.2373053879,0.0394671224,-0.2218977958,-0.0148762157,-0.1446876973,0.1181574464,-0.0024681343,-0.0502193533,0.0985896364,0.143421784,0.0863526538,0.1571967155,0.0158531591,-0.0381097384,0.5339030027,0.6092336774,-0.0233167801,0.0006790083,0.039184723,-0.0617796928,-0.4591567516,0.1990256011,0.2507909536,-0.0377940126,-0.8460482359,-0.0552419797,0.2072765827,0.3129008114,-0.0364866778,0.0903270096,-0.2212751806,-0.198105529,0.2404571027,-0.3323580325,0.5323820114,0.0944028273,0.242668733,-0.1755241752,0.0201539863,0.3630640805,0.193492949,-0.1754037291,0.0536112003,0.0945937634,-0.140398562,0.3290589154,0.4565214515,0.8081385493,-0.1172459051,-0.1190668195,-0.004075788,-0.1455143094,0.2800195217,-0.4670466781,-0.0107088545,-0.1926065832,-0.2586920857,-0.0441653058,-0.1439217329,0.0402363092,-0.1787968129,-0.0212660842,0.1624367386,0.2555801868,0.2075008601,0.1338519007,0.2047467232,0.0642844215,-0.3234063983,0.0317041911,0.0718120337,-0.1916462928,0.1762247086,-0.1123823598,0.0228485987,-0.3125870824,0.1763160229,-0.2226732969,-0.0464611761,-0.0862539858,0.2179839611,0.1812766343,-0.1283545941,0.0708476827,0.3582497239,0.4691965282,-0.1416616887,0.0318295993,0.068314895,-0.2951416373,0.1271838248,0.0667114928,-0.243466109,0.2872308195,-0.0701186284,0.3233158588,0.2023533434,0.4158208072,-0.2841750979,0.1560437828,-0.1622921377,0.1703691185,-0.0544659719,0.3462355733,0.1270444989,-0.1362367719,-0.0584033504,-0.0049250461,0.1038166434,-0.2249019891,0.3124018013,-0.0346105397,-0.2160986811,0.0163747892,0.0837359354,-0.0703182966,-0.1638978422,-0.0417792015,-0.1517426968,-0.1967330426,-0.0304503366,-0.2186263055,-0.1701902896,-0.2886288762,0.3089745939,-0.2554734945,-0.5010094643,-0.1971584707,0.3870409429,0.4179735184,0.2863822281,-0.2421090454,-0.2874953747,-0.3281961977,0.2762742639,-0.2103974372,0.1733352989,-0.2154820561,0.5901417732,-0.1272023469,0.2967890203,-0.2152383029,-0.2923658788,-0.2511003613,-0.0724317804,-0.4433409572,-0.368193388,-0.0441001505,-0.027109256,-0.0854171142,0.2223088294,-0.2309468836,-0.0428764671,-0.2431994379,0.1782479733,0.031459678,0.4418120086,-0.158716768,0.3344663382,0.0551942959,-0.3866641521,0.0841009766,0.1035967097,-0.0637550652,0.2835543454,0.1996849626,0.1840124279,0.1139559224,-0.055301778,0.3162204623,0.3359111249,0.4048947692,0.3138521016,-0.3291019201,-0.0748057216,0.1486317962,0.2220020592,0.1404710114,-0.1018305048,0.43600896,0.2986540496,-0.0777134746,0.0970304385,0.3574431241,0.1552157104,0.1049595624,0.1615591347,-0.1594380736,0.0691542849,-0.0715450495,-0.0564126186,0.2643130124,-0.1573741883,0.2371173352,0.4199499786,0.3266837001,0.2741064131,0.167464748,0.0666891187,0.2360154688,0.5075060725,0.1998953819,-0.0234128181,-0.0090770898,0.1328239292,0.4826538563,-0.2623148859,0.4836318493,-0.0574854203,-0.1937593371,0.2043046653,-0.4279995263,0.2465712577,0.355962038,0.0289880782,-0.1492527872,-0.2426814735,0.3121583462,0.3237943649,-0.0378412679,-0.1072400287,-0.0338828638,-0.1182374954,-0.3678840399,0.2005825043,0.1597318351,-0.1120311841,-0.144638598,-0.0882376805,-0.197604388,0.0771286637,0.0307872314,-0.4916616082,-0.1184270009,0.013035018,0.0191560742,0.3470639288,-0.1806201786,0.2940875292,0.7280279398,-0.1368328929,-0.1232587695,0.4176488817,0.0290640611,-0.1232912093,0.3599896133,0.3029301763,-0.1313479692,-0.4097949564,0.2434921116,-0.0126560861,0.0090096202,-0.0021371448,0.1232979968,0.0344431102,0.2410011888,0.2508681118,0.0535558537,-0.1434519738,0.0682976544,-0.3407967687,0.3618963659,-0.4263160825,0.1210775673,0.300806433,0.0308796056,-0.0831633434,0.0318329893,-0.058763776,-0.4580427706,0.0310644582,-0.304872483,-0.108624436,-0.0532159656,0.007626717,-0.1413908601,0.285179615,0.3214081228,0.2369333208,-0.3483337462,-0.1536585838,-0.4194567204,0.0428625196,0.1830455065,-0.3505311012,-0.0381963067,0.2528027296,-0.0319824368,0.5803541541,-0.0283607878,0.3007442057,0.1152474955,0.0816766173,-0.4270045757,0.1995345503,-0.4688553512,-0.1081947237,0.1317014843,0.0474087,-0.2442966104,0.0516680889,-0.0914225504,0.1239640415,-0.1751901209,-0.0596671924,0.5259096622,0.3599401414,0.1547087431,0.3691020012,0.1324088424,0.1214027926,0.0797875896,-0.2308139652,0.1155482158,0.0566168614,-0.1507091671,0.3563115895,-0.1916333139,0.053953249,-0.0148481783,0.2521485984,0.5013907552,-0.4979832768,-0.0838880241,-0.029899396,-0.2199155837,0.0648839697,0.1440890133,0.5455387831,0.1903114319,0.2297285646,0.1160851568,-0.3590548635,0.411984086,-0.2448017299,0.0657234788,-0.2228418589,0.0472108349,0.3749010265,-0.0942502841,-0.574844718,-0.1309122592,0.1755871624,-0.1069520116,-0.1329507828,0.2344389856,-0.188247472,0.0649432316,-0.1123963296,0.1239950135,-0.045083072,0.0527138598,-0.1798617542,-0.2336570174]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2308","title":"Add COCO evaluation metrics","comments":"I went through the code you've [mentioned](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py) and I think there are 2 options on how we can go ahead:\r\n\r\n1) Implement how DETR people have done this (they're relying very heavily on the official implementation and they're focussing on torch dataset here. I feel ours should be something generic instead of pytorch specific.\r\n2) Do this [implementation](https:\/\/github.com\/cocodataset\/cocoapi\/blob\/ed842bffd41f6ff38707c4f0968d2cfd91088688\/PythonAPI\/pycocoEvalDemo.ipynb) where user can convert its output and ground truth annotation to pre-defined format and then feed it into our function to calculate metrics (looks very similar to you wanted above)\r\n\r\nIn my opinion, 2nd option looks very clean but I'm still figuring out how's it transforming the box co-ordinates of `coco_gt` which you've passed to `CocoEvaluator` (ground truth for evaluation). Since your model output was already converted to COCO api, I faced little problems there.","body":"I'm currently working on adding Facebook AI's DETR model (end-to-end object detection with Transformers) to HuggingFace Transformers. The model is working fine, but regarding evaluation, I'm currently relying on external `CocoEvaluator` and `PanopticEvaluator` objects which are defined in the original repository ([here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py#L22) and [here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/panoptic_eval.py#L13) respectively). \r\n\r\nRunning these in a notebook gives you nice summaries like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/48327001\/116878842-326f0680-ac20-11eb-9061-d6da02193694.png)\r\n\r\nIt would be great if we could import these metrics from the Datasets library, something like this:\r\n\r\n```\r\nimport datasets\r\n\r\nmetric = datasets.load_metric('coco')\r\n\r\nfor model_input, gold_references in evaluation_dataset:\r\n model_predictions = model(model_inputs)\r\n metric.add_batch(predictions=model_predictions, references=gold_references)\r\n\r\nfinal_score = metric.compute()\r\n```\r\n\r\nI think this would be great for object detection and semantic\/panoptic segmentation in general, not just for DETR. Reproducing results of object detection papers would be way easier.\r\n\r\nHowever, object detection and panoptic segmentation evaluation is a bit more complex than accuracy (it's more like a summary of metrics at different thresholds rather than a single one). I'm not sure how to proceed here, but happy to help making this possible.\r\n\r\n\r\n\r\n","comment_length":133,"text":"Add COCO evaluation metrics\nI'm currently working on adding Facebook AI's DETR model (end-to-end object detection with Transformers) to HuggingFace Transformers. The model is working fine, but regarding evaluation, I'm currently relying on external `CocoEvaluator` and `PanopticEvaluator` objects which are defined in the original repository ([here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py#L22) and [here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/panoptic_eval.py#L13) respectively). \r\n\r\nRunning these in a notebook gives you nice summaries like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/48327001\/116878842-326f0680-ac20-11eb-9061-d6da02193694.png)\r\n\r\nIt would be great if we could import these metrics from the Datasets library, something like this:\r\n\r\n```\r\nimport datasets\r\n\r\nmetric = datasets.load_metric('coco')\r\n\r\nfor model_input, gold_references in evaluation_dataset:\r\n model_predictions = model(model_inputs)\r\n metric.add_batch(predictions=model_predictions, references=gold_references)\r\n\r\nfinal_score = metric.compute()\r\n```\r\n\r\nI think this would be great for object detection and semantic\/panoptic segmentation in general, not just for DETR. Reproducing results of object detection papers would be way easier.\r\n\r\nHowever, object detection and panoptic segmentation evaluation is a bit more complex than accuracy (it's more like a summary of metrics at different thresholds rather than a single one). I'm not sure how to proceed here, but happy to help making this possible.\r\n\r\n\r\n\r\n\nI went through the code you've [mentioned](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py) and I think there are 2 options on how we can go ahead:\r\n\r\n1) Implement how DETR people have done this (they're relying very heavily on the official implementation and they're focussing on torch dataset here. I feel ours should be something generic instead of pytorch specific.\r\n2) Do this [implementation](https:\/\/github.com\/cocodataset\/cocoapi\/blob\/ed842bffd41f6ff38707c4f0968d2cfd91088688\/PythonAPI\/pycocoEvalDemo.ipynb) where user can convert its output and ground truth annotation to pre-defined format and then feed it into our function to calculate metrics (looks very similar to you wanted above)\r\n\r\nIn my opinion, 2nd option looks very clean but I'm still figuring out how's it transforming the box co-ordinates of `coco_gt` which you've passed to `CocoEvaluator` (ground truth for evaluation). Since your model output was already converted to COCO api, I faced little problems there.","embeddings":[-0.2762456536,-0.2085913718,-0.0654032603,0.0466305688,0.1744097322,-0.1518914849,0.0246486589,-0.1304698735,-0.1931971908,0.1273570806,-0.6784852743,0.121309571,-0.1210910976,0.1138263941,-0.2690066397,-0.0950239822,-0.0252741873,-0.01117709,-0.2147166431,0.1027091295,0.0485205166,0.178263396,0.2275513709,0.1101235598,-0.254538089,-0.1438167244,-0.2657630742,-0.2837206721,-0.1923244447,-0.3163909316,0.3075261712,0.4554100037,0.2024868876,0.4298277497,-0.0001223113,0.1656180918,0.0208981652,0.0250817984,-0.0517429598,-0.0965127945,0.0708127618,-0.3158518374,0.2736356556,-0.1634330601,0.1322087049,-0.030235678,-0.1248590201,0.0449080653,0.4284770191,0.1474260241,0.0711562485,0.1883551627,0.0393480249,-0.0047205747,-0.3366113007,0.3078396916,-0.3132876158,0.3043533266,0.059667293,-0.1810681075,-0.1572528481,0.223778367,0.3409006596,-0.0695617199,0.5765433311,0.1170461252,0.3985585272,-0.3481795788,-0.4635713398,0.1570440233,0.1507800668,-0.2379888147,-0.5474994183,-0.3472571671,-0.1965019703,-0.3565271199,-0.3971282542,-0.0950896218,-0.1564535499,0.1816256791,-0.3316313028,0.0121770641,-0.3044273555,-0.0409061536,0.0005170868,0.3494045436,0.1487829834,-0.1873195916,0.4057869911,-0.0739512742,0.3195967674,-0.2628585696,0.2078426033,0.0378848389,-0.2627312541,-0.2232255489,0.2751401961,0.1383701563,-0.07271377,0.0197973493,0.277000308,0.1162491813,-0.2116607279,0.116253823,0.1057365835,0.6059837341,0.0628726855,0.5722205639,0.1813307703,0.3665733635,0.0421056487,-0.0075578792,0.2369519323,0.1503079236,0.1127964556,0.1642229259,-0.1079887077,-0.2825087905,-0.1727300137,0.1893318892,0.1075053364,-0.0790875033,0.3042195141,0.4782568216,-0.1811585277,0.4461401999,-0.0655577332,-0.0462577306,-0.018244978,0.0664228797,-0.0574939139,0.3409990072,-0.3272207677,0.0691410154,0.1938776076,0.1533528864,-0.0313309766,-0.4772887826,0.7550008893,-0.3323637247,-0.1679587811,-0.302188158,0.3394156098,0.1996418089,-0.09726087,-0.191221267,0.0545454659,-0.2480310798,-0.2544415593,-0.2019887716,-0.0508468933,-0.2689056396,0.1182743087,0.0738486573,-0.5843091011,0.1008130983,-0.2262386978,0.5677810907,-0.3436725438,0.122991465,-0.059731476,-0.2086481154,-0.1660012752,-0.0124743292,0.3085112274,0.2649511993,-0.1796625555,-0.1830184162,-0.08096762,-0.2382341176,-0.1266950816,0.2226281017,-0.0105148107,0.2006147951,0.0851476267,0.1084154248,0.1714395881,-0.7114329338,-0.2342544049,0.0404349901,0.2236281931,0.2831639946,-0.0782680884,0.3078074157,0.1145276651,-0.1691482514,-0.1307482719,0.4090504944,-0.0740684122,-0.2533506751,-0.1601086408,0.0621848404,0.1803666949,0.1431153119,0.4445335865,-0.1584167629,0.0657703355,0.0254786722,-0.1756358147,-0.3219895959,-0.1698574573,-0.0325578898,0.3594143987,-0.023432564,0.1073376313,-0.2093330622,-0.5224545598,0.1673995852,-0.1160116866,0.2480333596,0.0971881971,-0.3190678954,-0.1431769878,0.4123532474,-0.049163118,-0.1736114025,0.069818534,-0.1345931441,0.2121600658,-0.0539133549,-0.0922610238,0.2222616971,0.0908731595,0.2821025252,0.4027760029,0.3636613786,-0.1111014411,0.129260987,0.1170272678,0.4381550252,0.0024813227,-0.0212981626,0.0223128274,0.3827012181,-0.0947424024,0.2323689014,0.2379073203,0.4944932461,0.1189072803,-0.3212482631,0.1284971535,0.0009626867,0.0421741977,-0.0464833528,-0.0077398974,0.7312797904,0.0620470941,0.4585793018,0.0933523029,-0.1022493839,0.0576270409,0.0553310104,-0.2785558403,-0.3295013309,0.1663157344,-0.2630715966,0.1256216168,-0.1224567592,-0.0244950354,-0.051908657,-0.1257094443,-0.1292593032,0.1346466094,0.281737864,0.0859710351,-0.0932163969,-0.0398116745,-0.2153113186,0.2656779289,0.0848730206,-0.0105626592,0.2854541838,-0.1838683784,0.1853946894,0.1745005697,0.4714135528,-0.1456243992,0.2780250311,0.2813168466,-0.0237598997,-0.3009242415,-0.4294913709,-0.2931693494,0.0122506646,-0.1763392687,0.3256577849,0.1141820624,-0.3490326405,0.1567971557,-0.4588406384,-0.2433834523,-0.5207465887,0.1033495888,0.0190505199,0.0706163645,0.127628088,0.0369858518,0.5012657046,-0.2499660105,0.0926333591,-0.120343402,-0.2922961414,0.0737061277,0.0129695507,0.2553323507,-0.0234923922,0.2086292505,0.0735825971,0.3236071169,0.0638396591,-0.5853328109,0.1565502584,-0.049568966,0.1875982732,-0.0077075786,-0.5228638649,0.1215294227,-0.3543279171,0.1721247137,-0.2101464123,-0.0749879554,-0.0530057326,-0.1912761033,-0.1462313235,-0.1577598602,0.0053714877,-0.4186987579,-0.1929976493,0.3601029813,0.2147980928,-0.0213449579,0.149769038,0.0865374133,0.0260949209,0.0550642833,-0.1355027854,-0.0529952385,0.0406923369,0.4118338525,-0.1769706011,-0.0537597276,-0.2154047787,-0.4016530812,-0.0702753589,-0.2474472076,-0.3203711212,-0.7439361811,-0.1434044391,0.1265176237,0.018342752,-0.0163928438,0.0766172707,-0.2344217002,0.1522072256,-0.0657699034,-0.4089482725,0.1194923073,-0.0777283534,-0.0890451148,-0.5065962672,0.4495178461,0.2294935733,0.3933584094,0.0016589224,-0.3807325065,0.1166252047,0.0910189077,0.3588932753,0.035747353,-0.1737159342,0.2155748159,-0.1794087738,-0.2432072461,0.2479097694,0.0934002101,-0.1802215725,-0.3338061273,0.0302223694,-0.399254024,-0.2653988004,0.0268932581,0.2293666154,0.0543949865,0.0034985505,-0.0015959748,-0.5473843217,-0.0930737406,0.3901608288,0.2279331833,0.0608800054,-0.0934580415,-0.2330212444,-0.1388532221,-0.4689036608,0.3694581389,0.0699460134,-0.0054084756,0.1233939752,-0.022870047,0.0018488296,-0.1014420837,0.2154691666,-0.0550017729,-0.0120993573,0.0832607523,-0.4966391325,-0.0782086551,-0.287144959,-0.2765392959,-0.0057883696,0.2186926454,0.533506453,-0.3527657986,0.1291580349,0.2480496913,0.0165713876,-0.2094554901,-0.0192452986,-0.1649112999,0.0822109208,-0.0150646707,-0.0420377925,0.1270928532,0.164976418,0.0591276698,0.1609514207,0.0661304295,-0.0418822616,0.5437110066,0.59553653,0.0093810745,0.0182909183,0.0254678056,-0.0320287049,-0.4661464989,0.2059058994,0.2957652509,-0.0609991252,-0.8166982532,-0.0561887659,0.1940588802,0.29348737,-0.0195895508,0.0721370801,-0.2421950996,-0.1871278584,0.2542257607,-0.3209154606,0.5390093327,0.0880365521,0.2292512208,-0.1826346666,0.0141430488,0.3457551599,0.164149791,-0.1617700309,0.0721413866,0.1148476005,-0.1693145037,0.339941591,0.4461795092,0.8220878839,-0.112610966,-0.1548721641,0.0120479055,-0.1374432892,0.2894862592,-0.4529245496,-0.0084894663,-0.1882472187,-0.2486756146,-0.0401295349,-0.140699327,0.0263495948,-0.1941782683,-0.0595137998,0.1681930423,0.2482341826,0.1855237186,0.1249211803,0.2243400812,0.0713229254,-0.3313355148,0.0026446516,0.0945768282,-0.2098754942,0.2094292045,-0.1125850976,-0.0089000026,-0.3395108879,0.1401900351,-0.204075262,-0.0296323858,-0.12020237,0.2145760357,0.1627174318,-0.1346902251,0.0599305555,0.3501694798,0.4952365756,-0.1618831307,0.0012825513,0.0753727928,-0.2775743008,0.1320809722,0.065620847,-0.2626317739,0.2951783538,-0.0912602842,0.3007409275,0.1701383144,0.4086494744,-0.3036220372,0.1506336033,-0.1786529273,0.1701086909,-0.0626603812,0.3288612366,0.1114547774,-0.1378707588,-0.0825314671,0.0148895541,0.1026236564,-0.2490640432,0.3069474399,-0.0335751921,-0.2077295631,0.0239762757,0.0965177938,-0.0818433538,-0.1717668474,-0.0118890805,-0.1071853712,-0.1967215985,-0.0395887978,-0.2562928796,-0.1372626871,-0.3169880211,0.3116516471,-0.2514774203,-0.4688404799,-0.1731033623,0.4210797846,0.3939843774,0.3023222983,-0.1934247464,-0.3022736311,-0.3581427932,0.2372599691,-0.2229090631,0.1673686504,-0.2183483392,0.6170164943,-0.0994330496,0.3165110052,-0.2374030799,-0.2778232396,-0.2744908929,-0.0647457689,-0.4368366897,-0.3990370631,-0.0308971293,-0.018898258,-0.0737770945,0.2288740277,-0.2616608441,-0.0601002723,-0.2391661257,0.1694385707,0.0471810848,0.4099207222,-0.1718661487,0.3195155561,0.0353315994,-0.3725660145,0.0807255283,0.1193138659,-0.0700906068,0.2858245671,0.2136671841,0.1655386835,0.1410432756,-0.1001109406,0.3156053424,0.3201066554,0.424900949,0.2872233391,-0.331237644,-0.0810121372,0.1463563889,0.2061989158,0.1439628303,-0.0891077369,0.4217730761,0.2530753911,-0.1003192812,0.1175420433,0.3994005024,0.1551519781,0.0832899287,0.1619013846,-0.1746638417,0.0905901343,-0.0762656629,-0.0285780448,0.2756691277,-0.1571265161,0.2356152236,0.4269887209,0.3106445968,0.2569349706,0.1443135589,0.1029462814,0.2596435547,0.4861937761,0.210677892,-0.0611570217,-0.0575119257,0.1315545142,0.5030164123,-0.2669906318,0.4692579508,-0.027661128,-0.1999390423,0.2299840003,-0.4348881543,0.2262986898,0.3229152262,-0.005621355,-0.1496713012,-0.2351872176,0.2930427492,0.313059032,-0.0489825122,-0.0969538167,-0.0488335676,-0.1319038868,-0.3544022143,0.214426294,0.137099117,-0.1006890461,-0.1435352415,-0.0954158753,-0.2012464553,0.0558300838,0.050099168,-0.4797582328,-0.1086416543,0.0039207037,-0.0289103873,0.3397587836,-0.2136209756,0.27994138,0.7104994059,-0.1439491808,-0.0775690004,0.4368622303,0.0029714939,-0.1155458838,0.3416868448,0.3122869432,-0.0868648067,-0.4118300676,0.23394382,0.0113726119,-0.0028079466,0.0047154492,0.1632648706,0.0655135214,0.2462731302,0.2556034327,0.0660330355,-0.1703211814,0.0462552942,-0.3037078083,0.3906089664,-0.4059385955,0.1243999228,0.2919084132,0.0214647446,-0.1243564785,0.0087587144,-0.1108527854,-0.4503015578,0.0219856892,-0.2958690524,-0.0949489623,-0.0922806635,0.0145536,-0.1507119089,0.2853779495,0.2920782268,0.232316494,-0.4331855178,-0.160640046,-0.4389683008,0.0494956225,0.1921201199,-0.3307118118,-0.0261436477,0.2662698627,-0.0240799729,0.5436844826,-0.0163942128,0.2819663584,0.1184482723,0.0744233951,-0.4420082569,0.2170403153,-0.4874477088,-0.1042252034,0.1522231847,0.0420709848,-0.2175891995,0.0488942452,-0.0720038861,0.1187461987,-0.1851815432,-0.0586597733,0.5239102244,0.3720209897,0.1598097086,0.3868485987,0.113906607,0.1370787472,0.0372333378,-0.2163701802,0.0815648735,0.047448054,-0.1319821924,0.4094475508,-0.1868492961,0.0475280993,-0.0603240505,0.2450965047,0.4844947159,-0.4972520769,-0.0790045634,0.0072275829,-0.212162599,0.0728155524,0.137725383,0.5201488733,0.1925054044,0.2557738125,0.0572960041,-0.3900304735,0.4627178609,-0.2369894385,0.0301202536,-0.2028691322,0.0776493326,0.377949357,-0.1164872572,-0.6244449615,-0.0953428969,0.1603278667,-0.087628603,-0.1636816114,0.2436120659,-0.1967924535,0.0784676895,-0.1009132639,0.128602922,-0.029001547,0.0615757182,-0.1632719785,-0.2158156335]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2308","title":"Add COCO evaluation metrics","comments":"Ok, thanks for the update.\r\n\r\nIndeed, the metrics API of Datasets is framework agnostic, so we can't rely on a PyTorch-only implementation.\r\n\r\n[This file](https:\/\/github.com\/cocodataset\/cocoapi\/blob\/ed842bffd41f6ff38707c4f0968d2cfd91088688\/PythonAPI\/pycocotools\/cocoeval.py) is probably want we need to implement.\r\n\r\n","body":"I'm currently working on adding Facebook AI's DETR model (end-to-end object detection with Transformers) to HuggingFace Transformers. The model is working fine, but regarding evaluation, I'm currently relying on external `CocoEvaluator` and `PanopticEvaluator` objects which are defined in the original repository ([here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py#L22) and [here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/panoptic_eval.py#L13) respectively). \r\n\r\nRunning these in a notebook gives you nice summaries like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/48327001\/116878842-326f0680-ac20-11eb-9061-d6da02193694.png)\r\n\r\nIt would be great if we could import these metrics from the Datasets library, something like this:\r\n\r\n```\r\nimport datasets\r\n\r\nmetric = datasets.load_metric('coco')\r\n\r\nfor model_input, gold_references in evaluation_dataset:\r\n model_predictions = model(model_inputs)\r\n metric.add_batch(predictions=model_predictions, references=gold_references)\r\n\r\nfinal_score = metric.compute()\r\n```\r\n\r\nI think this would be great for object detection and semantic\/panoptic segmentation in general, not just for DETR. Reproducing results of object detection papers would be way easier.\r\n\r\nHowever, object detection and panoptic segmentation evaluation is a bit more complex than accuracy (it's more like a summary of metrics at different thresholds rather than a single one). I'm not sure how to proceed here, but happy to help making this possible.\r\n\r\n\r\n\r\n","comment_length":31,"text":"Add COCO evaluation metrics\nI'm currently working on adding Facebook AI's DETR model (end-to-end object detection with Transformers) to HuggingFace Transformers. The model is working fine, but regarding evaluation, I'm currently relying on external `CocoEvaluator` and `PanopticEvaluator` objects which are defined in the original repository ([here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/coco_eval.py#L22) and [here](https:\/\/github.com\/facebookresearch\/detr\/blob\/a54b77800eb8e64e3ad0d8237789fcbf2f8350c5\/datasets\/panoptic_eval.py#L13) respectively). \r\n\r\nRunning these in a notebook gives you nice summaries like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/48327001\/116878842-326f0680-ac20-11eb-9061-d6da02193694.png)\r\n\r\nIt would be great if we could import these metrics from the Datasets library, something like this:\r\n\r\n```\r\nimport datasets\r\n\r\nmetric = datasets.load_metric('coco')\r\n\r\nfor model_input, gold_references in evaluation_dataset:\r\n model_predictions = model(model_inputs)\r\n metric.add_batch(predictions=model_predictions, references=gold_references)\r\n\r\nfinal_score = metric.compute()\r\n```\r\n\r\nI think this would be great for object detection and semantic\/panoptic segmentation in general, not just for DETR. Reproducing results of object detection papers would be way easier.\r\n\r\nHowever, object detection and panoptic segmentation evaluation is a bit more complex than accuracy (it's more like a summary of metrics at different thresholds rather than a single one). I'm not sure how to proceed here, but happy to help making this possible.\r\n\r\n\r\n\r\n\nOk, thanks for the update.\r\n\r\nIndeed, the metrics API of Datasets is framework agnostic, so we can't rely on a PyTorch-only implementation.\r\n\r\n[This file](https:\/\/github.com\/cocodataset\/cocoapi\/blob\/ed842bffd41f6ff38707c4f0968d2cfd91088688\/PythonAPI\/pycocotools\/cocoeval.py) is probably want we need to implement.\r\n\r\n","embeddings":[-0.2992970049,-0.2228718996,-0.0646218807,0.0151259722,0.1335908324,-0.1412784159,0.0264432374,-0.1363361776,-0.1954925209,0.1230499446,-0.6802905798,0.1259326339,-0.1547850221,0.1198722944,-0.2277210802,-0.0956800133,-0.0067188987,-0.0257028192,-0.206568256,0.1308602691,0.045062229,0.1559381485,0.2409396619,0.1375779659,-0.2349680364,-0.1268833131,-0.2791191638,-0.3148986101,-0.1528187543,-0.3324477971,0.3127663136,0.4507707655,0.2063035816,0.3868991137,-0.0001238944,0.1604953408,0.0293454695,0.0402950086,-0.0567350462,-0.0817935616,0.0817746967,-0.3192557096,0.2854306102,-0.1532307118,0.1690668762,-0.0341490284,-0.11895331,0.0350222848,0.4042483866,0.110640578,0.0524125993,0.1972407103,0.0121194068,-0.0126454048,-0.3311860859,0.3505337834,-0.3129050732,0.3000722826,0.0955421627,-0.1801473647,-0.1500790715,0.1911480427,0.3429262936,-0.0921719521,0.5646221042,0.094734095,0.3990193009,-0.3180494606,-0.4919623733,0.1922196746,0.1793985963,-0.2508137524,-0.5659953356,-0.3535729647,-0.1606238782,-0.3407167792,-0.4232775569,-0.0717138872,-0.1374007165,0.2017319947,-0.3214557171,-0.0216319673,-0.2929254472,-0.0346351229,-0.0127414381,0.3427121341,0.1697211266,-0.2110771686,0.3974739313,-0.0821206421,0.3321637213,-0.2659060657,0.1673607081,0.0565473512,-0.2368429601,-0.2462360114,0.2739785612,0.1305557489,-0.0841672197,0.0217711162,0.2886641026,0.1091312766,-0.2407827526,0.1194920912,0.1254581213,0.5836359262,0.0670502782,0.5859928131,0.1789158881,0.3602632582,0.0608827919,-0.0031642781,0.2547972798,0.1796337515,0.0882038474,0.1490274221,-0.1660265923,-0.2710515559,-0.1633441895,0.1672856361,0.1109902188,-0.0777895302,0.2944928706,0.4754862189,-0.1721894145,0.470359683,-0.0810871497,-0.0446160659,0.0116869276,0.038016744,-0.0310281925,0.3311302364,-0.2840276361,0.0755267441,0.1878756732,0.1341210902,-0.0454441309,-0.5317236781,0.7972844243,-0.3050076365,-0.1892516315,-0.2818904817,0.3368848562,0.2114264816,-0.1106370166,-0.2374489009,0.0274331123,-0.2493315488,-0.2812485099,-0.198466301,-0.0533035733,-0.2920435965,0.1029669642,0.051463183,-0.5960325599,0.1069889516,-0.2322064489,0.5729320645,-0.3790602386,0.1045930684,-0.0657642111,-0.2093566507,-0.1626275927,-0.007087301,0.2824583054,0.2480089068,-0.1690669358,-0.2179758549,-0.1003717855,-0.2434242368,-0.1601533294,0.2256184071,-0.025373321,0.1645831317,0.1001148745,0.0776345432,0.1797301024,-0.7268228531,-0.2085731477,0.0272011049,0.2609668076,0.2839909494,-0.0744253024,0.3155295551,0.1378124654,-0.176649332,-0.1323247701,0.4194304943,-0.0769564807,-0.2528687119,-0.1330254376,0.04196826,0.1788371056,0.1517906487,0.4420607388,-0.1673374772,0.1027848423,0.0223796275,-0.1959409565,-0.3134689927,-0.1980535686,-0.0567649007,0.3496710062,-0.0370448418,0.1241204813,-0.2290176004,-0.5125864744,0.1457272917,-0.1106102318,0.2121107131,0.1256605089,-0.331579417,-0.0971532613,0.4307197332,0.0061918478,-0.1665689349,0.041584827,-0.1597297192,0.2058889121,-0.0072757523,-0.1105474383,0.2088841498,0.0971468911,0.2919770777,0.4504459202,0.3401712477,-0.0702545047,0.1461795121,0.1189462095,0.461900115,-0.0725690126,-0.0189571343,0.0365773179,0.3616077006,-0.0942518041,0.267055124,0.2508037388,0.5370210409,0.1577999443,-0.3334538937,0.1407443285,-0.0011267818,0.0168407932,-0.0349564403,-0.0293977167,0.7423948646,0.0886614025,0.4548160136,0.0976241603,-0.1052287593,0.0497463532,0.070672594,-0.2982068658,-0.3200121224,0.1700922996,-0.2512491941,0.1522044092,-0.1100988314,-0.0280575752,-0.0896346048,-0.1401914209,-0.1429973096,0.1518868804,0.287394762,0.0649263859,-0.1160390601,-0.0440655425,-0.216845125,0.2574723661,0.0891266763,-0.0151801826,0.269939959,-0.1740833968,0.1887930185,0.1882830113,0.4376374483,-0.1654495001,0.2808382213,0.2957637906,-0.0424122177,-0.2989651859,-0.4278821945,-0.3164432645,-0.0064289253,-0.1498689502,0.3005524576,0.1229160354,-0.3501483798,0.1867867112,-0.4390273988,-0.2368394285,-0.5237368941,0.1158800051,0.0094083678,0.0711854696,0.1030559763,0.0307732392,0.5409748554,-0.2853987813,0.0967011452,-0.1340499669,-0.3083228469,0.0993752778,0.0013545768,0.2395742238,-0.0604021735,0.2097821832,0.081299834,0.334824115,0.0728450418,-0.5948604345,0.1543280631,0.0096929101,0.1603826582,-0.0054531153,-0.5465567708,0.1438091099,-0.3508621454,0.1390538365,-0.213007018,-0.0825375468,-0.0648088157,-0.1963885576,-0.1453529745,-0.1572336257,0.0272405203,-0.422347486,-0.1761989892,0.3842809498,0.2135610282,-0.0255784877,0.125026837,0.0594722554,-0.0020090819,0.0204948448,-0.1462385207,-0.0204305947,0.0450165421,0.4003348351,-0.1566701084,-0.0140681108,-0.2479526252,-0.4340081215,-0.1029934213,-0.2318248153,-0.3111585081,-0.7258253694,-0.1302177012,0.1542342007,0.0390790775,-0.0355025753,0.073373124,-0.2161287218,0.172402218,-0.0547054857,-0.3833325803,0.0750109255,-0.0586798415,-0.0996433944,-0.4895223677,0.4194797575,0.2412148565,0.3879185915,-0.0364152007,-0.3896676898,0.0873220563,0.076990582,0.3686606288,0.0446510874,-0.1522578746,0.199431017,-0.1945101917,-0.2475325465,0.2475544065,0.0997975916,-0.198965773,-0.3358202875,0.0328706503,-0.3914494216,-0.2428810894,0.0035526403,0.2393059134,0.0837491378,0.0150698368,-0.0241549946,-0.5422309041,-0.1110889167,0.4076268077,0.2536967993,0.0903865919,-0.1011144966,-0.2331892848,-0.176485315,-0.4281748533,0.3581779599,0.0973489657,-0.0104470532,0.1392206699,-0.0305083171,-0.0058139153,-0.0933935493,0.1738209277,-0.0761902854,-0.0466233268,0.0657274649,-0.5003890395,-0.0531978533,-0.2916437984,-0.2771713734,-0.0271761119,0.1963841468,0.5380575657,-0.3245322406,0.1257040054,0.2381215096,0.0402097926,-0.2187519073,-0.0233529229,-0.1505873203,0.1194386333,-0.00511995,-0.0420500748,0.0908680558,0.1580136865,0.0723910853,0.1715500504,0.0252887439,-0.0449913368,0.5337710977,0.5966694355,-0.0149759138,0.009342852,0.0359757133,-0.0603422299,-0.4722699523,0.1994300038,0.2425858676,-0.0456342958,-0.8425169587,-0.0807879567,0.1971932203,0.3074395955,-0.029213313,0.0756809711,-0.2079823464,-0.2092396021,0.2380547225,-0.3294453621,0.5146920085,0.0873055607,0.2418275326,-0.16962111,0.0292584207,0.3550982177,0.1916737109,-0.1663481444,0.0585922226,0.0968186408,-0.1540482193,0.3305730522,0.4545246065,0.8181722164,-0.1192755327,-0.1273750216,-0.0061082221,-0.1531973034,0.2818471193,-0.4569680393,-0.0054496606,-0.1914543062,-0.256526351,-0.0477844737,-0.1412290186,0.0512917265,-0.1973418295,-0.016496595,0.1554923058,0.2559522092,0.2121570706,0.1271374971,0.216220811,0.0843381509,-0.3273906708,0.0339085385,0.0684917644,-0.2080328017,0.1866787821,-0.1093768552,0.0122277392,-0.3065899014,0.1729826033,-0.2339274585,-0.049613487,-0.078851603,0.2229246199,0.1845922768,-0.1346129775,0.0835814625,0.3559841812,0.4857665598,-0.1557325721,0.0248111561,0.0844810382,-0.2943358123,0.1334506124,0.0661283508,-0.2414879948,0.2837930024,-0.0743486285,0.3252245784,0.1920771301,0.4238166511,-0.2977109253,0.1496539563,-0.1671896875,0.1661153287,-0.0519203134,0.3414984643,0.1268112361,-0.12878187,-0.0565661564,-0.0042446391,0.1037535593,-0.221602872,0.3126445413,-0.0375188664,-0.2261631936,0.0197013114,0.0981373861,-0.0708272159,-0.1594260186,-0.0357656404,-0.1458827257,-0.2006532848,-0.022468755,-0.2284067571,-0.172006458,-0.3008998632,0.3120031357,-0.2381599545,-0.4723114967,-0.206920445,0.3734064996,0.4046835899,0.2938787341,-0.219394967,-0.2839435041,-0.3402779996,0.2772217989,-0.2316181064,0.1614444852,-0.2128948867,0.6064299345,-0.1181394085,0.2897872627,-0.2133000642,-0.2823089361,-0.2256616652,-0.0892311037,-0.4386066496,-0.3682974279,-0.0456229635,-0.0401674956,-0.0864479169,0.2165942639,-0.2286456078,-0.0405447967,-0.2426799834,0.1813552976,0.0370746627,0.4435949922,-0.1740806401,0.3260143995,0.050713487,-0.3729524612,0.0612895153,0.0980289206,-0.0713128895,0.300606668,0.210142538,0.1781324446,0.1156240404,-0.0500001609,0.3227716088,0.3425932825,0.4007493556,0.3192232847,-0.3186663687,-0.0631362125,0.1415849477,0.2257972956,0.1379989535,-0.0919605419,0.4440926313,0.2987148464,-0.0951711163,0.0900461376,0.3543199301,0.1519923657,0.099279739,0.1784839928,-0.1570458263,0.0648357272,-0.0882317647,-0.0521351658,0.2720232904,-0.1416468769,0.2238745391,0.41028139,0.3358770609,0.2722454667,0.1774859577,0.0816791654,0.2443526387,0.4868412018,0.210097447,-0.0321222357,0.003197033,0.1310182512,0.4857233763,-0.265627116,0.4763305783,-0.0616062395,-0.199505955,0.2071076483,-0.4220941961,0.2439747155,0.3563007116,0.0155826518,-0.1459272653,-0.2326164246,0.3047082424,0.3196395636,-0.0256583132,-0.0917966887,-0.0305182524,-0.1278381199,-0.3761759698,0.2058968395,0.1671888977,-0.1115561798,-0.1389711052,-0.0944052935,-0.2057560682,0.0793882459,0.035280291,-0.4773696661,-0.1281571984,0.0162588861,0.0206923727,0.3455472589,-0.19322142,0.2982699871,0.7178528309,-0.1315582693,-0.1319897473,0.4260736704,0.0161097776,-0.1136231422,0.3595230281,0.3010449111,-0.1424921453,-0.4123658836,0.2452271134,-0.0206109043,0.0123118348,0.0134780174,0.1357758641,0.0556338578,0.2323544919,0.2414001822,0.0532793477,-0.1330929697,0.0651907921,-0.3356484473,0.363342613,-0.4270902872,0.1345568895,0.3058608472,0.0306401011,-0.1007793918,0.0238449331,-0.0561517328,-0.4548441768,0.0233866572,-0.2907950878,-0.1105620041,-0.0574185178,0.0041975784,-0.1385185122,0.2731828094,0.3027709723,0.2290837467,-0.3531990349,-0.1669386327,-0.4210653007,0.0429389849,0.2039078474,-0.3283086419,-0.0392365493,0.2470375299,-0.0261073261,0.5584407449,-0.0272402931,0.2986466587,0.1250827909,0.0796056092,-0.4260629117,0.2039917558,-0.4621121585,-0.1067151278,0.1302575916,0.0617681779,-0.2476505488,0.0420593955,-0.0984423757,0.1118496284,-0.1738926023,-0.0613050312,0.5322957039,0.3644721508,0.1501181275,0.3697048426,0.1324491352,0.1355475336,0.0834814683,-0.2399796098,0.1174196973,0.0658707544,-0.1526995599,0.3562642038,-0.1907481551,0.0470767617,-0.028796073,0.251963824,0.4928355217,-0.4842080772,-0.103069365,-0.0302183591,-0.2237473726,0.0572255738,0.1399781555,0.5265302658,0.1946202219,0.237290591,0.1010437459,-0.3527945876,0.4254373908,-0.2472471744,0.0713141114,-0.2376305163,0.0519838408,0.3838028014,-0.0993833542,-0.5770970583,-0.1266644001,0.1802999079,-0.0949134752,-0.1268654168,0.2416342944,-0.185522601,0.085640505,-0.1126959696,0.1083847433,-0.0421001166,0.0535944961,-0.1877756864,-0.2174891084]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2301","title":"Unable to setup dev env on Windows","comments":"Hi @gchhablani, \r\n\r\nThere are some 3rd-party dependencies that require to build code in C. In this case, it is the library `python-Levenshtein`.\r\n\r\nOn Windows, in order to be able to build C code, you need to install at least `Microsoft C++ Build Tools` version 14. You can find more info here: https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/","body":"Hi\r\n\r\nI tried installing the `\".[dev]\"` version on Windows 10 after cloning.\r\n\r\nHere is the error I'm facing:\r\n\r\n```bat\r\n(env) C:\\testing\\datasets>pip install -e \".[dev]\"\r\nObtaining file:\/\/\/C:\/testing\/datasets\r\nRequirement already satisfied: numpy>=1.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.19.5)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-4.0.0-cp37-cp37m-win_amd64.whl (13.3 MB)\r\nRequirement already satisfied: dill in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.3.1.1)\r\nCollecting pandas\r\n Using cached pandas-1.2.4-cp37-cp37m-win_amd64.whl (9.1 MB)\r\nRequirement already satisfied: requests>=2.19.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2.25.1)\r\nRequirement already satisfied: tqdm<4.50.0,>=4.27 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.49.0)\r\nRequirement already satisfied: xxhash in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2.0.2)\r\nCollecting multiprocess\r\n Using cached multiprocess-0.70.11.1-py37-none-any.whl (108 kB)\r\nRequirement already satisfied: fsspec in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2021.4.0)\r\nCollecting huggingface_hub<0.1.0\r\n Using cached huggingface_hub-0.0.8-py3-none-any.whl (34 kB)\r\nRequirement already satisfied: importlib_metadata in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.0.1)\r\nRequirement already satisfied: absl-py in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.12.0)\r\nRequirement already satisfied: pytest in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (6.2.3)\r\nCollecting pytest-xdist\r\n Using cached pytest_xdist-2.2.1-py3-none-any.whl (37 kB)\r\nCollecting apache-beam>=2.24.0\r\n Using cached apache_beam-2.29.0-cp37-cp37m-win_amd64.whl (3.7 MB)\r\nCollecting elasticsearch\r\n Using cached elasticsearch-7.12.1-py2.py3-none-any.whl (339 kB)\r\nRequirement already satisfied: boto3==1.16.43 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.16.43)\r\nRequirement already satisfied: botocore==1.19.43 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.19.43)\r\nCollecting moto[s3]==1.3.16\r\n Using cached moto-1.3.16-py2.py3-none-any.whl (879 kB)\r\nCollecting rarfile>=4.0\r\n Using cached rarfile-4.0-py3-none-any.whl (28 kB)\r\nCollecting tensorflow>=2.3\r\n Using cached tensorflow-2.4.1-cp37-cp37m-win_amd64.whl (370.7 MB)\r\nRequirement already satisfied: torch in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.8.1)\r\nRequirement already satisfied: transformers in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.5.1)\r\nCollecting bs4\r\n Using cached bs4-0.0.1-py3-none-any.whl\r\nCollecting conllu\r\n Using cached conllu-4.4-py2.py3-none-any.whl (15 kB)\r\nCollecting langdetect\r\n Using cached langdetect-1.0.8-py3-none-any.whl\r\nCollecting lxml\r\n Using cached lxml-4.6.3-cp37-cp37m-win_amd64.whl (3.5 MB)\r\nCollecting mwparserfromhell\r\n Using cached mwparserfromhell-0.6-cp37-cp37m-win_amd64.whl (101 kB)\r\nCollecting nltk\r\n Using cached nltk-3.6.2-py3-none-any.whl (1.5 MB)\r\nCollecting openpyxl\r\n Using cached openpyxl-3.0.7-py2.py3-none-any.whl (243 kB)\r\nCollecting py7zr\r\n Using cached py7zr-0.15.2-py3-none-any.whl (66 kB)\r\nCollecting tldextract\r\n Using cached tldextract-3.1.0-py2.py3-none-any.whl (87 kB)\r\nCollecting zstandard\r\n Using cached zstandard-0.15.2-cp37-cp37m-win_amd64.whl (582 kB)\r\nCollecting bert_score>=0.3.6\r\n Using cached bert_score-0.3.9-py3-none-any.whl (59 kB)\r\nCollecting rouge_score\r\n Using cached rouge_score-0.0.4-py2.py3-none-any.whl (22 kB)\r\nCollecting sacrebleu\r\n Using cached sacrebleu-1.5.1-py3-none-any.whl (54 kB)\r\nRequirement already satisfied: scipy in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.6.3)\r\nCollecting seqeval\r\n Using cached seqeval-1.2.2-py3-none-any.whl\r\nCollecting sklearn\r\n Using cached sklearn-0.0-py2.py3-none-any.whl\r\nCollecting jiwer\r\n Using cached jiwer-2.2.0-py3-none-any.whl (13 kB)\r\nRequirement already satisfied: toml>=0.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.10.2)\r\nRequirement already satisfied: requests_file>=1.5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.5.1)\r\nRequirement already satisfied: texttable>=1.6.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.6.3)\r\nRequirement already satisfied: s3fs>=0.4.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.4.2)\r\nRequirement already satisfied: Werkzeug>=1.0.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.0.1)\r\nCollecting black\r\n Using cached black-21.4b2-py3-none-any.whl (130 kB)\r\nCollecting isort\r\n Using cached isort-5.8.0-py3-none-any.whl (103 kB)\r\nCollecting flake8==3.7.9\r\n Using cached flake8-3.7.9-py2.py3-none-any.whl (69 kB)\r\nRequirement already satisfied: jmespath<1.0.0,>=0.7.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.10.0)\r\nRequirement already satisfied: s3transfer<0.4.0,>=0.3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.3.7)\r\nRequirement already satisfied: urllib3<1.27,>=1.25.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (1.26.4)\r\nRequirement already satisfied: python-dateutil<3.0.0,>=2.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (2.8.1)\r\nCollecting entrypoints<0.4.0,>=0.3.0\r\n Using cached entrypoints-0.3-py2.py3-none-any.whl (11 kB)\r\nCollecting pyflakes<2.2.0,>=2.1.0\r\n Using cached pyflakes-2.1.1-py2.py3-none-any.whl (59 kB)\r\nCollecting pycodestyle<2.6.0,>=2.5.0\r\n Using cached pycodestyle-2.5.0-py2.py3-none-any.whl (51 kB)\r\nCollecting mccabe<0.7.0,>=0.6.0\r\n Using cached mccabe-0.6.1-py2.py3-none-any.whl (8.6 kB)\r\nRequirement already satisfied: jsondiff>=1.1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.3.0)\r\nRequirement already satisfied: pytz in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2021.1)\r\nRequirement already satisfied: mock in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.0.3)\r\nRequirement already satisfied: MarkupSafe<2.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.1.1)\r\nRequirement already satisfied: python-jose[cryptography]<4.0.0,>=3.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0)\r\nRequirement already satisfied: aws-xray-sdk!=0.96,>=0.93 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.8.0)\r\nRequirement already satisfied: cryptography>=2.3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.7)\r\nRequirement already satisfied: more-itertools in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (8.7.0)\r\nRequirement already satisfied: PyYAML>=5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.4.1)\r\nRequirement already satisfied: boto>=2.36.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.49.0)\r\nRequirement already satisfied: idna<3,>=2.5 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.10)\r\nRequirement already satisfied: sshpubkeys>=3.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.3.1)\r\nRequirement already satisfied: responses>=0.9.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.13.3)\r\nRequirement already satisfied: xmltodict in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.12.0)\r\nRequirement already satisfied: setuptools in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (52.0.0.post20210125)\r\nRequirement already satisfied: Jinja2>=2.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.11.3)\r\nRequirement already satisfied: zipp in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.1)\r\nRequirement already satisfied: six>1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.15.0)\r\nRequirement already satisfied: ecdsa<0.15 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.14.1)\r\nRequirement already satisfied: docker>=2.5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.0.0)\r\nRequirement already satisfied: cfn-lint>=0.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.49.0)\r\nRequirement already satisfied: grpcio<2,>=1.29.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (1.32.0)\r\nCollecting hdfs<3.0.0,>=2.1.0\r\n Using cached hdfs-2.6.0-py3-none-any.whl (33 kB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-3.0.0-cp37-cp37m-win_amd64.whl (12.6 MB)\r\nCollecting fastavro<2,>=0.21.4\r\n Using cached fastavro-1.4.0-cp37-cp37m-win_amd64.whl (394 kB)\r\nRequirement already satisfied: httplib2<0.18.0,>=0.8 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.17.4)\r\nCollecting pymongo<4.0.0,>=3.8.0\r\n Using cached pymongo-3.11.3-cp37-cp37m-win_amd64.whl (382 kB)\r\nCollecting crcmod<2.0,>=1.7\r\n Using cached crcmod-1.7-py3-none-any.whl\r\nCollecting avro-python3!=1.9.2,<1.10.0,>=1.8.1\r\n Using cached avro_python3-1.9.2.1-py3-none-any.whl\r\nRequirement already satisfied: typing-extensions<3.8.0,>=3.7.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.7.4.3)\r\nRequirement already satisfied: future<1.0.0,>=0.18.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.18.2)\r\nCollecting oauth2client<5,>=2.0.1\r\n Using cached oauth2client-4.1.3-py2.py3-none-any.whl (98 kB)\r\nCollecting pydot<2,>=1.2.0\r\n Using cached pydot-1.4.2-py2.py3-none-any.whl (21 kB)\r\nRequirement already satisfied: protobuf<4,>=3.12.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.15.8)\r\nRequirement already satisfied: wrapt in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from aws-xray-sdk!=0.96,>=0.93->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.12.1)\r\nCollecting matplotlib\r\n Using cached matplotlib-3.4.1-cp37-cp37m-win_amd64.whl (7.1 MB)\r\nRequirement already satisfied: junit-xml~=1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.9)\r\nRequirement already satisfied: jsonpatch in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.32)\r\nRequirement already satisfied: jsonschema~=3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0)\r\nRequirement already satisfied: networkx~=2.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.5.1)\r\nRequirement already satisfied: aws-sam-translator>=1.35.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.35.0)\r\nRequirement already satisfied: cffi>=1.12 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.14.5)\r\nRequirement already satisfied: pycparser in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cffi>=1.12->cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.20)\r\nRequirement already satisfied: pywin32==227 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (227)\r\nRequirement already satisfied: websocket-client>=0.32.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.58.0)\r\nRequirement already satisfied: docopt in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from hdfs<3.0.0,>=2.1.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.6.2)\r\nRequirement already satisfied: filelock in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from huggingface_hub<0.1.0->datasets==1.5.0.dev0) (3.0.12)\r\nRequirement already satisfied: pyrsistent>=0.14.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.17.3)\r\nRequirement already satisfied: attrs>=17.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (20.3.0)\r\nRequirement already satisfied: decorator<5,>=4.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from networkx~=2.4->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.4.2)\r\nRequirement already satisfied: rsa>=3.1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (4.7.2)\r\nRequirement already satisfied: pyasn1-modules>=0.0.5 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.2.8)\r\nRequirement already satisfied: pyasn1>=0.1.7 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.4.8)\r\nRequirement already satisfied: pyparsing>=2.1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pydot<2,>=1.2.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (2.4.7)\r\nRequirement already satisfied: certifi>=2017.4.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (2020.12.5)\r\nRequirement already satisfied: chardet<5,>=3.0.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (4.0.0)\r\nCollecting keras-preprocessing~=1.1.2\r\n Using cached Keras_Preprocessing-1.1.2-py2.py3-none-any.whl (42 kB)\r\nRequirement already satisfied: termcolor~=1.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (1.1.0)\r\nRequirement already satisfied: tensorboard~=2.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.5.0)\r\nRequirement already satisfied: wheel~=0.35 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (0.36.2)\r\nCollecting opt-einsum~=3.3.0\r\n Using cached opt_einsum-3.3.0-py3-none-any.whl (65 kB)\r\nCollecting gast==0.3.3\r\n Using cached gast-0.3.3-py2.py3-none-any.whl (9.7 kB)\r\nCollecting google-pasta~=0.2\r\n Using cached google_pasta-0.2.0-py3-none-any.whl (57 kB)\r\nRequirement already satisfied: tensorflow-estimator<2.5.0,>=2.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.4.0)\r\nCollecting astunparse~=1.6.3\r\n Using cached astunparse-1.6.3-py2.py3-none-any.whl (12 kB)\r\nCollecting flatbuffers~=1.12.0\r\n Using cached flatbuffers-1.12-py2.py3-none-any.whl (15 kB)\r\nCollecting h5py~=2.10.0\r\n Using cached h5py-2.10.0-cp37-cp37m-win_amd64.whl (2.5 MB)\r\nRequirement already satisfied: markdown>=2.6.8 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.3.4)\r\nRequirement already satisfied: tensorboard-plugin-wit>=1.6.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.8.0)\r\nRequirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.4.4)\r\nRequirement already satisfied: tensorboard-data-server<0.7.0,>=0.6.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.6.0)\r\nRequirement already satisfied: google-auth<2,>=1.6.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.30.0)\r\nRequirement already satisfied: cachetools<5.0,>=2.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from google-auth<2,>=1.6.3->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (4.2.2)\r\nRequirement already satisfied: requests-oauthlib>=0.7.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.3.0)\r\nRequirement already satisfied: oauthlib>=3.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.1.0)\r\nRequirement already satisfied: regex!=2019.12.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (2021.4.4)\r\nRequirement already satisfied: tokenizers<0.11,>=0.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (0.10.2)\r\nRequirement already satisfied: sacremoses in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (0.0.45)\r\nRequirement already satisfied: packaging in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (20.9)\r\nCollecting pathspec<1,>=0.8.1\r\n Using cached pathspec-0.8.1-py2.py3-none-any.whl (28 kB)\r\nRequirement already satisfied: click>=7.1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from black->datasets==1.5.0.dev0) (7.1.2)\r\nCollecting appdirs\r\n Using cached appdirs-1.4.4-py2.py3-none-any.whl (9.6 kB)\r\nCollecting mypy-extensions>=0.4.3\r\n Using cached mypy_extensions-0.4.3-py2.py3-none-any.whl (4.5 kB)\r\nRequirement already satisfied: typed-ast>=1.4.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from black->datasets==1.5.0.dev0) (1.4.3)\r\nCollecting beautifulsoup4\r\n Using cached beautifulsoup4-4.9.3-py3-none-any.whl (115 kB)\r\nRequirement already satisfied: soupsieve>1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from beautifulsoup4->bs4->datasets==1.5.0.dev0) (2.2.1)\r\nCollecting python-Levenshtein\r\n Using cached python-Levenshtein-0.12.2.tar.gz (50 kB)\r\nRequirement already satisfied: jsonpointer>=1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonpatch->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.1)\r\nRequirement already satisfied: pillow>=6.2.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (8.2.0)\r\nRequirement already satisfied: cycler>=0.10 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (0.10.0)\r\nRequirement already satisfied: kiwisolver>=1.0.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (1.3.1)\r\nCollecting multiprocess\r\n Using cached multiprocess-0.70.11-py3-none-any.whl (98 kB)\r\n Using cached multiprocess-0.70.10.zip (2.4 MB)\r\n Using cached multiprocess-0.70.9-py3-none-any.whl\r\nRequirement already satisfied: joblib in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from nltk->datasets==1.5.0.dev0) (1.0.1)\r\nCollecting et-xmlfile\r\n Using cached et_xmlfile-1.1.0-py3-none-any.whl (4.7 kB)\r\nRequirement already satisfied: pyzstd<0.15.0,>=0.14.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from py7zr->datasets==1.5.0.dev0) (0.14.4)\r\nCollecting pyppmd<0.13.0,>=0.12.1\r\n Using cached pyppmd-0.12.1-cp37-cp37m-win_amd64.whl (32 kB)\r\nCollecting pycryptodome>=3.6.6\r\n Using cached pycryptodome-3.10.1-cp35-abi3-win_amd64.whl (1.6 MB)\r\nCollecting bcj-cffi<0.6.0,>=0.5.1\r\n Using cached bcj_cffi-0.5.1-cp37-cp37m-win_amd64.whl (21 kB)\r\nCollecting multivolumefile<0.3.0,>=0.2.0\r\n Using cached multivolumefile-0.2.3-py3-none-any.whl (17 kB)\r\nRequirement already satisfied: iniconfig in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.1.1)\r\nRequirement already satisfied: py>=1.8.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.10.0)\r\nRequirement already satisfied: pluggy<1.0.0a1,>=0.12 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (0.13.1)\r\nRequirement already satisfied: atomicwrites>=1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.4.0)\r\nRequirement already satisfied: colorama in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (0.4.4)\r\nCollecting pytest-forked\r\n Using cached pytest_forked-1.3.0-py2.py3-none-any.whl (4.7 kB)\r\nCollecting execnet>=1.1\r\n Using cached execnet-1.8.0-py2.py3-none-any.whl (39 kB)\r\nRequirement already satisfied: apipkg>=1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from execnet>=1.1->pytest-xdist->datasets==1.5.0.dev0) (1.5)\r\nCollecting portalocker==2.0.0\r\n Using cached portalocker-2.0.0-py2.py3-none-any.whl (11 kB)\r\nRequirement already satisfied: scikit-learn>=0.21.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from seqeval->datasets==1.5.0.dev0) (0.24.2)\r\nRequirement already satisfied: threadpoolctl>=2.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from scikit-learn>=0.21.3->seqeval->datasets==1.5.0.dev0) (2.1.0)\r\nBuilding wheels for collected packages: python-Levenshtein\r\n Building wheel for python-Levenshtein (setup.py) ... error\r\n ERROR: Command errored out with exit status 1:\r\n command: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' bdist_wheel -d 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-wheel-8jh7fm18'\r\n cwd: C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\r\n Complete output (27 lines):\r\n running bdist_wheel\r\n running build\r\n running build_py\r\n creating build\r\n creating build\\lib.win-amd64-3.7\r\n creating build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\StringMatcher.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\__init__.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running egg_info\r\n writing python_Levenshtein.egg-info\\PKG-INFO\r\n writing dependency_links to python_Levenshtein.egg-info\\dependency_links.txt\r\n writing entry points to python_Levenshtein.egg-info\\entry_points.txt\r\n writing namespace_packages to python_Levenshtein.egg-info\\namespace_packages.txt\r\n writing requirements to python_Levenshtein.egg-info\\requires.txt\r\n writing top-level names to python_Levenshtein.egg-info\\top_level.txt\r\n reading manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n reading manifest template 'MANIFEST.in'\r\n warning: no previously-included files matching '*pyc' found anywhere in distribution\r\n warning: no previously-included files matching '*so' found anywhere in distribution\r\n warning: no previously-included files matching '.project' found anywhere in distribution\r\n warning: no previously-included files matching '.pydevproject' found anywhere in distribution\r\n writing manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n copying Levenshtein\\_levenshtein.c -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\_levenshtein.h -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running build_ext\r\n building 'Levenshtein._levenshtein' extension\r\n error: Microsoft Visual C++ 14.0 or greater is required. Get it with \"Microsoft C++ Build Tools\": https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/\r\n ----------------------------------------\r\n ERROR: Failed building wheel for python-Levenshtein\r\n Running setup.py clean for python-Levenshtein\r\nFailed to build python-Levenshtein\r\nInstalling collected packages: python-Levenshtein, pytest-forked, pyppmd, pymongo, pyflakes, pydot, pycryptodome, pycodestyle, pyarrow, portalocker, pathspec, pandas, opt-einsum, oauth2client, nltk, mypy-extensions, multivolumefile, multiprocess, moto, mccabe, matplotlib, keras-preprocessing, huggingface-hub, hdfs, h5py, google-pasta, gast, flatbuffers, fastavro, execnet, et-xmlfile, entrypoints, crcmod, beautifulsoup4, bcj-cffi, avro-python3, astunparse, appdirs, zstandard, tldextract, tensorflow, sklearn, seqeval, sacrebleu, rouge-score, rarfile, pytest-xdist, py7zr, openpyxl, mwparserfromhell, lxml, langdetect, jiwer, isort, flake8, elasticsearch, datasets, conllu, bs4, black, bert-score, apache-beam\r\n Running setup.py install for python-Levenshtein ... error\r\n ERROR: Command errored out with exit status 1:\r\n command: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' install --record 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-record-v7l7zitb\\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\\ProgramData\\Anaconda3\\envs\\env\\Include\\python-Levenshtein'\r\n cwd: C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\r\n Complete output (27 lines):\r\n running install\r\n running build\r\n running build_py\r\n creating build\r\n creating build\\lib.win-amd64-3.7\r\n creating build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\StringMatcher.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\__init__.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running egg_info\r\n writing python_Levenshtein.egg-info\\PKG-INFO\r\n writing dependency_links to python_Levenshtein.egg-info\\dependency_links.txt\r\n writing entry points to python_Levenshtein.egg-info\\entry_points.txt\r\n writing namespace_packages to python_Levenshtein.egg-info\\namespace_packages.txt\r\n writing requirements to python_Levenshtein.egg-info\\requires.txt\r\n writing top-level names to python_Levenshtein.egg-info\\top_level.txt\r\n reading manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n reading manifest template 'MANIFEST.in'\r\n warning: no previously-included files matching '*pyc' found anywhere in distribution\r\n warning: no previously-included files matching '*so' found anywhere in distribution\r\n warning: no previously-included files matching '.project' found anywhere in distribution\r\n warning: no previously-included files matching '.pydevproject' found anywhere in distribution\r\n writing manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n copying Levenshtein\\_levenshtein.c -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\_levenshtein.h -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running build_ext\r\n building 'Levenshtein._levenshtein' extension\r\n error: Microsoft Visual C++ 14.0 or greater is required. Get it with \"Microsoft C++ Build Tools\": https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/\r\n ----------------------------------------\r\nERROR: Command errored out with exit status 1: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' install --record 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-record-v7l7zitb\\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\\ProgramData\\Anaconda3\\envs\\env\\Include\\python-Levenshtein' Check the logs for full command output.\r\n```\r\n\r\nHere are conda and python versions:\r\n\r\n```bat\r\n(env) C:\\testing\\datasets>conda --version\r\nconda 4.9.2\r\n\r\n(env) C:\\testing\\datasets>python --version\r\nPython 3.7.10\r\n```\r\n\r\nPlease help me out. Thanks.","comment_length":52,"text":"Unable to setup dev env on Windows\nHi\r\n\r\nI tried installing the `\".[dev]\"` version on Windows 10 after cloning.\r\n\r\nHere is the error I'm facing:\r\n\r\n```bat\r\n(env) C:\\testing\\datasets>pip install -e \".[dev]\"\r\nObtaining file:\/\/\/C:\/testing\/datasets\r\nRequirement already satisfied: numpy>=1.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.19.5)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-4.0.0-cp37-cp37m-win_amd64.whl (13.3 MB)\r\nRequirement already satisfied: dill in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.3.1.1)\r\nCollecting pandas\r\n Using cached pandas-1.2.4-cp37-cp37m-win_amd64.whl (9.1 MB)\r\nRequirement already satisfied: requests>=2.19.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2.25.1)\r\nRequirement already satisfied: tqdm<4.50.0,>=4.27 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.49.0)\r\nRequirement already satisfied: xxhash in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2.0.2)\r\nCollecting multiprocess\r\n Using cached multiprocess-0.70.11.1-py37-none-any.whl (108 kB)\r\nRequirement already satisfied: fsspec in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2021.4.0)\r\nCollecting huggingface_hub<0.1.0\r\n Using cached huggingface_hub-0.0.8-py3-none-any.whl (34 kB)\r\nRequirement already satisfied: importlib_metadata in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.0.1)\r\nRequirement already satisfied: absl-py in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.12.0)\r\nRequirement already satisfied: pytest in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (6.2.3)\r\nCollecting pytest-xdist\r\n Using cached pytest_xdist-2.2.1-py3-none-any.whl (37 kB)\r\nCollecting apache-beam>=2.24.0\r\n Using cached apache_beam-2.29.0-cp37-cp37m-win_amd64.whl (3.7 MB)\r\nCollecting elasticsearch\r\n Using cached elasticsearch-7.12.1-py2.py3-none-any.whl (339 kB)\r\nRequirement already satisfied: boto3==1.16.43 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.16.43)\r\nRequirement already satisfied: botocore==1.19.43 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.19.43)\r\nCollecting moto[s3]==1.3.16\r\n Using cached moto-1.3.16-py2.py3-none-any.whl (879 kB)\r\nCollecting rarfile>=4.0\r\n Using cached rarfile-4.0-py3-none-any.whl (28 kB)\r\nCollecting tensorflow>=2.3\r\n Using cached tensorflow-2.4.1-cp37-cp37m-win_amd64.whl (370.7 MB)\r\nRequirement already satisfied: torch in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.8.1)\r\nRequirement already satisfied: transformers in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.5.1)\r\nCollecting bs4\r\n Using cached bs4-0.0.1-py3-none-any.whl\r\nCollecting conllu\r\n Using cached conllu-4.4-py2.py3-none-any.whl (15 kB)\r\nCollecting langdetect\r\n Using cached langdetect-1.0.8-py3-none-any.whl\r\nCollecting lxml\r\n Using cached lxml-4.6.3-cp37-cp37m-win_amd64.whl (3.5 MB)\r\nCollecting mwparserfromhell\r\n Using cached mwparserfromhell-0.6-cp37-cp37m-win_amd64.whl (101 kB)\r\nCollecting nltk\r\n Using cached nltk-3.6.2-py3-none-any.whl (1.5 MB)\r\nCollecting openpyxl\r\n Using cached openpyxl-3.0.7-py2.py3-none-any.whl (243 kB)\r\nCollecting py7zr\r\n Using cached py7zr-0.15.2-py3-none-any.whl (66 kB)\r\nCollecting tldextract\r\n Using cached tldextract-3.1.0-py2.py3-none-any.whl (87 kB)\r\nCollecting zstandard\r\n Using cached zstandard-0.15.2-cp37-cp37m-win_amd64.whl (582 kB)\r\nCollecting bert_score>=0.3.6\r\n Using cached bert_score-0.3.9-py3-none-any.whl (59 kB)\r\nCollecting rouge_score\r\n Using cached rouge_score-0.0.4-py2.py3-none-any.whl (22 kB)\r\nCollecting sacrebleu\r\n Using cached sacrebleu-1.5.1-py3-none-any.whl (54 kB)\r\nRequirement already satisfied: scipy in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.6.3)\r\nCollecting seqeval\r\n Using cached seqeval-1.2.2-py3-none-any.whl\r\nCollecting sklearn\r\n Using cached sklearn-0.0-py2.py3-none-any.whl\r\nCollecting jiwer\r\n Using cached jiwer-2.2.0-py3-none-any.whl (13 kB)\r\nRequirement already satisfied: toml>=0.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.10.2)\r\nRequirement already satisfied: requests_file>=1.5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.5.1)\r\nRequirement already satisfied: texttable>=1.6.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.6.3)\r\nRequirement already satisfied: s3fs>=0.4.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.4.2)\r\nRequirement already satisfied: Werkzeug>=1.0.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.0.1)\r\nCollecting black\r\n Using cached black-21.4b2-py3-none-any.whl (130 kB)\r\nCollecting isort\r\n Using cached isort-5.8.0-py3-none-any.whl (103 kB)\r\nCollecting flake8==3.7.9\r\n Using cached flake8-3.7.9-py2.py3-none-any.whl (69 kB)\r\nRequirement already satisfied: jmespath<1.0.0,>=0.7.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.10.0)\r\nRequirement already satisfied: s3transfer<0.4.0,>=0.3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.3.7)\r\nRequirement already satisfied: urllib3<1.27,>=1.25.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (1.26.4)\r\nRequirement already satisfied: python-dateutil<3.0.0,>=2.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (2.8.1)\r\nCollecting entrypoints<0.4.0,>=0.3.0\r\n Using cached entrypoints-0.3-py2.py3-none-any.whl (11 kB)\r\nCollecting pyflakes<2.2.0,>=2.1.0\r\n Using cached pyflakes-2.1.1-py2.py3-none-any.whl (59 kB)\r\nCollecting pycodestyle<2.6.0,>=2.5.0\r\n Using cached pycodestyle-2.5.0-py2.py3-none-any.whl (51 kB)\r\nCollecting mccabe<0.7.0,>=0.6.0\r\n Using cached mccabe-0.6.1-py2.py3-none-any.whl (8.6 kB)\r\nRequirement already satisfied: jsondiff>=1.1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.3.0)\r\nRequirement already satisfied: pytz in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2021.1)\r\nRequirement already satisfied: mock in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.0.3)\r\nRequirement already satisfied: MarkupSafe<2.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.1.1)\r\nRequirement already satisfied: python-jose[cryptography]<4.0.0,>=3.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0)\r\nRequirement already satisfied: aws-xray-sdk!=0.96,>=0.93 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.8.0)\r\nRequirement already satisfied: cryptography>=2.3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.7)\r\nRequirement already satisfied: more-itertools in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (8.7.0)\r\nRequirement already satisfied: PyYAML>=5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.4.1)\r\nRequirement already satisfied: boto>=2.36.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.49.0)\r\nRequirement already satisfied: idna<3,>=2.5 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.10)\r\nRequirement already satisfied: sshpubkeys>=3.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.3.1)\r\nRequirement already satisfied: responses>=0.9.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.13.3)\r\nRequirement already satisfied: xmltodict in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.12.0)\r\nRequirement already satisfied: setuptools in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (52.0.0.post20210125)\r\nRequirement already satisfied: Jinja2>=2.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.11.3)\r\nRequirement already satisfied: zipp in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.1)\r\nRequirement already satisfied: six>1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.15.0)\r\nRequirement already satisfied: ecdsa<0.15 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.14.1)\r\nRequirement already satisfied: docker>=2.5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.0.0)\r\nRequirement already satisfied: cfn-lint>=0.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.49.0)\r\nRequirement already satisfied: grpcio<2,>=1.29.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (1.32.0)\r\nCollecting hdfs<3.0.0,>=2.1.0\r\n Using cached hdfs-2.6.0-py3-none-any.whl (33 kB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-3.0.0-cp37-cp37m-win_amd64.whl (12.6 MB)\r\nCollecting fastavro<2,>=0.21.4\r\n Using cached fastavro-1.4.0-cp37-cp37m-win_amd64.whl (394 kB)\r\nRequirement already satisfied: httplib2<0.18.0,>=0.8 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.17.4)\r\nCollecting pymongo<4.0.0,>=3.8.0\r\n Using cached pymongo-3.11.3-cp37-cp37m-win_amd64.whl (382 kB)\r\nCollecting crcmod<2.0,>=1.7\r\n Using cached crcmod-1.7-py3-none-any.whl\r\nCollecting avro-python3!=1.9.2,<1.10.0,>=1.8.1\r\n Using cached avro_python3-1.9.2.1-py3-none-any.whl\r\nRequirement already satisfied: typing-extensions<3.8.0,>=3.7.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.7.4.3)\r\nRequirement already satisfied: future<1.0.0,>=0.18.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.18.2)\r\nCollecting oauth2client<5,>=2.0.1\r\n Using cached oauth2client-4.1.3-py2.py3-none-any.whl (98 kB)\r\nCollecting pydot<2,>=1.2.0\r\n Using cached pydot-1.4.2-py2.py3-none-any.whl (21 kB)\r\nRequirement already satisfied: protobuf<4,>=3.12.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.15.8)\r\nRequirement already satisfied: wrapt in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from aws-xray-sdk!=0.96,>=0.93->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.12.1)\r\nCollecting matplotlib\r\n Using cached matplotlib-3.4.1-cp37-cp37m-win_amd64.whl (7.1 MB)\r\nRequirement already satisfied: junit-xml~=1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.9)\r\nRequirement already satisfied: jsonpatch in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.32)\r\nRequirement already satisfied: jsonschema~=3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0)\r\nRequirement already satisfied: networkx~=2.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.5.1)\r\nRequirement already satisfied: aws-sam-translator>=1.35.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.35.0)\r\nRequirement already satisfied: cffi>=1.12 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.14.5)\r\nRequirement already satisfied: pycparser in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cffi>=1.12->cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.20)\r\nRequirement already satisfied: pywin32==227 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (227)\r\nRequirement already satisfied: websocket-client>=0.32.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.58.0)\r\nRequirement already satisfied: docopt in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from hdfs<3.0.0,>=2.1.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.6.2)\r\nRequirement already satisfied: filelock in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from huggingface_hub<0.1.0->datasets==1.5.0.dev0) (3.0.12)\r\nRequirement already satisfied: pyrsistent>=0.14.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.17.3)\r\nRequirement already satisfied: attrs>=17.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (20.3.0)\r\nRequirement already satisfied: decorator<5,>=4.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from networkx~=2.4->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.4.2)\r\nRequirement already satisfied: rsa>=3.1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (4.7.2)\r\nRequirement already satisfied: pyasn1-modules>=0.0.5 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.2.8)\r\nRequirement already satisfied: pyasn1>=0.1.7 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.4.8)\r\nRequirement already satisfied: pyparsing>=2.1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pydot<2,>=1.2.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (2.4.7)\r\nRequirement already satisfied: certifi>=2017.4.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (2020.12.5)\r\nRequirement already satisfied: chardet<5,>=3.0.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (4.0.0)\r\nCollecting keras-preprocessing~=1.1.2\r\n Using cached Keras_Preprocessing-1.1.2-py2.py3-none-any.whl (42 kB)\r\nRequirement already satisfied: termcolor~=1.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (1.1.0)\r\nRequirement already satisfied: tensorboard~=2.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.5.0)\r\nRequirement already satisfied: wheel~=0.35 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (0.36.2)\r\nCollecting opt-einsum~=3.3.0\r\n Using cached opt_einsum-3.3.0-py3-none-any.whl (65 kB)\r\nCollecting gast==0.3.3\r\n Using cached gast-0.3.3-py2.py3-none-any.whl (9.7 kB)\r\nCollecting google-pasta~=0.2\r\n Using cached google_pasta-0.2.0-py3-none-any.whl (57 kB)\r\nRequirement already satisfied: tensorflow-estimator<2.5.0,>=2.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.4.0)\r\nCollecting astunparse~=1.6.3\r\n Using cached astunparse-1.6.3-py2.py3-none-any.whl (12 kB)\r\nCollecting flatbuffers~=1.12.0\r\n Using cached flatbuffers-1.12-py2.py3-none-any.whl (15 kB)\r\nCollecting h5py~=2.10.0\r\n Using cached h5py-2.10.0-cp37-cp37m-win_amd64.whl (2.5 MB)\r\nRequirement already satisfied: markdown>=2.6.8 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.3.4)\r\nRequirement already satisfied: tensorboard-plugin-wit>=1.6.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.8.0)\r\nRequirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.4.4)\r\nRequirement already satisfied: tensorboard-data-server<0.7.0,>=0.6.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.6.0)\r\nRequirement already satisfied: google-auth<2,>=1.6.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.30.0)\r\nRequirement already satisfied: cachetools<5.0,>=2.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from google-auth<2,>=1.6.3->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (4.2.2)\r\nRequirement already satisfied: requests-oauthlib>=0.7.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.3.0)\r\nRequirement already satisfied: oauthlib>=3.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.1.0)\r\nRequirement already satisfied: regex!=2019.12.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (2021.4.4)\r\nRequirement already satisfied: tokenizers<0.11,>=0.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (0.10.2)\r\nRequirement already satisfied: sacremoses in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (0.0.45)\r\nRequirement already satisfied: packaging in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (20.9)\r\nCollecting pathspec<1,>=0.8.1\r\n Using cached pathspec-0.8.1-py2.py3-none-any.whl (28 kB)\r\nRequirement already satisfied: click>=7.1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from black->datasets==1.5.0.dev0) (7.1.2)\r\nCollecting appdirs\r\n Using cached appdirs-1.4.4-py2.py3-none-any.whl (9.6 kB)\r\nCollecting mypy-extensions>=0.4.3\r\n Using cached mypy_extensions-0.4.3-py2.py3-none-any.whl (4.5 kB)\r\nRequirement already satisfied: typed-ast>=1.4.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from black->datasets==1.5.0.dev0) (1.4.3)\r\nCollecting beautifulsoup4\r\n Using cached beautifulsoup4-4.9.3-py3-none-any.whl (115 kB)\r\nRequirement already satisfied: soupsieve>1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from beautifulsoup4->bs4->datasets==1.5.0.dev0) (2.2.1)\r\nCollecting python-Levenshtein\r\n Using cached python-Levenshtein-0.12.2.tar.gz (50 kB)\r\nRequirement already satisfied: jsonpointer>=1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonpatch->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.1)\r\nRequirement already satisfied: pillow>=6.2.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (8.2.0)\r\nRequirement already satisfied: cycler>=0.10 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (0.10.0)\r\nRequirement already satisfied: kiwisolver>=1.0.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (1.3.1)\r\nCollecting multiprocess\r\n Using cached multiprocess-0.70.11-py3-none-any.whl (98 kB)\r\n Using cached multiprocess-0.70.10.zip (2.4 MB)\r\n Using cached multiprocess-0.70.9-py3-none-any.whl\r\nRequirement already satisfied: joblib in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from nltk->datasets==1.5.0.dev0) (1.0.1)\r\nCollecting et-xmlfile\r\n Using cached et_xmlfile-1.1.0-py3-none-any.whl (4.7 kB)\r\nRequirement already satisfied: pyzstd<0.15.0,>=0.14.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from py7zr->datasets==1.5.0.dev0) (0.14.4)\r\nCollecting pyppmd<0.13.0,>=0.12.1\r\n Using cached pyppmd-0.12.1-cp37-cp37m-win_amd64.whl (32 kB)\r\nCollecting pycryptodome>=3.6.6\r\n Using cached pycryptodome-3.10.1-cp35-abi3-win_amd64.whl (1.6 MB)\r\nCollecting bcj-cffi<0.6.0,>=0.5.1\r\n Using cached bcj_cffi-0.5.1-cp37-cp37m-win_amd64.whl (21 kB)\r\nCollecting multivolumefile<0.3.0,>=0.2.0\r\n Using cached multivolumefile-0.2.3-py3-none-any.whl (17 kB)\r\nRequirement already satisfied: iniconfig in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.1.1)\r\nRequirement already satisfied: py>=1.8.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.10.0)\r\nRequirement already satisfied: pluggy<1.0.0a1,>=0.12 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (0.13.1)\r\nRequirement already satisfied: atomicwrites>=1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.4.0)\r\nRequirement already satisfied: colorama in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (0.4.4)\r\nCollecting pytest-forked\r\n Using cached pytest_forked-1.3.0-py2.py3-none-any.whl (4.7 kB)\r\nCollecting execnet>=1.1\r\n Using cached execnet-1.8.0-py2.py3-none-any.whl (39 kB)\r\nRequirement already satisfied: apipkg>=1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from execnet>=1.1->pytest-xdist->datasets==1.5.0.dev0) (1.5)\r\nCollecting portalocker==2.0.0\r\n Using cached portalocker-2.0.0-py2.py3-none-any.whl (11 kB)\r\nRequirement already satisfied: scikit-learn>=0.21.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from seqeval->datasets==1.5.0.dev0) (0.24.2)\r\nRequirement already satisfied: threadpoolctl>=2.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from scikit-learn>=0.21.3->seqeval->datasets==1.5.0.dev0) (2.1.0)\r\nBuilding wheels for collected packages: python-Levenshtein\r\n Building wheel for python-Levenshtein (setup.py) ... error\r\n ERROR: Command errored out with exit status 1:\r\n command: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' bdist_wheel -d 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-wheel-8jh7fm18'\r\n cwd: C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\r\n Complete output (27 lines):\r\n running bdist_wheel\r\n running build\r\n running build_py\r\n creating build\r\n creating build\\lib.win-amd64-3.7\r\n creating build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\StringMatcher.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\__init__.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running egg_info\r\n writing python_Levenshtein.egg-info\\PKG-INFO\r\n writing dependency_links to python_Levenshtein.egg-info\\dependency_links.txt\r\n writing entry points to python_Levenshtein.egg-info\\entry_points.txt\r\n writing namespace_packages to python_Levenshtein.egg-info\\namespace_packages.txt\r\n writing requirements to python_Levenshtein.egg-info\\requires.txt\r\n writing top-level names to python_Levenshtein.egg-info\\top_level.txt\r\n reading manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n reading manifest template 'MANIFEST.in'\r\n warning: no previously-included files matching '*pyc' found anywhere in distribution\r\n warning: no previously-included files matching '*so' found anywhere in distribution\r\n warning: no previously-included files matching '.project' found anywhere in distribution\r\n warning: no previously-included files matching '.pydevproject' found anywhere in distribution\r\n writing manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n copying Levenshtein\\_levenshtein.c -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\_levenshtein.h -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running build_ext\r\n building 'Levenshtein._levenshtein' extension\r\n error: Microsoft Visual C++ 14.0 or greater is required. Get it with \"Microsoft C++ Build Tools\": https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/\r\n ----------------------------------------\r\n ERROR: Failed building wheel for python-Levenshtein\r\n Running setup.py clean for python-Levenshtein\r\nFailed to build python-Levenshtein\r\nInstalling collected packages: python-Levenshtein, pytest-forked, pyppmd, pymongo, pyflakes, pydot, pycryptodome, pycodestyle, pyarrow, portalocker, pathspec, pandas, opt-einsum, oauth2client, nltk, mypy-extensions, multivolumefile, multiprocess, moto, mccabe, matplotlib, keras-preprocessing, huggingface-hub, hdfs, h5py, google-pasta, gast, flatbuffers, fastavro, execnet, et-xmlfile, entrypoints, crcmod, beautifulsoup4, bcj-cffi, avro-python3, astunparse, appdirs, zstandard, tldextract, tensorflow, sklearn, seqeval, sacrebleu, rouge-score, rarfile, pytest-xdist, py7zr, openpyxl, mwparserfromhell, lxml, langdetect, jiwer, isort, flake8, elasticsearch, datasets, conllu, bs4, black, bert-score, apache-beam\r\n Running setup.py install for python-Levenshtein ... error\r\n ERROR: Command errored out with exit status 1:\r\n command: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' install --record 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-record-v7l7zitb\\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\\ProgramData\\Anaconda3\\envs\\env\\Include\\python-Levenshtein'\r\n cwd: C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\r\n Complete output (27 lines):\r\n running install\r\n running build\r\n running build_py\r\n creating build\r\n creating build\\lib.win-amd64-3.7\r\n creating build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\StringMatcher.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\__init__.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running egg_info\r\n writing python_Levenshtein.egg-info\\PKG-INFO\r\n writing dependency_links to python_Levenshtein.egg-info\\dependency_links.txt\r\n writing entry points to python_Levenshtein.egg-info\\entry_points.txt\r\n writing namespace_packages to python_Levenshtein.egg-info\\namespace_packages.txt\r\n writing requirements to python_Levenshtein.egg-info\\requires.txt\r\n writing top-level names to python_Levenshtein.egg-info\\top_level.txt\r\n reading manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n reading manifest template 'MANIFEST.in'\r\n warning: no previously-included files matching '*pyc' found anywhere in distribution\r\n warning: no previously-included files matching '*so' found anywhere in distribution\r\n warning: no previously-included files matching '.project' found anywhere in distribution\r\n warning: no previously-included files matching '.pydevproject' found anywhere in distribution\r\n writing manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n copying Levenshtein\\_levenshtein.c -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\_levenshtein.h -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running build_ext\r\n building 'Levenshtein._levenshtein' extension\r\n error: Microsoft Visual C++ 14.0 or greater is required. Get it with \"Microsoft C++ Build Tools\": https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/\r\n ----------------------------------------\r\nERROR: Command errored out with exit status 1: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' install --record 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-record-v7l7zitb\\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\\ProgramData\\Anaconda3\\envs\\env\\Include\\python-Levenshtein' Check the logs for full command output.\r\n```\r\n\r\nHere are conda and python versions:\r\n\r\n```bat\r\n(env) C:\\testing\\datasets>conda --version\r\nconda 4.9.2\r\n\r\n(env) C:\\testing\\datasets>python --version\r\nPython 3.7.10\r\n```\r\n\r\nPlease help me out. Thanks.\nHi @gchhablani, \r\n\r\nThere are some 3rd-party dependencies that require to build code in C. In this case, it is the library `python-Levenshtein`.\r\n\r\nOn Windows, in order to be able to build C code, you need to install at least `Microsoft C++ Build Tools` version 14. You can find more info here: https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/","embeddings":[-0.4121342301,-0.0855779797,-0.0977274776,-0.1387811452,0.3114477396,0.068374902,0.3448515236,0.0465872362,-0.1001304686,0.15623492,0.0433793887,0.2424268872,0.1581842005,0.2162972242,0.0545784347,0.0232140273,0.1360678375,0.4390758872,-0.2185967714,0.037439011,-0.0659957454,-0.0208863225,-0.1535535008,-0.0436114408,0.3134841323,-0.0264739394,-0.2292296886,0.1753225178,-0.3534286618,-0.3225588799,0.4030125737,-0.1064584628,0.4329041839,0.1327629685,-0.0001171767,0.2016099095,0.3223023117,-0.1553241611,-0.0909361094,0.0166196972,-0.0741046444,-0.4187884331,-0.0570751615,-0.1152198911,0.0013719843,-0.315310955,0.1467733234,-0.3042214811,0.0935275629,0.1765608937,0.1489830762,0.0158207696,0.3976289928,-0.0599443838,-0.0897259861,0.2228936255,-0.1047845185,0.2224471718,0.3808677495,-0.0767932087,0.2623265088,-0.0150638493,0.2122566849,0.0888135731,0.1605111212,0.0911397189,0.8259810209,-0.3334536552,-0.0124987392,-0.0583257899,0.4455650151,-0.1256540567,-0.2552556992,0.2472820431,0.1414016038,-0.325605303,0.261488229,0.3790609241,-0.3644766808,0.0435973518,-0.362721324,-0.2574132085,-0.4395217001,0.0149611542,0.1180061549,0.0826293603,0.1625450402,0.1449573189,-0.0660314187,0.0814976096,0.4332493246,0.0780942217,-0.1706448048,0.0194142777,-0.1333711296,-0.1515952498,-0.2876423001,0.3090680242,-0.0119880065,0.0120718013,-0.3012442887,-0.2072132975,0.2847516239,0.0740561634,-0.0211096983,0.3143170476,0.0671262816,-0.0794963837,0.1001376882,-0.1034927294,0.4024324119,0.0954257473,-0.220410049,-0.4841091931,-0.1915215552,0.502032876,0.0810901076,-0.1381357461,-0.2252849638,-0.0552830324,-0.278065145,0.1433484256,0.2262274027,0.247914955,-0.0905870795,-0.0975939855,0.0245632939,0.2494825125,-0.2053817958,0.111318931,0.1260858774,0.1531590074,-0.4578863382,-0.3628825843,0.2566353381,-0.2950874269,0.3838174641,0.2732485831,-0.0795328692,0.2728429437,-0.0912644342,0.190502882,0.1253395081,0.2146690041,0.192412436,0.1868555695,-0.0017933784,-0.2911956906,-0.0418691561,0.2853320539,0.0168284662,-0.2246380895,-0.3471564054,0.1082542017,-0.1117702425,0.017555505,0.0353535265,-0.0690591186,0.0887759179,0.1203753427,-0.0526402183,-0.1925052106,0.0195803028,-0.1289805025,0.3793908656,0.0687448531,-0.4063456953,0.2841972411,0.077341266,-0.1427752674,0.1688283831,0.032434091,-0.0016664289,0.1482707858,-0.2415356338,-0.2094384283,-0.1005012766,-0.4556683302,-0.1719743758,0.3552168608,-0.0029001641,-0.4598074555,0.2365486175,-0.120121628,0.0676399097,-0.1652117074,-0.088309899,0.2276965827,-0.035136126,-0.2145301253,-0.168868646,-0.2638670802,0.1783894897,-0.0220463779,-0.0101095643,-0.2324422002,-0.0198157672,-0.1845362335,-0.0388986133,-0.1567320377,0.2039210051,0.381216675,0.5579845905,-0.0816559568,-0.0684605464,0.0482342988,-0.1448994875,0.159021914,-0.4344110489,0.107544288,-0.0217336845,0.149328202,-0.0466755331,0.0853103474,-0.4496676028,-0.3711802661,0.116788134,0.0079455134,0.3306221664,-0.0743468627,0.1798881143,0.0693052262,0.0324366949,0.1059975326,-0.166411072,0.1343197078,-0.7459347248,-0.0168207139,-0.1043569595,0.1043144912,0.0355914086,-0.0769519955,0.0382833369,0.1590626389,-0.1584957093,-0.3072966635,-0.041364219,-0.3169023693,0.2125235647,-0.5566606522,0.2742674053,0.1649557948,0.0273893774,0.1721837968,-0.0810752809,-0.1246714965,-0.2803698778,-0.0341622308,-0.0767948776,0.297168076,0.3042367697,0.0494527295,0.0207218789,-0.1933784485,0.4338948429,0.575902164,0.0313213542,-0.3306666613,0.2104430646,0.0414407551,-0.1986906677,-0.0196556337,-0.142107293,-0.0658592954,-0.0807882249,0.0347912349,0.1587107182,-0.0406574272,0.5600038171,0.1366593093,-0.0128204282,0.2236149907,-0.2414536029,-0.0290230308,0.4613452852,-0.0718432441,-0.0432940163,0.1160499156,-0.3231308162,-0.1007453874,-0.0150394756,-0.2076589614,0.4360066056,0.3666128814,-0.414539963,-0.0872983187,-0.4251069129,0.0179413594,-0.3845640123,-0.5058633089,-0.1485284567,0.072969012,0.2087071687,-0.1159329861,0.2937576473,0.378765434,0.1709450483,-0.1817639172,-0.0483662486,-0.0111401537,-0.0018849734,-0.0123814195,-0.2988424599,0.0734532699,0.352761358,0.1322950125,0.4342385232,0.0266307555,0.0442592315,-0.1747365743,-0.3875649571,-0.0144538507,-0.1596354544,0.7915337682,0.1893676966,0.1281995326,-0.0219017398,-0.2373072803,0.2668535709,0.046285782,-0.2804477513,0.0047871685,-0.2273965627,-0.4810552001,-0.180219084,-0.3516542315,-0.1246601343,-0.0893754438,-0.107180655,-0.1341093034,-0.0714597255,0.4129341841,-0.0137822358,-0.0348017626,0.2470746636,-0.0075236536,-0.0635410845,0.111187987,0.2345679998,-0.4932157695,-0.4311946332,0.2324139327,0.0099825086,0.1128880084,0.1875332743,-0.2512127757,-0.1148042381,-0.1031353101,0.0084202457,-0.0645647049,0.2379448414,0.4715803862,0.2665568888,0.0352040157,-0.0425962657,-0.2322521061,0.1093985289,-0.0791745335,0.4247240722,0.2350768745,0.577147603,0.0135583226,0.4463897347,0.7228881121,0.0914104655,0.2535391748,-0.0413708538,0.3796830773,0.0905267894,-0.3451045454,0.2690578103,0.0630212799,-0.0072042472,0.1809092909,-0.0321068391,-0.1778542548,-0.2200647742,0.0465762503,-0.0596503131,0.0298709534,-0.1174279228,-0.3436402082,0.3009181321,-0.2163112164,0.1664456427,0.132733658,-0.0476562269,0.0549124517,0.462567538,-0.2057617009,0.2380177081,-0.4431191385,-0.55777812,-0.3410708904,0.2160663605,-0.0173758063,0.6229293942,-0.2246486992,0.4463389814,0.1381700486,-0.2857893705,0.4693852663,-0.4746426642,0.0524854362,-0.0733335987,0.0820763633,-0.3210382462,-0.2203683704,0.0427084416,0.1718384922,-0.0350638144,-0.1515042037,-0.1922936738,0.2373284698,0.2754104733,0.0534360409,-0.1614242047,-0.0936392322,-0.2664470971,-0.4962081015,-0.0505326651,-0.0729672164,0.0656276122,0.0504231825,0.2158268392,0.2125395685,-0.062741369,0.0499402396,0.1299736947,-0.002951686,0.0127464049,-0.0393292047,0.277925998,-0.0897330716,0.242069453,0.0891641825,0.2664802969,-0.214750424,-0.1888166517,-0.0411368757,-0.1824876517,0.4089705944,0.0675264001,-0.1009932458,-0.3996094167,-0.0261758715,-0.1057991907,-0.2291745842,-0.1686869115,0.2476257235,0.0220554098,-0.3125375807,-0.1065755039,0.0946872979,-0.0872679502,0.126850307,0.3581095934,0.2488824427,-0.2112712711,-0.0633669794,0.3672279716,0.6484234333,0.0318691768,-0.1837054044,0.0639485046,-0.5025157928,0.127110377,0.0638693869,0.1315517128,-0.1698898226,-0.1872195154,0.0249396823,-0.1965131164,0.324855268,0.0538668744,-0.4849865437,0.1431676745,0.4321363568,0.123522751,0.1298967749,0.103631869,0.0593833961,-0.1762978286,0.1346544772,0.1409142166,0.0725384578,0.1524269283,-0.1616719067,0.0688693598,-0.1432239711,-0.2424301654,-0.6267002225,0.112472631,-0.2296545804,0.277250886,-0.2013393044,-0.608169198,0.4104798436,0.3885934055,0.569214344,0.2539714575,-0.2626110911,0.0875652134,-0.0031408167,-0.3031441569,0.1424965262,-0.0968290716,0.0821555853,-0.2844583094,-0.3339810669,-0.1550788134,-0.06838236,0.0433429256,-0.3456445634,-0.0371471792,-0.1009884551,0.0593316406,0.2151547074,-0.12476971,0.2778620124,0.0004334133,0.0718653649,-0.0310869589,-0.0376362354,0.3993119001,-0.2985121012,-0.157242015,-0.0628597215,0.0892680213,-0.1416343004,-0.097849831,0.3126449585,0.0009065954,-0.1707309037,-0.1474594772,0.0576615967,-0.2658415139,-0.2694555819,-0.4231140912,0.154992193,0.0106100775,0.1373024285,-0.1414523721,0.1642699391,-0.2475146353,-0.0291168727,-0.3162648976,-0.1858838648,-0.0911561549,-0.2300188392,0.2858869433,0.1424523741,0.2312096506,0.2592296898,-0.0326089486,-0.2188580483,0.001480481,0.2834210098,0.1979220957,-0.0843515694,-0.3264656067,-0.0342857167,-0.0405877642,0.0829677805,-0.2389823347,-0.1508959979,-0.1201620698,0.1044125929,0.1108317301,0.0316658281,-0.2158530653,0.1758563519,-0.345400095,-0.3830916286,0.03885803,-0.0019292964,0.152979821,-0.1848097295,0.3930841386,0.045836769,-0.0579165481,-0.2661431134,0.0873562098,-0.2104155123,0.5279401541,0.0870263129,0.2425913066,0.0917045474,-0.0350838788,-0.6708033085,0.2103807628,0.0831123367,0.0742151961,-0.0349078625,-0.1580380648,0.1837056726,-0.221336782,0.2976179719,0.3265206516,-0.0052294247,0.1880482137,0.54577142,0.1286130399,-0.237014696,0.1696367413,0.1250644326,-0.078648977,-0.1110885218,0.1126404256,0.2368839383,-0.2054820806,0.3799754977,-0.0941831097,0.4906442463,-0.2079764307,0.004008227,0.5395568609,0.0277876202,0.0393729135,-0.0421883427,0.0318805948,0.1792411208,0.583144486,0.0031546124,0.0298069771,0.140637517,0.3013638258,-0.1679567248,-0.0243418384,-0.1806923449,-0.2011103779,0.2749938071,0.0183627587,0.2508787215,0.6245895624,-0.4090289176,-0.061805658,-0.5428097844,0.2052147239,-0.2572387755,-0.0793285444,-0.2710337043,0.209878087,0.0581043065,0.0631146058,0.1123113036,-0.1852890104,0.0515824072,0.2637988329,-0.4474705756,-0.2216476202,0.3423317373,0.3455912173,-0.1874864846,0.0017540216,0.287109375,0.2440224737,-0.1056914702,-0.0194142144,0.4746490121,0.6444396973,0.4055102468,-0.3922822475,-0.0329234637,0.0636443719,0.0366322398,-0.1734556854,0.2106984258,0.030856356,0.1918324083,0.1406432539,0.0972324535,0.0374468304,-0.0763897672,0.2183976471,-0.1176240891,0.2614663243,0.0845798329,0.1263475567,0.1881272048,0.1792557985,-0.0836737677,-0.3425420523,-0.0696679875,0.1219683215,-0.0048621199,-0.1159545034,-0.110576421,0.0403339155,-0.2090077251,0.3243708313,0.2986623645,0.032985013,-0.3327325284,-0.1645352095,-0.3989806473,-0.1721303314,-0.0934053212,-0.3542155325,-0.0387813784,-0.1040405557,0.0048022824,0.079297021,0.0293139573,0.0266957842,0.1461547613,0.6865946054,-0.2840730846,0.0265500713,0.3582297862,0.1359502375,0.0420462489,-0.330671519,0.0253973547,0.3507256508,-0.0839070082,-0.1968306303,0.2640077472,0.1867125332,-0.1484750211,0.3411541581,0.2205915898,0.7242153883,0.121033676,-0.0829702541,-0.3107076883,0.0463107191,-0.1874597371,0.3284565806,0.4097307622,-0.2244555354,-0.3802423477,-0.4079791903,-0.1977362782,0.1020833179,0.0908953398,0.4288522899,-0.3884783685,0.3272823095,-0.0742274299,0.1896982193,0.104075551,0.0265209787,-0.0568277091,0.2692764401,-0.465179652,-0.2999322116,0.0334029756,-0.0166977998,-0.0271500405,-0.0090621253,-0.0695152283,-0.3448971808,0.3211408854,-0.3420500457,0.0472643636,0.2058964074,0.026876729,-0.3330900371,0.2316471785,0.1748759449,-0.1278861016,-0.069692336,0.4330183864,-0.1066297889,-0.1177691668,-0.0149017265,-0.1021668166]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2301","title":"Unable to setup dev env on Windows","comments":"Hi @albertvillanova \r\n\r\nSorry for such a trivial issue ;-; \r\n\r\nThanks a lot.","body":"Hi\r\n\r\nI tried installing the `\".[dev]\"` version on Windows 10 after cloning.\r\n\r\nHere is the error I'm facing:\r\n\r\n```bat\r\n(env) C:\\testing\\datasets>pip install -e \".[dev]\"\r\nObtaining file:\/\/\/C:\/testing\/datasets\r\nRequirement already satisfied: numpy>=1.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.19.5)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-4.0.0-cp37-cp37m-win_amd64.whl (13.3 MB)\r\nRequirement already satisfied: dill in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.3.1.1)\r\nCollecting pandas\r\n Using cached pandas-1.2.4-cp37-cp37m-win_amd64.whl (9.1 MB)\r\nRequirement already satisfied: requests>=2.19.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2.25.1)\r\nRequirement already satisfied: tqdm<4.50.0,>=4.27 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.49.0)\r\nRequirement already satisfied: xxhash in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2.0.2)\r\nCollecting multiprocess\r\n Using cached multiprocess-0.70.11.1-py37-none-any.whl (108 kB)\r\nRequirement already satisfied: fsspec in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2021.4.0)\r\nCollecting huggingface_hub<0.1.0\r\n Using cached huggingface_hub-0.0.8-py3-none-any.whl (34 kB)\r\nRequirement already satisfied: importlib_metadata in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.0.1)\r\nRequirement already satisfied: absl-py in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.12.0)\r\nRequirement already satisfied: pytest in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (6.2.3)\r\nCollecting pytest-xdist\r\n Using cached pytest_xdist-2.2.1-py3-none-any.whl (37 kB)\r\nCollecting apache-beam>=2.24.0\r\n Using cached apache_beam-2.29.0-cp37-cp37m-win_amd64.whl (3.7 MB)\r\nCollecting elasticsearch\r\n Using cached elasticsearch-7.12.1-py2.py3-none-any.whl (339 kB)\r\nRequirement already satisfied: boto3==1.16.43 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.16.43)\r\nRequirement already satisfied: botocore==1.19.43 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.19.43)\r\nCollecting moto[s3]==1.3.16\r\n Using cached moto-1.3.16-py2.py3-none-any.whl (879 kB)\r\nCollecting rarfile>=4.0\r\n Using cached rarfile-4.0-py3-none-any.whl (28 kB)\r\nCollecting tensorflow>=2.3\r\n Using cached tensorflow-2.4.1-cp37-cp37m-win_amd64.whl (370.7 MB)\r\nRequirement already satisfied: torch in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.8.1)\r\nRequirement already satisfied: transformers in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.5.1)\r\nCollecting bs4\r\n Using cached bs4-0.0.1-py3-none-any.whl\r\nCollecting conllu\r\n Using cached conllu-4.4-py2.py3-none-any.whl (15 kB)\r\nCollecting langdetect\r\n Using cached langdetect-1.0.8-py3-none-any.whl\r\nCollecting lxml\r\n Using cached lxml-4.6.3-cp37-cp37m-win_amd64.whl (3.5 MB)\r\nCollecting mwparserfromhell\r\n Using cached mwparserfromhell-0.6-cp37-cp37m-win_amd64.whl (101 kB)\r\nCollecting nltk\r\n Using cached nltk-3.6.2-py3-none-any.whl (1.5 MB)\r\nCollecting openpyxl\r\n Using cached openpyxl-3.0.7-py2.py3-none-any.whl (243 kB)\r\nCollecting py7zr\r\n Using cached py7zr-0.15.2-py3-none-any.whl (66 kB)\r\nCollecting tldextract\r\n Using cached tldextract-3.1.0-py2.py3-none-any.whl (87 kB)\r\nCollecting zstandard\r\n Using cached zstandard-0.15.2-cp37-cp37m-win_amd64.whl (582 kB)\r\nCollecting bert_score>=0.3.6\r\n Using cached bert_score-0.3.9-py3-none-any.whl (59 kB)\r\nCollecting rouge_score\r\n Using cached rouge_score-0.0.4-py2.py3-none-any.whl (22 kB)\r\nCollecting sacrebleu\r\n Using cached sacrebleu-1.5.1-py3-none-any.whl (54 kB)\r\nRequirement already satisfied: scipy in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.6.3)\r\nCollecting seqeval\r\n Using cached seqeval-1.2.2-py3-none-any.whl\r\nCollecting sklearn\r\n Using cached sklearn-0.0-py2.py3-none-any.whl\r\nCollecting jiwer\r\n Using cached jiwer-2.2.0-py3-none-any.whl (13 kB)\r\nRequirement already satisfied: toml>=0.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.10.2)\r\nRequirement already satisfied: requests_file>=1.5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.5.1)\r\nRequirement already satisfied: texttable>=1.6.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.6.3)\r\nRequirement already satisfied: s3fs>=0.4.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.4.2)\r\nRequirement already satisfied: Werkzeug>=1.0.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.0.1)\r\nCollecting black\r\n Using cached black-21.4b2-py3-none-any.whl (130 kB)\r\nCollecting isort\r\n Using cached isort-5.8.0-py3-none-any.whl (103 kB)\r\nCollecting flake8==3.7.9\r\n Using cached flake8-3.7.9-py2.py3-none-any.whl (69 kB)\r\nRequirement already satisfied: jmespath<1.0.0,>=0.7.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.10.0)\r\nRequirement already satisfied: s3transfer<0.4.0,>=0.3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.3.7)\r\nRequirement already satisfied: urllib3<1.27,>=1.25.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (1.26.4)\r\nRequirement already satisfied: python-dateutil<3.0.0,>=2.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (2.8.1)\r\nCollecting entrypoints<0.4.0,>=0.3.0\r\n Using cached entrypoints-0.3-py2.py3-none-any.whl (11 kB)\r\nCollecting pyflakes<2.2.0,>=2.1.0\r\n Using cached pyflakes-2.1.1-py2.py3-none-any.whl (59 kB)\r\nCollecting pycodestyle<2.6.0,>=2.5.0\r\n Using cached pycodestyle-2.5.0-py2.py3-none-any.whl (51 kB)\r\nCollecting mccabe<0.7.0,>=0.6.0\r\n Using cached mccabe-0.6.1-py2.py3-none-any.whl (8.6 kB)\r\nRequirement already satisfied: jsondiff>=1.1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.3.0)\r\nRequirement already satisfied: pytz in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2021.1)\r\nRequirement already satisfied: mock in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.0.3)\r\nRequirement already satisfied: MarkupSafe<2.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.1.1)\r\nRequirement already satisfied: python-jose[cryptography]<4.0.0,>=3.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0)\r\nRequirement already satisfied: aws-xray-sdk!=0.96,>=0.93 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.8.0)\r\nRequirement already satisfied: cryptography>=2.3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.7)\r\nRequirement already satisfied: more-itertools in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (8.7.0)\r\nRequirement already satisfied: PyYAML>=5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.4.1)\r\nRequirement already satisfied: boto>=2.36.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.49.0)\r\nRequirement already satisfied: idna<3,>=2.5 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.10)\r\nRequirement already satisfied: sshpubkeys>=3.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.3.1)\r\nRequirement already satisfied: responses>=0.9.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.13.3)\r\nRequirement already satisfied: xmltodict in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.12.0)\r\nRequirement already satisfied: setuptools in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (52.0.0.post20210125)\r\nRequirement already satisfied: Jinja2>=2.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.11.3)\r\nRequirement already satisfied: zipp in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.1)\r\nRequirement already satisfied: six>1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.15.0)\r\nRequirement already satisfied: ecdsa<0.15 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.14.1)\r\nRequirement already satisfied: docker>=2.5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.0.0)\r\nRequirement already satisfied: cfn-lint>=0.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.49.0)\r\nRequirement already satisfied: grpcio<2,>=1.29.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (1.32.0)\r\nCollecting hdfs<3.0.0,>=2.1.0\r\n Using cached hdfs-2.6.0-py3-none-any.whl (33 kB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-3.0.0-cp37-cp37m-win_amd64.whl (12.6 MB)\r\nCollecting fastavro<2,>=0.21.4\r\n Using cached fastavro-1.4.0-cp37-cp37m-win_amd64.whl (394 kB)\r\nRequirement already satisfied: httplib2<0.18.0,>=0.8 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.17.4)\r\nCollecting pymongo<4.0.0,>=3.8.0\r\n Using cached pymongo-3.11.3-cp37-cp37m-win_amd64.whl (382 kB)\r\nCollecting crcmod<2.0,>=1.7\r\n Using cached crcmod-1.7-py3-none-any.whl\r\nCollecting avro-python3!=1.9.2,<1.10.0,>=1.8.1\r\n Using cached avro_python3-1.9.2.1-py3-none-any.whl\r\nRequirement already satisfied: typing-extensions<3.8.0,>=3.7.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.7.4.3)\r\nRequirement already satisfied: future<1.0.0,>=0.18.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.18.2)\r\nCollecting oauth2client<5,>=2.0.1\r\n Using cached oauth2client-4.1.3-py2.py3-none-any.whl (98 kB)\r\nCollecting pydot<2,>=1.2.0\r\n Using cached pydot-1.4.2-py2.py3-none-any.whl (21 kB)\r\nRequirement already satisfied: protobuf<4,>=3.12.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.15.8)\r\nRequirement already satisfied: wrapt in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from aws-xray-sdk!=0.96,>=0.93->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.12.1)\r\nCollecting matplotlib\r\n Using cached matplotlib-3.4.1-cp37-cp37m-win_amd64.whl (7.1 MB)\r\nRequirement already satisfied: junit-xml~=1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.9)\r\nRequirement already satisfied: jsonpatch in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.32)\r\nRequirement already satisfied: jsonschema~=3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0)\r\nRequirement already satisfied: networkx~=2.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.5.1)\r\nRequirement already satisfied: aws-sam-translator>=1.35.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.35.0)\r\nRequirement already satisfied: cffi>=1.12 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.14.5)\r\nRequirement already satisfied: pycparser in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cffi>=1.12->cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.20)\r\nRequirement already satisfied: pywin32==227 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (227)\r\nRequirement already satisfied: websocket-client>=0.32.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.58.0)\r\nRequirement already satisfied: docopt in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from hdfs<3.0.0,>=2.1.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.6.2)\r\nRequirement already satisfied: filelock in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from huggingface_hub<0.1.0->datasets==1.5.0.dev0) (3.0.12)\r\nRequirement already satisfied: pyrsistent>=0.14.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.17.3)\r\nRequirement already satisfied: attrs>=17.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (20.3.0)\r\nRequirement already satisfied: decorator<5,>=4.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from networkx~=2.4->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.4.2)\r\nRequirement already satisfied: rsa>=3.1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (4.7.2)\r\nRequirement already satisfied: pyasn1-modules>=0.0.5 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.2.8)\r\nRequirement already satisfied: pyasn1>=0.1.7 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.4.8)\r\nRequirement already satisfied: pyparsing>=2.1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pydot<2,>=1.2.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (2.4.7)\r\nRequirement already satisfied: certifi>=2017.4.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (2020.12.5)\r\nRequirement already satisfied: chardet<5,>=3.0.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (4.0.0)\r\nCollecting keras-preprocessing~=1.1.2\r\n Using cached Keras_Preprocessing-1.1.2-py2.py3-none-any.whl (42 kB)\r\nRequirement already satisfied: termcolor~=1.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (1.1.0)\r\nRequirement already satisfied: tensorboard~=2.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.5.0)\r\nRequirement already satisfied: wheel~=0.35 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (0.36.2)\r\nCollecting opt-einsum~=3.3.0\r\n Using cached opt_einsum-3.3.0-py3-none-any.whl (65 kB)\r\nCollecting gast==0.3.3\r\n Using cached gast-0.3.3-py2.py3-none-any.whl (9.7 kB)\r\nCollecting google-pasta~=0.2\r\n Using cached google_pasta-0.2.0-py3-none-any.whl (57 kB)\r\nRequirement already satisfied: tensorflow-estimator<2.5.0,>=2.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.4.0)\r\nCollecting astunparse~=1.6.3\r\n Using cached astunparse-1.6.3-py2.py3-none-any.whl (12 kB)\r\nCollecting flatbuffers~=1.12.0\r\n Using cached flatbuffers-1.12-py2.py3-none-any.whl (15 kB)\r\nCollecting h5py~=2.10.0\r\n Using cached h5py-2.10.0-cp37-cp37m-win_amd64.whl (2.5 MB)\r\nRequirement already satisfied: markdown>=2.6.8 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.3.4)\r\nRequirement already satisfied: tensorboard-plugin-wit>=1.6.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.8.0)\r\nRequirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.4.4)\r\nRequirement already satisfied: tensorboard-data-server<0.7.0,>=0.6.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.6.0)\r\nRequirement already satisfied: google-auth<2,>=1.6.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.30.0)\r\nRequirement already satisfied: cachetools<5.0,>=2.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from google-auth<2,>=1.6.3->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (4.2.2)\r\nRequirement already satisfied: requests-oauthlib>=0.7.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.3.0)\r\nRequirement already satisfied: oauthlib>=3.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.1.0)\r\nRequirement already satisfied: regex!=2019.12.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (2021.4.4)\r\nRequirement already satisfied: tokenizers<0.11,>=0.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (0.10.2)\r\nRequirement already satisfied: sacremoses in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (0.0.45)\r\nRequirement already satisfied: packaging in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (20.9)\r\nCollecting pathspec<1,>=0.8.1\r\n Using cached pathspec-0.8.1-py2.py3-none-any.whl (28 kB)\r\nRequirement already satisfied: click>=7.1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from black->datasets==1.5.0.dev0) (7.1.2)\r\nCollecting appdirs\r\n Using cached appdirs-1.4.4-py2.py3-none-any.whl (9.6 kB)\r\nCollecting mypy-extensions>=0.4.3\r\n Using cached mypy_extensions-0.4.3-py2.py3-none-any.whl (4.5 kB)\r\nRequirement already satisfied: typed-ast>=1.4.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from black->datasets==1.5.0.dev0) (1.4.3)\r\nCollecting beautifulsoup4\r\n Using cached beautifulsoup4-4.9.3-py3-none-any.whl (115 kB)\r\nRequirement already satisfied: soupsieve>1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from beautifulsoup4->bs4->datasets==1.5.0.dev0) (2.2.1)\r\nCollecting python-Levenshtein\r\n Using cached python-Levenshtein-0.12.2.tar.gz (50 kB)\r\nRequirement already satisfied: jsonpointer>=1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonpatch->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.1)\r\nRequirement already satisfied: pillow>=6.2.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (8.2.0)\r\nRequirement already satisfied: cycler>=0.10 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (0.10.0)\r\nRequirement already satisfied: kiwisolver>=1.0.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (1.3.1)\r\nCollecting multiprocess\r\n Using cached multiprocess-0.70.11-py3-none-any.whl (98 kB)\r\n Using cached multiprocess-0.70.10.zip (2.4 MB)\r\n Using cached multiprocess-0.70.9-py3-none-any.whl\r\nRequirement already satisfied: joblib in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from nltk->datasets==1.5.0.dev0) (1.0.1)\r\nCollecting et-xmlfile\r\n Using cached et_xmlfile-1.1.0-py3-none-any.whl (4.7 kB)\r\nRequirement already satisfied: pyzstd<0.15.0,>=0.14.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from py7zr->datasets==1.5.0.dev0) (0.14.4)\r\nCollecting pyppmd<0.13.0,>=0.12.1\r\n Using cached pyppmd-0.12.1-cp37-cp37m-win_amd64.whl (32 kB)\r\nCollecting pycryptodome>=3.6.6\r\n Using cached pycryptodome-3.10.1-cp35-abi3-win_amd64.whl (1.6 MB)\r\nCollecting bcj-cffi<0.6.0,>=0.5.1\r\n Using cached bcj_cffi-0.5.1-cp37-cp37m-win_amd64.whl (21 kB)\r\nCollecting multivolumefile<0.3.0,>=0.2.0\r\n Using cached multivolumefile-0.2.3-py3-none-any.whl (17 kB)\r\nRequirement already satisfied: iniconfig in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.1.1)\r\nRequirement already satisfied: py>=1.8.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.10.0)\r\nRequirement already satisfied: pluggy<1.0.0a1,>=0.12 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (0.13.1)\r\nRequirement already satisfied: atomicwrites>=1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.4.0)\r\nRequirement already satisfied: colorama in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (0.4.4)\r\nCollecting pytest-forked\r\n Using cached pytest_forked-1.3.0-py2.py3-none-any.whl (4.7 kB)\r\nCollecting execnet>=1.1\r\n Using cached execnet-1.8.0-py2.py3-none-any.whl (39 kB)\r\nRequirement already satisfied: apipkg>=1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from execnet>=1.1->pytest-xdist->datasets==1.5.0.dev0) (1.5)\r\nCollecting portalocker==2.0.0\r\n Using cached portalocker-2.0.0-py2.py3-none-any.whl (11 kB)\r\nRequirement already satisfied: scikit-learn>=0.21.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from seqeval->datasets==1.5.0.dev0) (0.24.2)\r\nRequirement already satisfied: threadpoolctl>=2.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from scikit-learn>=0.21.3->seqeval->datasets==1.5.0.dev0) (2.1.0)\r\nBuilding wheels for collected packages: python-Levenshtein\r\n Building wheel for python-Levenshtein (setup.py) ... error\r\n ERROR: Command errored out with exit status 1:\r\n command: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' bdist_wheel -d 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-wheel-8jh7fm18'\r\n cwd: C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\r\n Complete output (27 lines):\r\n running bdist_wheel\r\n running build\r\n running build_py\r\n creating build\r\n creating build\\lib.win-amd64-3.7\r\n creating build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\StringMatcher.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\__init__.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running egg_info\r\n writing python_Levenshtein.egg-info\\PKG-INFO\r\n writing dependency_links to python_Levenshtein.egg-info\\dependency_links.txt\r\n writing entry points to python_Levenshtein.egg-info\\entry_points.txt\r\n writing namespace_packages to python_Levenshtein.egg-info\\namespace_packages.txt\r\n writing requirements to python_Levenshtein.egg-info\\requires.txt\r\n writing top-level names to python_Levenshtein.egg-info\\top_level.txt\r\n reading manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n reading manifest template 'MANIFEST.in'\r\n warning: no previously-included files matching '*pyc' found anywhere in distribution\r\n warning: no previously-included files matching '*so' found anywhere in distribution\r\n warning: no previously-included files matching '.project' found anywhere in distribution\r\n warning: no previously-included files matching '.pydevproject' found anywhere in distribution\r\n writing manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n copying Levenshtein\\_levenshtein.c -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\_levenshtein.h -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running build_ext\r\n building 'Levenshtein._levenshtein' extension\r\n error: Microsoft Visual C++ 14.0 or greater is required. Get it with \"Microsoft C++ Build Tools\": https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/\r\n ----------------------------------------\r\n ERROR: Failed building wheel for python-Levenshtein\r\n Running setup.py clean for python-Levenshtein\r\nFailed to build python-Levenshtein\r\nInstalling collected packages: python-Levenshtein, pytest-forked, pyppmd, pymongo, pyflakes, pydot, pycryptodome, pycodestyle, pyarrow, portalocker, pathspec, pandas, opt-einsum, oauth2client, nltk, mypy-extensions, multivolumefile, multiprocess, moto, mccabe, matplotlib, keras-preprocessing, huggingface-hub, hdfs, h5py, google-pasta, gast, flatbuffers, fastavro, execnet, et-xmlfile, entrypoints, crcmod, beautifulsoup4, bcj-cffi, avro-python3, astunparse, appdirs, zstandard, tldextract, tensorflow, sklearn, seqeval, sacrebleu, rouge-score, rarfile, pytest-xdist, py7zr, openpyxl, mwparserfromhell, lxml, langdetect, jiwer, isort, flake8, elasticsearch, datasets, conllu, bs4, black, bert-score, apache-beam\r\n Running setup.py install for python-Levenshtein ... error\r\n ERROR: Command errored out with exit status 1:\r\n command: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' install --record 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-record-v7l7zitb\\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\\ProgramData\\Anaconda3\\envs\\env\\Include\\python-Levenshtein'\r\n cwd: C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\r\n Complete output (27 lines):\r\n running install\r\n running build\r\n running build_py\r\n creating build\r\n creating build\\lib.win-amd64-3.7\r\n creating build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\StringMatcher.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\__init__.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running egg_info\r\n writing python_Levenshtein.egg-info\\PKG-INFO\r\n writing dependency_links to python_Levenshtein.egg-info\\dependency_links.txt\r\n writing entry points to python_Levenshtein.egg-info\\entry_points.txt\r\n writing namespace_packages to python_Levenshtein.egg-info\\namespace_packages.txt\r\n writing requirements to python_Levenshtein.egg-info\\requires.txt\r\n writing top-level names to python_Levenshtein.egg-info\\top_level.txt\r\n reading manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n reading manifest template 'MANIFEST.in'\r\n warning: no previously-included files matching '*pyc' found anywhere in distribution\r\n warning: no previously-included files matching '*so' found anywhere in distribution\r\n warning: no previously-included files matching '.project' found anywhere in distribution\r\n warning: no previously-included files matching '.pydevproject' found anywhere in distribution\r\n writing manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n copying Levenshtein\\_levenshtein.c -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\_levenshtein.h -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running build_ext\r\n building 'Levenshtein._levenshtein' extension\r\n error: Microsoft Visual C++ 14.0 or greater is required. Get it with \"Microsoft C++ Build Tools\": https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/\r\n ----------------------------------------\r\nERROR: Command errored out with exit status 1: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' install --record 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-record-v7l7zitb\\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\\ProgramData\\Anaconda3\\envs\\env\\Include\\python-Levenshtein' Check the logs for full command output.\r\n```\r\n\r\nHere are conda and python versions:\r\n\r\n```bat\r\n(env) C:\\testing\\datasets>conda --version\r\nconda 4.9.2\r\n\r\n(env) C:\\testing\\datasets>python --version\r\nPython 3.7.10\r\n```\r\n\r\nPlease help me out. Thanks.","comment_length":12,"text":"Unable to setup dev env on Windows\nHi\r\n\r\nI tried installing the `\".[dev]\"` version on Windows 10 after cloning.\r\n\r\nHere is the error I'm facing:\r\n\r\n```bat\r\n(env) C:\\testing\\datasets>pip install -e \".[dev]\"\r\nObtaining file:\/\/\/C:\/testing\/datasets\r\nRequirement already satisfied: numpy>=1.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.19.5)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-4.0.0-cp37-cp37m-win_amd64.whl (13.3 MB)\r\nRequirement already satisfied: dill in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.3.1.1)\r\nCollecting pandas\r\n Using cached pandas-1.2.4-cp37-cp37m-win_amd64.whl (9.1 MB)\r\nRequirement already satisfied: requests>=2.19.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2.25.1)\r\nRequirement already satisfied: tqdm<4.50.0,>=4.27 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.49.0)\r\nRequirement already satisfied: xxhash in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2.0.2)\r\nCollecting multiprocess\r\n Using cached multiprocess-0.70.11.1-py37-none-any.whl (108 kB)\r\nRequirement already satisfied: fsspec in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (2021.4.0)\r\nCollecting huggingface_hub<0.1.0\r\n Using cached huggingface_hub-0.0.8-py3-none-any.whl (34 kB)\r\nRequirement already satisfied: importlib_metadata in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.0.1)\r\nRequirement already satisfied: absl-py in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.12.0)\r\nRequirement already satisfied: pytest in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (6.2.3)\r\nCollecting pytest-xdist\r\n Using cached pytest_xdist-2.2.1-py3-none-any.whl (37 kB)\r\nCollecting apache-beam>=2.24.0\r\n Using cached apache_beam-2.29.0-cp37-cp37m-win_amd64.whl (3.7 MB)\r\nCollecting elasticsearch\r\n Using cached elasticsearch-7.12.1-py2.py3-none-any.whl (339 kB)\r\nRequirement already satisfied: boto3==1.16.43 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.16.43)\r\nRequirement already satisfied: botocore==1.19.43 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.19.43)\r\nCollecting moto[s3]==1.3.16\r\n Using cached moto-1.3.16-py2.py3-none-any.whl (879 kB)\r\nCollecting rarfile>=4.0\r\n Using cached rarfile-4.0-py3-none-any.whl (28 kB)\r\nCollecting tensorflow>=2.3\r\n Using cached tensorflow-2.4.1-cp37-cp37m-win_amd64.whl (370.7 MB)\r\nRequirement already satisfied: torch in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.8.1)\r\nRequirement already satisfied: transformers in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (4.5.1)\r\nCollecting bs4\r\n Using cached bs4-0.0.1-py3-none-any.whl\r\nCollecting conllu\r\n Using cached conllu-4.4-py2.py3-none-any.whl (15 kB)\r\nCollecting langdetect\r\n Using cached langdetect-1.0.8-py3-none-any.whl\r\nCollecting lxml\r\n Using cached lxml-4.6.3-cp37-cp37m-win_amd64.whl (3.5 MB)\r\nCollecting mwparserfromhell\r\n Using cached mwparserfromhell-0.6-cp37-cp37m-win_amd64.whl (101 kB)\r\nCollecting nltk\r\n Using cached nltk-3.6.2-py3-none-any.whl (1.5 MB)\r\nCollecting openpyxl\r\n Using cached openpyxl-3.0.7-py2.py3-none-any.whl (243 kB)\r\nCollecting py7zr\r\n Using cached py7zr-0.15.2-py3-none-any.whl (66 kB)\r\nCollecting tldextract\r\n Using cached tldextract-3.1.0-py2.py3-none-any.whl (87 kB)\r\nCollecting zstandard\r\n Using cached zstandard-0.15.2-cp37-cp37m-win_amd64.whl (582 kB)\r\nCollecting bert_score>=0.3.6\r\n Using cached bert_score-0.3.9-py3-none-any.whl (59 kB)\r\nCollecting rouge_score\r\n Using cached rouge_score-0.0.4-py2.py3-none-any.whl (22 kB)\r\nCollecting sacrebleu\r\n Using cached sacrebleu-1.5.1-py3-none-any.whl (54 kB)\r\nRequirement already satisfied: scipy in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.6.3)\r\nCollecting seqeval\r\n Using cached seqeval-1.2.2-py3-none-any.whl\r\nCollecting sklearn\r\n Using cached sklearn-0.0-py2.py3-none-any.whl\r\nCollecting jiwer\r\n Using cached jiwer-2.2.0-py3-none-any.whl (13 kB)\r\nRequirement already satisfied: toml>=0.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.10.2)\r\nRequirement already satisfied: requests_file>=1.5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.5.1)\r\nRequirement already satisfied: texttable>=1.6.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.6.3)\r\nRequirement already satisfied: s3fs>=0.4.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (0.4.2)\r\nRequirement already satisfied: Werkzeug>=1.0.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from datasets==1.5.0.dev0) (1.0.1)\r\nCollecting black\r\n Using cached black-21.4b2-py3-none-any.whl (130 kB)\r\nCollecting isort\r\n Using cached isort-5.8.0-py3-none-any.whl (103 kB)\r\nCollecting flake8==3.7.9\r\n Using cached flake8-3.7.9-py2.py3-none-any.whl (69 kB)\r\nRequirement already satisfied: jmespath<1.0.0,>=0.7.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.10.0)\r\nRequirement already satisfied: s3transfer<0.4.0,>=0.3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.3.7)\r\nRequirement already satisfied: urllib3<1.27,>=1.25.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (1.26.4)\r\nRequirement already satisfied: python-dateutil<3.0.0,>=2.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (2.8.1)\r\nCollecting entrypoints<0.4.0,>=0.3.0\r\n Using cached entrypoints-0.3-py2.py3-none-any.whl (11 kB)\r\nCollecting pyflakes<2.2.0,>=2.1.0\r\n Using cached pyflakes-2.1.1-py2.py3-none-any.whl (59 kB)\r\nCollecting pycodestyle<2.6.0,>=2.5.0\r\n Using cached pycodestyle-2.5.0-py2.py3-none-any.whl (51 kB)\r\nCollecting mccabe<0.7.0,>=0.6.0\r\n Using cached mccabe-0.6.1-py2.py3-none-any.whl (8.6 kB)\r\nRequirement already satisfied: jsondiff>=1.1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.3.0)\r\nRequirement already satisfied: pytz in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2021.1)\r\nRequirement already satisfied: mock in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.0.3)\r\nRequirement already satisfied: MarkupSafe<2.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.1.1)\r\nRequirement already satisfied: python-jose[cryptography]<4.0.0,>=3.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0)\r\nRequirement already satisfied: aws-xray-sdk!=0.96,>=0.93 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.8.0)\r\nRequirement already satisfied: cryptography>=2.3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.7)\r\nRequirement already satisfied: more-itertools in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (8.7.0)\r\nRequirement already satisfied: PyYAML>=5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.4.1)\r\nRequirement already satisfied: boto>=2.36.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.49.0)\r\nRequirement already satisfied: idna<3,>=2.5 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.10)\r\nRequirement already satisfied: sshpubkeys>=3.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.3.1)\r\nRequirement already satisfied: responses>=0.9.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.13.3)\r\nRequirement already satisfied: xmltodict in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.12.0)\r\nRequirement already satisfied: setuptools in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (52.0.0.post20210125)\r\nRequirement already satisfied: Jinja2>=2.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.11.3)\r\nRequirement already satisfied: zipp in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.1)\r\nRequirement already satisfied: six>1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.15.0)\r\nRequirement already satisfied: ecdsa<0.15 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.14.1)\r\nRequirement already satisfied: docker>=2.5.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.0.0)\r\nRequirement already satisfied: cfn-lint>=0.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.49.0)\r\nRequirement already satisfied: grpcio<2,>=1.29.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (1.32.0)\r\nCollecting hdfs<3.0.0,>=2.1.0\r\n Using cached hdfs-2.6.0-py3-none-any.whl (33 kB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-3.0.0-cp37-cp37m-win_amd64.whl (12.6 MB)\r\nCollecting fastavro<2,>=0.21.4\r\n Using cached fastavro-1.4.0-cp37-cp37m-win_amd64.whl (394 kB)\r\nRequirement already satisfied: httplib2<0.18.0,>=0.8 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.17.4)\r\nCollecting pymongo<4.0.0,>=3.8.0\r\n Using cached pymongo-3.11.3-cp37-cp37m-win_amd64.whl (382 kB)\r\nCollecting crcmod<2.0,>=1.7\r\n Using cached crcmod-1.7-py3-none-any.whl\r\nCollecting avro-python3!=1.9.2,<1.10.0,>=1.8.1\r\n Using cached avro_python3-1.9.2.1-py3-none-any.whl\r\nRequirement already satisfied: typing-extensions<3.8.0,>=3.7.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.7.4.3)\r\nRequirement already satisfied: future<1.0.0,>=0.18.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.18.2)\r\nCollecting oauth2client<5,>=2.0.1\r\n Using cached oauth2client-4.1.3-py2.py3-none-any.whl (98 kB)\r\nCollecting pydot<2,>=1.2.0\r\n Using cached pydot-1.4.2-py2.py3-none-any.whl (21 kB)\r\nRequirement already satisfied: protobuf<4,>=3.12.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.15.8)\r\nRequirement already satisfied: wrapt in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from aws-xray-sdk!=0.96,>=0.93->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.12.1)\r\nCollecting matplotlib\r\n Using cached matplotlib-3.4.1-cp37-cp37m-win_amd64.whl (7.1 MB)\r\nRequirement already satisfied: junit-xml~=1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.9)\r\nRequirement already satisfied: jsonpatch in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.32)\r\nRequirement already satisfied: jsonschema~=3.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0)\r\nRequirement already satisfied: networkx~=2.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.5.1)\r\nRequirement already satisfied: aws-sam-translator>=1.35.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.35.0)\r\nRequirement already satisfied: cffi>=1.12 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.14.5)\r\nRequirement already satisfied: pycparser in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from cffi>=1.12->cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.20)\r\nRequirement already satisfied: pywin32==227 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (227)\r\nRequirement already satisfied: websocket-client>=0.32.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.58.0)\r\nRequirement already satisfied: docopt in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from hdfs<3.0.0,>=2.1.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.6.2)\r\nRequirement already satisfied: filelock in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from huggingface_hub<0.1.0->datasets==1.5.0.dev0) (3.0.12)\r\nRequirement already satisfied: pyrsistent>=0.14.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.17.3)\r\nRequirement already satisfied: attrs>=17.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (20.3.0)\r\nRequirement already satisfied: decorator<5,>=4.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from networkx~=2.4->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.4.2)\r\nRequirement already satisfied: rsa>=3.1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (4.7.2)\r\nRequirement already satisfied: pyasn1-modules>=0.0.5 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.2.8)\r\nRequirement already satisfied: pyasn1>=0.1.7 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.4.8)\r\nRequirement already satisfied: pyparsing>=2.1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pydot<2,>=1.2.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (2.4.7)\r\nRequirement already satisfied: certifi>=2017.4.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (2020.12.5)\r\nRequirement already satisfied: chardet<5,>=3.0.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (4.0.0)\r\nCollecting keras-preprocessing~=1.1.2\r\n Using cached Keras_Preprocessing-1.1.2-py2.py3-none-any.whl (42 kB)\r\nRequirement already satisfied: termcolor~=1.1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (1.1.0)\r\nRequirement already satisfied: tensorboard~=2.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.5.0)\r\nRequirement already satisfied: wheel~=0.35 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (0.36.2)\r\nCollecting opt-einsum~=3.3.0\r\n Using cached opt_einsum-3.3.0-py3-none-any.whl (65 kB)\r\nCollecting gast==0.3.3\r\n Using cached gast-0.3.3-py2.py3-none-any.whl (9.7 kB)\r\nCollecting google-pasta~=0.2\r\n Using cached google_pasta-0.2.0-py3-none-any.whl (57 kB)\r\nRequirement already satisfied: tensorflow-estimator<2.5.0,>=2.4.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.4.0)\r\nCollecting astunparse~=1.6.3\r\n Using cached astunparse-1.6.3-py2.py3-none-any.whl (12 kB)\r\nCollecting flatbuffers~=1.12.0\r\n Using cached flatbuffers-1.12-py2.py3-none-any.whl (15 kB)\r\nCollecting h5py~=2.10.0\r\n Using cached h5py-2.10.0-cp37-cp37m-win_amd64.whl (2.5 MB)\r\nRequirement already satisfied: markdown>=2.6.8 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.3.4)\r\nRequirement already satisfied: tensorboard-plugin-wit>=1.6.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.8.0)\r\nRequirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.4.4)\r\nRequirement already satisfied: tensorboard-data-server<0.7.0,>=0.6.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.6.0)\r\nRequirement already satisfied: google-auth<2,>=1.6.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.30.0)\r\nRequirement already satisfied: cachetools<5.0,>=2.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from google-auth<2,>=1.6.3->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (4.2.2)\r\nRequirement already satisfied: requests-oauthlib>=0.7.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.3.0)\r\nRequirement already satisfied: oauthlib>=3.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.1.0)\r\nRequirement already satisfied: regex!=2019.12.17 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (2021.4.4)\r\nRequirement already satisfied: tokenizers<0.11,>=0.10.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (0.10.2)\r\nRequirement already satisfied: sacremoses in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (0.0.45)\r\nRequirement already satisfied: packaging in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from transformers->datasets==1.5.0.dev0) (20.9)\r\nCollecting pathspec<1,>=0.8.1\r\n Using cached pathspec-0.8.1-py2.py3-none-any.whl (28 kB)\r\nRequirement already satisfied: click>=7.1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from black->datasets==1.5.0.dev0) (7.1.2)\r\nCollecting appdirs\r\n Using cached appdirs-1.4.4-py2.py3-none-any.whl (9.6 kB)\r\nCollecting mypy-extensions>=0.4.3\r\n Using cached mypy_extensions-0.4.3-py2.py3-none-any.whl (4.5 kB)\r\nRequirement already satisfied: typed-ast>=1.4.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from black->datasets==1.5.0.dev0) (1.4.3)\r\nCollecting beautifulsoup4\r\n Using cached beautifulsoup4-4.9.3-py3-none-any.whl (115 kB)\r\nRequirement already satisfied: soupsieve>1.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from beautifulsoup4->bs4->datasets==1.5.0.dev0) (2.2.1)\r\nCollecting python-Levenshtein\r\n Using cached python-Levenshtein-0.12.2.tar.gz (50 kB)\r\nRequirement already satisfied: jsonpointer>=1.9 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from jsonpatch->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.1)\r\nRequirement already satisfied: pillow>=6.2.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (8.2.0)\r\nRequirement already satisfied: cycler>=0.10 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (0.10.0)\r\nRequirement already satisfied: kiwisolver>=1.0.1 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (1.3.1)\r\nCollecting multiprocess\r\n Using cached multiprocess-0.70.11-py3-none-any.whl (98 kB)\r\n Using cached multiprocess-0.70.10.zip (2.4 MB)\r\n Using cached multiprocess-0.70.9-py3-none-any.whl\r\nRequirement already satisfied: joblib in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from nltk->datasets==1.5.0.dev0) (1.0.1)\r\nCollecting et-xmlfile\r\n Using cached et_xmlfile-1.1.0-py3-none-any.whl (4.7 kB)\r\nRequirement already satisfied: pyzstd<0.15.0,>=0.14.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from py7zr->datasets==1.5.0.dev0) (0.14.4)\r\nCollecting pyppmd<0.13.0,>=0.12.1\r\n Using cached pyppmd-0.12.1-cp37-cp37m-win_amd64.whl (32 kB)\r\nCollecting pycryptodome>=3.6.6\r\n Using cached pycryptodome-3.10.1-cp35-abi3-win_amd64.whl (1.6 MB)\r\nCollecting bcj-cffi<0.6.0,>=0.5.1\r\n Using cached bcj_cffi-0.5.1-cp37-cp37m-win_amd64.whl (21 kB)\r\nCollecting multivolumefile<0.3.0,>=0.2.0\r\n Using cached multivolumefile-0.2.3-py3-none-any.whl (17 kB)\r\nRequirement already satisfied: iniconfig in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.1.1)\r\nRequirement already satisfied: py>=1.8.2 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.10.0)\r\nRequirement already satisfied: pluggy<1.0.0a1,>=0.12 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (0.13.1)\r\nRequirement already satisfied: atomicwrites>=1.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (1.4.0)\r\nRequirement already satisfied: colorama in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from pytest->datasets==1.5.0.dev0) (0.4.4)\r\nCollecting pytest-forked\r\n Using cached pytest_forked-1.3.0-py2.py3-none-any.whl (4.7 kB)\r\nCollecting execnet>=1.1\r\n Using cached execnet-1.8.0-py2.py3-none-any.whl (39 kB)\r\nRequirement already satisfied: apipkg>=1.4 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from execnet>=1.1->pytest-xdist->datasets==1.5.0.dev0) (1.5)\r\nCollecting portalocker==2.0.0\r\n Using cached portalocker-2.0.0-py2.py3-none-any.whl (11 kB)\r\nRequirement already satisfied: scikit-learn>=0.21.3 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from seqeval->datasets==1.5.0.dev0) (0.24.2)\r\nRequirement already satisfied: threadpoolctl>=2.0.0 in c:\\programdata\\anaconda3\\envs\\env\\lib\\site-packages (from scikit-learn>=0.21.3->seqeval->datasets==1.5.0.dev0) (2.1.0)\r\nBuilding wheels for collected packages: python-Levenshtein\r\n Building wheel for python-Levenshtein (setup.py) ... error\r\n ERROR: Command errored out with exit status 1:\r\n command: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' bdist_wheel -d 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-wheel-8jh7fm18'\r\n cwd: C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\r\n Complete output (27 lines):\r\n running bdist_wheel\r\n running build\r\n running build_py\r\n creating build\r\n creating build\\lib.win-amd64-3.7\r\n creating build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\StringMatcher.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\__init__.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running egg_info\r\n writing python_Levenshtein.egg-info\\PKG-INFO\r\n writing dependency_links to python_Levenshtein.egg-info\\dependency_links.txt\r\n writing entry points to python_Levenshtein.egg-info\\entry_points.txt\r\n writing namespace_packages to python_Levenshtein.egg-info\\namespace_packages.txt\r\n writing requirements to python_Levenshtein.egg-info\\requires.txt\r\n writing top-level names to python_Levenshtein.egg-info\\top_level.txt\r\n reading manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n reading manifest template 'MANIFEST.in'\r\n warning: no previously-included files matching '*pyc' found anywhere in distribution\r\n warning: no previously-included files matching '*so' found anywhere in distribution\r\n warning: no previously-included files matching '.project' found anywhere in distribution\r\n warning: no previously-included files matching '.pydevproject' found anywhere in distribution\r\n writing manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n copying Levenshtein\\_levenshtein.c -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\_levenshtein.h -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running build_ext\r\n building 'Levenshtein._levenshtein' extension\r\n error: Microsoft Visual C++ 14.0 or greater is required. Get it with \"Microsoft C++ Build Tools\": https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/\r\n ----------------------------------------\r\n ERROR: Failed building wheel for python-Levenshtein\r\n Running setup.py clean for python-Levenshtein\r\nFailed to build python-Levenshtein\r\nInstalling collected packages: python-Levenshtein, pytest-forked, pyppmd, pymongo, pyflakes, pydot, pycryptodome, pycodestyle, pyarrow, portalocker, pathspec, pandas, opt-einsum, oauth2client, nltk, mypy-extensions, multivolumefile, multiprocess, moto, mccabe, matplotlib, keras-preprocessing, huggingface-hub, hdfs, h5py, google-pasta, gast, flatbuffers, fastavro, execnet, et-xmlfile, entrypoints, crcmod, beautifulsoup4, bcj-cffi, avro-python3, astunparse, appdirs, zstandard, tldextract, tensorflow, sklearn, seqeval, sacrebleu, rouge-score, rarfile, pytest-xdist, py7zr, openpyxl, mwparserfromhell, lxml, langdetect, jiwer, isort, flake8, elasticsearch, datasets, conllu, bs4, black, bert-score, apache-beam\r\n Running setup.py install for python-Levenshtein ... error\r\n ERROR: Command errored out with exit status 1:\r\n command: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' install --record 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-record-v7l7zitb\\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\\ProgramData\\Anaconda3\\envs\\env\\Include\\python-Levenshtein'\r\n cwd: C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\r\n Complete output (27 lines):\r\n running install\r\n running build\r\n running build_py\r\n creating build\r\n creating build\\lib.win-amd64-3.7\r\n creating build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\StringMatcher.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\__init__.py -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running egg_info\r\n writing python_Levenshtein.egg-info\\PKG-INFO\r\n writing dependency_links to python_Levenshtein.egg-info\\dependency_links.txt\r\n writing entry points to python_Levenshtein.egg-info\\entry_points.txt\r\n writing namespace_packages to python_Levenshtein.egg-info\\namespace_packages.txt\r\n writing requirements to python_Levenshtein.egg-info\\requires.txt\r\n writing top-level names to python_Levenshtein.egg-info\\top_level.txt\r\n reading manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n reading manifest template 'MANIFEST.in'\r\n warning: no previously-included files matching '*pyc' found anywhere in distribution\r\n warning: no previously-included files matching '*so' found anywhere in distribution\r\n warning: no previously-included files matching '.project' found anywhere in distribution\r\n warning: no previously-included files matching '.pydevproject' found anywhere in distribution\r\n writing manifest file 'python_Levenshtein.egg-info\\SOURCES.txt'\r\n copying Levenshtein\\_levenshtein.c -> build\\lib.win-amd64-3.7\\Levenshtein\r\n copying Levenshtein\\_levenshtein.h -> build\\lib.win-amd64-3.7\\Levenshtein\r\n running build_ext\r\n building 'Levenshtein._levenshtein' extension\r\n error: Microsoft Visual C++ 14.0 or greater is required. Get it with \"Microsoft C++ Build Tools\": https:\/\/visualstudio.microsoft.com\/visual-cpp-build-tools\/\r\n ----------------------------------------\r\nERROR: Command errored out with exit status 1: 'C:\\ProgramData\\Anaconda3\\envs\\env\\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"'; __file__='\"'\"'C:\\\\Users\\\\VKC~1\\\\AppData\\\\Local\\\\Temp\\\\pip-install-ynt_dbm4\\\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\\\setup.py'\"'\"';f=getattr(tokenize, '\"'\"'open'\"'\"', open)(__file__);code=f.read().replace('\"'\"'\\r\\n'\"'\"', '\"'\"'\\n'\"'\"');f.close();exec(compile(code, __file__, '\"'\"'exec'\"'\"'))' install --record 'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-record-v7l7zitb\\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\\ProgramData\\Anaconda3\\envs\\env\\Include\\python-Levenshtein' Check the logs for full command output.\r\n```\r\n\r\nHere are conda and python versions:\r\n\r\n```bat\r\n(env) C:\\testing\\datasets>conda --version\r\nconda 4.9.2\r\n\r\n(env) C:\\testing\\datasets>python --version\r\nPython 3.7.10\r\n```\r\n\r\nPlease help me out. Thanks.\nHi @albertvillanova \r\n\r\nSorry for such a trivial issue ;-; \r\n\r\nThanks a lot.","embeddings":[-0.4121342301,-0.0855779797,-0.0977274776,-0.1387811452,0.3114477396,0.068374902,0.3448515236,0.0465872362,-0.1001304686,0.15623492,0.0433793887,0.2424268872,0.1581842005,0.2162972242,0.0545784347,0.0232140273,0.1360678375,0.4390758872,-0.2185967714,0.037439011,-0.0659957454,-0.0208863225,-0.1535535008,-0.0436114408,0.3134841323,-0.0264739394,-0.2292296886,0.1753225178,-0.3534286618,-0.3225588799,0.4030125737,-0.1064584628,0.4329041839,0.1327629685,-0.0001171767,0.2016099095,0.3223023117,-0.1553241611,-0.0909361094,0.0166196972,-0.0741046444,-0.4187884331,-0.0570751615,-0.1152198911,0.0013719843,-0.315310955,0.1467733234,-0.3042214811,0.0935275629,0.1765608937,0.1489830762,0.0158207696,0.3976289928,-0.0599443838,-0.0897259861,0.2228936255,-0.1047845185,0.2224471718,0.3808677495,-0.0767932087,0.2623265088,-0.0150638493,0.2122566849,0.0888135731,0.1605111212,0.0911397189,0.8259810209,-0.3334536552,-0.0124987392,-0.0583257899,0.4455650151,-0.1256540567,-0.2552556992,0.2472820431,0.1414016038,-0.325605303,0.261488229,0.3790609241,-0.3644766808,0.0435973518,-0.362721324,-0.2574132085,-0.4395217001,0.0149611542,0.1180061549,0.0826293603,0.1625450402,0.1449573189,-0.0660314187,0.0814976096,0.4332493246,0.0780942217,-0.1706448048,0.0194142777,-0.1333711296,-0.1515952498,-0.2876423001,0.3090680242,-0.0119880065,0.0120718013,-0.3012442887,-0.2072132975,0.2847516239,0.0740561634,-0.0211096983,0.3143170476,0.0671262816,-0.0794963837,0.1001376882,-0.1034927294,0.4024324119,0.0954257473,-0.220410049,-0.4841091931,-0.1915215552,0.502032876,0.0810901076,-0.1381357461,-0.2252849638,-0.0552830324,-0.278065145,0.1433484256,0.2262274027,0.247914955,-0.0905870795,-0.0975939855,0.0245632939,0.2494825125,-0.2053817958,0.111318931,0.1260858774,0.1531590074,-0.4578863382,-0.3628825843,0.2566353381,-0.2950874269,0.3838174641,0.2732485831,-0.0795328692,0.2728429437,-0.0912644342,0.190502882,0.1253395081,0.2146690041,0.192412436,0.1868555695,-0.0017933784,-0.2911956906,-0.0418691561,0.2853320539,0.0168284662,-0.2246380895,-0.3471564054,0.1082542017,-0.1117702425,0.017555505,0.0353535265,-0.0690591186,0.0887759179,0.1203753427,-0.0526402183,-0.1925052106,0.0195803028,-0.1289805025,0.3793908656,0.0687448531,-0.4063456953,0.2841972411,0.077341266,-0.1427752674,0.1688283831,0.032434091,-0.0016664289,0.1482707858,-0.2415356338,-0.2094384283,-0.1005012766,-0.4556683302,-0.1719743758,0.3552168608,-0.0029001641,-0.4598074555,0.2365486175,-0.120121628,0.0676399097,-0.1652117074,-0.088309899,0.2276965827,-0.035136126,-0.2145301253,-0.168868646,-0.2638670802,0.1783894897,-0.0220463779,-0.0101095643,-0.2324422002,-0.0198157672,-0.1845362335,-0.0388986133,-0.1567320377,0.2039210051,0.381216675,0.5579845905,-0.0816559568,-0.0684605464,0.0482342988,-0.1448994875,0.159021914,-0.4344110489,0.107544288,-0.0217336845,0.149328202,-0.0466755331,0.0853103474,-0.4496676028,-0.3711802661,0.116788134,0.0079455134,0.3306221664,-0.0743468627,0.1798881143,0.0693052262,0.0324366949,0.1059975326,-0.166411072,0.1343197078,-0.7459347248,-0.0168207139,-0.1043569595,0.1043144912,0.0355914086,-0.0769519955,0.0382833369,0.1590626389,-0.1584957093,-0.3072966635,-0.041364219,-0.3169023693,0.2125235647,-0.5566606522,0.2742674053,0.1649557948,0.0273893774,0.1721837968,-0.0810752809,-0.1246714965,-0.2803698778,-0.0341622308,-0.0767948776,0.297168076,0.3042367697,0.0494527295,0.0207218789,-0.1933784485,0.4338948429,0.575902164,0.0313213542,-0.3306666613,0.2104430646,0.0414407551,-0.1986906677,-0.0196556337,-0.142107293,-0.0658592954,-0.0807882249,0.0347912349,0.1587107182,-0.0406574272,0.5600038171,0.1366593093,-0.0128204282,0.2236149907,-0.2414536029,-0.0290230308,0.4613452852,-0.0718432441,-0.0432940163,0.1160499156,-0.3231308162,-0.1007453874,-0.0150394756,-0.2076589614,0.4360066056,0.3666128814,-0.414539963,-0.0872983187,-0.4251069129,0.0179413594,-0.3845640123,-0.5058633089,-0.1485284567,0.072969012,0.2087071687,-0.1159329861,0.2937576473,0.378765434,0.1709450483,-0.1817639172,-0.0483662486,-0.0111401537,-0.0018849734,-0.0123814195,-0.2988424599,0.0734532699,0.352761358,0.1322950125,0.4342385232,0.0266307555,0.0442592315,-0.1747365743,-0.3875649571,-0.0144538507,-0.1596354544,0.7915337682,0.1893676966,0.1281995326,-0.0219017398,-0.2373072803,0.2668535709,0.046285782,-0.2804477513,0.0047871685,-0.2273965627,-0.4810552001,-0.180219084,-0.3516542315,-0.1246601343,-0.0893754438,-0.107180655,-0.1341093034,-0.0714597255,0.4129341841,-0.0137822358,-0.0348017626,0.2470746636,-0.0075236536,-0.0635410845,0.111187987,0.2345679998,-0.4932157695,-0.4311946332,0.2324139327,0.0099825086,0.1128880084,0.1875332743,-0.2512127757,-0.1148042381,-0.1031353101,0.0084202457,-0.0645647049,0.2379448414,0.4715803862,0.2665568888,0.0352040157,-0.0425962657,-0.2322521061,0.1093985289,-0.0791745335,0.4247240722,0.2350768745,0.577147603,0.0135583226,0.4463897347,0.7228881121,0.0914104655,0.2535391748,-0.0413708538,0.3796830773,0.0905267894,-0.3451045454,0.2690578103,0.0630212799,-0.0072042472,0.1809092909,-0.0321068391,-0.1778542548,-0.2200647742,0.0465762503,-0.0596503131,0.0298709534,-0.1174279228,-0.3436402082,0.3009181321,-0.2163112164,0.1664456427,0.132733658,-0.0476562269,0.0549124517,0.462567538,-0.2057617009,0.2380177081,-0.4431191385,-0.55777812,-0.3410708904,0.2160663605,-0.0173758063,0.6229293942,-0.2246486992,0.4463389814,0.1381700486,-0.2857893705,0.4693852663,-0.4746426642,0.0524854362,-0.0733335987,0.0820763633,-0.3210382462,-0.2203683704,0.0427084416,0.1718384922,-0.0350638144,-0.1515042037,-0.1922936738,0.2373284698,0.2754104733,0.0534360409,-0.1614242047,-0.0936392322,-0.2664470971,-0.4962081015,-0.0505326651,-0.0729672164,0.0656276122,0.0504231825,0.2158268392,0.2125395685,-0.062741369,0.0499402396,0.1299736947,-0.002951686,0.0127464049,-0.0393292047,0.277925998,-0.0897330716,0.242069453,0.0891641825,0.2664802969,-0.214750424,-0.1888166517,-0.0411368757,-0.1824876517,0.4089705944,0.0675264001,-0.1009932458,-0.3996094167,-0.0261758715,-0.1057991907,-0.2291745842,-0.1686869115,0.2476257235,0.0220554098,-0.3125375807,-0.1065755039,0.0946872979,-0.0872679502,0.126850307,0.3581095934,0.2488824427,-0.2112712711,-0.0633669794,0.3672279716,0.6484234333,0.0318691768,-0.1837054044,0.0639485046,-0.5025157928,0.127110377,0.0638693869,0.1315517128,-0.1698898226,-0.1872195154,0.0249396823,-0.1965131164,0.324855268,0.0538668744,-0.4849865437,0.1431676745,0.4321363568,0.123522751,0.1298967749,0.103631869,0.0593833961,-0.1762978286,0.1346544772,0.1409142166,0.0725384578,0.1524269283,-0.1616719067,0.0688693598,-0.1432239711,-0.2424301654,-0.6267002225,0.112472631,-0.2296545804,0.277250886,-0.2013393044,-0.608169198,0.4104798436,0.3885934055,0.569214344,0.2539714575,-0.2626110911,0.0875652134,-0.0031408167,-0.3031441569,0.1424965262,-0.0968290716,0.0821555853,-0.2844583094,-0.3339810669,-0.1550788134,-0.06838236,0.0433429256,-0.3456445634,-0.0371471792,-0.1009884551,0.0593316406,0.2151547074,-0.12476971,0.2778620124,0.0004334133,0.0718653649,-0.0310869589,-0.0376362354,0.3993119001,-0.2985121012,-0.157242015,-0.0628597215,0.0892680213,-0.1416343004,-0.097849831,0.3126449585,0.0009065954,-0.1707309037,-0.1474594772,0.0576615967,-0.2658415139,-0.2694555819,-0.4231140912,0.154992193,0.0106100775,0.1373024285,-0.1414523721,0.1642699391,-0.2475146353,-0.0291168727,-0.3162648976,-0.1858838648,-0.0911561549,-0.2300188392,0.2858869433,0.1424523741,0.2312096506,0.2592296898,-0.0326089486,-0.2188580483,0.001480481,0.2834210098,0.1979220957,-0.0843515694,-0.3264656067,-0.0342857167,-0.0405877642,0.0829677805,-0.2389823347,-0.1508959979,-0.1201620698,0.1044125929,0.1108317301,0.0316658281,-0.2158530653,0.1758563519,-0.345400095,-0.3830916286,0.03885803,-0.0019292964,0.152979821,-0.1848097295,0.3930841386,0.045836769,-0.0579165481,-0.2661431134,0.0873562098,-0.2104155123,0.5279401541,0.0870263129,0.2425913066,0.0917045474,-0.0350838788,-0.6708033085,0.2103807628,0.0831123367,0.0742151961,-0.0349078625,-0.1580380648,0.1837056726,-0.221336782,0.2976179719,0.3265206516,-0.0052294247,0.1880482137,0.54577142,0.1286130399,-0.237014696,0.1696367413,0.1250644326,-0.078648977,-0.1110885218,0.1126404256,0.2368839383,-0.2054820806,0.3799754977,-0.0941831097,0.4906442463,-0.2079764307,0.004008227,0.5395568609,0.0277876202,0.0393729135,-0.0421883427,0.0318805948,0.1792411208,0.583144486,0.0031546124,0.0298069771,0.140637517,0.3013638258,-0.1679567248,-0.0243418384,-0.1806923449,-0.2011103779,0.2749938071,0.0183627587,0.2508787215,0.6245895624,-0.4090289176,-0.061805658,-0.5428097844,0.2052147239,-0.2572387755,-0.0793285444,-0.2710337043,0.209878087,0.0581043065,0.0631146058,0.1123113036,-0.1852890104,0.0515824072,0.2637988329,-0.4474705756,-0.2216476202,0.3423317373,0.3455912173,-0.1874864846,0.0017540216,0.287109375,0.2440224737,-0.1056914702,-0.0194142144,0.4746490121,0.6444396973,0.4055102468,-0.3922822475,-0.0329234637,0.0636443719,0.0366322398,-0.1734556854,0.2106984258,0.030856356,0.1918324083,0.1406432539,0.0972324535,0.0374468304,-0.0763897672,0.2183976471,-0.1176240891,0.2614663243,0.0845798329,0.1263475567,0.1881272048,0.1792557985,-0.0836737677,-0.3425420523,-0.0696679875,0.1219683215,-0.0048621199,-0.1159545034,-0.110576421,0.0403339155,-0.2090077251,0.3243708313,0.2986623645,0.032985013,-0.3327325284,-0.1645352095,-0.3989806473,-0.1721303314,-0.0934053212,-0.3542155325,-0.0387813784,-0.1040405557,0.0048022824,0.079297021,0.0293139573,0.0266957842,0.1461547613,0.6865946054,-0.2840730846,0.0265500713,0.3582297862,0.1359502375,0.0420462489,-0.330671519,0.0253973547,0.3507256508,-0.0839070082,-0.1968306303,0.2640077472,0.1867125332,-0.1484750211,0.3411541581,0.2205915898,0.7242153883,0.121033676,-0.0829702541,-0.3107076883,0.0463107191,-0.1874597371,0.3284565806,0.4097307622,-0.2244555354,-0.3802423477,-0.4079791903,-0.1977362782,0.1020833179,0.0908953398,0.4288522899,-0.3884783685,0.3272823095,-0.0742274299,0.1896982193,0.104075551,0.0265209787,-0.0568277091,0.2692764401,-0.465179652,-0.2999322116,0.0334029756,-0.0166977998,-0.0271500405,-0.0090621253,-0.0695152283,-0.3448971808,0.3211408854,-0.3420500457,0.0472643636,0.2058964074,0.026876729,-0.3330900371,0.2316471785,0.1748759449,-0.1278861016,-0.069692336,0.4330183864,-0.1066297889,-0.1177691668,-0.0149017265,-0.1021668166]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2300","title":"Add VoxPopuli","comments":"I'm happy to take this on:) One question: The original unlabelled data is stored unsegmented (see e.g. https:\/\/github.com\/facebookresearch\/voxpopuli\/blob\/main\/voxpopuli\/get_unlabelled_data.py#L30), but segmenting the audio in the dataset would require a dependency on something like soundfile or torchaudio. An alternative could be to provide the segments start and end times as a Sequence and then it's up to the user to perform the segmentation on-the-fly if they wish?","body":"## Adding a Dataset\r\n- **Name:** Voxpopuli\r\n- **Description:** VoxPopuli is raw data is collected from 2009-2020 European Parliament event recordings\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2101.00390\r\n- **Data:** https:\/\/github.com\/facebookresearch\/voxpopuli\r\n- **Motivation:** biggest unlabeled speech dataset\r\n\r\n**Note**: Since the dataset is so huge, we should only add the config `10k` in the beginning.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":65,"text":"Add VoxPopuli\n## Adding a Dataset\r\n- **Name:** Voxpopuli\r\n- **Description:** VoxPopuli is raw data is collected from 2009-2020 European Parliament event recordings\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2101.00390\r\n- **Data:** https:\/\/github.com\/facebookresearch\/voxpopuli\r\n- **Motivation:** biggest unlabeled speech dataset\r\n\r\n**Note**: Since the dataset is so huge, we should only add the config `10k` in the beginning.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nI'm happy to take this on:) One question: The original unlabelled data is stored unsegmented (see e.g. https:\/\/github.com\/facebookresearch\/voxpopuli\/blob\/main\/voxpopuli\/get_unlabelled_data.py#L30), but segmenting the audio in the dataset would require a dependency on something like soundfile or torchaudio. An alternative could be to provide the segments start and end times as a Sequence and then it's up to the user to perform the segmentation on-the-fly if they wish?","embeddings":[-0.2925267816,0.2328721285,-0.0486936197,-0.075351961,-0.1558787078,-0.1852182746,0.3809297681,0.2100731581,-0.0260843243,0.2548180521,-0.2955557704,0.0939441025,-0.5140134096,0.2282823771,-0.0125598451,-0.3366333544,0.0352492854,0.1592858881,0.1379971355,-0.164063409,-0.1692886204,-0.1792166829,-0.3667643964,-0.3199285865,-0.2815556526,-0.0437956937,-0.1714102328,-0.2002765089,-0.2472260892,-0.4485911429,-0.2518495023,0.0144170914,0.2051721662,0.7153629661,-0.0001053368,-0.2415970266,-0.0325996578,-0.2413592637,-0.1185308173,0.1629431397,-0.1493930668,-0.00040988,-0.5305423737,-0.1669961512,-0.2364865094,-0.1963237971,-0.0257730559,-0.3211324215,0.2954051197,0.1716529131,0.2048163861,-0.2917951643,-0.0625556633,-0.0694862232,0.2304438353,0.0651996434,-0.0031293798,-0.092455022,0.0819319785,0.1289016902,0.0632565841,0.4567219317,0.048035495,-0.2020329684,0.1953832358,-0.2308383435,-0.3671851754,-0.4931638539,-0.1287543923,0.5815055966,0.6085772514,-0.2704017162,-0.3155983984,-0.1887135506,0.017296046,-0.3431016505,-0.3455108106,0.3020210564,-0.1481228024,0.1379403323,-0.0139239151,-0.3403760493,-0.139805913,0.09456072,0.0610956997,0.3523587286,0.0075402851,-0.0270667337,-0.0055152192,-0.2872108817,-0.4808892012,0.0372110829,0.2388414741,0.4193492532,-0.0955717415,-0.3250354528,-0.1058607474,-0.1411993951,0.1988049746,0.0749901086,-0.0428450331,0.0520790927,0.1216937453,-0.0533867478,0.2828865647,0.0018168306,-0.0796325281,0.06189657,0.106637165,-0.3097324371,0.0792736784,0.07797914,-0.207936123,0.0881983042,-0.2041271627,0.0717201829,-0.0003148082,-0.0222755671,-0.036746826,0.0304692071,-0.3467599452,-0.0008884933,-0.0230709445,0.0554772355,-0.2168714851,-0.0929948613,-0.1684412062,0.2591876388,-0.0406800956,-0.0559704751,0.0351964869,0.1009159461,-0.1506544203,0.2815228701,0.222290054,-0.0228016302,-0.0549538657,0.1347343624,0.2096423507,0.0041599385,-0.1650160104,-0.0868010521,0.2816715837,-0.1880036741,-0.0206576772,0.0136978477,-0.4079439044,0.2230240554,-0.2369116098,-0.0792811513,0.2260556221,-0.2475363463,-0.3368155956,0.2681538761,-0.2152198553,0.0765840858,0.3830368519,0.4839094579,0.0821424276,-0.4656518102,0.1455120295,0.1029925793,-0.0668397248,-0.0582668446,0.2434944659,0.1746969968,-0.4136834145,-0.1219652221,0.1921569556,0.1991189867,0.0906349123,0.2032759637,-0.0010413016,0.1211841404,0.0335892662,0.3394368589,-0.0144694233,-0.2765723765,-0.2051727325,-0.003992503,-0.070061788,0.0412313417,0.0307181757,0.1075341478,0.7296260595,-0.2890873849,0.293964684,0.3556069434,-0.0669099241,-0.0297854375,-0.135075286,-0.3457042277,0.1745294929,0.2376138568,-0.0093147606,-0.0910741836,-0.2259638458,0.4240358472,0.279981941,-0.1462579668,0.2492959648,0.0866705701,0.4744336605,0.0259649679,0.0415912792,-0.5515688062,-0.0342338122,-0.0477552228,-0.0555826649,0.3149366677,0.0476558357,-0.4406020045,-0.0364435166,-0.3342244029,0.0223343838,-0.001787528,0.1680096388,0.0543685779,0.1334091276,0.0449331328,-0.5313402414,-0.3334075511,0.3088917732,-0.0396379344,-0.094154723,0.0056672106,0.0476123542,-0.0103824763,0.1713317186,0.411060065,0.203959614,-0.1203439832,0.3768965304,0.3011947274,0.09832187,-0.0104773622,-0.3251622915,0.1853755116,0.2171079665,-0.3233448565,0.1568763852,0.3292605579,0.1090728939,-0.223283127,-0.0851829723,0.2577805519,0.4536969066,0.1915163845,0.0335362069,-0.175367713,-0.1085515618,-0.0654786825,-0.2391389757,-0.1759951264,-0.066945523,0.0106972661,0.135772109,0.1745662391,-0.3150488734,0.1541749984,0.4628355205,-0.2876065373,0.1285879761,-0.0026041064,-0.3632547557,-0.2040610015,-0.016408056,-0.4502245784,0.0546335839,0.2087284923,0.2128758132,0.0812147558,0.3439451456,-0.0911908746,0.017957285,0.1442548782,-0.0896611065,0.3438565731,-0.0715648979,0.0123671293,-0.0352526866,-0.3124836981,-0.0005498257,-0.0457016751,0.0406434536,-0.1953156739,0.0758681744,-0.2181125283,0.0700938851,-0.6984212995,-0.1143324971,0.1414815634,0.3873268664,-0.1372511089,-0.1164514124,0.3872788846,-0.0124242669,0.6995114684,-0.0220749285,-0.1269961894,0.0451534875,-0.0926728398,0.1592124701,0.1802596897,-0.0271615721,0.2576533854,-0.0187917836,0.1747030169,0.0832940489,0.2450941056,-0.0755416676,0.2807094455,-0.0216200445,0.1031171903,0.0060365158,-0.034200836,0.2087468207,-0.3511863351,-0.1773940772,-0.0107193682,0.0613081343,-0.1447474808,-0.2861626446,0.2674386501,-0.0853765234,-0.2339542061,-0.6628136635,-0.4774103761,0.3154145181,0.3436959088,0.1062131599,0.0269194786,-0.1817506254,0.0004490287,-0.0913956165,0.1175121516,-0.194193244,-0.0850744992,0.3067539334,-0.3033916652,-0.1011050642,-0.2366933972,-0.1138288975,0.0029071958,0.025774572,-0.0672825202,0.3524060249,0.0338186473,-0.0124957217,-0.0293374248,-0.1199404076,0.1766284406,-0.0328032449,-0.0975781158,-0.0202307887,0.1687162369,-0.1106215864,0.1685243696,0.0209706128,-0.0062213643,-0.1654156148,0.1592139006,0.9956915975,0.4840340614,0.0175147466,-0.0846361965,-0.0097522605,0.0726347044,-0.1796179265,-0.0587085858,0.5532771349,-0.0530377254,-0.1393408775,0.3943901062,0.0936160982,-0.3328875601,-0.0402027108,0.1430505216,-0.1820315123,-0.2135009021,0.122331202,0.0459815748,0.1855963916,-0.0658073574,0.1167552322,-0.2406722158,-0.0907292515,0.0895983353,-0.0965596512,0.2592721581,0.0846506953,-0.3579294682,0.0299741309,-0.4366183281,-0.0892414004,0.1967768222,-0.0965186432,-0.0927780569,-0.3198323548,-0.1560835391,0.1696870774,-0.1161262766,-0.0806686878,-0.0641258582,0.2182721049,0.6738339663,0.0340830386,0.4116018116,-0.1827397346,-0.0359324217,0.4898314774,0.2419969589,-0.3840941787,0.2102400661,0.2967365384,0.5018628836,0.0241964962,-0.047139246,0.03320802,0.0032868469,-0.1292332113,0.2143203616,0.2693782449,0.2161967903,0.117103599,0.0696158335,0.1879124343,0.1066986993,0.2010374218,0.0410475247,0.109163709,-0.0508437455,0.1418712139,0.1368931681,-0.3504539728,0.427106142,0.3146428764,0.1403210014,-0.0551570803,0.1650426239,-0.0094345603,0.2653466165,0.2739956975,0.2458314449,0.1712578237,-0.0407835133,0.3296459019,-0.3226228356,-0.0500217415,0.0061406684,-0.0895618945,-0.7040663362,-0.4799272418,0.2087133527,0.0070537487,-0.1786894947,0.1951508075,0.0504535772,-0.1807205528,0.6409842968,-0.1242557243,1.1574864388,0.4827519059,0.1591488868,0.1125205755,0.2006054372,0.237384364,-0.3035926223,0.0677878708,-0.0783930421,0.1987785548,-0.0498534553,0.0647743046,-0.1585078537,0.0622847155,-0.4324495494,-0.1228512377,-0.0842516795,0.0584112704,-0.0590635389,0.3170300126,-0.0231893137,-0.4224412143,-0.0628534481,0.217955485,0.0070259352,0.1343671978,0.0345465653,-0.1556110084,0.2441345751,0.0544787534,-0.3005514741,0.0694260523,-0.1215177998,-0.2669512331,-0.2369612455,-0.1326867789,0.0626224801,0.1947985142,0.0110709239,0.2505174577,-0.0093569318,0.1420045793,0.210490793,0.0371185839,0.2002182007,-0.0400160551,0.3876713514,-0.1026768833,-0.2133580744,0.2087536901,0.084586814,-0.2225948572,-0.208950609,-0.1261921674,0.2238057852,-0.2065971196,0.133875072,0.24416852,0.0451116264,-0.161676392,0.1447420418,0.0308184493,-0.3660282493,0.2378792465,-0.1021035761,-0.0897374973,-0.2988923192,0.2991769016,0.199194327,-0.0660865307,0.2654159963,0.1631921977,-0.2442837059,-0.3363744915,0.3560642898,0.0272337329,-0.2357182652,-0.0897472799,0.0112669393,0.0744597539,0.0584660992,0.3196860552,0.1003474742,-0.1523532122,-0.2279850096,-0.1858519763,-0.41594854,0.106522426,0.1921975166,0.1433023512,0.0104509536,0.0516233146,-0.1151609272,0.0882368088,-0.4500927925,0.0385353677,-0.1323826462,0.1920957863,-0.1528332233,-0.3040905893,-0.0052291746,-0.0048891669,0.1730174869,0.209867835,-0.145431295,-0.2495696694,-0.3390628994,0.1457418054,0.2924629748,-0.1402598768,-0.1638766974,0.0836843699,-0.2418487519,-0.1068436205,0.3897120953,0.3422495127,0.088857539,-0.2039665282,-0.0724264607,-0.038363073,-0.0480956845,0.0086663514,0.2314363271,0.1936449409,0.2700588107,0.2792682052,-0.1034337804,-0.119839415,0.1107210219,0.4209379256,0.460962981,0.0164932571,0.2232061327,-0.2835409641,0.3192583323,0.1052826196,0.4450283945,0.2600056231,-0.2056766748,-0.004945667,0.3935936987,0.3718560636,-0.1773594022,0.0193234701,0.0810720474,-0.2412969917,0.3103743494,0.2030960321,0.1488631368,0.1200706586,0.067681551,-0.0242296252,0.0328606591,-0.0844698995,0.287440151,0.3628749251,-0.1632513553,-0.0739804208,0.1752865613,0.3543741703,0.3791209757,0.5762962103,-0.1655828655,0.4650764763,0.30727157,-0.3316735327,0.2191117406,-0.1568671018,-0.1609803289,0.2824746668,0.0721072778,0.3706011176,-0.0795156509,0.3249427378,-0.6310242414,-0.3518262208,0.080315873,0.1161889583,0.1487323642,-0.1080908328,-0.2929435074,-0.041825965,-0.0189676695,-0.1837135106,0.0820153654,-0.1899735034,-0.1283593923,-0.0696764514,0.0318355225,-0.0394864343,-0.3224305212,-0.1900803,0.420560807,-0.20988141,0.0343906246,-0.0753983557,0.015079489,0.1595045924,0.2452432513,0.1217797473,-0.0540571548,-0.4147152603,0.2664084136,-0.205263719,-0.0796635598,0.0155249676,0.1405997425,0.2044271678,0.2787300348,0.3516688645,0.1802438051,-0.2170608938,0.0712030381,0.1786266565,0.0285099912,-0.107989639,0.021035919,-0.1912674904,-0.0878531337,-0.2785838842,-0.281231463,-0.3557817638,-0.1439835727,0.1353619546,-0.196617499,0.1673760265,-0.0332600139,0.1274765283,-0.1395255625,0.0831318274,0.1159180924,0.0855446905,0.0874626487,-0.0615021624,-0.216003269,-0.0088678561,-0.1558317691,-0.0121774981,-0.2357166857,0.1340303719,0.2658376396,0.3944415748,-0.3273127675,0.1216310188,-0.4167952836,0.2804063261,-0.3058728874,0.4049426019,-0.3673978448,0.5024877787,0.0545746014,-0.3051643372,-0.0807004273,0.0088899937,0.1484775245,0.0989567265,-0.281157136,-0.0034874806,-0.3702619672,0.4069079459,-0.2077631801,0.2602316141,-0.2546191216,-0.1644987315,-0.0612240061,0.1538012177,-0.1738909334,-0.1906511188,0.1229967177,0.4134189785,-0.135033533,0.2048758566,-0.0354044996,0.126711905,-0.0008092611,0.0561117306,-0.1075561941,0.1782752275,-0.1608628482,0.2360178381,0.1800747514,0.1975188404,-0.2519268692,0.0901321992,-0.1539792866,-0.248271659,0.5132052898,-0.2454801053,-0.3263628185,-0.2705000043,0.202494964,0.3115450144,0.0868823826,-0.5332134366,0.1467756182,0.207487762,-0.0956714377,-0.0311994758,-0.0753796771,-0.3398642242,-0.3039981723,-0.2820351124,0.2117494643,0.054130964,-0.1605184525,-0.0391393676,-0.2843498886]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2300","title":"Add VoxPopuli","comments":"Hey @jfainberg,\r\n\r\nThis sounds great! I think adding a dependency would not be a big problem, however automatically segmenting the data probably means that it would take a very long time to do:\r\n\r\n```python\r\ndataset = load_dataset(\"voxpopuli\", \"french\")\r\n```\r\n\r\n=> so as a start I think your option 2 is the way to go!","body":"## Adding a Dataset\r\n- **Name:** Voxpopuli\r\n- **Description:** VoxPopuli is raw data is collected from 2009-2020 European Parliament event recordings\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2101.00390\r\n- **Data:** https:\/\/github.com\/facebookresearch\/voxpopuli\r\n- **Motivation:** biggest unlabeled speech dataset\r\n\r\n**Note**: Since the dataset is so huge, we should only add the config `10k` in the beginning.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":54,"text":"Add VoxPopuli\n## Adding a Dataset\r\n- **Name:** Voxpopuli\r\n- **Description:** VoxPopuli is raw data is collected from 2009-2020 European Parliament event recordings\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2101.00390\r\n- **Data:** https:\/\/github.com\/facebookresearch\/voxpopuli\r\n- **Motivation:** biggest unlabeled speech dataset\r\n\r\n**Note**: Since the dataset is so huge, we should only add the config `10k` in the beginning.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHey @jfainberg,\r\n\r\nThis sounds great! I think adding a dependency would not be a big problem, however automatically segmenting the data probably means that it would take a very long time to do:\r\n\r\n```python\r\ndataset = load_dataset(\"voxpopuli\", \"french\")\r\n```\r\n\r\n=> so as a start I think your option 2 is the way to go!","embeddings":[-0.3622044623,0.2060509473,-0.0382300317,0.0014168692,-0.1214067563,-0.1296067238,0.2858929932,0.3488551676,0.1810037941,0.2673174739,-0.1220442802,0.1734330058,-0.5791192651,0.248657003,0.1030062363,-0.1610254347,0.1628734767,0.2866100371,0.095686391,-0.1880006641,-0.1290652156,-0.0335456841,-0.2270617336,-0.2975462079,-0.3531153798,0.0726715773,-0.1824733466,-0.2084129304,-0.1638894826,-0.2925849855,-0.2777968347,0.0334550031,0.2236543447,0.5653854609,-0.0001063294,-0.1449197084,-0.0884732082,-0.3044509292,0.0654772148,-0.0746489689,-0.0837001726,-0.1278287917,-0.4064903855,-0.0930253044,-0.2304193527,-0.14254345,-0.0099482238,-0.2797588408,0.219080478,0.2948248982,0.2410483956,-0.0154664135,-0.0557157435,-0.1818623245,0.1919789463,0.0053661554,-0.0167084429,-0.1235181242,0.0366831571,0.0193053875,0.0953444839,0.5634723306,-0.0203446001,-0.2623254657,-0.0066587911,-0.0895410776,-0.12987113,-0.4109855294,-0.0241050459,0.4772322774,0.4497810006,-0.2047097981,-0.2180188149,-0.1389561445,0.0822957903,-0.2967887819,-0.1726078689,0.2062319666,-0.0013450619,0.155995369,0.1247609481,-0.3047011495,-0.0445498005,0.2276000381,-0.0649357587,0.4114827216,-0.0107397391,-0.1321664602,0.0144946594,-0.2984534502,-0.3767269552,0.2829953432,0.2045624703,0.2323184609,-0.1000517011,-0.3026239574,0.1004269496,-0.0496396273,0.1847789288,-0.0104825487,-0.1364692301,0.1110458747,0.097645916,0.0904042274,0.2405513525,-0.0651146472,-0.0079654716,-0.1032886803,0.1962272078,-0.2124027163,-0.1570116132,0.1893924028,-0.2619791031,0.115580909,-0.0177203119,0.0266590323,0.0966506302,-0.1736463159,0.017109666,0.2102119476,-0.1320615262,0.0434758067,-0.0822213665,0.3474678397,-0.1171204597,0.2644560337,-0.0594007261,0.3889173269,-0.1311837435,-0.2085464597,-0.033891499,0.1506564766,-0.0409684964,0.2504803836,0.182081297,0.1834037602,0.0814338103,0.0801912248,0.4036087692,0.0462034941,0.0749738216,-0.0386664495,0.1496947408,-0.2442623377,0.0992058143,-0.0477506556,-0.0776870847,-0.0466320142,-0.2552615106,-0.0784299299,0.2434822172,-0.2535855472,-0.2981429398,0.2990464866,-0.1131108627,-0.1176313832,0.2105854005,0.6592457294,-0.0265016109,-0.3424615562,-0.0631873459,0.0526253656,-0.1718238741,-0.0150967557,0.149996385,0.2983689904,-0.317938596,-0.198401913,0.0946157202,0.2202681154,0.0514327325,0.2936134636,-0.0016879287,0.0959733129,-0.0085592382,0.223138392,0.0867227092,-0.3522504866,-0.3762060404,0.0857585296,-0.0944087133,0.037445046,0.1689456701,0.205665946,0.5524084568,-0.2033003867,0.342663765,0.4188266098,-0.0198924486,-0.0120015536,-0.093754299,-0.3128604293,0.2254299968,0.3783738017,0.0043753078,0.0188300516,-0.266597122,0.3279888034,0.3670642376,-0.2761883736,0.1107815653,0.1414887607,0.4262054265,0.0683253035,-0.0993256941,-0.454015255,-0.0165982731,-0.0257257037,-0.0956509784,0.3194210231,-0.0702558458,-0.5392349958,0.0250341278,-0.3108381629,0.0048708576,0.0713820234,0.2480750084,0.1039580703,0.1168510467,0.0260644965,-0.4303718209,-0.031660486,0.1560540497,-0.1150336713,-0.1255701929,0.1169828102,-0.019477766,-0.1644623131,0.2081956714,0.2909441292,0.0755520463,0.0528743863,0.2668861747,0.0709270313,0.1357492059,-0.1132397428,-0.0462313034,0.0530095659,-0.0085947737,-0.384639442,0.0634869039,0.3921803832,-0.0621207058,-0.0536217839,0.0895175114,0.213598296,0.497148931,0.0607669316,-0.0248055328,-0.0618537515,0.0949490741,-0.0250037238,-0.1085492298,-0.2827538848,0.1236067936,0.1179459691,0.1139038429,0.1439710408,-0.433542192,0.0984679684,0.4523300827,-0.1298229843,0.385292381,0.048118487,-0.3084796667,-0.01942176,-0.0551850833,-0.3254057765,0.1220445335,0.3093282282,0.0621400326,-0.0152722774,0.2443618625,-0.0327318683,0.0469609983,0.0170565862,-0.198391661,0.2099432796,0.0601658225,-0.1522661448,-0.2575821877,-0.3199756145,-0.0662788898,-0.0185559131,0.097553961,-0.2175047547,0.0117299501,-0.0706099719,0.0489459261,-0.3854290545,-0.1949515641,0.1977466047,0.3800057173,-0.0833938792,-0.1135616079,0.2648079097,-0.0519088171,0.4458293021,-0.0833611116,-0.3190987706,0.1689811498,-0.1556269825,0.0674682334,0.1898882389,0.003277397,0.2206191272,0.0939170271,-0.0385639891,0.0312705263,0.1009206176,-0.1914321482,0.1785158366,-0.1373093128,0.0531251431,0.0208643079,0.0789514706,0.0419875495,-0.2782662809,-0.0537762046,-0.1046979651,0.1359056979,-0.0085710324,-0.2253914326,0.1661050171,0.0294249989,-0.3284982443,-0.7819321752,-0.564748764,0.2430546135,0.3136271536,0.1024924964,-0.0683986768,-0.1409730017,-0.0563666038,-0.1719581187,0.2470577806,-0.2032077312,0.0324627794,0.4612770379,-0.2707383037,-0.1865020841,-0.2920601368,-0.2375630289,0.0957440287,-0.1553489566,-0.2603883743,0.2003104389,-0.1632536799,0.0965487212,-0.0052643623,-0.026227301,0.3263149858,0.1320692152,-0.1316966712,-0.0994005501,0.1153095961,-0.0704140589,0.13906084,-0.0816311762,0.0134751415,-0.1746074855,0.031437844,0.6643480659,0.3344547451,0.0125743132,0.1750955582,-0.001384374,0.1119310111,-0.122755155,-0.1437574029,0.1638988107,-0.0561096221,-0.1187385321,0.5490953326,0.0184465703,-0.4654216468,-0.1587921083,0.1038697883,-0.2134498656,-0.2118079662,0.2113968581,0.0388385728,0.1743007153,-0.1160275936,-0.0555653609,-0.2088219821,-0.0120235076,0.1845556647,-0.0107939653,0.1448997855,-0.1179814115,-0.2399490625,0.0103754746,-0.4669721425,0.1290613413,0.0955209583,-0.1345151216,-0.119046554,-0.2967168093,-0.2244537026,0.0841989741,-0.0949912965,-0.2629108727,-0.1887690872,0.1808092743,0.3611077368,-0.1731304973,0.4213995934,-0.3377197087,-0.1228825301,0.3760777116,0.2507035136,-0.4433840513,0.3351083994,0.1823346466,0.5083808899,0.0011387859,-0.0721249804,-0.2182147205,0.02664252,-0.2039164901,0.4047904909,0.1798480302,0.0870758444,0.2208205014,0.0615780167,0.0703690499,0.1028543562,0.3789216876,0.077046752,0.2333376408,0.0289199166,0.1345978975,0.1121224239,-0.1881216615,-0.029536549,0.4321112633,0.0630023926,-0.2604467273,0.0249346551,0.0034138754,0.3576073349,0.3181078732,0.2150359303,0.4056583047,0.0334019959,0.1880118847,-0.2066317499,0.0695324019,0.2023233175,-0.1256695092,-0.4516105354,-0.5478141308,0.1800533235,0.1455732137,-0.137280494,0.0861364082,0.0228809398,-0.118779704,0.6165671945,-0.3423854411,1.1753385067,0.3344912529,0.2076600492,0.048014123,0.1837857962,0.3083378077,-0.3009546697,0.0103776958,0.0684071705,-0.0885392725,-0.0191612281,0.0707406327,-0.1447669715,0.1776678264,-0.7377361059,0.0661364421,-0.1844971925,-0.0211432669,-0.1358198971,0.2500279248,-0.1251886189,-0.5233939886,-0.4447706938,0.2471960187,-0.0130944075,0.1376274228,-0.0007742986,-0.0571646318,0.0897976831,-0.0959877223,-0.2057290524,0.0694097877,-0.1310457736,-0.15450643,-0.3360479474,-0.0461532772,0.0329301432,0.2875268161,0.1190964803,0.4786844254,-0.1238556728,0.0854962319,0.3140234053,0.1191115528,0.1748796403,-0.0558027439,0.2491304576,-0.0389485955,-0.3561419547,0.2521935701,0.1491468102,-0.2774612606,-0.2780540884,-0.1171088442,0.1806351542,-0.0457689613,-0.0004737449,0.307536155,0.0031316371,-0.2630973756,0.20511356,0.0770249143,-0.4349069893,0.0878979042,0.0002250245,-0.0886179358,-0.1464882493,0.245290786,0.307666868,-0.0739828721,0.1029378176,0.3325701654,-0.129441604,-0.3821825385,0.1188283116,0.0627315342,-0.2772650123,-0.1331376582,0.0890587121,-0.1484768689,-0.1134964749,0.2436072379,0.1177873015,-0.2202747464,-0.1120534316,0.0209175143,-0.6115119457,0.2955403924,0.36338377,0.1040994599,0.2315910757,-0.1309386194,-0.2784081697,0.1934369951,-0.4642868936,-0.0632442832,-0.0965128466,0.1608902216,0.0074284812,-0.245955497,0.1158762649,-0.0405655541,0.2241312563,0.1675934345,-0.0176859535,-0.325505048,-0.2224099189,0.0868890584,0.23665075,0.1457112879,-0.0152936652,0.0368036069,-0.1100719348,-0.1060400158,0.44591555,0.2392483801,0.0525279231,-0.088047266,-0.1513987929,-0.2160506845,-0.1253124177,-0.0663155317,0.1494654268,0.1749302745,0.230473727,0.0673189536,0.0116901863,-0.1654163301,0.0794728175,0.3628195226,0.3833221197,0.1561152041,0.0436235666,-0.2703102827,0.222749725,0.1688221544,0.4512782395,0.1104411259,-0.1209568456,0.0941596702,0.4990374744,0.4179678559,-0.4108498096,-0.0577696003,0.2328460813,-0.3552444875,0.383977592,0.3740125597,0.0952963009,0.1067197695,0.0465501361,0.0371718444,-0.1151010841,-0.0377179794,0.2571498156,0.1460974962,0.102478154,-0.0748103708,-0.0013248236,0.343977958,0.3266224265,0.530084908,-0.2825890183,0.3886515796,0.0956526473,-0.2597624063,0.2736119032,0.0642057508,0.1377782077,0.2306453139,0.1587521732,0.2691073716,-0.0515620522,0.4701898396,-0.5321553946,-0.2839860916,-0.0702507719,0.0982946306,0.065747954,-0.1043236479,-0.0354628302,0.0261128116,-0.2292875499,-0.1708310246,0.2301443219,-0.1359268576,-0.268232882,0.2988194823,0.1705299318,-0.2401818484,-0.1490207314,-0.1470341086,0.3594738543,-0.2478773594,0.167184189,-0.1194426343,0.0528101772,0.2467084676,0.2892059088,0.1465043575,-0.0964315534,-0.3751784265,0.329867512,-0.2860944271,-0.0447454974,-0.1460156441,0.0783331022,0.1508556157,0.4221980274,0.276132673,0.2086019814,-0.3079877198,-0.1333698481,0.1586046964,0.0423179455,-0.1505169421,0.0643145889,-0.1524146646,0.0442225635,-0.4095975161,-0.2510286272,-0.6997446418,-0.1190193817,0.1614104658,-0.052483663,0.071913071,-0.0839184076,0.1417285353,-0.0368319117,-0.0125219626,0.0663312823,0.2649344504,-0.0075703906,-0.0811253041,-0.3581408858,0.2474453896,-0.2162035108,-0.0841441602,-0.2664990425,0.0812276304,0.3445623815,0.4079996943,-0.2888800204,0.0208227206,-0.3841980994,0.1748024225,-0.2963863909,0.292953223,-0.2503388524,0.4722779393,0.1480895728,-0.3501276374,-0.3310790658,0.0409936123,0.173099637,0.0031631566,-0.4065484107,-0.0937643796,-0.3867294788,0.405582726,-0.1437370181,0.5092089772,-0.3020283282,-0.2368839383,0.033028055,-0.0799093693,-0.2482530326,-0.1539028138,0.032241568,0.4596185684,-0.1901183128,0.2216008455,-0.0705072582,0.1909005344,0.0626067147,-0.0679495558,-0.1536718905,0.2455637753,-0.0057615242,0.1625899225,-0.0941634476,0.2533111274,-0.2927538455,0.1010814011,-0.126283139,-0.2569379508,0.3687209785,-0.3402694762,-0.2988111973,-0.2196720988,0.3579612374,0.1641681641,0.0845543966,-0.2945986986,0.1834283918,0.0427230485,0.1098536178,-0.0096263913,0.0706700683,-0.1355887055,-0.2739542425,-0.2798532844,0.1673711687,0.0584491082,-0.0878947005,-0.1146517247,-0.3189821541]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2294","title":"Slow #0 when using map to tokenize.","comments":"Hi ! Have you tried other values for `preprocessing_num_workers` ? Is it always process 0 that is slower ?\r\nThere are no difference between process 0 and the others except that it processes the first shard of the dataset.","body":"Hi, _datasets_ is really amazing! I am following [run_mlm_no_trainer.py](url) to pre-train BERT, and it uses `tokenized_datasets = raw_datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=args.preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=not args.overwrite_cache,\r\n )` to tokenize by multiprocessing. However, I have found that when `num_proc`>1\uff0cthe process _#0_ is much slower than others.\r\nIt looks like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/31714566\/116665555-81246280-a9cc-11eb-8a37-6e608ab310d0.png)\r\nIt takes more than 12 hours for #0, while others just about half an hour. Could anyone tell me it is normal or not, and is there any methods to speed up it?\r\n","comment_length":39,"text":"Slow #0 when using map to tokenize.\nHi, _datasets_ is really amazing! I am following [run_mlm_no_trainer.py](url) to pre-train BERT, and it uses `tokenized_datasets = raw_datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=args.preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=not args.overwrite_cache,\r\n )` to tokenize by multiprocessing. However, I have found that when `num_proc`>1\uff0cthe process _#0_ is much slower than others.\r\nIt looks like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/31714566\/116665555-81246280-a9cc-11eb-8a37-6e608ab310d0.png)\r\nIt takes more than 12 hours for #0, while others just about half an hour. Could anyone tell me it is normal or not, and is there any methods to speed up it?\r\n\nHi ! Have you tried other values for `preprocessing_num_workers` ? Is it always process 0 that is slower ?\r\nThere are no difference between process 0 and the others except that it processes the first shard of the dataset.","embeddings":[-0.4360868633,-0.3340421319,-0.0331812017,-0.0555871613,0.0697014108,-0.1944787651,0.3812371194,0.2073157132,-0.2748001516,0.0517776236,0.3157570064,0.3878782392,-0.2385890484,-0.0081708841,-0.2220509499,0.1165851951,0.1717216223,0.1350863129,0.3698132932,-0.1361592263,-0.2459045947,-0.0289522111,-0.4537950158,0.1861464381,-0.5515934229,-0.0939204022,0.0168486442,0.0239229016,-0.1578884572,-0.332066685,-0.2777600884,0.0381636694,0.0004328316,0.3452011943,-0.0001291091,-0.2028736472,0.3343115449,0.3039593399,0.0683973506,-0.0113468543,0.1007517427,-0.4201109111,-0.0934073478,-0.0518576428,0.0113793248,0.0090428917,0.2106573284,-0.080221355,0.1935958713,0.0336161815,0.0578775331,0.3355197012,-0.4118835628,0.3002010286,-0.0285426844,0.1999465078,-0.0942220986,-0.0267875493,0.1064961553,-0.1693147868,0.1627754867,0.1288421005,-0.2492694855,-0.0105363671,0.2192675769,0.1379474849,0.5077482462,-0.4718617797,0.341483444,0.242095232,-0.1116657183,0.1791999191,-0.3060411811,-0.3064305782,0.0190430563,-0.4115245938,-0.002620022,0.0045570042,0.0472666025,-0.0208514314,-0.4781558216,0.1738860011,0.1423054338,0.0237278379,-0.2842411101,0.6366424561,0.0835530832,0.120980911,0.1686219275,-0.1988871247,-0.2149642408,-0.2544279397,0.1541729122,0.5229039788,-0.1893164515,-0.1005733088,-0.1133739948,-0.0243343841,-0.2357409447,-0.3004845083,-0.3913839161,0.3871785998,0.0365640409,0.1213994995,0.0062340447,-0.0791992918,0.0439269617,0.2379095554,0.2816818357,-0.4815149307,-0.2453261167,0.0616535023,-0.2930653095,-0.1480608732,-0.0174206942,0.1158762053,-0.21604608,-0.0180586483,-0.0678548515,0.0642534271,-0.3971650302,0.0320900157,0.3310677409,0.1310614496,-0.055100929,0.5344244838,-0.35351336,-0.0543107614,-0.512452662,0.0594564341,-0.0384577923,-0.1878851056,-0.2003136426,0.1468878388,-0.0343806148,0.1941287965,0.1312454641,0.0913665071,0.175832212,0.0764411837,0.553352952,-0.2733448148,0.2202516645,0.2896812558,-0.2941050529,0.4735400081,-0.0568000562,0.2583467662,-0.2132853121,0.4684457183,-0.513767004,-0.4455690384,0.1040069386,-0.0502087399,-0.0349195972,0.2365925312,-0.2229654044,0.3019234538,0.3209953606,-0.0065404489,-0.1902328581,-0.2581340671,-0.569391191,-0.0931628868,0.2414494753,0.2863372564,-0.128432557,0.0288027264,0.1292352378,0.2965081632,0.7083096504,0.3495642841,-0.1610260755,0.1975356489,-0.1121664122,0.5817081928,-0.0573083386,-0.0024948099,-0.1221966818,0.331412971,-0.3430331349,-0.1398524791,-0.0191545077,0.0702573508,0.2119109631,-0.1398646832,0.207643643,0.0084966999,-0.0263468977,0.3012176156,-0.4460811317,0.0836742297,0.2569980323,-0.0578046776,-0.1963777393,-0.1404869556,-0.3174522519,0.0300512332,0.0254321173,-0.0975899547,0.1209555939,0.0960759223,0.151537329,-0.1494010538,0.0074114809,0.0793573335,0.0448698588,0.1637542546,-0.1777203679,0.057131514,0.0228353757,-0.0804902911,0.1946886182,-0.0073325098,-0.0777315572,0.0006750442,-0.1042289138,-0.1330376565,-0.171391651,-0.0241134223,-0.0527573675,0.0256196428,0.0285961777,-0.0098101199,0.1717209369,-0.1269735247,0.0856135413,-0.1991586685,-0.3183270991,0.1766962707,0.293682754,-0.1260847151,0.0239189956,0.3493389785,0.339012146,0.0682437867,-0.0969691351,0.2577645183,0.3222962618,0.225922659,0.2996601462,-0.2068662196,-0.0924339071,-0.2335605621,0.2150356472,0.869019866,0.3758693337,0.2990359664,-0.0244135521,-0.2960143089,-0.0128675206,0.282496959,-0.0023054213,-0.0397009999,0.2367777079,0.3676816523,0.3466972709,0.2306513637,-0.0149930604,0.1968118399,0.136352554,-0.2129914761,-0.1881495118,0.0376621708,-0.034360297,-0.1988098323,0.1022654697,-0.5859146714,0.2442331463,-0.0352273881,0.1711592525,-0.0063750111,-0.1822370589,-0.0630133227,0.1467841268,0.1539622694,0.0412814133,-0.0272227563,0.1279747784,0.070094496,-0.0916842371,-0.1123032123,0.220798403,0.5185967088,-0.22422041,0.1264094859,-0.1815513223,0.3484581411,-0.0092580831,-0.1181666404,-0.1135653481,-0.2855651975,-0.0233045779,0.0805970505,0.1668619514,0.1096139997,-0.014418059,-0.0733010918,-0.1003982276,0.2277217209,-0.0665105879,-0.0254166946,-0.2751339674,-0.1501680911,0.3499274254,0.0754394457,0.0420038253,-0.0100306198,-0.3488819301,0.0729177147,-0.5219573975,-0.0330346897,0.0633742213,-0.0331862979,0.0179083683,0.0217051264,-0.5248930454,0.1516859829,0.1539483964,-0.1330841929,-0.2944366634,-0.0114098564,-0.2839923501,-0.23933658,-0.0805285797,0.1574743837,-0.0880601853,0.3033662438,0.1242295206,-0.1432494968,0.1995020509,-0.1756274104,-0.1001195386,0.1543434262,-0.0461929254,-0.3905456066,-0.0550690331,-0.2851268351,0.3271590769,-0.1026911139,0.0898140296,0.031749934,-0.0419975035,-0.0872765854,0.1167670116,0.0802402645,0.0108845262,-0.0897263885,0.0274961051,-0.0580018535,-0.0622239858,0.3971161842,0.1535115093,0.0334778503,0.1151597127,-0.3876470029,-0.1599521935,0.0064627104,-0.0498174839,0.0998841226,0.3346404731,0.0874356255,0.9783514738,0.1062373146,-0.2804869413,-0.1336958259,-0.1836688668,-0.2738057971,0.0565302372,0.0110868458,0.2926231027,0.3481584489,-0.1060491726,0.3516156673,0.0360320918,-0.4280943274,0.1520730853,0.010627389,-0.2330413759,-0.1345241815,0.1594465971,-0.2357102931,0.299619168,-0.0050081126,0.2305292189,-0.4275159538,-0.3596069217,-0.1457601339,0.0265662484,0.0506896004,-0.33763659,-0.6599957943,-0.1241075769,-0.8557753563,0.3928841054,0.0668693036,0.400065273,0.3554160893,-0.041925326,0.2548026145,-0.1705655903,0.9261282086,-0.3752493858,-0.0479906276,0.0609394945,-0.3622165322,-0.1824952066,-0.2055875212,-0.0673929602,0.5040717721,0.4771130979,0.6695144176,-0.0840027109,-0.3767013252,0.0516610146,-0.160721764,0.1572792381,-0.1000597179,-0.3202103972,0.0091814203,0.1780499965,0.3636065125,-0.132689625,0.121794343,-0.1157529727,-0.1865264326,-0.0926803127,-0.0315532908,0.3476769626,0.0403212085,0.1968769878,0.2026277483,0.2127318531,0.3579660058,0.2923240066,0.3474156559,-0.3887328506,-0.0742835924,0.092195496,0.174922958,-0.1078387797,0.4646156728,0.5473309755,0.1076299623,0.1817895621,0.0966072083,0.3447217047,-0.1060336828,0.250379324,0.5614129305,0.2437326312,-0.5444434881,-0.2288586944,0.074421443,0.5478266478,-0.0730599388,-0.0137783233,-0.4667111635,-0.0549075864,0.115440473,0.3533850014,1.1370749474,-0.4572671354,0.2235987931,-0.3345710039,0.18654567,0.1391558349,-0.3053277135,0.169403702,-0.3431823552,0.126192078,-0.0519645885,-0.0698041022,0.2564783096,0.470136255,0.1489556879,0.1898185015,-0.0283735953,0.2957143486,-0.1705434322,0.093322888,0.541778326,-0.3341778219,0.1757439822,-0.0575115904,0.20419541,0.0147024924,-0.0116626685,0.0275173057,-0.029948812,-0.1083595231,-0.1351928264,-0.3347499669,-0.2640464902,-0.104440622,0.2398075014,-0.0443115681,-0.0281098969,0.0686316341,0.2051471621,0.2910551131,-0.0669766888,0.0809767246,0.0974567235,0.1381753087,0.1266136467,-0.374538511,-0.1369054019,-0.0491354614,0.0354471989,-0.1650296003,-0.1853073984,0.0037485177,-0.2561003864,0.0792189091,0.6586583853,0.3335108459,-0.1097470075,0.1790461987,-0.1239253506,-0.1210858449,-0.0148874987,-0.0461560786,-0.1962199509,0.4305231869,-0.1915863305,-0.491784066,-0.0159026235,0.4187040627,0.1013255492,-0.1454494894,0.5520439148,-0.2703860104,-0.2152642906,-0.0652342737,0.4115815461,0.1103633791,0.1579066813,0.0483676381,0.3706420362,-0.0252780039,0.3468013406,-0.0522842519,-0.2708886266,0.0632002875,-0.1935213506,-0.1205903143,-0.3543397188,-0.3424687684,-0.187629059,-0.0933290422,0.0107331239,0.2203310579,-0.2272357494,-0.1343101114,-0.1209257469,0.0782522038,-0.0859043002,0.2770214379,-0.2059434354,-0.0328886062,-0.0524546467,-0.0511375293,0.0184968393,0.2247282267,-0.2035122365,-0.0793928653,-0.0961862057,0.2424827963,0.0696056783,-0.4496526122,0.3252120614,0.1513261348,-0.3381516039,-0.2613519132,-0.1411797404,0.2994889021,-0.0444702767,-0.1229884773,0.4639209807,0.1736130714,-0.1211529672,0.1722890884,-0.0991256684,0.150899291,-0.0524269231,0.3517229259,0.3136869669,-0.088929601,0.2567594349,-0.0925771073,0.1440130025,0.1110242605,0.0540262461,-0.0857150853,0.0341706946,0.1923605651,-0.0414828882,0.1802287698,0.0338949561,-0.3946097791,0.6167111397,0.000125676,0.0131852487,-0.1981575787,0.3172850907,0.0790933073,-0.0049079745,0.1999772489,0.3037670255,0.3584080935,-0.2044161707,0.0103774155,-0.1436354667,-0.2623726428,0.0120437844,0.1228772849,0.1161635667,0.1627409756,0.4875868559,0.0196338426,-0.0677000508,0.4756557941,0.3589251935,0.5300147533,0.3057105243,0.4101434648,-0.4583235085,-0.3641141355,-0.2764772177,0.6516803503,0.023685677,0.2694939375,-0.0620531216,0.2542199194,-0.1229631081,-0.4720273912,0.1158060953,0.3828210831,-0.3418947756,-0.5146676898,-0.2581237853,-0.0828671157,-0.0889221802,0.1356114,-0.2706575394,0.1188920289,0.3547161222,0.0894194767,0.0913793668,0.0061074374,-0.1751971692,-0.1463090032,0.3146529496,-0.0293302555,0.1819389611,-0.6590737104,0.1216758788,0.1522235274,-0.0928068459,0.3953402042,0.62741822,-0.4116170108,-0.0309420004,0.0173123777,0.0739464983,0.1964253336,0.2626115084,0.0273683835,-0.1165741384,-0.0630779639,-0.2329760939,-0.1438362002,-0.3250848055,0.0415995903,0.0322756842,0.1660140604,0.2918975353,-0.051323466,-0.2218306959,-0.0675225556,0.2138188779,-0.1194086,0.006276805,0.1002995521,-0.1701913923,0.1046152785,0.108951889,-0.0085929437,-0.2315937281,0.4062277377,0.207452178,-0.0814038441,-0.4438832104,-0.1291895211,-0.4720833898,0.1414043903,-0.3913087249,0.1325625181,-0.1618442237,0.0186502747,-0.0297398139,-0.0175823495,0.115549393,-0.0919885412,-0.0879414752,0.2840396166,-0.2083985656,0.4887945652,-0.3455667794,-0.2010160387,-0.0177066643,-0.3488202393,0.455925256,-0.0328089707,-0.1777759194,-0.0711356178,0.1870520115,-0.1703267395,0.3382768035,0.3156832755,0.1269315481,0.2096211314,0.0929424912,0.1342771351,0.0061375098,0.3694486916,-0.2378084064,0.216237843,-0.0895204321,-0.0584537312,-0.082097806,0.0643766597,-0.069773376,-0.0684572607,0.1957748085,-0.0802553222,-0.1768534184,0.2231896222,0.0027977468,0.0258168038,-0.1533599794,0.1623668373,0.2284028232,-0.0751340091,-0.3145273924,0.1471103579,0.2970348299,-0.5313223004,-0.126343295,-0.7054094672,0.0914079174,0.0993219391,0.1780920327,-0.1492673755,-0.2474097759,0.4167242646,-0.1845205128,-0.3637081683,0.2536584735,-0.062303789,-0.2088903934,-0.3710052669,0.0534602962,-0.1362582147,-0.0812289566,0.0470621102,-0.0913573876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2294","title":"Slow #0 when using map to tokenize.","comments":"Hi, I have found the reason of it. Before using the map function to tokenize the data, I concatenate the wikipedia and bookcorpus first, like this:\r\n```if args.dataset_name1 is not None:\r\n dataset1 = load_dataset(args.dataset_name1, args.dataset_config_name1, split=\"train\")\r\n dataset1 = dataset1.remove_columns('title')\r\n if args.dataset_name2 is not None:\r\n dataset2 = load_dataset(args.dataset_name2, args.dataset_config_name2,split=\"train\")\r\n assert dataset1.features.type == dataset2.features.type, str(dataset1.features.type)+';'+str(dataset2.features.type)\r\n datasets12 = concatenate_datasets([dataset1, dataset2], split='train')\r\n```\r\nWhen I just use one datasets, e.g. wikipedia, the problem seems no longer exist:\r\n![image](https:\/\/user-images.githubusercontent.com\/31714566\/116967059-13d24380-ace4-11eb-8d14-b7b9c9a275cc.png)\r\n\r\nBookcorpus has more row numbers than Wikipedia, however, it takes much more time to process each batch of wiki than that of bookcorpus. When we first concatenate two datasets and then use _map_ to process the concatenated datasets, e.g. `num_proc=5`, process 0 has to process all of the wikipedia data, causing the problem that #0 takes a longer time to finish the job. \r\n\r\nThe problem is caused by the different characteristic of different datasets. One solution might be using _map_ first to process two datasets seperately, then concatenate the tokenized and processed datasets before input to the `Dataloader`.\r\n\r\n","body":"Hi, _datasets_ is really amazing! I am following [run_mlm_no_trainer.py](url) to pre-train BERT, and it uses `tokenized_datasets = raw_datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=args.preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=not args.overwrite_cache,\r\n )` to tokenize by multiprocessing. However, I have found that when `num_proc`>1\uff0cthe process _#0_ is much slower than others.\r\nIt looks like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/31714566\/116665555-81246280-a9cc-11eb-8a37-6e608ab310d0.png)\r\nIt takes more than 12 hours for #0, while others just about half an hour. Could anyone tell me it is normal or not, and is there any methods to speed up it?\r\n","comment_length":172,"text":"Slow #0 when using map to tokenize.\nHi, _datasets_ is really amazing! I am following [run_mlm_no_trainer.py](url) to pre-train BERT, and it uses `tokenized_datasets = raw_datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=args.preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=not args.overwrite_cache,\r\n )` to tokenize by multiprocessing. However, I have found that when `num_proc`>1\uff0cthe process _#0_ is much slower than others.\r\nIt looks like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/31714566\/116665555-81246280-a9cc-11eb-8a37-6e608ab310d0.png)\r\nIt takes more than 12 hours for #0, while others just about half an hour. Could anyone tell me it is normal or not, and is there any methods to speed up it?\r\n\nHi, I have found the reason of it. Before using the map function to tokenize the data, I concatenate the wikipedia and bookcorpus first, like this:\r\n```if args.dataset_name1 is not None:\r\n dataset1 = load_dataset(args.dataset_name1, args.dataset_config_name1, split=\"train\")\r\n dataset1 = dataset1.remove_columns('title')\r\n if args.dataset_name2 is not None:\r\n dataset2 = load_dataset(args.dataset_name2, args.dataset_config_name2,split=\"train\")\r\n assert dataset1.features.type == dataset2.features.type, str(dataset1.features.type)+';'+str(dataset2.features.type)\r\n datasets12 = concatenate_datasets([dataset1, dataset2], split='train')\r\n```\r\nWhen I just use one datasets, e.g. wikipedia, the problem seems no longer exist:\r\n![image](https:\/\/user-images.githubusercontent.com\/31714566\/116967059-13d24380-ace4-11eb-8d14-b7b9c9a275cc.png)\r\n\r\nBookcorpus has more row numbers than Wikipedia, however, it takes much more time to process each batch of wiki than that of bookcorpus. When we first concatenate two datasets and then use _map_ to process the concatenated datasets, e.g. `num_proc=5`, process 0 has to process all of the wikipedia data, causing the problem that #0 takes a longer time to finish the job. \r\n\r\nThe problem is caused by the different characteristic of different datasets. One solution might be using _map_ first to process two datasets seperately, then concatenate the tokenized and processed datasets before input to the `Dataloader`.\r\n\r\n","embeddings":[-0.4216788113,-0.2463773936,-0.0001557727,0.0670946091,0.0309756473,-0.2085657269,0.4310260415,0.2237498313,-0.2545164824,0.0450726971,0.2334742099,0.3255723715,-0.141471222,0.0176063161,-0.2975897789,0.0944023654,0.1700220108,0.1978500187,0.3180062473,-0.106922403,-0.1142958179,-0.0325482376,-0.513414681,0.0937799811,-0.516996026,-0.094234772,-0.052355025,-0.008509526,-0.087839283,-0.3727582395,-0.1796108931,0.0533387512,0.0209058262,0.3229079843,-0.0001312324,-0.2167607844,0.321980685,0.251037389,-0.077723369,0.0198327024,0.0318119526,-0.3308933377,-0.077887468,-0.1234237701,0.0175319202,0.0237414055,0.0749674439,-0.2054030895,0.0913552269,0.0141805075,0.0364925042,0.3499347866,-0.3504897058,0.2594860792,-0.0747995898,0.2026159912,-0.0626254454,-0.0495990664,0.1012017727,-0.1707479507,0.1734009683,0.2813491523,-0.2406679839,-0.0495627448,0.1356945932,0.1603334695,0.4377095103,-0.4210774601,0.3177548945,0.2465801984,-0.0422619022,0.097834304,-0.3110384643,-0.2917485833,0.0804972127,-0.2755731344,-0.0260986835,0.0504077449,0.0243027862,-0.0064566298,-0.427964747,0.1010868847,0.1111115143,0.1455284357,-0.2755873501,0.5521758199,0.1067677513,0.2040685415,0.1689036191,-0.2546339631,-0.3334892988,-0.4314102232,0.1836215258,0.477684617,-0.287428081,-0.1189518645,-0.155375123,-0.2407781035,-0.1358047873,-0.3256503642,-0.4295946956,0.358568728,0.031497594,0.0367216021,0.1139290407,0.0111999372,0.1415977031,0.3352188766,0.2813811004,-0.3873303533,-0.2329002619,0.0947895348,-0.3194718063,0.0019109213,-0.1017474458,0.0573491007,-0.1384474933,-0.0624518916,-0.0863149762,0.0299601927,-0.4705274999,0.0324194096,0.3583807647,0.1112597138,-0.0891091153,0.556709826,-0.3564697802,-0.0558779985,-0.4742580652,0.1034273878,-0.0102328192,-0.1693373024,-0.2180451304,0.1857894063,0.0246342495,0.0998852178,0.0603745207,0.1137021706,0.1227449849,0.0288294069,0.5027255416,-0.393191725,0.2299547344,0.2137600332,-0.2416318357,0.5415733457,-0.0803955272,0.0517807752,-0.2654265463,0.4353518784,-0.5407435894,-0.4055198133,-0.0378012881,-0.0362009592,-0.0213706996,0.2882744372,-0.2898297608,0.3646245897,0.3560539782,-0.0592743903,-0.2194041908,-0.1098218486,-0.4298143387,-0.0662313327,0.1920270473,0.3820966482,-0.08586815,-0.0251053963,0.2118619382,0.364710182,0.7044425607,0.4722286463,-0.2808420956,0.1810291409,-0.1422220021,0.6345632076,0.1009041518,-0.2661673129,-0.1717864126,0.4077458084,-0.3184235096,-0.0377777442,0.0876825601,0.0555645339,0.1874470711,-0.1130772382,0.2441991717,0.1069103107,0.0766589344,0.3537717462,-0.4231739938,0.0233156718,0.3323550224,-0.0467140079,-0.2578987777,-0.1601985991,-0.2706966102,0.0933790952,0.073723577,-0.1934672594,0.1603425294,0.243669048,0.1488006413,-0.1971042007,0.0145643512,0.1307732165,-0.1473884583,0.1625458449,-0.0961497277,0.1292523295,-0.0469140559,-0.2002084106,0.1997577399,-0.1230454147,-0.1088268906,-0.0366090201,-0.1214845404,-0.1047692671,-0.1175522581,0.0085352669,0.0045816242,0.0854114592,0.0250084065,-0.0044681425,0.0385741219,-0.0387144461,0.1060211733,-0.0975360945,-0.2942665517,0.2391216606,0.2922899127,-0.1670206636,0.0466096327,0.3752070367,0.2724765241,0.0439519696,-0.0118864747,0.1884023398,0.3430142701,0.1958228052,0.2216542363,-0.1940376312,0.005526871,-0.2589847744,0.0628024861,0.8033874035,0.4351347685,0.3787606359,0.0590223223,-0.310072422,0.0016223151,0.1633840501,-0.0873767063,-0.1206291765,0.1915403157,0.3726432323,0.3853441179,0.3481548131,-0.0425087102,0.2032209188,0.2762518227,-0.1238583177,-0.1909551322,0.0339910053,-0.1010523364,-0.2211484015,0.1905425787,-0.5782694817,0.3371144533,0.0041400166,0.1818534732,-0.0196053367,-0.1397617459,-0.0293810517,0.208451733,0.1059032381,-0.0174112022,-0.0055940892,0.2592740357,0.0706541762,-0.0856020451,-0.2151544392,0.2720209062,0.4728945494,-0.173538819,0.1442514658,-0.1528486609,0.2518606484,-0.0047391434,-0.2206519246,-0.08354415,-0.2356707305,-0.0855709612,0.1491017044,0.1920591593,0.1763487011,0.0780320093,-0.0344592072,-0.0481045842,0.079381533,-0.0425899513,-0.0353820994,-0.2902143598,-0.1774063557,0.3210521638,0.0619787723,0.0119390367,0.0396472737,-0.311113894,-0.0545602813,-0.5426447392,-0.0045083473,0.0185830761,-0.1482398212,-0.00013115,-0.0861114264,-0.5322773457,0.0132138543,0.1309877932,0.1258892864,-0.313397795,0.1124276817,-0.2538447082,-0.3051471412,-0.123465769,0.1048952416,-0.0565843396,0.2436976582,0.1446895301,-0.1881213784,0.194577843,-0.0590647049,-0.0902348459,0.1419096142,-0.0224593785,-0.348313868,-0.0931343883,-0.212852478,0.2848571539,-0.0747310668,0.093404673,-0.0979233235,-0.0383979939,-0.0665078163,0.2167899758,-0.0050204648,-0.0108644944,-0.1919534802,0.1102564186,-0.1700980514,-0.0288522206,0.3901928663,0.2208699882,0.0089336941,0.186425671,-0.2493094802,-0.1036614627,-0.0369431302,-0.0451333262,0.0267364755,0.2068388909,0.138674438,0.8986776471,0.2649667263,-0.3478029668,-0.1407364905,-0.1692233533,-0.1681494415,0.0339967646,-0.0702737719,0.1933213323,0.2878656983,0.0653432906,0.3337760866,0.0477307178,-0.4255113006,0.0757419541,0.0679423735,-0.200989306,-0.1883058697,0.1098383293,-0.3344598711,0.3503893316,0.036955066,0.1809231788,-0.3938423395,-0.3281774223,-0.1703460217,-0.012721193,-0.0705465972,-0.4317488074,-0.7411492467,-0.1603211761,-0.8604405522,0.3774619997,0.1138947085,0.4273203611,0.4424976707,-0.0496721156,0.20070301,-0.1443955451,0.9118052125,-0.235482201,-0.1320429593,0.0651303753,-0.2948298156,-0.2825039029,-0.1848840415,-0.135006249,0.5606465936,0.4132285416,0.8910484314,-0.2881854177,-0.3910267651,0.0918288901,-0.0542336218,0.2368577868,-0.0117616151,-0.2989796996,0.0217598006,0.0537792146,0.3063639998,-0.1271662861,0.2147834748,-0.1256704926,-0.2243696153,-0.1045870408,-0.0579674095,0.322206825,0.0160490572,0.2576271594,0.2704817355,0.2573385835,0.2526003718,0.2224618196,0.4124114215,-0.3274652362,-0.0457701869,-0.0760485604,0.0832426548,-0.0973727107,0.4350222051,0.5226225853,0.0872128531,0.273335129,0.0553377569,0.4113700688,-0.0891622379,0.3040414453,0.5810800791,0.2233854085,-0.67451334,-0.1365586221,0.054813128,0.4843259752,-0.117183961,0.1027116552,-0.4373593926,-0.1425437331,0.0541655086,0.0916486159,1.1344281435,-0.3912650347,0.2987220585,-0.335300684,0.2050500661,0.2643357217,-0.2932571173,0.1459595114,-0.2696007788,0.0886798277,-0.0223327633,-0.0719436854,0.2274925858,0.4449830353,-0.0161430705,0.2005712986,-0.0961376056,0.3847248554,-0.1948966384,0.2793486416,0.5195389986,-0.360670805,0.0977197438,-0.0962956697,0.1068647578,-0.0045545292,0.0402074046,0.070455201,-0.123521328,-0.2152327895,-0.1378235668,-0.2195775062,-0.2363108993,0.1335582435,0.1438664049,-0.0820709616,-0.0600903854,0.1567383856,0.1476810873,0.3711024821,-0.0738878325,0.0373583362,0.0624030977,0.1400870383,0.1169228777,-0.4759300351,-0.1420603395,0.0166001283,0.0216810405,-0.051134374,-0.210413307,0.0044902824,-0.2595064938,0.1267155856,0.6846943498,0.2058439106,-0.0773865134,0.2871789932,-0.1554269344,-0.1291231662,-0.0603239052,-0.0261623003,-0.1896750629,0.4373886287,-0.0392108485,-0.5450547338,-0.0115869036,0.5462257862,0.1766413599,-0.1305155754,0.5346004367,-0.3273348808,-0.2365207523,-0.0676724985,0.3629681468,0.0064804605,0.0368449613,0.0092674037,0.3748605549,-0.0010217542,0.3156609237,0.0579906814,-0.3333629668,0.1151412949,-0.133559078,-0.1369019896,-0.2969364822,-0.2305311412,-0.1561617404,-0.0337158777,0.0789394081,0.2386427075,-0.3597070873,-0.1117150113,-0.1031237841,0.1722776145,-0.1716838628,0.3474082351,-0.1470310986,-0.0098967729,-0.0367307924,-0.0710790306,0.0576293804,0.1206152588,-0.0915222168,-0.0774667338,-0.1411896795,0.2497136146,0.0052958173,-0.3966466486,0.2849988937,0.0630081668,-0.3061583638,-0.3085699379,-0.0362928286,0.313231945,-0.0114058862,-0.1323159933,0.3741413355,0.1712273061,-0.1985106468,0.236048609,-0.1497827917,0.0487589575,-0.0171281993,0.3821556866,0.2901320159,-0.0776533559,0.1234232485,-0.0667800903,0.2899846733,0.1601424515,-0.0129377712,-0.0384360589,-0.0291012861,0.1701569855,-0.0386611484,0.2133072764,0.0034069312,-0.4337036014,0.5494694114,-0.0024498249,-0.0012663696,-0.2942692339,0.2426232249,-0.0435251296,0.020686239,0.1721123308,0.2906180322,0.3161354661,-0.1133483127,-0.0637254491,-0.0128461933,-0.0209684223,0.083392337,0.1557510346,0.1571674049,0.0776256546,0.3752671778,0.00210487,-0.0302339066,0.3816519082,0.263407141,0.529810071,0.4195980728,0.4187683761,-0.4285306036,-0.2606981695,-0.1719605029,0.5843432546,-0.0457376949,0.1622938365,-0.1014736518,0.2947352529,-0.1984384358,-0.4319773912,0.1838297844,0.381729275,-0.2996103764,-0.4777976573,-0.1748476624,-0.0874692425,-0.0691168904,0.0974077582,-0.2252933681,0.1270912588,0.3971890807,0.105320178,0.0954690129,-0.1167217717,-0.0512644649,-0.0514086597,0.356826216,-0.0892328545,0.2188121527,-0.5762058496,0.1682055295,0.0820517316,0.0075193387,0.4137625396,0.6114664674,-0.3368606865,-0.0594518669,0.0809926391,0.0561208613,0.2254053205,0.2264375091,0.1319906712,-0.3040263355,-0.0107807396,-0.2309982926,-0.1472589672,-0.3216590881,0.073569715,-0.118131198,-0.0113394884,0.1541370004,-0.096338585,-0.2762731016,-0.0361941829,0.1771428585,-0.1627211273,-0.0665556043,0.2220119983,-0.2043383271,0.0730709359,0.0440650694,-0.0161616225,-0.2055457532,0.436306417,0.1931465715,0.0087999217,-0.5149211287,-0.2020198852,-0.42839396,0.1290433556,-0.3554841876,0.1211769581,-0.208888486,0.1035784781,-0.0451578014,0.0453556515,0.0580852255,-0.0680925101,-0.1239583194,0.2289163321,-0.3094981909,0.5009469986,-0.2956531346,-0.2376540899,-0.0127015999,-0.4991868138,0.4623540342,0.0846226886,-0.1847005337,-0.0811260194,0.1874538809,-0.0802225471,0.2932308614,0.3498600721,0.1159156635,0.3047508001,0.0878460407,0.0769958571,0.0760035738,0.3194638789,-0.1371795982,0.2736652792,-0.0183784813,-0.089285925,-0.145326972,0.0888000652,0.0161174592,0.009906644,0.2922938466,-0.0905246958,-0.1410578042,0.2335031629,-0.0231722835,0.0249358434,-0.1679146439,0.2333766818,0.2141359448,-0.0457365587,-0.2683860064,0.0445177704,0.3343248963,-0.4197914004,-0.1525999159,-0.6207870841,0.1547615528,0.1683097482,0.1688883007,-0.1889675856,-0.2531668842,0.5244745612,-0.1408542097,-0.4507493973,0.2982175946,-0.0810266361,-0.237361908,-0.3806575835,0.0726999566,-0.1702402085,-0.1334683746,0.009952019,-0.1878964752]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2294","title":"Slow #0 when using map to tokenize.","comments":"That makes sense ! You can indeed use `map` on both datasets separately and then concatenate.\r\nAnother option is to concatenate, then shuffle, and then `map`.","body":"Hi, _datasets_ is really amazing! I am following [run_mlm_no_trainer.py](url) to pre-train BERT, and it uses `tokenized_datasets = raw_datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=args.preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=not args.overwrite_cache,\r\n )` to tokenize by multiprocessing. However, I have found that when `num_proc`>1\uff0cthe process _#0_ is much slower than others.\r\nIt looks like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/31714566\/116665555-81246280-a9cc-11eb-8a37-6e608ab310d0.png)\r\nIt takes more than 12 hours for #0, while others just about half an hour. Could anyone tell me it is normal or not, and is there any methods to speed up it?\r\n","comment_length":26,"text":"Slow #0 when using map to tokenize.\nHi, _datasets_ is really amazing! I am following [run_mlm_no_trainer.py](url) to pre-train BERT, and it uses `tokenized_datasets = raw_datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=args.preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=not args.overwrite_cache,\r\n )` to tokenize by multiprocessing. However, I have found that when `num_proc`>1\uff0cthe process _#0_ is much slower than others.\r\nIt looks like this:\r\n![image](https:\/\/user-images.githubusercontent.com\/31714566\/116665555-81246280-a9cc-11eb-8a37-6e608ab310d0.png)\r\nIt takes more than 12 hours for #0, while others just about half an hour. Could anyone tell me it is normal or not, and is there any methods to speed up it?\r\n\nThat makes sense ! You can indeed use `map` on both datasets separately and then concatenate.\r\nAnother option is to concatenate, then shuffle, and then `map`.","embeddings":[-0.4388077557,-0.275254041,-0.0035829467,-0.0195050742,0.037655551,-0.1688462198,0.384619832,0.2470972836,-0.2695715129,0.0726477057,0.1870826185,0.3446471393,-0.1986020654,0.0340104513,-0.3061349988,0.0878567547,0.116552256,0.1894627661,0.2597584724,-0.0976634547,-0.1504402161,-0.0872968882,-0.4416112602,0.0910414532,-0.4734842181,-0.08449734,-0.0814721584,0.0456183404,-0.1439362466,-0.2998284101,-0.22646299,0.1087268293,-0.0366152525,0.3386248946,-0.0001279823,-0.232973814,0.2126871049,0.2498805076,0.0457996801,-0.0689349771,0.0412353016,-0.3254992962,-0.1309286058,-0.0187025275,-0.0309802908,-0.0179256815,0.1673597693,-0.1152555197,0.1905249357,-0.0400749259,0.0446927622,0.3407954872,-0.3702794313,0.2314551175,-0.0591440089,0.1832049042,-0.0711284131,-0.0657924339,0.1532546282,-0.2268456072,0.1455459893,0.1906764358,-0.2239545584,-0.0393581092,0.1851701438,0.1328697056,0.4181729555,-0.427369833,0.2814035416,0.2699203491,-0.1735964715,0.1514287293,-0.3362787962,-0.3355287313,0.0419164561,-0.3565230668,-0.0188186411,-0.0346324518,0.0667284355,-0.0074808202,-0.5250094533,0.1170422137,0.1661503464,0.0660812631,-0.3220513761,0.6341160536,0.1833461523,0.1864078939,0.1982421875,-0.2674390376,-0.325796634,-0.3126784563,0.1464710087,0.4829999804,-0.2674062848,-0.1194259301,-0.0261692889,-0.2011550963,-0.1995183825,-0.3869157434,-0.4060808718,0.3477786183,0.0074185971,0.0792026371,0.1178821549,0.0061186613,0.0867659748,0.3740830421,0.256185919,-0.4760053456,-0.2496020198,0.1057588756,-0.3221828341,-0.0576486103,-0.1828142852,0.0616205037,-0.2615204453,-0.0441681147,-0.1043375283,-0.0334302895,-0.4193731844,0.0477108434,0.3961853385,0.0800352916,0.0011543407,0.5042101145,-0.3501891494,0.0108959228,-0.4182539284,0.1281474382,-0.0084887184,-0.1850924343,-0.2621052563,0.1198893636,-0.0490981117,0.2026842833,0.0051389625,0.1334311366,0.1958023161,0.2225049287,0.4963684976,-0.2874108255,0.2397117168,0.2287209332,-0.1848647147,0.4846333861,-0.0675850436,0.1625760794,-0.2317890823,0.4496634603,-0.4836660624,-0.3581535518,0.0886901096,-0.0503153019,-0.0290916655,0.2750347555,-0.2045658231,0.4464249611,0.3693268895,-0.0993712246,-0.2557810247,-0.1525740325,-0.5478855968,-0.0989280045,0.2410939187,0.2521885633,-0.1584029645,-0.0110206315,0.1865359247,0.36947456,0.6669709682,0.5067364573,-0.2878709435,0.1310139894,-0.1607699543,0.6380207539,0.0407219082,-0.1091928631,-0.1088482216,0.3064481616,-0.4410433471,-0.0714501962,0.0128975352,-0.0292178765,0.1942563504,-0.1717973202,0.3794989586,0.0860753357,-0.1242235005,0.28827402,-0.4205456674,0.0442052595,0.3506424725,-0.083741039,-0.2194923162,-0.1622108221,-0.2516686916,-0.0266855694,0.0405645408,-0.1992124617,0.1669294089,0.0959291086,0.1650446355,-0.1816592366,-0.0439745672,0.1550924629,-0.0143420119,0.189425081,-0.1078041866,0.1165861934,-0.0158081129,-0.118895188,0.2822369337,-0.1358155012,-0.0231440477,0.0286241043,-0.1187128574,-0.1488008797,-0.1367196739,-0.0344713107,-0.0475839451,0.0045012212,0.0934303701,-0.0588060208,0.1493183523,-0.0622188859,0.1166628376,-0.1637578905,-0.2954397202,0.2489109933,0.2838457525,-0.1420406699,0.1078457758,0.3962527812,0.2024593502,0.0529389866,-0.0186137334,0.1946263313,0.3567746282,0.2995521426,0.2858372629,-0.3156941235,-0.0897186249,-0.2797776461,0.1507816315,0.8951528072,0.3780685961,0.3151248097,0.056814909,-0.2766118646,-0.004333124,0.2213652581,-0.0341794752,-0.0283556078,0.2033922672,0.3341964185,0.3051337004,0.2553846836,-0.0430152416,0.2804568112,0.2164853215,-0.09988828,-0.195873484,0.0009245285,-0.0914533958,-0.1556072384,0.1186439469,-0.4987288415,0.311635524,0.0389192775,0.249767676,0.001279352,-0.1723957956,-0.0137270037,0.1406293809,0.1262094676,-0.0237505399,-0.0348012596,0.2234736234,0.0336094499,-0.0484149568,-0.1124185473,0.3035728931,0.4295854867,-0.2028124034,0.0842254907,-0.1635441333,0.2961604595,-0.0375076123,-0.2037374824,-0.0970498621,-0.2546182871,0.0398239121,0.1681345105,0.0835213065,0.1886325032,0.0789038762,-0.1020120904,-0.1018423811,0.1730253696,0.0364545546,-0.0307044536,-0.2684114575,-0.1226452887,0.3741773069,0.0297645666,0.0512430444,0.0608648248,-0.3349949121,0.081458278,-0.5531005859,-0.0262767766,-0.0215872861,-0.1566880047,0.0560895838,-0.0829023719,-0.6097153425,0.0379511863,0.1915591359,0.0100364015,-0.3099003136,0.0293184351,-0.2165517062,-0.3161248863,-0.1767939627,0.1241433322,-0.0743780211,0.2454777658,0.1708730608,-0.2345411479,0.2101939917,-0.1212064549,-0.0949990898,0.1285719723,-0.1188468859,-0.4825815558,-0.078733772,-0.2240341902,0.3454927504,-0.0826594755,0.0813403726,-0.0067228428,-0.0524938256,-0.1030180603,0.2512447536,0.1546482295,0.0125445165,-0.0915634334,0.0958880857,-0.0917695835,0.0141655933,0.4137443304,0.1844600439,0.028602099,0.1182156801,-0.3019368351,-0.0392446704,0.0395163372,0.0175780077,0.0223320518,0.2818388343,0.1994910836,0.9364292026,0.1769333631,-0.2471613586,-0.1466753483,-0.1311273128,-0.1864523888,0.0355857946,-0.0280281082,0.2754463255,0.3319622278,-0.065449208,0.3055554032,0.0692320168,-0.4447918832,0.0683288649,0.065526031,-0.1484705061,-0.1748000681,0.1976065487,-0.292121172,0.2222035378,0.0639033243,0.2113481611,-0.4197668135,-0.2751148641,-0.1789607257,-0.0849712342,-0.0347942412,-0.3862010539,-0.7608810663,-0.1851507127,-0.8920628428,0.4031721056,0.1285987943,0.4056979418,0.4040029049,-0.1473227143,0.1809919626,-0.1462071389,0.9515538216,-0.3148618937,-0.101819329,0.0942880958,-0.2907163203,-0.2164988965,-0.1962091923,-0.054416351,0.4924846292,0.4774306715,0.7733253241,-0.1962785423,-0.3250955045,0.0093723079,-0.1400545835,0.2199729085,-0.0067697,-0.2870981991,0.0898113474,0.1091538891,0.2544772923,-0.0445640944,0.1162588522,-0.1389713734,-0.1802916825,-0.0156381689,-0.0378983915,0.3263306618,0.0452067964,0.2420582175,0.2547127306,0.2709084451,0.4144828618,0.2233682722,0.3825790286,-0.3118927777,-0.0680493638,0.0341820978,0.1353441924,-0.092942059,0.4539326727,0.549808085,0.1277031898,0.1894687563,0.0446932912,0.3629505932,-0.0970467925,0.2474793941,0.5302976966,0.2229895294,-0.6152862906,-0.1870729476,-0.0418677777,0.500403285,-0.0977363512,0.1358559132,-0.4847735763,-0.1182470769,0.0849323645,0.0773488358,1.0334544182,-0.5148515701,0.2964819074,-0.3219099641,0.1757460684,0.2100131363,-0.2687895894,0.1971257329,-0.2429300398,0.0711006224,-0.0187859759,-0.0614188462,0.2567293346,0.4876919687,0.0485899113,0.1938780844,-0.0276017319,0.3981177211,-0.2274305373,0.0628070086,0.5450242758,-0.3392467797,0.0779131725,-0.0579876192,0.2245115936,-0.0643875822,0.0137376739,0.000811765,-0.1434345245,-0.1669957638,-0.1103082374,-0.2305851579,-0.2144282758,0.0163937099,0.1533913016,0.00400159,-0.0636673793,0.0405462682,0.1298915446,0.2966968417,-0.0708946288,0.0827691555,0.1352577507,0.1543612927,0.1289828718,-0.4702272713,-0.1566708237,0.0467622802,0.0551219508,-0.1249579042,-0.1654784679,-0.0354792401,-0.3376195431,0.1171024442,0.6711376905,0.283816576,-0.0546955727,0.3056975007,-0.0777949989,-0.1577309519,-0.0371344499,-0.0486759245,-0.1908678859,0.4021508992,-0.2150208354,-0.4364954531,0.0072038961,0.5445734262,0.1661619246,-0.1699452996,0.4865759611,-0.2501218319,-0.1797344834,-0.0319511108,0.3752322793,0.1716775447,0.1371528655,0.0203050282,0.3294658661,-0.0705557615,0.3392579556,0.0268393289,-0.2536615729,0.0797123536,-0.1415092498,-0.1076898202,-0.3482828438,-0.2523837388,-0.1526265442,-0.0975807831,0.0162703879,0.2962122262,-0.2722663581,-0.0762864128,-0.129494071,0.0739231557,-0.0707213506,0.3378512561,-0.1955773234,-0.0156830996,-0.0153119499,-0.1194767132,0.0632915795,0.195832625,-0.1811698824,-0.0880896151,-0.0486477092,0.2328052521,0.0942495763,-0.413282901,0.3264922202,0.0742444471,-0.3295325935,-0.295763135,-0.0713587627,0.376973331,-0.0888149291,-0.2104261518,0.4065176845,0.131271258,-0.0231064595,0.1755511761,-0.1344916075,0.0307519604,-0.0640719831,0.3533712029,0.271412462,-0.080688566,0.2277092636,-0.147152856,0.2102837116,0.1438198388,-0.0448984839,-0.0448772088,-0.0620039664,0.1628134698,0.0859895498,0.2296501249,0.0315996706,-0.3968471289,0.6052674651,0.002599247,-0.0253975112,-0.2075587809,0.304780364,0.1090018079,0.044123739,0.1556784064,0.2953764498,0.2974738479,-0.2237291485,-0.0349692851,-0.1236262023,-0.1364785731,0.0851320922,0.2014510632,0.0931705758,0.0903859809,0.3620444834,-0.0252307151,-0.047321517,0.422205627,0.2469168603,0.5134435296,0.3462409973,0.3657308519,-0.3900961876,-0.3400643468,-0.200199604,0.6330743432,0.0776188225,0.2425194383,-0.0193521883,0.277751416,-0.1924544424,-0.4998669922,0.1787226051,0.4119102359,-0.3552468419,-0.4772709608,-0.1415457875,-0.1361977458,-0.0916011035,0.1225053221,-0.2288359553,0.1500305235,0.4793517888,0.0331876092,0.1832872182,-0.0671393275,-0.180501774,-0.1006516665,0.3039552271,-0.0721868053,0.1637684852,-0.5421501398,0.0823062137,0.2256791443,-0.052955173,0.351405412,0.6039911509,-0.3567196429,0.0075332387,0.0174760725,0.0715858117,0.1786415875,0.2215700001,0.0940288827,-0.2806364894,-0.0183808506,-0.2068704814,-0.1642174572,-0.2842901349,0.0092654601,0.0029797929,0.1242677793,0.1480992138,-0.0966764465,-0.2055516243,-0.0035365555,0.1343883425,-0.1567650884,-0.0454964899,0.0952081978,-0.2115180641,0.0895369723,0.0734497756,-0.0024370567,-0.1321172565,0.3498377204,0.220988825,-0.1230856851,-0.4596059322,-0.1800464243,-0.5000241399,0.1425477266,-0.4191339016,0.1426107138,-0.2712028027,0.1326361597,0.0476999991,0.0183829311,0.1398370862,-0.0291795842,-0.078336142,0.2604596913,-0.2813761234,0.5002378821,-0.2853158116,-0.2422318161,-0.0138350744,-0.4122861028,0.4706130326,0.0631023198,-0.1594393849,-0.0784423947,0.2328397632,-0.1661154628,0.2907161117,0.2665555477,0.1540990174,0.3074619472,0.0896098167,0.1229654551,0.0203498863,0.3362579048,-0.244001016,0.2080632001,-0.0524313524,-0.0398024693,-0.2057079822,0.0260170996,-0.0121332956,-0.1082280502,0.2541697919,-0.0816439167,-0.2034932524,0.2863993943,0.0140948109,0.0633784458,-0.2183950841,0.2291814536,0.2644093931,-0.0783021674,-0.2523019612,0.1403753608,0.3205935061,-0.5568313599,-0.2029646039,-0.6513800025,0.0344411917,0.1288051754,0.1286983043,-0.2269717008,-0.2931537926,0.3912018239,-0.1233502701,-0.3294096291,0.2369915247,-0.1250512004,-0.2932345271,-0.3721488714,0.0813701674,-0.1436474174,-0.1949796677,0.0596726052,-0.1216476262]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2288","title":"Load_dataset for local CSV files","comments":"Hi,\r\n\r\nthis is not a standard CSV file (requires additional preprocessing) so I wouldn't label this as s bug. You could parse the examples with the regex module or the string API to extract the data, but the following approach is probably the easiest (once you load the data):\r\n```python\r\nimport ast\r\n# load the dataset and copy the features\r\ndef process(ex):\r\n return {\"tokens\": ast.literal_eval(ex[\"tokens\"]), \"labels\": ast.literal_eval(ex[\"labels\"])}\r\ndataset = dataset.map(process, features=new_features)\r\n```\r\n","body":"The method load_dataset fails to correctly load a dataset from csv. \r\n\r\nMoreover, I am working on a token-classification task ( POS tagging) , where each row in my CSV contains two columns each of them having a list of strings.\r\nrow example:\r\n```tokens | labels\r\n['I' , 'am', 'John'] | ['PRON', 'AUX', 'PROPN' ] \r\n```\r\nThe method, loads each list as a string: (i.g \"['I' , 'am', 'John']\").\r\nTo solve this issue, I copied the Datasets.Features, created Sequence types ( instead of Value) and tried to cast the features type\r\n```\r\nnew_features['tokens'] = Sequence(feature=Value(dtype='string', id=None))\r\nnew_features['labels'] = Sequence(feature=ClassLabel(num_classes=len(tag2idx), names=list(unique_tags)))\r\ndataset = dataset.cast(new_features)\r\n```\r\nbut I got the following error \r\n```\r\nArrowNotImplementedError: Unsupported cast from string to list using function cast_list\r\n```\r\nMoreover, I tried to set feature parameter in load_dataset method, to my new_features, but this fails as well.\r\nHow can this be solved ?","comment_length":72,"text":"Load_dataset for local CSV files\nThe method load_dataset fails to correctly load a dataset from csv. \r\n\r\nMoreover, I am working on a token-classification task ( POS tagging) , where each row in my CSV contains two columns each of them having a list of strings.\r\nrow example:\r\n```tokens | labels\r\n['I' , 'am', 'John'] | ['PRON', 'AUX', 'PROPN' ] \r\n```\r\nThe method, loads each list as a string: (i.g \"['I' , 'am', 'John']\").\r\nTo solve this issue, I copied the Datasets.Features, created Sequence types ( instead of Value) and tried to cast the features type\r\n```\r\nnew_features['tokens'] = Sequence(feature=Value(dtype='string', id=None))\r\nnew_features['labels'] = Sequence(feature=ClassLabel(num_classes=len(tag2idx), names=list(unique_tags)))\r\ndataset = dataset.cast(new_features)\r\n```\r\nbut I got the following error \r\n```\r\nArrowNotImplementedError: Unsupported cast from string to list using function cast_list\r\n```\r\nMoreover, I tried to set feature parameter in load_dataset method, to my new_features, but this fails as well.\r\nHow can this be solved ?\nHi,\r\n\r\nthis is not a standard CSV file (requires additional preprocessing) so I wouldn't label this as s bug. You could parse the examples with the regex module or the string API to extract the data, but the following approach is probably the easiest (once you load the data):\r\n```python\r\nimport ast\r\n# load the dataset and copy the features\r\ndef process(ex):\r\n return {\"tokens\": ast.literal_eval(ex[\"tokens\"]), \"labels\": ast.literal_eval(ex[\"labels\"])}\r\ndataset = dataset.map(process, features=new_features)\r\n```\r\n","embeddings":[-0.110432297,-0.0460996181,0.0017928476,0.0747474656,0.4480468929,0.0627773106,0.4552136958,0.219682008,0.299490124,-0.1329664886,0.206075266,0.4624345005,-0.0728465244,0.0797176287,-0.0546839014,-0.225862965,0.1463976949,0.121658057,-0.068167083,-0.0514059886,-0.2214132249,0.1172490567,-0.2457354516,0.0626452342,0.1473498046,0.067499727,0.0444802754,0.1194713861,0.1582894623,-0.4119055867,0.1597045362,-0.1485556513,0.3543725908,0.2177936286,-0.0001225974,-0.0308731589,0.2000939697,-0.0455260463,-0.1343298107,-0.4299833477,-0.3686624765,-0.4428124428,0.6035311222,-0.2337204367,0.0679722503,0.0058746655,-0.1318276227,-0.4414084256,0.3704528511,0.4917436838,0.1028463691,0.2604970336,-0.2128971964,0.0683712736,0.2531427145,0.172960192,-0.2683466673,0.0115083726,0.2413832992,0.2831138074,0.0202589948,-0.251404494,-0.3278875947,-0.1090300754,0.4502304494,0.1004333869,-0.1728934497,-0.2152026296,-0.0185692199,0.0967391878,0.6940967441,-0.0439019613,-0.3437823057,-0.1345560402,0.1029636487,-0.3265565336,0.0719097704,0.0394793227,-0.0936724469,0.2009950578,-0.2288368642,0.0179859232,-0.0210507773,0.2387349755,0.0580496825,-0.0684755296,-0.1830670089,0.2108053714,-0.0479418114,-0.2082449645,0.158677578,-0.2253193408,0.1418036073,0.307048142,-0.4954456389,0.1484595686,0.1250777394,-0.1021132842,-0.0821916088,0.2027681023,0.314424485,0.1013435349,-0.0861858726,0.2051599771,0.2729567289,0.2770818174,0.2978345454,0.0810159221,-0.0635052919,-0.0946139917,-0.1663808078,0.0363045745,-0.3185822666,-0.1923478991,0.3789396882,0.1587801576,0.1375382096,-0.3657591343,-0.362699002,0.2351652086,-0.2413615137,-0.0889683217,0.0760768577,0.283634156,-0.129787311,0.4884750247,0.1027978286,0.3484078944,0.1101016402,-0.1406168938,-0.161834538,-0.1079706252,-0.0891124681,-0.0602726191,0.0965371057,-0.3631424606,0.1043661162,-0.0351545699,0.0497357994,-0.1586531699,-0.2356229573,-0.0306247156,0.003865544,0.2680171132,0.0140141835,0.0956334621,0.3562788367,-0.04595403,-0.2184284329,0.3119084239,-0.4620773494,-0.2153678834,0.0045838011,-0.0065716007,-0.2278196365,-0.0331734493,-0.4317633212,0.0690495893,0.1968415082,0.1540489942,-0.1759973019,-0.4027112424,-0.2068364173,-0.4180796742,0.1779424399,0.3996336758,-0.5163905621,0.0751355812,0.0115570929,0.0245534964,0.1286622733,0.4650031328,-0.1831560582,0.296323359,-0.1501670331,0.3829417825,0.5047297478,0.099883981,-0.0535182878,0.5566630363,0.063063778,0.0640882254,0.1053316817,-0.0127036516,0.3253757954,-0.1700782776,0.0269855931,0.4483290911,-0.1635106355,-0.0150684919,-0.1051956415,-0.1559162289,0.2475551218,0.1033510715,-0.1184359044,0.3426096737,0.1636872888,-0.0604913868,0.0419215485,-0.3142814338,0.1542034149,0.2057690024,-0.0115860142,0.1584032178,-0.049541723,-0.0531680845,-0.636657536,-0.0379679389,0.2213625461,-0.1010948569,-0.3996529281,-0.0541028827,-0.3353928924,0.2223334759,-0.0879020244,0.5470007658,-0.1008092389,0.1392119527,-0.0799395666,0.0280992948,-0.0997856706,0.0700337514,-0.2976451218,0.0392469056,-0.2093299776,0.0670426711,0.2348161191,-0.1278736293,-0.0986958519,0.0919288248,0.3017797172,-0.0927237272,-0.2805355489,0.3237709701,0.0724094808,-0.1806598604,-0.4057722092,0.0707104579,0.1225665361,-0.4118911028,-0.0960494876,0.3675079346,0.298291862,-0.1217671931,-0.1948543936,0.5355858803,-0.1639039963,0.3591678143,-0.1003533378,0.0166162923,0.4018522501,0.0990034863,-0.2214927226,-0.0735374466,0.1187479869,0.1199110448,0.2753313482,0.0832990184,-0.4843612611,-0.107410118,0.2010634094,-0.1397026777,0.1424913108,0.3510163426,-0.0368599556,0.1023043171,0.0236881711,-0.2591193318,0.3383775949,-0.0355219319,-0.0439354815,-0.0970787778,-0.0985544994,-0.0578270108,0.1779477745,-0.03336142,0.1140087545,0.237608552,0.0508311465,-0.0402042083,-0.2066465169,-0.1779197305,0.1371901035,0.1496103406,-0.6657977104,0.1570387781,-0.4116671085,0.2593963742,-0.283072114,-0.0852635726,0.1355614513,-0.1846075803,-0.2917081416,0.2667333186,-0.1641355902,0.1261749268,-0.5663655996,-0.0178402979,0.2130135149,-0.6803790927,0.1811012328,0.0629809722,-0.1782786697,-0.076444611,0.3991116583,0.2297205627,0.0199138876,-0.0418583006,0.0285445899,-0.1528597027,-0.1876204312,0.1119922847,-0.1340637058,0.188344717,0.1034669206,0.060844101,-0.2700412869,-0.5033234954,0.2780170739,-0.1225630417,-0.1226960272,0.4148798585,-0.0341416374,-0.3602371514,-0.2041586041,-0.4714085758,-0.1735008657,-0.3078493476,0.1805062592,0.1036359519,-0.0178940371,0.0935265049,0.2594175637,0.0780491531,0.2604344487,0.043624308,-0.0324880071,0.0536315925,0.5690494776,-0.1156726182,-0.4092385471,0.0243679937,-0.202160418,-0.1955375969,0.3714095056,-0.1427942961,0.1118415743,-0.0826797634,0.2352448106,0.0215764083,-0.0280712508,0.3271402717,-0.030869728,0.0953569934,-0.1797703654,-0.2646719515,0.1653976887,0.2676552832,0.0848531201,0.5000003576,0.6963592172,-0.2381380945,0.4044570327,-0.0274050087,-0.1215998381,0.4540215731,-0.1748410761,0.0772479624,-0.1671363413,-0.2304948568,-0.2067226321,-0.2361323386,-0.3253380358,0.3207266629,-0.1002926975,-0.3117117882,-0.1034622118,0.3674573302,-0.3126691282,-0.203471303,0.3986499906,-0.2290899605,0.3204396665,-0.0259240717,0.1603925079,-0.0733196661,-0.149263978,-0.0054993816,0.1686309427,0.1123344451,-0.1163133383,-0.4571547806,-0.0342421904,-0.3056702912,0.1775715351,0.3312333822,0.5592125058,-0.0358422957,-0.1711426079,0.1063516662,0.108345829,0.390116781,-0.2086946517,-0.0011614894,0.2523574829,0.039778281,-0.0906488672,0.009187324,-0.1796163321,-0.0382781699,-0.0817602873,0.133976087,-0.236657232,-0.1357567161,-0.1948538572,0.3967040181,-0.2197418213,-0.0785213411,-0.1300492734,-0.3525593877,-0.2233395278,-0.0077143512,0.2062229812,0.4148727357,-0.0111159123,0.1392092109,-0.4173452854,-0.0830807686,0.3357973397,0.0164366197,0.3778498769,0.0730972961,0.2380180657,-0.2853320539,0.0951169431,-0.0547423065,0.7008427382,0.018754473,-0.8055835962,0.1155392826,-0.234724462,0.610119462,0.1196998581,-0.3118107915,0.0897006691,0.0616618507,-0.0284202173,0.0239184536,0.0324660204,0.1782879233,-0.1129926071,-0.0735789314,-0.5334360003,0.2675793171,-0.0179276858,-0.0700028837,-0.1552603096,-0.0905959159,-0.3439866006,0.5822557807,0.1116436869,0.8149862289,-0.158609122,0.0223722607,0.1512666494,0.1094425544,0.3992291987,0.4279824793,-0.0887042731,-0.3721043468,-0.132212922,-0.1304240972,-0.0532289445,0.2831767499,0.3073975444,-0.1827316731,0.3941601813,-0.2625230849,0.4788019359,0.1579052806,0.0136174373,0.3976051807,-0.561057508,-0.4361240268,0.0566091835,-0.0795366243,-0.0500404686,0.1793603152,-0.1025282666,-0.0147744771,-0.0690700263,-0.167157799,0.2023209184,-0.0245170947,0.0333176404,0.0225514397,-0.3911602497,0.3321498334,0.3873806596,0.21401079,-0.123612076,0.0530750118,-0.1053180099,0.0653138831,0.1610916257,-0.1094210446,-0.172766,0.3233717084,0.2814852893,0.0633920208,0.1883973777,-0.1877054572,-0.4563293755,0.2772749066,0.0200755987,0.2457721233,-0.6801099777,-0.6067524552,-0.1620600373,0.1821791977,-0.0854545832,-0.0015260165,0.0692153573,-0.0136250416,0.104377985,0.0404799618,-0.1276173443,0.0908666253,0.143096894,0.0589276031,-0.1127821058,0.4481777549,0.3223393261,-0.0844491273,-0.0717108026,0.1627796441,0.2564382851,-0.5918012261,0.0548073389,-0.1395562887,-0.1452335268,0.0483500175,0.6411305666,0.0509413965,0.1388819218,-0.0750026777,-0.4760088325,-0.3894129694,0.1771916449,0.1975643933,0.0376206264,-0.053180743,-0.0236834586,-0.1459553093,-0.2336322665,-0.1492191702,-0.2112296969,-0.0947558433,0.1627303809,0.0821603015,0.2315099239,0.1650505364,0.0571824312,0.0496041849,0.0728442743,0.0216940213,-0.1048053131,-0.0801290944,0.2456323057,0.1700915992,-0.013329953,-0.2409716994,-0.2888040245,-0.1671981514,-0.3431462646,-0.1311102211,0.1529919058,-0.1177067831,0.0611686148,0.1232551485,0.1416044086,-0.2951571941,0.2083006948,-0.2207979411,0.3090303838,-0.0685164183,0.2035964131,0.0723570585,0.0867720842,-0.1694903076,0.1411750019,0.2775630057,0.0172086265,0.2658576667,-0.3038050532,0.10086371,0.1427964717,0.2968776226,0.4185534716,-0.0852780938,0.0391541459,0.3072123826,0.0333190337,-0.1571903676,-0.1819154024,0.186019823,-0.0447643958,-0.0816647336,0.354286164,0.1773703247,-0.1022181138,0.123380661,-0.0286831781,0.5112878084,-0.1949408054,-0.0704223737,0.6916535497,0.291329205,0.1611916423,0.1603681147,-0.013248085,0.1994768083,0.368275404,-0.1253437251,-0.0963837057,0.2946385443,0.4194568694,0.0165845435,-0.4054487348,0.2480338663,0.2353260219,-0.0219285153,0.2139557302,0.1303156614,0.2696859837,-0.1014121249,-0.2234062701,-0.1427471191,0.3004921973,-0.138213113,-0.3201159239,-0.3272539377,-0.1077628285,-0.2257195413,-0.0402907133,-0.1128794104,-0.1230407953,0.3907460868,0.1284091622,0.2790490389,-0.367477566,-0.307592839,0.054378204,0.0201761182,-0.2398434132,0.3100335896,-0.1282327175,0.0060277823,0.3204963207,-0.0867378339,0.3156220317,0.5036233664,0.079475753,-0.0671956763,-0.0450188108,-0.141057238,-0.0617945939,0.2614477575,-0.18359828,0.3418868482,0.2386571467,0.0206035972,-0.0517917424,0.4905465245,-0.1871995479,0.0214823429,-0.1375136822,0.3965120614,0.2406617999,-0.1396111399,-0.0916913226,-0.1429029107,-0.0435063988,-0.0867452621,0.4419136047,0.0340844952,0.2552061975,0.2211965024,0.0060107666,-0.080828622,0.1633162647,0.2484547496,0.1098901182,-0.2191248685,0.1393919289,-0.7467547655,0.0526853502,0.1535263509,-0.2677980959,0.4478868842,0.2577357888,0.2434037477,0.2736611068,-0.4172680974,-0.1067539603,0.1898745596,0.2864471078,0.0744639561,-0.2029294223,0.179116115,0.2065355331,-0.2200174779,-0.5413213968,0.0190438144,-0.1167133078,-0.0261460003,-0.1522412598,-0.2528543174,-0.0122338729,-0.0994630307,0.2671571076,-0.0228849594,0.5270901322,-0.0979368761,0.0292960182,-0.0202064756,-0.3900432885,-0.229401648,0.4323118329,-0.2207965702,0.494822681,-0.2900181711,0.4376683533,-0.2634708285,-0.0292548109,0.152831167,-0.1753899455,-0.4586544037,0.4043571055,-0.5012464523,0.1835056543,0.0770833194,0.1095325947,0.0006309397,0.2872672975,0.0253669955,-0.3708912432,0.3762868941,-0.557035923,-0.3053693473,-0.0174342673,-0.0726699233,0.0540473498,-0.0671416372,-0.4190599918,-0.1293233633,0.2275843769,-0.1773216724,-0.2662453949,0.0923742354,-0.0665552542,0.0066878968,-0.2397062182,0.2292441577,0.1236816496,-0.1978785694,-0.1641547978,-0.4658317268]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2288","title":"Load_dataset for local CSV files","comments":"Hi,\r\n\r\nThanks for the reply.\r\nI have already used ```ast.literal_eval``` to evaluate the string into list, but I was getting another error:\r\n```\r\nArrowInvalid: Could not convert X with type str: tried to convert to int\r\n```\r\nWhy this happens ? Should labels be mapped to their ids and use int instead of str ?","body":"The method load_dataset fails to correctly load a dataset from csv. \r\n\r\nMoreover, I am working on a token-classification task ( POS tagging) , where each row in my CSV contains two columns each of them having a list of strings.\r\nrow example:\r\n```tokens | labels\r\n['I' , 'am', 'John'] | ['PRON', 'AUX', 'PROPN' ] \r\n```\r\nThe method, loads each list as a string: (i.g \"['I' , 'am', 'John']\").\r\nTo solve this issue, I copied the Datasets.Features, created Sequence types ( instead of Value) and tried to cast the features type\r\n```\r\nnew_features['tokens'] = Sequence(feature=Value(dtype='string', id=None))\r\nnew_features['labels'] = Sequence(feature=ClassLabel(num_classes=len(tag2idx), names=list(unique_tags)))\r\ndataset = dataset.cast(new_features)\r\n```\r\nbut I got the following error \r\n```\r\nArrowNotImplementedError: Unsupported cast from string to list using function cast_list\r\n```\r\nMoreover, I tried to set feature parameter in load_dataset method, to my new_features, but this fails as well.\r\nHow can this be solved ?","comment_length":55,"text":"Load_dataset for local CSV files\nThe method load_dataset fails to correctly load a dataset from csv. \r\n\r\nMoreover, I am working on a token-classification task ( POS tagging) , where each row in my CSV contains two columns each of them having a list of strings.\r\nrow example:\r\n```tokens | labels\r\n['I' , 'am', 'John'] | ['PRON', 'AUX', 'PROPN' ] \r\n```\r\nThe method, loads each list as a string: (i.g \"['I' , 'am', 'John']\").\r\nTo solve this issue, I copied the Datasets.Features, created Sequence types ( instead of Value) and tried to cast the features type\r\n```\r\nnew_features['tokens'] = Sequence(feature=Value(dtype='string', id=None))\r\nnew_features['labels'] = Sequence(feature=ClassLabel(num_classes=len(tag2idx), names=list(unique_tags)))\r\ndataset = dataset.cast(new_features)\r\n```\r\nbut I got the following error \r\n```\r\nArrowNotImplementedError: Unsupported cast from string to list using function cast_list\r\n```\r\nMoreover, I tried to set feature parameter in load_dataset method, to my new_features, but this fails as well.\r\nHow can this be solved ?\nHi,\r\n\r\nThanks for the reply.\r\nI have already used ```ast.literal_eval``` to evaluate the string into list, but I was getting another error:\r\n```\r\nArrowInvalid: Could not convert X with type str: tried to convert to int\r\n```\r\nWhy this happens ? Should labels be mapped to their ids and use int instead of str ?","embeddings":[-0.1745119542,-0.1091788337,-0.0231068395,0.0980234146,0.4885460138,0.0214575231,0.4982636869,0.2223207355,0.3661426306,-0.0579060018,0.0228598081,0.5268843174,-0.0411034487,0.1786160618,-0.0871043503,-0.1635926515,0.1504180282,0.1215566248,-0.0954685286,-0.054811608,-0.2822064757,0.1153573394,-0.2327804267,0.1028901264,0.0758110136,-0.0212439336,0.1522944868,0.0589686707,0.1551152021,-0.46901685,0.0427116603,-0.1593906432,0.4239130914,0.1603819579,-0.0001222864,-0.0723953396,0.2014042437,-0.0017853752,-0.1700443029,-0.2721410692,-0.3334534466,-0.331261456,0.5561583042,-0.2470588535,0.1309711039,-0.0750061497,-0.0999568924,-0.4044760168,0.2009782791,0.496066004,0.1088149324,0.1556666493,-0.0642046854,0.0743320733,0.3011790812,0.1582485288,-0.2453445643,0.0072984123,0.2836193144,0.4058943391,-0.0172992852,-0.1583520472,-0.2349089384,-0.1760405749,0.5206285119,0.0213500671,-0.1103872508,-0.2337289751,-0.0291656051,-0.0270526726,0.824239254,-0.0337931812,-0.33137393,0.0284710489,0.0084755486,-0.4012273252,-0.0091966204,-0.0444252566,-0.0706204623,0.1964198351,-0.2307465374,-0.0202314667,-0.1065665931,0.2094974965,0.1264789253,0.0024114156,-0.1604218334,0.252599299,-0.0231931992,-0.3424344659,0.1272237003,-0.1139903888,0.0617731921,0.2001869828,-0.5123823285,0.0699720904,-0.0475859046,-0.0426837727,-0.0511405244,0.1655862331,0.213074103,0.1673607081,-0.0533780903,0.1695119143,0.2711187899,0.4393441677,0.3845515549,0.1459635645,-0.0772828162,-0.2162695378,-0.1859956682,0.0367971063,-0.3532292843,-0.1677093655,0.4503811598,0.1206073463,0.0453948155,-0.472183913,-0.3626649082,0.1549423635,-0.2265360206,-0.0372204147,0.1171711385,0.376806885,-0.1251847893,0.5413228869,0.1200025603,0.2771107554,0.1804116219,-0.1838897914,-0.1071730107,-0.095461227,-0.145713225,-0.0957704335,0.038471695,-0.3362178206,0.1143923029,-0.0716412216,0.0394222774,-0.1666861027,-0.2015030086,-0.1003284752,0.1008358672,0.2000790834,-0.0727956071,0.1101015136,0.2909413874,-0.0541023985,-0.170156166,0.3316213489,-0.5061308742,-0.2791664302,-0.0096540516,0.0039627338,-0.0279720332,-0.0403443314,-0.3437798023,0.1190858558,0.1428308189,0.0864046663,-0.2007548809,-0.4404556453,-0.1076720208,-0.3787831068,0.096455805,0.2499113828,-0.5306060314,0.1216105968,0.0757080913,0.0141585395,0.2510149181,0.5593862534,-0.146987617,0.2511620224,-0.1406387985,0.3639690876,0.5061859488,-0.0752420351,-0.068733789,0.5408250093,0.1516232789,-0.1511905044,0.0947824046,0.1008558571,0.3450995982,-0.1787752211,0.0259270333,0.3236461878,-0.109979175,-0.0535836369,-0.2369864881,-0.0890310258,0.2300912738,0.1359630972,-0.1365004629,0.327465266,0.1866754591,0.0384798944,-0.0829609558,-0.301882267,0.1766661853,0.1386438608,-0.1012358144,0.1233697087,-0.0199299064,0.0147775253,-0.6875210404,-0.1371833086,0.1024215445,-0.0769350678,-0.2303434759,-0.0767372102,-0.3774006963,0.232375443,-0.0619983934,0.5937697887,-0.1051987484,0.1890155822,-0.090512082,-0.0568342432,-0.1481353045,-0.0236116908,-0.3030292392,0.0087987315,-0.1092748046,-0.134216845,0.1625541449,-0.1477347761,-0.1353454292,0.0716530308,0.3193821609,-0.058781907,-0.3308487833,0.3488802314,0.0969928429,-0.2780359983,-0.3008566499,0.1070826054,0.0893097818,-0.5350093842,-0.0993565544,0.264462769,0.3447467685,-0.156527102,-0.1530912966,0.4296925068,-0.2420479357,0.4259821475,-0.1141403764,-0.0087803118,0.3594492078,0.131830126,-0.2028185427,-0.0955797583,0.0416045077,0.0982768312,0.3094275296,0.1073539555,-0.5325565338,-0.0246956144,0.2461336553,-0.1355324388,0.1485410035,0.3439357579,0.0264561884,0.058364328,0.1440448761,-0.377235949,0.3147376776,0.0168545246,0.0069153598,-0.1343993247,-0.1218231022,-0.0391922072,0.244708553,-0.0561150871,0.0706914142,0.2142902464,0.033025153,-0.1653952301,-0.1804197729,-0.2334323823,0.0790087804,0.1729777008,-0.6239628792,0.0942308307,-0.3296726048,0.1743192524,-0.237088114,-0.2676620185,0.1135046855,-0.2016695589,-0.2089027762,0.2420663089,-0.120127514,0.1385365874,-0.403313905,0.0975712985,0.1901009232,-0.5237894654,-0.0100405533,0.0020129271,-0.2710900605,-0.0818773434,0.3792653382,0.2798444629,0.0619555041,-0.1255289763,0.152325511,-0.1678494662,-0.2552996874,0.0680264086,-0.0739110559,0.1182409003,0.0621922649,0.0708518252,-0.324840039,-0.3971909881,0.2296409458,-0.10517288,-0.1522288769,0.3554173708,-0.104435727,-0.3573527336,-0.1748083234,-0.5161355734,-0.2136065066,-0.4151930213,0.1933624744,0.139304474,0.0469273292,0.0435752273,0.2623166144,0.1231550351,0.3312580287,0.1154701263,-0.0740698054,0.0053718071,0.5278599858,-0.1268101931,-0.3519244492,0.0330701508,-0.1132661849,-0.0612165332,0.310936898,-0.0418369621,0.1912265122,-0.0278596263,0.2648984194,-0.0023115284,0.0812982619,0.2223408967,0.0311561543,0.0761905238,-0.1425583065,-0.2828026116,0.3175474405,0.1232309341,0.1157069057,0.5898655057,0.6031092405,-0.3257270455,0.3851868212,-0.0492634512,-0.1964689195,0.4677959681,-0.1519373357,0.1013601273,-0.231947571,-0.1796561927,-0.1222321317,-0.1774987131,-0.2684704959,0.307085216,-0.1230671629,-0.3154534101,-0.1293951422,0.3553231359,-0.2967695892,-0.0969944522,0.3734416366,-0.2136843801,0.2447087914,-0.1190203875,0.0842543244,-0.0764391348,-0.1087259501,0.0913234055,0.1315589994,0.1068794727,-0.0873861685,-0.5244096518,-0.1038899049,-0.2857050896,0.2556534708,0.3292594254,0.655788362,-0.1432271898,-0.1253608465,0.0668188259,0.1294065714,0.3343352675,-0.2652457356,-0.0857283026,0.2183834314,0.0955885202,0.0212632064,0.0911455452,-0.2352445424,-0.1220686436,-0.0932944193,0.1723745167,-0.3121429682,-0.1891805232,-0.1928462833,0.3368740082,-0.1755696684,-0.0354147218,-0.1357356012,-0.3079069853,-0.2543926239,0.0501426011,0.2633685172,0.3873698115,0.064897716,0.0652583763,-0.4860838056,-0.0539547056,0.3168369234,0.0468862243,0.3901622593,0.0490335636,0.1119226813,-0.3145260811,0.1846647859,0.0357678123,0.6665095687,-0.0731327534,-0.8118278384,0.1926188022,-0.2284466177,0.6646200418,0.0639152378,-0.2622721493,0.0321485586,-0.0808480158,-0.0392323956,0.0516320169,0.1157314405,0.1197475493,-0.0257609952,-0.1189155206,-0.4395480156,0.268492043,-0.030725956,-0.0509762913,-0.0223328378,-0.0436791293,-0.3465599418,0.5494551659,0.2992221713,0.9601728916,-0.1665427983,-0.0754781514,0.1074845493,0.0600274466,0.4184290171,0.4510524273,-0.1104425415,-0.4783469141,-0.0814775601,-0.0965282694,-0.1091406643,0.1392843425,0.3080063164,-0.2980454564,0.4427348375,-0.1956773996,0.3273774981,0.1547433138,0.1308043748,0.421651721,-0.6057389975,-0.3888364732,0.0494441167,-0.2001888305,-0.0558025576,0.1221092045,-0.1107899249,0.0048690289,-0.0194987934,-0.0054047145,0.1903212667,0.0289518889,0.0167375281,0.0370974429,-0.2289071232,0.2931969464,0.4237119257,0.323954314,-0.1413274109,0.0419277176,-0.1754638702,0.0567614362,0.2114453316,-0.1153764948,-0.1104580313,0.375310123,0.2046166211,0.1186157614,0.2687427998,-0.200861901,-0.4143680036,0.1782288253,-0.0760684162,0.3111491203,-0.7221404314,-0.4602564871,-0.128583163,0.0333975516,-0.0694584176,-0.0068865153,0.0017650393,0.0250652768,0.1748291254,0.1566285342,-0.1613202691,0.1177105606,0.0944817215,0.0566830151,-0.0683545917,0.3383947909,0.212665692,-0.1351852566,-0.0864232704,0.2136965394,0.2606021464,-0.6703618765,0.1074096337,-0.1867013723,-0.2304186821,0.0568599924,0.6951977611,0.0893610418,0.2368177176,-0.1200567856,-0.3224506974,-0.5295479298,0.232519269,0.0771251172,0.1245122999,-0.0255271997,-0.0955997035,-0.1672355682,-0.2512124777,-0.1790679991,-0.2260364443,-0.2032000422,0.1978922486,-0.0268370993,0.1749275029,0.0735571235,0.1984947324,0.0286735687,0.1221107095,-0.0104309097,-0.0668749884,-0.0072935936,0.2638356388,0.2219250053,0.0452938154,-0.3669616878,-0.2743262649,-0.1671470702,-0.302028656,-0.1595329046,0.0931516662,-0.1908308864,0.0694066063,0.122177124,0.1618801802,-0.2803739309,0.2299583405,-0.0745547265,0.2896379828,-0.1253440827,0.1830162108,0.0148315756,-0.0342968032,-0.1972382218,0.2030228525,0.4118275642,-0.076239042,0.2323230356,-0.2946207225,0.1638775766,0.1537060589,0.2169139832,0.3342391849,-0.0989016891,-0.0225505512,0.3155774474,0.0290143359,-0.0749651492,-0.2232451141,0.0144935297,-0.0943720043,-0.1142105386,0.4057324827,0.2134423554,-0.1341504157,0.1725117117,-0.1364741772,0.4199763536,-0.2052786201,-0.2592469454,0.6168783307,0.1744426936,0.173277989,0.3604376018,0.0263558775,0.1716962904,0.3486081362,-0.1904983073,0.0358635224,0.2933235168,0.5195006132,0.0373278186,-0.2363219112,0.2139700055,0.3423799872,0.016566908,0.1976490468,0.1567900926,0.2483701855,-0.1659100205,-0.3684499264,-0.0531707704,0.268332541,-0.212314561,-0.2660959065,-0.4559815824,-0.1595574319,-0.2920824885,-0.0755719095,-0.2318983227,-0.0610434078,0.4262933135,0.157483533,0.3302525282,-0.3787958026,-0.4230478108,0.1484677941,0.0860213339,-0.2165574878,0.4323184788,-0.1213348508,0.017309539,0.3313530087,0.0118939374,0.348197341,0.5415486693,0.1784790158,-0.1693001837,0.055514276,-0.1464320272,-0.010035431,0.2831676602,-0.1197287217,0.2042817026,0.2606139779,0.0198879447,-0.0287935212,0.4744407535,-0.2728385031,0.0609481484,0.0328227542,0.3194430768,0.2836459875,-0.2071183324,-0.1013968214,-0.1426229775,-0.1514361799,-0.0781531185,0.3565252125,-0.0592491478,0.2234139144,0.1519929618,0.0196995977,-0.0528681912,0.2357492298,0.1952368021,0.0818016604,-0.2347944826,0.0796157643,-0.6122213006,0.1297814101,0.3405697644,-0.1923160255,0.477017045,0.2007792592,0.2242653966,0.3185648918,-0.3486528397,-0.1267262846,0.2773860395,0.3379606307,0.0695361719,-0.0097363638,0.2614303529,0.2672411203,-0.2102860063,-0.5415986776,0.0850290209,-0.0747512206,-0.038054388,-0.1727861613,-0.1443443,-0.0024616374,-0.0389548242,0.2549988627,0.0003754296,0.5277963877,-0.0800938085,0.0247710012,0.1740865707,-0.4902578592,-0.1550229341,0.3757013679,-0.1601576656,0.4425735474,-0.1881638914,0.5270153284,-0.2822114527,-0.1201009974,0.1412360668,-0.0103471503,-0.4238345921,0.3756207824,-0.5192402601,0.0653479621,0.1564181,0.1585126072,0.1654796898,0.1913678795,-0.1295810789,-0.3327223659,0.4483350813,-0.3660878539,-0.3723253608,0.1123897135,-0.0695723221,0.1881098002,-0.1034300774,-0.5125928521,-0.1575303227,0.2134394348,-0.1684264392,-0.3530288041,0.0400134251,-0.1029169858,0.1037679166,-0.2346630991,0.1878359765,0.1710439473,-0.0456501469,-0.1793133914,-0.4535680711]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2288","title":"Load_dataset for local CSV files","comments":"Yes, just map the labels to their ids.","body":"The method load_dataset fails to correctly load a dataset from csv. \r\n\r\nMoreover, I am working on a token-classification task ( POS tagging) , where each row in my CSV contains two columns each of them having a list of strings.\r\nrow example:\r\n```tokens | labels\r\n['I' , 'am', 'John'] | ['PRON', 'AUX', 'PROPN' ] \r\n```\r\nThe method, loads each list as a string: (i.g \"['I' , 'am', 'John']\").\r\nTo solve this issue, I copied the Datasets.Features, created Sequence types ( instead of Value) and tried to cast the features type\r\n```\r\nnew_features['tokens'] = Sequence(feature=Value(dtype='string', id=None))\r\nnew_features['labels'] = Sequence(feature=ClassLabel(num_classes=len(tag2idx), names=list(unique_tags)))\r\ndataset = dataset.cast(new_features)\r\n```\r\nbut I got the following error \r\n```\r\nArrowNotImplementedError: Unsupported cast from string to list using function cast_list\r\n```\r\nMoreover, I tried to set feature parameter in load_dataset method, to my new_features, but this fails as well.\r\nHow can this be solved ?","comment_length":8,"text":"Load_dataset for local CSV files\nThe method load_dataset fails to correctly load a dataset from csv. \r\n\r\nMoreover, I am working on a token-classification task ( POS tagging) , where each row in my CSV contains two columns each of them having a list of strings.\r\nrow example:\r\n```tokens | labels\r\n['I' , 'am', 'John'] | ['PRON', 'AUX', 'PROPN' ] \r\n```\r\nThe method, loads each list as a string: (i.g \"['I' , 'am', 'John']\").\r\nTo solve this issue, I copied the Datasets.Features, created Sequence types ( instead of Value) and tried to cast the features type\r\n```\r\nnew_features['tokens'] = Sequence(feature=Value(dtype='string', id=None))\r\nnew_features['labels'] = Sequence(feature=ClassLabel(num_classes=len(tag2idx), names=list(unique_tags)))\r\ndataset = dataset.cast(new_features)\r\n```\r\nbut I got the following error \r\n```\r\nArrowNotImplementedError: Unsupported cast from string to list using function cast_list\r\n```\r\nMoreover, I tried to set feature parameter in load_dataset method, to my new_features, but this fails as well.\r\nHow can this be solved ?\nYes, just map the labels to their ids.","embeddings":[-0.1388803273,-0.0401049778,0.0073327473,0.0986296311,0.4620278478,0.0569799803,0.5705062151,0.2671584189,0.3436504304,-0.1559716314,0.0872420818,0.4712432921,-0.0202665385,0.1968052238,-0.0105222641,-0.2581174076,0.1815401465,0.1017879844,-0.0888823867,-0.0140854809,-0.265306443,0.0297483131,-0.2757693827,0.0774082467,0.0188080203,0.0558445565,0.0704049245,0.0675991476,0.1964584589,-0.4892320037,0.1505249292,-0.0290140659,0.419410646,0.1575927734,-0.0001215104,-0.0449704044,0.2264862955,-0.0928434655,-0.0991415158,-0.3086333871,-0.3979975283,-0.3734481335,0.5364148021,-0.2948939502,0.0769859925,-0.0498431586,-0.1112524197,-0.3619641364,0.2983614206,0.466909647,0.1113494858,0.1790606529,-0.0950245634,0.0220356546,0.2970421314,0.1666090786,-0.3054665625,0.0439160615,0.3263615072,0.28355667,0.0069031273,-0.1696840972,-0.3025632203,-0.1681576073,0.5296937227,-0.0029835391,-0.1353193372,-0.2289123088,-0.0080692833,0.0546618626,0.8512439728,-0.0655780137,-0.3174843788,-0.0556070693,0.1188972071,-0.3209231794,0.0828735903,-0.0465023592,-0.019771114,0.1295439452,-0.3139625788,-0.0257110652,-0.063229695,0.1979211122,0.1089067012,-0.0047050873,-0.1449606121,0.2381021529,-0.0455971807,-0.246349737,0.1898722947,-0.1060462222,0.1307644248,0.3022912741,-0.5439129472,0.0710766017,0.0494397879,-0.0261806287,-0.0199912935,0.2273358107,0.1598479897,0.1270178705,-0.0784855336,0.1969529986,0.3010213375,0.3397428989,0.408610791,0.1141363829,-0.0410167128,-0.1475486159,-0.1959067881,0.0590959601,-0.3804418445,-0.2067958564,0.415792495,0.1841147989,0.0212657005,-0.3639909923,-0.2976562679,0.1796044856,-0.2297486663,-0.0191783905,0.1556813568,0.3728463054,-0.1203068048,0.4791846573,0.0721626282,0.3590092063,0.1449205875,-0.1524095237,-0.1414850354,-0.1371960193,-0.1543769389,0.008441586,0.1182378009,-0.289044708,0.1557031423,-0.0547972508,-0.0088920249,-0.1468625516,-0.1606102288,-0.046229504,-0.0362359285,0.2038841099,-0.0020975906,0.1045088843,0.3188068569,-0.0441455506,-0.2403986901,0.3084672391,-0.5463098288,-0.2748061419,0.0330101363,0.0050530387,-0.0781800225,-0.0026862421,-0.4287045896,0.1554680914,0.186692819,0.0771248192,-0.1957833618,-0.3678832054,-0.1541469097,-0.3626759648,0.1283688098,0.3547216058,-0.5119929314,0.0066447565,0.14186652,-0.0210155938,0.1716755927,0.5199783444,-0.1762490124,0.3015101552,-0.1285159439,0.4106365442,0.5060756207,-0.0338434391,-0.0298448578,0.5519416332,0.1131788939,-0.1030353457,0.1634783894,0.0922084898,0.3472668231,-0.2060607076,0.0305681527,0.3761134446,-0.1770423353,-0.0034449017,-0.1056538001,-0.0893396437,0.1895926297,0.143212229,-0.1129426733,0.3034436107,0.1644529551,0.0937674716,-0.0160749611,-0.3421500921,0.1629131585,0.1678912938,-0.0714517981,0.1407141089,-0.054875534,0.0279276576,-0.6750742197,-0.0627555028,0.1736579984,-0.0551608391,-0.2942739129,-0.1143781096,-0.3666611016,0.2243435383,-0.1347509325,0.5390368104,-0.1195896864,0.1483447105,-0.0677247792,-0.0207550786,-0.1186659113,0.0157153476,-0.2867617607,-0.0057711513,-0.1817723662,-0.0089156348,0.1940152794,-0.0778708979,-0.1352500916,0.0802610368,0.3492612541,-0.1332118213,-0.2605765462,0.3393309116,0.116673328,-0.2432789057,-0.2810564935,0.1435894966,0.0990139842,-0.4612132609,-0.1063275039,0.1947364509,0.2915081084,-0.152764231,-0.2321550846,0.4877677262,-0.2353849113,0.3657234311,-0.1577934772,-0.0660731792,0.4375196397,0.1111076847,-0.2379331589,-0.1260700524,0.0408088341,0.1650858223,0.3136439919,0.1375090629,-0.5187990665,-0.0716838315,0.2044140249,-0.1342459768,0.2067052573,0.2200645804,-0.0218848828,0.0319216959,0.0900188461,-0.3492116034,0.3582169116,-0.0023714006,-0.0156186586,-0.1039296463,-0.1042924151,-0.0540095344,0.144881323,-0.1594783664,0.0568043329,0.2334430218,-0.0215914883,-0.1038801,-0.2456496656,-0.2005769014,0.1676684171,0.1005436555,-0.6145338416,0.0915315077,-0.3506908417,0.1618985683,-0.3309769332,-0.1316244006,0.1012099832,-0.1461797953,-0.1814760715,0.2523818016,-0.1322276592,0.2271693796,-0.4267176986,0.1042948291,0.224630788,-0.6247831583,0.1197305396,0.0085232593,-0.161377266,-0.0740837231,0.4018029571,0.2442578226,0.09583579,-0.1229092851,0.0631320924,-0.1799059957,-0.2662219107,0.0742151812,-0.0774365366,0.1374429464,0.0291435979,0.0887903273,-0.2239456475,-0.4154924452,0.2995001674,-0.1145169958,-0.1069549918,0.293356508,-0.120276317,-0.3609604537,-0.1789625734,-0.5238916874,-0.1772275567,-0.3606033623,0.1726552099,0.0582549572,0.0145348404,-0.014653367,0.2213191092,0.0485523157,0.241593495,0.059528891,-0.0612722784,-0.0603131615,0.52131778,-0.1779292524,-0.349855572,0.0433087833,-0.1740531772,-0.0765377358,0.353487432,-0.105741553,0.1217470169,0.0647713915,0.2287265807,0.0773731992,0.0270154383,0.2280739099,0.0154172713,0.1046875268,-0.1033003777,-0.2784180641,0.1656719595,0.1823016852,0.165149793,0.5380093455,0.6245655417,-0.3417063653,0.4897071719,-0.0236125682,-0.1534687877,0.5532031655,-0.1173391789,0.1568087339,-0.2416640818,-0.1868686676,-0.096366927,-0.1747822165,-0.2962107062,0.2897308469,-0.1399980485,-0.3728736937,-0.1297984868,0.2868448198,-0.2724949718,-0.1262346953,0.4241885841,-0.1894708127,0.2839106917,-0.1352625936,0.1317236573,-0.1291665584,-0.0776041076,0.0184686612,0.0850209445,0.1163556874,-0.1206441671,-0.5616517663,-0.0725406557,-0.3594123125,0.1460547149,0.3399963975,0.5603936315,-0.1166767702,-0.1003928185,0.0346360281,0.1336451024,0.4168894589,-0.269351095,-0.0079664076,0.2469153255,0.0580906905,-0.0099919122,0.0341808572,-0.0994799212,-0.1274863929,-0.0791370422,0.2226148099,-0.2520556748,-0.0843331069,-0.170707792,0.355289191,-0.2046826333,-0.1399080753,-0.0804230794,-0.2858299315,-0.2760542333,0.0216608234,0.219936654,0.3978689313,0.0723162442,0.0022357036,-0.3762727976,-0.0708855465,0.227933079,0.1078421995,0.3083104789,0.1432077438,0.149901405,-0.2820409834,0.1116770431,0.0330076069,0.6782884598,-0.0456600226,-0.7954154015,0.1204169691,-0.2073794007,0.6535369158,0.0497287177,-0.2294091582,0.0397833399,-0.0512345694,-0.0498508699,-0.0398616344,0.0868426785,0.1615515053,-0.0539683215,-0.1774223894,-0.5492207408,0.2650476098,0.0305996779,-0.0830051973,-0.1254773289,-0.1391492635,-0.3866341412,0.5347246528,0.2208793014,0.8540100455,-0.2388675511,-0.0463832542,0.0921668559,0.1404347718,0.5190047622,0.309820652,-0.11055509,-0.3964279592,-0.1515434682,-0.1364042014,-0.0539759621,0.2086142898,0.3160041571,-0.1601401269,0.4872645736,-0.2127080858,0.439607352,0.0418783575,-0.0130293164,0.4669885039,-0.6458953619,-0.4755726457,0.054120224,-0.1049827039,-0.101189509,0.0802981332,-0.0951977,0.0580863506,-0.0186385121,-0.089026846,0.2352293879,0.0348372422,0.0374829695,0.0534700118,-0.1980630755,0.2764098346,0.4387976527,0.3136426508,-0.1643570215,0.0207351353,-0.0686580837,0.0736736804,0.1599728763,-0.1780272722,-0.1565582901,0.3948527575,0.2541754246,0.0965078548,0.2322876006,-0.1918466091,-0.4143607318,0.1782682836,-0.0613216013,0.306476742,-0.6674274802,-0.5202156305,-0.1517400593,0.1747418195,-0.1233638972,0.0201624129,0.0598368905,0.0062150746,0.1535511911,0.0157634225,-0.0958042443,0.0663392469,0.1762379259,0.0484669395,-0.0553527363,0.3412354589,0.2650476694,-0.0656440035,-0.0719221607,0.2334278971,0.372422874,-0.6674916148,0.0154414047,-0.0666331947,-0.244904995,0.0471147448,0.6547221541,0.0338335931,0.1416633576,-0.0125148771,-0.3671209514,-0.509290278,0.2249250561,0.1408665031,0.0833328515,0.0407133587,-0.0220042355,-0.164154008,-0.1311155111,-0.1656915545,-0.3051869273,-0.2293694019,0.203243807,0.0534344912,0.2660323381,0.0442485102,0.1434679627,0.0610247068,0.0781268999,-0.0119190672,-0.1008349136,-0.0502775349,0.2555093765,0.1959720254,0.0323247835,-0.3029441237,-0.3193968832,-0.2138054371,-0.3958051801,-0.0758564323,0.1394772679,-0.1228130013,0.0195258297,0.1431389153,0.1249037534,-0.3485674262,0.2649866641,-0.091639258,0.2596299648,-0.0716663152,0.1230633631,-0.047132235,-0.0200489368,-0.2020561844,0.1874117404,0.3760364652,-0.0354164839,0.2561130822,-0.2895854115,0.155799374,0.2813562751,0.2985320985,0.3319421709,-0.0814257264,0.0695355982,0.3238965869,0.0500175729,-0.1580191702,-0.2804535925,0.1302655339,-0.0802938417,-0.1017569304,0.3384002149,0.2153161168,-0.1213904172,0.1428238153,-0.1495504081,0.418907702,-0.2220739722,-0.0847322494,0.6308858991,0.2657082081,0.165602386,0.2684025466,0.0399863422,0.2300924063,0.3502397537,-0.1036003381,0.0077904104,0.2748799622,0.5153800249,0.0563174672,-0.3315949142,0.1519663036,0.2500823438,-0.0324847624,0.1646799296,0.1034299731,0.1814925373,-0.1289484948,-0.351169765,-0.1183633432,0.3060260713,-0.1988455206,-0.3068078458,-0.3921069503,-0.1316393167,-0.2082474232,-0.0249367114,-0.2161444128,-0.1015988961,0.3670547605,0.0763504654,0.2720611989,-0.3966534734,-0.3619816899,0.107235916,0.0445240401,-0.2152524143,0.3686611652,-0.1606802493,-0.0280077849,0.3348024786,-0.051344201,0.3396236002,0.5325984359,0.2173131108,-0.1536403149,0.0020342311,-0.1255990267,-0.0740706623,0.2313316315,-0.0952768177,0.2553022504,0.2197247893,0.0199538972,-0.034362305,0.5285341144,-0.2380533963,0.0307692569,-0.0093052033,0.3264406025,0.2005064487,-0.1400732398,-0.112926729,-0.073358722,-0.1513550729,-0.0406086557,0.4469990134,0.0220801868,0.2557076216,0.1727833301,0.0141773876,-0.0910224169,0.2451270074,0.129195258,0.0642499179,-0.2624595463,0.1390377283,-0.7283061147,0.1021011397,0.2481797189,-0.1394657493,0.4437143803,0.1925298423,0.2926445007,0.241633445,-0.4724876881,-0.170827955,0.3081249893,0.2676199377,0.0409435742,-0.1145605594,0.0919301063,0.2381150126,-0.2232520729,-0.532024622,0.0579008386,-0.0815610439,-0.0500896685,-0.1488395333,-0.2053172588,-0.00415202,-0.0230095573,0.3032039404,-0.1019407958,0.6126285195,-0.1006588861,0.1281505674,0.0621540844,-0.454149276,-0.1813659817,0.3549093604,-0.1751691103,0.4126894474,-0.2983413339,0.5648124218,-0.2985903621,-0.0969273373,0.1082600728,-0.0672043115,-0.3865168095,0.3776716292,-0.5142288208,0.0608757436,0.1740602404,0.0916009173,0.0708455369,0.2044568956,0.0512611568,-0.3087092638,0.4359409213,-0.4125050604,-0.3238154948,-0.0239668582,-0.1182068139,0.1072359905,-0.0477088019,-0.4732357562,-0.0330391899,0.2402113229,-0.1443374902,-0.3313264549,0.0897337273,-0.0607230067,0.0791433975,-0.2525192499,0.1944455802,0.1315929741,-0.0575893782,-0.1591366976,-0.4853348732]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2285","title":"Help understanding how to build a dataset for language modeling as with the old TextDataset","comments":"\r\nI received an answer for this question on the HuggingFace Datasets forum by @lhoestq\r\n\r\nHi !\r\n\r\nIf you want to tokenize line by line, you can use this:\r\n\r\n```\r\nmax_seq_length = 512\r\nnum_proc = 4\r\n\r\ndef tokenize_function(examples):\r\n# Remove empty lines\r\nexamples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\nreturn tokenizer(\r\n examples[\"text\"],\r\n truncation=True,\r\n max_length=max_seq_length,\r\n)\r\n\r\ntokenized_dataset = dataset.map(\r\ntokenize_function,\r\nbatched=True,\r\nnum_proc=num_proc,\r\nremove_columns=[\"text\"],\r\n)\r\n```\r\n\r\nThough the TextDataset was doing a different processing by concatenating all the texts and building blocks of size 512. If you need this behavior, then you must apply an additional map function after the tokenization:\r\n\r\n```\r\n# Main data processing function that will concatenate all texts from\r\n# our dataset and generate chunks of max_seq_length.\r\ndef group_texts(examples):\r\n# Concatenate all texts.\r\nconcatenated_examples = {k: sum(examples[k], []) for k in examples.keys()}\r\ntotal_length = len(concatenated_examples[list(examples.keys())[0]])\r\n# We drop the small remainder, we could add padding if the model supported it instead of this drop,\r\n# you can customize this part to your needs.\r\ntotal_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n# Split by chunks of max_len.\r\nresult = {\r\n k: [t[i : i + max_seq_length] for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n}\r\nreturn result\r\n\r\n# Note that with `batched=True`, this map processes 1,000 texts together,\r\n# so group_texts throws away a remainder for each of those groups of 1,000 texts.\r\n# You can adjust that batch_size here but a higher value might be slower to preprocess.\r\n\r\ntokenized_dataset = tokenized_dataset.map(\r\ngroup_texts,\r\nbatched=True,\r\nnum_proc=num_proc,\r\n)\r\n```\r\n\r\nThis code comes from the processing of the run_mlm.py example script of transformers\r\n\r\n","body":"Hello,\r\n\r\nI am trying to load a custom dataset that I will then use for language modeling. The dataset consists of a text file that has a whole document in each line, meaning that each line overpasses the normal 512 tokens limit of most tokenizers.\r\n\r\nI would like to understand what is the process to build a text dataset that tokenizes each line, having previously split the documents in the dataset into lines of a \"tokenizable\" size, as the old TextDataset class would do, where you only had to do the following, and a tokenized dataset without text loss would be available to pass to a DataCollator:\r\n\r\n```\r\nmodel_checkpoint = 'distilbert-base-uncased'\r\n\r\nfrom transformers import AutoTokenizer\r\ntokenizer = AutoTokenizer.from_pretrained(model_checkpoint)\r\n\r\nfrom transformers import TextDataset\r\n\r\ndataset = TextDataset(\r\n tokenizer=tokenizer,\r\n file_path=\"path\/to\/text_file.txt\",\r\n block_size=512,\r\n)\r\n```\r\n\r\nFor now, what I have is the following, which, of course, throws an error because each line is longer than the maximum block size in the tokenizer:\r\n\r\n```\r\nimport datasets\r\ndataset = datasets.load_dataset('path\/to\/text_file.txt')\r\n\r\nmodel_checkpoint = 'distilbert-base-uncased'\r\ntokenizer = AutoTokenizer.from_pretrained(model_checkpoint)\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[\"text\"])\r\n\r\ntokenized_datasets = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=[\"text\"])\r\n\r\ntokenized_datasets\r\n```\r\n\r\nSo what would be the \"standard\" way of creating a dataset in the way it was done before?\r\n\r\nThank you very much for the help :))","comment_length":270,"text":"Help understanding how to build a dataset for language modeling as with the old TextDataset\nHello,\r\n\r\nI am trying to load a custom dataset that I will then use for language modeling. The dataset consists of a text file that has a whole document in each line, meaning that each line overpasses the normal 512 tokens limit of most tokenizers.\r\n\r\nI would like to understand what is the process to build a text dataset that tokenizes each line, having previously split the documents in the dataset into lines of a \"tokenizable\" size, as the old TextDataset class would do, where you only had to do the following, and a tokenized dataset without text loss would be available to pass to a DataCollator:\r\n\r\n```\r\nmodel_checkpoint = 'distilbert-base-uncased'\r\n\r\nfrom transformers import AutoTokenizer\r\ntokenizer = AutoTokenizer.from_pretrained(model_checkpoint)\r\n\r\nfrom transformers import TextDataset\r\n\r\ndataset = TextDataset(\r\n tokenizer=tokenizer,\r\n file_path=\"path\/to\/text_file.txt\",\r\n block_size=512,\r\n)\r\n```\r\n\r\nFor now, what I have is the following, which, of course, throws an error because each line is longer than the maximum block size in the tokenizer:\r\n\r\n```\r\nimport datasets\r\ndataset = datasets.load_dataset('path\/to\/text_file.txt')\r\n\r\nmodel_checkpoint = 'distilbert-base-uncased'\r\ntokenizer = AutoTokenizer.from_pretrained(model_checkpoint)\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[\"text\"])\r\n\r\ntokenized_datasets = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=[\"text\"])\r\n\r\ntokenized_datasets\r\n```\r\n\r\nSo what would be the \"standard\" way of creating a dataset in the way it was done before?\r\n\r\nThank you very much for the help :))\n\r\nI received an answer for this question on the HuggingFace Datasets forum by @lhoestq\r\n\r\nHi !\r\n\r\nIf you want to tokenize line by line, you can use this:\r\n\r\n```\r\nmax_seq_length = 512\r\nnum_proc = 4\r\n\r\ndef tokenize_function(examples):\r\n# Remove empty lines\r\nexamples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\nreturn tokenizer(\r\n examples[\"text\"],\r\n truncation=True,\r\n max_length=max_seq_length,\r\n)\r\n\r\ntokenized_dataset = dataset.map(\r\ntokenize_function,\r\nbatched=True,\r\nnum_proc=num_proc,\r\nremove_columns=[\"text\"],\r\n)\r\n```\r\n\r\nThough the TextDataset was doing a different processing by concatenating all the texts and building blocks of size 512. If you need this behavior, then you must apply an additional map function after the tokenization:\r\n\r\n```\r\n# Main data processing function that will concatenate all texts from\r\n# our dataset and generate chunks of max_seq_length.\r\ndef group_texts(examples):\r\n# Concatenate all texts.\r\nconcatenated_examples = {k: sum(examples[k], []) for k in examples.keys()}\r\ntotal_length = len(concatenated_examples[list(examples.keys())[0]])\r\n# We drop the small remainder, we could add padding if the model supported it instead of this drop,\r\n# you can customize this part to your needs.\r\ntotal_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n# Split by chunks of max_len.\r\nresult = {\r\n k: [t[i : i + max_seq_length] for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n}\r\nreturn result\r\n\r\n# Note that with `batched=True`, this map processes 1,000 texts together,\r\n# so group_texts throws away a remainder for each of those groups of 1,000 texts.\r\n# You can adjust that batch_size here but a higher value might be slower to preprocess.\r\n\r\ntokenized_dataset = tokenized_dataset.map(\r\ngroup_texts,\r\nbatched=True,\r\nnum_proc=num_proc,\r\n)\r\n```\r\n\r\nThis code comes from the processing of the run_mlm.py example script of transformers\r\n\r\n","embeddings":[-0.2392032892,-0.0001858806,0.0175925288,0.1512313336,0.1369385719,-0.1672839969,0.5160788298,0.0982798114,-0.0554774702,-0.1596171558,0.1894440204,-0.1941065639,-0.1646524668,0.1340248436,0.1435168535,-0.0841108859,0.1339856535,0.0562884994,0.1950121075,0.023583632,0.0756416768,0.0561196543,-0.0015379904,-0.0821405351,-0.5577692389,-0.1091072038,-0.1831939965,-0.1227100119,-0.3284448087,-0.4212741554,0.0153291784,0.3389095366,0.4065946937,0.4958460927,-0.0001238081,-0.2122489214,-0.177432999,-0.2544007003,-0.0525416359,-0.237539798,0.3744877279,-0.4061252177,0.107468687,-0.1326390505,0.0795988813,-0.2096719146,-0.0318121016,-0.2318027467,0.5191584826,0.4787013531,0.0087787909,0.0259202924,0.0743123814,0.1130469888,-0.3428885639,0.1746680588,-0.1214065105,0.1298532933,0.3205685914,-0.0565026291,-0.0402820073,-0.0930833444,0.0166849941,-0.3108217418,0.3316012919,0.1152971238,-0.1415722817,-0.4516028762,0.0755913779,0.2538356185,0.6707497239,-0.3457719982,-0.3893296719,-0.4439043105,-0.1682328582,0.0374632366,-0.0296423137,0.0730164275,-0.2541878819,0.1022670195,-0.1240605265,-0.294654727,-0.2216565907,0.0544245765,-0.3207271695,0.3257496655,-0.0336121321,0.0357048847,-0.1194261163,-0.2430410087,-0.059906356,-0.2204925269,0.2210455984,0.4883925915,-0.0837700814,-0.3066206872,-0.2297293544,-0.3395672441,0.5583885908,0.0868365467,-0.0990155861,0.0349288397,-0.2218428552,-0.0805967897,0.0350312404,0.3476090431,0.5046224594,-0.1744385362,0.0108416593,-0.169294998,-0.1438457072,-0.2762911618,-0.6274005175,0.117275551,-0.059924271,-0.0570774451,0.0392435677,-0.0520313345,0.2307766676,-0.0069928747,-0.2350392342,-0.0044798492,0.0452586971,0.0356492549,-0.0675020069,-0.0861181617,-0.2376567274,0.4212774634,-0.2759157717,-0.2418381125,0.0703430623,-0.1863947809,-0.1865321845,0.1939846724,0.1796313524,0.0179879908,0.3778651059,0.1087445989,-0.1177096665,-0.2375998646,0.1789949238,-0.2524816096,-0.0524344705,-0.0748906136,0.1091084182,0.2168307751,0.1097023636,-0.417730391,-0.2348013818,0.033199966,-0.0249291044,-0.3025587797,-0.0333838165,0.028493451,-0.192048341,-0.1472052634,-0.2405373156,0.6766546369,0.4798668027,0.0126460642,-0.0017274155,-0.0322487764,-0.1381386518,-0.0999684483,0.2362454236,0.4747399986,-0.7666448951,0.0063369167,0.0026421065,-0.0365215726,0.0965150073,0.3493333459,-0.1648329794,0.6068561077,-0.0613503121,0.5180051923,0.3503054678,0.130518645,-0.2049496472,0.1372380257,0.0292102527,0.1867638379,-0.107311368,-0.1562440097,0.5612369776,-0.2129071951,-0.0361982398,0.1182904318,-0.1476489156,-0.0669578165,0.0326253697,0.0857679993,0.2753346562,-0.1395994723,0.0697687194,-0.0398723632,-0.0444107465,0.3139216304,0.1966240555,-0.1594019681,0.2759898007,0.0362149328,0.2324165106,0.4073259532,-0.0153670628,-0.269325763,-0.0759757906,0.0567638688,0.2049039155,0.2655689716,-0.0570136979,-0.1916297227,-0.247473672,-0.1511307061,0.0207675304,-0.1977970451,-0.0181167405,0.0179733969,-0.1353481561,0.0234728567,-0.3133320808,0.1876451969,-0.2843897641,0.3427190185,-0.2325659841,-0.0005955913,0.2443991154,-0.1013842002,-0.1625291705,0.3830213845,-0.0273513813,-0.1313119382,0.0279479399,0.2366939336,0.3679019213,-0.1408003867,-0.2275568247,0.4600908458,0.3784703314,0.0309689473,0.457837671,-0.2016495168,0.019380128,-0.1889169216,0.0346812047,0.3599475324,0.209063679,0.2274591774,-0.0112377815,-0.0714983791,0.3128148019,0.0040353718,-0.0427536964,-0.0139180552,-0.015475913,0.1952089816,0.3493808806,0.0107706487,0.0465696417,0.3449261189,0.2432102263,-0.1532547027,0.1287156045,0.2910188138,-0.2863344848,-0.20529975,-0.1908234656,0.0046362383,0.0556731634,0.0476993956,0.3897702694,0.0664353073,-0.1433897018,-0.1299940348,0.2656420469,0.0605866201,-0.0630757362,0.310069561,-0.1549257338,0.1154171452,-0.2331500798,-0.4740202129,-0.1258013844,0.2036343068,-0.1853826642,0.1338459551,0.04872486,-0.0661096945,-0.3965555429,-0.1830105335,-0.371871978,0.0566233136,-0.0349516608,-0.1158750281,0.149379164,0.0335758068,0.2499835938,0.506326735,0.009573034,0.0583958924,0.2063226849,-0.2811831832,-0.25026232,-0.0346411243,0.0933179334,0.2416541576,0.1541722715,-0.087047115,0.0865254253,-0.0161855165,-0.5051578879,0.1515581161,-0.1692162603,0.4367188811,0.055425223,0.1744432747,0.0144784721,-0.0833800212,-0.2698356509,0.0895950049,-0.2077247351,-0.229113996,-0.11113929,-0.0045480668,0.0718693659,-0.1308140904,-0.3728190958,-0.0504081286,0.2722548544,0.1782402098,0.1835107654,-0.0798517168,-0.0299037285,0.2354943305,0.1032655314,0.0373405814,0.0435234234,-0.0512595996,0.2787410617,-0.2501344979,-0.2366040498,-0.0474127717,-0.1261128634,0.2636546195,0.1433220655,-0.2231925428,0.2140240073,-0.1526659727,0.0093198819,0.0154485377,-0.02381845,0.405169636,-0.0522558428,0.2788363993,-0.1214466766,0.018085232,0.0984245241,-0.2761836052,0.1969474107,-0.0060328697,0.6503129601,0.0369096696,0.6386395693,0.3684069216,-0.1325151026,0.0752433315,-0.0477322675,0.0422274098,-0.3112989068,-0.22771281,0.0006069641,0.0090353573,-0.0301074646,0.395670861,0.242307812,0.1143696606,0.2021980137,-0.1181013361,-0.139484942,-0.2757112384,0.2836950719,-0.46955809,0.4671967626,-0.0721840262,0.0116693815,-0.331728518,-0.1169725806,-0.2223072201,0.3786241412,-0.0282127056,0.0598835312,-0.5507213473,0.3868112564,-0.6631576419,0.0291504581,0.2206850499,0.3547828197,0.1103871018,0.146371454,-0.0709168762,-0.0203723554,0.4026373029,-0.0227485932,-0.289421916,-0.1082241312,-0.1320896596,-0.2538170516,0.2933549583,-0.0395808145,0.1975641847,0.531175077,0.6587136388,-0.3302847147,-0.2319725305,-0.378287971,0.5049983263,-0.1509028375,-0.0131164957,0.1996587813,-0.0180583633,-0.2463341355,0.0441162623,0.1279946417,0.1873219609,0.0992732048,0.0426309332,-0.1333944649,-0.2861351967,0.256357193,0.0594392344,0.1277616918,-0.2765386999,0.0873952433,0.2656547129,-0.1350411922,-0.0495324619,0.1073107272,0.175916627,-0.4010824561,-0.0129603427,-0.1064639986,0.498267442,0.007294646,0.1161932051,0.1759970039,-0.0717409551,-0.1283391565,-0.2516773939,0.499040693,-0.0004323443,0.1072119102,-0.4950771332,-0.5806419849,0.1287848949,0.0022292859,-0.095541425,0.3580082059,-0.0112047428,-0.5861917138,0.5469619632,0.233993724,1.0034718513,0.0364242531,0.1965809017,-0.0692434162,0.2333040237,0.3526033163,-0.5463705063,-0.2250963002,-0.4049321413,0.2398339212,-0.1035824865,0.0531465448,0.2031105161,0.4196674228,-0.0978061557,0.0434216522,0.1982338428,-0.0273168627,0.1032379866,0.4068479836,-0.3265764415,-0.2755764723,0.0517287888,0.0033750557,-0.0178098381,-0.20658499,-0.1173749566,0.184115693,0.0330491252,-0.0679970905,-0.4975017607,-0.3805792928,-0.3608533144,-0.0133375917,0.1129649282,-0.207826212,0.4366981089,0.2826575935,0.6320579648,0.3912512958,-0.2255332321,0.0295785163,-0.2768983245,-0.1763713658,0.0498523898,-0.2072148919,0.3300938308,0.0513834357,-0.1469299644,-0.3165063262,0.0341586694,0.0248636659,-0.3118890524,-0.3918748796,0.2977168262,-0.5495624542,0.0693327859,0.1410159767,0.1742270589,-0.162275359,-0.0320815705,0.2196674347,-0.0508986674,-0.0259331148,-0.1620739102,-0.0592637658,-0.1033973843,0.0747573376,-0.0118697779,-0.2910359204,0.5136494637,0.2392341048,-0.1066969857,-0.1358098537,0.3822306097,0.2690787315,-0.4311947525,0.0318620317,-0.2794513702,-0.1974163204,0.0385380872,0.3607567847,0.2582133412,-0.1642000079,-0.0545679629,-0.4993221462,-0.4086224437,0.0049348236,-0.0258853119,0.276212275,-0.1327503771,0.3132956326,-0.1386003196,0.1803968847,-0.1185475066,0.2509557605,0.2383137494,0.2692071199,-0.0381936654,-0.1312283278,-0.1876063943,0.0220616329,-0.1256458461,0.3032785952,-0.1072252169,-0.0504502952,-0.481066376,0.2701501548,0.2685157955,-0.3452085555,0.1435485631,0.0769631341,-0.2080500871,-0.3401215374,0.1666087359,0.0652571321,0.0499008596,0.1358625591,0.4429149032,0.3076679707,-0.3088450134,-0.1887063533,0.1123204976,0.2102125734,0.426361084,0.1400438994,-0.1221103892,-0.0386325642,-0.2016714662,0.5881649256,0.5096353889,-0.083517313,0.0431773737,0.0823903754,0.140650183,0.1542516947,0.1059381589,0.4155412912,0.0251163319,-0.2322398424,0.6475023031,0.0915134102,0.030358579,0.1065910757,0.1562035084,-0.1744714379,0.0069195586,0.3582223356,0.0635342672,0.1894368231,-0.2472031116,-0.2593580186,0.530336082,0.0892974883,0.05789157,0.1401211321,-0.1871411651,0.0493195616,0.219301641,0.1716695279,0.3259340525,0.5564154983,-0.0698604062,0.3214527965,0.053129226,0.004704359,0.0471904427,-0.3095895648,-0.0800243914,0.0480233878,0.1656663567,-0.1105846837,-0.1947387904,0.1978369206,0.0363031961,-0.108343102,0.1280542314,0.2772510052,-0.1708551347,-0.4737078547,-0.3481684327,-0.2068168223,0.0164129939,-0.1115106046,0.0134600056,-0.3089475036,-0.3214224279,-0.1910711229,0.0930105522,-0.4402127862,0.2264525592,-0.1502142847,-0.0590888076,-0.0950689316,0.4051769674,0.0410051644,0.0708771497,-0.1157145202,0.2022528946,0.1273886412,0.1905830801,-0.1700425446,0.1962477714,-0.4156353772,0.0052444702,-0.0055853971,0.374794513,-0.0177562311,0.3423497379,0.1662877202,-0.0815358832,0.0093167415,-0.072930038,0.3396167159,-0.4478395581,-0.1336389482,0.4081301689,0.1175649539,-0.0746408477,-0.0261989478,0.0599855669,-0.2737570703,-0.0408930965,0.503592968,-0.0996121541,0.1589269489,-0.1605332345,-0.0049139112,-0.053294532,0.3754828572,0.4513396025,0.2204478234,-0.1742960364,-0.2868292332,-0.4054223597,-0.0375688449,0.0112631656,-0.079519026,-0.0870297626,0.0210078433,0.1426892877,0.1683030427,-0.3433713615,0.3283092082,-0.2516362369,-0.0704329312,-0.0427422076,-0.0357308909,-0.033896707,0.4775058925,-0.1342855245,-0.1031242236,-0.0185356308,-0.034754429,-0.1216445789,0.0555412397,-0.0064598508,0.2513418496,-0.2258577645,0.1185980141,0.2096252888,0.3537658453,0.0451680124,0.2406837493,-0.2795905769,0.1464356184,-0.2605933547,0.3197108507,0.1615745574,0.371563226,-0.3493005335,0.0580704398,-0.0444214493,0.1008119211,0.197295174,-0.2761369944,-0.3211239576,0.1628666073,-0.2799776495,-0.2471338212,0.1270712912,0.624832809,-0.1231764629,0.360670656,0.0793342963,0.0421331674,0.1611795276,-0.4122857451,-0.4679849446,-0.0765423626,-0.0395694673,0.649461329,-0.2072230726,-0.6433721781,-0.022690136,0.5512036681,0.0085754329,-0.2864916027,0.2637565434,0.2395220548,-0.28322348,-0.2148469687,0.5683005452,-0.1135209873,-0.0341540612,0.0079850303,-0.3772534728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2285","title":"Help understanding how to build a dataset for language modeling as with the old TextDataset","comments":"Resolved","body":"Hello,\r\n\r\nI am trying to load a custom dataset that I will then use for language modeling. The dataset consists of a text file that has a whole document in each line, meaning that each line overpasses the normal 512 tokens limit of most tokenizers.\r\n\r\nI would like to understand what is the process to build a text dataset that tokenizes each line, having previously split the documents in the dataset into lines of a \"tokenizable\" size, as the old TextDataset class would do, where you only had to do the following, and a tokenized dataset without text loss would be available to pass to a DataCollator:\r\n\r\n```\r\nmodel_checkpoint = 'distilbert-base-uncased'\r\n\r\nfrom transformers import AutoTokenizer\r\ntokenizer = AutoTokenizer.from_pretrained(model_checkpoint)\r\n\r\nfrom transformers import TextDataset\r\n\r\ndataset = TextDataset(\r\n tokenizer=tokenizer,\r\n file_path=\"path\/to\/text_file.txt\",\r\n block_size=512,\r\n)\r\n```\r\n\r\nFor now, what I have is the following, which, of course, throws an error because each line is longer than the maximum block size in the tokenizer:\r\n\r\n```\r\nimport datasets\r\ndataset = datasets.load_dataset('path\/to\/text_file.txt')\r\n\r\nmodel_checkpoint = 'distilbert-base-uncased'\r\ntokenizer = AutoTokenizer.from_pretrained(model_checkpoint)\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[\"text\"])\r\n\r\ntokenized_datasets = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=[\"text\"])\r\n\r\ntokenized_datasets\r\n```\r\n\r\nSo what would be the \"standard\" way of creating a dataset in the way it was done before?\r\n\r\nThank you very much for the help :))","comment_length":1,"text":"Help understanding how to build a dataset for language modeling as with the old TextDataset\nHello,\r\n\r\nI am trying to load a custom dataset that I will then use for language modeling. The dataset consists of a text file that has a whole document in each line, meaning that each line overpasses the normal 512 tokens limit of most tokenizers.\r\n\r\nI would like to understand what is the process to build a text dataset that tokenizes each line, having previously split the documents in the dataset into lines of a \"tokenizable\" size, as the old TextDataset class would do, where you only had to do the following, and a tokenized dataset without text loss would be available to pass to a DataCollator:\r\n\r\n```\r\nmodel_checkpoint = 'distilbert-base-uncased'\r\n\r\nfrom transformers import AutoTokenizer\r\ntokenizer = AutoTokenizer.from_pretrained(model_checkpoint)\r\n\r\nfrom transformers import TextDataset\r\n\r\ndataset = TextDataset(\r\n tokenizer=tokenizer,\r\n file_path=\"path\/to\/text_file.txt\",\r\n block_size=512,\r\n)\r\n```\r\n\r\nFor now, what I have is the following, which, of course, throws an error because each line is longer than the maximum block size in the tokenizer:\r\n\r\n```\r\nimport datasets\r\ndataset = datasets.load_dataset('path\/to\/text_file.txt')\r\n\r\nmodel_checkpoint = 'distilbert-base-uncased'\r\ntokenizer = AutoTokenizer.from_pretrained(model_checkpoint)\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[\"text\"])\r\n\r\ntokenized_datasets = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=[\"text\"])\r\n\r\ntokenized_datasets\r\n```\r\n\r\nSo what would be the \"standard\" way of creating a dataset in the way it was done before?\r\n\r\nThank you very much for the help :))\nResolved","embeddings":[-0.2752788067,0.0790361613,0.0096118012,0.1541674137,0.1268851757,-0.1365696043,0.4942955971,0.1448565125,-0.073795788,-0.187264666,0.2369309813,-0.1595558077,-0.1842525154,0.0847825333,0.12803711,-0.0277594738,0.1333022565,0.0742345601,0.2444006056,-0.0158724058,0.0821128935,0.0102104302,0.0182104483,-0.0977365673,-0.5597465038,-0.0959972069,-0.1499238461,-0.1552750766,-0.3058024943,-0.4282675087,0.001219679,0.3046123683,0.4055964351,0.5166521668,-0.000123806,-0.20167467,-0.1488274634,-0.2814053297,-0.0576476939,-0.243695721,0.2896805406,-0.3781110644,0.1133231595,-0.1281047761,0.0750633478,-0.2468433082,-0.0296042114,-0.2605940104,0.4783642292,0.5464983582,0.0059826281,-0.0601763837,0.0621914342,0.1094443575,-0.3377287686,0.1991954595,-0.1042982489,0.1258754432,0.3530709743,-0.0627444685,-0.0231459439,-0.1415844858,0.0234376658,-0.3365864456,0.3195755482,0.1166460365,-0.0918433741,-0.471491158,0.085002929,0.2460734099,0.7488407493,-0.3618007302,-0.3559001982,-0.3971188664,-0.1411824971,0.1069322899,-0.0116968984,0.132369563,-0.282450974,0.1292226315,-0.0685148016,-0.3066227436,-0.2532494068,0.0661680102,-0.2912592888,0.37689358,0.0371059924,0.0462543555,-0.1763399988,-0.2117870748,0.0755449608,-0.2519206703,0.2068387568,0.4757409096,-0.0536321811,-0.3308727145,-0.2510897517,-0.3509254754,0.5114939213,0.0582129806,-0.1234411448,0.0374809243,-0.2315993756,-0.0505577847,0.0228005331,0.2612951398,0.4988461137,-0.1544501036,-0.0043467344,-0.2505922019,-0.1972514242,-0.2595581412,-0.6558189988,0.1323345304,0.0001985186,-0.0590696856,0.0154642668,-0.0473506525,0.2304743975,-0.0202295464,-0.2316259593,-0.0082493853,0.0318281576,0.0463619269,-0.0746882632,-0.0885469392,-0.2603935003,0.4612663686,-0.2586448491,-0.2531812787,0.0744549409,-0.1491504014,-0.1911728531,0.1480299085,0.2090711147,0.0875900313,0.3608373702,0.1072096601,-0.1648861319,-0.2150407135,0.2460973412,-0.2748446465,-0.0529989079,-0.0701284707,0.1442082226,0.1895029545,0.1028898135,-0.4417264462,-0.2342241704,0.1257224679,-0.0374698527,-0.29451105,-0.0237014927,0.0386206619,-0.1980319023,-0.153936699,-0.1440274268,0.6503120661,0.5139029622,-0.0389661789,-0.0411656685,-0.0200424772,-0.0888185203,-0.1168877855,0.189748168,0.471029669,-0.8301251531,0.0581733771,-0.073723793,-0.0327455476,0.0786821321,0.290413022,-0.1963496655,0.6375905871,-0.0681122839,0.4610254467,0.4078597426,0.1795091927,-0.209945485,0.129501909,0.0335644297,0.1563652307,-0.1035326198,-0.0628024265,0.482629478,-0.217130661,-0.0457123183,0.1129511371,-0.1889507473,-0.0673819408,0.0496089943,0.0885250792,0.3160016835,-0.1392380297,0.1117528677,-0.0378336944,-0.012170149,0.3528766036,0.1717862189,-0.1578956246,0.2773495018,0.0704207122,0.2738774419,0.3906757236,-0.0257772449,-0.279006958,-0.0295180939,0.0410264358,0.1781962514,0.2417199165,-0.0416110381,-0.1750018299,-0.2224463373,-0.1559338272,0.0179277286,-0.1450502425,-0.0055101663,0.0471382737,-0.1559309512,0.0154537167,-0.3025246561,0.1742354631,-0.2926347852,0.2892850637,-0.186340034,0.0022218288,0.2258528769,-0.1295737922,-0.174726516,0.3517545462,-0.0479675345,-0.0908577517,-0.0018246861,0.2156728357,0.3708452582,-0.1235960498,-0.2311401814,0.4332295954,0.3695605695,-0.0585544668,0.4894850254,-0.1707594991,0.0279553961,-0.2128297389,0.0199520849,0.3229399323,0.1864878386,0.2040638477,0.0193661116,-0.0827098936,0.320053041,-0.0255262554,-0.0114450492,-0.0107290959,-0.0304203089,0.1878430843,0.3391094804,0.0203353558,0.1094944999,0.3720138669,0.2509760261,-0.0894872174,0.1424987316,0.3258264363,-0.2863969803,-0.194758594,-0.238205269,0.0374981053,0.0781139135,0.0403730087,0.3694118857,0.0393930748,-0.1495297104,-0.1556714773,0.2969370782,0.0737481341,-0.0004074918,0.3609405756,-0.1661285311,0.0918624327,-0.2067154944,-0.5002880692,-0.084069699,0.2055984288,-0.1579047889,0.1245988756,0.0868013799,-0.0448261052,-0.4290056527,-0.1948902905,-0.3251187205,0.0576876961,-0.0241645072,-0.139638111,0.1719357371,0.0046177586,0.2202244699,0.4302629828,0.0097249439,0.0247084051,0.2405432761,-0.3586658835,-0.2610946894,-0.0238927603,0.0922227502,0.2567632496,0.1506670117,-0.0544148572,0.0770779848,0.0201701764,-0.4987143278,0.120447278,-0.1143906936,0.4396464825,-0.0026855643,0.1807170063,0.0598223917,-0.0829074606,-0.2721565664,0.1053874642,-0.1648614854,-0.1816079021,-0.0923047438,-0.0269051287,0.1183580607,-0.1891198158,-0.4272451997,-0.0454994552,0.1483100355,0.197660014,0.2029997408,-0.0891488716,0.0353996269,0.2241721898,0.1744184196,0.0173479877,0.0927595347,0.0055636792,0.2995705903,-0.2556292713,-0.2299671471,-0.0042510084,-0.1577004939,0.2659344077,0.1460161656,-0.1974955797,0.3054780066,-0.1486777365,-0.0172893461,0.0027437482,-0.0622545667,0.4194029272,-0.0075166789,0.2778610587,-0.1237753928,0.0801520348,0.1142040789,-0.2657427788,0.2310200185,-0.0213963818,0.6587687731,0.0096473778,0.595318675,0.3725365698,-0.1631973237,0.0512430891,-0.0020027994,-0.0079434486,-0.2485302985,-0.199775964,-0.0131069031,0.0494425632,-0.0750421286,0.3964889646,0.2265223712,0.0258632787,0.1640276462,-0.098757945,-0.1659605801,-0.2516511977,0.3336081207,-0.5507493615,0.4910554886,-0.0608590022,-0.0358718038,-0.3362339735,-0.159062624,-0.2755900919,0.3481835127,-0.0206212532,0.0376950167,-0.5603550076,0.3448073864,-0.5934057832,-0.0002354426,0.2043755203,0.3582433164,0.0887544379,0.1698133796,-0.0974612013,-0.0033032198,0.3416797817,-0.1204902902,-0.259375006,-0.0677659288,-0.0194223635,-0.2325031757,0.2992912531,-0.0677912161,0.2290980369,0.5192552209,0.6033684611,-0.2561033666,-0.194517687,-0.3954640627,0.5209569931,-0.1781931818,-0.0339437202,0.2066590488,-0.018199563,-0.2318158001,0.0242538378,0.1156837791,0.167842716,0.0679865256,0.0055928454,-0.1464352459,-0.2678794861,0.2991909385,0.0570978001,0.148682937,-0.2839533091,0.1000145525,0.1886838973,-0.1420979798,0.0104451096,0.1083703265,0.2093972564,-0.3746312857,-0.0051910356,-0.1714086086,0.5225891471,-0.0267724954,0.1416906267,0.1670989245,-0.0657753348,-0.1928976029,-0.235263139,0.5005987883,-0.0690164492,0.0530901887,-0.5153008103,-0.613602221,0.1214954928,-0.0229012929,-0.0899649933,0.3768248558,-0.103215158,-0.6049001217,0.5901966095,0.260150075,1.0232452154,0.0431065857,0.1503717452,-0.1089649871,0.2251690775,0.3162970543,-0.6155140996,-0.2120399177,-0.4247459769,0.2241093963,-0.1044999585,0.0016026107,0.1995890439,0.492274344,-0.1408635527,0.0061774817,0.1284854859,-0.0184874684,0.1666332632,0.4509717822,-0.2921358049,-0.2491112798,0.103267774,0.0125362873,-0.0104395552,-0.2567545474,-0.1371281594,0.1455800533,0.0159281511,-0.0207165126,-0.5003291965,-0.3555761874,-0.3252560794,0.0169255249,0.0235843081,-0.1969290972,0.480707705,0.2973553836,0.6376609802,0.4177748859,-0.2174187899,0.0545308106,-0.2332278937,-0.2094623446,0.0870260149,-0.1930740923,0.3340639174,0.0465449542,-0.206200704,-0.2373727262,0.0127107017,-0.0060689356,-0.323587507,-0.3988482952,0.3249945939,-0.5288798809,0.0179121494,0.1287324578,0.1066614166,-0.1481942236,-0.0217012689,0.2406314313,-0.057206098,0.0093886005,-0.1806469113,-0.0652483478,-0.0881990492,0.0306702368,0.0721116588,-0.2707574069,0.5002795458,0.2290663272,-0.0856099948,-0.1685410142,0.393024087,0.2021217644,-0.4059944153,-0.0106952945,-0.3149010837,-0.1877924502,0.0534737706,0.2731820643,0.253331095,-0.0948422328,-0.0701783001,-0.52008003,-0.388482511,-0.0219455827,-0.0008830926,0.2809630334,-0.1774866134,0.2880315185,-0.1049582139,0.2328548431,-0.1262019575,0.2043265849,0.2575089931,0.3019227386,-0.0981618986,-0.0938295946,-0.258790195,0.0029152541,-0.1224558502,0.2662453055,-0.0864104331,-0.0585898198,-0.4252178967,0.2733770609,0.2728996277,-0.3364344835,0.0994903296,0.0131537002,-0.209160164,-0.3021591902,0.0985011458,-0.0063386844,0.0383670591,0.1578012407,0.4322455525,0.2658196688,-0.3363594115,-0.1921830177,0.0864653289,0.2084959596,0.3757403791,0.1418391913,-0.0435213149,-0.0524396412,-0.2586334348,0.564406395,0.4757580459,-0.0372255519,0.0306963399,0.0540924221,0.1962610781,0.1489271075,0.0322389938,0.4324266911,0.0279383026,-0.1799346805,0.6269698739,0.1011174396,0.0088850595,0.098473154,0.1509217769,-0.1831855029,0.0132506499,0.350134939,0.0708893016,0.1682900935,-0.2239841372,-0.237785086,0.5152842402,0.0842829272,0.0611526892,0.2125634253,-0.1789619923,0.0388643704,0.1503523737,0.1340723634,0.3176992238,0.6118144989,-0.0381808616,0.3145724833,0.0784453303,-0.0766486749,0.0345227011,-0.2716097236,-0.0748949796,0.0597057715,0.1436005235,-0.1116012409,-0.1553779542,0.1224191114,0.0364551283,-0.0922944918,0.1190306619,0.2025168091,-0.1732835174,-0.4344192743,-0.3709191382,-0.2609013021,0.0200018529,-0.1455055326,0.0323614068,-0.256087929,-0.3558675945,-0.1335449964,0.0917848125,-0.4081516266,0.3311243355,-0.1059882715,-0.1241022497,-0.1314977109,0.4013444185,0.0515605696,0.0637335181,-0.145848915,0.1716765463,0.1098219976,0.2321894169,-0.2296707332,0.2294413894,-0.3835394382,-0.04324558,-0.0119285658,0.3488596976,0.0152149349,0.3369501531,0.1603622288,-0.0498322584,0.0187156517,-0.0504738055,0.3851917684,-0.4717318118,-0.1144078746,0.4172973335,0.206670627,-0.1017840132,-0.0110934908,0.0506956913,-0.2818850577,-0.0741807371,0.5695211887,-0.1110723838,0.1680381596,-0.1478542984,-0.0066866805,-0.0798197389,0.388597995,0.4119670391,0.223066628,-0.2060917467,-0.3186214864,-0.4061474204,-0.0327869654,-0.0393070802,-0.0504405722,-0.0943737552,0.0469556227,0.1439488679,0.1674648225,-0.3382814229,0.3574922383,-0.2727070153,-0.0328164585,-0.0343278423,-0.0779801607,-0.0296435617,0.5165071487,-0.1650056243,-0.1361413747,-0.0299267042,-0.0273407493,-0.094295226,0.0314810909,-0.0648434162,0.3325419128,-0.188897565,0.1950983852,0.1513123661,0.4023987949,0.0242118556,0.2504161298,-0.2604077756,0.144603923,-0.2374410331,0.3401325643,0.2497676164,0.3828650415,-0.3925804496,0.1149356663,-0.0287863556,0.0491124503,0.1830530316,-0.2654310465,-0.3182499111,0.1140787005,-0.3002091944,-0.1886045486,0.1572879106,0.5893192291,-0.1131974906,0.3527789116,0.0976468623,0.0711608753,0.0813257545,-0.3661095202,-0.5172597766,-0.0777519792,-0.0363784693,0.5826897025,-0.1148343682,-0.5886265039,-0.0188335087,0.5914517045,0.009118895,-0.257129997,0.2188809216,0.3040905893,-0.2758349478,-0.2128051072,0.5650271773,-0.1329900026,-0.0251236074,-0.0471128672,-0.3848664761]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2279","title":"Compatibility with Ubuntu 18 and GLIBC 2.27?","comments":"From the trace this seems like an error in the tokenizer library instead.\r\n\r\nDo you mind opening an issue at https:\/\/github.com\/huggingface\/tokenizers instead?","body":"## Describe the bug\r\nFor use on Ubuntu systems, it seems that datasets requires GLIBC 2.29. However, Ubuntu 18 runs with GLIBC 2.27 and it seems [non-trivial to upgrade GLIBC to 2.29 for Ubuntu 18 users](https:\/\/www.digitalocean.com\/community\/questions\/how-install-glibc-2-29-or-higher-in-ubuntu-18-04). \r\n\r\nI'm not sure if there is anything that can be done about this, but I'd like to confirm that using huggingface\/datasets requires either an upgrade to Ubuntu 19\/20 or a hand-rolled install of a higher version of GLIBC.\r\n\r\n## Steps to reproduce the bug\r\n1. clone the transformers repo\r\n2. move to examples\/pytorch\/language-modeling\r\n3. run example command:\r\n```python run_clm.py --model_name_or_path gpt2 --dataset_name wikitext --dataset_config_name wikitext-2-raw-v1 --do_train --do_eval --output_dir \/tmp\/test-clm```\r\n\r\n\r\n## Expected results\r\nAs described in the transformers repo.\r\n\r\n## Actual results\r\n```Traceback (most recent call last):\r\n File \"run_clm.py\", line 34, in \r\n from transformers import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/__init__.py\", line 2487, in __getattr__\r\n return super().__getattr__(name)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/file_utils.py\", line 1699, in __getattr__\r\n module = self._get_module(self._class_to_module[name])\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/__init__.py\", line 2481, in _get_module\r\n return importlib.import_module(\".\" + module_name, self.__name__)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/__init__.py\", line 19, in \r\n from . import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/layoutlm\/__init__.py\", line 23, in \r\n from .tokenization_layoutlm import LayoutLMTokenizer\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/layoutlm\/tokenization_layoutlm.py\", line 19, in \r\n from ..bert.tokenization_bert import BertTokenizer\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/bert\/tokenization_bert.py\", line 23, in \r\n from ...tokenization_utils import PreTrainedTokenizer, _is_control, _is_punctuation, _is_whitespace\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils.py\", line 26, in \r\n from .tokenization_utils_base import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 68, in \r\n from tokenizers import AddedToken\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/tokenizers\/__init__.py\", line 79, in \r\n from .tokenizers import (\r\nImportError: \/lib\/x86_64-linux-gnu\/libm.so.6: version `GLIBC_2.29' not found (required by \/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/tokenizers\/tokenizers.cpython-37m-x86_64-linux-gnu.so)\r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```\r\n- Datasets: 1.6.1\r\n- Python: 3.7.10 (default, Feb 26 2021, 18:47:35) \r\n[GCC 7.3.0]\r\n- Platform: Linux-4.15.0-128-generic-x86_64-with-debian-buster-sid\r\n\r\n```\r\n","comment_length":22,"text":"Compatibility with Ubuntu 18 and GLIBC 2.27?\n## Describe the bug\r\nFor use on Ubuntu systems, it seems that datasets requires GLIBC 2.29. However, Ubuntu 18 runs with GLIBC 2.27 and it seems [non-trivial to upgrade GLIBC to 2.29 for Ubuntu 18 users](https:\/\/www.digitalocean.com\/community\/questions\/how-install-glibc-2-29-or-higher-in-ubuntu-18-04). \r\n\r\nI'm not sure if there is anything that can be done about this, but I'd like to confirm that using huggingface\/datasets requires either an upgrade to Ubuntu 19\/20 or a hand-rolled install of a higher version of GLIBC.\r\n\r\n## Steps to reproduce the bug\r\n1. clone the transformers repo\r\n2. move to examples\/pytorch\/language-modeling\r\n3. run example command:\r\n```python run_clm.py --model_name_or_path gpt2 --dataset_name wikitext --dataset_config_name wikitext-2-raw-v1 --do_train --do_eval --output_dir \/tmp\/test-clm```\r\n\r\n\r\n## Expected results\r\nAs described in the transformers repo.\r\n\r\n## Actual results\r\n```Traceback (most recent call last):\r\n File \"run_clm.py\", line 34, in \r\n from transformers import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/__init__.py\", line 2487, in __getattr__\r\n return super().__getattr__(name)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/file_utils.py\", line 1699, in __getattr__\r\n module = self._get_module(self._class_to_module[name])\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/__init__.py\", line 2481, in _get_module\r\n return importlib.import_module(\".\" + module_name, self.__name__)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/__init__.py\", line 19, in \r\n from . import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/layoutlm\/__init__.py\", line 23, in \r\n from .tokenization_layoutlm import LayoutLMTokenizer\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/layoutlm\/tokenization_layoutlm.py\", line 19, in \r\n from ..bert.tokenization_bert import BertTokenizer\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/bert\/tokenization_bert.py\", line 23, in \r\n from ...tokenization_utils import PreTrainedTokenizer, _is_control, _is_punctuation, _is_whitespace\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils.py\", line 26, in \r\n from .tokenization_utils_base import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 68, in \r\n from tokenizers import AddedToken\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/tokenizers\/__init__.py\", line 79, in \r\n from .tokenizers import (\r\nImportError: \/lib\/x86_64-linux-gnu\/libm.so.6: version `GLIBC_2.29' not found (required by \/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/tokenizers\/tokenizers.cpython-37m-x86_64-linux-gnu.so)\r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```\r\n- Datasets: 1.6.1\r\n- Python: 3.7.10 (default, Feb 26 2021, 18:47:35) \r\n[GCC 7.3.0]\r\n- Platform: Linux-4.15.0-128-generic-x86_64-with-debian-buster-sid\r\n\r\n```\r\n\nFrom the trace this seems like an error in the tokenizer library instead.\r\n\r\nDo you mind opening an issue at https:\/\/github.com\/huggingface\/tokenizers instead?","embeddings":[-0.142811805,-0.1723928154,0.1075327396,0.1302344054,0.0596383885,-0.0772048682,0.2908620238,0.2598784566,0.2259389609,0.0111382147,0.1740464121,0.2267897278,-0.2015923262,0.0207950752,0.0591229573,-0.2397232503,0.4002413452,0.1342248768,-0.3602030575,-0.0450363941,-0.1554612368,0.5982704759,-0.3146861196,-0.010957269,-0.4952140749,0.0392144546,-0.0519102439,0.1466761082,0.13439174,-0.1309381276,0.4062684178,-0.1249714494,-0.0909759551,0.754386127,-0.0001224395,0.0493884832,0.2803669572,-0.0356524996,-0.3091667891,-0.3431194127,0.4603341222,-0.2034225613,0.1984533519,0.0210759174,-0.1604524255,-0.0312387682,0.0431692265,-0.1770747751,0.306265533,0.2292727828,0.1635718495,0.6086139679,0.2872385085,-0.2450601906,0.1327299774,0.0671043396,-0.3019824922,0.1775183678,0.3917792141,0.0825865865,0.1014894694,0.3350143731,0.1052379236,-0.2016983181,0.4753361344,-0.0923256204,-0.2559214234,-0.2584765255,0.0923650339,0.1190377399,0.1131736115,-0.3634867966,-0.4058305025,-0.5498496294,-0.2388646007,0.2090977579,0.367434144,-0.2292598337,-0.0580108985,0.1378621012,-0.3331091702,-0.3827241659,-0.1197224706,-0.0331870615,-0.012718291,0.8859995604,-0.0173584167,0.1676160395,0.0610819794,0.0065577272,0.00012063,-0.1602582037,0.00854606,0.1647605896,-0.1325274855,-0.0888525993,0.053081993,-0.1233673468,0.0066031292,-0.1705616266,-0.2632810771,-0.1685592979,-0.1061792672,0.1360406727,0.0399863385,0.2783112526,-0.1579051763,0.4739059508,0.2186329961,0.2473488599,0.321674943,-0.1070386618,0.1026504189,-0.2637830079,-0.1310983449,0.1177573726,0.2092296034,-0.2898231447,-0.017997589,-0.0566714592,-0.0108650476,0.2226473093,0.0681747422,0.1660168469,-0.0328398757,0.3804472387,0.0928115323,0.2229461521,-0.1679702252,0.2382942885,-0.1164787933,-0.0974209011,-0.0144361146,-0.0716307014,0.1691595614,-0.0391608514,0.2538377941,0.2059005052,0.4203241169,0.0874161497,-0.1992437989,0.0457497984,0.0518945269,0.3462992609,-0.4304388165,0.3338389397,0.1684209108,-0.3088178933,-0.3119699657,-0.0098933876,-0.0638596117,-0.0914926156,-0.0469189622,0.0632655919,-0.2492465824,0.013064472,0.0124965664,0.1144580767,0.172650516,-0.4946889877,0.2769922614,-0.233783111,-0.2119734287,-0.0241523031,0.1076139584,0.0356387608,-0.2591195107,-0.4424846768,0.3240971267,-0.0839912519,-0.0312667713,0.243970871,-0.1142867506,-0.3914651573,-0.1942076236,-0.0358673967,-0.2203388661,-0.1526675969,-0.5307197571,0.0207097698,0.0859973133,0.2394689918,-0.0067258072,-0.0262469035,0.0615262613,-0.2693711817,-0.1275818497,0.3495996594,0.1197857261,0.0980074033,-0.4085882008,-0.3331645727,-0.0149042811,-0.0940913484,0.1942503303,-0.164009288,-0.1923215687,0.1174797788,0.4155255854,-0.2786252499,-0.0290605333,0.1464767158,0.5196980834,0.000521138,0.0618164539,-0.1114572659,-0.6179240942,0.4469006062,-0.2873841226,0.1566895992,-0.3138538003,-0.1721369773,-0.0007825637,0.1277842969,0.0836197957,-0.3957600594,-0.0320510939,0.0547227673,0.1871065795,0.0526270345,-0.1427202672,0.2477567494,0.0274521876,0.0078920322,-0.4396381378,0.3152775764,-0.0824749023,-0.1737359762,0.1070349589,0.2113476843,0.0512974896,-0.329780519,-0.1780619919,0.3240359724,-0.127081722,0.1900217682,-0.3105565012,0.0714245811,0.2983856797,-0.0419499762,-0.0758858919,0.3169097304,-0.007165622,0.0673383176,0.086399734,0.1811774969,0.0145425787,0.1595082581,0.2673688233,0.0799954161,0.0134186652,-0.0280264802,-0.0551781021,-0.205769375,0.053303048,-0.0545434989,-0.0722139925,0.1719166487,-0.1459628791,-0.0196434744,0.6664274335,0.0451898836,0.0067488635,0.1859145761,-0.1284980327,-0.104209736,0.1641649604,0.3101860285,0.2909154594,0.1102724299,-0.0396555513,0.2677046955,-0.0750832558,-0.1431828141,0.2155104727,0.096880801,0.4194385409,0.0191845484,0.0385328494,0.30566594,0.0614310354,-0.0603859276,-0.263718158,0.0914352238,-0.3121880293,0.4507681429,-0.0368034206,0.0784676224,-0.1586845964,-0.1577325314,-0.7852627635,-0.1883606911,0.0067408956,0.0399592854,-0.1980149597,0.3214106858,-0.0444885045,0.1915934533,-0.2949773371,-0.2638169527,-0.2137812376,0.0314920433,-0.5112403631,-0.0231350772,0.2785100937,0.1089708582,0.0357541703,-0.0728168488,-0.1063587368,0.0394180641,-0.5937504172,0.2073419094,-0.096671924,0.3363634646,0.0931136757,-0.1605436951,-0.1574152261,-0.2098421901,0.2231027633,-0.8244650364,-0.0332416594,-0.0670602322,-0.033060573,-0.2129868716,-0.310906589,0.0425588973,-0.1937351674,-0.2245925218,0.0641586706,0.0267934725,0.1504383236,0.6384630203,0.0632362068,0.0668419898,-0.368848443,-0.0192970764,0.0950938016,0.1094495058,0.264444232,0.0237147622,-0.464327395,-0.1210149899,-0.048923403,0.1750032455,0.1410595775,-0.3063699901,-0.2472435981,-0.4217748642,0.0800869763,0.193205893,-0.0237948094,0.4895001054,-0.2125200182,0.1190253422,-0.2503902018,-0.1644293964,0.1279211491,-0.0880082622,0.1465210468,0.0695360452,0.1001236588,0.1234058142,0.7045325637,0.4875889719,-0.0884720162,0.1642232537,0.1532943696,0.4174870849,-0.2423343807,-0.3420737088,-0.0094390912,0.0148862777,-0.1092205122,0.1582188308,-0.0186669808,-0.1451962143,-0.3119723201,-0.0997494906,-0.0315329805,-0.3864445686,0.1610625535,0.3849331141,0.4469650388,0.1315151006,0.0702414885,-0.0262658466,-0.0061874171,0.2651171982,0.2288968265,0.100570485,0.1592470706,0.0106114158,0.0563791692,-0.4902714789,0.1502764821,0.1207980216,0.2744883895,-0.0123045016,-0.2598343492,-0.0926972628,-0.0416162908,0.3547393978,0.2080987543,-0.1358809322,0.003782179,-0.0252398606,-0.5615041256,-0.0845361948,-0.1922897249,0.0295988917,0.6906397343,0.7015839219,-0.2574883401,-0.2235772908,-0.0012592168,0.198622033,-0.0859060064,-0.124964878,-0.4714793563,0.0180772059,-0.2427761555,-0.0761247426,0.2081302255,0.0937619209,-0.0123717142,0.1114062369,-0.079252474,-0.3343803585,-0.1157324389,0.2385465652,0.2372391522,-0.085375011,0.1637600958,0.1426896751,-0.0472987443,-0.0937493518,0.5736646056,0.0868205428,-0.1973429024,-0.0140719935,0.2528957725,0.382376194,0.3725439012,-0.3145151734,-0.0523282252,0.0576482527,0.1869302541,0.071708411,-0.2260339409,0.1847774684,0.1969469488,-0.3472759426,-0.2714084983,0.3154604435,0.2166334838,-0.0438227467,-0.0723231733,0.3505695164,-0.2092162371,0.2476397902,0.0264697745,0.8544824123,0.1244776174,0.2327861786,0.2973266244,-0.025873825,0.7188073397,0.3994665146,0.1941176206,-0.4109279513,-0.3418685794,-0.0140615785,-0.0808945149,0.4022877216,-0.2950657606,-0.3223613203,0.3390083611,0.2339359522,0.1027220339,0.0795767307,0.0588903092,-0.1981144547,-0.1559313834,-0.291967243,0.0332191549,0.0540284626,0.3022700548,-0.0552458391,-0.1685113311,-0.3493580818,-0.4462107718,-0.2820329368,0.2805140316,-0.327734679,-0.0141637884,0.5910345912,-0.2505148351,0.4873715937,-0.0331625193,0.1778804213,0.2963130772,-0.2713108957,0.2396015674,-0.2851697505,0.0019540354,0.1223955229,0.0028118626,0.3139875829,-0.1776818484,-0.1181388199,-0.2422830164,-0.1403474063,-0.1115921363,0.0921525061,-0.1329473555,-0.0848161876,-0.0531583615,-0.1113059521,0.0065422328,0.1808992475,-0.286334008,0.1043158323,-0.0331024565,-0.1111448929,-0.1462763101,-0.080900386,-0.2520555258,-0.068032667,0.1273733526,-0.1710668951,0.0880030319,0.4218143523,0.4625130594,-0.1962947994,0.0047671441,-0.2927006483,0.1706172973,-1.0560743809,0.043469809,0.1184038892,-0.0590664968,-0.2321721762,0.4432278574,0.3857446909,-0.2091692239,-0.1934963167,-0.1018265486,-0.268640697,0.0025457847,-0.0818316191,0.1668409109,-0.1092051417,-0.1131565645,0.0659626871,0.2004723847,-0.218396917,-0.0878222957,0.0255133975,-0.1071741208,-0.0951175913,-0.3705217242,0.2673198581,-0.2820159495,-0.029769456,0.1741184592,-0.101277329,-0.0719066858,-0.2767718732,0.1109216884,0.3009899259,-0.189226374,-0.159265101,0.0971671641,-0.0445916131,-0.251255244,0.3204421699,0.3702948391,-0.0054153316,0.1673935503,0.1635505855,0.0742891654,0.005182656,0.1371880323,0.000412804,0.1637768447,0.2630919516,0.2655125558,0.0087822285,0.127626285,0.0735779777,0.240582481,-0.0842642337,0.1357299834,0.0359867997,-0.1764383316,0.0797156915,0.2588960826,0.3712697625,0.2570674419,-0.093263872,-0.1037778854,0.2864119709,0.0947101712,-0.2520166934,-0.081054166,0.1429417729,0.225741744,0.0062077292,0.2939155102,0.3655900061,-0.0921466872,0.292607218,0.2397045046,0.2022410482,0.312107712,0.4906820357,0.3819250762,0.2148486078,0.1730962843,0.3125342131,0.0882166103,0.1764625162,-0.0680328235,-0.2254639566,0.0927049667,-0.3624399602,0.0612281598,0.2626800835,-0.3448006213,-0.1063368097,-0.2771709561,0.1597803831,0.1789622754,0.0619780459,0.3423141837,-0.1999441236,-0.1018967852,-0.2147206068,0.1653594822,-0.1557972133,-0.1243789792,0.3625316322,-0.057580851,-0.2877358198,0.014200896,-0.1381665319,-0.1885128617,0.0023840433,0.0862934291,-0.1487503946,-0.1796547621,0.0309437681,0.1107827649,0.2017217278,0.0674140453,0.2881188989,0.2268684953,-0.1005053893,-0.2474524081,0.2419550121,0.4578117132,-0.0087836161,0.0842525363,0.1260661036,-0.1481120437,0.1541395187,0.0521552935,0.3505434096,-0.0320188329,0.1383468062,0.1348989457,0.0451738723,-0.0982146561,-0.0344973505,-0.3612898886,0.3455745578,-0.2912254632,0.7918094993,-0.4724257886,-0.0623592213,-0.141964376,0.2556201518,-0.4287148714,-0.1302291751,0.3787559569,-0.0227484945,0.145622462,-0.2917760909,0.0362585858,0.0979721546,0.3941867054,0.3059559464,0.2157992274,0.093060106,-0.0810708702,-0.4841988087,0.238054812,0.0206123721,0.3427098095,0.214920342,-0.2224812061,0.0794468522,0.0694118366,0.21201922,0.4279299974,-0.2316808552,-0.2786812782,-0.1207926124,0.0128874909,-0.4145508707,-0.3672766387,0.1829686463,-0.1834723204,-0.0652342141,-0.3103603423,-0.0375364646,-0.2787756026,0.0337750688,-0.1651547551,-0.319103539,0.2987806499,0.3485632241,0.2140877694,-0.0922433287,-0.3349895775,-0.2125594765,-0.060501691,-0.1788991988,0.2837646306,-0.2700270116,0.3132869601,-0.2699929476,-0.1873683184,-0.1783807427,0.0091541251,0.0317165777,-0.1672866344,-0.0790593848,0.0237620622,-0.2985110581,0.0984308943,0.3186790645,0.3591607213,0.053078372,-0.0591421835,-0.3668980896,-0.2386033386,0.6176970601,-0.3460363746,-0.1892908812,-0.1798475683,0.1215003207,0.1456499249,-0.4370312095,-0.6966891289,-0.0974395275,0.0974253118,0.0474987887,0.074125506,0.059034422,-0.077841416,-0.0506133586,-0.0690909028,0.6845296025,0.0159644112,-0.2760543227,0.2355748713,-0.1013072059]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2279","title":"Compatibility with Ubuntu 18 and GLIBC 2.27?","comments":"Hi @tginart, thanks for reporting.\r\n\r\nI think this issue is already open at `tokenizers` library: https:\/\/github.com\/huggingface\/tokenizers\/issues\/685","body":"## Describe the bug\r\nFor use on Ubuntu systems, it seems that datasets requires GLIBC 2.29. However, Ubuntu 18 runs with GLIBC 2.27 and it seems [non-trivial to upgrade GLIBC to 2.29 for Ubuntu 18 users](https:\/\/www.digitalocean.com\/community\/questions\/how-install-glibc-2-29-or-higher-in-ubuntu-18-04). \r\n\r\nI'm not sure if there is anything that can be done about this, but I'd like to confirm that using huggingface\/datasets requires either an upgrade to Ubuntu 19\/20 or a hand-rolled install of a higher version of GLIBC.\r\n\r\n## Steps to reproduce the bug\r\n1. clone the transformers repo\r\n2. move to examples\/pytorch\/language-modeling\r\n3. run example command:\r\n```python run_clm.py --model_name_or_path gpt2 --dataset_name wikitext --dataset_config_name wikitext-2-raw-v1 --do_train --do_eval --output_dir \/tmp\/test-clm```\r\n\r\n\r\n## Expected results\r\nAs described in the transformers repo.\r\n\r\n## Actual results\r\n```Traceback (most recent call last):\r\n File \"run_clm.py\", line 34, in \r\n from transformers import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/__init__.py\", line 2487, in __getattr__\r\n return super().__getattr__(name)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/file_utils.py\", line 1699, in __getattr__\r\n module = self._get_module(self._class_to_module[name])\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/__init__.py\", line 2481, in _get_module\r\n return importlib.import_module(\".\" + module_name, self.__name__)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/__init__.py\", line 19, in \r\n from . import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/layoutlm\/__init__.py\", line 23, in \r\n from .tokenization_layoutlm import LayoutLMTokenizer\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/layoutlm\/tokenization_layoutlm.py\", line 19, in \r\n from ..bert.tokenization_bert import BertTokenizer\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/bert\/tokenization_bert.py\", line 23, in \r\n from ...tokenization_utils import PreTrainedTokenizer, _is_control, _is_punctuation, _is_whitespace\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils.py\", line 26, in \r\n from .tokenization_utils_base import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 68, in \r\n from tokenizers import AddedToken\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/tokenizers\/__init__.py\", line 79, in \r\n from .tokenizers import (\r\nImportError: \/lib\/x86_64-linux-gnu\/libm.so.6: version `GLIBC_2.29' not found (required by \/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/tokenizers\/tokenizers.cpython-37m-x86_64-linux-gnu.so)\r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```\r\n- Datasets: 1.6.1\r\n- Python: 3.7.10 (default, Feb 26 2021, 18:47:35) \r\n[GCC 7.3.0]\r\n- Platform: Linux-4.15.0-128-generic-x86_64-with-debian-buster-sid\r\n\r\n```\r\n","comment_length":16,"text":"Compatibility with Ubuntu 18 and GLIBC 2.27?\n## Describe the bug\r\nFor use on Ubuntu systems, it seems that datasets requires GLIBC 2.29. However, Ubuntu 18 runs with GLIBC 2.27 and it seems [non-trivial to upgrade GLIBC to 2.29 for Ubuntu 18 users](https:\/\/www.digitalocean.com\/community\/questions\/how-install-glibc-2-29-or-higher-in-ubuntu-18-04). \r\n\r\nI'm not sure if there is anything that can be done about this, but I'd like to confirm that using huggingface\/datasets requires either an upgrade to Ubuntu 19\/20 or a hand-rolled install of a higher version of GLIBC.\r\n\r\n## Steps to reproduce the bug\r\n1. clone the transformers repo\r\n2. move to examples\/pytorch\/language-modeling\r\n3. run example command:\r\n```python run_clm.py --model_name_or_path gpt2 --dataset_name wikitext --dataset_config_name wikitext-2-raw-v1 --do_train --do_eval --output_dir \/tmp\/test-clm```\r\n\r\n\r\n## Expected results\r\nAs described in the transformers repo.\r\n\r\n## Actual results\r\n```Traceback (most recent call last):\r\n File \"run_clm.py\", line 34, in \r\n from transformers import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/__init__.py\", line 2487, in __getattr__\r\n return super().__getattr__(name)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/file_utils.py\", line 1699, in __getattr__\r\n module = self._get_module(self._class_to_module[name])\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/__init__.py\", line 2481, in _get_module\r\n return importlib.import_module(\".\" + module_name, self.__name__)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/__init__.py\", line 19, in \r\n from . import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/layoutlm\/__init__.py\", line 23, in \r\n from .tokenization_layoutlm import LayoutLMTokenizer\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/layoutlm\/tokenization_layoutlm.py\", line 19, in \r\n from ..bert.tokenization_bert import BertTokenizer\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/models\/bert\/tokenization_bert.py\", line 23, in \r\n from ...tokenization_utils import PreTrainedTokenizer, _is_control, _is_punctuation, _is_whitespace\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils.py\", line 26, in \r\n from .tokenization_utils_base import (\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 68, in \r\n from tokenizers import AddedToken\r\n File \"\/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/tokenizers\/__init__.py\", line 79, in \r\n from .tokenizers import (\r\nImportError: \/lib\/x86_64-linux-gnu\/libm.so.6: version `GLIBC_2.29' not found (required by \/home\/tginart\/anaconda3\/envs\/huggingface\/lib\/python3.7\/site-packages\/tokenizers\/tokenizers.cpython-37m-x86_64-linux-gnu.so)\r\n```\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```\r\n- Datasets: 1.6.1\r\n- Python: 3.7.10 (default, Feb 26 2021, 18:47:35) \r\n[GCC 7.3.0]\r\n- Platform: Linux-4.15.0-128-generic-x86_64-with-debian-buster-sid\r\n\r\n```\r\n\nHi @tginart, thanks for reporting.\r\n\r\nI think this issue is already open at `tokenizers` library: https:\/\/github.com\/huggingface\/tokenizers\/issues\/685","embeddings":[-0.142811805,-0.1723928154,0.1075327396,0.1302344054,0.0596383885,-0.0772048682,0.2908620238,0.2598784566,0.2259389609,0.0111382147,0.1740464121,0.2267897278,-0.2015923262,0.0207950752,0.0591229573,-0.2397232503,0.4002413452,0.1342248768,-0.3602030575,-0.0450363941,-0.1554612368,0.5982704759,-0.3146861196,-0.010957269,-0.4952140749,0.0392144546,-0.0519102439,0.1466761082,0.13439174,-0.1309381276,0.4062684178,-0.1249714494,-0.0909759551,0.754386127,-0.0001224395,0.0493884832,0.2803669572,-0.0356524996,-0.3091667891,-0.3431194127,0.4603341222,-0.2034225613,0.1984533519,0.0210759174,-0.1604524255,-0.0312387682,0.0431692265,-0.1770747751,0.306265533,0.2292727828,0.1635718495,0.6086139679,0.2872385085,-0.2450601906,0.1327299774,0.0671043396,-0.3019824922,0.1775183678,0.3917792141,0.0825865865,0.1014894694,0.3350143731,0.1052379236,-0.2016983181,0.4753361344,-0.0923256204,-0.2559214234,-0.2584765255,0.0923650339,0.1190377399,0.1131736115,-0.3634867966,-0.4058305025,-0.5498496294,-0.2388646007,0.2090977579,0.367434144,-0.2292598337,-0.0580108985,0.1378621012,-0.3331091702,-0.3827241659,-0.1197224706,-0.0331870615,-0.012718291,0.8859995604,-0.0173584167,0.1676160395,0.0610819794,0.0065577272,0.00012063,-0.1602582037,0.00854606,0.1647605896,-0.1325274855,-0.0888525993,0.053081993,-0.1233673468,0.0066031292,-0.1705616266,-0.2632810771,-0.1685592979,-0.1061792672,0.1360406727,0.0399863385,0.2783112526,-0.1579051763,0.4739059508,0.2186329961,0.2473488599,0.321674943,-0.1070386618,0.1026504189,-0.2637830079,-0.1310983449,0.1177573726,0.2092296034,-0.2898231447,-0.017997589,-0.0566714592,-0.0108650476,0.2226473093,0.0681747422,0.1660168469,-0.0328398757,0.3804472387,0.0928115323,0.2229461521,-0.1679702252,0.2382942885,-0.1164787933,-0.0974209011,-0.0144361146,-0.0716307014,0.1691595614,-0.0391608514,0.2538377941,0.2059005052,0.4203241169,0.0874161497,-0.1992437989,0.0457497984,0.0518945269,0.3462992609,-0.4304388165,0.3338389397,0.1684209108,-0.3088178933,-0.3119699657,-0.0098933876,-0.0638596117,-0.0914926156,-0.0469189622,0.0632655919,-0.2492465824,0.013064472,0.0124965664,0.1144580767,0.172650516,-0.4946889877,0.2769922614,-0.233783111,-0.2119734287,-0.0241523031,0.1076139584,0.0356387608,-0.2591195107,-0.4424846768,0.3240971267,-0.0839912519,-0.0312667713,0.243970871,-0.1142867506,-0.3914651573,-0.1942076236,-0.0358673967,-0.2203388661,-0.1526675969,-0.5307197571,0.0207097698,0.0859973133,0.2394689918,-0.0067258072,-0.0262469035,0.0615262613,-0.2693711817,-0.1275818497,0.3495996594,0.1197857261,0.0980074033,-0.4085882008,-0.3331645727,-0.0149042811,-0.0940913484,0.1942503303,-0.164009288,-0.1923215687,0.1174797788,0.4155255854,-0.2786252499,-0.0290605333,0.1464767158,0.5196980834,0.000521138,0.0618164539,-0.1114572659,-0.6179240942,0.4469006062,-0.2873841226,0.1566895992,-0.3138538003,-0.1721369773,-0.0007825637,0.1277842969,0.0836197957,-0.3957600594,-0.0320510939,0.0547227673,0.1871065795,0.0526270345,-0.1427202672,0.2477567494,0.0274521876,0.0078920322,-0.4396381378,0.3152775764,-0.0824749023,-0.1737359762,0.1070349589,0.2113476843,0.0512974896,-0.329780519,-0.1780619919,0.3240359724,-0.127081722,0.1900217682,-0.3105565012,0.0714245811,0.2983856797,-0.0419499762,-0.0758858919,0.3169097304,-0.007165622,0.0673383176,0.086399734,0.1811774969,0.0145425787,0.1595082581,0.2673688233,0.0799954161,0.0134186652,-0.0280264802,-0.0551781021,-0.205769375,0.053303048,-0.0545434989,-0.0722139925,0.1719166487,-0.1459628791,-0.0196434744,0.6664274335,0.0451898836,0.0067488635,0.1859145761,-0.1284980327,-0.104209736,0.1641649604,0.3101860285,0.2909154594,0.1102724299,-0.0396555513,0.2677046955,-0.0750832558,-0.1431828141,0.2155104727,0.096880801,0.4194385409,0.0191845484,0.0385328494,0.30566594,0.0614310354,-0.0603859276,-0.263718158,0.0914352238,-0.3121880293,0.4507681429,-0.0368034206,0.0784676224,-0.1586845964,-0.1577325314,-0.7852627635,-0.1883606911,0.0067408956,0.0399592854,-0.1980149597,0.3214106858,-0.0444885045,0.1915934533,-0.2949773371,-0.2638169527,-0.2137812376,0.0314920433,-0.5112403631,-0.0231350772,0.2785100937,0.1089708582,0.0357541703,-0.0728168488,-0.1063587368,0.0394180641,-0.5937504172,0.2073419094,-0.096671924,0.3363634646,0.0931136757,-0.1605436951,-0.1574152261,-0.2098421901,0.2231027633,-0.8244650364,-0.0332416594,-0.0670602322,-0.033060573,-0.2129868716,-0.310906589,0.0425588973,-0.1937351674,-0.2245925218,0.0641586706,0.0267934725,0.1504383236,0.6384630203,0.0632362068,0.0668419898,-0.368848443,-0.0192970764,0.0950938016,0.1094495058,0.264444232,0.0237147622,-0.464327395,-0.1210149899,-0.048923403,0.1750032455,0.1410595775,-0.3063699901,-0.2472435981,-0.4217748642,0.0800869763,0.193205893,-0.0237948094,0.4895001054,-0.2125200182,0.1190253422,-0.2503902018,-0.1644293964,0.1279211491,-0.0880082622,0.1465210468,0.0695360452,0.1001236588,0.1234058142,0.7045325637,0.4875889719,-0.0884720162,0.1642232537,0.1532943696,0.4174870849,-0.2423343807,-0.3420737088,-0.0094390912,0.0148862777,-0.1092205122,0.1582188308,-0.0186669808,-0.1451962143,-0.3119723201,-0.0997494906,-0.0315329805,-0.3864445686,0.1610625535,0.3849331141,0.4469650388,0.1315151006,0.0702414885,-0.0262658466,-0.0061874171,0.2651171982,0.2288968265,0.100570485,0.1592470706,0.0106114158,0.0563791692,-0.4902714789,0.1502764821,0.1207980216,0.2744883895,-0.0123045016,-0.2598343492,-0.0926972628,-0.0416162908,0.3547393978,0.2080987543,-0.1358809322,0.003782179,-0.0252398606,-0.5615041256,-0.0845361948,-0.1922897249,0.0295988917,0.6906397343,0.7015839219,-0.2574883401,-0.2235772908,-0.0012592168,0.198622033,-0.0859060064,-0.124964878,-0.4714793563,0.0180772059,-0.2427761555,-0.0761247426,0.2081302255,0.0937619209,-0.0123717142,0.1114062369,-0.079252474,-0.3343803585,-0.1157324389,0.2385465652,0.2372391522,-0.085375011,0.1637600958,0.1426896751,-0.0472987443,-0.0937493518,0.5736646056,0.0868205428,-0.1973429024,-0.0140719935,0.2528957725,0.382376194,0.3725439012,-0.3145151734,-0.0523282252,0.0576482527,0.1869302541,0.071708411,-0.2260339409,0.1847774684,0.1969469488,-0.3472759426,-0.2714084983,0.3154604435,0.2166334838,-0.0438227467,-0.0723231733,0.3505695164,-0.2092162371,0.2476397902,0.0264697745,0.8544824123,0.1244776174,0.2327861786,0.2973266244,-0.025873825,0.7188073397,0.3994665146,0.1941176206,-0.4109279513,-0.3418685794,-0.0140615785,-0.0808945149,0.4022877216,-0.2950657606,-0.3223613203,0.3390083611,0.2339359522,0.1027220339,0.0795767307,0.0588903092,-0.1981144547,-0.1559313834,-0.291967243,0.0332191549,0.0540284626,0.3022700548,-0.0552458391,-0.1685113311,-0.3493580818,-0.4462107718,-0.2820329368,0.2805140316,-0.327734679,-0.0141637884,0.5910345912,-0.2505148351,0.4873715937,-0.0331625193,0.1778804213,0.2963130772,-0.2713108957,0.2396015674,-0.2851697505,0.0019540354,0.1223955229,0.0028118626,0.3139875829,-0.1776818484,-0.1181388199,-0.2422830164,-0.1403474063,-0.1115921363,0.0921525061,-0.1329473555,-0.0848161876,-0.0531583615,-0.1113059521,0.0065422328,0.1808992475,-0.286334008,0.1043158323,-0.0331024565,-0.1111448929,-0.1462763101,-0.080900386,-0.2520555258,-0.068032667,0.1273733526,-0.1710668951,0.0880030319,0.4218143523,0.4625130594,-0.1962947994,0.0047671441,-0.2927006483,0.1706172973,-1.0560743809,0.043469809,0.1184038892,-0.0590664968,-0.2321721762,0.4432278574,0.3857446909,-0.2091692239,-0.1934963167,-0.1018265486,-0.268640697,0.0025457847,-0.0818316191,0.1668409109,-0.1092051417,-0.1131565645,0.0659626871,0.2004723847,-0.218396917,-0.0878222957,0.0255133975,-0.1071741208,-0.0951175913,-0.3705217242,0.2673198581,-0.2820159495,-0.029769456,0.1741184592,-0.101277329,-0.0719066858,-0.2767718732,0.1109216884,0.3009899259,-0.189226374,-0.159265101,0.0971671641,-0.0445916131,-0.251255244,0.3204421699,0.3702948391,-0.0054153316,0.1673935503,0.1635505855,0.0742891654,0.005182656,0.1371880323,0.000412804,0.1637768447,0.2630919516,0.2655125558,0.0087822285,0.127626285,0.0735779777,0.240582481,-0.0842642337,0.1357299834,0.0359867997,-0.1764383316,0.0797156915,0.2588960826,0.3712697625,0.2570674419,-0.093263872,-0.1037778854,0.2864119709,0.0947101712,-0.2520166934,-0.081054166,0.1429417729,0.225741744,0.0062077292,0.2939155102,0.3655900061,-0.0921466872,0.292607218,0.2397045046,0.2022410482,0.312107712,0.4906820357,0.3819250762,0.2148486078,0.1730962843,0.3125342131,0.0882166103,0.1764625162,-0.0680328235,-0.2254639566,0.0927049667,-0.3624399602,0.0612281598,0.2626800835,-0.3448006213,-0.1063368097,-0.2771709561,0.1597803831,0.1789622754,0.0619780459,0.3423141837,-0.1999441236,-0.1018967852,-0.2147206068,0.1653594822,-0.1557972133,-0.1243789792,0.3625316322,-0.057580851,-0.2877358198,0.014200896,-0.1381665319,-0.1885128617,0.0023840433,0.0862934291,-0.1487503946,-0.1796547621,0.0309437681,0.1107827649,0.2017217278,0.0674140453,0.2881188989,0.2268684953,-0.1005053893,-0.2474524081,0.2419550121,0.4578117132,-0.0087836161,0.0842525363,0.1260661036,-0.1481120437,0.1541395187,0.0521552935,0.3505434096,-0.0320188329,0.1383468062,0.1348989457,0.0451738723,-0.0982146561,-0.0344973505,-0.3612898886,0.3455745578,-0.2912254632,0.7918094993,-0.4724257886,-0.0623592213,-0.141964376,0.2556201518,-0.4287148714,-0.1302291751,0.3787559569,-0.0227484945,0.145622462,-0.2917760909,0.0362585858,0.0979721546,0.3941867054,0.3059559464,0.2157992274,0.093060106,-0.0810708702,-0.4841988087,0.238054812,0.0206123721,0.3427098095,0.214920342,-0.2224812061,0.0794468522,0.0694118366,0.21201922,0.4279299974,-0.2316808552,-0.2786812782,-0.1207926124,0.0128874909,-0.4145508707,-0.3672766387,0.1829686463,-0.1834723204,-0.0652342141,-0.3103603423,-0.0375364646,-0.2787756026,0.0337750688,-0.1651547551,-0.319103539,0.2987806499,0.3485632241,0.2140877694,-0.0922433287,-0.3349895775,-0.2125594765,-0.060501691,-0.1788991988,0.2837646306,-0.2700270116,0.3132869601,-0.2699929476,-0.1873683184,-0.1783807427,0.0091541251,0.0317165777,-0.1672866344,-0.0790593848,0.0237620622,-0.2985110581,0.0984308943,0.3186790645,0.3591607213,0.053078372,-0.0591421835,-0.3668980896,-0.2386033386,0.6176970601,-0.3460363746,-0.1892908812,-0.1798475683,0.1215003207,0.1456499249,-0.4370312095,-0.6966891289,-0.0974395275,0.0974253118,0.0474987887,0.074125506,0.059034422,-0.077841416,-0.0506133586,-0.0690909028,0.6845296025,0.0159644112,-0.2760543227,0.2355748713,-0.1013072059]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2278","title":"Loss result inGptNeoForCasual","comments":"Hi ! I think you might have to ask on the `transformers` repo on or the forum at https:\/\/discuss.huggingface.co\/\r\n\r\nClosing since it's not related to this library","body":"Is there any way you give the \" loss\" and \"logits\" results in the gpt neo api? ","comment_length":27,"text":"Loss result inGptNeoForCasual\nIs there any way you give the \" loss\" and \"logits\" results in the gpt neo api? \nHi ! I think you might have to ask on the `transformers` repo on or the forum at https:\/\/discuss.huggingface.co\/\r\n\r\nClosing since it's not related to this library","embeddings":[-0.1725184619,-0.5047546625,-0.0494229384,0.473176837,-0.0721403807,-0.3464438319,0.0108577479,0.0527497567,-0.4997354448,0.1422903687,-0.0995590985,-0.0388611034,0.100904882,0.2542966306,0.0646138266,-0.2345316112,-0.0951776728,0.298414439,-0.113592945,-0.2847176492,0.046722319,0.7516429424,-0.0423596501,0.4498414397,-0.1351107955,0.049132552,0.1681937128,-0.1435233206,-0.1015470177,-0.1026617214,0.0283567552,-0.2763161659,-0.1589080542,0.2016085386,-0.0001213429,-0.0264125504,0.2445292324,-0.0716523081,-0.1024020538,0.0959482342,0.2629530728,0.116295673,0.1316123009,-0.1478680223,-0.4186839163,-0.4525968432,-0.1050605103,-0.3300442994,0.2786361277,0.156503126,0.0447498858,0.4286803603,-0.2391529232,0.0296922661,0.1258453876,0.2987284064,-0.3401528001,-0.2796235979,0.0260950234,-0.3103776276,0.3510200679,0.2686969936,0.3863429129,-0.3036620319,0.0181093048,0.3710387051,0.1963018924,0.0109900814,0.0251023266,0.2208728939,0.2916791737,-0.1006681174,-0.1725571603,-0.3272690475,-0.076347895,-0.5108529329,-0.0946470425,0.1157223061,-0.2886104286,0.1027034521,-0.0897850245,0.1514853686,-0.3256933987,-0.1401705146,-0.1286067069,-0.0098543847,0.101522021,0.0668269917,0.2126813978,-0.0489526466,-0.4141369164,0.4112629294,-0.1428751051,-0.2921434939,-0.200718388,-0.5218161941,0.4599569738,0.2730282247,0.0312607288,-0.0208967198,0.1309318691,-0.2003554553,-0.2176462859,0.0279037207,0.137572065,-0.0961225256,0.4073675275,0.1480400711,0.1019013077,-0.0475528203,0.1873628795,-0.152867496,0.4590636492,0.3763070703,0.0797442868,0.1429805011,0.0296133645,-0.4953493178,-0.280303359,-0.2522744238,0.1631973982,0.1406692266,-0.1088396013,-0.1343104988,-0.0160850268,0.1368656904,0.4740391672,-0.1410940289,-0.2202246934,0.1094303131,-0.0504105948,0.0339513905,-0.4868873358,0.1971952766,0.1035983264,0.3923437893,0.0333321542,-0.3985877931,-0.2281475812,0.2114372551,-0.0236336533,-0.0581281483,0.431316793,0.2363340706,-0.2228953391,-0.2575081289,-0.1625892967,-0.2158327103,-0.3757222593,-0.2422768623,0.4703900218,0.1956524402,-0.167727679,0.1122743413,0.0742414594,-0.1150673479,0.0617525205,0.3946146965,-0.0748204216,-0.1125263274,0.228046447,-0.0367226638,-0.3630042672,-0.1009183377,0.147413373,0.201251477,-0.0140398229,-0.5118097067,0.2428103238,-0.364155829,-0.1003240347,0.6593738198,0.0223975312,-0.0343961418,-0.1571909338,0.066836521,0.1328684986,-0.129877761,-0.1844610423,-0.4084619284,-0.1652341485,-0.1120061278,0.0725027546,0.0194830596,0.0134648941,-0.1131599545,0.1517076939,0.2277487367,-0.1084786206,-0.0292873103,-0.1077157259,-0.1577295959,0.2575592995,0.077276364,0.0432767384,-0.0455071479,-0.0621432848,-0.0667331219,0.094162479,-0.1415484548,-0.0810557082,-0.2618052661,0.3792394102,-0.3635045588,0.2026861459,-0.2679693103,-0.2763743401,0.0226156209,-0.5987124443,0.4693303406,0.0762628838,-0.2173783481,0.1839211583,-0.0067559402,0.4335826039,0.1400045753,0.1324246228,0.1296623051,0.0034162169,0.2564793527,-0.0129541969,0.0517177582,0.0345509052,-0.0483000986,-0.3150790632,0.1414650828,0.2187710553,-0.2970020473,-0.0528162047,0.0064465385,0.2238237113,0.1665391624,0.0170899089,-0.013207946,-0.3755169809,0.4678097367,0.1800196171,0.8736052513,0.2860046923,-0.0812116936,-0.2098022401,0.3937200904,-0.1191533208,-0.1728307009,-0.1444349438,0.1783284396,0.4140585661,-0.0591887087,0.2585327625,-0.0206662919,-0.1965887547,-0.1870826781,-0.3010665178,0.1405651867,0.1782164425,-0.3403763175,-0.0610642657,-0.1732201576,-0.4480157495,-0.0855614915,0.5458114743,-0.1134902462,0.1407276392,0.1678895503,-0.163473323,0.2742599249,0.1547592133,-0.1610339582,0.5066561103,0.0889239088,0.4276523292,0.1344052404,-0.0072405036,-0.1800989658,0.270796001,0.1540868878,0.0585375801,0.1390319467,0.1704631597,0.0915151015,-0.2852141559,0.0512629338,-0.384313643,-0.2931557,-0.0383235775,-0.1311173439,0.1286748946,-0.4033598006,-0.2206862718,-0.1276967674,0.1210237071,-0.0528600886,0.2827812433,-0.0695349947,-0.2162185758,-0.2116764337,0.262898773,0.5175610185,-0.0787364468,0.5037275553,-0.4499675035,-0.0716039613,-0.2682846189,0.1174156219,-0.4420118332,-0.3176285326,0.3655708134,-0.2907857001,-0.0914599374,-0.0759961829,-0.1990365237,0.3708024025,0.0023099952,-0.3136318922,0.1398541629,0.0063221497,0.0689287037,0.0739045367,0.1504748613,-0.6341795325,-0.0714748725,-0.2168478519,-0.0190161038,0.1884606332,-0.4838357568,-0.059614066,-0.3014622927,-0.2533233464,0.5574224591,-0.0429198705,0.0389955826,0.8306107521,-0.5347166061,0.0162806753,0.089945823,0.2708842456,0.0525015965,-0.0101972176,-0.0211191736,-0.1664485037,-0.0324043334,-0.1454047263,0.5164799094,-0.063463375,0.0934325904,-0.2375232875,-0.3187120557,0.0204699021,-0.0377339609,-0.1212196723,0.0364743285,0.6025368571,0.0533900931,-0.1724061519,-0.0387756303,-0.1411359906,0.1813826114,0.1829167157,0.2384239882,-0.0252660289,0.1747956872,0.1527781337,0.2597800791,0.1726745963,-0.5284258127,0.0049804514,0.0444116071,-0.1989776343,0.1114135459,-0.061512202,-0.2533995509,0.026897192,-0.0674713179,0.092611365,0.1822878569,-0.4473478198,-0.2816170752,-0.0643139258,-0.4277186394,-0.226035431,-0.2386196405,0.3652978241,0.0498350039,-0.3288972378,0.0844350979,0.1946460903,0.0669984147,0.5939670205,0.1519534439,0.1812232733,-0.1348908693,-0.1274597496,-0.0367073119,-0.4204012752,0.1611368507,-0.35095644,-0.2312136889,-0.3339869976,0.0014866377,-0.049428381,0.1004787162,0.4226041436,0.1340404451,-0.0104381163,0.0897459686,-0.3559645414,0.141196087,0.0403474346,-0.2209928185,-0.1935627609,0.731577158,0.3855091929,-0.4366104901,-0.2343100011,0.3028858602,-0.105660677,-0.0878251493,0.0988214463,-0.0412238613,-0.1869122684,0.1169939414,-0.0313217677,0.239899084,-0.0064875637,-0.1730325073,-0.2637092769,0.1085904762,-0.1694783717,0.0811264142,0.0226026084,0.2244395763,-0.106565021,-0.1305893809,0.1412334591,0.5329356194,-0.2432237864,0.148433432,0.6979504228,-0.0205304194,0.0702241138,0.1610849649,0.0437210724,0.2752022445,0.038240511,-0.034476418,-0.2715015411,0.3770997524,-0.0776867345,0.216234982,0.0598708652,-0.0110788392,-0.1531212777,0.5143990517,0.4952771664,0.2290320992,-0.3437059224,0.5232909918,0.2530077994,-0.599412322,-0.042360343,0.0932850689,1.2044788599,0.0724358559,0.2235270292,0.0314199328,-0.3536097407,0.40672791,0.1242623702,0.0078917043,-0.0750905573,-0.2230889201,-0.0305301584,0.0422625057,0.0691007897,-0.1389264464,-0.0713878796,0.132372424,0.4221555293,-0.0152443163,0.0837335736,0.402140379,0.3814328015,-0.2051498443,-0.0883516669,-0.0082328906,-0.0135498326,0.4488619864,-0.0580671728,-0.3236905932,0.0099247936,0.0581387244,-0.4257675409,0.0460798368,-0.2219448388,0.1018632427,0.1866083145,-0.4215055704,-0.1442057192,-0.0665630177,0.3502733111,0.1162467673,-0.2288773507,0.20555152,-0.6553243399,0.1337067932,-0.1111477688,-0.0223606639,0.3049784601,-0.1374572814,-0.2320296317,0.2556152642,0.2460429966,-0.0239813048,-0.3673838973,-0.3370626271,-0.1786551178,0.0387853682,0.2585060298,0.0523240045,-0.37593171,0.0676309466,0.0332308114,0.46895051,-0.2207185924,0.1279876679,0.0771204829,-0.1748057157,-0.1103821993,0.1873653382,0.3356001973,-0.0485889018,-0.0124244187,0.1048637927,0.0102064377,-0.034971159,-0.5182816982,0.0962407216,-0.4879133403,0.0929771513,0.1731506884,-0.0834302828,0.0521249063,0.4990685284,0.157914564,0.3171496093,-0.281981647,-0.025607137,-0.2516598701,0.2031749189,-0.0348430276,0.5519405007,0.378483057,0.2551999688,0.2363877147,-0.0135566816,-0.273529321,-0.2293474525,-0.1390469223,0.104696773,0.3563306928,-0.1656013727,-0.1046919376,-0.0334438421,-0.0603813343,0.4253439307,-0.0404491387,-0.015908625,0.112092793,0.149815321,0.1353460848,0.0684866831,-0.1953890026,-0.0157234091,0.2045608759,0.0757187009,0.171763733,-0.2325238436,0.2176932544,0.5747404099,0.3914261162,0.2458034158,0.2725383937,0.1071837172,0.3449466825,-0.0029915364,-0.1375877708,0.1985436529,-0.0066131069,-0.1798162609,-0.2169688791,0.3240689635,0.0533198342,-0.0288359337,0.2950301468,0.3801261783,-0.4418635666,-0.059984073,0.0650190264,-0.4060010016,0.0577241145,-0.2160560638,-0.2143867612,0.0848919824,-0.03001724,-0.0424114577,0.1603688151,0.0849685222,0.1243043244,0.1374167055,0.1534514129,0.0862917304,0.2262290269,-0.1230447739,0.2260157019,-0.0153785078,0.0753926635,-0.5670807362,-0.1740172952,-0.027114803,-0.000351358,-0.1932015717,0.2002460659,-0.1940700859,0.0959669873,0.1983021796,-0.1597925872,0.0894910172,-0.1692624092,0.4100303352,-0.1874136776,-0.0285619218,0.0504292883,0.4167961776,0.4355935156,-0.3474953771,0.0660831705,-0.133455351,-0.0092373369,0.4053116143,-0.0700920597,-0.0862438232,-0.2403082997,-0.1606181264,0.1411544979,-0.0841543302,-0.0286082607,0.0938247144,0.1536001712,-0.2205559313,0.1354841292,0.5547671914,-0.256906271,0.4653145969,0.530571878,-0.193166256,-0.1419741809,0.4079794884,0.1885657459,-0.5202434063,0.6899908185,0.0674935952,-0.190423578,-0.2995374501,0.4237954319,0.1942037195,0.0212338436,0.0459429882,0.0550730415,0.0988913625,-0.0965901315,0.0894840136,0.0906344578,-0.0860464945,-0.0521743223,-0.1964584589,0.1391919404,-0.7284783125,0.124876298,-0.231243819,-0.3695222437,-0.3456446528,0.0206570867,-0.1900036335,-0.1387187541,-0.0223027077,0.0365557484,0.015828969,-0.2200501114,0.0230328962,-0.0675529391,0.2398566157,0.1897879094,0.4737430811,-0.0260189176,-0.3410238922,0.1662025005,0.0969296917,0.3595590591,-0.0218514502,0.0608391464,0.0311376732,0.3090901971,0.2028237283,0.1055342779,0.6015223265,-0.0134021053,-0.0921448991,-0.3707886934,0.3012417853,-0.1956473738,-0.1594762951,0.1906780601,0.1231779307,0.106744267,-0.1550716907,0.069680728,-0.258523345,0.1412055194,-0.0900472924,0.5153779387,0.2933413982,0.28586483,-0.0991802812,-0.1435767859,-0.384188503,0.4304681122,-0.0727626458,-0.1152705923,-0.1751074642,-0.0556308366,0.2661163807,0.115607135,-0.0146457897,-0.0104153994,0.0005027702,0.0609776303,-0.1666046381,0.1279964447,-0.0895332992,-0.0738784298,-0.4382946193,0.0800397769,0.3747513592,-0.040132463,-0.1245538667,-0.2125291079,-0.40792799,0.322906673,0.052931346,-0.1005448774,-0.0648525655,0.177710399,0.0544739068,0.2989241183,-0.5815803409,-0.1861034632,0.1627860963,-0.1288931966,-0.0801475942,0.2587035298,0.2808337808,-0.3622457385,0.0584177673,0.2710856199,0.3739876151,-0.08818385,0.0602197424,-0.268589288]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2276","title":"concatenate_datasets loads all the data into memory","comments":"Therefore, when I try to concatenate larger datasets (5x 35GB data sets) I also get an out of memory error, since over 90GB of swap space was used at the time of the crash:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nMemoryError Traceback (most recent call last)\r\n in \r\n 20 print(file_name)\r\n 21 cv_batch = load_from_disk(file_name)\r\n---> 22 cv_sampled_train = concatenate_datasets([cv_sampled_train, cv_batch])\r\n 23 \r\n 24 print(\"Saving to disk!\")\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\arrow_dataset.py in concatenate_datasets(dsets, info, split, axis)\r\n 2891 \r\n 2892 # Concatenate tables\r\n-> 2893 table = concat_tables([dset._data for dset in dsets if len(dset._data) > 0], axis=axis)\r\n 2894 table = update_metadata_with_features(table, None)\r\n 2895 \r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in concat_tables(tables, axis)\r\n 837 if len(tables) == 1:\r\n 838 return tables[0]\r\n--> 839 return ConcatenationTable.from_tables(tables, axis=axis)\r\n 840 \r\n 841 \r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in from_tables(cls, tables, axis)\r\n 697 return result\r\n 698 \r\n--> 699 blocks = to_blocks(tables[0])\r\n 700 for table in tables[1:]:\r\n 701 table_blocks = to_blocks(table)\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in to_blocks(table)\r\n 669 return [[InMemoryTable(table)]]\r\n 670 elif isinstance(table, ConcatenationTable):\r\n--> 671 return copy.deepcopy(table.blocks)\r\n 672 else:\r\n 673 return [[table]]\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 151 copier = getattr(x, \"__deepcopy__\", None)\r\n 152 if copier is not None:\r\n--> 153 y = copier(memo)\r\n 154 else:\r\n 155 reductor = dispatch_table.get(cls)\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in __deepcopy__(self, memo)\r\n 143 # by adding it to the memo, self.table won't be copied\r\n 144 memo[id(self.table)] = self.table\r\n--> 145 return _deepcopy(self, memo)\r\n 146 \r\n 147 def __getstate__(self):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in _deepcopy(x, memo)\r\n 62 memo[id(x)] = result\r\n 63 for k, v in x.__dict__.items():\r\n---> 64 setattr(result, k, copy.deepcopy(v, memo))\r\n 65 return result\r\n 66 \r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 170 y = x\r\n 171 else:\r\n--> 172 y = _reconstruct(x, memo, *rv)\r\n 173 \r\n 174 # If is its own copy, don't memoize.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _reconstruct(x, memo, func, args, state, listiter, dictiter, deepcopy)\r\n 262 if deep and args:\r\n 263 args = (deepcopy(arg, memo) for arg in args)\r\n--> 264 y = func(*args)\r\n 265 if deep:\r\n 266 memo[id(x)] = y\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in (.0)\r\n 261 deep = memo is not None\r\n 262 if deep and args:\r\n--> 263 args = (deepcopy(arg, memo) for arg in args)\r\n 264 y = func(*args)\r\n 265 if deep:\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 170 y = x\r\n 171 else:\r\n--> 172 y = _reconstruct(x, memo, *rv)\r\n 173 \r\n 174 # If is its own copy, don't memoize.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _reconstruct(x, memo, func, args, state, listiter, dictiter, deepcopy)\r\n 262 if deep and args:\r\n 263 args = (deepcopy(arg, memo) for arg in args)\r\n--> 264 y = func(*args)\r\n 265 if deep:\r\n 266 memo[id(x)] = y\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in (.0)\r\n 261 deep = memo is not None\r\n 262 if deep and args:\r\n--> 263 args = (deepcopy(arg, memo) for arg in args)\r\n 264 y = func(*args)\r\n 265 if deep:\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_tuple(x, memo, deepcopy)\r\n 208 \r\n 209 def _deepcopy_tuple(x, memo, deepcopy=deepcopy):\r\n--> 210 y = [deepcopy(a, memo) for a in x]\r\n 211 # We're not going to put the tuple in the memo, but it's still important we\r\n 212 # check for it, in case the tuple contains recursive mutable structures.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in (.0)\r\n 208 \r\n 209 def _deepcopy_tuple(x, memo, deepcopy=deepcopy):\r\n--> 210 y = [deepcopy(a, memo) for a in x]\r\n 211 # We're not going to put the tuple in the memo, but it's still important we\r\n 212 # check for it, in case the tuple contains recursive mutable structures.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_tuple(x, memo, deepcopy)\r\n 208 \r\n 209 def _deepcopy_tuple(x, memo, deepcopy=deepcopy):\r\n--> 210 y = [deepcopy(a, memo) for a in x]\r\n 211 # We're not going to put the tuple in the memo, but it's still important we\r\n 212 # check for it, in case the tuple contains recursive mutable structures.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in (.0)\r\n 208 \r\n 209 def _deepcopy_tuple(x, memo, deepcopy=deepcopy):\r\n--> 210 y = [deepcopy(a, memo) for a in x]\r\n 211 # We're not going to put the tuple in the memo, but it's still important we\r\n 212 # check for it, in case the tuple contains recursive mutable structures.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 159 reductor = getattr(x, \"__reduce_ex__\", None)\r\n 160 if reductor is not None:\r\n--> 161 rv = reductor(4)\r\n 162 else:\r\n 163 reductor = getattr(x, \"__reduce__\", None)\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\pyarrow\\io.pxi in pyarrow.lib.Buffer.__reduce_ex__()\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\pyarrow\\io.pxi in pyarrow.lib.Buffer.to_pybytes()\r\n\r\nMemoryError: \r\n\r\n```","body":"## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n","comment_length":1031,"text":"concatenate_datasets loads all the data into memory\n## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n\nTherefore, when I try to concatenate larger datasets (5x 35GB data sets) I also get an out of memory error, since over 90GB of swap space was used at the time of the crash:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nMemoryError Traceback (most recent call last)\r\n in \r\n 20 print(file_name)\r\n 21 cv_batch = load_from_disk(file_name)\r\n---> 22 cv_sampled_train = concatenate_datasets([cv_sampled_train, cv_batch])\r\n 23 \r\n 24 print(\"Saving to disk!\")\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\arrow_dataset.py in concatenate_datasets(dsets, info, split, axis)\r\n 2891 \r\n 2892 # Concatenate tables\r\n-> 2893 table = concat_tables([dset._data for dset in dsets if len(dset._data) > 0], axis=axis)\r\n 2894 table = update_metadata_with_features(table, None)\r\n 2895 \r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in concat_tables(tables, axis)\r\n 837 if len(tables) == 1:\r\n 838 return tables[0]\r\n--> 839 return ConcatenationTable.from_tables(tables, axis=axis)\r\n 840 \r\n 841 \r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in from_tables(cls, tables, axis)\r\n 697 return result\r\n 698 \r\n--> 699 blocks = to_blocks(tables[0])\r\n 700 for table in tables[1:]:\r\n 701 table_blocks = to_blocks(table)\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in to_blocks(table)\r\n 669 return [[InMemoryTable(table)]]\r\n 670 elif isinstance(table, ConcatenationTable):\r\n--> 671 return copy.deepcopy(table.blocks)\r\n 672 else:\r\n 673 return [[table]]\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 151 copier = getattr(x, \"__deepcopy__\", None)\r\n 152 if copier is not None:\r\n--> 153 y = copier(memo)\r\n 154 else:\r\n 155 reductor = dispatch_table.get(cls)\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in __deepcopy__(self, memo)\r\n 143 # by adding it to the memo, self.table won't be copied\r\n 144 memo[id(self.table)] = self.table\r\n--> 145 return _deepcopy(self, memo)\r\n 146 \r\n 147 def __getstate__(self):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\datasets\\table.py in _deepcopy(x, memo)\r\n 62 memo[id(x)] = result\r\n 63 for k, v in x.__dict__.items():\r\n---> 64 setattr(result, k, copy.deepcopy(v, memo))\r\n 65 return result\r\n 66 \r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 170 y = x\r\n 171 else:\r\n--> 172 y = _reconstruct(x, memo, *rv)\r\n 173 \r\n 174 # If is its own copy, don't memoize.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _reconstruct(x, memo, func, args, state, listiter, dictiter, deepcopy)\r\n 262 if deep and args:\r\n 263 args = (deepcopy(arg, memo) for arg in args)\r\n--> 264 y = func(*args)\r\n 265 if deep:\r\n 266 memo[id(x)] = y\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in (.0)\r\n 261 deep = memo is not None\r\n 262 if deep and args:\r\n--> 263 args = (deepcopy(arg, memo) for arg in args)\r\n 264 y = func(*args)\r\n 265 if deep:\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 170 y = x\r\n 171 else:\r\n--> 172 y = _reconstruct(x, memo, *rv)\r\n 173 \r\n 174 # If is its own copy, don't memoize.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _reconstruct(x, memo, func, args, state, listiter, dictiter, deepcopy)\r\n 262 if deep and args:\r\n 263 args = (deepcopy(arg, memo) for arg in args)\r\n--> 264 y = func(*args)\r\n 265 if deep:\r\n 266 memo[id(x)] = y\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in (.0)\r\n 261 deep = memo is not None\r\n 262 if deep and args:\r\n--> 263 args = (deepcopy(arg, memo) for arg in args)\r\n 264 y = func(*args)\r\n 265 if deep:\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_tuple(x, memo, deepcopy)\r\n 208 \r\n 209 def _deepcopy_tuple(x, memo, deepcopy=deepcopy):\r\n--> 210 y = [deepcopy(a, memo) for a in x]\r\n 211 # We're not going to put the tuple in the memo, but it's still important we\r\n 212 # check for it, in case the tuple contains recursive mutable structures.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in (.0)\r\n 208 \r\n 209 def _deepcopy_tuple(x, memo, deepcopy=deepcopy):\r\n--> 210 y = [deepcopy(a, memo) for a in x]\r\n 211 # We're not going to put the tuple in the memo, but it's still important we\r\n 212 # check for it, in case the tuple contains recursive mutable structures.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_tuple(x, memo, deepcopy)\r\n 208 \r\n 209 def _deepcopy_tuple(x, memo, deepcopy=deepcopy):\r\n--> 210 y = [deepcopy(a, memo) for a in x]\r\n 211 # We're not going to put the tuple in the memo, but it's still important we\r\n 212 # check for it, in case the tuple contains recursive mutable structures.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in (.0)\r\n 208 \r\n 209 def _deepcopy_tuple(x, memo, deepcopy=deepcopy):\r\n--> 210 y = [deepcopy(a, memo) for a in x]\r\n 211 # We're not going to put the tuple in the memo, but it's still important we\r\n 212 # check for it, in case the tuple contains recursive mutable structures.\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 144 copier = _deepcopy_dispatch.get(cls)\r\n 145 if copier is not None:\r\n--> 146 y = copier(x, memo)\r\n 147 else:\r\n 148 if issubclass(cls, type):\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in _deepcopy_list(x, memo, deepcopy)\r\n 203 append = y.append\r\n 204 for a in x:\r\n--> 205 append(deepcopy(a, memo))\r\n 206 return y\r\n 207 d[list] = _deepcopy_list\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\copy.py in deepcopy(x, memo, _nil)\r\n 159 reductor = getattr(x, \"__reduce_ex__\", None)\r\n 160 if reductor is not None:\r\n--> 161 rv = reductor(4)\r\n 162 else:\r\n 163 reductor = getattr(x, \"__reduce__\", None)\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\pyarrow\\io.pxi in pyarrow.lib.Buffer.__reduce_ex__()\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\pyarrow\\io.pxi in pyarrow.lib.Buffer.to_pybytes()\r\n\r\nMemoryError: \r\n\r\n```","embeddings":[-0.2084973752,-0.1142902002,0.0484714955,0.4346452057,0.199329704,0.1706529409,-0.1373960376,0.2591967583,-0.1694815457,0.0429848805,0.0269325897,0.2109090835,0.0059097265,-0.2283868343,-0.1217163354,0.1358806938,0.2059148848,0.0483620204,-0.2932789326,-0.0278221704,-0.2548550665,0.2438369095,-0.3456242681,-0.3321818709,-0.3617053926,0.268399775,-0.288836211,0.4278264642,0.0539543182,0.0645484328,0.3253515661,-0.0345083736,0.0323292613,0.4721652567,-0.0001129602,-0.0248592626,0.1940784007,-0.1648728251,-0.3701837957,0.0790072531,-0.257725656,-0.4280099869,-0.1000079662,0.0440540612,0.4190388918,0.033717785,-0.1139069051,-0.7214249969,0.0311022121,0.1015497446,0.2013147473,0.0450461656,0.3114587069,-0.1664170623,-0.0396206118,0.0367016122,-0.1191298142,0.3378038704,0.1667080969,-0.2502472401,0.0814066082,0.0372247323,-0.2164464295,-0.1925283819,0.0617970489,0.2860530615,-0.2610839307,-0.4039141536,0.024543019,0.029022174,0.2679313123,-0.410095185,-0.042093657,-0.4409756958,0.036623124,-0.1752433777,0.2727940977,0.4027750194,-0.2462693602,0.293792367,-0.0894597396,0.0641691759,0.0416414775,0.0172061753,-0.2132966369,0.1253425181,-0.0891545489,0.0814755037,0.4102660716,-0.0002172078,0.29209885,-0.3168378174,-0.3524680436,0.0377512462,-0.5595110655,-0.0498269126,0.0755836442,-0.4589984715,0.2216620147,0.1062295064,0.1199405491,0.0183505844,0.212127164,-0.0187555477,0.3851090074,0.2461845875,-0.0689737573,0.4001848698,0.0538023338,-0.1622482836,0.302623421,0.0824345574,-0.029857846,-0.2151603252,0.2167406082,-0.2746641338,0.2562694252,-0.0793108121,-0.4918817282,0.2032411695,-0.1662214547,0.1221663281,-0.0328876488,0.1849793792,0.0313825198,0.5110003352,0.1733138263,0.1208142936,-0.0608106479,0.1380167902,-0.1753146946,-0.1013809815,-0.2131145895,0.3016291857,0.2818205357,-0.0465134867,0.0303951073,0.2547302544,0.1015235111,-0.2291039824,0.0861584991,-0.3414562643,0.0732797608,0.1529537141,0.0300752893,0.3761548698,0.1255546063,-0.0855323374,-0.0447999351,0.3257975876,-0.312535882,-0.1928355694,-0.2904739082,0.1899365634,0.0848515853,0.1171091199,-0.2278673947,0.2889300883,0.3921437562,0.2276145965,-0.0246095154,-0.0160901472,0.1302628517,-0.3428185284,0.3278464675,0.4563962221,-0.1643471569,0.2442180216,0.1593949497,0.2009991705,0.5149561167,0.3146007061,-0.0787222609,0.0898945928,-0.4410283864,-0.0834654197,0.1891031414,-0.1437196136,-0.263667345,0.2389305979,-0.250733614,0.3553431928,-0.0772488788,0.1755164564,-0.0911854059,0.1544353962,0.7130596638,0.4093846977,0.0317519046,0.1562935859,-0.3548541367,-0.0897677019,0.1346036643,-0.0315746814,-0.0920929909,0.2037222236,-0.0326738805,0.101136595,0.4664764702,-0.1077108309,0.153782919,0.4310880899,0.0536223911,-0.0714828223,-0.2064581066,0.0748017207,-0.5007076859,0.0443876721,0.0569930151,-0.2531939447,-0.0204966217,-0.0257828552,-0.0208263192,-0.0269897208,0.0041292487,0.0003340359,0.0507822856,0.2245372832,-0.1111161187,-0.05244793,-0.0241559371,0.8368053436,-0.1254532039,0.0507056601,-0.4194497764,0.1712570935,-0.0288675167,-0.2300039381,-0.0660286769,0.0523214936,0.1158133075,0.0188080091,-0.1096438766,0.5126361251,-0.1974750161,-0.052632004,-0.2070373297,-0.255222857,0.0165734403,-0.0363014676,0.0188739039,-0.2550829351,0.0981568024,-0.147838518,-0.0155783817,0.3320586979,-0.0144178234,0.0340214893,0.0131641245,-0.0984588116,-0.0694328323,-0.1997836083,0.1187121272,-0.187878564,0.1486685127,0.2488634884,0.2260706872,0.2236911505,-0.4840830564,-0.0810607523,0.2194009572,0.0873406306,0.2150361985,0.114058435,-0.3331508636,-0.1729366928,-0.0043807761,0.1619129777,0.7703347206,0.158680141,0.2190234959,0.027602585,-0.0670482293,-0.2215753347,0.2817302942,-0.0253939908,0.0702340901,0.2413841486,0.2463647574,-0.004090522,-0.2063647807,0.1619835943,0.1250704378,0.006698038,-0.2925804555,0.1093799099,-0.2098116875,0.0715502128,-0.3482342362,-0.3516216278,-0.0114572477,-0.4532111287,-0.3251947463,0.395349741,-0.0905613005,-0.1436620951,0.0558752455,0.2527936399,-0.0280837454,0.0283187088,0.0868103728,0.2992718816,-0.3061571419,0.0406388193,0.0782498419,-0.2633136213,0.0490673073,0.1617473811,-0.1413246542,0.0032180639,-0.0775131136,-0.0589703247,-0.0648287833,0.002731751,-0.1926749498,0.1106752828,0.0238208659,-0.3615097702,0.0093959169,0.4291680455,0.0770543218,0.4083506465,0.2699891925,-0.1056147888,-0.3779594004,-0.1318799257,-0.3104640245,-0.3968418539,0.0882834047,-0.0485187098,0.0247029588,0.3051521182,0.1884802282,-0.0642439499,0.0278732907,0.0880296603,-0.1925894171,-0.2163109332,0.4186181128,-0.0596594103,-0.2062426209,-0.1150502935,0.1457549334,-0.2462955415,0.2110211849,-0.4671842754,0.1191657484,-0.5144683123,0.036162667,-0.2598979175,0.0987710059,0.0204598419,0.0733611211,-0.07519418,-0.087361522,-0.0732144713,-0.1453312933,-0.0290519334,0.3612685502,-0.2521463335,-0.0733989552,0.2915198207,-0.0696369186,0.4133873582,0.0289635099,0.3018461764,0.1516514271,0.5536729693,-0.2386159748,-0.4298308492,-0.122861959,-0.1587635279,-0.0061717248,-0.1209721193,-0.0357560627,-0.1237010583,-0.1062171757,0.2274902016,0.1769467145,-0.1957773268,0.2441173196,-0.2653460801,-0.0477347709,-0.059861239,0.0071465205,-0.0704567209,-0.0330448933,-0.1874562204,0.0097139329,0.0092439549,-0.1222308725,-0.2583731413,0.0481114388,-0.6078559756,0.0813036561,-0.1773091108,0.5477053523,0.1794641465,-0.3359860778,-0.0903060958,0.0117468294,0.7129833102,0.3859626651,-0.4184196889,0.2137069702,0.1342884004,-0.508459568,-0.1242008954,-0.1015698165,0.2427203059,0.1661406308,0.2337725461,-0.1978385299,0.0345439278,0.139410764,0.3654304445,-0.1443135142,-0.0190237183,-0.2345568985,-0.2062204629,-0.1828336567,-0.2722642124,0.1025900245,0.3259936273,-0.2024847418,-0.3330004811,-0.0266627241,-0.1827154458,0.2898464501,-0.1630952954,0.4102861881,-0.001857593,0.3690981567,-0.2302062064,0.2487228215,0.3287046552,0.5444668531,0.1445554197,-0.1886040568,0.2795898914,0.0070743593,0.3126012087,0.2015107274,-0.1283898056,0.0842770636,-0.2613304853,0.1809066683,-0.5846990943,0.1711177528,0.0644457564,0.0806887522,-0.6066932082,-0.1957490295,0.2711178064,0.1606961936,-0.0131714921,0.5433198214,-0.4463484287,-0.4786521196,0.1642616093,-0.1519677639,0.8347200155,0.0207112469,0.2568376362,0.0044035409,-0.0597236678,0.2216639519,0.0106041245,0.2319178581,-0.1823973805,-0.2544629574,0.1474297792,-0.2599653304,0.0836360008,-0.2718308866,-0.4256110489,0.1060700491,-0.3712700307,-0.022935044,-0.2223657668,0.0113987736,-0.2986971438,-0.3575039506,-0.0757511482,0.1579180807,0.2478518784,-0.0842900276,0.0014606151,-0.0135904988,-0.1467341632,-0.1307817101,-0.3107578754,0.3131834269,-0.1598427147,0.4684114754,-0.2722895443,-0.1829643548,-0.1438465863,0.1320194006,0.2478927672,0.3407207727,0.0169652775,-0.0956034511,-0.1326029152,0.2782518268,0.0808181018,-0.2257166803,0.0548750013,0.3345298469,-0.1280462891,0.1026321873,-0.0138710132,-0.2697596252,-0.2939833701,0.285615474,0.0197919011,-0.2719836235,-0.1172960401,0.275182724,0.2105696648,-0.1763989776,0.1346274018,0.2218135148,-0.1078091413,0.2202524394,-0.42332533,-0.3826931715,-0.1413326412,0.3322012424,0.2150717229,-0.0476863086,0.2619226277,0.0064737303,0.037330512,-0.1044234186,0.2194757164,0.021449374,-0.1695188582,0.2213663161,-0.1948362887,-0.0924790353,0.0339002721,0.2063981295,0.1196031645,-0.0841680691,-0.0830367357,-0.1579917669,-0.4656386673,0.077698566,0.0747643486,0.352053225,-0.0351995528,0.0434461758,-0.119748801,0.1185342893,-0.2961999774,-0.0317852087,-0.2356138378,0.177330032,-0.1427203566,0.0333633386,-0.0078617241,-0.0961168706,0.1403389871,0.2220603824,0.087529622,-0.201487124,0.081839554,0.0819422975,0.0577591695,-0.1826287061,-0.2424244881,-0.4744129777,0.0761530846,-0.3522908986,0.2711901665,0.2841579616,-0.0429666601,-0.0853778198,0.0960474387,0.0166845378,0.0894248933,0.0766519085,-0.0908352807,0.055822555,-0.0828101411,0.2883355021,-0.0166554023,-0.0094453935,-0.5261986256,0.1355681568,0.0457160659,0.0580819808,0.0829283744,-0.2150712162,-0.2936805189,-0.0760245845,0.0833572075,0.4004903734,-0.198356241,-0.0106341718,-0.0482330769,0.188016966,0.2002944052,-0.3394595683,-0.1354779303,-0.054134015,-0.075500913,0.1981995404,0.1597377807,-0.3077151179,-0.2801880836,-0.0657666102,0.5569904447,0.2812485397,0.3075849116,0.2442364395,0.2052915096,0.1028156802,0.1233467907,0.0502338931,0.2296522409,0.3803371191,-0.0192458276,0.2851464152,0.1478357166,-0.0647624433,-0.1967901736,-0.4574253857,0.2853144705,0.4371398985,-0.2440536022,0.3275417387,0.0515150577,-0.0573755801,-0.4356239438,-0.2360715866,-0.045751255,0.1197241321,-0.3212194145,-0.0957368389,0.2676940858,-0.085853599,0.0629815236,0.1351405531,0.1661013216,-0.3355976045,0.4730648696,0.0138576394,-0.2500505447,-0.1553128958,0.2223838121,0.503497541,0.0119024562,-0.3562343121,0.0451845713,0.1132243872,-0.0502090007,-0.0544717014,0.4576027095,0.5256034136,0.5564421415,0.2286735326,0.2088666707,0.3797374964,-0.0563464388,-0.301923573,0.2601626217,-0.0157608539,-0.0937593579,0.0445714071,0.1994582862,-0.1327547729,0.0180627219,0.4137890339,0.1692031175,-0.3755024374,0.0050654048,-0.0561922714,-0.2994552553,0.071939677,0.1382433027,-0.2935512662,0.0199723486,0.4786643386,-0.1998849064,0.1544065028,0.087119773,0.0902591422,0.307133764,0.3065686822,0.2413416505,-0.0728603974,-0.3713642657,-0.3787606657,-0.2045230865,0.2920801044,0.0131600825,0.2769449055,-0.0518896952,0.3205883801,0.0929842889,-0.0158902146,-0.156793505,0.2470312119,-0.1267678291,0.0205865763,-0.530344367,-0.170669198,0.2856861353,-0.0350447521,-0.0198058244,-0.4125314355,0.413952738,0.0030132688,0.088518694,-0.1432080716,0.2161065489,0.1710315049,0.2663677335,0.4304056168,0.2482922524,0.1741651893,-0.1628273576,-0.3097068965,-0.2974675298,-0.3777626753,0.0783981755,-0.1954627335,0.0234466679,0.0784439892,-0.378097415,0.0148945497,-0.1069144085,0.0385914817,0.2003040761,0.1180154532,-0.1220133305,0.0405565724,0.0305430796,0.140604198,0.1762432307,0.3873684406,-0.080844976,0.3279305696,-0.0769979581,-0.2976796627,-0.0052264887,0.0882843435,-0.2614132166,-0.2003346086,0.205604732,-0.0715865046,0.123528555,-0.5712339282,-0.1176659539,0.3252485693,-0.1334315538,-0.4380504787,0.4799673557,-0.1831926703,-0.0903029144,-0.013281811,0.4617712796,0.0376565307,-0.4231987596,0.3124876916,-0.177028209]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2276","title":"concatenate_datasets loads all the data into memory","comments":"Hi ! this looks like an important issue. Let me try to reproduce this.\r\nCc @samsontmr this might be related to the memory issue you have in #2134 ","body":"## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n","comment_length":28,"text":"concatenate_datasets loads all the data into memory\n## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n\nHi ! this looks like an important issue. Let me try to reproduce this.\r\nCc @samsontmr this might be related to the memory issue you have in #2134 ","embeddings":[-0.1655793339,0.0086438237,0.025085656,0.3855163455,0.2144166231,0.1329604685,0.0000971354,0.2391076833,-0.2387411892,0.0435207263,0.1007450595,0.1849094927,0.1349058151,-0.1963892877,-0.1764700562,0.2529346645,0.2075009346,0.1770739257,-0.3354966044,-0.0526397601,-0.2628764212,0.2321233749,-0.3049420416,-0.4224450588,-0.4580206871,0.2613813579,-0.3565160334,0.341798991,0.0743395239,-0.0279053077,0.3124286532,0.0880531073,0.0267186351,0.4711241126,-0.0001092321,-0.0790268779,0.223182857,-0.1731623858,-0.3741953671,-0.0006446507,-0.206647858,-0.4626248479,0.0248347633,-0.0448434614,0.3222416937,0.0223689992,-0.1336586475,-0.5861000419,0.0342048481,-0.0428405628,0.2663333714,-0.0470443964,0.1371284276,-0.0761274025,-0.0676253289,0.1555089206,-0.0315027833,0.193926543,0.074843064,-0.0581822544,0.1401505172,0.1468088478,-0.2224315256,-0.0578752197,0.0761224627,0.3823084235,-0.1556165069,-0.3912760019,0.1003550813,0.013922058,0.3194790483,-0.3667294383,-0.1034772471,-0.3596833348,0.0806934461,-0.1792730987,0.2723421454,0.4094532728,-0.1917519122,0.1821502,-0.0955394208,0.0062458869,-0.0361389369,-0.0321247801,-0.1807833016,0.1143901795,-0.1381177902,0.0591760613,0.2896771133,0.0071404162,0.2253641188,-0.4683856964,-0.2991128564,0.1013631374,-0.4785361886,-0.0062481333,-0.0034642192,-0.3938516378,0.1206180453,0.1632405818,0.0700611696,0.0221848357,0.233151868,0.0234280191,0.3330491781,0.1919004023,-0.0601132661,0.3533694148,0.1468001902,-0.1694046706,0.1771907061,0.0226119347,-0.1276815534,-0.2536406517,0.2378406227,-0.2868468761,0.22037597,-0.141863212,-0.4976771474,0.1240119934,-0.1110918671,0.067440249,-0.0226403922,0.2313437909,-0.0302149653,0.468762815,0.225509122,0.2407607287,-0.0569588691,0.19360511,-0.1889927983,-0.1315045953,-0.1324183047,0.2119251192,0.2942709327,-0.2170787156,-0.0015042638,0.3486470282,0.0658855066,-0.158025533,0.0930391476,-0.3391183317,0.1912353784,0.152496323,0.1535510719,0.3738382757,0.0881197229,-0.1477621347,-0.0577585511,0.4209817052,-0.2102910131,-0.1051318422,-0.2599584162,0.2226831019,0.0764071643,0.1525421441,-0.1444146186,0.2782827914,0.3907812834,0.2938722372,-0.1349306852,0.0010443953,-0.0442564189,-0.4204100668,0.4669266343,0.4033464491,-0.1331076026,0.2231316715,0.0950644165,0.2323825359,0.5047422647,0.3542568386,-0.0591240339,0.0130126346,-0.3014681935,0.0535793453,0.165953964,-0.1710008979,-0.1954697371,0.2662730813,-0.2416405976,0.3232174516,-0.0223679878,0.1308974028,-0.0406770594,0.115315333,0.5678358674,0.4280857742,0.1203774065,0.1059351116,-0.3661514819,-0.1243509054,0.1430986077,-0.148582831,-0.2223394215,0.2367296219,-0.0745880604,-0.1166882887,0.4902270734,-0.1383096278,0.229373157,0.3590711057,0.2224223614,-0.0895943269,-0.2062208205,0.1898645014,-0.4102156758,0.0116716828,0.1662635952,-0.2398364544,-0.0783679038,-0.094877623,0.0968281999,0.0360408165,-0.0288469698,-0.0119433971,0.0789150894,0.2847230434,-0.0222637076,-0.0659041554,0.0368670821,0.832049489,-0.0992904529,-0.0373105071,-0.2667585313,0.1807724684,-0.0037860416,-0.2975318432,-0.0234784819,0.1025625765,0.2381665558,0.0583726317,-0.0710332096,0.4748388827,-0.2936935723,-0.0626462176,-0.1457628757,-0.2413434833,-0.0273022801,0.0979812965,0.0734949335,-0.1472253054,0.1992408633,-0.1914403141,0.0033334445,0.3507956862,0.1001457423,0.1305978745,0.0190079231,-0.1095671505,0.0522732697,-0.1929912567,0.1165606454,-0.2192910612,0.1902213991,0.2370195985,0.1897492558,0.2867828906,-0.4739089906,0.0853219479,0.2243829519,0.0859741122,0.134588331,0.1528250128,-0.3762894571,-0.1647811234,-0.118295379,0.2007670254,0.7927389741,0.1822930276,0.1886880547,-0.0461789109,-0.0195291508,-0.2375982851,0.3301806152,-0.0596759357,0.0493188575,0.2444156408,0.263857156,-0.0619980805,-0.1087931916,0.2657482624,0.0917521417,-0.0722228587,-0.2444089055,0.0349161737,-0.2231038809,0.1363499761,-0.3180537224,-0.2835676074,-0.0104357926,-0.3355413973,-0.3896481097,0.4484146237,-0.0530083776,-0.1751144677,0.0252497699,0.2732765079,-0.0759382099,0.01656227,-0.0371091701,0.2551343143,-0.4038929343,0.051539015,0.0634798408,-0.2436132729,0.0719503239,0.1281689554,-0.0348507985,-0.0180502981,-0.0568420403,-0.1370334029,-0.0814142302,0.1027673185,-0.2358754277,0.1462717056,0.0238742474,-0.3808460832,0.0596417561,0.3619405329,0.0492811613,0.3570196033,0.311745584,-0.2407876849,-0.3567046821,-0.1098106131,-0.3115668893,-0.3086031675,0.0800438076,0.0158426128,0.1094315872,0.1882324517,0.0879658759,-0.108084932,0.1119806617,0.1333832145,-0.2219352573,-0.2804148495,0.5735882521,-0.1309706718,-0.3618882895,-0.0370361879,0.0512450486,-0.1881371886,0.1929485947,-0.5059261918,0.1453736126,-0.4991856515,0.0533202402,-0.2712647915,0.1151775867,-0.0531395935,0.1162872836,-0.1049734801,-0.1924835593,-0.1470502615,0.0249408055,0.0444609448,0.2594724596,-0.2077739388,-0.1047967151,0.1587515473,-0.1050413921,0.5049679875,-0.0508894585,0.3199081421,0.2215720713,0.4598450363,-0.2015454173,-0.4354869723,-0.1995421499,-0.2189509273,0.0760166794,0.0926732868,0.0842148215,-0.1030303463,-0.0882743448,0.188662827,0.1009895429,-0.2592321336,0.1957737058,-0.1638156921,0.0225308258,-0.0280095432,-0.0647343546,-0.1412207931,-0.0814821124,-0.2522289753,0.0069883112,0.0326024704,-0.0900954977,-0.2615724504,0.0022692308,-0.6186498404,0.1515329629,-0.0837033913,0.4394445717,0.1057544649,-0.3408277631,0.0717101097,-0.0512027554,0.7416384816,0.3063717186,-0.4622938335,0.2142731994,0.1128516346,-0.4659021795,-0.0900797397,-0.0442951098,0.1861101389,0.1753768325,0.1690665185,-0.2508674562,-0.0587699674,0.0620602556,0.3879022598,-0.1628969163,-0.0929847211,-0.3386979997,-0.201175943,-0.1498248577,-0.3379560113,0.0504356958,0.4113532305,-0.2342184931,-0.2840871215,-0.0269441567,-0.1616677195,0.2719671428,-0.2362777591,0.5309255719,0.0917564183,0.403606087,-0.0857964978,0.2549717426,0.2840248644,0.5802751184,0.0843181089,-0.0551957637,0.2512030303,-0.2015672177,0.3931274712,0.2361795306,-0.1646124572,0.2078143805,-0.3127156794,0.151119709,-0.6362518072,0.136729911,0.1120366752,0.1204809844,-0.614120841,-0.256608814,0.2589708269,0.0819217116,-0.0361666828,0.558214426,-0.5020986795,-0.3950890303,0.0777554438,-0.1777589172,0.8481449485,0.0699067786,0.2565594018,0.0027996053,-0.2009338588,0.1584954709,0.1039952263,0.2797349989,-0.1751070917,-0.1814471632,0.1522355527,-0.2464660853,0.069038555,-0.1220559329,-0.4632975161,0.0091000479,-0.5243902206,-0.0217028167,-0.1231712401,0.1032260582,-0.2120978683,-0.3028658926,-0.0358666405,0.2050860077,0.3308335543,-0.0084452843,0.0268148221,-0.0175384264,-0.2335453033,-0.1823991835,-0.1997535378,0.3178455234,-0.1743880063,0.4426388144,-0.2106556743,-0.3191386461,-0.103240788,0.095793508,0.3153251708,0.2377961427,-0.0095438994,-0.0968259051,-0.0856418684,0.3278057277,0.0898515061,-0.2695420384,0.0056702425,0.2372932881,-0.1189018041,-0.0290074963,-0.0130606564,-0.1951521337,-0.3066833317,0.2422951013,0.0326611027,-0.3036009371,-0.0849661306,0.3106265068,0.2804531753,-0.1517194062,0.1858512163,0.2436746955,-0.1264245808,0.2694484591,-0.3472909927,-0.3536490202,-0.1472583562,0.354677707,0.3105773926,-0.0207007378,0.3795063198,-0.0259320978,-0.046152778,-0.139271751,0.1437416077,-0.0492220074,-0.1437403411,0.2533928752,-0.0723611265,-0.0245765708,-0.0736332312,0.2642334402,-0.0107502379,-0.0522374101,-0.0499433689,-0.2327727973,-0.3944982588,-0.0881241336,0.0223935191,0.2742875218,-0.025922887,0.0368151143,-0.096118249,0.2236864865,-0.3511759341,-0.1262113601,-0.187538147,0.2015400827,-0.0892664939,-0.0035379468,0.1460658014,0.0338843949,0.1957268268,0.1709041446,0.0599287599,-0.249528423,0.0352633633,0.0864397734,-0.0465427116,-0.2119366825,-0.2039418817,-0.3957461715,0.0185144525,-0.3560476601,0.2193858325,0.2203253508,-0.0527057312,-0.1323361397,-0.0683471113,0.0547501966,0.0508900173,0.0306921974,-0.1557452232,0.0050947252,-0.0430691838,0.1874603331,-0.0439723656,-0.0179947596,-0.4997471571,0.0306652617,0.0877424404,-0.0207827706,-0.0201893244,-0.1921583563,-0.2383553833,-0.1208875552,0.131478712,0.3539880514,-0.2005269229,0.0096165165,-0.0026400869,0.2471840978,0.0900178179,-0.2926798463,-0.2121765614,-0.1438190043,-0.1670495123,0.1733814627,0.1612119377,-0.2851781249,-0.262860626,-0.0674715415,0.5141931176,0.1537663192,0.2914773822,0.186492756,0.2115906328,0.1594833434,0.0799725652,0.145088315,0.0747695789,0.2733060718,-0.0731557906,0.2683878541,0.2369282991,-0.1638639271,-0.1892028153,-0.4537916481,0.286906004,0.4056817591,-0.133587718,0.2038646936,-0.0452832244,0.0616544746,-0.395190686,-0.2849051356,-0.1014392823,0.1666636169,-0.3034034967,-0.1454534978,0.1830058992,-0.0824944004,0.0329313353,0.1180342808,0.192132324,-0.2573126256,0.4321693778,0.1411121935,-0.1275139302,-0.2316100597,0.293130815,0.3955428898,0.0570614561,-0.325753808,0.0650611669,0.092005305,0.0739112049,0.0520133376,0.3636522293,0.5503244996,0.6578983665,0.2299029976,0.1052021906,0.2951344848,-0.0638787076,-0.2650016248,0.2549946308,-0.0617018752,-0.0989862531,0.1175218374,0.2143211812,-0.1710015982,0.0013895237,0.3623443246,0.1524405181,-0.382571429,0.1151847243,-0.0261548683,-0.2708133161,0.0641742796,0.0035999911,-0.2895315289,-0.0531748533,0.5402069092,-0.2178838402,0.1411162168,0.0234850552,0.1104842275,0.166243732,0.3059573174,0.0939162374,-0.0284780748,-0.4043636918,-0.3820244968,-0.2501634061,0.2439369708,0.0739299878,0.2556680441,-0.124522455,0.3312850595,-0.0306098107,-0.0976012498,-0.1264604628,0.1275250167,-0.1415269673,-0.0024958712,-0.5370720625,-0.0231017545,0.2088389695,0.0114367157,-0.0010783328,-0.3491337895,0.3652962744,-0.1066995114,0.1608670503,-0.1346103251,0.227335453,0.1830498576,0.3793092966,0.4042256474,0.2506742775,0.2433475107,-0.2374553978,-0.2472519875,-0.2280054092,-0.2799972594,-0.009912462,-0.1266322434,0.0126282591,0.0822140276,-0.3894753158,0.0802964866,-0.1074458435,0.0438102894,0.1785130501,0.1614345461,-0.0175018162,0.0740023628,0.0483424328,0.1693348587,0.2136311829,0.248434037,-0.0600391105,0.2875373363,-0.1768734902,-0.3219001293,0.1524873972,0.0063738967,-0.2955347896,-0.1489922851,0.312076807,-0.0680463314,0.158091411,-0.63853544,-0.1486531794,0.3362166286,-0.1242278218,-0.4552138448,0.4964091778,-0.159198314,-0.116471,-0.0237516239,0.3737540841,0.0552553758,-0.4150145352,0.2294659466,-0.2355408072]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2276","title":"concatenate_datasets loads all the data into memory","comments":"@lhoestq Just went to open a similar issue.\r\n\r\nIt seems like deep copying (tested on master) the dataset object writes the table's record batches (`dset._data._batches`) into RAM.\r\n\r\nTo find the bug, I modified the `_deepcopy` function in `table.py` as follows:\r\n```python\r\ndef _deepcopy(x, memo: dict):\r\n \"\"\"deepcopy a regular class instance\"\"\"\r\n import psutil # pip install this package\r\n import time\r\n cls = x.__class__\r\n result = cls.__new__(cls)\r\n memo[id(x)] = result\r\n for k, v in x.__dict__.items():\r\n print(\"=\"* 50)\r\n print(\"Current memory:\", psutil.virtual_memory().percent)\r\n print(f\"Saving object {k} with value {v}\")\r\n setattr(result, k, copy.deepcopy(v, memo))\r\n time.sleep(5)\r\n print(\"Memory after copy:\", psutil.virtual_memory().percent)\r\n return result\r\n```\r\nTest script:\r\n```python\r\nimport copy\r\nfrom datasets import load_dataset\r\nbk = load_dataset(\"bookcorpus\", split=\"train\")\r\nbk_copy = copy.deepcopy(bk)\r\n```","body":"## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n","comment_length":113,"text":"concatenate_datasets loads all the data into memory\n## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n\n@lhoestq Just went to open a similar issue.\r\n\r\nIt seems like deep copying (tested on master) the dataset object writes the table's record batches (`dset._data._batches`) into RAM.\r\n\r\nTo find the bug, I modified the `_deepcopy` function in `table.py` as follows:\r\n```python\r\ndef _deepcopy(x, memo: dict):\r\n \"\"\"deepcopy a regular class instance\"\"\"\r\n import psutil # pip install this package\r\n import time\r\n cls = x.__class__\r\n result = cls.__new__(cls)\r\n memo[id(x)] = result\r\n for k, v in x.__dict__.items():\r\n print(\"=\"* 50)\r\n print(\"Current memory:\", psutil.virtual_memory().percent)\r\n print(f\"Saving object {k} with value {v}\")\r\n setattr(result, k, copy.deepcopy(v, memo))\r\n time.sleep(5)\r\n print(\"Memory after copy:\", psutil.virtual_memory().percent)\r\n return result\r\n```\r\nTest script:\r\n```python\r\nimport copy\r\nfrom datasets import load_dataset\r\nbk = load_dataset(\"bookcorpus\", split=\"train\")\r\nbk_copy = copy.deepcopy(bk)\r\n```","embeddings":[-0.1189593524,0.0833053589,0.1260726601,0.371997565,0.0171169862,0.113386929,-0.1137914583,0.3422017992,-0.3763219416,-0.0313398018,-0.0029891629,0.2833094299,0.1033000275,-0.1568289697,0.0146102719,0.4244330823,0.2601239979,0.134578377,-0.4625256956,-0.1483727992,-0.3425718844,-0.0025916975,-0.1786786169,-0.4195239842,-0.4456115961,0.3286666274,-0.3202904463,0.2377286553,-0.0560326912,0.128633216,0.355545193,0.1864214092,-0.0315982215,0.5348338485,-0.0001191608,0.0577579923,0.2564501762,-0.1635587364,-0.4494656026,0.1362714171,-0.3920058906,-0.3536658585,0.1314734966,0.0160404183,0.416490078,0.0663911626,-0.2298910618,-0.4634824991,-0.0232501589,0.0652262866,0.1593083292,0.1110112965,-0.0350349993,-0.1213506758,0.0089878831,0.1179197505,-0.1411336064,0.2453012168,-0.1013238281,-0.116763331,0.0576568209,0.1684359312,-0.35906896,-0.0874381438,-0.0095486175,0.2900185585,-0.1620877683,-0.2823731303,0.2129003853,0.0464695394,0.2146040499,-0.4365739226,-0.0415480658,-0.4749776125,-0.0196938086,-0.3048731089,0.0995297506,0.308213383,-0.1890996844,0.1121600196,-0.13280949,0.0581193231,0.0337472819,-0.073737599,-0.1217983887,-0.1695670187,0.0101737473,0.0271479599,0.4104741812,-0.0549752563,0.4511062503,-0.3960840404,-0.2956274152,0.0855898634,-0.5441936851,0.0141931763,0.1144322529,-0.2700991631,0.1956006885,0.2327721417,0.2149356455,0.0469081774,0.1379220337,-0.022849787,0.4365807772,0.1996461302,-0.1025633216,0.5009908676,0.1304700375,-0.1632460356,0.3409542143,0.110141784,-0.1132148653,-0.0781086311,0.3009509146,-0.3656769097,0.229241237,-0.1927221417,-0.3715546727,0.1397656053,0.0065254094,0.0778659061,-0.0496670604,0.3585070074,0.0671438053,0.2547057867,0.15389283,0.3614987135,-0.06985154,0.1949400306,-0.2237622738,-0.1218684241,-0.16035375,0.2349952906,0.366168052,0.0074927118,-0.0572899692,0.3013069332,0.0723042563,-0.1387106925,-0.0609996095,-0.3262606561,0.3241434395,0.2703713477,0.0587168597,0.2103704959,0.2266677767,-0.1304647774,-0.1423020065,0.4503260255,-0.2546396255,-0.0735558346,-0.069354929,0.10335695,0.0362258926,0.0443489477,-0.4758992195,0.3106041253,0.434776932,0.2327105105,0.0285809096,-0.1107441112,-0.1464839727,-0.3564536572,0.4472286999,0.4729628861,-0.0541462898,0.0487051979,0.0353441387,0.2649788857,0.5495262742,0.3402062953,0.0434603579,0.0879209563,-0.2372004837,0.0179053303,0.0983412564,-0.0005004104,-0.2484383583,0.1872132421,-0.2704435587,0.335116595,-0.0241550449,0.1016249433,0.1242662966,0.097319752,0.660245955,0.2771522999,0.0627577081,0.1526489705,-0.3004397154,-0.1258177757,0.0933524594,0.008635669,-0.1207739562,0.3070910871,-0.0019083718,-0.2292074412,0.386333853,0.0090980995,0.1405576915,0.2855954766,0.1549436897,-0.0899411812,-0.1746754348,0.2379614115,-0.3544645011,0.097163029,0.1835031509,-0.0397738367,-0.1856774241,-0.1383266896,0.134820208,0.2285040617,-0.2306197435,0.0922537223,0.0155475968,0.0989256203,-0.0327142216,0.1316239089,0.1348375231,0.6968744397,-0.1539064348,0.0587004051,-0.4625235796,0.1415293664,0.1405992806,-0.2519921958,-0.0621952377,0.2103710771,0.1967872679,0.0732989833,-0.1008108929,0.4235238731,-0.3193057775,-0.1450687796,-0.2693098783,0.0403387994,-0.0298414659,-0.0516968966,0.1415467411,-0.2254734337,0.2247688621,-0.1563021243,-0.056741938,0.450727582,0.0112392036,0.1874462962,-0.0045901737,-0.1919718832,0.0793231726,-0.1642476618,0.0901824981,-0.1587000936,0.1115505174,0.3809700608,0.1517383903,0.332141012,-0.3299784958,0.1686086655,0.19155626,-0.0641717762,0.2080635428,0.2489986569,-0.3479628861,-0.2832609713,-0.1409808546,0.3459080458,0.7657057047,0.0833002329,0.2235913277,-0.1126445308,-0.0013695674,-0.1795598119,0.3187161684,-0.0918825939,0.1735778749,0.2070209831,0.2243025899,-0.1472703516,-0.2065800428,0.4320197701,0.1452322006,-0.0747572333,-0.2772232294,0.0283994805,-0.3362882733,0.0518880039,-0.3733009398,-0.1222399995,0.0817801803,-0.3311992884,-0.2710944712,0.3929058015,-0.0935641974,-0.1096558943,-0.0498728529,0.3842968047,-0.0503481664,-0.1125858277,0.00265779,0.1787295491,-0.2788047791,-0.0525808372,-0.0056801997,-0.3215699792,0.0520076565,0.2059115171,-0.1652086973,0.0415372588,0.0537941642,0.026079135,-0.195922628,0.0513004847,-0.1727582514,0.1507132947,0.0790431798,-0.2274551392,-0.016985558,0.3729391992,0.0160092954,0.5513109565,0.2199639529,-0.2465394437,-0.1986986995,0.0650124103,-0.3557266295,-0.1634889692,0.2513190806,-0.1487260908,0.0252158549,0.1737631708,0.0827576742,-0.1229356602,0.0731968656,0.1564647704,-0.1991321892,-0.1466024965,0.422041297,-0.0023801727,-0.3424716592,-0.2232850343,0.0801083818,-0.3310308456,0.151091069,-0.5484544635,0.2285906076,-0.5363147259,0.2024652958,0.0190964006,0.0740161762,0.1360855848,0.1101197228,-0.0381845795,-0.1437601894,-0.2721313536,0.0334944762,0.006990639,0.0654356256,-0.0745998323,0.2158823609,0.2305157036,0.0186750889,0.3907527328,-0.0177374594,0.3915452063,0.1259574443,0.5700924397,-0.3393678367,-0.4471695721,-0.3153416514,-0.3729497492,-0.055265449,-0.0380563438,-0.081765987,-0.2075938284,0.05664571,0.1470752656,0.0643212721,-0.2225518376,0.1592552066,-0.2326899916,-0.1527368575,0.1568510383,-0.1012194008,-0.135378927,-0.0830879807,-0.2535193264,-0.0893896669,-0.14918378,-0.1135687903,-0.3458332121,-0.1257225722,-0.6547678709,0.0768065229,-0.1478189826,0.6040266752,0.1357030571,-0.3745177686,0.039974872,0.0413492471,0.8027870655,0.3686524332,-0.2305441052,0.1310169101,-0.0224331059,-0.3862507939,-0.0591944568,0.0813653618,0.2802149057,0.1229078919,0.095016323,-0.2442751676,0.0788232312,0.233195588,0.2669423223,-0.2513377666,0.0450551994,-0.2758742273,-0.2127845287,0.0257187411,-0.2337641567,0.203343749,0.2741473913,-0.3531470001,-0.2460872829,-0.0267989673,-0.204359144,0.2672508359,-0.2123702168,0.4381570518,0.1416172683,0.4936629832,-0.1587229222,0.2963472903,0.3730778694,0.7436975241,-0.1091234162,-0.1179474518,0.1731951386,-0.2392701358,0.3826468289,0.2412312627,-0.1666252762,0.2367009521,-0.2310508639,-0.0033404664,-0.537243247,-0.0061614173,-0.0372622088,0.0735994279,-0.5417613387,-0.273545146,0.4134725928,0.1998048574,0.0134619046,0.4708473384,-0.5638400912,-0.3500669003,0.1962456554,-0.3353712857,0.8098754883,0.0702949613,0.2955168486,-0.0543087274,-0.0189364068,-0.0495379157,0.1149437949,0.1686782837,-0.0994602218,-0.3948722184,0.2031390369,-0.2248120457,-0.1883260161,-0.2532236576,-0.2801165581,0.0907682627,-0.4150789678,-0.1333607584,0.0245230738,-0.1456512511,0.0061809081,-0.2912026644,0.073474966,0.1313215047,0.4107922018,-0.2013477981,0.1182061508,0.0275958497,-0.3105555177,-0.3001871705,-0.2445830405,0.2241717279,-0.2056496292,0.5132964253,-0.2175204009,-0.2047210038,-0.1777374595,-0.0141704688,0.2302371413,0.1844240129,0.0014675306,-0.2093808651,0.216916889,0.3816982806,0.2162962705,-0.3113329113,0.0308292303,0.2869880199,-0.0235171746,-0.0897099376,-0.0129283471,-0.2515476644,-0.3127513528,0.16242221,0.1311502755,-0.2297178954,-0.2221891433,0.2884562612,0.2736463547,-0.2233535498,0.1020718068,0.1862878799,-0.1079064459,0.3019550443,-0.4452914596,-0.4008279741,-0.1504719853,0.5288501382,0.4275746047,0.137094304,0.3589572608,-0.0255245045,-0.0275361314,-0.0926667601,0.3038037419,0.2948084474,-0.3083602488,0.3327580988,-0.2347115725,-0.0889829695,-0.1223549992,0.1906555444,-0.0902995616,-0.1498221904,-0.2004105598,-0.1588423252,-0.1792373508,0.0921351314,0.0629845485,0.2616425455,0.0333868824,0.0792109817,-0.1321287453,0.2453097403,-0.2345609516,-0.1031526476,-0.0935181156,0.1942147017,0.0110161463,0.1036981791,0.2282829583,-0.0901651829,0.0720124245,0.2308770418,0.1835256815,-0.1775111556,0.1119421273,0.1160487011,-0.0784042701,-0.0639154539,-0.3406570256,-0.4153370559,-0.1584494412,-0.2976614237,0.2195536345,0.1389335245,-0.0571225286,-0.2025782764,0.2531115413,0.0496828146,0.0989655331,0.1189783439,-0.1670356095,-0.0721581429,-0.0590221323,0.1988938749,-0.1856142431,-0.0442988016,-0.4569975734,0.0580303818,0.1112943888,-0.1273508668,0.0608634464,-0.2804783285,-0.3058772981,-0.1940655559,0.0891561136,0.4233353138,-0.1628593057,-0.1574442983,-0.0450260341,0.1168304905,0.032736104,-0.3386598825,-0.2539919317,-0.211680308,-0.0826829746,0.310847491,0.1218883395,-0.3003925979,-0.2763796151,-0.0822505578,0.512406826,0.1108015776,0.3124137819,0.1702504307,0.2720431089,0.2492618412,0.0365549028,0.1158342734,0.0244238116,0.2108104229,0.0569148324,0.3601379097,0.2659353018,-0.236796841,-0.1599995196,-0.49415344,0.3613446653,0.5100719929,-0.1413559169,0.3520647287,-0.0123265898,0.0729844421,-0.356743902,-0.2225819528,-0.2216585129,0.1330315918,-0.1914551556,-0.2346656024,0.2077041566,-0.1291694045,-0.0449781045,0.1002355963,0.0568048023,-0.2421501428,0.3202767372,0.1047628298,-0.2098561525,-0.0811471343,0.1302895546,0.3109115958,-0.060378205,-0.2923100293,0.0793742016,-0.0151286246,-0.023587089,0.0448475592,0.4151218534,0.4353760481,0.5306361914,0.2169145495,0.1128671691,0.3910498023,-0.1669711024,-0.2181597352,0.2484214157,-0.132676214,-0.201514855,0.0555638149,0.1131807789,-0.1079042479,-0.0361441188,0.4862872958,0.1950152367,-0.3314703703,0.1184647083,0.0301655084,-0.2474866658,0.112563096,0.0522432476,-0.3138692379,0.0204907227,0.654902339,-0.3941720128,0.0542714521,0.0577680729,0.0555417128,0.1530132443,0.4801187813,0.1707013398,-0.0453491583,-0.3510207534,-0.2779957354,-0.1362156868,0.1387810856,-0.0248796865,0.1667618752,0.0925180987,0.2837331593,0.0989436582,-0.1669981331,-0.2755042613,0.093592003,-0.2863233984,0.0450815372,-0.4407545626,-0.204511717,0.1137031242,0.1588921249,-0.0348877795,-0.3842607737,0.3081074059,-0.0702193603,0.0374397263,-0.0346696153,0.1880494356,0.112813361,0.4133765101,0.2729741633,0.2253101617,0.1860002875,-0.1833369434,-0.1867899746,-0.1063160077,-0.2303209156,-0.1898794472,-0.1697673798,0.0219486728,-0.0460425913,-0.2856986821,0.1003535911,-0.0203452799,-0.0850032046,0.212313652,0.0074449866,-0.0953532085,0.2097866088,0.0374383479,0.2626575232,0.1896249801,0.190323323,-0.0774146542,0.4393135309,0.0394965187,-0.3373165429,0.1008078232,-0.0407974496,-0.1056732908,-0.1429012269,0.3319269419,-0.1910954565,0.1146391034,-0.4796127379,-0.1773150116,0.2614832819,-0.1287756413,-0.5020694137,0.6252390146,-0.1250656098,-0.0407961383,-0.1198756322,0.4905556142,0.0633976683,-0.4901542962,0.2776883841,-0.3202514052]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2276","title":"concatenate_datasets loads all the data into memory","comments":"Thanks for the insights @mariosasko ! I'm working on a fix.\r\nSince this is a big issue I'll make a patch release as soon as this is fixed","body":"## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n","comment_length":28,"text":"concatenate_datasets loads all the data into memory\n## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n\nThanks for the insights @mariosasko ! I'm working on a fix.\r\nSince this is a big issue I'll make a patch release as soon as this is fixed","embeddings":[-0.1398513168,0.015114448,0.0576703213,0.3696739078,0.1839567572,0.1596759409,-0.033678297,0.2669216692,-0.2130460292,0.0089295153,0.0735274181,0.1821824461,0.1092261598,-0.2046645433,-0.1817605793,0.3304117918,0.2417273819,0.2320262641,-0.2873950303,-0.0829890594,-0.2531095147,0.2507131398,-0.311899215,-0.3932681084,-0.4115408361,0.2931381166,-0.3173240721,0.2868468463,0.0731343403,0.0163131412,0.3663010895,0.0684919283,-0.0108415475,0.4899155796,-0.0001141718,-0.0399469957,0.2374247015,-0.1666595042,-0.371073395,0.0382387303,-0.2115313262,-0.4239183962,0.0949019715,0.0163905304,0.3227258921,0.0276343487,-0.1392493546,-0.6195906401,-0.036331322,-0.0184868425,0.2055884451,-0.0736077577,0.0873234943,-0.0346819982,-0.1545480192,0.1501190811,-0.0562109314,0.2053209096,0.0462260507,-0.0868438259,0.1201179102,0.1036038622,-0.2861646712,-0.0662713721,0.0869725496,0.3886128962,-0.1644797772,-0.4153285623,0.0623523779,0.0289843343,0.2989785969,-0.3810659349,-0.0750826746,-0.4106537998,0.0723076537,-0.151874572,0.2585920095,0.5170176029,-0.2185684741,0.2292109579,-0.1242389977,-0.025238242,-0.016185388,-0.0290718637,-0.2198185027,0.1054319218,-0.126125142,0.0533120893,0.3340789378,0.0692817196,0.2617450953,-0.4625049829,-0.3334731758,0.1024646461,-0.5122957826,-0.0164026413,-0.0102712503,-0.3762557805,0.1318039745,0.2016307712,0.0878395736,0.0035140365,0.184177652,0.0197742544,0.3769460618,0.1765152961,0.0080639152,0.3474484384,0.1052436158,-0.2071619928,0.2690299451,0.0588003919,-0.0470570698,-0.2088435888,0.2501572967,-0.3447650969,0.1984031647,-0.1314506382,-0.4553020895,0.1100055501,-0.1005401239,0.0687937289,-0.0308024473,0.2494181991,-0.0282951146,0.4881114662,0.1856010258,0.3167209327,-0.0345314294,0.1992320865,-0.1725660563,-0.1550679654,-0.0954498351,0.2142083198,0.3256211579,-0.236500591,-0.0088531934,0.4181314409,0.0340125747,-0.1766606718,0.0817503259,-0.3128965497,0.2096921206,0.1971650124,0.1583949178,0.422079891,0.1548349708,-0.1589883715,-0.0694989264,0.4355956316,-0.287933737,-0.0658500716,-0.2847761214,0.1766027957,0.0320008993,0.1077617332,-0.2073271722,0.3298118412,0.4095009267,0.2896985114,-0.0911844745,0.0171630085,-0.0444950312,-0.3863945007,0.4746124744,0.374691695,-0.1228752956,0.1804333627,0.0796554461,0.2555893958,0.5120504498,0.3279236257,-0.0482980199,0.0531703196,-0.3730937541,0.0125809368,0.1686255187,-0.1546408236,-0.1727763116,0.3032649457,-0.2416497469,0.3287750483,-0.0439332575,0.169486925,-0.0733178183,0.1647847146,0.5839229226,0.465687722,0.1425248533,0.1275511533,-0.2993758321,-0.1339702904,0.1769673079,-0.1294547766,-0.225111872,0.3253209889,-0.0589603558,-0.1444763392,0.5275813937,-0.1006270573,0.2085408717,0.3714689314,0.150481984,-0.1359591931,-0.2112352699,0.194219932,-0.4309019148,0.0013527095,0.1675441265,-0.2653865218,-0.0621719025,-0.0578743853,0.1020050868,0.0258222912,-0.0636250824,0.0238730665,0.012160602,0.2641777694,-0.099379316,0.0258629303,-0.0086658709,0.8541288972,-0.033142563,0.0026922363,-0.2960159183,0.1831163168,0.0733396783,-0.26403597,-0.0746999756,0.0341198519,0.1943829954,0.0817593113,-0.0983499363,0.4858409762,-0.3071314394,-0.0535058901,-0.1756159216,-0.2325031608,-0.030167209,0.0834080726,0.0084746219,-0.1559773535,0.1944572628,-0.2200031728,-0.0202957094,0.3703812063,0.0729342252,0.1607596129,-0.0248842351,-0.0984202996,0.0550480001,-0.1947636753,0.1069363207,-0.2017446756,0.1457116306,0.2696819603,0.1671738029,0.3178861141,-0.469243139,0.0399977118,0.218240574,0.0924291983,0.1402144581,0.1689800322,-0.4341333807,-0.1735431701,-0.1096332297,0.18923226,0.7930139899,0.1422207355,0.2025080621,-0.081556946,0.0044435863,-0.2626176775,0.3425416946,-0.0330345668,0.0056893597,0.2706080377,0.2729274631,-0.1077963933,-0.1397445947,0.3924454451,0.0544332303,-0.0995829925,-0.2963497937,0.1051601619,-0.2472942919,0.1749789566,-0.3388817012,-0.2995420396,-0.0429706611,-0.3729386628,-0.3870545626,0.4966091514,-0.0050710961,-0.2058933675,-0.0000230926,0.3029479086,-0.0802884251,-0.1078599021,-0.0437195785,0.2978871167,-0.4031485617,0.0023266287,0.0329105556,-0.2414670736,0.0255984701,0.1616852582,-0.0746065453,0.0062471745,-0.0376013555,-0.1460683942,-0.1108777747,0.1193870306,-0.2633400559,0.144841373,0.0067907702,-0.3595245779,0.0139946369,0.4124944806,0.0607036278,0.4417824745,0.3276702166,-0.2216087729,-0.357884407,-0.0738942698,-0.2946987748,-0.2672674358,0.099407427,0.0303421058,0.0763324648,0.1941413432,0.0754337311,-0.164186433,0.0795483217,0.1035593972,-0.1700496078,-0.3065524995,0.5985969305,-0.0416410975,-0.3260031939,-0.0542533733,0.0742413849,-0.2196608633,0.1507494897,-0.5239768624,0.2077150792,-0.5243376493,0.0357964411,-0.2324332595,0.0915300697,-0.0867217407,0.0982720628,-0.0368086696,-0.1558751166,-0.1413527578,0.0125797698,0.0044497326,0.2630653977,-0.1881356984,-0.1222265065,0.1747363359,-0.1320964843,0.5040397644,-0.0255294964,0.3300635219,0.1965639591,0.5206709504,-0.2252723277,-0.4929818213,-0.2043385804,-0.2016645223,0.1007428095,0.0577180497,0.077843219,-0.0946503133,-0.0817828402,0.2344908565,0.0969502479,-0.222257182,0.1670328081,-0.205409646,-0.0506107397,0.0042486354,-0.0312077925,-0.1099209264,-0.1497731358,-0.237592563,0.0199278742,0.0114061795,-0.061896462,-0.2488289028,0.0346120447,-0.5779016018,0.0993207991,-0.1357795596,0.4480731487,0.1299209297,-0.3381295204,0.0687865466,0.0045524505,0.7800455689,0.3826602399,-0.4314598739,0.2051055282,0.1531833559,-0.4868685603,-0.106903173,0.0146364663,0.1714553237,0.1549955308,0.1559254676,-0.1871196926,-0.0236669127,0.0183730777,0.3927375674,-0.2104063481,-0.1052649915,-0.3675717413,-0.1692112088,-0.1150122732,-0.3382618427,0.0803163871,0.3705011904,-0.2373082787,-0.3466645479,-0.0292506367,-0.1737226993,0.2647711933,-0.2550118566,0.5063934326,0.0859555304,0.3880677819,-0.1179890558,0.2238128483,0.2197091281,0.5938230157,0.1231938154,-0.1026300043,0.2181450874,-0.225262776,0.3985524178,0.1742021143,-0.184655726,0.1840730011,-0.3219740391,0.0856892392,-0.6387634873,0.1511772126,0.0732672811,0.1199123412,-0.6324182749,-0.231316939,0.2561938465,0.1056603864,-0.0216331612,0.552747786,-0.5034878254,-0.4059453309,0.040590506,-0.2261582017,0.8108160496,0.0540830195,0.2876097858,-0.0171815865,-0.1803123951,0.1524910778,0.111667335,0.273644954,-0.2010595649,-0.2040135562,0.1287796348,-0.2660422921,0.0044186455,-0.160061568,-0.452701062,0.0691425204,-0.5905037522,-0.0506209359,-0.080623731,0.0679040328,-0.1449848562,-0.3061988652,0.0529611669,0.1591800004,0.2841213346,0.0000369279,0.0886321366,-0.0578101389,-0.2199354023,-0.2148439884,-0.1946781874,0.3777413368,-0.1425355673,0.4557338059,-0.2058804929,-0.2754249573,-0.1219581068,0.1634797305,0.3350863755,0.2765885293,-0.0273095686,-0.1375316679,-0.0673851147,0.3632540405,0.130204767,-0.2657195032,-0.0197195895,0.2913518548,-0.0664217919,-0.0717533752,0.0024251875,-0.2321026921,-0.2963326275,0.2948572338,0.0133136017,-0.2735061049,-0.118689768,0.3011372089,0.3806626499,-0.1508241296,0.1304541677,0.2537243068,-0.0948471725,0.2317899168,-0.3779788613,-0.3909690976,-0.136513859,0.3814929128,0.3144072592,-0.0888853297,0.3437522054,-0.0430450663,-0.0188746955,-0.1143520996,0.2174154669,-0.0165064335,-0.1529435068,0.2706179619,-0.0953308493,0.0080777779,-0.0805858448,0.2164027393,-0.0599567965,-0.079811506,-0.0548536517,-0.1989965737,-0.3820281923,-0.0718343481,0.0063271932,0.2280543149,0.054697834,0.0106074223,-0.0638366714,0.2170872986,-0.2776566744,-0.1158027127,-0.1540311426,0.146075815,-0.0868835524,0.0577560998,0.1480903476,-0.0008431305,0.1407097578,0.105820924,0.106678538,-0.2063993067,0.0332690738,0.1143854931,-0.0116976034,-0.1314940453,-0.2068292797,-0.4186026454,0.0010654363,-0.3671497107,0.1847518831,0.2422655076,-0.0208056588,-0.1279929876,-0.008668663,0.0280055925,0.03711918,-0.0186052453,-0.1917670071,-0.0357006937,-0.0672508329,0.1754504591,-0.0418209881,0.0059081819,-0.5367488861,0.0873259977,0.1139786392,0.0057071242,-0.0187885389,-0.2345141321,-0.2203796357,-0.1694924682,0.1231051907,0.3735879064,-0.2358668447,0.0285465978,-0.0396065824,0.1834036112,0.1108333394,-0.3205860257,-0.1787668914,-0.0526628979,-0.1876270771,0.1849693954,0.1193197593,-0.3131456673,-0.2715880275,-0.0768368989,0.5262079835,0.1202998906,0.292023778,0.211880818,0.2025566548,0.1655825526,0.0613665469,0.1247074157,0.0348170698,0.3070790768,-0.0530882925,0.2144795656,0.2865367532,-0.2327163815,-0.235666126,-0.4283558726,0.2475953102,0.4237318039,-0.155030638,0.2289580405,0.0401220247,0.0791492611,-0.4021759331,-0.3282314539,-0.0945644006,0.1622190624,-0.2592235506,-0.1383008361,0.2126486599,-0.0485826135,0.0484618135,0.0994850248,0.2145466954,-0.2807631791,0.4253142178,0.1433030367,-0.1721548736,-0.242456004,0.2480066866,0.4640237689,0.0112851439,-0.3268963993,0.0366256833,0.027784789,0.0765656829,0.0569226742,0.3208045661,0.4883131683,0.6327383518,0.2889400125,0.1040662974,0.3102713823,-0.0706405193,-0.2328088433,0.2654061913,-0.1099807471,-0.0806352273,0.089370586,0.1750108451,-0.1282295734,0.0753762797,0.3534379005,0.2108038515,-0.4426002204,0.1605287194,-0.0011849416,-0.2379377484,0.0257542785,0.0223686974,-0.2537686825,0.004722802,0.5532049537,-0.1821507066,0.1118662432,0.013844368,0.0740121827,0.2559785247,0.2634495497,0.0847179368,-0.0089979451,-0.3793722391,-0.3257670105,-0.191137895,0.2649611533,0.1162207127,0.365485847,-0.1273177117,0.2802459896,0.0141838295,-0.0619113147,-0.174649924,0.1437336206,-0.158567369,-0.0395416357,-0.4842324555,-0.0827421024,0.2124892324,0.0345615298,-0.0472925231,-0.3593175709,0.3947598636,-0.1068708971,0.1028013602,-0.0799385384,0.2341654152,0.1995652169,0.337266475,0.374832958,0.2423322201,0.2296585739,-0.2148532569,-0.1963131726,-0.2213146687,-0.3038081229,-0.0730425864,-0.1656857282,0.0013532216,0.0800111517,-0.4104301035,0.1225764751,-0.1289905757,0.0115279211,0.2242868096,0.1620289236,-0.0820090771,0.0895310044,0.0539804809,0.2117806375,0.1964697987,0.174843058,-0.0805256516,0.3490605056,-0.1060525477,-0.3138890862,0.0989739075,0.0467913002,-0.2664640844,-0.1155917197,0.3089934587,-0.0543666258,0.1614176482,-0.6315418482,-0.1777596772,0.3585123122,-0.1232246235,-0.4799721241,0.5117151141,-0.1925513297,-0.1067776084,-0.0470472686,0.361828506,0.1069896296,-0.4236905873,0.281674087,-0.2509839535]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2276","title":"concatenate_datasets loads all the data into memory","comments":"Hi @samsontmr @TaskManager91 the fix is on the master branch, feel free to install `datasets` from source and let us know if you still have issues","body":"## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n","comment_length":26,"text":"concatenate_datasets loads all the data into memory\n## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n\nHi @samsontmr @TaskManager91 the fix is on the master branch, feel free to install `datasets` from source and let us know if you still have issues","embeddings":[-0.170591861,0.0092170332,0.0292842202,0.3677366972,0.1755022407,0.1506508887,-0.0363330878,0.2557009161,-0.2184762359,0.0205546785,0.0756376907,0.2006986439,0.1137187034,-0.1913734525,-0.1878060699,0.3212848306,0.2241285145,0.2327459753,-0.3028806746,-0.0799712464,-0.2509787977,0.2532582581,-0.3152192235,-0.4041965008,-0.3926775455,0.2633895576,-0.324208647,0.2930183411,0.0551784709,0.0143002868,0.3530985415,0.0911309645,-0.006137263,0.5166870356,-0.0001119337,-0.0348054878,0.219003275,-0.1603055149,-0.3798380494,0.0134006506,-0.2082094103,-0.431055218,0.0528312698,-0.0045533706,0.3232043386,0.0165413711,-0.1144325137,-0.6221048832,-0.0218277741,-0.0040224274,0.2312553972,-0.0567462742,0.1109552085,-0.0584037527,-0.1406242996,0.1617460996,-0.0351833552,0.1927781105,0.0371735543,-0.0715287179,0.1079412773,0.079248555,-0.2629720271,-0.083159931,0.0752241835,0.3842883706,-0.1346914023,-0.4134614468,0.0804955363,0.0150397122,0.3004926145,-0.3991719782,-0.0938755423,-0.3797347844,0.0740216374,-0.1438546181,0.2679314315,0.4909189045,-0.2104186267,0.219563514,-0.1331172436,-0.036526829,-0.0239779484,-0.028899638,-0.2043293566,0.0931631997,-0.1354396194,0.0305273775,0.3242896497,0.0352999233,0.2702701688,-0.4690794647,-0.3298483789,0.1103057042,-0.513263464,-0.0276157558,-0.0030134087,-0.3875360489,0.1243834794,0.2151877582,0.0875878707,0.0227854196,0.184498772,0.0349109657,0.3635174632,0.1804081649,0.0184642617,0.3571968675,0.1284895092,-0.1866255999,0.2579454184,0.0578463078,-0.0656249374,-0.2233371288,0.2364633381,-0.3508513868,0.1808035523,-0.1380872279,-0.4598098695,0.1174360663,-0.0863997713,0.0496026762,-0.0094316872,0.2542226017,-0.0292180497,0.4726689756,0.1849807799,0.2863076925,-0.0344359241,0.201805681,-0.1823039949,-0.1430383474,-0.1071337834,0.208167702,0.318336159,-0.237367779,-0.0012824204,0.3865213394,0.03396466,-0.1422039121,0.0747801885,-0.3283980787,0.1817576289,0.1874959767,0.1599019468,0.385514915,0.1233887896,-0.1299316287,-0.0734612346,0.4197448492,-0.2636742592,-0.0535451323,-0.2766641974,0.2080066651,0.0323732533,0.0900532603,-0.1872912943,0.3255513608,0.3913214207,0.2665786147,-0.1022031605,0.0159123689,-0.0254238937,-0.4058591723,0.4793983996,0.3491703272,-0.1226651818,0.1716086864,0.0541150719,0.2214607894,0.5023226738,0.3170529008,-0.0411457643,0.0464436114,-0.3630587161,0.0208721384,0.172096312,-0.1713977009,-0.1892300695,0.2992790341,-0.232709378,0.305789113,-0.0190437678,0.1393660903,-0.0547649898,0.1674485207,0.5857989192,0.4550622106,0.1370738745,0.1111657768,-0.3130151033,-0.1439277977,0.1525891423,-0.1146862805,-0.2094824761,0.3083630502,-0.0232087485,-0.1329599023,0.5200488567,-0.1009353697,0.2115394026,0.374530077,0.174204886,-0.1261503696,-0.2095542401,0.176323697,-0.4116720557,-0.0128653375,0.1423327327,-0.2536487579,-0.0743055195,-0.058770448,0.0862646177,0.0054154764,-0.0461516567,0.0131203309,0.0468461588,0.2829904258,-0.0541425571,0.0100656962,-0.0051468834,0.8445510864,-0.0522520691,-0.0038746768,-0.2735114694,0.2024548203,0.0517794602,-0.2781853676,-0.0604393296,0.0601664297,0.1768258363,0.0623119436,-0.0969210938,0.5086948872,-0.3162319064,-0.0530032888,-0.128217712,-0.2386722118,-0.0197944585,0.0813364983,0.0303822011,-0.1581412107,0.1951512247,-0.2013555616,-0.0485713333,0.3627675772,0.0637389719,0.1470878571,0.0014080026,-0.1114584878,0.0617574602,-0.1842131168,0.1107689291,-0.1929574013,0.1656011343,0.2717431486,0.1882426143,0.2941269875,-0.4811876416,0.0245005563,0.2233474255,0.0795772821,0.1472449899,0.1627761126,-0.4270969927,-0.1794689596,-0.1047101915,0.1944568157,0.8014233112,0.1726982445,0.2021740973,-0.0856779665,-0.0116208401,-0.2446803749,0.3470253944,-0.0430199094,0.0150324805,0.2690575123,0.2735055089,-0.1009962186,-0.1482656449,0.3355564773,0.0742284507,-0.1128370687,-0.2773726285,0.0631775036,-0.2430235893,0.1607270092,-0.315087527,-0.2844523787,-0.0286442153,-0.3622499704,-0.3986128569,0.4999139607,0.0003713467,-0.1869661957,0.0068637296,0.2999919355,-0.0691002831,-0.0762333125,-0.0526445471,0.2924110591,-0.3996618092,0.0342250839,0.0327193178,-0.2350403666,0.059115652,0.1551668793,-0.063314639,-0.0152570223,-0.0391109101,-0.1443186551,-0.1046109498,0.1309558153,-0.2461868972,0.1443766654,0.0002403225,-0.3521909714,0.0151467007,0.386020869,0.0331956856,0.4261025488,0.317624867,-0.2234506309,-0.367921561,-0.089445807,-0.3102516234,-0.2938085794,0.0898835808,0.0360095501,0.0871712938,0.1884695143,0.0940233693,-0.1125219539,0.0752632767,0.1272690892,-0.1948869228,-0.3146141171,0.5944067836,-0.0795404613,-0.3260225058,-0.0507530794,0.0738560334,-0.2051508129,0.1580086946,-0.518168509,0.1775104702,-0.5287163258,0.0553962812,-0.2294151932,0.1287552863,-0.0822000131,0.1013927013,-0.0731060654,-0.1697247773,-0.1496968269,-0.0004368192,0.0265975446,0.2668580711,-0.1863807142,-0.1166978627,0.1746390909,-0.1422039568,0.5094496608,-0.0357053131,0.3200167716,0.1902724653,0.5234614611,-0.1986839771,-0.4719843864,-0.1932121515,-0.2005855143,0.0893358141,0.0563366376,0.0636828318,-0.0920831263,-0.0806076303,0.224447608,0.0911798105,-0.2175606191,0.1473002434,-0.1959721297,-0.0329134054,-0.019227732,-0.0387291312,-0.1207239106,-0.1115191355,-0.2408657521,0.0230470635,0.0003594924,-0.0564636961,-0.2714478076,0.0201331005,-0.5876372457,0.1094759107,-0.1373864412,0.4535588622,0.109008424,-0.3208844066,0.0583272651,-0.0161342192,0.7636983395,0.3329622149,-0.4363057911,0.2071588933,0.1552901417,-0.4658376276,-0.083580777,-0.0213175043,0.189979732,0.1312688738,0.1583263725,-0.2055954486,-0.0296573471,0.0358316898,0.3744699061,-0.2042749673,-0.0955858901,-0.3499988914,-0.173755601,-0.1365813762,-0.3245686889,0.0489245579,0.4076435864,-0.2329463363,-0.3107894063,-0.0368187875,-0.1849568933,0.2744167745,-0.2308510393,0.514585793,0.0639485866,0.3858354986,-0.097893931,0.2156769484,0.2473815978,0.5988701582,0.1038330123,-0.1060993075,0.2526395619,-0.2163180858,0.4124968648,0.1633570343,-0.1602391005,0.1917658001,-0.3098959029,0.0977840647,-0.6230969429,0.1641239822,0.072556667,0.1276913732,-0.6346238852,-0.2413502336,0.2271974981,0.1082774401,-0.0185788926,0.5124950409,-0.4901216626,-0.4179080129,0.0697128475,-0.2118566781,0.8130754232,0.0547282808,0.2585769892,-0.0296547245,-0.1845354289,0.1582954228,0.0853048041,0.2678154707,-0.1852398664,-0.1961282641,0.1438766271,-0.2712043226,0.0100168055,-0.1378272325,-0.484480679,0.0686793253,-0.548830092,-0.0551482253,-0.0857354552,0.0833575502,-0.1527262628,-0.304567188,0.0371952727,0.1895385087,0.3127615154,0.0013445584,0.0653884336,-0.0448101163,-0.2109622061,-0.1944261789,-0.1934619695,0.3593723476,-0.1455610245,0.4600490034,-0.2225938439,-0.2794795334,-0.127181679,0.1378703564,0.3171218932,0.2699973583,-0.0230862778,-0.1305085272,-0.0730518624,0.3410402834,0.1088792607,-0.2430868298,0.0008913178,0.2779831886,-0.0806919783,-0.0385391377,-0.0032455535,-0.2296980023,-0.287956655,0.2690279484,0.0109510059,-0.2714538574,-0.0941431373,0.304525435,0.3451378644,-0.1455772221,0.158016488,0.2575798631,-0.087366432,0.241808489,-0.3634423912,-0.3893838525,-0.1616250426,0.3766770065,0.2974401712,-0.0833572224,0.3261156678,-0.0495862477,-0.0365192518,-0.1432907283,0.1960612983,-0.0354609899,-0.1487310827,0.2714607418,-0.0911768228,-0.0164950769,-0.0638385192,0.2176407129,-0.0276381187,-0.0734695271,-0.0515613109,-0.2201011926,-0.3964553475,-0.0557610579,-0.0022764294,0.2487777621,0.0545786619,0.004968402,-0.0857975855,0.2177583128,-0.3150094748,-0.108596094,-0.1695530415,0.1493176371,-0.0745879263,0.043076396,0.1462408453,0.0055682268,0.1722056568,0.1193140298,0.1077749208,-0.2334418297,0.0453122333,0.0964168757,-0.0259043649,-0.1499037147,-0.2246697545,-0.4041139781,0.0112567395,-0.3703247309,0.1964097172,0.2312370539,-0.035397768,-0.1200981513,-0.0055265063,0.0413132198,0.0197682343,-0.012156304,-0.1634091437,-0.0472184978,-0.0828909725,0.1975089014,-0.061721243,-0.0061473679,-0.5149455667,0.0845452026,0.1428639591,-0.0128747625,0.0052393684,-0.2167884111,-0.2100987136,-0.1587082446,0.1227068603,0.372314781,-0.2173629552,0.0193792246,-0.0240696073,0.2157365978,0.0988935903,-0.325053364,-0.2007527202,-0.0765904188,-0.158984527,0.1698937565,0.1420832574,-0.2991505265,-0.2625758946,-0.0591584258,0.5463074446,0.1247389838,0.2897948325,0.2192467898,0.2227225006,0.1619433165,0.082105495,0.1254656613,0.0443909802,0.3199918866,-0.0474389158,0.2595846653,0.2789168358,-0.2171848565,-0.2115840465,-0.4280100167,0.2837201357,0.4275855422,-0.1443542093,0.2253155112,0.0233595446,0.0850126371,-0.4194725752,-0.2932543755,-0.1019086838,0.139627561,-0.2674476206,-0.1335678697,0.1960607469,-0.0838549137,0.0306117889,0.0907151401,0.2005171627,-0.2689688504,0.4259611964,0.1229153648,-0.1554669291,-0.2405553758,0.2527740896,0.4658412039,0.0022952091,-0.3160019815,0.0356349386,0.0271701142,0.072089687,0.0554053411,0.3411989808,0.4986339808,0.6147717834,0.2721056044,0.1044659242,0.3074151874,-0.0678053722,-0.2478257865,0.2650793195,-0.1127168089,-0.108143419,0.0906782597,0.2065341324,-0.1485038698,0.05666361,0.3594036996,0.1688804328,-0.4155342877,0.1474209577,0.0089649726,-0.2527569234,0.0678588897,0.0108298007,-0.2431136817,-0.0130260754,0.5400262475,-0.2046988457,0.1184391156,0.0085532758,0.0944712311,0.2409147024,0.2731514871,0.104712449,-0.0267013013,-0.371048063,-0.3345951736,-0.2055735439,0.2501990795,0.1388462931,0.341666013,-0.1382288486,0.2902057469,0.018928444,-0.0795916021,-0.1497365385,0.1210184693,-0.1521126926,-0.0313671753,-0.4907976687,-0.0563847609,0.2144361138,0.0261179022,-0.0263207555,-0.3581999242,0.3783017397,-0.1205616295,0.1346715093,-0.0968298614,0.2296245843,0.2119969577,0.3251651525,0.3890555203,0.2415724695,0.2294231951,-0.2480791807,-0.2008237839,-0.2242737561,-0.2944231033,-0.0670720637,-0.1548818648,-0.0179351382,0.081275247,-0.410318166,0.1024460718,-0.1002154648,0.0182395708,0.2129152268,0.1546316594,-0.0696231797,0.0950295702,0.0535516702,0.1901407689,0.2027446181,0.1909311712,-0.0671830177,0.3385685682,-0.1113278568,-0.3111752272,0.0780565441,0.0500970446,-0.2669709325,-0.1360529661,0.2945317328,-0.0741825327,0.172639966,-0.6362550259,-0.1700377613,0.3600722253,-0.1359412968,-0.4746211171,0.5072944164,-0.1810392737,-0.1051709279,-0.0325118937,0.3822626173,0.0733084306,-0.4318178594,0.2638352811,-0.2444105744]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2276","title":"concatenate_datasets loads all the data into memory","comments":"We just released `datasets` 1.6.2 that includes the fix :)","body":"## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n","comment_length":10,"text":"concatenate_datasets loads all the data into memory\n## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n\nWe just released `datasets` 1.6.2 that includes the fix :)","embeddings":[-0.146085307,0.0303163044,0.0393807888,0.37336725,0.1540361941,0.1677799672,-0.0191981252,0.2605399191,-0.2435812205,0.0131694656,0.0582248494,0.1784075946,0.1040864363,-0.2164498121,-0.1754347682,0.3355038166,0.2411968261,0.2202045918,-0.2739174366,-0.081243366,-0.247351408,0.2451900095,-0.3168361783,-0.3823642433,-0.404481709,0.2859959006,-0.3224494159,0.3059146404,0.0710617155,0.0027194559,0.3710920513,0.1032611355,-0.0133597776,0.5149241686,-0.0001145863,-0.0350632146,0.2296015173,-0.1475828141,-0.3935889602,0.0388054028,-0.2348852754,-0.4296608269,0.0643651858,0.0113204699,0.351051271,0.0300263707,-0.105051361,-0.6652596593,-0.0510699488,-0.0006860563,0.204287678,-0.0676139519,0.0894967616,-0.0582116693,-0.1576556563,0.1622509509,-0.0454800017,0.2094069272,0.0348502882,-0.0815880299,0.1016634107,0.0735454112,-0.2592453361,-0.083422415,0.0838360563,0.3897694647,-0.1617908627,-0.4274959862,0.0775155872,0.0278428141,0.3115271926,-0.3802059293,-0.0855419859,-0.4082713723,0.0351825841,-0.1192924976,0.2642156184,0.4899975061,-0.208709538,0.2289491892,-0.1450100839,-0.040071778,-0.0170303341,-0.0325643979,-0.2294583172,0.0599999093,-0.1318623871,0.0454784892,0.3391783237,0.0404771827,0.2950280011,-0.4626585245,-0.3034632802,0.1171669066,-0.5105712414,-0.0146053387,-0.0022978722,-0.3861888647,0.1485217959,0.1896292567,0.0787433833,0.0199447237,0.1630929559,0.0273114238,0.3908552229,0.1668454558,0.0355937965,0.3791486323,0.1024676487,-0.2043576092,0.2730085552,0.0578394048,-0.0675644204,-0.2072717845,0.288554877,-0.3823831379,0.2072774023,-0.1418024451,-0.46190992,0.1306217015,-0.0932404697,0.0679046586,-0.0276516881,0.2162925154,-0.0290461797,0.4888408184,0.1765419543,0.2931123078,-0.0212719161,0.213665396,-0.1775455624,-0.1645159125,-0.108286038,0.2193517238,0.2996461987,-0.2361288667,-0.0052076289,0.3944636881,0.0334782824,-0.1527783871,0.070829019,-0.3145301938,0.1959461123,0.1908815652,0.1478322297,0.4091849029,0.1418364793,-0.1418253481,-0.0673570037,0.4310062826,-0.2935177684,-0.0608428419,-0.2894254029,0.1760149896,-0.0010470352,0.0943759456,-0.2068603933,0.3254600167,0.3998029828,0.2474611849,-0.0759163946,0.016514793,-0.0340044461,-0.393991977,0.4684164822,0.38313362,-0.1283294111,0.1598059237,0.0329959765,0.2219454199,0.5018486381,0.298343122,-0.0404887646,0.0517255925,-0.3872821033,-0.0047220616,0.2006138861,-0.166792348,-0.1872843504,0.2823456824,-0.2405069023,0.3492953479,-0.0322700702,0.1580548882,-0.0687379986,0.1636402011,0.6182198524,0.4394689202,0.1273228526,0.1044442132,-0.3026207983,-0.1490970701,0.1667072028,-0.1141086668,-0.2076791674,0.331204623,-0.0363298766,-0.1279819012,0.5139526725,-0.0696302205,0.1982998848,0.3732087016,0.1381919682,-0.1206969842,-0.1970086843,0.1754201949,-0.4346022606,0.0105003538,0.1559218466,-0.2795760632,-0.0721661448,-0.0661409199,0.0855706483,0.0090863323,-0.0518898033,0.0327324718,0.0094780372,0.2759043276,-0.0956747308,0.0339707471,-0.0208134204,0.8513176441,-0.0183306504,0.0110458573,-0.3118620813,0.1761735976,0.0813087448,-0.2455288619,-0.0944070891,0.0308486987,0.1673162729,0.0638437271,-0.0808518156,0.5145437121,-0.2904238403,-0.046295438,-0.1567268074,-0.247727707,-0.0174368024,0.1033771262,0.0049385857,-0.1660743952,0.1794856042,-0.2128377408,-0.0577742122,0.3720332384,0.0502871051,0.1592637599,-0.0229110382,-0.0989993289,0.0417482518,-0.182227999,0.0942653418,-0.1888293326,0.1494719833,0.2801570892,0.197989732,0.3298128545,-0.504645884,0.0431764536,0.2108942121,0.0757964849,0.1313936412,0.1668796986,-0.4245755076,-0.1886354685,-0.1013943031,0.2027505636,0.8031783104,0.1458126903,0.22337161,-0.0848751217,-0.0018002355,-0.2375929952,0.3312143981,-0.0177040417,0.015200139,0.2630984485,0.2749147713,-0.0764445662,-0.1345153898,0.343257457,0.0659588128,-0.1058452353,-0.3088314235,0.1153797805,-0.2474131733,0.18136549,-0.330003798,-0.2960261405,-0.0473113991,-0.3744719028,-0.4023845494,0.4886190295,0.0118872421,-0.2103978395,-0.0347145945,0.3259149492,-0.061882697,-0.1085805222,-0.0441844165,0.3093555868,-0.40001297,0.0094066365,0.0326704644,-0.2525604963,0.0343113542,0.143248111,-0.0666693002,0.0061636879,-0.0288937259,-0.1482231617,-0.0957856029,0.1328398138,-0.228372395,0.1458331943,-0.0103470385,-0.3338783085,0.0027693885,0.4412018061,0.0596523061,0.431646347,0.3384381831,-0.2106789052,-0.345797509,-0.0829713419,-0.2933303714,-0.2672246397,0.1154412404,0.0112220841,0.0766462162,0.2112600654,0.0889114067,-0.102959238,0.0734848753,0.1053248793,-0.1736071557,-0.3370990455,0.5983121395,-0.0522048436,-0.3185575902,-0.0751680359,0.0761140287,-0.2258685082,0.1817123145,-0.52673316,0.2026055455,-0.5129346848,0.0418146849,-0.238417998,0.0937045738,-0.0856186897,0.1104968637,-0.0454135798,-0.1601970345,-0.1419030428,-0.0155328987,0.0088963397,0.2737543285,-0.191119805,-0.1045082137,0.1917799115,-0.1346291602,0.4946791828,-0.0189970769,0.3127444088,0.1742968261,0.5493310094,-0.1909850389,-0.4891180396,-0.1991260052,-0.2027278394,0.0672693774,0.0392175205,0.0630422831,-0.0816236734,-0.0738252625,0.2203604281,0.1107147932,-0.2125081867,0.1451679617,-0.2453724593,-0.0341036208,-0.0124681592,-0.04899811,-0.0992197245,-0.1203306243,-0.2529835105,0.018745346,0.0239825584,-0.0597277842,-0.2866056561,0.0603080913,-0.5661479235,0.0979921445,-0.1318306178,0.462436378,0.112574771,-0.3182171881,0.0611944422,0.0039839312,0.7987990975,0.3678527474,-0.418738842,0.2144976705,0.1745514423,-0.4922047257,-0.0962373391,0.0090806428,0.2236403376,0.1503154933,0.1722679585,-0.1644340008,-0.0265480354,0.0367304608,0.3763964772,-0.2109329253,-0.1065799519,-0.3564773202,-0.1727278233,-0.1035264507,-0.315217346,0.0746355504,0.3847097456,-0.2132800519,-0.3279328346,-0.043916475,-0.172435686,0.2849879563,-0.2279546857,0.498228699,0.0745735466,0.3677732646,-0.0690821111,0.1667140275,0.2413741201,0.6092144251,0.1187990829,-0.1422341019,0.2336911559,-0.2056414336,0.4180021882,0.1551944464,-0.1727444679,0.159621194,-0.2977648377,0.0706533343,-0.6056122184,0.1712537259,0.0811932236,0.1103947684,-0.6752091646,-0.2507430315,0.264081955,0.125460878,-0.0079996856,0.5352156162,-0.4725586176,-0.4208785594,0.0324072912,-0.2185956836,0.8024559617,0.0591279194,0.2772392333,-0.0236520432,-0.1729031503,0.1592616439,0.1194466427,0.2508530021,-0.1763953269,-0.1816692203,0.1189554185,-0.279179424,0.0062207053,-0.152235657,-0.4545689821,0.0815640837,-0.590891242,-0.0632290021,-0.0497704521,0.0536888242,-0.1327498704,-0.2972792685,0.0628771707,0.1649300754,0.300249666,-0.0053678132,0.0732641593,-0.0471684709,-0.2146418393,-0.1917070746,-0.2013403624,0.3595133722,-0.1367607415,0.4503142238,-0.1830509305,-0.2777614295,-0.1192270517,0.1635042429,0.3133012354,0.2647567093,-0.0124345254,-0.1491117477,-0.0700133964,0.3407376707,0.1421868652,-0.2702345848,0.0193626955,0.3116307259,-0.0735309273,-0.0639819205,-0.006706825,-0.2351511419,-0.2951131761,0.3103396297,0.0294970684,-0.2865570188,-0.1334168911,0.2644022107,0.3649924099,-0.1343639344,0.131828621,0.2601546645,-0.0651954934,0.2325612158,-0.3796776235,-0.3829919398,-0.1229825467,0.3753474951,0.2949557602,-0.0880788267,0.3312292099,-0.0769923478,-0.0202310868,-0.1133533344,0.2349284142,-0.0091423625,-0.1263387501,0.2885500193,-0.1224999577,0.0105288569,-0.0817368627,0.1991022229,-0.0575954914,-0.1053360924,-0.0505529754,-0.2170135528,-0.3766368628,-0.0481182747,0.0077124587,0.238842234,0.0774783343,-0.0364678837,-0.0711938515,0.1764170229,-0.2784692347,-0.0972827151,-0.1587462425,0.1446032673,-0.0649604797,0.0697043389,0.1514972597,-0.0308738016,0.1392356753,0.1122694016,0.1191973686,-0.2109419107,0.0298956856,0.1140178815,-0.0045726867,-0.1399390399,-0.204909265,-0.4105929136,0.0155567853,-0.3584765494,0.1795067489,0.2618009448,-0.0140834916,-0.1208126545,-0.0160247553,0.0247588232,0.0084594553,-0.0079256427,-0.1736847311,-0.0606392063,-0.069584839,0.2251535952,-0.0654274449,0.0024717234,-0.5385620594,0.0975952595,0.1374663562,-0.0200991668,-0.0195518658,-0.2432418764,-0.2072783262,-0.1582346261,0.1120764613,0.4056002796,-0.2355064303,0.0174850095,-0.0509884059,0.1806000918,0.1059770137,-0.3319741488,-0.203995347,-0.0569857433,-0.1875429899,0.1790538877,0.1091598272,-0.3194824457,-0.2719939947,-0.0638241693,0.5600719452,0.102411218,0.3140778542,0.2181774527,0.2060091347,0.1540130675,0.1062925309,0.1303851306,0.0393968225,0.3244773746,-0.0354144387,0.2368234247,0.312630862,-0.2264216542,-0.2344570756,-0.4063408375,0.270221293,0.4634734988,-0.1539798677,0.233292371,0.0432224795,0.0769078434,-0.4293982089,-0.2940843999,-0.0777337477,0.1194512844,-0.2436008155,-0.1387564689,0.2311366946,-0.0775968879,0.0302764419,0.087949194,0.2085721195,-0.2891420424,0.441149801,0.1082124561,-0.1773457825,-0.2399639785,0.2512705624,0.4726603031,0.0003655879,-0.3271406293,0.0260710213,0.0085721379,0.0567881465,0.0767262354,0.3657236993,0.5011808872,0.6203553081,0.2993020415,0.1005006954,0.2926928103,-0.0833655596,-0.2433958352,0.2662456334,-0.1352584809,-0.1203578785,0.0633338168,0.1852027625,-0.1237171963,0.1027999595,0.36296013,0.182573989,-0.4694604576,0.1580235064,-0.013873782,-0.2385495007,0.0627761781,0.0178940501,-0.2360942066,-0.0334428959,0.5596913099,-0.1926750243,0.1074113399,0.0259626396,0.0729196668,0.2706975341,0.2489480823,0.1113268659,-0.0397801958,-0.3467857838,-0.3167397678,-0.219942376,0.2453430742,0.1430887431,0.3409044743,-0.1036614999,0.289575845,0.0103284596,-0.0587072894,-0.1764861792,0.1355377734,-0.1941390187,-0.0412973128,-0.4622561336,-0.0790119544,0.2165543735,0.0323442072,-0.0521448106,-0.3643663228,0.4012223184,-0.1272708476,0.0986743048,-0.0783838108,0.2219251394,0.2145172358,0.3178834915,0.3873594105,0.2301864177,0.2174175233,-0.2095179856,-0.2128257155,-0.2615278661,-0.3039789796,-0.0763782784,-0.1526922137,-0.0235496853,0.0663091242,-0.434410274,0.1159046367,-0.0861134976,0.0142864482,0.2260772437,0.151303336,-0.0612495653,0.0992964059,0.0530648939,0.19969818,0.1994184405,0.193300575,-0.0803988278,0.3557089567,-0.1120845377,-0.3084890246,0.0725625083,0.0570947267,-0.2754215598,-0.1187255904,0.2882756889,-0.0642946586,0.153945446,-0.652515173,-0.1796154976,0.3844854832,-0.1337441206,-0.4622994363,0.5038222075,-0.1887709945,-0.1100367308,-0.049169682,0.399654597,0.0728746429,-0.4120744169,0.3105203807,-0.2367001027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2276","title":"concatenate_datasets loads all the data into memory","comments":"thanks it works like a charm! :)","body":"## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n","comment_length":7,"text":"concatenate_datasets loads all the data into memory\n## Describe the bug\r\nWhen I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk.\r\n\r\nInterestingly, this happens when trying to save the new dataset to disk or concatenating it again.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/7063207\/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png)\r\n\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import concatenate_datasets, load_from_disk\r\n\r\ntest_sampled_pro = load_from_disk(\"test_sampled_pro\")\r\nval_sampled_pro = load_from_disk(\"val_sampled_pro\")\r\n\r\nbig_set = concatenate_datasets([test_sampled_pro, val_sampled_pro])\r\n\r\n# Loaded to memory\r\nbig_set.save_to_disk(\"big_set\")\r\n\r\n# Loaded to memory\r\nbig_set = concatenate_datasets([big_set, val_sampled_pro])\r\n```\r\n\r\n## Expected results\r\nThe data should be loaded into memory in batches and then saved directly to disk.\r\n\r\n## Actual results\r\nThe entire data set is loaded into the memory and then saved to the hard disk.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\n- Datasets: 1.6.1\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26) \r\n[GCC 7.3.0]\r\n- Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10\r\n```\r\n\nthanks it works like a charm! :)","embeddings":[-0.1167225093,-0.0080092484,0.0535871461,0.3752789199,0.1577809155,0.1856212616,-0.019155113,0.2346013635,-0.2243559659,-0.0018983706,0.1053597033,0.1621640027,0.1124006584,-0.2112618983,-0.1625767648,0.3219698668,0.2721738219,0.2222565562,-0.253498733,-0.1009221002,-0.2356301844,0.2408419251,-0.3372508287,-0.3942461312,-0.4050992727,0.3053331375,-0.3452889025,0.3094511628,0.0814065188,0.0507833101,0.3862079978,0.0803597942,0.0124108959,0.515994668,-0.0001169711,-0.0523180701,0.2316369116,-0.1693799943,-0.3562711179,0.0401990712,-0.2190032899,-0.4345083833,0.0656195953,0.0082894927,0.3437852561,0.0373588502,-0.1173345,-0.6581455469,-0.03234284,-0.012283057,0.1861512363,-0.0906370133,0.0894812942,-0.0270233285,-0.1552164555,0.1803214103,-0.0426061414,0.2046961337,0.0510974713,-0.0702268928,0.1384816319,0.083417207,-0.281293869,-0.1057186127,0.1118245646,0.4073778987,-0.1832231283,-0.4277389646,0.0738076568,0.0069117546,0.3404434323,-0.398108989,-0.074041687,-0.3997927904,0.029276168,-0.1350212693,0.2655061483,0.5107787251,-0.2003498673,0.2314664423,-0.136675939,-0.0321017727,-0.0191402677,-0.0189272929,-0.2352524251,0.1058334485,-0.1288841367,0.0699277148,0.3259629011,0.0677116588,0.2919636667,-0.4761316776,-0.2934914529,0.1359986812,-0.5336111188,-0.019332597,-0.0254567172,-0.3772266507,0.1380503476,0.2001530528,0.0949466079,0.00416971,0.1880039424,0.0188994054,0.3976982534,0.2006783485,0.0227573235,0.3687833846,0.1115019172,-0.2260210365,0.259790808,0.0465009511,-0.071944952,-0.1945050508,0.2944571078,-0.3931571543,0.1997938305,-0.1336738616,-0.4689389765,0.1150570437,-0.1067241579,0.0567336641,-0.0325712934,0.2244647592,-0.0395090841,0.4825879037,0.194268778,0.3203171492,-0.0309646986,0.2252405286,-0.1624368131,-0.1627715528,-0.1227209345,0.229839474,0.2963659763,-0.2257878333,-0.0012392853,0.3977627754,0.0341472104,-0.1602352709,0.0786955878,-0.2974763215,0.2031440437,0.1862682998,0.1467393488,0.4229363501,0.1424358785,-0.1629554182,-0.0681586415,0.4663750827,-0.2982423902,-0.0548323095,-0.2848940492,0.1560093313,-0.0159195103,0.0912825987,-0.2143431455,0.3195360303,0.4283790886,0.2450360358,-0.0627542064,0.0250492506,-0.0308533814,-0.3976528645,0.481418699,0.3787437975,-0.1417109072,0.1621032357,0.0706790388,0.2488301843,0.5283635855,0.3223563433,-0.0110058021,0.0500100777,-0.3806521297,0.0191081297,0.2036043406,-0.1712295562,-0.1846901476,0.2914278805,-0.2437743843,0.3502888381,-0.0252080299,0.1873915642,-0.0932048485,0.1625858098,0.5992077589,0.4449378848,0.1289371997,0.1050088629,-0.2894152105,-0.1448588222,0.194350183,-0.1379926652,-0.2184612602,0.3541206121,-0.0413334817,-0.1134321615,0.5286946893,-0.0890727788,0.2068559527,0.3683291972,0.134045288,-0.1195921749,-0.1863797903,0.2027831376,-0.4209855199,0.0073826625,0.1694492847,-0.2962956727,-0.0453006066,-0.0442033261,0.0995031893,0.0130316475,-0.04120408,0.0189379621,-0.0242272802,0.2638458908,-0.0932803974,0.0042991699,-0.0419820547,0.8521857858,-0.0285237432,0.0143380929,-0.3257662654,0.1837539524,0.0812143832,-0.2645001709,-0.1246016473,0.0365919322,0.1752195954,0.0625497475,-0.0666665062,0.5093409419,-0.3066211343,-0.0443679951,-0.186389178,-0.2372592837,-0.0133146346,0.0984589681,0.0109574469,-0.1552259326,0.1962587088,-0.2384228259,-0.0418432914,0.3409346342,0.0618557669,0.1764806211,-0.0300016552,-0.1084779575,0.0311522558,-0.1847686023,0.0894537643,-0.1647641063,0.1790914685,0.2616814673,0.1716295481,0.30290398,-0.5096773505,0.0224628635,0.175759986,0.0812191293,0.1270955652,0.1700863391,-0.4266649485,-0.1730842143,-0.0970294401,0.1823716313,0.8289055228,0.1243809313,0.2062119246,-0.0796269923,-0.0050468333,-0.2430296987,0.331882298,-0.0344498418,0.0030591802,0.2680387199,0.260374397,-0.0785360411,-0.103666544,0.3687556684,0.0533525907,-0.1344305277,-0.3210855722,0.1311758608,-0.2336866409,0.1769188493,-0.3360182047,-0.3163892031,-0.0478822254,-0.3653653562,-0.4228371382,0.48461169,0.0054239803,-0.2313988209,-0.0186035931,0.3134132922,-0.0864253417,-0.1130876541,-0.0460065939,0.3021612167,-0.4040479362,-0.0170489829,0.0174638499,-0.2499818206,0.0296230931,0.1452656239,-0.0594116002,-0.0046151374,-0.0322169513,-0.1490756124,-0.087769568,0.1386556476,-0.2469571084,0.1380443126,-0.0049916459,-0.3445608616,0.0177875254,0.4462214112,0.0750546604,0.4430394173,0.3652442396,-0.2264184207,-0.3506116867,-0.0410537869,-0.3046951592,-0.2401395589,0.1040015593,0.0301224757,0.074003689,0.1952189058,0.0917406306,-0.1269474626,0.0820422471,0.1156724319,-0.1546797603,-0.3211847246,0.6022117734,-0.0352159701,-0.3301559687,-0.0542485863,0.084681645,-0.2458117455,0.1921924949,-0.5279544592,0.2095107138,-0.5149428844,-0.0037138499,-0.2298366427,0.1102603674,-0.0887409002,0.0611490235,-0.013563483,-0.1757887304,-0.1614419818,0.00714926,0.0188188422,0.3083603978,-0.1789971441,-0.1153696626,0.1698780805,-0.1366402507,0.496100992,-0.0113820368,0.3250823021,0.1795650274,0.5144249797,-0.2075134069,-0.4881798625,-0.216506198,-0.2150973529,0.0731395259,0.0451912582,0.0922927558,-0.0848304927,-0.101517044,0.2298579961,0.0924943909,-0.2157504261,0.1363271028,-0.2266135812,-0.0476102084,-0.0163082927,-0.0536607765,-0.0955723971,-0.1286571175,-0.2513400018,0.0488615185,0.030549407,-0.0572599061,-0.2592467368,0.014791606,-0.5861258507,0.108908914,-0.1426069587,0.4584695697,0.1247305721,-0.3492514193,0.077487655,-0.0040636505,0.8123036027,0.3860254586,-0.4249907434,0.2087303251,0.161239177,-0.4800817668,-0.1180629283,0.0392922759,0.1970301121,0.1414536387,0.1690787673,-0.1643148065,0.0010853469,0.0077003189,0.4012598097,-0.2055786103,-0.1325112283,-0.3660716414,-0.1706568897,-0.1111255139,-0.3518340588,0.0683570951,0.3903689086,-0.2050547302,-0.3221436739,-0.0270483773,-0.1736686081,0.2890754938,-0.2555242777,0.4978435636,0.0639476404,0.3590512872,-0.0990240574,0.2204165459,0.2005787045,0.6041920781,0.1415048689,-0.1160833314,0.2325547636,-0.2334441096,0.4253969193,0.15366745,-0.1774602979,0.1846618354,-0.3063538373,0.0623705797,-0.6292974353,0.1372658908,0.0835631117,0.1273139864,-0.6609066725,-0.2396929711,0.2759853899,0.1260470599,-0.0093756327,0.5547523499,-0.4829413295,-0.4192562699,0.0228494927,-0.2087607533,0.8359116912,0.0784083381,0.2901652455,-0.0413948707,-0.1757697314,0.16838862,0.1147715449,0.2761633396,-0.1819667667,-0.1819718182,0.1086787507,-0.2836016119,0.0008317158,-0.1632947177,-0.4674455523,0.0696807876,-0.5994870067,-0.0358432755,-0.0585695915,0.0280443728,-0.1377072483,-0.2819499969,0.0484310798,0.1442808956,0.3401872814,0.0152559718,0.0530012697,-0.0407070294,-0.2414348423,-0.2039785534,-0.2047843039,0.3587834835,-0.1134456843,0.4281314015,-0.1926320642,-0.3019204438,-0.1091828793,0.1882483661,0.3401083648,0.2536093891,-0.0197980125,-0.1888896823,-0.0746568814,0.3582473397,0.1298969835,-0.2858204544,-0.0109678404,0.2939290106,-0.0674559325,-0.0751140267,-0.0013207556,-0.2357942015,-0.3105884194,0.3293305933,0.02666061,-0.2815686464,-0.1262613386,0.2754585445,0.3877061605,-0.1352788508,0.1117381155,0.2743307054,-0.0859046355,0.2338092327,-0.3917926848,-0.3844957352,-0.1263804287,0.3706642091,0.2973603606,-0.0862310603,0.3365399241,-0.0669705495,-0.0116312252,-0.0913813561,0.2213664055,-0.0240565632,-0.1430523843,0.2779606879,-0.1249851435,0.0069170427,-0.1064779088,0.2098962963,-0.0299549475,-0.0986287221,-0.0535643175,-0.2412130684,-0.3641799688,-0.0453237519,0.0391194746,0.2200305164,0.038252987,-0.0242585,-0.0460399464,0.1655549109,-0.2522574961,-0.1117301285,-0.1623183489,0.1536125541,-0.0723139271,0.0969959646,0.1246826723,-0.0116844745,0.1176169366,0.113846682,0.1276338696,-0.1913759112,0.0300988574,0.1258280724,0.0010018613,-0.1368049532,-0.1848359853,-0.416451335,0.013141403,-0.362434417,0.1861450225,0.2538997829,-0.0068259505,-0.1294531524,-0.0106560355,0.0510179773,-0.0291851517,0.0000012882,-0.1761193573,-0.0423247926,-0.06870763,0.1989370286,-0.0383233428,0.0228256043,-0.5458562374,0.0885853767,0.1324744374,-0.0169544034,-0.0168429073,-0.2495429218,-0.1824599802,-0.1766408235,0.1295222491,0.422098577,-0.2391796559,0.0384817235,-0.0376551263,0.1606242806,0.1274296343,-0.3098876178,-0.2013649195,-0.0710810348,-0.2180210948,0.1810366213,0.1253010482,-0.3079228699,-0.297603786,-0.0918723345,0.5634431243,0.1357669681,0.296987772,0.2297581285,0.1857327521,0.1896266341,0.0608837046,0.1428316832,0.009580018,0.2860203683,-0.033013016,0.2338412404,0.3007438481,-0.2161950916,-0.2589811087,-0.4246033132,0.2635904551,0.4456207156,-0.1426684558,0.2175603062,0.0316087678,0.0760748535,-0.4294070005,-0.2934633791,-0.0733328164,0.1539726704,-0.2382424772,-0.1437871605,0.2208970338,-0.087860316,0.0379114933,0.0662917048,0.2104749978,-0.2721458077,0.4234369099,0.0972390994,-0.1549433917,-0.2214446217,0.2564543784,0.4673560858,-0.0163847078,-0.3316155374,0.0328796841,0.0106146792,0.0827289522,0.0978991538,0.3364855945,0.4978487194,0.6439464688,0.2906396687,0.1160516813,0.2907954752,-0.089612104,-0.2331551611,0.2924584448,-0.1056129038,-0.1254767179,0.0711359903,0.1555305272,-0.0977445915,0.1009698138,0.3750605583,0.196749419,-0.4795674682,0.1680236012,-0.0276461802,-0.2430644929,0.0472542904,0.0190920811,-0.2540749907,-0.009599709,0.5448389053,-0.2125525177,0.1086790562,0.0513919443,0.0601780042,0.2609449923,0.2647172809,0.1125678569,-0.0039215414,-0.3516901135,-0.3238888979,-0.2068302333,0.2484186292,0.1412200183,0.3366371095,-0.1144526526,0.2920978069,0.0077583562,-0.0554190464,-0.1685561985,0.148997739,-0.1944966465,-0.0188313145,-0.4414308369,-0.1060183495,0.2233436108,0.0613439232,-0.0632368475,-0.3611062169,0.4058023095,-0.1078401506,0.0809046552,-0.0746583194,0.2291673273,0.2022636533,0.3237789273,0.3758161068,0.2293279916,0.2445682585,-0.2135000229,-0.1909113973,-0.2504677176,-0.2796136141,-0.0660589263,-0.1707462519,-0.0220373608,0.044608254,-0.4420309961,0.1230458394,-0.0918971673,0.0272557996,0.2325251549,0.151297152,-0.0656670928,0.0845914185,0.0432244949,0.183258906,0.2209240347,0.1861538738,-0.0906412527,0.3497987986,-0.1204052716,-0.2887835801,0.0790932402,0.0344266444,-0.2711871266,-0.1265544295,0.2884162068,-0.0720805898,0.1832310259,-0.6375777721,-0.1870618612,0.3745397925,-0.1326142251,-0.4474212527,0.5150055289,-0.185078606,-0.1208837554,-0.0386036783,0.3966582119,0.0742228851,-0.4025908411,0.2796815634,-0.25908795]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2275","title":"SNLI dataset has labels of -1 ","comments":"Hi @puzzler10, \r\nThose examples where `gold_label` field was empty, -1 label was alloted to it. In order to remove it you can filter the samples from train\/val\/test splits. Here's how you can drop those rows from the dataset:\r\n`dataset = load_dataset(\"snli\")`\r\n`dataset_test_filter = dataset['test'].filter(lambda example: example['label'] != -1)`\r\n\r\nI agree it should have been mentioned in the documentation. I'll raise a PR regarding the same. Thanks for pointing out!","body":"There are a number of rows with a label of -1 in the SNLI dataset. The dataset descriptions [here](https:\/\/nlp.stanford.edu\/projects\/snli\/) and [here](https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/snli) don't list -1 as a label possibility, and neither does the dataset viewer. As examples, see index 107 or 124 of the test set.\r\n\r\nIt isn't clear what these labels mean. I found a [line of code](https:\/\/github.com\/huggingface\/datasets\/blob\/80e59ef178d3bb2090d091bc32315c655eb0633d\/datasets\/snli\/snli.py#L94) that seems to put them in but it seems still unclear why they are there. The current workaround is to just drop the rows from any model being trained. \r\n\r\nPerhaps the documentation should be updated.","comment_length":69,"text":"SNLI dataset has labels of -1 \nThere are a number of rows with a label of -1 in the SNLI dataset. The dataset descriptions [here](https:\/\/nlp.stanford.edu\/projects\/snli\/) and [here](https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/snli) don't list -1 as a label possibility, and neither does the dataset viewer. As examples, see index 107 or 124 of the test set.\r\n\r\nIt isn't clear what these labels mean. I found a [line of code](https:\/\/github.com\/huggingface\/datasets\/blob\/80e59ef178d3bb2090d091bc32315c655eb0633d\/datasets\/snli\/snli.py#L94) that seems to put them in but it seems still unclear why they are there. The current workaround is to just drop the rows from any model being trained. \r\n\r\nPerhaps the documentation should be updated.\nHi @puzzler10, \r\nThose examples where `gold_label` field was empty, -1 label was alloted to it. In order to remove it you can filter the samples from train\/val\/test splits. Here's how you can drop those rows from the dataset:\r\n`dataset = load_dataset(\"snli\")`\r\n`dataset_test_filter = dataset['test'].filter(lambda example: example['label'] != -1)`\r\n\r\nI agree it should have been mentioned in the documentation. I'll raise a PR regarding the same. Thanks for pointing out!","embeddings":[0.2989758253,-0.3761280775,-0.0328568369,0.2222822458,0.0694199353,0.0815882981,0.3312395513,0.1691942811,0.181328088,0.2444586903,-0.3545390666,0.4890278876,-0.1160793751,0.264321506,-0.0070854737,0.1549568772,0.2352825403,0.303981185,0.2841205299,-0.3298657238,-0.2717858553,0.0914113894,-0.405192703,0.2479602247,-0.3731481433,0.0205870774,0.1299512088,-0.0096964277,0.0335911103,-0.5129570961,0.0948818326,0.1173787862,-0.0772679076,-0.0724165887,-0.0001088527,-0.2403392494,0.0697496906,-0.047023192,-0.6223359704,-0.2529820502,-0.1181870475,0.2077092826,0.0751572996,0.0172909703,-0.0961224213,-0.052774895,0.0744386092,0.1242415681,-0.0809797347,0.3326574266,0.1794922054,0.3804579675,-0.1768749058,0.0707381293,0.3439856172,-0.1560855955,0.017882321,0.2107768506,-0.0133288531,0.0425255261,0.2260413021,0.5354082584,-0.0452154428,0.0348852649,0.3393327594,0.022858398,-0.0777820721,-0.4266029894,0.0421027653,0.3638975322,0.0810523927,-0.230555594,-0.3671469688,-0.2749627829,0.2244778425,-0.2685078979,0.0582703799,0.2668513954,0.0697489232,0.2614788413,-0.1460727155,-0.0509525016,-0.175192982,0.0005081995,-0.1057194099,0.6413038969,-0.1593694389,0.1289835721,0.5075032115,0.0728355199,-0.4432066679,0.0394593552,0.1568491906,0.4317371845,-0.569457233,-0.1717996895,-0.0282681026,0.1632059813,0.0481878258,0.2185510546,0.0437853001,0.0269279908,-0.0140581802,0.2916793823,-0.0694846883,0.1436824799,0.3065781593,-0.0466930643,0.211473763,0.0162288174,0.3600986302,0.0481924452,-0.2132719457,-0.139418155,0.1308576465,-0.2398141474,0.1998532116,-0.2626758516,-0.6473467946,0.2097320706,-0.3907648623,0.1524562687,0.086687237,0.0645216629,0.2203213722,-0.0598432161,-0.0791026875,0.0456990339,-0.1378145665,-0.4292081594,-0.1970970631,0.0903697908,-0.0988867879,-0.1167682186,0.2107839584,-0.3131540716,0.3315541744,0.0532458797,-0.2020506561,-0.0207850803,0.1787769049,0.1645315588,0.2473968714,0.5771651268,-0.3488188982,0.3330381811,0.0659814328,-0.196420148,-0.1017569751,-0.0393126309,-0.1723008603,-0.2989450991,-0.0818260387,0.1699967086,-0.1952711642,-0.1519026756,-0.0912291631,0.0878947675,-0.1778071076,0.1813961267,0.2566719353,-0.2120781988,-0.0585161708,-0.0010470236,0.3605412543,0.1575626135,-0.4851129651,-0.2034799308,0.1237736493,-0.4227319062,-0.0173462499,0.2736925185,0.0932890847,-0.1971938461,-0.0713968575,0.252091527,0.2139832377,-0.2968644202,-0.3827925324,0.0271746255,-0.3269345462,-0.1678779125,-0.1895414591,0.3081360459,-0.1457919627,0.1059518978,-0.0698814094,-0.2019356936,0.0900856256,-0.2835516632,-0.5512373447,-0.2125910074,0.2463142276,0.3127197921,-0.0492930077,0.0508235693,-0.1639029682,-0.0070770928,0.2720735669,0.0131315337,0.2313565612,0.1837443113,0.430899471,0.1810058206,0.0823249146,-0.0849987566,-0.3912090957,0.2321951538,0.0310569219,0.3658102751,0.1825841963,-0.2678924203,-0.2410059422,-0.2028992325,-0.2412320226,-0.2026890516,0.0869416073,0.0619796813,0.2216905057,-0.0590649284,-0.3432810307,0.0738289878,-0.3657287955,0.2240730226,-0.4974218607,-0.1394050717,0.1059826463,0.1703776419,-0.1773592383,0.2051088363,0.2067948431,0.0261948612,-0.0004640424,0.3204998672,-0.1998345405,-0.2124389261,0.230879724,0.4564017951,0.1108060181,-0.4229474962,-0.0603930503,0.3066322505,0.0346611701,0.0723259747,-0.3624300063,0.3699859083,-0.1780641973,0.1303417832,-0.157143116,0.1258876771,0.103882812,0.0623039193,-0.3292539716,-0.2804180086,-0.1012573317,-0.3951129913,0.3665444851,0.1523706019,-0.6327460408,0.1134551987,0.0347607322,-0.3291116953,-0.08874356,0.0729339942,-0.2424841821,-0.1196305528,0.3244645596,0.2411278337,0.1161471382,0.2301638573,0.1370828748,0.1427655667,-0.3513398767,-0.2686562538,0.0397718363,0.1523792148,-0.1068359688,-0.1208846271,0.1304067224,-0.0173292551,-0.3405174911,-0.2308853567,-0.0885479227,0.2723142207,-0.4020915031,0.0290394183,-0.2172335535,-0.1394995749,-0.3488059044,-0.0022709807,-0.2464655787,-0.2919360399,0.31280002,0.004745502,-0.203044042,0.1004573926,0.0316309668,0.5093470216,-0.1519939601,0.2992749512,-0.3002235591,-0.2652912438,-0.2074290961,0.0390334986,-0.2587574124,0.2907176912,0.3018003702,-0.1576564014,-0.333899647,-0.0033467682,-0.3023775518,0.0480288491,-0.1426198483,0.1002896801,0.1844928414,0.0393273644,-0.0511753,0.3166732192,0.0049948175,-0.2142882794,-0.1116836742,-0.0925555825,-0.0810804591,0.1272654086,-0.5270533562,-0.3587163091,-0.1944821924,-0.0670323074,0.2488209158,-0.199238494,0.0729883239,0.3046570718,-0.0580612198,0.30747962,-0.1491309702,0.1342582852,-0.1992041171,-0.4086114168,0.1899916232,0.0326557942,-0.2425484657,0.1930107474,0.2684447169,-0.0160765853,-0.2244117707,-0.3304808736,-0.0894588381,0.1758215427,0.0823932365,0.1061904654,-0.0795192942,-0.1110929474,0.0465240292,0.0067266622,-0.2253056914,0.0642919987,0.3514791429,-0.4254739881,0.2542331219,-0.1100141183,0.1659719795,0.010234803,0.2668569982,0.4718406796,0.0371448435,-0.1448502392,0.02754738,0.1412937492,-0.0483857915,0.174921751,0.1657844931,0.2737950683,-0.0954464749,0.4123096764,0.0016736649,0.0519004501,0.0887970105,-0.0069065201,0.0350570567,-0.1055696607,-0.1193992347,-0.0342235789,-0.0516616218,-0.002548394,-0.0705692694,-0.0076269954,-0.0360234603,-0.1251480728,0.1299681664,0.0240500402,0.0580964237,-0.3151971102,0.0465812311,-0.2711879909,-0.0921700448,-0.1488165855,0.2241202444,0.0887396708,-0.0866958201,0.4816256464,0.2233085781,0.6515523791,-0.2569717765,0.1707719415,0.1465686262,0.3238312304,-0.2031408101,-0.2338120937,-0.4656892717,-0.286024332,0.3575042784,0.319435358,-0.1423304379,-0.5077242255,0.4295351207,0.0348507464,-0.1356572807,0.0319492817,-0.2754111588,-0.2216804922,-0.0745625794,0.2017955333,0.0877771303,0.0055557871,-0.2659326792,0.1474954486,0.2062808722,0.1089142114,0.4171215296,0.3996474445,0.0480077863,0.1336193532,-0.0788364485,0.4705051184,0.2033641338,0.1280690283,0.5412425995,-0.0320213586,-0.4326098561,0.1952376664,-0.30174914,0.4824659228,0.271944046,0.1515711397,0.0913848281,-0.24348557,0.3418335021,-0.255014956,0.0671939552,0.268576324,-0.05620116,-0.1702016294,-0.2825742662,0.5954551697,-0.0494576171,-0.3141204417,0.0247504879,0.4294599593,-0.236809954,0.2148871124,0.3691088259,1.0459475517,-0.0521243699,0.0110982461,0.3327120543,-0.5709407926,0.5602008104,-0.1948108971,-0.0722167045,-0.2561849356,0.0398421884,0.123963438,0.2006249279,-0.1609837115,0.4800284803,0.1506713182,0.1787363142,-0.2310985029,0.183551684,0.1390640587,-0.0544175282,0.2079624385,-0.0026003963,-0.2688750923,0.1545749605,0.0909163207,0.137251243,-0.1643724293,0.0725014806,0.0128596518,-0.0502959341,-0.0006518889,-0.2077049315,0.2996355891,-0.016824251,0.1648656577,-0.0320592783,-0.1578754634,0.3023853004,0.6914594173,-0.1512018144,-0.0364852026,0.0953297839,0.0168681405,0.3420286477,0.0187349599,0.1816219687,0.567946732,0.0848258585,-0.3281464279,-0.0055455812,0.0921958685,0.1986774057,-0.3963276148,-0.0232436024,-0.2978699505,0.109867394,0.0510895029,-0.2207343578,0.2528729141,-0.1529622227,0.164469108,0.0812319517,-0.2304650992,-0.2184703499,0.0326976925,-0.4425136447,-0.2420793623,0.2558957338,-0.1108561978,0.2954883277,0.1164970174,-0.2047530264,-0.1425510794,-0.2471474111,0.0117650004,0.4294540584,-0.6151472926,0.1111146286,0.2049765438,-0.10366752,0.030526014,0.3982726336,0.0723205209,-0.0505544655,-0.0751855522,-0.1825867742,-0.2546253204,0.0325339921,-0.0457650684,0.4078810811,-0.0299972221,0.2271571308,-0.1240512729,-0.1573385,-0.3544267118,-0.0750734955,-0.4251217842,0.2522546053,0.1342241913,0.3542677462,-0.0405721292,-0.0353892073,0.0837234333,-0.0655838922,-0.0464521162,-0.1373874843,-0.3091201782,0.1497508287,0.0685371459,-0.1972481161,-0.0483176783,0.2364224792,-0.0852420852,-0.0070723095,0.3630233705,0.237792179,0.3187184334,0.863564074,-0.0884645209,0.2512857616,0.0050875149,-0.1277726889,0.1021060348,0.0536886007,0.1734299958,-0.0303517822,-0.18726632,-0.0045387805,-0.1407217532,0.1538668871,-0.2063364536,-0.1397120804,0.0252372008,-0.2607496679,0.284309268,-0.0573380701,0.3268589675,0.1496073157,-0.195259124,-0.2138921767,0.2366321087,0.2653482258,-0.1799706519,0.0295334477,0.2653588057,0.0234092772,0.0868367776,0.1953811646,0.1683211476,0.0460222177,0.0783969909,-0.0725749806,0.2347391844,-0.2773815691,0.100228779,0.3798257411,-0.1474040449,0.0760474503,0.4589995444,-0.1008728072,0.1375338733,0.3812130094,-0.0973137096,0.6287071705,-0.2104889601,0.30151546,-0.1653841287,0.0593548194,-0.0893737525,0.2047702223,-0.3985283673,0.1160482243,-0.0631506518,0.317056179,-0.2719576359,-0.2514201999,0.1825874895,0.3171639442,-0.2034422755,-0.36770913,-0.6252554655,-0.093905285,-0.1328828335,0.300460726,-0.1347045004,-0.1321982592,0.2144760191,-0.0496293232,-0.116557464,-0.3096385002,-0.3505442739,0.2089573741,0.4235255122,0.0082153948,-0.0533895306,0.1345664114,0.1074732468,0.2328767627,-0.0599111393,0.2295925617,0.0902943611,0.3245148957,-0.0794222653,-0.1306712031,-0.1107167974,-0.1173749119,0.1408378929,0.1603417099,-0.0306679532,0.1941063553,0.0937063396,-0.1574204117,0.3419394195,-0.0722276568,0.4403964877,-0.3386833072,0.1590570956,-0.1795813292,-0.290394634,-0.2629449069,-0.1401441842,-0.0519083962,0.3314613104,-0.0606240742,0.2896050811,0.1731745005,-0.1960676908,0.0994277075,0.2572852373,0.3898850977,0.0960836634,-0.0798738971,-0.1623767316,-0.0282623507,-0.8415147662,-0.0190440267,-0.0237928033,-0.0034853979,0.2146442235,0.0383529589,-0.2455152124,-0.2468423992,-0.0246144999,-0.0062903836,0.1036540121,0.0396053232,-0.0095444499,0.0344998091,-0.2680270374,-0.0806607008,0.1091817692,-0.0008418867,0.060521841,-0.1485964954,-0.0186847504,-0.0961735547,0.3204795718,-0.1085058004,0.1731996685,0.123069413,-0.0154400989,0.3519541919,0.2310056239,-0.07901752,-0.3477729559,0.1440770328,-0.4776942432,-0.0361928269,-0.224714905,0.3466446102,0.0888938084,-0.287078917,-0.0480909757,0.1668858975,0.0773954615,-0.0106210168,-0.5227891207,0.2769576609,-0.1839490235,0.1439634115,0.2457573414,0.1054013371,0.1986226588,-0.119909361,-0.3376211226,-0.2967925966,0.4267443717,-0.3606746793,-0.0565035753,0.0524062626,0.3034640253,0.4309638739,-0.3343274891,-0.9000049829,-0.0720251352,0.2416552901,-0.1272199899,-0.4244158268,0.2112102509,0.2320234329,0.090948388,-0.100270465,0.0472114794,0.3323106766,0.0538824908,0.3208736181,-0.4091674984]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2272","title":"Bug in Dataset.class_encode_column","comments":"This has been fixed in this commit: https:\/\/github.com\/huggingface\/datasets\/pull\/2254\/commits\/88676c930216cd4cc31741b99827b477d2b46cb6\r\n\r\nIt was introduced in #2246 : using map with `input_columns` doesn't return the other columns anymore","body":"## Describe the bug\r\n\r\nAll the rest of the columns except the one passed to `Dataset.class_encode_column` are discarded.\r\n\r\n## Expected results\r\n\r\nAll the original columns should be kept.\r\n\r\nThis needs regression tests.\r\n","comment_length":24,"text":"Bug in Dataset.class_encode_column\n## Describe the bug\r\n\r\nAll the rest of the columns except the one passed to `Dataset.class_encode_column` are discarded.\r\n\r\n## Expected results\r\n\r\nAll the original columns should be kept.\r\n\r\nThis needs regression tests.\r\n\nThis has been fixed in this commit: https:\/\/github.com\/huggingface\/datasets\/pull\/2254\/commits\/88676c930216cd4cc31741b99827b477d2b46cb6\r\n\r\nIt was introduced in #2246 : using map with `input_columns` doesn't return the other columns anymore","embeddings":[-0.0357847661,-0.1575217843,-0.0868108794,0.2376332283,0.5289875269,0.1175254658,0.5686455369,0.3704491556,0.2597455382,0.1947687119,-0.1332879066,0.54302001,0.0994263068,0.3341946304,0.0166783128,-0.1077984869,0.1042396128,0.1556778699,-0.5332588553,-0.2485715598,-0.3539150953,0.1785128117,-0.189593628,0.0363146365,-0.0856321976,0.2268483937,-0.0752511993,0.0718287379,-0.1968088746,0.0467561819,0.2146316022,-0.1659525633,-0.1237033755,0.077156879,-0.0000990271,0.0114663513,0.0859174654,-0.0091559635,-0.1566771418,0.0474112406,-0.1982406974,-0.0677877367,-0.0882979035,-0.2672233284,-0.1891156733,-0.2229386121,-0.1765221953,-0.2252686918,0.1340979189,0.0211990178,0.3522591293,0.2214994878,-0.1476594359,-0.0600136518,0.2655742168,0.217487067,-0.0632185414,-0.1774149835,0.0207797941,-0.0078307251,-0.0524007455,0.4466290176,0.0209640879,-0.1336341947,-0.0051338705,0.203453064,-0.0406310447,-0.0303325169,0.3821034133,0.1162569374,0.2409487814,-0.0864889845,-0.1750107706,-0.2274489552,0.0348636024,-0.2414517403,0.1153323129,-0.0931093097,0.1492735445,0.1759854406,-0.38226071,-0.0089526512,-0.0970892534,0.1939299107,-0.283518225,0.1710606366,-0.2821823657,0.0852564499,0.0107296398,0.0057511083,-0.2462015003,-0.1706468612,-0.0102349473,0.1799632013,-0.0588390231,-0.2139725536,0.0842067003,0.162787661,0.2372650951,-0.1169941127,-0.0369562693,0.2024584413,-0.0170683879,0.1216709465,0.3319074214,0.1637552977,0.1191615686,0.3312908411,0.2741263211,0.0048443577,0.0452919528,-0.1319956183,0.3326645195,0.2048241645,0.083530806,-0.113155812,0.0843298137,-0.2055879533,-0.1399455965,0.1899603307,0.0204549339,-0.0261583794,-0.0671980083,0.3910754621,0.0939425975,0.1522224694,0.0091073848,0.1897300631,0.0487884507,-0.1829747856,-0.4172065556,-0.0833830386,-0.1617066562,-0.1496147066,0.0571537018,-0.0215368681,0.3924635053,0.2146307975,-0.1008040458,-0.0021817882,-0.1540977359,-0.0079119075,0.4681972563,0.0567120984,-0.1498759538,0.2009750605,0.1821885258,-0.2646003664,0.0076709664,-0.0745368004,-0.1921508163,-0.0650472268,-0.1382558644,0.3105876446,0.0283279158,0.1095171571,-0.1155268252,0.2912951708,0.3388133943,-0.175012961,0.1155868396,-0.3287636936,-0.2149400711,-0.1349684894,0.1861959696,0.069545947,-0.2191279531,-0.2320273221,0.3701491952,-0.1202198938,0.0886877477,0.2754109204,-0.1826442331,-0.0655736551,-0.1955094934,0.2797731757,0.1806037277,-0.083446838,-0.4100649953,0.1806074828,0.097192347,0.0853633285,-0.1506409347,-0.0745889172,0.2699152827,-0.0252983179,0.0880026147,0.1082192063,0.0613477007,0.1780363172,-0.1506055444,-0.0585445203,0.1019947827,-0.1354887635,-0.0785999149,0.1987272352,0.1226853356,-0.1103497744,0.3551034033,-0.2854625881,0.165890336,0.2804603577,0.0421767086,-0.1347810328,0.0597593002,-0.2607401311,-0.4284385741,-0.0199495479,0.2042558193,-0.1423262358,-0.1081608459,-0.3205409646,-0.2251669466,-0.0325765759,-0.1822501272,0.0456575565,0.255636692,0.0194956847,0.019729564,0.0542693362,-0.2197064161,0.1615279764,-0.2340631187,0.0589018464,-0.4512899518,0.1625316143,0.0268572047,0.0533040054,-0.0592216887,0.2286079675,0.1907048225,-0.1196359992,0.0058782022,0.16816625,-0.200497523,0.1012768596,-0.2552154362,0.0251861475,0.2435627282,0.063422665,-0.007960083,-0.1525691003,0.1898518801,-0.0640749186,-0.3242453635,0.3277555704,0.1052784696,0.0502937287,0.0834908485,-0.0356507935,0.0753540844,-0.1134135425,-0.2598536313,-0.4904225469,-0.1066021249,-0.1074671745,-0.011583305,0.0517564677,-0.3236914575,0.1402301788,0.9188784361,-0.0623417795,-0.0147714987,-0.0939866975,-0.3445657492,0.0729151815,0.1259706467,0.3295140862,0.3965092301,0.2283821553,0.2659594119,0.0557789616,-0.1804527491,-0.0716970339,0.2371605188,0.0097744968,-0.0305779949,0.2169628143,0.2506073415,-0.0417708606,-0.5560484529,0.4048532844,0.0905832201,0.0257220827,-0.3956535459,0.0577441193,-0.5252432227,-0.0719339624,0.0748953447,-0.0868828818,0.1610769182,-0.4229637384,0.0589337647,0.0997522622,-0.4003640413,0.1637859195,-0.0271169152,-0.0133928992,0.0148535231,0.0539087318,-0.1009223387,-0.1556568444,-0.0931526572,0.1648465395,-0.0498673879,-0.1709077507,0.1399349868,-0.0500380546,-0.2620826364,-0.28654176,-0.4427244365,0.2523182929,-0.3048395813,0.0588300489,0.1076537743,0.1922371984,-0.3203525841,0.0516176634,0.3080891967,-0.1982468665,-0.3238245249,0.1872299761,0.1801666021,-0.1789220124,-0.3981729448,-0.4971632063,0.1300304234,-0.1350488067,0.1815885752,-0.1555291861,-0.063083142,0.0275862422,0.0029281843,0.0385403186,-0.3008101881,0.0293904543,-0.6573472023,-0.2217814475,0.3632721007,-0.2966222167,-0.5062741041,-0.1254555881,0.097908698,0.1011653394,0.0171907227,-0.3091311753,-0.3677557111,-0.0832215101,0.1842756569,-0.0637969449,0.2143393755,0.3514961302,-0.006168298,-0.1157656461,-0.1629763246,-0.1138976514,0.1560923457,0.2188880295,0.0149244191,-0.2122608274,0.4909074605,0.1053582579,0.2070960253,0.2912396789,0.0119976345,0.2866948545,-0.0825411975,0.5360159874,-0.1932152659,-0.2198107541,0.1119850427,-0.2987490594,0.2717894614,0.2111551017,0.0109972646,-0.1812853366,0.0832057297,0.2468914241,-0.0337267853,-0.3023664355,0.0729222521,0.1730727404,0.081915386,0.3885632455,0.2997320592,-0.0978402346,-0.0457863435,0.1973557919,-0.2071418017,0.0532040372,-0.1990231574,-0.5546408892,-0.1234234124,-0.4953662157,0.2916838825,0.0468144603,0.0535965301,-0.0667382628,-0.3529548347,0.0443254933,0.102957271,0.447245717,-0.2107103318,-0.105663538,-0.0093874717,0.020498503,-0.30861938,-0.0045578578,-0.2849847972,0.0043289047,0.1969816685,0.5901811719,-0.3379146755,-0.1179241389,0.4202672839,0.1328514665,-0.1366242617,0.0133246602,-0.1167633832,-0.2524291873,-0.1269264221,-0.0232580714,0.2281379551,0.1141954437,-0.117924951,-0.3082918823,0.0313449465,-0.2168823183,0.0741278455,0.1820000708,0.2910559773,0.0874118581,0.153372094,0.0328697711,0.2015525997,-0.0967381373,0.4839460552,-0.1953886449,-0.1645174772,0.1696023494,-0.3728142977,0.380232513,0.3001912832,-0.0873068273,0.0630682707,-0.0684944987,-0.0157052614,-0.218679145,-0.0503551327,0.47104761,0.1342893392,-0.0705217272,-0.2062286884,0.2522025108,-0.1254615784,-0.141760841,0.296375066,0.2017965615,-0.3386618495,0.2593962252,0.0027739855,0.6919936538,0.0131439166,0.1439873576,0.1355214566,-0.0851740539,0.2800991237,0.0058680139,0.1299838871,-0.3186072409,-0.4095243514,0.0413689315,0.1173001975,0.259893626,0.0215861928,-0.0184926428,0.1189649552,0.0529269502,0.5219535828,-0.1091255173,-0.1962874085,0.0781304538,-0.2820718288,-0.2800424993,0.2990988195,0.1022767499,-0.1271913052,0.1078757942,0.0645862743,0.0232882854,-0.2749243677,0.0634105206,0.1309594065,-0.0393980481,0.1502915025,0.241662398,-0.325486958,-0.0723647028,0.3597698212,0.3453976214,0.116304107,0.0017102532,0.0124283414,0.0512763783,0.4156190753,-0.0058010942,0.137318626,0.1973049641,0.0137810288,-0.0776174217,0.1323333681,0.0666218027,-0.0871185511,-0.1424063295,0.0675827488,0.0647140965,-0.5439127088,-0.0392957442,-0.0201069005,0.3515423834,-0.4296716154,0.225039199,0.0568185113,-0.1846995205,0.30392012,0.3186178803,-0.2402119637,-0.1534907371,0.4318380952,0.0366681032,0.1157804355,0.4951030016,-0.0813175514,-0.2604911625,-0.2637657523,0.033414036,0.3343142271,-0.2594819963,-0.0055013332,-0.1497611851,-0.0482240915,-0.083014749,0.1514921039,-0.2278896719,-0.0814110935,-0.1304176301,-0.1993328929,-0.3260846436,0.1621860862,0.1392475367,0.2407764941,0.1289318949,-0.0627673194,-0.2677092254,0.1025732756,-0.4605872035,-0.0582601503,-0.0802952573,0.2680490613,0.3867271543,0.0230644848,0.1932047457,-0.2736717761,0.3206257224,0.3222122192,-0.0779009759,-0.3718768954,-0.2599527538,0.045859836,0.1076996401,0.101824306,-0.0309779402,0.0240448453,-0.2011662722,-0.3315497935,0.3971156478,0.0532397963,0.0873995051,0.3024343848,-0.086398609,0.0604441538,0.2415836453,-0.0798689127,-0.0203662831,-0.0690585673,0.0304768346,-0.0131826624,-0.0466163456,0.0655996576,0.0671435073,0.0691497847,0.0119934492,0.0322482847,0.5499171019,-0.30133304,0.0264211651,0.2025975734,0.2565907538,-0.0716707408,-0.3626756668,0.001368675,0.1364887506,0.3815198839,-0.3571908474,-0.1150803939,0.4075306952,0.1416531652,0.1356966645,0.1534600109,0.1509614587,-0.0615384094,0.3127653897,0.0162298158,0.1870701015,-0.1838772595,0.1210737228,0.3842481673,-0.0590949506,0.1473535299,0.3182504177,0.2877363265,-0.0495941862,0.2911371887,0.0663147941,-0.0297498908,-0.1791940182,0.2915977836,-0.0968018174,-0.2994911969,0.3273139894,0.1368472278,-0.2408308983,0.3581093848,0.170830667,0.2096785754,-0.1921666861,-0.2833569348,-0.2604634464,0.3138661087,-0.4095665514,-0.2418336719,-0.1223418638,-0.1359011084,0.0042578112,0.1382559687,-0.1698545963,-0.0716608912,0.4369072318,-0.0963112935,-0.1244532168,-0.271304369,-0.0526374467,-0.1426900029,0.2615812421,0.0010815491,0.2203297019,-0.0193681251,0.0602744743,0.1005869806,0.0835645944,-0.0154712284,-0.1095956713,0.207482487,0.1429909319,-0.1010160372,-0.2637554705,-0.0882178769,0.3772134185,0.1925729215,0.0887079909,0.2937474847,0.2467878908,-0.2389924228,-0.0411044136,0.1555913836,0.4759061635,-0.2429932803,0.2295663506,-0.1152866185,-0.3539078534,-0.1549537331,-0.329559952,-0.1710232794,0.0375801958,0.1887456179,0.0451500751,0.3476096392,-0.1241079643,0.158897832,0.2408941835,0.166962862,0.0467541665,-0.4708096683,-0.1284573227,-0.2433347702,-0.5270622373,-0.0078691673,0.0818517283,-0.0226913225,-0.1237663403,0.1700597852,0.4372745156,-0.0557214357,0.1674906909,-0.2988936007,0.0277436413,-0.0554372258,-0.1381869763,-0.2429211885,-0.0012598912,-0.0526818149,0.094369486,-0.2242772728,0.180094555,0.1935569793,0.1312591732,-0.1239113808,0.3565212488,-0.5263646841,-0.3031449616,0.161817193,0.3016983271,0.0715803951,-0.2473286837,-0.0058233691,-0.0275590066,-0.1512510628,-0.0774732977,0.2821397781,-0.1465465277,0.3806613088,-0.0500939675,-0.3484113216,-0.0685103536,0.1027244106,0.2090615034,-0.0204189271,-0.2849866748,0.315435946,-0.0171894133,-0.1848484725,0.0949905291,0.227932632,-0.0925654843,0.0231071506,0.0586345233,-0.421102792,0.4729938507,-0.4269503057,-0.3738433123,0.0709978938,0.1301802397,0.1701504737,-0.2235271931,-0.6250747442,-0.0060521318,0.1710981876,-0.3379549086,-0.142561838,0.3280798793,-0.1809660494,0.1133664474,-0.1432526857,0.4612301588,0.2436235845,-0.2658311427,0.223811239,-0.2294272333]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2271","title":"Synchronize table metadata with features","comments":"See PR #2274 ","body":"**Is your feature request related to a problem? Please describe.**\r\n\r\nAs pointed out in this [comment](https:\/\/github.com\/huggingface\/datasets\/pull\/2145#discussion_r621326767):\r\n> Metadata stored in the schema is just a redundant information regarding the feature types.\r\nIt is used when calling Dataset.from_file to know which feature types to use.\r\nThese metadata are stored in the schema of the pyarrow table by using `update_metadata_with_features`.\r\nHowever this something that's almost never tested properly.\r\n\r\n**Describe the solution you'd like**\r\n\r\nWe should find a way to always make sure that the metadata (in `self.data.schema.metadata`) are synced with the actual feature types (in `self.info.features`).","comment_length":3,"text":"Synchronize table metadata with features\n**Is your feature request related to a problem? Please describe.**\r\n\r\nAs pointed out in this [comment](https:\/\/github.com\/huggingface\/datasets\/pull\/2145#discussion_r621326767):\r\n> Metadata stored in the schema is just a redundant information regarding the feature types.\r\nIt is used when calling Dataset.from_file to know which feature types to use.\r\nThese metadata are stored in the schema of the pyarrow table by using `update_metadata_with_features`.\r\nHowever this something that's almost never tested properly.\r\n\r\n**Describe the solution you'd like**\r\n\r\nWe should find a way to always make sure that the metadata (in `self.data.schema.metadata`) are synced with the actual feature types (in `self.info.features`).\nSee PR #2274 ","embeddings":[-0.2881360054,0.1180547923,-0.0211803056,0.06773258,0.1028694063,0.0052841227,0.0582284667,0.3161320388,-0.4829489887,0.09251187,-0.2342810333,0.4763721824,0.2071071118,0.4053275287,0.0442276523,0.0835805833,0.1723809689,0.2394901514,-0.0341262855,0.0822491497,-0.206051603,-0.1669519693,0.0548219793,0.1356379539,-0.4717291594,0.2137147337,-0.1756486893,0.0865190402,-0.1925352365,-0.3045569956,-0.0175290182,0.3195476532,0.0110489149,0.0880506635,-0.0001043164,0.0125347152,0.1511658132,-0.1455724984,-0.3487788737,0.0224558767,-0.0158479456,-0.0015682513,-0.0243505668,-0.2648689747,-0.2071035504,-0.1340334117,-0.1325446814,0.2959061861,-0.2695070505,-0.0094669331,0.3088229299,0.3877404928,0.2223859727,-0.1683766544,0.5512846112,0.3795274496,-0.2507541478,0.1725381166,0.025579229,-0.1389248669,-0.0836352184,0.4501466453,-0.0876277313,0.0236299466,0.4468856454,0.2651268244,-0.1040656269,0.0432814471,0.2293899357,0.093298018,0.6977802515,-0.5567823052,-0.2485389411,-0.1628249437,0.0231583342,-0.3190426826,0.2554807663,0.2397920489,0.1417824626,0.238776058,0.0691277832,-0.0520002581,-0.0215367638,-0.1184824631,-0.1425298303,0.2927440405,-0.0106465248,-0.0032003596,0.0350309014,-0.1869364381,-0.0514439382,-0.1656515598,-0.190924257,0.1369708478,0.0018582122,-0.269965291,0.0554644279,0.0183290895,0.3501729667,0.1691530198,-0.0370516852,0.1239333302,-0.212467283,0.1633651257,0.2042058706,-0.0586006232,0.1858647466,-0.1078470647,0.4003216624,0.0284366589,-0.0097025111,-0.1166795269,0.2134320736,0.2952618599,0.2974167168,-0.2367533445,0.4665791094,-0.218463555,-0.2727993727,0.0329844132,0.0114271641,-0.0891684741,-0.1515221447,0.1900088638,-0.093215026,-0.1224464327,-0.1433990449,-0.0148432702,-0.0378422253,-0.1686026603,-0.2404534072,-0.2021797597,-0.0324065723,0.2836404443,0.0640895367,-0.098185569,0.215813756,0.2486098111,0.0330070481,0.2068184465,-0.2417211831,-0.1018081009,-0.0980570763,0.3425076902,-0.1191794574,0.0051927282,-0.1595097035,0.0200326908,-0.4245601296,0.2023955286,-0.2193795294,-0.3522664905,-0.2184506059,0.2983688414,-0.2188726813,-0.0646206737,-0.2814308405,0.0292564388,0.1831764579,-0.1872867644,-0.0741998702,0.10549815,-0.1515197903,-0.3375084996,0.0123066017,0.0076611233,-0.2969875336,-0.0257737618,0.7464125156,-0.1620088816,-0.0016132504,0.2054653466,-0.2168090492,0.0416051596,-0.0819282681,0.3349730968,0.0464042388,-0.0937227756,-0.3041495383,-0.1934304237,-0.0423996858,0.0419379808,0.0143921804,0.1581041217,0.1810355335,-0.0028768142,-0.0406962633,0.3260696828,0.0142067634,0.133696571,-0.1215620637,-0.1551046371,-0.0188180599,0.0675799325,-0.333804369,-0.0230264235,0.3393026888,-0.1830267012,0.2329021245,0.0250037294,-0.0695544332,0.1942501217,0.5088823438,0.4032215774,0.0229921285,-0.3098955452,-0.3919469714,0.2513093948,0.0926643983,0.2788232565,-0.3305763304,-0.5951313972,0.0068512498,0.1779347956,-0.1967167258,-0.1117925942,0.2409357578,-0.0357994474,-0.0148149384,0.101873748,-0.1017679945,0.0130261285,-0.0996628627,0.0302328169,-0.196213305,0.3771903217,-0.0938956365,-0.2423710674,0.1775732189,0.193265602,0.0928678587,0.1112196967,-0.0568774827,0.4009172618,0.1037588939,0.0939048305,-0.1700885892,0.5726002455,0.1565981805,0.0123594878,0.0254133753,-0.2128992975,0.0942977145,-0.0799508914,-0.4386568964,0.3183524609,0.2513358593,-0.112135984,0.1609641463,-0.1232240126,0.3500747681,0.0230723042,-0.0072627072,-0.0604562573,-0.2467954308,0.1558616608,-0.2843956947,0.4195638001,-0.3810902238,0.3999258876,0.7764489651,-0.1560044885,0.1713652313,-0.1170514748,-0.2846693397,-0.2338415384,0.3565846682,-0.3278133273,0.4097588062,0.2777570486,-0.0063340198,0.0825636312,-0.2458533943,-0.0261780284,-0.0497433282,-0.11299254,0.1443508267,-0.0034186381,0.3312661052,0.0493552014,-0.3020992875,0.0982103422,-0.0294354837,0.0693605244,-0.3273467422,-0.1212126985,-0.3845911026,-0.1708239913,-0.0680791587,-0.2618167996,-0.5203061104,-0.1757748276,0.2897361815,0.4068333209,-0.1468291581,0.309330374,-0.038574148,-0.0329051912,0.0436744615,-0.2785004377,-0.1532853097,-0.2287638634,-0.0400561951,0.0750362054,0.0732795149,-0.0052951258,0.4641815722,-0.0022502425,0.1519991159,-0.6802784204,-0.473344177,0.0840288177,-0.2056848705,-0.0486513339,0.2262002826,0.0626894534,0.2781907916,-0.1447208971,0.2996420264,-0.1620816737,-0.4031334221,0.2327687293,-0.007145775,-0.1776651293,-0.1027975604,-0.1642436683,-0.1317765713,-0.2586527169,0.3394889534,0.1039418653,0.2430388778,0.3338436186,-0.2467125803,-0.3113057911,-0.4495480955,0.0442892723,-0.2910484672,-0.1248508096,0.1720740646,-0.1467701048,-0.3874178529,0.1640409827,0.2470157444,0.011176751,0.0969819501,-0.0221590903,-0.3590673208,-0.1349354833,0.6837050319,0.0349755287,0.0996851176,0.3763024211,0.248992756,-0.1639707237,0.0523946472,-0.1609209031,0.1335293502,0.153619498,-0.0907430947,-0.2001406699,0.1594102383,-0.0074113193,0.4082185328,-0.3072856367,-0.1957496554,0.2216151655,-0.2123871446,0.3044979572,-0.3032551408,0.0554274246,-0.0604760461,-0.1269134283,0.1568930298,0.3893688619,-0.0225412007,0.0486232378,-0.0496726744,-0.2010573,-0.0236200467,-0.1561306268,0.0057639191,0.2109344006,0.2635809481,0.102842249,0.0003887987,-0.019447241,-0.2165945172,0.1699026525,0.2262698114,-0.1779562086,-0.0199850947,-0.5554490685,-0.1556279659,-0.1175692454,0.1459205747,0.0021232134,-0.0070385216,0.0304995663,-0.3132740259,0.1955276877,0.0729124099,0.253665626,0.1454724818,-0.054671023,-0.1000713632,0.0661664233,-0.0093577132,-0.3289385736,-0.0207320172,0.1130430549,0.0207769386,0.6400614381,-0.4492131174,-0.2573181689,0.307040602,-0.0250773598,-0.1181173027,-0.1795246005,-0.2105956227,-0.2292033583,-0.1766012311,0.0051170127,0.1020613164,0.0355266854,-0.2350165993,-0.0485333763,-0.0270678494,0.0428155996,-0.0118836267,0.2585461736,0.1686478853,-0.0260559823,0.1178047583,-0.1448598206,0.316162318,0.5785091519,0.4808707833,0.0788376182,-0.3196054995,-0.2125454098,-0.2709399164,0.160468623,-0.0093442509,0.0478099883,0.4624549747,-0.2919570208,-0.1322780252,-0.4939796925,-0.1672905236,0.1401831657,-0.1046605036,-0.0962432101,-0.3462351859,0.6149224043,-0.0302579645,-0.0710579455,-0.2293384373,0.1140173972,-0.1686303169,0.1102095395,0.1894570142,0.8457510471,0.1275084764,0.2707035542,0.288095057,-0.1495175809,0.2505521178,-0.4978858829,0.0187607631,-0.5265628695,-0.4086212814,-0.0954308882,0.0984366164,0.1693208367,0.1123944893,-0.1294770688,0.1108301803,0.0357501358,0.4131039083,0.053419672,0.1512884647,-0.0864907876,-0.1260152757,-0.3559901118,0.1786296517,-0.1013751552,0.0156433452,-0.0460137911,-0.1344208121,-0.0526553355,-0.0423964113,0.0855489299,0.1042961106,-0.1499035954,0.1096436679,0.0425572172,0.1174283698,-0.3474507332,-0.2249004394,-0.0249483138,0.0371647812,-0.10845422,0.1062157974,0.2817443609,0.2170052081,0.0414486527,-0.0350459255,0.3293689191,-0.1038320959,-0.2002739459,0.0581161864,0.0360528119,-0.2856272459,-0.2746345997,-0.0969852731,-0.0687771961,-0.3122511506,-0.0199133269,-0.119329609,0.0235386174,-0.1862946153,0.2302729338,0.1392652988,-0.129980877,0.3052870929,-0.1592465639,-0.132174477,-0.2477110922,0.3575157821,-0.00541953,0.1954321861,0.0886364952,-0.0434438996,-0.2756918371,-0.1029109582,0.0314615592,0.123274073,-0.5028450489,-0.2503345311,0.3520486653,0.1256342679,-0.037215963,0.4547479451,0.1240689978,-0.2604032159,-0.1789672971,-0.0730282813,-0.0979611948,0.2673958242,0.1341844499,0.2278362215,0.3146077693,0.2345346063,0.0457058251,0.1299871653,-0.4657669067,-0.2373444289,0.0462498218,0.2089922279,0.2865159512,0.0657574013,-0.0542127788,-0.2466214597,0.2245371044,-0.0358305089,0.0843595713,-0.3113168776,-0.0103122313,0.0270531159,0.0570452474,0.4196210802,-0.2781018615,-0.237405166,-0.1760390848,-0.1193124354,0.3183794022,0.1268950403,-0.0346343853,0.1802446693,0.3967500925,0.0981612653,0.0672032684,0.2597390413,-0.1071970016,-0.0100817885,-0.1427340209,0.0240971074,-0.623719275,0.0238774009,0.168792665,0.2465204149,-0.0907627493,-0.0735516474,0.2183016241,-0.0879491419,-0.0873015523,0.0902017727,0.4165029526,0.1280513555,0.0046533826,-0.3966577947,0.3936778009,0.3727865815,-0.479932785,-0.3399995565,0.1198803261,-0.0947221592,0.0685214251,0.2428156137,0.4668470323,0.3719100058,-0.0338991694,0.1124114096,0.1629515737,0.0719987899,-0.2065877914,0.4341675639,0.1545177996,0.1350210458,0.5374164581,0.3496403098,0.2582811713,0.2894730866,0.1913210005,0.3428260684,0.3019797206,-0.2157768756,0.3110262156,-0.3073815703,-0.2231420428,-0.1211472899,0.293112725,-0.0845679045,0.0782494247,0.2458322048,0.1436698586,-0.2390608639,-0.2041135728,0.2708345652,-0.1052985266,0.0385499485,0.0359999985,-0.1007841602,0.1256666034,-0.1924640834,-0.1707131416,0.0638412312,0.127992034,-0.2906766236,0.1427144855,-0.1931964606,0.2291094959,0.0541823655,0.1543953717,0.0018173003,0.053101521,-0.1342222095,-0.0832557455,0.1065541282,0.1921656877,0.2035865933,0.2331084162,0.1610043049,0.1273412853,-0.2823653817,-0.0628243387,-0.0541241281,0.0770191774,-0.0962953269,-0.1063890606,0.2488613427,0.2356479168,-0.1912014186,-0.0755228847,-0.0412583128,0.1198134795,-0.1794039309,0.1888580024,0.2575950325,-0.4126694798,0.2683780789,0.0605057701,-0.0354039632,-0.2043492496,0.3461007476,0.2488128841,0.0453642197,-0.2868385911,0.1606932729,0.1086233705,0.3538560271,0.256937176,-0.0021773567,0.005224152,0.0484411977,-0.6575188637,0.1009089127,-0.1800340414,-0.2279761881,0.2322481424,0.153499037,0.2263403088,0.0394856967,0.0951284021,-0.1696953028,0.0414140821,0.0097557809,-0.2493214458,-0.0991540402,-0.3645996749,-0.2486034185,-0.1149652749,-0.1234720126,0.3196343184,0.0660425425,0.1481649578,-0.4441376626,0.2159958631,0.252526015,0.1264215112,0.2850132585,-0.1738402992,0.2099661678,0.0264970604,0.1580566466,-0.0754922628,-0.0239331331,-0.3281131089,0.0772437006,-0.2242805064,0.0636623651,0.1762051582,-0.1232104897,-0.2757920623,0.3898660839,-0.0620970167,-0.1616279781,-0.1121152639,0.03523295,-0.080341965,-0.0727276206,0.0368596204,0.178663224,0.0808179453,0.0923893377,-0.2817398012,-0.3263230324,0.4138689041,0.0445496999,-0.2350711226,0.0070933527,0.188403219,-0.0104134129,-0.0862070248,-0.5174303055,-0.0145224491,0.3597051501,0.0391372442,-0.1102216169,0.2936047018,0.0026670576,0.0011376708,-0.2047587335,0.2907767296,-0.1970711499,-0.1182257533,-0.161347568,-0.1659584343]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2267","title":"DatasetDict save load Failing test in 1.6 not in 1.5","comments":"Thanks for reporting ! We're looking into it","body":"## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n","comment_length":8,"text":"DatasetDict save load Failing test in 1.6 not in 1.5\n## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n\nThanks for reporting ! We're looking into it","embeddings":[-0.1946798116,0.1140982956,-0.0127640385,0.284942925,0.064592123,-0.0614173673,0.1323396713,0.4565899968,0.4422373772,0.1792287529,0.2358967513,0.380851388,0.0545869619,-0.0561681502,-0.1504484713,-0.0622494444,0.2574026883,0.1362523735,0.047262188,0.1269330233,-0.1914962828,0.2465310097,-0.1496895254,-0.0032418445,-0.1895998418,0.1437868327,-0.2581939399,0.2942799032,-0.3553473651,-0.6636254787,0.2940472364,0.4080147743,0.3425835669,0.5095262527,-0.0001234102,0.0544177964,0.2841644287,-0.2096645385,-0.4484797716,-0.4640186727,-0.1993898153,0.0136942323,0.2473115027,-0.1567811072,0.174591288,-0.0732848793,-0.0321101137,-0.4220353365,0.1694841832,0.3417609334,0.1407151371,0.2266554832,0.2764903307,-0.2219138741,-0.2547798157,0.3302724957,-0.0204170793,0.48288396,-0.0186927021,0.0731682926,0.0414443091,0.0742254928,-0.0195427742,-0.1017797217,0.093925342,-0.0247307811,0.162043348,-0.0513432026,0.240814209,0.1979950964,0.6499521136,-0.2013998032,-0.5342013836,0.0352447331,-0.0611302778,0.1786230356,0.3324406445,-0.0698243231,0.0839635506,0.3244956732,-0.0997817516,-0.3653488159,-0.2351392955,-0.1203212589,0.0710190833,-0.2877634466,-0.0193996113,0.1131953076,-0.0397098362,0.0214464869,0.4245892167,-0.3535504639,-0.3489905298,-0.1019979045,-0.0459269546,-0.1144007817,-0.0653342977,-0.2706740201,0.113346532,-0.0720998496,0.2517627776,0.1316185147,-0.0391709432,0.0602462031,0.2115243673,0.2749886513,0.4198958874,0.2743373513,0.0357514173,0.0714356154,-0.2357528359,0.0838931948,-0.0404101573,-0.15166004,0.2088321447,-0.1358045042,0.5703294277,-0.4004871547,-0.3274163008,0.2602038383,-0.2743057013,-0.1729606539,-0.1039793417,0.2240318358,0.09327925,0.1972419173,0.1087378412,0.3317791224,0.0468955897,0.0277815927,-0.1801332235,0.0353045389,0.0400684513,0.011786541,0.0179779567,-0.0999579355,0.267460227,0.015178483,-0.4577178359,-0.1766552329,-0.0413619019,-0.0473274663,0.2514302135,0.0383211523,-0.2181576937,-0.056682542,0.2363063097,0.1262998134,-0.2239949852,0.2120421976,0.0718438476,-0.1589062363,-0.1686418802,0.1228035986,-0.3944003284,-0.2602534592,-0.2819331884,-0.0493995957,-0.083663933,-0.2920115888,-0.1123725697,-0.3287760615,-0.237355873,-0.1225714758,0.1428985149,0.4307113886,-0.5336277485,0.1942441016,0.1860763729,-0.0290210564,0.0966063365,-0.0230782609,0.0003097686,0.1380230188,-0.1793204844,0.1278733611,0.4110462368,-0.5454348922,-0.6159434319,0.307710886,-0.0049888324,0.1782058775,-0.1437655389,-0.0217472035,0.2127342075,-0.0530057736,0.0544378124,0.0589728355,0.0534363128,-0.0575243719,-0.4947072268,-0.3328706622,0.1166524291,-0.0132593643,-0.0556115769,0.091937542,0.1581820548,-0.2094933689,0.4844789207,-0.0858414844,-0.1812547147,0.1170089394,0.4515592456,-0.0738887712,0.0274728462,-0.2316350937,-0.5508966446,0.4312414229,0.0377286226,-0.0222507045,-0.0489411242,0.0583773702,-0.4379805923,0.1203429848,-0.3182476759,0.046511244,0.0106168631,0.078376852,-0.0866948366,-0.0088240383,-0.1149406061,0.140570417,-0.1519619673,0.0698583946,0.004358958,0.336599648,-0.2037377357,-0.2205159664,0.170929715,-0.0127892671,0.2898364365,-0.2678136826,-0.3201774955,0.4959192574,0.2282747328,0.0215910152,-0.0704142451,-0.3370976448,-0.0950707644,-0.0260207877,0.1928254664,0.4983442724,0.0574062653,0.0559969395,-0.2572018802,0.3138726652,0.0657231435,0.0568134151,0.0700780526,0.1781899929,0.2946604192,-0.024419995,-0.0170402769,-0.3440820575,0.2174295187,0.018817924,0.0102335596,0.0753727928,-0.1335324794,0.1482347995,0.25894925,0.0366390534,0.1828769594,0.0873260871,0.0021229947,-0.4102320969,0.0591992363,0.3296465278,0.6812970042,-0.0120210182,-0.0898986459,-0.0542867593,-0.0393865295,0.0392726399,0.1494913846,-0.0172953773,0.2711475194,0.1783984751,0.1901718229,0.0689663813,-0.2123125196,-0.1901509315,-0.0073718987,0.051064495,-0.3820844293,0.1315279007,-0.1379302591,0.5604397058,-0.204323709,-0.5258817077,-0.0548608974,-0.3896036744,-0.1108690277,0.3982457519,-0.1051947698,0.0420946889,-0.5863333344,0.042864982,-0.0554989949,-0.325607568,-0.1113469601,-0.1961999387,-0.1263737977,-0.1553942561,0.1137430668,0.1928962618,0.2179133743,-0.1094706878,-0.1682274193,-0.4347150326,-0.3339819312,0.1541959494,-0.1012643725,0.050694488,0.1850075871,-0.0659798011,0.1797925979,-0.0064853802,0.0098090507,-0.1555842906,-0.3324509263,0.1918925792,0.2623599172,-0.1795146763,0.0405300371,-0.4744684994,-0.317047894,-0.3201007545,0.2410695106,-0.0716331676,0.0295918006,0.0510351174,0.1749565452,0.152455315,0.0156619512,0.102222845,-0.0204779282,-0.1949142367,0.1171747521,-0.0141771985,-0.4500836134,0.1088241562,0.0357693024,0.1472648084,0.2089921087,-0.2778184414,-0.1442975402,0.0047041876,0.414567858,-0.0329649821,-0.2965220213,0.4994246364,0.335587889,0.0297882631,-0.1571239531,-0.0598418005,0.1395886838,0.0997576341,0.2127372026,-0.3571592867,0.4544912279,-0.3967370093,0.2623894513,-0.0561215617,-0.1933416575,0.2192241848,0.1355145276,0.3633723259,-0.4195737839,-0.2501191199,-0.0181212332,-0.2273002863,-0.1420907974,0.0491867177,-0.1898749918,0.3078674674,-0.0113018882,-0.2713414431,-0.1212768629,-0.3088128567,-0.0883086398,-0.6213800311,0.2525457442,0.129497841,0.2512174249,-0.1520728171,0.0570483357,0.0548919328,0.1056846306,0.3619897962,-0.1298970282,-0.6689382195,0.4119164944,-0.2728250027,0.3531420827,0.0170031525,0.6950641274,0.0251035877,-0.1877912432,-0.074011609,-0.0956236944,0.3716633618,0.1301261038,0.0777470544,0.1875228435,-0.15032655,-0.6920769215,-0.1083943248,0.0503957905,0.3830853999,0.1262316108,0.3629608154,-0.0177808609,-0.1991339624,0.2198234499,0.3881291151,-0.1670640707,-0.274376452,-0.2853179872,-0.1365243644,-0.1515447497,-0.2810799479,-0.0419628099,0.2196073383,-0.2597106695,-0.0625315532,-0.5216527581,-0.3158732355,0.064146623,0.1999338716,0.4470637441,-0.3979617953,0.2182785869,0.2503279746,0.169705525,0.5388915539,0.4505360723,-0.0512378775,-0.4795738757,-0.039661333,-0.1913179606,0.0461554788,0.0478836857,-0.0462604798,-0.1848695874,0.0630236641,-0.2803143561,-0.3086195886,0.0911357701,0.0995180681,0.0440459028,-0.3165179789,-0.1771359891,0.5262101889,-0.0977220014,-0.1301518232,0.4414811134,0.3330871165,-0.236470148,0.5683749914,-0.0350576416,0.6247905493,0.2005376965,0.126446858,0.2727004886,-0.515450716,0.0626663193,0.5921412706,0.0294377375,-0.5868449211,-0.1553604603,0.0331250206,-0.12538293,0.0326819122,0.036870379,-0.3613365293,0.0716338158,0.0084466087,-0.2822274566,0.1183375046,0.3593575358,-0.2079016119,0.1589399576,-0.3033201694,0.0912196338,0.1137904301,0.1030311212,-0.0877372548,-0.0660760701,-0.1419514716,-0.3079771698,-0.1325310767,0.2350479513,-0.1199391335,0.4439127445,0.2897050679,-0.4493189156,0.2196412235,0.2174482346,0.4070452452,-0.0582864359,-0.1289931089,-0.044320479,-0.253841579,-0.0569630153,0.2750828266,-0.0208601356,0.2357589453,0.0243200008,-0.2413082123,0.0149238911,-0.2202434391,-0.3488709629,-0.2136550844,-0.0252611972,0.0917013511,-0.403341949,-0.5704155564,-0.3656395674,-0.1338036209,-0.3513588607,0.045161888,0.1369814575,0.0903127491,0.1345508993,-0.030370215,-0.3727883101,0.1286842227,0.4785127938,-0.1395132095,-0.2365066856,0.5704622269,-0.0664498061,0.1163967997,-0.2200823426,0.1668679267,0.2937476337,-0.2160442621,0.4291682541,-0.2352191359,-0.1053846255,-0.033432588,0.5656456947,0.1970895678,-0.1196097955,0.0568211302,-0.2952887118,-0.3237188458,0.1376452446,-0.038047269,0.4634029865,-0.0172834657,-0.3075003326,-0.2098487467,0.1128287613,-0.2028447539,0.0374976136,0.0159998704,0.0449076295,0.1605370641,-0.0462051779,0.1810541749,-0.311786443,0.0975406766,-0.1628718376,0.0355669707,-0.0206244513,-0.2262709588,0.1717493087,0.1430513859,0.2631170452,-0.0958966464,-0.1393290013,0.0043114931,-0.0736732259,-0.0378092565,0.3947651386,-0.1223446727,0.2778863609,0.0895340815,0.2064663917,-0.1312061846,0.2292111069,0.064420931,0.3075909615,-0.0757777393,0.2196436971,-0.0794219077,-0.0892363861,-0.2326027453,0.1826041937,0.029328797,0.0955826789,0.1823251098,-0.1802758723,-0.0411883555,0.1104308739,0.2567685246,0.5237183571,-0.1497170478,-0.2371406406,-0.0213249139,0.1600279957,-0.1601502299,-0.2017437965,0.3003623784,-0.0835295469,-0.1162904724,-0.0765228197,-0.2054964602,-0.1942750812,-0.1985163838,0.3335354626,0.5401267409,-0.0722699091,0.2756753266,0.1050364822,-0.1331185549,0.1698137224,0.4339623749,-0.026871765,0.2728070617,0.2969137728,-0.0106996065,0.4752700925,0.1081370488,0.1763004363,0.1444735676,0.029552646,0.0501384512,-0.2383726835,0.0339755304,0.0337763727,0.0379040502,0.207487449,-0.5338286757,-0.0095709851,-0.0529127605,-0.3925668597,-0.0503622741,-0.2166739851,-0.2375059426,-0.1166066527,0.1916654706,-0.0871482491,-0.1870366335,0.2569023669,0.1854578108,0.1341342032,-0.1262688935,-0.1991169453,0.2447045147,0.128683418,-0.1152322665,-0.2423141152,0.1165309399,0.0655258819,-0.0270040352,0.2843689919,0.3048303723,0.5131988525,0.4630441964,-0.1235537753,-0.0936735943,-0.0676186979,-0.0890747234,0.160659641,0.1176436469,-0.2288692296,0.1419881284,0.3497951031,0.1042521968,0.0525661483,0.3554263413,0.1828850508,0.2552399039,-0.0793646127,0.8174863458,0.0652648807,-0.401847899,0.1853944808,0.2061248571,-0.0562279373,-0.2423319072,0.2547922134,0.2379752696,0.0354117416,0.043099571,0.0429509506,-0.0761299208,0.4061383009,0.3945253193,-0.009129893,-0.253089875,0.2011501044,-0.5567246675,0.1267212033,0.1150677055,-0.0242360272,0.1735586077,0.1886163652,0.1791179031,-0.0079982691,0.1234039813,0.3020167649,-0.2655394077,-0.0732762218,-0.2039487809,-0.0483581908,0.1700234711,-0.0384172276,-0.0339166112,-0.4353083074,0.2317860276,-0.2152300924,-0.0408024713,-0.1203102246,-0.1444326192,-0.0316804796,-0.0343607739,0.4194562733,0.1286736578,0.4079838991,0.0496628247,-0.4076710939,-0.2884944081,-0.1973595172,-0.1579232812,0.2628087997,0.0948492214,0.6385340095,-0.0972695649,0.0452125445,-0.2792316675,0.2594893575,-0.1286927909,-0.0773257017,-0.2083820403,0.1652529091,0.1398079991,0.0103357537,0.2678919435,0.1165606603,-0.0676965714,0.37588045,-0.3817226589,-0.2833266556,0.4406756163,0.0210840479,-0.1905985326,-0.0162194837,0.3005398512,-0.0426337682,-0.181629017,-0.7152320743,0.0028659408,0.415509671,-0.0063177161,-0.0824656785,0.1022371128,0.2898762226,0.1865238696,-0.0089995535,0.5413072109,-0.0108727999,0.0113003505,-0.0968637392,-0.072947748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2267","title":"DatasetDict save load Failing test in 1.6 not in 1.5","comments":"I'm not able to reproduce this, do you think you can provide a code that creates a DatasetDict that has this issue when saving and reloading ?","body":"## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n","comment_length":27,"text":"DatasetDict save load Failing test in 1.6 not in 1.5\n## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n\nI'm not able to reproduce this, do you think you can provide a code that creates a DatasetDict that has this issue when saving and reloading ?","embeddings":[-0.1946798116,0.1140982956,-0.0127640385,0.284942925,0.064592123,-0.0614173673,0.1323396713,0.4565899968,0.4422373772,0.1792287529,0.2358967513,0.380851388,0.0545869619,-0.0561681502,-0.1504484713,-0.0622494444,0.2574026883,0.1362523735,0.047262188,0.1269330233,-0.1914962828,0.2465310097,-0.1496895254,-0.0032418445,-0.1895998418,0.1437868327,-0.2581939399,0.2942799032,-0.3553473651,-0.6636254787,0.2940472364,0.4080147743,0.3425835669,0.5095262527,-0.0001234102,0.0544177964,0.2841644287,-0.2096645385,-0.4484797716,-0.4640186727,-0.1993898153,0.0136942323,0.2473115027,-0.1567811072,0.174591288,-0.0732848793,-0.0321101137,-0.4220353365,0.1694841832,0.3417609334,0.1407151371,0.2266554832,0.2764903307,-0.2219138741,-0.2547798157,0.3302724957,-0.0204170793,0.48288396,-0.0186927021,0.0731682926,0.0414443091,0.0742254928,-0.0195427742,-0.1017797217,0.093925342,-0.0247307811,0.162043348,-0.0513432026,0.240814209,0.1979950964,0.6499521136,-0.2013998032,-0.5342013836,0.0352447331,-0.0611302778,0.1786230356,0.3324406445,-0.0698243231,0.0839635506,0.3244956732,-0.0997817516,-0.3653488159,-0.2351392955,-0.1203212589,0.0710190833,-0.2877634466,-0.0193996113,0.1131953076,-0.0397098362,0.0214464869,0.4245892167,-0.3535504639,-0.3489905298,-0.1019979045,-0.0459269546,-0.1144007817,-0.0653342977,-0.2706740201,0.113346532,-0.0720998496,0.2517627776,0.1316185147,-0.0391709432,0.0602462031,0.2115243673,0.2749886513,0.4198958874,0.2743373513,0.0357514173,0.0714356154,-0.2357528359,0.0838931948,-0.0404101573,-0.15166004,0.2088321447,-0.1358045042,0.5703294277,-0.4004871547,-0.3274163008,0.2602038383,-0.2743057013,-0.1729606539,-0.1039793417,0.2240318358,0.09327925,0.1972419173,0.1087378412,0.3317791224,0.0468955897,0.0277815927,-0.1801332235,0.0353045389,0.0400684513,0.011786541,0.0179779567,-0.0999579355,0.267460227,0.015178483,-0.4577178359,-0.1766552329,-0.0413619019,-0.0473274663,0.2514302135,0.0383211523,-0.2181576937,-0.056682542,0.2363063097,0.1262998134,-0.2239949852,0.2120421976,0.0718438476,-0.1589062363,-0.1686418802,0.1228035986,-0.3944003284,-0.2602534592,-0.2819331884,-0.0493995957,-0.083663933,-0.2920115888,-0.1123725697,-0.3287760615,-0.237355873,-0.1225714758,0.1428985149,0.4307113886,-0.5336277485,0.1942441016,0.1860763729,-0.0290210564,0.0966063365,-0.0230782609,0.0003097686,0.1380230188,-0.1793204844,0.1278733611,0.4110462368,-0.5454348922,-0.6159434319,0.307710886,-0.0049888324,0.1782058775,-0.1437655389,-0.0217472035,0.2127342075,-0.0530057736,0.0544378124,0.0589728355,0.0534363128,-0.0575243719,-0.4947072268,-0.3328706622,0.1166524291,-0.0132593643,-0.0556115769,0.091937542,0.1581820548,-0.2094933689,0.4844789207,-0.0858414844,-0.1812547147,0.1170089394,0.4515592456,-0.0738887712,0.0274728462,-0.2316350937,-0.5508966446,0.4312414229,0.0377286226,-0.0222507045,-0.0489411242,0.0583773702,-0.4379805923,0.1203429848,-0.3182476759,0.046511244,0.0106168631,0.078376852,-0.0866948366,-0.0088240383,-0.1149406061,0.140570417,-0.1519619673,0.0698583946,0.004358958,0.336599648,-0.2037377357,-0.2205159664,0.170929715,-0.0127892671,0.2898364365,-0.2678136826,-0.3201774955,0.4959192574,0.2282747328,0.0215910152,-0.0704142451,-0.3370976448,-0.0950707644,-0.0260207877,0.1928254664,0.4983442724,0.0574062653,0.0559969395,-0.2572018802,0.3138726652,0.0657231435,0.0568134151,0.0700780526,0.1781899929,0.2946604192,-0.024419995,-0.0170402769,-0.3440820575,0.2174295187,0.018817924,0.0102335596,0.0753727928,-0.1335324794,0.1482347995,0.25894925,0.0366390534,0.1828769594,0.0873260871,0.0021229947,-0.4102320969,0.0591992363,0.3296465278,0.6812970042,-0.0120210182,-0.0898986459,-0.0542867593,-0.0393865295,0.0392726399,0.1494913846,-0.0172953773,0.2711475194,0.1783984751,0.1901718229,0.0689663813,-0.2123125196,-0.1901509315,-0.0073718987,0.051064495,-0.3820844293,0.1315279007,-0.1379302591,0.5604397058,-0.204323709,-0.5258817077,-0.0548608974,-0.3896036744,-0.1108690277,0.3982457519,-0.1051947698,0.0420946889,-0.5863333344,0.042864982,-0.0554989949,-0.325607568,-0.1113469601,-0.1961999387,-0.1263737977,-0.1553942561,0.1137430668,0.1928962618,0.2179133743,-0.1094706878,-0.1682274193,-0.4347150326,-0.3339819312,0.1541959494,-0.1012643725,0.050694488,0.1850075871,-0.0659798011,0.1797925979,-0.0064853802,0.0098090507,-0.1555842906,-0.3324509263,0.1918925792,0.2623599172,-0.1795146763,0.0405300371,-0.4744684994,-0.317047894,-0.3201007545,0.2410695106,-0.0716331676,0.0295918006,0.0510351174,0.1749565452,0.152455315,0.0156619512,0.102222845,-0.0204779282,-0.1949142367,0.1171747521,-0.0141771985,-0.4500836134,0.1088241562,0.0357693024,0.1472648084,0.2089921087,-0.2778184414,-0.1442975402,0.0047041876,0.414567858,-0.0329649821,-0.2965220213,0.4994246364,0.335587889,0.0297882631,-0.1571239531,-0.0598418005,0.1395886838,0.0997576341,0.2127372026,-0.3571592867,0.4544912279,-0.3967370093,0.2623894513,-0.0561215617,-0.1933416575,0.2192241848,0.1355145276,0.3633723259,-0.4195737839,-0.2501191199,-0.0181212332,-0.2273002863,-0.1420907974,0.0491867177,-0.1898749918,0.3078674674,-0.0113018882,-0.2713414431,-0.1212768629,-0.3088128567,-0.0883086398,-0.6213800311,0.2525457442,0.129497841,0.2512174249,-0.1520728171,0.0570483357,0.0548919328,0.1056846306,0.3619897962,-0.1298970282,-0.6689382195,0.4119164944,-0.2728250027,0.3531420827,0.0170031525,0.6950641274,0.0251035877,-0.1877912432,-0.074011609,-0.0956236944,0.3716633618,0.1301261038,0.0777470544,0.1875228435,-0.15032655,-0.6920769215,-0.1083943248,0.0503957905,0.3830853999,0.1262316108,0.3629608154,-0.0177808609,-0.1991339624,0.2198234499,0.3881291151,-0.1670640707,-0.274376452,-0.2853179872,-0.1365243644,-0.1515447497,-0.2810799479,-0.0419628099,0.2196073383,-0.2597106695,-0.0625315532,-0.5216527581,-0.3158732355,0.064146623,0.1999338716,0.4470637441,-0.3979617953,0.2182785869,0.2503279746,0.169705525,0.5388915539,0.4505360723,-0.0512378775,-0.4795738757,-0.039661333,-0.1913179606,0.0461554788,0.0478836857,-0.0462604798,-0.1848695874,0.0630236641,-0.2803143561,-0.3086195886,0.0911357701,0.0995180681,0.0440459028,-0.3165179789,-0.1771359891,0.5262101889,-0.0977220014,-0.1301518232,0.4414811134,0.3330871165,-0.236470148,0.5683749914,-0.0350576416,0.6247905493,0.2005376965,0.126446858,0.2727004886,-0.515450716,0.0626663193,0.5921412706,0.0294377375,-0.5868449211,-0.1553604603,0.0331250206,-0.12538293,0.0326819122,0.036870379,-0.3613365293,0.0716338158,0.0084466087,-0.2822274566,0.1183375046,0.3593575358,-0.2079016119,0.1589399576,-0.3033201694,0.0912196338,0.1137904301,0.1030311212,-0.0877372548,-0.0660760701,-0.1419514716,-0.3079771698,-0.1325310767,0.2350479513,-0.1199391335,0.4439127445,0.2897050679,-0.4493189156,0.2196412235,0.2174482346,0.4070452452,-0.0582864359,-0.1289931089,-0.044320479,-0.253841579,-0.0569630153,0.2750828266,-0.0208601356,0.2357589453,0.0243200008,-0.2413082123,0.0149238911,-0.2202434391,-0.3488709629,-0.2136550844,-0.0252611972,0.0917013511,-0.403341949,-0.5704155564,-0.3656395674,-0.1338036209,-0.3513588607,0.045161888,0.1369814575,0.0903127491,0.1345508993,-0.030370215,-0.3727883101,0.1286842227,0.4785127938,-0.1395132095,-0.2365066856,0.5704622269,-0.0664498061,0.1163967997,-0.2200823426,0.1668679267,0.2937476337,-0.2160442621,0.4291682541,-0.2352191359,-0.1053846255,-0.033432588,0.5656456947,0.1970895678,-0.1196097955,0.0568211302,-0.2952887118,-0.3237188458,0.1376452446,-0.038047269,0.4634029865,-0.0172834657,-0.3075003326,-0.2098487467,0.1128287613,-0.2028447539,0.0374976136,0.0159998704,0.0449076295,0.1605370641,-0.0462051779,0.1810541749,-0.311786443,0.0975406766,-0.1628718376,0.0355669707,-0.0206244513,-0.2262709588,0.1717493087,0.1430513859,0.2631170452,-0.0958966464,-0.1393290013,0.0043114931,-0.0736732259,-0.0378092565,0.3947651386,-0.1223446727,0.2778863609,0.0895340815,0.2064663917,-0.1312061846,0.2292111069,0.064420931,0.3075909615,-0.0757777393,0.2196436971,-0.0794219077,-0.0892363861,-0.2326027453,0.1826041937,0.029328797,0.0955826789,0.1823251098,-0.1802758723,-0.0411883555,0.1104308739,0.2567685246,0.5237183571,-0.1497170478,-0.2371406406,-0.0213249139,0.1600279957,-0.1601502299,-0.2017437965,0.3003623784,-0.0835295469,-0.1162904724,-0.0765228197,-0.2054964602,-0.1942750812,-0.1985163838,0.3335354626,0.5401267409,-0.0722699091,0.2756753266,0.1050364822,-0.1331185549,0.1698137224,0.4339623749,-0.026871765,0.2728070617,0.2969137728,-0.0106996065,0.4752700925,0.1081370488,0.1763004363,0.1444735676,0.029552646,0.0501384512,-0.2383726835,0.0339755304,0.0337763727,0.0379040502,0.207487449,-0.5338286757,-0.0095709851,-0.0529127605,-0.3925668597,-0.0503622741,-0.2166739851,-0.2375059426,-0.1166066527,0.1916654706,-0.0871482491,-0.1870366335,0.2569023669,0.1854578108,0.1341342032,-0.1262688935,-0.1991169453,0.2447045147,0.128683418,-0.1152322665,-0.2423141152,0.1165309399,0.0655258819,-0.0270040352,0.2843689919,0.3048303723,0.5131988525,0.4630441964,-0.1235537753,-0.0936735943,-0.0676186979,-0.0890747234,0.160659641,0.1176436469,-0.2288692296,0.1419881284,0.3497951031,0.1042521968,0.0525661483,0.3554263413,0.1828850508,0.2552399039,-0.0793646127,0.8174863458,0.0652648807,-0.401847899,0.1853944808,0.2061248571,-0.0562279373,-0.2423319072,0.2547922134,0.2379752696,0.0354117416,0.043099571,0.0429509506,-0.0761299208,0.4061383009,0.3945253193,-0.009129893,-0.253089875,0.2011501044,-0.5567246675,0.1267212033,0.1150677055,-0.0242360272,0.1735586077,0.1886163652,0.1791179031,-0.0079982691,0.1234039813,0.3020167649,-0.2655394077,-0.0732762218,-0.2039487809,-0.0483581908,0.1700234711,-0.0384172276,-0.0339166112,-0.4353083074,0.2317860276,-0.2152300924,-0.0408024713,-0.1203102246,-0.1444326192,-0.0316804796,-0.0343607739,0.4194562733,0.1286736578,0.4079838991,0.0496628247,-0.4076710939,-0.2884944081,-0.1973595172,-0.1579232812,0.2628087997,0.0948492214,0.6385340095,-0.0972695649,0.0452125445,-0.2792316675,0.2594893575,-0.1286927909,-0.0773257017,-0.2083820403,0.1652529091,0.1398079991,0.0103357537,0.2678919435,0.1165606603,-0.0676965714,0.37588045,-0.3817226589,-0.2833266556,0.4406756163,0.0210840479,-0.1905985326,-0.0162194837,0.3005398512,-0.0426337682,-0.181629017,-0.7152320743,0.0028659408,0.415509671,-0.0063177161,-0.0824656785,0.1022371128,0.2898762226,0.1865238696,-0.0089995535,0.5413072109,-0.0108727999,0.0113003505,-0.0968637392,-0.072947748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2267","title":"DatasetDict save load Failing test in 1.6 not in 1.5","comments":"Hi, I just ran into a similar error. Here is the minimal code to reproduce:\r\n```python\r\nfrom datasets import load_dataset, DatasetDict\r\nds = load_dataset('super_glue', 'multirc')\r\n\r\nds.save_to_disk('tempds')\r\n\r\nds = DatasetDict.load_from_disk('tempds')\r\n\r\n```\r\n\r\n```bash\r\nReusing dataset super_glue (\/home\/idahl\/.cache\/huggingface\/datasets\/super_glue\/multirc\/1.0.2\/2fb163bca9085c1deb906aff20f00c242227ff704a4e8c9cfdfe820be3abfc83)\r\nTraceback (most recent call last):\r\n File \"\/home\/idahl\/eval-util-expl\/multirc\/tmp.py\", line 7, in \r\n ds = DatasetDict.load_from_disk('tempds')\r\n File \"\/home\/idahl\/miniconda3\/envs\/eval-util-expl\/lib\/python3.9\/site-packages\/datasets\/dataset_dict.py\", line 710, in load_from_disk\r\n dataset_dict[k] = Dataset.load_from_disk(dataset_dict_split_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/home\/idahl\/miniconda3\/envs\/eval-util-expl\/lib\/python3.9\/site-packages\/datasets\/arrow_dataset.py\", line 687, in load_from_disk\r\n return Dataset(\r\n File \"\/home\/idahl\/miniconda3\/envs\/eval-util-expl\/lib\/python3.9\/site-packages\/datasets\/arrow_dataset.py\", line 274, in __init__\r\n raise ValueError(\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'answer': Value(dtype='string', id=None), 'idx': {'answer': Value(dtype='int32', id=None), 'paragraph': Value(dtype='int32', id=None), 'question': Value(dtype='int32', id=None)}, 'label': ClassLabel(num_classes=2, names=['False', 'True'], names_file=None, id=None), 'paragraph': Value(dtype='string', id=None), 'question': Value(dtype='string', id=None)}\r\nwith type\r\nstruct, label: int64, paragraph: string, question: string>\r\n\r\nbut expected something like\r\n{'answer': Value(dtype='string', id=None), 'idx': {'paragraph': Value(dtype='int32', id=None), 'question': Value(dtype='int32', id=None), 'answer': Value(dtype='int32', id=None)}, 'label': Value(dtype='int64', id=None), 'paragraph': Value(dtype='string', id=None), 'question': Value(dtype='string', id=None)}\r\nwith type\r\nstruct, label: int64, paragraph: string, question: string>\r\n\r\n```\r\n\r\nThe non-matching part seems to be\r\n`'label': ClassLabel(num_classes=2, names=['False', 'True'], names_file=None, id=None),`\r\nvs \r\n`'label': Value(dtype='int64', id=None),`\r\n\r\nAnd the order in the `1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n","comment_length":226,"text":"DatasetDict save load Failing test in 1.6 not in 1.5\n## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n\nHi, I just ran into a similar error. Here is the minimal code to reproduce:\r\n```python\r\nfrom datasets import load_dataset, DatasetDict\r\nds = load_dataset('super_glue', 'multirc')\r\n\r\nds.save_to_disk('tempds')\r\n\r\nds = DatasetDict.load_from_disk('tempds')\r\n\r\n```\r\n\r\n```bash\r\nReusing dataset super_glue (\/home\/idahl\/.cache\/huggingface\/datasets\/super_glue\/multirc\/1.0.2\/2fb163bca9085c1deb906aff20f00c242227ff704a4e8c9cfdfe820be3abfc83)\r\nTraceback (most recent call last):\r\n File \"\/home\/idahl\/eval-util-expl\/multirc\/tmp.py\", line 7, in \r\n ds = DatasetDict.load_from_disk('tempds')\r\n File \"\/home\/idahl\/miniconda3\/envs\/eval-util-expl\/lib\/python3.9\/site-packages\/datasets\/dataset_dict.py\", line 710, in load_from_disk\r\n dataset_dict[k] = Dataset.load_from_disk(dataset_dict_split_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/home\/idahl\/miniconda3\/envs\/eval-util-expl\/lib\/python3.9\/site-packages\/datasets\/arrow_dataset.py\", line 687, in load_from_disk\r\n return Dataset(\r\n File \"\/home\/idahl\/miniconda3\/envs\/eval-util-expl\/lib\/python3.9\/site-packages\/datasets\/arrow_dataset.py\", line 274, in __init__\r\n raise ValueError(\r\nValueError: External features info don't match the dataset:\r\nGot\r\n{'answer': Value(dtype='string', id=None), 'idx': {'answer': Value(dtype='int32', id=None), 'paragraph': Value(dtype='int32', id=None), 'question': Value(dtype='int32', id=None)}, 'label': ClassLabel(num_classes=2, names=['False', 'True'], names_file=None, id=None), 'paragraph': Value(dtype='string', id=None), 'question': Value(dtype='string', id=None)}\r\nwith type\r\nstruct, label: int64, paragraph: string, question: string>\r\n\r\nbut expected something like\r\n{'answer': Value(dtype='string', id=None), 'idx': {'paragraph': Value(dtype='int32', id=None), 'question': Value(dtype='int32', id=None), 'answer': Value(dtype='int32', id=None)}, 'label': Value(dtype='int64', id=None), 'paragraph': Value(dtype='string', id=None), 'question': Value(dtype='string', id=None)}\r\nwith type\r\nstruct, label: int64, paragraph: string, question: string>\r\n\r\n```\r\n\r\nThe non-matching part seems to be\r\n`'label': ClassLabel(num_classes=2, names=['False', 'True'], names_file=None, id=None),`\r\nvs \r\n`'label': Value(dtype='int64', id=None),`\r\n\r\nAnd the order in the `1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n","comment_length":29,"text":"DatasetDict save load Failing test in 1.6 not in 1.5\n## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n\nMy current workaround is to remove the idx feature:\r\n\r\n```\r\n\r\nfrom datasets import load_dataset, DatasetDict, Value\r\nds = load_dataset('super_glue', 'multirc')\r\nds = ds.remove_columns('idx')\r\n\r\nds.save_to_disk('tempds')\r\n\r\nds = DatasetDict.load_from_disk('tempds')\r\n\r\n```\r\n\r\nworks.","embeddings":[-0.1946798116,0.1140982956,-0.0127640385,0.284942925,0.064592123,-0.0614173673,0.1323396713,0.4565899968,0.4422373772,0.1792287529,0.2358967513,0.380851388,0.0545869619,-0.0561681502,-0.1504484713,-0.0622494444,0.2574026883,0.1362523735,0.047262188,0.1269330233,-0.1914962828,0.2465310097,-0.1496895254,-0.0032418445,-0.1895998418,0.1437868327,-0.2581939399,0.2942799032,-0.3553473651,-0.6636254787,0.2940472364,0.4080147743,0.3425835669,0.5095262527,-0.0001234102,0.0544177964,0.2841644287,-0.2096645385,-0.4484797716,-0.4640186727,-0.1993898153,0.0136942323,0.2473115027,-0.1567811072,0.174591288,-0.0732848793,-0.0321101137,-0.4220353365,0.1694841832,0.3417609334,0.1407151371,0.2266554832,0.2764903307,-0.2219138741,-0.2547798157,0.3302724957,-0.0204170793,0.48288396,-0.0186927021,0.0731682926,0.0414443091,0.0742254928,-0.0195427742,-0.1017797217,0.093925342,-0.0247307811,0.162043348,-0.0513432026,0.240814209,0.1979950964,0.6499521136,-0.2013998032,-0.5342013836,0.0352447331,-0.0611302778,0.1786230356,0.3324406445,-0.0698243231,0.0839635506,0.3244956732,-0.0997817516,-0.3653488159,-0.2351392955,-0.1203212589,0.0710190833,-0.2877634466,-0.0193996113,0.1131953076,-0.0397098362,0.0214464869,0.4245892167,-0.3535504639,-0.3489905298,-0.1019979045,-0.0459269546,-0.1144007817,-0.0653342977,-0.2706740201,0.113346532,-0.0720998496,0.2517627776,0.1316185147,-0.0391709432,0.0602462031,0.2115243673,0.2749886513,0.4198958874,0.2743373513,0.0357514173,0.0714356154,-0.2357528359,0.0838931948,-0.0404101573,-0.15166004,0.2088321447,-0.1358045042,0.5703294277,-0.4004871547,-0.3274163008,0.2602038383,-0.2743057013,-0.1729606539,-0.1039793417,0.2240318358,0.09327925,0.1972419173,0.1087378412,0.3317791224,0.0468955897,0.0277815927,-0.1801332235,0.0353045389,0.0400684513,0.011786541,0.0179779567,-0.0999579355,0.267460227,0.015178483,-0.4577178359,-0.1766552329,-0.0413619019,-0.0473274663,0.2514302135,0.0383211523,-0.2181576937,-0.056682542,0.2363063097,0.1262998134,-0.2239949852,0.2120421976,0.0718438476,-0.1589062363,-0.1686418802,0.1228035986,-0.3944003284,-0.2602534592,-0.2819331884,-0.0493995957,-0.083663933,-0.2920115888,-0.1123725697,-0.3287760615,-0.237355873,-0.1225714758,0.1428985149,0.4307113886,-0.5336277485,0.1942441016,0.1860763729,-0.0290210564,0.0966063365,-0.0230782609,0.0003097686,0.1380230188,-0.1793204844,0.1278733611,0.4110462368,-0.5454348922,-0.6159434319,0.307710886,-0.0049888324,0.1782058775,-0.1437655389,-0.0217472035,0.2127342075,-0.0530057736,0.0544378124,0.0589728355,0.0534363128,-0.0575243719,-0.4947072268,-0.3328706622,0.1166524291,-0.0132593643,-0.0556115769,0.091937542,0.1581820548,-0.2094933689,0.4844789207,-0.0858414844,-0.1812547147,0.1170089394,0.4515592456,-0.0738887712,0.0274728462,-0.2316350937,-0.5508966446,0.4312414229,0.0377286226,-0.0222507045,-0.0489411242,0.0583773702,-0.4379805923,0.1203429848,-0.3182476759,0.046511244,0.0106168631,0.078376852,-0.0866948366,-0.0088240383,-0.1149406061,0.140570417,-0.1519619673,0.0698583946,0.004358958,0.336599648,-0.2037377357,-0.2205159664,0.170929715,-0.0127892671,0.2898364365,-0.2678136826,-0.3201774955,0.4959192574,0.2282747328,0.0215910152,-0.0704142451,-0.3370976448,-0.0950707644,-0.0260207877,0.1928254664,0.4983442724,0.0574062653,0.0559969395,-0.2572018802,0.3138726652,0.0657231435,0.0568134151,0.0700780526,0.1781899929,0.2946604192,-0.024419995,-0.0170402769,-0.3440820575,0.2174295187,0.018817924,0.0102335596,0.0753727928,-0.1335324794,0.1482347995,0.25894925,0.0366390534,0.1828769594,0.0873260871,0.0021229947,-0.4102320969,0.0591992363,0.3296465278,0.6812970042,-0.0120210182,-0.0898986459,-0.0542867593,-0.0393865295,0.0392726399,0.1494913846,-0.0172953773,0.2711475194,0.1783984751,0.1901718229,0.0689663813,-0.2123125196,-0.1901509315,-0.0073718987,0.051064495,-0.3820844293,0.1315279007,-0.1379302591,0.5604397058,-0.204323709,-0.5258817077,-0.0548608974,-0.3896036744,-0.1108690277,0.3982457519,-0.1051947698,0.0420946889,-0.5863333344,0.042864982,-0.0554989949,-0.325607568,-0.1113469601,-0.1961999387,-0.1263737977,-0.1553942561,0.1137430668,0.1928962618,0.2179133743,-0.1094706878,-0.1682274193,-0.4347150326,-0.3339819312,0.1541959494,-0.1012643725,0.050694488,0.1850075871,-0.0659798011,0.1797925979,-0.0064853802,0.0098090507,-0.1555842906,-0.3324509263,0.1918925792,0.2623599172,-0.1795146763,0.0405300371,-0.4744684994,-0.317047894,-0.3201007545,0.2410695106,-0.0716331676,0.0295918006,0.0510351174,0.1749565452,0.152455315,0.0156619512,0.102222845,-0.0204779282,-0.1949142367,0.1171747521,-0.0141771985,-0.4500836134,0.1088241562,0.0357693024,0.1472648084,0.2089921087,-0.2778184414,-0.1442975402,0.0047041876,0.414567858,-0.0329649821,-0.2965220213,0.4994246364,0.335587889,0.0297882631,-0.1571239531,-0.0598418005,0.1395886838,0.0997576341,0.2127372026,-0.3571592867,0.4544912279,-0.3967370093,0.2623894513,-0.0561215617,-0.1933416575,0.2192241848,0.1355145276,0.3633723259,-0.4195737839,-0.2501191199,-0.0181212332,-0.2273002863,-0.1420907974,0.0491867177,-0.1898749918,0.3078674674,-0.0113018882,-0.2713414431,-0.1212768629,-0.3088128567,-0.0883086398,-0.6213800311,0.2525457442,0.129497841,0.2512174249,-0.1520728171,0.0570483357,0.0548919328,0.1056846306,0.3619897962,-0.1298970282,-0.6689382195,0.4119164944,-0.2728250027,0.3531420827,0.0170031525,0.6950641274,0.0251035877,-0.1877912432,-0.074011609,-0.0956236944,0.3716633618,0.1301261038,0.0777470544,0.1875228435,-0.15032655,-0.6920769215,-0.1083943248,0.0503957905,0.3830853999,0.1262316108,0.3629608154,-0.0177808609,-0.1991339624,0.2198234499,0.3881291151,-0.1670640707,-0.274376452,-0.2853179872,-0.1365243644,-0.1515447497,-0.2810799479,-0.0419628099,0.2196073383,-0.2597106695,-0.0625315532,-0.5216527581,-0.3158732355,0.064146623,0.1999338716,0.4470637441,-0.3979617953,0.2182785869,0.2503279746,0.169705525,0.5388915539,0.4505360723,-0.0512378775,-0.4795738757,-0.039661333,-0.1913179606,0.0461554788,0.0478836857,-0.0462604798,-0.1848695874,0.0630236641,-0.2803143561,-0.3086195886,0.0911357701,0.0995180681,0.0440459028,-0.3165179789,-0.1771359891,0.5262101889,-0.0977220014,-0.1301518232,0.4414811134,0.3330871165,-0.236470148,0.5683749914,-0.0350576416,0.6247905493,0.2005376965,0.126446858,0.2727004886,-0.515450716,0.0626663193,0.5921412706,0.0294377375,-0.5868449211,-0.1553604603,0.0331250206,-0.12538293,0.0326819122,0.036870379,-0.3613365293,0.0716338158,0.0084466087,-0.2822274566,0.1183375046,0.3593575358,-0.2079016119,0.1589399576,-0.3033201694,0.0912196338,0.1137904301,0.1030311212,-0.0877372548,-0.0660760701,-0.1419514716,-0.3079771698,-0.1325310767,0.2350479513,-0.1199391335,0.4439127445,0.2897050679,-0.4493189156,0.2196412235,0.2174482346,0.4070452452,-0.0582864359,-0.1289931089,-0.044320479,-0.253841579,-0.0569630153,0.2750828266,-0.0208601356,0.2357589453,0.0243200008,-0.2413082123,0.0149238911,-0.2202434391,-0.3488709629,-0.2136550844,-0.0252611972,0.0917013511,-0.403341949,-0.5704155564,-0.3656395674,-0.1338036209,-0.3513588607,0.045161888,0.1369814575,0.0903127491,0.1345508993,-0.030370215,-0.3727883101,0.1286842227,0.4785127938,-0.1395132095,-0.2365066856,0.5704622269,-0.0664498061,0.1163967997,-0.2200823426,0.1668679267,0.2937476337,-0.2160442621,0.4291682541,-0.2352191359,-0.1053846255,-0.033432588,0.5656456947,0.1970895678,-0.1196097955,0.0568211302,-0.2952887118,-0.3237188458,0.1376452446,-0.038047269,0.4634029865,-0.0172834657,-0.3075003326,-0.2098487467,0.1128287613,-0.2028447539,0.0374976136,0.0159998704,0.0449076295,0.1605370641,-0.0462051779,0.1810541749,-0.311786443,0.0975406766,-0.1628718376,0.0355669707,-0.0206244513,-0.2262709588,0.1717493087,0.1430513859,0.2631170452,-0.0958966464,-0.1393290013,0.0043114931,-0.0736732259,-0.0378092565,0.3947651386,-0.1223446727,0.2778863609,0.0895340815,0.2064663917,-0.1312061846,0.2292111069,0.064420931,0.3075909615,-0.0757777393,0.2196436971,-0.0794219077,-0.0892363861,-0.2326027453,0.1826041937,0.029328797,0.0955826789,0.1823251098,-0.1802758723,-0.0411883555,0.1104308739,0.2567685246,0.5237183571,-0.1497170478,-0.2371406406,-0.0213249139,0.1600279957,-0.1601502299,-0.2017437965,0.3003623784,-0.0835295469,-0.1162904724,-0.0765228197,-0.2054964602,-0.1942750812,-0.1985163838,0.3335354626,0.5401267409,-0.0722699091,0.2756753266,0.1050364822,-0.1331185549,0.1698137224,0.4339623749,-0.026871765,0.2728070617,0.2969137728,-0.0106996065,0.4752700925,0.1081370488,0.1763004363,0.1444735676,0.029552646,0.0501384512,-0.2383726835,0.0339755304,0.0337763727,0.0379040502,0.207487449,-0.5338286757,-0.0095709851,-0.0529127605,-0.3925668597,-0.0503622741,-0.2166739851,-0.2375059426,-0.1166066527,0.1916654706,-0.0871482491,-0.1870366335,0.2569023669,0.1854578108,0.1341342032,-0.1262688935,-0.1991169453,0.2447045147,0.128683418,-0.1152322665,-0.2423141152,0.1165309399,0.0655258819,-0.0270040352,0.2843689919,0.3048303723,0.5131988525,0.4630441964,-0.1235537753,-0.0936735943,-0.0676186979,-0.0890747234,0.160659641,0.1176436469,-0.2288692296,0.1419881284,0.3497951031,0.1042521968,0.0525661483,0.3554263413,0.1828850508,0.2552399039,-0.0793646127,0.8174863458,0.0652648807,-0.401847899,0.1853944808,0.2061248571,-0.0562279373,-0.2423319072,0.2547922134,0.2379752696,0.0354117416,0.043099571,0.0429509506,-0.0761299208,0.4061383009,0.3945253193,-0.009129893,-0.253089875,0.2011501044,-0.5567246675,0.1267212033,0.1150677055,-0.0242360272,0.1735586077,0.1886163652,0.1791179031,-0.0079982691,0.1234039813,0.3020167649,-0.2655394077,-0.0732762218,-0.2039487809,-0.0483581908,0.1700234711,-0.0384172276,-0.0339166112,-0.4353083074,0.2317860276,-0.2152300924,-0.0408024713,-0.1203102246,-0.1444326192,-0.0316804796,-0.0343607739,0.4194562733,0.1286736578,0.4079838991,0.0496628247,-0.4076710939,-0.2884944081,-0.1973595172,-0.1579232812,0.2628087997,0.0948492214,0.6385340095,-0.0972695649,0.0452125445,-0.2792316675,0.2594893575,-0.1286927909,-0.0773257017,-0.2083820403,0.1652529091,0.1398079991,0.0103357537,0.2678919435,0.1165606603,-0.0676965714,0.37588045,-0.3817226589,-0.2833266556,0.4406756163,0.0210840479,-0.1905985326,-0.0162194837,0.3005398512,-0.0426337682,-0.181629017,-0.7152320743,0.0028659408,0.415509671,-0.0063177161,-0.0824656785,0.1022371128,0.2898762226,0.1865238696,-0.0089995535,0.5413072109,-0.0108727999,0.0113003505,-0.0968637392,-0.072947748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2267","title":"DatasetDict save load Failing test in 1.6 not in 1.5","comments":"It looks like this issue comes from the order of the fields in the 'idx' struct that is different for some reason.\r\nI'm looking into it. Note that as a workaround you can also flatten the nested features with `ds = ds.flatten()`","body":"## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n","comment_length":42,"text":"DatasetDict save load Failing test in 1.6 not in 1.5\n## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n\nIt looks like this issue comes from the order of the fields in the 'idx' struct that is different for some reason.\r\nI'm looking into it. Note that as a workaround you can also flatten the nested features with `ds = ds.flatten()`","embeddings":[-0.1946798116,0.1140982956,-0.0127640385,0.284942925,0.064592123,-0.0614173673,0.1323396713,0.4565899968,0.4422373772,0.1792287529,0.2358967513,0.380851388,0.0545869619,-0.0561681502,-0.1504484713,-0.0622494444,0.2574026883,0.1362523735,0.047262188,0.1269330233,-0.1914962828,0.2465310097,-0.1496895254,-0.0032418445,-0.1895998418,0.1437868327,-0.2581939399,0.2942799032,-0.3553473651,-0.6636254787,0.2940472364,0.4080147743,0.3425835669,0.5095262527,-0.0001234102,0.0544177964,0.2841644287,-0.2096645385,-0.4484797716,-0.4640186727,-0.1993898153,0.0136942323,0.2473115027,-0.1567811072,0.174591288,-0.0732848793,-0.0321101137,-0.4220353365,0.1694841832,0.3417609334,0.1407151371,0.2266554832,0.2764903307,-0.2219138741,-0.2547798157,0.3302724957,-0.0204170793,0.48288396,-0.0186927021,0.0731682926,0.0414443091,0.0742254928,-0.0195427742,-0.1017797217,0.093925342,-0.0247307811,0.162043348,-0.0513432026,0.240814209,0.1979950964,0.6499521136,-0.2013998032,-0.5342013836,0.0352447331,-0.0611302778,0.1786230356,0.3324406445,-0.0698243231,0.0839635506,0.3244956732,-0.0997817516,-0.3653488159,-0.2351392955,-0.1203212589,0.0710190833,-0.2877634466,-0.0193996113,0.1131953076,-0.0397098362,0.0214464869,0.4245892167,-0.3535504639,-0.3489905298,-0.1019979045,-0.0459269546,-0.1144007817,-0.0653342977,-0.2706740201,0.113346532,-0.0720998496,0.2517627776,0.1316185147,-0.0391709432,0.0602462031,0.2115243673,0.2749886513,0.4198958874,0.2743373513,0.0357514173,0.0714356154,-0.2357528359,0.0838931948,-0.0404101573,-0.15166004,0.2088321447,-0.1358045042,0.5703294277,-0.4004871547,-0.3274163008,0.2602038383,-0.2743057013,-0.1729606539,-0.1039793417,0.2240318358,0.09327925,0.1972419173,0.1087378412,0.3317791224,0.0468955897,0.0277815927,-0.1801332235,0.0353045389,0.0400684513,0.011786541,0.0179779567,-0.0999579355,0.267460227,0.015178483,-0.4577178359,-0.1766552329,-0.0413619019,-0.0473274663,0.2514302135,0.0383211523,-0.2181576937,-0.056682542,0.2363063097,0.1262998134,-0.2239949852,0.2120421976,0.0718438476,-0.1589062363,-0.1686418802,0.1228035986,-0.3944003284,-0.2602534592,-0.2819331884,-0.0493995957,-0.083663933,-0.2920115888,-0.1123725697,-0.3287760615,-0.237355873,-0.1225714758,0.1428985149,0.4307113886,-0.5336277485,0.1942441016,0.1860763729,-0.0290210564,0.0966063365,-0.0230782609,0.0003097686,0.1380230188,-0.1793204844,0.1278733611,0.4110462368,-0.5454348922,-0.6159434319,0.307710886,-0.0049888324,0.1782058775,-0.1437655389,-0.0217472035,0.2127342075,-0.0530057736,0.0544378124,0.0589728355,0.0534363128,-0.0575243719,-0.4947072268,-0.3328706622,0.1166524291,-0.0132593643,-0.0556115769,0.091937542,0.1581820548,-0.2094933689,0.4844789207,-0.0858414844,-0.1812547147,0.1170089394,0.4515592456,-0.0738887712,0.0274728462,-0.2316350937,-0.5508966446,0.4312414229,0.0377286226,-0.0222507045,-0.0489411242,0.0583773702,-0.4379805923,0.1203429848,-0.3182476759,0.046511244,0.0106168631,0.078376852,-0.0866948366,-0.0088240383,-0.1149406061,0.140570417,-0.1519619673,0.0698583946,0.004358958,0.336599648,-0.2037377357,-0.2205159664,0.170929715,-0.0127892671,0.2898364365,-0.2678136826,-0.3201774955,0.4959192574,0.2282747328,0.0215910152,-0.0704142451,-0.3370976448,-0.0950707644,-0.0260207877,0.1928254664,0.4983442724,0.0574062653,0.0559969395,-0.2572018802,0.3138726652,0.0657231435,0.0568134151,0.0700780526,0.1781899929,0.2946604192,-0.024419995,-0.0170402769,-0.3440820575,0.2174295187,0.018817924,0.0102335596,0.0753727928,-0.1335324794,0.1482347995,0.25894925,0.0366390534,0.1828769594,0.0873260871,0.0021229947,-0.4102320969,0.0591992363,0.3296465278,0.6812970042,-0.0120210182,-0.0898986459,-0.0542867593,-0.0393865295,0.0392726399,0.1494913846,-0.0172953773,0.2711475194,0.1783984751,0.1901718229,0.0689663813,-0.2123125196,-0.1901509315,-0.0073718987,0.051064495,-0.3820844293,0.1315279007,-0.1379302591,0.5604397058,-0.204323709,-0.5258817077,-0.0548608974,-0.3896036744,-0.1108690277,0.3982457519,-0.1051947698,0.0420946889,-0.5863333344,0.042864982,-0.0554989949,-0.325607568,-0.1113469601,-0.1961999387,-0.1263737977,-0.1553942561,0.1137430668,0.1928962618,0.2179133743,-0.1094706878,-0.1682274193,-0.4347150326,-0.3339819312,0.1541959494,-0.1012643725,0.050694488,0.1850075871,-0.0659798011,0.1797925979,-0.0064853802,0.0098090507,-0.1555842906,-0.3324509263,0.1918925792,0.2623599172,-0.1795146763,0.0405300371,-0.4744684994,-0.317047894,-0.3201007545,0.2410695106,-0.0716331676,0.0295918006,0.0510351174,0.1749565452,0.152455315,0.0156619512,0.102222845,-0.0204779282,-0.1949142367,0.1171747521,-0.0141771985,-0.4500836134,0.1088241562,0.0357693024,0.1472648084,0.2089921087,-0.2778184414,-0.1442975402,0.0047041876,0.414567858,-0.0329649821,-0.2965220213,0.4994246364,0.335587889,0.0297882631,-0.1571239531,-0.0598418005,0.1395886838,0.0997576341,0.2127372026,-0.3571592867,0.4544912279,-0.3967370093,0.2623894513,-0.0561215617,-0.1933416575,0.2192241848,0.1355145276,0.3633723259,-0.4195737839,-0.2501191199,-0.0181212332,-0.2273002863,-0.1420907974,0.0491867177,-0.1898749918,0.3078674674,-0.0113018882,-0.2713414431,-0.1212768629,-0.3088128567,-0.0883086398,-0.6213800311,0.2525457442,0.129497841,0.2512174249,-0.1520728171,0.0570483357,0.0548919328,0.1056846306,0.3619897962,-0.1298970282,-0.6689382195,0.4119164944,-0.2728250027,0.3531420827,0.0170031525,0.6950641274,0.0251035877,-0.1877912432,-0.074011609,-0.0956236944,0.3716633618,0.1301261038,0.0777470544,0.1875228435,-0.15032655,-0.6920769215,-0.1083943248,0.0503957905,0.3830853999,0.1262316108,0.3629608154,-0.0177808609,-0.1991339624,0.2198234499,0.3881291151,-0.1670640707,-0.274376452,-0.2853179872,-0.1365243644,-0.1515447497,-0.2810799479,-0.0419628099,0.2196073383,-0.2597106695,-0.0625315532,-0.5216527581,-0.3158732355,0.064146623,0.1999338716,0.4470637441,-0.3979617953,0.2182785869,0.2503279746,0.169705525,0.5388915539,0.4505360723,-0.0512378775,-0.4795738757,-0.039661333,-0.1913179606,0.0461554788,0.0478836857,-0.0462604798,-0.1848695874,0.0630236641,-0.2803143561,-0.3086195886,0.0911357701,0.0995180681,0.0440459028,-0.3165179789,-0.1771359891,0.5262101889,-0.0977220014,-0.1301518232,0.4414811134,0.3330871165,-0.236470148,0.5683749914,-0.0350576416,0.6247905493,0.2005376965,0.126446858,0.2727004886,-0.515450716,0.0626663193,0.5921412706,0.0294377375,-0.5868449211,-0.1553604603,0.0331250206,-0.12538293,0.0326819122,0.036870379,-0.3613365293,0.0716338158,0.0084466087,-0.2822274566,0.1183375046,0.3593575358,-0.2079016119,0.1589399576,-0.3033201694,0.0912196338,0.1137904301,0.1030311212,-0.0877372548,-0.0660760701,-0.1419514716,-0.3079771698,-0.1325310767,0.2350479513,-0.1199391335,0.4439127445,0.2897050679,-0.4493189156,0.2196412235,0.2174482346,0.4070452452,-0.0582864359,-0.1289931089,-0.044320479,-0.253841579,-0.0569630153,0.2750828266,-0.0208601356,0.2357589453,0.0243200008,-0.2413082123,0.0149238911,-0.2202434391,-0.3488709629,-0.2136550844,-0.0252611972,0.0917013511,-0.403341949,-0.5704155564,-0.3656395674,-0.1338036209,-0.3513588607,0.045161888,0.1369814575,0.0903127491,0.1345508993,-0.030370215,-0.3727883101,0.1286842227,0.4785127938,-0.1395132095,-0.2365066856,0.5704622269,-0.0664498061,0.1163967997,-0.2200823426,0.1668679267,0.2937476337,-0.2160442621,0.4291682541,-0.2352191359,-0.1053846255,-0.033432588,0.5656456947,0.1970895678,-0.1196097955,0.0568211302,-0.2952887118,-0.3237188458,0.1376452446,-0.038047269,0.4634029865,-0.0172834657,-0.3075003326,-0.2098487467,0.1128287613,-0.2028447539,0.0374976136,0.0159998704,0.0449076295,0.1605370641,-0.0462051779,0.1810541749,-0.311786443,0.0975406766,-0.1628718376,0.0355669707,-0.0206244513,-0.2262709588,0.1717493087,0.1430513859,0.2631170452,-0.0958966464,-0.1393290013,0.0043114931,-0.0736732259,-0.0378092565,0.3947651386,-0.1223446727,0.2778863609,0.0895340815,0.2064663917,-0.1312061846,0.2292111069,0.064420931,0.3075909615,-0.0757777393,0.2196436971,-0.0794219077,-0.0892363861,-0.2326027453,0.1826041937,0.029328797,0.0955826789,0.1823251098,-0.1802758723,-0.0411883555,0.1104308739,0.2567685246,0.5237183571,-0.1497170478,-0.2371406406,-0.0213249139,0.1600279957,-0.1601502299,-0.2017437965,0.3003623784,-0.0835295469,-0.1162904724,-0.0765228197,-0.2054964602,-0.1942750812,-0.1985163838,0.3335354626,0.5401267409,-0.0722699091,0.2756753266,0.1050364822,-0.1331185549,0.1698137224,0.4339623749,-0.026871765,0.2728070617,0.2969137728,-0.0106996065,0.4752700925,0.1081370488,0.1763004363,0.1444735676,0.029552646,0.0501384512,-0.2383726835,0.0339755304,0.0337763727,0.0379040502,0.207487449,-0.5338286757,-0.0095709851,-0.0529127605,-0.3925668597,-0.0503622741,-0.2166739851,-0.2375059426,-0.1166066527,0.1916654706,-0.0871482491,-0.1870366335,0.2569023669,0.1854578108,0.1341342032,-0.1262688935,-0.1991169453,0.2447045147,0.128683418,-0.1152322665,-0.2423141152,0.1165309399,0.0655258819,-0.0270040352,0.2843689919,0.3048303723,0.5131988525,0.4630441964,-0.1235537753,-0.0936735943,-0.0676186979,-0.0890747234,0.160659641,0.1176436469,-0.2288692296,0.1419881284,0.3497951031,0.1042521968,0.0525661483,0.3554263413,0.1828850508,0.2552399039,-0.0793646127,0.8174863458,0.0652648807,-0.401847899,0.1853944808,0.2061248571,-0.0562279373,-0.2423319072,0.2547922134,0.2379752696,0.0354117416,0.043099571,0.0429509506,-0.0761299208,0.4061383009,0.3945253193,-0.009129893,-0.253089875,0.2011501044,-0.5567246675,0.1267212033,0.1150677055,-0.0242360272,0.1735586077,0.1886163652,0.1791179031,-0.0079982691,0.1234039813,0.3020167649,-0.2655394077,-0.0732762218,-0.2039487809,-0.0483581908,0.1700234711,-0.0384172276,-0.0339166112,-0.4353083074,0.2317860276,-0.2152300924,-0.0408024713,-0.1203102246,-0.1444326192,-0.0316804796,-0.0343607739,0.4194562733,0.1286736578,0.4079838991,0.0496628247,-0.4076710939,-0.2884944081,-0.1973595172,-0.1579232812,0.2628087997,0.0948492214,0.6385340095,-0.0972695649,0.0452125445,-0.2792316675,0.2594893575,-0.1286927909,-0.0773257017,-0.2083820403,0.1652529091,0.1398079991,0.0103357537,0.2678919435,0.1165606603,-0.0676965714,0.37588045,-0.3817226589,-0.2833266556,0.4406756163,0.0210840479,-0.1905985326,-0.0162194837,0.3005398512,-0.0426337682,-0.181629017,-0.7152320743,0.0028659408,0.415509671,-0.0063177161,-0.0824656785,0.1022371128,0.2898762226,0.1865238696,-0.0089995535,0.5413072109,-0.0108727999,0.0113003505,-0.0968637392,-0.072947748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2267","title":"DatasetDict save load Failing test in 1.6 not in 1.5","comments":"I just pushed a fix on `master`. We'll do a new release soon !\r\n\r\nThanks for reporting","body":"## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n","comment_length":17,"text":"DatasetDict save load Failing test in 1.6 not in 1.5\n## Describe the bug\r\n\r\nWe have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema.\r\n\r\n\r\n\r\n\r\nDowngrading to `>1.6` -- fixes the problem.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\n\r\n### Load a dataset dict from jsonl \r\n\r\npath = '\/test\/foo'\r\n\r\nds_dict.save_to_disk(path)\r\n\r\nds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6\r\n```\r\n\r\n## Expected results\r\n\r\nUpgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk.\r\n\r\n## Actual results\r\n```\r\n # Infer features if None\r\n inferred_features = Features.from_arrow_schema(arrow_table.schema)\r\n if self.info.features is None:\r\n self.info.features = inferred_features\r\n \r\n # Infer fingerprint if None\r\n \r\n if self._fingerprint is None:\r\n self._fingerprint = generate_fingerprint(self)\r\n \r\n # Sanity checks\r\n \r\n assert self.features is not None, \"Features can't be None in a Dataset object\"\r\n assert self._fingerprint is not None, \"Fingerprint can't be None in a Dataset object\"\r\n if self.info.features.type != inferred_features.type:\r\n> raise ValueError(\r\n \"External features info don't match the dataset:\\nGot\\n{}\\nwith type\\n{}\\n\\nbut expected something like\\n{}\\nwith type\\n{}\".format(\r\n self.info.features, self.info.features.type, inferred_features, inferred_features.type\r\n )\r\n )\r\nE ValueError: External features info don't match the dataset:\r\nE Got\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, color: string, head: int64, head_span: struct, label: string>>, spans: list>, text: string, tokens: list>>\r\nE \r\nE but expected something like\r\nE {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]}\r\nE with type\r\nE struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list, encoding__offsets: list>, encoding__overflowing: list, encoding__tokens: list, encoding__words: list, ner_ids: list, ner_labels: list, relations: list, child_span: struct, color: string, label: string>>, spans: list>, text: string, tokens: list>>\r\n\r\n..\/..\/..\/..\/..\/.virtualenvs\/tf_ner_rel_lib\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py:274: ValueError\r\n```\r\n## Versions\r\n- Datasets: 1.6.1\r\n- Python: 3.8.5 (default, Jan 26 2021, 10:01:04) \r\n[Clang 12.0.0 (clang-1200.0.32.2)]\r\n- Platform: macOS-10.15.7-x86_64-i386-64bit\r\n\r\n```\r\n\nI just pushed a fix on `master`. We'll do a new release soon !\r\n\r\nThanks for reporting","embeddings":[-0.1946798116,0.1140982956,-0.0127640385,0.284942925,0.064592123,-0.0614173673,0.1323396713,0.4565899968,0.4422373772,0.1792287529,0.2358967513,0.380851388,0.0545869619,-0.0561681502,-0.1504484713,-0.0622494444,0.2574026883,0.1362523735,0.047262188,0.1269330233,-0.1914962828,0.2465310097,-0.1496895254,-0.0032418445,-0.1895998418,0.1437868327,-0.2581939399,0.2942799032,-0.3553473651,-0.6636254787,0.2940472364,0.4080147743,0.3425835669,0.5095262527,-0.0001234102,0.0544177964,0.2841644287,-0.2096645385,-0.4484797716,-0.4640186727,-0.1993898153,0.0136942323,0.2473115027,-0.1567811072,0.174591288,-0.0732848793,-0.0321101137,-0.4220353365,0.1694841832,0.3417609334,0.1407151371,0.2266554832,0.2764903307,-0.2219138741,-0.2547798157,0.3302724957,-0.0204170793,0.48288396,-0.0186927021,0.0731682926,0.0414443091,0.0742254928,-0.0195427742,-0.1017797217,0.093925342,-0.0247307811,0.162043348,-0.0513432026,0.240814209,0.1979950964,0.6499521136,-0.2013998032,-0.5342013836,0.0352447331,-0.0611302778,0.1786230356,0.3324406445,-0.0698243231,0.0839635506,0.3244956732,-0.0997817516,-0.3653488159,-0.2351392955,-0.1203212589,0.0710190833,-0.2877634466,-0.0193996113,0.1131953076,-0.0397098362,0.0214464869,0.4245892167,-0.3535504639,-0.3489905298,-0.1019979045,-0.0459269546,-0.1144007817,-0.0653342977,-0.2706740201,0.113346532,-0.0720998496,0.2517627776,0.1316185147,-0.0391709432,0.0602462031,0.2115243673,0.2749886513,0.4198958874,0.2743373513,0.0357514173,0.0714356154,-0.2357528359,0.0838931948,-0.0404101573,-0.15166004,0.2088321447,-0.1358045042,0.5703294277,-0.4004871547,-0.3274163008,0.2602038383,-0.2743057013,-0.1729606539,-0.1039793417,0.2240318358,0.09327925,0.1972419173,0.1087378412,0.3317791224,0.0468955897,0.0277815927,-0.1801332235,0.0353045389,0.0400684513,0.011786541,0.0179779567,-0.0999579355,0.267460227,0.015178483,-0.4577178359,-0.1766552329,-0.0413619019,-0.0473274663,0.2514302135,0.0383211523,-0.2181576937,-0.056682542,0.2363063097,0.1262998134,-0.2239949852,0.2120421976,0.0718438476,-0.1589062363,-0.1686418802,0.1228035986,-0.3944003284,-0.2602534592,-0.2819331884,-0.0493995957,-0.083663933,-0.2920115888,-0.1123725697,-0.3287760615,-0.237355873,-0.1225714758,0.1428985149,0.4307113886,-0.5336277485,0.1942441016,0.1860763729,-0.0290210564,0.0966063365,-0.0230782609,0.0003097686,0.1380230188,-0.1793204844,0.1278733611,0.4110462368,-0.5454348922,-0.6159434319,0.307710886,-0.0049888324,0.1782058775,-0.1437655389,-0.0217472035,0.2127342075,-0.0530057736,0.0544378124,0.0589728355,0.0534363128,-0.0575243719,-0.4947072268,-0.3328706622,0.1166524291,-0.0132593643,-0.0556115769,0.091937542,0.1581820548,-0.2094933689,0.4844789207,-0.0858414844,-0.1812547147,0.1170089394,0.4515592456,-0.0738887712,0.0274728462,-0.2316350937,-0.5508966446,0.4312414229,0.0377286226,-0.0222507045,-0.0489411242,0.0583773702,-0.4379805923,0.1203429848,-0.3182476759,0.046511244,0.0106168631,0.078376852,-0.0866948366,-0.0088240383,-0.1149406061,0.140570417,-0.1519619673,0.0698583946,0.004358958,0.336599648,-0.2037377357,-0.2205159664,0.170929715,-0.0127892671,0.2898364365,-0.2678136826,-0.3201774955,0.4959192574,0.2282747328,0.0215910152,-0.0704142451,-0.3370976448,-0.0950707644,-0.0260207877,0.1928254664,0.4983442724,0.0574062653,0.0559969395,-0.2572018802,0.3138726652,0.0657231435,0.0568134151,0.0700780526,0.1781899929,0.2946604192,-0.024419995,-0.0170402769,-0.3440820575,0.2174295187,0.018817924,0.0102335596,0.0753727928,-0.1335324794,0.1482347995,0.25894925,0.0366390534,0.1828769594,0.0873260871,0.0021229947,-0.4102320969,0.0591992363,0.3296465278,0.6812970042,-0.0120210182,-0.0898986459,-0.0542867593,-0.0393865295,0.0392726399,0.1494913846,-0.0172953773,0.2711475194,0.1783984751,0.1901718229,0.0689663813,-0.2123125196,-0.1901509315,-0.0073718987,0.051064495,-0.3820844293,0.1315279007,-0.1379302591,0.5604397058,-0.204323709,-0.5258817077,-0.0548608974,-0.3896036744,-0.1108690277,0.3982457519,-0.1051947698,0.0420946889,-0.5863333344,0.042864982,-0.0554989949,-0.325607568,-0.1113469601,-0.1961999387,-0.1263737977,-0.1553942561,0.1137430668,0.1928962618,0.2179133743,-0.1094706878,-0.1682274193,-0.4347150326,-0.3339819312,0.1541959494,-0.1012643725,0.050694488,0.1850075871,-0.0659798011,0.1797925979,-0.0064853802,0.0098090507,-0.1555842906,-0.3324509263,0.1918925792,0.2623599172,-0.1795146763,0.0405300371,-0.4744684994,-0.317047894,-0.3201007545,0.2410695106,-0.0716331676,0.0295918006,0.0510351174,0.1749565452,0.152455315,0.0156619512,0.102222845,-0.0204779282,-0.1949142367,0.1171747521,-0.0141771985,-0.4500836134,0.1088241562,0.0357693024,0.1472648084,0.2089921087,-0.2778184414,-0.1442975402,0.0047041876,0.414567858,-0.0329649821,-0.2965220213,0.4994246364,0.335587889,0.0297882631,-0.1571239531,-0.0598418005,0.1395886838,0.0997576341,0.2127372026,-0.3571592867,0.4544912279,-0.3967370093,0.2623894513,-0.0561215617,-0.1933416575,0.2192241848,0.1355145276,0.3633723259,-0.4195737839,-0.2501191199,-0.0181212332,-0.2273002863,-0.1420907974,0.0491867177,-0.1898749918,0.3078674674,-0.0113018882,-0.2713414431,-0.1212768629,-0.3088128567,-0.0883086398,-0.6213800311,0.2525457442,0.129497841,0.2512174249,-0.1520728171,0.0570483357,0.0548919328,0.1056846306,0.3619897962,-0.1298970282,-0.6689382195,0.4119164944,-0.2728250027,0.3531420827,0.0170031525,0.6950641274,0.0251035877,-0.1877912432,-0.074011609,-0.0956236944,0.3716633618,0.1301261038,0.0777470544,0.1875228435,-0.15032655,-0.6920769215,-0.1083943248,0.0503957905,0.3830853999,0.1262316108,0.3629608154,-0.0177808609,-0.1991339624,0.2198234499,0.3881291151,-0.1670640707,-0.274376452,-0.2853179872,-0.1365243644,-0.1515447497,-0.2810799479,-0.0419628099,0.2196073383,-0.2597106695,-0.0625315532,-0.5216527581,-0.3158732355,0.064146623,0.1999338716,0.4470637441,-0.3979617953,0.2182785869,0.2503279746,0.169705525,0.5388915539,0.4505360723,-0.0512378775,-0.4795738757,-0.039661333,-0.1913179606,0.0461554788,0.0478836857,-0.0462604798,-0.1848695874,0.0630236641,-0.2803143561,-0.3086195886,0.0911357701,0.0995180681,0.0440459028,-0.3165179789,-0.1771359891,0.5262101889,-0.0977220014,-0.1301518232,0.4414811134,0.3330871165,-0.236470148,0.5683749914,-0.0350576416,0.6247905493,0.2005376965,0.126446858,0.2727004886,-0.515450716,0.0626663193,0.5921412706,0.0294377375,-0.5868449211,-0.1553604603,0.0331250206,-0.12538293,0.0326819122,0.036870379,-0.3613365293,0.0716338158,0.0084466087,-0.2822274566,0.1183375046,0.3593575358,-0.2079016119,0.1589399576,-0.3033201694,0.0912196338,0.1137904301,0.1030311212,-0.0877372548,-0.0660760701,-0.1419514716,-0.3079771698,-0.1325310767,0.2350479513,-0.1199391335,0.4439127445,0.2897050679,-0.4493189156,0.2196412235,0.2174482346,0.4070452452,-0.0582864359,-0.1289931089,-0.044320479,-0.253841579,-0.0569630153,0.2750828266,-0.0208601356,0.2357589453,0.0243200008,-0.2413082123,0.0149238911,-0.2202434391,-0.3488709629,-0.2136550844,-0.0252611972,0.0917013511,-0.403341949,-0.5704155564,-0.3656395674,-0.1338036209,-0.3513588607,0.045161888,0.1369814575,0.0903127491,0.1345508993,-0.030370215,-0.3727883101,0.1286842227,0.4785127938,-0.1395132095,-0.2365066856,0.5704622269,-0.0664498061,0.1163967997,-0.2200823426,0.1668679267,0.2937476337,-0.2160442621,0.4291682541,-0.2352191359,-0.1053846255,-0.033432588,0.5656456947,0.1970895678,-0.1196097955,0.0568211302,-0.2952887118,-0.3237188458,0.1376452446,-0.038047269,0.4634029865,-0.0172834657,-0.3075003326,-0.2098487467,0.1128287613,-0.2028447539,0.0374976136,0.0159998704,0.0449076295,0.1605370641,-0.0462051779,0.1810541749,-0.311786443,0.0975406766,-0.1628718376,0.0355669707,-0.0206244513,-0.2262709588,0.1717493087,0.1430513859,0.2631170452,-0.0958966464,-0.1393290013,0.0043114931,-0.0736732259,-0.0378092565,0.3947651386,-0.1223446727,0.2778863609,0.0895340815,0.2064663917,-0.1312061846,0.2292111069,0.064420931,0.3075909615,-0.0757777393,0.2196436971,-0.0794219077,-0.0892363861,-0.2326027453,0.1826041937,0.029328797,0.0955826789,0.1823251098,-0.1802758723,-0.0411883555,0.1104308739,0.2567685246,0.5237183571,-0.1497170478,-0.2371406406,-0.0213249139,0.1600279957,-0.1601502299,-0.2017437965,0.3003623784,-0.0835295469,-0.1162904724,-0.0765228197,-0.2054964602,-0.1942750812,-0.1985163838,0.3335354626,0.5401267409,-0.0722699091,0.2756753266,0.1050364822,-0.1331185549,0.1698137224,0.4339623749,-0.026871765,0.2728070617,0.2969137728,-0.0106996065,0.4752700925,0.1081370488,0.1763004363,0.1444735676,0.029552646,0.0501384512,-0.2383726835,0.0339755304,0.0337763727,0.0379040502,0.207487449,-0.5338286757,-0.0095709851,-0.0529127605,-0.3925668597,-0.0503622741,-0.2166739851,-0.2375059426,-0.1166066527,0.1916654706,-0.0871482491,-0.1870366335,0.2569023669,0.1854578108,0.1341342032,-0.1262688935,-0.1991169453,0.2447045147,0.128683418,-0.1152322665,-0.2423141152,0.1165309399,0.0655258819,-0.0270040352,0.2843689919,0.3048303723,0.5131988525,0.4630441964,-0.1235537753,-0.0936735943,-0.0676186979,-0.0890747234,0.160659641,0.1176436469,-0.2288692296,0.1419881284,0.3497951031,0.1042521968,0.0525661483,0.3554263413,0.1828850508,0.2552399039,-0.0793646127,0.8174863458,0.0652648807,-0.401847899,0.1853944808,0.2061248571,-0.0562279373,-0.2423319072,0.2547922134,0.2379752696,0.0354117416,0.043099571,0.0429509506,-0.0761299208,0.4061383009,0.3945253193,-0.009129893,-0.253089875,0.2011501044,-0.5567246675,0.1267212033,0.1150677055,-0.0242360272,0.1735586077,0.1886163652,0.1791179031,-0.0079982691,0.1234039813,0.3020167649,-0.2655394077,-0.0732762218,-0.2039487809,-0.0483581908,0.1700234711,-0.0384172276,-0.0339166112,-0.4353083074,0.2317860276,-0.2152300924,-0.0408024713,-0.1203102246,-0.1444326192,-0.0316804796,-0.0343607739,0.4194562733,0.1286736578,0.4079838991,0.0496628247,-0.4076710939,-0.2884944081,-0.1973595172,-0.1579232812,0.2628087997,0.0948492214,0.6385340095,-0.0972695649,0.0452125445,-0.2792316675,0.2594893575,-0.1286927909,-0.0773257017,-0.2083820403,0.1652529091,0.1398079991,0.0103357537,0.2678919435,0.1165606603,-0.0676965714,0.37588045,-0.3817226589,-0.2833266556,0.4406756163,0.0210840479,-0.1905985326,-0.0162194837,0.3005398512,-0.0426337682,-0.181629017,-0.7152320743,0.0028659408,0.415509671,-0.0063177161,-0.0824656785,0.1022371128,0.2898762226,0.1865238696,-0.0089995535,0.5413072109,-0.0108727999,0.0113003505,-0.0968637392,-0.072947748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2262","title":"NewsPH NLI dataset script fails to access test data.","comments":"Thanks @bhavitvyamalik for the fix !\r\nThe fix will be available in the next release.\r\nIt's already available on the `master` branch. For now you can either install `datasets` from source or use `script_version=\"master\"` in `load_dataset` to use the fixed version of this dataset.","body":"In Newsph-NLI Dataset (#1192), it fails to access test data.\r\n\r\nAccording to the script below, the download manager will download the train data when trying to download the test data. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/2a2dd6316af2cc7fdf24e4779312e8ee0c7ed98b\/datasets\/newsph_nli\/newsph_nli.py#L71\r\n\r\nIf you download it according to the script above, you can see that train and test receive the same data as shown below.\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> newsph_nli = load_dataset(path=\".\/datasets\/newsph_nli.py\")\r\n>>> newsph_nli\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 420000\r\n })\r\n test: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 420000\r\n })\r\n validation: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 90000\r\n })\r\n})\r\n>>> newsph_nli[\"train\"][0]\r\n{'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).',\r\n 'label': 1,\r\n 'premise': '\"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa,\" ayon kay Robredo sa inilabas nitong statement.'}\r\n>>> newsph_nli[\"test\"][0]\r\n{'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).',\r\n 'label': 1,\r\n 'premise': '\"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa,\" ayon kay Robredo sa inilabas nitong statement.'}\r\n```\r\n\r\nIn local, I modified the code of the source as below and got the correct result.\r\n```python\r\n71 test_path = os.path.join(download_path, \"test.csv\") \r\n```\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> newsph_nli = load_dataset(path=\".\/datasets\/newsph_nli.py\")\r\n>>> newsph_nli\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 420000\r\n })\r\n test: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 9000\r\n })\r\n validation: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 90000\r\n })\r\n})\r\n>>> newsph_nli[\"train\"][0]\r\n{'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).',\r\n 'label': 1,\r\n 'premise': '\"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa,\" ayon kay Robredo sa inilabas nitong statement.'}\r\n>>> newsph_nli[\"test\"][0]\r\n{'hypothesis': '-- JAI (@JaiPaller) September 13, 2019',\r\n 'label': 1,\r\n 'premise': 'Pinag-iingat ng Konsulado ng Pilipinas sa Dubai ang publiko, partikular ang mga donor, laban sa mga scam na gumagamit ng mga charitable organization.'}\r\n```\r\n\r\nI don't have experience with open source pull requests, so I suggest that you reflect them in the source.\r\n\r\nThank you for reading :)","comment_length":44,"text":"NewsPH NLI dataset script fails to access test data.\nIn Newsph-NLI Dataset (#1192), it fails to access test data.\r\n\r\nAccording to the script below, the download manager will download the train data when trying to download the test data. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/2a2dd6316af2cc7fdf24e4779312e8ee0c7ed98b\/datasets\/newsph_nli\/newsph_nli.py#L71\r\n\r\nIf you download it according to the script above, you can see that train and test receive the same data as shown below.\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> newsph_nli = load_dataset(path=\".\/datasets\/newsph_nli.py\")\r\n>>> newsph_nli\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 420000\r\n })\r\n test: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 420000\r\n })\r\n validation: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 90000\r\n })\r\n})\r\n>>> newsph_nli[\"train\"][0]\r\n{'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).',\r\n 'label': 1,\r\n 'premise': '\"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa,\" ayon kay Robredo sa inilabas nitong statement.'}\r\n>>> newsph_nli[\"test\"][0]\r\n{'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).',\r\n 'label': 1,\r\n 'premise': '\"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa,\" ayon kay Robredo sa inilabas nitong statement.'}\r\n```\r\n\r\nIn local, I modified the code of the source as below and got the correct result.\r\n```python\r\n71 test_path = os.path.join(download_path, \"test.csv\") \r\n```\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> newsph_nli = load_dataset(path=\".\/datasets\/newsph_nli.py\")\r\n>>> newsph_nli\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 420000\r\n })\r\n test: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 9000\r\n })\r\n validation: Dataset({\r\n features: ['premise', 'hypothesis', 'label'],\r\n num_rows: 90000\r\n })\r\n})\r\n>>> newsph_nli[\"train\"][0]\r\n{'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).',\r\n 'label': 1,\r\n 'premise': '\"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa,\" ayon kay Robredo sa inilabas nitong statement.'}\r\n>>> newsph_nli[\"test\"][0]\r\n{'hypothesis': '-- JAI (@JaiPaller) September 13, 2019',\r\n 'label': 1,\r\n 'premise': 'Pinag-iingat ng Konsulado ng Pilipinas sa Dubai ang publiko, partikular ang mga donor, laban sa mga scam na gumagamit ng mga charitable organization.'}\r\n```\r\n\r\nI don't have experience with open source pull requests, so I suggest that you reflect them in the source.\r\n\r\nThank you for reading :)\nThanks @bhavitvyamalik for the fix !\r\nThe fix will be available in the next release.\r\nIt's already available on the `master` branch. For now you can either install `datasets` from source or use `script_version=\"master\"` in `load_dataset` to use the fixed version of this dataset.","embeddings":[-0.1607158035,0.2849110961,-0.126976192,0.1755043417,0.085902229,0.0402895994,0.2562074959,0.3658847213,0.0419336632,0.3072821796,0.0052477997,0.0433721282,0.0158911105,0.2191219032,0.177515775,-0.1738474071,-0.0040245214,0.0109901149,-0.1373095661,0.0521646887,-0.1775496453,0.0072262986,-0.1995907426,0.158338502,-0.0375777408,0.210018754,-0.2031048238,0.094160974,-0.1932428777,-0.6549795866,0.3090721965,0.1273378432,0.0832721815,0.2888227701,-0.0001082165,0.0894125178,0.2988508344,0.0091615925,-0.3292865455,-0.3701016903,-0.3256571591,-0.4836532176,0.0113816774,-0.3868236542,-0.1000621468,-0.1852469742,0.2667526305,0.109197624,0.2585923076,0.6185494065,0.2178416103,0.360341996,0.0827492103,-0.0977324769,0.1449719667,0.0725211352,-0.1622267962,0.1413548589,-0.0237706862,-0.1449209899,0.0799242333,0.1947094202,0.0518151969,0.1217470467,-0.187107414,0.1651480198,0.2658438981,-0.2221500427,-0.1103747487,0.2396497428,0.0484495461,-0.249437809,-0.2226448357,-0.0337351002,-0.0844576284,0.0159997903,0.1275511682,0.3240195215,-0.2225623131,0.1198315471,-0.304782033,-0.0031713874,-0.3851747811,0.395362854,-0.2153151929,0.21622172,0.0354196467,0.1695628762,0.1948063374,-0.1787180752,-0.1353119016,-0.1662853956,-0.1664285809,0.1104919314,0.0447671115,-0.0354129523,-0.0712503046,-0.3175800145,0.3023943007,0.1509393603,0.0641392469,0.0615024716,-0.0090623898,0.0035719338,0.1476139724,-0.1303728968,0.0290067215,0.3485050499,0.3061073124,0.2977993488,0.074188903,-0.0587520935,0.0372149087,-0.0911941156,0.1269004345,-0.0198108908,0.4139211178,-0.2075270712,-0.123758696,0.2419610322,-0.0031197101,-0.1290283054,-0.0752904266,0.4222076237,0.0038069515,0.0440091752,0.0618460029,0.3608467877,-0.035158217,-0.089692682,-0.2308822125,0.147298187,-0.3299790025,0.1613569409,0.6004884839,0.0971651897,0.2637774348,-0.0439442508,-0.195943445,-0.2917523086,0.1034094617,-0.2452872843,0.5001855493,0.1316228062,0.2623940408,0.1821569502,-0.0875076056,0.3129864931,0.0101262797,0.0853112414,-0.0186631531,-0.3564785719,0.2350665033,0.2354106456,-0.2452390343,-0.026149353,0.0424883924,0.0778634995,-0.1504396498,-0.3769740462,-0.0665062815,-0.1449884027,-0.2322539538,-0.1142747253,0.3140326738,0.4287765622,-0.0391391627,-0.0388351269,0.1627981365,-0.032857772,0.1943426877,0.0196333602,-0.3911494613,0.2049051523,-0.1807556301,0.0412888266,0.5587879419,-0.4500208795,-0.724973321,0.2535991371,-0.0212613605,0.053040307,0.1778284758,0.0625050664,-0.0860289708,-0.0278928783,-0.4274306595,0.3901510537,0.0546880439,0.222521916,0.0165248718,-0.1682703942,0.3594576418,0.2673202455,0.0448145457,-0.0441087037,-0.0306841321,0.4716745019,0.6009884477,0.0481058396,0.0090541067,0.0842781141,-0.0217513461,-0.1325216293,-0.0750061423,-0.0794651657,-0.0566266403,0.2953828871,0.0179411843,-0.017959062,-0.1477219313,-0.0005189619,-0.5050708652,-0.0385373309,-0.3645445406,-0.2266291231,0.2683858871,0.3846640289,0.3528345525,0.3000055254,-0.030288389,0.3504017889,-0.3305170238,-0.0496247225,-0.2497683465,0.2820423841,-0.2013546824,0.0091113858,-0.022729218,0.1534704417,0.1481286883,-0.0519124605,-0.0547741354,0.2898336053,0.1164437756,0.0854344741,-0.0616064072,0.0667308643,0.1511046588,-0.5583309531,0.271253109,0.3992883861,0.1664095074,-0.0469950549,-0.2413630188,0.2244085968,-0.029788319,0.1660888046,0.2944079936,-0.0828003511,0.2006493509,-0.0955019072,-0.2641208172,-0.1278713197,0.0308986194,-0.0712233111,-0.085645549,-0.0873884857,-0.0694778562,-0.0240070913,0.5533158183,-0.0228692107,-0.0859912261,-0.142210111,-0.1544724405,-0.0401732996,-0.0457055308,0.3417987823,0.3394460082,-0.0052406257,0.1332755834,0.2720524967,0.0106293047,-0.2517020702,0.111766167,0.125965625,0.1259822994,0.2157384753,0.0531439707,0.0178054795,-0.1025190577,0.0333006196,0.3690502644,0.1830672324,-0.2688636184,-0.019761784,-0.2633472979,-0.5984284282,-0.2817599177,-0.1618253291,0.0276547931,-0.4178620577,-0.0583133735,0.2076059729,0.0333635993,-0.1740947664,-0.3731132448,-0.0507754274,0.0232869387,-0.1845424771,-0.067758441,-0.111016877,-0.2914595008,0.0881278068,0.0988045037,0.2324982733,0.185199663,-0.3372068107,-0.1936274469,-0.4294444621,-0.0837894827,0.2052208036,0.038776271,0.3248253167,-0.0089297742,0.4082071781,0.127725631,0.0729600936,0.2689097524,-0.2258544266,-0.2140462101,0.3920098543,0.0003068156,-0.0243679658,-0.0420403257,-0.8750798106,-0.3868276775,-0.2710677683,-0.0702921376,-0.1641078442,0.1371508837,0.4277605116,0.0564923584,0.1112268642,-0.0629068166,0.1891739666,-0.2624030709,0.1213424057,0.1968318522,-0.4102002084,-0.6001107693,-0.0918635651,0.1056960821,0.1412737817,0.0692282766,-0.4383095801,-0.099446483,-0.2035561502,0.2111938596,0.065518029,-0.1395415962,0.0986180305,-0.1876580715,0.0152319307,-0.0599374846,0.0633013919,0.1874912083,0.0681978017,0.2779685557,0.1245734245,0.4347033203,0.1053015515,0.5339272022,0.2907461822,-0.0869296715,0.4084816575,-0.4260172844,0.0973081514,0.0522852205,-0.3037233949,0.1839884967,0.034941949,0.1958121359,0.0272643194,-0.2361386418,-0.0680467859,-0.2292046696,0.0158425793,-0.2501107454,-0.274654001,0.1084882617,0.1041522473,0.2464464009,0.2721261382,0.2970673144,-0.0250744727,-0.2863848507,-0.0831079558,0.1753787845,0.0077823699,0.1302408278,-0.1485901177,-0.3568096459,-0.5432794094,0.049815964,0.1777336448,0.6533555388,-0.0952230394,-0.2328331172,0.0570928976,-0.0656629726,0.2139716148,-0.2622179389,0.2860746086,0.0941676721,-0.1925501376,-0.4614230096,-0.1490027755,-0.5367030501,0.2200160772,0.045881208,0.0783033445,-0.3128925264,-0.0865690485,0.4102016389,0.0667632595,-0.089586474,-0.1199110597,-0.0781766772,-0.2306940705,-0.1983647645,-0.0634454638,-0.0272966288,0.1544400007,-0.4726415873,0.0435138531,0.1438605338,0.1500722319,0.0452246964,0.4804131687,0.1437704861,0.1152134314,0.1626639962,0.1206894517,0.0033895818,0.048385296,0.4401526153,0.3329812586,-0.088432841,0.2296990007,-0.2514719963,0.2321724296,0.4090436697,-0.0843231902,-0.080255188,0.0124578318,-0.4772042036,-0.1453658789,0.1055897623,0.4880534112,-0.1975769699,-0.1639469266,-0.3932226598,0.2943482697,0.1879693419,-0.0693177506,-0.1881128401,-0.3113544583,-0.0901663005,0.19939439,-0.3372785151,0.7343437076,0.2030811608,0.1060781851,0.3501306176,0.0492577516,0.1445159614,-0.2173543423,0.1401653439,-0.2893861234,-0.1896621436,0.0507207923,-0.1337523162,-0.0575896278,0.0699674413,-0.1388361156,0.1918223798,0.0765191466,0.0465892926,-0.0903612822,0.4218005836,-0.1952953041,0.2279953361,-0.3704454899,0.1807519048,0.0409516506,0.1949239224,-0.1249005422,-0.125885576,-0.211458236,-0.0805524439,-0.2322449237,0.3438405991,-0.1014807746,-0.0080386642,-0.1548868269,-0.0715998858,-0.2092353106,0.3940388858,0.1898008138,0.1455554813,-0.2332072109,-0.0151821962,-0.1218621582,-0.0989138037,0.170207426,0.0720782727,0.2749034166,-0.2374999374,-0.5205233097,0.2717046142,0.0122125456,0.0115169166,-0.3011000156,-0.0515832491,0.2221486717,-0.4237349033,-0.2757753432,-0.1270696968,-0.2994165421,-0.0844061524,0.1799011379,-0.1724430621,-0.3016172647,0.0538752154,0.1061510071,-0.3442393243,0.0102176126,0.267234087,-0.0498558991,0.2570135891,0.5883638263,0.1426815093,0.0978076607,-0.2515437007,-0.1344152838,0.3967268169,-0.4480822384,-0.2421755344,-0.0587321594,0.2318981588,0.4225235283,0.497972846,-0.0524656288,-0.2359661609,-0.0647889748,-0.4515605569,-0.1405136287,0.0455117673,0.0813049525,0.2086680233,0.1423360109,0.0296594761,-0.0092559867,-0.1055456921,-0.3727012873,-0.1019631773,-0.2531993687,0.1466193795,-0.067255035,0.0076765744,0.1220961288,0.0045650154,0.1222153828,0.1053171903,-0.1673088521,-0.2406251132,-0.0903889313,0.1120863333,0.207494095,-0.0463491939,-0.3955715299,-0.2170511037,-0.1175274327,-0.0634542555,0.1516934037,-0.1089366376,0.0791212693,0.4872291088,0.1247822493,0.1911436319,0.0628994033,0.1000806019,-0.2500253618,0.3012390137,-0.0008740956,0.3668177426,-0.050457906,-0.028464485,-0.3968820572,-0.1286951005,0.3236761987,0.1369549483,0.2293805778,-0.2778627872,0.1718880534,-0.2285359949,0.3552609086,0.4184894562,-0.3320914209,0.0760074183,0.030481251,0.2555727959,-0.1592840701,0.0507189296,0.1791630238,-0.0031618788,0.142982915,0.2469055355,0.3377937078,-0.0695366785,0.2741985619,0.0299772155,0.273721993,-0.1217315271,0.0021439353,0.3989951313,-0.1240557358,-0.0454700254,0.0245486237,0.0185880214,0.0841747895,0.1809131652,0.0145675307,0.6020000577,0.3479761481,0.1010265127,0.1548650712,-0.0158827975,-0.1489730775,0.0707309842,-0.2232420743,0.2216875106,-0.1234169081,0.3565532565,-0.1334574074,0.0958144441,-0.5014275312,0.1163443401,-0.0066388519,-0.0972699896,-0.3805087507,-0.221804738,-0.2661810815,-0.1269684285,-0.071346581,-0.1114912406,0.3025799692,0.1585452259,-0.3330277205,-0.3868171275,0.0582638271,0.1791599542,0.1049534157,-0.2697232068,0.3325802386,-0.0474270098,-0.0690362602,0.3110888004,0.5155497193,0.337205261,0.1309101582,0.0748459622,0.0987794921,0.0350755267,-0.1985813081,0.0018601255,0.4188431501,0.25980708,-0.0176604372,0.3898780942,0.2189407945,-0.0906144381,-0.2487785518,0.2301792949,0.0657978132,-0.2974543273,0.1374529451,-0.2030493915,0.053777948,-0.3047029674,-0.1069199815,-0.5738534927,-0.249407202,0.2485083938,0.1277544647,0.3732635975,-0.2122015506,0.1067163125,-0.0765890852,0.4998399317,0.2393326014,0.2802219987,-0.2526634634,-0.2962815464,-0.6371301413,0.2800095081,-0.0336490087,-0.375767529,0.018093681,0.0966579989,-0.1841700226,0.0458851941,0.0163000617,-0.0502737463,-0.0921424106,0.4529127777,-0.4418610334,-0.3507741988,-0.0523808897,0.1537087113,-0.256706506,-0.1359141469,0.106962502,-0.1130834818,0.0478983708,0.0103551792,-0.2944969535,0.0620652214,0.4316852987,0.406312108,0.196434319,0.6059331298,-0.2052845359,-0.0306185391,-0.4789448082,-0.0331004746,-0.05877674,0.0631799623,-0.0728438944,0.1951271296,0.1172641814,0.1936030239,-0.0026944131,-0.0902222618,-0.0918910801,-0.4591364264,-0.0700070709,-0.1678907722,-0.1076351032,0.0234892555,0.0124001093,0.1525815129,-0.0149437031,0.0027827548,-0.0432892703,-0.2013232112,0.3291140199,-0.1706297249,-0.2205204964,-0.001282815,0.0587952435,-0.4319295287,0.1373668164,-0.6566751599,0.143506214,0.1344794929,0.0965302065,-0.0786257237,0.1540100724,-0.0446563661,0.2452207804,-0.0960332081,0.4119135439,0.0631083697,-0.2670820653,-0.0691991225,0.0340400375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2256","title":"Running `datase.map` with `num_proc > 1` uses a lot of memory","comments":"Thanks for reporting ! We are working on this and we'll do a patch release very soon.","body":"## Describe the bug\r\nRunning `datase.map` with `num_proc > 1` leads to a tremendous memory usage that requires swapping on disk and it becomes very slow.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndstc8_datset = load_dataset(\"roskoN\/dstc8-reddit-corpus\", keep_in_memory=False)\r\n\r\n\r\ndef _prepare_sample(batch):\r\n return {\"input_ids\": list(), \"attention_mask\": list()}\r\n\r\n\r\nfor split_name, dataset_split in list(dstc8_datset.items()):\r\n print(f\"Processing {split_name}\")\r\n encoded_dataset_split = dataset_split.map(\r\n function=_prepare_sample,\r\n batched=True,\r\n num_proc=4,\r\n remove_columns=dataset_split.column_names,\r\n batch_size=10,\r\n writer_batch_size=10,\r\n keep_in_memory=False,\r\n )\r\n print(encoded_dataset_split)\r\n\r\n path = f\".\/data\/encoded_{split_name}\"\r\n\r\n encoded_dataset_split.save_to_disk(path)\r\n```\r\n\r\n## Expected results\r\nMemory usage should stay within reasonable boundaries.\r\n\r\n\r\n## Actual results\r\nThis is htop-output from running the provided script.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/8143425\/115954836-66954980-a4f3-11eb-8340-0153bdc3a475.png)\r\n\r\n## Versions\r\n```\r\n- Datasets: 1.6.0\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26)\r\n[GCC 7.3.0]\r\n- Platform: Linux-4.19.128-microsoft-standard-x86_64-with-glibc2.10\r\n```\r\nRunning on WSL2\r\n","comment_length":17,"text":"Running `datase.map` with `num_proc > 1` uses a lot of memory\n## Describe the bug\r\nRunning `datase.map` with `num_proc > 1` leads to a tremendous memory usage that requires swapping on disk and it becomes very slow.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndstc8_datset = load_dataset(\"roskoN\/dstc8-reddit-corpus\", keep_in_memory=False)\r\n\r\n\r\ndef _prepare_sample(batch):\r\n return {\"input_ids\": list(), \"attention_mask\": list()}\r\n\r\n\r\nfor split_name, dataset_split in list(dstc8_datset.items()):\r\n print(f\"Processing {split_name}\")\r\n encoded_dataset_split = dataset_split.map(\r\n function=_prepare_sample,\r\n batched=True,\r\n num_proc=4,\r\n remove_columns=dataset_split.column_names,\r\n batch_size=10,\r\n writer_batch_size=10,\r\n keep_in_memory=False,\r\n )\r\n print(encoded_dataset_split)\r\n\r\n path = f\".\/data\/encoded_{split_name}\"\r\n\r\n encoded_dataset_split.save_to_disk(path)\r\n```\r\n\r\n## Expected results\r\nMemory usage should stay within reasonable boundaries.\r\n\r\n\r\n## Actual results\r\nThis is htop-output from running the provided script.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/8143425\/115954836-66954980-a4f3-11eb-8340-0153bdc3a475.png)\r\n\r\n## Versions\r\n```\r\n- Datasets: 1.6.0\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26)\r\n[GCC 7.3.0]\r\n- Platform: Linux-4.19.128-microsoft-standard-x86_64-with-glibc2.10\r\n```\r\nRunning on WSL2\r\n\nThanks for reporting ! We are working on this and we'll do a patch release very soon.","embeddings":[-0.0754546002,-0.3018320203,-0.0413843729,0.3514220417,0.2028204054,0.0986170247,0.099690333,0.2842758894,0.2723552883,0.2235053182,0.1852624714,0.4316777885,-0.1457952112,-0.0435094424,-0.0000909037,0.1481419951,0.2869782448,0.0148507189,0.1028161347,0.0590187646,-0.4523685277,0.0157497246,-0.3991191387,-0.2898879349,-0.1479185075,0.1475247145,-0.0865113661,0.3391438127,0.2111929506,-0.5017675161,0.1326698512,-0.0852004588,0.1664270163,0.2540663183,-0.0001137038,-0.1059106588,0.1337454021,0.1001989022,-0.1821791977,0.1486949176,-0.312330991,-0.3803421557,-0.0605341941,-0.3717800379,0.0739793628,0.1645597368,-0.0146035962,-0.6698246598,0.218203634,0.3299291134,0.1970496178,-0.1352159828,-0.1267959476,0.1426888108,0.2613807321,0.2144849747,-0.0019844794,-0.2373128384,0.3164677024,0.1026870683,0.0636079609,0.4374418557,-0.2772944868,-0.031458009,-0.0467965528,0.1114465892,0.3358644247,-0.2466217428,0.3791429102,-0.1982060522,0.1761236638,-0.3240182698,0.1939211786,-0.1367563456,-0.2850831747,-0.4452186823,0.2263071686,0.0385255329,-0.1923348159,0.1316108555,-0.020382246,-0.0772100911,0.2251264602,-0.0430459939,-0.0519027896,0.1416192502,0.0057180547,0.1725652665,0.2679631412,-0.1135372818,-0.2476183027,-0.0225866605,-0.034672074,0.1145174801,-0.3829443753,-0.1603149325,0.0903514326,-0.1762013882,0.0510158837,-0.3032523692,0.105065465,0.1592099816,0.2167821974,0.1538217217,0.3814410865,0.0126948627,0.036227949,0.3134065866,0.2402250022,-0.1346106976,0.2550569773,0.057650283,0.171500355,-0.2757337689,0.3283064663,-0.0324563347,0.2573720813,-0.1818683892,-0.2745570242,0.2707969248,-0.0707862526,0.2562495172,0.0107705668,0.2392870039,0.2339676768,0.4569321871,-0.1290821284,-0.1677623093,-0.2951617241,-0.1780389547,-0.1612370014,-0.0624380484,-0.4973858595,0.0127899535,0.2001049221,-0.0157730225,-0.0672496557,0.0855232179,-0.100447394,-0.3198694289,0.0025576681,-0.1037027985,0.2624628842,0.2540854812,-0.2491627932,0.4018828571,-0.1385018826,-0.0014938404,-0.3082485199,0.4717456698,-0.3724940121,-0.2667652369,-0.1029589772,0.0955395624,0.1612251699,0.2892300487,-0.2312066108,0.2438077778,0.4335412681,0.0352739953,-0.0686667413,-0.1108095795,-0.6351304054,-0.2619717121,-0.0059178141,0.4431113899,-0.3248881996,0.2043526769,0.0911757201,0.061570663,0.5199885368,0.312188983,0.0134994769,-0.2704667449,0.0342344344,0.1604535282,0.2634508014,-0.3385006785,-0.8233773708,0.4366506338,-0.2258637249,-0.0950899646,0.1500333548,0.2552035451,0.3965847194,-0.1706774384,0.0508164763,0.2231238633,0.0575615577,0.6018776894,-0.4744095802,-0.2935760915,0.0439664386,-0.0241866279,-0.1145748198,-0.1487894803,0.0661745071,0.1963353306,0.4563109577,-0.0463559777,0.2246581763,0.1608531475,-0.1028402224,-0.0563536361,0.1254666001,-0.2527486086,-0.4038699269,0.122095786,0.0962012112,-0.1537615955,-0.0750262067,-0.1954279989,0.0505759828,0.0149783427,0.1189678013,-0.3708422184,0.0518171303,-0.0682420433,0.0159427635,-0.2603600323,0.0931864381,0.2902766168,0.1489875317,-0.0118339593,-0.2216127217,0.0043831752,-0.0359767675,-0.1321116835,-0.1899850667,0.112056978,0.1009918153,0.0459345132,-0.1223837584,0.2798475623,0.2700845599,-0.0917511955,-0.1688500196,0.0286816098,0.3951385319,-0.2478909791,0.3301876783,0.1503428668,0.2383357137,-0.1706946343,-0.1315209866,-0.0679994896,0.1871339828,0.2635293603,-0.3016976118,0.1257996261,0.2288258821,0.0110698612,0.233426705,-0.3601567745,0.0444171987,0.2666493952,0.0942667574,0.1847881675,0.0502932593,0.2274681032,0.3800434768,0.0071574715,-0.0674424171,0.2289534658,-0.4205145836,-0.2080101073,0.1267896742,-0.1767047048,0.4225719273,0.1683559418,0.0327091739,-0.0163709726,-0.0035293209,-0.0506971441,0.1758865863,0.087111719,-0.0744864345,-0.3462857008,0.2362720221,-0.135760501,-0.0663327277,-0.1010182872,0.0710001513,0.3797438443,-0.3099096715,0.0203586984,-0.1911893785,0.3776617348,0.0741064847,0.1036665365,0.1740548313,-0.1810127795,-0.0455531105,0.0940028802,0.010912885,0.0489136353,0.0440295264,0.133071959,0.0573309995,0.0461419038,-0.1617720425,-0.2432823479,-0.2543354332,-0.0349646062,0.2633143365,0.0199793596,0.35756585,0.0410643853,-0.1443333775,-0.0342577919,-0.1015553176,-0.0204920657,0.1299987137,-0.0371873863,-0.1992779672,-0.076335974,-0.0712473765,0.1608318388,-0.0548366718,-0.2768733203,-0.1413891464,0.3183433115,0.0626070201,-0.0981762037,-0.2886623442,-0.2188269645,-0.1263393611,-0.1334591806,0.1696945578,0.1252622306,0.4080451429,-0.281423986,0.2976653278,0.0794598535,0.121788308,-0.0141471475,-0.3536334038,-0.0541282296,0.0513544865,-0.0443673357,-0.1495989263,0.033932291,0.120275192,-0.1339851171,0.1744812578,-0.4615814388,0.1104643792,-0.4668396115,0.1116042063,0.0113620861,0.0800106376,0.5854110718,0.3444525599,-0.1051352322,0.0483481474,-0.2294131815,0.0313920677,-0.0876305178,0.0033826383,0.2333286107,0.4563247263,0.2585018575,0.1466429532,0.5971662402,-0.1571900547,0.0591799617,0.0931066796,-0.1019993052,-0.0068084695,-0.2207559496,0.1000622585,-0.1552474648,-0.141707629,0.4348702431,0.1866174489,-0.5755245686,0.0369556881,0.0787440538,-0.0339536555,-0.2688444257,0.3238787949,-0.2240194082,0.1153782532,0.0575222597,0.0397316925,-0.4204181731,-0.4140310287,-0.1729495525,-0.2100835294,-0.1254701316,-0.21117571,-0.2528620362,-0.0087267328,-0.9209541678,0.4346578717,0.0423067585,0.2663880587,0.0291176382,-0.1270744354,0.2269104719,-0.0960791036,0.5157929063,-0.0552822687,0.0635642707,-0.0331190601,0.0716398507,-0.3244416118,0.1178434938,-0.370285511,0.2683574855,0.1121725962,0.4310158789,-0.0707379505,-0.2245576978,0.3284598887,-0.0398592912,-0.1784522831,-0.1064338386,-0.2224118114,-0.436483264,-0.3006706238,0.1323173791,0.1993526667,-0.0298142284,0.0594165735,-0.2182418555,-0.0568515733,-0.1921623945,0.3076767623,-0.0633955076,0.2152741104,0.1270607263,0.0936274603,0.1779280901,0.1557342708,0.2029785216,0.3056217134,-0.0692676157,0.0772814527,0.1778837889,-0.0201269351,0.2897514999,0.3837610483,-0.096614413,-0.1651208997,-0.4494531751,0.4448655248,-0.5467810035,-0.0021801861,0.3576762378,0.1874077916,-0.3642303646,-0.4681982398,0.3083806634,0.4301122725,-0.2308666855,0.3238501549,-0.371575743,-0.3439403474,0.4432556331,0.246260494,1.0496554375,-0.4170783162,0.1567764878,-0.1183276325,-0.1092586815,0.2459249198,-0.2475515306,0.2015374154,-0.1053607389,-0.1511635333,0.0992517918,-0.0712988675,0.2511323392,0.289203912,-0.3414063752,0.1495854408,0.0113055753,0.1863741875,-0.0279172771,0.0384020917,0.2189945877,-0.3954831064,-0.0976110101,0.1587067395,0.0917466581,-0.2410398573,0.0605530143,0.1979291886,0.0437090173,-0.406267941,-0.2723635733,-0.0460996889,-0.2185847759,0.4199738801,-0.3964389265,-0.0363087133,0.1137499809,0.293200165,0.0169061851,0.3196717799,0.0498148277,-0.2268603444,0.3166848719,0.3254150152,0.2444178462,-0.029882187,0.0344419852,0.1223586723,-0.3180612028,0.1635264456,0.0399918817,-0.1702682376,-0.2749531269,0.2188417166,0.1691807657,-0.0348456353,0.0755603984,0.1349300742,-0.2710989714,-0.208531633,0.1263069957,0.0586598068,-0.1913861483,0.6787996292,-0.1512288302,-0.5358080268,-0.0390403531,0.4033436477,0.3998194933,-0.0419904664,0.2359113246,-0.0529590622,-0.2472295463,-0.2729047239,0.0615736246,-0.1938860565,-0.213328138,0.1621176451,-0.2020886987,-0.1738423109,-0.1614867747,-0.1788173616,-0.1046938077,0.0417481847,0.0245129839,-0.1625879556,-0.5511971116,-0.0844382569,-0.0048970077,0.3056385219,0.0284461156,0.1492988765,-0.405338943,0.2631620467,-0.2854908407,0.1681712866,0.0311893634,0.2942487895,0.1396812052,0.1356910467,0.0485884249,-0.1166589484,0.1171823442,0.4313510358,-0.0945000425,-0.240700528,0.1281098872,0.1215903983,-0.2400650382,-0.4046664536,-0.0769409537,-0.1059115082,-0.1060749739,-0.3666528463,0.3326639235,0.2881289124,0.0065845423,0.033172749,0.3520736992,0.0294058137,0.3683872819,0.4676436782,-0.3181140721,0.4390540719,-0.1310398132,0.4889287055,-0.0095687695,-0.1177618504,-0.0541601293,0.1230434552,0.1878221184,0.2196976691,0.0221043099,-0.234637022,-0.0886606798,0.387162894,0.0116693759,0.1940530688,-0.2117689848,-0.0995645821,0.2190718651,0.1965896189,-0.0653882623,-0.1499190331,0.3283310831,0.102489613,0.0533681326,0.3175069094,0.1553063393,-0.1566402465,-0.2167559862,0.0569214001,0.1864658892,-0.2504196763,0.0445000455,0.4319393337,0.1219553277,-0.0511190072,0.1328743994,0.00608304,0.339192301,0.7737321854,0.0299080331,0.4731138945,0.2530523241,0.3822734058,-0.2586421371,-0.2549137175,0.0199648254,0.5909483433,-0.5202937722,0.3448863924,0.1492221504,0.1271319091,-0.1794496477,-0.0138027109,-0.2726608813,0.2004192173,-0.2477800995,-0.3404521644,0.2359280586,-0.059883114,-0.2130357176,0.1699009091,-0.1578850448,-0.1862811297,0.7662403584,0.1137115136,-0.2051296085,-0.0337388031,-0.0958880782,0.0757601261,0.2537487745,-0.4055742025,0.0447936729,-0.1128054708,0.1331862211,-0.0065248525,-0.0909330845,0.3416788578,0.4272848666,-0.3083944023,0.0091654342,0.0735299215,-0.0213933568,-0.2865777314,0.2351248562,-0.1621107906,0.1921808422,-0.0053282925,0.027760433,-0.1431485713,-0.3150235713,0.3541503251,0.2710649371,0.0317653008,0.2356154323,0.0002751994,-0.2144954056,0.0864685178,-0.122608833,-0.234495163,-0.4195281267,0.4404408634,-0.4686650932,0.4340609014,-0.1359061003,0.0892813206,-0.0249708313,0.3473840356,-0.2726144195,-0.3021366596,-0.3087101877,-0.1953905076,-0.2418971509,0.0147772608,-0.2225594968,0.013464272,-0.1187452227,0.200883314,-0.2130833268,0.2623043954,0.0436940789,0.0240494031,0.0742783993,0.3759109974,-0.2080882043,0.1624818146,-0.0778405815,0.0834022984,0.2171081901,-0.709338367,0.3202190101,-0.0587175451,0.0275209695,-0.4466195107,0.1894747019,-0.0185406841,0.2479845583,0.583525598,0.3166791201,0.1492120028,-0.0907255262,-0.2716040909,-0.243096903,0.1842841059,-0.138626799,-0.1415862441,-0.0156178689,-0.079265371,-0.3890057504,-0.0809703246,-0.0648031607,-0.1123294756,0.0023693123,0.0472192317,-0.3245970011,-0.0081050498,0.0594383925,0.2595493793,-0.2023595273,0.2692157328,0.0353789516,0.0073948177,-0.0858302191,-0.4433559775,0.2853235304,-0.1633032858,-0.3462834656,-0.1708748341,0.4053235054,-0.2097498178,-0.1706225276,-0.2371229827,-0.023558883,0.160202831,-0.0396217145,-0.4610652626,0.2239741832,-0.1126031429,-0.2044065446,-0.0654520988,0.28561759,0.007112721,0.0262755733,0.22555767,-0.5136157274]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2256","title":"Running `datase.map` with `num_proc > 1` uses a lot of memory","comments":"We did a patch release to fix this issue.\r\nIt should be fixed in the new version 1.6.1\r\n\r\nThanks again for reporting and for the details :)","body":"## Describe the bug\r\nRunning `datase.map` with `num_proc > 1` leads to a tremendous memory usage that requires swapping on disk and it becomes very slow.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndstc8_datset = load_dataset(\"roskoN\/dstc8-reddit-corpus\", keep_in_memory=False)\r\n\r\n\r\ndef _prepare_sample(batch):\r\n return {\"input_ids\": list(), \"attention_mask\": list()}\r\n\r\n\r\nfor split_name, dataset_split in list(dstc8_datset.items()):\r\n print(f\"Processing {split_name}\")\r\n encoded_dataset_split = dataset_split.map(\r\n function=_prepare_sample,\r\n batched=True,\r\n num_proc=4,\r\n remove_columns=dataset_split.column_names,\r\n batch_size=10,\r\n writer_batch_size=10,\r\n keep_in_memory=False,\r\n )\r\n print(encoded_dataset_split)\r\n\r\n path = f\".\/data\/encoded_{split_name}\"\r\n\r\n encoded_dataset_split.save_to_disk(path)\r\n```\r\n\r\n## Expected results\r\nMemory usage should stay within reasonable boundaries.\r\n\r\n\r\n## Actual results\r\nThis is htop-output from running the provided script.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/8143425\/115954836-66954980-a4f3-11eb-8340-0153bdc3a475.png)\r\n\r\n## Versions\r\n```\r\n- Datasets: 1.6.0\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26)\r\n[GCC 7.3.0]\r\n- Platform: Linux-4.19.128-microsoft-standard-x86_64-with-glibc2.10\r\n```\r\nRunning on WSL2\r\n","comment_length":27,"text":"Running `datase.map` with `num_proc > 1` uses a lot of memory\n## Describe the bug\r\nRunning `datase.map` with `num_proc > 1` leads to a tremendous memory usage that requires swapping on disk and it becomes very slow.\r\n\r\n## Steps to reproduce the bug\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndstc8_datset = load_dataset(\"roskoN\/dstc8-reddit-corpus\", keep_in_memory=False)\r\n\r\n\r\ndef _prepare_sample(batch):\r\n return {\"input_ids\": list(), \"attention_mask\": list()}\r\n\r\n\r\nfor split_name, dataset_split in list(dstc8_datset.items()):\r\n print(f\"Processing {split_name}\")\r\n encoded_dataset_split = dataset_split.map(\r\n function=_prepare_sample,\r\n batched=True,\r\n num_proc=4,\r\n remove_columns=dataset_split.column_names,\r\n batch_size=10,\r\n writer_batch_size=10,\r\n keep_in_memory=False,\r\n )\r\n print(encoded_dataset_split)\r\n\r\n path = f\".\/data\/encoded_{split_name}\"\r\n\r\n encoded_dataset_split.save_to_disk(path)\r\n```\r\n\r\n## Expected results\r\nMemory usage should stay within reasonable boundaries.\r\n\r\n\r\n## Actual results\r\nThis is htop-output from running the provided script.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/8143425\/115954836-66954980-a4f3-11eb-8340-0153bdc3a475.png)\r\n\r\n## Versions\r\n```\r\n- Datasets: 1.6.0\r\n- Python: 3.8.8 (default, Apr 13 2021, 19:58:26)\r\n[GCC 7.3.0]\r\n- Platform: Linux-4.19.128-microsoft-standard-x86_64-with-glibc2.10\r\n```\r\nRunning on WSL2\r\n\nWe did a patch release to fix this issue.\r\nIt should be fixed in the new version 1.6.1\r\n\r\nThanks again for reporting and for the details :)","embeddings":[-0.0729513094,-0.2929819822,-0.042427361,0.3707073927,0.2069268823,0.0879345685,0.0891324282,0.2784173787,0.2625880837,0.2296098918,0.179768458,0.4309554696,-0.1489770412,-0.0277889445,-0.0165795069,0.1438741237,0.283003211,0.0198419057,0.0992945209,0.0716441944,-0.448911339,0.0283552222,-0.4113303423,-0.2822188139,-0.1598487943,0.1489114016,-0.0890552849,0.3382924199,0.2094049156,-0.516469419,0.1316249669,-0.0899930447,0.173614651,0.2649437785,-0.0001142843,-0.101619862,0.125217557,0.0971597806,-0.1957461089,0.1276341528,-0.3146615922,-0.3634471297,-0.0303165559,-0.3590409458,0.0726523474,0.1657172143,-0.0194390193,-0.6739723682,0.2170994878,0.3392878473,0.1941734254,-0.1366488934,-0.1196864545,0.1301043332,0.2364715487,0.2139768153,0.0010632157,-0.2308275849,0.322963804,0.1149242744,0.0747516453,0.4133724272,-0.2654496431,-0.0384335257,-0.0535099283,0.0997086763,0.352129519,-0.2720297873,0.3823864758,-0.1950076967,0.1814183444,-0.3327334523,0.1823688596,-0.1348747164,-0.2793178558,-0.4510781169,0.2298989743,0.0367099978,-0.1882492006,0.1366421878,-0.0380606093,-0.0894393548,0.2267846614,-0.0533736944,-0.0368233137,0.1144272089,0.0036473363,0.1740719527,0.2728739977,-0.111607343,-0.2295513302,-0.0282947272,-0.0394031405,0.1216240227,-0.3987645805,-0.157946229,0.0689608529,-0.1813658923,0.0463564768,-0.3080103993,0.1184394658,0.1578158587,0.2054126263,0.1522976756,0.3860761821,0.0093933782,0.0597691946,0.3085868657,0.2237630188,-0.1464236379,0.2582480013,0.0417789295,0.1672824472,-0.2809308171,0.3191932142,-0.0285182949,0.23656708,-0.1937867105,-0.2665585876,0.2586752474,-0.0610516109,0.2611827254,0.0222862605,0.2262551188,0.2448637784,0.4470606744,-0.1463556588,-0.160926789,-0.2841985822,-0.1833258569,-0.1512222886,-0.0574940294,-0.5046676397,0.0224807914,0.2060202956,-0.028394334,-0.0664362833,0.0933883861,-0.1053046659,-0.3432498574,0.012175451,-0.1125842556,0.2572804689,0.2620687485,-0.2398930192,0.3993423283,-0.1505595595,0.0162432101,-0.3013105094,0.4810847938,-0.3707550764,-0.2709364891,-0.0990084037,0.0852084234,0.1575821787,0.2947646379,-0.2420880646,0.2445336729,0.4286451042,0.0269888714,-0.0889528766,-0.121341981,-0.6368365288,-0.2681862414,0.0030347933,0.4324857593,-0.3264067769,0.2170574516,0.0809328929,0.0344491489,0.5263162851,0.310947299,0.0118526854,-0.2675107419,0.0316987745,0.1533008516,0.2658886909,-0.3270070851,-0.817191422,0.4371910393,-0.2157791704,-0.109528482,0.1383834928,0.2508608699,0.40468961,-0.1663723886,0.0347509049,0.2282755971,0.0608967543,0.5903041363,-0.4691990316,-0.303893894,0.0413282737,-0.0195751209,-0.1173685193,-0.1451271623,0.0748494491,0.1778643429,0.453548342,-0.058543466,0.2303673923,0.1774595976,-0.1110148206,-0.0563489832,0.1232414767,-0.2524731159,-0.4027755857,0.13394095,0.0832437128,-0.1526621878,-0.0804286152,-0.1751124114,0.0650854781,0.0063811243,0.12697725,-0.3695948124,0.0472450666,-0.078113012,0.0223031454,-0.2606379986,0.0992855877,0.2954145074,0.1608096063,-0.0029197617,-0.2119587362,0.0069003459,-0.0256237015,-0.1433457285,-0.1732806712,0.1090763211,0.112046659,0.0333948098,-0.1382002831,0.2837573588,0.2842199206,-0.0985985622,-0.1563302875,0.0272796862,0.4024820328,-0.2332773954,0.327091068,0.1425988823,0.2420417517,-0.1731173992,-0.1279882044,-0.0569084659,0.1828204393,0.2651683688,-0.2958455682,0.1233256534,0.2298754156,0.0121203372,0.2454758883,-0.36111781,0.0471434109,0.287186861,0.098257944,0.1909164488,0.0465373211,0.2114551216,0.3700146973,0.0057400702,-0.0826599076,0.2261688709,-0.4248207808,-0.2112520337,0.1328227967,-0.1818520278,0.4213223457,0.1641579717,0.0243873633,-0.0241139587,0.012031504,-0.0545138419,0.1692528725,0.0867379382,-0.0513383336,-0.367634356,0.2332722545,-0.1403825134,-0.0653906837,-0.1040041819,0.0889222696,0.3924361169,-0.2967655957,0.0110161183,-0.1935511976,0.400585711,0.0906683803,0.1055033356,0.1667192876,-0.18145518,-0.056737192,0.0982364044,0.0294012669,0.051611118,0.0656295717,0.1132944152,0.0556617863,0.0510638542,-0.1494139284,-0.2518106997,-0.2523031831,-0.0411089212,0.2673361897,0.0278153326,0.3511581421,0.0504587665,-0.155298993,-0.0317021236,-0.0926315561,-0.0218256712,0.1374751627,-0.034724433,-0.2257715315,-0.0832699984,-0.0728773773,0.1783143878,-0.0700538903,-0.2762166262,-0.1268468648,0.3321932256,0.061236538,-0.113702178,-0.2935315967,-0.2158631831,-0.1409101784,-0.1193530709,0.1708788723,0.1151362881,0.4196763933,-0.284565419,0.3068885803,0.0737346858,0.1142024025,-0.0205499046,-0.3569996059,-0.0585363396,0.0571643636,-0.0496483184,-0.1584663987,0.0290774722,0.1449582577,-0.1220377386,0.1793140471,-0.4542498887,0.118075192,-0.469053477,0.1137285158,0.0141398842,0.062865831,0.5845873952,0.3504353762,-0.0964849144,0.0529159978,-0.230626896,0.0342685841,-0.0814838186,-0.0005931606,0.2426695526,0.4532476664,0.2545185089,0.1341665834,0.6049479246,-0.1597821862,0.0588254035,0.0959706977,-0.1034300774,-0.0146367466,-0.2227414548,0.0980299488,-0.1607872099,-0.1319242567,0.4267861545,0.1794440001,-0.5726996064,0.0305252746,0.0714418888,-0.0304647796,-0.2655984461,0.3138512373,-0.2360403687,0.139657259,0.0529653691,0.0380796231,-0.421166569,-0.4224815369,-0.1897607744,-0.2102629244,-0.1148622185,-0.2045514584,-0.2550401092,0.0069437809,-0.9173231125,0.4260064065,0.0354426391,0.2873561084,0.0209309533,-0.0870488882,0.2374596894,-0.0823640972,0.5041686893,-0.04623276,0.0520478897,-0.036997892,0.0529942587,-0.3321003616,0.1269351244,-0.369209528,0.2951182723,0.1372307539,0.4450311363,-0.0686329678,-0.2359070182,0.3105686009,-0.028055314,-0.1798984855,-0.1254401505,-0.2167558521,-0.420091778,-0.3174932301,0.121516265,0.2025130391,-0.0256984681,0.0626203567,-0.2181415707,-0.052546218,-0.1870114952,0.3138545454,-0.0666481107,0.2187060267,0.1215391383,0.1070262715,0.1821810901,0.1268317699,0.2172592729,0.2916457653,-0.0685903803,0.087760292,0.181286335,-0.019848872,0.2986559868,0.3787651956,-0.0977903008,-0.1653563678,-0.4598185122,0.4437204599,-0.5339167118,-0.0056926981,0.3493115008,0.1893457025,-0.3774055541,-0.4680443108,0.3122807145,0.4378613532,-0.2264335901,0.3244328797,-0.3689555228,-0.3400514126,0.4310781658,0.245957002,1.0425153971,-0.4180926681,0.1474301815,-0.121995762,-0.1226690039,0.2334078848,-0.2235253155,0.2028811276,-0.1204934418,-0.1521508247,0.0998859331,-0.0788350925,0.2469507754,0.302749753,-0.343583703,0.1628810614,0.012138796,0.1544173062,-0.0279074013,0.0670087785,0.2419487536,-0.3951832652,-0.0779892579,0.1558913887,0.0974499583,-0.2486918271,0.0699923635,0.1840825081,0.0488288589,-0.384226203,-0.2763758004,-0.0557901673,-0.2391079217,0.4249266386,-0.3855213523,-0.0357600413,0.1384730786,0.2917613387,0.020908447,0.3187533319,0.0544690154,-0.2226583511,0.3064108193,0.3159710765,0.237069279,-0.0073574483,0.0240911674,0.1245163977,-0.3129646182,0.1590856314,0.0319005959,-0.1547683477,-0.277349472,0.2015645355,0.1497800946,-0.0377496295,0.0626067892,0.1266821474,-0.2779558003,-0.2009113133,0.1245696843,0.0692260563,-0.1845234185,0.6873885989,-0.151524201,-0.5412964821,-0.041894637,0.3999683559,0.392899543,-0.0489329547,0.237181291,-0.0453164354,-0.2420300543,-0.2633185983,0.0626353547,-0.1891954243,-0.2120862156,0.1704222709,-0.1979534626,-0.164414227,-0.1494859457,-0.1847037524,-0.1102135554,0.0472379588,0.0317659713,-0.1628386527,-0.5576214194,-0.0959425047,-0.0174774528,0.3096831441,0.0278614052,0.1491663307,-0.3958244324,0.2738790512,-0.279393822,0.1749602258,0.0430202335,0.295163393,0.1685283184,0.1238424927,0.0481739417,-0.1176057234,0.1080875918,0.4275873303,-0.0902685449,-0.2311120927,0.1356230974,0.121494621,-0.2221672386,-0.4117948711,-0.1003170684,-0.11029616,-0.1172305495,-0.3700533509,0.318164736,0.2941945791,-0.0046399604,0.0415020473,0.3446668088,0.0228504445,0.3782466054,0.4740272462,-0.3175374269,0.4398245811,-0.1240398958,0.5029127598,-0.0006961742,-0.1194653139,-0.0664824545,0.1135382354,0.189324975,0.2019136101,0.0142613556,-0.2180069983,-0.0792208016,0.3881903589,0.0022820702,0.1882443577,-0.2010737062,-0.1041810438,0.2134892792,0.1874675602,-0.0533839613,-0.1512509733,0.3261016011,0.1152948067,0.0401337855,0.3144670427,0.1612166315,-0.153558284,-0.2242771089,0.0629224703,0.1983852834,-0.2453487515,0.0340042077,0.4191627204,0.1287705749,-0.0589253865,0.1465151757,0.008379125,0.3284142017,0.7951183319,0.0146453623,0.4674240947,0.2519786954,0.3775933981,-0.2475756258,-0.2453172207,-0.0036997523,0.597800076,-0.5198103189,0.3401630819,0.137522012,0.1434893757,-0.184596777,-0.0176327974,-0.2620611787,0.2212814689,-0.2488320619,-0.3359470367,0.2347176671,-0.0679704547,-0.2230425924,0.1814351976,-0.1633814424,-0.1857769638,0.7715320587,0.1256975532,-0.2111807317,-0.0299296807,-0.105704084,0.1067803204,0.2385019809,-0.4037965238,0.0527578518,-0.1270991266,0.1256876141,-0.0163104963,-0.0895857811,0.3479725122,0.4311940074,-0.3091953099,0.0066910223,0.0831447318,-0.0184263065,-0.2655335069,0.2327060848,-0.1633505225,0.2119304985,-0.0132778455,0.0194175057,-0.1292643845,-0.323941946,0.3484843373,0.2820705175,0.0261901785,0.2455248088,0.0140513228,-0.2189703286,0.091001451,-0.1264245957,-0.2209575027,-0.4252358377,0.4463094175,-0.4778807163,0.4292029142,-0.1338777691,0.0863556042,-0.0299511682,0.3474974632,-0.2645900249,-0.3152249455,-0.3324307203,-0.2019028664,-0.2512768805,-0.0072482284,-0.2229041159,0.0300769731,-0.1103314981,0.1783263832,-0.2234199196,0.2558409274,0.0423539877,0.0398418941,0.0681416467,0.367983371,-0.2168523371,0.1783957034,-0.0945814848,0.0818415433,0.2187575102,-0.6929849386,0.3271977901,-0.0742616504,0.0175315607,-0.4543448985,0.1899351627,-0.0266178995,0.2580189705,0.5818484426,0.3229673505,0.1431878656,-0.0801820457,-0.2579418421,-0.2280545831,0.1943673491,-0.1347552538,-0.1406232566,-0.0131774964,-0.0856256187,-0.3904081285,-0.0767765865,-0.0737424493,-0.1178444102,-0.016984567,0.0490505435,-0.3196547031,-0.0116922231,0.0686318651,0.2735214531,-0.1925982982,0.2733894885,0.0418226682,0.0107282931,-0.0904776305,-0.4277671576,0.2736369967,-0.1582601517,-0.3398911953,-0.1713929772,0.389619112,-0.1916779578,-0.1781104654,-0.2426636815,-0.0195113942,0.1601627916,-0.0438182428,-0.463844955,0.2188513875,-0.1155027375,-0.1995620579,-0.0619690754,0.2797385454,0.0230919011,0.0117796483,0.218198657,-0.4928128719]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Hi ! Sorry to hear that. This may come from another issue then.\r\n\r\nFirst can we check if this latency comes from the dataset itself ?\r\nYou can try to load your dataset and benchmark the speed of querying random examples inside it ?\r\n```python\r\nimport time\r\nimport numpy as np\r\n\r\nfrom datasets import load_from_disk\r\n\r\ndataset = load_from_disk(...) # or from load_dataset...\r\n\r\n_start = time.time()\r\nn = 100\r\nfor i in np.random.default_rng(42).integers(0, len(dataset), size=n):\r\n _ = dataset[i]\r\nprint(time.time() - _start)\r\n```\r\n\r\nIf we see a significant speed difference between your two datasets then it would mean that there's an issue somewhere","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":101,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHi ! Sorry to hear that. This may come from another issue then.\r\n\r\nFirst can we check if this latency comes from the dataset itself ?\r\nYou can try to load your dataset and benchmark the speed of querying random examples inside it ?\r\n```python\r\nimport time\r\nimport numpy as np\r\n\r\nfrom datasets import load_from_disk\r\n\r\ndataset = load_from_disk(...) # or from load_dataset...\r\n\r\n_start = time.time()\r\nn = 100\r\nfor i in np.random.default_rng(42).integers(0, len(dataset), size=n):\r\n _ = dataset[i]\r\nprint(time.time() - _start)\r\n```\r\n\r\nIf we see a significant speed difference between your two datasets then it would mean that there's an issue somewhere","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Hi @lhoestq, here is the result. I additionally measured time to `load_from_disk`:\r\n* 60GB\r\n```\r\nloading took: 22.618776321411133\r\nramdom indexing 100 times took: 0.10214924812316895\r\n```\r\n\r\n* 600GB\r\n```\r\nloading took: 1176.1764674186707\r\nramdom indexing 100 times took: 2.853600025177002\r\n```\r\n\r\nHmm.. I double checked that it's version 1.6.0. The difference seems quite big, could it be related to the running environment? \r\n","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":59,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHi @lhoestq, here is the result. I additionally measured time to `load_from_disk`:\r\n* 60GB\r\n```\r\nloading took: 22.618776321411133\r\nramdom indexing 100 times took: 0.10214924812316895\r\n```\r\n\r\n* 600GB\r\n```\r\nloading took: 1176.1764674186707\r\nramdom indexing 100 times took: 2.853600025177002\r\n```\r\n\r\nHmm.. I double checked that it's version 1.6.0. The difference seems quite big, could it be related to the running environment? \r\n","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"I'm surprised by the speed change. Can you give more details about your dataset ?\r\nThe speed depends on the number of batches in the arrow tables and the distribution of the lengths of the batches.\r\nYou can access the batches by doing `dataset.data.to_batches()` (use only for debugging) (it doesn't bring data in memory).\r\n\r\nAlso can you explain what parameters you used if you used `map` calls ?\r\nAlso if you have some code that reproduces the issue I'd be happy to investigate it.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":84,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nI'm surprised by the speed change. Can you give more details about your dataset ?\r\nThe speed depends on the number of batches in the arrow tables and the distribution of the lengths of the batches.\r\nYou can access the batches by doing `dataset.data.to_batches()` (use only for debugging) (it doesn't bring data in memory).\r\n\r\nAlso can you explain what parameters you used if you used `map` calls ?\r\nAlso if you have some code that reproduces the issue I'd be happy to investigate it.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Also if you could give us more info about your env like your OS, version of pyarrow and if you're using an HDD or a SSD","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":26,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nAlso if you could give us more info about your env like your OS, version of pyarrow and if you're using an HDD or a SSD","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Here are some details of my 600GB dataset. This is a dataset AFTER the `map` function and once I load this dataset, I do not use `map` anymore in the training. Regarding the distribution of the lengths, it is almost uniform (90% is 512 tokens, and 10% is randomly shorter than that -- typical setting for language modeling).\r\n```\r\nlen(batches):\r\n492763\r\n\r\nbatches[0]: \r\npyarrow.RecordBatch\r\nattention_mask: list\r\n child 0, item: uint8\r\ninput_ids: list\r\n child 0, item: int16\r\nspecial_tokens_mask: list\r\n child 0, item: uint8\r\ntoken_type_ids: list\r\n child 0, item: uint8\r\n```\r\n\r\nHere the some parameters to `map` function just in case it is relevant:\r\n```\r\nnum_proc=1 # as multi processing is slower in my case\r\nload_from_cache_file=False\r\n```\r\n","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":118,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHere are some details of my 600GB dataset. This is a dataset AFTER the `map` function and once I load this dataset, I do not use `map` anymore in the training. Regarding the distribution of the lengths, it is almost uniform (90% is 512 tokens, and 10% is randomly shorter than that -- typical setting for language modeling).\r\n```\r\nlen(batches):\r\n492763\r\n\r\nbatches[0]: \r\npyarrow.RecordBatch\r\nattention_mask: list\r\n child 0, item: uint8\r\ninput_ids: list\r\n child 0, item: int16\r\nspecial_tokens_mask: list\r\n child 0, item: uint8\r\ntoken_type_ids: list\r\n child 0, item: uint8\r\n```\r\n\r\nHere the some parameters to `map` function just in case it is relevant:\r\n```\r\nnum_proc=1 # as multi processing is slower in my case\r\nload_from_cache_file=False\r\n```\r\n","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Regarding the environment, I am running the code on a cloud server. Here are some info:\r\n```\r\nUbuntu 18.04.5 LTS # cat \/etc\/issue\r\npyarrow 3.0.0 # pip list | grep pyarrow\r\n```\r\nThe data is stored in SSD and it is mounted to the machine via Network File System.\r\n\r\nIf you could point me to some of the commands to check the details of the environment, I would be happy to provide relevant information @lhoestq !","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":76,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nRegarding the environment, I am running the code on a cloud server. Here are some info:\r\n```\r\nUbuntu 18.04.5 LTS # cat \/etc\/issue\r\npyarrow 3.0.0 # pip list | grep pyarrow\r\n```\r\nThe data is stored in SSD and it is mounted to the machine via Network File System.\r\n\r\nIf you could point me to some of the commands to check the details of the environment, I would be happy to provide relevant information @lhoestq !","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"I am not sure how I could provide you with the reproducible code, since the problem only arises when the data is big. For the moment, I would share the part that I think is relevant. Feel free to ask me for more info.\r\n\r\n```python\r\nclass MyModel(pytorch_lightning.LightningModule)\r\n def setup(self, stage):\r\n self.dataset = datasets.load_from_disk(path)\r\n self.dataset.set_format(\"torch\")\r\n\r\n def train_dataloader(self):\r\n collate_fn = transformers.DataCollatorForLanguageModeling(\r\n tokenizer=transformers.ElectraTokenizerFast.from_pretrained(tok_path)\r\n )\r\n dataloader = torch.utils.DataLoader(\r\n self.dataset,\r\n batch_size=32,\r\n collate_fn=collate_fn,\r\n num_workers=8,\r\n pin_memory=True,\r\n )\r\n```","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":71,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nI am not sure how I could provide you with the reproducible code, since the problem only arises when the data is big. For the moment, I would share the part that I think is relevant. Feel free to ask me for more info.\r\n\r\n```python\r\nclass MyModel(pytorch_lightning.LightningModule)\r\n def setup(self, stage):\r\n self.dataset = datasets.load_from_disk(path)\r\n self.dataset.set_format(\"torch\")\r\n\r\n def train_dataloader(self):\r\n collate_fn = transformers.DataCollatorForLanguageModeling(\r\n tokenizer=transformers.ElectraTokenizerFast.from_pretrained(tok_path)\r\n )\r\n dataloader = torch.utils.DataLoader(\r\n self.dataset,\r\n batch_size=32,\r\n collate_fn=collate_fn,\r\n num_workers=8,\r\n pin_memory=True,\r\n )\r\n```","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Hi ! Sorry for the delay I haven't had a chance to take a look at this yet. Are you still experiencing this issue ?\r\nI'm asking because the latest patch release 1.6.2 fixed a few memory issues that could have lead to slow downs","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":45,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHi ! Sorry for the delay I haven't had a chance to take a look at this yet. Are you still experiencing this issue ?\r\nI'm asking because the latest patch release 1.6.2 fixed a few memory issues that could have lead to slow downs","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Hi! I just ran the same code with different datasets (one is 60 GB and another 600 GB), and the latter runs much slower. ETA differs by 10x.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":28,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHi! I just ran the same code with different datasets (one is 60 GB and another 600 GB), and the latter runs much slower. ETA differs by 10x.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq and @hwijeen\r\n\r\nDespite upgrading to datasets 1.6.2, still experiencing extremely slow (2h00) loading for a 300Gb local dataset shard size 1.1Gb on local HDD (40Mb\/s read speed). This corresponds almost exactly to total data divided by reading speed implying that it reads the entire dataset at each load.\r\n\r\nStack details:\r\n=========\r\n\r\n> GCC version: Could not collect\r\n> Clang version: Could not collect\r\n> CMake version: Could not collect\r\n> \r\n> Python version: 3.7 (64-bit runtime)\r\n> Is CUDA available: True\r\n> CUDA runtime version: 10.2.89\r\n> GPU models and configuration: GPU 0: GeForce GTX 1050\r\n> Nvidia driver version: 457.63\r\n> cuDNN version: C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\cudnn64_7.dll\r\n> HIP runtime version: N\/A\r\n> MIOpen runtime version: N\/A\r\n> \r\n> Versions of relevant libraries:\r\n> [pip3] datasets==1.6.2\r\n> [pip3] transformers==4.5.1\r\n> [pip3] numpy==1.19.1\r\n> [pip3] numpydoc==1.1.0\r\n> [pip3] pytorch-metric-learning==0.9.98\r\n> [pip3] torch==1.8.1\r\n> [pip3] torchaudio==0.8.1\r\n> [pip3] torchvision==0.2.2\r\n> [conda] blas 2.16 mkl conda-forge\r\n> [conda] cudatoolkit 10.2.89 hb195166_8 conda-forge\r\n> [conda] libblas 3.8.0 16_mkl conda-forge\r\n> [conda] libcblas 3.8.0 16_mkl conda-forge\r\n> [conda] liblapack 3.8.0 16_mkl conda-forge\r\n> [conda] liblapacke 3.8.0 16_mkl conda-forge\r\n> [conda] mkl 2020.1 216\r\n> [conda] numpy 1.19.1 py37hae9e721_0 conda-forge\r\n> [conda] numpydoc 1.1.0 py_1 conda-forge\r\n> [conda] pytorch 1.8.1 py3.7_cuda10.2_cudnn7_0 pytorch\r\n> [conda] pytorch-metric-learning 0.9.98 pyh39e3cac_0 metric-learning\r\n> [conda] torchaudio 0.8.1 py37 pytorch\r\n> [conda] torchvision 0.2.2 py_3 pytorch","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":227,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq and @hwijeen\r\n\r\nDespite upgrading to datasets 1.6.2, still experiencing extremely slow (2h00) loading for a 300Gb local dataset shard size 1.1Gb on local HDD (40Mb\/s read speed). This corresponds almost exactly to total data divided by reading speed implying that it reads the entire dataset at each load.\r\n\r\nStack details:\r\n=========\r\n\r\n> GCC version: Could not collect\r\n> Clang version: Could not collect\r\n> CMake version: Could not collect\r\n> \r\n> Python version: 3.7 (64-bit runtime)\r\n> Is CUDA available: True\r\n> CUDA runtime version: 10.2.89\r\n> GPU models and configuration: GPU 0: GeForce GTX 1050\r\n> Nvidia driver version: 457.63\r\n> cuDNN version: C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\cudnn64_7.dll\r\n> HIP runtime version: N\/A\r\n> MIOpen runtime version: N\/A\r\n> \r\n> Versions of relevant libraries:\r\n> [pip3] datasets==1.6.2\r\n> [pip3] transformers==4.5.1\r\n> [pip3] numpy==1.19.1\r\n> [pip3] numpydoc==1.1.0\r\n> [pip3] pytorch-metric-learning==0.9.98\r\n> [pip3] torch==1.8.1\r\n> [pip3] torchaudio==0.8.1\r\n> [pip3] torchvision==0.2.2\r\n> [conda] blas 2.16 mkl conda-forge\r\n> [conda] cudatoolkit 10.2.89 hb195166_8 conda-forge\r\n> [conda] libblas 3.8.0 16_mkl conda-forge\r\n> [conda] libcblas 3.8.0 16_mkl conda-forge\r\n> [conda] liblapack 3.8.0 16_mkl conda-forge\r\n> [conda] liblapacke 3.8.0 16_mkl conda-forge\r\n> [conda] mkl 2020.1 216\r\n> [conda] numpy 1.19.1 py37hae9e721_0 conda-forge\r\n> [conda] numpydoc 1.1.0 py_1 conda-forge\r\n> [conda] pytorch 1.8.1 py3.7_cuda10.2_cudnn7_0 pytorch\r\n> [conda] pytorch-metric-learning 0.9.98 pyh39e3cac_0 metric-learning\r\n> [conda] torchaudio 0.8.1 py37 pytorch\r\n> [conda] torchvision 0.2.2 py_3 pytorch","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Hi @BenoitDalFerro how do your load your dataset ?","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":9,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHi @BenoitDalFerro how do your load your dataset ?","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Hi @lhoestq thanks for the quick turn-around, actually the plain vanilla way, without an particular knack or fashion, I tried to look into the documentation for some alternative but couldn't find any\r\n\r\n> dataset = load_from_disk(dataset_path=os.path.join(datasets_dir,dataset_dir))","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":36,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHi @lhoestq thanks for the quick turn-around, actually the plain vanilla way, without an particular knack or fashion, I tried to look into the documentation for some alternative but couldn't find any\r\n\r\n> dataset = load_from_disk(dataset_path=os.path.join(datasets_dir,dataset_dir))","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"I\u2019m facing the same issue when loading a 900GB dataset (stored via `save_to_disk`): `load_from_disk(path_to_dir)` takes 1.5 hours and htop consistently shows high IO rates > 120 M\/s.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":27,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nI\u2019m facing the same issue when loading a 900GB dataset (stored via `save_to_disk`): `load_from_disk(path_to_dir)` takes 1.5 hours and htop consistently shows high IO rates > 120 M\/s.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@tsproisl same here, smells like ~~teen spirit~~ intended generator inadvertently ending up iterator\r\n\r\n@lhoestq perhaps solution to detect bug location in code is to track its signature via HD read usage monitoring, option is to add tracking decorator on top each function and sequentially close all hatches from top to bottom, suggest PySmart https:\/\/pypi.org\/project\/pySMART\/ a Smartmontools implementation","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":57,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@tsproisl same here, smells like ~~teen spirit~~ intended generator inadvertently ending up iterator\r\n\r\n@lhoestq perhaps solution to detect bug location in code is to track its signature via HD read usage monitoring, option is to add tracking decorator on top each function and sequentially close all hatches from top to bottom, suggest PySmart https:\/\/pypi.org\/project\/pySMART\/ a Smartmontools implementation","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"I wasn't able to reproduce this on a toy dataset of around 300GB:\r\n\r\n```python\r\nimport datasets as ds\r\n\r\ns = ds.load_dataset(\"squad\", split=\"train\")\r\ns4000 = ds.concatenate_datasets([s] * 4000)\r\nprint(ds.utils.size_str(s4000.data.nbytes)) # '295.48 GiB'\r\n\r\ns4000.save_to_disk(\"tmp\/squad_4000\")\r\n```\r\n\r\n```python\r\nimport psutil\r\nimport time\r\nfrom datasets import load_from_disk\r\n\r\ndisk = \"disk0\" # You may have to change your disk here\r\niocnt1 = psutil.disk_io_counters(perdisk=True)[disk]\r\ntime1 = time.time()\r\n\r\ns4000_reloaded = load_from_disk(\"tmp\/squad_4000\")\r\n\r\ntime2 = time.time()\r\niocnt2 = psutil.disk_io_counters(perdisk=True)[disk]\r\n\r\nprint(f\"Blocks read {iocnt2.read_count - iocnt1.read_count}\") # Blocks read 18\r\nprint(f\"Elapsed time: {time2 - time1:.02f}s\") # Elapsed time: 14.60s\r\n```\r\n\r\nCould you run this on your side and tell me if how much time it takes ? Please run this when your machine is idle so that other processes don't interfere.\r\n\r\nI got these results on my macbook pro on datasets 1.6.2","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":130,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nI wasn't able to reproduce this on a toy dataset of around 300GB:\r\n\r\n```python\r\nimport datasets as ds\r\n\r\ns = ds.load_dataset(\"squad\", split=\"train\")\r\ns4000 = ds.concatenate_datasets([s] * 4000)\r\nprint(ds.utils.size_str(s4000.data.nbytes)) # '295.48 GiB'\r\n\r\ns4000.save_to_disk(\"tmp\/squad_4000\")\r\n```\r\n\r\n```python\r\nimport psutil\r\nimport time\r\nfrom datasets import load_from_disk\r\n\r\ndisk = \"disk0\" # You may have to change your disk here\r\niocnt1 = psutil.disk_io_counters(perdisk=True)[disk]\r\ntime1 = time.time()\r\n\r\ns4000_reloaded = load_from_disk(\"tmp\/squad_4000\")\r\n\r\ntime2 = time.time()\r\niocnt2 = psutil.disk_io_counters(perdisk=True)[disk]\r\n\r\nprint(f\"Blocks read {iocnt2.read_count - iocnt1.read_count}\") # Blocks read 18\r\nprint(f\"Elapsed time: {time2 - time1:.02f}s\") # Elapsed time: 14.60s\r\n```\r\n\r\nCould you run this on your side and tell me if how much time it takes ? Please run this when your machine is idle so that other processes don't interfere.\r\n\r\nI got these results on my macbook pro on datasets 1.6.2","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq thanks, test running as we speak, bear with me","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":10,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq thanks, test running as we speak, bear with me","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Just tried on google colab and got ~1min for a 15GB dataset (only 200 times SQuAD), while it should be instantaneous. The time is spent reading the Apache Arrow table from the memory mapped file. This might come a virtual disk management issue. I'm trying to see if I can still speed it up on colab.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":56,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nJust tried on google colab and got ~1min for a 15GB dataset (only 200 times SQuAD), while it should be instantaneous. The time is spent reading the Apache Arrow table from the memory mapped file. This might come a virtual disk management issue. I'm trying to see if I can still speed it up on colab.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq what is Google Colab's HD read speed, is it possible to introspect incl. make like SSD or HDD ?","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":20,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq what is Google Colab's HD read speed, is it possible to introspect incl. make like SSD or HDD ?","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq Thank you! The issue is getting more interesting. The second script is still running, but it's definitely taking much longer than 15 seconds.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":24,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq Thank you! The issue is getting more interesting. The second script is still running, but it's definitely taking much longer than 15 seconds.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Okay, here\u2019s the ouput:\r\nBlocks read 158396\r\nElapsed time: 529.10s\r\n\r\nAlso using datasets 1.6.2. Do you have any ideas, how to pinpoint the problem?","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":24,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nOkay, here\u2019s the ouput:\r\nBlocks read 158396\r\nElapsed time: 529.10s\r\n\r\nAlso using datasets 1.6.2. Do you have any ideas, how to pinpoint the problem?","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq, @tsproisl mmmh still writing on my side about 1h to go, thinking on it are your large datasets all monoblock unsharded ? mine is 335 times 1.18Gb shards.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":29,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq, @tsproisl mmmh still writing on my side about 1h to go, thinking on it are your large datasets all monoblock unsharded ? mine is 335 times 1.18Gb shards.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"The 529.10s was a bit too optimistic. I cancelled the reading process once before running it completely, therefore the harddrive cache probably did its work.\r\n\r\nHere are three consecutive runs\r\nFirst run (freshly written to disk):\r\nBlocks read 309702\r\nElapsed time: 1267.74s\r\nSecond run (immediately after):\r\nBlocks read 113944\r\nElapsed time: 417.55s\r\nThird run (immediately after):\r\nBlocks read 42518\r\nElapsed time: 199.19s\r\n","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":62,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nThe 529.10s was a bit too optimistic. I cancelled the reading process once before running it completely, therefore the harddrive cache probably did its work.\r\n\r\nHere are three consecutive runs\r\nFirst run (freshly written to disk):\r\nBlocks read 309702\r\nElapsed time: 1267.74s\r\nSecond run (immediately after):\r\nBlocks read 113944\r\nElapsed time: 417.55s\r\nThird run (immediately after):\r\nBlocks read 42518\r\nElapsed time: 199.19s\r\n","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq \r\nFirst test\r\n> elapsed time: 11219.05s\r\n\r\nSecond test running bear with me, for Windows users slight trick to modify original \"disk0\" string:\r\n\r\nFirst find physical unit relevant key in dictionnary\r\n```\r\nimport psutil\r\npsutil.disk_io_counters(perdisk=True)\r\n```\r\n\r\n> {'PhysicalDrive0': sdiskio(read_count=18453286, write_count=4075333, read_bytes=479546467840, write_bytes=161590275072, read_time=20659, write_time=2464),\r\n> 'PhysicalDrive1': sdiskio(read_count=1495778, write_count=388781, read_bytes=548628622336, write_bytes=318234849280, read_time=426066, write_time=19085)}\r\n\r\nIn my case it's _PhysicalDrive1_\r\n\r\nThen insert relevant key's string as _disk_ variable\r\n\r\n```\r\npsutil.disk_io_counters()\r\ndisk = 'PhysicalDrive1' # You may have to change your disk here\r\niocnt1 = psutil.disk_io_counters(perdisk=True)[disk]\r\ntime1 = time.time()\r\ns4000_reloaded = load_from_disk(\"your path here\")\r\ntime2 = time.time()\r\niocnt2 = psutil.disk_io_counters(perdisk=True)[disk]\r\nprint(f\"Blocks read {iocnt2.read_count - iocnt1.read_count}\") # Blocks read 18\r\nprint(f\"Elapsed time: {time2 - time1:.02f}s\") # Elapsed time: 14.60s\r\n```","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":115,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq \r\nFirst test\r\n> elapsed time: 11219.05s\r\n\r\nSecond test running bear with me, for Windows users slight trick to modify original \"disk0\" string:\r\n\r\nFirst find physical unit relevant key in dictionnary\r\n```\r\nimport psutil\r\npsutil.disk_io_counters(perdisk=True)\r\n```\r\n\r\n> {'PhysicalDrive0': sdiskio(read_count=18453286, write_count=4075333, read_bytes=479546467840, write_bytes=161590275072, read_time=20659, write_time=2464),\r\n> 'PhysicalDrive1': sdiskio(read_count=1495778, write_count=388781, read_bytes=548628622336, write_bytes=318234849280, read_time=426066, write_time=19085)}\r\n\r\nIn my case it's _PhysicalDrive1_\r\n\r\nThen insert relevant key's string as _disk_ variable\r\n\r\n```\r\npsutil.disk_io_counters()\r\ndisk = 'PhysicalDrive1' # You may have to change your disk here\r\niocnt1 = psutil.disk_io_counters(perdisk=True)[disk]\r\ntime1 = time.time()\r\ns4000_reloaded = load_from_disk(\"your path here\")\r\ntime2 = time.time()\r\niocnt2 = psutil.disk_io_counters(perdisk=True)[disk]\r\nprint(f\"Blocks read {iocnt2.read_count - iocnt1.read_count}\") # Blocks read 18\r\nprint(f\"Elapsed time: {time2 - time1:.02f}s\") # Elapsed time: 14.60s\r\n```","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq\r\nSecond test\r\n\r\n> Blocks read 1265609\r\n> Elapsed time: 11216.55s","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":11,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq\r\nSecond test\r\n\r\n> Blocks read 1265609\r\n> Elapsed time: 11216.55s","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq any luck ?","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":4,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq any luck ?","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Unfortunately no. Thanks for running the benchmark though, it shows that you machine does a lot of read operations. This is not expected: in other machines it does almost no read operations which enables a very fast loading.\r\n\r\nI did some tests on google colab and have the same issue. The first time the dataset arrow file is memory mapped takes always a lot of time (time seems linear with respect to the dataset size). Reloading the dataset is then instantaneous since the arrow file has already been memory mapped.\r\n\r\nI also tried using the Arrow IPC file format (see #1933) instead of the current streaming format that we use but it didn't help.\r\n\r\nMemory mapping is handled by the OS and depends on the disk you're using, so I'm not sure we can do much about it. I'll continue to investigate anyway, because I still don't know why in some cases it would go through the entire file (high `Blocks read ` as in your tests) and in other cases it would do almost no reading.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":177,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nUnfortunately no. Thanks for running the benchmark though, it shows that you machine does a lot of read operations. This is not expected: in other machines it does almost no read operations which enables a very fast loading.\r\n\r\nI did some tests on google colab and have the same issue. The first time the dataset arrow file is memory mapped takes always a lot of time (time seems linear with respect to the dataset size). Reloading the dataset is then instantaneous since the arrow file has already been memory mapped.\r\n\r\nI also tried using the Arrow IPC file format (see #1933) instead of the current streaming format that we use but it didn't help.\r\n\r\nMemory mapping is handled by the OS and depends on the disk you're using, so I'm not sure we can do much about it. I'll continue to investigate anyway, because I still don't know why in some cases it would go through the entire file (high `Blocks read ` as in your tests) and in other cases it would do almost no reading.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"@lhoestq thanks for the effort, let's stay in touch","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":9,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\n@lhoestq thanks for the effort, let's stay in touch","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Just want to say that I am seeing the same issue. Dataset size if 268GB and it takes **3 hours** to load `load_from_disk`, using dataset version `1.9.0`. Filesystem underneath is `Lustre` ","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":31,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nJust want to say that I am seeing the same issue. Dataset size if 268GB and it takes **3 hours** to load `load_from_disk`, using dataset version `1.9.0`. Filesystem underneath is `Lustre` ","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Hi @lhoestq, confirmed Windows issue, exact same code running on Linux OS total loading time about 3 minutes.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":18,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHi @lhoestq, confirmed Windows issue, exact same code running on Linux OS total loading time about 3 minutes.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2252","title":"Slow dataloading with big datasets issue persists","comments":"Hmm that's different from what I got. I was on Ubuntu when reporting the initial issue.","body":"Hi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.","comment_length":16,"text":"Slow dataloading with big datasets issue persists\nHi,\r\n\r\nI reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122).\r\nHowever, the problem seems to persist. Here is the profiled results:\r\n\r\n\r\n1) Running with 60GB\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 517.96 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nmodel_backward \t| 0.26144 \t|100 \t| 26.144 \t| 5.0475 \t|\r\nmodel_forward \t| 0.11123 \t|100 \t| 11.123 \t| 2.1474 \t|\r\nget_train_batch \t| 0.097121 \t|100 \t| 9.7121 \t| 1.8751 \t|\r\n```\r\n\r\n\r\n3) Running with 600GB, datasets==1.6.0\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 4563.2 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nget_train_batch \t| 5.1279 \t|100 \t| 512.79 \t| 11.237 \t|\r\nmodel_backward \t| 4.8394 \t|100 \t| 483.94 \t| 10.605 \t|\r\nmodel_forward \t| 0.12162 \t|100 \t| 12.162 \t| 0.26652 \t|\r\n```\r\n\r\nI see that `get_train_batch` lags when data is large. Could this be related to different issues?\r\nI would be happy to provide necessary information to investigate.\nHmm that's different from what I got. I was on Ubuntu when reporting the initial issue.","embeddings":[-0.4040590525,0.2139611691,-0.134309724,0.1632939279,0.2269002348,-0.1808893532,0.2195484489,0.4007899463,0.1186829358,-0.0092018284,-0.3094537258,0.0660658404,0.1796627045,0.1480765939,-0.1164040118,0.1400319189,0.0100115351,0.0462014377,-0.1018281952,-0.0599145927,0.07170894,-0.1256888807,0.0162408222,-0.2149125636,-0.321829021,-0.035173133,0.2733658254,0.0361014716,-0.2680936158,-0.1831433326,0.2045997828,0.2419507354,0.0882460326,0.3962605298,-0.0001050314,-0.0951045752,0.3049376905,0.0726464018,-0.2473200262,0.1302319169,0.0542529263,-0.336699158,-0.1207264289,0.1083628237,-0.0350148529,0.1210587919,-0.2277355492,-0.0067853495,0.3558999598,0.1554210037,0.258639276,0.0054166568,-0.2216917872,-0.3251929879,0.1556022167,0.2288568765,-0.2061296552,0.316334933,0.4986956418,-0.1909870356,-0.5347768068,0.0420746282,-0.0375930667,0.1030172035,-0.2523976266,-0.064847365,0.1814326197,-0.0184340309,0.2288667262,0.2305987775,0.5177352428,0.2850895524,-0.3112405837,-0.283450067,-0.3231844306,-0.3884500265,0.209275797,-0.0874323472,-0.0877031907,0.0355130248,-0.2518314719,-0.202842623,0.0418954939,-0.069468081,-0.3781086206,-0.1201944426,-0.0730070472,-0.0856000185,0.1283757091,-0.2761492431,0.2117173225,-0.0886533707,0.1636075974,0.101612933,-0.5672925115,-0.0537961908,0.3513720036,0.0552260801,0.152962625,0.2054211497,0.1399331689,0.2763732672,0.0849820599,-0.2167821079,0.3756537437,0.2172034979,-0.207066685,-0.0734169632,0.4972681701,0.1715171486,-0.2257164568,0.0879854187,-0.145631671,-0.0804974213,0.0692875683,-0.2820042074,-0.170788154,-0.3726626635,-0.3188099563,0.252097398,0.1120118052,0.0482922681,0.1105851084,0.3219352365,-0.0813050419,0.4326556325,0.0139108989,-0.1857730746,-0.1517120451,-0.2132649273,-0.2128109038,-0.2773059905,-0.4498109221,0.1449789405,0.4436470568,0.0247756429,0.0010015079,0.2090330273,0.1274003536,-0.010391362,-0.2007555664,-0.3551882505,-0.2446888238,0.1974934787,-0.1794242114,0.3334898651,-0.1749725044,0.150991112,-0.1955755651,0.1619163901,-0.3553528786,-0.2452231944,0.3079734445,0.3025393784,-0.1045797467,-0.044610966,-0.4129046202,0.2642707825,-0.1100947186,0.0590913631,-0.1334376335,-0.0237634797,-0.1063722819,-0.0018752494,0.0808164775,0.2770811319,-0.3943965137,0.1528885067,-0.2211903185,0.1494279951,0.1306662112,0.3977420032,-0.3263542652,0.0876103118,-0.1106215045,0.0166786872,0.1702519059,-0.3009161949,-0.5397740602,0.399830848,-0.2098787725,0.1034381092,0.1209401488,-0.0182545036,0.3900322914,-0.2471283823,0.2662606239,0.5497125983,0.0612694584,0.1136698574,-0.3252187371,-0.1047572196,-0.033708971,0.5643473864,0.1143966317,-0.1104968041,0.1361210346,0.0966812074,0.1597174853,0.052768454,-0.0750206038,0.1345390379,0.1969802827,-0.0893592462,-0.0232280102,0.1839111,-0.4957300127,0.4001031518,-0.0450965427,-0.2439492494,0.3706954122,0.079882279,-0.1051796675,-0.1792069077,-0.1787629128,0.1509504467,0.1051254645,-0.1905723512,-0.0538962521,0.1902869493,0.1383858174,0.3167603016,-0.3021217287,-0.001923556,0.0116819637,0.1035133526,0.1402072906,-0.0813204572,0.1752838492,-0.0305479225,0.0584349036,-0.0661685467,-0.089132227,0.0895663351,0.1268130094,0.3413229883,-0.01244081,0.2423082441,0.1456023753,0.3362042606,0.1086204872,-0.0293664243,0.0543257706,-0.3189198077,-0.2518000305,0.4823022485,-0.0920576677,0.1507884413,0.116451405,-0.326992631,0.4123485386,-0.1484995186,0.255930841,0.2874782383,0.5536789298,0.0951023251,0.4025269747,0.2306759208,-0.2004444003,0.2244458497,0.2324167639,-0.0279092733,-0.3425456583,0.1649187654,-0.0515673012,-0.2804048061,0.0083010131,0.0587804243,0.552093327,0.1125910804,0.0406083763,-0.1931914985,0.4311437011,-0.305651933,0.276478231,-0.0137504013,0.081766732,0.4084943235,0.4574069381,-0.2098321319,-0.4086069167,-0.1135862768,0.3819160461,0.1469848007,-0.07280422,-0.1258427799,-0.203995809,-0.0578330941,-0.0548907034,-0.1194069013,-0.2367260009,-0.2680535913,0.0132276183,0.3044422567,-0.0600870922,0.0084868325,-0.2665054798,0.5148743987,0.0486832932,-0.0543113649,-0.4178192317,-0.0225873962,-0.215959236,0.1583644599,0.162519753,-0.2993919551,0.2113576233,0.1774643362,-0.0072441734,-0.0192260314,-0.3806327879,0.0831742585,0.2474463731,0.047911603,-0.234609589,0.0143209007,0.1991361976,0.0063428385,0.0695415363,-0.3881141245,-0.1075753048,0.0090646623,-0.0538933873,0.1783518642,0.0542122833,-0.2898902297,-0.1657841504,-0.4625604451,0.2143458873,-0.1899102479,0.00080242,0.2282609493,-0.1258067638,0.0544250906,0.0254087951,-0.1182290763,-0.2713723481,-0.5140330195,0.3919060528,-0.0605122671,-0.3088650703,-0.222576797,0.1292104721,-0.0127305621,0.27438429,-0.4091230333,0.1720881313,-0.2697909772,0.1982387453,0.0265607238,-0.0886877701,0.0365646072,-0.2912847102,-0.22062096,-0.0252424628,-0.0509685203,0.1468975991,0.2361966521,0.3168797791,-0.2697795033,0.3415651023,0.1393641233,0.4849933684,0.0297220852,-0.0464570634,0.3064985573,0.3087260723,0.1859564185,-0.2166093439,-0.1917747706,0.0240004919,-0.2419841737,-0.1816830188,0.3324069977,-0.0094316844,-0.4836246669,-0.0519542992,-0.0279166847,0.2211064696,-0.1555524319,0.2269752175,-0.1468743533,0.0044934303,0.1748711467,0.0506794341,-0.2485346347,-0.2570213675,0.0509831756,-0.2897282839,0.1129226014,-0.2460878342,-0.1308770627,0.3246513903,-0.6654969454,0.0506278425,-0.1267722398,0.3393508792,0.1319963634,-0.0208719932,0.2086361498,0.0468039475,0.5165268183,0.089820005,0.1393014491,0.000371122,-0.4541063905,0.0226785857,-0.2141716033,-0.0514461212,0.0486918949,0.3995553851,0.3858709335,-0.3827715218,-0.1480104923,-0.0237370655,-0.0528077073,-0.2226080447,-0.2605672181,-0.0055550253,0.1305884421,0.1284982562,0.2395779788,0.1507945657,0.1218816489,-0.2085506022,-0.0549896881,-0.0681530759,0.182902202,0.2443775684,-0.1095462143,0.1049960256,0.0346527584,0.5330203772,0.2021953464,0.1620398462,0.3916485906,0.5336763263,-0.0214422848,-0.0590329692,0.1207185611,0.0848849565,0.1329710782,0.5663316846,0.1114223227,0.0659486204,-0.3181549907,0.0906942859,-0.1745761782,0.1184808984,0.15540649,-0.1474261582,-0.3094096184,-0.1591176242,0.5107486844,0.0106088221,0.04199218,0.1951280236,-0.2514148653,-0.0468579568,0.0357323438,-0.1320942342,0.8368296623,-0.0260927342,0.104034707,-0.0680799782,-0.0040179966,0.2523504496,-0.3498376012,-0.0505505539,-0.206683591,-0.162843585,0.2075821906,-0.0653010234,-0.039212957,0.2160114795,0.1681978554,0.1074097082,-0.0210481845,0.3766250908,-0.1330876946,0.2871600389,0.1428226978,-0.1990962178,-0.1988931298,0.2173184305,-0.1519214362,-0.0115478151,-0.0593674667,-0.0738370195,-0.0321625434,0.0265141409,-0.0823038891,0.0418742187,-0.1302343309,0.3452497125,-0.0467488356,-0.3351005316,0.1721290499,-0.2237442732,-0.4176707566,-0.0647134706,-0.2056355625,0.3521544039,-0.1278087944,0.3836575449,0.3575843275,-0.2768963873,0.2228883952,-0.0734646693,0.0860969946,-0.1276029646,-0.0200733636,-0.2176635563,-0.2408339232,0.0453494154,0.1819564849,-0.2663728595,-0.1333111078,0.0646921769,0.1159255952,-0.1442123652,0.207165584,0.1312051564,0.1609460413,0.3172070384,0.0431217514,-0.1773989499,-0.2193928957,0.6642938852,-0.0262714047,-0.2106984556,0.3752241135,0.1622522771,-0.3655776381,-0.1216234639,0.031635534,0.3744303882,-0.4561277926,0.3150388896,-0.2884168625,-0.0785708949,0.2572209239,0.0436805077,-0.0155680412,-0.4285440743,0.0068618208,-0.2145426422,-0.2761790454,0.0276390333,-0.0778127387,0.1018890813,-0.3613103628,-0.1052535027,-0.1253169328,0.3255432248,-0.3826213181,0.2146103084,0.0679275692,0.1204808801,-0.1848234534,-0.422363013,-0.0113013424,-0.0841144547,0.107070066,0.1057987958,-0.1574967355,-0.3501004279,-0.0694133639,0.1125117615,-0.0117709804,-0.0029523456,0.1273091137,-0.3357498646,-0.0294344909,-0.1573072821,0.0086406795,0.1867414415,0.0523041375,-0.3707129359,0.2300569266,0.078483887,-0.0203799624,0.368542254,0.0107488539,0.1475582272,0.1166396588,0.1964166164,-0.1105829552,-0.2409083843,-0.1624659747,-0.1184795052,0.2805272639,-0.04670408,0.3215598762,-0.3202127516,-0.095098421,0.2012293041,0.3612818122,0.5668721199,-0.0242242217,-0.3804124892,0.0732547194,0.2727664411,-0.1539495438,-0.1301027834,0.0856587365,0.0380627438,-0.1332578957,0.0265709255,0.3532831669,-0.1669161916,0.1675893515,0.1402467042,0.446441859,0.0974353179,-0.034831509,0.1065549701,0.1278485805,0.0280681606,0.5872793794,0.259762615,0.0650208294,0.4543121755,0.0269574523,0.3439672291,0.0912302285,0.1599013209,0.0254587959,-0.4160406291,-0.109056361,0.3545891047,-0.3630832434,0.0509817824,-0.0628620014,0.1160134971,0.0191873536,0.0246345028,-0.3914385438,0.0497355871,-0.2921129763,-0.054396823,0.1644137949,-0.1611653119,-0.0829837248,0.2432962358,-0.232341215,0.050789129,0.4543318152,0.290466249,-0.0613234267,-0.4144056737,-0.2544569671,-0.1119512022,0.1596449316,-0.24500525,0.1634246707,-0.0468118787,-0.1688562781,0.1048564762,0.2044881284,0.1834546179,0.3097057343,0.1034324914,-0.0217743237,0.0743550509,-0.0863781124,-0.137756303,0.5175192952,-0.0090704281,0.1067919508,0.1602637172,0.1014398485,-0.3488075435,-0.2175482661,-0.0757729784,-0.1216346845,-0.2066874355,0.1268776357,0.2536277473,-0.102071695,-0.3327085972,-0.203865543,0.0937746689,-0.2198894024,0.2739774883,-0.2362271249,0.0659081489,-0.1743789613,0.1296192706,-0.3521494567,0.5997630954,0.0962916389,0.0063369884,-0.4889294207,-0.0470266528,-0.4901943505,0.1136222556,-0.2887199223,0.3211907744,0.1183495745,0.5177027583,-0.0065310695,0.036375612,-0.0883134902,0.1146806479,0.0202270877,0.0085204644,-0.4558708072,0.2321362942,-0.0280456115,0.0410675928,-0.1092434376,-0.4895465672,0.3563454747,-0.0153234694,0.0828864276,-0.2241468132,-0.1388386935,0.184108153,0.4946778715,0.3996113539,0.0950392932,0.2468270808,0.1787800342,-0.1524983644,-0.3793302476,-0.2371514887,-0.0333090425,-0.1625075787,0.0524967685,-0.0921576396,0.071034506,-0.011771665,0.0052402313,0.1160590723,-0.2415509075,-0.0463599265,-0.168895185,-0.3302365839,0.0503661968,0.3088819385,0.0663308799,0.3367395699,-0.1124826446,0.0192811303,-0.4977087677,-0.3860772848,0.3325393796,-0.361992985,-0.1143822744,-0.2368952483,0.4293966591,0.3034186661,0.2264493108,-0.296094656,-0.0187425856,0.1793894023,-0.1053812802,-0.0332277678,0.4777288139,-0.0485651679,0.0319637731,-0.0503286235,0.2442066073,0.0583079048,-0.2495268136,-0.1617683172,-0.0234537106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2250","title":"some issue in loading local txt file as Dataset for run_mlm.py","comments":"Hi,\r\n\r\n1. try\r\n ```python\r\n dataset = load_dataset(\"text\", data_files={\"train\": [\"a1.txt\", \"b1.txt\"], \"test\": [\"c1.txt\"]})\r\n ```\r\n instead.\r\n\r\n Sadly, I can't reproduce the error on my machine. If the above code doesn't resolve the issue, try to update the library to the \r\n newest version (`pip install datasets --upgrade`).\r\n\r\n2. https:\/\/github.com\/huggingface\/transformers\/blob\/3ed5e97ba04ce9b24b4a7161ea74572598a4c480\/examples\/pytorch\/language-modeling\/run_mlm.py#L258-L259\r\nThis is the original code. You'll have to modify the example source to work with multiple train files. To make it easier, let's say \"|\" will act as a delimiter between files:\r\n ```python\r\n if data_args.train_file is not None:\r\n data_files[\"train\"] = data_args.train_file.split(\"|\") # + .split(\"|\")\r\n ```\r\n Then call the script as follows (**dataset_name must be None**):\r\n ```bash\r\n python run_mlm.py [... other args] --train_file a1.txt|b1.txt\r\n ```","body":"![image](https:\/\/user-images.githubusercontent.com\/14968123\/115773877-18cef300-a3c6-11eb-8e58-a9cbfd1001ec.png)\r\n\r\nfirst of all, I tried to load 3 .txt files as a dataset (sure that the directory and permission is OK.), I face with the below error.\r\n\r\n> FileNotFoundError: [Errno 2] No such file or directory: 'c'\r\n\r\nby removing one of the training .txt files It's fixed and although if I put all file as training it's ok\r\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115774207-867b1f00-a3c6-11eb-953b-905cfb112d25.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115774264-9b57b280-a3c6-11eb-9f36-7b109f0e5a31.png)\r\n\r\n\r\nafter this, my question is how could I use this defined Dataset for run_mlm.py for from scratch pretraining.\r\nby using --train_file path_to_train_file just can use one .txt , .csv or, .json file. I tried to set my defined Dataset as --dataset_name but the below issue occurs.\r\n\r\n\r\n> Traceback (most recent call last):\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 336, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/dataset\/dataset.py\r\n\r\n> During handling of the above exception, another exception occurred:\r\n\r\n> Traceback (most recent call last):\r\n File \"run_mlm.py\", line 486, in \r\n main()\r\n File \"run_mlm.py\", line 242, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir)\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 719, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 347, in prepare_module\r\n combined_path, github_file_path\r\nFileNotFoundError: Couldn't find file locally at dataset\/dataset.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.6.0\/datasets\/dataset\/dataset.py.\r\nThe file is also not present on the master branch on github.\r\n","comment_length":110,"text":"some issue in loading local txt file as Dataset for run_mlm.py\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115773877-18cef300-a3c6-11eb-8e58-a9cbfd1001ec.png)\r\n\r\nfirst of all, I tried to load 3 .txt files as a dataset (sure that the directory and permission is OK.), I face with the below error.\r\n\r\n> FileNotFoundError: [Errno 2] No such file or directory: 'c'\r\n\r\nby removing one of the training .txt files It's fixed and although if I put all file as training it's ok\r\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115774207-867b1f00-a3c6-11eb-953b-905cfb112d25.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115774264-9b57b280-a3c6-11eb-9f36-7b109f0e5a31.png)\r\n\r\n\r\nafter this, my question is how could I use this defined Dataset for run_mlm.py for from scratch pretraining.\r\nby using --train_file path_to_train_file just can use one .txt , .csv or, .json file. I tried to set my defined Dataset as --dataset_name but the below issue occurs.\r\n\r\n\r\n> Traceback (most recent call last):\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 336, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/dataset\/dataset.py\r\n\r\n> During handling of the above exception, another exception occurred:\r\n\r\n> Traceback (most recent call last):\r\n File \"run_mlm.py\", line 486, in \r\n main()\r\n File \"run_mlm.py\", line 242, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir)\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 719, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 347, in prepare_module\r\n combined_path, github_file_path\r\nFileNotFoundError: Couldn't find file locally at dataset\/dataset.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.6.0\/datasets\/dataset\/dataset.py.\r\nThe file is also not present on the master branch on github.\r\n\nHi,\r\n\r\n1. try\r\n ```python\r\n dataset = load_dataset(\"text\", data_files={\"train\": [\"a1.txt\", \"b1.txt\"], \"test\": [\"c1.txt\"]})\r\n ```\r\n instead.\r\n\r\n Sadly, I can't reproduce the error on my machine. If the above code doesn't resolve the issue, try to update the library to the \r\n newest version (`pip install datasets --upgrade`).\r\n\r\n2. https:\/\/github.com\/huggingface\/transformers\/blob\/3ed5e97ba04ce9b24b4a7161ea74572598a4c480\/examples\/pytorch\/language-modeling\/run_mlm.py#L258-L259\r\nThis is the original code. You'll have to modify the example source to work with multiple train files. To make it easier, let's say \"|\" will act as a delimiter between files:\r\n ```python\r\n if data_args.train_file is not None:\r\n data_files[\"train\"] = data_args.train_file.split(\"|\") # + .split(\"|\")\r\n ```\r\n Then call the script as follows (**dataset_name must be None**):\r\n ```bash\r\n python run_mlm.py [... other args] --train_file a1.txt|b1.txt\r\n ```","embeddings":[-0.2187605202,-0.1667418778,0.0313900374,0.4202873111,0.4102846384,0.2086267918,0.3599502444,0.3321821988,0.0959176049,-0.0865045115,0.1432959884,0.1121851429,-0.313026011,0.1154056862,0.2711088955,-0.1855070144,-0.0543438829,0.1169156954,0.0061769388,-0.0527790189,-0.0572860017,0.2120701969,-0.0856259763,0.0367862619,-0.5357351303,-0.1124623269,-0.0895917416,0.5331773758,-0.1113560572,-0.4210807085,0.2054494172,-0.2190006226,0.5652353764,0.3786007762,-0.0001178408,0.1366185844,0.0195880178,-0.3046211302,-0.2785406709,-0.3674179018,-0.0473928452,-0.1517332792,0.3131980002,-0.2982816398,-0.2849579453,-0.2986627519,0.1183683649,-0.4172201753,0.5750330687,0.5418713689,0.1464993954,-0.1126833931,-0.1135130003,0.1661650389,0.1643294096,0.1989790797,0.0532440953,0.4563586712,0.235948056,-0.186613515,0.1223256141,0.0028428503,-0.0856467262,0.2835535109,0.4207870364,0.2291142344,0.2676959634,-0.3368067145,-0.0659586936,0.4293656945,0.5069271922,-0.4630822241,-0.1613177806,-0.2148654163,0.0965309665,-0.5334337354,0.1466606557,0.1705652922,-0.169303,0.1613130867,-0.0401136689,-0.1284488291,-0.2960370183,0.3532934189,0.1373363286,-0.004697883,-0.2373589724,0.1342300028,-0.0491270348,0.2122362554,0.1465490609,-0.1308979094,0.116060406,0.352633208,-0.2304416448,0.0151426448,-0.3056645989,-0.228691265,-0.104920052,-0.2218109965,0.0616082363,-0.1123377234,0.0158279967,0.1886262298,-0.0349827595,0.1783737838,-0.0147651555,0.2518513799,0.1695749611,-0.0863323212,-0.2277290672,-0.1104219034,-0.4364836216,-0.3626224995,-0.1716032326,0.186109066,-0.0098632099,-0.2603826821,-0.1868661791,-0.0135928411,-0.4629870653,-0.1196219325,0.050188873,0.4206337333,0.0471919477,-0.2701141834,0.1909436882,0.3114246428,-0.2982201278,-0.0205369294,-0.1354185492,0.2676131725,-0.1667536795,0.0067309653,0.2710702419,-0.0559694432,0.5489364862,-0.1468862593,0.0182242636,-0.1580575705,0.2986269891,-0.3071927726,0.1116289347,0.3208245039,-0.136329487,-0.114771083,0.2910399139,-0.0443102866,-0.1029344946,0.1777272969,-0.3631927967,-0.3632000387,0.0566768944,0.0550087988,-0.0763354525,0.0673879683,-0.0280140806,0.1393100917,0.3043964207,0.0021818108,-0.1248049214,-0.4227061868,-0.2112958878,-0.1633839011,0.3459704518,0.5008915663,-0.327234298,0.1389868855,-0.0614091307,0.0824354514,0.1293088347,0.1791763306,-0.4740315676,0.5258911252,-0.3562895656,0.2602771819,0.3359387815,-0.3661050797,0.2029674649,0.4296671152,-0.0621659532,-0.0405628383,0.1338306367,-0.0786804929,-0.1721276343,-0.0613215342,0.1770077497,0.224434793,-0.100840956,0.0733385086,-0.1971501857,-0.2577080429,-0.0527974516,0.1942562163,-0.0366877504,0.076797992,0.1875154078,-0.0004553866,0.1252678931,-0.267814666,0.120769605,0.3422435224,0.0662675872,0.1107200086,0.0467288233,0.1225750372,-0.3514122665,0.1639356613,-0.0799138099,-0.0736305416,-0.1430327296,-0.0644586608,-0.1137371659,-0.2062303871,-0.3944917619,-0.2405236363,0.0265963133,0.021793874,0.0545740873,0.0177385323,-0.3444501162,0.1442458928,-0.487413466,-0.0421785899,-0.0992305279,0.2021943778,0.0164213609,0.0035270979,0.0593575947,0.0558211654,0.0060588713,-0.32654953,-0.1184618026,0.3234590888,0.0528179258,0.0771333128,-0.1485342085,0.2100295573,0.1040451527,-0.0613660179,0.1945244819,0.0633509085,0.3194554448,-0.1345395893,-0.199500218,0.3969226778,0.0695660785,0.0165768638,0.1155702099,-0.2274149358,0.1046558321,-0.0991497263,-0.0072395988,-0.1597729623,0.2439163327,0.0610595495,0.1981470138,-0.1272440404,-0.3091569841,-0.2142795324,0.3216104507,0.254540652,0.0373857655,0.0955028683,-0.1733588725,0.0505982302,0.0459097587,0.2077760249,0.4979125857,0.0192748941,-0.2003566772,0.3725386858,0.0216428265,-0.2439142764,0.1191153452,-0.0843787342,0.3388143182,0.4956532717,-0.2445803583,-0.0676739663,-0.214152649,-0.0274478383,0.303414017,0.4202139974,-0.5494006276,0.2633036077,-0.3707825541,-0.1403633952,-0.4049875438,0.1483595371,-0.033233989,0.133253485,-0.3153639436,0.1717948765,0.248615548,0.0067794635,-0.1144893393,-0.0167532135,0.0440500714,-0.3137128651,0.09483096,-0.1302260309,-0.2783370316,-0.0179274641,0.5486425757,0.1192936152,0.0776710883,-0.1489702761,-0.3432919681,0.0173715539,-0.0160412099,0.1359623522,0.038618587,0.3947208822,-0.000171923,0.3190338016,-0.2006282955,-0.0312068965,0.1276067644,0.1451542974,-0.1676795632,-0.1626083255,0.004552512,-0.2500949204,-0.1370775998,-0.6858372688,-0.5413176417,-0.1693502814,0.0753475353,0.2003324926,0.194726333,0.1813874841,0.1762078702,0.2450316846,0.0561480038,0.1711566001,-0.0381061994,-0.4465243518,0.226453945,-0.2394654751,-0.2247730196,0.3580351472,-0.0442435667,0.021543989,0.1005166247,-0.5381856561,-0.1136736795,0.2504937351,-0.0290173553,-0.0123384399,-0.0333243571,0.2848323584,-0.0811643004,0.1176243722,-0.0744209588,-0.0750229433,0.3046918213,0.1528561562,0.2563612759,0.3320654631,0.531327486,-0.0368128642,0.7044878006,0.236127004,-0.2257723808,0.3754787147,-0.30376333,0.1068046167,-0.1427258849,-0.1135770157,0.2218452841,0.0835338235,-0.2613663375,0.179017365,0.047193829,0.0657031909,-0.2824084461,0.3268254995,-0.0857608318,-0.1656234562,0.2103683352,-0.1576741338,0.4189042449,0.0357233547,0.1311354786,0.1456913054,-0.2054309994,-0.0827773064,0.56676507,0.1291620135,0.2777432203,-0.557723701,-0.3043447435,-0.3477812111,0.1395135224,0.1647335589,0.487090826,-0.1520963311,-0.0405335687,0.3218385875,0.0231596995,0.4743595421,-0.2974193096,-0.0139402058,0.3608810306,-0.0608358569,0.0712911412,-0.1211801544,0.1311718971,0.5619038343,-0.0362299606,0.4016287327,-0.1029695719,-0.2938934267,0.0835931525,0.2028924078,-0.1985018253,-0.2078430355,-0.1692018211,-0.1432550997,-0.299744904,-0.2505711019,0.0571173355,0.1843994558,-0.0075563649,0.2715389729,-0.0176868737,0.1675961465,0.2338578403,0.0625933111,0.2957535386,0.1207596287,0.1210656241,0.266847223,0.3667298257,0.4993268549,0.5277525783,-0.4694200456,-0.1614117175,0.1477746516,-0.1777692586,0.3248339593,0.2907809913,-0.2040279508,-0.0262716487,0.2625283599,-0.2953740954,-0.0472016893,0.2233002335,0.2479669005,-0.0097966846,-0.0658067763,-0.5862109661,0.2136285752,0.052174639,0.1908984035,0.1354257166,-0.6723163128,-0.44629246,0.2254086584,-0.012819292,0.7663846612,-0.2099230289,0.4583087265,0.1677959114,0.2719201148,-0.0255540535,-0.4253872037,0.274053216,-0.4830499887,0.0642511547,-0.1064785421,-0.1500697732,0.2763530314,0.1562802792,0.0164983142,0.1594835073,0.0549421683,0.1809711456,-0.0730056837,0.3219032288,0.2434393018,-0.1929605007,-0.2359692454,0.0102940034,-0.1801214814,0.210058108,-0.1129737347,-0.1887555271,-0.1296449006,0.0862196535,0.0653886721,-0.09072043,-0.6141549349,0.1236642599,-0.1575998515,-0.3965589702,0.2871089578,0.3647828996,0.1706037819,-0.0710478276,-0.295419544,0.1321610063,0.112256825,-0.1842178553,-0.03295191,-0.0516680852,0.2976560891,-0.0925220028,0.0413704999,0.2050380111,-0.3047454953,-0.3144062459,-0.0661679506,-0.2993863523,0.5104281306,-0.1897246838,-0.4631552398,0.0700422227,0.0860522017,0.0621171445,0.0842691138,0.1648372412,-0.238874197,0.1868060082,-0.0325410627,-0.4317283332,0.0277171861,0.5185140371,-0.0289326627,0.0790055841,0.565784812,0.3568657935,-0.0299055073,-0.2044344395,0.3942812085,0.0715577677,-0.0699861944,0.1138019711,0.4307899475,0.241881907,0.0622125529,0.2950501144,0.2360558957,-0.1236714348,0.2050521076,-0.4611515403,-0.5064795613,-0.101691626,-0.0835034996,0.2456608713,0.171657607,0.1761143655,0.0406590216,-0.2078078687,-0.2155595422,0.2344577909,-0.0199173559,0.1433347613,0.2096408606,0.081258744,0.3789624274,0.0672319233,0.0618059337,-0.0272171125,-0.3651145995,-0.1301807314,-0.0792092979,0.18247585,0.0090726735,-0.3380416036,-0.0017499232,-0.5721231103,-0.2706252337,-0.0220801271,0.0168050844,0.0830389261,-0.0859040692,0.0943692848,-0.1830237806,-0.0719616339,-0.1091131717,-0.0538006276,-0.1991327852,0.0114911059,0.0659253523,0.1101860553,-0.0955982879,0.0075461208,-0.0074562207,-0.09169361,-0.2438067943,0.1240446195,0.232803762,-0.4102424085,0.189899981,-0.2137137949,0.2356956154,0.4120396972,-0.4174543321,0.2346376926,0.5493932366,0.0715426058,-0.1191133782,0.1600171775,0.3307400644,-0.0316046625,-0.1106172353,0.1972023696,0.3799904883,-0.0352505222,-0.1929016262,0.0609401949,0.3784879148,-0.1502983421,0.0640601888,0.297532171,-0.1603610069,0.1333838701,0.2091663182,0.3601654768,0.1545427144,0.524029851,-0.0822410136,-0.0931553021,0.0597373247,0.0534929745,-0.3930665851,-0.6014754176,0.1731493026,0.2167733014,-0.1283470839,0.1240105629,-0.1549286842,0.4967992008,-0.0010901182,-0.2926058471,0.0142848277,0.0785399377,-0.2275688648,-0.1339007467,-0.4213933349,-0.1551742405,0.0502549112,0.0063855299,-0.1967864782,-0.1983109117,-0.1814775318,0.2520407736,-0.1487800628,-0.1792172939,0.092628859,0.0937754661,-0.1402390301,-0.2473496497,0.303843528,-0.1089146882,-0.0788340047,0.1995709985,-0.1509103328,0.2978542447,0.3143269718,-0.1724802405,-0.21891509,0.0805384442,0.0941954106,0.2014834434,0.0283289105,0.1573800296,0.1723032594,0.3426735699,-0.0058866218,0.0086851269,0.316616565,0.0011797708,-0.0813740045,-0.0921972319,-0.058974985,0.2395663857,-0.0096713621,-0.0538717359,0.1466577351,-0.3901590109,0.0225809664,0.7142034173,-0.0171017461,0.3614580333,-0.0683395937,0.0325171649,-0.1164210215,0.3474167287,0.1298307776,-0.0834638029,-0.406519711,-0.0060042422,-0.6180542111,-0.0460275933,-0.1617492586,-0.1863463521,-0.1596548855,0.0043891231,0.1381011903,-0.0324810892,-0.1281869113,-0.0901290774,0.1696688384,0.3214869201,-0.0856050849,-0.2258813679,-0.1679627597,0.3584050536,-0.0031588958,-0.3093914986,-0.0893948153,-0.1936803013,-0.0117186802,-0.2749819756,0.0232634321,0.1074019521,0.2027430683,0.2969765067,-0.2101576924,0.4309168756,0.0065391446,0.1876882464,-0.1705281138,0.0723379701,-0.0811829865,0.1116091833,0.0496192425,0.4050960243,-0.529505074,-0.11736615,-0.5370402336,-0.0946171582,0.0156083265,-0.2459225059,-0.1738643795,0.0654882789,-0.1364307702,0.2675381899,0.2362120152,0.3413186967,-0.0979718938,-0.0337588191,-0.3247717917,-0.0526789241,0.5630751252,-0.4386031926,0.0079312734,0.1063509285,0.0169814751,0.1237151548,0.1071063057,-0.3041403294,0.0648378134,0.4830067158,-0.1889516562,-0.2679226398,0.3511603475,-0.0312894993,0.0399591774,0.023728447,0.0085910037,-0.0142722279,-0.3185524046,-0.3784615397,-0.4073459506]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2250","title":"some issue in loading local txt file as Dataset for run_mlm.py","comments":"i meet the same error with datasets 1.11.0, is there any insight about this?","body":"![image](https:\/\/user-images.githubusercontent.com\/14968123\/115773877-18cef300-a3c6-11eb-8e58-a9cbfd1001ec.png)\r\n\r\nfirst of all, I tried to load 3 .txt files as a dataset (sure that the directory and permission is OK.), I face with the below error.\r\n\r\n> FileNotFoundError: [Errno 2] No such file or directory: 'c'\r\n\r\nby removing one of the training .txt files It's fixed and although if I put all file as training it's ok\r\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115774207-867b1f00-a3c6-11eb-953b-905cfb112d25.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115774264-9b57b280-a3c6-11eb-9f36-7b109f0e5a31.png)\r\n\r\n\r\nafter this, my question is how could I use this defined Dataset for run_mlm.py for from scratch pretraining.\r\nby using --train_file path_to_train_file just can use one .txt , .csv or, .json file. I tried to set my defined Dataset as --dataset_name but the below issue occurs.\r\n\r\n\r\n> Traceback (most recent call last):\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 336, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/dataset\/dataset.py\r\n\r\n> During handling of the above exception, another exception occurred:\r\n\r\n> Traceback (most recent call last):\r\n File \"run_mlm.py\", line 486, in \r\n main()\r\n File \"run_mlm.py\", line 242, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir)\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 719, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 347, in prepare_module\r\n combined_path, github_file_path\r\nFileNotFoundError: Couldn't find file locally at dataset\/dataset.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.6.0\/datasets\/dataset\/dataset.py.\r\nThe file is also not present on the master branch on github.\r\n","comment_length":14,"text":"some issue in loading local txt file as Dataset for run_mlm.py\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115773877-18cef300-a3c6-11eb-8e58-a9cbfd1001ec.png)\r\n\r\nfirst of all, I tried to load 3 .txt files as a dataset (sure that the directory and permission is OK.), I face with the below error.\r\n\r\n> FileNotFoundError: [Errno 2] No such file or directory: 'c'\r\n\r\nby removing one of the training .txt files It's fixed and although if I put all file as training it's ok\r\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115774207-867b1f00-a3c6-11eb-953b-905cfb112d25.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/14968123\/115774264-9b57b280-a3c6-11eb-9f36-7b109f0e5a31.png)\r\n\r\n\r\nafter this, my question is how could I use this defined Dataset for run_mlm.py for from scratch pretraining.\r\nby using --train_file path_to_train_file just can use one .txt , .csv or, .json file. I tried to set my defined Dataset as --dataset_name but the below issue occurs.\r\n\r\n\r\n> Traceback (most recent call last):\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 336, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py\", line 291, in cached_path\r\n use_auth_token=download_config.use_auth_token,\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py\", line 621, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/dataset\/dataset.py\r\n\r\n> During handling of the above exception, another exception occurred:\r\n\r\n> Traceback (most recent call last):\r\n File \"run_mlm.py\", line 486, in \r\n main()\r\n File \"run_mlm.py\", line 242, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir)\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 719, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py\", line 347, in prepare_module\r\n combined_path, github_file_path\r\nFileNotFoundError: Couldn't find file locally at dataset\/dataset.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.6.0\/datasets\/dataset\/dataset.py.\r\nThe file is also not present on the master branch on github.\r\n\ni meet the same error with datasets 1.11.0, is there any insight about this?","embeddings":[-0.2187605202,-0.1667418778,0.0313900374,0.4202873111,0.4102846384,0.2086267918,0.3599502444,0.3321821988,0.0959176049,-0.0865045115,0.1432959884,0.1121851429,-0.313026011,0.1154056862,0.2711088955,-0.1855070144,-0.0543438829,0.1169156954,0.0061769388,-0.0527790189,-0.0572860017,0.2120701969,-0.0856259763,0.0367862619,-0.5357351303,-0.1124623269,-0.0895917416,0.5331773758,-0.1113560572,-0.4210807085,0.2054494172,-0.2190006226,0.5652353764,0.3786007762,-0.0001178408,0.1366185844,0.0195880178,-0.3046211302,-0.2785406709,-0.3674179018,-0.0473928452,-0.1517332792,0.3131980002,-0.2982816398,-0.2849579453,-0.2986627519,0.1183683649,-0.4172201753,0.5750330687,0.5418713689,0.1464993954,-0.1126833931,-0.1135130003,0.1661650389,0.1643294096,0.1989790797,0.0532440953,0.4563586712,0.235948056,-0.186613515,0.1223256141,0.0028428503,-0.0856467262,0.2835535109,0.4207870364,0.2291142344,0.2676959634,-0.3368067145,-0.0659586936,0.4293656945,0.5069271922,-0.4630822241,-0.1613177806,-0.2148654163,0.0965309665,-0.5334337354,0.1466606557,0.1705652922,-0.169303,0.1613130867,-0.0401136689,-0.1284488291,-0.2960370183,0.3532934189,0.1373363286,-0.004697883,-0.2373589724,0.1342300028,-0.0491270348,0.2122362554,0.1465490609,-0.1308979094,0.116060406,0.352633208,-0.2304416448,0.0151426448,-0.3056645989,-0.228691265,-0.104920052,-0.2218109965,0.0616082363,-0.1123377234,0.0158279967,0.1886262298,-0.0349827595,0.1783737838,-0.0147651555,0.2518513799,0.1695749611,-0.0863323212,-0.2277290672,-0.1104219034,-0.4364836216,-0.3626224995,-0.1716032326,0.186109066,-0.0098632099,-0.2603826821,-0.1868661791,-0.0135928411,-0.4629870653,-0.1196219325,0.050188873,0.4206337333,0.0471919477,-0.2701141834,0.1909436882,0.3114246428,-0.2982201278,-0.0205369294,-0.1354185492,0.2676131725,-0.1667536795,0.0067309653,0.2710702419,-0.0559694432,0.5489364862,-0.1468862593,0.0182242636,-0.1580575705,0.2986269891,-0.3071927726,0.1116289347,0.3208245039,-0.136329487,-0.114771083,0.2910399139,-0.0443102866,-0.1029344946,0.1777272969,-0.3631927967,-0.3632000387,0.0566768944,0.0550087988,-0.0763354525,0.0673879683,-0.0280140806,0.1393100917,0.3043964207,0.0021818108,-0.1248049214,-0.4227061868,-0.2112958878,-0.1633839011,0.3459704518,0.5008915663,-0.327234298,0.1389868855,-0.0614091307,0.0824354514,0.1293088347,0.1791763306,-0.4740315676,0.5258911252,-0.3562895656,0.2602771819,0.3359387815,-0.3661050797,0.2029674649,0.4296671152,-0.0621659532,-0.0405628383,0.1338306367,-0.0786804929,-0.1721276343,-0.0613215342,0.1770077497,0.224434793,-0.100840956,0.0733385086,-0.1971501857,-0.2577080429,-0.0527974516,0.1942562163,-0.0366877504,0.076797992,0.1875154078,-0.0004553866,0.1252678931,-0.267814666,0.120769605,0.3422435224,0.0662675872,0.1107200086,0.0467288233,0.1225750372,-0.3514122665,0.1639356613,-0.0799138099,-0.0736305416,-0.1430327296,-0.0644586608,-0.1137371659,-0.2062303871,-0.3944917619,-0.2405236363,0.0265963133,0.021793874,0.0545740873,0.0177385323,-0.3444501162,0.1442458928,-0.487413466,-0.0421785899,-0.0992305279,0.2021943778,0.0164213609,0.0035270979,0.0593575947,0.0558211654,0.0060588713,-0.32654953,-0.1184618026,0.3234590888,0.0528179258,0.0771333128,-0.1485342085,0.2100295573,0.1040451527,-0.0613660179,0.1945244819,0.0633509085,0.3194554448,-0.1345395893,-0.199500218,0.3969226778,0.0695660785,0.0165768638,0.1155702099,-0.2274149358,0.1046558321,-0.0991497263,-0.0072395988,-0.1597729623,0.2439163327,0.0610595495,0.1981470138,-0.1272440404,-0.3091569841,-0.2142795324,0.3216104507,0.254540652,0.0373857655,0.0955028683,-0.1733588725,0.0505982302,0.0459097587,0.2077760249,0.4979125857,0.0192748941,-0.2003566772,0.3725386858,0.0216428265,-0.2439142764,0.1191153452,-0.0843787342,0.3388143182,0.4956532717,-0.2445803583,-0.0676739663,-0.214152649,-0.0274478383,0.303414017,0.4202139974,-0.5494006276,0.2633036077,-0.3707825541,-0.1403633952,-0.4049875438,0.1483595371,-0.033233989,0.133253485,-0.3153639436,0.1717948765,0.248615548,0.0067794635,-0.1144893393,-0.0167532135,0.0440500714,-0.3137128651,0.09483096,-0.1302260309,-0.2783370316,-0.0179274641,0.5486425757,0.1192936152,0.0776710883,-0.1489702761,-0.3432919681,0.0173715539,-0.0160412099,0.1359623522,0.038618587,0.3947208822,-0.000171923,0.3190338016,-0.2006282955,-0.0312068965,0.1276067644,0.1451542974,-0.1676795632,-0.1626083255,0.004552512,-0.2500949204,-0.1370775998,-0.6858372688,-0.5413176417,-0.1693502814,0.0753475353,0.2003324926,0.194726333,0.1813874841,0.1762078702,0.2450316846,0.0561480038,0.1711566001,-0.0381061994,-0.4465243518,0.226453945,-0.2394654751,-0.2247730196,0.3580351472,-0.0442435667,0.021543989,0.1005166247,-0.5381856561,-0.1136736795,0.2504937351,-0.0290173553,-0.0123384399,-0.0333243571,0.2848323584,-0.0811643004,0.1176243722,-0.0744209588,-0.0750229433,0.3046918213,0.1528561562,0.2563612759,0.3320654631,0.531327486,-0.0368128642,0.7044878006,0.236127004,-0.2257723808,0.3754787147,-0.30376333,0.1068046167,-0.1427258849,-0.1135770157,0.2218452841,0.0835338235,-0.2613663375,0.179017365,0.047193829,0.0657031909,-0.2824084461,0.3268254995,-0.0857608318,-0.1656234562,0.2103683352,-0.1576741338,0.4189042449,0.0357233547,0.1311354786,0.1456913054,-0.2054309994,-0.0827773064,0.56676507,0.1291620135,0.2777432203,-0.557723701,-0.3043447435,-0.3477812111,0.1395135224,0.1647335589,0.487090826,-0.1520963311,-0.0405335687,0.3218385875,0.0231596995,0.4743595421,-0.2974193096,-0.0139402058,0.3608810306,-0.0608358569,0.0712911412,-0.1211801544,0.1311718971,0.5619038343,-0.0362299606,0.4016287327,-0.1029695719,-0.2938934267,0.0835931525,0.2028924078,-0.1985018253,-0.2078430355,-0.1692018211,-0.1432550997,-0.299744904,-0.2505711019,0.0571173355,0.1843994558,-0.0075563649,0.2715389729,-0.0176868737,0.1675961465,0.2338578403,0.0625933111,0.2957535386,0.1207596287,0.1210656241,0.266847223,0.3667298257,0.4993268549,0.5277525783,-0.4694200456,-0.1614117175,0.1477746516,-0.1777692586,0.3248339593,0.2907809913,-0.2040279508,-0.0262716487,0.2625283599,-0.2953740954,-0.0472016893,0.2233002335,0.2479669005,-0.0097966846,-0.0658067763,-0.5862109661,0.2136285752,0.052174639,0.1908984035,0.1354257166,-0.6723163128,-0.44629246,0.2254086584,-0.012819292,0.7663846612,-0.2099230289,0.4583087265,0.1677959114,0.2719201148,-0.0255540535,-0.4253872037,0.274053216,-0.4830499887,0.0642511547,-0.1064785421,-0.1500697732,0.2763530314,0.1562802792,0.0164983142,0.1594835073,0.0549421683,0.1809711456,-0.0730056837,0.3219032288,0.2434393018,-0.1929605007,-0.2359692454,0.0102940034,-0.1801214814,0.210058108,-0.1129737347,-0.1887555271,-0.1296449006,0.0862196535,0.0653886721,-0.09072043,-0.6141549349,0.1236642599,-0.1575998515,-0.3965589702,0.2871089578,0.3647828996,0.1706037819,-0.0710478276,-0.295419544,0.1321610063,0.112256825,-0.1842178553,-0.03295191,-0.0516680852,0.2976560891,-0.0925220028,0.0413704999,0.2050380111,-0.3047454953,-0.3144062459,-0.0661679506,-0.2993863523,0.5104281306,-0.1897246838,-0.4631552398,0.0700422227,0.0860522017,0.0621171445,0.0842691138,0.1648372412,-0.238874197,0.1868060082,-0.0325410627,-0.4317283332,0.0277171861,0.5185140371,-0.0289326627,0.0790055841,0.565784812,0.3568657935,-0.0299055073,-0.2044344395,0.3942812085,0.0715577677,-0.0699861944,0.1138019711,0.4307899475,0.241881907,0.0622125529,0.2950501144,0.2360558957,-0.1236714348,0.2050521076,-0.4611515403,-0.5064795613,-0.101691626,-0.0835034996,0.2456608713,0.171657607,0.1761143655,0.0406590216,-0.2078078687,-0.2155595422,0.2344577909,-0.0199173559,0.1433347613,0.2096408606,0.081258744,0.3789624274,0.0672319233,0.0618059337,-0.0272171125,-0.3651145995,-0.1301807314,-0.0792092979,0.18247585,0.0090726735,-0.3380416036,-0.0017499232,-0.5721231103,-0.2706252337,-0.0220801271,0.0168050844,0.0830389261,-0.0859040692,0.0943692848,-0.1830237806,-0.0719616339,-0.1091131717,-0.0538006276,-0.1991327852,0.0114911059,0.0659253523,0.1101860553,-0.0955982879,0.0075461208,-0.0074562207,-0.09169361,-0.2438067943,0.1240446195,0.232803762,-0.4102424085,0.189899981,-0.2137137949,0.2356956154,0.4120396972,-0.4174543321,0.2346376926,0.5493932366,0.0715426058,-0.1191133782,0.1600171775,0.3307400644,-0.0316046625,-0.1106172353,0.1972023696,0.3799904883,-0.0352505222,-0.1929016262,0.0609401949,0.3784879148,-0.1502983421,0.0640601888,0.297532171,-0.1603610069,0.1333838701,0.2091663182,0.3601654768,0.1545427144,0.524029851,-0.0822410136,-0.0931553021,0.0597373247,0.0534929745,-0.3930665851,-0.6014754176,0.1731493026,0.2167733014,-0.1283470839,0.1240105629,-0.1549286842,0.4967992008,-0.0010901182,-0.2926058471,0.0142848277,0.0785399377,-0.2275688648,-0.1339007467,-0.4213933349,-0.1551742405,0.0502549112,0.0063855299,-0.1967864782,-0.1983109117,-0.1814775318,0.2520407736,-0.1487800628,-0.1792172939,0.092628859,0.0937754661,-0.1402390301,-0.2473496497,0.303843528,-0.1089146882,-0.0788340047,0.1995709985,-0.1509103328,0.2978542447,0.3143269718,-0.1724802405,-0.21891509,0.0805384442,0.0941954106,0.2014834434,0.0283289105,0.1573800296,0.1723032594,0.3426735699,-0.0058866218,0.0086851269,0.316616565,0.0011797708,-0.0813740045,-0.0921972319,-0.058974985,0.2395663857,-0.0096713621,-0.0538717359,0.1466577351,-0.3901590109,0.0225809664,0.7142034173,-0.0171017461,0.3614580333,-0.0683395937,0.0325171649,-0.1164210215,0.3474167287,0.1298307776,-0.0834638029,-0.406519711,-0.0060042422,-0.6180542111,-0.0460275933,-0.1617492586,-0.1863463521,-0.1596548855,0.0043891231,0.1381011903,-0.0324810892,-0.1281869113,-0.0901290774,0.1696688384,0.3214869201,-0.0856050849,-0.2258813679,-0.1679627597,0.3584050536,-0.0031588958,-0.3093914986,-0.0893948153,-0.1936803013,-0.0117186802,-0.2749819756,0.0232634321,0.1074019521,0.2027430683,0.2969765067,-0.2101576924,0.4309168756,0.0065391446,0.1876882464,-0.1705281138,0.0723379701,-0.0811829865,0.1116091833,0.0496192425,0.4050960243,-0.529505074,-0.11736615,-0.5370402336,-0.0946171582,0.0156083265,-0.2459225059,-0.1738643795,0.0654882789,-0.1364307702,0.2675381899,0.2362120152,0.3413186967,-0.0979718938,-0.0337588191,-0.3247717917,-0.0526789241,0.5630751252,-0.4386031926,0.0079312734,0.1063509285,0.0169814751,0.1237151548,0.1071063057,-0.3041403294,0.0648378134,0.4830067158,-0.1889516562,-0.2679226398,0.3511603475,-0.0312894993,0.0399591774,0.023728447,0.0085910037,-0.0142722279,-0.3185524046,-0.3784615397,-0.4073459506]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2243","title":"Map is slow and processes batches one after another","comments":"Hi @villmow, thanks for reporting.\r\n\r\nCould you please try with the Datasets version 1.6? We released it yesterday and it fixes some issues about the processing speed. You can see the fix implemented by @lhoestq here: #2122.\r\n\r\nOnce you update Datasets, please confirm if the problem persists.","body":"## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!","comment_length":47,"text":"Map is slow and processes batches one after another\n## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!\nHi @villmow, thanks for reporting.\r\n\r\nCould you please try with the Datasets version 1.6? We released it yesterday and it fixes some issues about the processing speed. You can see the fix implemented by @lhoestq here: #2122.\r\n\r\nOnce you update Datasets, please confirm if the problem persists.","embeddings":[-0.3536394835,-0.4107639492,-0.0976410061,0.2570492923,-0.0076191481,0.1119764224,0.1884649545,0.3910495639,0.4355854094,-0.0132389804,0.2567101121,0.3540889323,0.0090388684,-0.1517525315,-0.1746495366,0.0698189065,0.1166413799,-0.1431869268,-0.0087070642,-0.1623785645,-0.3209817111,-0.087763235,-0.2003731877,0.0438098162,-0.1860580742,-0.2596023977,0.0139482338,0.0891774371,-0.0040700217,-0.3334619105,-0.0190602932,-0.0149467504,-0.1927878559,0.9497959614,-0.0001168566,-0.0805333108,0.2479249239,0.2797504961,0.1089458242,0.0850166529,-0.2313876301,-0.3576023281,0.0878629759,-0.2359792441,0.2497991621,-0.023537647,-0.002994285,-0.5800064802,0.2692072093,-0.0939939618,0.140329212,0.0948103815,-0.386369735,-0.0908237621,0.0266346149,0.005713955,0.0010937327,0.1683355421,0.5363388658,-0.2120462209,-0.1499995291,0.1861923039,-0.4690407515,0.300696224,0.0931109935,0.0692812651,0.5572910309,-0.561075449,0.2869141996,0.1746804267,0.0806175396,0.0566017814,-0.044365216,-0.401881963,-0.4251192212,0.0211951807,0.063075155,0.1158909574,-0.1820195317,-0.2456142008,-0.4943926632,0.016884638,0.2267080694,0.0516579226,0.0052897572,0.0966013446,0.1384706944,0.2925825417,0.3433318436,0.1091784909,0.0677273944,-0.305072695,0.0327481292,0.3052342236,-0.452498883,0.0874829516,0.4808760583,0.0541621819,0.0520278104,-0.3966451585,-0.142978698,0.027827343,0.0229985379,0.1050668061,0.3621997237,-0.006308923,-0.0987884104,0.2177391052,0.4536742568,-0.2545010448,-0.3706454635,0.2139307857,0.0409928337,-0.5764334202,0.1754998714,0.0035549165,0.0092065679,-0.0291023236,-0.0612536371,-0.0754595697,-0.1532066166,-0.1927638799,0.1660397798,0.3424512148,-0.1398621947,0.2019235492,-0.1730198562,-0.1404955387,-0.2200604081,0.2372357398,-0.1123882532,-0.3543464839,-0.3067362607,0.1507669836,0.1014041901,-0.2737501264,0.0847366825,0.3837851286,-0.3222764432,-0.1553747356,0.1739759445,-0.3959164023,0.0064279926,0.02302121,0.0425842926,0.6243890524,0.0900540799,0.2955989838,-0.2339507639,0.3710044324,-0.597797811,-0.1593910456,0.2099527568,0.0487274565,-0.0191689916,0.2650373578,-0.2873978615,0.0804849267,0.2995231748,-0.175204441,-0.2621099949,-0.1640048623,-0.3243804276,-0.309782505,-0.1098135337,0.3768040538,-0.0898602307,0.0984198675,-0.250891,0.1711792499,0.472150296,0.5991899967,-0.2346330732,0.4373857975,-0.2421700507,-0.0288157668,0.018366918,0.0336230397,-0.2913622558,0.7230955958,-0.3079105616,0.0267542675,-0.0977236032,0.2283003926,0.3933739364,-0.2332687676,0.269977957,0.3129475713,-0.0474728532,0.4035684764,-0.3195688128,-0.1058711931,0.1694390625,-0.0341540128,0.0220105406,0.0716377869,0.1143119931,-0.3669935763,0.3592448533,-0.0605864935,0.0328281857,0.2090796232,0.0447492786,-0.0553572476,-0.0184040796,0.0925919116,-0.0947196558,0.2243772,-0.0192848165,-0.3248773515,0.0324211195,0.0841892362,-0.0510729142,0.0634693652,-0.148111701,0.0944968536,-0.0638687611,0.1023333073,0.0228009783,-0.0569532141,0.2669195533,0.1030330062,0.1032134891,-0.2411195487,-0.0628770217,0.0730087459,0.0916984603,-0.0767954811,-0.0768885463,-0.2617317736,0.3339428306,-0.0912025124,-0.1781997085,0.1648807675,0.0898478627,0.3029913902,-0.0419128984,0.1013589427,0.1004371494,0.262160331,0.1643080562,0.0480548963,0.2535101473,-0.3618002236,-0.0108606694,0.1457659006,-0.0321851522,0.4392972291,-0.1151999235,-0.0318247154,-0.0059306752,-0.0788996592,0.2529892325,0.1770585775,0.2835728228,-0.0793612152,0.4447599351,0.2781111002,-0.0302263368,0.0132176057,0.4852698743,-0.0271275472,-0.1436029375,0.1334326118,-0.1315696836,-0.0328650586,-0.0420117863,0.1701156944,0.4768700898,0.0805022568,0.0264230464,-0.182655856,0.1206792891,-0.0661399812,0.1008415818,0.1482352167,0.4915498495,-0.045618128,0.3280079663,-0.0262983516,0.0204266366,-0.1683085412,0.2249910086,0.2838003635,-0.1066473126,0.1500249803,-0.0864218175,0.1899553239,-0.3173114359,-0.0530077182,0.0569251329,-0.2011275887,-0.1721757054,0.2013168633,0.0083301719,0.1018987522,0.1917135566,-0.2170332074,-0.0424208865,-0.105635047,-0.0220116004,-0.3020041585,-0.0366279744,-0.0245760642,0.3743671179,-0.0097525688,0.2599737346,0.1229065731,-0.4351427257,-0.0318548083,0.0146431364,-0.134053126,-0.1755057871,0.0385618508,-0.2328603119,0.2902824283,-0.071351856,-0.1962250173,0.2703505158,-0.6561154127,0.0064780097,-0.0568646416,-0.1580043286,-0.3678793013,-0.0483247302,-0.1501354724,-0.1791179329,-0.1662816256,0.3194873929,-0.1182533354,0.3504863977,-0.0193890464,0.1067479774,0.0565150231,0.1134434715,-0.1843942404,-0.1836296767,-0.2651648223,0.1830536276,0.0533219241,-0.0865589529,-0.2310346961,-0.0336551405,-0.103563711,0.2471540272,-0.391320765,0.1752493232,-0.5987238884,0.2207243592,0.0769206658,0.2120436877,0.3499593139,-0.1178209186,-0.0693313479,-0.1490348577,-0.2949303687,-0.0052095717,0.2313658893,-0.0322886072,0.1110155582,0.504396975,-0.0653442889,0.6194151044,0.4053158462,-0.060055986,0.0572079122,0.0808383003,-0.1393208802,-0.2665077448,-0.3724047542,0.0360680856,-0.3106940389,-0.3482807875,0.2595496476,-0.0779712871,-0.4955922663,0.037873622,0.0569978207,-0.2354692668,-0.0812703744,0.1077042893,-0.0398104563,0.2836271524,0.0451275744,0.0951742455,-0.4651483893,-0.0807362273,0.0391063616,-0.3701034188,0.3640816808,-0.2868103981,-0.1681772023,0.1631643027,-0.5535615087,0.2083932161,0.0754104033,0.1869859546,0.0137673067,0.1490963846,0.3038130105,0.1498131603,0.747386992,-0.0480228178,0.1053793058,-0.00722,-0.187079072,-0.5962153673,-0.025959529,-0.0231607836,0.4650937915,0.6051085591,0.3453222215,-0.1446372569,-0.0277191028,-0.1587144136,0.24467884,0.0206058659,-0.2795672119,-0.0973887816,-0.1563323885,-0.1371879131,0.1207817048,-0.1132946834,-0.0592597984,0.0447899997,-0.2501924932,0.1766347438,0.225139007,0.1085796878,0.1332273334,0.2369526178,0.0792281181,0.5708370209,0.2187570482,0.1466367841,-0.0046405476,0.3554345071,-0.1292726696,0.2139156759,0.1814136654,-0.0028032565,0.2090529799,0.4299541116,-0.0406555235,0.0522600859,-0.1365888566,0.2162789851,-0.2567058206,-0.1218845323,0.5087291598,0.2650709748,-0.4859510362,-0.5629432797,0.2916388512,0.2924588323,-0.380158484,0.639980197,-0.8032797575,-0.0330473483,0.4559307992,-0.0678517297,0.694691956,-0.2801558077,-0.1290728301,-0.1350941956,0.1064754426,-0.1222082749,0.26741153,0.2052533776,-0.2929269969,-0.1009617895,0.2086085081,-0.2684565187,0.0551342331,0.1981160939,-0.0062499782,0.060606882,-0.0408778042,0.0415203422,-0.2543274164,0.0650403127,0.4701123834,-0.3301798701,-0.1192822531,0.0843546093,0.1271931529,0.2608578503,0.0197539236,-0.0918400362,-0.2189222723,-0.1566542834,-0.1526544392,0.0026953819,-0.3488563597,0.347224772,-0.0512613878,0.1078181714,0.1518945098,0.1402824819,-0.1792607009,0.1257934868,-0.0836189613,0.16624403,0.2662297189,0.3825931549,0.2885946929,-0.2326244861,-0.1720510125,0.0450579599,-0.0883711502,-0.1523728222,-0.1140356138,-0.1603781134,-0.4679456651,0.5018332601,0.0943726227,0.1071598604,-0.4344460964,0.2077492774,-0.0158790983,-0.1838546246,0.0703701153,0.155168429,0.1048135906,0.7553778887,-0.4236602485,-0.3901266158,-0.00198007,0.2198662609,0.2952713668,0.1123985201,0.3343268931,0.1712135375,-0.222226277,-0.0866997987,0.1048377678,-0.0797601938,-0.2804398239,0.2056123912,-0.2834074199,0.0297720879,0.2937417924,-0.0060889614,-0.187464416,0.0622981936,-0.1230188981,-0.1706308126,-0.2203829885,-0.1845600009,-0.0944474414,-0.0320843458,-0.1417267621,0.1171070263,0.0346491188,0.5534904599,-0.2105933428,0.086045377,0.1453553587,0.2014833391,0.0312301125,-0.0578449331,0.1517126709,-0.1721321046,0.0156973191,0.1662783772,-0.0927256271,-0.1589340717,0.0091945874,0.176161781,-0.2031961679,-0.3424268961,0.2239152938,-0.3177878559,0.0368953794,-0.4057893753,0.208641082,0.3461664319,-0.0236161593,-0.4323787391,0.3018778563,-0.0884994939,0.1120628193,0.2977092266,-0.4248302579,-0.0268875025,0.0234790854,0.0996798128,0.3500134945,-0.1138402671,0.008737809,-0.1823055744,0.2913302779,0.2199276835,0.3804685771,-0.2574824989,0.0662407205,0.4824460447,0.2524221241,0.2404263616,-0.0137446644,-0.0694172531,0.1133243889,0.118562907,-0.086099118,-0.1395348758,0.6526375413,0.083264105,0.0199283287,0.0320323296,0.1843595505,0.0023583334,-0.4783991873,0.0849459246,0.1564912647,-0.2350760102,0.1105499044,0.3106342852,0.1898123175,-0.1293052733,0.2571067214,-0.0139380181,0.0622463897,0.3669970334,0.221698612,0.2367869318,0.2245042324,0.1035286933,0.0251101535,-0.3650442362,0.0451269373,0.2423908263,-0.3808351457,0.1161130518,0.1658350378,0.2006110102,-0.3007858098,-0.2111154795,-0.2413771302,0.2320920527,-0.2003004402,-0.175361082,0.1362297535,-0.0020117345,0.1566546112,0.1407229304,-0.1213386282,0.0112219928,0.5113050342,0.1351642311,-0.0666334778,-0.140653491,-0.4248694479,-0.0930648074,0.2990778089,-0.3026794195,0.0119955381,-0.321685195,-0.1741578728,-0.270011574,0.2303567082,0.4243108332,0.578343153,-0.0638668537,0.0594046228,0.0387760364,0.058683075,-0.1515976787,0.3526741564,-0.153096661,0.1793697923,0.0488222577,-0.018849846,-0.1761784554,-0.0927365869,-0.0101099117,0.2458303273,0.162916109,0.1476048529,-0.2401455343,-0.2276580632,-0.0430455841,0.0435930006,-0.1089703739,-0.1449199468,0.4527194798,-0.4046346247,0.1082326025,-0.1527310163,0.0367451198,-0.3230798841,0.4421977103,0.1769668162,0.2170802653,-0.4274473488,-0.1645277888,-0.4656027853,0.0945194215,-0.6270409822,0.3479112089,0.0184438173,0.1136777252,-0.044115264,0.01815979,0.193802923,-0.0754053518,-0.167051211,0.2149129212,-0.2346523851,0.0765882283,-0.353376776,-0.1058027223,0.1517272443,-0.4540385306,0.3884385228,-0.3258297443,0.047073774,-0.1807992607,0.1086731255,0.1300684065,0.1781799346,0.5276604891,0.4160920084,0.2875406742,-0.1532476395,-0.0357660241,-0.204975903,-0.2076935917,-0.0986897573,0.153901279,0.1099425033,0.1592293531,-0.2325558364,-0.0640316233,-0.2706333697,0.0089948308,-0.116637215,0.1578444541,0.0779863968,-0.2959678471,-0.1835751981,0.2445849031,0.0585401617,0.2674297988,0.0500095747,0.0709704831,-0.1508758962,-0.1668469906,0.3609440625,-0.5754396319,-0.3871363401,-0.3030738533,0.2370181829,-0.0861409083,0.1355550289,-0.0060564461,-0.1824787557,0.2757093608,0.0892306641,-0.2111136466,0.0828522965,-0.2617931962,-0.0277418401,0.0513776094,0.0487921871,0.0564324781,-0.5505631566,0.2456963956,-0.1178314462]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2243","title":"Map is slow and processes batches one after another","comments":"Hi @albertvillanova, thanks for the reply. I just tried the new version and the problem still persists. \r\n\r\nDo I need to rebuild the saved dataset (which I load from disk) with the 1.6.0 version of datasets? My script loads this dataset and creates new datasets from it. I tried it without rebuilding.\r\n\r\nSee this short video of what happens. It does not create all processes at the same time:\r\n\r\nhttps:\/\/user-images.githubusercontent.com\/2743060\/115720139-0da3a500-a37d-11eb-833a-9bbacc70868d.mp4\r\n\r\n","body":"## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!","comment_length":70,"text":"Map is slow and processes batches one after another\n## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!\nHi @albertvillanova, thanks for the reply. I just tried the new version and the problem still persists. \r\n\r\nDo I need to rebuild the saved dataset (which I load from disk) with the 1.6.0 version of datasets? My script loads this dataset and creates new datasets from it. I tried it without rebuilding.\r\n\r\nSee this short video of what happens. It does not create all processes at the same time:\r\n\r\nhttps:\/\/user-images.githubusercontent.com\/2743060\/115720139-0da3a500-a37d-11eb-833a-9bbacc70868d.mp4\r\n\r\n","embeddings":[-0.3630149364,-0.3355692923,-0.0869320706,0.2914981842,0.0048044329,0.0850246921,0.207652986,0.3980056643,0.4348949194,-0.0089639695,0.2125943303,0.2832153738,-0.0250464026,-0.2042584866,-0.197465688,0.1578701138,0.1223637983,-0.0891835243,-0.0129013015,-0.1942545027,-0.3704572618,-0.0611239597,-0.1589694023,-0.0333990157,-0.203557536,-0.2246247679,0.0031420246,0.16731745,-0.0278627519,-0.3802036345,-0.0222847667,0.0346784741,-0.217319414,0.9670494795,-0.0001179558,-0.0346966349,0.2712252736,0.2329895794,0.0776725635,0.0469267331,-0.1714724004,-0.3223571479,0.1062251329,-0.192093581,0.2716529667,-0.0017312861,0.0206621755,-0.6094627976,0.2712439895,-0.0779175013,0.1323134154,0.1193444803,-0.4176820219,-0.1101857424,0.0098892516,0.0443693511,-0.0237345956,0.2211578041,0.469622016,-0.1314185411,-0.161554426,0.1670174003,-0.4216374755,0.2660012245,0.0706350505,0.0413721614,0.4938414395,-0.5665130019,0.3148688078,0.1981596947,0.0256941523,0.0241917334,-0.0996279344,-0.3593217731,-0.3677815795,0.0037667628,0.0352485627,0.1015741229,-0.1780721992,-0.2604908943,-0.510027349,-0.0142801199,0.2391904593,0.0101411361,0.0763001591,0.0058418438,0.1450753808,0.2875227332,0.352347672,0.1838088632,0.1186252385,-0.304667592,0.0057900851,0.2401407808,-0.4155762494,0.1519878656,0.4824928343,0.0279344805,0.0871175379,-0.4267273247,-0.089911215,0.0575835854,0.1102226451,0.0385423489,0.3332891166,-0.0898994356,-0.0239036288,0.1651181281,0.4020159245,-0.2371514887,-0.3967393339,0.1732116193,0.0284360144,-0.5089076757,0.1731099337,0.0160155334,0.0584516115,-0.0073794699,-0.0503870435,-0.0698949844,-0.1287699044,-0.2418865561,0.1157804355,0.3481166065,-0.0990552083,0.284419179,-0.1221458018,-0.1511015296,-0.3075758219,0.2642687559,-0.1004589126,-0.3427415788,-0.3130770326,0.1872937977,0.1027255654,-0.3253945112,0.1008893326,0.3663766086,-0.3238576651,-0.140540868,0.1640705168,-0.3921980262,0.037114948,0.0362593681,0.1015313864,0.5760555267,0.0617110655,0.3559787571,-0.2052507102,0.3554876149,-0.5027271509,-0.1576422155,0.1760053337,0.0561755635,-0.0610033758,0.2625415623,-0.230644688,0.077102907,0.3370042145,-0.2698493302,-0.3164979219,-0.1621470898,-0.2763829827,-0.3266085684,-0.1250043511,0.4855922461,-0.0969214812,0.0926420689,-0.2499944568,0.1078091785,0.4460745156,0.5792755485,-0.2000685632,0.3872777224,-0.223335743,-0.0858245641,0.1105750427,0.109143056,-0.3711589277,0.62790519,-0.2107079178,0.0252087936,-0.0302220359,0.2141833156,0.4351965487,-0.2300636619,0.189813152,0.2136393934,-0.0715662986,0.3919412196,-0.3567687869,-0.1234346107,0.1199619621,-0.068618618,0.0712119192,0.1128558591,0.1466659009,-0.3748902678,0.3419593573,-0.0813911706,0.0750169531,0.2327743322,0.0045928643,-0.0640248284,0.0254024249,0.0408823788,-0.1903897971,0.1950108856,-0.0402653478,-0.2926433086,0.0948730782,0.0526052527,-0.0963645726,0.0147375111,-0.1551770866,0.0935113654,-0.0817369744,0.2289909571,0.0444578193,-0.0817886963,0.1954099834,0.0643505752,0.1163239852,-0.2580953538,-0.1073692366,0.1122938022,0.0365022756,-0.1578089893,-0.0445042215,-0.2471504062,0.2896203101,-0.089943707,-0.1257686168,0.1543934196,0.0985815451,0.3515720069,-0.0858734176,0.0658476278,0.0880118981,0.3965590596,0.1759848893,0.0508555397,0.2638603151,-0.3270398974,0.0296379291,0.1832247823,0.0414749235,0.4627265632,-0.0223166067,-0.0116066085,-0.0076325648,-0.0596297979,0.2356954217,0.1006035656,0.2856350243,0.0174683537,0.465131402,0.2683877945,-0.0083425092,0.0978523716,0.5448834896,-0.0384720117,-0.145064041,0.1689181477,-0.0935998186,-0.0727545768,-0.1138985455,0.1575489193,0.4817344844,0.0790671259,0.0289869066,-0.1687620282,0.1559609026,-0.058006186,0.0718887746,0.1646239311,0.6082304716,0.0435824767,0.32114169,-0.0311715584,-0.0458855256,-0.1859902292,0.2890871763,0.2805094421,-0.129435271,0.1533901244,-0.0461231247,0.2741508484,-0.2902233005,-0.0880478173,0.0753923729,-0.2380780727,-0.1974068284,0.2284224778,-0.0029088338,0.1274891794,0.182867825,-0.1537574083,-0.0923857987,-0.0000270022,-0.0489275306,-0.3085848093,-0.0368410014,-0.0362476669,0.3555816114,-0.0531523712,0.2529759705,0.1141378954,-0.4489083588,-0.0389914289,-0.0068281908,-0.0905419141,-0.1739006042,-0.0252901316,-0.1881171018,0.2716783583,-0.029089747,-0.2099588066,0.265679419,-0.6992890835,-0.0694216564,-0.0357697308,-0.1853723973,-0.3714575171,-0.0621029101,-0.2339767069,-0.1610837579,-0.2183563262,0.2947669029,-0.0674247816,0.3748463988,-0.1149966195,0.09592437,0.0573428273,0.1013476327,-0.1709908843,-0.1748352498,-0.2321131527,0.1582455933,-0.0439944305,-0.1131773889,-0.247648403,-0.1134029776,-0.0751528069,0.3227633238,-0.3404302895,0.1396199167,-0.5815432072,0.172369644,0.0529167652,0.2146683037,0.4270578325,-0.0608473867,-0.0778912902,-0.1548994482,-0.3055674434,-0.0078540025,0.2052788138,-0.0036146445,0.0674420595,0.542740345,-0.0926426798,0.6695802212,0.4467199445,-0.0754954964,0.0411668457,0.0264881235,-0.0547487661,-0.3249762058,-0.3788352907,0.0915603787,-0.2707596123,-0.2935105264,0.2437550277,-0.1172546521,-0.4212988913,0.0439122878,0.0312338304,-0.2758792043,-0.1213838831,0.1162857637,-0.097058177,0.3035881519,0.0317237936,0.1417817026,-0.4291802645,-0.0093306648,0.0107623069,-0.3043084145,0.4023874402,-0.2962892354,-0.1551970989,0.3099076152,-0.6096820831,0.2347208858,0.0806418806,0.1860010028,0.0709650367,0.107090719,0.2777571082,0.1059552133,0.7209101915,-0.0790826008,0.0211906172,0.0412268005,-0.2007650584,-0.6698828936,-0.0049527586,-0.0127181616,0.4702877104,0.6730164289,0.3351783156,-0.1638124883,-0.0326173492,-0.1395627856,0.2700152993,-0.014502828,-0.249907136,-0.1622543186,-0.2093793899,-0.196888566,0.0287488811,-0.1429502219,-0.0143247778,0.0607077517,-0.2408839166,0.0920096859,0.1469636261,0.1134049967,0.1214263216,0.2578723431,0.0983288661,0.5815179348,0.1797896028,0.0700602233,0.0515779778,0.4057936966,-0.1279024929,0.1545745879,0.2507947981,-0.0008213657,0.2090462893,0.4520723224,0.0144893602,-0.002907292,-0.0988428369,0.1922035515,-0.194029808,-0.1173432097,0.4256081283,0.171086967,-0.4726230502,-0.5922845006,0.2535350323,0.2536030412,-0.3607655168,0.6391512752,-0.7487567663,-0.0183293447,0.4669686258,0.002820835,0.7358861566,-0.2567670941,-0.0418370292,-0.1000896618,0.0824595913,-0.1197908372,0.4190860093,0.2596223056,-0.3722594678,-0.0804279521,0.2213825434,-0.2510465682,0.0735090971,0.1787227839,0.0123321461,0.0808367878,-0.0396678112,-0.0299986228,-0.286565423,0.1244358644,0.3993237019,-0.3570459187,-0.0643467158,0.0637205467,0.123383306,0.2416669726,0.0296138953,-0.0855247825,-0.1752415746,-0.134915486,-0.1914401054,-0.0151506066,-0.3884584308,0.3374822736,-0.0356906503,-0.001097044,0.1039383635,0.1001684964,-0.18673639,0.1002775356,-0.049924735,0.1501244456,0.2366527766,0.3369306624,0.2823454738,-0.221893236,-0.1344579458,0.030493753,-0.0960116386,-0.2380340099,-0.1103519127,-0.193684876,-0.4630658627,0.4645528495,0.0995132923,0.0769588128,-0.4267886281,0.2048383653,-0.1343347281,-0.1900669038,0.0477786697,0.140486896,0.1448500156,0.7204334736,-0.449521035,-0.367592901,-0.0362619683,0.2923368812,0.3315473497,0.046317894,0.3831964731,0.123812221,-0.1957349032,-0.1185149848,0.0998443365,-0.0271352939,-0.2387663871,0.2741428614,-0.2905699015,-0.0018503349,0.3436768949,0.015529817,-0.2062702179,0.122430712,-0.1491795331,-0.151700139,-0.2442890406,-0.1842517108,-0.0479684509,0.0064313617,-0.1502271295,0.0969573185,0.0614564531,0.4971451163,-0.208967641,0.0648500696,0.1968674958,0.1899017096,0.0246660355,-0.1359559894,0.1353496313,-0.1885118484,0.0163046531,0.1631894261,-0.090997152,-0.1194695011,0.0107555231,0.1842518896,-0.2022351325,-0.2798935175,0.18416363,-0.2515424788,-0.0125244539,-0.4043359458,0.2153512388,0.3910016418,-0.0911776572,-0.4116526246,0.2231805474,-0.121409595,0.1379814893,0.2327386439,-0.4405063391,0.0219244119,0.0274142586,0.1154976711,0.403211534,-0.1132010221,0.0035117108,-0.1495000869,0.2916156352,0.2376497835,0.4089950323,-0.2148096263,0.0811433122,0.4790180624,0.2529851496,0.3715296984,-0.0216074679,-0.0751505569,0.0790517107,0.1053319871,-0.0986377001,-0.118040435,0.5559914708,0.017260991,0.0223236587,0.0815002695,0.1109004691,-0.0734092742,-0.4644205272,0.1309803575,0.1541843414,-0.3088000715,0.1640720516,0.2352608889,0.2134702653,-0.1326985955,0.3234117925,-0.0104671465,0.0613101237,0.3002292514,0.1901672632,0.2820367813,0.1583477259,0.1097849533,0.0129354373,-0.3917321265,0.0778551996,0.1975888014,-0.37114048,0.1210035905,0.1877914667,0.1419266313,-0.3178578019,-0.2040488571,-0.2610981166,0.2047228068,-0.2230825126,-0.1804422289,0.0873599872,-0.0069226404,0.1763298959,0.1609487981,-0.110495545,0.0136724543,0.5182111859,0.141936034,-0.0955726504,-0.1300569028,-0.3889630139,-0.1098817736,0.3121516407,-0.3008531928,0.0523625016,-0.2596473992,-0.1228784621,-0.2750819325,0.3463408947,0.4599645436,0.6527571678,-0.0124106845,0.0608471744,0.0710591674,0.0507835113,-0.1679679155,0.3546045721,-0.0640357584,0.2595690489,0.0531159043,-0.0321906172,-0.1525261849,-0.0521172099,-0.0238505658,0.2068840861,0.1658694744,0.2682729959,-0.2395357639,-0.2659416497,-0.017229246,0.1227830499,-0.1800326407,-0.2328299284,0.4231389165,-0.426163584,0.1212930307,-0.2007842064,0.0358420834,-0.2684952915,0.4545712173,0.1397160292,0.1488043517,-0.4117129147,-0.0982580036,-0.4501784742,0.0493731983,-0.5862381458,0.3814853728,-0.034618143,0.1783493012,-0.0473565422,-0.0081328601,0.1628370285,-0.0443746373,-0.2253890485,0.1940117031,-0.2575012445,0.0605098568,-0.301499337,-0.1696663648,0.1718636155,-0.4466189146,0.3505331278,-0.3587097824,0.0564386584,-0.1126198247,0.088916555,0.137510404,0.142740801,0.5240485668,0.4821200669,0.2463360876,-0.2212333083,-0.1244490296,-0.2973881662,-0.2201163024,-0.1331464946,0.1974972039,0.1863687485,0.1875919402,-0.2295739204,-0.0875499323,-0.2807492912,-0.0312832966,-0.1266412288,0.2273052782,0.1401259601,-0.2672667801,-0.1147851124,0.2475760132,0.0845567733,0.3016546071,0.0440084077,0.0786036327,-0.2230159938,-0.1386023611,0.3557642102,-0.4948846996,-0.3726125658,-0.3232242763,0.3176662326,-0.0697514489,0.1108843461,-0.0645007789,-0.1400405914,0.2673558295,0.0445397571,-0.2483783662,0.1095242128,-0.2917210162,-0.0558233708,0.0421828702,0.003717208,0.0795395821,-0.6037822962,0.2104653269,-0.1166232824]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2243","title":"Map is slow and processes batches one after another","comments":"There can be a bit of delay between the creations of the processes but this delay should be the same for both your `map` calls. We should look into this.\r\nAlso if you hav some code that reproduces this issue on google colab that'd be really useful !\r\n\r\nRegarding the speed differences:\r\nThis looks like a similar issue as https:\/\/github.com\/huggingface\/datasets\/issues\/1992 who is experiencing the same speed differences between processes.\r\nThis is a known bug that we are investigating. As of now I've never managed to reproduce it on my machine so it's pretty hard for me to find where this issue comes from.\r\n","body":"## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!","comment_length":103,"text":"Map is slow and processes batches one after another\n## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!\nThere can be a bit of delay between the creations of the processes but this delay should be the same for both your `map` calls. We should look into this.\r\nAlso if you hav some code that reproduces this issue on google colab that'd be really useful !\r\n\r\nRegarding the speed differences:\r\nThis looks like a similar issue as https:\/\/github.com\/huggingface\/datasets\/issues\/1992 who is experiencing the same speed differences between processes.\r\nThis is a known bug that we are investigating. As of now I've never managed to reproduce it on my machine so it's pretty hard for me to find where this issue comes from.\r\n","embeddings":[-0.3194858134,-0.4774059355,-0.0747577175,0.3302206993,-0.0672015771,0.0412484743,0.2097329944,0.3263496161,0.4367236495,0.0293657854,0.241692856,0.3490332365,0.0190935712,-0.0388512351,-0.1365028918,0.0055769235,0.1733696014,-0.1191435233,0.0165218115,-0.1376280785,-0.2373458147,0.0043261829,-0.1967942417,-0.0159254391,-0.1699439734,-0.2258180082,-0.0241215285,0.0833455771,0.0180470161,-0.3501836061,0.0165381711,0.0091593005,-0.1990933716,0.9538552165,-0.0001199121,-0.0984314978,0.2415305674,0.3196500838,0.0868761763,0.0823288038,-0.2088344395,-0.3429727256,0.1231218204,-0.2194910794,0.181776464,0.0489583462,-0.0054669217,-0.478476882,0.2912492752,-0.1400743425,0.1226795465,0.1224082932,-0.4304054081,-0.1170066521,0.0643731058,0.0136844898,-0.0449960642,0.2221290916,0.5488747358,-0.1446338445,-0.1135171801,0.1787206531,-0.4208791554,0.3248113394,0.1550359428,0.1488950402,0.4124362469,-0.5443951488,0.2542956769,0.2439431697,0.0642348453,0.0644448474,-0.0699718073,-0.4524817467,-0.4557455182,-0.050525751,0.1526226699,0.1117550284,-0.2006653398,-0.2737897933,-0.5146392584,0.0925396383,0.2486539781,0.0496206582,-0.0354445167,0.0554922223,0.0845619068,0.298102051,0.3523456752,0.1011010781,0.0175358225,-0.359507978,0.0585528389,0.3016782999,-0.4295680523,0.0651175007,0.4697278142,0.0874668062,0.0563356057,-0.3640021086,-0.1127560064,-0.025833508,-0.0637136623,0.1232847795,0.3699847162,0.0635845363,-0.0990267172,0.1580916196,0.449812144,-0.2064523548,-0.3955384791,0.2026739419,0.099490352,-0.5509958863,0.1231858656,0.0072417259,-0.0061324919,-0.1270748079,-0.0732416138,-0.0996087193,-0.1479617655,-0.2157738209,0.1760329157,0.3118009865,-0.1720112413,0.1390254498,-0.1780219823,-0.1064369604,-0.2601937652,0.2096254528,-0.109958075,-0.3654551804,-0.316978395,0.2297674567,0.091924414,-0.2789958715,0.0632554516,0.4462325573,-0.2974608541,-0.1287810951,0.192067802,-0.4455226064,0.019282937,0.0177791025,0.0631104931,0.5827058554,0.0641357452,0.2910842896,-0.2870440185,0.352948457,-0.5255614519,-0.1047502756,0.247007072,0.0056428602,-0.0970230699,0.2554524839,-0.3724861443,0.1563412249,0.2468484342,-0.2550650239,-0.2419194132,-0.1594786346,-0.3996351361,-0.3362621963,-0.1038314849,0.4296418726,-0.1361075491,0.0145078488,-0.2160125673,0.1814296246,0.449703455,0.6606137156,-0.2419025302,0.405659467,-0.2201427519,-0.0463531241,-0.0111754192,0.0401429497,-0.2736548781,0.7009415627,-0.2806129754,0.0992984623,-0.0818606764,0.2187545002,0.3799466491,-0.245737955,0.2991667688,0.3218560517,-0.0316918343,0.4197546542,-0.3054518104,-0.1577080786,0.1585372537,-0.0863017142,-0.0191195291,0.1247162446,0.1028204411,-0.4522362649,0.3015806377,-0.0758424997,0.0183351822,0.218925029,0.1022760868,-0.0690943077,-0.0285861827,0.0735594779,-0.1101508141,0.2620704472,-0.0402679481,-0.2302188426,0.0483768396,0.0963158086,-0.0515501797,0.0079332693,-0.1040668786,0.093493022,-0.0941341966,0.1083738655,0.0511480682,-0.0304974932,0.2722037733,0.167321071,0.1698075831,-0.240647912,-0.0940886959,0.1516323239,0.1160434783,-0.1340617687,-0.0750783384,-0.2594904304,0.3234907091,-0.1187682226,-0.1830591112,0.1087518632,0.0736615136,0.3307231963,0.0041519548,0.1491952538,0.1749664396,0.283064276,0.2505117059,-0.0231627729,0.2347990423,-0.3815715909,-0.0548163727,0.2052121609,-0.0218360126,0.4643540382,-0.08364705,-0.0821953043,-0.0320651978,-0.0591447763,0.2360588312,0.1716521084,0.2913886011,-0.0511242673,0.4382725656,0.2802740932,-0.0629544333,0.0940982848,0.519520998,0.0014685434,-0.1753924787,0.1385180652,-0.1514688581,-0.0686138868,-0.0401510224,0.1637530774,0.5669997334,0.0724177659,0.0455683433,-0.1423933357,0.1140161902,-0.0465193503,0.0907405913,0.1230156347,0.4285607636,-0.07515122,0.3441388607,0.0090245297,0.0327259861,-0.1939031333,0.2061123252,0.2143487781,-0.0936268643,0.1492774636,-0.0833012834,0.1406968385,-0.3772099614,-0.0402711816,-0.0436655693,-0.1907573044,-0.217360422,0.2685177326,0.0243420787,0.053916961,0.2391970307,-0.1273527294,-0.1300688684,-0.100606665,-0.0242543276,-0.3237657547,-0.0682418197,-0.0410738252,0.3971996307,0.023496164,0.2353716493,0.0678013116,-0.4454363585,0.0030340324,-0.0445876271,-0.0715801045,-0.1857948899,0.1254251003,-0.2303630412,0.3192312121,-0.1090103835,-0.1898678392,0.3258904219,-0.6125127673,-0.0445981584,-0.1552238762,-0.1521812528,-0.3659999967,-0.0362102091,-0.1075744405,-0.1633157134,-0.1247296035,0.3783959448,-0.1027492285,0.3412996531,0.0129914852,0.0697473735,0.0358464234,0.1879738271,-0.0981814191,-0.2270577103,-0.2855250835,0.0979163498,-0.0202624034,-0.1072789356,-0.2148885131,-0.0279016811,-0.1364365667,0.249072507,-0.4367383122,0.0953364149,-0.6324816346,0.2248194814,0.112212576,0.1835850924,0.3638401926,-0.1243840903,-0.043315161,-0.1775204986,-0.4602483511,-0.0299754571,0.289239794,0.019463215,0.1156247705,0.4566768408,-0.1499717683,0.694293201,0.3712833524,-0.0745497346,0.0624874718,0.0717327595,-0.126137495,-0.3553828299,-0.3348673284,0.0974921957,-0.2912041247,-0.3932519257,0.3314828873,-0.0566935875,-0.4954518378,-0.0110010151,0.053455092,-0.2503325939,-0.1332394183,0.0763655603,0.0955021605,0.2046999782,0.0583501868,0.0604519658,-0.5040941834,-0.0661611781,-0.004187671,-0.275547564,0.3255921304,-0.3012377024,-0.1762015224,0.0937422886,-0.558195591,0.2284720689,0.0903511792,0.1152762622,-0.0442058071,0.1032496691,0.3380590975,0.1329394579,0.7330361009,-0.0418952331,0.0436713547,-0.0261804424,-0.2548162341,-0.5318097472,-0.0710283443,-0.0235804822,0.4319599569,0.5820283294,0.3616741002,-0.1606953591,-0.0542591698,-0.098722823,0.3022754788,0.0390292853,-0.2510151267,-0.1459540427,-0.1491783559,-0.0676597357,0.0988165289,-0.1765549779,-0.033134982,-0.0086119846,-0.2094713002,0.1804425865,0.2442979366,0.1138890013,0.1692844778,0.2778656483,0.0919654444,0.5838769078,0.2333565056,0.1503582746,-0.0255127326,0.3816792369,-0.1456775069,0.2088950872,0.2213505208,0.0017615907,0.1407615691,0.5219025016,0.0168228764,0.0616738424,-0.0942675173,0.1803191155,-0.2476207912,-0.1468975395,0.4836440682,0.3026020825,-0.4802187979,-0.5349165201,0.3012941182,0.3897208571,-0.4364249706,0.642596066,-0.8030806184,-0.03487169,0.4852522612,-0.1162199229,0.791017592,-0.2726522386,-0.0640692562,-0.0942941904,0.1040655151,-0.0798156559,0.3242757618,0.2351139337,-0.3184455037,-0.1684567332,0.1873153746,-0.300688833,0.0404379517,0.1464457512,-0.0381777883,0.0571545549,-0.0338263027,0.0669829398,-0.2781297863,-0.0156273637,0.4708388448,-0.2859353125,-0.1470903158,0.0611566082,0.1181217283,0.3203131855,0.0010162162,-0.0786054432,-0.2998782396,-0.1230902448,-0.2078882158,0.0150032947,-0.3579515219,0.2997009754,-0.0144960387,0.0686281621,0.1134933755,0.1601995826,-0.0663905442,0.0720735937,-0.1098365039,0.1361834109,0.2847070694,0.4006179273,0.2626697123,-0.2824150324,-0.1337514371,0.0433796272,-0.0740462169,-0.2143929899,-0.1205190122,-0.1313812882,-0.4387509823,0.5310998559,0.0010711717,0.0806290433,-0.4185936749,0.2369568199,0.0427126922,-0.1557958573,0.0442956574,0.1502563804,0.1130943596,0.7382258773,-0.4055370092,-0.4140591323,-0.0183124784,0.2062739283,0.3746128082,0.077230297,0.3092643023,0.1556944698,-0.1866631955,-0.0473578349,0.0582417399,-0.1584671736,-0.2826387882,0.2330803871,-0.2658883035,0.0336041898,0.2847352624,0.0910620168,-0.1708763093,0.1311800033,-0.0772812739,-0.1855891645,-0.1639417708,-0.1798882037,-0.0723745525,-0.0252358355,-0.1616580039,0.1054512858,0.0400090739,0.4681864977,-0.1916241646,0.096151337,0.1173207536,0.1907104254,0.0411051512,-0.0549228303,0.2053995579,-0.186453715,-0.0231093504,0.1262726337,-0.1104048565,-0.1304396093,0.0198093597,0.198534146,-0.1801177859,-0.3488309681,0.140638724,-0.2483157068,0.1380702406,-0.4525107145,0.2766251862,0.3829393983,-0.0181527399,-0.4886367321,0.3218188882,-0.0284851342,0.0680268109,0.3769405782,-0.3603972197,0.0083349906,0.0633944348,0.0989835337,0.299576968,-0.0980074853,0.0845323205,-0.1769239604,0.2232199162,0.2317640781,0.3134753406,-0.2220516503,0.0455839373,0.4461660981,0.2535208166,0.2841158807,0.0314728953,-0.0645191148,0.0951099694,0.0759103224,-0.0357556753,-0.1070595533,0.6998181343,0.0796739161,0.0061073015,0.0736855268,0.2419078946,0.0853193477,-0.4488469958,0.0580467209,0.1969972849,-0.1601164639,0.1415559351,0.2287372798,0.2382544279,-0.0594731048,0.2625474632,-0.0134179797,0.0154396798,0.2349777371,0.1971016824,0.141542986,0.198015824,0.0884738415,0.0863742381,-0.4029288292,0.0977751464,0.2925219238,-0.3569515347,0.0906481594,0.1319441944,0.2310852557,-0.3770206571,-0.2672022879,-0.2398117036,0.2937868536,-0.1931099445,-0.1836691797,0.1328747869,-0.0047694137,0.1510374993,0.1858956367,-0.1211304665,0.0269820169,0.5510709882,0.1247075871,-0.029057743,-0.1535293162,-0.488003701,-0.09871687,0.3059157729,-0.2534528375,0.0592577644,-0.3243092,-0.1986314654,-0.2053003758,0.2730235755,0.4457674325,0.5597249866,-0.0652721897,0.0134479208,0.0529808961,0.0990436971,-0.1248611063,0.2919400632,-0.1225785688,0.1639741063,0.053507328,-0.0336245038,-0.1490589678,-0.053845942,-0.0791351572,0.2418648601,0.1019078419,0.1280561984,-0.3148501813,-0.1739650518,-0.0542586856,0.0480022579,-0.1228197664,-0.1263462603,0.3955925405,-0.3989728689,0.1068118662,-0.0944724455,0.0192604363,-0.2891869545,0.4734735191,0.2145934403,0.2468398213,-0.3985393047,-0.1295465231,-0.4535676241,0.057049863,-0.6260553598,0.3307388127,-0.0605844446,0.1703847349,-0.0630168915,0.006990273,0.1890663803,-0.0693134367,-0.1687947959,0.2257017046,-0.2695319951,0.0663509518,-0.3263704181,-0.176891163,0.1285068244,-0.4083438516,0.3732220531,-0.2949460149,-0.0018864212,-0.1522426307,0.1093186215,0.0873928517,0.1328686029,0.5245754123,0.440628022,0.3837631941,-0.1255714148,-0.1049544662,-0.138208881,-0.2574093044,-0.0852334723,0.1945772171,0.0846089795,0.1691830903,-0.2388857603,-0.0160217006,-0.2639772296,0.0940710828,-0.0943159312,0.113681823,0.0826008096,-0.3479435742,-0.2195760608,0.2355388701,0.0721242949,0.322988838,0.0230129939,0.1062776446,-0.1316958815,-0.1287914366,0.3937507868,-0.6275826097,-0.3203854561,-0.3296598494,0.2324418873,-0.088657707,0.0667531192,-0.0019408643,-0.202151835,0.2370859385,0.0712474436,-0.1932217926,0.1157781109,-0.271718353,-0.0457119755,0.032885436,0.0542979389,0.0397300646,-0.5831233859,0.2457560003,-0.100156337]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2243","title":"Map is slow and processes batches one after another","comments":"Upgrade to 1.6.1 solved my problem somehow. I did not change any of my code, but now it starts all processes around the same time.","body":"## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!","comment_length":25,"text":"Map is slow and processes batches one after another\n## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!\nUpgrade to 1.6.1 solved my problem somehow. I did not change any of my code, but now it starts all processes around the same time.","embeddings":[-0.328520298,-0.4251143336,-0.0902183875,0.3159681857,0.017065281,0.090156652,0.2094428092,0.4447200894,0.4825794697,0.0649622902,0.2510423362,0.31609267,-0.0560902543,-0.3143905997,-0.2324120849,0.1299518049,0.1601340771,-0.1192271411,-0.0551301017,-0.2028851807,-0.3048771024,-0.0078666881,-0.1871987879,0.0967328474,-0.1404462755,-0.1534426957,-0.0539782532,0.1025865152,0.0483303927,-0.3943070173,0.0211049281,0.0591858663,-0.1813183576,0.8799492717,-0.000122085,0.0114948386,0.2477268279,0.2849186957,0.1633939743,-0.0141195599,-0.1903142631,-0.2765859067,0.2130437046,-0.1274119765,0.2456354648,0.0258935522,0.0164517108,-0.640250206,0.1821404994,-0.032140784,0.1067631543,0.0738543719,-0.3857874274,-0.0927315801,0.0135707865,-0.0621790886,0.018127799,0.236090228,0.508998394,-0.1530715078,-0.167480737,0.1986735314,-0.3881816566,0.3281824887,0.052534923,0.0307636634,0.6446434855,-0.5962138772,0.2733977437,0.1532078832,-0.0808097795,0.0810292587,-0.0253483243,-0.3553411067,-0.3839282691,0.0868265256,0.0888893381,0.1113566384,-0.1912720352,-0.1945640147,-0.4781170189,0.070794791,0.2309859097,0.0300617702,0.1173126921,0.1442994177,0.2024504691,0.3093335927,0.4033643007,0.2462476939,0.2411026657,-0.2332062125,0.0064064409,0.2732770443,-0.4459080398,0.1241496801,0.5200048685,0.0867764279,0.0475485101,-0.5806805491,-0.0912430659,0.0393422171,0.1242073253,0.0633936077,0.3477093577,-0.0904643461,-0.0357493237,0.2498846352,0.3441907465,-0.2958385944,-0.4390595257,0.2486191094,-0.0160911288,-0.5093580484,0.1471351683,0.0583608672,0.0214539897,-0.0482897162,0.0136080198,-0.1149354875,-0.1971804351,-0.2436129451,0.1031602472,0.3249608278,-0.1104527339,0.239444837,-0.1344807595,-0.1778879017,-0.2257106304,0.289318949,-0.1174995601,-0.2958429456,-0.3196435571,0.138715297,0.0025330682,-0.373770535,0.035556972,0.4177811444,-0.3017961085,-0.1876014322,0.1808341444,-0.3083860576,0.0056096427,0.0124672046,0.0029921175,0.5556014776,0.1374043971,0.5318370461,-0.2708052695,0.3144215345,-0.5807901621,-0.1458477825,0.1188044995,0.0506162941,-0.0251623224,0.2462059408,-0.1311655939,0.0963201597,0.2827989459,-0.329559505,-0.2998921871,-0.1696284413,-0.287966311,-0.2514474988,-0.1645826995,0.3703288734,0.0015978087,0.0699285194,-0.2103019953,0.0787313804,0.4496004581,0.4156296253,-0.207585305,0.3538426459,-0.2240495533,-0.062329974,0.132104829,0.0309717897,-0.4005418718,0.667723,-0.2599672675,-0.0095557217,-0.063424781,0.2496449798,0.4455076754,-0.1556868404,0.282045275,0.26270926,-0.0572414771,0.4519650042,-0.4122684896,-0.1481003761,0.0155626237,-0.0891285837,0.0221562926,0.1694601625,0.1445702612,-0.4743440747,0.3899276555,-0.0761157721,0.035759937,0.1839105785,-0.0694310591,-0.0629471987,0.0335050672,0.0339341462,-0.1193147749,0.2032585144,-0.04320715,-0.2309648842,0.1026155353,0.0101115936,-0.0872514844,-0.0492902808,-0.1813875735,0.0792013332,-0.0921608955,0.2140888721,-0.0339442566,-0.0652823299,0.1847362667,0.0358002447,0.1898382902,-0.2237684876,0.0226210859,0.0642823428,0.0673079491,-0.1615779996,-0.0262198243,-0.2895718813,0.3231557608,-0.1598036289,-0.2133892477,0.1712712497,0.0485898443,0.4096708298,-0.0624726862,-0.0313412324,0.0597195365,0.3930762112,0.1482026279,0.0926675573,0.1199301705,-0.310262084,0.0450549014,0.2158518285,0.0695550591,0.461722523,-0.0773711205,0.0262373984,0.0060173627,-0.0828284845,0.2806044817,0.0949622914,0.3291273713,0.0061603384,0.4355294704,0.3036996126,0.0188760124,0.0724018663,0.5454025269,-0.0023864422,-0.2093969584,0.1821486056,-0.0453135297,-0.0745837241,-0.0186945293,0.145009473,0.4704556167,0.0592017323,0.0415220857,-0.1976518631,0.1342905611,-0.0113459537,0.0264968798,0.1439185143,0.5454495549,-0.0772547647,0.3817724288,0.0127196899,-0.0481063426,-0.2288660705,0.1870306283,0.2764610648,-0.0724953115,0.2033232301,-0.0034867437,0.3172149062,-0.2773835361,-0.2160998434,0.0942140892,-0.1857625246,-0.2104178518,0.1738044173,0.0202491563,0.1362210959,0.1513073444,-0.0940554738,-0.0637962818,0.0024563335,0.0040113325,-0.3196435273,-0.0449066982,-0.0813286752,0.2629683018,-0.049146641,0.2252518833,0.150646612,-0.4977270663,0.0932104513,-0.0467607193,-0.0155137442,-0.1538104713,-0.1058646813,-0.2013936043,0.1916435808,-0.0757521987,-0.2173840702,0.2132003754,-0.630905211,-0.0781372637,-0.0248469114,-0.1241870075,-0.271828115,-0.1186771095,-0.1993278116,-0.2023314089,-0.1979486197,0.2467657775,-0.1272894144,0.3811453581,-0.2140100747,0.1172325164,0.0577185117,0.0452684984,-0.1253619492,-0.1413497925,-0.3023563921,0.1250578463,0.0086253742,-0.0639722943,-0.3012583852,-0.1058022976,-0.0266058855,0.3626348376,-0.3844183087,0.1069599986,-0.5251045227,0.1232716441,0.0808826983,0.1221255064,0.4194758534,0.0025602442,-0.0889196843,-0.1034303755,-0.225945279,0.0017521264,0.1917863339,-0.0445029251,0.0387105793,0.5795582533,-0.1471131295,0.6572436094,0.3843574822,-0.0609936826,-0.0111356899,-0.0183652584,-0.1962141842,-0.2524475157,-0.3930990398,0.0927753448,-0.2513141334,-0.3111329079,0.2499520332,-0.06590195,-0.4047551751,0.0030452067,0.0311716069,-0.23516877,-0.1224618107,0.0686495081,-0.1459804028,0.3705112338,-0.0389427021,0.1677023619,-0.4189141691,-0.06357757,0.0560183711,-0.285654217,0.3517279923,-0.3162992299,-0.053567376,0.3908151984,-0.5518405437,0.253433913,0.0308441035,0.1092310622,0.0284962803,0.1204988807,0.3649800122,0.1873167902,0.6455700397,0.0569411367,0.0430919193,0.0095905224,-0.250856787,-0.6014393568,-0.0165803246,0.0175076909,0.473849535,0.8081660271,0.4271784127,-0.1105487123,-0.0770256892,-0.0832907036,0.1911305934,0.0233745817,-0.2846399248,-0.1913906187,-0.1285206228,-0.1593597382,0.1185397357,-0.1695539057,-0.0058777272,0.033845678,-0.3186742365,0.1142633036,0.0943030715,0.105259195,0.1232985929,0.2086758614,0.0323736817,0.5327065587,0.2790551782,0.0622501224,-0.015738653,0.3771702945,-0.1509032696,0.0702421293,0.174802497,-0.0256217886,0.1866864562,0.4035952091,-0.0072396826,0.0265930798,-0.099100098,0.2236520499,-0.1720789373,-0.0332782604,0.4676235616,0.1350873858,-0.4751652479,-0.5731217861,0.2930725217,0.2454826087,-0.4326726794,0.7999383211,-0.7203829288,-0.0233216044,0.4647375941,-0.0582020693,0.7366470695,-0.2252163589,-0.0468834043,-0.1134913191,0.0016395918,-0.0966490507,0.545013845,0.2873944938,-0.3282624185,-0.0226928256,0.1914458722,-0.2536578178,-0.06440337,0.160286665,-0.0610740073,0.0084821619,-0.051601693,-0.0880756676,-0.3143183291,0.0132995453,0.4626286924,-0.282548964,-0.0696026906,0.0295955986,0.0774158686,0.3475759625,0.0496011339,-0.1678541005,-0.1329222769,-0.1412420124,-0.1106954366,-0.0064310972,-0.3458260894,0.3355688453,-0.0181876477,0.0017519674,0.0954991654,0.0754738525,-0.1531455815,0.0275327135,-0.010044991,0.1465115249,0.3092437685,0.3101764917,0.2597781122,-0.1874192804,-0.2010074556,0.0728460327,-0.0879890174,-0.2149649262,-0.105313018,-0.2370033115,-0.514030993,0.4933771491,0.0763710216,0.1374213099,-0.4340177476,0.1306513548,-0.1092270687,-0.1804885715,0.0389337279,0.1331738234,0.1753908396,0.6333762407,-0.3870874345,-0.478504926,-0.0471135229,0.2769545317,0.2634620368,-0.0876094848,0.2662312984,0.0843804553,-0.1896448284,-0.0717234537,0.0679041818,-0.0488685407,-0.2146081477,0.3136089444,-0.3062498271,-0.0101046348,0.3257450759,0.0531307757,-0.1956655532,0.1932851821,-0.1045657024,-0.1416771859,-0.3051638007,-0.2252979726,-0.17719239,0.0403807648,-0.2287755311,0.101802595,0.0249871686,0.5330877304,-0.1708766669,0.0492288619,0.2193306535,0.1439285874,0.0015927353,-0.0504938662,0.1143140793,-0.1421094388,-0.0102518545,0.1347477585,-0.070913069,-0.0711639449,-0.014539497,0.2056803852,-0.2066223472,-0.2353150249,0.1806330085,-0.2686134577,0.0379463769,-0.3392085433,0.2073043585,0.3485957384,-0.0793942958,-0.3916223645,0.2095500678,-0.0793947503,0.2350161076,0.1846557409,-0.4053307176,-0.0395529568,0.0769044384,0.0990409404,0.3507372439,-0.1054120138,0.0943303928,-0.1459418833,0.3196319938,0.1992683858,0.3192611933,-0.1064087376,0.0653424114,0.5074803233,0.1796132475,0.3057312071,-0.0781085342,-0.038844265,-0.0058617806,0.0930525884,-0.0194858536,-0.1145032495,0.545017302,0.0727393329,-0.0282820798,0.0170117095,0.0484888442,-0.0953810662,-0.5180805326,0.1688509732,0.2100942433,-0.2860741317,0.1390002817,0.2218402922,0.1580220461,-0.0890189856,0.3816064596,-0.0422558077,0.0520954058,0.3182891607,0.1820548922,0.2886223793,0.1145864725,0.1258770078,0.0638243333,-0.3519522846,0.0876094848,0.2372218817,-0.3662941754,0.079491362,0.2309834659,0.1598940939,-0.4054525793,-0.1986223906,-0.2021134943,0.1854862869,-0.1995558292,-0.1512770504,0.1649578661,0.0245289728,0.1848349571,0.1754570901,-0.0838196203,0.1255090684,0.5904154181,0.1692120731,-0.0504487492,-0.0849395916,-0.4554617405,-0.0362176411,0.316429466,-0.2982967496,0.0353239365,-0.3432677984,-0.0946320221,-0.2026021332,0.2515508235,0.4208081961,0.7341746688,-0.0967566893,-0.0124222822,-0.0009803929,0.0785954595,-0.171117425,0.3257127702,-0.0512995683,0.3729535639,-0.0278777834,-0.0585723184,-0.1342336386,0.05441311,-0.0365598723,0.170639649,0.1491724104,0.3705941439,-0.2488196641,-0.274838537,0.0195744801,0.0687912256,-0.1372283101,-0.2307241559,0.367639631,-0.3615260124,0.0863070861,-0.1536700726,0.0120410407,-0.238663733,0.3800439835,0.1147632971,0.2215904742,-0.4694232643,-0.1976955831,-0.450178206,0.1214562058,-0.5382359028,0.4328245819,-0.0062182159,0.1362378746,-0.0949910656,-0.0379765332,0.1784360856,-0.0368308499,-0.2325476408,0.1282736957,-0.2308744043,0.1267989278,-0.3797703683,-0.1577134132,0.2243397236,-0.3874178529,0.3878501952,-0.3890949488,0.0122346971,-0.1171746477,0.0684707686,0.1629689187,0.1270407587,0.5165662766,0.4254449904,0.2855575383,-0.2117873728,-0.0606583357,-0.2839683592,-0.2531270683,-0.1093201488,0.1243836805,0.1298407316,0.1171877235,-0.2266769856,-0.2070804685,-0.3209629655,-0.014978407,-0.0732587501,0.2023950964,0.1799793094,-0.2647532225,-0.0994758159,0.2698439956,0.0632701963,0.2581601739,0.1208816171,0.0523911975,-0.212166369,-0.165854305,0.3300718665,-0.4964949489,-0.3814640641,-0.3555073738,0.3037727177,-0.084585011,0.1005764008,-0.1509470195,-0.1820100993,0.3137876391,0.019148821,-0.2466621548,0.0723781362,-0.1470407397,0.0328962989,0.0417607687,-0.0022273243,0.1815617234,-0.5332587361,0.2090578377,-0.1272593439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2243","title":"Map is slow and processes batches one after another","comments":"Nice ! I'm glad this works now.\r\nClosing for now, but feel free to re-open if you experience this issue again.","body":"## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!","comment_length":21,"text":"Map is slow and processes batches one after another\n## Describe the bug\r\n\r\nI have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. \r\n\r\nI process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps.\r\n\r\npseudo code:\r\n```python\r\nds = datasets.load_from_disk(\"path\")\r\nnew_dataset = ds.map(work, batched=True, ...) # fast uses all processes\r\nfinal_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another\r\n```\r\n\r\n## Expected results\r\nSecond stage should be as fast as the first stage.\r\n\r\n## Versions\r\nPaste the output of the following code:\r\n- Datasets: 1.5.0\r\n- Python: 3.8.8 (default, Feb 24 2021, 21:46:12)\r\n- Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 \r\n\r\nDo you guys have any idea? Thanks a lot!\nNice ! I'm glad this works now.\r\nClosing for now, but feel free to re-open if you experience this issue again.","embeddings":[-0.3070743978,-0.4695039988,-0.092293784,0.2605443895,0.0237791054,0.0816692859,0.2060968131,0.3872794509,0.4719318151,0.020735044,0.2926188111,0.3568664789,0.0152331358,-0.172343865,-0.1459616125,0.0627288818,0.1288415045,-0.169583872,0.0138117224,-0.1798467487,-0.2991749048,-0.0734263286,-0.1795146763,0.0421460606,-0.1804162115,-0.2228629589,-0.0007136218,0.1077165678,0.0487654805,-0.2912753522,-0.0189660471,-0.0574136265,-0.217557162,0.954675138,-0.0001176089,-0.0876431689,0.2521365583,0.2681464255,0.147297591,0.0785760432,-0.2093132585,-0.3373840749,0.1125773117,-0.2512634993,0.2299133539,0.0075566242,0.0000873652,-0.5254435539,0.278719008,-0.1038165018,0.1410560757,0.0903238654,-0.3887793422,-0.0974326655,0.0356710441,-0.0157806017,-0.0041974084,0.1936461031,0.5403661728,-0.2044119835,-0.1470499635,0.165630877,-0.478243053,0.2883205414,0.1425646096,0.1045709103,0.579603076,-0.5662628412,0.2586008608,0.1645043492,0.0364158973,0.0676730499,0.0002429153,-0.4046181738,-0.4485390782,0.0004350235,0.0631367639,0.1191911995,-0.2013661861,-0.2541602552,-0.4854084253,0.060443718,0.2294803113,0.0657129213,-0.0226833243,0.1426103115,0.1721695662,0.2969888151,0.3728101552,0.1315265447,-0.0061089615,-0.2827489972,0.0329689942,0.2808256447,-0.4053996205,0.0805034265,0.5078934431,0.0328569859,0.0499742068,-0.354185909,-0.1188242659,-0.0054993713,0.0185323618,0.1302485466,0.3539407551,0.0411518216,-0.1375973225,0.1864664257,0.4516462684,-0.2629924417,-0.3996079564,0.1982381195,0.0668713972,-0.566060245,0.168046385,0.0080164503,0.0050520482,-0.0405716002,-0.0489652902,-0.0865609571,-0.1470603645,-0.1776747704,0.1613954008,0.3504637778,-0.1179773435,0.1744433045,-0.1617615074,-0.0972028673,-0.2360582799,0.2652748227,-0.1058576852,-0.3608027697,-0.3060231507,0.1375770718,0.0916159824,-0.223560378,0.1049929038,0.3826281726,-0.3034338951,-0.1787350923,0.1791311353,-0.3981319368,0.0234944597,0.0205251202,0.0504818633,0.5985868573,0.1087485179,0.2657260597,-0.2467353791,0.3976507485,-0.57773453,-0.1601953059,0.2145015895,0.0432265699,0.0021722559,0.2794611752,-0.2627557814,0.0550751425,0.3159916103,-0.1885834187,-0.2529248595,-0.1632301658,-0.2909041643,-0.3229499161,-0.1161959544,0.3648350835,-0.0555359013,0.0846190229,-0.2270486057,0.1989284754,0.4607825875,0.6413199902,-0.2464523613,0.4451713562,-0.2116091698,-0.0282693449,0.0078554638,0.0238374546,-0.2697035968,0.7397873998,-0.3511784077,0.0436996631,-0.0849282965,0.2499584705,0.3816208243,-0.2272116244,0.2943175733,0.3158895075,-0.0448896587,0.4030993581,-0.3237964213,-0.0970086679,0.1848305166,-0.0664784908,0.0195168704,0.0924740434,0.1116137728,-0.3846982121,0.3901068568,-0.0824221596,0.0202758014,0.2032290846,0.0583020747,-0.0374687687,-0.0450531319,0.146766901,-0.0363038629,0.2148035765,-0.0297569036,-0.3130533397,0.0280722175,0.0901294872,-0.0566683635,0.0405774973,-0.1481070369,0.0768140927,-0.069609493,0.1082078964,0.0066256733,-0.053109277,0.2858844399,0.0768032819,0.1393086016,-0.2579503953,-0.0606353693,0.0649239197,0.0832973793,-0.0895802304,-0.1150141731,-0.247845158,0.3522396386,-0.077013202,-0.1728483886,0.1359662265,0.0681286901,0.2770183682,-0.0817129314,0.1085997447,0.0848038718,0.2572034895,0.1620967239,0.0714403614,0.2367134541,-0.3935838342,0.0078216819,0.1139740646,-0.0127198147,0.4513719976,-0.1187484935,-0.0240847208,-0.0157214329,-0.1067117453,0.2500286996,0.1847309768,0.2829656899,-0.1002832726,0.4187237322,0.2549993992,-0.0326462835,0.0118966671,0.4661342502,-0.0203538276,-0.1252837181,0.1260020733,-0.1191269979,-0.0074430616,-0.0529824682,0.1171952114,0.4750802815,0.0915320888,0.0083646923,-0.1771562546,0.1136810407,-0.0731731281,0.079966262,0.1351906955,0.4936067164,-0.0609047115,0.3400739729,-0.0251336861,0.0226746611,-0.1699529886,0.2145279348,0.2423357219,-0.0909221172,0.1643041819,-0.0658701807,0.1948194802,-0.3574422002,-0.0718456432,0.0449446961,-0.1869829446,-0.1754426956,0.231722489,-0.0079136491,0.0853914693,0.2350057214,-0.2120740265,-0.0586857758,-0.121783942,0.0293068215,-0.3171491325,-0.0466824733,-0.0308324005,0.3746777475,-0.0422525331,0.2445873618,0.1350599974,-0.4562498331,-0.0100961933,0.0141648315,-0.1387724727,-0.1915162355,0.0532073863,-0.2786862254,0.2900403142,-0.0724834427,-0.2245515436,0.2720659077,-0.6565895677,0.007991367,-0.0480761863,-0.1245818883,-0.3801763952,-0.0833954886,-0.1291444749,-0.1720814258,-0.164166823,0.3399759233,-0.0990640596,0.33451882,-0.0075923009,0.0989273489,0.0189516041,0.1534696668,-0.1666975021,-0.1813574433,-0.2220367491,0.1597072631,0.0227446314,-0.0923017189,-0.2212070227,-0.0145056015,-0.1139128134,0.2172623575,-0.4146239161,0.1890673637,-0.5987946391,0.2085665315,0.0747189671,0.1890527457,0.3789047003,-0.1311908364,-0.0708762929,-0.1619340628,-0.3247306943,0.0190550517,0.2639814913,-0.0260156337,0.1003705859,0.4866039455,-0.0761096552,0.6302146912,0.3575815558,-0.0333175883,0.0737168267,0.0584787093,-0.1835186034,-0.2872954309,-0.3533495069,0.0379477702,-0.3034133017,-0.3304222226,0.2818643749,-0.0600946583,-0.5309894085,0.0297307577,0.0659345463,-0.2661901116,-0.0936015844,0.1138191521,-0.0231470428,0.2349762321,0.0273402408,0.1085844487,-0.4608954787,-0.0767427981,0.0334723331,-0.3773456514,0.3688298762,-0.2808011174,-0.1387191862,0.1179228351,-0.5799161792,0.1996640712,0.0383365378,0.1366108209,0.0005053294,0.1017595157,0.3043722808,0.1429289281,0.7509541512,-0.0038133243,0.1073134914,0.0068127341,-0.2112995535,-0.5945114493,-0.0318583809,-0.0196369663,0.4358693659,0.6071033478,0.3249939382,-0.1549628228,0.0118084429,-0.1762377173,0.2500935793,0.0162843317,-0.2643999457,-0.1277375519,-0.1821655482,-0.137760058,0.1234620437,-0.1214194298,-0.0392844528,0.0229101032,-0.2260066122,0.2043337226,0.2336878926,0.0908891633,0.1068893299,0.2693607509,0.0611444823,0.5416282415,0.2114551067,0.1885899901,-0.0642715394,0.350592792,-0.1188224927,0.2857779562,0.1973268837,-0.0285961237,0.2222449929,0.4491458237,-0.0043714866,0.0523670502,-0.1520204395,0.218400836,-0.2671606541,-0.1502926946,0.5050221086,0.2374799699,-0.45411551,-0.5346496105,0.2677076757,0.2974103391,-0.3881889582,0.6406934857,-0.8524453044,-0.052479241,0.4936686456,-0.0822669566,0.7306090593,-0.2935535312,-0.0979340374,-0.1213341057,0.1403482407,-0.1560688764,0.2540215552,0.2495049387,-0.3049434125,-0.1173626333,0.2125660926,-0.2820887566,0.0246986877,0.16196841,-0.03268563,0.0415708199,-0.0802704543,0.0633880049,-0.281693697,0.0205937698,0.483093828,-0.3060741127,-0.1351395696,0.0746344626,0.1742425114,0.2771479189,-0.0018766554,-0.0944878757,-0.2551518679,-0.1927969307,-0.1149889901,0.011113408,-0.3075273335,0.3527854979,-0.0638325363,0.1049792618,0.1013514623,0.159316808,-0.1946602464,0.1353972256,-0.0978001356,0.1479282379,0.3188362718,0.406047523,0.280249387,-0.2338894606,-0.1790453494,0.0335420817,-0.0941350684,-0.1612543762,-0.1206197813,-0.1833854467,-0.5209077001,0.5433419943,0.0646672174,0.1150690019,-0.4308284223,0.2202363312,-0.0204852559,-0.1794325113,0.0661323294,0.1343456209,0.0650712848,0.7289732099,-0.4065171182,-0.39237234,-0.0028013331,0.2055414617,0.3045361936,0.1434856057,0.3214297593,0.1601703912,-0.1942386627,-0.0825454444,0.0780684575,-0.0789752752,-0.2984722853,0.1853135526,-0.2877359092,0.0244949237,0.2984972298,0.0183137972,-0.1707783639,0.0987780765,-0.1191510782,-0.1615107358,-0.2252186835,-0.1782317609,-0.0896639302,-0.0534119718,-0.1646291763,0.1401908249,0.0422300473,0.5476230383,-0.2087198943,0.0895352885,0.1256801784,0.2231359631,0.0012663828,-0.0650873855,0.1448468268,-0.1340653002,0.0154731106,0.1855998039,-0.0939445198,-0.1516582668,0.0192496255,0.1828035116,-0.2072361857,-0.3256991506,0.1942347884,-0.3056369126,0.0746300295,-0.416416347,0.2432856113,0.3443967104,-0.0149540035,-0.4733923376,0.3334658444,-0.0491185673,0.102640979,0.3028348088,-0.4161450267,-0.0191366449,0.0372498184,0.0622454248,0.3463118076,-0.0950258672,0.0308497176,-0.1575402021,0.2799505889,0.2421229929,0.3736259341,-0.2309811115,0.0847024992,0.4543590248,0.2566924989,0.1965305954,-0.0343706198,-0.0319284573,0.0931535736,0.1203756109,-0.0710514933,-0.1435079873,0.6602487564,0.0670212507,0.0054216692,0.055053819,0.2134027183,-0.0290263575,-0.4931353927,0.0877278,0.139703393,-0.195875138,0.0922127143,0.2872327566,0.1633993536,-0.1455360055,0.2326078713,-0.0387281738,0.054710079,0.3058010042,0.2399388552,0.2004847974,0.205123812,0.0565315075,0.0162157305,-0.3659247458,0.0470152013,0.2293115109,-0.3461771905,0.1094781533,0.1486751437,0.1915073246,-0.3070209324,-0.2250908017,-0.2419805527,0.2735802233,-0.2309373915,-0.1511179656,0.1628222913,-0.0024767944,0.1675643772,0.1298668385,-0.0982036144,0.0116576217,0.5087919235,0.1298430115,-0.0827079415,-0.1088689044,-0.4329910278,-0.0931456983,0.2954479754,-0.3275838792,0.0223977827,-0.2966714203,-0.1493007243,-0.2413934618,0.1802442521,0.4253827035,0.6085185409,-0.0797764212,0.0685900822,0.0358746015,0.0571244322,-0.1609904766,0.3388271332,-0.1143610626,0.205250591,0.0680826232,-0.0217146706,-0.1720154285,-0.0640570968,0.0177934449,0.2640702724,0.1710684597,0.1635029316,-0.2422294021,-0.2544052005,-0.053501755,0.0303050037,-0.1356640458,-0.128720209,0.3978264332,-0.3950270712,0.1284874082,-0.1210020781,0.0311057437,-0.3333319128,0.4219708145,0.1288329065,0.228645429,-0.4557029307,-0.169599846,-0.4521678984,0.1259710342,-0.6484364867,0.3481283486,0.0013070634,0.1190322712,-0.0279904492,-0.0126905395,0.1784439981,-0.0967017338,-0.1622726917,0.1998983026,-0.2243278772,0.0600521863,-0.3359034956,-0.1105844602,0.141398713,-0.4349131286,0.378295809,-0.338589102,0.0600012727,-0.1619616151,0.0991833806,0.1155298129,0.2375809997,0.5331740975,0.431579262,0.3331642151,-0.1517585069,-0.0499521941,-0.1837177873,-0.2088989466,-0.0783288181,0.1293917894,0.1191177219,0.1649010628,-0.2282634825,-0.0550477952,-0.2991815507,0.0292446539,-0.0940079316,0.1457363814,0.0831633434,-0.3136825562,-0.2114542723,0.2175403386,0.0734681189,0.2862501442,0.0508431755,0.0616763979,-0.1026221737,-0.1564363092,0.3905712366,-0.5741915107,-0.3926385045,-0.3124676645,0.2573976815,-0.1148566678,0.1415224373,0.030655602,-0.1947878897,0.2509527802,0.118468456,-0.2298499197,0.0855281204,-0.2733574212,-0.054997649,0.0728235468,0.0326524265,0.070257701,-0.5396556854,0.239121139,-0.1208748668]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2242","title":"Link to datasets viwer on Quick Tour page returns \"502 Bad Gateway\"","comments":"This should be fixed now!\r\n\r\ncc @srush ","body":"Link to datasets viwer (https:\/\/huggingface.co\/datasets\/viewer\/) on Quick Tour page (https:\/\/huggingface.co\/docs\/datasets\/quicktour.html) returns \"502 Bad Gateway\"\r\n\r\nThe same error with https:\/\/huggingface.co\/datasets\/viewer\/?dataset=glue&config=mrpc ","comment_length":7,"text":"Link to datasets viwer on Quick Tour page returns \"502 Bad Gateway\"\nLink to datasets viwer (https:\/\/huggingface.co\/datasets\/viewer\/) on Quick Tour page (https:\/\/huggingface.co\/docs\/datasets\/quicktour.html) returns \"502 Bad Gateway\"\r\n\r\nThe same error with https:\/\/huggingface.co\/datasets\/viewer\/?dataset=glue&config=mrpc \nThis should be fixed now!\r\n\r\ncc @srush ","embeddings":[-0.2079320401,-0.4101823568,0.0818817839,0.183934167,0.3960956335,-0.2639939785,-0.025460504,0.0469439775,0.0375418961,-0.008497593,-0.2796851993,0.1796281934,0.4520506263,0.3836477995,0.178095445,-0.2308237106,0.0239983127,-0.1310635358,-0.3405912519,-0.0222028308,0.0464713834,0.1375623345,-0.1632145196,0.2611606419,-0.2712629139,0.1765830368,0.2629891634,0.2453493774,-0.2306219488,-0.5001212358,-0.0341178775,0.0697853863,0.1482548714,0.2818796039,-0.0001180756,0.1617345959,0.3973474801,0.193399474,-0.1244315356,0.0997093096,-0.5719077587,0.1414133906,0.0958769545,0.0627506599,0.0173608437,0.036552269,-0.3819513917,-0.2680610716,0.275500536,0.3098993003,0.1725411564,0.4018753469,0.0386901498,-0.3952960074,0.1691518873,-0.2286975831,-0.1305597574,0.0025716699,0.0725834817,-0.1062147692,-0.1978020817,0.3601199687,0.0352375619,0.0734086782,0.228143543,0.0490455031,-0.0516000688,-0.0528142005,0.025954973,0.4438382685,0.3182025552,0.0418138243,-0.1818636656,0.0548902601,-0.2071683109,0.3691947758,0.3216128647,-0.0768591762,-0.0347135961,0.4126830697,-0.194318518,-0.2949140966,0.0789057538,0.116567336,0.0289431661,-0.2193193883,-0.0387039669,0.1307688355,0.052110102,-0.2654346526,-0.2360070497,0.2029292285,-0.2673762441,-0.1569788307,-0.0897103697,-0.1051315516,0.2300347835,0.2421507388,0.3731906712,0.0436973162,0.0867728814,0.0485616624,-0.3023752272,-0.0150146764,0.1735997945,-0.0345129371,-0.2432551086,0.2066839933,0.3693188429,0.4046919346,0.0674674064,-0.0733760223,0.3246539235,-0.3338608444,-0.2971211672,-0.1475563794,0.0070449281,-0.459458977,-0.2007873952,0.081704855,0.2628745735,-0.1435906291,0.3435126543,0.4790741503,-0.2292525172,0.323226124,-0.5227351785,0.2160989493,-0.2436471283,-0.2039170563,-0.1611839235,0.073665984,-0.2443223447,0.2381793261,0.0595992804,0.1577541232,0.1179390401,0.1437665522,-0.1353629082,-0.3113238513,-0.1998685747,0.0085873781,0.1246219873,0.0555654429,0.186201781,0.4582801759,0.112808764,0.0399351157,0.097472094,-0.2610835433,-0.5629953146,-0.3686422706,0.3090554774,0.1173115671,0.0016180564,-0.1533698738,0.2060573399,-0.0479184501,-0.3529824913,-0.2393184006,-0.2527753115,0.3820371032,-0.1379339695,-0.0729718879,0.2998264134,0.3335144222,-0.0499699153,-0.258829087,-0.062283963,-0.1070018709,-0.1055706367,0.5090881586,-0.1061995253,-0.0569400676,-0.3307220936,0.0174819045,0.3710862696,-0.2424529344,-0.3206376433,0.3183476627,-0.1523713768,-0.2610199749,-0.1152958572,-0.0381245166,-0.1764903218,-0.0402926207,0.2197269946,0.1357414126,0.3019686341,0.0924974978,-0.3829333782,-0.073464185,-0.0539848544,0.055449754,0.1867210269,0.2088122964,0.1351392716,-0.4407036602,0.1912810951,0.0267505515,0.066483058,0.3745447993,0.312361449,-0.1794121563,0.204395026,0.2270141393,-0.0970791429,0.0331842192,-0.5256299973,-0.0648622364,-0.0202909261,-0.0250747558,-0.5287338495,-0.1067927703,-0.1149797514,0.025174601,0.0929314345,0.0118820649,-0.1798830777,0.2005951852,0.0048188148,0.0810085014,0.1312836707,0.1629065871,0.0951458737,0.3051271737,-0.1235362738,0.0695176646,0.1846170425,0.0832014382,0.4866995513,0.1958869696,-0.098736681,0.3124420345,0.0523275062,0.1254543662,0.387686044,-0.0049198992,0.2900629342,-0.5319990516,-0.1598387361,0.0364452824,-0.0189711191,0.1197652966,-0.2924672961,0.3572997153,-0.1236580685,0.3233401179,0.1206020191,0.4198075235,0.1894301921,-0.0784971416,-0.0171637684,-0.2803044915,0.3305770457,0.4617204964,0.2312447131,-0.1662032604,0.095467478,0.1469719112,-0.2532685995,0.1283091456,0.1988462955,0.0959067866,-0.1562442333,-0.1989060193,0.2815254033,-0.0148951644,0.0614652894,0.1402249336,0.1553698182,0.0958775431,-0.2424084395,-0.0771387443,0.0549034886,0.073890686,0.0000885097,0.1008636951,-0.1333822757,-0.20862104,-0.165943414,0.0219390951,-0.0667650402,0.1623274535,-0.5543406606,-0.1519482434,-0.2580004632,-0.4605607092,0.1408337951,-0.2011793703,-0.2062727809,-0.4525601268,-0.1347370148,-0.275002718,-0.4346733689,0.3653812408,0.0634351522,0.1317479759,-0.0738493279,0.3182646036,-0.143980667,-0.1561900824,-0.110997498,0.0341072045,0.0607987754,-0.0331952423,-0.1492891163,-0.120872654,-0.0798691735,-0.6939417124,-0.2089890242,0.3171244562,0.1947188526,0.4093862772,0.2355354279,0.279705137,0.0286503267,0.2312418669,0.3997167647,0.1907959133,0.1143829748,-0.1229714826,-0.2088721246,-0.1017529368,0.2608142197,-0.0327783264,0.2564601302,-0.3273408711,-0.1465438306,-0.0479792431,-0.0309316758,-0.1511523575,0.1942840666,-0.0368867293,-0.1872651875,0.2652763426,-0.3118976355,-0.4000118375,0.2376099527,-0.3980326056,-0.3438014388,0.0376916714,0.1258838624,0.1085850224,0.0403067768,-0.3281690776,-0.2229660004,-0.4650966227,0.0485233106,0.2034138441,0.1602338701,-0.150378719,-0.1265727878,-0.1645099968,-0.1102415696,-0.1578408927,0.0052933488,-0.3174500167,0.2889996469,-0.1255759597,0.2435949147,-0.0854845643,-0.0651963353,0.4545230865,0.3975493908,0.6014127135,-0.2267560959,0.4388616979,-0.1246186346,-0.070751965,0.0072592916,-0.074300237,0.0947192535,0.1978999525,0.3506551981,-0.1227606162,-0.415561974,-0.1896948069,-0.1154398173,-0.2990134358,-0.2446358055,0.0634351149,-0.0598841533,0.3705915213,0.2663883567,-0.0702877119,0.0173647385,0.3566446006,0.4829532802,0.0094538685,0.0422423296,-0.2975625992,0.2095445991,-0.3748435378,0.1082524136,0.0091652973,0.267801851,-0.1798118949,-0.2642815709,0.0160205811,-0.0403488055,0.2724252045,-0.4338154495,0.1370411962,0.0627663732,0.2488050759,-0.2807280421,0.0812875256,0.1163399443,0.0939239264,0.1293500215,0.0716144294,-0.2875830829,-0.1363586038,0.3011300862,-0.3630709052,0.0563064553,0.0265148077,-0.1800764948,-0.034293212,-0.0451079905,-0.229602322,-0.1661673337,0.1585776061,-0.0682724789,-0.1866075397,-0.0156070963,-0.2020759434,-0.0611592866,0.2085707486,0.4501496851,-0.1304760575,0.2716609538,0.3416740298,0.6558120847,-0.5079265833,0.2435099483,-0.1308071762,-0.227682218,0.1397461593,0.1148009077,0.0457772762,0.3514863253,0.1402000636,0.2410416901,0.2638845742,0.1799126118,-0.2170899063,0.3915747404,0.3607663512,0.0632261634,-0.6307272315,-0.0843646154,0.4638182819,-0.0320598595,-0.1760094017,0.3973103166,0.3934896588,0.1263943911,-0.1848055869,-0.0096169217,0.9444585443,0.2112264186,0.2680863738,0.2481930107,-0.0662590042,0.402936101,0.3032189906,-0.2240569144,-0.2987601459,0.0322895832,-0.0561766736,-0.0883690938,-0.1458416134,-0.3275798261,-0.1602214426,0.2338517457,0.2442379296,0.195832476,0.1308906823,0.0846396983,0.1978765726,-0.1322104186,0.0027065005,0.0868004858,0.0188061427,0.3331851959,-0.2114306092,-0.0708282962,0.0465612113,-0.2756946683,-0.1765013039,0.0777264237,-0.2166629732,0.4022295475,0.0443333052,-0.3876002431,-0.0366374217,0.033249829,0.3029325902,0.1374761313,-0.5003725886,0.4761455357,-0.162969023,0.029490903,-0.0463933423,-0.0956027955,-0.0604321882,-0.1060382575,-0.1456770748,-0.0301105175,-0.06406115,0.0199543927,-0.514090538,-0.1136839986,0.1842166483,-0.318634361,0.137521714,0.1066552103,0.1586422324,-0.3674170971,0.1002271399,0.1755866855,0.2190704197,0.1842354983,0.4654400349,-0.2538778782,-0.0649075806,0.3570060432,-0.3679180741,-0.3236111701,-0.0178516395,0.069313623,-0.3411992192,-0.1149816141,0.1765311807,0.2943526208,-0.3439387977,-0.0616217591,-0.0282991901,-0.0887426063,-0.0927288607,0.2250243723,0.038462542,0.0512735248,0.1737211645,-0.2161150873,-0.3811458051,0.1505220383,-0.2591615021,0.1264010519,-0.0718334988,-0.0886772648,0.2845101357,-0.1845457405,-0.2497955114,0.0688232556,-0.0572861247,-0.2266394794,-0.0648026168,-0.1372515261,0.2760649323,0.0107907373,-0.0191867612,-0.2863526344,-0.2185751796,-0.0594246238,-0.2420533001,0.1071575657,-0.000905184,-0.5251541138,0.0050465334,0.1836104691,0.0066178082,-0.0585086271,0.1773568988,0.1318379939,-0.0631385595,-0.3853952289,-0.3221164942,-0.2939242423,-0.4506383538,0.1906626523,0.4017994106,0.1436276883,-0.014217196,0.0243041161,-0.2104182988,-0.2284806222,0.255400151,0.1850674897,0.4798246622,0.0438669808,0.1784622967,-0.2365131378,0.1532271653,-0.0525121242,0.4528961778,0.2027381808,-0.0905317739,0.092914395,0.3229187429,0.1891653389,-0.0546307303,-0.099378176,0.1412086785,0.0762074143,0.1447910964,-0.2180961967,0.0148720248,0.2393725365,0.1338307559,0.01124781,0.2116000056,0.1254700124,-0.004041261,-0.0575794056,0.1391838193,0.1483383775,0.1036859974,-0.112348251,-0.0263438597,0.3227481544,-0.3542326987,0.0388513766,-0.1711325049,0.2618943453,0.0939052254,-0.2240329981,-0.06621968,0.137763083,-0.1826940477,-0.1178653836,-0.114590764,0.6248869896,-0.0682868212,-0.1213186979,-0.3798995912,0.141945079,-0.161689654,-0.399299711,-0.493452996,-0.2575891018,-0.2345027328,0.2185015827,-0.0772320852,0.1716244221,0.098044619,0.0576578118,0.1379882395,-0.6326800585,0.0263387747,0.3971800506,0.1744186282,-0.0480005555,0.084994249,0.0362885557,-0.0116089936,0.3378307521,0.2313068062,0.4022939205,0.1313348711,-0.0141365873,-0.1688605398,0.0623384975,-0.0883982033,-0.1138087735,-0.0677224621,0.0734540299,-0.1646361798,0.4268012941,0.1038761064,-0.2406455427,0.1747628003,-0.2475519478,0.3287488222,0.115453355,-0.0165495072,-0.3491408825,0.0462122895,-0.3172910511,-0.1954585761,-0.4131455719,0.410364002,0.049866613,-0.2069744617,0.1333594024,-0.2918366492,0.0438432992,0.1241832525,0.7378137708,0.0839133114,0.262976706,-0.3460187912,-0.4527204931,-0.4788594544,0.2479510754,0.0308975466,0.2380198538,-0.0056160353,0.0455634259,-0.5047969818,0.2425878048,0.4905583262,0.3964122534,-0.1493782401,0.1419239342,-0.2062953413,0.1359035671,0.434333235,0.023781037,-0.0065944092,0.0079047475,0.3816730082,0.1048034132,-0.0556481928,0.1382640898,0.1360291988,-0.2895013094,-0.3299596012,0.3706074953,0.2669917345,0.4297234118,0.1327373534,0.0167262331,-0.5268521309,-0.8276218176,0.068534404,-0.2457037121,0.131221965,0.0583419427,0.134956494,0.0848993734,-0.3310419023,0.2704519033,0.1296138912,0.4241008461,-0.1451950371,0.0598052032,0.3798525035,0.1495601237,-0.0889399126,0.0750699341,0.0603789762,0.2781733871,-0.3128536344,-0.6135975718,0.4709577262,-0.0012537525,0.2841418684,-0.3031010032,0.3841099739,0.3007038832,-0.1658782512,-0.315553844,-0.1297975034,0.0449210256,0.2161149085,-0.2282931507,0.1704607755,-0.0879188478,0.0531039461,-0.0970190391,0.5301214457,0.2182455361,-0.368303895,0.3664596677,0.1011485457]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2239","title":"Error loading wikihow dataset","comments":"Hi @odellus, thanks for reporting.\r\n\r\nThe `wikihow` dataset has 2 versions:\r\n- `all`: Consisting of the concatenation of all paragraphs as the articles and the bold lines as the reference summaries.\r\n- `sep`: Consisting of each paragraph and its summary.\r\n\r\nTherefore, in order to load it, you have to specify which version you would like, for example:\r\n```python\r\ndataset = load_dataset('wikihow', 'all')\r\n```\r\n\r\nPlease, tell me if this solves your problem.","body":"## Describe the bug\r\n\r\nWhen attempting to load wikihow into a dataset with\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\nI get the message:\r\n```\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\nat the end of a [full stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2).\r\n\r\n## Steps to reproduce the bug\r\n\r\nI have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https:\/\/huggingface.co\/datasets\/wikihow) says to use \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow')\r\n```\r\nto load the dataset. I do so and I get the message\r\n```\r\nAssertionError: The dataset wikihow with config all requires manual data.\r\n Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https:\/\/github.com\/mahnazkoupaee\/WikiHow-Dataset.\r\n You need to download the following two files manually:\r\n 1) https:\/\/ucsb.app.box.com\/s\/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under \/wikihowAll.csv\r\n 2) https:\/\/ucsb.app.box.com\/s\/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under \/wikihowSep.csv\r\n\r\n The can e.g. be \"~\/manual_wikihow_data\".\r\n\r\n Wikihow can then be loaded using the following command `datasets.load_dataset(\"wikihow\", data_dir=\"\")`.\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='')\r\n```\r\n\r\nSo I create a directory `.\/wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory.\r\n\r\nThen I run \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\n\r\nthat's when I get the [stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2)\r\n\r\n## Expected results\r\nI expected it to load the downloaded files into a dataset.\r\n\r\n## Actual results\r\n```python\r\nUsing custom data configuration default-data_dir=.%2Fwikihow\r\nDownloading and preparing dataset wikihow\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/azureuser\/.cache\/huggingface\/datasets\/wikihow\/default-data_dir=.%2Fwikihow\/0.0.0\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... ---------------------------------------------------------------------------\r\nAttributeError\r\nTraceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset('wikihow',data_dir='.\/wikihow')\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n745 try_from_hf_gcs=try_from_hf_gcs,\r\n746 base_path=base_path,--> \r\n747 use_auth_token=use_auth_token,\r\n748 )\r\n749 \r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n577 if not downloaded_from_gcs:\r\n578 self._download_and_prepare( -->\r\n579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs \r\n580 ) \r\n581 # Sync info\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n632 split_dict = SplitDict(dataset_name=self.name)\r\n633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) -->\r\n634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) \r\n635 \r\n636 # Checksums verification\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikihow\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2\/wikihow.py in _split_generators(self, dl_manager)\r\n132\r\n133 path_to_manual_file = os.path.join(\r\n--> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename \r\n135 ) \r\n136\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n```\r\n- Datasets: 1.5.0\r\n- Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0]\r\n- Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic\r\n```","comment_length":71,"text":"Error loading wikihow dataset\n## Describe the bug\r\n\r\nWhen attempting to load wikihow into a dataset with\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\nI get the message:\r\n```\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\nat the end of a [full stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2).\r\n\r\n## Steps to reproduce the bug\r\n\r\nI have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https:\/\/huggingface.co\/datasets\/wikihow) says to use \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow')\r\n```\r\nto load the dataset. I do so and I get the message\r\n```\r\nAssertionError: The dataset wikihow with config all requires manual data.\r\n Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https:\/\/github.com\/mahnazkoupaee\/WikiHow-Dataset.\r\n You need to download the following two files manually:\r\n 1) https:\/\/ucsb.app.box.com\/s\/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under \/wikihowAll.csv\r\n 2) https:\/\/ucsb.app.box.com\/s\/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under \/wikihowSep.csv\r\n\r\n The can e.g. be \"~\/manual_wikihow_data\".\r\n\r\n Wikihow can then be loaded using the following command `datasets.load_dataset(\"wikihow\", data_dir=\"\")`.\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='')\r\n```\r\n\r\nSo I create a directory `.\/wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory.\r\n\r\nThen I run \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\n\r\nthat's when I get the [stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2)\r\n\r\n## Expected results\r\nI expected it to load the downloaded files into a dataset.\r\n\r\n## Actual results\r\n```python\r\nUsing custom data configuration default-data_dir=.%2Fwikihow\r\nDownloading and preparing dataset wikihow\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/azureuser\/.cache\/huggingface\/datasets\/wikihow\/default-data_dir=.%2Fwikihow\/0.0.0\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... ---------------------------------------------------------------------------\r\nAttributeError\r\nTraceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset('wikihow',data_dir='.\/wikihow')\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n745 try_from_hf_gcs=try_from_hf_gcs,\r\n746 base_path=base_path,--> \r\n747 use_auth_token=use_auth_token,\r\n748 )\r\n749 \r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n577 if not downloaded_from_gcs:\r\n578 self._download_and_prepare( -->\r\n579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs \r\n580 ) \r\n581 # Sync info\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n632 split_dict = SplitDict(dataset_name=self.name)\r\n633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) -->\r\n634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) \r\n635 \r\n636 # Checksums verification\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikihow\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2\/wikihow.py in _split_generators(self, dl_manager)\r\n132\r\n133 path_to_manual_file = os.path.join(\r\n--> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename \r\n135 ) \r\n136\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n```\r\n- Datasets: 1.5.0\r\n- Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0]\r\n- Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic\r\n```\nHi @odellus, thanks for reporting.\r\n\r\nThe `wikihow` dataset has 2 versions:\r\n- `all`: Consisting of the concatenation of all paragraphs as the articles and the bold lines as the reference summaries.\r\n- `sep`: Consisting of each paragraph and its summary.\r\n\r\nTherefore, in order to load it, you have to specify which version you would like, for example:\r\n```python\r\ndataset = load_dataset('wikihow', 'all')\r\n```\r\n\r\nPlease, tell me if this solves your problem.","embeddings":[-0.2274995297,0.370223105,0.0238096733,0.3987950683,0.2537099421,0.2747787535,0.4276534021,0.4391774237,0.2458434254,0.0935082212,0.216285497,0.3853774071,-0.0128944451,0.1863868237,0.1205022633,-0.2686142325,0.0110668847,0.1045249701,0.223709926,0.1890229732,-0.3448888063,0.0065287249,-0.2329775989,0.1040147245,-0.3400695622,0.0130667957,-0.0981442481,0.1813103855,-0.0533038154,-0.541544497,0.4757557511,-0.1185788065,0.2518153787,0.4039177299,-0.0001126044,0.1877029687,0.5060688257,0.0342030711,-0.442166388,-0.0944597572,-0.3161360323,-0.1228394881,0.2420690507,-0.1337201148,-0.0326552615,-0.0258759465,0.004085809,0.0358592607,0.2355376929,0.2652969062,0.1653248966,-0.1196185276,0.3745595217,-0.1360630244,0.1942486763,-0.0933305025,-0.0869166702,0.1622466892,0.0739811137,-0.2382172346,0.1115441099,0.20557262,-0.1832122952,0.0634829551,0.5792307854,-0.0081012985,0.2564974129,-0.332991004,0.2632296085,0.3839665055,0.8194904923,-0.2249974608,-0.2384388745,-0.2182331532,0.1727716476,0.1644721776,0.3875032365,0.3061026633,-0.3748465776,0.0360186175,-0.0377139971,-0.2557303309,-0.0580877513,0.2808242142,-0.294668287,0.4108183086,-0.0778242275,0.1745568216,-0.2258381695,-0.0642693043,0.2035174668,-0.32481125,-0.2376295626,0.3015760481,-0.1384118348,0.0768692344,0.1585353613,0.1928257942,0.0994419307,-0.0371334255,-0.100838013,-0.1908228844,0.1564052403,0.1384836882,0.4442620873,0.2339715958,-0.0133317849,0.2062765062,0.1207495183,0.3048851788,-0.1984937936,-0.1141341776,-0.201289773,-0.2741613984,0.1328350455,-0.2296912819,0.3544214666,-0.0338293388,-0.2546579242,0.2001613677,0.0116372267,0.0730370209,0.0347945094,0.4159566164,-0.2064811736,0.1446116865,0.2597591877,0.2927667499,0.1025156602,0.1234883666,-0.1005089432,0.0907599106,-0.0558772162,-0.0318382233,0.1500617266,-0.1930603236,0.2553026974,0.1958255172,0.125093013,-0.2846936584,-0.0093548056,0.2284626216,-0.2356490493,0.3494037092,0.0726296008,0.1431750059,0.1701949537,-0.6475993991,-0.1152566895,0.1633387059,-0.394199729,-0.3855848312,-0.2756711245,0.2184655666,-0.0403429158,0.0450250953,-0.419618398,-0.1785893738,0.2975123823,-0.2910596132,-0.2390632927,-0.0699154362,0.1259678304,-0.1256434321,0.356915772,0.4782387316,-0.3908683658,-0.0335086882,-0.0930062383,0.0438776985,-0.0572247431,-0.3101461828,-0.3554069996,0.4971389174,-0.288457185,-0.2874653041,0.2937496006,-0.4576177597,-0.2892730534,0.2900746465,0.0192084331,0.3332006037,-0.0229105316,-0.1692107767,-0.168219924,-0.0836994424,0.2327568531,0.2221877873,-0.0055906828,0.0550968647,-0.1940852553,-0.2364349663,0.1046095416,0.186486274,0.0511644967,0.1748686433,0.0437120758,0.0242446903,0.1256706417,-0.1626351625,-0.0451247096,0.1203386486,-0.0281695016,-0.0047448212,0.048327703,0.0023430884,-0.4610753655,0.3484283388,0.0433676019,0.1083413661,-0.1505929083,-0.0570079163,-0.2748941183,0.1563291699,-0.5284996033,-0.1772395521,0.16646716,0.0618430004,-0.149765268,-0.018513212,-0.124017477,0.0431065783,-0.3334240615,0.1222898439,-0.2070953697,0.1706795096,-0.1306974888,-0.0565901399,-0.0111799212,0.1489347219,0.0346692316,0.0848748386,-0.2042343467,0.3257358968,0.0568369739,0.1124035195,-0.0444702618,-0.0688852519,-0.037389446,-0.1064234599,0.0548359565,0.2269100845,0.2477425188,-0.0861583799,-0.092016995,0.230902642,0.0540100299,0.3870119154,0.0292328987,0.267403245,0.2057381421,-0.0143849896,-0.024044076,0.0058025098,0.1215947866,0.2089033723,0.0798321962,-0.0836231709,-0.1614168137,-0.1593151838,0.098150149,0.0621439032,0.0763312802,0.0264563933,-0.2656008005,0.0542387515,0.0102328975,0.2679121494,0.2297856659,0.0542744882,-0.2823029459,0.029405009,-0.0894653425,-0.0224468429,0.3617697656,0.1321124285,0.1846684366,0.3180979192,-0.1216696352,0.1428943425,-0.0601892583,-0.2058867961,-0.1090199202,0.2014214098,-0.4196126759,0.0011625303,0.0554270037,-0.12873137,-0.0488140583,-0.2332871109,-0.2695047557,-0.4261551201,-0.0986412242,0.0017243456,-0.1481575668,0.1272243708,-0.1493606865,0.1027415693,0.0276107434,-0.2686441839,-0.2308505625,-0.3650170863,-0.3169379532,-0.010910823,0.4290106595,0.014576084,0.1516437829,-0.3627824783,-0.041841913,-0.0665280297,-0.182035476,0.1500556469,0.0805360898,0.7029446959,0.1669664234,0.4550088048,0.0690727234,-0.4021815658,0.3672315776,0.0559912547,0.1430447549,0.2535715699,-0.0262755174,-0.233836621,0.0019271201,-0.326227963,-0.1746632308,-0.2757624686,-0.2295166552,0.3282880485,0.0036655893,0.5006689429,0.2981620729,0.1741961688,0.0721937418,0.293499738,-0.0413388088,-0.6339867711,0.3558411598,-0.1906136423,-0.2751670182,0.2133964598,-0.1321064532,0.2231930494,-0.0869277865,-0.5945403576,-0.0181757268,-0.0856766701,0.2774372101,-0.0212452579,0.174653545,0.0803521276,0.1810375452,0.0438931286,-0.16777201,0.0480528399,-0.1401127279,-0.1015582308,0.1961410195,0.1260029525,0.1382963806,-0.3169760108,0.6540639997,0.0909557492,0.1336974353,0.282076925,-0.2309667915,0.237269327,-0.2938368618,-0.366079241,0.142866239,-0.1669014841,-0.1178768277,0.2168298066,0.0458054617,-0.2475858182,-0.3451890051,0.0167480763,-0.0648541227,-0.3360706866,0.00500212,-0.0207578167,0.1135326698,0.1858856827,0.1822646409,-0.2703818381,-0.2345915437,0.1548732817,0.4231767952,0.0087489625,-0.1173151657,0.0347412154,-0.1464680284,-0.2632380128,0.0673621967,0.143509835,0.0301828831,0.0692543685,0.160138011,-0.0753667355,-0.043373514,0.4331784248,-0.2390737832,0.2342213988,0.2200020254,0.2694022655,-0.3107819855,-0.1034184247,-0.0251532011,0.0629202202,0.0118524972,0.1688247323,-0.0887500793,-0.0934559628,0.131155178,0.4144935906,-0.2235663086,-0.2698959708,-0.2487764359,-0.0687507242,-0.4285556972,-0.2023697495,-0.0320565999,0.4673660994,-0.0326748192,-0.0647634789,-0.2586669624,0.2108554989,-0.050994385,0.1131491736,0.3406184614,0.0321993232,0.1324106604,0.1004277468,0.1380321234,0.1158991605,0.534376204,0.2168563753,-0.2238845527,-0.2228387147,-0.0851792991,0.0313075818,0.1784212738,-0.2225720435,-0.4367471039,-0.152959764,0.0778846219,-0.2856153548,0.0102684777,0.1609386653,-0.0096937818,-0.2836462259,-0.5890882015,0.6666560173,-0.0377636254,-0.0160097722,0.377890408,0.2793262601,-0.468231082,0.291544944,0.0237020031,0.7756505013,0.1787301749,0.189420104,0.385335654,0.2011995018,0.0771394819,-0.3319946826,-0.0282286443,-0.2672969997,-0.1985023022,-0.0653544441,-0.0603929572,0.2244365662,0.0444639623,-0.3415676653,0.1563438028,-0.2356469035,0.1824591309,0.1031999588,0.1254787147,-0.2654340863,-0.2216459364,-0.3295260668,0.103697449,-0.0262821745,0.1340884715,-0.1326717138,0.0420874394,0.03195123,-0.2009951323,-0.276255846,0.2567481101,-0.1589357853,0.1236921176,0.0484220237,-0.1378914118,0.2281662375,0.5829567909,-0.1593587995,0.244093135,-0.3477098942,0.3507631421,-0.1915881932,-0.0595127791,-0.0050645182,0.118300125,0.3890743256,-0.0086640865,-0.1870107055,0.0776970685,-0.3867884576,-0.0793544203,-0.0003725718,-0.0602900982,0.2495678067,-0.3778285384,-0.3229056597,-0.1403657049,0.1496346146,-0.2204141617,0.1755172312,0.0848179311,-0.3581290841,0.0449242666,0.0820429102,-0.2780804634,0.061380256,0.1504953504,-0.1318150908,0.1769619435,0.6715326309,0.1515841633,-0.0835839659,-0.1752031296,0.3402463794,0.239140898,-0.3221225441,0.0764362887,0.1584532559,0.1720089465,-0.1837609559,0.5057669878,0.0111206537,-0.0361346528,0.1237044409,-0.611213088,-0.180048719,0.1154958606,0.0847896934,0.2263802886,-0.0776381344,-0.0945743844,0.1279538423,-0.0588704124,-0.2863120735,-0.058799658,-0.2254420966,0.1243875474,0.184392944,0.3527343273,-0.0772996247,-0.1034431532,0.1885668933,-0.2939728498,0.0775585771,-0.2647897005,-0.1814705282,0.1575973481,0.1562161744,-0.0526112169,0.1496885568,-0.1911634952,-0.1454014778,-0.3024875522,0.2301179171,-0.1878865212,-0.1893222928,0.1273608804,0.0039846329,0.2513929605,-0.2547197342,0.2307413518,-0.1212229207,0.3786002398,-0.069414407,0.0638326257,0.0754180849,0.0700532496,0.0636797994,0.1784889549,-0.4604460895,0.2478851378,0.4021214545,-0.5224174261,0.3102975786,0.2887887061,0.0420287549,0.4635989666,-0.4118162394,0.2062280476,0.2217870057,0.2036857456,-0.2438907027,-0.0228031687,0.3716575503,0.0349622183,-0.0560302511,-0.0056387219,-0.0307278559,-0.3672403395,-0.3781550229,0.0917167813,0.2237689197,-0.3046437204,0.3425470591,0.6575306654,0.0451318212,0.1759259701,-0.1087863594,0.2009283304,-0.1053792685,0.6286644936,-0.2232085168,0.0522453375,0.0775191337,0.0544747785,0.021548694,-0.3388665318,0.403706938,0.2471379489,-0.2143837512,0.0779276788,-0.052623488,0.0070789112,-0.1041948274,0.1115579233,-0.2287227958,-0.0595837533,-0.1917319596,-0.1866235137,-0.321981132,-0.2954736054,0.1318513751,0.2057689726,0.0225878377,-0.2227572054,-0.0292872693,0.1319862455,-0.1630318463,-0.4736092091,0.3834295273,0.1882407069,0.031339258,-0.2784054279,0.0510508195,0.1162586808,0.0456838869,0.1139037758,0.2092193663,0.6908439398,0.4375931025,-0.2205272019,0.1245531514,0.0927388892,-0.2373757958,-0.1092979833,0.0666442737,0.0102185681,0.169852078,0.327993542,0.1993826479,-0.1710319817,-0.1310653538,0.1197896376,0.1861197352,-0.3883250058,0.2811678052,-0.189319104,-0.1461223364,-0.1223476902,0.0990120694,-0.2563290596,0.0708758011,0.4077413082,0.2742966413,0.0725743026,-0.1249823272,0.0650696903,-0.1079402864,0.4987427592,0.3013652265,0.1351226717,-0.3123226762,-0.0683784559,-0.5570700169,0.0616605058,0.0642265305,-0.2956343293,0.1404156685,-0.0501559041,-0.131200552,0.1987145394,0.0020378774,-0.2134056091,0.2585369945,0.2531548738,-0.1139301062,-0.5248924494,0.1135269031,0.2256429642,-0.1041887999,-0.3573681712,-0.0946138576,-0.3640032113,0.0264801551,0.0779690817,0.0237678513,0.204964146,-0.183862105,0.3086179793,-0.0612024888,0.515040338,-0.0827326775,-0.0476548225,-0.3289958835,-0.285235256,-0.2313295454,0.2595256865,-0.0318625979,0.3192126155,-0.2210430503,0.0657729283,-0.2718800604,-0.0785162896,0.0239786264,0.3921006918,-0.3312179148,-0.0785401911,-0.0629318506,0.1922969073,0.2366017401,0.0647772327,-0.1289997995,0.2295152694,-0.0741891116,-0.3985043168,0.3420298994,-0.5386348367,-0.4959172904,-0.0198907666,-0.0243965182,-0.4006351829,-0.0844035372,-0.5008683205,0.1480123401,0.3623288572,-0.0370375477,-0.2316140682,-0.0645410791,0.0931225866,0.1666582376,-0.0898157209,0.1690057218,-0.0272076726,-0.1126220226,-0.110605292,-0.254230231]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2239","title":"Error loading wikihow dataset","comments":"Good call out. I did try that and that's when it told me to download the\ndataset. Don't believe I have tried it with local files. Will try first\nthing in the morning and get back to you.\n\nOn Mon, Apr 19, 2021, 11:17 PM Albert Villanova del Moral <\n***@***.***> wrote:\n\n> Hi @odellus , thanks for reporting.\n>\n> The wikihow dataset has 2 versions:\n>\n> - all: Consisting of the concatenation of all paragraphs as the\n> articles and the bold lines as the reference summaries.\n> - sep: Consisting of each paragraph and its summary.\n>\n> Therefore, in order to load it, you have to specify which version you\n> would like, for example:\n>\n> dataset = load_dataset('wikihow', 'all')\n>\n> Please, tell me if this solves your problem.\n>\n> \u2014\n> You are receiving this because you were mentioned.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","body":"## Describe the bug\r\n\r\nWhen attempting to load wikihow into a dataset with\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\nI get the message:\r\n```\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\nat the end of a [full stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2).\r\n\r\n## Steps to reproduce the bug\r\n\r\nI have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https:\/\/huggingface.co\/datasets\/wikihow) says to use \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow')\r\n```\r\nto load the dataset. I do so and I get the message\r\n```\r\nAssertionError: The dataset wikihow with config all requires manual data.\r\n Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https:\/\/github.com\/mahnazkoupaee\/WikiHow-Dataset.\r\n You need to download the following two files manually:\r\n 1) https:\/\/ucsb.app.box.com\/s\/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under \/wikihowAll.csv\r\n 2) https:\/\/ucsb.app.box.com\/s\/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under \/wikihowSep.csv\r\n\r\n The can e.g. be \"~\/manual_wikihow_data\".\r\n\r\n Wikihow can then be loaded using the following command `datasets.load_dataset(\"wikihow\", data_dir=\"\")`.\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='')\r\n```\r\n\r\nSo I create a directory `.\/wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory.\r\n\r\nThen I run \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\n\r\nthat's when I get the [stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2)\r\n\r\n## Expected results\r\nI expected it to load the downloaded files into a dataset.\r\n\r\n## Actual results\r\n```python\r\nUsing custom data configuration default-data_dir=.%2Fwikihow\r\nDownloading and preparing dataset wikihow\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/azureuser\/.cache\/huggingface\/datasets\/wikihow\/default-data_dir=.%2Fwikihow\/0.0.0\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... ---------------------------------------------------------------------------\r\nAttributeError\r\nTraceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset('wikihow',data_dir='.\/wikihow')\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n745 try_from_hf_gcs=try_from_hf_gcs,\r\n746 base_path=base_path,--> \r\n747 use_auth_token=use_auth_token,\r\n748 )\r\n749 \r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n577 if not downloaded_from_gcs:\r\n578 self._download_and_prepare( -->\r\n579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs \r\n580 ) \r\n581 # Sync info\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n632 split_dict = SplitDict(dataset_name=self.name)\r\n633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) -->\r\n634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) \r\n635 \r\n636 # Checksums verification\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikihow\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2\/wikihow.py in _split_generators(self, dl_manager)\r\n132\r\n133 path_to_manual_file = os.path.join(\r\n--> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename \r\n135 ) \r\n136\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n```\r\n- Datasets: 1.5.0\r\n- Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0]\r\n- Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic\r\n```","comment_length":168,"text":"Error loading wikihow dataset\n## Describe the bug\r\n\r\nWhen attempting to load wikihow into a dataset with\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\nI get the message:\r\n```\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\nat the end of a [full stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2).\r\n\r\n## Steps to reproduce the bug\r\n\r\nI have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https:\/\/huggingface.co\/datasets\/wikihow) says to use \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow')\r\n```\r\nto load the dataset. I do so and I get the message\r\n```\r\nAssertionError: The dataset wikihow with config all requires manual data.\r\n Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https:\/\/github.com\/mahnazkoupaee\/WikiHow-Dataset.\r\n You need to download the following two files manually:\r\n 1) https:\/\/ucsb.app.box.com\/s\/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under \/wikihowAll.csv\r\n 2) https:\/\/ucsb.app.box.com\/s\/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under \/wikihowSep.csv\r\n\r\n The can e.g. be \"~\/manual_wikihow_data\".\r\n\r\n Wikihow can then be loaded using the following command `datasets.load_dataset(\"wikihow\", data_dir=\"\")`.\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='')\r\n```\r\n\r\nSo I create a directory `.\/wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory.\r\n\r\nThen I run \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\n\r\nthat's when I get the [stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2)\r\n\r\n## Expected results\r\nI expected it to load the downloaded files into a dataset.\r\n\r\n## Actual results\r\n```python\r\nUsing custom data configuration default-data_dir=.%2Fwikihow\r\nDownloading and preparing dataset wikihow\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/azureuser\/.cache\/huggingface\/datasets\/wikihow\/default-data_dir=.%2Fwikihow\/0.0.0\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... ---------------------------------------------------------------------------\r\nAttributeError\r\nTraceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset('wikihow',data_dir='.\/wikihow')\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n745 try_from_hf_gcs=try_from_hf_gcs,\r\n746 base_path=base_path,--> \r\n747 use_auth_token=use_auth_token,\r\n748 )\r\n749 \r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n577 if not downloaded_from_gcs:\r\n578 self._download_and_prepare( -->\r\n579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs \r\n580 ) \r\n581 # Sync info\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n632 split_dict = SplitDict(dataset_name=self.name)\r\n633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) -->\r\n634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) \r\n635 \r\n636 # Checksums verification\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikihow\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2\/wikihow.py in _split_generators(self, dl_manager)\r\n132\r\n133 path_to_manual_file = os.path.join(\r\n--> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename \r\n135 ) \r\n136\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n```\r\n- Datasets: 1.5.0\r\n- Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0]\r\n- Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic\r\n```\nGood call out. I did try that and that's when it told me to download the\ndataset. Don't believe I have tried it with local files. Will try first\nthing in the morning and get back to you.\n\nOn Mon, Apr 19, 2021, 11:17 PM Albert Villanova del Moral <\n***@***.***> wrote:\n\n> Hi @odellus , thanks for reporting.\n>\n> The wikihow dataset has 2 versions:\n>\n> - all: Consisting of the concatenation of all paragraphs as the\n> articles and the bold lines as the reference summaries.\n> - sep: Consisting of each paragraph and its summary.\n>\n> Therefore, in order to load it, you have to specify which version you\n> would like, for example:\n>\n> dataset = load_dataset('wikihow', 'all')\n>\n> Please, tell me if this solves your problem.\n>\n> \u2014\n> You are receiving this because you were mentioned.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","embeddings":[-0.2274995297,0.370223105,0.0238096733,0.3987950683,0.2537099421,0.2747787535,0.4276534021,0.4391774237,0.2458434254,0.0935082212,0.216285497,0.3853774071,-0.0128944451,0.1863868237,0.1205022633,-0.2686142325,0.0110668847,0.1045249701,0.223709926,0.1890229732,-0.3448888063,0.0065287249,-0.2329775989,0.1040147245,-0.3400695622,0.0130667957,-0.0981442481,0.1813103855,-0.0533038154,-0.541544497,0.4757557511,-0.1185788065,0.2518153787,0.4039177299,-0.0001126044,0.1877029687,0.5060688257,0.0342030711,-0.442166388,-0.0944597572,-0.3161360323,-0.1228394881,0.2420690507,-0.1337201148,-0.0326552615,-0.0258759465,0.004085809,0.0358592607,0.2355376929,0.2652969062,0.1653248966,-0.1196185276,0.3745595217,-0.1360630244,0.1942486763,-0.0933305025,-0.0869166702,0.1622466892,0.0739811137,-0.2382172346,0.1115441099,0.20557262,-0.1832122952,0.0634829551,0.5792307854,-0.0081012985,0.2564974129,-0.332991004,0.2632296085,0.3839665055,0.8194904923,-0.2249974608,-0.2384388745,-0.2182331532,0.1727716476,0.1644721776,0.3875032365,0.3061026633,-0.3748465776,0.0360186175,-0.0377139971,-0.2557303309,-0.0580877513,0.2808242142,-0.294668287,0.4108183086,-0.0778242275,0.1745568216,-0.2258381695,-0.0642693043,0.2035174668,-0.32481125,-0.2376295626,0.3015760481,-0.1384118348,0.0768692344,0.1585353613,0.1928257942,0.0994419307,-0.0371334255,-0.100838013,-0.1908228844,0.1564052403,0.1384836882,0.4442620873,0.2339715958,-0.0133317849,0.2062765062,0.1207495183,0.3048851788,-0.1984937936,-0.1141341776,-0.201289773,-0.2741613984,0.1328350455,-0.2296912819,0.3544214666,-0.0338293388,-0.2546579242,0.2001613677,0.0116372267,0.0730370209,0.0347945094,0.4159566164,-0.2064811736,0.1446116865,0.2597591877,0.2927667499,0.1025156602,0.1234883666,-0.1005089432,0.0907599106,-0.0558772162,-0.0318382233,0.1500617266,-0.1930603236,0.2553026974,0.1958255172,0.125093013,-0.2846936584,-0.0093548056,0.2284626216,-0.2356490493,0.3494037092,0.0726296008,0.1431750059,0.1701949537,-0.6475993991,-0.1152566895,0.1633387059,-0.394199729,-0.3855848312,-0.2756711245,0.2184655666,-0.0403429158,0.0450250953,-0.419618398,-0.1785893738,0.2975123823,-0.2910596132,-0.2390632927,-0.0699154362,0.1259678304,-0.1256434321,0.356915772,0.4782387316,-0.3908683658,-0.0335086882,-0.0930062383,0.0438776985,-0.0572247431,-0.3101461828,-0.3554069996,0.4971389174,-0.288457185,-0.2874653041,0.2937496006,-0.4576177597,-0.2892730534,0.2900746465,0.0192084331,0.3332006037,-0.0229105316,-0.1692107767,-0.168219924,-0.0836994424,0.2327568531,0.2221877873,-0.0055906828,0.0550968647,-0.1940852553,-0.2364349663,0.1046095416,0.186486274,0.0511644967,0.1748686433,0.0437120758,0.0242446903,0.1256706417,-0.1626351625,-0.0451247096,0.1203386486,-0.0281695016,-0.0047448212,0.048327703,0.0023430884,-0.4610753655,0.3484283388,0.0433676019,0.1083413661,-0.1505929083,-0.0570079163,-0.2748941183,0.1563291699,-0.5284996033,-0.1772395521,0.16646716,0.0618430004,-0.149765268,-0.018513212,-0.124017477,0.0431065783,-0.3334240615,0.1222898439,-0.2070953697,0.1706795096,-0.1306974888,-0.0565901399,-0.0111799212,0.1489347219,0.0346692316,0.0848748386,-0.2042343467,0.3257358968,0.0568369739,0.1124035195,-0.0444702618,-0.0688852519,-0.037389446,-0.1064234599,0.0548359565,0.2269100845,0.2477425188,-0.0861583799,-0.092016995,0.230902642,0.0540100299,0.3870119154,0.0292328987,0.267403245,0.2057381421,-0.0143849896,-0.024044076,0.0058025098,0.1215947866,0.2089033723,0.0798321962,-0.0836231709,-0.1614168137,-0.1593151838,0.098150149,0.0621439032,0.0763312802,0.0264563933,-0.2656008005,0.0542387515,0.0102328975,0.2679121494,0.2297856659,0.0542744882,-0.2823029459,0.029405009,-0.0894653425,-0.0224468429,0.3617697656,0.1321124285,0.1846684366,0.3180979192,-0.1216696352,0.1428943425,-0.0601892583,-0.2058867961,-0.1090199202,0.2014214098,-0.4196126759,0.0011625303,0.0554270037,-0.12873137,-0.0488140583,-0.2332871109,-0.2695047557,-0.4261551201,-0.0986412242,0.0017243456,-0.1481575668,0.1272243708,-0.1493606865,0.1027415693,0.0276107434,-0.2686441839,-0.2308505625,-0.3650170863,-0.3169379532,-0.010910823,0.4290106595,0.014576084,0.1516437829,-0.3627824783,-0.041841913,-0.0665280297,-0.182035476,0.1500556469,0.0805360898,0.7029446959,0.1669664234,0.4550088048,0.0690727234,-0.4021815658,0.3672315776,0.0559912547,0.1430447549,0.2535715699,-0.0262755174,-0.233836621,0.0019271201,-0.326227963,-0.1746632308,-0.2757624686,-0.2295166552,0.3282880485,0.0036655893,0.5006689429,0.2981620729,0.1741961688,0.0721937418,0.293499738,-0.0413388088,-0.6339867711,0.3558411598,-0.1906136423,-0.2751670182,0.2133964598,-0.1321064532,0.2231930494,-0.0869277865,-0.5945403576,-0.0181757268,-0.0856766701,0.2774372101,-0.0212452579,0.174653545,0.0803521276,0.1810375452,0.0438931286,-0.16777201,0.0480528399,-0.1401127279,-0.1015582308,0.1961410195,0.1260029525,0.1382963806,-0.3169760108,0.6540639997,0.0909557492,0.1336974353,0.282076925,-0.2309667915,0.237269327,-0.2938368618,-0.366079241,0.142866239,-0.1669014841,-0.1178768277,0.2168298066,0.0458054617,-0.2475858182,-0.3451890051,0.0167480763,-0.0648541227,-0.3360706866,0.00500212,-0.0207578167,0.1135326698,0.1858856827,0.1822646409,-0.2703818381,-0.2345915437,0.1548732817,0.4231767952,0.0087489625,-0.1173151657,0.0347412154,-0.1464680284,-0.2632380128,0.0673621967,0.143509835,0.0301828831,0.0692543685,0.160138011,-0.0753667355,-0.043373514,0.4331784248,-0.2390737832,0.2342213988,0.2200020254,0.2694022655,-0.3107819855,-0.1034184247,-0.0251532011,0.0629202202,0.0118524972,0.1688247323,-0.0887500793,-0.0934559628,0.131155178,0.4144935906,-0.2235663086,-0.2698959708,-0.2487764359,-0.0687507242,-0.4285556972,-0.2023697495,-0.0320565999,0.4673660994,-0.0326748192,-0.0647634789,-0.2586669624,0.2108554989,-0.050994385,0.1131491736,0.3406184614,0.0321993232,0.1324106604,0.1004277468,0.1380321234,0.1158991605,0.534376204,0.2168563753,-0.2238845527,-0.2228387147,-0.0851792991,0.0313075818,0.1784212738,-0.2225720435,-0.4367471039,-0.152959764,0.0778846219,-0.2856153548,0.0102684777,0.1609386653,-0.0096937818,-0.2836462259,-0.5890882015,0.6666560173,-0.0377636254,-0.0160097722,0.377890408,0.2793262601,-0.468231082,0.291544944,0.0237020031,0.7756505013,0.1787301749,0.189420104,0.385335654,0.2011995018,0.0771394819,-0.3319946826,-0.0282286443,-0.2672969997,-0.1985023022,-0.0653544441,-0.0603929572,0.2244365662,0.0444639623,-0.3415676653,0.1563438028,-0.2356469035,0.1824591309,0.1031999588,0.1254787147,-0.2654340863,-0.2216459364,-0.3295260668,0.103697449,-0.0262821745,0.1340884715,-0.1326717138,0.0420874394,0.03195123,-0.2009951323,-0.276255846,0.2567481101,-0.1589357853,0.1236921176,0.0484220237,-0.1378914118,0.2281662375,0.5829567909,-0.1593587995,0.244093135,-0.3477098942,0.3507631421,-0.1915881932,-0.0595127791,-0.0050645182,0.118300125,0.3890743256,-0.0086640865,-0.1870107055,0.0776970685,-0.3867884576,-0.0793544203,-0.0003725718,-0.0602900982,0.2495678067,-0.3778285384,-0.3229056597,-0.1403657049,0.1496346146,-0.2204141617,0.1755172312,0.0848179311,-0.3581290841,0.0449242666,0.0820429102,-0.2780804634,0.061380256,0.1504953504,-0.1318150908,0.1769619435,0.6715326309,0.1515841633,-0.0835839659,-0.1752031296,0.3402463794,0.239140898,-0.3221225441,0.0764362887,0.1584532559,0.1720089465,-0.1837609559,0.5057669878,0.0111206537,-0.0361346528,0.1237044409,-0.611213088,-0.180048719,0.1154958606,0.0847896934,0.2263802886,-0.0776381344,-0.0945743844,0.1279538423,-0.0588704124,-0.2863120735,-0.058799658,-0.2254420966,0.1243875474,0.184392944,0.3527343273,-0.0772996247,-0.1034431532,0.1885668933,-0.2939728498,0.0775585771,-0.2647897005,-0.1814705282,0.1575973481,0.1562161744,-0.0526112169,0.1496885568,-0.1911634952,-0.1454014778,-0.3024875522,0.2301179171,-0.1878865212,-0.1893222928,0.1273608804,0.0039846329,0.2513929605,-0.2547197342,0.2307413518,-0.1212229207,0.3786002398,-0.069414407,0.0638326257,0.0754180849,0.0700532496,0.0636797994,0.1784889549,-0.4604460895,0.2478851378,0.4021214545,-0.5224174261,0.3102975786,0.2887887061,0.0420287549,0.4635989666,-0.4118162394,0.2062280476,0.2217870057,0.2036857456,-0.2438907027,-0.0228031687,0.3716575503,0.0349622183,-0.0560302511,-0.0056387219,-0.0307278559,-0.3672403395,-0.3781550229,0.0917167813,0.2237689197,-0.3046437204,0.3425470591,0.6575306654,0.0451318212,0.1759259701,-0.1087863594,0.2009283304,-0.1053792685,0.6286644936,-0.2232085168,0.0522453375,0.0775191337,0.0544747785,0.021548694,-0.3388665318,0.403706938,0.2471379489,-0.2143837512,0.0779276788,-0.052623488,0.0070789112,-0.1041948274,0.1115579233,-0.2287227958,-0.0595837533,-0.1917319596,-0.1866235137,-0.321981132,-0.2954736054,0.1318513751,0.2057689726,0.0225878377,-0.2227572054,-0.0292872693,0.1319862455,-0.1630318463,-0.4736092091,0.3834295273,0.1882407069,0.031339258,-0.2784054279,0.0510508195,0.1162586808,0.0456838869,0.1139037758,0.2092193663,0.6908439398,0.4375931025,-0.2205272019,0.1245531514,0.0927388892,-0.2373757958,-0.1092979833,0.0666442737,0.0102185681,0.169852078,0.327993542,0.1993826479,-0.1710319817,-0.1310653538,0.1197896376,0.1861197352,-0.3883250058,0.2811678052,-0.189319104,-0.1461223364,-0.1223476902,0.0990120694,-0.2563290596,0.0708758011,0.4077413082,0.2742966413,0.0725743026,-0.1249823272,0.0650696903,-0.1079402864,0.4987427592,0.3013652265,0.1351226717,-0.3123226762,-0.0683784559,-0.5570700169,0.0616605058,0.0642265305,-0.2956343293,0.1404156685,-0.0501559041,-0.131200552,0.1987145394,0.0020378774,-0.2134056091,0.2585369945,0.2531548738,-0.1139301062,-0.5248924494,0.1135269031,0.2256429642,-0.1041887999,-0.3573681712,-0.0946138576,-0.3640032113,0.0264801551,0.0779690817,0.0237678513,0.204964146,-0.183862105,0.3086179793,-0.0612024888,0.515040338,-0.0827326775,-0.0476548225,-0.3289958835,-0.285235256,-0.2313295454,0.2595256865,-0.0318625979,0.3192126155,-0.2210430503,0.0657729283,-0.2718800604,-0.0785162896,0.0239786264,0.3921006918,-0.3312179148,-0.0785401911,-0.0629318506,0.1922969073,0.2366017401,0.0647772327,-0.1289997995,0.2295152694,-0.0741891116,-0.3985043168,0.3420298994,-0.5386348367,-0.4959172904,-0.0198907666,-0.0243965182,-0.4006351829,-0.0844035372,-0.5008683205,0.1480123401,0.3623288572,-0.0370375477,-0.2316140682,-0.0645410791,0.0931225866,0.1666582376,-0.0898157209,0.1690057218,-0.0272076726,-0.1126220226,-0.110605292,-0.254230231]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2239","title":"Error loading wikihow dataset","comments":"Hi @odellus, yes you are right.\r\n\r\nDue to the server where the `wikihow` dataset is hosted, the dataset can't be downloaded automatically by `huggingface` and you have to download it manually as you did.\r\n\r\nNevertheless, you have to specify which dataset version you would like to load anyway:\r\n```python\r\ndataset = load_dataset('wikihow', 'all', data_dir='.\/wikihow')\r\n```\r\nor\r\n```python\r\ndataset = load_dataset('wikihow', 'sep', data_dir='.\/wikihow')\r\n```\r\nI find that the instructions given by `huggingface` are not clear enough: I am going to fix this.\r\nPlease tell me if this eventually works for you.","body":"## Describe the bug\r\n\r\nWhen attempting to load wikihow into a dataset with\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\nI get the message:\r\n```\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\nat the end of a [full stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2).\r\n\r\n## Steps to reproduce the bug\r\n\r\nI have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https:\/\/huggingface.co\/datasets\/wikihow) says to use \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow')\r\n```\r\nto load the dataset. I do so and I get the message\r\n```\r\nAssertionError: The dataset wikihow with config all requires manual data.\r\n Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https:\/\/github.com\/mahnazkoupaee\/WikiHow-Dataset.\r\n You need to download the following two files manually:\r\n 1) https:\/\/ucsb.app.box.com\/s\/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under \/wikihowAll.csv\r\n 2) https:\/\/ucsb.app.box.com\/s\/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under \/wikihowSep.csv\r\n\r\n The can e.g. be \"~\/manual_wikihow_data\".\r\n\r\n Wikihow can then be loaded using the following command `datasets.load_dataset(\"wikihow\", data_dir=\"\")`.\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='')\r\n```\r\n\r\nSo I create a directory `.\/wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory.\r\n\r\nThen I run \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\n\r\nthat's when I get the [stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2)\r\n\r\n## Expected results\r\nI expected it to load the downloaded files into a dataset.\r\n\r\n## Actual results\r\n```python\r\nUsing custom data configuration default-data_dir=.%2Fwikihow\r\nDownloading and preparing dataset wikihow\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/azureuser\/.cache\/huggingface\/datasets\/wikihow\/default-data_dir=.%2Fwikihow\/0.0.0\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... ---------------------------------------------------------------------------\r\nAttributeError\r\nTraceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset('wikihow',data_dir='.\/wikihow')\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n745 try_from_hf_gcs=try_from_hf_gcs,\r\n746 base_path=base_path,--> \r\n747 use_auth_token=use_auth_token,\r\n748 )\r\n749 \r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n577 if not downloaded_from_gcs:\r\n578 self._download_and_prepare( -->\r\n579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs \r\n580 ) \r\n581 # Sync info\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n632 split_dict = SplitDict(dataset_name=self.name)\r\n633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) -->\r\n634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) \r\n635 \r\n636 # Checksums verification\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikihow\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2\/wikihow.py in _split_generators(self, dl_manager)\r\n132\r\n133 path_to_manual_file = os.path.join(\r\n--> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename \r\n135 ) \r\n136\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n```\r\n- Datasets: 1.5.0\r\n- Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0]\r\n- Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic\r\n```","comment_length":90,"text":"Error loading wikihow dataset\n## Describe the bug\r\n\r\nWhen attempting to load wikihow into a dataset with\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\nI get the message:\r\n```\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\nat the end of a [full stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2).\r\n\r\n## Steps to reproduce the bug\r\n\r\nI have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https:\/\/huggingface.co\/datasets\/wikihow) says to use \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow')\r\n```\r\nto load the dataset. I do so and I get the message\r\n```\r\nAssertionError: The dataset wikihow with config all requires manual data.\r\n Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https:\/\/github.com\/mahnazkoupaee\/WikiHow-Dataset.\r\n You need to download the following two files manually:\r\n 1) https:\/\/ucsb.app.box.com\/s\/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under \/wikihowAll.csv\r\n 2) https:\/\/ucsb.app.box.com\/s\/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under \/wikihowSep.csv\r\n\r\n The can e.g. be \"~\/manual_wikihow_data\".\r\n\r\n Wikihow can then be loaded using the following command `datasets.load_dataset(\"wikihow\", data_dir=\"\")`.\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='')\r\n```\r\n\r\nSo I create a directory `.\/wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory.\r\n\r\nThen I run \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\n\r\nthat's when I get the [stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2)\r\n\r\n## Expected results\r\nI expected it to load the downloaded files into a dataset.\r\n\r\n## Actual results\r\n```python\r\nUsing custom data configuration default-data_dir=.%2Fwikihow\r\nDownloading and preparing dataset wikihow\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/azureuser\/.cache\/huggingface\/datasets\/wikihow\/default-data_dir=.%2Fwikihow\/0.0.0\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... ---------------------------------------------------------------------------\r\nAttributeError\r\nTraceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset('wikihow',data_dir='.\/wikihow')\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n745 try_from_hf_gcs=try_from_hf_gcs,\r\n746 base_path=base_path,--> \r\n747 use_auth_token=use_auth_token,\r\n748 )\r\n749 \r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n577 if not downloaded_from_gcs:\r\n578 self._download_and_prepare( -->\r\n579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs \r\n580 ) \r\n581 # Sync info\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n632 split_dict = SplitDict(dataset_name=self.name)\r\n633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) -->\r\n634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) \r\n635 \r\n636 # Checksums verification\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikihow\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2\/wikihow.py in _split_generators(self, dl_manager)\r\n132\r\n133 path_to_manual_file = os.path.join(\r\n--> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename \r\n135 ) \r\n136\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n```\r\n- Datasets: 1.5.0\r\n- Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0]\r\n- Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic\r\n```\nHi @odellus, yes you are right.\r\n\r\nDue to the server where the `wikihow` dataset is hosted, the dataset can't be downloaded automatically by `huggingface` and you have to download it manually as you did.\r\n\r\nNevertheless, you have to specify which dataset version you would like to load anyway:\r\n```python\r\ndataset = load_dataset('wikihow', 'all', data_dir='.\/wikihow')\r\n```\r\nor\r\n```python\r\ndataset = load_dataset('wikihow', 'sep', data_dir='.\/wikihow')\r\n```\r\nI find that the instructions given by `huggingface` are not clear enough: I am going to fix this.\r\nPlease tell me if this eventually works for you.","embeddings":[-0.2274995297,0.370223105,0.0238096733,0.3987950683,0.2537099421,0.2747787535,0.4276534021,0.4391774237,0.2458434254,0.0935082212,0.216285497,0.3853774071,-0.0128944451,0.1863868237,0.1205022633,-0.2686142325,0.0110668847,0.1045249701,0.223709926,0.1890229732,-0.3448888063,0.0065287249,-0.2329775989,0.1040147245,-0.3400695622,0.0130667957,-0.0981442481,0.1813103855,-0.0533038154,-0.541544497,0.4757557511,-0.1185788065,0.2518153787,0.4039177299,-0.0001126044,0.1877029687,0.5060688257,0.0342030711,-0.442166388,-0.0944597572,-0.3161360323,-0.1228394881,0.2420690507,-0.1337201148,-0.0326552615,-0.0258759465,0.004085809,0.0358592607,0.2355376929,0.2652969062,0.1653248966,-0.1196185276,0.3745595217,-0.1360630244,0.1942486763,-0.0933305025,-0.0869166702,0.1622466892,0.0739811137,-0.2382172346,0.1115441099,0.20557262,-0.1832122952,0.0634829551,0.5792307854,-0.0081012985,0.2564974129,-0.332991004,0.2632296085,0.3839665055,0.8194904923,-0.2249974608,-0.2384388745,-0.2182331532,0.1727716476,0.1644721776,0.3875032365,0.3061026633,-0.3748465776,0.0360186175,-0.0377139971,-0.2557303309,-0.0580877513,0.2808242142,-0.294668287,0.4108183086,-0.0778242275,0.1745568216,-0.2258381695,-0.0642693043,0.2035174668,-0.32481125,-0.2376295626,0.3015760481,-0.1384118348,0.0768692344,0.1585353613,0.1928257942,0.0994419307,-0.0371334255,-0.100838013,-0.1908228844,0.1564052403,0.1384836882,0.4442620873,0.2339715958,-0.0133317849,0.2062765062,0.1207495183,0.3048851788,-0.1984937936,-0.1141341776,-0.201289773,-0.2741613984,0.1328350455,-0.2296912819,0.3544214666,-0.0338293388,-0.2546579242,0.2001613677,0.0116372267,0.0730370209,0.0347945094,0.4159566164,-0.2064811736,0.1446116865,0.2597591877,0.2927667499,0.1025156602,0.1234883666,-0.1005089432,0.0907599106,-0.0558772162,-0.0318382233,0.1500617266,-0.1930603236,0.2553026974,0.1958255172,0.125093013,-0.2846936584,-0.0093548056,0.2284626216,-0.2356490493,0.3494037092,0.0726296008,0.1431750059,0.1701949537,-0.6475993991,-0.1152566895,0.1633387059,-0.394199729,-0.3855848312,-0.2756711245,0.2184655666,-0.0403429158,0.0450250953,-0.419618398,-0.1785893738,0.2975123823,-0.2910596132,-0.2390632927,-0.0699154362,0.1259678304,-0.1256434321,0.356915772,0.4782387316,-0.3908683658,-0.0335086882,-0.0930062383,0.0438776985,-0.0572247431,-0.3101461828,-0.3554069996,0.4971389174,-0.288457185,-0.2874653041,0.2937496006,-0.4576177597,-0.2892730534,0.2900746465,0.0192084331,0.3332006037,-0.0229105316,-0.1692107767,-0.168219924,-0.0836994424,0.2327568531,0.2221877873,-0.0055906828,0.0550968647,-0.1940852553,-0.2364349663,0.1046095416,0.186486274,0.0511644967,0.1748686433,0.0437120758,0.0242446903,0.1256706417,-0.1626351625,-0.0451247096,0.1203386486,-0.0281695016,-0.0047448212,0.048327703,0.0023430884,-0.4610753655,0.3484283388,0.0433676019,0.1083413661,-0.1505929083,-0.0570079163,-0.2748941183,0.1563291699,-0.5284996033,-0.1772395521,0.16646716,0.0618430004,-0.149765268,-0.018513212,-0.124017477,0.0431065783,-0.3334240615,0.1222898439,-0.2070953697,0.1706795096,-0.1306974888,-0.0565901399,-0.0111799212,0.1489347219,0.0346692316,0.0848748386,-0.2042343467,0.3257358968,0.0568369739,0.1124035195,-0.0444702618,-0.0688852519,-0.037389446,-0.1064234599,0.0548359565,0.2269100845,0.2477425188,-0.0861583799,-0.092016995,0.230902642,0.0540100299,0.3870119154,0.0292328987,0.267403245,0.2057381421,-0.0143849896,-0.024044076,0.0058025098,0.1215947866,0.2089033723,0.0798321962,-0.0836231709,-0.1614168137,-0.1593151838,0.098150149,0.0621439032,0.0763312802,0.0264563933,-0.2656008005,0.0542387515,0.0102328975,0.2679121494,0.2297856659,0.0542744882,-0.2823029459,0.029405009,-0.0894653425,-0.0224468429,0.3617697656,0.1321124285,0.1846684366,0.3180979192,-0.1216696352,0.1428943425,-0.0601892583,-0.2058867961,-0.1090199202,0.2014214098,-0.4196126759,0.0011625303,0.0554270037,-0.12873137,-0.0488140583,-0.2332871109,-0.2695047557,-0.4261551201,-0.0986412242,0.0017243456,-0.1481575668,0.1272243708,-0.1493606865,0.1027415693,0.0276107434,-0.2686441839,-0.2308505625,-0.3650170863,-0.3169379532,-0.010910823,0.4290106595,0.014576084,0.1516437829,-0.3627824783,-0.041841913,-0.0665280297,-0.182035476,0.1500556469,0.0805360898,0.7029446959,0.1669664234,0.4550088048,0.0690727234,-0.4021815658,0.3672315776,0.0559912547,0.1430447549,0.2535715699,-0.0262755174,-0.233836621,0.0019271201,-0.326227963,-0.1746632308,-0.2757624686,-0.2295166552,0.3282880485,0.0036655893,0.5006689429,0.2981620729,0.1741961688,0.0721937418,0.293499738,-0.0413388088,-0.6339867711,0.3558411598,-0.1906136423,-0.2751670182,0.2133964598,-0.1321064532,0.2231930494,-0.0869277865,-0.5945403576,-0.0181757268,-0.0856766701,0.2774372101,-0.0212452579,0.174653545,0.0803521276,0.1810375452,0.0438931286,-0.16777201,0.0480528399,-0.1401127279,-0.1015582308,0.1961410195,0.1260029525,0.1382963806,-0.3169760108,0.6540639997,0.0909557492,0.1336974353,0.282076925,-0.2309667915,0.237269327,-0.2938368618,-0.366079241,0.142866239,-0.1669014841,-0.1178768277,0.2168298066,0.0458054617,-0.2475858182,-0.3451890051,0.0167480763,-0.0648541227,-0.3360706866,0.00500212,-0.0207578167,0.1135326698,0.1858856827,0.1822646409,-0.2703818381,-0.2345915437,0.1548732817,0.4231767952,0.0087489625,-0.1173151657,0.0347412154,-0.1464680284,-0.2632380128,0.0673621967,0.143509835,0.0301828831,0.0692543685,0.160138011,-0.0753667355,-0.043373514,0.4331784248,-0.2390737832,0.2342213988,0.2200020254,0.2694022655,-0.3107819855,-0.1034184247,-0.0251532011,0.0629202202,0.0118524972,0.1688247323,-0.0887500793,-0.0934559628,0.131155178,0.4144935906,-0.2235663086,-0.2698959708,-0.2487764359,-0.0687507242,-0.4285556972,-0.2023697495,-0.0320565999,0.4673660994,-0.0326748192,-0.0647634789,-0.2586669624,0.2108554989,-0.050994385,0.1131491736,0.3406184614,0.0321993232,0.1324106604,0.1004277468,0.1380321234,0.1158991605,0.534376204,0.2168563753,-0.2238845527,-0.2228387147,-0.0851792991,0.0313075818,0.1784212738,-0.2225720435,-0.4367471039,-0.152959764,0.0778846219,-0.2856153548,0.0102684777,0.1609386653,-0.0096937818,-0.2836462259,-0.5890882015,0.6666560173,-0.0377636254,-0.0160097722,0.377890408,0.2793262601,-0.468231082,0.291544944,0.0237020031,0.7756505013,0.1787301749,0.189420104,0.385335654,0.2011995018,0.0771394819,-0.3319946826,-0.0282286443,-0.2672969997,-0.1985023022,-0.0653544441,-0.0603929572,0.2244365662,0.0444639623,-0.3415676653,0.1563438028,-0.2356469035,0.1824591309,0.1031999588,0.1254787147,-0.2654340863,-0.2216459364,-0.3295260668,0.103697449,-0.0262821745,0.1340884715,-0.1326717138,0.0420874394,0.03195123,-0.2009951323,-0.276255846,0.2567481101,-0.1589357853,0.1236921176,0.0484220237,-0.1378914118,0.2281662375,0.5829567909,-0.1593587995,0.244093135,-0.3477098942,0.3507631421,-0.1915881932,-0.0595127791,-0.0050645182,0.118300125,0.3890743256,-0.0086640865,-0.1870107055,0.0776970685,-0.3867884576,-0.0793544203,-0.0003725718,-0.0602900982,0.2495678067,-0.3778285384,-0.3229056597,-0.1403657049,0.1496346146,-0.2204141617,0.1755172312,0.0848179311,-0.3581290841,0.0449242666,0.0820429102,-0.2780804634,0.061380256,0.1504953504,-0.1318150908,0.1769619435,0.6715326309,0.1515841633,-0.0835839659,-0.1752031296,0.3402463794,0.239140898,-0.3221225441,0.0764362887,0.1584532559,0.1720089465,-0.1837609559,0.5057669878,0.0111206537,-0.0361346528,0.1237044409,-0.611213088,-0.180048719,0.1154958606,0.0847896934,0.2263802886,-0.0776381344,-0.0945743844,0.1279538423,-0.0588704124,-0.2863120735,-0.058799658,-0.2254420966,0.1243875474,0.184392944,0.3527343273,-0.0772996247,-0.1034431532,0.1885668933,-0.2939728498,0.0775585771,-0.2647897005,-0.1814705282,0.1575973481,0.1562161744,-0.0526112169,0.1496885568,-0.1911634952,-0.1454014778,-0.3024875522,0.2301179171,-0.1878865212,-0.1893222928,0.1273608804,0.0039846329,0.2513929605,-0.2547197342,0.2307413518,-0.1212229207,0.3786002398,-0.069414407,0.0638326257,0.0754180849,0.0700532496,0.0636797994,0.1784889549,-0.4604460895,0.2478851378,0.4021214545,-0.5224174261,0.3102975786,0.2887887061,0.0420287549,0.4635989666,-0.4118162394,0.2062280476,0.2217870057,0.2036857456,-0.2438907027,-0.0228031687,0.3716575503,0.0349622183,-0.0560302511,-0.0056387219,-0.0307278559,-0.3672403395,-0.3781550229,0.0917167813,0.2237689197,-0.3046437204,0.3425470591,0.6575306654,0.0451318212,0.1759259701,-0.1087863594,0.2009283304,-0.1053792685,0.6286644936,-0.2232085168,0.0522453375,0.0775191337,0.0544747785,0.021548694,-0.3388665318,0.403706938,0.2471379489,-0.2143837512,0.0779276788,-0.052623488,0.0070789112,-0.1041948274,0.1115579233,-0.2287227958,-0.0595837533,-0.1917319596,-0.1866235137,-0.321981132,-0.2954736054,0.1318513751,0.2057689726,0.0225878377,-0.2227572054,-0.0292872693,0.1319862455,-0.1630318463,-0.4736092091,0.3834295273,0.1882407069,0.031339258,-0.2784054279,0.0510508195,0.1162586808,0.0456838869,0.1139037758,0.2092193663,0.6908439398,0.4375931025,-0.2205272019,0.1245531514,0.0927388892,-0.2373757958,-0.1092979833,0.0666442737,0.0102185681,0.169852078,0.327993542,0.1993826479,-0.1710319817,-0.1310653538,0.1197896376,0.1861197352,-0.3883250058,0.2811678052,-0.189319104,-0.1461223364,-0.1223476902,0.0990120694,-0.2563290596,0.0708758011,0.4077413082,0.2742966413,0.0725743026,-0.1249823272,0.0650696903,-0.1079402864,0.4987427592,0.3013652265,0.1351226717,-0.3123226762,-0.0683784559,-0.5570700169,0.0616605058,0.0642265305,-0.2956343293,0.1404156685,-0.0501559041,-0.131200552,0.1987145394,0.0020378774,-0.2134056091,0.2585369945,0.2531548738,-0.1139301062,-0.5248924494,0.1135269031,0.2256429642,-0.1041887999,-0.3573681712,-0.0946138576,-0.3640032113,0.0264801551,0.0779690817,0.0237678513,0.204964146,-0.183862105,0.3086179793,-0.0612024888,0.515040338,-0.0827326775,-0.0476548225,-0.3289958835,-0.285235256,-0.2313295454,0.2595256865,-0.0318625979,0.3192126155,-0.2210430503,0.0657729283,-0.2718800604,-0.0785162896,0.0239786264,0.3921006918,-0.3312179148,-0.0785401911,-0.0629318506,0.1922969073,0.2366017401,0.0647772327,-0.1289997995,0.2295152694,-0.0741891116,-0.3985043168,0.3420298994,-0.5386348367,-0.4959172904,-0.0198907666,-0.0243965182,-0.4006351829,-0.0844035372,-0.5008683205,0.1480123401,0.3623288572,-0.0370375477,-0.2316140682,-0.0645410791,0.0931225866,0.1666582376,-0.0898157209,0.1690057218,-0.0272076726,-0.1126220226,-0.110605292,-0.254230231]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2239","title":"Error loading wikihow dataset","comments":"That was it. Thank you Albert!","body":"## Describe the bug\r\n\r\nWhen attempting to load wikihow into a dataset with\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\nI get the message:\r\n```\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\nat the end of a [full stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2).\r\n\r\n## Steps to reproduce the bug\r\n\r\nI have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https:\/\/huggingface.co\/datasets\/wikihow) says to use \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow')\r\n```\r\nto load the dataset. I do so and I get the message\r\n```\r\nAssertionError: The dataset wikihow with config all requires manual data.\r\n Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https:\/\/github.com\/mahnazkoupaee\/WikiHow-Dataset.\r\n You need to download the following two files manually:\r\n 1) https:\/\/ucsb.app.box.com\/s\/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under \/wikihowAll.csv\r\n 2) https:\/\/ucsb.app.box.com\/s\/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under \/wikihowSep.csv\r\n\r\n The can e.g. be \"~\/manual_wikihow_data\".\r\n\r\n Wikihow can then be loaded using the following command `datasets.load_dataset(\"wikihow\", data_dir=\"\")`.\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='')\r\n```\r\n\r\nSo I create a directory `.\/wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory.\r\n\r\nThen I run \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\n\r\nthat's when I get the [stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2)\r\n\r\n## Expected results\r\nI expected it to load the downloaded files into a dataset.\r\n\r\n## Actual results\r\n```python\r\nUsing custom data configuration default-data_dir=.%2Fwikihow\r\nDownloading and preparing dataset wikihow\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/azureuser\/.cache\/huggingface\/datasets\/wikihow\/default-data_dir=.%2Fwikihow\/0.0.0\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... ---------------------------------------------------------------------------\r\nAttributeError\r\nTraceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset('wikihow',data_dir='.\/wikihow')\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n745 try_from_hf_gcs=try_from_hf_gcs,\r\n746 base_path=base_path,--> \r\n747 use_auth_token=use_auth_token,\r\n748 )\r\n749 \r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n577 if not downloaded_from_gcs:\r\n578 self._download_and_prepare( -->\r\n579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs \r\n580 ) \r\n581 # Sync info\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n632 split_dict = SplitDict(dataset_name=self.name)\r\n633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) -->\r\n634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) \r\n635 \r\n636 # Checksums verification\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikihow\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2\/wikihow.py in _split_generators(self, dl_manager)\r\n132\r\n133 path_to_manual_file = os.path.join(\r\n--> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename \r\n135 ) \r\n136\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n```\r\n- Datasets: 1.5.0\r\n- Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0]\r\n- Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic\r\n```","comment_length":6,"text":"Error loading wikihow dataset\n## Describe the bug\r\n\r\nWhen attempting to load wikihow into a dataset with\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\nI get the message:\r\n```\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\nat the end of a [full stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2).\r\n\r\n## Steps to reproduce the bug\r\n\r\nI have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https:\/\/huggingface.co\/datasets\/wikihow) says to use \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow')\r\n```\r\nto load the dataset. I do so and I get the message\r\n```\r\nAssertionError: The dataset wikihow with config all requires manual data.\r\n Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https:\/\/github.com\/mahnazkoupaee\/WikiHow-Dataset.\r\n You need to download the following two files manually:\r\n 1) https:\/\/ucsb.app.box.com\/s\/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under \/wikihowAll.csv\r\n 2) https:\/\/ucsb.app.box.com\/s\/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under \/wikihowSep.csv\r\n\r\n The can e.g. be \"~\/manual_wikihow_data\".\r\n\r\n Wikihow can then be loaded using the following command `datasets.load_dataset(\"wikihow\", data_dir=\"\")`.\r\n .\r\n Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='')\r\n```\r\n\r\nSo I create a directory `.\/wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory.\r\n\r\nThen I run \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wikihow', data_dir='.\/wikihow')\r\n```\r\n\r\nthat's when I get the [stack trace](https:\/\/gist.github.com\/odellus\/602c3b2de52f541d353b1022f320ffc2)\r\n\r\n## Expected results\r\nI expected it to load the downloaded files into a dataset.\r\n\r\n## Actual results\r\n```python\r\nUsing custom data configuration default-data_dir=.%2Fwikihow\r\nDownloading and preparing dataset wikihow\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/azureuser\/.cache\/huggingface\/datasets\/wikihow\/default-data_dir=.%2Fwikihow\/0.0.0\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... ---------------------------------------------------------------------------\r\nAttributeError\r\nTraceback (most recent call last)\r\n in \r\n----> 1 dataset = load_dataset('wikihow',data_dir='.\/wikihow')\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n745 try_from_hf_gcs=try_from_hf_gcs,\r\n746 base_path=base_path,--> \r\n747 use_auth_token=use_auth_token,\r\n748 )\r\n749 \r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n577 if not downloaded_from_gcs:\r\n578 self._download_and_prepare( -->\r\n579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs \r\n580 ) \r\n581 # Sync info\r\n~\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n632 split_dict = SplitDict(dataset_name=self.name)\r\n633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) -->\r\n634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) \r\n635 \r\n636 # Checksums verification\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikihow\/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2\/wikihow.py in _split_generators(self, dl_manager)\r\n132\r\n133 path_to_manual_file = os.path.join(\r\n--> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename \r\n135 ) \r\n136\r\nAttributeError: 'BuilderConfig' object has no attribute 'filename'\r\n```\r\n## Versions\r\nPaste the output of the following code:\r\n```python\r\nimport datasets\r\nimport sys\r\nimport platform\r\n\r\nprint(f\"\"\"\r\n- Datasets: {datasets.__version__}\r\n- Python: {sys.version}\r\n- Platform: {platform.platform()}\r\n\"\"\")\r\n```\r\n```\r\n- Datasets: 1.5.0\r\n- Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0]\r\n- Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic\r\n```\nThat was it. Thank you Albert!","embeddings":[-0.2274995297,0.370223105,0.0238096733,0.3987950683,0.2537099421,0.2747787535,0.4276534021,0.4391774237,0.2458434254,0.0935082212,0.216285497,0.3853774071,-0.0128944451,0.1863868237,0.1205022633,-0.2686142325,0.0110668847,0.1045249701,0.223709926,0.1890229732,-0.3448888063,0.0065287249,-0.2329775989,0.1040147245,-0.3400695622,0.0130667957,-0.0981442481,0.1813103855,-0.0533038154,-0.541544497,0.4757557511,-0.1185788065,0.2518153787,0.4039177299,-0.0001126044,0.1877029687,0.5060688257,0.0342030711,-0.442166388,-0.0944597572,-0.3161360323,-0.1228394881,0.2420690507,-0.1337201148,-0.0326552615,-0.0258759465,0.004085809,0.0358592607,0.2355376929,0.2652969062,0.1653248966,-0.1196185276,0.3745595217,-0.1360630244,0.1942486763,-0.0933305025,-0.0869166702,0.1622466892,0.0739811137,-0.2382172346,0.1115441099,0.20557262,-0.1832122952,0.0634829551,0.5792307854,-0.0081012985,0.2564974129,-0.332991004,0.2632296085,0.3839665055,0.8194904923,-0.2249974608,-0.2384388745,-0.2182331532,0.1727716476,0.1644721776,0.3875032365,0.3061026633,-0.3748465776,0.0360186175,-0.0377139971,-0.2557303309,-0.0580877513,0.2808242142,-0.294668287,0.4108183086,-0.0778242275,0.1745568216,-0.2258381695,-0.0642693043,0.2035174668,-0.32481125,-0.2376295626,0.3015760481,-0.1384118348,0.0768692344,0.1585353613,0.1928257942,0.0994419307,-0.0371334255,-0.100838013,-0.1908228844,0.1564052403,0.1384836882,0.4442620873,0.2339715958,-0.0133317849,0.2062765062,0.1207495183,0.3048851788,-0.1984937936,-0.1141341776,-0.201289773,-0.2741613984,0.1328350455,-0.2296912819,0.3544214666,-0.0338293388,-0.2546579242,0.2001613677,0.0116372267,0.0730370209,0.0347945094,0.4159566164,-0.2064811736,0.1446116865,0.2597591877,0.2927667499,0.1025156602,0.1234883666,-0.1005089432,0.0907599106,-0.0558772162,-0.0318382233,0.1500617266,-0.1930603236,0.2553026974,0.1958255172,0.125093013,-0.2846936584,-0.0093548056,0.2284626216,-0.2356490493,0.3494037092,0.0726296008,0.1431750059,0.1701949537,-0.6475993991,-0.1152566895,0.1633387059,-0.394199729,-0.3855848312,-0.2756711245,0.2184655666,-0.0403429158,0.0450250953,-0.419618398,-0.1785893738,0.2975123823,-0.2910596132,-0.2390632927,-0.0699154362,0.1259678304,-0.1256434321,0.356915772,0.4782387316,-0.3908683658,-0.0335086882,-0.0930062383,0.0438776985,-0.0572247431,-0.3101461828,-0.3554069996,0.4971389174,-0.288457185,-0.2874653041,0.2937496006,-0.4576177597,-0.2892730534,0.2900746465,0.0192084331,0.3332006037,-0.0229105316,-0.1692107767,-0.168219924,-0.0836994424,0.2327568531,0.2221877873,-0.0055906828,0.0550968647,-0.1940852553,-0.2364349663,0.1046095416,0.186486274,0.0511644967,0.1748686433,0.0437120758,0.0242446903,0.1256706417,-0.1626351625,-0.0451247096,0.1203386486,-0.0281695016,-0.0047448212,0.048327703,0.0023430884,-0.4610753655,0.3484283388,0.0433676019,0.1083413661,-0.1505929083,-0.0570079163,-0.2748941183,0.1563291699,-0.5284996033,-0.1772395521,0.16646716,0.0618430004,-0.149765268,-0.018513212,-0.124017477,0.0431065783,-0.3334240615,0.1222898439,-0.2070953697,0.1706795096,-0.1306974888,-0.0565901399,-0.0111799212,0.1489347219,0.0346692316,0.0848748386,-0.2042343467,0.3257358968,0.0568369739,0.1124035195,-0.0444702618,-0.0688852519,-0.037389446,-0.1064234599,0.0548359565,0.2269100845,0.2477425188,-0.0861583799,-0.092016995,0.230902642,0.0540100299,0.3870119154,0.0292328987,0.267403245,0.2057381421,-0.0143849896,-0.024044076,0.0058025098,0.1215947866,0.2089033723,0.0798321962,-0.0836231709,-0.1614168137,-0.1593151838,0.098150149,0.0621439032,0.0763312802,0.0264563933,-0.2656008005,0.0542387515,0.0102328975,0.2679121494,0.2297856659,0.0542744882,-0.2823029459,0.029405009,-0.0894653425,-0.0224468429,0.3617697656,0.1321124285,0.1846684366,0.3180979192,-0.1216696352,0.1428943425,-0.0601892583,-0.2058867961,-0.1090199202,0.2014214098,-0.4196126759,0.0011625303,0.0554270037,-0.12873137,-0.0488140583,-0.2332871109,-0.2695047557,-0.4261551201,-0.0986412242,0.0017243456,-0.1481575668,0.1272243708,-0.1493606865,0.1027415693,0.0276107434,-0.2686441839,-0.2308505625,-0.3650170863,-0.3169379532,-0.010910823,0.4290106595,0.014576084,0.1516437829,-0.3627824783,-0.041841913,-0.0665280297,-0.182035476,0.1500556469,0.0805360898,0.7029446959,0.1669664234,0.4550088048,0.0690727234,-0.4021815658,0.3672315776,0.0559912547,0.1430447549,0.2535715699,-0.0262755174,-0.233836621,0.0019271201,-0.326227963,-0.1746632308,-0.2757624686,-0.2295166552,0.3282880485,0.0036655893,0.5006689429,0.2981620729,0.1741961688,0.0721937418,0.293499738,-0.0413388088,-0.6339867711,0.3558411598,-0.1906136423,-0.2751670182,0.2133964598,-0.1321064532,0.2231930494,-0.0869277865,-0.5945403576,-0.0181757268,-0.0856766701,0.2774372101,-0.0212452579,0.174653545,0.0803521276,0.1810375452,0.0438931286,-0.16777201,0.0480528399,-0.1401127279,-0.1015582308,0.1961410195,0.1260029525,0.1382963806,-0.3169760108,0.6540639997,0.0909557492,0.1336974353,0.282076925,-0.2309667915,0.237269327,-0.2938368618,-0.366079241,0.142866239,-0.1669014841,-0.1178768277,0.2168298066,0.0458054617,-0.2475858182,-0.3451890051,0.0167480763,-0.0648541227,-0.3360706866,0.00500212,-0.0207578167,0.1135326698,0.1858856827,0.1822646409,-0.2703818381,-0.2345915437,0.1548732817,0.4231767952,0.0087489625,-0.1173151657,0.0347412154,-0.1464680284,-0.2632380128,0.0673621967,0.143509835,0.0301828831,0.0692543685,0.160138011,-0.0753667355,-0.043373514,0.4331784248,-0.2390737832,0.2342213988,0.2200020254,0.2694022655,-0.3107819855,-0.1034184247,-0.0251532011,0.0629202202,0.0118524972,0.1688247323,-0.0887500793,-0.0934559628,0.131155178,0.4144935906,-0.2235663086,-0.2698959708,-0.2487764359,-0.0687507242,-0.4285556972,-0.2023697495,-0.0320565999,0.4673660994,-0.0326748192,-0.0647634789,-0.2586669624,0.2108554989,-0.050994385,0.1131491736,0.3406184614,0.0321993232,0.1324106604,0.1004277468,0.1380321234,0.1158991605,0.534376204,0.2168563753,-0.2238845527,-0.2228387147,-0.0851792991,0.0313075818,0.1784212738,-0.2225720435,-0.4367471039,-0.152959764,0.0778846219,-0.2856153548,0.0102684777,0.1609386653,-0.0096937818,-0.2836462259,-0.5890882015,0.6666560173,-0.0377636254,-0.0160097722,0.377890408,0.2793262601,-0.468231082,0.291544944,0.0237020031,0.7756505013,0.1787301749,0.189420104,0.385335654,0.2011995018,0.0771394819,-0.3319946826,-0.0282286443,-0.2672969997,-0.1985023022,-0.0653544441,-0.0603929572,0.2244365662,0.0444639623,-0.3415676653,0.1563438028,-0.2356469035,0.1824591309,0.1031999588,0.1254787147,-0.2654340863,-0.2216459364,-0.3295260668,0.103697449,-0.0262821745,0.1340884715,-0.1326717138,0.0420874394,0.03195123,-0.2009951323,-0.276255846,0.2567481101,-0.1589357853,0.1236921176,0.0484220237,-0.1378914118,0.2281662375,0.5829567909,-0.1593587995,0.244093135,-0.3477098942,0.3507631421,-0.1915881932,-0.0595127791,-0.0050645182,0.118300125,0.3890743256,-0.0086640865,-0.1870107055,0.0776970685,-0.3867884576,-0.0793544203,-0.0003725718,-0.0602900982,0.2495678067,-0.3778285384,-0.3229056597,-0.1403657049,0.1496346146,-0.2204141617,0.1755172312,0.0848179311,-0.3581290841,0.0449242666,0.0820429102,-0.2780804634,0.061380256,0.1504953504,-0.1318150908,0.1769619435,0.6715326309,0.1515841633,-0.0835839659,-0.1752031296,0.3402463794,0.239140898,-0.3221225441,0.0764362887,0.1584532559,0.1720089465,-0.1837609559,0.5057669878,0.0111206537,-0.0361346528,0.1237044409,-0.611213088,-0.180048719,0.1154958606,0.0847896934,0.2263802886,-0.0776381344,-0.0945743844,0.1279538423,-0.0588704124,-0.2863120735,-0.058799658,-0.2254420966,0.1243875474,0.184392944,0.3527343273,-0.0772996247,-0.1034431532,0.1885668933,-0.2939728498,0.0775585771,-0.2647897005,-0.1814705282,0.1575973481,0.1562161744,-0.0526112169,0.1496885568,-0.1911634952,-0.1454014778,-0.3024875522,0.2301179171,-0.1878865212,-0.1893222928,0.1273608804,0.0039846329,0.2513929605,-0.2547197342,0.2307413518,-0.1212229207,0.3786002398,-0.069414407,0.0638326257,0.0754180849,0.0700532496,0.0636797994,0.1784889549,-0.4604460895,0.2478851378,0.4021214545,-0.5224174261,0.3102975786,0.2887887061,0.0420287549,0.4635989666,-0.4118162394,0.2062280476,0.2217870057,0.2036857456,-0.2438907027,-0.0228031687,0.3716575503,0.0349622183,-0.0560302511,-0.0056387219,-0.0307278559,-0.3672403395,-0.3781550229,0.0917167813,0.2237689197,-0.3046437204,0.3425470591,0.6575306654,0.0451318212,0.1759259701,-0.1087863594,0.2009283304,-0.1053792685,0.6286644936,-0.2232085168,0.0522453375,0.0775191337,0.0544747785,0.021548694,-0.3388665318,0.403706938,0.2471379489,-0.2143837512,0.0779276788,-0.052623488,0.0070789112,-0.1041948274,0.1115579233,-0.2287227958,-0.0595837533,-0.1917319596,-0.1866235137,-0.321981132,-0.2954736054,0.1318513751,0.2057689726,0.0225878377,-0.2227572054,-0.0292872693,0.1319862455,-0.1630318463,-0.4736092091,0.3834295273,0.1882407069,0.031339258,-0.2784054279,0.0510508195,0.1162586808,0.0456838869,0.1139037758,0.2092193663,0.6908439398,0.4375931025,-0.2205272019,0.1245531514,0.0927388892,-0.2373757958,-0.1092979833,0.0666442737,0.0102185681,0.169852078,0.327993542,0.1993826479,-0.1710319817,-0.1310653538,0.1197896376,0.1861197352,-0.3883250058,0.2811678052,-0.189319104,-0.1461223364,-0.1223476902,0.0990120694,-0.2563290596,0.0708758011,0.4077413082,0.2742966413,0.0725743026,-0.1249823272,0.0650696903,-0.1079402864,0.4987427592,0.3013652265,0.1351226717,-0.3123226762,-0.0683784559,-0.5570700169,0.0616605058,0.0642265305,-0.2956343293,0.1404156685,-0.0501559041,-0.131200552,0.1987145394,0.0020378774,-0.2134056091,0.2585369945,0.2531548738,-0.1139301062,-0.5248924494,0.1135269031,0.2256429642,-0.1041887999,-0.3573681712,-0.0946138576,-0.3640032113,0.0264801551,0.0779690817,0.0237678513,0.204964146,-0.183862105,0.3086179793,-0.0612024888,0.515040338,-0.0827326775,-0.0476548225,-0.3289958835,-0.285235256,-0.2313295454,0.2595256865,-0.0318625979,0.3192126155,-0.2210430503,0.0657729283,-0.2718800604,-0.0785162896,0.0239786264,0.3921006918,-0.3312179148,-0.0785401911,-0.0629318506,0.1922969073,0.2366017401,0.0647772327,-0.1289997995,0.2295152694,-0.0741891116,-0.3985043168,0.3420298994,-0.5386348367,-0.4959172904,-0.0198907666,-0.0243965182,-0.4006351829,-0.0844035372,-0.5008683205,0.1480123401,0.3623288572,-0.0370375477,-0.2316140682,-0.0645410791,0.0931225866,0.1666582376,-0.0898157209,0.1690057218,-0.0272076726,-0.1126220226,-0.110605292,-0.254230231]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2237","title":"Update Dataset.dataset_size after transformed with map","comments":"@albertvillanova I would like to take this up. It would be great if you could point me as to how the dataset size is calculated in HF. Thanks!","body":"After loading a dataset, if we transform it by using `.map` its `dataset_size` attirbute is not updated.","comment_length":28,"text":"Update Dataset.dataset_size after transformed with map\nAfter loading a dataset, if we transform it by using `.map` its `dataset_size` attirbute is not updated.\n@albertvillanova I would like to take this up. It would be great if you could point me as to how the dataset size is calculated in HF. Thanks!","embeddings":[-0.2140057236,-0.3152703345,-0.122815989,0.151918903,0.0606090762,0.0196822062,0.2813269794,-0.1209019348,0.1866587102,0.1171898469,-0.1872747988,0.0038696581,0.380846709,0.1846315265,0.271148622,0.0805765986,0.2649388611,0.1283430755,-0.5549340844,-0.0820946544,-0.2858287394,-0.054799363,-0.3831388056,-0.339327246,0.0922732353,0.0161281005,-0.373599112,0.0223031156,-0.0533075258,-0.1949787438,0.0718323886,0.1569414437,0.2523393631,0.3328717947,-0.000118608,0.0030874119,0.060776107,-0.232634902,0.0674877316,0.0612015128,-0.3378659785,-0.1122008637,-0.3726660013,-0.131082654,0.0208221152,-0.1059127972,-0.2514362633,-0.1718560904,-0.0952340588,0.0019068832,0.126869157,0.0038471692,0.1123806685,-0.0862614885,0.0469178446,0.4319821894,0.3182981312,0.1597346365,-0.125337854,0.0477172844,-0.1166615114,0.4219639897,0.3868218064,-0.1872308403,0.6096593738,-0.2739829421,0.0580312908,-0.2759192884,0.25853163,-0.0472459495,0.6375643611,-0.0424682684,-0.1704120487,0.2470779419,-0.0291200113,0.2279429287,0.0077345422,-0.1512331367,0.2990824282,-0.0861105248,-0.4554169178,-0.4315140843,0.1040445045,-0.0273132809,-0.2427789867,0.1446979046,-0.1112796515,0.2554887235,-0.2085964233,-0.1983199716,-0.0714238435,-0.1080984697,-0.1298040152,0.2719272971,-0.1660077125,-0.2704578936,0.0243236981,0.1149551943,0.2199000716,-0.5036532879,-0.3162200153,-0.1402003467,-0.2914329171,0.0798045769,0.1231794283,0.2272793502,-0.0206621755,0.1592146009,-0.106652841,-0.1457856894,-0.1415425539,-0.1551551223,0.0363503546,-0.1320218295,0.1413971633,-0.3502228856,-0.1277054697,-0.2908428311,0.0824375153,-0.1878947318,0.0590823814,-0.1257807612,-0.0484554656,0.4238687456,-0.0765653327,0.2112364769,0.0476730317,-0.070592925,0.0623574853,-0.3581084013,-0.1582989246,-0.2173956931,-0.0436136425,-0.040028967,-0.2888955772,-0.0153968921,0.192159012,-0.0682718679,-0.0100626312,0.0613143258,-0.2472406179,-0.3080363274,0.5430862308,0.3725977242,-0.0346395187,-0.1020848155,-0.2056690007,0.0348296911,-0.2220863402,0.5323480368,-0.4249657094,-0.0220034309,-0.0164283272,0.1099156067,0.370747149,-0.142919302,-0.3712674379,0.4043624699,0.4439758658,-0.233330965,0.1766771525,-0.144543007,-0.3452298343,-0.1946550608,-0.0265192091,0.2341801673,-0.4227820039,-0.0769671574,0.1293610781,-0.0949333832,-0.0758656785,0.1869470328,-0.2029879838,0.2052725405,-0.2751541734,-0.0264795963,0.4559625387,0.153468594,-0.4716664255,0.2193999738,0.1093273908,-0.422233969,-0.1315321624,0.179771021,0.3205760121,-0.1876436174,-0.1563882232,0.2847704291,-0.151896745,-0.0051316624,-0.2183376849,-0.0154411839,0.2294495851,-0.272349447,0.3975720108,0.2733157873,0.4171893895,0.1851175278,0.2256759852,0.0936024413,0.2195407003,0.488866359,-0.1095969379,0.0856005922,0.3812716007,0.1324256659,-0.1740548611,-0.0752758533,-0.3204288483,-0.3319666684,0.6687418222,-0.0966060609,-0.3359660804,-0.0507548861,0.196964249,-0.0666560531,-0.0605126657,0.1799889356,0.0005359737,-0.4350314438,-0.0211310685,0.3384574652,-0.1690157652,-0.0516393818,-0.1935639381,0.4136397243,0.0334193856,0.1814790964,-0.2950874269,-0.2474314272,0.1697004884,-0.129627645,-0.5138614178,0.5568604469,0.0231895167,0.1645977944,0.3235615492,-0.0610960908,0.0794885308,-0.1397193819,0.4504624009,-0.211895898,0.2293570489,-0.0954312384,-0.0957633182,-0.0936714187,-0.0251375474,0.063264586,0.0550186671,-0.221438691,0.0508439057,0.0463702865,0.0053123394,-0.1153997034,-0.4647921622,0.0697127953,0.1271218359,0.0914869159,0.29965204,0.0864425153,0.5922946334,-0.1145482883,-0.2592746913,0.4422220886,-0.6018925905,-0.3079836965,0.0779067576,0.0987894163,0.4737581909,0.1595906168,0.3140643537,-0.0679319277,0.1276168078,0.1151645109,0.1253499538,-0.2597392201,0.2894563377,0.0159956608,0.3577781916,0.0495108292,-0.1046204194,0.1467944831,0.1935370415,0.000960484,-0.2141058296,0.0306788273,0.0610924847,-0.1605760008,-0.1327034384,-0.0776216462,-0.4361529052,-0.1390067041,-0.2273796648,-0.1222077459,0.2565586269,-0.0839830562,0.1252574474,-0.059715353,0.13939327,-0.0056539536,0.1181034297,0.0697114989,-0.2667838335,0.040638987,-0.1827833503,-0.2520087957,0.3343188167,-0.1925526261,0.0872487053,-0.2872765064,-0.5419437289,0.0671591088,0.0726977438,-0.1776107103,-0.0480507761,0.0545799658,-0.2404439151,0.5561314225,-0.0330331065,-0.2603099346,-0.1486041546,-0.0934144333,-0.269133836,0.1325567812,0.0069176992,-0.1663765758,-0.1284467578,-0.0060933586,-0.1103033349,-0.0678273439,0.0013865456,-0.0018110225,0.162728861,-0.1070974469,0.0917880684,-0.2905718088,-0.4420056641,-0.3465592563,0.2333111316,-0.1668551564,-0.2456224412,0.1360714436,-0.0397342369,0.0531751961,0.285584569,-0.4390060604,0.001907522,-0.2569933236,0.1336506903,-0.3816582561,0.4650990665,0.4649777412,-0.1931083053,-0.0217304453,-0.1601399332,-0.3421410322,0.0221876036,0.5710698962,0.4459757507,0.1922606379,0.4896537066,0.032642547,0.3656321466,0.035914652,-0.2391735464,0.3064498007,-0.018202303,0.2634079456,-0.3918576539,0.1460926384,0.1333952099,0.0588989072,0.014639846,0.2835103273,0.1154615358,-0.1875747591,0.1867989302,-0.6332862973,-0.0856014192,0.1739153862,0.2057460696,-0.2999028265,0.5226066709,0.2023053765,-0.1186945662,-0.3370678425,-0.1614811569,0.0884562358,0.2191433012,0.3309368193,0.0755060688,-0.1272507012,0.1783542037,-0.0374495164,0.4621969759,0.1265323013,0.1225271299,0.1360724717,-0.0558733344,0.0783823878,0.0126248542,0.5038232207,0.0572215952,-0.0133887175,-0.0276392233,-0.2663757503,-0.1733079702,-0.0302931871,-0.1200951561,-0.0064266543,-0.0297501814,0.5657557845,0.1122528017,-0.2860833108,0.1884264648,0.1513282508,-0.2347458154,-0.11281351,-0.0953423083,-0.0633944124,-0.238775745,0.0356184542,-0.2177069783,-0.2891511023,-0.0141806034,-0.0635060668,-0.2371362448,-0.0348993391,-0.2310150415,0.246176526,0.3423182368,0.3496639729,0.0119774994,0.1120425537,0.1592477411,0.3382798731,0.3827717602,-0.1593349725,-0.0139456578,-0.0429118983,0.1522669494,0.4659046531,0.0274841953,0.0271212123,-0.0773006007,0.1354242116,0.273098886,-0.3703922331,0.1427563727,0.2864748538,-0.0103275906,-0.3279107511,-0.3974000812,0.4388547838,-0.0843491703,-0.0292805694,0.1944139153,0.2325008065,-0.32964167,0.4847520292,0.257383436,0.8518754244,-0.1079274863,-0.2634683251,-0.0353159457,-0.2792352736,-0.3146860301,-0.4570840895,0.0338804759,-0.3315185606,-0.2903365791,-0.0247495193,0.0222284608,0.1811502874,0.3551355004,-0.407571733,0.1832198054,0.170309335,0.3071151972,-0.1503184587,0.0610583276,0.018247949,-0.1289482713,0.0756897554,0.1380480379,0.0011310094,0.1588470191,-0.1131140143,0.1153471172,0.1345825791,-0.1179264039,0.0256785266,-0.2325983495,-0.3113340437,0.0803499073,-0.0858614817,0.204209581,0.2555887103,0.2275423855,-0.0294648297,0.2428283095,0.2011037469,0.2134095281,0.2661733627,0.4326908588,0.0473842844,0.314541012,0.1325956732,-0.0504244529,-0.3777057528,-0.1642896086,-0.1882668138,-0.3265549839,-0.3648972809,-0.1243494898,0.286233902,-0.4843809307,0.0331741422,0.130997628,0.0992185175,0.1860062033,0.0607794337,-0.0608307607,-0.0425184965,0.5782653093,0.3905580044,-0.3896168172,-0.2133714259,0.4047325552,0.0968294814,0.1713435054,0.3150086999,-0.2766409814,-0.1403547972,-0.0606271513,0.2169501632,0.0243307762,-0.1240531728,-0.0700631291,0.0081484728,0.1974965483,-0.1024969146,-0.0588436499,0.0068691093,0.1679727137,-0.3675274551,-0.3850823641,-0.560977459,0.0693347901,-0.1425025314,0.0989469066,0.2754642069,-0.1074578315,0.211594969,0.6015930772,-0.2662201524,-0.0550230108,0.3377180099,0.4226115644,0.0877096355,0.312435478,-0.0761968046,-0.0099370442,-0.0009489558,-0.1354351938,-0.1441879869,-0.1140036285,-0.0733992532,0.1309792101,-0.1095949858,0.1733467579,-0.2824149728,-0.5237739086,-0.0834713206,-0.2588953078,0.0914598927,-0.1052842662,-0.0736050382,0.3677444458,0.4271843135,-0.1542941332,0.1063289568,0.2279397845,0.3831100762,0.1607558429,-0.3790391386,0.2424962819,-0.0449334942,-0.1546946466,-0.4974563718,0.1371769756,0.4143356383,-0.2048913836,0.174521029,-0.0415025577,0.5330858827,-0.1215084717,0.1827476174,0.0399593189,0.2078929842,-0.1102114692,0.0375649147,0.1143459976,-0.2187923193,0.1581638008,0.0632972643,0.1472353339,-0.0786529705,0.2484925389,0.2706505358,0.470492214,-0.0549882725,0.1468248218,-0.0151100652,-0.0116155297,-0.236306712,0.7342295051,-0.1024148241,0.2639229596,0.2599751651,0.2073660642,-0.0222291052,0.5434538126,0.4413153231,0.3745322824,0.0129652452,0.2651264071,0.3152928948,-0.4953007996,0.1709749848,0.319352001,0.0870403722,0.0203999262,0.3167509139,-0.18710953,0.4784356058,-0.5166832805,-0.0488771833,0.3402392268,-0.0470013805,-0.1142579019,0.1654684246,-0.0926523283,-0.2543624938,0.071370922,-0.0918413699,-0.0569909923,0.5444126725,0.0117150834,0.0043229046,0.0162324831,-0.0063114245,0.0247222707,0.0959702209,-0.1606917083,0.4082081318,-0.0393902622,0.2547602654,0.4199392796,0.3694162667,-0.0672288388,0.1705738604,0.0450081341,0.3584500253,0.2183001637,-0.2190245688,0.0465470739,-0.1061874032,-0.1234612092,-0.2593729198,-0.0262956955,0.0656529143,-0.1066657826,0.1110750586,-0.2047566772,0.222638309,-0.2474738061,0.5339394212,0.1450458914,-0.19860138,-0.261089921,-0.2012497783,-0.2419426441,-0.1441998482,0.4378477931,-0.2818748355,0.1111840308,-0.0101880999,0.0446583964,0.0988991857,0.4319502711,0.2890982628,-0.28081128,-0.1705264598,-0.4154456258,-0.3810479045,-0.0544290207,0.1726622432,-0.0116676567,0.0804960877,0.1426573545,0.0874043629,-0.1501584649,0.6539110541,-0.5256928802,0.075252831,0.5172635317,0.0599824674,-0.2514145672,-0.6457121968,0.1792384833,0.1014218107,-0.1655282825,0.1507623792,0.2050513923,-0.0268513672,-0.2309288979,0.1188352257,0.2486523241,-0.1183630005,0.4065017104,0.0029659683,0.1948698312,-0.304128319,0.2305038571,0.2251961231,-0.0709353462,-0.3524971008,0.3481687009,-0.1475655288,0.1479864568,-0.0994005054,-0.4449288249,0.1051518396,-0.1708311886,-0.1205765307,-0.14268969,-0.2406900823,-0.0736687779,-0.1960358173,-0.0846508443,0.3271540105,0.3446578979,0.220822677,0.0302764978,-0.3022896647,-0.0665532947,0.2736282051,-0.2532921731,-0.3596785665,-0.1206596121,-0.0495742708,-0.2490158975,-0.5401294827,-0.3720544279,0.0165632349,0.5108923316,0.0008996485,0.3345389366,0.2416678965,0.1318820119,0.3801806271,-0.2322729975,0.4988072217,-0.1960777044,-0.0948148817,-0.4374320805,-0.076536037]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"Hi ! Indeed there's no verification on the uniqueness nor the types of the keys.\r\nDo you already have some ideas of what you would like to implement and how ?","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":31,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\nHi ! Indeed there's no verification on the uniqueness nor the types of the keys.\r\nDo you already have some ideas of what you would like to implement and how ?","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"Hey @lhoestq, thank you so much for the opportunity.\r\nAlthough I haven't had much experience with the HF Datasets code, after a careful look at how the `ArrowWriter` functions, I think we can implement this as follows:\r\n\r\n1. First, we would have to update the `ArrowWriter.write()` function here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fcd3c3c8e3b1d9a2f3686a496082e21f06591380\/src\/datasets\/arrow_writer.py#L296\r\nso that it accepts an additional argument `key` which would be appended along with the example here after hashing.\r\n\r\n2. Then, we would need to create a `Hasher` class which will take the key as its input and return a hash for it (We might need to use some hash salt which can be passed to the ArrowWriter.writer() with value equal to the `split_name` for differentiating between same keys of different splits)\r\n\r\n We can use the `hashlib.md5` function for hashing which will conert each key to its byte code before hashing (depending on the data type of the key) **Thus, the `key` type will be verified here**.\r\n\r\n3. Now, we would have to edit this\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fcd3c3c8e3b1d9a2f3686a496082e21f06591380\/src\/datasets\/arrow_writer.py#L257\r\n so that it iterates over each `(hash, example)` pair (sorted according to hash). We can then simply **check whether each hash is different from the previous hash** (since they will be sorted)\r\n\r\nHowever, since I'm not very familiar with how the data is being written on disk in the form of a table, I might need some guidance for Step 3. \r\nPlease let me know your thought on this. Thanks!","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":235,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\nHey @lhoestq, thank you so much for the opportunity.\r\nAlthough I haven't had much experience with the HF Datasets code, after a careful look at how the `ArrowWriter` functions, I think we can implement this as follows:\r\n\r\n1. First, we would have to update the `ArrowWriter.write()` function here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fcd3c3c8e3b1d9a2f3686a496082e21f06591380\/src\/datasets\/arrow_writer.py#L296\r\nso that it accepts an additional argument `key` which would be appended along with the example here after hashing.\r\n\r\n2. Then, we would need to create a `Hasher` class which will take the key as its input and return a hash for it (We might need to use some hash salt which can be passed to the ArrowWriter.writer() with value equal to the `split_name` for differentiating between same keys of different splits)\r\n\r\n We can use the `hashlib.md5` function for hashing which will conert each key to its byte code before hashing (depending on the data type of the key) **Thus, the `key` type will be verified here**.\r\n\r\n3. Now, we would have to edit this\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fcd3c3c8e3b1d9a2f3686a496082e21f06591380\/src\/datasets\/arrow_writer.py#L257\r\n so that it iterates over each `(hash, example)` pair (sorted according to hash). We can then simply **check whether each hash is different from the previous hash** (since they will be sorted)\r\n\r\nHowever, since I'm not very familiar with how the data is being written on disk in the form of a table, I might need some guidance for Step 3. \r\nPlease let me know your thought on this. Thanks!","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"Interesting !\r\nWe keep the dataset sorted in the order examples are generated by the builder (we expect the dataset builders to generate examples in deterministic order). Therefore I don't think we should shuffle the examples with the hashing. Let me know what you think.\r\nOther that that, I really like the idea of checking for keys duplicates in `write_examples_on_file` :)\r\n\r\nThis looks like a great plan ! Feel free to open a PR and ping me if you have questions or if I can help\r\n","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":86,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\nInteresting !\r\nWe keep the dataset sorted in the order examples are generated by the builder (we expect the dataset builders to generate examples in deterministic order). Therefore I don't think we should shuffle the examples with the hashing. Let me know what you think.\r\nOther that that, I really like the idea of checking for keys duplicates in `write_examples_on_file` :)\r\n\r\nThis looks like a great plan ! Feel free to open a PR and ping me if you have questions or if I can help\r\n","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"@lhoestq I'm glad you liked the idea!\r\nI think that since the keys will be unique and deterministic in the nature themselves, so even if we shuffle the examples according to the hash, a deterministic order would still be maintained (as the keys will always have the same hash, whenever the dataset is generated). \r\nAnd since, we are not dealing with time series data (which would require the data to be in original order), I don't think the order of examples would matter much, as long as the order is deterministic and constant for all users.\r\n\r\nI think that this is also what was originally envisioned as mentioned in the documentation here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/6775661b19d2ec339784f3d84553a3996a1d86c3\/src\/datasets\/builder.py#L973\r\n\r\nAlso, if we avoid this, we would need to keep track of all the hashed keys in some place and compare each individual key with all others. This can cause some major overhead as each dataset consists of tens of thousands of examples.\r\nLet me know your thoughts in it! I would be opening a PR soon :)","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":171,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\n@lhoestq I'm glad you liked the idea!\r\nI think that since the keys will be unique and deterministic in the nature themselves, so even if we shuffle the examples according to the hash, a deterministic order would still be maintained (as the keys will always have the same hash, whenever the dataset is generated). \r\nAnd since, we are not dealing with time series data (which would require the data to be in original order), I don't think the order of examples would matter much, as long as the order is deterministic and constant for all users.\r\n\r\nI think that this is also what was originally envisioned as mentioned in the documentation here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/6775661b19d2ec339784f3d84553a3996a1d86c3\/src\/datasets\/builder.py#L973\r\n\r\nAlso, if we avoid this, we would need to keep track of all the hashed keys in some place and compare each individual key with all others. This can cause some major overhead as each dataset consists of tens of thousands of examples.\r\nLet me know your thoughts in it! I would be opening a PR soon :)","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"When users load their own data, they expect the order to stay the same. I think that shuffling the data can make things inconvenient.\r\n\r\n> I think that this is also what was originally envisioned as mentioned in the documentation here:\r\n\r\nThis part was originally developed by tensorflow datasets, and tensorflow datasets indeed does the shuffling. However in this library this is probably not what we want in the general case. But if @albertvillanova and @thomwolf you have opinions on this please let us know.\r\n\r\n> Also, if we avoid this, we would need to keep track of all the hashed keys in some place and compare each individual key with all others. This can cause some major overhead as each dataset consists of tens of thousands of examples.\r\n\r\nMaybe we cam simply keep track of the hashes of of each batch being written ? The size of the batch when the data are save in arrow is 10 000 examples. This would only ensure that we don't have duplicates in each batch, but there might still be duplicates across batches. For 10 000 examples the hashes can just be stored as a python `set`.\r\n\r\nOtherwise if we want full deduplication, we need an extra tool that allows to temporarily save and query hashes that may need to use disk space rather than memory.","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":224,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\nWhen users load their own data, they expect the order to stay the same. I think that shuffling the data can make things inconvenient.\r\n\r\n> I think that this is also what was originally envisioned as mentioned in the documentation here:\r\n\r\nThis part was originally developed by tensorflow datasets, and tensorflow datasets indeed does the shuffling. However in this library this is probably not what we want in the general case. But if @albertvillanova and @thomwolf you have opinions on this please let us know.\r\n\r\n> Also, if we avoid this, we would need to keep track of all the hashed keys in some place and compare each individual key with all others. This can cause some major overhead as each dataset consists of tens of thousands of examples.\r\n\r\nMaybe we cam simply keep track of the hashes of of each batch being written ? The size of the batch when the data are save in arrow is 10 000 examples. This would only ensure that we don't have duplicates in each batch, but there might still be duplicates across batches. For 10 000 examples the hashes can just be stored as a python `set`.\r\n\r\nOtherwise if we want full deduplication, we need an extra tool that allows to temporarily save and query hashes that may need to use disk space rather than memory.","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"Yes I think we want to keep the original order by default and only shuffle when the user ask for it (for instance by calling `dataset.shuffle()`). That\u2019s how I had it in mind originally.","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":34,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\nYes I think we want to keep the original order by default and only shuffle when the user ask for it (for instance by calling `dataset.shuffle()`). That\u2019s how I had it in mind originally.","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"Hey @lhoestq, I just had a more in-depth look at the original TFDS code about why the keys and hash were used in the first place.\r\n\r\nIn my opinion, the only use that the `hash(key)` serves is that it allows us to shuffle the examples in a deterministic order (as each example will always yield the same key and thus, the same hash on every system) so that the same dataset is generated for each user, irrespective of the order the examples are yielded by the dataset builder on different user systems.\r\n\r\nOtherwise, if we are not shuffling, then while yielding and writing the data, after getting the key and hashing it for an example, I can't quite see the use of the hash or the key. The hash will simply be generated for each example but not actually used anywhere?\r\n\r\n@lhoestq @thomwolf It would be great if you could explain a bit more about the usage of keys. Thanks!\r\n","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":160,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\nHey @lhoestq, I just had a more in-depth look at the original TFDS code about why the keys and hash were used in the first place.\r\n\r\nIn my opinion, the only use that the `hash(key)` serves is that it allows us to shuffle the examples in a deterministic order (as each example will always yield the same key and thus, the same hash on every system) so that the same dataset is generated for each user, irrespective of the order the examples are yielded by the dataset builder on different user systems.\r\n\r\nOtherwise, if we are not shuffling, then while yielding and writing the data, after getting the key and hashing it for an example, I can't quite see the use of the hash or the key. The hash will simply be generated for each example but not actually used anywhere?\r\n\r\n@lhoestq @thomwolf It would be great if you could explain a bit more about the usage of keys. Thanks!\r\n","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"In `datasets` the keys are currently ignored.\r\nFor shuffling we don't use the keys. Instead we shuffle an array of indices. Since both the original order of the dataset and the indices shuffling are deterministic, then `dataset.shuffle` is deterministic as well.\r\nWe can use it to:\r\n1. detect duplicates\r\n2. verify that the generation order is indeed deterministic\r\n3. maybe more ?","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":62,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\nIn `datasets` the keys are currently ignored.\r\nFor shuffling we don't use the keys. Instead we shuffle an array of indices. Since both the original order of the dataset and the indices shuffling are deterministic, then `dataset.shuffle` is deterministic as well.\r\nWe can use it to:\r\n1. detect duplicates\r\n2. verify that the generation order is indeed deterministic\r\n3. maybe more ?","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2230","title":"Keys yielded while generating dataset are not being checked","comments":"Thanks a lot @lhoestq. I think I understand what we need to do now. The keys can indeed be used for detecting duplicates in generated examples as well as ensuring the order.\r\n\r\n> Maybe we cam simply keep track of the hashes of of each batch being written ? The size of the batch when the data are save in arrow is 10 000 examples. This would only ensure that we don't have duplicates in each batch,\r\n\r\nI think that checking for duplicates in every batch independently would be sufficient as the probability of collisions using something like `MD5` is very low. I would be opening a draft PR soon. It would be great to have your guidance. Thanks!","body":"The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!","comment_length":119,"text":"Keys yielded while generating dataset are not being checked\nThe keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not.\r\nCurrently, the keys are not being checked for any of these, as evident from `xnli' dataset generation:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\nEven after having a tuple as key, the dataset is generated without any warning.\r\n\r\nAlso, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example):\r\n```\r\n>>> import datasets\r\n>>> nik = datasets.load_dataset('anli')\r\nDownloading and preparing dataset anli\/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\\Users\\nikhil\\.cache\\huggingface\\datasets\\anli\\plain_text\\0.1.0\\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299...\r\n0 examples [00:00, ? examples\/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: \"Rete filoviaria di Parma\" ) forms part of the public transport network of the city and \"comune\" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''}\r\n2021-04-16 12:38:14.483968: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll\r\n1 examples [00:01, 1.87s\/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Sharron Macready was a popular character through the 1980's.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': \"Bastedo didn't keep any pets because of her views on animal rights.\", 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''}\r\n1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 \u2013 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage\/science fiction adventure series \"The Champions\". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''}\r\n```\r\nHere also, the dataset was generated successfuly even hough it had same keys without any warning.\r\n\r\nThe reason appears to stem from here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L988\r\nHere, although it has access to every key, but it is not being checked and the example is written directly:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/src\/datasets\/builder.py#L992\r\n\r\nI would like to take this issue if you allow me. Thank You!\nThanks a lot @lhoestq. I think I understand what we need to do now. The keys can indeed be used for detecting duplicates in generated examples as well as ensuring the order.\r\n\r\n> Maybe we cam simply keep track of the hashes of of each batch being written ? The size of the batch when the data are save in arrow is 10 000 examples. This would only ensure that we don't have duplicates in each batch,\r\n\r\nI think that checking for duplicates in every batch independently would be sufficient as the probability of collisions using something like `MD5` is very low. I would be opening a draft PR soon. It would be great to have your guidance. Thanks!","embeddings":[0.0275365282,-0.2159314305,0.0291701369,0.4762756228,0.0684194565,-0.236362353,0.4168978631,0.0804130509,0.43124336,0.1339114755,0.1672184616,0.3502569795,-0.0206804313,0.2065166831,-0.0032002178,0.4027804732,0.0770776644,0.0507332645,-0.3874769807,-0.1746384501,-0.5882232189,0.0452296324,-0.1532139629,-0.0614637099,-0.1678526849,0.1618548632,-0.0681429282,-0.0138252173,-0.1972754598,-0.4447283149,0.2861754596,0.395427525,-0.1313844472,0.3610020876,-0.0001243995,-0.031345699,0.0754736513,-0.1916624457,-0.5255250931,0.0262024067,-0.1937693655,0.0950022191,-0.0176387988,-0.5688117146,-0.053775806,-0.1629761904,-0.0972570553,-0.4674580991,0.2114742845,0.2158976346,0.0896028355,0.4042153656,0.3006412983,0.2042028606,0.2016513646,0.6070529819,-0.2093221992,-0.3547791243,0.0443218388,0.3141504526,0.081910789,0.177616924,0.0606224015,-0.17011033,0.200685814,0.1464149356,-0.2246280462,-0.323405236,-0.0920045674,0.3267407119,0.3529065847,-0.1715489775,-0.3823952675,-0.3127988577,-0.0313100591,0.1885313839,0.4732134342,0.1420394927,-0.3297613263,0.0042914129,-0.2719512284,0.3002656102,-0.0179662202,-0.0041730828,0.258790642,0.1244735271,-0.0329472497,0.0638200045,0.2017904222,-0.1953783631,-0.1708811522,-0.4019609094,0.0959564373,0.3071315587,-0.1999058276,-0.2576290667,0.2660217881,0.194044441,0.4387064874,0.0169971455,-0.0074185338,0.0199898295,-0.1039318219,-0.1370512396,-0.0529770032,0.2784969211,0.1445692927,0.0533135198,0.4216781557,-0.0334113911,-0.1720716804,0.0693692267,-0.0535320565,0.0950190723,0.4306455255,0.2696842551,0.335506022,0.025251301,-0.1482149214,0.0974797383,-0.1257552505,-0.2656867802,0.2667299807,0.3143821061,-0.0797205865,0.0989429504,-0.1258902401,0.0694410652,-0.3220945597,-0.0343445353,-0.0853379816,-0.1496426165,-0.0802808627,-0.0657494068,0.364279598,-0.3379724324,0.0242795926,0.2694472075,-0.2059150934,-0.288422972,-0.1518755108,-0.0767865106,0.5315922499,0.2941078246,-0.1158165112,0.024007896,0.0686455518,-0.4160591662,-0.309743762,0.3407129049,-0.2412345409,-0.1555794626,0.2009787858,0.0987635627,-0.3604650795,-0.1261386573,-0.1306612045,-0.0645978823,0.3372377157,0.2071813345,0.0784945637,-0.1708253771,0.1355777979,-0.4528718889,0.2073252201,0.8914080262,-0.1023142859,-0.0063826209,-0.1526415646,-0.0278965682,0.1243735701,0.3553079963,-0.0161852706,0.3325898349,-0.3431203961,0.092856735,0.1716639251,-0.10762842,-0.2139098197,0.0121155484,0.1867583841,0.0967974141,0.5034436584,0.2218383104,0.2685669065,-0.0544432886,-0.2298940867,0.0698378459,0.0156069836,0.0602037869,-0.2132327706,-0.1308508813,0.0318849385,0.0194860678,-0.0369520262,0.2195641398,0.1360975951,-0.2552815676,0.0255509987,-0.2306274325,0.0509529673,-0.0563530996,0.1623330116,0.0917921513,0.0456483886,0.4361868799,-0.4905918837,0.2230523229,-0.0022498632,0.1985981464,-0.4391951263,-0.2743707597,-0.0583684258,0.2331923544,-0.3910266459,0.1591251194,0.0221546218,0.3127879202,0.3133720756,0.2550627887,-0.0731581673,0.0802614763,-0.0975416601,0.1954384744,-0.3723435104,0.0034481185,-0.071176298,0.068382673,-0.0470243059,0.264408648,0.103337653,-0.1461705863,-0.042986434,0.2023597211,0.2752099931,-0.0221164934,0.0455042571,0.3107895553,0.2325396389,-0.0491344407,-0.038399078,0.1895335317,0.1686585546,0.0750721842,-0.2943840027,0.8053556681,-0.2581625283,0.028535137,-0.0023022937,-0.0071980618,0.0217753276,-0.1347505897,-0.1365982592,-0.1884437501,-0.0732932538,-0.0163164865,0.1072857827,0.4089407623,-0.2961963713,0.157638669,0.7058497667,0.0941348001,0.2832787037,-0.0321235694,0.1921659261,-0.08089865,-0.2282799184,0.3368208706,0.1522255242,0.1965065747,0.1185858622,-0.3717745543,-0.0730491057,-0.1460594982,0.0427722856,-0.1891172081,-0.3544773459,0.3523737192,-0.3063002825,-0.1038096994,-0.5338432193,0.2326877713,0.1194282547,0.251612097,-0.5340343118,-0.2392493188,-0.1169456393,-0.0837991312,-0.228394866,-0.179580465,-0.0798715502,-0.3404607177,0.1645958573,0.0483226292,-0.0415060073,0.2251764238,-0.0276001878,0.177685082,0.0363353752,-0.5385163426,-0.4368271828,0.0511809215,-0.145413205,-0.0151514607,0.2481909692,-0.1409252733,0.3679054081,-0.3222305775,0.2253864855,-0.2745192051,-0.3014227152,0.0428631417,-0.1585089266,0.119652167,-0.0956903771,-0.0779617354,-0.0854469389,-0.0822699219,0.2423255742,-0.2479203939,-0.5795606375,0.2682214081,0.069850184,-0.187886402,-0.2038371116,-0.5165557861,-0.0041087638,0.0617712699,0.1320423484,-0.0713649914,0.1752530336,0.0461078286,0.1794478446,-0.1891895682,0.0000351128,-0.0443066023,-0.2693847418,-0.3226259351,0.6126976013,-0.0839342177,-0.2482605726,-0.0997935906,-0.1879334301,-0.2220841497,0.148715809,-0.3126063943,0.1589475423,-0.5614319444,0.6340059042,0.2777655721,0.2941862047,0.4636279643,0.096457094,0.0151649397,-0.231760785,0.0259325672,0.3365894854,-0.0143845584,0.2851168215,0.1721878797,0.2982102633,-0.2947055399,0.4157005548,0.3359111547,-0.1027194783,0.3535640538,-0.1526918858,0.3624245822,0.0208121259,-0.3244894147,-0.0466853455,0.197660163,0.1448141932,0.1876156032,0.0363316685,0.3886367083,-0.1072061956,0.4754819572,-0.7206183672,-0.3073938191,0.0303063747,-0.1915806681,0.2396026254,0.008560624,0.3000758886,-0.1423638612,-0.2402999252,-0.2210789323,0.3997723162,0.4897811115,-0.0650007203,-0.4595610499,-0.0350506678,-0.5776282549,0.0720781311,-0.034360297,0.2558515072,0.0583175905,-0.291287601,0.1583037823,-0.1152778119,0.4106994271,-0.2006324828,0.0915498957,-0.1241664141,0.1059207991,-0.146541208,0.074377127,-0.1692644656,-0.1423255056,-0.1802759916,0.5174462199,-0.5598748922,-0.1423892826,0.0348059162,-0.0109509742,-0.1608135849,-0.4294210672,-0.436886102,-0.1710219383,-0.1745037585,0.1889982969,-0.0279020853,0.1397201866,-0.5259871483,-0.0763651356,-0.2007005066,0.1436228901,0.3456246853,0.1947358698,0.3810791671,0.1793131679,0.2992999256,0.1381073594,0.0127522396,0.107730642,0.6457735896,-0.020930469,-0.1307189018,-0.1229427606,-0.3637976646,-0.2348939329,0.0491311997,-0.3108313084,0.1380220354,0.0838794634,0.0392896496,-0.4141130447,-0.2273482084,0.2334343642,-0.0790460333,-0.3273983002,-0.4119492173,0.2217229903,-0.0523264408,-0.1161649898,0.0675454289,0.3800842464,-0.1981739551,0.1788538694,0.4224412739,0.6792136431,-0.3589565158,0.0781869739,0.132688418,0.0002009634,0.3454467952,0.1435309201,0.1005206108,-0.6070398688,0.0357894339,-0.0309856087,-0.1770862639,-0.1437167078,-0.0286894664,-0.3284612596,-0.2210587412,0.0313082188,0.2751153111,-0.2847685814,0.4140976667,-0.1807782054,-0.2084251791,0.1361821294,0.0359391384,0.1278978288,0.6129814982,0.1220616102,0.0329466015,-0.2541007102,-0.385980159,-0.1389316916,0.0961476639,-0.1959287524,0.1010784358,0.139915064,-0.1529215872,0.0752090141,0.5738867521,0.6691455841,0.1494533122,0.1169735789,-0.1957574785,0.2974543571,0.1938553005,-0.0434503295,-0.0402528569,0.0629057363,0.2413102537,-0.2347972989,-0.239189446,0.1737246811,-0.2943579257,-0.4648035765,-0.2237489969,-0.2652669251,-0.3984834254,0.4013697207,-0.1159311086,-0.2875409424,-0.1351932138,0.0114428457,0.178158313,0.1493752003,0.6839717627,-0.2366015464,-0.2745065689,-0.0755618811,0.2371231169,0.1613749564,0.185430184,0.3100966513,0.0506317988,-0.1598343551,-0.2896625102,0.4080948234,0.1270555854,-0.653804481,0.1584935784,0.0069575831,0.1260238588,0.1320844442,0.4580089748,0.0831548646,0.3045461178,0.0452790745,-0.4174903333,-0.3791794181,0.1421855837,0.1229511201,0.1663134545,-0.2781091034,-0.1295073479,0.2371924818,-0.0970914885,-0.2319469154,-0.2686059475,-0.0118640829,-0.0829752013,0.2023100704,-0.0000615233,0.2469593138,0.0123412842,0.1196076274,-0.3138856888,-0.2301682234,-0.0468295142,-0.0646984801,0.1968930215,-0.0667049885,0.2460963726,-0.3984269202,-0.2937814891,-0.0716146603,-0.2857213616,0.1225824282,0.2313950062,-0.1323323697,0.359755069,0.0998113006,0.1995752007,-0.1428146511,0.1152760461,-0.1334348172,-0.0616847426,0.1036160141,0.11604096,-0.0581950694,-0.0126068667,-0.3222407699,0.2602240443,0.1303788126,-0.0295774769,0.0631705523,-0.3162484467,0.130463466,0.1253684312,0.2124871016,0.4128771424,0.0213230364,-0.1599720865,0.1262924522,0.1095462069,-0.0524808653,-0.0774541423,-0.1057515517,0.4045765102,0.153943792,0.5210471153,-0.0272808038,-0.0759405568,-0.003981431,-0.1401819736,0.0401138961,-0.0553556532,0.037016768,0.7215251923,0.0033207084,0.0236803107,0.5608684421,0.0468490086,-0.3103269041,0.3334859312,0.0794760138,0.2596439719,-0.0594843738,0.331603229,0.0038180638,-0.2403466403,0.2577210665,0.0598208308,-0.3180663586,0.0827418491,0.2135326862,0.2427790165,-0.0996311232,0.0139412601,-0.6906050444,-0.0261755772,-0.2743199766,-0.2494876981,-0.3948239684,-0.4349914193,-0.1296161264,-0.0516813658,0.0653561354,0.0837114975,0.1000914723,0.1421042085,-0.0959611237,-0.3503395021,0.113084428,0.1265848577,0.315223217,-0.2974818051,0.1314507276,0.2653692067,0.1533068866,0.0023728472,0.307109803,0.2919712663,0.3392054439,0.0291712601,0.2081508487,-0.1007880941,-0.1049738303,-0.1104123071,0.5324909687,0.0612865537,0.0125477416,0.26301229,0.0473063402,-0.0439803712,-0.1078521237,0.0949217305,0.0566299222,0.2469438612,0.378051728,0.0062833428,-0.2240938395,0.1664492786,0.2008960992,-0.1284392178,-0.2256353199,0.0752758831,0.267449826,0.0800943002,-0.4431453943,0.0444709212,0.0107641723,0.397333473,0.3050801456,-0.223571226,-0.2529635429,-0.1469859332,-0.7591509223,-0.0828499123,-0.0884770826,-0.1693403572,-0.0722601265,0.3721694052,-0.0035544587,0.0900173038,-0.0112864031,-0.168304652,-0.0287195705,0.4541615248,-0.1920486838,-0.1011465341,-0.0437029339,-0.0857811421,-0.2175277174,-0.4019389749,0.2391231656,-0.1568981856,-0.0509429872,0.0202913303,0.4017066061,0.1667452157,-0.1145086065,0.1857395619,0.4368845224,0.520812273,0.0604049638,0.0136404643,-0.1431025565,0.0854348764,-0.0707814693,0.246494323,0.1811536103,0.3095553815,0.0287799686,0.0299920663,-0.323503226,-0.0564056188,0.2536708415,-0.1377624273,-0.0269036349,0.1137028709,-0.0217052866,-0.317278415,0.122966975,0.2679323852,0.0417121612,0.1168053448,-0.1234356388,-0.2284724712,0.0967701748,-0.2478921562,0.0329242647,-0.3760192394,0.2189298421,0.157173723,-0.091869846,-0.2250888497,0.312261045,0.3311510682,-0.1604458839,0.0366785154,0.1225570515,-0.0565136746,0.2555027008,-0.0898036584,0.1513250023,0.0428157747,-0.0526729524,0.3032369018,-0.1087279245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2229","title":"`xnli` dataset creating a tuple key while yielding instead of `str` or `int`","comments":"Hi ! Sure sounds good. Also if you find other datasets that use tuples instead of str\/int, you can also fix them !\r\nthanks :)","body":"When using `ds = datasets.load_dataset('xnli', 'ar')`, the dataset generation script uses the following section of code in the egging, which yields a tuple key instead of the specified `str` or `int` key:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\n\r\nSince, community datasets in Tensorflow Datasets also use HF datasets, this causes a Tuple key error while loading HF's `xnli` dataset. \r\nI'm up for sending a fix for this, I think we can simply use `file_idx + \"_\" + row_idx` as a unique key instead of a tuple.","comment_length":25,"text":"`xnli` dataset creating a tuple key while yielding instead of `str` or `int`\nWhen using `ds = datasets.load_dataset('xnli', 'ar')`, the dataset generation script uses the following section of code in the egging, which yields a tuple key instead of the specified `str` or `int` key:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\n\r\nSince, community datasets in Tensorflow Datasets also use HF datasets, this causes a Tuple key error while loading HF's `xnli` dataset. \r\nI'm up for sending a fix for this, I think we can simply use `file_idx + \"_\" + row_idx` as a unique key instead of a tuple.\nHi ! Sure sounds good. Also if you find other datasets that use tuples instead of str\/int, you can also fix them !\r\nthanks :)","embeddings":[-0.0539387055,0.0509747416,0.0451682322,0.1262666285,0.2513764799,0.0173006449,0.4449461997,0.2987630963,0.6008678675,0.2404715419,0.1092354357,0.4075706303,0.0195070487,0.2159906626,-0.0843307152,-0.0561311804,-0.0785400942,0.2352073789,-0.2244517952,-0.1137270257,-0.5031872392,0.121004805,0.0294014718,0.2606830299,-0.2010565847,0.0400427841,0.1937448382,0.1910863668,-0.0840927213,-0.252540797,0.4881781042,0.0085605886,0.0756866187,0.1829353571,-0.0001135974,0.0208701994,0.2497083247,-0.0734922364,-0.468265444,-0.2629375756,-0.2167611122,0.2286738306,-0.0234127901,-0.2467126846,-0.0012098321,-0.0569351614,-0.0673003197,-0.261190474,0.2860363722,0.1601759195,0.1450576633,0.4815440476,0.0250980984,-0.138239041,0.3516049385,0.2497404218,-0.1786918938,0.1878612489,0.1838114411,-0.2809710205,-0.0479889512,0.2159287632,0.206644848,0.1380616128,0.3269072473,0.2719631493,0.0823418498,-0.2623186409,-0.0092996899,0.3619779944,0.1439969987,-0.1113360003,-0.1411518604,0.0992559567,0.1987168491,-0.2580661774,0.1839321256,0.0912623852,-0.4491142333,-0.0272621475,0.1974137127,-0.1070303172,-0.2575809062,0.2337970138,0.0154427402,0.0173182525,0.0591291599,0.1151014492,0.0064529758,-0.3979426324,-0.1829753369,0.0894437879,0.2683078051,0.1556880623,-0.3238301873,-0.031098729,0.1199393868,-0.3682415187,0.3530477881,-0.0212667901,0.1273285896,0.0893297866,-0.3878022432,0.1040480211,-0.0136975339,0.0737677738,0.0256564636,0.0063200607,0.331030786,-0.1388094574,-0.053821817,0.0349739715,-0.0082260771,0.0409489498,-0.0041416218,-0.1950260252,0.4094110429,0.0185403768,-0.3756034076,0.203909114,0.0801996663,-0.0092292009,-0.0260075964,0.367501229,0.2517862022,0.1281282455,-0.032025855,0.1792397201,-0.2550680041,-0.4093573689,-0.2063461691,-0.0526213236,-0.0127909714,0.0976488143,0.2107612491,-0.3897715211,0.0392958447,0.3133696616,0.25810498,-0.0556731485,-0.0147759169,0.0536787212,-0.0691242367,0.2281856686,0.1762597859,0.129270792,0.2934645712,-0.6828520298,-0.0041344548,0.169910565,-0.5129545927,-0.3525079787,-0.2636779845,0.1806254387,0.1077662855,-0.0586749725,-0.3332086802,-0.0793849826,0.0376669317,0.2584780753,-0.0324632302,-0.3713117838,-0.0931690559,-0.3605386615,0.1431995332,0.3623300493,-0.2543599606,-0.0424474962,0.0729089081,-0.0691700354,0.1776707172,0.6391195059,-0.1121900827,0.1469333768,-0.2595930099,0.279083252,0.4829978645,-0.2314783931,-0.4842320085,0.4521036148,-0.1890470535,-0.0753735006,0.0764777213,0.3655757904,-0.0417426042,0.0079897316,0.2896290123,0.2973432541,-0.0450821184,0.0502691045,-0.3076674342,-0.2173958719,0.1462385952,0.2962858975,-0.0078178467,0.2304442376,-0.0029742457,0.1335355192,0.222263366,-0.3926811516,0.130096823,-0.0981249958,0.4231389761,0.1318524778,0.0693470463,0.1597889811,-0.5406081676,0.2111240923,-0.0602562279,0.1853376627,-0.1759155393,-0.2527042925,-0.2737975121,0.1418508142,-0.3421969414,0.2631456554,0.0014369828,-0.1117820516,-0.0338203795,-0.0560011975,-0.1530446261,0.1724900454,-0.3591039777,0.1286535114,-0.4930763245,0.3700057864,-0.0432023853,-0.0199131649,-0.0034728546,0.1393671334,-0.0881554857,-0.1522151828,-0.1518664956,0.2224430293,0.0088391555,0.0696530566,0.0315627344,0.1348724663,0.2902286649,-0.2376737893,0.0486095995,0.4267165363,0.079688929,-0.1021378264,-0.2944548726,0.6457003951,-0.2812168002,0.1006652564,0.068083182,0.1324638277,0.1876256913,-0.0304539204,-0.1810988188,-0.4469127655,0.1374972761,-0.0984272584,0.1333730817,-0.0339754373,-0.5189490914,0.1274168342,0.475038141,0.0838380754,0.1823892891,0.3433338404,-0.0326617137,0.0029704338,-0.2844299376,0.3413448036,0.3974466622,0.0594391115,-0.1956835389,-0.0011278983,-0.3301291764,-0.2890697122,0.0736719295,-0.3107881546,-0.0680122003,0.1784962118,0.0737751871,-0.0021064084,-0.3311927021,-0.0381827056,0.0670965463,0.3988737464,-0.4998731017,-0.2168361992,-0.2133771032,-0.6131717563,-0.2112469077,0.0100624897,0.1155965626,-0.3808224201,0.0488855802,0.1438613832,-0.0783905312,-0.0846129805,0.0677924603,0.3079422414,0.1276211441,-0.1672043353,-0.0977199301,-0.1523034275,-0.173420772,0.0324888304,0.1174714863,-0.0184826218,0.3863341212,-0.0920158625,0.0982084647,-0.4978992641,-0.3933636844,0.0216765832,-0.1079630181,0.1131493822,0.2016664892,0.4266102314,-0.1588403881,-0.3726124167,0.4755396247,-0.077461496,-0.3099055886,0.2457005084,-0.1300873607,-0.0226199105,-0.2289124876,-0.5481781363,0.0909843743,-0.3140561581,0.1613312066,0.0778830871,0.2737427354,0.1724705994,0.0965671316,0.0370361693,0.4510773122,-0.4143409729,-0.4166817963,-0.1695429236,0.3733316362,-0.339122355,-0.4973729551,-0.1566298753,-0.0387999229,0.1970476806,0.1855373234,-0.3948681951,0.0366139039,-0.1812382042,0.5950683951,-0.0583579578,0.0350835882,0.0124826701,-0.0552740805,-0.0447310694,-0.0964671671,0.1253857613,0.3055225313,0.1056367159,0.111787416,0.4784660637,0.4796535075,-0.044311326,0.6656321287,0.3739884198,-0.1023455635,0.4134171009,-0.2668378651,-0.0025653716,-0.2134555727,-0.109148778,-0.0565790012,0.0691796765,0.2225812376,0.0634512231,-0.0777253211,0.1510552019,0.0654857159,0.3232292831,0.0722200871,-0.3624207079,0.2906660736,-0.2244563699,0.0627596378,-0.0329024047,-0.0834364444,-0.1726052612,-0.18635647,0.1566853821,0.2086079419,0.2589940429,0.1364474148,-0.2403875589,-0.0165744498,-0.5030446053,0.1778529137,-0.018808851,0.4084614813,-0.1785601676,-0.0530699864,0.1145969331,0.155585438,0.4162330627,-0.2936872542,0.0790834948,-0.0931698829,0.063003242,-0.342048794,-0.0460149013,-0.1662649214,-0.229414776,-0.083254084,0.8664246798,-0.1325515807,-0.1677978486,0.0213307589,-0.0565345287,-0.3647374511,-0.2877304852,-0.1746883988,-0.2655872703,-0.3925242126,-0.1595258564,-0.0759658217,0.0552915782,-0.0399297215,0.0405507945,-0.2076729536,0.1380955577,0.2477145642,0.0824548751,0.1984301955,0.2575535774,0.2018103302,-0.1347910464,0.0796287581,0.1580952406,0.4207831323,-0.0389668308,-0.3720773757,-0.0269153435,-0.328648448,0.0262582824,0.1237998903,-0.0286088903,0.2414384335,-0.1652685106,0.1863650829,-0.410451442,0.0624074377,0.3004697263,-0.1325950772,-0.2000101358,-0.3252111375,0.5260792971,-0.0417512283,-0.0446025357,0.3378832042,0.3154117763,-0.1332259327,0.614861846,0.1929963529,0.7376439571,-0.2670452595,0.1003568172,0.4510695338,0.0471986234,0.6227918863,-0.0247875322,-0.1593766063,-0.3689332604,-0.3104618192,-0.0982002541,-0.1037102044,-0.1837731451,0.2103724033,-0.1627420038,0.178095907,-0.1227848083,-0.0312059391,0.0015532505,0.4075968564,-0.0931715295,-0.3137898445,-0.0394699834,0.104388155,-0.0838218182,0.4041211307,0.0119635044,-0.0912307873,-0.0310470462,-0.2687156796,-0.2670804262,0.1149845272,-0.02497207,0.2723202705,0.1781940907,-0.3479052186,0.1799740493,0.2616091669,0.5482526422,-0.1015658155,-0.0035117085,-0.0186356958,0.0723798275,0.1811472923,-0.0954222605,0.0240164511,0.4749388099,0.0008447532,-0.0845989659,-0.0245378651,0.1507389545,-0.1788440347,-0.4634506404,-0.1758327931,-0.4185681641,-0.196942687,-0.1440832913,-0.1827930659,-0.1211714894,0.0190529246,0.0539088845,-0.0202353559,0.156940341,0.0018287511,0.1977655739,-0.3082784712,0.0764584392,0.116061151,0.1974799037,-0.0349118486,0.5787038803,0.0987795219,0.0342713036,-0.1857767552,-0.022236919,0.4961244464,-0.4041717649,-0.0059912028,-0.1479429454,0.1436947137,0.1393912584,0.211880222,0.2655131519,0.195661068,0.0390240289,-0.4466374218,-0.4397516549,0.1911924034,-0.1775909662,0.0994990766,-0.2370747179,0.1441639662,0.0405682251,0.0892712325,-0.2811643779,0.0547878705,-0.0115701025,-0.0834743604,0.4311925173,0.2089464366,-0.0237138458,0.1546169221,0.089448154,0.0425377116,-0.2589315176,-0.1430261135,-0.3218593895,0.1314428002,-0.0175115708,0.0800636634,0.0103839701,-0.1721907854,0.0177123882,-0.0119695039,0.3084239364,0.0717864037,-0.2504442632,0.425216198,-0.1632285118,0.2524295449,-0.0664052963,0.1118718088,0.0872990713,0.4252062738,0.011878714,0.0856326222,-0.0836329609,-0.0515937433,-0.7974537611,0.1531335115,0.0613584593,-0.1603004783,0.0742032528,-0.3296672702,0.0964330807,-0.1546334475,0.0025849687,0.1083012596,-0.3812048733,0.0113719665,0.1066748798,0.2220423967,-0.1231629401,-0.2336769849,-0.1267967522,0.0827572197,-0.0241378359,0.2546176314,0.1124987602,0.0503082499,-0.2499805093,-0.2119123489,0.189677,-0.6406252384,-0.0183209237,0.548376441,-0.1121439561,-0.0709947795,-0.0248101465,0.1557555497,-0.0849545896,0.5093197227,-0.113964729,0.1051708683,0.0673351735,0.2445564568,0.2413163632,-0.1795834154,-0.079084985,0.5982223749,-0.4374122918,0.191398561,0.3534727097,0.5119328499,-0.0397308506,-0.1067262217,-0.2132662535,-0.0657821,-0.2211241424,-0.2181427628,-0.4431684911,-0.1813251972,-0.4223663509,0.1216450334,-0.1381988674,0.1312886029,0.0607585683,0.2641534805,-0.0350223444,-0.3320475221,0.0163129177,0.2452542782,0.0997766182,-0.2861545086,0.1145690531,0.1956957877,-0.1253987253,0.2547958493,0.1425401866,0.4778693318,0.4655953944,0.1104236916,0.3127039671,-0.0918578655,-0.1138936579,0.004037776,0.0501489453,0.2707979679,-0.1446049064,0.3734244406,0.0871923119,-0.0824730545,-0.0947782844,0.12439619,0.2726529241,-0.1429151446,0.3773798943,-0.1605291814,-0.1043976173,0.0695940927,0.0593551099,-0.2156261504,0.0690355971,0.3900226057,0.3381557167,0.2874902487,-0.1088164225,0.0806248635,-0.1326265633,0.5759512782,0.2976171076,-0.0099782106,-0.2852285802,-0.3854357004,-0.7895969152,0.158408314,-0.0122163892,-0.0574749708,0.1738658994,0.3346887529,0.1757188886,0.1005631611,-0.142514661,-0.3924269378,-0.0489995219,0.488373667,-0.2203855664,-0.0647056624,0.126144439,-0.0484355316,-0.1326726824,-0.4188384116,0.1404303312,0.2042955905,-0.0421488956,-0.3748367131,0.4626344442,0.1275418252,0.1240850985,0.1936294138,0.232998848,0.3457583487,0.0560125932,-0.1624628454,-0.1033500209,-0.2894288301,-0.3063622415,0.1979949623,0.1222503185,0.288235873,0.1899720132,-0.2254144847,-0.4239665866,0.044234246,0.1041088477,-0.4373058677,-0.3394853473,0.1927913725,-0.142690286,-0.1098595783,0.1777238995,0.3548058867,0.174344793,0.113003172,-0.0001791974,-0.4871050715,0.2116997689,-0.299128741,-0.2027867883,-0.0066586155,0.2682013512,-0.1667630076,0.0059399209,-0.3672613204,0.1299049258,0.1774536669,-0.2218391299,-0.3701560199,0.1563970745,0.1018588841,0.2022552639,-0.2157152742,0.3182597756,0.1456280947,-0.2757139802,0.173557654,-0.2715713382]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2229","title":"`xnli` dataset creating a tuple key while yielding instead of `str` or `int`","comments":"@lhoestq I have sent a PR for fixing the issue. Would be great if you could have a look! Thanks!","body":"When using `ds = datasets.load_dataset('xnli', 'ar')`, the dataset generation script uses the following section of code in the egging, which yields a tuple key instead of the specified `str` or `int` key:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\n\r\nSince, community datasets in Tensorflow Datasets also use HF datasets, this causes a Tuple key error while loading HF's `xnli` dataset. \r\nI'm up for sending a fix for this, I think we can simply use `file_idx + \"_\" + row_idx` as a unique key instead of a tuple.","comment_length":20,"text":"`xnli` dataset creating a tuple key while yielding instead of `str` or `int`\nWhen using `ds = datasets.load_dataset('xnli', 'ar')`, the dataset generation script uses the following section of code in the egging, which yields a tuple key instead of the specified `str` or `int` key:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/56346791aed417306d054d89bd693d6b7eab17f7\/datasets\/xnli\/xnli.py#L196\r\n\r\nSince, community datasets in Tensorflow Datasets also use HF datasets, this causes a Tuple key error while loading HF's `xnli` dataset. \r\nI'm up for sending a fix for this, I think we can simply use `file_idx + \"_\" + row_idx` as a unique key instead of a tuple.\n@lhoestq I have sent a PR for fixing the issue. Would be great if you could have a look! Thanks!","embeddings":[-0.0702305734,0.0961900726,0.0778552964,0.1273570806,0.1900975257,0.0126102399,0.449765563,0.2994563282,0.6488567591,0.1779723167,0.0878629908,0.4414196014,0.0609022342,0.2075343877,-0.0426086709,-0.0038821478,-0.0590648577,0.2475437671,-0.270088613,-0.1440696567,-0.4942212999,0.1290112287,0.0590818264,0.2485807985,-0.1913190633,0.0619562566,0.2353725433,0.2271416336,-0.0578649268,-0.231456086,0.5725955963,0.0114073344,0.0351137668,0.2210238129,-0.0001182187,0.0057875635,0.235965848,-0.0812724978,-0.4586208165,-0.2777384222,-0.1937724501,0.2858184576,-0.0185022056,-0.1956772506,-0.0241901614,-0.0318657011,-0.0890224576,-0.2912680805,0.3045458198,0.1411118954,0.1012161225,0.5076861978,-0.0015598519,-0.1039695516,0.2827213705,0.3210789859,-0.1544620097,0.2094181031,0.1812248081,-0.2964422405,-0.0707189068,0.2409769893,0.2307791561,0.1223838925,0.3430539966,0.2623575032,0.0293429811,-0.2340623736,0.0456491373,0.3561387956,0.152802363,-0.158778891,-0.0808353052,0.1110712066,0.1855397373,-0.1912935376,0.2102704644,0.085092403,-0.4934516847,-0.0450517051,0.171436578,-0.1658021212,-0.2685093582,0.2046644688,0.0310249925,0.0256582052,0.0522603653,0.1541383266,-0.0537844896,-0.3798465729,-0.180950731,0.0566228628,0.2785687745,0.186706081,-0.3462030292,-0.0397396572,0.1426015645,-0.3174471557,0.4102703035,0.0238242261,0.0719214156,0.065637365,-0.4437951744,0.0565213561,0.0189705268,0.0064283973,0.0634887815,0.0258461908,0.3252578676,-0.1443240941,-0.0521679446,0.0622096881,0.0331528038,0.0548977889,-0.0047640242,-0.2382578254,0.4292405844,-0.0054675061,-0.3453379869,0.2238122076,0.1361610293,-0.0346963927,-0.0265237689,0.3213921487,0.2820681334,0.1355551928,-0.064050734,0.201431334,-0.2725867033,-0.4351241291,-0.1945147365,-0.06521862,-0.0201770738,0.1389552504,0.2019282579,-0.4108029306,0.025599461,0.3576527536,0.2412360311,-0.0624580346,-0.0854442045,0.0695406795,-0.0439071059,0.2065785527,0.1528370529,0.1327440888,0.2927932143,-0.6724526882,-0.0264560003,0.1488302201,-0.4627384841,-0.3408790231,-0.2876138985,0.1304148287,0.1075053364,-0.0338144116,-0.3273074925,-0.0331979506,0.0768313408,0.3027576804,-0.0145910513,-0.3540088832,-0.1421949714,-0.3390742242,0.1374310553,0.4490889013,-0.3026025891,-0.0341366641,0.1084006056,-0.039577052,0.1453768313,0.6637606025,-0.1132184714,0.1557540148,-0.2688517272,0.2169886976,0.4657373726,-0.2953676581,-0.4891296327,0.5142925382,-0.244742021,-0.0906946883,0.0965175927,0.3413774371,-0.0515389852,0.0072896979,0.280374229,0.2847390175,-0.0704721808,0.0136054931,-0.2741324902,-0.2110413462,0.1374817789,0.2579405606,0.0063828318,0.2574510872,-0.0397424847,0.1318580806,0.2551363111,-0.3740133047,0.1646961719,-0.1387715042,0.442397356,0.133719027,0.0754648671,0.1614752561,-0.5667373538,0.2011476904,-0.0271508321,0.1870721132,-0.2295346558,-0.2263766229,-0.3379972875,0.1709311455,-0.3567269146,0.2806505263,-0.0644999817,-0.1205384731,-0.0141854128,-0.0527514927,-0.119579725,0.217986241,-0.3730079234,0.1698513329,-0.513522923,0.3820358515,0.0151542323,-0.0033492115,-0.0094868299,0.1608085185,-0.0974757448,-0.1528729945,-0.1390003413,0.2045148015,-0.0318154059,0.0998232067,0.0785870627,0.1124184728,0.3132004738,-0.2474359721,0.0973865911,0.420461148,0.1018082201,-0.1078212857,-0.2806727886,0.6413202882,-0.265537709,0.1465586722,0.023377249,0.0997146517,0.1449086219,-0.0183355734,-0.156501472,-0.456420511,0.0755424201,-0.086596258,0.1723823696,-0.0415025093,-0.5107375383,0.1601004303,0.4665774107,0.0978005826,0.1503409594,0.3520168066,-0.0461252257,-0.0372990482,-0.2707844973,0.3472121358,0.4186730981,0.0066455025,-0.2218784839,0.0232526213,-0.3446961641,-0.2819954157,0.0617921017,-0.3365485668,-0.08019384,0.2138845026,0.0640868768,-0.0274222363,-0.3360664845,-0.0535807647,0.0640417561,0.35137555,-0.5140116215,-0.1834724396,-0.1738094836,-0.6247996092,-0.2197042555,0.0682751611,0.1395656765,-0.3413124979,0.0422627516,0.1681332439,-0.0272621587,-0.1199104786,0.0574179366,0.3083279431,0.1335732192,-0.1832105517,-0.0579218417,-0.1691828668,-0.1520417184,-0.0155690648,0.1158931032,-0.0518910177,0.3363802433,-0.0917729735,0.0576459654,-0.5386189222,-0.3667121828,0.0629966781,-0.1329158992,0.096696198,0.1896066815,0.4263233542,-0.1497290581,-0.3702141047,0.4545177221,-0.1328839958,-0.2982057333,0.210893482,-0.1885073185,0.0058833337,-0.1925978214,-0.5648372769,0.1293588877,-0.2554282844,0.1841124594,0.1010496616,0.2554558814,0.1573609859,0.0598342307,-0.0213592052,0.4327559769,-0.4605207741,-0.4601733088,-0.1789412349,0.3468564749,-0.3206701279,-0.4748046994,-0.1672707796,-0.0536935776,0.2448612005,0.1639297158,-0.3726163805,0.0362476744,-0.1461523473,0.6259244084,-0.0784577876,0.056387011,0.003346428,-0.0454089269,-0.0069139008,-0.1038952619,0.1286156476,0.3307235241,0.1364132166,0.1208571345,0.5113132,0.457803607,-0.0726824403,0.7307398319,0.3878905475,-0.0939111114,0.3839215636,-0.2607840598,0.0283578411,-0.1875331849,-0.0771837384,-0.0698270202,0.089350909,0.2247118801,0.075324662,-0.0563580878,0.2010137588,0.1043952703,0.3235479593,0.0841046497,-0.3629881442,0.2535820901,-0.2235082388,0.034705054,-0.002383945,-0.0588001497,-0.1934445351,-0.1966071874,0.1551623642,0.2092352509,0.2664136887,0.1493084431,-0.2545771599,0.0415279903,-0.5325753689,0.1423542053,-0.0342222489,0.4112564623,-0.1906432956,0.0005642162,0.1026364267,0.1655980945,0.4364417493,-0.2808634043,0.1209340319,-0.1033122987,0.0926193446,-0.3388012946,-0.0176090356,-0.1397709548,-0.2068332583,-0.1103481352,0.8588706255,-0.1347868294,-0.1666597575,0.0238158833,-0.0798937604,-0.3452742696,-0.3179731965,-0.1265582889,-0.2647454739,-0.341058284,-0.1245568916,-0.0742843747,0.0384419039,-0.0410095491,0.052635625,-0.1289285123,0.1312093735,0.2488941848,0.0797342658,0.1856541336,0.2549162507,0.2163953483,-0.1367268115,0.0458278842,0.1889093518,0.4144462645,-0.0076748924,-0.3783010244,-0.0580680333,-0.294980973,0.0499798395,0.1743704528,-0.0649333075,0.2576627731,-0.145915255,0.1320279837,-0.3990792334,0.0040973895,0.2709631026,-0.1203640029,-0.1829957962,-0.3075173497,0.555834651,-0.0454289131,-0.0611481406,0.3498488367,0.3161206841,-0.1454130411,0.6077867746,0.1557311863,0.7579222322,-0.2664358914,0.0856710821,0.4314405918,0.0131014222,0.688924551,-0.0332963541,-0.2075321525,-0.380461216,-0.2582828104,-0.1141497791,-0.0892563164,-0.1735268235,0.2237879336,-0.1204526722,0.2171323746,-0.1698041111,0.0164465662,0.0066384189,0.4152493179,-0.1245249584,-0.3207077682,-0.0214952994,0.056551192,-0.047884848,0.4819027781,-0.0044260956,-0.0755513012,-0.061983481,-0.2659850121,-0.3030018508,0.1360493004,-0.0069829435,0.2537105381,0.2293514609,-0.3981375098,0.2583366334,0.2784709632,0.6117494106,-0.1454027891,-0.0412950143,0.0058252877,0.1002017483,0.1814714223,-0.0793799758,0.0165113714,0.4528204501,0.0360200778,-0.1176940203,-0.0992383957,0.1585685909,-0.176034838,-0.4426099956,-0.1919261217,-0.3512182236,-0.2392002046,-0.1000855118,-0.1758451164,-0.0948926583,0.0353994444,0.0091581615,-0.0081356652,0.1550484002,-0.0343562476,0.1885125935,-0.2985710204,0.0654935539,0.0980413407,0.2255681306,-0.0593976155,0.5643968582,0.1239501759,0.0638021678,-0.1736727059,-0.0883423686,0.5117588639,-0.435198307,-0.0157871693,-0.1168153062,0.1396972984,0.1275131106,0.2093440294,0.2713354826,0.2161752433,0.0168284811,-0.422685951,-0.4231400788,0.1834543645,-0.2383051813,0.0536579341,-0.219431892,0.1762333512,0.0467203036,0.131757468,-0.2271301597,0.0118755801,0.0329937525,-0.0884260982,0.4423232079,0.1974610835,-0.0189120974,0.1438004524,0.0411593243,0.0292593688,-0.2791414559,-0.092680566,-0.29723683,0.1576782316,0.0006593319,0.0654929653,0.0041939737,-0.1479567587,-0.0172117837,-0.0138336169,0.3544194996,0.0950808823,-0.2435282022,0.3452404439,-0.1411787719,0.2481026351,-0.1061432213,0.1093548611,0.0990749374,0.457087934,-0.0064911167,0.0978495553,-0.0868633687,-0.057925459,-0.8441802859,0.1883462369,0.0923516974,-0.1496966928,0.0531364828,-0.3454817832,0.0581447966,-0.1585340798,0.0026861529,0.1437492073,-0.3473712504,-0.0282755364,0.0961244851,0.1619206816,-0.0917378962,-0.2351212054,-0.1227781549,0.0681864396,-0.0477737747,0.2856597006,0.189525038,0.0495388061,-0.2613171339,-0.2105496228,0.2116844803,-0.6142406464,0.0084571093,0.5428471565,-0.1165461093,-0.0191935934,0.0088783847,0.1817708015,-0.1159737632,0.4868438244,-0.0580273978,0.0777162388,0.0864651352,0.2649025023,0.2574900985,-0.1879974455,-0.0466611348,0.5726473331,-0.4538517296,0.1783535928,0.3731364608,0.5612061024,-0.0177509934,-0.1054079533,-0.2318201065,-0.0712556019,-0.2318294644,-0.2709780931,-0.4251090288,-0.1855171472,-0.3763929904,0.1644724458,-0.1437493861,0.0870296583,0.0379903764,0.2201379985,-0.0508414097,-0.3381506801,0.0129773626,0.2290822864,0.07972271,-0.2487999052,0.1437992305,0.1467214227,-0.1164534763,0.2759447396,0.129640907,0.4179956317,0.4494100511,0.149542883,0.2901523709,-0.0621726513,-0.0972077623,0.0316521339,0.0411583856,0.2337774485,-0.2043859512,0.3510064483,0.0440291837,-0.0471319854,-0.1450905651,0.1875878274,0.2525775135,-0.1295835972,0.3547744453,-0.209831804,-0.1167025045,0.0316069089,0.0820184499,-0.1671605855,0.0567485951,0.3998742104,0.3546751738,0.2771970928,-0.1219091713,0.051608216,-0.152956605,0.5715521574,0.276827693,-0.0432030708,-0.2789267302,-0.4268169999,-0.7540883422,0.1668043137,-0.0210878719,-0.1192840338,0.155103907,0.3425880969,0.1806433648,0.0964585766,-0.1303140372,-0.4154455066,-0.0724780187,0.4477213621,-0.177474767,-0.0388854556,0.0758356825,-0.0611814708,-0.1408303082,-0.4151671231,0.1325246692,0.2436689585,-0.0892501324,-0.3685547113,0.4689706564,0.1193311512,0.0884076953,0.1792702228,0.2448937893,0.3682719171,0.0684039816,-0.2043875009,-0.1018876582,-0.269026041,-0.3508322537,0.2523497045,0.1051632687,0.2737438381,0.1665946394,-0.2177688479,-0.3813337982,0.0766947344,0.1570960581,-0.4420635104,-0.3601760268,0.169743076,-0.1750987917,-0.1018458903,0.1536202878,0.3485894203,0.1436231732,0.1322574466,0.0354235433,-0.5088876486,0.1832260936,-0.3091132939,-0.243133232,-0.0341390073,0.2914814949,-0.1942309141,-0.063518934,-0.339112401,0.1603287011,0.1409399062,-0.2260430753,-0.3101964891,0.1264636517,0.0668042749,0.225973919,-0.2413260788,0.3000396192,0.137790069,-0.2491010427,0.1746743917,-0.2720994055]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2226","title":"Batched map fails when removing all columns","comments":"I found the problem. I called `set_format` on some columns before. This makes it crash. Here is a complete example to reproduce:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nsst = load_dataset(\"sst\")\r\nsst.set_format(\"torch\", columns=[\"label\"], output_all_columns=True)\r\nds = sst[\"train\"]\r\n\r\n# crashes\r\nds.map(\r\n lambda x: {\"a\": list(range(20))},\r\n remove_columns=ds.column_names,\r\n load_from_cache_file=False,\r\n num_proc=1,\r\n batched=True,\r\n)\r\n```","body":"Hi @lhoestq ,\r\n\r\nI'm hijacking this issue, because I'm currently trying to do the approach you recommend:\r\n\r\n> Currently the optimal setup for single-column computations is probably to do something like\r\n> \r\n> ```python\r\n> result = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n> ```\r\n\r\nHere is my code: (see edit, in which I added a simplified version\r\n\r\n```\r\nThis is the error:\r\n```bash\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000\r\n```\r\nI wonder why this error occurs, when I delete every column? Can you give me a hint?\r\n\r\n### Edit:\r\nI preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the\r\ncomplete dataset and print every sample before calling map. There seems to be no other problem with the dataset.\r\n\r\nI tried to simplify the code that crashes:\r\n\r\n```python\r\n# works\r\nlog.debug(dataset.column_names)\r\nlog.debug(dataset)\r\nfor i, sample in enumerate(dataset):\r\n log.debug(i, sample)\r\n\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n)\r\n```\r\n\r\n```\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000\r\n```\r\n\r\nEdit2: \r\n\r\nMay this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error:\r\n\r\n```python\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n features=datasets.Features(\r\n {\r\n \"a\": datasets.Sequence(datasets.Value(\"int32\"))\r\n }\r\n )\r\n)\r\n```\r\n\r\n```\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1704, in _map_single\r\n writer.write_batch(batch)\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 312, in write_batch\r\n col_type = schema.field(col).type if schema is not None else None\r\n File \"pyarrow\/types.pxi\", line 1341, in pyarrow.lib.Schema.field\r\nKeyError: 'Column tokens does not exist in schema'\r\n```\r\n\r\n_Originally posted by @villmow in https:\/\/github.com\/huggingface\/datasets\/issues\/2193#issuecomment-820230874_","comment_length":49,"text":"Batched map fails when removing all columns\nHi @lhoestq ,\r\n\r\nI'm hijacking this issue, because I'm currently trying to do the approach you recommend:\r\n\r\n> Currently the optimal setup for single-column computations is probably to do something like\r\n> \r\n> ```python\r\n> result = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n> ```\r\n\r\nHere is my code: (see edit, in which I added a simplified version\r\n\r\n```\r\nThis is the error:\r\n```bash\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000\r\n```\r\nI wonder why this error occurs, when I delete every column? Can you give me a hint?\r\n\r\n### Edit:\r\nI preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the\r\ncomplete dataset and print every sample before calling map. There seems to be no other problem with the dataset.\r\n\r\nI tried to simplify the code that crashes:\r\n\r\n```python\r\n# works\r\nlog.debug(dataset.column_names)\r\nlog.debug(dataset)\r\nfor i, sample in enumerate(dataset):\r\n log.debug(i, sample)\r\n\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n)\r\n```\r\n\r\n```\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000\r\n```\r\n\r\nEdit2: \r\n\r\nMay this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error:\r\n\r\n```python\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n features=datasets.Features(\r\n {\r\n \"a\": datasets.Sequence(datasets.Value(\"int32\"))\r\n }\r\n )\r\n)\r\n```\r\n\r\n```\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1704, in _map_single\r\n writer.write_batch(batch)\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 312, in write_batch\r\n col_type = schema.field(col).type if schema is not None else None\r\n File \"pyarrow\/types.pxi\", line 1341, in pyarrow.lib.Schema.field\r\nKeyError: 'Column tokens does not exist in schema'\r\n```\r\n\r\n_Originally posted by @villmow in https:\/\/github.com\/huggingface\/datasets\/issues\/2193#issuecomment-820230874_\nI found the problem. I called `set_format` on some columns before. This makes it crash. Here is a complete example to reproduce:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nsst = load_dataset(\"sst\")\r\nsst.set_format(\"torch\", columns=[\"label\"], output_all_columns=True)\r\nds = sst[\"train\"]\r\n\r\n# crashes\r\nds.map(\r\n lambda x: {\"a\": list(range(20))},\r\n remove_columns=ds.column_names,\r\n load_from_cache_file=False,\r\n num_proc=1,\r\n batched=True,\r\n)\r\n```","embeddings":[-0.1897988468,0.1186695099,0.0169567857,0.0345561393,0.2957544327,0.1997561753,0.7953292131,0.3469538689,0.2681218088,0.5083371997,0.1278548092,0.3986316621,-0.2221594751,-0.1497610509,-0.2561763227,-0.2258543968,0.180315271,0.1939368397,-0.1543451697,0.1138284206,-0.3259884417,-0.074356012,-0.5531980395,0.1137052774,0.0998652279,-0.281159699,-0.2459139675,-0.2607480586,0.0405273177,-0.3617533743,0.0900468752,-0.2502871752,0.1128109172,0.4839785695,-0.0001168795,0.0218302999,0.1716712415,-0.0334697515,-0.1326156855,0.0640774146,-0.1687964052,-0.1176634431,-0.0496759899,-0.3645480275,0.5358587503,-0.2108844817,-0.3315514624,-0.381175071,-0.0405727848,0.1007553786,0.169487834,0.1469649225,-0.0236559641,0.1243837774,0.1997407079,-0.0343382359,-0.1041169465,-0.008293177,0.4995136261,-0.4657591283,-0.0907840803,0.2961730361,-0.2830587029,-0.0230159666,0.1004072875,-0.0464141592,0.2637531757,-0.4090450108,0.1707249731,0.1164800003,-0.0180447102,-0.3283998966,-0.0548055284,-0.184510991,-0.1374211758,-0.5785285234,-0.1803824157,0.0719124228,-0.2919106185,-0.1150881946,-0.457372129,0.159992218,0.0326213799,0.1772422045,-0.0578017868,0.3028842211,0.151088357,0.5388099551,0.0138682658,-0.0603085943,-0.0003699868,-0.0276932251,0.0329861157,0.367708981,-0.493185699,-0.1327268332,-0.0320847854,-0.1082744449,0.2840714455,-0.2979654074,0.094810091,-0.0292264242,0.5917070508,0.1859064102,0.4440813661,0.2730583549,0.021104537,0.1322196424,0.0927797928,-0.053203471,0.0608418174,-0.0009549259,0.2914585471,-0.081994772,0.5020079613,0.2725434303,0.0259534437,-0.0072190715,-0.075207144,-0.158616364,-0.3384186327,0.204627499,-0.0911616161,0.0677755699,0.2737882435,0.0878871754,-0.3580185771,0.1353775859,-0.0966755003,-0.1938144863,-0.1133242175,0.1158341393,-0.3451522887,-0.1509786099,0.3529440463,0.1030584052,0.1164731979,-0.0047300588,-0.010748487,-0.1193328947,-0.2432302088,-0.1289957613,0.2258381397,0.02893872,0.1047094688,0.1680702269,0.2093173712,0.033626534,-0.0138567258,0.2573041618,-0.0536132939,-0.3020163178,-0.1343097836,0.1198306754,0.0918057412,0.2040550411,-0.3438605666,0.0384183526,0.4002182484,0.0612391569,0.0008474094,-0.290682137,0.220656991,-0.2849721313,-0.1069678515,0.2588521838,-0.5109861493,0.158715412,0.0789557025,0.0771709159,0.2378589064,0.2749264538,-0.2287099212,0.1179107651,-0.1785375029,0.287909776,0.1158813164,-0.121906437,-0.1291884631,0.2582710683,-0.1710102111,0.0061789071,-0.2403122038,-0.183501333,0.6115528941,-0.2941825986,0.1319783777,-0.0414869897,-0.3053473234,0.1779927462,-0.1582277268,-0.2711120248,-0.0471705981,-0.0894282609,0.1199706793,-0.01697715,0.302390933,-0.451308012,0.2286178917,-0.3265037835,0.339217186,0.4150714874,0.0918067545,-0.035193719,0.166398108,-0.1615534425,-0.7281162143,0.1981071383,0.1456974298,-0.0879978985,-0.3534245789,-0.2648589909,0.0765319467,0.1501424313,0.0226836484,0.1255517751,0.0711953044,-0.2024617195,0.0510789268,-0.3482041061,-0.2364073396,-0.2209415287,-0.1390867531,0.1733426154,-0.1894963533,-0.3441658318,-0.0279753543,-0.3571982086,-0.0204043388,0.1966486275,0.239356637,-0.1307834387,-0.0186508615,0.4590831697,-0.1979596764,-0.0776889548,-0.4247601926,-0.0810873657,-0.0454146862,-0.0406259447,-0.0342591852,-0.1599503905,0.142451182,-0.144739449,-0.3011455238,0.0029289648,-0.2381207794,0.5006266832,-0.1498275399,-0.0373831503,-0.0561359078,0.0231335517,-0.0771553442,-0.153137818,-0.1096757501,-0.0489469655,-0.2542680204,-0.0305233877,0.0260866899,-0.1773487478,0.221904546,-0.0419845395,0.0335538946,-0.0073624742,-0.2589181364,0.0027919528,0.339520216,-0.0899918675,0.4603516459,0.1365824491,0.0819970593,-0.0418755561,0.1916162968,0.0838154778,0.1106001586,0.2706128061,0.0925673023,0.0756221414,0.3986678123,0.0161472093,-0.2661878169,-0.351077348,0.1753147542,0.4957439005,-0.2044825107,-0.043800015,-0.0830236226,0.0995051414,0.2236653417,-0.1830222905,0.0713529959,-0.4394610822,-0.1476138383,0.1156395972,-0.1264729351,0.2542539537,-0.1595221162,-0.0608217195,0.1876654774,-0.0264898054,-0.1460788846,-0.1495200843,-0.1710772216,0.0146625154,-0.03508294,-0.144578144,0.0407584049,0.4789995551,-0.3435450494,-0.2255327553,-0.0635434538,-0.0063208234,-0.5026287436,-0.0688748881,0.3389750123,0.2059350759,-0.3035939336,-0.1560991108,0.0831781253,-0.1239539161,0.0785744712,0.2210019976,0.1313579977,-0.1047570184,-0.2336309254,0.1117874458,0.0731343701,-0.2843957543,0.1609170586,-0.1200424731,0.2389975786,-0.2521990538,0.2278801054,-0.2472717613,-0.0553630665,-0.40374735,0.0974972919,-0.010698528,0.1542848647,0.1120021343,-0.0112327524,0.0451950282,-0.0525667556,0.032991603,0.3330432177,-0.3355278969,0.046339605,-0.1026915088,0.135222882,-0.0032963625,0.232637465,0.3257702589,0.2052081823,0.0374394618,-0.0169250686,-0.164213717,0.0817854702,-0.0682317913,-0.003880508,0.0999058709,0.4984065294,0.2369377017,0.5979442,0.204280436,-0.0856824219,0.0948240235,-0.1327353269,0.0399422608,-0.1180508658,-0.3469358385,0.0949326307,-0.3334673345,-0.0153741492,0.0916314721,-0.142918691,-0.5661892295,0.0262778308,0.3182716966,0.0356097259,-0.2705049813,0.2294884175,-0.242961511,0.2372483611,0.1168125421,0.0199956484,-0.2669150531,-0.1107122079,0.2998273075,-0.3833626807,0.1725095212,-0.2118994147,-0.3650291264,-0.0065127322,-0.7440903187,0.4507893026,0.0828428343,0.3523317575,0.1741855294,-0.0121682016,0.0209380426,0.1607970893,0.4166469872,-0.2360412925,-0.1881221533,0.1594663262,0.1826755702,-0.4863575101,-0.057204552,-0.3145472705,0.1810837686,0.2628193498,0.8223790526,-0.3449789286,-0.0785233304,0.3561532199,0.1908120662,0.081973806,-0.1069040298,-0.1756309569,-0.2924115956,-0.1121415347,0.1398420483,0.3096801043,0.1881920844,0.1740519851,-0.2482351512,0.0171870999,-0.1449332535,0.1038188264,0.0200537276,-0.0551291481,-0.1025983393,0.0748112649,-0.1875823587,0.019703459,0.09852276,0.1518204361,-0.2780478299,-0.1183985844,0.0980392471,-0.1654563546,0.5431213975,0.242498979,0.0117663573,0.1635087579,-0.3425570726,0.3598184288,0.1224122271,-0.4356709719,0.3494389653,0.0364778265,-0.3460505009,0.1196431071,0.2144177407,0.2063299268,-0.03236733,0.7606868744,0.0132842269,-0.2412766218,0.7118184566,0.1966302842,0.6759189367,-0.1432673782,-0.0712213963,0.290207535,0.360881716,0.2842663527,0.1126970053,0.381161958,-0.3241282701,-0.0603737421,0.00332267,-0.0351574942,0.1903509349,0.4072134197,0.0292364396,0.3826394081,-0.1521288007,0.4469073713,-0.0252744313,-0.1773910522,0.1236897781,-0.3599649966,0.0880283713,0.0208682176,0.046122279,-0.1213788539,-0.0826372132,0.0984830633,0.0594394431,-0.3395923972,-0.1458568722,-0.0059602247,-0.2831459045,0.2795832455,-0.0178861022,-0.023591958,-0.0276551321,0.1857581586,-0.1375504732,0.0231043696,0.2528838813,-0.0234127026,0.5783449411,0.1532377601,0.0718974993,0.0801609978,-0.3442005217,0.0582830235,-0.0581079088,0.1441702545,-0.2681709528,-0.2917725146,-0.3978101015,0.302022934,0.2278286517,-0.4201090634,-0.174985975,-0.1819713861,-0.028987037,-0.2493113279,0.0179118775,-0.112072818,-0.1037063524,0.7665064931,-0.1971287727,-0.2018837929,0.0001165,0.054283157,-0.0103726638,-0.0073020007,0.509395957,0.1933065951,-0.1906852573,-0.1447668821,-0.1147615388,0.2298499942,-0.4227213264,0.1481416821,-0.3345050812,0.3107252121,0.137876451,0.3870111406,-0.0561545715,0.1035897061,-0.0587295517,-0.1376464814,-0.3454743624,0.2268033475,0.2731829584,0.2479774058,0.009982734,0.0238754451,-0.2870259881,0.1395941824,-0.2310718894,0.1292947829,0.0296753552,0.4705128074,0.2248261571,0.1841877252,-0.063720718,0.1222635955,0.1711612195,0.3236610889,0.2282157838,-0.1890603006,0.087144509,0.0845169425,0.0414255783,-0.0440464094,0.0407923609,-0.1662178338,-0.2258954048,-0.2493456602,0.2834926546,0.2077129483,-0.1348534673,-0.051100973,0.2674014568,0.0906685889,-0.0091010891,0.1165961102,-0.2273985595,-0.082839407,0.1396965832,0.3365561068,0.1446776241,-0.0922280326,0.0380749702,0.0792174712,-0.1854953766,0.1861371994,0.2741543055,-0.3076931536,-0.0741831362,-0.0205289815,0.3571772575,0.1077172607,-0.265585959,-0.1578714103,0.094328694,0.1932017356,-0.0930128843,0.0147829317,0.3097126186,0.3044721186,0.2403047383,0.4139226079,0.0311292689,-0.4228920341,-0.1355920285,-0.0868847966,0.5410777926,-0.1324590594,-0.2924458683,0.6839470863,-0.148261413,-0.0108291414,0.2884373665,0.1919950396,0.3235805035,0.5047323108,-0.0714931935,0.0853694305,0.3333715498,0.104234606,-0.2822675109,-0.4482128918,0.2085962892,0.2343000472,0.0912758559,0.1438355148,0.1783404946,0.0890338719,-0.0002732429,-0.216752544,0.2180004567,0.2582809031,-0.3013063371,-0.025725428,-0.364698261,-0.0765605271,0.1926115453,-0.0516907535,-0.1579218358,-0.1624597609,0.5937283039,0.0620205924,-0.0564188994,-0.3340180814,-0.1528503001,-0.0166738778,0.3545717895,-0.1517580003,0.1551139057,0.0096705584,0.0895556584,-0.2272724062,0.0879208595,0.4263053536,0.3567664027,0.0829301924,0.0477266125,-0.1091136932,-0.1831457168,-0.0300317425,0.1075926647,0.2164026946,0.3985658288,0.3411513269,0.0027499029,-0.0674818233,-0.0332265981,0.3050508499,0.1356130689,-0.0164149981,-0.033591345,0.0403946526,-0.4961536229,-0.1215939075,-0.2080625743,-0.0198331699,-0.2020421326,0.3985048234,-0.1470507532,0.2387565076,0.0221792795,0.0588729046,-0.0984587073,0.0365584604,0.4919873178,-0.3134085834,-0.549947083,-0.205067113,-0.3719426394,-0.0011593878,-0.7319380045,-0.0049579004,0.2208574563,0.1944051087,0.1252499968,0.0747248083,0.0763463005,0.1676183045,-0.2411892414,0.8675721288,-0.3371376693,0.0864991546,-0.0816274583,0.0118992459,0.0936869979,-0.5087841749,0.4204275906,0.219004184,0.0183358509,-0.1565003097,-0.014172188,0.1934941858,-0.1132708788,0.3862443864,0.0413534828,0.0781790391,0.185065344,-0.3118218482,-0.3471717834,0.0969083011,-0.1761681736,0.1207804456,0.21304138,0.1897467226,-0.058404211,0.0200456567,-0.0797780529,0.054864984,-0.0549621284,0.0551304221,-0.1536683887,0.1488116533,0.0846773535,0.1772117466,0.0675943792,0.0562539361,0.1116035134,0.2221062779,-0.3874033988,-0.3115005195,0.4752250314,-0.7607836127,-0.6713259816,-0.0875397548,0.0665985942,0.0903250277,-0.1569895744,-0.3100399077,-0.0200973954,0.262804538,-0.1931921691,-0.1906295866,-0.005261525,-0.3285938501,0.0072535705,-0.0165186673,0.0777009577,0.0896192864,-0.1323938072,-0.0028246206,-0.3992703259]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2226","title":"Batched map fails when removing all columns","comments":"Thanks for reporting and for providing this code to reproduce the issue, this is really helpful !","body":"Hi @lhoestq ,\r\n\r\nI'm hijacking this issue, because I'm currently trying to do the approach you recommend:\r\n\r\n> Currently the optimal setup for single-column computations is probably to do something like\r\n> \r\n> ```python\r\n> result = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n> ```\r\n\r\nHere is my code: (see edit, in which I added a simplified version\r\n\r\n```\r\nThis is the error:\r\n```bash\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000\r\n```\r\nI wonder why this error occurs, when I delete every column? Can you give me a hint?\r\n\r\n### Edit:\r\nI preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the\r\ncomplete dataset and print every sample before calling map. There seems to be no other problem with the dataset.\r\n\r\nI tried to simplify the code that crashes:\r\n\r\n```python\r\n# works\r\nlog.debug(dataset.column_names)\r\nlog.debug(dataset)\r\nfor i, sample in enumerate(dataset):\r\n log.debug(i, sample)\r\n\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n)\r\n```\r\n\r\n```\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000\r\n```\r\n\r\nEdit2: \r\n\r\nMay this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error:\r\n\r\n```python\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n features=datasets.Features(\r\n {\r\n \"a\": datasets.Sequence(datasets.Value(\"int32\"))\r\n }\r\n )\r\n)\r\n```\r\n\r\n```\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1704, in _map_single\r\n writer.write_batch(batch)\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 312, in write_batch\r\n col_type = schema.field(col).type if schema is not None else None\r\n File \"pyarrow\/types.pxi\", line 1341, in pyarrow.lib.Schema.field\r\nKeyError: 'Column tokens does not exist in schema'\r\n```\r\n\r\n_Originally posted by @villmow in https:\/\/github.com\/huggingface\/datasets\/issues\/2193#issuecomment-820230874_","comment_length":17,"text":"Batched map fails when removing all columns\nHi @lhoestq ,\r\n\r\nI'm hijacking this issue, because I'm currently trying to do the approach you recommend:\r\n\r\n> Currently the optimal setup for single-column computations is probably to do something like\r\n> \r\n> ```python\r\n> result = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n> ```\r\n\r\nHere is my code: (see edit, in which I added a simplified version\r\n\r\n```\r\nThis is the error:\r\n```bash\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000\r\n```\r\nI wonder why this error occurs, when I delete every column? Can you give me a hint?\r\n\r\n### Edit:\r\nI preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the\r\ncomplete dataset and print every sample before calling map. There seems to be no other problem with the dataset.\r\n\r\nI tried to simplify the code that crashes:\r\n\r\n```python\r\n# works\r\nlog.debug(dataset.column_names)\r\nlog.debug(dataset)\r\nfor i, sample in enumerate(dataset):\r\n log.debug(i, sample)\r\n\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n)\r\n```\r\n\r\n```\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000\r\n```\r\n\r\nEdit2: \r\n\r\nMay this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error:\r\n\r\n```python\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n features=datasets.Features(\r\n {\r\n \"a\": datasets.Sequence(datasets.Value(\"int32\"))\r\n }\r\n )\r\n)\r\n```\r\n\r\n```\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1704, in _map_single\r\n writer.write_batch(batch)\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 312, in write_batch\r\n col_type = schema.field(col).type if schema is not None else None\r\n File \"pyarrow\/types.pxi\", line 1341, in pyarrow.lib.Schema.field\r\nKeyError: 'Column tokens does not exist in schema'\r\n```\r\n\r\n_Originally posted by @villmow in https:\/\/github.com\/huggingface\/datasets\/issues\/2193#issuecomment-820230874_\nThanks for reporting and for providing this code to reproduce the issue, this is really helpful !","embeddings":[-0.1897988468,0.1186695099,0.0169567857,0.0345561393,0.2957544327,0.1997561753,0.7953292131,0.3469538689,0.2681218088,0.5083371997,0.1278548092,0.3986316621,-0.2221594751,-0.1497610509,-0.2561763227,-0.2258543968,0.180315271,0.1939368397,-0.1543451697,0.1138284206,-0.3259884417,-0.074356012,-0.5531980395,0.1137052774,0.0998652279,-0.281159699,-0.2459139675,-0.2607480586,0.0405273177,-0.3617533743,0.0900468752,-0.2502871752,0.1128109172,0.4839785695,-0.0001168795,0.0218302999,0.1716712415,-0.0334697515,-0.1326156855,0.0640774146,-0.1687964052,-0.1176634431,-0.0496759899,-0.3645480275,0.5358587503,-0.2108844817,-0.3315514624,-0.381175071,-0.0405727848,0.1007553786,0.169487834,0.1469649225,-0.0236559641,0.1243837774,0.1997407079,-0.0343382359,-0.1041169465,-0.008293177,0.4995136261,-0.4657591283,-0.0907840803,0.2961730361,-0.2830587029,-0.0230159666,0.1004072875,-0.0464141592,0.2637531757,-0.4090450108,0.1707249731,0.1164800003,-0.0180447102,-0.3283998966,-0.0548055284,-0.184510991,-0.1374211758,-0.5785285234,-0.1803824157,0.0719124228,-0.2919106185,-0.1150881946,-0.457372129,0.159992218,0.0326213799,0.1772422045,-0.0578017868,0.3028842211,0.151088357,0.5388099551,0.0138682658,-0.0603085943,-0.0003699868,-0.0276932251,0.0329861157,0.367708981,-0.493185699,-0.1327268332,-0.0320847854,-0.1082744449,0.2840714455,-0.2979654074,0.094810091,-0.0292264242,0.5917070508,0.1859064102,0.4440813661,0.2730583549,0.021104537,0.1322196424,0.0927797928,-0.053203471,0.0608418174,-0.0009549259,0.2914585471,-0.081994772,0.5020079613,0.2725434303,0.0259534437,-0.0072190715,-0.075207144,-0.158616364,-0.3384186327,0.204627499,-0.0911616161,0.0677755699,0.2737882435,0.0878871754,-0.3580185771,0.1353775859,-0.0966755003,-0.1938144863,-0.1133242175,0.1158341393,-0.3451522887,-0.1509786099,0.3529440463,0.1030584052,0.1164731979,-0.0047300588,-0.010748487,-0.1193328947,-0.2432302088,-0.1289957613,0.2258381397,0.02893872,0.1047094688,0.1680702269,0.2093173712,0.033626534,-0.0138567258,0.2573041618,-0.0536132939,-0.3020163178,-0.1343097836,0.1198306754,0.0918057412,0.2040550411,-0.3438605666,0.0384183526,0.4002182484,0.0612391569,0.0008474094,-0.290682137,0.220656991,-0.2849721313,-0.1069678515,0.2588521838,-0.5109861493,0.158715412,0.0789557025,0.0771709159,0.2378589064,0.2749264538,-0.2287099212,0.1179107651,-0.1785375029,0.287909776,0.1158813164,-0.121906437,-0.1291884631,0.2582710683,-0.1710102111,0.0061789071,-0.2403122038,-0.183501333,0.6115528941,-0.2941825986,0.1319783777,-0.0414869897,-0.3053473234,0.1779927462,-0.1582277268,-0.2711120248,-0.0471705981,-0.0894282609,0.1199706793,-0.01697715,0.302390933,-0.451308012,0.2286178917,-0.3265037835,0.339217186,0.4150714874,0.0918067545,-0.035193719,0.166398108,-0.1615534425,-0.7281162143,0.1981071383,0.1456974298,-0.0879978985,-0.3534245789,-0.2648589909,0.0765319467,0.1501424313,0.0226836484,0.1255517751,0.0711953044,-0.2024617195,0.0510789268,-0.3482041061,-0.2364073396,-0.2209415287,-0.1390867531,0.1733426154,-0.1894963533,-0.3441658318,-0.0279753543,-0.3571982086,-0.0204043388,0.1966486275,0.239356637,-0.1307834387,-0.0186508615,0.4590831697,-0.1979596764,-0.0776889548,-0.4247601926,-0.0810873657,-0.0454146862,-0.0406259447,-0.0342591852,-0.1599503905,0.142451182,-0.144739449,-0.3011455238,0.0029289648,-0.2381207794,0.5006266832,-0.1498275399,-0.0373831503,-0.0561359078,0.0231335517,-0.0771553442,-0.153137818,-0.1096757501,-0.0489469655,-0.2542680204,-0.0305233877,0.0260866899,-0.1773487478,0.221904546,-0.0419845395,0.0335538946,-0.0073624742,-0.2589181364,0.0027919528,0.339520216,-0.0899918675,0.4603516459,0.1365824491,0.0819970593,-0.0418755561,0.1916162968,0.0838154778,0.1106001586,0.2706128061,0.0925673023,0.0756221414,0.3986678123,0.0161472093,-0.2661878169,-0.351077348,0.1753147542,0.4957439005,-0.2044825107,-0.043800015,-0.0830236226,0.0995051414,0.2236653417,-0.1830222905,0.0713529959,-0.4394610822,-0.1476138383,0.1156395972,-0.1264729351,0.2542539537,-0.1595221162,-0.0608217195,0.1876654774,-0.0264898054,-0.1460788846,-0.1495200843,-0.1710772216,0.0146625154,-0.03508294,-0.144578144,0.0407584049,0.4789995551,-0.3435450494,-0.2255327553,-0.0635434538,-0.0063208234,-0.5026287436,-0.0688748881,0.3389750123,0.2059350759,-0.3035939336,-0.1560991108,0.0831781253,-0.1239539161,0.0785744712,0.2210019976,0.1313579977,-0.1047570184,-0.2336309254,0.1117874458,0.0731343701,-0.2843957543,0.1609170586,-0.1200424731,0.2389975786,-0.2521990538,0.2278801054,-0.2472717613,-0.0553630665,-0.40374735,0.0974972919,-0.010698528,0.1542848647,0.1120021343,-0.0112327524,0.0451950282,-0.0525667556,0.032991603,0.3330432177,-0.3355278969,0.046339605,-0.1026915088,0.135222882,-0.0032963625,0.232637465,0.3257702589,0.2052081823,0.0374394618,-0.0169250686,-0.164213717,0.0817854702,-0.0682317913,-0.003880508,0.0999058709,0.4984065294,0.2369377017,0.5979442,0.204280436,-0.0856824219,0.0948240235,-0.1327353269,0.0399422608,-0.1180508658,-0.3469358385,0.0949326307,-0.3334673345,-0.0153741492,0.0916314721,-0.142918691,-0.5661892295,0.0262778308,0.3182716966,0.0356097259,-0.2705049813,0.2294884175,-0.242961511,0.2372483611,0.1168125421,0.0199956484,-0.2669150531,-0.1107122079,0.2998273075,-0.3833626807,0.1725095212,-0.2118994147,-0.3650291264,-0.0065127322,-0.7440903187,0.4507893026,0.0828428343,0.3523317575,0.1741855294,-0.0121682016,0.0209380426,0.1607970893,0.4166469872,-0.2360412925,-0.1881221533,0.1594663262,0.1826755702,-0.4863575101,-0.057204552,-0.3145472705,0.1810837686,0.2628193498,0.8223790526,-0.3449789286,-0.0785233304,0.3561532199,0.1908120662,0.081973806,-0.1069040298,-0.1756309569,-0.2924115956,-0.1121415347,0.1398420483,0.3096801043,0.1881920844,0.1740519851,-0.2482351512,0.0171870999,-0.1449332535,0.1038188264,0.0200537276,-0.0551291481,-0.1025983393,0.0748112649,-0.1875823587,0.019703459,0.09852276,0.1518204361,-0.2780478299,-0.1183985844,0.0980392471,-0.1654563546,0.5431213975,0.242498979,0.0117663573,0.1635087579,-0.3425570726,0.3598184288,0.1224122271,-0.4356709719,0.3494389653,0.0364778265,-0.3460505009,0.1196431071,0.2144177407,0.2063299268,-0.03236733,0.7606868744,0.0132842269,-0.2412766218,0.7118184566,0.1966302842,0.6759189367,-0.1432673782,-0.0712213963,0.290207535,0.360881716,0.2842663527,0.1126970053,0.381161958,-0.3241282701,-0.0603737421,0.00332267,-0.0351574942,0.1903509349,0.4072134197,0.0292364396,0.3826394081,-0.1521288007,0.4469073713,-0.0252744313,-0.1773910522,0.1236897781,-0.3599649966,0.0880283713,0.0208682176,0.046122279,-0.1213788539,-0.0826372132,0.0984830633,0.0594394431,-0.3395923972,-0.1458568722,-0.0059602247,-0.2831459045,0.2795832455,-0.0178861022,-0.023591958,-0.0276551321,0.1857581586,-0.1375504732,0.0231043696,0.2528838813,-0.0234127026,0.5783449411,0.1532377601,0.0718974993,0.0801609978,-0.3442005217,0.0582830235,-0.0581079088,0.1441702545,-0.2681709528,-0.2917725146,-0.3978101015,0.302022934,0.2278286517,-0.4201090634,-0.174985975,-0.1819713861,-0.028987037,-0.2493113279,0.0179118775,-0.112072818,-0.1037063524,0.7665064931,-0.1971287727,-0.2018837929,0.0001165,0.054283157,-0.0103726638,-0.0073020007,0.509395957,0.1933065951,-0.1906852573,-0.1447668821,-0.1147615388,0.2298499942,-0.4227213264,0.1481416821,-0.3345050812,0.3107252121,0.137876451,0.3870111406,-0.0561545715,0.1035897061,-0.0587295517,-0.1376464814,-0.3454743624,0.2268033475,0.2731829584,0.2479774058,0.009982734,0.0238754451,-0.2870259881,0.1395941824,-0.2310718894,0.1292947829,0.0296753552,0.4705128074,0.2248261571,0.1841877252,-0.063720718,0.1222635955,0.1711612195,0.3236610889,0.2282157838,-0.1890603006,0.087144509,0.0845169425,0.0414255783,-0.0440464094,0.0407923609,-0.1662178338,-0.2258954048,-0.2493456602,0.2834926546,0.2077129483,-0.1348534673,-0.051100973,0.2674014568,0.0906685889,-0.0091010891,0.1165961102,-0.2273985595,-0.082839407,0.1396965832,0.3365561068,0.1446776241,-0.0922280326,0.0380749702,0.0792174712,-0.1854953766,0.1861371994,0.2741543055,-0.3076931536,-0.0741831362,-0.0205289815,0.3571772575,0.1077172607,-0.265585959,-0.1578714103,0.094328694,0.1932017356,-0.0930128843,0.0147829317,0.3097126186,0.3044721186,0.2403047383,0.4139226079,0.0311292689,-0.4228920341,-0.1355920285,-0.0868847966,0.5410777926,-0.1324590594,-0.2924458683,0.6839470863,-0.148261413,-0.0108291414,0.2884373665,0.1919950396,0.3235805035,0.5047323108,-0.0714931935,0.0853694305,0.3333715498,0.104234606,-0.2822675109,-0.4482128918,0.2085962892,0.2343000472,0.0912758559,0.1438355148,0.1783404946,0.0890338719,-0.0002732429,-0.216752544,0.2180004567,0.2582809031,-0.3013063371,-0.025725428,-0.364698261,-0.0765605271,0.1926115453,-0.0516907535,-0.1579218358,-0.1624597609,0.5937283039,0.0620205924,-0.0564188994,-0.3340180814,-0.1528503001,-0.0166738778,0.3545717895,-0.1517580003,0.1551139057,0.0096705584,0.0895556584,-0.2272724062,0.0879208595,0.4263053536,0.3567664027,0.0829301924,0.0477266125,-0.1091136932,-0.1831457168,-0.0300317425,0.1075926647,0.2164026946,0.3985658288,0.3411513269,0.0027499029,-0.0674818233,-0.0332265981,0.3050508499,0.1356130689,-0.0164149981,-0.033591345,0.0403946526,-0.4961536229,-0.1215939075,-0.2080625743,-0.0198331699,-0.2020421326,0.3985048234,-0.1470507532,0.2387565076,0.0221792795,0.0588729046,-0.0984587073,0.0365584604,0.4919873178,-0.3134085834,-0.549947083,-0.205067113,-0.3719426394,-0.0011593878,-0.7319380045,-0.0049579004,0.2208574563,0.1944051087,0.1252499968,0.0747248083,0.0763463005,0.1676183045,-0.2411892414,0.8675721288,-0.3371376693,0.0864991546,-0.0816274583,0.0118992459,0.0936869979,-0.5087841749,0.4204275906,0.219004184,0.0183358509,-0.1565003097,-0.014172188,0.1934941858,-0.1132708788,0.3862443864,0.0413534828,0.0781790391,0.185065344,-0.3118218482,-0.3471717834,0.0969083011,-0.1761681736,0.1207804456,0.21304138,0.1897467226,-0.058404211,0.0200456567,-0.0797780529,0.054864984,-0.0549621284,0.0551304221,-0.1536683887,0.1488116533,0.0846773535,0.1772117466,0.0675943792,0.0562539361,0.1116035134,0.2221062779,-0.3874033988,-0.3115005195,0.4752250314,-0.7607836127,-0.6713259816,-0.0875397548,0.0665985942,0.0903250277,-0.1569895744,-0.3100399077,-0.0200973954,0.262804538,-0.1931921691,-0.1906295866,-0.005261525,-0.3285938501,0.0072535705,-0.0165186673,0.0777009577,0.0896192864,-0.1323938072,-0.0028246206,-0.3992703259]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2226","title":"Batched map fails when removing all columns","comments":"I merged a fix, it should work on `master` now :)\r\nWe'll do a new release soon !","body":"Hi @lhoestq ,\r\n\r\nI'm hijacking this issue, because I'm currently trying to do the approach you recommend:\r\n\r\n> Currently the optimal setup for single-column computations is probably to do something like\r\n> \r\n> ```python\r\n> result = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n> ```\r\n\r\nHere is my code: (see edit, in which I added a simplified version\r\n\r\n```\r\nThis is the error:\r\n```bash\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000\r\n```\r\nI wonder why this error occurs, when I delete every column? Can you give me a hint?\r\n\r\n### Edit:\r\nI preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the\r\ncomplete dataset and print every sample before calling map. There seems to be no other problem with the dataset.\r\n\r\nI tried to simplify the code that crashes:\r\n\r\n```python\r\n# works\r\nlog.debug(dataset.column_names)\r\nlog.debug(dataset)\r\nfor i, sample in enumerate(dataset):\r\n log.debug(i, sample)\r\n\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n)\r\n```\r\n\r\n```\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000\r\n```\r\n\r\nEdit2: \r\n\r\nMay this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error:\r\n\r\n```python\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n features=datasets.Features(\r\n {\r\n \"a\": datasets.Sequence(datasets.Value(\"int32\"))\r\n }\r\n )\r\n)\r\n```\r\n\r\n```\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1704, in _map_single\r\n writer.write_batch(batch)\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 312, in write_batch\r\n col_type = schema.field(col).type if schema is not None else None\r\n File \"pyarrow\/types.pxi\", line 1341, in pyarrow.lib.Schema.field\r\nKeyError: 'Column tokens does not exist in schema'\r\n```\r\n\r\n_Originally posted by @villmow in https:\/\/github.com\/huggingface\/datasets\/issues\/2193#issuecomment-820230874_","comment_length":18,"text":"Batched map fails when removing all columns\nHi @lhoestq ,\r\n\r\nI'm hijacking this issue, because I'm currently trying to do the approach you recommend:\r\n\r\n> Currently the optimal setup for single-column computations is probably to do something like\r\n> \r\n> ```python\r\n> result = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n> ```\r\n\r\nHere is my code: (see edit, in which I added a simplified version\r\n\r\n```\r\nThis is the error:\r\n```bash\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000\r\n```\r\nI wonder why this error occurs, when I delete every column? Can you give me a hint?\r\n\r\n### Edit:\r\nI preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the\r\ncomplete dataset and print every sample before calling map. There seems to be no other problem with the dataset.\r\n\r\nI tried to simplify the code that crashes:\r\n\r\n```python\r\n# works\r\nlog.debug(dataset.column_names)\r\nlog.debug(dataset)\r\nfor i, sample in enumerate(dataset):\r\n log.debug(i, sample)\r\n\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n)\r\n```\r\n\r\n```\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000\r\n```\r\n\r\nEdit2: \r\n\r\nMay this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error:\r\n\r\n```python\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n features=datasets.Features(\r\n {\r\n \"a\": datasets.Sequence(datasets.Value(\"int32\"))\r\n }\r\n )\r\n)\r\n```\r\n\r\n```\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1704, in _map_single\r\n writer.write_batch(batch)\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 312, in write_batch\r\n col_type = schema.field(col).type if schema is not None else None\r\n File \"pyarrow\/types.pxi\", line 1341, in pyarrow.lib.Schema.field\r\nKeyError: 'Column tokens does not exist in schema'\r\n```\r\n\r\n_Originally posted by @villmow in https:\/\/github.com\/huggingface\/datasets\/issues\/2193#issuecomment-820230874_\nI merged a fix, it should work on `master` now :)\r\nWe'll do a new release soon !","embeddings":[-0.1897988468,0.1186695099,0.0169567857,0.0345561393,0.2957544327,0.1997561753,0.7953292131,0.3469538689,0.2681218088,0.5083371997,0.1278548092,0.3986316621,-0.2221594751,-0.1497610509,-0.2561763227,-0.2258543968,0.180315271,0.1939368397,-0.1543451697,0.1138284206,-0.3259884417,-0.074356012,-0.5531980395,0.1137052774,0.0998652279,-0.281159699,-0.2459139675,-0.2607480586,0.0405273177,-0.3617533743,0.0900468752,-0.2502871752,0.1128109172,0.4839785695,-0.0001168795,0.0218302999,0.1716712415,-0.0334697515,-0.1326156855,0.0640774146,-0.1687964052,-0.1176634431,-0.0496759899,-0.3645480275,0.5358587503,-0.2108844817,-0.3315514624,-0.381175071,-0.0405727848,0.1007553786,0.169487834,0.1469649225,-0.0236559641,0.1243837774,0.1997407079,-0.0343382359,-0.1041169465,-0.008293177,0.4995136261,-0.4657591283,-0.0907840803,0.2961730361,-0.2830587029,-0.0230159666,0.1004072875,-0.0464141592,0.2637531757,-0.4090450108,0.1707249731,0.1164800003,-0.0180447102,-0.3283998966,-0.0548055284,-0.184510991,-0.1374211758,-0.5785285234,-0.1803824157,0.0719124228,-0.2919106185,-0.1150881946,-0.457372129,0.159992218,0.0326213799,0.1772422045,-0.0578017868,0.3028842211,0.151088357,0.5388099551,0.0138682658,-0.0603085943,-0.0003699868,-0.0276932251,0.0329861157,0.367708981,-0.493185699,-0.1327268332,-0.0320847854,-0.1082744449,0.2840714455,-0.2979654074,0.094810091,-0.0292264242,0.5917070508,0.1859064102,0.4440813661,0.2730583549,0.021104537,0.1322196424,0.0927797928,-0.053203471,0.0608418174,-0.0009549259,0.2914585471,-0.081994772,0.5020079613,0.2725434303,0.0259534437,-0.0072190715,-0.075207144,-0.158616364,-0.3384186327,0.204627499,-0.0911616161,0.0677755699,0.2737882435,0.0878871754,-0.3580185771,0.1353775859,-0.0966755003,-0.1938144863,-0.1133242175,0.1158341393,-0.3451522887,-0.1509786099,0.3529440463,0.1030584052,0.1164731979,-0.0047300588,-0.010748487,-0.1193328947,-0.2432302088,-0.1289957613,0.2258381397,0.02893872,0.1047094688,0.1680702269,0.2093173712,0.033626534,-0.0138567258,0.2573041618,-0.0536132939,-0.3020163178,-0.1343097836,0.1198306754,0.0918057412,0.2040550411,-0.3438605666,0.0384183526,0.4002182484,0.0612391569,0.0008474094,-0.290682137,0.220656991,-0.2849721313,-0.1069678515,0.2588521838,-0.5109861493,0.158715412,0.0789557025,0.0771709159,0.2378589064,0.2749264538,-0.2287099212,0.1179107651,-0.1785375029,0.287909776,0.1158813164,-0.121906437,-0.1291884631,0.2582710683,-0.1710102111,0.0061789071,-0.2403122038,-0.183501333,0.6115528941,-0.2941825986,0.1319783777,-0.0414869897,-0.3053473234,0.1779927462,-0.1582277268,-0.2711120248,-0.0471705981,-0.0894282609,0.1199706793,-0.01697715,0.302390933,-0.451308012,0.2286178917,-0.3265037835,0.339217186,0.4150714874,0.0918067545,-0.035193719,0.166398108,-0.1615534425,-0.7281162143,0.1981071383,0.1456974298,-0.0879978985,-0.3534245789,-0.2648589909,0.0765319467,0.1501424313,0.0226836484,0.1255517751,0.0711953044,-0.2024617195,0.0510789268,-0.3482041061,-0.2364073396,-0.2209415287,-0.1390867531,0.1733426154,-0.1894963533,-0.3441658318,-0.0279753543,-0.3571982086,-0.0204043388,0.1966486275,0.239356637,-0.1307834387,-0.0186508615,0.4590831697,-0.1979596764,-0.0776889548,-0.4247601926,-0.0810873657,-0.0454146862,-0.0406259447,-0.0342591852,-0.1599503905,0.142451182,-0.144739449,-0.3011455238,0.0029289648,-0.2381207794,0.5006266832,-0.1498275399,-0.0373831503,-0.0561359078,0.0231335517,-0.0771553442,-0.153137818,-0.1096757501,-0.0489469655,-0.2542680204,-0.0305233877,0.0260866899,-0.1773487478,0.221904546,-0.0419845395,0.0335538946,-0.0073624742,-0.2589181364,0.0027919528,0.339520216,-0.0899918675,0.4603516459,0.1365824491,0.0819970593,-0.0418755561,0.1916162968,0.0838154778,0.1106001586,0.2706128061,0.0925673023,0.0756221414,0.3986678123,0.0161472093,-0.2661878169,-0.351077348,0.1753147542,0.4957439005,-0.2044825107,-0.043800015,-0.0830236226,0.0995051414,0.2236653417,-0.1830222905,0.0713529959,-0.4394610822,-0.1476138383,0.1156395972,-0.1264729351,0.2542539537,-0.1595221162,-0.0608217195,0.1876654774,-0.0264898054,-0.1460788846,-0.1495200843,-0.1710772216,0.0146625154,-0.03508294,-0.144578144,0.0407584049,0.4789995551,-0.3435450494,-0.2255327553,-0.0635434538,-0.0063208234,-0.5026287436,-0.0688748881,0.3389750123,0.2059350759,-0.3035939336,-0.1560991108,0.0831781253,-0.1239539161,0.0785744712,0.2210019976,0.1313579977,-0.1047570184,-0.2336309254,0.1117874458,0.0731343701,-0.2843957543,0.1609170586,-0.1200424731,0.2389975786,-0.2521990538,0.2278801054,-0.2472717613,-0.0553630665,-0.40374735,0.0974972919,-0.010698528,0.1542848647,0.1120021343,-0.0112327524,0.0451950282,-0.0525667556,0.032991603,0.3330432177,-0.3355278969,0.046339605,-0.1026915088,0.135222882,-0.0032963625,0.232637465,0.3257702589,0.2052081823,0.0374394618,-0.0169250686,-0.164213717,0.0817854702,-0.0682317913,-0.003880508,0.0999058709,0.4984065294,0.2369377017,0.5979442,0.204280436,-0.0856824219,0.0948240235,-0.1327353269,0.0399422608,-0.1180508658,-0.3469358385,0.0949326307,-0.3334673345,-0.0153741492,0.0916314721,-0.142918691,-0.5661892295,0.0262778308,0.3182716966,0.0356097259,-0.2705049813,0.2294884175,-0.242961511,0.2372483611,0.1168125421,0.0199956484,-0.2669150531,-0.1107122079,0.2998273075,-0.3833626807,0.1725095212,-0.2118994147,-0.3650291264,-0.0065127322,-0.7440903187,0.4507893026,0.0828428343,0.3523317575,0.1741855294,-0.0121682016,0.0209380426,0.1607970893,0.4166469872,-0.2360412925,-0.1881221533,0.1594663262,0.1826755702,-0.4863575101,-0.057204552,-0.3145472705,0.1810837686,0.2628193498,0.8223790526,-0.3449789286,-0.0785233304,0.3561532199,0.1908120662,0.081973806,-0.1069040298,-0.1756309569,-0.2924115956,-0.1121415347,0.1398420483,0.3096801043,0.1881920844,0.1740519851,-0.2482351512,0.0171870999,-0.1449332535,0.1038188264,0.0200537276,-0.0551291481,-0.1025983393,0.0748112649,-0.1875823587,0.019703459,0.09852276,0.1518204361,-0.2780478299,-0.1183985844,0.0980392471,-0.1654563546,0.5431213975,0.242498979,0.0117663573,0.1635087579,-0.3425570726,0.3598184288,0.1224122271,-0.4356709719,0.3494389653,0.0364778265,-0.3460505009,0.1196431071,0.2144177407,0.2063299268,-0.03236733,0.7606868744,0.0132842269,-0.2412766218,0.7118184566,0.1966302842,0.6759189367,-0.1432673782,-0.0712213963,0.290207535,0.360881716,0.2842663527,0.1126970053,0.381161958,-0.3241282701,-0.0603737421,0.00332267,-0.0351574942,0.1903509349,0.4072134197,0.0292364396,0.3826394081,-0.1521288007,0.4469073713,-0.0252744313,-0.1773910522,0.1236897781,-0.3599649966,0.0880283713,0.0208682176,0.046122279,-0.1213788539,-0.0826372132,0.0984830633,0.0594394431,-0.3395923972,-0.1458568722,-0.0059602247,-0.2831459045,0.2795832455,-0.0178861022,-0.023591958,-0.0276551321,0.1857581586,-0.1375504732,0.0231043696,0.2528838813,-0.0234127026,0.5783449411,0.1532377601,0.0718974993,0.0801609978,-0.3442005217,0.0582830235,-0.0581079088,0.1441702545,-0.2681709528,-0.2917725146,-0.3978101015,0.302022934,0.2278286517,-0.4201090634,-0.174985975,-0.1819713861,-0.028987037,-0.2493113279,0.0179118775,-0.112072818,-0.1037063524,0.7665064931,-0.1971287727,-0.2018837929,0.0001165,0.054283157,-0.0103726638,-0.0073020007,0.509395957,0.1933065951,-0.1906852573,-0.1447668821,-0.1147615388,0.2298499942,-0.4227213264,0.1481416821,-0.3345050812,0.3107252121,0.137876451,0.3870111406,-0.0561545715,0.1035897061,-0.0587295517,-0.1376464814,-0.3454743624,0.2268033475,0.2731829584,0.2479774058,0.009982734,0.0238754451,-0.2870259881,0.1395941824,-0.2310718894,0.1292947829,0.0296753552,0.4705128074,0.2248261571,0.1841877252,-0.063720718,0.1222635955,0.1711612195,0.3236610889,0.2282157838,-0.1890603006,0.087144509,0.0845169425,0.0414255783,-0.0440464094,0.0407923609,-0.1662178338,-0.2258954048,-0.2493456602,0.2834926546,0.2077129483,-0.1348534673,-0.051100973,0.2674014568,0.0906685889,-0.0091010891,0.1165961102,-0.2273985595,-0.082839407,0.1396965832,0.3365561068,0.1446776241,-0.0922280326,0.0380749702,0.0792174712,-0.1854953766,0.1861371994,0.2741543055,-0.3076931536,-0.0741831362,-0.0205289815,0.3571772575,0.1077172607,-0.265585959,-0.1578714103,0.094328694,0.1932017356,-0.0930128843,0.0147829317,0.3097126186,0.3044721186,0.2403047383,0.4139226079,0.0311292689,-0.4228920341,-0.1355920285,-0.0868847966,0.5410777926,-0.1324590594,-0.2924458683,0.6839470863,-0.148261413,-0.0108291414,0.2884373665,0.1919950396,0.3235805035,0.5047323108,-0.0714931935,0.0853694305,0.3333715498,0.104234606,-0.2822675109,-0.4482128918,0.2085962892,0.2343000472,0.0912758559,0.1438355148,0.1783404946,0.0890338719,-0.0002732429,-0.216752544,0.2180004567,0.2582809031,-0.3013063371,-0.025725428,-0.364698261,-0.0765605271,0.1926115453,-0.0516907535,-0.1579218358,-0.1624597609,0.5937283039,0.0620205924,-0.0564188994,-0.3340180814,-0.1528503001,-0.0166738778,0.3545717895,-0.1517580003,0.1551139057,0.0096705584,0.0895556584,-0.2272724062,0.0879208595,0.4263053536,0.3567664027,0.0829301924,0.0477266125,-0.1091136932,-0.1831457168,-0.0300317425,0.1075926647,0.2164026946,0.3985658288,0.3411513269,0.0027499029,-0.0674818233,-0.0332265981,0.3050508499,0.1356130689,-0.0164149981,-0.033591345,0.0403946526,-0.4961536229,-0.1215939075,-0.2080625743,-0.0198331699,-0.2020421326,0.3985048234,-0.1470507532,0.2387565076,0.0221792795,0.0588729046,-0.0984587073,0.0365584604,0.4919873178,-0.3134085834,-0.549947083,-0.205067113,-0.3719426394,-0.0011593878,-0.7319380045,-0.0049579004,0.2208574563,0.1944051087,0.1252499968,0.0747248083,0.0763463005,0.1676183045,-0.2411892414,0.8675721288,-0.3371376693,0.0864991546,-0.0816274583,0.0118992459,0.0936869979,-0.5087841749,0.4204275906,0.219004184,0.0183358509,-0.1565003097,-0.014172188,0.1934941858,-0.1132708788,0.3862443864,0.0413534828,0.0781790391,0.185065344,-0.3118218482,-0.3471717834,0.0969083011,-0.1761681736,0.1207804456,0.21304138,0.1897467226,-0.058404211,0.0200456567,-0.0797780529,0.054864984,-0.0549621284,0.0551304221,-0.1536683887,0.1488116533,0.0846773535,0.1772117466,0.0675943792,0.0562539361,0.1116035134,0.2221062779,-0.3874033988,-0.3115005195,0.4752250314,-0.7607836127,-0.6713259816,-0.0875397548,0.0665985942,0.0903250277,-0.1569895744,-0.3100399077,-0.0200973954,0.262804538,-0.1931921691,-0.1906295866,-0.005261525,-0.3285938501,0.0072535705,-0.0165186673,0.0777009577,0.0896192864,-0.1323938072,-0.0028246206,-0.3992703259]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2218","title":"Duplicates in the LAMA dataset","comments":"Hi,\r\n\r\ncurrently the datasets API doesn't have a dedicated function to remove duplicate rows, but since the LAMA dataset is not too big (it fits in RAM), we can leverage pandas to help us remove duplicates:\r\n```python\r\n>>> from datasets import load_dataset, Dataset\r\n>>> dataset = load_dataset('lama', split='train')\r\n>>> dataset = Dataset.from_pandas(dataset.to_pandas().drop_duplicates(subset=...)) # specify a subset of the columns to consider in a list or use all of the columns if None\r\n```\r\n\r\nNote that the same can be achieved with the `Dataset.filter` method but this would requrie some extra work (filter function, speed?).","body":"I observed duplicates in the LAMA probing dataset, see a minimal code below. \r\n\r\n```\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('lama')\r\nNo config specified, defaulting to: lama\/trex\r\nReusing dataset lama (\/home\/anam\/.cache\/huggingface\/datasets\/lama\/trex\/1.1.0\/97deffae13eca0a18e77dfb3960bb31741e973586f5c1fe1ec0d6b5eece7bddc)\r\n>>> train_dataset = dataset['train']\r\n>>> train_dataset[0]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n>>> train_dataset[1]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n```\r\n\r\nI checked the original data available at https:\/\/dl.fbaipublicfiles.com\/LAMA\/data.zip. This particular duplicated comes from:\r\n```\r\n{\"uuid\": \"40b2ed1c-0961-482e-844e-32596b6117c8\", \"obj_uri\": \"Q150\", \"obj_label\": \"French\", \"sub_uri\": \"Q441235\", \"sub_label\": \"Louis Jules Trochu\", \"predicate_id\": \"P103\", \"evidences\": [{\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}, {\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}]}\r\n``` \r\n\r\nWhat is the best way to deal with these duplicates if I want to use `datasets` to probe with LAMA? ","comment_length":94,"text":"Duplicates in the LAMA dataset\nI observed duplicates in the LAMA probing dataset, see a minimal code below. \r\n\r\n```\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('lama')\r\nNo config specified, defaulting to: lama\/trex\r\nReusing dataset lama (\/home\/anam\/.cache\/huggingface\/datasets\/lama\/trex\/1.1.0\/97deffae13eca0a18e77dfb3960bb31741e973586f5c1fe1ec0d6b5eece7bddc)\r\n>>> train_dataset = dataset['train']\r\n>>> train_dataset[0]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n>>> train_dataset[1]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n```\r\n\r\nI checked the original data available at https:\/\/dl.fbaipublicfiles.com\/LAMA\/data.zip. This particular duplicated comes from:\r\n```\r\n{\"uuid\": \"40b2ed1c-0961-482e-844e-32596b6117c8\", \"obj_uri\": \"Q150\", \"obj_label\": \"French\", \"sub_uri\": \"Q441235\", \"sub_label\": \"Louis Jules Trochu\", \"predicate_id\": \"P103\", \"evidences\": [{\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}, {\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}]}\r\n``` \r\n\r\nWhat is the best way to deal with these duplicates if I want to use `datasets` to probe with LAMA? \nHi,\r\n\r\ncurrently the datasets API doesn't have a dedicated function to remove duplicate rows, but since the LAMA dataset is not too big (it fits in RAM), we can leverage pandas to help us remove duplicates:\r\n```python\r\n>>> from datasets import load_dataset, Dataset\r\n>>> dataset = load_dataset('lama', split='train')\r\n>>> dataset = Dataset.from_pandas(dataset.to_pandas().drop_duplicates(subset=...)) # specify a subset of the columns to consider in a list or use all of the columns if None\r\n```\r\n\r\nNote that the same can be achieved with the `Dataset.filter` method but this would requrie some extra work (filter function, speed?).","embeddings":[0.261572361,-0.321171701,-0.0303781144,0.6542945504,0.3174540102,-0.1388740391,0.3127517998,0.3269729912,-0.5472849011,0.3365436196,-0.3454262018,0.3600648046,0.0819626823,-0.2813018262,0.1604101509,-0.1338147521,0.0254947934,-0.1661694944,-0.2045628428,-0.2063404322,-0.1825976968,0.3395481408,0.0372470431,0.0838601887,-0.140867278,0.1054256782,-0.1408223361,0.2696481645,0.0211215485,-0.2721820772,0.135984987,0.1195656359,-0.0423955284,0.498601675,-0.0001062553,0.0266553126,-0.1191423163,-0.0516993366,-0.1489317864,-0.0307918806,-0.0281007513,-0.0206731278,-0.0044797799,-0.2624208331,-0.3085130155,-0.0679863617,-0.0779155344,-0.110377565,0.4661053121,0.2096013427,0.2672528923,0.3489761353,-0.2831966579,-0.1589336097,0.222759366,-0.1836934388,-0.0866325945,0.4994784296,0.1833483577,0.6046422124,-0.0669347644,0.5685862303,0.0555699319,-0.0770117268,-0.280671984,-0.319239378,0.3182042837,-0.1115747392,0.3222734928,0.3282349706,-0.0091443872,-0.2400610596,-0.1954755038,-0.1231129169,-0.0279283728,-0.1791888773,0.0652221888,0.1354205757,0.0618838258,0.145541653,0.1373835504,0.0252608377,0.2321779728,-0.1204136834,-0.1461453438,-0.1684123576,-0.020073479,0.1156065762,0.1672807336,-0.1583189964,0.2028746605,-0.1564389169,-0.1728953719,0.049390059,-0.2232096195,-0.1347196251,0.0018672978,-0.1056643799,0.1160554811,0.5150147676,0.2339853346,0.3465538025,-0.1307944506,0.0239981376,0.154215306,0.0706345588,-0.0500430986,0.0645817295,-0.1759255826,0.0218920615,-0.316541791,0.0213840082,-0.3417175412,-0.1691877395,0.077304475,-0.0114144515,0.0543418713,-0.1062709838,-0.0838243812,-0.2307511121,-0.403701812,-0.1084920391,0.0605471805,0.1502104402,0.1372216493,0.2108773738,-0.3826012313,0.0159680881,-0.203060329,-0.1918776035,-0.3366514444,0.1416363716,-0.1806383282,0.0794324577,-0.0046100086,-0.722656846,0.4088579118,0.0900722593,-0.1415320188,-0.2692626417,0.4366317391,-0.2164307982,0.1583116204,0.0111389691,-0.1688192636,0.1273399889,-0.0186237115,-0.0218234546,-0.1080174148,0.0238687955,-0.1405326873,-0.0906960964,-0.1264307946,0.2938247025,0.1999355406,0.2575154901,0.3168151379,0.3918619454,-0.0317622386,0.0356455855,0.0930770412,0.0182362515,-0.3235778809,-0.1311048865,0.0814261809,0.2557787299,-0.0844862163,-0.0342722572,0.0129466234,0.0059579285,0.2059483379,0.5270101428,-0.1314258724,0.2728064358,-0.3161910474,0.3068459034,0.1753151566,-0.4883659184,-0.005071064,-0.1671351492,0.0342875831,0.1037544161,0.2280557156,0.0257393103,0.2874675095,-0.102095753,0.2206647247,-0.1189499497,0.2129841596,-0.1990242004,-0.2243700176,-0.0217814241,-0.04410078,-0.1115867645,-0.1224101856,-0.2407581955,0.023219645,-0.0264477897,0.1633731127,-0.087478511,-0.1633484215,-0.0287171733,0.0382602774,-0.0036900432,0.0643826351,0.0341986902,-0.1461330056,0.2015026361,-0.177235052,0.3269667625,0.0302657578,-0.1839377433,-0.2615669072,-0.0959582105,-0.2973302901,-0.1768952906,0.2863488495,0.2737017274,0.0176793355,0.1425696164,0.0079255672,0.1106468365,-0.0368708037,0.1908319294,-0.5088585019,0.0802482441,-0.1847524792,0.0603999868,0.1334294826,0.2846055627,0.1789463311,0.0156143932,0.2225701809,0.0092979688,0.312587589,0.027524583,0.2928505242,-0.1491547674,0.0233941413,-0.1707874835,-0.0926245153,0.0833111554,0.1865539849,-0.0546854064,-0.0379610248,0.4738602638,0.5329996943,-0.2722349763,0.1020209789,-0.1302971393,0.2794715166,-0.13579759,-0.0729421526,-0.492949456,0.0669534355,0.556897819,0.240272522,0.4626099169,-0.2420150191,0.089189589,0.1307507306,-0.1754649132,0.0280328002,-0.1890137345,-0.213729918,-0.111845158,0.1290919334,0.3554101884,0.0149280503,0.2737828195,0.5646850467,0.1218466386,-0.1524690241,-0.2303453982,0.1902780086,-0.0628233552,-0.1223017126,0.1247760057,0.1089129746,-0.0174058173,-0.7163102627,0.1091497615,0.0674972758,0.1782831848,-0.1040224731,0.0458259918,-0.5186646581,-0.2792091668,-0.3423459828,0.116162993,-0.2141042203,-0.2102277726,0.2235587537,0.0833830535,-0.0791669711,0.2216530442,-0.0587286018,0.4643632472,-0.1414313018,0.1901517808,-0.2832052708,0.0621855333,-0.2202775031,0.2213906348,0.1573350579,-0.0116416793,0.3668330908,-0.1453346759,-0.0091756685,-0.4053198099,-0.5439637899,0.1548528969,0.1620249599,0.1663461775,-0.0150466356,-0.2284674048,-0.0598770604,-0.2287738174,-0.092636548,0.2382824421,-0.418710351,-0.0588348731,-0.160943985,0.3222578466,-0.1648000032,-0.3569099605,0.1333421022,-0.1296956241,0.1131868586,0.094897002,0.2506995499,0.1955591291,-0.0962625518,0.1708019823,-0.4386238456,0.0114025883,-0.5250063539,-0.3347619176,0.1283046454,-0.1221257597,-0.0773989931,-0.1265262663,-0.4503448606,0.5347155333,-0.2043711841,-0.3808073103,-0.1040007174,0.0391999371,0.0052048257,0.3159893453,-0.072947599,-0.0524006747,0.2954608798,-0.31335783,-0.0572949871,0.005322624,0.0534760132,0.0567474142,0.1919592917,-0.3597275317,-0.0272062924,0.1268526465,0.176952526,0.4423536062,-0.0578040108,0.2711303532,0.1016453654,0.4149212241,-0.0350889452,-0.096319668,-0.1374595016,-0.1580803245,0.1463189572,0.5981575251,0.1985949129,0.0929000229,-0.0192282498,0.0230381601,-0.1344581395,-0.382353127,-0.2087143064,-0.2636793554,-0.0069357581,0.211635828,-0.1811709702,0.1093775481,0.1607611477,0.159393236,0.0727887377,-0.1498091817,-0.0968033001,-0.0381266437,0.5167205334,-0.616617322,0.1802100837,0.06245967,-0.0113827176,0.0461531989,-0.0188277774,0.1572398096,-0.0100983195,0.4922399521,-0.3488265276,0.0443124697,0.033238437,-0.1674066931,-0.311411798,0.1224629208,-0.2982160747,-0.1219374016,0.1731457859,0.5565639734,-0.2510905564,-0.0345062241,0.2294424772,0.1539774239,-0.2126265317,0.0874856263,-0.2398451716,-0.4762768745,-0.1906112731,-0.0675283596,-0.0483932123,0.0847018883,0.1393900961,-0.3079848588,-0.1403479427,0.0805948973,0.3862339258,0.1830413789,0.0926794559,0.2992742062,0.2168322802,0.0538575388,0.21787332,0.0334332548,0.7148370743,-0.3341753781,-0.269846648,-0.2647650242,-0.0753840581,0.0212232675,0.3166486919,0.0391311459,0.4542929828,-0.0619201884,-0.0403190628,0.0117502259,0.1106813103,0.4621829093,0.0956039578,-0.2127039135,-0.3641094565,0.0299233105,-0.0071278801,-0.2579858303,0.333319217,-0.1612223983,-0.0604873262,0.0208554249,0.0301933251,1.0644155741,-0.276058495,0.0928506255,0.3648729622,-0.2055189162,0.2576840818,0.1336090118,-0.0888972655,-0.1576398164,0.2409190089,0.0959125683,0.1458934397,0.0026951078,0.2030563504,-0.2064414918,0.0662465319,0.1770918667,-0.1697546989,-0.0333000459,0.0339081325,0.0557188503,0.1788266599,0.1361746192,0.2064751983,0.1101291254,-0.027352158,0.0567529835,-0.0005138547,0.1192090437,-0.1942103654,0.0687476695,0.0199436266,-0.1854947358,0.2835600376,-0.1523345113,-0.4586697817,-0.1404190361,-0.2169040293,0.192853272,-0.090867728,0.0777017772,0.1705893278,0.3158831298,0.1821529269,-0.1296540797,-0.0478682071,0.3358630836,-0.0105488608,-0.2502413392,0.0827576965,0.283759594,0.0486860313,0.0348307118,0.2801772654,-0.3211697042,-0.0737142861,0.0834078416,0.2097002417,-0.2965657711,-0.4031594694,0.2004607916,-0.1632791758,-0.2333471477,-0.1487857103,0.0982618406,-0.2829054594,-0.0062638968,0.1913605332,0.4226550162,-0.0867739469,0.1324677616,0.0786751211,-0.0891379341,-0.2560742795,0.1318191588,0.146703586,-0.56168437,0.1952316016,-0.1307239085,0.0196078084,0.1153238937,0.1314592063,-0.0170001648,-0.2400068194,-0.2057884187,0.070640862,-0.3164937794,0.0977499112,-0.0716601387,0.3519402742,-0.1970597953,0.022794025,-0.1280143261,0.2551706135,-0.4023962617,0.3481814563,-0.2751218379,0.2651209831,-0.227521494,-0.0737769604,-0.1020166501,0.3335079253,0.22485663,-0.0967155173,-0.4004729092,-0.3237605691,-0.2450777441,0.0638765991,-0.1113529056,-0.0443416536,-0.0186431129,0.1241746396,0.2122274935,-0.1670002192,0.1350725144,0.1513061374,0.2527486682,0.1451545805,0.100498341,-0.0807351843,-0.2317061573,-0.0962019637,0.1128522903,0.0605732575,0.3657427728,-0.2195559889,-0.2831239104,-0.237146154,-0.0200081486,0.0086322976,0.1730715334,-0.16219154,-0.0964880809,0.1578146368,-0.2682702839,-0.0104041034,0.2486469448,-0.0059935804,-0.2193805575,-0.2179911286,0.1154690906,0.2430663109,-0.5087182522,-0.1043305472,0.0029253184,-0.3091639876,0.2455791235,0.1804054528,0.0081359968,-0.0319934897,0.0910799056,-0.2570297718,0.0060194312,-0.1729642004,0.3735982478,0.3542261422,0.0548061058,-0.1256594211,0.5434408188,0.0092391018,0.2257671654,0.477034241,-0.0859541371,0.5685110092,-0.3942370117,0.160040766,0.2826440334,-0.045967754,0.1593913585,0.2637465298,-0.0594224706,0.0169202499,0.0836917013,0.1143938377,0.249516502,0.0125526711,-0.2026275694,0.1779238284,-0.1874750704,-0.3193872869,0.1066904142,-0.2320482433,-0.0901398659,0.1481957734,0.0413206108,-0.0848239511,0.0190776847,0.2085656524,-0.0676684007,0.0382270813,-0.3484534621,-0.1815468669,0.1049735621,-0.1133373976,0.3851994276,0.4369290471,-0.0377188399,-0.0744031221,0.3042879701,0.3492517471,-0.3513217568,0.1408329606,-0.0486053973,-0.3223835528,0.001642033,-0.1893455237,0.1748964638,0.0700880736,-0.0769213364,0.2669952214,0.1355818212,-0.2319869697,-0.2327668965,0.4252729118,-0.0695289895,-0.4111594856,0.4312680066,-0.3308849633,-0.0459831581,-0.0697187111,0.2132516056,-0.4768787026,0.3899275959,0.421694845,-0.1598174721,0.0607748292,-0.0812868401,0.1262682229,0.0143226059,0.5851635337,0.1267192513,0.0722896978,-0.2126637548,-0.3589540422,-0.5591656566,0.3465291858,-0.3802582324,0.0131842159,0.012630661,0.0681148767,0.0360699855,0.1930892169,-0.0475786217,0.1963954121,0.1265958846,-0.2434819788,-0.4807630181,-0.0214080419,0.0113793993,-0.1753039956,0.0147315301,-0.2825924754,0.0951528624,-0.412732631,0.0645713285,0.1924360543,-0.124110207,0.231443122,0.2564783096,0.289527148,0.1387013346,0.2878736854,0.0106290877,0.0304810777,0.0776627883,-0.1701426059,0.1227281541,0.2852529287,0.1638192385,-0.0594854429,0.0061943526,-0.2301675528,0.1879753768,0.370940119,0.2256373018,0.1016526669,-0.0982463285,0.1454761922,0.1742744148,-0.1492628008,-0.0882016644,0.3099316359,0.0910006911,-0.046500016,-0.2668633759,-0.5187711716,0.5504893661,-0.3391304612,-0.057701081,-0.3835508227,0.2319307625,-0.1272619665,0.0766809583,-0.7147963643,0.2014071047,0.3444114923,0.0844140276,-0.2033613026,0.2627139688,-0.2223798335,0.0071993386,0.005952938,0.011704131,0.2193950266,-0.054083433,0.105964303,-0.2452144027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2218","title":"Duplicates in the LAMA dataset","comments":"Oh, seems like my question wasn't specified well. I'm _not_ asking how to remove duplicates, but whether duplicates should be removed if I want to do the evaluation on the LAMA dataset as it was proposed in the original paper\/repository? In other words, will I get the same result if evaluate on the de-duplicated dataset loaded from HF's `datasets` as the results I'd get if I use the original data format and data processing script in https:\/\/github.com\/facebookresearch\/LAMA? ","body":"I observed duplicates in the LAMA probing dataset, see a minimal code below. \r\n\r\n```\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('lama')\r\nNo config specified, defaulting to: lama\/trex\r\nReusing dataset lama (\/home\/anam\/.cache\/huggingface\/datasets\/lama\/trex\/1.1.0\/97deffae13eca0a18e77dfb3960bb31741e973586f5c1fe1ec0d6b5eece7bddc)\r\n>>> train_dataset = dataset['train']\r\n>>> train_dataset[0]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n>>> train_dataset[1]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n```\r\n\r\nI checked the original data available at https:\/\/dl.fbaipublicfiles.com\/LAMA\/data.zip. This particular duplicated comes from:\r\n```\r\n{\"uuid\": \"40b2ed1c-0961-482e-844e-32596b6117c8\", \"obj_uri\": \"Q150\", \"obj_label\": \"French\", \"sub_uri\": \"Q441235\", \"sub_label\": \"Louis Jules Trochu\", \"predicate_id\": \"P103\", \"evidences\": [{\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}, {\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}]}\r\n``` \r\n\r\nWhat is the best way to deal with these duplicates if I want to use `datasets` to probe with LAMA? ","comment_length":77,"text":"Duplicates in the LAMA dataset\nI observed duplicates in the LAMA probing dataset, see a minimal code below. \r\n\r\n```\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('lama')\r\nNo config specified, defaulting to: lama\/trex\r\nReusing dataset lama (\/home\/anam\/.cache\/huggingface\/datasets\/lama\/trex\/1.1.0\/97deffae13eca0a18e77dfb3960bb31741e973586f5c1fe1ec0d6b5eece7bddc)\r\n>>> train_dataset = dataset['train']\r\n>>> train_dataset[0]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n>>> train_dataset[1]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n```\r\n\r\nI checked the original data available at https:\/\/dl.fbaipublicfiles.com\/LAMA\/data.zip. This particular duplicated comes from:\r\n```\r\n{\"uuid\": \"40b2ed1c-0961-482e-844e-32596b6117c8\", \"obj_uri\": \"Q150\", \"obj_label\": \"French\", \"sub_uri\": \"Q441235\", \"sub_label\": \"Louis Jules Trochu\", \"predicate_id\": \"P103\", \"evidences\": [{\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}, {\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}]}\r\n``` \r\n\r\nWhat is the best way to deal with these duplicates if I want to use `datasets` to probe with LAMA? \nOh, seems like my question wasn't specified well. I'm _not_ asking how to remove duplicates, but whether duplicates should be removed if I want to do the evaluation on the LAMA dataset as it was proposed in the original paper\/repository? In other words, will I get the same result if evaluate on the de-duplicated dataset loaded from HF's `datasets` as the results I'd get if I use the original data format and data processing script in https:\/\/github.com\/facebookresearch\/LAMA? ","embeddings":[0.261572361,-0.321171701,-0.0303781144,0.6542945504,0.3174540102,-0.1388740391,0.3127517998,0.3269729912,-0.5472849011,0.3365436196,-0.3454262018,0.3600648046,0.0819626823,-0.2813018262,0.1604101509,-0.1338147521,0.0254947934,-0.1661694944,-0.2045628428,-0.2063404322,-0.1825976968,0.3395481408,0.0372470431,0.0838601887,-0.140867278,0.1054256782,-0.1408223361,0.2696481645,0.0211215485,-0.2721820772,0.135984987,0.1195656359,-0.0423955284,0.498601675,-0.0001062553,0.0266553126,-0.1191423163,-0.0516993366,-0.1489317864,-0.0307918806,-0.0281007513,-0.0206731278,-0.0044797799,-0.2624208331,-0.3085130155,-0.0679863617,-0.0779155344,-0.110377565,0.4661053121,0.2096013427,0.2672528923,0.3489761353,-0.2831966579,-0.1589336097,0.222759366,-0.1836934388,-0.0866325945,0.4994784296,0.1833483577,0.6046422124,-0.0669347644,0.5685862303,0.0555699319,-0.0770117268,-0.280671984,-0.319239378,0.3182042837,-0.1115747392,0.3222734928,0.3282349706,-0.0091443872,-0.2400610596,-0.1954755038,-0.1231129169,-0.0279283728,-0.1791888773,0.0652221888,0.1354205757,0.0618838258,0.145541653,0.1373835504,0.0252608377,0.2321779728,-0.1204136834,-0.1461453438,-0.1684123576,-0.020073479,0.1156065762,0.1672807336,-0.1583189964,0.2028746605,-0.1564389169,-0.1728953719,0.049390059,-0.2232096195,-0.1347196251,0.0018672978,-0.1056643799,0.1160554811,0.5150147676,0.2339853346,0.3465538025,-0.1307944506,0.0239981376,0.154215306,0.0706345588,-0.0500430986,0.0645817295,-0.1759255826,0.0218920615,-0.316541791,0.0213840082,-0.3417175412,-0.1691877395,0.077304475,-0.0114144515,0.0543418713,-0.1062709838,-0.0838243812,-0.2307511121,-0.403701812,-0.1084920391,0.0605471805,0.1502104402,0.1372216493,0.2108773738,-0.3826012313,0.0159680881,-0.203060329,-0.1918776035,-0.3366514444,0.1416363716,-0.1806383282,0.0794324577,-0.0046100086,-0.722656846,0.4088579118,0.0900722593,-0.1415320188,-0.2692626417,0.4366317391,-0.2164307982,0.1583116204,0.0111389691,-0.1688192636,0.1273399889,-0.0186237115,-0.0218234546,-0.1080174148,0.0238687955,-0.1405326873,-0.0906960964,-0.1264307946,0.2938247025,0.1999355406,0.2575154901,0.3168151379,0.3918619454,-0.0317622386,0.0356455855,0.0930770412,0.0182362515,-0.3235778809,-0.1311048865,0.0814261809,0.2557787299,-0.0844862163,-0.0342722572,0.0129466234,0.0059579285,0.2059483379,0.5270101428,-0.1314258724,0.2728064358,-0.3161910474,0.3068459034,0.1753151566,-0.4883659184,-0.005071064,-0.1671351492,0.0342875831,0.1037544161,0.2280557156,0.0257393103,0.2874675095,-0.102095753,0.2206647247,-0.1189499497,0.2129841596,-0.1990242004,-0.2243700176,-0.0217814241,-0.04410078,-0.1115867645,-0.1224101856,-0.2407581955,0.023219645,-0.0264477897,0.1633731127,-0.087478511,-0.1633484215,-0.0287171733,0.0382602774,-0.0036900432,0.0643826351,0.0341986902,-0.1461330056,0.2015026361,-0.177235052,0.3269667625,0.0302657578,-0.1839377433,-0.2615669072,-0.0959582105,-0.2973302901,-0.1768952906,0.2863488495,0.2737017274,0.0176793355,0.1425696164,0.0079255672,0.1106468365,-0.0368708037,0.1908319294,-0.5088585019,0.0802482441,-0.1847524792,0.0603999868,0.1334294826,0.2846055627,0.1789463311,0.0156143932,0.2225701809,0.0092979688,0.312587589,0.027524583,0.2928505242,-0.1491547674,0.0233941413,-0.1707874835,-0.0926245153,0.0833111554,0.1865539849,-0.0546854064,-0.0379610248,0.4738602638,0.5329996943,-0.2722349763,0.1020209789,-0.1302971393,0.2794715166,-0.13579759,-0.0729421526,-0.492949456,0.0669534355,0.556897819,0.240272522,0.4626099169,-0.2420150191,0.089189589,0.1307507306,-0.1754649132,0.0280328002,-0.1890137345,-0.213729918,-0.111845158,0.1290919334,0.3554101884,0.0149280503,0.2737828195,0.5646850467,0.1218466386,-0.1524690241,-0.2303453982,0.1902780086,-0.0628233552,-0.1223017126,0.1247760057,0.1089129746,-0.0174058173,-0.7163102627,0.1091497615,0.0674972758,0.1782831848,-0.1040224731,0.0458259918,-0.5186646581,-0.2792091668,-0.3423459828,0.116162993,-0.2141042203,-0.2102277726,0.2235587537,0.0833830535,-0.0791669711,0.2216530442,-0.0587286018,0.4643632472,-0.1414313018,0.1901517808,-0.2832052708,0.0621855333,-0.2202775031,0.2213906348,0.1573350579,-0.0116416793,0.3668330908,-0.1453346759,-0.0091756685,-0.4053198099,-0.5439637899,0.1548528969,0.1620249599,0.1663461775,-0.0150466356,-0.2284674048,-0.0598770604,-0.2287738174,-0.092636548,0.2382824421,-0.418710351,-0.0588348731,-0.160943985,0.3222578466,-0.1648000032,-0.3569099605,0.1333421022,-0.1296956241,0.1131868586,0.094897002,0.2506995499,0.1955591291,-0.0962625518,0.1708019823,-0.4386238456,0.0114025883,-0.5250063539,-0.3347619176,0.1283046454,-0.1221257597,-0.0773989931,-0.1265262663,-0.4503448606,0.5347155333,-0.2043711841,-0.3808073103,-0.1040007174,0.0391999371,0.0052048257,0.3159893453,-0.072947599,-0.0524006747,0.2954608798,-0.31335783,-0.0572949871,0.005322624,0.0534760132,0.0567474142,0.1919592917,-0.3597275317,-0.0272062924,0.1268526465,0.176952526,0.4423536062,-0.0578040108,0.2711303532,0.1016453654,0.4149212241,-0.0350889452,-0.096319668,-0.1374595016,-0.1580803245,0.1463189572,0.5981575251,0.1985949129,0.0929000229,-0.0192282498,0.0230381601,-0.1344581395,-0.382353127,-0.2087143064,-0.2636793554,-0.0069357581,0.211635828,-0.1811709702,0.1093775481,0.1607611477,0.159393236,0.0727887377,-0.1498091817,-0.0968033001,-0.0381266437,0.5167205334,-0.616617322,0.1802100837,0.06245967,-0.0113827176,0.0461531989,-0.0188277774,0.1572398096,-0.0100983195,0.4922399521,-0.3488265276,0.0443124697,0.033238437,-0.1674066931,-0.311411798,0.1224629208,-0.2982160747,-0.1219374016,0.1731457859,0.5565639734,-0.2510905564,-0.0345062241,0.2294424772,0.1539774239,-0.2126265317,0.0874856263,-0.2398451716,-0.4762768745,-0.1906112731,-0.0675283596,-0.0483932123,0.0847018883,0.1393900961,-0.3079848588,-0.1403479427,0.0805948973,0.3862339258,0.1830413789,0.0926794559,0.2992742062,0.2168322802,0.0538575388,0.21787332,0.0334332548,0.7148370743,-0.3341753781,-0.269846648,-0.2647650242,-0.0753840581,0.0212232675,0.3166486919,0.0391311459,0.4542929828,-0.0619201884,-0.0403190628,0.0117502259,0.1106813103,0.4621829093,0.0956039578,-0.2127039135,-0.3641094565,0.0299233105,-0.0071278801,-0.2579858303,0.333319217,-0.1612223983,-0.0604873262,0.0208554249,0.0301933251,1.0644155741,-0.276058495,0.0928506255,0.3648729622,-0.2055189162,0.2576840818,0.1336090118,-0.0888972655,-0.1576398164,0.2409190089,0.0959125683,0.1458934397,0.0026951078,0.2030563504,-0.2064414918,0.0662465319,0.1770918667,-0.1697546989,-0.0333000459,0.0339081325,0.0557188503,0.1788266599,0.1361746192,0.2064751983,0.1101291254,-0.027352158,0.0567529835,-0.0005138547,0.1192090437,-0.1942103654,0.0687476695,0.0199436266,-0.1854947358,0.2835600376,-0.1523345113,-0.4586697817,-0.1404190361,-0.2169040293,0.192853272,-0.090867728,0.0777017772,0.1705893278,0.3158831298,0.1821529269,-0.1296540797,-0.0478682071,0.3358630836,-0.0105488608,-0.2502413392,0.0827576965,0.283759594,0.0486860313,0.0348307118,0.2801772654,-0.3211697042,-0.0737142861,0.0834078416,0.2097002417,-0.2965657711,-0.4031594694,0.2004607916,-0.1632791758,-0.2333471477,-0.1487857103,0.0982618406,-0.2829054594,-0.0062638968,0.1913605332,0.4226550162,-0.0867739469,0.1324677616,0.0786751211,-0.0891379341,-0.2560742795,0.1318191588,0.146703586,-0.56168437,0.1952316016,-0.1307239085,0.0196078084,0.1153238937,0.1314592063,-0.0170001648,-0.2400068194,-0.2057884187,0.070640862,-0.3164937794,0.0977499112,-0.0716601387,0.3519402742,-0.1970597953,0.022794025,-0.1280143261,0.2551706135,-0.4023962617,0.3481814563,-0.2751218379,0.2651209831,-0.227521494,-0.0737769604,-0.1020166501,0.3335079253,0.22485663,-0.0967155173,-0.4004729092,-0.3237605691,-0.2450777441,0.0638765991,-0.1113529056,-0.0443416536,-0.0186431129,0.1241746396,0.2122274935,-0.1670002192,0.1350725144,0.1513061374,0.2527486682,0.1451545805,0.100498341,-0.0807351843,-0.2317061573,-0.0962019637,0.1128522903,0.0605732575,0.3657427728,-0.2195559889,-0.2831239104,-0.237146154,-0.0200081486,0.0086322976,0.1730715334,-0.16219154,-0.0964880809,0.1578146368,-0.2682702839,-0.0104041034,0.2486469448,-0.0059935804,-0.2193805575,-0.2179911286,0.1154690906,0.2430663109,-0.5087182522,-0.1043305472,0.0029253184,-0.3091639876,0.2455791235,0.1804054528,0.0081359968,-0.0319934897,0.0910799056,-0.2570297718,0.0060194312,-0.1729642004,0.3735982478,0.3542261422,0.0548061058,-0.1256594211,0.5434408188,0.0092391018,0.2257671654,0.477034241,-0.0859541371,0.5685110092,-0.3942370117,0.160040766,0.2826440334,-0.045967754,0.1593913585,0.2637465298,-0.0594224706,0.0169202499,0.0836917013,0.1143938377,0.249516502,0.0125526711,-0.2026275694,0.1779238284,-0.1874750704,-0.3193872869,0.1066904142,-0.2320482433,-0.0901398659,0.1481957734,0.0413206108,-0.0848239511,0.0190776847,0.2085656524,-0.0676684007,0.0382270813,-0.3484534621,-0.1815468669,0.1049735621,-0.1133373976,0.3851994276,0.4369290471,-0.0377188399,-0.0744031221,0.3042879701,0.3492517471,-0.3513217568,0.1408329606,-0.0486053973,-0.3223835528,0.001642033,-0.1893455237,0.1748964638,0.0700880736,-0.0769213364,0.2669952214,0.1355818212,-0.2319869697,-0.2327668965,0.4252729118,-0.0695289895,-0.4111594856,0.4312680066,-0.3308849633,-0.0459831581,-0.0697187111,0.2132516056,-0.4768787026,0.3899275959,0.421694845,-0.1598174721,0.0607748292,-0.0812868401,0.1262682229,0.0143226059,0.5851635337,0.1267192513,0.0722896978,-0.2126637548,-0.3589540422,-0.5591656566,0.3465291858,-0.3802582324,0.0131842159,0.012630661,0.0681148767,0.0360699855,0.1930892169,-0.0475786217,0.1963954121,0.1265958846,-0.2434819788,-0.4807630181,-0.0214080419,0.0113793993,-0.1753039956,0.0147315301,-0.2825924754,0.0951528624,-0.412732631,0.0645713285,0.1924360543,-0.124110207,0.231443122,0.2564783096,0.289527148,0.1387013346,0.2878736854,0.0106290877,0.0304810777,0.0776627883,-0.1701426059,0.1227281541,0.2852529287,0.1638192385,-0.0594854429,0.0061943526,-0.2301675528,0.1879753768,0.370940119,0.2256373018,0.1016526669,-0.0982463285,0.1454761922,0.1742744148,-0.1492628008,-0.0882016644,0.3099316359,0.0910006911,-0.046500016,-0.2668633759,-0.5187711716,0.5504893661,-0.3391304612,-0.057701081,-0.3835508227,0.2319307625,-0.1272619665,0.0766809583,-0.7147963643,0.2014071047,0.3444114923,0.0844140276,-0.2033613026,0.2627139688,-0.2223798335,0.0071993386,0.005952938,0.011704131,0.2193950266,-0.054083433,0.105964303,-0.2452144027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2218","title":"Duplicates in the LAMA dataset","comments":"So it looks like the person who added LAMA to the library chose to have one item per piece of evidence rather than one per relation - and in this case, there are duplicate pieces of evidence for the target relation\r\n\r\nIf I understand correctly, to reproduce reported results, you would have to aggregate predictions for the several pieces of evidence provided for each relation (each unique `uuid`), but the original authors will know better \r\n\r\ncc @fabiopetroni ","body":"I observed duplicates in the LAMA probing dataset, see a minimal code below. \r\n\r\n```\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('lama')\r\nNo config specified, defaulting to: lama\/trex\r\nReusing dataset lama (\/home\/anam\/.cache\/huggingface\/datasets\/lama\/trex\/1.1.0\/97deffae13eca0a18e77dfb3960bb31741e973586f5c1fe1ec0d6b5eece7bddc)\r\n>>> train_dataset = dataset['train']\r\n>>> train_dataset[0]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n>>> train_dataset[1]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n```\r\n\r\nI checked the original data available at https:\/\/dl.fbaipublicfiles.com\/LAMA\/data.zip. This particular duplicated comes from:\r\n```\r\n{\"uuid\": \"40b2ed1c-0961-482e-844e-32596b6117c8\", \"obj_uri\": \"Q150\", \"obj_label\": \"French\", \"sub_uri\": \"Q441235\", \"sub_label\": \"Louis Jules Trochu\", \"predicate_id\": \"P103\", \"evidences\": [{\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}, {\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}]}\r\n``` \r\n\r\nWhat is the best way to deal with these duplicates if I want to use `datasets` to probe with LAMA? ","comment_length":77,"text":"Duplicates in the LAMA dataset\nI observed duplicates in the LAMA probing dataset, see a minimal code below. \r\n\r\n```\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('lama')\r\nNo config specified, defaulting to: lama\/trex\r\nReusing dataset lama (\/home\/anam\/.cache\/huggingface\/datasets\/lama\/trex\/1.1.0\/97deffae13eca0a18e77dfb3960bb31741e973586f5c1fe1ec0d6b5eece7bddc)\r\n>>> train_dataset = dataset['train']\r\n>>> train_dataset[0]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n>>> train_dataset[1]\r\n{'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'}\r\n```\r\n\r\nI checked the original data available at https:\/\/dl.fbaipublicfiles.com\/LAMA\/data.zip. This particular duplicated comes from:\r\n```\r\n{\"uuid\": \"40b2ed1c-0961-482e-844e-32596b6117c8\", \"obj_uri\": \"Q150\", \"obj_label\": \"French\", \"sub_uri\": \"Q441235\", \"sub_label\": \"Louis Jules Trochu\", \"predicate_id\": \"P103\", \"evidences\": [{\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}, {\"sub_surface\": \"Louis Jules Trochu\", \"obj_surface\": \"French\", \"masked_sentence\": \"Louis Jules Trochu ([lwi \\u0292yl t\\u0281\\u0254\\u0283y]; 12 March 1815 \\u2013 7 October 1896) was a [MASK] military leader and politician.\"}]}\r\n``` \r\n\r\nWhat is the best way to deal with these duplicates if I want to use `datasets` to probe with LAMA? \nSo it looks like the person who added LAMA to the library chose to have one item per piece of evidence rather than one per relation - and in this case, there are duplicate pieces of evidence for the target relation\r\n\r\nIf I understand correctly, to reproduce reported results, you would have to aggregate predictions for the several pieces of evidence provided for each relation (each unique `uuid`), but the original authors will know better \r\n\r\ncc @fabiopetroni ","embeddings":[0.261572361,-0.321171701,-0.0303781144,0.6542945504,0.3174540102,-0.1388740391,0.3127517998,0.3269729912,-0.5472849011,0.3365436196,-0.3454262018,0.3600648046,0.0819626823,-0.2813018262,0.1604101509,-0.1338147521,0.0254947934,-0.1661694944,-0.2045628428,-0.2063404322,-0.1825976968,0.3395481408,0.0372470431,0.0838601887,-0.140867278,0.1054256782,-0.1408223361,0.2696481645,0.0211215485,-0.2721820772,0.135984987,0.1195656359,-0.0423955284,0.498601675,-0.0001062553,0.0266553126,-0.1191423163,-0.0516993366,-0.1489317864,-0.0307918806,-0.0281007513,-0.0206731278,-0.0044797799,-0.2624208331,-0.3085130155,-0.0679863617,-0.0779155344,-0.110377565,0.4661053121,0.2096013427,0.2672528923,0.3489761353,-0.2831966579,-0.1589336097,0.222759366,-0.1836934388,-0.0866325945,0.4994784296,0.1833483577,0.6046422124,-0.0669347644,0.5685862303,0.0555699319,-0.0770117268,-0.280671984,-0.319239378,0.3182042837,-0.1115747392,0.3222734928,0.3282349706,-0.0091443872,-0.2400610596,-0.1954755038,-0.1231129169,-0.0279283728,-0.1791888773,0.0652221888,0.1354205757,0.0618838258,0.145541653,0.1373835504,0.0252608377,0.2321779728,-0.1204136834,-0.1461453438,-0.1684123576,-0.020073479,0.1156065762,0.1672807336,-0.1583189964,0.2028746605,-0.1564389169,-0.1728953719,0.049390059,-0.2232096195,-0.1347196251,0.0018672978,-0.1056643799,0.1160554811,0.5150147676,0.2339853346,0.3465538025,-0.1307944506,0.0239981376,0.154215306,0.0706345588,-0.0500430986,0.0645817295,-0.1759255826,0.0218920615,-0.316541791,0.0213840082,-0.3417175412,-0.1691877395,0.077304475,-0.0114144515,0.0543418713,-0.1062709838,-0.0838243812,-0.2307511121,-0.403701812,-0.1084920391,0.0605471805,0.1502104402,0.1372216493,0.2108773738,-0.3826012313,0.0159680881,-0.203060329,-0.1918776035,-0.3366514444,0.1416363716,-0.1806383282,0.0794324577,-0.0046100086,-0.722656846,0.4088579118,0.0900722593,-0.1415320188,-0.2692626417,0.4366317391,-0.2164307982,0.1583116204,0.0111389691,-0.1688192636,0.1273399889,-0.0186237115,-0.0218234546,-0.1080174148,0.0238687955,-0.1405326873,-0.0906960964,-0.1264307946,0.2938247025,0.1999355406,0.2575154901,0.3168151379,0.3918619454,-0.0317622386,0.0356455855,0.0930770412,0.0182362515,-0.3235778809,-0.1311048865,0.0814261809,0.2557787299,-0.0844862163,-0.0342722572,0.0129466234,0.0059579285,0.2059483379,0.5270101428,-0.1314258724,0.2728064358,-0.3161910474,0.3068459034,0.1753151566,-0.4883659184,-0.005071064,-0.1671351492,0.0342875831,0.1037544161,0.2280557156,0.0257393103,0.2874675095,-0.102095753,0.2206647247,-0.1189499497,0.2129841596,-0.1990242004,-0.2243700176,-0.0217814241,-0.04410078,-0.1115867645,-0.1224101856,-0.2407581955,0.023219645,-0.0264477897,0.1633731127,-0.087478511,-0.1633484215,-0.0287171733,0.0382602774,-0.0036900432,0.0643826351,0.0341986902,-0.1461330056,0.2015026361,-0.177235052,0.3269667625,0.0302657578,-0.1839377433,-0.2615669072,-0.0959582105,-0.2973302901,-0.1768952906,0.2863488495,0.2737017274,0.0176793355,0.1425696164,0.0079255672,0.1106468365,-0.0368708037,0.1908319294,-0.5088585019,0.0802482441,-0.1847524792,0.0603999868,0.1334294826,0.2846055627,0.1789463311,0.0156143932,0.2225701809,0.0092979688,0.312587589,0.027524583,0.2928505242,-0.1491547674,0.0233941413,-0.1707874835,-0.0926245153,0.0833111554,0.1865539849,-0.0546854064,-0.0379610248,0.4738602638,0.5329996943,-0.2722349763,0.1020209789,-0.1302971393,0.2794715166,-0.13579759,-0.0729421526,-0.492949456,0.0669534355,0.556897819,0.240272522,0.4626099169,-0.2420150191,0.089189589,0.1307507306,-0.1754649132,0.0280328002,-0.1890137345,-0.213729918,-0.111845158,0.1290919334,0.3554101884,0.0149280503,0.2737828195,0.5646850467,0.1218466386,-0.1524690241,-0.2303453982,0.1902780086,-0.0628233552,-0.1223017126,0.1247760057,0.1089129746,-0.0174058173,-0.7163102627,0.1091497615,0.0674972758,0.1782831848,-0.1040224731,0.0458259918,-0.5186646581,-0.2792091668,-0.3423459828,0.116162993,-0.2141042203,-0.2102277726,0.2235587537,0.0833830535,-0.0791669711,0.2216530442,-0.0587286018,0.4643632472,-0.1414313018,0.1901517808,-0.2832052708,0.0621855333,-0.2202775031,0.2213906348,0.1573350579,-0.0116416793,0.3668330908,-0.1453346759,-0.0091756685,-0.4053198099,-0.5439637899,0.1548528969,0.1620249599,0.1663461775,-0.0150466356,-0.2284674048,-0.0598770604,-0.2287738174,-0.092636548,0.2382824421,-0.418710351,-0.0588348731,-0.160943985,0.3222578466,-0.1648000032,-0.3569099605,0.1333421022,-0.1296956241,0.1131868586,0.094897002,0.2506995499,0.1955591291,-0.0962625518,0.1708019823,-0.4386238456,0.0114025883,-0.5250063539,-0.3347619176,0.1283046454,-0.1221257597,-0.0773989931,-0.1265262663,-0.4503448606,0.5347155333,-0.2043711841,-0.3808073103,-0.1040007174,0.0391999371,0.0052048257,0.3159893453,-0.072947599,-0.0524006747,0.2954608798,-0.31335783,-0.0572949871,0.005322624,0.0534760132,0.0567474142,0.1919592917,-0.3597275317,-0.0272062924,0.1268526465,0.176952526,0.4423536062,-0.0578040108,0.2711303532,0.1016453654,0.4149212241,-0.0350889452,-0.096319668,-0.1374595016,-0.1580803245,0.1463189572,0.5981575251,0.1985949129,0.0929000229,-0.0192282498,0.0230381601,-0.1344581395,-0.382353127,-0.2087143064,-0.2636793554,-0.0069357581,0.211635828,-0.1811709702,0.1093775481,0.1607611477,0.159393236,0.0727887377,-0.1498091817,-0.0968033001,-0.0381266437,0.5167205334,-0.616617322,0.1802100837,0.06245967,-0.0113827176,0.0461531989,-0.0188277774,0.1572398096,-0.0100983195,0.4922399521,-0.3488265276,0.0443124697,0.033238437,-0.1674066931,-0.311411798,0.1224629208,-0.2982160747,-0.1219374016,0.1731457859,0.5565639734,-0.2510905564,-0.0345062241,0.2294424772,0.1539774239,-0.2126265317,0.0874856263,-0.2398451716,-0.4762768745,-0.1906112731,-0.0675283596,-0.0483932123,0.0847018883,0.1393900961,-0.3079848588,-0.1403479427,0.0805948973,0.3862339258,0.1830413789,0.0926794559,0.2992742062,0.2168322802,0.0538575388,0.21787332,0.0334332548,0.7148370743,-0.3341753781,-0.269846648,-0.2647650242,-0.0753840581,0.0212232675,0.3166486919,0.0391311459,0.4542929828,-0.0619201884,-0.0403190628,0.0117502259,0.1106813103,0.4621829093,0.0956039578,-0.2127039135,-0.3641094565,0.0299233105,-0.0071278801,-0.2579858303,0.333319217,-0.1612223983,-0.0604873262,0.0208554249,0.0301933251,1.0644155741,-0.276058495,0.0928506255,0.3648729622,-0.2055189162,0.2576840818,0.1336090118,-0.0888972655,-0.1576398164,0.2409190089,0.0959125683,0.1458934397,0.0026951078,0.2030563504,-0.2064414918,0.0662465319,0.1770918667,-0.1697546989,-0.0333000459,0.0339081325,0.0557188503,0.1788266599,0.1361746192,0.2064751983,0.1101291254,-0.027352158,0.0567529835,-0.0005138547,0.1192090437,-0.1942103654,0.0687476695,0.0199436266,-0.1854947358,0.2835600376,-0.1523345113,-0.4586697817,-0.1404190361,-0.2169040293,0.192853272,-0.090867728,0.0777017772,0.1705893278,0.3158831298,0.1821529269,-0.1296540797,-0.0478682071,0.3358630836,-0.0105488608,-0.2502413392,0.0827576965,0.283759594,0.0486860313,0.0348307118,0.2801772654,-0.3211697042,-0.0737142861,0.0834078416,0.2097002417,-0.2965657711,-0.4031594694,0.2004607916,-0.1632791758,-0.2333471477,-0.1487857103,0.0982618406,-0.2829054594,-0.0062638968,0.1913605332,0.4226550162,-0.0867739469,0.1324677616,0.0786751211,-0.0891379341,-0.2560742795,0.1318191588,0.146703586,-0.56168437,0.1952316016,-0.1307239085,0.0196078084,0.1153238937,0.1314592063,-0.0170001648,-0.2400068194,-0.2057884187,0.070640862,-0.3164937794,0.0977499112,-0.0716601387,0.3519402742,-0.1970597953,0.022794025,-0.1280143261,0.2551706135,-0.4023962617,0.3481814563,-0.2751218379,0.2651209831,-0.227521494,-0.0737769604,-0.1020166501,0.3335079253,0.22485663,-0.0967155173,-0.4004729092,-0.3237605691,-0.2450777441,0.0638765991,-0.1113529056,-0.0443416536,-0.0186431129,0.1241746396,0.2122274935,-0.1670002192,0.1350725144,0.1513061374,0.2527486682,0.1451545805,0.100498341,-0.0807351843,-0.2317061573,-0.0962019637,0.1128522903,0.0605732575,0.3657427728,-0.2195559889,-0.2831239104,-0.237146154,-0.0200081486,0.0086322976,0.1730715334,-0.16219154,-0.0964880809,0.1578146368,-0.2682702839,-0.0104041034,0.2486469448,-0.0059935804,-0.2193805575,-0.2179911286,0.1154690906,0.2430663109,-0.5087182522,-0.1043305472,0.0029253184,-0.3091639876,0.2455791235,0.1804054528,0.0081359968,-0.0319934897,0.0910799056,-0.2570297718,0.0060194312,-0.1729642004,0.3735982478,0.3542261422,0.0548061058,-0.1256594211,0.5434408188,0.0092391018,0.2257671654,0.477034241,-0.0859541371,0.5685110092,-0.3942370117,0.160040766,0.2826440334,-0.045967754,0.1593913585,0.2637465298,-0.0594224706,0.0169202499,0.0836917013,0.1143938377,0.249516502,0.0125526711,-0.2026275694,0.1779238284,-0.1874750704,-0.3193872869,0.1066904142,-0.2320482433,-0.0901398659,0.1481957734,0.0413206108,-0.0848239511,0.0190776847,0.2085656524,-0.0676684007,0.0382270813,-0.3484534621,-0.1815468669,0.1049735621,-0.1133373976,0.3851994276,0.4369290471,-0.0377188399,-0.0744031221,0.3042879701,0.3492517471,-0.3513217568,0.1408329606,-0.0486053973,-0.3223835528,0.001642033,-0.1893455237,0.1748964638,0.0700880736,-0.0769213364,0.2669952214,0.1355818212,-0.2319869697,-0.2327668965,0.4252729118,-0.0695289895,-0.4111594856,0.4312680066,-0.3308849633,-0.0459831581,-0.0697187111,0.2132516056,-0.4768787026,0.3899275959,0.421694845,-0.1598174721,0.0607748292,-0.0812868401,0.1262682229,0.0143226059,0.5851635337,0.1267192513,0.0722896978,-0.2126637548,-0.3589540422,-0.5591656566,0.3465291858,-0.3802582324,0.0131842159,0.012630661,0.0681148767,0.0360699855,0.1930892169,-0.0475786217,0.1963954121,0.1265958846,-0.2434819788,-0.4807630181,-0.0214080419,0.0113793993,-0.1753039956,0.0147315301,-0.2825924754,0.0951528624,-0.412732631,0.0645713285,0.1924360543,-0.124110207,0.231443122,0.2564783096,0.289527148,0.1387013346,0.2878736854,0.0106290877,0.0304810777,0.0776627883,-0.1701426059,0.1227281541,0.2852529287,0.1638192385,-0.0594854429,0.0061943526,-0.2301675528,0.1879753768,0.370940119,0.2256373018,0.1016526669,-0.0982463285,0.1454761922,0.1742744148,-0.1492628008,-0.0882016644,0.3099316359,0.0910006911,-0.046500016,-0.2668633759,-0.5187711716,0.5504893661,-0.3391304612,-0.057701081,-0.3835508227,0.2319307625,-0.1272619665,0.0766809583,-0.7147963643,0.2014071047,0.3444114923,0.0844140276,-0.2033613026,0.2627139688,-0.2223798335,0.0071993386,0.005952938,0.011704131,0.2193950266,-0.054083433,0.105964303,-0.2452144027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2214","title":"load_metric error: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'","comments":"Hi @nsaphra, thanks for reporting.\r\n\r\nThis issue was fixed in `datasets` version 1.3.0. Could you please update `datasets` and tell me if the problem persists?\r\n```shell\r\npip install -U datasets\r\n```","body":"I'm having the same problem as [Notebooks issue 10](https:\/\/github.com\/huggingface\/notebooks\/issues\/10) on datasets 1.2.1, and it seems to be an issue with the datasets package.\r\n\r\n```python\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric(\"glue\", \"sst2\")\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 502, in load_metric\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 66, in import_main_class\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\", line 1014, in _gcd_import\r\n File \"\", line 991, in _find_and_load\r\n File \"\", line 975, in _find_and_load_unlocked\r\n File \"\", line 671, in _load_unlocked\r\n File \"\", line 783, in exec_module\r\n File \"\", line 219, in _call_with_frames_removed\r\n File \"\/home\/ns4008\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/glue\/e4606ab9804a36bcd5a9cebb2cb65bb14b6ac78ee9e6d5981fa679a495dd55de\/glue.py\", line 105, in \r\n @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)\r\nAttributeError: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\r\n```","comment_length":31,"text":"load_metric error: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\nI'm having the same problem as [Notebooks issue 10](https:\/\/github.com\/huggingface\/notebooks\/issues\/10) on datasets 1.2.1, and it seems to be an issue with the datasets package.\r\n\r\n```python\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric(\"glue\", \"sst2\")\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 502, in load_metric\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 66, in import_main_class\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\", line 1014, in _gcd_import\r\n File \"\", line 991, in _find_and_load\r\n File \"\", line 975, in _find_and_load_unlocked\r\n File \"\", line 671, in _load_unlocked\r\n File \"\", line 783, in exec_module\r\n File \"\", line 219, in _call_with_frames_removed\r\n File \"\/home\/ns4008\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/glue\/e4606ab9804a36bcd5a9cebb2cb65bb14b6ac78ee9e6d5981fa679a495dd55de\/glue.py\", line 105, in \r\n @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)\r\nAttributeError: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\r\n```\nHi @nsaphra, thanks for reporting.\r\n\r\nThis issue was fixed in `datasets` version 1.3.0. Could you please update `datasets` and tell me if the problem persists?\r\n```shell\r\npip install -U datasets\r\n```","embeddings":[-0.272324115,-0.2135577202,0.0193226859,0.1852925718,0.4195098579,0.0630552247,0.2729536891,0.1801064909,0.0756967589,-0.0524457283,-0.1940380335,0.1545963436,-0.0744487196,0.2974149585,0.0652143732,-0.0471376516,-0.0315006822,0.0150832096,-0.262601763,0.0296820384,-0.3187800348,0.293637991,-0.1672726125,-0.1764028966,-0.4021217525,-0.0105557591,0.0437473059,0.2441147864,-0.2862702608,-0.5002132654,0.205599755,0.0462033674,0.2963437438,0.3439409733,-0.0001162612,-0.0840861276,0.2146514505,-0.0279947501,-0.5544195771,-0.1524079293,0.0528011806,-0.2994415164,0.2512985468,-0.1053964123,-0.026472494,-0.1421283036,-0.0965214372,-0.2981289327,0.3209556639,0.2984568179,0.2062049806,0.6519038081,0.1126084998,-0.3246344924,0.0756547675,-0.1170940176,-0.1389452964,0.5830135942,0.2155656517,-0.2970149517,0.0055574444,0.1997525692,-0.0833818167,0.2960377336,0.5132633448,0.0072548999,0.0939969942,-0.0558887757,0.1015874147,0.059948612,0.484824419,-0.3377355635,-0.3453686535,-0.2876414061,0.157133773,-0.317251116,0.2258811444,-0.1266947538,0.0611915551,0.1738776267,-0.1425379813,-0.2746182978,0.027385395,0.2549823523,-0.0566361211,-0.1325753331,-0.3288052678,-0.0292661712,0.3993551731,0.0067997873,0.0061149034,0.1106405407,0.0018888788,0.2454485446,-0.2312647104,0.1604048014,0.2210600972,0.0638860911,0.1995056421,0.1169555858,0.1385359019,-0.1091200337,0.2827876508,0.1445126384,0.0451146699,0.455306083,0.4384199381,0.0501264371,0.2727783024,0.3831758201,0.2283397615,-0.1032075882,-0.0001541478,-0.4193750322,-0.1396079361,-0.1696378291,0.3145554066,-0.0437624045,-0.339749068,0.2789826691,0.225033164,-0.1100322455,0.0559781827,0.166669026,0.0389566608,0.0680720434,0.3797704279,0.1804562062,-0.2567597926,0.0695359856,-0.2539496422,0.019209398,-0.2204780877,0.2743342817,0.2387660295,-0.1765834093,0.2692468762,0.1394805461,0.0810193792,0.0131133115,-0.1423486918,0.1412217319,-0.2486168146,0.2955060303,0.013345235,0.1941344142,0.3053650856,-0.3389752507,-0.0738557056,-0.2483933419,-0.2457935363,-0.0913328826,-0.2465759516,0.1607254595,-0.3777996302,0.1018846259,-0.1548740715,0.0269365031,0.1158767045,0.0955258533,-0.0533686727,-0.0219097435,-0.4126269221,-0.0537653193,0.3398079276,0.407076776,-0.2471117377,-0.4057824016,0.1503967047,-0.1388975978,-0.2114148438,0.0392381735,0.0853885636,0.2430645525,-0.2263119221,0.0026322715,0.1370444,-0.5137932301,-0.4204660356,0.1396769583,-0.1487744749,0.0827302411,0.0750858337,-0.1714534611,0.1146221235,0.183054179,0.4504910111,0.0914153084,0.1167476848,-0.1548795253,-0.290825367,-0.1838073581,-0.2629747391,0.2544320524,0.1082778797,0.0519852117,0.175834924,0.1697672904,0.0678901598,-0.0798934996,0.011294866,0.4932676554,0.1543736309,-0.0364747122,0.1711865067,-0.2008412182,-0.5201351643,0.2156841606,-0.0183965508,0.0993434042,-0.0899934471,-0.0668857396,-0.374559015,-0.0463788509,-0.0753306672,-0.1765413284,0.0452711396,-0.0109367613,-0.1645413339,0.3287498355,-0.1420029849,0.2980311811,-0.3087519705,0.3473673463,-0.3626068532,0.2142395675,0.0971087739,-0.1683486402,0.1748615503,0.1577347219,0.1103116572,-0.1772711575,-0.0718013197,0.4855265319,0.1505092531,0.0629772693,-0.0285120029,0.2165285498,0.1915651858,-0.1421236694,-0.1667533368,-0.0733661801,0.035070125,-0.0022641413,-0.0186236035,0.1603875458,-0.1758860499,0.0595270544,0.1123987362,0.1688022166,0.048878897,-0.0778958648,-0.1880658865,-0.3839529753,0.479842335,-0.0180930253,0.3275276124,-0.0315788649,0.1737259477,-0.1127126664,0.3555789888,0.0096923541,0.0877412483,0.1162704453,-0.308801055,0.1370434165,-0.0036531514,0.011657536,0.4693245888,0.2006489336,-0.0045757908,0.2940479517,0.0089875711,-0.0964115486,0.12861377,0.0178792831,-0.1744787693,0.2901944816,-0.0639035255,-0.0595521629,-0.3733405769,0.1549218893,-0.0791442096,0.2749669254,-0.5363030434,-0.1757649183,-0.141107738,0.0296398327,0.0277284384,-0.0767671391,-0.2003962547,-0.2440668344,-0.0292488467,0.325810194,0.1955318749,0.3660383523,0.2152761966,0.4211014807,-0.0384670012,0.0846681148,-0.1243118271,-0.2874588072,-0.0785380751,0.0081650047,0.1657315195,-0.0335034579,0.1387250274,-0.6127305627,0.06443654,-0.236285001,-0.3688507378,0.0849087164,-0.0047380268,0.6359252334,0.3205825984,0.1494532824,-0.0961013958,0.0279980153,0.3684045076,0.0242663678,-0.2071206421,0.0681154877,-0.0857632682,-0.2336559445,-0.3043382466,-0.1047391817,0.0186322648,-0.503063798,0.3876099288,0.0402854495,-0.2336942852,0.3373192847,0.0795239285,0.2915133536,0.0523024015,0.2416902632,-0.261346221,-0.3416886032,0.2696045637,-0.2346160859,-0.4145871997,-0.037765462,0.1373154223,0.1808706522,-0.0452895761,-0.4627621174,-0.4108830392,-0.1143371314,0.1598947495,-0.1194815487,0.3886238635,0.1906858832,0.0357468687,0.0353719592,-0.3328506052,-0.1745150089,0.0547360964,-0.1960035712,0.0167503189,-0.2609972954,0.1186313927,0.0603463016,0.3715052307,0.2424533367,-0.0580785982,0.2844280005,-0.1521651447,0.6900770664,-0.1171550304,-0.5205394626,0.070948638,0.0032520145,0.2066455036,0.294349432,0.0683957711,0.3357643485,-0.2086580098,0.0799778104,-0.107449159,-0.1985200495,-0.1780722588,0.0575307794,0.2056672275,0.043477606,0.2509680092,-0.1185469553,0.0546496622,0.1190925911,0.4457848966,-0.0582257807,0.0248067491,-0.396876663,0.1530397236,-0.4186137915,0.3114287555,-0.0233412813,0.2591668963,-0.2117097825,-0.1854931265,0.0008127582,0.0003988862,0.4743218124,-0.0232882611,-0.1957427561,0.2003183365,-0.0690099299,-0.685212791,-0.0052718553,-0.0417399704,-0.0789852813,0.1828616858,0.7977381349,-0.2625100315,-0.3014018238,0.1301297694,0.1023930162,-0.0043754182,-0.0999475941,-0.3193716407,-0.3368156254,-0.1378110796,-0.0916927457,-0.0276012328,0.2983665466,0.1582421064,0.2091055214,-0.1144397929,-0.2217674553,0.1712339669,0.1866385192,0.1943307966,-0.0234470051,0.3947012126,0.3202777803,-0.0763602033,0.4250121713,0.4037992954,-0.4309894443,-0.6466447711,0.1787409633,-0.0535439402,0.1202298105,0.157498464,-0.2035825402,0.0360497124,-0.0532159917,0.2204992473,-0.4408127964,0.2652892768,0.4414024055,0.1872788072,-0.0870289728,-0.1726748943,0.306450218,0.0870726556,0.00999304,0.4308449328,0.0634560212,-0.2938775718,-0.0701917261,-0.2305243164,0.5940442681,0.0290086195,0.1729602814,0.2764141262,-0.1854176819,0.4218051434,-0.0393425412,-0.0951728448,-0.2183059305,-0.4836227894,0.0405451246,-0.0413951837,0.2808120549,-0.2143423706,-0.1116682589,0.0950956792,-0.2373245806,0.0704354271,-0.0563032404,0.2277140021,-0.2661421895,-0.1809367836,-0.2217793018,0.122424826,-0.1408931613,0.2309868932,-0.1256341487,0.1210680231,-0.0697817206,-0.2212692052,-0.1227326468,-0.1291439235,-0.2800857723,0.144272238,0.2374107093,-0.1632544696,-0.0273615066,0.3203982413,0.4834209681,0.0730526522,-0.3700918257,0.20750314,-0.3207891285,-0.0973202959,-0.0424925014,0.0715611503,0.2013940662,0.0530964024,0.0595653243,-0.1689048558,0.0645233318,0.2031840235,0.0174677745,-0.0456067212,-0.0876971036,-0.4919409156,-0.1246381253,-0.0510219261,0.2836449742,-0.327545166,0.0981150642,0.0562658198,0.014764308,0.1047228202,-0.0331803784,-0.242884919,-0.0733277053,0.6832268834,-0.361394614,-0.1414557099,0.1263989955,0.1018077284,-0.318993628,-0.1406425238,0.2113707811,0.2389927953,-0.3433932364,0.2778002918,0.5266761184,0.0435032248,-0.0637036711,0.4369817674,0.2357449532,-0.3535464406,0.1269493997,-0.2212594002,-0.3677404225,0.2803766131,-0.1661068797,0.1610450149,0.0869790614,0.1942923218,-0.1050552502,-0.0148057314,-0.2793801129,0.211676687,-0.0993371308,-0.1408014148,0.6002429724,-0.2447780818,0.2816815078,-0.1756057292,0.0771706775,0.015653275,-0.3317786157,-0.1820966899,-0.4139865637,0.1064253449,-0.0538477972,-0.1300275624,0.2956237793,0.0028885128,-0.0922225863,-0.2934277952,0.1935119033,0.4534095228,-0.046222426,0.2377129346,-0.122749649,0.232391119,0.1859098375,0.1401135325,-0.100802131,0.0997574255,-0.0606916659,0.0412759297,-0.1847326607,0.2028342634,0.0222671926,0.2372679859,0.1615038514,-0.1148739532,0.1546191126,-0.1380564868,0.0113244327,-0.0168448221,0.3367970586,0.2945950031,-0.2881359756,-0.0667470321,0.1050204039,0.1873946935,-0.2524309754,-0.2942762375,0.0593587644,0.0206503924,0.1127656773,0.1540552378,0.1536152661,-0.1455436349,0.0944109932,0.0923186019,0.3824128211,0.1105610281,0.170265317,0.1636682898,0.1364786774,-0.0244052391,0.3797616065,0.0976049304,0.2229423076,0.0688975453,-0.4428275526,0.0882378817,-0.2641712427,0.3839266896,-0.2432175875,-0.4694573879,0.0781102702,0.0745861009,-0.0688807964,-0.0325670131,-0.3320102096,0.7181387544,-0.3015743792,-0.0596848205,-0.0467802659,0.0559419021,-0.1177831888,-0.1937913001,-0.0985112488,-0.1023545563,-0.0581151396,-0.0594300516,-0.0467668734,-0.1459453106,0.3599132895,-0.1085797548,-0.0336290449,-0.3310728967,-0.3075093031,0.2592685521,-0.1198569462,-0.132297948,0.2339993864,-0.1117595658,0.2395399362,0.0096039288,0.6236209869,0.4861671031,0.0578343943,0.1642263532,-0.1270774752,-0.1182659939,0.0369492695,-0.0820896849,0.4027970135,0.0168094747,0.1449293196,0.1828604192,0.1365623325,-0.2118962109,0.1529088318,-0.2281762213,-0.3224211037,-0.3317995965,0.0626964718,-0.2673926651,-0.0225482862,-0.130914852,0.114061676,-0.3242834806,0.1154515967,0.408370167,0.1985117197,0.08757101,-0.1523372978,0.0868726298,-0.053498473,0.1909273863,0.400077939,0.0367470011,-0.2721703053,-0.2734406888,-0.6236268878,0.0099143265,0.1216766313,-0.0570753515,0.1204302981,-0.0348658748,0.0610775165,-0.0391001515,0.1405400932,0.1027174741,-0.0041313292,-0.2952067256,-0.4408206344,0.0197271816,-0.2038558275,-0.101773493,0.0670962855,-0.2293940634,-0.0436051041,-0.1830174476,-0.02075647,-0.1345573664,0.2511404753,0.0660701469,0.0035569319,0.4722623229,-0.0868728086,0.3599009216,-0.0522920601,-0.0728004128,-0.1637538671,-0.4566641748,-0.0161271878,0.1952701509,-0.0664701834,0.2887929678,-0.4623144269,-0.3876182139,-0.2914842367,0.5101401806,0.2518330514,-0.0921489224,-0.2270249575,0.337551713,-0.1708122492,0.0662507117,0.1981481314,0.3309055269,-0.1810210496,0.3260757923,-0.1189147606,-0.5026894212,0.6233594418,-0.3565818667,-0.0996360332,0.2245274782,0.288562417,0.3277512789,-0.3560940027,-0.60240376,-0.057727553,0.3717492819,0.0284310561,-0.506750524,0.591793716,-0.271661669,0.1035222337,-0.0543052293,0.1474553049,0.2477992475,-0.2005041838,0.2222824991,-0.2424782515]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2214","title":"load_metric error: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'","comments":"There might be a bug in the conda version of `datasets` 1.2.1 where the datasets\/metric scripts are downloaded from `master` instead of the `1.2.1` repo.\r\n\r\nYou can try setting the env var `HF_SCRIPTS_VERSION=\"1.2.1\"` as a workaround. Let me know if that helps.","body":"I'm having the same problem as [Notebooks issue 10](https:\/\/github.com\/huggingface\/notebooks\/issues\/10) on datasets 1.2.1, and it seems to be an issue with the datasets package.\r\n\r\n```python\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric(\"glue\", \"sst2\")\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 502, in load_metric\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 66, in import_main_class\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\", line 1014, in _gcd_import\r\n File \"\", line 991, in _find_and_load\r\n File \"\", line 975, in _find_and_load_unlocked\r\n File \"\", line 671, in _load_unlocked\r\n File \"\", line 783, in exec_module\r\n File \"\", line 219, in _call_with_frames_removed\r\n File \"\/home\/ns4008\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/glue\/e4606ab9804a36bcd5a9cebb2cb65bb14b6ac78ee9e6d5981fa679a495dd55de\/glue.py\", line 105, in \r\n @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)\r\nAttributeError: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\r\n```","comment_length":42,"text":"load_metric error: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\nI'm having the same problem as [Notebooks issue 10](https:\/\/github.com\/huggingface\/notebooks\/issues\/10) on datasets 1.2.1, and it seems to be an issue with the datasets package.\r\n\r\n```python\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric(\"glue\", \"sst2\")\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 502, in load_metric\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 66, in import_main_class\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\", line 1014, in _gcd_import\r\n File \"\", line 991, in _find_and_load\r\n File \"\", line 975, in _find_and_load_unlocked\r\n File \"\", line 671, in _load_unlocked\r\n File \"\", line 783, in exec_module\r\n File \"\", line 219, in _call_with_frames_removed\r\n File \"\/home\/ns4008\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/glue\/e4606ab9804a36bcd5a9cebb2cb65bb14b6ac78ee9e6d5981fa679a495dd55de\/glue.py\", line 105, in \r\n @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)\r\nAttributeError: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\r\n```\nThere might be a bug in the conda version of `datasets` 1.2.1 where the datasets\/metric scripts are downloaded from `master` instead of the `1.2.1` repo.\r\n\r\nYou can try setting the env var `HF_SCRIPTS_VERSION=\"1.2.1\"` as a workaround. Let me know if that helps.","embeddings":[-0.272324115,-0.2135577202,0.0193226859,0.1852925718,0.4195098579,0.0630552247,0.2729536891,0.1801064909,0.0756967589,-0.0524457283,-0.1940380335,0.1545963436,-0.0744487196,0.2974149585,0.0652143732,-0.0471376516,-0.0315006822,0.0150832096,-0.262601763,0.0296820384,-0.3187800348,0.293637991,-0.1672726125,-0.1764028966,-0.4021217525,-0.0105557591,0.0437473059,0.2441147864,-0.2862702608,-0.5002132654,0.205599755,0.0462033674,0.2963437438,0.3439409733,-0.0001162612,-0.0840861276,0.2146514505,-0.0279947501,-0.5544195771,-0.1524079293,0.0528011806,-0.2994415164,0.2512985468,-0.1053964123,-0.026472494,-0.1421283036,-0.0965214372,-0.2981289327,0.3209556639,0.2984568179,0.2062049806,0.6519038081,0.1126084998,-0.3246344924,0.0756547675,-0.1170940176,-0.1389452964,0.5830135942,0.2155656517,-0.2970149517,0.0055574444,0.1997525692,-0.0833818167,0.2960377336,0.5132633448,0.0072548999,0.0939969942,-0.0558887757,0.1015874147,0.059948612,0.484824419,-0.3377355635,-0.3453686535,-0.2876414061,0.157133773,-0.317251116,0.2258811444,-0.1266947538,0.0611915551,0.1738776267,-0.1425379813,-0.2746182978,0.027385395,0.2549823523,-0.0566361211,-0.1325753331,-0.3288052678,-0.0292661712,0.3993551731,0.0067997873,0.0061149034,0.1106405407,0.0018888788,0.2454485446,-0.2312647104,0.1604048014,0.2210600972,0.0638860911,0.1995056421,0.1169555858,0.1385359019,-0.1091200337,0.2827876508,0.1445126384,0.0451146699,0.455306083,0.4384199381,0.0501264371,0.2727783024,0.3831758201,0.2283397615,-0.1032075882,-0.0001541478,-0.4193750322,-0.1396079361,-0.1696378291,0.3145554066,-0.0437624045,-0.339749068,0.2789826691,0.225033164,-0.1100322455,0.0559781827,0.166669026,0.0389566608,0.0680720434,0.3797704279,0.1804562062,-0.2567597926,0.0695359856,-0.2539496422,0.019209398,-0.2204780877,0.2743342817,0.2387660295,-0.1765834093,0.2692468762,0.1394805461,0.0810193792,0.0131133115,-0.1423486918,0.1412217319,-0.2486168146,0.2955060303,0.013345235,0.1941344142,0.3053650856,-0.3389752507,-0.0738557056,-0.2483933419,-0.2457935363,-0.0913328826,-0.2465759516,0.1607254595,-0.3777996302,0.1018846259,-0.1548740715,0.0269365031,0.1158767045,0.0955258533,-0.0533686727,-0.0219097435,-0.4126269221,-0.0537653193,0.3398079276,0.407076776,-0.2471117377,-0.4057824016,0.1503967047,-0.1388975978,-0.2114148438,0.0392381735,0.0853885636,0.2430645525,-0.2263119221,0.0026322715,0.1370444,-0.5137932301,-0.4204660356,0.1396769583,-0.1487744749,0.0827302411,0.0750858337,-0.1714534611,0.1146221235,0.183054179,0.4504910111,0.0914153084,0.1167476848,-0.1548795253,-0.290825367,-0.1838073581,-0.2629747391,0.2544320524,0.1082778797,0.0519852117,0.175834924,0.1697672904,0.0678901598,-0.0798934996,0.011294866,0.4932676554,0.1543736309,-0.0364747122,0.1711865067,-0.2008412182,-0.5201351643,0.2156841606,-0.0183965508,0.0993434042,-0.0899934471,-0.0668857396,-0.374559015,-0.0463788509,-0.0753306672,-0.1765413284,0.0452711396,-0.0109367613,-0.1645413339,0.3287498355,-0.1420029849,0.2980311811,-0.3087519705,0.3473673463,-0.3626068532,0.2142395675,0.0971087739,-0.1683486402,0.1748615503,0.1577347219,0.1103116572,-0.1772711575,-0.0718013197,0.4855265319,0.1505092531,0.0629772693,-0.0285120029,0.2165285498,0.1915651858,-0.1421236694,-0.1667533368,-0.0733661801,0.035070125,-0.0022641413,-0.0186236035,0.1603875458,-0.1758860499,0.0595270544,0.1123987362,0.1688022166,0.048878897,-0.0778958648,-0.1880658865,-0.3839529753,0.479842335,-0.0180930253,0.3275276124,-0.0315788649,0.1737259477,-0.1127126664,0.3555789888,0.0096923541,0.0877412483,0.1162704453,-0.308801055,0.1370434165,-0.0036531514,0.011657536,0.4693245888,0.2006489336,-0.0045757908,0.2940479517,0.0089875711,-0.0964115486,0.12861377,0.0178792831,-0.1744787693,0.2901944816,-0.0639035255,-0.0595521629,-0.3733405769,0.1549218893,-0.0791442096,0.2749669254,-0.5363030434,-0.1757649183,-0.141107738,0.0296398327,0.0277284384,-0.0767671391,-0.2003962547,-0.2440668344,-0.0292488467,0.325810194,0.1955318749,0.3660383523,0.2152761966,0.4211014807,-0.0384670012,0.0846681148,-0.1243118271,-0.2874588072,-0.0785380751,0.0081650047,0.1657315195,-0.0335034579,0.1387250274,-0.6127305627,0.06443654,-0.236285001,-0.3688507378,0.0849087164,-0.0047380268,0.6359252334,0.3205825984,0.1494532824,-0.0961013958,0.0279980153,0.3684045076,0.0242663678,-0.2071206421,0.0681154877,-0.0857632682,-0.2336559445,-0.3043382466,-0.1047391817,0.0186322648,-0.503063798,0.3876099288,0.0402854495,-0.2336942852,0.3373192847,0.0795239285,0.2915133536,0.0523024015,0.2416902632,-0.261346221,-0.3416886032,0.2696045637,-0.2346160859,-0.4145871997,-0.037765462,0.1373154223,0.1808706522,-0.0452895761,-0.4627621174,-0.4108830392,-0.1143371314,0.1598947495,-0.1194815487,0.3886238635,0.1906858832,0.0357468687,0.0353719592,-0.3328506052,-0.1745150089,0.0547360964,-0.1960035712,0.0167503189,-0.2609972954,0.1186313927,0.0603463016,0.3715052307,0.2424533367,-0.0580785982,0.2844280005,-0.1521651447,0.6900770664,-0.1171550304,-0.5205394626,0.070948638,0.0032520145,0.2066455036,0.294349432,0.0683957711,0.3357643485,-0.2086580098,0.0799778104,-0.107449159,-0.1985200495,-0.1780722588,0.0575307794,0.2056672275,0.043477606,0.2509680092,-0.1185469553,0.0546496622,0.1190925911,0.4457848966,-0.0582257807,0.0248067491,-0.396876663,0.1530397236,-0.4186137915,0.3114287555,-0.0233412813,0.2591668963,-0.2117097825,-0.1854931265,0.0008127582,0.0003988862,0.4743218124,-0.0232882611,-0.1957427561,0.2003183365,-0.0690099299,-0.685212791,-0.0052718553,-0.0417399704,-0.0789852813,0.1828616858,0.7977381349,-0.2625100315,-0.3014018238,0.1301297694,0.1023930162,-0.0043754182,-0.0999475941,-0.3193716407,-0.3368156254,-0.1378110796,-0.0916927457,-0.0276012328,0.2983665466,0.1582421064,0.2091055214,-0.1144397929,-0.2217674553,0.1712339669,0.1866385192,0.1943307966,-0.0234470051,0.3947012126,0.3202777803,-0.0763602033,0.4250121713,0.4037992954,-0.4309894443,-0.6466447711,0.1787409633,-0.0535439402,0.1202298105,0.157498464,-0.2035825402,0.0360497124,-0.0532159917,0.2204992473,-0.4408127964,0.2652892768,0.4414024055,0.1872788072,-0.0870289728,-0.1726748943,0.306450218,0.0870726556,0.00999304,0.4308449328,0.0634560212,-0.2938775718,-0.0701917261,-0.2305243164,0.5940442681,0.0290086195,0.1729602814,0.2764141262,-0.1854176819,0.4218051434,-0.0393425412,-0.0951728448,-0.2183059305,-0.4836227894,0.0405451246,-0.0413951837,0.2808120549,-0.2143423706,-0.1116682589,0.0950956792,-0.2373245806,0.0704354271,-0.0563032404,0.2277140021,-0.2661421895,-0.1809367836,-0.2217793018,0.122424826,-0.1408931613,0.2309868932,-0.1256341487,0.1210680231,-0.0697817206,-0.2212692052,-0.1227326468,-0.1291439235,-0.2800857723,0.144272238,0.2374107093,-0.1632544696,-0.0273615066,0.3203982413,0.4834209681,0.0730526522,-0.3700918257,0.20750314,-0.3207891285,-0.0973202959,-0.0424925014,0.0715611503,0.2013940662,0.0530964024,0.0595653243,-0.1689048558,0.0645233318,0.2031840235,0.0174677745,-0.0456067212,-0.0876971036,-0.4919409156,-0.1246381253,-0.0510219261,0.2836449742,-0.327545166,0.0981150642,0.0562658198,0.014764308,0.1047228202,-0.0331803784,-0.242884919,-0.0733277053,0.6832268834,-0.361394614,-0.1414557099,0.1263989955,0.1018077284,-0.318993628,-0.1406425238,0.2113707811,0.2389927953,-0.3433932364,0.2778002918,0.5266761184,0.0435032248,-0.0637036711,0.4369817674,0.2357449532,-0.3535464406,0.1269493997,-0.2212594002,-0.3677404225,0.2803766131,-0.1661068797,0.1610450149,0.0869790614,0.1942923218,-0.1050552502,-0.0148057314,-0.2793801129,0.211676687,-0.0993371308,-0.1408014148,0.6002429724,-0.2447780818,0.2816815078,-0.1756057292,0.0771706775,0.015653275,-0.3317786157,-0.1820966899,-0.4139865637,0.1064253449,-0.0538477972,-0.1300275624,0.2956237793,0.0028885128,-0.0922225863,-0.2934277952,0.1935119033,0.4534095228,-0.046222426,0.2377129346,-0.122749649,0.232391119,0.1859098375,0.1401135325,-0.100802131,0.0997574255,-0.0606916659,0.0412759297,-0.1847326607,0.2028342634,0.0222671926,0.2372679859,0.1615038514,-0.1148739532,0.1546191126,-0.1380564868,0.0113244327,-0.0168448221,0.3367970586,0.2945950031,-0.2881359756,-0.0667470321,0.1050204039,0.1873946935,-0.2524309754,-0.2942762375,0.0593587644,0.0206503924,0.1127656773,0.1540552378,0.1536152661,-0.1455436349,0.0944109932,0.0923186019,0.3824128211,0.1105610281,0.170265317,0.1636682898,0.1364786774,-0.0244052391,0.3797616065,0.0976049304,0.2229423076,0.0688975453,-0.4428275526,0.0882378817,-0.2641712427,0.3839266896,-0.2432175875,-0.4694573879,0.0781102702,0.0745861009,-0.0688807964,-0.0325670131,-0.3320102096,0.7181387544,-0.3015743792,-0.0596848205,-0.0467802659,0.0559419021,-0.1177831888,-0.1937913001,-0.0985112488,-0.1023545563,-0.0581151396,-0.0594300516,-0.0467668734,-0.1459453106,0.3599132895,-0.1085797548,-0.0336290449,-0.3310728967,-0.3075093031,0.2592685521,-0.1198569462,-0.132297948,0.2339993864,-0.1117595658,0.2395399362,0.0096039288,0.6236209869,0.4861671031,0.0578343943,0.1642263532,-0.1270774752,-0.1182659939,0.0369492695,-0.0820896849,0.4027970135,0.0168094747,0.1449293196,0.1828604192,0.1365623325,-0.2118962109,0.1529088318,-0.2281762213,-0.3224211037,-0.3317995965,0.0626964718,-0.2673926651,-0.0225482862,-0.130914852,0.114061676,-0.3242834806,0.1154515967,0.408370167,0.1985117197,0.08757101,-0.1523372978,0.0868726298,-0.053498473,0.1909273863,0.400077939,0.0367470011,-0.2721703053,-0.2734406888,-0.6236268878,0.0099143265,0.1216766313,-0.0570753515,0.1204302981,-0.0348658748,0.0610775165,-0.0391001515,0.1405400932,0.1027174741,-0.0041313292,-0.2952067256,-0.4408206344,0.0197271816,-0.2038558275,-0.101773493,0.0670962855,-0.2293940634,-0.0436051041,-0.1830174476,-0.02075647,-0.1345573664,0.2511404753,0.0660701469,0.0035569319,0.4722623229,-0.0868728086,0.3599009216,-0.0522920601,-0.0728004128,-0.1637538671,-0.4566641748,-0.0161271878,0.1952701509,-0.0664701834,0.2887929678,-0.4623144269,-0.3876182139,-0.2914842367,0.5101401806,0.2518330514,-0.0921489224,-0.2270249575,0.337551713,-0.1708122492,0.0662507117,0.1981481314,0.3309055269,-0.1810210496,0.3260757923,-0.1189147606,-0.5026894212,0.6233594418,-0.3565818667,-0.0996360332,0.2245274782,0.288562417,0.3277512789,-0.3560940027,-0.60240376,-0.057727553,0.3717492819,0.0284310561,-0.506750524,0.591793716,-0.271661669,0.1035222337,-0.0543052293,0.1474553049,0.2477992475,-0.2005041838,0.2222824991,-0.2424782515]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2214","title":"load_metric error: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'","comments":"I just faced the same issue. I was using 1.2.1 from conda and received the same AttributeError complaining about 'add_start_docstrings'. Uninstalling the conda installed datasets and then installing the latest datasets (version 1.5.0) using pip install solved the issue for me. I don't like mixing up conda and pip installs in the same environments but this will have to do for now, until 1.5.0 is made available through conda.","body":"I'm having the same problem as [Notebooks issue 10](https:\/\/github.com\/huggingface\/notebooks\/issues\/10) on datasets 1.2.1, and it seems to be an issue with the datasets package.\r\n\r\n```python\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric(\"glue\", \"sst2\")\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 502, in load_metric\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 66, in import_main_class\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\", line 1014, in _gcd_import\r\n File \"\", line 991, in _find_and_load\r\n File \"\", line 975, in _find_and_load_unlocked\r\n File \"\", line 671, in _load_unlocked\r\n File \"\", line 783, in exec_module\r\n File \"\", line 219, in _call_with_frames_removed\r\n File \"\/home\/ns4008\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/glue\/e4606ab9804a36bcd5a9cebb2cb65bb14b6ac78ee9e6d5981fa679a495dd55de\/glue.py\", line 105, in \r\n @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)\r\nAttributeError: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\r\n```","comment_length":69,"text":"load_metric error: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\nI'm having the same problem as [Notebooks issue 10](https:\/\/github.com\/huggingface\/notebooks\/issues\/10) on datasets 1.2.1, and it seems to be an issue with the datasets package.\r\n\r\n```python\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric(\"glue\", \"sst2\")\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 502, in load_metric\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 66, in import_main_class\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\", line 1014, in _gcd_import\r\n File \"\", line 991, in _find_and_load\r\n File \"\", line 975, in _find_and_load_unlocked\r\n File \"\", line 671, in _load_unlocked\r\n File \"\", line 783, in exec_module\r\n File \"\", line 219, in _call_with_frames_removed\r\n File \"\/home\/ns4008\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/glue\/e4606ab9804a36bcd5a9cebb2cb65bb14b6ac78ee9e6d5981fa679a495dd55de\/glue.py\", line 105, in \r\n @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)\r\nAttributeError: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\r\n```\nI just faced the same issue. I was using 1.2.1 from conda and received the same AttributeError complaining about 'add_start_docstrings'. Uninstalling the conda installed datasets and then installing the latest datasets (version 1.5.0) using pip install solved the issue for me. I don't like mixing up conda and pip installs in the same environments but this will have to do for now, until 1.5.0 is made available through conda.","embeddings":[-0.272324115,-0.2135577202,0.0193226859,0.1852925718,0.4195098579,0.0630552247,0.2729536891,0.1801064909,0.0756967589,-0.0524457283,-0.1940380335,0.1545963436,-0.0744487196,0.2974149585,0.0652143732,-0.0471376516,-0.0315006822,0.0150832096,-0.262601763,0.0296820384,-0.3187800348,0.293637991,-0.1672726125,-0.1764028966,-0.4021217525,-0.0105557591,0.0437473059,0.2441147864,-0.2862702608,-0.5002132654,0.205599755,0.0462033674,0.2963437438,0.3439409733,-0.0001162612,-0.0840861276,0.2146514505,-0.0279947501,-0.5544195771,-0.1524079293,0.0528011806,-0.2994415164,0.2512985468,-0.1053964123,-0.026472494,-0.1421283036,-0.0965214372,-0.2981289327,0.3209556639,0.2984568179,0.2062049806,0.6519038081,0.1126084998,-0.3246344924,0.0756547675,-0.1170940176,-0.1389452964,0.5830135942,0.2155656517,-0.2970149517,0.0055574444,0.1997525692,-0.0833818167,0.2960377336,0.5132633448,0.0072548999,0.0939969942,-0.0558887757,0.1015874147,0.059948612,0.484824419,-0.3377355635,-0.3453686535,-0.2876414061,0.157133773,-0.317251116,0.2258811444,-0.1266947538,0.0611915551,0.1738776267,-0.1425379813,-0.2746182978,0.027385395,0.2549823523,-0.0566361211,-0.1325753331,-0.3288052678,-0.0292661712,0.3993551731,0.0067997873,0.0061149034,0.1106405407,0.0018888788,0.2454485446,-0.2312647104,0.1604048014,0.2210600972,0.0638860911,0.1995056421,0.1169555858,0.1385359019,-0.1091200337,0.2827876508,0.1445126384,0.0451146699,0.455306083,0.4384199381,0.0501264371,0.2727783024,0.3831758201,0.2283397615,-0.1032075882,-0.0001541478,-0.4193750322,-0.1396079361,-0.1696378291,0.3145554066,-0.0437624045,-0.339749068,0.2789826691,0.225033164,-0.1100322455,0.0559781827,0.166669026,0.0389566608,0.0680720434,0.3797704279,0.1804562062,-0.2567597926,0.0695359856,-0.2539496422,0.019209398,-0.2204780877,0.2743342817,0.2387660295,-0.1765834093,0.2692468762,0.1394805461,0.0810193792,0.0131133115,-0.1423486918,0.1412217319,-0.2486168146,0.2955060303,0.013345235,0.1941344142,0.3053650856,-0.3389752507,-0.0738557056,-0.2483933419,-0.2457935363,-0.0913328826,-0.2465759516,0.1607254595,-0.3777996302,0.1018846259,-0.1548740715,0.0269365031,0.1158767045,0.0955258533,-0.0533686727,-0.0219097435,-0.4126269221,-0.0537653193,0.3398079276,0.407076776,-0.2471117377,-0.4057824016,0.1503967047,-0.1388975978,-0.2114148438,0.0392381735,0.0853885636,0.2430645525,-0.2263119221,0.0026322715,0.1370444,-0.5137932301,-0.4204660356,0.1396769583,-0.1487744749,0.0827302411,0.0750858337,-0.1714534611,0.1146221235,0.183054179,0.4504910111,0.0914153084,0.1167476848,-0.1548795253,-0.290825367,-0.1838073581,-0.2629747391,0.2544320524,0.1082778797,0.0519852117,0.175834924,0.1697672904,0.0678901598,-0.0798934996,0.011294866,0.4932676554,0.1543736309,-0.0364747122,0.1711865067,-0.2008412182,-0.5201351643,0.2156841606,-0.0183965508,0.0993434042,-0.0899934471,-0.0668857396,-0.374559015,-0.0463788509,-0.0753306672,-0.1765413284,0.0452711396,-0.0109367613,-0.1645413339,0.3287498355,-0.1420029849,0.2980311811,-0.3087519705,0.3473673463,-0.3626068532,0.2142395675,0.0971087739,-0.1683486402,0.1748615503,0.1577347219,0.1103116572,-0.1772711575,-0.0718013197,0.4855265319,0.1505092531,0.0629772693,-0.0285120029,0.2165285498,0.1915651858,-0.1421236694,-0.1667533368,-0.0733661801,0.035070125,-0.0022641413,-0.0186236035,0.1603875458,-0.1758860499,0.0595270544,0.1123987362,0.1688022166,0.048878897,-0.0778958648,-0.1880658865,-0.3839529753,0.479842335,-0.0180930253,0.3275276124,-0.0315788649,0.1737259477,-0.1127126664,0.3555789888,0.0096923541,0.0877412483,0.1162704453,-0.308801055,0.1370434165,-0.0036531514,0.011657536,0.4693245888,0.2006489336,-0.0045757908,0.2940479517,0.0089875711,-0.0964115486,0.12861377,0.0178792831,-0.1744787693,0.2901944816,-0.0639035255,-0.0595521629,-0.3733405769,0.1549218893,-0.0791442096,0.2749669254,-0.5363030434,-0.1757649183,-0.141107738,0.0296398327,0.0277284384,-0.0767671391,-0.2003962547,-0.2440668344,-0.0292488467,0.325810194,0.1955318749,0.3660383523,0.2152761966,0.4211014807,-0.0384670012,0.0846681148,-0.1243118271,-0.2874588072,-0.0785380751,0.0081650047,0.1657315195,-0.0335034579,0.1387250274,-0.6127305627,0.06443654,-0.236285001,-0.3688507378,0.0849087164,-0.0047380268,0.6359252334,0.3205825984,0.1494532824,-0.0961013958,0.0279980153,0.3684045076,0.0242663678,-0.2071206421,0.0681154877,-0.0857632682,-0.2336559445,-0.3043382466,-0.1047391817,0.0186322648,-0.503063798,0.3876099288,0.0402854495,-0.2336942852,0.3373192847,0.0795239285,0.2915133536,0.0523024015,0.2416902632,-0.261346221,-0.3416886032,0.2696045637,-0.2346160859,-0.4145871997,-0.037765462,0.1373154223,0.1808706522,-0.0452895761,-0.4627621174,-0.4108830392,-0.1143371314,0.1598947495,-0.1194815487,0.3886238635,0.1906858832,0.0357468687,0.0353719592,-0.3328506052,-0.1745150089,0.0547360964,-0.1960035712,0.0167503189,-0.2609972954,0.1186313927,0.0603463016,0.3715052307,0.2424533367,-0.0580785982,0.2844280005,-0.1521651447,0.6900770664,-0.1171550304,-0.5205394626,0.070948638,0.0032520145,0.2066455036,0.294349432,0.0683957711,0.3357643485,-0.2086580098,0.0799778104,-0.107449159,-0.1985200495,-0.1780722588,0.0575307794,0.2056672275,0.043477606,0.2509680092,-0.1185469553,0.0546496622,0.1190925911,0.4457848966,-0.0582257807,0.0248067491,-0.396876663,0.1530397236,-0.4186137915,0.3114287555,-0.0233412813,0.2591668963,-0.2117097825,-0.1854931265,0.0008127582,0.0003988862,0.4743218124,-0.0232882611,-0.1957427561,0.2003183365,-0.0690099299,-0.685212791,-0.0052718553,-0.0417399704,-0.0789852813,0.1828616858,0.7977381349,-0.2625100315,-0.3014018238,0.1301297694,0.1023930162,-0.0043754182,-0.0999475941,-0.3193716407,-0.3368156254,-0.1378110796,-0.0916927457,-0.0276012328,0.2983665466,0.1582421064,0.2091055214,-0.1144397929,-0.2217674553,0.1712339669,0.1866385192,0.1943307966,-0.0234470051,0.3947012126,0.3202777803,-0.0763602033,0.4250121713,0.4037992954,-0.4309894443,-0.6466447711,0.1787409633,-0.0535439402,0.1202298105,0.157498464,-0.2035825402,0.0360497124,-0.0532159917,0.2204992473,-0.4408127964,0.2652892768,0.4414024055,0.1872788072,-0.0870289728,-0.1726748943,0.306450218,0.0870726556,0.00999304,0.4308449328,0.0634560212,-0.2938775718,-0.0701917261,-0.2305243164,0.5940442681,0.0290086195,0.1729602814,0.2764141262,-0.1854176819,0.4218051434,-0.0393425412,-0.0951728448,-0.2183059305,-0.4836227894,0.0405451246,-0.0413951837,0.2808120549,-0.2143423706,-0.1116682589,0.0950956792,-0.2373245806,0.0704354271,-0.0563032404,0.2277140021,-0.2661421895,-0.1809367836,-0.2217793018,0.122424826,-0.1408931613,0.2309868932,-0.1256341487,0.1210680231,-0.0697817206,-0.2212692052,-0.1227326468,-0.1291439235,-0.2800857723,0.144272238,0.2374107093,-0.1632544696,-0.0273615066,0.3203982413,0.4834209681,0.0730526522,-0.3700918257,0.20750314,-0.3207891285,-0.0973202959,-0.0424925014,0.0715611503,0.2013940662,0.0530964024,0.0595653243,-0.1689048558,0.0645233318,0.2031840235,0.0174677745,-0.0456067212,-0.0876971036,-0.4919409156,-0.1246381253,-0.0510219261,0.2836449742,-0.327545166,0.0981150642,0.0562658198,0.014764308,0.1047228202,-0.0331803784,-0.242884919,-0.0733277053,0.6832268834,-0.361394614,-0.1414557099,0.1263989955,0.1018077284,-0.318993628,-0.1406425238,0.2113707811,0.2389927953,-0.3433932364,0.2778002918,0.5266761184,0.0435032248,-0.0637036711,0.4369817674,0.2357449532,-0.3535464406,0.1269493997,-0.2212594002,-0.3677404225,0.2803766131,-0.1661068797,0.1610450149,0.0869790614,0.1942923218,-0.1050552502,-0.0148057314,-0.2793801129,0.211676687,-0.0993371308,-0.1408014148,0.6002429724,-0.2447780818,0.2816815078,-0.1756057292,0.0771706775,0.015653275,-0.3317786157,-0.1820966899,-0.4139865637,0.1064253449,-0.0538477972,-0.1300275624,0.2956237793,0.0028885128,-0.0922225863,-0.2934277952,0.1935119033,0.4534095228,-0.046222426,0.2377129346,-0.122749649,0.232391119,0.1859098375,0.1401135325,-0.100802131,0.0997574255,-0.0606916659,0.0412759297,-0.1847326607,0.2028342634,0.0222671926,0.2372679859,0.1615038514,-0.1148739532,0.1546191126,-0.1380564868,0.0113244327,-0.0168448221,0.3367970586,0.2945950031,-0.2881359756,-0.0667470321,0.1050204039,0.1873946935,-0.2524309754,-0.2942762375,0.0593587644,0.0206503924,0.1127656773,0.1540552378,0.1536152661,-0.1455436349,0.0944109932,0.0923186019,0.3824128211,0.1105610281,0.170265317,0.1636682898,0.1364786774,-0.0244052391,0.3797616065,0.0976049304,0.2229423076,0.0688975453,-0.4428275526,0.0882378817,-0.2641712427,0.3839266896,-0.2432175875,-0.4694573879,0.0781102702,0.0745861009,-0.0688807964,-0.0325670131,-0.3320102096,0.7181387544,-0.3015743792,-0.0596848205,-0.0467802659,0.0559419021,-0.1177831888,-0.1937913001,-0.0985112488,-0.1023545563,-0.0581151396,-0.0594300516,-0.0467668734,-0.1459453106,0.3599132895,-0.1085797548,-0.0336290449,-0.3310728967,-0.3075093031,0.2592685521,-0.1198569462,-0.132297948,0.2339993864,-0.1117595658,0.2395399362,0.0096039288,0.6236209869,0.4861671031,0.0578343943,0.1642263532,-0.1270774752,-0.1182659939,0.0369492695,-0.0820896849,0.4027970135,0.0168094747,0.1449293196,0.1828604192,0.1365623325,-0.2118962109,0.1529088318,-0.2281762213,-0.3224211037,-0.3317995965,0.0626964718,-0.2673926651,-0.0225482862,-0.130914852,0.114061676,-0.3242834806,0.1154515967,0.408370167,0.1985117197,0.08757101,-0.1523372978,0.0868726298,-0.053498473,0.1909273863,0.400077939,0.0367470011,-0.2721703053,-0.2734406888,-0.6236268878,0.0099143265,0.1216766313,-0.0570753515,0.1204302981,-0.0348658748,0.0610775165,-0.0391001515,0.1405400932,0.1027174741,-0.0041313292,-0.2952067256,-0.4408206344,0.0197271816,-0.2038558275,-0.101773493,0.0670962855,-0.2293940634,-0.0436051041,-0.1830174476,-0.02075647,-0.1345573664,0.2511404753,0.0660701469,0.0035569319,0.4722623229,-0.0868728086,0.3599009216,-0.0522920601,-0.0728004128,-0.1637538671,-0.4566641748,-0.0161271878,0.1952701509,-0.0664701834,0.2887929678,-0.4623144269,-0.3876182139,-0.2914842367,0.5101401806,0.2518330514,-0.0921489224,-0.2270249575,0.337551713,-0.1708122492,0.0662507117,0.1981481314,0.3309055269,-0.1810210496,0.3260757923,-0.1189147606,-0.5026894212,0.6233594418,-0.3565818667,-0.0996360332,0.2245274782,0.288562417,0.3277512789,-0.3560940027,-0.60240376,-0.057727553,0.3717492819,0.0284310561,-0.506750524,0.591793716,-0.271661669,0.1035222337,-0.0543052293,0.1474553049,0.2477992475,-0.2005041838,0.2222824991,-0.2424782515]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2214","title":"load_metric error: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'","comments":"Yep, seems to have fixed things! The conda package could really do with an update. Thanks!","body":"I'm having the same problem as [Notebooks issue 10](https:\/\/github.com\/huggingface\/notebooks\/issues\/10) on datasets 1.2.1, and it seems to be an issue with the datasets package.\r\n\r\n```python\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric(\"glue\", \"sst2\")\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 502, in load_metric\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 66, in import_main_class\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\", line 1014, in _gcd_import\r\n File \"\", line 991, in _find_and_load\r\n File \"\", line 975, in _find_and_load_unlocked\r\n File \"\", line 671, in _load_unlocked\r\n File \"\", line 783, in exec_module\r\n File \"\", line 219, in _call_with_frames_removed\r\n File \"\/home\/ns4008\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/glue\/e4606ab9804a36bcd5a9cebb2cb65bb14b6ac78ee9e6d5981fa679a495dd55de\/glue.py\", line 105, in \r\n @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)\r\nAttributeError: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\r\n```","comment_length":16,"text":"load_metric error: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\nI'm having the same problem as [Notebooks issue 10](https:\/\/github.com\/huggingface\/notebooks\/issues\/10) on datasets 1.2.1, and it seems to be an issue with the datasets package.\r\n\r\n```python\r\n>>> from datasets import load_metric\r\n>>> metric = load_metric(\"glue\", \"sst2\")\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 502, in load_metric\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/site-packages\/datasets-1.2.1-py3.8.egg\/datasets\/load.py\", line 66, in import_main_class\r\n File \"\/ext3\/miniconda3\/lib\/python3.8\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"\", line 1014, in _gcd_import\r\n File \"\", line 991, in _find_and_load\r\n File \"\", line 975, in _find_and_load_unlocked\r\n File \"\", line 671, in _load_unlocked\r\n File \"\", line 783, in exec_module\r\n File \"\", line 219, in _call_with_frames_removed\r\n File \"\/home\/ns4008\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/glue\/e4606ab9804a36bcd5a9cebb2cb65bb14b6ac78ee9e6d5981fa679a495dd55de\/glue.py\", line 105, in \r\n @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)\r\nAttributeError: module 'datasets.utils.file_utils' has no attribute 'add_start_docstrings'\r\n```\nYep, seems to have fixed things! The conda package could really do with an update. Thanks!","embeddings":[-0.272324115,-0.2135577202,0.0193226859,0.1852925718,0.4195098579,0.0630552247,0.2729536891,0.1801064909,0.0756967589,-0.0524457283,-0.1940380335,0.1545963436,-0.0744487196,0.2974149585,0.0652143732,-0.0471376516,-0.0315006822,0.0150832096,-0.262601763,0.0296820384,-0.3187800348,0.293637991,-0.1672726125,-0.1764028966,-0.4021217525,-0.0105557591,0.0437473059,0.2441147864,-0.2862702608,-0.5002132654,0.205599755,0.0462033674,0.2963437438,0.3439409733,-0.0001162612,-0.0840861276,0.2146514505,-0.0279947501,-0.5544195771,-0.1524079293,0.0528011806,-0.2994415164,0.2512985468,-0.1053964123,-0.026472494,-0.1421283036,-0.0965214372,-0.2981289327,0.3209556639,0.2984568179,0.2062049806,0.6519038081,0.1126084998,-0.3246344924,0.0756547675,-0.1170940176,-0.1389452964,0.5830135942,0.2155656517,-0.2970149517,0.0055574444,0.1997525692,-0.0833818167,0.2960377336,0.5132633448,0.0072548999,0.0939969942,-0.0558887757,0.1015874147,0.059948612,0.484824419,-0.3377355635,-0.3453686535,-0.2876414061,0.157133773,-0.317251116,0.2258811444,-0.1266947538,0.0611915551,0.1738776267,-0.1425379813,-0.2746182978,0.027385395,0.2549823523,-0.0566361211,-0.1325753331,-0.3288052678,-0.0292661712,0.3993551731,0.0067997873,0.0061149034,0.1106405407,0.0018888788,0.2454485446,-0.2312647104,0.1604048014,0.2210600972,0.0638860911,0.1995056421,0.1169555858,0.1385359019,-0.1091200337,0.2827876508,0.1445126384,0.0451146699,0.455306083,0.4384199381,0.0501264371,0.2727783024,0.3831758201,0.2283397615,-0.1032075882,-0.0001541478,-0.4193750322,-0.1396079361,-0.1696378291,0.3145554066,-0.0437624045,-0.339749068,0.2789826691,0.225033164,-0.1100322455,0.0559781827,0.166669026,0.0389566608,0.0680720434,0.3797704279,0.1804562062,-0.2567597926,0.0695359856,-0.2539496422,0.019209398,-0.2204780877,0.2743342817,0.2387660295,-0.1765834093,0.2692468762,0.1394805461,0.0810193792,0.0131133115,-0.1423486918,0.1412217319,-0.2486168146,0.2955060303,0.013345235,0.1941344142,0.3053650856,-0.3389752507,-0.0738557056,-0.2483933419,-0.2457935363,-0.0913328826,-0.2465759516,0.1607254595,-0.3777996302,0.1018846259,-0.1548740715,0.0269365031,0.1158767045,0.0955258533,-0.0533686727,-0.0219097435,-0.4126269221,-0.0537653193,0.3398079276,0.407076776,-0.2471117377,-0.4057824016,0.1503967047,-0.1388975978,-0.2114148438,0.0392381735,0.0853885636,0.2430645525,-0.2263119221,0.0026322715,0.1370444,-0.5137932301,-0.4204660356,0.1396769583,-0.1487744749,0.0827302411,0.0750858337,-0.1714534611,0.1146221235,0.183054179,0.4504910111,0.0914153084,0.1167476848,-0.1548795253,-0.290825367,-0.1838073581,-0.2629747391,0.2544320524,0.1082778797,0.0519852117,0.175834924,0.1697672904,0.0678901598,-0.0798934996,0.011294866,0.4932676554,0.1543736309,-0.0364747122,0.1711865067,-0.2008412182,-0.5201351643,0.2156841606,-0.0183965508,0.0993434042,-0.0899934471,-0.0668857396,-0.374559015,-0.0463788509,-0.0753306672,-0.1765413284,0.0452711396,-0.0109367613,-0.1645413339,0.3287498355,-0.1420029849,0.2980311811,-0.3087519705,0.3473673463,-0.3626068532,0.2142395675,0.0971087739,-0.1683486402,0.1748615503,0.1577347219,0.1103116572,-0.1772711575,-0.0718013197,0.4855265319,0.1505092531,0.0629772693,-0.0285120029,0.2165285498,0.1915651858,-0.1421236694,-0.1667533368,-0.0733661801,0.035070125,-0.0022641413,-0.0186236035,0.1603875458,-0.1758860499,0.0595270544,0.1123987362,0.1688022166,0.048878897,-0.0778958648,-0.1880658865,-0.3839529753,0.479842335,-0.0180930253,0.3275276124,-0.0315788649,0.1737259477,-0.1127126664,0.3555789888,0.0096923541,0.0877412483,0.1162704453,-0.308801055,0.1370434165,-0.0036531514,0.011657536,0.4693245888,0.2006489336,-0.0045757908,0.2940479517,0.0089875711,-0.0964115486,0.12861377,0.0178792831,-0.1744787693,0.2901944816,-0.0639035255,-0.0595521629,-0.3733405769,0.1549218893,-0.0791442096,0.2749669254,-0.5363030434,-0.1757649183,-0.141107738,0.0296398327,0.0277284384,-0.0767671391,-0.2003962547,-0.2440668344,-0.0292488467,0.325810194,0.1955318749,0.3660383523,0.2152761966,0.4211014807,-0.0384670012,0.0846681148,-0.1243118271,-0.2874588072,-0.0785380751,0.0081650047,0.1657315195,-0.0335034579,0.1387250274,-0.6127305627,0.06443654,-0.236285001,-0.3688507378,0.0849087164,-0.0047380268,0.6359252334,0.3205825984,0.1494532824,-0.0961013958,0.0279980153,0.3684045076,0.0242663678,-0.2071206421,0.0681154877,-0.0857632682,-0.2336559445,-0.3043382466,-0.1047391817,0.0186322648,-0.503063798,0.3876099288,0.0402854495,-0.2336942852,0.3373192847,0.0795239285,0.2915133536,0.0523024015,0.2416902632,-0.261346221,-0.3416886032,0.2696045637,-0.2346160859,-0.4145871997,-0.037765462,0.1373154223,0.1808706522,-0.0452895761,-0.4627621174,-0.4108830392,-0.1143371314,0.1598947495,-0.1194815487,0.3886238635,0.1906858832,0.0357468687,0.0353719592,-0.3328506052,-0.1745150089,0.0547360964,-0.1960035712,0.0167503189,-0.2609972954,0.1186313927,0.0603463016,0.3715052307,0.2424533367,-0.0580785982,0.2844280005,-0.1521651447,0.6900770664,-0.1171550304,-0.5205394626,0.070948638,0.0032520145,0.2066455036,0.294349432,0.0683957711,0.3357643485,-0.2086580098,0.0799778104,-0.107449159,-0.1985200495,-0.1780722588,0.0575307794,0.2056672275,0.043477606,0.2509680092,-0.1185469553,0.0546496622,0.1190925911,0.4457848966,-0.0582257807,0.0248067491,-0.396876663,0.1530397236,-0.4186137915,0.3114287555,-0.0233412813,0.2591668963,-0.2117097825,-0.1854931265,0.0008127582,0.0003988862,0.4743218124,-0.0232882611,-0.1957427561,0.2003183365,-0.0690099299,-0.685212791,-0.0052718553,-0.0417399704,-0.0789852813,0.1828616858,0.7977381349,-0.2625100315,-0.3014018238,0.1301297694,0.1023930162,-0.0043754182,-0.0999475941,-0.3193716407,-0.3368156254,-0.1378110796,-0.0916927457,-0.0276012328,0.2983665466,0.1582421064,0.2091055214,-0.1144397929,-0.2217674553,0.1712339669,0.1866385192,0.1943307966,-0.0234470051,0.3947012126,0.3202777803,-0.0763602033,0.4250121713,0.4037992954,-0.4309894443,-0.6466447711,0.1787409633,-0.0535439402,0.1202298105,0.157498464,-0.2035825402,0.0360497124,-0.0532159917,0.2204992473,-0.4408127964,0.2652892768,0.4414024055,0.1872788072,-0.0870289728,-0.1726748943,0.306450218,0.0870726556,0.00999304,0.4308449328,0.0634560212,-0.2938775718,-0.0701917261,-0.2305243164,0.5940442681,0.0290086195,0.1729602814,0.2764141262,-0.1854176819,0.4218051434,-0.0393425412,-0.0951728448,-0.2183059305,-0.4836227894,0.0405451246,-0.0413951837,0.2808120549,-0.2143423706,-0.1116682589,0.0950956792,-0.2373245806,0.0704354271,-0.0563032404,0.2277140021,-0.2661421895,-0.1809367836,-0.2217793018,0.122424826,-0.1408931613,0.2309868932,-0.1256341487,0.1210680231,-0.0697817206,-0.2212692052,-0.1227326468,-0.1291439235,-0.2800857723,0.144272238,0.2374107093,-0.1632544696,-0.0273615066,0.3203982413,0.4834209681,0.0730526522,-0.3700918257,0.20750314,-0.3207891285,-0.0973202959,-0.0424925014,0.0715611503,0.2013940662,0.0530964024,0.0595653243,-0.1689048558,0.0645233318,0.2031840235,0.0174677745,-0.0456067212,-0.0876971036,-0.4919409156,-0.1246381253,-0.0510219261,0.2836449742,-0.327545166,0.0981150642,0.0562658198,0.014764308,0.1047228202,-0.0331803784,-0.242884919,-0.0733277053,0.6832268834,-0.361394614,-0.1414557099,0.1263989955,0.1018077284,-0.318993628,-0.1406425238,0.2113707811,0.2389927953,-0.3433932364,0.2778002918,0.5266761184,0.0435032248,-0.0637036711,0.4369817674,0.2357449532,-0.3535464406,0.1269493997,-0.2212594002,-0.3677404225,0.2803766131,-0.1661068797,0.1610450149,0.0869790614,0.1942923218,-0.1050552502,-0.0148057314,-0.2793801129,0.211676687,-0.0993371308,-0.1408014148,0.6002429724,-0.2447780818,0.2816815078,-0.1756057292,0.0771706775,0.015653275,-0.3317786157,-0.1820966899,-0.4139865637,0.1064253449,-0.0538477972,-0.1300275624,0.2956237793,0.0028885128,-0.0922225863,-0.2934277952,0.1935119033,0.4534095228,-0.046222426,0.2377129346,-0.122749649,0.232391119,0.1859098375,0.1401135325,-0.100802131,0.0997574255,-0.0606916659,0.0412759297,-0.1847326607,0.2028342634,0.0222671926,0.2372679859,0.1615038514,-0.1148739532,0.1546191126,-0.1380564868,0.0113244327,-0.0168448221,0.3367970586,0.2945950031,-0.2881359756,-0.0667470321,0.1050204039,0.1873946935,-0.2524309754,-0.2942762375,0.0593587644,0.0206503924,0.1127656773,0.1540552378,0.1536152661,-0.1455436349,0.0944109932,0.0923186019,0.3824128211,0.1105610281,0.170265317,0.1636682898,0.1364786774,-0.0244052391,0.3797616065,0.0976049304,0.2229423076,0.0688975453,-0.4428275526,0.0882378817,-0.2641712427,0.3839266896,-0.2432175875,-0.4694573879,0.0781102702,0.0745861009,-0.0688807964,-0.0325670131,-0.3320102096,0.7181387544,-0.3015743792,-0.0596848205,-0.0467802659,0.0559419021,-0.1177831888,-0.1937913001,-0.0985112488,-0.1023545563,-0.0581151396,-0.0594300516,-0.0467668734,-0.1459453106,0.3599132895,-0.1085797548,-0.0336290449,-0.3310728967,-0.3075093031,0.2592685521,-0.1198569462,-0.132297948,0.2339993864,-0.1117595658,0.2395399362,0.0096039288,0.6236209869,0.4861671031,0.0578343943,0.1642263532,-0.1270774752,-0.1182659939,0.0369492695,-0.0820896849,0.4027970135,0.0168094747,0.1449293196,0.1828604192,0.1365623325,-0.2118962109,0.1529088318,-0.2281762213,-0.3224211037,-0.3317995965,0.0626964718,-0.2673926651,-0.0225482862,-0.130914852,0.114061676,-0.3242834806,0.1154515967,0.408370167,0.1985117197,0.08757101,-0.1523372978,0.0868726298,-0.053498473,0.1909273863,0.400077939,0.0367470011,-0.2721703053,-0.2734406888,-0.6236268878,0.0099143265,0.1216766313,-0.0570753515,0.1204302981,-0.0348658748,0.0610775165,-0.0391001515,0.1405400932,0.1027174741,-0.0041313292,-0.2952067256,-0.4408206344,0.0197271816,-0.2038558275,-0.101773493,0.0670962855,-0.2293940634,-0.0436051041,-0.1830174476,-0.02075647,-0.1345573664,0.2511404753,0.0660701469,0.0035569319,0.4722623229,-0.0868728086,0.3599009216,-0.0522920601,-0.0728004128,-0.1637538671,-0.4566641748,-0.0161271878,0.1952701509,-0.0664701834,0.2887929678,-0.4623144269,-0.3876182139,-0.2914842367,0.5101401806,0.2518330514,-0.0921489224,-0.2270249575,0.337551713,-0.1708122492,0.0662507117,0.1981481314,0.3309055269,-0.1810210496,0.3260757923,-0.1189147606,-0.5026894212,0.6233594418,-0.3565818667,-0.0996360332,0.2245274782,0.288562417,0.3277512789,-0.3560940027,-0.60240376,-0.057727553,0.3717492819,0.0284310561,-0.506750524,0.591793716,-0.271661669,0.1035222337,-0.0543052293,0.1474553049,0.2477992475,-0.2005041838,0.2222824991,-0.2424782515]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2212","title":"Can't reach \"https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\" when trying to load fquad dataset","comments":"Hi ! Apparently the data are not available from this url anymore. We'll replace it with the new url when it's available","body":"I'm trying to load the [fquad dataset](https:\/\/huggingface.co\/datasets\/fquad) by running: \r\n\r\n```Python\r\nfquad = load_dataset(\"fquad\")\r\n```\r\n\r\nwhich produces the following error:\r\n\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset fquad\/default (download: 3.14 MiB, generated: 6.62 MiB, post-processed: Unknown size, total: 9.76 MiB) to \/root\/.cache\/huggingface\/datasets\/fquad\/default\/0.1.0\/778dc2c85813d05ddd0c17087294d5f8f24820752340958070876b677af9f061...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nConnectionError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 fquad = load_dataset(\"fquad\")\r\n\r\n11 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 617 \r\n 618 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\r\n```\r\n\r\nDoes anyone know why that is and how to fix it? ","comment_length":22,"text":"Can't reach \"https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\" when trying to load fquad dataset\nI'm trying to load the [fquad dataset](https:\/\/huggingface.co\/datasets\/fquad) by running: \r\n\r\n```Python\r\nfquad = load_dataset(\"fquad\")\r\n```\r\n\r\nwhich produces the following error:\r\n\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset fquad\/default (download: 3.14 MiB, generated: 6.62 MiB, post-processed: Unknown size, total: 9.76 MiB) to \/root\/.cache\/huggingface\/datasets\/fquad\/default\/0.1.0\/778dc2c85813d05ddd0c17087294d5f8f24820752340958070876b677af9f061...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nConnectionError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 fquad = load_dataset(\"fquad\")\r\n\r\n11 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 617 \r\n 618 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\r\n```\r\n\r\nDoes anyone know why that is and how to fix it? \nHi ! Apparently the data are not available from this url anymore. We'll replace it with the new url when it's available","embeddings":[-0.3520766795,0.1850519031,-0.1066456437,0.2455794364,0.3998158276,0.03360755,0.3876356781,0.2061039656,0.3057492673,0.1453926563,-0.2683598399,-0.1908080578,0.2827515602,0.0175514612,0.027606599,0.0714482665,-0.0717816278,-0.0440939777,-0.1596578062,0.0708816275,-0.3542164862,0.3704529405,-0.1904580444,0.1628842503,-0.3339352906,0.0633746386,0.0055798003,0.3590935767,-0.3199560642,-0.1079104543,0.286209017,0.0416278467,0.004151077,0.4852188826,-0.0001157215,0.1660642475,0.3586290479,-0.0317648053,-0.4298180938,-0.5632392168,-0.3045651317,-0.072570838,0.2153185606,-0.2087870389,-0.1616279483,-0.0290350895,0.1196179837,-0.4198773503,0.2603788972,0.4918957353,0.1842911541,-0.0472500958,0.2975516617,-0.1803262234,-0.0591369905,-0.1362888217,-0.0200078152,0.6450526714,-0.0140431896,-0.2045283467,-0.0508274771,0.0286189858,0.1182557642,0.1726587862,0.1585940719,0.2011903822,-0.3372362256,-0.1506722569,0.2118362486,0.0146889491,0.5337827802,-0.2849702537,-0.4723348618,-0.1233135611,-0.0491205491,-0.3767210245,0.5024482012,0.0925156698,-0.2628643513,0.1907122433,-0.4664973021,-0.2508746386,-0.1997162253,0.3678171635,-0.3860652447,0.2701286077,-0.1864978373,0.2325517088,0.2770298421,-0.1557939649,-0.2593966126,0.1098564789,0.1527526826,0.3186956048,-0.314109236,0.1022981256,-0.0440244749,-0.2830099761,0.1964606792,0.3815691173,-0.015078485,0.0076378272,-0.0969404653,0.164527759,0.4020113349,0.3189195991,-0.1120543703,0.2087809891,0.1873425245,0.6325322986,-0.1311301142,-0.1770190001,-0.1378113925,-0.081372045,-0.1658248156,-0.183514595,0.1537760198,-0.1985757202,-0.1576269418,0.1740346998,-0.2624880672,-0.0844819471,0.0354961865,0.3024871647,-0.0897951052,0.1091868505,0.0184984915,0.2455940992,-0.1006541178,0.1638454646,-0.1644030064,0.2479628474,-0.218215555,0.231559068,0.2416772842,0.1246844679,0.3092523515,-0.3974604607,0.0709448978,-0.0954510868,0.2211935818,-0.2454142272,-0.156402275,0.3248185813,0.1361620575,-0.019331608,0.1892515123,-0.0819873288,-0.120350033,-0.0591272786,-0.238079235,-0.2690150142,-0.1610825509,0.1693099141,-0.1822194904,0.0440686196,-0.3180564344,0.0211405642,-0.163014695,-0.2929230332,-0.02649194,0.1383045167,-0.1317382902,-0.1248099804,0.3689207137,0.5510037541,-0.2626015544,-0.0661597773,-0.2078201175,-0.0995471701,-0.0146748368,0.4244317114,-0.174391821,0.2035565674,-0.4347251058,0.2382106483,0.4783934355,-0.2472981364,-0.8428885341,0.427974999,-0.2450462729,0.1419134438,-0.091932185,0.0322814882,-0.0358391963,0.1951994896,0.6311691403,0.2926326096,-0.0661307499,-0.127931118,-0.0659748837,-0.3087247014,-0.0573887713,0.2728063464,-0.0072478638,0.173291862,0.2028902769,-0.11355827,0.203204602,0.1464847624,-0.1294753551,0.2984036505,0.0654904172,0.0501435176,-0.1333141923,-0.0529266,-0.6362524629,0.2412666529,-0.2172224224,-0.0310093351,-0.5615538359,0.2495313436,-0.4475069642,0.0381114259,-0.0719014481,0.0171304978,0.0995740071,0.0793238357,0.1675628722,0.1997745037,-0.2326557487,0.3575578332,0.0593538806,0.2285772711,-0.4177794456,0.3038277328,-0.068603389,-0.0128734708,-0.0401704386,0.0375095345,0.0876060575,-0.1907199472,-0.1723519415,0.3521072268,-0.0981610119,0.1596617848,0.3979538083,0.4166022539,0.1592304558,-0.0419741049,0.0967004448,-0.1133250967,0.0562817417,-0.0873886049,-0.1806227118,0.1824068427,0.1226353049,0.3023006022,0.140295893,0.033637926,0.365693301,0.1180057302,-0.0833135769,-0.1543994844,0.15414083,0.4664791822,0.2170068771,0.0557061583,-0.1387757808,-0.0494213514,0.3819881976,-0.2163211554,0.0133876475,0.2899906933,0.0413509086,-0.2723265886,0.1269237697,0.0862024873,0.4915738106,0.0586339347,0.1429887861,0.1387392431,-0.2469654232,-0.1638064682,0.0784770474,0.0087313019,0.0382009223,0.0595250651,0.1666518003,-0.0206933133,-0.1088888422,-0.3694197536,0.0472553745,0.1783639491,-0.1222572029,0.3422809839,-0.1195476055,-0.1523807496,0.1046303585,-0.3231037259,-0.2455987036,-0.4834875762,-0.1251465082,0.2941145301,0.0920701548,0.0488251857,-0.3312270641,0.0984709337,0.1856976151,-0.5195731521,-0.1338463426,-0.1513882279,-0.1231954619,0.0563571267,0.3074066341,-0.2006785274,0.2499470264,-0.2296645343,-0.100894548,-0.4725251794,-0.1700989753,0.0898433775,-0.0312449019,0.40792799,0.2447232157,0.3579797447,-0.0674479678,-0.0232894085,0.4561803341,0.0728968084,-0.0803323686,-0.0389703028,-0.1058985814,0.1410915405,0.054669302,-0.263803184,-0.2067806423,-0.3587554097,0.4501703084,0.0336139351,0.0696525946,0.1306530684,0.1599377245,0.2007167935,0.1608746797,-0.1496057361,-0.3580869138,-0.5675964355,0.4216380119,-0.2252491564,-0.062960811,0.0954560116,0.1260264516,0.4072681069,0.3108044267,-0.5964596272,-0.4336869121,-0.2441263199,0.2438689023,-0.0759851113,-0.2841826975,0.2059355229,-0.2813442647,-0.0455122963,0.0795237049,-0.3435887992,0.1072266921,0.0380031541,0.4218061566,0.1905358434,0.5441442132,-0.1006031409,0.596581161,0.0506761186,0.1118795276,0.3819426298,-0.1311579049,0.2393558174,-0.4368827343,-0.2892046869,0.1354883611,-0.0735363513,-0.3199809194,-0.1010950878,0.1610036194,-0.0667655841,-0.2432231307,0.2578684688,-0.4428741932,-0.2013871372,0.043550387,-0.0229954161,0.0392960124,0.1748021841,0.0631838068,-0.1840065271,-0.0513392054,-0.0495904051,0.1898236424,0.0814525932,0.1691627055,-0.0057752947,-0.3805061877,-0.5242769718,0.408682853,0.0688265115,0.3191053867,-0.2343497723,0.0901854932,0.159983471,-0.0714382604,0.5216056108,0.0438570678,-0.0242689401,0.2260752767,-0.0639998466,-0.2260826379,0.0408335701,0.1604545265,0.2876725197,0.2784393728,0.5854113698,-0.4355755746,-0.0223598778,0.2736801207,0.0492445305,-0.0857091174,-0.0519774668,-0.1963580996,-0.4926885366,-0.345927,-0.2309070528,0.0583768375,0.3225263953,0.2719903886,-0.1396180838,0.1985048801,-0.2339979112,-0.123792477,0.0856428072,0.2614714205,-0.1848341525,-0.0721978843,0.446174711,0.0551759638,0.0625965148,0.9526379108,0.2309765965,-0.1918456703,0.3530420959,0.2041083425,-0.0470772423,0.2335586697,-0.0544416867,-0.2276651412,0.0475472696,0.3281215131,-0.0701255649,0.3138429821,0.3014744818,-0.0341497324,-0.2183242589,-0.3774331212,0.5831079483,-0.0610893443,0.0716049373,0.5631733537,-0.0031045412,-0.1457767785,-0.0282859821,-0.11133793,0.8322306871,-0.2192073613,0.1396558434,0.3237110972,-0.2258144915,0.2849935293,-0.3626310825,0.1815250069,-0.1902487278,0.055939123,-0.1027311981,-0.1027690098,0.0169316493,0.2292079628,0.1400343627,0.4701394737,-0.0622956306,-0.0413589627,0.2022134811,0.2214944363,-0.1957224309,-0.03502509,-0.4555773437,0.1509607136,-0.063500762,0.3048473299,0.0043505183,-0.0404348522,-0.2007136643,-0.1768927574,-0.3790899217,0.1039981544,-0.4450157583,0.3993927836,0.064871639,-0.1194170266,-0.3323346078,0.1331066787,-0.1718951911,0.0374773331,-0.377406925,0.2002423555,-0.250582844,-0.1914522499,0.0667510033,0.2090218812,0.1672664881,-0.0286223646,-0.1082126722,0.2219368368,-0.1580397338,-0.4935772419,-0.0861230642,-0.0270098094,0.1342190057,-0.1137048602,-0.1250256896,-0.3386538327,-0.204316318,-0.0527786091,0.0808490962,0.028858887,0.2523483336,-0.2069956809,0.0026370073,-0.2918313742,-0.1399932802,0.5058478117,-0.2566277981,-0.3374788463,0.4631577432,0.0977263302,-0.0907071531,-0.1474872082,-0.1863509417,-0.0134128286,-0.1456059068,0.0908238962,0.0119757038,0.1186038107,-0.2104348689,-0.0244327877,0.2548632026,-0.1683793068,0.046724651,-0.5118241906,-0.4071712792,0.2717998028,-0.2288030982,0.1913926899,0.0034555777,-0.0871694088,0.173056066,-0.2681245506,-0.2592228055,0.1471401006,-0.4422644079,-0.2400080115,0.2411803603,-0.1960831136,0.4321490824,0.0995835066,0.1983939856,-0.0372604281,-0.2293649018,-0.1816976815,0.0783660114,0.1327116042,0.1031492278,-0.4042631388,0.0659462512,-0.0905284211,-0.1013431326,-0.1443170309,0.0145540675,0.2522515655,0.1482552588,-0.2715089619,0.1424651891,0.0433753803,-0.1513764858,0.0312853009,0.1540951729,0.3858402669,0.1071474254,0.0385405049,-0.012002415,0.0421502218,-0.1947081685,0.0288459379,0.0366724245,0.1114400104,0.5246019363,-0.2630036473,0.0688438863,0.1733348519,0.2766209543,0.3434119523,-0.393322587,-0.0504622869,0.0861113518,0.1886861473,-0.1005256698,-0.0409392379,0.1158134937,-0.0200303271,-0.0530431531,0.0752635002,0.0275851823,0.0262580942,0.1938834339,0.0303886868,0.4034167826,0.0132632684,0.0182413943,-0.0686375201,-0.1445128173,0.0048207832,0.3360909522,0.1343934685,0.3272197843,0.4190325737,-0.1450155079,-0.0987357721,-0.3855920732,0.2788904011,-0.2114679515,-0.4335516095,-0.2508144975,0.2423272133,-0.3047325015,0.2247778773,-0.2435577214,0.2378086001,-0.4140954912,0.2134464532,-0.2503335774,0.0139560569,-0.0381372944,-0.1202731133,0.0460858718,-0.1928994507,-0.1554569602,0.0292251222,-0.0773304701,-0.1959445477,0.1673088223,-0.1012982503,-0.1600542814,-0.2908314466,0.1489149928,0.0585049614,0.1058554426,-0.2569085658,0.500236094,0.1337188333,-0.1236687973,0.259260118,0.2547582388,0.3921833038,0.2354182303,-0.0563612543,0.118781276,-0.1520817727,-0.1580698192,-0.1649380028,0.22200571,0.2826267481,-0.2399268299,0.40833956,0.220150277,-0.1681924611,0.3380620778,-0.0190949738,0.4372432232,-0.2133324146,0.3192174733,-0.3780584931,0.0575311445,-0.3101490736,-0.0832276717,-0.5760593414,-0.2530725598,-0.1635162681,-0.101671882,0.1617380083,-0.0138189169,0.0595093146,0.0480281673,0.3779025376,0.4761712253,0.1966776699,-0.2976574898,-0.2617687881,-0.5369645953,0.0845123827,-0.0744423419,0.0407859534,-0.0126415752,0.3252759278,-0.143486008,0.1179323345,0.1704975367,0.175069809,-0.2017744631,0.1189973131,-0.2758839428,-0.416370213,-0.0354897715,0.1237288564,-0.0925103053,-0.1620142162,0.1815119684,-0.2884228528,-0.0075483653,0.0063050175,0.1264898777,0.0303840991,-0.0279789791,0.4688521624,0.1719368994,0.4205750823,0.1091431305,-0.2500298619,-0.3299544752,-0.1993986368,-0.0289107151,0.0117749656,-0.0593902618,0.456469506,-0.0293870457,0.0285883415,-0.2617828548,0.1135850996,-0.0730579421,-0.1677081287,0.0184928607,-0.1089656875,0.0220410693,0.140310213,-0.00121685,0.3501479924,0.0503059253,0.2221424431,-0.0946951881,-0.1574622244,0.5106208324,-0.0712286085,-0.2704927921,0.2154466212,0.0052094804,-0.0617787167,-0.0020324003,-0.3609396815,0.2056361735,0.4209412038,0.079865478,-0.1565104872,0.0797880292,-0.1106195226,-0.0691273436,-0.1145044044,0.2245735824,-0.104605332,-0.1141530871,0.05216261,0.0729801282]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2212","title":"Can't reach \"https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\" when trying to load fquad dataset","comments":"I saw this on their website when we request to download the dataset:\r\n![image](https:\/\/user-images.githubusercontent.com\/19718818\/114879600-fa458680-9e1e-11eb-9e05-f0963d68ff0f.png)\r\n\r\nCan we still request them link for the dataset and make a PR? @lhoestq @yjernite ","body":"I'm trying to load the [fquad dataset](https:\/\/huggingface.co\/datasets\/fquad) by running: \r\n\r\n```Python\r\nfquad = load_dataset(\"fquad\")\r\n```\r\n\r\nwhich produces the following error:\r\n\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset fquad\/default (download: 3.14 MiB, generated: 6.62 MiB, post-processed: Unknown size, total: 9.76 MiB) to \/root\/.cache\/huggingface\/datasets\/fquad\/default\/0.1.0\/778dc2c85813d05ddd0c17087294d5f8f24820752340958070876b677af9f061...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nConnectionError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 fquad = load_dataset(\"fquad\")\r\n\r\n11 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 617 \r\n 618 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\r\n```\r\n\r\nDoes anyone know why that is and how to fix it? ","comment_length":29,"text":"Can't reach \"https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\" when trying to load fquad dataset\nI'm trying to load the [fquad dataset](https:\/\/huggingface.co\/datasets\/fquad) by running: \r\n\r\n```Python\r\nfquad = load_dataset(\"fquad\")\r\n```\r\n\r\nwhich produces the following error:\r\n\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset fquad\/default (download: 3.14 MiB, generated: 6.62 MiB, post-processed: Unknown size, total: 9.76 MiB) to \/root\/.cache\/huggingface\/datasets\/fquad\/default\/0.1.0\/778dc2c85813d05ddd0c17087294d5f8f24820752340958070876b677af9f061...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nConnectionError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 fquad = load_dataset(\"fquad\")\r\n\r\n11 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 617 \r\n 618 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\r\n```\r\n\r\nDoes anyone know why that is and how to fix it? \nI saw this on their website when we request to download the dataset:\r\n![image](https:\/\/user-images.githubusercontent.com\/19718818\/114879600-fa458680-9e1e-11eb-9e05-f0963d68ff0f.png)\r\n\r\nCan we still request them link for the dataset and make a PR? @lhoestq @yjernite ","embeddings":[-0.3520766795,0.1850519031,-0.1066456437,0.2455794364,0.3998158276,0.03360755,0.3876356781,0.2061039656,0.3057492673,0.1453926563,-0.2683598399,-0.1908080578,0.2827515602,0.0175514612,0.027606599,0.0714482665,-0.0717816278,-0.0440939777,-0.1596578062,0.0708816275,-0.3542164862,0.3704529405,-0.1904580444,0.1628842503,-0.3339352906,0.0633746386,0.0055798003,0.3590935767,-0.3199560642,-0.1079104543,0.286209017,0.0416278467,0.004151077,0.4852188826,-0.0001157215,0.1660642475,0.3586290479,-0.0317648053,-0.4298180938,-0.5632392168,-0.3045651317,-0.072570838,0.2153185606,-0.2087870389,-0.1616279483,-0.0290350895,0.1196179837,-0.4198773503,0.2603788972,0.4918957353,0.1842911541,-0.0472500958,0.2975516617,-0.1803262234,-0.0591369905,-0.1362888217,-0.0200078152,0.6450526714,-0.0140431896,-0.2045283467,-0.0508274771,0.0286189858,0.1182557642,0.1726587862,0.1585940719,0.2011903822,-0.3372362256,-0.1506722569,0.2118362486,0.0146889491,0.5337827802,-0.2849702537,-0.4723348618,-0.1233135611,-0.0491205491,-0.3767210245,0.5024482012,0.0925156698,-0.2628643513,0.1907122433,-0.4664973021,-0.2508746386,-0.1997162253,0.3678171635,-0.3860652447,0.2701286077,-0.1864978373,0.2325517088,0.2770298421,-0.1557939649,-0.2593966126,0.1098564789,0.1527526826,0.3186956048,-0.314109236,0.1022981256,-0.0440244749,-0.2830099761,0.1964606792,0.3815691173,-0.015078485,0.0076378272,-0.0969404653,0.164527759,0.4020113349,0.3189195991,-0.1120543703,0.2087809891,0.1873425245,0.6325322986,-0.1311301142,-0.1770190001,-0.1378113925,-0.081372045,-0.1658248156,-0.183514595,0.1537760198,-0.1985757202,-0.1576269418,0.1740346998,-0.2624880672,-0.0844819471,0.0354961865,0.3024871647,-0.0897951052,0.1091868505,0.0184984915,0.2455940992,-0.1006541178,0.1638454646,-0.1644030064,0.2479628474,-0.218215555,0.231559068,0.2416772842,0.1246844679,0.3092523515,-0.3974604607,0.0709448978,-0.0954510868,0.2211935818,-0.2454142272,-0.156402275,0.3248185813,0.1361620575,-0.019331608,0.1892515123,-0.0819873288,-0.120350033,-0.0591272786,-0.238079235,-0.2690150142,-0.1610825509,0.1693099141,-0.1822194904,0.0440686196,-0.3180564344,0.0211405642,-0.163014695,-0.2929230332,-0.02649194,0.1383045167,-0.1317382902,-0.1248099804,0.3689207137,0.5510037541,-0.2626015544,-0.0661597773,-0.2078201175,-0.0995471701,-0.0146748368,0.4244317114,-0.174391821,0.2035565674,-0.4347251058,0.2382106483,0.4783934355,-0.2472981364,-0.8428885341,0.427974999,-0.2450462729,0.1419134438,-0.091932185,0.0322814882,-0.0358391963,0.1951994896,0.6311691403,0.2926326096,-0.0661307499,-0.127931118,-0.0659748837,-0.3087247014,-0.0573887713,0.2728063464,-0.0072478638,0.173291862,0.2028902769,-0.11355827,0.203204602,0.1464847624,-0.1294753551,0.2984036505,0.0654904172,0.0501435176,-0.1333141923,-0.0529266,-0.6362524629,0.2412666529,-0.2172224224,-0.0310093351,-0.5615538359,0.2495313436,-0.4475069642,0.0381114259,-0.0719014481,0.0171304978,0.0995740071,0.0793238357,0.1675628722,0.1997745037,-0.2326557487,0.3575578332,0.0593538806,0.2285772711,-0.4177794456,0.3038277328,-0.068603389,-0.0128734708,-0.0401704386,0.0375095345,0.0876060575,-0.1907199472,-0.1723519415,0.3521072268,-0.0981610119,0.1596617848,0.3979538083,0.4166022539,0.1592304558,-0.0419741049,0.0967004448,-0.1133250967,0.0562817417,-0.0873886049,-0.1806227118,0.1824068427,0.1226353049,0.3023006022,0.140295893,0.033637926,0.365693301,0.1180057302,-0.0833135769,-0.1543994844,0.15414083,0.4664791822,0.2170068771,0.0557061583,-0.1387757808,-0.0494213514,0.3819881976,-0.2163211554,0.0133876475,0.2899906933,0.0413509086,-0.2723265886,0.1269237697,0.0862024873,0.4915738106,0.0586339347,0.1429887861,0.1387392431,-0.2469654232,-0.1638064682,0.0784770474,0.0087313019,0.0382009223,0.0595250651,0.1666518003,-0.0206933133,-0.1088888422,-0.3694197536,0.0472553745,0.1783639491,-0.1222572029,0.3422809839,-0.1195476055,-0.1523807496,0.1046303585,-0.3231037259,-0.2455987036,-0.4834875762,-0.1251465082,0.2941145301,0.0920701548,0.0488251857,-0.3312270641,0.0984709337,0.1856976151,-0.5195731521,-0.1338463426,-0.1513882279,-0.1231954619,0.0563571267,0.3074066341,-0.2006785274,0.2499470264,-0.2296645343,-0.100894548,-0.4725251794,-0.1700989753,0.0898433775,-0.0312449019,0.40792799,0.2447232157,0.3579797447,-0.0674479678,-0.0232894085,0.4561803341,0.0728968084,-0.0803323686,-0.0389703028,-0.1058985814,0.1410915405,0.054669302,-0.263803184,-0.2067806423,-0.3587554097,0.4501703084,0.0336139351,0.0696525946,0.1306530684,0.1599377245,0.2007167935,0.1608746797,-0.1496057361,-0.3580869138,-0.5675964355,0.4216380119,-0.2252491564,-0.062960811,0.0954560116,0.1260264516,0.4072681069,0.3108044267,-0.5964596272,-0.4336869121,-0.2441263199,0.2438689023,-0.0759851113,-0.2841826975,0.2059355229,-0.2813442647,-0.0455122963,0.0795237049,-0.3435887992,0.1072266921,0.0380031541,0.4218061566,0.1905358434,0.5441442132,-0.1006031409,0.596581161,0.0506761186,0.1118795276,0.3819426298,-0.1311579049,0.2393558174,-0.4368827343,-0.2892046869,0.1354883611,-0.0735363513,-0.3199809194,-0.1010950878,0.1610036194,-0.0667655841,-0.2432231307,0.2578684688,-0.4428741932,-0.2013871372,0.043550387,-0.0229954161,0.0392960124,0.1748021841,0.0631838068,-0.1840065271,-0.0513392054,-0.0495904051,0.1898236424,0.0814525932,0.1691627055,-0.0057752947,-0.3805061877,-0.5242769718,0.408682853,0.0688265115,0.3191053867,-0.2343497723,0.0901854932,0.159983471,-0.0714382604,0.5216056108,0.0438570678,-0.0242689401,0.2260752767,-0.0639998466,-0.2260826379,0.0408335701,0.1604545265,0.2876725197,0.2784393728,0.5854113698,-0.4355755746,-0.0223598778,0.2736801207,0.0492445305,-0.0857091174,-0.0519774668,-0.1963580996,-0.4926885366,-0.345927,-0.2309070528,0.0583768375,0.3225263953,0.2719903886,-0.1396180838,0.1985048801,-0.2339979112,-0.123792477,0.0856428072,0.2614714205,-0.1848341525,-0.0721978843,0.446174711,0.0551759638,0.0625965148,0.9526379108,0.2309765965,-0.1918456703,0.3530420959,0.2041083425,-0.0470772423,0.2335586697,-0.0544416867,-0.2276651412,0.0475472696,0.3281215131,-0.0701255649,0.3138429821,0.3014744818,-0.0341497324,-0.2183242589,-0.3774331212,0.5831079483,-0.0610893443,0.0716049373,0.5631733537,-0.0031045412,-0.1457767785,-0.0282859821,-0.11133793,0.8322306871,-0.2192073613,0.1396558434,0.3237110972,-0.2258144915,0.2849935293,-0.3626310825,0.1815250069,-0.1902487278,0.055939123,-0.1027311981,-0.1027690098,0.0169316493,0.2292079628,0.1400343627,0.4701394737,-0.0622956306,-0.0413589627,0.2022134811,0.2214944363,-0.1957224309,-0.03502509,-0.4555773437,0.1509607136,-0.063500762,0.3048473299,0.0043505183,-0.0404348522,-0.2007136643,-0.1768927574,-0.3790899217,0.1039981544,-0.4450157583,0.3993927836,0.064871639,-0.1194170266,-0.3323346078,0.1331066787,-0.1718951911,0.0374773331,-0.377406925,0.2002423555,-0.250582844,-0.1914522499,0.0667510033,0.2090218812,0.1672664881,-0.0286223646,-0.1082126722,0.2219368368,-0.1580397338,-0.4935772419,-0.0861230642,-0.0270098094,0.1342190057,-0.1137048602,-0.1250256896,-0.3386538327,-0.204316318,-0.0527786091,0.0808490962,0.028858887,0.2523483336,-0.2069956809,0.0026370073,-0.2918313742,-0.1399932802,0.5058478117,-0.2566277981,-0.3374788463,0.4631577432,0.0977263302,-0.0907071531,-0.1474872082,-0.1863509417,-0.0134128286,-0.1456059068,0.0908238962,0.0119757038,0.1186038107,-0.2104348689,-0.0244327877,0.2548632026,-0.1683793068,0.046724651,-0.5118241906,-0.4071712792,0.2717998028,-0.2288030982,0.1913926899,0.0034555777,-0.0871694088,0.173056066,-0.2681245506,-0.2592228055,0.1471401006,-0.4422644079,-0.2400080115,0.2411803603,-0.1960831136,0.4321490824,0.0995835066,0.1983939856,-0.0372604281,-0.2293649018,-0.1816976815,0.0783660114,0.1327116042,0.1031492278,-0.4042631388,0.0659462512,-0.0905284211,-0.1013431326,-0.1443170309,0.0145540675,0.2522515655,0.1482552588,-0.2715089619,0.1424651891,0.0433753803,-0.1513764858,0.0312853009,0.1540951729,0.3858402669,0.1071474254,0.0385405049,-0.012002415,0.0421502218,-0.1947081685,0.0288459379,0.0366724245,0.1114400104,0.5246019363,-0.2630036473,0.0688438863,0.1733348519,0.2766209543,0.3434119523,-0.393322587,-0.0504622869,0.0861113518,0.1886861473,-0.1005256698,-0.0409392379,0.1158134937,-0.0200303271,-0.0530431531,0.0752635002,0.0275851823,0.0262580942,0.1938834339,0.0303886868,0.4034167826,0.0132632684,0.0182413943,-0.0686375201,-0.1445128173,0.0048207832,0.3360909522,0.1343934685,0.3272197843,0.4190325737,-0.1450155079,-0.0987357721,-0.3855920732,0.2788904011,-0.2114679515,-0.4335516095,-0.2508144975,0.2423272133,-0.3047325015,0.2247778773,-0.2435577214,0.2378086001,-0.4140954912,0.2134464532,-0.2503335774,0.0139560569,-0.0381372944,-0.1202731133,0.0460858718,-0.1928994507,-0.1554569602,0.0292251222,-0.0773304701,-0.1959445477,0.1673088223,-0.1012982503,-0.1600542814,-0.2908314466,0.1489149928,0.0585049614,0.1058554426,-0.2569085658,0.500236094,0.1337188333,-0.1236687973,0.259260118,0.2547582388,0.3921833038,0.2354182303,-0.0563612543,0.118781276,-0.1520817727,-0.1580698192,-0.1649380028,0.22200571,0.2826267481,-0.2399268299,0.40833956,0.220150277,-0.1681924611,0.3380620778,-0.0190949738,0.4372432232,-0.2133324146,0.3192174733,-0.3780584931,0.0575311445,-0.3101490736,-0.0832276717,-0.5760593414,-0.2530725598,-0.1635162681,-0.101671882,0.1617380083,-0.0138189169,0.0595093146,0.0480281673,0.3779025376,0.4761712253,0.1966776699,-0.2976574898,-0.2617687881,-0.5369645953,0.0845123827,-0.0744423419,0.0407859534,-0.0126415752,0.3252759278,-0.143486008,0.1179323345,0.1704975367,0.175069809,-0.2017744631,0.1189973131,-0.2758839428,-0.416370213,-0.0354897715,0.1237288564,-0.0925103053,-0.1620142162,0.1815119684,-0.2884228528,-0.0075483653,0.0063050175,0.1264898777,0.0303840991,-0.0279789791,0.4688521624,0.1719368994,0.4205750823,0.1091431305,-0.2500298619,-0.3299544752,-0.1993986368,-0.0289107151,0.0117749656,-0.0593902618,0.456469506,-0.0293870457,0.0285883415,-0.2617828548,0.1135850996,-0.0730579421,-0.1677081287,0.0184928607,-0.1089656875,0.0220410693,0.140310213,-0.00121685,0.3501479924,0.0503059253,0.2221424431,-0.0946951881,-0.1574622244,0.5106208324,-0.0712286085,-0.2704927921,0.2154466212,0.0052094804,-0.0617787167,-0.0020324003,-0.3609396815,0.2056361735,0.4209412038,0.079865478,-0.1565104872,0.0797880292,-0.1106195226,-0.0691273436,-0.1145044044,0.2245735824,-0.104605332,-0.1141530871,0.05216261,0.0729801282]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2212","title":"Can't reach \"https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\" when trying to load fquad dataset","comments":"I've contacted Martin (first author of the fquad paper) regarding a possible new url. Hopefully we can get one soon !","body":"I'm trying to load the [fquad dataset](https:\/\/huggingface.co\/datasets\/fquad) by running: \r\n\r\n```Python\r\nfquad = load_dataset(\"fquad\")\r\n```\r\n\r\nwhich produces the following error:\r\n\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset fquad\/default (download: 3.14 MiB, generated: 6.62 MiB, post-processed: Unknown size, total: 9.76 MiB) to \/root\/.cache\/huggingface\/datasets\/fquad\/default\/0.1.0\/778dc2c85813d05ddd0c17087294d5f8f24820752340958070876b677af9f061...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nConnectionError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 fquad = load_dataset(\"fquad\")\r\n\r\n11 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 617 \r\n 618 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\r\n```\r\n\r\nDoes anyone know why that is and how to fix it? ","comment_length":21,"text":"Can't reach \"https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\" when trying to load fquad dataset\nI'm trying to load the [fquad dataset](https:\/\/huggingface.co\/datasets\/fquad) by running: \r\n\r\n```Python\r\nfquad = load_dataset(\"fquad\")\r\n```\r\n\r\nwhich produces the following error:\r\n\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset fquad\/default (download: 3.14 MiB, generated: 6.62 MiB, post-processed: Unknown size, total: 9.76 MiB) to \/root\/.cache\/huggingface\/datasets\/fquad\/default\/0.1.0\/778dc2c85813d05ddd0c17087294d5f8f24820752340958070876b677af9f061...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nConnectionError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 fquad = load_dataset(\"fquad\")\r\n\r\n11 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 617 \r\n 618 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\r\n```\r\n\r\nDoes anyone know why that is and how to fix it? \nI've contacted Martin (first author of the fquad paper) regarding a possible new url. Hopefully we can get one soon !","embeddings":[-0.3520766795,0.1850519031,-0.1066456437,0.2455794364,0.3998158276,0.03360755,0.3876356781,0.2061039656,0.3057492673,0.1453926563,-0.2683598399,-0.1908080578,0.2827515602,0.0175514612,0.027606599,0.0714482665,-0.0717816278,-0.0440939777,-0.1596578062,0.0708816275,-0.3542164862,0.3704529405,-0.1904580444,0.1628842503,-0.3339352906,0.0633746386,0.0055798003,0.3590935767,-0.3199560642,-0.1079104543,0.286209017,0.0416278467,0.004151077,0.4852188826,-0.0001157215,0.1660642475,0.3586290479,-0.0317648053,-0.4298180938,-0.5632392168,-0.3045651317,-0.072570838,0.2153185606,-0.2087870389,-0.1616279483,-0.0290350895,0.1196179837,-0.4198773503,0.2603788972,0.4918957353,0.1842911541,-0.0472500958,0.2975516617,-0.1803262234,-0.0591369905,-0.1362888217,-0.0200078152,0.6450526714,-0.0140431896,-0.2045283467,-0.0508274771,0.0286189858,0.1182557642,0.1726587862,0.1585940719,0.2011903822,-0.3372362256,-0.1506722569,0.2118362486,0.0146889491,0.5337827802,-0.2849702537,-0.4723348618,-0.1233135611,-0.0491205491,-0.3767210245,0.5024482012,0.0925156698,-0.2628643513,0.1907122433,-0.4664973021,-0.2508746386,-0.1997162253,0.3678171635,-0.3860652447,0.2701286077,-0.1864978373,0.2325517088,0.2770298421,-0.1557939649,-0.2593966126,0.1098564789,0.1527526826,0.3186956048,-0.314109236,0.1022981256,-0.0440244749,-0.2830099761,0.1964606792,0.3815691173,-0.015078485,0.0076378272,-0.0969404653,0.164527759,0.4020113349,0.3189195991,-0.1120543703,0.2087809891,0.1873425245,0.6325322986,-0.1311301142,-0.1770190001,-0.1378113925,-0.081372045,-0.1658248156,-0.183514595,0.1537760198,-0.1985757202,-0.1576269418,0.1740346998,-0.2624880672,-0.0844819471,0.0354961865,0.3024871647,-0.0897951052,0.1091868505,0.0184984915,0.2455940992,-0.1006541178,0.1638454646,-0.1644030064,0.2479628474,-0.218215555,0.231559068,0.2416772842,0.1246844679,0.3092523515,-0.3974604607,0.0709448978,-0.0954510868,0.2211935818,-0.2454142272,-0.156402275,0.3248185813,0.1361620575,-0.019331608,0.1892515123,-0.0819873288,-0.120350033,-0.0591272786,-0.238079235,-0.2690150142,-0.1610825509,0.1693099141,-0.1822194904,0.0440686196,-0.3180564344,0.0211405642,-0.163014695,-0.2929230332,-0.02649194,0.1383045167,-0.1317382902,-0.1248099804,0.3689207137,0.5510037541,-0.2626015544,-0.0661597773,-0.2078201175,-0.0995471701,-0.0146748368,0.4244317114,-0.174391821,0.2035565674,-0.4347251058,0.2382106483,0.4783934355,-0.2472981364,-0.8428885341,0.427974999,-0.2450462729,0.1419134438,-0.091932185,0.0322814882,-0.0358391963,0.1951994896,0.6311691403,0.2926326096,-0.0661307499,-0.127931118,-0.0659748837,-0.3087247014,-0.0573887713,0.2728063464,-0.0072478638,0.173291862,0.2028902769,-0.11355827,0.203204602,0.1464847624,-0.1294753551,0.2984036505,0.0654904172,0.0501435176,-0.1333141923,-0.0529266,-0.6362524629,0.2412666529,-0.2172224224,-0.0310093351,-0.5615538359,0.2495313436,-0.4475069642,0.0381114259,-0.0719014481,0.0171304978,0.0995740071,0.0793238357,0.1675628722,0.1997745037,-0.2326557487,0.3575578332,0.0593538806,0.2285772711,-0.4177794456,0.3038277328,-0.068603389,-0.0128734708,-0.0401704386,0.0375095345,0.0876060575,-0.1907199472,-0.1723519415,0.3521072268,-0.0981610119,0.1596617848,0.3979538083,0.4166022539,0.1592304558,-0.0419741049,0.0967004448,-0.1133250967,0.0562817417,-0.0873886049,-0.1806227118,0.1824068427,0.1226353049,0.3023006022,0.140295893,0.033637926,0.365693301,0.1180057302,-0.0833135769,-0.1543994844,0.15414083,0.4664791822,0.2170068771,0.0557061583,-0.1387757808,-0.0494213514,0.3819881976,-0.2163211554,0.0133876475,0.2899906933,0.0413509086,-0.2723265886,0.1269237697,0.0862024873,0.4915738106,0.0586339347,0.1429887861,0.1387392431,-0.2469654232,-0.1638064682,0.0784770474,0.0087313019,0.0382009223,0.0595250651,0.1666518003,-0.0206933133,-0.1088888422,-0.3694197536,0.0472553745,0.1783639491,-0.1222572029,0.3422809839,-0.1195476055,-0.1523807496,0.1046303585,-0.3231037259,-0.2455987036,-0.4834875762,-0.1251465082,0.2941145301,0.0920701548,0.0488251857,-0.3312270641,0.0984709337,0.1856976151,-0.5195731521,-0.1338463426,-0.1513882279,-0.1231954619,0.0563571267,0.3074066341,-0.2006785274,0.2499470264,-0.2296645343,-0.100894548,-0.4725251794,-0.1700989753,0.0898433775,-0.0312449019,0.40792799,0.2447232157,0.3579797447,-0.0674479678,-0.0232894085,0.4561803341,0.0728968084,-0.0803323686,-0.0389703028,-0.1058985814,0.1410915405,0.054669302,-0.263803184,-0.2067806423,-0.3587554097,0.4501703084,0.0336139351,0.0696525946,0.1306530684,0.1599377245,0.2007167935,0.1608746797,-0.1496057361,-0.3580869138,-0.5675964355,0.4216380119,-0.2252491564,-0.062960811,0.0954560116,0.1260264516,0.4072681069,0.3108044267,-0.5964596272,-0.4336869121,-0.2441263199,0.2438689023,-0.0759851113,-0.2841826975,0.2059355229,-0.2813442647,-0.0455122963,0.0795237049,-0.3435887992,0.1072266921,0.0380031541,0.4218061566,0.1905358434,0.5441442132,-0.1006031409,0.596581161,0.0506761186,0.1118795276,0.3819426298,-0.1311579049,0.2393558174,-0.4368827343,-0.2892046869,0.1354883611,-0.0735363513,-0.3199809194,-0.1010950878,0.1610036194,-0.0667655841,-0.2432231307,0.2578684688,-0.4428741932,-0.2013871372,0.043550387,-0.0229954161,0.0392960124,0.1748021841,0.0631838068,-0.1840065271,-0.0513392054,-0.0495904051,0.1898236424,0.0814525932,0.1691627055,-0.0057752947,-0.3805061877,-0.5242769718,0.408682853,0.0688265115,0.3191053867,-0.2343497723,0.0901854932,0.159983471,-0.0714382604,0.5216056108,0.0438570678,-0.0242689401,0.2260752767,-0.0639998466,-0.2260826379,0.0408335701,0.1604545265,0.2876725197,0.2784393728,0.5854113698,-0.4355755746,-0.0223598778,0.2736801207,0.0492445305,-0.0857091174,-0.0519774668,-0.1963580996,-0.4926885366,-0.345927,-0.2309070528,0.0583768375,0.3225263953,0.2719903886,-0.1396180838,0.1985048801,-0.2339979112,-0.123792477,0.0856428072,0.2614714205,-0.1848341525,-0.0721978843,0.446174711,0.0551759638,0.0625965148,0.9526379108,0.2309765965,-0.1918456703,0.3530420959,0.2041083425,-0.0470772423,0.2335586697,-0.0544416867,-0.2276651412,0.0475472696,0.3281215131,-0.0701255649,0.3138429821,0.3014744818,-0.0341497324,-0.2183242589,-0.3774331212,0.5831079483,-0.0610893443,0.0716049373,0.5631733537,-0.0031045412,-0.1457767785,-0.0282859821,-0.11133793,0.8322306871,-0.2192073613,0.1396558434,0.3237110972,-0.2258144915,0.2849935293,-0.3626310825,0.1815250069,-0.1902487278,0.055939123,-0.1027311981,-0.1027690098,0.0169316493,0.2292079628,0.1400343627,0.4701394737,-0.0622956306,-0.0413589627,0.2022134811,0.2214944363,-0.1957224309,-0.03502509,-0.4555773437,0.1509607136,-0.063500762,0.3048473299,0.0043505183,-0.0404348522,-0.2007136643,-0.1768927574,-0.3790899217,0.1039981544,-0.4450157583,0.3993927836,0.064871639,-0.1194170266,-0.3323346078,0.1331066787,-0.1718951911,0.0374773331,-0.377406925,0.2002423555,-0.250582844,-0.1914522499,0.0667510033,0.2090218812,0.1672664881,-0.0286223646,-0.1082126722,0.2219368368,-0.1580397338,-0.4935772419,-0.0861230642,-0.0270098094,0.1342190057,-0.1137048602,-0.1250256896,-0.3386538327,-0.204316318,-0.0527786091,0.0808490962,0.028858887,0.2523483336,-0.2069956809,0.0026370073,-0.2918313742,-0.1399932802,0.5058478117,-0.2566277981,-0.3374788463,0.4631577432,0.0977263302,-0.0907071531,-0.1474872082,-0.1863509417,-0.0134128286,-0.1456059068,0.0908238962,0.0119757038,0.1186038107,-0.2104348689,-0.0244327877,0.2548632026,-0.1683793068,0.046724651,-0.5118241906,-0.4071712792,0.2717998028,-0.2288030982,0.1913926899,0.0034555777,-0.0871694088,0.173056066,-0.2681245506,-0.2592228055,0.1471401006,-0.4422644079,-0.2400080115,0.2411803603,-0.1960831136,0.4321490824,0.0995835066,0.1983939856,-0.0372604281,-0.2293649018,-0.1816976815,0.0783660114,0.1327116042,0.1031492278,-0.4042631388,0.0659462512,-0.0905284211,-0.1013431326,-0.1443170309,0.0145540675,0.2522515655,0.1482552588,-0.2715089619,0.1424651891,0.0433753803,-0.1513764858,0.0312853009,0.1540951729,0.3858402669,0.1071474254,0.0385405049,-0.012002415,0.0421502218,-0.1947081685,0.0288459379,0.0366724245,0.1114400104,0.5246019363,-0.2630036473,0.0688438863,0.1733348519,0.2766209543,0.3434119523,-0.393322587,-0.0504622869,0.0861113518,0.1886861473,-0.1005256698,-0.0409392379,0.1158134937,-0.0200303271,-0.0530431531,0.0752635002,0.0275851823,0.0262580942,0.1938834339,0.0303886868,0.4034167826,0.0132632684,0.0182413943,-0.0686375201,-0.1445128173,0.0048207832,0.3360909522,0.1343934685,0.3272197843,0.4190325737,-0.1450155079,-0.0987357721,-0.3855920732,0.2788904011,-0.2114679515,-0.4335516095,-0.2508144975,0.2423272133,-0.3047325015,0.2247778773,-0.2435577214,0.2378086001,-0.4140954912,0.2134464532,-0.2503335774,0.0139560569,-0.0381372944,-0.1202731133,0.0460858718,-0.1928994507,-0.1554569602,0.0292251222,-0.0773304701,-0.1959445477,0.1673088223,-0.1012982503,-0.1600542814,-0.2908314466,0.1489149928,0.0585049614,0.1058554426,-0.2569085658,0.500236094,0.1337188333,-0.1236687973,0.259260118,0.2547582388,0.3921833038,0.2354182303,-0.0563612543,0.118781276,-0.1520817727,-0.1580698192,-0.1649380028,0.22200571,0.2826267481,-0.2399268299,0.40833956,0.220150277,-0.1681924611,0.3380620778,-0.0190949738,0.4372432232,-0.2133324146,0.3192174733,-0.3780584931,0.0575311445,-0.3101490736,-0.0832276717,-0.5760593414,-0.2530725598,-0.1635162681,-0.101671882,0.1617380083,-0.0138189169,0.0595093146,0.0480281673,0.3779025376,0.4761712253,0.1966776699,-0.2976574898,-0.2617687881,-0.5369645953,0.0845123827,-0.0744423419,0.0407859534,-0.0126415752,0.3252759278,-0.143486008,0.1179323345,0.1704975367,0.175069809,-0.2017744631,0.1189973131,-0.2758839428,-0.416370213,-0.0354897715,0.1237288564,-0.0925103053,-0.1620142162,0.1815119684,-0.2884228528,-0.0075483653,0.0063050175,0.1264898777,0.0303840991,-0.0279789791,0.4688521624,0.1719368994,0.4205750823,0.1091431305,-0.2500298619,-0.3299544752,-0.1993986368,-0.0289107151,0.0117749656,-0.0593902618,0.456469506,-0.0293870457,0.0285883415,-0.2617828548,0.1135850996,-0.0730579421,-0.1677081287,0.0184928607,-0.1089656875,0.0220410693,0.140310213,-0.00121685,0.3501479924,0.0503059253,0.2221424431,-0.0946951881,-0.1574622244,0.5106208324,-0.0712286085,-0.2704927921,0.2154466212,0.0052094804,-0.0617787167,-0.0020324003,-0.3609396815,0.2056361735,0.4209412038,0.079865478,-0.1565104872,0.0797880292,-0.1106195226,-0.0691273436,-0.1145044044,0.2245735824,-0.104605332,-0.1141530871,0.05216261,0.0729801282]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2212","title":"Can't reach \"https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\" when trying to load fquad dataset","comments":"They now made a website to force people who want to use the dataset for commercial purposes to seek a commercial license from them ...","body":"I'm trying to load the [fquad dataset](https:\/\/huggingface.co\/datasets\/fquad) by running: \r\n\r\n```Python\r\nfquad = load_dataset(\"fquad\")\r\n```\r\n\r\nwhich produces the following error:\r\n\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset fquad\/default (download: 3.14 MiB, generated: 6.62 MiB, post-processed: Unknown size, total: 9.76 MiB) to \/root\/.cache\/huggingface\/datasets\/fquad\/default\/0.1.0\/778dc2c85813d05ddd0c17087294d5f8f24820752340958070876b677af9f061...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nConnectionError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 fquad = load_dataset(\"fquad\")\r\n\r\n11 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 617 \r\n 618 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\r\n```\r\n\r\nDoes anyone know why that is and how to fix it? ","comment_length":25,"text":"Can't reach \"https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\" when trying to load fquad dataset\nI'm trying to load the [fquad dataset](https:\/\/huggingface.co\/datasets\/fquad) by running: \r\n\r\n```Python\r\nfquad = load_dataset(\"fquad\")\r\n```\r\n\r\nwhich produces the following error:\r\n\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset fquad\/default (download: 3.14 MiB, generated: 6.62 MiB, post-processed: Unknown size, total: 9.76 MiB) to \/root\/.cache\/huggingface\/datasets\/fquad\/default\/0.1.0\/778dc2c85813d05ddd0c17087294d5f8f24820752340958070876b677af9f061...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nConnectionError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 fquad = load_dataset(\"fquad\")\r\n\r\n11 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n--> 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 617 \r\n 618 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/illuin\/fquad\/train.json.zip\r\n```\r\n\r\nDoes anyone know why that is and how to fix it? \nThey now made a website to force people who want to use the dataset for commercial purposes to seek a commercial license from them ...","embeddings":[-0.3520766795,0.1850519031,-0.1066456437,0.2455794364,0.3998158276,0.03360755,0.3876356781,0.2061039656,0.3057492673,0.1453926563,-0.2683598399,-0.1908080578,0.2827515602,0.0175514612,0.027606599,0.0714482665,-0.0717816278,-0.0440939777,-0.1596578062,0.0708816275,-0.3542164862,0.3704529405,-0.1904580444,0.1628842503,-0.3339352906,0.0633746386,0.0055798003,0.3590935767,-0.3199560642,-0.1079104543,0.286209017,0.0416278467,0.004151077,0.4852188826,-0.0001157215,0.1660642475,0.3586290479,-0.0317648053,-0.4298180938,-0.5632392168,-0.3045651317,-0.072570838,0.2153185606,-0.2087870389,-0.1616279483,-0.0290350895,0.1196179837,-0.4198773503,0.2603788972,0.4918957353,0.1842911541,-0.0472500958,0.2975516617,-0.1803262234,-0.0591369905,-0.1362888217,-0.0200078152,0.6450526714,-0.0140431896,-0.2045283467,-0.0508274771,0.0286189858,0.1182557642,0.1726587862,0.1585940719,0.2011903822,-0.3372362256,-0.1506722569,0.2118362486,0.0146889491,0.5337827802,-0.2849702537,-0.4723348618,-0.1233135611,-0.0491205491,-0.3767210245,0.5024482012,0.0925156698,-0.2628643513,0.1907122433,-0.4664973021,-0.2508746386,-0.1997162253,0.3678171635,-0.3860652447,0.2701286077,-0.1864978373,0.2325517088,0.2770298421,-0.1557939649,-0.2593966126,0.1098564789,0.1527526826,0.3186956048,-0.314109236,0.1022981256,-0.0440244749,-0.2830099761,0.1964606792,0.3815691173,-0.015078485,0.0076378272,-0.0969404653,0.164527759,0.4020113349,0.3189195991,-0.1120543703,0.2087809891,0.1873425245,0.6325322986,-0.1311301142,-0.1770190001,-0.1378113925,-0.081372045,-0.1658248156,-0.183514595,0.1537760198,-0.1985757202,-0.1576269418,0.1740346998,-0.2624880672,-0.0844819471,0.0354961865,0.3024871647,-0.0897951052,0.1091868505,0.0184984915,0.2455940992,-0.1006541178,0.1638454646,-0.1644030064,0.2479628474,-0.218215555,0.231559068,0.2416772842,0.1246844679,0.3092523515,-0.3974604607,0.0709448978,-0.0954510868,0.2211935818,-0.2454142272,-0.156402275,0.3248185813,0.1361620575,-0.019331608,0.1892515123,-0.0819873288,-0.120350033,-0.0591272786,-0.238079235,-0.2690150142,-0.1610825509,0.1693099141,-0.1822194904,0.0440686196,-0.3180564344,0.0211405642,-0.163014695,-0.2929230332,-0.02649194,0.1383045167,-0.1317382902,-0.1248099804,0.3689207137,0.5510037541,-0.2626015544,-0.0661597773,-0.2078201175,-0.0995471701,-0.0146748368,0.4244317114,-0.174391821,0.2035565674,-0.4347251058,0.2382106483,0.4783934355,-0.2472981364,-0.8428885341,0.427974999,-0.2450462729,0.1419134438,-0.091932185,0.0322814882,-0.0358391963,0.1951994896,0.6311691403,0.2926326096,-0.0661307499,-0.127931118,-0.0659748837,-0.3087247014,-0.0573887713,0.2728063464,-0.0072478638,0.173291862,0.2028902769,-0.11355827,0.203204602,0.1464847624,-0.1294753551,0.2984036505,0.0654904172,0.0501435176,-0.1333141923,-0.0529266,-0.6362524629,0.2412666529,-0.2172224224,-0.0310093351,-0.5615538359,0.2495313436,-0.4475069642,0.0381114259,-0.0719014481,0.0171304978,0.0995740071,0.0793238357,0.1675628722,0.1997745037,-0.2326557487,0.3575578332,0.0593538806,0.2285772711,-0.4177794456,0.3038277328,-0.068603389,-0.0128734708,-0.0401704386,0.0375095345,0.0876060575,-0.1907199472,-0.1723519415,0.3521072268,-0.0981610119,0.1596617848,0.3979538083,0.4166022539,0.1592304558,-0.0419741049,0.0967004448,-0.1133250967,0.0562817417,-0.0873886049,-0.1806227118,0.1824068427,0.1226353049,0.3023006022,0.140295893,0.033637926,0.365693301,0.1180057302,-0.0833135769,-0.1543994844,0.15414083,0.4664791822,0.2170068771,0.0557061583,-0.1387757808,-0.0494213514,0.3819881976,-0.2163211554,0.0133876475,0.2899906933,0.0413509086,-0.2723265886,0.1269237697,0.0862024873,0.4915738106,0.0586339347,0.1429887861,0.1387392431,-0.2469654232,-0.1638064682,0.0784770474,0.0087313019,0.0382009223,0.0595250651,0.1666518003,-0.0206933133,-0.1088888422,-0.3694197536,0.0472553745,0.1783639491,-0.1222572029,0.3422809839,-0.1195476055,-0.1523807496,0.1046303585,-0.3231037259,-0.2455987036,-0.4834875762,-0.1251465082,0.2941145301,0.0920701548,0.0488251857,-0.3312270641,0.0984709337,0.1856976151,-0.5195731521,-0.1338463426,-0.1513882279,-0.1231954619,0.0563571267,0.3074066341,-0.2006785274,0.2499470264,-0.2296645343,-0.100894548,-0.4725251794,-0.1700989753,0.0898433775,-0.0312449019,0.40792799,0.2447232157,0.3579797447,-0.0674479678,-0.0232894085,0.4561803341,0.0728968084,-0.0803323686,-0.0389703028,-0.1058985814,0.1410915405,0.054669302,-0.263803184,-0.2067806423,-0.3587554097,0.4501703084,0.0336139351,0.0696525946,0.1306530684,0.1599377245,0.2007167935,0.1608746797,-0.1496057361,-0.3580869138,-0.5675964355,0.4216380119,-0.2252491564,-0.062960811,0.0954560116,0.1260264516,0.4072681069,0.3108044267,-0.5964596272,-0.4336869121,-0.2441263199,0.2438689023,-0.0759851113,-0.2841826975,0.2059355229,-0.2813442647,-0.0455122963,0.0795237049,-0.3435887992,0.1072266921,0.0380031541,0.4218061566,0.1905358434,0.5441442132,-0.1006031409,0.596581161,0.0506761186,0.1118795276,0.3819426298,-0.1311579049,0.2393558174,-0.4368827343,-0.2892046869,0.1354883611,-0.0735363513,-0.3199809194,-0.1010950878,0.1610036194,-0.0667655841,-0.2432231307,0.2578684688,-0.4428741932,-0.2013871372,0.043550387,-0.0229954161,0.0392960124,0.1748021841,0.0631838068,-0.1840065271,-0.0513392054,-0.0495904051,0.1898236424,0.0814525932,0.1691627055,-0.0057752947,-0.3805061877,-0.5242769718,0.408682853,0.0688265115,0.3191053867,-0.2343497723,0.0901854932,0.159983471,-0.0714382604,0.5216056108,0.0438570678,-0.0242689401,0.2260752767,-0.0639998466,-0.2260826379,0.0408335701,0.1604545265,0.2876725197,0.2784393728,0.5854113698,-0.4355755746,-0.0223598778,0.2736801207,0.0492445305,-0.0857091174,-0.0519774668,-0.1963580996,-0.4926885366,-0.345927,-0.2309070528,0.0583768375,0.3225263953,0.2719903886,-0.1396180838,0.1985048801,-0.2339979112,-0.123792477,0.0856428072,0.2614714205,-0.1848341525,-0.0721978843,0.446174711,0.0551759638,0.0625965148,0.9526379108,0.2309765965,-0.1918456703,0.3530420959,0.2041083425,-0.0470772423,0.2335586697,-0.0544416867,-0.2276651412,0.0475472696,0.3281215131,-0.0701255649,0.3138429821,0.3014744818,-0.0341497324,-0.2183242589,-0.3774331212,0.5831079483,-0.0610893443,0.0716049373,0.5631733537,-0.0031045412,-0.1457767785,-0.0282859821,-0.11133793,0.8322306871,-0.2192073613,0.1396558434,0.3237110972,-0.2258144915,0.2849935293,-0.3626310825,0.1815250069,-0.1902487278,0.055939123,-0.1027311981,-0.1027690098,0.0169316493,0.2292079628,0.1400343627,0.4701394737,-0.0622956306,-0.0413589627,0.2022134811,0.2214944363,-0.1957224309,-0.03502509,-0.4555773437,0.1509607136,-0.063500762,0.3048473299,0.0043505183,-0.0404348522,-0.2007136643,-0.1768927574,-0.3790899217,0.1039981544,-0.4450157583,0.3993927836,0.064871639,-0.1194170266,-0.3323346078,0.1331066787,-0.1718951911,0.0374773331,-0.377406925,0.2002423555,-0.250582844,-0.1914522499,0.0667510033,0.2090218812,0.1672664881,-0.0286223646,-0.1082126722,0.2219368368,-0.1580397338,-0.4935772419,-0.0861230642,-0.0270098094,0.1342190057,-0.1137048602,-0.1250256896,-0.3386538327,-0.204316318,-0.0527786091,0.0808490962,0.028858887,0.2523483336,-0.2069956809,0.0026370073,-0.2918313742,-0.1399932802,0.5058478117,-0.2566277981,-0.3374788463,0.4631577432,0.0977263302,-0.0907071531,-0.1474872082,-0.1863509417,-0.0134128286,-0.1456059068,0.0908238962,0.0119757038,0.1186038107,-0.2104348689,-0.0244327877,0.2548632026,-0.1683793068,0.046724651,-0.5118241906,-0.4071712792,0.2717998028,-0.2288030982,0.1913926899,0.0034555777,-0.0871694088,0.173056066,-0.2681245506,-0.2592228055,0.1471401006,-0.4422644079,-0.2400080115,0.2411803603,-0.1960831136,0.4321490824,0.0995835066,0.1983939856,-0.0372604281,-0.2293649018,-0.1816976815,0.0783660114,0.1327116042,0.1031492278,-0.4042631388,0.0659462512,-0.0905284211,-0.1013431326,-0.1443170309,0.0145540675,0.2522515655,0.1482552588,-0.2715089619,0.1424651891,0.0433753803,-0.1513764858,0.0312853009,0.1540951729,0.3858402669,0.1071474254,0.0385405049,-0.012002415,0.0421502218,-0.1947081685,0.0288459379,0.0366724245,0.1114400104,0.5246019363,-0.2630036473,0.0688438863,0.1733348519,0.2766209543,0.3434119523,-0.393322587,-0.0504622869,0.0861113518,0.1886861473,-0.1005256698,-0.0409392379,0.1158134937,-0.0200303271,-0.0530431531,0.0752635002,0.0275851823,0.0262580942,0.1938834339,0.0303886868,0.4034167826,0.0132632684,0.0182413943,-0.0686375201,-0.1445128173,0.0048207832,0.3360909522,0.1343934685,0.3272197843,0.4190325737,-0.1450155079,-0.0987357721,-0.3855920732,0.2788904011,-0.2114679515,-0.4335516095,-0.2508144975,0.2423272133,-0.3047325015,0.2247778773,-0.2435577214,0.2378086001,-0.4140954912,0.2134464532,-0.2503335774,0.0139560569,-0.0381372944,-0.1202731133,0.0460858718,-0.1928994507,-0.1554569602,0.0292251222,-0.0773304701,-0.1959445477,0.1673088223,-0.1012982503,-0.1600542814,-0.2908314466,0.1489149928,0.0585049614,0.1058554426,-0.2569085658,0.500236094,0.1337188333,-0.1236687973,0.259260118,0.2547582388,0.3921833038,0.2354182303,-0.0563612543,0.118781276,-0.1520817727,-0.1580698192,-0.1649380028,0.22200571,0.2826267481,-0.2399268299,0.40833956,0.220150277,-0.1681924611,0.3380620778,-0.0190949738,0.4372432232,-0.2133324146,0.3192174733,-0.3780584931,0.0575311445,-0.3101490736,-0.0832276717,-0.5760593414,-0.2530725598,-0.1635162681,-0.101671882,0.1617380083,-0.0138189169,0.0595093146,0.0480281673,0.3779025376,0.4761712253,0.1966776699,-0.2976574898,-0.2617687881,-0.5369645953,0.0845123827,-0.0744423419,0.0407859534,-0.0126415752,0.3252759278,-0.143486008,0.1179323345,0.1704975367,0.175069809,-0.2017744631,0.1189973131,-0.2758839428,-0.416370213,-0.0354897715,0.1237288564,-0.0925103053,-0.1620142162,0.1815119684,-0.2884228528,-0.0075483653,0.0063050175,0.1264898777,0.0303840991,-0.0279789791,0.4688521624,0.1719368994,0.4205750823,0.1091431305,-0.2500298619,-0.3299544752,-0.1993986368,-0.0289107151,0.0117749656,-0.0593902618,0.456469506,-0.0293870457,0.0285883415,-0.2617828548,0.1135850996,-0.0730579421,-0.1677081287,0.0184928607,-0.1089656875,0.0220410693,0.140310213,-0.00121685,0.3501479924,0.0503059253,0.2221424431,-0.0946951881,-0.1574622244,0.5106208324,-0.0712286085,-0.2704927921,0.2154466212,0.0052094804,-0.0617787167,-0.0020324003,-0.3609396815,0.2056361735,0.4209412038,0.079865478,-0.1565104872,0.0797880292,-0.1106195226,-0.0691273436,-0.1145044044,0.2245735824,-0.104605332,-0.1141530871,0.05216261,0.0729801282]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2211","title":"Getting checksum error when trying to load lc_quad dataset","comments":"Hi,\r\n\r\nI've already opened a PR with the fix. If you are in a hurry, just build the project from source and run:\r\n```bash\r\ndatasets-cli test datasets\/lc_quad --save_infos --all_configs --ignore_verifications\r\n```\r\n\r\n","body":"I'm having issues loading the [lc_quad](https:\/\/huggingface.co\/datasets\/fquad) dataset by running:\r\n\r\n```Python\r\nlc_quad = load_dataset(\"lc_quad\")\r\n```\r\n\r\nwhich is giving me the following error:\r\n\r\n``` \r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset lc_quad\/default (download: 3.69 MiB, generated: 19.77 MiB, post-processed: Unknown size, total: 23.46 MiB) to \/root\/.cache\/huggingface\/datasets\/lc_quad\/default\/2.0.0\/5a98fe174603f5dec6df07edf1c2b4d2317210d2ad61f5a393839bca4d64e5a7...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 lc_quad = load_dataset(\"lc_quad\")\r\n\r\n3 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/AskNowQA\/LC-QuAD2.0\/archive\/master.zip']\r\n```\r\n\r\nDoes anyone know why this could be and how I fix it? ","comment_length":31,"text":"Getting checksum error when trying to load lc_quad dataset\nI'm having issues loading the [lc_quad](https:\/\/huggingface.co\/datasets\/fquad) dataset by running:\r\n\r\n```Python\r\nlc_quad = load_dataset(\"lc_quad\")\r\n```\r\n\r\nwhich is giving me the following error:\r\n\r\n``` \r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset lc_quad\/default (download: 3.69 MiB, generated: 19.77 MiB, post-processed: Unknown size, total: 23.46 MiB) to \/root\/.cache\/huggingface\/datasets\/lc_quad\/default\/2.0.0\/5a98fe174603f5dec6df07edf1c2b4d2317210d2ad61f5a393839bca4d64e5a7...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 lc_quad = load_dataset(\"lc_quad\")\r\n\r\n3 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/AskNowQA\/LC-QuAD2.0\/archive\/master.zip']\r\n```\r\n\r\nDoes anyone know why this could be and how I fix it? \nHi,\r\n\r\nI've already opened a PR with the fix. If you are in a hurry, just build the project from source and run:\r\n```bash\r\ndatasets-cli test datasets\/lc_quad --save_infos --all_configs --ignore_verifications\r\n```\r\n\r\n","embeddings":[-0.1586478055,0.0430786312,-0.0340241343,0.3563593924,0.2677668631,0.0147292903,0.0714035481,0.2469018549,0.3710919619,-0.0422729291,-0.1049166024,0.0677453876,-0.0377763398,0.0753146783,-0.2280279398,0.2225681841,0.0536352806,0.0427623615,-0.0752972588,0.0635507628,-0.2016302496,0.2370710671,-0.08427459,-0.2540154159,-0.2298655808,0.022695167,0.2053581625,0.4523075819,-0.284686029,-0.2342202663,0.2958003879,0.2990624905,0.177793175,0.3523058593,-0.0001155879,0.1017175913,0.3779894412,-0.189946264,-0.3943250775,-0.0775870606,-0.2616369724,-0.1055975109,0.1051906645,-0.1757149994,0.0675687641,0.3183248341,-0.1755350679,-0.1860815436,0.1201474518,0.0313655734,0.2525502741,0.6949155927,0.0769914463,0.0821324289,-0.0173101816,-0.0665880516,0.0181370638,0.6490388513,0.2040345371,-0.1024491787,-0.0439264029,0.1912636757,-0.0535609722,0.1119764149,0.171669215,-0.1965045333,-0.0742644519,-0.0029104268,0.2049059272,0.2949047983,0.4116024971,-0.354295969,-0.4142789841,-0.0578518435,0.0658886731,-0.19241032,0.3854560852,0.2196364552,-0.202465862,-0.1056818217,-0.3435885608,0.0057747285,0.1171856374,0.143759042,0.02749338,0.0133236069,0.011992489,0.0433482118,0.1741916984,-0.1444574744,-0.1835757196,-0.0199342594,-0.2033450156,0.2262380272,-0.4441525638,-0.0280742981,-0.0538369827,0.5078675747,0.161497131,0.5987932682,0.1114636809,0.2344520241,-0.1141757444,0.131139487,0.0641107485,0.1764589846,0.0813860819,0.2512382567,0.1025379524,0.5696989894,-0.112466,-0.0161850434,-0.1434008032,-0.067008093,0.020459095,0.0047425097,0.1683295965,-0.5566789508,-0.2990564108,0.4156864285,0.1240056381,-0.0489496663,0.4083915353,0.4536544979,-0.1148815006,0.0788542703,0.0255795624,0.0592117757,-0.2122116536,0.0015811836,-0.2094918638,0.095049344,-0.109906666,0.1747077703,0.3324213028,-0.2124649584,0.3945674002,-0.1254246235,0.554668963,-0.0824102387,0.0168444403,-0.0795143917,-0.2176980674,0.3631320894,-0.1268157661,0.0099507356,0.4028933644,-0.0756156445,-0.2388128042,-0.0873798952,-0.1864018887,-0.1896510571,-0.2478010505,0.1476690024,-0.4457603991,-0.0073195416,-0.1004635543,-0.431081146,0.3045098782,-0.3342050016,0.0233886093,-0.1273536235,-0.1298718005,-0.1921387166,0.2523182333,0.3476815522,-0.2570971847,-0.042105291,0.005207208,-0.2823810875,0.191449672,0.2836829424,-0.0853362828,0.018048035,-0.2128898054,-0.1154889539,0.0440893285,-0.3135988712,-0.6612861156,0.1054850817,0.2083634287,0.2534271777,-0.1629279107,0.0760661811,-0.1316621155,-0.0755772591,0.0276353955,-0.0193063058,0.0347143225,0.0147750927,-0.2605696619,-0.1714076847,0.1294886768,0.2703022659,-0.0975577682,-0.0796645805,0.0913568959,0.0269035045,0.195572868,-0.0815563351,-0.1271460056,0.1410221159,0.4066483974,0.0088428976,0.0124131655,-0.1226483509,-0.5203519464,0.4184086621,-0.1758986264,0.1054993346,-0.3010833561,0.0123756966,-0.3993290365,-0.0077080829,-0.2525680065,0.0424720533,0.0912322775,0.1547874808,0.1401317716,-0.018431779,-0.2219920009,0.1539941281,-0.402622968,0.1927557141,-0.4083478153,0.4389024079,0.1484152824,-0.0811729208,0.0649157166,0.2987393141,0.0440049544,0.0688854307,-0.1830139458,0.3952317238,0.3292908669,-0.0046892441,0.1685605198,0.2685255408,-0.0258053299,-0.1721272618,-0.0555889159,0.2175353318,0.1912800372,-0.0338720679,-0.0383775122,0.4533189535,-0.1471694559,0.0875585973,-0.0493916832,0.0198792871,0.2120854259,-0.037548665,-0.2479397207,-0.2896395624,0.3542730808,0.3663347065,0.2394928038,0.0791928396,-0.0366405286,-0.2723244131,-0.0387272686,0.0369359143,-0.0535523482,0.0033647083,0.2176830769,0.0531175286,0.1653119475,0.3118388951,0.3474906087,-0.0092862491,-0.0535392649,0.2015189081,-0.3057135344,-0.0120725799,-0.0608281977,0.0010351073,0.089135319,0.2957101464,0.0235628765,-0.0257526413,-0.3373190165,-0.1105120778,0.0253474694,0.3205064833,-0.3456208706,0.0741355494,-0.3163037002,-0.1975536197,-0.1722880751,-0.24867028,-0.4937612116,-0.4542938173,-0.1639761925,0.2879142463,0.0259698164,0.1399546862,-0.5169958472,0.1362466961,0.0900542811,-0.512042582,0.0927218571,-0.1727410704,-0.0236470792,0.0141025204,0.5645958185,0.020563338,0.4354997575,-0.1627611667,-0.027611088,-0.3428322673,-0.1681953967,-0.0328796282,-0.1144415289,0.2417946756,0.2336486727,0.2427727729,-0.1269131154,-0.21963498,0.3777884245,0.1318973303,-0.2311527729,0.1286144406,-0.1610604674,-0.1313485354,-0.0604946129,0.2755798995,-0.1098041609,-0.3019474447,-0.0892819986,0.175815016,0.092692107,-0.0439409241,-0.0829568282,0.1223537475,-0.0465585142,0.0810393393,-0.3157858253,-0.6449924707,0.3277232647,-0.0472498387,-0.3249155879,0.0022676487,-0.0587228388,0.4990760386,0.0536289699,-0.4397414625,-0.3454911113,-0.177422449,0.2491053939,0.0834853798,-0.0003843263,0.0818123147,-0.1677067727,0.0547439717,-0.0944872871,-0.252823323,-0.1147165969,-0.0507771038,0.5432264805,-0.1026443392,0.2152421921,-0.1030042171,0.6116770506,0.4520657063,-0.0781401247,0.3794070482,0.0450162813,0.4519749284,-0.1160534918,-0.335739255,0.0971652195,-0.1612164825,-0.0847993344,0.0252150185,0.0292716045,0.024734132,-0.2515132129,0.151915893,-0.2503833771,-0.2537184358,0.0410110839,-0.1620973051,0.1288381964,-0.0400340371,0.0869282484,-0.0380054191,-0.3001151383,0.0917965695,0.4121662974,-0.10287752,-0.1234615371,-0.4588961005,-0.2496635169,-0.2685909271,0.3270353079,0.2161614746,0.4495323598,-0.0150478901,-0.2056308389,-0.0443210639,-0.1235325411,0.148912549,0.0390379429,0.0833090246,0.2485013604,-0.1471063793,-0.3670141697,-0.1597008407,0.0344158411,0.1818722337,0.179887116,0.3147481084,-0.4724255502,0.06753131,-0.0433376431,0.2448161542,-0.2145601064,-0.3109270334,-0.524887085,-0.2773876786,-0.5080869794,-0.0672974661,-0.2400072515,0.4325858951,0.107441999,0.3124433458,0.2348861098,-0.1548345834,0.3811735213,0.2378400713,0.3265141249,0.1357013583,0.1744468063,0.0411280133,-0.0249344725,0.0622662269,0.9562077522,-0.2117443532,-0.210804984,-0.035291288,0.1248347238,-0.0898462981,0.1483587772,-0.0194941238,-0.0160934944,0.4172479212,-0.0226287469,-0.2142560035,0.0240513384,-0.1016139016,-0.0430673063,-0.3212284744,-0.1270317286,0.1637931764,-0.2481814027,0.1020817682,0.5275672674,0.1273375899,-0.0017748867,0.0106437057,0.1958931088,0.7999645472,-0.067967616,0.048691377,0.4638361633,-0.0942289159,0.1278882027,0.267272979,-0.16252698,-0.3938178122,-0.2919070721,-0.0014132104,-0.1252771169,0.1884137541,-0.157389611,0.0349529311,0.2362201661,-0.1836348176,0.0361152813,0.0520516597,0.2275846303,-0.2709960639,-0.2156780213,-0.3438726068,0.1822007,-0.1196344048,0.2140209526,-0.2056079358,0.1336970627,-0.084583126,-0.2432812154,-0.4584517181,0.08267048,-0.1393226534,0.2479300946,0.5432108641,-0.01165043,-0.1105084345,0.0510684736,0.13976641,0.1708892137,-0.3202243745,0.1759998649,0.2094753236,-0.026594596,0.0044733309,-0.0108374041,0.3899156749,-0.1165590361,-0.1128100604,0.0313871019,0.0539408252,-0.4626779258,0.158140406,-0.1413083076,-0.1339838803,-0.4125747085,-0.3186384141,-0.2078332901,0.084862031,-0.1035225838,0.1193909347,0.4159374833,-0.0739142373,0.0533740297,0.1596258134,-0.2478346229,-0.1074660197,0.4095676839,-0.2446775883,-0.3135176003,0.4543264508,0.072319366,-0.1784508973,-0.1679846197,-0.0204058178,-0.0449930541,-0.3731469214,0.0616806075,-0.0245231371,-0.0404421724,-0.1267807186,0.3055093288,0.4896071851,-0.0096833957,0.2847531438,-0.7144518495,-0.3556802571,0.2366777062,-0.0546412282,0.2952449024,-0.2526121438,0.1385347694,-0.2104969025,0.0593775921,-0.2816969454,0.0230787657,-0.4564669728,-0.0837881491,0.1001639888,-0.0127250785,0.261464417,0.0349851549,0.1742554307,0.0547779761,-0.3107987046,-0.1835141629,-0.2588793635,0.1122425124,0.1209283322,-0.280680269,0.1347668916,0.0320408158,0.0065012369,-0.1869190335,-0.0410554186,0.1059171185,-0.1052688062,0.2443249673,0.2662098706,-0.0443941802,-0.2006893754,0.1606378704,-0.0766754299,0.0761699304,0.2360352427,-0.0732889846,-0.1039145216,0.0062031471,0.096316725,0.1853497028,0.001875477,-0.1533675939,0.3341062367,-0.1458846629,0.0689527541,0.2138649076,0.3672226369,0.2346559465,-0.3093515038,0.0030957852,0.2768591344,0.1573718935,-0.2205290496,0.0155065032,0.2803596258,0.2246055603,0.0225315634,0.2279590815,0.0356610417,-0.2950276732,0.3212125897,-0.072273463,0.2576577365,-0.1907857805,0.1527574956,-0.0063177482,-0.0556527823,-0.0287510473,0.1048015133,-0.1620068848,-0.0459872484,-0.0354439802,-0.1378435045,0.1757641137,-0.2960186005,0.17698735,-0.1391239464,-0.6813548803,0.3512970805,0.145282194,0.0003174874,0.0817152932,-0.0337929539,0.3823237121,-0.244302094,0.1538819671,-0.4536537826,0.1196980551,-0.1372100264,-0.1535899937,-0.2521290481,-0.2382503301,0.1049422771,0.1202083305,-0.015179147,-0.104823634,0.1012709513,0.0977588519,-0.1928468496,-0.4656092525,-0.1910358667,0.1723300517,-0.0744981617,-0.1790537536,0.3651694059,0.250138998,0.0551218837,0.3380270898,0.1576621383,0.464658469,0.3421230018,0.1389377862,-0.0273248088,-0.1392035633,0.1697890311,-0.0492099263,0.6106901765,0.1171118468,0.0790215358,0.4796912074,0.1567838937,-0.1200560853,-0.1497459859,-0.2036273479,0.0458743684,-0.0536595434,0.3299311996,-0.164848581,0.285926789,-0.2320778072,0.283267647,-0.4769468606,-0.1319950521,-0.0052832728,0.1365718246,0.2171823233,0.086733833,0.0526953824,-0.1251878291,0.6108255386,0.479727447,0.2195465416,-0.2921740413,-0.3415738642,-0.8700946569,0.3062995076,0.0001155213,-0.2973013818,0.1456142217,0.1620195806,0.0235687755,0.2645165622,0.2131074369,-0.1798981577,-0.0352095179,-0.1901183277,-0.1012415364,-0.1103344336,-0.1084368825,-0.2594906092,-0.1512116194,-0.2713238895,0.0640031695,0.1597411036,0.0773380622,0.051146511,-0.3196358383,-0.1865685284,0.2459578961,0.0364952013,0.1474298835,0.7800116539,0.054971572,-0.1918807626,-0.1947782189,-0.1534125954,0.0571291186,0.1269258112,-0.1336616576,0.5662155151,0.1243901476,0.1683933437,-0.2983309031,0.7553335428,0.1155050993,-0.4614267051,-0.2976111472,0.1096067652,-0.0259118788,0.0050343927,0.0148074282,0.3162671626,0.0250104703,0.5059561729,0.0054063075,-0.1460597217,0.7090224624,-0.3143617511,-0.1443075538,0.0277506523,0.1365534812,-0.2922633588,-0.1482906938,-0.7186961174,0.123646237,0.1957412958,0.0238483585,-0.130585745,0.2036420852,-0.1852646768,0.1725731194,-0.0994895622,0.1660661101,-0.1630265862,-0.1523593068,0.0270941108,-0.0499184653]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2211","title":"Getting checksum error when trying to load lc_quad dataset","comments":"Ah sorry, I tried searching but couldn't find any related PR. \r\n\r\nThank you! ","body":"I'm having issues loading the [lc_quad](https:\/\/huggingface.co\/datasets\/fquad) dataset by running:\r\n\r\n```Python\r\nlc_quad = load_dataset(\"lc_quad\")\r\n```\r\n\r\nwhich is giving me the following error:\r\n\r\n``` \r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset lc_quad\/default (download: 3.69 MiB, generated: 19.77 MiB, post-processed: Unknown size, total: 23.46 MiB) to \/root\/.cache\/huggingface\/datasets\/lc_quad\/default\/2.0.0\/5a98fe174603f5dec6df07edf1c2b4d2317210d2ad61f5a393839bca4d64e5a7...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 lc_quad = load_dataset(\"lc_quad\")\r\n\r\n3 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/AskNowQA\/LC-QuAD2.0\/archive\/master.zip']\r\n```\r\n\r\nDoes anyone know why this could be and how I fix it? ","comment_length":13,"text":"Getting checksum error when trying to load lc_quad dataset\nI'm having issues loading the [lc_quad](https:\/\/huggingface.co\/datasets\/fquad) dataset by running:\r\n\r\n```Python\r\nlc_quad = load_dataset(\"lc_quad\")\r\n```\r\n\r\nwhich is giving me the following error:\r\n\r\n``` \r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset lc_quad\/default (download: 3.69 MiB, generated: 19.77 MiB, post-processed: Unknown size, total: 23.46 MiB) to \/root\/.cache\/huggingface\/datasets\/lc_quad\/default\/2.0.0\/5a98fe174603f5dec6df07edf1c2b4d2317210d2ad61f5a393839bca4d64e5a7...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n\r\n in ()\r\n----> 1 lc_quad = load_dataset(\"lc_quad\")\r\n\r\n3 frames\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/AskNowQA\/LC-QuAD2.0\/archive\/master.zip']\r\n```\r\n\r\nDoes anyone know why this could be and how I fix it? \nAh sorry, I tried searching but couldn't find any related PR. \r\n\r\nThank you! ","embeddings":[-0.1586478055,0.0430786312,-0.0340241343,0.3563593924,0.2677668631,0.0147292903,0.0714035481,0.2469018549,0.3710919619,-0.0422729291,-0.1049166024,0.0677453876,-0.0377763398,0.0753146783,-0.2280279398,0.2225681841,0.0536352806,0.0427623615,-0.0752972588,0.0635507628,-0.2016302496,0.2370710671,-0.08427459,-0.2540154159,-0.2298655808,0.022695167,0.2053581625,0.4523075819,-0.284686029,-0.2342202663,0.2958003879,0.2990624905,0.177793175,0.3523058593,-0.0001155879,0.1017175913,0.3779894412,-0.189946264,-0.3943250775,-0.0775870606,-0.2616369724,-0.1055975109,0.1051906645,-0.1757149994,0.0675687641,0.3183248341,-0.1755350679,-0.1860815436,0.1201474518,0.0313655734,0.2525502741,0.6949155927,0.0769914463,0.0821324289,-0.0173101816,-0.0665880516,0.0181370638,0.6490388513,0.2040345371,-0.1024491787,-0.0439264029,0.1912636757,-0.0535609722,0.1119764149,0.171669215,-0.1965045333,-0.0742644519,-0.0029104268,0.2049059272,0.2949047983,0.4116024971,-0.354295969,-0.4142789841,-0.0578518435,0.0658886731,-0.19241032,0.3854560852,0.2196364552,-0.202465862,-0.1056818217,-0.3435885608,0.0057747285,0.1171856374,0.143759042,0.02749338,0.0133236069,0.011992489,0.0433482118,0.1741916984,-0.1444574744,-0.1835757196,-0.0199342594,-0.2033450156,0.2262380272,-0.4441525638,-0.0280742981,-0.0538369827,0.5078675747,0.161497131,0.5987932682,0.1114636809,0.2344520241,-0.1141757444,0.131139487,0.0641107485,0.1764589846,0.0813860819,0.2512382567,0.1025379524,0.5696989894,-0.112466,-0.0161850434,-0.1434008032,-0.067008093,0.020459095,0.0047425097,0.1683295965,-0.5566789508,-0.2990564108,0.4156864285,0.1240056381,-0.0489496663,0.4083915353,0.4536544979,-0.1148815006,0.0788542703,0.0255795624,0.0592117757,-0.2122116536,0.0015811836,-0.2094918638,0.095049344,-0.109906666,0.1747077703,0.3324213028,-0.2124649584,0.3945674002,-0.1254246235,0.554668963,-0.0824102387,0.0168444403,-0.0795143917,-0.2176980674,0.3631320894,-0.1268157661,0.0099507356,0.4028933644,-0.0756156445,-0.2388128042,-0.0873798952,-0.1864018887,-0.1896510571,-0.2478010505,0.1476690024,-0.4457603991,-0.0073195416,-0.1004635543,-0.431081146,0.3045098782,-0.3342050016,0.0233886093,-0.1273536235,-0.1298718005,-0.1921387166,0.2523182333,0.3476815522,-0.2570971847,-0.042105291,0.005207208,-0.2823810875,0.191449672,0.2836829424,-0.0853362828,0.018048035,-0.2128898054,-0.1154889539,0.0440893285,-0.3135988712,-0.6612861156,0.1054850817,0.2083634287,0.2534271777,-0.1629279107,0.0760661811,-0.1316621155,-0.0755772591,0.0276353955,-0.0193063058,0.0347143225,0.0147750927,-0.2605696619,-0.1714076847,0.1294886768,0.2703022659,-0.0975577682,-0.0796645805,0.0913568959,0.0269035045,0.195572868,-0.0815563351,-0.1271460056,0.1410221159,0.4066483974,0.0088428976,0.0124131655,-0.1226483509,-0.5203519464,0.4184086621,-0.1758986264,0.1054993346,-0.3010833561,0.0123756966,-0.3993290365,-0.0077080829,-0.2525680065,0.0424720533,0.0912322775,0.1547874808,0.1401317716,-0.018431779,-0.2219920009,0.1539941281,-0.402622968,0.1927557141,-0.4083478153,0.4389024079,0.1484152824,-0.0811729208,0.0649157166,0.2987393141,0.0440049544,0.0688854307,-0.1830139458,0.3952317238,0.3292908669,-0.0046892441,0.1685605198,0.2685255408,-0.0258053299,-0.1721272618,-0.0555889159,0.2175353318,0.1912800372,-0.0338720679,-0.0383775122,0.4533189535,-0.1471694559,0.0875585973,-0.0493916832,0.0198792871,0.2120854259,-0.037548665,-0.2479397207,-0.2896395624,0.3542730808,0.3663347065,0.2394928038,0.0791928396,-0.0366405286,-0.2723244131,-0.0387272686,0.0369359143,-0.0535523482,0.0033647083,0.2176830769,0.0531175286,0.1653119475,0.3118388951,0.3474906087,-0.0092862491,-0.0535392649,0.2015189081,-0.3057135344,-0.0120725799,-0.0608281977,0.0010351073,0.089135319,0.2957101464,0.0235628765,-0.0257526413,-0.3373190165,-0.1105120778,0.0253474694,0.3205064833,-0.3456208706,0.0741355494,-0.3163037002,-0.1975536197,-0.1722880751,-0.24867028,-0.4937612116,-0.4542938173,-0.1639761925,0.2879142463,0.0259698164,0.1399546862,-0.5169958472,0.1362466961,0.0900542811,-0.512042582,0.0927218571,-0.1727410704,-0.0236470792,0.0141025204,0.5645958185,0.020563338,0.4354997575,-0.1627611667,-0.027611088,-0.3428322673,-0.1681953967,-0.0328796282,-0.1144415289,0.2417946756,0.2336486727,0.2427727729,-0.1269131154,-0.21963498,0.3777884245,0.1318973303,-0.2311527729,0.1286144406,-0.1610604674,-0.1313485354,-0.0604946129,0.2755798995,-0.1098041609,-0.3019474447,-0.0892819986,0.175815016,0.092692107,-0.0439409241,-0.0829568282,0.1223537475,-0.0465585142,0.0810393393,-0.3157858253,-0.6449924707,0.3277232647,-0.0472498387,-0.3249155879,0.0022676487,-0.0587228388,0.4990760386,0.0536289699,-0.4397414625,-0.3454911113,-0.177422449,0.2491053939,0.0834853798,-0.0003843263,0.0818123147,-0.1677067727,0.0547439717,-0.0944872871,-0.252823323,-0.1147165969,-0.0507771038,0.5432264805,-0.1026443392,0.2152421921,-0.1030042171,0.6116770506,0.4520657063,-0.0781401247,0.3794070482,0.0450162813,0.4519749284,-0.1160534918,-0.335739255,0.0971652195,-0.1612164825,-0.0847993344,0.0252150185,0.0292716045,0.024734132,-0.2515132129,0.151915893,-0.2503833771,-0.2537184358,0.0410110839,-0.1620973051,0.1288381964,-0.0400340371,0.0869282484,-0.0380054191,-0.3001151383,0.0917965695,0.4121662974,-0.10287752,-0.1234615371,-0.4588961005,-0.2496635169,-0.2685909271,0.3270353079,0.2161614746,0.4495323598,-0.0150478901,-0.2056308389,-0.0443210639,-0.1235325411,0.148912549,0.0390379429,0.0833090246,0.2485013604,-0.1471063793,-0.3670141697,-0.1597008407,0.0344158411,0.1818722337,0.179887116,0.3147481084,-0.4724255502,0.06753131,-0.0433376431,0.2448161542,-0.2145601064,-0.3109270334,-0.524887085,-0.2773876786,-0.5080869794,-0.0672974661,-0.2400072515,0.4325858951,0.107441999,0.3124433458,0.2348861098,-0.1548345834,0.3811735213,0.2378400713,0.3265141249,0.1357013583,0.1744468063,0.0411280133,-0.0249344725,0.0622662269,0.9562077522,-0.2117443532,-0.210804984,-0.035291288,0.1248347238,-0.0898462981,0.1483587772,-0.0194941238,-0.0160934944,0.4172479212,-0.0226287469,-0.2142560035,0.0240513384,-0.1016139016,-0.0430673063,-0.3212284744,-0.1270317286,0.1637931764,-0.2481814027,0.1020817682,0.5275672674,0.1273375899,-0.0017748867,0.0106437057,0.1958931088,0.7999645472,-0.067967616,0.048691377,0.4638361633,-0.0942289159,0.1278882027,0.267272979,-0.16252698,-0.3938178122,-0.2919070721,-0.0014132104,-0.1252771169,0.1884137541,-0.157389611,0.0349529311,0.2362201661,-0.1836348176,0.0361152813,0.0520516597,0.2275846303,-0.2709960639,-0.2156780213,-0.3438726068,0.1822007,-0.1196344048,0.2140209526,-0.2056079358,0.1336970627,-0.084583126,-0.2432812154,-0.4584517181,0.08267048,-0.1393226534,0.2479300946,0.5432108641,-0.01165043,-0.1105084345,0.0510684736,0.13976641,0.1708892137,-0.3202243745,0.1759998649,0.2094753236,-0.026594596,0.0044733309,-0.0108374041,0.3899156749,-0.1165590361,-0.1128100604,0.0313871019,0.0539408252,-0.4626779258,0.158140406,-0.1413083076,-0.1339838803,-0.4125747085,-0.3186384141,-0.2078332901,0.084862031,-0.1035225838,0.1193909347,0.4159374833,-0.0739142373,0.0533740297,0.1596258134,-0.2478346229,-0.1074660197,0.4095676839,-0.2446775883,-0.3135176003,0.4543264508,0.072319366,-0.1784508973,-0.1679846197,-0.0204058178,-0.0449930541,-0.3731469214,0.0616806075,-0.0245231371,-0.0404421724,-0.1267807186,0.3055093288,0.4896071851,-0.0096833957,0.2847531438,-0.7144518495,-0.3556802571,0.2366777062,-0.0546412282,0.2952449024,-0.2526121438,0.1385347694,-0.2104969025,0.0593775921,-0.2816969454,0.0230787657,-0.4564669728,-0.0837881491,0.1001639888,-0.0127250785,0.261464417,0.0349851549,0.1742554307,0.0547779761,-0.3107987046,-0.1835141629,-0.2588793635,0.1122425124,0.1209283322,-0.280680269,0.1347668916,0.0320408158,0.0065012369,-0.1869190335,-0.0410554186,0.1059171185,-0.1052688062,0.2443249673,0.2662098706,-0.0443941802,-0.2006893754,0.1606378704,-0.0766754299,0.0761699304,0.2360352427,-0.0732889846,-0.1039145216,0.0062031471,0.096316725,0.1853497028,0.001875477,-0.1533675939,0.3341062367,-0.1458846629,0.0689527541,0.2138649076,0.3672226369,0.2346559465,-0.3093515038,0.0030957852,0.2768591344,0.1573718935,-0.2205290496,0.0155065032,0.2803596258,0.2246055603,0.0225315634,0.2279590815,0.0356610417,-0.2950276732,0.3212125897,-0.072273463,0.2576577365,-0.1907857805,0.1527574956,-0.0063177482,-0.0556527823,-0.0287510473,0.1048015133,-0.1620068848,-0.0459872484,-0.0354439802,-0.1378435045,0.1757641137,-0.2960186005,0.17698735,-0.1391239464,-0.6813548803,0.3512970805,0.145282194,0.0003174874,0.0817152932,-0.0337929539,0.3823237121,-0.244302094,0.1538819671,-0.4536537826,0.1196980551,-0.1372100264,-0.1535899937,-0.2521290481,-0.2382503301,0.1049422771,0.1202083305,-0.015179147,-0.104823634,0.1012709513,0.0977588519,-0.1928468496,-0.4656092525,-0.1910358667,0.1723300517,-0.0744981617,-0.1790537536,0.3651694059,0.250138998,0.0551218837,0.3380270898,0.1576621383,0.464658469,0.3421230018,0.1389377862,-0.0273248088,-0.1392035633,0.1697890311,-0.0492099263,0.6106901765,0.1171118468,0.0790215358,0.4796912074,0.1567838937,-0.1200560853,-0.1497459859,-0.2036273479,0.0458743684,-0.0536595434,0.3299311996,-0.164848581,0.285926789,-0.2320778072,0.283267647,-0.4769468606,-0.1319950521,-0.0052832728,0.1365718246,0.2171823233,0.086733833,0.0526953824,-0.1251878291,0.6108255386,0.479727447,0.2195465416,-0.2921740413,-0.3415738642,-0.8700946569,0.3062995076,0.0001155213,-0.2973013818,0.1456142217,0.1620195806,0.0235687755,0.2645165622,0.2131074369,-0.1798981577,-0.0352095179,-0.1901183277,-0.1012415364,-0.1103344336,-0.1084368825,-0.2594906092,-0.1512116194,-0.2713238895,0.0640031695,0.1597411036,0.0773380622,0.051146511,-0.3196358383,-0.1865685284,0.2459578961,0.0364952013,0.1474298835,0.7800116539,0.054971572,-0.1918807626,-0.1947782189,-0.1534125954,0.0571291186,0.1269258112,-0.1336616576,0.5662155151,0.1243901476,0.1683933437,-0.2983309031,0.7553335428,0.1155050993,-0.4614267051,-0.2976111472,0.1096067652,-0.0259118788,0.0050343927,0.0148074282,0.3162671626,0.0250104703,0.5059561729,0.0054063075,-0.1460597217,0.7090224624,-0.3143617511,-0.1443075538,0.0277506523,0.1365534812,-0.2922633588,-0.1482906938,-0.7186961174,0.123646237,0.1957412958,0.0238483585,-0.130585745,0.2036420852,-0.1852646768,0.1725731194,-0.0994895622,0.1660661101,-0.1630265862,-0.1523593068,0.0270941108,-0.0499184653]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2210","title":"dataloading slow when using HUGE dataset","comments":"Hi ! Yes this is an issue with `datasets<=1.5.0`\r\nThis issue has been fixed by #2122 , we'll do a new release soon :)\r\nFor now you can test it on the `master` branch.","body":"Hi,\r\n\r\nWhen I use datasets with 600GB data, the dataloading speed increases significantly. \r\nI am experimenting with two datasets, and one is about 60GB and the other 600GB.\r\nSimply speaking, my code uses `datasets.set_format(\"torch\")` function and let pytorch-lightning handle ddp training.\r\nWhen looking at the pytorch-lightning supported profile of two different runs, I see that fetching a batch(`get_train_batch`) consumes an unreasonable amount of time when data is large. What could be the cause?\r\n\r\n* 60GB data\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 200.33 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nrun_training_epoch \t| 71.994 \t|1 \t| 71.994 \t| 35.937 \t|\r\nrun_training_batch \t| 0.64373 \t|100 \t| 64.373 \t| 32.133 \t|\r\noptimizer_step_and_closure_0 \t| 0.64322 \t|100 \t| 64.322 \t| 32.108 \t|\r\ntraining_step_and_backward \t| 0.61004 \t|100 \t| 61.004 \t| 30.452 \t|\r\nmodel_backward \t| 0.37552 \t|100 \t| 37.552 \t| 18.745 \t|\r\nmodel_forward \t| 0.22813 \t|100 \t| 22.813 \t| 11.387 \t|\r\ntraining_step \t| 0.22759 \t|100 \t| 22.759 \t| 11.361 \t|\r\nget_train_batch \t| 0.066385 \t|100 \t| 6.6385 \t| 3.3138 \t|\r\n```\r\n\r\n* 600GB data\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 3285.6 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nrun_training_epoch \t| 1397.9 \t|1 \t| 1397.9 \t| 42.546 \t|\r\nrun_training_batch \t| 7.2596 \t|100 \t| 725.96 \t| 22.095 \t|\r\noptimizer_step_and_closure_0 \t| 7.2589 \t|100 \t| 725.89 \t| 22.093 \t|\r\ntraining_step_and_backward \t| 7.223 \t|100 \t| 722.3 \t| 21.984 \t|\r\nmodel_backward \t| 6.9662 \t|100 \t| 696.62 \t| 21.202 \t|\r\nget_train_batch \t| 6.322 \t|100 \t| 632.2 \t| 19.241 \t|\r\nmodel_forward \t| 0.24902 \t|100 \t| 24.902 \t| 0.75789 \t|\r\ntraining_step \t| 0.2485 \t|100 \t| 24.85 \t| 0.75633 \t|\r\n```\r\n","comment_length":34,"text":"dataloading slow when using HUGE dataset\nHi,\r\n\r\nWhen I use datasets with 600GB data, the dataloading speed increases significantly. \r\nI am experimenting with two datasets, and one is about 60GB and the other 600GB.\r\nSimply speaking, my code uses `datasets.set_format(\"torch\")` function and let pytorch-lightning handle ddp training.\r\nWhen looking at the pytorch-lightning supported profile of two different runs, I see that fetching a batch(`get_train_batch`) consumes an unreasonable amount of time when data is large. What could be the cause?\r\n\r\n* 60GB data\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 200.33 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nrun_training_epoch \t| 71.994 \t|1 \t| 71.994 \t| 35.937 \t|\r\nrun_training_batch \t| 0.64373 \t|100 \t| 64.373 \t| 32.133 \t|\r\noptimizer_step_and_closure_0 \t| 0.64322 \t|100 \t| 64.322 \t| 32.108 \t|\r\ntraining_step_and_backward \t| 0.61004 \t|100 \t| 61.004 \t| 30.452 \t|\r\nmodel_backward \t| 0.37552 \t|100 \t| 37.552 \t| 18.745 \t|\r\nmodel_forward \t| 0.22813 \t|100 \t| 22.813 \t| 11.387 \t|\r\ntraining_step \t| 0.22759 \t|100 \t| 22.759 \t| 11.361 \t|\r\nget_train_batch \t| 0.066385 \t|100 \t| 6.6385 \t| 3.3138 \t|\r\n```\r\n\r\n* 600GB data\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 3285.6 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nrun_training_epoch \t| 1397.9 \t|1 \t| 1397.9 \t| 42.546 \t|\r\nrun_training_batch \t| 7.2596 \t|100 \t| 725.96 \t| 22.095 \t|\r\noptimizer_step_and_closure_0 \t| 7.2589 \t|100 \t| 725.89 \t| 22.093 \t|\r\ntraining_step_and_backward \t| 7.223 \t|100 \t| 722.3 \t| 21.984 \t|\r\nmodel_backward \t| 6.9662 \t|100 \t| 696.62 \t| 21.202 \t|\r\nget_train_batch \t| 6.322 \t|100 \t| 632.2 \t| 19.241 \t|\r\nmodel_forward \t| 0.24902 \t|100 \t| 24.902 \t| 0.75789 \t|\r\ntraining_step \t| 0.2485 \t|100 \t| 24.85 \t| 0.75633 \t|\r\n```\r\n\nHi ! Yes this is an issue with `datasets<=1.5.0`\r\nThis issue has been fixed by #2122 , we'll do a new release soon :)\r\nFor now you can test it on the `master` branch.","embeddings":[-0.5256726146,-0.2135471553,-0.0631118715,0.2832506299,0.1389209926,-0.0277841985,0.1501031071,0.2327974141,-0.1576929986,-0.0921895951,-0.1191496626,0.1439053416,-0.1486592591,-0.1823080629,-0.02395886,-0.1902338415,-0.0351155736,0.0441081598,-0.3937166631,-0.0954213142,0.1610000432,-0.3550180793,-0.0351877548,-0.194183588,-0.5536121726,-0.0915223435,0.1496039629,0.1078931913,-0.0339777619,-0.1050066277,0.2057130784,0.09889777,0.2247546464,0.5077366829,-0.0001180999,-0.0244024172,0.2768439651,0.1152795404,-0.3997972012,0.3242514133,-0.1600245386,-0.4594145417,-0.0535578541,-0.062167231,0.0956807137,-0.1884867251,-0.1830473393,-0.0993874297,0.1756862849,0.2129599899,0.1189069375,0.1577775031,-0.5473752618,-0.098498717,0.214314267,0.2566957474,-0.0703741089,0.4807890058,0.5785629749,-0.1316340566,-0.4877013564,0.0276801568,-0.1057128459,0.202533558,0.1750495285,-0.1479503065,-0.0157900527,-0.0179220717,0.3070149422,0.3041422069,0.574791491,-0.0630354658,-0.0341611616,-0.4913966954,-0.1859330386,-0.1462961733,-0.1616671532,0.1183803156,-0.2198844403,-0.1053002179,-0.5135316253,-0.0290877912,-0.1425199062,-0.0743970945,-0.1110334173,0.1562515646,0.084657304,0.1382816136,0.1856726706,-0.0827668831,0.2695870399,-0.2561670244,0.1405552477,0.0753811151,-0.4388132691,-0.0728772581,0.0968031213,-0.1686056852,0.1457460523,0.1857771128,-0.0019715128,0.1145246327,0.3466461301,-0.1527279168,0.3864695728,0.2220082432,-0.2823094428,0.1429652274,0.212145865,-0.1111238822,-0.0888329148,0.1805128306,-0.1373773515,-0.0715581998,0.2868857384,-0.3479177356,-0.244884029,-0.0368539765,-0.1311723292,-0.1438372582,-0.0216162466,-0.0731065199,0.3328087032,0.3250092864,-0.2769701183,0.1984865367,-0.3478924632,-0.0923691541,-0.1456362307,-0.0397834033,-0.2537195981,-0.362110883,-0.226942569,0.1369737536,0.3024580479,0.2183676064,0.0834764764,0.1128564104,0.0392978042,0.0382257886,-0.0897479132,-0.2322904319,-0.1428294927,0.1894834638,-0.2574872673,0.172670126,-0.128473267,0.543684423,-0.1996114999,0.403834343,-0.3958274126,-0.5039075017,-0.0904548243,0.151175335,-0.1105475575,-0.097300157,-0.4331704974,0.1526718438,0.0872960091,-0.1316269189,-0.0676238239,-0.3260264397,-0.0606062897,0.0065886779,0.1086093262,0.2436362356,-0.1952825487,-0.11479453,0.3166126609,0.1327327788,0.2860880494,0.5053688884,-0.57882303,0.3097110987,-0.0662712753,-0.4181604087,0.3517036438,-0.2286948413,-0.7667885423,0.4073599577,-0.2922475934,0.1128956899,0.3479835689,0.3371129036,0.4691828489,-0.1738813967,0.3076795042,0.6321914196,-0.0391158387,0.264008224,-0.2744717598,0.0676410496,0.0770711377,0.43430686,0.0485435091,-0.052612301,0.1456944644,0.3421171904,0.3252030611,-0.0484418273,-0.1815531403,0.2315501422,-0.1892012954,-0.0710435957,0.0750142857,0.0952446461,-0.535097301,0.3407512903,-0.1621402353,-0.3661987484,0.0473462529,-0.0397824273,0.0118515175,-0.0142923547,-0.2345995009,-0.0129343364,-0.0938502401,0.0316641331,0.3591991663,-0.0247140378,0.2586245239,0.1560186744,-0.3389982879,0.0163940713,-0.0187629238,0.0116387429,0.3819453716,0.0331695415,-0.0902024359,0.1316494346,0.0178084336,-0.174774617,-0.1438285261,0.012032887,0.0568178743,0.0211333353,-0.0174807515,0.2743739784,0.1342590451,0.1359640509,0.3300759792,-0.1267269701,0.1150928661,-0.3905344307,-0.2507505119,0.2684234679,0.050589174,0.3714226782,0.2174327075,-0.4488625526,0.1981152892,0.0361007266,0.2560522258,0.3012628257,0.3356685042,0.0393964238,0.6574735045,0.275136888,-0.0801873729,-0.0228258912,0.1576235443,-0.4038502276,-0.1537788063,0.1729964763,-0.3585538268,-0.3292834163,0.0519678108,0.0320587009,0.4785901904,0.2494711429,0.2282233983,-0.300632596,0.2224225104,-0.2358573377,0.1176818758,0.1153726503,0.1814310551,0.1138052791,0.3615718186,-0.0831156597,-0.1852358133,-0.1550077498,0.2703311145,0.2636873424,0.0484489426,0.365884006,-0.0708064735,-0.1336396188,-0.1987583339,0.0870623589,-0.0340273455,-0.1547109932,-0.0423091017,0.2943165004,0.1146710441,-0.0428337418,0.0574255437,0.3087708652,0.1741379648,-0.5386882424,-0.0388458744,-0.2046868652,-0.3223607838,0.0653300807,0.1915878803,-0.3348327875,0.380828768,0.2125113904,0.0180717967,-0.1169625893,-0.1210707054,0.0996205956,0.1340198964,-0.1610566974,-0.3777614832,-0.013924717,0.2444612384,0.1310233325,0.1397853047,-0.2243206054,0.056504719,0.1865109801,-0.1861602962,0.0061066365,0.2705686092,0.020120617,-0.2299532443,-0.2449507713,-0.0008155367,-0.2585439682,0.1199163124,0.1637964398,0.0298895016,0.1776873469,0.0702078491,0.0482572876,0.0167482868,-0.2887303829,0.1760679632,0.0177291259,-0.244820565,-0.4982016981,0.0225961376,-0.1001201794,0.5587286353,-0.693610847,-0.0073417486,-0.4541683793,0.2404493988,-0.1954780966,-0.0072444114,0.0014597039,-0.2785138786,-0.030139897,0.1731698513,0.0741404891,-0.1601118296,0.0552857369,0.3070278168,-0.2037931979,0.1371228695,0.2046261579,0.8623281717,-0.0597996116,-0.3136715591,0.0234497041,0.0580130182,0.3383568227,-0.1843931675,-0.1994634122,0.2740378082,-0.1152115315,-0.1137010902,0.2909384966,-0.0939478278,-0.3744614422,0.0598384477,-0.1998437792,0.3293919265,-0.093162775,0.3659172952,-0.2859895229,-0.0506083146,0.0342718847,0.1668292135,-0.1679098159,-0.2601915598,-0.0100645125,-0.2233889848,-0.0259368271,-0.0258858558,0.0280825291,0.1818678975,-0.5656908751,0.1588510722,0.0670242161,0.2626863718,0.1121223718,-0.0408620648,0.3483414948,0.1353435367,0.6014813781,0.2783255279,0.2870131731,-0.1728001386,-0.4035538435,-0.2081205398,-0.0747035518,-0.3570012748,0.3095758557,0.4312022924,0.2041463703,-0.1900150478,-0.2005497515,0.0862455741,0.2199070752,-0.1666173935,-0.3529470861,-0.1158731058,-0.1915997267,-0.0119944857,0.3463253677,-0.0044022431,0.1742774397,-0.0511138812,-0.0573318228,-0.1515638381,0.2218330353,0.1605645567,0.0436300375,-0.1091810837,0.1068404838,0.1961916536,-0.0122198854,0.3635070622,0.4482749403,0.3003789485,0.0266805906,-0.2349670529,0.1634908617,0.2717039287,0.2000735998,0.3689030707,0.222268939,-0.0828681439,-0.2962101102,0.2456645668,-0.1849709898,0.2391969115,0.385158211,-0.0200758316,-0.6416810751,-0.4975689948,0.5110851526,0.0045210938,0.0198011585,0.2292898744,-0.4414499402,0.0129967798,-0.0448420718,0.019173054,0.6361575127,-0.1764858067,0.3652829826,-0.0125756171,0.1191395521,0.3268478513,-0.0260471795,0.0624023974,-0.1856810749,-0.0536914207,0.2275849134,-0.260628283,-0.0678394213,0.2028576881,0.3317076564,0.1882179976,-0.1313769221,0.1802570075,0.0128784003,0.4488943219,0.087104477,-0.2177276313,-0.0922996104,0.1461530626,-0.3234034479,0.0433328524,-0.1008133143,0.0351995863,-0.010129232,0.0288976077,-0.2235299349,0.1035109088,-0.2309993953,0.4226004183,-0.2515100241,-0.5213246942,-0.0270561986,0.0712961853,-0.5702292323,-0.0277219117,-0.065616183,0.3650149703,-0.275878042,0.2803942561,0.3480331004,-0.2569478452,0.2451663017,0.0102142747,0.1187238097,0.0089513343,-0.0229716953,-0.3396007121,-0.0272119641,0.3641217351,0.2775537074,-0.0798749328,-0.2256363928,0.0126863727,0.0174238011,-0.0971890166,0.036334645,0.1916294843,0.0123243462,0.383453548,-0.030107094,-0.5099642873,-0.0553476103,0.3252414465,0.207098335,-0.2156796902,0.3971477151,0.0324988365,-0.1932039857,-0.0374405347,0.247715205,0.1495770961,-0.3545184433,0.1787432134,-0.0600161627,0.1370621324,0.1939854324,-0.186659798,-0.0886498168,-0.3123488724,-0.2700841427,-0.0955469832,-0.3872537315,-0.1503923386,-0.036697384,0.0782166198,0.0422228314,-0.1261561811,-0.0374239087,0.434088558,-0.2102997601,-0.0549222231,0.2392007858,0.058531478,-0.0265866295,-0.2505361736,0.2138557732,0.0045713675,-0.056154374,-0.0516307205,0.0408308543,-0.2512097359,-0.0473772548,0.1242967919,-0.0042895847,-0.1202432066,-0.2368971705,-0.3456606567,0.0548121035,-0.2050169408,0.4311216772,0.2175818384,0.0157253165,-0.3140423,0.303971976,0.0252249595,-0.0269068684,0.4463458657,-0.1477811933,0.2223430872,0.1358648092,0.0677679479,0.2876966,-0.2494032234,-0.2128688991,0.0116770547,0.590990901,-0.1453072578,0.1489440948,-0.3075108528,0.1673520356,-0.0027907775,0.0725504234,0.3108545244,0.013528306,-0.4140635431,0.1785523295,0.0868745297,0.0393486954,-0.3019305766,-0.0945117697,-0.0371351987,-0.0766192675,0.1334467083,0.0568924695,0.2192476839,-0.1308561713,-0.1306972504,0.269646287,0.1258151531,-0.0478358045,0.3882276118,0.2731565833,-0.1047928333,0.4876898527,0.3903231919,0.2716769874,0.3355735242,0.102062434,0.6923543215,0.1055385768,-0.0799944848,0.0108039565,-0.3727954924,-0.1548216492,0.2638239264,-0.166650176,0.1343443096,-0.1949355602,0.0091241905,-0.1912466586,-0.0653578565,-0.3557800055,-0.0264231265,-0.2079644948,0.0053799045,0.1997800916,0.0551402569,0.2186171561,0.3261700571,-0.3942027688,-0.1147039756,-0.0518027246,0.0310508795,-0.1218374372,-0.0492100157,-0.0098884813,0.1416735053,-0.0238095969,-0.1151937395,0.0647130162,-0.0013871252,-0.0761072934,0.0374418758,0.1494815648,0.2420556843,0.3533017337,0.058909595,-0.0446690582,0.126529485,-0.0062668235,-0.1860655695,0.5170593262,0.0138476584,-0.0091504036,0.1525837928,0.0038943272,-0.2196478546,-0.0569683984,0.0233472437,-0.2085929066,0.0347209871,0.189181909,0.1516201794,-0.2049027234,-0.261921227,0.1431287825,-0.2022099644,-0.2747009397,0.3893353343,-0.25304389,-0.0148345986,-0.0078042955,0.0281694066,-0.3048426509,0.907115221,0.2623467743,0.3463222682,-0.3564275503,0.1867588162,-0.419162631,0.0239930898,-0.5187041759,0.202119261,0.1043012068,0.231370464,-0.0152566805,0.1377824843,-0.1181806847,0.0812142044,0.0129073299,-0.2207535207,-0.255010128,-0.0874242559,-0.1549421102,0.0752035081,-0.0481376573,-0.6022773981,0.3446797729,-0.0124616958,-0.015062266,-0.2360092402,0.0408715643,0.1365883052,0.1809746772,0.2468440086,0.1607690006,0.5449459553,0.2268916816,-0.0201756246,-0.1396606416,0.030689491,-0.1491610259,0.0218766741,0.0283089224,-0.0148983467,0.0355800688,-0.0963340849,0.1079289988,0.4588612318,-0.0429329462,-0.2395263016,-0.03266792,-0.3271843195,-0.0190351605,0.2352484912,0.2845640779,0.2960159779,-0.0742801875,-0.1876186728,-0.2534402907,-0.1082600877,0.3916227818,-0.4769768119,-0.2205241024,-0.2020839453,0.3011841476,0.1523760557,0.1260053962,-0.1114242747,0.0943015814,0.1843781769,-0.0639625713,0.003885668,0.3781554401,-0.2533888519,0.1732249409,-0.1296792328,0.501232028,-0.1671091467,-0.3956730068,-0.0955530182,-0.1395307034]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2210","title":"dataloading slow when using HUGE dataset","comments":"Hi, thank you for your answer. I did not realize that my issue stems from the same problem. ","body":"Hi,\r\n\r\nWhen I use datasets with 600GB data, the dataloading speed increases significantly. \r\nI am experimenting with two datasets, and one is about 60GB and the other 600GB.\r\nSimply speaking, my code uses `datasets.set_format(\"torch\")` function and let pytorch-lightning handle ddp training.\r\nWhen looking at the pytorch-lightning supported profile of two different runs, I see that fetching a batch(`get_train_batch`) consumes an unreasonable amount of time when data is large. What could be the cause?\r\n\r\n* 60GB data\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 200.33 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nrun_training_epoch \t| 71.994 \t|1 \t| 71.994 \t| 35.937 \t|\r\nrun_training_batch \t| 0.64373 \t|100 \t| 64.373 \t| 32.133 \t|\r\noptimizer_step_and_closure_0 \t| 0.64322 \t|100 \t| 64.322 \t| 32.108 \t|\r\ntraining_step_and_backward \t| 0.61004 \t|100 \t| 61.004 \t| 30.452 \t|\r\nmodel_backward \t| 0.37552 \t|100 \t| 37.552 \t| 18.745 \t|\r\nmodel_forward \t| 0.22813 \t|100 \t| 22.813 \t| 11.387 \t|\r\ntraining_step \t| 0.22759 \t|100 \t| 22.759 \t| 11.361 \t|\r\nget_train_batch \t| 0.066385 \t|100 \t| 6.6385 \t| 3.3138 \t|\r\n```\r\n\r\n* 600GB data\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 3285.6 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nrun_training_epoch \t| 1397.9 \t|1 \t| 1397.9 \t| 42.546 \t|\r\nrun_training_batch \t| 7.2596 \t|100 \t| 725.96 \t| 22.095 \t|\r\noptimizer_step_and_closure_0 \t| 7.2589 \t|100 \t| 725.89 \t| 22.093 \t|\r\ntraining_step_and_backward \t| 7.223 \t|100 \t| 722.3 \t| 21.984 \t|\r\nmodel_backward \t| 6.9662 \t|100 \t| 696.62 \t| 21.202 \t|\r\nget_train_batch \t| 6.322 \t|100 \t| 632.2 \t| 19.241 \t|\r\nmodel_forward \t| 0.24902 \t|100 \t| 24.902 \t| 0.75789 \t|\r\ntraining_step \t| 0.2485 \t|100 \t| 24.85 \t| 0.75633 \t|\r\n```\r\n","comment_length":18,"text":"dataloading slow when using HUGE dataset\nHi,\r\n\r\nWhen I use datasets with 600GB data, the dataloading speed increases significantly. \r\nI am experimenting with two datasets, and one is about 60GB and the other 600GB.\r\nSimply speaking, my code uses `datasets.set_format(\"torch\")` function and let pytorch-lightning handle ddp training.\r\nWhen looking at the pytorch-lightning supported profile of two different runs, I see that fetching a batch(`get_train_batch`) consumes an unreasonable amount of time when data is large. What could be the cause?\r\n\r\n* 60GB data\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 200.33 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nrun_training_epoch \t| 71.994 \t|1 \t| 71.994 \t| 35.937 \t|\r\nrun_training_batch \t| 0.64373 \t|100 \t| 64.373 \t| 32.133 \t|\r\noptimizer_step_and_closure_0 \t| 0.64322 \t|100 \t| 64.322 \t| 32.108 \t|\r\ntraining_step_and_backward \t| 0.61004 \t|100 \t| 61.004 \t| 30.452 \t|\r\nmodel_backward \t| 0.37552 \t|100 \t| 37.552 \t| 18.745 \t|\r\nmodel_forward \t| 0.22813 \t|100 \t| 22.813 \t| 11.387 \t|\r\ntraining_step \t| 0.22759 \t|100 \t| 22.759 \t| 11.361 \t|\r\nget_train_batch \t| 0.066385 \t|100 \t| 6.6385 \t| 3.3138 \t|\r\n```\r\n\r\n* 600GB data\r\n```\r\nAction \t| Mean duration (s)\t|Num calls \t| Total time (s) \t| Percentage % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nTotal \t| - \t|_ \t| 3285.6 \t| 100 % \t|\r\n------------------------------------------------------------------------------------------------------------------------------------\r\nrun_training_epoch \t| 1397.9 \t|1 \t| 1397.9 \t| 42.546 \t|\r\nrun_training_batch \t| 7.2596 \t|100 \t| 725.96 \t| 22.095 \t|\r\noptimizer_step_and_closure_0 \t| 7.2589 \t|100 \t| 725.89 \t| 22.093 \t|\r\ntraining_step_and_backward \t| 7.223 \t|100 \t| 722.3 \t| 21.984 \t|\r\nmodel_backward \t| 6.9662 \t|100 \t| 696.62 \t| 21.202 \t|\r\nget_train_batch \t| 6.322 \t|100 \t| 632.2 \t| 19.241 \t|\r\nmodel_forward \t| 0.24902 \t|100 \t| 24.902 \t| 0.75789 \t|\r\ntraining_step \t| 0.2485 \t|100 \t| 24.85 \t| 0.75633 \t|\r\n```\r\n\nHi, thank you for your answer. I did not realize that my issue stems from the same problem. ","embeddings":[-0.5256726146,-0.2135471553,-0.0631118715,0.2832506299,0.1389209926,-0.0277841985,0.1501031071,0.2327974141,-0.1576929986,-0.0921895951,-0.1191496626,0.1439053416,-0.1486592591,-0.1823080629,-0.02395886,-0.1902338415,-0.0351155736,0.0441081598,-0.3937166631,-0.0954213142,0.1610000432,-0.3550180793,-0.0351877548,-0.194183588,-0.5536121726,-0.0915223435,0.1496039629,0.1078931913,-0.0339777619,-0.1050066277,0.2057130784,0.09889777,0.2247546464,0.5077366829,-0.0001180999,-0.0244024172,0.2768439651,0.1152795404,-0.3997972012,0.3242514133,-0.1600245386,-0.4594145417,-0.0535578541,-0.062167231,0.0956807137,-0.1884867251,-0.1830473393,-0.0993874297,0.1756862849,0.2129599899,0.1189069375,0.1577775031,-0.5473752618,-0.098498717,0.214314267,0.2566957474,-0.0703741089,0.4807890058,0.5785629749,-0.1316340566,-0.4877013564,0.0276801568,-0.1057128459,0.202533558,0.1750495285,-0.1479503065,-0.0157900527,-0.0179220717,0.3070149422,0.3041422069,0.574791491,-0.0630354658,-0.0341611616,-0.4913966954,-0.1859330386,-0.1462961733,-0.1616671532,0.1183803156,-0.2198844403,-0.1053002179,-0.5135316253,-0.0290877912,-0.1425199062,-0.0743970945,-0.1110334173,0.1562515646,0.084657304,0.1382816136,0.1856726706,-0.0827668831,0.2695870399,-0.2561670244,0.1405552477,0.0753811151,-0.4388132691,-0.0728772581,0.0968031213,-0.1686056852,0.1457460523,0.1857771128,-0.0019715128,0.1145246327,0.3466461301,-0.1527279168,0.3864695728,0.2220082432,-0.2823094428,0.1429652274,0.212145865,-0.1111238822,-0.0888329148,0.1805128306,-0.1373773515,-0.0715581998,0.2868857384,-0.3479177356,-0.244884029,-0.0368539765,-0.1311723292,-0.1438372582,-0.0216162466,-0.0731065199,0.3328087032,0.3250092864,-0.2769701183,0.1984865367,-0.3478924632,-0.0923691541,-0.1456362307,-0.0397834033,-0.2537195981,-0.362110883,-0.226942569,0.1369737536,0.3024580479,0.2183676064,0.0834764764,0.1128564104,0.0392978042,0.0382257886,-0.0897479132,-0.2322904319,-0.1428294927,0.1894834638,-0.2574872673,0.172670126,-0.128473267,0.543684423,-0.1996114999,0.403834343,-0.3958274126,-0.5039075017,-0.0904548243,0.151175335,-0.1105475575,-0.097300157,-0.4331704974,0.1526718438,0.0872960091,-0.1316269189,-0.0676238239,-0.3260264397,-0.0606062897,0.0065886779,0.1086093262,0.2436362356,-0.1952825487,-0.11479453,0.3166126609,0.1327327788,0.2860880494,0.5053688884,-0.57882303,0.3097110987,-0.0662712753,-0.4181604087,0.3517036438,-0.2286948413,-0.7667885423,0.4073599577,-0.2922475934,0.1128956899,0.3479835689,0.3371129036,0.4691828489,-0.1738813967,0.3076795042,0.6321914196,-0.0391158387,0.264008224,-0.2744717598,0.0676410496,0.0770711377,0.43430686,0.0485435091,-0.052612301,0.1456944644,0.3421171904,0.3252030611,-0.0484418273,-0.1815531403,0.2315501422,-0.1892012954,-0.0710435957,0.0750142857,0.0952446461,-0.535097301,0.3407512903,-0.1621402353,-0.3661987484,0.0473462529,-0.0397824273,0.0118515175,-0.0142923547,-0.2345995009,-0.0129343364,-0.0938502401,0.0316641331,0.3591991663,-0.0247140378,0.2586245239,0.1560186744,-0.3389982879,0.0163940713,-0.0187629238,0.0116387429,0.3819453716,0.0331695415,-0.0902024359,0.1316494346,0.0178084336,-0.174774617,-0.1438285261,0.012032887,0.0568178743,0.0211333353,-0.0174807515,0.2743739784,0.1342590451,0.1359640509,0.3300759792,-0.1267269701,0.1150928661,-0.3905344307,-0.2507505119,0.2684234679,0.050589174,0.3714226782,0.2174327075,-0.4488625526,0.1981152892,0.0361007266,0.2560522258,0.3012628257,0.3356685042,0.0393964238,0.6574735045,0.275136888,-0.0801873729,-0.0228258912,0.1576235443,-0.4038502276,-0.1537788063,0.1729964763,-0.3585538268,-0.3292834163,0.0519678108,0.0320587009,0.4785901904,0.2494711429,0.2282233983,-0.300632596,0.2224225104,-0.2358573377,0.1176818758,0.1153726503,0.1814310551,0.1138052791,0.3615718186,-0.0831156597,-0.1852358133,-0.1550077498,0.2703311145,0.2636873424,0.0484489426,0.365884006,-0.0708064735,-0.1336396188,-0.1987583339,0.0870623589,-0.0340273455,-0.1547109932,-0.0423091017,0.2943165004,0.1146710441,-0.0428337418,0.0574255437,0.3087708652,0.1741379648,-0.5386882424,-0.0388458744,-0.2046868652,-0.3223607838,0.0653300807,0.1915878803,-0.3348327875,0.380828768,0.2125113904,0.0180717967,-0.1169625893,-0.1210707054,0.0996205956,0.1340198964,-0.1610566974,-0.3777614832,-0.013924717,0.2444612384,0.1310233325,0.1397853047,-0.2243206054,0.056504719,0.1865109801,-0.1861602962,0.0061066365,0.2705686092,0.020120617,-0.2299532443,-0.2449507713,-0.0008155367,-0.2585439682,0.1199163124,0.1637964398,0.0298895016,0.1776873469,0.0702078491,0.0482572876,0.0167482868,-0.2887303829,0.1760679632,0.0177291259,-0.244820565,-0.4982016981,0.0225961376,-0.1001201794,0.5587286353,-0.693610847,-0.0073417486,-0.4541683793,0.2404493988,-0.1954780966,-0.0072444114,0.0014597039,-0.2785138786,-0.030139897,0.1731698513,0.0741404891,-0.1601118296,0.0552857369,0.3070278168,-0.2037931979,0.1371228695,0.2046261579,0.8623281717,-0.0597996116,-0.3136715591,0.0234497041,0.0580130182,0.3383568227,-0.1843931675,-0.1994634122,0.2740378082,-0.1152115315,-0.1137010902,0.2909384966,-0.0939478278,-0.3744614422,0.0598384477,-0.1998437792,0.3293919265,-0.093162775,0.3659172952,-0.2859895229,-0.0506083146,0.0342718847,0.1668292135,-0.1679098159,-0.2601915598,-0.0100645125,-0.2233889848,-0.0259368271,-0.0258858558,0.0280825291,0.1818678975,-0.5656908751,0.1588510722,0.0670242161,0.2626863718,0.1121223718,-0.0408620648,0.3483414948,0.1353435367,0.6014813781,0.2783255279,0.2870131731,-0.1728001386,-0.4035538435,-0.2081205398,-0.0747035518,-0.3570012748,0.3095758557,0.4312022924,0.2041463703,-0.1900150478,-0.2005497515,0.0862455741,0.2199070752,-0.1666173935,-0.3529470861,-0.1158731058,-0.1915997267,-0.0119944857,0.3463253677,-0.0044022431,0.1742774397,-0.0511138812,-0.0573318228,-0.1515638381,0.2218330353,0.1605645567,0.0436300375,-0.1091810837,0.1068404838,0.1961916536,-0.0122198854,0.3635070622,0.4482749403,0.3003789485,0.0266805906,-0.2349670529,0.1634908617,0.2717039287,0.2000735998,0.3689030707,0.222268939,-0.0828681439,-0.2962101102,0.2456645668,-0.1849709898,0.2391969115,0.385158211,-0.0200758316,-0.6416810751,-0.4975689948,0.5110851526,0.0045210938,0.0198011585,0.2292898744,-0.4414499402,0.0129967798,-0.0448420718,0.019173054,0.6361575127,-0.1764858067,0.3652829826,-0.0125756171,0.1191395521,0.3268478513,-0.0260471795,0.0624023974,-0.1856810749,-0.0536914207,0.2275849134,-0.260628283,-0.0678394213,0.2028576881,0.3317076564,0.1882179976,-0.1313769221,0.1802570075,0.0128784003,0.4488943219,0.087104477,-0.2177276313,-0.0922996104,0.1461530626,-0.3234034479,0.0433328524,-0.1008133143,0.0351995863,-0.010129232,0.0288976077,-0.2235299349,0.1035109088,-0.2309993953,0.4226004183,-0.2515100241,-0.5213246942,-0.0270561986,0.0712961853,-0.5702292323,-0.0277219117,-0.065616183,0.3650149703,-0.275878042,0.2803942561,0.3480331004,-0.2569478452,0.2451663017,0.0102142747,0.1187238097,0.0089513343,-0.0229716953,-0.3396007121,-0.0272119641,0.3641217351,0.2775537074,-0.0798749328,-0.2256363928,0.0126863727,0.0174238011,-0.0971890166,0.036334645,0.1916294843,0.0123243462,0.383453548,-0.030107094,-0.5099642873,-0.0553476103,0.3252414465,0.207098335,-0.2156796902,0.3971477151,0.0324988365,-0.1932039857,-0.0374405347,0.247715205,0.1495770961,-0.3545184433,0.1787432134,-0.0600161627,0.1370621324,0.1939854324,-0.186659798,-0.0886498168,-0.3123488724,-0.2700841427,-0.0955469832,-0.3872537315,-0.1503923386,-0.036697384,0.0782166198,0.0422228314,-0.1261561811,-0.0374239087,0.434088558,-0.2102997601,-0.0549222231,0.2392007858,0.058531478,-0.0265866295,-0.2505361736,0.2138557732,0.0045713675,-0.056154374,-0.0516307205,0.0408308543,-0.2512097359,-0.0473772548,0.1242967919,-0.0042895847,-0.1202432066,-0.2368971705,-0.3456606567,0.0548121035,-0.2050169408,0.4311216772,0.2175818384,0.0157253165,-0.3140423,0.303971976,0.0252249595,-0.0269068684,0.4463458657,-0.1477811933,0.2223430872,0.1358648092,0.0677679479,0.2876966,-0.2494032234,-0.2128688991,0.0116770547,0.590990901,-0.1453072578,0.1489440948,-0.3075108528,0.1673520356,-0.0027907775,0.0725504234,0.3108545244,0.013528306,-0.4140635431,0.1785523295,0.0868745297,0.0393486954,-0.3019305766,-0.0945117697,-0.0371351987,-0.0766192675,0.1334467083,0.0568924695,0.2192476839,-0.1308561713,-0.1306972504,0.269646287,0.1258151531,-0.0478358045,0.3882276118,0.2731565833,-0.1047928333,0.4876898527,0.3903231919,0.2716769874,0.3355735242,0.102062434,0.6923543215,0.1055385768,-0.0799944848,0.0108039565,-0.3727954924,-0.1548216492,0.2638239264,-0.166650176,0.1343443096,-0.1949355602,0.0091241905,-0.1912466586,-0.0653578565,-0.3557800055,-0.0264231265,-0.2079644948,0.0053799045,0.1997800916,0.0551402569,0.2186171561,0.3261700571,-0.3942027688,-0.1147039756,-0.0518027246,0.0310508795,-0.1218374372,-0.0492100157,-0.0098884813,0.1416735053,-0.0238095969,-0.1151937395,0.0647130162,-0.0013871252,-0.0761072934,0.0374418758,0.1494815648,0.2420556843,0.3533017337,0.058909595,-0.0446690582,0.126529485,-0.0062668235,-0.1860655695,0.5170593262,0.0138476584,-0.0091504036,0.1525837928,0.0038943272,-0.2196478546,-0.0569683984,0.0233472437,-0.2085929066,0.0347209871,0.189181909,0.1516201794,-0.2049027234,-0.261921227,0.1431287825,-0.2022099644,-0.2747009397,0.3893353343,-0.25304389,-0.0148345986,-0.0078042955,0.0281694066,-0.3048426509,0.907115221,0.2623467743,0.3463222682,-0.3564275503,0.1867588162,-0.419162631,0.0239930898,-0.5187041759,0.202119261,0.1043012068,0.231370464,-0.0152566805,0.1377824843,-0.1181806847,0.0812142044,0.0129073299,-0.2207535207,-0.255010128,-0.0874242559,-0.1549421102,0.0752035081,-0.0481376573,-0.6022773981,0.3446797729,-0.0124616958,-0.015062266,-0.2360092402,0.0408715643,0.1365883052,0.1809746772,0.2468440086,0.1607690006,0.5449459553,0.2268916816,-0.0201756246,-0.1396606416,0.030689491,-0.1491610259,0.0218766741,0.0283089224,-0.0148983467,0.0355800688,-0.0963340849,0.1079289988,0.4588612318,-0.0429329462,-0.2395263016,-0.03266792,-0.3271843195,-0.0190351605,0.2352484912,0.2845640779,0.2960159779,-0.0742801875,-0.1876186728,-0.2534402907,-0.1082600877,0.3916227818,-0.4769768119,-0.2205241024,-0.2020839453,0.3011841476,0.1523760557,0.1260053962,-0.1114242747,0.0943015814,0.1843781769,-0.0639625713,0.003885668,0.3781554401,-0.2533888519,0.1732249409,-0.1296792328,0.501232028,-0.1671091467,-0.3956730068,-0.0955530182,-0.1395307034]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2207","title":"making labels consistent across the datasets","comments":"Hi ! The ClassLabel feature type encodes the labels as integers.\r\nThe integer corresponds to the index of the label name in the `names` list of the ClassLabel.\r\nHere that means that the labels are 'entailment' (0), 'neutral' (1), 'contradiction' (2).\r\n\r\nYou can get the label names back by using `a.features['label'].int2str(i)`.\r\n","body":"Hi\r\nFor accessing the labels one can type \r\n```\r\n>>> a.features['label']\r\nClassLabel(num_classes=3, names=['entailment', 'neutral', 'contradiction'], names_file=None, id=None)\r\n```\r\nThe labels however are not consistent with the actual labels sometimes, for instance in case of XNLI, the actual labels are 0,1,2, but if one try to access as above they are entailment, neutral,contradiction,\r\nit would be great to have the labels consistent.\r\n\r\nthanks \r\n","comment_length":51,"text":"making labels consistent across the datasets\nHi\r\nFor accessing the labels one can type \r\n```\r\n>>> a.features['label']\r\nClassLabel(num_classes=3, names=['entailment', 'neutral', 'contradiction'], names_file=None, id=None)\r\n```\r\nThe labels however are not consistent with the actual labels sometimes, for instance in case of XNLI, the actual labels are 0,1,2, but if one try to access as above they are entailment, neutral,contradiction,\r\nit would be great to have the labels consistent.\r\n\r\nthanks \r\n\nHi ! The ClassLabel feature type encodes the labels as integers.\r\nThe integer corresponds to the index of the label name in the `names` list of the ClassLabel.\r\nHere that means that the labels are 'entailment' (0), 'neutral' (1), 'contradiction' (2).\r\n\r\nYou can get the label names back by using `a.features['label'].int2str(i)`.\r\n","embeddings":[0.0162600745,-0.1287364215,-0.0699902028,0.4029694796,0.3827131689,-0.1300277263,0.4260282815,0.0234175641,0.0852748901,0.2710852325,-0.2305020094,0.5326738358,-0.0153415753,0.4047870934,-0.307893008,0.0154723981,-0.1154663414,0.1015716866,0.1249587983,-0.3230445385,-0.2213249803,-0.1315994114,0.0152088413,0.334883064,-0.4379475415,-0.141682148,0.0278720055,-0.1909615248,0.0741544813,-0.4989104271,0.1208871752,0.3788155615,0.0031806116,0.1615455449,-0.0000959714,-0.2662477195,0.1202937737,0.0277738199,-0.0765430629,0.0507424362,-0.2216129452,-0.1730420589,0.0833134875,-0.4307701588,-0.2084742486,0.0238299482,-0.1198960394,-0.2397072762,-0.1590719521,-0.0710135326,0.2925275564,0.0040540034,0.128914088,0.1718682945,0.4010023177,-0.0080782948,0.0416165479,0.1494404376,0.3026364148,0.129017666,0.0510599241,0.5405575037,-0.035616409,-0.062250562,0.3284755647,0.1424699873,0.0399955511,-0.3465792239,0.0029805398,0.4912868738,0.584572196,-0.249494642,-0.4491060674,-0.1393752396,0.1123988628,-0.2547758222,-0.1436599791,0.0708718002,0.2149394006,0.042105861,0.0539676882,-0.0519852489,-0.2780491412,0.1498251706,0.0187883321,0.4695444107,0.0171505082,0.2028816044,-0.0887358561,-0.4577319026,0.0376456343,-0.0076666591,0.0038778041,0.2679313123,-0.1903167069,0.0004920989,-0.1286011934,0.0033978471,0.0820229799,0.1513404101,-0.2002418488,0.1516584754,-0.2954198718,0.1294176728,0.0067377677,0.3571250737,0.5226082802,0.2573030591,0.3497829437,-0.0255378429,-0.3429965973,-0.0779767856,-0.1289377362,0.117494978,0.3556327224,-0.0392704681,0.0160166845,-0.3400921524,-0.3314759433,0.0508840345,-0.1754248589,-0.0531823412,0.1985355914,0.2774687409,0.1758970618,0.3547516167,0.11419563,-0.0679105893,-0.129161194,-0.3238432109,-0.2257992774,-0.0322754085,-0.0664247125,-0.172146827,-0.0451963358,0.2057716101,0.2654462755,-0.1270953119,0.0177972671,-0.1221840158,-0.0213277731,-0.1247066259,0.013071551,-0.1685440987,-0.3194271922,0.0438333489,0.0006262753,-0.1108302698,-0.2505707741,-0.2797500491,-0.2588024437,0.0304975491,0.2304798216,0.3659622669,0.1701172888,-0.1080110818,0.3435677886,0.1185834035,-0.1111146957,-0.1016408578,0.1304856837,-0.3175900578,0.1881766468,-0.1673260629,-0.0616379194,0.0561994761,-0.3011079431,-0.1929347962,0.3234800994,-0.1957866251,-0.1605420858,0.0398119129,-0.0043440871,-0.1616312861,-0.0097828945,0.081334129,0.4260672331,-0.4890017807,-0.3455292881,-0.215180397,0.1158952788,-0.1208518222,0.2246493697,0.4517965615,0.092608735,0.0097813914,0.1628989875,0.0739813149,0.1442049444,-0.0268066023,-0.1295728534,-0.0836474597,0.0303375926,0.1009498239,-0.2006149888,-0.2347719818,0.1300798506,-0.0363650061,0.2824650407,-0.4692776203,0.0849545226,0.1094819158,0.4002465308,0.2273927182,-0.0841675252,-0.2679437995,-0.2897003591,0.2414544821,-0.0737488717,0.2531753778,0.2952412963,-0.4176500738,-0.0334187858,-0.114094846,-0.2889352143,-0.0800449848,0.3252169788,0.3039772511,-0.1158038303,-0.069671005,-0.1151366979,-0.0504070669,-0.1801474392,0.0385653004,-0.1042997018,-0.1267308742,0.0450599007,0.172812894,-0.1324340999,0.4198814332,0.2545403242,0.0150632067,-0.0720482096,0.2207140476,0.2466808558,-0.198717162,0.0628409237,0.3595350087,0.2418501973,-0.2225042135,0.04424062,0.1010803655,0.1455687732,0.1002046764,-0.1473449767,0.5054873228,-0.0140543981,0.0740434527,0.0558408797,0.0776267722,0.2117099166,-0.0607625097,-0.3135550618,-0.1297812462,0.0235370044,-0.0802810043,0.0891641453,0.1970622987,-0.6030899286,0.5452006459,0.6344435811,0.0285052974,0.1303443313,-0.2016640157,-0.2598124444,0.2631831467,0.137886703,0.1743709147,-0.091809541,0.3492028117,0.0007843775,-0.1269412488,-0.2725159824,-0.0351667553,0.032810647,-0.1560471207,0.1233802065,-0.066625379,0.1052386835,-0.1327948123,0.0184839554,-0.4001693428,-0.1800195128,-0.0406782813,-0.3130125701,0.1659746319,-0.1016380936,-0.2682479322,-0.3333453238,-0.4044164419,-0.3766966164,-0.4287503958,0.0814472511,-0.0757482573,-0.2525373995,0.2243506312,-0.0716915578,0.2579423189,0.0325993337,-0.0418787524,0.2593669593,-0.4869864881,-0.4448615611,0.1094455943,-0.0474651419,0.0781192034,0.2551839054,-0.1009775475,0.193041876,-0.0853098035,-0.4455027878,-0.0569377728,-0.1742358953,0.1129001975,0.1211117879,-0.0299212802,0.0745403618,0.0304516181,0.1996329427,-0.1402624696,0.0097622732,-0.0211205278,-0.0075489497,-0.3436497152,-0.0661106184,-0.6197527051,-0.3925741315,-0.2936830521,0.0688273609,-0.0980795622,0.095512405,0.060406141,-0.2549695671,0.0536748916,0.0824274421,0.5573016405,-0.296551615,-0.1041312292,0.1520818025,-0.1441338211,-0.1998024136,0.0779300928,-0.0933245122,0.1299971789,-0.0859044269,0.018262472,-0.3870120645,-0.096018739,0.2507855296,0.1855876893,0.2257487923,0.0180490837,0.2406856716,-0.1593569666,-0.1146705598,-0.1340183765,0.3696860373,0.0592280179,0.2220904082,0.0443476737,-0.170018658,-0.2130917609,0.0299039409,0.1449424475,-0.1307631284,0.0554137751,-0.0990678445,0.1191018224,0.282528013,-0.0518533587,0.053147953,0.0907538906,-0.0187692717,0.1328069419,-0.1242526397,-0.1352366507,-0.1660631001,0.0671977103,-0.1685390323,-0.2314542681,-0.0948908627,0.149899736,0.0634464473,-0.0005060376,0.1474652141,-0.1570693105,-0.1121991575,0.0188881885,-0.0085746795,-0.3010180593,0.0863813832,-0.668708086,-0.370092541,0.1302803159,0.3246813416,0.1557057649,0.0092274249,0.1074413955,-0.1605980545,0.0526586436,0.1092467457,0.3452030718,-0.1231956333,0.0597348362,0.1564569473,-0.0374144427,-0.1318472475,-0.0317303166,-0.3407370746,-0.4124085307,0.2561511397,0.1757475138,-0.4678797424,-0.1738601327,0.032206662,0.0417838767,-0.3526409268,0.1511853933,0.016568793,0.1702745557,-0.1063529328,-0.1102944016,-0.2012555599,0.2618975043,-0.301119417,0.1390096247,-0.1100598052,0.1207371131,0.3727037907,0.5779052377,0.4539884031,-0.053326901,-0.1763789058,0.1368285418,0.3725435734,0.1078229994,0.5516909361,-0.3065680563,-0.3936004043,-0.1471506655,-0.3717603087,0.4085085094,0.2649833262,0.234719947,-0.0482175499,-0.4544080198,-0.0339655727,-0.1824530363,0.0902625918,0.0834134221,0.2831938863,-0.100549832,-0.3547981977,0.4368868768,-0.0167233329,-0.2030842155,-0.0688422397,0.3948056698,-0.3977599144,0.2816140652,0.300632149,0.8588835597,0.170946151,0.2171094716,0.189544186,-0.0330892913,0.7430707216,0.0384738967,0.295150727,-0.3186835945,-0.1728597432,0.0220728032,0.1678132713,-0.2842455506,0.1117521226,-0.0204857849,0.5322527885,-0.0484973378,-0.0407954827,-0.2019559592,0.0385865234,0.1554113477,-0.0493461266,-0.0613518618,0.1877093762,-0.2476736456,0.0227982625,-0.0917987972,-0.1111849025,-0.2335607857,0.0792325884,0.1505281329,-0.0150572518,0.1452060938,0.0287628844,0.0681326687,0.0846666396,-0.2142507881,-0.0143809915,0.4202456176,0.1051282585,-0.1228862032,-0.0165364202,0.2023112923,0.3643657565,-0.0171934124,0.0700298995,0.5418530703,0.1744167209,-0.3010494411,0.1652928889,0.1161206216,0.0518671833,-0.3705786467,-0.2109765857,-0.027045615,-0.1738397926,0.1271543056,0.1300356835,-0.0139370691,-0.1672594547,0.2107219845,-0.0289873593,-0.2216912955,0.158717379,0.3067651987,-0.102967605,-0.0033234989,-0.0494807251,0.0802812874,0.2584996223,0.1753513962,-0.0618236698,-0.1246364862,-0.1845169514,0.0350933,0.322817266,-0.1774902344,0.0141988955,-0.0346613862,-0.2759725451,0.037059661,0.366019994,0.1767404974,0.4073539078,-0.1362541467,-0.334679395,-0.2940627337,0.2053038925,0.1035849676,0.5411198735,-0.1021241993,0.003515742,-0.084688589,0.0362085477,-0.4692041874,-0.1788289994,-0.2643841505,0.0300266892,0.2223577648,0.1943110675,-0.0616183504,0.1931773871,0.1847544611,-0.1003611162,-0.0763816088,-0.2584423125,-0.2580343187,0.0372569151,-0.0724827573,0.2678268254,-0.1898682117,-0.2380986065,0.3634139895,-0.267234832,0.0849276334,0.0334907398,0.0160590187,0.6699377298,-0.2746170461,0.1906252205,-0.0161331464,0.2923656702,0.2016665041,-0.2945486903,0.2462823093,0.0245086774,-0.2802989781,-0.0878335536,-0.0912393853,0.1401887685,-0.1739381105,0.0800962076,-0.1157909706,-0.0404338539,0.2326632887,0.3198111951,0.1912134141,0.2833020091,-0.0796788186,-0.1135381535,0.0008900691,0.4026626647,-0.2881116271,-0.1653479487,-0.0160354637,-0.0607308894,0.1993918568,0.3537395,0.1719893813,0.1471735686,0.0368238315,-0.1169243082,0.1564733982,-0.1589807123,0.1727954596,0.2296063155,-0.0663468391,0.0153561197,0.5013181567,0.1384329498,0.198976621,0.3526783884,0.0669980347,0.2979585826,0.0530522615,0.0318398811,0.1548145711,0.2173383385,-0.0670629814,0.1344039142,-0.1077987999,0.1149797961,-0.0083511919,-0.1065097228,-0.3142518401,-0.4529090822,0.0083612427,0.2475609779,-0.1106022671,-0.2584577799,-0.3169406354,-0.3145022392,0.0295492876,-0.2020742297,0.0204117484,-0.1642256528,0.2425481826,0.0104015945,0.0582943186,-0.6072797179,-0.1158910766,0.2009637207,0.4825261235,-0.0470445231,-0.0451543815,0.254163444,-0.1045296937,0.2959363759,0.1769165844,-0.0101825064,0.1673233956,-0.0632340983,-0.0442407839,-0.1472975314,-0.2732391357,-0.0197547171,-0.2685239315,0.0655189082,-0.2591530085,0.4273434579,0.1967945099,-0.1959593296,0.2888393998,-0.2689484954,0.2470980734,-0.0967142135,0.3938010037,0.2482170463,-0.0834991708,-0.1637675166,-0.2074750215,-0.418908596,-0.0337539017,0.1735800207,0.119015947,0.3294913769,-0.230452776,0.1744326651,0.0467711017,0.1016180441,0.0656000227,-0.1012194529,-0.0420516282,0.1775422692,-0.596978426,0.0986245945,0.1457094103,-0.4883113801,-0.0039019678,0.0689069927,0.0076408484,-0.3794311285,0.2487165332,-0.1756595075,0.3450993896,-0.1183805987,-0.5937594175,-0.1662165225,0.1165785864,-0.0916212797,0.1675551534,-0.1217336282,-0.018831104,-0.2435672581,0.1032894179,0.0916004106,0.2222185582,0.1852480471,0.219761759,0.097519435,0.0069883256,0.3506559432,0.0557341762,0.0438976735,-0.0433487818,-0.0668633953,-0.2903592885,0.0402562171,0.0220586844,0.4526719451,-0.0101347705,0.1919436902,-0.1298294216,0.0125701958,0.1213474274,-0.2020455748,-0.0882043615,0.2974076867,-0.0348802432,-0.2018259019,0.2351933271,0.3749948144,0.1204511151,0.0227135085,0.0916107818,-0.3557882607,0.4579107165,-0.0742051601,-0.2912693918,0.011748354,0.2051733732,0.2138843238,-0.0128113236,-0.4568787515,0.0492446497,0.3367868364,-0.1434984356,-0.4387976825,0.1279618144,0.1306618452,0.0891050622,-0.1594522595,0.004113914,0.1579317451,0.0543661229,-0.1419683546,-0.1255166531]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2206","title":"Got pyarrow error when loading a dataset while adding special tokens into the tokenizer","comments":"Hi,\r\n\r\nthe output of the tokenizers is treated specially in the lib to optimize the dataset size (see the code [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L138-L141)). It looks like that one of the values in a dictionary returned by the tokenizer is out of the assumed range.\r\nCan you please provide a minimal reproducible example for more help?","body":"I added five more special tokens into the GPT2 tokenizer. But after that, when I try to pre-process the data using my previous code, I got an error shown below:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1687, in _map_single\r\n writer.write(example)\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 296, in write\r\n self.write_on_file()\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 270, in write_on_file\r\n pa_array = pa.array(typed_sequence)\r\n File \"pyarrow\/array.pxi\", line 222, in pyarrow.lib.array\r\n File \"pyarrow\/array.pxi\", line 110, in pyarrow.lib._handle_arrow_array_protocol\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 108, in __arrow_array__\r\n out = out.cast(pa.list_(self.optimized_int_type))\r\n File \"pyarrow\/array.pxi\", line 810, in pyarrow.lib.Array.cast\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/pyarrow\/compute.py\", line 281, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 465, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 294, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Integer value 50259 not in range: -128 to 127\r\n\r\nDo you have any idea about it?","comment_length":53,"text":"Got pyarrow error when loading a dataset while adding special tokens into the tokenizer\nI added five more special tokens into the GPT2 tokenizer. But after that, when I try to pre-process the data using my previous code, I got an error shown below:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1687, in _map_single\r\n writer.write(example)\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 296, in write\r\n self.write_on_file()\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 270, in write_on_file\r\n pa_array = pa.array(typed_sequence)\r\n File \"pyarrow\/array.pxi\", line 222, in pyarrow.lib.array\r\n File \"pyarrow\/array.pxi\", line 110, in pyarrow.lib._handle_arrow_array_protocol\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 108, in __arrow_array__\r\n out = out.cast(pa.list_(self.optimized_int_type))\r\n File \"pyarrow\/array.pxi\", line 810, in pyarrow.lib.Array.cast\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/pyarrow\/compute.py\", line 281, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 465, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 294, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Integer value 50259 not in range: -128 to 127\r\n\r\nDo you have any idea about it?\nHi,\r\n\r\nthe output of the tokenizers is treated specially in the lib to optimize the dataset size (see the code [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L138-L141)). It looks like that one of the values in a dictionary returned by the tokenizer is out of the assumed range.\r\nCan you please provide a minimal reproducible example for more help?","embeddings":[-0.2338283807,0.2921876907,-0.0588289574,0.0690962002,0.2734426856,-0.0860713869,0.18364425,0.3055635691,-0.5368508101,-0.1932197064,-0.0461710617,0.435470134,-0.0010597197,-0.2381744683,0.2536684275,-0.0995298922,0.0430259258,0.2507618368,0.2605471909,0.1317316443,-0.0321432874,0.1118264794,-0.1489618123,0.5977945924,-0.3350471854,-0.1270231009,0.0943686217,0.0601615049,-0.0006362203,-0.7611901164,0.0361957289,0.0767358392,0.257945925,0.1267624646,-0.000113376,0.082067512,0.3061074615,0.0346666314,-0.0352177434,0.0471045561,0.1980308145,-0.0341250151,0.2838692069,-0.2447410524,0.0418412797,-1.0216341019,-0.092064172,0.1748209447,0.4970366061,0.2085008621,0.2248863131,0.2016527355,0.420830071,0.0445980728,0.4089097083,0.0463489853,0.1187575832,0.2101746053,0.0648392066,-0.0744946226,-0.0082713105,-0.1188545451,-0.3061366975,0.2222594768,0.0819868073,0.0422179736,-0.0128988633,-0.0767085329,-0.0784930363,-0.0311290734,0.4616255462,-0.4693733454,-0.1958956271,-0.0923196748,0.0183173362,-0.416379869,0.1107923314,0.1142617688,-0.2171587646,0.0793027058,0.1194188371,-0.0217523016,-0.2067582905,0.1848450452,-0.0662417859,0.3541361094,0.030876739,0.2433264703,-0.0098536797,0.0171071067,0.2675454915,0.1362328231,-0.20636563,0.3336641192,-0.3007293344,0.0695281476,-0.0275461059,-0.2159134448,0.130602181,0.174859941,0.1709089279,0.104464978,0.1628208607,0.0762209892,-0.1299583167,0.1835837215,-0.162752375,0.2325523049,0.1585679799,-0.3149089813,0.0370789021,-0.1459881067,-0.3320124447,-0.2384871691,0.3329922855,0.1477688849,0.338866353,-0.0093660438,-0.1467920095,0.0947749093,-0.4526158869,-0.1450776309,0.0061007231,0.2455308884,0.2182272822,-0.227198109,-0.1537809521,0.2055501193,-0.1003332511,-0.2328752279,-0.1099831611,-0.0415037572,-0.0182201341,0.1874153167,0.2010739595,0.0286410004,0.2526602447,0.191785872,-0.0618622452,0.0152938822,0.3261609972,-0.1774689406,0.00584799,0.353584677,0.0630945936,0.3427803516,0.0158788543,-0.0727052465,-0.035782557,0.2693840265,-0.3565315902,-0.4401058257,-0.2793906927,0.1600381881,-0.1012900174,0.0013683235,-0.2858616412,-0.0320901535,0.4216962755,-0.3251367807,-0.1027349234,-0.3298664689,0.0337792262,-0.3652558625,0.3561752439,-0.0902508795,-0.56661731,0.1339240223,-0.0229863543,-0.1644707471,0.1516373754,0.3593240678,-0.3796893954,0.6116597056,-0.3605375588,0.6695796251,0.2312913537,-0.0378579535,-0.2667081356,-0.1009031981,-0.1092613116,0.0314189866,-0.1686679721,-0.1111759916,0.1504587829,-0.0762138218,-0.1399538964,0.0632124171,0.0288358945,0.0341261104,-0.1693430841,-0.0971364975,0.2992917001,0.1167512313,0.0325557068,-0.131849736,0.0156451128,-0.2472018301,0.0563441142,-0.3048322201,0.135184437,0.0759484395,0.2232836485,-0.0552587323,0.089476116,-0.1198254526,-0.2498248518,0.0050441306,-0.5045512915,-0.0529679134,-0.3933269978,-0.0097885858,-0.2112289518,0.0880161747,-0.1989298612,0.1071834713,0.1984060556,0.1060441881,0.0044009401,0.1505635381,0.0360580832,-0.1274581701,0.1176428646,0.1194809228,0.0944710746,0.2554818988,-0.2053661793,-0.614557445,0.12766774,-0.0362825505,0.2126549035,-0.0232843962,-0.0731346458,0.2727456093,-0.1284492314,0.0718912333,-0.2981536388,0.0743246377,0.0266837943,-0.4504473209,-0.1293709427,-0.0512205735,0.1151850969,-0.0185076017,0.4564298093,0.3656798005,0.2342091799,-0.0699911118,-0.1562936753,-0.2149705142,0.2021137327,0.032652162,0.2847647667,0.1573234648,0.0747532099,0.3221545815,0.1628095061,-0.0037151151,-0.0672149807,0.1072628424,0.3034688532,-0.1084505245,0.0763668269,0.0472967774,-0.1052669585,-0.1009280533,0.0544268526,-0.4296977818,0.2058745176,0.1628007293,-0.3397504389,0.0139675438,-0.073378697,-0.1370283961,0.1553959846,0.0631728098,0.3842323422,0.3814519346,0.2945330143,0.0729723871,-0.1637671739,0.017035896,0.2265655994,0.3600274324,-0.4554604888,-0.0783177093,-0.1630183607,-0.050847929,0.1702332795,-0.3216295242,-0.0486299023,-0.1523302495,0.0268299188,0.086485222,-0.0517295189,0.2023847401,0.0673771575,0.0459718928,0.1958207637,-0.0231608637,-0.3977649808,-0.2804777026,-0.2186540812,0.0824890435,0.3015551269,0.0403780341,0.2039325982,0.237473011,0.0278314594,-0.1771772653,-0.3392232656,0.0429519191,-0.0736969858,0.2346140146,0.0524714328,-0.0311724246,-0.080389455,-0.4751782417,0.2225087285,-0.236471504,-0.2954795361,0.1646377295,-0.1729840785,-0.0004738634,-0.0085444367,-0.0560015626,-0.2244289666,-0.1134307832,0.128232941,0.0839182213,0.1650997847,-0.2448323816,-0.0034249236,0.173871085,-0.203290984,-0.1970703602,-0.1134401187,0.045367118,0.351980567,-0.1946189553,-0.2235365659,-0.1290646791,-0.0890230834,0.3335036635,0.1064963043,-0.2684744,0.136640355,-0.0324387327,0.4266342223,-0.3216074407,-0.3929841518,0.3826549947,0.0959021151,-0.1160591617,0.0203609541,-0.00924493,0.0358997472,-0.1952968538,-0.1222014427,-0.0495722108,0.3201514781,-0.1264334768,0.8966103196,-0.2778320611,-0.1123260111,0.4765632749,-0.0585326925,-0.0492790081,-0.1197143942,-0.1173318997,-0.1053839251,-0.1219811961,0.1585762054,0.0415137038,-0.3985213339,-0.3152047396,0.1176416501,-0.0033431409,-0.4849431515,-0.1623238623,0.0248305611,-0.0060156621,0.3399605453,0.0808941498,-0.1960046142,-0.2349725515,-0.1044546589,0.1002516001,-0.0964429155,0.1460984498,-0.0823539123,-0.6153333187,-0.0263252631,-0.3447532356,0.3508977592,0.2303440422,0.2787733972,-0.0252001882,0.0546588935,0.155444786,-0.2645351589,0.3481028676,0.0460813679,-0.0793132409,0.3778907359,0.0673645288,-0.4065200388,-0.2591381967,-0.1735210121,0.2162076682,0.173527807,0.208809793,-0.3339239359,-0.1532545686,0.1459283829,0.0317485854,-0.0758192018,-0.1607360393,-0.2423786819,-0.3862481415,-0.275254488,0.0777349249,0.0905977935,0.2722552717,-0.1924265176,-0.0251263361,-0.1741384417,-0.1011114791,-0.1357075274,-0.0230219364,0.2261856645,-0.0508405715,0.0109645054,0.0407166444,-0.0713443011,0.5171343088,-0.0105129406,0.6572525501,-0.3270750046,0.0146129578,0.0358019173,0.3106359839,-0.103694804,-0.2722659409,0.4553168416,-0.2256218642,0.0886269733,0.0143841766,0.2432913929,0.3678915799,-0.2389955521,-0.3546925485,-0.0285969768,0.2601580024,0.068147175,-0.1354558915,0.3313287199,-0.1181410179,-0.3191924691,0.3394183517,0.2960068882,0.7895694375,-0.1261397749,0.1547693461,0.5485855341,-0.2694839835,0.3345366418,-0.1798016578,0.0359759852,-0.4415816665,-0.0063238158,0.0751746818,-0.3206635714,0.0590695068,0.134945944,-0.1340192407,-0.1622877121,-0.1321251094,0.3438943326,0.3400494456,0.3328998387,0.1474967152,-0.0178198032,0.0948307887,0.1308038235,-0.4046937525,-0.1341082156,-0.2259301543,-0.1723212004,-0.0335794762,-0.1221877038,-0.1576716453,0.0173198264,-0.4561391175,0.3478041291,0.084916614,-0.3162203133,0.1610312462,-0.0474384762,0.1019051149,0.1037969887,-0.171671927,0.1528247148,-0.0826575905,-0.1429954618,-0.2142309099,-0.2495845109,0.0375227518,-0.0866126716,-0.2825962007,0.4322399795,-0.1139802113,-0.2102710754,-0.1728677452,0.0125369374,0.3119156063,-0.2836611867,0.2026781142,-0.2300821245,0.0316063166,-0.1942098588,0.121195063,0.0395354256,-0.03126809,0.0140519878,-0.2241231501,-0.053376846,-0.0647389516,0.1019981131,0.1302523315,0.071900323,0.8524980545,-0.0117236096,-0.016146237,-0.1319328547,-0.0278517641,0.1447279006,-0.6926686764,0.2088905275,0.1893457174,0.174597308,-0.03625229,0.2144402564,-0.0601530075,0.1539870352,0.1137562767,-0.2568929195,-0.3925020993,0.1009477973,-0.0155634563,0.1732783318,0.2280417383,0.4849222302,-0.2576970458,0.244118616,-0.2813732922,0.2138562202,-0.1269195676,0.1881315708,-0.3539942503,-0.2457783073,-0.241872862,0.1209030896,0.1827508658,0.2270849347,-0.088000685,-0.2621850967,-0.0628427193,0.1547938138,0.1043961123,-0.4758487344,0.138458997,-0.077649042,0.0856483877,0.0443274826,0.141410917,0.0083140191,-0.0238952506,-0.0294425413,0.6121737957,0.0333852693,-0.1780305803,0.1745882183,-0.3252797425,-0.0355582461,0.0940314233,0.42152071,-0.0784160569,0.040830642,0.07821922,0.0332826301,0.265899092,-0.1068007648,0.1057710126,-0.5149422884,-0.3337253034,-0.2798563242,0.2233311981,0.0947135165,-0.3138941526,-0.2480107546,0.2871521115,0.2126106173,-0.2153721452,-0.0119206766,0.1517178267,-0.053314317,-0.249914661,-0.0180371497,0.3401537836,-0.1459798664,0.0140544623,0.0499458984,0.524061203,0.1642270982,-0.14197056,0.067836225,-0.1499078274,-0.1343789846,0.3621260226,0.0754969418,0.0589414909,0.0791232809,-0.2144823223,0.1044686735,0.2081271261,0.1839293092,0.0116759604,-0.4132245183,-0.2656545937,0.2523524463,0.0547421128,-0.0737463683,-0.0049849255,0.3377439976,0.6447910666,-0.1422982663,-0.0968451202,0.5756676197,-0.2329218984,0.0020321843,-0.3756772876,-0.146589905,0.0745915547,-0.0359977074,-0.1046378836,-0.1838692874,-0.0185516682,0.224923104,-0.3663147688,-0.3276275694,0.5081475377,0.2541369796,-0.0847616643,-0.1901150644,0.3602640033,0.1809311509,-0.2443731874,0.0219395347,0.4166369438,0.6981303096,0.3844665289,-0.262109369,0.0661902353,-0.1517230421,0.1134638041,0.1007219851,0.2648787796,0.3564858437,0.199668929,0.085015811,0.1405355036,-0.157891497,-0.2345954776,0.3939746916,0.0582140014,-0.0315158926,-0.0618402436,-0.2130903453,-0.4125773907,0.0771784261,0.2182449847,-0.0915769264,0.2303873599,0.3385384679,0.0952060446,-0.0094057582,-0.0438190065,0.0980002508,-0.1631917804,0.5890395641,0.3657805324,0.2943979204,-0.4726662636,-0.4595135152,-0.4318799078,0.4607594013,-0.1200262234,0.1326164156,0.2335555851,0.1743251979,0.0586354211,0.0616109781,0.1084995121,0.4466968477,-0.2370323986,0.2657864988,-0.3800933361,0.4180729091,0.0149965854,-0.0360371694,-0.1832068413,-0.2183639705,0.3100320995,-0.2936545014,0.1718758345,-0.1804483533,-0.1550525129,0.2990249395,0.1778461337,0.5840939283,0.0202660132,0.3855083883,0.0433578752,0.3265528977,-0.0895396397,-0.3381975889,-0.1921401769,0.3090176284,-0.0329326577,0.1928832084,0.1407828629,0.0347659402,-0.3030672669,0.0247733202,-0.2038728297,0.1529365629,-0.1174189821,0.1256079674,-0.1389215142,-0.0102254199,-0.0201867335,-0.107367985,0.1551196277,0.3255605996,-0.5399785638,-0.3889168501,0.5106248856,-0.3500748575,-0.1078936905,0.0033394035,0.1297105253,0.1302658916,-0.0042215339,-0.142142728,-0.058534693,0.7115308642,-0.0309066363,-0.2054666877,0.300884217,0.0833453834,0.064668268,-0.0842685923,0.2101774365,0.0909491405,-0.1234653071,-0.237041086,-0.2266750038]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2206","title":"Got pyarrow error when loading a dataset while adding special tokens into the tokenizer","comments":"Hi @yana-xuyan, thanks for reporting.\r\n\r\nAs clearly @mariosasko explained, `datasets` performs some optimizations in order to reduce the size of the dataset cache files. And one of them is storing the field `special_tokens_mask` as `int8`, which means that this field can only contain integers between `-128` to `127`. As your message error states, one of the values of this field is `50259`, and therefore it cannot be stored as an `int8`.\r\n\r\nMaybe we could implement a way to disable this optimization and allow using any integer value; although the size of the cache files would be much larger.","body":"I added five more special tokens into the GPT2 tokenizer. But after that, when I try to pre-process the data using my previous code, I got an error shown below:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1687, in _map_single\r\n writer.write(example)\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 296, in write\r\n self.write_on_file()\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 270, in write_on_file\r\n pa_array = pa.array(typed_sequence)\r\n File \"pyarrow\/array.pxi\", line 222, in pyarrow.lib.array\r\n File \"pyarrow\/array.pxi\", line 110, in pyarrow.lib._handle_arrow_array_protocol\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 108, in __arrow_array__\r\n out = out.cast(pa.list_(self.optimized_int_type))\r\n File \"pyarrow\/array.pxi\", line 810, in pyarrow.lib.Array.cast\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/pyarrow\/compute.py\", line 281, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 465, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 294, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Integer value 50259 not in range: -128 to 127\r\n\r\nDo you have any idea about it?","comment_length":98,"text":"Got pyarrow error when loading a dataset while adding special tokens into the tokenizer\nI added five more special tokens into the GPT2 tokenizer. But after that, when I try to pre-process the data using my previous code, I got an error shown below:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1687, in _map_single\r\n writer.write(example)\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 296, in write\r\n self.write_on_file()\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 270, in write_on_file\r\n pa_array = pa.array(typed_sequence)\r\n File \"pyarrow\/array.pxi\", line 222, in pyarrow.lib.array\r\n File \"pyarrow\/array.pxi\", line 110, in pyarrow.lib._handle_arrow_array_protocol\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 108, in __arrow_array__\r\n out = out.cast(pa.list_(self.optimized_int_type))\r\n File \"pyarrow\/array.pxi\", line 810, in pyarrow.lib.Array.cast\r\n File \"\/home\/xuyan\/anaconda3\/envs\/convqa\/lib\/python3.7\/site-packages\/pyarrow\/compute.py\", line 281, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 465, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 294, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Integer value 50259 not in range: -128 to 127\r\n\r\nDo you have any idea about it?\nHi @yana-xuyan, thanks for reporting.\r\n\r\nAs clearly @mariosasko explained, `datasets` performs some optimizations in order to reduce the size of the dataset cache files. And one of them is storing the field `special_tokens_mask` as `int8`, which means that this field can only contain integers between `-128` to `127`. As your message error states, one of the values of this field is `50259`, and therefore it cannot be stored as an `int8`.\r\n\r\nMaybe we could implement a way to disable this optimization and allow using any integer value; although the size of the cache files would be much larger.","embeddings":[-0.2338283807,0.2921876907,-0.0588289574,0.0690962002,0.2734426856,-0.0860713869,0.18364425,0.3055635691,-0.5368508101,-0.1932197064,-0.0461710617,0.435470134,-0.0010597197,-0.2381744683,0.2536684275,-0.0995298922,0.0430259258,0.2507618368,0.2605471909,0.1317316443,-0.0321432874,0.1118264794,-0.1489618123,0.5977945924,-0.3350471854,-0.1270231009,0.0943686217,0.0601615049,-0.0006362203,-0.7611901164,0.0361957289,0.0767358392,0.257945925,0.1267624646,-0.000113376,0.082067512,0.3061074615,0.0346666314,-0.0352177434,0.0471045561,0.1980308145,-0.0341250151,0.2838692069,-0.2447410524,0.0418412797,-1.0216341019,-0.092064172,0.1748209447,0.4970366061,0.2085008621,0.2248863131,0.2016527355,0.420830071,0.0445980728,0.4089097083,0.0463489853,0.1187575832,0.2101746053,0.0648392066,-0.0744946226,-0.0082713105,-0.1188545451,-0.3061366975,0.2222594768,0.0819868073,0.0422179736,-0.0128988633,-0.0767085329,-0.0784930363,-0.0311290734,0.4616255462,-0.4693733454,-0.1958956271,-0.0923196748,0.0183173362,-0.416379869,0.1107923314,0.1142617688,-0.2171587646,0.0793027058,0.1194188371,-0.0217523016,-0.2067582905,0.1848450452,-0.0662417859,0.3541361094,0.030876739,0.2433264703,-0.0098536797,0.0171071067,0.2675454915,0.1362328231,-0.20636563,0.3336641192,-0.3007293344,0.0695281476,-0.0275461059,-0.2159134448,0.130602181,0.174859941,0.1709089279,0.104464978,0.1628208607,0.0762209892,-0.1299583167,0.1835837215,-0.162752375,0.2325523049,0.1585679799,-0.3149089813,0.0370789021,-0.1459881067,-0.3320124447,-0.2384871691,0.3329922855,0.1477688849,0.338866353,-0.0093660438,-0.1467920095,0.0947749093,-0.4526158869,-0.1450776309,0.0061007231,0.2455308884,0.2182272822,-0.227198109,-0.1537809521,0.2055501193,-0.1003332511,-0.2328752279,-0.1099831611,-0.0415037572,-0.0182201341,0.1874153167,0.2010739595,0.0286410004,0.2526602447,0.191785872,-0.0618622452,0.0152938822,0.3261609972,-0.1774689406,0.00584799,0.353584677,0.0630945936,0.3427803516,0.0158788543,-0.0727052465,-0.035782557,0.2693840265,-0.3565315902,-0.4401058257,-0.2793906927,0.1600381881,-0.1012900174,0.0013683235,-0.2858616412,-0.0320901535,0.4216962755,-0.3251367807,-0.1027349234,-0.3298664689,0.0337792262,-0.3652558625,0.3561752439,-0.0902508795,-0.56661731,0.1339240223,-0.0229863543,-0.1644707471,0.1516373754,0.3593240678,-0.3796893954,0.6116597056,-0.3605375588,0.6695796251,0.2312913537,-0.0378579535,-0.2667081356,-0.1009031981,-0.1092613116,0.0314189866,-0.1686679721,-0.1111759916,0.1504587829,-0.0762138218,-0.1399538964,0.0632124171,0.0288358945,0.0341261104,-0.1693430841,-0.0971364975,0.2992917001,0.1167512313,0.0325557068,-0.131849736,0.0156451128,-0.2472018301,0.0563441142,-0.3048322201,0.135184437,0.0759484395,0.2232836485,-0.0552587323,0.089476116,-0.1198254526,-0.2498248518,0.0050441306,-0.5045512915,-0.0529679134,-0.3933269978,-0.0097885858,-0.2112289518,0.0880161747,-0.1989298612,0.1071834713,0.1984060556,0.1060441881,0.0044009401,0.1505635381,0.0360580832,-0.1274581701,0.1176428646,0.1194809228,0.0944710746,0.2554818988,-0.2053661793,-0.614557445,0.12766774,-0.0362825505,0.2126549035,-0.0232843962,-0.0731346458,0.2727456093,-0.1284492314,0.0718912333,-0.2981536388,0.0743246377,0.0266837943,-0.4504473209,-0.1293709427,-0.0512205735,0.1151850969,-0.0185076017,0.4564298093,0.3656798005,0.2342091799,-0.0699911118,-0.1562936753,-0.2149705142,0.2021137327,0.032652162,0.2847647667,0.1573234648,0.0747532099,0.3221545815,0.1628095061,-0.0037151151,-0.0672149807,0.1072628424,0.3034688532,-0.1084505245,0.0763668269,0.0472967774,-0.1052669585,-0.1009280533,0.0544268526,-0.4296977818,0.2058745176,0.1628007293,-0.3397504389,0.0139675438,-0.073378697,-0.1370283961,0.1553959846,0.0631728098,0.3842323422,0.3814519346,0.2945330143,0.0729723871,-0.1637671739,0.017035896,0.2265655994,0.3600274324,-0.4554604888,-0.0783177093,-0.1630183607,-0.050847929,0.1702332795,-0.3216295242,-0.0486299023,-0.1523302495,0.0268299188,0.086485222,-0.0517295189,0.2023847401,0.0673771575,0.0459718928,0.1958207637,-0.0231608637,-0.3977649808,-0.2804777026,-0.2186540812,0.0824890435,0.3015551269,0.0403780341,0.2039325982,0.237473011,0.0278314594,-0.1771772653,-0.3392232656,0.0429519191,-0.0736969858,0.2346140146,0.0524714328,-0.0311724246,-0.080389455,-0.4751782417,0.2225087285,-0.236471504,-0.2954795361,0.1646377295,-0.1729840785,-0.0004738634,-0.0085444367,-0.0560015626,-0.2244289666,-0.1134307832,0.128232941,0.0839182213,0.1650997847,-0.2448323816,-0.0034249236,0.173871085,-0.203290984,-0.1970703602,-0.1134401187,0.045367118,0.351980567,-0.1946189553,-0.2235365659,-0.1290646791,-0.0890230834,0.3335036635,0.1064963043,-0.2684744,0.136640355,-0.0324387327,0.4266342223,-0.3216074407,-0.3929841518,0.3826549947,0.0959021151,-0.1160591617,0.0203609541,-0.00924493,0.0358997472,-0.1952968538,-0.1222014427,-0.0495722108,0.3201514781,-0.1264334768,0.8966103196,-0.2778320611,-0.1123260111,0.4765632749,-0.0585326925,-0.0492790081,-0.1197143942,-0.1173318997,-0.1053839251,-0.1219811961,0.1585762054,0.0415137038,-0.3985213339,-0.3152047396,0.1176416501,-0.0033431409,-0.4849431515,-0.1623238623,0.0248305611,-0.0060156621,0.3399605453,0.0808941498,-0.1960046142,-0.2349725515,-0.1044546589,0.1002516001,-0.0964429155,0.1460984498,-0.0823539123,-0.6153333187,-0.0263252631,-0.3447532356,0.3508977592,0.2303440422,0.2787733972,-0.0252001882,0.0546588935,0.155444786,-0.2645351589,0.3481028676,0.0460813679,-0.0793132409,0.3778907359,0.0673645288,-0.4065200388,-0.2591381967,-0.1735210121,0.2162076682,0.173527807,0.208809793,-0.3339239359,-0.1532545686,0.1459283829,0.0317485854,-0.0758192018,-0.1607360393,-0.2423786819,-0.3862481415,-0.275254488,0.0777349249,0.0905977935,0.2722552717,-0.1924265176,-0.0251263361,-0.1741384417,-0.1011114791,-0.1357075274,-0.0230219364,0.2261856645,-0.0508405715,0.0109645054,0.0407166444,-0.0713443011,0.5171343088,-0.0105129406,0.6572525501,-0.3270750046,0.0146129578,0.0358019173,0.3106359839,-0.103694804,-0.2722659409,0.4553168416,-0.2256218642,0.0886269733,0.0143841766,0.2432913929,0.3678915799,-0.2389955521,-0.3546925485,-0.0285969768,0.2601580024,0.068147175,-0.1354558915,0.3313287199,-0.1181410179,-0.3191924691,0.3394183517,0.2960068882,0.7895694375,-0.1261397749,0.1547693461,0.5485855341,-0.2694839835,0.3345366418,-0.1798016578,0.0359759852,-0.4415816665,-0.0063238158,0.0751746818,-0.3206635714,0.0590695068,0.134945944,-0.1340192407,-0.1622877121,-0.1321251094,0.3438943326,0.3400494456,0.3328998387,0.1474967152,-0.0178198032,0.0948307887,0.1308038235,-0.4046937525,-0.1341082156,-0.2259301543,-0.1723212004,-0.0335794762,-0.1221877038,-0.1576716453,0.0173198264,-0.4561391175,0.3478041291,0.084916614,-0.3162203133,0.1610312462,-0.0474384762,0.1019051149,0.1037969887,-0.171671927,0.1528247148,-0.0826575905,-0.1429954618,-0.2142309099,-0.2495845109,0.0375227518,-0.0866126716,-0.2825962007,0.4322399795,-0.1139802113,-0.2102710754,-0.1728677452,0.0125369374,0.3119156063,-0.2836611867,0.2026781142,-0.2300821245,0.0316063166,-0.1942098588,0.121195063,0.0395354256,-0.03126809,0.0140519878,-0.2241231501,-0.053376846,-0.0647389516,0.1019981131,0.1302523315,0.071900323,0.8524980545,-0.0117236096,-0.016146237,-0.1319328547,-0.0278517641,0.1447279006,-0.6926686764,0.2088905275,0.1893457174,0.174597308,-0.03625229,0.2144402564,-0.0601530075,0.1539870352,0.1137562767,-0.2568929195,-0.3925020993,0.1009477973,-0.0155634563,0.1732783318,0.2280417383,0.4849222302,-0.2576970458,0.244118616,-0.2813732922,0.2138562202,-0.1269195676,0.1881315708,-0.3539942503,-0.2457783073,-0.241872862,0.1209030896,0.1827508658,0.2270849347,-0.088000685,-0.2621850967,-0.0628427193,0.1547938138,0.1043961123,-0.4758487344,0.138458997,-0.077649042,0.0856483877,0.0443274826,0.141410917,0.0083140191,-0.0238952506,-0.0294425413,0.6121737957,0.0333852693,-0.1780305803,0.1745882183,-0.3252797425,-0.0355582461,0.0940314233,0.42152071,-0.0784160569,0.040830642,0.07821922,0.0332826301,0.265899092,-0.1068007648,0.1057710126,-0.5149422884,-0.3337253034,-0.2798563242,0.2233311981,0.0947135165,-0.3138941526,-0.2480107546,0.2871521115,0.2126106173,-0.2153721452,-0.0119206766,0.1517178267,-0.053314317,-0.249914661,-0.0180371497,0.3401537836,-0.1459798664,0.0140544623,0.0499458984,0.524061203,0.1642270982,-0.14197056,0.067836225,-0.1499078274,-0.1343789846,0.3621260226,0.0754969418,0.0589414909,0.0791232809,-0.2144823223,0.1044686735,0.2081271261,0.1839293092,0.0116759604,-0.4132245183,-0.2656545937,0.2523524463,0.0547421128,-0.0737463683,-0.0049849255,0.3377439976,0.6447910666,-0.1422982663,-0.0968451202,0.5756676197,-0.2329218984,0.0020321843,-0.3756772876,-0.146589905,0.0745915547,-0.0359977074,-0.1046378836,-0.1838692874,-0.0185516682,0.224923104,-0.3663147688,-0.3276275694,0.5081475377,0.2541369796,-0.0847616643,-0.1901150644,0.3602640033,0.1809311509,-0.2443731874,0.0219395347,0.4166369438,0.6981303096,0.3844665289,-0.262109369,0.0661902353,-0.1517230421,0.1134638041,0.1007219851,0.2648787796,0.3564858437,0.199668929,0.085015811,0.1405355036,-0.157891497,-0.2345954776,0.3939746916,0.0582140014,-0.0315158926,-0.0618402436,-0.2130903453,-0.4125773907,0.0771784261,0.2182449847,-0.0915769264,0.2303873599,0.3385384679,0.0952060446,-0.0094057582,-0.0438190065,0.0980002508,-0.1631917804,0.5890395641,0.3657805324,0.2943979204,-0.4726662636,-0.4595135152,-0.4318799078,0.4607594013,-0.1200262234,0.1326164156,0.2335555851,0.1743251979,0.0586354211,0.0616109781,0.1084995121,0.4466968477,-0.2370323986,0.2657864988,-0.3800933361,0.4180729091,0.0149965854,-0.0360371694,-0.1832068413,-0.2183639705,0.3100320995,-0.2936545014,0.1718758345,-0.1804483533,-0.1550525129,0.2990249395,0.1778461337,0.5840939283,0.0202660132,0.3855083883,0.0433578752,0.3265528977,-0.0895396397,-0.3381975889,-0.1921401769,0.3090176284,-0.0329326577,0.1928832084,0.1407828629,0.0347659402,-0.3030672669,0.0247733202,-0.2038728297,0.1529365629,-0.1174189821,0.1256079674,-0.1389215142,-0.0102254199,-0.0201867335,-0.107367985,0.1551196277,0.3255605996,-0.5399785638,-0.3889168501,0.5106248856,-0.3500748575,-0.1078936905,0.0033394035,0.1297105253,0.1302658916,-0.0042215339,-0.142142728,-0.058534693,0.7115308642,-0.0309066363,-0.2054666877,0.300884217,0.0833453834,0.064668268,-0.0842685923,0.2101774365,0.0909491405,-0.1234653071,-0.237041086,-0.2266750038]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2200","title":"_prepare_split will overwrite DatasetBuilder.info.features","comments":"Hi ! This might be related to #2153 \r\n\r\nYou're right the ArrowWriter should be initialized with `features=self.info.features` ! Good catch\r\nI'm opening a PR to fix this and also to figure out how it was not caught in the tests\r\n\r\nEDIT: opened #2201","body":"Hi, here is my issue:\r\nI initialized a Csv datasetbuilder with specific features:\r\n```\r\ndef get_dataset_features(data_args):\r\n features = {}\r\n if data_args.text_features:\r\n features.update({text_feature: hf_features.Value(\"string\") for text_feature in data_args.text_features.strip().split(\",\")})\r\n if data_args.num_features:\r\n features.update({text_feature: hf_features.Value(\"float32\") for text_feature in data_args.num_features.strip().split(\",\")})\r\n if data_args.label_classes:\r\n features[\"label\"] = hf_features.ClassLabel(names=data_args.label_classes.strip().split(\",\"))\r\n else:\r\n features[\"label\"] = hf_features.Value(\"float32\")\r\n return hf_features.Features(features)\r\n\r\ndatasets = load_dataset(extension,\r\n data_files=data_files,\r\n sep=data_args.delimiter,\r\n header=data_args.header,\r\n column_names=data_args.column_names.split(\",\") if data_args.column_names else None,\r\n features=get_dataset_features(data_args=data_args))\r\n```\r\nThe `features` is printout as below before `builder_instance.as_dataset` is called:\r\n```\r\n{'label': ClassLabel(num_classes=2, names=['unacceptable', 'acceptable'], names_file=None, id=None), 'notated': Value(dtype='string', id=None), 'sentence': Value(dtype='string', id=None), 'src_code': Value(dtype='string', id=None)}\r\n````\r\n\r\nBut after the `builder_instance.as_dataset` is called for Csv dataset builder, the `features` is changed to:\r\n```\r\n{'label': Value(dtype='int64', id=None), 'notated': Value(dtype='string', id=None), 'sentence': Value(dtype='string', id=None), 'src_code': Value(dtype='string', id=None)}\r\n```\r\n\r\nAfter digged into the code, I releazed that in `ArrowBasedBuilder._prepare_split`, the DatasetBuilder's info's features will be overwrited by `ArrowWriter`'s `_features`. \r\nBut `ArrowWriter` is initailized without passing `features`.\r\nSo my concern is:\r\nIt's this overwrite must be done, or, should it be an option to pass features in `_prepare_split` function?","comment_length":43,"text":"_prepare_split will overwrite DatasetBuilder.info.features\nHi, here is my issue:\r\nI initialized a Csv datasetbuilder with specific features:\r\n```\r\ndef get_dataset_features(data_args):\r\n features = {}\r\n if data_args.text_features:\r\n features.update({text_feature: hf_features.Value(\"string\") for text_feature in data_args.text_features.strip().split(\",\")})\r\n if data_args.num_features:\r\n features.update({text_feature: hf_features.Value(\"float32\") for text_feature in data_args.num_features.strip().split(\",\")})\r\n if data_args.label_classes:\r\n features[\"label\"] = hf_features.ClassLabel(names=data_args.label_classes.strip().split(\",\"))\r\n else:\r\n features[\"label\"] = hf_features.Value(\"float32\")\r\n return hf_features.Features(features)\r\n\r\ndatasets = load_dataset(extension,\r\n data_files=data_files,\r\n sep=data_args.delimiter,\r\n header=data_args.header,\r\n column_names=data_args.column_names.split(\",\") if data_args.column_names else None,\r\n features=get_dataset_features(data_args=data_args))\r\n```\r\nThe `features` is printout as below before `builder_instance.as_dataset` is called:\r\n```\r\n{'label': ClassLabel(num_classes=2, names=['unacceptable', 'acceptable'], names_file=None, id=None), 'notated': Value(dtype='string', id=None), 'sentence': Value(dtype='string', id=None), 'src_code': Value(dtype='string', id=None)}\r\n````\r\n\r\nBut after the `builder_instance.as_dataset` is called for Csv dataset builder, the `features` is changed to:\r\n```\r\n{'label': Value(dtype='int64', id=None), 'notated': Value(dtype='string', id=None), 'sentence': Value(dtype='string', id=None), 'src_code': Value(dtype='string', id=None)}\r\n```\r\n\r\nAfter digged into the code, I releazed that in `ArrowBasedBuilder._prepare_split`, the DatasetBuilder's info's features will be overwrited by `ArrowWriter`'s `_features`. \r\nBut `ArrowWriter` is initailized without passing `features`.\r\nSo my concern is:\r\nIt's this overwrite must be done, or, should it be an option to pass features in `_prepare_split` function?\nHi ! This might be related to #2153 \r\n\r\nYou're right the ArrowWriter should be initialized with `features=self.info.features` ! Good catch\r\nI'm opening a PR to fix this and also to figure out how it was not caught in the tests\r\n\r\nEDIT: opened #2201","embeddings":[-0.2406021953,-0.0382910147,-0.1077674851,0.1783376485,0.2982976437,0.2197924554,0.4546083808,0.2158689797,-0.3327544034,0.1149650663,0.1291757524,0.1715127975,0.0863348842,0.4873531759,-0.1215008423,0.0767409503,0.0831202567,0.2866099775,0.1698089838,-0.0732159466,-0.2161129415,0.103940621,-0.260414362,0.0780965611,-0.4611605406,0.0068310783,0.0738492385,0.1830849648,0.1115719378,-0.4215717316,0.1278963238,0.1931540817,-0.0024506638,0.0135782873,-0.0001165283,0.0921400413,0.0531196035,-0.0490681157,-0.3561661243,0.1506268084,-0.3260130882,-0.2654621005,-0.0111547243,-0.2360556424,-0.1424901187,-0.3270657361,-0.3870328367,0.1431457102,0.21410124,0.2266549468,0.1773981899,-0.2564507723,0.1165271029,0.1099403724,0.3127687573,0.1972708553,-0.0730632916,-0.2695540786,-0.294752568,0.1149955839,-0.1780861318,0.4079653323,-0.1879745275,-0.0244833566,0.1067840531,0.0501348674,-0.0720179081,-0.1327121854,0.0906457528,-0.3222067356,0.2972629964,0.0031607789,-0.0745959058,-0.5507095456,-0.1756774783,-0.3726785183,0.127729401,-0.167284295,0.1768467575,0.1145290211,-0.106956929,-0.3455787599,0.1335982829,-0.0906985849,-0.0378318988,0.3073068261,-0.4004644454,0.3126069605,-0.1145926118,0.0700759366,0.295794636,-0.2810826898,-0.2957600653,-0.1158553436,-0.1413079649,-0.1271789819,0.1206175536,-0.1135420054,-0.0204882305,0.0391933806,-0.183370471,0.1696168929,-0.0626039505,-0.0579128414,0.2710881531,0.3043972552,0.2000663131,0.1332727075,-0.0148484511,0.3680326641,-0.3029093444,-0.1120791882,0.3196536303,-0.1899855882,0.3789294064,0.3041920364,0.5231932402,-0.191226095,-0.2840594351,0.2589049637,0.3041548729,-0.047680594,-0.0207292493,0.1026611105,0.0128632467,0.2322100401,-0.2137506455,0.0731487572,-0.302153945,-0.1779461801,-0.2684747875,-0.2020440698,-0.0143179186,0.1958020329,-0.0162585042,-0.028942639,0.4025155902,0.1338067055,0.059362866,-0.5640695095,-0.0786658973,-0.141408965,0.245393455,0.2372240126,0.0063599204,0.297984153,0.345356226,-0.1086787507,0.0423199162,0.0928914696,-0.4464731514,-0.3818405867,0.0246932805,0.204917416,-0.038765531,0.2318276614,0.0056333705,0.0674563348,0.3656390905,-0.0792343542,-0.0102581428,-0.184376806,-0.4282066822,-0.0858234316,0.0689308494,0.3378699124,-0.3634487987,0.0270513836,-0.0374605916,-0.0010612832,0.2343657315,0.0017081597,-0.0472922921,0.2207314819,-0.2923869193,-0.1538804322,0.4493572712,0.0422428735,-0.2183953077,0.4702954888,-0.0512736104,0.1960454583,0.097857289,-0.1573392004,0.312523365,0.0569648445,-0.0447400771,0.0004948836,-0.1355825812,-0.0460937694,-0.2612141669,-0.2650542259,-0.008415495,-0.2596246004,0.0484578907,0.2520490289,-0.149836421,-0.2065731287,0.2138381004,0.1278960705,0.0759673044,0.238937363,-0.0274179019,0.4694034457,0.1447722316,-0.1496132612,-0.2516558468,-0.0256625861,-0.006287029,-0.226612851,-0.1490036845,-0.5273110271,-0.4477546513,-0.0709879026,-0.376218468,-0.264528662,0.0848521516,0.1711636782,-0.0287180431,-0.1455246806,0.0106613887,0.0743661597,-0.1461340338,0.1596860439,-0.0771052167,0.0408394523,-0.0278835911,0.0059103705,-0.0508349724,0.3137435317,0.2422779202,-0.1615428478,-0.3486573398,0.5029213428,0.129590854,0.0253561251,-0.3426887095,-0.3701828122,0.1229716316,-0.0126654766,-0.0751680434,-0.0887555107,0.0540345162,-0.0973859057,-0.0829878077,0.4442037046,-0.1351927668,0.2727766335,0.0633149743,-0.0458262675,0.1429138184,-0.0051077781,0.2960257828,-0.4752125442,-0.4301364422,-0.1807939857,-0.1933782995,0.1361899674,-0.0568668954,0.4161864221,0.7362260222,0.0912750587,-0.2430435121,-0.0015031998,-0.1585697085,-0.0849274471,0.1719686985,-0.1211564466,0.3886586726,0.0281732567,0.0056694192,0.0048922412,-0.0323953554,0.0279556587,0.2369142026,0.195208624,0.1960088462,0.2906233966,0.0994656309,-0.0508444123,-0.3068084419,0.2607759833,0.1012084484,0.1346651912,-0.3549932539,0.3007321954,-0.3476017416,0.1741188616,-0.3198778927,-0.0015288031,-0.0686357468,-0.4444669485,0.0085163433,0.3251644373,-0.04318358,0.1285805553,-0.1287238449,-0.1116976589,0.0461462475,-0.2226932645,0.1801130623,-0.0527548119,-0.3112846911,0.0221238621,-0.0914603621,0.1510246545,0.0681759343,0.0183007941,-0.2734510303,-0.2057698667,-0.3139711022,0.1297929287,-0.2200113833,0.1257251501,0.1559238434,-0.2314231992,0.3351610899,-0.3764411211,0.1105968058,0.1745934039,-0.030335892,-0.0731975138,0.1376205236,0.0059762136,-0.1190199405,-0.5209413767,0.0441726148,-0.2698732316,-0.0336812027,0.2164280266,0.0426517986,0.0496085621,0.011808469,-0.1766927093,0.179286629,0.0049384478,-0.4112112522,-0.2042838186,-0.0725068152,-0.1624400169,0.2003741264,0.031419199,-0.1869898736,0.0107457153,-0.0141782928,-0.3892456591,0.0628277361,-0.2297626436,0.1830192059,-0.0313399471,-0.0654165223,0.0715479031,0.3412663043,-0.0353141055,0.0644659922,-0.2030811757,0.2194034606,-0.0548892543,0.0521089919,-0.1899810135,0.3345891535,-0.1968087107,0.7819837332,-0.0370066687,0.1000483185,0.0815292224,-0.0416024178,0.2504915297,-0.2591479421,-0.3987992406,0.1645832807,-0.1963440776,-0.0728487447,0.3796070814,-0.0990734175,0.147952497,0.2129001766,0.0369535461,-0.1447470188,-0.3877590001,-0.0382353552,-0.1339042634,0.4046775997,0.1702878326,0.1226609796,-0.2520200908,-0.0809987932,-0.1148270741,0.2434997261,0.2173370421,-0.2415080369,-0.4367235601,-0.0477512889,0.1094470844,0.1142058969,0.1452283263,0.1801118702,-0.0691871345,-0.2261043489,0.1469920278,0.0663821921,0.6020619869,-0.001593545,0.0893499106,0.2863385677,0.1414212584,0.4394608438,-0.385630995,-0.0263817832,0.6457603574,0.0466839336,0.3440138102,0.0570220351,-0.3574425578,0.7359841466,0.2237891555,-0.4519286752,0.1409161091,-0.1648079157,0.0155695621,-0.1326268613,0.0458727777,-0.0465164036,0.0182099994,-0.3915722966,-0.2023957968,-0.2415165305,-0.1870998442,-0.0574111529,0.0098595377,0.2120065838,-0.0419113189,0.0273001771,-0.0304028541,0.1045274809,0.1013760343,0.6346020699,-0.1218815744,-0.6454026103,0.0526319705,-0.2799099088,0.3864447474,0.2780576646,-0.197907716,0.3460095227,-0.0478184596,0.2940057516,-0.5682277679,0.2830225527,0.4415433705,0.09933386,-0.3798827827,-0.5284830332,0.3364790678,-0.1991058439,-0.1632905006,-0.1263380796,0.1568223983,-0.5815492272,0.502946794,-0.0480983034,0.552575767,0.2817796171,0.1456730217,-0.0596217476,0.4062951207,0.5035848022,-0.2715296149,0.2468828708,-0.3581906259,-0.4022682309,-0.0234562978,0.1249708086,0.26547575,0.1090989336,-0.2776915729,0.493137151,-0.364053756,0.2984279096,-0.1434563249,0.0993173942,0.1148136482,-0.1004862264,0.2887810171,0.0711408406,0.0594277829,0.0769228488,0.1941160709,-0.0005984192,0.3359731436,-0.0402314253,-0.0544736534,0.0335327834,-0.0611637719,0.150058493,0.013336163,-0.276491791,0.3465050757,-0.3510287702,0.3947290778,0.2172810286,0.2850671411,0.0914980397,0.3242627978,0.2868179977,-0.1777195185,0.2628845572,0.2786632478,0.3406605721,-0.1953748465,0.1444916278,-0.1367163658,0.0413951091,-0.232319966,-0.1036987081,0.2827704251,-0.8835759163,-0.094782792,0.0146639636,0.1324677169,-0.3969879448,0.0564999543,-0.3819342554,-0.2622256577,0.1641947329,-0.0680011436,-0.3061917126,-0.0870294347,-0.1133318171,0.3726708889,0.1711788028,0.6264430881,-0.0985798687,-0.157681331,-0.2450169921,-0.0059849536,-0.123159118,-0.2279707938,0.1567362398,0.0155930594,-0.0709494203,0.0796243772,0.1542987227,-0.1563897431,0.3370626569,-0.3017419875,-0.2852936387,-0.3449874222,-0.2147877216,-0.0338958241,0.0869262815,0.1602318138,0.185390383,-0.2084304392,0.4056222141,-0.2619144022,-0.0314986371,-0.0364254713,0.3956214488,0.0656920671,0.0869985148,-0.0181533862,-0.0262147076,0.121575959,-0.1006318554,0.268035084,-0.2361129671,-0.3126168251,0.1382649839,-0.1827713102,0.0544784032,-0.1590766162,0.0578636527,0.0218237881,-0.1035419852,0.3592787683,-0.141206041,0.3238607943,0.4083138108,-0.2750033438,0.0652092248,-0.0785179883,-0.0183321945,-0.12258178,0.3001945615,0.1198046282,0.2510439456,-0.1364616901,0.0107375849,0.0105193527,-0.0630402565,-0.2360347658,0.0126777636,0.2943050563,-0.1888019741,0.141473949,0.1635720581,0.2870686948,0.4440115094,0.1440840364,-0.2545003295,0.2238463461,0.1845123768,-0.1594916284,-0.179832682,0.2453946173,-0.2276021987,0.2211372554,0.4496669471,0.2195438743,0.2532426417,0.1180917397,0.2186754942,0.1186124831,-0.0602947362,0.4032922685,0.6268619895,0.1083770767,0.1862609684,0.3067202866,0.0096247802,0.1607291102,0.4856569767,0.130896315,0.2337167114,0.1021460816,-0.1254928708,-0.0429431535,-0.5238930583,0.0189285502,0.1090182289,-0.2705442309,0.3666851223,0.0564689152,0.336540252,-0.2169646323,-0.3691895008,0.0142576322,0.2399218231,-0.2655659914,-0.3388785422,0.2285338789,-0.1335350424,-0.4065559506,0.1406855881,0.0178984497,-0.0867012292,0.3517054021,-0.2302890867,-0.2614565492,-0.2153623849,0.1352092475,-0.1417113096,0.1342934072,-0.0578665435,0.2423514128,-0.0093197543,0.1342033297,0.0953271687,0.1112266481,0.2807966471,0.1006871313,-0.0732277855,-0.2079860419,0.0077028018,-0.1528949142,-0.1608589292,-0.0776968226,-0.2984029949,0.120289363,0.1377218962,0.1062908396,-0.1541672498,-0.0458696485,0.4999829531,-0.2237764001,-0.2473223358,0.4840558469,0.4217514992,-0.1941654831,-0.0158371665,0.0170503035,-0.2240185291,-0.3644656241,0.6477417946,0.0759228989,0.4203593433,0.2216892093,0.0639371499,-0.0400408022,0.2455180138,-0.053374514,-0.2108443826,-0.4087748528,0.1788351536,-0.3057606816,0.0167723335,0.1303603649,-0.1314755231,0.2277019322,0.0077145533,-0.1295635849,-0.1021613181,-0.029392371,-0.3368740678,0.4206036627,-0.0728648081,-0.1466155946,-0.1230785623,-0.2190882713,-0.0881382748,-0.0774828047,-0.3402249813,0.1750461608,0.2217362076,-0.0822981894,0.134310171,-0.0083880648,0.2225129157,-0.3318454325,0.0433524027,0.144706443,0.1511901766,-0.2609167397,0.2926473916,-0.0769182816,-0.2004084438,-0.3502568901,0.2879448831,0.0859967098,0.3944647908,-0.1279842257,0.1640851647,-0.3171319366,0.0340158939,0.3650172353,-0.1681365222,-0.3235811889,0.4275168478,-0.0395362936,0.2235413492,0.6821812391,0.3631203473,-0.0285397172,0.2105820626,-0.1586071253,-0.3186752796,0.459502548,0.276570797,-0.4551957548,-0.1983164251,0.2825817764,0.4700420201,-0.2568918169,-0.4221356213,-0.2443165183,0.4609529972,-0.1922482997,0.0276600011,0.2145583183,0.0945017487,-0.09905193,-0.0457403697,0.0095974375,0.1780732423,-0.2173994482,0.1988812834,-0.1616548449]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2200","title":"_prepare_split will overwrite DatasetBuilder.info.features","comments":"> Hi ! This might be related to #2153\r\n> \r\n> You're right the ArrowWriter should be initialized with `features=self.info.features` ! Good catch\r\n> I'm opening a PR to fix this and also to figure out how it was not caught in the tests\r\n> \r\n> EDIT: opened #2201\r\n\r\nGlad to hear that! Thank you for your fix, I'm new to huggingface, it's a fantastic project \ud83d\ude01","body":"Hi, here is my issue:\r\nI initialized a Csv datasetbuilder with specific features:\r\n```\r\ndef get_dataset_features(data_args):\r\n features = {}\r\n if data_args.text_features:\r\n features.update({text_feature: hf_features.Value(\"string\") for text_feature in data_args.text_features.strip().split(\",\")})\r\n if data_args.num_features:\r\n features.update({text_feature: hf_features.Value(\"float32\") for text_feature in data_args.num_features.strip().split(\",\")})\r\n if data_args.label_classes:\r\n features[\"label\"] = hf_features.ClassLabel(names=data_args.label_classes.strip().split(\",\"))\r\n else:\r\n features[\"label\"] = hf_features.Value(\"float32\")\r\n return hf_features.Features(features)\r\n\r\ndatasets = load_dataset(extension,\r\n data_files=data_files,\r\n sep=data_args.delimiter,\r\n header=data_args.header,\r\n column_names=data_args.column_names.split(\",\") if data_args.column_names else None,\r\n features=get_dataset_features(data_args=data_args))\r\n```\r\nThe `features` is printout as below before `builder_instance.as_dataset` is called:\r\n```\r\n{'label': ClassLabel(num_classes=2, names=['unacceptable', 'acceptable'], names_file=None, id=None), 'notated': Value(dtype='string', id=None), 'sentence': Value(dtype='string', id=None), 'src_code': Value(dtype='string', id=None)}\r\n````\r\n\r\nBut after the `builder_instance.as_dataset` is called for Csv dataset builder, the `features` is changed to:\r\n```\r\n{'label': Value(dtype='int64', id=None), 'notated': Value(dtype='string', id=None), 'sentence': Value(dtype='string', id=None), 'src_code': Value(dtype='string', id=None)}\r\n```\r\n\r\nAfter digged into the code, I releazed that in `ArrowBasedBuilder._prepare_split`, the DatasetBuilder's info's features will be overwrited by `ArrowWriter`'s `_features`. \r\nBut `ArrowWriter` is initailized without passing `features`.\r\nSo my concern is:\r\nIt's this overwrite must be done, or, should it be an option to pass features in `_prepare_split` function?","comment_length":67,"text":"_prepare_split will overwrite DatasetBuilder.info.features\nHi, here is my issue:\r\nI initialized a Csv datasetbuilder with specific features:\r\n```\r\ndef get_dataset_features(data_args):\r\n features = {}\r\n if data_args.text_features:\r\n features.update({text_feature: hf_features.Value(\"string\") for text_feature in data_args.text_features.strip().split(\",\")})\r\n if data_args.num_features:\r\n features.update({text_feature: hf_features.Value(\"float32\") for text_feature in data_args.num_features.strip().split(\",\")})\r\n if data_args.label_classes:\r\n features[\"label\"] = hf_features.ClassLabel(names=data_args.label_classes.strip().split(\",\"))\r\n else:\r\n features[\"label\"] = hf_features.Value(\"float32\")\r\n return hf_features.Features(features)\r\n\r\ndatasets = load_dataset(extension,\r\n data_files=data_files,\r\n sep=data_args.delimiter,\r\n header=data_args.header,\r\n column_names=data_args.column_names.split(\",\") if data_args.column_names else None,\r\n features=get_dataset_features(data_args=data_args))\r\n```\r\nThe `features` is printout as below before `builder_instance.as_dataset` is called:\r\n```\r\n{'label': ClassLabel(num_classes=2, names=['unacceptable', 'acceptable'], names_file=None, id=None), 'notated': Value(dtype='string', id=None), 'sentence': Value(dtype='string', id=None), 'src_code': Value(dtype='string', id=None)}\r\n````\r\n\r\nBut after the `builder_instance.as_dataset` is called for Csv dataset builder, the `features` is changed to:\r\n```\r\n{'label': Value(dtype='int64', id=None), 'notated': Value(dtype='string', id=None), 'sentence': Value(dtype='string', id=None), 'src_code': Value(dtype='string', id=None)}\r\n```\r\n\r\nAfter digged into the code, I releazed that in `ArrowBasedBuilder._prepare_split`, the DatasetBuilder's info's features will be overwrited by `ArrowWriter`'s `_features`. \r\nBut `ArrowWriter` is initailized without passing `features`.\r\nSo my concern is:\r\nIt's this overwrite must be done, or, should it be an option to pass features in `_prepare_split` function?\n> Hi ! This might be related to #2153\r\n> \r\n> You're right the ArrowWriter should be initialized with `features=self.info.features` ! Good catch\r\n> I'm opening a PR to fix this and also to figure out how it was not caught in the tests\r\n> \r\n> EDIT: opened #2201\r\n\r\nGlad to hear that! Thank you for your fix, I'm new to huggingface, it's a fantastic project \ud83d\ude01","embeddings":[-0.2406021953,-0.0382910147,-0.1077674851,0.1783376485,0.2982976437,0.2197924554,0.4546083808,0.2158689797,-0.3327544034,0.1149650663,0.1291757524,0.1715127975,0.0863348842,0.4873531759,-0.1215008423,0.0767409503,0.0831202567,0.2866099775,0.1698089838,-0.0732159466,-0.2161129415,0.103940621,-0.260414362,0.0780965611,-0.4611605406,0.0068310783,0.0738492385,0.1830849648,0.1115719378,-0.4215717316,0.1278963238,0.1931540817,-0.0024506638,0.0135782873,-0.0001165283,0.0921400413,0.0531196035,-0.0490681157,-0.3561661243,0.1506268084,-0.3260130882,-0.2654621005,-0.0111547243,-0.2360556424,-0.1424901187,-0.3270657361,-0.3870328367,0.1431457102,0.21410124,0.2266549468,0.1773981899,-0.2564507723,0.1165271029,0.1099403724,0.3127687573,0.1972708553,-0.0730632916,-0.2695540786,-0.294752568,0.1149955839,-0.1780861318,0.4079653323,-0.1879745275,-0.0244833566,0.1067840531,0.0501348674,-0.0720179081,-0.1327121854,0.0906457528,-0.3222067356,0.2972629964,0.0031607789,-0.0745959058,-0.5507095456,-0.1756774783,-0.3726785183,0.127729401,-0.167284295,0.1768467575,0.1145290211,-0.106956929,-0.3455787599,0.1335982829,-0.0906985849,-0.0378318988,0.3073068261,-0.4004644454,0.3126069605,-0.1145926118,0.0700759366,0.295794636,-0.2810826898,-0.2957600653,-0.1158553436,-0.1413079649,-0.1271789819,0.1206175536,-0.1135420054,-0.0204882305,0.0391933806,-0.183370471,0.1696168929,-0.0626039505,-0.0579128414,0.2710881531,0.3043972552,0.2000663131,0.1332727075,-0.0148484511,0.3680326641,-0.3029093444,-0.1120791882,0.3196536303,-0.1899855882,0.3789294064,0.3041920364,0.5231932402,-0.191226095,-0.2840594351,0.2589049637,0.3041548729,-0.047680594,-0.0207292493,0.1026611105,0.0128632467,0.2322100401,-0.2137506455,0.0731487572,-0.302153945,-0.1779461801,-0.2684747875,-0.2020440698,-0.0143179186,0.1958020329,-0.0162585042,-0.028942639,0.4025155902,0.1338067055,0.059362866,-0.5640695095,-0.0786658973,-0.141408965,0.245393455,0.2372240126,0.0063599204,0.297984153,0.345356226,-0.1086787507,0.0423199162,0.0928914696,-0.4464731514,-0.3818405867,0.0246932805,0.204917416,-0.038765531,0.2318276614,0.0056333705,0.0674563348,0.3656390905,-0.0792343542,-0.0102581428,-0.184376806,-0.4282066822,-0.0858234316,0.0689308494,0.3378699124,-0.3634487987,0.0270513836,-0.0374605916,-0.0010612832,0.2343657315,0.0017081597,-0.0472922921,0.2207314819,-0.2923869193,-0.1538804322,0.4493572712,0.0422428735,-0.2183953077,0.4702954888,-0.0512736104,0.1960454583,0.097857289,-0.1573392004,0.312523365,0.0569648445,-0.0447400771,0.0004948836,-0.1355825812,-0.0460937694,-0.2612141669,-0.2650542259,-0.008415495,-0.2596246004,0.0484578907,0.2520490289,-0.149836421,-0.2065731287,0.2138381004,0.1278960705,0.0759673044,0.238937363,-0.0274179019,0.4694034457,0.1447722316,-0.1496132612,-0.2516558468,-0.0256625861,-0.006287029,-0.226612851,-0.1490036845,-0.5273110271,-0.4477546513,-0.0709879026,-0.376218468,-0.264528662,0.0848521516,0.1711636782,-0.0287180431,-0.1455246806,0.0106613887,0.0743661597,-0.1461340338,0.1596860439,-0.0771052167,0.0408394523,-0.0278835911,0.0059103705,-0.0508349724,0.3137435317,0.2422779202,-0.1615428478,-0.3486573398,0.5029213428,0.129590854,0.0253561251,-0.3426887095,-0.3701828122,0.1229716316,-0.0126654766,-0.0751680434,-0.0887555107,0.0540345162,-0.0973859057,-0.0829878077,0.4442037046,-0.1351927668,0.2727766335,0.0633149743,-0.0458262675,0.1429138184,-0.0051077781,0.2960257828,-0.4752125442,-0.4301364422,-0.1807939857,-0.1933782995,0.1361899674,-0.0568668954,0.4161864221,0.7362260222,0.0912750587,-0.2430435121,-0.0015031998,-0.1585697085,-0.0849274471,0.1719686985,-0.1211564466,0.3886586726,0.0281732567,0.0056694192,0.0048922412,-0.0323953554,0.0279556587,0.2369142026,0.195208624,0.1960088462,0.2906233966,0.0994656309,-0.0508444123,-0.3068084419,0.2607759833,0.1012084484,0.1346651912,-0.3549932539,0.3007321954,-0.3476017416,0.1741188616,-0.3198778927,-0.0015288031,-0.0686357468,-0.4444669485,0.0085163433,0.3251644373,-0.04318358,0.1285805553,-0.1287238449,-0.1116976589,0.0461462475,-0.2226932645,0.1801130623,-0.0527548119,-0.3112846911,0.0221238621,-0.0914603621,0.1510246545,0.0681759343,0.0183007941,-0.2734510303,-0.2057698667,-0.3139711022,0.1297929287,-0.2200113833,0.1257251501,0.1559238434,-0.2314231992,0.3351610899,-0.3764411211,0.1105968058,0.1745934039,-0.030335892,-0.0731975138,0.1376205236,0.0059762136,-0.1190199405,-0.5209413767,0.0441726148,-0.2698732316,-0.0336812027,0.2164280266,0.0426517986,0.0496085621,0.011808469,-0.1766927093,0.179286629,0.0049384478,-0.4112112522,-0.2042838186,-0.0725068152,-0.1624400169,0.2003741264,0.031419199,-0.1869898736,0.0107457153,-0.0141782928,-0.3892456591,0.0628277361,-0.2297626436,0.1830192059,-0.0313399471,-0.0654165223,0.0715479031,0.3412663043,-0.0353141055,0.0644659922,-0.2030811757,0.2194034606,-0.0548892543,0.0521089919,-0.1899810135,0.3345891535,-0.1968087107,0.7819837332,-0.0370066687,0.1000483185,0.0815292224,-0.0416024178,0.2504915297,-0.2591479421,-0.3987992406,0.1645832807,-0.1963440776,-0.0728487447,0.3796070814,-0.0990734175,0.147952497,0.2129001766,0.0369535461,-0.1447470188,-0.3877590001,-0.0382353552,-0.1339042634,0.4046775997,0.1702878326,0.1226609796,-0.2520200908,-0.0809987932,-0.1148270741,0.2434997261,0.2173370421,-0.2415080369,-0.4367235601,-0.0477512889,0.1094470844,0.1142058969,0.1452283263,0.1801118702,-0.0691871345,-0.2261043489,0.1469920278,0.0663821921,0.6020619869,-0.001593545,0.0893499106,0.2863385677,0.1414212584,0.4394608438,-0.385630995,-0.0263817832,0.6457603574,0.0466839336,0.3440138102,0.0570220351,-0.3574425578,0.7359841466,0.2237891555,-0.4519286752,0.1409161091,-0.1648079157,0.0155695621,-0.1326268613,0.0458727777,-0.0465164036,0.0182099994,-0.3915722966,-0.2023957968,-0.2415165305,-0.1870998442,-0.0574111529,0.0098595377,0.2120065838,-0.0419113189,0.0273001771,-0.0304028541,0.1045274809,0.1013760343,0.6346020699,-0.1218815744,-0.6454026103,0.0526319705,-0.2799099088,0.3864447474,0.2780576646,-0.197907716,0.3460095227,-0.0478184596,0.2940057516,-0.5682277679,0.2830225527,0.4415433705,0.09933386,-0.3798827827,-0.5284830332,0.3364790678,-0.1991058439,-0.1632905006,-0.1263380796,0.1568223983,-0.5815492272,0.502946794,-0.0480983034,0.552575767,0.2817796171,0.1456730217,-0.0596217476,0.4062951207,0.5035848022,-0.2715296149,0.2468828708,-0.3581906259,-0.4022682309,-0.0234562978,0.1249708086,0.26547575,0.1090989336,-0.2776915729,0.493137151,-0.364053756,0.2984279096,-0.1434563249,0.0993173942,0.1148136482,-0.1004862264,0.2887810171,0.0711408406,0.0594277829,0.0769228488,0.1941160709,-0.0005984192,0.3359731436,-0.0402314253,-0.0544736534,0.0335327834,-0.0611637719,0.150058493,0.013336163,-0.276491791,0.3465050757,-0.3510287702,0.3947290778,0.2172810286,0.2850671411,0.0914980397,0.3242627978,0.2868179977,-0.1777195185,0.2628845572,0.2786632478,0.3406605721,-0.1953748465,0.1444916278,-0.1367163658,0.0413951091,-0.232319966,-0.1036987081,0.2827704251,-0.8835759163,-0.094782792,0.0146639636,0.1324677169,-0.3969879448,0.0564999543,-0.3819342554,-0.2622256577,0.1641947329,-0.0680011436,-0.3061917126,-0.0870294347,-0.1133318171,0.3726708889,0.1711788028,0.6264430881,-0.0985798687,-0.157681331,-0.2450169921,-0.0059849536,-0.123159118,-0.2279707938,0.1567362398,0.0155930594,-0.0709494203,0.0796243772,0.1542987227,-0.1563897431,0.3370626569,-0.3017419875,-0.2852936387,-0.3449874222,-0.2147877216,-0.0338958241,0.0869262815,0.1602318138,0.185390383,-0.2084304392,0.4056222141,-0.2619144022,-0.0314986371,-0.0364254713,0.3956214488,0.0656920671,0.0869985148,-0.0181533862,-0.0262147076,0.121575959,-0.1006318554,0.268035084,-0.2361129671,-0.3126168251,0.1382649839,-0.1827713102,0.0544784032,-0.1590766162,0.0578636527,0.0218237881,-0.1035419852,0.3592787683,-0.141206041,0.3238607943,0.4083138108,-0.2750033438,0.0652092248,-0.0785179883,-0.0183321945,-0.12258178,0.3001945615,0.1198046282,0.2510439456,-0.1364616901,0.0107375849,0.0105193527,-0.0630402565,-0.2360347658,0.0126777636,0.2943050563,-0.1888019741,0.141473949,0.1635720581,0.2870686948,0.4440115094,0.1440840364,-0.2545003295,0.2238463461,0.1845123768,-0.1594916284,-0.179832682,0.2453946173,-0.2276021987,0.2211372554,0.4496669471,0.2195438743,0.2532426417,0.1180917397,0.2186754942,0.1186124831,-0.0602947362,0.4032922685,0.6268619895,0.1083770767,0.1862609684,0.3067202866,0.0096247802,0.1607291102,0.4856569767,0.130896315,0.2337167114,0.1021460816,-0.1254928708,-0.0429431535,-0.5238930583,0.0189285502,0.1090182289,-0.2705442309,0.3666851223,0.0564689152,0.336540252,-0.2169646323,-0.3691895008,0.0142576322,0.2399218231,-0.2655659914,-0.3388785422,0.2285338789,-0.1335350424,-0.4065559506,0.1406855881,0.0178984497,-0.0867012292,0.3517054021,-0.2302890867,-0.2614565492,-0.2153623849,0.1352092475,-0.1417113096,0.1342934072,-0.0578665435,0.2423514128,-0.0093197543,0.1342033297,0.0953271687,0.1112266481,0.2807966471,0.1006871313,-0.0732277855,-0.2079860419,0.0077028018,-0.1528949142,-0.1608589292,-0.0776968226,-0.2984029949,0.120289363,0.1377218962,0.1062908396,-0.1541672498,-0.0458696485,0.4999829531,-0.2237764001,-0.2473223358,0.4840558469,0.4217514992,-0.1941654831,-0.0158371665,0.0170503035,-0.2240185291,-0.3644656241,0.6477417946,0.0759228989,0.4203593433,0.2216892093,0.0639371499,-0.0400408022,0.2455180138,-0.053374514,-0.2108443826,-0.4087748528,0.1788351536,-0.3057606816,0.0167723335,0.1303603649,-0.1314755231,0.2277019322,0.0077145533,-0.1295635849,-0.1021613181,-0.029392371,-0.3368740678,0.4206036627,-0.0728648081,-0.1466155946,-0.1230785623,-0.2190882713,-0.0881382748,-0.0774828047,-0.3402249813,0.1750461608,0.2217362076,-0.0822981894,0.134310171,-0.0083880648,0.2225129157,-0.3318454325,0.0433524027,0.144706443,0.1511901766,-0.2609167397,0.2926473916,-0.0769182816,-0.2004084438,-0.3502568901,0.2879448831,0.0859967098,0.3944647908,-0.1279842257,0.1640851647,-0.3171319366,0.0340158939,0.3650172353,-0.1681365222,-0.3235811889,0.4275168478,-0.0395362936,0.2235413492,0.6821812391,0.3631203473,-0.0285397172,0.2105820626,-0.1586071253,-0.3186752796,0.459502548,0.276570797,-0.4551957548,-0.1983164251,0.2825817764,0.4700420201,-0.2568918169,-0.4221356213,-0.2443165183,0.4609529972,-0.1922482997,0.0276600011,0.2145583183,0.0945017487,-0.09905193,-0.0457403697,0.0095974375,0.1780732423,-0.2173994482,0.1988812834,-0.1616548449]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2196","title":"`load_dataset` caches two arrow files?","comments":"Hi ! Files that starts with `cache-*` are cached computation files, i.e. they are the cached results of map\/filter\/cast\/etc. operations. For example if you used `map` on your dataset to transform it, then the resulting dataset is going to be stored and cached in a `cache-*` file. These files are used to avoid having to load the dataset in RAM, even after many transforms","body":"Hi,\r\n\r\nI am using datasets to load large json file of 587G.\r\nI checked the cached folder and found that there are two arrow files created:\r\n* `cache-ed205e500a7dc44c.arrow` - 355G\r\n* `json-train.arrow` - 582G\r\n\r\nWhy is the first file created?\r\nIf I delete it, would I still be able to `load_from_disk`?","comment_length":64,"text":"`load_dataset` caches two arrow files?\nHi,\r\n\r\nI am using datasets to load large json file of 587G.\r\nI checked the cached folder and found that there are two arrow files created:\r\n* `cache-ed205e500a7dc44c.arrow` - 355G\r\n* `json-train.arrow` - 582G\r\n\r\nWhy is the first file created?\r\nIf I delete it, would I still be able to `load_from_disk`?\nHi ! Files that starts with `cache-*` are cached computation files, i.e. they are the cached results of map\/filter\/cast\/etc. operations. For example if you used `map` on your dataset to transform it, then the resulting dataset is going to be stored and cached in a `cache-*` file. These files are used to avoid having to load the dataset in RAM, even after many transforms","embeddings":[-0.0206235927,-0.1851760596,-0.1328714788,0.6645821929,-0.0811096057,0.305211246,0.1752955765,0.2564817667,0.287450105,-0.1578645408,-0.0090863518,0.1946638376,0.0801829025,-0.5130008459,0.192416504,0.1686628014,0.1584644467,0.0357156284,-0.1609626412,-0.1386406422,-0.1706289351,0.1584541202,0.2008032501,0.1132956147,-0.4863940179,-0.1746705472,0.1140891612,0.2621504366,-0.0464333221,-0.3418199718,0.2564128041,-0.0493696518,0.0540107265,0.3561965227,-0.0001198421,0.0101960944,0.3606316149,0.1277110875,-0.4809896946,0.0153871085,-0.3229597509,-0.0753125027,0.1952300072,-0.2579166293,0.45841977,-0.1872411072,-0.0695052296,-0.7284278274,0.568451345,0.1373000294,0.1673657745,-0.1836172938,-0.4365001917,0.1170695126,0.2427159101,0.0862704366,0.0348213725,-0.0274123438,-0.1152970567,0.2550847828,0.0084783314,0.0654598102,-0.1409609914,-0.0061692763,0.4315126836,0.1002563015,-0.0336375944,-0.2234607637,0.2855522633,0.0902169719,0.8578144312,-0.2979494035,0.0392456874,-0.3473233283,-0.2356774956,-0.1443538517,0.2823199034,0.3157087266,0.1556548923,-0.0248261206,-0.1394176036,-0.4467857182,0.0990005285,-0.1330397129,0.1985005438,-0.3570916951,-0.1704366952,0.3212152719,0.3046756983,0.3051571548,0.3443063498,-0.4836414456,-0.2567694485,0.3919183314,-0.1676268131,0.1453921944,-0.1216804758,0.019767141,-0.0845500678,0.3244434297,0.3523723185,-0.2121753246,-0.1888590902,0.1681446284,0.1398879141,0.3988686204,0.1835668087,-0.002278852,0.011346532,-0.3571019769,-0.0108401524,-0.260830164,-0.0843963921,-0.1345771551,0.3956592977,-0.3732519448,-0.0711927935,-0.1004829705,0.0157761239,0.0031898427,0.0849955454,-0.3467419446,0.1040635109,0.0538092293,0.2754336596,0.2759617269,-0.1532218903,0.038046509,-0.0484487377,-0.0527621694,-0.1766158938,-0.0172601081,-0.1700626463,0.2704500854,0.2655587494,0.0037329083,0.2545683384,0.0916006714,-0.0608627461,-0.0543895699,0.386248976,-0.2153563052,0.1529203653,0.4383428395,-0.1737543792,0.427498728,0.0190216415,-0.01698591,-0.1211496517,0.4724342525,-0.48855865,-0.1591059566,0.1287668943,0.1424016058,-0.0551954471,0.3164665103,-0.808244586,-0.0007383485,0.2645081878,-0.1677342653,0.2512048483,0.0491757728,-0.1883518994,-0.3580837846,-0.2017025948,0.2732269466,-0.4757975936,-0.1344007403,-0.212552309,-0.184701845,-0.1997340471,0.347836107,-0.5759468675,0.3453568816,-0.5347441435,0.1664724648,0.5266288519,-0.3847642839,-0.3605982959,0.2506011426,-0.0380812287,0.1663451195,0.2721282244,0.2182006985,-0.1023307294,-0.2018231302,-0.1041522101,-0.0861043632,0.1661472917,-0.3134887815,-0.1434256732,0.0190952569,0.0897763893,-0.0663595349,-0.1878647953,0.0354379825,0.0699583888,-0.0580334328,0.217008248,0.0795840099,0.1795778871,0.1891771555,-0.143675983,-0.2049163133,0.2398435622,0.1060039401,-0.6893283129,0.2930278182,-0.1043398604,-0.7563843131,-0.0923603848,-0.0757525414,-0.1200728491,-0.065023765,-0.2157051563,0.0706047565,-0.047408089,0.1403094828,0.1338894069,-0.2436747402,-0.3252306581,0.7150179744,-0.0651439577,-0.0319744498,-0.4741214216,0.0818874761,0.1955753863,0.2808515728,-0.3693420887,-0.1733142585,0.0616052411,-0.1553861052,-0.0872999877,0.3423824906,0.1587363183,0.0816379339,0.1597303897,0.3373652697,0.0528742336,-0.094502151,0.2905022204,-0.1088058278,0.2044790536,-0.3820786178,-0.1535615325,0.0764505342,-0.2378111482,0.1004443243,-0.0944696069,-0.3120369613,0.1867325604,0.0675016269,-0.0250660274,-0.1098622233,0.1858586967,0.3055247366,0.4112285078,0.4338188767,0.1313607842,-0.0053280345,0.4112856388,-0.2373322994,-0.1887231618,0.1868093908,-0.3113767505,-0.3228855133,-0.1912425011,0.4735212326,0.6378601789,0.0563691482,0.1754252315,0.0024439227,0.1258141994,-0.1620593071,0.2027002424,-0.0622796863,0.3740920424,0.0654769167,-0.152791813,-0.1322551817,-0.3498427868,-0.0258125234,0.2369617969,0.0013321505,-0.2612858117,0.1710428894,-0.0301582124,-0.0766577274,-0.2585294247,0.1396564245,-0.1742328107,-0.1237514243,-0.0833469331,0.1740720868,0.3223650753,-0.1156030968,-0.301685214,0.5464680791,-0.0441392511,0.0205827486,-0.4892040789,-0.1473871469,-0.0474743657,0.0060407575,0.0929661244,-0.1244915575,0.0971613377,-0.1376733631,0.2625198364,-0.251023978,0.1507316679,-0.0135110701,0.1546670645,-0.0067826696,-0.1983197331,0.0210131202,-0.3504289985,0.0431183763,-0.1127622873,0.0284714885,-0.2904050648,-0.053298939,-0.084039934,0.0628023222,-0.0875963718,-0.0336913802,0.0717709288,0.0021123542,0.2212624103,-0.054850772,0.1340098977,-0.032885246,-0.3017032444,0.1539708674,0.1445062757,0.108298853,-0.4748567045,-0.7183316946,0.2579834163,-0.1595869958,0.0216958281,0.1440955251,0.0385784842,0.2280184478,0.0514161065,-0.5628803372,0.1586019695,-0.1018931195,0.1548069417,-0.4036144912,-0.0201202519,0.1646022201,0.139987886,0.0117779756,-0.2070706785,-0.079028517,0.1412537992,0.02384221,0.2856884599,-0.0039138086,0.2602865398,-0.1735490113,0.6865373254,0.0560432598,0.2372025996,0.2109369487,0.4149290621,0.4307758212,-0.215962261,0.0449297912,-0.1597059071,-0.1557262838,-0.1205705032,0.1386100352,0.1522235274,-0.1358102709,0.1498071402,0.1156611219,-0.1890235096,0.1638130844,0.1285749078,-0.5458725095,-0.0936674774,0.0829505175,-0.2028476894,-0.1122084111,0.1043317541,-0.0637394413,0.1197961792,0.3924949169,0.0245919824,-0.3626145124,0.0418764092,-0.1019601971,0.1704232544,-0.1276641339,-0.1241331026,0.0466336906,0.1456721276,0.1839406788,0.1173278466,0.7556931973,0.0689093098,-0.0042753359,0.2504169047,0.1635794193,-0.3476664126,0.0561862774,0.1012089327,0.3155828118,-0.0724934191,0.3185842633,-0.0702631548,-0.08725968,-0.4280514121,0.4414302707,-0.3793692589,-0.228442356,-0.1763138473,-0.166102007,-0.006511244,-0.2366746515,-0.0372680537,-0.0689697191,-0.1220226586,0.0307865851,-0.2609102726,-0.219263494,-0.0041815769,-0.1828789413,0.4164488912,0.06281773,-0.0163298585,0.3135465384,0.4170598984,0.1798792183,0.5180498362,-0.0844290331,-0.0402149372,-0.0343479924,-0.0298483167,0.0105954269,0.0120353494,-0.1521036774,-0.0000464499,0.0288307611,-0.2253950387,0.019068636,-0.0389360748,0.047871124,-0.0167067796,-0.2692253888,-0.5570721626,0.5188219547,0.2009653598,-0.329957664,0.5775389671,-0.1945197135,-0.2601770461,0.1905091256,-0.179751426,0.8575187922,-0.2320058644,0.1530697644,-0.0931503251,-0.3358079791,0.2799904048,-0.4325314164,-0.0136612877,-0.0686786175,0.0142071908,-0.0639224425,-0.319373548,0.107494913,0.5550762415,0.0391736701,0.0568457469,-0.1903911531,-0.0959137902,0.0254774168,0.234972477,0.1144002229,-0.1327078342,-0.0066075278,0.0858254507,-0.1981879324,-0.1961918473,-0.0707825646,0.1385371387,0.1334469914,-0.1467639804,0.1382436752,-0.053993497,-0.0569482855,0.5380523205,-0.1739407331,-0.2969219983,-0.1323028058,0.4597611427,0.0470264666,-0.3815936744,-0.0140121365,0.4528037608,0.2481702566,0.1352085769,-0.349324435,-0.1970515698,0.1152032316,-0.0626428276,-0.0331331231,0.1704402268,-0.3907630742,-0.520293057,0.0758625269,0.383943975,0.0440740585,0.0700656623,-0.2338163257,0.1925329715,-0.0393804386,0.1221188083,0.0389428549,0.3972509205,0.07539846,0.0174660962,-0.0109421117,-0.4464409053,0.0197867658,0.2537942231,0.3440080583,-0.2685213387,0.4966940284,-0.2944134176,-0.2041380852,-0.0906013697,0.1410117596,0.3131948113,-0.4004871547,0.3007013202,0.0699777305,0.2724123597,-0.0875865668,-0.1474675536,0.2230041027,0.180605039,-0.0779224411,-0.2757705748,-0.3409194648,0.0647654459,-0.2518686354,0.1543305367,0.2471299171,-0.7568339705,0.0359936841,0.2772006989,-0.1806924939,0.4340602756,0.1817559451,0.1086573824,0.0465390533,0.0083201788,0.203563109,-0.0067307614,-0.0142896101,-0.1547662318,-0.281559974,-0.177363798,0.0064882566,0.1880190372,-0.1101575643,-0.1239975467,-0.3026311994,-0.2136261612,0.2996172011,-0.0924047157,0.1689259112,0.0763957426,0.0154713187,-0.0232408587,0.314001739,-0.1672372669,-0.4207923412,0.1020236984,-0.0712582543,0.3043347001,0.0287030507,0.1461309195,0.0224611647,-0.10863401,-0.3524737656,-0.0193178747,0.1491248459,-0.0739312395,0.0535204075,-0.365635246,0.1887578815,0.1058717743,0.142425403,0.2639564574,-0.1918773204,-0.1659752429,-0.0014203729,0.0538501479,0.1590597183,-0.2506591976,0.1905309409,0.0957181156,-0.3110486865,0.0160594899,0.3221443295,-0.1197286248,-0.0757652447,0.0208637435,0.4545165598,-0.2129799873,0.2867920995,0.1155598685,0.1423378736,-0.0306055825,0.4425140023,0.3101650178,0.07612928,0.491967082,-0.0862210616,0.3369765282,0.5232391953,0.0696397275,-0.150447011,-0.4384793937,-0.016478816,0.3598088026,-0.4211256802,0.0403464958,0.1181389317,0.152103141,0.2226317376,-0.1888797432,-0.100547567,0.0676699579,-0.2423571497,-0.2012050897,0.0536787733,-0.2410685122,0.1944410503,0.2857096493,-0.0437985919,0.1294073015,0.3107999563,0.0396747366,-0.2294274718,0.0982371867,0.4913252592,0.1306879371,-0.0662854537,-0.1000393927,-0.0126050161,-0.0685262755,-0.0895672366,0.1754461378,-0.0227649976,0.0520485044,0.1915345788,0.0926473662,0.1363093406,0.2215892673,-0.0598476864,-0.1363717169,0.2799260318,0.1623349935,-0.3710960448,0.0641643628,0.1114165261,-0.0446158536,0.1663723588,0.0484724864,0.3469825983,-0.2062878907,-0.1631243229,-0.3431838155,-0.0560138971,-0.1490937173,0.2874445319,-0.148705855,-0.0673667043,0.5298345685,-0.1145623848,-0.119595781,-0.2176063806,-0.0045799292,-0.1946097165,0.494995296,0.3005730808,0.0117114224,-0.1784257442,-0.2971549332,-0.4798891842,-0.0468932353,-0.2442493886,0.3913602531,0.0142155774,0.3093050122,-0.0688220412,-0.0186910331,0.4066348672,-0.04410211,0.1097253188,-0.0233554225,-0.2180859596,0.0973769575,0.4327713251,0.2211844474,-0.0986054465,-0.358175993,0.0113265608,0.174404189,0.0111418189,0.0408059843,0.0218481384,0.21596995,0.5090534687,0.6151248813,0.1940452307,0.1792437285,0.099276863,0.0417762212,-0.0513151214,-0.1942852885,0.1955423802,0.1394195408,0.0097879432,0.1717472672,-0.5143799186,0.328879416,-0.0932545662,0.0292315073,-0.1276391298,-0.1343791634,-0.1802267283,-0.1930216551,-0.3331318796,0.1861663312,0.195795089,0.2248795182,0.0822166726,0.1197485551,-0.0191670004,0.0241712816,0.4215191901,-0.1253902167,-0.1972171962,0.2731103003,0.2348908931,0.2969862223,-0.1469351202,-0.3629284501,-0.051629141,0.5664259195,-0.2388070375,-0.2849764228,-0.0066741304,-0.2340872437,0.07934618,0.0303652436,0.5920177698,0.0352816135,-0.1686136425,-0.0769130439,-0.3721188903]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2196","title":"`load_dataset` caches two arrow files?","comments":"Thanks @lhoestq! Hmm.. that's strange because I specifically turned off auto caching, and saved mapped result, using `save_to_disk`, to another location. At this location, the following file is created:`355G\tcache-ed205e500a7dc44c.arrow`\r\n\r\nTo my observation, both `load_dataset` and `map` creates `cache-*` files, and I wonder what the `cache-*` file from `load_dataset` is for (as I believe the same information is stored in `json-train.arrow`.","body":"Hi,\r\n\r\nI am using datasets to load large json file of 587G.\r\nI checked the cached folder and found that there are two arrow files created:\r\n* `cache-ed205e500a7dc44c.arrow` - 355G\r\n* `json-train.arrow` - 582G\r\n\r\nWhy is the first file created?\r\nIf I delete it, would I still be able to `load_from_disk`?","comment_length":61,"text":"`load_dataset` caches two arrow files?\nHi,\r\n\r\nI am using datasets to load large json file of 587G.\r\nI checked the cached folder and found that there are two arrow files created:\r\n* `cache-ed205e500a7dc44c.arrow` - 355G\r\n* `json-train.arrow` - 582G\r\n\r\nWhy is the first file created?\r\nIf I delete it, would I still be able to `load_from_disk`?\nThanks @lhoestq! Hmm.. that's strange because I specifically turned off auto caching, and saved mapped result, using `save_to_disk`, to another location. At this location, the following file is created:`355G\tcache-ed205e500a7dc44c.arrow`\r\n\r\nTo my observation, both `load_dataset` and `map` creates `cache-*` files, and I wonder what the `cache-*` file from `load_dataset` is for (as I believe the same information is stored in `json-train.arrow`.","embeddings":[0.0172763765,-0.1464014351,-0.1149083897,0.6492552161,-0.0665469244,0.3106843829,0.2409403175,0.2578835487,0.3021347225,-0.2563412189,-0.0356172957,0.2787819207,0.1161758229,-0.4971514642,0.2449564934,0.1934699416,0.1645396501,0.0410898104,-0.1407982856,-0.102236554,-0.1449192762,0.1404834986,0.222994253,0.0501033925,-0.4586227834,-0.1467895508,0.0726128966,0.1264030486,-0.0452624112,-0.4037206173,0.3527000844,-0.0786419064,0.0307803452,0.4275515378,-0.0001186709,0.0302080251,0.3575187624,0.0993109867,-0.3851467669,-0.0286421832,-0.4544894695,-0.045694299,0.1494431496,-0.1819910556,0.380058825,-0.3371661901,-0.0662364289,-0.6853268743,0.4682787955,0.1548903733,0.179665491,-0.1966743469,-0.4313079417,0.0719178542,0.293679744,0.1138556749,-0.0295685101,-0.0345636867,-0.1491800249,0.2294983268,-0.035603445,0.1215262935,-0.1098485813,-0.0463744439,0.4478723705,0.1780072302,-0.1123236567,-0.2661146522,0.3580049574,0.053959474,0.8774031997,-0.3279373348,0.0023098926,-0.3756660819,-0.2317557335,-0.0114048133,0.2280352265,0.3111110628,0.1924596429,-0.0084913597,-0.201050505,-0.4259337783,0.0605824068,-0.1790846288,0.1664555967,-0.3395424187,-0.2024616152,0.3108737767,0.239737466,0.2673693299,0.2836093903,-0.4972212017,-0.3389441669,0.4030974209,-0.1985778362,0.0318644531,-0.0466743521,-0.0403825156,-0.03329641,0.29533723,0.4044393599,-0.2616675794,-0.1658657342,0.1207460389,0.2782330513,0.3537445664,0.2547233999,0.0268986337,-0.0461516008,-0.339974016,-0.0218213089,-0.208674103,-0.0490394048,-0.0859611183,0.4438293576,-0.4126614332,-0.0931672007,-0.0895797983,0.0455818772,0.0914719701,0.0405215658,-0.3477860391,0.0488015935,0.0804478601,0.182041809,0.2340861708,-0.233913824,0.0636636317,-0.0101311086,-0.0300818793,-0.1814797521,-0.0180175155,-0.1444652081,0.265342623,0.273581475,0.0721837953,0.2267827243,0.144211337,-0.2142865658,-0.0831868052,0.459084928,-0.2324572206,0.1928142756,0.4072645307,-0.163272813,0.4347712696,-0.060497269,0.1209551468,-0.1324381232,0.5263389349,-0.4890322983,-0.2337027639,0.1063235626,0.1635842174,-0.1018485352,0.2115263194,-0.6663516164,0.0656587705,0.323399514,-0.2094890773,0.2708799541,0.0823254585,-0.2421623766,-0.3829841912,-0.1247377023,0.3206803501,-0.3754127622,-0.1680775881,-0.1721181124,-0.1740692705,-0.2347567976,0.2553412616,-0.5216798782,0.3648763299,-0.4526238739,0.1768103391,0.5462340117,-0.3530583382,-0.4222028553,0.2921740115,-0.0394608714,0.1639520079,0.1800560951,0.2358596921,-0.129404977,-0.1869406253,-0.0265118349,-0.0481117629,0.1490021348,-0.252238214,-0.0990578309,-0.0420263819,0.0355101973,-0.0342949294,-0.2268188149,0.0001690536,0.0800405741,-0.0962591246,0.1690088212,0.1523696631,0.2140231878,0.2017194182,-0.1144552901,-0.2469071895,0.2298972309,0.1480555683,-0.7635045648,0.3079406023,-0.1594994515,-0.7619137168,-0.0747809261,-0.1076103002,-0.1010557115,-0.0893118978,-0.2672033906,-0.0377834477,-0.0343671031,0.12902309,0.0996729136,-0.189741984,-0.2920210958,0.5691560507,0.0448395796,-0.0608553216,-0.5005730391,0.0350248888,0.235583216,0.2694952488,-0.3725483716,-0.1414360106,0.0084840851,-0.1698793769,-0.0955605134,0.2989889085,0.1324427426,0.2216421515,0.1800314784,0.2426699847,0.0743947551,-0.0678143129,0.2297535837,-0.1207529604,0.1633173823,-0.3794222772,-0.285648793,0.0898486823,-0.154698506,0.0992564335,-0.1303553134,-0.2762314975,0.1194211543,0.0226738583,0.0369797982,-0.1994014829,0.0540000275,0.2509945035,0.3887884319,0.4738147259,0.1165863201,0.1842743903,0.3787924647,-0.208732456,-0.1766487062,0.1748841703,-0.3041078746,-0.4068354666,-0.1282897145,0.5813217759,0.6566796899,0.0224437937,0.2662492394,-0.0583369695,0.205867365,-0.1070435047,0.2185213268,-0.1296971589,0.3127490282,0.1337821037,0.031962011,-0.0801476166,-0.314345032,0.0470698923,0.223111093,0.0093092341,-0.1905515492,0.1201205775,-0.0449030362,0.1156192794,-0.2616874874,0.1045317799,-0.1441285312,-0.1079823747,-0.0595245697,0.1823108494,0.2603966594,-0.1355582327,-0.3985460103,0.521171093,-0.0771162659,-0.0677677765,-0.4286523759,-0.1353064775,0.002325743,0.0156934615,0.0473663174,-0.1873987019,0.0727301985,-0.074298352,0.2118711323,-0.2597361803,0.1914972216,0.0028847721,0.1121254787,-0.0621776842,-0.1794524193,-0.0543337725,-0.3448907435,0.0772262439,-0.136048317,-0.0512222014,-0.3175914884,-0.07461638,-0.1079362258,0.1040938348,-0.0770919472,-0.0732302144,0.047450684,0.015551934,0.2687015533,-0.0390584171,0.0824766755,-0.0717997998,-0.3182495534,0.2286307365,0.0234945621,0.0890922099,-0.5059283376,-0.6877738237,0.3267150521,-0.2294784486,0.0065221018,0.1465352327,0.025075784,0.1810816526,0.0638732538,-0.6293523312,0.1646163762,-0.1493868381,0.1181597039,-0.3938648701,-0.0835525468,0.1742088348,0.1690076292,0.0234810337,-0.2177424729,-0.0835142136,0.2013215125,0.0812743753,0.3253596425,-0.0374762565,0.1487569213,-0.1572584808,0.684986949,0.1634379178,0.2408478856,0.1585013419,0.3259359896,0.4771516621,-0.2095793933,0.0312422663,-0.1076795757,-0.1500517875,-0.0661489666,0.1684324145,0.159466818,-0.1389693916,0.179469943,0.058818575,-0.1254652888,0.0917907357,0.1231938973,-0.51541394,-0.0883924961,0.0399133451,-0.1238299087,-0.1138720959,0.0483702011,-0.0279455911,0.0119859204,0.4236934781,0.0629793927,-0.3388803601,0.0458814651,-0.2007436454,0.1100528315,-0.0827294216,-0.0739983916,-0.0187990684,0.0797373801,0.1910790503,0.090039894,0.6883507371,-0.0268438291,-0.0061348584,0.1408070773,0.2076436132,-0.3507490158,0.0895611569,0.1844308823,0.3399956524,-0.0810737014,0.2973805964,-0.071422942,-0.0677842647,-0.3177435398,0.4325808287,-0.302020818,-0.2460191697,-0.1424422562,-0.1041009575,0.0907698646,-0.274502784,-0.0453043096,-0.1495144665,-0.1426372528,-0.0587911531,-0.1975502074,-0.0950500816,-0.031632442,-0.1203270406,0.4342670739,0.0305791982,0.0851259828,0.3868001401,0.3048560619,0.1845519692,0.5441380143,-0.118554607,-0.0482981205,0.0191380046,-0.1058676466,0.0057778936,-0.0007476753,-0.1402359456,0.05428873,-0.0021305187,-0.3232123852,0.0462113172,-0.1101758704,0.1355286837,-0.0768929347,-0.3059686124,-0.5708690286,0.497803539,0.1715555936,-0.2917498052,0.6982319355,-0.2817908525,-0.2447005659,0.1893452555,-0.1960577518,0.8117073774,-0.0243190844,0.2095356882,-0.0690740496,-0.31748721,0.2996848226,-0.3061925769,0.0094765751,-0.0919216573,0.0470288806,-0.0310890041,-0.2858136296,0.0259552691,0.5388069153,0.0217838753,-0.0291946363,-0.1891044527,0.0749527588,-0.0560940094,0.1630775034,0.0999965668,-0.1319270134,0.0483126305,0.1163753867,-0.1305357069,-0.1711046696,0.00165418,0.0836491659,0.0644804165,-0.1286579669,0.1008419991,0.0099275876,-0.008881269,0.5330633521,-0.0894700959,-0.3523549438,-0.1207580864,0.4396194518,0.0384690352,-0.438791424,-0.0710992366,0.4661090076,0.3222037554,0.2295128554,-0.2802366018,-0.310546875,0.1266949922,-0.0466571674,-0.0474574789,0.1918982118,-0.3758213818,-0.426864326,0.0417725593,0.4487209916,0.0104914811,0.066850245,-0.2820750475,0.2307265848,-0.0048331344,0.0974725261,0.0629568771,0.366207689,0.0885037184,0.029982293,-0.1947008371,-0.4164140224,0.0288325362,0.3071613908,0.401267767,-0.2311730832,0.5021203756,-0.315040946,-0.1916740686,-0.1308929175,0.2161295712,0.3525097072,-0.4439143538,0.1858521849,-0.0059823166,0.2789995074,-0.0248101577,-0.1342747211,0.1655061245,0.1758145988,-0.038897533,-0.2350428998,-0.3576944172,0.0346172899,-0.2280533463,0.1640298516,0.1866438389,-0.7211914062,0.0638425574,0.2420561016,-0.1967248768,0.3830199838,0.2259819061,0.0917929783,-0.0010107105,0.043557018,0.2381801754,-0.0527113266,-0.0253878571,-0.169288516,-0.2760158777,-0.16658099,-0.0204819366,0.1742615253,-0.085363023,-0.0603221916,-0.3280655146,-0.2206213325,0.1910265088,-0.0904949158,0.2266635001,0.116812855,0.0530368648,-0.1274449527,0.2419679314,-0.1703642458,-0.391895771,0.0772186369,-0.0460781716,0.2237461805,0.0458695069,0.1744982153,-0.0128612174,-0.1075060517,-0.1906539798,0.029228963,0.116123572,-0.0261584315,0.0317417681,-0.317215234,0.1668661535,0.052086547,0.1516219229,0.3093539178,-0.2213665992,-0.1850498468,-0.0657379404,0.085671477,0.1444754153,-0.2371694744,0.198524341,0.1094164848,-0.280184567,-0.0305117443,0.2940535247,-0.0821697414,-0.0816449374,0.1285952926,0.4023344815,-0.2552741766,0.3821407557,0.127399683,0.2081644982,-0.0227917787,0.397985667,0.2878246903,0.12268392,0.3527514935,-0.0447155833,0.3252020478,0.5679380298,0.01788662,-0.09070158,-0.414513886,-0.034138184,0.3077028692,-0.4453383088,0.0572337136,0.2392231226,0.2952447832,0.0887698159,-0.1675038487,-0.1869409531,0.0120238494,-0.172514081,-0.1931069642,0.1219040155,-0.3153280616,0.2408369482,0.2190691978,-0.0109222513,0.1480471641,0.3217511773,0.0533426367,-0.1923637688,-0.0535540208,0.4184318781,0.1564745158,-0.0252307411,-0.0622273535,0.0233765151,-0.0746111125,-0.1131288037,0.1222534105,-0.0186673887,-0.0272835493,0.1956581473,0.1093571186,0.1342213154,0.2877740562,-0.0464050397,-0.1271412224,0.3520168364,0.1325871944,-0.6029282212,0.1157257035,0.1073772833,-0.0242128745,0.1924082786,0.1308597475,0.3290371299,-0.1656909734,-0.0892708525,-0.3895645142,-0.0767214522,-0.0849275291,0.2677378953,-0.1598470062,-0.1504536718,0.5574601293,-0.1089921966,-0.0884576738,-0.2930591106,0.0069544609,-0.1473171711,0.5219276547,0.1645273119,-0.0285376105,-0.1346675754,-0.2367607206,-0.5200847983,0.0423934162,-0.2269410938,0.4772405028,-0.0421165079,0.3377253711,-0.1472652704,-0.0670681968,0.3972636759,0.0708963647,0.0967219174,-0.0152607448,-0.2222390324,0.10804943,0.3750351369,0.2518652081,-0.103304565,-0.3137207627,0.0358816199,0.1303761899,0.0149558121,0.038794674,-0.081535928,0.1898534149,0.4535370469,0.6795006394,0.1503202766,0.0926537588,0.0682909116,-0.0049876436,-0.1192306057,-0.2603259981,0.1461518556,0.1206531599,0.0472406894,0.1330359578,-0.4705204368,0.3438639939,-0.0892518312,-0.0315642022,-0.1404587328,-0.1297427863,-0.1468491405,-0.1365944147,-0.3660115004,0.2043118477,0.1580005735,0.1742909402,0.0356454886,0.1503091902,-0.0317372009,0.0285410304,0.3936862946,-0.1578481942,-0.1641329974,0.2927773297,0.2191907465,0.2682179213,-0.2180180103,-0.3902973831,-0.0239676498,0.5540385842,-0.2181408405,-0.2416341901,-0.0371698439,-0.2585568428,0.1128587797,0.034694232,0.667511344,0.1050681397,-0.162990734,0.0088216057,-0.3090760708]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2196","title":"`load_dataset` caches two arrow files?","comments":"This is a wrong report -- `cache-*` files are created only my `map`, not by `load_dataset`. ","body":"Hi,\r\n\r\nI am using datasets to load large json file of 587G.\r\nI checked the cached folder and found that there are two arrow files created:\r\n* `cache-ed205e500a7dc44c.arrow` - 355G\r\n* `json-train.arrow` - 582G\r\n\r\nWhy is the first file created?\r\nIf I delete it, would I still be able to `load_from_disk`?","comment_length":16,"text":"`load_dataset` caches two arrow files?\nHi,\r\n\r\nI am using datasets to load large json file of 587G.\r\nI checked the cached folder and found that there are two arrow files created:\r\n* `cache-ed205e500a7dc44c.arrow` - 355G\r\n* `json-train.arrow` - 582G\r\n\r\nWhy is the first file created?\r\nIf I delete it, would I still be able to `load_from_disk`?\nThis is a wrong report -- `cache-*` files are created only my `map`, not by `load_dataset`. ","embeddings":[-0.0013235379,-0.2079309672,-0.1252864003,0.7411247492,-0.1488292068,0.2700351179,0.2812337577,0.2348195761,0.3605192006,-0.2358909547,-0.0116218273,0.206755802,0.1341191679,-0.4797021151,0.1827089787,0.2512860894,0.2364553213,0.0539577678,-0.151545018,-0.1846313179,-0.1211207286,0.1802385449,0.1390398592,0.0808782056,-0.4849410951,-0.1015205383,0.1177735776,0.1523867548,-0.1212622374,-0.3925187588,0.337383002,-0.0722730532,0.0059329565,0.4784198701,-0.0001242887,0.0004778891,0.3728579283,0.0655030608,-0.276564151,0.0549495369,-0.3939551115,-0.1579756439,0.2119982988,-0.2245272249,0.3471077085,-0.3236004114,-0.1582753956,-0.6293208003,0.522464335,0.2015789598,0.1405273974,-0.2274221629,-0.4012844861,0.1085536778,0.2451905608,0.1741999984,0.0215819236,0.0170855094,-0.2218236476,0.257229954,-0.0648386553,0.1377795935,-0.0872283727,0.0792113394,0.4362153113,0.1867984235,-0.014609525,-0.2205650061,0.3009508252,0.0931445509,0.9287804961,-0.3096571565,-0.028492162,-0.3948613703,-0.2127708942,-0.0490798801,0.2915646136,0.3378342986,0.0747810751,-0.0071365684,-0.1541530341,-0.4959351122,0.1060581282,-0.1487178653,0.1050290167,-0.3734419048,-0.2028996199,0.3179076612,0.2064976096,0.310079813,0.3849569857,-0.4313015342,-0.3036452532,0.3299284577,-0.2232827693,0.0779077634,-0.0468327627,-0.0520557538,0.0498712175,0.2283989042,0.3403443396,-0.3397465944,-0.2176031619,0.0817578509,0.2542223036,0.3434743583,0.2462421507,0.1183107793,-0.0782853365,-0.3428346515,-0.0451073311,-0.2441424876,-0.0325658433,-0.1506538987,0.3954870701,-0.358263433,-0.0603703484,-0.1741026193,0.1083906516,0.0602608658,0.0574860647,-0.4677244723,0.0272937678,0.10050679,0.1709105372,0.3066871166,-0.1498608887,0.1831598878,-0.0014764139,-0.0591002554,-0.1941262484,-0.0403829813,-0.1624013633,0.2471527606,0.241208002,0.0371519849,0.2517976463,0.1460357904,-0.1485764086,-0.1187229529,0.3654903173,-0.1904856116,0.1931566298,0.3797282577,-0.0661892295,0.4204210639,0.0000280261,0.0562608205,-0.1615721583,0.5288683176,-0.5380883813,-0.2114272416,0.0173040088,0.1125977561,-0.135336414,0.2883089781,-0.690656662,0.0383638814,0.3620166183,-0.1594407409,0.2597061098,0.1579668224,-0.2269218713,-0.3358771205,-0.1293586195,0.4049054086,-0.4220991731,-0.1894842237,-0.2243154943,-0.1474877894,-0.0534971245,0.2821417451,-0.5153653026,0.4251534939,-0.5116041303,0.1266720742,0.5514280796,-0.3688095808,-0.4169304967,0.3923598826,-0.0574928485,0.0401594453,0.3320165873,0.1701670736,-0.1255356371,-0.1716515571,-0.0239928253,-0.0854510888,0.157312125,-0.1866728812,-0.1431426108,-0.0213273112,0.1980864406,-0.1080158204,-0.1761618704,0.0533616468,0.116089195,-0.07612703,0.2507043183,0.1806489676,0.1387550384,0.2073521912,-0.1048336476,-0.2661542594,0.2222424448,0.2363766581,-0.7416957021,0.2548560202,-0.1476524025,-0.7274240255,0.0331725664,-0.06951949,-0.1548352242,-0.0574831553,-0.2202033401,0.0517605878,-0.0655089021,0.119630985,0.0642776117,-0.2124178112,-0.2887091935,0.6405111551,-0.0720655993,-0.0352000184,-0.4461592436,0.121358484,0.2529789805,0.2813433409,-0.2945739031,-0.2160004824,0.0829773843,-0.0783961192,-0.1387543976,0.271342814,0.1348098814,0.1769557893,0.225570932,0.1641738862,0.0336066745,-0.1351716369,0.2525275648,-0.06468115,0.1850169897,-0.3635964096,-0.2463095784,0.0732182413,-0.1819232404,0.083435528,-0.172975257,-0.2876412272,0.2575789392,0.0241394993,0.0010644716,-0.2000224292,0.1289936006,0.2738920152,0.3292220235,0.4564264417,0.186922729,0.0687473938,0.4439372718,-0.1352470666,-0.2231046706,0.2293696553,-0.3471647501,-0.4305856228,-0.0428596996,0.4795961976,0.6346254349,0.0524272956,0.2430172116,-0.0588974655,0.2254140079,-0.1210730821,0.2058811486,-0.1258021146,0.2707575858,0.2693890929,-0.0361580253,-0.19101955,-0.3453933001,0.1269650161,0.2785906494,0.0752327964,-0.3163998723,0.022024421,-0.0005510038,0.0009883376,-0.2347723842,0.155482024,-0.1514724195,-0.0571271069,-0.0359404348,0.101245299,0.3760837913,-0.1483340412,-0.3818002641,0.4429107606,-0.0317033567,0.0120561374,-0.4359982908,-0.0773774683,0.0062160776,-0.0201945174,0.0612643436,-0.0691547915,0.1264527142,-0.1897959411,0.2256261557,-0.2579315007,0.2238742113,0.0586705804,0.1991300136,-0.0722822696,-0.0932445899,-0.0063899853,-0.3061403036,0.0402889103,-0.1290145665,-0.0858143046,-0.3608880639,-0.0347954482,-0.2356941849,0.1552826315,-0.0740585327,-0.1209929809,0.086680375,0.0295676254,0.2045621872,0.04503702,0.1327867359,-0.0457830429,-0.3009677231,0.1279265136,0.0439737439,0.1046284512,-0.5376971364,-0.7024547458,0.2538052499,-0.1244832799,-0.0008702041,0.1646632403,0.0071904599,0.1146520227,0.0239121784,-0.6430913806,0.0776421204,-0.1291880608,0.1001469493,-0.4377544522,0.0315030515,0.2254792005,0.0922549739,0.0403308496,-0.2202931643,-0.1112453341,0.147397846,0.1006074846,0.364870429,0.0325328112,0.3112600744,-0.1857202947,0.607345283,0.2476839423,0.2405098528,0.19819507,0.2462696284,0.4493066967,-0.1849257499,0.005221433,-0.1256704628,-0.095861949,-0.0745300427,0.1621602178,0.1993874907,-0.148602888,0.1217992082,0.1278694421,-0.3149566054,0.1301389337,0.0233463552,-0.4966554642,-0.0734961778,0.0481187813,-0.1772634238,-0.1590229422,0.0179506186,-0.0380287319,0.0554810092,0.4832533896,0.0229955688,-0.3681592643,0.0419559106,-0.2726768255,0.0945149884,-0.1444545686,-0.0615256131,0.0473862886,0.1296505332,0.2777794898,0.1239323542,0.7186539769,-0.0076628155,0.0801283419,0.1951043159,0.2021508813,-0.3993442655,0.0818049163,0.0721869096,0.3692381978,-0.0813333765,0.3179687262,-0.1737516075,-0.0363956355,-0.3998003304,0.4534441531,-0.3241198361,-0.3078626096,-0.1848668605,-0.128023535,0.0372166634,-0.2122396678,-0.0612461343,-0.1030318961,-0.0534729175,-0.1018499807,-0.1828900129,-0.1325999051,-0.0106485924,-0.1577121019,0.5028899312,0.0727730095,0.1424803436,0.326443702,0.2844718993,0.1006738618,0.6179882884,-0.0507175028,0.0110502373,-0.0069833584,-0.1388685554,-0.0391130745,0.015824493,-0.2123361081,0.0428085439,-0.0525368862,-0.2650168836,0.1317896098,-0.0345823504,0.0665621534,-0.1058306322,-0.2407782525,-0.4946712852,0.4943987727,0.221894592,-0.2921310663,0.7134968638,-0.2678271532,-0.2052764148,0.1416470855,-0.2339219451,0.8169119954,-0.1017170474,0.0754900202,-0.0712855831,-0.2967706025,0.2919248343,-0.3984555602,-0.0409769565,-0.1412089616,-0.01109007,-0.045951996,-0.283377856,-0.0148655633,0.5497389436,0.0746046603,0.0041173017,-0.2025575042,0.0279839821,-0.0603116453,0.2044803351,0.1568813026,-0.1264695376,-0.0402072445,0.0246014148,-0.1684395522,-0.1137954518,0.0108401449,0.1596240848,0.1040772796,-0.302012831,0.0986558646,-0.050376337,0.0130163385,0.4422308803,-0.0722521618,-0.3880963922,-0.1589611173,0.4739687145,0.0072584176,-0.384429574,-0.0712875277,0.3765245974,0.2823236585,0.2021597624,-0.3082776666,-0.1891047359,-0.0379737206,-0.1022074372,-0.0399360918,0.1005175486,-0.3779098392,-0.5546373129,0.0911148265,0.4258129895,0.0130576147,0.0490886904,-0.2318656743,0.2566057444,0.0273942687,0.1178207472,0.0247645993,0.4632155895,0.0083302809,0.0266215671,-0.0653418452,-0.4270563722,0.0455710329,0.3402090967,0.4247439206,-0.1800312996,0.4000791311,-0.28071931,-0.178923592,-0.1061398238,0.1377500594,0.3976520002,-0.4959985316,0.1845720857,0.0446942933,0.303674221,-0.0690995976,-0.1082929373,0.2218023986,0.2455172688,-0.0082300669,-0.2096360028,-0.3218845427,0.0871958882,-0.2704619467,0.1277165264,0.2162760198,-0.7216903567,0.0634290203,0.3088502884,-0.1518501639,0.3722361326,0.2691301405,0.0742092878,-0.0098019177,0.0939171687,0.2346450239,-0.0647759661,-0.0302754752,-0.1650409847,-0.2098020315,-0.1362677068,0.018925678,0.2167364955,-0.0432213657,-0.0407196581,-0.2283444107,-0.2332425863,0.0817332119,-0.084475182,0.1314611286,0.0993550718,-0.0014922614,-0.061376337,0.3157812953,-0.2510995567,-0.5019800663,0.0555392876,-0.0692731962,0.2259491533,-0.0372875668,0.1839094162,0.0821005479,-0.0535609908,-0.2882064879,0.0313708298,0.1032677516,0.0403593294,0.0333200544,-0.3170485198,0.2241150141,0.1492220461,0.2395610511,0.2552303672,-0.2194753885,-0.0856135264,-0.027629178,0.037344683,0.0298782494,-0.2787132561,0.1930206269,0.1288041174,-0.31747365,-0.0012822878,0.3024915159,-0.1544518471,0.0009872118,0.0811040848,0.3936609328,-0.2308752835,0.3140845895,0.2147596478,0.1096523628,0.0043123499,0.4021009207,0.2612257004,0.0197194014,0.4292603135,0.0874666274,0.3169120252,0.5410943031,0.0506456047,-0.2002192438,-0.4718766809,0.0128363315,0.303082794,-0.4003796875,-0.0461870134,0.170055449,0.256385386,0.1450048238,-0.1513087749,-0.2486271858,-0.0435139537,-0.1459732652,-0.1100511402,0.0446665548,-0.2654961646,0.2061276585,0.2095094174,-0.055697374,0.1547969431,0.2726155818,0.0718443766,-0.2558670938,-0.0111093149,0.4257176816,0.1911603361,-0.028907204,-0.1679758132,0.0847853944,-0.1648778766,-0.0742945075,0.2118581235,-0.0026375635,-0.0213929787,0.167654261,0.2285775095,0.0593461916,0.270326674,-0.0494426787,-0.1172117367,0.3435636461,0.1106570661,-0.5399623513,0.0261542909,0.0732434094,-0.0348529369,0.2287302017,0.0645062178,0.2918296456,-0.175920859,-0.0906456783,-0.4421067536,-0.0463006385,-0.1719378382,0.2332713455,-0.1420274526,-0.1617720723,0.499633491,-0.0603867136,-0.1057106405,-0.3211374879,-0.0186657105,-0.2300728858,0.3697824478,0.1730164289,0.0391445719,-0.1572854221,-0.2795975506,-0.5812830925,0.0849058852,-0.2160646021,0.4904789031,-0.0278432313,0.3362504244,-0.1336788833,0.0438884459,0.4437274635,0.0589886308,0.1473845541,-0.0009261878,-0.1656163633,0.0090929018,0.3874020278,0.311599642,-0.1493280083,-0.338329941,0.05092315,0.1351896673,-0.026980944,0.0927214548,-0.1029011533,0.2029485106,0.5067508221,0.6720576286,0.1903526932,0.2355190665,0.0533221886,0.0357366912,-0.0773627311,-0.2588543296,0.10171175,0.1399802119,0.0133877303,0.1179542243,-0.4730008841,0.3262010217,-0.0933826119,0.0197413843,-0.1346081644,-0.1239301562,-0.2015507072,-0.1333200485,-0.3255666196,0.1312024593,0.1506091356,0.1228453964,0.0316739902,0.0998735428,0.0197815336,-0.0221258383,0.3603230715,-0.1492314637,-0.1052179933,0.2237624973,0.325286001,0.3091470897,-0.2567374408,-0.330399394,-0.0470994562,0.5866264701,-0.2299798578,-0.2018119693,0.004030982,-0.2191023082,0.1260613054,0.0399794988,0.6796936393,0.0515370928,-0.1524148732,-0.0491036549,-0.3140853345]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2195","title":"KeyError: '_indices_files' in `arrow_dataset.py`","comments":"Thanks for reporting @samsontmr.\r\n\r\nIt seems a backward compatibility issue...","body":"After pulling the latest master, I'm getting a crash when `load_from_disk` tries to load my local dataset.\r\n\r\nTrace:\r\n```\r\nTraceback (most recent call last):\r\n File \"load_data.py\", line 11, in \r\n dataset = load_from_disk(SRC)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 784, in load_from_disk\r\n return DatasetDict.load_from_disk(dataset_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/dataset_dict.py\", line 692, in load_from_disk\r\n dataset_dict[k] = Dataset.load_from_disk(dataset_dict_split_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 634, in load_from_disk\r\n if state[\"_indices_files\"]:\r\nKeyError: '_indices_files'\r\n```\r\n\r\nI believe this is the line causing the error since there may not be a `_indices_files` key in the older versions:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/b70141e3c5149430951773aaa0155555c5fb3e76\/src\/datasets\/arrow_dataset.py#L634\r\n\r\nMay I suggest using `state.get()` instead of directly indexing the dictionary?\r\n\r\n@lhoestq ","comment_length":10,"text":"KeyError: '_indices_files' in `arrow_dataset.py`\nAfter pulling the latest master, I'm getting a crash when `load_from_disk` tries to load my local dataset.\r\n\r\nTrace:\r\n```\r\nTraceback (most recent call last):\r\n File \"load_data.py\", line 11, in \r\n dataset = load_from_disk(SRC)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 784, in load_from_disk\r\n return DatasetDict.load_from_disk(dataset_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/dataset_dict.py\", line 692, in load_from_disk\r\n dataset_dict[k] = Dataset.load_from_disk(dataset_dict_split_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 634, in load_from_disk\r\n if state[\"_indices_files\"]:\r\nKeyError: '_indices_files'\r\n```\r\n\r\nI believe this is the line causing the error since there may not be a `_indices_files` key in the older versions:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/b70141e3c5149430951773aaa0155555c5fb3e76\/src\/datasets\/arrow_dataset.py#L634\r\n\r\nMay I suggest using `state.get()` instead of directly indexing the dictionary?\r\n\r\n@lhoestq \nThanks for reporting @samsontmr.\r\n\r\nIt seems a backward compatibility issue...","embeddings":[-0.3480816782,0.0301481821,-0.0518501848,0.687351048,-0.0778651163,0.1504861116,0.2006231993,0.5257608294,0.5055568814,0.1736582965,0.0381601229,0.1658373624,-0.3785415292,0.0226946659,-0.1071279943,0.0356825553,0.1004338786,0.282892406,-0.0493085869,-0.0393888056,-0.3141395152,0.1039332449,-0.1718684882,0.2657938004,-0.1816494316,0.1458103657,-0.0758230314,0.4838101268,-0.174485743,-0.6979006529,0.4852072299,-0.0263064764,0.190176487,0.565951705,-0.000124887,0.2735367715,0.3680529594,-0.0113450019,-0.2807968557,-0.299108237,-0.2398884445,-0.1354429275,0.3450046778,-0.1531854421,0.114507392,-0.4507660568,-0.3216248453,-0.3750584126,0.2612743378,0.0499283336,0.0946163759,0.3247491419,0.2301473171,-0.0989642441,0.2426403463,0.0146979121,0.1046656221,0.5075047016,0.2338429391,-0.2900749147,-0.099991709,0.1540052742,-0.1082652509,-0.0291712657,0.3468365967,0.0130410986,0.5942838192,-0.0651985556,-0.0802534074,0.1592507809,0.7142473459,-0.1073516756,-0.4755650163,-0.3687905967,-0.2322052419,-0.2909124792,0.2785750926,-0.3243991733,-0.2490376085,0.1780139953,0.2001084387,-0.2026610225,0.0134048583,0.3170470297,-0.2520062327,0.3052549064,0.0472134277,0.0036801184,0.6091058254,-0.061000485,0.1262109727,0.1920224577,-0.1498491764,0.0957965776,-0.1897277534,-0.041876059,0.1698119342,-0.3847656548,-0.2043846548,0.0489039756,-0.0418217592,0.2194772512,-0.1582194269,0.0171149075,0.629352808,0.260322839,0.1601459086,0.214642629,0.2806347311,0.1875303984,-0.0073226728,-0.0076226173,0.0274700113,-0.3619118035,0.0581549071,0.1007895991,0.3251579702,-0.2840910554,-0.0880516618,0.0626633167,-0.083983466,-0.2105324566,0.1065892428,0.3161305785,0.1580457836,0.5748215914,0.049242612,0.1754967123,-0.0530194975,0.3905651867,-0.0970358104,-0.2789994776,-0.0437029526,-0.0068811807,0.0663796216,-0.5058432817,0.2061230987,-0.1107687652,0.0559109449,0.0362657681,-0.1990457475,0.1008710414,0.2090242356,0.2979105115,-0.2459095865,0.1798747182,0.3856786788,-0.4363466799,-0.1850633323,0.1282753795,-0.2739574015,-0.6071556211,-0.3050348759,0.0792611912,0.1114556864,0.0256855655,-0.1950393021,-0.1063293368,0.0980099887,-0.0508782715,-0.0039300323,-0.0844877064,-0.1053850725,-0.1728766412,-0.0275600366,0.428765893,-0.5936437845,-0.0153986374,-0.263846904,-0.0973801911,-0.2817415297,0.2694525123,-0.4251615405,0.1883685291,-0.31475842,-0.1505842805,0.3692338765,-0.4750757217,-0.4828775823,0.3147622645,-0.0382291079,-0.1063909456,0.1441060752,-0.0863948464,0.1846064776,-0.0052482411,0.466676861,0.0335273594,0.0927503631,-0.1058450192,-0.0376387611,-0.1766355038,0.0593333989,0.332072556,0.0822735056,-0.1246179491,-0.0246208739,-0.1893011481,0.323769033,-0.0127456924,-0.0979578942,0.2707363665,0.2576822937,0.0074365577,0.0877735019,-0.0281076506,-0.4311747253,0.4764530659,0.1269202828,-0.1436425447,-0.3046075404,0.0576095879,-0.204906702,0.1430316567,-0.3838434815,0.1653743982,-0.0443980098,-0.1903958172,-0.0764399022,-0.0511777475,-0.3738670349,-0.2628928423,0.1099626943,0.2170253992,-0.2331440747,0.2441045195,-0.1233719364,-0.0302373338,-0.0408792719,0.1042789221,-0.0457759835,-0.2924667299,-0.1727135032,0.3905315995,0.2238410711,0.0658984408,-0.1758957952,0.0357518271,0.2583751082,0.0278369691,0.3661861718,0.3360883594,0.0165141542,-0.1566290259,0.1172321141,0.4590587914,-0.0945639461,0.3437943161,-0.0652319565,0.051303111,-0.0177963804,0.1191935018,0.1210875735,-0.5758320689,-0.0400060527,-0.1425597668,0.39732337,-0.0807479247,-0.0859616548,0.0228064544,0.2661430836,-0.1102821752,0.1839565784,0.0259291921,-0.1689128727,0.0070391516,0.0593138412,0.1893244088,0.2473946214,0.1097217202,-0.0139364824,0.1107767224,0.2238194346,-0.1325857192,0.3747676909,0.1114111021,0.3317106068,0.234944582,-0.2750403881,-0.1318517476,-0.2031774223,-0.0844189525,-0.1289204657,0.3822287917,-0.2688304484,0.0265973192,-0.4335649908,-0.2848981023,-0.1864100695,-0.3646925986,-0.2046783715,-0.4418205917,-0.0877743214,0.2054905146,-0.2337587923,0.0226692744,-0.2112057507,-0.0433308855,0.2200845331,-0.0792160109,0.0495815538,-0.3740925491,-0.2058417946,-0.0658296794,0.4681040645,-0.1801248044,-0.0266625676,-0.0118920337,0.0514297001,-0.069472827,-0.0585763492,0.1620065421,-0.1573341042,-0.0491236337,0.1702230126,0.2771518528,-0.0541904345,-0.2197736949,0.1480592042,-0.1676279604,-0.2757572234,0.1259573698,0.0824958831,0.1308053285,-0.0153264962,-0.1050438434,-0.256447643,-0.4464187324,0.1979564726,0.1310247481,0.349481076,0.3548307121,0.2260243446,0.2607572079,0.1880302131,-0.1089764163,-0.1276111752,-0.1270029694,0.0782293379,0.032040026,-0.1577864438,-0.1632757038,0.0306339953,0.0753113255,-0.1218411922,-0.5037189722,0.0332688689,0.0571271293,0.3437405527,-0.0155809047,0.1028240621,0.3205411434,0.2148469985,-0.1056693271,0.0830094665,-0.0750605091,-0.0415828452,0.1516057551,0.0661027282,0.1585153639,0.6146366596,-0.1701132655,0.957798779,-0.0745384395,0.0697051436,0.3080660701,0.0293767415,0.4093284309,-0.3344971538,-0.3089015782,-0.407135278,0.0184642132,-0.2617631853,-0.0429820903,-0.0400111675,-0.2466940284,-0.2750674784,-0.1290609688,-0.1997899711,-0.3002964556,0.0833064318,-0.44956249,-0.1304944605,-0.2399997264,0.3223652542,-0.0078917993,0.0547102317,0.2369860709,0.131813392,0.1603437662,-0.1751424521,-0.1319512576,-0.1989596635,-0.3255288303,0.393552959,0.122965306,0.3423699439,0.1506834626,-0.3161849976,-0.1449414939,-0.0195870623,0.7621179223,0.061246451,0.012730943,0.33651039,-0.1485461295,-0.7080937028,-0.1192511916,-0.0560953058,0.3735026717,-0.2649310231,0.6748162508,-0.0742974058,-0.3542921543,-0.1659762859,0.2461484522,-0.2924389839,-0.0550532602,-0.3419442475,0.0815438703,-0.3460504413,0.0639629066,-0.1591597199,0.1581749022,-0.0191009194,-0.0445058271,-0.0326955356,-0.235807687,-0.035015922,0.1593347341,0.4954830706,-0.0824304745,0.1458149552,0.4296375513,0.1442330629,0.3694687784,0.5261553526,-0.0331965685,0.0194161031,0.0734528229,0.0128910355,0.2490457296,0.4406180084,-0.0068795276,0.0693465769,0.2043870389,-0.2127927542,-0.1926091313,-0.4150504768,0.0652363449,-0.0488939546,-0.220489949,-0.3871481717,0.5498907566,0.0861166567,-0.1760122478,0.4220814407,0.281739831,-0.2140474468,0.6807106137,0.024600504,0.642595768,-0.1670988649,0.2859301567,0.248958379,0.0498712324,0.0153907416,0.3780663908,-0.0968997031,-0.4184154868,-0.068861708,-0.1497574598,-0.1061723009,0.0339443609,0.1142887175,0.0365409367,0.3101278543,-0.327786088,0.0529310256,-0.3975891471,0.0796739981,-0.1314951926,-0.1473574042,-0.5137034059,0.0027177772,0.2049192786,0.0761613324,-0.2041634917,-0.1107491031,-0.1203300953,-0.1795095652,-0.2387708426,-0.1850522757,-0.076307781,0.6282818317,-0.0168834571,-0.4393633604,-0.1686239243,0.209059462,0.0869143978,-0.139463529,-0.1496883035,-0.1534985453,-0.0662438348,0.0074201259,-0.0145891951,0.0308864266,0.4155854583,-0.0129757756,-0.3383813202,0.0680078045,-0.1483605206,-0.6900101304,0.1057152599,0.149383083,-0.2563326061,0.0320978798,-0.341440171,-0.1307484955,-0.0628225356,-0.0921189338,0.0061053522,0.1477670223,0.0940438658,-0.1980633885,0.0928475484,-0.1108570546,0.1095291078,0.698933661,0.1389992088,0.2271331251,0.619118154,0.1571891159,0.0089586256,-0.2093888074,0.1239051595,0.3681029379,-0.214083612,0.1730223149,-0.1477832198,-0.0965241492,0.2239040136,0.1674374342,0.4405491054,0.0770314187,-0.3431201577,-0.3073432744,-0.6235188842,0.5413686633,-0.074769944,0.1858645678,-0.3402197659,-0.3440027833,-0.2508101761,-0.2206686735,-0.1856478751,0.3617193103,0.0838924348,-0.0288724694,-0.03655269,0.0107049784,0.1918310523,-0.2414353937,0.0198692307,-0.0943368077,-0.2587106526,-0.0766820163,0.0562419482,0.1737300307,0.0901723281,-0.217159763,0.0039293738,-0.3660217226,-0.067943871,-0.0450003073,0.2554685473,0.0824197531,0.0866506845,0.1789336652,0.4045010507,0.1148699,-0.0660946742,0.385858953,-0.0267959163,0.368645072,0.019213846,0.1183529124,-0.2701329291,-0.1279647648,-0.2024184614,0.1645312756,-0.3105652332,0.1701408178,0.3367153108,-0.3421877325,-0.1756334305,0.4158471227,0.5223764777,0.4512340128,-0.122582972,0.0779046416,0.1465541124,0.1080762669,-0.07864362,-0.3219598234,-0.0935813338,0.0375579409,-0.1394735277,0.1992657185,0.2288283259,-0.2687583268,-0.0704419389,0.2031307518,0.0198372621,0.180831641,0.2249736935,0.6219126582,0.0738738477,-0.1563950628,0.2886601985,0.1099570915,0.5546107888,0.4484431744,-0.2012695223,0.0436727181,0.1578819752,0.0959322453,-0.0405788831,-0.5247774124,0.3907715678,0.1752471924,-0.1346987337,-0.0889799073,0.2275410444,0.6251251101,-0.2984883189,-0.0718875527,-0.1815495938,-0.136833936,-0.2062932402,-0.1555327475,0.1915982217,-0.2265865952,0.0818603337,0.043541044,-0.2121332139,-0.174438104,-0.0359164849,0.0084872255,-0.0996171311,-0.1674369425,-0.1139471605,0.2915246487,0.2020365,-0.155770421,-0.0815690532,0.3587991297,0.0282530673,0.1051047742,0.3302351236,0.4493142068,0.196667105,-0.5093146563,0.0844841823,0.1356437802,-0.1825339049,-0.0663520545,-0.1588062197,0.038896028,0.1339813471,0.2705523968,0.0723135695,-0.0030160444,0.015240252,0.250774622,0.4278231859,-0.0579017326,-0.036996834,-0.0296438895,-0.1185248271,0.1220521778,0.1367431879,-0.2676424384,-0.1819089204,0.3129284978,0.1661720276,0.1299787611,0.1005945206,0.020352317,0.0084774671,0.5679678917,0.511331737,-0.0583874434,0.0061489358,-0.015576602,-0.7016690969,-0.0853484273,0.0421518236,0.3114493787,0.3209134936,0.1056280285,0.3163285553,0.0884422585,0.2337159216,0.0014980639,0.0591553375,-0.0384278856,-0.2517872155,-0.2747831941,0.1708573848,-0.1513152868,-0.160095796,-0.3507173955,0.0556477904,-0.1797157824,-0.0180437881,-0.1010022685,0.094534941,-0.1381376535,0.0486550778,0.5888919234,-0.1210164428,0.1094570011,0.1548878551,-0.2974712849,-0.1896977276,-0.2960760593,-0.1095006838,0.2015025765,-0.0242954064,0.474952668,-0.2087959945,-0.0462660827,-0.2998072803,0.314171046,-0.0397408679,-0.0403813384,-0.4732936919,0.1886202097,-0.0997236744,-0.2107943445,0.1146276966,0.2282400578,0.1070369929,0.1430939883,-0.1357293576,-0.1627197862,0.4633956254,-0.2617483437,0.0302433502,0.004329782,0.1836889982,0.2498479486,-0.0837386772,-0.5364425778,0.0102361236,0.2373892367,-0.0949755013,-0.2901373506,-0.0906274021,0.0212726425,0.2023250759,-0.1543517113,0.2582116425,-0.0602571033,-0.2688776255,-0.1068650112,-0.1343672425]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2195","title":"KeyError: '_indices_files' in `arrow_dataset.py`","comments":"Thanks @samsontmr this should be fixed on master now\r\n\r\nFeel free to reopen if you're still having issues","body":"After pulling the latest master, I'm getting a crash when `load_from_disk` tries to load my local dataset.\r\n\r\nTrace:\r\n```\r\nTraceback (most recent call last):\r\n File \"load_data.py\", line 11, in \r\n dataset = load_from_disk(SRC)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 784, in load_from_disk\r\n return DatasetDict.load_from_disk(dataset_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/dataset_dict.py\", line 692, in load_from_disk\r\n dataset_dict[k] = Dataset.load_from_disk(dataset_dict_split_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 634, in load_from_disk\r\n if state[\"_indices_files\"]:\r\nKeyError: '_indices_files'\r\n```\r\n\r\nI believe this is the line causing the error since there may not be a `_indices_files` key in the older versions:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/b70141e3c5149430951773aaa0155555c5fb3e76\/src\/datasets\/arrow_dataset.py#L634\r\n\r\nMay I suggest using `state.get()` instead of directly indexing the dictionary?\r\n\r\n@lhoestq ","comment_length":18,"text":"KeyError: '_indices_files' in `arrow_dataset.py`\nAfter pulling the latest master, I'm getting a crash when `load_from_disk` tries to load my local dataset.\r\n\r\nTrace:\r\n```\r\nTraceback (most recent call last):\r\n File \"load_data.py\", line 11, in \r\n dataset = load_from_disk(SRC)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 784, in load_from_disk\r\n return DatasetDict.load_from_disk(dataset_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/dataset_dict.py\", line 692, in load_from_disk\r\n dataset_dict[k] = Dataset.load_from_disk(dataset_dict_split_path, fs, keep_in_memory=keep_in_memory)\r\n File \"\/opt\/conda\/envs\/py38\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 634, in load_from_disk\r\n if state[\"_indices_files\"]:\r\nKeyError: '_indices_files'\r\n```\r\n\r\nI believe this is the line causing the error since there may not be a `_indices_files` key in the older versions:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/b70141e3c5149430951773aaa0155555c5fb3e76\/src\/datasets\/arrow_dataset.py#L634\r\n\r\nMay I suggest using `state.get()` instead of directly indexing the dictionary?\r\n\r\n@lhoestq \nThanks @samsontmr this should be fixed on master now\r\n\r\nFeel free to reopen if you're still having issues","embeddings":[-0.3437005281,0.0625090078,-0.0619883277,0.6979857087,-0.0762992576,0.1504234225,0.1940686256,0.4931738079,0.5131545663,0.1480900198,0.0192217361,0.1839222014,-0.3918602765,0.0680304319,-0.1045260802,0.0790594593,0.0853407234,0.2412601411,-0.1544322819,-0.052290827,-0.2806541324,0.1007366776,-0.1309833378,0.211374253,-0.2218931764,0.1231649667,-0.0895246267,0.4890262485,-0.1853998005,-0.7146970034,0.4884228408,-0.0012378489,0.198797375,0.6145881414,-0.0001211874,0.2563541532,0.328388989,0.0076413802,-0.3190935552,-0.3326137662,-0.2397470772,-0.1572865993,0.3840711117,-0.1370294094,0.1061333418,-0.4474045634,-0.3572835922,-0.3314718306,0.2223311365,0.0783566535,0.1171720922,0.3026997149,0.2074174583,-0.1436059624,0.228953734,0.0077808006,0.1049866378,0.5003935099,0.1849455237,-0.2934077084,-0.115201056,0.1408634186,-0.1057001576,-0.0338394083,0.3557126224,-0.0116487294,0.5778528452,-0.1533564031,-0.0614438169,0.1703200489,0.7022152543,-0.145922184,-0.4514220357,-0.3354025185,-0.1894246638,-0.2096370906,0.2590704858,-0.2555756569,-0.2330313176,0.1550564617,0.1738686711,-0.1955357045,0.056436453,0.2905206382,-0.2734226584,0.3019711673,0.0345148742,-0.030322887,0.6314465404,-0.0803371742,0.0911605954,0.1656604558,-0.1702350974,0.0807329342,-0.2672142982,-0.038882982,0.1909446716,-0.3841376007,-0.2094092965,0.0528946966,-0.0255248528,0.1885953397,-0.1497028768,0.0399960838,0.6687847376,0.2489634603,0.1215692312,0.2297081053,0.2717272341,0.2171474695,-0.0350570753,0.0183774922,0.0184037704,-0.347192049,0.0448060893,0.119112514,0.3113000393,-0.3126894236,-0.0694165379,0.0593046881,-0.0624833927,-0.201540783,0.1370553374,0.3150915802,0.1558444947,0.5801064968,0.0374670699,0.1970073879,-0.0338273272,0.3713238239,-0.1358800232,-0.2581335306,-0.0995747894,-0.0013663978,0.1081545055,-0.5095399022,0.2359856814,-0.047541324,0.0749816075,0.0100532221,-0.1632342041,0.084071897,0.1596419215,0.3046089411,-0.1820002645,0.2101746202,0.3970351517,-0.3919407427,-0.1523308307,0.1238470376,-0.2913793623,-0.5782806873,-0.3056536615,0.1312686503,0.1295812875,0.0252597015,-0.2260646373,-0.1392920017,0.1114809364,0.0196710825,-0.0214505251,-0.0969825238,-0.0833036155,-0.1675645262,0.0159698743,0.3993794322,-0.5810675621,-0.0023690723,-0.2711961567,-0.1212873608,-0.2390404493,0.2816366851,-0.4063346982,0.2037551999,-0.3347817361,-0.1348029077,0.3043564558,-0.4426818788,-0.4974357784,0.3520651162,-0.0488600992,-0.1117116734,0.1396108866,-0.1120444015,0.1941486746,-0.0449751951,0.4687681496,0.0551270992,0.0712261647,-0.0960199609,-0.0715619698,-0.1324856579,0.03400857,0.3673783243,0.0757355466,-0.1486506313,-0.0234469827,-0.1910144389,0.3385557532,-0.033754088,-0.0971989632,0.31829229,0.2205266505,-0.0087158261,0.0638513342,-0.0000374727,-0.4078004658,0.4620512724,0.1059394255,-0.0443780683,-0.3058128953,0.0868807361,-0.2338584512,0.1178293079,-0.3985089958,0.0907850042,0.0071391193,-0.1555865258,-0.1538615823,-0.1076429859,-0.3897000253,-0.2656809688,0.0552962534,0.2186610997,-0.1987503469,0.2255682945,-0.1129231304,-0.0787864849,-0.0303891562,0.1103976592,-0.0472365431,-0.2692546546,-0.1994897723,0.3676323295,0.2034950703,0.0388879552,-0.1302076131,0.043396201,0.2617529631,-0.0059622438,0.3687925935,0.3228161037,0.0084586469,-0.1577733159,0.0768168792,0.4362381697,-0.1849120855,0.3767441213,-0.0632533655,0.0499109142,-0.0009461989,0.0831125975,0.1180704609,-0.5903175473,-0.0254220888,-0.1207477748,0.3981439769,-0.0711293966,-0.0318615586,0.011218044,0.2276003212,-0.0984686315,0.1662933677,0.001818024,-0.1907745749,0.0243224613,0.0431337841,0.2312740833,0.1958069205,0.137964353,-0.0030731107,0.0805642009,0.1951951087,-0.104629606,0.3564743698,0.0686343387,0.2952162921,0.2632842064,-0.2636420131,-0.1510628462,-0.2133145332,-0.0609140843,-0.1362898201,0.3755882978,-0.2155833691,-0.0201139748,-0.4201763272,-0.2314468026,-0.1256988645,-0.3376940191,-0.1507581621,-0.436881274,-0.088378273,0.2107362598,-0.2474728227,0.0550901107,-0.197035864,-0.1132625863,0.1789672524,-0.0340482183,0.060162805,-0.3995396197,-0.2059826851,-0.0277684163,0.496370852,-0.1698491424,-0.0224506948,0.0206515752,0.0296302382,-0.0853550285,-0.0005940204,0.1873171329,-0.1565771103,-0.0260392744,0.1472346932,0.3131296933,-0.0509984456,-0.209030956,0.1667183638,-0.1248663589,-0.2623379827,0.1546895355,0.0747201294,0.1215401292,-0.0419443063,-0.1364700198,-0.2812353671,-0.4659274518,0.1132291034,0.1329471916,0.3265402317,0.305771172,0.2569748461,0.2501551211,0.1969927549,-0.1252462119,-0.1024906486,-0.1609047353,0.0975885391,0.0345190391,-0.1683220714,-0.147705704,0.0333468765,0.1063582823,-0.1455202252,-0.5225371718,0.0735288337,0.0396972634,0.3337232172,-0.0432079136,0.1500978172,0.2738485038,0.2244729549,-0.1346719116,0.0823900551,-0.0716342703,-0.0528472476,0.1690415889,0.0709919184,0.0759250894,0.6219985485,-0.1741819829,0.9047240615,-0.0850608423,0.0983518362,0.3507145047,0.027119277,0.3714459538,-0.3195245266,-0.3450508714,-0.3971735537,-0.0128928469,-0.2690455914,-0.02587674,-0.0536878593,-0.2634806633,-0.2424750775,-0.106241934,-0.2012027651,-0.267886579,0.0891782492,-0.4607056379,-0.1519239247,-0.2918173969,0.2894775271,0.0338952914,0.0784962624,0.2496539354,0.0718304664,0.163762331,-0.1985615045,-0.1715665162,-0.2360578775,-0.302815944,0.3233646154,0.0326958373,0.3189825416,0.1509532183,-0.278213948,-0.1228729784,-0.027419949,0.7755700946,0.0694629252,-0.0135410577,0.3416607082,-0.1828559637,-0.6849226356,-0.1430061609,-0.076991722,0.3777199984,-0.2822335958,0.6447867751,-0.1052529067,-0.3734832406,-0.1347847879,0.2421916425,-0.2975417674,-0.0471377745,-0.3241860569,0.0681183115,-0.361435473,0.0424428061,-0.1753165722,0.1782323122,-0.0182590056,-0.040976651,-0.023354765,-0.2156773657,-0.0189865213,0.1708407104,0.5178511143,-0.0917809978,0.1552842557,0.3910789192,0.151252225,0.3668096066,0.5727211237,-0.052418042,0.0227869544,0.1233680993,0.0127731636,0.2191072702,0.3768678308,0.0017909455,0.0071477345,0.1853459477,-0.1399025172,-0.2090144455,-0.3456370234,0.0707189366,-0.072574541,-0.2323499322,-0.3987552822,0.5538119674,0.0757332072,-0.1451486647,0.4524231255,0.2212210745,-0.2190643996,0.7223880291,0.0420392677,0.5834644437,-0.1472351551,0.303397119,0.2070568055,0.0408477113,-0.0204752106,0.2904047966,-0.0836610943,-0.4212234318,-0.0757740512,-0.1213997304,-0.099199295,-0.0378701501,0.1712074131,0.0080721928,0.308340162,-0.3715083897,0.0482658409,-0.4021244645,0.1216111481,-0.1254554391,-0.1721249521,-0.4360770881,0.0483664833,0.2044882625,0.0633208379,-0.2060789913,-0.1036889255,-0.0989707336,-0.1534956694,-0.2596629858,-0.1319444925,-0.1039303839,0.5987221599,-0.0494348295,-0.4428333342,-0.1699003279,0.1780454814,0.0864562243,-0.1195992306,-0.1258827001,-0.1075543389,-0.0314495042,0.041904483,-0.040888641,0.0695694536,0.4228502512,-0.0175479762,-0.3506642282,0.0891879648,-0.1480855495,-0.6180958152,0.1087119579,0.1479555666,-0.2532766759,-0.0102044772,-0.3166955113,-0.1118125692,-0.0347662643,-0.1134832501,0.0442433394,0.1400861293,0.0375358015,-0.1662015021,0.0954251289,-0.108862251,0.0871869102,0.7104115486,0.1615451425,0.225993067,0.6737828255,0.1651811302,0.0447911955,-0.2560970485,0.1196541563,0.3818442225,-0.2010110468,0.2045415789,-0.1781062931,-0.1663603485,0.2491137236,0.1407563984,0.4852972329,0.1138321757,-0.3393975496,-0.3412427902,-0.6491513848,0.4758007824,-0.0777494833,0.1905664206,-0.3575003743,-0.3366330266,-0.2618661523,-0.1594820321,-0.2298216224,0.348959893,0.0203115139,-0.0432253778,-0.0170165729,0.0294098407,0.1890240312,-0.2335283309,0.0664042011,-0.121415779,-0.2289002985,-0.126657322,0.0812334418,0.1456775367,0.0704901442,-0.1864243746,-0.0125319399,-0.3193384707,-0.0384719484,-0.0330956914,0.282033354,0.0154565927,0.1098285019,0.1792058796,0.4029449821,0.0735877082,-0.1137402505,0.3437511027,-0.0137062296,0.2762866914,0.0530834049,0.0968638808,-0.2084213048,-0.1503098458,-0.1973834038,0.1348847747,-0.3143283129,0.2006953061,0.3316076696,-0.3723166585,-0.1694096625,0.4137821794,0.5258206129,0.4191300273,-0.1160741523,0.0613585413,0.1539830118,0.1574775577,-0.0676562563,-0.2825668454,-0.0349910706,0.0515097827,-0.1126385704,0.1593417972,0.2207184285,-0.2603266239,-0.0983498991,0.2220219672,0.009967898,0.1790398955,0.2652374506,0.6549831033,0.047775384,-0.1828883737,0.2611658275,0.0806480572,0.5296794176,0.5002399087,-0.1772082299,0.0948737338,0.147034362,0.058011502,-0.0320027955,-0.5207428932,0.3944750726,0.16123043,-0.1278056651,-0.1419652402,0.1691355705,0.5864527225,-0.2946018875,-0.1435379833,-0.1679213196,-0.1049019173,-0.2334730923,-0.0906336159,0.0862782374,-0.2682772875,0.0853686705,0.0653264821,-0.201841563,-0.1600526422,-0.0437693596,0.034818694,-0.1453925073,-0.1557482034,-0.1444471329,0.3033106327,0.1579716653,-0.1481506675,-0.0349099897,0.384344995,0.0172674488,0.0584639721,0.3044855595,0.4444730282,0.265829891,-0.5644999743,0.0691369623,0.1930319518,-0.2052133977,-0.0465575904,-0.1699369401,0.0112724286,0.148055777,0.2356417477,0.1047225595,-0.0508823693,0.0374308154,0.2313921452,0.3739520609,-0.0057793828,-0.0465185903,0.0601953715,-0.1229652837,0.1386374235,0.1046861261,-0.2637210786,-0.1810970008,0.40951249,0.1460425407,0.131470412,0.093633689,0.049232997,-0.0019208867,0.6029649973,0.5017882586,-0.0730915442,-0.063491255,-0.020015236,-0.6808801889,-0.0904058665,0.0202098954,0.2912875712,0.3152634501,0.1324048787,0.3313698173,0.1017869487,0.2425287068,0.0359325707,0.086395286,0.0192449205,-0.3052066863,-0.2474491596,0.1221120656,-0.0869731158,-0.1743315309,-0.3412002325,0.0169100184,-0.1808812469,0.0343128443,-0.0641596839,0.0510385819,-0.0993934199,0.0412078351,0.5967946053,-0.1437921971,0.0915654376,0.1601386964,-0.296128124,-0.1572458148,-0.3000837862,-0.1445714384,0.1659157425,0.0093030231,0.488232255,-0.1628598869,-0.0143571785,-0.3650959134,0.3171492815,-0.0384442732,-0.0233008657,-0.4206443429,0.1915092766,-0.0203533415,-0.2299865037,0.1428321302,0.2392257452,0.104466401,0.1673978269,-0.1479468346,-0.1747443378,0.4662326574,-0.2733976245,0.0169343352,0.0283580888,0.2047582269,0.1471784413,-0.0636845231,-0.4889203906,0.0350649022,0.2195564061,-0.0950229242,-0.3031611741,-0.0935754329,0.1114848256,0.2328037173,-0.1336058229,0.2096349001,-0.0311653856,-0.2814044356,-0.139311552,-0.1226979941]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2194","title":"py3.7: TypeError: can't pickle _LazyModule objects","comments":"\r\nThis wasn't a `datasets` problem, but `transformers`' and it was solved here https:\/\/github.com\/huggingface\/transformers\/pull\/11168\r\n","body":"While this works fine with py3.8, under py3.7, with a totally new conda env and transformers install:\r\n\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/transformers\r\ncd transformers\r\npip install -e .[testing]\r\n\r\nexport BS=1; rm -rf \/tmp\/test-clm; PYTHONPATH=src USE_TF=0 CUDA_VISIBLE_DEVICES=0 python \\\r\nexamples\/language-modeling\/run_clm.py --model_name_or_path distilgpt2 --dataset_name wikitext \\\r\n--dataset_config_name wikitext-2-raw-v1 --do_train --max_train_samples 1 \\\r\n--per_device_train_batch_size $BS --output_dir \/tmp\/test-clm --block_size 128 --logging_steps 1 \\\r\n--fp16\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_clm.py\", line 453, in \r\n main()\r\n File \"examples\/language-modeling\/run_clm.py\", line 336, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 303, in \r\n for k, dataset in self.items()\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1259, in map\r\n update_data=update_data,\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 157, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 389, in dumps\r\n dump(obj, file)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 361, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 556, in save_function\r\n obj=obj,\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 524, in save\r\n rv = reduce(self.proto)\r\nTypeError: can't pickle _LazyModule objects\r\n```\r\n```\r\n$ python --version\r\nPython 3.7.4\r\n\r\n$ python -m torch.utils.collect_env\r\nCollecting environment information...\r\nPyTorch version: 1.8.0.dev20210110+cu110\r\nIs debug build: False\r\nCUDA used to build PyTorch: 11.0\r\nROCM used to build PyTorch: N\/A\r\n\r\nOS: Ubuntu 20.04.2 LTS (x86_64)\r\nGCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0\r\nClang version: 10.0.0-4ubuntu1 \r\nCMake version: version 3.16.3\r\n```\r\n\r\nThanks.","comment_length":13,"text":"py3.7: TypeError: can't pickle _LazyModule objects\nWhile this works fine with py3.8, under py3.7, with a totally new conda env and transformers install:\r\n\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/transformers\r\ncd transformers\r\npip install -e .[testing]\r\n\r\nexport BS=1; rm -rf \/tmp\/test-clm; PYTHONPATH=src USE_TF=0 CUDA_VISIBLE_DEVICES=0 python \\\r\nexamples\/language-modeling\/run_clm.py --model_name_or_path distilgpt2 --dataset_name wikitext \\\r\n--dataset_config_name wikitext-2-raw-v1 --do_train --max_train_samples 1 \\\r\n--per_device_train_batch_size $BS --output_dir \/tmp\/test-clm --block_size 128 --logging_steps 1 \\\r\n--fp16\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_clm.py\", line 453, in \r\n main()\r\n File \"examples\/language-modeling\/run_clm.py\", line 336, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 303, in \r\n for k, dataset in self.items()\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1259, in map\r\n update_data=update_data,\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 157, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 389, in dumps\r\n dump(obj, file)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 361, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 556, in save_function\r\n obj=obj,\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/home\/stas\/anaconda3\/lib\/python3.7\/pickle.py\", line 524, in save\r\n rv = reduce(self.proto)\r\nTypeError: can't pickle _LazyModule objects\r\n```\r\n```\r\n$ python --version\r\nPython 3.7.4\r\n\r\n$ python -m torch.utils.collect_env\r\nCollecting environment information...\r\nPyTorch version: 1.8.0.dev20210110+cu110\r\nIs debug build: False\r\nCUDA used to build PyTorch: 11.0\r\nROCM used to build PyTorch: N\/A\r\n\r\nOS: Ubuntu 20.04.2 LTS (x86_64)\r\nGCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0\r\nClang version: 10.0.0-4ubuntu1 \r\nCMake version: version 3.16.3\r\n```\r\n\r\nThanks.\n\r\nThis wasn't a `datasets` problem, but `transformers`' and it was solved here https:\/\/github.com\/huggingface\/transformers\/pull\/11168\r\n","embeddings":[-0.1246638298,-0.2800029218,0.0102392863,0.0428071246,0.2114832401,-0.1000178456,0.4270269573,0.4956005216,-0.1175258905,-0.0975463167,-0.2169504166,0.6081893444,-0.2419368625,-0.2057418674,-0.2879943848,-0.1205900386,0.3193874657,0.1881065667,-0.2526328266,-0.2084699571,-0.1915359497,0.0964935049,-0.2404561043,-0.0917270407,-0.0389703438,0.0167094544,0.0361307412,0.0273612104,-0.100044407,-0.1091980487,0.484229058,0.1827317327,0.1594474912,0.4119527936,-0.0001136487,0.2710008621,0.1944552362,-0.0705220699,-0.3275661767,-0.06862735,-0.0967073143,-0.1153661013,0.1716420949,-0.1212027445,-0.0454877652,0.1072109267,-0.1677041501,-0.2127361745,0.3663665652,0.2743563056,0.2295114398,0.5361444354,-0.0980848595,-0.1675886661,0.1126614437,0.1677086651,-0.3499956727,0.1804063171,0.1615159512,0.0121087553,-0.1613077521,0.1032232419,0.0931350961,0.1365531534,-0.0104408683,0.0491099581,0.4215169847,-0.1219776273,0.1030768901,0.090032205,0.0869783759,-0.1059741154,-0.3644068539,-0.2949759066,-0.2756364346,-0.2603702247,-0.1180811226,-0.0281845499,-0.0963396952,0.2262700349,-0.1358830929,0.0184898172,0.0610882193,0.1370189488,0.1712328196,0.6562994123,-0.1202367768,0.1605676115,0.3113485575,0.0108865704,0.3688592613,0.1282702833,-0.1315863431,-0.0367902741,-0.0894225314,-0.2164620012,0.1220557168,-0.5101084709,-0.0777223334,-0.0080803595,-0.2572022676,0.1492491066,0.0274086967,0.1489541829,0.0482668579,0.3884806633,-0.1631519347,0.5761070848,0.2375507653,0.3265770078,0.1539948732,0.0540750511,-0.1017149165,-0.1881777346,0.0487313643,0.1151105314,0.0031679545,0.0475774519,-0.2759521604,-0.2183811814,0.339073807,0.1376144439,0.1156681851,0.3977783024,-0.1240018904,0.383954674,-0.0970213413,0.4125422537,-0.2263994515,0.1417227387,-0.1449755281,-0.0028276665,-0.0515903346,-0.2751283944,0.1202233806,0.3911178112,0.3049985766,-0.1079033315,-0.0449499749,0.0925625935,0.0728266984,-0.1927007586,-0.0330227166,0.3256643713,-0.2237992734,0.1721329987,0.5163032413,-0.4574613571,-0.2914823294,0.0910478979,-0.1048493609,-0.3494744897,-0.2063051164,0.1886192113,-0.0622352064,-0.0136821326,-0.0449214317,-0.127797544,0.2027472109,0.0558203347,-0.1022598743,-0.3338252008,-0.1498707682,-0.2585225999,0.1845505387,0.2395551056,-0.0339859724,-0.004735311,-0.3707762957,-0.0680421516,0.2212658077,0.0701747835,-0.2528327703,0.1884349585,-0.1253964901,-0.2098996043,0.263335973,-0.3431504667,-0.2949306667,-0.1351091862,0.1275581717,0.2736187279,-0.0262450427,-0.0961856768,0.0784693584,-0.1329758167,0.2919353247,0.2968382239,0.2626328468,-0.0166477859,-0.205741778,-0.5603092909,-0.0651898682,0.0512328707,0.3892809749,0.0931745395,-0.1049410179,-0.029855065,0.2972363234,-0.0911482871,0.1596625745,0.0928587914,0.6337450743,-0.2758597136,0.0112541998,0.1283127069,-0.6093167663,0.4746496081,0.1405604631,0.0243730284,-0.4269521534,0.10572128,-0.1153522357,0.1183628365,-0.4068238139,-0.3231731355,0.182220757,0.023026282,0.287913084,-0.1566914618,-0.0593380816,0.0663567781,0.0312237572,0.1007382274,0.0313730314,0.2697269022,-0.3124516904,-0.3386402726,-0.1074113548,0.1215039715,0.0859875754,-0.2611347437,-0.1384667307,0.0874169916,0.2593746781,-0.1114789397,-0.1742797643,0.1218882799,-0.0089789964,-0.054557059,0.1755213141,0.2047269344,0.0431650952,-0.0604723282,0.0899378434,0.242177695,0.0638434589,0.1741407365,0.1836080104,0.0880301967,0.219494164,-0.0424667709,0.3108650148,-0.3795113862,-0.1042753384,0.372900635,0.0313055925,-0.0523641184,0.1158421338,0.1402218789,0.4219222665,0.2693614662,0.0741915256,0.3096559644,-0.186752826,0.1333236098,0.0536410585,0.3026951253,0.3619996011,-0.0128942635,-0.0204405021,0.0284403097,0.0621747524,0.0847767517,0.1326806396,0.2454260737,0.5366300941,0.1241473034,0.0652650818,-0.0473853238,-0.1002717763,-0.4163165092,0.0097540822,0.2883231044,-0.0141193513,0.1562461853,-0.1763306707,0.4993186593,-0.1799680442,-0.2788597047,-0.0818994045,-0.279690057,-0.1846143007,0.3273070157,-0.2546777129,0.1746878028,-0.300951004,-0.031711556,0.1652612686,-0.1664562225,-0.0035739974,-0.1900844425,-0.3132197857,0.0044088508,0.0848389938,-0.2894690633,-0.1206057668,-0.2443773001,0.018369453,0.0341008157,-0.4039056003,0.1061803475,-0.0126184635,0.4389199615,0.0581261292,0.2609312832,0.0963692069,-0.4702159762,0.2330487818,-0.1071018055,0.0378948562,0.095486261,0.0428105071,-0.0555307232,-0.287247628,-0.2369063646,-0.432485044,-0.4679963589,0.2045193762,0.1448124945,0.0825234056,0.5685711503,0.4462990463,0.149677217,0.1759720892,0.3646390736,0.0516808182,0.0751727894,0.2222380042,-0.0071129678,-0.2227340192,-0.3978248835,-0.4485785961,0.2483640909,-0.1797801703,-0.2609122992,0.0856868848,-0.4601543844,0.4757069051,0.0603173152,-0.0055412501,0.4046200514,0.2869279087,-0.0960220546,-0.195090875,-0.0604257621,0.0840654001,0.1817073673,0.3081721365,-0.0606533773,-0.0371663831,0.0885410383,0.8457784057,-0.0625767782,-0.3090642393,0.3011673987,0.3228740394,0.27294752,-0.0641876236,-0.2895969748,-0.1856785417,-0.2493606657,-0.0598720014,0.1364802569,0.0215544868,-0.3702583313,-0.0289999228,0.1570182294,-0.0462725013,-0.312672019,0.2491835207,0.0102476068,-0.177133292,0.0849898607,0.244756937,-0.2118858546,0.2095586956,0.0581699982,0.0650189519,-0.2876721025,-0.1586962789,0.300427705,-0.1297270656,-0.3789155483,0.1120494008,-0.0692745298,0.4720147252,0.1006338745,-0.4139404595,-0.2330119312,-0.0538394563,0.5441067219,-0.1220101565,-0.0868830904,0.3643039167,0.0728127658,-0.2709541917,-0.4917675853,-0.0838948339,-0.1540626287,0.7009064555,0.6861712337,-0.2182397395,-0.1529875845,-0.1375271827,0.1658508629,-0.0912574902,-0.0068578008,-0.3302050531,-0.2411468625,-0.3341560364,0.0093185455,0.1262239069,0.1666361392,-0.2387672961,-0.1119827926,-0.1109151021,-0.2365115285,-0.1636469364,0.2372399867,0.1029471606,-0.0714579076,0.2733432651,-0.1276386529,0.1258046776,-0.1808055192,0.2268607169,-0.1409585774,-0.2824903727,-0.0307270139,0.0622915886,0.2199910134,0.0986911431,0.055458691,0.2082612216,0.032189358,0.2097079009,0.1064751148,0.1325461715,0.2600807548,-0.0244177915,-0.2099417001,-0.2663866282,0.2285377681,-0.0081828991,-0.0247436725,0.0827490836,-0.2503434122,-0.2823473811,0.5823017955,0.1690817773,0.6868520975,0.0852042958,0.0475268178,0.3833718598,-0.0521970168,0.2409619987,-0.1178976148,0.2398637235,-0.3718243539,-0.3116846681,0.1513225883,-0.2429838181,0.1880078763,-0.270113498,-0.3233343065,0.1445527971,-0.3372011781,-0.1086111888,0.0764989704,-0.0708046779,0.2236171514,-0.227024734,0.1372250468,0.0892493725,0.042409528,0.3420085013,-0.0152181145,-0.161697045,-0.4075741172,-0.3274033368,-0.2037391514,0.1840066314,0.2488791049,0.6642631888,0.42277354,-0.119093284,0.1431236714,-0.2073970139,0.2056651264,-0.0087762177,-0.1615579873,0.1751471311,0.3712835908,-0.069802627,0.1066500321,-0.1801020503,0.4022938907,0.0204899535,-0.2296086848,-0.0510105006,-0.2294132262,-0.1136185378,0.30971995,-0.1394521892,0.3802990615,-0.0715741292,-0.1588303596,0.126853466,-0.2669717669,-0.3018001318,0.1626711488,-0.0498556346,-0.3256384432,0.2904969752,-0.0575344488,-0.1602350622,-0.0252050068,0.2240256965,0.2699185014,0.3075921237,0.5213400126,0.1849194169,-0.174479872,-0.1993798465,-0.2696540654,-0.2202759981,-0.5101467371,0.1317055672,-0.1350959688,-0.3405151665,-0.1299384832,0.3094701767,0.163970843,0.1065331176,-0.3221516311,-0.1614484936,-0.4098920226,-0.0104663102,-0.2776601613,0.1512472928,-0.3070999086,0.0470277444,-0.0198470596,0.1643857807,-0.2827276289,-0.1477009952,-0.4392418265,0.0497462451,-0.3415777385,-0.5661296248,0.163542375,0.0938234329,0.1375854015,0.0936436653,-0.3022043407,-0.1747609973,-0.3933849931,0.1003511623,0.103014715,-0.1769881099,0.1329164803,0.1678612083,-0.0250160862,-0.0924355984,-0.0806395933,-0.2091103643,-0.0869269222,0.0664323345,0.1076698005,0.1641331762,0.1170005202,-0.0290333256,-0.0086802486,0.1577731967,0.1494337618,0.2259128094,-0.178435728,0.1162505224,-0.0568199083,-0.1012946963,-0.0642998442,-0.0355465785,0.378677696,-0.2920702994,0.046568688,-0.3024606407,0.1472044289,0.3194357753,-0.1190284342,-0.1465221792,0.3752413392,0.2219667137,-0.2978767455,-0.0025534513,0.0142148491,0.0764803588,0.0501164533,0.182024464,-0.0263383463,-0.2110521197,0.2065535635,0.2589918971,0.1118551642,-0.0120119732,0.3028580248,0.32057181,0.1013267934,0.0900926217,0.4885220826,-0.0741285533,0.2885040641,0.1695934534,-0.4038478136,-0.0279918108,-0.3101570904,-0.0101930276,0.1660609543,-0.2809512913,0.3844383657,-0.0233361945,0.1436888129,0.0331227146,-0.1440892816,0.1975586116,-0.2035194486,0.1216147915,-0.3576218188,0.1515701264,-0.2187893242,-0.1585212201,0.3434265852,0.0080273226,-0.1734503955,-0.0092228455,0.068846941,-0.2523231208,0.1803003103,0.1108043566,-0.1307085007,-0.1216261983,-0.0187929682,0.0433084443,0.1381439418,-0.1616659611,0.0370781571,0.3141900599,-0.2092274874,-0.6543345451,0.143428117,0.458979547,0.4952997267,-0.3055261672,0.0206241775,0.167655468,-0.1292135119,-0.1672609448,0.2107074857,0.1043298692,0.3307001293,0.1919810623,0.1544502974,-0.2581092417,-0.1369844973,-0.0878850594,0.1116116047,0.0719296038,0.6385579109,0.0498813838,-0.1997868866,0.0062698862,0.0683505237,-0.3507208526,0.3475797772,0.3512928188,-0.1270153821,0.2664192319,-0.1019331813,0.0735419989,0.0441734865,0.4316432774,0.03533208,0.0454868004,-0.3527221084,0.0880128443,-0.1400937736,0.1562740058,-0.2091967016,0.0526903272,0.0915924236,0.1963043809,0.0217033625,0.0744331405,-0.1930602044,0.3219523728,-0.0383204632,0.1419627517,-0.3500450253,-0.0792523697,-0.372980535,-0.1365805715,0.1833984256,-0.5927073359,-0.0071959565,-0.0129067888,0.1352603436,-0.1617542356,-0.0071374322,0.235932678,-0.2311911881,0.2665962577,0.0265700165,0.2900746763,0.0479745194,-0.3419520259,-0.2662654817,-0.3045827448,0.0816580877,-0.2666801214,-0.0141738635,0.3200130463,-0.1123692244,0.3886220455,-0.4038390219,0.1768711209,0.2671312094,-0.1090052575,0.0495304205,0.2721484005,-0.1504925042,0.0949938223,0.2195362449,0.2015395314,-0.0126836402,-0.0605703965,-0.4030297399,-0.2274160683,0.4964508116,-0.5431749225,-0.5018324852,-0.1688915044,-0.0841216817,-0.1501206458,-0.0443166308,-0.3090223074,0.0871141627,0.1143873259,-0.1513744742,-0.1797164679,0.1392956972,0.315246731,-0.1256095171,-0.0975036919,0.6928526163,-0.1009183303,-0.1336501539,0.0122063607,0.0877374113]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"Hi ! Yes we are working on making `filter` significantly faster. You can look at related PRs here: #2060 #2178 \r\n\r\nI think you can expect to have the fast version of `filter` available next week.\r\n\r\nWe'll make it only select one column, and we'll also make the overall filtering operation way faster by avoiding many arrow<->python conversions especially during writing.\r\n\r\nI'll let you know how it goes !","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":68,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\nHi ! Yes we are working on making `filter` significantly faster. You can look at related PRs here: #2060 #2178 \r\n\r\nI think you can expect to have the fast version of `filter` available next week.\r\n\r\nWe'll make it only select one column, and we'll also make the overall filtering operation way faster by avoiding many arrow<->python conversions especially during writing.\r\n\r\nI'll let you know how it goes !","embeddings":[-0.1237306297,0.2983857691,-0.0208494086,-0.253370285,0.1016575247,-0.1387869567,0.3027839959,0.6129099131,0.2476259619,-0.0328550227,-0.0985994413,0.486082375,0.0381354578,-0.1835789829,-0.0157278683,0.1854575276,0.090091534,0.3238477707,0.3035545349,0.102088511,-0.0896562859,-0.1577695161,-0.1881163716,-0.0899896622,-0.0092690233,-0.0232012812,0.2395873368,-0.4195279777,-0.3287965357,-0.1773727238,0.297545433,0.2998765111,-0.172678113,0.1115236431,-0.0001252739,-0.0511702336,0.0480971336,0.0077384748,-0.2174410671,-0.0757367685,0.1736849248,-0.2435381562,0.1387726963,-0.1812337488,0.0790456235,-0.0927500203,0.0311846491,-0.0050550313,0.0181248859,-0.0703878924,0.0492277034,-0.1021525934,-0.2875941098,0.2743896842,0.4761992991,0.2047463655,-0.0083874306,-0.1396258622,0.3575642705,-0.3982299864,-0.1014354154,0.56949085,-0.4970218539,-0.0589742512,0.3547709584,0.0128873754,0.1070796624,-0.2253790498,0.3161904514,0.1574979126,0.1816737056,-0.2657525241,-0.1376526058,-0.239519313,-0.0685056448,-0.2304995209,-0.0747090057,0.0416634269,-0.4605602026,0.0094686523,-0.2457598597,-0.2786821127,-0.1020751223,0.4110223055,-0.4475055039,0.4533467591,0.1792585999,0.2715730965,0.2482569963,-0.2148012519,0.0170701668,-0.0379099101,0.435307622,0.5598415732,-0.4327193499,-0.1885328293,0.0590466596,0.0518121608,0.1920053661,-0.2098816931,-0.1892127842,0.4369458258,0.3875004351,0.0874744132,0.3505068421,0.0362566859,-0.0684560388,0.6006133556,0.3717557192,-0.1598750949,0.1043104678,0.0937727764,0.0169144273,0.2993757427,0.1783020943,-0.383556962,-0.2213908136,-0.1921917498,-0.0098191882,-0.0626161993,-0.2706885636,0.0100912945,0.1224384084,0.3963264525,0.2601259053,0.2985363007,-0.1823651344,-0.0098769339,-0.2351266295,0.0012536942,0.0751003325,0.0637567192,-0.014473903,0.1764781773,0.1271897256,0.1445113122,-0.1922730356,0.0070491503,-0.0052531864,0.1212490126,-0.0002686502,-0.19125323,0.2603036165,0.4908867478,-0.1221283674,0.3843762577,0.2474316657,-0.2563829124,-0.3247012794,0.2783824205,-0.1615214497,-0.2274027765,0.090019308,-0.0268314276,-0.0578177981,0.2852585614,-0.1246034876,0.6093722582,0.352546066,-0.1533310562,-0.1372533739,-0.1558424532,-0.2802358568,-0.2027676404,0.3308411539,-0.0845789611,-0.4108805656,-0.1949425042,-0.2197655439,0.293651849,0.4690000713,0.3666896224,-0.1443485767,0.0736395568,0.3656388521,0.3744418025,0.6236249208,-0.0126945423,-0.5791273713,0.071481213,-0.1642435491,0.1080868095,-0.184307307,0.3767103553,0.682488203,0.1427904814,0.2814699709,0.193589896,-0.0623287596,0.2590735257,-0.1493914872,-0.1364693344,0.1592231095,0.057282757,-0.0175766721,-0.1859794855,-0.0208692737,0.038418375,0.0531386174,0.0237401836,0.1548101455,0.0545606986,0.1606440097,-0.065180853,0.1213701963,-0.271641165,-0.250651449,0.0501639433,0.4665891826,0.1352522224,-0.3585747182,-0.3654567599,-0.1568947732,0.1310864836,0.4029460251,0.1300919205,-0.0620348081,-0.3364857435,0.2275574654,-0.2044117004,-0.2051123977,-0.0941829234,-0.0602979325,0.0857655853,0.0432424918,0.0083482834,0.2371554971,-0.0684362352,-0.0852610618,-0.1317295581,0.2426041067,0.2441398352,0.1396604031,0.0777215436,0.2486257702,-0.1695636213,-0.2477496117,0.4770077765,0.0623661391,0.0963865966,0.1250981987,-0.1157191172,0.1698826253,-0.1404646337,-0.5081629753,0.2969232798,0.1103785634,0.6598653197,0.0501083769,0.1540759057,0.1797201782,0.0642933697,-0.1062137708,-0.3445337713,0.0360425897,0.0097844936,-0.1476871222,0.1191625595,-0.0164020862,0.311193794,0.2018047124,-0.0426718444,0.0906062052,0.3051792383,-0.1029429734,-0.1918518841,0.1773736328,-0.1114645153,-0.0233574919,0.2024938911,0.1365882754,-0.2860154808,-0.1919526011,-0.0219912548,0.2376988083,0.2129504681,-0.1489863843,-0.2452257872,0.2079773545,-0.052284766,-0.3534881473,0.0330158956,-0.049060367,0.3363072276,-0.2336784303,-0.1825104654,-0.2598515153,-0.1426429898,0.3249364793,-0.1080732942,0.0476651415,-0.2792333961,0.3316101432,0.1832596362,0.1082328632,0.1978942901,-0.1676245481,0.2757034004,0.0065290718,-0.1713165492,-0.3011424839,-0.467705965,0.0129964156,-0.0658765882,0.0555221029,0.4877010584,0.2073121965,0.3975659907,-0.3270418644,-0.2520027757,-0.4476266503,0.0363562815,-0.0938580483,0.2299013734,0.0694783255,0.532861948,-0.3596934676,0.0280111916,-0.0339045972,-0.0420932844,-0.0976932645,0.1281557679,-0.0612140521,0.2300920337,-0.1059907079,0.2067679018,0.0523473881,-0.1154341474,-0.1115759239,-0.0638523325,0.3551960588,-0.4245603979,0.0063691381,-0.298771739,-0.1701822728,-0.0975964367,-0.0731262565,0.2677829266,0.428162694,0.2718480527,-0.0998913348,-0.1488031298,-0.0964542255,-0.197841391,0.5476942062,-0.2510440946,-0.0182163846,-0.2387063205,-0.0076119909,0.1322064698,0.1117509678,0.4148824513,-0.0027590583,0.0404852815,0.0841110796,-0.2670807838,-0.2934962213,-0.1214538887,-0.1879380196,0.2416297793,0.6311585903,0.2827468514,0.7758919001,0.1585040838,0.2212490886,-0.0352504142,0.0000667495,-0.0199694689,-0.026555093,-0.2585273087,-0.2028662711,-0.2288312018,0.00177074,0.15331541,0.2056681812,-0.5505740643,-0.0349599235,0.1599182039,-0.1243148968,-0.3024810851,0.3074432909,0.0429903194,0.129929781,0.3579552472,0.0900822803,-0.4467230439,-0.4382316768,-0.0686128139,-0.1267530918,0.009031524,-0.1507368088,-0.6066063046,-0.2665180862,-0.6999788284,0.2868987918,0.1216505021,0.1199997887,0.2492075861,-0.2503612339,0.2562128305,-0.120631516,0.6425053477,-0.0756214038,-0.2868643701,0.1585032493,-0.0402051136,-0.6319999099,0.2532657087,-0.1052063629,0.1121054888,0.1583260894,0.4987238348,-0.539742887,-0.2433731407,0.0737013668,0.1668108404,-0.1530797333,0.0764809176,-0.0852575153,-0.3236049116,-0.155081436,0.2726483345,0.0012735343,0.135393098,0.4726767838,-0.0452169627,0.120014824,-0.1234330535,0.1013451368,0.3883758485,-0.0829367191,0.203412652,0.0647977516,0.2924255431,-0.3008713722,-0.0861680731,0.0826908574,-0.0724688992,-0.1351945549,-0.0698229596,0.1214748248,0.2875911593,0.5704543591,0.0263033304,0.3826789558,-0.1464604884,0.2083724886,-0.0090282056,-0.0144843515,-0.2250097543,0.0013283582,-0.2829307914,-0.3367445171,0.4052058756,0.2880266607,-0.2896946967,0.6709111333,0.2133115232,-0.1537069082,0.4451875389,0.3529538214,1.0079770088,-0.4848758578,0.1318350583,0.0028191938,0.0846739039,0.3737158477,-0.3971540332,0.2397607565,-0.16377832,-0.0604240187,-0.0320760459,-0.0690370128,-0.1288758516,0.1712004393,0.1313425601,0.1395166814,-0.0944087729,0.46356076,0.1023129523,0.1462407708,0.2702816427,-0.4677024782,-0.0909074396,-0.0055961641,-0.0260197446,-0.2302765846,0.0807614625,0.3179075718,-0.1683582515,-0.4044848382,-0.2705263793,-0.3155388534,-0.1868193299,0.0461292155,-0.1413326114,-0.0896037817,-0.1633402407,-0.1179498732,-0.3856491745,0.1154306009,0.1286413968,0.1160673276,0.4527563453,0.2938825786,0.1255627871,-0.1649270952,0.0901222304,0.0425589457,-0.0523623042,0.2120885849,-0.086741738,-0.3952055871,-0.2785109282,0.1678526551,0.6052414179,0.1069878787,-0.0429318883,-0.0353726707,0.1367548108,-0.2750187218,-0.0082141338,-0.0976498574,0.0350237302,0.4042137563,0.0630764216,-0.2621450424,-0.0026682711,0.3417323828,0.3048818111,0.0388075374,0.1938853562,-0.1132600233,-0.5539798141,-0.0429438315,-0.0578192323,0.1937415302,0.0155416215,0.0119008422,-0.346998632,-0.1012728438,0.0332415849,0.1828553677,-0.0593058728,-0.0367570147,-0.2799467742,-0.4299452007,-0.0247346703,0.4117118418,0.1347901225,0.0436267778,-0.0224000644,-0.1254225969,-0.4318522215,0.4749076962,-0.1643548757,-0.1609070301,-0.0701550022,0.3995199203,-0.1253770143,0.2816865742,-0.101436086,-0.0485595874,0.0809797496,-0.0200034901,-0.0000313137,-0.1230206043,0.101277329,0.2107244134,-0.1763364226,-0.1292170882,0.1146144792,-0.1573331058,-0.1441316754,-0.4329552352,0.1776144207,0.0241644364,-0.0330143049,-0.3705619276,0.3870609999,-0.1899461746,-0.0303929672,0.1495629996,-0.2027793974,-0.1067697778,0.0635471568,0.339833051,0.0760192126,0.0470574982,-0.0957734063,0.1340186298,0.0065673143,-0.2462806553,0.0288025457,0.1185765564,0.0413293578,0.202600345,0.1136428937,0.3781630993,-0.0227025263,-0.0405412726,0.3674143553,0.0015126691,-0.2030047774,-0.3245555758,0.3961252272,-0.0828287378,0.2127986103,0.3578576744,-0.0304116253,0.1189253181,0.0621135607,-0.2317251265,0.5323970914,-0.2480443716,-0.3125813007,0.3860908747,0.2828720808,-0.1621648222,0.1166021302,0.1299548149,-0.1021374092,0.2557518482,-0.1180983633,0.2251511812,0.1457099468,0.1153860167,-0.4327330887,-0.0669913366,0.0590895489,0.4809488356,-0.1702383906,0.0535878092,0.0727178529,-0.1600968689,0.5276103616,-0.0994011089,0.2219124436,0.277836889,0.1126371399,-0.2351590246,-0.3627433777,-0.2615092397,0.0753773525,0.0202132035,-0.0492385626,-0.3334949911,0.4740839005,0.0513870083,0.254230231,-0.417531848,0.1141129881,-0.2918193936,0.2974714935,-0.1703625917,0.0429236777,-0.4668008089,0.3129241765,0.0506234616,-0.3800785244,0.1181158125,0.3163570464,0.0049841343,-0.0587676801,-0.3002197444,-0.2340755165,-0.0624200031,0.037731573,0.089023076,0.0551040918,-0.0459939167,-0.0437043644,-0.0474134013,-0.2555032074,0.3206149638,0.1943210363,-0.0108285584,-0.0755018815,0.1178065985,-0.2721933424,0.0277991556,-0.0904481933,0.2527711391,0.0565551817,0.0049547357,-0.1734624505,0.1501864046,0.0472349077,0.009915269,0.046812959,0.2418235838,-0.0901392475,-0.0123431887,-0.1436452866,-0.3169914186,-0.4253723919,-0.0527304932,-0.7110068202,-0.3438414633,0.3633756042,0.325609833,0.4668024778,0.385676384,-0.0218875036,-0.1853245795,-0.3303778172,0.35460338,-0.1234021708,0.4170963764,-0.0245323479,0.1251488924,-0.0768925473,-0.5126946568,0.3432998359,-0.1628069133,-0.0592578799,-0.0112874368,0.0658438876,0.2929888964,0.2172849774,0.2603945732,-0.1421017945,0.1607438326,0.0618622825,0.1467075646,-0.1186076477,0.3497376442,-0.4658871293,0.1488122493,-0.0536530875,0.186391592,-0.1162667498,-0.1365158707,0.0254056025,0.0742522404,-0.2378194183,-0.1673729122,0.2110529095,0.0836493596,0.2566668689,0.2304351032,-0.3541635275,-0.0726132318,0.00521565,0.072804682,-0.2791232169,-0.2218242139,0.5476240516,-0.8053920269,-0.4464309514,-0.2124727219,0.0680942088,-0.0251128655,-0.1262598932,-0.1320752949,-0.3772806525,0.2484419495,-0.1559031755,-0.3053336143,0.0773434639,0.1260427237,-0.1556928903,-0.3427354395,-0.0300387908,-0.0093355086,-0.0752346218,-0.5701998472,-0.5153660774]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"@lhoestq Thanks for the response\u2014 it's great to hear that we'll be getting a much faster `filter` method soon. However, my use case does also involve using `map` over a single column in order to pre-compute roughly uniformly sized batches, and right now that is also very slow. Is there any plan to make `map` faster for single column operations?\r\n\r\nIf that's not a priority for the maintainers right now, I could try my hand at adding the feature, but I can't guarantee I would do a good job given my lack of familiarity with pyarrow.","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":96,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\n@lhoestq Thanks for the response\u2014 it's great to hear that we'll be getting a much faster `filter` method soon. However, my use case does also involve using `map` over a single column in order to pre-compute roughly uniformly sized batches, and right now that is also very slow. Is there any plan to make `map` faster for single column operations?\r\n\r\nIf that's not a priority for the maintainers right now, I could try my hand at adding the feature, but I can't guarantee I would do a good job given my lack of familiarity with pyarrow.","embeddings":[-0.1885621548,0.3007333875,-0.0062208734,-0.2915562987,0.1139059737,-0.125191018,0.377454102,0.6256682873,0.2875320315,0.0172812026,-0.0364471562,0.5041142702,0.0290438402,-0.1641895324,-0.0369139165,0.1709661037,0.1048239693,0.2638177872,0.2762956917,0.1285973787,-0.105499059,-0.230432421,-0.2212545425,-0.0562877022,0.0027704337,-0.1056099907,0.1589284986,-0.4839156866,-0.2446201891,-0.1703580618,0.2296721339,0.3419600129,-0.1645797342,0.1854377538,-0.0001265591,-0.015295011,0.0672803298,0.0046598678,-0.1326862276,-0.0347867832,0.1036187708,-0.2461567819,0.0807675719,-0.2282919735,0.1449373364,-0.0729279965,0.0921783447,-0.0480944179,-0.0118833063,-0.1497037262,0.0260400884,-0.1085724011,-0.2673380375,0.2866571844,0.393298775,0.1476784497,-0.0333684795,-0.1472428292,0.3618592322,-0.4993890524,-0.1028713882,0.5538045764,-0.4942291677,-0.0468575619,0.3240176439,0.0596823879,0.1114338115,-0.1639341563,0.2983200252,0.1498267651,0.1279602647,-0.2527841032,-0.2238488048,-0.2600064576,-0.0924418569,-0.1685850471,-0.0613482371,-0.0341120027,-0.395265609,-0.0476741418,-0.2808194458,-0.3235608637,-0.0451245233,0.3968591094,-0.4360876679,0.4768274724,0.2962536216,0.2747288644,0.3442022502,-0.1562515199,0.002886483,0.0076632174,0.4512862265,0.5650333166,-0.4838467538,-0.1833854765,0.0761876479,0.0020666937,0.1733239442,-0.2895022035,-0.1337257326,0.4674760103,0.3749387562,0.1282308549,0.3100210726,0.0715184212,-0.152591899,0.5785233378,0.449947834,-0.129087761,0.0587535352,0.1226570308,-0.0138666965,0.227632001,0.1756646484,-0.3141234815,-0.3155645132,-0.129306972,0.1225796044,-0.1538142413,-0.2330203354,-0.0025992324,0.0703268573,0.3658011854,0.2422051579,0.2920158505,-0.2204377204,-0.0425460264,-0.1804530472,0.0192145091,0.1090091318,0.0703660175,-0.0252906997,0.2317008525,0.1311382651,0.1572612971,-0.2627331614,0.022589691,-0.0427703559,0.1445612758,0.002202641,-0.1336308271,0.2234643698,0.3903025687,-0.1576905549,0.3427307308,0.1693411916,-0.1029923633,-0.3292446434,0.2667769492,-0.1952231824,-0.2769999504,0.1620929837,-0.042051293,-0.0549543053,0.2621858418,-0.1286166608,0.5985473394,0.3813329637,-0.2264772654,-0.161817044,-0.0694191754,-0.3165629208,-0.1783338338,0.2890984416,-0.107911706,-0.3547855914,-0.2313257307,-0.1878269315,0.3709045947,0.41609025,0.3849086165,-0.1198677272,0.1343068928,0.326631248,0.3321783543,0.5479459167,-0.1042645127,-0.5639656186,0.0655858293,-0.2416366339,0.0511492491,-0.132146135,0.3440311849,0.6694209576,0.0902323499,0.3183021545,0.1980740428,-0.1306650639,0.3279159963,-0.1836322248,-0.2531143427,0.1532042623,0.0489023924,0.0427230671,-0.1249033883,-0.0157539696,0.044599507,-0.012951361,-0.0185180809,0.1647890955,0.060226623,0.0655706674,-0.0913712457,0.1133537441,-0.2321182787,-0.2884664237,0.0555403233,0.3865182698,0.1459059417,-0.3229031563,-0.3966253698,-0.039270483,0.0619345754,0.4316823184,0.1690214723,-0.0822235197,-0.2904554009,0.3275698423,-0.2072603405,-0.1705351621,-0.1253021806,0.0233977586,0.0944539383,0.0706552267,-0.019744765,0.2438316643,-0.0827937871,-0.1178131625,-0.0614261217,0.2267122865,0.1529109329,0.1414231658,0.0943184793,0.1466474533,-0.0662407354,-0.17863442,0.5075376034,0.0700756013,0.2343178391,0.1177433059,-0.1211806536,0.1545407176,-0.1719129384,-0.5523522496,0.2861994505,0.1465551257,0.6651866436,0.0841302425,0.1220776737,0.1631964594,0.0618514493,-0.0704910606,-0.3236384392,0.0100724185,0.0261883866,-0.1350475997,0.1274299622,0.0202217549,0.2641682625,0.2843721807,-0.0682591498,0.1308690459,0.3047645986,-0.0523111336,-0.2207185328,0.2096385062,-0.1084976941,0.0057851104,0.1629600376,0.1217780188,-0.2879891992,-0.1651800722,0.0005724233,0.2164832354,0.179439202,-0.1415688396,-0.2166480571,0.259044528,-0.0143977152,-0.2989133596,-0.0106078032,0.0033571064,0.3582737446,-0.1539956331,-0.263288945,-0.2762300372,-0.0915408656,0.3394620717,-0.0426127873,0.0372456685,-0.309864819,0.307069093,0.1898735166,0.1026942208,0.2113774717,-0.1225055754,0.2939247191,-0.0268258285,-0.1728902757,-0.3328283131,-0.5393310785,0.0738206059,-0.0852858722,0.0609383211,0.3177292645,0.1754732579,0.3763103485,-0.3800204098,-0.223446101,-0.4325364828,0.0378246717,-0.0858775079,0.0967652947,0.0850851089,0.427698195,-0.3566629291,0.0667089149,-0.0191574004,-0.130415082,-0.0418006591,0.0976940244,0.0027200289,0.2611276507,-0.1275754124,0.2719992101,0.0730158091,-0.1125070751,-0.0697457045,-0.1818050444,0.3236169219,-0.4373997152,0.0072227512,-0.2794033587,-0.1387145519,-0.1494147927,-0.0410582535,0.195117414,0.4307050407,0.271338582,-0.0999782458,-0.2389670312,-0.1302749813,-0.1822680235,0.571095109,-0.2226636559,-0.0169394407,-0.204339087,0.0421478339,0.12186528,0.0817672759,0.4627899826,0.0375514291,0.0420409404,0.1029868349,-0.2466260791,-0.281311512,-0.0661424845,-0.2550459504,0.2800433934,0.56834656,0.3082821369,0.7266680598,0.0995223299,0.1849897355,-0.0433402807,-0.0169090759,-0.0830331743,0.0083542401,-0.242262274,-0.1618505865,-0.2642281055,0.0175665431,0.1437634826,0.1747601926,-0.5115405321,-0.0572758839,0.1953601837,-0.0673126057,-0.2947244644,0.3095500767,0.0013085535,0.1959345192,0.3222126365,0.1213774905,-0.5042794943,-0.384335041,0.0024958167,-0.1383252591,0.0402848572,-0.2440773994,-0.5901769996,-0.2019113451,-0.7772264481,0.3221868575,0.1993768066,0.1177185625,0.2021474391,-0.204020679,0.192175895,-0.1418825388,0.631762445,-0.1416506469,-0.265416503,0.1212948561,-0.06587217,-0.5496025681,0.1885109991,-0.1124100462,0.1748047322,0.1434688717,0.6085749269,-0.5473339558,-0.2011221349,0.0751436576,0.1330287158,-0.1246811002,0.088488549,-0.0653651431,-0.2612018585,-0.158933267,0.2646705508,-0.0285974871,0.0804991052,0.4317059219,-0.0816911906,0.1488001645,-0.0747326985,0.0849915147,0.3523804247,-0.0977422372,0.1742149889,0.0473537296,0.3824132681,-0.3722229898,-0.0965272933,0.0484999605,-0.0851116255,-0.1281922907,-0.0551700816,0.0793755203,0.2917897105,0.5293624997,0.0455495901,0.4108189046,-0.1715718061,0.2378075719,-0.0840340033,-0.119588919,-0.120784767,0.0624491982,-0.2965333462,-0.3106878102,0.3419625461,0.3424166739,-0.3179953098,0.7173678279,0.0572209321,-0.1750314385,0.4352219105,0.3283764422,0.9149349332,-0.5607410669,0.1422017664,-0.0261075366,0.1689550728,0.3355853856,-0.3939293623,0.2626981139,-0.1080000028,-0.0748513043,-0.0103270467,-0.011386998,-0.1241636723,0.2502501905,0.0524287969,0.1525126547,-0.0156947691,0.5056079626,0.0496899225,0.1444932967,0.3504320085,-0.5191715956,-0.0689044893,-0.0282111764,0.052250579,-0.2357691228,0.0680917427,0.326651752,-0.1470910311,-0.3844144642,-0.2798959315,-0.3103632033,-0.1332495958,0.0652173683,-0.1632840633,0.0363338441,-0.2110343724,-0.0992980078,-0.3842190504,0.1073063239,0.1282978803,0.027497245,0.4499252737,0.3063403964,0.1378822029,-0.1738569885,0.032499928,0.071831964,-0.0743707865,0.1883494556,-0.1334447265,-0.4151903391,-0.3370565176,0.1551696807,0.5482294559,0.0980455577,0.0058046854,-0.065120399,0.071111016,-0.3431614339,-0.0311352629,-0.094201602,0.0626096576,0.4800369442,0.0268999171,-0.2448760718,0.0245454609,0.3853031397,0.2751039863,0.0067283632,0.2028694898,-0.1543508023,-0.5062571168,-0.0726518184,-0.0528793707,0.1456578225,0.0340627804,0.0834869146,-0.3774341047,-0.1931779385,0.0424489751,0.207409218,-0.1004489437,-0.0333153158,-0.3305812478,-0.4232656956,-0.0712771788,0.4162345827,0.2322441638,0.0238806587,0.0567832179,-0.1345335245,-0.4628409445,0.5339362025,-0.139524743,-0.1586112678,-0.0122109829,0.3892353773,-0.0375511944,0.2797936797,-0.081059359,-0.0000286426,0.0742138773,0.0025763838,0.0576240197,-0.0986017659,0.1095756218,0.2336111218,-0.2305233777,-0.1115553603,0.1880575866,-0.1836698055,-0.1581714898,-0.4777017832,0.2033780962,0.0635814667,-0.0704593658,-0.3293753862,0.3687349856,-0.159183383,-0.0086849164,0.1842011213,-0.1702012122,-0.1051285863,0.0022570339,0.363632828,0.0898703188,0.045148734,-0.0060331151,0.1071893796,0.0324232616,-0.1778638512,0.1130439043,0.1745919734,0.028748516,0.2241037339,0.1080931127,0.278498441,-0.0144753847,-0.0748520195,0.2693952322,-0.0352370292,-0.2041104436,-0.3471260071,0.4577021003,-0.0549655892,0.2916862667,0.3726937175,0.0294675771,0.1111128405,0.0538082644,-0.2809929252,0.4950441718,-0.2272440791,-0.266541779,0.4440737963,0.3002642989,-0.1433021426,0.1471917778,0.0509747379,-0.0858997777,0.2235788256,-0.0608492903,0.2653491199,0.1429326087,0.1563304365,-0.3695169389,-0.103869319,0.1376334578,0.4576835632,-0.0954738036,0.1299730539,0.0675002709,-0.1041138768,0.3830341995,-0.0945932865,0.2381734103,0.2927366495,0.0518166572,-0.2534310222,-0.3368816674,-0.2654070258,0.0807973593,0.0562043637,-0.0898934826,-0.2092670649,0.6242589355,0.0587936677,0.3631591201,-0.3136518598,0.0905774683,-0.3196901083,0.4099572301,-0.1601746231,0.098274976,-0.4550583661,0.2978042662,0.0457877666,-0.25660941,0.1654209942,0.3744315207,-0.0223907921,0.0160352662,-0.3487819135,-0.2221980542,-0.0723251402,0.0730130598,0.058225356,-0.0306647643,-0.0743525922,-0.0727119073,-0.0337033384,-0.205810681,0.2698540092,0.1440162212,0.0443663299,-0.1139964983,0.2203514725,-0.3583990335,0.0894811079,-0.0856514722,0.3273695707,0.0101838373,-0.0486244075,-0.2447471768,0.0983688459,0.065892905,-0.00354799,0.1109228507,0.3180923462,-0.0563119501,-0.0112249441,-0.1563532054,-0.3178821802,-0.436607182,-0.0439393297,-0.7775238156,-0.3388490081,0.3656428158,0.2502533793,0.4327199161,0.3576798141,0.0594947375,-0.254188031,-0.3120018542,0.3594669402,-0.1705431938,0.4331060946,-0.0254811421,0.0763301849,-0.0630652905,-0.611269474,0.4074570537,-0.0563078076,-0.0882132724,-0.0075217159,0.058953397,0.2174141854,0.1186416671,0.3269604743,-0.1543505341,0.1257506162,0.0076241908,0.1343767345,-0.1755220592,0.4124649167,-0.4547960162,0.1857274771,-0.0626004562,0.1543040723,-0.1882402003,-0.1462156028,-0.0242185555,0.0578407235,-0.2225520611,-0.1250486523,0.2061223388,0.1301379651,0.20737423,0.1720453352,-0.3358873427,-0.0327047147,0.0141503084,0.0218519587,-0.298791647,-0.1553069651,0.4840515256,-0.7961956859,-0.4655472338,-0.2211012989,0.0340085812,-0.1171962172,-0.1112646684,-0.1842517108,-0.372202307,0.2183485925,-0.0547117665,-0.2753545046,0.0352672711,0.104565531,-0.1995068789,-0.3591188788,-0.0041877283,-0.0378078856,-0.073004663,-0.5274661183,-0.5042610765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"Currently the optimal setup for single-column computations is probably to do something like\r\n```python\r\nresult = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n```\r\nThis has two advantages:\r\n- input_columns=\"my_col\" allows to only read the column \"my_col\"\r\n- remove_columns=dataset.column_names makes `map` only keep the output of your function `f`, and it drops the other columns of the dataset instead of keeping them.\r\n\r\nLet me know if it improves speed on your side.\r\n\r\nYou can also get more speed by using `batched=True` and setting `num_proc=` for multiprocessing","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":82,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\nCurrently the optimal setup for single-column computations is probably to do something like\r\n```python\r\nresult = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n```\r\nThis has two advantages:\r\n- input_columns=\"my_col\" allows to only read the column \"my_col\"\r\n- remove_columns=dataset.column_names makes `map` only keep the output of your function `f`, and it drops the other columns of the dataset instead of keeping them.\r\n\r\nLet me know if it improves speed on your side.\r\n\r\nYou can also get more speed by using `batched=True` and setting `num_proc=` for multiprocessing","embeddings":[-0.1294831634,0.3391118646,0.010745544,-0.2183184326,0.0991218835,-0.0936415046,0.4080079794,0.5750510097,0.2324593961,0.0821521357,-0.0695115253,0.4634191394,0.0280998684,-0.1797308326,-0.0288197026,0.2203612626,0.0991646647,0.3066155314,0.3233721256,0.0617822073,-0.1302561164,-0.2138473243,-0.1926399022,-0.0478350148,-0.0517784096,-0.13717103,0.1724527031,-0.4266522527,-0.2462246418,-0.1342870593,0.3095401824,0.3318148851,-0.1605596989,0.1470036358,-0.0001253524,-0.0406684466,0.0192112252,-0.0276141986,-0.1973944306,-0.0811643004,0.1680786312,-0.2034993023,0.123624377,-0.2444311082,0.0947886631,-0.130144909,0.0337789245,-0.0696482509,-0.0037102471,-0.1487485468,0.0337340087,-0.1216606945,-0.3191208243,0.2640785575,0.4368256032,0.1192009076,0.029285932,-0.1973635852,0.2728891373,-0.3919943571,-0.1034414023,0.5533384681,-0.4737558067,-0.0747155324,0.3379896581,0.0730806664,0.1299578995,-0.208292231,0.3107067943,0.1420645118,0.1558232456,-0.2844787538,-0.2145438343,-0.2602309287,-0.0650966913,-0.2818907201,-0.0909782499,0.0619444288,-0.4108319283,-0.0512523502,-0.257570684,-0.2819582522,-0.1185125634,0.4137207568,-0.5403962731,0.4640475512,0.1953418553,0.2909294367,0.1714852899,-0.172450617,-0.0003867647,-0.0425582677,0.4799712002,0.5920433998,-0.4159093201,-0.1651493013,0.0710029304,-0.0379414223,0.2057011127,-0.2445922941,-0.1593009233,0.4084405601,0.3382352889,0.1511622667,0.3294962347,0.0680308565,-0.0791234747,0.5972884893,0.4532561898,-0.1002956107,0.1745669097,0.0794557855,-0.034586709,0.3752577901,0.2228336632,-0.317723155,-0.190140903,-0.2148229778,0.0109900394,-0.0076941638,-0.2603752911,0.0728768334,0.0439946204,0.3605624139,0.2810072303,0.3244585991,-0.1984217316,0.0045822505,-0.1865285635,-0.0042484505,0.0518905334,0.1187332347,-0.0268213823,0.1671210378,0.1928978711,0.1814216822,-0.1461576223,0.0213639326,-0.0123185273,0.0233145654,-0.0547806099,-0.2017881423,0.2513333559,0.4559927285,-0.0558084361,0.3779971898,0.2080499232,-0.2995767593,-0.3604398668,0.2355452329,-0.1126081422,-0.2335226983,0.1333259791,-0.0212276354,-0.0874711126,0.3423869312,-0.1555574834,0.5529490709,0.3655698597,-0.0892596096,-0.1648725569,-0.1550862938,-0.2858027816,-0.1917106062,0.3239213228,-0.1121316329,-0.3836997449,-0.1963758022,-0.1693079323,0.2926425636,0.4127456844,0.3783282936,-0.0817165896,0.1228251457,0.3835410476,0.3930479884,0.5803683996,-0.1034062952,-0.5420395732,0.0425314419,-0.209037751,0.1126556844,-0.1759260744,0.3581134379,0.7156688571,0.1406585574,0.3412389457,0.1964561343,-0.0874867216,0.235885337,-0.1443508565,-0.1909597665,0.0868707895,0.0933989286,0.0729549527,-0.1783013195,0.0013807012,0.0436715521,0.0437361449,-0.0355439745,0.202437222,0.0922088176,0.1031505316,-0.0610237718,0.088407591,-0.2997893393,-0.4036322534,0.0803653449,0.4533421993,0.1321221441,-0.318423301,-0.4617904127,-0.1174544021,0.0867154822,0.4272183776,0.1464284211,-0.0560345873,-0.2921648622,0.2219502777,-0.1985518336,-0.2281971276,-0.1070432216,-0.0053619174,0.1575302482,0.0297097508,-0.0352749787,0.2604260147,-0.0967653841,-0.0356778838,-0.044092834,0.2602740526,0.2296839356,0.1976291388,0.0972480699,0.2005938888,-0.1151910499,-0.2620398104,0.4657939076,0.0628671423,0.2084219903,0.073745966,-0.1619641632,0.161136806,-0.1653588116,-0.571092844,0.2745729983,0.0802082568,0.6571453214,0.0523994528,0.1148218513,0.1762300879,0.0400787815,-0.1104949936,-0.3743601739,0.0251731388,0.0522628874,-0.1696225405,0.0909887254,-0.0342410728,0.236560002,0.2350629866,-0.0920958295,0.1094538271,0.2645371258,-0.0854295865,-0.1937573552,0.2097469121,-0.0749384165,-0.0109995492,0.1932286769,0.1502155662,-0.2479687184,-0.1998437047,0.0215491112,0.1492290497,0.2268427014,-0.1879795641,-0.1777149588,0.2505490184,-0.0165851973,-0.3266107738,0.0432185978,-0.0124858161,0.358207047,-0.2027126849,-0.2141210735,-0.2567456663,-0.1363751441,0.3243538737,-0.083675459,0.0062987399,-0.3239730299,0.2826361954,0.2296165973,0.0995098501,0.2094608694,-0.143508032,0.3257274628,-0.0073616165,-0.1692798138,-0.2807351649,-0.4938735068,0.0047150804,-0.0653579831,-0.0360509157,0.4081599712,0.2390294373,0.3726723194,-0.3311499655,-0.23395437,-0.3956118524,0.0620728694,-0.111167945,0.170346126,0.0949766487,0.4669963419,-0.3315181434,0.0566570573,-0.0435329117,-0.0225022677,-0.021113079,0.0907096788,0.0183935575,0.2062325031,-0.144601658,0.2502629161,0.0640242919,-0.1108155474,-0.0677070394,-0.0787984654,0.3661225438,-0.4411123395,-0.0846926495,-0.3175471723,-0.0489359237,-0.0837778226,-0.1027952135,0.246325776,0.4151182771,0.3067745566,-0.0962076411,-0.1359816641,-0.1121387705,-0.1576292366,0.6070975661,-0.2828311622,-0.0360731483,-0.1843346655,-0.0031887321,0.1298577785,0.1505562365,0.4464666843,0.0410566591,0.0530790649,0.0689035133,-0.2580910027,-0.2522442639,-0.1246067956,-0.2612768412,0.2092583179,0.5654310584,0.320635587,0.7258679271,0.1424076557,0.1577912718,-0.0668367669,-0.0433975235,-0.0177733805,0.0256421641,-0.1868588477,-0.202644214,-0.2535117567,-0.0260585435,0.1770983487,0.1996235549,-0.5090255141,-0.075621441,0.1815344989,-0.1252081394,-0.3385491371,0.2375819087,0.0109047946,0.191151455,0.3765664995,0.1332771331,-0.4534001946,-0.4018692076,-0.0390888415,-0.0890869498,-0.0157257151,-0.1809643656,-0.5789369941,-0.1956321299,-0.7405368686,0.2922461629,0.1310438663,0.0847021788,0.2528220415,-0.2260348797,0.1941351891,-0.136633873,0.5798608661,-0.082505472,-0.2240753919,0.1575079113,-0.0124611193,-0.6134457588,0.1907949746,-0.1679947078,0.1235227585,0.1410180032,0.6254450083,-0.5886576176,-0.2078297734,0.0685646906,0.1660812497,-0.1751503944,0.0899863243,-0.1042120904,-0.3079536557,-0.1583369374,0.2863936722,0.008871438,0.0992233977,0.4500450194,-0.0252194256,0.0567622818,-0.0831183493,0.1457530409,0.4208121598,-0.0992939919,0.174905777,0.0397949032,0.3401392996,-0.3165377676,-0.0387107059,0.0795936808,-0.0590486526,-0.1550715715,-0.017339265,0.0238963999,0.3053067029,0.5189320445,0.0065893461,0.3439621329,-0.2133316696,0.2341223508,-0.09517093,-0.0126866642,-0.1836408973,0.025321262,-0.3267571926,-0.3109101653,0.4088344276,0.3397986889,-0.3074758053,0.6517520547,0.1854652166,-0.2204728425,0.4290182889,0.3837018311,0.9632584453,-0.5312177539,0.1537014246,0.0519704297,0.1483111978,0.4319674075,-0.4409230649,0.2472386211,-0.1535259932,-0.0678972974,-0.0565943345,-0.0382300764,-0.1370845139,0.224066034,0.121888794,0.2110776454,-0.0712221637,0.4896341264,0.1225952283,0.1071582139,0.3036717474,-0.5117375851,-0.1088012084,-0.0205419585,-0.0519753247,-0.1958558857,0.0174114723,0.440587908,-0.1976066083,-0.3893168867,-0.2506988347,-0.3046148717,-0.1109618917,0.0474569425,-0.2182463557,-0.0003561124,-0.2450839132,-0.0840412825,-0.310416609,0.1532732248,0.1600227803,0.0433561169,0.4482523501,0.2986348867,0.1100801528,-0.1192908809,0.0398938693,0.0810145512,-0.0170776062,0.2253186107,-0.0950900763,-0.2768141925,-0.2861355841,0.1231281683,0.5650429726,0.0438576117,0.0576898232,-0.0678035989,0.1874904186,-0.3127431273,-0.0283286907,-0.0911498815,0.0332965106,0.4166702926,0.0199962407,-0.2361268252,0.0099517964,0.340190798,0.3162249029,0.0180036929,0.1913575679,-0.1530790925,-0.4885406792,-0.0606079511,-0.0689181834,0.1285406053,0.0677284226,0.0568904057,-0.3558070958,-0.1211188361,-0.0483741648,0.2245283723,-0.0695583969,0.0142334644,-0.2056830525,-0.4497348368,-0.0702765733,0.4310661852,0.2193412632,0.1123792157,0.0700829253,-0.1089040712,-0.5081058741,0.4486719072,-0.1567644179,-0.1713134497,-0.1201055273,0.4468562305,-0.0262385346,0.2871848345,-0.1272630692,-0.070899263,0.0952433273,-0.0560168773,0.0530656725,-0.1106326506,0.0162165184,0.2148144692,-0.1731273979,-0.093467325,0.1737358272,-0.1355854571,-0.1125467122,-0.4212664366,0.2841621339,0.0927283168,-0.0122849727,-0.2881916463,0.3486792147,-0.0896970406,-0.1130394414,0.1339292079,-0.1787056178,-0.0896433443,0.0537479743,0.3764321804,0.0356944464,0.0578478351,-0.0855772942,0.1384302378,0.0209030341,-0.2208879441,0.0139914388,0.1287961304,0.0483882762,0.2053155303,0.1468271315,0.3270042539,-0.0773038566,-0.0844794735,0.3128420115,0.0135721508,-0.2022168636,-0.279389143,0.4110395312,-0.1318184286,0.2403960675,0.410023421,-0.0333230495,0.1102989763,-0.0313256681,-0.2870815694,0.5734832883,-0.2287258208,-0.2615214288,0.3842894137,0.2309413254,-0.1336114705,0.0989186317,0.1595322341,-0.1070569977,0.2939474583,-0.0647357255,0.2178291231,0.1471167952,0.0862330198,-0.4544852972,-0.1338356733,0.1037918702,0.4685186446,-0.1169150248,0.0719066188,0.0245756004,-0.1990139782,0.4639554322,-0.0639429688,0.211917147,0.3197755814,0.0774644688,-0.2459576875,-0.3654267192,-0.2698218822,0.0696310177,0.0449944623,-0.0696029887,-0.3188228309,0.4536851645,0.0790147409,0.2794358432,-0.3466840982,0.099940896,-0.2763340175,0.3596677482,-0.1741621941,0.0696488097,-0.4111563563,0.2742773592,0.0356762446,-0.3283545673,0.1366632134,0.2756025493,0.0271963533,-0.0081175296,-0.3326180577,-0.2413622439,-0.1090972126,-0.0015370671,0.0518128723,0.0962474868,0.0000790643,-0.0458264425,-0.0548204519,-0.1857498288,0.2693297267,0.1811865717,-0.0363212079,-0.0674098805,0.1468579173,-0.3282646537,0.066836752,-0.1062715128,0.308883369,0.0234817211,-0.018852761,-0.1414920241,0.1683291048,0.0559606999,0.0073317168,0.0538591444,0.2261193097,-0.0254492722,-0.0580108464,-0.1612190157,-0.3147875965,-0.4354113936,-0.0354606025,-0.7556962967,-0.4821601808,0.3296384811,0.2776812613,0.4234730601,0.3575217426,-0.038102746,-0.1960168481,-0.3290936649,0.3978025615,-0.134613663,0.4052017927,-0.0876087993,0.0390781648,-0.0431056991,-0.571862042,0.3007232249,-0.0505402237,-0.0730253085,0.049357336,0.0969880745,0.2359211147,0.1218985394,0.2772589028,-0.1566973329,0.1534968913,0.0622584671,0.0690810978,-0.1328702867,0.4088369012,-0.5085700154,0.1719652116,-0.092055507,0.1988611668,-0.1279172897,-0.1135878935,-0.0078774327,0.0722817928,-0.1708055586,-0.1761433184,0.1931818128,0.1397044063,0.196831435,0.2184041589,-0.2370515168,0.0053675361,-0.0120601114,0.0910359323,-0.2862004638,-0.2247099876,0.5315035582,-0.7931774259,-0.4790951908,-0.2157015651,0.1210274845,-0.0571405627,-0.1644340158,-0.186448738,-0.3395096958,0.2887536287,-0.1777638048,-0.3770124018,0.0114195328,0.0803592056,-0.2801436186,-0.3226734996,-0.0827904418,-0.0371742509,-0.0847026035,-0.5192147493,-0.5552364588]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"Hi @lhoestq ,\r\n\r\nI'm hijacking this issue, because I'm currently trying to do the approach you recommend:\r\n\r\n> Currently the optimal setup for single-column computations is probably to do something like\r\n> \r\n> ```python\r\n> result = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n> ```\r\n\r\nHere is my code: (see edit, in which I added a simplified version\r\n\r\n```\r\nThis is the error:\r\n```bash\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000\r\n```\r\nI wonder why this error occurs, when I delete every column? Can you give me a hint?\r\n\r\n### Edit:\r\nI preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the\r\ncomplete dataset and print every sample before calling map. There seems to be no other problem with the dataset.\r\n\r\nI tried to simplify the code that crashes:\r\n\r\n```python\r\n# works\r\nlog.debug(dataset.column_names)\r\nlog.debug(dataset)\r\nfor i, sample in enumerate(dataset):\r\n log.debug(i, sample)\r\n\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n)\r\n```\r\n\r\n```\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000\r\n```\r\n\r\nEdit2: \r\n\r\nMay this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error:\r\n\r\n```python\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n features=datasets.Features(\r\n {\r\n \"a\": datasets.Sequence(datasets.Value(\"int32\"))\r\n }\r\n )\r\n)\r\n```\r\n\r\n```\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1704, in _map_single\r\n writer.write_batch(batch)\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 312, in write_batch\r\n col_type = schema.field(col).type if schema is not None else None\r\n File \"pyarrow\/types.pxi\", line 1341, in pyarrow.lib.Schema.field\r\nKeyError: 'Column tokens does not exist in schema'\r\n```","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":285,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\nHi @lhoestq ,\r\n\r\nI'm hijacking this issue, because I'm currently trying to do the approach you recommend:\r\n\r\n> Currently the optimal setup for single-column computations is probably to do something like\r\n> \r\n> ```python\r\n> result = dataset.map(f, input_columns=\"my_col\", remove_columns=dataset.column_names)\r\n> ```\r\n\r\nHere is my code: (see edit, in which I added a simplified version\r\n\r\n```\r\nThis is the error:\r\n```bash\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000\r\n```\r\nI wonder why this error occurs, when I delete every column? Can you give me a hint?\r\n\r\n### Edit:\r\nI preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the\r\ncomplete dataset and print every sample before calling map. There seems to be no other problem with the dataset.\r\n\r\nI tried to simplify the code that crashes:\r\n\r\n```python\r\n# works\r\nlog.debug(dataset.column_names)\r\nlog.debug(dataset)\r\nfor i, sample in enumerate(dataset):\r\n log.debug(i, sample)\r\n\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n)\r\n```\r\n\r\n```\r\npyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000\r\n```\r\n\r\nEdit2: \r\n\r\nMay this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error:\r\n\r\n```python\r\n# crashes\r\ncounted_dataset = dataset.map(\r\n lambda x: {\"a\": list(range(20))},\r\n input_columns=column,\r\n remove_columns=dataset.column_names,\r\n load_from_cache_file=False,\r\n num_proc=num_workers,\r\n batched=True,\r\n features=datasets.Features(\r\n {\r\n \"a\": datasets.Sequence(datasets.Value(\"int32\"))\r\n }\r\n )\r\n)\r\n```\r\n\r\n```\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1704, in _map_single\r\n writer.write_batch(batch)\r\n File \"env\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 312, in write_batch\r\n col_type = schema.field(col).type if schema is not None else None\r\n File \"pyarrow\/types.pxi\", line 1341, in pyarrow.lib.Schema.field\r\nKeyError: 'Column tokens does not exist in schema'\r\n```","embeddings":[-0.1297885925,0.3438464403,0.0056276908,-0.2346831262,0.1094863191,-0.1253667325,0.395519793,0.5808492303,0.2243223041,0.0691065192,-0.0648775175,0.487503618,0.0244312882,-0.1736370772,-0.0233188793,0.2062173933,0.1099480912,0.3038486242,0.3059216738,0.0706640705,-0.0972660705,-0.196017921,-0.196914956,-0.0473396219,-0.0616178773,-0.0969334841,0.1700523049,-0.4098371267,-0.2656676769,-0.1343441457,0.3225792348,0.3406042457,-0.1604736745,0.1512905806,-0.0001253172,-0.0479153842,0.0150217032,-0.0417702869,-0.1860088855,-0.0878247544,0.1646131724,-0.21570611,0.1242413521,-0.2357680947,0.0668423474,-0.1099051833,0.0184353553,-0.0702940747,0.0177385099,-0.1491788179,0.037824627,-0.1153193414,-0.3119522929,0.2564554811,0.4294274449,0.1337286085,0.0135107469,-0.188696146,0.3076252043,-0.4169775546,-0.1078032851,0.5532121062,-0.4494970739,-0.0867570266,0.3468251824,0.0665670112,0.1131561846,-0.2048595995,0.2837946713,0.1231319457,0.1522666067,-0.2929821312,-0.2057834417,-0.2492104173,-0.0813231394,-0.2832041383,-0.076393187,0.0381773002,-0.4349862635,-0.056203261,-0.2655761242,-0.2840900421,-0.1205125153,0.4191443324,-0.5168408155,0.4453457296,0.2073737234,0.2835462093,0.1737061739,-0.1643485427,-0.0238912906,-0.0488698743,0.491484642,0.5992772579,-0.4381175041,-0.1678372473,0.0869320035,-0.0217091981,0.2280963212,-0.2405296862,-0.1810513586,0.4192118943,0.3674304485,0.1295282692,0.3315454721,0.0803944319,-0.0740255713,0.5757723451,0.4447900057,-0.0879587531,0.1487137228,0.0883329585,-0.0260100309,0.3440795839,0.2287721187,-0.2832647264,-0.200966239,-0.2048863471,0.0376366712,-0.0027047105,-0.2731019855,0.0661597475,0.0608097203,0.3648388684,0.2870669365,0.28793782,-0.1871757805,0.0238858033,-0.220812887,0.0010180156,0.055280339,0.1150854006,-0.0744161308,0.1614280641,0.2063453197,0.1675103307,-0.1636311561,0.0466469601,0.0027351519,0.0358745679,-0.059275195,-0.1801027209,0.2324701101,0.4378067553,-0.065989852,0.3828446567,0.2269887775,-0.2831005454,-0.3427932858,0.2607621253,-0.0827737451,-0.2477099746,0.1496957093,-0.0216771159,-0.096440427,0.3550901413,-0.1667314619,0.5624736547,0.3837341666,-0.1071810946,-0.1672737449,-0.1610994935,-0.2877018452,-0.1954727471,0.3321774006,-0.0849982426,-0.3995158374,-0.1915385127,-0.1742316931,0.2851774693,0.4178023338,0.3706712723,-0.0936105773,0.1220273376,0.3517781496,0.4231685996,0.5522978902,-0.0770867094,-0.5348977447,0.0360717662,-0.224387303,0.1220817044,-0.1777932346,0.3341093957,0.7191378474,0.1400877684,0.3616188765,0.2063519806,-0.1054341793,0.2566282451,-0.1739203483,-0.1875509322,0.1082142964,0.1010326296,0.0737975165,-0.1837998182,-0.0069080633,0.0134579642,0.0576997884,-0.0499883816,0.2077866495,0.0752976984,0.1167883947,-0.0806707665,0.0855278522,-0.2948940098,-0.3675554097,0.1026487499,0.4557942748,0.1674704254,-0.3501776755,-0.4428637624,-0.1378015727,0.0900194496,0.408793956,0.1375357062,-0.0490505621,-0.3041940629,0.2227518409,-0.1991464794,-0.2078983039,-0.1095270663,0.000580881,0.1515118182,0.0227787439,-0.0070037679,0.2466126382,-0.0959864631,-0.0374679118,-0.0547052398,0.2666029632,0.2137798518,0.1916841865,0.0870014504,0.1957302094,-0.117195189,-0.2693024278,0.4498797059,0.0647738278,0.201766178,0.0769662634,-0.1623327285,0.1554093212,-0.1814915091,-0.5328130722,0.3011566699,0.0802621767,0.6450633407,0.0309948269,0.1309301704,0.1760700792,0.0390137956,-0.1072938442,-0.3742184043,0.0231643748,0.0713966042,-0.1661333442,0.0815336928,-0.0277556721,0.2639753819,0.190789327,-0.0964990854,0.0930827558,0.262673378,-0.0679099038,-0.1828457117,0.226233691,-0.0757433847,0.020079948,0.1824525893,0.1405412853,-0.251835078,-0.1768807322,0.0360903814,0.1681702137,0.2130683362,-0.1704660207,-0.1802038252,0.2561203241,-0.0382318646,-0.3358976841,0.0600581691,0.0034637861,0.3546962142,-0.2224044949,-0.2182590067,-0.2511993051,-0.1353276521,0.3232792318,-0.0828273892,0.0066931015,-0.2888740599,0.2953870893,0.2205252796,0.1083198264,0.2351681292,-0.1334855705,0.3227449954,-0.0152808521,-0.1918017268,-0.2713219225,-0.4650273025,-0.0009278756,-0.0592176616,-0.0043337569,0.424478054,0.2421155274,0.390532434,-0.3245842457,-0.2451094389,-0.4025381804,0.0513647608,-0.1198890358,0.1941350847,0.109929733,0.4666310847,-0.3420136273,0.0523041189,-0.0229985379,-0.0488411263,-0.0397746973,0.0918895826,0.0137830339,0.2035859972,-0.1543302685,0.2484562695,0.0468144976,-0.1075429842,-0.0762664899,-0.0906670019,0.3686672151,-0.4598110914,-0.0609560199,-0.3338884711,-0.0207138788,-0.1050695851,-0.1179809868,0.2465784848,0.4201830029,0.2896746397,-0.1053839698,-0.1386381239,-0.1294739395,-0.1599316448,0.6129691005,-0.2647590637,-0.0381004252,-0.1784407198,0.0093415137,0.1334380805,0.1467615217,0.4467785656,0.0264446288,0.0575173497,0.0608379468,-0.2759607732,-0.2290013134,-0.1141933575,-0.2464484125,0.2201988995,0.5735800862,0.3240065277,0.7751105428,0.1520693004,0.1780719012,-0.051378455,-0.0341672599,-0.048240494,-0.0013013742,-0.2136194706,-0.1924018711,-0.2690327764,-0.0360371023,0.1694476306,0.2143824697,-0.5216483474,-0.0756610408,0.1756911278,-0.1282862574,-0.3454465866,0.2591948509,0.0102546867,0.1966409832,0.3793596029,0.1261278838,-0.4769870639,-0.387755543,-0.0377535447,-0.0887961537,-0.0059976741,-0.1740470529,-0.5687568188,-0.1773722917,-0.7664279342,0.2751237154,0.1402772516,0.0863992423,0.2430044711,-0.2164093256,0.2072665095,-0.1244175658,0.6251431704,-0.067073144,-0.2090532035,0.1638413221,-0.0260799322,-0.6028417945,0.1863063127,-0.135117501,0.1393923908,0.1597448736,0.5891287327,-0.6044453979,-0.200225696,0.0272709578,0.1462673545,-0.1492120177,0.0842126235,-0.0748717338,-0.3018998206,-0.1549275666,0.297275126,0.0412316322,0.1232822612,0.4362004399,-0.0189577918,0.0935054496,-0.0593143627,0.1476088464,0.3977612555,-0.0844948664,0.1703260541,0.0470631942,0.329328984,-0.331858784,-0.0449535623,0.0953390822,-0.0621816665,-0.1341530979,-0.0250850618,0.0251078587,0.2897004783,0.5526978374,0.0097050378,0.3432308137,-0.2175799012,0.2243480384,-0.0855379999,-0.0361493304,-0.1952147484,0.0326361135,-0.3006547093,-0.3092266321,0.3831386864,0.3159516454,-0.3090573549,0.6614328027,0.1723917574,-0.2290065587,0.4607008994,0.3709950745,0.9752556086,-0.5129275322,0.164195478,0.068820864,0.1409442276,0.4058244824,-0.4298228025,0.2410654277,-0.1704446077,-0.0654887632,-0.0597464256,-0.048200082,-0.1479976475,0.2164142132,0.1176715866,0.1895917803,-0.059713535,0.4660337567,0.1189339757,0.1150355637,0.3079616427,-0.5089708567,-0.1210521385,-0.0193635933,-0.0442143977,-0.1744911373,0.0146516385,0.3953194022,-0.2156654596,-0.3960534036,-0.2694065869,-0.3082792759,-0.1529436707,0.0520541668,-0.1991599202,-0.006883007,-0.2173346281,-0.0921342149,-0.3370637894,0.1279224306,0.1543779373,0.0470732749,0.474373281,0.3202131689,0.128852129,-0.1476274282,0.039413318,0.069922924,-0.0086879451,0.1831962913,-0.1061242521,-0.2918617129,-0.296605587,0.1253706068,0.5700806975,0.0377567075,0.0270299315,-0.0743003339,0.184021011,-0.3169250786,-0.0260619372,-0.0776811391,0.0350937285,0.4307331443,0.0009047168,-0.2017875463,0.0209382772,0.3638753593,0.3006711006,0.0080587501,0.2062954158,-0.1151027679,-0.488286674,-0.0527709946,-0.0696430132,0.1498470306,0.0410360694,0.047644835,-0.365226835,-0.1214287877,-0.0371270999,0.2336244732,-0.048573412,0.0217262208,-0.1903949231,-0.4643524885,-0.0804060698,0.4110324383,0.2061308026,0.0998191535,0.0502310991,-0.0862697735,-0.4749446809,0.4416937828,-0.1632984132,-0.1535954773,-0.0957395583,0.4342618287,-0.0482689403,0.2723000944,-0.1264176816,-0.0808697641,0.0900956243,-0.0530615896,0.012638066,-0.1142887995,0.0320478044,0.2085016966,-0.1704892069,-0.1092702448,0.1840132326,-0.1446888149,-0.1090037748,-0.4130848646,0.2695925236,0.0926885381,-0.0381622054,-0.3210903108,0.3805538714,-0.1182939559,-0.1029509827,0.1259240359,-0.1646890491,-0.1021379977,0.0821017548,0.3613914847,0.0257422701,0.0502002798,-0.0561585352,0.1461238414,0.0020497399,-0.1900008917,0.0345628336,0.131363824,0.0368833616,0.1965006143,0.1575856358,0.3292462528,-0.0979901105,-0.0720921084,0.3070291281,0.0204177052,-0.1972117424,-0.2758509815,0.4279967844,-0.1215464398,0.2314718366,0.4082900584,-0.0206354652,0.0943659991,-0.0256963391,-0.272007674,0.5573038459,-0.2318074405,-0.2705376446,0.3658281565,0.2138030231,-0.1216494963,0.0923692137,0.1510497183,-0.1193656847,0.2973027229,-0.0749467164,0.2211153805,0.1364606917,0.0955066755,-0.4624134302,-0.148496598,0.0889827833,0.4589466751,-0.1110121533,0.0684063956,0.0383403786,-0.167571485,0.4820982218,-0.0855100453,0.2303491831,0.3141022921,0.0710153729,-0.2378041446,-0.3685976565,-0.2726429105,0.0645476878,0.0334118269,-0.0883200765,-0.3147929013,0.4633413255,0.0729860738,0.2775478959,-0.3419552743,0.0653923973,-0.2865811884,0.353949666,-0.1692410558,0.0790553316,-0.4063327312,0.2762991488,0.03643528,-0.3371809721,0.1465807855,0.3046513796,0.0188882947,-0.0193372369,-0.3063296676,-0.2333146781,-0.1015716866,0.0132324453,0.0755577162,0.0966714695,0.0049315877,-0.0488101169,-0.0566601343,-0.2034194767,0.2960623503,0.1627049446,-0.014427674,-0.0858473033,0.1537098885,-0.320094347,0.0671884418,-0.1206936911,0.3084366918,0.0249780081,-0.0243552458,-0.1464885026,0.1711666286,0.0486627892,0.007186071,0.0341341645,0.2107687891,-0.0506976433,-0.0489929914,-0.2168074697,-0.3345028162,-0.4528573453,-0.0412931144,-0.7500758171,-0.4522311091,0.33539325,0.2778217196,0.433178395,0.3551092446,-0.042516157,-0.2030104548,-0.3337236643,0.4259215593,-0.1469492614,0.4066602886,-0.0923305675,0.0507738106,-0.04516799,-0.5560513139,0.3085948229,-0.0577566773,-0.0624177344,0.0469005294,0.0726417899,0.2363189161,0.1304043382,0.2742682397,-0.1506303847,0.1588125378,0.0602741688,0.0844879672,-0.1354198307,0.4218728542,-0.5123556256,0.1772307158,-0.086148873,0.229760766,-0.1129934564,-0.114366293,-0.0214145407,0.0569009744,-0.167748794,-0.175637126,0.1880337149,0.1494572461,0.1913279593,0.2306420803,-0.2836628258,0.0284294151,-0.0291078109,0.1143977717,-0.2971754372,-0.2301122695,0.5320161581,-0.8063476086,-0.4839922488,-0.2087682039,0.107952036,-0.0636689812,-0.1608317941,-0.1949755996,-0.3483868539,0.2565123737,-0.1921050251,-0.3577033579,0.022179408,0.0823838562,-0.2597562671,-0.3264108002,-0.0814186558,-0.03755638,-0.0764013082,-0.5289854407,-0.563018024]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"Hi ! Can you open a separate issue for that ?\r\nAlso if you could provide a google colab or a sample code to reproduce this issue that would be helpful.\r\nOn my side I was not able to reproduce this error.","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":42,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\nHi ! Can you open a separate issue for that ?\r\nAlso if you could provide a google colab or a sample code to reproduce this issue that would be helpful.\r\nOn my side I was not able to reproduce this error.","embeddings":[-0.0739275962,0.2622617483,0.010453444,-0.1460129768,0.1726150662,-0.0921144709,0.3422678113,0.5855397582,0.2409633398,0.0138788251,-0.0288185868,0.4858111739,0.0747995228,-0.2420258522,-0.0771973953,0.2121567279,0.0645875633,0.3524308801,0.3226831257,0.0753293633,-0.1231598556,-0.0958926827,-0.2396637797,-0.027308302,-0.1052388102,-0.0756350234,0.2402052283,-0.3865373731,-0.3200216889,-0.1982570887,0.3132258952,0.2609861195,-0.2287148386,0.1817971021,-0.0001276501,-0.0581163168,0.1082803756,-0.0010412417,-0.1381468177,-0.0250625256,0.1099642962,-0.1435803622,0.1595816016,-0.2177058607,0.060991589,-0.1062188968,0.0007782824,0.0064213825,0.0571450256,-0.0390551314,0.0349226594,-0.0823546201,-0.2686341107,0.2206536531,0.5313364863,0.1463045031,0.0038702819,-0.1021067426,0.3115445077,-0.3857913911,-0.1269279718,0.5586366057,-0.4722703993,-0.01489692,0.3759162426,0.0052528437,0.1003192887,-0.2585339546,0.3394195139,0.1346260309,0.2470730841,-0.252373606,-0.074480705,-0.2339121252,-0.0812008679,-0.242116794,0.0057814936,0.0697179139,-0.4549058974,-0.0236453619,-0.2457007021,-0.2589138448,-0.1095361933,0.3977832198,-0.5061330199,0.4292740226,0.1459528804,0.2955040932,0.1822728217,-0.2157247365,-0.0307177976,0.0070946617,0.385294944,0.5473134518,-0.5153126121,-0.1205781698,0.130554989,-0.0011697254,0.2056518495,-0.2471275777,-0.2166626006,0.3844571114,0.3991676569,0.1074465886,0.3933253586,0.1232143193,-0.0664458051,0.616337657,0.3975470662,-0.0870264173,0.0749649554,0.0728815496,-0.0070858588,0.3263713717,0.2278842181,-0.3896753788,-0.152371332,-0.2447281778,-0.0461859107,-0.0112265376,-0.2754251957,0.0753837675,0.1497301012,0.4403749406,0.2603336275,0.2501383126,-0.2040504664,0.0413907431,-0.2743712366,0.0689503774,0.0466026627,0.0728697851,-0.1434799582,0.1552459449,0.123694703,0.1516343951,-0.1084835753,0.0524867922,-0.0320192575,0.048827745,-0.0413984954,-0.2150832564,0.228081733,0.5221583247,-0.1386476308,0.4291599095,0.2292030752,-0.2557504475,-0.3305651546,0.2962751389,-0.2245358974,-0.2434123755,0.0945887342,-0.0493415967,-0.0717157423,0.3505977392,-0.112861231,0.5079697967,0.4323571324,-0.1694267094,-0.0945496261,-0.1912849247,-0.3159533441,-0.2156187296,0.2965351939,0.0022861755,-0.4213631749,-0.1952866614,-0.2165256441,0.2711444497,0.5304780006,0.4404331148,-0.124631308,0.0786661133,0.2929182649,0.4428164065,0.5207013488,-0.0072946679,-0.6128538251,0.1197995842,-0.2150262445,0.0924033374,-0.20398058,0.3250977099,0.6797481179,0.1578985155,0.3617906272,0.1648360789,-0.0221153032,0.2446768731,-0.1920106411,-0.0925857127,0.1381864846,0.0686163083,0.0109994439,-0.1415878683,0.0202317592,0.0064281835,-0.0117277279,-0.0229040515,0.1731658876,0.1595007777,0.1444867849,-0.0474596471,0.1431256384,-0.2540426254,-0.2715231478,0.0644922182,0.3877613842,0.1844186038,-0.3709639311,-0.3565759957,-0.2604984939,0.1445243061,0.3517885804,0.1155855134,-0.0966623873,-0.3039396107,0.1562776417,-0.1436887085,-0.1705742478,-0.0988533571,-0.0689976588,0.0802975222,0.0136118205,-0.0565731153,0.1670723706,-0.1067387238,-0.0188671555,-0.1260797679,0.2792257965,0.2390753031,0.0818327218,0.0696512014,0.2735039592,-0.1891265959,-0.2591295242,0.4771980047,0.0727245584,0.0756566003,0.1348820776,-0.1028706133,0.1460470408,-0.2113201916,-0.442625761,0.2642633021,0.125878945,0.6324936152,0.0156597514,0.1364632845,0.2505248487,0.0159851648,-0.075626336,-0.3630045652,0.0850607082,0.0069791414,-0.0781270936,0.0950096846,-0.0385916159,0.2597822249,0.2616235316,-0.0216930602,0.1133338213,0.365085125,-0.1223695502,-0.2129504681,0.1746501625,-0.1219652891,0.030689029,0.1632504463,0.1556094438,-0.209340632,-0.1445565969,-0.0418502688,0.1894768327,0.2225824594,-0.1069726273,-0.1777325273,0.2512494624,-0.0443398133,-0.3463544846,0.0572749078,0.0166495703,0.3730432689,-0.2785396874,-0.1976841539,-0.2901167274,-0.1320257038,0.304094553,-0.1016241536,-0.0022889264,-0.3022094965,0.3303010166,0.1749198735,0.1361000091,0.2405017763,-0.1292820424,0.3036357164,0.0538430288,-0.1591786146,-0.2672785819,-0.5161728263,-0.0140732713,-0.0943457559,0.0272433329,0.4950301051,0.2363532782,0.3811714947,-0.2868723571,-0.1996533573,-0.4601122141,0.0901708603,-0.1253234446,0.2369723916,0.0723383054,0.5490533113,-0.3248769045,0.0652616397,0.0097933291,-0.0948728323,-0.1028262749,0.181861043,-0.0713994503,0.1781483144,-0.1312668771,0.2205970287,0.0725918785,-0.1176746786,-0.1060684621,-0.116534695,0.3740614057,-0.3518956006,-0.0339260362,-0.2666487694,-0.1013841182,-0.1289845556,-0.1214599535,0.2334732413,0.4066542387,0.2464244366,-0.0668164641,-0.0993939638,-0.048872035,-0.1490194052,0.5926892757,-0.3425274789,0.0154618174,-0.2961544693,-0.0117681092,0.0977891088,0.0548851155,0.5126541853,-0.0307217874,0.0438489467,0.0353364907,-0.3410559297,-0.2484962046,-0.1060235426,-0.1593159586,0.2240360975,0.6323900223,0.2220179737,0.8341838121,0.1447575539,0.2216544896,0.011950735,-0.0426444821,0.026710676,-0.0763693005,-0.2541618943,-0.1907681227,-0.2126714587,-0.1055905893,0.1826873422,0.1845564842,-0.6039087772,-0.0340679176,0.1278661638,-0.1548551917,-0.2923816442,0.2587528527,0.0390253104,0.1774883121,0.3480659723,0.1299901158,-0.4443022907,-0.4553514421,-0.0618755594,-0.1176093742,-0.0040127751,-0.1838155687,-0.6025506258,-0.2310373932,-0.7433963418,0.2902022004,0.0570852757,0.185245201,0.1480816305,-0.1874945015,0.3190456033,-0.1060250476,0.6654680967,-0.088331528,-0.2052764297,0.2013680339,0.0137275709,-0.6145784855,0.2049715072,-0.1121036559,0.1562265009,0.1993316561,0.5296481848,-0.5742838383,-0.2019665688,0.0575856604,0.1104301214,-0.1621996611,0.0468525179,-0.1731902212,-0.3812516034,-0.1731962115,0.2843076289,0.0413601808,0.148856014,0.4511500299,-0.0074324617,0.1518259346,-0.0816291422,0.070124276,0.3923295438,-0.0236002766,0.1812709421,0.0604774579,0.2708532214,-0.2992085218,-0.0366488621,0.1903488785,-0.0141537311,-0.1260766238,0.0524566919,0.0761803389,0.2690075934,0.5618201494,-0.0107324962,0.3450727463,-0.1956552565,0.2114363611,-0.0274570007,0.0478351191,-0.1078942642,-0.0300552789,-0.3140520453,-0.2903091609,0.3591682315,0.269841522,-0.2953546643,0.6984674931,0.1862786859,-0.1885776669,0.5103732347,0.3394595385,1.0705478191,-0.5219205022,0.1128328145,0.0534609854,0.1015219837,0.3691101074,-0.3902153671,0.2997530997,-0.2043590546,-0.0891994312,-0.0158888344,-0.1083137468,-0.1283222437,0.1757344007,0.1561743468,0.1369960755,-0.1475343406,0.5164923072,0.104354091,0.113605395,0.3319136202,-0.454778105,-0.1312395632,-0.0187242012,-0.0495642088,-0.1737531722,0.0326616466,0.3124896884,-0.1600409299,-0.4126213789,-0.3091057241,-0.2302286923,-0.1692227572,0.0307715945,-0.1888955981,-0.0267335996,-0.153873682,-0.1176723093,-0.3874808848,0.1451998651,0.0607582666,0.1745344251,0.4286745489,0.2183490694,0.1385302246,-0.0833631903,0.0809015706,0.0445018746,-0.0414429307,0.1639076173,-0.1121714711,-0.3609417081,-0.3497936726,0.1797660589,0.5157939196,0.0604193099,-0.0660663322,-0.0889270604,0.1701791137,-0.267387569,-0.029057432,-0.1226562336,0.0278317239,0.4215622246,-0.0281728301,-0.32027179,0.0122577492,0.3402383924,0.3003230989,0.0592587441,0.2638503313,-0.1190320924,-0.532137692,-0.0510331653,-0.107472308,0.1750036031,-0.0514305271,0.0395166688,-0.365822643,-0.0924720615,0.0257020537,0.160528481,-0.0497595519,-0.0211828239,-0.1983176172,-0.49512586,-0.0815920159,0.3578316867,0.1081925482,0.0487751067,-0.004460969,-0.0874785557,-0.4045781493,0.4990330935,-0.1465853453,-0.1367083043,-0.1801359504,0.4135910869,-0.0844877288,0.2446781099,-0.0438890569,-0.0772114471,0.0335730016,-0.015907336,0.0133826453,-0.0906627476,0.0758159012,0.242848888,-0.1449742466,-0.1170032993,0.0985661373,-0.2193202525,-0.1534190923,-0.4113562107,0.2653709352,0.0858705491,-0.0441930927,-0.3793169558,0.3709975481,-0.151446268,-0.0596790612,0.1474844813,-0.1912383586,-0.080193758,0.1186013892,0.2735243738,0.0598946027,0.0670486763,-0.1254298687,0.1386251152,0.0044344175,-0.2434275001,0.0321888253,0.012022079,0.0822523162,0.19385086,0.1171853244,0.3672375977,-0.1102286056,-0.076758571,0.3427892923,-0.0209614262,-0.2146289349,-0.3234681785,0.4670250416,-0.0744038895,0.1410490721,0.3350793421,0.0295642633,0.114427574,0.0224723369,-0.2200226188,0.5852678418,-0.2514870167,-0.3143302202,0.3529465795,0.2052788436,-0.1159589216,0.1966900527,0.1865798682,-0.1286758482,0.2911342978,-0.0344511829,0.1995794773,0.1337287426,0.0727759898,-0.4731949866,-0.1582411528,0.0127447266,0.5391823649,-0.125097096,0.0851221085,0.0537658446,-0.1533163637,0.4815630615,-0.1459686309,0.1932910681,0.320451647,0.0353758857,-0.1911688447,-0.3031992912,-0.2898069322,0.0900629759,0.0373030007,-0.0463064574,-0.2341216356,0.4957736433,0.0178611726,0.1024547294,-0.3974160254,0.0679676086,-0.2486336082,0.3070704341,-0.2145027071,0.090138942,-0.4960911572,0.2895284593,0.0633159876,-0.3176859617,0.1985583603,0.361395061,0.0273041129,-0.1305040568,-0.2642959654,-0.2342980504,-0.0443602912,0.0992717892,0.1438566744,0.1085511521,-0.0531789027,-0.0908092633,-0.0422813669,-0.2001724392,0.3480274081,0.2581951618,0.0295270309,-0.1088978872,0.0416664481,-0.3389949501,0.0077637183,-0.0848872289,0.1925149411,0.1259734184,0.0392456651,-0.1577073634,0.1781602055,-0.0043512178,-0.002116984,0.0137805305,0.2925411165,-0.0124635231,-0.0241862163,-0.2196210027,-0.3401702344,-0.5015767217,0.0472729132,-0.8133047819,-0.3624006808,0.4201374352,0.3210954368,0.4156699777,0.280028075,-0.0129214274,-0.2078255713,-0.3595885634,0.3166342974,-0.1748295128,0.4127577245,-0.0687724873,0.078263469,-0.0556000434,-0.4907157719,0.3658676445,-0.1014886126,-0.0907519013,-0.0139151979,0.0890985802,0.2314491421,0.1879885048,0.2940822244,-0.1103710681,0.26828143,0.0706141591,0.108813107,-0.1420528144,0.2802864015,-0.4864358306,0.1273785383,-0.003676598,0.183431074,-0.0712678209,-0.1329747438,-0.0715484992,0.0855358914,-0.2166691571,-0.1765100062,0.2054857165,0.0757352039,0.2305584699,0.2810198069,-0.2449574471,-0.0296314899,0.0169442762,0.0897105187,-0.2991917431,-0.2776012421,0.6093685627,-0.7927552462,-0.5057577491,-0.1406860799,0.1032546461,-0.0435934551,-0.208520472,-0.2053461671,-0.3382758796,0.2539304793,-0.1302684546,-0.3445855379,0.0682523549,0.0589740947,-0.1691321731,-0.2935920954,-0.0291197523,-0.0129508348,-0.1184020638,-0.5401830673,-0.4864801168]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"@lhoestq Sorry I'm just responding now. I'm currently using your recommendation for the map on a single column, and I've gotten it to be fast enough to sort of work for my use case by just setting `num_proc=10`, although it's still quite slow. It's clear that it is still loading the entirety of each row into memory and then discarding everything except the selected column, instead of exploiting the columnar data format to only load the selected column.\r\n\r\nMy code is like this:\r\n```\r\n self.dataset = self.dataset.sort('num_tokens')\r\n batch_dataset = self.dataset.map(\r\n\tcompute_uniform_sized_batches,\r\n\tbatched=True, batch_size=10_000, num_proc=10, input_columns=['num_tokens'],\r\n\tremove_columns=get_columns_all_equal(self.dataset),\r\n\twith_indices=True,\r\n\tfn_kwargs=dict(max_size=tokens_per_batch)\r\n)\r\nself.batches = {\r\n\tname: list(zip(split['start'], split['length']))\r\n\tfor name, split in batch_dataset.items()\r\n}\r\n```\r\nI find that the processes with higher IDs take significantly longer to complete, presumably because the dataset is sorted by article length and they're loading the entire article text into memory, instead of just the 'num_tokens' column.\r\n\r\nI should note that my batching procedure would work best if I just used `batch_size=None` and loaded the whole column into memory at once, but I found that this was intolerably slow and gave me no progress information, so I'm using the less than ideal `batch_size=10_000`.","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":195,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\n@lhoestq Sorry I'm just responding now. I'm currently using your recommendation for the map on a single column, and I've gotten it to be fast enough to sort of work for my use case by just setting `num_proc=10`, although it's still quite slow. It's clear that it is still loading the entirety of each row into memory and then discarding everything except the selected column, instead of exploiting the columnar data format to only load the selected column.\r\n\r\nMy code is like this:\r\n```\r\n self.dataset = self.dataset.sort('num_tokens')\r\n batch_dataset = self.dataset.map(\r\n\tcompute_uniform_sized_batches,\r\n\tbatched=True, batch_size=10_000, num_proc=10, input_columns=['num_tokens'],\r\n\tremove_columns=get_columns_all_equal(self.dataset),\r\n\twith_indices=True,\r\n\tfn_kwargs=dict(max_size=tokens_per_batch)\r\n)\r\nself.batches = {\r\n\tname: list(zip(split['start'], split['length']))\r\n\tfor name, split in batch_dataset.items()\r\n}\r\n```\r\nI find that the processes with higher IDs take significantly longer to complete, presumably because the dataset is sorted by article length and they're loading the entire article text into memory, instead of just the 'num_tokens' column.\r\n\r\nI should note that my batching procedure would work best if I just used `batch_size=None` and loaded the whole column into memory at once, but I found that this was intolerably slow and gave me no progress information, so I'm using the less than ideal `batch_size=10_000`.","embeddings":[-0.1100344881,0.2694859207,-0.0035479604,-0.2267886102,0.1348388344,-0.1574557573,0.3228950202,0.5779768229,0.2516382039,0.0171539988,-0.0084351329,0.5262537003,-0.0003768195,-0.2199472189,-0.0524406247,0.232875973,0.0863020569,0.2980628312,0.323153466,0.1050170362,-0.1054408401,-0.174614951,-0.1898676455,-0.0851241946,-0.080841817,-0.0479384959,0.2107251585,-0.3849202096,-0.2996502519,-0.199813664,0.2933183908,0.2998622954,-0.1688535213,0.1480011642,-0.0001249938,-0.0329319611,0.0593611151,-0.0150163192,-0.1549152732,-0.0841840953,0.1353837103,-0.2224764675,0.0945757329,-0.2144913375,0.050780829,-0.1102484018,0.0543325469,-0.0347234122,0.0267525464,-0.0667863414,0.0478332639,-0.08839605,-0.311880976,0.2385984659,0.4412196577,0.1660660058,0.0005605074,-0.1601057649,0.338162303,-0.3996120691,-0.1329923868,0.4940267503,-0.4622115195,-0.0765878409,0.3967831135,0.0319406912,0.0792658851,-0.2212283015,0.3087268472,0.1258919239,0.1974623054,-0.281825453,-0.0885776803,-0.2058185786,-0.0772382393,-0.2547152042,-0.0588703156,0.0190499406,-0.4656490982,-0.0028989639,-0.2404753566,-0.2685699463,-0.1122994125,0.4409021735,-0.5037668943,0.4319310188,0.2216345668,0.2682172656,0.2149718702,-0.2014353722,-0.0521682873,-0.0211346615,0.4715501964,0.5985535383,-0.4138296247,-0.1830009073,0.1064458266,0.0227847714,0.2341974825,-0.2586019635,-0.1823876202,0.4355725646,0.3492282629,0.1186510623,0.351398319,0.0338762328,-0.0562201031,0.5463657975,0.4081636667,-0.0972744748,0.1370310038,0.0992179886,0.0056041181,0.3168942928,0.2251746953,-0.3960950673,-0.2358844727,-0.1837448627,0.0351988114,-0.0079395445,-0.2091797292,0.0445251577,0.0754380748,0.4448008835,0.2787966132,0.3084744513,-0.2266307771,-0.0192539264,-0.235152185,0.0280104876,0.0535937063,0.0838450044,-0.0949299857,0.180284515,0.1423449516,0.1708377749,-0.1454723626,0.0148306256,-0.0590241738,0.1263824701,-0.0375047661,-0.1837397963,0.2449113876,0.4446518421,-0.149041757,0.3734372258,0.2086372077,-0.1996190995,-0.3181154132,0.3277811706,-0.1003847346,-0.2305619121,0.1103366688,-0.0251126606,-0.0854818076,0.3163959086,-0.0983737484,0.5966129303,0.3636236191,-0.1567471176,-0.1268096268,-0.1794048399,-0.3244422972,-0.2377982885,0.3155627251,-0.0559911542,-0.4163646698,-0.2329698056,-0.1773415655,0.3014839888,0.4253210723,0.3679703176,-0.1312419176,0.0589311086,0.3872956634,0.4368954301,0.5277869105,-0.0276141763,-0.6045199037,0.0641094819,-0.1901936978,0.0697050691,-0.1573721021,0.3717439175,0.6730760336,0.17122823,0.3508662283,0.2020464092,-0.043672543,0.2744268775,-0.1774673164,-0.2010703832,0.1675621867,0.0992528051,0.0320842452,-0.2219464779,-0.0036628288,0.0244350694,0.0568580665,0.0123090195,0.1843397766,0.0709385499,0.1305354238,-0.0961132273,0.1153456345,-0.2739713192,-0.2958954573,0.0815804675,0.384316206,0.1637853831,-0.3264375329,-0.3511686027,-0.181150943,0.0926743522,0.3969911933,0.0968651399,-0.0634365082,-0.3534477055,0.1967143863,-0.1489762217,-0.1934258789,-0.1466518492,0.0024978118,0.104288131,0.0114268316,-0.0093639866,0.2064909786,-0.1165341735,-0.0471429154,-0.1154092625,0.2352760136,0.2410050929,0.1672466546,0.0823833495,0.2372433841,-0.1647633761,-0.3285937905,0.4697247148,0.1023412496,0.1482241452,0.1163063049,-0.1425990164,0.1648341864,-0.1604898721,-0.5011028051,0.2884663641,0.1480735689,0.613444984,0.0310936701,0.1114454791,0.1854691356,0.0051052296,-0.0606541075,-0.3591530323,0.0357489549,0.097132057,-0.1129213646,0.1017361134,-0.0033517568,0.2966711223,0.1944944263,-0.0443623774,0.0852872208,0.3028745055,-0.127077952,-0.1970970631,0.2345538884,-0.1604048014,0.0320270136,0.1957661211,0.0991943032,-0.2818477154,-0.1702195704,0.0116474247,0.1980623305,0.1764254868,-0.1119158268,-0.1793755591,0.2269996107,-0.0561372489,-0.345241338,0.0782023072,-0.0096061574,0.3562093079,-0.2056916207,-0.2208713293,-0.2761364281,-0.1030508727,0.3162029684,-0.1026304811,-0.0073960093,-0.2565502524,0.3267335892,0.2095459253,0.133966729,0.2215714157,-0.1613433212,0.2975682616,0.0058617084,-0.1849832982,-0.2740886509,-0.5016184449,0.0176578164,-0.0662833825,0.0460664444,0.4536599815,0.2173223048,0.4591349065,-0.3362983763,-0.1795960516,-0.3963326812,0.0620957129,-0.125780344,0.2041710764,0.0657641962,0.5384756327,-0.3111215234,0.0434608497,-0.0108580422,-0.1486571133,-0.0590963587,0.1332432777,-0.0285292696,0.1866470128,-0.1014212742,0.213221401,0.0452919416,-0.1045640036,-0.0941851884,-0.0347685255,0.3498975039,-0.41978845,0.0379531607,-0.2923451662,-0.06746158,-0.1448165923,-0.1258461475,0.2819057703,0.433894366,0.2763317823,-0.1003727242,-0.1484422088,-0.0836175829,-0.1696136892,0.5442781448,-0.2675815225,-0.008610039,-0.2266428918,0.0212720837,0.1478111744,0.0819820985,0.5152600408,0.0026169308,0.0346801281,0.0704884082,-0.293710202,-0.2269754857,-0.0938408449,-0.1578365266,0.2495295107,0.5338845849,0.3260044158,0.7934350371,0.165519312,0.1323277801,-0.0513937585,-0.0149193248,-0.041881118,-0.0051874169,-0.1926889122,-0.1869391948,-0.2549144924,-0.0303341802,0.1843656003,0.1866320968,-0.5919137597,-0.0216599684,0.1671774238,-0.0917177424,-0.3089050949,0.2933984399,0.0843449533,0.1379101425,0.3415698707,0.1316806823,-0.4720871449,-0.4646476209,-0.0276712459,-0.1265916973,-0.0257080626,-0.1839925945,-0.5926154852,-0.1910198778,-0.7978946567,0.2661314607,0.1182484254,0.1145836785,0.211372152,-0.2573486269,0.2431063205,-0.1332240552,0.6605279446,-0.0738667622,-0.2266158164,0.1603893191,0.0352549888,-0.6176286936,0.2366262078,-0.084761247,0.1380902678,0.1354446858,0.5151340365,-0.5808650851,-0.1864307225,-0.0015591632,0.1045730412,-0.1565127671,0.0389330797,-0.1103153676,-0.2894618213,-0.1778404713,0.2542255521,0.0160685442,0.1212344542,0.4547694921,0.0177706238,0.1650839448,-0.1009040773,0.1025078446,0.3889807761,-0.0775690004,0.1896748394,0.0409724675,0.3473918438,-0.3493169546,-0.0358289666,0.107276395,-0.0601994842,-0.092104584,-0.0145633426,0.078551963,0.2629893422,0.575407505,0.0660439357,0.3655546606,-0.1661543101,0.1807442307,-0.0545161292,-0.0270435419,-0.1794118881,-0.0380456001,-0.295917958,-0.2920765281,0.3691782951,0.2954030633,-0.300976932,0.6721137166,0.1318915039,-0.1752949506,0.4907180369,0.3872979879,1.0202457905,-0.5151749253,0.1723365486,0.0334805474,0.1508064419,0.3322605193,-0.3836838305,0.2518134713,-0.220879674,-0.0677120015,-0.0308350865,-0.0808012858,-0.1512115151,0.2147798091,0.0942203477,0.1282655448,-0.0474693738,0.5101552606,0.0963324532,0.1301908046,0.308034867,-0.4673937559,-0.0732318461,-0.0013813671,-0.0190655626,-0.2065097243,0.0376517512,0.353654772,-0.2031346858,-0.3892402649,-0.2511255443,-0.2868540287,-0.2169196308,0.0470669083,-0.1523421407,-0.0498477295,-0.1734866649,-0.057511773,-0.426120311,0.0908014253,0.0768958777,0.1112436354,0.466680944,0.2333110571,0.1339196116,-0.1408437788,0.0918302983,0.0707336217,-0.0359872691,0.2393046767,-0.0997866988,-0.3629197776,-0.3013522625,0.1551561505,0.5723900199,0.0323223621,-0.0212125164,-0.0529675782,0.1434834898,-0.2560228705,-0.006222249,-0.1069572344,0.0837673992,0.4150019586,0.0248397943,-0.2222280353,0.0027733098,0.3920274079,0.2615508139,0.0367485322,0.2135914415,-0.1123167872,-0.5209222436,-0.0463464819,-0.0708209127,0.1633602083,0.0231660586,-0.0265082866,-0.3171157539,-0.1171124652,0.0195751768,0.1998368353,-0.063565284,-0.0542227477,-0.2133342475,-0.4395647645,-0.0956658944,0.4088517129,0.1223778799,0.0639256537,-0.0121866511,-0.0877251327,-0.4353480935,0.4807341695,-0.176166296,-0.1444407403,-0.0618591458,0.3912008703,-0.1011151597,0.2439957112,-0.0958538577,-0.1171397939,0.0701480657,-0.0281525888,0.0058826194,-0.1220167428,0.0532571636,0.2138762921,-0.1364699751,-0.1235966235,0.1105300337,-0.1912611425,-0.1196228787,-0.4049907625,0.2113597989,0.0740714297,-0.0717808157,-0.3361969888,0.4533841014,-0.1535314471,-0.054048419,0.1692801416,-0.1950561851,-0.055762507,0.1005159467,0.3331463337,0.0248843431,0.0515390411,-0.0787418187,0.1919612885,0.0267479699,-0.2102126479,0.0367629789,0.1012551486,0.044450663,0.1786355078,0.1139469445,0.3573235273,-0.1206504554,-0.0631423444,0.3542405069,0.0239211097,-0.1925200373,-0.3094851077,0.4062330127,-0.0966023952,0.2209085971,0.3701665998,0.0328850485,0.0865340233,0.003562019,-0.2328653932,0.5390011072,-0.2743962109,-0.3363249302,0.3552633822,0.2743499577,-0.170166865,0.1596386135,0.1653812379,-0.0558429249,0.2616038918,-0.04249046,0.2302852124,0.116556108,0.0800242275,-0.4942270517,-0.1290124655,0.0336452872,0.5174564719,-0.1757681966,0.0897554606,0.0952242911,-0.0697376132,0.4977814555,-0.0909489691,0.2404045761,0.274841398,0.0831557065,-0.2219694406,-0.3533059657,-0.2782159448,0.0571993962,0.0608380996,-0.0785314143,-0.3210173547,0.513825953,0.0657213405,0.1888090372,-0.3860408366,0.1302873045,-0.2775847912,0.3409267962,-0.1915676743,0.0772369131,-0.4894484282,0.2793100774,0.0836868584,-0.3684571385,0.1497601122,0.3017207086,-0.0487072319,-0.0419643708,-0.3215754926,-0.226353839,-0.0916243419,0.0591271445,0.0959443003,0.0303700101,-0.0479739457,-0.0444662906,-0.0575386323,-0.2394274026,0.3788043261,0.1721910089,0.0436850078,-0.111356765,0.1614056826,-0.3382180929,0.0898445472,-0.1136470512,0.2642892897,0.0856457502,-0.0494222566,-0.1938244849,0.1661048979,0.0220229737,0.0127768228,0.017533442,0.2497019619,-0.1321570873,-0.040403422,-0.1920133829,-0.3410747647,-0.4634925723,-0.0235710517,-0.7484756708,-0.3901863992,0.3368365765,0.2894300222,0.4407713413,0.3460671306,-0.0725476518,-0.1604891568,-0.3481918275,0.3894012272,-0.1362974197,0.4280390143,-0.0501279384,0.1046733856,-0.0555283092,-0.5257041454,0.342748642,-0.1517550796,-0.0740507692,-0.027201375,0.0453908481,0.2259249687,0.216302067,0.2984825969,-0.1649972349,0.1771684289,0.0547536016,0.0956382379,-0.1418911517,0.4147330523,-0.5043898821,0.1182054877,-0.0312295575,0.1689738929,-0.1039680913,-0.1184643432,-0.002100338,0.0843592882,-0.2054902315,-0.1556859761,0.1682890356,0.0690801069,0.2219384909,0.2420840561,-0.3701490164,0.0222020391,-0.0190853439,0.1013263986,-0.2939059734,-0.2079076767,0.4910782576,-0.8021829724,-0.4429042935,-0.1894975156,0.0487756915,-0.0894781128,-0.1487003267,-0.1571356207,-0.3269068301,0.2224559933,-0.1530997902,-0.389398545,0.0274792947,0.0684685111,-0.2264362723,-0.3367836177,0.0118871294,-0.0051301918,-0.0711205676,-0.5593317747,-0.5152246952]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"Hi @norabelrose ! I'm glad you managed to make this work on your side.\r\nRegarding memory usage, you can try to drop the columns that you don't want to use for your `map` for now.\r\n\r\nIn the future we'll try to find a way to not load unnecessary columns in memory in `map`. Currently the way it works is that it gets the batch as a python dict, then it updates it using the output of your mapping function, and finally it removes columns from `remove_columns`. Therefore for a moment some columns are loaded in memory even if you remove them or don't use them for your mapping function.\r\n\r\nIt would be nice to have a way to optimize memory for cases such as yours !","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":126,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\nHi @norabelrose ! I'm glad you managed to make this work on your side.\r\nRegarding memory usage, you can try to drop the columns that you don't want to use for your `map` for now.\r\n\r\nIn the future we'll try to find a way to not load unnecessary columns in memory in `map`. Currently the way it works is that it gets the batch as a python dict, then it updates it using the output of your mapping function, and finally it removes columns from `remove_columns`. Therefore for a moment some columns are loaded in memory even if you remove them or don't use them for your mapping function.\r\n\r\nIt would be nice to have a way to optimize memory for cases such as yours !","embeddings":[-0.1090221256,0.2856530845,-0.0173122119,-0.1360481232,0.1743156165,-0.084663257,0.276823163,0.5508022904,0.2953394353,-0.0280731265,-0.0430036187,0.4847539961,-0.0104954401,-0.2168293297,-0.0766484737,0.1780526191,0.0879167616,0.2981182933,0.2514612377,0.1253352016,-0.1527184248,-0.1559893638,-0.1923488528,-0.1331228912,-0.1440608799,-0.07994055,0.186787352,-0.3805733323,-0.2664017379,-0.2372178882,0.2627159357,0.3139121234,-0.1195456535,0.1400264651,-0.000123181,-0.0655137748,0.0379224308,0.0187601745,-0.1874784827,-0.0091844425,0.1147256866,-0.2367014587,0.1002808437,-0.2135514319,0.1023752615,-0.1358706057,-0.0096762376,-0.0671397895,0.0681037381,-0.0473486073,0.0690712109,-0.0756958723,-0.2891823649,0.2594843507,0.4704771042,0.1618333608,0.0502097532,-0.1637221426,0.3472459912,-0.3854161501,-0.192709133,0.5658636093,-0.4568247497,0.0031641799,0.3684160411,0.0001994617,0.1213868111,-0.2628812492,0.35361889,0.120161213,0.2966677547,-0.3108711541,-0.033721447,-0.2469123304,-0.0874592736,-0.2455392182,-0.0481368154,0.0224935692,-0.4655238986,-0.0642245263,-0.2485087365,-0.2779921591,-0.1007028446,0.3768830001,-0.412034899,0.4298382401,0.1846134812,0.2694784403,0.2300573289,-0.2425142676,-0.0344193503,-0.0763403475,0.3317642808,0.5829606652,-0.436930716,-0.1821168661,0.081949614,-0.0360533856,0.2144129574,-0.2783696949,-0.2595427632,0.3622379899,0.3828552961,0.0960842073,0.4457158446,0.1008039266,-0.0724330992,0.5937908888,0.3524028659,-0.1336074769,0.0922416672,0.0355100818,-0.0083410451,0.2875795066,0.30770123,-0.422405988,-0.146506682,-0.1777960509,-0.0036207726,-0.0233573094,-0.2148676962,0.0137573378,0.1195844784,0.462803036,0.2295093238,0.2768382728,-0.1977340877,-0.0248471405,-0.2141761929,0.0515227802,0.0254434757,0.0953226611,-0.1413809061,0.179748863,0.1237722933,0.2349213064,-0.0849079117,0.0241599642,-0.0277580358,0.0716111884,0.0253313668,-0.2372814268,0.2751714289,0.4669004977,-0.1456158608,0.3750057518,0.1949282438,-0.2103823274,-0.3377526104,0.2783410251,-0.1568898857,-0.2745813727,0.1174590066,-0.0101031,-0.0357580669,0.3091068566,-0.1921065599,0.5416850448,0.390511781,-0.1119554713,-0.140958339,-0.1778873354,-0.2912268341,-0.2638456225,0.330737561,0.0047742035,-0.4009059966,-0.2362161875,-0.2102873772,0.2777677476,0.4510690868,0.4538514912,-0.1082937941,0.1115136668,0.3450726867,0.3644188046,0.524923563,-0.1185906529,-0.5643758774,0.0911747441,-0.0897123888,0.1206969321,-0.1512248665,0.4033991098,0.6525670886,0.1655063033,0.3023563027,0.1939819157,-0.0593865886,0.3020563722,-0.154225558,-0.1881451011,0.1357928663,0.074010022,0.0358703509,-0.2453615069,-0.0491573364,0.0437317491,0.0210190117,-0.081031464,0.2167941034,0.0880567208,0.1659963876,-0.0188013967,0.1049728543,-0.251210928,-0.3391766846,0.0935987979,0.3977253735,0.1697494388,-0.3231752217,-0.3669006228,-0.1637741029,0.0814843327,0.3438125551,0.0508630276,-0.0270686317,-0.368604511,0.2010230273,-0.1962224245,-0.145414263,-0.140348047,-0.0127806105,0.0920539126,0.0132701211,-0.0675174594,0.1958551705,-0.0875226855,-0.015383983,-0.047894828,0.2411514074,0.2710720599,0.1745267212,0.0639069453,0.276348114,-0.1552943438,-0.2778624892,0.4768836498,0.0327292457,0.1264176071,0.1915228069,-0.1480120122,0.1535405517,-0.1665282249,-0.4134834111,0.2336407751,0.1651546955,0.6565909386,0.0086745974,0.133165881,0.1882093251,-0.0068121045,-0.0356929004,-0.3979204893,0.0318740234,0.021727547,-0.0778418183,0.1329749674,0.0074173394,0.1980525404,0.2051876485,-0.007843554,0.0878041536,0.3099796474,-0.2001462728,-0.2421217114,0.1981220096,-0.1498928368,0.0176193211,0.2220949829,0.0836346447,-0.255923301,-0.101659596,0.0087207844,0.2313448042,0.2254356146,-0.0668537766,-0.2380709052,0.1779941618,-0.0571115725,-0.306778729,0.0590055138,-0.0237057526,0.3478134871,-0.2445156276,-0.2061146051,-0.2547019422,-0.1980947107,0.34061113,-0.0816507787,0.0087110465,-0.2504313588,0.2866850793,0.1881872863,0.1070009843,0.2564488351,-0.1198146418,0.3113119602,0.029083034,-0.2222064137,-0.2729599178,-0.5268831849,-0.0290297773,-0.0305163953,0.1108557805,0.3845769167,0.3267253637,0.4251878262,-0.2760056257,-0.1865748167,-0.3498575091,0.0308897831,-0.1473956853,0.2165368795,0.0386563912,0.5493985415,-0.3426373005,0.0491896309,-0.0027287239,-0.1080833748,-0.0340949297,0.129964456,-0.0471408553,0.1786493808,-0.0442438684,0.1530379057,0.0591475293,-0.1323919296,-0.1048721001,-0.0440553986,0.3658530414,-0.3394891918,0.0361534506,-0.2583793402,-0.1067750528,-0.1361564994,-0.1385722011,0.2278966755,0.4098497033,0.2368452698,-0.0491841398,-0.0461336486,-0.0541709848,-0.0778101012,0.5534914136,-0.3383617699,0.0059875748,-0.242931217,-0.00154214,0.1468024552,0.1258050203,0.4377745986,0.0251519214,0.0017539191,0.0893941745,-0.2488978058,-0.2838338912,-0.1181722432,-0.1418951005,0.2551566362,0.5586386919,0.2956329584,0.8239872456,0.1558609158,0.2164775133,0.0134823853,0.0023206158,0.0195042118,-0.0187361687,-0.205739364,-0.2271466404,-0.2426349372,-0.0351009481,0.1851104796,0.2311797887,-0.5529456139,-0.0075122123,0.1159802377,-0.0865242779,-0.3321227133,0.3729735315,0.0185178891,0.1834984869,0.3319443166,0.0540244803,-0.4832007289,-0.4939587414,-0.0439363196,-0.1727934629,-0.0043341648,-0.1839491278,-0.59751302,-0.2067459971,-0.7730185986,0.2885383666,0.0515270196,0.0626806095,0.1622111499,-0.1980967373,0.2604497671,-0.0520281345,0.6324747205,-0.1064108163,-0.288744241,0.164031446,-0.0244975109,-0.6307044625,0.2311373055,-0.1395322978,0.1357935071,0.145519495,0.5015463829,-0.5444862843,-0.2731395662,0.022779163,0.160159111,-0.2109158039,0.0412265547,-0.0777449161,-0.3415340483,-0.2475187778,0.2455600053,0.0549145378,0.1496191472,0.5006216764,0.0109202676,0.1289727837,-0.0532472618,0.1641123891,0.3766265512,-0.0041641151,0.2145359665,0.0170877576,0.3097978532,-0.2957212329,0.0482379124,0.09501151,-0.0958029032,-0.1314442903,-0.0914095864,0.0859841257,0.342150569,0.5212577581,0.0167562384,0.3089970648,-0.1656335592,0.1997567862,-0.0987809598,0.0243706238,-0.1273795217,0.0123826982,-0.3444961309,-0.3687979579,0.348382175,0.3398561478,-0.2826129198,0.7005108595,0.1345766187,-0.2693924606,0.4919739664,0.4547047913,1.0030083656,-0.5092764497,0.1918296069,0.1041995287,0.1524888575,0.3764023781,-0.3755291998,0.2738724649,-0.1744574904,0.0193383526,0.0076186005,-0.0686173812,-0.103651464,0.1873567998,0.1057574898,0.1570839733,-0.1172845438,0.4785140157,-0.0023792356,0.16668275,0.2566238344,-0.5155163407,-0.125604555,0.0300085954,-0.0233561751,-0.2395553291,-0.0016504325,0.360432893,-0.1163911521,-0.4067580402,-0.2394730747,-0.3339540064,-0.176479727,0.0764963999,-0.2180508077,-0.0000617843,-0.0966370031,-0.0341447815,-0.3839152157,0.0999607369,0.1196131259,0.1239837408,0.4367550313,0.2180314809,0.0789832473,-0.1147110462,0.0638253689,0.0545119457,-0.030629525,0.2301749736,-0.0844710618,-0.3699976206,-0.3310753107,0.1422875971,0.5918899775,-0.0156173985,-0.0570480376,-0.0731878728,0.1384318322,-0.1705451459,0.0139001859,-0.1172701642,0.0375611857,0.4882902801,0.0093265092,-0.2604594231,0.0049470593,0.2434514612,0.3391028047,0.0581786484,0.2930648923,-0.0815019757,-0.498718977,-0.0857777074,-0.0699472949,0.2039607167,0.0262368321,0.0186411776,-0.2155817002,-0.0963979438,-0.0305516906,0.1936423779,-0.0464116298,-0.0499742255,-0.1869866997,-0.4001560807,-0.1264634579,0.412748307,0.156798631,0.1442384571,-0.0115681849,-0.1527849436,-0.4431920052,0.5253605247,-0.1937604249,-0.1797101945,-0.0994731039,0.4249873161,-0.0759303644,0.2984488904,-0.0215442237,-0.1423902363,0.0567687936,-0.0781058595,-0.0279935822,-0.1443944573,0.0546853468,0.2024955899,-0.167041868,-0.2009327114,0.1073825657,-0.2026349157,-0.0884774774,-0.4500786662,0.2485097051,0.0877198875,-0.0673259795,-0.3250293434,0.3965335488,-0.2247069776,-0.0517582186,0.217476204,-0.165226683,-0.0836824402,0.1474944353,0.3286175132,0.1112592295,0.032710772,-0.0768908262,0.1567069143,0.0865010172,-0.2299770266,-0.0654055178,0.0911867768,-0.0222369004,0.2265074104,0.0777576193,0.3255521059,-0.1188305095,-0.0701999217,0.3611966968,0.0489182211,-0.1887324005,-0.2899717391,0.3760865331,-0.1546105742,0.1473887861,0.353082329,-0.0040374245,0.0852662474,-0.0218479596,-0.2005933523,0.5574969053,-0.2597492635,-0.2453697771,0.2899138033,0.2669285238,-0.186950013,0.1434138864,0.2223601192,-0.0730319917,0.2246967256,-0.0497388914,0.1983326226,0.1444982439,0.0901541561,-0.5114418268,-0.121196568,0.0495831594,0.5318866372,-0.1830295026,0.0672331676,0.0398365296,-0.1401380897,0.5727429986,-0.1642997414,0.1831101775,0.2923385203,0.0271730442,-0.2031528056,-0.3153848648,-0.2637041509,0.1272265613,0.0498836115,-0.0062657227,-0.401766032,0.4743148983,0.0501123033,0.1437917352,-0.3458489478,0.1734967977,-0.180464223,0.3057919443,-0.2012145519,-0.0011616766,-0.4704732895,0.2981130481,0.053574793,-0.2909857035,0.2042212635,0.3116222322,-0.0056867856,-0.045510944,-0.2424696833,-0.2219530493,-0.1456355304,0.0876774117,0.0622467957,0.0538623258,-0.081843473,-0.0303669032,-0.0568741634,-0.2863596976,0.3476958871,0.1824982911,0.0264550615,-0.081361331,0.1229305565,-0.3137011826,0.0713341907,-0.0916880518,0.202140376,0.0537368655,0.0780517384,-0.193275094,0.1744870394,0.0030703393,0.0275521334,0.0026182549,0.2995877266,-0.0786050111,-0.0460828729,-0.2731840014,-0.3212508559,-0.4879871309,-0.0796559751,-0.8037527204,-0.3432793319,0.317146033,0.2904569507,0.3600548506,0.33194682,-0.0512447841,-0.1530701369,-0.4010699689,0.3075761497,-0.1426282823,0.3985101283,-0.0483067706,0.1764246076,-0.0004775208,-0.4729259908,0.3123589456,-0.1248280779,-0.0591295399,-0.0413188227,0.0913440734,0.25510934,0.2520376742,0.3186087012,-0.1116294265,0.2177822888,0.0628048033,0.0907694772,-0.1198382154,0.3120602071,-0.4494713247,0.0692379028,0.0040289224,0.1768660396,-0.2057840228,-0.1238448322,-0.0555479676,0.1101205572,-0.2011115253,-0.1833705902,0.2157503068,0.0367242172,0.261010468,0.2835199237,-0.3485640287,0.023250265,-0.025329046,0.0744886771,-0.3010801077,-0.2312053442,0.5163487196,-0.8135402799,-0.4936999381,-0.1943000704,0.1085390449,-0.0466180593,-0.2379191816,-0.1640188545,-0.2879093885,0.2436066568,-0.1206541806,-0.334729135,0.0481671579,0.0469665863,-0.1878629625,-0.3019767702,-0.0490140542,-0.0736180022,-0.1330049783,-0.5351458192,-0.4960491061]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"@lhoestq After looking through the source code, it looks like the following solution has at least some chance of working:\r\n- refactor `Dataset.map()` so that the `input_columns` parameter is implemented by using the `self.formatted_as()` context manager with `columns=input_columns`\r\n- change `Dataset._getitem()` so that it passes `self._data.drop(drop_columns)` to the `query_table()` function whenever `format_columns` is non-None and `output_all_columns` is False, instead of `self._data` itself","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":62,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\n@lhoestq After looking through the source code, it looks like the following solution has at least some chance of working:\r\n- refactor `Dataset.map()` so that the `input_columns` parameter is implemented by using the `self.formatted_as()` context manager with `columns=input_columns`\r\n- change `Dataset._getitem()` so that it passes `self._data.drop(drop_columns)` to the `query_table()` function whenever `format_columns` is non-None and `output_all_columns` is False, instead of `self._data` itself","embeddings":[-0.1273223609,0.3481048346,-0.0064757606,-0.1535289884,0.143802911,-0.1467355937,0.3634429872,0.6069020033,0.1660746187,-0.0057872017,-0.1035190299,0.5548501611,0.0025605711,-0.195535183,-0.0578183644,0.1826296151,0.0585309118,0.2830992043,0.2737965584,0.1041020751,-0.0951171964,-0.170508489,-0.2150659114,-0.0666617304,-0.0642238408,-0.0613076687,0.1991731524,-0.3850052655,-0.2869391739,-0.2271060199,0.2922720015,0.3593518436,-0.1801968366,0.1267358363,-0.0001259425,-0.0458019748,0.0389490537,-0.0200659167,-0.1705733687,-0.0551256351,0.1479955465,-0.1761647463,0.0846670493,-0.2403679639,0.0551327839,-0.132731244,-0.010619347,-0.0117523298,0.0406139717,-0.1128470451,0.0341307744,-0.0340769887,-0.30176121,0.2565586567,0.4595094919,0.1940970868,0.0270513613,-0.1416135132,0.2707462907,-0.3485600352,-0.1681900769,0.5096774101,-0.4922267795,-0.0580939725,0.3593020141,0.0253371317,0.1058124229,-0.2550091445,0.2683979869,0.1529864371,0.2921063602,-0.3296048343,-0.1517493427,-0.3154854774,-0.0439147316,-0.2649434507,-0.0707568526,0.0200015157,-0.4546419084,-0.015496376,-0.2215270847,-0.2403392196,-0.1327635348,0.4017274976,-0.4890250266,0.4664961398,0.1749093831,0.2699730396,0.200843811,-0.2302258313,0.0042692474,-0.0303562582,0.4081297517,0.5703794956,-0.4117674828,-0.2143025249,0.1393150836,0.0036534108,0.2306438237,-0.192380473,-0.2197024673,0.4396784306,0.300208807,0.1032507345,0.3937817514,0.0823249146,-0.0117408941,0.6130168438,0.4194803238,-0.0690276697,0.1520483792,0.1106830686,-0.0377936661,0.2828658521,0.191223368,-0.2766187787,-0.1490796655,-0.1855109036,-0.0168785043,-0.0178032033,-0.2466075867,0.0188309085,0.0866866261,0.3693024218,0.255395025,0.3317075372,-0.1854817718,-0.0037354205,-0.2259202898,0.0245300177,0.0904608294,0.0658389032,-0.0753162652,0.1368270069,0.1243725345,0.2003501356,-0.1117541268,0.019388197,-0.01117648,0.1090581715,-0.0483329743,-0.1529540718,0.2417969853,0.4411389232,-0.1661824584,0.376127094,0.2155161649,-0.3126173317,-0.3832078576,0.243093282,-0.1104761735,-0.1868739128,0.0666751862,-0.0289041717,-0.0771972388,0.3230990767,-0.1639081389,0.5878397226,0.3482172489,-0.1493635476,-0.0980278626,-0.1733254343,-0.2845931351,-0.2081663311,0.3253329992,-0.0149536151,-0.4749377966,-0.2395092696,-0.1989283413,0.2556532323,0.4086407721,0.3316078484,-0.1002431139,0.1415997148,0.3397714496,0.4450962842,0.5255149603,-0.0780069679,-0.5603324771,0.0560347475,-0.195227325,0.0996111184,-0.1634822041,0.3420075178,0.7027949691,0.1154097021,0.2961075008,0.2033743709,-0.0511499047,0.2437443584,-0.1246561781,-0.2059466541,0.175549373,0.1367104352,0.0292530227,-0.2447059751,0.0477044843,-0.0039334428,0.0220564716,-0.0478291288,0.2118877769,0.085820958,0.1237950474,-0.0884239301,0.0869387686,-0.2819356322,-0.3313639462,0.0774918869,0.4104803205,0.224770993,-0.3545203507,-0.4367652237,-0.184704259,0.1123172939,0.3844044507,0.0683844462,-0.0719927326,-0.3210816681,0.1843955815,-0.1295587718,-0.1944354177,-0.1407885402,-0.0147879655,0.1092911884,-0.0360376053,-0.0412913896,0.2478711307,-0.0388693698,-0.0346045941,-0.0625139251,0.2811250389,0.2168859392,0.1725888848,0.0954233855,0.242354542,-0.1273519695,-0.3383031189,0.451720655,0.1155748665,0.1782616079,0.124825798,-0.1141844913,0.1300780028,-0.1701489985,-0.5284741521,0.3342117369,0.1444132924,0.6386406422,0.0327836908,0.1267504692,0.1661235839,0.0019883774,-0.1329389066,-0.3860221803,-0.0053127944,0.0512732752,-0.0726254508,0.0961115584,-0.0275738686,0.3275288939,0.2445507646,-0.0626856685,0.1090940237,0.2772999704,-0.1371352375,-0.1556412578,0.2355410755,-0.0841586217,0.0132603263,0.2041249573,0.1494021118,-0.1876815856,-0.1209459975,0.0262486003,0.2114543319,0.2394544929,-0.1365422308,-0.1613265574,0.2345717698,-0.0368097015,-0.3467926383,0.0188298747,0.0038079813,0.3276860118,-0.2645940483,-0.2206591964,-0.3071566522,-0.1773902923,0.3076581359,-0.1017003134,-0.0127009246,-0.2653778493,0.3173015714,0.2154940963,0.0401482843,0.3119371831,-0.1661814749,0.3087298572,0.0243713856,-0.1637331843,-0.2447717935,-0.5096173286,0.0633461922,-0.0686664283,0.0641943738,0.4734525084,0.2361041605,0.4196451008,-0.3037250042,-0.2142433971,-0.432030201,0.0937002823,-0.1424224228,0.2290282249,0.1384034902,0.5033736825,-0.3042500615,0.0128717776,-0.0201114863,-0.0442610271,-0.0917531401,0.0790370256,-0.03497307,0.2020378113,-0.1302694082,0.2140203863,0.0292718131,-0.1585879326,-0.0912216082,-0.0908362567,0.3771420419,-0.3267349005,-0.0654540583,-0.2527085543,-0.0476640612,-0.1487681717,-0.0874359533,0.2180919051,0.4490581453,0.2227489203,-0.0433771126,-0.1537193656,-0.0969225988,-0.1233660877,0.5705345869,-0.2659854889,-0.0307171028,-0.1824514568,0.0891898721,0.1391223222,0.1379883587,0.4991467595,0.0553051382,0.0428186394,0.0685393363,-0.2809600532,-0.2711595297,-0.0988806561,-0.2212449312,0.2532505989,0.6047789454,0.2968128026,0.8350998163,0.1069739163,0.1818827987,-0.0387490764,-0.0171391442,0.0519235581,0.0148791205,-0.20782125,-0.1637065411,-0.2221999913,-0.0640083924,0.1698118597,0.1435193866,-0.551061213,0.0033139947,0.1110485718,-0.1422188729,-0.3159610927,0.257117033,0.0388079472,0.1549913436,0.3217479885,0.1841271818,-0.4510055184,-0.414239794,-0.0209762584,-0.1442963779,0.0042879302,-0.1905401498,-0.6768003702,-0.1820839942,-0.7496144176,0.272426039,0.1283871382,0.1044469997,0.2136636674,-0.2561153769,0.2302795351,-0.1493795365,0.6817163825,-0.1297953129,-0.2012693733,0.2135451138,0.0631826594,-0.6210018396,0.2086020261,-0.0594422929,0.176992029,0.1192829534,0.498986721,-0.5984188914,-0.242306143,0.0295483638,0.1255254745,-0.1567388773,0.0966097638,-0.0112243993,-0.3419350684,-0.1949516088,0.2824064195,0.068662405,0.1367495805,0.4177076817,0.0197617076,0.1251194179,-0.0417043008,0.0763844177,0.4052557051,-0.0119333155,0.2226386219,0.0638302639,0.3394083083,-0.3068763316,-0.0153823197,0.0988523737,-0.0776726827,-0.1732518077,-0.0052550342,0.0673845112,0.2934234142,0.5596963167,0.0236438736,0.3625144958,-0.1349499822,0.2042379081,-0.0689314455,0.0086079463,-0.1927107871,-0.0298148226,-0.3097382486,-0.3103672862,0.3756297529,0.2721712291,-0.2979207337,0.6662807465,0.1824174523,-0.2572315931,0.4809890091,0.3857996762,0.9818708897,-0.4391013682,0.1944669187,0.0656752214,0.12380822,0.4268319309,-0.3922558427,0.2636387646,-0.2277778983,-0.0156035116,-0.0482393429,-0.066324383,-0.1452963203,0.2381993383,0.1275870949,0.1684179306,-0.1042028144,0.5085713267,0.0684886351,0.1353707165,0.2738368511,-0.524356246,-0.1250846684,-0.015376471,-0.0862962604,-0.2240331471,0.029722549,0.3337557912,-0.1380545795,-0.3959973156,-0.2638469636,-0.2780079246,-0.1923871636,0.0325113088,-0.172145173,-0.0582662299,-0.1527361274,-0.0771735162,-0.3341743946,0.083720386,0.1058537886,0.1438482255,0.4252270758,0.3069601953,0.1827412695,-0.1586415321,0.0996921733,0.0837217793,-0.0102693681,0.1840320081,-0.0719709992,-0.3719846904,-0.2996486723,0.1372822672,0.5914369226,0.0275001563,-0.0068545407,-0.0893053487,0.1800944954,-0.2673337162,-0.0323387869,-0.1305841357,0.0310469195,0.4102390707,-0.051787205,-0.2097820193,0.0114972983,0.3610177338,0.2665208876,0.0627219602,0.2423955351,-0.0836216658,-0.5424603224,-0.0597628243,-0.02989427,0.1459826529,-0.0243045613,0.0082402425,-0.3483137488,-0.1090193018,-0.0436425097,0.1970876157,-0.108927846,-0.0193238575,-0.2348047197,-0.4801256359,-0.1004069224,0.4124408662,0.1363637,0.1101143286,0.0255638845,-0.1032995209,-0.4717288315,0.4218456149,-0.1611301154,-0.1528005302,-0.0824985728,0.4136499166,-0.0525787361,0.2995499372,-0.0834406465,-0.1278388947,0.0404472947,-0.0157471914,0.0060281726,-0.1082221121,-0.0180257838,0.2158908695,-0.1520169526,-0.1767698824,0.1595075876,-0.161010161,-0.1283530891,-0.4170399308,0.2688875198,0.0739142969,-0.024800064,-0.2930168211,0.3756003976,-0.1098565757,-0.0850409865,0.1856269836,-0.0925673842,-0.0441623554,0.1314783394,0.3624938428,0.0163754504,0.102339074,-0.0212152377,0.1830400079,0.0764910951,-0.2310966104,0.0298056118,0.0579439253,-0.0244719405,0.2433531582,0.17060332,0.3392596841,-0.1147817895,-0.0680904239,0.4051098824,0.0090434328,-0.2317363322,-0.3160851896,0.4245131612,-0.1597135365,0.2071770281,0.3458247781,0.0047682021,0.078890264,0.0172809251,-0.202501148,0.6030319333,-0.2612032294,-0.2481488436,0.3932270408,0.236740306,-0.1405761838,0.1525176466,0.1961874217,-0.1034132317,0.209977597,0.0134104369,0.2117497772,0.1433825195,0.0495016836,-0.4571825862,-0.1390480697,0.071952574,0.4697214961,-0.0802283287,0.0991027653,0.103409566,-0.1375617087,0.5090923309,-0.1485086381,0.2394255996,0.2918792367,0.0518301502,-0.2284321934,-0.388102442,-0.271740973,0.0887424052,0.0431637391,-0.0877015442,-0.3937123716,0.4382043481,0.0317190476,0.1805929244,-0.449778825,0.0854621679,-0.2634504437,0.3338375092,-0.2152819633,0.0529084243,-0.4307364523,0.2762869895,0.0397213101,-0.3055504858,0.1705830693,0.2844312489,-0.0580198579,-0.1117850989,-0.2988744974,-0.1893373877,-0.0749962032,0.0306660067,0.0724835098,-0.0202625375,-0.0377770886,-0.0616244003,-0.0431645736,-0.2216572464,0.3707210422,0.1556831151,-0.0224830359,-0.1269090772,0.1448960304,-0.3175268471,0.1032469198,-0.1658274978,0.2675752342,0.0687755048,0.0433632247,-0.1203856245,0.1901073307,-0.0016944879,0.0097624958,0.0085388869,0.2075282186,-0.0391885601,-0.0659175292,-0.1844328344,-0.2827776968,-0.5424228311,-0.0785276145,-0.6783168912,-0.4402120709,0.3769552112,0.2896398008,0.4799543619,0.3359672129,-0.0373755395,-0.1987977922,-0.3016084731,0.3532307148,-0.154874146,0.3796235025,-0.0948214382,0.069667384,-0.0461302139,-0.5607630014,0.3027586043,-0.0647452846,-0.0920319706,-0.0119486665,0.107428506,0.251044184,0.1420669258,0.2661870122,-0.1468736529,0.1501161307,0.0665975139,0.1073784083,-0.1670599431,0.3415957093,-0.5109817982,0.1398368925,-0.0354842208,0.2615741789,-0.1015487537,-0.1498741657,-0.0644224957,0.0485894941,-0.1707939804,-0.1590428501,0.1794200987,0.1091797426,0.1894636303,0.2357088327,-0.3155451715,-0.0114716906,-0.0222352296,0.1398939937,-0.2780249119,-0.2221183777,0.549346745,-0.8324962258,-0.479265362,-0.1653926075,0.070106715,-0.0641957745,-0.1915352792,-0.1486912072,-0.319442451,0.2589017153,-0.1735186875,-0.3030003607,0.0436290056,0.0690993443,-0.2117331773,-0.3254095614,-0.0479791276,0.0009032061,-0.1098849773,-0.5369403958,-0.534882009]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"Looks like a great direction :)\r\nNote that `query_table` doesn't bring data into memory. Only `format_table` does.\r\nAlso the dataset may already have a format with `columns=` already defined so we would need to define the formatted `input_dataset` like:\r\n```python\r\n# before the `map` main for loop\r\ninput_columns = input_columns if input_columns is not None else self.column_names\r\nif not self._output_all_columns:\r\n columns = [col for col in input_columns if self._format_columns is None or col in self._format_columns]\r\n input_dataset = self.with_format(\r\n type=self._format_type,\r\n columns=columns\r\n )\r\nelse:\r\n # in this case we could find a way to filter both format_columns and unformatted columns eventually\r\n input_dataset = self\r\n# then input_dataset can be used in the main for loop of `map`\r\n```\r\n\r\nEDIT: oh and regarding streaming format versus file format for arrow, we plan to start using the file format #1933 at one point (though I'm not sure if it would improve performance)","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":148,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\nLooks like a great direction :)\r\nNote that `query_table` doesn't bring data into memory. Only `format_table` does.\r\nAlso the dataset may already have a format with `columns=` already defined so we would need to define the formatted `input_dataset` like:\r\n```python\r\n# before the `map` main for loop\r\ninput_columns = input_columns if input_columns is not None else self.column_names\r\nif not self._output_all_columns:\r\n columns = [col for col in input_columns if self._format_columns is None or col in self._format_columns]\r\n input_dataset = self.with_format(\r\n type=self._format_type,\r\n columns=columns\r\n )\r\nelse:\r\n # in this case we could find a way to filter both format_columns and unformatted columns eventually\r\n input_dataset = self\r\n# then input_dataset can be used in the main for loop of `map`\r\n```\r\n\r\nEDIT: oh and regarding streaming format versus file format for arrow, we plan to start using the file format #1933 at one point (though I'm not sure if it would improve performance)","embeddings":[-0.1205226481,0.2916710377,-0.0068580708,-0.1955762655,0.1112996936,-0.1071509719,0.2545422614,0.599486053,0.2043481618,-0.0285109952,-0.0340009406,0.4793538153,0.0217377339,-0.2131702602,0.0204914026,0.225736469,0.1094413996,0.354948014,0.2541983724,0.0966203585,-0.1538394839,-0.1982913762,-0.1836565286,-0.1087915748,-0.0737467185,-0.0132568274,0.1467769593,-0.3689957857,-0.3287634552,-0.1571341157,0.298129797,0.335116595,-0.1511103958,0.1153092608,-0.0001276158,-0.0219644252,-0.0039140563,-0.0525664724,-0.1830680817,-0.0591208749,0.1285026968,-0.2101277411,0.0795893446,-0.2305582464,0.0204804,-0.0886209607,0.0288893171,-0.0736887231,0.0148101766,-0.1144906878,0.0261801705,-0.0833349451,-0.3116624951,0.2302701473,0.4493208528,0.1998949051,0.0042839465,-0.1619109958,0.2894332707,-0.4011121392,-0.1172273755,0.5342761278,-0.4963223338,-0.0241851322,0.3617823124,0.0233744979,0.0134989619,-0.1979040802,0.3108482361,0.1128364354,0.225044027,-0.3431988657,-0.1086047068,-0.2292713672,-0.0838759914,-0.269415468,-0.0652150139,0.0585792288,-0.4512800276,-0.0089030201,-0.2625634074,-0.2576969862,-0.0991568193,0.4092383683,-0.5199321508,0.4321989715,0.1646626443,0.2750136256,0.1986683756,-0.219545126,-0.0431810357,-0.0297076963,0.3957521319,0.5969600677,-0.4020283818,-0.2055867016,0.1378192455,-0.0239012912,0.2302290201,-0.2287324965,-0.1678272933,0.3932457268,0.3833607137,0.1029802188,0.3581872582,0.006378524,-0.0522425734,0.6364861727,0.3577840924,-0.1339600235,0.1019664109,0.0639413446,-0.032211531,0.3444749415,0.2406603545,-0.3781087399,-0.1642122865,-0.1726717353,0.0016429852,-0.0048700697,-0.2048109621,0.050323423,0.075477168,0.4536677897,0.2588711381,0.2259848565,-0.1784389615,0.0110360645,-0.2188385576,0.0681120902,0.066275686,0.058855217,-0.0594325513,0.2203586847,0.1725183129,0.1975576878,-0.1392242461,0.0358778499,-0.0783822462,0.0963132307,-0.0028510378,-0.2358976901,0.3039046824,0.4720265269,-0.1210360602,0.3792858422,0.2375580519,-0.216723755,-0.347665906,0.3116398752,-0.1199537367,-0.1991361976,0.1178496629,-0.039899908,-0.1022518724,0.3331950009,-0.0838698894,0.6144601107,0.3624878526,-0.1243353635,-0.1075508147,-0.1345472783,-0.32340464,-0.2182026207,0.373061657,-0.037768282,-0.4412716627,-0.2263315171,-0.150397718,0.2870781124,0.401794672,0.3283833265,-0.1041111872,0.1448107213,0.3399391472,0.4092049301,0.5807511806,0.0020424912,-0.5462443829,0.0586960576,-0.1678364724,0.1042349115,-0.1550773829,0.3828926384,0.6717818975,0.1935446262,0.338942349,0.2051123232,-0.0929735601,0.2500671148,-0.0852625221,-0.1095489711,0.1974383742,0.1217210665,0.0077488422,-0.1483389735,-0.0093657142,0.0399956107,0.0263680834,-0.0251079723,0.1214189082,0.0646109208,0.1171696782,-0.0716163442,0.0811888129,-0.2945955396,-0.2906388342,0.0719172806,0.4683172405,0.1420728415,-0.3408866823,-0.3782500625,-0.1787259281,0.143328771,0.4013044834,0.1306817532,-0.0783435181,-0.3235485256,0.2239701152,-0.1672380716,-0.2348143458,-0.1474268734,-0.0294382703,0.0988392904,0.003334074,0.0025427374,0.2784744501,-0.0976318642,-0.032074526,-0.0948892683,0.2482024878,0.2585181594,0.2177055478,0.0754668862,0.2817989886,-0.1683757156,-0.2904419005,0.4979357719,0.0264392029,0.1684324294,0.1303661019,-0.20779562,0.1808964908,-0.1695896238,-0.4422305822,0.3461422622,0.1428615749,0.6261826158,0.0383623801,0.0879711881,0.1729707569,0.0151635632,-0.1029490754,-0.3706105947,0.0246630087,0.0675802454,-0.1296297461,0.1175779104,-0.0286361109,0.262838006,0.2400072962,-0.0757464692,0.1183066368,0.2568657994,-0.1232646853,-0.1520133466,0.1783413589,-0.1688497663,0.0163515192,0.1746796221,0.138066262,-0.2322896272,-0.1784208417,0.0018266228,0.2231835127,0.1911960244,-0.1504746377,-0.2071059197,0.1783090085,-0.0749041289,-0.3234237432,0.0882717744,-0.0030709892,0.3133063614,-0.2282298654,-0.2052068114,-0.255549252,-0.1922771633,0.3304744959,-0.1014569253,0.0395308957,-0.2535791099,0.3107281327,0.1738023013,0.0868997648,0.2332748622,-0.1845206171,0.3074072599,0.0328975953,-0.2708673775,-0.2649327815,-0.4880022407,0.0622917265,-0.0660041645,0.0105123185,0.434378624,0.2581012845,0.3665306568,-0.2448257804,-0.2507920265,-0.3883439302,0.0792067349,-0.136558488,0.2166884094,0.1166438758,0.5547269583,-0.3538579941,0.0443881303,-0.0130543523,-0.0474988818,-0.0637936145,0.1721643656,-0.0247787423,0.1852326244,-0.0968935937,0.2254283726,0.0812081844,-0.0963438004,-0.1229767874,-0.0644629002,0.3656049371,-0.3716361523,-0.0098516112,-0.2970950603,-0.1427563876,-0.1317302138,-0.1191973835,0.2455678284,0.4853591323,0.2650068104,-0.0503653735,-0.1151808277,-0.1063332558,-0.155970037,0.6325244904,-0.3378854692,0.0153188799,-0.2393226922,0.0202769823,0.1652296036,0.123026289,0.4506605566,-0.0520643629,0.0475989059,0.0480048731,-0.2972130179,-0.2780140638,-0.0886455327,-0.2216757834,0.2424536943,0.644724071,0.3033609986,0.7878698707,0.0838349983,0.1795894206,0.0335374475,-0.0619435869,0.0475945324,0.0101660993,-0.2111126184,-0.2400131226,-0.2072144598,-0.0753114,0.1416136771,0.2063371986,-0.5454421043,-0.0550727807,0.164457351,-0.140335694,-0.3017390668,0.3235955834,0.0665494725,0.190079689,0.3558623195,0.1475870162,-0.4759306312,-0.4431441128,-0.0849097669,-0.116017133,0.0462106094,-0.1550174952,-0.6074303389,-0.2667904198,-0.7115900517,0.2854957879,0.0791328996,0.1452513635,0.2168409824,-0.2947318554,0.220400393,-0.0925423652,0.6694387794,-0.0520819314,-0.2414198071,0.1431227475,-0.0561401509,-0.6001130939,0.2190056592,-0.076991424,0.1723325402,0.0756415948,0.5250889659,-0.5840340853,-0.1827317327,0.0584263168,0.1571159512,-0.1851092279,0.0542646348,-0.1085367873,-0.3439580202,-0.1471361369,0.2608181834,0.0282211285,0.107977502,0.4737846255,-0.0192326121,0.0674177408,-0.0903822929,0.0966001898,0.4231258035,-0.0435207076,0.2096159458,0.0681550577,0.3000501394,-0.3356435895,-0.0294003654,0.1067852154,-0.0257600751,-0.0809923336,0.0154738165,0.0973517373,0.3004748523,0.5683048964,0.0342305414,0.3881701231,-0.1295009404,0.1878234148,-0.1390419751,0.0039841267,-0.200626716,-0.0432831645,-0.3406222463,-0.3927215636,0.3689990342,0.3045246303,-0.2760222852,0.653116107,0.1705692708,-0.2066047788,0.4587948024,0.3477600515,1.0195086002,-0.5214763284,0.1699344665,0.0322636887,0.1440646201,0.3666465282,-0.4250974357,0.250038147,-0.1836036146,-0.0954937115,-0.0526848584,-0.1069662198,-0.1003898755,0.1931835115,0.1298272759,0.1475315243,-0.099160254,0.5432235599,0.1006645039,0.0851835832,0.2250944376,-0.442933172,-0.1512704343,-0.019622799,-0.0472354665,-0.2359918058,0.0416934676,0.3622329831,-0.1557518095,-0.4161446691,-0.2355289459,-0.2978554964,-0.2282820195,-0.0113713564,-0.2222597748,-0.0533126891,-0.2118134052,-0.081831798,-0.4060822129,0.1428719163,0.1504830271,0.0990865454,0.4586839378,0.2568593919,0.1095416844,-0.1178553626,0.1019914895,0.1032886654,-0.0114836963,0.2659314573,-0.0809335932,-0.3675947189,-0.3423034549,0.1643647552,0.5911531448,0.034018185,-0.0623520873,-0.0798481032,0.2068804204,-0.2147004455,-0.0293558445,-0.1153373271,0.0567737073,0.3959183693,0.0095646121,-0.285538435,0.0040001692,0.3544893265,0.2855494022,0.0208347552,0.2044816315,-0.1106403247,-0.5134525895,-0.0425503701,-0.0506711416,0.2281655669,0.0060095573,-0.0121159358,-0.27960971,-0.0588046275,-0.0362700224,0.176183328,-0.049972184,-0.0529454574,-0.1996813118,-0.4386970997,-0.0182059035,0.4425365329,0.1786849052,0.1022828147,0.0539023019,-0.1254320294,-0.4615599513,0.4586753547,-0.1486272812,-0.192113027,-0.0985347927,0.4340789616,-0.1451818645,0.3470753431,-0.124218367,-0.1238164902,0.059339527,-0.0372590385,0.0099964384,-0.1110816523,0.0633869171,0.2302578092,-0.1380207241,-0.1275823563,0.0664629713,-0.1967458427,-0.1203837693,-0.4393717349,0.2562833726,0.0678952336,-0.0392500274,-0.3404693007,0.4545376301,-0.166731596,-0.1042162031,0.1259800792,-0.1754609197,-0.0643463358,0.069110848,0.3398545384,0.0253191218,0.0868800953,-0.0635239258,0.1758936048,0.0753097981,-0.2827733755,0.0058806534,0.0976229906,0.0337164141,0.1995992064,0.1574257463,0.3863253593,-0.1026553139,-0.0637379438,0.3855548501,-0.0013763288,-0.1756060719,-0.2820312083,0.3754214048,-0.1159967333,0.2112380117,0.4110926986,-0.015505285,0.1346479505,0.0378637724,-0.2541955411,0.5481324792,-0.1878698468,-0.2755916119,0.3920185566,0.2921043038,-0.1242124364,0.1140775234,0.2238339335,-0.1109483019,0.2314634025,-0.038114056,0.2487496138,0.1422787011,0.0404577851,-0.4667989314,-0.1206577495,0.0343435891,0.4787987769,-0.1020074263,0.1082389131,0.0011627502,-0.1533785164,0.538546443,-0.0922688544,0.1676845849,0.2412602156,0.0828594863,-0.1849400103,-0.3271013796,-0.2880984545,0.0628865808,0.0237304084,-0.0346031375,-0.3504155874,0.409735322,0.0497023873,0.2264173031,-0.3599361181,0.2024379671,-0.2352502197,0.2522643507,-0.2211434394,0.0708500966,-0.4758047163,0.2940942049,0.1046742722,-0.344080478,0.109756425,0.3184894919,-0.0143800341,-0.0367730372,-0.283141017,-0.2473113388,-0.1265820265,0.0822507441,0.044978071,0.043346826,-0.0522308685,-0.0464488082,-0.0263266973,-0.2710572183,0.3402980268,0.1830480695,-0.0185470264,-0.1184859946,0.0944026113,-0.2621901333,0.1225149408,-0.1087833568,0.2234492749,0.0629801676,0.0031002876,-0.1863150448,0.154445976,0.0642118156,0.0005111973,0.006081365,0.2094147205,-0.0861231983,0.0205599107,-0.2128866911,-0.3577113152,-0.4625985324,-0.0634597018,-0.7106280327,-0.4246838987,0.3601848483,0.286570996,0.4639976919,0.3493735194,-0.073007606,-0.1968182474,-0.4219459593,0.4059285522,-0.094086282,0.325902462,-0.093716681,0.1215585545,-0.0954417735,-0.490640372,0.2663345039,-0.149153322,-0.0930133685,-0.0397672504,0.0610733218,0.2796287239,0.2296252549,0.2637280226,-0.1626534015,0.2027767152,0.0390215777,0.1617889553,-0.1238076016,0.3242393732,-0.471222192,0.1185456738,-0.0569969565,0.1671810001,-0.1209937707,-0.1547784805,-0.0034629118,0.0753991455,-0.2015694976,-0.1348519176,0.2402377278,0.0925468951,0.1910641342,0.2353417277,-0.3011476994,0.0027838186,-0.0161230434,0.1544591039,-0.2255385369,-0.2050644904,0.4881336093,-0.8047510386,-0.4586584866,-0.1668249369,0.0745781884,-0.0560602993,-0.1588740796,-0.1639579087,-0.3209580779,0.2653920352,-0.1165277213,-0.2881031334,0.0956863463,0.0898057744,-0.1982935965,-0.3410841823,0.0114867613,-0.0575587898,-0.132000789,-0.5714484453,-0.5315229297]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"Good to know about `query_table` not bringing anything into memory. I was under the impression that it did because a while back I looked at my `map` operation in pdb and it looked like it was spending forever in line 93 of formatting.py, `return pa.concat_tables(....)`, although that was before the `fast_slice` interpolation search was implemented, so it may have had more to do with the slow ChunkedArray slice implementation than anything else.\r\n\r\nIf `query_table` is I\/O free then the fix may be as simple as just adding this to line 1779 of arrow_dataset.py:\r\n```python\r\n# Only load the columns we actually need\r\nif input_columns:\r\n stack.enter_context(self.formatted_as(\r\n self._format_type,\r\n columns=input_columns,\r\n output_all_columns=False,\r\n **self._format_kwargs\r\n ))\r\n```\r\nIt's not clear to me why the `[col for col in input_columns if self._format_columns is None or col in self._format_columns]` check would be necessary\u2014 it seems like either `input_columns` should simply temporarily override the `_format_columns` within the `map` operation, or we should throw an error if there are any conflicts. Currently it doesn't look like this case is checked for at all within `map`, but maybe I'm just missing it.","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":181,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\nGood to know about `query_table` not bringing anything into memory. I was under the impression that it did because a while back I looked at my `map` operation in pdb and it looked like it was spending forever in line 93 of formatting.py, `return pa.concat_tables(....)`, although that was before the `fast_slice` interpolation search was implemented, so it may have had more to do with the slow ChunkedArray slice implementation than anything else.\r\n\r\nIf `query_table` is I\/O free then the fix may be as simple as just adding this to line 1779 of arrow_dataset.py:\r\n```python\r\n# Only load the columns we actually need\r\nif input_columns:\r\n stack.enter_context(self.formatted_as(\r\n self._format_type,\r\n columns=input_columns,\r\n output_all_columns=False,\r\n **self._format_kwargs\r\n ))\r\n```\r\nIt's not clear to me why the `[col for col in input_columns if self._format_columns is None or col in self._format_columns]` check would be necessary\u2014 it seems like either `input_columns` should simply temporarily override the `_format_columns` within the `map` operation, or we should throw an error if there are any conflicts. Currently it doesn't look like this case is checked for at all within `map`, but maybe I'm just missing it.","embeddings":[-0.1046057194,0.3224443197,-0.0158340186,-0.1899836212,0.0996766686,-0.0926468894,0.2255839109,0.5739011168,0.1345910579,-0.0051863473,-0.0915138796,0.457932353,0.0769863501,-0.2375071645,0.0071079414,0.256791234,0.0978264213,0.3570863008,0.3204873502,0.1195372492,-0.1944888085,-0.2011037469,-0.1671691835,-0.093112193,-0.1043966487,-0.0672499016,0.1529695094,-0.3571105897,-0.3311984241,-0.1571363956,0.2827627957,0.320168972,-0.2016059607,0.12886329,-0.000126842,-0.0501120053,0.0091096619,-0.0208219625,-0.1327055693,0.0189617556,0.204142198,-0.2184470594,0.0786347389,-0.2216875106,0.0527093895,-0.0613879487,0.0214829259,-0.0298988875,-0.0216247085,-0.0822018981,0.0331359431,-0.0527435802,-0.3226946592,0.211495012,0.4797375798,0.1571221501,0.0235043522,-0.2126512825,0.2892075479,-0.4368453622,-0.1270746738,0.5308563709,-0.5316395164,-0.0464950874,0.3427626789,0.049161043,0.0022185585,-0.2007324398,0.3583235145,0.0518088304,0.1605584472,-0.296064347,-0.0898218527,-0.2226341069,-0.0841382965,-0.2170465142,-0.066492714,0.0542102233,-0.456191659,-0.0182688907,-0.2613746524,-0.2506407201,-0.0523681492,0.3157967031,-0.5186735988,0.423607558,0.1705063879,0.2459537834,0.3076338768,-0.2111773044,-0.0427854173,0.0078033293,0.4094476998,0.5418298244,-0.3882897198,-0.1447073519,0.2166091651,-0.0046076714,0.2440500259,-0.2781878114,-0.1783222258,0.3794896901,0.4178121984,0.0922844037,0.3731509149,0.0572141968,-0.1314653903,0.6420698166,0.3668268919,-0.159287259,0.0897910744,0.082753852,0.0434356108,0.379861623,0.1911718845,-0.433431685,-0.1847686768,-0.1714393944,-0.0253485925,-0.0084876474,-0.1762731075,0.1174384728,0.0855485722,0.4891795516,0.2316110134,0.225787282,-0.1678627431,-0.0405790918,-0.2845214903,0.0734631494,0.0280695148,0.0345130339,-0.0512352437,0.2307008952,0.1601486057,0.1830994487,-0.1646120995,0.0152628208,-0.0722632706,0.1157179847,-0.0101245558,-0.2799978554,0.3596175015,0.5151501298,-0.1593509465,0.3464233577,0.21702604,-0.2327194214,-0.3488967121,0.3377682865,-0.132456243,-0.2263544947,0.1486907899,-0.0374891236,-0.0669678897,0.3339708149,-0.0396511108,0.6287926435,0.3945141435,-0.1595231593,-0.1377298385,-0.1415912062,-0.3631263077,-0.2324801534,0.4056436718,0.0062037702,-0.413715601,-0.266879946,-0.1262893677,0.3257215023,0.4485503733,0.4173039198,-0.0687523037,0.1465699822,0.3328170776,0.4406389892,0.4901854396,0.0109500466,-0.5700111389,0.0405080914,-0.1876828223,0.0902732089,-0.1475191861,0.3816785216,0.636492312,0.2377496064,0.3549273908,0.1779022068,-0.0855331644,0.2613480687,-0.1656879336,-0.1175130457,0.2443193942,0.1028595716,0.0680371821,-0.1726322323,-0.0305090565,0.0317019559,0.0234105289,0.0258355457,0.0950574651,0.0940840617,0.1272767931,-0.0584882908,0.0846121833,-0.2734279931,-0.2215220779,0.0759864673,0.4218752384,0.1808428466,-0.3522837162,-0.3539554179,-0.146749258,0.1585353613,0.3680203855,0.0825599059,-0.0650566891,-0.3180902302,0.1961794198,-0.1326883584,-0.226525411,-0.1426741481,-0.0524415523,0.0599178001,-0.0610491037,-0.0308259036,0.2068485469,-0.1080193147,-0.0747326314,-0.0408544727,0.1979223639,0.3113134205,0.1773786992,0.0573525243,0.2771246731,-0.2035668045,-0.2185436636,0.4961541891,0.0398581401,0.1094273478,0.1492642164,-0.2200749069,0.1322612911,-0.2009371817,-0.4113146961,0.3036414981,0.1208436266,0.6467670202,0.0456714258,0.0394136831,0.1866068095,0.0499052294,-0.0209066793,-0.346706599,0.0559415258,0.1037927717,-0.0900080651,0.185476467,0.0254997667,0.2736612856,0.2416229844,-0.0639337227,0.0869315565,0.2822878957,-0.0984455645,-0.2455402613,0.1920063645,-0.2486240715,0.0268249512,0.2137017101,0.1858732104,-0.2289258838,-0.1750144958,-0.0132448021,0.2093530297,0.2164669335,-0.185696736,-0.2287490219,0.1939637363,-0.051694721,-0.3180813193,0.1175254956,-0.0644953325,0.2998574376,-0.2273163348,-0.2105186582,-0.2503575683,-0.1601746827,0.377016753,-0.0750863627,-0.012997997,-0.2862330973,0.3427686095,0.1238389239,0.1124928966,0.2113857716,-0.2102439255,0.3423673511,0.0275171399,-0.2496317774,-0.2695373595,-0.5140575767,0.0300203655,-0.0644229054,-0.0016973639,0.4669359028,0.2170314342,0.3613241315,-0.2467996925,-0.1738606691,-0.4038601816,0.058160685,-0.1109809652,0.2604952157,0.1097898111,0.481426537,-0.3645493686,0.0615465492,-0.0480229072,-0.1158276647,-0.0675555468,0.1509262174,-0.0420305543,0.1865849793,-0.0737376511,0.2256578952,0.0664133132,-0.1036398038,-0.067420207,-0.1274869144,0.3572297692,-0.3141646981,-0.0175234694,-0.217801407,-0.1374755055,-0.0886398554,-0.1316664964,0.2906396687,0.4122536182,0.2820613086,-0.0256648976,-0.1306436658,-0.0831021518,-0.1489645243,0.5557096601,-0.3557236195,0.0762016997,-0.2389134169,0.0555393249,0.1153746843,0.1321066916,0.430894047,-0.0446397066,0.0358335897,0.0726099983,-0.3384998441,-0.2934176326,-0.0454604551,-0.1858828515,0.2047918439,0.5843628049,0.2935068607,0.7721477747,0.0733563155,0.2170889527,0.0129311299,-0.0168373156,0.0995332152,-0.067567043,-0.2521345615,-0.2696581185,-0.2066072673,-0.1512663066,0.1877740473,0.2309909612,-0.5564442277,-0.0622038282,0.1338438243,-0.0898364708,-0.2923491597,0.3094570041,0.137782678,0.1617710441,0.3674435616,0.1456876546,-0.4637149572,-0.4544992447,-0.0988754183,-0.1702692211,0.0169055816,-0.1957838386,-0.5587356091,-0.2824582458,-0.7603421211,0.2780366242,0.0803083926,0.1472099125,0.1619992852,-0.3202675581,0.246556893,-0.0714159608,0.6872801781,-0.0104190735,-0.319770366,0.1650805771,-0.0399809442,-0.5558211803,0.2617914975,-0.0753829479,0.1620527953,0.1187624708,0.5174376965,-0.5288392305,-0.1607707739,0.0649641454,0.1406831145,-0.1829618514,0.0820181295,-0.176884979,-0.4218448102,-0.0769358128,0.2829673588,0.0335437767,0.0901070163,0.4562004507,-0.0388985462,0.1134735569,-0.1031082273,0.1004560292,0.3970749974,-0.0307701211,0.1656958461,0.0527826659,0.3227798641,-0.3491329849,-0.0113402288,0.1368063539,-0.1160186827,-0.0272126906,0.0151860453,0.1074855253,0.3381061256,0.5932183266,0.0359438695,0.3888600469,-0.1304178238,0.1923792213,-0.1143197566,-0.0236530453,-0.1840966642,-0.0883662403,-0.2721232176,-0.3375445306,0.3402324915,0.2814843357,-0.2885201573,0.6071960926,0.1475522071,-0.1815362722,0.479670167,0.3666977286,1.0546474457,-0.5467887521,0.190688923,0.0070271008,0.1227302104,0.3516569734,-0.4375026524,0.2877714634,-0.1726364195,-0.1859958321,0.0071153911,-0.0953332186,-0.1211981475,0.1627475768,0.1264435202,0.1171510741,-0.0559112877,0.5475067496,0.1061089933,0.0435002148,0.2786797285,-0.4434577823,-0.1778914928,-0.0017231858,-0.011087737,-0.2943689823,0.0281584207,0.3908119202,-0.146688953,-0.372906059,-0.2168047726,-0.2689218521,-0.2742671669,0.0710576773,-0.2382919341,0.0106421402,-0.2562418282,-0.0652529746,-0.4524874091,0.1239846721,0.1430531144,0.0936328992,0.4910560846,0.2933094501,0.1694731563,-0.0947185159,0.0999560282,0.0882193372,-0.049051553,0.2174963951,-0.1155950502,-0.3844929338,-0.3565834165,0.1957243532,0.5410517454,0.0535354279,-0.0135002751,-0.0737215281,0.2004080415,-0.204274714,-0.0176929161,-0.104129523,0.0264188722,0.4646103382,-0.008595502,-0.3243326247,-0.0151910251,0.3858380914,0.2920014858,0.0562836826,0.1858604848,-0.1234939471,-0.5478109121,-0.0282436348,-0.0300151538,0.1806606948,0.0569259264,0.0046833125,-0.2886976302,-0.1033007801,-0.0024997501,0.1498054117,-0.012430747,-0.1362506151,-0.2823067009,-0.356369555,-0.0686285123,0.4347744584,0.2062357962,0.1038126573,0.0580777861,-0.1695519239,-0.4256277978,0.5064364672,-0.1533578783,-0.2173798233,-0.1600067466,0.4036403298,-0.2167794257,0.2944271266,-0.0808223188,-0.1172749847,0.0371399671,0.0098441988,0.0286148973,-0.1156682894,0.1221238524,0.2382363826,-0.1640378088,-0.0912226886,0.0329155736,-0.1731162965,-0.1281941533,-0.4739854634,0.2547739744,0.0811576992,-0.0105227027,-0.3706906438,0.4491297901,-0.1695689708,-0.0471708179,0.1714463681,-0.1817426085,-0.0698383451,0.0787623078,0.3315686285,0.0072500682,0.0656058416,-0.0210312176,0.2076462507,0.0554323979,-0.3246943355,-0.0300754681,0.0958479568,-0.0197713301,0.2034463584,0.1249017641,0.4191643,-0.0637691617,-0.0818422511,0.3650665581,-0.0094159087,-0.1428363323,-0.3040504754,0.4143345058,-0.1500126421,0.1602364182,0.4338808358,0.046345558,0.1727833301,0.0426422544,-0.2313518226,0.5329794288,-0.1897871047,-0.2840983272,0.371732831,0.2719367743,-0.1213118955,0.1473492682,0.1920560598,-0.0682762191,0.2579434514,-0.0198372342,0.297333777,0.0691000521,0.0615576059,-0.4594765902,-0.1466951519,0.0268919002,0.4805705547,-0.1504916549,0.1339513063,-0.0326375216,-0.0540251844,0.4316004217,-0.1216271669,0.1498163939,0.256814748,0.0415894464,-0.17480883,-0.2070778757,-0.2911296189,0.0330205187,0.0350493081,0.0190116446,-0.3109900951,0.4285987318,-0.0039090463,0.1566877961,-0.3297465146,0.1822060347,-0.2607098222,0.2837090492,-0.2353382558,0.0545073859,-0.4959184229,0.3333556354,0.0746462718,-0.3306196034,0.1200232655,0.3252446055,-0.0473216176,-0.079077594,-0.259773165,-0.2576856315,-0.1574633867,0.1137760356,0.024979122,0.0598857887,-0.0813167244,-0.0540571772,-0.0252097473,-0.2818568647,0.3496991396,0.2898426652,-0.003158801,-0.1131201461,0.0707380325,-0.3256591558,0.1132195815,-0.1082583293,0.2076455802,0.096694842,-0.0187659897,-0.2357936352,0.1526088715,0.0616221279,0.0044912044,0.033621002,0.2721969783,-0.0687187389,0.0013840544,-0.2263070047,-0.3218940496,-0.4398283958,-0.0292007532,-0.6811189651,-0.356535852,0.3824746609,0.2835347652,0.4024675786,0.2974587977,-0.0878105164,-0.2426253855,-0.3541280925,0.2983641028,-0.1090958863,0.3451604843,-0.1417726576,0.1289978474,-0.0833110139,-0.4821410775,0.2896772325,-0.1179274395,-0.088659361,-0.0606942214,0.0506663062,0.2764613628,0.3113325536,0.3039519489,-0.1210212782,0.1925037652,0.0212958548,0.1368174106,-0.1140910238,0.3005502522,-0.4723969102,0.0878935903,-0.0672235489,0.1004680321,-0.1225905418,-0.1978385448,-0.0027770794,0.1293994933,-0.161832124,-0.1365453303,0.2370935082,0.0221230853,0.224718675,0.2171257287,-0.2882349193,-0.036039196,-0.0173349939,0.1500177085,-0.2527909577,-0.2092111111,0.5330389738,-0.7193199396,-0.4463279247,-0.1595333964,0.091413103,-0.0455705263,-0.1843872964,-0.2144661397,-0.3904004991,0.2421592772,-0.0916078985,-0.3103508353,0.1091041863,0.0371695943,-0.1971423775,-0.3341417015,-0.0097178845,-0.1004761532,-0.1067952886,-0.5324395299,-0.4777884483]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"`query_table` simply slices\/concatenates parts of the table. The actual data inside the table is not brought in memory.\r\nAlso I'm more in favor of declaring `input_dataset = self.with_format(...)` since `formatted_as` may update the dataset fingerprint of `self`, which is not expected when someone runs `map`.\r\n\r\n> It's not clear to me why the [col for col in input_columns if self._format_columns is None or col in self._format_columns] check would be necessary\u2014 it seems like either input_columns should simply temporarily override the _format_columns within the map operation, or we should throw an error if there are any conflicts. Currently it doesn't look like this case is checked for at all within map, but maybe I'm just missing it.\r\n\r\nActually yes we can just use input_columns. And we do need to add a check to make sure there are not conflicts or this could lead to confusing errors.","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":145,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\n`query_table` simply slices\/concatenates parts of the table. The actual data inside the table is not brought in memory.\r\nAlso I'm more in favor of declaring `input_dataset = self.with_format(...)` since `formatted_as` may update the dataset fingerprint of `self`, which is not expected when someone runs `map`.\r\n\r\n> It's not clear to me why the [col for col in input_columns if self._format_columns is None or col in self._format_columns] check would be necessary\u2014 it seems like either input_columns should simply temporarily override the _format_columns within the map operation, or we should throw an error if there are any conflicts. Currently it doesn't look like this case is checked for at all within map, but maybe I'm just missing it.\r\n\r\nActually yes we can just use input_columns. And we do need to add a check to make sure there are not conflicts or this could lead to confusing errors.","embeddings":[-0.0905099362,0.2783811986,-0.0073054857,-0.1781735718,0.1408671141,-0.0961291492,0.3184326887,0.5706532598,0.1868227869,-0.0280762687,-0.0586419851,0.4846483767,0.0940666646,-0.2658150494,-0.0045916475,0.2481260002,0.100163646,0.3244252205,0.2597131133,0.0609557107,-0.1716758907,-0.2206242532,-0.2154829949,-0.0906187817,-0.0869437084,-0.0568239987,0.1601068228,-0.375895083,-0.3585896492,-0.1589115411,0.2664770782,0.3161299825,-0.2051470429,0.1349192262,-0.0001272232,-0.0201998558,0.0467341878,-0.037327908,-0.1995202601,-0.0664869398,0.1474398226,-0.1848529577,0.1072840989,-0.2459902614,0.059843298,-0.1259113997,0.0318901353,-0.0521755069,0.0090082446,-0.1114037856,0.0332686491,-0.0471821167,-0.3189130723,0.2509446144,0.4844817817,0.2142710388,-0.0029356712,-0.1754944623,0.2836517394,-0.3993281424,-0.1439495385,0.5009637475,-0.4719651043,-0.0450852662,0.3814987838,0.0277342685,0.0133877341,-0.197485283,0.3342888355,0.0723303556,0.2053318173,-0.3297653198,-0.0984181762,-0.2013444602,-0.0895540863,-0.2471139282,-0.0652233213,0.0728401765,-0.419649601,-0.006134755,-0.2666519284,-0.2152370214,-0.0831532106,0.3996171355,-0.4749021232,0.4343481064,0.1701668203,0.3172897398,0.1965943128,-0.1460077763,0.0056473636,-0.0809224248,0.4420045912,0.5432040691,-0.4257611334,-0.139276132,0.1378888637,0.019359678,0.2514255643,-0.2275911272,-0.1442237943,0.3652479053,0.3820867538,0.1383820772,0.3823768198,0.0552436523,-0.0678869411,0.5654723048,0.340166539,-0.1287790686,0.099251166,0.0866193026,0.0072611244,0.3771412671,0.2361530215,-0.397176981,-0.1501549929,-0.1441699862,-0.0033347334,-0.0359203704,-0.2364974469,0.0379734635,0.1146311089,0.4467284083,0.217115432,0.2219400555,-0.2395924926,0.0069658682,-0.2428150475,0.0167142022,0.0263271052,0.0361496843,-0.0885544717,0.1784718335,0.1263100952,0.1700144112,-0.1283931285,0.0416778587,-0.0802163482,0.1100666523,0.017849965,-0.2410555631,0.304330349,0.5060296655,-0.1270371079,0.4151405692,0.2337826341,-0.2551225126,-0.354177922,0.3156724274,-0.2056185156,-0.2452472001,0.1567534506,-0.0432886854,-0.0533226468,0.3144853413,-0.0819184855,0.5806527138,0.3763392568,-0.1514410079,-0.1035860702,-0.1623917818,-0.3125413656,-0.2653497756,0.3400167227,-0.0321212113,-0.4576417804,-0.2199665606,-0.1748510599,0.2500417233,0.4678746164,0.4039694667,-0.1157769188,0.1306847185,0.3230548203,0.4230688512,0.5786601305,-0.0044513447,-0.6051669121,0.0456339903,-0.1910752654,0.1302179098,-0.1258923858,0.3633518517,0.6630903482,0.1699456722,0.3502970338,0.208381623,-0.0832398906,0.2332906574,-0.1320047528,-0.0992477015,0.2240942717,0.0698581189,-0.0314692035,-0.1580111235,0.0432081893,-0.0146944486,0.012032832,-0.0307444688,0.1495976001,0.0352642536,0.1225036234,-0.0793454126,0.0968361646,-0.1701586097,-0.3151438236,0.0833316073,0.4757758677,0.1626962572,-0.3184887767,-0.4162273705,-0.1717945784,0.1589773893,0.3474315107,0.0717163458,-0.0883723497,-0.3062795699,0.1927574575,-0.1530602127,-0.2230236679,-0.1582785845,-0.0786756203,0.0918850824,-0.0018242518,-0.0461274311,0.2228624672,-0.0545991175,-0.0833940431,-0.0686212406,0.2227339447,0.2268548459,0.169095695,0.1006862223,0.3023964465,-0.2007376552,-0.3022830486,0.4918423891,0.0368731432,0.1524703801,0.1140558422,-0.2299974263,0.1563223302,-0.2239600271,-0.4133352041,0.2558535933,0.1995146126,0.6288716793,0.0527716987,0.0841597617,0.1763401031,0.0580587834,-0.0930086747,-0.3284853697,-0.0237520132,0.0519340262,-0.0684548616,0.1705927998,-0.0424999483,0.2811124325,0.3036043346,-0.0635403022,0.1217754781,0.2883030176,-0.1449168772,-0.220800966,0.2054200619,-0.1809400618,0.0339856185,0.1745779663,0.1639837474,-0.2133012861,-0.1876698881,-0.0083754892,0.199229762,0.2124735266,-0.1250440478,-0.2343129963,0.2074477971,-0.0626442581,-0.3354994357,0.0732227564,0.0212960802,0.2981348336,-0.2283177823,-0.1893731803,-0.2869470417,-0.1307526976,0.3210741282,-0.0955272242,0.0365262292,-0.2719177008,0.3166380525,0.1489036679,0.1104646474,0.2225836664,-0.0962288603,0.3324805498,0.0444538556,-0.2473369241,-0.2861946821,-0.4948458374,0.0235192161,-0.0699812695,0.0105353817,0.3992695808,0.2383036762,0.3589071631,-0.2722924352,-0.2329825312,-0.4402537048,0.0664399862,-0.1753516942,0.2203930467,0.1014787927,0.5082466602,-0.3530637026,0.0723143592,0.0008109203,-0.0886769667,-0.1040697321,0.1476087868,-0.0548230857,0.1584558636,-0.1324396282,0.2103565186,0.0645837039,-0.1025527269,-0.088569425,-0.1085603088,0.3564195335,-0.3576383591,-0.0122366296,-0.2859368622,-0.115928553,-0.0990879908,-0.1391461939,0.211953029,0.443318665,0.2500690818,-0.0770412982,-0.1258710325,-0.0833321437,-0.1012141854,0.6114694476,-0.364541024,-0.0142550459,-0.2090164423,0.0716235638,0.1647153497,0.1409079283,0.4514619112,-0.0423866212,0.0590075627,0.0534386039,-0.3214302361,-0.2903229892,-0.0478871129,-0.2004417479,0.2138630152,0.6448985338,0.309388876,0.8403735161,0.1130487919,0.1725102067,0.0337036103,-0.0510696247,0.0815417692,-0.0633322522,-0.2036994398,-0.2114545405,-0.2350517064,-0.0964301303,0.183326453,0.1717270017,-0.5537528992,-0.0195636004,0.1473821253,-0.1255721599,-0.2622869611,0.304364413,0.1472882032,0.168082267,0.3741904497,0.1579210609,-0.4175935388,-0.4230319858,-0.0638339892,-0.1295709312,0.0286786966,-0.1340877861,-0.5923666358,-0.2360268086,-0.7482565045,0.3018723428,0.0829417109,0.1522829682,0.1874333173,-0.2930601537,0.2830201089,-0.0755596608,0.6906268001,-0.045545578,-0.2970350087,0.1469210982,-0.0236814693,-0.5939133763,0.2209278196,-0.0780473799,0.1537737101,0.1613260508,0.5281014442,-0.5880943537,-0.1972346902,0.0655304939,0.1682630777,-0.1555670202,0.0721529424,-0.1180700436,-0.3797647059,-0.1486627758,0.224503994,-0.0023846847,0.0908385515,0.4537904561,-0.0277318675,0.110886693,-0.0929073021,0.0876652449,0.4273621738,-0.0526528731,0.1802545488,0.0472498015,0.2657177746,-0.2962566316,-0.0124261081,0.2026898414,-0.0458398983,-0.1055439636,0.0209687483,0.1237254664,0.3193656206,0.5515719056,0.0181521103,0.3794238567,-0.1454445124,0.1872474104,-0.1112791672,0.0134002697,-0.1301664859,-0.0295656696,-0.3471386731,-0.3774890006,0.3290825486,0.2539483905,-0.2744832635,0.6376811862,0.1579562575,-0.1558041722,0.4983794987,0.3384491205,1.0316679478,-0.5237039328,0.1810263246,0.0649036169,0.1576909274,0.3389491737,-0.4578652978,0.2421647608,-0.1825539619,-0.1255982518,-0.0275493562,-0.1177877411,-0.0809304044,0.2127182335,0.1301225275,0.1471439004,-0.0913341194,0.5559559464,0.1121522561,0.1031305566,0.3103384376,-0.4191634655,-0.0604601726,-0.0199587252,-0.0042771376,-0.2660847306,0.0550704487,0.3650122285,-0.1820290834,-0.4132028222,-0.179347232,-0.2829311192,-0.2284823656,0.0317845307,-0.1688814908,-0.034445677,-0.1854680181,-0.0827976987,-0.4226719439,0.1179892719,0.165964365,0.1124050394,0.5068206191,0.2643394768,0.1579757482,-0.1196403205,0.1131157652,0.0955267176,-0.021248037,0.2237527221,-0.1186210066,-0.4042244554,-0.3700035214,0.1789553463,0.5880467892,0.0173713583,-0.0404587463,-0.057955455,0.1724025905,-0.2196785063,-0.0317841731,-0.1428275406,0.0773450211,0.4151414037,-0.0195986312,-0.3192186952,-0.0349478237,0.3392710984,0.2927271724,0.0734843537,0.1987022907,-0.1479400247,-0.530084312,-0.0362331085,-0.0716006681,0.2113047242,-0.0061027091,-0.0332694314,-0.3178265989,-0.0642098039,-0.0575801358,0.1637305915,-0.053069599,-0.038555488,-0.2860912681,-0.4268892109,-0.0263917986,0.3972118795,0.1831769198,0.0775503442,0.1022703573,-0.1148219332,-0.4360665381,0.5048314929,-0.1520837992,-0.2250878513,-0.0594025478,0.4260328412,-0.1437666267,0.3060754538,-0.0553408377,-0.0834064409,0.0524670444,-0.0266078394,-0.0302524157,-0.1135834455,0.0803570226,0.2378629297,-0.1588044167,-0.0903760195,0.0642995387,-0.1745073199,-0.1558818817,-0.4696463645,0.2788101137,0.0590898544,-0.0569162071,-0.3633279204,0.3943197429,-0.151709035,-0.0716864094,0.1434642226,-0.1927836835,-0.045546405,0.0723629966,0.2985589802,0.0111879613,0.052919209,-0.0957496837,0.1875260025,0.0862702057,-0.2638135254,0.0045443019,0.0805196837,0.0404200256,0.144331634,0.1404486448,0.3716208935,-0.0964741185,-0.1018291041,0.3553173244,-0.0119078364,-0.1676963717,-0.2848278582,0.4189099371,-0.0634245127,0.2279410064,0.41563043,0.0333216265,0.1718685031,0.0038923374,-0.2800189853,0.5418035984,-0.1752542108,-0.2971153855,0.3870226443,0.2368099391,-0.1032685786,0.1635177881,0.1996377259,-0.097458072,0.2522516549,-0.0348037742,0.2527009249,0.1692421138,0.0488701053,-0.444273442,-0.1520387381,0.1098210514,0.4503726661,-0.1263173521,0.109526448,0.0359788164,-0.1239174008,0.5140132904,-0.096478723,0.1751797497,0.2371955216,0.0678901225,-0.153294012,-0.3148585558,-0.253097266,0.0662415475,0.0134201283,-0.0530078635,-0.333481282,0.4227046669,0.0035088598,0.160949856,-0.3389717042,0.1463305503,-0.2596039474,0.2895329297,-0.207603097,0.0851739794,-0.5017508268,0.29166013,0.085510537,-0.3297107816,0.1519297212,0.3821591735,0.0403681397,-0.0721257329,-0.2558274567,-0.2519416511,-0.1346752495,0.0848696604,0.0511255153,0.0589232817,-0.0254165139,-0.0539004095,-0.0340353362,-0.2723685801,0.3616445661,0.1971458942,-0.0291687381,-0.1404814422,0.0581687577,-0.310474962,0.120283179,-0.0976939127,0.1871075928,0.0659422204,0.0450585969,-0.2126419246,0.1698902398,-0.0271920357,-0.0016710379,-0.007414978,0.2734506428,-0.0535080843,-0.0002118102,-0.2139617503,-0.3434696794,-0.4677125812,0.0057679876,-0.7461003661,-0.3828416169,0.3602566719,0.3030540347,0.4502289891,0.3592269719,-0.0408315919,-0.2498151213,-0.3973614275,0.3535872698,-0.1052342728,0.3524170816,-0.1102694198,0.1047954857,-0.0784879401,-0.5093358755,0.3277420998,-0.141918391,-0.0881255344,-0.0596211143,0.0847385824,0.288751334,0.2048827708,0.3212364316,-0.1280442923,0.1876013577,0.0583338216,0.1533363461,-0.1223327518,0.3114006519,-0.4641097188,0.1369898021,-0.0773245469,0.1348330677,-0.0685960799,-0.1925361902,-0.0155427856,0.1387905329,-0.1513771862,-0.1522450447,0.265410006,0.1188470647,0.1583606005,0.2228540182,-0.2743425369,-0.0407455787,0.0157607328,0.1023176759,-0.2132022828,-0.2157047093,0.5233751535,-0.7930595279,-0.4431898594,-0.162582159,0.0579366572,-0.052723527,-0.17479904,-0.1622184813,-0.3733665049,0.2869019508,-0.1278045624,-0.3367220163,0.1035346463,0.0237301495,-0.1718414724,-0.3246551156,-0.0177163817,-0.1019672155,-0.1382971704,-0.5111670494,-0.5413414836]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2193","title":"Filtering\/mapping on one column is very slow","comments":"That sounds good to me! I just submitted a PR (#2246) implementing your approach. I also changed how `_query_table` handles Iterable keys since it still seemed like `pa.concat_tables` was taking a long time to create the table for each batch. Now my whole `map()` operation takes 1 min 46 seconds where it used to take somewhere on the order of 10 minutes.","body":"I'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.","comment_length":62,"text":"Filtering\/mapping on one column is very slow\nI'm currently using the `wikipedia` dataset\u2014 I'm tokenizing the articles with the `tokenizers` library using `map()` and also adding a new `num_tokens` column to the dataset as part of that map operation.\r\n\r\nI want to be able to _filter_ the dataset based on this `num_tokens` column, but even when I specify `input_columns=['num_tokens']`, it seems that the entirety of each row is loaded into memory, which makes the operation take much longer than it should. Indeed, `filter` currently just calls `map`, and I found that in `_map_single` on lines 1690-1704 of `arrow_dataset.py`, the method is just grabbing slices of _all the rows_ of the dataset and then passing only the specified columns to the map function. It seems that, when the user passes a value for `input_columns`, the `map` function should create a temporary pyarrow table by selecting just those columns, and then get slices from that table. Or something like that\u2014 I'm not very familiar with the pyarrow API.\r\n\r\nI know that in the meantime I can sort of get around this by simply only returning the rows that match my filter criterion from the tokenizing function I pass to `map()`, but I actually _also_ want to map on just the `num_tokens` column in order to compute batches with a roughly uniform number of tokens per batch. I would also ideally like to be able to change my minimum and maximum article lengths without having to re-tokenize the entire dataset.\r\n\r\nPS: This is definitely not a \"dataset request.\" I'm realizing that I don't actually know how to remove labels from my own issues on other people's repos, if that is even possible.\nThat sounds good to me! I just submitted a PR (#2246) implementing your approach. I also changed how `_query_table` handles Iterable keys since it still seemed like `pa.concat_tables` was taking a long time to create the table for each batch. Now my whole `map()` operation takes 1 min 46 seconds where it used to take somewhere on the order of 10 minutes.","embeddings":[-0.1737178862,0.330304265,-0.0151500572,-0.2352599353,0.0967264846,-0.1373569518,0.2451967895,0.6020462513,0.121574536,-0.0561254472,-0.0701213703,0.5131320357,0.0704808384,-0.2048014849,-0.0179261304,0.2067434043,0.0910430774,0.3320176601,0.3077172935,0.1418736279,-0.12450964,-0.1790834367,-0.1606733501,-0.0990440995,-0.0650921762,-0.0865211338,0.1429021358,-0.4043100178,-0.3323493898,-0.2287281901,0.3221342862,0.3873366117,-0.1767820269,0.1415216476,-0.0001248356,-0.0549316108,0.014044716,0.0137422355,-0.1824400872,-0.0295471847,0.1309021264,-0.2167410254,0.0755152702,-0.1730611324,0.0476248302,-0.1083694547,0.0306255519,-0.0478271581,0.0215574764,-0.0948138908,0.0412021503,-0.0476073287,-0.2681420147,0.2097008824,0.4645839036,0.1935361326,-0.0502106026,-0.1640952975,0.3533452749,-0.4462712109,-0.1411142349,0.5246965885,-0.4593464434,-0.0403555036,0.3705934584,0.0064206352,0.0693932325,-0.2104827464,0.2937604785,0.1523418128,0.1825716347,-0.3065946698,-0.1353577971,-0.2616470158,-0.0473709777,-0.2595102489,-0.060145285,0.045671083,-0.4410809577,-0.0086555965,-0.2520543635,-0.2057042569,-0.0832181573,0.3169912994,-0.4233549237,0.4046776891,0.2104633749,0.2437730879,0.2672819197,-0.2661036551,0.0122516463,0.0007127369,0.3663119674,0.5617972016,-0.4556772411,-0.1941936314,0.1578668803,-0.0031945738,0.3044346571,-0.243067503,-0.2232784033,0.4180018902,0.3920199871,0.0452779494,0.3689549565,0.0754629374,-0.0427866392,0.5948009491,0.3502455354,-0.09752886,0.0398624092,0.0979136527,0.0036888297,0.2870084643,0.2049080431,-0.3669087291,-0.1987364441,-0.1619669795,-0.0005876391,-0.0147799784,-0.169925496,0.022249965,0.1173086539,0.455984652,0.2418352216,0.2840332687,-0.2097379118,0.0206912067,-0.2422883809,0.0668707862,0.065980494,-0.0030266629,-0.1116452366,0.2152013183,0.1996005774,0.1780086756,-0.2033906281,0.0543159805,-0.0201782323,0.120637022,-0.0231195688,-0.2095009536,0.2548909187,0.4586892128,-0.1261474192,0.3494060338,0.2085407227,-0.2365360111,-0.3625700772,0.2813225985,-0.1292225868,-0.2616797686,0.1350033134,-0.0050355126,-0.0949928686,0.2969782948,-0.0937212929,0.5968532562,0.4027432203,-0.105794318,-0.1509811282,-0.1515808553,-0.303047806,-0.1957915276,0.3636776209,-0.0224481896,-0.4293302298,-0.1934003085,-0.130592674,0.3211943209,0.4375316501,0.4082219005,-0.1090678722,0.1430213749,0.369586736,0.3730922341,0.5056322217,0.0012932747,-0.5878244042,0.0985966027,-0.1475342661,0.077212885,-0.0818872824,0.3269074261,0.6758551002,0.1778357327,0.3605339825,0.1743607372,-0.0981857553,0.269028157,-0.1825914681,-0.1666394174,0.2102351338,0.1413112283,-0.0145347081,-0.1957734823,-0.0167291071,-0.0238866657,0.0183546543,-0.0340095572,0.1450583637,0.0538885221,0.1372446567,-0.0590700805,0.1375439316,-0.2296515107,-0.2415961921,0.0802753344,0.3720583916,0.2029228061,-0.394905448,-0.4120774269,-0.1733501405,0.1774894595,0.340257585,0.0905264392,-0.046565529,-0.3512777388,0.2362405062,-0.1481235325,-0.2082928121,-0.1503553391,-0.0387831889,0.0938901827,-0.0214913618,-0.0832755193,0.1942667216,-0.0556214489,-0.011073133,-0.0742620826,0.2472742945,0.25627473,0.158920303,0.0734024346,0.252461493,-0.1650274694,-0.1919422597,0.4814751446,0.0881001502,0.1320402473,0.1218481883,-0.2094859928,0.0983595178,-0.1908242404,-0.4711273611,0.3587895632,0.1109310985,0.683293283,0.0562724248,0.077021949,0.2420536578,0.0231919345,-0.0858704448,-0.3186715543,0.053463567,0.0744462162,-0.0931241438,0.134412393,0.0041623744,0.3699188828,0.2097867131,-0.048275914,0.0577067174,0.2880360782,-0.11212942,-0.2076292783,0.1912911087,-0.134046793,0.0256762952,0.1986670345,0.12500301,-0.2327360362,-0.1954657435,-0.0191463344,0.2358425856,0.1909601688,-0.1508255452,-0.213441506,0.2186042517,-0.0437209606,-0.3614638448,0.0391206034,-0.0381608158,0.3463041484,-0.2217038721,-0.1943615675,-0.2746430635,-0.1993744671,0.3418702781,-0.0573148467,0.0268589407,-0.2546793818,0.3636475205,0.147611931,0.1041288599,0.2501795888,-0.1866698563,0.2861469984,0.0559879802,-0.2362312078,-0.3087218106,-0.5034463406,0.0297077745,-0.0576346852,0.0336736441,0.3986271918,0.2478462309,0.3540845811,-0.2575275004,-0.235790506,-0.425932169,0.0515650474,-0.1574469656,0.2065596879,0.1195913553,0.5191636086,-0.3392128348,-0.0219038222,0.014687025,-0.1095936075,-0.0687238351,0.1560565531,-0.1134048998,0.1369801909,-0.0930724591,0.1554458141,0.0553491674,-0.1131982654,-0.0398740992,-0.0548620597,0.3683741987,-0.3320026994,0.0155543955,-0.3100392818,-0.1693540215,-0.1076868698,-0.1219562516,0.246271342,0.4157121778,0.2412174493,-0.0508850962,-0.1137241051,-0.0976725444,-0.1423163414,0.5603023767,-0.288372308,0.0036245259,-0.2501584888,0.0819394365,0.1499505639,0.1256531924,0.41136989,0.0233376138,0.0292191356,0.0727467611,-0.2747860849,-0.265996784,-0.1089745685,-0.1483224332,0.2204151005,0.5995185375,0.3056702018,0.8259323835,0.1461455673,0.2041177303,0.0113568408,-0.0393553376,0.0351681225,-0.0617045499,-0.2360727042,-0.2448732853,-0.2205196619,-0.0732967407,0.1571108252,0.224080354,-0.5832890272,-0.0232148953,0.1542590559,-0.0750843808,-0.3146774769,0.3322229981,0.0594746172,0.1893119961,0.3733280897,0.1218842492,-0.4728974998,-0.4238558412,-0.0668869391,-0.1457630247,0.0048059467,-0.1921513826,-0.6434128881,-0.2553215623,-0.8011319041,0.288248986,0.085426487,0.1900915354,0.2013129592,-0.2591617405,0.2519071698,-0.1105630398,0.6910763979,-0.0863247961,-0.2742422223,0.2116947174,0.0120939016,-0.5982701182,0.2258569449,-0.049823001,0.179419443,0.1453026086,0.5133736134,-0.5888466239,-0.1962420344,0.012588325,0.1193638891,-0.1398930699,0.0663581192,-0.0807125792,-0.3828619123,-0.1426973939,0.2985942662,0.0233363286,0.1394144148,0.4171300232,-0.0118875559,0.1098664626,-0.0434922352,0.0566923209,0.4345966876,-0.0511907004,0.1705740392,0.1386630535,0.3155053556,-0.340742439,0.0302705914,0.0975705832,-0.0715781152,-0.1321180761,-0.0456639193,0.1029339805,0.2442724556,0.5856772661,0.0237268526,0.37126863,-0.1306110173,0.2108144313,-0.0397328436,-0.0472761728,-0.172222361,-0.0398457274,-0.3455403745,-0.4100881219,0.2944976389,0.2769503295,-0.2877586782,0.5986565351,0.108362183,-0.1681361049,0.5244889855,0.3236609995,1.0571986437,-0.5189003348,0.1504935175,0.0092878658,0.1536120921,0.3635289669,-0.4427518249,0.254961282,-0.2084512711,-0.1057655364,-0.0030289332,-0.075724557,-0.1230415031,0.2054664344,0.0988989621,0.1527143419,-0.0495307334,0.4358771443,0.0650617033,0.1470280439,0.2220462859,-0.5009416342,-0.1556420326,-0.0032372046,-0.0110685276,-0.2385223806,0.0460571796,0.3009208739,-0.1505009234,-0.3813907504,-0.2751620114,-0.2833554149,-0.2573634982,0.0087856594,-0.1616133749,-0.0412304327,-0.1846195757,-0.1071350574,-0.3667344153,0.1103556976,0.1170675457,0.1808414906,0.4582231045,0.2670478225,0.1729890108,-0.1589831263,0.0368415862,0.0842364877,-0.04478487,0.1267386973,-0.1029131487,-0.4008793533,-0.3160424829,0.1403674632,0.5727041364,0.033795815,0.0078351861,-0.0225751493,0.1333252192,-0.2516576946,-0.0018535985,-0.1188054681,0.0824054033,0.4565769434,-0.027230002,-0.2644684017,0.0116972476,0.3394796252,0.2911001146,0.059803538,0.1668428183,-0.0950083509,-0.5276139379,-0.0585710257,-0.0580889359,0.1581247002,-0.0314791575,-0.0513166115,-0.2905528843,-0.1098838001,0.0539725833,0.1859714091,-0.026362706,-0.0672652051,-0.1996268332,-0.395063132,-0.0409961827,0.3940740526,0.1895653903,0.0728791133,0.0619028322,-0.105757609,-0.4257470071,0.4849445522,-0.1800254136,-0.1897769421,-0.1056913584,0.3836824,-0.1667005569,0.2562185228,-0.0814872608,-0.12832807,0.0692461729,-0.005573981,-0.0075382902,-0.1338732243,0.0889949799,0.1997961104,-0.1707245111,-0.1075979397,0.0280460082,-0.1739907861,-0.1748982519,-0.4453567266,0.1998779327,0.0480562523,-0.0530804992,-0.4007871151,0.3999560475,-0.171092391,-0.0482889302,0.1756545305,-0.1579749137,-0.0703105032,0.1083418876,0.3241197169,-0.0289290342,0.0329184979,-0.0359421745,0.1754429042,0.0681950822,-0.2027570456,0.0485401042,0.1107329726,-0.0451443978,0.2096825838,0.1707156003,0.3930060863,-0.0566607602,-0.0269629657,0.371415168,0.0193829518,-0.1883515567,-0.3401808143,0.4027529955,-0.1269009113,0.1954554617,0.3642883003,0.0686026365,0.135190621,0.0908435732,-0.2223103195,0.6175413728,-0.260884434,-0.288140595,0.3428646028,0.2779901624,-0.12723203,0.1567322016,0.2082572281,-0.0815167129,0.2100405246,-0.0578179806,0.25132671,0.1970292032,0.064988032,-0.4335648119,-0.0970000401,0.0148964841,0.4516619742,-0.0940737873,0.0669019297,0.028723238,-0.0680379644,0.5121017098,-0.1372922212,0.2153731883,0.2418622226,0.0878523365,-0.217823267,-0.3109594584,-0.2795116305,0.0975806341,-0.0165839586,-0.0641463473,-0.2964713871,0.4300550222,0.0347177684,0.2284208983,-0.43785128,0.1886383444,-0.2358880192,0.3433355987,-0.203097403,0.0802779719,-0.4757507145,0.2709186673,0.0657155365,-0.2793754339,0.1804791838,0.3388805687,-0.0215541236,-0.0323866233,-0.2759718299,-0.2054631561,-0.1266843975,0.0848994553,0.0418897793,0.0050509619,-0.0361385383,-0.0458714217,-0.0400178954,-0.2692870498,0.3161686957,0.2101806998,0.0587210581,-0.1758021563,0.1078010872,-0.3133994043,0.133627519,-0.0684288964,0.2310837209,0.0788872018,0.0219287351,-0.1923197806,0.1817888767,-0.0308100041,0.0192396119,0.0393720344,0.2859229147,-0.0379655659,0.0100691868,-0.2129465044,-0.3246408701,-0.4845193326,-0.0508021936,-0.7596924901,-0.3530135155,0.3888367414,0.3353849649,0.4507119954,0.370835036,-0.040823888,-0.2208570838,-0.3869852722,0.4041162729,-0.1300005466,0.3830999732,-0.0210199896,0.1431985199,-0.0822181553,-0.5314817429,0.341771245,-0.1066255569,-0.0619855486,-0.0291415397,0.0667937174,0.3297044337,0.210501045,0.2903521061,-0.0934270546,0.1745530963,0.0669198334,0.1510739475,-0.1747182608,0.2675829828,-0.4393056929,0.0974083021,-0.0462264195,0.0972075462,-0.1002977639,-0.1852229983,-0.0050044591,0.0859042332,-0.1603618115,-0.1441853344,0.1892619133,0.0854240134,0.2521850467,0.2280613482,-0.37901178,0.006634424,-0.0243187658,0.1235873476,-0.2594006062,-0.2491070777,0.4583288431,-0.7934114337,-0.4857645035,-0.1487043947,0.0693741888,-0.03476182,-0.1215015724,-0.1182699725,-0.3490398526,0.2194595337,-0.1350418329,-0.3207785487,0.0985648707,0.0716264024,-0.215152666,-0.3376730084,-0.0159622151,-0.0440616906,-0.1244518384,-0.5616627932,-0.5143708587]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2190","title":"News_commentary Dataset Translation Pairs are of Incorrect Language Specified Pairs","comments":"Hi @anassalamah,\r\n\r\nCould you please try with this:\r\n```python\r\ntrain_ds = load_dataset(\"news_commentary\", lang1=\"ar\", lang2=\"en\", split='train[:98%]')\r\nval_ds = load_dataset(\"news_commentary\", lang1=\"ar\", lang2=\"en\", split='train[98%:]')\r\n```","body":"I used load_dataset to load the news_commentary dataset for \"ar-en\" translation pairs but found translations from Arabic to Hindi. \r\n\r\n```\r\ntrain_ds = load_dataset(\"news_commentary\", \"ar-en\", split='train[:98%]')\r\nval_ds = load_dataset(\"news_commentary\", \"ar-en\", split='train[98%:]')\r\n\r\n# filtering out examples that are not ar-en translations but ar-hi\r\nval_ds = val_ds.filter(lambda example, indice: indice not in chain(range(1312,1327) ,range(1384,1399), range(1030,1042)), with_indices=True)\r\n```\r\n\r\n* I'm fairly new to using datasets so I might be doing something wrong","comment_length":22,"text":"News_commentary Dataset Translation Pairs are of Incorrect Language Specified Pairs\nI used load_dataset to load the news_commentary dataset for \"ar-en\" translation pairs but found translations from Arabic to Hindi. \r\n\r\n```\r\ntrain_ds = load_dataset(\"news_commentary\", \"ar-en\", split='train[:98%]')\r\nval_ds = load_dataset(\"news_commentary\", \"ar-en\", split='train[98%:]')\r\n\r\n# filtering out examples that are not ar-en translations but ar-hi\r\nval_ds = val_ds.filter(lambda example, indice: indice not in chain(range(1312,1327) ,range(1384,1399), range(1030,1042)), with_indices=True)\r\n```\r\n\r\n* I'm fairly new to using datasets so I might be doing something wrong\nHi @anassalamah,\r\n\r\nCould you please try with this:\r\n```python\r\ntrain_ds = load_dataset(\"news_commentary\", lang1=\"ar\", lang2=\"en\", split='train[:98%]')\r\nval_ds = load_dataset(\"news_commentary\", lang1=\"ar\", lang2=\"en\", split='train[98%:]')\r\n```","embeddings":[-0.1603863835,0.0044462634,-0.2170423716,0.2492842227,-0.0537915379,0.08853852,0.3200059533,0.2293999642,-0.025391113,-0.235799253,-0.2461355627,0.1912741661,0.1839801073,0.2749135792,-0.0599651821,-0.3076826334,0.0750906318,-0.1572856307,-0.0141737703,-0.1957393885,-0.1186241284,0.4341784418,-0.185517773,0.0662630275,-0.0329696573,-0.0440256409,0.0081097521,-0.0337008461,0.1106164008,-0.1234146133,0.0804368034,0.2322270125,0.1669614017,0.0935731381,-0.000112882,0.0626912713,0.1198943853,-0.1591492891,-0.1555098444,-0.29826653,-0.1329162717,-0.2726258934,0.1110204533,-0.2844849527,-0.1342525035,-0.4882697463,-0.2545556724,-0.5316694975,0.5619793534,0.4927171171,0.2283712476,0.1886770427,-0.0293710399,-0.0626298636,0.0920654163,-0.1563192457,0.1393924356,0.3267943859,0.3788948655,0.233580485,-0.1417146772,0.4323812127,-0.2732176185,-0.1382027715,-0.3607785702,-0.0472382531,0.2317581177,-0.4424891472,0.2594132721,0.209234193,0.4308646619,-0.197879985,-0.3025541902,-0.1495145559,-0.124740459,0.0732997134,0.0783393085,0.1791186482,-0.2363340408,0.4208522141,0.1015166715,-0.1204006001,0.2136509866,0.3579363525,0.0695387945,0.0383974724,0.0634348392,0.1177145913,0.0055171782,-0.2202454656,0.1597228795,-0.1910922527,0.0891226754,0.3495159745,-0.3829525113,0.0538152009,-0.2302294374,-0.1596407443,-0.1328743845,-0.2417428195,-0.2591854632,0.2534895241,-0.1386710703,0.0434368029,0.0932815373,0.0334969424,0.2530211508,0.2435534298,-0.0637139454,-0.158026129,-0.1277534217,0.3061584532,-0.0908975825,-0.3678646386,-0.2495773733,-0.0187953915,0.010044408,-0.1621397883,-0.4117812812,0.1917952746,-0.540441215,-0.4620901346,-0.0832575038,0.1982965171,0.1635905951,0.1188690662,0.1320028305,0.3674795926,-0.1292291433,-0.3483611643,-0.2025290281,0.1246428639,-0.2258711755,-0.0809420794,0.1155193523,-0.1200007945,0.3928953409,0.0889490098,0.0468147174,-0.2268930525,-0.0558341555,-0.2146737576,-0.0459015965,-0.0501574203,0.1979777366,0.3317025602,0.2548564374,-0.2848145664,0.0236388408,0.3267103136,-0.4435573518,0.310405314,-0.1882556379,0.1667381823,-0.1077538133,-0.0085980082,-0.1697164774,0.6289420128,0.3114931583,-0.0870607495,-0.0063190805,-0.0197960045,-0.146292299,-0.0890332833,0.0914800093,0.268612206,-0.7350516915,-0.1571094692,0.2101731151,-0.0272703432,0.1780678034,0.4736939371,-0.2419959754,0.2404444814,-0.2692958415,-0.1279771924,0.35891819,-0.0183163602,-0.1526155323,0.0628183484,0.1232641637,0.1729451716,0.2040342242,0.1173776165,0.2221214473,-0.0477190502,0.2874537408,0.3555460572,0.182118699,0.0138756931,-0.1741799116,-0.1975986958,0.8043532968,0.1991026849,0.3475913703,0.0937266052,0.151637882,0.0677664429,0.4992110729,0.0608986244,0.0599411763,0.2085180134,-0.3930371106,-0.0078786854,0.0847025439,-0.0669237226,-0.1784451306,0.0225504972,-0.1481734514,0.2782815993,0.0313991494,0.0713068023,-0.5101252198,-0.2441058159,-0.228397429,0.1128571257,0.174093917,-0.0561848059,-0.2003445178,0.3978535533,0.0161955133,0.3494778872,-0.395316422,-0.0775607303,-0.4699580371,0.5394413471,0.1512416005,-0.2488870919,0.0358652435,0.1230579317,0.232458055,0.0670663863,-0.2001377642,0.033492513,0.2786709666,0.3482088149,-0.0507935137,-0.0064074779,0.2219891548,-0.5121762156,0.1271721572,0.3674887419,0.0434743389,-0.0011217565,0.0626947284,0.5724955201,-0.068531014,0.1539771706,-0.0435155146,0.0383259021,0.4436298907,0.0863988623,-0.189714998,-0.3766207993,0.150515601,0.1976629198,0.3643490076,0.4173545837,-0.1284819394,-0.0297192428,0.4801755846,0.0456367433,0.1539311409,0.2072928995,-0.0988580957,-0.10108006,-0.0438236892,0.1428357959,0.0806832388,0.2600463331,0.0762111992,-0.0822421461,0.1274817735,-0.1235573962,0.2154759467,-0.0048621707,0.2227825224,-0.0436737426,-0.0389345624,-0.1674813032,-0.3917497694,0.2549133599,0.091762282,0.2471168339,-0.3253046572,0.0379047133,-0.3931705654,-0.4345546663,-0.3691056073,-0.337320596,-0.1522034109,-0.2440908402,0.0691450164,-0.3370831907,-0.0516580306,0.273757726,0.0989735276,-0.1359253079,-0.0055580237,-0.1260436922,-0.0266693514,-0.281082809,-0.1618700176,0.1251725554,0.2314505726,0.0707543939,0.1313434243,-0.4163937867,-0.474951297,-0.119828105,-0.4053294361,-0.0478653796,-0.2889056802,0.1879488379,-0.1676564068,0.3506744206,-0.121194616,-0.3291274309,0.1718019843,0.3448718786,-0.173918128,0.0501891524,-0.1361021101,0.1287710518,0.0959296823,-0.7553024888,-0.4478923082,-0.18588081,-0.2732867599,-0.1029555127,0.137365818,-0.2782262862,0.1378390044,-0.0699487999,-0.0187814608,0.0390545577,-0.4745860696,-0.1095813587,0.2636788785,0.0078690508,-0.2437195331,-0.1009240299,-0.0027440989,0.1687009335,-0.1457637101,-0.1177996546,0.1816825122,-0.1753924638,0.2152718604,0.0845397934,0.0549144745,-0.0229640584,-0.071428068,0.065899387,-0.0468740687,-0.2789111137,0.1696225405,-0.0225339681,0.1660208255,0.0673169866,0.3073378503,-0.2233282328,0.4089117348,0.2515799999,-0.1983568519,0.4292526543,-0.0283505823,0.0025143488,-0.1306345463,-0.3173961639,-0.128792122,0.120776765,-0.0733568221,0.3530524373,-0.2160768807,-0.5282369256,-0.164227888,-0.1892632544,-0.6420202851,-0.1335216165,0.0415200889,-0.5166899562,0.1264352947,0.0708089843,-0.0150549738,0.0889474526,-0.1692957431,0.0185513981,-0.0479631796,-0.1658101082,0.2463035285,-0.3809500039,-0.2149633765,-0.046963077,0.1062623113,-0.0010408181,0.45055601,-0.1928527951,0.0754275694,0.2239798456,0.2311145067,0.5450607538,-0.0847501531,-0.0873195082,0.1134038493,0.0887208059,-0.227688536,-0.1550697535,-0.337934345,0.2419273257,0.2359014452,0.2066494077,-0.2130774409,-0.201326251,-0.0040750727,0.3539861441,-0.2952656448,-0.0009571852,-0.2916831672,-0.2589244246,-0.2787749767,0.045990929,0.1793077886,0.1624131352,-0.1072355509,-0.0954643041,0.1414389759,0.1182458103,0.2900578678,0.2143933177,0.3774513304,0.2690688372,0.0625307709,0.0205370151,0.1457179934,-0.1414898187,0.4112642705,0.0600505024,-0.395006001,-0.1237328798,-0.3656805456,0.3682533205,0.1666402966,-0.1312101781,0.3721060455,0.1139569581,-0.1772342324,0.1557319015,-0.1156454459,0.4012765288,0.2329642773,-0.3053250909,-0.67359519,0.148820132,-0.0954890773,-0.1972216219,0.1967376173,-0.0047772638,-0.3827113211,0.4975559711,0.0628824905,1.0744348764,-0.0280651897,-0.0391543172,0.1542208046,-0.1277807504,0.123649165,0.2234504819,-0.2471264005,-0.2681130469,0.2510400414,0.0152117703,0.1283028275,0.1363772601,0.2492529303,-0.3342450559,0.1956286877,-0.1802161485,-0.3561236262,0.1753016263,0.093161799,0.1528226584,-0.1095376164,-0.5600212812,0.0805029571,-0.1436259151,-0.033795286,-0.0346261039,-0.0843458325,-0.0435274541,-0.1221409515,-0.234213084,0.0002533506,-0.3061943948,0.1942933202,-0.1266780198,-0.099163793,0.209623456,0.589877367,0.3825435638,0.0900685415,0.1290688366,0.0459588282,-0.0162966698,0.1403831542,0.1570024639,-0.0772414207,0.2753791511,0.0077806418,-0.4079748094,0.3461227417,0.0816487372,0.0866824239,-0.3897942901,-0.0271482561,0.108677797,-0.055128932,-0.2895777524,0.2188413441,0.0350418016,-0.2126814425,0.0971647725,0.1919373274,-0.1699073017,0.154595688,-0.1013789028,-0.4117565453,-0.0543459952,0.4425744712,0.5296670794,-0.1368502527,0.5471823215,0.3692955375,-0.2138710171,-0.1849889755,0.062248148,-0.1013345644,-0.4705782831,0.1147586554,0.0867466629,-0.0498025306,-0.2640342414,0.1789015234,-0.0241441745,0.2092762887,0.0941926166,-0.4357054234,-0.0603149757,0.1707262695,0.2167465538,-0.0992684737,0.0820893124,-0.3421545327,0.0714788958,0.0785902068,-0.2994539738,0.0846595839,-0.1085873693,0.2960164547,-0.2808215618,0.1872196048,0.0795302391,0.0016341434,0.0794659853,-0.3297418952,0.0653547645,-0.1954997927,-0.1920432746,0.0868405402,-0.1420460194,0.0796724111,0.0120769506,-0.2142714262,0.2390350103,-0.0211505443,0.2231057286,0.4757255018,0.1401373446,0.2635378838,-0.0364012681,-0.178924799,-0.3181318641,-0.1975480467,-0.0137958443,-0.1672095209,0.2358393073,0.1261312664,0.1884119064,-0.0956990719,-0.0267545078,-0.0117564239,-0.0181864537,0.1746972948,0.0015369947,-0.2785703242,0.1076115966,0.036491096,0.1765878946,0.2028993666,-0.2308778167,-0.079600811,0.1853612214,0.1772670001,-0.6147273183,-0.021338122,0.5288526416,-0.2750108838,0.144755587,0.2450878471,0.2031955421,0.2351166606,0.1567688584,-0.190795213,0.1878733635,-0.0836343691,0.2569765747,0.4046105742,0.2117767632,0.1322848499,0.0986598954,-0.0933855698,-0.1742670834,0.1631149352,-0.3341971338,0.5841134787,0.1820128113,0.0717125535,0.0548619442,-0.0903353542,0.1808563322,0.2378703803,-0.3248430192,0.0884138644,0.1094144285,0.2814013958,0.3797245622,-0.2312443256,-0.2180895954,0.0207410604,-0.2322667092,-0.0766195357,-0.2955450118,-0.2896881998,-0.4726287127,-0.1764667779,0.0326334462,-0.2177877575,0.3586580157,0.1761458963,-0.0041052201,-0.3209910691,0.0277706925,-0.2209038883,-0.079335466,-0.0303452145,0.3630647063,0.114558354,-0.1080797687,0.204437077,0.3449500203,0.1811066121,0.2049921453,0.0664626509,0.0409103855,0.1917154938,-0.0250575133,-0.0758972615,0.3343344033,-0.0258315634,0.246072337,0.1930473149,0.1758646369,-0.2258744687,0.1317781061,0.4836496115,0.1724402159,-0.1379546225,0.1059488654,-0.4913064837,0.1809763461,-0.3183490336,-0.0482853279,-0.379301399,-0.1332992613,0.3305273056,-0.1352650225,0.1345623583,0.076838553,0.1104673073,0.0673258603,0.4936191142,0.3853738606,0.332190901,-0.4657478929,-0.1169254929,-0.5569061041,0.1085769534,0.0508563854,0.083995752,-0.0349523127,0.0864709467,0.5793499351,0.4553441405,0.2812592685,0.1297930479,-0.2663858235,0.1700767875,-0.3078098893,-0.0728084147,0.0753231868,0.3558697402,-0.1880237609,-0.1337626278,0.0448234379,-0.0883316919,0.0428946465,-0.1545836627,-0.2890212238,0.2012955546,0.0260725114,0.0915936828,0.0575332232,0.3487926126,-0.1918726563,0.3104370534,0.1170603037,-0.283810854,-0.0762260407,0.3920159042,-0.0468336716,0.4638170004,-0.1197871119,0.1886683404,-0.01522604,0.1342437416,0.0824551806,-0.1377569288,-0.4191016555,0.1097268686,0.1462793201,0.005187511,-0.0509907305,0.342994839,-0.0020058656,0.1670835614,-0.0962200537,-0.151856184,0.2185222208,-0.365632236,-0.3418230712,-0.0787499622,0.109906055,0.2867902219,0.0220973454,-0.2997660637,0.1584622562,0.3886457086,0.0586730316,0.2360069901,0.305385232,-0.0925673172,0.285692513,-0.0411746912,-0.4088045955,0.23993586,-0.009883482,-0.1149385199,-0.4027515352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2190","title":"News_commentary Dataset Translation Pairs are of Incorrect Language Specified Pairs","comments":"Hello @albertvillanova, \r\n\r\nThanks for the suggestion. I didn't know you could do that. however, it didn't resolve the issue\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/8571003\/114169966-ec819400-993a-11eb-8a67-930f9a9b2290.png)\r\n","body":"I used load_dataset to load the news_commentary dataset for \"ar-en\" translation pairs but found translations from Arabic to Hindi. \r\n\r\n```\r\ntrain_ds = load_dataset(\"news_commentary\", \"ar-en\", split='train[:98%]')\r\nval_ds = load_dataset(\"news_commentary\", \"ar-en\", split='train[98%:]')\r\n\r\n# filtering out examples that are not ar-en translations but ar-hi\r\nval_ds = val_ds.filter(lambda example, indice: indice not in chain(range(1312,1327) ,range(1384,1399), range(1030,1042)), with_indices=True)\r\n```\r\n\r\n* I'm fairly new to using datasets so I might be doing something wrong","comment_length":20,"text":"News_commentary Dataset Translation Pairs are of Incorrect Language Specified Pairs\nI used load_dataset to load the news_commentary dataset for \"ar-en\" translation pairs but found translations from Arabic to Hindi. \r\n\r\n```\r\ntrain_ds = load_dataset(\"news_commentary\", \"ar-en\", split='train[:98%]')\r\nval_ds = load_dataset(\"news_commentary\", \"ar-en\", split='train[98%:]')\r\n\r\n# filtering out examples that are not ar-en translations but ar-hi\r\nval_ds = val_ds.filter(lambda example, indice: indice not in chain(range(1312,1327) ,range(1384,1399), range(1030,1042)), with_indices=True)\r\n```\r\n\r\n* I'm fairly new to using datasets so I might be doing something wrong\nHello @albertvillanova, \r\n\r\nThanks for the suggestion. I didn't know you could do that. however, it didn't resolve the issue\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/8571003\/114169966-ec819400-993a-11eb-8a67-930f9a9b2290.png)\r\n","embeddings":[-0.2070920914,0.1082228944,-0.1750500053,0.325787425,-0.1173716411,0.1063093469,0.2994769812,0.2640418708,-0.0422107913,-0.2637794614,-0.2876567245,0.2302991599,0.2286696136,0.1873274744,-0.0158621632,-0.187350899,0.0602838621,-0.1162408143,-0.0766713321,-0.2078409791,-0.1206930727,0.4833289087,-0.1554276943,0.0874612927,-0.0198556017,-0.075730741,0.1117976755,-0.0085346177,0.0858545899,-0.0731238946,0.0822786689,0.2614829242,0.1084483415,0.0069688694,-0.000112457,0.0388533138,0.1481903791,-0.1804726124,-0.1509228647,-0.2352927923,-0.189953357,-0.1916122437,0.0494393408,-0.2331949621,-0.0486273542,-0.5036552548,-0.259227097,-0.6031265855,0.4965833724,0.4825668931,0.2310813069,0.1592324823,-0.080241695,-0.0408687107,0.113849245,-0.0614906475,0.1242374107,0.3094267547,0.4024916589,0.284763515,-0.2044501007,0.4041460156,-0.2282712162,-0.1140778661,-0.3524553478,-0.1065054387,0.1520601958,-0.3482451439,0.2825841308,0.2758645415,0.5080477595,-0.1678182334,-0.1993252486,-0.163642168,-0.1394638419,0.2181403637,0.1707418412,0.1603756845,-0.2545471489,0.4165382981,0.0235113278,-0.1333047599,0.2564955056,0.2740854621,0.0819528922,0.0065442226,0.0261745285,0.1473970413,-0.0265232809,-0.2402652949,0.2194836289,-0.1835362762,-0.0394179709,0.3342429698,-0.4208230078,0.0948475897,-0.2715766132,-0.1111835316,-0.0361876003,-0.2702115476,-0.2159671932,0.229037717,-0.2124342471,0.0176820308,0.1101568937,0.0204241443,0.2749475539,0.1409305483,-0.1145376712,-0.2219070792,-0.0550830439,0.3491464257,-0.0747398958,-0.3854444325,-0.3508569896,-0.0075297602,-0.0393993296,-0.1918295175,-0.4289531708,0.1837588251,-0.4606366456,-0.4549331665,-0.0740870312,0.1213783026,0.1983214766,0.0735761821,0.1456255168,0.3672660589,-0.1184895337,-0.5046819448,-0.1860050559,0.0585050359,-0.2199442685,-0.0350082293,0.1084139049,-0.038663514,0.3271984756,0.0896235108,0.0091534154,-0.2344892323,-0.067808263,-0.207544446,-0.0875821188,-0.0745326057,0.1588890702,0.3607528806,0.1669569463,-0.2650854588,0.0145604722,0.3440955579,-0.4482043982,0.3166036308,-0.207299605,0.135767594,-0.2083462775,-0.0145419855,-0.211677447,0.7168802023,0.2936292589,-0.1661977172,-0.0369467214,-0.0093934815,-0.1457163543,-0.0237391237,0.0767944381,0.336155355,-0.8569787145,-0.1413990408,0.2179490775,-0.0559222661,0.1649335325,0.4651129246,-0.252130419,0.2206267416,-0.2703158855,-0.0620368198,0.2639403343,0.1395991594,-0.1155772582,0.0304961447,0.1279390603,0.1764543802,0.1702403426,0.0675706938,0.2028357983,-0.0669738427,0.2557018101,0.3319889009,0.204748407,0.0191822704,-0.2515000105,-0.2176717073,0.7432216406,0.1392657906,0.3220184743,0.0856108293,0.2037649751,0.1042031348,0.4789544344,0.0701840892,0.1100987792,0.2135412693,-0.4879893363,0.0136606405,0.1181341186,-0.0065688314,-0.2372851819,0.052592013,-0.1922469139,0.2402575761,0.0976162478,0.0946607143,-0.5183297396,-0.2205010653,-0.1496953517,0.1412650645,0.1538337171,-0.0047622141,-0.3061053753,0.4099654555,0.0166393053,0.3181897402,-0.4055170715,-0.141106382,-0.5065098405,0.5023491383,0.2046554685,-0.1949003339,0.0360704511,0.0416821353,0.2079787701,0.0284688175,-0.2528675497,0.0272006933,0.190356642,0.4163331389,-0.0680365562,0.0285491943,0.1770003736,-0.5377523899,0.1077106446,0.270442754,-0.0166983884,-0.0243043397,0.0595322587,0.5245122313,-0.0873628557,0.1252810508,-0.1490870118,0.0368919112,0.4804258645,0.106395483,-0.1384664774,-0.3812029958,0.1789486408,0.2783130705,0.3480254412,0.5137357116,-0.0483732559,0.009833049,0.4501891434,0.1022343412,0.1360550225,0.2385698706,-0.1030573994,-0.1707395166,-0.037024498,0.1684930772,0.0658082291,0.1991298646,0.0308090486,-0.0914862007,0.1579202414,-0.1339720339,0.1961496025,-0.0373551548,0.3182069063,-0.014592614,-0.0755701885,-0.1697480381,-0.4137766659,0.2959269881,0.2264300138,0.2362146825,-0.3682640195,0.0184783004,-0.4315112829,-0.4144338667,-0.3770637512,-0.2337869108,-0.2051527649,-0.2250044793,0.1289830655,-0.3780675232,-0.0888224021,0.2565278709,0.1377760768,-0.0182585418,-0.0960767642,-0.0535562709,-0.1146815792,-0.2418716401,-0.2124231458,0.123713918,0.2201943398,0.0656659007,0.0410545357,-0.3897136152,-0.4834827483,-0.1878939718,-0.410392344,-0.0015650811,-0.2613317966,0.1828640848,-0.1427705139,0.3141559064,-0.0985723808,-0.2879113555,0.0858922899,0.3214446604,-0.1512869149,0.076855205,-0.1768757403,0.142793119,0.0454226062,-0.7031232119,-0.3522700667,-0.165628776,-0.2152815461,-0.1919798553,0.1147099882,-0.3081685901,0.0880977362,-0.0563028604,-0.0772619024,0.0004884935,-0.5082883239,-0.1679225266,0.2492828369,0.0321363918,-0.230691582,-0.1206715107,0.0442688651,0.1717723012,-0.1149341464,-0.1849770099,0.2866252661,-0.1248209924,0.2030000091,0.0720046535,-0.15070647,-0.0101025114,-0.1209734455,0.0821487382,-0.0942245722,-0.2967146933,0.2079440951,-0.0237842016,0.2100845128,0.0603336543,0.2478287071,-0.202966556,0.3356182873,0.1705829948,-0.2107934803,0.4412063658,0.0188660696,0.0747187957,-0.1253014356,-0.2115646601,-0.0604497083,0.171161741,-0.0759840533,0.4531770945,-0.2013954818,-0.5693998933,-0.0856505781,-0.1551740021,-0.5203014612,-0.1276294142,0.0447622277,-0.5405315161,0.1157745123,0.1255958378,-0.0463960692,0.0934115872,-0.2187827229,-0.0159637984,-0.108444117,-0.0868310183,0.3002156913,-0.4213475287,-0.1029061824,-0.0097134784,0.085602954,0.0254894719,0.4773219526,-0.2653152049,0.1275766492,0.2303052098,0.2291091084,0.5325816274,-0.0887982547,-0.0412771069,0.1172797307,0.1922929436,-0.2439574599,-0.162951827,-0.288443625,0.1725551486,0.2764640152,0.1522154957,-0.197830379,-0.1684870422,-0.0418223739,0.3355362117,-0.3215472102,0.0712494105,-0.2177109867,-0.2328757644,-0.2509149015,0.0198228881,0.1966995299,0.1071152017,-0.0507809035,-0.0372980721,0.131295234,0.1556154042,0.2322296202,0.2158607095,0.3609078825,0.3124877512,0.0315983258,0.0424511246,0.09424676,-0.1116561666,0.427366823,0.1121565551,-0.4271694422,-0.0642731413,-0.4153336585,0.3232229352,0.1768146008,-0.1840081662,0.3443755209,0.1214610562,-0.225170508,0.171686098,-0.106177479,0.3313925266,0.1941072643,-0.2231347114,-0.606009841,0.1368162185,-0.1493895948,-0.2557846904,0.2395674437,-0.0266222134,-0.3724319637,0.485204339,0.167867288,1.0589587688,-0.0351571813,-0.1065522209,0.0805298015,-0.1452812552,0.1684014946,0.1785505265,-0.2806615829,-0.2597343624,0.2365649939,0.0398072153,0.140850544,0.1140275523,0.3115155101,-0.230071336,0.046827171,-0.1089319065,-0.3886446357,0.2423501015,0.0483884364,0.1796866804,-0.1412625313,-0.4036936164,0.0855781958,-0.2182266414,-0.0689880401,-0.0555506386,-0.0272183493,0.0629645064,-0.1132560745,-0.2976603508,0.0212450419,-0.2934215963,0.1737210453,-0.1397776604,0.0001138605,0.2800719738,0.5999950171,0.3800118864,0.0967751518,0.1027367935,0.1409987062,0.0695201084,0.1410850435,0.1837740391,-0.0612153001,0.213112399,0.0452286899,-0.3418715298,0.2424247712,0.0667412952,0.0292607322,-0.4200558662,-0.0834331587,0.145311594,0.0165917035,-0.1668879837,0.1984101832,0.0586388037,-0.1752560586,0.0911570713,0.1827024221,-0.1086889133,0.1305872202,-0.1456711888,-0.4750136435,-0.1150655001,0.4618605375,0.504003942,-0.1515927315,0.5482198596,0.3820110857,-0.1995596588,-0.1921114773,0.0485898219,-0.1049567685,-0.6128003001,0.1195227131,0.0028249731,-0.1312386096,-0.2441658527,0.1862153113,-0.0041138758,0.1908037663,0.0972852334,-0.4100600481,0.0260743164,0.1470229477,0.1694381237,-0.1030798107,0.1504966766,-0.3970626295,0.0595293343,0.1124793217,-0.2846499681,0.0799205601,-0.1439437717,0.3035819232,-0.2976050675,0.2210137695,0.0754348189,-0.019517988,0.0359818824,-0.2703358829,0.077673614,-0.1676853001,-0.192303732,0.0855726153,-0.1444277465,0.1185435578,-0.0103992773,-0.2335662395,0.2074431181,-0.0152421286,0.2201704979,0.472327292,0.1144115403,0.1701891273,-0.0476433784,-0.2262441814,-0.3558191657,-0.1969031096,0.0618406087,-0.156218037,0.1322892755,0.1447282284,0.1705607623,-0.0989715084,-0.0260477159,0.0426081233,-0.0046605151,0.2062832564,-0.0048490739,-0.2688602507,0.1709917188,0.0225155652,0.129785493,0.0968352631,-0.2307147384,-0.1970711052,0.18620947,0.1450797766,-0.5371164083,0.0039228485,0.4899635315,-0.2088151723,0.1046070904,0.2514387071,0.2955908775,0.1768141091,0.1996255666,-0.2214526385,0.2599437535,-0.0738697425,0.2285894752,0.402623862,0.1189539209,0.1579728574,0.1433981359,-0.081046842,-0.1764240712,0.1547015607,-0.3103887141,0.5803027749,0.2122041136,0.0943989977,0.0573751181,-0.158953771,0.2136092186,0.247156173,-0.2911059856,0.1603900045,0.1672289968,0.2373123765,0.3992871642,-0.2398177534,-0.2351872921,0.0065201987,-0.2358171195,-0.0484422483,-0.3694187105,-0.3110376,-0.4732294381,-0.1640121937,-0.0363926925,-0.1810350418,0.3520997763,0.1379123032,-0.0226739924,-0.3563269675,0.0450657904,-0.2528059781,-0.149470672,0.0124072889,0.4107272327,0.0713987499,-0.0816570669,0.3319815695,0.4165544808,0.1129616052,0.2215241939,0.2037568986,-0.00758423,0.2002260983,0.0748287663,-0.1067427844,0.3502701521,-0.0256736968,0.1786249727,0.2434384674,0.1610470414,-0.1904512197,0.0812072828,0.4890146554,0.2658087015,-0.033288613,0.0065444284,-0.5003945827,0.2084724903,-0.2983183861,-0.0586425252,-0.2563425601,-0.1303331703,0.3003174663,-0.1850001365,0.1647413522,0.1264184117,0.1150685772,0.065795891,0.6010003686,0.3236118555,0.2309814543,-0.3963485062,-0.1894579828,-0.5788682699,0.1657877117,0.0847200006,0.1364896595,0.0024742214,0.1010726541,0.6250315905,0.3840386868,0.2882041037,0.1861670017,-0.3260056376,0.1504499167,-0.3130934536,-0.0815064609,0.1224748194,0.3754122257,-0.1949742883,-0.0970989838,0.1322989166,0.0054934691,0.0285953395,-0.1130395085,-0.3280518353,0.1813624203,0.0295792278,0.0740896761,0.03200325,0.2794815898,-0.1914895028,0.342944324,0.0464442261,-0.3566104472,-0.1167313233,0.4083713889,-0.076836668,0.4339626729,-0.1212077215,0.2614528239,0.0233436469,0.0535383485,0.0698970854,-0.0115457634,-0.5153927803,0.0900912285,0.0836986676,0.0525213256,-0.0656459779,0.3392994702,0.0372986011,0.1431979537,-0.1179038808,-0.1691694707,0.231799826,-0.3227152228,-0.3061877787,-0.0129739074,0.1450244933,0.3385321796,-0.0158714801,-0.2886614501,0.1198952347,0.3938474655,0.0513232686,0.2838503718,0.2297877073,-0.1901123375,0.2200772911,-0.0667837486,-0.4433006644,0.288659215,-0.0213217307,-0.0940582529,-0.3443768322]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2189","title":"save_to_disk doesn't work when we use concatenate_datasets function before creating the final dataset_object.","comments":"Hi ! We refactored save_to_disk in #2025 so this doesn't happen.\r\nFeel free to try it on master for now\r\nWe'll do a new release soon","body":"As you can see, it saves the entire dataset.\r\n\r\n@lhoestq \r\n\r\nYou can check by going through the following example,\r\n\r\n```\r\nfrom datasets import load_from_disk,concatenate_datasets\r\n\r\nloaded_data=load_from_disk('\/home\/gsir059\/HNSW-ori\/my_knowledge_dataset')\r\nn=20\r\nkb_list=[loaded_data.shard(n, i, contiguous=True) for i in range(n)]\r\nfinal_dataset=concatenate_datasets([kb_list[1],kb_list[2]])\r\nfinal_dataset.save_to_disk('\/home\/gsir059\/haha\/k.arrow')\r\n```","comment_length":26,"text":"save_to_disk doesn't work when we use concatenate_datasets function before creating the final dataset_object.\nAs you can see, it saves the entire dataset.\r\n\r\n@lhoestq \r\n\r\nYou can check by going through the following example,\r\n\r\n```\r\nfrom datasets import load_from_disk,concatenate_datasets\r\n\r\nloaded_data=load_from_disk('\/home\/gsir059\/HNSW-ori\/my_knowledge_dataset')\r\nn=20\r\nkb_list=[loaded_data.shard(n, i, contiguous=True) for i in range(n)]\r\nfinal_dataset=concatenate_datasets([kb_list[1],kb_list[2]])\r\nfinal_dataset.save_to_disk('\/home\/gsir059\/haha\/k.arrow')\r\n```\nHi ! We refactored save_to_disk in #2025 so this doesn't happen.\r\nFeel free to try it on master for now\r\nWe'll do a new release soon","embeddings":[-0.2544762492,0.0369030796,-0.0747913346,0.0235322844,0.2022773474,0.2077016681,0.3147445321,0.3745852709,-0.0703066215,0.2908636332,-0.1010363847,0.1949494332,0.0023286699,0.2046323568,-0.0519777238,0.1872817427,0.3261983991,0.2545649409,-0.202490136,-0.1318568736,-0.3687126338,0.3375607431,-0.0871662647,-0.3545360267,-0.2599395514,0.1303224117,-0.3702034056,0.1801269501,0.0867946893,-0.0932735503,0.1130666807,0.0303959753,0.1794944555,0.1613758355,-0.0001024196,-0.0183926225,0.0984345675,-0.2301510423,-0.5061350465,0.0002731246,-0.0454339124,-0.4297147691,-0.0318013839,-0.2701205909,0.0554985516,0.0306889862,-0.0264874194,-0.3492635787,0.1595520228,0.1346326768,0.309262991,-0.0615131408,0.2263928801,-0.1241282299,0.0925930217,-0.061951369,-0.1597601026,-0.0174898487,-0.3068611026,0.0766253173,0.1838666946,0.2538936436,-0.0520645045,-0.1528071463,0.102972433,0.3315379024,0.2000031471,-0.3587841392,0.1949589849,0.0508012176,0.3223497868,-0.4228012264,-0.3215260804,-0.1421272755,-0.0708394945,-0.4146020412,0.0688096136,0.2725207508,-0.0372716561,0.1376816183,-0.2104525864,-0.0943455771,-0.069819577,0.0858969092,-0.0970790237,-0.0578370914,-0.2149947584,0.0483856872,0.4175906181,-0.2034444958,0.0756108984,-0.4940950274,-0.0186775941,-0.0991517976,-0.371789217,-0.0662881136,-0.0382048488,-0.1302770078,0.1177785024,0.1654810011,0.1567116082,0.0172697473,0.2839658558,0.1399288625,0.1291773468,0.1396843344,0.2109784186,0.068145901,0.2035749704,-0.1210986599,0.0115744416,0.0075942227,0.1597272307,-0.0621643066,0.3084388375,-0.1679115593,0.2425519675,-0.0319233574,-0.0998061746,0.2859278023,0.2349762172,0.065560393,-0.2424149215,0.3184429109,0.0408475026,0.1560940892,-0.0597217456,0.3190596104,0.1788872778,0.0181819517,-0.3325875998,-0.1384913027,0.0419947393,0.1650955081,0.1061084345,-0.1681782007,0.2190976292,-0.1321576238,0.1250537783,0.0135273319,-0.0349260196,-0.2764335275,0.3223589957,0.2634705901,0.3512492776,-0.0020948618,-0.0024642614,-0.3695912659,-0.1427434683,0.2944760919,-0.0911261812,-0.1671527624,-0.2289693654,0.3325921893,-0.0890039057,0.158902213,-0.3130002916,0.1394113004,0.3160459995,0.0520765781,-0.1731426567,-0.1179567501,-0.1283720881,-0.5556748509,0.377820909,0.1488147974,-0.0492435843,0.141371876,-0.0439600162,-0.0754638687,0.0834014639,0.1271142513,0.0235083513,0.0756958425,-0.3151823878,0.4141362607,0.3489595652,-0.0966070369,-0.3720923662,0.16935651,-0.0126689365,-0.0344887562,0.0872744471,0.1582108438,0.2718800902,-0.0721340999,0.248126477,0.2817783058,0.0836539865,0.0335968472,-0.3789056838,0.1018801481,-0.042809505,-0.1003163457,-0.2233697176,0.2506659329,0.0871279091,-0.1296687871,0.3815090358,-0.2515321672,0.2696840763,0.438035965,0.2545845807,-0.2948097885,-0.1404696703,0.125088796,-0.6262158751,0.1282477081,0.2535210252,-0.1669909507,0.0085176928,-0.1840660721,-0.0659888312,0.1249769032,-0.2429518849,0.0508549958,0.2963756621,0.2682271004,0.1049840152,-0.0379334837,-0.1391397417,0.1948613375,-0.0677439421,0.0089129601,-0.256256789,0.2754266262,-0.166095227,-0.3440093994,0.0040625175,0.0395993441,0.1255230457,-0.0486763008,-0.0014976056,0.5335175395,-0.1285197139,0.064571552,-0.0740877315,-0.0528816469,-0.033474654,0.0302071832,0.0748106986,0.0701766685,0.1700238585,-0.0647116527,-0.3401039243,0.3452366889,0.1025321856,0.2119351029,0.2155987918,-0.0051220949,0.1724714935,-0.1435659975,0.0114306547,-0.2770319581,-0.0996472463,0.1160653681,0.1534866542,-0.0418848917,-0.1850672215,0.2772960663,0.5342244506,0.0047868043,0.2684757113,0.1592027992,-0.3228473365,-0.2200796157,-0.0512727015,0.2949929833,0.5205633044,0.2487683296,0.2373423725,-0.0739356056,-0.0226968527,-0.1503005177,0.3239525259,-0.0092433114,0.1937918961,0.4255512059,0.0722199753,-0.1734611094,-0.2458447516,0.216474548,-0.0804639757,0.0885643438,-0.228697747,-0.1394912004,-0.1880013943,0.1280589253,-0.1593550593,-0.1150401607,-0.1817211807,-0.5245651007,-0.132096678,0.4993870854,-0.1842717975,-0.0487550646,0.1641299576,0.0706619993,-0.0039135288,-0.1958074868,-0.0752318576,0.0643517524,-0.1250968724,0.2130450606,-0.0566554144,-0.181535989,0.2067794353,-0.109968923,0.1588941365,-0.3810654283,-0.0727515593,0.0060042944,-0.1633942723,0.0072642621,0.0531139337,0.2406397462,-0.0669149756,-0.2053179592,0.157913655,0.0342975482,-0.2009773403,0.322337389,0.1404600888,-0.1537690461,-0.1958235949,-0.2867220938,-0.1278878003,-0.3667713702,0.2408263385,0.0425303727,0.1054977402,0.0470765643,0.3343971968,0.0507683307,0.1508998424,-0.1325421482,-0.2667574883,-0.4190298021,0.5335767269,-0.3335801363,-0.4781666398,0.2454314977,-0.0832561553,0.0287268851,0.0784360543,-0.4200796187,-0.051572755,-0.0873609185,0.041065041,-0.0841630474,0.2444756329,0.290635258,0.1483316123,-0.1550348252,-0.1538979858,-0.2125998288,0.257129252,0.1114040986,0.3442560136,-0.1954375505,0.1338014156,0.1407623738,0.1544995606,0.1809569895,0.1047078893,0.3695558608,0.101584442,0.5001899004,-0.3050990701,-0.2166190594,-0.1263912469,-0.1631068587,-0.0570841208,0.1586762816,0.014714404,0.0458634757,-0.011292317,-0.0382191204,-0.1091692969,-0.317756027,0.0651757792,0.1224088594,0.1376289725,-0.0586149432,-0.0128374845,0.1147705689,0.0393821411,-0.1008008122,0.1192648783,0.1517088711,-0.1176706254,-0.498912394,-0.0271281824,-0.5107488036,0.3123718202,-0.0537328608,0.0665755942,0.1716881096,-0.3388547003,-0.1255565137,-0.1214795411,0.5472689867,-0.1033595502,-0.1469325274,0.2647749484,-0.0213549659,-0.3900559545,0.0049922774,0.113587141,0.0322965607,-0.091412507,0.4053320587,-0.2943360806,-0.0033263345,-0.0025433777,0.5030466318,-0.2902576327,0.0526016317,-0.3101170957,-0.1842461526,-0.5458337665,-0.2709190547,-0.1587413549,0.2314002365,-0.1468465179,-0.1265230179,-0.1977181435,-0.555107832,0.1449787915,-0.1743404269,0.5804968476,0.0656959042,0.3546785116,-0.1276699305,0.2457726598,0.2979593277,0.5010453463,-0.2553141117,-0.2001609653,0.083232291,-0.2154494524,0.3007062078,0.2000466436,-0.1495770961,-0.191849038,-0.2488209754,0.0843841508,-0.1546165496,0.0426697433,0.0459745675,-0.1262765229,-0.3833072782,-0.0453319699,0.3572864532,-0.1167091057,0.1183473542,-0.0412221216,0.0118026566,-0.3274763525,0.4180617332,-0.0403686948,0.7842453718,0.2104879767,0.4815624952,0.0452086255,-0.0732950866,0.0327204429,0.0996644795,0.1373757422,-0.4184630513,-0.1445324272,0.0372516401,-0.0154119851,0.0415755212,0.1034914926,-0.4483249485,0.1474428624,-0.3108688295,0.0624462217,-0.2164748013,0.1173091307,-0.1834235638,-0.4166025817,-0.1221856698,0.2904290259,0.2758099139,-0.0950816944,0.1101950258,-0.189283371,-0.2790108621,-0.0953041464,0.059640225,0.284381628,-0.1293841153,0.3061514795,-0.3621830642,-0.4505599439,-0.5449163318,0.3328428864,0.3889661729,0.1015559882,-0.1239499003,-0.1875551194,0.0174622089,0.0126239685,-0.1784996986,-0.1696206629,0.2127749473,-0.0724227726,0.0666665882,0.0399627276,0.0372539312,-0.1777832061,-0.0693111345,0.0032457844,0.0849566609,-0.2288416326,-0.2024765611,0.1848591864,-0.2149599195,-0.3011002839,0.2421291471,0.1823656559,-0.2337661088,0.2335448861,-0.2284679413,-0.0486505404,-0.0894216299,0.4123314023,0.3013586104,0.0898723379,0.3267374039,-0.1710891426,-0.0307906959,-0.2798591554,0.2657255828,-0.0210408643,-0.0508862212,0.3845867217,-0.1155890226,0.0356925055,-0.1603339016,0.3472803831,-0.0324776098,0.043126639,-0.0502112918,-0.1576770693,-0.405652374,-0.0964472368,0.1587664783,0.3276346326,0.0187358726,0.0800635666,-0.0844473243,0.1526060998,-0.4565459788,-0.0644526184,-0.0799507648,0.0790908113,0.1535951942,-0.0079387343,0.2262550443,-0.1351850778,0.2722092569,0.1451885998,-0.0681955889,-0.4086963534,-0.0664042234,0.0627454668,0.0560973659,0.0908121243,-0.0868617073,-0.2218414247,-0.0972003192,-0.3641072512,-0.0661831796,0.0047877589,-0.0787914991,-0.0098902462,0.2135493755,0.2411317825,0.1394850165,-0.2031902671,-0.0645401105,0.1836986095,-0.2426444888,0.1101425439,-0.2292162329,-0.2055710852,-0.1175106391,-0.0215054397,-0.0502378158,0.0065395003,0.1916694641,-0.194959864,-0.3101609647,-0.2141376287,0.4370488822,0.1747402102,-0.2056213766,-0.0758357346,0.0738587976,0.3058640659,-0.1324885488,-0.31512025,0.1374548376,-0.1827314198,0.0190911405,0.1574217677,0.1232967079,-0.0573543757,-0.277420938,0.1031668708,0.1321495175,-0.0668754801,0.094344385,0.3355685174,0.0363536924,0.4215303659,-0.0219184645,0.1519083232,0.2003871948,0.4646340311,-0.2015645206,0.4510605037,0.3531264961,-0.2047065496,0.0448162779,-0.2510586381,0.3148807883,-0.0124377338,-0.1210377887,0.2766667604,0.0886270031,-0.0924845189,-0.1561163068,-0.2416667193,-0.1868081838,0.2985247076,-0.1953546852,-0.1328549832,0.1940386593,-0.0372732133,0.0200108197,0.0573377497,0.0927765071,-0.3903914094,0.2720750868,-0.0313368104,0.2085283995,0.0938594043,0.1055419743,0.0666661561,0.0330835171,-0.1981871277,0.0135862371,0.1995747983,0.0195501316,-0.0696282387,0.1215837598,0.3711819351,0.3659860492,0.0907474086,0.2423739731,0.14318569,-0.1749037504,-0.1607588828,-0.0036746915,-0.0122542148,0.0423722938,0.4947176874,0.2761059403,-0.2584260106,0.1484420747,0.2257326394,0.0316300094,-0.2331473082,0.1536015421,0.0357049182,-0.4102100134,-0.0663207397,-0.0989831612,-0.2807653844,-0.2876633406,0.4731296897,-0.1005418003,0.1434698254,-0.1104007736,0.1654275358,0.0616376698,0.4518494904,-0.069725506,-0.0446479917,-0.3877370059,-0.2134353966,-0.4200601578,0.0223822054,0.1277050972,0.2152198553,-0.201420337,0.4583127797,0.0815735683,0.1868253052,-0.1017854959,-0.0872710422,-0.0183258746,0.3318609893,-0.2051381022,-0.109272778,0.0771486908,0.1314721853,0.1458146721,-0.4644191861,0.3085884452,-0.0104239546,0.2647193372,0.0284816232,0.242205888,0.1414976418,0.1359400153,0.3980335891,0.1207735315,0.1030346081,-0.3297482133,-0.2140914798,0.1160089001,0.0324001648,-0.0965127945,-0.0115340818,0.1447606087,0.358196348,-0.2636270523,0.1054234356,-0.0446590632,-0.0452230498,0.3222957253,0.1339650452,-0.0896758288,0.2718452811,0.1822461635,0.113340579,0.1858025193,0.255870223,-0.1761450469,0.2921958268,-0.2606454492,-0.4027805328,0.3502841592,-0.0406880118,-0.2211070806,-0.0238411557,0.2712753415,0.1372812688,0.123991102,-0.5584985614,-0.0826585665,0.3341221213,-0.1519653201,-0.3571297228,0.353792578,-0.1059932485,-0.005755438,0.0049680094,0.3333816528,0.1975370497,-0.4678483009,-0.032501433,-0.459418565]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2188","title":"Duplicate data in Timit dataset","comments":"Hi ! Thanks for reporting\r\nIf I recall correctly this has been recently fixed #1995\r\nCan you try to upgrade your local version of `datasets` ?\r\n```\r\npip install --upgrade datasets\r\n```","body":"I ran a simple code to list all texts in Timit dataset and the texts were all the same.\r\nIs this dataset corrupted?\r\n**Code:**\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(*timit['train']['text'], sep='\\n')\r\n**Result:**\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\n...\r\n...\r\nWould such an act of refusal be useful?","comment_length":32,"text":"Duplicate data in Timit dataset\nI ran a simple code to list all texts in Timit dataset and the texts were all the same.\r\nIs this dataset corrupted?\r\n**Code:**\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(*timit['train']['text'], sep='\\n')\r\n**Result:**\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\n...\r\n...\r\nWould such an act of refusal be useful?\nHi ! Thanks for reporting\r\nIf I recall correctly this has been recently fixed #1995\r\nCan you try to upgrade your local version of `datasets` ?\r\n```\r\npip install --upgrade datasets\r\n```","embeddings":[-0.0085601602,-0.3116270602,-0.0845909119,0.6408435106,0.2972834706,0.1798605472,0.2127219588,0.3483661413,-0.3550944924,0.0640646145,-0.1848578006,0.3607736826,-0.0370692685,0.1377495974,0.1514657885,0.1737731546,-0.0231208708,-0.0077550919,-0.289678067,-0.2847210169,-0.0019159283,0.1685191691,-0.0928682387,0.0616810247,-0.0929964259,0.2439683527,-0.1059634238,-0.1654433757,-0.0110924039,-0.2916881442,0.039128527,0.1391637921,-0.1956939995,0.4862783551,-0.0001211682,0.1176715791,0.141956076,0.0384408645,-0.2202286571,-0.2915332615,-0.1174837649,-0.1514540017,0.2768204212,-0.093014434,-0.1665944755,0.0084072202,0.0813450813,-0.1515229493,0.3422286212,0.1550573856,0.1326091737,0.1251025796,-0.0319642946,0.1261979192,0.2245519608,0.1391940564,0.0429698601,0.1158418208,0.351942867,0.365419805,0.4090771675,0.1617848128,-0.2078804225,0.2010411173,-0.3571011424,0.0607093535,0.0050468207,-0.1050854921,-0.0295727327,0.3857053518,0.8181933165,-0.3492590189,-0.1290716231,0.0266013034,-0.0413215868,-0.1824618429,0.1129803434,0.1279145926,-0.2183217853,0.161939308,-0.1877855808,0.1523911208,0.041346591,0.1306083202,0.081719704,-0.3430477679,-0.0316619277,0.0668796524,-0.0958121046,0.0588132851,0.12808685,-0.1315314174,-0.0388975516,-0.0530780815,-0.0954163894,-0.0338919684,-0.5064440966,-0.1193039939,0.172377333,0.2862872481,0.4710802138,0.0988899469,-0.2946043015,-0.0915646106,0.475292325,-0.06704925,-0.1020936668,-0.0530314967,-0.1388146281,-0.2779713571,-0.396797061,-0.026198633,0.0038941933,0.0036781437,0.2771374285,-0.0767455623,0.3120041192,-0.2282845825,-0.5888941884,-0.0935210958,-0.4488693178,-0.0916466266,-0.1306605637,-0.1027447209,0.0775114745,0.0874411389,0.0335715637,0.0318364874,-0.2102265209,-0.1265277565,-0.1079942286,-0.306145221,-0.0174625721,-0.077755779,-0.009289748,-0.3747313321,0.1959086359,0.29171893,-0.0306807812,-0.2282073945,0.0096422089,-0.1862995327,-0.0141302561,0.0944495574,0.0140463412,0.3920698464,0.2078878731,-0.0813206136,0.0554698072,0.2834269106,-0.2466738671,-0.1098549291,0.113766335,0.1410531104,-0.0766588002,-0.1108217612,-0.1702792197,0.1028332189,0.2403246611,-0.1323535144,0.03633729,0.0477540791,-0.218952477,-0.0610455163,-0.18714948,0.343459785,-0.5317342877,0.0615807585,-0.0243831296,0.0334308296,0.0333389305,0.3163551986,-0.1876560152,0.369735986,-0.2389821559,-0.152196452,0.1076035202,-0.2316105068,-0.3582291901,0.0533435643,-0.0566046461,-0.0179928467,0.0473609604,0.2656922042,0.1468071342,0.0304086693,0.0650881603,-0.2567756176,0.2930332124,-0.3015697896,-0.4290743172,-0.06268657,0.4001121819,-0.0568456091,-0.2479376495,0.0383779816,0.073491253,0.0613337196,0.3726989031,0.2081892043,-0.1574118286,0.2859969735,0.2329321504,0.3830018938,0.2396134436,-0.1696216166,0.2872486711,-0.0753625929,0.1877301931,0.2143067569,0.2333484143,-0.2413323671,-0.1373885572,-0.0262933951,-0.3445990086,0.1045035645,0.1936203688,0.2245903313,-0.3319903016,0.1008642688,-0.097961545,0.6168999672,-0.3779688478,0.0188989546,-0.1586885154,0.1941652894,-0.1101722419,0.1503379047,0.0451617874,0.0427666083,0.0110732764,0.0053405897,-0.1109578982,0.1506658643,0.1766404063,-0.1716476828,-0.2193176895,-0.3255206048,0.2579401433,-0.5105289817,-0.2207472622,0.5468989611,0.2125024647,-0.1158464402,-0.0344049409,-0.0573118851,-0.222957477,-0.0068441378,-0.0107683185,0.1831531972,0.0709391758,-0.1089939922,0.0113869375,-0.2626825273,0.3656535149,0.1338628232,0.1802449822,0.3277192414,-0.0700456724,0.1865557283,0.0942412168,-0.0637165084,0.0107965237,0.0091217747,-0.0198521186,0.3328455687,-0.1506571621,0.1400274634,0.2939414978,0.1768110245,-0.0199784096,0.1302333921,0.029967472,-0.0383365676,0.2424828559,0.179170996,-0.1587680727,0.2704196572,-0.0308715347,-0.0532163829,-0.5526506305,0.2271169424,0.2784294486,0.3408263326,-0.2932356596,-0.0568935461,-0.3629149199,-0.1330238432,-0.5408290625,-0.0371635184,-0.2454103529,-0.1514032781,0.0793837905,0.0795682669,0.2793279588,0.1364261955,-0.0461639129,-0.1637203842,-0.0556844063,0.0606978349,-0.1150866449,0.097170651,-0.3198128343,0.0955677778,0.225795716,0.1798101664,0.238954559,-0.3659195304,0.0553943254,-0.343419522,-0.312902838,0.1037420407,-0.1895084232,0.3461570144,0.1730522215,0.1027155668,-0.3056551516,-0.2064390182,0.0690842494,-0.1115272939,-0.3633112013,0.0304775815,0.0058882791,-0.1495912969,-0.0017719446,-0.6204960346,-0.2918468714,0.062333867,0.3842011392,0.0756763145,0.1663363874,0.3559519947,-0.0517888367,-0.0216832906,0.2291805744,0.2381192893,-0.5113918781,-0.2132319957,-0.0249327552,0.1379188448,-0.2106077671,-0.1239940599,0.0099257389,0.109306246,0.0942362323,-0.5886982679,0.0314737521,-0.0134940101,-0.1855831742,0.0318211131,-0.0464713387,0.2385500818,0.0799124837,-0.0116049452,-0.0772038549,0.0090622306,0.0986079127,-0.1817134172,0.3093614876,0.0258249231,0.1898767948,0.0985339135,0.2369150072,0.63784796,0.0301537644,0.2011714876,-0.2357314974,0.495657295,-0.0320334621,-0.4854995012,-0.2749164402,-0.1550837308,-0.2376657426,0.2133058459,-0.0062685562,0.1792380214,0.0842150226,0.1413427442,-0.4024018943,-0.1958308071,-0.209246099,-0.2632051706,0.1199411079,0.1868512779,-0.1118007377,0.1284536719,-0.3919721842,-0.128160879,0.2763438821,-0.2326485366,-0.0987101793,-0.5177311301,-0.3856748343,-0.157224983,0.2232970595,0.3914359808,0.3871856034,0.0129344016,0.1227082983,0.3271722794,0.0899288654,0.6433966756,-0.4577264488,-0.225581497,0.2776937485,0.1923351139,-0.3458144963,0.0036671567,-0.2985073626,0.1101403013,0.1936447322,0.0200168416,-0.1690909714,-0.057779327,0.2804934084,0.3153574467,-0.3922641873,-0.0374699645,-0.2901182771,-0.3878030777,-0.0596343167,-0.2520772815,0.0017683633,0.0339680389,-0.2614023387,0.0808122233,-0.1597418934,0.2822009325,0.4743410647,-0.2297744155,0.5287470222,0.0183319636,0.3186362684,0.038356889,-0.0256382376,0.0346134342,0.794262588,-0.1817183197,-0.0518700927,-0.2749042511,-0.5175102353,0.3264536858,0.2323887795,0.1002618149,0.2010602951,-0.3210880458,-0.2144898474,0.0877313614,0.0226800777,0.4050423503,0.0403066725,-0.3864827752,-0.5881889462,0.1536230743,0.2283320725,-0.212581411,0.4853883982,-0.2419734746,-0.1396205723,0.2389875948,0.0198102687,1.0326571465,-0.3713350892,0.2594498992,-0.0434149094,0.1413666457,0.1241544411,-0.2882463634,-0.0225446727,-0.2724564075,-0.2003300339,-0.1822969466,-0.001937166,-0.1124415845,0.3325400949,-0.2659353614,0.1491214484,0.0046444484,0.4775801599,-0.0332907364,-0.0351792984,-0.0386375189,0.2886991203,0.1627724767,0.1069134101,-0.1234183684,0.1513611674,0.1287584603,-0.0298846513,0.1740528792,-0.2242054045,-0.6097682118,-0.2163857967,-0.2782771289,0.2034094334,-0.1606040597,-0.5160952806,-0.0829311684,0.2894191146,0.407776475,0.2702653408,-0.1323209703,-0.0527479164,0.3445211649,0.1857606769,-0.1239494383,-0.043912787,0.0764124095,0.1406891346,-0.2715586722,0.0058448263,0.3306662738,-0.3195129633,-0.4827293754,0.2105579972,-0.0424486548,-0.5394480824,0.2258798033,0.3129874766,-0.1576694548,-0.2404380888,0.0607078373,-0.1167857051,-0.1838815957,0.2974874079,0.0382159352,-0.3845880926,-0.0466835387,0.3951343596,0.6507580876,-0.0362954661,0.5934002995,-0.0796872899,0.032568533,-0.2607859969,0.2576854527,-0.0785075352,-0.409745276,0.053650897,-0.2556084394,0.4455704391,0.1403000653,0.232606411,0.0997582451,0.1492038071,-0.2308324724,-0.2159941792,-0.2349487543,0.1496259719,0.3456688821,0.4682068229,-0.0830982029,0.1155192703,0.0606956035,-0.162424311,-0.2127869576,0.4024803638,0.0188986622,0.2500678599,-0.1053192914,-0.0827531815,0.1098582521,-0.0133245485,0.0944070891,0.1693226248,0.0219434462,-0.1436608285,-0.1135645658,0.1853685677,-0.0106775472,-0.0386763029,-0.4134532511,-0.2445090562,0.2411625832,0.0747652128,-0.1334320307,0.3666664362,0.1888755858,0.1262804866,0.0740828663,-0.0769250914,-0.0899789259,-0.1799811572,-0.1865112782,0.2855626941,0.326489687,-0.030424241,0.1567704529,0.0246068798,-0.3166654408,0.0635849461,0.2275947332,-0.0632548481,0.0260907356,-0.5054306388,0.1208677664,0.1265207827,0.4913758039,0.4850880504,-0.2948520482,-0.0272346251,0.0519161522,0.1407959461,-0.2662788332,-0.1509933025,0.0380292647,0.0395000018,-0.0849535316,0.2165424526,0.2520009875,-0.3805686533,0.3474817574,-0.0954950303,0.3520767093,-0.3051479757,0.174756363,0.4739437103,0.2302883714,-0.2237274349,-0.0315309241,0.1291671246,-0.0878963992,0.5185220242,-0.0401065685,0.1377223581,-0.3086026013,-0.0972528383,0.15413405,0.0535965413,0.1003666893,0.2339375764,-0.2164604664,0.2790325284,0.2260108739,0.4698835313,0.0167086124,-0.1919236779,-0.383007288,-0.1604190767,-0.240869388,0.0084330617,-0.1273938864,-0.0666005835,-0.0035137967,0.0644068569,-0.0064332173,-0.1346313357,0.2504635453,0.2350131273,-0.0574409217,-0.7289216518,-0.1670006812,0.2327004224,-0.0303845927,-0.2189621329,0.2196619511,0.2771153152,0.0529672317,0.1772486269,0.5775400996,0.3732342124,0.0633336157,0.0164404772,0.0206761975,-0.448653698,-0.011061376,-0.0365728661,0.4057448804,-0.0987177268,-0.06709975,0.1446692497,0.0500505231,0.0595599897,0.1273137629,0.2412568629,0.0006836643,-0.5757218003,0.3191532493,-0.0570687018,-0.0943147391,-0.1890242845,-0.0539693311,-0.4186169803,0.1573536545,0.7430905104,0.1384953856,0.1193907261,-0.0645652339,0.0558846369,-0.0465303585,0.1426379979,0.291202575,0.0155902831,-0.2468739003,-0.0864060819,-0.6696778536,0.2713157535,-0.4794603288,0.2054276019,0.092501007,-0.0102427732,0.0602673441,0.4230240583,0.2584810257,0.0538322926,-0.6601259112,0.2825868428,-0.4204730093,-0.3259839416,0.1669091582,0.094342418,-0.088654682,-0.1949442029,0.1824276745,-0.1806586832,-0.0176795591,-0.246689871,-0.0139554422,0.0855448321,0.354644835,-0.0831909701,0.0379626714,0.3189561367,0.1843013316,-0.0486692749,-0.0238231011,-0.4219785333,-0.1447987854,0.6116062999,0.1478206664,-0.135345906,-0.1828469932,-0.1231040955,0.083442308,-0.0091937017,0.2873488963,0.1239854395,-0.2562179267,-0.2145985961,-0.1380327344,0.189276427,0.0116369082,0.6060588956,0.2208260149,0.3174074888,-0.1575764269,-0.5446339846,0.4519626796,-0.564443171,-0.2094043344,-0.0854327902,0.0419431329,0.1815060824,0.1017209068,-0.5763419271,-0.1168334037,0.6631140709,-0.0973281488,-0.0683502406,0.3364093304,0.1263561994,0.0116316183,-0.0046614367,0.5800077319,0.1627821773,-0.2072148621,0.4560902715,-0.2182007581]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2188","title":"Duplicate data in Timit dataset","comments":"Hi Ihoestq,\r\n\r\nThank you. It works after upgrading the datasets\r\n","body":"I ran a simple code to list all texts in Timit dataset and the texts were all the same.\r\nIs this dataset corrupted?\r\n**Code:**\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(*timit['train']['text'], sep='\\n')\r\n**Result:**\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\n...\r\n...\r\nWould such an act of refusal be useful?","comment_length":10,"text":"Duplicate data in Timit dataset\nI ran a simple code to list all texts in Timit dataset and the texts were all the same.\r\nIs this dataset corrupted?\r\n**Code:**\r\ntimit = load_dataset(\"timit_asr\")\r\nprint(*timit['train']['text'], sep='\\n')\r\n**Result:**\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\nWould such an act of refusal be useful?\r\n...\r\n...\r\nWould such an act of refusal be useful?\nHi Ihoestq,\r\n\r\nThank you. It works after upgrading the datasets\r\n","embeddings":[0.121791631,-0.3184637129,-0.0837452337,0.7355444431,0.2122436166,0.2201525122,0.1826737523,0.3428937197,-0.395878464,-0.0273982454,-0.227510944,0.2680452466,-0.0534021705,0.1220525876,0.1580957472,0.215275377,0.0122375684,-0.0032453658,-0.3113051057,-0.3277761936,-0.1055735052,0.1642697304,-0.1561585367,0.0678436235,-0.0755195022,0.2314480841,-0.2280749083,-0.0841313899,0.0218036342,-0.1920224875,-0.019182682,0.1936765909,-0.1815677136,0.4815202653,-0.0001220929,0.1414220482,0.06391415,0.0211020615,-0.1527284086,-0.1534835696,-0.1733815968,0.0280395988,0.1926930249,-0.1067950651,-0.1812706739,0.0580070391,0.0265970584,-0.2345749736,0.3397452235,0.1985892653,0.1082001999,0.0638841689,-0.0227684118,0.2139966488,0.2378153354,0.2042028159,0.075853534,0.1509079188,0.3090035617,0.4288197756,0.4270784557,0.1161005944,-0.1331561208,0.1282091737,-0.3832941949,0.0398590304,-0.0435583591,-0.0968168676,0.0140783014,0.3854770958,0.7993019819,-0.3010451198,-0.0815758407,0.0735850334,-0.0135764331,-0.1503840685,0.1358372569,0.0800255761,-0.1193960831,0.1904778481,-0.3148603439,0.1584532112,-0.0027344904,0.110743925,0.1111518294,-0.3964847624,-0.0579684265,0.0808081031,-0.214551717,0.0530832969,0.1133664995,-0.2595157325,-0.031556353,-0.129690513,-0.1341753602,-0.0557578318,-0.5319383144,-0.1631933302,0.1726861745,0.3063330948,0.4566307366,0.09537334,-0.2838374674,-0.1934747994,0.4608413875,-0.0704361647,-0.092596449,-0.0700230151,-0.1404509991,-0.2514720857,-0.4507479668,0.0252217297,-0.034152735,0.01156954,0.3143877089,-0.1505611241,0.3315055966,-0.2477264553,-0.5225642323,-0.1016354635,-0.524723649,-0.1063947007,-0.1115421131,-0.1247172877,0.1364661157,0.0171263367,0.0458240397,-0.0004205337,-0.1505548954,-0.1328650415,-0.1135696173,-0.3031435907,-0.0310825873,-0.0042018476,-0.0722745731,-0.3445942402,0.1633076668,0.1930935085,-0.0089735826,-0.2643534541,0.043051336,-0.1618561,0.014490311,-0.0109156556,0.0019330108,0.3867740035,0.1493273228,0.0237817001,0.0878724083,0.3101830184,-0.1984889507,-0.0915261284,0.2498639673,0.1400207579,-0.051782649,-0.0625877082,-0.1361445785,0.1183945686,0.2674144208,-0.1176976338,0.0893815383,0.0719639584,-0.2298419178,-0.0477525927,-0.2034578025,0.3960454464,-0.4948565662,0.0786115229,0.0162065458,0.0337894484,0.034657456,0.3321011961,-0.183625102,0.3942420781,-0.2592499554,-0.1296638399,0.0664178133,-0.1495415121,-0.3100149333,0.088939704,-0.0407875441,0.0156823974,0.0800389498,0.3058664203,0.1410998404,0.0347495489,0.0616623126,-0.3148037195,0.3214288354,-0.3420423865,-0.4005632102,-0.0065164063,0.5016216636,-0.0290462095,-0.2721210122,0.0807286873,0.0709930956,0.0724740699,0.383205533,0.1725751758,-0.1685519367,0.2326631546,0.140047729,0.2826251388,0.2657576799,-0.1284310669,0.2700634897,-0.0918077901,0.172363162,0.1829301417,0.3704684973,-0.1863702536,-0.1236576661,-0.0987772569,-0.3953477442,0.0766916871,0.1678383797,0.2849641144,-0.3120980561,0.086503312,-0.1258276403,0.630196631,-0.4772153795,-0.0370175987,-0.1528442204,0.2421980053,-0.0845591947,0.1587687135,0.0835612863,0.0383204818,0.003617218,-0.0424087159,-0.1133001074,0.1204374731,0.1610830128,-0.2377107888,-0.1829768121,-0.3222615123,0.2777871788,-0.5195671916,-0.2598215938,0.536167562,0.2229965627,-0.1567326635,0.0017963346,0.0020314748,-0.1889690459,-0.0058292802,0.0235361662,0.1106101125,-0.0326552205,-0.118028149,0.0782117322,-0.2172927707,0.3532127142,0.2253308445,0.2155521363,0.3836679161,-0.1081461087,0.1796711385,0.1273904294,-0.1755486727,-0.0253297798,0.0006597688,-0.0057662791,0.293767035,-0.1932585984,0.0427354947,0.3237374127,0.1725208461,-0.0720744357,0.1383266002,0.0582093447,-0.0374686867,0.2538720369,0.1355242282,-0.079929024,0.238954246,-0.0972284526,-0.0146348709,-0.5417348146,0.1679864973,0.3951684535,0.2930653989,-0.3293259144,-0.1234518215,-0.3082023561,-0.1475128978,-0.5244300961,0.0563639924,-0.213119328,-0.122541897,0.0725020319,0.0268609077,0.3663319945,0.0909325927,-0.0805545226,-0.1099076718,-0.050438337,0.0734489784,-0.1156585068,0.1098556444,-0.3590046167,0.1102239117,0.2428261042,0.1567851156,0.1719490886,-0.420194,0.0858454853,-0.3456212878,-0.2438534051,0.1249740124,-0.2393015325,0.2362561524,0.1147377044,0.0217624027,-0.3126451373,-0.1216524243,-0.0164061747,-0.0572288483,-0.4619587064,0.0513837747,0.0080047613,-0.1998727769,0.0490282252,-0.6930491328,-0.1604251564,0.1659904867,0.4014916122,0.0683657676,0.181348592,0.4118545651,-0.1101761982,0.0235635433,0.2946431935,0.2877390981,-0.4960514903,-0.2778726816,-0.0439483225,0.1706279814,-0.2388302237,-0.1834415644,0.0588556491,0.1230088547,0.0693408474,-0.6029294133,0.0319383703,0.0126364008,-0.2367658019,0.0625396147,-0.1345989704,0.2208443731,0.046646107,0.006309534,-0.0321749225,-0.0500874147,0.1555198282,-0.1667348891,0.2932550013,0.0909528509,0.2083243877,0.0390193835,0.1763544232,0.6163191795,0.0060264347,0.1735677123,-0.2361683398,0.4764597416,-0.0050493991,-0.412419945,-0.1453075111,-0.1004958376,-0.2170898914,0.2544211745,-0.0227841828,0.1646554917,0.1114331558,0.0489147343,-0.4283204377,-0.184713766,-0.2330495417,-0.2766540945,0.0402159691,0.1628396362,-0.1644940972,0.1851639301,-0.3634041548,-0.0772562847,0.3257243335,-0.1623855531,-0.1239447743,-0.530310452,-0.3468016982,-0.1239233539,0.2441472113,0.3599576354,0.4005360007,0.0173417926,0.1214714646,0.3590462506,0.0485660434,0.7170503736,-0.4396269321,-0.1889392436,0.2935365438,0.1260062307,-0.2703633904,0.0548045672,-0.3124841154,0.070222117,0.2293069214,-0.0030849138,-0.159537524,-0.0594488159,0.2312138528,0.2800856829,-0.4290494621,-0.0975411162,-0.2848863602,-0.3667180538,-0.0520198569,-0.2940282524,-0.0740057901,0.0353336632,-0.2780554295,-0.0081194239,-0.2394134402,0.1655859947,0.5383916497,-0.2100996077,0.4524310827,0.0552068353,0.3000417948,0.0232622363,0.0127987489,0.0305280536,0.7020853758,-0.1101509929,-0.0451546088,-0.1985320747,-0.5856593251,0.3347100317,0.2310136408,0.1147217155,0.2241978347,-0.2905902863,-0.1790776104,0.1204068214,0.0273514744,0.433991611,0.0127339279,-0.486137569,-0.5610536337,0.0554112829,0.1826503873,-0.205021739,0.4627156556,-0.1538569331,-0.1886115819,0.1586632878,0.0859329179,1.1006901264,-0.4329958856,0.2531535625,-0.073350139,0.0999082923,0.1174926981,-0.2297163308,-0.0277672149,-0.2634464502,-0.2014838457,-0.229747951,-0.0085547576,-0.1096983626,0.3428320587,-0.2375071049,0.1767548919,-0.0516323224,0.5246623158,-0.0028527312,-0.0008790351,0.0843031928,0.3533161283,0.1530059129,0.0788141936,-0.047905989,0.0731123909,0.1572681069,-0.0269659217,0.2098098695,-0.1584460139,-0.5043452382,-0.1796045899,-0.2620665431,0.1764969677,-0.1518255919,-0.4737109542,-0.0605670288,0.3486547172,0.4583916068,0.2403914183,-0.1163769588,-0.1692850143,0.3445759714,0.1797903925,-0.1559415013,-0.0777325705,0.1159558743,0.1414407939,-0.31671229,0.0747641101,0.3337687254,-0.3598433435,-0.526697278,0.1839559525,-0.0913669392,-0.5437738299,0.248123154,0.3065992594,-0.283551544,-0.220449239,0.0445863865,-0.1765464693,-0.169092983,0.280954212,0.0601843521,-0.4300897717,-0.0871656835,0.3611609638,0.692802012,0.0956327841,0.6303321719,-0.1235147119,0.076324746,-0.2532035112,0.2561169863,0.0495282076,-0.3795239031,0.023546584,-0.2158819884,0.5026550889,0.1490798593,0.2425636649,0.0512004346,0.2165057659,-0.1872843504,-0.2298112512,-0.1804947704,0.1434120089,0.3687047362,0.4434487522,-0.0461830422,0.1900094301,0.0636095852,-0.1359469742,-0.1933095306,0.3780862987,-0.0019007497,0.2200290412,-0.1212748587,-0.0404932871,0.0924244598,-0.0071384758,0.0839809254,0.1706783921,0.1007460952,-0.118489258,-0.0947092995,0.2013962865,0.0202111173,-0.010909115,-0.4141482413,-0.254029125,0.1932858229,0.1224947497,-0.1357152015,0.3085146248,0.2210880369,0.058151789,0.0265158545,-0.1402291954,-0.2675738633,-0.1897925138,-0.2437104732,0.2894000113,0.2869648039,-0.0705280229,0.1887313575,-0.0016365205,-0.3734047115,0.076104857,0.2457874566,-0.1281601936,0.0467050076,-0.4984141588,0.1470416486,0.0950459912,0.4649443626,0.4928767681,-0.2499477416,-0.0774486065,0.0038336227,0.0757430792,-0.2520690262,-0.1493497789,-0.0446838737,-0.0313462503,-0.1036655754,0.239677161,0.1789068133,-0.3461972475,0.3411019444,-0.1517432332,0.3311131299,-0.2414622754,0.096116446,0.3485895097,0.2501909733,-0.2042519599,0.0667930245,0.1583615243,-0.1566104144,0.4312984943,-0.0198155101,0.2452923805,-0.3857079446,-0.0174342226,0.1345437169,0.0873983353,0.1646578908,0.3448945582,-0.2775501907,0.3103102446,0.1840661764,0.4008575678,0.1085733399,-0.1794613153,-0.3515607417,-0.1591020077,-0.1669953316,0.0303800814,-0.2340048105,-0.1190362498,-0.0091166245,0.0484344102,-0.0464260057,-0.176853627,0.2419453561,0.1843149811,-0.0496766903,-0.7318539023,-0.1252370328,0.1987216622,-0.051255215,-0.206410259,0.2810963094,0.3177622259,0.0713610575,0.2980912626,0.6033411026,0.3172428906,0.0470911302,-0.0055156155,0.0216098465,-0.3519901037,-0.0397840254,0.0464709029,0.4243768752,-0.0998865813,-0.176900059,0.1051216498,0.0322888009,0.0886779875,0.1579322219,0.3321921527,-0.0405231416,-0.5797713399,0.3257236183,-0.0545559041,-0.025573926,-0.1996345967,-0.0380653441,-0.3977474272,0.1864604503,0.684025526,0.0529942438,0.1020229831,0.0346850082,0.0461436734,-0.0434360541,0.1858994067,0.2662366927,-0.0200973675,-0.2706245184,-0.039699249,-0.6192758679,0.3042189777,-0.3095858395,0.1548147202,0.0984999388,-0.036193233,0.048557274,0.2920229137,0.2161684781,0.073474355,-0.7003994584,0.317879349,-0.4329651296,-0.2622369528,0.0957976952,0.1801714003,-0.1712171286,-0.1396931261,0.2415354252,-0.1758356094,-0.0587715246,-0.2171988636,0.049713362,0.0680280924,0.3595534265,-0.117767632,0.0166701414,0.3342880011,0.1487689018,-0.0173841342,0.1082159132,-0.3788925111,-0.1546898484,0.6345922351,0.1702946275,-0.1933920979,-0.1580063701,-0.1129394546,0.1664950848,0.0150336837,0.3186981678,0.1345500797,-0.3001073897,-0.187917307,-0.1406303495,0.1389978826,0.0711054131,0.6141666174,0.2124284655,0.2912798226,-0.1787165552,-0.4764049649,0.4350118339,-0.6599905491,-0.1407947093,-0.0828272551,0.0394951142,0.1680328995,0.0906006768,-0.5330011845,-0.1055724025,0.7180587649,-0.0783677101,-0.0672939345,0.3506139815,0.1317193359,0.0015961054,-0.000466497,0.5324160457,0.1221246943,-0.2277929932,0.4219360054,-0.1893929988]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"An educated guess: does this refer to the fact that depending on the custom column names in the dataset files (csv in this case), there is a dataset loader being created? and this dataset loader - using the \"custom data configuration\" is used among all jobs running using this particular csv files? (thinking out loud here...)\r\n\r\nIf this is the case, it may be ok for my use case (have to think about it more), still a bit surprising given that datasets caching is disabled (or so I hope) by the lines I pasted above. ","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":95,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nAn educated guess: does this refer to the fact that depending on the custom column names in the dataset files (csv in this case), there is a dataset loader being created? and this dataset loader - using the \"custom data configuration\" is used among all jobs running using this particular csv files? (thinking out loud here...)\r\n\r\nIf this is the case, it may be ok for my use case (have to think about it more), still a bit surprising given that datasets caching is disabled (or so I hope) by the lines I pasted above. ","embeddings":[-0.0747397617,0.0329241566,-0.1112524569,0.1840382367,0.2081030607,0.1995864511,0.3412374258,0.0353642963,0.3006587923,-0.1421154141,0.170811981,0.0910153612,-0.2320550829,0.0387957729,-0.0910931453,0.4722394943,0.0865925699,0.0711622313,-0.1339131147,0.0554632209,-0.0745019615,-0.0144607704,-0.1561083049,0.0823811889,-0.483264178,-0.0868270323,0.1125857383,-0.024432132,0.0212619007,-0.505325973,0.4776338041,0.2666884959,0.1665527225,0.3351033628,-0.0001139642,0.0305280704,0.171191588,-0.0332141183,-0.2615791559,-0.0518321469,-0.5447487235,-0.4565433264,0.2222072333,-0.1231800094,0.0039046986,-0.0361423828,0.0507738255,-0.6826250553,0.4117220044,0.2966955602,0.238175109,-0.1285123676,-0.4286361337,0.0785945728,0.01174457,0.1088230461,-0.0908934623,-0.0873219147,0.1938952953,-0.1112779677,-0.0507549793,0.3164396286,-0.0971551985,0.3233141899,0.5042588711,0.0171202812,-0.2412169576,-0.2511727512,0.3183849752,0.0218120925,0.9332623482,-0.3090761602,-0.1231772974,-0.4111324251,-0.0318885073,-0.2988231182,0.2127162963,-0.0283942204,0.0482818186,0.1094920039,-0.4467254579,-0.2396516651,-0.0837241411,-0.2604036629,-0.2173207551,-0.1060871184,-0.3423906267,0.1185139641,-0.0954381078,-0.0773529485,0.4783639312,-0.4291200638,0.0521481335,0.1758250296,-0.4903399944,-0.039084278,0.1224549711,0.3279424608,-0.1147692502,0.1350920796,0.1050546095,0.0122758495,-0.0907438695,0.0799774453,0.4582875967,0.2901148796,0.0588694811,0.1618072242,0.1719782352,-0.160741657,-0.241159752,-0.1606788486,0.1068607494,-0.413952738,0.518558681,0.0468939357,0.1974547058,-0.2680057287,-0.3001054823,0.1110601425,0.0948532373,-0.1812060475,0.0791819766,0.1522895098,0.0211147256,0.0074778986,-0.1893983632,0.0162826721,-0.1804979295,-0.1794106215,-0.1986326873,-0.1786896884,-0.2716118991,0.0370225273,0.1595566422,-0.1190396696,0.0555740073,0.3713513315,0.1261793673,-0.2491984069,0.1370883286,-0.0088045169,0.1465437859,0.2985467911,-0.2595076263,0.2414861768,0.3928395212,-0.0316358693,-0.2273064703,0.2966907322,-0.4408548176,-0.5244545937,0.0838082433,0.1510600746,-0.5156723261,-0.174303174,-0.1659304649,-0.0480842665,0.4366083145,0.0913138241,0.1581714302,-0.221163705,-0.1835799515,-0.2487584054,-0.0916242599,0.4071259201,-0.6090791225,0.0580475777,-0.2302497029,-0.0686956272,0.1020542011,0.2040497959,-0.4850606322,0.1768309772,-0.1797080636,-0.2035737634,0.0928329453,-0.1709412187,-0.2905771732,0.2443111539,0.1496290416,0.4461428523,0.1504223347,0.0730721578,0.098458387,0.0018103365,-0.020994097,-0.0479279384,0.1157689691,-0.1945623904,-0.054495886,-0.2685349584,0.2093880475,-0.1184311211,0.0805206001,0.1271192729,-0.0038219944,-0.1973346323,0.0553921089,-0.0919834897,0.0335476026,0.2680851817,-0.0206702296,0.2839364707,0.1013465747,0.1371282637,-0.5190863609,0.3768135607,0.1972192526,-0.3630043864,-0.1326486319,-0.2342196554,-0.1636397392,-0.0399625488,-0.4563782513,-0.1838131249,0.0451876856,0.0128975371,-0.0490491129,-0.1572533846,-0.230217278,0.4298688769,-0.3645770252,-0.1283554733,-0.1698278934,0.1035878733,0.0485141464,0.3051691949,-0.2094290257,0.1295059621,0.1013689786,0.119615294,0.0973193347,0.1836394817,0.1609486639,0.4847697914,0.0469985195,0.3916104734,-0.0876893848,0.0587859266,0.3191917539,-0.0920874327,0.1048939899,-0.1198229045,-0.1035623699,0.4611107409,-0.1929136962,0.2234096676,-0.231130749,-0.2143968493,0.1486461163,-0.0641654804,-0.0321682282,-0.1272863746,-0.0385257341,0.0578000844,0.0152750537,0.3464541137,-0.3043401241,0.1205620766,0.2261316627,0.0875343531,-0.0441640876,-0.017838316,-0.0513337664,-0.2315881252,0.1401090026,0.4445009828,0.5561187863,0.0778593346,0.0189793948,0.05105865,0.087690331,-0.1661075503,0.2612457275,-0.1427205205,0.0464981422,0.1394626796,0.0001477332,-0.1083628237,-0.2499698997,0.0795986578,0.1397586018,0.1755785197,-0.5349952579,0.2427573055,-0.4415971935,-0.1433641464,-0.324346453,0.1396015882,-0.2296126336,-0.3920868337,-0.0558286458,0.0654042587,0.1098907217,0.0578589179,-0.2731127143,0.405615896,-0.1743054092,-0.0612870902,-0.067523621,-0.162519455,-0.1320426911,0.0824302956,0.1212183237,-0.1581968665,0.2031899542,0.0007400502,0.0459611975,-0.2996458113,0.1214200854,-0.051458206,0.0274218526,0.2571710646,-0.0936723799,0.0155690722,-0.0305417627,-0.0444958881,-0.0695528686,0.1485839933,0.1438077092,-0.159375295,-0.0334503315,0.0861432403,0.0265934449,-0.193444103,-0.3771933019,-0.1723975837,-0.1766427755,0.0177744422,-0.0237497091,0.4363087416,-0.0280570723,-0.0785347149,0.1182727218,0.0119348159,-0.5252438188,-0.5456594229,0.2012652755,-0.079322502,0.0201251339,0.1922750175,0.1350903511,0.3536137938,0.011069349,-0.6578817368,0.0101154717,-0.136392951,0.086420849,0.1211572513,-0.0074003763,0.1934141368,0.0741366372,-0.0120978104,-0.1191144586,-0.1167680994,-0.1516049504,-0.1109017357,0.3235321939,0.1593184918,-0.0324482247,0.036751993,0.9366265535,0.3841562271,-0.121468313,0.1396234781,-0.108836323,0.622118175,0.0644044429,-0.2527451515,-0.1152320057,-0.2590205669,-0.3898253739,0.0313536115,-0.0384908281,0.0496639051,0.1796972454,0.3303647935,-0.051457271,-0.2739363015,0.2169685066,-0.4276330471,0.3161676228,0.1154605374,-0.1289369911,-0.2765468657,-0.0896553025,-0.3233951628,0.2079395652,0.2542685568,-0.0912257358,-0.1865716726,-0.1520025879,-0.2801687419,0.2399471402,-0.1212937534,0.1672365814,-0.1665854752,0.1557669938,0.2747409344,0.2953509986,0.3989815712,-0.5094984174,-0.0547716469,0.2473613918,-0.0301233884,-0.0246781446,-0.2044211328,0.1053705364,0.3304403722,-0.0029191168,0.1057159528,0.2741363049,-0.1856383234,-0.1463248134,0.1368196309,-0.2211832106,-0.2176621407,-0.2295271009,-0.0647614673,-0.1236507371,-0.0071743256,-0.0346280523,-0.1003785953,-0.2729247212,0.1329510659,-0.1421343237,0.1635401398,0.2158608288,0.0707106814,0.1651082188,0.165827617,0.0806579962,0.3721834421,-0.0893875808,0.3206748366,0.5537675023,-0.159621194,-0.1702971607,-0.0921291709,0.0285201687,0.2239181697,0.2774191201,-0.3186157644,0.2311957181,0.0150815025,-0.1968954206,-0.4145432413,-0.247533977,0.3081844449,0.2830772102,-0.2407494485,-0.6778427362,0.4185904264,0.2034242898,-0.2414294928,0.0114896726,-0.1155929789,-0.3438018858,-0.0826011077,0.0311502051,0.6868385077,0.0536781922,0.1511935443,-0.08733318,-0.0505860299,0.4624560773,-0.2295815796,0.217359215,-0.1653167009,-0.0813855305,-0.1341671795,-0.1254744977,0.3942568004,0.1997703612,-0.0706662312,0.3692277372,-0.3337557912,0.4619781971,-0.1979233027,-0.0477025136,-0.1497226954,0.0053808251,0.0632505789,0.1764404178,-0.0409989581,0.378221482,-0.1098464653,0.2351148725,0.2955272794,0.158552587,-0.3830777407,-0.2306856364,0.0952183753,0.2589576542,-0.1065820754,-0.3580361307,0.4108014703,0.393088311,0.2602231801,-0.0794414952,-0.1240395904,0.2739798129,0.2207245529,0.226437524,-0.1794320345,-0.034283895,0.2966901362,0.1402222216,-0.3095149696,0.1744709164,-0.0203972254,-0.2564274371,-0.0291427393,0.4295302927,0.0357082225,-0.6212303042,-0.1937509328,0.286085844,0.2380784601,0.140977636,0.1657323837,0.0975857899,0.1726160347,0.5359575748,-0.4709506929,-0.3526124358,-0.0289779603,0.2171926498,0.5373673439,-0.1034850031,0.4559842944,-0.2134130001,-0.2968732715,-0.2070980817,0.0560980588,0.1261419356,-0.0860368684,0.0951053724,0.121282883,0.3870361149,0.1258447915,0.0527884625,0.0906804651,0.0326523371,-0.1889146864,-0.1910330504,-0.0908899531,-0.0370000564,-0.0190094225,0.4248803258,-0.1085397899,-0.1448487937,0.0165486783,0.0053797495,-0.2818008065,0.2052350044,-0.1517514884,0.1312711686,0.1264013052,0.0177501347,0.1077294499,-0.2185231745,-0.0413200818,-0.1166727245,-0.3801874816,-0.1926355213,0.0644762963,0.142687723,-0.0326852985,-0.1700797528,0.0117397802,-0.2476841956,0.0291255098,-0.1259061247,0.2186978906,0.363579303,0.0053541292,0.2291691005,-0.2157220393,-0.1443708837,-0.0980219096,-0.0127811143,0.0226474926,0.408683449,0.3274835944,0.3122511804,-0.112781927,0.1433492899,-0.2429715544,0.2534457743,0.0494389087,0.1417547315,0.1378201544,-0.3263118863,0.1708707362,0.21786654,0.3389469683,0.2562046349,-0.3132695556,-0.1951802373,0.1058207378,0.1471500695,-0.1450524181,-0.1846280694,0.1522567123,0.1968634278,-0.1703169644,0.2726588547,0.5628316998,-0.0311788991,0.4801337123,-0.0712575242,0.6081089973,-0.464114368,0.4534156621,0.128799364,0.0151933311,-0.1285957545,0.4395719767,0.5133951306,0.037492983,0.5916284323,0.0516391136,0.1914888322,0.4715460837,0.1481672227,-0.3705318272,-0.8666630387,0.4774235189,0.4776128829,-0.5016353726,0.1033874974,0.0431574397,0.0473804362,-0.1856457889,0.2252573669,-0.1247980297,0.1617403924,-0.2156005949,-0.1505635977,0.4182736874,-0.2759999335,0.1143250316,0.1305047274,-0.0543568395,-0.0419634432,0.1496677101,-0.2406262457,-0.1486199647,-0.3659248054,0.4533606172,-0.0087456042,0.3205533326,-0.1618720591,0.1212193072,-0.086369358,0.1227119863,0.3090470433,0.309104085,0.5185569525,0.1061776131,0.2864176929,0.0258777756,-0.0574803129,0.0429528281,-0.2092999965,0.219723016,-0.1978299618,-0.0892031714,0.1855271012,0.1404807419,0.0371158868,0.1310311258,0.1015050039,0.2059725672,-0.4896321297,-0.1068909094,0.1763571948,0.0596949682,-0.011909862,0.2690345347,-0.266155839,-0.2663977146,0.7305529714,0.0587829053,0.3269719481,-0.1290088147,0.0735100359,-0.133599773,0.3655230105,0.1005996019,0.0710482299,-0.2286351174,-0.0154606206,-0.6817538738,-0.0881665871,-0.313960433,0.144767344,0.1171007529,0.111854367,-0.1319794357,-0.0138003156,0.0433720164,-0.2220948637,-0.1332458705,-0.1272137463,-0.0476978682,-0.1649046689,-0.0732101426,-0.0108080376,-0.1269413978,-0.2174556106,0.1390997022,-0.0047837943,0.003728277,-0.1258691847,0.0488723144,0.1192159727,0.0717008784,0.4264052808,0.0970844626,0.3955483139,0.303019166,0.1015703231,-0.2562287748,-0.2654801607,-0.1641040295,0.2406622022,0.1155837178,-0.1395573914,-0.542992115,0.1047074497,-0.2227845341,0.0548640825,-0.0120233577,-0.1266849488,0.0643173456,-0.1498863846,-0.0647278801,0.2170044631,-0.0487025827,0.2689602375,0.0162306558,0.3161885738,-0.2850948274,-0.0212373734,0.309437871,-0.1760539412,-0.2819661498,-0.0532595962,0.1987364292,-0.0369926393,-0.1815896332,-0.5344362259,0.0498642288,0.5088617206,-0.283326745,-0.0162014589,0.0728510246,0.0203093551,0.0472291037,-0.0840392634,0.2260671407,-0.0358380489,-0.0223107133,0.0305849668,-0.3441295922]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"Hi ! Currently disabling the caching means that all the dataset transform like `map`, `filter` etc. ignore the cache: it doesn't write nor read processed cache files.\r\nHowever `load_dataset` reuses datasets that have already been prepared: it does reload prepared dataset files.\r\n\r\nIndeed from the documentation:\r\n> datasets.set_caching_enabled(boolean: bool)\r\n\r\n> When applying transforms on a dataset, the data are stored in cache files. The caching mechanism allows to reload an existing cache file if it\u2019s already been computed.\r\n> Reloading a dataset is possible since the cache files are named using the dataset fingerprint, which is updated after each transform.\r\n> If disabled, the library will no longer reload cached datasets files when applying transforms to the datasets. More precisely, if the caching is disabled:\r\n> - cache files are always recreated\r\n> - cache files are written to a temporary directory that is deleted when session closes\r\n> - cache files are named using a random hash instead of the dataset fingerprint - use datasets.Dataset.save_to_disk() to save a transformed dataset or it will be deleted when session closes\r\n> - caching doesn\u2019t affect datasets.load_dataset(). If you want to regenerate a dataset from scratch you should use the download_mode parameter in datasets.load_dataset().","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":202,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nHi ! Currently disabling the caching means that all the dataset transform like `map`, `filter` etc. ignore the cache: it doesn't write nor read processed cache files.\r\nHowever `load_dataset` reuses datasets that have already been prepared: it does reload prepared dataset files.\r\n\r\nIndeed from the documentation:\r\n> datasets.set_caching_enabled(boolean: bool)\r\n\r\n> When applying transforms on a dataset, the data are stored in cache files. The caching mechanism allows to reload an existing cache file if it\u2019s already been computed.\r\n> Reloading a dataset is possible since the cache files are named using the dataset fingerprint, which is updated after each transform.\r\n> If disabled, the library will no longer reload cached datasets files when applying transforms to the datasets. More precisely, if the caching is disabled:\r\n> - cache files are always recreated\r\n> - cache files are written to a temporary directory that is deleted when session closes\r\n> - cache files are named using a random hash instead of the dataset fingerprint - use datasets.Dataset.save_to_disk() to save a transformed dataset or it will be deleted when session closes\r\n> - caching doesn\u2019t affect datasets.load_dataset(). If you want to regenerate a dataset from scratch you should use the download_mode parameter in datasets.load_dataset().","embeddings":[-0.0997084454,0.0613426492,-0.0944522992,0.1375739574,0.2868357897,0.1569391489,0.2946323752,0.0576292314,0.2189560533,-0.2274291068,0.0563382395,0.1891585737,-0.1538448781,-0.1349174827,-0.0155598177,0.4577814043,0.1715121865,0.0599948317,-0.1155785844,0.0047335713,-0.0356277116,-0.0026572377,-0.2127188146,0.0402786806,-0.4570401311,-0.1324051023,0.1229298636,-0.0716498047,-0.0130503355,-0.5536743402,0.4886825979,0.3283195496,0.0741478354,0.2518526912,-0.0001118015,0.0727999434,0.2476903051,-0.0178316571,-0.2822506726,-0.003201406,-0.5275136232,-0.3997315764,0.2077168673,-0.1217703149,0.0088245841,-0.2282435447,0.0214830358,-0.7317454815,0.3986606002,0.2624377608,0.2537710965,-0.1417690814,-0.3225730658,0.0698815659,0.0294989012,0.1056568027,-0.0790080056,-0.1197099462,0.1755229682,0.00115097,-0.1297950745,0.3401147127,-0.1820656359,0.2184822261,0.5376929641,-0.0251180753,-0.3009690344,-0.3138095438,0.3533349931,-0.0475770123,0.8824431896,-0.3004612327,-0.2132212669,-0.3719502687,-0.1303938329,-0.107222043,0.2515766323,-0.0626551956,0.1068924516,0.1626930237,-0.5890986919,-0.3423182666,-0.0354443565,-0.1216804236,-0.1883628964,-0.0987214446,-0.3428514898,0.1247541532,-0.0124882208,0.0298062265,0.5160682201,-0.4871827662,-0.0295811631,0.2198566943,-0.4583883584,0.0142551577,0.1422318816,0.4166750908,-0.0496474281,0.220106706,0.1075856835,0.0910666659,-0.1975068301,0.0559330285,0.3907082677,0.382031858,0.0675789565,0.1710542142,0.0845614672,-0.2150832415,-0.1033617258,-0.1998728067,0.1032040641,-0.3857403398,0.5634826422,-0.0266759973,0.2618582845,-0.294590503,-0.3204870224,0.0342997648,0.0142830722,-0.2044890523,0.1050424129,0.1328129321,-0.0418563187,0.1020384952,-0.1786978096,-0.0918801203,-0.1994803548,-0.2397157699,-0.2297836989,-0.1517885029,-0.3549550474,0.0677673593,0.2348688841,-0.3219405413,0.0271079615,0.3597248495,0.0614182539,-0.2675002515,0.1516847908,-0.0396935157,0.3635239005,0.3102005124,-0.3273587525,0.2836120129,0.3906999528,-0.0564260893,-0.1284382939,0.382796973,-0.4262167811,-0.3516856134,0.1925227195,0.1590859294,-0.4444234371,-0.1102425605,-0.2061909288,0.0902894586,0.4188187122,-0.0466482677,0.1753661036,-0.1439128071,-0.130047664,-0.2852768302,0.1402714401,0.4280160069,-0.5497478247,-0.0674493983,-0.2498912662,-0.0562166758,0.1099889353,0.2396276444,-0.4815126359,0.2011151314,-0.2263178527,-0.1987856477,0.1428757459,-0.121588923,-0.4337772429,0.1803838015,0.2499530017,0.4296301901,0.0207902826,0.078282021,0.0647751093,-0.0365483053,-0.1067260504,-0.1406600326,0.061826434,-0.245730862,-0.1365717649,-0.2602417171,0.3213842511,-0.1354840398,0.1037089154,0.2238005251,0.0367317945,0.0181506891,0.0311799441,-0.0368100591,0.1045531631,0.3294175267,-0.0188392252,0.103748925,0.1263507605,0.080727458,-0.4650685489,0.3508911133,0.1209326386,-0.4675593674,-0.14911443,-0.1868676096,-0.1537768841,-0.121195741,-0.4467225373,-0.1915006787,0.0879468173,-0.0069218948,-0.0161998272,-0.0809430704,-0.2429459542,0.4812088311,-0.2107368857,-0.1352239698,-0.2489376366,0.141337797,0.0420287326,0.2810896337,-0.2694530785,0.054643441,0.1063114479,0.0800701752,0.0314386189,0.2419422418,0.0877580643,0.4027651846,-0.0050890767,0.3685146272,0.0253670849,-0.1032399833,0.2767234445,-0.0941211581,0.1489203721,-0.0821669027,-0.1542397439,0.2779749334,-0.2706847191,0.2121043652,-0.2125898302,-0.2112362087,0.124009423,-0.0346608013,-0.0884961858,-0.1701375246,-0.0698364154,0.0679313987,0.1711432636,0.3167246878,-0.0430301689,0.0704462975,0.3305062056,0.0603819676,-0.1028052941,0.0156823378,-0.1570168436,-0.3179258704,0.0985313952,0.4447603226,0.4994432926,0.1275836676,0.066603303,-0.0098827332,0.0787229836,-0.1767649353,0.249533385,-0.0649069324,0.0468865484,0.1298066676,0.0407715552,-0.0312101245,-0.3414979577,0.11968337,0.2410701066,0.1237346604,-0.4768824279,0.3093511462,-0.446213007,-0.0721967071,-0.139826566,0.1368020177,-0.2670971155,-0.4040007889,-0.1407969892,0.0495771766,0.2279472202,0.121240437,-0.2031304836,0.3428511024,-0.1336136758,-0.022177333,-0.1573786438,-0.1400097758,-0.1562621593,0.0989416838,0.1103478596,-0.3031523824,0.2201781422,-0.0197522976,0.1069621742,-0.3282643855,0.0002519308,0.0028273987,0.0342083722,0.1573209912,-0.2203421295,-0.0943113267,-0.1018785164,0.0736058652,-0.0442598574,0.0304602515,0.1118764803,-0.1155054942,-0.0266527645,0.0439222828,-0.059039779,-0.2237265259,-0.2553488612,-0.1147144139,-0.1969298869,-0.0365621708,-0.1048907787,0.5499766469,-0.0062835664,-0.0027899391,0.0241824035,-0.0475359708,-0.5478751063,-0.4771881998,0.2433176488,-0.0103913406,-0.0927400067,0.1800292134,0.1821801513,0.3579396904,0.209037587,-0.5898662806,0.071619831,-0.1588022709,0.0233939569,0.029422611,-0.0425002202,0.2604376376,0.1229924932,-0.0197214708,-0.1496986896,-0.0818621293,-0.1456715614,-0.1006152257,0.2926935852,-0.014061884,-0.0258151423,0.1166400015,0.8728512526,0.3354340494,-0.174358353,0.1497428715,0.0036005187,0.6802711487,-0.0236994382,-0.2721688449,-0.1048575491,-0.3019877374,-0.2880857289,0.0078901676,-0.0682234541,-0.0739514679,0.1186829805,0.2826231718,-0.0740708262,-0.254660368,0.1120001674,-0.470066309,0.3354950547,0.1904261112,-0.0317074768,-0.2627396882,-0.0764461234,-0.2539058924,0.1647631675,0.4068376422,-0.112213701,-0.1564539224,-0.0322538875,-0.2148509622,0.2287481129,-0.1043402329,0.1942424625,-0.1866880953,0.1193086132,0.2706303596,0.2362115234,0.441640377,-0.3354632258,-0.1284669787,0.2465899736,0.0022430148,-0.1495687068,-0.1514665782,0.0276498832,0.3585686982,-0.074028559,0.0572513193,0.2113767117,-0.2003385127,-0.3200255334,0.169825688,-0.1995245665,-0.1700764447,-0.2334340513,-0.0489894859,-0.2833401263,-0.0769561976,-0.0371066928,-0.1540483087,-0.2317907214,0.0298896786,-0.1639719158,0.1110302433,0.1563223004,0.0292974133,0.3012327552,0.2159175426,0.0202350374,0.39444682,-0.0352937505,0.3171941936,0.4560652077,-0.0167690553,-0.1099895239,-0.024350293,0.0099987201,0.0677469894,0.1511399597,-0.4170918465,0.1462960392,0.0736841708,-0.1210204512,-0.387739867,-0.0547519177,0.2709854841,0.2128453255,-0.3550459743,-0.6746981144,0.4435946643,0.1464603841,-0.2401805967,0.1383092701,-0.0883962512,-0.2996916175,-0.0610640571,0.1230763346,0.6396497488,-0.0497897342,0.0999635756,-0.1400749385,0.0719694644,0.50225389,-0.3129159212,0.3041851223,-0.0861971527,-0.1927681416,-0.1455082595,-0.1197860166,0.3632362485,0.2697044909,-0.1359048635,0.3079724014,-0.3161233068,0.3013362586,-0.1505202204,-0.0549886413,-0.0285897348,-0.0406887531,0.1442474425,0.2281893492,-0.0997174308,0.3356190324,-0.0689985976,0.2048369795,0.2493768781,0.1936885864,-0.3374401629,-0.1257372648,0.0690717176,0.3066958487,-0.0890401378,-0.2670940161,0.4590046704,0.350666523,0.3423378468,-0.0457615443,-0.1775251627,0.3148001432,0.1381987631,0.2338989228,-0.1689061671,-0.0510084145,0.3056409955,0.1594471633,-0.2640140355,0.1524901688,-0.0882909372,-0.224095583,-0.0509482324,0.4085111618,0.0640561953,-0.5559608936,-0.0908329263,0.3273923397,0.2179462612,0.1672279835,0.1639008671,0.1661961228,0.2095592916,0.4357341528,-0.3501187861,-0.3544806838,-0.0227383878,0.2921639681,0.4367716908,-0.1963724941,0.4934601188,-0.2784533203,-0.2668257058,-0.1840011626,-0.0593237504,0.0768895894,-0.1987026185,0.032051567,0.0503813997,0.5129694343,0.1112082303,0.0021714908,-0.0117472233,0.0196253974,-0.218228668,-0.0911413655,-0.1519612074,-0.025247531,-0.0733809918,0.3485889733,-0.0597647466,-0.1408622563,0.0822803602,0.0742959008,-0.3046661317,0.1040654108,-0.0914507806,0.1877194345,0.0248712301,-0.0595068336,0.1020690948,-0.3310244381,-0.0254216529,-0.1343944222,-0.2966209352,-0.2075813562,0.1033255458,0.1513067633,0.0808229148,-0.2305877507,-0.1054556891,-0.178361401,-0.0189114083,-0.1884192675,0.2050105333,0.3741443157,0.0511078686,0.2064264566,-0.1330082268,-0.1382596493,-0.1291483492,-0.1326714754,-0.0152387237,0.2860151231,0.3478200436,0.2652789056,0.0007204367,0.1014991105,-0.3858973384,0.3128343523,0.1614358276,0.1342080384,0.1904036254,-0.3635286093,0.2479934841,0.1711934656,0.2598030567,0.1603669822,-0.2352412492,-0.2446569204,0.047794316,0.173289001,-0.1072360873,-0.1930183768,0.1676573902,0.2809453905,-0.0967925116,0.2038504034,0.4800886214,-0.0323564745,0.5502661467,-0.0269420575,0.6699725986,-0.4500140548,0.4508243203,0.1529663354,0.1055638045,-0.0680244043,0.4651838541,0.4063000977,-0.02370595,0.6273493171,0.0144168343,0.2521226406,0.4498822689,0.2904343009,-0.3405666053,-0.819208324,0.3104337454,0.398214519,-0.4488942623,0.1168086901,0.0262725595,-0.0100036366,-0.0453770682,0.0183689967,-0.0888939202,0.2448589504,-0.1834665835,-0.1232040748,0.3968523443,-0.2634819448,0.0694973692,0.0968418047,-0.0434626602,-0.0514188036,0.164625138,-0.158807084,-0.2100954205,-0.31027475,0.4182795584,0.0607177615,0.3171711564,-0.0884847715,0.2217873335,-0.0683781952,0.0083076693,0.2621364892,0.3208215237,0.3727636039,0.0957438871,0.246052146,0.1270030886,-0.0210857857,0.0010045533,-0.1751105487,0.310511142,-0.1886152029,-0.0904983133,0.0898072273,0.1573914737,0.0561004691,0.1040536612,0.001008379,0.2535036802,-0.4775228202,-0.0056232768,0.167027548,0.0225883927,-0.0631172881,0.2347108126,-0.3002535105,-0.2069145143,0.7661941648,-0.0268255994,0.2581926882,-0.2106379718,0.0933529958,-0.0331006758,0.3607287407,0.1303011328,0.1006149203,-0.1831120998,-0.0573928282,-0.6422398686,-0.0962606445,-0.1662902832,0.2937772274,0.1058248132,0.0946621001,-0.0822318047,0.0579708293,0.1092484146,-0.1807011068,-0.0588789843,-0.2368717492,-0.0097362166,-0.0535416827,-0.1769293249,0.0126564046,-0.209571138,-0.1916644424,0.1834266037,-0.0666880235,0.0385413989,-0.0850876719,0.0116733173,0.0822245181,0.1533953547,0.448131144,0.2328588367,0.2836711407,0.2613926232,0.1431041956,-0.3198749721,-0.2190731317,-0.1147533879,0.2478031367,0.0711662546,-0.1148603186,-0.4961168766,0.0229275506,-0.2213889807,0.1173789576,0.0895465687,-0.016669618,-0.0308832619,-0.212569803,-0.0893308297,0.1652230471,0.0440973602,0.2976623774,0.0289042443,0.3217282593,-0.3588088453,-0.0783866569,0.2933993042,-0.3242954612,-0.2258917242,0.060728956,0.1499200463,0.0497913957,-0.2368643433,-0.5234311819,-0.0139792291,0.584651053,-0.3617408276,-0.0685681999,0.0632003248,-0.096434325,0.093758218,-0.1103020161,0.2539664209,0.0622684285,0.0009359301,0.039420519,-0.2846994698]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"Thank you for the clarification. \r\n\r\nThis is a bit confusing. On one hand, it says that cache files are always recreated and written to a temporary directory that is removed; on the other hand the last bullet point makes me think that since the default according to the docs for `download_mode (Optional datasets.GenerateMode) \u2013 select the download\/generate mode - Default to REUSE_DATASET_IF_EXISTS` => it almost sounds that it could reload prepared dataset files. Where are these files stored? I guess not in the temporary directory that is removed... \r\n\r\nI find this type of api design error-prone. When I see as a programmer `datasets.set_caching_enabled(False)` I expect no reuse of anything in the cache. ","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":112,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nThank you for the clarification. \r\n\r\nThis is a bit confusing. On one hand, it says that cache files are always recreated and written to a temporary directory that is removed; on the other hand the last bullet point makes me think that since the default according to the docs for `download_mode (Optional datasets.GenerateMode) \u2013 select the download\/generate mode - Default to REUSE_DATASET_IF_EXISTS` => it almost sounds that it could reload prepared dataset files. Where are these files stored? I guess not in the temporary directory that is removed... \r\n\r\nI find this type of api design error-prone. When I see as a programmer `datasets.set_caching_enabled(False)` I expect no reuse of anything in the cache. ","embeddings":[0.0234997645,-0.0067316932,-0.1006589457,0.2295337319,0.2193898708,0.164606899,0.1426831633,-0.1221596375,0.2681152225,-0.2213968635,0.1463908553,0.0476379991,-0.0811795667,-0.0989200249,0.0419844203,0.4397714734,0.1416729987,0.1220702529,-0.1241388991,-0.024245074,-0.1132137179,-0.0720723197,-0.2296927422,-0.0093612997,-0.3864178956,-0.1857894808,-0.013625782,-0.0095074028,-0.1264188141,-0.6089860797,0.5007253289,0.3236505389,0.1929257661,0.199494645,-0.0001181951,0.0677730143,0.2361237705,-0.1209783182,-0.2735072374,-0.0727832243,-0.6044927835,-0.2693714797,0.2145034671,-0.1004407555,0.1328168362,-0.3968074322,0.148782894,-0.6915858984,0.4652371705,0.3504002988,0.1721186787,-0.2823828757,-0.4284880459,0.0757464468,-0.1182394847,0.1931481808,-0.0309059098,-0.1335907429,0.1799318194,0.085323751,0.0445982106,0.1798013896,-0.0702087879,0.1382529289,0.6103155613,0.0010878311,-0.3598736227,-0.4836100638,0.3094478846,-0.037922617,1.1107507944,-0.2843290269,-0.2158844173,-0.3374539614,-0.1147955805,-0.1629478633,0.2030141205,0.0141905863,0.0817528069,0.2514043152,-0.4645248353,-0.4127131104,-0.1039394662,-0.1423556656,-0.1313592643,-0.1855160594,-0.3275108635,0.0561177395,-0.2420201749,0.0381601825,0.4648624063,-0.5910779834,0.0029419353,0.3164913058,-0.2159607857,-0.0329931527,0.1887451857,0.3879157007,0.0005817513,0.3070461154,0.2229655087,0.0402304269,-0.2444278896,0.0954841971,0.4019175768,0.3555066586,0.217130065,0.0207256023,0.1586512774,-0.1143525541,-0.067549713,-0.2549190223,0.0378968045,-0.1489846706,0.614471972,0.0111030219,0.2402537614,-0.2406780273,-0.2174336463,0.1185985059,-0.0145712188,-0.0895531178,0.0217408594,0.1052210703,-0.0551350154,0.0611126386,-0.199483946,0.0208433475,-0.1996301115,-0.2355587631,-0.1645979881,-0.1472320855,-0.2722468972,0.0655939877,0.2371266335,-0.1996525228,0.2041181326,0.3499111831,0.0142246773,-0.2714629173,0.3167570829,0.0534754209,0.3164253831,0.3470377922,-0.1522307396,0.2555709481,0.2580889463,-0.026440518,-0.1713908911,0.4140850008,-0.4340630472,-0.4625374973,0.206080988,0.1405343562,-0.4565962255,-0.0531906709,-0.1370828152,-0.0015279182,0.4295649529,0.0103738168,0.1819979399,-0.0979933366,-0.0836929306,-0.3451650441,0.0018541116,0.5554590225,-0.6028236151,0.0171681494,-0.1815545112,-0.1036814004,-0.0311860926,0.209923774,-0.3868829012,0.2438092381,-0.2664463818,-0.1506946385,0.2950621545,-0.2705557048,-0.3486518562,-0.0232020412,0.3061289191,0.3065513968,0.0396657102,0.1131661087,0.170281291,-0.0200742036,-0.0307632834,-0.045878578,0.0205203015,-0.3010168672,-0.1418018192,-0.2315362245,0.102464579,-0.2727431059,0.0124846771,0.1822210848,0.026256429,-0.130549103,0.0150086768,0.0417236872,0.065758951,0.2498109788,0.0437640697,0.1365919858,0.0974456966,0.0889289752,-0.5970900655,0.3292792439,0.0235638749,-0.4680375755,-0.0944330841,-0.2102169096,-0.145463109,-0.1103564501,-0.4886603057,-0.1875457615,0.0347709917,-0.001873165,0.1523391008,-0.0182379112,-0.2908667922,0.3823859394,-0.2144011408,-0.044017043,-0.2194145322,0.1462047696,0.0243482534,0.2098653615,-0.3131634891,0.1022916138,0.1038090438,-0.0374070853,0.0518371761,0.2952741086,0.1787336469,0.4258133769,-0.0014279156,0.4329093993,0.1385597587,0.0109969154,0.2834740877,-0.1397245824,0.0873665959,-0.0346955732,-0.1987511665,0.2592175901,-0.2250606865,0.0767403543,-0.0830584839,-0.2420131564,0.0986187756,0.0978424773,-0.0763717741,-0.2709977329,-0.1368174404,0.2496686578,0.1502288431,0.3429254889,-0.0461412594,0.2502219677,0.301848352,-0.0566386543,-0.0451386981,-0.1202748641,-0.1615779251,-0.3995043337,0.1606617421,0.4072237313,0.5993422866,0.0622534417,0.2182065845,-0.0258428548,0.1038823947,-0.1643979102,0.2832359672,-0.0532159172,-0.0447376035,0.1709440202,0.0531280302,-0.0749831498,-0.3602162898,0.1397021264,0.2798999548,0.1886998415,-0.3671920598,0.2529411018,-0.4431505799,-0.1817049384,-0.1705857217,0.1654935181,-0.1021194756,-0.4035171866,-0.0525826439,-0.0047760652,0.1409047395,0.0820238814,-0.1954796016,0.3975166976,-0.1860222369,-0.005379091,-0.1806028038,-0.0211061686,-0.2057397515,0.0298245233,0.12935929,-0.2706010938,0.289229095,-0.076741524,0.103508994,-0.3255177736,0.0774658695,0.0174133573,0.2044340074,0.2359988093,-0.0490939468,0.0545111187,0.1179451719,0.2206740081,-0.1058522761,-0.0016604846,0.060298264,-0.193171531,0.0858834907,0.1474360675,-0.0783885345,-0.2904154658,-0.3152566552,-0.1024320126,-0.2619400918,0.0277815145,-0.0481702909,0.3808300197,0.0195101779,-0.0984848812,-0.0039829542,-0.0292455964,-0.5016860366,-0.570994854,0.2443315834,-0.130773142,-0.004299453,0.2075239122,0.0845579058,0.3798983395,0.1197345182,-0.5160893798,-0.1788179725,-0.0568521693,0.1516496837,0.0886654109,0.0310542062,0.33483693,0.077580303,0.0335270651,-0.1269313991,-0.0595849231,-0.0351212993,-0.0185819566,0.1683895737,0.1260145009,-0.069397755,0.0441119075,0.9790992141,0.4086240232,-0.1154412255,0.1760876179,0.0696031675,0.7738029361,-0.0377138443,-0.0741373077,-0.2078894079,-0.1884602457,-0.3139920831,0.0167483687,0.0886378735,-0.0161532909,0.186817348,0.3473849893,-0.131339103,-0.216522783,0.0515593179,-0.3629049957,0.3482807279,0.2650031745,-0.0842528641,-0.3296344578,-0.0185541194,-0.2946543396,0.1830747873,0.4402164519,0.0292307734,-0.1951730847,-0.1712912321,-0.2636550963,0.1743567884,-0.0061189667,0.2029758096,-0.0061155255,0.2090798914,0.333534956,0.1802455932,0.3525334001,-0.5026122332,-0.182457149,0.1346820593,0.1880277842,0.0194836687,-0.1133231446,0.1127954572,0.2617640495,0.0060895775,-0.0353843495,0.2537998557,-0.233689785,-0.3589023948,0.2316759527,-0.2619051635,-0.0648513213,-0.1865345538,-0.1826255918,-0.1798884571,-0.1747107059,0.0035716051,-0.3195203245,-0.2146860361,0.0176030751,-0.0704080015,0.1649591774,0.036995925,-0.0473923907,0.2551009655,0.0189983733,0.0032832674,0.3809720278,-0.1457580775,0.2899613976,0.5449878573,-0.1722263545,-0.0084875179,-0.0830257237,0.0414407142,-0.0148431882,0.1966684163,-0.3733760715,0.1374657303,-0.0585765354,-0.2101402879,-0.438280046,-0.1286466569,0.2698004246,0.0860324129,-0.1452623904,-0.8525411487,0.4882219732,0.1349603385,-0.3002378643,0.2561308444,-0.0263156816,-0.2982075214,-0.029187182,0.1631516516,0.6705079675,0.0342867002,0.253272295,-0.1095070392,-0.0696867034,0.6333460212,-0.3704907298,0.2600890994,-0.1389879137,-0.0534678884,-0.2236573845,-0.1458281875,0.3658932447,0.159662798,-0.007823064,0.378811866,-0.1656179577,0.394069016,-0.0691816881,-0.074151054,-0.290684104,0.0225861669,0.17847462,0.1118525341,0.1364024282,0.3256077766,-0.2210949212,0.2563667893,0.3178254068,0.1809582114,-0.3007849455,-0.2886204123,-0.0643324256,0.2830480039,-0.2062586993,-0.2831354141,0.29361552,0.4367158711,0.194254756,-0.0588092171,-0.1622582376,0.2445443869,0.1702603847,0.1599704772,-0.1203072071,-0.0548461191,0.2731928825,0.1304356456,-0.2121498883,0.1657131761,-0.1185848191,-0.0801921859,-0.2377203107,0.3219054341,0.1242242232,-0.5290641785,-0.2564133704,0.2405350506,0.1354174614,0.1218912527,0.0849765614,0.2668370306,0.2513145804,0.4207593501,-0.5648021102,-0.1899154931,-0.0505779237,0.2930498123,0.5396540761,-0.221786052,0.4010860026,-0.3772263229,-0.2443060279,-0.1587748528,0.0501658805,-0.0241275411,-0.1426791102,0.0522158444,0.0530313626,0.3743497133,0.0272890702,0.0046014311,0.0209170468,0.0095929736,-0.2336743176,-0.1644587964,-0.0868571997,0.0352885984,0.0572942719,0.4075863063,-0.0326914042,-0.326832056,0.0959439874,-0.1341905147,-0.2389647663,0.1581134498,0.0122103654,0.1036856771,0.0062939455,0.048702091,0.007785215,-0.2578079402,-0.1230218038,-0.136540845,-0.3432871401,-0.1299782395,0.1342009306,0.1895481497,0.0832495168,-0.2429346591,0.026556544,-0.2787335813,-0.0062215654,-0.1361238062,0.222990185,0.2821002007,0.0375672095,0.1426472813,-0.2779729366,-0.0224557817,-0.1634532958,0.044871591,0.0462444983,0.2892729342,0.3116780818,0.3197031915,-0.1628529727,0.0154927084,-0.3095220923,0.376835674,0.0573940352,0.1486513466,0.2582587898,-0.3487161994,0.2262749672,0.1172134653,0.1990614533,0.350258857,-0.285261035,-0.1070061028,0.027433956,0.1120249406,-0.0820120052,-0.1429763138,0.0683031306,0.284656167,-0.1225742549,0.2854655385,0.5204182863,-0.1154556796,0.4649974108,-0.0577532947,0.6504695415,-0.4031918943,0.3757524788,0.119037956,0.068590574,0.0092724366,0.3274647295,0.4481209219,0.1422035098,0.6588886976,0.1218707189,0.3879487216,0.4673338234,0.2016203403,-0.3833975792,-0.8955919147,0.3078376353,0.3818647861,-0.3685449362,0.2639086843,0.1064828485,-0.0312504619,-0.1380837113,0.0902555883,-0.0998127684,0.1968070567,-0.1688738763,-0.1832996309,0.3603134155,-0.3866859972,0.1872093529,0.1168715879,-0.0066795922,-0.0898091197,0.0141190253,-0.2020324022,-0.2183419615,-0.3180062771,0.484737277,0.0347975977,0.2326506972,-0.0617071427,0.2855441868,0.0759228915,-0.0152517427,0.1597117186,0.2611503005,0.3983959258,0.118130371,0.1448617578,0.127605468,0.0456531681,0.0126993256,-0.1188007742,0.2274229825,-0.1583292484,-0.1021404415,0.070680052,0.1190618202,0.1113485917,0.0680559427,0.07853131,0.0670678243,-0.4187906086,0.0865258425,0.1118575931,0.0541385561,0.047094468,0.2602119744,-0.282061547,-0.2008925527,0.5995694399,-0.0185128711,0.1889125407,-0.2611022592,0.0533789545,-0.1386907548,0.206206277,0.2603817582,0.1407028437,-0.1800919473,-0.060456723,-0.6016354561,-0.0857688636,-0.1399373412,0.2087332606,0.0244958941,0.1647185236,-0.1989481449,-0.0473623835,0.0839218199,-0.0454461277,0.0549813807,-0.0095489174,0.0370474048,-0.1825579703,-0.1941013932,-0.0226569548,-0.2907926142,-0.3785873055,0.0232156254,-0.2155515999,-0.0512065105,-0.0924677625,0.0003644662,0.1837017238,0.0372892879,0.5205756426,0.2066607922,0.1702024639,0.3203724623,-0.017144667,-0.2609784603,-0.0301309433,-0.1241852567,0.1429869831,0.0451505259,-0.1089266539,-0.5915424824,0.0786102563,-0.2193136811,-0.00565978,0.1123646051,0.0419399925,-0.0896682739,-0.1870545,-0.2448749393,0.2106907815,0.2306343913,0.2907436788,-0.0177011788,0.303009361,-0.2997810841,0.0832378417,0.2538387179,-0.1835063696,-0.1611135751,-0.0786268488,0.2735330164,0.1603915989,-0.2609639764,-0.4676229954,0.025445275,0.5799104571,-0.3833327889,-0.1049101353,0.0173238888,-0.0641884655,-0.0145640727,-0.1484284103,0.5461590886,0.0278828703,0.0070441538,0.1480559707,-0.2995674014]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"It would be nice if the documentation elaborated on all the possible values for `download_mode` and\/or a link to `datasets.GenerateMode`. \r\nThis info here:\r\n```\r\n \"\"\"`Enum` for how to treat pre-existing downloads and data.\r\n The default mode is `REUSE_DATASET_IF_EXISTS`, which will reuse both\r\n raw downloads and the prepared dataset if they exist.\r\n The generations modes:\r\n | | Downloads | Dataset |\r\n | -----------------------------------|-----------|---------|\r\n | `REUSE_DATASET_IF_EXISTS` (default)| Reuse | Reuse |\r\n | `REUSE_CACHE_IF_EXISTS` | Reuse | Fresh |\r\n | `FORCE_REDOWNLOAD` | Fresh | Fresh |\r\n```","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":84,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nIt would be nice if the documentation elaborated on all the possible values for `download_mode` and\/or a link to `datasets.GenerateMode`. \r\nThis info here:\r\n```\r\n \"\"\"`Enum` for how to treat pre-existing downloads and data.\r\n The default mode is `REUSE_DATASET_IF_EXISTS`, which will reuse both\r\n raw downloads and the prepared dataset if they exist.\r\n The generations modes:\r\n | | Downloads | Dataset |\r\n | -----------------------------------|-----------|---------|\r\n | `REUSE_DATASET_IF_EXISTS` (default)| Reuse | Reuse |\r\n | `REUSE_CACHE_IF_EXISTS` | Reuse | Fresh |\r\n | `FORCE_REDOWNLOAD` | Fresh | Fresh |\r\n```","embeddings":[-0.132575646,0.0042611659,-0.1003257781,0.1450196654,0.2611149251,0.0885635614,0.1497801691,-0.0054396302,0.0523649976,-0.2209995091,-0.0105340136,-0.0093214773,-0.0788026974,0.0148025285,-0.0911452025,0.4515508115,0.1717670411,0.0544777438,-0.2703152001,-0.0126867108,-0.0795434117,-0.0508989282,-0.1528280079,-0.0177352671,-0.32577914,-0.0017379133,0.0608107559,-0.0484852456,-0.2403380722,-0.6755395532,0.5190625191,0.4494843185,0.2083913535,0.1146773845,-0.0001168151,0.0489492677,0.3039731085,-0.1338789612,-0.2875078917,-0.0103915669,-0.545085609,-0.2912785709,0.1124305576,-0.1120601073,0.0395753756,-0.1538888067,0.0811221823,-0.5785891414,0.3715526164,0.3695131838,0.1818338633,-0.1832768321,-0.2001060545,0.0893235952,0.141222015,0.2437101752,-0.1237794831,0.037241336,0.322075516,0.1247805357,-0.018199617,0.2879489958,-0.0560561381,0.2960295379,0.565648675,0.0093442956,0.1636774093,-0.4565844536,0.2388889641,0.0250146352,1.1493660212,-0.2332352996,-0.2873546481,-0.3733200729,-0.1109540015,-0.3186913729,0.2588652074,0.0140755,-0.0272201523,0.1969658136,-0.6656895876,-0.2369510233,-0.1510466486,-0.1439849883,-0.1140457541,0.0083224606,-0.3288093209,0.0601481088,0.0228604544,-0.044869937,0.5436966419,-0.4406124651,0.0177962817,0.2235455811,-0.3665752411,-0.135038197,0.1707547605,0.3412370086,0.0559608191,0.2809710801,0.004347065,-0.1090332493,-0.1632771045,0.079607442,0.4099765122,0.1847312152,0.153617993,-0.1211039722,0.1026657671,-0.0653599501,-0.0127923414,-0.1522945166,0.0578929633,-0.3024713397,0.5569894314,0.0259583648,0.2239093632,-0.2749815285,-0.3039260209,-0.0038776384,0.014193425,-0.2060308754,0.0928864032,0.1861323118,-0.0754157305,0.0621719472,-0.2922911942,0.0188443791,-0.159011811,-0.3619897962,-0.1578511894,-0.1793255508,-0.3269338608,0.0204795729,0.2394786775,-0.2574568391,0.1005438492,0.4393438399,0.0420365855,-0.2138965577,0.1889699399,0.0254874099,0.2010563165,0.4977919161,-0.2865719199,0.1973752677,0.2685185969,0.0111293374,-0.1918101013,0.4145025015,-0.3691080213,-0.5318116546,0.1474086493,0.139004916,-0.506051302,-0.1422891319,-0.0900613442,0.1105869412,0.3779514432,-0.0056704073,0.1612610966,-0.1432572454,-0.1368408501,-0.3684626222,0.0003973177,0.4416205287,-0.611340344,-0.0170365162,-0.3705367744,-0.2274078131,0.1591456085,0.1475977004,-0.3789155781,0.1754211485,-0.1863431185,-0.2089995593,0.1826339066,-0.1241744831,-0.3571071923,0.1453820467,0.3975531161,0.2618071437,0.0976000652,0.1303774863,0.1016698256,-0.1206594333,-0.1181316376,-0.0377236307,-0.0049879448,-0.2503035367,-0.0772855729,-0.3503139913,0.1276770532,-0.1639273614,0.0155469934,0.1499371231,0.0802270621,-0.1014925092,-0.0588772818,-0.0537634157,0.0857708976,0.1736115813,0.1347942948,0.1471327841,0.0282387193,0.1696176082,-0.6066907644,0.3852394223,0.0662049577,-0.3480160236,-0.0871552899,-0.211859405,-0.2996636629,-0.0250247512,-0.3529517949,-0.2120274156,0.0458297171,-0.0506738834,0.1571803093,-0.1518154591,-0.2582426965,0.3438586891,-0.2423756868,0.0148016233,-0.1462842226,0.1495354921,-0.0488027707,0.2019355744,-0.2094453871,-0.0156090166,0.1156178489,0.0086224116,0.0717082545,0.4083191454,0.1737720817,0.4718440175,-0.1032984555,0.4203929007,0.1377690732,0.0736767352,0.2135801762,-0.1835939586,0.1024537906,0.0142307077,-0.2718121707,0.3134011626,-0.1931469142,0.0768746585,-0.1140330881,-0.2736133039,0.1306158602,0.0810394958,-0.0460012183,-0.2751426995,-0.0808304176,0.1798795462,0.0037942186,0.3058403432,-0.0288724918,0.2879081964,0.2200863957,-0.2195736766,-0.0610304438,-0.0124928104,0.0235614646,-0.4724953473,0.1849091053,0.4928191006,0.555147469,0.0222601853,0.2973548174,0.043381393,-0.0461560786,-0.2960595489,0.3671460152,-0.1907936782,0.0465036333,0.1567800939,-0.0506649427,-0.1499492526,-0.3461315334,0.1256599128,0.3300360441,0.3210259676,-0.4505416453,0.1189449579,-0.4110972881,-0.0787225813,-0.0761827454,0.1549040824,-0.2310667783,-0.3646591306,-0.1081455722,0.1242578849,0.1314709485,-0.0163651314,-0.1997042,0.2560541332,-0.1282802969,0.0514581017,-0.1284702718,-0.1792527735,-0.1449583173,0.065004088,0.0717630014,-0.3010616899,0.258638978,-0.1313634813,0.094339475,-0.2228563428,-0.0122406129,0.0962457359,0.0774576664,0.2347013056,-0.131571576,0.0296801832,0.1202929989,0.1396782696,-0.081918329,-0.0028210666,-0.0017424873,-0.2095082849,-0.0061241915,0.1070558205,-0.1420656741,-0.3035123646,-0.2983780503,-0.0950419977,-0.2578513622,0.0004816628,-0.0052524256,0.5572144389,-0.0099968771,0.0066042226,0.0939478353,0.015162576,-0.5945007205,-0.4954152405,0.2906995714,-0.1030748338,-0.0394016244,0.2776832581,0.0995816216,0.3611543477,0.2926027477,-0.5287021995,-0.0356636159,-0.1569282264,0.1076141,0.1552859247,0.0289056674,0.2721232176,0.0868229419,-0.0061669108,-0.1301562637,0.0074913995,-0.1374365091,-0.0472735874,0.2741394937,0.0605774149,0.0238111783,0.0636950731,0.9581979513,0.4418125749,-0.0832928121,0.1660141349,0.0828534365,0.6536433697,0.088015765,-0.0483274721,0.0106890174,-0.2249955535,-0.2996979952,0.0261038616,0.0156354215,-0.0034118362,0.0951902941,0.4420705438,-0.1336169839,-0.2970892191,0.055939272,-0.3286455274,0.4383737445,0.1199656427,-0.0246776566,-0.1355341524,-0.0386380069,-0.316853255,0.2735345364,0.3320982456,0.0051927073,-0.2060045451,-0.0337153226,-0.2731868625,0.198052749,0.0885881707,0.1384667605,-0.0764437094,0.2041453421,0.3053606749,0.0551660396,0.4123055935,-0.6105519533,-0.0105049368,0.1238894761,0.0826096013,-0.0288583264,-0.1715600789,-0.0109607941,0.1350925714,0.0507854223,-0.120547317,0.0707469806,-0.2454445362,-0.2502867579,0.0775805339,-0.0945285782,-0.1838328242,-0.2000323981,-0.1234360188,-0.2955692708,-0.0423199981,-0.0776298717,-0.2414809763,-0.3587051034,0.1035591364,-0.0516121611,0.2731730938,0.109730497,0.1589665711,0.2044169307,0.1661432385,-0.018342277,0.3403426111,-0.4006919861,0.3688643277,0.507334888,-0.0451886132,0.0010621722,-0.0655848309,0.0574607439,-0.0769571513,0.2200240195,-0.3538410068,0.0212958548,-0.0334314182,-0.2535518706,-0.4195622802,-0.0767133906,0.3313021064,0.0247687735,-0.2231519222,-0.8084747791,0.5134011507,0.1843257099,-0.2629623711,0.06299714,0.0593877621,-0.2338336408,-0.3189501464,0.2463939488,0.6856463552,0.0211212765,0.1604952514,0.0181620531,-0.0361167938,0.5817766786,-0.2473655343,0.1859795153,-0.0498430058,0.0205071904,-0.1884727478,-0.1235983148,0.4794451594,0.2287913412,-0.1546936929,0.3027621508,-0.186023742,0.39604792,-0.1069144681,-0.1233858317,-0.2225198895,0.0205066446,0.1676725745,0.1771556884,-0.0417632349,0.3428207934,-0.1581338197,0.2205705345,0.2329137474,0.2162550837,-0.4692493081,-0.251429975,0.127587989,0.2413775176,-0.1821361482,-0.2639633417,0.3876970708,0.4182442129,0.3648182452,-0.1093133315,-0.2285253704,0.2783448398,0.0890798271,0.1886729449,-0.1542813331,-0.0647230744,0.3571134806,0.0443256348,-0.2755440772,0.2178252488,-0.0168756675,-0.038072221,-0.0511509366,0.182441473,-0.0326609425,-0.5251840353,-0.1205990389,0.2801264524,0.127941668,0.1533234864,0.0972220972,0.1801127046,0.1680440456,0.3907105029,-0.575872004,-0.2164125443,-0.1327159107,0.2938949764,0.4757038355,-0.2232360095,0.2769854963,-0.3090809286,-0.2733973265,-0.1132372767,0.086471945,-0.1512522995,-0.2568321824,0.0809842497,0.0562291369,0.3315257728,0.0911223367,0.0808098242,0.0851493627,0.1114174575,-0.1219341606,-0.0954745188,-0.1219083369,-0.0759996325,-0.0955836624,0.3711849451,-0.0930886492,-0.2647480071,0.1706953645,-0.1295152456,-0.2550218403,0.0876841694,-0.0030135259,0.16419366,0.0623026229,-0.1257254928,0.0457171537,-0.3645147383,-0.1038499698,-0.1774617434,-0.524784863,-0.1387736499,-0.0082629593,0.1929164529,0.111476779,-0.3349023461,0.0274395365,-0.2507593036,-0.1026532203,-0.0996653736,0.1932182759,0.2394442409,0.0643022209,0.2433053553,-0.127657786,0.0600160733,-0.1710183918,-0.0793192238,0.0448628739,0.3459503353,0.3399980962,0.3575984538,-0.1772592366,-0.0374859795,-0.3184013069,0.3974718451,0.193723008,0.0382943898,0.2872058749,-0.2777395248,0.2308436781,0.078010723,0.3234718442,0.3506149352,-0.1437326521,-0.1888293326,0.0375485085,0.133792758,-0.0908211619,-0.1459820122,-0.0368662253,0.3552684188,-0.08665663,0.2495104373,0.5172488093,0.0803849325,0.5661527514,-0.0108084474,0.5127908587,-0.3276567459,0.4368333519,0.3536316752,0.1125664636,-0.0184973702,0.4581286609,0.3813281059,0.1320946813,0.5072706938,0.0320479348,0.22377792,0.3761743009,0.2062734067,-0.3331028521,-0.7990939617,0.2745979428,0.4237221777,-0.4052152038,0.2453992069,0.1098915488,-0.0461164415,-0.0603363179,0.0330282226,-0.196010083,0.1885323226,-0.1355808973,-0.231207177,0.2734331787,-0.3501799703,-0.072510846,0.2056947052,-0.2398762256,-0.1312088221,0.1012741327,-0.1173487306,-0.1434180439,-0.4577664733,0.4708824456,-0.0328092687,0.2585250735,-0.0714808777,0.2669600546,0.2390209287,-0.0912995338,0.2848149538,0.2631331086,0.5140573382,0.1025296524,0.0745016485,0.1179382429,0.0744434148,0.027463235,-0.1108152941,0.236211434,-0.0721882209,0.02717104,0.0854701921,0.0888835192,0.0558627918,0.1025628075,-0.0258004982,0.0754956156,-0.4381799102,0.1496443152,0.1288441122,0.0857993066,0.137563616,0.2785962224,-0.3378023207,-0.2569725513,0.5932246447,-0.046314612,0.2142388225,-0.1442986131,0.0704877004,-0.0217846278,0.4156266153,0.1063924357,0.0766220093,-0.1099088714,-0.0176006798,-0.7548527122,-0.2957159877,-0.1501213759,0.1948822737,0.21563299,0.1462046802,-0.110119164,0.2309693247,-0.0136859566,-0.1420092583,-0.0463029891,-0.0584754013,0.0807637647,-0.1495244354,-0.1093722582,-0.1099742278,-0.2897756994,-0.2760651112,0.2154343873,-0.136738658,-0.0569443665,-0.0696993247,-0.0235586669,0.284776628,0.1172648966,0.4224748313,0.2611921132,0.2404793799,0.2846589983,0.0044919611,-0.3045017123,-0.0507765487,-0.0870600119,0.1580365896,0.0261008032,-0.0540098064,-0.5192809105,0.1436507106,-0.1828611791,0.2056902796,0.1607756764,-0.11458119,-0.0355270281,-0.1985425651,-0.1444984972,0.1064965874,0.116549179,0.2339967042,-0.0426434875,0.420055598,-0.3301743865,-0.0878036022,0.2840035558,-0.1189451143,-0.1438233405,-0.0291912314,0.1371345669,0.2747971117,-0.1208777204,-0.533290565,0.1804154217,0.4839946032,-0.32937482,-0.1950064152,0.0920226648,0.0073721446,-0.0753997192,-0.1456322968,0.5694743395,0.1046599448,-0.1223846376,0.1178272292,-0.3137853146]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"I have another question. Assuming that I understood correctly and there is reuse of datasets files when caching is disabled (!), I'm guessing there is a directory that is created based on some information on the dataset file. I'm interested in the situation where I'm loading a (custom) dataset from local disk. What information is used to create the directory\/filenames where the files are stored?\r\n\r\nI'm concerned about the following scenario: if I have a file, let's say `train.csv` at path `the_path`, run once, the dataset is prepared, some models are run, etc. Now let's say there is an issue and I recreate `train.csv` at the same path `the_path`. Is there enough information in the temporary name\/hash to *not* reload the *old* prepared dataset (e.g., timestamp of the file)? Or is it going to reload the *old* prepared file? ","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":139,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nI have another question. Assuming that I understood correctly and there is reuse of datasets files when caching is disabled (!), I'm guessing there is a directory that is created based on some information on the dataset file. I'm interested in the situation where I'm loading a (custom) dataset from local disk. What information is used to create the directory\/filenames where the files are stored?\r\n\r\nI'm concerned about the following scenario: if I have a file, let's say `train.csv` at path `the_path`, run once, the dataset is prepared, some models are run, etc. Now let's say there is an issue and I recreate `train.csv` at the same path `the_path`. Is there enough information in the temporary name\/hash to *not* reload the *old* prepared dataset (e.g., timestamp of the file)? Or is it going to reload the *old* prepared file? ","embeddings":[0.0396139957,0.1744792014,-0.1220532879,0.3688902855,0.122898601,0.0972481668,0.3478941023,-0.043518573,0.1747947037,-0.1183368042,0.0815539956,-0.0739900917,-0.0967100561,-0.1997417659,0.1834369302,0.4033867419,0.2352863103,0.0705635548,-0.1633934826,-0.2112264335,-0.2158203423,-0.1136627346,0.0287037473,0.0257110316,-0.4653145969,-0.0532943495,-0.1611276567,0.3412358463,0.0454519801,-0.5218764544,0.1373799741,0.3655670285,0.263463825,0.391907841,-0.0001192328,0.09008874,0.0751684755,-0.2078173459,-0.48993361,-0.2357729673,-0.6089838743,-0.3132255077,0.105227448,-0.2591819763,0.1430513114,-0.2797461152,0.1961197108,-0.5753272176,0.3598959744,0.2168545723,0.131611675,-0.3067780137,-0.4450264573,0.1514536887,0.1042250842,0.3387694061,-0.0052347714,0.096113719,0.1016787887,-0.0347756483,-0.0932452232,0.078442499,0.0769727826,0.0982000157,0.5855482221,0.1981262565,-0.1865921915,-0.1474257708,0.1589407474,0.0758652613,0.8340936899,-0.3309270442,-0.440697819,-0.3639886379,-0.040344812,-0.0652806982,0.3427962363,0.0169863086,-0.0998441353,0.1036839411,-0.3586764932,-0.306771785,-0.1540292352,-0.1644052267,0.14031896,-0.174760744,-0.2014998049,0.0858067721,-0.1667185724,0.0152185252,0.4679963589,-0.6780709028,0.1247476116,0.3843238056,-0.103743203,0.0395144485,0.0338585563,0.4278025329,-0.1384791732,-0.0401636139,0.320382148,-0.0879884437,-0.2387324572,0.1491919756,0.3794382811,0.1677575111,-0.0275194291,0.0827718228,0.1578300893,-0.1401789784,-0.261226505,-0.2612173855,-0.2172830701,-0.2086579353,0.4720521867,-0.0031830173,0.1987889558,-0.2219714075,0.0133667858,0.0747099891,0.0689277053,-0.1600145251,0.0717227608,0.1632653624,-0.1020670086,-0.044686228,-0.3007525504,-0.197535947,-0.0810197964,0.034819033,-0.1088869572,-0.2815064192,-0.347922802,0.0796191245,0.2255404294,-0.2305431217,0.2035108358,0.2208004296,-0.1127785593,-0.1163142696,0.0974741876,-0.0763822719,0.3487943411,0.4564616382,-0.2026843429,0.1994194984,0.0794386044,-0.119977653,-0.2430878878,0.3181140721,-0.5094616413,-0.4343141615,0.1935459524,0.0951597244,-0.2897161245,0.1831614375,-0.009421505,-0.1164964139,0.2761576176,-0.0150649846,0.1290667355,-0.1315943897,-0.1630372703,-0.3629796207,0.0344613865,0.5511756539,-0.7481017113,0.0441918857,-0.0422222428,-0.1912372559,-0.2221763283,0.3882530034,-0.4213086963,0.2873209119,-0.2788359523,-0.0958224535,0.3518692851,-0.1698625386,-0.1403254718,0.0512461476,0.0592428669,0.2768931687,0.1626450419,0.2749512792,-0.2181053758,-0.1565304697,-0.036564514,0.0067199864,-0.1402028352,-0.2560178936,-0.051706586,-0.0527877994,0.0278318748,-0.1478358358,-0.1344298273,0.2383815646,0.1785310954,-0.039702978,0.0203971695,0.0235946756,0.0684596598,0.4347875118,0.0913381577,0.2837566435,0.2616576254,0.246180132,-0.6048414707,0.3834663033,-0.0249567237,-0.5644062757,-0.1697784811,-0.2210454494,-0.1049434915,-0.0595596395,-0.3477904797,-0.020488428,-0.0005916337,0.2033090293,0.1174134389,-0.1571396738,-0.3319161534,0.2712101042,-0.443734169,-0.0700598657,-0.2444700301,-0.0145351104,-0.0495776311,0.1621655226,-0.3741001487,0.1364878565,0.0358717032,-0.1138863191,0.0252653491,0.4073795974,0.1449673325,0.5493708253,0.244940713,0.5470482111,-0.0219655391,0.286717236,0.4552431107,-0.2784743607,0.2444137186,-0.0584344417,-0.3102046549,0.37250337,-0.5067824721,0.1116784215,0.0574625209,-0.4737315476,0.2037767023,-0.003811205,0.0048187231,-0.1751117408,-0.0792367235,0.2939703763,0.3907528818,0.3153144121,-0.074699685,0.0330108739,0.2318620533,-0.1005252525,-0.2088135481,-0.0361079834,-0.1767385155,-0.3452965617,0.0123527506,0.5395867825,0.5537014008,0.0178910829,0.3181603551,-0.0884636715,-0.0170324892,-0.2547692358,0.0898669884,-0.1171046495,-0.1454311311,0.3790723681,-0.0249046329,-0.0418304466,-0.2419884205,0.0240629688,0.2772979438,0.0776957348,-0.2950513363,0.2372207344,-0.115787223,-0.1908859909,-0.2870165408,0.1896909028,-0.0941100568,-0.1602423191,-0.023361193,0.0946948603,0.0126699526,0.0027123757,-0.1074908301,0.3336413205,-0.1697277129,-0.1477844566,-0.1375887841,-0.2733901441,-0.0900296643,0.0388462469,0.1775089353,-0.1694718003,0.0805361569,-0.0859056041,0.0763437748,-0.4038631618,0.2571551204,-0.0204170477,0.0533889793,0.2470944077,-0.1725019366,0.0728258789,0.0072484068,0.0569072478,0.0198910814,0.0912606046,-0.01183521,-0.2892057598,0.1912009567,0.0224021245,-0.0786828771,-0.1699850857,-0.3214595914,0.0282163844,0.0466776974,0.0863009393,0.0366173238,0.4797021747,-0.1061562598,-0.0073569715,-0.028706776,0.0170078054,-0.3930701613,-0.7377177477,0.1330609024,-0.1119260639,0.0786107704,0.0585448779,-0.0297630616,0.1121716872,0.3737277389,-0.504437089,-0.1825035065,0.210889712,0.294691056,0.1026549786,-0.0055985181,0.3985081911,0.0343538299,-0.0381631702,-0.2396537215,0.1046172008,0.0721841827,0.056721352,0.2137652338,0.1187241301,-0.0438153706,-0.1013187319,1.0561300516,0.065662019,-0.095050402,0.198408857,0.0055845962,0.7749556899,-0.0111275185,0.0602477901,-0.2000258118,-0.1283800006,-0.56447649,0.1442537606,0.1790536493,0.1041959748,0.0803923979,0.3580323458,-0.3325107396,-0.247289449,0.2445425689,-0.3849089742,0.4336819351,0.1737383008,-0.2445264757,-0.2122984529,-0.146805957,-0.1929415166,0.2209748924,0.3969123662,-0.1111954972,-0.3127847016,-0.0740492716,0.0500066727,-0.0133446837,0.2047041506,0.2459461093,-0.0905716568,0.0419184789,0.1568480283,0.185742408,0.3920873106,-0.406494379,0.0583009832,0.2035020441,0.0753713325,0.0133512849,-0.0592624284,0.0275285225,0.1789463609,-0.2179643959,0.2962104082,0.3291772306,-0.2602208555,-0.299903214,0.4384298623,-0.3279012144,-0.0225322563,-0.1101692095,0.090969108,0.0019616573,-0.2577278614,-0.2456548959,-0.3908724785,-0.1066592261,0.1209404618,-0.1705131829,0.1837504953,0.0841669142,0.084081009,0.259039551,0.0627165437,0.0964842141,0.4471685588,0.036060147,0.4149880409,0.4983239472,-0.4157919288,0.0295081493,-0.0692331418,0.075657472,0.0925726369,0.4459820092,-0.2754133344,-0.0583277233,0.0116503611,-0.3149535656,-0.5572513342,0.1016678885,0.2380487174,0.1453149319,-0.2457350343,-0.7298757434,0.4616045654,0.0442892872,-0.4533214569,0.1364473403,-0.0430615284,-0.4334998131,0.0092919618,0.0433704257,0.7851680517,-0.137771666,0.3983212113,-0.2703844309,0.0335509665,0.2256225497,-0.3931541741,-0.0027617195,-0.1618190408,-0.1194536984,-0.2122356296,-0.1050596461,0.4006747305,0.4143877625,0.2862845361,0.4097852409,-0.0912253112,0.4684946537,-0.1631309986,0.1035575271,-0.100899294,-0.0715286508,0.1278520375,0.0950082242,0.2018307298,0.2626236677,-0.1316262633,0.2657703459,0.1029041559,0.2221252471,-0.0917192176,-0.2430613786,-0.1948888153,0.311228931,-0.3205515742,-0.4092464149,-0.227576077,0.4445063472,0.3319456279,-0.0499682948,-0.0220468696,0.3189427853,0.0405939035,0.2928324342,-0.2071420848,-0.0023629742,0.2992729247,0.2187440842,0.0425270312,-0.0308257565,-0.3231315315,-0.065374516,-0.2003763169,0.2053970844,0.2990846038,-0.4685586095,-0.2136063874,0.0552726984,-0.0002940916,0.1352769583,0.0712408125,0.1472804695,0.200704664,0.2619278729,-0.3904380202,-0.2276108712,-0.0714148805,0.3039796054,0.3663794994,-0.1873263568,0.4460788965,-0.5390532613,-0.1543507129,-0.1125841811,0.230624184,-0.0064290543,-0.0702816993,0.1091833413,0.1758557856,0.2060984522,0.0015502564,0.064585045,0.2470375001,0.0373875387,-0.2635701001,-0.1857196391,-0.1792279035,0.0767905712,0.2825473249,0.5055933595,0.049781356,-0.326533556,-0.0195428859,-0.0958999246,-0.233286351,0.1494162977,-0.0492360108,0.3271291852,0.0763111338,0.1545310467,0.1579227,-0.307128042,-0.0096324105,-0.2811216414,-0.2304947078,-0.1399681121,0.0323975421,0.1811800301,0.000329007,-0.1888404936,-0.1868029088,-0.4582824111,0.1054100618,-0.166519165,0.0591925569,0.3447114229,0.1310682744,0.2091640681,0.1343014538,0.1419612616,-0.0352532677,0.1301222742,-0.1393749416,0.4549197555,0.2578461766,0.1962607056,-0.3528995514,0.0233943556,-0.3583092093,0.2972155511,0.0489135571,-0.0411170088,0.1730858088,-0.0618501641,0.1825910956,0.1147224009,0.3548704982,0.4522508383,-0.1832130253,-0.3053302467,0.3238880038,0.0878386945,-0.0550881363,-0.1283651441,0.0072762822,0.087558195,-0.1063701957,0.241677627,0.4929354489,-0.0420583449,0.1961424649,-0.1603521258,0.4638437927,-0.2952373028,0.2642432451,0.2453193814,0.1432914138,0.0542682037,0.2146519423,0.4828643203,0.1201797128,0.7716601491,-0.016387634,0.2793602645,0.373804003,0.2739771307,-0.2553504407,-0.6836423278,0.2907900512,0.3437066078,-0.4014596045,0.2547961175,-0.0376695842,0.1597912908,-0.1283015758,0.1722543687,0.0633054003,0.1602606475,-0.1695295423,-0.4084586203,0.2957370281,-0.3943382204,0.1876867712,0.1556524187,-0.0171741936,-0.018581422,0.0033890561,-0.0121166278,0.1542948037,-0.1750787348,0.4936147034,0.1188831106,0.178524971,-0.1551939696,0.1975999027,0.2845064402,0.0919049755,0.2843462825,0.0773596019,0.2557862401,0.1358577311,0.0959464312,0.1754267961,-0.0277326237,0.0102462946,-0.0954828262,0.1879786998,-0.1585476995,-0.1055002287,0.1334713846,0.1134380028,0.0749187469,0.2434409857,0.0278371368,0.1547426432,-0.2178726494,0.0431210995,0.1784570068,0.0230485499,0.0505611263,0.2182837874,-0.1799523979,-0.289468199,0.575160563,0.0298290849,0.2406537682,0.04169292,0.0304981135,-0.1033739597,0.3388955891,0.2730054259,-0.0102566648,-0.2040196657,-0.1510638446,-0.5926517844,-0.1760676354,-0.147804305,0.1158547848,-0.1470402479,0.1717142463,-0.0730914623,-0.0551582649,-0.0662258938,-0.1378733516,-0.1780122519,0.0832472071,-0.0565333962,-0.3101798594,-0.3974449635,0.0148078464,-0.1445543021,-0.4541039169,-0.0841965377,-0.0954917818,-0.1085955799,0.0121787619,-0.0570548363,0.1640075743,0.3761210144,0.408875525,-0.0518202521,0.0744909048,0.2234586179,0.0243090056,-0.2947926819,-0.3059476912,-0.0678379908,0.2062373459,0.0328585096,-0.0975559652,-0.6697198749,0.1179779544,-0.1831289381,-0.1426717788,-0.0076597063,-0.0265190452,-0.0208753776,-0.2445758283,-0.1177628636,0.260057807,0.2057588995,0.3443100452,0.040831387,0.2696208954,-0.1667965353,0.0884371027,0.4007047713,-0.0664077029,-0.1173945665,0.0312732346,0.2196320742,-0.0166659225,-0.1236056611,-0.4434629381,0.1094242558,0.6644132137,-0.2521063089,-0.0318181925,-0.138445735,-0.057436984,0.040493492,-0.1720846444,0.1647000313,0.043586079,-0.1685229093,0.0237753633,-0.2692070305]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"Thanks for the feedback, we'll work in improving this aspect of the documentation.\r\n\r\n> Where are these files stored? I guess not in the temporary directory that is removed...\r\n\r\nWe're using the Arrow file format to load datasets. Therefore each time you load a dataset, it is prepared as an arrow file on your disk. By default the file is located in the ~\/.cache\/huggingface\/datasets\/\/\/ directory.\r\n\r\n> What information is used to create the directory\/filenames where the files are stored?\r\n\r\nThe config_id contains a hash that takes into account:\r\n- the dataset loader used and its source code (e.g. the \"csv\" loader)\r\n- the arguments passed to the loader (e.g. the csv delimiter)\r\n- metadata of the local data files if any (e.g. their timestamps)\r\n\r\n> I'm concerned about the following scenario: if I have a file, let's say train.csv at path the_path, run once, the dataset is prepared, some models are run, etc. Now let's say there is an issue and I recreate train.csv at the same path the_path. Is there enough information in the temporary name\/hash to not reload the old prepared dataset (e.g., timestamp of the file)? Or is it going to reload the old prepared file?\r\n\r\nYes the timestamp of the local csv file is taken into account. If you edit your csv file, the config_id will change and loading the dataset will create a new arrow file.","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":231,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nThanks for the feedback, we'll work in improving this aspect of the documentation.\r\n\r\n> Where are these files stored? I guess not in the temporary directory that is removed...\r\n\r\nWe're using the Arrow file format to load datasets. Therefore each time you load a dataset, it is prepared as an arrow file on your disk. By default the file is located in the ~\/.cache\/huggingface\/datasets\/\/\/ directory.\r\n\r\n> What information is used to create the directory\/filenames where the files are stored?\r\n\r\nThe config_id contains a hash that takes into account:\r\n- the dataset loader used and its source code (e.g. the \"csv\" loader)\r\n- the arguments passed to the loader (e.g. the csv delimiter)\r\n- metadata of the local data files if any (e.g. their timestamps)\r\n\r\n> I'm concerned about the following scenario: if I have a file, let's say train.csv at path the_path, run once, the dataset is prepared, some models are run, etc. Now let's say there is an issue and I recreate train.csv at the same path the_path. Is there enough information in the temporary name\/hash to not reload the old prepared dataset (e.g., timestamp of the file)? Or is it going to reload the old prepared file?\r\n\r\nYes the timestamp of the local csv file is taken into account. If you edit your csv file, the config_id will change and loading the dataset will create a new arrow file.","embeddings":[0.0426528901,0.1128463596,-0.0747745112,0.3651132286,0.187513724,0.16579099,0.3618480265,-0.0763975754,0.1920935363,-0.145717144,0.1244352832,0.0901360363,-0.269436568,-0.1203824431,0.0926404372,0.3132811189,0.1218126342,0.0018566946,-0.2674063146,-0.0612732209,-0.0238487758,0.0139037659,-0.0476134755,0.0775936395,-0.5502688885,-0.1778058559,-0.0284800045,0.1388774067,-0.04224471,-0.5476482511,0.4293833077,0.2526207268,0.1022150964,0.4431468546,-0.0001136524,0.1564302742,0.1653490961,-0.0595049635,-0.3835014999,-0.0884687677,-0.4759688675,-0.2783870995,0.2284000963,-0.1791343987,0.0559503101,-0.1754128486,0.0664183646,-0.7208995819,0.4306547642,0.3034115732,0.2095777094,-0.1128296256,-0.3428081572,0.16692698,0.1212248057,0.1176112965,-0.0647921115,0.0958380178,0.0171205495,0.0299993977,-0.1908132881,0.3489200175,0.0284836553,0.0995286107,0.6798650026,0.1164378077,-0.3021637797,-0.2816374302,0.1710201502,-0.0168927088,0.8757190704,-0.3772922456,-0.3937720656,-0.4215785861,-0.0722437426,-0.2910638452,0.3705019653,-0.0349943116,0.0338183269,0.1965864748,-0.4137375951,-0.2978313863,-0.1873830855,-0.1798488945,-0.0950306356,-0.2104989588,-0.3281470239,0.0431772061,0.0529060252,0.0894156471,0.3412467837,-0.4796886146,0.0544818491,0.3145650029,-0.3024866879,0.0585052669,0.0669266209,0.3536751568,-0.1032308489,0.1801586002,0.3130632043,-0.0729674026,-0.2701882124,0.1462639421,0.3341298699,0.3940928578,-0.0065452317,0.0667705387,0.2119108289,-0.1026698872,-0.1689470112,-0.2856127322,-0.0797221437,-0.4382008612,0.3683880568,0.000018508,0.182117328,-0.2402320653,-0.0969628394,0.0762550831,0.064712882,-0.1605023444,0.2484934181,0.1427246332,-0.0343203582,0.0168877076,-0.1806506366,-0.1179083884,-0.1633675843,-0.0465737991,-0.1743831933,-0.1365133673,-0.3817654252,0.1212826893,0.2318619043,-0.2882330716,0.172854811,0.2895810008,0.1174064279,-0.1914015263,0.1421865225,-0.099582918,0.3003712893,0.3959777057,-0.3659523129,0.2547310293,0.2532998025,0.0470304228,-0.238289535,0.2351591587,-0.5508418083,-0.5308260322,0.1095277816,0.1173829436,-0.418703407,-0.0184019562,-0.2455897331,-0.1156271249,0.3187911212,0.062400654,0.2418775707,-0.2037982196,-0.1772338003,-0.3270478845,0.0544245467,0.5265339613,-0.6201192737,0.0139235519,-0.1013620198,-0.056978371,-0.0566182621,0.4388082325,-0.564440012,0.2282956541,-0.2668422163,-0.0418815315,0.2101627588,-0.2359636277,-0.2370516509,0.0387334414,0.0486462861,0.3653502166,0.1788061261,0.06647484,0.0901674479,-0.13607499,-0.0383855477,-0.0327425189,0.0925370231,-0.2657940686,-0.1053954437,-0.2953509092,0.0831988603,-0.2103252411,-0.0497050881,-0.013660498,0.0565380715,-0.1861411333,-0.0096127726,-0.1338639259,0.0811727941,0.4035136104,0.0486800149,0.284064889,0.095466435,0.2152553946,-0.5267915726,0.423304379,0.1852604151,-0.4941213131,-0.2006000727,-0.2615771294,-0.0718733817,-0.0862336755,-0.4411376417,-0.1706931144,0.0341377407,0.0446686447,0.1159935743,-0.0514420792,-0.3322877586,0.4101415873,-0.4049244225,-0.0417120084,-0.3104243279,0.084500432,0.0234978925,0.2354417592,-0.3102506399,0.1405351907,0.0299967621,-0.06298998,0.0427050665,0.36036098,0.1344787329,0.4914358556,0.1914765388,0.519780755,0.0840321481,-0.0144377733,0.3538705409,-0.1910290867,0.1165380031,0.0307780337,-0.3368668258,0.3540522158,-0.3387058377,0.1638696492,-0.0121469386,-0.2903886735,0.1131076068,0.0169298779,-0.0755761564,-0.0768131614,-0.0301896594,0.0985094756,0.254365176,0.4224379957,-0.2281018943,0.1754329354,0.3237643838,-0.0215610415,-0.261690706,0.0235331543,-0.1400740743,-0.2476538271,0.1388082355,0.4800477028,0.5014015436,0.1201579794,0.2383859158,-0.0533888675,0.0216731839,-0.2430947423,0.2449923307,-0.0331893004,0.083022736,0.1908868253,-0.0131332744,0.0383751877,-0.2990606129,-0.0289780442,0.1093659773,0.0501198731,-0.4698907733,0.3693457246,-0.3929538131,-0.0820747241,-0.3823751807,0.0865272582,-0.227621913,-0.2523648739,-0.0142976884,0.0020540985,0.0541889369,0.0533842295,-0.2774982452,0.4074819684,-0.1261828691,-0.0919074938,-0.1295308024,-0.1039870307,-0.1594325751,0.0164022036,0.2174411863,-0.1975784004,0.0960609838,-0.0307019521,0.0838793591,-0.3365899026,0.1275617778,0.0149003174,0.0186424032,0.2037233859,-0.1277354062,0.0353868119,-0.1253300309,-0.0770093277,-0.1005413979,0.0471430346,0.0399333239,-0.1819608957,0.1469050795,0.1284316927,-0.1161590591,-0.2377016544,-0.3192602694,-0.132194519,0.0855249316,0.0390010215,0.0027544945,0.5872468352,-0.0016312777,0.1230578125,-0.0244505946,0.0663004667,-0.441637069,-0.6039819121,0.3160150647,-0.1407755613,-0.0384192504,0.1388538927,0.0660286695,0.2734756768,0.1964789331,-0.6239000559,-0.1205821261,-0.0345267914,0.1573712379,0.1162360385,0.0700723305,0.309738785,0.0374815576,-0.0466117337,-0.150802061,0.0029912887,0.0816945359,-0.0783292428,0.2549778819,0.1132200956,-0.0473794565,0.0169874933,1.1582586765,0.2377635986,-0.0377980545,0.0956422314,0.0073740934,0.7471949458,-0.0396334566,-0.0848495737,-0.0874110982,-0.2571058571,-0.4753899872,0.0898691118,0.1911808103,0.1373717189,0.1831166595,0.3430119455,-0.1310899109,-0.3341440856,0.3024993837,-0.4036674201,0.3176871836,0.0099281697,-0.0714738816,-0.2217425406,-0.0479290299,-0.1426384598,0.3482542336,0.3176629543,0.0198274124,-0.1918480098,-0.1668291986,-0.2231999636,0.0548327565,-0.0498776957,0.1097148955,-0.1733345836,0.1313461065,0.2701214552,0.1460508108,0.4943933487,-0.2504502833,-0.0759193897,0.1707909554,0.0724025816,-0.0138058085,-0.1366952658,0.0794685706,0.2714104652,-0.0392791629,0.1366447061,0.224317655,-0.2971727252,-0.1961429417,0.2696304023,-0.2419202626,-0.052734971,-0.1840977371,-0.0568296313,-0.0981646478,-0.1347899884,0.0195159707,-0.287350744,-0.2387722731,0.0516465753,-0.1101530343,0.0859108567,0.1384114623,0.0411047004,0.3774224818,0.1493372172,-0.0511121638,0.6349805593,0.0398975611,0.3729882538,0.5501739383,-0.1853513718,-0.0917050168,-0.0443244129,-0.0234640315,0.1364943385,0.4127084613,-0.4581158757,0.0470288843,0.0467636958,-0.2780455053,-0.3946865797,-0.0761097297,0.3407290578,0.2532005906,-0.1485254616,-0.6971559525,0.5179476738,0.2509444952,-0.2987609804,0.2439565063,0.0462419838,-0.4107718766,-0.0370483585,0.1446019858,0.7890865207,-0.0115420539,0.353428632,-0.0712912828,-0.0696333274,0.3899291158,-0.3019837141,0.1157951951,-0.0856978297,-0.0856731832,-0.2145989686,-0.1433061659,0.3582910299,0.2706644833,0.0100116991,0.3308504522,-0.2380569726,0.2818346024,-0.2557389438,0.0498599149,-0.0701776817,-0.118982628,-0.0440059528,0.1469250023,0.1189132407,0.2945714891,-0.1537289172,0.1041405499,0.1377119124,0.194072336,-0.2254681587,-0.2816457152,-0.23212336,0.3175226152,-0.1013315618,-0.443872571,0.1494942456,0.4084871411,0.2014108747,-0.1299447268,-0.1075018495,0.2657452226,0.1492578387,0.1895849407,-0.1285016388,-0.1794518977,0.3612165749,0.1541674137,-0.0561227389,0.0460044146,-0.2187594622,-0.2864958346,0.1323195845,0.2304904759,0.1065843031,-0.4199064076,-0.1353827566,0.1652855128,0.082154803,0.1623176038,0.1254360676,0.1747611761,0.1839187741,0.3437950313,-0.3938896656,-0.2065222561,-0.0358070247,0.1985306889,0.2643139958,-0.1702116877,0.4318950474,-0.2670847178,-0.3354786634,-0.0778516605,0.1133329049,0.0880197138,-0.2367671728,0.0549247973,0.1126627922,0.3235100508,-0.0188434124,0.0040825726,0.2732168734,0.0082892356,-0.3613339663,-0.2348080575,-0.2408811897,0.0922416821,0.0175370164,0.4436586499,-0.0335394181,-0.2856045365,0.0626607761,-0.1526922435,-0.28901577,0.1619475484,-0.0786772147,0.0942884013,0.2255488485,0.0305917356,0.2534170747,-0.303198725,-0.0509899482,-0.2666368186,-0.3077819645,-0.1608223319,0.010461214,0.1589038074,-0.1022016779,-0.2055808604,-0.0441840813,-0.3034949303,0.0315058231,-0.1685610265,0.1219393983,0.3316381574,0.0964756161,0.0860096514,-0.205196619,0.0163839534,-0.0808411911,0.0041200947,0.0177726466,0.3359734714,0.3906490505,0.2816237807,-0.2716905773,0.0024271177,-0.2317798138,0.3015431464,-0.0651254877,0.1148612499,0.1655110568,-0.286347121,0.1028351411,0.199158296,0.3281067908,0.3717802167,-0.2649432719,-0.2294768244,0.1357257664,0.1309862286,-0.0262098722,-0.2149589509,0.0542246252,0.1898081005,-0.1325531751,0.2585157454,0.5278128386,-0.0127964038,0.2924180329,0.0566799156,0.4073824286,-0.2825934291,0.3838898242,0.135954231,0.0083706053,0.0159257986,0.4384000301,0.4868390262,0.222441569,0.6142733693,-0.0598112829,0.1821415126,0.4852375984,0.1930830777,-0.3621421456,-0.7948089242,0.1853654385,0.2902176976,-0.3958214223,0.1470615119,-0.0087451264,0.3306702673,-0.1507887691,-0.0326174982,-0.0210003015,0.207501173,-0.2223577499,-0.187792331,0.3327509463,-0.3243267238,0.138526544,0.1337724179,-0.0966908783,-0.0753626823,0.1147793606,-0.1287962794,-0.0892727152,-0.2601792812,0.4980159104,0.0721445531,0.3708717227,-0.0935227722,0.2920441628,0.0470705442,0.0807700083,0.1934072375,0.1907147914,0.4301682413,0.0948305801,0.2269955128,0.0762035251,-0.002200213,0.0119244009,-0.0995878875,0.197194919,-0.1713890284,-0.1878993958,0.1820628345,0.1363786608,0.0194269884,0.1632918417,-0.0603631511,0.1767074466,-0.3804508746,-0.0838854462,0.2030134946,-0.0060393554,-0.0172501523,0.1702433079,-0.2853621244,-0.2765758634,0.7414007187,0.0565528311,0.2432411015,-0.0573861338,0.0709838122,0.052309759,0.4304016531,0.2606304586,0.1481004953,-0.1230071858,-0.0288083162,-0.770713985,-0.1841091365,-0.1568199694,0.1471705437,0.0206822045,0.1961885393,-0.0730563998,0.0521332324,0.1274416149,-0.2160123587,-0.018225003,-0.1002864018,-0.026457794,-0.0534801111,-0.1319114119,0.0254150108,-0.1393777281,-0.2766142786,0.0604270212,-0.0223571528,-0.0560375452,0.0712019801,0.0394560695,0.0316875912,0.1538897753,0.4297095239,0.0328071006,0.1282011569,0.3001413047,0.0915507227,-0.3685914874,-0.1902424097,-0.1202827618,0.1641093493,-0.025724927,-0.0122083593,-0.6058148742,0.0169119574,-0.3258044422,0.0379636772,0.1219577715,-0.1550280601,-0.0037435445,-0.1074890941,-0.1270571053,0.2228344083,0.1746289879,0.3728477657,-0.0318343267,0.2100907266,-0.2636838257,-0.0118863629,0.419793427,-0.2700089514,-0.1785029173,0.1446351707,0.1899061352,0.2657697797,-0.117907308,-0.6296887994,0.1073625833,0.5160404444,-0.2359795272,-0.1337089539,-0.0295837782,-0.082080096,0.0360580012,-0.106757462,0.3352447748,-0.0085469084,-0.0522506908,0.0372903198,-0.349930197]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"Thank you for all your clarifications, really helpful! \r\n\r\nIf you have the bandwidth, please do revisit the api wrt cache disabling. Anywhere in the computer stack (hardware included) where you disable the cache, one assumes there is no caching that happens. ","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":41,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nThank you for all your clarifications, really helpful! \r\n\r\nIf you have the bandwidth, please do revisit the api wrt cache disabling. Anywhere in the computer stack (hardware included) where you disable the cache, one assumes there is no caching that happens. ","embeddings":[-0.0806554928,-0.0190928429,-0.0784587637,0.1829437613,0.2073648721,0.2329128832,0.171755597,0.0126285432,0.2039582878,-0.1684737206,0.1043677405,0.0651792958,-0.1479164511,-0.0486048199,-0.045972012,0.4469532669,0.1602656096,0.1373850703,-0.1307970285,-0.0402544253,-0.0730126277,-0.0087378025,-0.2274751663,0.1030513942,-0.5465358496,-0.1996650994,0.0297564063,-0.0258189291,-0.05394952,-0.4566874802,0.4985801578,0.1906560063,0.1309051365,0.2862364948,-0.0001159255,0.0028495502,0.1477109641,0.0074455054,-0.2455655932,0.1221066862,-0.5522729754,-0.3764937818,0.2959463298,-0.073448509,-0.0296691135,-0.2271424979,0.0803217441,-0.6296426654,0.4789683521,0.3128571212,0.2188513279,-0.0277704578,-0.3452477157,0.1580730528,-0.0628924295,-0.026200084,-0.1437474936,-0.0680628419,0.2651391625,-0.1496252716,-0.054432191,0.3541885018,-0.0805519745,0.4096233845,0.5019232631,0.0515010767,-0.3145160675,-0.3835275471,0.307474494,0.0309551489,0.9555873871,-0.2958950102,-0.1510228068,-0.32016325,-0.0614940412,-0.2522046268,0.2153982073,-0.0423638187,0.0364885852,0.2256952673,-0.5020576119,-0.2865841091,-0.0761595368,-0.1078191325,-0.2340299338,0.0283327401,-0.2219641507,0.079621084,-0.1803814024,-0.0442373604,0.4569609165,-0.4619446397,-0.0086373063,0.1626583785,-0.5250949264,-0.1069003791,0.1738810688,0.4802712202,-0.1630517542,0.1826247722,0.0631816164,0.0179973114,-0.0638797656,0.0455498733,0.442483753,0.2882967591,0.0465804674,0.0637087673,0.1841566712,-0.2733981907,-0.1512683928,-0.1937046945,0.1448383778,-0.318026036,0.559448123,-0.0312632173,0.3589950204,-0.3831406236,-0.3112589717,0.0587365441,-0.0189174302,-0.1682016999,0.0525005125,0.1131047755,-0.0656816438,-0.0090201236,-0.1372025758,-0.0024412859,-0.2364441603,-0.260489285,-0.2496363521,-0.1963600367,-0.2484804541,0.0164702274,0.2276224047,-0.1754544377,-0.0450234786,0.3477347791,0.1634486616,-0.218427971,0.1509355456,0.0006865881,0.2434147596,0.303089112,-0.2945020497,0.3164536655,0.4273717999,0.0003933724,-0.1998229325,0.4188993573,-0.4340383708,-0.4555507898,0.0640675202,0.131456539,-0.4850300848,-0.2716362178,-0.1847335845,-0.0189423263,0.408110261,-0.0017416151,0.154373616,-0.1329516023,-0.2522609234,-0.2450333536,-0.0040233508,0.3853209019,-0.6156472564,0.0083780978,-0.2004260421,-0.0571059883,0.1146910414,0.3705382943,-0.4463194013,0.1517270952,-0.1681305915,-0.1725257784,0.1848114729,-0.1718278974,-0.5081831217,0.2176715583,0.2019001395,0.3354839981,-0.0228355769,0.0741344318,0.1265327334,0.0012812277,-0.0319760963,-0.0523547828,0.0952003226,-0.2044354528,-0.2159064561,-0.4232357442,0.267224133,-0.1377511024,0.0715645999,0.1494498998,-0.0501820594,-0.1708871424,0.0718718693,-0.0569422841,0.0981407762,0.1958926469,0.1950851679,0.0945544764,0.1072174534,0.0522578545,-0.4603861868,0.3345037699,0.1416356415,-0.3668768704,-0.0786483139,-0.2039870173,-0.1024142355,-0.1082033962,-0.4313141406,-0.0742346272,0.0494787507,-0.0529498905,0.0037209645,-0.072317414,-0.2836421132,0.4135521054,-0.3291513622,-0.1031679735,-0.1165905744,0.2108018547,0.0391822495,0.1926305294,-0.2228360474,0.1413662583,0.0376594476,0.1866347492,0.0262457784,0.1384474486,0.042536214,0.3538772166,0.0919521824,0.3583668768,0.0069250194,-0.0616680384,0.3143136203,0.0478435978,0.2054936439,-0.0972665846,-0.1094152629,0.2266597301,-0.1106720343,0.1286316067,-0.2301409841,-0.1533769369,0.1183973104,-0.0293303467,0.056472823,-0.1460091174,-0.0302134585,0.0089361845,-0.0356497467,0.3004308045,-0.1030885875,0.1872988641,0.3256934881,0.0171542596,0.032515116,0.0430144332,-0.135059759,-0.3747719228,0.1514116228,0.3989087045,0.5773319006,0.1253362894,0.1422910392,0.0576081835,0.0563110709,-0.1647052318,0.3104420304,-0.0427394696,-0.0521648675,0.0985169932,0.0859911889,-0.1160690188,-0.3053369224,0.0906663463,0.2317037135,0.2590830624,-0.5257883072,0.228161633,-0.5218098164,-0.1893330365,-0.1227461696,0.0145789841,-0.3121325076,-0.400919795,-0.0026238887,-0.0152529413,0.1822064221,0.0625660196,-0.2080118507,0.4678928256,-0.1262197644,0.0463774577,-0.1139295697,-0.156369403,-0.2914364934,0.0924053863,0.0868111625,-0.1507676095,0.26876387,0.0293926895,0.0568593442,-0.2252899408,0.0984303877,-0.0356405303,0.0268333107,0.2886548638,-0.2070231885,0.0631752908,-0.0600750931,0.0365455747,-0.0255814735,0.0490395986,0.1152663082,-0.2090692669,-0.0139637794,0.1001366302,-0.0165686682,-0.167108044,-0.3912502229,-0.1389485747,-0.1975909472,-0.0063309944,-0.0405979976,0.4335287511,-0.0328270309,-0.0438484624,0.0445838422,-0.0258184075,-0.5719744563,-0.4883940518,0.2505269349,0.0060659922,-0.1540320218,0.2757163644,0.2353700846,0.3690859377,0.0922650024,-0.7071725726,-0.1561724097,-0.2046680748,0.1184478998,0.1058258265,0.0241531953,0.2398289591,0.0240915995,-0.0268915929,-0.1140977368,-0.1467262655,-0.1291036457,-0.0374850184,0.3500134349,0.0910490453,0.0001158654,0.0416235328,0.9497952461,0.3943794668,-0.1414885819,0.1439827532,-0.0240650773,0.498060286,0.0316481814,-0.1973225623,-0.0223429315,-0.2218565941,-0.326559931,0.1050060615,0.0302491765,-0.0344939381,0.079308942,0.3164926469,-0.0557476692,-0.2625714839,0.1485172212,-0.3140106201,0.2633895874,0.1785900295,-0.0668597594,-0.2687122524,-0.1759432256,-0.3116210103,0.2126884311,0.2438860238,-0.0834522918,-0.0698425099,-0.1346961409,-0.3145463467,0.2725902796,-0.0989718363,0.2434129715,-0.2030457705,0.2065507174,0.3731815815,0.2971287668,0.2507959902,-0.5127696395,-0.0630538389,0.2290844768,0.0103920419,-0.0986902043,-0.1902768314,-0.042101305,0.1808192879,0.0313361324,-0.0466323085,0.292732805,-0.206123203,-0.1982197613,0.0727549717,-0.2074224651,-0.1631847024,-0.2685338259,-0.0736334324,-0.1776900142,-0.003478274,-0.0291659385,-0.159392938,-0.2659488618,0.0489089489,-0.2242682129,0.2267317176,0.1672305763,0.0370891541,0.2579162121,0.1197768152,0.069313854,0.2803872526,0.0117839426,0.2370014042,0.479149133,0.0039309044,-0.1519687176,-0.0573556237,0.0520149618,0.1470309198,0.2802855372,-0.359736681,0.2402039766,0.1647098362,-0.0412786938,-0.4268030226,-0.1556103826,0.3061187565,0.2207678854,-0.2533014119,-0.6095197797,0.5321125388,0.1871464103,-0.2384274751,0.2558771074,-0.0618053973,-0.2736566663,-0.0648569614,0.0623318963,0.8591033816,0.0976355895,0.1440969706,-0.0029542181,-0.1681729853,0.5544694066,-0.2190593928,0.2768955529,-0.1044926643,-0.1345790923,-0.1915024668,-0.052045051,0.3022085428,0.2040858865,-0.1602407247,0.3003246486,-0.2708013356,0.4485932589,-0.1492143124,-0.1534205228,-0.1431580484,0.0090884203,-0.0060550324,0.1852538437,-0.1069364101,0.3841551244,-0.1476826817,0.1969138533,0.2623200715,0.1741456836,-0.4287788272,-0.2461885512,0.0591168702,0.2823919952,-0.1715060622,-0.3929845989,0.4317492247,0.3194620907,0.2764995098,-0.053546086,-0.2857945859,0.2667242885,0.063453421,0.1829374582,-0.2152649462,-0.0134184258,0.2422488481,0.1089961752,-0.404791832,0.0933835208,-0.0107914116,-0.2129336149,-0.0464651473,0.4137536883,0.0230957959,-0.5987148881,-0.1530689001,0.2957655489,0.2590725422,0.1559877992,0.155702889,0.1950618327,0.1615442932,0.5383930206,-0.3727496564,-0.3319196105,-0.0751143843,0.3418464065,0.4800575972,-0.1337217838,0.5248402953,-0.1896690577,-0.3283007145,-0.2143830061,-0.1198878214,0.0537749827,-0.1673796475,0.0300609469,0.1847631931,0.4469096065,0.0859865472,0.0942791253,0.0386935398,0.1691406071,-0.2848289311,-0.1499036402,-0.0869991854,-0.023355335,-0.0660739541,0.4195612669,-0.1543795466,-0.2241765112,0.1199375018,-0.0192358308,-0.2836767435,0.2019927204,-0.1278946549,0.0934849083,0.0326919556,-0.1223993376,0.1731600314,-0.2500205934,-0.0944233462,-0.1227531508,-0.3531879783,-0.1902602166,0.1031416431,0.190219298,0.064419277,-0.284771204,0.0072887596,-0.224674806,-0.0313429162,-0.114037022,0.2164696604,0.3175140917,0.0416080095,0.2557589114,-0.1455190182,-0.1382402033,-0.0408191197,0.0162845086,0.0620762594,0.3474431932,0.3701218665,0.2468786389,-0.0009160005,0.1782296151,-0.2215669155,0.2649018168,0.0477011018,0.0994335189,0.180384323,-0.4075655341,0.1558149308,0.2178879678,0.3242298365,0.2714052796,-0.2259005904,-0.3062483966,0.0838246793,0.1407383531,-0.1409043819,-0.2209397256,0.2345395386,0.3185212314,-0.1707043797,0.1732645929,0.4352169931,-0.099724479,0.3946253657,-0.0300415233,0.626105547,-0.5539909601,0.37934497,0.0789631456,0.069297716,-0.1200839207,0.3651635051,0.4278164804,0.0224862844,0.5699408054,0.0848642364,0.156887427,0.5307418704,0.2910475731,-0.388859272,-0.9131240249,0.4202112556,0.4402659535,-0.5396683812,0.1072005257,0.0947705954,-0.0534236617,-0.1686119139,0.1495589763,-0.1600314528,0.2045980543,-0.2233534902,-0.0710277185,0.4242421687,-0.1915762275,0.2048639208,0.065175429,-0.062195491,-0.0769083723,0.0825557336,-0.1635192484,-0.1894571781,-0.3350575268,0.3617977798,0.0639959946,0.3503903449,-0.0971988589,0.1679084748,0.0634869412,0.0434228256,0.2047022879,0.4444728196,0.5200179815,0.0289697219,0.1914118528,0.2181697935,0.0260916967,0.0139706451,-0.1626804173,0.2969217598,-0.1348585337,-0.0214275476,0.1584324092,0.1294741184,0.0683175623,0.0819698423,-0.0369922481,0.1760775149,-0.4534128606,-0.0189833213,0.1762523204,-0.0180624221,-0.0692690983,0.2122768611,-0.2905001938,-0.144781813,0.7292047739,0.0316708088,0.2395653129,-0.2009863555,0.0622640774,-0.1819661707,0.4538908303,0.3087110519,0.2311467379,-0.1840930134,-0.0565365925,-0.6199725866,-0.0325406678,-0.2790194154,0.1169055402,0.1015521064,0.1611326039,-0.1948215216,0.0374085344,0.0814649835,-0.1598072946,-0.0768684521,-0.0674438998,-0.163355276,-0.0760748684,-0.0253170338,-0.0049998309,-0.1898803413,-0.2545613348,0.2036049813,-0.028941853,-0.0221752897,-0.1108200401,-0.0322997756,0.0367441513,0.0414098501,0.4271948338,0.2819315493,0.404743731,0.3656843603,0.1175662577,-0.250697881,-0.1911176592,-0.1692751497,0.2737903595,0.0763390139,-0.1567207724,-0.4409970641,0.0766617283,-0.1389280707,0.1922304481,0.0059981667,-0.0247931648,0.0739348456,-0.2451407462,-0.0166010391,0.1591039598,0.1393741071,0.2356667817,-0.0348674953,0.2864537239,-0.3084515631,-0.0712316707,0.2981557548,-0.2871443629,-0.2383147627,-0.1336086094,0.3350433409,0.0122034708,-0.2180282772,-0.531155467,-0.0422041491,0.5525933504,-0.3209537268,-0.0749308318,0.1568257958,0.0607789867,0.1323459744,-0.0880957395,0.3794916868,-0.0187160093,-0.0245170649,0.0621992946,-0.3847378492]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"That makes total sense indeed !\r\nI think we can do the change","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":13,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nThat makes total sense indeed !\r\nI think we can do the change","embeddings":[-0.03455019,-0.0147938421,-0.138871327,0.1486412734,0.3467749059,0.1663145125,0.2095449418,0.0184292626,0.1497073621,-0.1232775971,0.1502337903,0.0608753413,-0.2619658113,0.0266031548,-0.0535789505,0.4429046214,0.1626019329,0.1055143028,-0.077376157,0.0023205795,-0.0461694226,-0.0181370284,-0.2040186822,0.0004582883,-0.4875939786,-0.0888759047,0.1054878831,-0.0982578173,-0.0416800752,-0.6186156869,0.4342827201,0.3009496033,0.0873402953,0.2949176133,-0.0001131926,0.0165395681,0.2299327552,-0.0200202093,-0.2306292355,0.0702981427,-0.533420682,-0.4411410391,0.2137729675,-0.0746457949,-0.0468381792,-0.1602111161,0.0167643968,-0.7276179194,0.4109701216,0.3747263253,0.2619732022,-0.1324798912,-0.2851992846,0.1013571769,0.0566277727,0.0908736661,-0.1052105054,-0.0983291864,0.1983121037,-0.0774817169,0.0027197448,0.3598616421,-0.099390395,0.2863802612,0.5238748789,-0.0048457757,-0.255230695,-0.3250645995,0.3102321625,-0.059833426,0.9308157563,-0.3203135729,-0.1747311056,-0.3114352524,-0.0364902355,-0.3494378924,0.2042980939,-0.08270742,0.0800688043,0.2081457227,-0.5194151402,-0.3289479613,-0.0809053555,-0.1129984185,-0.2162674367,-0.0671373755,-0.3081649244,0.0517842919,-0.0885095298,-0.0624832548,0.4980713725,-0.41317904,-0.0025784958,0.1555081308,-0.4899553359,-0.0998102203,0.16408059,0.4158254564,-0.1229362339,0.1979296505,0.05868496,0.0573499873,-0.0846228525,0.0060948292,0.3720418513,0.3438482285,0.0945631787,0.1824902743,0.1579174995,-0.182393536,-0.1667100489,-0.1692570597,0.1789198965,-0.4318241179,0.565854609,0.1086699218,0.2776173949,-0.3017982543,-0.3192349076,0.1147619933,0.0482082628,-0.1614865512,0.1093379855,0.1229516566,-0.0315505713,-0.0299025644,-0.2115404308,0.0168891568,-0.2191731483,-0.2214260697,-0.2290153056,-0.1216325834,-0.2757024169,-0.0321235023,0.2050339729,-0.1819254011,-0.0036441816,0.3446848094,0.1499624401,-0.2178937942,0.1231318563,-0.0435689203,0.193503052,0.2639785409,-0.3555396795,0.1971267164,0.4169443846,-0.0173726957,-0.2284673154,0.3116857111,-0.3756107986,-0.4972402453,0.04335678,0.1679980457,-0.5166566968,-0.2079642415,-0.1030537933,0.0450875722,0.3863268495,0.0693372786,0.1596076787,-0.1276215315,-0.1729646176,-0.2505604327,0.0417078733,0.3511242867,-0.6129997969,0.0846285075,-0.2809143662,-0.0469953194,0.1316187382,0.2888280153,-0.4532000422,0.1468031108,-0.1379461586,-0.2152214497,0.0629354045,-0.1855588108,-0.4398142397,0.1060394421,0.1643172652,0.3207904398,0.1121058837,0.065936543,0.1155766994,0.0136937257,0.0073259114,-0.0837385803,0.1065610424,-0.1991099715,-0.1599497795,-0.3544819355,0.2656930089,-0.1645237803,0.076752767,0.1455747336,0.0132258469,-0.1599065214,0.0127401063,-0.0784296542,0.0749644935,0.3082188368,0.1546087414,0.1749906838,0.0860967934,0.0404105149,-0.4320345521,0.3331376612,0.184808597,-0.3447685242,-0.1878937483,-0.2888640165,-0.2012565732,-0.0674480498,-0.4215483963,-0.1489227563,0.104598932,-0.0700554401,0.0065854266,-0.1207851171,-0.2173055112,0.3896846473,-0.2646053731,-0.122856304,-0.1352908313,0.2276156694,-0.0002022734,0.238354966,-0.2278940231,0.1694687605,0.1042602435,0.0881888717,0.0257168896,0.2118017524,0.0800356418,0.4110904038,0.0717688426,0.3596872687,-0.0464937165,-0.1112140566,0.262347728,-0.0027538447,0.0699130818,-0.0355405472,-0.1469029039,0.3687850833,-0.1720052809,0.1506101936,-0.2259668857,-0.1835734546,0.1470349431,-0.0792780519,-0.0362565666,-0.2321133465,-0.0152807198,0.0248663351,-0.0126529457,0.3453137875,-0.2121369094,0.1584947854,0.3114901483,0.0920178443,-0.0326281376,0.0141492393,-0.0570825227,-0.3143423498,0.1790846586,0.3978915513,0.5105689764,0.1328355372,0.0202333592,0.0568740219,0.1222767606,-0.139988482,0.3418778181,-0.1145679578,0.0256613344,0.1245098859,0.088033244,-0.1024969369,-0.3082797825,0.0700362772,0.163352102,0.2660385072,-0.486610353,0.2012961954,-0.5197749734,-0.1306241453,-0.18676126,0.080916442,-0.2153603137,-0.3887993991,-0.0344297551,0.00048529,0.1308405697,0.0790657699,-0.1806554049,0.3265492916,-0.1071886569,0.0045367321,-0.1171158925,-0.1626316458,-0.1651272178,0.100165911,0.1011125967,-0.200261429,0.2353281826,-0.0567039512,0.0331017524,-0.2998912036,-0.0285752937,-0.0611394756,0.0882371068,0.2746521831,-0.070799686,-0.0163356867,-0.0060690045,0.0029168038,-0.0370006263,0.1118388698,0.1148406193,-0.1382845789,-0.0434656888,0.1105080172,-0.0298351888,-0.1851750016,-0.4213707149,-0.1562539637,-0.1734531373,0.0418298244,-0.0643082634,0.4752353132,0.0571163893,-0.0568747669,0.0533649847,-0.0142008355,-0.5257066488,-0.4446227252,0.2137199193,-0.0680302456,-0.0406988561,0.2123416066,0.1608870029,0.3784689307,0.0468977578,-0.6304756403,-0.0359384492,-0.1784816533,0.0870537236,0.0828940496,-0.0005229731,0.2080489993,0.1039546803,-0.04966373,-0.1120039672,-0.0907804295,-0.2079023272,-0.1046424136,0.3226632178,0.0617637523,-0.0051495852,0.1053937525,0.9334228635,0.3847083151,-0.1217864081,0.1473627985,-0.0855629593,0.5959040523,0.0108502666,-0.3384212554,-0.11083024,-0.224912867,-0.3940272927,0.0198065639,-0.010373909,0.0267241821,0.2016488165,0.3435433507,-0.0467997827,-0.2882842124,0.1301000118,-0.393134743,0.3005768955,0.1689594239,-0.1172653362,-0.2971144021,-0.1292117834,-0.3510816991,0.1541136056,0.1952911764,-0.0997291878,-0.2233729511,-0.2424323559,-0.3320572674,0.2557167709,-0.11954844,0.1719949096,-0.1618530601,0.2650561631,0.3114858568,0.267668128,0.3922453225,-0.4935994446,-0.0796867684,0.2288913876,0.0367283672,-0.0829264149,-0.1470775008,0.0580843389,0.3810375631,0.06419909,0.0502194464,0.2478955984,-0.2434576452,-0.1333393753,0.1113604456,-0.1368133277,-0.1319113076,-0.2253515124,-0.0817769021,-0.2324474901,-0.0376838818,0.0092505766,-0.0790520236,-0.2510252595,0.0593320355,-0.1529600322,0.161991924,0.1513072252,-0.0175558608,0.2782713771,0.1633750349,0.0203611944,0.3279664516,-0.1296378523,0.3107233346,0.4954526424,-0.0663075298,-0.1678617448,-0.0866028592,0.0549437702,0.1543833166,0.2875729203,-0.3686639369,0.2665334046,0.0494387262,-0.0699947625,-0.3320601881,-0.1258289367,0.3047307134,0.2599732578,-0.299790293,-0.6187777519,0.4312044084,0.1956469864,-0.2415124178,0.0614468493,-0.0282334685,-0.3354884982,-0.066390872,0.1044163033,0.6944807172,0.1040131599,0.1699845791,-0.0801120922,-0.0463144928,0.4731561542,-0.2602503598,0.2527127266,-0.1356863976,-0.0924410224,-0.1281187385,-0.0629291385,0.4217448831,0.2091314048,-0.1656903327,0.3344279528,-0.3537070751,0.4707083106,-0.2273286283,-0.0691240877,-0.0925846919,0.0506759584,0.0032890905,0.2127732486,-0.0353476889,0.3638131618,-0.1186719611,0.200530827,0.2367412448,0.1378042251,-0.3743269145,-0.2772500813,0.0704605281,0.2513110638,-0.1309867948,-0.3703690469,0.4281009138,0.3540003002,0.3355400562,0.0259957835,-0.1997558475,0.2408926934,0.2147402912,0.2037094533,-0.1434255689,-0.0279739406,0.1965621412,0.1342436224,-0.3343711197,0.153423205,-0.0477450937,-0.273034513,-0.0018337984,0.3558602035,0.0541085899,-0.5879560709,-0.1468234807,0.2811291516,0.2099912614,0.1146701723,0.1943554282,0.1338423938,0.1811138541,0.4862014055,-0.4472766221,-0.2799202502,-0.0514795631,0.2803092599,0.5611675382,-0.1329568624,0.4843144119,-0.2270376831,-0.2945597768,-0.2380455285,0.0054353508,0.0759276673,-0.1477794945,0.1142512336,0.0802929103,0.3738029897,0.1422393322,0.0113793388,0.0951904953,0.0934043974,-0.1997875422,-0.1263113618,-0.1966894269,-0.0276525207,-0.0180741232,0.4218913317,-0.1365539581,-0.1179774851,-0.0039709513,-0.0307819247,-0.3128466606,0.1646522284,-0.1319115758,0.1118609533,0.072494626,-0.0555324666,0.1041595042,-0.3033548892,-0.020023061,-0.1502714902,-0.3637976348,-0.2214694619,0.0956357047,0.1371311694,-0.0140836732,-0.2484774292,0.04851266,-0.1768335402,-0.0595012009,-0.1115428284,0.1401793957,0.3946370482,0.0736429468,0.2450297624,-0.1630962342,-0.1917691678,-0.000529821,-0.1228862405,0.0635180473,0.3433628976,0.3169111311,0.3575587273,-0.0828785077,0.1345311403,-0.2419004142,0.2756068707,0.0367012955,0.130525142,0.1511329859,-0.3299919963,0.1696421057,0.2754589617,0.3543921113,0.1914708614,-0.3189962804,-0.1959727407,0.0752073675,0.1775970757,-0.1508308202,-0.2385061532,0.202519536,0.2176849693,-0.1203488857,0.1946585923,0.5315614939,0.0310010016,0.5700542331,0.009639929,0.6164848804,-0.4582871795,0.5060157776,0.0911132246,0.0165002048,-0.1536714435,0.4479502439,0.4311118722,0.0225328617,0.6574629545,0.0347627699,0.1810201854,0.4646798074,0.2123666108,-0.4065099955,-0.7890630364,0.4118677974,0.4221393466,-0.484388113,0.124790132,0.0929044038,0.0578513332,-0.1298530549,0.150941208,-0.0717293993,0.207966134,-0.1792134047,-0.1294819266,0.4117643237,-0.2030182928,0.1183642671,0.0574765578,-0.1050736308,-0.056201797,0.0959202275,-0.2050378174,-0.1542004198,-0.3847836852,0.5370151401,0.0183359869,0.3984482586,-0.1324335486,0.1580950171,-0.0869883448,0.0723196566,0.2755470574,0.3706969917,0.4889999628,0.0875288248,0.2787902057,0.0252813399,0.0206554905,-0.0442627519,-0.1936439425,0.226250872,-0.1574830711,-0.0364830755,0.0995935723,0.1572615504,0.0048864172,0.1177123711,0.0322541483,0.1919691116,-0.4827860594,-0.022504719,0.2209869176,-0.0271745995,-0.0198364519,0.2180216759,-0.3062412143,-0.277274549,0.7084780335,0.0746294633,0.2999086976,-0.2239238471,0.0919123739,-0.1172775701,0.3570028841,0.1163681075,0.1515190005,-0.2021076679,-0.0350086354,-0.6460978985,-0.0998864546,-0.2353311777,0.1814557314,0.1474145204,0.1088811606,-0.2161194831,0.0545687191,0.1149785444,-0.2125983685,-0.0945838988,-0.1358339936,-0.066582188,-0.052700635,-0.084394075,0.0007061761,-0.1457008868,-0.2065367252,0.2211224586,-0.0052862703,0.0305016618,-0.1791211665,0.0161176845,0.105743818,0.0547240414,0.4212237597,0.1778118014,0.3382136822,0.2721798122,0.083602488,-0.2898478508,-0.25389871,-0.1843435019,0.2502782643,0.1208034158,-0.0477865785,-0.5011557341,-0.001980132,-0.1684196144,0.0933073983,0.0463407747,-0.107870549,0.024808798,-0.0940984637,-0.0132636419,0.1736363769,-0.0207990538,0.26593104,0.0069398782,0.3325314522,-0.3395596743,-0.0875209719,0.2996442914,-0.202181384,-0.2556758821,-0.0377801955,0.2093232721,0.1057971939,-0.3124691844,-0.4907820821,-0.0015977659,0.5052515268,-0.2824791074,-0.0170078818,0.1524447203,-0.0266486164,0.0687652826,-0.0720192268,0.3035476208,0.0087688686,0.0236688778,0.0506304577,-0.3298139274]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"I have another question about caching, this time in the case where FORCE_REDOWNLOAD is used to load the dataset, the datasets cache is one directory as defined by HF_HOME and there are multiple concurrent jobs running in a cluster using the same local dataset (i.e., same local files in the cluster). Does anything in the naming convention and\/or file access\/locking that you're using prevent race conditions between the concurrent jobs on the caching of the local dataset they all use?\r\n\r\nI noticed some errors (can provide more details if helpful) in load_dataset\/prepare_split that lead to my question above. \r\n\r\nLet me know if my question is clear, I can elaborate more if needed @lhoestq Thank you!","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":115,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nI have another question about caching, this time in the case where FORCE_REDOWNLOAD is used to load the dataset, the datasets cache is one directory as defined by HF_HOME and there are multiple concurrent jobs running in a cluster using the same local dataset (i.e., same local files in the cluster). Does anything in the naming convention and\/or file access\/locking that you're using prevent race conditions between the concurrent jobs on the caching of the local dataset they all use?\r\n\r\nI noticed some errors (can provide more details if helpful) in load_dataset\/prepare_split that lead to my question above. \r\n\r\nLet me know if my question is clear, I can elaborate more if needed @lhoestq Thank you!","embeddings":[-0.1976588219,0.0997744873,-0.1261471659,0.2399428338,0.0289438497,0.0434310474,0.4564794302,0.0984457806,0.4220702648,-0.2127026469,-0.02498319,-0.0400107726,-0.0365918837,0.0134087205,-0.1342043132,0.4030806124,0.2076878548,-0.0396345407,-0.024214901,-0.1413462311,-0.1500361115,-0.0289533883,-0.0475723408,0.0606783032,-0.5617077947,-0.0671415403,0.0013730753,0.1623043418,0.1725726724,-0.4235329926,0.3519337773,0.5294938087,0.0670450926,0.4914126098,-0.000113091,-0.0403809808,0.1958641857,-0.0477524549,-0.2622745633,-0.1576190591,-0.5350163579,-0.4387889802,0.1720770895,-0.0988316759,0.0574336424,-0.1514247209,0.0411747545,-0.7131090164,0.4519171119,0.1427449733,0.2041774243,-0.1091207042,-0.4887529612,0.0160306338,0.0962762758,0.3353848159,0.0161248278,0.1014839038,0.2383863479,-0.3198069036,-0.3291200399,0.2688158154,-0.0138629032,0.3965907693,0.3954433203,0.0086095035,-0.2763716578,-0.2745757997,0.1655485034,0.0608960241,0.7798292041,-0.2836916447,-0.3395363092,-0.4090892673,-0.1475395113,-0.179307729,0.269965589,0.0716021284,0.0429681391,0.084578447,-0.3964806497,-0.2547957301,0.0036561524,-0.2672636509,-0.0383914709,-0.0404894166,-0.1567974687,0.08330708,-0.0747073144,0.0514648147,0.6233451962,-0.5045340657,0.1790409088,0.1976583749,-0.461689353,0.0237274785,-0.0364603847,0.391843915,-0.0540191829,-0.0245083682,-0.1536819339,0.0867669061,-0.1915480793,0.1589517742,0.4464426339,0.1503168792,0.1327717155,0.2022628486,0.2322850674,-0.3030024469,-0.3496911526,-0.0695606917,0.0665362924,-0.4245012701,0.2996263802,0.0340163559,0.1059436426,-0.3955444396,-0.2909142077,0.0781235546,0.1319535226,-0.278221339,0.0458954424,0.1907658726,0.0533931516,0.0282913912,-0.251444906,0.083957538,-0.3625871241,-0.2256822139,-0.2220804542,-0.1252384633,-0.2524294257,0.1785207093,0.2062439173,-0.2326714993,0.0402708203,0.3950311542,0.0962525681,-0.1698834002,0.1404356211,-0.1678614914,0.2335447967,0.3442668915,-0.2546127737,0.2355287969,0.3324483335,0.0308906045,-0.1443060786,0.2100905776,-0.4729039967,-0.4184192717,0.1001290902,0.1033481881,-0.4051647186,-0.024724545,-0.1482537091,0.0726757571,0.5164402723,0.0497851707,0.0943618119,-0.2699440718,-0.2512399554,-0.1872328371,-0.0906452164,0.5964114666,-0.6261852384,0.0243644491,-0.1429244578,-0.0997800827,0.0721887052,0.3109852076,-0.4317257702,0.1771506667,-0.2471140921,-0.0300256945,0.156193763,-0.2819254994,-0.2977422774,0.2853240967,-0.0048165028,0.2871124446,0.1105008647,-0.0748473853,0.0774108246,-0.2064444572,0.0731695145,0.0241682958,-0.0004380147,-0.1278619617,-0.2169781923,-0.2795419097,0.1265176088,-0.0591096692,0.1558092833,0.1283837259,0.1315277815,-0.0641909987,0.1397758573,-0.042271547,0.1320656836,0.3305953741,-0.0562164523,0.2266031504,0.1514285058,0.2036530674,-0.6013004184,0.4022013545,0.0565555915,-0.4018244743,-0.0155669907,-0.19565323,-0.1518979967,-0.1913250834,-0.4529682398,-0.0661183074,0.0631238446,-0.1243939474,-0.0742722526,-0.2170352936,-0.2007468343,0.8151642084,-0.3109962046,-0.031359341,-0.0666529536,0.2341920882,-0.065775536,0.2016776949,-0.1721449941,0.0347544812,0.2445912063,-0.0093639335,-0.0629991591,0.3307782412,0.2265687883,0.5461758375,0.1839093864,0.2227417678,0.1407313198,0.1049825251,0.3516898155,-0.2002376169,0.1083542332,-0.1767371595,-0.2546150982,0.3825460374,-0.3531452715,0.2602759302,-0.1692652553,-0.2800016105,0.0722134933,-0.0253565293,0.057882674,-0.0620063394,0.1634283662,-0.0544249155,0.2642069757,0.215580821,-0.1668545306,0.1455173194,0.198378399,0.0810150802,-0.1046497077,-0.1036524251,0.0311687831,-0.3026696742,0.1909436136,0.4321049452,0.6589485407,0.0841457546,0.1966480017,-0.0270573683,-0.0880918503,-0.2778379321,0.1220179349,-0.0720074773,-0.0023052332,0.2029411942,0.0612138882,-0.1524753571,-0.3158774972,-0.0474574044,0.289884001,0.0772339255,-0.4448517859,0.1399446577,-0.2746499181,0.0384695977,-0.1693123728,0.1227735728,-0.3491450548,-0.3374187946,-0.0173502117,0.1677834988,0.0397741683,0.023976896,-0.3611675203,0.3172857463,-0.3596293926,-0.0370979793,-0.1784277409,-0.3712448776,-0.1486219168,0.0270703863,0.2050141543,0.0677456036,0.2444338351,0.0925387964,-0.1358068585,-0.1985537261,0.2229715288,-0.0905946866,-0.0714234337,0.1459326744,-0.3722260296,-0.0143959671,-0.041953478,-0.0836177021,0.001237568,-0.0441604704,0.0252002217,-0.2760803103,0.076716952,0.1379478723,-0.0329541266,-0.126061365,-0.3585487604,-0.297344327,-0.1416451782,-0.0665722191,-0.0133502455,0.381219238,-0.1233073846,-0.109098427,0.2232977748,-0.2107465267,-0.5925027132,-0.4445326626,0.0657245293,-0.1588762552,-0.1897361875,-0.0171294678,0.1173106879,0.1880111396,0.1741952747,-0.5063966513,0.0306788776,-0.0522000492,0.1078686342,0.1149188355,0.0119206095,0.2383215278,0.0561734252,0.0068713585,-0.1866166741,0.0011006157,0.0650862604,-0.0169898253,0.215507552,0.031028172,-0.1636013985,-0.1278615892,0.9876939058,0.2820958197,-0.0200459696,0.2478423119,0.2097233236,0.5658015609,0.0727661252,-0.2682198286,-0.0198029596,-0.3216585219,-0.4030307531,0.0986347124,-0.0202114023,-0.0255670156,0.0158950444,0.2516833842,-0.1507887542,-0.3202366531,0.0968192294,-0.401637882,0.2247706354,0.1057193056,-0.157984972,-0.2311588228,-0.0638039112,-0.2034137398,0.1252926439,0.2671797872,-0.1133419871,-0.3760961294,-0.0122120418,-0.150981456,0.2174095064,0.0030068979,0.2297933847,-0.2213808745,0.1541533023,0.242559135,0.2520405352,0.5395837426,-0.4943318665,-0.0120087722,0.3081458211,0.0935060382,-0.0350710377,-0.3314194083,0.0597195439,0.3186377585,-0.0601289533,0.1686540842,0.1850640178,-0.3548405468,-0.2019772828,0.1204962209,-0.1876204014,-0.3386184573,-0.1475478858,0.0515872426,-0.0989235342,0.1706444323,0.0041627516,-0.1042352915,-0.4475524724,0.1424416751,-0.0375594832,0.1645590216,0.1822348833,-0.0063507338,0.2474911362,0.0672517121,0.1224882454,0.475134939,-0.1162652448,0.391204536,0.513235569,-0.2522020936,-0.0745842457,0.0655106977,0.0436777771,0.2390205562,0.4920995235,-0.1698341072,0.0112334071,0.049134966,-0.1993610859,-0.3951824009,-0.0089399246,0.0891337246,0.1115366444,-0.2740194499,-0.5477327108,0.4944396019,0.1882710308,-0.2931149006,0.2271384448,-0.0794415176,-0.2718234062,0.1646752954,0.0818441063,0.7546726465,-0.0053866003,0.0857155323,-0.2928617597,-0.1523361206,0.361589998,-0.4735431671,0.1916038692,-0.2058301717,-0.1356478631,-0.0661596134,-0.1759931892,0.3035153151,0.4233146608,0.0698853359,0.4099542499,-0.1637893319,0.3477377594,-0.3406014442,0.1299733371,-0.0802544877,-0.097459428,0.1579228044,0.1904121637,-0.0052903593,0.4484213591,-0.1342687011,0.2542358637,0.1499658823,0.2168230712,-0.3202621937,-0.0961832553,0.0148670068,0.3158015311,-0.3518647254,-0.4861323833,0.3081658185,0.2944430709,0.4739224911,-0.1809300482,-0.1067484766,0.2365146428,0.1502017081,0.3250730634,-0.1963798255,0.015832739,0.3516878188,0.1532214582,-0.2486715317,-0.0203191154,-0.1314673126,-0.1292048097,-0.053615462,0.3481574655,0.1101598665,-0.5068115592,-0.0620738901,0.3035106361,0.2639786005,0.1314888895,0.126232639,0.1465285718,0.1877068877,0.4004671574,-0.4862881899,-0.269787997,-0.1348173767,0.3867883384,0.4538158178,-0.3098533154,0.4357559681,-0.1971489787,-0.2967898548,-0.1681284308,-0.0168313961,0.2014452666,-0.1655529588,0.0588928498,-0.0716054961,0.0945137367,0.108180359,-0.0181551687,0.1278892308,0.1276206076,-0.0724169537,-0.1381508261,-0.2885971069,-0.0941319168,-0.1662011296,0.3729237318,-0.0076631485,-0.2614076436,-0.0035443401,0.1206896529,-0.2857523263,0.0724877119,-0.1828872561,0.1312214881,0.0144410226,-0.066926986,0.2288741916,-0.1978757232,-0.0724170431,-0.0604826435,-0.3882534504,-0.1609352231,0.1040430218,0.1701238751,0.0574103333,-0.1091029122,-0.1158858985,-0.2992547452,0.1175630838,-0.149661988,0.2170011401,0.3762131929,-0.007314784,0.2057655156,0.20460926,0.0075703785,-0.0796658471,0.0969998688,0.1692903638,0.2467493117,0.2833224833,0.2763617039,-0.0494927093,0.1230390742,-0.2391476929,0.1309688836,0.2271893173,0.1936818808,0.0724753216,-0.2769876122,0.1188425496,0.0833599865,0.4556765854,0.362321943,-0.0685254708,-0.3023668826,0.2224196941,0.1409721673,-0.0703122541,-0.2093816698,0.4214825034,0.2121277601,-0.1001908258,0.2593785524,0.6226680279,-0.0030641619,0.2651771605,0.1284811497,0.6269985437,-0.3409231901,0.31788674,0.1241050139,0.140591681,0.0725372955,0.4786079228,0.4627037644,0.0455111563,0.4786471128,0.1187354028,0.2241659313,0.3307625651,0.2659928799,-0.2395766675,-0.7666537762,0.3386404812,0.4435625374,-0.3572900891,0.1059565693,-0.1620753855,0.0412574708,-0.2385780811,-0.022323709,-0.0198922306,0.1341738403,-0.1757948548,-0.1652769744,0.5611439347,-0.2321899235,0.0724461675,0.0717932656,-0.0872568414,0.031054331,0.2954609394,-0.1402337849,-0.0206406321,-0.1417829543,0.3359812796,0.0895525143,0.2814581394,-0.1195767447,0.1500709653,0.0193081591,0.0229670033,0.2433671355,0.3205235004,0.4312110245,0.073075518,0.056789171,-0.0419804417,0.0803494528,0.0306584127,-0.1244742796,0.2402953506,-0.2892828584,-0.1411246359,0.2796352804,0.1205690801,0.0544575416,0.0997401997,0.0347342156,0.0522233993,-0.2454397529,0.0413270816,0.1616995633,-0.1541542411,0.0413748324,0.0872313902,-0.2158909142,-0.213801682,0.6434292793,-0.0411731787,0.3312615454,-0.087418817,0.0671206564,-0.0899093673,0.4424495995,0.3044471741,0.0016845698,-0.2855187356,-0.0738333464,-0.6275610924,-0.0998570174,-0.339946866,0.309122473,0.0127602927,0.1506415308,-0.0677453727,0.0091883345,-0.0018432206,-0.115077883,-0.2683277428,0.16037184,-0.1871931255,-0.0789659321,-0.1476559043,-0.0946875215,-0.0910220742,-0.2468649149,0.1848932505,0.0806933641,-0.069491595,-0.0171885621,-0.0482560918,0.341827929,-0.0492819212,0.3211779594,0.0896967947,0.314481318,0.3501135111,0.1428947598,-0.3316807747,-0.253038615,-0.2174379826,0.1777808666,-0.0698350593,-0.0495416075,-0.6060162783,0.0150945904,-0.1568980664,0.1981024444,-0.098649323,-0.1057599187,0.103886582,-0.1709438562,-0.0705884919,0.2016068548,0.2013843656,0.3171714842,0.0810509473,0.4173973501,-0.4404121637,-0.0809613094,0.2961800098,-0.2702329457,-0.2628027499,-0.0478427298,0.3122309446,0.0949088037,-0.2099296004,-0.417320013,0.00378488,0.4315959811,-0.2807969153,-0.0492599085,0.0411787629,0.1083437726,0.0290373601,-0.0964805633,-0.0139839873,0.0313168541,-0.0698076338,0.1674192101,-0.1769052148]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"I got another error that convinces me there is a race condition (one of the test files had zero samples at prediction time). I think it comes down to the fact that the `config_id` above (used in the naming for the cache) has no information on who's touching the data. If I have 2 concurrent jobs, both loading the same dataset and forcing redownload, they may step on each other foot\/caching of the dataset. ","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":74,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nI got another error that convinces me there is a race condition (one of the test files had zero samples at prediction time). I think it comes down to the fact that the `config_id` above (used in the naming for the cache) has no information on who's touching the data. If I have 2 concurrent jobs, both loading the same dataset and forcing redownload, they may step on each other foot\/caching of the dataset. ","embeddings":[-0.2517684996,0.1913679242,-0.1285429001,0.2252946794,-0.0270277169,0.077824086,0.3286910057,0.100526467,0.3222955167,-0.1650660783,0.1239636615,-0.1279569864,-0.1924358457,-0.0144361462,-0.0895964056,0.5366289616,0.1078998223,-0.0252203513,-0.0901686996,0.0365105681,-0.1266117245,0.0053364378,-0.0214975122,0.0335966013,-0.5114260316,-0.1194659993,0.0914082453,0.1005366817,0.0635159165,-0.4727329612,0.4316722751,0.4998078942,0.1185006723,0.5272336006,-0.0001120525,-0.0287346337,0.275087297,-0.0498848557,-0.1975311935,-0.0401539877,-0.5221555233,-0.4225539863,0.1734901369,-0.1565448493,-0.0192890484,0.1251882911,0.099964276,-0.6649633646,0.4458995759,0.2068891823,0.2239559889,0.0063834074,-0.3632511795,0.0565340295,0.1321967095,0.0316346288,0.1112322435,0.0179962404,0.2499786168,-0.276514709,-0.1764031351,0.3918348253,-0.1496852785,0.2700966001,0.3394783139,-0.0495601483,-0.1456187963,-0.2498754561,0.1793821454,0.0785242692,0.7550192475,-0.2804460227,-0.2590423226,-0.3573783636,-0.0003891959,-0.3408976793,0.2716992795,0.0059854067,-0.0621335283,0.1301233768,-0.5325371027,-0.0447265208,0.0544280186,-0.2020443976,-0.0876097754,-0.0325516425,-0.182489112,0.1206842959,-0.1120894924,0.0778149813,0.5639139414,-0.4694675207,-0.0030593129,0.1652616411,-0.5255541205,-0.0425236449,0.114429757,0.3191137612,-0.0718690082,0.0461115763,-0.0258326959,-0.0575423278,-0.0753804296,0.0962906182,0.4880084395,0.2864927351,-0.0163678918,0.1548945904,0.3341031969,-0.2074424922,-0.2292799056,-0.0197030753,0.1273514628,-0.4306164682,0.5709225535,0.0303970724,0.2376965582,-0.2874131203,-0.4723013341,0.1600819677,-0.0511509441,-0.1926122755,0.1289615929,0.1974543035,-0.0455315076,-0.0731103644,-0.0361734256,-0.0604148321,-0.3438555002,-0.2678798139,-0.2570454478,-0.1845753789,-0.2981892824,0.0259228144,0.142329514,-0.131467849,0.0474159718,0.3375694752,0.0919070169,-0.2276921421,0.1491729617,-0.1079981923,0.1522395909,0.4109433293,-0.2356166542,0.2177990526,0.2515305281,-0.0982941389,-0.1494981498,0.231740579,-0.4703750908,-0.3041459918,0.209346965,0.1603921801,-0.6245672107,-0.0508760177,-0.1154582947,0.1069257632,0.528937161,-0.0188324042,0.1880380958,-0.146504879,-0.1347687244,-0.1819501221,-0.0632875562,0.6424156427,-0.4017948508,0.0320385881,-0.0944020227,0.0119825583,0.0594302453,0.2395131588,-0.4469092786,0.1544790119,-0.148510471,-0.1909796596,0.1386921704,-0.1890893579,-0.4584935308,0.1903180331,-0.0515882634,0.3528974056,0.1084554121,0.1193268001,-0.0222294219,-0.1931741983,0.0405067056,-0.1391272694,0.0323301293,-0.0980474725,-0.1925178468,-0.4559210837,0.2004025578,-0.0915422142,0.1057980657,0.0321127512,-0.0470400043,-0.2174394727,0.0696061254,0.0122231357,0.0625416562,0.3167732656,-0.0313241109,0.0907533765,0.0103259739,0.0960560963,-0.6243484616,0.3583464324,0.1563860029,-0.3539857566,0.0139278984,-0.1468540579,-0.1040834561,-0.0771554485,-0.4389853776,-0.1530054808,0.0888962299,-0.0018820949,-0.0361587815,-0.1967172474,-0.1879837662,0.6546053886,-0.3568730056,-0.1490763128,-0.2713383138,0.1059929579,0.0194851961,0.0749570951,-0.2156205773,-0.0297371186,0.1446014792,0.1075262949,0.0449019,0.2335270643,0.0997379944,0.4624310434,0.0397954583,0.4178235531,-0.0245628506,0.1138097942,0.1684465259,-0.0570358597,0.1276118159,-0.1137056127,-0.0868541896,0.4415454566,-0.3481494784,0.2157451212,-0.2746936381,-0.2525871396,0.157769531,-0.0914995745,-0.0684843212,-0.0187276658,0.1852874011,0.0674142092,0.1038760766,0.2266269028,-0.3037698865,0.1307943165,0.1964209825,0.0860399902,-0.0355747789,-0.0536786467,0.0118191577,-0.1683056653,0.1427569836,0.4865145087,0.6019807458,0.1609792262,0.1027020812,-0.0308136325,-0.0692350566,-0.2510652542,0.179191038,-0.084663108,-0.0707233027,0.1648723483,0.0941823348,-0.0786523372,-0.2352936864,-0.002498521,0.2213748395,0.1878318042,-0.5408714414,0.2193582505,-0.1885175556,-0.064324744,-0.2079839557,0.0157936048,-0.2783078253,-0.271194756,0.029967973,0.2203122675,0.0512737148,-0.0009051731,-0.4029424489,0.3308380842,-0.2962324917,-0.0193091258,-0.1532669812,-0.3106408417,-0.1549325883,0.0505178161,0.2070084214,-0.0508817285,0.2605375648,0.0185471196,-0.0519024506,-0.0746047199,0.1386124641,-0.1986941397,0.0166718028,0.2831512392,-0.271815598,0.0186259579,0.0076696053,-0.1196956038,-0.0598488674,0.0348810554,-0.0434903726,-0.2097879499,0.0749027655,0.0885233954,-0.0312413871,-0.1403440684,-0.3080449998,-0.2072532028,-0.132946521,-0.0004844931,-0.0717660263,0.5000891685,-0.0366152897,-0.1307935268,0.1735211909,-0.0806697309,-0.5959798694,-0.6015761495,0.0083438754,-0.0080362139,-0.1165084615,0.0793122277,0.0809104815,0.33736974,0.2010284066,-0.4741949737,0.0586170219,-0.1161702797,0.1523581743,0.1011249572,-0.0133294221,0.3284638524,0.1294866651,-0.0322445258,-0.1425666064,0.0085591236,-0.0020877535,-0.0802742615,0.3332745731,-0.0510389507,-0.0800951794,-0.0817691311,1.0259200335,0.4451493323,-0.2207205594,0.1197505444,0.1114840284,0.5954015255,0.0992717668,-0.2981829047,-0.1415586025,-0.2105116397,-0.4468586445,0.0789374486,-0.0813032016,-0.0102540273,0.0256450679,0.3378879428,-0.2246461213,-0.3094711304,0.1535270363,-0.3852465749,0.211097911,0.1298295259,-0.1262348294,-0.2641899288,-0.1565332264,-0.3046491146,0.1566655636,0.2108722776,-0.129270345,-0.2809834778,-0.0529128797,-0.2109042257,0.2140878439,-0.1104004532,0.2288006991,-0.1527272612,0.1393655837,0.1894436926,0.1234557107,0.5261524916,-0.5408663154,0.018712353,0.3258866966,0.0366335139,-0.0141301546,-0.2502390742,-0.0214142688,0.2749581337,0.0569709092,0.1040941328,0.1127330735,-0.1625840366,-0.1295721084,0.038636297,-0.2036714554,-0.3574638069,-0.3348812461,0.021614423,-0.0356379189,0.2071159929,0.0349811725,-0.1706222892,-0.4067677557,0.0633658543,-0.1790162176,0.2599106729,0.2379893214,0.0964000374,0.1390655339,0.1510614455,0.1921012253,0.5285314322,-0.1943765134,0.3289651871,0.4249373078,-0.02883357,-0.0411116034,-0.0333525687,0.2387886047,0.1610037237,0.4802272916,-0.1635887772,0.1184633002,0.1118430421,-0.2244086266,-0.4298040867,-0.0539079271,0.1832758337,0.3433671892,-0.2981332242,-0.5379780531,0.4298603833,0.1825302988,-0.2888793945,0.2886286676,-0.2511372268,-0.3057352304,-0.0478732325,0.0606491081,0.9063932896,0.0238127075,0.0896112323,-0.1680405289,-0.2043424994,0.3515007496,-0.2665155232,0.2621555924,-0.1686121225,-0.1190141961,-0.0768342242,-0.1368049383,0.3444722891,0.2039647102,-0.0256587975,0.3815742135,-0.2514645159,0.3177064955,-0.2536364794,0.0545574613,-0.0312776417,0.0408976637,0.0482793041,0.2011442631,-0.0915900692,0.4211692512,-0.1481292993,0.1660478115,0.1732250005,0.2095961571,-0.3672165573,-0.0758556202,-0.1288067997,0.2412811518,-0.3847011924,-0.4085422754,0.2100903243,0.3001431227,0.2872572243,-0.0402122438,-0.0879063085,0.3305515051,0.1718912423,0.2496002316,-0.1143169999,-0.0921343789,0.2531345189,0.0928103626,-0.4002296329,0.1381380558,-0.1176001802,-0.0976189375,0.0476060994,0.405823946,0.1681503803,-0.4719682932,-0.0661570951,0.3432350159,0.2380196154,0.122524783,0.1486919522,0.21730946,0.1809560508,0.4563590288,-0.493185699,-0.3539918065,-0.0980028585,0.454185307,0.4964257181,-0.1159436032,0.4686361253,-0.3675578535,-0.1963132769,-0.1657242775,0.0218101908,0.2290471196,-0.1015461385,0.1191608906,-0.00793391,0.1827759296,0.1918762773,0.0799608082,0.1043664217,0.0297296271,-0.0993023515,-0.1074315459,-0.2284149975,0.0016842351,-0.0087534944,0.4354404211,-0.0215753112,-0.2398881912,-0.1233078837,0.0886602774,-0.3268211484,0.0402751938,-0.3119004071,0.1347809583,0.0895455703,-0.0904190168,0.1559416205,-0.3188300729,-0.0550835244,-0.1348194778,-0.4211334884,-0.1929088384,0.1011249349,0.1473773718,-0.0453899615,-0.1048486978,0.0359335765,-0.3042224944,0.115548566,-0.1196432859,0.1508763582,0.2795186043,0.0709541813,0.1298112422,0.0144620659,0.0089743286,-0.1445267946,0.0718488321,0.0618245378,0.0933989212,0.2500111461,0.2964651883,-0.0790922195,0.1362194717,-0.2720768154,0.1582303643,0.179656595,0.1432493776,0.0405362882,-0.2879132032,0.0595656671,0.2745726407,0.3348704278,0.3194316924,-0.1518950015,-0.2957105041,0.1598290503,0.18146725,-0.1841985434,-0.068137303,0.2696793079,0.291885525,-0.029099131,0.2303137034,0.4189582169,0.0286829118,0.4258458912,0.0884434879,0.6176128387,-0.4926784337,0.4518832862,-0.003626979,0.1652176529,-0.0140561974,0.4809515774,0.3415840566,0.1257572472,0.3665905595,0.0908314735,0.2661991119,0.4168555439,0.179576546,-0.3835105002,-0.7729428411,0.3268946409,0.4059698582,-0.4040574431,0.1575590074,0.0464886203,0.0945227221,-0.2413250059,0.0773721486,-0.1642310321,0.1319763511,-0.1541714519,-0.1720743179,0.465079546,-0.2134919763,0.1724028736,0.0140795568,-0.0729571953,-0.0196417551,0.2851194739,-0.1888890117,-0.1902803928,-0.3847339749,0.1992869973,0.1175816208,0.313601315,-0.2144120485,0.1406805664,0.170244351,0.0135516999,0.2077638954,0.4880630672,0.6151871681,0.0458032452,0.2123407274,0.0407727808,0.1567066163,-0.0211978648,-0.2275059819,0.2416978478,-0.1272059083,-0.1402416527,0.1282978952,0.1625059545,0.0109188044,0.0844229236,0.1475846916,0.1768210232,-0.3624552786,-0.1245504171,0.0180375762,-0.0987361968,-0.0387730785,0.1679379344,-0.223694697,-0.0650017634,0.570633173,-0.0086031193,0.1929667741,-0.1190561131,0.0842364579,-0.1275200993,0.451287061,0.196003288,0.0122875776,-0.2609120905,0.0184162017,-0.7642455697,-0.0648425147,-0.3708969951,0.2149638683,0.019599136,0.1942862123,-0.1366360635,0.027583899,0.0445959605,-0.1766609401,-0.1415237635,0.168988958,-0.2013436258,-0.0542339422,-0.0901088789,-0.189970389,0.0224072691,-0.2785151899,0.1121837869,0.0400693379,0.0102654956,0.0393891297,0.024368817,0.2430414706,0.1524775624,0.4254672229,0.1209515855,0.4215009212,0.2823090851,0.0887164697,-0.3945355117,-0.271237731,-0.2611302435,0.2192792147,0.0499732457,-0.0604159348,-0.5520471931,0.1047298387,-0.1174669787,0.1339379102,-0.1022370681,-0.2000134289,0.1163014695,-0.2459812015,-0.0709169507,0.3233185112,0.0463845395,0.3567880392,0.047515709,0.4030321836,-0.451166749,-0.0404338576,0.3104499578,-0.2887721062,-0.209782213,-0.0825660825,0.2860711813,0.0124396328,-0.2090423405,-0.448982805,0.0666356906,0.4374011755,-0.2764775753,-0.0816244781,0.1540213823,0.0156007502,0.0642776117,-0.0405627303,0.1321404725,0.0544446707,-0.1222036704,0.2334980816,-0.2742744386]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"We're using a locking mechanism to prevent two processes from writing at the same time. The locking is based on the `filelock` module.\r\nAlso directories that are being written use a suffix \".incomplete\" so that reading is not possible on a dataset being written.\r\n\r\nDo you think you could provide a simple code to reproduce the race condition you experienced ?","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":61,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nWe're using a locking mechanism to prevent two processes from writing at the same time. The locking is based on the `filelock` module.\r\nAlso directories that are being written use a suffix \".incomplete\" so that reading is not possible on a dataset being written.\r\n\r\nDo you think you could provide a simple code to reproduce the race condition you experienced ?","embeddings":[-0.196753189,0.1176110059,-0.0897247791,0.355255723,0.1577601433,0.1552672088,0.1879122555,0.1608638912,0.3102245629,-0.1881488115,0.1910189688,0.0179731119,-0.2734490633,0.0268451404,-0.2476358712,0.4049519598,0.1267285794,-0.0034802724,-0.1405780464,0.0179454051,0.0459334515,0.0447995812,-0.1037570462,-0.0012473409,-0.476983726,-0.2035485059,0.0959886834,0.1866528988,-0.0399648771,-0.5564548969,0.3019807339,0.4502322674,0.0678038895,0.5293136835,-0.000115094,-0.127984643,0.2920048237,-0.0531719662,-0.2578074932,-0.072666049,-0.5385324359,-0.4082248807,0.2471218407,-0.146911934,-0.0855540335,-0.0813922212,0.1438263208,-0.8633799553,0.5533196926,0.2262484878,0.1847568601,0.0490379669,-0.3496796489,0.0953061506,0.0084356172,0.0291794706,-0.0481614172,0.0714802593,0.4345098138,-0.2140529752,-0.1860693246,0.3216883838,-0.2358110845,0.2497453392,0.4829368591,0.0574803166,-0.1865171492,-0.2884808779,0.1342118382,0.0174988154,0.7710599303,-0.4393549562,-0.1225506067,-0.3344056606,-0.0687558502,-0.3336839378,0.3206148446,-0.0728345737,-0.0339725092,0.2561246455,-0.472356528,-0.0451729968,0.0078581031,-0.1658639163,0.0376529098,-0.0919709876,-0.1253417432,0.0373961031,-0.1179847047,0.1216066778,0.5495470166,-0.5747789741,0.0110015748,0.2291449308,-0.5444290638,-0.0410546325,0.0324345529,0.4762230217,-0.1241324171,0.1248387769,-0.0110148611,0.0557427444,-0.0400535017,0.1130376309,0.3680425584,0.2238933593,0.121543318,0.2314200252,0.2183250636,-0.1693474203,-0.2075107992,-0.0547740422,0.0730549321,-0.3701848686,0.5901768804,0.0753738359,0.1406158656,-0.3174619675,-0.3990462422,0.2899316251,-0.0197662972,-0.0911198407,0.1224696338,0.1468481123,-0.0552359708,-0.2533983588,-0.0159429125,-0.0094256988,-0.2785454988,-0.1450182348,-0.1850530058,-0.1182996035,-0.3526631296,-0.0679666027,0.1608492732,-0.1622107476,-0.0802540854,0.4727271199,-0.029606238,-0.1480693519,0.2859733105,-0.1099013165,0.0946576819,0.2131463587,-0.1637274921,0.3256706893,0.4162673056,-0.0792467371,-0.1719822437,0.3135366142,-0.3567288816,-0.2460056245,0.0298705455,0.1221244261,-0.5536221266,-0.0403471068,0.0136223137,0.0955767781,0.4750459194,0.0389860831,0.2495601326,-0.1382685006,-0.2923187017,-0.1906098276,-0.0443615206,0.731172502,-0.5330130458,0.1285101622,-0.1396491379,-0.0286400672,0.1445684135,0.2841942608,-0.2962820232,0.1323597878,-0.2115534991,-0.131062746,0.0573135279,-0.1184853688,-0.4381595552,0.174296245,-0.1678357869,0.4191295505,0.2183451355,0.302973032,0.0468535759,-0.0431203954,0.0522445664,-0.0761143863,0.0460267179,-0.2244427949,-0.1492369473,-0.3316139579,0.221999228,-0.1010166705,0.0453742519,0.0104914447,0.0069762175,-0.1455114931,0.1469559371,-0.0052159191,0.1594039649,0.3471348882,0.0670275539,0.2349253297,0.0162089616,-0.139325738,-0.4302188456,0.3824398816,0.1700233966,-0.3273444176,-0.0764160827,-0.1682446301,-0.274194032,-0.1506723911,-0.4196013212,-0.0648655742,0.0559284091,-0.1632685661,-0.0997350514,0.0063885967,-0.0581865646,0.7144823074,-0.3550771773,-0.1178708747,-0.1787383705,0.1654230952,-0.0695138872,-0.0650652945,-0.1282577813,0.016238397,0.1677399576,0.088434346,0.0401568003,0.1977181882,0.1437592059,0.4981746078,-0.0173988808,0.4018269181,0.0288784131,0.0421065651,0.1149366274,0.0580227301,0.1100393608,-0.0902630389,-0.2157793194,0.4076112211,-0.3188075125,0.0771591738,-0.2222145051,-0.1766803712,0.0708627626,-0.1201454252,0.0369205289,-0.053223297,0.326100409,0.0933621824,0.0394372642,0.2987610996,-0.1140406504,0.1542569548,0.3554931581,0.1300945282,0.1081680655,-0.0364470668,0.0031142693,-0.1081881672,0.0884905532,0.3828621507,0.4414940476,0.1244364381,0.1364682615,0.0507013127,-0.0579844825,-0.2591862381,0.3113619685,-0.1765068322,-0.0299759656,0.2807720006,0.0883563086,-0.1850343645,-0.2827694416,-0.1310247183,0.1508902758,0.2686749697,-0.414002955,0.189224571,-0.4363931417,-0.006558218,-0.2460869253,0.1756597012,-0.1134246811,-0.2515414357,0.0480956212,0.3083199263,0.1618435681,-0.0087755835,-0.363363117,0.1862072349,-0.0862537995,-0.0292842239,-0.1167898998,-0.2754992545,-0.1771163046,0.0540667474,0.2507048845,0.0493864864,0.2407139838,0.0357820429,-0.0555945002,-0.1726720184,0.0869023576,-0.1923737973,0.0038501304,0.3356839418,-0.1077344865,0.0945145786,-0.0795742273,-0.0116854757,-0.147387445,0.0740206838,-0.0783890486,-0.125515759,0.0772518516,0.0360645577,-0.1096629277,-0.1280042827,-0.308237344,-0.1841048002,-0.0727252215,-0.0917507634,0.01081663,0.4160496593,0.0477247052,0.0082806358,0.0806196406,-0.0969172567,-0.4676979482,-0.3725499511,0.1579973996,0.0714701563,-0.1243049502,0.0383487493,0.1377870291,0.1091660932,0.0639515966,-0.4974830449,-0.0647069365,-0.2150494456,0.1265069991,0.0788128972,-0.0567570329,0.3888295591,0.1091576368,-0.034947034,-0.2198407203,0.1228038892,0.0552483983,-0.2224344611,0.1876205802,-0.0696273968,0.1301476955,0.0980787948,0.9211750031,0.4421302378,-0.1490120441,0.1426526755,0.0367858633,0.5512974858,0.0714660138,-0.2881068885,-0.2727802992,-0.201275602,-0.5301855206,0.0737027228,-0.1222468168,0.1210196689,0.1622659862,0.256395191,-0.0245226901,-0.2634499371,0.0854231045,-0.3224430084,0.3055689633,0.1176087633,-0.1135855764,-0.1386606395,-0.159452796,-0.2369191945,0.3065062165,0.1802804619,-0.0642330647,-0.2745895982,-0.0641615838,-0.3204016387,0.1976513118,0.0092161708,0.1392200738,-0.3340420127,0.2803924382,0.3066825867,0.1070355996,0.5290564895,-0.4673745334,0.1154226437,0.2394134998,0.1314843893,-0.0981873125,-0.1250732839,-0.1551833749,0.3858980238,0.1222370043,0.1415533125,0.1297630221,-0.1207309067,-0.335288763,0.0299772061,-0.1702425927,-0.2717116177,-0.3003409803,-0.0513489954,-0.2229291499,0.2612918913,0.1674815416,-0.1090227216,-0.3220362067,-0.124456659,-0.1182028651,0.2126154602,0.2658466399,-0.012299954,0.2714411914,0.0407203771,0.2027820647,0.3392231762,-0.3385327756,0.2675183415,0.4162417054,0.0638293177,0.030167019,0.0778313056,0.0254294984,0.0431312472,0.3892775178,-0.2104763687,0.3602577448,0.31591326,-0.2228939682,-0.4493178427,-0.1846023351,0.18635571,0.3459112346,-0.2317754328,-0.5013896823,0.3324224055,0.2090015113,-0.2888940275,0.3288908601,-0.2253593504,-0.3025459051,-0.0679496303,0.0877480507,0.8110818863,-0.0809318274,0.1577160358,-0.1047106087,-0.3233383596,0.4059961736,-0.2327055186,0.2079770416,-0.2531996369,-0.2549883723,-0.1602258384,-0.2344596684,0.4729737043,0.1805520654,-0.1103876159,0.3075832725,-0.4579268396,0.3586802483,-0.1905742139,0.1103417277,-0.245228529,0.0999716893,0.0334637202,0.1664407402,-0.0294464082,0.3087554276,-0.1583581418,0.1298674196,0.0857040957,0.2419232577,-0.4965526164,-0.1260896921,0.0283941142,0.2636705935,-0.3742273152,-0.4050726891,0.4280455112,0.2767014205,0.3951699734,0.0871892571,-0.2225192636,0.3622687757,0.1929384172,0.1235759035,-0.1346281171,-0.1545104831,0.1473842412,0.13906461,-0.3111176789,0.0902867466,-0.0450944118,-0.1597726643,0.0432387441,0.3093547225,0.1202122197,-0.34978351,-0.2250769585,0.2266209126,0.1959666163,0.0931312665,0.1234253794,0.1805720925,0.2332053483,0.3488671184,-0.5268515348,-0.2841077745,-0.0600972325,0.4590164125,0.5724371672,-0.1469995081,0.5827165842,-0.0361847393,-0.3093143106,-0.2024856955,-0.1136060655,0.1153994948,-0.1437118649,0.146532461,-0.0635502711,0.2393677682,0.1057119817,0.0575048141,0.1236136854,0.0426457748,-0.1881812215,-0.2267202586,-0.2250871211,-0.130401805,-0.079543516,0.4237511158,-0.1307204068,-0.1249592081,-0.1547616273,0.0462993942,-0.2889908254,0.0783739164,-0.2232812494,0.1012440622,0.1913743466,-0.0936262757,0.1094516069,-0.3080514669,-0.1040450856,-0.0126841776,-0.3979837,-0.1693216115,0.0203968734,0.1602524221,-0.1382474303,-0.2311486602,0.0558580384,-0.3819752336,0.0380377956,-0.0533149838,0.1250467598,0.4172584713,0.0579878725,-0.0271963347,-0.0339336842,-0.0306240786,-0.2523109913,-0.1675626338,-0.1354081035,0.0299070161,0.2597557008,0.1947025061,0.0351850279,0.1711546779,-0.3019530475,0.2385359704,0.259829551,0.1949338019,0.1614032239,-0.1459692568,0.0144771663,0.3086936772,0.4575980902,0.1588773727,-0.1488347948,-0.3826071322,0.0354971625,0.1678587943,-0.1444538534,-0.2038942128,0.1804102808,0.1259716302,-0.1261832118,0.1846870482,0.4293412268,0.0687757581,0.3359305561,0.1370522827,0.6464172006,-0.4157039523,0.4020079374,-0.0214492809,0.1241483837,-0.1396703273,0.4943686724,0.1702246368,0.1256908774,0.4902687967,0.2014727145,0.1294519901,0.5204962492,0.2998254299,-0.5431535244,-0.6889321208,0.2738269866,0.3488745391,-0.3763978481,0.2423752695,0.0201690961,0.0024440051,-0.2609206438,0.1678765416,-0.1381923854,0.1791601777,-0.1543649733,-0.1634513736,0.5487616658,-0.2669768631,0.1158447936,0.0672507212,-0.0067788926,0.1863047332,0.1822021753,-0.1288622767,-0.0684223026,-0.4192673564,0.2929189801,0.2478071153,0.359895438,-0.2545068562,0.1231478974,0.0996151045,-0.0398908332,0.1149677336,0.3805103302,0.5170983672,0.093573302,0.1812502444,0.0517078564,0.0577839278,0.0136836823,-0.1843722314,0.3988665044,-0.1164815873,-0.0592681915,0.010174267,0.1127576903,-0.0093597937,0.0828230828,0.1456762105,0.0465975888,-0.4272485375,0.01604582,0.1558159441,-0.0448789522,0.0233874694,0.2268603295,-0.2670728564,-0.137284711,0.619040072,0.0123454137,0.2194406986,-0.161592856,0.0717888921,-0.0920788273,0.3734674156,0.1822925657,0.1486894637,-0.3880926073,-0.0459001847,-0.7635447383,-0.0579608157,-0.433213383,0.0615117364,0.0375061482,0.1487486362,-0.2146013379,-0.0109138135,0.1640014648,-0.0992856696,-0.2824617922,0.0789492503,-0.1491823941,-0.0327432975,-0.0718953386,-0.2404256761,-0.1748695821,-0.3282741308,0.3216022551,-0.0223113615,0.0536647849,-0.0545273274,-0.1020410731,0.1952946782,-0.0020405883,0.3635303378,0.1444205642,0.4088543952,0.2295743227,0.0669232011,-0.3549906909,-0.2211503386,-0.2582732737,0.2048445046,0.0921686962,0.1315241605,-0.621353507,-0.1107707918,-0.164438948,0.3291623294,0.0122512793,-0.2006853223,0.0399549864,-0.1441812962,-0.1794246137,0.4260330796,-0.0349136926,0.3897731006,0.1582591087,0.437587887,-0.3942762911,-0.0755254701,0.2275890112,-0.3375058174,-0.3425861895,-0.0241612867,0.2251670659,0.1019811705,-0.1575053632,-0.4065971375,-0.0406342484,0.4557797313,-0.256590724,-0.0541679747,0.2129300833,-0.052550327,0.0668796897,0.0248620864,0.1761422008,0.0781632289,-0.1209693477,0.1881721914,-0.2954846025]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2187","title":"Question (potential issue?) related to datasets caching","comments":"I can provide details about the code I'm running (it's really-really close to some official samples from the huggingface transformers examples, I can point to the exact sample file, I kept a record of that). I can also describe in which conditions this race occurs (I'm convinced it has to do with forcing the redownloading of the dataset, I've been running hundreds of experiments before and didn't have a problem before I forced the redownload). I also can provide samples of the different stack errors I get and some details about the level of concurrency of jobs I was running. I can also try to imagine how the race manifests (I'm fairly sure that it's a combo of one job cleaning up and another job being in the middle of the run).\r\n\r\nHowever, I have to cleanup all this to make sure I'm no spilling any info I shouldn't be spilling. I'll try to do it by the end of the week, if you think all this is helpful. \r\n\r\nFor now, I have a workaround. Don't use forcing redownloading. And to be ultra careful (although I don't think this is a problem), I run a series of jobs that will prepare the datasets and I know there is no concurrency wrt the dataset. Once that's done (and I believe even having multiple jobs loading the datasets at the same time doesn't create problems, as long as REUSE_DATASET_IF_EXISTS is the policy for loading the dataset, so the filelock mechanism you're using is working in that scenario), the prepared datasets will be reused, no race possible in any way. \r\n\r\nThanks for all the details you provided, it helped me understand the underlying implementation and coming up with workarounds when I ran into issues. ","body":"I thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!","comment_length":291,"text":"Question (potential issue?) related to datasets caching\nI thought I had disabled datasets caching in my code, as follows:\r\n```\r\nfrom datasets import set_caching_enabled\r\n...\r\ndef main():\r\n\r\n # disable caching in datasets\r\n set_caching_enabled(False)\r\n```\r\nHowever, in my log files I see messages like the following:\r\n\r\n```\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Using custom data configuration default-888a87931cbc5877\r\n04\/07\/2021 18:34:42 - WARNING - datasets.builder - Reusing dataset csv (xxxx\/cache-transformers\/datasets\/csv\/default-888a87931cbc5877\/0.0.0\/965b6429be0fc05f975b608ce64e1fa941cc8fb4f30629b523d2390f3c0e1a93\r\n```\r\nCan you please let me know what this reusing dataset csv means? I wouldn't expect any reusing with the datasets caching disabled. Thank you!\nI can provide details about the code I'm running (it's really-really close to some official samples from the huggingface transformers examples, I can point to the exact sample file, I kept a record of that). I can also describe in which conditions this race occurs (I'm convinced it has to do with forcing the redownloading of the dataset, I've been running hundreds of experiments before and didn't have a problem before I forced the redownload). I also can provide samples of the different stack errors I get and some details about the level of concurrency of jobs I was running. I can also try to imagine how the race manifests (I'm fairly sure that it's a combo of one job cleaning up and another job being in the middle of the run).\r\n\r\nHowever, I have to cleanup all this to make sure I'm no spilling any info I shouldn't be spilling. I'll try to do it by the end of the week, if you think all this is helpful. \r\n\r\nFor now, I have a workaround. Don't use forcing redownloading. And to be ultra careful (although I don't think this is a problem), I run a series of jobs that will prepare the datasets and I know there is no concurrency wrt the dataset. Once that's done (and I believe even having multiple jobs loading the datasets at the same time doesn't create problems, as long as REUSE_DATASET_IF_EXISTS is the policy for loading the dataset, so the filelock mechanism you're using is working in that scenario), the prepared datasets will be reused, no race possible in any way. \r\n\r\nThanks for all the details you provided, it helped me understand the underlying implementation and coming up with workarounds when I ran into issues. ","embeddings":[-0.2184490561,-0.0956564993,-0.0137409028,0.3255733848,0.1443524659,0.0401230156,0.2102061063,-0.000614408,0.2891818583,-0.2389941514,-0.0172556043,-0.1222877279,-0.2192769796,0.066736266,-0.0892674029,0.3100433648,0.1153514162,-0.0467662029,-0.2045849413,0.0261501055,-0.0979608744,0.0027088388,-0.0717039779,0.073559247,-0.5078132153,-0.1520064026,0.1791333556,0.1339304447,0.1075709537,-0.5074865818,0.4659926295,0.3332926929,0.1759842485,0.5821760297,-0.0001166996,0.0438799523,0.2641133964,-0.0501921922,-0.265378803,0.0882521272,-0.2523038387,-0.2569718063,0.2132923901,-0.0493345,-0.1088622063,0.0157411043,0.1350114197,-0.6665105224,0.7769575715,0.193910718,0.1887287349,0.2240133584,-0.3726796508,0.0592404231,0.0492232554,0.202809155,-0.0299787186,-0.0593019426,0.3182838261,-0.1498446018,-0.2916332185,0.3643889427,-0.1399885416,0.2117864341,0.5098317862,-0.0168476291,-0.2728926241,-0.2045925707,0.1750508696,0.135580644,0.6188479662,-0.220178768,-0.2451947331,-0.4416951537,-0.2497267574,-0.29144454,0.2839868665,-0.0784026533,-0.0207538381,0.2485278845,-0.5805151463,-0.0892047659,-0.0174449738,-0.2691357136,-0.0459529795,-0.009221822,-0.2377458811,0.1776919663,0.0724423304,0.1365374327,0.4682551622,-0.4004385471,0.0528008752,0.1464362293,-0.5663049817,-0.0044937041,0.18407996,0.4157847762,-0.0239245631,0.0996473432,-0.1134606078,0.0065553929,-0.0779411793,-0.0695116743,0.4441516101,0.3384639025,-0.1350044608,0.1796701252,0.2509441972,0.0115088243,-0.0748262256,-0.0447793789,0.1517451555,-0.3178058267,0.4623070359,0.0269136187,0.2113283426,-0.3225102723,-0.3603189588,0.0980270281,0.1430568844,-0.1305947453,0.1650646329,0.2228051126,-0.0674749017,-0.1174937934,0.0155920358,-0.0915749073,-0.3483349383,-0.3060492873,-0.2344367355,-0.198652342,-0.3293975294,0.0883615986,0.1922550052,-0.2432707846,-0.0837706253,0.288639605,0.2113749236,-0.1393365413,0.0559378006,-0.0209430251,0.1351681799,0.3564774096,-0.3113028109,0.3331823349,0.384408474,0.0039459649,-0.1590268016,0.1457968503,-0.2420632094,-0.4122130573,0.0911552534,0.0609010383,-0.5861328244,0.0046572331,-0.2381560057,0.1079232246,0.4921090305,0.1390797049,0.1639644504,-0.1510176957,-0.1433896124,-0.1491028219,-0.0015660542,0.7684624195,-0.408189714,-0.0344140343,-0.1229348034,0.0282869227,0.1170176044,0.3622053564,-0.5105870962,0.1386979371,-0.1987221986,-0.2503762543,-0.160459131,-0.2873807251,-0.4558068812,0.1284737587,-0.071611695,0.4496453702,0.1800094098,0.0716421455,0.0661644936,-0.1947529018,-0.0151455263,0.0183860734,0.0790265277,-0.0493962541,-0.1938403249,-0.5073305964,0.1770637929,0.0141504128,0.1087328941,0.0926752761,-0.1015163511,-0.0482048206,0.0621732622,-0.1136097237,0.1263996214,0.2767601907,-0.0203790423,0.0538357496,0.0563505962,0.0637686253,-0.7198367715,0.3640692532,0.0621958598,-0.1775560081,-0.0647809282,-0.1762681305,-0.2811445594,-0.0513734072,-0.4152140915,-0.2084294558,-0.0077643017,-0.0453630984,0.0941106528,-0.1766506881,-0.1650436074,0.6631064415,-0.2613233626,-0.0377372317,-0.4285814166,0.2097108066,-0.0056174709,0.1183557138,-0.1302143037,0.0049774274,0.1483415067,-0.0457024388,0.0555208698,0.2990195751,-0.0373432413,0.4231193066,0.056237407,0.3884788454,0.0602316298,-0.0962392837,0.2031462342,0.0306118168,0.0598438643,0.0012683826,-0.2350691259,0.3701173961,-0.3914734125,0.2245015502,-0.1687200218,-0.1664040685,0.1337195635,-0.1492458135,-0.088498801,0.0524171777,0.1954354197,0.1490417421,0.0420340262,0.2648719251,-0.1553646773,0.0830846056,0.3847407103,-0.0000495887,-0.0716430098,0.0249559227,-0.1766460538,-0.2358375788,0.1340906322,0.4145579636,0.477411747,0.0927886143,0.0733265579,0.0247047525,-0.1316232234,-0.2188059688,0.2385985702,-0.0333118029,-0.0063819648,0.1322800815,0.1501534879,-0.1033414379,-0.507316947,-0.0526213869,0.1299577653,0.2476085126,-0.4868394434,0.3516368568,-0.242655322,0.0176855884,-0.2625905573,0.0853152871,-0.3560926616,-0.2444316596,0.0534923337,0.2039736658,0.0836242139,0.0649746358,-0.1669404656,0.4773253798,-0.1676080972,0.0121735958,-0.2627763748,-0.2786838114,-0.2398258746,-0.0141076585,0.2255124003,-0.1726500839,0.2520121634,0.0302066114,0.0040558991,-0.1593365073,-0.0121277738,-0.1062167436,-0.0289151799,0.3015747666,-0.1936888695,0.1197013855,-0.0767588615,-0.1719681025,-0.0160153694,-0.1238480955,-0.0658681691,-0.1261281222,-0.0156724937,0.1032822952,-0.0041502584,-0.1276306659,-0.197197184,-0.233464554,0.0609404296,-0.0719271898,-0.1150187701,0.6053476334,-0.0297701266,-0.0719685704,0.0590292402,-0.1717266142,-0.644600153,-0.344782263,0.1292091161,0.0916936547,-0.1958631724,0.0055275764,0.0989459008,0.2120895833,0.1458248645,-0.6791308522,0.0002875104,-0.2181429416,-0.0537371598,0.1720138341,0.0500238538,0.3800395429,0.1637678444,-0.0125736455,-0.0002627061,-0.0968451053,-0.0199534707,-0.0650363788,0.3751189411,-0.0797761902,0.0927563012,0.0115720211,0.9484989047,0.3921333849,-0.1921592355,0.0230165888,0.1780514121,0.5541974902,-0.0524036326,-0.3691071868,-0.0638644248,-0.3495127261,-0.403097868,0.0950814188,-0.072701171,0.1678203195,0.1396745145,0.3958053291,-0.0177404601,-0.3547945917,0.1807866096,-0.1882044077,0.2042851597,0.174690634,-0.1109636277,-0.3441926539,-0.2060511261,-0.1491052508,0.1773611158,0.2643487453,-0.055145625,-0.3965284526,0.0762187168,-0.4206959903,0.3041411638,-0.1695515364,0.1284060031,-0.2336470485,0.1175068095,0.3223249316,0.155005604,0.6023414731,-0.3451421261,0.0229813065,0.0821799934,-0.0549186431,-0.2506531179,-0.2354979813,-0.063493982,0.1432371736,0.0611968599,0.2134844065,-0.0873489529,-0.1546728313,-0.0518676788,0.0747046247,-0.2156245708,-0.174305439,-0.365637064,-0.2099665254,-0.2125594318,0.2507587671,0.1931668222,-0.1734550446,-0.3332675993,0.0755575672,-0.1186509803,0.1591359526,0.2175832987,0.0495363399,0.2550600171,0.2173022777,0.0375316776,0.5109981298,-0.2194156349,0.163136363,0.4903517962,0.0542380661,-0.2091220319,0.081709601,0.1843613535,0.0765582174,0.5475765467,-0.2433158755,0.2171182781,0.3171908855,-0.0675292686,-0.5061047077,-0.0654043555,0.2310700864,0.3086543381,-0.2962409556,-0.3967838585,0.4350885153,0.1611391008,-0.2826729119,0.2136973143,-0.0432992429,-0.2798189819,0.1505182534,0.0942792892,0.9982904196,-0.1949733347,0.2090241313,-0.0191534907,-0.1586573124,0.6049423814,-0.2817754447,0.2646333575,-0.1085772216,-0.2103364468,-0.1136727184,-0.1672250479,0.2565815747,0.2159817517,-0.0513340533,0.3694311082,-0.238775894,0.4395717382,-0.1976953149,0.0810442343,0.056327153,-0.1476301998,-0.1512925774,0.1492900997,-0.1385562867,0.4672683179,-0.1525168568,0.1019643992,0.1897746623,0.1193229258,-0.4207965732,-0.0937194526,-0.0053804652,0.2596659958,-0.0195944309,-0.2217673957,0.3719568253,0.2365526259,0.4461587369,-0.0853945389,-0.1838989407,0.3200930357,-0.0009723672,0.1532409787,-0.1228828579,-0.0911116675,0.2610341609,0.177786991,-0.3299684525,0.1710199267,0.0224033389,-0.3243248463,0.0540726148,0.2893796563,0.0432862528,-0.5138397813,0.0632029921,0.321831733,0.1608247012,0.0619603172,0.0875606462,0.1086739302,0.1139117628,0.3072549403,-0.2990091145,-0.3236632049,-0.1141427681,0.4120743275,0.49700737,-0.2377083302,0.5076001287,-0.0371108502,-0.2190788388,-0.1393059939,-0.1854185015,0.2526694834,-0.5142325759,0.1806246787,-0.0955933556,0.1075655222,0.1148710921,0.1174005121,0.2029696554,-0.0126268705,-0.2576303184,-0.0290427208,-0.3065875471,-0.0058658654,-0.0665067732,0.3970097005,0.0137835778,-0.067938678,-0.1032980084,0.1723506302,-0.2670114338,-0.002637086,-0.2696271539,0.1142376289,0.1528417021,-0.3196125925,0.2932244539,-0.2459312975,-0.1411475539,-0.0386363976,-0.4665690064,-0.1316976696,0.0504765958,0.1920237243,-0.0836309418,-0.2960078716,-0.1130534187,-0.1597792953,0.1190349311,-0.1632992029,0.2240612358,0.3359929025,0.0280140098,-0.0113086468,0.0494611785,-0.0827817544,-0.0830925629,-0.0817228258,0.055062037,0.2035272121,0.3943223059,0.2428922653,-0.0679363832,0.0793475956,-0.3300131261,0.2858979404,0.3381083012,0.2159092426,0.1693324447,-0.2660115659,-0.0638589934,0.1596682966,0.3380587101,0.2196959406,-0.0931827053,-0.3879364431,-0.0108066434,0.1076499,-0.1728515625,-0.1275527924,0.3048496544,0.2739822268,-0.1411838233,0.3339981735,0.4254006445,0.0563323013,0.5450056791,0.1230318621,0.8266119361,-0.4311626852,0.4370393455,-0.0314341187,0.0473012924,-0.0394713171,0.6299881339,0.2780952156,0.2424534112,0.3363348246,0.1222376525,0.003255601,0.056930799,0.3076259196,-0.3858354092,-0.7054358721,0.1817716956,0.4742477536,-0.3341670334,0.1956605613,0.0491503142,0.1231237203,-0.2674402595,0.0476267338,-0.3517276645,0.2956910729,-0.1455845535,-0.2735616267,0.5554332137,-0.1987349838,0.1009414792,0.1675454974,-0.0342251845,-0.0270057525,0.2948549688,-0.2003130913,-0.1536576003,-0.282043606,0.1737097949,0.1838050485,0.3505780101,-0.1240313128,0.2311261147,0.2219638079,0.0254687052,0.1306252778,0.4619572461,0.5259422064,-0.0516012534,0.0839899778,0.2164376676,0.0298777651,0.1085668653,-0.1861699075,0.3402243853,-0.1844017357,-0.0597759411,0.0274575613,0.1202937514,0.0130256098,-0.0148177696,-0.1308914274,0.1223370135,-0.3265521824,-0.0460151844,0.0294297915,-0.0492129773,-0.070707269,0.1818908304,-0.2867631018,-0.0290515665,0.5673057437,-0.1833664924,0.2569792271,-0.1006048843,0.0487734377,-0.1099192128,0.5449647903,0.3410850167,0.1019505933,-0.2991896272,-0.0146673042,-0.6810794473,-0.0785881132,-0.2810330093,0.2189967632,0.1051527336,0.2027107626,-0.1195542365,0.0607268699,0.0004609286,-0.1750670075,-0.319589287,-0.0104579404,-0.0973802879,0.0199282933,0.004805584,-0.0699740425,-0.1038163006,-0.2499149591,0.2984412014,0.2389940768,-0.0354395844,-0.0437826477,0.0643723458,0.1109832153,0.0058752508,0.2429562509,0.2596434057,0.2820842564,0.3239372373,0.00575657,-0.4033367336,-0.2470770478,-0.1818036735,0.164651677,0.0001342228,0.1116916239,-0.4221593738,-0.0303590689,-0.2969040573,0.1451149285,0.1499473751,-0.2467023581,-0.0663996786,-0.1645839959,-0.214569509,0.2962663174,0.1407629699,0.4002628922,0.099928081,0.3906538188,-0.4157352448,-0.1032342389,0.4054407775,-0.337623626,-0.3259881139,0.0948297381,0.1858668625,0.1200378239,-0.2982679904,-0.5435605049,0.0448629409,0.2724534869,-0.203610003,-0.1230369285,0.1326122284,-0.0984333307,0.0288571119,-0.0705798343,0.2866345048,0.0962144807,-0.1128979996,0.2418969274,-0.2190691382]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2185","title":".map() and distributed training","comments":"Hi, one workaround would be to save the mapped(tokenized in your case) file using `save_to_disk`, and having each process load this file using `load_from_disk`. This is what I am doing, and in this case, I turn off the ability to automatically load from the cache.\r\n\r\nAlso, multiprocessing the map function seems to be slower at the moment (#1992), hope this helps you.","body":"Hi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.","comment_length":62,"text":".map() and distributed training\nHi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.\nHi, one workaround would be to save the mapped(tokenized in your case) file using `save_to_disk`, and having each process load this file using `load_from_disk`. This is what I am doing, and in this case, I turn off the ability to automatically load from the cache.\r\n\r\nAlso, multiprocessing the map function seems to be slower at the moment (#1992), hope this helps you.","embeddings":[-0.2537618279,-0.2207167447,0.1177767888,0.0745745301,0.2116166502,-0.1097597107,0.5170732141,0.0102053266,0.2801567912,0.0237633772,0.1822478175,0.3434494138,-0.2290642709,-0.5855442286,0.0941431373,0.0615340844,0.1861155033,0.0366594605,-0.0377356932,-0.2008032501,-0.1529330909,0.0759166479,-0.1542706788,0.1170513257,-0.7540709376,-0.1523153037,-0.1834535897,0.1027398556,0.2253180742,-0.3499087691,0.2342278361,0.2213348448,0.4013623595,0.5336914062,-0.0001272091,0.0312626697,0.1878611445,-0.1983309686,0.0397178754,-0.0476766787,0.1641900688,0.0050253472,0.1936113387,-0.2953366041,0.1310725659,-0.1476733088,0.3262221217,-0.4418162107,0.4140013158,-0.1077297255,-0.0018774115,0.0943896994,-0.5714957118,0.1398764253,-0.2214399129,0.1234803572,0.1711140275,0.2180822492,0.3029266596,-0.3416551054,-0.3789148629,0.1582272202,-0.1466236264,0.1340594739,0.4135627449,0.0366018042,-0.0789881125,-0.3487006128,-0.075750038,-0.1218584478,0.1076876521,-0.0743067637,-0.1881792545,-0.4665382504,-0.2733976543,-0.0016580709,-0.0024628954,0.116243735,-0.0233727768,-0.1197138727,-0.5124091506,-0.0625246316,0.0552944317,0.1326544136,-0.2732613087,0.4564465284,-0.0254738089,0.3163393438,0.1419928521,0.0543378443,0.0837460086,-0.1719923168,0.1296369135,0.3715832233,0.0275528189,0.0141743347,0.0489273742,-0.3223410249,-0.0232889,0.0356097072,0.2321069092,0.2598097324,-0.0479859076,0.265766263,0.4007704556,0.0927509293,-0.1391413957,0.6415418983,0.0301112514,-0.2004719824,-0.6418218613,-0.1724125892,-0.3519452512,-0.0627355203,0.3335886896,0.0427098982,-0.2098451257,0.0722271129,0.4604047239,-0.2394168377,-0.0065294988,-0.1325436831,0.0212120153,0.2508155704,0.1287937164,0.437037617,-0.0555269644,0.1014266536,-0.190160498,0.1589486897,-0.0596041419,-0.3029897511,-0.3049105704,0.316068083,0.1618765295,0.1877589226,0.4239696264,-0.1001108065,0.289442122,-0.1613698006,0.3981889188,-0.1120462343,0.6585771441,0.1044589132,-0.2513783276,0.2223655581,0.1862223744,0.3770996928,-0.1535631716,0.335416913,-0.5022829771,-0.4383485317,0.4798748195,-0.0732970536,-0.0095938472,0.0692022368,-0.1767223477,0.0427591875,0.5391663909,-0.2560784817,0.134424001,-0.3575458229,-0.2090658844,-0.1776281595,0.2568801343,0.4551616609,0.3137147129,-0.341816932,0.1241682246,0.3162250221,0.2081830502,0.2290944606,-0.6239327788,0.3898105621,-0.2199796587,0.2747520506,0.3872870505,-0.3571222723,-0.1435228735,0.2945982218,-0.0989739373,0.0233595725,-0.1570941806,-0.0249821246,0.4641161859,-0.0252317898,0.1258820295,0.2964549661,-0.216262728,0.2001035213,-0.261736244,0.0169189703,0.0804381073,-0.0070854248,0.1947211921,0.0033709346,-0.0128012626,0.1196192652,0.2070769817,0.0102299741,0.2283787429,0.0252541639,-0.3850671053,0.0924091339,0.165482685,0.0665195286,-0.1504980773,0.1814241707,-0.0332348272,-0.3959296942,-0.1602503955,-0.0522115901,0.1628748178,-0.2076937854,-0.3108922839,-0.1402078569,-0.1141550615,0.1380293667,-0.0407926142,-0.2170617431,-0.3000540435,-0.0936286971,0.1410432905,-0.0588696413,0.2177682668,0.1037803516,0.1914524436,-0.1203961521,-0.4317591786,-0.1487240344,0.1081416756,-0.2924800515,-0.1076147482,0.3776341677,0.0455733202,0.2080194801,-0.1105848923,0.3216450512,0.2376917005,0.066899173,0.1610087454,0.1279921681,-0.1245717034,-0.1512021422,-0.142078042,0.3465422094,0.128989622,0.1471359581,-0.0155485664,0.0163485743,0.0418233871,0.0829828158,-0.2396878004,-0.0824884102,-0.3200368583,-0.1921786666,0.2612736225,0.1546629518,0.0379151292,0.0008637675,0.3312492073,0.0261272565,0.0083661247,-0.1484815776,-0.1806361079,0.0400650427,0.1552491784,0.1572336107,0.3422781229,-0.0142571228,0.3932816386,-0.1144815236,0.1034137383,0.06225916,-0.0817419663,-0.0038878322,0.2324126065,0.15785712,-0.0431561358,0.183517918,-0.2471131682,-0.104632467,0.3485348523,-0.2459665388,0.0869296491,0.2329536527,-0.0123324441,0.1808802485,-0.5816081762,0.0918158442,0.0024792906,0.0669043064,-0.1619982123,0.2604064047,-0.0950392038,0.4181236029,-0.0660849512,0.2435841858,-0.1192760319,-0.2782994807,0.1805499345,-0.15312244,-0.012159599,-0.1923269033,0.1284361035,-0.2399803996,-0.0226501599,0.0775606632,-0.0577873811,-0.1429297924,0.1827362925,0.1133465022,0.2074275613,-0.5552704334,-0.2969094515,-0.0320735425,-0.378795594,-0.1486826241,0.1686326563,-0.4979081154,0.1192707866,-0.055137407,0.245264411,-0.1717564315,-0.1245706826,-0.3170382977,-0.2305708379,-0.0193637144,0.133881554,-0.0549433865,0.1418401748,0.0033795778,-0.1053651124,-0.2393789291,-0.1764801741,-0.2918813229,0.0651777461,-0.5251649618,0.3886308968,-0.1867223978,-0.0471369028,-0.109347865,-0.0461086854,-0.0654087663,0.6399704814,-0.334898591,0.0144050214,0.3095326722,-0.010113867,-0.0965969041,0.1116476655,0.5424685478,0.1066478863,0.1266251504,-0.0549990423,0.2443153262,0.2560414374,0.1710574329,0.0513743982,0.1595277488,-0.1193900257,0.1637128294,1.2782388926,0.1522191167,-0.0518915094,0.0262432639,0.0314820632,-0.0140471822,0.0686671063,-0.1058131158,0.0613840669,-0.3951035738,-0.1308125407,0.179306373,0.1871938407,-0.5139589906,0.0521272793,0.1936313659,-0.0122206649,-0.3019163013,0.3315780163,-0.7171853185,0.5827730298,0.0590841807,0.0178885721,-0.4254792333,0.045399826,0.0945204347,-0.1513158679,0.7113601565,0.0534209572,-0.4360212088,0.0862669274,-0.1918416172,0.0394076109,0.2148355395,0.2028769851,0.0952703208,-0.1703162938,0.147795856,0.1729288846,0.4053848684,-0.0639535859,0.0224064104,0.1169934794,-0.2179405391,-0.1327845156,-0.1501162648,0.064364247,0.3700457215,0.0236489978,0.430801183,-0.0303367078,-0.1774694622,-0.4406905174,0.0884211138,-0.3515691161,-0.0433642752,-0.1870866269,0.0759658292,-0.2960950434,0.2095098794,0.3047899604,-0.1735835522,-0.0213712044,-0.262239337,-0.0159953311,0.3147200942,-0.1036774591,0.037082281,0.2442333847,0.1481226534,0.013170097,0.4719136059,0.2137568444,0.2138777673,0.3546659946,-0.3274190426,-0.1393887997,0.0368320271,0.0073549538,0.234070614,0.1898229718,0.0306465756,-0.2361022979,0.1773872077,-0.167742312,-0.2283785045,0.2950416803,0.1178728044,0.0684980303,-0.1757792085,-0.4709654748,0.0921665505,0.1839773506,-0.2220710367,0.4868579209,-0.8516537547,-0.261679858,0.4181653559,0.0608363412,0.6925716996,-0.2607029974,0.227264449,-0.250442028,0.4874329865,0.020454742,-0.6066097617,0.4200372398,-0.254858017,-0.2041648924,-0.0412223898,-0.1717183441,-0.1559633017,0.7070311904,0.0804174393,0.1725239158,0.1165403426,0.3262236714,-0.0640336424,0.0934551284,0.2966890037,-0.4377833605,0.229028821,-0.0624263473,0.1829893738,0.156969294,-0.0804343894,0.0565100312,0.124040477,-0.2584005296,0.2086893916,0.0717218742,-0.3600730598,0.3634980619,0.0398199335,0.0521443076,-0.243333742,0.6011244059,0.0109047834,-0.2319959551,0.09407603,0.0436988212,0.193117395,-0.0126967691,-0.262334615,-0.1697430909,0.1262998283,0.0967982411,-0.1459383518,-0.1205745712,-0.3000965714,-0.4416499734,-0.223753795,0.360080421,0.6041749716,-0.0505638681,-0.0588486269,0.2081897855,-0.0861713141,0.166884169,-0.0069433674,0.0640020221,-0.1381703764,0.3867056668,-0.0235097799,-0.1242907271,0.2014888227,0.3131961524,-0.1574299932,-0.1682617664,0.2568942606,-0.2654665112,-0.09693712,-0.0573922507,-0.0330521911,0.1389321983,-0.3625167906,0.0169916097,-0.1065589935,0.0166316852,0.0993066281,0.0837628245,0.1839834154,0.107527487,-0.323993057,-0.1815751493,-0.3717830181,-0.0278916862,-0.1365002245,0.0505666025,0.273403585,0.1305761337,0.1053174138,0.3764922321,-0.1317270249,0.1228380725,0.0817987695,0.3973923624,-0.05186712,0.0050210641,-0.1158289313,-0.0291919559,-0.1875379533,0.0908713639,-0.1421469152,0.004255462,0.0848780647,0.2250854522,0.1945449114,-0.0718456656,-0.0755309612,-0.2173605263,0.116221495,-0.4500408173,0.0487676337,0.0160334427,-0.1881897151,0.0467121117,0.5341085196,0.1623094976,-0.4041575789,0.1451513916,-0.0866433457,-0.0505961627,-0.0320249684,0.1041378677,0.1630800366,-0.0049900212,-0.0814534128,-0.0345540866,0.3659764528,0.1690802127,0.4754377306,-0.2272117883,0.0704821721,-0.0292813536,0.1734144688,0.2325785905,-0.3147429526,-0.0599202737,0.3415470421,0.0241225902,0.0282654651,-0.1990050822,0.3771487474,0.0867178366,-0.1169262379,0.269580394,0.2913129628,-0.3984209597,0.1071690992,0.13878043,0.5225743651,-0.3287658095,0.2685762048,0.026906088,0.085723117,0.1760590225,0.2698514163,0.310449481,0.1782119125,-0.0747818798,0.0293806326,0.2496602833,0.3985043764,-0.1035515368,-0.1874758154,-0.4894893765,-0.0736385658,-0.0044096862,-0.0369611047,0.1242682934,-0.2778055072,0.2728984356,-0.2847654819,-0.5857912898,0.1254214495,0.3274540007,-0.2013130635,-0.0886794105,0.0926787481,0.0113809267,-0.04623463,0.0941591561,0.0441685989,0.2097102404,0.5593304038,0.0858864933,0.0181098767,0.1541578323,-0.0717386082,0.0019633719,0.4593269229,-0.3261109591,0.1913291067,-0.1725729853,0.0807391182,0.0200430546,-0.0315704718,0.0498550087,0.5228732228,-0.1188720688,0.1938308328,0.0442509204,0.1886543483,-0.1131996587,0.1667529047,-0.1233745888,-0.3006847203,0.0307742395,-0.0814752579,-0.0131051978,0.028823575,-0.0665220618,0.0668584257,0.2088633776,0.2820904851,0.1485458016,-0.1038457751,0.1028903425,0.0694101751,-0.3599323332,0.1259170175,0.2366672009,-0.2897271514,0.1959314346,0.0505576171,-0.0399407148,0.0708528757,0.2542789876,0.1748011559,0.0303770751,-0.2747746408,-0.3522980511,-0.6728181839,0.302385211,0.0174807664,0.25087744,-0.1023810804,0.0117317839,0.0563799366,-0.0531251766,-0.2462201864,-0.2524305582,-0.038776882,0.3524783552,-0.0957507938,0.0024702032,-0.2875679433,0.399572432,0.1458623558,-0.1972410381,0.065390259,-0.0597794913,-0.0863967463,0.2657533586,-0.4236015975,-0.1760620475,0.0676908493,0.4964187145,0.0649721324,0.2099227756,0.0600785837,0.1801740825,-0.0981862769,0.0451122038,-0.0188293029,-0.0528834648,-0.0837273076,0.2601968348,-0.528018117,-0.0028348397,-0.4991467893,-0.4764001071,0.1246198639,-0.0155331492,0.0108440826,-0.2105778903,-0.2984160483,0.0919267014,0.2645703852,0.4401881397,-0.0507179983,-0.0598962829,-0.2187685072,0.0092310868,0.266627878,-0.6458785534,-0.28847,-0.2124379426,0.1345545799,-0.0403628573,-0.0596350729,-0.4026406705,-0.0723963454,0.2834527791,-0.0711669028,-0.4361755252,-0.1718805879,-0.0504207276,-0.1707316041,-0.2380543798,0.4838347733,-0.0408494323,-0.2001532614,-0.0964371935,-0.1717197299]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2185","title":".map() and distributed training","comments":"Thanks @hwijeen for the workaround, feels a bit prototypical but it works! (it seems files are written twice then though)\r\n\r\n(I haven't observed slowness using multiprocessed map function but I could be wrong)","body":"Hi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.","comment_length":33,"text":".map() and distributed training\nHi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.\nThanks @hwijeen for the workaround, feels a bit prototypical but it works! (it seems files are written twice then though)\r\n\r\n(I haven't observed slowness using multiprocessed map function but I could be wrong)","embeddings":[-0.2537618279,-0.2207167447,0.1177767888,0.0745745301,0.2116166502,-0.1097597107,0.5170732141,0.0102053266,0.2801567912,0.0237633772,0.1822478175,0.3434494138,-0.2290642709,-0.5855442286,0.0941431373,0.0615340844,0.1861155033,0.0366594605,-0.0377356932,-0.2008032501,-0.1529330909,0.0759166479,-0.1542706788,0.1170513257,-0.7540709376,-0.1523153037,-0.1834535897,0.1027398556,0.2253180742,-0.3499087691,0.2342278361,0.2213348448,0.4013623595,0.5336914062,-0.0001272091,0.0312626697,0.1878611445,-0.1983309686,0.0397178754,-0.0476766787,0.1641900688,0.0050253472,0.1936113387,-0.2953366041,0.1310725659,-0.1476733088,0.3262221217,-0.4418162107,0.4140013158,-0.1077297255,-0.0018774115,0.0943896994,-0.5714957118,0.1398764253,-0.2214399129,0.1234803572,0.1711140275,0.2180822492,0.3029266596,-0.3416551054,-0.3789148629,0.1582272202,-0.1466236264,0.1340594739,0.4135627449,0.0366018042,-0.0789881125,-0.3487006128,-0.075750038,-0.1218584478,0.1076876521,-0.0743067637,-0.1881792545,-0.4665382504,-0.2733976543,-0.0016580709,-0.0024628954,0.116243735,-0.0233727768,-0.1197138727,-0.5124091506,-0.0625246316,0.0552944317,0.1326544136,-0.2732613087,0.4564465284,-0.0254738089,0.3163393438,0.1419928521,0.0543378443,0.0837460086,-0.1719923168,0.1296369135,0.3715832233,0.0275528189,0.0141743347,0.0489273742,-0.3223410249,-0.0232889,0.0356097072,0.2321069092,0.2598097324,-0.0479859076,0.265766263,0.4007704556,0.0927509293,-0.1391413957,0.6415418983,0.0301112514,-0.2004719824,-0.6418218613,-0.1724125892,-0.3519452512,-0.0627355203,0.3335886896,0.0427098982,-0.2098451257,0.0722271129,0.4604047239,-0.2394168377,-0.0065294988,-0.1325436831,0.0212120153,0.2508155704,0.1287937164,0.437037617,-0.0555269644,0.1014266536,-0.190160498,0.1589486897,-0.0596041419,-0.3029897511,-0.3049105704,0.316068083,0.1618765295,0.1877589226,0.4239696264,-0.1001108065,0.289442122,-0.1613698006,0.3981889188,-0.1120462343,0.6585771441,0.1044589132,-0.2513783276,0.2223655581,0.1862223744,0.3770996928,-0.1535631716,0.335416913,-0.5022829771,-0.4383485317,0.4798748195,-0.0732970536,-0.0095938472,0.0692022368,-0.1767223477,0.0427591875,0.5391663909,-0.2560784817,0.134424001,-0.3575458229,-0.2090658844,-0.1776281595,0.2568801343,0.4551616609,0.3137147129,-0.341816932,0.1241682246,0.3162250221,0.2081830502,0.2290944606,-0.6239327788,0.3898105621,-0.2199796587,0.2747520506,0.3872870505,-0.3571222723,-0.1435228735,0.2945982218,-0.0989739373,0.0233595725,-0.1570941806,-0.0249821246,0.4641161859,-0.0252317898,0.1258820295,0.2964549661,-0.216262728,0.2001035213,-0.261736244,0.0169189703,0.0804381073,-0.0070854248,0.1947211921,0.0033709346,-0.0128012626,0.1196192652,0.2070769817,0.0102299741,0.2283787429,0.0252541639,-0.3850671053,0.0924091339,0.165482685,0.0665195286,-0.1504980773,0.1814241707,-0.0332348272,-0.3959296942,-0.1602503955,-0.0522115901,0.1628748178,-0.2076937854,-0.3108922839,-0.1402078569,-0.1141550615,0.1380293667,-0.0407926142,-0.2170617431,-0.3000540435,-0.0936286971,0.1410432905,-0.0588696413,0.2177682668,0.1037803516,0.1914524436,-0.1203961521,-0.4317591786,-0.1487240344,0.1081416756,-0.2924800515,-0.1076147482,0.3776341677,0.0455733202,0.2080194801,-0.1105848923,0.3216450512,0.2376917005,0.066899173,0.1610087454,0.1279921681,-0.1245717034,-0.1512021422,-0.142078042,0.3465422094,0.128989622,0.1471359581,-0.0155485664,0.0163485743,0.0418233871,0.0829828158,-0.2396878004,-0.0824884102,-0.3200368583,-0.1921786666,0.2612736225,0.1546629518,0.0379151292,0.0008637675,0.3312492073,0.0261272565,0.0083661247,-0.1484815776,-0.1806361079,0.0400650427,0.1552491784,0.1572336107,0.3422781229,-0.0142571228,0.3932816386,-0.1144815236,0.1034137383,0.06225916,-0.0817419663,-0.0038878322,0.2324126065,0.15785712,-0.0431561358,0.183517918,-0.2471131682,-0.104632467,0.3485348523,-0.2459665388,0.0869296491,0.2329536527,-0.0123324441,0.1808802485,-0.5816081762,0.0918158442,0.0024792906,0.0669043064,-0.1619982123,0.2604064047,-0.0950392038,0.4181236029,-0.0660849512,0.2435841858,-0.1192760319,-0.2782994807,0.1805499345,-0.15312244,-0.012159599,-0.1923269033,0.1284361035,-0.2399803996,-0.0226501599,0.0775606632,-0.0577873811,-0.1429297924,0.1827362925,0.1133465022,0.2074275613,-0.5552704334,-0.2969094515,-0.0320735425,-0.378795594,-0.1486826241,0.1686326563,-0.4979081154,0.1192707866,-0.055137407,0.245264411,-0.1717564315,-0.1245706826,-0.3170382977,-0.2305708379,-0.0193637144,0.133881554,-0.0549433865,0.1418401748,0.0033795778,-0.1053651124,-0.2393789291,-0.1764801741,-0.2918813229,0.0651777461,-0.5251649618,0.3886308968,-0.1867223978,-0.0471369028,-0.109347865,-0.0461086854,-0.0654087663,0.6399704814,-0.334898591,0.0144050214,0.3095326722,-0.010113867,-0.0965969041,0.1116476655,0.5424685478,0.1066478863,0.1266251504,-0.0549990423,0.2443153262,0.2560414374,0.1710574329,0.0513743982,0.1595277488,-0.1193900257,0.1637128294,1.2782388926,0.1522191167,-0.0518915094,0.0262432639,0.0314820632,-0.0140471822,0.0686671063,-0.1058131158,0.0613840669,-0.3951035738,-0.1308125407,0.179306373,0.1871938407,-0.5139589906,0.0521272793,0.1936313659,-0.0122206649,-0.3019163013,0.3315780163,-0.7171853185,0.5827730298,0.0590841807,0.0178885721,-0.4254792333,0.045399826,0.0945204347,-0.1513158679,0.7113601565,0.0534209572,-0.4360212088,0.0862669274,-0.1918416172,0.0394076109,0.2148355395,0.2028769851,0.0952703208,-0.1703162938,0.147795856,0.1729288846,0.4053848684,-0.0639535859,0.0224064104,0.1169934794,-0.2179405391,-0.1327845156,-0.1501162648,0.064364247,0.3700457215,0.0236489978,0.430801183,-0.0303367078,-0.1774694622,-0.4406905174,0.0884211138,-0.3515691161,-0.0433642752,-0.1870866269,0.0759658292,-0.2960950434,0.2095098794,0.3047899604,-0.1735835522,-0.0213712044,-0.262239337,-0.0159953311,0.3147200942,-0.1036774591,0.037082281,0.2442333847,0.1481226534,0.013170097,0.4719136059,0.2137568444,0.2138777673,0.3546659946,-0.3274190426,-0.1393887997,0.0368320271,0.0073549538,0.234070614,0.1898229718,0.0306465756,-0.2361022979,0.1773872077,-0.167742312,-0.2283785045,0.2950416803,0.1178728044,0.0684980303,-0.1757792085,-0.4709654748,0.0921665505,0.1839773506,-0.2220710367,0.4868579209,-0.8516537547,-0.261679858,0.4181653559,0.0608363412,0.6925716996,-0.2607029974,0.227264449,-0.250442028,0.4874329865,0.020454742,-0.6066097617,0.4200372398,-0.254858017,-0.2041648924,-0.0412223898,-0.1717183441,-0.1559633017,0.7070311904,0.0804174393,0.1725239158,0.1165403426,0.3262236714,-0.0640336424,0.0934551284,0.2966890037,-0.4377833605,0.229028821,-0.0624263473,0.1829893738,0.156969294,-0.0804343894,0.0565100312,0.124040477,-0.2584005296,0.2086893916,0.0717218742,-0.3600730598,0.3634980619,0.0398199335,0.0521443076,-0.243333742,0.6011244059,0.0109047834,-0.2319959551,0.09407603,0.0436988212,0.193117395,-0.0126967691,-0.262334615,-0.1697430909,0.1262998283,0.0967982411,-0.1459383518,-0.1205745712,-0.3000965714,-0.4416499734,-0.223753795,0.360080421,0.6041749716,-0.0505638681,-0.0588486269,0.2081897855,-0.0861713141,0.166884169,-0.0069433674,0.0640020221,-0.1381703764,0.3867056668,-0.0235097799,-0.1242907271,0.2014888227,0.3131961524,-0.1574299932,-0.1682617664,0.2568942606,-0.2654665112,-0.09693712,-0.0573922507,-0.0330521911,0.1389321983,-0.3625167906,0.0169916097,-0.1065589935,0.0166316852,0.0993066281,0.0837628245,0.1839834154,0.107527487,-0.323993057,-0.1815751493,-0.3717830181,-0.0278916862,-0.1365002245,0.0505666025,0.273403585,0.1305761337,0.1053174138,0.3764922321,-0.1317270249,0.1228380725,0.0817987695,0.3973923624,-0.05186712,0.0050210641,-0.1158289313,-0.0291919559,-0.1875379533,0.0908713639,-0.1421469152,0.004255462,0.0848780647,0.2250854522,0.1945449114,-0.0718456656,-0.0755309612,-0.2173605263,0.116221495,-0.4500408173,0.0487676337,0.0160334427,-0.1881897151,0.0467121117,0.5341085196,0.1623094976,-0.4041575789,0.1451513916,-0.0866433457,-0.0505961627,-0.0320249684,0.1041378677,0.1630800366,-0.0049900212,-0.0814534128,-0.0345540866,0.3659764528,0.1690802127,0.4754377306,-0.2272117883,0.0704821721,-0.0292813536,0.1734144688,0.2325785905,-0.3147429526,-0.0599202737,0.3415470421,0.0241225902,0.0282654651,-0.1990050822,0.3771487474,0.0867178366,-0.1169262379,0.269580394,0.2913129628,-0.3984209597,0.1071690992,0.13878043,0.5225743651,-0.3287658095,0.2685762048,0.026906088,0.085723117,0.1760590225,0.2698514163,0.310449481,0.1782119125,-0.0747818798,0.0293806326,0.2496602833,0.3985043764,-0.1035515368,-0.1874758154,-0.4894893765,-0.0736385658,-0.0044096862,-0.0369611047,0.1242682934,-0.2778055072,0.2728984356,-0.2847654819,-0.5857912898,0.1254214495,0.3274540007,-0.2013130635,-0.0886794105,0.0926787481,0.0113809267,-0.04623463,0.0941591561,0.0441685989,0.2097102404,0.5593304038,0.0858864933,0.0181098767,0.1541578323,-0.0717386082,0.0019633719,0.4593269229,-0.3261109591,0.1913291067,-0.1725729853,0.0807391182,0.0200430546,-0.0315704718,0.0498550087,0.5228732228,-0.1188720688,0.1938308328,0.0442509204,0.1886543483,-0.1131996587,0.1667529047,-0.1233745888,-0.3006847203,0.0307742395,-0.0814752579,-0.0131051978,0.028823575,-0.0665220618,0.0668584257,0.2088633776,0.2820904851,0.1485458016,-0.1038457751,0.1028903425,0.0694101751,-0.3599323332,0.1259170175,0.2366672009,-0.2897271514,0.1959314346,0.0505576171,-0.0399407148,0.0708528757,0.2542789876,0.1748011559,0.0303770751,-0.2747746408,-0.3522980511,-0.6728181839,0.302385211,0.0174807664,0.25087744,-0.1023810804,0.0117317839,0.0563799366,-0.0531251766,-0.2462201864,-0.2524305582,-0.038776882,0.3524783552,-0.0957507938,0.0024702032,-0.2875679433,0.399572432,0.1458623558,-0.1972410381,0.065390259,-0.0597794913,-0.0863967463,0.2657533586,-0.4236015975,-0.1760620475,0.0676908493,0.4964187145,0.0649721324,0.2099227756,0.0600785837,0.1801740825,-0.0981862769,0.0451122038,-0.0188293029,-0.0528834648,-0.0837273076,0.2601968348,-0.528018117,-0.0028348397,-0.4991467893,-0.4764001071,0.1246198639,-0.0155331492,0.0108440826,-0.2105778903,-0.2984160483,0.0919267014,0.2645703852,0.4401881397,-0.0507179983,-0.0598962829,-0.2187685072,0.0092310868,0.266627878,-0.6458785534,-0.28847,-0.2124379426,0.1345545799,-0.0403628573,-0.0596350729,-0.4026406705,-0.0723963454,0.2834527791,-0.0711669028,-0.4361755252,-0.1718805879,-0.0504207276,-0.1707316041,-0.2380543798,0.4838347733,-0.0408494323,-0.2001532614,-0.0964371935,-0.1717197299]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2185","title":".map() and distributed training","comments":"To my understanding, files are written twice anyhow(one after load_dataset, another aftet map). It's just that you now have it at the location where you can see, whereas it was secretlely saved at caching folder(.cache\/huggingface\/datasets by default)! Correct me if I'm wrong!","body":"Hi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.","comment_length":42,"text":".map() and distributed training\nHi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.\nTo my understanding, files are written twice anyhow(one after load_dataset, another aftet map). It's just that you now have it at the location where you can see, whereas it was secretlely saved at caching folder(.cache\/huggingface\/datasets by default)! Correct me if I'm wrong!","embeddings":[-0.2537618279,-0.2207167447,0.1177767888,0.0745745301,0.2116166502,-0.1097597107,0.5170732141,0.0102053266,0.2801567912,0.0237633772,0.1822478175,0.3434494138,-0.2290642709,-0.5855442286,0.0941431373,0.0615340844,0.1861155033,0.0366594605,-0.0377356932,-0.2008032501,-0.1529330909,0.0759166479,-0.1542706788,0.1170513257,-0.7540709376,-0.1523153037,-0.1834535897,0.1027398556,0.2253180742,-0.3499087691,0.2342278361,0.2213348448,0.4013623595,0.5336914062,-0.0001272091,0.0312626697,0.1878611445,-0.1983309686,0.0397178754,-0.0476766787,0.1641900688,0.0050253472,0.1936113387,-0.2953366041,0.1310725659,-0.1476733088,0.3262221217,-0.4418162107,0.4140013158,-0.1077297255,-0.0018774115,0.0943896994,-0.5714957118,0.1398764253,-0.2214399129,0.1234803572,0.1711140275,0.2180822492,0.3029266596,-0.3416551054,-0.3789148629,0.1582272202,-0.1466236264,0.1340594739,0.4135627449,0.0366018042,-0.0789881125,-0.3487006128,-0.075750038,-0.1218584478,0.1076876521,-0.0743067637,-0.1881792545,-0.4665382504,-0.2733976543,-0.0016580709,-0.0024628954,0.116243735,-0.0233727768,-0.1197138727,-0.5124091506,-0.0625246316,0.0552944317,0.1326544136,-0.2732613087,0.4564465284,-0.0254738089,0.3163393438,0.1419928521,0.0543378443,0.0837460086,-0.1719923168,0.1296369135,0.3715832233,0.0275528189,0.0141743347,0.0489273742,-0.3223410249,-0.0232889,0.0356097072,0.2321069092,0.2598097324,-0.0479859076,0.265766263,0.4007704556,0.0927509293,-0.1391413957,0.6415418983,0.0301112514,-0.2004719824,-0.6418218613,-0.1724125892,-0.3519452512,-0.0627355203,0.3335886896,0.0427098982,-0.2098451257,0.0722271129,0.4604047239,-0.2394168377,-0.0065294988,-0.1325436831,0.0212120153,0.2508155704,0.1287937164,0.437037617,-0.0555269644,0.1014266536,-0.190160498,0.1589486897,-0.0596041419,-0.3029897511,-0.3049105704,0.316068083,0.1618765295,0.1877589226,0.4239696264,-0.1001108065,0.289442122,-0.1613698006,0.3981889188,-0.1120462343,0.6585771441,0.1044589132,-0.2513783276,0.2223655581,0.1862223744,0.3770996928,-0.1535631716,0.335416913,-0.5022829771,-0.4383485317,0.4798748195,-0.0732970536,-0.0095938472,0.0692022368,-0.1767223477,0.0427591875,0.5391663909,-0.2560784817,0.134424001,-0.3575458229,-0.2090658844,-0.1776281595,0.2568801343,0.4551616609,0.3137147129,-0.341816932,0.1241682246,0.3162250221,0.2081830502,0.2290944606,-0.6239327788,0.3898105621,-0.2199796587,0.2747520506,0.3872870505,-0.3571222723,-0.1435228735,0.2945982218,-0.0989739373,0.0233595725,-0.1570941806,-0.0249821246,0.4641161859,-0.0252317898,0.1258820295,0.2964549661,-0.216262728,0.2001035213,-0.261736244,0.0169189703,0.0804381073,-0.0070854248,0.1947211921,0.0033709346,-0.0128012626,0.1196192652,0.2070769817,0.0102299741,0.2283787429,0.0252541639,-0.3850671053,0.0924091339,0.165482685,0.0665195286,-0.1504980773,0.1814241707,-0.0332348272,-0.3959296942,-0.1602503955,-0.0522115901,0.1628748178,-0.2076937854,-0.3108922839,-0.1402078569,-0.1141550615,0.1380293667,-0.0407926142,-0.2170617431,-0.3000540435,-0.0936286971,0.1410432905,-0.0588696413,0.2177682668,0.1037803516,0.1914524436,-0.1203961521,-0.4317591786,-0.1487240344,0.1081416756,-0.2924800515,-0.1076147482,0.3776341677,0.0455733202,0.2080194801,-0.1105848923,0.3216450512,0.2376917005,0.066899173,0.1610087454,0.1279921681,-0.1245717034,-0.1512021422,-0.142078042,0.3465422094,0.128989622,0.1471359581,-0.0155485664,0.0163485743,0.0418233871,0.0829828158,-0.2396878004,-0.0824884102,-0.3200368583,-0.1921786666,0.2612736225,0.1546629518,0.0379151292,0.0008637675,0.3312492073,0.0261272565,0.0083661247,-0.1484815776,-0.1806361079,0.0400650427,0.1552491784,0.1572336107,0.3422781229,-0.0142571228,0.3932816386,-0.1144815236,0.1034137383,0.06225916,-0.0817419663,-0.0038878322,0.2324126065,0.15785712,-0.0431561358,0.183517918,-0.2471131682,-0.104632467,0.3485348523,-0.2459665388,0.0869296491,0.2329536527,-0.0123324441,0.1808802485,-0.5816081762,0.0918158442,0.0024792906,0.0669043064,-0.1619982123,0.2604064047,-0.0950392038,0.4181236029,-0.0660849512,0.2435841858,-0.1192760319,-0.2782994807,0.1805499345,-0.15312244,-0.012159599,-0.1923269033,0.1284361035,-0.2399803996,-0.0226501599,0.0775606632,-0.0577873811,-0.1429297924,0.1827362925,0.1133465022,0.2074275613,-0.5552704334,-0.2969094515,-0.0320735425,-0.378795594,-0.1486826241,0.1686326563,-0.4979081154,0.1192707866,-0.055137407,0.245264411,-0.1717564315,-0.1245706826,-0.3170382977,-0.2305708379,-0.0193637144,0.133881554,-0.0549433865,0.1418401748,0.0033795778,-0.1053651124,-0.2393789291,-0.1764801741,-0.2918813229,0.0651777461,-0.5251649618,0.3886308968,-0.1867223978,-0.0471369028,-0.109347865,-0.0461086854,-0.0654087663,0.6399704814,-0.334898591,0.0144050214,0.3095326722,-0.010113867,-0.0965969041,0.1116476655,0.5424685478,0.1066478863,0.1266251504,-0.0549990423,0.2443153262,0.2560414374,0.1710574329,0.0513743982,0.1595277488,-0.1193900257,0.1637128294,1.2782388926,0.1522191167,-0.0518915094,0.0262432639,0.0314820632,-0.0140471822,0.0686671063,-0.1058131158,0.0613840669,-0.3951035738,-0.1308125407,0.179306373,0.1871938407,-0.5139589906,0.0521272793,0.1936313659,-0.0122206649,-0.3019163013,0.3315780163,-0.7171853185,0.5827730298,0.0590841807,0.0178885721,-0.4254792333,0.045399826,0.0945204347,-0.1513158679,0.7113601565,0.0534209572,-0.4360212088,0.0862669274,-0.1918416172,0.0394076109,0.2148355395,0.2028769851,0.0952703208,-0.1703162938,0.147795856,0.1729288846,0.4053848684,-0.0639535859,0.0224064104,0.1169934794,-0.2179405391,-0.1327845156,-0.1501162648,0.064364247,0.3700457215,0.0236489978,0.430801183,-0.0303367078,-0.1774694622,-0.4406905174,0.0884211138,-0.3515691161,-0.0433642752,-0.1870866269,0.0759658292,-0.2960950434,0.2095098794,0.3047899604,-0.1735835522,-0.0213712044,-0.262239337,-0.0159953311,0.3147200942,-0.1036774591,0.037082281,0.2442333847,0.1481226534,0.013170097,0.4719136059,0.2137568444,0.2138777673,0.3546659946,-0.3274190426,-0.1393887997,0.0368320271,0.0073549538,0.234070614,0.1898229718,0.0306465756,-0.2361022979,0.1773872077,-0.167742312,-0.2283785045,0.2950416803,0.1178728044,0.0684980303,-0.1757792085,-0.4709654748,0.0921665505,0.1839773506,-0.2220710367,0.4868579209,-0.8516537547,-0.261679858,0.4181653559,0.0608363412,0.6925716996,-0.2607029974,0.227264449,-0.250442028,0.4874329865,0.020454742,-0.6066097617,0.4200372398,-0.254858017,-0.2041648924,-0.0412223898,-0.1717183441,-0.1559633017,0.7070311904,0.0804174393,0.1725239158,0.1165403426,0.3262236714,-0.0640336424,0.0934551284,0.2966890037,-0.4377833605,0.229028821,-0.0624263473,0.1829893738,0.156969294,-0.0804343894,0.0565100312,0.124040477,-0.2584005296,0.2086893916,0.0717218742,-0.3600730598,0.3634980619,0.0398199335,0.0521443076,-0.243333742,0.6011244059,0.0109047834,-0.2319959551,0.09407603,0.0436988212,0.193117395,-0.0126967691,-0.262334615,-0.1697430909,0.1262998283,0.0967982411,-0.1459383518,-0.1205745712,-0.3000965714,-0.4416499734,-0.223753795,0.360080421,0.6041749716,-0.0505638681,-0.0588486269,0.2081897855,-0.0861713141,0.166884169,-0.0069433674,0.0640020221,-0.1381703764,0.3867056668,-0.0235097799,-0.1242907271,0.2014888227,0.3131961524,-0.1574299932,-0.1682617664,0.2568942606,-0.2654665112,-0.09693712,-0.0573922507,-0.0330521911,0.1389321983,-0.3625167906,0.0169916097,-0.1065589935,0.0166316852,0.0993066281,0.0837628245,0.1839834154,0.107527487,-0.323993057,-0.1815751493,-0.3717830181,-0.0278916862,-0.1365002245,0.0505666025,0.273403585,0.1305761337,0.1053174138,0.3764922321,-0.1317270249,0.1228380725,0.0817987695,0.3973923624,-0.05186712,0.0050210641,-0.1158289313,-0.0291919559,-0.1875379533,0.0908713639,-0.1421469152,0.004255462,0.0848780647,0.2250854522,0.1945449114,-0.0718456656,-0.0755309612,-0.2173605263,0.116221495,-0.4500408173,0.0487676337,0.0160334427,-0.1881897151,0.0467121117,0.5341085196,0.1623094976,-0.4041575789,0.1451513916,-0.0866433457,-0.0505961627,-0.0320249684,0.1041378677,0.1630800366,-0.0049900212,-0.0814534128,-0.0345540866,0.3659764528,0.1690802127,0.4754377306,-0.2272117883,0.0704821721,-0.0292813536,0.1734144688,0.2325785905,-0.3147429526,-0.0599202737,0.3415470421,0.0241225902,0.0282654651,-0.1990050822,0.3771487474,0.0867178366,-0.1169262379,0.269580394,0.2913129628,-0.3984209597,0.1071690992,0.13878043,0.5225743651,-0.3287658095,0.2685762048,0.026906088,0.085723117,0.1760590225,0.2698514163,0.310449481,0.1782119125,-0.0747818798,0.0293806326,0.2496602833,0.3985043764,-0.1035515368,-0.1874758154,-0.4894893765,-0.0736385658,-0.0044096862,-0.0369611047,0.1242682934,-0.2778055072,0.2728984356,-0.2847654819,-0.5857912898,0.1254214495,0.3274540007,-0.2013130635,-0.0886794105,0.0926787481,0.0113809267,-0.04623463,0.0941591561,0.0441685989,0.2097102404,0.5593304038,0.0858864933,0.0181098767,0.1541578323,-0.0717386082,0.0019633719,0.4593269229,-0.3261109591,0.1913291067,-0.1725729853,0.0807391182,0.0200430546,-0.0315704718,0.0498550087,0.5228732228,-0.1188720688,0.1938308328,0.0442509204,0.1886543483,-0.1131996587,0.1667529047,-0.1233745888,-0.3006847203,0.0307742395,-0.0814752579,-0.0131051978,0.028823575,-0.0665220618,0.0668584257,0.2088633776,0.2820904851,0.1485458016,-0.1038457751,0.1028903425,0.0694101751,-0.3599323332,0.1259170175,0.2366672009,-0.2897271514,0.1959314346,0.0505576171,-0.0399407148,0.0708528757,0.2542789876,0.1748011559,0.0303770751,-0.2747746408,-0.3522980511,-0.6728181839,0.302385211,0.0174807664,0.25087744,-0.1023810804,0.0117317839,0.0563799366,-0.0531251766,-0.2462201864,-0.2524305582,-0.038776882,0.3524783552,-0.0957507938,0.0024702032,-0.2875679433,0.399572432,0.1458623558,-0.1972410381,0.065390259,-0.0597794913,-0.0863967463,0.2657533586,-0.4236015975,-0.1760620475,0.0676908493,0.4964187145,0.0649721324,0.2099227756,0.0600785837,0.1801740825,-0.0981862769,0.0451122038,-0.0188293029,-0.0528834648,-0.0837273076,0.2601968348,-0.528018117,-0.0028348397,-0.4991467893,-0.4764001071,0.1246198639,-0.0155331492,0.0108440826,-0.2105778903,-0.2984160483,0.0919267014,0.2645703852,0.4401881397,-0.0507179983,-0.0598962829,-0.2187685072,0.0092310868,0.266627878,-0.6458785534,-0.28847,-0.2124379426,0.1345545799,-0.0403628573,-0.0596350729,-0.4026406705,-0.0723963454,0.2834527791,-0.0711669028,-0.4361755252,-0.1718805879,-0.0504207276,-0.1707316041,-0.2380543798,0.4838347733,-0.0408494323,-0.2001532614,-0.0964371935,-0.1717197299]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2185","title":".map() and distributed training","comments":"Slowness in multiprocessing has been observed in certain environments but not others. We're investigating ;)","body":"Hi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.","comment_length":15,"text":".map() and distributed training\nHi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.\nSlowness in multiprocessing has been observed in certain environments but not others. We're investigating ;)","embeddings":[-0.2537618279,-0.2207167447,0.1177767888,0.0745745301,0.2116166502,-0.1097597107,0.5170732141,0.0102053266,0.2801567912,0.0237633772,0.1822478175,0.3434494138,-0.2290642709,-0.5855442286,0.0941431373,0.0615340844,0.1861155033,0.0366594605,-0.0377356932,-0.2008032501,-0.1529330909,0.0759166479,-0.1542706788,0.1170513257,-0.7540709376,-0.1523153037,-0.1834535897,0.1027398556,0.2253180742,-0.3499087691,0.2342278361,0.2213348448,0.4013623595,0.5336914062,-0.0001272091,0.0312626697,0.1878611445,-0.1983309686,0.0397178754,-0.0476766787,0.1641900688,0.0050253472,0.1936113387,-0.2953366041,0.1310725659,-0.1476733088,0.3262221217,-0.4418162107,0.4140013158,-0.1077297255,-0.0018774115,0.0943896994,-0.5714957118,0.1398764253,-0.2214399129,0.1234803572,0.1711140275,0.2180822492,0.3029266596,-0.3416551054,-0.3789148629,0.1582272202,-0.1466236264,0.1340594739,0.4135627449,0.0366018042,-0.0789881125,-0.3487006128,-0.075750038,-0.1218584478,0.1076876521,-0.0743067637,-0.1881792545,-0.4665382504,-0.2733976543,-0.0016580709,-0.0024628954,0.116243735,-0.0233727768,-0.1197138727,-0.5124091506,-0.0625246316,0.0552944317,0.1326544136,-0.2732613087,0.4564465284,-0.0254738089,0.3163393438,0.1419928521,0.0543378443,0.0837460086,-0.1719923168,0.1296369135,0.3715832233,0.0275528189,0.0141743347,0.0489273742,-0.3223410249,-0.0232889,0.0356097072,0.2321069092,0.2598097324,-0.0479859076,0.265766263,0.4007704556,0.0927509293,-0.1391413957,0.6415418983,0.0301112514,-0.2004719824,-0.6418218613,-0.1724125892,-0.3519452512,-0.0627355203,0.3335886896,0.0427098982,-0.2098451257,0.0722271129,0.4604047239,-0.2394168377,-0.0065294988,-0.1325436831,0.0212120153,0.2508155704,0.1287937164,0.437037617,-0.0555269644,0.1014266536,-0.190160498,0.1589486897,-0.0596041419,-0.3029897511,-0.3049105704,0.316068083,0.1618765295,0.1877589226,0.4239696264,-0.1001108065,0.289442122,-0.1613698006,0.3981889188,-0.1120462343,0.6585771441,0.1044589132,-0.2513783276,0.2223655581,0.1862223744,0.3770996928,-0.1535631716,0.335416913,-0.5022829771,-0.4383485317,0.4798748195,-0.0732970536,-0.0095938472,0.0692022368,-0.1767223477,0.0427591875,0.5391663909,-0.2560784817,0.134424001,-0.3575458229,-0.2090658844,-0.1776281595,0.2568801343,0.4551616609,0.3137147129,-0.341816932,0.1241682246,0.3162250221,0.2081830502,0.2290944606,-0.6239327788,0.3898105621,-0.2199796587,0.2747520506,0.3872870505,-0.3571222723,-0.1435228735,0.2945982218,-0.0989739373,0.0233595725,-0.1570941806,-0.0249821246,0.4641161859,-0.0252317898,0.1258820295,0.2964549661,-0.216262728,0.2001035213,-0.261736244,0.0169189703,0.0804381073,-0.0070854248,0.1947211921,0.0033709346,-0.0128012626,0.1196192652,0.2070769817,0.0102299741,0.2283787429,0.0252541639,-0.3850671053,0.0924091339,0.165482685,0.0665195286,-0.1504980773,0.1814241707,-0.0332348272,-0.3959296942,-0.1602503955,-0.0522115901,0.1628748178,-0.2076937854,-0.3108922839,-0.1402078569,-0.1141550615,0.1380293667,-0.0407926142,-0.2170617431,-0.3000540435,-0.0936286971,0.1410432905,-0.0588696413,0.2177682668,0.1037803516,0.1914524436,-0.1203961521,-0.4317591786,-0.1487240344,0.1081416756,-0.2924800515,-0.1076147482,0.3776341677,0.0455733202,0.2080194801,-0.1105848923,0.3216450512,0.2376917005,0.066899173,0.1610087454,0.1279921681,-0.1245717034,-0.1512021422,-0.142078042,0.3465422094,0.128989622,0.1471359581,-0.0155485664,0.0163485743,0.0418233871,0.0829828158,-0.2396878004,-0.0824884102,-0.3200368583,-0.1921786666,0.2612736225,0.1546629518,0.0379151292,0.0008637675,0.3312492073,0.0261272565,0.0083661247,-0.1484815776,-0.1806361079,0.0400650427,0.1552491784,0.1572336107,0.3422781229,-0.0142571228,0.3932816386,-0.1144815236,0.1034137383,0.06225916,-0.0817419663,-0.0038878322,0.2324126065,0.15785712,-0.0431561358,0.183517918,-0.2471131682,-0.104632467,0.3485348523,-0.2459665388,0.0869296491,0.2329536527,-0.0123324441,0.1808802485,-0.5816081762,0.0918158442,0.0024792906,0.0669043064,-0.1619982123,0.2604064047,-0.0950392038,0.4181236029,-0.0660849512,0.2435841858,-0.1192760319,-0.2782994807,0.1805499345,-0.15312244,-0.012159599,-0.1923269033,0.1284361035,-0.2399803996,-0.0226501599,0.0775606632,-0.0577873811,-0.1429297924,0.1827362925,0.1133465022,0.2074275613,-0.5552704334,-0.2969094515,-0.0320735425,-0.378795594,-0.1486826241,0.1686326563,-0.4979081154,0.1192707866,-0.055137407,0.245264411,-0.1717564315,-0.1245706826,-0.3170382977,-0.2305708379,-0.0193637144,0.133881554,-0.0549433865,0.1418401748,0.0033795778,-0.1053651124,-0.2393789291,-0.1764801741,-0.2918813229,0.0651777461,-0.5251649618,0.3886308968,-0.1867223978,-0.0471369028,-0.109347865,-0.0461086854,-0.0654087663,0.6399704814,-0.334898591,0.0144050214,0.3095326722,-0.010113867,-0.0965969041,0.1116476655,0.5424685478,0.1066478863,0.1266251504,-0.0549990423,0.2443153262,0.2560414374,0.1710574329,0.0513743982,0.1595277488,-0.1193900257,0.1637128294,1.2782388926,0.1522191167,-0.0518915094,0.0262432639,0.0314820632,-0.0140471822,0.0686671063,-0.1058131158,0.0613840669,-0.3951035738,-0.1308125407,0.179306373,0.1871938407,-0.5139589906,0.0521272793,0.1936313659,-0.0122206649,-0.3019163013,0.3315780163,-0.7171853185,0.5827730298,0.0590841807,0.0178885721,-0.4254792333,0.045399826,0.0945204347,-0.1513158679,0.7113601565,0.0534209572,-0.4360212088,0.0862669274,-0.1918416172,0.0394076109,0.2148355395,0.2028769851,0.0952703208,-0.1703162938,0.147795856,0.1729288846,0.4053848684,-0.0639535859,0.0224064104,0.1169934794,-0.2179405391,-0.1327845156,-0.1501162648,0.064364247,0.3700457215,0.0236489978,0.430801183,-0.0303367078,-0.1774694622,-0.4406905174,0.0884211138,-0.3515691161,-0.0433642752,-0.1870866269,0.0759658292,-0.2960950434,0.2095098794,0.3047899604,-0.1735835522,-0.0213712044,-0.262239337,-0.0159953311,0.3147200942,-0.1036774591,0.037082281,0.2442333847,0.1481226534,0.013170097,0.4719136059,0.2137568444,0.2138777673,0.3546659946,-0.3274190426,-0.1393887997,0.0368320271,0.0073549538,0.234070614,0.1898229718,0.0306465756,-0.2361022979,0.1773872077,-0.167742312,-0.2283785045,0.2950416803,0.1178728044,0.0684980303,-0.1757792085,-0.4709654748,0.0921665505,0.1839773506,-0.2220710367,0.4868579209,-0.8516537547,-0.261679858,0.4181653559,0.0608363412,0.6925716996,-0.2607029974,0.227264449,-0.250442028,0.4874329865,0.020454742,-0.6066097617,0.4200372398,-0.254858017,-0.2041648924,-0.0412223898,-0.1717183441,-0.1559633017,0.7070311904,0.0804174393,0.1725239158,0.1165403426,0.3262236714,-0.0640336424,0.0934551284,0.2966890037,-0.4377833605,0.229028821,-0.0624263473,0.1829893738,0.156969294,-0.0804343894,0.0565100312,0.124040477,-0.2584005296,0.2086893916,0.0717218742,-0.3600730598,0.3634980619,0.0398199335,0.0521443076,-0.243333742,0.6011244059,0.0109047834,-0.2319959551,0.09407603,0.0436988212,0.193117395,-0.0126967691,-0.262334615,-0.1697430909,0.1262998283,0.0967982411,-0.1459383518,-0.1205745712,-0.3000965714,-0.4416499734,-0.223753795,0.360080421,0.6041749716,-0.0505638681,-0.0588486269,0.2081897855,-0.0861713141,0.166884169,-0.0069433674,0.0640020221,-0.1381703764,0.3867056668,-0.0235097799,-0.1242907271,0.2014888227,0.3131961524,-0.1574299932,-0.1682617664,0.2568942606,-0.2654665112,-0.09693712,-0.0573922507,-0.0330521911,0.1389321983,-0.3625167906,0.0169916097,-0.1065589935,0.0166316852,0.0993066281,0.0837628245,0.1839834154,0.107527487,-0.323993057,-0.1815751493,-0.3717830181,-0.0278916862,-0.1365002245,0.0505666025,0.273403585,0.1305761337,0.1053174138,0.3764922321,-0.1317270249,0.1228380725,0.0817987695,0.3973923624,-0.05186712,0.0050210641,-0.1158289313,-0.0291919559,-0.1875379533,0.0908713639,-0.1421469152,0.004255462,0.0848780647,0.2250854522,0.1945449114,-0.0718456656,-0.0755309612,-0.2173605263,0.116221495,-0.4500408173,0.0487676337,0.0160334427,-0.1881897151,0.0467121117,0.5341085196,0.1623094976,-0.4041575789,0.1451513916,-0.0866433457,-0.0505961627,-0.0320249684,0.1041378677,0.1630800366,-0.0049900212,-0.0814534128,-0.0345540866,0.3659764528,0.1690802127,0.4754377306,-0.2272117883,0.0704821721,-0.0292813536,0.1734144688,0.2325785905,-0.3147429526,-0.0599202737,0.3415470421,0.0241225902,0.0282654651,-0.1990050822,0.3771487474,0.0867178366,-0.1169262379,0.269580394,0.2913129628,-0.3984209597,0.1071690992,0.13878043,0.5225743651,-0.3287658095,0.2685762048,0.026906088,0.085723117,0.1760590225,0.2698514163,0.310449481,0.1782119125,-0.0747818798,0.0293806326,0.2496602833,0.3985043764,-0.1035515368,-0.1874758154,-0.4894893765,-0.0736385658,-0.0044096862,-0.0369611047,0.1242682934,-0.2778055072,0.2728984356,-0.2847654819,-0.5857912898,0.1254214495,0.3274540007,-0.2013130635,-0.0886794105,0.0926787481,0.0113809267,-0.04623463,0.0941591561,0.0441685989,0.2097102404,0.5593304038,0.0858864933,0.0181098767,0.1541578323,-0.0717386082,0.0019633719,0.4593269229,-0.3261109591,0.1913291067,-0.1725729853,0.0807391182,0.0200430546,-0.0315704718,0.0498550087,0.5228732228,-0.1188720688,0.1938308328,0.0442509204,0.1886543483,-0.1131996587,0.1667529047,-0.1233745888,-0.3006847203,0.0307742395,-0.0814752579,-0.0131051978,0.028823575,-0.0665220618,0.0668584257,0.2088633776,0.2820904851,0.1485458016,-0.1038457751,0.1028903425,0.0694101751,-0.3599323332,0.1259170175,0.2366672009,-0.2897271514,0.1959314346,0.0505576171,-0.0399407148,0.0708528757,0.2542789876,0.1748011559,0.0303770751,-0.2747746408,-0.3522980511,-0.6728181839,0.302385211,0.0174807664,0.25087744,-0.1023810804,0.0117317839,0.0563799366,-0.0531251766,-0.2462201864,-0.2524305582,-0.038776882,0.3524783552,-0.0957507938,0.0024702032,-0.2875679433,0.399572432,0.1458623558,-0.1972410381,0.065390259,-0.0597794913,-0.0863967463,0.2657533586,-0.4236015975,-0.1760620475,0.0676908493,0.4964187145,0.0649721324,0.2099227756,0.0600785837,0.1801740825,-0.0981862769,0.0451122038,-0.0188293029,-0.0528834648,-0.0837273076,0.2601968348,-0.528018117,-0.0028348397,-0.4991467893,-0.4764001071,0.1246198639,-0.0155331492,0.0108440826,-0.2105778903,-0.2984160483,0.0919267014,0.2645703852,0.4401881397,-0.0507179983,-0.0598962829,-0.2187685072,0.0092310868,0.266627878,-0.6458785534,-0.28847,-0.2124379426,0.1345545799,-0.0403628573,-0.0596350729,-0.4026406705,-0.0723963454,0.2834527791,-0.0711669028,-0.4361755252,-0.1718805879,-0.0504207276,-0.1707316041,-0.2380543798,0.4838347733,-0.0408494323,-0.2001532614,-0.0964371935,-0.1717197299]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2185","title":".map() and distributed training","comments":"So to answer my initial question, I was just doing something stupid as I was not re-giving the `preprocessing_num_workers` arguments when launching the distributed training (and it was then set to `None`). I initially thought the hash was computed only with the `tokenize_function` but it's all arguments. Thanks @lhoestq for clarifying!","body":"Hi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.","comment_length":51,"text":".map() and distributed training\nHi,\r\nI have a question regarding distributed training and the `.map` call on a dataset.\r\n\r\nI have a local dataset \"my_custom_dataset\" that I am loading with `datasets = load_from_disk(dataset_path=my_path)`.\r\n`dataset` is then tokenized:\r\n```python\r\ndatasets = load_from_disk(dataset_path=my_path)\r\n\r\n[...]\r\n\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name])\r\n\r\nlogger.info(\"Mapping dataset to tokenized dataset.\")\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=preprocessing_num_workers,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n)\r\n```\r\nI am using 31 workers (`preprocessing_num_workers=31`) and thus it creates 31 `cache*.arrow` files in `my_path\/train` (there is only a train split).\r\nWhen I relaunch the script, the map is tokenization is skipped in favor of loading the 31 previously cached files, and that's perfect.\r\n\r\nEverything so far was done by launching a **single process script**.\r\nI now launch the same training script in **distributed mode** (`pytorch -m torch.distributed.launch --nproc_per_node 2`). However, once it reaches the map call, it re-does the tokenization... instead of loading the 31 cached files. \r\n\r\nI tried adding the `cache_file_name` argument: `cache_file_name={\"train\": my_path\/one_of_the_arrow_file}`, but I can't give the 31 cached files, so it probably isn't the right way to do it.\r\n\r\n**My question: what is the best way to load cached files if they were pre-processed and dumped in multiple arrow files?** It seems automatically handled for single processes but fails on distributed training.\r\n\r\n- I am following the same structure as the examples of transformers (more specifically [run_clm.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_clm.py) in my case)\r\n- I am using 1.5.0 version of datasets if that matters.\nSo to answer my initial question, I was just doing something stupid as I was not re-giving the `preprocessing_num_workers` arguments when launching the distributed training (and it was then set to `None`). I initially thought the hash was computed only with the `tokenize_function` but it's all arguments. Thanks @lhoestq for clarifying!","embeddings":[-0.2537618279,-0.2207167447,0.1177767888,0.0745745301,0.2116166502,-0.1097597107,0.5170732141,0.0102053266,0.2801567912,0.0237633772,0.1822478175,0.3434494138,-0.2290642709,-0.5855442286,0.0941431373,0.0615340844,0.1861155033,0.0366594605,-0.0377356932,-0.2008032501,-0.1529330909,0.0759166479,-0.1542706788,0.1170513257,-0.7540709376,-0.1523153037,-0.1834535897,0.1027398556,0.2253180742,-0.3499087691,0.2342278361,0.2213348448,0.4013623595,0.5336914062,-0.0001272091,0.0312626697,0.1878611445,-0.1983309686,0.0397178754,-0.0476766787,0.1641900688,0.0050253472,0.1936113387,-0.2953366041,0.1310725659,-0.1476733088,0.3262221217,-0.4418162107,0.4140013158,-0.1077297255,-0.0018774115,0.0943896994,-0.5714957118,0.1398764253,-0.2214399129,0.1234803572,0.1711140275,0.2180822492,0.3029266596,-0.3416551054,-0.3789148629,0.1582272202,-0.1466236264,0.1340594739,0.4135627449,0.0366018042,-0.0789881125,-0.3487006128,-0.075750038,-0.1218584478,0.1076876521,-0.0743067637,-0.1881792545,-0.4665382504,-0.2733976543,-0.0016580709,-0.0024628954,0.116243735,-0.0233727768,-0.1197138727,-0.5124091506,-0.0625246316,0.0552944317,0.1326544136,-0.2732613087,0.4564465284,-0.0254738089,0.3163393438,0.1419928521,0.0543378443,0.0837460086,-0.1719923168,0.1296369135,0.3715832233,0.0275528189,0.0141743347,0.0489273742,-0.3223410249,-0.0232889,0.0356097072,0.2321069092,0.2598097324,-0.0479859076,0.265766263,0.4007704556,0.0927509293,-0.1391413957,0.6415418983,0.0301112514,-0.2004719824,-0.6418218613,-0.1724125892,-0.3519452512,-0.0627355203,0.3335886896,0.0427098982,-0.2098451257,0.0722271129,0.4604047239,-0.2394168377,-0.0065294988,-0.1325436831,0.0212120153,0.2508155704,0.1287937164,0.437037617,-0.0555269644,0.1014266536,-0.190160498,0.1589486897,-0.0596041419,-0.3029897511,-0.3049105704,0.316068083,0.1618765295,0.1877589226,0.4239696264,-0.1001108065,0.289442122,-0.1613698006,0.3981889188,-0.1120462343,0.6585771441,0.1044589132,-0.2513783276,0.2223655581,0.1862223744,0.3770996928,-0.1535631716,0.335416913,-0.5022829771,-0.4383485317,0.4798748195,-0.0732970536,-0.0095938472,0.0692022368,-0.1767223477,0.0427591875,0.5391663909,-0.2560784817,0.134424001,-0.3575458229,-0.2090658844,-0.1776281595,0.2568801343,0.4551616609,0.3137147129,-0.341816932,0.1241682246,0.3162250221,0.2081830502,0.2290944606,-0.6239327788,0.3898105621,-0.2199796587,0.2747520506,0.3872870505,-0.3571222723,-0.1435228735,0.2945982218,-0.0989739373,0.0233595725,-0.1570941806,-0.0249821246,0.4641161859,-0.0252317898,0.1258820295,0.2964549661,-0.216262728,0.2001035213,-0.261736244,0.0169189703,0.0804381073,-0.0070854248,0.1947211921,0.0033709346,-0.0128012626,0.1196192652,0.2070769817,0.0102299741,0.2283787429,0.0252541639,-0.3850671053,0.0924091339,0.165482685,0.0665195286,-0.1504980773,0.1814241707,-0.0332348272,-0.3959296942,-0.1602503955,-0.0522115901,0.1628748178,-0.2076937854,-0.3108922839,-0.1402078569,-0.1141550615,0.1380293667,-0.0407926142,-0.2170617431,-0.3000540435,-0.0936286971,0.1410432905,-0.0588696413,0.2177682668,0.1037803516,0.1914524436,-0.1203961521,-0.4317591786,-0.1487240344,0.1081416756,-0.2924800515,-0.1076147482,0.3776341677,0.0455733202,0.2080194801,-0.1105848923,0.3216450512,0.2376917005,0.066899173,0.1610087454,0.1279921681,-0.1245717034,-0.1512021422,-0.142078042,0.3465422094,0.128989622,0.1471359581,-0.0155485664,0.0163485743,0.0418233871,0.0829828158,-0.2396878004,-0.0824884102,-0.3200368583,-0.1921786666,0.2612736225,0.1546629518,0.0379151292,0.0008637675,0.3312492073,0.0261272565,0.0083661247,-0.1484815776,-0.1806361079,0.0400650427,0.1552491784,0.1572336107,0.3422781229,-0.0142571228,0.3932816386,-0.1144815236,0.1034137383,0.06225916,-0.0817419663,-0.0038878322,0.2324126065,0.15785712,-0.0431561358,0.183517918,-0.2471131682,-0.104632467,0.3485348523,-0.2459665388,0.0869296491,0.2329536527,-0.0123324441,0.1808802485,-0.5816081762,0.0918158442,0.0024792906,0.0669043064,-0.1619982123,0.2604064047,-0.0950392038,0.4181236029,-0.0660849512,0.2435841858,-0.1192760319,-0.2782994807,0.1805499345,-0.15312244,-0.012159599,-0.1923269033,0.1284361035,-0.2399803996,-0.0226501599,0.0775606632,-0.0577873811,-0.1429297924,0.1827362925,0.1133465022,0.2074275613,-0.5552704334,-0.2969094515,-0.0320735425,-0.378795594,-0.1486826241,0.1686326563,-0.4979081154,0.1192707866,-0.055137407,0.245264411,-0.1717564315,-0.1245706826,-0.3170382977,-0.2305708379,-0.0193637144,0.133881554,-0.0549433865,0.1418401748,0.0033795778,-0.1053651124,-0.2393789291,-0.1764801741,-0.2918813229,0.0651777461,-0.5251649618,0.3886308968,-0.1867223978,-0.0471369028,-0.109347865,-0.0461086854,-0.0654087663,0.6399704814,-0.334898591,0.0144050214,0.3095326722,-0.010113867,-0.0965969041,0.1116476655,0.5424685478,0.1066478863,0.1266251504,-0.0549990423,0.2443153262,0.2560414374,0.1710574329,0.0513743982,0.1595277488,-0.1193900257,0.1637128294,1.2782388926,0.1522191167,-0.0518915094,0.0262432639,0.0314820632,-0.0140471822,0.0686671063,-0.1058131158,0.0613840669,-0.3951035738,-0.1308125407,0.179306373,0.1871938407,-0.5139589906,0.0521272793,0.1936313659,-0.0122206649,-0.3019163013,0.3315780163,-0.7171853185,0.5827730298,0.0590841807,0.0178885721,-0.4254792333,0.045399826,0.0945204347,-0.1513158679,0.7113601565,0.0534209572,-0.4360212088,0.0862669274,-0.1918416172,0.0394076109,0.2148355395,0.2028769851,0.0952703208,-0.1703162938,0.147795856,0.1729288846,0.4053848684,-0.0639535859,0.0224064104,0.1169934794,-0.2179405391,-0.1327845156,-0.1501162648,0.064364247,0.3700457215,0.0236489978,0.430801183,-0.0303367078,-0.1774694622,-0.4406905174,0.0884211138,-0.3515691161,-0.0433642752,-0.1870866269,0.0759658292,-0.2960950434,0.2095098794,0.3047899604,-0.1735835522,-0.0213712044,-0.262239337,-0.0159953311,0.3147200942,-0.1036774591,0.037082281,0.2442333847,0.1481226534,0.013170097,0.4719136059,0.2137568444,0.2138777673,0.3546659946,-0.3274190426,-0.1393887997,0.0368320271,0.0073549538,0.234070614,0.1898229718,0.0306465756,-0.2361022979,0.1773872077,-0.167742312,-0.2283785045,0.2950416803,0.1178728044,0.0684980303,-0.1757792085,-0.4709654748,0.0921665505,0.1839773506,-0.2220710367,0.4868579209,-0.8516537547,-0.261679858,0.4181653559,0.0608363412,0.6925716996,-0.2607029974,0.227264449,-0.250442028,0.4874329865,0.020454742,-0.6066097617,0.4200372398,-0.254858017,-0.2041648924,-0.0412223898,-0.1717183441,-0.1559633017,0.7070311904,0.0804174393,0.1725239158,0.1165403426,0.3262236714,-0.0640336424,0.0934551284,0.2966890037,-0.4377833605,0.229028821,-0.0624263473,0.1829893738,0.156969294,-0.0804343894,0.0565100312,0.124040477,-0.2584005296,0.2086893916,0.0717218742,-0.3600730598,0.3634980619,0.0398199335,0.0521443076,-0.243333742,0.6011244059,0.0109047834,-0.2319959551,0.09407603,0.0436988212,0.193117395,-0.0126967691,-0.262334615,-0.1697430909,0.1262998283,0.0967982411,-0.1459383518,-0.1205745712,-0.3000965714,-0.4416499734,-0.223753795,0.360080421,0.6041749716,-0.0505638681,-0.0588486269,0.2081897855,-0.0861713141,0.166884169,-0.0069433674,0.0640020221,-0.1381703764,0.3867056668,-0.0235097799,-0.1242907271,0.2014888227,0.3131961524,-0.1574299932,-0.1682617664,0.2568942606,-0.2654665112,-0.09693712,-0.0573922507,-0.0330521911,0.1389321983,-0.3625167906,0.0169916097,-0.1065589935,0.0166316852,0.0993066281,0.0837628245,0.1839834154,0.107527487,-0.323993057,-0.1815751493,-0.3717830181,-0.0278916862,-0.1365002245,0.0505666025,0.273403585,0.1305761337,0.1053174138,0.3764922321,-0.1317270249,0.1228380725,0.0817987695,0.3973923624,-0.05186712,0.0050210641,-0.1158289313,-0.0291919559,-0.1875379533,0.0908713639,-0.1421469152,0.004255462,0.0848780647,0.2250854522,0.1945449114,-0.0718456656,-0.0755309612,-0.2173605263,0.116221495,-0.4500408173,0.0487676337,0.0160334427,-0.1881897151,0.0467121117,0.5341085196,0.1623094976,-0.4041575789,0.1451513916,-0.0866433457,-0.0505961627,-0.0320249684,0.1041378677,0.1630800366,-0.0049900212,-0.0814534128,-0.0345540866,0.3659764528,0.1690802127,0.4754377306,-0.2272117883,0.0704821721,-0.0292813536,0.1734144688,0.2325785905,-0.3147429526,-0.0599202737,0.3415470421,0.0241225902,0.0282654651,-0.1990050822,0.3771487474,0.0867178366,-0.1169262379,0.269580394,0.2913129628,-0.3984209597,0.1071690992,0.13878043,0.5225743651,-0.3287658095,0.2685762048,0.026906088,0.085723117,0.1760590225,0.2698514163,0.310449481,0.1782119125,-0.0747818798,0.0293806326,0.2496602833,0.3985043764,-0.1035515368,-0.1874758154,-0.4894893765,-0.0736385658,-0.0044096862,-0.0369611047,0.1242682934,-0.2778055072,0.2728984356,-0.2847654819,-0.5857912898,0.1254214495,0.3274540007,-0.2013130635,-0.0886794105,0.0926787481,0.0113809267,-0.04623463,0.0941591561,0.0441685989,0.2097102404,0.5593304038,0.0858864933,0.0181098767,0.1541578323,-0.0717386082,0.0019633719,0.4593269229,-0.3261109591,0.1913291067,-0.1725729853,0.0807391182,0.0200430546,-0.0315704718,0.0498550087,0.5228732228,-0.1188720688,0.1938308328,0.0442509204,0.1886543483,-0.1131996587,0.1667529047,-0.1233745888,-0.3006847203,0.0307742395,-0.0814752579,-0.0131051978,0.028823575,-0.0665220618,0.0668584257,0.2088633776,0.2820904851,0.1485458016,-0.1038457751,0.1028903425,0.0694101751,-0.3599323332,0.1259170175,0.2366672009,-0.2897271514,0.1959314346,0.0505576171,-0.0399407148,0.0708528757,0.2542789876,0.1748011559,0.0303770751,-0.2747746408,-0.3522980511,-0.6728181839,0.302385211,0.0174807664,0.25087744,-0.1023810804,0.0117317839,0.0563799366,-0.0531251766,-0.2462201864,-0.2524305582,-0.038776882,0.3524783552,-0.0957507938,0.0024702032,-0.2875679433,0.399572432,0.1458623558,-0.1972410381,0.065390259,-0.0597794913,-0.0863967463,0.2657533586,-0.4236015975,-0.1760620475,0.0676908493,0.4964187145,0.0649721324,0.2099227756,0.0600785837,0.1801740825,-0.0981862769,0.0451122038,-0.0188293029,-0.0528834648,-0.0837273076,0.2601968348,-0.528018117,-0.0028348397,-0.4991467893,-0.4764001071,0.1246198639,-0.0155331492,0.0108440826,-0.2105778903,-0.2984160483,0.0919267014,0.2645703852,0.4401881397,-0.0507179983,-0.0598962829,-0.2187685072,0.0092310868,0.266627878,-0.6458785534,-0.28847,-0.2124379426,0.1345545799,-0.0403628573,-0.0596350729,-0.4026406705,-0.0723963454,0.2834527791,-0.0711669028,-0.4361755252,-0.1718805879,-0.0504207276,-0.1707316041,-0.2380543798,0.4838347733,-0.0408494323,-0.2001532614,-0.0964371935,-0.1717197299]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"Hi ! Can you try to increase the block size ? For example\r\n```python\r\nblock_size_10MB = 10<<20\r\nload_dataset(\"json\", ..., block_size=block_size_10MB)\r\n```\r\nThe block size corresponds to how much bytes to process at a time from the input stream.\r\nThis will determine multi-threading granularity as well as the size of individual chunks in the dataset.\r\n\r\nYou can also try with bigger block sizes if needed","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":64,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nHi ! Can you try to increase the block size ? For example\r\n```python\r\nblock_size_10MB = 10<<20\r\nload_dataset(\"json\", ..., block_size=block_size_10MB)\r\n```\r\nThe block size corresponds to how much bytes to process at a time from the input stream.\r\nThis will determine multi-threading granularity as well as the size of individual chunks in the dataset.\r\n\r\nYou can also try with bigger block sizes if needed","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"Hi @lhoestq! Thank you for your prompt reply.\r\nI have experimented with (10<<20, 10<<28, 10<<30, 10<<33, 10<<34), since my machine has 192G of memory, but it's either the above-mentioned error or processed killed because of OOM.\r\n\r\nCould you give me a bit of background on why block size needs to be exactly calibrated?\r\nTo my understanding, small block sized should run just fine despite its slowness..\r\n\r\n\r\n","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":66,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nHi @lhoestq! Thank you for your prompt reply.\r\nI have experimented with (10<<20, 10<<28, 10<<30, 10<<33, 10<<34), since my machine has 192G of memory, but it's either the above-mentioned error or processed killed because of OOM.\r\n\r\nCould you give me a bit of background on why block size needs to be exactly calibrated?\r\nTo my understanding, small block sized should run just fine despite its slowness..\r\n\r\n\r\n","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"We're using the JSON loader of pyarrow. It parses the file chunk by chunk to load the dataset.\r\nThis issue happens when there's no delimiter in one chunk of data. For json line, the delimiter is the end of line.\r\nSo with a big value for chunk_size this should have worked unless you have one extremely long line in your file.\r\n\r\nAlso what version of pyarrow are you using ?\r\n\r\nFInally I wonder if it could be an issue on pyarrow's side when using big json files. (I haven't tested big json files like yours)","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":95,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nWe're using the JSON loader of pyarrow. It parses the file chunk by chunk to load the dataset.\r\nThis issue happens when there's no delimiter in one chunk of data. For json line, the delimiter is the end of line.\r\nSo with a big value for chunk_size this should have worked unless you have one extremely long line in your file.\r\n\r\nAlso what version of pyarrow are you using ?\r\n\r\nFInally I wonder if it could be an issue on pyarrow's side when using big json files. (I haven't tested big json files like yours)","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"I'm using `pyarrow==3.0.0` with `datasets==1.5.0`.\r\n\r\nYour point totally makes sense. I will check if my jsonl file contains an extremely long file and let you know. \r\n\r\nHere are some different error messages that I got when tweaking `block_size`. I also suspect that this is related to the pyarrow... but I guess it would be wonderful if datasesets could give a clear guide on how to play with large datasets! (I am suddenly experiencing various issue when working with large datasets.. e.g. #1992 )\r\n```python\r\n return paj.ReadOptions(use_threads=self.use_threads, block_size=self.block_size)\r\n File \"pyarrow\/_json.pyx\", line 56, in pyarrow._json.ReadOptions.__init__\r\n File \"pyarrow\/_json.pyx\", line 81, in pyarrow._json.ReadOptions.block_size.__set__\r\nOverflowError: value too large to convert to int32_t\r\n```\r\n\r\n```python\r\n\r\nline 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Exceeded maximum rows\r\n```","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":137,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nI'm using `pyarrow==3.0.0` with `datasets==1.5.0`.\r\n\r\nYour point totally makes sense. I will check if my jsonl file contains an extremely long file and let you know. \r\n\r\nHere are some different error messages that I got when tweaking `block_size`. I also suspect that this is related to the pyarrow... but I guess it would be wonderful if datasesets could give a clear guide on how to play with large datasets! (I am suddenly experiencing various issue when working with large datasets.. e.g. #1992 )\r\n```python\r\n return paj.ReadOptions(use_threads=self.use_threads, block_size=self.block_size)\r\n File \"pyarrow\/_json.pyx\", line 56, in pyarrow._json.ReadOptions.__init__\r\n File \"pyarrow\/_json.pyx\", line 81, in pyarrow._json.ReadOptions.block_size.__set__\r\nOverflowError: value too large to convert to int32_t\r\n```\r\n\r\n```python\r\n\r\nline 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Exceeded maximum rows\r\n```","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"I am getting the same error. When I tweak the block_size, I also find:\r\n`OverflowError: value too large to convert to int32_t`\r\nand \r\n`pyarrow.lib.ArrowInvalid: Exceeded maximum rows`\r\n","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":27,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nI am getting the same error. When I tweak the block_size, I also find:\r\n`OverflowError: value too large to convert to int32_t`\r\nand \r\n`pyarrow.lib.ArrowInvalid: Exceeded maximum rows`\r\n","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"I made more tests. I used a smaller dataset and I was getting the same error, which means that it was not necessarily linked to the dataset size. To make both my smaller and larger datasets work, I got rid of lists with the json file. I had the following data format:\r\n```python\r\n[\r\n {'key': \"a\", 'value': ['one', 'two', 'three']},\r\n {'key': \"b\", 'value': ['four', 'five', 'six']}\r\n]\r\n```\r\nI changed to:\r\n\r\n```python\r\n {'key': \"a\", 'value': 'one\\ntwo\\nthree'},\r\n {'key': \"b\", 'value': 'four\\nfive\\nsix']}\r\n```\r\nand that worked!\r\n\r\nI used the following to reformat my json file:\r\n```python\r\nwith open(file_name, \"w\", encoding=\"utf-8\") as f:\r\n for item in list_:\r\n f.write(json.dumps(item) + \"\\n\")\r\n```\r\nThis works with `block_size_10MB = 10 << 20` or without specifying `block_size`.","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":120,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nI made more tests. I used a smaller dataset and I was getting the same error, which means that it was not necessarily linked to the dataset size. To make both my smaller and larger datasets work, I got rid of lists with the json file. I had the following data format:\r\n```python\r\n[\r\n {'key': \"a\", 'value': ['one', 'two', 'three']},\r\n {'key': \"b\", 'value': ['four', 'five', 'six']}\r\n]\r\n```\r\nI changed to:\r\n\r\n```python\r\n {'key': \"a\", 'value': 'one\\ntwo\\nthree'},\r\n {'key': \"b\", 'value': 'four\\nfive\\nsix']}\r\n```\r\nand that worked!\r\n\r\nI used the following to reformat my json file:\r\n```python\r\nwith open(file_name, \"w\", encoding=\"utf-8\") as f:\r\n for item in list_:\r\n f.write(json.dumps(item) + \"\\n\")\r\n```\r\nThis works with `block_size_10MB = 10 << 20` or without specifying `block_size`.","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"Thanks @hwijeen for reporting and thanks @jpilaul for pointing this out.\r\n\r\nIndeed, those are different JSON-like formats:\r\n- the first one is the **standard JSON** format: all the file content is JSON-valid, thus all content is either a JSON object (between curly brackets `{...}`) or a JSON array (between square brackets `[...]`)\r\n- the second one is called **JSON Lines**: the entire file content is not JSON-valid, but only every line (newline-delimited) is JSON-valid\r\n\r\nCurrently PyArrow only supports **JSON Lines** format: \r\n- https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.json.read_json.html\r\n > Currently only the line-delimited JSON format is supported.\r\n- https:\/\/arrow.apache.org\/docs\/python\/json.html\r\n > Arrow supports reading columnar data from line-delimited JSON files.","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":104,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nThanks @hwijeen for reporting and thanks @jpilaul for pointing this out.\r\n\r\nIndeed, those are different JSON-like formats:\r\n- the first one is the **standard JSON** format: all the file content is JSON-valid, thus all content is either a JSON object (between curly brackets `{...}`) or a JSON array (between square brackets `[...]`)\r\n- the second one is called **JSON Lines**: the entire file content is not JSON-valid, but only every line (newline-delimited) is JSON-valid\r\n\r\nCurrently PyArrow only supports **JSON Lines** format: \r\n- https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.json.read_json.html\r\n > Currently only the line-delimited JSON format is supported.\r\n- https:\/\/arrow.apache.org\/docs\/python\/json.html\r\n > Arrow supports reading columnar data from line-delimited JSON files.","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"Thanks @albertvillanova for your explanation, it is helpful to know (maybe add to docs?)!\r\nHowever, the problem I described above happened when I was dealing with jsonl files \ud83d\ude3f\r\nAlthough I did not thoroughly inspect, I suspect the cause was the one extremely long document in my case.","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":48,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nThanks @albertvillanova for your explanation, it is helpful to know (maybe add to docs?)!\r\nHowever, the problem I described above happened when I was dealing with jsonl files \ud83d\ude3f\r\nAlthough I did not thoroughly inspect, I suspect the cause was the one extremely long document in my case.","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2181","title":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)","comments":"I see... I guess there is another problem going one then, related to the size.","body":"Hi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!","comment_length":15,"text":"Error when loading a HUGE json file (pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries)\nHi, thanks for the great library. I have used the brilliant library for a couple of small projects, and now using it for a fairly big project.\r\nWhen loading a huge json file of 500GB, pyarrow complains as follows:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1027, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/user\/.pyenv\/versions\/3.7.9\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/app\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/json\/9498524fd296a6cca99c66d6c5be507d1c0991f5a814e535b507f4a66096a641\/json.py\", line 83, in _generate_tables\r\n parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 247, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\nWhen using only a small portion of the sample file, say first 100 lines, it works perfectly well..\r\n\r\nI see that it is the error from pyarrow, but could you give me a hint or possible solutions?\r\n#369 describes the same error and #372 claims to have fixed the issue, but I have no clue why I am still getting this one. Thanks in advance!\nI see... I guess there is another problem going one then, related to the size.","embeddings":[0.016556887,0.0126864687,-0.0075106705,0.3409382999,0.1604535431,-0.2647139132,0.193632707,0.5859469771,-0.1621605307,-0.0815067366,0.1654730737,0.1441630423,-0.0150832608,0.0145624503,-0.1295208484,-0.1911917031,0.033368215,0.0782742128,-0.0815535039,0.1355070025,-0.0368864611,0.2328725159,-0.0935274884,0.2323912382,-0.0870280042,-0.0781508908,0.1980634928,0.3374814093,-0.3348842263,-0.4495861232,0.1612249315,-0.2951855063,0.0062585096,0.3340712488,-0.0001156308,0.0471057519,0.2495900393,0.109179385,0.0187429283,-0.2191216499,0.2808351219,-0.3987070322,0.228011921,-0.1674175858,0.1295851022,-0.4462138414,-0.3999225199,0.3737807572,0.4277296364,-0.125555113,0.1651013941,0.091210708,0.2657488286,0.2669147551,0.4368439317,-0.0652506948,0.027398346,0.5041721463,0.3052019775,-0.2382042259,-0.3354803026,-0.1370317638,-0.249592483,0.1381146312,0.3779326081,-0.1172010303,0.076280944,0.0589680932,-0.0799430832,0.0377216749,0.2263211757,-0.4086019993,-0.1033591554,-0.0781874731,-0.0856484845,-0.1830741614,0.3079182208,0.3882312179,-0.2103946954,0.0698304623,-0.1914447248,-0.1000016332,-0.1991208196,0.1786511391,-0.3580315709,-0.1259495318,-0.0024537381,0.3322300911,0.2310532629,-0.007610932,-0.035575673,0.0311062783,-0.2734938562,-0.056598112,-0.2094626725,0.0166391041,0.0015142764,-0.3171832263,0.1906490326,0.2260884643,0.1960860789,-0.0655944422,0.2463149428,0.2574926019,0.6111616492,0.059475515,-0.3522540629,-0.1677471697,0.2054002136,0.237815097,-0.0194872171,0.0609876625,-0.0592074208,-0.5538618565,0.0594820604,-0.2798270881,0.259775579,-0.2160245329,-0.2103198022,0.1333221942,-0.7950463891,0.0500115268,0.1662532538,0.2496998012,0.0272156876,0.2363475859,0.0485518463,0.1581959277,0.0089593437,-0.1194209307,-0.0078674201,0.1263681799,-0.152511999,0.0385965146,0.1579921395,-0.1335816085,-0.0064160214,0.1193348318,0.2173471451,-0.2116392553,0.1862066239,0.0166174266,0.0477700792,0.3288215995,-0.0516762882,0.1116407961,-0.0548598059,-0.0126550142,-0.2687699497,0.4798909724,-0.0626880005,-0.3448120952,-0.1742749214,0.100412406,-0.2606137395,0.2655539811,0.0049829497,0.0375169143,0.1504191756,-0.3691319823,0.2161603868,-0.131151095,0.1778304875,-0.3457722068,0.1031698063,-0.0813739225,-0.686923027,0.017364502,0.2560487986,0.1754170209,-0.094436802,0.5446502566,-0.415045321,0.1447616071,-0.045954112,0.1615658551,0.1449172199,0.160628736,-0.6060854793,0.3728240132,-0.2745345831,0.057643231,0.0307355803,-0.0146053843,-0.1264383495,-0.0241248291,-0.0301004108,0.3515685797,0.0832778588,0.2309474498,-0.3654409647,-0.2369496226,0.1233631745,0.3634355962,-0.1145457625,-0.4116987586,-0.0493318588,-0.1916890144,0.055319462,-0.1157415733,0.1517809778,0.4218284488,0.135918498,-0.1193930358,0.102686882,0.1130747423,-0.4654299319,0.021460088,0.0717728287,-0.0631102473,-0.3461884856,-0.0881970525,-0.030149851,0.2487495095,0.0945114121,0.2060095221,0.0340390354,-0.1731155217,0.0721888989,-0.0024395175,-0.0986388773,-0.413590461,0.1334338039,0.0734615549,-0.0101108309,0.390627265,0.0637331009,-0.293379277,0.0692314729,0.1497817785,-0.0336900316,-0.1447110027,-0.0944388956,0.184556067,-0.0829884708,0.0026114385,-0.2447776794,0.1293199956,0.2801520824,-0.3557440937,0.0277640875,0.1032500342,0.0765697882,-0.0654545948,0.2241742462,0.4178517759,-0.328894347,0.203213647,0.0963192433,-0.2922596335,0.0688130632,-0.0792855769,-0.0184273757,0.1651138216,0.2211024761,0.3922338784,0.2243866175,0.1382108033,-0.1602222621,-0.119418323,0.398599565,-0.1834302247,-0.0815003961,0.152345404,-0.1335780174,-0.0331332907,0.0964195654,-0.1294155121,0.1355017573,0.2478066981,-0.1427013129,-0.0006218139,0.0272530131,-0.1798278987,0.2222775072,0.197826013,0.1286401451,0.2018653899,0.525388062,-0.1262486577,-0.2340527326,-0.2239226252,0.24420017,0.1961286515,-0.2908231318,0.1298102885,-0.0687725171,0.0243006777,-0.0847825631,-0.679587245,-0.3061884344,-0.0921970904,0.0006510711,0.4704107642,-0.330901593,0.0706555918,0.1599271297,0.3159408569,0.0793664604,-0.1885332614,-0.044373747,-0.3438287079,-0.2535299659,-0.0260646604,0.4630136192,0.1464138925,0.0704189911,0.3701916039,0.0221871901,-0.0610447712,-0.4288595021,0.218477577,-0.0467988625,0.0778331682,0.0034065631,0.1291554272,-0.247211799,-0.2504459023,0.3068675399,0.0360249728,-0.205761373,0.327116698,-0.1920830458,0.0744263157,-0.1781024784,0.1151923761,0.2128263414,-0.5418013334,0.3973288238,-0.1855192035,0.0039909063,-0.2795964479,0.2635209858,0.0960103422,0.0679152757,-0.0404972062,0.0594006777,-0.4201915264,0.1961016506,-0.0403325036,-0.1881746352,0.0418398418,0.1097168624,0.1519098282,0.1650311649,-0.6058769226,-0.0128622605,-0.0775761679,0.5512671471,-0.34305498,-0.2375435829,0.1149812415,0.1875803918,-0.0166976135,0.1244238466,-0.1434655637,0.0150683811,-0.007390561,0.2049880326,-0.1010244936,0.5279217958,0.1304708272,0.7018238306,-0.1654545516,-0.1430081576,0.3313239813,0.2419932485,0.0773653984,0.0481638126,-0.0284133386,-0.1179990172,-0.1809873581,-0.082166642,-0.0068381014,0.0487272777,-0.1101650149,-0.0276327617,-0.1180320904,-0.2268259227,-0.1588106602,0.3197793961,0.0610067919,-0.1138895303,-0.2007410973,-0.2847957313,-0.0170084145,-0.0803113654,0.058490932,0.0380901061,0.109308593,0.1145236641,-0.0157520957,-0.3002438545,-0.03336972,0.0799987689,0.3141287267,-0.1444272548,0.05180379,-0.1972083151,0.1843090355,-0.4777928293,0.7541434765,0.46496436,-0.0762622878,0.1359017193,-0.0663604811,-0.3780810833,-0.0619557053,0.0836297646,0.19823578,0.5710027814,0.1032129526,-0.4526485801,-0.1059838161,-0.1052925736,0.3224133551,0.0146235097,-0.4784038067,-0.4063741565,-0.4959048629,-0.4578323662,-0.0559185967,0.0945278183,0.1901103556,0.1029651314,-0.0987389386,0.2796903253,-0.2547020614,-0.4795063436,0.0978191793,0.1785515547,-0.5503336787,0.1491027772,0.0512285754,0.3193382919,0.5359148383,0.6146833301,0.5909862518,-0.2183253914,-0.0538363568,0.3960570991,0.179886803,0.167946279,0.1672121137,0.2085579187,-0.1727021188,0.2022818327,0.2214758396,-0.0104978262,0.1908637285,-0.1210765243,-0.0008764418,0.2911641896,0.502509892,-0.2172303349,0.1578591317,0.4766687453,0.283444345,-0.0467166416,0.222686559,-0.0596349575,0.9461208582,-0.0210727379,0.1325546056,0.5243240595,-0.4252023101,0.5806598067,-0.3188984394,0.0976480842,-0.4774041772,0.3479599357,0.0661804825,-0.2101878226,-0.0861986279,0.0646333247,-0.0427132286,-0.1986050904,-0.0862123966,0.0907088295,-0.1165641248,0.072005637,-0.0206812564,0.0002179008,-0.324554503,-0.0062565482,-0.3452311456,-0.1282550842,-0.1717916131,-0.3250171542,0.0421548039,-0.1757344902,-0.1917931437,0.27093485,0.1396908313,0.3771396279,0.2748383582,-0.142274633,-0.0074551897,-0.4400272965,-0.2658154666,-0.1649732739,-0.038579531,0.2240040004,0.0084361359,-0.125425905,-0.0218853652,0.1243404746,0.1479468495,-0.1720808744,-0.1758476347,0.1463666111,-0.1700342,-0.1859632432,-0.1407295018,0.1831440181,-0.1518963724,-0.031130895,-0.0627598912,-0.1478973478,0.0817598999,-0.1284501851,0.0683539659,0.0026659416,0.0737182721,-0.3161939085,0.0839205012,-0.3015911281,-0.0840510428,0.2698033452,-0.0315242931,0.1463471204,0.4538902342,0.4027526081,-0.0726219341,-0.0956394076,-0.0567291044,0.5515409708,-0.6051450968,0.2913151979,-0.1565793902,0.1126884371,0.0965897813,0.3177764714,0.2544378042,0.0715060085,0.0265478864,-0.3478247821,-0.228525877,0.49165681,-0.0185184702,0.0122019807,0.1027065217,0.4117230773,-0.0780539811,0.1973409653,-0.2198727131,0.3020847738,0.0287230853,0.0564546213,-0.4602073133,-0.2132687718,-0.2015920132,0.2550780177,0.0270884354,0.1679867655,-0.0712248683,-0.1259257495,-0.1276164204,0.1012032777,0.102016978,-0.0968398228,-0.0645024255,-0.1031718776,0.1211112067,0.1297840625,0.1317270547,-0.2099898309,-0.0535009541,-0.3004156053,0.0015181198,-0.1877655536,-0.3642434776,0.0790169537,-0.0337792262,0.2890965343,0.0804840401,0.2754154503,-0.0237067323,-0.0644059181,-0.22985816,0.1052854285,0.152957812,-0.0467558056,0.401498884,-0.3238122761,-0.0387646407,-0.2653609514,0.2756453156,0.4593812525,-0.2252635211,-0.1238568723,0.0152619174,0.1677667648,0.0016635651,-0.215190962,-0.219523713,-0.0497761928,-0.1538697779,-0.0746731684,-0.0904239416,-0.2073939592,-0.1787260175,0.1219562218,0.2947433591,0.241909951,-0.346365124,0.3792241216,-0.2789323628,-0.1749604493,0.2887688577,0.135514155,0.5593860149,-0.0907415226,-0.0720371902,0.1808650494,-0.0511423796,0.0833588466,-0.1531466097,-0.4281859994,-0.1786465049,0.3637129962,0.2060844451,-0.2417384535,0.3044683337,0.5248377919,0.4010774493,-0.1126888245,-0.0198079515,0.2256049365,-0.130985409,0.2223398685,-0.5097849965,-0.1980471462,-0.1236797422,0.3870281577,-0.1253376007,0.0557185821,0.4912979305,0.1782965064,-0.0879884064,-0.6212043166,-0.1130404621,0.0727156624,-0.0714724213,-0.2394380271,0.2298415005,0.6816926003,-0.4974334538,-0.0561945774,-0.215879634,0.4886175394,0.2710981369,-0.2111812979,-0.4033637941,-0.2183916718,-0.0720908716,-0.0950081572,0.1022535264,0.5172538757,0.349851042,0.2333961427,0.0662924498,-0.1418414563,0.2303020805,0.1463665068,0.138451755,0.0005854817,-0.2333464324,-0.2354170978,-0.3443225026,-0.1802339256,-0.2282650024,-0.3921266198,-0.331035316,0.1211898774,-0.0312109645,-0.1318797469,-0.1365348548,0.0630984902,-0.0726336241,0.6978250146,0.3485355973,0.2058713883,-0.2465088516,-0.2613237798,-0.2471017987,0.4796885848,-0.4091098011,0.1983162165,0.2130555511,0.183383733,0.1070473716,0.3368973434,-0.065203689,0.5276255608,-0.1808633357,-0.2147909552,-0.3851124942,0.0368262939,0.2935440242,-0.1667665094,-0.1844618171,-0.151456967,0.3386736214,-0.1768078953,0.1125646606,-0.1182276607,0.0232722685,0.053921964,0.0855706111,0.3299843371,0.0818731636,0.3291729093,0.0286539588,-0.1429241747,-0.10798347,-0.2546292245,-0.1726094484,0.3846240938,0.0581517741,0.4116558135,-0.0787665993,0.0795214325,-0.1733229309,0.1052340493,-0.4348300099,-0.0979113281,0.1218618304,-0.2352110445,0.024975352,0.1028489694,-0.0171118379,0.1146759838,0.0885695368,0.1137052253,-0.494494319,-0.1414470226,0.4697202742,-0.3657928407,-0.2979660034,0.4970999062,0.1034523398,0.3452362418,-0.1185090169,-0.5068212748,0.0891442373,0.1896468252,0.0586640462,-0.3722429276,0.2245831043,0.0482143424,-0.0609631985,0.0519516729,0.2235715687,0.1836720407,-0.3651713729,-0.0317124315,-0.1290486157]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2176","title":"Converting a Value to a ClassLabel","comments":"Hi @nelson-liu!\r\nHere is what I do to convert a string to class label:\r\n\r\n```python\r\nfrom datasets import load_dataset, features\r\n\r\n\r\ndset = load_dataset(...)\r\ncol_name = \"the string column name\"\r\n\r\nclass_names = dset.unique(col_name)\r\nclass_feature = features.ClassLabel(names=sorted(class_names))\r\ndset = dset.map(lambda str_value: {col_name: class_feature.str2int(str_value)}, input_columns=col_name)\r\n\r\ndset = dset.cast(features.Features({\r\n ...\r\n col_name: class_feature\r\n})\r\n```\r\n","body":"Hi!\r\n\r\nIn the docs for `cast`, it's noted that `For non-trivial conversion, e.g. string <-> ClassLabel you should use map() to update the Dataset.`\r\n\r\nWould it be possible to have an example that demonstrates such a string <-> ClassLabel conversion using `map`? Thanks!","comment_length":50,"text":"Converting a Value to a ClassLabel\nHi!\r\n\r\nIn the docs for `cast`, it's noted that `For non-trivial conversion, e.g. string <-> ClassLabel you should use map() to update the Dataset.`\r\n\r\nWould it be possible to have an example that demonstrates such a string <-> ClassLabel conversion using `map`? Thanks!\nHi @nelson-liu!\r\nHere is what I do to convert a string to class label:\r\n\r\n```python\r\nfrom datasets import load_dataset, features\r\n\r\n\r\ndset = load_dataset(...)\r\ncol_name = \"the string column name\"\r\n\r\nclass_names = dset.unique(col_name)\r\nclass_feature = features.ClassLabel(names=sorted(class_names))\r\ndset = dset.map(lambda str_value: {col_name: class_feature.str2int(str_value)}, input_columns=col_name)\r\n\r\ndset = dset.cast(features.Features({\r\n ...\r\n col_name: class_feature\r\n})\r\n```\r\n","embeddings":[-0.0255655907,-0.1936047971,0.0500335023,0.0593827851,0.6273636222,0.2115247846,0.2829759717,0.0898772702,0.0759942457,-0.0568674728,0.1029048711,0.6814887524,-0.0339189284,0.115833059,-0.063829571,-0.2042801529,0.1750881374,0.1827311069,-0.1002876386,-0.056887541,-0.3199301064,-0.0221358184,-0.4753821194,0.1178039089,-0.0183532331,-0.0131625459,-0.0521399416,-0.1745000482,-0.0095149893,-0.2616071403,-0.0025995478,-0.2266906947,-0.0235290825,0.4040020704,-0.0001230374,-0.325283885,0.1713490933,0.1276385337,-0.1027286276,-0.0012642887,-0.5446994901,0.1509112716,-0.0493944585,-0.3124809265,-0.0679486319,0.4585186839,0.0102069946,-0.3518183827,-0.0797238797,-0.0385747217,0.0643964559,-0.2685542405,0.0433487855,0.1502618194,0.1173128039,-0.0639188141,0.1979943216,0.1961912811,0.5107662082,0.3812748194,-0.0233184416,0.2310218215,-0.2420650274,-0.4479733109,0.2709723711,0.1409620345,0.0147539843,-0.3441451192,0.2141891122,0.0697048306,0.5089139938,-0.2787736952,-0.2406190187,-0.0289247558,0.2034195364,-0.1926503479,-0.1329814643,0.0453148559,0.1556026638,0.0525094084,-0.5872043371,-0.0947453901,0.0096688913,0.0064496761,-0.04198296,0.3577371538,-0.0681064501,0.2042177618,-0.183602497,-0.3301814795,-0.0656691194,-0.1423177719,0.053425204,0.451168865,0.0285003148,-0.1622573733,-0.1803563833,0.308008492,-0.3923470676,-0.4623276889,-0.0825618654,0.2044612765,-0.1535297781,0.2360501885,0.1304843426,-0.0718885139,0.6325418353,0.2619615495,-0.0377869047,-0.3972043097,-0.3506874442,0.0470989533,-0.1964601278,-0.0970517769,0.3145183623,0.2756204009,-0.0859175399,-0.208613649,0.060328491,-0.0629566833,-0.6144182086,0.1692011654,0.0577347204,0.102768749,0.159675926,0.15300861,0.3831623495,0.0127394684,0.3372881114,-0.5151320696,0.0256724358,0.1687627882,-0.0594784804,-0.2546179891,-0.2046683133,0.1303680539,-0.1657950729,0.1104661152,-0.1394721568,-0.2411892265,-0.1913698465,0.3533140719,0.1423953176,0.0091943825,-0.3899411559,0.0919626057,0.362169534,-0.3446409106,-0.3618513048,0.4646984339,-0.3172389567,0.105220139,0.2430622429,0.0626072958,0.3431931734,-0.2344117314,-0.0157063436,0.1214720458,0.2205845118,-0.323260963,0.2305161804,-0.3888455629,-0.2717774212,-0.1573465765,0.1096925884,0.1466993243,-0.5220822692,-0.172278434,0.220476523,0.1450605094,-0.2246677876,-0.303283751,-0.0844213963,0.4340304732,0.1755695194,0.0679722875,0.743145287,-0.4040966332,-0.1180322319,0.2816625834,-0.044224117,-0.5518223047,-0.1832592934,0.4364493787,0.1886939406,-0.1759246737,0.1285496354,0.1844903529,-0.2581253946,0.1220353022,0.0849035829,0.0374753475,0.1188872978,-0.1132329702,-0.204303816,0.4207707644,0.1227368861,0.4091431797,-0.0281125456,-0.2878473997,0.3498987257,0.0733438507,-0.0331092849,0.0556454845,-0.1548636705,-0.233484447,0.1583824158,0.0418301709,0.3316720724,-0.0388476215,0.2497375906,-0.2951244116,-0.0320845544,-0.0257879458,0.237743035,0.2743281126,0.0450926907,-0.2332085818,0.057657294,-0.3211953342,0.0445083641,-0.2223040015,-0.1409013718,0.2288301438,-0.0443542078,-0.2813130021,0.1206276938,-0.1339090914,-0.5128315687,-0.1144197211,0.2241096944,0.0126015656,-0.1962092072,0.3364670575,0.0232319161,-0.1875057369,-0.1845719963,0.2407033145,0.192868799,-0.2362461537,0.2976700068,0.1409193724,0.1962014437,-0.2853609324,-0.3096443117,0.2476768792,-0.0560093597,0.2904866934,-0.0988103375,0.2705557048,-0.028501587,0.0112132011,-0.0903075114,0.1164753959,-0.4181664586,-0.2635629475,0.1799531579,0.0261154603,-0.2478145808,0.2360076606,0.1281989664,-0.5026358962,0.2705974281,0.1162729636,-0.2514476478,0.3145844638,-0.0029368745,0.1560785621,0.1207884029,-0.0455898792,0.1869260818,0.11836081,-0.0162551049,0.1488968283,0.3097302616,0.0846782476,0.1536530405,-0.380091697,0.0024057084,0.1814293265,0.2393162549,-0.0628101677,0.1377518922,0.0139249545,-0.3364920318,0.0362231247,0.0175071452,0.225125134,-0.1867703199,-0.1059157103,0.1685806066,-0.3640716374,-0.1875962168,-0.0140826553,0.06448742,0.1266998947,-0.2037019283,-0.0895199254,0.1548196673,-0.4496595562,0.098924078,-0.2649113536,-0.142023325,-0.0725619942,-0.0077444417,-0.0715042129,0.0643922165,-0.142561093,0.185032472,-0.104815796,-0.5802630186,-0.0710165054,-0.2839055955,-0.1934550405,0.3171650171,-0.2148442715,-0.4116519094,0.1804641038,0.1438295096,-0.3101681173,-0.0612025261,0.1244389638,0.0559946187,-0.154782474,-0.1652976722,-0.0611049905,-0.0006229725,-0.0131550021,0.1271002442,-0.1438927799,-0.1116666272,-0.2295333743,0.1373449564,0.4809709489,0.2863416672,0.1686650068,0.1029805243,-0.1149902642,0.0788241178,0.0629107952,0.0188949015,0.0958622918,-0.0502325632,-0.1277494729,-0.040573366,0.009583707,-0.0468978845,-0.0986372828,0.2942809463,0.0259362422,0.1530192196,0.2664952874,0.3972101808,0.0180965643,-0.0478277467,-0.3397721052,-0.0846751928,0.0018226623,-0.0003129006,0.2170868665,0.4192286134,0.0019848177,0.1187851056,-0.1266331971,-0.3776563704,-0.2086678445,-0.210274443,-0.1063516438,-0.0071304459,-0.0226576813,-0.1015788913,-0.1895440072,-0.3195202649,-0.088262789,0.0029031951,-0.1653680205,0.1434974223,0.1068401635,-0.4341834188,-0.1911122948,0.1427051872,-0.2788325846,-0.0014220425,-0.2721328437,-0.1227785349,-0.1991496831,-0.1161536947,0.0036394214,-0.3099345565,-0.0294060819,-0.1133377776,-0.2600283325,-0.1979280412,0.105628483,0.3141315877,0.2474207431,-0.0124007883,-0.1005084366,-0.0259748586,0.1225774661,0.0789211914,0.5989871025,-0.3749782145,0.5384542942,0.3970617056,0.1023102105,0.0220939778,0.0963059962,-0.5713668466,0.2216720432,0.5073204041,0.2008759081,-0.0468512848,-0.0228673052,0.2609052062,0.1410344541,-0.0725183785,0.2324441075,0.1165143996,0.0093542626,-0.1606273353,0.1001963541,0.3373531997,0.1128802672,0.2035348117,0.1686590165,-0.2296980321,0.253090471,0.2245436013,-0.0782187209,-0.0245438796,0.2291560024,0.0548910499,-0.2350263447,-0.0252665523,-0.5373609066,0.4027567804,-0.1361222714,-0.3011232018,0.0273725223,0.2237888724,0.6616911888,0.0804578215,0.2298316807,-0.2139789909,0.0041389698,0.047728125,-0.2515520751,-0.0259645097,0.1131285727,0.227032572,-0.4178067446,-0.266572684,0.7585983872,-0.0291903988,-0.3857869506,0.1773231477,0.3100294769,-0.555798769,0.3557275236,0.5645313859,1.0125486851,-0.3431296647,0.1454716921,-0.2484121025,-0.1621375829,0.5679709911,0.2519029379,0.0880277604,-0.5652980208,0.0916982666,-0.1793633401,-0.0906066895,-0.0089547252,0.5965073705,-0.609790504,0.2107544094,-0.2679894567,0.1973945349,0.2196619362,-0.2279046327,0.1847220063,-0.2431535274,0.0256851595,0.1945687383,0.0789450034,-0.2898958027,0.0106718494,0.1411578208,-0.2379117161,-0.0533445515,-0.2061766386,0.0650886297,-0.0482121445,-0.0813584253,0.2329940498,0.0485210568,0.3747678697,0.6419818401,0.1609936059,0.1660036892,-0.0272732023,0.0487698279,0.0272181164,0.2853085697,0.2338415086,0.1786558032,0.2727909386,0.4090893865,-0.0385717154,0.3694918752,-0.2036484182,-0.0529256389,-0.1899779886,0.0168209635,0.0808304623,-0.5354503989,-0.288388133,0.1659200042,-0.0394937024,-0.0337104313,0.0246262401,-0.1796377003,-0.1675097346,0.1386393011,0.0447593853,-0.3176604211,0.3458250761,0.3615108728,-0.1783432364,0.1018877774,0.1702836305,-0.044767525,0.0922074914,-0.0773491859,0.4468341768,0.0818070993,-0.0976535976,0.0416785069,-0.099543497,-0.2006267905,0.1619394124,0.158271715,0.078700006,0.3553593457,-0.265642941,-0.1228698269,-0.2867258191,-0.0527212694,0.4271603227,0.0177223887,-0.2163178921,0.024953898,0.0018833075,-0.0971542746,-0.1648704708,0.1144624799,0.0444035381,0.3307688236,0.2599433661,0.5231438279,-0.2434344888,-0.0348922946,0.0268337745,0.0751657635,0.3260428011,-0.0992771089,-0.018600652,0.1596021056,0.2455032915,0.1522901207,-0.1651834249,0.1069822311,-0.0486880094,-0.3041029871,-0.1637863964,0.0876576751,-0.0636787266,0.2773742378,0.0776106417,0.2589271665,-0.0397043973,-0.0828627795,0.058986567,0.2501312196,-0.16561158,0.2256467342,0.3910964727,0.0243667569,0.0973727033,0.2583626211,0.1075731367,-0.4333234727,0.3887710571,-0.2362117171,0.4785032272,0.1618556529,0.1450473368,-0.1302812696,-0.1566516459,0.1163124815,0.1600237787,0.0387980342,0.0055738054,-0.0567881763,0.4063164592,-0.077764608,-0.1374928802,0.0881364569,-0.4882024229,0.1837023646,-0.2927996218,-0.1953432113,0.6786150336,-0.4554275274,-0.0774901509,0.5122261047,-0.0090326257,0.0824035034,0.3932998776,0.2641244531,-0.0176844168,0.1688565463,0.1542432755,0.1913851649,-0.0663403645,0.2061370015,0.1363457888,0.1509128511,0.4715078473,0.4203931391,-0.0547293164,0.2940470278,0.0100152781,-0.1320383996,-0.6468144655,-0.386957109,0.0624807961,0.2653108239,-0.1460632235,-0.2055393755,-0.1703096926,0.0567072034,0.0826713592,0.0244151689,-0.206188783,-0.1274563819,0.7491557598,0.0376739278,0.4000746012,-0.1511315405,-0.4842125475,-0.2126098573,0.254324466,-0.2470190525,-0.0663001239,-0.189968735,0.1491680592,0.3778536916,0.1601914465,0.2957721353,0.3590630293,-0.3054682314,0.2664515078,-0.0995986462,-0.1622082293,0.2465300113,0.0287704952,-0.0433145277,-0.1728661656,-0.133203119,0.0702796802,0.0104690278,0.1741269529,0.0883012414,0.3612296581,0.1379585415,0.3692828417,0.1229498014,-0.3744261861,0.0683621168,0.0596387386,-0.1512699276,0.0250933282,0.347661376,-0.3015752733,0.1799566597,0.2458891273,0.0189051758,0.5963383913,0.1128668487,0.2405885607,-0.1761267632,-0.1012260839,0.1970670521,-0.3364038169,-0.2154930532,0.1898294538,-0.3442457616,-0.0179292057,-0.1552858949,0.190863505,0.0833527297,0.1257834882,-0.4912692904,-0.1025471166,0.1621189564,0.362932235,-0.3980661333,0.2992558479,0.181675002,-0.0522304997,-0.2583378553,0.1431859732,-0.0370673202,0.0055397716,-0.173643738,0.0744590759,0.0152735347,-0.2734265327,-0.0163699631,-0.0003851333,0.3927971125,-0.0876839161,0.109340474,0.4171192348,-0.1196040735,-0.2028664201,0.3020249605,-0.5358947515,0.0188857075,-0.0343907885,0.0679788515,-0.0220972486,-0.4184687138,-0.0114096804,-0.1128929779,-0.6668360829,0.149390474,-0.4806389511,0.0590933152,0.117983222,0.1565638632,-0.0001420217,0.4983156621,0.0398741774,-0.2799821794,0.2041750699,-0.6204312444,-0.4009134173,0.1542604119,0.0369955115,-0.1645594984,0.0300338045,-0.3541930914,-0.0812166408,0.3870242536,-0.0242356583,-0.3789287806,0.1472871602,-0.1123757362,-0.2048938721,-0.2524279654,0.1457889676,-0.118119821,0.1565184891,-0.1777057201,-0.2537243366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2175","title":"dataset.search_batch() function outputs all -1 indices sometime.","comments":"Actually, I found the answer [here](https:\/\/github.com\/facebookresearch\/faiss\/wiki\/FAQ#what-does-it-mean-when-a-search-returns--1-ids). \r\n\r\nSo we have to do some modifications to the code for instances where the index doesn't retrieve any IDs.","body":"I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n ","comment_length":25,"text":"dataset.search_batch() function outputs all -1 indices sometime.\nI am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n \nActually, I found the answer [here](https:\/\/github.com\/facebookresearch\/faiss\/wiki\/FAQ#what-does-it-mean-when-a-search-returns--1-ids). \r\n\r\nSo we have to do some modifications to the code for instances where the index doesn't retrieve any IDs.","embeddings":[0.0112970611,-0.3721057475,-0.1019985154,0.0192688778,0.2021879554,-0.0966877714,0.2803093195,0.2968439758,0.1391415298,0.4209129214,-0.2324287444,-0.2434533834,0.1058326885,-0.033898145,-0.1621748805,0.1081455201,0.2741164267,0.3812984526,-0.0525572002,-0.5196725726,-0.3187375963,0.1191092208,-0.0284683872,0.3729512095,-0.2878592014,0.2902801335,-0.0275247227,-0.0893982723,-0.2744064629,-0.4506147206,0.5258822441,-0.3038473129,0.4114505351,0.2260930985,-0.0001312564,0.1222993061,0.3880340755,-0.0330385678,-0.0978071392,-0.1585950553,-0.0568627752,0.1535648406,0.0832173079,-0.0496736802,0.0529693365,-0.293507278,0.2583491504,-0.0760006383,-0.0392876305,0.2710106373,0.0165874027,-0.1001899764,-0.2166370302,-0.0440472066,0.6340439916,-0.316478014,0.0808619559,-0.143078208,0.4359707832,-0.0289208461,0.1490160525,0.1251526773,0.0537862107,-0.0263432525,-0.2640728354,0.1519919336,0.6509360671,-0.347019285,0.1948388517,0.2429174185,0.2132443488,-0.0165178366,-0.3425406516,0.1010914296,0.1615769565,0.0643818378,-0.0493507236,0.1699866951,-0.1023728475,0.2492177188,-0.0735962987,0.3226192296,-0.2642556131,0.1283892244,0.0921010301,0.3048032522,0.0520673916,0.1365374923,0.1266665608,0.1793742031,0.2614558935,0.1650856435,0.1424205452,0.2667565942,-0.545294106,-0.0396536961,0.3013203442,-0.2635917068,0.1599813849,-0.3341228962,-0.295882225,0.0122053837,-0.1906159222,-0.0811451674,0.1348101348,0.0300397296,-0.0409570709,0.0879250839,0.2758273184,-0.4434856772,0.0225716718,0.05646386,0.1405127645,0.0172800384,-0.0684079379,0.1226767525,0.0707820579,-0.2924739122,-0.3751615286,-0.092208527,-0.2851830423,0.2287971377,-0.2117733508,0.1873225123,0.3979750276,0.0808830336,-0.0434568711,0.0965683982,-0.1920486689,-0.0721826628,-0.1490160227,-0.1684361994,-0.1290246695,0.0257082675,0.1512941122,-0.4258549809,-0.1803211719,0.0613467805,-0.1417217553,0.1943153292,0.1223837063,-0.1732915044,0.4533530176,0.5615368485,-0.1166198105,0.3323795497,0.1357719451,0.0466702506,-0.0244340524,0.0082993126,0.0602010712,-0.5894426703,0.1787384748,0.0659676343,0.0158018582,0.3439203501,0.3202601373,0.3334341943,-0.2957113981,0.3365435004,-0.0159025975,-0.4066883028,0.0313610584,-0.0263190251,0.2536776662,0.1740864664,-0.0535256527,0.1447275877,0.0188022442,-0.2671530545,0.10422948,0.3047020435,0.0776820034,0.2836076021,-0.4879200757,0.1991422325,0.1316751391,-0.2821635902,-0.3218086958,0.056944754,-0.2449695468,-0.6043919325,0.0168585777,0.2755216658,0.2623558939,0.202829197,0.2860118151,0.1991926134,-0.0223771054,-0.1068089604,-0.4190152586,-0.0749165043,0.262950778,-0.0806498677,0.2959195077,0.2607781291,0.0897124112,-0.7707069516,0.3443847895,0.078310065,0.097485669,-0.0759550408,0.3600547612,0.0039185164,0.717531085,-0.0647906512,-0.0875052363,0.1142708212,-0.3557227254,0.0644310415,0.0200779531,-0.2087035775,-0.0145462397,-0.0329615809,0.060866762,0.2663134634,-0.0613433495,-0.1442528665,0.2852407098,-0.1506960839,-0.2166908532,0.0279039238,-0.1892898977,-0.1727768183,-0.4619542658,0.2869715989,-0.0044568004,-0.2153081298,0.0080345124,-0.0235068556,0.3108555675,-0.0657237023,0.0083285766,0.1663467586,-0.2015215904,-0.2778989971,0.7204379439,0.1057652831,-0.2275300771,-0.4191123843,0.1112755015,0.474425137,0.1263320744,-0.1979949772,0.1549974382,0.2241897285,-0.2468896806,0.4735146165,-0.122557722,-0.001431486,0.0749269798,0.142181322,-0.1148776785,-0.2041202635,0.1019529179,-0.1453208774,-0.1148640439,0.0365648121,-0.251742065,0.180736661,-0.1566976607,-0.4020548463,0.0373723991,0.2453040183,0.0468486212,-0.0122243585,0.123898834,-0.2407312244,0.3852651417,0.1465954036,-0.1858467013,-0.2387987822,-0.1886484921,-0.1494654566,0.0910928547,0.1543564796,-0.24721466,0.1434974521,0.3117801845,-0.0742414072,-0.3294375837,-0.2253194153,-0.3847643137,0.2618565857,-0.2814090848,-0.0455774181,0.0504941493,-0.027146671,-0.3396145701,-0.0936510414,-0.0362692289,-0.076493375,0.2038690597,-0.3289221823,-0.0500363447,0.0073315878,-0.1228785291,0.2336872816,0.0564926751,0.3149169087,-0.6124581099,0.0611079,-0.3827643991,-0.2278803438,-0.5138926506,0.306232959,-0.0604019165,-0.0381238796,-0.5279856324,-0.5108689666,0.1429802775,0.0286729913,-0.1098232344,-0.0350111946,0.3234111667,0.0188210737,-0.2142340839,0.3020818233,-0.0022866556,-0.0831172541,-0.1791341454,0.1730704904,-0.2697455883,0.3766937852,-0.2327563167,-0.2679131627,-0.0313489214,-0.0388202593,0.1331948042,-0.2206423879,0.12472675,0.0276666228,0.1237579361,-0.1222937331,-0.0160318147,-0.0337288529,-0.2221769094,-0.2021137774,0.2577733994,0.3072982728,-0.3648935258,-0.2410871536,-0.1581915617,-0.5619202256,0.4506982565,-0.135999456,-0.1375068575,0.0701930895,0.0001695525,0.0629063025,0.441385746,0.1164398417,-0.069123663,-0.0520515628,-0.2459105402,-0.0521875471,0.3002716005,-0.0251200367,0.2634778321,0.0242301766,-0.0936778486,-0.2716232836,0.7259996533,0.1539808065,-0.071441479,-0.0122895008,0.1312197149,0.1544188857,0.1646025777,0.0283248667,0.2517497838,0.2088071257,-0.3656691015,-0.0824299678,-0.0206991583,0.1367794573,0.279705435,0.1323606372,-0.1434504688,-0.1201965883,-0.056984406,0.2925826013,0.218982473,0.0246624034,0.1509403288,-0.0940014496,0.1088427231,0.1891393214,0.0688245595,0.0959357694,0.2566945255,0.3786235452,-0.4972277582,-0.4568194747,0.224278003,-0.0808367282,0.3757855594,0.2393745035,0.1027819514,0.4788725972,0.2590561807,1.25680089,-0.1759703457,0.2236908525,0.223304525,0.114690423,-0.209475562,-0.4277406633,-0.3497859836,0.1456702352,0.2816675305,0.3042633832,-0.134570986,-0.282990247,0.3241880834,-0.0831251293,-0.0247051474,-0.4840517342,-0.4144079089,-0.3878677487,0.3596350849,0.1768422425,0.0704084188,-0.0257625058,-0.1287026703,-0.3183708787,-0.2886581123,0.0842141435,-0.0209282916,0.0136784101,0.1088996753,-0.0100783827,0.0540215075,0.5879078507,-0.0352985598,0.028829148,0.0063315849,-0.2897151411,0.1097691134,0.1316749901,-0.2149085999,0.13006109,0.40254125,-0.0378961377,-0.0379729606,-0.0699877664,0.1739650071,0.0251751803,-0.1724173129,0.3342431188,0.0776080042,-0.2163945735,-0.2567570508,0.6706346273,-0.1897262037,-0.2443475574,-0.0455895253,0.3609122038,-0.1271772832,0.8627623916,0.2036354542,1.0294399261,-0.1593864262,0.0449495576,-0.0330660231,-0.0576934963,0.3887160718,-0.1368834227,0.2788418233,-0.3323223591,-0.0388177596,-0.1417314708,-0.1017492414,-0.3878503442,0.4730277061,0.0222503878,0.3106590211,-0.0436043702,-0.2067585886,0.1425085515,0.0420786217,0.4215950072,-0.0997207686,-0.2761313319,-0.0268138554,0.2104556113,0.3512393832,-0.1016684547,-0.0206809547,0.0860846937,-0.0166559014,-0.2263438106,-0.2110869139,0.0651955605,0.2944015563,0.6096583605,-0.2668304741,-0.1942780912,-0.158507809,0.5273296833,-0.3109761775,-0.0649512708,-0.0632395297,0.426078558,0.0647376031,0.065994218,-0.2863041162,0.3530517817,0.1669599414,-0.4066240191,-0.0286879875,0.0466228686,0.1825835705,-0.2095660269,0.0837354362,-0.3531548679,-0.1128302887,0.1354658157,0.0735301003,0.1463965327,-0.1096654162,-0.0329746306,0.344363749,-0.2488816679,0.522822082,0.0379229747,-0.5396445394,0.0018793594,0.0687790439,0.3131364882,0.1026276425,0.0284283515,-0.0910012871,-0.106788449,0.0047644316,-0.3982792795,0.2628671229,-0.0657194853,0.3646074831,0.0066468124,-0.0524577722,0.0149269346,-0.0651229545,-0.0535121672,-0.0393876843,-0.6319127083,-0.1118278801,-0.0674781799,0.002082479,0.0426682048,0.2827430069,0.0436288416,-0.2224809229,0.04941006,-0.4360682666,-0.1517593861,-0.307020843,-0.3297475278,0.3338544965,-0.4554995298,0.0233901292,-0.1727028191,-0.0945394933,-0.0404481627,-0.0134357754,-0.1461308002,-0.0257276483,0.0994813815,0.2227835953,0.1763508022,0.259028852,-0.2526920438,-0.0301912706,-0.1449386925,0.0389373563,0.0625407919,0.6113522053,0.0664352626,0.2071974874,0.1814807802,-0.1188334525,0.0674207285,0.1603710651,0.220181495,0.2092126608,0.0098267654,0.3986610174,-0.24791646,-0.0045660739,-0.0807426125,0.1079753712,-0.0862931237,0.1245352775,0.0638480037,-0.1448547393,-0.1157495379,0.0802883431,0.0730740055,0.2872047722,-0.3682045043,-0.2889963984,0.0876005515,0.0878476351,-0.073615022,0.0310945921,0.1891530901,-0.0096661272,-0.0857616812,0.1576618254,-0.085710898,-0.2770795226,-0.082984753,0.1212407425,0.3878120184,0.0845260769,-0.4620716572,-0.4683160186,0.2818698287,0.1255486161,0.2182914764,-0.0888202339,0.1867126673,0.5988236666,0.1066996157,0.1163441464,0.0078103645,0.1931095421,-0.3157951236,-0.1006947234,0.357945621,0.4545297623,-0.3513323069,0.0446386226,0.2346816957,0.1333734095,-0.113142252,0.0405173823,0.1585411876,-0.2260121852,0.0537519716,0.0783934668,-0.1673153788,0.2790884078,0.0128534809,0.1342286468,0.0158728231,0.0545016751,0.0646355078,0.2016745806,-0.0954063162,-0.0659713596,-0.123331368,0.4243821204,0.2690967023,0.0256963335,-0.1349169165,-0.1839722097,0.0624125004,0.3232342005,0.3626470864,0.2544467449,0.405059725,-0.2183091938,0.1943673939,-0.0896439627,-0.171577394,-0.1543341875,-0.1575698107,-0.2221920192,-0.0709798038,-0.0498288646,0.0249194112,0.0257747006,-0.2098159939,-0.0548168235,0.3915155232,-0.140857622,-0.0055022761,0.0987098888,-0.0644803494,-0.2178876847,0.1257738173,-0.1987023801,0.233371824,0.1554546803,-0.0973494723,0.1492017061,0.0634052604,-0.0192187224,-0.2836737335,0.1690780222,0.3973787129,0.3210483789,-0.140442118,0.1825247407,-0.1147373766,0.0534048304,-0.0813331753,0.2980788648,0.4000996053,0.3792742491,-0.3778193593,-0.2069658488,0.4430241585,0.5543782711,-0.0169583336,0.5559698343,0.1320232004,-0.0057812151,-0.4808116853,-0.0163145717,0.0246724691,-0.2159026116,0.3507216871,0.0773191229,-0.0402618498,0.1147016659,0.0085252952,0.1612529904,-0.2590675652,0.3424575627,0.2182440758,-0.1450224817,0.0004632165,0.1746752411,-0.3607175052,-0.1042837128,-0.4627586901,-0.1236031055,-0.2472831607,-0.1884256303,-0.2096784562,-0.289891988,0.0043421714,0.2782425284,0.2054184675,0.0261446349,-0.1434522122,0.3118313849,-0.5026423931,0.0435682908,-0.1125665382,-0.0771072879,0.2508091033,-0.0200798791,-0.3463618755,-0.3060626984,0.4788448215,-0.3345266879,-0.2170309573,-0.4942217171,0.4417169392,0.3645997941,0.1880706996,-0.6248781681,-0.2481432408,0.1785907149,0.0756682828,-0.3999008536,-0.5030996799,0.116796568,0.1236272827,-0.133301273,-0.3742966056,0.4390973747,0.0781260803,0.2793938518,-0.3112007678]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2175","title":"dataset.search_batch() function outputs all -1 indices sometime.","comments":"@lhoestq @patrickvonplaten \r\n\r\nI also found another short bug in the retrieval part. Especially, when retrieving documents. If Faiss returns the -1 as the index, the retriever will always use the last element in the dataset.\r\n\r\nplease check [def get_doc_dicts function](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L222)\r\n\r\n\r\nDoes the use of the HNSW guarantee to retrieve valid indexes always? \r\n\r\n","body":"I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n ","comment_length":52,"text":"dataset.search_batch() function outputs all -1 indices sometime.\nI am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n \n@lhoestq @patrickvonplaten \r\n\r\nI also found another short bug in the retrieval part. Especially, when retrieving documents. If Faiss returns the -1 as the index, the retriever will always use the last element in the dataset.\r\n\r\nplease check [def get_doc_dicts function](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L222)\r\n\r\n\r\nDoes the use of the HNSW guarantee to retrieve valid indexes always? \r\n\r\n","embeddings":[0.015852442,-0.2806913555,-0.0632928833,0.0849646404,0.0853546485,-0.1038461328,0.2369085401,0.302310586,0.105233565,0.3437609076,-0.2517732382,-0.215343684,0.1649906933,-0.2617835999,-0.1169081405,0.0578847006,0.2875569463,0.3696747422,-0.0135725662,-0.5771781802,-0.4364783466,0.0454011261,-0.1087439135,0.359913975,-0.2922560573,0.1714099944,-0.0935413465,-0.0432227775,-0.2799686193,-0.4463257194,0.5693718195,-0.2051214129,0.4998085797,0.1934945583,-0.0001310245,0.0607319064,0.3989880085,-0.0673881844,-0.117808044,-0.1064763665,-0.1419556737,0.0830873996,0.0018900594,-0.0588475429,0.0330884568,-0.2911155522,0.2227630317,-0.1077727899,-0.0043290067,0.3318609297,0.0216136575,0.0569206215,-0.173826471,0.0023263954,0.7081764936,-0.3659300208,0.0985162109,-0.0367784202,0.48124367,-0.0528162047,0.1526664197,0.1807488054,0.0019854784,0.0009164838,-0.3093779683,0.2561921775,0.6104996204,-0.2700276077,0.1526672244,0.3558495343,0.3618957698,0.0523847118,-0.406441927,0.0502593704,0.1637898088,0.0432621539,-0.0340016484,0.2212761492,-0.0714895874,0.2482031435,0.0166672189,0.3187111318,-0.3073361218,0.2191102803,0.0984985083,0.2952780128,0.045627635,0.0873526782,0.1593000591,0.2445999533,0.3256033659,0.0117048277,0.0640360117,0.2494962066,-0.4395670891,0.0244554505,0.2559680939,-0.1571497023,0.1326080412,-0.3659854531,-0.2643686235,-0.0438166745,-0.1686107218,-0.0382474698,0.2108128071,0.098118037,-0.1500431895,0.1708924472,0.3238219619,-0.3653721213,0.0207301285,0.0379847735,0.1482292414,0.0402758159,-0.0610235743,0.0817494988,0.0378922187,-0.3919434249,-0.3166395426,-0.0747113004,-0.2071287632,0.1259294152,-0.3126271069,0.181250006,0.325061202,0.1952412128,-0.0645725131,0.1933427453,-0.2017333806,0.0103716953,-0.196739316,-0.1962479651,-0.1573488712,0.082716845,0.0814234093,-0.3686146736,-0.1698286235,0.0812144578,-0.0815586746,0.1601490825,0.0442280285,-0.2015522122,0.4831160605,0.4733822942,-0.1933667213,0.2601576149,0.0769247487,0.0277600214,-0.0702724233,0.0068190293,0.0162304454,-0.5032625198,0.166158244,0.0640356243,0.1187813208,0.3264333308,0.3128567636,0.251963079,-0.1443451345,0.3916178942,0.0385279916,-0.5134455562,0.0387769341,-0.0768570602,0.195645988,0.1660141647,0.0098022418,0.1663014144,0.0855415612,-0.1749097705,0.0766915604,0.3393226862,0.0318907797,0.3072890937,-0.5202445388,0.2322781831,0.1017867401,-0.4268246889,-0.4370739162,0.0217671208,-0.0934108943,-0.4778693914,0.0489125997,0.2361738682,0.4215180576,0.0948141888,0.1668855101,0.2142251134,-0.0300276577,-0.1694520712,-0.5196028948,-0.0608971231,0.1508563161,-0.1074654385,0.4080958366,0.305515945,0.0542486794,-0.6813327074,0.4600915015,0.0416386276,0.0762428343,-0.0491395555,0.3157004416,0.0517031178,0.6557026505,-0.0099551706,-0.0952553526,0.1157822162,-0.3426854312,0.099101387,0.0702580735,-0.1577310711,0.0223703794,-0.0416803434,-0.0441979878,0.2016412467,-0.0683345348,-0.0571705922,0.2374501228,-0.0488145016,-0.1406459659,0.0389309078,-0.2022221386,-0.1388723701,-0.5094869733,0.3016099036,0.0281648319,-0.1289514303,0.0058402522,0.0292830504,0.2807450891,-0.0501279756,-0.0498349071,0.0834944844,-0.1135855839,-0.2807566524,0.5495071411,0.1666018218,-0.1528518647,-0.5171636939,0.2230280787,0.515527904,0.1301402003,-0.2427323759,0.1156460494,0.2706033885,-0.2726711035,0.3828047514,-0.1083431393,-0.0549600907,0.0667175129,0.0663514361,-0.1498882324,-0.2030669451,0.0003471868,-0.2384688407,-0.0611852445,-0.0635788962,-0.2246356755,0.0834550411,-0.0093103051,-0.3920833766,0.0029282621,0.2973248065,0.0369533524,-0.0580252632,0.0672341213,-0.2013530284,0.3888042271,0.1604969949,-0.1281766891,-0.20559448,-0.220016703,-0.1835314184,0.0934477225,0.1413725913,-0.2600218654,0.2491225153,0.1945117414,-0.0841558129,-0.2929005027,-0.1100271419,-0.4757437706,0.1957865506,-0.256444633,-0.0163515359,0.0148865031,-0.0253463387,-0.4278128147,-0.1020294353,0.0343128778,-0.1144874915,0.1415252835,-0.2802917063,-0.0328773819,0.0512912758,-0.2847647965,0.1777866781,-0.0399978012,0.3042396605,-0.5487369895,0.066704534,-0.462308228,-0.2274498343,-0.4807326198,0.2997175455,0.0594249144,-0.025593726,-0.5072212219,-0.416665107,0.13378793,0.1072622463,-0.0663938969,0.000182708,0.2799879909,0.0276467688,-0.1641700566,0.3296678066,0.0342072994,-0.11002177,-0.098552607,0.1459679157,-0.2561155558,0.2692086101,-0.1441011876,-0.3244903386,-0.0767010301,-0.0238161776,0.0231356528,-0.2285491526,0.0347294435,0.0748743042,0.0923198536,-0.1048884764,-0.0713524669,0.0395773128,-0.2044111043,-0.1953291893,0.3142088652,0.335285753,-0.4002545178,-0.351967901,-0.2729283273,-0.5548422337,0.4818032384,-0.1490300447,-0.092409797,0.0499774329,0.0470276959,0.1055967361,0.4930113852,0.1570604444,-0.1823891848,-0.0377443954,-0.3242684007,0.0488731004,0.3383647203,0.0189887714,0.3005694151,0.1381306946,-0.1576920152,-0.3023749292,0.7045295238,0.2232642472,-0.0359439254,0.0347887501,0.2212922722,0.2890367806,0.1963255107,-0.005157379,0.179197818,0.1580584049,-0.3494434953,0.0392721295,-0.0221236814,0.1128418222,0.2741543651,0.0199249536,-0.1492028683,-0.1857704967,-0.0051587936,0.2051899433,0.2168476582,0.0696570426,0.2705717683,0.0154572148,0.0147531675,0.1499039531,0.0868911445,0.1883235127,0.2707948685,0.4090035558,-0.3244124949,-0.4757452905,0.1463624835,-0.0082476633,0.4967692494,0.3124474287,0.1192591116,0.411200434,0.2638064623,1.1679080725,-0.2185720354,0.2166092247,0.2469864637,0.0877017975,-0.199864462,-0.4820183516,-0.4562118649,-0.0837844834,0.265109092,0.2621896267,-0.1808608621,-0.1812041253,0.2296392918,-0.0946379378,-0.1061576754,-0.4746522009,-0.3367076218,-0.3955332041,0.3641491234,0.1785661876,0.0283321049,-0.0206217729,-0.2030138224,-0.2718679309,-0.2557529211,0.0980790704,-0.0298793744,-0.0057380684,0.1531223655,-0.1616355032,0.0615710616,0.6109381914,0.064232789,0.2519602776,0.0642024726,-0.5061047077,0.1954033971,-0.0409363098,-0.3315085471,0.1486862302,0.4675429761,-0.0435562171,-0.0231483877,-0.0333590694,0.1408854723,-0.0037418972,-0.2420521528,0.368968606,0.0055337413,-0.2917936742,-0.2453865409,0.6936020255,-0.1850001216,-0.3081679046,-0.0043120282,0.3812161982,-0.2878372371,0.8606350422,0.2484531254,1.0878503323,-0.0684335381,0.1362705082,-0.0076686223,-0.1831636131,0.4024516046,-0.2085948288,0.2347267419,-0.4163199365,-0.0998442993,-0.1737681031,-0.1350580305,-0.3575216234,0.5084710121,-0.0502661169,0.3532286584,-0.0885593519,-0.0280104429,0.1135660112,0.0916360989,0.4454284608,-0.1015605181,-0.2362879217,-0.0383846499,0.2732487321,0.34717834,-0.1548326313,-0.0595596284,0.1147310659,-0.0937660411,-0.3056170642,-0.1230990067,0.1350201368,0.2377492487,0.596858561,-0.2270078063,-0.166798383,-0.1047330052,0.5608096719,-0.3332360685,-0.1165125817,-0.0511674955,0.4098786712,0.0548220985,0.0997841805,-0.2400351167,0.41379565,0.1398809105,-0.390586257,-0.1648500413,0.0024678423,0.1172022671,-0.293343693,0.04266968,-0.3670483828,-0.1194263622,0.1269759089,0.0542828403,0.1731961519,-0.1075911149,-0.0152581232,0.2759130001,-0.2465018332,0.6874191165,0.053256385,-0.519998014,-0.0296495222,0.0724037439,0.3525430262,0.133305341,0.0242212564,-0.0723083317,-0.2190024257,-0.0332484841,-0.411999464,0.2119603604,-0.0836990848,0.3959652185,0.0127277542,0.0144810667,-0.0114262309,-0.1274555027,0.0600479245,-0.1624428779,-0.582880795,-0.1820282042,0.0213951636,-0.0175693482,0.0856904536,0.3645204008,-0.0187087208,-0.2293539792,0.0390162654,-0.3160556555,-0.1595782936,-0.3216889501,-0.2893075347,0.2833944559,-0.3228815794,0.1756567508,-0.122678265,-0.055429507,-0.0547152795,0.0461011343,-0.1090926528,-0.0041226093,0.0084140534,0.221054703,0.1829682589,0.1836368889,-0.1699388027,-0.0438014977,-0.1710023135,-0.0237588994,-0.0287474338,0.5818440318,0.0738729537,0.3687915802,0.2246167362,-0.0456679091,0.1085254475,0.2580301762,0.125996232,0.304102689,0.0026267129,0.4426372349,-0.1826833934,-0.0350029096,-0.1058422923,0.127791211,-0.027740363,0.1321172863,0.1057634205,-0.1753789485,-0.0949508697,-0.0052630217,0.1087535843,0.3664783835,-0.3425274789,-0.3485372365,0.1282233894,0.0778543577,-0.0981620178,-0.0787719637,0.1822721064,-0.0353828967,-0.1590308994,0.1678360999,0.0353359357,-0.3870280087,-0.1926213801,0.0282447394,0.3948737383,0.0312690288,-0.4899244606,-0.223772794,0.2902981937,0.156305328,0.1586345136,-0.1249935701,0.1711173207,0.5818685293,0.1624415368,0.1391313076,-0.0360258296,0.249774918,-0.2962177098,-0.1385935396,0.3591981828,0.4962324202,-0.317792803,0.0701233521,0.1964791119,0.073666051,0.0865558162,0.0735378638,0.0206594113,-0.2292126417,0.0370219499,0.0223149955,-0.1508489996,0.1682177186,0.1052332222,0.1328173578,0.0021786969,0.1194750816,-0.0277927537,0.1732125878,-0.0450627767,-0.0709883347,-0.0273774061,0.2930558324,0.2030259371,-0.0183952115,-0.1430911273,-0.1980174333,0.1577096879,0.3388198614,0.2720306516,0.2620838881,0.4358717203,-0.1678258032,0.0527261794,-0.0910179913,-0.1290218085,-0.1540765166,0.0157375857,-0.2296796888,-0.1882425845,0.0609381571,0.0027873844,0.0417086296,-0.1089112237,-0.1202692389,0.2378078997,-0.1023316085,0.073685348,0.1758266091,-0.1136450395,-0.1925267279,0.1888753772,-0.0533290468,0.3200596571,0.3009596467,-0.0698483288,0.241229102,0.127177313,-0.0245255455,-0.2553524077,0.2151845396,0.4262211323,0.3479794264,-0.1311557293,0.1695004106,-0.1988391429,0.0531281903,-0.1644836217,0.2272004187,0.3382124305,0.3163152933,-0.3079973459,-0.2095260918,0.4614076614,0.5569142699,0.0321503468,0.5386589766,0.1523463726,-0.1554842889,-0.463039279,0.0073626284,-0.003043212,-0.3373909593,0.2270667255,0.0924114957,-0.0356258862,-0.010150644,0.1722424775,0.1312748939,-0.2741596401,0.3313329518,0.333825767,-0.073292315,-0.0481176153,0.1569272131,-0.3059238493,-0.0399724394,-0.5270509124,-0.1893174052,-0.285797298,-0.1078230813,-0.2614035308,-0.3345785737,-0.1015726924,0.2463639528,0.2668548226,0.0416913554,-0.2155611217,0.2800278366,-0.4384513497,0.0002979704,-0.0189952329,-0.0894304216,0.179212749,0.0236870125,-0.2966307998,-0.2883657515,0.5323396921,-0.3443340361,-0.1925091445,-0.5042878389,0.3939205408,0.2632346749,0.1275881529,-0.6056138873,-0.2095988989,0.2696826458,0.0522328764,-0.4514890015,-0.4445721209,0.2302814722,0.1378475428,-0.133625865,-0.3892811239,0.4234298766,-0.0135805011,0.1584173292,-0.2815927267]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2175","title":"dataset.search_batch() function outputs all -1 indices sometime.","comments":"Hi !\r\nNo it happens sometimes to return -1, especially if your dataset is small.\r\nIf your dataset is big enough it shouldn't happen in my experience.\r\n\r\nIdeally we should ignore all the -1 that are returned. It should be possible to change that in RAG's code ","body":"I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n ","comment_length":47,"text":"dataset.search_batch() function outputs all -1 indices sometime.\nI am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n \nHi !\r\nNo it happens sometimes to return -1, especially if your dataset is small.\r\nIf your dataset is big enough it shouldn't happen in my experience.\r\n\r\nIdeally we should ignore all the -1 that are returned. It should be possible to change that in RAG's code ","embeddings":[-0.0197226517,-0.4322832227,-0.0812413171,0.0478306413,0.2197973728,-0.0908182114,0.3115728199,0.2803561389,0.1310122609,0.4431156814,-0.2505156994,-0.1531092376,0.0910181254,-0.0791352913,-0.135779947,0.0532653742,0.2860962749,0.3947462738,-0.1084839106,-0.548168242,-0.3425454497,0.0654460639,-0.088732712,0.3915581703,-0.2586427331,0.2340943366,0.0112102134,-0.1145115867,-0.2590104938,-0.4190974534,0.5721322894,-0.3347861469,0.5025041699,0.1503544003,-0.00013128,0.0766315088,0.3688726127,-0.0625466332,-0.1117210239,-0.161178425,-0.0739837587,0.1744759977,0.0725012347,-0.0385839008,0.0562549792,-0.2273409665,0.2482057959,-0.1148414165,-0.0062826844,0.2285684794,0.0220436715,0.0227524079,-0.2356109619,0.0053634937,0.6570580006,-0.3688427806,0.0722842291,-0.0749816,0.4281029403,0.0015933035,0.0991462916,0.1038106307,0.0210785158,-0.0032237314,-0.2851937711,0.2080062032,0.6046985388,-0.320632875,0.1641495377,0.2689990401,0.3400613964,-0.0508612692,-0.3605320454,0.1214827746,0.1793317497,-0.0378344953,-0.0686209723,0.1795603186,-0.0993587226,0.2582895756,-0.03508237,0.3449515402,-0.2938513756,0.118688114,0.057106927,0.3514524102,0.0614332967,0.1458185017,0.1700643599,0.2510413229,0.2481168211,0.1212699786,0.110793218,0.235305652,-0.5594887733,-0.014438441,0.2380897552,-0.2268468589,0.122181192,-0.3115229309,-0.2844860852,-0.0163496621,-0.156818971,-0.0694277585,0.1400136203,0.0411851294,-0.0960753113,0.1236311123,0.3022830486,-0.4449905157,0.0296696313,0.0645202473,0.142378822,0.0233239252,-0.0470923856,0.1370096654,0.0471978262,-0.3374822736,-0.3623555303,-0.1418831348,-0.3032471836,0.1603071094,-0.2624214888,0.0967039913,0.3457147479,0.0832174495,-0.1061626524,0.1414313018,-0.1822333634,-0.032826338,-0.1593857408,-0.1636707783,-0.1820599884,0.0722587928,0.0887960568,-0.4495787024,-0.2080674469,0.1057759598,-0.1052311733,0.1818165034,0.0921698287,-0.1559053659,0.3855243325,0.5239189267,-0.1450588852,0.3324652612,0.1163736358,0.0880202129,-0.0329306945,0.0009646671,0.0366859287,-0.5844788551,0.2138009369,0.0606951825,0.0174864233,0.3481500447,0.2698480785,0.2174264938,-0.2486609221,0.3894504011,-0.0429959819,-0.4340237975,0.0442993492,-0.0562191308,0.2539737225,0.1359872222,-0.0387292206,0.236968115,0.041999355,-0.2390343398,0.1674460173,0.2964871824,0.0506457426,0.2838562727,-0.4702191651,0.1772668213,0.0893879831,-0.3046462536,-0.3688875139,0.0061005177,-0.2453841865,-0.5936496258,0.051695548,0.2751006484,0.3244095147,0.1700236797,0.2676326036,0.2598537505,-0.0128020877,-0.1208562329,-0.4410265088,-0.0023128004,0.2443336248,-0.0659471452,0.3060710132,0.2945915163,0.0656388551,-0.7135805488,0.4023995101,0.0280543528,0.0839299932,-0.1237923503,0.3473557234,0.0027055452,0.6791746616,-0.0388453305,-0.0912725404,0.0993104801,-0.3615148664,0.0821184367,0.0451403633,-0.1641352177,-0.0131609319,-0.0566073433,0.0350172631,0.3031967878,-0.0796024427,-0.0947854519,0.2517355382,-0.1156874001,-0.2033242732,0.1115549132,-0.1613625139,-0.1303514689,-0.5231329799,0.2891619205,0.0498317219,-0.1892002821,0.0171600692,0.0105069587,0.2565825582,-0.0857755467,-0.0181058496,0.0827768147,-0.2165863514,-0.2311838567,0.6152418852,0.0245226808,-0.2053008229,-0.43799299,0.1390234083,0.4631363451,0.1479940563,-0.1994883716,0.2235160619,0.2833438814,-0.2941742539,0.4283117354,-0.1830681562,0.0042766277,0.0652105734,0.120762527,-0.1361808181,-0.1715675443,0.0841850936,-0.1806734055,-0.0789749324,0.0343611129,-0.2360083461,0.0575818382,-0.0935290903,-0.3933068812,0.0064138346,0.2676650882,0.0374962464,-0.00070187,0.131628722,-0.2041493952,0.3502353132,0.1771640331,-0.1632460952,-0.2239397317,-0.2237665355,-0.1491775513,0.0740049258,0.1370210201,-0.1925206929,0.1644243449,0.2711198032,-0.1044159979,-0.3035338819,-0.2041675895,-0.4468974471,0.2802048326,-0.2310415059,-0.0213748813,-0.015734259,0.0536570661,-0.398863256,-0.0704355016,0.0189246088,-0.0486695729,0.1597715169,-0.3013108671,-0.0147574795,0.0004013052,-0.1252846271,0.206812039,-0.0207491387,0.330542922,-0.5354221463,0.053976126,-0.4510928094,-0.2239985764,-0.5155104995,0.2345588505,0.0440515354,-0.0445373058,-0.487423569,-0.4051345885,0.1648562998,0.0389442332,-0.1175602376,-0.089323774,0.2565197945,0.0194461737,-0.1756545454,0.2952637076,-0.0040650079,0.0231820066,-0.1310558915,0.2128733248,-0.2988213003,0.3243871927,-0.2268588841,-0.3331537247,-0.0371429808,-0.0438878536,0.0810980797,-0.307970345,0.0639729798,0.0162995048,0.0980020836,-0.1462934613,-0.0720574185,-0.029819211,-0.1622832268,-0.2097996026,0.2825422287,0.3656990826,-0.3144036829,-0.3057961464,-0.2115171403,-0.5170463324,0.5374627709,-0.1317461282,-0.1213972792,0.0975359753,-0.0199507587,0.0872214958,0.4435061514,0.097622402,-0.0927802995,-0.0323477611,-0.2207289338,-0.0921771079,0.3034675121,0.0361799859,0.3236656487,0.071344167,-0.042862054,-0.2759235799,0.6984179616,0.1545129865,-0.1129388362,-0.0368444473,0.1865950525,0.1347173452,0.1537353545,0.0395465009,0.2100217938,0.1850563288,-0.4141756892,-0.0594649389,-0.06316486,0.1115485802,0.2609257102,0.1311180741,-0.0800763145,-0.125422731,-0.0258482341,0.2546736598,0.2224895805,0.0465416126,0.1755637527,-0.0444884375,0.0648728162,0.1899704784,0.0601817518,0.1394727081,0.2521349192,0.3575158119,-0.5147714615,-0.4327399135,0.1961583495,-0.0379747525,0.3839322627,0.2927122712,0.1412981302,0.4802467227,0.2687455416,1.2012723684,-0.1994317025,0.196893096,0.2090404779,0.121208325,-0.2111752331,-0.4662036598,-0.3762189448,0.1179503575,0.2625866234,0.3140963018,-0.121398665,-0.2203767598,0.3250303864,0.0044487729,-0.0341629684,-0.5240986943,-0.399772048,-0.3855193257,0.4010043442,0.2155383527,-0.0039579985,-0.0078607835,-0.1093083397,-0.3480362892,-0.2681781054,0.0921772644,0.0093285609,0.0210951976,0.1582893133,-0.0985485166,0.0732622668,0.5732539296,0.0143446969,0.0252200663,0.006418908,-0.3854689896,0.1318612546,0.051940456,-0.2821500301,0.1654088497,0.3774569333,-0.0721023381,-0.047655683,-0.111011222,0.1519591957,-0.0352654047,-0.2369250357,0.320753485,0.10408438,-0.2557940483,-0.2596049309,0.6877688766,-0.2166750133,-0.2821053863,0.0394140929,0.2832063735,-0.1812400967,0.8766024113,0.2455625832,1.0091710091,-0.1843691617,0.0763754621,0.004868608,-0.0780311748,0.3902865648,-0.1610203683,0.3290749788,-0.3745317459,0.0178982895,-0.157352224,-0.1409616768,-0.430848062,0.4691513777,0.0101766456,0.3317351043,-0.0496725477,-0.1418050081,0.1141154915,0.0149967112,0.4707395732,-0.1438022256,-0.265011996,-0.0456860997,0.2187989503,0.3855186403,-0.1249697581,-0.0634903386,0.1340590268,0.0123969512,-0.2642601132,-0.2053083926,0.174822256,0.247835204,0.6476370096,-0.2919166386,-0.1821454912,-0.1085373759,0.5909499526,-0.2857289314,-0.0287666935,-0.1408561319,0.3821340799,0.0609416179,0.0609080605,-0.2160615921,0.3016310334,0.1459929943,-0.4138074219,-0.0620545521,0.0529666655,0.1421032995,-0.218496263,0.1249319836,-0.2946150005,-0.0588992648,0.1669796556,0.1228135973,0.2005666643,-0.1120925844,-0.0280543733,0.2976492345,-0.2442921549,0.6148080826,0.0253070127,-0.5982121229,0.0017284206,0.0058569959,0.3548895419,0.1154065803,-0.0000142509,-0.0747389123,-0.1368252933,-0.0188424028,-0.4253597558,0.2005566955,-0.1114421412,0.3863447905,0.017143935,-0.0314246602,-0.0305697676,-0.1784035116,0.0075116558,-0.1205601618,-0.6033821702,-0.1258958876,-0.0730070993,-0.041494187,0.0507578552,0.2873975933,0.0348717123,-0.2175994217,0.0542430207,-0.423666358,-0.1410948187,-0.3288079798,-0.2730073333,0.3394035995,-0.3267074823,0.0706435516,-0.2071160376,-0.0144034205,-0.0259376224,0.0175403003,-0.1718308628,-0.0019626678,0.0633735359,0.2224036455,0.1406713724,0.2365950942,-0.2138760686,-0.0150444815,-0.1310726255,0.0232862644,0.071883291,0.622710526,0.1043881923,0.2232316136,0.1787254065,-0.1164205298,0.070897378,0.1596698761,0.1794396788,0.1937541068,0.0192631744,0.4420147538,-0.1910339445,-0.0397484675,-0.166454345,0.0411365889,-0.1019664556,0.1038786098,0.0475215428,-0.1832726896,-0.0966234654,0.0550627708,0.0586792901,0.2841099799,-0.3265295327,-0.3252603412,0.1167268008,0.0748208836,-0.0746403709,0.0229984187,0.1192850396,0.0479479134,-0.1677703857,0.180318743,-0.0906993598,-0.3160063922,-0.1355917305,0.078251794,0.4261501729,0.0611958057,-0.4527646601,-0.3825871348,0.2901516557,0.1490912735,0.1967364997,-0.0591326617,0.145151481,0.6113203764,0.1217516363,0.1107536033,-0.0113499593,0.1995959878,-0.371522069,-0.1344605386,0.3394747674,0.5218444467,-0.3490316272,0.0508743078,0.2930487394,0.1431473643,-0.0056756362,0.0780321434,0.0807053894,-0.2815238535,0.0190337095,0.029193962,-0.1254896522,0.2548183799,0.1008758545,0.2151561379,0.0169997159,0.1244815588,0.0994328111,0.1951429546,-0.059592776,-0.0192557182,-0.1417937726,0.4372681677,0.1537416577,0.042686075,-0.1426152885,-0.2369130254,0.0918060616,0.333809644,0.246155858,0.2434301525,0.430619508,-0.1936028302,0.1557523757,-0.1435735077,-0.1213178784,-0.1625489444,-0.1095952094,-0.2291338593,-0.0706929117,-0.0257985704,0.0164091177,0.0409961343,-0.0840604678,-0.0974233299,0.3490952253,-0.1354080439,0.0347179025,0.1858801544,-0.0455698669,-0.2363215387,0.1576511115,-0.1187545434,0.3110816181,0.2012786567,-0.0969728604,0.1226819083,0.0982028022,-0.0262509342,-0.2952477932,0.1485583931,0.402797699,0.3613175452,-0.1547199488,0.2238681763,-0.1267787218,0.0900346413,-0.1411443204,0.2908811271,0.3600698113,0.3362483084,-0.3147233725,-0.1618635803,0.4526500404,0.5721437931,-0.0233682171,0.5273842812,0.1373144984,-0.0644497424,-0.4178000987,-0.0203019921,-0.0066611231,-0.2597434819,0.3239433467,0.1009857878,-0.0186073389,0.045934312,0.0946535096,0.175449118,-0.2559640408,0.3542015553,0.2255945355,-0.1032011732,0.0405598991,0.1807113141,-0.3297010958,-0.0746265054,-0.4674162269,-0.2327865958,-0.2586240768,-0.1115681008,-0.2286154628,-0.2888632417,-0.0895061269,0.3101375997,0.2296524942,0.0700833499,-0.1351113468,0.2806587815,-0.4428521097,-0.0036483794,-0.0204245336,-0.0758653283,0.2608934343,-0.0316250548,-0.3340324163,-0.2992790937,0.4845566452,-0.3352626562,-0.2552256584,-0.4549149275,0.4566248357,0.3513981402,0.1830566078,-0.6376389861,-0.2375078648,0.1805979908,0.0340805613,-0.3883374631,-0.4541911781,0.2193976045,0.1335379332,-0.1433109641,-0.4149546623,0.4335042834,0.034969762,0.238224104,-0.3442326486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2175","title":"dataset.search_batch() function outputs all -1 indices sometime.","comments":"I also checked with some indexes it returns more -1s. Specially with IVF\nwhen nprobr is very low. It doesn't happen when using HNSW though. But at\nthe moment if it happens, dataset will always return the last element.\nMaybe we should change it to repeat the most last valid retrieved doc id.\nWhat do you think?\n\nOn Wed, Apr 7, 2021, 21:09 Quentin Lhoest ***@***.***> wrote:\n\n> Hi !\n> No it happens sometimes to return -1, especially if your dataset is small.\n> If your dataset is big enough it shouldn't happen.\n>\n> Ideally we should ignore all the -1 that are returned. It should be\n> possible to change that in RAG's code\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","body":"I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n ","comment_length":150,"text":"dataset.search_batch() function outputs all -1 indices sometime.\nI am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n \nI also checked with some indexes it returns more -1s. Specially with IVF\nwhen nprobr is very low. It doesn't happen when using HNSW though. But at\nthe moment if it happens, dataset will always return the last element.\nMaybe we should change it to repeat the most last valid retrieved doc id.\nWhat do you think?\n\nOn Wed, Apr 7, 2021, 21:09 Quentin Lhoest ***@***.***> wrote:\n\n> Hi !\n> No it happens sometimes to return -1, especially if your dataset is small.\n> If your dataset is big enough it shouldn't happen.\n>\n> Ideally we should ignore all the -1 that are returned. It should be\n> possible to change that in RAG's code\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","embeddings":[-0.0352195762,-0.3799373209,-0.0478167236,0.0425724685,0.1534609348,-0.0822973475,0.2974610329,0.3594945967,0.1545493454,0.4000560939,-0.2719640434,-0.1355304122,0.1801845729,-0.1162477732,-0.1042535901,0.0176872201,0.2510522008,0.411700666,-0.0238933545,-0.565504849,-0.3472929001,0.0554176494,-0.1621703207,0.3566616476,-0.2797409892,0.2079564035,-0.0260151736,-0.1103101596,-0.256462276,-0.4766809642,0.5943052769,-0.2659983039,0.4845439494,0.1558329761,-0.0001315398,0.0446879305,0.3873681426,-0.0088379504,-0.1438093483,-0.1818877459,-0.1090587005,0.1446238309,0.043235112,-0.0554544628,0.0236524157,-0.2460289001,0.2300864756,-0.1483960003,-0.0233964622,0.2602907419,0.0044122054,0.1053060442,-0.1652548313,0.041457992,0.6535948515,-0.2939594686,0.0850197971,-0.0507773273,0.4610896111,-0.0346091203,0.0753294975,0.1881507784,-0.001329746,0.0161234234,-0.2380205691,0.2958970666,0.5326974988,-0.2921397388,0.2075950205,0.3135941327,0.4642091095,-0.0597214401,-0.3771156371,0.0416517034,0.1993481219,-0.0407039151,-0.0111020561,0.1670993567,-0.0956297889,0.2709483504,-0.04237305,0.3521004915,-0.3264372051,0.2192137539,0.0522135571,0.2759846747,0.0092749791,0.1088937595,0.2060456723,0.2290040255,0.2474695891,0.0632587597,0.0785056502,0.2596684694,-0.5432364941,-0.0447989218,0.2311130315,-0.1491761059,0.168373853,-0.3305642009,-0.2783356011,-0.0451584458,-0.1775639057,-0.1095416769,0.2284875512,0.0602514334,-0.0660796091,0.1812396348,0.2954419553,-0.4274535179,0.0121696023,0.0389140509,0.1758261919,0.0083056372,-0.0358683094,0.0813368633,0.0619083866,-0.3679016829,-0.3316133022,-0.1007401869,-0.2808183134,0.1629868299,-0.2730956376,0.1317038685,0.3634670377,0.146359548,-0.0804848224,0.109670341,-0.2068977803,-0.0622215904,-0.1117540151,-0.1769572496,-0.2272161841,0.0936456844,0.1030369103,-0.4674134552,-0.2122441828,0.1181561574,-0.1083407924,0.1680172682,0.0629030019,-0.1599711627,0.4342661798,0.5060639977,-0.1651233733,0.3369477689,0.1049577147,0.0215298515,-0.0737526789,-0.0155770238,0.0040881461,-0.5431575179,0.1411886364,0.0422844291,0.086526081,0.2734597325,0.2544466853,0.2392120212,-0.1934601814,0.3121080399,0.0012689596,-0.4304847717,-0.0292749032,-0.018434424,0.2137634903,0.2421736568,-0.0626737326,0.2080489844,-0.0002855672,-0.1212485656,0.1199433059,0.3533531725,0.0338832662,0.302108109,-0.5125042796,0.2115641385,0.0529593378,-0.4102325737,-0.4800129235,0.0852876529,-0.2135843337,-0.4950402677,0.0450820029,0.2997730374,0.3912861049,0.1524201483,0.1798935682,0.219024986,-0.0094275316,-0.1238508746,-0.473811537,-0.0203025304,0.1659637392,-0.0641442537,0.2935773134,0.298910439,0.0591122285,-0.6651149392,0.5013080835,0.043033544,0.065996252,-0.1130268052,0.3581508398,0.0074451803,0.6426554918,-0.0502744801,-0.1780302674,0.0940999091,-0.2164572924,0.0910365358,0.030363787,-0.1974472106,-0.0065674917,-0.076018557,-0.0133750765,0.2669549882,-0.1085205749,-0.0845133141,0.2347446084,-0.1179971248,-0.2005395442,0.1396357268,-0.1923646331,-0.116550453,-0.5885253549,0.2917744815,0.0471416079,-0.1216971278,0.0123132244,-0.0259520076,0.2452685088,-0.0673241243,-0.014078211,0.0787069723,-0.1622734964,-0.2337723523,0.6361717582,0.0879751518,-0.1430283189,-0.4997425377,0.1686100215,0.422323674,0.1080616117,-0.1674082726,0.0855955258,0.2818098962,-0.2499457151,0.3955381513,-0.1710406095,0.0001389042,0.0027462172,0.109056823,-0.1679994464,-0.2835789025,0.0963128433,-0.1664516926,0.0276063457,0.0002169652,-0.2364962101,0.1579594463,0.0416519232,-0.4263933599,-0.0353489667,0.318325907,0.0457656831,-0.0429300852,0.0815228671,-0.181658417,0.3200334609,0.2101315409,-0.1190136001,-0.1937002242,-0.2115126401,-0.1423906386,0.1488105804,0.1308148801,-0.2229634821,0.219523862,0.2386860996,-0.0670403913,-0.3496905267,-0.1729846001,-0.4460815489,0.2864308655,-0.2928532362,-0.0258862637,-0.0175163746,0.0072540739,-0.39909181,-0.1714131981,0.0129362922,-0.095446676,0.1556204557,-0.296256125,0.0372462012,0.0285777561,-0.1710748076,0.2711132765,-0.0011822265,0.3199294508,-0.5916768312,0.0757883191,-0.4536616802,-0.2365648597,-0.4173086882,0.2427259088,0.0915850624,-0.0534466542,-0.4779871702,-0.4758246839,0.0625490919,0.0882033035,-0.0828444362,-0.0033015423,0.2853538692,0.0243894327,-0.1791019738,0.2776325643,-0.0406959765,-0.0596278124,-0.1343443692,0.1894851327,-0.2842754424,0.3387922943,-0.1425896436,-0.3792662621,-0.0355894528,-0.048461888,0.0492439345,-0.3014895022,0.084129557,0.1359289885,0.0926205143,-0.1223763973,-0.0579570457,-0.0589886941,-0.1991788,-0.2659586668,0.276599884,0.3525114059,-0.3270429075,-0.3649878502,-0.2087554485,-0.4985119998,0.4951010942,-0.214250505,-0.0898166597,0.049526196,0.0754159689,0.0916238055,0.4582168758,0.0985497013,-0.1285074949,-0.0332209542,-0.2222563326,-0.006869453,0.2888504863,0.0057800626,0.2680354118,0.1182754189,-0.0085810991,-0.2806569934,0.6924853325,0.2454247773,-0.1094336212,0.0033003264,0.1748466045,0.1834654957,0.0850424543,-0.051114887,0.1842916459,0.124213554,-0.3602141738,-0.0015486865,-0.022711914,0.1254473627,0.262141794,0.0651473328,-0.1408729851,-0.1906814873,0.0160578098,0.2723436356,0.1903143525,0.0644935146,0.2008590102,-0.0713967234,0.0626240969,0.1590752453,0.130851388,0.1889720708,0.2279018909,0.3748849928,-0.3430370688,-0.4709903896,0.2146635205,-0.0379671901,0.3748730123,0.2928202152,0.16982463,0.4761978984,0.2678497136,1.2380464077,-0.2207324654,0.2397738099,0.2109358013,0.0684705079,-0.1333337873,-0.3933833838,-0.3835386336,0.0539187752,0.2844854891,0.2951929867,-0.1307265759,-0.2451580912,0.3851926327,-0.0432937481,-0.0225203838,-0.5602548718,-0.3447216153,-0.3303626776,0.3607947826,0.2142215073,-0.0167532843,-0.0356720164,-0.0863249302,-0.3638100922,-0.260625422,0.0319279507,-0.0040611345,-0.0172438174,0.1569556147,-0.035199035,0.1268289685,0.5927976966,0.0563239045,0.1354800165,0.0466015004,-0.4220663309,0.1149935275,0.0003929326,-0.3203994632,0.1439979374,0.4571455419,-0.0528085008,-0.0479089916,-0.0234420802,0.1816634536,-0.0425869152,-0.2462247759,0.3428120315,0.1262637079,-0.3059759736,-0.306414932,0.7411090136,-0.185535118,-0.2672069371,0.1069349945,0.4109170437,-0.22649692,0.7764762044,0.2357637733,1.0908918381,-0.1659912318,0.0971567035,0.0304792561,-0.1151242405,0.445453465,-0.1803417802,0.2873794138,-0.4076926112,-0.0588939562,-0.1609715819,-0.122031942,-0.386790663,0.4929742217,-0.0134289348,0.2915371358,-0.0327848159,-0.1302913427,0.1176492944,0.135237664,0.4263617992,-0.1774723679,-0.3494233787,-0.0646451786,0.1964097917,0.3611923456,-0.0817258433,-0.0556192733,0.1696477234,-0.0655207187,-0.3079976737,-0.1957835257,0.1668441743,0.2483810037,0.605091691,-0.2735135555,-0.0926155746,-0.1107756197,0.5516929626,-0.2731856406,-0.0825296342,-0.1057430506,0.3689369559,0.08189556,0.1000140682,-0.2633995116,0.4028351903,0.1485235244,-0.4183520675,-0.126414448,0.0918532759,0.1044737473,-0.1990514845,0.0399839431,-0.2782661021,-0.1274038404,0.1478634328,0.1152883992,0.1704014242,-0.1256527603,-0.0417334251,0.2570219934,-0.2283674181,0.6009514928,0.0229783542,-0.5520160794,-0.0258276351,0.0755628124,0.3389896452,0.0900315046,0.0240294728,-0.0063298633,-0.1868756264,-0.0373293236,-0.373390317,0.2717144191,-0.1918026358,0.364373982,0.0309785083,0.0142212426,-0.0025919967,-0.166615665,-0.0068242112,-0.1368847489,-0.6075827479,-0.1545444578,-0.0980522931,0.0054501649,0.058507707,0.3379846811,0.0525969975,-0.171304971,0.0245960541,-0.3661860526,-0.1417574883,-0.3039373755,-0.2061253041,0.3066472709,-0.2834742963,0.0753316432,-0.1492100358,-0.0744471848,-0.0382262357,0.0128114242,-0.1530418247,0.0059130616,0.0284617618,0.2310505211,0.1268340647,0.1678782403,-0.1572222412,-0.0414525159,-0.1851231158,-0.0511417948,0.0929044783,0.6247398853,0.138448745,0.2274681926,0.171216771,-0.1496953517,0.0927992389,0.1731610745,0.1757945865,0.1908447295,0.0336913764,0.4374406934,-0.2129032463,-0.0112238359,-0.1696378142,0.1339258105,-0.0512631983,0.0851517841,0.0876811966,-0.1978400648,-0.1283259392,0.0589114428,0.1164155677,0.3177773952,-0.3164216876,-0.3802131414,0.0690348372,0.058925081,-0.1155012771,-0.0540943816,0.1288800389,0.00897454,-0.1514481902,0.1617999822,-0.0815635175,-0.3397396803,-0.145252496,0.0311050545,0.481471926,0.0169146024,-0.4494760633,-0.2749432623,0.255040735,0.1599330753,0.1452344209,-0.0567985848,0.1440955848,0.65049541,0.0611681379,0.0951553136,-0.0044952747,0.2546862066,-0.3507919312,-0.1205265,0.3938989639,0.5549640656,-0.3658357859,0.1144748554,0.3032388687,0.0590790175,0.0498141609,0.0347512215,0.0258075707,-0.2654338181,0.053526748,0.045060914,-0.1681104749,0.2421640903,0.0718769133,0.2219418734,-0.0612585023,0.1321687698,0.0572973639,0.1927283853,-0.0593719222,-0.0958711877,-0.1027271822,0.3466031253,0.2414034307,-0.0203101542,-0.0924160108,-0.2260829806,0.1413005441,0.2834900618,0.2858725786,0.2577607334,0.4385697246,-0.1007146612,0.1342716962,-0.1403617859,-0.1125695705,-0.1482642591,-0.0706932917,-0.2538696826,-0.1201609001,0.0064018872,-0.0004376241,0.0668118,-0.1140036657,-0.0617006533,0.3319973052,-0.2130603343,-0.0438318551,0.1763200313,-0.0799642205,-0.2092103809,0.1814578176,-0.0862186998,0.3858225048,0.2377939373,-0.0724797249,0.1383315027,0.1076776162,-0.0228732117,-0.2899659872,0.1609802842,0.4326782823,0.3409773111,-0.1819424778,0.193661496,-0.1942362785,0.0695758164,-0.1200070605,0.1969024986,0.375702709,0.2980351448,-0.265416652,-0.1884822398,0.4569741488,0.5234504938,-0.0345627889,0.5137478709,0.1578451693,-0.0858115032,-0.4482064545,0.0031463085,-0.0429220684,-0.3278622329,0.3053078055,0.1365034133,-0.0628277361,0.0044304468,0.213394016,0.1614729762,-0.2566328943,0.3023566604,0.2806190848,-0.08122278,-0.0041028387,0.1284438521,-0.3599644005,-0.0700793117,-0.536372602,-0.1271946281,-0.1923541278,-0.0623675995,-0.2078746706,-0.3585233092,-0.105616577,0.2334912121,0.2771334946,0.1271458566,-0.181960851,0.2790573239,-0.4390176833,0.0158227142,-0.0636896044,-0.017800685,0.2450154126,0.0175689068,-0.343360275,-0.3317607641,0.4912696481,-0.41079548,-0.2306312323,-0.4675190449,0.4894466102,0.360394448,0.124417603,-0.6537602544,-0.2610960603,0.2394976616,0.0226534978,-0.4419298172,-0.3901976645,0.2465598285,0.1512916684,-0.1403435022,-0.312725991,0.4764627218,-0.0298661496,0.2219779491,-0.3673062027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2175","title":"dataset.search_batch() function outputs all -1 indices sometime.","comments":"That would be an easy way to workaround this issue. Feel free to open a PR on `transformers` and ping me ! :)","body":"I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n ","comment_length":23,"text":"dataset.search_batch() function outputs all -1 indices sometime.\nI am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n \nThat would be an easy way to workaround this issue. Feel free to open a PR on `transformers` and ping me ! :)","embeddings":[-0.0255081374,-0.43428877,-0.0677888989,0.0737091675,0.2557210028,-0.086884737,0.3684980273,0.2831262648,0.0772832334,0.4528709054,-0.2223181874,-0.1984312534,0.0448379964,-0.063660197,-0.1256999224,-0.0068469043,0.28163445,0.4138695896,-0.1538365483,-0.5060524344,-0.2846685946,0.0919231623,-0.0996468067,0.3383357823,-0.2817484736,0.2458025515,0.0199471842,-0.1887715161,-0.2253594995,-0.4320185483,0.5846902728,-0.3023687601,0.4834949374,0.2854872942,-0.0001320533,0.1387141198,0.4092568159,-0.0608325861,-0.0931295976,-0.1343977004,-0.1120465994,0.1853378564,0.0772110745,-0.0690410882,0.0472479463,-0.2181160152,0.2536381185,-0.0957304761,0.0374494381,0.2091830373,0.0222841147,0.0297923461,-0.2128657401,-0.0140263839,0.6179968715,-0.2804864943,0.0285131484,-0.1167524979,0.4519714713,-0.0117919669,0.1072011068,0.1190893203,0.0261775609,-0.0273777936,-0.2742781341,0.1751958728,0.6338202953,-0.3051923811,0.1454397738,0.2770226002,0.3461788893,-0.0626919419,-0.3898919225,0.113210842,0.129207179,0.0541254319,-0.0549930856,0.1394111216,-0.1022167653,0.2307329923,-0.0631995201,0.3316141963,-0.3248741031,0.0969753265,0.0660300031,0.4078654051,0.0728807971,0.118746601,0.1552821845,0.2217082083,0.3452387452,0.1692684144,0.138444528,0.2564408183,-0.5704571009,-0.0328681879,0.2476677895,-0.3040004373,0.0914045572,-0.3667513728,-0.2984798849,-0.0011986484,-0.1310694665,-0.0730982348,0.1572881192,0.0958087891,-0.1555739641,0.19059591,0.289927572,-0.4025508165,0.0473156571,0.0571678989,0.133002311,-0.0341476277,-0.0552361384,0.163936168,0.0571805947,-0.3014554381,-0.3065543473,-0.187353,-0.2294137776,0.177884087,-0.2336438745,0.1199391782,0.3631218076,0.103890717,-0.1139714867,0.128608495,-0.2179741114,0.0080415085,-0.1513518244,-0.1766064316,-0.1628477275,0.0268835928,0.1277550906,-0.3845344484,-0.2226812243,0.0798276961,-0.0499852449,0.1883440316,0.0910435617,-0.1283472031,0.3765000999,0.5422126651,-0.1805920154,0.3457104862,0.1473206729,0.064840287,-0.0523054376,0.0569305569,0.0623312034,-0.5635715723,0.2131013423,0.0429791138,-0.0099195102,0.3326792121,0.2444730401,0.2408253849,-0.2267525345,0.3575468659,-0.0646775663,-0.4154815972,0.0750563741,-0.0482546538,0.2719373107,0.1569995582,-0.075921081,0.1987424791,0.0016797713,-0.2054322809,0.1476551443,0.3099027276,0.0314142443,0.2916763723,-0.4581266642,0.1396230012,0.098017633,-0.3492688537,-0.4021477401,-0.0138706183,-0.2424459308,-0.6214386821,0.057207074,0.2899440825,0.3333925307,0.2120535523,0.1955876499,0.2302583605,-0.0019288264,-0.0502748601,-0.434353888,-0.0875273794,0.2455367744,-0.0673291385,0.3384167552,0.3327716887,0.0542202145,-0.6487197876,0.3746267259,0.0277143046,0.1122425422,-0.0425458848,0.3574846685,-0.0490810797,0.6802942753,-0.0569766685,-0.0994707271,0.1098419875,-0.4044047594,0.1155566126,-0.0753009692,-0.1674618721,-0.0443620533,-0.0288728736,0.0516013429,0.2574236095,-0.0835033953,-0.135890305,0.2232373357,-0.1738039851,-0.2003886998,0.0807734653,-0.0530159362,-0.185188964,-0.5225940943,0.2937067449,0.0043019275,-0.2040100396,-0.0028494082,-0.0113272816,0.2704211771,-0.1043976992,-0.0470289476,0.1162900105,-0.2604333758,-0.1867754757,0.6034044027,0.0949822813,-0.2119152099,-0.5101861954,0.1308421791,0.5453127623,0.1332515329,-0.1913255304,0.1168153957,0.280608505,-0.2984347939,0.4549020231,-0.1223077849,0.0325306058,0.0806694478,0.0950451866,-0.129133448,-0.1698790491,0.0585401244,-0.1538615376,-0.1018987522,0.0576151647,-0.1809526682,0.0724459067,-0.0708554685,-0.3715221584,0.0288233273,0.237229839,0.0213536881,0.0044943471,0.1027630717,-0.1742998809,0.3418673277,0.1591805518,-0.1636027843,-0.2321522981,-0.1992494911,-0.1335535496,0.0652231798,0.1495561004,-0.2263540924,0.1603021175,0.2735206187,-0.1297739893,-0.2528555691,-0.2256877571,-0.3778072298,0.292773813,-0.2615428567,0.0061667948,-0.0009528858,0.0838354155,-0.4358869493,-0.0872685313,-0.0372384042,-0.0413084663,0.1497715116,-0.2632548809,-0.0305020865,0.0376622565,-0.1437921822,0.2031647116,0.0334750637,0.2562789321,-0.5662103891,0.0405999459,-0.4012899101,-0.2267560363,-0.5071386099,0.2500789762,-0.0428261831,-0.0016554183,-0.4777206182,-0.3937184811,0.1553553641,0.0370763019,-0.0991733,-0.0453234911,0.2388147116,0.0361723267,-0.1780598909,0.2443041503,-0.001067506,-0.0549965166,-0.1276286095,0.2173933536,-0.3013999462,0.328219533,-0.2105446607,-0.3242880106,-0.0426462144,-0.090696454,0.0754310787,-0.2990777194,0.0941386297,0.0747863278,0.128862083,-0.1666218042,-0.0098447958,-0.0191105604,-0.1045617387,-0.1794235855,0.338285774,0.3333112001,-0.3282365203,-0.3087780774,-0.2162129283,-0.4937134087,0.5283079743,-0.0928187445,-0.0157318972,0.0018385454,-0.0691943467,0.0996076465,0.4600244164,0.1996829659,-0.0514195599,-0.038545128,-0.2258525044,-0.0831207037,0.3009174466,0.0581529327,0.3111126125,0.0482883938,-0.0592646599,-0.273632735,0.7122138739,0.1361288577,-0.0920717418,-0.0344325975,0.1491027772,0.1136701852,0.159275949,0.0019229664,0.1652656645,0.22888349,-0.4103986323,-0.065682292,-0.0133136138,0.077308163,0.2662028074,0.1429796964,-0.1251465231,-0.1810502112,-0.0159985013,0.25155285,0.2669729888,0.0143951364,0.1463373005,-0.1024174318,0.0564950109,0.1641007662,0.0309964679,0.0924770012,0.2320384383,0.3905269504,-0.4612025619,-0.4391341805,0.1856734902,-0.040725138,0.4382944405,0.2652526498,0.1048692018,0.4620369375,0.2753231823,1.2533450127,-0.1682037264,0.2443877161,0.2285845429,0.1121662185,-0.247984007,-0.4534047544,-0.3643168211,0.099153772,0.243336916,0.2862571776,-0.1009635404,-0.2710575461,0.3411564827,-0.0417643264,-0.077075325,-0.5282347202,-0.3852576911,-0.3984296024,0.4009665847,0.1933477968,0.1002765074,-0.0016870468,-0.0981549323,-0.3518532217,-0.2977444232,0.0587855056,-0.0235365499,0.0146293957,0.1803554147,-0.0844778791,0.0557962023,0.6131926179,-0.0663987398,-0.0055677569,-0.0284146108,-0.3356005847,0.1228953227,0.1098737046,-0.2699326277,0.123008579,0.3907278776,-0.0632374063,-0.0347634703,-0.0324393846,0.1370742619,-0.0038248,-0.2497377545,0.3184871078,0.109429352,-0.2624953091,-0.2656925619,0.6580673456,-0.1930513531,-0.2695932388,-0.0207187813,0.2656034529,-0.1994875222,0.9289184213,0.2211869061,1.0775786638,-0.2336357385,0.1243536845,-0.0402222052,-0.0385025106,0.3738060892,-0.1477863789,0.3885957897,-0.4005363882,-0.0579814836,-0.1568462849,-0.1692636907,-0.4292779863,0.4705719054,-0.0453675799,0.3410059512,-0.0935931578,-0.0992786288,0.1152340844,-0.0186626371,0.5046880841,-0.1451409459,-0.2361819744,-0.0453217626,0.2728790343,0.3918407261,-0.1352116466,-0.082715787,0.0907285586,-0.0187595915,-0.2966330945,-0.1587718725,0.2126326561,0.3333903253,0.6535942554,-0.2758546472,-0.1297228485,-0.095034793,0.566845119,-0.2924180925,-0.0711946934,-0.092734538,0.3423767686,0.0585613288,0.0793097615,-0.2069378942,0.2932891548,0.173027873,-0.4413439631,-0.074992992,0.0813832507,0.0970349908,-0.2128097862,0.071943365,-0.3438452482,-0.1062026247,0.1761126369,0.0801264867,0.1485583782,-0.1392734498,-0.0243772026,0.2996040285,-0.254024446,0.634608984,0.0144603057,-0.5885603428,0.0614552163,0.055413343,0.3862685561,0.0927913263,0.0231979676,-0.0478761047,-0.1428459287,-0.0228151269,-0.4327024221,0.1647472531,-0.1825357229,0.4008662701,-0.0338533521,-0.071781598,-0.0131973941,-0.1236332729,0.0177667513,-0.0629776344,-0.6734098792,-0.1147424281,-0.1720388085,-0.0326645151,0.0396826342,0.3249145746,-0.0044476865,-0.2103179693,0.0549472347,-0.3948972225,-0.1477694958,-0.337363869,-0.3186050355,0.3224571347,-0.3186080456,0.0102541484,-0.1454121768,-0.0666963533,-0.0523236059,0.0033243876,-0.1456394345,-0.0006992357,0.0434132777,0.2336117029,0.1369237751,0.2370668054,-0.2467978001,-0.0148838693,-0.1199731529,0.0132372407,0.0394404419,0.6291817427,0.1146616489,0.2691942751,0.1905390918,-0.0973767489,0.0076064388,0.1805887818,0.2111609131,0.1675792336,0.0501722954,0.4540923536,-0.2053153962,-0.0286083557,-0.171770215,0.0796856806,-0.0547103658,0.1496181041,0.0422053784,-0.1746305972,-0.1131614819,0.0815504268,0.0865568146,0.2704322636,-0.3065719604,-0.3265143037,0.1269027889,0.0752308294,-0.1225903556,-0.0393654332,0.1416326165,0.0366361327,-0.1312530488,0.1692129821,-0.0751828328,-0.3183493614,-0.0408226326,0.1197020635,0.4419667125,0.0297926273,-0.4090233147,-0.3178193569,0.2936320901,0.1212728843,0.2321816236,-0.1055873111,0.1887025684,0.6046782136,0.0992951617,0.0593043081,-0.0757178441,0.2035538703,-0.3737027347,-0.0705192611,0.3196417391,0.4683903158,-0.2669393718,0.0369865038,0.2698330879,0.1102557406,0.0049410416,0.1236929372,0.1029183269,-0.2615015209,0.0346320868,-0.0050767437,-0.0716744363,0.239576146,0.0433638282,0.1515957862,0.0264622141,0.1367262602,0.1029605642,0.2031820863,-0.0519417524,-0.0167800058,-0.0875003114,0.4751597643,0.1755359024,0.0319664516,-0.1565060169,-0.2082107067,0.1234945357,0.2431231886,0.3446089327,0.2604810894,0.4134622514,-0.2396590412,0.2142131329,-0.1261841208,-0.1410274208,-0.1568867564,-0.110990651,-0.2464742213,-0.0485986173,-0.0899071544,0.0161397588,0.040525008,-0.097103782,-0.1191771179,0.3489648402,-0.0887707099,0.0838767812,0.1934058815,-0.1072785631,-0.1946724951,0.1110503376,-0.1409388334,0.2484157979,0.261174053,-0.1299318224,0.1621003449,0.1149600893,-0.0317187272,-0.2445857972,0.1948191673,0.3826380074,0.4384099543,-0.141425401,0.251077801,-0.171644032,0.0784000829,-0.0957523286,0.3263388574,0.3865191638,0.3334309757,-0.3761411905,-0.191206187,0.4479921162,0.5177015662,-0.0460338108,0.5572063923,0.130053699,-0.030659603,-0.371810019,-0.029229667,0.0206283797,-0.2748633623,0.3173436522,0.1746455878,-0.0274725631,0.0188972075,0.0632615611,0.1854253709,-0.2322801203,0.2986271381,0.2376524061,-0.0691029727,0.012623772,0.1746597439,-0.4002309442,-0.1194841936,-0.4669219851,-0.2581314445,-0.2126856297,-0.1061553583,-0.2768254578,-0.2955435514,-0.1151280478,0.2654784024,0.2090074271,0.0321962275,-0.1818615198,0.2971213162,-0.4628486931,0.0397841372,-0.0534088947,-0.0780032426,0.2556306422,-0.0288738236,-0.3407840729,-0.3159100711,0.4795515239,-0.3244588971,-0.305586338,-0.4587934017,0.3932493031,0.2971420288,0.195014447,-0.6422244906,-0.2094750404,0.1584324092,0.0496233068,-0.3588386476,-0.4520335793,0.2327465266,0.12982665,-0.1456965655,-0.3039952517,0.4553862512,0.0320599377,0.1474997699,-0.340958178]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2175","title":"dataset.search_batch() function outputs all -1 indices sometime.","comments":"Sure. Will push everything together with RAG end to end. :) thanks a lot.\n\nOn Wed, Apr 7, 2021, 21:16 Quentin Lhoest ***@***.***> wrote:\n\n> That would be an easy way to workaround this issue. Feel free to open a PR\n> on transformers and ping me ! :)\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","body":"I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n ","comment_length":82,"text":"dataset.search_batch() function outputs all -1 indices sometime.\nI am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, \"IVF65536_HNSW32,Flat\")**.\r\n\r\nDuring the retrieval phase exactly in [this line of retrieval_rag.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/models\/rag\/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/113782387-37a67600-9786-11eb-9c29-acad661a9648.png)\r\n\r\n\r\nHere, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ?\r\n\r\nIs this a problem of the index, where the faiss can't find any similar vector?\r\nIs there documentation on the output index being -1?\r\n\r\n@lhoestq \r\n \nSure. Will push everything together with RAG end to end. :) thanks a lot.\n\nOn Wed, Apr 7, 2021, 21:16 Quentin Lhoest ***@***.***> wrote:\n\n> That would be an easy way to workaround this issue. Feel free to open a PR\n> on transformers and ping me ! :)\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","embeddings":[-0.0283135884,-0.4459782839,-0.0621183515,0.1084139571,0.292105943,-0.0717789158,0.3549713194,0.3072511852,0.0994754583,0.437363416,-0.2883694172,-0.1778719276,0.0475873761,-0.0405465253,-0.118691586,0.0098077962,0.2470591068,0.4063918591,-0.1565123051,-0.5071401,-0.2862778902,0.1293407232,-0.0813589767,0.3280839324,-0.3517280519,0.2273842394,0.0281685721,-0.139819786,-0.2862049341,-0.4498890936,0.6027758718,-0.2830130458,0.4530468285,0.2541905642,-0.0001301647,0.1176812127,0.4395870864,-0.0651393086,-0.151789546,-0.1811907887,-0.0812221095,0.2138450742,0.0726928785,-0.0371777378,0.0060919323,-0.2177123427,0.2247651815,-0.1663434207,0.0418865085,0.1865682751,0.0233882628,0.1017680019,-0.1806958616,-0.0772505105,0.599283874,-0.2709929645,0.0486720465,-0.0523496829,0.456910044,0.0019479635,0.092516318,0.1397999376,0.0461416878,-0.0342070013,-0.2073286623,0.1863569766,0.5528726578,-0.2722924352,0.1710795015,0.2689873576,0.3746112585,-0.0904797465,-0.4090772271,0.0736160576,0.1362152398,0.0564786717,-0.0152638722,0.1417938769,-0.11472141,0.258849293,-0.10165108,0.3451104462,-0.3150436282,0.1087806076,0.0793952867,0.3879116178,-0.0039707585,0.1147775874,0.1775084287,0.2319929898,0.2189952284,0.1565219313,0.14337641,0.2508933246,-0.5799228549,-0.0268856771,0.2490704209,-0.2496403754,0.1563377976,-0.2913855016,-0.3163481653,0.00535121,-0.1451295167,-0.1174501553,0.1691085994,0.1386014223,-0.0916091129,0.1230222657,0.2601278722,-0.3778172135,0.0464000925,0.0165737811,0.1478918493,-0.0251178313,-0.1337244511,0.1508807987,0.069909282,-0.3211227953,-0.3362391889,-0.1822969168,-0.2384527177,0.156142503,-0.1856631786,0.1956372261,0.3701727986,0.0615759417,-0.1258408278,0.1341767162,-0.1884582341,-0.0083691636,-0.1735410243,-0.1935446411,-0.2100252956,0.0248548873,0.1446507871,-0.408821106,-0.1844087988,0.1028482765,-0.0403670706,0.2042436898,0.0240630526,-0.1293121874,0.381536752,0.5448859334,-0.1957634091,0.3435320258,0.1752250791,0.0462620445,-0.0807228684,0.0104634566,0.0561413988,-0.5754235387,0.1869510561,0.0365568846,-0.0247184113,0.2832264006,0.2651228607,0.2736901343,-0.2275352329,0.3701698482,-0.0708191022,-0.3711774945,0.04962001,-0.0310726147,0.2711126506,0.1841756701,-0.1032493189,0.17016913,-0.001558539,-0.2113306969,0.1576528698,0.355874896,0.0396361016,0.2429683954,-0.4880305827,0.2115790546,0.0413329266,-0.3464721441,-0.4149978459,0.0347531475,-0.2874720395,-0.5506218672,0.0943671912,0.2539378405,0.2913822234,0.2167929858,0.221115917,0.1762186736,-0.0184136033,-0.0822382942,-0.4449312091,-0.1094930619,0.1820261031,-0.0668349788,0.2814899087,0.301538825,0.0567572825,-0.6469029784,0.4162867665,-0.0154198837,0.1333107352,-0.0414281078,0.3924898803,0.0089213466,0.6352488399,-0.0670065507,-0.1997414529,0.1173175722,-0.3723261058,0.1516669691,-0.1041777655,-0.2065367252,-0.0788395107,-0.0319018774,0.0432047695,0.2024579644,-0.0909036323,-0.1343251765,0.2712657154,-0.1845749021,-0.2099106908,0.1425682306,-0.0775697306,-0.16261518,-0.5799489617,0.302626133,0.0197168775,-0.1899344325,0.0344842225,0.0379857793,0.2533740699,-0.1179477647,-0.0377238281,0.1215903759,-0.2508451939,-0.1928158849,0.6424354315,0.0720804334,-0.1853332669,-0.5010784268,0.1635936052,0.4550594389,0.1374626607,-0.1593974382,0.0422651842,0.3477858305,-0.3275406957,0.4250102639,-0.1345338076,0.035993889,0.0596402772,0.0846265033,-0.1595563889,-0.2088905871,0.0791566595,-0.145915702,-0.032150086,0.0721165463,-0.1810097545,0.1149869263,-0.0065552909,-0.3482501507,-0.0046556154,0.2404882163,-0.0307829194,0.0100295199,0.07758867,-0.1530270427,0.3684752285,0.1996846646,-0.1538051069,-0.2296267748,-0.1705859303,-0.1761599332,0.1038918048,0.1313128024,-0.2240185589,0.1574562341,0.2851550877,-0.1116567329,-0.2977469563,-0.1965698749,-0.4191932678,0.2730156183,-0.2794298828,-0.0117355324,-0.0167913362,0.0561585575,-0.451954931,-0.118975684,-0.084305644,-0.0798701793,0.143268466,-0.2832213342,-0.0262453835,0.0633004606,-0.1149929687,0.2100354284,0.0624083057,0.233408764,-0.6038302183,0.0384087674,-0.4086795151,-0.2158354968,-0.4646937549,0.250713408,0.0166676193,0.0315372348,-0.4577505291,-0.421887219,0.0902834013,0.0586623549,-0.1387365311,-0.0181439128,0.2579153776,0.0200823043,-0.2282142788,0.2372665852,0.0528728962,-0.0917406157,-0.1654131562,0.19085069,-0.2420116514,0.3424965739,-0.1977430731,-0.36896047,-0.012199576,-0.1239932403,0.1559957266,-0.3126060963,0.0653308406,0.1334224492,0.1033990234,-0.1387468874,-0.0690943077,-0.0376080275,-0.1779902577,-0.1865139157,0.3142649829,0.3371855915,-0.3545959592,-0.3330664933,-0.1831586212,-0.4907527864,0.5181811452,-0.1064134985,-0.0649667308,-0.0219290555,-0.0372577161,0.1350492388,0.4749009609,0.1957831234,-0.0861116499,-0.0533937924,-0.2279218733,-0.1095679924,0.3319778442,0.0107251741,0.2729150951,0.060941264,-0.0408687852,-0.2583755255,0.7406029105,0.2493864,-0.0772702396,-0.0038316017,0.1467562765,0.1696023941,0.1052189171,-0.0348764062,0.2000178993,0.1997401714,-0.3633256555,-0.0231813397,-0.0181912053,0.1379132718,0.2525853515,0.1331998408,-0.1205030531,-0.2060520202,0.0133614587,0.2651364207,0.2366911769,0.0554277636,0.1293682158,-0.1069733128,0.0863468871,0.1686658412,0.0602950118,0.1176761612,0.219700247,0.3878758848,-0.385327965,-0.5046994686,0.2042442858,-0.0252878256,0.3977499902,0.2302015573,0.0910199508,0.4760736823,0.2946364582,1.2821089029,-0.1257203072,0.2100656927,0.2024417073,0.0521936119,-0.2569007277,-0.384518981,-0.3927997053,0.1071787402,0.2438451201,0.3198533654,-0.1448898166,-0.3078741133,0.3683764338,-0.0393618532,-0.0377905779,-0.5277113318,-0.3560360074,-0.3641533554,0.3309241533,0.211603269,0.1023004353,0.037355192,-0.0724366158,-0.3457964063,-0.2657880485,0.0430624038,0.0285140872,-0.0076313294,0.1817750037,-0.0545258895,0.0753951073,0.6259134412,0.004059209,0.0351529978,0.0407277755,-0.3378148377,0.1058724448,0.0830655992,-0.2505222857,0.1562091708,0.4710416496,-0.0778070837,-0.0682384297,-0.0047484706,0.1829999983,-0.0190184861,-0.2835802436,0.2942579091,0.1262440234,-0.2307731062,-0.2612527311,0.7115757465,-0.199929595,-0.2380958647,-0.0338412896,0.3728461266,-0.1944930255,0.8509473205,0.1994220763,1.1119365692,-0.2306782156,0.1241329834,0.0136608826,-0.0229497552,0.4233249724,-0.1478209198,0.3599698246,-0.4168886542,-0.0745230019,-0.1267657727,-0.1625534594,-0.3965692222,0.5001984835,-0.048585169,0.3075248301,-0.0509979501,-0.0995346233,0.1122645363,0.0322758742,0.43194139,-0.1958491802,-0.3082029521,-0.0494016297,0.2895002365,0.4239201546,-0.1144874766,-0.130663693,0.0609413087,-0.0434389301,-0.3266913295,-0.1706413925,0.1495880485,0.3132311404,0.6839635372,-0.2813191712,-0.0597472079,-0.1292348951,0.5940539837,-0.2949202061,-0.0975086987,-0.0932810828,0.2968186736,0.0448546074,0.087817423,-0.2085441351,0.3264082074,0.1658751667,-0.4087222517,-0.118134357,0.0887144431,0.1004107669,-0.1945730746,0.039905522,-0.3729729652,-0.1658736169,0.134651795,0.0667743608,0.1419640929,-0.1279295832,-0.0285451673,0.2953920364,-0.2502564788,0.5640613437,0.0565908775,-0.5724158883,0.0212692767,0.0987817124,0.3291194439,0.0724500194,0.0501314737,0.00084865,-0.1735296249,-0.0331755728,-0.4581155777,0.2036810666,-0.2213502824,0.4114436507,0.0110916402,-0.0499434769,0.0116131147,-0.1267791241,0.0472033173,-0.087239176,-0.6709446311,-0.1352551579,-0.2030446678,-0.0041326014,0.0396138355,0.3588506281,-0.0133222211,-0.1280950606,0.0652358234,-0.3754579723,-0.166390568,-0.3348731399,-0.2888953984,0.2904349566,-0.2414183617,-0.0278278422,-0.1007804573,-0.0519361906,-0.0417564362,-0.0030182505,-0.2036745548,-0.0052222391,0.0491063111,0.2304776758,0.1264705956,0.2335477918,-0.2158169746,0.024791142,-0.1576772481,-0.0441032983,0.1189129576,0.6434118748,0.116807647,0.2623495758,0.2124279886,-0.1490022689,0.0374207161,0.1779783666,0.2131088823,0.1592005193,0.0470374934,0.3956907988,-0.2537586093,-0.0300068911,-0.1442241371,0.1031638384,-0.0178730078,0.1729230285,0.0132562434,-0.1739442497,-0.1685121655,0.125555858,0.1405684501,0.2671737075,-0.2686234117,-0.3306474388,0.100789696,0.0750902146,-0.1332664937,-0.0001478569,0.1461533457,0.0295173004,-0.1004157215,0.1746884435,-0.0606295168,-0.2794895172,-0.0094918553,0.1139700487,0.469147712,0.049252443,-0.404843688,-0.282820493,0.3373284638,0.1346959174,0.2748322189,-0.0744078159,0.2376716882,0.623999536,0.0158929117,0.0542862341,-0.1314190626,0.2484118789,-0.3704633415,-0.1058534756,0.3135043979,0.5167598128,-0.3027828336,0.0218010843,0.2416456044,0.2011872381,-0.0204311032,0.1160022467,0.0708290562,-0.2149723917,0.0191192348,-0.037917722,-0.0710589066,0.214370653,0.0164968297,0.1525220126,-0.021142507,0.1446314752,0.1464318484,0.1926598847,-0.0222269129,-0.1095518693,-0.1289520264,0.4419502616,0.2132030725,0.0392160937,-0.1258063167,-0.2113787234,0.1281358898,0.241447553,0.3575932086,0.313642323,0.395290345,-0.146236673,0.2081685811,-0.1472530067,-0.1377396286,-0.1769449115,-0.081948489,-0.2318558544,-0.0910958275,-0.0467151366,0.0268883929,0.0375411324,-0.1590422988,-0.1036705971,0.349565655,-0.1056792885,0.0548236035,0.1510559469,-0.0851700008,-0.1924754381,0.1248358339,-0.1775255352,0.2826818228,0.2329522669,-0.1194288656,0.1866091788,0.1175417006,-0.0125525119,-0.2626390457,0.1983684897,0.3687415123,0.3809488416,-0.1608276218,0.2081639916,-0.2221247554,0.0555270463,-0.0466012284,0.2873368859,0.3420760036,0.3129171431,-0.318891108,-0.1757490486,0.4365797043,0.4873604178,-0.0404291935,0.5094999075,0.1159966066,-0.0109986337,-0.3908467889,-0.0488444939,0.0390698314,-0.2555005848,0.3060807586,0.2037486434,-0.0464076214,-0.000629782,0.1717254221,0.1663027853,-0.2835230529,0.2716397941,0.2651441991,-0.0683029741,-0.0019562952,0.1600393057,-0.4410412014,-0.1277185977,-0.4847532213,-0.1873944998,-0.198332876,-0.059410885,-0.2815765738,-0.3265518844,-0.1115984246,0.2839322686,0.2401632369,0.0392763875,-0.1936526,0.3200109303,-0.4926871359,0.0414746404,-0.061620634,-0.001981572,0.2617309988,-0.0137672815,-0.3590401113,-0.3816456795,0.5597449541,-0.3355350196,-0.3149106205,-0.4205088913,0.4017616212,0.3072033226,0.1281175464,-0.6617913842,-0.1746369302,0.1516131461,0.0570147224,-0.3895004988,-0.4049233198,0.2210484594,0.0872343332,-0.142376855,-0.3343044817,0.4810689092,0.0113392966,0.1717561483,-0.3487364948]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2170","title":"Wikipedia historic dumps are deleted but hf\/datasets hardcodes dump date","comments":"It seems that this can be fixed from user's end by including a `date` argument, like this:\r\n\r\n`dataset = datasets.load_dataset('wikipedia', '20200501.en', date='20210420')`\r\n\r\nYou can get available dates from [here](https:\/\/dumps.wikimedia.org\/enwiki\/).\r\n\r\nThis is not a proper fix however as all the files will still have '20200501' in their file names.","body":"Wikimedia does not keep all historical dumps. For example, as of today https:\/\/dumps.wikimedia.org\/kowiki\/ only provides\r\n\r\n```\r\n20201220\/ 02-Feb-2021 01:36 -\r\n20210101\/ 21-Feb-2021 01:26 -\r\n20210120\/ 02-Mar-2021 01:25 -\r\n20210201\/ 21-Mar-2021 01:26 -\r\n20210220\/ 02-Apr-2021 01:26 -\r\n20210301\/ 03-Mar-2021 08:10 -\r\n20210320\/ 21-Mar-2021 18:13 -\r\n20210401\/ 03-Apr-2021 10:08 -\r\nlatest\/ 03-Apr-2021 10:08 -\r\n```\r\n\r\nHowever, the wikipedia dataset provided in the library, only supports the following configs, none of which are applicable anymore when disregarding the cached datasets:\r\n\r\n```\r\nValueError: BuilderConfig 20210401.ko not found. Available: ['20200501.aa', '20200501.ab', '20200501.ace', '20200501.ady', '20200501.af', '20200501.ak', '20200501.als', '20200501.am', '20200501.an', '20200501.ang', '20200501.ar', '20200501.arc', '20200501.arz', '20200501.as', '20200501.ast', '20200501.atj', '20200501.av', '20200501.ay', '20200501.az', '20200501.azb', '20200501.ba', '20200501.bar', '20200501.bat-smg', '20200501.bcl', '20200501.be', '20200501.be-x-old', '20200501.bg', '20200501.bh', '20200501.bi', '20200501.bjn', '20200501.bm', '20200501.bn', '20200501.bo', '20200501.bpy', '20200501.br', '20200501.bs', '20200501.bug', '20200501.bxr', '20200501.ca', '20200501.cbk-zam', '20200501.cdo', '20200501.ce', '20200501.ceb', '20200501.ch', '20200501.cho', '20200501.chr', '20200501.chy', '20200501.ckb', '20200501.co', '20200501.cr', '20200501.crh', '20200501.cs', '20200501.csb', '20200501.cu', '20200501.cv', '20200501.cy', '20200501.da', '20200501.de', '20200501.din', '20200501.diq', '20200501.dsb', '20200501.dty', '20200501.dv', '20200501.dz', '20200501.ee', '20200501.el', '20200501.eml', '20200501.en', '20200501.eo', '20200501.es', '20200501.et', '20200501.eu', '20200501.ext', '20200501.fa', '20200501.ff', '20200501.fi', '20200501.fiu-vro', '20200501.fj', '20200501.fo', '20200501.fr', '20200501.frp', '20200501.frr', '20200501.fur', '20200501.fy', '20200501.ga', '20200501.gag', '20200501.gan', '20200501.gd', '20200501.gl', '20200501.glk', '20200501.gn', '20200501.gom', '20200501.gor', '20200501.got', '20200501.gu', '20200501.gv', '20200501.ha', '20200501.hak', '20200501.haw', '20200501.he', '20200501.hi', '20200501.hif', '20200501.ho', '20200501.hr', '20200501.hsb', '20200501.ht', '20200501.hu', '20200501.hy', '20200501.ia', '20200501.id', '20200501.ie', '20200501.ig', '20200501.ii', '20200501.ik', '20200501.ilo', '20200501.inh', '20200501.io', '20200501.is', '20200501.it', '20200501.iu', '20200501.ja', '20200501.jam', '20200501.jbo', '20200501.jv', '20200501.ka', '20200501.kaa', '20200501.kab', '20200501.kbd', '20200501.kbp', '20200501.kg', '20200501.ki', '20200501.kj', '20200501.kk', '20200501.kl', '20200501.km', '20200501.kn', '20200501.ko', '20200501.koi', '20200501.krc', '20200501.ks', '20200501.ksh', '20200501.ku', '20200501.kv', '20200501.kw', '20200501.ky', '20200501.la', '20200501.lad', '20200501.lb', '20200501.lbe', '20200501.lez', '20200501.lfn', '20200501.lg', '20200501.li', '20200501.lij', '20200501.lmo', '20200501.ln', '20200501.lo', '20200501.lrc', '20200501.lt', '20200501.ltg', '20200501.lv', '20200501.mai', '20200501.map-bms', '20200501.mdf', '20200501.mg', '20200501.mh', '20200501.mhr', '20200501.mi', '20200501.min', '20200501.mk', '20200501.ml', '20200501.mn', '20200501.mr', '20200501.mrj', '20200501.ms', '20200501.mt', '20200501.mus', '20200501.mwl', '20200501.my', '20200501.myv', '20200501.mzn', '20200501.na', '20200501.nah', '20200501.nap', '20200501.nds', '20200501.nds-nl', '20200501.ne', '20200501.new', '20200501.ng', '20200501.nl', '20200501.nn', '20200501.no', '20200501.nov', '20200501.nrm', '20200501.nso', '20200501.nv', '20200501.ny', '20200501.oc', '20200501.olo', '20200501.om', '20200501.or', '20200501.os', '20200501.pa', '20200501.pag', '20200501.pam', '20200501.pap', '20200501.pcd', '20200501.pdc', '20200501.pfl', '20200501.pi', '20200501.pih', '20200501.pl', '20200501.pms', '20200501.pnb', '20200501.pnt', '20200501.ps', '20200501.pt', '20200501.qu', '20200501.rm', '20200501.rmy', '20200501.rn', '20200501.ro', '20200501.roa-rup', '20200501.roa-tara', '20200501.ru', '20200501.rue', '20200501.rw', '20200501.sa', '20200501.sah', '20200501.sat', '20200501.sc', '20200501.scn', '20200501.sco', '20200501.sd', '20200501.se', '20200501.sg', '20200501.sh', '20200501.si', '20200501.simple', '20200501.sk', '20200501.sl', '20200501.sm', '20200501.sn', '20200501.so', '20200501.sq', '20200501.sr', '20200501.srn', '20200501.ss', '20200501.st', '20200501.stq', '20200501.su', '20200501.sv', '20200501.sw', '20200501.szl', '20200501.ta', '20200501.tcy', '20200501.te', '20200501.tet', '20200501.tg', '20200501.th', '20200501.ti', '20200501.tk', '20200501.tl', '20200501.tn', '20200501.to', '20200501.tpi', '20200501.tr', '20200501.ts', '20200501.tt', '20200501.tum', '20200501.tw', '20200501.ty', '20200501.tyv', '20200501.udm', '20200501.ug', '20200501.uk', '20200501.ur', '20200501.uz', '20200501.ve', '20200501.vec', '20200501.vep', '20200501.vi', '20200501.vls', '20200501.vo', '20200501.wa', '20200501.war', '20200501.wo', '20200501.wuu', '20200501.xal', '20200501.xh', '20200501.xmf', '20200501.yi', '20200501.yo', '20200501.za', '20200501.zea', '20200501.zh', '20200501.zh-classical', '20200501.zh-min-nan', '20200501.zh-yue', '20200501.zu']\r\n```\r\n\r\nThe cached datasets:\r\n\r\n```\r\n% aws s3 --no-sign-request --endpoint-url https:\/\/storage.googleapis.com ls s3:\/\/huggingface-nlp\/cache\/datasets\/wikipedia\/\r\n PRE 20200501.de\/\r\n PRE 20200501.en\/\r\n PRE 20200501.fr\/\r\n PRE 20200501.frr\/\r\n PRE 20200501.it\/\r\n PRE 20200501.simple\/\r\n```","comment_length":48,"text":"Wikipedia historic dumps are deleted but hf\/datasets hardcodes dump date\nWikimedia does not keep all historical dumps. For example, as of today https:\/\/dumps.wikimedia.org\/kowiki\/ only provides\r\n\r\n```\r\n20201220\/ 02-Feb-2021 01:36 -\r\n20210101\/ 21-Feb-2021 01:26 -\r\n20210120\/ 02-Mar-2021 01:25 -\r\n20210201\/ 21-Mar-2021 01:26 -\r\n20210220\/ 02-Apr-2021 01:26 -\r\n20210301\/ 03-Mar-2021 08:10 -\r\n20210320\/ 21-Mar-2021 18:13 -\r\n20210401\/ 03-Apr-2021 10:08 -\r\nlatest\/ 03-Apr-2021 10:08 -\r\n```\r\n\r\nHowever, the wikipedia dataset provided in the library, only supports the following configs, none of which are applicable anymore when disregarding the cached datasets:\r\n\r\n```\r\nValueError: BuilderConfig 20210401.ko not found. Available: ['20200501.aa', '20200501.ab', '20200501.ace', '20200501.ady', '20200501.af', '20200501.ak', '20200501.als', '20200501.am', '20200501.an', '20200501.ang', '20200501.ar', '20200501.arc', '20200501.arz', '20200501.as', '20200501.ast', '20200501.atj', '20200501.av', '20200501.ay', '20200501.az', '20200501.azb', '20200501.ba', '20200501.bar', '20200501.bat-smg', '20200501.bcl', '20200501.be', '20200501.be-x-old', '20200501.bg', '20200501.bh', '20200501.bi', '20200501.bjn', '20200501.bm', '20200501.bn', '20200501.bo', '20200501.bpy', '20200501.br', '20200501.bs', '20200501.bug', '20200501.bxr', '20200501.ca', '20200501.cbk-zam', '20200501.cdo', '20200501.ce', '20200501.ceb', '20200501.ch', '20200501.cho', '20200501.chr', '20200501.chy', '20200501.ckb', '20200501.co', '20200501.cr', '20200501.crh', '20200501.cs', '20200501.csb', '20200501.cu', '20200501.cv', '20200501.cy', '20200501.da', '20200501.de', '20200501.din', '20200501.diq', '20200501.dsb', '20200501.dty', '20200501.dv', '20200501.dz', '20200501.ee', '20200501.el', '20200501.eml', '20200501.en', '20200501.eo', '20200501.es', '20200501.et', '20200501.eu', '20200501.ext', '20200501.fa', '20200501.ff', '20200501.fi', '20200501.fiu-vro', '20200501.fj', '20200501.fo', '20200501.fr', '20200501.frp', '20200501.frr', '20200501.fur', '20200501.fy', '20200501.ga', '20200501.gag', '20200501.gan', '20200501.gd', '20200501.gl', '20200501.glk', '20200501.gn', '20200501.gom', '20200501.gor', '20200501.got', '20200501.gu', '20200501.gv', '20200501.ha', '20200501.hak', '20200501.haw', '20200501.he', '20200501.hi', '20200501.hif', '20200501.ho', '20200501.hr', '20200501.hsb', '20200501.ht', '20200501.hu', '20200501.hy', '20200501.ia', '20200501.id', '20200501.ie', '20200501.ig', '20200501.ii', '20200501.ik', '20200501.ilo', '20200501.inh', '20200501.io', '20200501.is', '20200501.it', '20200501.iu', '20200501.ja', '20200501.jam', '20200501.jbo', '20200501.jv', '20200501.ka', '20200501.kaa', '20200501.kab', '20200501.kbd', '20200501.kbp', '20200501.kg', '20200501.ki', '20200501.kj', '20200501.kk', '20200501.kl', '20200501.km', '20200501.kn', '20200501.ko', '20200501.koi', '20200501.krc', '20200501.ks', '20200501.ksh', '20200501.ku', '20200501.kv', '20200501.kw', '20200501.ky', '20200501.la', '20200501.lad', '20200501.lb', '20200501.lbe', '20200501.lez', '20200501.lfn', '20200501.lg', '20200501.li', '20200501.lij', '20200501.lmo', '20200501.ln', '20200501.lo', '20200501.lrc', '20200501.lt', '20200501.ltg', '20200501.lv', '20200501.mai', '20200501.map-bms', '20200501.mdf', '20200501.mg', '20200501.mh', '20200501.mhr', '20200501.mi', '20200501.min', '20200501.mk', '20200501.ml', '20200501.mn', '20200501.mr', '20200501.mrj', '20200501.ms', '20200501.mt', '20200501.mus', '20200501.mwl', '20200501.my', '20200501.myv', '20200501.mzn', '20200501.na', '20200501.nah', '20200501.nap', '20200501.nds', '20200501.nds-nl', '20200501.ne', '20200501.new', '20200501.ng', '20200501.nl', '20200501.nn', '20200501.no', '20200501.nov', '20200501.nrm', '20200501.nso', '20200501.nv', '20200501.ny', '20200501.oc', '20200501.olo', '20200501.om', '20200501.or', '20200501.os', '20200501.pa', '20200501.pag', '20200501.pam', '20200501.pap', '20200501.pcd', '20200501.pdc', '20200501.pfl', '20200501.pi', '20200501.pih', '20200501.pl', '20200501.pms', '20200501.pnb', '20200501.pnt', '20200501.ps', '20200501.pt', '20200501.qu', '20200501.rm', '20200501.rmy', '20200501.rn', '20200501.ro', '20200501.roa-rup', '20200501.roa-tara', '20200501.ru', '20200501.rue', '20200501.rw', '20200501.sa', '20200501.sah', '20200501.sat', '20200501.sc', '20200501.scn', '20200501.sco', '20200501.sd', '20200501.se', '20200501.sg', '20200501.sh', '20200501.si', '20200501.simple', '20200501.sk', '20200501.sl', '20200501.sm', '20200501.sn', '20200501.so', '20200501.sq', '20200501.sr', '20200501.srn', '20200501.ss', '20200501.st', '20200501.stq', '20200501.su', '20200501.sv', '20200501.sw', '20200501.szl', '20200501.ta', '20200501.tcy', '20200501.te', '20200501.tet', '20200501.tg', '20200501.th', '20200501.ti', '20200501.tk', '20200501.tl', '20200501.tn', '20200501.to', '20200501.tpi', '20200501.tr', '20200501.ts', '20200501.tt', '20200501.tum', '20200501.tw', '20200501.ty', '20200501.tyv', '20200501.udm', '20200501.ug', '20200501.uk', '20200501.ur', '20200501.uz', '20200501.ve', '20200501.vec', '20200501.vep', '20200501.vi', '20200501.vls', '20200501.vo', '20200501.wa', '20200501.war', '20200501.wo', '20200501.wuu', '20200501.xal', '20200501.xh', '20200501.xmf', '20200501.yi', '20200501.yo', '20200501.za', '20200501.zea', '20200501.zh', '20200501.zh-classical', '20200501.zh-min-nan', '20200501.zh-yue', '20200501.zu']\r\n```\r\n\r\nThe cached datasets:\r\n\r\n```\r\n% aws s3 --no-sign-request --endpoint-url https:\/\/storage.googleapis.com ls s3:\/\/huggingface-nlp\/cache\/datasets\/wikipedia\/\r\n PRE 20200501.de\/\r\n PRE 20200501.en\/\r\n PRE 20200501.fr\/\r\n PRE 20200501.frr\/\r\n PRE 20200501.it\/\r\n PRE 20200501.simple\/\r\n```\nIt seems that this can be fixed from user's end by including a `date` argument, like this:\r\n\r\n`dataset = datasets.load_dataset('wikipedia', '20200501.en', date='20210420')`\r\n\r\nYou can get available dates from [here](https:\/\/dumps.wikimedia.org\/enwiki\/).\r\n\r\nThis is not a proper fix however as all the files will still have '20200501' in their file names.","embeddings":[-0.0480471887,0.339981854,-0.0221922565,0.0315804072,-0.3193795681,0.156512931,0.3078280389,0.5392638445,0.176421985,0.1010856703,-0.0753563866,0.0723977834,0.2024516314,-0.2449931204,-0.1113701835,-0.182833761,0.0309121516,0.0772673413,-0.1100769192,-0.2934595644,-0.1943483204,0.1416632831,-0.297540009,-0.1001333967,-0.1351926029,0.1282684058,-0.1212271452,-0.1731771827,-0.236763224,-0.2825134099,0.2279289067,0.2149960697,0.2150947005,0.2445408404,-0.0001126101,-0.0224174578,0.4396784306,-0.0729151815,-0.6998178959,0.3375677466,-0.4934145212,-0.2081644088,-0.2113407701,-0.2803381681,0.0249623079,0.0599011295,0.1523051113,-0.0141651891,0.2861972451,0.1448759735,0.2360973507,-0.0227914248,0.2254752666,-0.168529287,0.3714103401,0.313983798,-0.1705524623,-0.1458384842,-0.3366081715,-0.0145526333,-0.08039698,0.5216310024,-0.0483497046,-0.2389984727,0.1012657657,-0.1834843755,-0.1664302051,-0.1672370434,0.4476556182,0.3567122519,0.5580341816,-0.1420494467,-0.1736138016,0.0477034599,-0.2691888511,-0.3119944036,0.3322288692,0.1879585236,-0.0743197575,0.086736396,0.0434375294,-0.1867388338,-0.1235563383,0.4326101243,-0.2945711315,0.6903668046,0.0475939922,-0.0200513583,-0.2650398314,-0.0862197578,-0.0129311718,-0.316567719,0.0335615203,0.3412547708,0.1046115011,-0.1992751062,-0.0340665504,0.1706593335,0.1952961534,-0.2298509181,-0.1247055903,0.1728707105,0.1436997503,-0.084455736,0.2835518122,-0.2561607063,0.2181194872,0.14792867,0.4882195592,0.3513464332,-0.199255392,-0.0234843548,0.3428494632,0.0529373288,-0.0747671202,-0.1216740608,0.1743231118,-0.3396600187,0.1949049085,0.1505921632,0.1067213863,-0.3127460182,-0.3135461807,0.3275847137,-0.1723534465,0.0550002493,-0.0030665204,-0.1820926815,-0.1246303394,-0.1386279911,-0.1526300758,-0.0057441792,0.1183300242,0.1284084022,0.2885450423,-0.1358129829,0.3161966801,0.2317119241,-0.0991327614,0.1145829186,-0.1741470397,0.0152740525,0.2022375613,0.3797014952,0.0382675529,0.1678479016,-0.1536102742,-0.2450581938,-0.3076083064,-0.0909029543,-0.2692331076,0.003493526,-0.3159883618,0.2036642581,0.1309011728,0.023593137,0.0376099013,0.1419020146,0.1953950673,-0.2002358288,0.0092894314,-0.0318655558,-0.0295625664,-0.1204090193,0.2662154734,0.1527356058,-0.1783032268,-0.1148426086,-0.0741789192,0.3705844283,0.1859222949,0.1107274964,-0.0983730704,0.0929739103,-0.1122706607,-0.2385077775,0.2569867074,-0.0488516539,-0.2477300316,-0.142287612,0.5185168386,0.1292577237,-0.0351007208,-0.1889841408,0.5140385032,-0.2660749555,0.0054178634,0.1362960041,0.1500758231,-0.0966951251,-0.1010837778,-0.332380712,0.0940822363,-0.1349024326,0.5809412599,-0.0463306271,0.3364216387,0.6471557617,0.377738893,0.0675285086,0.2076520026,0.3722668886,-0.1182639599,-0.0895943046,0.2840391397,-0.2195201367,-0.3384561539,0.2157478929,-0.1119534969,0.1423028857,0.0572796129,-0.2991670668,-0.2290911525,0.0036074684,-0.1578158438,-0.1922062188,0.2709792554,0.1160458103,0.2696571946,0.3821341693,0.0707488135,-0.2911677361,-0.0803748667,0.0224439856,-0.1743042469,0.2483406663,-0.3143912852,0.0948141441,0.0104760071,0.1629978865,0.2624762356,0.0467962511,0.1376544833,-0.1132883802,0.1075474173,0.1248564869,-0.0343472473,0.0228770245,0.2571222782,-0.2377163917,0.1016275734,0.3862405419,-0.0660331622,-0.1089660376,-0.3348828554,0.2010368705,0.3396961689,-0.0081818094,-0.096274212,-0.0002625089,0.2850736976,-0.0728488266,0.1175118685,-0.4912272096,-0.1852969378,0.2930184901,-0.1906876117,0.0877466351,-0.094600752,0.2002782524,0.6361352801,-0.0283581223,0.0360681862,0.3035472631,-0.3020724654,-0.3853172362,0.3101589084,-0.1874304563,-0.3706717193,0.1790653914,0.2985033989,-0.18427293,-0.191159457,-0.100862518,0.3368108273,0.0546802357,0.2021879703,0.0554028116,0.3110338151,0.2944017947,-0.3105643094,0.1880522519,-0.1699729711,0.085625492,-0.001388579,-0.0112569062,-0.5644822717,0.0784604028,0.1482223868,-0.0970541537,-0.5503886342,-0.6236146092,0.2578138113,0.0797467902,-0.121780701,0.1836832017,-0.2991448939,0.2259609252,0.1115831584,0.2293134332,-0.359590143,-0.3663398325,-0.1580598056,0.0799806118,0.0267514233,-0.1627448797,0.2104156166,-0.1402296871,-0.5127974749,-0.4433472753,-0.6245263815,0.3214212656,-0.0778710693,-0.0595676601,0.1482465565,0.2876193821,0.0446178094,-0.3598255515,-0.0445629284,-0.0845200792,-0.0961177647,-0.1044833139,-0.3406780362,0.240445748,-0.0595720522,0.1670760661,-0.1340950131,-0.1953468472,-0.1555752754,0.4140326381,0.0772166699,0.2195165008,-0.0850676894,-0.3220025897,0.0055803121,0.208600238,-0.4126555026,-0.3447987735,0.5001904368,-0.3294726312,-0.3419498801,0.169386819,-0.0701691285,-0.0787318423,0.0163808577,-0.4505988955,0.1796373427,0.0271939728,0.3788494468,0.3891842067,-0.0260350276,0.2755755782,0.0348579735,-0.0446544662,-0.0865289196,0.0076209502,-0.111386463,-0.0580644794,0.3612657189,0.1532097161,0.0982783735,0.046685338,0.9649457932,0.129607752,0.175260812,0.2703441381,0.0933707729,0.422911495,-0.296841979,0.031347923,0.0888592005,0.0506782047,0.1441750526,0.2668755651,0.114292033,-0.3368980885,0.0401475616,-0.0561869442,-0.1404110491,-0.4507347941,-0.2183957249,0.1395625025,0.1739774793,0.2208260894,0.1641782522,-0.3181213737,-0.4052608609,0.3040072024,0.1644185334,0.0470861085,0.1440111399,-0.2075382769,0.1267812997,-0.5879554152,0.1699393988,-0.0450913571,0.0630301237,-0.1274249703,0.0217734277,-0.0962825865,0.0595289394,-0.0420555696,-0.2388567328,0.2718627751,0.2691740096,0.0239767078,-0.1405697912,0.0205693413,-0.0479287058,0.1030628458,0.2742254138,0.2927269638,-0.3884562552,-0.1339472234,0.4520693123,0.314946115,-0.1313464493,0.0320040882,0.1103868708,-0.1165558547,-0.1930273771,-0.3179684281,0.0428573973,0.148812741,-0.2054591775,0.1010592878,-0.2121364921,0.0956856534,-0.0005831206,-0.0347558893,0.1180442274,0.3438915312,-0.0756219551,-0.0243572369,0.2225073129,0.136364758,-0.0072103338,0.0231025796,-0.0037992529,-0.1215571314,-0.2037767023,-0.0587098747,0.199733004,-0.0520161353,0.0165819898,-0.044390168,0.0522500724,-0.1027558595,-0.2034795731,-0.1904218495,-0.0853029341,-0.3912639618,0.0338613093,0.4699323177,-0.0343349203,0.0513630472,0.3707818389,0.2155039757,-0.4664890766,0.3854550123,0.2939129174,0.8347738981,0.1898766011,0.1881596148,0.1820242256,0.1101787612,0.3675718009,0.0527264886,0.0521540716,-0.1155320555,0.075399071,-0.0228813402,0.3158735335,-0.0470613502,-0.2119174451,-0.147010684,0.3688459694,-0.1156375334,0.495500952,-0.0295316186,0.1209016219,0.0672810897,-0.107918784,-0.2309720069,0.154519394,0.1146324128,0.20689632,-0.1153967083,-0.0914429575,0.289586246,-0.1751039475,-0.4062292874,-0.0956724882,0.0261741728,0.342623502,-0.0390423611,-0.2834788263,-0.2238362283,0.3323227763,0.2916031778,0.0975844637,-0.0602427796,0.1861873418,-0.2744443417,-0.2148988992,-0.1029837132,0.1444678307,0.0479049087,-0.1917316318,-0.0940221474,0.234339267,-0.267280519,-0.3892906606,-0.1091276556,-0.0915101543,-0.1745405048,0.3141992688,-0.0561185144,-0.0780313537,-0.1603638381,-0.452439487,0.1582749635,-0.1562495679,0.269826591,0.0558287986,-0.0296710748,-0.177858308,-0.2163670361,0.3298382163,0.0466847606,0.1715169549,0.6110993028,0.2423764467,-0.3313790262,-0.1585514545,-0.1488048583,-0.3503169715,-0.1226206645,0.0027593679,-0.4130464196,0.0118887136,-0.1797307879,0.3767560124,0.0177583471,-0.2951879203,-0.1263607144,-0.3745446503,-0.0287250374,0.2725760937,-0.1091428623,0.0510793999,0.1647903621,-0.0170247741,-0.1053795889,0.1381785125,-0.3056139946,0.0877611712,0.0292149857,0.1146594137,-0.2499316782,-0.0376227461,-0.1758767664,-0.0679048374,0.1059217453,-0.1951123029,0.0084639983,-0.2114462554,-0.2095867991,0.1348883361,-0.0234818533,-0.0472017601,-0.0292450767,-0.1437854022,-0.0996534601,-0.2585893273,0.0201124642,0.0672349706,0.0766697377,0.2918261588,0.3037975729,0.0459492244,0.0538113303,0.0972014219,0.1759545952,0.5515096188,0.0319131315,0.3699150085,-0.0455829389,-0.1020636261,-0.1002447903,0.2322753221,-0.5310224295,-0.1332032979,0.6618318558,-0.0191819761,-0.1777704656,0.5489698648,0.0790044218,-0.0890685022,-0.0443039052,-0.1241976172,-0.0672596917,0.21554856,-0.415181011,-0.1710607409,0.3353061676,0.0469865836,0.1376532465,0.0936234593,0.2512539923,-0.4399034977,0.213267386,0.2694577873,0.088052474,-0.2221907675,0.3708510697,0.217308104,0.1169304922,0.2881016731,-0.0560477823,0.171509698,-0.0184213854,-0.0354767479,0.254760623,0.100226745,0.361926496,0.0141441878,0.2070838362,-0.2588253915,0.3601000905,0.2614317238,-0.074890621,0.1805955917,0.2045027763,-0.4046829343,0.3852643967,0.0291205328,-0.3738873601,0.2746961117,-0.1384236664,-0.2239904255,-0.2610627711,-0.2730075717,0.2544862032,0.0904904827,0.13105084,-0.3175448179,0.2628822029,0.0146097606,-0.1453618258,-0.2687106729,0.6404503584,-0.1838212311,0.3839745522,-0.1375797838,0.0341050737,-0.0687104389,-0.0616256744,0.0359619111,0.0860251114,0.2202849388,-0.141144827,-0.1487487853,0.0484409109,-0.0286176223,-0.1927867383,-0.1696994007,0.0362559855,0.0899391249,-0.2654728591,0.3961621225,0.1518195122,-0.1117652431,-0.538356185,0.2200966924,0.1398579776,0.0246102381,-0.225658074,0.0294108,-0.1924510747,0.3682516515,0.0251918919,-0.033152163,-0.0186040737,0.1754935831,0.0818287581,0.3540491462,-0.1348263472,0.1112554446,0.0014079694,0.2195776254,0.3295511305,-0.302410543,0.0246559456,-0.3585800827,-0.2694030702,-0.1162963733,-0.4224758744,-0.1305844933,0.3918165863,0.356528163,0.1863863021,-0.0947760716,0.2789004445,-0.2075873464,0.0550676435,-0.0140282437,-0.3055956662,0.0359509848,0.1721840054,0.1491940469,-0.2695627809,-0.4008760452,0.3634429276,-0.1732610315,0.0856611505,-0.0778291598,0.0711339042,0.2079639435,0.0433589295,-0.026394194,0.212843433,-0.0716938376,-0.2881034315,-0.322717011,-0.1422344297,0.1642449796,0.0366430692,0.1561230421,-0.3485835791,0.2843595743,-0.0423893407,-0.1827995777,-0.0463385098,0.715236485,-0.0573209599,0.1616110951,-0.2237075567,0.0877657831,-0.0971618518,0.1019454077,0.2723842859,0.1690089554,-0.1890601069,0.1129491031,-0.0942370296,-0.5099195242,0.0974450707,-0.402715385,-0.4580297768,-0.3972794116,-0.0640225187,0.076579757,0.1758395731,-0.246899426,0.0494573116,0.3798077404,-0.2447845638,0.0399638563,0.003836469,-0.0131425038,0.1720331162,-0.088394098,0.2015592009,0.0890823305,-0.4260413051,-0.5251040459,-0.2948700786]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2166","title":"Regarding Test Sets for the GEM datasets","comments":"Hi @vyraun ! The test references for CommonGen are not publicly available: you can reach out to the original dataset authors if you would like to ask for them, but we will not be releasing them as part of GEM (March 31st was the release date for the test set inputs, references are incidentally released for some of the test sets but shouldn't really be used for benchmark submissions)\r\n\r\ncc @sebastiangehrmann","body":"@yjernite Hi, are the test sets for the GEM datasets scheduled to be [added soon](https:\/\/gem-benchmark.com\/shared_task)? \r\n\r\ne.g.\r\n\r\n```\r\nfrom datasets import load_dataset\r\nDATASET_NAME=\"common_gen\"\r\ndata = load_dataset(\"gem\", DATASET_NAME)\r\n```\r\n\r\nThe test set doesn't have the target or references.\r\n\r\n```\r\ndata['test'][0]\r\n{'concept_set_id': 0, 'concepts': ['drill', 'field', 'run', 'team'], 'gem_id': 'common_gen-test-0', 'gem_parent_id': 'common_gen-test-0', 'references': [], 'target': ''}\r\n```\r\n\r\n","comment_length":71,"text":"Regarding Test Sets for the GEM datasets\n@yjernite Hi, are the test sets for the GEM datasets scheduled to be [added soon](https:\/\/gem-benchmark.com\/shared_task)? \r\n\r\ne.g.\r\n\r\n```\r\nfrom datasets import load_dataset\r\nDATASET_NAME=\"common_gen\"\r\ndata = load_dataset(\"gem\", DATASET_NAME)\r\n```\r\n\r\nThe test set doesn't have the target or references.\r\n\r\n```\r\ndata['test'][0]\r\n{'concept_set_id': 0, 'concepts': ['drill', 'field', 'run', 'team'], 'gem_id': 'common_gen-test-0', 'gem_parent_id': 'common_gen-test-0', 'references': [], 'target': ''}\r\n```\r\n\r\n\nHi @vyraun ! The test references for CommonGen are not publicly available: you can reach out to the original dataset authors if you would like to ask for them, but we will not be releasing them as part of GEM (March 31st was the release date for the test set inputs, references are incidentally released for some of the test sets but shouldn't really be used for benchmark submissions)\r\n\r\ncc @sebastiangehrmann","embeddings":[-0.3199131787,-0.0961244851,-0.1973158568,0.1844352782,-0.0919602215,0.118466258,0.2837320268,0.3940259218,-0.1052865312,-0.0161253773,0.1892587394,0.2699712515,-0.3344815075,0.1294022053,-0.0042844992,0.2062070072,0.0137714744,0.0532648787,-0.1366805881,-0.1827852875,-0.0030422248,0.2782809138,-0.1043493971,0.0674216449,-0.4564504027,-0.2105980515,-0.1491971314,0.127491951,-0.0292934906,-0.2702152729,0.3068476915,0.1725850701,-0.1328514069,0.4753216803,-0.0001048454,-0.092475757,0.2723983228,-0.1354853958,-0.2651570439,-0.3887256682,-0.3688611686,-0.0995736867,-0.0375091918,-0.1074555814,-0.1079825461,-0.0469908938,-0.1061154976,-0.4838304222,0.0354261696,0.2401734143,0.2840694189,0.3282464445,-0.0739153102,-0.2256850451,0.1541750729,-0.2887515128,-0.2577032745,0.2372087538,0.0459526889,0.0304730348,0.2049218714,0.1084306687,0.0687660798,0.1535712928,0.1378137618,0.1191932037,-0.0138435252,-0.373249799,-0.1442776918,0.2480443418,0.6280167699,-0.3502284884,-0.481459111,-0.1563065499,-0.2806761265,-0.0516629964,0.023211265,0.2364029884,-0.1203913689,0.1719749123,-0.3285424411,-0.3881807923,0.0009274193,-0.2865280509,0.1206907853,0.0589674823,-0.0700026527,-0.1194067672,-0.0599666685,-0.0071771275,-0.1204124317,-0.1025010645,-0.1917668432,0.0383693352,-0.3621035516,-0.4651756883,-0.175957039,-0.1125507876,0.3332552314,0.4292494655,0.5199062228,0.1588800102,-0.1818336248,0.1502227336,0.3433420062,0.1893132925,0.3401162326,-0.1188675836,0.1506562531,-0.0951440632,-0.1557407379,0.2906402647,-0.0817789659,-0.1565802395,-0.0089309104,0.1021690294,0.0856286511,-0.1682278216,-0.2093770653,0.1630986035,0.2056977898,-0.260371536,-0.1541120708,0.1218300983,-0.0569119677,-0.1605733037,-0.3958009481,0.1040320247,-0.118862398,-0.3285265565,-0.1978198588,0.0562764294,0.0054480582,0.0675543547,0.1492599696,-0.2012851387,0.4375038743,-0.0745779499,0.2316796184,-0.0567061342,0.3254747093,-0.0189853609,0.0689285025,0.2251051366,-0.0693525597,0.2178825587,-0.0323822051,-0.1107460484,-0.3100337088,0.3561433256,-0.129467085,-0.1646644026,0.3178631961,0.2424749583,-0.6210737228,-0.2433426082,-0.3021156192,0.4153852463,-0.0981868654,-0.204107523,-0.0203391463,-0.1193988323,0.068235524,-0.2143627405,0.30456388,0.1103176698,-0.1111432463,-0.0800725073,0.2217071205,-0.0407881364,0.0517827086,0.0540543795,-0.0427649952,-0.2430043966,-0.0572339632,0.173601687,0.3769561052,-0.1589796394,-0.1349399835,-0.0331564583,-0.2171144187,-0.1006632671,0.1498170048,-0.0805038661,0.1672935635,-0.1374400407,-0.1856816113,0.121845156,-0.2028912157,0.0184544958,-0.4238097966,-0.2053615302,0.3437220752,-0.0682144165,0.0103603704,-0.2939474583,0.0078066834,0.0155284181,0.1362079531,0.1302019209,-0.1227962673,0.210458681,0.1868480742,-0.3276837766,0.0944227278,-0.0963011757,-0.199502781,0.3398899734,0.3615368009,0.162085548,0.6675425172,-0.130902335,-0.3988662064,0.1048480421,-0.1167538464,-0.2448287904,0.1276991069,0.0224810317,0.0640118048,-0.0409484543,-0.2648723722,-0.1205672026,-0.2782312632,-0.0588954277,0.1076681316,0.4514991045,-0.1003340185,-0.0928502679,0.1537717283,0.0597490519,-0.059230525,-0.1679751277,-0.044983279,0.2149646729,0.2345026881,0.3967015743,0.5926070809,0.4832264483,0.1302341968,-0.4008479714,0.0054555545,-0.142151922,-0.0121017983,0.0782554522,-0.0040949523,0.5423492193,-0.0047594677,-0.0244002175,0.1844103038,-0.0400171652,0.0160118099,0.022838803,-0.1425108165,0.0788014531,0.006102806,-0.2962937653,-0.1734932065,0.0747587383,-0.1455050558,0.1093975306,0.3203474283,0.0163721777,-0.0421183705,0.0658583194,-0.0669589639,-0.352091074,-0.0452134982,0.2316674739,0.2718405426,0.2748383582,0.5183141828,-0.1675246209,0.0718117133,0.1177208573,0.2747019231,0.1620261371,0.3612320721,0.0641383231,0.0139861517,-0.0303531978,0.0787399039,0.0004979396,0.254319191,-0.0126081267,-0.265535146,-0.1824519783,-0.0529922061,0.4955043197,-0.1487344205,-0.2569100559,-0.2974275351,-0.1817663759,0.1917511225,-0.2066358626,0.118453294,0.1440537721,-0.1022746935,0.1683882326,-0.3601239324,-0.0104556242,-0.362883091,-0.2907291055,-0.0691317543,0.2497561425,0.0202234685,-0.1943970025,0.5272140503,-0.0249321256,-0.0333685093,0.0251066126,-0.3380875587,-0.0206466988,-0.2334231883,0.6151625514,-0.0213523917,0.1114930809,-0.1227341518,-0.0504871979,0.1510163397,-0.367094487,-0.2025052458,-0.4389244616,0.0394201949,0.1674714088,0.0141842114,-0.7287330627,-0.313970834,-0.136432901,0.0277883224,0.078133367,0.0207945723,0.0937430039,-0.0762842596,-0.0136159062,0.2463950664,0.1230628118,-0.1112482697,-0.2469578534,0.0709656775,-0.2565820813,-0.2834366262,0.0082686692,-0.2342488468,0.5369318128,-0.2553298175,-0.4300388098,0.1072039083,0.138503328,0.1700685471,0.3131311834,-0.3208166063,-0.0004081299,0.0782382041,-0.2634627521,-0.2218330652,-0.2145446092,0.0192224532,0.1301267594,0.0001412471,0.2200954109,-0.0504713729,0.1129408255,1.2106736898,0.3273599744,-0.2195055038,0.0636019483,-0.1927153468,0.0751804858,-0.0221974738,-0.1251813173,-0.1512246877,0.0874205679,-0.1855769008,0.1736409366,0.2079459876,-0.046727296,-0.120227471,0.0999011993,-0.2788953483,-0.0751182586,-0.2139841914,-0.3132952154,0.1975212544,0.1157158837,-0.1154130325,-0.0477227755,-0.101353161,-0.0383366346,0.1806686968,0.013311334,-0.1800635904,-0.5068635941,-0.0113041503,-0.0431137271,0.2794252932,0.0795165449,0.1619376391,-0.1619589776,-0.0535285436,-0.0329803079,-0.133048445,-0.0355496705,-0.3102973104,0.0261688884,0.1630448252,-0.2168928981,-0.0081138387,0.0866068006,0.0115358261,0.0566007458,0.4807618558,-0.0489896312,0.0497589782,-0.370719254,0.0235234927,0.0425225161,-0.188998118,0.2285545766,-0.1610232294,-0.1048052683,-0.162328124,0.0350560024,0.0179963801,-0.0748319626,-0.5981101394,-0.1424155384,0.1041496396,0.0443027243,0.4162410498,-0.0220279172,0.275978893,0.2045426071,0.4596998692,0.4485518932,-0.1858811826,0.0823455006,0.3713404238,-0.0798934028,0.3677316308,0.0714714602,-0.1460967362,0.5774689913,0.2619875669,0.0281138085,-0.1851836294,-0.1039742976,-0.288544029,-0.3033469915,0.1477827132,0.1347528696,0.0864683613,-0.2001955211,-0.6305391788,0.1975009888,0.2010252923,-0.3536801338,0.3593759537,-0.0789748207,0.0408884883,0.0714415088,0.0161137655,0.7185249329,0.0311300289,0.0408372208,-0.063367866,-0.0831397548,-0.0582954288,-0.1969101429,-0.0650126413,-0.4512869418,-0.3052611649,0.0646354929,-0.239998579,0.0965367705,0.0461733341,0.0533526614,0.213724941,0.0748995617,-0.0926873386,-0.0279138479,0.1169246063,0.191501677,0.2568719387,0.0779549032,0.1929308027,0.0691705123,0.1767724305,0.1489313692,-0.0896165818,-0.2012271881,-0.3708186448,-0.51761657,-0.0316124335,-0.0510129444,-0.1126781106,0.0256291013,-0.4053792357,0.2161786109,0.5042461157,0.1690028012,-0.0211101677,-0.0940543562,0.2346985489,-0.2069260478,0.1775067747,0.2685363889,0.264398694,0.298704356,-0.1736251712,-0.2950647175,0.2239082158,-0.2379511148,0.1217780113,-0.2031573355,-0.0356719978,0.2882803679,-0.2387649566,0.1726543009,0.0429667048,-0.0536635704,-0.2230569571,0.2215283662,0.1827590168,-0.0591833033,0.148422569,0.2733097374,-0.1622299701,-0.0331391171,0.3201951981,0.1813633144,-0.0981443748,0.4658169746,-0.0653994456,-0.0376856253,-0.3925319612,0.3945834637,0.3197169602,-0.1613172144,0.1343164444,-0.0827924311,0.081957005,0.1526437998,0.557266295,0.3154766262,0.0091164233,-0.1529158056,-0.1201261356,-0.0160592124,0.0721236095,-0.3067631125,0.4167079329,-0.1884607077,-0.5972592831,0.0360752381,0.2838351727,-0.422000736,0.0986538529,-0.1859423071,0.1016609892,0.1157085523,0.0426247567,0.310792625,0.0213165805,0.2126501501,-0.2934315801,-0.1701072752,-0.3000238538,-0.0817709565,0.1047785878,-0.1281003207,0.0412937328,-0.3866678774,0.0008957569,-0.0565720201,-0.0235377308,0.0773107111,-0.1803407967,0.0496796407,0.4335500598,0.1328221112,0.1139926836,-0.0313406624,0.098703526,0.0590467043,0.6497112513,0.0540724508,0.1072622687,0.2401765883,-0.117683433,-0.0970347896,0.1032772958,0.0982732773,-0.1566594243,0.2506431639,-0.1824530512,0.1813058853,0.152157262,0.3487755954,-0.0782396123,-0.2179406583,0.1871303469,0.0157740247,0.2566324472,-0.0791184679,-0.1002468169,0.1507938504,0.0553888753,0.1126978919,0.1759254783,0.5118083358,-0.0121664787,0.1419258714,0.160520792,0.2285176069,-0.2283696085,0.3504428864,-0.1774766445,-0.0784470066,0.039711725,0.1137494072,0.0338494815,0.1301718652,-0.1471461803,0.119816497,0.6880673766,0.3455751836,-0.0218679197,0.002970939,-0.1939599663,0.2410517484,-0.136192739,0.1080638766,0.0413656756,0.0950077698,0.6839952469,-0.0069868783,0.1802154183,-0.2407154143,-0.2542263567,0.04743395,-0.0468393527,0.1862078309,-0.3324164748,-0.2405874729,-0.0999501199,-0.0467075035,0.1115542799,0.2917162776,-0.0582833923,0.1853178889,-0.1413296461,0.3238018751,0.2330872715,-0.0464062281,-0.2948651314,0.225022912,-0.0163244214,-0.0854441151,0.1405863017,0.3934528232,0.371073246,-0.1066667885,-0.3742781281,-0.2025521994,0.0403250679,-0.1668348014,-0.0226200391,0.3400896788,-0.1154897362,-0.3748776317,0.3489947021,0.2770542204,-0.0606157817,-0.0007212026,0.1701216698,-0.1206800044,-0.2029059529,0.540869832,0.0171917956,-0.0640412048,-0.0150463115,0.0213842131,-0.3221749067,-0.2637946308,0.3815341592,-0.0299417209,-0.0720614344,-0.0675551221,0.1355533153,-0.0533799566,0.4942827821,-0.0191160161,0.3656122386,-0.2919708788,-0.1607242078,-0.2264883071,0.0716511011,0.5041439533,0.1339232773,0.1225116104,0.0957553089,0.1908079833,0.1724904329,0.1969130635,-0.2208015621,-0.086912021,0.3794533014,-0.4213513136,-0.0196723007,-0.3491999209,0.1296296567,0.0191059299,0.0250672884,-0.1146988571,-0.2825491428,-0.0121566476,-0.2955873013,-0.2437657863,0.4059177637,0.0561938398,0.4025733173,-0.1456017643,0.8773338795,0.0132251019,-0.1577969342,-0.005294316,-0.4694607556,0.1708764285,-0.0260084011,-0.2699880004,0.0613111295,-0.0630879402,-0.0651198477,-0.0958216935,-0.2657929659,0.0501174293,0.0394811332,-0.0188562069,-0.2808692157,-0.196735397,-0.0119320564,0.2182398736,0.2776321471,-0.059858758,0.1244896948,-0.1524220705,0.0226656906,0.4759706259,0.0715650916,0.0338383168,-0.4323130548,0.2209634781,0.2272646576,-0.0178584773,-0.4435823262,-0.1076604426,0.5402593017,0.1289527267,-0.0759133101,0.197856158,0.3213538229,0.1204367355,-0.0154199433,0.159460187,-0.3088413477,0.073769547,-0.1196222827,-0.05943501]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2166","title":"Regarding Test Sets for the GEM datasets","comments":"Oh okay, thanks @yjernite ! ","body":"@yjernite Hi, are the test sets for the GEM datasets scheduled to be [added soon](https:\/\/gem-benchmark.com\/shared_task)? \r\n\r\ne.g.\r\n\r\n```\r\nfrom datasets import load_dataset\r\nDATASET_NAME=\"common_gen\"\r\ndata = load_dataset(\"gem\", DATASET_NAME)\r\n```\r\n\r\nThe test set doesn't have the target or references.\r\n\r\n```\r\ndata['test'][0]\r\n{'concept_set_id': 0, 'concepts': ['drill', 'field', 'run', 'team'], 'gem_id': 'common_gen-test-0', 'gem_parent_id': 'common_gen-test-0', 'references': [], 'target': ''}\r\n```\r\n\r\n","comment_length":5,"text":"Regarding Test Sets for the GEM datasets\n@yjernite Hi, are the test sets for the GEM datasets scheduled to be [added soon](https:\/\/gem-benchmark.com\/shared_task)? \r\n\r\ne.g.\r\n\r\n```\r\nfrom datasets import load_dataset\r\nDATASET_NAME=\"common_gen\"\r\ndata = load_dataset(\"gem\", DATASET_NAME)\r\n```\r\n\r\nThe test set doesn't have the target or references.\r\n\r\n```\r\ndata['test'][0]\r\n{'concept_set_id': 0, 'concepts': ['drill', 'field', 'run', 'team'], 'gem_id': 'common_gen-test-0', 'gem_parent_id': 'common_gen-test-0', 'references': [], 'target': ''}\r\n```\r\n\r\n\nOh okay, thanks @yjernite ! ","embeddings":[-0.2423841953,-0.1023766324,-0.217669487,0.1744783521,-0.1093867794,0.1701618731,0.3166179955,0.3741603494,-0.1163754463,-0.1348387599,0.0985358581,0.2754116654,-0.3876177073,0.2263205647,0.0858924463,0.178005591,0.0664892569,-0.0329669416,-0.1635664254,-0.1446855366,-0.0818025693,0.3256413043,-0.199252665,-0.024741644,-0.4385986626,-0.2250252366,-0.2253356725,0.1641076505,-0.1128951237,-0.3215231597,0.3283916712,0.179784745,-0.0996069014,0.4712410569,-0.0001076743,-0.0972846746,0.2767040133,-0.1622172296,-0.317045033,-0.3581338823,-0.4203817844,-0.1976664662,0.0584175177,-0.1561119407,-0.1007136926,-0.0338723585,-0.1044883281,-0.41962865,-0.0485681295,0.3183373511,0.2614703476,0.3422113359,-0.168430075,-0.206740573,0.100496076,-0.202301085,-0.2290571034,0.2590492964,-0.0036295655,0.0146770943,0.2504741549,0.108254686,0.0099685304,0.2085843682,0.1962733418,0.0742454901,-0.1285590976,-0.3590754271,-0.2475716174,0.2237913758,0.6646274328,-0.2694216371,-0.4922915697,-0.2300522476,-0.3657217324,-0.0867066905,-0.020350473,0.2318538576,-0.0406990536,0.1278847754,-0.3301391304,-0.3869197667,0.0566709451,-0.1941863,0.0448369905,0.0647994652,-0.0694870949,-0.0819677636,-0.0866768733,-0.0183276609,-0.0812759697,-0.1602267921,-0.2641840279,0.1328713745,-0.3610768318,-0.4818922579,-0.16437006,0.0325490311,0.2423200756,0.5460938215,0.3721685112,0.0975797996,-0.2214623094,0.1585015059,0.3444314301,0.1498176008,0.4101448059,0.0020652504,0.113901481,-0.060280595,-0.0620663539,0.2714732587,-0.1152183712,-0.1931308061,-0.0412456319,0.0606600195,0.1818767637,-0.1293684095,-0.0888859704,0.1507906914,0.2213567048,-0.3887292147,-0.1611794084,0.1389253289,-0.076852195,-0.220299691,-0.4117494524,0.2171884477,-0.108342804,-0.2158867568,-0.1641413867,0.1334959418,0.1112094447,0.0063169217,0.1500688642,-0.2264512181,0.4878134131,-0.1891827285,0.1338797212,-0.1075418815,0.310664624,-0.0525461622,0.1371249557,0.313678205,0.0014869614,0.2313804328,-0.0101169851,-0.0957005844,-0.2816625237,0.4174501896,-0.0844348893,-0.1403403431,0.3260136545,0.2179326713,-0.6480875611,-0.3426914811,-0.4409269989,0.4562060833,-0.0726919174,-0.0866437927,0.0010074677,-0.0760070235,0.0493583381,-0.1768169999,0.2236549705,0.1824904531,-0.1040423438,-0.1220225021,0.2557303607,-0.1101243347,-0.0128679294,0.0272417068,-0.0268896762,-0.0884684771,-0.1480271965,0.1559095234,0.3396210968,-0.0909944698,0.0054599331,-0.0389249064,-0.1280315965,-0.0808430836,0.0731780455,-0.0658615977,0.140271008,-0.1050296873,-0.2023899108,0.0484571904,-0.2042267323,0.0621597245,-0.3554079533,-0.0917549133,0.4161292613,-0.118581295,0.0548699051,-0.2754030526,-0.1292601675,0.0226227306,0.1523156315,0.246056661,-0.1572167277,0.1179024801,0.138050139,-0.2577794194,0.165394187,-0.0543292686,-0.2266522199,0.2926003039,0.4630917907,0.1552581489,0.6165255308,-0.0197397117,-0.4112021625,0.1325315684,-0.1713446677,-0.2885832787,0.0873449147,-0.0380512103,0.0679682568,-0.0011630601,-0.3801616728,-0.0069703828,-0.2620060146,0.0471681505,0.0741082802,0.4704503715,-0.032580521,-0.0400076173,0.0903007686,0.003024162,0.0007513227,-0.2652706504,-0.0595927909,0.1209909543,0.3538026512,0.3784921169,0.5874525309,0.4715746045,0.1596785784,-0.3193691373,-0.0531943366,-0.1695429534,0.0464828871,0.0333368704,0.0624672286,0.5468242168,-0.057462696,0.0292226151,0.2033577412,-0.0329630673,0.0773953721,0.0111302352,-0.1555080563,0.0791121945,0.0931912139,-0.2571781576,-0.1899287403,0.1378961951,-0.0793496892,-0.0108187459,0.2884023488,-0.0489106812,-0.0768434852,0.0511127934,-0.0774789527,-0.3378792703,-0.1238125265,0.1757437289,0.2695752978,0.2081926465,0.4295229912,-0.1366821975,0.0922183394,0.0615546517,0.3187260628,0.208220467,0.2941581607,-0.0143210189,-0.0979151726,-0.1275053024,0.0084085511,-0.0256667845,0.2870330811,0.0690075085,-0.3527412713,-0.2411694974,0.0353988297,0.4352717996,-0.1880109012,-0.3663303256,-0.2434812784,-0.2509459555,0.1398120672,-0.2313388586,0.0890713632,0.1112853661,-0.1346618086,0.1627374887,-0.3668240309,-0.0418982692,-0.3192468584,-0.2659401298,-0.0115655065,0.2661609054,0.0108366385,-0.2414150238,0.482952863,-0.1147330701,-0.0997469276,0.0392699093,-0.2686074972,-0.002009474,-0.2684287727,0.6333341002,-0.0489863046,0.1413705498,-0.0494341552,0.0311500188,0.1084482893,-0.2383004278,-0.1952569783,-0.4336464703,0.0728639662,0.1691495925,0.1271078289,-0.7435390949,-0.3394576907,-0.0765550062,0.0162114743,0.1574529558,-0.0002869496,0.0942920744,-0.1012117863,-0.0215031393,0.167936638,0.0154978875,-0.0399794057,-0.3016294539,0.0708208978,-0.2649869919,-0.2239613831,0.013753986,-0.2723499238,0.4887934029,-0.3745372891,-0.389136225,0.0822196379,0.0977241695,0.1549926251,0.3050180376,-0.2843656838,-0.0118690552,0.0009402948,-0.1852923632,-0.1644885391,-0.2195757627,0.087600179,0.1418282688,-0.1675267965,0.2256317586,-0.0641520768,0.0036173768,1.1235276461,0.3761463165,-0.0976822078,0.090744473,-0.1921256185,0.0980248377,0.0176402219,-0.1276723146,-0.0904772356,0.1699707061,-0.2317606956,0.2355778217,0.1805891842,-0.0781990737,-0.1508647799,0.0354186334,-0.3323262632,0.0385095477,-0.1480282545,-0.449986726,0.1340408176,0.0916982889,-0.1472626328,-0.0099298796,-0.0776266977,-0.118315272,0.2561381161,0.0488387495,-0.2107637376,-0.4824265838,-0.0811571255,-0.1126161292,0.2567873597,0.0278465152,0.1567437351,-0.0780244321,-0.0672651455,-0.1326884031,-0.0545878597,0.0095183356,-0.2564626038,0.0011359822,0.1716350466,-0.2818217874,-0.0177372545,0.0534250885,-0.0037381577,0.0549856536,0.410864383,0.0343624242,0.2062767148,-0.292733103,0.1214742064,0.0698959604,-0.2000947297,0.239045769,-0.1397431493,-0.1538927406,-0.1010514051,0.0578954555,0.0962918252,-0.106695734,-0.5020112395,-0.2167346478,0.0401528217,0.0125346594,0.3897722065,0.0197634939,0.2125525922,0.1928144097,0.4744117558,0.4390246868,-0.2473863065,0.1152563766,0.4609874785,-0.0942766964,0.3276560605,0.0299793892,-0.1290735751,0.7060554028,0.1394203454,0.0100270966,-0.1630588025,-0.1691524982,-0.2590617836,-0.3880655468,0.0559254661,0.2126480639,0.1318241209,-0.2346412987,-0.6614342332,0.1600463092,0.1458366662,-0.267487973,0.334718883,-0.1563024521,0.056931708,0.0734440014,-0.0784479678,0.587878406,-0.0263324697,0.0052569304,-0.0358346216,-0.1176423281,0.0162486229,-0.2876574099,-0.1037281379,-0.4601440728,-0.2651166916,0.0152361747,-0.2826154828,0.1203997135,0.0913908705,0.0287181642,0.2308914363,0.0541749112,-0.0432442427,-0.0623367131,0.1452530175,0.2839135528,0.2635532618,0.0259858891,0.1732038707,0.0787612051,0.1829503477,0.0577129982,0.0713508055,-0.1932549626,-0.3488192558,-0.5013494492,-0.0975398347,-0.0426052399,-0.0877976865,0.0013205342,-0.3374206722,0.1223287806,0.5082534552,0.0892930478,-0.0395386219,-0.0614010207,0.2257808894,-0.2055913359,0.1667618304,0.2053854913,0.286871314,0.3576684892,-0.1740204543,-0.231291607,0.1559573412,-0.29163149,0.1534805596,-0.1931941807,0.0883589685,0.3014695048,-0.3021629453,0.0503396876,0.0827376992,0.0293787122,-0.2278704345,0.1918619871,0.1991957128,-0.1003631353,0.1996196061,0.2926394641,-0.2038927376,-0.0734696463,0.2191781998,0.2075717449,-0.0697601363,0.5267481208,-0.0641176403,-0.0210035164,-0.3711678386,0.4586166441,0.3629185855,-0.0935426131,0.2044804245,0.048518762,0.2352854162,0.1266881078,0.5553035736,0.2434555143,-0.0026014566,-0.216337502,-0.1254675388,-0.0033486513,0.0282716621,-0.2517240942,0.4081941843,-0.124096483,-0.4919240773,0.1638921052,0.3203667104,-0.3693027794,0.1198542044,-0.1860360056,0.1726948023,0.1175436154,0.0809121951,0.2862860262,-0.0302093942,0.1970680505,-0.3148397505,-0.1181829125,-0.3064861596,-0.1270081848,0.124756068,-0.0879337043,0.01665828,-0.3571140468,0.0314227231,-0.0493953004,-0.0327785276,-0.0108839348,-0.1247078106,0.0140176183,0.4455219209,0.2648967206,-0.0765834153,-0.0381363221,0.1012413874,0.0643522814,0.5857334137,0.0588346012,0.0585767329,0.2972370684,-0.0872379169,-0.0453938283,0.0534144193,0.0051430301,-0.1502200961,0.14338018,-0.2270802706,0.242521584,0.0567022189,0.39082551,0.0116552003,-0.1269391328,0.1307407618,0.0467529632,0.2341223657,-0.1462973207,-0.1169025898,0.228650257,0.052427724,0.1046299711,0.2123345584,0.4200643003,0.1374329478,0.0808187425,0.117830053,0.2334485948,-0.2122330517,0.4198739827,-0.2058554292,-0.0251762047,0.0355252586,0.1302659959,-0.0288727451,0.1566043049,-0.1507196724,0.1293049753,0.7019695044,0.3885045052,-0.0194920357,0.0467907041,-0.2703139186,0.2231947035,-0.1742964536,0.0343311653,0.0480217487,0.0725196823,0.6498006582,-0.0155960862,0.200154677,-0.2073507607,-0.2209618241,0.0398434289,-0.0793305114,0.284883678,-0.3558932543,-0.2320420444,-0.1094711572,-0.029965952,-0.0086153662,0.3335649669,-0.0424160399,0.1917505115,-0.2016485333,0.4628675878,0.1446469277,-0.1250423044,-0.2848543227,0.2009953558,-0.0792948455,-0.0879373625,0.1058873311,0.4516940713,0.3615199924,-0.0822006464,-0.2585630417,-0.2058700323,0.0182868056,-0.1388419569,-0.0225006118,0.3705675304,-0.1667077541,-0.2863453627,0.2888634503,0.2660790682,-0.0304039493,0.1066838875,0.2693699002,-0.2611867189,-0.1910546124,0.5365919471,0.0436533429,-0.0168486834,-0.0715954006,0.0421487726,-0.3554580212,-0.2933987677,0.3491539359,-0.0768215582,-0.0885860622,-0.0018137464,0.1164170057,-0.0856504068,0.5615434647,0.0866252258,0.452311337,-0.3718660176,-0.1578377932,-0.1196479574,-0.022114601,0.4476617277,0.1438634247,0.1038672328,0.0712988675,0.2078461498,0.1950039268,0.1720747054,-0.1557200402,-0.0506509207,0.3701468706,-0.5040840507,-0.0466757901,-0.4101790488,0.2122748196,0.0429467745,0.0920756161,-0.1230631173,-0.274817735,-0.0783856213,-0.2575726807,-0.281874299,0.464350611,-0.0355455428,0.4087870419,-0.1938895285,0.9285750985,-0.0025003387,-0.0546292327,0.0316271223,-0.4337864816,0.0981931165,0.0532942563,-0.2437218279,0.0943811014,-0.1235307306,-0.0095839212,-0.0401926301,-0.2664000988,-0.0313069299,0.0755785629,0.0582112484,-0.2889856398,-0.1217343435,-0.0523798987,0.1828547716,0.2885090113,-0.0836410001,0.1550344974,-0.1540772021,0.0365874805,0.4523286521,0.083438985,0.096148923,-0.3872427642,0.1266459674,0.290073514,-0.0037518565,-0.4590023458,-0.0597380362,0.6511249542,0.112332873,-0.0011547854,0.2160808295,0.3314760029,0.0996291265,-0.0601890683,0.1595898718,-0.3141887486,0.0967245772,-0.1521959603,-0.1008653417]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2165","title":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset","comments":"Hi,\r\n\r\na HF dataset can be converted to a Torch Dataset with a simple wrapper as follows:\r\n```python\r\nfrom torch.utils.data import Dataset\r\n \r\nclass HFDataset(Dataset):\r\n def __init__(self, dset):\r\n self.dset = dset\r\n\r\n def __getitem__(self, idx):\r\n return self.dset[idx]\r\n\r\n def __len__(self):\r\n return len(self.dset)\r\n\r\ntrain_ds = HFDataset(train_ds)\r\n```\r\n@lhoestq Since the Arrow Dataset already provides `__getitem__` and `__len__`, I think we could use the [virtual subclass](https:\/\/docs.python.org\/3\/library\/abc.html#abc.ABCMeta.register) mechanism from the `abc` module to elegantly solve this issue. This mechanism would allow the Arrow Dataset to be used in place of the Torch Dataset because the `isinstance(instance of Arrow Dataset, TorchDataset)` check would return True (DeepSpeed has this check [here](https:\/\/github.com\/microsoft\/DeepSpeed\/blob\/ab5534fc4c0f8ca21ada321f9730d723aa31288b\/deepspeed\/runtime\/engine.py#L823)).\r\n\r\nAnd it requires a minimal change in the `arrow_dataset.py` file:\r\n```python\r\nif config.TORCH_AVAILABLE:\r\n from torch.utils.data import Dataset as TorchDataset\r\n TorchDataset.register(Dataset)\r\n```","body":"Hi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n","comment_length":124,"text":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset\nHi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n\nHi,\r\n\r\na HF dataset can be converted to a Torch Dataset with a simple wrapper as follows:\r\n```python\r\nfrom torch.utils.data import Dataset\r\n \r\nclass HFDataset(Dataset):\r\n def __init__(self, dset):\r\n self.dset = dset\r\n\r\n def __getitem__(self, idx):\r\n return self.dset[idx]\r\n\r\n def __len__(self):\r\n return len(self.dset)\r\n\r\ntrain_ds = HFDataset(train_ds)\r\n```\r\n@lhoestq Since the Arrow Dataset already provides `__getitem__` and `__len__`, I think we could use the [virtual subclass](https:\/\/docs.python.org\/3\/library\/abc.html#abc.ABCMeta.register) mechanism from the `abc` module to elegantly solve this issue. This mechanism would allow the Arrow Dataset to be used in place of the Torch Dataset because the `isinstance(instance of Arrow Dataset, TorchDataset)` check would return True (DeepSpeed has this check [here](https:\/\/github.com\/microsoft\/DeepSpeed\/blob\/ab5534fc4c0f8ca21ada321f9730d723aa31288b\/deepspeed\/runtime\/engine.py#L823)).\r\n\r\nAnd it requires a minimal change in the `arrow_dataset.py` file:\r\n```python\r\nif config.TORCH_AVAILABLE:\r\n from torch.utils.data import Dataset as TorchDataset\r\n TorchDataset.register(Dataset)\r\n```","embeddings":[-0.2578476369,-0.2652207911,0.092290394,0.3382236362,0.1922441572,0.2227143943,-0.0302729923,0.3457017541,-0.1513580829,-0.1886630952,-0.3486473262,0.3248988092,-0.1942691356,-0.1420536041,0.1868406981,-0.2680391371,0.1673274487,-0.0773859099,-0.2466962934,-0.1544105113,-0.1386658102,-0.1202884912,0.0764644146,0.0747125149,0.0671005175,0.0934039727,0.0947778746,0.1692148447,-0.0237060338,-0.3378914595,0.2977361679,0.0648861304,0.5751037598,0.2561476827,-0.0001273232,0.276031971,0.0935578793,-0.0755770281,0.0312795937,-0.0670314357,-0.0523281954,-0.0347759612,0.1930178553,-0.1016476229,-0.3808221221,-0.5978340507,-0.1727228761,-0.5317306519,0.0468852669,0.4796395302,0.0522026569,0.2359526902,-0.0159182791,0.0726409778,0.2887389064,-0.0978339612,-0.1821518838,0.1489152312,0.0003680434,0.1105434671,-0.1440718323,0.1899665147,-0.1628380269,-0.0048651923,0.4622430801,-0.1983766258,-0.3828516006,-0.1405878067,-0.111037761,0.4482490122,0.178531304,-0.3304377496,-0.1622341275,0.0913779661,-0.0402276292,-0.1471044719,-0.3329949081,0.1591297686,0.0133263832,0.1457206905,0.0341933519,-0.0348534435,-0.1720402837,0.0566778407,0.1703047007,0.2567848563,0.1148063466,0.2005096525,0.1122348979,-0.3005607724,0.4763327241,-0.2886344492,0.0199535806,-0.1253504753,-0.1023310423,-0.0023473741,-0.3257407248,-0.0770284161,0.4265949726,0.1055244431,0.3500558436,0.3115786612,-0.1508314162,0.0488586538,-0.0783415884,-0.1055723429,0.1002681479,0.2584201097,-0.0067850319,-0.2417301983,0.0126299374,0.2741792202,-0.0550226197,-0.3046224415,0.0989658833,-0.0002228799,-0.1225692928,-0.1470832229,-0.2946826518,-0.0963819325,-0.3344509602,0.0045026164,0.1658029407,0.1950969696,-0.0069162836,0.1559909135,0.4160038233,0.5736421943,-0.2732748389,-0.0442828089,0.0392627008,0.1037559137,-0.0204075892,-0.2133430839,-0.0619503036,-0.1895148009,0.0589849092,-0.0272746123,-0.1381994635,0.2087249905,0.4373679757,-0.1368232667,0.2677766681,0.0341860503,0.0437844358,0.010727806,0.3737203777,0.4464449286,-0.2962799966,0.2520726621,-0.3542585075,-0.2294124216,-0.029993346,0.0291095488,0.1518912315,-0.3381064236,-0.2596776485,0.3441300392,-0.0471921749,0.0570967868,0.0422882177,-0.4522698224,-0.3920982778,-0.1631955504,0.2866369784,0.2642970383,-0.4543376863,-0.1082593873,0.0669646636,-0.1349182129,0.1149821132,0.2226646841,-0.2385523468,-0.0212371852,-0.3799104691,-0.1190298796,0.5021675825,-0.3953184485,0.050411053,0.0612166077,0.137086466,-0.0883586109,0.0613233782,0.0608317517,-0.1293748021,-0.1211572438,0.177742213,0.1931716949,-0.1514971703,-0.2278325409,-0.1285722256,-0.1981902272,0.3369712532,-0.0489436053,0.2017852515,0.2132306546,-0.1483509839,-0.3161423504,0.2712277472,-0.0189265776,0.1627959609,0.2168645859,-0.2954394817,0.1737281829,-0.051365871,0.1118234918,0.0711634085,0.130041793,0.0647766069,0.0738745108,0.1080077663,0.0362750329,-0.1997978538,0.1803220958,-0.113247104,0.0170497075,-0.0658317581,-0.0846983716,-0.1495205015,-0.2479619533,-0.0476644002,0.2230951041,-0.4051263928,0.1576890349,-0.5375214219,0.2642347515,0.1351727545,0.1516149044,-0.0461898148,0.2652033567,0.0887988359,0.0306286942,-0.1722036302,0.2192764282,0.0369970165,-0.0704309493,0.2098776251,0.2918401062,0.1083689183,-0.9760195017,0.4055995643,0.4391723871,0.1616518945,-0.0263196845,-0.0654944032,0.737255156,0.2774629295,0.3000271916,-0.0249516424,0.013337099,0.0203084964,0.1796132028,-0.2465766966,0.0484425835,-0.1590321511,0.064657785,0.335036546,-0.067366749,-0.260027945,0.072083801,-0.0637842268,-0.0697551519,0.0892389715,0.2696089745,-0.5429385304,-0.0200525057,0.2840873897,-0.1175016835,0.2478155345,0.1257991046,0.2453291118,0.0192358196,-0.1548355669,0.116677314,0.3136957884,0.1029244885,0.3733651638,-0.0712378472,0.0262523629,-0.060589917,-0.2755173743,-0.0274274237,-0.1287209988,-0.1086189672,-0.4903784692,0.2309375554,-0.3518230617,0.0477069952,-0.3123164177,-0.1259673834,0.0694135576,-0.1767664701,-0.1042157784,-0.0295040123,0.1408050656,0.1130039543,-0.415553093,-0.2949005365,0.1695128828,-0.414301157,0.0517658107,-0.1680516899,-0.1829601973,-0.0821089074,0.1017923206,-0.0141242957,0.3234539926,0.2352633625,-0.0364735238,-0.4555365443,-0.1237668544,0.0857319236,-0.1682258546,0.1182565764,0.166331172,0.0459988341,0.2455699295,0.0315160081,0.0844996497,0.0641371161,-0.0313020274,0.0945117623,-0.4060159624,0.1735185236,0.3860553205,-0.1754043847,-0.3225271702,-0.2034178078,-0.4023958445,0.0576415658,0.0536065735,-0.0176006444,0.3570226431,0.1686804444,0.2799854577,-0.0048589534,-0.1254453808,-0.0701352805,0.3763884008,-0.0555457734,-0.2689588964,-0.1432141811,-0.3228366673,0.2073334157,0.3074100316,-0.3161610961,0.0621641614,-0.2311657071,0.3600255251,0.0136123896,-0.0250633042,0.3019002378,-0.1703789532,0.1023959517,-0.1909873933,-0.2149832547,-0.0230281558,-0.2790466547,0.295226723,0.3544196188,0.4846105278,-0.1506832093,1.0412967205,-0.0465684608,-0.3697224557,0.3973759711,-0.1264249682,0.1498175263,-0.1833273023,-0.2263837755,-0.1105482951,-0.2265174389,0.2802512944,0.0799794495,-0.1003458723,-0.1705220342,0.0693363398,0.1214482859,-0.0306395572,-0.1660863906,0.232955277,-0.8961458802,0.0176867079,0.0021336996,0.1300476044,-0.3229356408,0.0231794436,0.0732279941,0.2152174264,0.1483139843,0.1805096418,-0.3862232566,0.205311358,-0.237146914,0.4286137521,0.0562744103,0.4741934836,0.1464484036,-0.1573108286,-0.0251614973,-0.0147807105,0.2404384911,-0.1279577911,-0.0947264358,0.0771026537,-0.0921520293,-0.5414339304,-0.0246009305,0.0846695304,-0.1194922253,-0.1428970546,-0.0692430884,-0.1483660936,-0.3878341913,0.0475911237,-0.0577733852,-0.0982008204,0.3453828394,-0.0070090569,-0.2675894797,0.0000853262,-0.0048480048,0.112253204,-0.0689446554,-0.1097260937,-0.2179275751,-0.4557943344,-0.0097421855,0.2941734195,-0.200243488,0.2585789561,-0.1388330311,0.0753416792,-0.1527341753,0.2849926054,0.2438029498,0.4379942417,-0.1970749497,-0.2902020812,0.0096696522,0.3952714503,0.3297528028,0.1525477469,-0.1225221083,0.2307644933,0.0732720643,-0.2015485018,-0.3593078256,0.0972255245,0.2285035104,0.0436530299,-0.3781014979,-0.5780895352,0.7419283986,0.2688182294,0.1454413682,0.1953775883,0.3553934395,-0.3011116683,0.6170405746,0.3395324051,0.9867911935,-0.0691077113,0.3258782029,0.0542070791,-0.2959991693,0.0721049011,-0.0504346639,-0.0582276843,-0.4126892686,-0.1477094591,-0.1338684112,-0.162208885,0.0057246727,0.0689140409,0.3153631687,0.1780281067,-0.0744580105,-0.0383907743,0.0962402672,0.07236889,-0.179056704,-0.2455155253,0.0499621592,0.1340464801,-0.1460223049,0.0349265821,0.0267107952,-0.2325405926,-0.2148473412,-0.0474874638,-0.0926460177,0.0968861729,-0.4661181867,0.1713736653,-0.1045225412,-0.7122539878,0.3375083506,0.2854370773,0.3511965275,-0.1926299483,-0.1350196004,0.1517783552,0.3961049616,-0.0373029187,0.0267662629,-0.2233536094,0.4106802046,0.1068917438,0.0353528671,0.0175526235,-0.0071006259,-0.3052248359,-0.3145102262,-0.1245238855,0.5930759311,-0.2820001245,-0.3719059825,0.2380301058,-0.0927092135,-0.0675932318,0.0090437345,0.0696061403,-0.110625118,-0.0107094133,0.1345484555,-0.302125752,0.0600137897,0.3590584993,0.2400671244,0.3162533343,0.3917658031,0.1692462862,-0.0556959063,-0.0065321997,0.2975601852,1.0077922344,-0.2458742261,-0.0099407826,-0.1170474365,-0.16097112,0.1037977487,-0.163428396,0.0081031797,-0.0553746447,-0.1087672338,-0.1185110211,-0.4555997849,0.074856773,-0.4061071575,0.2491332144,-0.1225532144,0.0172263179,-0.3211313784,0.0155149829,-0.1448336393,0.1127302647,0.2651410699,0.1370833963,-0.1197657213,0.1266102493,0.0069290418,0.1773068458,0.0005030292,0.0791052952,-0.1828219891,-0.1051900312,-0.2731465101,0.1961015165,0.0683137029,-0.1551606655,0.0994860828,-0.2853499651,-0.2229064405,-0.0364955515,0.1683729887,0.206748873,0.0059138825,0.2294740081,0.3034988046,-0.0259202421,-0.2105904669,0.0383487381,0.3125870228,0.394697994,-0.0605469644,0.0965473056,-0.1757992506,-0.2931242287,-0.3873532116,0.175218299,0.4720820785,0.0273497552,0.231171906,-0.0398828909,0.1024805009,-0.2299898714,0.2298423052,0.3792033494,-0.0611143634,0.0471935533,0.0231125634,-0.0648273826,-0.3075178564,-0.0568648688,-0.3594819605,-0.4126648307,0.0003112403,0.1142260581,0.0826150998,0.6105093956,-0.1618748456,-0.2463416755,0.0081118178,-0.2925339341,-0.0741097331,0.4579829872,0.2106727362,0.0325755589,0.3636946082,0.091152437,0.3156626225,0.256274879,0.3217669129,0.9033663869,0.5328700542,0.0519999862,0.4340225756,-0.3633095026,-0.2659887373,0.234275952,0.1408527493,0.2021415234,0.1176667586,0.4040745497,-0.0051704934,-0.4713328779,0.0642769039,0.0514754988,0.0454387479,-0.381562978,-0.1736598462,0.0116795227,-0.3381462395,-0.1598154455,-0.3009896874,-0.2025426328,-0.0273909885,0.1743180901,-0.0134382686,-0.1040218771,0.0547183603,0.2319033444,-0.1501416415,-0.04295481,0.0767042488,-0.0977955014,-0.0906883553,0.0501100756,0.2382849604,0.179313466,0.0387448892,-0.3101733923,-0.0302372128,0.5008476377,-0.1578499079,0.2633668482,-0.1167042628,0.0330316834,-0.0895635337,0.3350349367,-0.0045014899,0.0456661023,-0.0018747948,-0.0360819772,-0.0118333492,-0.0806547925,0.4686767161,0.0381002016,0.0139767174,-0.1736445725,0.0824355558,-0.2970457077,-0.144772023,0.6065617204,-0.1959421039,0.1275307834,0.0394565836,-0.0156108588,-0.0559972785,0.2712171972,0.2909684777,0.0128564928,-0.4178183973,0.0564243123,-0.2842056751,-0.060048271,-0.0816407651,0.0632245988,-0.0166597348,0.005641439,0.3884036243,0.0199822523,-0.0046376185,-0.0372499712,0.2183652669,0.4570036829,0.0243278556,-0.2359934151,-0.1582070887,0.3439809084,-0.0470928065,-0.3401724696,-0.183438167,0.2665222883,-0.0988293886,0.041692391,-0.2853511572,0.5871304274,-0.0558396839,0.3612969518,0.0134541765,0.4425632954,-0.1935126334,0.0562472902,0.0560615063,-0.0514196157,-0.4351491928,0.1867741793,-0.1683947146,0.3249020576,-0.0179732256,-0.2146579325,-0.0555233955,0.1130816638,-0.0192127246,-0.483169347,-0.3689863086,0.5198543668,-0.33630988,0.0545218214,0.2158741653,0.5405632257,0.1599622369,0.2815880775,-0.2493225336,-0.1977889538,0.5602641702,0.0879465565,-0.0127672628,0.1000927985,-0.0909548551,0.2509343624,0.0842080563,-0.304174602,-0.1044752821,0.1396904141,-0.2259104401,0.0817541331,0.13318941,0.2929913998,0.2363059074,-0.3110005856,0.4001900554,0.1546282619,0.0743895695,-0.1339149028,-0.2898061872]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2165","title":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset","comments":"Interesting ! Thanks for sharing this @mariosasko . I like the idea\r\nThis looks like something we should add IMO","body":"Hi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n","comment_length":20,"text":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset\nHi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n\nInteresting ! Thanks for sharing this @mariosasko . I like the idea\r\nThis looks like something we should add IMO","embeddings":[-0.2704315186,-0.3552313447,0.0578338169,0.3274024427,0.1308025718,0.2287478447,-0.0746707842,0.375287354,-0.1519821286,-0.2421408743,-0.3885174096,0.3905744851,-0.2419558316,0.0159070753,0.1749463826,-0.30655393,0.2215385139,0.0879765153,-0.2604885101,-0.1844500154,-0.1150370091,-0.0907219425,0.062386252,0.1066241488,0.0754081309,0.0218765084,0.0603702664,0.136170283,-0.0912094861,-0.3727336824,0.1620187461,0.0805812404,0.6023117304,0.2542231381,-0.000129465,0.1214095578,0.0471356772,-0.1353145391,0.1333484501,-0.064884387,-0.1909479499,-0.0160525888,0.1982826144,-0.1445407718,-0.3125005066,-0.5852587819,-0.0864551291,-0.3302887976,0.1317971796,0.5674702525,0.0448211171,0.2532970011,0.0411301218,0.0413687751,0.2709915936,-0.1340613514,-0.1894199848,0.051447209,-0.0449580252,0.1564192921,-0.1868779957,0.1760749966,-0.1330002695,0.026119912,0.3555385172,-0.3080777824,-0.3637715876,-0.1833691448,-0.0535963401,0.5436592698,0.2123097479,-0.2470743954,-0.0536052287,0.1764544398,-0.0406476855,-0.1832195371,-0.4047416747,0.2438120395,-0.034479212,0.1389831454,-0.0193724483,-0.2209188342,-0.1599256545,0.0496588573,0.2337450832,0.3325103819,0.0863249451,0.172097981,0.219557181,-0.3176527023,0.4350479543,-0.2300881743,0.0136854481,-0.1905343384,-0.0844793469,-0.0536000393,-0.367852509,-0.0870001391,0.4450832009,0.0914297029,0.1940545291,0.3110405505,-0.1877269894,0.0876536444,-0.1655006707,-0.2114872187,0.1320845187,0.2368789464,0.0151883224,-0.2638655305,-0.0350546204,0.2685378194,-0.1065920591,-0.4033831954,0.0268643498,0.0132904379,-0.1553608179,-0.1946214139,-0.178907603,-0.2201947719,-0.2744254768,-0.0415353402,0.0255338755,0.2432051748,-0.0197861511,0.0324828252,0.245116666,0.6499621868,-0.1770289391,-0.1200614497,0.023189323,0.1186215729,-0.0104452195,-0.178273648,-0.090177238,-0.1836858094,0.0831105039,-0.0311975162,-0.2837878168,0.1042146087,0.3755451441,-0.1094037741,0.2886730433,0.0350889266,0.1142901108,0.0078265797,0.278511256,0.411947608,-0.3310035467,0.2717525959,-0.5120928288,-0.1923474967,-0.1678929627,0.0259820148,0.1766342074,-0.3390814066,-0.4178642333,0.3362242877,0.0006454258,0.1393365562,0.0111863548,-0.4340932369,-0.3783961535,-0.1909058392,0.1955821961,0.261905849,-0.3799597919,-0.1437560171,0.1453683078,-0.1773046702,0.2109769136,0.3144453764,-0.2082760036,0.0645736456,-0.4396536946,-0.0454900004,0.6284146905,-0.3011912107,0.1494977623,0.1173237562,0.1322222054,-0.2706713676,0.116647616,-0.0181617606,-0.078445673,-0.0447085649,0.1449571401,0.2285384387,-0.13093701,-0.1571481675,-0.1241616905,-0.307241261,0.486638695,-0.0557226315,0.2174060345,0.2946627736,-0.1126179099,-0.2794396281,0.3269390762,-0.068717517,0.1383547783,0.2432400733,-0.3034929335,0.0261795148,-0.0794050321,0.1094832942,0.1647337675,-0.0427010991,0.0162464995,0.0033794018,0.0795040429,0.0811147019,-0.2249483317,0.1649547964,-0.1156474873,0.0749071389,-0.081822224,-0.0752734914,-0.1724797487,-0.1560641527,-0.0423299819,0.2044191957,-0.3709309101,0.1121349409,-0.5357997417,0.2550240457,0.0991600826,0.1160597727,0.0869624913,0.2077864558,0.0999309719,0.0034452002,-0.131325379,0.1819621921,0.0038002152,0.0547716357,0.3907215595,0.3828898668,0.0789416879,-1.0408308506,0.374701798,0.4230888784,0.1125190854,-0.1140040681,-0.0838940293,0.7048979998,0.198375091,0.2045114785,-0.0233915877,-0.0807897225,0.0893388912,0.1275876164,-0.2039730698,0.0450539328,-0.0958149061,-0.0324646607,0.2899934947,-0.0298492406,-0.3910634518,-0.0265778545,0.0611635745,0.0086546754,0.09781041,0.3565205038,-0.5009670854,-0.1333936006,0.2598287165,-0.326651305,0.1245565265,0.1496166587,0.166812256,0.0315002911,-0.0519470535,0.0357631817,0.2676283121,-0.0245042816,0.4564778209,0.0383480191,0.0988226682,-0.0827441067,-0.3403329551,0.0135738216,-0.1481552124,-0.0426874049,-0.4770848453,0.1296864003,-0.3740543127,-0.1077135056,-0.3232433498,0.016059909,0.0922948197,-0.0635075942,-0.1585211158,-0.0373835191,0.2656473517,0.1264510155,-0.4192354977,-0.2399662435,0.2190296352,-0.262365669,-0.0459454991,-0.1887426227,-0.079747051,-0.082025215,0.0914017707,0.0161286127,0.2941834629,0.1412696093,0.1444798708,-0.4242689312,-0.1420470774,0.002584188,-0.0993426144,0.0697790384,0.159506321,0.1411956102,0.2244947702,0.0567411259,0.04318434,0.0411016792,-0.040708743,0.1077585667,-0.4161676168,0.1413841844,0.4955780804,-0.1301068515,-0.3275340199,-0.2123824954,-0.3685065806,0.0493371673,0.1444416344,-0.1609321982,0.3607040048,0.1027723551,0.1735191792,-0.0248907655,-0.0694721416,-0.0942184925,0.3464500308,-0.077798374,-0.2930299938,-0.1745948941,-0.2663951218,0.1151683778,0.1990528703,-0.344843328,0.1793664396,-0.1567104161,0.3103542924,-0.0030000366,-0.0515287928,0.3110346496,-0.1472959667,0.1452459246,-0.1825834811,-0.2098716795,-0.0008980338,-0.1738931835,0.3357551396,0.4362910092,0.456735611,-0.2110413313,0.9687743783,-0.0148856388,-0.3842717409,0.5177223682,-0.1633149981,0.1034211069,-0.191820398,-0.2534873188,-0.0386898778,-0.0802293047,0.3667734265,0.0974166319,-0.1248238832,-0.1807949841,0.0865008384,0.1067031175,-0.1053733006,-0.0593368076,0.2345705032,-0.8194179535,-0.0726820379,0.0025314011,-0.006836833,-0.243728444,-0.0556770936,0.0777246729,0.3136205375,0.2601485252,0.2205914855,-0.3707065284,0.1837079823,-0.2120569646,0.3850882947,-0.0528433397,0.4841068387,0.1041782573,-0.0641131252,0.0318519622,-0.0306218918,0.3866941631,-0.1875582039,-0.0914019644,0.0194156561,-0.0259632897,-0.44393453,-0.0057255761,0.04131658,-0.0433616117,-0.1164484322,0.1831144392,-0.1546757817,-0.359623462,-0.0006945958,0.0029350657,-0.1138090938,0.2737464011,0.0768160075,-0.2045682669,0.0357518196,0.0536901876,-0.0164440665,-0.099000074,-0.1160668954,-0.2233144343,-0.4766241014,0.0645204186,0.2487376779,-0.2038895339,0.1920407563,-0.07865908,-0.0404606424,-0.2650974691,0.2962009311,0.2615866661,0.274735868,-0.269529134,-0.1795158833,-0.012042189,0.3011478186,0.3822052479,0.0853849053,-0.0686252266,0.2815308571,0.0628931895,-0.0958307087,-0.337895453,0.1491992325,0.290302664,-0.0034239993,-0.5377708673,-0.4617139995,0.7115878463,0.2423554808,0.1123874485,0.2173785865,0.2269866914,-0.2727288604,0.7211720943,0.2216558754,1.1427094936,-0.1634585857,0.3277299404,-0.0689024776,-0.3151945472,0.1440817714,-0.1366242468,-0.0438573807,-0.5092170238,-0.1337456256,-0.1724207848,-0.2102554739,0.0005089678,0.1115138754,0.3066324592,0.1736799479,-0.0907347575,-0.0430986919,0.0739057139,0.1339908987,-0.0127220284,-0.2152223885,-0.0496791638,0.0812612399,-0.2455009967,0.0428519771,0.0226380769,-0.3167784214,-0.0857654735,-0.0128047746,-0.0629137754,0.1333414167,-0.3916155398,0.208681345,-0.1020074487,-0.7172011733,0.2480072379,0.3210870624,0.3370936513,-0.1327753663,-0.1462028176,0.163606599,0.4743584692,-0.1735897362,-0.0122842034,-0.092538543,0.2693701684,0.0069938083,-0.005127138,0.0592472069,0.0436879508,-0.3624272943,-0.3166557252,-0.1638199687,0.4906712174,-0.2466342151,-0.5456697941,0.2004903108,-0.0626043677,-0.0492505133,-0.0078771356,0.0940499008,-0.0801145881,0.0802997947,0.2769869268,-0.3589984775,0.0691729337,0.3846994638,0.1874381006,0.378805995,0.3787675798,0.166243732,-0.0843464285,-0.0581624135,0.2729899883,0.8475379944,-0.2112778574,-0.0550370589,-0.1017795801,-0.0667243749,0.1764791608,-0.1538363248,-0.0242772251,-0.010791216,-0.0010594173,-0.168432653,-0.3776629269,-0.009420841,-0.5037513971,0.2098778933,0.0322902761,-0.0655000508,-0.3649929166,0.0716206357,-0.1308607161,0.1892534792,0.2305907756,0.1326996833,-0.1198969036,0.1519486904,-0.0308755022,0.1419837773,-0.0059039588,0.0213198662,-0.1980637163,-0.1242233142,-0.234045893,0.2140558064,0.101650767,-0.1233232841,0.108550936,-0.2530794144,-0.2654302418,0.0349137299,0.1609399766,0.2066160887,0.0862387493,0.2973311543,0.340244621,-0.1025866047,-0.2123933285,0.0754836351,0.3458487391,0.3790521026,-0.138481006,0.1107969284,-0.1813203841,-0.2900881171,-0.3661788702,0.1722217202,0.5551644564,0.0176636931,0.2392827272,-0.0788118839,0.1425167322,-0.2420075983,0.2221855968,0.3790283203,-0.1063592583,0.0677401274,0.0594033115,-0.0754961818,-0.2994278669,-0.0385040082,-0.2173163146,-0.3141108155,-0.0723757669,0.0272178091,0.1255375743,0.5410738587,-0.1436738521,-0.2965466082,0.0447992049,-0.345156759,-0.1395789683,0.3816301525,0.2292929888,0.0080817286,0.3678341806,0.2350911647,0.3199836612,0.1539038867,0.4099703133,0.8492857218,0.6417722702,0.0754529312,0.4385546148,-0.4119920433,-0.3295015395,0.2571868896,0.1432525516,0.1258504689,0.1502238959,0.4933296144,0.1703612357,-0.566595912,0.0716961101,0.1914031655,0.0346361548,-0.4513969421,-0.1302147508,0.0413922891,-0.3098055124,-0.124427475,-0.3104376793,-0.1233868897,-0.0551060177,0.225851953,0.0566277243,-0.0724428669,0.1049896181,0.2062295675,-0.1139234453,-0.0713274106,0.1549415439,-0.1350672543,-0.0542063937,0.0578979477,0.1662692726,0.2439159751,-0.0082278848,-0.3018837869,-0.0454342403,0.4646241367,-0.164520219,0.3403019905,-0.073619388,0.2081261128,-0.1007426679,0.3429599404,-0.0059633348,0.0626542494,0.135430038,0.0287158787,-0.0066623315,0.0180635471,0.2481313199,0.1146365255,0.0294339657,-0.1742191762,-0.0112429895,-0.3586736023,-0.0496014133,0.6691644788,-0.2396361381,0.1143987924,0.0804413855,-0.0062691486,-0.0650279373,0.1841631532,0.3714635372,0.1079923138,-0.3948980272,0.0223799199,-0.1956920475,0.0392694622,0.0264249668,0.1066620275,-0.0834692866,-0.0030416602,0.3395575285,0.0002309813,-0.0614797175,-0.0334356353,0.2042383999,0.5422481894,0.0192900803,-0.1077275351,-0.2591682374,0.3638250232,-0.1441082805,-0.3204354942,-0.0903427601,0.2091027349,-0.0847721323,-0.0010740056,-0.2335554361,0.5360317826,-0.0856618583,0.3362798989,-0.0733782947,0.4900251329,-0.2959281206,0.1195824072,0.173921302,-0.0841972753,-0.4016540945,0.2119205743,-0.0927228108,0.2605184615,-0.0029802273,-0.1320884079,-0.0440982506,0.1206762418,-0.1207277179,-0.4478740692,-0.2386040837,0.4975051582,-0.2389293015,-0.096418716,0.2845879793,0.4631789625,0.1880175471,0.2012938708,-0.160835892,-0.1496623158,0.5671274066,0.1585633159,-0.0479432419,0.0831967369,-0.1093915924,0.2654272914,0.0603676438,-0.3633694649,-0.1027078852,0.1557672769,-0.2713345587,0.1184527799,0.2174606919,0.2340915352,0.27198264,-0.300362736,0.4093898535,0.2130363137,0.0507346727,-0.273974061,-0.3515224457]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2165","title":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset","comments":"@mariosasko \r\nThx for your code!\r\nIt perfectly works with a small modification for HF NLP dataset:\r\n```\r\noriginal_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds = HFDataset(train_ds['train']) # needs splitting\r\n```","body":"Hi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n","comment_length":28,"text":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset\nHi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n\n@mariosasko \r\nThx for your code!\r\nIt perfectly works with a small modification for HF NLP dataset:\r\n```\r\noriginal_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds = HFDataset(train_ds['train']) # needs splitting\r\n```","embeddings":[-0.2263557464,-0.3335738182,0.070314914,0.3271583915,0.1331761032,0.1971441209,-0.0642952695,0.3733654618,-0.1391186118,-0.2662570179,-0.3984828591,0.3849843442,-0.179449752,-0.009580628,0.1554363668,-0.300897032,0.1781212389,0.076728873,-0.2033017427,-0.2192837894,-0.0984729603,-0.0496377274,0.0211220514,0.125883013,0.0060907789,0.0596893653,0.0468233041,0.2068376541,-0.0479721501,-0.3290063739,0.1856233627,0.0352230929,0.5887832046,0.2546223998,-0.0001306744,0.1232334152,0.0475138538,-0.1539012492,0.1102606431,-0.1059945002,-0.1251382828,-0.037439622,0.1923864782,-0.1235969514,-0.3527246416,-0.5381259322,-0.0982355997,-0.290035665,0.1821317077,0.5682997108,0.0439835005,0.2380842417,-0.0001738147,0.0502141267,0.2654594481,-0.1158183888,-0.1378005445,0.0770626739,-0.0951834396,0.1144737378,-0.2686012983,0.1428055316,-0.1503308266,0.072083056,0.3074949682,-0.3364239931,-0.4187689722,-0.1907426119,-0.0029426359,0.5164994001,0.1975763738,-0.2164663225,-0.0486005694,0.1461015195,-0.1083505079,-0.1705802828,-0.36707443,0.2305695266,-0.0336455964,0.1331619173,0.037428882,-0.1993561387,-0.1586773843,0.0975329578,0.2056356072,0.3252075911,0.0735005364,0.1826860458,0.2181069851,-0.3050098419,0.3931107819,-0.2096302658,-0.032205753,-0.1784204692,-0.1522565931,-0.0226949342,-0.3895271122,-0.0581793077,0.4548136294,0.0893317983,0.1379813701,0.3401524723,-0.1559869051,0.0358810574,-0.1179903522,-0.1961131692,0.1375550181,0.2798586786,-0.0167926978,-0.3020294011,-0.0787237957,0.233344838,-0.0840417594,-0.412138164,0.053519275,0.0179046523,-0.1774987131,-0.2150649875,-0.2046911567,-0.2606694996,-0.2944220603,-0.0563340746,0.0382383801,0.2656323016,0.0401877016,0.0500871055,0.1952546835,0.6915662289,-0.2502996325,-0.1751054674,0.0112741599,0.1056316867,-0.0220319144,-0.182876125,-0.1040798649,-0.2147393376,0.1593292803,-0.0057802135,-0.2855864167,0.1035448611,0.3521379828,-0.1735322922,0.2845810354,0.0553727895,0.1202842668,0.0789878294,0.3318808675,0.3952591419,-0.3046818376,0.2600634694,-0.5100749135,-0.2010753751,-0.1981972903,0.0106506571,0.1503171921,-0.2748128176,-0.4242849052,0.3067432344,0.0271716546,0.1364007741,0.0167120453,-0.4317055047,-0.3108105063,-0.1870919764,0.2145445198,0.2237268388,-0.3245928586,-0.1132323965,0.1227968559,-0.1630419791,0.2324233502,0.3118992746,-0.2270777225,0.0936800241,-0.4663161039,0.0080991257,0.6656390429,-0.3257720172,0.1022588164,0.1635378003,0.0983478501,-0.2311872244,0.1086594462,-0.0553934127,-0.0797104612,-0.0385858119,0.1303543001,0.2308624685,-0.094621554,-0.1622785926,-0.1658529043,-0.2639196813,0.5142816901,-0.0872293562,0.2348762602,0.3476688564,-0.0908901617,-0.2052481771,0.3716835678,-0.0414164327,0.1036917642,0.2142174244,-0.3256297112,0.0535456091,-0.073021993,0.0921149254,0.1350721419,-0.0252536591,0.0498598292,-0.033963386,0.0881131068,0.1063458622,-0.2813718021,0.1484549791,-0.1486712545,0.0176728163,-0.1103216186,-0.0677547008,-0.2025362998,-0.1761354953,-0.0231417269,0.286152482,-0.3305152357,0.12785092,-0.4923216701,0.2535890639,0.0675133988,0.1056316346,0.1362594068,0.1969215572,0.1536073983,-0.0305043459,-0.1565905809,0.2011708468,0.0490445942,0.0091487216,0.359028995,0.2326785326,0.0966187045,-1.0042270422,0.3522100151,0.4355178177,0.1550198644,-0.1452267319,-0.0021778455,0.6890761852,0.2010133117,0.2365346253,-0.0610618405,-0.0508690514,0.0942555219,0.1049328148,-0.1251616329,0.0396533199,-0.0850838497,-0.0466094166,0.2758455575,-0.0639292672,-0.3997496367,-0.0149441082,0.1547486633,-0.0097320555,0.0751002729,0.3563731611,-0.5196492672,-0.1247848123,0.3244453073,-0.3547014892,0.1424794346,0.1628667712,0.1763049215,0.0010702891,-0.0438614339,0.0199416392,0.2627765834,0.0312216226,0.4762645066,0.0328029841,0.1253219992,-0.0937413201,-0.320377022,0.0578687899,-0.1390546262,-0.0352983363,-0.4970809519,0.1807901859,-0.4068142176,-0.1239830256,-0.2987782061,-0.0224426761,0.0396812484,-0.109131597,-0.2386263758,0.0302582346,0.3040401042,0.0953090936,-0.417355597,-0.3343822062,0.219783932,-0.2575615644,-0.0490855873,-0.1521484107,-0.1422960311,-0.0792501494,0.0874339938,0.0642262474,0.2967342436,0.1606131494,0.117771633,-0.3894567788,-0.1165060624,0.0104811285,-0.1332972944,0.0998919457,0.1726034433,0.1334909797,0.1722554117,0.0761121884,0.0592288338,0.0346762761,-0.0320955515,0.1489344388,-0.3590313792,0.1663262397,0.4372859299,-0.1095225364,-0.2783699036,-0.2167597115,-0.3826257586,0.0475437827,0.1732869744,-0.1151548102,0.3042676449,0.1316343397,0.211764127,-0.0758673996,-0.121579662,-0.0538332798,0.360645473,-0.0776255131,-0.3281585276,-0.1559724212,-0.2494729608,0.0884207562,0.2179971188,-0.3521410227,0.1593417078,-0.1564198434,0.3416546881,-0.0553478934,-0.0349266231,0.3035986125,-0.1843936294,0.1471060514,-0.2114264071,-0.2488058656,0.061330162,-0.1483648866,0.3629367054,0.3649741709,0.4517152905,-0.1608110219,0.9838785529,-0.0303140003,-0.3359048367,0.5201300979,-0.1261212975,0.086239405,-0.2030792385,-0.2672263384,-0.0298658479,-0.0799623504,0.4489382803,0.1056021154,-0.1356224567,-0.147509858,0.0822239891,0.0455772169,-0.1306184828,-0.0660999492,0.2251461297,-0.84493047,-0.1056759581,0.0727220327,0.0569750778,-0.2449705154,-0.0726714134,0.0564687438,0.3254698515,0.2142682523,0.2519320548,-0.3784071505,0.1734532416,-0.2046069056,0.4338724613,-0.0459957384,0.5219185948,0.1248015389,-0.0683132038,0.0906441063,-0.0323842317,0.4775116146,-0.1738080084,-0.0706753358,0.0359209105,-0.025581032,-0.4465414882,0.0172570571,0.0562241934,0.0697536767,-0.0893032551,0.1419761926,-0.1395868808,-0.4577145875,-0.0526844971,0.0157329999,-0.0781969503,0.2679589689,0.0275434498,-0.2418017089,0.0301985741,0.0325047299,0.0150478547,-0.0481001846,-0.1440620124,-0.2042286098,-0.4858337641,0.0070133517,0.2883121967,-0.1957104057,0.2373309433,-0.088448666,-0.0736440942,-0.2873155773,0.3368117213,0.2056099474,0.2770309746,-0.1854200661,-0.1713092029,0.0132958274,0.2324945778,0.3811320066,0.0995606706,-0.0634262264,0.2771746516,0.0227152798,-0.0596839488,-0.3146493137,0.1943282485,0.317125231,-0.0140805636,-0.5217674971,-0.4427949786,0.6736053824,0.1517965347,0.1031016335,0.1824989915,0.290576309,-0.2887245715,0.6902830005,0.1923063844,1.0491645336,-0.1496044248,0.3091436327,-0.019248344,-0.3549844623,0.1914817691,-0.0620067716,-0.0127197495,-0.5064066648,-0.0988877341,-0.1696773767,-0.1890803277,0.045667775,0.1084339693,0.3459700048,0.211841315,-0.1081131622,-0.03563058,0.0932142287,0.1610936075,0.0246001612,-0.2186434567,-0.0476582982,0.0867055953,-0.2787378728,0.1409859508,0.0454186499,-0.3511820138,-0.1316142976,-0.0653837547,-0.0387333818,0.135840416,-0.3878106773,0.1912806183,-0.0955749899,-0.722196281,0.3027074933,0.2570642829,0.3729010224,-0.1525343508,-0.0759721845,0.151722312,0.4544072449,-0.1220795289,-0.0165471155,-0.1029004231,0.2754592597,0.0021330949,-0.0121416757,0.0144783119,0.0496026464,-0.354049325,-0.3075190783,-0.1008297652,0.4927372038,-0.2824708223,-0.5382110476,0.180941835,-0.028042227,-0.0404096283,-0.0311694853,0.0760166347,-0.1006569266,0.0231765117,0.219609797,-0.3577341735,0.0479486436,0.3908985853,0.2444619834,0.3545530438,0.4260512888,0.1283814311,-0.0747564286,-0.0551734678,0.2427574694,0.7975454926,-0.2431243211,-0.0955932811,-0.0828490183,-0.0964286476,0.1759666055,-0.1229593977,-0.1000612453,0.0187568218,-0.0037023118,-0.2240096331,-0.3872649074,-0.0501669534,-0.5921089053,0.1292785108,0.0445958339,0.0095894188,-0.3298636377,0.1190791354,-0.119034566,0.1844636947,0.2547365129,0.1583484709,-0.1665880084,0.1188053861,-0.0005616251,0.1166002825,-0.0072357031,0.0770435557,-0.1851371974,-0.123597756,-0.2432995141,0.2295061201,0.0940965638,-0.1306687444,0.0751342177,-0.2411138862,-0.2445299327,0.0727368072,0.1853953004,0.1755363345,0.1117695123,0.3480849862,0.3641879261,-0.0677340701,-0.2277558595,0.0497227348,0.3066651821,0.3425446153,-0.1606393158,0.1104381084,-0.1617767662,-0.2147651315,-0.4247927368,0.1144338623,0.5022952557,-0.0028918253,0.1857275069,-0.1026440933,0.1288243532,-0.255785048,0.1771665812,0.392242223,-0.0652695969,-0.0092794746,0.0954942256,-0.0923827291,-0.3175997734,-0.0011595017,-0.2000226527,-0.3673337698,-0.1038198248,0.0687591732,0.1370665878,0.5545260906,-0.125981763,-0.3066502512,0.0544887483,-0.3451065719,-0.1011897624,0.4092557132,0.1976185441,0.0214364864,0.3923998773,0.2195708603,0.2428042144,0.2284490913,0.3846330941,0.8282759786,0.6093149781,0.0693631917,0.4093192518,-0.439285785,-0.3943210542,0.2902410626,0.1612641215,0.095513016,0.0909392014,0.5308842659,0.1345909238,-0.6116360426,0.1243948266,0.1417175233,0.0394368619,-0.3871904016,-0.0720665306,0.0942411348,-0.3360298574,-0.1689391434,-0.3047473133,-0.1319286674,-0.0257730056,0.2366738319,0.0376940444,-0.077047877,0.0455885418,0.1857755631,-0.1383960843,-0.0701644942,0.1703161746,-0.1693473309,-0.0163693279,0.1089865044,0.1222435981,0.2206779122,-0.0236985516,-0.3086427748,-0.102700606,0.4592534006,-0.1692235917,0.3425976038,-0.0366069861,0.202995047,-0.0599706247,0.3750295937,-0.032713186,0.0678919256,0.1241498739,0.0405827314,0.0066335215,-0.0242575742,0.3539984226,0.0674851388,0.0228485074,-0.1710181385,-0.0509920754,-0.3561062515,-0.0426161289,0.5597941875,-0.2557029426,0.1359279901,0.0525568686,-0.0106771076,-0.1017212644,0.1591456681,0.3113599122,0.113558583,-0.3916855454,0.0128198992,-0.202885583,0.0193147697,-0.0143975215,0.0728556216,-0.0521284938,-0.0008855254,0.3236848712,-0.0529138148,-0.0882758051,-0.0321378298,0.2348828167,0.4995101392,0.0454158857,-0.1006599367,-0.3244436383,0.3683188856,-0.1543809474,-0.3648245931,-0.0524975024,0.220898509,-0.1017179638,0.0308083557,-0.2137626708,0.5447202325,-0.0581353046,0.2709105015,-0.0344842896,0.4849650264,-0.2545777857,0.1276676655,0.1671371758,-0.0664449707,-0.4173105359,0.1933073848,-0.1038347483,0.3298132122,-0.0180293787,-0.0523737781,-0.0486999936,0.1613539457,-0.1031826958,-0.3979177177,-0.2104855031,0.5480275154,-0.2608939409,-0.0680663213,0.3363808393,0.513948679,0.1706792861,0.2059370279,-0.2271478176,-0.1567133963,0.5854461789,0.1026785001,-0.0993436798,0.1165449545,-0.0490630977,0.3045933247,0.0171859581,-0.3708270788,-0.1677887142,0.1889812946,-0.2670780122,0.1175101474,0.2422823012,0.2049200088,0.2456115186,-0.2822860181,0.3716866374,0.2165394425,0.0090612173,-0.1876170784,-0.3382449448]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2165","title":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset","comments":"@lhoestq Sadly, from Python 3.7 onwards `torch.utils.data.Dataset` doesn't support the virtual subclass mechanism due to `typing.Generic` type no longer having `abc.ABCMeta` as its metaclass.\r\n\r\nWith that in mind, another option is to remove a direct type check (`isinstance(dataset, torch.utils.data.Dataset)`) in `deepspeed.initalize` and to rewrite the checks in a manner similar to `torch.utils.data.DataLoader` ([link](https:\/\/github.com\/pytorch\/pytorch\/blob\/b80c6f863f2327c712c478f67c248b94d66b65ac\/torch\/utils\/data\/dataloader.py#L197-L239)). This is exactly why the `DataLoader` works with arbitrary objects that provide `__getitem__` and `__len__` (and in our case, the `ArrowDataset`). By doing so, their code wouldn't be any stricter in comparison to the `DataLoader`.\r\n\r\nSo if you agree, I can open an issue in their repo and fix this if they like the idea.","body":"Hi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n","comment_length":108,"text":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset\nHi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n\n@lhoestq Sadly, from Python 3.7 onwards `torch.utils.data.Dataset` doesn't support the virtual subclass mechanism due to `typing.Generic` type no longer having `abc.ABCMeta` as its metaclass.\r\n\r\nWith that in mind, another option is to remove a direct type check (`isinstance(dataset, torch.utils.data.Dataset)`) in `deepspeed.initalize` and to rewrite the checks in a manner similar to `torch.utils.data.DataLoader` ([link](https:\/\/github.com\/pytorch\/pytorch\/blob\/b80c6f863f2327c712c478f67c248b94d66b65ac\/torch\/utils\/data\/dataloader.py#L197-L239)). This is exactly why the `DataLoader` works with arbitrary objects that provide `__getitem__` and `__len__` (and in our case, the `ArrowDataset`). By doing so, their code wouldn't be any stricter in comparison to the `DataLoader`.\r\n\r\nSo if you agree, I can open an issue in their repo and fix this if they like the idea.","embeddings":[-0.1920574605,-0.2704025805,0.0988816097,0.3232178986,0.2382786572,0.1883175522,-0.045659896,0.3633685112,-0.1526031047,-0.2291758657,-0.2917279899,0.3924194276,-0.2503360212,-0.1537752897,0.1516395509,-0.2407059222,0.181158945,0.0032430561,-0.269872725,-0.199164629,-0.1094516814,-0.1029918268,0.0290850047,0.0508622639,0.0944085643,0.1040174365,0.0816430524,0.1808154732,-0.1128340736,-0.3632012308,0.3348354399,0.0631520003,0.568320632,0.3380843997,-0.0001285925,0.2495380491,0.1516090184,-0.0954126567,-0.048618298,0.0058680433,-0.0310990922,-0.1078799888,0.2537959218,-0.0793687776,-0.2585174739,-0.6241518259,-0.1227698997,-0.5026624203,0.0664021,0.4283711314,0.0463510714,0.3282434344,0.0120554427,0.0952218473,0.2861973643,-0.0377576649,-0.1647085696,0.1939546615,0.058613088,0.1633573174,-0.1833067387,0.0865228325,-0.1490781903,0.0236566886,0.4681111276,-0.2537768781,-0.3934955001,-0.1979160607,-0.1612958759,0.5067217946,0.2559040487,-0.3469001651,-0.2240120173,0.0779629573,-0.0618266314,-0.1476538777,-0.217859596,0.1044944972,-0.0763134807,0.0751041695,0.0757466853,0.0165292565,-0.1876958758,0.1010662317,0.1476613879,0.1680387855,0.1117822528,0.1688792557,0.1994540989,-0.3436609209,0.3780586123,-0.2171828002,0.0021823584,-0.1452053636,-0.1031779498,-0.1081254631,-0.3531705737,-0.1528615355,0.383035928,0.1301053762,0.3453951478,0.3318163455,-0.2277630717,0.0814984962,-0.0249896515,-0.1177566424,0.1623562276,0.2724095583,0.0154545484,-0.1962260902,0.0820427164,0.2925750017,-0.1423644423,-0.3170528114,0.1186282784,0.0623504221,-0.1232669055,-0.1359723061,-0.1813091934,-0.1504846066,-0.2349579036,0.0214676932,0.1157316044,0.1644117832,-0.0010171535,0.218443051,0.3691661656,0.5446549654,-0.2553099394,-0.1770908087,0.0780243725,0.1107357293,-0.0750111192,-0.2217584997,0.0215201508,-0.1792351902,0.0170285013,-0.0673136413,-0.1304965019,0.3081122637,0.2849660218,-0.1399293393,0.2640922964,0.066293478,-0.0842040703,0.0411605984,0.3640844822,0.4290799797,-0.3253016174,0.2448960245,-0.3367206454,-0.1767986864,-0.1608814746,0.0041704481,0.1324421465,-0.3356076777,-0.3940821886,0.1987404376,0.0049035777,0.0119573334,0.0823145136,-0.5586115122,-0.2367926985,-0.1953954101,0.2141402662,0.2105928212,-0.4652603567,-0.1148305535,-0.032348454,-0.1110006422,0.1935072243,0.307397604,-0.2178398818,-0.02567886,-0.3970097005,-0.0204563402,0.5086792111,-0.2867502272,0.0247857943,0.1531435698,0.1048132852,-0.0957697034,0.1215051711,0.0464643911,-0.0526720993,-0.1397288442,0.1875391155,0.1678770036,-0.1229540184,-0.2729192674,-0.1068385616,-0.2577292621,0.3619751334,-0.0371052884,0.1300515831,0.21104303,-0.1213669851,-0.2831217051,0.2862468362,-0.0242851339,0.1068466604,0.1502738148,-0.201292932,0.1783106625,0.0183900129,0.0233058948,0.0052240025,0.1541561186,0.0099737821,0.0297068898,-0.0348640531,0.0375255942,-0.1254970282,0.1646582633,-0.0355101675,0.0571016632,-0.0778035596,-0.0694773495,-0.2363839,-0.2742439806,-0.1225562841,0.0787640288,-0.4409392476,0.178393513,-0.5549850464,0.1552928835,0.1712520868,0.118306607,-0.0533446819,0.2489914149,0.0549134687,-0.0293417964,-0.1652685553,0.179641068,0.123894982,-0.0634167492,0.2238185704,0.2895556986,0.0666508451,-0.8759514093,0.3386188447,0.4776403606,0.1758344322,-0.0606287606,-0.0306877978,0.6746484637,0.269923985,0.2855539024,-0.0329919308,-0.009948181,0.0626338199,0.2108589262,-0.2279220968,0.0115737105,-0.1417211294,0.178337425,0.2537435591,0.038325455,-0.243582055,0.0231189635,0.0286006462,-0.1075292602,0.1185367703,0.3567152619,-0.5336498618,-0.017127106,0.2257860303,-0.1542957425,0.2203960866,0.079782635,0.2486361116,-0.0573246554,-0.1059990004,0.0514930636,0.3088961244,0.0781626776,0.3587369621,-0.0323414095,0.0559454672,-0.055926349,-0.3572993875,0.0482363254,-0.1054936945,-0.0915639922,-0.511428237,0.2086586803,-0.4489869177,0.0558845885,-0.3178323209,-0.11460834,-0.0138481874,-0.1688977927,-0.1546970457,0.0197159052,0.1207129732,0.1362329274,-0.4001930952,-0.1381525248,0.2074002773,-0.4670451283,0.0483384393,-0.1286058426,-0.2392186224,-0.11042054,0.1444949061,-0.1327437758,0.2613075376,0.2039724141,0.0211454481,-0.4411420822,-0.1653214991,0.084304221,-0.165503338,0.0724387318,0.2261061519,0.0766591281,0.3106871247,-0.0076395953,0.0699216053,-0.0379846878,-0.0956487656,0.1708254069,-0.3793530464,0.2229391485,0.3711147308,-0.1360011548,-0.3164051175,-0.2379695773,-0.3310390413,0.0463848375,0.1217815578,-0.0089485813,0.3633452654,0.1398113519,0.2525959611,0.0753093213,-0.0746526346,-0.0394742712,0.4306192398,0.0210851952,-0.309174329,-0.1776057482,-0.3541427255,0.1667936444,0.3370664716,-0.3112230897,0.0742642134,-0.2507444024,0.3438512087,0.0475702249,-0.0304733571,0.3099016547,-0.1462945938,0.0725222453,-0.2030642629,-0.1782543659,-0.018077448,-0.2516228259,0.3121638298,0.3920455873,0.4929171503,-0.1153860688,1.0432752371,-0.1220421493,-0.5147123337,0.4142384827,-0.1147439703,0.1429576278,-0.1226013973,-0.2998675406,-0.085366033,-0.2901656628,0.2939036191,0.0313187093,-0.1601836979,-0.2480983585,0.1413031518,0.1854305267,-0.0607377812,-0.1736496687,0.3794947565,-0.8608574867,-0.080937393,-0.0547402278,0.145854488,-0.3503080606,-0.0442955829,0.0788310766,0.2806054354,0.1299971789,0.1442867965,-0.2763812244,0.127389133,-0.1922665685,0.4928195477,0.0262334682,0.6061600447,0.1682636142,-0.2415487468,-0.0266088825,-0.0479809493,0.3368054032,-0.1201387271,-0.0284864809,0.1070727929,-0.153774634,-0.5539861321,-0.0416317023,-0.0091988444,0.0008516716,-0.0880936012,-0.0115801767,-0.2062880099,-0.3495067656,-0.0042238361,-0.0204436574,-0.0742843971,0.2395645976,-0.0618916377,-0.1756765097,0.0153800696,-0.0085249795,0.1251471043,-0.0059991619,-0.1849310398,-0.1844238639,-0.4665316045,-0.072802715,0.319319427,-0.195409745,0.2066604942,-0.0913769752,-0.0270608813,-0.1993328929,0.3776907325,0.3188765347,0.4460636675,-0.1958580911,-0.314585954,0.1087746099,0.3574344218,0.3197109401,0.1773001999,-0.1695223302,0.2552005649,0.1032042056,-0.1656397283,-0.399043113,0.1485093832,0.3337714076,-0.0139878299,-0.384973526,-0.5564731359,0.7538570762,0.3035254478,0.1560622752,0.3382978141,0.2668631077,-0.2794843614,0.5248183012,0.4079276323,1.0512201786,-0.0509784594,0.3578878343,0.059565682,-0.1912040859,0.173831597,-0.0342134126,0.0107817082,-0.4742813408,-0.1591959447,-0.1777703911,-0.1990585178,0.0807511881,0.0677642226,0.3001704216,0.1468989402,-0.1568988115,0.1233604848,0.1443847865,0.137003541,-0.2108580917,-0.2943648398,0.0387039557,0.0807715952,-0.1526237279,0.0323701017,0.0839411318,-0.2476135045,-0.1776064634,-0.0623997487,-0.174337104,0.0983655006,-0.3862039745,0.2085451782,-0.0430509038,-0.6213812828,0.3718841374,0.2089663446,0.334646523,-0.2072170228,-0.1145251095,0.0898869485,0.4293406904,-0.0604890697,-0.0188369993,-0.1988706738,0.4383510351,0.1158377007,0.1028564349,-0.0041483352,0.0040067574,-0.371676147,-0.1788451374,-0.1418672204,0.4884434938,-0.3600063324,-0.3033034503,0.0732410103,-0.1057725325,-0.0269546546,-0.0135335717,0.0660573542,-0.0794526264,-0.029825164,0.2205516845,-0.3070401847,0.1049227268,0.3794654608,0.2592764199,0.2761076689,0.4509197772,0.1658732593,-0.0842058882,-0.084031418,0.2752795815,0.9381102324,-0.2417883426,0.1131504551,-0.0707230046,-0.1823946089,0.1009256095,-0.1334833205,0.105553098,0.0117177842,-0.1784546226,-0.2008000016,-0.4730121195,0.106684871,-0.4885850847,0.252222389,-0.1127750725,-0.0003833159,-0.2478294522,0.0050857691,-0.1294638664,0.1128219366,0.208162114,0.0766926259,-0.1583962142,0.1341680586,-0.001301205,0.1834419668,-0.0373182595,0.1008868068,-0.2249795645,-0.0719658136,-0.2094803751,0.2116009593,0.0788350627,-0.0487045869,0.1003225073,-0.2600993812,-0.2248568386,-0.0455035642,0.1009886414,0.0768968686,-0.0250033606,0.336188823,0.3974800408,-0.0176139921,-0.2065616697,0.0856440812,0.2487569749,0.4111220241,-0.0284354817,0.0919955149,-0.1686856896,-0.252120167,-0.3187128603,0.2463505715,0.4437052608,0.047410816,0.2199501842,-0.0745144039,0.0695232973,-0.3206493855,0.2149876654,0.4866550565,-0.0578187034,-0.0545588881,-0.0114212651,-0.0757067949,-0.2705475688,-0.045503471,-0.274125576,-0.3820186853,0.0093019353,0.092543073,0.1122258082,0.5620799661,-0.089305304,-0.256685555,0.0706409439,-0.2954809666,-0.028384611,0.4573873281,0.1571798772,0.0178327952,0.455740422,0.1562418342,0.3069688082,0.2131180465,0.3235092759,0.8402161598,0.5083611608,0.0241861492,0.416328609,-0.4468284547,-0.1702664346,0.3761992455,0.2840809226,0.1153379902,0.1381605268,0.4704501331,-0.0412484966,-0.48365134,0.0164676309,-0.014304745,0.0409532711,-0.3381249011,-0.3427492678,-0.0142067028,-0.4155673087,-0.2136502713,-0.3284617364,-0.1840654463,-0.0310501195,0.1535822153,-0.0573646948,-0.2048217654,0.0162403565,0.1845861226,-0.1490480304,-0.0861728787,0.0984741598,-0.0619472601,-0.0533340313,0.0844013989,0.1812232286,0.2301818579,0.1560850888,-0.3443406522,-0.0796003714,0.3975009918,-0.1332998127,0.285816133,-0.0066142892,0.0329288505,-0.050483413,0.3187774122,-0.0040306086,0.0654802248,0.0210266821,-0.0072604404,0.0509192869,-0.0689643919,0.5326979756,0.1324308664,0.0325450748,-0.1482269466,0.054076422,-0.3688842058,-0.155706808,0.6477063298,-0.1662311852,0.1110663414,0.0502600819,-0.0144342249,-0.0233093407,0.3688507974,0.3647022843,0.0189680792,-0.3466326594,0.0738323927,-0.3001281023,0.0229688585,-0.0210429132,0.0045116739,0.033984486,0.0613775253,0.5213210583,0.0317596011,-0.0474660359,0.0604745783,0.1181068942,0.3871954978,-0.0312981941,-0.2424382567,-0.1258305758,0.3430368602,-0.1222838983,-0.3957667351,-0.0955137089,0.1792186797,-0.0882429704,0.0777847841,-0.3348729014,0.6279917955,0.0028238064,0.3965129852,0.0139712077,0.4773536623,-0.1456064731,-0.0347250588,0.0490610227,-0.0706805289,-0.4631290436,0.2110556066,-0.1499648541,0.327979207,-0.0359973423,-0.1442195177,-0.1181407347,0.1551754624,-0.1119900793,-0.4463451803,-0.3844429851,0.5214521885,-0.3229981065,-0.0472746082,0.1672801673,0.4711968899,0.1661391109,0.2436231673,-0.2501737773,-0.2161932886,0.578556776,-0.0363396071,-0.1052244306,0.0236454699,-0.0473427959,0.2287595272,0.1911903918,-0.3108547926,-0.1146501154,0.1373117715,-0.2019237727,0.1190903336,0.1711321026,0.214611575,0.2068580687,-0.3751349747,0.4397037029,0.1238738894,0.0825854987,-0.2764619887,-0.3310411274]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2165","title":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset","comments":"That makes sense ! Feel free to open an issue on their repo and discuss this idea","body":"Hi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n","comment_length":17,"text":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset\nHi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n\nThat makes sense ! Feel free to open an issue on their repo and discuss this idea","embeddings":[-0.2620015442,-0.305977881,0.0482731722,0.3386891484,0.1174888611,0.2266668677,-0.0971872956,0.3648378253,-0.1662661433,-0.218519181,-0.3497868478,0.394482851,-0.2521584034,-0.0014390136,0.1679954678,-0.2934633195,0.2071099281,0.0787692592,-0.2426628768,-0.2180524319,-0.1317133754,-0.0721179247,0.0561505035,0.1099010855,0.0453758053,0.0387697108,0.0631345063,0.1353642642,-0.1149804741,-0.3563358188,0.2166966498,0.0383347869,0.5877937675,0.2521918118,-0.0001278631,0.1164884567,0.0586811751,-0.1147667915,0.1068665981,-0.0929902941,-0.174374029,-0.0208212119,0.2010273784,-0.1319154799,-0.307541728,-0.5764225125,-0.0690180138,-0.3266124725,0.1039493755,0.5637236238,0.0582137741,0.2589324415,0.0555239804,0.0262063984,0.3165854514,-0.147557795,-0.1901795268,0.0691053122,-0.0712911859,0.1284115613,-0.1900322437,0.1756014824,-0.1578796208,0.0448201932,0.3596791625,-0.3286312819,-0.3591727018,-0.1928064525,-0.0682306364,0.5270514488,0.2203394622,-0.2135134488,-0.0389554873,0.1802363545,-0.0732751936,-0.1967220008,-0.3787529469,0.198263824,-0.0162514597,0.1270776987,-0.0345080607,-0.2056979984,-0.1423495859,0.071584031,0.1724095792,0.3733929396,0.0766004547,0.1745888442,0.2233617008,-0.3262197375,0.4041501284,-0.2009280473,0.0032965213,-0.1929715872,-0.1247138083,-0.0328153223,-0.3565469682,-0.0698654428,0.4570572376,0.1322319508,0.2172506899,0.308917284,-0.1965114176,0.0849978849,-0.1674623042,-0.2068859786,0.1313789189,0.2410633713,0.0171281621,-0.270380646,-0.016132243,0.2711999416,-0.1041060314,-0.3959648609,0.0039959676,0.0003850678,-0.1091186702,-0.2102986723,-0.1964522302,-0.2079259306,-0.268462956,-0.050248567,0.0188779403,0.2163887322,-0.0120289233,0.0506806783,0.2428289056,0.6414881349,-0.2044611126,-0.097200729,0.0029477645,0.0924842507,-0.0101870429,-0.1769014001,-0.0899683461,-0.149987042,0.1072871983,-0.0319974124,-0.2831275761,0.1549769044,0.3588211536,-0.097734794,0.2597135603,0.0583751425,0.0769039541,0.0361856818,0.2901774347,0.4265041947,-0.3359179497,0.2660285532,-0.5100514293,-0.1835344583,-0.1684627682,0.0424546376,0.1579413116,-0.3772470057,-0.4059728384,0.3280169666,-0.0173317287,0.1169557199,0.061781045,-0.4518920183,-0.337308079,-0.1928305179,0.1911066025,0.2245470732,-0.3925149441,-0.1291614622,0.1234330535,-0.1684450209,0.2170879543,0.3301728964,-0.2159967273,0.0313050635,-0.4144066274,-0.0111348229,0.6295618415,-0.2943707407,0.1210268736,0.0977080241,0.1029670089,-0.2821547985,0.1250987202,-0.0195885003,-0.1065179259,-0.0370885618,0.1584386826,0.2269485444,-0.1056133434,-0.1747419834,-0.1187010482,-0.320887357,0.4772762954,-0.066561155,0.2013028115,0.281778127,-0.1068628952,-0.2452245504,0.3179291785,-0.072518222,0.1154307649,0.2461644113,-0.2807471454,0.0252963305,-0.0772584677,0.1072368175,0.1306121349,-0.0304074716,0.0351776928,0.0028217356,0.0903309956,0.069700934,-0.2287643403,0.1778737903,-0.109227404,0.0447425656,-0.0674027279,-0.0622881912,-0.1602267325,-0.1794353276,-0.0604325496,0.1831260026,-0.3872933984,0.1212605238,-0.5793866515,0.2193060666,0.0925740972,0.1212818176,0.0703063309,0.2091723382,0.1279124767,-0.0046756081,-0.1303329319,0.187411204,-0.022134142,0.0734135732,0.3958806992,0.385607183,0.0828461125,-1.0582391024,0.3952005804,0.4471340179,0.1119966432,-0.0961246639,-0.0695654154,0.6733705401,0.1918676794,0.2059757859,-0.0288748946,-0.0767917708,0.1104539037,0.1275686473,-0.2036572844,0.0659687892,-0.0860861093,-0.0166577548,0.3008368611,-0.0184975974,-0.3902417719,-0.025535984,0.1084865928,0.0096155014,0.0751611888,0.3450949192,-0.5225219727,-0.1504500955,0.2817334235,-0.314238131,0.1200883761,0.1677169204,0.2351465672,0.0198727716,-0.0947349891,-0.0013714206,0.2785582244,-0.0029330628,0.4571766257,0.0140758725,0.0845476538,-0.0685095936,-0.3553752005,0.0135916043,-0.1594937146,-0.0283026639,-0.4566035569,0.1471638829,-0.4094292521,-0.1325123608,-0.2963459492,-0.0165236592,0.0800621137,-0.0948295295,-0.15250732,-0.0411940999,0.2538252771,0.0888529047,-0.4279693067,-0.2591484189,0.2284206301,-0.2501128316,-0.0765269697,-0.1794051677,-0.1035007462,-0.0624710321,0.0933521017,0.0121554593,0.2956854999,0.1202921867,0.1410046667,-0.4283608794,-0.1182050705,0.0054063243,-0.1314729303,0.0446436219,0.1953770965,0.1535257101,0.191829294,0.0539567173,0.0574833304,0.0092918808,-0.0366963334,0.1131214499,-0.4322033525,0.1673461497,0.4726242721,-0.1572847366,-0.3136512339,-0.218298316,-0.3373828232,0.0825211182,0.1418906897,-0.1249805465,0.3699382544,0.1498870701,0.1845361292,-0.0177838095,-0.0911418125,-0.0760118142,0.344949007,-0.0636730939,-0.2971725762,-0.1309769303,-0.2379546762,0.1214394495,0.2287795842,-0.3549236953,0.1684113592,-0.1870066673,0.3227963746,0.0330061279,-0.0170318428,0.3250388801,-0.1784799695,0.1134302318,-0.1828920245,-0.2321367115,-0.0109383613,-0.1593890637,0.3865181208,0.4134480357,0.4659550786,-0.2001300901,0.9793554544,-0.0141021693,-0.3777912557,0.5141610503,-0.1381682456,0.1083688661,-0.1941893399,-0.2245259285,-0.0368074775,-0.1004952565,0.3606392443,0.0945610106,-0.1277509779,-0.1838702112,0.1019632369,0.1188513041,-0.134158954,-0.0463943556,0.2515123188,-0.8417859674,-0.1052167714,0.0155228293,-0.0152480295,-0.2573592067,-0.0736730248,0.0508027673,0.2974078953,0.2227784395,0.2117643356,-0.4028938115,0.1966193318,-0.2293698639,0.438888967,-0.0812731683,0.4864448905,0.0975403488,-0.0795007944,0.0515801683,-0.0645755306,0.3845900595,-0.1764783859,-0.1200195774,0.044498302,-0.0103808688,-0.4325417876,0.0031332045,0.0112128798,-0.0199566986,-0.0984150469,0.1689682603,-0.1514193118,-0.365614444,0.0058574942,0.0054320022,-0.0978402942,0.3364621103,0.0604638681,-0.2607177794,0.0461719632,0.0432053357,-0.0223428383,-0.0855612978,-0.1329709589,-0.2263627499,-0.4594130218,0.0577057824,0.2551597357,-0.2010529488,0.2151525915,-0.1025263146,-0.0914755464,-0.2516608834,0.3307195306,0.2434736043,0.3105360866,-0.2324685156,-0.1777080595,0.0115964161,0.329194665,0.3991323411,0.1161225662,-0.048603978,0.2798422873,0.0726907402,-0.108251214,-0.3122854829,0.1556276828,0.3349766135,-0.006093238,-0.5391897559,-0.4697954357,0.7293050885,0.2588543594,0.1164349988,0.1840459704,0.2263998687,-0.2594000399,0.6836552024,0.2476070374,1.1105152369,-0.1505958289,0.325476408,-0.0553832017,-0.304946512,0.1974311173,-0.1390781105,-0.0372846946,-0.4860218763,-0.1130753085,-0.1740113497,-0.202886343,0.0212463252,0.1251732111,0.3290598691,0.184454143,-0.0773122832,-0.0389896519,0.0793218985,0.1365746558,-0.0199592486,-0.184709549,-0.0652611479,0.1012168229,-0.2433091551,0.0594144873,0.0162273645,-0.3013008237,-0.1052429453,-0.0508493856,-0.0633065701,0.1296350062,-0.3991996348,0.1822525263,-0.0962809995,-0.7509379387,0.2139521986,0.2963104248,0.3420215249,-0.1596007049,-0.1480835825,0.1683668792,0.457670182,-0.1613238603,-0.0013215624,-0.0938171148,0.2785462737,0.0020784598,0.0218586959,0.0619198717,0.0477983505,-0.3985947967,-0.3425138593,-0.1530277282,0.4711922705,-0.2364471555,-0.5164051652,0.2092851847,-0.0526448563,-0.061727114,0.0075550373,0.1004681215,-0.0716265664,0.0738552958,0.2897489369,-0.3646855652,0.0436401293,0.3805157542,0.2072926909,0.3583519161,0.3787144721,0.1604646295,-0.0877015218,-0.0713327825,0.248845458,0.8871480823,-0.244599551,-0.0749843717,-0.0994515717,-0.0787958577,0.2101045698,-0.1447379291,0.0072239065,0.0007943599,-0.0470401049,-0.1608392298,-0.3713755906,0.0323585756,-0.5008922815,0.2149920911,0.0250551645,-0.085782297,-0.3493961096,0.054962039,-0.1430900097,0.186542958,0.2190351337,0.1205224991,-0.1012581587,0.1273954809,-0.0131586837,0.1512590051,0.0032078139,0.0194382593,-0.1998090297,-0.1410713047,-0.2307722569,0.2048526555,0.1036021262,-0.1494746804,0.0814371631,-0.2535155118,-0.2487972528,0.0392641202,0.1388670653,0.1848515272,0.1014311239,0.2950353324,0.3234407604,-0.0625920221,-0.1849877238,0.0561650284,0.3580034673,0.3486047983,-0.159319073,0.116733402,-0.1748815924,-0.2942559421,-0.3523997664,0.1472955197,0.5240929723,0.0277557429,0.2010703236,-0.0655578226,0.1321330518,-0.2486328781,0.212868154,0.4067132771,-0.0934373438,0.0560362302,0.0784911588,-0.0621011257,-0.3087408841,-0.0584247261,-0.2066038847,-0.3107801676,-0.0911205038,0.0011045972,0.1392413378,0.5642006397,-0.113952212,-0.2999230921,0.0501896255,-0.3612253964,-0.117828764,0.4179739356,0.1900862455,0.0165006742,0.3872236907,0.2478434145,0.3203352094,0.1747997999,0.3979457915,0.8507425785,0.6488271952,0.0659124628,0.4314879477,-0.3954394758,-0.3699268699,0.2533235252,0.1470923722,0.1307246238,0.1713160276,0.5190113783,0.1368676126,-0.5436130166,0.0644952655,0.1760166585,0.0215304829,-0.4329630435,-0.1627655625,0.0446734168,-0.3134481311,-0.14613612,-0.3626264632,-0.1190508083,-0.0226943307,0.1918986738,0.0151507026,-0.0894166231,0.1009465382,0.1629132181,-0.1449505687,-0.0583668426,0.1631779373,-0.1606595367,-0.0456536487,0.0503520817,0.185474202,0.2472784966,-0.0190694649,-0.2948700488,-0.0716006458,0.4638655484,-0.1851460785,0.3130325675,-0.0494706109,0.1871951371,-0.1125678718,0.3059999347,0.0126176123,0.0503679253,0.1502809525,0.0073726559,0.0001854458,-0.0016840299,0.2825617492,0.0821032524,0.0098572876,-0.164013505,-0.0108397556,-0.389570266,-0.0628776401,0.6542888284,-0.2125016302,0.127077952,0.0697131678,0.0010596421,-0.0440519489,0.1833190471,0.3698509634,0.1064888239,-0.3608699441,0.0290924646,-0.2346725911,0.0396539979,0.0235653017,0.1058874875,-0.0842465088,-0.0097637232,0.3654394448,0.0080633387,-0.0615913086,-0.0478135906,0.1948836148,0.5103932023,0.0251611602,-0.1059492454,-0.2576786578,0.3504765928,-0.1483099014,-0.3171130419,-0.0563678518,0.2402727455,-0.0793898925,-0.0301052574,-0.2307637483,0.5451130867,-0.0586870313,0.3227196336,-0.0544054471,0.5014358759,-0.2844734192,0.1069041416,0.1247162521,-0.0837067813,-0.4017588794,0.2189140022,-0.1264355779,0.2732658982,0.0107221026,-0.1001706049,-0.0257066786,0.1373885572,-0.1320918798,-0.4687426388,-0.2303670794,0.4813528955,-0.2396201193,-0.1400437504,0.2907838821,0.469376117,0.2053034157,0.1607203633,-0.1676559597,-0.1560054868,0.5634088516,0.1537971944,-0.0410719961,0.0822807178,-0.1121467501,0.3160589337,0.065918684,-0.3561655581,-0.1133613214,0.1487045735,-0.2926995456,0.1281116605,0.2271982431,0.2410165817,0.2620629966,-0.3164125383,0.4251031578,0.2194572836,0.0516381226,-0.247576803,-0.3496411741]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2165","title":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset","comments":"@y-rokutan Hi, now if you install `deepspeed` from master (this feature will be available in the next official release), the code should work without subclassing. Let us know if you still have any issues.","body":"Hi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n","comment_length":34,"text":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset\nHi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n\n@y-rokutan Hi, now if you install `deepspeed` from master (this feature will be available in the next official release), the code should work without subclassing. Let us know if you still have any issues.","embeddings":[-0.2635490596,-0.3762305975,0.0858724415,0.3691848218,0.1816754192,0.3435536921,-0.0518239327,0.4010279179,-0.1650431603,-0.258449614,-0.3508144617,0.3204566538,-0.2445640415,0.0142735196,0.1966751963,-0.3255868852,0.2107604295,0.0282448214,-0.2117948681,-0.2563457787,-0.2183745801,-0.0660058558,0.0091975173,0.1386984736,-0.0008858729,0.0312691592,0.0924518183,0.1584639698,-0.1116146445,-0.366057694,0.2941603065,0.0659787208,0.6609077454,0.2473991662,-0.000130302,0.1516997963,0.0485581644,-0.0997549444,0.1597805619,-0.0868282244,-0.2239148766,0.0278999675,0.2070776224,-0.1289735138,-0.4146131575,-0.6126383543,-0.0347185321,-0.3231691718,0.1267071068,0.452029109,0.0357436016,0.1441907734,0.0045099747,0.0439588875,0.2477651834,-0.1934699863,-0.1854258925,0.0936603993,-0.0743704736,0.1410143077,-0.2250965089,0.0856280103,-0.1529570222,0.0256859586,0.4744634628,-0.305649519,-0.3226049542,-0.2647331357,-0.0733929202,0.5432499051,0.2157951146,-0.2298208177,-0.1046039164,0.1822216362,-0.055059012,-0.2311099917,-0.3596684635,0.2333776355,0.0170585625,0.130522728,0.0112073291,-0.1299144477,-0.1863081157,0.0875596553,0.2806100547,0.3632269502,0.0851095244,0.1512035578,0.2155201882,-0.2419120222,0.3436709046,-0.2262788415,0.0070464769,-0.1812538952,-0.1346868873,-0.0533127859,-0.3148994446,-0.0849068016,0.378131032,0.1288087815,0.174003005,0.3770180345,-0.1272423118,0.0730314553,-0.1347578317,-0.1246929839,0.2171281427,0.2373773903,0.0385422036,-0.2505725026,-0.0205379482,0.2715170383,-0.1502734572,-0.2897736728,-0.0468763784,0.0514920764,-0.1515406519,-0.176536873,-0.1665547341,-0.1864985526,-0.3177583218,-0.0514321253,0.0475088358,0.2343060076,0.0550113358,0.0473903492,0.2683138549,0.6924920082,-0.2033870816,-0.1185874566,0.0322334804,0.1174828932,-0.0959676951,-0.1980102062,-0.0800251067,-0.1564632803,0.0190549064,0.0263228212,-0.3015548289,0.1301091909,0.3758248091,-0.0996755585,0.2288138419,0.0459041074,0.0318006389,0.073764205,0.3789966106,0.4413585663,-0.3008752167,0.2715044022,-0.4328229725,-0.2426431626,-0.1156899855,-0.0130986301,0.1889711767,-0.2756026387,-0.4165625274,0.3720363975,0.0163513198,0.1343854517,0.010878602,-0.4601106048,-0.4295722246,-0.1702985466,0.2251150608,0.1835807115,-0.4052045047,-0.1438096017,0.062004108,-0.1478810906,0.2841463983,0.2974840403,-0.1768792123,0.016512746,-0.3843535483,-0.0178858098,0.5604600906,-0.2282059491,0.1074003354,0.1166640818,0.1041468829,-0.219372645,0.1730726361,0.0408265479,-0.0429625809,-0.0621203855,0.122775875,0.222984612,-0.121819593,-0.2111681253,-0.104038842,-0.2291263938,0.4387815595,-0.0603312403,0.2165647894,0.3418134451,-0.1543974578,-0.2581012845,0.3750075698,-0.0382846296,0.1416975558,0.2198246866,-0.2698795497,0.1213159412,-0.0675660819,0.0889883339,0.2008627802,0.0510114655,0.0607858002,0.0659811869,0.1656821668,0.0506542847,-0.1664160639,0.1074292883,-0.1352767646,0.0021145626,-0.1062234044,-0.0797208324,-0.2741826177,-0.2691041529,-0.090402782,0.2340280414,-0.4500222206,0.1428385675,-0.5754755735,0.161221534,0.1305392385,0.1429700106,0.0562027805,0.1802558452,0.0934509635,-0.0646034479,-0.1435524672,0.2007333338,0.048501309,-0.0976667851,0.3783784807,0.2502427399,0.0699367747,-1.0497536659,0.3431996405,0.4532094896,0.0923509821,-0.1317552924,-0.0381977931,0.6848375201,0.2695319653,0.2214422524,-0.0369539931,-0.0438535325,0.0043841512,0.1337317079,-0.2001353651,0.0169994719,-0.1230770424,-0.0709825307,0.2662944794,-0.0146301854,-0.3610474467,-0.0365428105,0.0466169044,-0.0357978307,0.1359192878,0.3326887786,-0.5043021441,-0.0583159514,0.2491913438,-0.1609594822,0.1241280884,0.1438811123,0.2088501453,0.0072885808,-0.101140283,0.0282054152,0.2873184085,0.0236929804,0.377846688,-0.0637540445,-0.0188888684,-0.0708828643,-0.3746263683,0.0985848084,-0.1449115276,-0.0693301633,-0.4895896316,0.1309631765,-0.4484529495,-0.071146965,-0.2224132419,-0.0177480634,0.0441772714,-0.049512472,-0.1779140681,-0.0210448932,0.2825018764,0.1720361263,-0.4338253736,-0.2607576847,0.2005131692,-0.2989144623,0.0215104632,-0.143920064,-0.1155608967,-0.0867560953,0.129149735,-0.0038978767,0.2609847784,0.1594065875,0.0566514991,-0.2899437845,-0.1649594158,-0.02743572,-0.2137408108,0.1000065282,0.1556865275,0.1725544333,0.1556279212,0.0278074201,0.1050844342,0.0160122514,-0.0390340053,0.0990422294,-0.3585253656,0.1676232368,0.4255625308,-0.1689360291,-0.2158085406,-0.235413447,-0.3735868931,0.030824298,0.1254537106,-0.1326868385,0.2924011052,0.0698174909,0.2185610384,0.0046675834,-0.0764476508,-0.0743279159,0.3837733269,-0.0212695897,-0.3230701983,-0.1505278647,-0.2576108873,0.106747739,0.154496491,-0.315159291,0.1952720284,-0.2176488936,0.2480791807,0.0706432834,-0.1014247611,0.3441767693,-0.1542800218,0.141318664,-0.1726019233,-0.2886224389,0.0019523023,-0.1891165525,0.3890924156,0.3815199733,0.5699663162,-0.186128214,0.9883284569,-0.0373085476,-0.3312836289,0.5028914809,-0.0849152431,0.0929300338,-0.1694969386,-0.3107759058,-0.0327687375,-0.1209060475,0.3107076883,0.1139407754,-0.1250899583,-0.1988011897,0.1134589836,0.167877242,-0.0181296971,-0.0517735481,0.2330697328,-0.8297374249,-0.0912510902,-0.0216027834,0.1228052303,-0.2273715883,-0.0501350164,0.0121529875,0.3278545439,0.1979079545,0.216331467,-0.3806886077,0.1577455699,-0.2193287909,0.4085028768,-0.0704607219,0.4912365973,0.09337423,-0.0949968696,0.0861580074,0.0220691599,0.3633362651,-0.1793453097,-0.0308415424,0.1117618531,-0.1261214763,-0.4755630493,-0.0558814518,0.0296045206,-0.0462330393,-0.0690710396,0.0467016958,-0.1427495033,-0.4097144306,0.065143317,-0.1442108899,-0.0870091319,0.2494536191,-0.0412376523,-0.2178957611,0.0369093977,0.077864632,0.0283411946,-0.0556211919,-0.152257055,-0.2364598811,-0.4242433012,-0.0366366096,0.3029746413,-0.235510096,0.2133246958,-0.0416630022,-0.0484553427,-0.2323370427,0.3077423573,0.2700440884,0.2854973972,-0.2015432268,-0.1715502441,0.0880808234,0.3263977468,0.4192618728,0.1228455976,-0.1259831339,0.2496310174,0.085406214,-0.1699382216,-0.3876940608,0.1204575077,0.3908457458,0.0005144631,-0.5109210014,-0.5484149456,0.7228651643,0.2101601064,0.1095934734,0.1810742319,0.1794039607,-0.2531594336,0.6627234221,0.2255906463,1.1227077246,-0.1567255706,0.3088030815,-0.0989400744,-0.2254077196,0.1381103992,-0.0955765471,0.0803964809,-0.4583054483,-0.2231978029,-0.2168099433,-0.1797730476,0.0752980933,0.1701574475,0.3530699611,0.1755508482,-0.1129942834,0.0249556676,0.1306478977,0.1007460579,-0.0128154438,-0.2109470367,-0.0458246842,0.0714127868,-0.1820750237,0.0481949486,0.0235543493,-0.3256043494,-0.0910585895,-0.0490233265,-0.0971552357,0.143970266,-0.3985098302,0.2058293521,-0.1375064254,-0.7237468362,0.3812147975,0.3280155957,0.3723204732,-0.16270262,-0.0940276906,0.1356170923,0.3901247382,-0.1575114578,-0.038642142,-0.1192772537,0.2890608907,0.0428449288,0.010075978,-0.0111499727,0.0320055,-0.410613507,-0.3022831678,-0.1543992609,0.4988342822,-0.2720614672,-0.4807466269,0.1492034495,-0.0385665372,-0.0865035206,-0.0084987134,0.040865656,-0.1314422637,0.0605041534,0.2716684341,-0.3586280644,0.0340238437,0.4236020744,0.2787997425,0.4550118446,0.3825232089,0.217340976,-0.031176623,-0.0744189247,0.3133806288,0.9300650358,-0.1052509397,-0.0382367186,-0.0390760303,-0.1128386557,0.2114227712,-0.1719279885,0.0587480552,0.0183199085,-0.0646202043,-0.1465440542,-0.3974838257,0.035099525,-0.5616660714,0.2371178567,-0.0131261572,-0.0443713926,-0.2860120535,0.1106179655,-0.1105932444,0.1045890003,0.1652841419,0.1090122536,-0.0535767004,0.177802071,0.0395174176,0.1697932482,-0.0288619362,0.0214286074,-0.2437858433,-0.0856211185,-0.2043776512,0.2034528702,0.1092087328,-0.0705147609,0.1099070087,-0.2208592594,-0.3213769495,0.0353788324,0.1673647761,0.1290812492,0.0602772757,0.3388198018,0.3320457041,-0.0992205366,-0.2505174577,0.0117371287,0.2476679832,0.3977492154,-0.1052652597,0.0479069017,-0.1343837678,-0.2938901484,-0.3598111868,0.128236711,0.4391294122,0.0138453627,0.240182519,-0.1184680983,0.1339882165,-0.209809199,0.2071189284,0.3725752831,0.0018703699,0.0330054499,-0.0609828345,-0.0930862501,-0.3016681075,-0.042255614,-0.1800869405,-0.3002391458,-0.0902887061,0.028997248,0.1032248586,0.553503871,-0.1385714859,-0.2539667487,0.0602764785,-0.3215013444,-0.0014764753,0.4246539772,0.1985871643,0.0489681214,0.4345265627,0.2401871234,0.3080203235,0.2676519454,0.4743523896,0.8348222375,0.5884381533,0.0774723738,0.368068099,-0.4591241479,-0.2820629179,0.3882302046,0.1568257958,0.0790390372,0.0688012093,0.5054250956,0.0001397798,-0.6188648939,0.0159465652,0.0685019419,0.0143842138,-0.3718113303,-0.2234374732,0.0995883197,-0.2777837813,-0.1721456349,-0.3405427635,-0.1036872864,-0.0677410066,0.1796111166,-0.0009894598,-0.0781408697,-0.0416699424,0.1533586085,-0.1269313991,-0.0614897422,0.1845899224,-0.2294329405,-0.0330865011,0.0950707197,0.1865214109,0.1773936152,0.0341180153,-0.2639219165,-0.0837148353,0.4728598595,-0.1462549567,0.3952938616,-0.057339482,0.224815622,-0.031453006,0.3200482428,-0.0349097066,0.0801534429,0.0922416672,0.0229728427,0.043974489,-0.0226472113,0.3486649096,0.0821723938,0.0068174186,-0.1428256035,0.0078708846,-0.3138804138,-0.0179931931,0.6848994493,-0.2469412237,0.0772482306,0.0536717959,-0.0274706818,-0.0573344231,0.2777770758,0.310070008,-0.0601119101,-0.3868315816,-0.0088447081,-0.2415472269,0.0332234316,0.0033746113,0.1420803964,-0.0431406796,-0.0373945236,0.4060946107,-0.028631378,-0.0117791193,-0.032829944,0.1235852763,0.5267846584,0.0577883497,-0.0698453411,-0.2128749192,0.3947785199,-0.1954042763,-0.2832946777,-0.0827915221,0.1895259768,-0.1277540922,0.1008323729,-0.2137658298,0.5065088868,-0.1021379381,0.3509518504,-0.0359173492,0.5236024261,-0.25780797,0.1195465103,0.1606182456,-0.0764293298,-0.4228771627,0.2357124239,-0.1327616721,0.3454509377,-0.0522755198,-0.1511160284,-0.0333356559,0.1769847572,-0.0893204436,-0.4472331107,-0.2921124697,0.543843329,-0.3005524278,-0.0901238695,0.3235526085,0.4553418159,0.1936599165,0.146396026,-0.1793061048,-0.1881177574,0.5304230452,0.0859737024,-0.088576667,0.0214537028,-0.0912607908,0.2373923808,0.1098616198,-0.24356848,-0.1449953169,0.0951248109,-0.1903585792,0.0469833054,0.2565312684,0.2410611063,0.213490814,-0.2993459404,0.3301889598,0.1673750579,0.0171649866,-0.2032989562,-0.3716502786]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2165","title":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset","comments":"Worth mentioning that any function that expects a `torch..Dataset` (like `torch..DataLoader`) will fail a mypy-esque typecheck if a `datasets.Dataset` is passed, even though it implements the interface correctly (I think). The virtual subclass idea was a good one- I wonder if there's another workaround given the Generic issue. What we're really talking about is something similar to the structural subtyping semantics that `typing.Protocol` defines. If `torch..DataLoader` accepted anything that supports `__getitem__` and `__len__` methods this would be much easier. Not sure if there's a way to do this without the wrapper from the perspective of `datasets`.","body":"Hi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n","comment_length":96,"text":"How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset\nHi,\r\n\r\nI'm trying to pretraine deep-speed model using HF arxiv dataset like:\r\n```\r\ntrain_ds = nlp.load_dataset('scientific_papers', 'arxiv')\r\ntrain_ds.set_format(\r\n type=\"torch\",\r\n columns=[\"input_ids\", \"attention_mask\", \"global_attention_mask\", \"labels\"],\r\n )\r\nengine, _, _, _ = deepspeed.initialize(\r\n args=args,\r\n model=model,\r\n model_parameters=[p for p in model.parameters() if p.requires_grad],\r\n training_data=train_ds)\r\n```\r\nbut deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?\r\n\nWorth mentioning that any function that expects a `torch..Dataset` (like `torch..DataLoader`) will fail a mypy-esque typecheck if a `datasets.Dataset` is passed, even though it implements the interface correctly (I think). The virtual subclass idea was a good one- I wonder if there's another workaround given the Generic issue. What we're really talking about is something similar to the structural subtyping semantics that `typing.Protocol` defines. If `torch..DataLoader` accepted anything that supports `__getitem__` and `__len__` methods this would be much easier. Not sure if there's a way to do this without the wrapper from the perspective of `datasets`.","embeddings":[-0.2078215033,-0.2331490517,0.1384682655,0.2725795507,0.2668266296,0.1537621021,-0.0031080558,0.3240680695,-0.0748968348,-0.2203996629,-0.3576208353,0.3906598687,-0.2636044323,0.105207786,0.2797994614,-0.2113587856,0.1772631854,-0.0501310565,-0.1659512818,-0.1595477462,-0.177148059,-0.1184021086,0.0547709689,0.0935725495,0.1425067335,0.0650051683,0.0786012039,0.2379954904,-0.1792033017,-0.2768820822,0.3760774732,0.1968694329,0.4361511767,0.1904756278,-0.0001295324,0.2556096911,0.1330806464,-0.1514336169,-0.0350119844,0.0231361929,-0.0956846103,-0.0058039296,0.1660961956,-0.0608776361,-0.263053596,-0.6010507941,-0.0813809037,-0.3644055724,-0.0128699131,0.3584641814,0.0206217598,0.3547851443,0.0011108372,0.0559762754,0.2998205423,0.0063132565,-0.1517211646,0.0891334862,-0.1101552844,0.1564475596,-0.2003160864,0.017144179,-0.1828868091,-0.0504500717,0.4789533019,-0.2569850683,-0.411652118,-0.1110196561,-0.1148878634,0.5390319228,0.1896052957,-0.3263989389,-0.2952817082,0.0647178367,-0.1126721874,-0.1941776574,-0.2395136654,0.0374426469,0.0168021489,0.1477460563,0.1189094782,-0.0822050944,-0.1484340727,0.1507809758,0.1586577445,0.189727515,0.0554301552,0.2007675916,0.1981996745,-0.3664046228,0.3242422044,-0.232834965,-0.0055380194,-0.1253769249,-0.018496098,-0.1147717088,-0.2362650335,0.0687458962,0.4641350806,0.0720357373,0.1962969005,0.4362742901,-0.3657406569,0.1652643383,-0.0035399643,-0.1884425133,0.1268772334,0.3190458119,0.0136149088,-0.1865618527,0.1171434075,0.2503778934,-0.0632261857,-0.1424945146,-0.0214009043,0.0687717944,-0.0095397802,-0.1376754642,-0.1544179618,-0.2173983157,-0.2425918877,0.0071896617,0.1731628776,0.1989182383,0.0464286655,0.2016894817,0.3246091306,0.5662387609,-0.2975242734,-0.1407991648,0.0386501402,0.0852948204,-0.099816747,-0.2087254822,-0.0367854573,-0.1110672727,-0.1289435029,0.0083112745,-0.0774281099,0.2705898583,0.2934509814,0.0112030096,0.2924002409,-0.0088668102,-0.0506427325,0.033993464,0.3592982888,0.301910013,-0.3759589195,0.1356443316,-0.4547368288,-0.1987839192,-0.0828806162,-0.0187599361,0.0401504599,-0.2016354203,-0.3906699419,0.2585842013,-0.0183183569,-0.0244549625,0.0184878334,-0.4866313636,-0.3768112063,-0.2238227576,0.2163504511,0.188343659,-0.4176258147,-0.1818936318,0.0440067463,-0.198169589,0.1914093047,0.289255172,-0.2226304263,-0.0521310382,-0.3271151185,0.0597645231,0.5643676519,-0.1906437278,-0.0303818788,0.1291262209,0.1200063378,-0.003957768,0.0622874126,0.0052822344,-0.0918734893,-0.1014146954,0.219894439,0.2000405341,-0.1793326139,-0.2974882126,-0.130935818,-0.2400948554,0.2908289135,-0.1476475745,0.1456648707,0.2855872512,-0.1942010373,-0.2734147608,0.3236292005,-0.0721281692,0.1452775002,0.0780832916,-0.289785862,0.3068268895,0.0280978065,-0.0039131898,0.1166118532,0.1188362762,0.0460418984,0.0801185444,0.0156897474,-0.1031302959,-0.128483668,0.1592297107,-0.0084317997,0.1284552217,-0.0995795503,0.0432025939,-0.3955700994,-0.2482409775,-0.1125764847,0.080000706,-0.3220640123,0.1985419691,-0.616733849,0.1133341938,0.1474821866,0.1528816372,-0.082584694,0.2855966389,0.0134976311,-0.0205444321,-0.0565258302,0.2021771818,0.2115097344,-0.0618809462,0.1903974563,0.3349079788,0.0072285938,-0.840179503,0.3379893899,0.415142566,0.1432085782,-0.1016327068,-0.0548334569,0.7059512734,0.260623306,0.2426980883,-0.0817961991,-0.012058422,0.0649009347,0.1966552585,-0.3872092068,-0.0461578555,-0.2101971507,0.0573000461,0.2363507599,0.1334766895,-0.3134972453,0.0263527222,0.1032713428,-0.1176700667,0.2139063478,0.3229413927,-0.3895614743,0.0043239971,0.2237343639,-0.155978471,0.178251341,0.1281371415,0.2365185022,-0.0146142421,-0.1825609654,0.0960272625,0.2678771913,0.1079960316,0.2084406018,-0.1348013878,-0.010651906,-0.0330474526,-0.316788733,0.1150294915,-0.2100258321,-0.2024738938,-0.6494216323,0.2454880923,-0.644500792,-0.1679469794,-0.2117986679,-0.1283808202,0.0738043562,-0.1325402856,-0.1838198006,-0.1610394269,0.1574227959,0.1804305613,-0.4207696915,-0.087619856,0.1453914642,-0.3617405593,0.1635016203,-0.1736216247,-0.1349204183,-0.1402754635,0.1672249287,-0.0116370115,0.3143476546,0.2745062113,-0.0749648735,-0.2894769907,-0.2304533869,0.156988278,-0.1682230383,0.1051212847,0.4179945886,0.1087098941,0.2576506734,-0.0116583053,0.135540545,0.0952822044,-0.1142037585,0.1711758375,-0.3422296047,0.1466582417,0.2856680751,-0.0005159274,-0.271135062,-0.2474175394,-0.1315757483,-0.1056080535,0.1040342152,0.1052230671,0.2524413764,0.0772711784,0.1599101573,0.0024250087,-0.1160770804,-0.1452409625,0.2762694359,0.0449747592,-0.1961497366,-0.2027565241,-0.3262172937,0.1203276515,0.3483853638,-0.1995397061,0.1215454936,-0.088067241,0.2964408696,0.1406215578,-0.0328138955,0.2701464891,-0.0281772967,0.119058691,-0.2390896678,-0.2776736915,-0.0470402837,-0.3374869823,0.2847621143,0.5173725486,0.4786030352,-0.1577197015,0.8596717119,-0.0879080296,-0.4088988006,0.4252443612,-0.1249101609,0.1847854853,-0.1103778034,-0.4013188481,-0.0969165787,-0.2668886185,0.239632979,0.0914087966,-0.1537971497,-0.2068269551,0.1639364362,0.1523761451,-0.0264299698,-0.261315316,0.4399176836,-0.8675211072,-0.1050403044,-0.0578205176,0.0795483366,-0.3812004924,0.0011201695,0.0686211362,0.2858522236,0.1166149303,0.1581166536,-0.3939729333,0.1059401929,-0.1882931292,0.4757519662,0.0191639587,0.5118653178,0.0879123956,-0.1726588756,-0.0696582943,0.0268167816,0.2896457911,-0.2780473232,-0.0805945173,0.1641328633,-0.039040003,-0.5347517133,0.0066226781,-0.0439321063,0.1003816202,-0.0451431125,0.1103137508,-0.2395056337,-0.3378883004,0.0294973124,-0.0553753488,-0.0326881148,0.2306826115,-0.1315283328,-0.2091811746,0.0484417714,-0.0030786924,0.116532363,-0.1095057726,-0.1023848504,-0.2505788803,-0.413895905,-0.0436236709,0.3718596101,-0.2287891805,0.2082284242,-0.0586988516,-0.0537223667,-0.2406348586,0.4764309824,0.2851330638,0.3794768155,-0.3600588143,-0.3047217131,0.0356576592,0.3138700426,0.4174485505,0.187698096,-0.1444206238,0.3934052289,0.1332859695,-0.2115071565,-0.5201492906,0.1842911988,0.332847774,-0.0139124114,-0.4928371906,-0.6445838809,0.7689057589,0.2218127847,0.0544247739,0.3121086657,0.3377504051,-0.3339827359,0.6794016361,0.2577805221,1.0996062756,-0.0346682854,0.3274339437,-0.0318780169,-0.1232236177,0.2990600765,-0.0514365733,0.037419036,-0.3626796901,-0.192016229,-0.231072247,-0.1870166361,-0.0066353763,0.1129599884,0.1887165606,0.1578836143,-0.2065266073,0.0865199342,0.1863582432,0.1434390694,-0.1430522799,-0.3133465648,-0.0725741461,0.0737974048,-0.2018469423,-0.0247666053,0.0696131736,-0.2688633502,-0.1569671631,-0.0939245746,-0.1439598948,0.0402995385,-0.4678852558,0.1700505465,0.0067252028,-0.4388856292,0.3773150742,0.2177981436,0.3210292161,-0.084062919,-0.1062156931,0.0549592599,0.3936531544,-0.0869994536,-0.0647535771,-0.1884710789,0.4313010573,0.1897947788,0.0387761109,-0.1223741919,0.0134799806,-0.265140444,-0.3395312428,-0.1420839131,0.5508438349,-0.4496554434,-0.3465381265,0.0302718412,0.0008457564,-0.0114150653,-0.0343917497,0.088135913,-0.0147055201,0.0936475918,0.2609142065,-0.3039735854,0.0750020891,0.3815479577,0.1440957785,0.3418433964,0.3346377611,0.1366581619,-0.0858721808,-0.0354809575,0.3121761084,0.9939111471,-0.1061718315,0.0196709055,-0.131787166,-0.2180085629,0.0851488113,-0.2070150375,0.1507610977,0.0491243154,-0.1872912943,-0.1722064763,-0.3375902474,0.2550194561,-0.3870527148,0.298512727,-0.0794733688,0.2165443003,-0.3035988808,0.0311469305,-0.134667486,0.0342314541,0.1797302812,0.0705113783,-0.0291183293,0.2059076726,0.0574300922,0.0696485564,-0.0349054448,0.0624160655,-0.1884785593,-0.0458090119,-0.2914164066,0.2247946113,0.0019303944,0.0069705406,0.2205756903,-0.1755093038,-0.2088712454,-0.1003458723,0.0976725966,0.0996841043,0.0047066817,0.3070245385,0.3442031145,0.0926198661,-0.1428008527,0.0978469998,0.337038219,0.4264219403,-0.0894038007,0.0438061357,-0.2029936165,-0.2533398271,-0.1870264262,0.3083278537,0.5461503863,0.0302509759,0.1883122474,0.0196390543,0.0237200074,-0.2299456298,0.2487044483,0.5658827424,-0.0058658617,0.0509624109,-0.0675119534,-0.1053529233,-0.2270379663,-0.0321216099,-0.0863212124,-0.3270879984,-0.0131164137,0.1574241221,0.016194487,0.5776231289,-0.1346444786,-0.3196704388,0.0066970359,-0.2881914377,-0.0220321268,0.5073307753,0.0964724123,0.0737790689,0.3497420847,0.1650417447,0.3461249173,0.2299318463,0.4368546605,0.7877060175,0.461609602,0.1401689053,0.4382709861,-0.5139477253,-0.195902124,0.4214673936,0.2564806342,0.2017282546,0.0750500336,0.6161073446,-0.0600150637,-0.4812819362,0.1460745186,0.1253003627,0.000547051,-0.4632918835,-0.337798804,-0.0295960549,-0.4231980145,-0.2846585214,-0.3085650206,-0.1266383231,-0.024193516,0.0743197575,0.0771802962,-0.1394361109,-0.0729415268,0.0746720657,-0.0849191844,-0.1124606952,0.1132898331,-0.1118329391,-0.0214532856,0.1088947728,0.1355242878,0.2491071522,0.0923753306,-0.2736178041,-0.0808614939,0.347168982,-0.1542700678,0.2459336072,-0.1530474573,0.0572599508,-0.0728862658,0.3427268565,-0.0630103424,0.0821188688,-0.0155275697,-0.0536864027,0.0572774634,-0.0625022873,0.554709971,0.1199889034,0.0866110399,-0.1235129535,0.0593173616,-0.4033994675,-0.1010095403,0.6918380857,-0.0552567393,0.155853942,0.0879931748,-0.0238864403,0.0356972367,0.3571387827,0.2786395252,-0.0012066903,-0.314378202,0.0869773924,-0.2824189961,0.0191213526,-0.0538673885,0.0133895604,0.0472504199,0.0194263086,0.5246506929,-0.0339059383,-0.0880302489,-0.0512899235,0.1709915847,0.3318522871,0.0924044028,-0.2448701262,-0.174009487,0.2448079735,-0.0757258683,-0.3458242714,-0.1614000797,0.310713619,-0.1198501512,0.0009036235,-0.2722069025,0.4618014097,-0.1719902605,0.3024443686,0.0348052233,0.5514158607,-0.240949899,-0.0486092158,0.1155570224,-0.0889051259,-0.4955530465,0.2672635615,-0.1680183113,0.2821672857,-0.0952110142,-0.1633874327,-0.1532776952,0.1078039929,-0.0438831449,-0.4155344367,-0.3930428922,0.5580598116,-0.3865450621,-0.1510735303,0.1841348559,0.5083801746,0.1409491599,0.2737412453,-0.1228882596,-0.2094222903,0.5849864483,0.0536794141,0.0076913075,-0.0119049596,-0.1061060727,0.1365028769,0.1702827662,-0.3620127738,-0.1443612576,0.1439766735,-0.2070527375,0.0495489873,0.1811148226,0.2289016843,0.1970977038,-0.4207523465,0.4470783174,0.1964857131,0.0416078158,-0.1021852493,-0.3120927215]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2162","title":"visualization for cc100 is broken ","comments":"This looks like an issue with the cc100 dataset itself but not sure\r\nDid you try loading cc100 on your machine ?","body":"Hi\r\nvisualization through dataset viewer for cc100 is broken\r\nhttps:\/\/huggingface.co\/datasets\/viewer\/\r\n\r\nthanks a lot\r\n","comment_length":22,"text":"visualization for cc100 is broken \nHi\r\nvisualization through dataset viewer for cc100 is broken\r\nhttps:\/\/huggingface.co\/datasets\/viewer\/\r\n\r\nthanks a lot\r\n\nThis looks like an issue with the cc100 dataset itself but not sure\r\nDid you try loading cc100 on your machine ?","embeddings":[-0.5656793714,-0.2004439533,-0.0850355178,0.1293200999,0.2095071673,-0.0013308327,0.1641536355,0.1368134022,-0.0659237504,0.4231900573,-0.0523652881,0.2376998067,0.1289270371,0.4269973636,-0.0176101644,-0.2411688119,0.0532902479,0.4060421884,-0.3985911012,0.0212525502,0.043411959,0.0523336455,-0.1739806384,-0.0693603009,-0.1812584847,-0.2919968665,-0.0957440063,0.1488842666,-0.2506378889,-0.4322847128,0.060996592,0.1807223707,0.1114192307,0.3614214361,-0.0000948288,0.0520825982,0.3273299932,0.1721242219,-0.1750353873,0.1617924869,-0.3436121643,-0.0873224363,0.2742396593,-0.0548348837,-0.1783528179,0.1911072731,-0.0204064362,-0.0307972282,0.3275450766,-0.2472634166,0.4222187698,0.2349870056,0.0398907512,-0.3050333858,-0.1605335921,0.1931785196,-0.113192156,0.1625883579,0.1607860476,-0.0157570317,-0.0894084945,0.2910957634,0.1937401295,-0.006556401,0.0340785794,-0.1106692553,-0.0367662013,-0.2360448241,0.247420311,0.0474843569,0.4210410118,0.0793646798,0.148647204,0.1815606654,-0.0695406497,-0.6380495429,0.1570040882,0.1009481773,0.0964078605,0.2430051565,-0.3762997985,-0.0677687675,-0.2017343193,-0.0138401808,-0.2421446592,-0.1251512617,-0.2139435709,0.0061438913,0.0715346336,0.0346728414,-0.0787408054,0.2379066795,-0.3031328917,-0.0922413692,-0.2987854779,-0.0159615185,0.1302254945,0.4907619655,0.0583911687,0.1188843995,0.3125003278,0.195605576,0.2119922638,0.2065149695,0.1031227857,0.0831065848,-0.2976173162,0.1005259529,0.4791635871,0.3510072231,-0.0793915838,-0.259262532,-0.116581358,-0.2494165301,-0.1120600998,-0.0014789108,0.0683400854,-0.194302693,-0.2914254069,0.0996225923,0.0628406182,0.183207646,0.0965739191,0.4411553144,0.0086648585,0.1357439011,0.0215744972,0.1259842366,-0.1545628309,-0.2653722465,-0.2535617352,-0.0937746614,-0.2249134183,-0.0452786833,0.1392562985,-0.3511639833,0.1486142576,0.2843308151,0.24389036,-0.0997054949,-0.0786505267,0.0231300183,-0.0841180608,0.2223770916,0.092014499,-0.0371528268,0.1359326839,-0.0663182363,0.3401287496,-0.1389588863,-0.1046525538,-0.1058119312,-0.3097951114,0.3111568689,-0.0393325984,0.0318809152,-0.1348148137,0.0415361747,0.0054447893,-0.0063716546,-0.1271549761,0.0690330788,-0.0485788174,-0.1065335572,0.2176987976,0.2176955491,-0.5703334212,0.0275736544,-0.1413705796,-0.1999806762,0.0043481258,0.1873580962,-0.1815028191,-0.0876335129,-0.1811662763,-0.0983159542,-0.1626347303,-0.2828272283,-0.3876275718,0.1866451055,-0.0017750179,-0.0798946992,-0.1160927862,-0.2375925779,0.2893235385,-0.0153164174,-0.3430396616,-0.2110937834,0.0176128596,-0.0127845118,-0.2409080863,-0.2167172283,0.0838877037,-0.0764978975,0.147901848,0.1105542108,0.1725352705,-0.2935748994,0.2661367953,-0.1485100538,0.1161036119,0.0912747309,0.2704190314,-0.1600845456,-0.1128767356,-0.0416106172,-0.0674103796,0.0204777047,0.0606135689,0.271689415,0.039632272,-0.0088672508,-0.2549166679,-0.0335104018,-0.3341198862,-0.1676502228,0.3803723454,0.092168957,-0.2315768003,0.0005186008,-0.0421158932,0.1758345664,0.0126305791,0.0483755507,0.1761740148,0.3433649838,-0.275255084,-0.063771978,0.1673242152,-0.0719604343,0.2455675006,-0.1013275832,-0.1643158644,0.4407145083,-0.1486031413,0.4009417593,0.1041888744,-0.1629094034,0.3613379896,-0.6228360534,0.110934779,-0.0870218277,0.0760375559,0.1592278779,-0.0845578164,-0.096132867,0.0773143694,-0.0738332346,-0.0045655984,0.1905115545,0.2280788422,0.003077413,-0.0194355231,-0.2421627194,0.1882217675,0.2159336805,0.3058051169,-0.1830739081,-0.3681041896,0.1584586501,0.1430668533,0.1536124945,0.0109324045,-0.0728268772,-0.3498611748,0.0213345792,0.3850712478,0.0080434829,-0.1321379095,0.1849578768,-0.0770443752,0.0816132873,-0.1488431245,-0.0276763439,0.0297795273,-0.0575451069,0.3187336326,0.0401709229,0.0173402485,-0.1541346163,-0.5369300246,0.0395917781,0.0058500143,0.2060564309,-0.2513286173,-0.0886813924,-0.3233671188,-0.1130416617,0.2310751677,0.0394948013,0.0922844857,-0.2377570421,0.0139981238,0.1969210356,-0.1181684583,0.2624606788,0.0012004502,0.213672623,0.0281767324,0.5410407782,-0.2618633509,-0.1442385167,-0.1840663254,0.2897135019,0.0871278495,-0.0790610313,0.1842187643,-0.081015341,0.2908611596,-0.1730745733,-0.3851382732,0.2910663486,-0.1283418387,0.1904563755,-0.0026750055,0.2665259242,-0.3870860636,-0.0203630459,0.1880902201,-0.0413361788,-0.0606841072,-0.0031808945,-0.0159594659,-0.3020131588,-0.2241060138,0.075142391,-0.1735683084,-0.3436998725,0.0384506099,0.1464444995,0.0640092716,0.174707979,0.2289257199,-0.097976543,0.148127839,-0.0595277585,-0.3003802299,-0.5974050164,0.1632573009,-0.5660870671,-0.4630771577,0.1409201324,0.1243423894,0.4092950523,0.0419113412,-0.374995172,0.1733655035,-0.0801885501,-0.0343607329,0.1711721718,-0.1160889268,0.1028820798,0.1523438841,-0.2272927761,-0.1422841102,-0.2734931707,-0.0979778916,-0.1062644497,0.1159405336,-0.2867873311,0.0807190239,0.1047687456,0.4419476986,0.1123191491,-0.0741317198,0.4239270389,-0.0862234309,0.5601354241,-0.1611284167,-0.3377251327,0.1469471008,-0.1507562697,-0.0988627151,0.2195901573,0.0591342784,-0.3533902168,-0.3039927781,0.0354865491,-0.2814315856,-0.2238369733,-0.3538106978,0.0835166872,0.2517737746,0.0645770133,0.2820880711,0.1081139371,-0.1622391194,0.1294884235,0.120091632,0.0353275388,-0.1414877921,-0.4134144485,-0.2049735337,-0.3734684587,0.1895397902,-0.1099827439,-0.0502126366,-0.1996527463,-0.0842040554,0.0514093265,-0.1056896597,0.3423094749,-0.0683498681,-0.0004602485,0.095474124,-0.1732041687,0.0656740814,0.0295124929,-0.1048086509,0.1432252377,-0.1582610309,-0.1050961465,0.0037603052,0.0784186795,0.1847960651,0.0099398009,-0.1699629128,-0.0204203669,-0.0938846469,0.097890161,-0.2449435294,-0.2076078355,-0.2207294405,-0.0100953858,-0.2272652239,-0.266187489,-0.1380919367,-0.3244821429,0.0185235981,0.4938951433,0.3863057792,-0.0652048364,0.086012207,0.2924594283,0.3374376893,0.330375433,0.3128893673,0.1370004117,-0.2071926147,0.2907999456,-0.1313313991,0.0928649157,0.2617779076,0.0430267267,-0.0152969966,0.2100190818,0.2225783318,-0.3074879348,0.1980711967,0.2529595792,0.1679596454,-0.4825790226,0.0580452047,0.3693285882,0.0141713787,-0.0597247481,0.1262945086,0.210071668,-0.0813223273,-0.1351283491,0.1255878061,0.72099334,0.0001049565,-0.2122114748,0.1130356714,-0.1008676738,0.3043750226,0.042439118,0.0194628164,-0.281711489,-0.2667110264,-0.0736100525,0.0370212682,0.2209498733,-0.0685559437,-0.2538509667,0.0554414764,-0.1838659644,0.3419280946,-0.0456507802,-0.0496994257,0.0574332066,-0.0669077933,0.0541138016,0.3630203605,-0.0797085837,0.09391132,-0.2961657047,0.0133893806,0.0031091413,0.0670562387,-0.2429053187,0.0920199007,-0.2276492566,0.1686824709,-0.1923021972,-0.5009307861,0.1603826284,0.0637821779,0.0975629166,0.3254849613,-0.3234696388,0.4022997618,-0.0457815789,-0.1814325303,-0.1833239496,0.3055182099,0.0101979934,-0.1501729637,-0.1779603213,0.1741853654,0.1115857959,-0.0194320455,0.0475308895,0.1179086417,0.1931281537,-0.0222880729,0.1926682293,0.0071358569,0.0033140753,-0.3757091761,0.2785083652,-0.0511808582,0.0666463226,-0.0360996425,0.4344739616,0.0669221953,-0.1159676686,0.0637562796,-0.1367863715,-0.1317863911,0.1777618527,0.2601834834,-0.2765630186,-0.3988590539,-0.0656845197,0.1571887136,-0.2817639709,0.0723021999,0.0805480182,0.0077480683,0.0140639627,0.3513558507,0.1392953694,-0.1685831696,-0.072309114,-0.2136258483,-0.4982856512,-0.0773467943,-0.2293803841,0.2354177684,0.1222251803,0.2731162012,0.0696441829,0.0219165217,-0.4724562466,-0.0437902845,-0.1993090212,-0.0929690972,0.1779615879,-0.0277194474,0.0122174826,-0.1333376616,0.2816725671,0.2427574396,-0.1752961576,-0.3853913546,-0.0153475031,0.021483006,0.0950367451,-0.3926620483,0.1759342104,0.2973404527,-0.0333288014,-0.0578467101,0.0739079118,-0.1647577733,0.0900411308,0.0186314099,-0.2337357998,-0.0474628694,0.0239735395,-0.1795483828,0.2719549537,0.2520774603,-0.0699139908,-0.0998038724,0.1286891103,-0.1220529079,0.1165040061,0.166445747,-0.0365914516,0.195877865,0.4097904563,-0.2700436413,0.267236501,0.0104015451,0.3953765631,0.0123087456,0.0397565924,0.0226721168,0.1892243326,0.4355754852,-0.2832293808,0.1362297833,0.0024187684,0.0065332432,0.2610806823,-0.1032957509,0.2708553076,-0.1254646629,-0.061753232,0.0812321231,0.3503269255,-0.1907881796,0.3412300646,0.4016758204,-0.0528335012,0.1506222337,0.0642256811,0.2355763614,-0.0181958675,0.4345915616,-0.16188851,-0.0178998653,0.1488087475,0.1640518904,0.2202476412,-0.1720549017,0.0562382489,0.1335775554,-0.1476082057,-0.0115501005,0.1621236503,0.4196664691,0.0223253332,-0.3306180537,-0.3100748658,0.1664960533,-0.1759346724,0.1114474386,-0.4876584113,-0.1929874569,-0.0160589367,0.1783457994,0.0034387344,0.0139339194,0.3328846991,-0.034872219,-0.0487208366,-0.5029998422,0.0837433562,0.2177887857,0.2639810741,0.0242695939,0.2061741501,0.2265587747,-0.0002994029,0.1532242447,0.4285952747,0.3149451613,0.067712076,0.0359154046,-0.0532670543,0.108960405,-0.2051901966,0.027342597,-0.1149929315,0.212964341,0.0950714722,0.1885064244,0.2647565603,-0.2827623785,0.0825788677,-0.0216897465,0.2163794488,-0.1022081524,-0.3874949515,-0.2982395887,-0.0977330655,-0.0861080438,-0.3890854716,-0.3332912922,0.1323380321,0.2015293092,-0.2986508906,0.0781787485,-0.3521553278,0.1648675054,0.0640510768,0.4023211002,0.243168205,0.2007891536,-0.1250182092,-0.4483349621,-0.3504098654,0.1526513994,0.1033011079,-0.0398853607,-0.3057733774,0.051186014,0.0918323025,0.1235852838,0.4952954948,0.000179103,0.1944515109,-0.0594194978,-0.2752925158,0.1799764037,0.0094617642,0.0195162296,-0.1272803247,-0.0661201403,0.258279562,-0.0021932588,0.2210864127,-0.1490803361,0.2031294852,-0.4349627197,-0.2029470503,0.3208516836,0.1792405397,0.3766888976,-0.1100781411,-0.0005273503,-0.0858119205,-0.2289458811,-0.0935204923,0.2492439598,0.2223969102,0.3229876757,0.1318791807,-0.1263436675,-0.0290869307,0.049321793,0.2770048678,0.3405130208,-0.2955278456,0.1076351404,0.0787425414,-0.227883175,0.2594730258,0.1059644744,-0.0409698449,-0.0755646676,-0.1921973526,-0.3657118678,0.1892424226,-0.260374248,-0.3046204746,-0.0842138305,-0.0746047348,0.3660804033,-0.085108161,-0.2530287206,0.0696015358,0.073270157,-0.1121167168,-0.0957397595,0.3657014966,-0.0892180279,-0.0652138665,0.1243475229,0.5360304713,0.3529991508,-0.188002333,0.0473649092,-0.0968632847]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2162","title":"visualization for cc100 is broken ","comments":"Hi\nloading works fine, but the viewer only is broken\nthanks\n\nOn Wed, Apr 7, 2021 at 12:17 PM Quentin Lhoest ***@***.***>\nwrote:\n\n> This looks like an issue with the cc100 dataset itself but not sure\n> Did you try loading cc100 on your machine ?\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","body":"Hi\r\nvisualization through dataset viewer for cc100 is broken\r\nhttps:\/\/huggingface.co\/datasets\/viewer\/\r\n\r\nthanks a lot\r\n","comment_length":80,"text":"visualization for cc100 is broken \nHi\r\nvisualization through dataset viewer for cc100 is broken\r\nhttps:\/\/huggingface.co\/datasets\/viewer\/\r\n\r\nthanks a lot\r\n\nHi\nloading works fine, but the viewer only is broken\nthanks\n\nOn Wed, Apr 7, 2021 at 12:17 PM Quentin Lhoest ***@***.***>\nwrote:\n\n> This looks like an issue with the cc100 dataset itself but not sure\n> Did you try loading cc100 on your machine ?\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","embeddings":[-0.4892333746,-0.25718835,-0.0154404212,0.1678069085,0.2279232144,0.036741484,0.1436953247,0.0804585591,-0.0384785756,0.3180071712,-0.1425651461,0.2456283122,0.141941458,0.4538696408,0.0949068367,-0.320658803,0.1359129101,0.3763961196,-0.4416952729,0.0868928432,0.0558291823,0.103545703,-0.1063319966,-0.1049200669,-0.1652607024,-0.2434975803,-0.1958783269,0.2551199198,-0.2148875594,-0.4934449494,0.0909606665,0.2749294937,0.1034732461,0.3986742496,-0.0001022205,0.049955342,0.3971211314,0.1768862605,-0.2313628942,-0.0155357495,-0.3341015279,-0.0857342929,0.3056445122,0.0706033558,-0.0899706483,0.1896803677,-0.0461142957,-0.0780905783,0.4632003605,-0.3263448775,0.3314780593,0.1898190677,0.2163901627,-0.23562482,-0.2317939252,0.3152169585,-0.1130151227,0.3244764805,0.1339665651,0.0188412573,-0.1388449818,0.281008184,0.1435005665,-0.0581777468,0.2358365953,-0.1443402469,-0.1305725127,-0.2075549364,0.131921187,0.1367862076,0.4491756856,-0.0257950295,0.0039736619,0.1525943726,-0.1334011406,-0.7048443556,0.2510020733,0.0759367272,0.1342730075,0.2552893162,-0.4646982849,-0.0914586186,-0.1806110889,-0.0053140786,-0.2083681375,-0.2731879056,-0.2704397142,0.0497461669,0.084737502,0.0378440917,-0.1317539513,0.2829217315,-0.3070276678,-0.0304717589,-0.332672745,-0.0233380254,0.0595650822,0.608674407,0.1661481112,0.1431820542,0.262011677,0.1786367744,0.1625328362,0.1637374759,0.1870710552,0.1150897145,-0.2575576305,0.004103967,0.3975895941,0.3946669102,-0.0249199551,-0.2377850562,-0.1014918461,-0.1884737164,-0.2739528716,-0.052520074,0.0503012761,-0.2563163936,-0.283516854,0.0952253789,-0.0298537444,0.2155107856,0.1024071574,0.4667984545,-0.0222323146,0.1667989045,-0.0565112494,0.0894866958,-0.1122331619,-0.2835863531,-0.1714645773,-0.1027710214,-0.2645992339,0.0969906002,0.1056806892,-0.5895859599,0.1090206057,0.3119139075,0.2893698215,-0.1508764178,-0.2520390153,0.0349118412,-0.1531484574,0.2180142552,-0.0182143059,-0.0133201024,0.1382204294,-0.0576597825,0.295964092,-0.1241456196,-0.1790524572,-0.1323901862,-0.2980608642,0.1988736987,-0.2029173374,0.0301478654,-0.2107234746,0.1286042333,-0.0801802799,-0.0655269474,-0.1545435488,0.1827645004,-0.0919286162,-0.0699687004,0.2726301849,0.3414158523,-0.6863186955,0.0113491025,-0.2106210291,-0.214264214,-0.1314727962,0.222207129,-0.1535880864,-0.2164105475,-0.1873219162,-0.1039628461,-0.379512012,-0.3386236429,-0.3993170261,0.2244603336,0.0147255166,0.1020725742,-0.1624567062,-0.3465771675,0.2197372764,-0.1073613912,-0.4080575109,-0.3395632505,0.0483307168,-0.090697363,-0.2202285975,-0.2207838148,-0.0248326976,-0.1567666829,0.0327054337,0.0524218418,0.2398315221,-0.2631926239,0.2512972951,-0.1058086976,0.0842958912,0.1203728765,0.2442514449,-0.035289973,-0.0389813371,-0.0316407159,-0.1632596105,0.093528159,0.1100458354,0.1650076658,0.0794903114,0.0177719798,-0.2871534228,0.0784739479,-0.2831757069,-0.2864531875,0.229266569,-0.0078505753,-0.220486477,-0.0568295754,-0.1291802078,0.2829442918,-0.0350263044,0.0403471477,0.0812886879,0.3608579338,-0.2117709816,0.0663953274,0.1169618443,0.0159762464,0.2351304442,-0.1477355957,-0.182675764,0.486557126,-0.135331735,0.493232578,0.0639591962,-0.1806655824,0.4445699751,-0.686383605,0.1042585969,-0.1436603665,0.0319808349,0.1143553928,-0.1996471435,-0.0951006711,-0.0258608237,-0.0602661707,-0.004764393,0.3002632558,0.175249055,0.0208751317,-0.1129434332,-0.2815241814,0.3239423037,0.1032822952,0.3746682107,-0.1104339957,-0.3645000756,0.094450295,0.0589414947,0.0663930848,-0.1505609751,-0.0212418884,-0.4246389866,0.058792945,0.4740256369,0.0202400573,-0.1344058961,0.1091591269,-0.1784779876,0.1640637815,-0.1147888899,-0.0054770997,0.0562664084,0.0153738195,0.3779697716,-0.05947439,-0.0090441294,-0.0412558727,-0.4889648259,-0.0065910919,-0.036040578,0.1884359568,-0.3148462474,-0.0544636995,-0.2563505173,-0.1135910228,0.1614991277,-0.0664468929,-0.0294059757,-0.2248456478,-0.1075395644,0.2443359047,-0.088251017,0.2095916569,0.0373778045,0.2259053886,-0.0769371539,0.5092431307,-0.383086592,-0.0880490392,-0.1897036582,0.207261771,0.151835829,-0.1588685811,0.1316144913,-0.0541120805,0.2230404466,-0.1354923695,-0.4229271412,0.351303339,-0.2789322436,0.2825937569,0.0180322565,0.1744456142,-0.5256007314,0.0149826678,0.2019056976,-0.0165836718,-0.0259550773,-0.0310120787,0.0307590403,-0.2149371505,-0.1770557761,0.1357313395,-0.0972959399,-0.2872205079,0.2448734641,0.1140967384,0.0695603862,0.299531132,0.1981417537,-0.00537088,0.0894727111,-0.1570617557,-0.261768043,-0.6501930356,0.1145755798,-0.580360055,-0.4522836208,0.1640467793,0.183376506,0.4095659554,0.0223093852,-0.4418850243,0.20043239,-0.0492386557,-0.0774395615,0.1694942713,-0.2371670902,-0.0520734861,0.1274531931,-0.1467647105,-0.1221009195,-0.2460592538,-0.2049303055,-0.1848915219,0.1114116535,-0.2083913535,0.1352379024,0.1887238622,0.4692262709,0.3177906871,0.0179005563,0.3638769984,-0.0256513823,0.6060584188,-0.0940382704,-0.3954617381,0.1610117108,-0.2635265291,-0.1106085256,0.2245452702,0.1638233066,-0.2813156247,-0.3329944313,-0.0213941876,-0.1901987344,-0.2044192702,-0.4022081494,-0.0952065662,0.2365483344,0.0945554748,0.3012013435,0.1872934699,-0.2381776869,0.1218914017,0.2228388935,0.1378754973,-0.0563399568,-0.3333063126,-0.1550253034,-0.4598313272,0.1912705004,-0.119205907,-0.0245974492,-0.1621480137,0.0897188559,0.043134708,-0.1166873351,0.4323546886,0.08344201,-0.0657962188,0.0206605867,-0.2793537676,0.0274505634,0.1344770789,0.0036596982,0.1915417612,-0.1758197248,0.007738295,0.0332585722,-0.022678297,0.2441917509,0.1092428043,-0.1373887062,-0.1233561039,-0.0592083074,0.0876370147,-0.2600350082,-0.2787789106,-0.1868244261,-0.0090487953,-0.2768268883,-0.2453651577,-0.0309770498,-0.270595938,0.1344106644,0.406396687,0.3771792054,-0.056455154,0.1148965359,0.4774945378,0.3243748546,0.352869302,0.3981165886,0.1790292561,-0.2628018558,0.1974925399,-0.0177555606,0.1450859904,0.2858116627,-0.0325285383,-0.065377228,0.3229691982,0.1560714394,-0.4014391899,0.0218198393,0.28145805,0.2107096016,-0.5869054198,-0.0320110172,0.4865449667,0.0425229743,0.0244373474,0.1429231018,0.4273278713,-0.0154762976,-0.1793307364,0.2195118666,0.7822980285,-0.0852204487,-0.2441665977,0.1609069556,-0.0724488795,0.5282046795,0.0920659229,-0.0052015036,-0.2392439991,-0.2635155022,-0.1081514135,0.0192674212,0.3487837017,-0.0347120166,-0.1978699416,0.0767204389,-0.1154656112,0.3633242548,0.0274878759,-0.0209578238,-0.0317621082,-0.1243717521,0.052406583,0.2741874456,-0.0710231066,0.1198073328,-0.3207534552,0.0189971346,-0.0395037904,0.0727083385,-0.4134102464,0.066816777,-0.2139827609,0.0972498357,-0.1545440108,-0.5218372941,0.3360894322,0.1267322302,0.052815605,0.3062734902,-0.4049156904,0.3605416417,-0.1570579112,-0.2065270841,-0.1192581579,0.3092532158,0.0769666731,-0.1147351936,-0.1159133986,0.1432215571,0.0744677037,0.0528133661,-0.0045484863,0.139068082,0.1204417795,-0.0198727995,0.1986436844,-0.0818978995,0.0448006615,-0.353790313,0.1825722456,-0.0236160792,0.098224178,-0.0461268649,0.3708560169,0.044084534,-0.1025626212,0.0899795666,-0.2272089422,-0.2361452132,0.2592030168,0.3711191416,-0.2947306037,-0.2603667378,-0.0373257063,0.1620597988,-0.4218553305,0.0710398853,0.0879322439,0.1964734942,-0.0413637459,0.3043957949,0.1934485137,-0.2146189213,0.0477550216,-0.2464826405,-0.5051091909,0.0095528904,-0.2503125668,0.2647609115,0.2067702264,0.3871720135,0.0558869913,0.1240466684,-0.3818436861,-0.0157490708,-0.1078446433,-0.1650280505,0.2736258507,0.068039611,0.0908037275,-0.1712992936,0.1991847008,0.3404519558,-0.1769588441,-0.2423436791,-0.0868763477,0.0664085224,0.0810746402,-0.4429329336,0.1842288077,0.2947709262,0.0168445799,-0.100709185,0.1171609312,-0.0522598475,0.1438386589,-0.0333774239,-0.2583406568,-0.0354276523,-0.090100579,-0.1844912469,0.3827036023,0.3077846766,-0.1213059947,-0.0769079253,0.0812678859,-0.1150851995,0.0831537545,0.2045481205,-0.0781326592,0.1085335985,0.5183010697,-0.2559981644,0.2800973654,-0.0350436755,0.4954096675,0.0593205728,0.0501791835,0.0827177763,0.2220927179,0.2963466346,-0.2159973681,0.2225351334,0.0296186134,0.0777670518,0.1235868409,-0.097855635,0.2934837937,-0.1248734668,-0.0230830926,-0.0129558407,0.4266720414,-0.139371261,0.4329891503,0.464949131,-0.0962626636,0.1818495095,0.067554988,0.258151263,-0.0089332154,0.4877112508,-0.198983416,-0.1394606978,-0.0077661034,0.316293329,0.1967287809,-0.2686849535,-0.0073733721,0.1334370822,-0.1398731023,0.0344949365,0.1266100109,0.5635016561,-0.0401284397,-0.3392046392,-0.2308545709,0.1940721571,-0.0562957712,0.0696077347,-0.4285204113,-0.1924719661,-0.0327321514,0.203067407,-0.0358550027,-0.0769791603,0.3292371333,0.0425864868,-0.0600816868,-0.5967036486,0.0047177561,0.2601430118,0.3136415482,0.029153442,0.19828251,0.2266047746,0.0434601642,0.0961250886,0.4073176384,0.3040822148,0.0233556628,0.1639563143,-0.0363513045,0.0734031647,-0.147789076,0.1805889755,-0.0004239941,0.1901205182,0.057721857,0.2234319001,0.166074127,-0.1932645589,0.0568624698,-0.0404099002,0.3502347469,-0.2300453633,-0.2944568694,-0.3778141439,-0.0521235503,-0.0427737385,-0.3763996661,-0.3126113415,0.1954146177,0.2065277547,-0.3035633564,0.0059867185,-0.3602223694,0.1030218154,0.1652546972,0.3752549291,0.3739386499,0.2587443888,-0.1840346456,-0.4179108441,-0.3763251901,0.1055428162,0.1215544194,0.1609395295,-0.2228065282,-0.0635875538,0.1066199765,0.1546380371,0.404945612,-0.0531757511,0.2147499174,-0.0518633798,-0.1292672902,0.1912103891,-0.0255182795,0.0425749756,-0.1327254325,0.0717547759,0.319734931,0.0825479925,0.0700987279,-0.2058374286,0.2681626678,-0.4323088825,-0.3078066409,0.2139192671,0.21493797,0.3701926768,-0.0298632719,-0.1319437474,-0.2877645195,-0.258418262,-0.1755834669,0.2879810035,0.1683018655,0.3712053299,0.0934679732,-0.2876015604,-0.04175115,0.0986436084,0.2029402703,0.3742450178,-0.4230095744,0.098016724,-0.0467309095,-0.2117277086,0.3160829544,0.2755284905,-0.0386101864,0.0674131066,-0.2401773632,-0.3253991008,0.2884454727,-0.4104400575,-0.173800081,0.015976062,-0.0622032657,0.4638101757,-0.2478466034,-0.3457044363,0.0219633635,0.0650144294,-0.0513223298,-0.0197088718,0.3315581083,-0.1098205,-0.061832618,0.0984673426,0.66055125,0.3390100896,-0.2001922727,0.1868810207,-0.0298749059]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2161","title":"any possibility to download part of large datasets only?","comments":"Not yet but it\u2019s on the short\/mid-term roadmap (requested by many indeed).","body":"Hi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks","comment_length":12,"text":"any possibility to download part of large datasets only?\nHi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks\nNot yet but it\u2019s on the short\/mid-term roadmap (requested by many indeed).","embeddings":[-0.3891194463,-0.3740123808,-0.1162739247,0.1506759524,0.0622137636,0.1887900233,-0.304078877,0.4447672665,0.0378322266,0.3542444408,-0.3564593792,-0.1513887942,-0.0974688753,0.3621045351,0.2103431225,-0.0131300651,-0.1141289249,0.1904921532,-0.1827059239,-0.1419407725,-0.1062881202,-0.2634531856,-0.1772791147,-0.2474129349,0.1011999547,0.0521883741,-0.0610785075,-0.2129583508,-0.4846319258,0.1679270566,0.1850956529,0.1347322315,0.2132475674,0.0874543041,-0.0001165178,-0.3638610244,0.249820888,-0.1420041621,-0.0546267964,-0.075972423,-0.2661541998,0.0980023369,-0.2611363232,-0.3607740402,0.1278627813,0.2774968147,0.153834343,0.0642965436,0.2317324281,0.0465363264,0.1569114625,0.0504524708,-0.09192729,-0.1059827283,0.2745202482,0.1316850781,-0.0167731717,0.0890930369,0.6529842019,0.2955078483,0.1596417129,0.0165712982,0.0486342907,0.3312644064,-0.0742083937,-0.2447949201,-0.0676132068,-0.667817235,0.3225148022,0.7190986872,0.6390171051,0.0586327203,-0.2068556398,-0.1973976791,0.0950743407,-0.1318095624,-0.2059930861,0.5990070105,-0.3209141195,0.2095693499,-0.4649315476,-0.304790318,-0.0938758403,0.2228641361,-0.3349883556,0.2093671858,-0.0607747696,-0.0444552787,0.4219534993,0.0172877535,0.2261251062,0.0613452829,-0.1840499192,0.3042449355,-0.1875890046,-0.4578171372,-0.3579727411,0.3052995205,0.4713473022,0.1654957682,-0.026111763,0.1588607877,0.0159724895,-0.0204095244,0.528472662,-0.2225199193,-0.4207628965,0.1492383182,0.2825590968,-0.1604557037,0.1035159901,0.0383179039,-0.0514938496,0.1244315282,-0.5423872471,-0.0376945101,-0.287822932,-0.4864363372,0.0762605742,-0.3568837941,0.2287694514,0.0679216832,-0.1987922937,0.059343487,0.0698556975,0.1869078577,-0.3365315199,-0.0574789718,0.0091133658,-0.3153178692,-0.0668499246,-0.0282008797,0.0415485092,0.0218406226,0.2869832218,-0.1902919412,0.2831380665,-0.1725798249,0.2274955064,0.0860829502,-0.0981448144,-0.1438060999,-0.0826147795,0.1885789335,-0.0069073299,0.2825871706,-0.2084194422,0.5895011425,-0.2004659772,0.3146482706,-0.2555077076,-0.2006597817,0.1155236065,0.0904464796,-0.3533132076,-0.1260456741,-0.3621942103,0.1638825089,-0.1060635373,0.0444506183,0.0426647104,0.0674075186,-0.1413932294,-0.0674550906,0.1211105958,0.0898060948,-0.1766871363,-0.0851314068,-0.1094345227,-0.2077217251,0.3175319433,0.369419843,-0.176935941,-0.1691922098,-0.13284944,-0.1729665101,0.1642738432,0.0458409674,-0.5530946851,0.3733063638,0.0181443859,-0.2508170307,-0.033175528,0.2059596628,0.573400557,-0.0056808363,-0.1946263313,0.5374069214,-0.1190647855,0.0720527694,-0.1475657076,-0.3201547861,-0.1750618219,0.2306001931,0.3784197569,0.4108273089,0.2228132784,-0.0331756957,0.3630105555,-0.1209552363,0.3794414997,0.1146710962,0.1130558923,-0.1809627116,-0.0378118604,-0.3441828787,-0.2006198466,0.184596926,-0.0362634212,-0.53636235,0.3140614331,-0.1476005167,0.059953168,-0.3719056845,0.4457927644,0.0963291675,-0.1260570437,-0.3031656742,0.3468135595,-0.2120360434,-0.3672791421,-0.2690413296,-0.0109913694,0.1107423678,0.2041404396,0.1209547296,0.2861880362,0.233896777,0.270506978,-0.2418699116,-0.0350423418,-0.1135994717,0.1303821951,0.1256029755,0.0259893723,0.0659341663,0.1047680229,0.0390891805,0.3482429385,-0.0960348025,0.048871696,0.101464726,0.0510292314,-0.085973762,-0.4007906318,0.4246424437,-0.3485095501,0.1527459323,0.0928489566,0.0650995821,0.338024199,-0.2173324674,0.1352971494,0.1284496039,0.3347902298,0.1501485407,0.1884484142,0.0666100308,-0.3190064728,-0.2428981215,0.1351118237,-0.1440277696,-0.2521216869,0.2183995545,-0.3129914105,0.1080330536,0.0762421787,0.2197606713,0.1191248745,0.2111122161,0.4136463702,-0.093166776,0.5867982507,-0.1284908503,-0.0642246231,-0.0496462248,-0.1699490249,-0.0376071222,0.055077631,-0.2037945092,0.0120354425,0.1388000399,0.2614860833,0.1480385214,0.0902386159,-0.4748574197,-0.2198837996,-0.6576402187,0.1051030084,0.1890158802,-0.0106627727,-0.0324440114,0.0493316054,0.310688436,-0.068180725,-0.1266114861,-0.0047410997,0.5340299606,-0.20363608,0.3075397611,-0.1434357911,-0.0763658509,0.2102896422,0.2144121528,0.0947407708,0.0464815386,0.2278739065,0.0323682465,-0.1463812143,-0.3229003847,-0.0072430852,0.01388318,0.2544018328,-0.2540077269,-0.2031039298,0.3431735933,-0.0518687516,0.1341795772,-0.2050896585,-0.1062456369,-0.1747316569,-0.0296900682,0.0750913471,0.1570331603,-0.2507067025,0.1538674384,-0.5111762285,-0.3047043979,0.5068899989,0.2649386823,0.211191982,-0.1921357661,-0.1236038134,-0.3133017719,-0.4235978127,-0.4293907881,0.094215937,-0.778712213,0.3560622633,-0.1063195243,-0.2906838059,-0.0174535327,0.1595924795,-0.1259342581,0.5196785927,-0.3482219577,0.1467086524,-0.2033713162,0.0317676216,0.1693217754,0.0550416075,-0.1508952528,-0.6785643697,-0.0147130685,0.0755637214,0.3077225685,-0.1279938519,0.172923848,0.3239328265,-0.0536141582,-0.1130114272,0.1195136011,0.6544164419,0.0438829623,0.2231645286,-0.0120178061,0.1469647437,0.067150332,-0.0335654654,0.1249981672,0.3182806969,-0.0588897578,-0.1626066118,0.210560903,0.0549813733,-0.2403685898,-0.2935972214,0.0229169447,-0.0570403561,-0.1307528615,-0.0944484919,-0.0409435816,0.0148520637,0.1669400483,-0.0710615516,0.0352375545,-0.1958815455,0.1172664762,0.023394458,0.3955272138,-0.0682183951,-0.5285246968,0.0079901163,-0.3258945346,0.1046456993,-0.0632990152,-0.0630554408,0.1187950149,0.1397503465,0.0980643481,0.2250496,0.3740623891,0.1097024679,0.0903314799,-0.1371557713,-0.2282398492,-0.1578129977,0.0423299596,-0.0177873001,0.0173638016,0.2937908173,-0.1104017422,-0.3869293034,0.227535367,0.0321265534,0.4309997559,-0.126136601,-0.1096326113,0.2762153447,-0.1079813093,-0.2097551525,0.2629416585,-0.0617206134,-0.2768288553,-0.1175235957,-0.091641888,-0.1372774541,0.0540449508,-0.0734338239,-0.1729569435,-0.006562226,0.0645111278,0.467261374,0.0704179332,0.0186738577,0.3044774234,0.4561192989,0.1728293896,-0.0904219747,0.0264460314,-0.2682951987,0.0627148226,0.1863460541,0.0902078599,0.306923151,0.1236800775,0.2023658007,-0.1539227664,-0.0656309351,-0.4275345206,-0.1237248629,-0.5411214232,-0.4703683257,0.4501516819,0.1469973177,-0.0150968386,0.381930083,-0.1726711541,0.1714622825,-0.1046068519,-0.091600813,0.8376738429,-0.1282040328,0.1089364141,-0.582687676,0.0880468786,0.6016821265,0.0844000205,0.0178139545,-0.2287988514,-0.4036796391,-0.1265609562,-0.0773871988,-0.1865012497,0.2191276401,-0.0140324878,0.3342232108,-0.0188354775,0.1209898144,0.0318458118,-0.0042693317,-0.254205972,-0.1926446557,0.3182453513,0.0610096864,-0.1544341296,0.3923353553,-0.0458212718,0.0103616156,-0.0599656478,0.1627359092,-0.2029775381,0.1303596795,-0.2572265565,-0.1610218734,-0.2619208694,-0.0703883246,0.0926846564,-0.1712713093,-0.052617915,0.2570390701,-0.368799001,0.2093354315,0.1777641028,0.0735390857,0.0427497625,0.2596936226,-0.2962136567,-0.0398654826,-0.038773451,0.1279263049,-0.0318852179,-0.2183395624,-0.1323293298,0.0740629137,0.0375743248,0.0915791243,-0.1063560173,0.1396319717,-0.0045523169,-0.1219030023,0.071765326,0.2285137624,-0.1336793154,0.4617174268,0.2624637485,-0.2429394573,-0.2160748094,0.6025977731,0.1481419504,-0.1980640292,0.0398264565,-0.0848970115,-0.2438188195,-0.1730263382,-0.0219368897,-0.2924412787,-0.2294911146,0.061611414,-0.3568263054,0.3281305432,-0.0630736053,0.0522400774,0.2148600966,-0.0970679373,-0.1036716998,-0.3015722334,-0.3558741808,0.1707814336,-0.0284618288,0.2075101733,-0.1949706525,0.0554932617,-0.0556042679,0.3422850668,-0.2016983628,0.3914819956,0.1816868037,0.309051007,0.295317471,0.3341222107,-0.2802690864,-0.0450773351,-0.0452751853,0.3191940784,-0.028446639,-0.1390519738,0.0395360626,0.2113285512,0.0684082806,-0.3105314374,0.3305314183,-0.1679070294,-0.1765465587,0.0448859856,0.0605307855,0.0797420889,0.221266821,-0.3298292458,0.5457164049,-0.1837894469,-0.4162542224,0.5804182291,-0.0668462217,0.1755859107,-0.1050952822,0.4465003908,0.2378045171,-0.1020478681,-0.2331247181,-0.2606301308,0.1281630844,-0.2289606482,0.3784428537,0.2253852189,0.3213331997,0.2980000675,0.3070169985,0.5308384299,0.0150588378,0.1693071276,0.4281780422,0.188030079,-0.1285158694,-0.0209528767,-0.0240430199,0.1259854287,-0.067188777,0.1150439382,0.0692101941,-0.2892436683,-0.1110152155,-0.3718354702,0.2256902009,0.2351754308,0.0984020531,0.7093648314,0.2885873318,-0.2875523269,-0.0796873495,0.2237575352,0.2400835156,0.3412955105,-0.0477847941,0.311876446,0.211027354,0.0116077755,0.0214039925,-0.7416121364,-0.0230804496,0.5278957486,0.0306334253,0.1104904339,-0.0399600938,-0.0021563931,0.0671603903,-0.1244516894,-0.1484514028,0.34768188,0.2081785053,0.0069056619,-0.1138776764,-0.3308625519,0.0599435382,0.4651026726,0.0268055573,-0.3721131384,0.5097936988,0.2622542679,0.0755279213,-0.2458222657,-0.0134865288,0.0329282396,0.1590388566,-0.4868361354,-0.0863443762,-0.1593943089,-0.0344498903,0.4787340462,0.139201954,-0.2080733478,0.0867780149,0.1971832067,-0.0136836627,0.0832031518,0.0530152172,-0.1707907766,0.0867154151,0.3282063603,0.091768302,0.3761356175,0.0357240327,-0.081651181,0.160697028,-0.1577219814,-0.2033814937,-0.3718720973,-0.2033306509,0.2540257573,-0.0361650698,-0.0504299365,-0.1201861799,-0.041457966,-0.0005966536,-0.4159591496,-0.5178381801,0.1937783808,0.3726863861,0.0314733498,-0.2502913475,0.3110057116,0.3613951206,0.1025884449,-0.2219725549,-0.4834929407,-0.133455053,0.0313245319,-0.2214999944,0.3481009901,-0.0325207151,0.3166560829,-0.0116619626,0.1680830568,-0.1648943126,0.1120261103,0.0061327443,0.3994951248,-0.1755858362,-0.1069868356,-0.0376345813,0.430949688,-0.0844210312,-0.4166844487,0.359965384,-0.0607664585,-0.0608775727,0.2279116213,-0.0132843265,0.1545913666,-0.274656415,0.2712777853,-0.1917614788,0.2938014865,-0.0138398651,0.0413306318,-0.0191197041,-0.2151006013,-0.2527847588,-0.0442254767,0.2071425468,-0.1137406975,0.0390558504,0.162722975,0.0366047136,-0.0107226651,-0.2080032378,0.3357404768,-0.082815744,-0.5599879622,-0.2229252011,0.0774425492,0.1879588962,-0.0176749043,-0.2571412325,-0.3028224111,-0.2137773782,-0.0043448927,0.517881453,-0.0673079416,0.2112175524,-0.4311704636,0.2439931184,0.1319239438,0.0757227615,-0.1829138547,0.068117626,0.4186392128,-0.3070172071,0.2027669698,0.1270377487,0.0668979958,-0.2542068362,-0.142417118,0.3023631871,0.2283589393,-0.4080889523,-0.2699455917,-0.249566406]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2161","title":"any possibility to download part of large datasets only?","comments":"oh, great, really awesome feature to have, thank you very much for the great, fabulous work","body":"Hi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks","comment_length":16,"text":"any possibility to download part of large datasets only?\nHi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks\noh, great, really awesome feature to have, thank you very much for the great, fabulous work","embeddings":[-0.4569222927,-0.4417874813,-0.1896090657,0.1484071463,0.1038997695,0.1881608218,-0.2670145929,0.3661693335,0.0632605478,0.4351860881,-0.4470095634,-0.1280029565,-0.1039806157,0.3638656139,0.19023633,-0.0035727338,-0.1473829001,0.2523769736,-0.2892186642,-0.1919610351,-0.0600638241,-0.1989702284,-0.2322041541,-0.2747715414,0.1167170703,0.0897206813,-0.0433795117,-0.1757752746,-0.5081452131,0.0985448956,0.141211316,0.2492479384,0.1027179956,0.0950109512,-0.0001107031,-0.4209389687,0.2668626904,-0.0625784174,-0.0843795389,-0.0519629829,-0.2210000753,0.1284930706,-0.2507516444,-0.3328807056,-0.0196087305,0.2339521646,0.1992979199,0.052536834,0.1239218563,0.1056275144,0.1930296421,0.1519487053,-0.0723707303,-0.1902453154,0.4425436258,0.0775688216,-0.081824325,0.0459746085,0.5918859243,0.2943479419,0.1667646766,0.0849437714,0.1034169719,0.2473489344,-0.1054589748,-0.2039708793,-0.0984312892,-0.6417449117,0.233724162,0.6659565568,0.6920181513,0.0556284226,-0.1506160647,-0.1202960387,0.0236584172,-0.1743654162,-0.2018057406,0.6123729944,-0.3871223927,0.2348982692,-0.4387528002,-0.2116997689,-0.12346486,0.2909560204,-0.3018239737,0.2000410855,-0.1299531013,-0.1054302454,0.407800734,0.0596905127,0.1735424995,0.1433303207,-0.1963621229,0.3420022428,-0.1567712128,-0.536655128,-0.1837792397,0.2887184918,0.5101740956,0.2040085793,0.1140416861,0.1894533783,-0.0184952524,-0.101585336,0.5733489394,-0.112977244,-0.3923663497,0.0809294656,0.3171570301,-0.0752295479,-0.0010931335,-0.0670251325,0.0463561118,0.1651315689,-0.5736999512,0.0215226673,-0.1362490803,-0.4492550492,0.0948014408,-0.281965673,0.3471706808,0.0344671607,-0.0805370361,0.1907563657,-0.0205787905,0.0423795208,-0.4645516276,-0.1681990176,-0.0162997488,-0.1974528283,-0.0918180943,0.0224358495,0.0416284464,-0.0880819559,0.3229663074,-0.1877465099,0.2957335114,-0.1981063187,0.3834774196,0.2267553657,-0.1125748008,-0.247535646,-0.0029422953,0.198540315,0.0700587556,0.1578546464,-0.2934513986,0.5020813346,-0.1349240392,0.3260570765,-0.2535348535,-0.2201768011,-0.0583021417,0.1515716463,-0.2562196553,-0.1333431751,-0.2288068384,0.1576193869,-0.1931798905,0.0939430147,0.079911761,0.1832103282,-0.0579355024,-0.0438142307,0.0031955966,-0.0112947142,-0.1513257176,-0.0708584413,-0.1550827473,-0.2979604602,0.3592858016,0.1999580413,-0.1971529722,-0.2521312237,-0.1948246062,-0.1017807499,0.2613486648,0.0787256286,-0.5193734169,0.2332689762,0.0885039642,-0.2687841952,0.0295890868,0.1556360126,0.5315999389,-0.0201996379,-0.262334317,0.5017807484,0.0501082316,0.0913201496,-0.0916647464,-0.2254313082,-0.1961180717,0.2768270373,0.2638245523,0.3479859829,0.4208673835,-0.0123440474,0.5024280548,-0.1851747334,0.2358278185,0.0584069751,0.2728781402,-0.1755978167,-0.0143288821,-0.4393552542,-0.1206988469,0.1615196615,-0.0966031253,-0.2584325671,0.3596479297,-0.1486888677,0.0447295383,-0.3190504313,0.2947393954,-0.0419411473,-0.0287704561,-0.2965403795,0.409374088,-0.0676431283,-0.3393349648,-0.2134828418,-0.0733152181,0.0736941621,0.1995432824,-0.0110144354,0.3500831127,0.2485636324,0.2441171408,-0.181439355,0.0056323633,-0.0823098347,0.0218137093,0.1309656054,0.0395607837,0.1180215552,0.0199108571,-0.0270482022,0.3809484243,-0.113807179,0.0639724284,0.1126640588,0.1333455592,0.0193930604,-0.4595884085,0.345484674,-0.2954245508,0.0264134519,0.16872558,-0.0491245538,0.3897067606,-0.2340516895,0.1279972494,0.0550441071,0.3604438603,0.163096264,0.0675400347,-0.0170681085,-0.2695915401,-0.1503736973,0.0879044235,-0.0890912414,-0.1837015003,0.2362305224,-0.2942356765,0.1283306181,0.1056605652,0.1332110316,0.2384630591,0.2614440024,0.419123292,-0.0596985407,0.5338898301,-0.0798916295,0.0344999954,-0.1112052947,-0.2185872793,0.0271370411,0.1192569062,-0.2720867693,-0.1960972995,0.1185508817,0.1599407047,0.1313765645,0.1610589474,-0.5328859687,-0.1614165753,-0.5867491364,0.201430887,0.1765574366,0.0589453951,-0.0102747884,0.1366430372,0.3728674352,-0.1443712711,-0.1607402563,-0.0089779543,0.4862968624,-0.1067721546,0.356928736,-0.1209159717,0.0533811711,0.2583846748,0.252482444,0.0410920382,0.0816703141,0.2867903113,0.1602697074,-0.1303244531,-0.4161487818,-0.0071792677,0.089368321,0.3147806525,-0.1491581202,-0.0932761505,0.2668730617,-0.0285225995,0.1035330892,-0.2054284811,-0.1214779317,-0.2091581225,-0.144128561,0.0973671749,0.0649711415,-0.2281395048,0.0051542521,-0.5068331957,-0.3248071373,0.5314695835,0.3946836591,0.2537349463,-0.1494753212,-0.1242601499,-0.251788795,-0.3573119938,-0.4232786,-0.1491891146,-0.78258425,0.4113961458,-0.2570545971,-0.3438847661,-0.0755715668,0.1506991684,-0.1608594209,0.4088255167,-0.247415781,0.0621273853,-0.2387616783,0.02575019,0.1554755718,0.1434172243,-0.1240387261,-0.7096162438,-0.1004810557,0.1014398709,0.3139617145,-0.2614524961,0.250983268,0.3713076413,-0.0902180597,-0.1903648376,0.0399543159,0.5214747787,-0.0268349163,0.2538589537,0.13070409,-0.0290280152,0.0504805818,-0.0862841979,0.0495517105,0.3566077054,-0.0647659302,-0.2368633002,0.240204528,0.0547218695,-0.4161052108,-0.3159688711,0.0295124669,0.02450151,-0.0842438787,-0.1255429685,0.0592539608,-0.1105325446,0.1262482852,-0.0890258551,0.0932171941,-0.2466400564,0.1455629319,-0.0436632186,0.2179503888,-0.1918685138,-0.5488732457,-0.0880172253,-0.3732866943,0.0275459848,-0.0887911245,-0.2204172313,0.1540469527,0.0118220067,0.2002578527,0.1046784669,0.5495660305,0.1924218535,0.0441883281,-0.1874068826,-0.1772257984,0.0666364059,0.0763562247,-0.0516126789,-0.0296916068,0.1616582572,-0.118829608,-0.3217067122,0.2123565972,0.1185419187,0.3280698955,-0.1773505658,-0.1671269834,0.2144457698,-0.1191549301,-0.2957466841,0.2118777335,-0.0948872268,-0.1333677024,-0.0956818089,-0.2035372555,-0.0976695046,0.0478804335,-0.0564669147,-0.0978599265,0.0401680954,0.113626644,0.4426668584,0.0353579447,0.0149895009,0.1644071043,0.5174337626,0.230738923,0.0437976196,0.0463366397,-0.2539515197,0.0378039666,0.1980101764,0.0985030308,0.2301355451,0.0144507103,0.3215952814,-0.1256861985,0.0775860325,-0.4251402318,-0.168773666,-0.4518716335,-0.3695443571,0.42149055,0.0701667964,0.0333634578,0.362665385,-0.1292976141,0.1164963841,-0.1096644774,-0.051914189,0.9109153152,0.000767161,0.0757828504,-0.4442178905,0.0725759715,0.5207017064,-0.0016891884,0.0603379309,-0.2513405979,-0.4070748091,-0.1607123017,-0.0440550782,-0.1269133836,0.1275818199,-0.0306454264,0.2152334303,0.0330473818,0.0687594265,-0.1061895713,0.0132145481,-0.2566154897,-0.2052282244,0.2999649048,0.1174326092,0.0099960705,0.407209307,-0.0089249872,-0.0222103558,0.0564134642,0.0811619014,-0.1732973605,0.0782127529,-0.2402407229,-0.1271180809,-0.215333879,-0.1490634084,0.0119206486,-0.2115807831,-0.091134727,0.2856214941,-0.4071881771,0.26731655,0.2349710763,0.1415298134,-0.0455124564,0.170298636,-0.245164752,-0.1385824382,-0.0636678636,0.130478844,0.0675147548,-0.2238564938,-0.1592640728,0.16303204,-0.0017760174,0.072884798,-0.1480656266,0.147682175,-0.01074233,-0.1184334755,0.1171132252,0.2189232856,-0.2218289971,0.411645472,0.2776658833,-0.1695398986,-0.3347319067,0.6299330592,0.2015761733,-0.2579894066,-0.0588951632,-0.0949731469,-0.2412516624,-0.202932477,0.0254393741,-0.3619157374,-0.2489606291,0.0684148371,-0.2431002259,0.2482887954,-0.1463270336,0.0954281613,0.0726269186,0.0065886979,-0.1642003655,-0.2006360441,-0.4192204475,0.0153944353,-0.0635218769,0.2517635524,-0.2178003192,0.0093263807,-0.0072475299,0.1959369928,-0.2795917392,0.3235084414,0.1009355709,0.1888236105,0.4417482913,0.2852602005,-0.2603090107,-0.0910110846,0.0227065403,0.3770690262,-0.0110532874,-0.2271911204,0.02128613,0.1728378981,0.0746538937,-0.2443025112,0.2491169423,0.0081321523,-0.1254450977,0.0937339962,0.1706279218,0.0679826289,0.2518122792,-0.3036060333,0.5130080581,-0.3523679674,-0.4174032211,0.5407012701,0.0125269257,0.0838822946,-0.0037675463,0.4048577547,0.1292494088,-0.0992391557,-0.1200565398,-0.2527925968,0.1453033686,-0.1895305961,0.3002199233,0.3162284195,0.1649176627,0.3023346961,0.2457776517,0.5084362626,-0.0152826328,0.1085115522,0.4348978102,0.2377790213,-0.2317810506,-0.1498553455,-0.2036657333,-0.0485083796,0.0155830039,0.0273264349,-0.0010940925,-0.2478485405,-0.069670938,-0.2985459864,0.1178782508,0.2388333827,0.0260508731,0.7074028254,0.1500037462,-0.3291203082,-0.0424694605,0.1737633497,0.2074752599,0.4642714858,-0.1255783439,0.2695052326,0.2516717911,-0.0656983703,0.02726309,-0.6322484612,0.0490346327,0.4742776155,-0.1550252438,0.1333580464,0.028125437,0.0583111234,0.1292907447,-0.0616119541,-0.2436540127,0.3722504079,0.2169623822,-0.0386298858,-0.0611900985,-0.3569636345,0.02983227,0.4323067069,-0.0329571515,-0.3165524006,0.4121624529,0.0604237728,0.0055015208,-0.1391835511,0.0401393361,0.0196078084,0.1764257252,-0.4152518511,-0.1744776815,-0.061294347,-0.0081064273,0.29788661,0.1160275117,-0.1798857003,0.1004530713,0.2251486033,-0.040642038,0.1101352349,0.0405046791,-0.2707314491,0.0644743517,0.3029298484,0.0313111357,0.4102818966,0.0723478869,-0.1175643653,0.2593792081,-0.0650700554,-0.1717311889,-0.3983203769,-0.2220225185,0.2776276171,-0.0164545309,0.0086402502,-0.1884790361,-0.0986082479,0.0294517335,-0.3627734482,-0.4593484998,0.247218132,0.3025139272,0.0553995855,-0.302854389,0.4305765331,0.232688427,0.0304532759,-0.1031257361,-0.4151988029,-0.2457071096,0.0794984773,-0.1626459658,0.2742976844,-0.0632451996,0.3734807074,0.0651796013,0.3338826895,-0.1283565909,0.0352491774,0.0676705539,0.414480269,-0.1857172251,-0.0640038773,0.0284610949,0.5671876669,-0.0703192055,-0.4127237201,0.3753577471,-0.1111045107,0.0034153999,0.2319036275,0.1265772432,0.0676539764,-0.211583063,0.271636337,-0.2187780291,0.1533116549,-0.1447098851,0.1211735383,0.0287775099,-0.0959732905,-0.1997182518,-0.196289286,0.3393843472,-0.1301870495,0.2296262383,0.1153035909,0.0823763758,0.1041029319,-0.0554955639,0.2501114011,0.0361601114,-0.4684174657,-0.0568600222,0.0001645721,0.1268473715,-0.0011090249,-0.2792007625,-0.3743989468,-0.1956621557,-0.1279824972,0.5047439337,0.0234211218,0.1405041516,-0.436586529,0.2666721642,0.1913579404,0.0209435225,-0.2351938486,0.1179645136,0.4022257924,-0.3174346089,0.2347785383,0.2826938331,0.142582044,-0.228366524,-0.1481786221,0.2692000866,0.3289658427,-0.3923661113,-0.2372132242,-0.2452916503]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2161","title":"any possibility to download part of large datasets only?","comments":"We'll work on dataset streaming soon. This should allow you to only load the examples you need ;)","body":"Hi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks","comment_length":18,"text":"any possibility to download part of large datasets only?\nHi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks\nWe'll work on dataset streaming soon. This should allow you to only load the examples you need ;)","embeddings":[-0.4160075784,-0.4227997363,-0.0912121311,0.2210884541,0.1178921759,0.208614096,-0.2158917636,0.3746378422,0.1075313836,0.3137337267,-0.2778739035,-0.2264849246,-0.0971202999,0.3623839021,0.2786155045,-0.1379799843,-0.1310558468,0.2409262955,-0.195934087,-0.1360583901,0.0351950452,-0.3711940944,-0.1557517797,-0.2762488127,0.1072821841,0.0065381951,-0.0016220268,-0.152801916,-0.3216457367,0.1749350876,0.1286829263,0.0949424505,0.2968363166,0.0225311723,-0.0001144941,-0.3012693226,0.3300118446,-0.2018281966,-0.2040800303,-0.0629979521,-0.3263174295,0.163317278,-0.2283700705,-0.2755313218,0.0555490777,0.1558898836,0.1559820175,-0.021303568,0.3270632327,0.1292208731,0.1475346684,0.0527245738,-0.1914350986,-0.0762383267,0.2807136178,0.1402224898,0.0194406081,0.0566957854,0.6661752462,0.300422132,0.1056198776,0.0831306353,0.0118729435,0.1913021952,-0.0942525864,-0.3307495117,-0.1941795945,-0.623032093,0.2782781124,0.8252424598,0.6387521029,0.0939345956,-0.1760771275,-0.2948754728,0.1539803594,-0.1587239504,-0.246307373,0.629265964,-0.3267938793,0.1728767008,-0.540204823,-0.2325883061,-0.1865912229,0.2556458712,-0.2171286941,0.0882261023,-0.079015933,-0.0391912796,0.3143946528,-0.0377393067,0.311180383,-0.0938632265,-0.2239779532,0.1800582856,-0.2369107753,-0.3743753433,-0.3241960108,0.3473267555,0.3826083839,0.146750167,0.2207317203,0.196091488,0.0698590428,0.0267213043,0.424017787,-0.1731639504,-0.2814107835,0.2035138011,0.2531688809,0.0371988602,0.0458770916,-0.0351084732,-0.1991999596,0.2605749965,-0.5001295209,-0.0200261753,-0.3018489778,-0.421323806,-0.0149688032,-0.264572084,0.1118759811,0.1028684527,-0.0745217651,0.1771867275,-0.0060785618,0.2028246224,-0.3606562912,0.0311386921,0.07468009,-0.50466609,-0.0934187546,-0.0063180411,0.0433680825,0.049343653,0.3772119284,-0.2916145325,0.379822284,-0.1353830248,0.2860985696,0.1216544881,-0.023139311,-0.1652133018,-0.0564624742,0.1320164204,0.1193649247,0.1957210898,-0.0756725222,0.4238185585,-0.1083752736,0.3343409002,-0.1200220659,-0.2665345371,0.1457662731,0.1189228818,-0.3860332072,-0.080825299,-0.5248143673,0.2487920821,-0.1913020462,-0.0301384795,0.0101986537,0.0602893978,-0.0674410462,-0.0665501058,0.2375759333,0.0933558792,-0.2284577638,-0.0791617408,-0.15456523,-0.1223678142,0.3714396358,0.2702901363,-0.2721643746,-0.2225671113,-0.1437160969,-0.205364868,0.245246917,0.0532231517,-0.4416381121,0.4437435567,0.0718872845,-0.0778924227,0.0095832935,0.1884521246,0.6336510777,0.0795476735,-0.2656627893,0.6198097467,-0.1294934899,0.101855278,-0.0497492477,-0.2907518148,-0.1316679418,0.2972227037,0.2475744635,0.391423285,0.1908108443,0.0140925888,0.3742015064,-0.203521505,0.2755431235,0.1488898546,0.1048043296,-0.121955663,-0.1044193357,-0.332970947,-0.2455965877,0.2019154578,0.1670384556,-0.5572571754,0.3659934402,-0.1457132399,0.0362955369,-0.3327440023,0.3766084015,0.0140642934,-0.1043354869,-0.3237961829,0.3023703098,-0.2443415821,-0.3178130388,-0.0696160868,-0.1488308907,0.0794271454,0.0792413726,0.1160821617,0.4168473184,0.1795063615,0.2777090371,-0.3217895031,-0.0120968772,-0.0677321255,0.0181381609,0.1762458533,0.0013726536,0.2247814983,0.1153095365,-0.0137184346,0.2783602178,-0.197149381,0.0457342044,0.2255697697,0.1472244412,-0.0683518946,-0.4450111091,0.3316784799,-0.2376499176,0.1782503575,0.1220900714,0.0174170621,0.2728701234,-0.1368841976,0.0763017237,0.1350092739,0.450481683,0.0766145438,0.2161708474,-0.0247611105,-0.4326867461,-0.2301254869,-0.0211267062,-0.263863951,-0.2799152434,0.2115961909,-0.413223654,0.0817955658,0.1887237579,0.207456246,0.0813945904,0.2328840345,0.3495491147,-0.082832031,0.6517193317,-0.1666290164,-0.0071176989,-0.0003751305,-0.1914367229,-0.1143666804,0.05064594,-0.2767645121,-0.0760612339,0.1437413245,0.2027590871,0.0762582123,-0.0086085144,-0.3400782645,-0.244123891,-0.6714405417,0.0762961507,0.1519193649,0.0577860847,0.0029210374,0.0193789694,0.4107501209,-0.0264356062,-0.0868318379,-0.0315064378,0.5220625997,-0.0951065347,0.2493910491,-0.1078799069,-0.0365835465,0.0866006464,0.2252970189,0.1121644601,0.1787572801,0.2441754937,-0.0232837871,-0.08037965,-0.3961844742,0.0118816802,-0.0368463323,0.3093235791,-0.0959338173,-0.1772044152,0.4233986437,-0.1610391736,-0.017416887,-0.2303730249,-0.1522228271,-0.0341773704,0.0224791374,0.0635860935,0.268858701,-0.1699955165,0.1525307298,-0.4189006686,-0.346806109,0.3461508155,0.1341480911,0.1633359194,-0.2418619841,-0.0921135768,-0.2322889268,-0.2644003332,-0.3277679384,0.0947527215,-0.8488109708,0.41301018,-0.1486281306,-0.3536370397,-0.0303917248,0.0645904839,-0.2282583863,0.5403364897,-0.3521283269,0.1787143648,-0.2025313079,-0.119592838,0.1085025817,0.0318410844,-0.2324289829,-0.7003283501,0.000242744,0.0069476576,0.2122762352,-0.1871311963,0.1414143294,0.3855127692,-0.0612121001,0.0452300757,0.1133166179,0.6742675304,0.1244411543,0.1787571311,0.0296519063,0.0931019261,-0.0170451719,0.0397601128,0.0937465951,0.3169905841,-0.0764262006,-0.1842021048,0.1732453853,0.0769554749,-0.2560747266,-0.3300912976,0.0518585742,-0.1199234799,-0.1267179251,-0.0537191294,-0.15822348,-0.0047133812,0.1484063715,-0.020493798,0.0621159971,-0.2487458438,0.049259413,0.0287837349,0.3928291202,-0.0162778739,-0.4690263271,-0.0255295653,-0.3079204559,0.0758049861,-0.0953145176,-0.100026384,0.0522106104,0.1006636396,0.073355034,0.2640865743,0.3989117444,0.0156194568,0.1200770065,-0.1917634606,-0.1950031072,-0.197348088,0.03585691,-0.0327315666,-0.0134560792,0.2429583818,-0.1197205409,-0.3848586082,0.2900092602,-0.0171892233,0.4906187952,-0.0699083805,-0.1322181225,0.180302009,-0.0493799634,-0.2061173916,0.2111072838,-0.0840540677,-0.2146123946,-0.2539407015,-0.0784406438,-0.1251194179,0.1815256327,0.056097921,-0.1242298186,0.051886335,0.134619385,0.3988798857,0.1409564912,0.1178589165,0.2378036976,0.4254065752,0.2991864681,-0.1155485511,0.0455166548,-0.0836470872,-0.0475634672,0.164222002,0.0294937696,0.2476634085,0.3052614033,0.2623026967,-0.1150103956,0.1111810133,-0.4343546331,-0.119833149,-0.4944825172,-0.6160652637,0.4447358549,0.1657908708,0.0423179604,0.426591754,-0.2824220061,0.1392943859,-0.1379765421,-0.1551315635,0.7866649032,-0.044668097,0.0566250086,-0.4420417845,0.0823454112,0.4862217605,0.0888805464,-0.0156090017,-0.1864544451,-0.5498301983,-0.1266833842,-0.1424373984,-0.1348602325,0.1910951436,-0.0513050333,0.3555556834,-0.1016653851,0.2523554862,0.0963890329,-0.0012904952,-0.3779738843,-0.220428735,0.0948168337,0.0760561749,-0.1214836314,0.3061072826,-0.0015229726,-0.007503327,-0.0485644415,0.0961134136,-0.1955926567,0.2064882964,-0.3052038848,-0.1430629939,-0.3039154112,-0.1405451596,0.1190783083,-0.0109744035,-0.1744748205,0.1768085361,-0.3655038774,0.2938465476,0.2110069394,0.1051473469,-0.0190242417,0.1918653846,-0.2770028412,-0.0221473742,-0.0550033711,0.3291868269,0.0409177206,-0.1103689894,-0.1155573875,-0.0003945179,0.1025445834,0.0863654241,-0.1962905675,0.0315900259,0.0739088506,-0.1351101547,0.0788360834,0.1967615038,-0.1610275656,0.3221873641,0.2844254375,-0.120534271,-0.2411416918,0.5982012749,0.1175037399,-0.1930471659,0.05672203,-0.0242088474,-0.2109339535,-0.1459590197,0.1243494675,-0.2360053509,-0.1271972805,-0.0554685555,-0.2937462628,0.3160286546,-0.091245681,-0.053296715,0.1353044063,-0.0562686175,-0.0229642857,-0.2877100706,-0.3725259006,0.2079887837,-0.0053924951,0.1750872284,-0.143056035,-0.0372063853,-0.0696321651,0.2916401327,-0.2089210153,0.3644315898,0.1304096878,0.2831199169,0.2882481813,0.34952268,-0.2443481386,-0.0633350611,-0.0329829715,0.2173096985,-0.0284552537,-0.1330003142,0.0020383443,0.1984715164,0.1089472249,-0.3337973952,0.1763237566,-0.1900523007,-0.0990203321,0.0433452539,0.1389496922,-0.0159239788,0.2052645683,-0.3531026244,0.5082846284,-0.1798098385,-0.5356280804,0.4620235562,-0.1171631142,0.1306315511,-0.0761510208,0.3220211267,0.321087569,-0.1224154308,-0.2100385427,-0.2230060101,0.2127708942,-0.1388964504,0.3544560373,0.1668493301,0.3332601488,0.2461369187,0.244517833,0.5342573524,-0.0433653668,0.2196805477,0.4412213564,0.2113067657,-0.0947918221,0.0131518543,-0.0713933781,0.0731760934,-0.0042212764,0.0684015751,-0.0928286016,-0.3367205262,-0.1804695725,-0.3591629863,0.1765959263,0.2496611327,0.22800228,0.7436214089,0.2074043304,-0.2657696009,-0.3291831911,0.2548859715,0.1826936454,0.2631572485,-0.0423873775,0.3229688704,0.177969262,-0.0002193833,0.007509294,-0.7347243428,-0.127799347,0.5495398045,-0.0863283128,0.1695782989,-0.0750985444,0.0266850572,0.1294783205,-0.129287079,-0.1657757461,0.2464837879,0.2347250581,0.0446922295,-0.0985675901,-0.2884527445,-0.0108831208,0.4674572051,0.0264977999,-0.3120701313,0.2790999115,0.2482138127,0.0383433662,-0.2654882967,-0.0185951181,0.152142778,0.1460036486,-0.4191277623,-0.1224376038,-0.0334549695,0.0014860475,0.4145693481,0.0735967159,-0.1740531325,0.0690135658,0.2184703499,0.0576532409,0.2066496015,0.0722353607,-0.1061973572,0.0054670982,0.3811666965,-0.0179976262,0.3822942078,0.066165857,-0.1120460406,0.2347371876,-0.1527878493,-0.1883722991,-0.4321079254,-0.2551423609,0.1460382193,0.0544812977,-0.0776702911,-0.1592255384,-0.0845691562,0.1961088777,-0.2716816366,-0.3965240419,0.1675684899,0.3351818621,0.0275686178,-0.2927804887,0.3786398768,0.2864269018,0.1504724324,-0.2301665545,-0.5346704721,-0.1499330997,0.0194744952,-0.2891178727,0.3038632572,0.0546502434,0.1894909143,0.018169295,0.2077967077,-0.2392201126,0.1931411475,-0.1128476784,0.4548941553,-0.1457685381,-0.1725870222,-0.0497344546,0.5787016153,-0.1278290749,-0.400983423,0.3574506044,-0.0456189327,-0.0238707513,0.2780319452,0.0456854478,0.0740412772,-0.2255824804,0.1281099319,-0.1869567931,0.2994700372,-0.0536473989,0.1234951839,0.033541657,-0.1741548926,-0.1832642555,-0.0315529518,0.1947118789,-0.0502502248,0.1012261212,0.1653437763,0.0438488126,0.0073732729,-0.104826957,0.2365268767,-0.1300774515,-0.4823198318,-0.1868067533,0.1812528223,0.1828771383,-0.0484647565,-0.3840535879,-0.3502146602,-0.1270209104,-0.0526531748,0.5525504351,-0.1313402653,0.1620955467,-0.4143194556,0.323641181,0.2318766713,0.1669839323,-0.1339512914,0.161023289,0.4107991755,-0.3414475918,0.2553164363,0.1687323302,0.2022461742,-0.1833287776,-0.1224539429,0.1712695956,0.251616776,-0.4405371249,-0.2349566221,-0.3314546347]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2161","title":"any possibility to download part of large datasets only?","comments":"thanks a lot Quentin, this would be really really a great feature to have\n\nOn Wed, Apr 7, 2021 at 12:14 PM Quentin Lhoest ***@***.***>\nwrote:\n\n> We'll work on dataset streaming soon. This should allow you to only load\n> the examples you need ;)\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","body":"Hi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks","comment_length":79,"text":"any possibility to download part of large datasets only?\nHi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks\nthanks a lot Quentin, this would be really really a great feature to have\n\nOn Wed, Apr 7, 2021 at 12:14 PM Quentin Lhoest ***@***.***>\nwrote:\n\n> We'll work on dataset streaming soon. This should allow you to only load\n> the examples you need ;)\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","embeddings":[-0.4333571494,-0.4534873664,-0.0487325303,0.2667760849,0.1081289425,0.2431801558,-0.1623352021,0.4673766494,0.0734750777,0.3738210499,-0.3642461896,-0.2017757744,-0.0959115475,0.457791537,0.3195568621,-0.1472721994,-0.1357191503,0.2194820344,-0.2238364071,-0.1688051075,0.017404547,-0.2593170106,-0.1123769209,-0.2532622814,0.0350642912,0.0898537561,-0.0617154092,-0.0844494626,-0.3761279285,0.1701032072,0.1983957291,0.2140855193,0.2366961986,-0.0048213038,-0.0001184534,-0.2889572978,0.2554901242,-0.1432120502,-0.1864690632,-0.0577484518,-0.1817248613,0.1063679606,-0.2372541577,-0.262868762,-0.0091608325,0.1348176152,0.0690670013,-0.0218835771,0.3579902649,0.0345511064,0.1137142628,0.1012044027,-0.1499210149,-0.1075810269,0.272598505,0.1650012732,-0.0292274952,0.060077142,0.5326678753,0.2669931352,0.1294612437,0.117367059,-0.0027756188,0.2834877372,-0.0907342881,-0.1894118488,-0.2640944421,-0.6127950549,0.2104036659,0.8935748935,0.5572637916,0.0032649203,-0.2365484685,-0.4016392529,0.1302787066,-0.3426574171,-0.1672671884,0.5966566801,-0.3093205392,0.192076847,-0.5190604329,-0.1575798392,-0.1254054755,0.2723096907,-0.2245434374,-0.0139281359,-0.1768504083,-0.0458901934,0.4784290195,-0.0008327375,0.1448497325,-0.0473284982,-0.1877050549,0.2302681953,-0.2143140286,-0.4565867186,-0.2157699317,0.2906162441,0.5468568802,0.2475865185,0.0848186165,0.1489576548,-0.017101923,-0.0635389686,0.491554141,-0.2148909718,-0.2800322473,0.1818933934,0.3234026134,0.083050549,0.1106132641,-0.0376580507,-0.0844548568,0.245326519,-0.5971102715,-0.0122542037,-0.2203571945,-0.4416057169,0.0325089097,-0.2539963424,0.1385335475,0.1137556285,-0.0686253831,0.1652995795,0.0227542017,0.1592404246,-0.3400886655,0.0705235898,-0.0038737166,-0.4409328699,-0.132912755,-0.0432235599,0.0712560043,0.1238532141,0.4013691545,-0.3420198262,0.3170058131,-0.1057771742,0.3391904235,0.1396165043,-0.1107201129,-0.1596134454,-0.0881129429,0.1561319083,0.0169006623,0.2920125127,-0.0551254116,0.4835698307,-0.1643662751,0.2821950614,-0.1610147059,-0.2401708961,0.1469040513,0.0407236703,-0.3738723397,-0.0618707091,-0.5082563758,0.358671546,-0.1699185371,0.0181915406,-0.0189901385,0.1638083309,-0.2160223573,-0.0025573042,0.2322248369,0.2576268911,-0.2118474096,-0.1752518415,-0.1312912852,-0.139586404,0.3125015497,0.3773928583,-0.2835659087,-0.2201153636,-0.2215556204,-0.0606914833,0.1919154823,0.0454829633,-0.4724436104,0.3615725636,0.0099886172,-0.030708842,0.0478569455,0.1477380991,0.6136130691,0.0301964954,-0.1679208726,0.5325431228,-0.08927086,0.0808937028,-0.049649179,-0.3678531349,-0.1579562426,0.2755120993,0.1951163858,0.4164232314,0.1727066934,-0.0402595215,0.4430029988,-0.2201928347,0.3368228376,0.1818815023,0.1613423824,-0.0847179294,-0.1551973224,-0.4076611102,-0.2964693308,0.2118466496,0.132121712,-0.4998221397,0.3217774034,-0.2148119807,0.0395307168,-0.3210993707,0.4084578753,0.005683789,-0.1502961963,-0.3171791434,0.3548722863,-0.1546446234,-0.3655125201,-0.0756156072,-0.1111734882,0.1478280872,0.0036148818,0.1395067871,0.4277284741,0.1841936409,0.2695959508,-0.2078377455,0.0122560803,-0.1331464052,0.0429858752,0.1440230161,-0.0245181285,0.1365618855,0.093528308,0.0826350451,0.3256608546,-0.0734846964,0.0104416581,0.0894771665,0.1080784276,-0.0458220914,-0.493114531,0.4346589446,-0.2044020891,0.1076002344,0.0549512468,0.0469650067,0.2766539752,-0.1762056351,0.0548627973,0.0512986891,0.4707574844,0.0086868182,0.1812570542,-0.0498716682,-0.4779136181,-0.152316913,0.0950541571,-0.18972978,-0.2867540419,0.2613023818,-0.3662053645,0.1668248773,0.1373097152,0.1499001682,0.1633578241,0.1919973493,0.3767030835,-0.0624540746,0.6231998205,-0.1187390387,-0.0556434728,-0.0444185175,-0.2490230203,-0.0204355139,0.0649898201,-0.1989006847,-0.133872211,0.1765610427,0.1989263147,0.1194389164,-0.0303290859,-0.34018448,-0.3156105876,-0.5664642453,0.0017028174,0.1305923015,-0.0309436936,0.0081284959,0.0766007081,0.4942077994,-0.0850624666,-0.1224118844,0.0192583464,0.5008317828,-0.1121080294,0.2183012664,-0.1350676566,0.0262115765,0.1769272685,0.1761347502,0.0995416939,0.1548149586,0.2831908166,0.0096908715,-0.1914437711,-0.402951777,-0.0166045222,0.0456123985,0.1962953359,-0.0449833237,-0.0982212797,0.392893374,-0.127578482,0.0537221506,-0.1555530727,-0.1446164101,-0.1366891265,-0.0164341182,0.056850113,0.3071407378,-0.2456603795,0.1228187382,-0.3954452872,-0.3361863494,0.519174993,0.2037495226,0.1333261281,-0.1553138196,-0.12507613,-0.193160221,-0.3779405951,-0.3520568311,-0.0067362194,-0.8144291043,0.3607504666,-0.1228069142,-0.3429990411,-0.0961320922,0.040607363,-0.2655555606,0.4734797776,-0.4235666096,-0.0079286043,-0.2472912967,0.0260360315,0.130742535,0.0601334423,-0.0914825872,-0.7272815108,0.013625415,0.0473928005,0.0997406766,-0.1694025844,0.1886521727,0.3197810948,-0.0790273175,0.0955648646,0.0600277446,0.7942323685,0.200170368,0.2377908826,0.0253898948,0.0094836866,0.1083216667,-0.0896689743,0.0220876019,0.3123306632,-0.0865291134,-0.1279292256,0.1711140275,0.0905011296,-0.1765900254,-0.289442569,0.1126209348,-0.0648026317,-0.2297687829,-0.0853299722,-0.0360889472,-0.0486704968,0.1773178726,-0.0450428054,0.1011379808,-0.2055665553,0.0768797249,0.1030454338,0.4261222184,-0.0032629361,-0.5111986399,0.0256013703,-0.4151403606,0.0298555158,-0.100816533,-0.1754585803,0.0461766273,0.1045521572,0.1455368847,0.2749370337,0.609443903,0.1287353486,0.0785779506,-0.1902297586,-0.3172528446,-0.1819413155,0.0737126619,-0.006906209,-0.0538485684,0.2234911621,-0.0362248383,-0.498930186,0.2183304876,0.1101668775,0.4760243893,-0.0809594169,-0.1244961992,0.1341005117,-0.0801891685,-0.3476271927,0.2703022659,-0.0027732239,-0.1474338174,-0.2224275917,-0.10434784,-0.0786637217,0.1369889081,-0.0433476046,-0.1367671788,0.0629204735,0.1720432937,0.5014215112,0.2424036264,0.0527436212,0.1936713457,0.5433934331,0.2950322628,-0.1828605086,-0.0381433852,-0.0484642908,-0.0636333227,0.2625685036,-0.002589142,0.3213341236,0.3059742451,0.2316221148,-0.1925803721,0.0793704018,-0.3739943802,-0.0820894912,-0.4709227085,-0.5614632964,0.4475450516,0.2028789073,0.1062162668,0.4346662462,-0.1795486957,0.1982417405,-0.1883296371,-0.21632348,0.8998525143,-0.0903342068,0.1644762456,-0.4076107442,0.1434990168,0.527906239,0.0885768384,0.037722297,-0.2836924195,-0.4837168455,-0.105460003,-0.1152042821,-0.1306108832,0.1632205993,0.0144835738,0.3080941439,-0.0136230011,0.2355792224,-0.1081839874,-0.0506039932,-0.324015379,-0.2965848446,0.0336049795,0.0432380699,-0.0538074598,0.4086197317,0.0487387516,-0.0098851696,-0.0367715284,-0.0023935605,-0.2605628669,0.1173681021,-0.3374474347,-0.1653789133,-0.1736391783,-0.1403459758,0.0817817152,-0.0643985868,-0.0680085123,0.2243033201,-0.4172194302,0.1950127184,0.0875793695,0.0486802608,-0.0289878733,0.1295095533,-0.2384096235,-0.0165440142,-0.0272004604,0.2514557838,0.0822015405,-0.1809625179,-0.0513045676,0.0272661652,0.0684541315,0.0034612508,-0.1599866003,0.0955808014,0.1455604434,-0.173611626,0.0297265463,0.2340492606,-0.206310153,0.2809705436,0.2066080272,-0.0870758295,-0.2332040071,0.6446015835,0.059290681,-0.1178617403,0.0574287847,0.0331476778,-0.1950786263,-0.1573961973,0.0382627547,-0.2670314014,-0.292029649,0.0826941505,-0.1529109478,0.3014318049,-0.1664447337,0.0802085549,0.1290063709,-0.0819444507,-0.0128515279,-0.2683204412,-0.3731096983,0.1780234128,-0.0694724098,0.2584632337,-0.0643843487,0.0884527564,-0.1560299695,0.2733323872,-0.1831398159,0.3751690686,0.0650057718,0.2167410553,0.4075687528,0.2854740918,-0.1137747988,-0.0564871393,-0.0738376677,0.3291239142,-0.1099058986,-0.1135168299,-0.0640317798,0.1994957626,0.0789248869,-0.3184599578,0.3013946414,-0.074977532,-0.1752388626,-0.0162115525,0.2714849412,0.0736441612,0.2812747955,-0.3258123696,0.4892233908,-0.1573141962,-0.5090850592,0.5505743623,-0.0986980274,0.1000480875,-0.0494870581,0.3082484603,0.1302594841,-0.0771683902,-0.186660111,-0.218367815,0.2284560204,-0.1828157306,0.3463941514,0.1696282625,0.2244622558,0.3715806901,0.3263930082,0.4821109176,-0.0676549748,0.2630613148,0.4612115622,0.150950864,-0.1444545239,0.0062383171,-0.0243892614,0.0559234284,0.0369053893,0.0973551571,-0.0270196684,-0.276356101,-0.1115550026,-0.3886429667,0.1320342571,0.266398102,0.2323323637,0.706525147,0.1635611057,-0.2393027693,-0.1509808153,0.2523436248,0.1701167375,0.2535544038,-0.1003313512,0.1628603041,0.0309789032,-0.0316493772,-0.0584557876,-0.8065460324,-0.0191517528,0.5419054627,-0.0379736684,0.1781440824,-0.1530634165,0.1885349303,0.0250429325,-0.1017902419,-0.2043039352,0.3760119081,0.2479342669,-0.0519262813,0.0246596392,-0.2595549822,0.0205932595,0.5150591731,0.0020620688,-0.3844750226,0.3281677365,0.172640577,0.0208287444,-0.3904669881,-0.0901890993,0.1719796658,0.1882950962,-0.3696435988,-0.132772997,-0.074330546,0.004708779,0.4083339572,0.1221458837,-0.1973831058,-0.0800892711,0.3171685934,0.0733503178,0.1180296615,0.1058228314,-0.1125661582,0.0590074956,0.2582847476,0.0294451602,0.4560612738,0.0057619996,-0.0917912945,0.1343683749,-0.1761388928,-0.1272948533,-0.5407747626,-0.2045726031,0.1228743643,0.0578004941,-0.162427485,-0.1606598198,0.0118197193,0.0922975689,-0.2890814245,-0.377273798,0.1872646213,0.3017647862,0.0084913578,-0.2976927161,0.3289000392,0.3352466524,0.0406833738,-0.2348483056,-0.409171313,-0.1602902114,0.0041144663,-0.2967252433,0.3039312065,0.0120667508,0.2342775911,0.0612704977,0.2136326432,-0.2468576729,0.1188308373,0.0262546763,0.4117548168,-0.1158751696,-0.1725891083,-0.1143698618,0.4909881353,-0.0807861239,-0.3751322627,0.306275934,0.0404988676,-0.0982870907,0.2537679374,0.1508229822,-0.0326509029,-0.2881296873,0.1547313929,-0.1046459824,0.2012624592,-0.0296063349,0.0327399299,-0.0694689974,-0.1801565886,-0.3215429783,-0.0871651843,0.126500681,0.0232674461,0.0851306468,0.103120327,-0.0183848832,0.0764909834,-0.0063398634,0.2128142565,-0.0935365707,-0.3386485875,-0.1426964402,0.1622718275,0.1784218103,0.0436683483,-0.3368487954,-0.2921509445,-0.0895113721,-0.13959001,0.6184888482,-0.1524662226,0.1947831362,-0.3990027905,0.350575,0.2527207136,0.0121101253,-0.2960340381,0.1454516053,0.340301156,-0.3048078418,0.232601583,0.2621837556,0.0720755011,-0.2393237352,-0.110546492,0.3002930284,0.3226251006,-0.4334451854,-0.1008864418,-0.3000040352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2161","title":"any possibility to download part of large datasets only?","comments":"Is streaming completed? On the 1.8.0 docs it is mentioned (https:\/\/huggingface.co\/docs\/datasets\/dataset_streaming.html), but when following the example I get the following error:\r\n\r\n```\r\n>>> dataset2 = load_dataset(\"amazon_us_reviews\", \"Pet_Products_v1_00\", split='train', streaming=True)\r\n\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in ()\r\n----> 1 en_dataset = load_dataset('oscar', \"unshuffled_deduplicated_en\", split='train', streaming=True)\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)\r\n 339 if value is not None:\r\n 340 if not hasattr(builder_config, key):\r\n--> 341 raise ValueError(f\"BuilderConfig {builder_config} doesn't have a '{key}' key.\")\r\n 342 setattr(builder_config, key, value)\r\n 343 \r\n\r\nValueError: BuilderConfig OscarConfig(name='unshuffled_deduplicated_en', version=1.0.0, data_dir=None, data_files=None, description='Unshuffled and deduplicated, English OSCAR dataset') doesn't have a 'streaming' key.\r\n```\r\n\r\nUPDATE: Managed to get streaming working by building from source and installing the additional `datasets[streaming]` package:\r\n\r\n```\r\n!pip install git+https:\/\/github.com\/huggingface\/datasets.git\r\n!pip install datasets[streaming]\r\n```","body":"Hi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks","comment_length":123,"text":"any possibility to download part of large datasets only?\nHi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks\nIs streaming completed? On the 1.8.0 docs it is mentioned (https:\/\/huggingface.co\/docs\/datasets\/dataset_streaming.html), but when following the example I get the following error:\r\n\r\n```\r\n>>> dataset2 = load_dataset(\"amazon_us_reviews\", \"Pet_Products_v1_00\", split='train', streaming=True)\r\n\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n in ()\r\n----> 1 en_dataset = load_dataset('oscar', \"unshuffled_deduplicated_en\", split='train', streaming=True)\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)\r\n 339 if value is not None:\r\n 340 if not hasattr(builder_config, key):\r\n--> 341 raise ValueError(f\"BuilderConfig {builder_config} doesn't have a '{key}' key.\")\r\n 342 setattr(builder_config, key, value)\r\n 343 \r\n\r\nValueError: BuilderConfig OscarConfig(name='unshuffled_deduplicated_en', version=1.0.0, data_dir=None, data_files=None, description='Unshuffled and deduplicated, English OSCAR dataset') doesn't have a 'streaming' key.\r\n```\r\n\r\nUPDATE: Managed to get streaming working by building from source and installing the additional `datasets[streaming]` package:\r\n\r\n```\r\n!pip install git+https:\/\/github.com\/huggingface\/datasets.git\r\n!pip install datasets[streaming]\r\n```","embeddings":[-0.4476095736,-0.3885074854,0.0095246909,0.370123446,0.1067445353,0.1929703951,-0.0340346843,0.5148774385,0.0819868594,0.2324690074,-0.4534536302,-0.1990633309,-0.1784165353,0.4000229239,0.3254315257,-0.1704499125,-0.1151958555,0.1878741086,-0.0883976594,-0.1764554977,-0.0659911856,-0.0661999807,-0.1601547897,-0.1373348683,0.0878684372,-0.0127935819,-0.0002598203,-0.0606895387,-0.39527753,-0.0482899174,0.2791904211,0.1899072081,0.1114025861,0.1363454163,-0.0001268418,-0.1690050513,0.3633389175,-0.1690449715,-0.3607497513,-0.1897484213,-0.1767892838,0.086456649,-0.0258887466,-0.1968064457,0.0642866492,0.0536655597,0.0196072031,-0.020212071,0.4206509292,0.0790272653,0.0756611899,0.2862719297,0.106442757,0.1084069982,0.1896714866,0.1896163821,0.0424217843,0.1744011492,0.263699919,0.2539024949,0.0544782504,0.2674317658,0.0258967616,0.3122105598,0.0494102202,-0.183249861,-0.1826818734,-0.668645978,0.1587405503,0.6345419288,0.2820595503,-0.0176110677,-0.5097762346,-0.4816657305,0.1284909695,-0.6643292904,-0.068087779,0.6299393773,-0.4362098575,0.0515867434,-0.6224960089,-0.2577052712,-0.2082426399,0.3043964505,-0.195134297,0.1637154073,-0.2123502642,0.1319192797,0.264487505,-0.0100984648,0.0959196463,-0.2319997996,-0.0489966199,0.3651181459,-0.1721380055,-0.3230677545,-0.130076468,0.282348156,0.3912600279,0.2919591069,-0.0171136837,0.1264038086,0.1318633109,-0.0820960104,0.5853521824,-0.1074894518,-0.3332640231,0.2814321518,0.3189247549,0.3265119493,0.135591343,-0.0895507261,-0.1213611066,0.2551878393,-0.3615022004,-0.0669895336,0.0641150624,-0.3408076465,-0.0459385924,-0.1280491054,0.0305283554,-0.0864463896,0.0000250813,0.226681754,-0.0999816954,0.2417342663,-0.2276192307,0.0628773347,-0.0823991746,-0.2827228606,-0.0952995569,0.0019261892,0.1059893444,0.1242706925,0.2675984502,-0.5731204152,0.3885329366,-0.1137729064,0.5026563406,0.0964687839,-0.1554783434,-0.1604901403,0.0389188565,0.3139612079,0.0650518611,0.4057741761,0.0124905873,0.2547515333,-0.2002497166,0.2239141315,-0.1098953262,-0.4199577868,0.0124073839,0.0053586732,-0.3765225708,-0.040492624,-0.397038579,0.2838448584,0.0285333991,0.106795311,-0.0624929741,0.1174524948,-0.091144003,-0.0263898205,0.3707153797,0.3494339287,-0.0923917964,-0.2054376155,-0.1573995799,-0.0737660825,0.2902519405,0.4422823787,-0.3034178317,-0.1268008649,-0.2115474045,-0.259855032,0.3551361859,-0.085635215,-0.352812171,0.3833649755,0.1783619076,0.1554275453,0.1417041719,-0.0104912426,0.6427505016,0.0784268603,0.0167188384,0.496611774,-0.1728039831,-0.0436564982,-0.0887445584,-0.3199162781,-0.1930649579,0.3292202353,0.1675110012,0.3437730968,-0.0655097291,0.0158571266,0.4304685295,-0.0927524418,0.4422924221,0.1339258403,0.2232974321,0.0517347157,-0.2251435518,-0.4198052585,-0.390024364,0.2023004889,0.2302104235,-0.3973640203,0.1322660744,-0.1938214302,-0.0703584552,-0.2523659766,0.1331303269,-0.1378821582,-0.1556593478,-0.1328205913,0.3140981197,-0.1343347132,-0.5509424806,-0.0244986136,-0.0420947075,0.2612622678,-0.1101006791,0.2821855545,0.1781553775,-0.0102132494,0.2728771865,-0.1592219174,-0.0508325063,-0.1912722141,-0.048191987,0.216403842,-0.0373281613,0.0216384586,-0.1655334532,0.0964832753,0.3940747976,-0.002895595,-0.0280518048,0.1597703993,0.0947019085,-0.0528160967,-0.288987875,0.5896885991,-0.0793756694,0.1561117172,0.2797580659,0.0629937351,0.2583141625,0.024599582,-0.1124245524,-0.0111213,0.346318692,-0.129313767,0.3581708372,-0.1327632219,-0.5335001349,-0.2308386415,0.3453401625,-0.1733855456,-0.3991067111,0.2580744624,-0.2973130941,0.1224728078,0.1615042388,0.0103752734,0.0803766623,0.110991329,0.3299811184,0.0556583405,0.4505049288,-0.0324967541,0.0974358246,0.1050534099,-0.2600086629,0.0505798906,0.0576691367,-0.1339040548,-0.323603034,-0.029413078,0.1493064016,0.2342587411,-0.1711518914,-0.1821264625,-0.2360212654,-0.4625363052,-0.1564715952,-0.1158612967,-0.1308811605,-0.0992546082,-0.0840450302,0.6108608246,0.0186103415,0.0107477754,-0.0253544003,0.1656914949,0.0254362002,0.2290917486,-0.224337846,0.0105298162,0.155006811,0.0568729527,0.139343068,-0.0375017673,0.3074813783,-0.0599760674,-0.2525286078,-0.2569149137,-0.0906152427,0.0830261782,0.0150333764,-0.0439837649,0.0417102128,0.5142203569,-0.180791989,-0.1234094873,0.0148858782,-0.0761067271,-0.1766090542,-0.0104610734,-0.0691785663,0.229768768,-0.0946160257,0.0419856496,-0.4150296152,-0.3144565821,0.5148342848,0.376521498,0.2471687347,0.0415063798,-0.1072001234,-0.1911929697,-0.4565024674,-0.1922227144,0.0175258256,-0.8641309142,0.370621711,-0.048012685,-0.2978565097,0.0169300493,0.0122651653,-0.1107021272,0.3640224338,-0.385035485,0.1171999052,-0.2212017477,0.206400007,0.0881161764,0.1745103002,0.0010338839,-0.5461443663,0.1023331657,0.0984678641,0.0187489185,-0.2533030212,0.2585525215,0.2878212035,-0.1134539098,0.3037965596,0.010864621,1.0486694574,0.3577305079,0.1942342073,0.115386501,-0.0145859458,0.1851329356,-0.1911917329,-0.1458620876,0.3457380831,-0.078946881,-0.0668450072,0.105147019,0.0415116511,-0.1388587207,-0.3954806328,0.2484030426,-0.0438915007,-0.2528288364,-0.1172267795,-0.2193701565,0.0995958894,0.0906495526,0.0542370565,0.1259046346,-0.1767571121,-0.1605113596,0.2112127244,0.4497786164,0.0769428238,-0.2806784213,-0.0941862687,-0.5484695435,-0.0183815509,-0.0968109891,0.0207686555,0.0716721565,0.1369583011,0.1776129156,0.199952513,0.4251608551,0.0150515763,-0.0851411894,-0.1003859043,-0.3912237287,-0.3792966008,-0.0583369322,-0.061639905,0.0325773843,0.2290611118,0.2159029841,-0.598977685,0.1283589602,0.2132702768,0.4295913577,-0.0855615512,-0.0359848998,-0.1612869352,-0.25177145,-0.4768922627,0.3392892182,-0.0018541905,-0.0085675251,-0.1374035776,-0.1114352196,0.0696501359,0.2064140588,-0.0332829095,-0.0765376389,-0.054753013,0.1620873809,0.5258145332,0.2096888721,0.1231717542,0.1552083194,0.54124403,0.2921205461,-0.3287693858,-0.1230672076,0.1314025223,0.0337632336,0.2663985491,-0.0707625151,0.2665625513,0.4562000036,0.2169492692,-0.056978412,0.0430734195,-0.1117791384,-0.0314420462,-0.4366402328,-0.6649062037,0.183385551,0.028956417,0.171999529,0.5021618009,-0.0893522128,0.0872228816,-0.1249871701,-0.3470994234,0.8403228521,0.0230596904,0.2463889718,-0.0838512704,-0.0915785357,0.4362722635,0.3076654673,0.0302496515,-0.3009406924,-0.2791674137,-0.0544980429,-0.089717932,0.0475792028,0.0608512536,0.004792734,0.4569303393,0.0234942455,0.4431339502,-0.1506615877,0.094379127,-0.3789501786,-0.2833108008,-0.1951287687,-0.0439592488,-0.2036373913,0.5190579295,-0.0276786182,-0.0529743358,0.0900138989,-0.1041011214,-0.2623258233,-0.030528333,-0.3110882342,-0.0773403198,-0.1820992827,-0.3320752978,0.0449800305,-0.2058797479,-0.1879916936,0.0597037524,-0.2648640275,0.1744456738,0.0788602158,0.0530681983,-0.0842145681,0.1663140357,-0.1594961882,-0.0951340795,-0.0796273053,0.2896769345,0.0260673352,-0.1259272695,0.0134977773,-0.043205861,0.2284062654,0.0498592891,-0.0776964948,0.0118712038,-0.0022437058,-0.2475992739,-0.0125002842,0.0600453056,-0.2825704515,0.2127388269,0.1388477087,-0.0268463381,-0.2350724787,0.7145009041,0.0906659663,0.0234102327,0.2410631627,-0.068965748,-0.2382346243,-0.0698973313,0.0650196075,-0.162776649,-0.3265155554,0.1060188189,-0.08222536,0.4989001155,-0.3390577137,0.0803500786,0.1533051282,-0.0455191545,-0.0521174371,-0.2859503627,-0.4637408257,0.2339494973,-0.0310895294,0.1447184682,-0.2097663581,0.2193263769,-0.0144758299,0.2021393627,-0.1464585811,0.1956728846,0.0090670278,0.2211418003,0.3122569025,0.1734316796,0.0412191153,0.0369422287,-0.0457013585,0.1359823644,-0.1084390134,-0.0496176444,-0.0313955434,0.2314130217,0.1327960938,-0.5099074244,0.3159249723,0.0068709329,-0.3046574891,0.0213364884,0.3179292679,0.0578481853,0.2031617612,-0.1624803096,0.4404973686,-0.0921650678,-0.4527811706,0.2782374918,-0.1194799319,0.288382113,-0.0676694214,0.2628517449,0.2084853202,-0.1120108142,-0.2517091334,-0.2253290117,0.0856988877,-0.1128218621,0.4809917808,-0.0277797394,0.1876188666,0.209119752,0.5395689011,0.6694285274,-0.0530165806,0.2618055642,0.44467327,0.0809475631,-0.1898657531,0.1110779271,-0.0003325033,-0.0231842492,0.0412070975,0.0932054147,-0.1043451428,-0.2828885019,-0.2034022957,-0.2817724645,0.1338910162,0.2542670071,0.2420221567,0.7089776397,0.1716317683,-0.1210034266,0.0419054069,0.2615546584,0.1352959573,0.2326254547,-0.228507176,0.1435540617,0.0493179262,0.0102536976,0.0283562932,-0.697786212,-0.1262253672,0.3491155207,0.0734075308,0.179918915,-0.1520055532,0.2228636146,-0.072261408,-0.0845616758,-0.2514719367,0.4807581306,0.2058861405,-0.1214078292,-0.0505675003,-0.2208913863,-0.0029639138,0.4985956848,-0.00061906,-0.3852562606,0.3646084964,0.2382334918,-0.0054047285,-0.5793502331,-0.0796300694,0.1195175797,0.2672905326,-0.3746913373,0.128902778,0.1669923812,0.0234023836,0.2007475197,0.2626673877,0.0054997704,0.0036873338,0.1897207052,0.1832103878,0.2477828264,0.0157303736,-0.0511982366,0.0728417337,0.2460759431,0.1826988161,0.4314577579,-0.0748720244,-0.0686684027,0.0753027722,-0.146651715,-0.2283034921,-0.5099493265,0.0423510857,-0.1217955351,-0.0097378166,-0.1389461309,-0.106256038,-0.1694819331,0.0534895286,-0.1733472198,-0.1776193827,0.0898381844,0.1535759121,-0.0405295528,-0.2815022767,0.4673684835,0.3016182184,0.0266278572,-0.2878842354,-0.4712398052,-0.2998331785,-0.0577357747,-0.2856626213,0.1777127832,0.0457236692,0.3271953166,-0.1426305026,0.1498187035,-0.1828512847,0.0241614375,0.030470537,0.4348773956,-0.0860762894,-0.1851689219,-0.0583442934,0.4265660942,-0.1282648891,-0.2810760438,0.3487751484,-0.0216725096,-0.1689999998,0.2171521336,0.1006583497,0.0364659466,-0.2001003325,0.0856572986,-0.0234422944,0.3311543763,-0.0020098526,-0.0788414925,0.0014260238,-0.1774407029,-0.3367545307,0.1356220692,0.0134055596,0.2751170099,0.2030151188,-0.0970651656,-0.2472084463,0.2357616872,-0.093008101,-0.1959493905,-0.0243992619,-0.2596942484,-0.1272242814,0.2460621893,0.1297127157,0.1128144339,-0.3708875477,-0.2948181033,-0.2648235857,-0.2004150152,0.7916323543,-0.2839847207,-0.0090406295,-0.3863787055,0.2918309867,0.2459666878,-0.0674035698,-0.3543948531,0.2752070427,0.4130686522,-0.3291516006,-0.0016187538,0.2715638876,-0.0741296262,-0.0661868677,-0.06462951,0.3913561106,0.2036433667,-0.4596029818,-0.1089767888,-0.230786413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2161","title":"any possibility to download part of large datasets only?","comments":"Hi ! Streaming is available on `master` only right now. We'll make a new release 1.9.0 on Monday :)","body":"Hi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks","comment_length":19,"text":"any possibility to download part of large datasets only?\nHi\r\nSome of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled\/unshuffled data without going through first downloading the whole data then sampling? thanks\nHi ! Streaming is available on `master` only right now. We'll make a new release 1.9.0 on Monday :)","embeddings":[-0.4722626507,-0.4255569279,-0.1131871417,0.1344200224,0.0880462602,0.1738644838,-0.3542895317,0.3757156134,-0.0502581708,0.3286562264,-0.4129242003,-0.227931425,-0.1485811472,0.3793706298,0.2008761317,-0.0660857111,-0.1095422283,0.2192226052,-0.1483973712,-0.113596186,0.0163903777,-0.2493553013,-0.2375249416,-0.3178961575,0.2018005103,0.0277642515,-0.0485128127,-0.2904225588,-0.5257316232,0.0757009089,0.1052192748,0.1233388856,0.213425979,0.1093875691,-0.000115171,-0.3466500938,0.416819036,-0.1254801452,-0.0490663573,-0.1248012558,-0.3066554964,0.130421102,-0.1911244839,-0.1625910848,0.0535930432,0.1501369327,0.2008933574,0.069564037,0.1805625111,0.1309281439,0.1482643336,0.0635441318,-0.1145861596,-0.0245387964,0.2895706594,0.067812182,-0.0297182202,0.0065403683,0.5009749532,0.3647523224,0.1164343059,0.0436584316,0.0524160825,0.1226052493,0.0101258187,-0.3156893551,-0.1707394719,-0.6554211974,0.3526365161,0.7226455212,0.6469684839,0.101222679,-0.2085970789,-0.0435767472,0.17564255,-0.2512713373,-0.1427433491,0.6370520592,-0.3258793354,0.1487766057,-0.5430160165,-0.3155020773,-0.1687672138,0.304164201,-0.2599721551,0.3214052022,-0.0001990961,-0.0499977507,0.360345006,-0.002015355,0.2264091223,0.0228020065,-0.3124683499,0.2853650451,-0.2960832715,-0.3794879317,-0.3436775804,0.2594523728,0.2998431623,0.0891993567,0.1823874265,0.2357151657,0.0800661519,0.0151364719,0.5451431274,-0.107617259,-0.323325932,0.0730135068,0.3837715685,0.0101135392,0.1688045412,-0.0248710178,-0.0028995492,0.2587714791,-0.5076771379,-0.0003627476,-0.2615731955,-0.4713282585,0.1559984088,-0.2777479291,0.1362636089,0.0707588866,-0.0810322016,0.107066147,-0.0906696469,0.241054073,-0.4281466305,-0.0736106336,0.0975691974,-0.322091043,-0.1091440171,-0.0105769159,0.0280527826,-0.0787927359,0.2643789351,-0.5082525611,0.3371801674,-0.0795179009,0.3361353576,0.1806773543,-0.189269647,-0.1488175243,-0.0888085067,0.2145675272,0.104994379,0.2417758852,-0.2071612328,0.5045181513,-0.0368065313,0.2993345857,-0.1207159311,-0.283534795,0.1389861703,0.1309602708,-0.2674580514,-0.062985599,-0.3869692683,0.2854522467,-0.1779265255,0.0000469603,-0.0142693501,0.0776571557,-0.0435365401,-0.1150779128,0.2379499823,-0.1307851672,-0.2373149991,0.0068901898,-0.1620619893,-0.1714998037,0.4258080125,0.3534020483,-0.1707287431,-0.3960533142,-0.0673998818,-0.1546263099,0.2148637474,0.0437496714,-0.538692832,0.3985548913,0.131785959,-0.2564815879,0.0821295381,0.152552709,0.6344279647,0.1193681955,-0.2234910131,0.5501548648,-0.0589089245,0.0575012118,-0.095140256,-0.3214320838,-0.2185790986,0.1635003984,0.3424953818,0.4317515194,0.2950881124,0.0950777158,0.4195162654,-0.0896089002,0.2856127918,0.1713019013,0.1827346236,-0.1341881752,-0.0874409676,-0.3088537753,-0.0879281759,0.1430228949,0.1239087284,-0.3638298512,0.4067093432,-0.1922576278,0.088481389,-0.2914109528,0.4060345888,-0.0655182824,-0.0696879029,-0.178422913,0.2960861921,-0.2208670527,-0.3730580509,-0.2006771713,-0.0802235082,0.0889411271,0.0760953501,0.0453677736,0.2278527915,0.1880146265,0.3012546897,-0.3396292031,0.018282637,-0.0757414475,0.032717552,0.2482907921,-0.0751573667,0.1401787251,0.0556409135,0.0482327528,0.4031480849,-0.1497116387,-0.0283271428,0.3423365057,0.1293026507,-0.0366101377,-0.3519751728,0.2771193087,-0.2177254111,0.1233278662,0.1699346602,-0.002215513,0.3464900255,-0.0757288709,0.0770131797,0.1272966266,0.3411634564,0.037285123,0.0722674429,-0.1566409618,-0.369949609,-0.257492274,0.0630850568,-0.1521055996,-0.26229617,0.1761300564,-0.3542080522,0.0377749689,0.1432314068,0.1573666632,0.0434849299,0.1828321368,0.5009159446,-0.0189324152,0.5529599786,-0.1913037747,0.0323056504,0.0135213435,-0.367751658,0.0108494833,0.0776748508,-0.3239637613,-0.1185696498,0.2415394038,0.2291006446,0.1548428237,0.1494725496,-0.4640162587,-0.3105860651,-0.5512290001,0.1515311897,0.0502730459,-0.0246748868,-0.0118748834,0.0819526687,0.4645318687,0.0342576206,-0.1236094907,0.0287054069,0.3785234392,-0.1320492923,0.4409327209,-0.1768302023,-0.0413004272,0.2356285155,0.2211539149,0.0218223985,0.0270522088,0.1783610731,0.0761639476,-0.0646819621,-0.3325956166,-0.063948743,-0.0094431862,0.2617275417,-0.1962979883,-0.1334728152,0.3212757707,-0.0662328526,0.0288736112,-0.2167477161,-0.3117702305,-0.1400713623,-0.0669214502,-0.0051100315,0.1507525146,-0.1814942956,0.1428144574,-0.4396058321,-0.3790428042,0.3784822226,0.2214270681,0.1600247025,-0.3005808592,-0.121088028,-0.187777549,-0.2930058539,-0.4380383193,0.0462037064,-0.8105823994,0.322945267,-0.1686301529,-0.3520539105,0.1457237601,0.2076426595,-0.1211160719,0.4795018733,-0.4310851395,0.2508367002,-0.11923296,-0.1019102409,0.187668398,-0.0328274444,-0.2147152424,-0.6559681296,-0.0396956056,0.0308427792,0.2113589942,-0.3204505146,0.3215064704,0.3877242208,-0.0777190253,0.0027181855,0.1230207086,0.6648045182,0.0610291958,0.3236869872,0.0348335579,0.1735765487,0.039022468,-0.0626723468,0.0875246301,0.3505371809,-0.0765626058,-0.180175066,0.2538129687,0.0923455134,-0.3871078789,-0.4046891332,-0.0492944643,0.055689659,-0.0435905159,-0.104250066,0.0307580549,-0.0279367976,0.1188505664,0.0599452928,0.1685609668,-0.3120358586,0.0869186819,0.0748439059,0.308852911,0.0074383579,-0.3863292038,-0.1485591829,-0.3613516092,0.0631624907,-0.1004730612,0.0825573504,0.0680252016,0.16436252,0.2066699564,0.1446602792,0.3816034794,-0.0374659821,0.0500220656,-0.1897474527,-0.2201120555,-0.0895118937,0.0268401802,-0.0417056754,-0.0965519771,0.3259879649,-0.1407384574,-0.3543405831,0.288957268,0.0056223888,0.3419517577,-0.0903491303,-0.0451270305,0.1183281243,-0.0427905992,-0.1847223043,0.252240032,-0.2072150707,-0.163226217,-0.156011194,-0.1090735793,-0.0318549573,0.0216325615,-0.0828885734,-0.1000516564,0.0308272485,0.045329418,0.3428443074,0.1778293997,0.0819712058,0.1284077168,0.4695055485,0.3658311963,-0.1156463996,0.0867108628,-0.1306062192,-0.1015803069,0.2093808353,0.0437398441,0.2808165252,0.4085558057,0.2124495953,-0.0312629268,0.0622299835,-0.2949734926,-0.094150953,-0.5202360153,-0.5601618886,0.3108978868,0.0965794548,0.0867082104,0.4958634973,-0.3096118569,0.1638137251,-0.0990545675,-0.1506313086,0.889123559,0.0754089206,-0.006490293,-0.5325066447,-0.0284913126,0.420017153,0.0276410989,0.135617733,-0.1975045353,-0.4614289105,-0.1423406899,-0.0167069957,-0.0932056531,0.0716485307,-0.066296339,0.3452250361,0.0734825581,0.426235944,0.0041123354,0.0317639485,-0.3003113568,-0.1115790382,0.2346546054,0.0693436041,-0.1433306038,0.3273025751,-0.0557925515,-0.0327083692,0.0900519565,0.1534625143,-0.1626336724,0.1220900789,-0.2919662893,-0.1220971197,-0.4284983277,-0.2095222324,-0.0900093392,-0.1598198265,-0.1767659187,0.224577859,-0.3927684724,0.277068913,0.2208954096,0.0614601895,-0.0097395172,0.2679329515,-0.3007406294,-0.1530026048,-0.1218153909,0.3161703348,0.016243482,-0.2604296505,-0.119807139,0.0095501887,0.0472750105,0.2062589228,-0.0787362382,0.0608953796,-0.0678573623,-0.1324035525,0.0844364613,0.1405863613,-0.2451874763,0.2926476002,0.2806528211,-0.20300816,-0.3127925098,0.7061346769,0.2217414826,-0.1477390528,-0.0309071057,-0.0348444916,-0.2691622972,-0.1094613299,0.0344101675,-0.2847452462,-0.0970422402,-0.0436131768,-0.3754636049,0.2451198101,-0.0366515853,0.0363682657,0.0863700286,-0.009860429,-0.2653485835,-0.3093075752,-0.3473717868,0.1356408596,-0.0308900215,0.1335929781,-0.1747771353,-0.0131152049,0.0001685877,0.2695533633,-0.2037924528,0.2694521248,0.1999200433,0.2387233526,0.2172998786,0.3061388135,-0.191740185,-0.019398639,-0.0467092842,0.2147588581,-0.0651555434,-0.1116472036,0.0771595314,0.186522603,0.1970794499,-0.3018146455,0.108591117,-0.1156174168,-0.211378783,0.1027678847,0.1767237484,-0.0382391699,0.2515648603,-0.3462998569,0.5911293626,-0.2328620255,-0.4694001377,0.4542548656,0.0165076051,0.0974758118,-0.0787201896,0.3140096068,0.2915429175,-0.1683987677,-0.1722799838,-0.2812139392,0.0852772892,-0.1030148268,0.374987781,0.2042786628,0.3214718699,0.2236579061,0.3470141292,0.5542971492,0.1461546868,0.1856752634,0.4167735279,0.1880724579,-0.1497303545,-0.1000321805,-0.1143854335,0.1440722197,-0.0257479697,0.0018495834,-0.0225771144,-0.3263655901,-0.0515492745,-0.2704437077,0.0888491347,0.1533000618,0.1793412119,0.8333560824,0.1531047523,-0.2371095866,-0.1882935911,0.2068772912,0.1592477113,0.3730365932,-0.0479566753,0.3059625328,0.3063521385,-0.0312642083,0.0032447195,-0.5970546603,-0.2141679972,0.4046422839,-0.0142368283,0.209254235,0.0344609767,-0.0481589027,-0.0106614325,-0.0931055099,-0.1816472709,0.2660160065,0.2127107531,-0.0039871996,-0.1150406525,-0.2785249352,0.109497197,0.5113762021,0.1322319657,-0.1980447471,0.2608272731,0.2490079403,0.0308741946,-0.2651609778,-0.1359802336,0.0332691222,0.1771013737,-0.3987109363,-0.1482173204,0.0234416183,0.0628857538,0.439426899,0.0644202083,-0.1614291966,0.058764942,0.2105546445,0.0568031482,0.1634789556,0.0355565771,-0.1301715374,0.0776371509,0.4430594444,-0.0092204195,0.2910925448,0.0585153438,-0.0959537104,0.3714140952,-0.2060945779,-0.0672065318,-0.3228953481,-0.1336226761,0.1930322796,-0.0391326472,-0.0094226161,-0.1811572611,-0.1606294662,0.1833277643,-0.3444312513,-0.4696622193,0.1643548459,0.1820184439,0.0193004627,-0.2334381491,0.4481795728,0.1579576731,0.0105547132,-0.0990974829,-0.4864680469,-0.2503871024,0.0959639773,-0.1351801902,0.3514735699,-0.0042077405,0.2430803329,0.0163555499,0.2670602798,-0.1235209629,-0.0131922262,-0.0454170778,0.4119945765,-0.1600772291,-0.0816618577,-0.0142872278,0.5789083242,-0.1545487195,-0.2790962458,0.443695128,-0.1308503598,-0.0096990392,0.3190704286,-0.0506643243,0.1381157637,-0.1956164986,0.1083929315,-0.2287507206,0.1634789705,-0.121740967,0.0967456773,-0.0047981553,-0.1323550642,-0.2796777487,-0.0493525416,0.2281526327,0.0808744878,0.1648784429,0.074176982,0.016876461,0.1159249619,-0.1904602498,0.1538941115,-0.0167449489,-0.5041579008,-0.2152561247,0.0947389379,0.1537460834,-0.0536784045,-0.2692047358,-0.5225881934,-0.1821423322,-0.0793164298,0.5730479956,-0.0971038118,0.2402117401,-0.4517636299,0.1212144718,0.3278099597,0.1860437691,-0.1825704724,0.136777088,0.2774830759,-0.2497419566,0.2136400491,0.1351708919,0.0612913035,-0.1135808229,-0.1727844626,0.0957112536,0.3469174504,-0.4755298197,-0.2574577928,-0.2557221651]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2160","title":"data_args.preprocessing_num_workers almost freezes ","comments":"Hi.\r\nI cannot always reproduce this issue, and on later runs I did not see it so far. Sometimes also I set 8 processes but I see less being showed, is this normal, here only 5 are shown for 8 being set, thanks\r\n\r\n```\r\n#3: 11%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 172\/1583 [00:46<06:21, 3.70ba\/s]\r\n#4: 9%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f | 143\/1583 [00:46<07:46, 3.09ba\/s]\r\n#7: 6%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 98\/1583 [00:45<11:34, 2.14ba\/s]\r\n#5: 8%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258d | 124\/1583 [00:46<09:03, 2.68ba\/s]\r\n#6: 7%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f \r\n```","body":"Hi @lhoestq \r\n\r\nI am running this code from huggingface transformers https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py \r\n\r\nto speed up tokenization, since I am running on multiple datasets, I am using data_args.preprocessing_num_workers = 4 with opus100 corpus but this moves on till a point and then this freezes almost for sometime during tokenization steps and then this is back again, overall to me taking more time than normal case, I appreciate your advice on how I can use this option properly to speed up.\r\n\r\nthanks","comment_length":71,"text":"data_args.preprocessing_num_workers almost freezes \nHi @lhoestq \r\n\r\nI am running this code from huggingface transformers https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py \r\n\r\nto speed up tokenization, since I am running on multiple datasets, I am using data_args.preprocessing_num_workers = 4 with opus100 corpus but this moves on till a point and then this freezes almost for sometime during tokenization steps and then this is back again, overall to me taking more time than normal case, I appreciate your advice on how I can use this option properly to speed up.\r\n\r\nthanks\nHi.\r\nI cannot always reproduce this issue, and on later runs I did not see it so far. Sometimes also I set 8 processes but I see less being showed, is this normal, here only 5 are shown for 8 being set, thanks\r\n\r\n```\r\n#3: 11%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 172\/1583 [00:46<06:21, 3.70ba\/s]\r\n#4: 9%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f | 143\/1583 [00:46<07:46, 3.09ba\/s]\r\n#7: 6%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 98\/1583 [00:45<11:34, 2.14ba\/s]\r\n#5: 8%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258d | 124\/1583 [00:46<09:03, 2.68ba\/s]\r\n#6: 7%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f \r\n```","embeddings":[-0.251410991,-0.2479194999,-0.1668762267,0.0695675686,0.1297327876,-0.1807381213,0.4334701002,0.0990605354,-0.3538162112,0.2440417558,0.0638783947,0.239945367,0.0941158682,-0.1330185235,-0.0470905453,0.1516224444,0.1583945155,0.01122397,-0.1086994112,0.1517171264,-0.3039160669,0.2973909676,-0.1818717718,-0.1495001316,-0.3720234931,0.0712001994,0.1070430055,-0.0386771485,0.1266263425,-0.3786482513,-0.3722394109,0.2621736825,-0.1717153788,0.3977265656,-0.0001282334,0.0551564693,0.3310023546,0.334130466,-0.0993837044,0.1854704767,0.5246136189,-0.3683948219,0.0892792493,0.0601530075,-0.0127817662,0.0405185968,-0.0577857569,-0.1024643332,0.4945966899,-0.0235326365,0.0106404126,0.5192914009,-0.1219599321,0.2335590869,-0.3205538392,0.1153099611,-0.0788028166,0.1173704341,0.4010337293,0.197347939,0.3235229254,0.1297678053,-0.1178333163,-0.0455494262,-0.0672251657,0.2181202024,0.4664070308,-0.646307528,0.254032135,0.1169221625,-0.0574613065,0.1287365109,-0.0659624562,-0.2802684307,-0.1109629199,-0.5052868128,0.0478969365,-0.286601454,-0.0035855249,-0.0845630616,-0.1638436168,-0.03434553,0.0380558297,0.0745908245,0.2919778228,0.402382344,-0.2579991817,0.1412986517,0.4859220386,0.0520768389,-0.1439908743,-0.2433384508,0.1748287976,0.1739303768,-0.3967978656,0.094879739,-0.2360816002,0.1659679711,-0.1208465323,0.0911727697,-0.1102677658,0.51345402,0.4665889442,-0.2004278302,-0.1314332783,-0.1684507728,0.0231272895,-0.3938961029,0.5149444938,0.0773059353,0.012577232,0.0659774691,-0.242899552,-0.5354476571,0.1613685787,0.1407512277,0.0415939949,-0.1196804419,0.209046483,0.080072321,0.1907866448,-0.2290394753,0.3454236686,0.3752199113,-0.2789570987,0.1992174536,-0.2575200498,0.0450633168,-0.5872555375,-0.0557829179,0.0518546924,-0.2311387658,-0.0153393801,0.1887843162,0.2071906477,-0.2020408809,0.196659416,0.2144078612,0.6740496755,-0.422600776,0.3631923199,-0.1636646241,-0.0096500507,0.2447361499,-0.0727551579,0.3624317944,0.0352872498,0.3586993515,-0.3110885918,0.0150666256,0.0074116229,-0.2275886089,0.1634588242,-0.0480351113,-0.0323655978,0.2433685958,0.0056008692,0.1432507187,0.5009929538,0.38757056,0.0300536565,-0.1195357144,-0.4238312244,0.0253996365,0.2834200859,0.4679894447,0.5233025551,-0.0402477235,-0.0364370383,-0.0225381311,0.7288458943,0.2094704658,-0.0243388563,-0.0221082605,-0.0677255467,0.1443639994,-0.0810829997,0.072325848,0.0567137636,0.2270413637,-0.224065423,-0.0860366151,0.2414156049,0.0726291463,0.1347102076,-0.0877699628,0.1700471938,0.1761258692,0.2823870778,0.1734233946,-0.2352311164,0.014629432,-0.0128331911,0.0171705708,-0.3240709007,-0.0673821121,-0.5180609226,-0.0416161567,0.0703363642,0.005442726,0.1118422002,0.1537818164,-0.0834670961,0.0560844056,0.327490896,0.0175311863,-0.1221803427,-0.037940491,-0.3193322718,0.2669684887,0.0784632713,-0.0842257962,0.0320651904,0.0909706578,-0.1867544204,-0.0062119798,-0.1380517185,-0.0347193703,0.1920596659,-0.0670048296,0.0034454467,-0.083128348,0.0475848839,0.2688089609,-0.2574956715,0.0044197752,0.0832877159,-0.1455596387,-0.1100080982,0.0941949785,0.0614338331,-0.3610614836,0.0165602807,0.2966277599,0.1466055065,0.3110172749,-0.4479559958,0.5583019853,0.2145093381,0.1290580332,0.0892316625,0.0092792502,0.021773586,-0.1451040804,0.4335660338,0.4299760759,-0.2997783124,0.390612781,0.1070610359,-0.1821053326,-0.0560044833,0.2489263564,-0.2439727336,-0.0261756238,0.454087466,-0.0477517769,0.2301163375,0.22236453,-0.1836317331,0.2125936747,0.123734735,0.0111783529,-0.1703724861,0.0403767228,-0.0734566823,-0.1334791332,-0.2499898225,-0.6115149856,0.1777780205,-0.0069288849,-0.0365603045,-0.0478599966,0.0042380751,-0.206107989,0.2798293531,0.4372163415,0.186759606,0.1711970568,-0.1949297041,-0.0732352957,-0.1825201064,-0.129712075,0.1186935604,0.1672266275,-0.1895805597,0.1366368383,-0.2387172282,0.2938965261,-0.166299969,-0.1849084646,-0.4397236109,-0.0257224683,-0.1499881893,0.0983165428,0.010332007,0.1029523015,0.0242338125,0.08819855,-0.1002230793,0.0622529797,-0.1612755358,0.0371611714,-0.0491185896,-0.245304063,0.2989142239,0.1687112451,-0.0080339629,0.1353286952,-0.3070052266,-0.0110725723,-0.2749910653,0.1400469542,0.1708554476,0.3440091908,-0.0735172853,0.1897306591,0.0009131254,0.0276050251,0.100685887,-0.1594691724,-0.0642061532,0.0215204302,-0.1502985209,-0.124828808,-0.1320841461,0.0152242919,-0.2438078076,-0.1898180544,0.1589920819,-0.0782173648,-0.0424328186,0.1317870468,-0.0138681279,0.1339338869,0.259021163,0.1452918202,-0.1662133634,-0.0018491204,0.3186675012,-0.102850832,-0.1215235814,-0.3235512972,-0.1072703972,0.1314612776,-0.2445207685,-0.060911905,-0.015192505,-0.5067573786,-0.242324397,-0.1723922938,0.0972394496,0.4248436689,-0.0066832025,-0.0099411821,0.1922746897,-0.226909861,0.257029742,-0.0619243681,-0.3097364902,0.1396402568,0.470805198,-0.2967545688,0.6741126776,0.0749658123,-0.1954611242,0.050569091,0.0368689932,-0.1507929116,-0.0553347506,-0.0852852687,0.1318517178,0.2249401212,-0.0552713312,0.3454117477,-0.1059589759,0.0814157128,0.1267296374,-0.0824176148,0.0170042682,-0.3068866134,0.3244253695,0.144149974,0.402965188,-0.0234128833,0.1108589619,-0.0862727836,-0.4016608596,-0.0962364674,0.1196677536,0.2414515764,-0.1695468873,-0.6161898971,0.0137220714,-0.7858516574,0.382829845,-0.0458903648,0.5199118853,0.1082283705,0.0056625181,0.3145137429,-0.3467053473,0.7670843005,0.1822910011,-0.1279033571,0.0962365493,-0.7098243833,-0.0582463518,0.0075299405,-0.0926197469,0.2724966705,0.356678158,0.7060548067,-0.3017215431,-0.3776777387,-0.0767088383,0.0563512817,0.0015053817,-0.4166152477,-0.4019541442,0.0501906723,-0.2863701582,0.5134526491,0.0906587094,0.1405481398,-0.2109410912,-0.033237692,0.1202216819,-0.2873510122,0.4236606658,0.2079537958,-0.0913097039,-0.0237024985,0.380682379,0.4786849022,-0.2290195525,0.1489964724,-0.1707217097,-0.0779557973,-0.2364804,0.3810600936,-0.0867978185,0.5450720787,0.547257483,0.1362940967,0.4063150287,0.3780595064,0.4374063313,-0.2749789953,0.1259635836,0.1547682583,0.5308867693,-0.413048625,-0.0818807632,0.1402872652,0.2081383616,-0.0229310151,0.056151133,-0.1762442738,-0.0634084716,0.2071321458,0.3916881979,1.1319563389,-0.3283146322,0.1856960654,-0.4327244461,-0.0603081286,0.5704905391,-0.3324199319,0.2379963249,-0.365690887,-0.2717865109,0.1850220263,-0.3282362223,0.3038600087,0.1189042404,-0.2628186047,0.1521394849,-0.3306327462,0.0432289056,-0.4142543077,0.3191807866,0.0404136591,-0.6195269227,-0.0495103709,-0.0282824598,0.0733607784,0.2955296636,-0.105653964,0.0835872665,0.0109440098,0.0497659594,0.0896113962,-0.1360662431,-0.1693646759,-0.0388213433,0.58214885,0.0324064977,0.3180068731,-0.0777362809,0.5598856211,0.2434166968,0.0780923292,-0.0341118984,-0.1588768512,-0.146806255,-0.051588621,-0.0682883933,0.0123428861,-0.2124676108,0.0511397682,-0.2381158769,0.0821539909,0.0058843084,-0.163651973,0.0166181326,0.0232439879,0.0467579216,-0.1821012348,0.3080946803,-0.3214458227,-0.1167131662,-0.0423149541,0.2404101789,-0.4403895438,0.3303352892,-0.0288399551,-0.1738937646,-0.1171052083,0.0892917216,-0.0264957119,-0.4030739963,0.6092650294,0.2181994766,-0.2666139901,-0.0854160339,-0.2937756777,-0.1068665311,-0.3492948711,0.3741152883,0.3237866461,-0.2295299321,-0.0780584738,-0.1986997277,-0.3508916497,-0.1182246059,-0.310280323,-0.2950769067,-0.4665006995,-0.3801025152,0.0411475524,0.1233233362,-0.0471605994,0.2644374371,-0.389664799,0.0624550432,-0.078579329,-0.1725250781,-0.4062476754,0.089981705,-0.1120749637,-0.5190102458,-0.0281905197,0.1142883301,-0.0714788884,0.4420267344,-0.1587402076,-0.0638981536,-0.2271942645,0.2349816114,-0.2800927758,-0.4436420798,0.3615058959,0.3001589775,-0.2331470698,-0.2703569829,-0.1595397741,-0.1643766761,-0.1993987709,-0.5072793961,0.3599963486,0.1198562086,0.0199191794,0.1914347708,-0.0674523637,-0.000858234,0.1864865869,0.3077498078,0.0294049997,-0.0558889285,-0.0084525533,0.0732794926,0.6394947767,-0.336330533,0.0188858919,0.0433635898,-0.1731341332,0.1465682834,-0.0477737747,0.5118207335,0.1813317686,-0.2633015215,0.3250022829,-0.021706989,0.0180752259,-0.0700403601,0.1865508556,-0.0227422509,-0.0238569491,0.4588422775,0.3922025561,0.2405369729,0.1528814733,0.255594492,0.1636410803,-0.3829538226,-0.1078640074,-0.1812138408,-0.0154022034,0.2450419664,0.4365407228,0.0375350676,0.150581494,-0.1797694117,0.1335874051,0.2507675588,0.2416340858,0.4925803244,-0.1382800937,-0.2254936099,-0.4174551666,0.5522565842,-0.0057640905,0.3790930808,-0.0028133672,0.3359464705,-0.0554628149,-0.0098855263,-0.1745776087,0.5061416626,-0.1261597276,-0.4475290775,0.1426978707,0.1112100109,-0.2979697287,0.0771852285,0.1318656653,0.0256801974,-0.0423040129,0.1080408618,0.0119534582,-0.1828671396,-0.2537564337,-0.1672242433,0.2278695256,-0.0983467847,0.0125127248,0.0452595428,0.0097416127,-0.0199356824,-0.0247959699,0.5128124952,0.2004427314,-0.153748557,0.0811640769,-0.1068250537,0.3018487394,-0.1513091326,0.4270680547,0.0418999605,0.300124079,-0.1047819778,-0.2305849791,-0.15402852,-0.5271900892,-0.0419679806,-0.043149557,0.0332791395,0.2451081425,0.1459529996,-0.175697431,-0.2253252715,0.1532504708,-0.2153151482,-0.0614371635,0.3286037445,-0.3338750601,-0.1148736998,-0.0067679039,-0.0201614015,-0.1126109436,0.6130872369,-0.0589852333,0.3415980637,-0.1647949815,-0.143643707,-0.3964852989,-0.1860513985,0.2374784648,0.3587355912,0.1647129953,-0.1060245037,0.0867478102,0.0671766847,-0.0712862536,-0.3454673886,-0.0114509035,0.3094774485,-0.3384152353,0.3538843393,-0.4342478812,0.0272095967,0.1970861852,-0.1175379008,0.2873236239,0.1683010757,-0.1999628097,-0.3245337307,0.0758442432,-0.2453704327,0.2596354783,0.0963902324,0.3184599876,0.1464158297,-0.015788896,-0.0109736165,0.0959520638,0.2608543932,-0.0697439387,0.0721201822,0.1049503461,-0.1648038626,-0.146745488,-0.1313747764,-0.2452248782,0.0567507409,-0.0118050212,-0.1745484024,-0.0568415485,0.046912957,-0.076870583,0.2094804049,0.1592630595,0.4812909067,0.2651977539,-0.0506084152,-0.3223831952,-0.0804887637,0.425472945,-0.2499652356,-0.392942518,-0.6206628084,0.1766658723,0.0027146258,0.1121188328,-0.2380702645,-0.1182984635,0.015711043,-0.1303830147,-0.2840644717,0.4506074786,-0.1115772501,-0.2529273629,-0.2448707223,0.159426868,0.1453019083,0.1756260693,-0.1135017946,-0.256218344]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2160","title":"data_args.preprocessing_num_workers almost freezes ","comments":"closing since I cannot reproduce it again, thanks ","body":"Hi @lhoestq \r\n\r\nI am running this code from huggingface transformers https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py \r\n\r\nto speed up tokenization, since I am running on multiple datasets, I am using data_args.preprocessing_num_workers = 4 with opus100 corpus but this moves on till a point and then this freezes almost for sometime during tokenization steps and then this is back again, overall to me taking more time than normal case, I appreciate your advice on how I can use this option properly to speed up.\r\n\r\nthanks","comment_length":8,"text":"data_args.preprocessing_num_workers almost freezes \nHi @lhoestq \r\n\r\nI am running this code from huggingface transformers https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py \r\n\r\nto speed up tokenization, since I am running on multiple datasets, I am using data_args.preprocessing_num_workers = 4 with opus100 corpus but this moves on till a point and then this freezes almost for sometime during tokenization steps and then this is back again, overall to me taking more time than normal case, I appreciate your advice on how I can use this option properly to speed up.\r\n\r\nthanks\nclosing since I cannot reproduce it again, thanks ","embeddings":[-0.289732486,-0.159812659,-0.1597204208,0.0191832576,0.1144234538,-0.1817658097,0.3889515698,0.1026587486,-0.3467479646,0.3008647561,0.0742085427,0.2872329354,0.017077446,-0.1225069314,-0.0719711259,0.1078776866,0.123131372,0.066957131,-0.1839236319,0.1967520416,-0.2445827574,0.2267065346,-0.2135546654,-0.2097691745,-0.3322158456,0.047991477,0.1297815144,-0.0806673691,0.1030090377,-0.3698817492,-0.2307940871,0.2753131092,-0.1602369845,0.3430786431,-0.0001276039,0.0871993527,0.3303315639,0.2890488207,-0.0870506018,0.111837998,0.4225260317,-0.2943085134,0.0401739255,0.0864176825,-0.0937952995,0.0992121324,-0.0106443455,-0.1057411805,0.4724691808,0.0045577637,0.0187484231,0.5315389633,-0.0726396888,0.2117922604,-0.390444845,0.1048824787,-0.0864721984,0.0996075571,0.4586623311,0.1920196265,0.2899696529,0.1865534782,-0.1239831522,-0.0892236009,0.001812232,0.1301543415,0.4036654532,-0.6181378365,0.2636239827,0.0923883691,0.0155627625,0.1136952862,-0.0967665017,-0.156528756,-0.0076287771,-0.3764639497,0.0585646778,-0.3168759644,-0.0250802264,-0.0525295734,-0.1769800633,-0.1159099191,-0.0528592281,0.1183252335,0.2295303941,0.4435490668,-0.2357753962,0.1289633662,0.5178619027,-0.0403327309,-0.2003175914,-0.2182147056,0.1802338511,0.1894382834,-0.3426503539,0.0215373803,-0.2802378535,0.1376956403,-0.0980366841,0.0990803093,-0.1663493365,0.4915067255,0.5625671744,-0.1862709224,-0.1565736234,-0.2199796587,0.0501049757,-0.4182260036,0.5298476219,0.0648964494,-0.0348339118,0.0603475682,-0.2815020382,-0.5361596942,0.1581793576,0.160339579,0.11943762,-0.1508014202,0.1970660985,0.038764257,0.1154877692,-0.1716931015,0.3841070831,0.3893547356,-0.2840166986,0.0653187335,-0.2809877396,0.0366551802,-0.5343164206,-0.0611388832,0.0477570593,-0.1661622971,0.0969536752,0.1505824775,0.1994995773,-0.2551729381,0.1551715136,0.2201011628,0.7152233124,-0.4002840519,0.3236929178,-0.1627223939,-0.0176958833,0.2033345103,-0.0905227512,0.3392412066,-0.0234102029,0.3217682242,-0.3174622357,-0.0208753254,0.0225392655,-0.2838594019,0.1334025264,-0.0416642167,0.0256379601,0.1733347327,-0.0249668062,0.190271616,0.4915558398,0.4496014118,0.0779029652,-0.0785411149,-0.440782398,0.0441308208,0.3039888442,0.4378511012,0.5238133669,0.0001719757,0.0905392841,-0.0069248835,0.6427304745,0.1498324573,-0.0883846581,-0.0949546993,-0.0170255546,0.1794534475,-0.0251392815,-0.1054470614,0.0608865656,0.2414659411,-0.1985666007,-0.1304470301,0.2712430358,0.0217383299,0.2006019503,-0.1616601795,0.1616558135,0.2688415051,0.2757568061,0.1569673568,-0.1697059572,-0.004055596,-0.0077820336,0.0585268401,-0.2947441041,-0.161985904,-0.5081425309,0.037770804,-0.0184634011,-0.043848604,0.1221693829,0.1954465657,-0.0525248721,-0.0582669154,0.2690705657,-0.0301431026,-0.1574964076,-0.0791351721,-0.2522331476,0.3092055619,-0.0335000604,-0.1218008995,0.0547138005,0.0923818648,-0.2261872888,-0.0194600876,-0.1298098266,-0.09631183,0.2501862049,-0.0036790015,0.0261796638,-0.0674595684,0.0343314596,0.2676851153,-0.2829970717,-0.0531992875,0.0664959401,-0.0833537802,-0.0800879076,0.1437057257,0.0693927258,-0.4156949818,0.0011951352,0.3203828633,0.0305032432,0.3538359106,-0.4491858482,0.508490324,0.1823734045,0.013491625,0.0888570547,0.0319666378,0.0448525846,-0.0693285093,0.4030996859,0.482395649,-0.2983867228,0.3422097564,0.1358788908,-0.1711564958,-0.0422863103,0.2095318139,-0.3388972282,-0.0261708274,0.3781140447,-0.1276304424,0.2739115655,0.2351719588,-0.2561182976,0.1661656648,0.1476952136,-0.0160888731,-0.0809772089,0.0476262495,-0.0714963749,-0.1313910633,-0.2348781526,-0.5552027822,0.1164977029,0.0013867415,-0.0232635438,-0.0545087643,0.0216566343,-0.1572434306,0.3525059223,0.34375301,0.1450271755,0.1492990702,-0.1920293719,-0.0355426781,-0.1705774367,-0.1088640243,0.0713137835,0.1552579254,-0.0907882154,0.1787135005,-0.2669935822,0.1986664385,-0.1382829547,-0.2146409154,-0.4423953295,-0.0005896058,-0.1357070059,0.0996641591,0.0083357049,0.1374203861,0.0659176633,0.0662306771,-0.0799295977,0.0320725366,-0.1635374129,0.0442070588,0.0026010263,-0.2587062716,0.3222123384,0.1722279638,-0.0051193414,0.1856897324,-0.2329492271,-0.0743535757,-0.253313154,0.1573008895,0.1893117875,0.395167917,-0.0172212478,0.1454005241,-0.0126460278,0.0085464222,0.0548397563,-0.1058779359,-0.0813929066,0.0681665987,-0.1739325523,-0.1048047468,-0.1257430017,0.0312265903,-0.301661998,-0.2763968408,0.0753877163,-0.1308687627,-0.0697796047,0.1207032949,0.0178133715,0.1717419177,0.2268356979,0.1557615697,-0.1230138913,-0.0647669658,0.3871436119,-0.0747450441,-0.1278488189,-0.3576082885,-0.0431129485,0.1861849874,-0.1928478032,-0.0354199037,-0.0323203281,-0.4579220414,-0.2585943043,-0.1971142441,0.067645371,0.4451829493,0.001533691,-0.009818919,0.1755239218,-0.1528435051,0.2090502232,-0.0589598231,-0.3336906731,0.1507889777,0.460193485,-0.229666844,0.7112486959,0.0886824206,-0.2320901304,0.0662529469,0.0745569542,-0.121672228,-0.0122246752,-0.176871568,0.0822219774,0.2394989878,0.0849888921,0.3006242216,-0.10343308,0.0722903907,0.1016179845,-0.2336832583,0.0608994104,-0.3511600494,0.3261652589,0.0607452393,0.3985971808,-0.0306407362,0.1116729528,-0.1026763245,-0.3349088728,-0.0500668436,0.2027682215,0.2166903615,-0.0853495002,-0.5692731738,0.0257503279,-0.8145981431,0.3692863882,0.0332481116,0.54682827,0.1562880874,0.0032422889,0.2857390046,-0.3994809091,0.7713319659,0.1870099604,-0.1161802113,0.0647540316,-0.6939154267,-0.0239436198,0.0463829301,-0.05709235,0.3465093672,0.2620872557,0.7865763307,-0.2753962874,-0.4964419603,-0.0875132307,0.0018747746,0.0433324091,-0.3471828699,-0.3430639207,0.0275686961,-0.3073930442,0.522231102,0.1112576649,0.1337032467,-0.1202624813,-0.0611820109,0.0788462013,-0.2576754391,0.3659957051,0.2644789815,-0.1498648673,-0.0997341052,0.391079247,0.4070549011,-0.3407425582,0.1719434857,-0.2235464156,-0.0998024866,-0.3073051572,0.2968035936,-0.0716782659,0.6282290816,0.4713291824,0.068876043,0.4053804874,0.3102551103,0.3968020976,-0.3236936629,0.1228616461,0.1179949418,0.5907544494,-0.5318713188,-0.0560969263,0.1252380162,0.1550688595,0.0055881883,0.0557686351,-0.0740546212,-0.1078336313,0.2111397833,0.4065735638,1.0874135494,-0.2946910858,0.1821778268,-0.4702728987,-0.0226199497,0.6241234541,-0.3409049213,0.2287949473,-0.3004300296,-0.2343696654,0.16611287,-0.2804966867,0.3048322201,0.0679806694,-0.3292179108,0.1750086546,-0.2850902081,0.043948818,-0.4569405913,0.2862700224,-0.054427579,-0.6450108886,-0.0855657533,-0.017169226,0.0532372147,0.3244040906,-0.1143907607,0.0358235873,0.0776088312,0.0334329531,0.1153236032,-0.1084811315,-0.1754150987,-0.0208408181,0.6226547956,0.0170716196,0.4309776723,-0.070902288,0.5701234341,0.3139301836,0.0473912023,-0.0434078611,-0.1491898,-0.1350712478,-0.1159293056,-0.0660991818,0.0885863155,-0.1951940656,0.1208703965,-0.3175597787,0.1195510253,-0.0472644269,-0.1077540964,0.0072704027,-0.0246081464,0.0567744225,-0.1909662485,0.2980084717,-0.284294188,-0.1888687164,-0.0388226546,0.1520333886,-0.4494116604,0.296469301,0.0646261945,-0.1676412821,-0.0522167236,0.0989340842,-0.1002616659,-0.3521370888,0.6038628817,0.1878899336,-0.3001303971,-0.1238514036,-0.2564106882,-0.0927500129,-0.4810035229,0.3884578645,0.304461211,-0.2149647176,-0.0681894571,-0.1190166026,-0.3499702513,-0.1710370034,-0.280313462,-0.3272756636,-0.4583595097,-0.3233031332,0.1127324179,0.1174817383,-0.0338897333,0.2789436877,-0.4293768406,0.0672998503,-0.0889486372,-0.0945839956,-0.4863044322,0.0874432549,-0.0663242936,-0.5179325938,0.0150980074,0.1795091033,-0.0553105026,0.4354676604,-0.2017278224,-0.0745576546,-0.2644640803,0.225027144,-0.2485759854,-0.371437788,0.402095139,0.2545945048,-0.2520881593,-0.2082113028,-0.1508308649,-0.1685199291,-0.2058845162,-0.5109720826,0.3025387526,0.1866175979,0.0084310584,0.1650598198,-0.0341629349,-0.0314061083,0.1390543431,0.3264312148,-0.0615503564,-0.0555565059,-0.0104537345,0.1412484199,0.6715649962,-0.3103087842,0.0295045879,0.1217535064,-0.1234535053,0.1414975822,-0.0809820667,0.4559877515,0.1064433008,-0.22545214,0.348131001,-0.0186994858,0.0116464095,-0.1176686212,0.0863973647,-0.0877341554,0.0035442498,0.4166715741,0.3665235341,0.3105360568,0.1519561112,0.1822441369,0.248035863,-0.3374418318,-0.1544873565,-0.1368290484,-0.0100884736,0.161121875,0.4471876919,0.1347372532,0.1822401583,-0.1351186037,0.0590223148,0.2496486604,0.2347100526,0.5030959845,-0.1264209151,-0.1973547786,-0.3373428881,0.5845699906,-0.0030648925,0.3970070481,0.0362674482,0.3866611421,-0.074223116,-0.0553878546,-0.0967658311,0.5879903436,-0.1176162958,-0.4829298556,0.1414034069,0.1150581688,-0.2666410804,0.0921209976,0.1258202046,0.0214633476,-0.0489653498,0.1047648489,0.0739526004,-0.168793425,-0.2197145075,-0.2017787695,0.2377809137,-0.0775487572,0.0991095752,0.0555666722,0.0671536773,-0.0164211653,-0.0771210268,0.5046649575,0.1273927838,-0.2176028341,0.120252192,-0.132604152,0.261149317,-0.0959257111,0.4205179811,0.0273375921,0.3024725318,-0.0926732123,-0.2097427994,-0.1238259003,-0.4595785737,-0.0456024595,-0.0957816914,0.0330639966,0.1907062531,0.2013796121,-0.1562980413,-0.3018085659,0.1854401231,-0.2207883894,-0.0498962551,0.3115551472,-0.3225317299,-0.1077683643,0.0225367714,-0.0081706913,-0.0293888915,0.6360401511,-0.0179871507,0.3895027339,-0.2058777362,-0.166184634,-0.377374649,-0.2126346529,0.3374999762,0.4094554484,0.1677934974,-0.1145356745,0.1391121596,0.0572184734,-0.061222475,-0.3062317669,0.0111917593,0.236387372,-0.3145105839,0.3704319,-0.4006152153,0.0352489017,0.2195592076,-0.1515675336,0.2366525084,0.2422895432,-0.1909923255,-0.3480926156,0.0752474964,-0.2067492008,0.1999884844,0.0626092777,0.2348814756,0.0900636911,-0.0405983552,-0.0252234619,0.1817329079,0.3266462684,-0.1005157009,0.0650938302,0.1300720274,-0.1851738989,-0.1384365857,-0.1151794866,-0.2090564072,-0.0090383282,-0.0461866595,-0.1627198011,-0.1346162111,0.0901249275,-0.1139101163,0.1425728947,0.104412131,0.4841727614,0.2690679729,-0.0817375556,-0.3441430926,-0.1082639843,0.5011172295,-0.2591926157,-0.3975765109,-0.5629040003,0.1348993033,0.0184611976,0.1122298911,-0.3204002082,-0.0949793085,-0.0139617147,-0.0993710235,-0.2857542634,0.5058297515,-0.0300001856,-0.1855188459,-0.2627213299,0.2818742096,0.1273734868,0.2065705359,-0.0973527506,-0.3075067699]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2159","title":"adding ccnet dataset","comments":"closing since I think this is cc100, just the name has been changed. thanks ","body":"## Adding a Dataset\r\n- **Name:** ccnet\r\n\r\n- **Description:** \r\nCommon Crawl\r\n\r\n- **Paper:** \r\nhttps:\/\/arxiv.org\/abs\/1911.00359\r\n\r\n- **Data:** \r\nhttps:\/\/github.com\/facebookresearch\/cc_net\r\n\r\n- **Motivation:**\r\nthis is one of the most comprehensive clean monolingual datasets across a variety of languages. Quite important for cross-lingual reseach\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nthanks","comment_length":14,"text":"adding ccnet dataset\n## Adding a Dataset\r\n- **Name:** ccnet\r\n\r\n- **Description:** \r\nCommon Crawl\r\n\r\n- **Paper:** \r\nhttps:\/\/arxiv.org\/abs\/1911.00359\r\n\r\n- **Data:** \r\nhttps:\/\/github.com\/facebookresearch\/cc_net\r\n\r\n- **Motivation:**\r\nthis is one of the most comprehensive clean monolingual datasets across a variety of languages. Quite important for cross-lingual reseach\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nthanks\nclosing since I think this is cc100, just the name has been changed. thanks ","embeddings":[-0.1904818863,-0.1757380068,-0.1641885787,-0.1790977269,0.028357856,-0.0581131354,0.0791954175,0.1407520473,0.0717295632,0.1240057722,-0.1063362658,-0.0534253754,-0.0881107002,0.4060057998,0.275740087,0.1043649912,0.1590127498,0.0418705903,-0.3196364939,-0.114847675,-0.1807442158,0.1922346652,0.138748318,-0.2338838875,-0.0712847188,-0.0380920134,-0.2617788017,0.0061948099,-0.2521851063,-0.2903493345,0.2684653103,0.4157442153,-0.0100912573,0.5332247615,-0.0000991827,-0.0869341567,0.1746727377,-0.1229330152,-0.1878489256,-0.0294467323,0.0337693729,-0.1837795079,0.0190661605,-0.1538279951,-0.314109683,0.2004223466,0.0845237225,-0.2129344195,0.2082117647,0.2350298613,0.337510705,0.0290042162,-0.1818660051,-0.2915087044,-0.1198543012,-0.0180006027,-0.3151050806,0.292665571,0.4156002402,0.0601789802,0.1099363193,0.3690298796,0.0668828636,-0.2279454768,0.0711580813,-0.0364486128,0.006268648,-0.2584452331,0.320287317,0.2316382974,0.3776721954,-0.2595231235,0.1185871735,0.0398695357,-0.1258427799,-0.4261802733,0.0742641985,0.2261038721,-0.0880824849,0.3159805834,-0.2444194555,-0.2289880067,-0.0295159947,0.1770612001,0.0944409817,0.2111172229,-0.1291156858,-0.0882126167,0.1784510016,-0.0530157611,-0.102905646,-0.0436569192,0.0730902702,0.0460937247,-0.2932653427,-0.0845345631,0.2650445402,0.1078460962,0.4247509241,0.0436740406,0.1987763494,0.0799396262,-0.1641611308,0.0135628181,-0.1532568634,-0.0486709587,0.0540612191,-0.059500467,0.2631003857,-0.0488465615,0.1333132982,0.0123112639,-0.0774599761,0.212498188,-0.0588154085,-0.06723243,-0.002756122,0.0697244704,-0.1796686649,0.1108991057,0.1863142252,0.0155419568,0.0885136798,0.5497040749,0.2330815941,-0.1261238754,0.0040592602,0.1615179032,-0.0100845788,-0.294452548,-0.2976145148,-0.0429085381,-0.1867261082,-0.0957629979,0.239390403,-0.0712967515,-0.007670728,-0.169215098,0.080508858,0.1282177866,0.0538440831,0.0607225671,0.1266824752,-0.077432923,0.0346952192,-0.0172091611,0.0931679904,-0.333684504,-0.2845564783,-0.0715449601,-0.1632344574,-0.0704342648,-0.4062544107,0.3201120496,0.097862415,-0.2241833657,0.0645935908,0.5243077278,-0.1689563841,0.050015077,0.2171370685,0.3235795498,-0.2786643207,-0.0856400356,0.0997954831,0.3374951482,-0.2000523806,-0.2366835624,-0.1698062718,-0.0778547376,0.2236400992,0.0932098776,-0.0797766224,-0.1940964907,-0.2328989059,0.0763257816,0.1962097287,-0.2441653907,-0.3140488565,0.122934483,-0.35147205,-0.1737954021,0.0261010211,0.0911286548,-0.0152316643,-0.0331133679,0.0301401876,0.1200179905,-0.0161448754,0.081054233,-0.1170622185,-0.2664486468,0.3370988369,0.0727473199,0.0311978832,-0.1358075589,0.2511894405,-0.2999977767,0.1952834576,-0.3060242236,0.2013707459,0.1280468553,0.3979385197,-0.1282156408,0.1054572314,-0.0776179656,-0.340383321,0.0558612496,-0.0148284538,0.3839975595,-0.0210046656,-0.3194343746,-0.4446518421,-0.090057455,-0.2989542484,-0.1015987694,0.2968000472,0.2105739117,0.0226006936,-0.0265593696,-0.0425922275,0.2824538052,-0.255161494,-0.0645493716,0.1636298895,0.4560249448,-0.2892094851,0.0836407915,0.118825309,0.2209792137,0.0857714489,-0.0837132707,-0.0053966078,0.2179364413,-0.4771177769,0.3257370889,0.6667903066,0.1830282807,0.2291882634,-0.609036684,0.1358554661,0.1629352421,0.1407422274,-0.099795945,-0.0397563055,0.0422879867,0.3284898996,-0.0143985571,0.0337560922,-0.0415114872,0.4215320349,-0.0446967743,0.0490810201,-0.1606881768,0.3694202602,0.0117413308,0.2099179178,-0.0154065546,-0.4150969982,0.2681397796,0.2451186031,-0.0215513259,0.3971554935,0.2922533751,-0.1288445741,-0.1581666619,0.0657154992,0.0039230231,-0.0971038938,0.2271773368,0.1075803488,0.1173025221,0.2812429368,-0.1158072725,0.1379514337,-0.0886861607,0.0330999158,0.064419575,0.3193721771,-0.0381701812,-0.5611578822,-0.2969507575,-0.0628563762,-0.1472924054,0.132076636,0.0682474896,-0.1209420413,-0.4004201591,-0.1091788337,0.1005355269,-0.3455258012,-0.2180407792,0.0485830605,-0.1284745783,-0.0634160638,0.2321055681,-0.4783982635,0.41673702,-0.030113494,0.0176250041,-0.0910437703,-0.1718742549,-0.0263904221,0.2868527472,-0.0529384762,-0.1599943191,0.5330930948,-0.153693229,0.3352717757,-0.1431749761,-0.6313955188,0.3790313303,-0.1427959502,-0.2095890641,0.0454898104,0.0447658487,-0.4265986681,-0.303537339,0.127556622,-0.1978254914,-0.1685204208,-0.0134113543,-0.1960620731,-0.045745939,-0.209809646,-0.2477647811,-0.5771508813,-0.3130738139,0.0926402882,0.1460674852,0.1165703908,0.4257984757,-0.2910262942,0.0751286298,-0.0100603495,0.2229466587,-0.4464685619,-0.1147911549,0.1674381047,-0.3826459646,-0.2028230131,-0.1091758311,-0.2802805901,0.300979346,-0.1788789779,-0.2073886693,0.1003152207,-0.1622810513,0.142248258,0.4034996629,-0.0279070735,0.1884663552,0.0674574226,-0.2546734512,-0.1040417179,-0.1267071962,-0.0426174253,0.1998573244,0.1469707042,-0.0078509441,-0.0638097823,0.0808434486,0.5089789629,0.1469099224,0.0339896828,0.5050978661,-0.0384911634,0.3856612742,-0.250805676,-0.3712059259,0.3186396658,0.0659652054,-0.0284033846,0.4144210517,0.1413975507,-0.4788587391,-0.1894003004,-0.1823982596,-0.5581129193,-0.1865194887,-0.0078284098,0.2107729763,0.1987776011,-0.0227375794,-0.2499545664,-0.1865322441,-0.3410499394,-0.02552348,0.2244223207,-0.4013786316,-0.0801257715,-0.5014060736,-0.0385154672,-0.9135110378,0.2954773307,0.0172365177,-0.1008691862,-0.269145906,-0.0010655164,-0.020571785,0.0409133211,0.196623221,-0.2550888956,-0.0603275597,0.0124294385,0.3115750253,0.0891897604,0.2294651419,-0.2271146178,0.2225152999,0.3560866714,-0.3603794277,-0.2986814678,0.0762603655,0.1898058802,0.0106250206,0.0495863743,-0.0063663647,-0.174650088,-0.2437425405,-0.3871002495,-0.001510779,0.0427937098,0.3271850049,-0.0941919088,-0.211187005,0.3022387922,-0.0760583803,0.2832804024,0.2507798374,0.1104143038,0.5031846762,0.0584371202,-0.1183520183,0.0870864764,-0.0459496193,0.5582690239,-0.0637085438,-0.1704296917,0.1558262706,-0.2056181878,0.3016305864,0.2617964745,0.2799315453,0.1404372752,0.3030343652,0.0751223788,-0.2662753761,0.1102475673,0.1184670627,0.1022777185,-0.5510281324,-0.1414399445,0.3413091898,0.0196363088,-0.1806606501,0.2422142178,0.2742539346,-0.2652024627,0.1172159463,-0.2099025697,1.17190063,0.0990774557,0.0010802234,-0.0247553196,0.1301018745,0.2207594961,-0.0897414014,-0.0291251112,0.0399878472,-0.1122837514,-0.0279677492,0.0692182258,-0.0820478871,0.0929079503,-0.3621407151,0.2358482182,0.0881597698,0.3319882751,-0.0131768091,0.2308558077,-0.072895512,-0.3335394561,0.0013775517,0.2970362008,-0.064824529,0.1885299832,-0.1550852209,-0.0042459033,-0.1416783184,-0.2890020311,-0.1907581687,0.009474135,-0.0575038493,-0.0087845782,0.0296460055,-0.2753745615,0.16260387,-0.1109115034,0.3813767731,0.1862536222,-0.3429589868,0.2513705492,0.0495789833,0.0050905701,0.0765016302,-0.0280689728,-0.027334312,-0.0672583207,-0.4044447839,0.141280964,0.2331899852,-0.1523361653,0.0832131878,0.0207650866,0.1910252869,0.0179144572,0.2037938237,0.403159976,-0.2051460147,-0.2750046551,0.2598474026,0.0878013149,-0.000164995,-0.234241873,0.349164933,0.0072379676,-0.2230293602,0.1038010269,0.3082381785,-0.2911821306,0.2088277787,0.244531557,-0.2984273136,-0.3344343901,-0.0782754347,0.1275189072,-0.2383675724,-0.068377465,-0.0205308292,-0.2086518109,-0.0665358081,0.4601989388,0.0598190054,0.280071348,0.0550818592,-0.0384353586,-0.2158329338,0.1243047491,-0.1753222197,0.2003572434,0.1379681081,0.140507251,0.2336245179,0.0111401593,-0.5365862846,-0.0511613861,-0.255276382,0.0449956954,0.3155133724,-0.2031163126,0.170934096,-0.1051301807,0.2096117139,0.0388075896,-0.2635464966,-0.3797216415,-0.0261144023,0.0646639615,-0.031384673,-0.1633013487,0.0211593211,0.0749337152,0.009838568,-0.1227443144,0.1611940712,-0.0302726664,0.1280805916,0.0150079494,-0.0074636033,-0.0970731303,0.0995933935,-0.1384761482,0.215832293,0.4224183261,0.0117152557,0.0792889968,0.0046042921,-0.0958892852,0.1545548588,0.3591847122,0.109012872,0.0756205767,0.2136611193,0.0476588309,0.1024431959,0.4377659857,0.2378504872,-0.1986328959,-0.4112121463,-0.0154808126,0.1586768925,0.3702749014,-0.4132627547,-0.0651948079,0.1986300945,0.0451204926,0.309063077,0.1823173314,0.3281461,-0.1863476783,0.0756209642,0.0717424974,0.3089416921,-0.19487001,0.3248741031,0.0110949893,0.2167992741,0.0952440947,0.3205267489,0.0745372549,0.124244608,0.360401541,-0.2262755185,0.1392046213,0.1888837516,0.1014009193,0.3760102987,-0.0897561461,0.3217447698,0.1613897979,-0.268019855,0.2363466173,0.162989825,0.3989804089,-0.0443202145,-0.0266532376,-0.3113904893,0.2637389004,-0.0594211891,0.0320174843,-0.2285154015,-0.1215686724,-0.3479126692,-0.1675973684,-0.004417845,-0.1925380826,0.1175366938,0.006320572,0.0418320782,-0.2598995566,0.193592906,0.1140202805,0.2856330574,-0.1933213472,0.294588834,0.2863384783,-0.1234579235,0.0580278859,0.3346599936,0.2227836847,-0.1857726127,0.0516795367,0.2871559262,-0.0043081823,-0.2444838881,-0.1151023656,-0.0903110355,0.2516677678,-0.0139584132,0.1039417163,0.2646565139,-0.2694880366,-0.1585574746,0.3876613975,0.0269265138,-0.0272009596,-0.1671232432,-0.0742601231,0.1359845698,-0.4363732934,0.0679837018,-0.4583877623,0.1190106347,0.3404445052,-0.2136583477,0.2102740556,-0.3699443638,0.1662225127,0.0954231173,0.2683947086,0.1971530914,0.1798005104,0.0345000178,-0.5041196942,-0.2173768431,0.1188932955,-0.1589849293,-0.0523559526,-0.2523369789,0.1198232397,0.1891051233,0.1241295412,0.088015303,0.1674622893,0.0350448191,-0.3167806566,-0.2012440264,0.1916324049,-0.0211865492,0.1992894709,0.0235322043,-0.1296901852,0.0871886462,-0.0827272087,0.1243674159,-0.1625115126,-0.1855503768,-0.1927827895,-0.0868204758,0.4196765721,0.0371391438,0.3781719506,-0.4199460447,-0.0686039552,0.0473279394,-0.403244257,0.0408975966,0.1860478222,0.1049694046,0.0114797205,-0.0625210479,-0.1177026108,0.27546978,0.1513539404,0.1870758384,0.0360152498,-0.1734415293,0.0067689139,-0.1050330698,0.0225360189,-0.1227762997,0.1863208413,-0.0161461588,-0.0742592812,0.1433715969,-0.6241073608,0.2464302629,0.0815759078,-0.2154729813,-0.3889060318,0.161518693,0.098898977,0.2299463749,-0.3280778825,0.2074619234,0.1762166172,0.0800129473,-0.0603164509,0.2326906919,-0.2047993839,-0.1872143894,-0.0854050592,0.3533765972,0.2653466463,-0.2864451408,-0.0452724136,-0.3209331036]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2158","title":"viewer \"fake_news_english\" error","comments":"Thanks for reporting !\r\nThe viewer doesn't have all the dependencies of the datasets. We may add openpyxl to be able to show this dataset properly","body":"When I visit the [Huggingface - viewer](https:\/\/huggingface.co\/datasets\/viewer\/) web site, under the dataset \"fake_news_english\" I've got this error:\r\n\r\n> ImportError: To be able to use this dataset, you need to install the following dependencies['openpyxl'] using 'pip install # noqa: requires this pandas optional dependency for reading xlsx files' for instance'\r\n\r\nas well as the error Traceback.\r\n\r\n","comment_length":26,"text":"viewer \"fake_news_english\" error\nWhen I visit the [Huggingface - viewer](https:\/\/huggingface.co\/datasets\/viewer\/) web site, under the dataset \"fake_news_english\" I've got this error:\r\n\r\n> ImportError: To be able to use this dataset, you need to install the following dependencies['openpyxl'] using 'pip install # noqa: requires this pandas optional dependency for reading xlsx files' for instance'\r\n\r\nas well as the error Traceback.\r\n\r\n\nThanks for reporting !\r\nThe viewer doesn't have all the dependencies of the datasets. We may add openpyxl to be able to show this dataset properly","embeddings":[-0.1493438631,-0.1812810451,0.0336430967,0.341632545,0.2343087941,0.2863430381,-0.0157952346,0.2244071364,0.1654276997,0.0734670162,-0.2225701958,-0.1259506196,-0.03064215,0.3286091685,-0.0807032138,-0.209744215,0.2249317169,0.1918756962,-0.0383199081,-0.2154147625,-0.1752157807,0.2497071475,-0.2327596545,0.033299882,-0.1657814384,-0.005650437,0.0485558584,-0.1095098481,-0.2481057048,-0.2244533449,0.2573483884,-0.0757726133,0.0078952471,0.4936810136,-0.0001168755,0.0112852957,0.2621284723,0.0009628056,-0.0758979172,-0.1726921201,0.1580370665,-0.5534417033,0.2592311502,0.005135417,-0.1263031811,-0.6198988557,0.1344301105,-0.0443131849,0.4444801807,0.4118565917,0.2116182446,0.4163538218,0.4214429855,0.0597534105,-0.1707170755,0.0774004906,-0.0333689302,0.2485256046,-0.112039946,0.2506457567,0.0464389957,0.3587521315,-0.0722557977,-0.1672137231,-0.1182158664,-0.2060662955,-0.0668232441,-0.3877575099,0.065018788,0.2956514955,0.2338842452,-0.1602156162,-0.1942190379,-0.1710206121,-0.0075114551,0.1529851109,0.234055236,0.3493386805,0.0310773924,0.2444141656,0.089173153,-0.3955403268,-0.0207314249,0.1518016458,-0.2865380049,0.271138519,-0.2161859572,0.1463052034,0.2850150168,-0.1231927201,0.1981233209,-0.0080020959,-0.0066175866,0.168779552,-0.1542187631,0.0125708319,0.0683273897,0.4823487401,-0.1209615543,-0.4073605537,-0.4050123692,0.0087040886,-0.2241734564,0.2380542606,0.0841645971,0.0210722256,-0.0141955065,0.2033031434,0.1795813739,0.3211095035,0.1527701914,0.0679236874,0.0083028544,-0.2380101979,-0.4805679023,-0.3320239782,0.3121840954,-0.2570629716,-0.3029878438,0.2578764558,-0.0671019778,-0.0249174535,0.0111848582,0.182961151,-0.1553817987,0.4950966239,0.2343169153,0.1126600429,-0.2211113721,-0.4079149067,-0.1736008525,0.2763749361,0.0510919429,-0.1620607972,-0.032807745,-0.6860463023,0.2591982484,0.1957404912,0.1480438709,0.047656849,-0.1888779849,-0.0919522792,-0.2200558782,0.3047617674,0.1786770523,0.131814003,0.553268671,-0.3143092096,-0.1588373482,0.0255600773,-0.0247138031,-0.1193820387,-0.3907499313,0.1378650069,-0.3698265254,-0.0951169133,0.1383287907,0.1952489018,-0.2921114266,-0.2234259546,-0.1163565814,0.0132624386,-0.0202204604,0.0533876978,0.0504159555,0.5927078128,-0.4545469284,-0.3469185829,0.0944002643,-0.2169104517,0.0935402066,0.1010093316,-0.0120336916,-0.0668989122,-0.1673156321,-0.3382666707,0.2603876591,-0.4653259516,-0.2659332752,0.3473155797,0.0669156909,0.2950734198,0.1359106153,-0.1164171621,-0.2745509148,0.2345121801,-0.4540861249,0.0552111603,0.0520400256,-0.0751473382,-0.2216024101,-0.2808434665,0.2575274706,0.3178930283,0.1873325706,0.0692965537,-0.0543510877,-0.1554214954,0.2374704778,0.1147569418,0.0038577621,0.1038566008,0.1040420681,0.3543663621,0.054034166,0.002796411,0.0732586458,-0.0716021359,-0.0064702518,0.290582031,-0.1062062085,-0.2331238538,-0.3494341373,0.0760537684,-0.1288511604,-0.3132869899,0.0632410869,0.1411925107,-0.0140754459,0.5048753619,-0.1575295925,0.27174595,-0.1571386158,0.1240423918,-0.5752680898,0.1061153635,-0.2694675624,0.1400140375,0.1747436374,0.0970163867,-0.0346790664,-0.0112214023,0.0091679152,0.1916556209,-0.151748389,0.2544415593,0.1007500738,0.0872532651,0.250277102,-0.4632273316,0.1977060437,0.1617235839,0.1071821675,0.1465789974,-0.0417563878,0.1484865248,0.053547699,0.0759226158,-0.0791275725,0.4016434252,0.3749584258,0.2591935396,-0.0877757892,-0.372130245,0.4974481463,-0.1468615532,0.3775921464,-0.1414720416,-0.2833206952,-0.0737213716,0.1207566187,0.0512172654,0.1060101315,0.2926082015,-0.2328044027,0.1253170669,0.2561503351,-0.0175857991,0.1818636805,0.0954405665,-0.1969885677,0.2725213766,0.0246928614,-0.239864707,0.1188213229,0.0407809541,-0.1291139424,-0.0179597847,-0.1501010209,-0.0485614575,-0.5117405057,0.124787122,0.0088861641,0.2497189641,-0.3337169588,0.0810149387,-0.1943718195,-0.3811690807,-0.209766373,-0.4099612236,-0.2709551156,-0.422414571,-0.0119396104,0.105178073,-0.0280012209,0.3039633632,-0.018855961,0.1006590724,-0.0498096198,0.0785682201,-0.2276708633,-0.165293023,-0.0761513337,0.1163732186,0.1672727913,0.0746080056,0.1345999092,-0.5980010629,0.1774813384,-0.4021346271,-0.3951074481,0.1637867838,-0.3234314024,0.4101296961,0.2941055596,0.1234956756,-0.0674271658,-0.0877377763,0.2502646148,-0.1842281669,0.1840976924,0.039619837,-0.1432317793,-0.1686982214,0.0873312205,-0.1414393485,-0.2506596744,-0.3166332245,0.2217575312,0.1017183363,0.1098301411,-0.0413943306,0.1959975064,0.2456980497,-0.09875191,0.2361411005,-0.2863528132,-0.5247047544,0.3783111572,-0.1568012238,-0.3085812926,0.2651654482,0.0860078037,0.1951528192,-0.4093189836,-0.733242929,-0.2308117151,-0.1923142374,-0.0415656306,0.033482153,0.1754434854,0.1946000755,0.0358054265,0.0847483799,-0.0624023117,-0.050844159,-0.2867293656,-0.3271091878,0.0163104553,-0.0000910936,0.5346378088,-0.0564330071,0.2596888244,0.4661255777,0.1804060042,0.3968822956,0.1061628312,0.788567245,0.0033063213,-0.493596822,-0.0231896807,0.1583252102,0.2208786458,0.3041403592,-0.0754182637,0.2468591332,-0.5103856921,-0.3159966469,-0.327647388,-0.1706674397,-0.2363375127,-0.1083346531,0.2060176879,0.0129451239,0.0467308499,0.0787633508,-0.1366557777,0.1077705771,0.3479911685,0.0770940781,0.036545597,-0.0738063082,-0.0912135765,-0.4185932875,0.3242528439,-0.2333341092,0.2012438774,-0.0008866744,0.12144541,0.168438971,0.0369146913,0.7025109529,0.0302966032,-0.3258080482,-0.0525401905,-0.1223508641,-0.6611144543,0.1079375222,-0.2527157366,-0.1881041825,0.0247112401,0.0034355619,-0.2656268775,-0.1243870854,0.4412812293,0.0804680064,-0.2095832527,-0.2385823578,-0.2916270196,-0.2214309573,-0.3525426686,-0.0845472217,-0.0090309419,0.080427058,-0.0301801786,0.4483981133,-0.0971261039,-0.0236033611,0.3017007709,0.0508194491,0.0081440499,0.0206430051,0.0002469991,0.4754725695,0.1500506997,-0.0532303341,0.7757796645,-0.063116245,-0.7709522843,-0.075382784,0.0536334105,0.0531327985,0.3070106208,-0.1265822649,0.0274935756,0.2437786013,0.1017827094,-0.2109919786,0.1763730645,0.3813941777,0.2676440775,-0.1815091074,-0.2140790969,0.3380755484,0.1158891171,0.1463100314,0.1849011034,0.5283310413,-0.0841910243,0.2648402452,-0.0969953164,0.9988307953,0.1961120814,0.2482183278,0.4241094589,-0.2455766797,0.4833488762,-0.0700263754,-0.029295586,-0.144412607,-0.3055582345,-0.2449561059,0.0516548976,0.4823301435,-0.2462415695,-0.3096330166,0.3389145434,-0.1209041923,-0.0700884014,0.1229674071,0.1574329734,-0.3809257448,-0.0709056631,-0.3266378343,0.1354926378,0.0377953984,0.3158827424,-0.1692900956,0.1198324934,0.0187711827,-0.1505932361,-0.2882561684,0.0519955754,0.0278499871,0.0180970337,0.0684051588,-0.0966806412,0.2472620159,0.5656706691,0.3669228554,0.1865282208,-0.4858180583,0.1671657264,-0.0459156483,-0.4441520572,0.1770738065,0.2046080828,0.3137399852,-0.0114793628,-0.2831753492,0.3268379867,-0.1206450388,0.2323897928,-0.197733745,-0.0111095719,-0.0575118437,-0.3667595983,-0.3333875835,-0.1771249175,0.0570142865,-0.183652252,0.1110577136,0.0103437249,-0.1201190352,-0.0232845098,-0.0562225766,-0.1417724937,-0.2431451827,0.195931226,0.0323257521,0.0144018242,0.2819169164,0.2788679898,-0.3665272892,-0.1169046611,0.1877423078,0.0454841219,-0.3456623852,-0.1701560616,0.2914132476,-0.1606871933,-0.2383991778,0.0027348187,0.1746844202,-0.0411227793,0.0126880733,-0.3946435153,-0.0076743392,0.1925798804,0.1038074493,-0.1245338321,0.1717464775,0.2776166797,0.1337158978,0.1878321022,-0.2733457983,0.3751013875,-0.1645789295,-0.0308964159,0.1304587871,0.2312323749,0.322848767,-0.1181521341,0.0373162255,0.1187984645,-0.1167223305,-0.1287341863,-0.1502200365,0.1250339597,0.2713003457,-0.4052733779,0.2548579574,0.0848725885,-0.2070201188,-0.0313606411,-0.1952662617,0.3474892378,-0.0754381493,0.1337978989,0.0290038213,0.0485038497,-0.0916017368,-0.0754815117,-0.0463015065,0.1908141524,0.0429237783,-0.0622406602,0.0316442363,0.0093684206,0.0889623463,0.2149967551,0.2068993896,0.059317831,0.3137523532,-0.1371273845,0.125619173,0.1611731499,0.4621786177,0.2891745567,-0.4220869839,0.1494948715,0.4947500825,0.0427788682,-0.5449191332,0.0137339644,0.2964161038,-0.0359675027,0.0966339931,0.0143322786,0.1898609847,0.0388370603,0.1929156035,0.1015236601,0.1329613328,-0.0326306038,0.3157618344,0.2818041742,-0.1478830427,0.1104032546,0.2457271665,0.1664688587,-0.0167652871,0.2051963657,-0.3525443375,0.0279038977,-0.2922762632,0.0825242698,-0.035552226,-0.1481751651,0.2657366991,0.0649835318,-0.1496372521,-0.0404282175,-0.1202601343,0.4565219283,-0.2054389715,0.0748907551,-0.2444639206,0.3438742161,0.0049212608,0.025891427,0.2523988485,-0.1254977733,-0.2077852786,0.0125957131,0.1929333955,-0.5256760716,-0.0626932755,0.0837934539,-0.1748882383,-0.4093302786,-0.0262105372,0.0982757956,0.1624487638,-0.070634596,0.165264383,0.3033758402,0.0084784832,0.1473626047,0.6045824885,0.4527468085,0.070211634,0.1158503592,0.1578057259,-0.1719761044,-0.046099212,0.078289099,0.2830427289,-0.0351974629,0.2945498228,0.307382524,0.1182827502,-0.2352012843,-0.1192615926,-0.0033581497,0.2711737752,-0.5004979968,0.1815778017,-0.4764452875,0.1020690501,-0.3031993806,-0.1503982842,-0.5871675014,0.2847024202,0.2014154792,-0.1557389647,-0.076107882,-0.3292551935,0.0349972397,0.1634239703,0.4378169775,0.4837254882,0.3563603461,-0.2145857364,-0.3759633601,-0.6806294322,0.2766789496,-0.0359322317,-0.1828470081,-0.0877023116,-0.242352277,0.0795431212,0.290181309,-0.1649817973,0.2002098709,-0.1299871355,-0.1870605648,-0.1312101483,-0.0886711925,-0.0190622043,0.1277426332,-0.0501923598,0.0209721047,0.2831502259,-0.081290938,0.0103785302,0.0650069565,0.1038181335,-0.3491299152,-0.2601621151,0.1418422014,0.3700292408,0.332368046,-0.1182963923,-0.0788224936,0.0622204803,-0.2541967928,-0.4052447081,0.4564958513,0.0300649386,0.1583380997,-0.301048547,-0.5846755505,-0.0435882956,0.3900039196,0.1314124614,0.1165824383,-0.4264739156,0.133436814,-0.0666235238,0.0021296439,0.1974082887,0.2926412821,-0.0800682232,0.203448534,-0.1025537625,-0.1717098206,0.475861609,-0.2589901984,0.0350629427,0.0025696813,-0.0480378158,0.2377600372,-0.4223098755,-0.601000011,0.2271063775,0.1251104623,0.1089573056,-0.0012159693,0.1362534612,-0.2136979103,-0.0375876613,-0.0279388782,0.3252096474,0.3717556894,-0.1575442255,0.3425120711,-0.0167051069]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2153","title":"load_dataset ignoring features","comments":"Hi ! Thanks for reporting. I opened a PR to fix this issue: #2201","body":"First of all, I'm sorry if it is a repeated issue or the changes are already in master, I searched and I didn't find anything. \r\n\r\nI'm using datasets 1.5.0\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/37592763\/113114369-8f376580-920b-11eb-900d-94365b59f04b.png)\r\n\r\nAs you can see, when I load the dataset, the ClassLabels are ignored, I have to cast the dataset in order to make it work.\r\n\r\nCode to reproduce:\r\n\r\n```python\r\nimport datasets\r\ndata_location = \"\/data\/prueba_multiclase\"\r\nfeatures = datasets.Features(\r\n {\"texto\": datasets.Value(\"string\"), \"label\": datasets.features.ClassLabel(names=[\"false\", \"true\"])}\r\n )\r\ndataset = datasets.load_dataset(\r\n \"csv\", data_files=data_location, delimiter=\"\\t\", features=features\r\n )\r\n```\r\n\r\nDataset I used:\r\n\r\n\r\n[prueba_multiclase.zip](https:\/\/github.com\/huggingface\/datasets\/files\/6235022\/prueba_multiclase.zip) (it has to be unzipped)\r\n\r\n\r\nThank you! \u2764\ufe0f \r\n","comment_length":14,"text":"load_dataset ignoring features\nFirst of all, I'm sorry if it is a repeated issue or the changes are already in master, I searched and I didn't find anything. \r\n\r\nI'm using datasets 1.5.0\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/37592763\/113114369-8f376580-920b-11eb-900d-94365b59f04b.png)\r\n\r\nAs you can see, when I load the dataset, the ClassLabels are ignored, I have to cast the dataset in order to make it work.\r\n\r\nCode to reproduce:\r\n\r\n```python\r\nimport datasets\r\ndata_location = \"\/data\/prueba_multiclase\"\r\nfeatures = datasets.Features(\r\n {\"texto\": datasets.Value(\"string\"), \"label\": datasets.features.ClassLabel(names=[\"false\", \"true\"])}\r\n )\r\ndataset = datasets.load_dataset(\r\n \"csv\", data_files=data_location, delimiter=\"\\t\", features=features\r\n )\r\n```\r\n\r\nDataset I used:\r\n\r\n\r\n[prueba_multiclase.zip](https:\/\/github.com\/huggingface\/datasets\/files\/6235022\/prueba_multiclase.zip) (it has to be unzipped)\r\n\r\n\r\nThank you! \u2764\ufe0f \r\n\nHi ! Thanks for reporting. I opened a PR to fix this issue: #2201","embeddings":[-0.1033394337,0.0052699288,0.001646272,0.3112616539,0.4842154384,0.2444421947,0.7037388682,-0.0102143651,0.1451043636,0.0337648131,0.112438418,0.3442732096,-0.1163506955,0.5018551946,-0.2144526392,-0.0656915605,0.0363927148,0.1277149469,0.0825558901,-0.1254534423,-0.3830315173,0.0947665945,-0.3583936393,-0.0296227578,-0.1907411069,0.2579049468,0.0995686427,0.0467885956,-0.0531538166,-0.4804023206,0.4141026139,0.1879521459,0.3708599508,0.267072171,-0.0001140255,-0.0320013277,0.1021172553,-0.130379796,-0.1889138818,-0.0945015028,-0.451018244,-0.2830277085,0.3010983765,-0.2223178148,-0.3194632828,0.117787838,-0.2708767354,-0.082234256,-0.1400141269,0.0542442203,0.2065468878,-0.0430936478,-0.0755620822,0.1961785406,0.0377836302,-0.0704700947,-0.3091857731,0.1963919997,0.2152681053,0.1631591767,-0.011278769,0.3061054647,-0.1882140189,-0.0302638523,0.4362982213,-0.1885432601,0.0673443228,-0.3027724326,0.0694636405,0.0976627991,0.7716204524,-0.1111688018,-0.438591212,-0.3013161123,0.074464649,-0.3795777857,0.1598971039,0.0636833981,0.3168964088,0.1488632113,-0.4841517508,-0.059389405,-0.0502799749,0.1002540514,0.0220795553,0.0277843028,-0.1355092078,0.1503535509,-0.1822707057,-0.0327311121,0.3617728055,-0.1877419651,-0.094187811,0.2044890523,-0.3130448461,0.19857575,0.092450805,0.1943721771,0.0334327333,0.2261740863,-0.0359313823,0.2582353652,-0.2649019063,0.1250318438,0.2212480903,-0.0578203797,0.46923244,0.1309579015,0.1515894681,0.1436612457,-0.3306213319,0.0340544432,-0.1437530667,-0.0522746257,0.3445692956,0.0538102984,0.1130916104,-0.3915534616,-0.0794459283,-0.0159461349,-0.1512431949,-0.0467433147,0.1328262687,0.2510562837,0.07644137,0.5234663486,0.2114429027,0.3024414182,-0.2871133983,-0.3321474195,-0.236430034,-0.2337980121,-0.0954631642,0.0485702194,0.3418631852,0.1388815194,0.245664075,-0.0603480712,-0.0736370608,-0.1640505344,-0.1830602288,0.0824486986,-0.0989325941,0.2236211747,-0.1693915278,0.1835951954,0.2156713307,-0.119066678,0.0475694425,0.211139068,-0.2447338849,-0.143623203,0.2280897349,0.1815046668,-0.3090977669,-0.1606350094,-0.4676074684,0.1636721343,0.0006917387,-0.1450906992,-0.0143249221,-0.4730603397,-0.0420022719,-0.1917028874,0.2643386126,0.4124704897,-0.3415287137,-0.1534119844,0.2800125182,-0.11994569,-0.0469520912,-0.1880365759,-0.1551235318,0.074260205,-0.1330570281,-0.2750838101,0.171737954,-0.1895433664,-0.3709957898,-0.1439062953,0.1278757155,0.0049147662,0.1431509107,0.2906523347,-0.0792217627,-0.0086136134,-0.1590105295,0.4087690115,0.1038845256,-0.0144917313,-0.2082740664,-0.0743877515,0.008678725,0.3856330216,0.0683340356,0.2459988743,0.1651167572,0.0074674706,0.133761391,-0.2022164911,0.1588031799,0.4151982069,-0.0385832936,0.1461935937,0.0371519141,-0.0913846567,-0.377373457,0.305387795,0.1640835404,-0.012699415,0.0598549917,-0.1064713523,-0.0100415116,-0.0625974461,-0.5984095335,-0.1028617099,-0.0021044002,0.1964169443,-0.2410085946,-0.0562836751,-0.0231443588,0.2144462615,-0.1415256411,0.2712519169,-0.2729840279,0.0999350548,0.1229031309,-0.1381412894,-0.1363510936,0.1891177744,0.4001704454,-0.1779057086,-0.1692353487,0.2228799462,0.2220045179,-0.2140925229,-0.0209324528,0.1004514769,0.1295421124,-0.3592019677,-0.0059432229,0.0517742038,0.2365554571,-0.1492530107,-0.3222787976,0.3925388455,-0.0788220242,0.1630931795,0.0532968938,-0.091198802,0.3404313624,-0.0934116468,-0.0792053118,-0.0355481096,0.0308770929,0.0222671814,0.2292415202,0.2900805473,-0.4528570175,0.1855775416,0.3603377044,0.0422672816,-0.1070472822,-0.0505884178,-0.2113155276,0.1097641215,0.2660451531,0.0171694662,0.2552281916,0.2585332096,-0.0552887172,-0.0098973624,-0.015245676,-0.0680928975,0.248331666,-0.0086790007,0.0108367251,-0.0158795584,0.239648819,0.0917256773,-0.3205383718,-0.0839395225,0.1613158286,-0.0996719599,-0.5439400077,0.0457939357,-0.2049046159,0.0803115591,-0.1209715158,-0.1362223625,-0.254388839,-0.305513829,-0.1566802412,0.4129647613,0.0485072508,0.0978234187,-0.4254234433,0.0714078099,0.0456126295,-0.6064157486,0.1760403067,0.1869795322,-0.4785401821,-0.0500864424,-0.0879824683,0.1087620258,0.1340588629,-0.0551655218,0.0784532875,-0.1519365013,-0.3282815814,0.0050127809,-0.2428476363,0.0504866205,-0.0988449752,-0.0512440391,0.0154880229,0.1621456742,0.4146180451,0.0877557248,0.0984076709,0.211123541,0.2462694496,-0.159410581,-0.3171158433,-0.4890909195,-0.1021151692,-0.3961871266,0.0288999658,0.0967718288,-0.3240273893,0.1721124798,0.0706713125,0.1004997715,0.2724615335,0.0799398795,-0.0700961128,-0.3131956756,0.3434098661,-0.0818506554,-0.2187512517,0.0646105707,-0.0199946016,-0.041221492,0.0055860472,-0.4025328755,-0.0173179265,-0.0952415615,0.3524102569,0.153582871,0.1142781302,0.3060990274,0.218013227,0.1051643118,-0.1882968545,-0.358125031,0.0599684678,-0.062183857,0.2379497588,-0.1494469941,0.0707032681,-0.2402392775,0.3310950398,-0.039748773,-0.0251437556,0.4948443472,-0.0624538399,0.2496257871,-0.1538940221,-0.2894426882,0.0671375766,-0.2886564434,-0.0073192879,0.2290700376,0.1774288118,-0.0556711331,-0.222347185,0.2790785432,-0.1675675362,-0.1187404767,-0.1401732862,-0.0459239669,0.0983776748,0.1066651717,0.2704550624,-0.1664309055,-0.197500512,-0.2019418776,0.235877648,-0.024390744,0.0262053143,-0.5049213767,-0.0597092994,0.0111521771,0.0812152848,-0.0589259975,0.1500604898,-0.0274778847,0.0030045691,0.1244363412,0.099973686,0.5367012024,-0.4157113135,0.2973086238,0.3973895311,0.0891042948,-0.0906744748,-0.2839477062,-0.1540869474,-0.200594902,-0.0065766769,0.163257286,-0.0668765903,-0.2404557765,0.1515976638,0.365986526,-0.21876809,-0.0511626303,-0.3161494732,-0.1380751878,-0.2562221289,-0.2125790119,0.0882643238,0.3685908318,-0.4332980216,-0.0125889629,-0.1596798003,-0.0536129735,0.120034866,0.2301114053,0.1758638173,0.1065046713,0.3351995647,-0.0570962392,0.1083757281,0.1964699626,0.8620957732,-0.0573833473,-0.9934465289,0.090698801,-0.3942218721,0.492284894,-0.0274748392,0.016765127,-0.2449574471,-0.1671850681,-0.0516933911,-0.2643908262,0.1462927461,0.2633690536,0.1000770628,-0.3819731176,-0.5160293579,0.4756683111,-0.2461893708,-0.0969645083,-0.0592087843,-0.0465219691,-0.4712139368,0.3574056029,-0.0067267171,0.5893378258,-0.0726002157,0.1728347689,0.1717263907,0.1215586364,0.499610424,0.116767697,0.1801219881,-0.2863987982,-0.2946312129,-0.0774504021,-0.0595248379,0.1275990456,0.5485593081,-0.0385149941,0.4438938498,-0.3171224594,-0.174442485,0.0331084169,0.1367392391,0.3974382281,-0.2745510936,-0.1454487145,0.1295360327,-0.114361912,0.0858687833,0.0166004486,0.0375486799,-0.0225881301,-0.0615290515,-0.0718796849,0.3462180793,-0.0188337453,0.0447236225,-0.0136865862,-0.1057608649,0.327524811,0.4943903387,0.2517470121,0.0506232865,-0.1453224421,0.1279631555,0.039678853,0.1788265854,-0.0392132625,0.094346866,0.4873352051,0.1597002745,-0.254563123,-0.0815099701,-0.1709547043,-0.1813268363,-0.0681543499,-0.0438257791,0.1353817284,-0.6524505019,-0.1674778163,-0.1625872105,0.3390182853,-0.3045727909,0.1125403717,0.0072535723,0.0064274524,0.106444791,0.0509610362,-0.3338098824,-0.1926460564,0.2594896257,-0.1386795342,-0.0191366766,0.5800497532,0.3252718151,0.0009354718,-0.1244798526,0.2300684601,0.4347567558,-0.4420942366,-0.1452020258,0.3532298207,-0.161407575,0.2840534151,0.46587044,-0.1701562107,-0.1912266016,-0.0054657138,-0.4269529879,-0.1847593784,0.0037809459,-0.182149902,0.1344572753,0.1535975337,0.2156386524,0.2857967615,-0.1988381594,-0.2078796327,-0.0805643648,-0.3368809819,0.0163721163,0.3262781501,0.2977215052,-0.1147119701,-0.0008336132,0.1481358409,-0.2394649088,-0.1024748236,-0.1282724887,-0.1731122583,0.1250972599,0.172049582,0.3393345773,-0.3055202365,-0.3491403461,-0.1068887189,-0.1783793718,0.1315125376,0.0440371074,0.1537177116,0.4538025856,-0.395747602,0.4355540574,-0.2115631253,0.0158355292,-0.2220691144,-0.0455821864,0.274946183,-0.0504714884,0.2393866479,0.1006870344,-0.3842930198,-0.096007973,-0.1773314029,0.1205312908,0.2916249335,-0.4087588489,0.46043998,0.3358896375,0.1042883173,0.1249525994,-0.0850475207,-0.0759077296,0.112982817,0.252614677,-0.4703255296,-0.3451437652,0.1556363553,0.1011682898,-0.0290430281,0.3054766357,-0.0505584031,0.1645443439,-0.0644585267,0.0535034873,0.5986958146,0.0578235239,0.2430479079,0.411824286,-0.0335598998,0.1771110445,0.3585046232,0.3556156754,-0.061169982,0.4472643733,0.025033582,0.149004057,0.1022187173,0.348338902,-0.1197239161,-0.4835218489,0.1913657635,-0.1287564188,-0.1737604737,0.1845176518,-0.2890424132,0.0850677341,-0.6706323624,-0.3802956641,0.010417331,-0.0672309771,-0.1164284199,-0.0429489948,-0.0779078826,-0.0563536957,0.0755830035,-0.0313477628,-0.1241706982,-0.0757489875,0.4103399515,-0.0710554346,-0.0373883285,-0.5398231745,-0.2022552788,0.0339049809,0.1158614978,-0.0406643972,0.2759805322,-0.1642215699,-0.1817212701,0.335588783,0.0391979739,0.1371332109,0.2013189346,0.1168810204,-0.1913292557,-0.2851189077,-0.0199396517,-0.1463630497,0.1693449318,0.1258102059,0.3290918469,0.0874165073,0.1058785021,-0.1126262546,0.1974074543,0.108852528,0.128621757,-0.4212910831,0.5933308601,0.1922209114,-0.3104308546,-0.0902053416,-0.1003228873,-0.1615604758,-0.0558249839,0.5143449306,0.1244833544,0.1219476163,0.0606370568,0.0574866123,0.2019485682,0.2364750654,0.3068465889,-0.0604300015,-0.1865032166,0.393781215,-0.7514324784,0.1447692662,0.3227195144,-0.3138853312,0.3249519467,-0.0227010418,0.3816988766,-0.0808976516,-0.0919441655,0.0213493649,0.4711137116,0.0127778947,-0.2808685899,-0.0835215002,0.0327222943,0.3435527384,-0.161164254,-0.3536480069,0.1811486483,-0.045255743,-0.0311834123,0.0338678993,0.0113208443,0.2255841643,-0.1359215826,0.1797354221,-0.1394975483,0.5507989526,0.0457676239,0.0606581382,-0.1024755388,-0.2361214161,-0.1025688127,0.2940258086,-0.1544910371,0.4881404936,-0.2493637949,0.2757075429,-0.2402333617,0.1390834451,0.3025153875,0.0936334208,-0.3302803934,0.2986646593,-0.2565042377,0.1079015285,0.3344231844,0.1657182127,-0.0981187746,0.0139027322,0.0120112048,-0.3429665565,0.3115893304,-0.4151758254,-0.3215866089,-0.0651791394,0.2448123693,-0.0958858505,-0.0107368976,-0.717559278,0.0315096751,0.3827619255,0.0138072371,-0.5683805943,0.2920523286,-0.1211605519,-0.0303033516,-0.054858692,0.0897371471,-0.0009762355,0.0274908394,0.1090102121,0.0407056101]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2153","title":"load_dataset ignoring features","comments":"Nice question which helped me a lot! I have wasted a lot of time to the `DatasetDict` creation from a csv file. Hope the document of this module add some simple examples.","body":"First of all, I'm sorry if it is a repeated issue or the changes are already in master, I searched and I didn't find anything. \r\n\r\nI'm using datasets 1.5.0\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/37592763\/113114369-8f376580-920b-11eb-900d-94365b59f04b.png)\r\n\r\nAs you can see, when I load the dataset, the ClassLabels are ignored, I have to cast the dataset in order to make it work.\r\n\r\nCode to reproduce:\r\n\r\n```python\r\nimport datasets\r\ndata_location = \"\/data\/prueba_multiclase\"\r\nfeatures = datasets.Features(\r\n {\"texto\": datasets.Value(\"string\"), \"label\": datasets.features.ClassLabel(names=[\"false\", \"true\"])}\r\n )\r\ndataset = datasets.load_dataset(\r\n \"csv\", data_files=data_location, delimiter=\"\\t\", features=features\r\n )\r\n```\r\n\r\nDataset I used:\r\n\r\n\r\n[prueba_multiclase.zip](https:\/\/github.com\/huggingface\/datasets\/files\/6235022\/prueba_multiclase.zip) (it has to be unzipped)\r\n\r\n\r\nThank you! \u2764\ufe0f \r\n","comment_length":32,"text":"load_dataset ignoring features\nFirst of all, I'm sorry if it is a repeated issue or the changes are already in master, I searched and I didn't find anything. \r\n\r\nI'm using datasets 1.5.0\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/37592763\/113114369-8f376580-920b-11eb-900d-94365b59f04b.png)\r\n\r\nAs you can see, when I load the dataset, the ClassLabels are ignored, I have to cast the dataset in order to make it work.\r\n\r\nCode to reproduce:\r\n\r\n```python\r\nimport datasets\r\ndata_location = \"\/data\/prueba_multiclase\"\r\nfeatures = datasets.Features(\r\n {\"texto\": datasets.Value(\"string\"), \"label\": datasets.features.ClassLabel(names=[\"false\", \"true\"])}\r\n )\r\ndataset = datasets.load_dataset(\r\n \"csv\", data_files=data_location, delimiter=\"\\t\", features=features\r\n )\r\n```\r\n\r\nDataset I used:\r\n\r\n\r\n[prueba_multiclase.zip](https:\/\/github.com\/huggingface\/datasets\/files\/6235022\/prueba_multiclase.zip) (it has to be unzipped)\r\n\r\n\r\nThank you! \u2764\ufe0f \r\n\nNice question which helped me a lot! I have wasted a lot of time to the `DatasetDict` creation from a csv file. Hope the document of this module add some simple examples.","embeddings":[-0.0856140181,-0.0304919872,0.0129400603,0.2842156291,0.4256821573,0.2531645894,0.6406395435,-0.0539474487,0.2234636098,0.0533237867,0.1543203592,0.3363256156,-0.1144230142,0.4599338472,-0.1519141644,-0.0312218927,0.063283965,0.1389202476,0.0868092924,-0.1416218579,-0.3790285289,0.0661422163,-0.3810799122,-0.0748045146,-0.199522078,0.2117881924,0.0234554317,0.07192076,-0.0297584608,-0.5245155096,0.413390696,0.2789027691,0.3400850296,0.251696527,-0.0001132447,-0.0907421485,0.0439932868,-0.1335797906,-0.1930736899,-0.1258175671,-0.4377946556,-0.3049478531,0.293554306,-0.3018359244,-0.3062906265,0.1707061231,-0.2648938298,-0.2134493887,-0.0529288314,0.1210677773,0.1923326999,0.0092524923,-0.0871038511,0.1758142114,0.0143903447,-0.0617420748,-0.325831145,0.2043394744,0.1856566221,0.2550998926,-0.0153446216,0.2552683055,-0.214363575,0.0363501497,0.5101963282,-0.1277292073,-0.0145440605,-0.3282902539,0.0238243286,0.1738422066,0.7974794507,-0.2552067935,-0.4845220149,-0.2741535008,0.1363139153,-0.4022521377,0.1647414267,0.0906636193,0.2678411901,0.1801911443,-0.4022525847,-0.0693216324,-0.1144453436,0.0422427021,0.031348858,-0.0237456132,-0.1550961435,0.1390875131,-0.1776063144,-0.0167391114,0.3612679541,-0.2586158216,-0.0753731653,0.2356112748,-0.2841072381,0.1693637818,0.091674529,0.1357545108,0.0742881224,0.1346576959,0.0195800122,0.2739371359,-0.3180241883,0.1293332875,0.1421793848,-0.0020369068,0.5196121335,0.1570585668,0.198042199,0.0485058203,-0.2741855383,0.0649611875,-0.156292364,-0.0674709305,0.2910913527,0.0448109359,0.1363649368,-0.3723415136,-0.1389972121,0.0318643376,-0.1400588751,-0.0162491165,0.1177361235,0.2474844903,0.0247120373,0.4246523976,0.2354547828,0.2400788367,-0.2873142064,-0.3280520141,-0.2301762104,-0.1649012864,-0.0516043603,0.0536998697,0.4024191797,0.172232613,0.1889072657,-0.0935541242,-0.1197241247,-0.1503179371,-0.1773230135,0.0644035339,-0.0126002934,0.2159597725,-0.1039229706,0.1599552333,0.1865864992,-0.2099940628,0.0202711225,0.1871026605,-0.2052912414,-0.0912482366,0.1930449158,0.1692913622,-0.3488231599,-0.0992196351,-0.4686614871,0.1708803028,-0.0383038186,-0.0585591123,-0.0010499486,-0.4762669206,-0.074391745,-0.2339225709,0.3303196728,0.4801045656,-0.4956296384,-0.1252076775,0.303334415,-0.1631097049,-0.0423816964,-0.1286754161,-0.0879168361,0.1719543934,-0.095555529,-0.3004548252,0.2273970246,-0.1882335097,-0.3485089242,-0.1945337504,0.0981055051,0.0058674407,0.1647294462,0.3072187304,-0.0137548149,0.0016859525,-0.098152332,0.4130121768,0.0815085173,-0.016189836,-0.1648076773,-0.0039023247,0.0153039144,0.327521801,-0.0128150014,0.2236014903,0.1535398513,0.0070145875,0.1234481335,-0.2657747567,0.2417103946,0.3734304309,0.0262524337,0.2091947347,0.0803009719,-0.0996135399,-0.4244221449,0.3336114585,0.1969466656,-0.0301368739,0.0143269869,-0.1335237324,0.0164890569,-0.0420469604,-0.599829495,-0.0444379859,0.0043741786,0.108144179,-0.1282560676,-0.0561452545,-0.0211245511,0.2936041355,-0.2329749018,0.2568300068,-0.3175535798,0.0984747335,0.1104398668,-0.0933684781,-0.1049847901,0.2762140632,0.3593247235,-0.2134838402,-0.139570713,0.252053529,0.2794742584,-0.2395739555,-0.003348226,0.0503881983,0.1536212564,-0.38654387,0.0303776786,0.0651876405,0.2595449984,-0.1391369104,-0.4149834514,0.4254066348,-0.1629067808,0.1436124742,0.0709132105,-0.0759841576,0.3986670077,-0.0515849888,-0.1249036938,-0.0875488296,0.0120539544,-0.0065012444,0.2523196042,0.2596165538,-0.4297820628,0.1778017133,0.3382619917,0.023081027,-0.0686432719,-0.0502457023,-0.177865088,0.0848932713,0.2156614512,0.0609677508,0.2490087152,0.2434667051,-0.0681903735,0.0026713256,-0.0539208427,-0.0883627459,0.2646454871,-0.0272719078,-0.0585923605,-0.0459214188,0.1281882077,0.0731413141,-0.2991208434,-0.1569283456,0.0864324719,-0.0542883426,-0.5587561131,-0.0162466988,-0.1895411015,0.0446084775,-0.1800445765,-0.1181470081,-0.2130714059,-0.2718034983,-0.1732316613,0.382527411,0.0305418987,0.0627364963,-0.4193843305,0.107765615,0.0894372389,-0.5412775874,0.2205042988,0.1355538666,-0.4152530134,-0.0298727639,-0.1084800959,0.2368005067,0.2161876708,-0.1324639916,0.1083870828,-0.2184620649,-0.2653663456,0.0387863144,-0.246820569,0.0459491313,0.0089157,-0.0679905936,-0.055145558,0.1420685053,0.3785454929,0.0763236508,0.0598126762,0.2437888384,0.207072407,-0.2385058254,-0.2769645751,-0.5045168996,-0.1614888161,-0.3960846364,0.041375909,0.0946876779,-0.251645565,0.138907209,0.057367716,0.1830513179,0.2809784114,0.1540104747,-0.0310248919,-0.235582009,0.3494182825,-0.1501153409,-0.2869130671,0.1117159054,-0.0587879494,-0.0539898053,0.0010019642,-0.3663876951,0.088975668,-0.0896046609,0.3901561499,0.2198274881,0.1605475247,0.3285789788,0.2460654676,0.1037363186,-0.2389996946,-0.3663343489,0.0832588673,-0.0620273054,0.191442579,-0.0777534172,0.110681124,-0.2450100929,0.3476117253,-0.0138874138,0.0028704084,0.4750258625,-0.0642099977,0.2776764333,-0.155767858,-0.2877364457,0.0507540666,-0.2303261608,0.0119082946,0.2468530089,0.1611822248,0.0501923338,-0.2115146518,0.2637025416,-0.2098139524,-0.1581947953,-0.1451286077,-0.121677123,0.103978999,0.0837814063,0.217070505,-0.1675320268,-0.1631676704,-0.2661850154,0.2213691175,-0.0045869183,0.0102267116,-0.5514551997,-0.1828896105,-0.055338461,0.0938134864,-0.0580313243,0.1872236282,0.0261194184,0.0560863242,0.0933768973,0.1057456061,0.5734018683,-0.3399971128,0.2928527296,0.3923082948,0.1141810119,-0.2291172296,-0.2148536146,-0.1900697351,-0.1029031947,-0.0881072506,0.0920740366,-0.1049287543,-0.2605796158,0.1314916611,0.4926408529,-0.2059617937,-0.0615345351,-0.2721278071,-0.1646257937,-0.2355960906,-0.2033087164,0.0396521129,0.3704892695,-0.4324612021,-0.0091455411,-0.1053399667,-0.0523020476,0.1303985417,0.2326041162,0.222434327,0.1291310787,0.4058957696,-0.022981897,0.0852615386,0.2154715359,0.7898676395,-0.0779162273,-0.9518975616,0.0684627593,-0.4182258844,0.5247336626,-0.0762836263,-0.0203667041,-0.2259805948,-0.0906923115,-0.0732131153,-0.2362547368,0.1195159256,0.1982825696,0.1051263511,-0.4362965822,-0.5934206843,0.4223225713,-0.1847919524,-0.0875430703,0.027515389,0.0065174778,-0.5263801217,0.3839455545,0.0043080975,0.546996057,-0.0652339235,0.2061707675,0.1642443836,0.1326800287,0.5669634938,0.1644974947,0.124931857,-0.3096993268,-0.2729044855,-0.0801166371,-0.0462646931,0.2010452151,0.5373557806,-0.124750711,0.3956956565,-0.2290387899,-0.2270821929,-0.0035677017,0.2181626856,0.3714715838,-0.3578473926,-0.1528427601,0.1190256402,-0.0880047455,0.0192840621,0.0827520415,0.1017649248,-0.0791933089,-0.0693394244,-0.0784781873,0.2588514984,0.0013757482,0.0804194212,-0.0061965948,-0.0992328748,0.330291301,0.4486837387,0.3506098688,0.1147854999,-0.1648313701,0.1178843752,0.0604453012,0.1746528149,-0.0124873519,0.0916752368,0.4582448006,0.1442276984,-0.2612518072,-0.08616025,-0.1898118109,-0.1628652513,-0.025283413,-0.0830285177,0.2034991086,-0.589533031,-0.2081590593,-0.1479993314,0.2661505044,-0.3238860965,0.1126799285,0.0957265198,-0.0205563288,0.1644132733,-0.0115257241,-0.316532135,-0.1929322779,0.2515877485,-0.0450098738,0.0347441211,0.5573702455,0.3530451357,-0.0319895484,-0.1456607729,0.2625561953,0.3961201906,-0.3901971281,-0.1040851846,0.4095885158,-0.1070691794,0.2239109427,0.5827170014,-0.1966189742,-0.1878956705,-0.0146093313,-0.4081913531,-0.1960131526,0.0054683639,-0.0846169814,0.168094188,0.2236246914,0.1663226038,0.2781651914,-0.2066334039,-0.2128274292,-0.1660310924,-0.2861822248,0.0190421138,0.3513837755,0.3476586938,-0.1403683126,-0.0089216763,0.1520306915,-0.2583584785,-0.0868976265,-0.1422889382,-0.1935822964,0.1273211688,0.1429844052,0.2857295275,-0.2640102506,-0.3534655869,-0.1489557177,-0.2576411366,0.1526486576,0.1386089027,0.1096442267,0.4580551684,-0.3482677341,0.4155032337,-0.2080925703,0.0065494319,-0.193661958,-0.0799274221,0.2459103167,0.006763862,0.144402355,0.0632147118,-0.3804649115,-0.0384013578,-0.2331903577,0.0461409613,0.2846542895,-0.3949282765,0.4182245135,0.3276236355,0.1750586033,0.1971658319,-0.0623294115,-0.0664044172,0.1412867755,0.2240267247,-0.4584309161,-0.3629932404,0.1607583463,-0.0078138914,0.0382106639,0.3136925697,-0.1171948537,0.1066633239,-0.0651332065,0.0340413861,0.6418434381,-0.0415939242,0.2116548866,0.3595901132,-0.0425313301,0.1777715832,0.3617372215,0.3674906492,-0.0481479801,0.496902734,0.0696472079,0.114356339,0.1496316493,0.3424573839,-0.1239908338,-0.4713843763,0.2929890752,-0.1653482765,-0.2640785873,0.147114262,-0.2910413146,0.0563963242,-0.7303279638,-0.3448594511,-0.0291979276,-0.0853532702,-0.0668609366,-0.075441286,-0.1401504278,-0.0652368292,0.2003003806,-0.1063700914,-0.079210557,-0.1153403446,0.3415636122,-0.1385676414,-0.0184504576,-0.5109069347,-0.0891852751,-0.0430593565,0.1336765736,-0.0293962155,0.233999908,-0.174999401,-0.0956815556,0.2588124275,0.0092802383,0.1054129824,0.2317337543,0.193195045,-0.1506307423,-0.3097062111,-0.0795972347,-0.139452219,0.2378837615,0.0547459908,0.2346520126,0.0977925435,0.0997683555,-0.0947734639,0.1536293924,0.1820670515,0.1523027122,-0.3820311725,0.544527173,0.2070104182,-0.341691196,-0.05597784,-0.0295071602,-0.1854070127,-0.0874872133,0.482419908,0.0996362045,0.1188280731,-0.0049259127,0.0616282858,0.2064937651,0.2476278991,0.2770631611,0.00555292,-0.1593330801,0.392847389,-0.7810738683,0.0616663694,0.2918814123,-0.2907621264,0.3185737729,0.0418267734,0.3377401233,-0.0268265847,-0.0879086033,0.0872561783,0.3598929048,-0.0520297773,-0.1859561801,-0.1271452308,0.0463942774,0.3435320854,-0.2045230716,-0.4161462486,0.2103686333,-0.0327190422,-0.0380158275,0.0165938288,0.0412060134,0.2931004465,-0.0927976072,0.188798368,-0.1301510483,0.556444943,0.0033667351,0.0893991068,-0.1195809841,-0.1518790871,-0.085153155,0.3002238572,-0.1061232835,0.5048050284,-0.2135784477,0.1592861861,-0.1607824862,0.1998846382,0.3464806676,0.0126962652,-0.3018126488,0.3784741163,-0.2702767551,0.1205833107,0.2680622339,0.1173260137,-0.1000175327,0.0406335853,0.045865342,-0.3424759507,0.2959663868,-0.4661597311,-0.3100436032,-0.1048113704,0.2468159348,-0.1140740365,-0.0544514507,-0.6054846644,0.0529157482,0.4427366853,-0.0387563445,-0.5960952044,0.3080435395,-0.1771589667,0.0096865864,-0.0662133917,0.2158688009,-0.0519974902,-0.0348670371,0.0966071412,-0.0649608895]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2153","title":"load_dataset ignoring features","comments":"Hi :) We're indeed working on tutorials that we will add to the docs !","body":"First of all, I'm sorry if it is a repeated issue or the changes are already in master, I searched and I didn't find anything. \r\n\r\nI'm using datasets 1.5.0\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/37592763\/113114369-8f376580-920b-11eb-900d-94365b59f04b.png)\r\n\r\nAs you can see, when I load the dataset, the ClassLabels are ignored, I have to cast the dataset in order to make it work.\r\n\r\nCode to reproduce:\r\n\r\n```python\r\nimport datasets\r\ndata_location = \"\/data\/prueba_multiclase\"\r\nfeatures = datasets.Features(\r\n {\"texto\": datasets.Value(\"string\"), \"label\": datasets.features.ClassLabel(names=[\"false\", \"true\"])}\r\n )\r\ndataset = datasets.load_dataset(\r\n \"csv\", data_files=data_location, delimiter=\"\\t\", features=features\r\n )\r\n```\r\n\r\nDataset I used:\r\n\r\n\r\n[prueba_multiclase.zip](https:\/\/github.com\/huggingface\/datasets\/files\/6235022\/prueba_multiclase.zip) (it has to be unzipped)\r\n\r\n\r\nThank you! \u2764\ufe0f \r\n","comment_length":15,"text":"load_dataset ignoring features\nFirst of all, I'm sorry if it is a repeated issue or the changes are already in master, I searched and I didn't find anything. \r\n\r\nI'm using datasets 1.5.0\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/37592763\/113114369-8f376580-920b-11eb-900d-94365b59f04b.png)\r\n\r\nAs you can see, when I load the dataset, the ClassLabels are ignored, I have to cast the dataset in order to make it work.\r\n\r\nCode to reproduce:\r\n\r\n```python\r\nimport datasets\r\ndata_location = \"\/data\/prueba_multiclase\"\r\nfeatures = datasets.Features(\r\n {\"texto\": datasets.Value(\"string\"), \"label\": datasets.features.ClassLabel(names=[\"false\", \"true\"])}\r\n )\r\ndataset = datasets.load_dataset(\r\n \"csv\", data_files=data_location, delimiter=\"\\t\", features=features\r\n )\r\n```\r\n\r\nDataset I used:\r\n\r\n\r\n[prueba_multiclase.zip](https:\/\/github.com\/huggingface\/datasets\/files\/6235022\/prueba_multiclase.zip) (it has to be unzipped)\r\n\r\n\r\nThank you! \u2764\ufe0f \r\n\nHi :) We're indeed working on tutorials that we will add to the docs !","embeddings":[-0.1101284996,-0.0165535659,-0.0179757886,0.2797108293,0.4765644968,0.265261054,0.7201102972,-0.0131259663,0.1362201869,0.070500277,0.0996232256,0.3421684504,-0.1340013444,0.496086657,-0.2072590292,-0.0867805704,0.0528041758,0.111113362,0.1122093722,-0.1322964877,-0.3251573145,0.0733093992,-0.3695145845,-0.0308829471,-0.1820509732,0.2237114161,0.0670666322,0.0586670786,-0.0488981046,-0.4741632044,0.4196299016,0.2079200894,0.3793597221,0.283688575,-0.0001124861,-0.0452884547,0.0771025494,-0.1252857745,-0.1733756214,-0.0956469774,-0.4460331202,-0.2817453742,0.2998546958,-0.2455118597,-0.3235130608,0.1322029084,-0.2634668052,-0.1049767137,-0.1022059172,0.0584639385,0.2196192592,-0.0430974551,-0.0819111094,0.1600189656,0.0449733287,-0.0786820352,-0.3150000572,0.2215148062,0.2376375198,0.1762765348,-0.0147699686,0.316854924,-0.1934015453,-0.0201525204,0.4284486473,-0.1611540169,0.0429694131,-0.3015797138,0.0482175946,0.13410528,0.7763341069,-0.12808837,-0.4381689131,-0.2878551185,0.0898619071,-0.3596651852,0.1029618979,0.0698901638,0.3103159964,0.163254112,-0.4930078387,-0.0496280454,-0.0567151718,0.1051056236,0.0556815751,0.064019382,-0.162890926,0.1445664018,-0.1779283583,-0.040462736,0.3343972862,-0.161106348,-0.0732579604,0.2229824513,-0.2745836675,0.1725874394,0.127717495,0.1979460567,0.0414624289,0.2100031227,-0.0282412171,0.2714694738,-0.2659796774,0.1349220425,0.196829319,-0.047863476,0.4790740609,0.1443423033,0.163932994,0.1646094471,-0.3273822069,0.0391400531,-0.1568366289,-0.0543333031,0.3065228164,0.0658551902,0.1219482645,-0.3954098225,-0.1182375699,-0.0236919373,-0.1659450233,-0.0576904267,0.1484488249,0.2453754842,0.0592296198,0.4688180387,0.2193441242,0.3083884418,-0.3062899411,-0.3301115334,-0.256524235,-0.1614272594,-0.100788489,0.037382111,0.3433778882,0.1542748213,0.2271299511,-0.0970910713,-0.0434364229,-0.1518768668,-0.1849520355,0.10105104,-0.1061021239,0.2132460177,-0.1630431712,0.1646627337,0.2479479313,-0.1889735758,0.0178690888,0.1786756814,-0.2419788837,-0.0984171256,0.2069288492,0.2055110037,-0.3074479401,-0.176834628,-0.464502275,0.1882935166,-0.0126087656,-0.1404078603,-0.0057892259,-0.4625445604,-0.0058728866,-0.2258603275,0.3081898093,0.3966841698,-0.3912998438,-0.1484602094,0.2467354238,-0.1100534946,-0.0495180376,-0.1903810203,-0.1465432495,0.1196286604,-0.1101499945,-0.2255262285,0.2026680112,-0.2136660963,-0.3061308563,-0.1531294733,0.1054261699,-0.0068966886,0.1309522688,0.278608799,-0.0636416674,-0.0429037288,-0.1270948946,0.4129882157,0.0713364705,-0.0263930075,-0.1944422573,-0.0832179114,-0.0315918848,0.3770879209,0.0244307071,0.2341848016,0.198701635,0.0289873891,0.1415964663,-0.2319101542,0.1617275923,0.4092968106,0.0047783451,0.1386485398,-0.0023470588,-0.0675798357,-0.3856048882,0.2912772894,0.2298685908,0.0040735519,0.0659552217,-0.143162623,-0.0088445209,-0.0906506404,-0.5928422809,-0.1307033747,0.0280415379,0.1792022884,-0.1877312064,-0.0550832786,-0.0225458685,0.2042459548,-0.1266520768,0.2621226311,-0.2874043584,0.1012905836,0.0960544497,-0.12577039,-0.1380851865,0.1942443103,0.3953321278,-0.1908032,-0.1289384365,0.241682142,0.1941593289,-0.1917328686,0.0210049357,0.1199663207,0.1206060573,-0.401109606,-0.004694894,0.0320029631,0.217098996,-0.1720080376,-0.3591335714,0.4095731974,-0.0882547274,0.1560614407,0.0773781836,-0.0941991359,0.3326980472,-0.0870577097,-0.1143335849,-0.0343905166,0.027576603,0.0109708449,0.2111823112,0.2330053896,-0.4458596408,0.2219580412,0.3417536616,0.0295754038,-0.0924588665,-0.0818196386,-0.188447088,0.111414358,0.2850052118,0.0147809265,0.246763736,0.2719568014,-0.0098607,-0.0316037647,-0.0386250354,-0.0925200209,0.2277563214,-0.0210565813,0.0104613258,-0.0778022707,0.1982870847,0.0832136273,-0.2898977101,-0.1477727443,0.1412748098,-0.0966997743,-0.5482537746,0.0512749813,-0.1949816048,0.0606523789,-0.1100825369,-0.1512790918,-0.2189670801,-0.323653847,-0.1520814449,0.4053588808,0.0155196507,0.1014578193,-0.4491036534,0.0661490783,0.0394778401,-0.596829474,0.1799115539,0.1921439618,-0.4924163222,-0.0202539228,-0.0755530968,0.1157299504,0.1531994492,-0.0729061589,0.0737800375,-0.1439742893,-0.3506554365,-0.0049858755,-0.2732565105,0.0676745847,-0.0627445504,-0.0270522069,-0.0320792757,0.1256537139,0.4145658016,0.0744104683,0.1143800244,0.1715951115,0.2483626008,-0.1622714549,-0.322429806,-0.4859111905,-0.1085481942,-0.4171247184,0.0299379732,0.0909535661,-0.3232811093,0.1717455834,0.1094829738,0.1379386634,0.2584349811,0.1070892066,-0.0584415086,-0.3195766807,0.3415904939,-0.1010416523,-0.2253235132,0.057789281,-0.0464476719,-0.0082781324,-0.022593813,-0.3960751295,-0.0321584903,-0.0758668706,0.3645590544,0.1930589974,0.1742613018,0.2934271991,0.2150864452,0.0775777549,-0.1758816391,-0.3686971962,0.0440893322,-0.0789209157,0.2178661823,-0.1536886394,0.062479917,-0.2159755528,0.3609847128,-0.0320744067,-0.0233732816,0.4889881015,-0.0680600405,0.2554812431,-0.1366110295,-0.2944389284,0.0880556107,-0.2743042111,-0.0394232608,0.2261357903,0.186196059,-0.0768842474,-0.2381356508,0.3083367646,-0.2151528895,-0.1370550543,-0.1372364163,-0.050394766,0.0414167754,0.0629622564,0.2441604733,-0.1469725519,-0.1642682105,-0.2322485745,0.2345104665,-0.0388022177,0.0462894477,-0.4844487906,-0.0783294141,-0.0047559054,0.0857186839,-0.0863597319,0.1045684218,-0.0515920706,-0.015573822,0.1343773901,0.107438907,0.5176264048,-0.4222331941,0.3002414107,0.4201526046,0.1266178787,-0.0896905214,-0.2732635736,-0.1770574749,-0.1889415532,0.0037427528,0.1668730974,-0.0802972242,-0.2427852303,0.1772786975,0.3553276658,-0.1823289841,0.0173248481,-0.2712345123,-0.2022671998,-0.2443502247,-0.19905819,0.1006729677,0.384804666,-0.432677865,0.0324723609,-0.132893309,-0.0396374092,0.1355068982,0.2314430326,0.1599614918,0.1454361528,0.3200745583,-0.0492962487,0.0892536119,0.1734659076,0.8486731648,-0.0452833287,-0.955955863,0.1004693136,-0.3483059108,0.5229463577,-0.0074358629,0.0445056222,-0.255695492,-0.1366504878,0.0034728686,-0.2546765506,0.1832210869,0.2387021929,0.0960648432,-0.3928614855,-0.5318900943,0.4889569879,-0.2094772011,-0.0885490701,-0.0548745394,0.0001114943,-0.4781556129,0.3307846785,0.0219831597,0.6259090304,-0.0560656302,0.2001547962,0.1918316334,0.1377542913,0.5266860127,0.0900274515,0.1580796391,-0.2767345309,-0.2649388611,-0.0725269988,-0.0483416431,0.1437093765,0.5265618563,-0.0970932618,0.4256350696,-0.3436410129,-0.1735013127,0.0397876762,0.1317888647,0.3891461492,-0.2915013731,-0.1448628306,0.1493273526,-0.0900186375,0.0754149854,0.0105134482,0.0636407137,-0.0582104921,-0.070318833,-0.0620040745,0.3409591317,0.0077541871,0.0586470179,-0.0607030168,-0.1310951263,0.3045372069,0.4982421994,0.2633920312,0.0565307923,-0.1455122381,0.1348523647,0.0412244052,0.1600468457,-0.042516049,0.0745493919,0.5274111629,0.170807451,-0.2459001243,-0.091140531,-0.1758617908,-0.1773047894,-0.0895947292,-0.0464729667,0.1206242889,-0.6423515081,-0.1631720364,-0.1329786181,0.3188717365,-0.2880198061,0.1285623014,0.0495399795,-0.0342959017,0.1092830151,0.0806765854,-0.3120598793,-0.191780895,0.2682262957,-0.1183934361,-0.0208775382,0.5706679225,0.3371063173,-0.0034398802,-0.1492721736,0.2188218385,0.475494802,-0.4076351523,-0.1573832631,0.3551806808,-0.1524493396,0.2621833086,0.4752847254,-0.1617374122,-0.1997615248,-0.0026189,-0.4006646276,-0.1837180108,0.0474876463,-0.1821367294,0.1501588374,0.1306287199,0.1864153892,0.3115646541,-0.2026365697,-0.2327709794,-0.0909498557,-0.3312087655,0.0060253316,0.3397599161,0.3087159395,-0.0952358097,-0.0058336142,0.1711035073,-0.2377941161,-0.1072670147,-0.1472233385,-0.1847183257,0.119682543,0.1633429229,0.3027683794,-0.2537974715,-0.3248551786,-0.1135122702,-0.2056871802,0.1338302493,0.0481276549,0.1570344418,0.4284861982,-0.3920551836,0.468038857,-0.2034371346,-0.0232587233,-0.1896020323,-0.0462596044,0.2881257832,-0.041485507,0.211826846,0.0815499276,-0.3303265274,-0.0679305196,-0.1636856943,0.1088427007,0.2616940737,-0.4040419459,0.4691996872,0.3071564436,0.1347811967,0.1262346357,-0.0940690413,-0.0415816829,0.1339409202,0.2635242343,-0.4445192516,-0.3198080361,0.1327798963,0.0631613657,-0.0030171627,0.2916866243,-0.1208295375,0.2009675503,-0.0774723589,0.0398358293,0.5913004875,0.0550097302,0.2770438492,0.3896345198,-0.1066487357,0.1651771814,0.3575794697,0.3497486711,-0.0698032379,0.4130588472,0.0043097818,0.1538921297,0.0762047246,0.3340315819,-0.1358464211,-0.4634916484,0.2300030142,-0.130370602,-0.1600193083,0.173613444,-0.3114627898,0.0788170174,-0.701090157,-0.3586507738,0.021385571,-0.0870584548,-0.1049008071,-0.0584225953,-0.0804537013,-0.0518061519,0.0733327568,-0.0759535134,-0.1054689884,-0.1205937415,0.4168224335,-0.1173778921,-0.013934548,-0.5179982185,-0.2002282143,0.012760452,0.1201717556,-0.0248931684,0.2940679193,-0.1779401749,-0.1676612496,0.2845403552,0.04597947,0.1549624056,0.1840842962,0.1487505287,-0.1786350012,-0.2951256335,-0.0440427028,-0.1504951417,0.160681203,0.1100400165,0.2926881015,0.0782975405,0.1348006874,-0.124407731,0.2310179472,0.0973657444,0.1215466484,-0.3980062902,0.5818288922,0.1567356884,-0.3121525347,-0.060622137,-0.1090661734,-0.182714656,-0.0839760005,0.4932121038,0.1194985136,0.1302676499,0.0485334508,0.0681612417,0.23305282,0.2064966857,0.3185217977,-0.0355030373,-0.1895506978,0.4021375179,-0.7697131634,0.1016959697,0.3061653674,-0.3341930509,0.2908139229,0.0207024105,0.4033479095,-0.041187644,-0.0780000612,-0.0190526024,0.4778026044,-0.0035764419,-0.2348988503,-0.1129125655,0.0011039692,0.3637060523,-0.1583269984,-0.3473722935,0.1625092328,-0.0593046211,-0.0084257713,0.0258052293,0.0365687683,0.2521599829,-0.2061217725,0.1941556931,-0.1550696492,0.5497813821,0.0285698436,0.0557505712,-0.1351564974,-0.2014583945,-0.1011020169,0.2844315767,-0.1642748863,0.5074603558,-0.264724046,0.2508611977,-0.2236375362,0.1299260557,0.3330513835,0.0988463238,-0.2951564491,0.2952831089,-0.2256390601,0.0782092437,0.3138689399,0.176275149,-0.1051341593,0.0302008539,-0.0238144845,-0.3872246444,0.3403587937,-0.4392800629,-0.3298446536,-0.066802457,0.2529737651,-0.0947848558,-0.020212831,-0.7089985013,0.065542616,0.3825773597,-0.0165914446,-0.6089789271,0.3142911792,-0.108610414,-0.046394784,-0.0619270466,0.129645139,-0.0136643322,0.0424155146,0.1107515842,0.0044255494]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2149","title":"Telugu subset missing for xtreme tatoeba dataset","comments":"Good catch ! Thanks for reporting\r\n\r\nI just opened #2180 to fix this","body":"from nlp import load_dataset\r\ntrain_dataset = load_dataset('xtreme', 'tatoeba.tel')['validation']\r\nValueError: BuilderConfig tatoeba.tel not found.\r\n\r\nbut language tel is actually included in xtreme:\r\nhttps:\/\/github.com\/google-research\/xtreme\/blob\/master\/utils_preprocess.py\r\ndef tatoeba_preprocess(args):\r\n lang3_dict = {\r\n 'afr':'af', 'ara':'ar', 'bul':'bg', 'ben':'bn',\r\n 'deu':'de', 'ell':'el', 'spa':'es', 'est':'et',\r\n 'eus':'eu', 'pes':'fa', 'fin':'fi', 'fra':'fr',\r\n 'heb':'he', 'hin':'hi', 'hun':'hu', 'ind':'id',\r\n 'ita':'it', 'jpn':'ja', 'jav':'jv', 'kat':'ka',\r\n 'kaz':'kk', 'kor':'ko', 'mal':'ml', 'mar':'mr',\r\n 'nld':'nl', 'por':'pt', 'rus':'ru', 'swh':'sw',\r\n 'tam':'ta', **_'tel':'te'_**, 'tha':'th', 'tgl':'tl', <----here\r\n 'tur':'tr', 'urd':'ur', 'vie':'vi', 'cmn':'zh',\r\n 'eng':'en',\r\n }","comment_length":13,"text":"Telugu subset missing for xtreme tatoeba dataset\nfrom nlp import load_dataset\r\ntrain_dataset = load_dataset('xtreme', 'tatoeba.tel')['validation']\r\nValueError: BuilderConfig tatoeba.tel not found.\r\n\r\nbut language tel is actually included in xtreme:\r\nhttps:\/\/github.com\/google-research\/xtreme\/blob\/master\/utils_preprocess.py\r\ndef tatoeba_preprocess(args):\r\n lang3_dict = {\r\n 'afr':'af', 'ara':'ar', 'bul':'bg', 'ben':'bn',\r\n 'deu':'de', 'ell':'el', 'spa':'es', 'est':'et',\r\n 'eus':'eu', 'pes':'fa', 'fin':'fi', 'fra':'fr',\r\n 'heb':'he', 'hin':'hi', 'hun':'hu', 'ind':'id',\r\n 'ita':'it', 'jpn':'ja', 'jav':'jv', 'kat':'ka',\r\n 'kaz':'kk', 'kor':'ko', 'mal':'ml', 'mar':'mr',\r\n 'nld':'nl', 'por':'pt', 'rus':'ru', 'swh':'sw',\r\n 'tam':'ta', **_'tel':'te'_**, 'tha':'th', 'tgl':'tl', <----here\r\n 'tur':'tr', 'urd':'ur', 'vie':'vi', 'cmn':'zh',\r\n 'eng':'en',\r\n }\nGood catch ! Thanks for reporting\r\n\r\nI just opened #2180 to fix this","embeddings":[0.0081472257,-0.4740612209,-0.1207175478,0.1003277153,0.3836947978,0.0219736528,-0.0741518587,0.3838887215,-0.0477930494,0.3231437504,-0.2175323963,-0.2406455725,0.1925039738,0.2936338186,0.4102523625,-0.2051515877,-0.1968390942,0.3289284408,-0.0433228835,-0.0294695795,-0.2333699167,0.3420246243,0.0100679379,0.1278151572,-0.1998337656,0.2883112431,0.191571787,-0.2423719168,-0.0300879795,-0.5815172195,0.3295686245,-0.3618333638,0.1036538705,0.2919726074,-0.0001150986,0.2549854219,0.0408828519,-0.2649563551,0.1293021888,-0.2724213302,-0.4591397643,-0.0538563468,-0.1495500654,0.0200889781,-0.4610052705,-0.058969073,0.0190169401,-0.2806847692,0.2631630599,0.5606830716,0.1647421271,0.4273344576,-0.0548489578,-0.0505464636,-0.0626819655,-0.0720730498,-0.2150718272,0.0581810996,0.3639322221,0.0138045149,0.1330360621,0.107224077,0.327891022,-0.1369395107,-0.1386733651,-0.2285147607,0.4280737042,-0.5590351224,0.1742521077,0.1858094186,0.2529549301,-0.2437177896,-0.2634807527,0.2399702072,-0.0928768963,-0.5383004546,0.2598460913,0.067157723,-0.0613438562,0.3037593365,0.1015027016,0.0488979854,-0.1178393215,0.2021034211,-0.14806889,0.5112597942,0.015614328,-0.1847814769,-0.3288085163,0.156325534,0.2538048029,0.0333610475,0.0375625044,0.3978129923,-0.3515993357,-0.1493224204,0.3362097144,-0.3389123678,-0.1326335073,-0.1219123527,0.1875628233,0.2261893749,-0.1832578033,-0.1011302993,0.0270444993,0.1670242548,0.4375785589,-0.0976029038,0.0706851333,-0.1731178313,0.2090131342,0.0814487338,-0.0283003524,-0.1652308702,0.1412807107,0.220321998,0.0247081202,-0.0580936335,-0.4059661329,-0.0169254486,-0.5235716701,-0.152658999,0.0051041637,-0.1041200832,-0.1197152212,-0.0660542175,0.0260300264,0.1090712696,-0.4609709978,-0.411137104,-0.2455268651,0.1454891115,-0.3769713342,-0.1765828431,0.1075834557,-0.3678260744,0.3345236182,-0.3309435844,-0.0518589765,0.078633301,-0.151731208,-0.2814720571,-0.1026903093,0.2382422984,0.0401342176,-0.0832504407,0.0154433725,-0.2886589468,-0.1299793422,0.0639405474,-0.0591060221,-0.1858469993,0.1508461088,0.1606275439,-0.2364718616,-0.0280952249,-0.0932319686,0.0671912283,0.3098608553,0.1633835137,-0.1237475276,-0.1541583538,-0.2995844781,0.0149218384,0.2554819286,0.1220898107,-0.4176228344,-0.2078413516,0.0385183841,-0.0657179803,0.0754471198,0.0212086029,0.082783699,0.7348763347,-0.1611014307,0.26870206,0.579397738,-0.0759604648,-0.3393458128,-0.0793261901,-0.0862817615,-0.0111270761,-0.0368292667,-0.233055234,0.0285503138,-0.2719120383,0.8242737055,0.266400665,-0.1268145591,0.1057820767,-0.0485045277,0.2281656861,-0.038500309,0.0690152645,0.1674057245,-0.0680348277,-0.0078273304,-0.1482862979,0.2530833185,0.0063676923,0.2534566224,0.2475417554,0.1489982009,-0.2633437812,-0.1275487393,-0.4068569541,0.0340735056,0.2033243775,-0.2149664611,0.5816870928,0.0504784808,0.2404974401,-0.4730454683,-0.0136559624,-0.2254167944,-0.1465106308,0.1232441962,0.2859692574,0.1320579499,0.2273974419,-0.1558450907,-0.0166615434,0.1121457815,0.0394762456,-0.0047545638,0.2167338282,-0.0334581882,-0.2427274883,0.0546892248,0.4098059833,-0.0124525381,-0.2784764767,-0.2280039638,0.1115998253,-0.0290106647,-0.1143924147,-0.2211723775,0.0905190632,0.2163776606,-0.3368979394,-0.0568291806,0.2261603028,-0.1100323796,-0.0297727864,0.4564161301,0.2776589692,0.3022156656,-0.0527978502,0.1982658356,0.175212428,0.3601967096,-0.0342874229,-0.2967642546,-0.2584277987,0.2293424308,0.0900747329,-0.050995212,0.1252831072,-0.3991313875,0.0598002188,0.4562910795,-0.0396084599,0.05126331,-0.1209940538,-0.2050481588,0.0876808912,-0.0499295145,0.3659425378,0.1882328689,0.0596695989,0.0593964979,-0.0952253491,-0.0094119404,-0.2487523109,0.1780026853,0.1008388847,-0.0050285733,0.2513669729,-0.0452748872,0.0060341903,-0.2725375593,0.1245651692,0.0423629992,0.1266268343,-0.1193586588,0.1266959757,-0.3887158632,-0.6041193008,-0.2400282621,-0.3747468889,0.0711577237,-0.1905244291,0.135339573,-0.3515725434,-0.0969452336,0.4300224483,-0.1701499671,-0.3398034573,-0.0489860885,-0.4517698586,-0.0085355258,-0.0580143854,-0.3656453788,0.1436698884,-0.0458059199,-0.1293352395,0.3429461122,-0.546661377,-0.0203602612,-0.080904454,-0.4865018725,0.2459937036,-0.332270354,0.1310341656,0.0137394909,0.1858745068,-0.12098784,0.1111608893,0.3854987323,0.3171402812,-0.2006774694,0.220505774,-0.1006027311,0.0236886311,0.0632848591,-0.6302369237,-0.2555051446,-0.2114951909,-0.1560374051,0.0867238939,0.2538009286,0.4947026968,-0.1437610984,0.0611954331,-0.028488284,-0.0368990712,0.0005143636,-0.2554267943,0.4332663715,-0.1734945178,-0.3294140697,0.1131780222,-0.1868522614,0.9218883514,0.2237723172,-0.214067176,0.2928204238,-0.1824646741,0.0751430914,-0.0240678713,-0.0602322184,0.380869329,0.0932757333,-0.0390465371,-0.0249775462,-0.0118661672,0.1148194224,0.3480728567,0.3682903349,-0.0034822235,0.3111259937,-0.0105685685,-0.0693781152,-0.0770203844,-0.0080637913,0.624569118,-0.0988233984,-0.0013535548,-0.1342616379,-0.2932055891,0.0180242453,0.1822457314,0.3139866889,0.0133474637,-0.0902340859,0.3163050711,-0.2724902034,-0.1016631126,-0.1091835201,-0.2127411813,0.1344108284,-0.2969148755,0.5625063181,0.082443364,-0.1488012671,-0.1505953372,-0.0509888679,-0.1240806654,-0.0020040891,0.1744022816,0.1120876074,-0.0425629877,-0.170651257,-0.3944185972,0.0655555353,0.1412004232,0.3264913857,0.0405161157,-0.0753007606,-0.0108901635,0.2815800905,-0.0593375601,0.0435079038,0.1077444702,0.1725022644,-0.3474192321,-0.2495247722,-0.0837301165,-0.061774265,0.0220573302,0.0992756486,0.3820450008,-0.0915149748,-0.3113657236,0.2438520044,0.3148612678,-0.2325810194,0.0490954146,-0.1814841777,-0.3758052886,-0.0177510679,0.0309470948,0.4228869677,0.4118698835,0.5924986005,0.1475509256,0.048172541,-0.1725339592,0.1250840873,0.0971712843,0.2283424437,-0.1191830859,0.1334037185,-0.4716374874,-0.0983823016,-0.004652332,0.3386758864,0.0819093585,-0.4305074215,0.0835795179,-0.0595371611,0.4451319873,0.0120673468,-0.0032998919,-0.0701207742,-0.3849641979,0.0199875142,0.112687856,0.1881052405,0.2446898371,0.2069455236,-0.2662058771,-0.2919672132,-0.0060067307,-0.1492649317,0.1614573151,-0.0129697695,0.2987875044,-0.5209451318,0.0000211158,0.0658341944,0.852961719,0.1445811242,0.165972814,0.4096148014,0.4359839559,0.0297565926,-0.2963291109,0.2020605803,-0.0071364851,-0.2488467693,-0.1031271368,0.0669144765,0.071935527,0.0148386741,-0.0885620043,0.3964858949,0.3691385686,0.1334099621,0.4344076812,-0.1111872643,0.0902026519,-0.2119510472,-0.2870597541,0.0719780475,-0.1609364003,0.3424044549,-0.0523457974,-0.0656778291,-0.1347440034,-0.0409246683,-0.1628625095,0.1812805384,-0.5029996634,-0.0319256894,0.1744598746,-0.1524999142,0.2616787851,0.1766181439,0.3356055915,0.2541533411,-0.1755714566,0.1771767437,0.1146403924,-0.1615727991,0.1176468059,0.1739595234,0.3560292423,-0.282307744,-0.2788254023,0.2218560576,-0.0368637852,0.2224075496,-0.3560603857,-0.1600732505,-0.4129335284,-0.1935932785,-0.1871946305,0.0586132742,-0.0119978543,-0.2696048915,0.1447538584,0.0576555841,-0.2839250863,-0.0139146354,-0.1300639659,-0.2805185318,-0.0423683226,0.1866814196,0.4432984591,-0.2136276215,0.5697958469,0.2289169729,-0.0134702157,-0.2021571845,-0.1881424785,0.2586969435,-0.3480764627,0.044895906,-0.1978939474,-0.1225335449,-0.0670215264,0.4852093756,-0.0824932754,-0.1656718403,0.0418132618,-0.1723080426,-0.3085631132,0.1666436791,0.4089539051,0.0486233011,-0.3338673115,0.8365927935,0.2375647128,-0.0278020706,-0.2647210658,-0.1866039783,-0.0480052568,0.2420064658,0.0257055443,-0.0112514263,0.3062949479,0.1018063799,0.1535425931,0.1329298764,-0.1876309067,-0.2041744143,0.0241574757,0.1198528856,0.155148223,-0.0772843435,0.3153624535,0.1816107929,-0.103470467,0.0366085842,0.0382168144,-0.0971308872,-0.0446023792,0.3217286468,0.3046946526,-0.0409155004,0.25280568,-0.011457203,-0.1206894815,0.1990374178,-0.0968796536,-0.0481387675,0.0828059241,-0.1704476625,-0.540635705,-0.0385873131,-0.1266419441,0.1291065663,0.0437033959,-0.2415408045,0.1321172565,0.0394318588,0.0668719634,0.0265806373,-0.1336541623,0.0154510709,0.013982119,0.1516285837,-0.5490376353,0.1983080953,0.1232938543,-0.0659800693,0.181345284,0.1536112428,-0.1797128916,0.2978903055,0.4117666781,0.0100858109,0.0468222909,-0.4622359872,0.3310923874,0.7521615624,0.1321132779,0.2750337422,-0.0710189566,-0.1720442772,-0.1037012786,0.4101836085,-0.273586452,0.2271741629,0.0203206614,-0.066973418,-0.0321497805,-0.235480994,-0.1814493984,0.1429905742,-0.16712825,-0.070802927,0.1316743344,0.1510990262,-0.017006794,0.1902512163,-0.2265704274,0.2596535385,-0.1839353144,0.1377488971,-0.1289617419,-0.0375373401,-0.2883542776,-0.1761026978,0.0330823101,0.2610011101,-0.1573628336,0.0062614577,-0.2152484953,0.1941943914,-0.4321984053,-0.1973073035,-0.1350215077,-0.1418539584,0.4517677724,0.1084383279,-0.2476790696,-0.5078828931,0.4655709863,0.1633500308,0.0663645789,0.124717705,0.2609624267,-0.1252962649,-0.1432977617,0.0688917041,0.2470301241,0.1971614659,0.5147587657,0.1388345808,0.0853874907,-0.1166845337,0.0226666182,0.0262726452,-0.1326970607,-0.0886279494,0.5451316833,0.0838164613,-0.1199633703,-0.2208171785,-0.1200228631,-0.4724487364,0.0417082421,0.2120845914,0.1089867949,0.1289449781,-0.2368172407,0.0736595914,-0.1525416225,0.5251851082,0.1492352784,-0.1000321433,-0.399572283,-0.2326316088,-0.3059510589,0.2283666879,-0.0169871394,-0.1106848046,0.0587329976,-0.2226564139,0.1736342311,0.208304137,0.0964925066,0.5351299644,-0.0264463574,-0.3213944137,-0.3405846357,-0.1751984358,0.1391472369,0.3621047139,0.0602483265,-0.1314232796,0.1394799501,0.1028284132,0.1433506608,-0.2918767035,-0.1402774304,0.4313609302,0.103055276,0.2454452068,0.0358265415,0.1644561887,-0.1526647657,-0.0247050691,-0.0477238484,-0.2288480699,-0.1149031818,0.3939548731,0.2463799417,0.3267915249,-0.0522970855,-0.1283489615,-0.4267992377,0.1538268179,-0.2292600572,0.0306395274,0.0512374155,0.2756350935,-0.1424334198,0.0027616089,0.2607244551,0.0278448109,0.0723159835,0.1573922336,-0.3989627659,-0.5171048045,0.1461491734,0.1133412272,-0.2171436995,-0.2345401049,-0.0695547387,-0.0000802124,-0.0108594159,-0.3946219683,0.1680442095,0.3049348295,0.0729677305,-0.2227362096,0.6062399745,-0.092063792,0.1142994538,0.1805490106,0.5679516196,0.0843261108,-0.0778560042,-0.072454825,-0.0360590927]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2148","title":"Add configurable options to `seqeval` metric","comments":"Hi @marrodion. \r\n\r\nThanks for pointing this out. It would be great to incorporate this metric-specific enhancement.\r\n\r\nAnother possibility would be to require the user to input the scheme as a string `mode=\"strict\", scheme=\"IOB2\"` and then dynamically import the corresponding module using Python `importlib`:\r\n```python\r\nif scheme:\r\n scheme = importlib.import_module(f\"seqeval.scheme.{scheme}\")\r\n```\r\n\r\nFeel free to create a Pull Request to make this contribution.","body":"Right now `load_metric(\"seqeval\")` only works in the default mode of evaluation (equivalent to conll evaluation).\r\n\r\nHowever, seqeval library [supports](https:\/\/github.com\/chakki-works\/seqeval#support-features) different evaluation schemes (IOB1, IOB2, etc.), which can be plugged in just by supporting additional kwargs in `Seqeval._compute`\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/85cf7ff920c90ca2e12bedca12b36d2a043c3da2\/metrics\/seqeval\/seqeval.py#L109\r\n\r\nThings that would be relevant are, for example, supporting `mode=\"strict\", scheme=IOB2` to count only full entity match as a true positive and omit partial matches.\r\n\r\nThe only problem I see is that the spirit of `metrics` seems to not require additional imports from user. `seqeval` only supports schemes as objects, without any string aliases. \r\n\r\nIt can be solved naively with mapping like `{\"IOB2\": seqeval.scheme.IOB2}`. Or just left as is and require user to explicitly import scheme from `seqeval` if he wants to configure it past the default implementation.\r\n\r\nIf that makes sense, I am happy to implement the change.","comment_length":61,"text":"Add configurable options to `seqeval` metric\nRight now `load_metric(\"seqeval\")` only works in the default mode of evaluation (equivalent to conll evaluation).\r\n\r\nHowever, seqeval library [supports](https:\/\/github.com\/chakki-works\/seqeval#support-features) different evaluation schemes (IOB1, IOB2, etc.), which can be plugged in just by supporting additional kwargs in `Seqeval._compute`\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/85cf7ff920c90ca2e12bedca12b36d2a043c3da2\/metrics\/seqeval\/seqeval.py#L109\r\n\r\nThings that would be relevant are, for example, supporting `mode=\"strict\", scheme=IOB2` to count only full entity match as a true positive and omit partial matches.\r\n\r\nThe only problem I see is that the spirit of `metrics` seems to not require additional imports from user. `seqeval` only supports schemes as objects, without any string aliases. \r\n\r\nIt can be solved naively with mapping like `{\"IOB2\": seqeval.scheme.IOB2}`. Or just left as is and require user to explicitly import scheme from `seqeval` if he wants to configure it past the default implementation.\r\n\r\nIf that makes sense, I am happy to implement the change.\nHi @marrodion. \r\n\r\nThanks for pointing this out. It would be great to incorporate this metric-specific enhancement.\r\n\r\nAnother possibility would be to require the user to input the scheme as a string `mode=\"strict\", scheme=\"IOB2\"` and then dynamically import the corresponding module using Python `importlib`:\r\n```python\r\nif scheme:\r\n scheme = importlib.import_module(f\"seqeval.scheme.{scheme}\")\r\n```\r\n\r\nFeel free to create a Pull Request to make this contribution.","embeddings":[-0.4404146075,0.1902088523,-0.0845154896,-0.1625581235,0.0749471933,-0.1645829231,0.1490213871,0.249832496,-0.0836082622,0.3530944288,-0.4670089781,0.2486553341,-0.0323203169,0.2693849504,0.0718923658,0.3165939152,-0.2818853259,-0.0426247083,0.0006639335,0.1087301672,-0.552981019,-0.0117782867,-0.1030309796,-0.15207313,0.0609822944,0.2906382382,0.0560519472,-0.022627607,-0.1254418641,-0.5280546546,0.0175919458,0.4460298717,-0.2212209105,-0.0601568706,-0.0001078836,-0.1581130475,0.1689354926,-0.2260208726,-0.209033519,-0.0844341889,-0.3633291125,-0.291166991,0.306224227,-0.2748051286,0.0388805196,0.2660091817,-0.0165568199,-0.1161660179,0.0370270275,0.1991532445,0.1833565831,0.1182956547,-0.1218209118,-0.0621905066,-0.209774971,-0.0393513143,-0.1454713047,0.3663655221,0.082575649,0.226062566,-0.1475075483,0.0548125058,-0.0825342089,-0.242446214,0.2761701941,-0.1193725616,0.6835107803,-0.0490721017,-0.0101480577,0.3530786335,0.2921997607,-0.1381694674,-0.3030595481,-0.1382803172,0.0993463695,-0.6759464741,0.1466280818,-0.096086137,-0.1030868888,-0.0826961473,-0.1479647011,-0.033458773,-0.3022654653,0.0747025311,-0.2574523985,0.3939167857,0.1031625718,-0.0649499893,0.3541351259,-0.0873658583,-0.2809019983,0.1270287186,-0.0665891171,0.1656541824,-0.7408687472,-0.1477435529,-0.0183482822,-0.1815031767,0.2883767486,0.1682549715,0.1724751145,0.4368014336,0.2517397404,0.3541308939,-0.021023307,0.343251586,0.3131861091,-0.0112347333,0.2474581152,0.0248375814,0.3714899421,-0.0004124012,0.0090165222,-0.659924984,0.0679211393,0.1159389913,-0.1485631764,-0.2398644835,-0.4179399014,-0.1582226902,-0.0586348698,-0.1113006771,0.5018102527,0.123510316,-0.1155994385,-0.0895082802,0.0192277562,0.4378791749,-0.1458706856,-0.0979548097,0.016352592,0.0411669351,0.204576686,0.0621869266,0.2487519979,-0.3678162396,0.1303182095,-0.0490685068,0.4444544017,-0.1659912616,0.1460129023,0.1006266922,-0.2069736719,0.0788970739,-0.1332622319,-0.421855092,0.0916285068,-0.2899059653,-0.4019014537,-0.0714191794,-0.1759268045,-0.4768915176,0.0616531745,0.2252954692,-0.5246810317,-0.0130113326,-0.008132752,0.8326738477,-0.3221261501,0.0141438488,-0.0476214364,0.1276525706,-0.4847742617,-0.0889536738,0.5480530858,0.3216397464,-0.22203134,-0.2398311794,-0.2551134527,-0.0794141144,-0.2599944174,-0.2022042722,-0.0602394193,0.0837334469,0.0311896801,-0.0284743086,0.7662084699,-0.4222574532,-0.0210452862,-0.0803242028,0.1119456962,-0.2272862643,0.1323760301,-0.0373109914,0.3576788902,0.0049101464,0.0843078047,0.2205335498,0.1616144478,-0.261389941,-0.2668775916,-0.2711868584,-0.2316744179,0.1784901768,0.3956591189,-0.0351171643,0.320366621,-0.0065064873,-0.2160911262,-0.0357711166,0.1151763126,0.0618478023,0.2012492418,-0.1130905524,0.0918779001,-0.1234125197,0.060049627,0.2182962447,-0.0904197395,0.0480323285,0.121668376,-0.1127960607,-0.4679356515,0.0559176914,-0.1898509711,0.055529885,0.2089476287,-0.3269338012,-0.0342498571,0.1159047559,0.0020938292,0.0913385451,0.2793194056,0.0862990469,-0.0139891878,-0.0003465335,0.211473912,0.0567865744,0.2087517381,0.4165763259,0.1913335472,0.2593639791,-0.109164238,0.4110058844,-0.050290972,0.1841282696,-0.0150859347,0.5559138656,0.0683776066,0.0065893224,0.0519290306,-0.051652614,0.0611868091,-0.0641518161,-0.2367987633,0.5868170857,-0.0794756785,0.2356154472,-0.3424955904,0.0648828372,0.1319375932,-0.095331423,-0.552742064,-0.3062807024,0.1004475355,-0.0132671436,0.2669738233,0.0566559285,0.0582371354,-0.3772333264,-0.1556093544,0.1259270161,0.1413586736,0.1581708342,0.0661022961,-0.2052513212,0.0382201746,-0.4662923217,0.2903058827,0.215023607,0.0154534616,-0.0470344312,-0.2343880236,-0.0794960707,0.3340309262,0.1572770625,0.0158218406,0.2130637616,0.3205015659,-0.2190813273,-0.3135822415,-0.0800573677,0.0409122705,-0.042243503,-0.3572460115,-0.094319433,-0.1978915185,0.2450003624,0.1717701554,0.0460357703,-0.202575013,-0.4122638106,0.5190328956,0.2028549761,-0.1720626652,0.1802853495,-0.1151275709,0.3317414224,-0.0443053022,-0.4356972575,-0.1409579813,-0.1817716956,0.0940515399,0.0956959128,0.0057353685,0.1029757261,0.3749670088,-0.0675891861,0.2991303205,0.0355553702,-0.6364486814,-0.0239804611,0.2268261462,0.5722032785,0.1552384496,-0.4056130052,-0.0692542568,-0.0509045385,0.2231383026,0.2037607431,-0.0061180177,0.4136137068,-0.0609643981,0.2502992749,-0.2219782174,0.0178008229,0.1000242978,-0.5327192545,0.2453567982,0.0594392419,-0.0078855697,0.0686219111,0.1769714504,0.0562076718,0.1112414449,0.1771620363,-0.1424997002,0.1927740872,0.4930687547,-0.2048598081,-0.0341397859,0.0772614628,-0.278519243,0.4551796317,0.0798746943,-0.1620344967,-0.5353206992,0.3236708641,0.2179788947,-0.0605301745,-0.4171498716,0.0311844852,0.1809620559,0.0233632885,-0.0555088222,-0.1814023256,0.2989302874,0.1583503783,0.1338731498,-0.0115191266,0.216440931,0.1655124575,0.0562133566,-0.0208813697,0.2980598509,0.0986942351,0.0352460034,0.2259937972,0.1511306763,-0.0841050372,0.1994831413,0.34873119,-0.1657783836,0.3828115463,0.267894119,0.1650658399,-0.2781140804,-0.1224360615,0.1228322983,-0.2981288433,0.1224511713,-0.2493659556,0.2479146421,-0.1362841576,-0.073262997,-0.3205817044,-0.0506801121,0.1360851973,0.3659796417,-0.0212371852,0.0480641089,0.1203468591,0.226745069,-0.1817407459,0.2449273616,-0.2822351456,-0.1319660097,0.1585785151,0.0025967341,0.1485465914,0.0906284973,0.4362581372,-0.1654890776,-0.2084530145,0.1888996512,-0.0820999593,-0.5329558849,0.035662204,-0.0349518172,-0.4825755954,0.2373024076,0.2342269719,-0.6754449606,-0.0595333017,0.1215938628,0.1488143653,-0.2491261959,-0.3720691502,-0.2491947562,-0.1661500931,-0.1850426495,0.1767451614,-0.0412340723,0.2086205781,0.0582409836,0.0278216004,-0.2370103747,-0.2062327117,0.2350769937,0.2506457567,0.179631874,-0.2191855758,0.2529246509,0.1343897879,-0.2753549218,-0.038240958,0.056091655,0.1188495681,-0.5144646168,0.1194530353,-0.0995482951,0.4271160066,-0.0142535213,-0.0842271373,0.1047257259,-0.4169352353,0.0466158949,-0.2416335791,0.2342704833,0.1022849157,0.1387833804,0.0718401819,-0.2300117016,0.716165185,0.033456631,-0.3667717576,-0.3400568068,-0.194125846,-0.2762601674,0.1269845665,0.537196219,0.6189163327,-0.2297585458,-0.4230226278,0.1905954778,-0.1576028019,0.1787083447,-0.2520562112,0.0649107993,-0.1693065315,0.3038463891,0.1148401797,0.0961574391,0.121356003,0.0285860412,-0.3229551315,0.2623012662,-0.087644875,-0.1266814321,0.0731281936,0.1338573545,-0.1751511842,-0.3573800325,-0.1863697469,0.1556977779,-0.0601053312,-0.151774615,0.0936809853,-0.1170482486,0.1334586442,0.0523656979,0.0779815838,-0.3301276863,-0.148598358,0.0718109831,-0.1227501556,-0.1340819448,0.1538654268,0.3677942455,0.3159064949,0.2506953776,-0.2192075849,0.0688008517,-0.1206844449,0.4233389199,0.1289048791,0.0205714796,0.0675243139,-0.1725013703,-0.0250886213,0.1401518434,0.0098187281,-0.2661958039,0.1412730813,0.0608119033,0.2566727698,0.0033595422,0.2854259908,-0.2257712781,0.2838926017,-0.1538018584,0.1529941708,0.2556661963,0.0278814826,0.2798300385,-0.0294745918,-0.2504778206,-0.0226105191,0.0359695777,-0.1192767695,-0.0034108604,0.1109194905,-0.5640182495,-0.0269925222,-0.0785509497,-0.215232417,-0.3589277565,-0.248124063,0.1937834322,0.2345780432,-0.2313770354,0.0745710582,0.3725794852,0.1676902771,-0.0997102559,0.0897696391,-0.1951106936,-0.0583488382,0.2283564061,-0.0534788556,0.1799024791,-0.3127804995,0.1394997686,-0.3911748528,0.1289292574,-0.3697018325,-0.2773509622,-0.0878168046,0.1667949557,-0.1807762235,0.1421960592,-0.0388468057,0.2374231517,0.1220599636,0.038196709,-0.4078208506,-0.1503542215,-0.4157817066,0.1182316765,-0.2259816378,0.2852988243,0.0547872372,0.1516323686,-0.1848931015,-0.3355185986,-0.1151427627,0.1512244344,-0.2504039705,0.0327739157,-0.0120636206,0.2532393634,0.0426800773,0.0884670243,0.0761528462,0.0691097751,-0.2445199639,0.4089044929,-0.4262246788,0.0013449463,0.3232171535,0.2378902137,0.1443791986,-0.2653337419,0.0735961124,0.4012955129,-0.0727842227,0.0548970625,0.1699572504,0.2629700899,0.0851075798,-0.0324252062,0.0058320998,0.213355571,-0.201592207,0.1152645499,-0.3776824176,-0.1356106251,0.3888047636,0.4075064361,0.4992360771,-0.2398553044,0.6050266623,-0.0692225695,0.0270401388,0.0179134589,0.0892326981,0.1072707623,-0.1823900193,0.1261910051,0.612381041,-0.0296222363,0.3305944502,-0.2045360506,-0.2899580896,0.0331656635,-0.0681630448,0.2578570843,0.0157055426,0.0174082071,0.2866869569,0.0368598737,-0.2332393825,0.0928656831,0.0050936537,0.3247486055,-0.1984857619,0.0808285475,-0.3066544831,0.0290483423,0.1574488729,-0.1903543025,-0.1305323392,-0.4349572659,-0.2236597091,-0.1116330475,0.1209933981,-0.2770488858,-0.0610651039,0.0820430368,0.4553857744,-0.1370194703,-0.0792978778,0.3357088268,0.2150689512,-0.3242680132,-0.0028113143,0.1278240234,0.0335069932,0.4683331847,0.3301254213,0.0862673521,-0.0277187917,0.0959929749,-0.2647826672,-0.1995587945,0.0546611883,-0.2562652826,0.0792407319,0.2268079817,-0.223317489,0.3783642948,0.1649148017,-0.1877298355,0.2148337662,-0.3457731605,0.0238631666,-0.6392326951,0.0873928964,0.4923708141,0.0124288946,0.0293330066,-0.4252647758,0.0293373261,-0.3556753695,-0.1338486224,-0.0051880777,0.0981243998,-0.084345296,0.1254210621,0.1143112555,-0.0538969189,-0.181490466,-0.3265304863,-0.1567293406,0.1709607989,-0.6458708644,0.0052651451,0.3642018139,0.032177031,0.1904265434,0.0564566553,-0.1080868617,0.4374260604,-0.2179744095,-0.3041597307,0.2744573355,-0.1127930284,-0.3386615813,-0.0245663002,0.1490684599,-0.0836659223,0.0014883829,0.0283097066,-0.0002028954,-0.3106565773,0.1391765326,-0.0663409233,-0.085462749,0.3035061955,0.2935439646,0.1796829104,0.1582034528,0.3154585063,-0.138074398,0.3068928719,-0.0322095156,0.0993027389,-0.1724681407,-0.0504151247,-0.1201308966,0.1211799234,-0.0257842075,-0.1413269341,0.0438551418,0.5314903259,0.0005547463,-0.6094732285,-0.0070513422,0.0545103252,0.1806107461,0.0297050979,-0.2594840229,0.2796793878,0.2102257311,0.1567269862,0.0636734515,-0.1860812753,0.276258707,-0.1737950295,-0.0388326682,-0.0562239364,0.3350422084,0.4395314157,0.1541651785,-0.1488437802,-0.0303662252,0.4258217216,0.2423169464,0.139175728,0.4387211204,0.035335809,-0.0228176657,-0.2492280304,-0.3756598532,0.1360049695,0.2666226923,-0.0287333801,-0.2172955126]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2146","title":"Dataset file size on disk is very large with 3D Array","comments":"Hi ! In the arrow file we store all the integers as uint8.\r\nSo your arrow file should weigh around `height x width x n_channels x n_images` bytes.\r\n\r\nWhat feature type do your TFDS dataset have ?\r\n\r\nIf it uses a `tfds.features.Image` type, then what is stored is the encoded data (as png or jpg for example). Since these encodings are made for compression, the resulting tfrecord is smaller that the arrow file.\r\n\r\nWe are working on adding a similar feature in `datasets`: the ability to store the encoded data instead of the raw integers for images, but also for audio data. This way, arrow files will have similar sizes as tfrecords for images.","body":"Hi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n","comment_length":114,"text":"Dataset file size on disk is very large with 3D Array\nHi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n\nHi ! In the arrow file we store all the integers as uint8.\r\nSo your arrow file should weigh around `height x width x n_channels x n_images` bytes.\r\n\r\nWhat feature type do your TFDS dataset have ?\r\n\r\nIf it uses a `tfds.features.Image` type, then what is stored is the encoded data (as png or jpg for example). Since these encodings are made for compression, the resulting tfrecord is smaller that the arrow file.\r\n\r\nWe are working on adding a similar feature in `datasets`: the ability to store the encoded data instead of the raw integers for images, but also for audio data. This way, arrow files will have similar sizes as tfrecords for images.","embeddings":[-0.1452738494,-0.1132628769,-0.1544179171,0.4201557934,0.2135634124,0.1139131263,0.5066809654,0.2730783522,0.0205232408,0.0358475372,-0.1937043518,0.0874153972,-0.1694584191,0.3150443435,0.0677503124,0.1006881595,-0.0745178536,0.2513983548,-0.1334785819,-0.0857200548,-0.0097005265,0.0191179775,-0.0179097727,-0.1838068217,-0.0615620911,-0.1613830328,0.0913496315,-0.0241849087,-0.0830982178,0.0797271058,0.2449781299,-0.0880117193,0.1781996787,0.5363877416,-0.0001173844,0.0502765737,0.108294405,-0.1655321717,0.0791868642,0.2893508673,-0.0925560445,-0.4095061719,-0.0746671557,-0.3576083779,0.0769588351,-0.3249549866,-0.1521175504,-0.1535450816,0.1848900914,0.0791001096,0.2110456824,0.1533048898,-0.0271132421,0.415045768,0.0993782207,0.5338887572,-0.1533458233,0.1268558055,-0.1435388923,0.4372014105,0.084297426,0.0795503631,0.1613585651,0.2877249718,0.5040504932,-0.055517599,0.2392112762,-0.246191889,0.1298937201,0.4001127183,0.8583656549,-0.2007374316,-0.1842802763,-0.3255720437,-0.1083301753,-0.0323392935,0.0225410033,0.3800633848,-0.049889937,0.021931123,-0.6522023082,-0.1392243952,-0.2269416898,0.1455018222,-0.2136535794,-0.4809675813,-0.1380903423,0.1503390074,-0.1890008897,-0.2658364773,0.2607783675,-0.3955922127,0.0734376535,-0.0997146145,-0.2659658492,-0.2101803869,-0.2905137539,-0.3803426623,0.3530271351,-0.1538060755,-0.0108280405,-0.151993528,-0.364893049,-0.1477414966,0.1039847285,0.3860372305,-0.4768467844,-0.138597995,-0.1418677419,-0.2950907052,-0.0408817492,-0.0522949062,-0.2465380579,-0.0907576829,0.1986332238,-0.4835402966,0.0639369488,-0.0618666336,0.0446808524,0.0238069706,0.0864812881,-0.0952553973,-0.075162895,0.1316531152,0.1410508007,0.135276109,-0.1400613934,-0.039203234,-0.1859151274,0.0416380167,-0.1229669973,-0.1046110019,0.0626311079,-0.081755273,0.1144735962,0.1505685151,0.052291356,-0.1904567331,0.0903884694,0.2001003176,0.2624739408,-0.3473196626,0.4073754847,0.3550085127,0.1589064002,0.1983184069,-0.2854207754,0.1398651004,-0.1997080892,0.4944943488,-0.5467603207,-0.0453600101,-0.4173412025,0.0906217396,-0.2234345526,-0.0543203168,-0.5728366375,0.0097432947,0.3395531476,-0.1186216325,0.0678242967,-0.2173455358,-0.3785401583,-0.294359833,-0.0371542424,0.0312655084,-0.7563121319,0.2612313628,-0.1514395922,0.0817364231,0.1952205747,0.5952161551,-0.0899516866,0.4252856076,-0.4255605638,-0.0484716371,0.0430041216,-0.0384001322,-0.6611084342,0.487341702,0.0706712455,0.0485031083,0.1124144495,0.320182234,0.1636859328,-0.016375415,-0.2228286564,0.3266060352,0.0910482332,0.0426690839,-0.1908326745,-0.3811230659,0.0882390067,0.0459806919,-0.1374954879,-0.1405356973,0.1314450502,0.0834871456,0.2107224613,-0.1292201728,0.0360603482,0.3705856502,0.2096632719,-0.2556730807,-0.0215890892,0.1423658431,-0.6317534447,0.1833813041,0.2472616285,-0.4363888502,0.1557535678,-0.1044431329,0.0382063761,-0.2296464294,-0.0895897746,0.2013011128,0.0237026345,-0.0005763058,-0.0091966204,-0.0104556819,-0.0085912012,-0.147752732,0.1148846447,-0.1881847829,-0.4114223123,0.356153816,0.213565439,-0.0350419208,-0.1056866199,-0.078861475,-0.3569892943,-0.341083467,-0.1865350902,0.0800963417,0.2511079609,0.1212477982,-0.0773310214,0.4837380052,0.2234310806,0.0761163831,0.3378689587,-0.1210732833,0.0945777297,-0.2206157446,-0.2020530403,0.216520682,-0.1918247491,0.0659026727,0.0202879235,-0.3593147099,0.2857810557,0.1333052814,0.1050533727,-0.2576690316,-0.0769441724,0.1929026395,0.3467417359,0.2029425353,0.1076560989,0.0467540659,0.6398187876,-0.1667219102,0.1070185304,0.364521265,-0.1265423596,-0.0783273652,0.1058712453,0.2452514023,0.5257766247,0.217031315,0.045964893,-0.0413736738,0.4925529659,-0.1443800777,0.0811916888,0.0678964555,0.5235673785,0.2810914516,0.1602386236,0.0027567402,-0.284455806,-0.1133067086,0.2043940723,0.2465661764,-0.303352654,0.1999263167,-0.1631348729,-0.2681356966,-0.0948000997,0.1902986467,-0.060111925,0.1248413771,-0.3277061582,-0.1944479644,0.0251019541,-0.0797675923,0.1439219266,0.0847972631,0.2937849462,-0.0679315031,0.2701685727,0.3433182836,-0.0878251344,0.0354879275,0.4142909348,-0.1485609859,0.3000029325,-0.1635025293,0.2096656561,-0.138071239,-0.2516806722,0.1204532534,0.0605554953,0.0706452951,-0.109574303,0.2740660906,-0.1995595098,0.2018909901,0.0286068246,0.0493995771,-0.2603569627,0.0002532348,-0.0481878854,-0.1286520362,-0.2174704075,0.0934461653,-0.2839459479,-0.050991375,0.2078645676,0.1816901416,0.2575588822,0.3705631793,0.1723432094,0.0412281081,0.1493278593,-0.0243310146,-0.1940806955,-0.4591633081,0.1459610909,-0.0282914434,-0.3725241125,0.1432911903,0.0114656771,0.192036286,0.2045514584,-0.6480360627,-0.0394913256,-0.293095082,0.0378666818,-0.17624484,0.1218835488,0.0200794712,-0.2159727663,0.0826531723,-0.0400286093,-0.1227325797,-0.0801262707,0.1514735818,0.1209952906,0.127742365,0.5383259654,0.2677577734,0.410494417,0.1294866651,-0.3016063273,0.1248357072,-0.1209927276,0.3700109422,-0.2481317222,0.0517883822,0.3423662782,0.1794074625,-0.2017464638,0.2401463687,-0.0048973123,-0.070583038,0.3367039561,0.0287895519,-0.1383411288,0.0239062179,0.1935496777,-0.0604689121,0.010493882,-0.0702693909,-0.0987723097,-0.1590381861,-0.3050752282,-0.087373741,0.1904460341,0.2198341787,0.0339709073,-0.3773390353,-0.114849709,-0.3479646742,0.1886437535,-0.1687081754,0.3023189008,0.0322840959,0.2741390169,0.2475180477,0.1615713686,0.9262683988,-0.0916073173,0.0714728311,-0.1972740144,-0.0435676649,-0.5894108415,-0.1542813331,-0.0789639726,0.0270803031,-0.0590391792,0.5538505316,-0.0631092489,-0.0686066449,-0.261759311,0.2991942167,-0.3168808818,-0.3877653778,-0.2755343318,-0.2933425307,-0.1356876642,-0.3935822845,0.0199943967,-0.0869563818,0.0755996183,-0.3193262815,-0.267570585,-0.1990125477,0.0541381128,0.0145461811,0.3740357757,0.3435080349,0.2086753249,-0.0869254544,0.3284698129,0.3662003279,0.2595766485,0.0718757212,0.0568668023,0.2052865177,-0.2167998701,0.0861621499,0.1083574519,-0.0250589009,-0.2877222598,-0.0670191571,0.1119900942,-0.2830277085,0.1024947241,0.3718264997,-0.158607617,-0.0864219889,-0.3929764926,0.3548669219,0.0121597042,0.0940138474,0.0360056087,-0.0519523099,-0.1494732946,0.1480955929,0.2831670642,0.5525882244,-0.3548772931,0.3033214211,0.2974138558,0.0017682018,0.0114030931,-0.0265308805,-0.168065086,-0.1445067227,-0.3122710884,0.0030846142,-0.1878083795,-0.0254937578,0.4849464893,0.1750966311,0.0439129099,0.2067071199,0.0071276291,-0.1149759442,0.5293208361,0.299548775,-0.2283003181,0.0325021446,0.0639349148,-0.254832387,-0.1903491616,-0.0622595958,-0.0863793641,-0.0984942764,0.2215803266,-0.0403396897,-0.2126670629,-0.123895891,0.3679513037,-0.1082212925,-0.2316224575,0.1815731823,0.0838665143,-0.2010200322,0.196623072,-0.157318905,0.1279699206,0.0030061631,0.3249221146,0.1413984448,-0.1144990996,-0.0310456324,-0.0913244933,0.0234271865,-0.1943868995,0.1888038814,-0.1482692063,-0.0629260316,0.0454889238,0.029789513,-0.096110031,-0.653003335,-0.3431927264,-0.2048236728,0.4010295868,0.0805630684,0.1309611499,0.2124831229,0.4965970218,0.0243887007,-0.3730357587,-0.1403264701,0.1895269006,0.2858310342,0.1799467951,0.3866920769,-0.1341089606,-0.1365939081,-0.0709895268,0.6510813832,0.1012729779,-0.1912544817,-0.057765983,-0.0403905585,0.3169389963,0.0208149981,-0.334382683,0.188358143,-0.0939157605,-0.3599133492,-0.2077349275,-0.137304768,0.1214710847,-0.1601488441,0.2943194509,0.0504378565,-0.1864531487,0.0016885888,-0.0764309838,-0.2049330324,0.1969459504,0.2957784832,0.0060840035,-0.3328675926,0.0645710304,-0.0876779109,-0.2682572603,-0.0669379458,-0.1471435428,-0.1714788526,-0.1753690541,0.0967811123,0.1115372702,0.1287063509,-0.2801164687,0.1016181931,-0.3750437796,-0.4788466394,-0.13488096,0.1319559813,-0.0679693297,-0.090987362,0.103313379,0.0220806133,-0.251005441,0.007965601,0.1416076869,0.0354233049,0.2988489568,0.1892450303,0.1612934768,0.1742737293,0.0425288305,-0.4761444926,0.0175630245,-0.0588393733,0.0995729342,0.1874175221,-0.1962634623,0.2503475547,0.2590202689,0.3225476444,0.1073599681,-0.4383759499,-0.2273101062,-0.0798544213,0.1225005761,0.0510145389,-0.1831585467,0.1945354044,0.28687644,-0.0688323528,0.3184705377,0.2352006137,0.3476952314,-0.4201479852,0.0468274653,0.4110737741,0.0671937168,-0.1297736764,0.4717316329,-0.2036245912,0.0523640476,0.0846017823,0.2946646512,0.1201955527,0.8587027788,0.12909168,0.1891805381,0.4139887989,-0.0303379316,0.0068449378,-0.4966722131,-0.0952328369,0.1984754503,-0.168075338,0.2741121948,0.170211941,-0.3590529263,0.1909864247,-0.3852767646,-0.2336922735,-0.0620264858,-0.2622532547,0.1335817128,-0.0030272065,0.0273247976,0.1620105356,0.2371406257,-0.3845196962,-0.3014771342,-0.2461311817,0.1889080554,-0.2023765892,0.2124773413,0.4214303195,0.0444422476,-0.0315243825,-0.0391092412,0.3646275699,0.0986841619,0.1010680571,-0.0881536752,0.0778101683,0.1303085536,0.1942219287,0.1864715815,-0.0245688483,0.229917556,-0.0123417871,-0.0638061985,0.0230799001,-0.1007118225,-0.0150630157,0.1352045685,0.0769988522,-0.0195251871,-0.1023717076,0.4225893915,-0.0051351315,-0.4866001904,-0.0333422124,-0.4977223873,-0.0050530042,-0.051477354,0.0287456643,-0.4116464853,-0.2353772074,0.5690826178,-0.2013360709,0.2559473515,0.0774657801,0.0456615984,-0.6079736948,0.6227496862,0.1627824455,0.0933043957,-0.0685406327,0.1594237983,-0.5500161648,0.0303858668,-0.2166630477,-0.1015652195,0.233997196,0.4140377045,-0.0308639854,-0.0486871824,0.1650849581,0.003842826,-0.1916438937,0.0894423947,-0.2841004431,-0.1390906125,-0.1443412304,0.1185550466,-0.1753247529,-0.4083871543,0.3195689321,-0.0093015609,-0.0767879412,-0.1066462398,0.2173395455,-0.1415829957,0.4878300726,0.4013835788,-0.0483230203,0.3463920951,0.0004342151,0.0175026599,0.0451892316,0.1241305843,-0.0279618651,0.7077245712,0.0045384844,0.1401189715,-0.0876325294,0.0933168158,0.1183361933,0.0912858322,-0.0013901879,-0.1583742946,-0.1358546168,0.0508204177,-0.1527412385,0.0265369955,0.1540860236,0.1232053787,0.0315913707,-0.0814922303,-0.0285235085,-0.0700531751,0.3188904226,0.0257114191,-0.0847275257,0.0533844195,-0.0156111252,-0.0210556425,-0.0187305622,-0.2966986299,0.201208815,0.4706384838,-0.2481852323,-0.0093808491,0.3947989643,-0.2101584971,0.0058913021,-0.0756310448,0.7668739557,-0.1517630368,-0.3578999937,-0.1258624494,-0.0418124832]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2146","title":"Dataset file size on disk is very large with 3D Array","comments":"Thanks for the prompt response. You're right about the encoding, I have the `tfds.features.Image` feature type you mentioned.\r\nHowever, as described in the `dataset_info.json`, my dataset is made of 1479 (224x224x3) images. 1479 x 224 x 224 x 3 = 222630912 bytes which is far from the actual size 520803408 bytes. \r\n\r\nAnyway I look forward to the Image feature type in `datasets`. ","body":"Hi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n","comment_length":62,"text":"Dataset file size on disk is very large with 3D Array\nHi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n\nThanks for the prompt response. You're right about the encoding, I have the `tfds.features.Image` feature type you mentioned.\r\nHowever, as described in the `dataset_info.json`, my dataset is made of 1479 (224x224x3) images. 1479 x 224 x 224 x 3 = 222630912 bytes which is far from the actual size 520803408 bytes. \r\n\r\nAnyway I look forward to the Image feature type in `datasets`. ","embeddings":[-0.1452738494,-0.1132628769,-0.1544179171,0.4201557934,0.2135634124,0.1139131263,0.5066809654,0.2730783522,0.0205232408,0.0358475372,-0.1937043518,0.0874153972,-0.1694584191,0.3150443435,0.0677503124,0.1006881595,-0.0745178536,0.2513983548,-0.1334785819,-0.0857200548,-0.0097005265,0.0191179775,-0.0179097727,-0.1838068217,-0.0615620911,-0.1613830328,0.0913496315,-0.0241849087,-0.0830982178,0.0797271058,0.2449781299,-0.0880117193,0.1781996787,0.5363877416,-0.0001173844,0.0502765737,0.108294405,-0.1655321717,0.0791868642,0.2893508673,-0.0925560445,-0.4095061719,-0.0746671557,-0.3576083779,0.0769588351,-0.3249549866,-0.1521175504,-0.1535450816,0.1848900914,0.0791001096,0.2110456824,0.1533048898,-0.0271132421,0.415045768,0.0993782207,0.5338887572,-0.1533458233,0.1268558055,-0.1435388923,0.4372014105,0.084297426,0.0795503631,0.1613585651,0.2877249718,0.5040504932,-0.055517599,0.2392112762,-0.246191889,0.1298937201,0.4001127183,0.8583656549,-0.2007374316,-0.1842802763,-0.3255720437,-0.1083301753,-0.0323392935,0.0225410033,0.3800633848,-0.049889937,0.021931123,-0.6522023082,-0.1392243952,-0.2269416898,0.1455018222,-0.2136535794,-0.4809675813,-0.1380903423,0.1503390074,-0.1890008897,-0.2658364773,0.2607783675,-0.3955922127,0.0734376535,-0.0997146145,-0.2659658492,-0.2101803869,-0.2905137539,-0.3803426623,0.3530271351,-0.1538060755,-0.0108280405,-0.151993528,-0.364893049,-0.1477414966,0.1039847285,0.3860372305,-0.4768467844,-0.138597995,-0.1418677419,-0.2950907052,-0.0408817492,-0.0522949062,-0.2465380579,-0.0907576829,0.1986332238,-0.4835402966,0.0639369488,-0.0618666336,0.0446808524,0.0238069706,0.0864812881,-0.0952553973,-0.075162895,0.1316531152,0.1410508007,0.135276109,-0.1400613934,-0.039203234,-0.1859151274,0.0416380167,-0.1229669973,-0.1046110019,0.0626311079,-0.081755273,0.1144735962,0.1505685151,0.052291356,-0.1904567331,0.0903884694,0.2001003176,0.2624739408,-0.3473196626,0.4073754847,0.3550085127,0.1589064002,0.1983184069,-0.2854207754,0.1398651004,-0.1997080892,0.4944943488,-0.5467603207,-0.0453600101,-0.4173412025,0.0906217396,-0.2234345526,-0.0543203168,-0.5728366375,0.0097432947,0.3395531476,-0.1186216325,0.0678242967,-0.2173455358,-0.3785401583,-0.294359833,-0.0371542424,0.0312655084,-0.7563121319,0.2612313628,-0.1514395922,0.0817364231,0.1952205747,0.5952161551,-0.0899516866,0.4252856076,-0.4255605638,-0.0484716371,0.0430041216,-0.0384001322,-0.6611084342,0.487341702,0.0706712455,0.0485031083,0.1124144495,0.320182234,0.1636859328,-0.016375415,-0.2228286564,0.3266060352,0.0910482332,0.0426690839,-0.1908326745,-0.3811230659,0.0882390067,0.0459806919,-0.1374954879,-0.1405356973,0.1314450502,0.0834871456,0.2107224613,-0.1292201728,0.0360603482,0.3705856502,0.2096632719,-0.2556730807,-0.0215890892,0.1423658431,-0.6317534447,0.1833813041,0.2472616285,-0.4363888502,0.1557535678,-0.1044431329,0.0382063761,-0.2296464294,-0.0895897746,0.2013011128,0.0237026345,-0.0005763058,-0.0091966204,-0.0104556819,-0.0085912012,-0.147752732,0.1148846447,-0.1881847829,-0.4114223123,0.356153816,0.213565439,-0.0350419208,-0.1056866199,-0.078861475,-0.3569892943,-0.341083467,-0.1865350902,0.0800963417,0.2511079609,0.1212477982,-0.0773310214,0.4837380052,0.2234310806,0.0761163831,0.3378689587,-0.1210732833,0.0945777297,-0.2206157446,-0.2020530403,0.216520682,-0.1918247491,0.0659026727,0.0202879235,-0.3593147099,0.2857810557,0.1333052814,0.1050533727,-0.2576690316,-0.0769441724,0.1929026395,0.3467417359,0.2029425353,0.1076560989,0.0467540659,0.6398187876,-0.1667219102,0.1070185304,0.364521265,-0.1265423596,-0.0783273652,0.1058712453,0.2452514023,0.5257766247,0.217031315,0.045964893,-0.0413736738,0.4925529659,-0.1443800777,0.0811916888,0.0678964555,0.5235673785,0.2810914516,0.1602386236,0.0027567402,-0.284455806,-0.1133067086,0.2043940723,0.2465661764,-0.303352654,0.1999263167,-0.1631348729,-0.2681356966,-0.0948000997,0.1902986467,-0.060111925,0.1248413771,-0.3277061582,-0.1944479644,0.0251019541,-0.0797675923,0.1439219266,0.0847972631,0.2937849462,-0.0679315031,0.2701685727,0.3433182836,-0.0878251344,0.0354879275,0.4142909348,-0.1485609859,0.3000029325,-0.1635025293,0.2096656561,-0.138071239,-0.2516806722,0.1204532534,0.0605554953,0.0706452951,-0.109574303,0.2740660906,-0.1995595098,0.2018909901,0.0286068246,0.0493995771,-0.2603569627,0.0002532348,-0.0481878854,-0.1286520362,-0.2174704075,0.0934461653,-0.2839459479,-0.050991375,0.2078645676,0.1816901416,0.2575588822,0.3705631793,0.1723432094,0.0412281081,0.1493278593,-0.0243310146,-0.1940806955,-0.4591633081,0.1459610909,-0.0282914434,-0.3725241125,0.1432911903,0.0114656771,0.192036286,0.2045514584,-0.6480360627,-0.0394913256,-0.293095082,0.0378666818,-0.17624484,0.1218835488,0.0200794712,-0.2159727663,0.0826531723,-0.0400286093,-0.1227325797,-0.0801262707,0.1514735818,0.1209952906,0.127742365,0.5383259654,0.2677577734,0.410494417,0.1294866651,-0.3016063273,0.1248357072,-0.1209927276,0.3700109422,-0.2481317222,0.0517883822,0.3423662782,0.1794074625,-0.2017464638,0.2401463687,-0.0048973123,-0.070583038,0.3367039561,0.0287895519,-0.1383411288,0.0239062179,0.1935496777,-0.0604689121,0.010493882,-0.0702693909,-0.0987723097,-0.1590381861,-0.3050752282,-0.087373741,0.1904460341,0.2198341787,0.0339709073,-0.3773390353,-0.114849709,-0.3479646742,0.1886437535,-0.1687081754,0.3023189008,0.0322840959,0.2741390169,0.2475180477,0.1615713686,0.9262683988,-0.0916073173,0.0714728311,-0.1972740144,-0.0435676649,-0.5894108415,-0.1542813331,-0.0789639726,0.0270803031,-0.0590391792,0.5538505316,-0.0631092489,-0.0686066449,-0.261759311,0.2991942167,-0.3168808818,-0.3877653778,-0.2755343318,-0.2933425307,-0.1356876642,-0.3935822845,0.0199943967,-0.0869563818,0.0755996183,-0.3193262815,-0.267570585,-0.1990125477,0.0541381128,0.0145461811,0.3740357757,0.3435080349,0.2086753249,-0.0869254544,0.3284698129,0.3662003279,0.2595766485,0.0718757212,0.0568668023,0.2052865177,-0.2167998701,0.0861621499,0.1083574519,-0.0250589009,-0.2877222598,-0.0670191571,0.1119900942,-0.2830277085,0.1024947241,0.3718264997,-0.158607617,-0.0864219889,-0.3929764926,0.3548669219,0.0121597042,0.0940138474,0.0360056087,-0.0519523099,-0.1494732946,0.1480955929,0.2831670642,0.5525882244,-0.3548772931,0.3033214211,0.2974138558,0.0017682018,0.0114030931,-0.0265308805,-0.168065086,-0.1445067227,-0.3122710884,0.0030846142,-0.1878083795,-0.0254937578,0.4849464893,0.1750966311,0.0439129099,0.2067071199,0.0071276291,-0.1149759442,0.5293208361,0.299548775,-0.2283003181,0.0325021446,0.0639349148,-0.254832387,-0.1903491616,-0.0622595958,-0.0863793641,-0.0984942764,0.2215803266,-0.0403396897,-0.2126670629,-0.123895891,0.3679513037,-0.1082212925,-0.2316224575,0.1815731823,0.0838665143,-0.2010200322,0.196623072,-0.157318905,0.1279699206,0.0030061631,0.3249221146,0.1413984448,-0.1144990996,-0.0310456324,-0.0913244933,0.0234271865,-0.1943868995,0.1888038814,-0.1482692063,-0.0629260316,0.0454889238,0.029789513,-0.096110031,-0.653003335,-0.3431927264,-0.2048236728,0.4010295868,0.0805630684,0.1309611499,0.2124831229,0.4965970218,0.0243887007,-0.3730357587,-0.1403264701,0.1895269006,0.2858310342,0.1799467951,0.3866920769,-0.1341089606,-0.1365939081,-0.0709895268,0.6510813832,0.1012729779,-0.1912544817,-0.057765983,-0.0403905585,0.3169389963,0.0208149981,-0.334382683,0.188358143,-0.0939157605,-0.3599133492,-0.2077349275,-0.137304768,0.1214710847,-0.1601488441,0.2943194509,0.0504378565,-0.1864531487,0.0016885888,-0.0764309838,-0.2049330324,0.1969459504,0.2957784832,0.0060840035,-0.3328675926,0.0645710304,-0.0876779109,-0.2682572603,-0.0669379458,-0.1471435428,-0.1714788526,-0.1753690541,0.0967811123,0.1115372702,0.1287063509,-0.2801164687,0.1016181931,-0.3750437796,-0.4788466394,-0.13488096,0.1319559813,-0.0679693297,-0.090987362,0.103313379,0.0220806133,-0.251005441,0.007965601,0.1416076869,0.0354233049,0.2988489568,0.1892450303,0.1612934768,0.1742737293,0.0425288305,-0.4761444926,0.0175630245,-0.0588393733,0.0995729342,0.1874175221,-0.1962634623,0.2503475547,0.2590202689,0.3225476444,0.1073599681,-0.4383759499,-0.2273101062,-0.0798544213,0.1225005761,0.0510145389,-0.1831585467,0.1945354044,0.28687644,-0.0688323528,0.3184705377,0.2352006137,0.3476952314,-0.4201479852,0.0468274653,0.4110737741,0.0671937168,-0.1297736764,0.4717316329,-0.2036245912,0.0523640476,0.0846017823,0.2946646512,0.1201955527,0.8587027788,0.12909168,0.1891805381,0.4139887989,-0.0303379316,0.0068449378,-0.4966722131,-0.0952328369,0.1984754503,-0.168075338,0.2741121948,0.170211941,-0.3590529263,0.1909864247,-0.3852767646,-0.2336922735,-0.0620264858,-0.2622532547,0.1335817128,-0.0030272065,0.0273247976,0.1620105356,0.2371406257,-0.3845196962,-0.3014771342,-0.2461311817,0.1889080554,-0.2023765892,0.2124773413,0.4214303195,0.0444422476,-0.0315243825,-0.0391092412,0.3646275699,0.0986841619,0.1010680571,-0.0881536752,0.0778101683,0.1303085536,0.1942219287,0.1864715815,-0.0245688483,0.229917556,-0.0123417871,-0.0638061985,0.0230799001,-0.1007118225,-0.0150630157,0.1352045685,0.0769988522,-0.0195251871,-0.1023717076,0.4225893915,-0.0051351315,-0.4866001904,-0.0333422124,-0.4977223873,-0.0050530042,-0.051477354,0.0287456643,-0.4116464853,-0.2353772074,0.5690826178,-0.2013360709,0.2559473515,0.0774657801,0.0456615984,-0.6079736948,0.6227496862,0.1627824455,0.0933043957,-0.0685406327,0.1594237983,-0.5500161648,0.0303858668,-0.2166630477,-0.1015652195,0.233997196,0.4140377045,-0.0308639854,-0.0486871824,0.1650849581,0.003842826,-0.1916438937,0.0894423947,-0.2841004431,-0.1390906125,-0.1443412304,0.1185550466,-0.1753247529,-0.4083871543,0.3195689321,-0.0093015609,-0.0767879412,-0.1066462398,0.2173395455,-0.1415829957,0.4878300726,0.4013835788,-0.0483230203,0.3463920951,0.0004342151,0.0175026599,0.0451892316,0.1241305843,-0.0279618651,0.7077245712,0.0045384844,0.1401189715,-0.0876325294,0.0933168158,0.1183361933,0.0912858322,-0.0013901879,-0.1583742946,-0.1358546168,0.0508204177,-0.1527412385,0.0265369955,0.1540860236,0.1232053787,0.0315913707,-0.0814922303,-0.0285235085,-0.0700531751,0.3188904226,0.0257114191,-0.0847275257,0.0533844195,-0.0156111252,-0.0210556425,-0.0187305622,-0.2966986299,0.201208815,0.4706384838,-0.2481852323,-0.0093808491,0.3947989643,-0.2101584971,0.0058913021,-0.0756310448,0.7668739557,-0.1517630368,-0.3578999937,-0.1258624494,-0.0418124832]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2146","title":"Dataset file size on disk is very large with 3D Array","comments":"@lhoestq I changed the data structure so I have a 2D Array feature type instead of a 3D Array by grouping the two last dimensions ( a 224x672 2D Array instead of a 224x224x3 3D Array). The file size is now 223973964 bytes, nearly half the previous size! Which is around of what I would expect.\r\nI found similar behavior in existing `datasets` collection, when comparing black and white vs color image, for example MNIST vs CIFAR. ","body":"Hi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n","comment_length":77,"text":"Dataset file size on disk is very large with 3D Array\nHi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n\n@lhoestq I changed the data structure so I have a 2D Array feature type instead of a 3D Array by grouping the two last dimensions ( a 224x672 2D Array instead of a 224x224x3 3D Array). The file size is now 223973964 bytes, nearly half the previous size! Which is around of what I would expect.\r\nI found similar behavior in existing `datasets` collection, when comparing black and white vs color image, for example MNIST vs CIFAR. ","embeddings":[-0.1452738494,-0.1132628769,-0.1544179171,0.4201557934,0.2135634124,0.1139131263,0.5066809654,0.2730783522,0.0205232408,0.0358475372,-0.1937043518,0.0874153972,-0.1694584191,0.3150443435,0.0677503124,0.1006881595,-0.0745178536,0.2513983548,-0.1334785819,-0.0857200548,-0.0097005265,0.0191179775,-0.0179097727,-0.1838068217,-0.0615620911,-0.1613830328,0.0913496315,-0.0241849087,-0.0830982178,0.0797271058,0.2449781299,-0.0880117193,0.1781996787,0.5363877416,-0.0001173844,0.0502765737,0.108294405,-0.1655321717,0.0791868642,0.2893508673,-0.0925560445,-0.4095061719,-0.0746671557,-0.3576083779,0.0769588351,-0.3249549866,-0.1521175504,-0.1535450816,0.1848900914,0.0791001096,0.2110456824,0.1533048898,-0.0271132421,0.415045768,0.0993782207,0.5338887572,-0.1533458233,0.1268558055,-0.1435388923,0.4372014105,0.084297426,0.0795503631,0.1613585651,0.2877249718,0.5040504932,-0.055517599,0.2392112762,-0.246191889,0.1298937201,0.4001127183,0.8583656549,-0.2007374316,-0.1842802763,-0.3255720437,-0.1083301753,-0.0323392935,0.0225410033,0.3800633848,-0.049889937,0.021931123,-0.6522023082,-0.1392243952,-0.2269416898,0.1455018222,-0.2136535794,-0.4809675813,-0.1380903423,0.1503390074,-0.1890008897,-0.2658364773,0.2607783675,-0.3955922127,0.0734376535,-0.0997146145,-0.2659658492,-0.2101803869,-0.2905137539,-0.3803426623,0.3530271351,-0.1538060755,-0.0108280405,-0.151993528,-0.364893049,-0.1477414966,0.1039847285,0.3860372305,-0.4768467844,-0.138597995,-0.1418677419,-0.2950907052,-0.0408817492,-0.0522949062,-0.2465380579,-0.0907576829,0.1986332238,-0.4835402966,0.0639369488,-0.0618666336,0.0446808524,0.0238069706,0.0864812881,-0.0952553973,-0.075162895,0.1316531152,0.1410508007,0.135276109,-0.1400613934,-0.039203234,-0.1859151274,0.0416380167,-0.1229669973,-0.1046110019,0.0626311079,-0.081755273,0.1144735962,0.1505685151,0.052291356,-0.1904567331,0.0903884694,0.2001003176,0.2624739408,-0.3473196626,0.4073754847,0.3550085127,0.1589064002,0.1983184069,-0.2854207754,0.1398651004,-0.1997080892,0.4944943488,-0.5467603207,-0.0453600101,-0.4173412025,0.0906217396,-0.2234345526,-0.0543203168,-0.5728366375,0.0097432947,0.3395531476,-0.1186216325,0.0678242967,-0.2173455358,-0.3785401583,-0.294359833,-0.0371542424,0.0312655084,-0.7563121319,0.2612313628,-0.1514395922,0.0817364231,0.1952205747,0.5952161551,-0.0899516866,0.4252856076,-0.4255605638,-0.0484716371,0.0430041216,-0.0384001322,-0.6611084342,0.487341702,0.0706712455,0.0485031083,0.1124144495,0.320182234,0.1636859328,-0.016375415,-0.2228286564,0.3266060352,0.0910482332,0.0426690839,-0.1908326745,-0.3811230659,0.0882390067,0.0459806919,-0.1374954879,-0.1405356973,0.1314450502,0.0834871456,0.2107224613,-0.1292201728,0.0360603482,0.3705856502,0.2096632719,-0.2556730807,-0.0215890892,0.1423658431,-0.6317534447,0.1833813041,0.2472616285,-0.4363888502,0.1557535678,-0.1044431329,0.0382063761,-0.2296464294,-0.0895897746,0.2013011128,0.0237026345,-0.0005763058,-0.0091966204,-0.0104556819,-0.0085912012,-0.147752732,0.1148846447,-0.1881847829,-0.4114223123,0.356153816,0.213565439,-0.0350419208,-0.1056866199,-0.078861475,-0.3569892943,-0.341083467,-0.1865350902,0.0800963417,0.2511079609,0.1212477982,-0.0773310214,0.4837380052,0.2234310806,0.0761163831,0.3378689587,-0.1210732833,0.0945777297,-0.2206157446,-0.2020530403,0.216520682,-0.1918247491,0.0659026727,0.0202879235,-0.3593147099,0.2857810557,0.1333052814,0.1050533727,-0.2576690316,-0.0769441724,0.1929026395,0.3467417359,0.2029425353,0.1076560989,0.0467540659,0.6398187876,-0.1667219102,0.1070185304,0.364521265,-0.1265423596,-0.0783273652,0.1058712453,0.2452514023,0.5257766247,0.217031315,0.045964893,-0.0413736738,0.4925529659,-0.1443800777,0.0811916888,0.0678964555,0.5235673785,0.2810914516,0.1602386236,0.0027567402,-0.284455806,-0.1133067086,0.2043940723,0.2465661764,-0.303352654,0.1999263167,-0.1631348729,-0.2681356966,-0.0948000997,0.1902986467,-0.060111925,0.1248413771,-0.3277061582,-0.1944479644,0.0251019541,-0.0797675923,0.1439219266,0.0847972631,0.2937849462,-0.0679315031,0.2701685727,0.3433182836,-0.0878251344,0.0354879275,0.4142909348,-0.1485609859,0.3000029325,-0.1635025293,0.2096656561,-0.138071239,-0.2516806722,0.1204532534,0.0605554953,0.0706452951,-0.109574303,0.2740660906,-0.1995595098,0.2018909901,0.0286068246,0.0493995771,-0.2603569627,0.0002532348,-0.0481878854,-0.1286520362,-0.2174704075,0.0934461653,-0.2839459479,-0.050991375,0.2078645676,0.1816901416,0.2575588822,0.3705631793,0.1723432094,0.0412281081,0.1493278593,-0.0243310146,-0.1940806955,-0.4591633081,0.1459610909,-0.0282914434,-0.3725241125,0.1432911903,0.0114656771,0.192036286,0.2045514584,-0.6480360627,-0.0394913256,-0.293095082,0.0378666818,-0.17624484,0.1218835488,0.0200794712,-0.2159727663,0.0826531723,-0.0400286093,-0.1227325797,-0.0801262707,0.1514735818,0.1209952906,0.127742365,0.5383259654,0.2677577734,0.410494417,0.1294866651,-0.3016063273,0.1248357072,-0.1209927276,0.3700109422,-0.2481317222,0.0517883822,0.3423662782,0.1794074625,-0.2017464638,0.2401463687,-0.0048973123,-0.070583038,0.3367039561,0.0287895519,-0.1383411288,0.0239062179,0.1935496777,-0.0604689121,0.010493882,-0.0702693909,-0.0987723097,-0.1590381861,-0.3050752282,-0.087373741,0.1904460341,0.2198341787,0.0339709073,-0.3773390353,-0.114849709,-0.3479646742,0.1886437535,-0.1687081754,0.3023189008,0.0322840959,0.2741390169,0.2475180477,0.1615713686,0.9262683988,-0.0916073173,0.0714728311,-0.1972740144,-0.0435676649,-0.5894108415,-0.1542813331,-0.0789639726,0.0270803031,-0.0590391792,0.5538505316,-0.0631092489,-0.0686066449,-0.261759311,0.2991942167,-0.3168808818,-0.3877653778,-0.2755343318,-0.2933425307,-0.1356876642,-0.3935822845,0.0199943967,-0.0869563818,0.0755996183,-0.3193262815,-0.267570585,-0.1990125477,0.0541381128,0.0145461811,0.3740357757,0.3435080349,0.2086753249,-0.0869254544,0.3284698129,0.3662003279,0.2595766485,0.0718757212,0.0568668023,0.2052865177,-0.2167998701,0.0861621499,0.1083574519,-0.0250589009,-0.2877222598,-0.0670191571,0.1119900942,-0.2830277085,0.1024947241,0.3718264997,-0.158607617,-0.0864219889,-0.3929764926,0.3548669219,0.0121597042,0.0940138474,0.0360056087,-0.0519523099,-0.1494732946,0.1480955929,0.2831670642,0.5525882244,-0.3548772931,0.3033214211,0.2974138558,0.0017682018,0.0114030931,-0.0265308805,-0.168065086,-0.1445067227,-0.3122710884,0.0030846142,-0.1878083795,-0.0254937578,0.4849464893,0.1750966311,0.0439129099,0.2067071199,0.0071276291,-0.1149759442,0.5293208361,0.299548775,-0.2283003181,0.0325021446,0.0639349148,-0.254832387,-0.1903491616,-0.0622595958,-0.0863793641,-0.0984942764,0.2215803266,-0.0403396897,-0.2126670629,-0.123895891,0.3679513037,-0.1082212925,-0.2316224575,0.1815731823,0.0838665143,-0.2010200322,0.196623072,-0.157318905,0.1279699206,0.0030061631,0.3249221146,0.1413984448,-0.1144990996,-0.0310456324,-0.0913244933,0.0234271865,-0.1943868995,0.1888038814,-0.1482692063,-0.0629260316,0.0454889238,0.029789513,-0.096110031,-0.653003335,-0.3431927264,-0.2048236728,0.4010295868,0.0805630684,0.1309611499,0.2124831229,0.4965970218,0.0243887007,-0.3730357587,-0.1403264701,0.1895269006,0.2858310342,0.1799467951,0.3866920769,-0.1341089606,-0.1365939081,-0.0709895268,0.6510813832,0.1012729779,-0.1912544817,-0.057765983,-0.0403905585,0.3169389963,0.0208149981,-0.334382683,0.188358143,-0.0939157605,-0.3599133492,-0.2077349275,-0.137304768,0.1214710847,-0.1601488441,0.2943194509,0.0504378565,-0.1864531487,0.0016885888,-0.0764309838,-0.2049330324,0.1969459504,0.2957784832,0.0060840035,-0.3328675926,0.0645710304,-0.0876779109,-0.2682572603,-0.0669379458,-0.1471435428,-0.1714788526,-0.1753690541,0.0967811123,0.1115372702,0.1287063509,-0.2801164687,0.1016181931,-0.3750437796,-0.4788466394,-0.13488096,0.1319559813,-0.0679693297,-0.090987362,0.103313379,0.0220806133,-0.251005441,0.007965601,0.1416076869,0.0354233049,0.2988489568,0.1892450303,0.1612934768,0.1742737293,0.0425288305,-0.4761444926,0.0175630245,-0.0588393733,0.0995729342,0.1874175221,-0.1962634623,0.2503475547,0.2590202689,0.3225476444,0.1073599681,-0.4383759499,-0.2273101062,-0.0798544213,0.1225005761,0.0510145389,-0.1831585467,0.1945354044,0.28687644,-0.0688323528,0.3184705377,0.2352006137,0.3476952314,-0.4201479852,0.0468274653,0.4110737741,0.0671937168,-0.1297736764,0.4717316329,-0.2036245912,0.0523640476,0.0846017823,0.2946646512,0.1201955527,0.8587027788,0.12909168,0.1891805381,0.4139887989,-0.0303379316,0.0068449378,-0.4966722131,-0.0952328369,0.1984754503,-0.168075338,0.2741121948,0.170211941,-0.3590529263,0.1909864247,-0.3852767646,-0.2336922735,-0.0620264858,-0.2622532547,0.1335817128,-0.0030272065,0.0273247976,0.1620105356,0.2371406257,-0.3845196962,-0.3014771342,-0.2461311817,0.1889080554,-0.2023765892,0.2124773413,0.4214303195,0.0444422476,-0.0315243825,-0.0391092412,0.3646275699,0.0986841619,0.1010680571,-0.0881536752,0.0778101683,0.1303085536,0.1942219287,0.1864715815,-0.0245688483,0.229917556,-0.0123417871,-0.0638061985,0.0230799001,-0.1007118225,-0.0150630157,0.1352045685,0.0769988522,-0.0195251871,-0.1023717076,0.4225893915,-0.0051351315,-0.4866001904,-0.0333422124,-0.4977223873,-0.0050530042,-0.051477354,0.0287456643,-0.4116464853,-0.2353772074,0.5690826178,-0.2013360709,0.2559473515,0.0774657801,0.0456615984,-0.6079736948,0.6227496862,0.1627824455,0.0933043957,-0.0685406327,0.1594237983,-0.5500161648,0.0303858668,-0.2166630477,-0.1015652195,0.233997196,0.4140377045,-0.0308639854,-0.0486871824,0.1650849581,0.003842826,-0.1916438937,0.0894423947,-0.2841004431,-0.1390906125,-0.1443412304,0.1185550466,-0.1753247529,-0.4083871543,0.3195689321,-0.0093015609,-0.0767879412,-0.1066462398,0.2173395455,-0.1415829957,0.4878300726,0.4013835788,-0.0483230203,0.3463920951,0.0004342151,0.0175026599,0.0451892316,0.1241305843,-0.0279618651,0.7077245712,0.0045384844,0.1401189715,-0.0876325294,0.0933168158,0.1183361933,0.0912858322,-0.0013901879,-0.1583742946,-0.1358546168,0.0508204177,-0.1527412385,0.0265369955,0.1540860236,0.1232053787,0.0315913707,-0.0814922303,-0.0285235085,-0.0700531751,0.3188904226,0.0257114191,-0.0847275257,0.0533844195,-0.0156111252,-0.0210556425,-0.0187305622,-0.2966986299,0.201208815,0.4706384838,-0.2481852323,-0.0093808491,0.3947989643,-0.2101584971,0.0058913021,-0.0756310448,0.7668739557,-0.1517630368,-0.3578999937,-0.1258624494,-0.0418124832]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2146","title":"Dataset file size on disk is very large with 3D Array","comments":"Interesting !\r\nThis may be because of the offsets that are stored with the array data.\r\n\r\nCurrently the offsets are stored even if the `shape` of the arrays is fixed. This was needed because of some issues with pyarrow a few months ago. I think these issues have been addressed now, so we can probably try to remove them to make the file lighter.\r\n\r\nIdeally in your case the floats data should be 220 MB for both Array2D and Array3D","body":"Hi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n","comment_length":80,"text":"Dataset file size on disk is very large with 3D Array\nHi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n\nInteresting !\r\nThis may be because of the offsets that are stored with the array data.\r\n\r\nCurrently the offsets are stored even if the `shape` of the arrays is fixed. This was needed because of some issues with pyarrow a few months ago. I think these issues have been addressed now, so we can probably try to remove them to make the file lighter.\r\n\r\nIdeally in your case the floats data should be 220 MB for both Array2D and Array3D","embeddings":[-0.1452738494,-0.1132628769,-0.1544179171,0.4201557934,0.2135634124,0.1139131263,0.5066809654,0.2730783522,0.0205232408,0.0358475372,-0.1937043518,0.0874153972,-0.1694584191,0.3150443435,0.0677503124,0.1006881595,-0.0745178536,0.2513983548,-0.1334785819,-0.0857200548,-0.0097005265,0.0191179775,-0.0179097727,-0.1838068217,-0.0615620911,-0.1613830328,0.0913496315,-0.0241849087,-0.0830982178,0.0797271058,0.2449781299,-0.0880117193,0.1781996787,0.5363877416,-0.0001173844,0.0502765737,0.108294405,-0.1655321717,0.0791868642,0.2893508673,-0.0925560445,-0.4095061719,-0.0746671557,-0.3576083779,0.0769588351,-0.3249549866,-0.1521175504,-0.1535450816,0.1848900914,0.0791001096,0.2110456824,0.1533048898,-0.0271132421,0.415045768,0.0993782207,0.5338887572,-0.1533458233,0.1268558055,-0.1435388923,0.4372014105,0.084297426,0.0795503631,0.1613585651,0.2877249718,0.5040504932,-0.055517599,0.2392112762,-0.246191889,0.1298937201,0.4001127183,0.8583656549,-0.2007374316,-0.1842802763,-0.3255720437,-0.1083301753,-0.0323392935,0.0225410033,0.3800633848,-0.049889937,0.021931123,-0.6522023082,-0.1392243952,-0.2269416898,0.1455018222,-0.2136535794,-0.4809675813,-0.1380903423,0.1503390074,-0.1890008897,-0.2658364773,0.2607783675,-0.3955922127,0.0734376535,-0.0997146145,-0.2659658492,-0.2101803869,-0.2905137539,-0.3803426623,0.3530271351,-0.1538060755,-0.0108280405,-0.151993528,-0.364893049,-0.1477414966,0.1039847285,0.3860372305,-0.4768467844,-0.138597995,-0.1418677419,-0.2950907052,-0.0408817492,-0.0522949062,-0.2465380579,-0.0907576829,0.1986332238,-0.4835402966,0.0639369488,-0.0618666336,0.0446808524,0.0238069706,0.0864812881,-0.0952553973,-0.075162895,0.1316531152,0.1410508007,0.135276109,-0.1400613934,-0.039203234,-0.1859151274,0.0416380167,-0.1229669973,-0.1046110019,0.0626311079,-0.081755273,0.1144735962,0.1505685151,0.052291356,-0.1904567331,0.0903884694,0.2001003176,0.2624739408,-0.3473196626,0.4073754847,0.3550085127,0.1589064002,0.1983184069,-0.2854207754,0.1398651004,-0.1997080892,0.4944943488,-0.5467603207,-0.0453600101,-0.4173412025,0.0906217396,-0.2234345526,-0.0543203168,-0.5728366375,0.0097432947,0.3395531476,-0.1186216325,0.0678242967,-0.2173455358,-0.3785401583,-0.294359833,-0.0371542424,0.0312655084,-0.7563121319,0.2612313628,-0.1514395922,0.0817364231,0.1952205747,0.5952161551,-0.0899516866,0.4252856076,-0.4255605638,-0.0484716371,0.0430041216,-0.0384001322,-0.6611084342,0.487341702,0.0706712455,0.0485031083,0.1124144495,0.320182234,0.1636859328,-0.016375415,-0.2228286564,0.3266060352,0.0910482332,0.0426690839,-0.1908326745,-0.3811230659,0.0882390067,0.0459806919,-0.1374954879,-0.1405356973,0.1314450502,0.0834871456,0.2107224613,-0.1292201728,0.0360603482,0.3705856502,0.2096632719,-0.2556730807,-0.0215890892,0.1423658431,-0.6317534447,0.1833813041,0.2472616285,-0.4363888502,0.1557535678,-0.1044431329,0.0382063761,-0.2296464294,-0.0895897746,0.2013011128,0.0237026345,-0.0005763058,-0.0091966204,-0.0104556819,-0.0085912012,-0.147752732,0.1148846447,-0.1881847829,-0.4114223123,0.356153816,0.213565439,-0.0350419208,-0.1056866199,-0.078861475,-0.3569892943,-0.341083467,-0.1865350902,0.0800963417,0.2511079609,0.1212477982,-0.0773310214,0.4837380052,0.2234310806,0.0761163831,0.3378689587,-0.1210732833,0.0945777297,-0.2206157446,-0.2020530403,0.216520682,-0.1918247491,0.0659026727,0.0202879235,-0.3593147099,0.2857810557,0.1333052814,0.1050533727,-0.2576690316,-0.0769441724,0.1929026395,0.3467417359,0.2029425353,0.1076560989,0.0467540659,0.6398187876,-0.1667219102,0.1070185304,0.364521265,-0.1265423596,-0.0783273652,0.1058712453,0.2452514023,0.5257766247,0.217031315,0.045964893,-0.0413736738,0.4925529659,-0.1443800777,0.0811916888,0.0678964555,0.5235673785,0.2810914516,0.1602386236,0.0027567402,-0.284455806,-0.1133067086,0.2043940723,0.2465661764,-0.303352654,0.1999263167,-0.1631348729,-0.2681356966,-0.0948000997,0.1902986467,-0.060111925,0.1248413771,-0.3277061582,-0.1944479644,0.0251019541,-0.0797675923,0.1439219266,0.0847972631,0.2937849462,-0.0679315031,0.2701685727,0.3433182836,-0.0878251344,0.0354879275,0.4142909348,-0.1485609859,0.3000029325,-0.1635025293,0.2096656561,-0.138071239,-0.2516806722,0.1204532534,0.0605554953,0.0706452951,-0.109574303,0.2740660906,-0.1995595098,0.2018909901,0.0286068246,0.0493995771,-0.2603569627,0.0002532348,-0.0481878854,-0.1286520362,-0.2174704075,0.0934461653,-0.2839459479,-0.050991375,0.2078645676,0.1816901416,0.2575588822,0.3705631793,0.1723432094,0.0412281081,0.1493278593,-0.0243310146,-0.1940806955,-0.4591633081,0.1459610909,-0.0282914434,-0.3725241125,0.1432911903,0.0114656771,0.192036286,0.2045514584,-0.6480360627,-0.0394913256,-0.293095082,0.0378666818,-0.17624484,0.1218835488,0.0200794712,-0.2159727663,0.0826531723,-0.0400286093,-0.1227325797,-0.0801262707,0.1514735818,0.1209952906,0.127742365,0.5383259654,0.2677577734,0.410494417,0.1294866651,-0.3016063273,0.1248357072,-0.1209927276,0.3700109422,-0.2481317222,0.0517883822,0.3423662782,0.1794074625,-0.2017464638,0.2401463687,-0.0048973123,-0.070583038,0.3367039561,0.0287895519,-0.1383411288,0.0239062179,0.1935496777,-0.0604689121,0.010493882,-0.0702693909,-0.0987723097,-0.1590381861,-0.3050752282,-0.087373741,0.1904460341,0.2198341787,0.0339709073,-0.3773390353,-0.114849709,-0.3479646742,0.1886437535,-0.1687081754,0.3023189008,0.0322840959,0.2741390169,0.2475180477,0.1615713686,0.9262683988,-0.0916073173,0.0714728311,-0.1972740144,-0.0435676649,-0.5894108415,-0.1542813331,-0.0789639726,0.0270803031,-0.0590391792,0.5538505316,-0.0631092489,-0.0686066449,-0.261759311,0.2991942167,-0.3168808818,-0.3877653778,-0.2755343318,-0.2933425307,-0.1356876642,-0.3935822845,0.0199943967,-0.0869563818,0.0755996183,-0.3193262815,-0.267570585,-0.1990125477,0.0541381128,0.0145461811,0.3740357757,0.3435080349,0.2086753249,-0.0869254544,0.3284698129,0.3662003279,0.2595766485,0.0718757212,0.0568668023,0.2052865177,-0.2167998701,0.0861621499,0.1083574519,-0.0250589009,-0.2877222598,-0.0670191571,0.1119900942,-0.2830277085,0.1024947241,0.3718264997,-0.158607617,-0.0864219889,-0.3929764926,0.3548669219,0.0121597042,0.0940138474,0.0360056087,-0.0519523099,-0.1494732946,0.1480955929,0.2831670642,0.5525882244,-0.3548772931,0.3033214211,0.2974138558,0.0017682018,0.0114030931,-0.0265308805,-0.168065086,-0.1445067227,-0.3122710884,0.0030846142,-0.1878083795,-0.0254937578,0.4849464893,0.1750966311,0.0439129099,0.2067071199,0.0071276291,-0.1149759442,0.5293208361,0.299548775,-0.2283003181,0.0325021446,0.0639349148,-0.254832387,-0.1903491616,-0.0622595958,-0.0863793641,-0.0984942764,0.2215803266,-0.0403396897,-0.2126670629,-0.123895891,0.3679513037,-0.1082212925,-0.2316224575,0.1815731823,0.0838665143,-0.2010200322,0.196623072,-0.157318905,0.1279699206,0.0030061631,0.3249221146,0.1413984448,-0.1144990996,-0.0310456324,-0.0913244933,0.0234271865,-0.1943868995,0.1888038814,-0.1482692063,-0.0629260316,0.0454889238,0.029789513,-0.096110031,-0.653003335,-0.3431927264,-0.2048236728,0.4010295868,0.0805630684,0.1309611499,0.2124831229,0.4965970218,0.0243887007,-0.3730357587,-0.1403264701,0.1895269006,0.2858310342,0.1799467951,0.3866920769,-0.1341089606,-0.1365939081,-0.0709895268,0.6510813832,0.1012729779,-0.1912544817,-0.057765983,-0.0403905585,0.3169389963,0.0208149981,-0.334382683,0.188358143,-0.0939157605,-0.3599133492,-0.2077349275,-0.137304768,0.1214710847,-0.1601488441,0.2943194509,0.0504378565,-0.1864531487,0.0016885888,-0.0764309838,-0.2049330324,0.1969459504,0.2957784832,0.0060840035,-0.3328675926,0.0645710304,-0.0876779109,-0.2682572603,-0.0669379458,-0.1471435428,-0.1714788526,-0.1753690541,0.0967811123,0.1115372702,0.1287063509,-0.2801164687,0.1016181931,-0.3750437796,-0.4788466394,-0.13488096,0.1319559813,-0.0679693297,-0.090987362,0.103313379,0.0220806133,-0.251005441,0.007965601,0.1416076869,0.0354233049,0.2988489568,0.1892450303,0.1612934768,0.1742737293,0.0425288305,-0.4761444926,0.0175630245,-0.0588393733,0.0995729342,0.1874175221,-0.1962634623,0.2503475547,0.2590202689,0.3225476444,0.1073599681,-0.4383759499,-0.2273101062,-0.0798544213,0.1225005761,0.0510145389,-0.1831585467,0.1945354044,0.28687644,-0.0688323528,0.3184705377,0.2352006137,0.3476952314,-0.4201479852,0.0468274653,0.4110737741,0.0671937168,-0.1297736764,0.4717316329,-0.2036245912,0.0523640476,0.0846017823,0.2946646512,0.1201955527,0.8587027788,0.12909168,0.1891805381,0.4139887989,-0.0303379316,0.0068449378,-0.4966722131,-0.0952328369,0.1984754503,-0.168075338,0.2741121948,0.170211941,-0.3590529263,0.1909864247,-0.3852767646,-0.2336922735,-0.0620264858,-0.2622532547,0.1335817128,-0.0030272065,0.0273247976,0.1620105356,0.2371406257,-0.3845196962,-0.3014771342,-0.2461311817,0.1889080554,-0.2023765892,0.2124773413,0.4214303195,0.0444422476,-0.0315243825,-0.0391092412,0.3646275699,0.0986841619,0.1010680571,-0.0881536752,0.0778101683,0.1303085536,0.1942219287,0.1864715815,-0.0245688483,0.229917556,-0.0123417871,-0.0638061985,0.0230799001,-0.1007118225,-0.0150630157,0.1352045685,0.0769988522,-0.0195251871,-0.1023717076,0.4225893915,-0.0051351315,-0.4866001904,-0.0333422124,-0.4977223873,-0.0050530042,-0.051477354,0.0287456643,-0.4116464853,-0.2353772074,0.5690826178,-0.2013360709,0.2559473515,0.0774657801,0.0456615984,-0.6079736948,0.6227496862,0.1627824455,0.0933043957,-0.0685406327,0.1594237983,-0.5500161648,0.0303858668,-0.2166630477,-0.1015652195,0.233997196,0.4140377045,-0.0308639854,-0.0486871824,0.1650849581,0.003842826,-0.1916438937,0.0894423947,-0.2841004431,-0.1390906125,-0.1443412304,0.1185550466,-0.1753247529,-0.4083871543,0.3195689321,-0.0093015609,-0.0767879412,-0.1066462398,0.2173395455,-0.1415829957,0.4878300726,0.4013835788,-0.0483230203,0.3463920951,0.0004342151,0.0175026599,0.0451892316,0.1241305843,-0.0279618651,0.7077245712,0.0045384844,0.1401189715,-0.0876325294,0.0933168158,0.1183361933,0.0912858322,-0.0013901879,-0.1583742946,-0.1358546168,0.0508204177,-0.1527412385,0.0265369955,0.1540860236,0.1232053787,0.0315913707,-0.0814922303,-0.0285235085,-0.0700531751,0.3188904226,0.0257114191,-0.0847275257,0.0533844195,-0.0156111252,-0.0210556425,-0.0187305622,-0.2966986299,0.201208815,0.4706384838,-0.2481852323,-0.0093808491,0.3947989643,-0.2101584971,0.0058913021,-0.0756310448,0.7668739557,-0.1517630368,-0.3578999937,-0.1258624494,-0.0418124832]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2146","title":"Dataset file size on disk is very large with 3D Array","comments":"Yeah for sure, can you be a bit more specific about where the offset is stored in the code base ? And any reference to pyarrow issues if you have some. I would be very interested in contributing to `datasets` by trying to fix this issue. ","body":"Hi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n","comment_length":46,"text":"Dataset file size on disk is very large with 3D Array\nHi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n\nYeah for sure, can you be a bit more specific about where the offset is stored in the code base ? And any reference to pyarrow issues if you have some. I would be very interested in contributing to `datasets` by trying to fix this issue. ","embeddings":[-0.1452738494,-0.1132628769,-0.1544179171,0.4201557934,0.2135634124,0.1139131263,0.5066809654,0.2730783522,0.0205232408,0.0358475372,-0.1937043518,0.0874153972,-0.1694584191,0.3150443435,0.0677503124,0.1006881595,-0.0745178536,0.2513983548,-0.1334785819,-0.0857200548,-0.0097005265,0.0191179775,-0.0179097727,-0.1838068217,-0.0615620911,-0.1613830328,0.0913496315,-0.0241849087,-0.0830982178,0.0797271058,0.2449781299,-0.0880117193,0.1781996787,0.5363877416,-0.0001173844,0.0502765737,0.108294405,-0.1655321717,0.0791868642,0.2893508673,-0.0925560445,-0.4095061719,-0.0746671557,-0.3576083779,0.0769588351,-0.3249549866,-0.1521175504,-0.1535450816,0.1848900914,0.0791001096,0.2110456824,0.1533048898,-0.0271132421,0.415045768,0.0993782207,0.5338887572,-0.1533458233,0.1268558055,-0.1435388923,0.4372014105,0.084297426,0.0795503631,0.1613585651,0.2877249718,0.5040504932,-0.055517599,0.2392112762,-0.246191889,0.1298937201,0.4001127183,0.8583656549,-0.2007374316,-0.1842802763,-0.3255720437,-0.1083301753,-0.0323392935,0.0225410033,0.3800633848,-0.049889937,0.021931123,-0.6522023082,-0.1392243952,-0.2269416898,0.1455018222,-0.2136535794,-0.4809675813,-0.1380903423,0.1503390074,-0.1890008897,-0.2658364773,0.2607783675,-0.3955922127,0.0734376535,-0.0997146145,-0.2659658492,-0.2101803869,-0.2905137539,-0.3803426623,0.3530271351,-0.1538060755,-0.0108280405,-0.151993528,-0.364893049,-0.1477414966,0.1039847285,0.3860372305,-0.4768467844,-0.138597995,-0.1418677419,-0.2950907052,-0.0408817492,-0.0522949062,-0.2465380579,-0.0907576829,0.1986332238,-0.4835402966,0.0639369488,-0.0618666336,0.0446808524,0.0238069706,0.0864812881,-0.0952553973,-0.075162895,0.1316531152,0.1410508007,0.135276109,-0.1400613934,-0.039203234,-0.1859151274,0.0416380167,-0.1229669973,-0.1046110019,0.0626311079,-0.081755273,0.1144735962,0.1505685151,0.052291356,-0.1904567331,0.0903884694,0.2001003176,0.2624739408,-0.3473196626,0.4073754847,0.3550085127,0.1589064002,0.1983184069,-0.2854207754,0.1398651004,-0.1997080892,0.4944943488,-0.5467603207,-0.0453600101,-0.4173412025,0.0906217396,-0.2234345526,-0.0543203168,-0.5728366375,0.0097432947,0.3395531476,-0.1186216325,0.0678242967,-0.2173455358,-0.3785401583,-0.294359833,-0.0371542424,0.0312655084,-0.7563121319,0.2612313628,-0.1514395922,0.0817364231,0.1952205747,0.5952161551,-0.0899516866,0.4252856076,-0.4255605638,-0.0484716371,0.0430041216,-0.0384001322,-0.6611084342,0.487341702,0.0706712455,0.0485031083,0.1124144495,0.320182234,0.1636859328,-0.016375415,-0.2228286564,0.3266060352,0.0910482332,0.0426690839,-0.1908326745,-0.3811230659,0.0882390067,0.0459806919,-0.1374954879,-0.1405356973,0.1314450502,0.0834871456,0.2107224613,-0.1292201728,0.0360603482,0.3705856502,0.2096632719,-0.2556730807,-0.0215890892,0.1423658431,-0.6317534447,0.1833813041,0.2472616285,-0.4363888502,0.1557535678,-0.1044431329,0.0382063761,-0.2296464294,-0.0895897746,0.2013011128,0.0237026345,-0.0005763058,-0.0091966204,-0.0104556819,-0.0085912012,-0.147752732,0.1148846447,-0.1881847829,-0.4114223123,0.356153816,0.213565439,-0.0350419208,-0.1056866199,-0.078861475,-0.3569892943,-0.341083467,-0.1865350902,0.0800963417,0.2511079609,0.1212477982,-0.0773310214,0.4837380052,0.2234310806,0.0761163831,0.3378689587,-0.1210732833,0.0945777297,-0.2206157446,-0.2020530403,0.216520682,-0.1918247491,0.0659026727,0.0202879235,-0.3593147099,0.2857810557,0.1333052814,0.1050533727,-0.2576690316,-0.0769441724,0.1929026395,0.3467417359,0.2029425353,0.1076560989,0.0467540659,0.6398187876,-0.1667219102,0.1070185304,0.364521265,-0.1265423596,-0.0783273652,0.1058712453,0.2452514023,0.5257766247,0.217031315,0.045964893,-0.0413736738,0.4925529659,-0.1443800777,0.0811916888,0.0678964555,0.5235673785,0.2810914516,0.1602386236,0.0027567402,-0.284455806,-0.1133067086,0.2043940723,0.2465661764,-0.303352654,0.1999263167,-0.1631348729,-0.2681356966,-0.0948000997,0.1902986467,-0.060111925,0.1248413771,-0.3277061582,-0.1944479644,0.0251019541,-0.0797675923,0.1439219266,0.0847972631,0.2937849462,-0.0679315031,0.2701685727,0.3433182836,-0.0878251344,0.0354879275,0.4142909348,-0.1485609859,0.3000029325,-0.1635025293,0.2096656561,-0.138071239,-0.2516806722,0.1204532534,0.0605554953,0.0706452951,-0.109574303,0.2740660906,-0.1995595098,0.2018909901,0.0286068246,0.0493995771,-0.2603569627,0.0002532348,-0.0481878854,-0.1286520362,-0.2174704075,0.0934461653,-0.2839459479,-0.050991375,0.2078645676,0.1816901416,0.2575588822,0.3705631793,0.1723432094,0.0412281081,0.1493278593,-0.0243310146,-0.1940806955,-0.4591633081,0.1459610909,-0.0282914434,-0.3725241125,0.1432911903,0.0114656771,0.192036286,0.2045514584,-0.6480360627,-0.0394913256,-0.293095082,0.0378666818,-0.17624484,0.1218835488,0.0200794712,-0.2159727663,0.0826531723,-0.0400286093,-0.1227325797,-0.0801262707,0.1514735818,0.1209952906,0.127742365,0.5383259654,0.2677577734,0.410494417,0.1294866651,-0.3016063273,0.1248357072,-0.1209927276,0.3700109422,-0.2481317222,0.0517883822,0.3423662782,0.1794074625,-0.2017464638,0.2401463687,-0.0048973123,-0.070583038,0.3367039561,0.0287895519,-0.1383411288,0.0239062179,0.1935496777,-0.0604689121,0.010493882,-0.0702693909,-0.0987723097,-0.1590381861,-0.3050752282,-0.087373741,0.1904460341,0.2198341787,0.0339709073,-0.3773390353,-0.114849709,-0.3479646742,0.1886437535,-0.1687081754,0.3023189008,0.0322840959,0.2741390169,0.2475180477,0.1615713686,0.9262683988,-0.0916073173,0.0714728311,-0.1972740144,-0.0435676649,-0.5894108415,-0.1542813331,-0.0789639726,0.0270803031,-0.0590391792,0.5538505316,-0.0631092489,-0.0686066449,-0.261759311,0.2991942167,-0.3168808818,-0.3877653778,-0.2755343318,-0.2933425307,-0.1356876642,-0.3935822845,0.0199943967,-0.0869563818,0.0755996183,-0.3193262815,-0.267570585,-0.1990125477,0.0541381128,0.0145461811,0.3740357757,0.3435080349,0.2086753249,-0.0869254544,0.3284698129,0.3662003279,0.2595766485,0.0718757212,0.0568668023,0.2052865177,-0.2167998701,0.0861621499,0.1083574519,-0.0250589009,-0.2877222598,-0.0670191571,0.1119900942,-0.2830277085,0.1024947241,0.3718264997,-0.158607617,-0.0864219889,-0.3929764926,0.3548669219,0.0121597042,0.0940138474,0.0360056087,-0.0519523099,-0.1494732946,0.1480955929,0.2831670642,0.5525882244,-0.3548772931,0.3033214211,0.2974138558,0.0017682018,0.0114030931,-0.0265308805,-0.168065086,-0.1445067227,-0.3122710884,0.0030846142,-0.1878083795,-0.0254937578,0.4849464893,0.1750966311,0.0439129099,0.2067071199,0.0071276291,-0.1149759442,0.5293208361,0.299548775,-0.2283003181,0.0325021446,0.0639349148,-0.254832387,-0.1903491616,-0.0622595958,-0.0863793641,-0.0984942764,0.2215803266,-0.0403396897,-0.2126670629,-0.123895891,0.3679513037,-0.1082212925,-0.2316224575,0.1815731823,0.0838665143,-0.2010200322,0.196623072,-0.157318905,0.1279699206,0.0030061631,0.3249221146,0.1413984448,-0.1144990996,-0.0310456324,-0.0913244933,0.0234271865,-0.1943868995,0.1888038814,-0.1482692063,-0.0629260316,0.0454889238,0.029789513,-0.096110031,-0.653003335,-0.3431927264,-0.2048236728,0.4010295868,0.0805630684,0.1309611499,0.2124831229,0.4965970218,0.0243887007,-0.3730357587,-0.1403264701,0.1895269006,0.2858310342,0.1799467951,0.3866920769,-0.1341089606,-0.1365939081,-0.0709895268,0.6510813832,0.1012729779,-0.1912544817,-0.057765983,-0.0403905585,0.3169389963,0.0208149981,-0.334382683,0.188358143,-0.0939157605,-0.3599133492,-0.2077349275,-0.137304768,0.1214710847,-0.1601488441,0.2943194509,0.0504378565,-0.1864531487,0.0016885888,-0.0764309838,-0.2049330324,0.1969459504,0.2957784832,0.0060840035,-0.3328675926,0.0645710304,-0.0876779109,-0.2682572603,-0.0669379458,-0.1471435428,-0.1714788526,-0.1753690541,0.0967811123,0.1115372702,0.1287063509,-0.2801164687,0.1016181931,-0.3750437796,-0.4788466394,-0.13488096,0.1319559813,-0.0679693297,-0.090987362,0.103313379,0.0220806133,-0.251005441,0.007965601,0.1416076869,0.0354233049,0.2988489568,0.1892450303,0.1612934768,0.1742737293,0.0425288305,-0.4761444926,0.0175630245,-0.0588393733,0.0995729342,0.1874175221,-0.1962634623,0.2503475547,0.2590202689,0.3225476444,0.1073599681,-0.4383759499,-0.2273101062,-0.0798544213,0.1225005761,0.0510145389,-0.1831585467,0.1945354044,0.28687644,-0.0688323528,0.3184705377,0.2352006137,0.3476952314,-0.4201479852,0.0468274653,0.4110737741,0.0671937168,-0.1297736764,0.4717316329,-0.2036245912,0.0523640476,0.0846017823,0.2946646512,0.1201955527,0.8587027788,0.12909168,0.1891805381,0.4139887989,-0.0303379316,0.0068449378,-0.4966722131,-0.0952328369,0.1984754503,-0.168075338,0.2741121948,0.170211941,-0.3590529263,0.1909864247,-0.3852767646,-0.2336922735,-0.0620264858,-0.2622532547,0.1335817128,-0.0030272065,0.0273247976,0.1620105356,0.2371406257,-0.3845196962,-0.3014771342,-0.2461311817,0.1889080554,-0.2023765892,0.2124773413,0.4214303195,0.0444422476,-0.0315243825,-0.0391092412,0.3646275699,0.0986841619,0.1010680571,-0.0881536752,0.0778101683,0.1303085536,0.1942219287,0.1864715815,-0.0245688483,0.229917556,-0.0123417871,-0.0638061985,0.0230799001,-0.1007118225,-0.0150630157,0.1352045685,0.0769988522,-0.0195251871,-0.1023717076,0.4225893915,-0.0051351315,-0.4866001904,-0.0333422124,-0.4977223873,-0.0050530042,-0.051477354,0.0287456643,-0.4116464853,-0.2353772074,0.5690826178,-0.2013360709,0.2559473515,0.0774657801,0.0456615984,-0.6079736948,0.6227496862,0.1627824455,0.0933043957,-0.0685406327,0.1594237983,-0.5500161648,0.0303858668,-0.2166630477,-0.1015652195,0.233997196,0.4140377045,-0.0308639854,-0.0486871824,0.1650849581,0.003842826,-0.1916438937,0.0894423947,-0.2841004431,-0.1390906125,-0.1443412304,0.1185550466,-0.1753247529,-0.4083871543,0.3195689321,-0.0093015609,-0.0767879412,-0.1066462398,0.2173395455,-0.1415829957,0.4878300726,0.4013835788,-0.0483230203,0.3463920951,0.0004342151,0.0175026599,0.0451892316,0.1241305843,-0.0279618651,0.7077245712,0.0045384844,0.1401189715,-0.0876325294,0.0933168158,0.1183361933,0.0912858322,-0.0013901879,-0.1583742946,-0.1358546168,0.0508204177,-0.1527412385,0.0265369955,0.1540860236,0.1232053787,0.0315913707,-0.0814922303,-0.0285235085,-0.0700531751,0.3188904226,0.0257114191,-0.0847275257,0.0533844195,-0.0156111252,-0.0210556425,-0.0187305622,-0.2966986299,0.201208815,0.4706384838,-0.2481852323,-0.0093808491,0.3947989643,-0.2101584971,0.0058913021,-0.0756310448,0.7668739557,-0.1517630368,-0.3578999937,-0.1258624494,-0.0418124832]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2146","title":"Dataset file size on disk is very large with 3D Array","comments":"Pyarrow has two types of lists: variable length lists and fixed size lists.\r\nCurrently we store the ArrayXD data as variable length lists. They take more disk space because they must store both actual data and offsets.\r\nIn the `datasets` code this is done here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/dbac87c8a083f806467f5afc4ec9b401a7e4c15c\/src\/datasets\/features.py#L346-L352\r\n\r\nTo use a fixed length list, one should use the `list_size` argument of `pyarrow.list_()`.\r\nI believe this would work directly modulo some changes in the numpy conversion here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/dbac87c8a083f806467f5afc4ec9b401a7e4c15c\/src\/datasets\/features.py#L381-L395","body":"Hi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n","comment_length":75,"text":"Dataset file size on disk is very large with 3D Array\nHi, \r\n\r\nI have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. \r\n\r\nThe actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. \r\n\r\n`{\r\n \"description\": \"\",\r\n \"citation\": \"\",\r\n \"homepage\": \"\",\r\n \"license\": \"\",\r\n \"features\": {\r\n \"image\": {\r\n \"shape\": [224, 224, 3],\r\n \"dtype\": \"uint8\",\r\n \"id\": null,\r\n \"_type\": \"Array3D\",\r\n }\r\n },\r\n \"post_processed\": null,\r\n \"supervised_keys\": null,\r\n \"builder_name\": \"shot_type_image_dataset\",\r\n \"config_name\": \"default\",\r\n \"version\": {\r\n \"version_str\": \"0.0.0\",\r\n \"description\": null,\r\n \"major\": 0,\r\n \"minor\": 0,\r\n \"patch\": 0,\r\n },\r\n \"splits\": {\r\n \"train\": {\r\n \"name\": \"train\",\r\n \"num_bytes\": 520803408,\r\n \"num_examples\": 1479,\r\n \"dataset_name\": \"shot_type_image_dataset\",\r\n }\r\n },\r\n \"download_checksums\": {\r\n \"\": {\r\n \"num_bytes\": 16940447118,\r\n \"checksum\": \"5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03\",\r\n }\r\n },\r\n \"download_size\": 16940447118,\r\n \"post_processing_size\": null,\r\n \"dataset_size\": 520803408,\r\n \"size_in_bytes\": 17461250526,\r\n}`\r\n\r\nI have created the same dataset with tensorflow_dataset and it takes only 125MB on disk.\r\n\r\nI am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records.\r\n\r\nThis might be a problem for large dataset. \r\n\r\nThanks for your help. \r\n\nPyarrow has two types of lists: variable length lists and fixed size lists.\r\nCurrently we store the ArrayXD data as variable length lists. They take more disk space because they must store both actual data and offsets.\r\nIn the `datasets` code this is done here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/dbac87c8a083f806467f5afc4ec9b401a7e4c15c\/src\/datasets\/features.py#L346-L352\r\n\r\nTo use a fixed length list, one should use the `list_size` argument of `pyarrow.list_()`.\r\nI believe this would work directly modulo some changes in the numpy conversion here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/dbac87c8a083f806467f5afc4ec9b401a7e4c15c\/src\/datasets\/features.py#L381-L395","embeddings":[-0.1452738494,-0.1132628769,-0.1544179171,0.4201557934,0.2135634124,0.1139131263,0.5066809654,0.2730783522,0.0205232408,0.0358475372,-0.1937043518,0.0874153972,-0.1694584191,0.3150443435,0.0677503124,0.1006881595,-0.0745178536,0.2513983548,-0.1334785819,-0.0857200548,-0.0097005265,0.0191179775,-0.0179097727,-0.1838068217,-0.0615620911,-0.1613830328,0.0913496315,-0.0241849087,-0.0830982178,0.0797271058,0.2449781299,-0.0880117193,0.1781996787,0.5363877416,-0.0001173844,0.0502765737,0.108294405,-0.1655321717,0.0791868642,0.2893508673,-0.0925560445,-0.4095061719,-0.0746671557,-0.3576083779,0.0769588351,-0.3249549866,-0.1521175504,-0.1535450816,0.1848900914,0.0791001096,0.2110456824,0.1533048898,-0.0271132421,0.415045768,0.0993782207,0.5338887572,-0.1533458233,0.1268558055,-0.1435388923,0.4372014105,0.084297426,0.0795503631,0.1613585651,0.2877249718,0.5040504932,-0.055517599,0.2392112762,-0.246191889,0.1298937201,0.4001127183,0.8583656549,-0.2007374316,-0.1842802763,-0.3255720437,-0.1083301753,-0.0323392935,0.0225410033,0.3800633848,-0.049889937,0.021931123,-0.6522023082,-0.1392243952,-0.2269416898,0.1455018222,-0.2136535794,-0.4809675813,-0.1380903423,0.1503390074,-0.1890008897,-0.2658364773,0.2607783675,-0.3955922127,0.0734376535,-0.0997146145,-0.2659658492,-0.2101803869,-0.2905137539,-0.3803426623,0.3530271351,-0.1538060755,-0.0108280405,-0.151993528,-0.364893049,-0.1477414966,0.1039847285,0.3860372305,-0.4768467844,-0.138597995,-0.1418677419,-0.2950907052,-0.0408817492,-0.0522949062,-0.2465380579,-0.0907576829,0.1986332238,-0.4835402966,0.0639369488,-0.0618666336,0.0446808524,0.0238069706,0.0864812881,-0.0952553973,-0.075162895,0.1316531152,0.1410508007,0.135276109,-0.1400613934,-0.039203234,-0.1859151274,0.0416380167,-0.1229669973,-0.1046110019,0.0626311079,-0.081755273,0.1144735962,0.1505685151,0.052291356,-0.1904567331,0.0903884694,0.2001003176,0.2624739408,-0.3473196626,0.4073754847,0.3550085127,0.1589064002,0.1983184069,-0.2854207754,0.1398651004,-0.1997080892,0.4944943488,-0.5467603207,-0.0453600101,-0.4173412025,0.0906217396,-0.2234345526,-0.0543203168,-0.5728366375,0.0097432947,0.3395531476,-0.1186216325,0.0678242967,-0.2173455358,-0.3785401583,-0.294359833,-0.0371542424,0.0312655084,-0.7563121319,0.2612313628,-0.1514395922,0.0817364231,0.1952205747,0.5952161551,-0.0899516866,0.4252856076,-0.4255605638,-0.0484716371,0.0430041216,-0.0384001322,-0.6611084342,0.487341702,0.0706712455,0.0485031083,0.1124144495,0.320182234,0.1636859328,-0.016375415,-0.2228286564,0.3266060352,0.0910482332,0.0426690839,-0.1908326745,-0.3811230659,0.0882390067,0.0459806919,-0.1374954879,-0.1405356973,0.1314450502,0.0834871456,0.2107224613,-0.1292201728,0.0360603482,0.3705856502,0.2096632719,-0.2556730807,-0.0215890892,0.1423658431,-0.6317534447,0.1833813041,0.2472616285,-0.4363888502,0.1557535678,-0.1044431329,0.0382063761,-0.2296464294,-0.0895897746,0.2013011128,0.0237026345,-0.0005763058,-0.0091966204,-0.0104556819,-0.0085912012,-0.147752732,0.1148846447,-0.1881847829,-0.4114223123,0.356153816,0.213565439,-0.0350419208,-0.1056866199,-0.078861475,-0.3569892943,-0.341083467,-0.1865350902,0.0800963417,0.2511079609,0.1212477982,-0.0773310214,0.4837380052,0.2234310806,0.0761163831,0.3378689587,-0.1210732833,0.0945777297,-0.2206157446,-0.2020530403,0.216520682,-0.1918247491,0.0659026727,0.0202879235,-0.3593147099,0.2857810557,0.1333052814,0.1050533727,-0.2576690316,-0.0769441724,0.1929026395,0.3467417359,0.2029425353,0.1076560989,0.0467540659,0.6398187876,-0.1667219102,0.1070185304,0.364521265,-0.1265423596,-0.0783273652,0.1058712453,0.2452514023,0.5257766247,0.217031315,0.045964893,-0.0413736738,0.4925529659,-0.1443800777,0.0811916888,0.0678964555,0.5235673785,0.2810914516,0.1602386236,0.0027567402,-0.284455806,-0.1133067086,0.2043940723,0.2465661764,-0.303352654,0.1999263167,-0.1631348729,-0.2681356966,-0.0948000997,0.1902986467,-0.060111925,0.1248413771,-0.3277061582,-0.1944479644,0.0251019541,-0.0797675923,0.1439219266,0.0847972631,0.2937849462,-0.0679315031,0.2701685727,0.3433182836,-0.0878251344,0.0354879275,0.4142909348,-0.1485609859,0.3000029325,-0.1635025293,0.2096656561,-0.138071239,-0.2516806722,0.1204532534,0.0605554953,0.0706452951,-0.109574303,0.2740660906,-0.1995595098,0.2018909901,0.0286068246,0.0493995771,-0.2603569627,0.0002532348,-0.0481878854,-0.1286520362,-0.2174704075,0.0934461653,-0.2839459479,-0.050991375,0.2078645676,0.1816901416,0.2575588822,0.3705631793,0.1723432094,0.0412281081,0.1493278593,-0.0243310146,-0.1940806955,-0.4591633081,0.1459610909,-0.0282914434,-0.3725241125,0.1432911903,0.0114656771,0.192036286,0.2045514584,-0.6480360627,-0.0394913256,-0.293095082,0.0378666818,-0.17624484,0.1218835488,0.0200794712,-0.2159727663,0.0826531723,-0.0400286093,-0.1227325797,-0.0801262707,0.1514735818,0.1209952906,0.127742365,0.5383259654,0.2677577734,0.410494417,0.1294866651,-0.3016063273,0.1248357072,-0.1209927276,0.3700109422,-0.2481317222,0.0517883822,0.3423662782,0.1794074625,-0.2017464638,0.2401463687,-0.0048973123,-0.070583038,0.3367039561,0.0287895519,-0.1383411288,0.0239062179,0.1935496777,-0.0604689121,0.010493882,-0.0702693909,-0.0987723097,-0.1590381861,-0.3050752282,-0.087373741,0.1904460341,0.2198341787,0.0339709073,-0.3773390353,-0.114849709,-0.3479646742,0.1886437535,-0.1687081754,0.3023189008,0.0322840959,0.2741390169,0.2475180477,0.1615713686,0.9262683988,-0.0916073173,0.0714728311,-0.1972740144,-0.0435676649,-0.5894108415,-0.1542813331,-0.0789639726,0.0270803031,-0.0590391792,0.5538505316,-0.0631092489,-0.0686066449,-0.261759311,0.2991942167,-0.3168808818,-0.3877653778,-0.2755343318,-0.2933425307,-0.1356876642,-0.3935822845,0.0199943967,-0.0869563818,0.0755996183,-0.3193262815,-0.267570585,-0.1990125477,0.0541381128,0.0145461811,0.3740357757,0.3435080349,0.2086753249,-0.0869254544,0.3284698129,0.3662003279,0.2595766485,0.0718757212,0.0568668023,0.2052865177,-0.2167998701,0.0861621499,0.1083574519,-0.0250589009,-0.2877222598,-0.0670191571,0.1119900942,-0.2830277085,0.1024947241,0.3718264997,-0.158607617,-0.0864219889,-0.3929764926,0.3548669219,0.0121597042,0.0940138474,0.0360056087,-0.0519523099,-0.1494732946,0.1480955929,0.2831670642,0.5525882244,-0.3548772931,0.3033214211,0.2974138558,0.0017682018,0.0114030931,-0.0265308805,-0.168065086,-0.1445067227,-0.3122710884,0.0030846142,-0.1878083795,-0.0254937578,0.4849464893,0.1750966311,0.0439129099,0.2067071199,0.0071276291,-0.1149759442,0.5293208361,0.299548775,-0.2283003181,0.0325021446,0.0639349148,-0.254832387,-0.1903491616,-0.0622595958,-0.0863793641,-0.0984942764,0.2215803266,-0.0403396897,-0.2126670629,-0.123895891,0.3679513037,-0.1082212925,-0.2316224575,0.1815731823,0.0838665143,-0.2010200322,0.196623072,-0.157318905,0.1279699206,0.0030061631,0.3249221146,0.1413984448,-0.1144990996,-0.0310456324,-0.0913244933,0.0234271865,-0.1943868995,0.1888038814,-0.1482692063,-0.0629260316,0.0454889238,0.029789513,-0.096110031,-0.653003335,-0.3431927264,-0.2048236728,0.4010295868,0.0805630684,0.1309611499,0.2124831229,0.4965970218,0.0243887007,-0.3730357587,-0.1403264701,0.1895269006,0.2858310342,0.1799467951,0.3866920769,-0.1341089606,-0.1365939081,-0.0709895268,0.6510813832,0.1012729779,-0.1912544817,-0.057765983,-0.0403905585,0.3169389963,0.0208149981,-0.334382683,0.188358143,-0.0939157605,-0.3599133492,-0.2077349275,-0.137304768,0.1214710847,-0.1601488441,0.2943194509,0.0504378565,-0.1864531487,0.0016885888,-0.0764309838,-0.2049330324,0.1969459504,0.2957784832,0.0060840035,-0.3328675926,0.0645710304,-0.0876779109,-0.2682572603,-0.0669379458,-0.1471435428,-0.1714788526,-0.1753690541,0.0967811123,0.1115372702,0.1287063509,-0.2801164687,0.1016181931,-0.3750437796,-0.4788466394,-0.13488096,0.1319559813,-0.0679693297,-0.090987362,0.103313379,0.0220806133,-0.251005441,0.007965601,0.1416076869,0.0354233049,0.2988489568,0.1892450303,0.1612934768,0.1742737293,0.0425288305,-0.4761444926,0.0175630245,-0.0588393733,0.0995729342,0.1874175221,-0.1962634623,0.2503475547,0.2590202689,0.3225476444,0.1073599681,-0.4383759499,-0.2273101062,-0.0798544213,0.1225005761,0.0510145389,-0.1831585467,0.1945354044,0.28687644,-0.0688323528,0.3184705377,0.2352006137,0.3476952314,-0.4201479852,0.0468274653,0.4110737741,0.0671937168,-0.1297736764,0.4717316329,-0.2036245912,0.0523640476,0.0846017823,0.2946646512,0.1201955527,0.8587027788,0.12909168,0.1891805381,0.4139887989,-0.0303379316,0.0068449378,-0.4966722131,-0.0952328369,0.1984754503,-0.168075338,0.2741121948,0.170211941,-0.3590529263,0.1909864247,-0.3852767646,-0.2336922735,-0.0620264858,-0.2622532547,0.1335817128,-0.0030272065,0.0273247976,0.1620105356,0.2371406257,-0.3845196962,-0.3014771342,-0.2461311817,0.1889080554,-0.2023765892,0.2124773413,0.4214303195,0.0444422476,-0.0315243825,-0.0391092412,0.3646275699,0.0986841619,0.1010680571,-0.0881536752,0.0778101683,0.1303085536,0.1942219287,0.1864715815,-0.0245688483,0.229917556,-0.0123417871,-0.0638061985,0.0230799001,-0.1007118225,-0.0150630157,0.1352045685,0.0769988522,-0.0195251871,-0.1023717076,0.4225893915,-0.0051351315,-0.4866001904,-0.0333422124,-0.4977223873,-0.0050530042,-0.051477354,0.0287456643,-0.4116464853,-0.2353772074,0.5690826178,-0.2013360709,0.2559473515,0.0774657801,0.0456615984,-0.6079736948,0.6227496862,0.1627824455,0.0933043957,-0.0685406327,0.1594237983,-0.5500161648,0.0303858668,-0.2166630477,-0.1015652195,0.233997196,0.4140377045,-0.0308639854,-0.0486871824,0.1650849581,0.003842826,-0.1916438937,0.0894423947,-0.2841004431,-0.1390906125,-0.1443412304,0.1185550466,-0.1753247529,-0.4083871543,0.3195689321,-0.0093015609,-0.0767879412,-0.1066462398,0.2173395455,-0.1415829957,0.4878300726,0.4013835788,-0.0483230203,0.3463920951,0.0004342151,0.0175026599,0.0451892316,0.1241305843,-0.0279618651,0.7077245712,0.0045384844,0.1401189715,-0.0876325294,0.0933168158,0.1183361933,0.0912858322,-0.0013901879,-0.1583742946,-0.1358546168,0.0508204177,-0.1527412385,0.0265369955,0.1540860236,0.1232053787,0.0315913707,-0.0814922303,-0.0285235085,-0.0700531751,0.3188904226,0.0257114191,-0.0847275257,0.0533844195,-0.0156111252,-0.0210556425,-0.0187305622,-0.2966986299,0.201208815,0.4706384838,-0.2481852323,-0.0093808491,0.3947989643,-0.2101584971,0.0058913021,-0.0756310448,0.7668739557,-0.1517630368,-0.3578999937,-0.1258624494,-0.0418124832]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2144","title":"Loading wikipedia 20200501.en throws pyarrow related error","comments":"That's how I loaded the dataset\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n```","body":"**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n","comment_length":17,"text":"Loading wikipedia 20200501.en throws pyarrow related error\n**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n\nThat's how I loaded the dataset\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n```","embeddings":[-0.1041540205,0.3526462018,0.0340582058,0.3633103967,0.2745778561,0.1802078485,0.2762243748,0.4623250365,-0.04348813,-0.113784872,0.0247374158,0.236021921,0.1520678252,-0.1178016663,0.1841091067,-0.1683138758,0.0038767338,0.0068803728,0.0847920552,0.1328444481,-0.2001205236,0.1146743447,-0.1824860871,0.1555170566,-0.5857800245,0.0881166533,-0.0660176575,0.0540414415,0.0833243653,-0.613804698,0.3904819489,-0.0500538088,0.2010909319,0.2121308595,-0.0001204597,0.0749956965,0.4643647373,-0.0041334573,-0.2960134745,-0.0811445788,0.186580047,-0.0205053128,0.3180885315,-0.3607697487,0.1008782089,-0.2803391814,0.2242598385,0.0471996665,0.1832180023,0.0086322986,0.168285802,-0.1830526292,0.3375393748,0.0306443255,0.7112432122,-0.1374686658,0.0815054998,0.4424607456,0.1225971282,-0.1607152224,-0.186434865,0.23629798,-0.2502577007,0.1203090623,0.2671839893,-0.0770430341,0.3848769963,-0.2004953474,0.3921535015,0.2679577172,0.7762554288,-0.4092206359,0.1730330735,-0.2313123494,-0.0328712203,-0.0229594614,0.0938252583,0.2307374179,-0.27676934,0.0545302071,0.1571857184,-0.1188708618,-0.2609159052,0.301315248,-0.0642476678,0.1997214258,0.1399375498,0.3910686672,0.0163352489,-0.2268839478,0.0019353561,0.0632643551,0.1672586203,0.2208643556,-0.3931137919,0.107257776,-0.0683028772,0.1139047593,0.3778621256,-0.1693618745,-0.021596102,0.0659415349,0.2391710877,0.2064648122,0.4192265868,0.1241850182,-0.3237407506,-0.2987220883,0.2721662819,0.1988998652,-0.254214257,-0.0364414603,-0.0612123869,-0.1137952879,0.1385727972,-0.1709839702,0.4291738868,-0.2699913085,-0.220579192,0.1195769534,-0.1446794569,-0.0715681612,-0.1552301943,0.3867201805,-0.0916787088,0.2424993664,0.0672103316,0.0622379445,-0.1821437031,-0.1878142953,0.0057542627,0.1186774671,-0.2124670595,0.1797935814,0.2939305305,-0.1538039595,0.2583542168,0.1435221881,-0.1506101638,-0.1206758171,0.1700648069,0.0130382087,-0.2326727659,0.309833169,0.0075005512,0.2468717843,-0.0328951664,-0.1449694484,-0.1790345013,0.3182585537,-0.2848282754,-0.5152549744,-0.262249738,0.1209049523,-0.0116632255,0.0844510719,0.1616870016,0.0576540641,0.4413489699,-0.5136592984,-0.1003230512,-0.1257850081,-0.1855283529,-0.4527531266,0.1706373543,0.3363130093,-0.5569394231,0.0798316821,-0.0507644974,0.0824038908,0.2173671722,0.4450024068,-0.2298555672,0.1757664233,-0.1654862761,0.1924015433,0.1095912755,-0.3121576309,-0.5411019325,0.2140131146,-0.0612866841,0.1959047168,0.0259571318,0.2906033099,0.0093369419,0.1717624664,0.4052648544,0.2492511719,-0.0467647091,0.0787921399,-0.4085267782,-0.3879698217,0.3385632038,0.2016654164,-0.0047450308,-0.1988982856,0.1968246847,0.3221473992,0.2584798336,-0.2035795897,0.1287960708,0.0218389891,0.0705177039,-0.0129466308,0.2327663898,-0.2238298804,-0.4262207448,0.11144609,-0.3431405723,-0.0235893708,-0.3043866754,-0.361228615,-0.5231873989,0.0663694814,-0.171464175,-0.2400977015,0.115050301,-0.0613334365,0.0953495279,0.2569484115,0.1027859375,-0.0896757692,-0.2042666227,-0.0175576173,-0.4030902386,0.240516305,-0.308798492,-0.1786494255,-0.0221906919,0.2451284826,0.324608773,0.1002937332,0.0431751907,0.1558428705,0.1467771679,0.0931873098,-0.1324361414,-0.0177001171,0.2272260636,-0.2668777108,0.2040083557,0.1240440756,0.1868998408,-0.221017614,0.05095274,0.1322418749,-0.0019998513,0.1718758643,-0.0822376162,-0.0712248981,0.1570295691,-0.0034186426,0.3967843354,-0.1355208606,0.2599032819,0.4711607099,0.0430019759,0.1037084162,-0.0057747457,0.0103615662,0.1143461317,-0.0471609905,0.1531215012,0.1461279243,-0.3416619301,-0.2851566672,0.2188828588,-0.1442351341,0.0782735273,0.0937128961,0.0280694515,0.1398488432,-0.0689809099,-0.195103392,0.1316027343,-0.1178800836,0.4590720534,0.310067445,0.3028900325,-0.029343931,-0.3765672147,0.0593944713,0.0831138715,0.379365176,-0.3040897548,0.0794889852,-0.171332255,-0.243459627,-0.0772669688,-0.013044945,-0.2915118039,-0.317315042,0.0815705061,0.5144311786,0.0328679904,0.0099860048,-0.0540035702,-0.0480350479,0.0796879381,-0.1308909357,-0.0505061187,-0.7057309747,-0.3859112859,0.0051475652,0.2574861944,-0.1339800954,0.2502458096,-0.0254044235,-0.1927474737,-0.3766916692,-0.5513484478,-0.0036723248,-0.1104055196,0.3178443015,0.0176331513,0.3957713842,0.0727890357,-0.2152879387,0.2103869319,0.0130584994,-0.0248649139,0.2813790739,-0.3631691337,0.0844142213,-0.0089278743,-0.2284815609,-0.2832888961,-0.1489331275,0.1993967593,0.0630166978,0.1870315522,0.2383005023,0.0628918335,0.1012383327,0.2758947015,0.0676462352,-0.3717182577,-0.1608327031,0.3060976565,-0.2184316963,-0.2773063481,0.1912308037,0.002722593,0.1241865531,0.2216507792,-0.5899142623,0.0072065173,0.0857219473,0.3271362782,0.031964194,-0.1462098956,0.2367205918,0.0211273748,0.0012416313,0.062059287,0.0056531909,-0.1206239983,-0.1537077129,0.166241616,0.1140279919,0.2094682604,0.1647175252,0.7836135626,0.0867718309,0.0424960814,0.4263213873,-0.188306883,0.0635278597,-0.1462810636,-0.0966464877,0.0314848572,-0.3452987075,-0.2316510081,0.1231539845,-0.1210027635,-0.3868949413,-0.0544536486,-0.2450863421,-0.2544884384,-0.2394929677,0.0445744172,-0.0382827111,0.379373312,0.1044166535,-0.1685189754,-0.1514848173,-0.4425003827,-0.028786093,0.1463311911,-0.0292478688,-0.0367810801,-0.1005667672,-0.2029286176,-0.5161777735,0.4155936837,0.2859541476,0.1105859429,-0.1718751192,0.0878228843,0.4478865266,-0.1853179783,0.6596498489,-0.1708530486,0.0438121483,0.0984816179,0.1837503612,-0.7251459956,-0.0194136817,0.147544682,0.1342761517,0.215312928,-0.1166655123,-0.5139329433,0.0979648083,0.3441002071,0.3732801974,-0.0886276141,-0.1792948097,-0.191272229,-0.3767274618,-0.5138627887,-0.1649242193,-0.2111011744,0.2683063149,-0.025560908,0.031048581,0.050385315,0.201478228,-0.0147314733,0.2821478844,0.2859053016,0.0165023264,0.0235097762,0.0959309414,0.0991042778,0.3470714688,0.3539325595,0.2881973684,-0.0423728712,-0.0889987722,0.0853011161,-0.0638199598,0.18121171,-0.1502871662,0.0666161627,-0.0016658714,-0.1978694946,-0.0734471083,-0.1572476029,0.3747443855,-0.221284762,-0.2982092798,-0.2389822006,0.7085088491,0.1589902639,-0.0960436761,0.1231104955,-0.1088686809,-0.1373421997,0.365239948,0.1177065223,0.9276636839,-0.1522064656,0.328930378,0.3286547959,-0.0396991223,0.5044326186,-0.2825063169,0.1339882761,-0.484410882,0.2856452167,-0.0108886296,-0.0980820805,-0.1546719223,0.0310812201,-0.1827827245,0.0733063594,-0.2207050174,0.0885881856,0.0106324665,0.3567408025,-0.008351001,0.1106858179,-0.0842744187,0.0562638715,-0.3195542693,0.2571137249,-0.2615233958,-0.0009585342,0.0979617238,-0.1563698649,-0.4544543624,0.0785484239,-0.1875041425,0.4533356726,-0.306944102,-0.3188786209,0.0176066961,0.0598248057,-0.0411643572,-0.0937833935,-0.3611280918,0.1809431165,-0.2230480611,-0.1647396535,-0.1060880497,0.0703283325,0.2264247984,-0.0705450103,-0.4650467038,0.3104069829,-0.0331905186,-0.2690660954,-0.195166409,0.0056172167,0.0766634122,-0.1938509792,-0.1571667492,-0.2530135214,-0.1191873401,-0.1358085871,0.0734140873,-0.1295603067,0.0084917909,0.0646586269,0.0685643181,-0.3869736493,0.1155601069,0.4719053209,0.3767473102,0.1686637253,0.6754417419,0.2683519423,-0.2663931847,-0.0136300661,-0.3504727185,-0.3611333072,-0.3933808804,-0.0750969127,0.1329663992,0.4606651962,-0.2361188233,0.1057871878,0.1565126032,-0.1004546285,0.1289465129,-0.4669057727,-0.2856480777,0.3458832204,-0.1558524072,0.2150657028,0.333362639,0.1371670812,0.0941938311,0.2817735076,-0.2074543387,0.4754821956,-0.0808619931,0.3691472411,0.1022728831,-0.1045177504,-0.1664323509,0.0043666679,0.0278084781,0.0733071864,-0.0842073485,-0.187368691,-0.059740562,0.1769646406,0.0595366582,-0.2068469971,-0.1221330166,-0.4088256657,0.0580776408,-0.010986723,0.3590635657,0.1042840257,-0.0777881294,0.1733734012,0.1805440485,0.1427157372,-0.2289880365,0.3619517982,-0.1334508955,0.3957297802,0.2669553459,0.2385378629,-0.0947233662,-0.0126234489,-0.2449375242,0.0805034339,0.1154414788,-0.1886783838,0.185227707,-0.2776901424,-0.2030886859,-0.0213793032,0.0164039209,0.3396817148,-0.1417613029,-0.054867357,-0.0608212724,0.1005864516,-0.2835876942,0.046061974,0.485242784,-0.1776818633,-0.2785930634,0.0687629804,0.5847008824,-0.0438929535,0.1663225442,0.0466528051,0.2393534482,0.0115823224,0.0854468048,0.2219160944,0.1098196283,-0.1377399415,0.0567916296,0.2720851004,0.18669644,0.4535924494,-0.1657377928,0.0975392014,-0.1927057505,0.0472147912,-0.1411454231,-0.3601846993,0.1597557068,0.4120803773,0.1318735033,0.0065879552,0.0914900675,0.3713221848,0.6368659735,0.0102780368,-0.0483850986,0.3622556925,-0.1486780196,-0.1804165244,-0.2978949547,-0.3707225621,0.0285330173,0.1508214623,-0.2404500395,-0.2233104259,0.2616764009,0.185880214,-0.3208230734,-0.6472076178,0.5188295245,0.2831636965,0.1351263523,-0.222577095,0.2115809023,0.1248400882,-0.0737039596,0.1012161523,0.2894658446,0.6575102806,0.3211216331,-0.7122462988,0.0241689552,-0.1587850153,0.0964294225,-0.0956941545,-0.0255603027,0.1547429264,-0.0726363733,0.2175964713,0.0557836443,-0.0313376933,-0.1637457311,0.440110445,0.0479708463,-0.214529708,-0.2602033913,-0.1867547482,-0.2116843313,0.0047944435,0.088947244,-0.372079432,0.2248482704,0.4339295626,0.0074966205,0.1416673213,-0.0861911252,0.0382655524,-0.2804442644,0.4323863685,0.2422127724,0.1094457954,-0.4453440011,-0.2630242109,-0.6950753927,0.3619217277,-0.4275679588,-0.004696277,0.3281984031,0.1819252819,-0.2007462084,0.1665137261,0.072097376,-0.1435965896,-0.185102731,0.3347555101,-0.2243230045,-0.1052198634,-0.1288038343,-0.2936487794,-0.25168854,-0.3236352205,0.214736551,-0.194979921,-0.04772361,-0.2938646376,-0.1443579942,0.1363732815,0.1835709214,0.508359313,0.0855292305,0.6090764999,0.0791577175,-0.0309701562,-0.2001253963,-0.1561925411,0.1780349463,0.296132803,0.0426095761,0.1194765791,-0.0950099453,-0.1640064567,-0.1698608994,0.5059783459,-0.0187877361,0.0761009902,-0.0015461057,-0.2256751806,0.0126297828,0.3037437499,-0.141428262,0.1818352789,-0.0693844333,0.3124789596,-0.2922698855,-0.4456368685,0.4373099804,-0.4284289479,-0.3710170984,0.0440436043,0.1237263009,-0.1983818561,-0.1003204212,-0.3580924273,0.1454563588,0.4109820127,-0.0160897728,-0.1379513741,-0.1122561172,-0.0849029645,-0.1042083427,-0.2067106366,0.0315526053,0.0420022234,-0.1544650644,-0.3315714896,-0.5234953761]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2144","title":"Loading wikipedia 20200501.en throws pyarrow related error","comments":"Hi ! It looks like the arrow file in the folder\r\n`\/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931` is corrupted.\r\n\r\nCan you take a look and check that it's 18.3GB ?\r\n\r\nIf not, then maybe you need to redownload it:\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache', download_mode=\"force_redownload\")\r\n```","body":"**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n","comment_length":46,"text":"Loading wikipedia 20200501.en throws pyarrow related error\n**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n\nHi ! It looks like the arrow file in the folder\r\n`\/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931` is corrupted.\r\n\r\nCan you take a look and check that it's 18.3GB ?\r\n\r\nIf not, then maybe you need to redownload it:\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache', download_mode=\"force_redownload\")\r\n```","embeddings":[-0.1041540205,0.3526462018,0.0340582058,0.3633103967,0.2745778561,0.1802078485,0.2762243748,0.4623250365,-0.04348813,-0.113784872,0.0247374158,0.236021921,0.1520678252,-0.1178016663,0.1841091067,-0.1683138758,0.0038767338,0.0068803728,0.0847920552,0.1328444481,-0.2001205236,0.1146743447,-0.1824860871,0.1555170566,-0.5857800245,0.0881166533,-0.0660176575,0.0540414415,0.0833243653,-0.613804698,0.3904819489,-0.0500538088,0.2010909319,0.2121308595,-0.0001204597,0.0749956965,0.4643647373,-0.0041334573,-0.2960134745,-0.0811445788,0.186580047,-0.0205053128,0.3180885315,-0.3607697487,0.1008782089,-0.2803391814,0.2242598385,0.0471996665,0.1832180023,0.0086322986,0.168285802,-0.1830526292,0.3375393748,0.0306443255,0.7112432122,-0.1374686658,0.0815054998,0.4424607456,0.1225971282,-0.1607152224,-0.186434865,0.23629798,-0.2502577007,0.1203090623,0.2671839893,-0.0770430341,0.3848769963,-0.2004953474,0.3921535015,0.2679577172,0.7762554288,-0.4092206359,0.1730330735,-0.2313123494,-0.0328712203,-0.0229594614,0.0938252583,0.2307374179,-0.27676934,0.0545302071,0.1571857184,-0.1188708618,-0.2609159052,0.301315248,-0.0642476678,0.1997214258,0.1399375498,0.3910686672,0.0163352489,-0.2268839478,0.0019353561,0.0632643551,0.1672586203,0.2208643556,-0.3931137919,0.107257776,-0.0683028772,0.1139047593,0.3778621256,-0.1693618745,-0.021596102,0.0659415349,0.2391710877,0.2064648122,0.4192265868,0.1241850182,-0.3237407506,-0.2987220883,0.2721662819,0.1988998652,-0.254214257,-0.0364414603,-0.0612123869,-0.1137952879,0.1385727972,-0.1709839702,0.4291738868,-0.2699913085,-0.220579192,0.1195769534,-0.1446794569,-0.0715681612,-0.1552301943,0.3867201805,-0.0916787088,0.2424993664,0.0672103316,0.0622379445,-0.1821437031,-0.1878142953,0.0057542627,0.1186774671,-0.2124670595,0.1797935814,0.2939305305,-0.1538039595,0.2583542168,0.1435221881,-0.1506101638,-0.1206758171,0.1700648069,0.0130382087,-0.2326727659,0.309833169,0.0075005512,0.2468717843,-0.0328951664,-0.1449694484,-0.1790345013,0.3182585537,-0.2848282754,-0.5152549744,-0.262249738,0.1209049523,-0.0116632255,0.0844510719,0.1616870016,0.0576540641,0.4413489699,-0.5136592984,-0.1003230512,-0.1257850081,-0.1855283529,-0.4527531266,0.1706373543,0.3363130093,-0.5569394231,0.0798316821,-0.0507644974,0.0824038908,0.2173671722,0.4450024068,-0.2298555672,0.1757664233,-0.1654862761,0.1924015433,0.1095912755,-0.3121576309,-0.5411019325,0.2140131146,-0.0612866841,0.1959047168,0.0259571318,0.2906033099,0.0093369419,0.1717624664,0.4052648544,0.2492511719,-0.0467647091,0.0787921399,-0.4085267782,-0.3879698217,0.3385632038,0.2016654164,-0.0047450308,-0.1988982856,0.1968246847,0.3221473992,0.2584798336,-0.2035795897,0.1287960708,0.0218389891,0.0705177039,-0.0129466308,0.2327663898,-0.2238298804,-0.4262207448,0.11144609,-0.3431405723,-0.0235893708,-0.3043866754,-0.361228615,-0.5231873989,0.0663694814,-0.171464175,-0.2400977015,0.115050301,-0.0613334365,0.0953495279,0.2569484115,0.1027859375,-0.0896757692,-0.2042666227,-0.0175576173,-0.4030902386,0.240516305,-0.308798492,-0.1786494255,-0.0221906919,0.2451284826,0.324608773,0.1002937332,0.0431751907,0.1558428705,0.1467771679,0.0931873098,-0.1324361414,-0.0177001171,0.2272260636,-0.2668777108,0.2040083557,0.1240440756,0.1868998408,-0.221017614,0.05095274,0.1322418749,-0.0019998513,0.1718758643,-0.0822376162,-0.0712248981,0.1570295691,-0.0034186426,0.3967843354,-0.1355208606,0.2599032819,0.4711607099,0.0430019759,0.1037084162,-0.0057747457,0.0103615662,0.1143461317,-0.0471609905,0.1531215012,0.1461279243,-0.3416619301,-0.2851566672,0.2188828588,-0.1442351341,0.0782735273,0.0937128961,0.0280694515,0.1398488432,-0.0689809099,-0.195103392,0.1316027343,-0.1178800836,0.4590720534,0.310067445,0.3028900325,-0.029343931,-0.3765672147,0.0593944713,0.0831138715,0.379365176,-0.3040897548,0.0794889852,-0.171332255,-0.243459627,-0.0772669688,-0.013044945,-0.2915118039,-0.317315042,0.0815705061,0.5144311786,0.0328679904,0.0099860048,-0.0540035702,-0.0480350479,0.0796879381,-0.1308909357,-0.0505061187,-0.7057309747,-0.3859112859,0.0051475652,0.2574861944,-0.1339800954,0.2502458096,-0.0254044235,-0.1927474737,-0.3766916692,-0.5513484478,-0.0036723248,-0.1104055196,0.3178443015,0.0176331513,0.3957713842,0.0727890357,-0.2152879387,0.2103869319,0.0130584994,-0.0248649139,0.2813790739,-0.3631691337,0.0844142213,-0.0089278743,-0.2284815609,-0.2832888961,-0.1489331275,0.1993967593,0.0630166978,0.1870315522,0.2383005023,0.0628918335,0.1012383327,0.2758947015,0.0676462352,-0.3717182577,-0.1608327031,0.3060976565,-0.2184316963,-0.2773063481,0.1912308037,0.002722593,0.1241865531,0.2216507792,-0.5899142623,0.0072065173,0.0857219473,0.3271362782,0.031964194,-0.1462098956,0.2367205918,0.0211273748,0.0012416313,0.062059287,0.0056531909,-0.1206239983,-0.1537077129,0.166241616,0.1140279919,0.2094682604,0.1647175252,0.7836135626,0.0867718309,0.0424960814,0.4263213873,-0.188306883,0.0635278597,-0.1462810636,-0.0966464877,0.0314848572,-0.3452987075,-0.2316510081,0.1231539845,-0.1210027635,-0.3868949413,-0.0544536486,-0.2450863421,-0.2544884384,-0.2394929677,0.0445744172,-0.0382827111,0.379373312,0.1044166535,-0.1685189754,-0.1514848173,-0.4425003827,-0.028786093,0.1463311911,-0.0292478688,-0.0367810801,-0.1005667672,-0.2029286176,-0.5161777735,0.4155936837,0.2859541476,0.1105859429,-0.1718751192,0.0878228843,0.4478865266,-0.1853179783,0.6596498489,-0.1708530486,0.0438121483,0.0984816179,0.1837503612,-0.7251459956,-0.0194136817,0.147544682,0.1342761517,0.215312928,-0.1166655123,-0.5139329433,0.0979648083,0.3441002071,0.3732801974,-0.0886276141,-0.1792948097,-0.191272229,-0.3767274618,-0.5138627887,-0.1649242193,-0.2111011744,0.2683063149,-0.025560908,0.031048581,0.050385315,0.201478228,-0.0147314733,0.2821478844,0.2859053016,0.0165023264,0.0235097762,0.0959309414,0.0991042778,0.3470714688,0.3539325595,0.2881973684,-0.0423728712,-0.0889987722,0.0853011161,-0.0638199598,0.18121171,-0.1502871662,0.0666161627,-0.0016658714,-0.1978694946,-0.0734471083,-0.1572476029,0.3747443855,-0.221284762,-0.2982092798,-0.2389822006,0.7085088491,0.1589902639,-0.0960436761,0.1231104955,-0.1088686809,-0.1373421997,0.365239948,0.1177065223,0.9276636839,-0.1522064656,0.328930378,0.3286547959,-0.0396991223,0.5044326186,-0.2825063169,0.1339882761,-0.484410882,0.2856452167,-0.0108886296,-0.0980820805,-0.1546719223,0.0310812201,-0.1827827245,0.0733063594,-0.2207050174,0.0885881856,0.0106324665,0.3567408025,-0.008351001,0.1106858179,-0.0842744187,0.0562638715,-0.3195542693,0.2571137249,-0.2615233958,-0.0009585342,0.0979617238,-0.1563698649,-0.4544543624,0.0785484239,-0.1875041425,0.4533356726,-0.306944102,-0.3188786209,0.0176066961,0.0598248057,-0.0411643572,-0.0937833935,-0.3611280918,0.1809431165,-0.2230480611,-0.1647396535,-0.1060880497,0.0703283325,0.2264247984,-0.0705450103,-0.4650467038,0.3104069829,-0.0331905186,-0.2690660954,-0.195166409,0.0056172167,0.0766634122,-0.1938509792,-0.1571667492,-0.2530135214,-0.1191873401,-0.1358085871,0.0734140873,-0.1295603067,0.0084917909,0.0646586269,0.0685643181,-0.3869736493,0.1155601069,0.4719053209,0.3767473102,0.1686637253,0.6754417419,0.2683519423,-0.2663931847,-0.0136300661,-0.3504727185,-0.3611333072,-0.3933808804,-0.0750969127,0.1329663992,0.4606651962,-0.2361188233,0.1057871878,0.1565126032,-0.1004546285,0.1289465129,-0.4669057727,-0.2856480777,0.3458832204,-0.1558524072,0.2150657028,0.333362639,0.1371670812,0.0941938311,0.2817735076,-0.2074543387,0.4754821956,-0.0808619931,0.3691472411,0.1022728831,-0.1045177504,-0.1664323509,0.0043666679,0.0278084781,0.0733071864,-0.0842073485,-0.187368691,-0.059740562,0.1769646406,0.0595366582,-0.2068469971,-0.1221330166,-0.4088256657,0.0580776408,-0.010986723,0.3590635657,0.1042840257,-0.0777881294,0.1733734012,0.1805440485,0.1427157372,-0.2289880365,0.3619517982,-0.1334508955,0.3957297802,0.2669553459,0.2385378629,-0.0947233662,-0.0126234489,-0.2449375242,0.0805034339,0.1154414788,-0.1886783838,0.185227707,-0.2776901424,-0.2030886859,-0.0213793032,0.0164039209,0.3396817148,-0.1417613029,-0.054867357,-0.0608212724,0.1005864516,-0.2835876942,0.046061974,0.485242784,-0.1776818633,-0.2785930634,0.0687629804,0.5847008824,-0.0438929535,0.1663225442,0.0466528051,0.2393534482,0.0115823224,0.0854468048,0.2219160944,0.1098196283,-0.1377399415,0.0567916296,0.2720851004,0.18669644,0.4535924494,-0.1657377928,0.0975392014,-0.1927057505,0.0472147912,-0.1411454231,-0.3601846993,0.1597557068,0.4120803773,0.1318735033,0.0065879552,0.0914900675,0.3713221848,0.6368659735,0.0102780368,-0.0483850986,0.3622556925,-0.1486780196,-0.1804165244,-0.2978949547,-0.3707225621,0.0285330173,0.1508214623,-0.2404500395,-0.2233104259,0.2616764009,0.185880214,-0.3208230734,-0.6472076178,0.5188295245,0.2831636965,0.1351263523,-0.222577095,0.2115809023,0.1248400882,-0.0737039596,0.1012161523,0.2894658446,0.6575102806,0.3211216331,-0.7122462988,0.0241689552,-0.1587850153,0.0964294225,-0.0956941545,-0.0255603027,0.1547429264,-0.0726363733,0.2175964713,0.0557836443,-0.0313376933,-0.1637457311,0.440110445,0.0479708463,-0.214529708,-0.2602033913,-0.1867547482,-0.2116843313,0.0047944435,0.088947244,-0.372079432,0.2248482704,0.4339295626,0.0074966205,0.1416673213,-0.0861911252,0.0382655524,-0.2804442644,0.4323863685,0.2422127724,0.1094457954,-0.4453440011,-0.2630242109,-0.6950753927,0.3619217277,-0.4275679588,-0.004696277,0.3281984031,0.1819252819,-0.2007462084,0.1665137261,0.072097376,-0.1435965896,-0.185102731,0.3347555101,-0.2243230045,-0.1052198634,-0.1288038343,-0.2936487794,-0.25168854,-0.3236352205,0.214736551,-0.194979921,-0.04772361,-0.2938646376,-0.1443579942,0.1363732815,0.1835709214,0.508359313,0.0855292305,0.6090764999,0.0791577175,-0.0309701562,-0.2001253963,-0.1561925411,0.1780349463,0.296132803,0.0426095761,0.1194765791,-0.0950099453,-0.1640064567,-0.1698608994,0.5059783459,-0.0187877361,0.0761009902,-0.0015461057,-0.2256751806,0.0126297828,0.3037437499,-0.141428262,0.1818352789,-0.0693844333,0.3124789596,-0.2922698855,-0.4456368685,0.4373099804,-0.4284289479,-0.3710170984,0.0440436043,0.1237263009,-0.1983818561,-0.1003204212,-0.3580924273,0.1454563588,0.4109820127,-0.0160897728,-0.1379513741,-0.1122561172,-0.0849029645,-0.1042083427,-0.2067106366,0.0315526053,0.0420022234,-0.1544650644,-0.3315714896,-0.5234953761]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2144","title":"Loading wikipedia 20200501.en throws pyarrow related error","comments":"> Hi ! It looks like the arrow file in the folder\r\n> `\/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931` is corrupted.\r\n> \r\n> Can you take a look and check that it's 18.3GB ?\r\n> \r\n> If not, then maybe you need to redownload it:\r\n> \r\n> ```python\r\n> from datasets import load_dataset\r\n> ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache', download_mode=\"force_redownload\")\r\n> ```\r\n\r\nHi Ihoestq, thanks for the reply! Actually i think my issue is i couldn't download the dataset beyond 10.7G. It feels like the whole dataset is split into different volumes and after the first one was downloaded it crashed before proceeding to the next one. I did try 'force_redownload' mode but still got the same issue.","body":"**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n","comment_length":113,"text":"Loading wikipedia 20200501.en throws pyarrow related error\n**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n\n> Hi ! It looks like the arrow file in the folder\r\n> `\/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931` is corrupted.\r\n> \r\n> Can you take a look and check that it's 18.3GB ?\r\n> \r\n> If not, then maybe you need to redownload it:\r\n> \r\n> ```python\r\n> from datasets import load_dataset\r\n> ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache', download_mode=\"force_redownload\")\r\n> ```\r\n\r\nHi Ihoestq, thanks for the reply! Actually i think my issue is i couldn't download the dataset beyond 10.7G. It feels like the whole dataset is split into different volumes and after the first one was downloaded it crashed before proceeding to the next one. I did try 'force_redownload' mode but still got the same issue.","embeddings":[-0.1041540205,0.3526462018,0.0340582058,0.3633103967,0.2745778561,0.1802078485,0.2762243748,0.4623250365,-0.04348813,-0.113784872,0.0247374158,0.236021921,0.1520678252,-0.1178016663,0.1841091067,-0.1683138758,0.0038767338,0.0068803728,0.0847920552,0.1328444481,-0.2001205236,0.1146743447,-0.1824860871,0.1555170566,-0.5857800245,0.0881166533,-0.0660176575,0.0540414415,0.0833243653,-0.613804698,0.3904819489,-0.0500538088,0.2010909319,0.2121308595,-0.0001204597,0.0749956965,0.4643647373,-0.0041334573,-0.2960134745,-0.0811445788,0.186580047,-0.0205053128,0.3180885315,-0.3607697487,0.1008782089,-0.2803391814,0.2242598385,0.0471996665,0.1832180023,0.0086322986,0.168285802,-0.1830526292,0.3375393748,0.0306443255,0.7112432122,-0.1374686658,0.0815054998,0.4424607456,0.1225971282,-0.1607152224,-0.186434865,0.23629798,-0.2502577007,0.1203090623,0.2671839893,-0.0770430341,0.3848769963,-0.2004953474,0.3921535015,0.2679577172,0.7762554288,-0.4092206359,0.1730330735,-0.2313123494,-0.0328712203,-0.0229594614,0.0938252583,0.2307374179,-0.27676934,0.0545302071,0.1571857184,-0.1188708618,-0.2609159052,0.301315248,-0.0642476678,0.1997214258,0.1399375498,0.3910686672,0.0163352489,-0.2268839478,0.0019353561,0.0632643551,0.1672586203,0.2208643556,-0.3931137919,0.107257776,-0.0683028772,0.1139047593,0.3778621256,-0.1693618745,-0.021596102,0.0659415349,0.2391710877,0.2064648122,0.4192265868,0.1241850182,-0.3237407506,-0.2987220883,0.2721662819,0.1988998652,-0.254214257,-0.0364414603,-0.0612123869,-0.1137952879,0.1385727972,-0.1709839702,0.4291738868,-0.2699913085,-0.220579192,0.1195769534,-0.1446794569,-0.0715681612,-0.1552301943,0.3867201805,-0.0916787088,0.2424993664,0.0672103316,0.0622379445,-0.1821437031,-0.1878142953,0.0057542627,0.1186774671,-0.2124670595,0.1797935814,0.2939305305,-0.1538039595,0.2583542168,0.1435221881,-0.1506101638,-0.1206758171,0.1700648069,0.0130382087,-0.2326727659,0.309833169,0.0075005512,0.2468717843,-0.0328951664,-0.1449694484,-0.1790345013,0.3182585537,-0.2848282754,-0.5152549744,-0.262249738,0.1209049523,-0.0116632255,0.0844510719,0.1616870016,0.0576540641,0.4413489699,-0.5136592984,-0.1003230512,-0.1257850081,-0.1855283529,-0.4527531266,0.1706373543,0.3363130093,-0.5569394231,0.0798316821,-0.0507644974,0.0824038908,0.2173671722,0.4450024068,-0.2298555672,0.1757664233,-0.1654862761,0.1924015433,0.1095912755,-0.3121576309,-0.5411019325,0.2140131146,-0.0612866841,0.1959047168,0.0259571318,0.2906033099,0.0093369419,0.1717624664,0.4052648544,0.2492511719,-0.0467647091,0.0787921399,-0.4085267782,-0.3879698217,0.3385632038,0.2016654164,-0.0047450308,-0.1988982856,0.1968246847,0.3221473992,0.2584798336,-0.2035795897,0.1287960708,0.0218389891,0.0705177039,-0.0129466308,0.2327663898,-0.2238298804,-0.4262207448,0.11144609,-0.3431405723,-0.0235893708,-0.3043866754,-0.361228615,-0.5231873989,0.0663694814,-0.171464175,-0.2400977015,0.115050301,-0.0613334365,0.0953495279,0.2569484115,0.1027859375,-0.0896757692,-0.2042666227,-0.0175576173,-0.4030902386,0.240516305,-0.308798492,-0.1786494255,-0.0221906919,0.2451284826,0.324608773,0.1002937332,0.0431751907,0.1558428705,0.1467771679,0.0931873098,-0.1324361414,-0.0177001171,0.2272260636,-0.2668777108,0.2040083557,0.1240440756,0.1868998408,-0.221017614,0.05095274,0.1322418749,-0.0019998513,0.1718758643,-0.0822376162,-0.0712248981,0.1570295691,-0.0034186426,0.3967843354,-0.1355208606,0.2599032819,0.4711607099,0.0430019759,0.1037084162,-0.0057747457,0.0103615662,0.1143461317,-0.0471609905,0.1531215012,0.1461279243,-0.3416619301,-0.2851566672,0.2188828588,-0.1442351341,0.0782735273,0.0937128961,0.0280694515,0.1398488432,-0.0689809099,-0.195103392,0.1316027343,-0.1178800836,0.4590720534,0.310067445,0.3028900325,-0.029343931,-0.3765672147,0.0593944713,0.0831138715,0.379365176,-0.3040897548,0.0794889852,-0.171332255,-0.243459627,-0.0772669688,-0.013044945,-0.2915118039,-0.317315042,0.0815705061,0.5144311786,0.0328679904,0.0099860048,-0.0540035702,-0.0480350479,0.0796879381,-0.1308909357,-0.0505061187,-0.7057309747,-0.3859112859,0.0051475652,0.2574861944,-0.1339800954,0.2502458096,-0.0254044235,-0.1927474737,-0.3766916692,-0.5513484478,-0.0036723248,-0.1104055196,0.3178443015,0.0176331513,0.3957713842,0.0727890357,-0.2152879387,0.2103869319,0.0130584994,-0.0248649139,0.2813790739,-0.3631691337,0.0844142213,-0.0089278743,-0.2284815609,-0.2832888961,-0.1489331275,0.1993967593,0.0630166978,0.1870315522,0.2383005023,0.0628918335,0.1012383327,0.2758947015,0.0676462352,-0.3717182577,-0.1608327031,0.3060976565,-0.2184316963,-0.2773063481,0.1912308037,0.002722593,0.1241865531,0.2216507792,-0.5899142623,0.0072065173,0.0857219473,0.3271362782,0.031964194,-0.1462098956,0.2367205918,0.0211273748,0.0012416313,0.062059287,0.0056531909,-0.1206239983,-0.1537077129,0.166241616,0.1140279919,0.2094682604,0.1647175252,0.7836135626,0.0867718309,0.0424960814,0.4263213873,-0.188306883,0.0635278597,-0.1462810636,-0.0966464877,0.0314848572,-0.3452987075,-0.2316510081,0.1231539845,-0.1210027635,-0.3868949413,-0.0544536486,-0.2450863421,-0.2544884384,-0.2394929677,0.0445744172,-0.0382827111,0.379373312,0.1044166535,-0.1685189754,-0.1514848173,-0.4425003827,-0.028786093,0.1463311911,-0.0292478688,-0.0367810801,-0.1005667672,-0.2029286176,-0.5161777735,0.4155936837,0.2859541476,0.1105859429,-0.1718751192,0.0878228843,0.4478865266,-0.1853179783,0.6596498489,-0.1708530486,0.0438121483,0.0984816179,0.1837503612,-0.7251459956,-0.0194136817,0.147544682,0.1342761517,0.215312928,-0.1166655123,-0.5139329433,0.0979648083,0.3441002071,0.3732801974,-0.0886276141,-0.1792948097,-0.191272229,-0.3767274618,-0.5138627887,-0.1649242193,-0.2111011744,0.2683063149,-0.025560908,0.031048581,0.050385315,0.201478228,-0.0147314733,0.2821478844,0.2859053016,0.0165023264,0.0235097762,0.0959309414,0.0991042778,0.3470714688,0.3539325595,0.2881973684,-0.0423728712,-0.0889987722,0.0853011161,-0.0638199598,0.18121171,-0.1502871662,0.0666161627,-0.0016658714,-0.1978694946,-0.0734471083,-0.1572476029,0.3747443855,-0.221284762,-0.2982092798,-0.2389822006,0.7085088491,0.1589902639,-0.0960436761,0.1231104955,-0.1088686809,-0.1373421997,0.365239948,0.1177065223,0.9276636839,-0.1522064656,0.328930378,0.3286547959,-0.0396991223,0.5044326186,-0.2825063169,0.1339882761,-0.484410882,0.2856452167,-0.0108886296,-0.0980820805,-0.1546719223,0.0310812201,-0.1827827245,0.0733063594,-0.2207050174,0.0885881856,0.0106324665,0.3567408025,-0.008351001,0.1106858179,-0.0842744187,0.0562638715,-0.3195542693,0.2571137249,-0.2615233958,-0.0009585342,0.0979617238,-0.1563698649,-0.4544543624,0.0785484239,-0.1875041425,0.4533356726,-0.306944102,-0.3188786209,0.0176066961,0.0598248057,-0.0411643572,-0.0937833935,-0.3611280918,0.1809431165,-0.2230480611,-0.1647396535,-0.1060880497,0.0703283325,0.2264247984,-0.0705450103,-0.4650467038,0.3104069829,-0.0331905186,-0.2690660954,-0.195166409,0.0056172167,0.0766634122,-0.1938509792,-0.1571667492,-0.2530135214,-0.1191873401,-0.1358085871,0.0734140873,-0.1295603067,0.0084917909,0.0646586269,0.0685643181,-0.3869736493,0.1155601069,0.4719053209,0.3767473102,0.1686637253,0.6754417419,0.2683519423,-0.2663931847,-0.0136300661,-0.3504727185,-0.3611333072,-0.3933808804,-0.0750969127,0.1329663992,0.4606651962,-0.2361188233,0.1057871878,0.1565126032,-0.1004546285,0.1289465129,-0.4669057727,-0.2856480777,0.3458832204,-0.1558524072,0.2150657028,0.333362639,0.1371670812,0.0941938311,0.2817735076,-0.2074543387,0.4754821956,-0.0808619931,0.3691472411,0.1022728831,-0.1045177504,-0.1664323509,0.0043666679,0.0278084781,0.0733071864,-0.0842073485,-0.187368691,-0.059740562,0.1769646406,0.0595366582,-0.2068469971,-0.1221330166,-0.4088256657,0.0580776408,-0.010986723,0.3590635657,0.1042840257,-0.0777881294,0.1733734012,0.1805440485,0.1427157372,-0.2289880365,0.3619517982,-0.1334508955,0.3957297802,0.2669553459,0.2385378629,-0.0947233662,-0.0126234489,-0.2449375242,0.0805034339,0.1154414788,-0.1886783838,0.185227707,-0.2776901424,-0.2030886859,-0.0213793032,0.0164039209,0.3396817148,-0.1417613029,-0.054867357,-0.0608212724,0.1005864516,-0.2835876942,0.046061974,0.485242784,-0.1776818633,-0.2785930634,0.0687629804,0.5847008824,-0.0438929535,0.1663225442,0.0466528051,0.2393534482,0.0115823224,0.0854468048,0.2219160944,0.1098196283,-0.1377399415,0.0567916296,0.2720851004,0.18669644,0.4535924494,-0.1657377928,0.0975392014,-0.1927057505,0.0472147912,-0.1411454231,-0.3601846993,0.1597557068,0.4120803773,0.1318735033,0.0065879552,0.0914900675,0.3713221848,0.6368659735,0.0102780368,-0.0483850986,0.3622556925,-0.1486780196,-0.1804165244,-0.2978949547,-0.3707225621,0.0285330173,0.1508214623,-0.2404500395,-0.2233104259,0.2616764009,0.185880214,-0.3208230734,-0.6472076178,0.5188295245,0.2831636965,0.1351263523,-0.222577095,0.2115809023,0.1248400882,-0.0737039596,0.1012161523,0.2894658446,0.6575102806,0.3211216331,-0.7122462988,0.0241689552,-0.1587850153,0.0964294225,-0.0956941545,-0.0255603027,0.1547429264,-0.0726363733,0.2175964713,0.0557836443,-0.0313376933,-0.1637457311,0.440110445,0.0479708463,-0.214529708,-0.2602033913,-0.1867547482,-0.2116843313,0.0047944435,0.088947244,-0.372079432,0.2248482704,0.4339295626,0.0074966205,0.1416673213,-0.0861911252,0.0382655524,-0.2804442644,0.4323863685,0.2422127724,0.1094457954,-0.4453440011,-0.2630242109,-0.6950753927,0.3619217277,-0.4275679588,-0.004696277,0.3281984031,0.1819252819,-0.2007462084,0.1665137261,0.072097376,-0.1435965896,-0.185102731,0.3347555101,-0.2243230045,-0.1052198634,-0.1288038343,-0.2936487794,-0.25168854,-0.3236352205,0.214736551,-0.194979921,-0.04772361,-0.2938646376,-0.1443579942,0.1363732815,0.1835709214,0.508359313,0.0855292305,0.6090764999,0.0791577175,-0.0309701562,-0.2001253963,-0.1561925411,0.1780349463,0.296132803,0.0426095761,0.1194765791,-0.0950099453,-0.1640064567,-0.1698608994,0.5059783459,-0.0187877361,0.0761009902,-0.0015461057,-0.2256751806,0.0126297828,0.3037437499,-0.141428262,0.1818352789,-0.0693844333,0.3124789596,-0.2922698855,-0.4456368685,0.4373099804,-0.4284289479,-0.3710170984,0.0440436043,0.1237263009,-0.1983818561,-0.1003204212,-0.3580924273,0.1454563588,0.4109820127,-0.0160897728,-0.1379513741,-0.1122561172,-0.0849029645,-0.1042083427,-0.2067106366,0.0315526053,0.0420022234,-0.1544650644,-0.3315714896,-0.5234953761]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2144","title":"Loading wikipedia 20200501.en throws pyarrow related error","comments":"I just tried on my side and got no issues.\r\nWhen downloading the dataset again, did it crash at 10.7GB as well ?","body":"**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n","comment_length":23,"text":"Loading wikipedia 20200501.en throws pyarrow related error\n**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n\nI just tried on my side and got no issues.\r\nWhen downloading the dataset again, did it crash at 10.7GB as well ?","embeddings":[-0.1041540205,0.3526462018,0.0340582058,0.3633103967,0.2745778561,0.1802078485,0.2762243748,0.4623250365,-0.04348813,-0.113784872,0.0247374158,0.236021921,0.1520678252,-0.1178016663,0.1841091067,-0.1683138758,0.0038767338,0.0068803728,0.0847920552,0.1328444481,-0.2001205236,0.1146743447,-0.1824860871,0.1555170566,-0.5857800245,0.0881166533,-0.0660176575,0.0540414415,0.0833243653,-0.613804698,0.3904819489,-0.0500538088,0.2010909319,0.2121308595,-0.0001204597,0.0749956965,0.4643647373,-0.0041334573,-0.2960134745,-0.0811445788,0.186580047,-0.0205053128,0.3180885315,-0.3607697487,0.1008782089,-0.2803391814,0.2242598385,0.0471996665,0.1832180023,0.0086322986,0.168285802,-0.1830526292,0.3375393748,0.0306443255,0.7112432122,-0.1374686658,0.0815054998,0.4424607456,0.1225971282,-0.1607152224,-0.186434865,0.23629798,-0.2502577007,0.1203090623,0.2671839893,-0.0770430341,0.3848769963,-0.2004953474,0.3921535015,0.2679577172,0.7762554288,-0.4092206359,0.1730330735,-0.2313123494,-0.0328712203,-0.0229594614,0.0938252583,0.2307374179,-0.27676934,0.0545302071,0.1571857184,-0.1188708618,-0.2609159052,0.301315248,-0.0642476678,0.1997214258,0.1399375498,0.3910686672,0.0163352489,-0.2268839478,0.0019353561,0.0632643551,0.1672586203,0.2208643556,-0.3931137919,0.107257776,-0.0683028772,0.1139047593,0.3778621256,-0.1693618745,-0.021596102,0.0659415349,0.2391710877,0.2064648122,0.4192265868,0.1241850182,-0.3237407506,-0.2987220883,0.2721662819,0.1988998652,-0.254214257,-0.0364414603,-0.0612123869,-0.1137952879,0.1385727972,-0.1709839702,0.4291738868,-0.2699913085,-0.220579192,0.1195769534,-0.1446794569,-0.0715681612,-0.1552301943,0.3867201805,-0.0916787088,0.2424993664,0.0672103316,0.0622379445,-0.1821437031,-0.1878142953,0.0057542627,0.1186774671,-0.2124670595,0.1797935814,0.2939305305,-0.1538039595,0.2583542168,0.1435221881,-0.1506101638,-0.1206758171,0.1700648069,0.0130382087,-0.2326727659,0.309833169,0.0075005512,0.2468717843,-0.0328951664,-0.1449694484,-0.1790345013,0.3182585537,-0.2848282754,-0.5152549744,-0.262249738,0.1209049523,-0.0116632255,0.0844510719,0.1616870016,0.0576540641,0.4413489699,-0.5136592984,-0.1003230512,-0.1257850081,-0.1855283529,-0.4527531266,0.1706373543,0.3363130093,-0.5569394231,0.0798316821,-0.0507644974,0.0824038908,0.2173671722,0.4450024068,-0.2298555672,0.1757664233,-0.1654862761,0.1924015433,0.1095912755,-0.3121576309,-0.5411019325,0.2140131146,-0.0612866841,0.1959047168,0.0259571318,0.2906033099,0.0093369419,0.1717624664,0.4052648544,0.2492511719,-0.0467647091,0.0787921399,-0.4085267782,-0.3879698217,0.3385632038,0.2016654164,-0.0047450308,-0.1988982856,0.1968246847,0.3221473992,0.2584798336,-0.2035795897,0.1287960708,0.0218389891,0.0705177039,-0.0129466308,0.2327663898,-0.2238298804,-0.4262207448,0.11144609,-0.3431405723,-0.0235893708,-0.3043866754,-0.361228615,-0.5231873989,0.0663694814,-0.171464175,-0.2400977015,0.115050301,-0.0613334365,0.0953495279,0.2569484115,0.1027859375,-0.0896757692,-0.2042666227,-0.0175576173,-0.4030902386,0.240516305,-0.308798492,-0.1786494255,-0.0221906919,0.2451284826,0.324608773,0.1002937332,0.0431751907,0.1558428705,0.1467771679,0.0931873098,-0.1324361414,-0.0177001171,0.2272260636,-0.2668777108,0.2040083557,0.1240440756,0.1868998408,-0.221017614,0.05095274,0.1322418749,-0.0019998513,0.1718758643,-0.0822376162,-0.0712248981,0.1570295691,-0.0034186426,0.3967843354,-0.1355208606,0.2599032819,0.4711607099,0.0430019759,0.1037084162,-0.0057747457,0.0103615662,0.1143461317,-0.0471609905,0.1531215012,0.1461279243,-0.3416619301,-0.2851566672,0.2188828588,-0.1442351341,0.0782735273,0.0937128961,0.0280694515,0.1398488432,-0.0689809099,-0.195103392,0.1316027343,-0.1178800836,0.4590720534,0.310067445,0.3028900325,-0.029343931,-0.3765672147,0.0593944713,0.0831138715,0.379365176,-0.3040897548,0.0794889852,-0.171332255,-0.243459627,-0.0772669688,-0.013044945,-0.2915118039,-0.317315042,0.0815705061,0.5144311786,0.0328679904,0.0099860048,-0.0540035702,-0.0480350479,0.0796879381,-0.1308909357,-0.0505061187,-0.7057309747,-0.3859112859,0.0051475652,0.2574861944,-0.1339800954,0.2502458096,-0.0254044235,-0.1927474737,-0.3766916692,-0.5513484478,-0.0036723248,-0.1104055196,0.3178443015,0.0176331513,0.3957713842,0.0727890357,-0.2152879387,0.2103869319,0.0130584994,-0.0248649139,0.2813790739,-0.3631691337,0.0844142213,-0.0089278743,-0.2284815609,-0.2832888961,-0.1489331275,0.1993967593,0.0630166978,0.1870315522,0.2383005023,0.0628918335,0.1012383327,0.2758947015,0.0676462352,-0.3717182577,-0.1608327031,0.3060976565,-0.2184316963,-0.2773063481,0.1912308037,0.002722593,0.1241865531,0.2216507792,-0.5899142623,0.0072065173,0.0857219473,0.3271362782,0.031964194,-0.1462098956,0.2367205918,0.0211273748,0.0012416313,0.062059287,0.0056531909,-0.1206239983,-0.1537077129,0.166241616,0.1140279919,0.2094682604,0.1647175252,0.7836135626,0.0867718309,0.0424960814,0.4263213873,-0.188306883,0.0635278597,-0.1462810636,-0.0966464877,0.0314848572,-0.3452987075,-0.2316510081,0.1231539845,-0.1210027635,-0.3868949413,-0.0544536486,-0.2450863421,-0.2544884384,-0.2394929677,0.0445744172,-0.0382827111,0.379373312,0.1044166535,-0.1685189754,-0.1514848173,-0.4425003827,-0.028786093,0.1463311911,-0.0292478688,-0.0367810801,-0.1005667672,-0.2029286176,-0.5161777735,0.4155936837,0.2859541476,0.1105859429,-0.1718751192,0.0878228843,0.4478865266,-0.1853179783,0.6596498489,-0.1708530486,0.0438121483,0.0984816179,0.1837503612,-0.7251459956,-0.0194136817,0.147544682,0.1342761517,0.215312928,-0.1166655123,-0.5139329433,0.0979648083,0.3441002071,0.3732801974,-0.0886276141,-0.1792948097,-0.191272229,-0.3767274618,-0.5138627887,-0.1649242193,-0.2111011744,0.2683063149,-0.025560908,0.031048581,0.050385315,0.201478228,-0.0147314733,0.2821478844,0.2859053016,0.0165023264,0.0235097762,0.0959309414,0.0991042778,0.3470714688,0.3539325595,0.2881973684,-0.0423728712,-0.0889987722,0.0853011161,-0.0638199598,0.18121171,-0.1502871662,0.0666161627,-0.0016658714,-0.1978694946,-0.0734471083,-0.1572476029,0.3747443855,-0.221284762,-0.2982092798,-0.2389822006,0.7085088491,0.1589902639,-0.0960436761,0.1231104955,-0.1088686809,-0.1373421997,0.365239948,0.1177065223,0.9276636839,-0.1522064656,0.328930378,0.3286547959,-0.0396991223,0.5044326186,-0.2825063169,0.1339882761,-0.484410882,0.2856452167,-0.0108886296,-0.0980820805,-0.1546719223,0.0310812201,-0.1827827245,0.0733063594,-0.2207050174,0.0885881856,0.0106324665,0.3567408025,-0.008351001,0.1106858179,-0.0842744187,0.0562638715,-0.3195542693,0.2571137249,-0.2615233958,-0.0009585342,0.0979617238,-0.1563698649,-0.4544543624,0.0785484239,-0.1875041425,0.4533356726,-0.306944102,-0.3188786209,0.0176066961,0.0598248057,-0.0411643572,-0.0937833935,-0.3611280918,0.1809431165,-0.2230480611,-0.1647396535,-0.1060880497,0.0703283325,0.2264247984,-0.0705450103,-0.4650467038,0.3104069829,-0.0331905186,-0.2690660954,-0.195166409,0.0056172167,0.0766634122,-0.1938509792,-0.1571667492,-0.2530135214,-0.1191873401,-0.1358085871,0.0734140873,-0.1295603067,0.0084917909,0.0646586269,0.0685643181,-0.3869736493,0.1155601069,0.4719053209,0.3767473102,0.1686637253,0.6754417419,0.2683519423,-0.2663931847,-0.0136300661,-0.3504727185,-0.3611333072,-0.3933808804,-0.0750969127,0.1329663992,0.4606651962,-0.2361188233,0.1057871878,0.1565126032,-0.1004546285,0.1289465129,-0.4669057727,-0.2856480777,0.3458832204,-0.1558524072,0.2150657028,0.333362639,0.1371670812,0.0941938311,0.2817735076,-0.2074543387,0.4754821956,-0.0808619931,0.3691472411,0.1022728831,-0.1045177504,-0.1664323509,0.0043666679,0.0278084781,0.0733071864,-0.0842073485,-0.187368691,-0.059740562,0.1769646406,0.0595366582,-0.2068469971,-0.1221330166,-0.4088256657,0.0580776408,-0.010986723,0.3590635657,0.1042840257,-0.0777881294,0.1733734012,0.1805440485,0.1427157372,-0.2289880365,0.3619517982,-0.1334508955,0.3957297802,0.2669553459,0.2385378629,-0.0947233662,-0.0126234489,-0.2449375242,0.0805034339,0.1154414788,-0.1886783838,0.185227707,-0.2776901424,-0.2030886859,-0.0213793032,0.0164039209,0.3396817148,-0.1417613029,-0.054867357,-0.0608212724,0.1005864516,-0.2835876942,0.046061974,0.485242784,-0.1776818633,-0.2785930634,0.0687629804,0.5847008824,-0.0438929535,0.1663225442,0.0466528051,0.2393534482,0.0115823224,0.0854468048,0.2219160944,0.1098196283,-0.1377399415,0.0567916296,0.2720851004,0.18669644,0.4535924494,-0.1657377928,0.0975392014,-0.1927057505,0.0472147912,-0.1411454231,-0.3601846993,0.1597557068,0.4120803773,0.1318735033,0.0065879552,0.0914900675,0.3713221848,0.6368659735,0.0102780368,-0.0483850986,0.3622556925,-0.1486780196,-0.1804165244,-0.2978949547,-0.3707225621,0.0285330173,0.1508214623,-0.2404500395,-0.2233104259,0.2616764009,0.185880214,-0.3208230734,-0.6472076178,0.5188295245,0.2831636965,0.1351263523,-0.222577095,0.2115809023,0.1248400882,-0.0737039596,0.1012161523,0.2894658446,0.6575102806,0.3211216331,-0.7122462988,0.0241689552,-0.1587850153,0.0964294225,-0.0956941545,-0.0255603027,0.1547429264,-0.0726363733,0.2175964713,0.0557836443,-0.0313376933,-0.1637457311,0.440110445,0.0479708463,-0.214529708,-0.2602033913,-0.1867547482,-0.2116843313,0.0047944435,0.088947244,-0.372079432,0.2248482704,0.4339295626,0.0074966205,0.1416673213,-0.0861911252,0.0382655524,-0.2804442644,0.4323863685,0.2422127724,0.1094457954,-0.4453440011,-0.2630242109,-0.6950753927,0.3619217277,-0.4275679588,-0.004696277,0.3281984031,0.1819252819,-0.2007462084,0.1665137261,0.072097376,-0.1435965896,-0.185102731,0.3347555101,-0.2243230045,-0.1052198634,-0.1288038343,-0.2936487794,-0.25168854,-0.3236352205,0.214736551,-0.194979921,-0.04772361,-0.2938646376,-0.1443579942,0.1363732815,0.1835709214,0.508359313,0.0855292305,0.6090764999,0.0791577175,-0.0309701562,-0.2001253963,-0.1561925411,0.1780349463,0.296132803,0.0426095761,0.1194765791,-0.0950099453,-0.1640064567,-0.1698608994,0.5059783459,-0.0187877361,0.0761009902,-0.0015461057,-0.2256751806,0.0126297828,0.3037437499,-0.141428262,0.1818352789,-0.0693844333,0.3124789596,-0.2922698855,-0.4456368685,0.4373099804,-0.4284289479,-0.3710170984,0.0440436043,0.1237263009,-0.1983818561,-0.1003204212,-0.3580924273,0.1454563588,0.4109820127,-0.0160897728,-0.1379513741,-0.1122561172,-0.0849029645,-0.1042083427,-0.2067106366,0.0315526053,0.0420022234,-0.1544650644,-0.3315714896,-0.5234953761]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2144","title":"Loading wikipedia 20200501.en throws pyarrow related error","comments":"> I just tried on my side and got no issues.\r\n> When downloading the dataset again, did it crash at 10.7GB as well ?\r\n\r\nYes i have tried it multiple times on different machines. I am wondering if you could share the screenshot of your dependency versions and i will try to make them the same as yours?","body":"**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n","comment_length":59,"text":"Loading wikipedia 20200501.en throws pyarrow related error\n**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n\n> I just tried on my side and got no issues.\r\n> When downloading the dataset again, did it crash at 10.7GB as well ?\r\n\r\nYes i have tried it multiple times on different machines. I am wondering if you could share the screenshot of your dependency versions and i will try to make them the same as yours?","embeddings":[-0.1041540205,0.3526462018,0.0340582058,0.3633103967,0.2745778561,0.1802078485,0.2762243748,0.4623250365,-0.04348813,-0.113784872,0.0247374158,0.236021921,0.1520678252,-0.1178016663,0.1841091067,-0.1683138758,0.0038767338,0.0068803728,0.0847920552,0.1328444481,-0.2001205236,0.1146743447,-0.1824860871,0.1555170566,-0.5857800245,0.0881166533,-0.0660176575,0.0540414415,0.0833243653,-0.613804698,0.3904819489,-0.0500538088,0.2010909319,0.2121308595,-0.0001204597,0.0749956965,0.4643647373,-0.0041334573,-0.2960134745,-0.0811445788,0.186580047,-0.0205053128,0.3180885315,-0.3607697487,0.1008782089,-0.2803391814,0.2242598385,0.0471996665,0.1832180023,0.0086322986,0.168285802,-0.1830526292,0.3375393748,0.0306443255,0.7112432122,-0.1374686658,0.0815054998,0.4424607456,0.1225971282,-0.1607152224,-0.186434865,0.23629798,-0.2502577007,0.1203090623,0.2671839893,-0.0770430341,0.3848769963,-0.2004953474,0.3921535015,0.2679577172,0.7762554288,-0.4092206359,0.1730330735,-0.2313123494,-0.0328712203,-0.0229594614,0.0938252583,0.2307374179,-0.27676934,0.0545302071,0.1571857184,-0.1188708618,-0.2609159052,0.301315248,-0.0642476678,0.1997214258,0.1399375498,0.3910686672,0.0163352489,-0.2268839478,0.0019353561,0.0632643551,0.1672586203,0.2208643556,-0.3931137919,0.107257776,-0.0683028772,0.1139047593,0.3778621256,-0.1693618745,-0.021596102,0.0659415349,0.2391710877,0.2064648122,0.4192265868,0.1241850182,-0.3237407506,-0.2987220883,0.2721662819,0.1988998652,-0.254214257,-0.0364414603,-0.0612123869,-0.1137952879,0.1385727972,-0.1709839702,0.4291738868,-0.2699913085,-0.220579192,0.1195769534,-0.1446794569,-0.0715681612,-0.1552301943,0.3867201805,-0.0916787088,0.2424993664,0.0672103316,0.0622379445,-0.1821437031,-0.1878142953,0.0057542627,0.1186774671,-0.2124670595,0.1797935814,0.2939305305,-0.1538039595,0.2583542168,0.1435221881,-0.1506101638,-0.1206758171,0.1700648069,0.0130382087,-0.2326727659,0.309833169,0.0075005512,0.2468717843,-0.0328951664,-0.1449694484,-0.1790345013,0.3182585537,-0.2848282754,-0.5152549744,-0.262249738,0.1209049523,-0.0116632255,0.0844510719,0.1616870016,0.0576540641,0.4413489699,-0.5136592984,-0.1003230512,-0.1257850081,-0.1855283529,-0.4527531266,0.1706373543,0.3363130093,-0.5569394231,0.0798316821,-0.0507644974,0.0824038908,0.2173671722,0.4450024068,-0.2298555672,0.1757664233,-0.1654862761,0.1924015433,0.1095912755,-0.3121576309,-0.5411019325,0.2140131146,-0.0612866841,0.1959047168,0.0259571318,0.2906033099,0.0093369419,0.1717624664,0.4052648544,0.2492511719,-0.0467647091,0.0787921399,-0.4085267782,-0.3879698217,0.3385632038,0.2016654164,-0.0047450308,-0.1988982856,0.1968246847,0.3221473992,0.2584798336,-0.2035795897,0.1287960708,0.0218389891,0.0705177039,-0.0129466308,0.2327663898,-0.2238298804,-0.4262207448,0.11144609,-0.3431405723,-0.0235893708,-0.3043866754,-0.361228615,-0.5231873989,0.0663694814,-0.171464175,-0.2400977015,0.115050301,-0.0613334365,0.0953495279,0.2569484115,0.1027859375,-0.0896757692,-0.2042666227,-0.0175576173,-0.4030902386,0.240516305,-0.308798492,-0.1786494255,-0.0221906919,0.2451284826,0.324608773,0.1002937332,0.0431751907,0.1558428705,0.1467771679,0.0931873098,-0.1324361414,-0.0177001171,0.2272260636,-0.2668777108,0.2040083557,0.1240440756,0.1868998408,-0.221017614,0.05095274,0.1322418749,-0.0019998513,0.1718758643,-0.0822376162,-0.0712248981,0.1570295691,-0.0034186426,0.3967843354,-0.1355208606,0.2599032819,0.4711607099,0.0430019759,0.1037084162,-0.0057747457,0.0103615662,0.1143461317,-0.0471609905,0.1531215012,0.1461279243,-0.3416619301,-0.2851566672,0.2188828588,-0.1442351341,0.0782735273,0.0937128961,0.0280694515,0.1398488432,-0.0689809099,-0.195103392,0.1316027343,-0.1178800836,0.4590720534,0.310067445,0.3028900325,-0.029343931,-0.3765672147,0.0593944713,0.0831138715,0.379365176,-0.3040897548,0.0794889852,-0.171332255,-0.243459627,-0.0772669688,-0.013044945,-0.2915118039,-0.317315042,0.0815705061,0.5144311786,0.0328679904,0.0099860048,-0.0540035702,-0.0480350479,0.0796879381,-0.1308909357,-0.0505061187,-0.7057309747,-0.3859112859,0.0051475652,0.2574861944,-0.1339800954,0.2502458096,-0.0254044235,-0.1927474737,-0.3766916692,-0.5513484478,-0.0036723248,-0.1104055196,0.3178443015,0.0176331513,0.3957713842,0.0727890357,-0.2152879387,0.2103869319,0.0130584994,-0.0248649139,0.2813790739,-0.3631691337,0.0844142213,-0.0089278743,-0.2284815609,-0.2832888961,-0.1489331275,0.1993967593,0.0630166978,0.1870315522,0.2383005023,0.0628918335,0.1012383327,0.2758947015,0.0676462352,-0.3717182577,-0.1608327031,0.3060976565,-0.2184316963,-0.2773063481,0.1912308037,0.002722593,0.1241865531,0.2216507792,-0.5899142623,0.0072065173,0.0857219473,0.3271362782,0.031964194,-0.1462098956,0.2367205918,0.0211273748,0.0012416313,0.062059287,0.0056531909,-0.1206239983,-0.1537077129,0.166241616,0.1140279919,0.2094682604,0.1647175252,0.7836135626,0.0867718309,0.0424960814,0.4263213873,-0.188306883,0.0635278597,-0.1462810636,-0.0966464877,0.0314848572,-0.3452987075,-0.2316510081,0.1231539845,-0.1210027635,-0.3868949413,-0.0544536486,-0.2450863421,-0.2544884384,-0.2394929677,0.0445744172,-0.0382827111,0.379373312,0.1044166535,-0.1685189754,-0.1514848173,-0.4425003827,-0.028786093,0.1463311911,-0.0292478688,-0.0367810801,-0.1005667672,-0.2029286176,-0.5161777735,0.4155936837,0.2859541476,0.1105859429,-0.1718751192,0.0878228843,0.4478865266,-0.1853179783,0.6596498489,-0.1708530486,0.0438121483,0.0984816179,0.1837503612,-0.7251459956,-0.0194136817,0.147544682,0.1342761517,0.215312928,-0.1166655123,-0.5139329433,0.0979648083,0.3441002071,0.3732801974,-0.0886276141,-0.1792948097,-0.191272229,-0.3767274618,-0.5138627887,-0.1649242193,-0.2111011744,0.2683063149,-0.025560908,0.031048581,0.050385315,0.201478228,-0.0147314733,0.2821478844,0.2859053016,0.0165023264,0.0235097762,0.0959309414,0.0991042778,0.3470714688,0.3539325595,0.2881973684,-0.0423728712,-0.0889987722,0.0853011161,-0.0638199598,0.18121171,-0.1502871662,0.0666161627,-0.0016658714,-0.1978694946,-0.0734471083,-0.1572476029,0.3747443855,-0.221284762,-0.2982092798,-0.2389822006,0.7085088491,0.1589902639,-0.0960436761,0.1231104955,-0.1088686809,-0.1373421997,0.365239948,0.1177065223,0.9276636839,-0.1522064656,0.328930378,0.3286547959,-0.0396991223,0.5044326186,-0.2825063169,0.1339882761,-0.484410882,0.2856452167,-0.0108886296,-0.0980820805,-0.1546719223,0.0310812201,-0.1827827245,0.0733063594,-0.2207050174,0.0885881856,0.0106324665,0.3567408025,-0.008351001,0.1106858179,-0.0842744187,0.0562638715,-0.3195542693,0.2571137249,-0.2615233958,-0.0009585342,0.0979617238,-0.1563698649,-0.4544543624,0.0785484239,-0.1875041425,0.4533356726,-0.306944102,-0.3188786209,0.0176066961,0.0598248057,-0.0411643572,-0.0937833935,-0.3611280918,0.1809431165,-0.2230480611,-0.1647396535,-0.1060880497,0.0703283325,0.2264247984,-0.0705450103,-0.4650467038,0.3104069829,-0.0331905186,-0.2690660954,-0.195166409,0.0056172167,0.0766634122,-0.1938509792,-0.1571667492,-0.2530135214,-0.1191873401,-0.1358085871,0.0734140873,-0.1295603067,0.0084917909,0.0646586269,0.0685643181,-0.3869736493,0.1155601069,0.4719053209,0.3767473102,0.1686637253,0.6754417419,0.2683519423,-0.2663931847,-0.0136300661,-0.3504727185,-0.3611333072,-0.3933808804,-0.0750969127,0.1329663992,0.4606651962,-0.2361188233,0.1057871878,0.1565126032,-0.1004546285,0.1289465129,-0.4669057727,-0.2856480777,0.3458832204,-0.1558524072,0.2150657028,0.333362639,0.1371670812,0.0941938311,0.2817735076,-0.2074543387,0.4754821956,-0.0808619931,0.3691472411,0.1022728831,-0.1045177504,-0.1664323509,0.0043666679,0.0278084781,0.0733071864,-0.0842073485,-0.187368691,-0.059740562,0.1769646406,0.0595366582,-0.2068469971,-0.1221330166,-0.4088256657,0.0580776408,-0.010986723,0.3590635657,0.1042840257,-0.0777881294,0.1733734012,0.1805440485,0.1427157372,-0.2289880365,0.3619517982,-0.1334508955,0.3957297802,0.2669553459,0.2385378629,-0.0947233662,-0.0126234489,-0.2449375242,0.0805034339,0.1154414788,-0.1886783838,0.185227707,-0.2776901424,-0.2030886859,-0.0213793032,0.0164039209,0.3396817148,-0.1417613029,-0.054867357,-0.0608212724,0.1005864516,-0.2835876942,0.046061974,0.485242784,-0.1776818633,-0.2785930634,0.0687629804,0.5847008824,-0.0438929535,0.1663225442,0.0466528051,0.2393534482,0.0115823224,0.0854468048,0.2219160944,0.1098196283,-0.1377399415,0.0567916296,0.2720851004,0.18669644,0.4535924494,-0.1657377928,0.0975392014,-0.1927057505,0.0472147912,-0.1411454231,-0.3601846993,0.1597557068,0.4120803773,0.1318735033,0.0065879552,0.0914900675,0.3713221848,0.6368659735,0.0102780368,-0.0483850986,0.3622556925,-0.1486780196,-0.1804165244,-0.2978949547,-0.3707225621,0.0285330173,0.1508214623,-0.2404500395,-0.2233104259,0.2616764009,0.185880214,-0.3208230734,-0.6472076178,0.5188295245,0.2831636965,0.1351263523,-0.222577095,0.2115809023,0.1248400882,-0.0737039596,0.1012161523,0.2894658446,0.6575102806,0.3211216331,-0.7122462988,0.0241689552,-0.1587850153,0.0964294225,-0.0956941545,-0.0255603027,0.1547429264,-0.0726363733,0.2175964713,0.0557836443,-0.0313376933,-0.1637457311,0.440110445,0.0479708463,-0.214529708,-0.2602033913,-0.1867547482,-0.2116843313,0.0047944435,0.088947244,-0.372079432,0.2248482704,0.4339295626,0.0074966205,0.1416673213,-0.0861911252,0.0382655524,-0.2804442644,0.4323863685,0.2422127724,0.1094457954,-0.4453440011,-0.2630242109,-0.6950753927,0.3619217277,-0.4275679588,-0.004696277,0.3281984031,0.1819252819,-0.2007462084,0.1665137261,0.072097376,-0.1435965896,-0.185102731,0.3347555101,-0.2243230045,-0.1052198634,-0.1288038343,-0.2936487794,-0.25168854,-0.3236352205,0.214736551,-0.194979921,-0.04772361,-0.2938646376,-0.1443579942,0.1363732815,0.1835709214,0.508359313,0.0855292305,0.6090764999,0.0791577175,-0.0309701562,-0.2001253963,-0.1561925411,0.1780349463,0.296132803,0.0426095761,0.1194765791,-0.0950099453,-0.1640064567,-0.1698608994,0.5059783459,-0.0187877361,0.0761009902,-0.0015461057,-0.2256751806,0.0126297828,0.3037437499,-0.141428262,0.1818352789,-0.0693844333,0.3124789596,-0.2922698855,-0.4456368685,0.4373099804,-0.4284289479,-0.3710170984,0.0440436043,0.1237263009,-0.1983818561,-0.1003204212,-0.3580924273,0.1454563588,0.4109820127,-0.0160897728,-0.1379513741,-0.1122561172,-0.0849029645,-0.1042083427,-0.2067106366,0.0315526053,0.0420022234,-0.1544650644,-0.3315714896,-0.5234953761]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2144","title":"Loading wikipedia 20200501.en throws pyarrow related error","comments":"I tried using `datasets` from `master` on macos with python 3.7.2\r\nI also have `requests==2.23.0` and `tqdm==4.45.0`.","body":"**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n","comment_length":17,"text":"Loading wikipedia 20200501.en throws pyarrow related error\n**Problem description**\r\nI am getting the following error when trying to load wikipedia\/20200501.en dataset.\r\n\r\n**Error log**\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.6k\/14.6k [00:00<00:00, 5.41MB\/s]\r\nDownloading: 59%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 10.7G\/18.3G [11:30<08:08, 15.5MB\/s]\r\nDataset wikipedia downloaded and prepared to \/usr\/local\/workspace\/NAS_NLP\/cache\/wikipedia\/20200501.en\/1.0.0\/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data.\r\nTraceback (most recent call last):\r\n File \"load_wiki.py\", line 2, in \r\n ds = load_dataset('wikipedia', '20200501.en', cache_dir='\/usr\/local\/workspace\/NAS_NLP\/cache')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 751, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 746, in as_dataset\r\n map_tuple=True,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 204, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 763, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py\", line 835, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 324, in read_table\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 544, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Expected to be able to read 9176784 bytes for message body, got 4918712\r\n\r\n**Detailed version info**\r\ndatasets==1.5.0\r\n - dataclasses [required: Any, installed: 0.8]\r\n - dill [required: Any, installed: 0.3.3]\r\n - fsspec [required: Any, installed: 0.8.7]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - huggingface-hub [required: <0.1.0, installed: 0.0.7]\r\n - filelock [required: Any, installed: 3.0.12]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - requests [required: Any, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: Any, installed: 4.49.0]\r\n - importlib-metadata [required: Any, installed: 1.7.0]\r\n - zipp [required: >=0.5, installed: 3.1.0]\r\n - multiprocess [required: Any, installed: 0.70.11.1]\r\n - dill [required: >=0.3.3, installed: 0.3.3]\r\n - numpy [required: >=1.17, installed: 1.17.0]\r\n - pandas [required: Any, installed: 1.1.5]\r\n - numpy [required: >=1.15.4, installed: 1.17.0]\r\n - python-dateutil [required: >=2.7.3, installed: 2.8.0]\r\n - six [required: >=1.5, installed: 1.15.0]\r\n - pytz [required: >=2017.2, installed: 2020.1]\r\n - pyarrow [required: >=0.17.1, installed: 3.0.0]\r\n - numpy [required: >=1.16.6, installed: 1.17.0]\r\n - requests [required: >=2.19.0, installed: 2.24.0]\r\n - certifi [required: >=2017.4.17, installed: 2020.6.20]\r\n - chardet [required: >=3.0.2,<4, installed: 3.0.4]\r\n - idna [required: >=2.5,<3, installed: 2.6]\r\n - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10]\r\n - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0]\r\n - xxhash [required: Any, installed: 2.0.0]\r\n\nI tried using `datasets` from `master` on macos with python 3.7.2\r\nI also have `requests==2.23.0` and `tqdm==4.45.0`.","embeddings":[-0.1041540205,0.3526462018,0.0340582058,0.3633103967,0.2745778561,0.1802078485,0.2762243748,0.4623250365,-0.04348813,-0.113784872,0.0247374158,0.236021921,0.1520678252,-0.1178016663,0.1841091067,-0.1683138758,0.0038767338,0.0068803728,0.0847920552,0.1328444481,-0.2001205236,0.1146743447,-0.1824860871,0.1555170566,-0.5857800245,0.0881166533,-0.0660176575,0.0540414415,0.0833243653,-0.613804698,0.3904819489,-0.0500538088,0.2010909319,0.2121308595,-0.0001204597,0.0749956965,0.4643647373,-0.0041334573,-0.2960134745,-0.0811445788,0.186580047,-0.0205053128,0.3180885315,-0.3607697487,0.1008782089,-0.2803391814,0.2242598385,0.0471996665,0.1832180023,0.0086322986,0.168285802,-0.1830526292,0.3375393748,0.0306443255,0.7112432122,-0.1374686658,0.0815054998,0.4424607456,0.1225971282,-0.1607152224,-0.186434865,0.23629798,-0.2502577007,0.1203090623,0.2671839893,-0.0770430341,0.3848769963,-0.2004953474,0.3921535015,0.2679577172,0.7762554288,-0.4092206359,0.1730330735,-0.2313123494,-0.0328712203,-0.0229594614,0.0938252583,0.2307374179,-0.27676934,0.0545302071,0.1571857184,-0.1188708618,-0.2609159052,0.301315248,-0.0642476678,0.1997214258,0.1399375498,0.3910686672,0.0163352489,-0.2268839478,0.0019353561,0.0632643551,0.1672586203,0.2208643556,-0.3931137919,0.107257776,-0.0683028772,0.1139047593,0.3778621256,-0.1693618745,-0.021596102,0.0659415349,0.2391710877,0.2064648122,0.4192265868,0.1241850182,-0.3237407506,-0.2987220883,0.2721662819,0.1988998652,-0.254214257,-0.0364414603,-0.0612123869,-0.1137952879,0.1385727972,-0.1709839702,0.4291738868,-0.2699913085,-0.220579192,0.1195769534,-0.1446794569,-0.0715681612,-0.1552301943,0.3867201805,-0.0916787088,0.2424993664,0.0672103316,0.0622379445,-0.1821437031,-0.1878142953,0.0057542627,0.1186774671,-0.2124670595,0.1797935814,0.2939305305,-0.1538039595,0.2583542168,0.1435221881,-0.1506101638,-0.1206758171,0.1700648069,0.0130382087,-0.2326727659,0.309833169,0.0075005512,0.2468717843,-0.0328951664,-0.1449694484,-0.1790345013,0.3182585537,-0.2848282754,-0.5152549744,-0.262249738,0.1209049523,-0.0116632255,0.0844510719,0.1616870016,0.0576540641,0.4413489699,-0.5136592984,-0.1003230512,-0.1257850081,-0.1855283529,-0.4527531266,0.1706373543,0.3363130093,-0.5569394231,0.0798316821,-0.0507644974,0.0824038908,0.2173671722,0.4450024068,-0.2298555672,0.1757664233,-0.1654862761,0.1924015433,0.1095912755,-0.3121576309,-0.5411019325,0.2140131146,-0.0612866841,0.1959047168,0.0259571318,0.2906033099,0.0093369419,0.1717624664,0.4052648544,0.2492511719,-0.0467647091,0.0787921399,-0.4085267782,-0.3879698217,0.3385632038,0.2016654164,-0.0047450308,-0.1988982856,0.1968246847,0.3221473992,0.2584798336,-0.2035795897,0.1287960708,0.0218389891,0.0705177039,-0.0129466308,0.2327663898,-0.2238298804,-0.4262207448,0.11144609,-0.3431405723,-0.0235893708,-0.3043866754,-0.361228615,-0.5231873989,0.0663694814,-0.171464175,-0.2400977015,0.115050301,-0.0613334365,0.0953495279,0.2569484115,0.1027859375,-0.0896757692,-0.2042666227,-0.0175576173,-0.4030902386,0.240516305,-0.308798492,-0.1786494255,-0.0221906919,0.2451284826,0.324608773,0.1002937332,0.0431751907,0.1558428705,0.1467771679,0.0931873098,-0.1324361414,-0.0177001171,0.2272260636,-0.2668777108,0.2040083557,0.1240440756,0.1868998408,-0.221017614,0.05095274,0.1322418749,-0.0019998513,0.1718758643,-0.0822376162,-0.0712248981,0.1570295691,-0.0034186426,0.3967843354,-0.1355208606,0.2599032819,0.4711607099,0.0430019759,0.1037084162,-0.0057747457,0.0103615662,0.1143461317,-0.0471609905,0.1531215012,0.1461279243,-0.3416619301,-0.2851566672,0.2188828588,-0.1442351341,0.0782735273,0.0937128961,0.0280694515,0.1398488432,-0.0689809099,-0.195103392,0.1316027343,-0.1178800836,0.4590720534,0.310067445,0.3028900325,-0.029343931,-0.3765672147,0.0593944713,0.0831138715,0.379365176,-0.3040897548,0.0794889852,-0.171332255,-0.243459627,-0.0772669688,-0.013044945,-0.2915118039,-0.317315042,0.0815705061,0.5144311786,0.0328679904,0.0099860048,-0.0540035702,-0.0480350479,0.0796879381,-0.1308909357,-0.0505061187,-0.7057309747,-0.3859112859,0.0051475652,0.2574861944,-0.1339800954,0.2502458096,-0.0254044235,-0.1927474737,-0.3766916692,-0.5513484478,-0.0036723248,-0.1104055196,0.3178443015,0.0176331513,0.3957713842,0.0727890357,-0.2152879387,0.2103869319,0.0130584994,-0.0248649139,0.2813790739,-0.3631691337,0.0844142213,-0.0089278743,-0.2284815609,-0.2832888961,-0.1489331275,0.1993967593,0.0630166978,0.1870315522,0.2383005023,0.0628918335,0.1012383327,0.2758947015,0.0676462352,-0.3717182577,-0.1608327031,0.3060976565,-0.2184316963,-0.2773063481,0.1912308037,0.002722593,0.1241865531,0.2216507792,-0.5899142623,0.0072065173,0.0857219473,0.3271362782,0.031964194,-0.1462098956,0.2367205918,0.0211273748,0.0012416313,0.062059287,0.0056531909,-0.1206239983,-0.1537077129,0.166241616,0.1140279919,0.2094682604,0.1647175252,0.7836135626,0.0867718309,0.0424960814,0.4263213873,-0.188306883,0.0635278597,-0.1462810636,-0.0966464877,0.0314848572,-0.3452987075,-0.2316510081,0.1231539845,-0.1210027635,-0.3868949413,-0.0544536486,-0.2450863421,-0.2544884384,-0.2394929677,0.0445744172,-0.0382827111,0.379373312,0.1044166535,-0.1685189754,-0.1514848173,-0.4425003827,-0.028786093,0.1463311911,-0.0292478688,-0.0367810801,-0.1005667672,-0.2029286176,-0.5161777735,0.4155936837,0.2859541476,0.1105859429,-0.1718751192,0.0878228843,0.4478865266,-0.1853179783,0.6596498489,-0.1708530486,0.0438121483,0.0984816179,0.1837503612,-0.7251459956,-0.0194136817,0.147544682,0.1342761517,0.215312928,-0.1166655123,-0.5139329433,0.0979648083,0.3441002071,0.3732801974,-0.0886276141,-0.1792948097,-0.191272229,-0.3767274618,-0.5138627887,-0.1649242193,-0.2111011744,0.2683063149,-0.025560908,0.031048581,0.050385315,0.201478228,-0.0147314733,0.2821478844,0.2859053016,0.0165023264,0.0235097762,0.0959309414,0.0991042778,0.3470714688,0.3539325595,0.2881973684,-0.0423728712,-0.0889987722,0.0853011161,-0.0638199598,0.18121171,-0.1502871662,0.0666161627,-0.0016658714,-0.1978694946,-0.0734471083,-0.1572476029,0.3747443855,-0.221284762,-0.2982092798,-0.2389822006,0.7085088491,0.1589902639,-0.0960436761,0.1231104955,-0.1088686809,-0.1373421997,0.365239948,0.1177065223,0.9276636839,-0.1522064656,0.328930378,0.3286547959,-0.0396991223,0.5044326186,-0.2825063169,0.1339882761,-0.484410882,0.2856452167,-0.0108886296,-0.0980820805,-0.1546719223,0.0310812201,-0.1827827245,0.0733063594,-0.2207050174,0.0885881856,0.0106324665,0.3567408025,-0.008351001,0.1106858179,-0.0842744187,0.0562638715,-0.3195542693,0.2571137249,-0.2615233958,-0.0009585342,0.0979617238,-0.1563698649,-0.4544543624,0.0785484239,-0.1875041425,0.4533356726,-0.306944102,-0.3188786209,0.0176066961,0.0598248057,-0.0411643572,-0.0937833935,-0.3611280918,0.1809431165,-0.2230480611,-0.1647396535,-0.1060880497,0.0703283325,0.2264247984,-0.0705450103,-0.4650467038,0.3104069829,-0.0331905186,-0.2690660954,-0.195166409,0.0056172167,0.0766634122,-0.1938509792,-0.1571667492,-0.2530135214,-0.1191873401,-0.1358085871,0.0734140873,-0.1295603067,0.0084917909,0.0646586269,0.0685643181,-0.3869736493,0.1155601069,0.4719053209,0.3767473102,0.1686637253,0.6754417419,0.2683519423,-0.2663931847,-0.0136300661,-0.3504727185,-0.3611333072,-0.3933808804,-0.0750969127,0.1329663992,0.4606651962,-0.2361188233,0.1057871878,0.1565126032,-0.1004546285,0.1289465129,-0.4669057727,-0.2856480777,0.3458832204,-0.1558524072,0.2150657028,0.333362639,0.1371670812,0.0941938311,0.2817735076,-0.2074543387,0.4754821956,-0.0808619931,0.3691472411,0.1022728831,-0.1045177504,-0.1664323509,0.0043666679,0.0278084781,0.0733071864,-0.0842073485,-0.187368691,-0.059740562,0.1769646406,0.0595366582,-0.2068469971,-0.1221330166,-0.4088256657,0.0580776408,-0.010986723,0.3590635657,0.1042840257,-0.0777881294,0.1733734012,0.1805440485,0.1427157372,-0.2289880365,0.3619517982,-0.1334508955,0.3957297802,0.2669553459,0.2385378629,-0.0947233662,-0.0126234489,-0.2449375242,0.0805034339,0.1154414788,-0.1886783838,0.185227707,-0.2776901424,-0.2030886859,-0.0213793032,0.0164039209,0.3396817148,-0.1417613029,-0.054867357,-0.0608212724,0.1005864516,-0.2835876942,0.046061974,0.485242784,-0.1776818633,-0.2785930634,0.0687629804,0.5847008824,-0.0438929535,0.1663225442,0.0466528051,0.2393534482,0.0115823224,0.0854468048,0.2219160944,0.1098196283,-0.1377399415,0.0567916296,0.2720851004,0.18669644,0.4535924494,-0.1657377928,0.0975392014,-0.1927057505,0.0472147912,-0.1411454231,-0.3601846993,0.1597557068,0.4120803773,0.1318735033,0.0065879552,0.0914900675,0.3713221848,0.6368659735,0.0102780368,-0.0483850986,0.3622556925,-0.1486780196,-0.1804165244,-0.2978949547,-0.3707225621,0.0285330173,0.1508214623,-0.2404500395,-0.2233104259,0.2616764009,0.185880214,-0.3208230734,-0.6472076178,0.5188295245,0.2831636965,0.1351263523,-0.222577095,0.2115809023,0.1248400882,-0.0737039596,0.1012161523,0.2894658446,0.6575102806,0.3211216331,-0.7122462988,0.0241689552,-0.1587850153,0.0964294225,-0.0956941545,-0.0255603027,0.1547429264,-0.0726363733,0.2175964713,0.0557836443,-0.0313376933,-0.1637457311,0.440110445,0.0479708463,-0.214529708,-0.2602033913,-0.1867547482,-0.2116843313,0.0047944435,0.088947244,-0.372079432,0.2248482704,0.4339295626,0.0074966205,0.1416673213,-0.0861911252,0.0382655524,-0.2804442644,0.4323863685,0.2422127724,0.1094457954,-0.4453440011,-0.2630242109,-0.6950753927,0.3619217277,-0.4275679588,-0.004696277,0.3281984031,0.1819252819,-0.2007462084,0.1665137261,0.072097376,-0.1435965896,-0.185102731,0.3347555101,-0.2243230045,-0.1052198634,-0.1288038343,-0.2936487794,-0.25168854,-0.3236352205,0.214736551,-0.194979921,-0.04772361,-0.2938646376,-0.1443579942,0.1363732815,0.1835709214,0.508359313,0.0855292305,0.6090764999,0.0791577175,-0.0309701562,-0.2001253963,-0.1561925411,0.1780349463,0.296132803,0.0426095761,0.1194765791,-0.0950099453,-0.1640064567,-0.1698608994,0.5059783459,-0.0187877361,0.0761009902,-0.0015461057,-0.2256751806,0.0126297828,0.3037437499,-0.141428262,0.1818352789,-0.0693844333,0.3124789596,-0.2922698855,-0.4456368685,0.4373099804,-0.4284289479,-0.3710170984,0.0440436043,0.1237263009,-0.1983818561,-0.1003204212,-0.3580924273,0.1454563588,0.4109820127,-0.0160897728,-0.1379513741,-0.1122561172,-0.0849029645,-0.1042083427,-0.2067106366,0.0315526053,0.0420022234,-0.1544650644,-0.3315714896,-0.5234953761]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2139","title":"TypeError when using save_to_disk in a dataset loaded with ReadInstruction split","comments":"Hi !\r\nI think this has been fixed recently on `master`.\r\nCan you try again by installing `datasets` from `master` ?\r\n```\r\npip install git+https:\/\/github.com\/huggingface\/datasets.git\r\n```","body":"Hi,\r\n\r\nLoading a dataset with `load_dataset` using a split defined via `ReadInstruction` and then saving it to disk results in the following error: `TypeError: Object of type ReadInstruction is not JSON serializable`.\r\n\r\nHere is the minimal reproducible example:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nfrom datasets import ReadInstruction\r\n\r\ndata_1 = load_dataset(\r\n \"wikiann\",\r\n \"en\",\r\n split=\"validation\",\r\n)\r\n\r\ndata_1.save_to_disk(\"temporary_path_1\")\r\n\r\nprint(\"Save with regular split works.\")\r\n\r\ndata_2 = load_dataset(\r\n \"wikiann\",\r\n \"en\",\r\n split=ReadInstruction(\"validation\", to=50, unit=\"%\"),\r\n)\r\n\r\ndata_2.save_to_disk(\"temporary_path_2\")\r\n```\r\n\r\nand the corresponding output:\r\n\r\n```\r\nReusing dataset wikiann (\/xxxxx\/.cache\/huggingface\/datasets\/wikiann\/en\/1.1.0\/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9)\r\nSave with regular split works.\r\nReusing dataset wikiann (\/xxxxx\/.cache\/huggingface\/datasets\/wikiann\/en\/1.1.0\/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9)\r\nTraceback (most recent call last):\r\n File \"bug.py\", line 20, in \r\n data_2.save_to_disk(\"temporary_path_2\")\r\n File \"\/xxxxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 645, in save_to_disk\r\n json.dump(state, state_file, indent=2, sort_keys=True)\r\n File \"\/usr\/lib\/python3.7\/json\/__init__.py\", line 179, in dump\r\n for chunk in iterable:\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 431, in _iterencode\r\n yield from _iterencode_dict(o, _current_indent_level)\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 405, in _iterencode_dict\r\n yield from chunks\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 438, in _iterencode\r\n o = _default(o)\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 179, in default\r\n raise TypeError(f'Object of type {o.__class__.__name__} '\r\nTypeError: Object of type ReadInstruction is not JSON serializable\r\n```\r\n\r\nLet me know if there is some misuse from my end.\r\n\r\nThanks in advance.\r\n ","comment_length":26,"text":"TypeError when using save_to_disk in a dataset loaded with ReadInstruction split\nHi,\r\n\r\nLoading a dataset with `load_dataset` using a split defined via `ReadInstruction` and then saving it to disk results in the following error: `TypeError: Object of type ReadInstruction is not JSON serializable`.\r\n\r\nHere is the minimal reproducible example:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nfrom datasets import ReadInstruction\r\n\r\ndata_1 = load_dataset(\r\n \"wikiann\",\r\n \"en\",\r\n split=\"validation\",\r\n)\r\n\r\ndata_1.save_to_disk(\"temporary_path_1\")\r\n\r\nprint(\"Save with regular split works.\")\r\n\r\ndata_2 = load_dataset(\r\n \"wikiann\",\r\n \"en\",\r\n split=ReadInstruction(\"validation\", to=50, unit=\"%\"),\r\n)\r\n\r\ndata_2.save_to_disk(\"temporary_path_2\")\r\n```\r\n\r\nand the corresponding output:\r\n\r\n```\r\nReusing dataset wikiann (\/xxxxx\/.cache\/huggingface\/datasets\/wikiann\/en\/1.1.0\/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9)\r\nSave with regular split works.\r\nReusing dataset wikiann (\/xxxxx\/.cache\/huggingface\/datasets\/wikiann\/en\/1.1.0\/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9)\r\nTraceback (most recent call last):\r\n File \"bug.py\", line 20, in \r\n data_2.save_to_disk(\"temporary_path_2\")\r\n File \"\/xxxxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 645, in save_to_disk\r\n json.dump(state, state_file, indent=2, sort_keys=True)\r\n File \"\/usr\/lib\/python3.7\/json\/__init__.py\", line 179, in dump\r\n for chunk in iterable:\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 431, in _iterencode\r\n yield from _iterencode_dict(o, _current_indent_level)\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 405, in _iterencode_dict\r\n yield from chunks\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 438, in _iterencode\r\n o = _default(o)\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 179, in default\r\n raise TypeError(f'Object of type {o.__class__.__name__} '\r\nTypeError: Object of type ReadInstruction is not JSON serializable\r\n```\r\n\r\nLet me know if there is some misuse from my end.\r\n\r\nThanks in advance.\r\n \nHi !\r\nI think this has been fixed recently on `master`.\r\nCan you try again by installing `datasets` from `master` ?\r\n```\r\npip install git+https:\/\/github.com\/huggingface\/datasets.git\r\n```","embeddings":[-0.1317065656,0.2105600834,0.0438051894,0.3900926709,0.3082846701,0.2542080283,0.4115675688,0.2183050513,0.1559664458,0.0927028954,-0.1810030639,0.3821820319,-0.1831880361,0.4692062438,-0.3343833089,-0.2261848152,0.1314814836,0.0324933976,0.0247836336,0.1516534537,-0.2917743325,0.2401537746,-0.2096995413,-0.0722319856,-0.1023984179,-0.2237329334,-0.0900937691,0.1784162223,-0.0479017384,-0.270075351,0.4981570244,-0.0863716081,0.2484552562,0.4595005214,-0.0001193325,0.2449737489,0.2752628922,-0.1217553765,-0.4869103432,-0.1442823708,-0.112406753,-0.3307709694,0.2289606631,-0.2849002481,0.2563339472,0.032430876,-0.096720852,-0.2834960222,0.5513250828,-0.1404765546,0.1310997903,0.2201003879,0.2116705775,-0.0764774606,-0.0196526162,0.818102777,-0.3116717935,0.1208766252,-0.1681930125,0.1594806761,0.1860604286,0.139774695,0.077550739,-0.1289093494,0.1089108884,-0.0839564577,0.0414062925,-0.1637930125,-0.0721092597,0.0359598212,0.6240739226,-0.2325592786,-0.5413184166,-0.5269168019,-0.0988926366,0.1037318781,0.1883218139,0.0515774116,0.1023936495,0.1159397885,-0.0749133602,-0.1888535172,-0.1452777237,-0.0556888357,-0.0028690889,-0.3264967203,-0.2233644128,0.3309184015,0.2039163411,-0.0945509523,0.1549262106,-0.2039530426,-0.092710495,-0.0095436256,-0.2591286004,-0.2582912147,-0.463026464,-0.1182582676,0.0588776506,0.2336980402,0.108112976,-0.1332108825,-0.2215162218,0.1171837002,0.6883972287,0.1997386664,0.1489496082,0.402256608,0.029157605,0.1394936591,0.0054768277,-0.0856491104,0.166863516,-0.07627327,0.3958076239,0.0007771892,0.2863453031,-0.1103602871,-0.5084636807,-0.023641672,-0.3658467531,-0.2045274526,-0.0995762274,0.112592347,0.2994084954,0.2819698453,0.1736928523,0.4408884645,-0.0008400108,-0.095121637,-0.0727198273,-0.0292542335,0.0549812205,0.2533734143,0.0195076615,-0.0273443554,0.0528054126,0.0067487173,-0.4714187086,-0.2150871903,0.1913157701,-0.2646383345,0.0638191029,0.2851462364,0.2803843021,0.1533473432,0.1773774326,-0.3570280969,-0.3106738031,0.1204815879,-0.0965005457,-0.0488858037,-0.0437003672,0.0947650969,-0.1267966479,0.0659134686,-0.5947527885,-0.0187632609,0.4436030686,-0.2635980844,-0.0719171688,-0.2288510203,-0.0116031589,-0.3403199315,0.0138408691,0.4004660845,-0.3379040956,-0.0681683049,-0.2169623524,-0.2334198356,0.1664637476,0.4936624169,-0.0835335925,0.516941011,-0.1955942512,0.3982065916,0.1584657282,-0.2379556447,-0.2000245303,0.5022702217,-0.1037771031,0.1734888703,-0.0952558294,-0.0241522361,0.4080353379,-0.173525393,0.1406641603,0.0812345371,-0.0405478813,-0.0777483135,-0.194061473,-0.2555298805,0.1296404302,0.0126433447,-0.1006332636,0.1342628598,0.1260567456,0.1880754828,0.3111828566,-0.0541514978,0.2493341416,0.3054505289,0.117798306,-0.1352878213,-0.1075367704,-0.0221280865,-0.3029607236,-0.0044431854,0.2780956626,-0.2112738043,-0.0934908092,-0.1717114598,-0.2598276436,0.230335027,-0.2763422132,0.2911271751,0.0529794879,0.1208057478,-0.0248321537,0.2489490956,-0.0919569507,0.1896861494,-0.035242334,0.1281069666,-0.4389435053,0.4238277376,-0.0409120433,-0.3451372385,0.0302613918,0.0569884367,0.2687718272,-0.134433791,-0.0043099276,0.3361161947,0.2367794812,0.13005279,-0.2204691917,-0.0598820932,0.2161560059,-0.0719890445,-0.1192336306,0.2150163651,0.3863373101,-0.1784895062,-0.1886806488,0.3041788638,-0.0705867633,0.1873155981,0.0063082632,0.112623103,0.0241007935,-0.002370018,-0.0940756425,-0.4081260264,0.0147502823,0.0237155166,-0.1132685915,-0.2551979721,-0.0995129123,0.1117107123,0.4493326843,-0.2675612867,0.2195772082,0.0182680245,-0.1571053565,-0.0029315124,-0.0543791577,0.4611753821,0.5188760757,-0.0728522763,0.2165549845,-0.1663918197,0.0560957193,-0.0409083143,0.1319316179,0.1509232819,0.3151162267,0.2516278028,0.0549508184,0.0588581562,-0.0803973228,0.149934113,-0.0055816928,0.1412248909,-0.404068172,0.1511809379,-0.2129691243,0.3331795335,-0.1819794029,-0.2257498801,0.0959367529,-0.6485622525,-0.4150827229,0.4897378385,-0.1861340553,-0.039807003,0.0067072879,-0.0427006856,-0.0165362135,-0.0674778298,-0.124694787,0.1129419208,-0.2868056893,-0.0729240328,0.2463184297,-0.2035422623,0.2206840366,-0.1972499937,0.0126172695,-0.2052572817,0.0368248969,0.1174176559,-0.1562771499,0.1773216873,0.1652359366,0.2820571661,0.0242906269,-0.2725773752,0.0964507759,0.1477789581,-0.1811760366,0.1133565828,-0.0025246169,-0.0518953949,-0.1257267296,-0.158946678,-0.3489079773,-0.1871428788,0.5054351687,-0.1241575256,-0.2554841042,-0.1415614784,0.1180824488,0.1391178817,0.2809080482,-0.2997899354,-0.1871420145,-0.129202038,0.3373076916,-0.0742208287,-0.1497434825,0.1376183778,0.04738646,-0.2108744681,-0.0256522689,-0.4338174462,0.2235617787,-0.1067928895,0.041715987,-0.1858791709,0.0072100954,0.2149301469,0.1670225263,0.119503133,-0.0201800372,-0.4066403508,0.2743183672,-0.1029174104,0.2617074251,-0.3179399371,0.3655718863,0.2267260849,-0.0936277136,0.1042671725,0.1052585691,0.4074470103,0.4023137391,0.2292523235,-0.2795910835,-0.0465680361,-0.0305438284,-0.4651505351,-0.3055941761,-0.0788278282,0.0058602551,-0.4329887033,0.0954170525,-0.416962713,-0.0798513517,-0.4240089357,-0.064945735,0.0411414132,0.2644873261,-0.1293402761,0.3300028443,0.121315293,0.2686508298,0.0854175612,0.4426595569,0.304084599,-0.0225938745,-0.6052621007,-0.3987894952,-0.1494777501,0.085171327,-0.0030239439,0.3307610452,-0.1128201112,-0.158019945,-0.068572633,-0.0389397144,0.5201874971,-0.0512404703,0.0012258509,0.1916232407,0.0426360592,-0.3706858754,-0.0845857188,0.1276453584,0.261156112,0.0996290743,0.7260839939,-0.2869675756,0.0848333985,0.0986012891,0.3608519137,-0.2686014771,-0.2606610656,0.0762175992,-0.6618020535,-0.473454237,-0.0654048696,0.0891384482,0.0840258971,0.0445675887,0.0613792241,-0.2863175273,-0.1974561363,-0.030964205,0.0094024306,0.2691719532,-0.1909771711,0.1744082868,0.273471117,0.0785733461,0.4405231774,0.4388534129,-0.0147153819,-0.352555275,0.1523869783,-0.0716162622,0.0187627245,0.1555835307,0.0508779213,-0.0821264014,0.0883737728,-0.2365548164,-0.4633461237,-0.1488452405,0.0904251486,-0.3374938071,-0.6183341146,-0.2228628546,0.3830707669,0.0004322185,0.0913360044,-0.0558939688,0.2712638974,-0.2566220462,0.4474663734,0.1223123968,0.6830673814,0.0007449057,0.0854566023,0.4916387796,-0.6043204069,-0.0060888282,0.1641178727,-0.0600038543,-0.4265597761,0.1930721849,-0.0806793496,-0.0681977421,0.1394303888,-0.0234502982,-0.3511241376,-0.115192242,-0.3808244169,-0.205246985,-0.1649704576,0.1392966807,-0.4981330931,-0.3532392979,-0.1283452064,0.1440775543,0.0093873618,-0.0194945317,0.140739888,-0.3155276179,-0.0234527495,-0.0966568813,-0.3905836344,0.0247118045,-0.0757293254,0.3466628492,-0.1778202206,-0.1701766551,0.2836356759,0.2573932409,0.514473021,0.0030206104,-0.0201175436,-0.1384010911,-0.2794148922,0.0727894604,-0.017847985,-0.235719949,0.4342162609,0.1702372879,0.0195976123,-0.2134186029,-0.1493427455,-0.1292095184,-0.1325327307,0.0547382496,-0.224358201,-0.5140143633,-0.2109909058,-0.1133698747,0.0389370359,-0.3093291819,0.012805759,0.1201058105,-0.0274823904,0.1266261339,-0.3678091764,-0.0474945344,0.0045270314,0.3145180345,0.2062477767,0.2634675801,0.6696696877,-0.0583996847,-0.1212425306,-0.2391156405,0.7491883039,0.2670490444,-0.0332002267,0.3828852177,-0.3566769361,0.0791084021,-0.1761876494,0.2747239172,0.2144100964,-0.3028134406,-0.0459919684,-0.4061584771,-0.3587571084,0.2031573206,0.0973908529,0.2266011536,0.1682340205,0.1613412797,-0.2150634527,-0.0841039568,-0.148704946,-0.0051133367,-0.2340533882,-0.0138233341,-0.0079671498,-0.0616992116,0.2066635936,-0.1218830571,0.0526377968,0.202642262,0.0701572746,-0.0461211875,-0.1765582114,0.1328735501,0.1106576174,0.0928322822,-0.128450051,-0.2062306404,0.0730832443,-0.138905406,-0.1202656105,0.2732168436,-0.0864552036,0.1101714894,-0.0113607114,0.3906763494,-0.1400344819,0.1156818345,-0.0184722356,0.1907476485,-0.0251064748,0.2376375496,0.0582986288,0.0307972077,-0.3982856572,0.0157231875,0.3234840035,0.3239723444,0.2735621929,-0.5291035771,0.0136452671,-0.1279325783,0.1995225102,0.3046544194,-0.2359331995,-0.0323292315,-0.0842189714,0.0597508997,0.2379079908,-0.124497585,0.3456408978,-0.3701761067,-0.2507579327,0.194592312,-0.0075365012,0.226304695,-0.5040649772,0.0490179099,0.5732997656,0.0245195534,0.2739393413,0.5586566329,-0.2972379029,0.2062015533,-0.0551292561,-0.0303288829,0.2992989421,0.148515299,-0.0736934394,0.3769677579,0.2046097368,0.1215097234,0.0530814901,-0.191453293,-0.0007979974,-0.0545356646,0.1000757143,0.3828053474,0.1555067599,-0.0381457657,-0.6030409336,-0.1078897268,-0.1749638319,0.0733095482,-0.2203655541,-0.0936853662,-0.0803633779,-0.2421419472,-0.0503131151,-0.0172618218,-0.1793982983,-0.1467364579,0.2449443489,-0.052128762,0.0955899954,-0.119679518,0.0544791631,0.1570000648,0.0720025674,-0.3394526541,-0.1150503308,0.0763376132,-0.0231554266,-0.0515820123,0.0262633376,0.4611822665,0.2230966091,-0.299393028,-0.168780759,-0.0456724763,-0.0893872008,0.194780454,0.1658117026,-0.2169560939,0.1562274396,0.589836359,0.0546889901,0.0335577279,0.2126052082,0.2061872333,0.167957589,-0.1390759796,0.373916775,-0.4525499344,-0.1057959199,0.0817324072,-0.0233930927,-0.1923722029,-0.0686579868,0.2831576467,-0.0889566764,0.2677441835,0.1079684049,0.0300730485,0.0063130576,0.4416630268,0.2064991742,-0.3510468304,-0.1622361094,0.1728037894,-0.1454878896,-0.1226277798,-0.0181306694,0.0106476657,-0.0071235853,0.1739094704,0.1665398031,0.1428555399,0.0265425052,0.0155514749,0.1411405951,0.4424415231,0.0177942589,-0.3375852108,0.5128375292,-0.1091878042,-0.0130157769,-0.4200245142,0.2337831259,0.3612352908,-0.0040313802,0.0197012033,-0.036572855,-0.0331173651,0.0064799469,0.1603878736,0.4409674704,0.2313555628,-0.0048692822,-0.3268092573,0.0847199857,-0.1212448403,-0.1091054529,0.1739548892,-0.1156606078,0.7781214118,-0.3791902065,0.4145943224,-0.2927051187,-0.2153362185,-0.0463361219,0.1783768833,-0.4962359369,0.2853748798,-0.0774828047,0.0432463624,0.3445824981,0.1797284037,0.0849786848,0.3774322867,-0.1755812764,-0.0792290568,0.3478639722,-0.0715023652,-0.3467016518,0.4105906487,-0.1564235538,0.349638164,0.2499444485,-0.6930737495,-0.066903837,0.3843768239,-0.2654726803,-0.1002497748,0.0628135875,0.2828069031,-0.1878263652,0.0814793333,0.4868473709,0.0644614995,-0.2565093935,0.3085873127,-0.0935587063]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2139","title":"TypeError when using save_to_disk in a dataset loaded with ReadInstruction split","comments":"Hi!\r\n\r\nUsing that version of the code solves the issue. Thanks!","body":"Hi,\r\n\r\nLoading a dataset with `load_dataset` using a split defined via `ReadInstruction` and then saving it to disk results in the following error: `TypeError: Object of type ReadInstruction is not JSON serializable`.\r\n\r\nHere is the minimal reproducible example:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nfrom datasets import ReadInstruction\r\n\r\ndata_1 = load_dataset(\r\n \"wikiann\",\r\n \"en\",\r\n split=\"validation\",\r\n)\r\n\r\ndata_1.save_to_disk(\"temporary_path_1\")\r\n\r\nprint(\"Save with regular split works.\")\r\n\r\ndata_2 = load_dataset(\r\n \"wikiann\",\r\n \"en\",\r\n split=ReadInstruction(\"validation\", to=50, unit=\"%\"),\r\n)\r\n\r\ndata_2.save_to_disk(\"temporary_path_2\")\r\n```\r\n\r\nand the corresponding output:\r\n\r\n```\r\nReusing dataset wikiann (\/xxxxx\/.cache\/huggingface\/datasets\/wikiann\/en\/1.1.0\/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9)\r\nSave with regular split works.\r\nReusing dataset wikiann (\/xxxxx\/.cache\/huggingface\/datasets\/wikiann\/en\/1.1.0\/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9)\r\nTraceback (most recent call last):\r\n File \"bug.py\", line 20, in \r\n data_2.save_to_disk(\"temporary_path_2\")\r\n File \"\/xxxxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 645, in save_to_disk\r\n json.dump(state, state_file, indent=2, sort_keys=True)\r\n File \"\/usr\/lib\/python3.7\/json\/__init__.py\", line 179, in dump\r\n for chunk in iterable:\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 431, in _iterencode\r\n yield from _iterencode_dict(o, _current_indent_level)\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 405, in _iterencode_dict\r\n yield from chunks\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 438, in _iterencode\r\n o = _default(o)\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 179, in default\r\n raise TypeError(f'Object of type {o.__class__.__name__} '\r\nTypeError: Object of type ReadInstruction is not JSON serializable\r\n```\r\n\r\nLet me know if there is some misuse from my end.\r\n\r\nThanks in advance.\r\n ","comment_length":11,"text":"TypeError when using save_to_disk in a dataset loaded with ReadInstruction split\nHi,\r\n\r\nLoading a dataset with `load_dataset` using a split defined via `ReadInstruction` and then saving it to disk results in the following error: `TypeError: Object of type ReadInstruction is not JSON serializable`.\r\n\r\nHere is the minimal reproducible example:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nfrom datasets import ReadInstruction\r\n\r\ndata_1 = load_dataset(\r\n \"wikiann\",\r\n \"en\",\r\n split=\"validation\",\r\n)\r\n\r\ndata_1.save_to_disk(\"temporary_path_1\")\r\n\r\nprint(\"Save with regular split works.\")\r\n\r\ndata_2 = load_dataset(\r\n \"wikiann\",\r\n \"en\",\r\n split=ReadInstruction(\"validation\", to=50, unit=\"%\"),\r\n)\r\n\r\ndata_2.save_to_disk(\"temporary_path_2\")\r\n```\r\n\r\nand the corresponding output:\r\n\r\n```\r\nReusing dataset wikiann (\/xxxxx\/.cache\/huggingface\/datasets\/wikiann\/en\/1.1.0\/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9)\r\nSave with regular split works.\r\nReusing dataset wikiann (\/xxxxx\/.cache\/huggingface\/datasets\/wikiann\/en\/1.1.0\/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9)\r\nTraceback (most recent call last):\r\n File \"bug.py\", line 20, in \r\n data_2.save_to_disk(\"temporary_path_2\")\r\n File \"\/xxxxx\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 645, in save_to_disk\r\n json.dump(state, state_file, indent=2, sort_keys=True)\r\n File \"\/usr\/lib\/python3.7\/json\/__init__.py\", line 179, in dump\r\n for chunk in iterable:\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 431, in _iterencode\r\n yield from _iterencode_dict(o, _current_indent_level)\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 405, in _iterencode_dict\r\n yield from chunks\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 438, in _iterencode\r\n o = _default(o)\r\n File \"\/usr\/lib\/python3.7\/json\/encoder.py\", line 179, in default\r\n raise TypeError(f'Object of type {o.__class__.__name__} '\r\nTypeError: Object of type ReadInstruction is not JSON serializable\r\n```\r\n\r\nLet me know if there is some misuse from my end.\r\n\r\nThanks in advance.\r\n \nHi!\r\n\r\nUsing that version of the code solves the issue. Thanks!","embeddings":[-0.1317065656,0.2105600834,0.0438051894,0.3900926709,0.3082846701,0.2542080283,0.4115675688,0.2183050513,0.1559664458,0.0927028954,-0.1810030639,0.3821820319,-0.1831880361,0.4692062438,-0.3343833089,-0.2261848152,0.1314814836,0.0324933976,0.0247836336,0.1516534537,-0.2917743325,0.2401537746,-0.2096995413,-0.0722319856,-0.1023984179,-0.2237329334,-0.0900937691,0.1784162223,-0.0479017384,-0.270075351,0.4981570244,-0.0863716081,0.2484552562,0.4595005214,-0.0001193325,0.2449737489,0.2752628922,-0.1217553765,-0.4869103432,-0.1442823708,-0.112406753,-0.3307709694,0.2289606631,-0.2849002481,0.2563339472,0.032430876,-0.096720852,-0.2834960222,0.5513250828,-0.1404765546,0.1310997903,0.2201003879,0.2116705775,-0.0764774606,-0.0196526162,0.818102777,-0.3116717935,0.1208766252,-0.1681930125,0.1594806761,0.1860604286,0.139774695,0.077550739,-0.1289093494,0.1089108884,-0.0839564577,0.0414062925,-0.1637930125,-0.0721092597,0.0359598212,0.6240739226,-0.2325592786,-0.5413184166,-0.5269168019,-0.0988926366,0.1037318781,0.1883218139,0.0515774116,0.1023936495,0.1159397885,-0.0749133602,-0.1888535172,-0.1452777237,-0.0556888357,-0.0028690889,-0.3264967203,-0.2233644128,0.3309184015,0.2039163411,-0.0945509523,0.1549262106,-0.2039530426,-0.092710495,-0.0095436256,-0.2591286004,-0.2582912147,-0.463026464,-0.1182582676,0.0588776506,0.2336980402,0.108112976,-0.1332108825,-0.2215162218,0.1171837002,0.6883972287,0.1997386664,0.1489496082,0.402256608,0.029157605,0.1394936591,0.0054768277,-0.0856491104,0.166863516,-0.07627327,0.3958076239,0.0007771892,0.2863453031,-0.1103602871,-0.5084636807,-0.023641672,-0.3658467531,-0.2045274526,-0.0995762274,0.112592347,0.2994084954,0.2819698453,0.1736928523,0.4408884645,-0.0008400108,-0.095121637,-0.0727198273,-0.0292542335,0.0549812205,0.2533734143,0.0195076615,-0.0273443554,0.0528054126,0.0067487173,-0.4714187086,-0.2150871903,0.1913157701,-0.2646383345,0.0638191029,0.2851462364,0.2803843021,0.1533473432,0.1773774326,-0.3570280969,-0.3106738031,0.1204815879,-0.0965005457,-0.0488858037,-0.0437003672,0.0947650969,-0.1267966479,0.0659134686,-0.5947527885,-0.0187632609,0.4436030686,-0.2635980844,-0.0719171688,-0.2288510203,-0.0116031589,-0.3403199315,0.0138408691,0.4004660845,-0.3379040956,-0.0681683049,-0.2169623524,-0.2334198356,0.1664637476,0.4936624169,-0.0835335925,0.516941011,-0.1955942512,0.3982065916,0.1584657282,-0.2379556447,-0.2000245303,0.5022702217,-0.1037771031,0.1734888703,-0.0952558294,-0.0241522361,0.4080353379,-0.173525393,0.1406641603,0.0812345371,-0.0405478813,-0.0777483135,-0.194061473,-0.2555298805,0.1296404302,0.0126433447,-0.1006332636,0.1342628598,0.1260567456,0.1880754828,0.3111828566,-0.0541514978,0.2493341416,0.3054505289,0.117798306,-0.1352878213,-0.1075367704,-0.0221280865,-0.3029607236,-0.0044431854,0.2780956626,-0.2112738043,-0.0934908092,-0.1717114598,-0.2598276436,0.230335027,-0.2763422132,0.2911271751,0.0529794879,0.1208057478,-0.0248321537,0.2489490956,-0.0919569507,0.1896861494,-0.035242334,0.1281069666,-0.4389435053,0.4238277376,-0.0409120433,-0.3451372385,0.0302613918,0.0569884367,0.2687718272,-0.134433791,-0.0043099276,0.3361161947,0.2367794812,0.13005279,-0.2204691917,-0.0598820932,0.2161560059,-0.0719890445,-0.1192336306,0.2150163651,0.3863373101,-0.1784895062,-0.1886806488,0.3041788638,-0.0705867633,0.1873155981,0.0063082632,0.112623103,0.0241007935,-0.002370018,-0.0940756425,-0.4081260264,0.0147502823,0.0237155166,-0.1132685915,-0.2551979721,-0.0995129123,0.1117107123,0.4493326843,-0.2675612867,0.2195772082,0.0182680245,-0.1571053565,-0.0029315124,-0.0543791577,0.4611753821,0.5188760757,-0.0728522763,0.2165549845,-0.1663918197,0.0560957193,-0.0409083143,0.1319316179,0.1509232819,0.3151162267,0.2516278028,0.0549508184,0.0588581562,-0.0803973228,0.149934113,-0.0055816928,0.1412248909,-0.404068172,0.1511809379,-0.2129691243,0.3331795335,-0.1819794029,-0.2257498801,0.0959367529,-0.6485622525,-0.4150827229,0.4897378385,-0.1861340553,-0.039807003,0.0067072879,-0.0427006856,-0.0165362135,-0.0674778298,-0.124694787,0.1129419208,-0.2868056893,-0.0729240328,0.2463184297,-0.2035422623,0.2206840366,-0.1972499937,0.0126172695,-0.2052572817,0.0368248969,0.1174176559,-0.1562771499,0.1773216873,0.1652359366,0.2820571661,0.0242906269,-0.2725773752,0.0964507759,0.1477789581,-0.1811760366,0.1133565828,-0.0025246169,-0.0518953949,-0.1257267296,-0.158946678,-0.3489079773,-0.1871428788,0.5054351687,-0.1241575256,-0.2554841042,-0.1415614784,0.1180824488,0.1391178817,0.2809080482,-0.2997899354,-0.1871420145,-0.129202038,0.3373076916,-0.0742208287,-0.1497434825,0.1376183778,0.04738646,-0.2108744681,-0.0256522689,-0.4338174462,0.2235617787,-0.1067928895,0.041715987,-0.1858791709,0.0072100954,0.2149301469,0.1670225263,0.119503133,-0.0201800372,-0.4066403508,0.2743183672,-0.1029174104,0.2617074251,-0.3179399371,0.3655718863,0.2267260849,-0.0936277136,0.1042671725,0.1052585691,0.4074470103,0.4023137391,0.2292523235,-0.2795910835,-0.0465680361,-0.0305438284,-0.4651505351,-0.3055941761,-0.0788278282,0.0058602551,-0.4329887033,0.0954170525,-0.416962713,-0.0798513517,-0.4240089357,-0.064945735,0.0411414132,0.2644873261,-0.1293402761,0.3300028443,0.121315293,0.2686508298,0.0854175612,0.4426595569,0.304084599,-0.0225938745,-0.6052621007,-0.3987894952,-0.1494777501,0.085171327,-0.0030239439,0.3307610452,-0.1128201112,-0.158019945,-0.068572633,-0.0389397144,0.5201874971,-0.0512404703,0.0012258509,0.1916232407,0.0426360592,-0.3706858754,-0.0845857188,0.1276453584,0.261156112,0.0996290743,0.7260839939,-0.2869675756,0.0848333985,0.0986012891,0.3608519137,-0.2686014771,-0.2606610656,0.0762175992,-0.6618020535,-0.473454237,-0.0654048696,0.0891384482,0.0840258971,0.0445675887,0.0613792241,-0.2863175273,-0.1974561363,-0.030964205,0.0094024306,0.2691719532,-0.1909771711,0.1744082868,0.273471117,0.0785733461,0.4405231774,0.4388534129,-0.0147153819,-0.352555275,0.1523869783,-0.0716162622,0.0187627245,0.1555835307,0.0508779213,-0.0821264014,0.0883737728,-0.2365548164,-0.4633461237,-0.1488452405,0.0904251486,-0.3374938071,-0.6183341146,-0.2228628546,0.3830707669,0.0004322185,0.0913360044,-0.0558939688,0.2712638974,-0.2566220462,0.4474663734,0.1223123968,0.6830673814,0.0007449057,0.0854566023,0.4916387796,-0.6043204069,-0.0060888282,0.1641178727,-0.0600038543,-0.4265597761,0.1930721849,-0.0806793496,-0.0681977421,0.1394303888,-0.0234502982,-0.3511241376,-0.115192242,-0.3808244169,-0.205246985,-0.1649704576,0.1392966807,-0.4981330931,-0.3532392979,-0.1283452064,0.1440775543,0.0093873618,-0.0194945317,0.140739888,-0.3155276179,-0.0234527495,-0.0966568813,-0.3905836344,0.0247118045,-0.0757293254,0.3466628492,-0.1778202206,-0.1701766551,0.2836356759,0.2573932409,0.514473021,0.0030206104,-0.0201175436,-0.1384010911,-0.2794148922,0.0727894604,-0.017847985,-0.235719949,0.4342162609,0.1702372879,0.0195976123,-0.2134186029,-0.1493427455,-0.1292095184,-0.1325327307,0.0547382496,-0.224358201,-0.5140143633,-0.2109909058,-0.1133698747,0.0389370359,-0.3093291819,0.012805759,0.1201058105,-0.0274823904,0.1266261339,-0.3678091764,-0.0474945344,0.0045270314,0.3145180345,0.2062477767,0.2634675801,0.6696696877,-0.0583996847,-0.1212425306,-0.2391156405,0.7491883039,0.2670490444,-0.0332002267,0.3828852177,-0.3566769361,0.0791084021,-0.1761876494,0.2747239172,0.2144100964,-0.3028134406,-0.0459919684,-0.4061584771,-0.3587571084,0.2031573206,0.0973908529,0.2266011536,0.1682340205,0.1613412797,-0.2150634527,-0.0841039568,-0.148704946,-0.0051133367,-0.2340533882,-0.0138233341,-0.0079671498,-0.0616992116,0.2066635936,-0.1218830571,0.0526377968,0.202642262,0.0701572746,-0.0461211875,-0.1765582114,0.1328735501,0.1106576174,0.0928322822,-0.128450051,-0.2062306404,0.0730832443,-0.138905406,-0.1202656105,0.2732168436,-0.0864552036,0.1101714894,-0.0113607114,0.3906763494,-0.1400344819,0.1156818345,-0.0184722356,0.1907476485,-0.0251064748,0.2376375496,0.0582986288,0.0307972077,-0.3982856572,0.0157231875,0.3234840035,0.3239723444,0.2735621929,-0.5291035771,0.0136452671,-0.1279325783,0.1995225102,0.3046544194,-0.2359331995,-0.0323292315,-0.0842189714,0.0597508997,0.2379079908,-0.124497585,0.3456408978,-0.3701761067,-0.2507579327,0.194592312,-0.0075365012,0.226304695,-0.5040649772,0.0490179099,0.5732997656,0.0245195534,0.2739393413,0.5586566329,-0.2972379029,0.2062015533,-0.0551292561,-0.0303288829,0.2992989421,0.148515299,-0.0736934394,0.3769677579,0.2046097368,0.1215097234,0.0530814901,-0.191453293,-0.0007979974,-0.0545356646,0.1000757143,0.3828053474,0.1555067599,-0.0381457657,-0.6030409336,-0.1078897268,-0.1749638319,0.0733095482,-0.2203655541,-0.0936853662,-0.0803633779,-0.2421419472,-0.0503131151,-0.0172618218,-0.1793982983,-0.1467364579,0.2449443489,-0.052128762,0.0955899954,-0.119679518,0.0544791631,0.1570000648,0.0720025674,-0.3394526541,-0.1150503308,0.0763376132,-0.0231554266,-0.0515820123,0.0262633376,0.4611822665,0.2230966091,-0.299393028,-0.168780759,-0.0456724763,-0.0893872008,0.194780454,0.1658117026,-0.2169560939,0.1562274396,0.589836359,0.0546889901,0.0335577279,0.2126052082,0.2061872333,0.167957589,-0.1390759796,0.373916775,-0.4525499344,-0.1057959199,0.0817324072,-0.0233930927,-0.1923722029,-0.0686579868,0.2831576467,-0.0889566764,0.2677441835,0.1079684049,0.0300730485,0.0063130576,0.4416630268,0.2064991742,-0.3510468304,-0.1622361094,0.1728037894,-0.1454878896,-0.1226277798,-0.0181306694,0.0106476657,-0.0071235853,0.1739094704,0.1665398031,0.1428555399,0.0265425052,0.0155514749,0.1411405951,0.4424415231,0.0177942589,-0.3375852108,0.5128375292,-0.1091878042,-0.0130157769,-0.4200245142,0.2337831259,0.3612352908,-0.0040313802,0.0197012033,-0.036572855,-0.0331173651,0.0064799469,0.1603878736,0.4409674704,0.2313555628,-0.0048692822,-0.3268092573,0.0847199857,-0.1212448403,-0.1091054529,0.1739548892,-0.1156606078,0.7781214118,-0.3791902065,0.4145943224,-0.2927051187,-0.2153362185,-0.0463361219,0.1783768833,-0.4962359369,0.2853748798,-0.0774828047,0.0432463624,0.3445824981,0.1797284037,0.0849786848,0.3774322867,-0.1755812764,-0.0792290568,0.3478639722,-0.0715023652,-0.3467016518,0.4105906487,-0.1564235538,0.349638164,0.2499444485,-0.6930737495,-0.066903837,0.3843768239,-0.2654726803,-0.1002497748,0.0628135875,0.2828069031,-0.1878263652,0.0814793333,0.4868473709,0.0644614995,-0.2565093935,0.3085873127,-0.0935587063]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2135","title":"en language data from MLQA dataset is missing","comments":"Hi ! Indeed only the languages of the `translate-train` data are included...\r\nI can't find a link to download the english train set on https:\/\/github.com\/facebookresearch\/MLQA though, do you know where we can download it ?","body":"Hi\r\nI need mlqa-translate-train.en dataset, but it is missing from the MLQA dataset. could you have a look please? @lhoestq thank you for your help to fix this issue. ","comment_length":35,"text":"en language data from MLQA dataset is missing\nHi\r\nI need mlqa-translate-train.en dataset, but it is missing from the MLQA dataset. could you have a look please? @lhoestq thank you for your help to fix this issue. \nHi ! Indeed only the languages of the `translate-train` data are included...\r\nI can't find a link to download the english train set on https:\/\/github.com\/facebookresearch\/MLQA though, do you know where we can download it ?","embeddings":[-0.0344628803,0.2046830654,-0.227187261,0.2308054119,0.0578236319,0.2997930646,-0.0395332575,-0.0461162217,-0.1498138011,0.1264657974,0.195892483,-0.1498139054,0.067390956,0.4813955724,0.2251041383,-0.2650109828,-0.009786183,-0.079442285,-0.1342740804,-0.4665370584,0.0198140983,0.5020956993,-0.0956803113,-0.031468194,-0.1739675552,0.1743480265,-0.3577691317,0.0911980495,-0.1458940953,0.0571778044,0.1609995961,-0.3259171247,0.2097357213,-0.1647292376,-0.0001091284,0.0488843881,0.0214317422,-0.2748510242,-0.0866407752,-0.3655617535,-0.119143106,-0.2956379354,-0.1769669056,-0.1298706681,-0.1838501245,-0.0736344382,0.1559998691,-0.2765039802,0.1190437302,0.3667834997,0.1788885742,-0.1254622489,0.1327679604,-0.2215016633,0.0960634872,-0.0015868166,0.1220990717,0.0806107,0.3105128109,0.0426357538,0.1654581577,0.5170339942,0.3657609224,-0.1951818019,-0.2743358314,0.0918550566,0.4387052655,-0.6130290627,0.2459109873,0.2323862463,0.3963029981,0.0642605722,-0.2204280496,-0.0049439119,0.2352886647,0.1712478101,0.0973937288,0.4704558551,-0.331916362,0.2780916095,-0.1069712043,-0.5191504955,-0.0931159258,0.2702248693,-0.1711803973,0.2614015341,0.0409468375,-0.0682276338,-0.0980315283,0.0551038012,0.1683304608,0.1132869795,-0.2043648809,0.2074112594,-0.0596287251,-0.0694435984,-0.1071814075,-0.0628479496,0.0442221314,-0.3017544746,-0.4466921389,0.1175957918,-0.098255001,0.0766113847,0.2756772339,-0.0042863861,0.0451873802,-0.1117994413,-0.0182722826,-0.4221897721,-0.0167216752,-0.1126443073,-0.2181791067,0.0549062714,-0.6116184592,0.2076267749,-0.1856248081,-0.3717593849,-0.2271015197,-0.0178472102,-0.548188448,-0.1329703033,-0.0842325613,0.1708070636,0.1391259432,0.1097304523,0.0387191661,0.0876052156,-0.2127081901,-0.5654237866,-0.1886694729,0.216659233,-0.2004490644,-0.2025626153,-0.0060594459,0.2342913002,0.4591828883,-0.1983031332,0.0367990211,0.0089186076,0.1423790157,-0.152129814,0.1630709767,0.1273078769,0.2819876373,0.1840767115,-0.0005702196,-0.2114921212,0.0147197312,0.2187458277,-0.3707552552,-0.1261291802,-0.3416649103,0.2081024051,-0.0157137103,-0.2832198441,0.2268716097,0.7006776929,-0.0515589416,0.017304562,0.2035379857,0.0312999897,-0.078900829,0.0856770352,0.200035125,0.2012119293,-0.601254046,-0.4723479152,0.132443592,-0.1496263295,0.2162726969,0.3347105086,-0.308137536,-0.0056157857,-0.2851573527,0.1473447233,0.3123139739,-0.090883866,-0.3621456027,0.0711625069,0.0378513075,-0.1875449121,-0.1637816429,0.2593780756,-0.0118397549,0.1117783114,0.2610759437,0.4686023593,-0.1223036051,-0.1898304373,-0.1053931713,-0.2279411852,0.622326076,0.2268796265,0.212640658,-0.1230190992,0.1040074155,0.1111715212,0.2606935799,-0.0795060396,0.1433201581,0.5985336304,-0.1020627618,0.0215906184,0.2536551654,-0.0001963689,0.0188051555,-0.0132762641,-0.6309732199,0.2035327554,0.0095851244,0.021999117,-0.0432276987,-0.1058528498,-0.1966770589,-0.2617295086,0.1333910227,-0.2965782881,-0.0784632564,0.4289409816,0.0495193824,-0.0088376524,-0.2400574237,-0.131760627,-0.0592113137,0.4036954343,-0.2992990315,0.0492517166,0.2209753692,0.4827614725,0.0691576824,-0.0927278921,-0.0542097725,-0.0900359303,-0.0186838806,0.0758046508,0.2721095383,0.1493699998,0.5323989391,-0.2613608241,0.2088598758,0.1365316361,-0.0400511809,0.0231652334,-0.0972541869,0.2465907186,0.3171747923,-0.021942256,0.0535087921,0.344753921,0.3326198161,-0.0062588919,-0.0123571251,-0.3073703945,0.3201687634,-0.072664842,0.136526227,-0.0757598355,-0.1469404548,0.2850211263,0.7387043238,0.1240657568,0.2098559886,0.084957771,-0.4218397439,-0.2090277374,0.297154665,0.0460442305,0.0749297887,0.1877487451,0.3617949188,-0.0734828711,0.4398923516,-0.140120551,0.2000077516,0.1594542116,0.1363021135,-0.2569343448,0.0658930317,-0.0054348838,-0.4006048739,0.1530468166,0.2428847998,-0.0181932673,0.1239670739,0.1371142566,-0.3840941191,-0.8181413412,-0.1210189834,-0.3285419047,-0.2786876559,-0.1606533974,0.1709367633,-0.7855653167,0.0042886627,0.2437331825,0.3373645544,-0.014211446,-0.1161951944,0.1168831736,-0.0756036043,-0.1245467365,-0.4825126827,0.1762280315,-0.0289322622,0.1055137739,0.0067073326,-0.1609124243,0.0697346777,-0.2360586375,-0.48951298,0.1627566963,-0.0336750783,0.1449293941,-0.0722268894,-0.0020543311,-0.4595686197,0.0561483987,0.0869889036,0.1313925833,-0.3589528501,-0.0264959875,-0.1333763897,-0.0539586432,-0.0296468325,-0.4481564164,-0.7942317724,0.0470998026,-0.0254445542,0.0235762578,0.0455225408,-0.0960215256,-0.181181848,-0.0012225162,-0.0862498581,-0.0588547997,-0.4134362042,-0.1808990538,0.1929139197,-0.089289166,-0.4240905344,0.3333979249,0.0575139672,0.1615899056,-0.2290339321,-0.4955675602,-0.2501125634,0.1038496494,0.0756998435,-0.2734216452,-0.0491298884,0.3173699081,-0.1497465223,-0.0121672777,-0.0257461332,-0.0596243441,0.0275728069,0.2583023906,0.4377896786,-0.4849637151,0.3810780644,0.2894515693,0.3349755108,0.3354417384,0.1928781867,0.1851141006,0.0049765268,0.204833433,0.0775878057,-0.093962878,0.4627104402,0.0153496601,0.0678962842,0.2450260073,-0.0310823452,-0.2595216036,-0.1282051504,-0.1448743194,-0.1577750444,-0.1185041443,0.0035322404,-0.3705375493,0.4462801814,0.278881669,0.2835541964,0.2334595621,-0.0040406385,0.310860306,0.0581936315,0.1092327237,0.1489842236,-0.5356492996,-0.6774756908,-0.4137742817,0.0178143326,-0.0460203327,0.3492558599,-0.2022689283,0.3159270883,0.1282984465,0.0692241564,0.653663516,-0.1781610847,-0.3705919087,-0.0675853193,0.0320616923,0.0259083938,0.0210636072,-0.067351222,0.3173502982,0.5049280524,-0.1271526068,-0.0750949532,-0.1299311072,0.1897286475,0.2722421587,-0.3018901944,0.2110725492,-0.0528509468,-0.2436062992,-0.1980119348,-0.0069027725,0.1067410707,-0.0079733189,0.0640232936,-0.0087010032,0.2498305291,0.2399794012,-0.0909939781,0.0296720564,0.2414085418,0.2405936718,-0.006944431,0.1101201922,0.167550683,-0.0403500348,0.3628755212,0.1135010421,-0.3394518793,0.2700650394,-0.2651729882,0.3873056173,0.1473499388,0.0750739947,-0.020373283,0.5720809698,-0.3262433112,0.035692811,0.0072496082,0.1877932698,0.1163316667,-0.1971900612,-0.3613518178,0.1051104814,-0.1618571877,0.1121185347,-0.2287811786,-0.0159631651,-0.1922406107,0.4723280072,0.1149993464,0.8131114244,0.172917828,0.2653674185,-0.2128789723,-0.1444637328,0.1843687892,-0.2363998443,0.1201042458,-0.0790588707,0.0088768303,-0.2042356282,0.0908279344,-0.046667017,0.2055428773,-0.0521624312,0.3298265338,0.3807366192,-0.3269473016,0.1866822541,0.0681792349,0.0490682125,-0.1569547653,-0.1394020319,0.1233368814,0.1078926697,0.4474700689,-0.1518150717,-0.195032686,0.0474469252,-0.0101229465,-0.4358277619,0.0342404842,-0.6181277037,0.2651330233,-0.2025089711,-0.4514231384,0.1985459328,0.2820284069,0.3624441028,0.2047536075,-0.3248588145,0.5658938289,0.2798181474,-0.0865012109,0.3733977079,-0.0790587515,0.1902603656,-0.2697784007,-0.2434542477,0.2396706194,-0.209874317,0.114332363,-0.6147165895,0.0983329937,0.0974824727,0.2014009804,-0.064360857,0.3105178475,0.0140710184,-0.2111014873,0.1187827736,0.303391248,-0.0453685485,0.0930122957,0.0480728336,-0.0674219653,-0.277444005,0.3492662013,0.211670503,-0.1594548821,0.5841690302,0.1649902314,0.0130453156,-0.1773336232,0.1023618281,0.0691415668,-0.1856683195,-0.2398852855,0.0521864854,0.1877983809,-0.1963648051,0.1287595928,-0.0096379379,-0.276881218,-0.1318552047,-0.4883030057,-0.0330955759,0.2231066674,0.0115031125,-0.2465585023,0.207194671,0.1146185175,0.1737968624,-0.0177275948,-0.2826580703,0.099017784,-0.3407903612,0.1170912981,-0.31098634,-0.0273568612,0.2107983232,0.0098882942,0.0411001332,-0.1670541316,0.0556563474,-0.1659139097,0.0944287106,0.0442335829,0.1019381732,0.2109967917,0.0532393269,-0.2525967062,-0.1904030293,0.2619977593,-0.0674166083,-0.027933076,0.0371682048,0.2888005376,-0.0973697156,-0.0009417373,-0.2473876476,0.0661412403,0.2350029498,0.1454191357,0.2674199939,0.2663162947,0.357034266,0.04464994,-0.1112377122,-0.1341848522,-0.1661472768,0.000379394,0.1076240912,-0.1778693348,0.1229848415,0.1210381091,-0.1532257795,0.004413127,-0.3717908859,0.2839588225,0.4176321924,0.2205511481,-0.4797022343,-0.0664165765,0.0775193274,-0.2523564696,-0.0204070508,-0.0181947779,0.4070229232,0.3347913921,0.5049914718,0.1084177792,-0.0866630897,-0.132261768,0.0833304301,-0.125188157,0.0438098982,0.1171376407,0.1153695807,0.1527492106,0.0121640703,0.4706031978,0.1794547737,0.2413914204,0.3241352737,0.1357879788,0.045819439,-0.1199910939,-0.1456768662,-0.0049325316,-0.1080216765,0.1095988601,0.0138996001,0.1421697587,-0.0046588746,-0.303781718,-0.1372004151,0.2461920381,-0.2555316091,0.1096114963,0.0571783036,-0.116688773,-0.0666640997,0.0646630377,0.0561560616,-0.2038900703,-0.2721970677,0.2663774788,-0.2021347731,-0.0640047863,-0.0655525774,0.2139686346,-0.2533403337,-0.0913543254,0.4351055324,0.091925025,-0.1498566121,0.1068016216,0.1184948161,-0.1101508215,-0.1591192186,-0.3091270626,0.0644036904,0.2533023655,-0.1815321743,-0.1403748095,0.2805759609,0.5617468357,0.2089321017,0.2611387968,0.1185271665,-0.0966019481,0.0250889901,0.2744982839,0.0262698475,-0.1715334505,0.1727131158,-0.1618003547,0.0789708644,-0.4373008311,-0.1165173128,-0.4430786073,-0.0360766053,-0.0994550362,-0.1601615846,-0.1082494706,-0.145209372,0.0955013931,-0.1738533378,0.2952964902,0.489910841,0.2807605267,-0.2185988426,-0.4658297002,-0.2478750497,-0.0861513019,-0.006959117,0.2598725259,0.0467687435,0.3083914518,0.337292701,0.2134710401,0.1028381288,0.3506148756,-0.2340524048,-0.0428212583,-0.4953738749,-0.2108680606,-0.0311658159,0.4508553147,-0.0775642991,0.0645223781,0.147777915,0.1814357787,-0.0305832271,-0.2744168341,-0.3053379655,-0.0724550113,0.0184186809,0.5293847919,0.1815387011,0.0855618864,-0.1395011097,0.0197933391,0.165784657,-0.3820600212,-0.0141180223,0.2694744766,0.0222815238,0.033462476,-0.0601960421,-0.1295230985,0.0392524414,0.1752390265,0.0319485404,-0.1724292189,-0.3260920346,0.2349205464,0.0191378389,0.1085358784,0.0108624687,0.3288311362,-0.0269602109,-0.2470650077,-0.2232985795,0.1386561096,0.0727492869,-0.2553524375,0.032401789,-0.2219076604,-0.0193355009,0.0205100328,-0.0427108482,-0.2704522312,-0.0044649085,0.3768491149,0.1395280659,0.0286062099,0.2439169139,0.0668071285,0.1847597957,-0.137300089,0.3015005589,0.2160288841,-0.2874204516,0.0556801967,-0.1076183021]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2135","title":"en language data from MLQA dataset is missing","comments":"Hi @lhoestq \r\nthank you very much for coming back to me, now I see, you are right, in the link you sent I see split of {split}-context-{context_language}-question-{question_language}.json with context_language=question_language=en, TFDS most probably has extracted english ones from these files as en language files, but translate-train\/test do not have en indeed. thanks a lot for the great explanations","body":"Hi\r\nI need mlqa-translate-train.en dataset, but it is missing from the MLQA dataset. could you have a look please? @lhoestq thank you for your help to fix this issue. ","comment_length":57,"text":"en language data from MLQA dataset is missing\nHi\r\nI need mlqa-translate-train.en dataset, but it is missing from the MLQA dataset. could you have a look please? @lhoestq thank you for your help to fix this issue. \nHi @lhoestq \r\nthank you very much for coming back to me, now I see, you are right, in the link you sent I see split of {split}-context-{context_language}-question-{question_language}.json with context_language=question_language=en, TFDS most probably has extracted english ones from these files as en language files, but translate-train\/test do not have en indeed. thanks a lot for the great explanations","embeddings":[-0.0408197865,-0.0362202525,-0.1773771346,0.3228131235,0.1323038191,0.2861721516,0.0952832699,0.1073280647,-0.2343607247,0.1195361987,0.0948156416,0.0513932146,0.1680396497,0.523902595,0.206296578,-0.401420325,-0.0716602728,0.0019192108,-0.1760003865,-0.3533481658,0.057527937,0.4546353221,-0.1336801648,0.097742781,-0.283597976,0.0972493514,-0.4078720808,0.208615616,0.1086850464,0.0607025623,0.2069737017,-0.3061185181,0.1648241729,0.0024818643,-0.0001026441,0.0339449085,0.0562221073,-0.3316811323,-0.1245716438,-0.3011199534,-0.2936724722,-0.1166031733,-0.1913672239,-0.2669698298,-0.2100294083,-0.1600869596,0.0102598798,-0.4025422931,0.294521451,0.3969225585,0.2244403213,0.0597102493,0.1115079373,-0.0750962645,0.1472627819,0.0179583561,0.1538949758,0.0909608826,0.1666130871,0.0342720151,0.1306973696,0.5047555566,0.2061388046,-0.1100770161,-0.2966980338,0.1319010258,0.2486949563,-0.6032835245,0.2411035001,0.3763130307,0.5104178786,-0.0404776447,-0.2418508083,-0.1448254138,-0.0136252465,0.0993714705,0.1884403527,0.2149871439,-0.1208995357,0.2852535546,-0.0185189284,-0.312222898,-0.0564261451,0.1563021839,-0.4507113099,0.2781139314,0.0598659143,-0.0272461213,-0.207635656,0.0720339715,0.214170292,0.0857241899,-0.2254954129,0.1790015846,-0.2212730348,-0.2123450488,0.0013374971,-0.2044260204,-0.0787948668,-0.1716720015,-0.5002161264,0.0687911063,-0.075840354,0.0553399287,0.1506575495,0.1425784379,0.0742118508,-0.0000407521,-0.0345675312,-0.5035304427,-0.1269602627,-0.1095474735,-0.0571455173,-0.047424417,-0.5842799544,0.2645859122,-0.0489339381,-0.2018536329,-0.2747574449,-0.0557357557,-0.6679561138,-0.2636979818,-0.1630070359,0.0584547967,0.207579568,0.1428805441,0.0831876993,0.131198138,-0.3226513565,-0.4529734254,-0.2833188772,0.2836948633,-0.0452450551,-0.1650166959,-0.0661564469,0.2656957805,0.5278736949,-0.2556885183,-0.0366106443,-0.0110005038,0.074529916,-0.1067346185,0.0769132599,0.088959977,0.1962748915,0.0763683543,0.0843277425,-0.1846405268,0.0152574796,0.2290404737,-0.4823003709,-0.0396410599,-0.1821142286,0.2690197527,-0.1023227274,-0.2360186428,0.1503035277,0.6440026164,0.119322069,-0.0179714523,0.1329053193,-0.0403433889,-0.1953152269,0.0868195295,0.2072445899,0.2232121378,-0.7090595365,-0.2352009565,0.2311965525,-0.1255780458,0.0649191663,0.2192748338,-0.1816485524,0.2974399328,-0.2996007204,0.3631656766,0.1245925054,-0.3740579188,-0.118547298,0.2619829774,0.0106617613,-0.1634265929,-0.1058031917,-0.1710813642,0.160108164,0.0318514332,0.0292906221,0.3716142774,-0.2197672874,-0.1256002933,-0.1732638627,-0.1923583448,0.5057458282,0.1941702515,0.1641436815,-0.1853484511,0.045119103,0.1144302264,0.2556923032,-0.2323063165,0.0895541832,0.4553655088,-0.1292886883,-0.0670070052,0.2314843684,-0.1167257577,-0.240120694,0.0309543479,-0.503665328,0.2067003399,-0.0675953329,0.0471132062,-0.1141039729,-0.1690893024,-0.1501901448,-0.266376555,0.2185669839,-0.128269881,-0.1501293778,0.5255807638,0.1210810244,-0.0323449001,0.0362350829,-0.108423762,-0.1963132471,0.4880462587,-0.1885199696,0.0312878825,0.1449316591,0.3960081935,0.1007948294,-0.2371941954,-0.1993572116,0.0358269177,0.01040811,0.2304920107,0.0124625033,-0.0573116876,0.4358989298,-0.2884270847,-0.0393668376,0.1249902397,0.0571261086,-0.0100623779,-0.1664896756,0.3173385859,0.1417885125,-0.0154546611,0.0820578039,0.1560006589,0.2744921148,-0.2671564519,0.1516212076,-0.2489779443,0.1365277469,-0.1008696184,-0.0263799652,-0.0152852172,-0.0500319637,0.3457328677,0.9597117901,-0.0387768224,0.0324349031,0.1088812649,-0.2438545823,-0.2485950589,0.3327144086,0.2219971269,0.3681366146,0.2550665736,0.27862221,-0.059629485,0.3849787116,-0.2170824558,0.2478822172,0.1016413942,0.1226856858,-0.0602648593,0.0111213792,-0.1178099215,-0.3725441396,0.2336734086,0.2865144312,0.2004962564,-0.0297941826,0.2682924867,-0.5310765505,-0.6335401535,-0.0419786051,-0.2432370484,-0.1427946985,-0.2481586635,0.272582382,-0.5574778914,-0.1085469127,0.2909506559,0.4410870671,-0.0919117853,-0.1788231879,0.1353015155,-0.2354326993,-0.2749875486,-0.567044735,0.2050751746,0.0931925327,0.1703082919,0.1206256896,-0.244024083,0.0810178593,-0.0437348634,-0.4981174469,0.1700586826,-0.2002723217,0.281632632,-0.0635293499,0.0004735475,-0.4595284462,-0.0440799706,0.0705253184,0.1709949672,-0.3872296214,0.0093745571,-0.0262725689,-0.0521978997,-0.157628715,-0.7814698219,-0.5482575297,-0.0636083931,-0.1323074549,-0.1232254654,0.0307012759,-0.022132799,-0.1314075291,0.0421980061,0.0472832769,-0.1079865843,-0.3893261254,-0.1267874241,0.1431223899,-0.1554432809,-0.3047045767,0.2366571873,0.0481761917,0.4225658774,-0.2196065784,-0.3472971618,-0.1204656363,0.056391757,0.1077346206,-0.3534255326,-0.1192000061,0.5090583563,-0.017927574,-0.0089256978,-0.1515680403,0.0382350422,0.1233563051,0.2782582045,0.3899996281,-0.4099954367,0.599019289,0.3242695332,0.2691341341,0.1265043616,0.1426517665,0.1432933211,-0.0658831447,0.1140460372,-0.0823448151,-0.0694597214,0.4828374386,0.0637095496,0.027254751,0.4788178504,-0.0245168749,-0.1713120788,-0.1844300181,-0.1877018213,-0.1659408063,-0.2019790113,-0.0622825101,-0.43255952,0.3646171391,0.2551710308,0.2961049378,0.1527626663,0.0155420098,0.21429649,0.0851204246,0.284989953,0.1448106617,-0.6591535807,-0.4503853321,-0.1865547001,0.1234263778,0.0855081454,0.2387643009,-0.2729607821,0.1774840504,0.2363274544,0.0362738147,0.6784834862,-0.2564089596,-0.2933887839,0.0935809761,0.1132447794,0.0536968783,0.0061186906,-0.0851854756,0.3044101298,0.3356390297,-0.07378003,-0.2037022859,-0.2334937304,0.0725352392,0.275606364,-0.3622870743,0.0017342512,0.0183094721,-0.3184652925,-0.2057559788,-0.1192359701,0.2498415262,-0.0027277947,0.0554810911,-0.0868656933,0.2147668153,0.0321339108,-0.026101185,0.1267828941,0.2505819499,0.0226904061,-0.0493066758,-0.0068113841,0.133060962,0.0133359013,0.1902782768,-0.0447801314,-0.1839900911,0.1142016351,-0.3886477351,0.4801160395,0.1302576661,-0.0223742835,0.021154169,0.224654451,-0.1637255251,-0.0612126626,-0.0564257838,0.3183376193,0.0541251972,-0.08471255,-0.2531456947,0.0157669522,-0.196872592,0.0739416257,-0.1369461119,-0.0791202933,-0.3975170851,0.4556827247,0.1212805733,0.7878761292,0.1390993297,0.1330154091,0.1484573185,-0.2146419883,0.3469111621,-0.199674964,0.0509752929,-0.08579088,0.080855377,-0.1684202552,0.211710766,-0.0176616237,0.3485617936,-0.1559047252,0.3202444911,0.3486588597,-0.2753963768,0.3074086308,0.0692005455,0.2205930352,-0.0945587233,-0.0573923923,0.1753158867,0.1122528762,0.2840489447,-0.0421051532,-0.2589123249,0.0853898376,-0.0583690256,-0.22405985,0.0409724303,-0.3553406298,0.2102057338,-0.117330946,-0.4628449976,0.3120923638,0.2903683186,0.3461298347,0.149078235,-0.2801378369,0.4850560427,0.0923698097,0.1833777428,0.3302989006,-0.0601475202,0.4175025225,-0.2193597853,-0.2475058734,0.13604258,-0.2273777574,0.1020635515,-0.6504183412,0.0804631487,0.008220654,0.1127505004,-0.1060498729,0.1084769592,0.0669878721,-0.2709407508,0.1626634747,0.1567842811,-0.1053351164,0.2285718769,-0.162286371,-0.0724320486,-0.1819044799,0.2074667811,0.2294254601,0.0350676812,0.6080000997,0.1775730848,-0.0141877178,-0.2541285455,0.1796358228,-0.0665777251,-0.3251992166,-0.0479490794,0.2189279497,0.2051668018,-0.085014604,0.0971975476,-0.1367193311,-0.2769072056,-0.0489877239,-0.4096363783,-0.0944938213,0.2128804922,-0.0337427109,-0.2400594056,0.1149062216,0.3987501264,0.1066140532,0.1629966348,-0.3291785121,0.2155857235,-0.2183161974,0.160819605,-0.1214574277,-0.0708763972,0.2643538117,-0.0111170663,0.0817813277,-0.1284231246,0.0617263131,-0.2129900455,-0.0690182447,0.0385485403,0.1084224209,0.1545952559,0.0985842943,-0.1769915074,-0.0905823186,0.2229056209,-0.0325191841,-0.0857415944,0.0031745262,0.3812493682,-0.0563385598,0.0034463208,-0.2623738945,0.0191896744,0.2412519306,-0.0005893053,0.1244042292,0.1977638304,0.2756618261,0.0010225967,-0.2014317662,-0.2218727469,-0.1708760411,0.0243087225,0.2697992921,-0.2374151051,0.0573116355,0.0547024347,-0.0651955232,-0.0607501008,-0.283990562,0.2237013727,0.3925498426,0.2249376029,-0.434430033,0.0377570949,0.0820379704,-0.2965414524,-0.0091945138,-0.0663655102,0.2731379271,0.2848387361,0.4747729897,0.0864744037,0.1784646809,-0.2362237275,0.0068645668,0.1435355842,-0.1436452121,0.1766200662,0.1010344177,-0.1597585529,0.026766276,0.3241795003,0.202207759,0.2690443099,0.1170288473,0.2710110545,0.0689442754,-0.193954289,-0.1826097369,-0.0439024121,0.00420553,0.0085673751,0.1069432646,0.2234367281,0.0276103299,-0.2771236002,-0.127393797,0.3374982774,-0.2935268879,-0.0225012824,0.1256628335,-0.0453941524,-0.0166810751,-0.0035410118,-0.1144258082,-0.0968774334,-0.2248788774,0.2747928202,-0.2744381428,0.0866866112,-0.2334063351,0.0101316404,-0.2466106415,-0.0226037651,0.5441874862,0.1120963395,-0.0789513439,0.0401854366,-0.0063090315,0.0046296078,-0.1731770933,-0.3625847101,-0.0365515724,0.0442523919,-0.1444823146,0.0325331502,0.4947677553,0.2612558603,0.1904446483,0.2604634166,0.1738150716,-0.1475002915,0.1729815751,0.3002625108,-0.0797089711,-0.2932745516,0.1676615626,-0.0861582085,0.0768183321,-0.3240535557,-0.252967298,-0.4617142975,0.0657513365,-0.1046212912,-0.1675472409,-0.0957542732,-0.0988186672,0.1194552258,-0.142617017,0.3566014767,0.3404727578,0.1341146082,-0.1007143259,-0.2554682195,-0.3600748777,-0.0783693045,0.0668039024,0.1585814357,-0.1292598397,0.1715533882,0.4762705564,0.0464805961,0.0491876677,0.1647145748,-0.0708705708,-0.0575922169,-0.4569231868,-0.2759056091,-0.1950778514,0.3763597012,-0.138044849,-0.0073044086,0.1454735547,0.4115004539,-0.0047126072,-0.3350498676,-0.0586324446,0.1186079755,-0.0255448222,0.3235163093,0.3283949196,0.2939720452,-0.1925685704,0.2201743126,0.1869622767,-0.2465390712,-0.096198909,0.287406832,0.086451076,0.1841493547,0.0069165742,-0.2039063424,-0.0543043874,0.2042929381,0.0147120832,-0.1221375912,-0.3691646457,0.2334661931,0.0279792193,-0.0248188451,0.2751075327,0.44263798,-0.0597257912,-0.1823959649,-0.3170250058,-0.0001685877,0.1756587625,-0.2104923129,-0.1348282099,0.0475173593,0.0052322671,-0.0084625836,-0.1049581617,-0.253235966,-0.038987197,0.4291750491,0.1130805984,0.130687803,0.4553578794,0.346768856,0.145632565,-0.052448187,0.1252946556,0.2982235253,-0.228976801,0.163567692,-0.1215906441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2135","title":"en language data from MLQA dataset is missing","comments":"I close the ticket, since I do not see any en existing, they have trained on \"SQuAD V1.1\" instead. Thanks. ","body":"Hi\r\nI need mlqa-translate-train.en dataset, but it is missing from the MLQA dataset. could you have a look please? @lhoestq thank you for your help to fix this issue. ","comment_length":20,"text":"en language data from MLQA dataset is missing\nHi\r\nI need mlqa-translate-train.en dataset, but it is missing from the MLQA dataset. could you have a look please? @lhoestq thank you for your help to fix this issue. \nI close the ticket, since I do not see any en existing, they have trained on \"SQuAD V1.1\" instead. Thanks. ","embeddings":[-0.0614050999,-0.0550391451,-0.1854838133,0.0839591473,0.1434618384,0.2676542103,0.1858885288,-0.0415909104,-0.1807008833,0.1482942402,0.2283990085,0.1972756088,0.1776254624,0.3383448124,0.3166697323,-0.247869283,-0.0484909713,-0.0179905575,-0.0484796688,-0.4285933077,0.0541552082,0.2352594584,-0.357439667,0.0050521591,-0.218355298,0.1566549689,-0.3980647027,0.1877370924,0.0598654859,-0.1359649152,0.2692220807,-0.3328851163,0.3129377663,-0.2540810406,-0.0001056774,0.0122205792,0.0811979845,-0.4159366488,-0.1993284971,-0.1626038104,-0.2880621254,-0.2823456824,-0.2393742204,-0.2648057938,-0.2193339765,0.2201204151,0.3443268538,-0.5083245635,0.3320854604,0.3818938136,0.1968887299,0.0395645015,0.0926605836,-0.1524075419,-0.0093364986,-0.0729986131,0.1497445405,-0.0014476244,0.2738120556,-0.0678198114,0.1563989073,0.4081826508,0.2631127834,-0.2174041569,-0.0966817588,0.1399407834,0.4089863896,-0.5473685265,0.1374904066,0.0993199497,0.371715039,-0.1057604551,-0.3758931458,-0.0194037091,0.2659439445,-0.0150941433,0.1812555939,0.3444670737,-0.1382599324,0.209009707,-0.0271232482,-0.302996099,-0.122995913,0.1439061761,-0.3219052851,0.2777846158,0.1805324256,-0.0605957657,-0.1598589718,0.1292635798,0.2392052859,0.2068473995,-0.1283605993,0.2811942995,-0.331851095,-0.2662265301,-0.0694770738,-0.1709136814,-0.0941285789,-0.3655155599,-0.3424856663,0.1182669103,0.1825283319,0.0852201208,0.1651428044,0.0044161966,0.2554106414,0.0712154061,-0.0691667348,-0.4249580204,-0.0405894779,-0.0316138789,-0.1165059879,-0.0557375476,-0.4660688639,0.3323318064,-0.0607328936,-0.2391403764,-0.3966356516,0.0730463043,-0.6150502563,-0.0928713828,-0.0460385531,0.0860478133,0.0689227358,-0.0985739455,-0.0307651572,0.0924641564,-0.2274388373,-0.4822268784,-0.2362546027,0.2626545429,-0.1490529329,-0.1555288583,-0.0512125827,0.1021778658,0.4211609066,-0.1078583375,-0.0333041847,0.0805079192,0.1542165428,-0.0777847096,0.094264783,-0.0650047362,0.182098642,0.0209613964,-0.005597428,-0.3666210175,0.1188591495,0.2223931253,-0.5026968718,-0.0869517028,-0.3409621119,0.2205947042,-0.0634133071,-0.2909999788,0.4000843167,0.4645343125,-0.0290918369,-0.0116881002,0.1916267574,0.0152749112,-0.1655332893,0.0587170981,0.1839257181,0.2479294986,-0.6743554473,-0.3153376579,0.2034638375,-0.0352101065,0.1968295425,0.2301511019,-0.2615668476,0.069613345,-0.1940802485,0.1121588349,0.1704313755,-0.6116746068,-0.3677040339,0.0222886056,-0.0779980421,-0.3776448369,-0.1448689699,0.0998805761,0.2610703707,-0.0236979052,0.2579838634,0.3530294597,-0.1783163399,-0.1971265227,-0.1738871336,-0.0527912304,0.2627733052,0.3654524386,0.148446247,-0.2631180286,0.0485103764,0.1426053643,0.1779536009,-0.2317244858,0.1725405008,0.5178675652,-0.0805445164,-0.0635037646,0.0584816262,0.1955318898,-0.2065805197,0.0163487215,-0.7800026536,0.2138622254,-0.0132620595,-0.121064119,-0.0612700582,-0.0453794524,-0.226060912,-0.24951379,0.1818477213,-0.3295162618,-0.1632551253,0.3572050333,0.1028450057,0.1105438024,-0.1355049014,0.0349664167,-0.1937730461,0.4484168589,-0.335169673,0.0082158577,0.1537080705,0.4459812939,0.1235243902,-0.0344227999,-0.00512549,-0.0497260913,-0.0372594558,0.0708749518,0.264909178,-0.2318826914,0.5788726807,-0.307380408,0.1909388304,0.126711145,0.0165009182,0.0356325582,-0.3685255945,0.3260649443,0.2498879582,-0.2839771211,0.0070402352,0.3589638472,0.1375349313,-0.1451908797,-0.0582831129,-0.1105434075,0.1184332818,0.0118993744,0.0044005611,0.0775820762,-0.112581946,0.1153248474,0.7596583962,0.2361644953,0.2454383522,0.051362671,-0.2074975371,-0.1795811355,0.3047199845,0.0839871243,0.2399901301,0.1667071581,0.0910311118,-0.0672457218,0.2820444107,-0.097026825,0.218987152,0.022077309,0.1279479712,-0.2176184058,0.0640218109,-0.0449231453,-0.3568538129,0.0635781884,0.2387513965,0.1498513222,0.0212786812,0.0785312429,-0.3708969951,-0.5073181391,-0.1364304423,-0.2138953209,-0.2516545355,-0.1184654385,0.2009530962,-0.7401012778,0.1231374517,0.3331766427,0.335840106,-0.0423550047,0.0280856397,0.0819606707,0.009855344,-0.1406043619,-0.34906739,0.1282382756,-0.0401220806,-0.0030068469,-0.0462615602,0.0058609713,0.176862061,-0.2486251444,-0.4871396124,0.0777586028,0.0057945508,0.2682496607,-0.0086447513,0.0188808274,-0.7526343465,0.0099313119,0.1184596121,0.2236755043,-0.3405580521,-0.1409540623,-0.1712234914,-0.1047271192,-0.0248475093,-0.6468246579,-0.7372848392,0.0305309724,-0.1459040195,-0.0095176212,0.0076139257,0.0457624756,-0.0772762448,0.0728076622,-0.0398349501,-0.2196210474,-0.4290323853,-0.2455874532,0.2960585952,-0.0311709791,-0.3616074324,0.1702855676,-0.0103291934,0.4533658028,-0.1900149435,-0.3280900419,-0.1765517741,0.3218617141,-0.0206318405,-0.3668645024,-0.0531160161,0.5032516718,-0.0359730758,0.007629755,-0.1105419323,-0.2325088233,0.0713728964,0.3401716053,0.5897239447,-0.5921173692,0.4528929889,0.2750665843,0.2953697443,0.11580953,0.1370684952,0.1581324339,-0.028745966,0.1398789883,0.1614728719,-0.1506821811,0.5780953765,0.071333766,0.0526375398,0.3350735307,-0.0279414132,-0.3929462731,-0.1691907495,-0.1285373718,-0.078694582,-0.1181916595,-0.0027435708,-0.1899477243,0.3600896001,0.2566945851,0.4055584669,-0.0171049815,-0.0736015216,-0.0285912305,0.040048413,0.3247719407,0.1903122365,-0.6184052825,-0.5788437128,-0.1188429892,0.1225241497,-0.0285572894,0.5415173769,-0.1055406705,0.2024658918,0.2616970539,0.1445031613,0.5555588603,-0.243068099,-0.3454247117,-0.1314902902,0.2881118655,-0.1918623447,0.0003883907,0.0294636525,0.3925947845,0.4314259887,-0.2314029634,-0.008748522,-0.0813297331,0.3801347017,-0.0343821049,-0.2577846348,0.0386747457,-0.2340866774,-0.1954464167,-0.2207710594,0.1956864446,0.0967207029,-0.1245613322,0.2882359028,-0.1875129789,0.2321540117,0.1594732106,-0.0067619258,0.1188303903,0.2844282091,0.09320236,-0.0574472174,0.1403884292,0.2352664322,-0.0003655487,0.1646569967,-0.0287074093,-0.2687040865,0.3077233434,-0.3710045815,0.6595478058,0.2251626551,0.0923566073,0.1156473532,0.3028041124,-0.2651201785,0.0741765052,-0.230508402,0.2964482009,0.2582289577,-0.2570072114,-0.2314118445,-0.0073866863,-0.0777817443,0.1871107817,-0.2283021212,-0.0687551647,-0.1671080589,0.4020507336,0.1825669706,0.8222812414,0.041186057,0.1029222086,-0.1251330674,-0.1052090526,0.263160497,-0.195330292,0.2854102254,-0.227368176,-0.0576222613,-0.1682812124,0.0172704086,0.1919307858,0.4000087976,-0.3338609636,0.2994181216,0.3004610538,-0.1776118428,0.2175974995,0.0608496964,0.0637610033,-0.2100925744,-0.3378998339,0.1408398598,0.0805336609,0.364249438,-0.244950667,-0.1419679224,0.1068861783,-0.0784856454,-0.1326034218,0.0144890482,-0.3802137077,0.187495023,-0.0113335103,-0.4910606742,0.2870213687,0.4773972332,0.3643172085,0.3550560474,-0.2171392739,0.5245866179,0.1262669861,0.014424026,0.3423662484,0.0047789975,0.2060928643,-0.2244211584,-0.3705042601,0.2566948831,-0.096159175,-0.0323477201,-0.5521182418,0.2968838811,0.1679379344,0.1947372407,0.0181638598,0.0970331579,0.0722471848,-0.2222664356,0.1337298304,0.2241173536,-0.0378129706,0.2564706206,0.1661891341,-0.1043714508,-0.1268913895,0.3687799275,0.2147459239,-0.0532688238,0.5588973761,0.0592473447,0.0342613608,-0.2540690005,0.1239341646,-0.0739686787,-0.137568593,-0.2623916566,0.1452253759,0.1997595578,-0.029299343,-0.2072233856,-0.0644626692,-0.3883130848,-0.0622428618,-0.3872435987,-0.2752884328,0.0754099935,-0.1133893058,-0.122931309,0.212403819,0.1649701595,0.2495986372,0.0781828687,-0.3227564394,0.1679518074,-0.397331059,0.2239514887,-0.0156782884,-0.2213269621,0.1106734127,-0.1090313047,0.1154885441,-0.0890560597,-0.0444083512,-0.146446377,0.212185517,0.0551630221,-0.0651092529,0.2640431225,0.0133864703,-0.1056011915,-0.2285670787,0.2656398118,-0.115811795,-0.00325076,0.0955168903,0.3260475695,-0.1100255474,0.0383170247,-0.0806022063,-0.1450048089,0.3147141337,-0.0462802872,0.1328765899,0.3403069079,0.2551241517,-0.036593955,-0.1440806836,-0.0459792838,-0.0227454267,0.2281866074,0.1698565334,-0.3522877097,0.0159063097,0.1860785186,-0.2250193357,-0.0564581342,-0.3032708764,0.2272311151,0.363715589,0.226995945,-0.5343983769,0.0066712894,0.2371629775,-0.1729505509,-0.0136545273,0.0114739258,0.3277379274,0.2638809383,0.4061787724,0.2645318508,0.1514029205,-0.3466751873,0.030294735,-0.0275240764,0.0107518882,-0.0151055707,0.225854516,-0.1222212166,-0.0129909609,0.45652318,0.0838124529,0.2531338036,0.3350505233,0.1993779838,0.1543693542,-0.1966710091,-0.2339617014,0.1665221304,-0.0543207861,-0.0105108432,0.1204886734,0.4063559175,0.0876704305,-0.4291986823,-0.2303833812,0.1484030336,-0.4381048381,0.0428760201,0.09610603,-0.0571990944,0.1284685135,-0.0854215994,0.0644628927,-0.061209131,-0.2336712182,0.1906644851,-0.1822262555,0.1038997993,0.0218742397,0.2557960749,-0.2054534853,-0.1098108441,0.3854257762,-0.1390029341,-0.1274639517,0.0623377152,-0.0519584417,-0.0155729828,-0.1701721698,-0.163791433,0.157417804,0.1103963405,-0.1693791151,0.0059025213,0.3001666069,0.4074698985,0.0573103763,0.1915333271,0.1729009002,-0.0654352978,0.0593618192,0.2738451064,-0.0784710646,-0.1262728125,-0.0235403683,0.0622575432,0.0210812017,-0.3662447333,-0.1674963683,-0.3806779385,-0.0889189392,0.0384859145,-0.0417643264,-0.0341717303,-0.1998297125,0.1103359386,-0.1457634419,0.2659996748,0.338637799,0.0957281813,-0.150683701,-0.20722498,-0.4805024564,-0.0058464799,0.0607658885,0.1942615658,-0.0576737076,0.326764673,0.3132946193,0.1676271409,0.2692517638,0.0158548374,-0.0096586794,-0.1015057862,-0.5356661677,-0.1428379416,0.0602541342,0.3365083635,-0.0301467422,-0.02764697,0.1668958068,0.2043222934,0.0097369244,-0.3536172807,-0.1598585397,-0.0763357505,-0.0097773364,0.3959164619,0.0836950615,0.3392042518,-0.0668989569,0.171293661,0.1589923203,-0.1978260428,-0.0206469782,0.2478114665,0.1892030388,0.1112448499,-0.0699778423,-0.2354997247,0.0189300086,0.3167833388,0.1663195342,-0.1713151634,-0.4148258567,0.368324846,0.0656671524,0.1231911182,-0.1091120839,0.2324760854,0.0158612132,-0.0653180107,-0.2101466656,0.0055043995,0.3044301271,-0.1566054672,0.0137655307,-0.3062791526,0.0810980648,-0.1050598249,0.0411877632,-0.2789291441,-0.089605093,0.3944664598,0.0517788492,0.0519359931,0.2999483645,0.0418099053,0.1855539829,-0.1216883287,0.1529166251,0.0919364169,-0.1553381234,0.1783640534,-0.1917356104]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2134","title":"Saving large in-memory datasets with save_to_disk crashes because of pickling","comments":"Hi !\r\nIndeed `save_to_disk` doesn't call pickle anymore. Though the `OverflowError` can still appear for in-memory datasets bigger than 4GB. This happens when doing this for example:\r\n```python\r\nimport pyarrow as pa\r\nimport pickle\r\n\r\narr = pa.array([0] * ((4 * 8 << 30) \/\/ 64))\r\ntable = pa.Table.from_arrays([a], names=[\"foo\"])\r\npickle.dumps(table) # fails with an OverflowError\r\npickle.dumps(table, 4) # works !\r\n```\r\nWe'll do the change to use `protocol=4`.\r\n\r\nMoreover I've also seen other users complain about this error\r\n```\r\nstruct.error: 'I' format requires 0 <= number <= 4294967295\r\n```\r\n\r\nIt looks like something related to the 4GB limit as well but I'm not able to reproduce on my side.\r\nDo you think you can provide a script that reproduces the issue ?\r\nHow big is your dataset ? (number of bytes, number of rows)\r\n\r\n","body":"Using Datasets 1.5.0 on Python 3.7.\r\nRecently I've been working on medium to large size datasets (pretokenized raw text sizes from few gigabytes to low tens of gigabytes), and have found out that several preprocessing steps are massively faster when done in memory, and I have the ability to requisition a lot of RAM, so I decided to do these steps completely out of the datasets library.\r\n\r\n So my workflow is to do several .map() on datasets object, then for the operation which is faster in memory to extract the necessary columns from the dataset and then drop it whole, do the transformation in memory, and then create a fresh Dataset object using .from_dict() or other method. \r\n\r\nWhen I then try to call save_to_disk(path) on the dataset, it crashes because of pickling, which appears to be because of using old pickle protocol which doesn't support large files (over 4 GiB).\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 80, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n large_dataset.map(function=None, num_proc=2)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n large_dataset.map(function=None, num_proc=2)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 75, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 60, in tokenize_and_chunkify\r\n contexts_dataset.save_to_disk(chunked_path)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 457, in save_to_disk\r\n self = pickle.loads(pickle.dumps(self))\r\nOverflowError: cannot serialize a bytes object larger than 4 GiB\r\n```\r\nFrom what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository.\r\n\r\nTo save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk.\r\n\r\nAdditional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that.\r\n```\r\nTraceback (most recent call last):\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 94, in \r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\"\r\n main()\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 89, in main\r\n tokenize_and_chunkify(config)\r\n File \".\/tokenize_and_chunkify_in_memory.py\", line 67, in tokenize_and_chunkify\r\n contexts_dataset.map(function=None, cache_file_name=str(output_dir_path \/ \"tmp.arrow\"), writer_batch_size=50000, num_proc=config.threads)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1485, in \r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 657, in get\r\n raise self._value\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/pool.py\", line 431, in _handle_tasks\r\n put(task)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 454, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 437, in dump\r\n self.save(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 662, in save_reduce\r\n save(state)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/cernypro\/dev\/envs\/huggingface_gpu\/lib\/python3.7\/site-packages\/dill\/_dill.py\", line 941, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 859, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 885, in _batch_setitems\r\n save(v)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 846, in _batch_appends\r\n save(tmp[0])\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 789, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 819, in save_list\r\n self._batch_appends(obj)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 843, in _batch_appends\r\n save(x)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 549, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 638, in save_reduce\r\n save(args)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 774, in save_tuple\r\n save(element)\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 504, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/mnt\/appl\/software\/Python\/3.7.4-GCCcore-8.3.0\/lib\/python3.7\/pickle.py\", line 732, in save_bytes\r\n self._write_large_bytes(BINBYTES + pack(\">> questions = [\r\n... \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631?\",\r\n... \"\\u0643\\u0645 \\u0645\\u0631\\u0629 \\u064a\\u062a\\u0645 \\u0646\\u0634\\u0631\\u0647\\u0627 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n... \"\\u0645\\u0627 \\u0647\\u064a \\u0627\\u0644\\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u064a\\u0648\\u0645\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n... \"\\u0643\\u0645 \\u0639\\u062f\\u062f \\u0627\\u0644\\u0627\\u0648\\u0631\\u0627\\u0642 \\u0627\\u0644\\u0627\\u062e\\u0628\\u0627\\u0631\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0627\\u0644\\u062a\\u064a \\u0648\\u062c\\u062f\\u062a \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n... \"\\u0641\\u064a \\u0627\\u064a \\u0633\\u0646\\u0629 \\u0628\\u062f\\u0627\\u062a \\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u0637\\u0627\\u0644\\u0628 \\u0627\\u0644\\u062d\\u0633 \\u0627\\u0644\\u0633\\u0644\\u064a\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\"\r\n... ]\r\n>>> print(questions)\r\n['\u0645\u062a\u0649 \u0628\u062f\u0627\u062a \u0627\u0644\u0645\u062c\u0644\u0629 \u0627\u0644\u0645\u062f\u0631\u0633\u064a\u0629 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645 \u0628\u0627\u0644\u0646\u0634\u0631?', '\u0643\u0645 \u0645\u0631\u0629 \u064a\u062a\u0645 \u0646\u0634\u0631\u0647\u0627 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?', '\u0645\u0627 \u0647\u064a \u0627\u0644\u0648\u0631\u0642\u0629 \u0627\u0644\u064a\u0648\u0645\u064a\u0629 \u0644\u0644\u0637\u0644\u0627\u0628 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?', '\u0643\u0645 \u0639\u062f\u062f \u0627\u0644\u0627\u0648\u0631\u0627\u0642 \u0627\u0644\u0627\u062e\u0628\u0627\u0631\u064a\u0629 \u0644\u0644\u0637\u0644\u0627\u0628 \u0627\u0644\u062a\u064a \u0648\u062c\u062f\u062a \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?', '\u0641\u064a \u0627\u064a \u0633\u0646\u0629 \u0628\u062f\u0627\u062a \u0648\u0631\u0642\u0629 \u0627\u0644\u0637\u0627\u0644\u0628 \u0627\u0644\u062d\u0633 \u0627\u0644\u0633\u0644\u064a\u0645 \u0628\u0627\u0644\u0646\u0634\u0631 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?']\r\n```\r\nI don't think we can change this","body":"Hi \r\nLooking into MLQA dataset for langauge \"ar\":\r\n\r\n```\r\n \"question\": [\r\n \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631?\",\r\n \"\\u0643\\u0645 \\u0645\\u0631\\u0629 \\u064a\\u062a\\u0645 \\u0646\\u0634\\u0631\\u0647\\u0627 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0645\\u0627 \\u0647\\u064a \\u0627\\u0644\\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u064a\\u0648\\u0645\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0643\\u0645 \\u0639\\u062f\\u062f \\u0627\\u0644\\u0627\\u0648\\u0631\\u0627\\u0642 \\u0627\\u0644\\u0627\\u062e\\u0628\\u0627\\u0631\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0627\\u0644\\u062a\\u064a \\u0648\\u062c\\u062f\\u062a \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0641\\u064a \\u0627\\u064a \\u0633\\u0646\\u0629 \\u0628\\u062f\\u0627\\u062a \\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u0637\\u0627\\u0644\\u0628 \\u0627\\u0644\\u062d\\u0633 \\u0627\\u0644\\u0633\\u0644\\u064a\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\"\r\n ]\r\n```\r\n\r\nthe questions are in the wrong format, and not readable, could you please have a look? thanks @lhoestq \r\n","comment_length":111,"text":"bug in mlqa dataset \nHi \r\nLooking into MLQA dataset for langauge \"ar\":\r\n\r\n```\r\n \"question\": [\r\n \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631?\",\r\n \"\\u0643\\u0645 \\u0645\\u0631\\u0629 \\u064a\\u062a\\u0645 \\u0646\\u0634\\u0631\\u0647\\u0627 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0645\\u0627 \\u0647\\u064a \\u0627\\u0644\\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u064a\\u0648\\u0645\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0643\\u0645 \\u0639\\u062f\\u062f \\u0627\\u0644\\u0627\\u0648\\u0631\\u0627\\u0642 \\u0627\\u0644\\u0627\\u062e\\u0628\\u0627\\u0631\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0627\\u0644\\u062a\\u064a \\u0648\\u062c\\u062f\\u062a \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0641\\u064a \\u0627\\u064a \\u0633\\u0646\\u0629 \\u0628\\u062f\\u0627\\u062a \\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u0637\\u0627\\u0644\\u0628 \\u0627\\u0644\\u062d\\u0633 \\u0627\\u0644\\u0633\\u0644\\u064a\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\"\r\n ]\r\n```\r\n\r\nthe questions are in the wrong format, and not readable, could you please have a look? thanks @lhoestq \r\n\nIf you print those questions, you get readable texts:\r\n```python\r\n>>> questions = [\r\n... \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631?\",\r\n... \"\\u0643\\u0645 \\u0645\\u0631\\u0629 \\u064a\\u062a\\u0645 \\u0646\\u0634\\u0631\\u0647\\u0627 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n... \"\\u0645\\u0627 \\u0647\\u064a \\u0627\\u0644\\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u064a\\u0648\\u0645\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n... \"\\u0643\\u0645 \\u0639\\u062f\\u062f \\u0627\\u0644\\u0627\\u0648\\u0631\\u0627\\u0642 \\u0627\\u0644\\u0627\\u062e\\u0628\\u0627\\u0631\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0627\\u0644\\u062a\\u064a \\u0648\\u062c\\u062f\\u062a \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n... \"\\u0641\\u064a \\u0627\\u064a \\u0633\\u0646\\u0629 \\u0628\\u062f\\u0627\\u062a \\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u0637\\u0627\\u0644\\u0628 \\u0627\\u0644\\u062d\\u0633 \\u0627\\u0644\\u0633\\u0644\\u064a\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\"\r\n... ]\r\n>>> print(questions)\r\n['\u0645\u062a\u0649 \u0628\u062f\u0627\u062a \u0627\u0644\u0645\u062c\u0644\u0629 \u0627\u0644\u0645\u062f\u0631\u0633\u064a\u0629 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645 \u0628\u0627\u0644\u0646\u0634\u0631?', '\u0643\u0645 \u0645\u0631\u0629 \u064a\u062a\u0645 \u0646\u0634\u0631\u0647\u0627 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?', '\u0645\u0627 \u0647\u064a \u0627\u0644\u0648\u0631\u0642\u0629 \u0627\u0644\u064a\u0648\u0645\u064a\u0629 \u0644\u0644\u0637\u0644\u0627\u0628 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?', '\u0643\u0645 \u0639\u062f\u062f \u0627\u0644\u0627\u0648\u0631\u0627\u0642 \u0627\u0644\u0627\u062e\u0628\u0627\u0631\u064a\u0629 \u0644\u0644\u0637\u0644\u0627\u0628 \u0627\u0644\u062a\u064a \u0648\u062c\u062f\u062a \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?', '\u0641\u064a \u0627\u064a \u0633\u0646\u0629 \u0628\u062f\u0627\u062a \u0648\u0631\u0642\u0629 \u0627\u0644\u0637\u0627\u0644\u0628 \u0627\u0644\u062d\u0633 \u0627\u0644\u0633\u0644\u064a\u0645 \u0628\u0627\u0644\u0646\u0634\u0631 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?']\r\n```\r\nI don't think we can change this","embeddings":[-0.115853101,-0.231567964,-0.2734039724,0.2112123966,0.3321486115,0.0711947754,0.3524691463,0.1213381737,-0.3227720559,0.2254540175,0.0618873164,0.2817641795,0.2518624663,0.2561215162,0.1166762114,-0.047785569,0.1292088628,0.179456532,0.0600869805,-0.197704494,-0.196472913,0.2453739792,-0.1337222606,0.0539799668,-0.0679920763,-0.033509817,-0.0207405165,0.220893696,-0.0425398685,0.0830165371,-0.1433827132,-0.2492368817,-0.1639159769,0.005175015,-0.0001021291,-0.1083842143,0.1797213703,-0.0530055612,-0.2548291981,0.3303485215,-0.4433495402,0.0542143062,-0.3197640181,-0.3245649338,-0.18316935,-0.1904883683,-0.05527585,-0.5454843044,0.3158854544,0.3931446373,0.3555622101,0.1169286147,-0.0093527725,-0.0591307953,0.2337930202,-0.0075344178,0.0532882288,-0.0125398654,0.0099399295,-0.0157065429,0.0644993559,0.5903977156,0.1528967172,-0.0744416192,-0.2841548622,0.0947690308,0.2282186002,-0.1410573274,0.3744464219,0.2255229354,0.2068464309,-0.2394515574,-0.1113861203,0.070110254,0.03870086,-0.1902514845,-0.0511132963,0.2683143914,0.0487997532,0.0880704224,0.1475322396,0.1784594804,-0.1978724748,0.0273552109,-0.4308957458,0.4159484506,-0.1553923637,-0.0334863551,-0.1144851521,-0.2438467145,-0.1090199426,0.1160906479,-0.1431100965,0.1030091941,-0.1794525683,-0.188223511,-0.1247729361,0.2676885426,0.013106185,-0.0319190025,0.1373087317,0.1452646405,0.2267717719,-0.0552490279,-0.1274060756,-0.010375008,-0.1002315283,-0.115046382,0.2245224118,-0.1293273121,-0.0769749582,0.015334459,0.0609565601,-0.2191800922,-0.1487826705,0.2285619527,0.1345374286,-0.1410000473,-0.3242001534,0.056123469,-0.3493770361,0.0548069552,-0.1677003652,0.391261518,0.2288828194,-0.0240960065,-0.139952004,-0.0753454566,-0.1917124987,-0.3327361643,-0.3898272812,0.1255715787,0.1248719692,-0.27682212,-0.0159336664,0.3214123547,0.215539068,-0.0244428106,0.1260242909,0.0829229057,0.0949713215,-0.1912811249,0.2312819362,0.2930064201,-0.1015833616,-0.0493383966,-0.0328686349,-0.0564297847,0.0095267622,0.1452571899,0.0762135983,0.0079340953,-0.3933087587,0.3797929287,0.395275563,-0.0949567109,0.1754642427,0.2335368693,0.2187478989,0.0496398285,0.1180229709,-0.1062783301,-0.0219002888,-0.1518765986,0.063460812,0.0898392871,-0.526848197,0.0284028426,-0.3689579368,-0.0933342502,0.3024438322,0.3060186207,0.0991310477,-0.0708538219,-0.170061931,0.3025590479,0.1694825292,-0.0779909417,-0.3400383294,0.2042782009,-0.159998998,-0.1875737309,0.2839224935,0.2088894844,-0.1037751511,-0.0227849204,0.1018831581,0.1511971503,-0.1055432111,0.0276853014,-0.123123616,-0.1423805505,0.1725804061,0.0685132965,-0.2583670616,-0.1137779951,0.2595126629,0.024177283,0.3915384114,-0.0147941019,-0.0749591962,0.0135183893,0.3583312035,-0.4317682683,0.162547648,-0.0722043589,-0.3222065568,-0.219033882,-0.609854579,0.0809970871,-0.233429864,-0.1778639853,-0.1760995686,-0.1031846404,0.0356507711,0.0966448039,0.3250931799,0.1980949938,-0.0805992335,0.2429176867,0.0244634617,-0.0335509703,-0.0382368453,-0.1474969685,-0.3175995946,0.2270222902,-0.1953133941,0.0706109107,0.1879199296,0.3434305489,0.1390678883,-0.0944974348,0.1414386928,0.0136982501,0.095292449,0.2897173464,-0.1747563481,0.1883360296,0.2945405543,-0.2585287094,0.0214189049,0.3033981323,0.3103998601,0.0042769727,-0.1195691824,0.3716930151,0.0615129918,-0.0408308692,0.0971447751,0.029447915,0.4241673052,-0.0936098993,0.1024363488,-0.5371354222,0.168956995,0.1346676648,0.2095032185,0.0624080636,-0.5084814429,0.439779222,0.4208018482,-0.0966070369,0.0676753074,-0.1444756836,-0.1639984995,0.0293982271,0.0471223593,0.1910907626,0.2587710917,0.2780264914,0.0747869611,0.0024714572,-0.0050052307,-0.4039029181,0.34732306,-0.0582332052,0.0497225896,0.2490378618,0.1140516698,-0.0766907632,-0.3842312396,0.1208591983,-0.0701796636,0.0866600648,-0.2455856353,0.1962551326,-0.3665637672,-0.5910026431,-0.3121281564,-0.2207443565,0.1649776548,-0.2732044756,0.28374663,-0.1319197714,-0.2716562748,0.2348254919,0.1886444241,0.3562226892,0.238636449,0.2746743858,-0.310839653,-0.0177151002,-0.1842351109,0.310754627,0.1267085671,0.2490748763,0.1786777973,-0.2989786267,-0.1730295867,-0.2467732877,-0.4817385077,0.089297086,-0.032913927,0.3612640798,0.0369145088,-0.0061595929,-0.1015347168,-0.0593845882,-0.0475925393,0.1554066539,-0.321716845,-0.099974297,-0.1066943109,-0.0564780273,-0.2586625516,-0.5629889369,-0.2710545957,-0.1909809411,-0.0850012749,-0.0118034007,0.2423169762,-0.1813126653,-0.0354815312,0.1197240576,0.090256907,-0.0801457763,-0.3614607751,0.2991599441,0.361500293,-0.153798148,-0.4440942407,-0.1383514553,-0.1057202816,0.2217350155,-0.3010010421,-0.3994612396,-0.0869593993,-0.140109688,-0.2230976373,-0.0174844526,-0.073769547,0.2164241225,-0.1130887046,-0.2328870147,-0.0172807127,-0.1985118836,0.1154638603,0.130371049,0.169789657,-0.1722306162,0.5826883912,0.0520810261,0.1483297944,0.1584417224,-0.1299767792,0.171134308,-0.1682365537,0.1268296391,-0.0480326973,-0.045324862,0.3635538816,0.1533102244,-0.0188578218,0.3690085411,-0.057263691,-0.2640039027,0.1094783768,0.0063495804,-0.0653221384,-0.1079196334,0.0292380881,-0.2916397452,-0.0198185183,0.083358869,0.0959740281,0.1944290251,-0.1557240486,0.113623023,0.0154413022,0.0269193277,-0.1321808696,-0.531431973,-0.2443500906,-0.2926125526,0.3258085847,0.0728762224,0.3792927265,-0.2615154386,-0.1328953952,0.1868347973,-0.1185835674,0.5819468498,0.1304207742,0.1935622245,0.2138529122,0.1884213239,0.0790649056,0.1521455199,-0.466581583,0.3041167557,0.1202671006,0.2619284689,-0.1860757321,0.0104635172,0.1320597976,0.2706112862,-0.4115587473,0.0313057527,-0.3207559586,-0.2958230376,-0.0329888202,0.1487252116,0.1233900636,0.3150461614,-0.0871901885,-0.0636227503,-0.1172879636,-0.4636466801,0.04573467,0.2088248581,0.3036416471,-0.0808656588,-0.0717990547,-0.1067918241,0.3183946908,-0.1104695424,0.330857873,0.1301068366,-0.0969371796,0.0098023294,-0.4112183452,0.2951388359,0.0367309451,-0.1786257476,-0.0091440324,0.4109277129,0.0753104836,-0.1215833947,-0.1314014047,0.0782288015,0.154828459,-0.1019289941,-0.066371046,0.210072428,-0.2370853871,-0.1019979715,-0.0266773477,0.3568893075,-0.3343978524,0.4691640139,0.230417937,0.9524292946,0.2300878614,0.004815958,0.2026661187,-0.2724691927,0.0495895818,0.2469626069,0.3125798702,-0.3602144122,0.0808099508,-0.1237326413,-0.0259009004,0.1456841528,0.2686609328,-0.2391446084,0.0368269384,-0.0991390571,0.0187147185,0.0578457825,-0.0420080163,0.0123539818,0.0667764693,-0.3098329008,0.2407772094,0.1604472995,-0.1262230128,-0.1036741138,-0.1553587019,0.0229907036,-0.1915262491,-0.0891853645,0.0344958864,-0.4054076672,0.0044303206,0.1254346073,-0.3809590042,-0.110140115,0.1247104555,0.2879429758,0.1120830551,0.0355701856,0.1754250377,0.2798658609,0.333129853,0.2754772604,0.0300889499,0.2048370242,-0.3041776121,-0.1624447554,0.2047544718,0.11769142,-0.1924744099,-0.2655920088,0.2533445358,0.0037148867,0.0201470908,-0.1261609644,-0.0415273048,-0.0486049242,-0.2539508045,0.2601626217,-0.1233119741,-0.1218632832,0.3450837135,-0.015102515,-0.2232404798,-0.1870571226,0.2631909251,0.0915394872,0.2749405801,0.4604128301,0.0735991597,-0.2701103687,-0.467040658,-0.1441354305,-0.0737694427,-0.2279803902,0.1191655472,-0.1223850623,-0.2254580855,-0.0895111784,0.0805105567,-0.0582050197,0.1274322569,-0.1490082294,-0.61260432,-0.2825827301,-0.0735413954,0.0543799847,0.1607298106,0.1171972975,0.0226549432,-0.1883123368,0.0475361571,-0.4655611515,0.1205714047,-0.1525133252,0.068241179,-0.3576040566,0.0517567545,-0.056235522,0.1945682317,0.4125513434,-0.0648631826,-0.2503980994,-0.4053132534,-0.0830439925,0.0675584599,-0.1761305034,-0.1299041361,0.0353784822,-0.0416139252,-0.2331759036,0.1856706142,0.0996722728,-0.1140601933,0.0705626085,0.0362370946,0.1259871125,-0.0754710361,-0.0776163861,-0.0489319339,0.1121360734,-0.1620777547,0.1734581739,0.2188940048,-0.0234114621,-0.0476469025,0.3097182214,-0.1309831887,-0.0507532395,0.0357708558,0.3090589345,-0.1389124244,-0.1823407859,0.2213166058,0.0059361998,0.0923449323,-0.3267617226,-0.1344147474,0.0450517796,0.4097962677,-0.5532665849,-0.0793981999,0.0638042241,-0.0084645161,0.1509258747,0.1421750039,0.2615834475,-0.1343318373,0.126941666,0.2209291607,0.3258605301,-0.2247743607,-0.1728530079,0.050757736,-0.1218095198,0.1692665368,0.112531893,0.0999643132,0.1192644089,0.6916362047,-0.0437443219,0.2923962772,0.1101408973,0.0679828003,0.0815226734,0.1393963844,0.1451892406,0.4456792772,-0.2920979559,0.1568640769,0.5386681557,-0.1165414453,0.2051559836,0.0246807523,-0.288428098,-0.148318544,-0.291851759,-0.2522684932,-0.0286990777,-0.1616626233,-0.0294934344,-0.1111187562,0.0027753818,0.0421171524,0.5325813293,-0.007183094,-0.0443251655,-0.1829901636,-0.0525388308,-0.0752628222,-0.0456416085,-0.1999215931,0.3746516109,0.3940322697,-0.0059070871,0.2505495846,-0.0202580262,0.1869512349,0.0737290159,-0.0990660191,0.0158184003,-0.1997042149,-0.2816879153,-0.184263289,0.2646845281,0.1495168954,0.3046125472,0.2148529142,0.3038095236,-0.2171559036,0.1580083668,0.5882334113,0.151739046,-0.0693793744,0.0392128937,-0.4237434268,-0.0257013272,-0.3209612668,0.0478476658,-0.2464501858,0.077816397,0.0034496307,-0.2867472768,0.1594921201,0.1201608777,0.1503569931,0.0604840592,0.6015758514,0.1133554652,-0.0661049932,-0.1337702423,-0.2346678823,-0.5674973726,0.1958338916,0.2363114953,-0.0519492887,-0.1713323146,0.3429603279,0.1597278863,-0.0710562691,0.0902101845,0.0962439999,-0.0821874961,-0.0162224844,-0.4123512506,-0.106794402,0.1234901175,0.2328350395,-0.1281647086,-0.3708758056,0.3792416453,0.0888491347,0.1948824972,-0.2314056754,-0.2158889771,-0.1824168116,0.2466814965,0.0318037346,0.3799424767,0.1526882946,-0.3498205245,0.0361984484,0.1120776013,0.0147677688,-0.1114462987,0.2188643962,0.2435276508,0.2173151821,0.1415945441,0.0435096994,0.1522723287,0.4297482073,0.0463595279,-0.1174224317,-0.2635605335,0.5547003746,-0.1008020863,-0.0520818941,-0.1178275943,0.1396929026,0.0085183904,-0.0532908775,-0.0157468654,-0.4543573856,0.3082257807,-0.5355829597,-0.2539766133,-0.0591680072,0.1182153746,0.1231312752,-0.0180409644,-0.2814457119,0.0923632756,0.3500164449,-0.2909099758,0.1102767512,0.4204139411,0.1702205241,-0.0142796729,-0.2048442066,0.1536104232,0.4446440935,-0.1842172146,-0.048598107,-0.4758795798]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2133","title":"bug in mlqa dataset ","comments":"Hi @dorost1234.\r\n\r\nIn Python 3, strings are sequences of Unicode _code points_. Unicode is a specification that maps all characters (and emoji symbols) with its unique representation in terms of code points. That is what you see: Unicode code points (represented by a \\u escaped sequence of 16-bit hex values).\r\n\r\nCharacters are usually represented (on screen and papers) with a graphical element called _glyph_. That is what you would like to see: glyphs. But Python does not care about glyphs: that is the job of the GUI or the terminal; glyphs are what you get with the `print` function (if your terminal is properly configured to display those glyphs).\r\n\r\nYou have more detailed information about Unicode in the Python documentation: https:\/\/docs.python.org\/3\/howto\/unicode.html","body":"Hi \r\nLooking into MLQA dataset for langauge \"ar\":\r\n\r\n```\r\n \"question\": [\r\n \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631?\",\r\n \"\\u0643\\u0645 \\u0645\\u0631\\u0629 \\u064a\\u062a\\u0645 \\u0646\\u0634\\u0631\\u0647\\u0627 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0645\\u0627 \\u0647\\u064a \\u0627\\u0644\\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u064a\\u0648\\u0645\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0643\\u0645 \\u0639\\u062f\\u062f \\u0627\\u0644\\u0627\\u0648\\u0631\\u0627\\u0642 \\u0627\\u0644\\u0627\\u062e\\u0628\\u0627\\u0631\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0627\\u0644\\u062a\\u064a \\u0648\\u062c\\u062f\\u062a \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0641\\u064a \\u0627\\u064a \\u0633\\u0646\\u0629 \\u0628\\u062f\\u0627\\u062a \\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u0637\\u0627\\u0644\\u0628 \\u0627\\u0644\\u062d\\u0633 \\u0627\\u0644\\u0633\\u0644\\u064a\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\"\r\n ]\r\n```\r\n\r\nthe questions are in the wrong format, and not readable, could you please have a look? thanks @lhoestq \r\n","comment_length":121,"text":"bug in mlqa dataset \nHi \r\nLooking into MLQA dataset for langauge \"ar\":\r\n\r\n```\r\n \"question\": [\r\n \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631?\",\r\n \"\\u0643\\u0645 \\u0645\\u0631\\u0629 \\u064a\\u062a\\u0645 \\u0646\\u0634\\u0631\\u0647\\u0627 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0645\\u0627 \\u0647\\u064a \\u0627\\u0644\\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u064a\\u0648\\u0645\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0643\\u0645 \\u0639\\u062f\\u062f \\u0627\\u0644\\u0627\\u0648\\u0631\\u0627\\u0642 \\u0627\\u0644\\u0627\\u062e\\u0628\\u0627\\u0631\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0627\\u0644\\u062a\\u064a \\u0648\\u062c\\u062f\\u062a \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0641\\u064a \\u0627\\u064a \\u0633\\u0646\\u0629 \\u0628\\u062f\\u0627\\u062a \\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u0637\\u0627\\u0644\\u0628 \\u0627\\u0644\\u062d\\u0633 \\u0627\\u0644\\u0633\\u0644\\u064a\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\"\r\n ]\r\n```\r\n\r\nthe questions are in the wrong format, and not readable, could you please have a look? thanks @lhoestq \r\n\nHi @dorost1234.\r\n\r\nIn Python 3, strings are sequences of Unicode _code points_. Unicode is a specification that maps all characters (and emoji symbols) with its unique representation in terms of code points. That is what you see: Unicode code points (represented by a \\u escaped sequence of 16-bit hex values).\r\n\r\nCharacters are usually represented (on screen and papers) with a graphical element called _glyph_. That is what you would like to see: glyphs. But Python does not care about glyphs: that is the job of the GUI or the terminal; glyphs are what you get with the `print` function (if your terminal is properly configured to display those glyphs).\r\n\r\nYou have more detailed information about Unicode in the Python documentation: https:\/\/docs.python.org\/3\/howto\/unicode.html","embeddings":[-0.115853101,-0.231567964,-0.2734039724,0.2112123966,0.3321486115,0.0711947754,0.3524691463,0.1213381737,-0.3227720559,0.2254540175,0.0618873164,0.2817641795,0.2518624663,0.2561215162,0.1166762114,-0.047785569,0.1292088628,0.179456532,0.0600869805,-0.197704494,-0.196472913,0.2453739792,-0.1337222606,0.0539799668,-0.0679920763,-0.033509817,-0.0207405165,0.220893696,-0.0425398685,0.0830165371,-0.1433827132,-0.2492368817,-0.1639159769,0.005175015,-0.0001021291,-0.1083842143,0.1797213703,-0.0530055612,-0.2548291981,0.3303485215,-0.4433495402,0.0542143062,-0.3197640181,-0.3245649338,-0.18316935,-0.1904883683,-0.05527585,-0.5454843044,0.3158854544,0.3931446373,0.3555622101,0.1169286147,-0.0093527725,-0.0591307953,0.2337930202,-0.0075344178,0.0532882288,-0.0125398654,0.0099399295,-0.0157065429,0.0644993559,0.5903977156,0.1528967172,-0.0744416192,-0.2841548622,0.0947690308,0.2282186002,-0.1410573274,0.3744464219,0.2255229354,0.2068464309,-0.2394515574,-0.1113861203,0.070110254,0.03870086,-0.1902514845,-0.0511132963,0.2683143914,0.0487997532,0.0880704224,0.1475322396,0.1784594804,-0.1978724748,0.0273552109,-0.4308957458,0.4159484506,-0.1553923637,-0.0334863551,-0.1144851521,-0.2438467145,-0.1090199426,0.1160906479,-0.1431100965,0.1030091941,-0.1794525683,-0.188223511,-0.1247729361,0.2676885426,0.013106185,-0.0319190025,0.1373087317,0.1452646405,0.2267717719,-0.0552490279,-0.1274060756,-0.010375008,-0.1002315283,-0.115046382,0.2245224118,-0.1293273121,-0.0769749582,0.015334459,0.0609565601,-0.2191800922,-0.1487826705,0.2285619527,0.1345374286,-0.1410000473,-0.3242001534,0.056123469,-0.3493770361,0.0548069552,-0.1677003652,0.391261518,0.2288828194,-0.0240960065,-0.139952004,-0.0753454566,-0.1917124987,-0.3327361643,-0.3898272812,0.1255715787,0.1248719692,-0.27682212,-0.0159336664,0.3214123547,0.215539068,-0.0244428106,0.1260242909,0.0829229057,0.0949713215,-0.1912811249,0.2312819362,0.2930064201,-0.1015833616,-0.0493383966,-0.0328686349,-0.0564297847,0.0095267622,0.1452571899,0.0762135983,0.0079340953,-0.3933087587,0.3797929287,0.395275563,-0.0949567109,0.1754642427,0.2335368693,0.2187478989,0.0496398285,0.1180229709,-0.1062783301,-0.0219002888,-0.1518765986,0.063460812,0.0898392871,-0.526848197,0.0284028426,-0.3689579368,-0.0933342502,0.3024438322,0.3060186207,0.0991310477,-0.0708538219,-0.170061931,0.3025590479,0.1694825292,-0.0779909417,-0.3400383294,0.2042782009,-0.159998998,-0.1875737309,0.2839224935,0.2088894844,-0.1037751511,-0.0227849204,0.1018831581,0.1511971503,-0.1055432111,0.0276853014,-0.123123616,-0.1423805505,0.1725804061,0.0685132965,-0.2583670616,-0.1137779951,0.2595126629,0.024177283,0.3915384114,-0.0147941019,-0.0749591962,0.0135183893,0.3583312035,-0.4317682683,0.162547648,-0.0722043589,-0.3222065568,-0.219033882,-0.609854579,0.0809970871,-0.233429864,-0.1778639853,-0.1760995686,-0.1031846404,0.0356507711,0.0966448039,0.3250931799,0.1980949938,-0.0805992335,0.2429176867,0.0244634617,-0.0335509703,-0.0382368453,-0.1474969685,-0.3175995946,0.2270222902,-0.1953133941,0.0706109107,0.1879199296,0.3434305489,0.1390678883,-0.0944974348,0.1414386928,0.0136982501,0.095292449,0.2897173464,-0.1747563481,0.1883360296,0.2945405543,-0.2585287094,0.0214189049,0.3033981323,0.3103998601,0.0042769727,-0.1195691824,0.3716930151,0.0615129918,-0.0408308692,0.0971447751,0.029447915,0.4241673052,-0.0936098993,0.1024363488,-0.5371354222,0.168956995,0.1346676648,0.2095032185,0.0624080636,-0.5084814429,0.439779222,0.4208018482,-0.0966070369,0.0676753074,-0.1444756836,-0.1639984995,0.0293982271,0.0471223593,0.1910907626,0.2587710917,0.2780264914,0.0747869611,0.0024714572,-0.0050052307,-0.4039029181,0.34732306,-0.0582332052,0.0497225896,0.2490378618,0.1140516698,-0.0766907632,-0.3842312396,0.1208591983,-0.0701796636,0.0866600648,-0.2455856353,0.1962551326,-0.3665637672,-0.5910026431,-0.3121281564,-0.2207443565,0.1649776548,-0.2732044756,0.28374663,-0.1319197714,-0.2716562748,0.2348254919,0.1886444241,0.3562226892,0.238636449,0.2746743858,-0.310839653,-0.0177151002,-0.1842351109,0.310754627,0.1267085671,0.2490748763,0.1786777973,-0.2989786267,-0.1730295867,-0.2467732877,-0.4817385077,0.089297086,-0.032913927,0.3612640798,0.0369145088,-0.0061595929,-0.1015347168,-0.0593845882,-0.0475925393,0.1554066539,-0.321716845,-0.099974297,-0.1066943109,-0.0564780273,-0.2586625516,-0.5629889369,-0.2710545957,-0.1909809411,-0.0850012749,-0.0118034007,0.2423169762,-0.1813126653,-0.0354815312,0.1197240576,0.090256907,-0.0801457763,-0.3614607751,0.2991599441,0.361500293,-0.153798148,-0.4440942407,-0.1383514553,-0.1057202816,0.2217350155,-0.3010010421,-0.3994612396,-0.0869593993,-0.140109688,-0.2230976373,-0.0174844526,-0.073769547,0.2164241225,-0.1130887046,-0.2328870147,-0.0172807127,-0.1985118836,0.1154638603,0.130371049,0.169789657,-0.1722306162,0.5826883912,0.0520810261,0.1483297944,0.1584417224,-0.1299767792,0.171134308,-0.1682365537,0.1268296391,-0.0480326973,-0.045324862,0.3635538816,0.1533102244,-0.0188578218,0.3690085411,-0.057263691,-0.2640039027,0.1094783768,0.0063495804,-0.0653221384,-0.1079196334,0.0292380881,-0.2916397452,-0.0198185183,0.083358869,0.0959740281,0.1944290251,-0.1557240486,0.113623023,0.0154413022,0.0269193277,-0.1321808696,-0.531431973,-0.2443500906,-0.2926125526,0.3258085847,0.0728762224,0.3792927265,-0.2615154386,-0.1328953952,0.1868347973,-0.1185835674,0.5819468498,0.1304207742,0.1935622245,0.2138529122,0.1884213239,0.0790649056,0.1521455199,-0.466581583,0.3041167557,0.1202671006,0.2619284689,-0.1860757321,0.0104635172,0.1320597976,0.2706112862,-0.4115587473,0.0313057527,-0.3207559586,-0.2958230376,-0.0329888202,0.1487252116,0.1233900636,0.3150461614,-0.0871901885,-0.0636227503,-0.1172879636,-0.4636466801,0.04573467,0.2088248581,0.3036416471,-0.0808656588,-0.0717990547,-0.1067918241,0.3183946908,-0.1104695424,0.330857873,0.1301068366,-0.0969371796,0.0098023294,-0.4112183452,0.2951388359,0.0367309451,-0.1786257476,-0.0091440324,0.4109277129,0.0753104836,-0.1215833947,-0.1314014047,0.0782288015,0.154828459,-0.1019289941,-0.066371046,0.210072428,-0.2370853871,-0.1019979715,-0.0266773477,0.3568893075,-0.3343978524,0.4691640139,0.230417937,0.9524292946,0.2300878614,0.004815958,0.2026661187,-0.2724691927,0.0495895818,0.2469626069,0.3125798702,-0.3602144122,0.0808099508,-0.1237326413,-0.0259009004,0.1456841528,0.2686609328,-0.2391446084,0.0368269384,-0.0991390571,0.0187147185,0.0578457825,-0.0420080163,0.0123539818,0.0667764693,-0.3098329008,0.2407772094,0.1604472995,-0.1262230128,-0.1036741138,-0.1553587019,0.0229907036,-0.1915262491,-0.0891853645,0.0344958864,-0.4054076672,0.0044303206,0.1254346073,-0.3809590042,-0.110140115,0.1247104555,0.2879429758,0.1120830551,0.0355701856,0.1754250377,0.2798658609,0.333129853,0.2754772604,0.0300889499,0.2048370242,-0.3041776121,-0.1624447554,0.2047544718,0.11769142,-0.1924744099,-0.2655920088,0.2533445358,0.0037148867,0.0201470908,-0.1261609644,-0.0415273048,-0.0486049242,-0.2539508045,0.2601626217,-0.1233119741,-0.1218632832,0.3450837135,-0.015102515,-0.2232404798,-0.1870571226,0.2631909251,0.0915394872,0.2749405801,0.4604128301,0.0735991597,-0.2701103687,-0.467040658,-0.1441354305,-0.0737694427,-0.2279803902,0.1191655472,-0.1223850623,-0.2254580855,-0.0895111784,0.0805105567,-0.0582050197,0.1274322569,-0.1490082294,-0.61260432,-0.2825827301,-0.0735413954,0.0543799847,0.1607298106,0.1171972975,0.0226549432,-0.1883123368,0.0475361571,-0.4655611515,0.1205714047,-0.1525133252,0.068241179,-0.3576040566,0.0517567545,-0.056235522,0.1945682317,0.4125513434,-0.0648631826,-0.2503980994,-0.4053132534,-0.0830439925,0.0675584599,-0.1761305034,-0.1299041361,0.0353784822,-0.0416139252,-0.2331759036,0.1856706142,0.0996722728,-0.1140601933,0.0705626085,0.0362370946,0.1259871125,-0.0754710361,-0.0776163861,-0.0489319339,0.1121360734,-0.1620777547,0.1734581739,0.2188940048,-0.0234114621,-0.0476469025,0.3097182214,-0.1309831887,-0.0507532395,0.0357708558,0.3090589345,-0.1389124244,-0.1823407859,0.2213166058,0.0059361998,0.0923449323,-0.3267617226,-0.1344147474,0.0450517796,0.4097962677,-0.5532665849,-0.0793981999,0.0638042241,-0.0084645161,0.1509258747,0.1421750039,0.2615834475,-0.1343318373,0.126941666,0.2209291607,0.3258605301,-0.2247743607,-0.1728530079,0.050757736,-0.1218095198,0.1692665368,0.112531893,0.0999643132,0.1192644089,0.6916362047,-0.0437443219,0.2923962772,0.1101408973,0.0679828003,0.0815226734,0.1393963844,0.1451892406,0.4456792772,-0.2920979559,0.1568640769,0.5386681557,-0.1165414453,0.2051559836,0.0246807523,-0.288428098,-0.148318544,-0.291851759,-0.2522684932,-0.0286990777,-0.1616626233,-0.0294934344,-0.1111187562,0.0027753818,0.0421171524,0.5325813293,-0.007183094,-0.0443251655,-0.1829901636,-0.0525388308,-0.0752628222,-0.0456416085,-0.1999215931,0.3746516109,0.3940322697,-0.0059070871,0.2505495846,-0.0202580262,0.1869512349,0.0737290159,-0.0990660191,0.0158184003,-0.1997042149,-0.2816879153,-0.184263289,0.2646845281,0.1495168954,0.3046125472,0.2148529142,0.3038095236,-0.2171559036,0.1580083668,0.5882334113,0.151739046,-0.0693793744,0.0392128937,-0.4237434268,-0.0257013272,-0.3209612668,0.0478476658,-0.2464501858,0.077816397,0.0034496307,-0.2867472768,0.1594921201,0.1201608777,0.1503569931,0.0604840592,0.6015758514,0.1133554652,-0.0661049932,-0.1337702423,-0.2346678823,-0.5674973726,0.1958338916,0.2363114953,-0.0519492887,-0.1713323146,0.3429603279,0.1597278863,-0.0710562691,0.0902101845,0.0962439999,-0.0821874961,-0.0162224844,-0.4123512506,-0.106794402,0.1234901175,0.2328350395,-0.1281647086,-0.3708758056,0.3792416453,0.0888491347,0.1948824972,-0.2314056754,-0.2158889771,-0.1824168116,0.2466814965,0.0318037346,0.3799424767,0.1526882946,-0.3498205245,0.0361984484,0.1120776013,0.0147677688,-0.1114462987,0.2188643962,0.2435276508,0.2173151821,0.1415945441,0.0435096994,0.1522723287,0.4297482073,0.0463595279,-0.1174224317,-0.2635605335,0.5547003746,-0.1008020863,-0.0520818941,-0.1178275943,0.1396929026,0.0085183904,-0.0532908775,-0.0157468654,-0.4543573856,0.3082257807,-0.5355829597,-0.2539766133,-0.0591680072,0.1182153746,0.1231312752,-0.0180409644,-0.2814457119,0.0923632756,0.3500164449,-0.2909099758,0.1102767512,0.4204139411,0.1702205241,-0.0142796729,-0.2048442066,0.1536104232,0.4446440935,-0.1842172146,-0.048598107,-0.4758795798]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2133","title":"bug in mlqa dataset ","comments":"thank you so much for the insightful comments. ","body":"Hi \r\nLooking into MLQA dataset for langauge \"ar\":\r\n\r\n```\r\n \"question\": [\r\n \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631?\",\r\n \"\\u0643\\u0645 \\u0645\\u0631\\u0629 \\u064a\\u062a\\u0645 \\u0646\\u0634\\u0631\\u0647\\u0627 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0645\\u0627 \\u0647\\u064a \\u0627\\u0644\\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u064a\\u0648\\u0645\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0643\\u0645 \\u0639\\u062f\\u062f \\u0627\\u0644\\u0627\\u0648\\u0631\\u0627\\u0642 \\u0627\\u0644\\u0627\\u062e\\u0628\\u0627\\u0631\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0627\\u0644\\u062a\\u064a \\u0648\\u062c\\u062f\\u062a \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0641\\u064a \\u0627\\u064a \\u0633\\u0646\\u0629 \\u0628\\u062f\\u0627\\u062a \\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u0637\\u0627\\u0644\\u0628 \\u0627\\u0644\\u062d\\u0633 \\u0627\\u0644\\u0633\\u0644\\u064a\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\"\r\n ]\r\n```\r\n\r\nthe questions are in the wrong format, and not readable, could you please have a look? thanks @lhoestq \r\n","comment_length":8,"text":"bug in mlqa dataset \nHi \r\nLooking into MLQA dataset for langauge \"ar\":\r\n\r\n```\r\n \"question\": [\r\n \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631?\",\r\n \"\\u0643\\u0645 \\u0645\\u0631\\u0629 \\u064a\\u062a\\u0645 \\u0646\\u0634\\u0631\\u0647\\u0627 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0645\\u0627 \\u0647\\u064a \\u0627\\u0644\\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u064a\\u0648\\u0645\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0643\\u0645 \\u0639\\u062f\\u062f \\u0627\\u0644\\u0627\\u0648\\u0631\\u0627\\u0642 \\u0627\\u0644\\u0627\\u062e\\u0628\\u0627\\u0631\\u064a\\u0629 \\u0644\\u0644\\u0637\\u0644\\u0627\\u0628 \\u0627\\u0644\\u062a\\u064a \\u0648\\u062c\\u062f\\u062a \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\",\r\n \"\\u0641\\u064a \\u0627\\u064a \\u0633\\u0646\\u0629 \\u0628\\u062f\\u0627\\u062a \\u0648\\u0631\\u0642\\u0629 \\u0627\\u0644\\u0637\\u0627\\u0644\\u0628 \\u0627\\u0644\\u062d\\u0633 \\u0627\\u0644\\u0633\\u0644\\u064a\\u0645 \\u0628\\u0627\\u0644\\u0646\\u0634\\u0631 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u0645?\"\r\n ]\r\n```\r\n\r\nthe questions are in the wrong format, and not readable, could you please have a look? thanks @lhoestq \r\n\nthank you so much for the insightful comments. ","embeddings":[-0.115853101,-0.231567964,-0.2734039724,0.2112123966,0.3321486115,0.0711947754,0.3524691463,0.1213381737,-0.3227720559,0.2254540175,0.0618873164,0.2817641795,0.2518624663,0.2561215162,0.1166762114,-0.047785569,0.1292088628,0.179456532,0.0600869805,-0.197704494,-0.196472913,0.2453739792,-0.1337222606,0.0539799668,-0.0679920763,-0.033509817,-0.0207405165,0.220893696,-0.0425398685,0.0830165371,-0.1433827132,-0.2492368817,-0.1639159769,0.005175015,-0.0001021291,-0.1083842143,0.1797213703,-0.0530055612,-0.2548291981,0.3303485215,-0.4433495402,0.0542143062,-0.3197640181,-0.3245649338,-0.18316935,-0.1904883683,-0.05527585,-0.5454843044,0.3158854544,0.3931446373,0.3555622101,0.1169286147,-0.0093527725,-0.0591307953,0.2337930202,-0.0075344178,0.0532882288,-0.0125398654,0.0099399295,-0.0157065429,0.0644993559,0.5903977156,0.1528967172,-0.0744416192,-0.2841548622,0.0947690308,0.2282186002,-0.1410573274,0.3744464219,0.2255229354,0.2068464309,-0.2394515574,-0.1113861203,0.070110254,0.03870086,-0.1902514845,-0.0511132963,0.2683143914,0.0487997532,0.0880704224,0.1475322396,0.1784594804,-0.1978724748,0.0273552109,-0.4308957458,0.4159484506,-0.1553923637,-0.0334863551,-0.1144851521,-0.2438467145,-0.1090199426,0.1160906479,-0.1431100965,0.1030091941,-0.1794525683,-0.188223511,-0.1247729361,0.2676885426,0.013106185,-0.0319190025,0.1373087317,0.1452646405,0.2267717719,-0.0552490279,-0.1274060756,-0.010375008,-0.1002315283,-0.115046382,0.2245224118,-0.1293273121,-0.0769749582,0.015334459,0.0609565601,-0.2191800922,-0.1487826705,0.2285619527,0.1345374286,-0.1410000473,-0.3242001534,0.056123469,-0.3493770361,0.0548069552,-0.1677003652,0.391261518,0.2288828194,-0.0240960065,-0.139952004,-0.0753454566,-0.1917124987,-0.3327361643,-0.3898272812,0.1255715787,0.1248719692,-0.27682212,-0.0159336664,0.3214123547,0.215539068,-0.0244428106,0.1260242909,0.0829229057,0.0949713215,-0.1912811249,0.2312819362,0.2930064201,-0.1015833616,-0.0493383966,-0.0328686349,-0.0564297847,0.0095267622,0.1452571899,0.0762135983,0.0079340953,-0.3933087587,0.3797929287,0.395275563,-0.0949567109,0.1754642427,0.2335368693,0.2187478989,0.0496398285,0.1180229709,-0.1062783301,-0.0219002888,-0.1518765986,0.063460812,0.0898392871,-0.526848197,0.0284028426,-0.3689579368,-0.0933342502,0.3024438322,0.3060186207,0.0991310477,-0.0708538219,-0.170061931,0.3025590479,0.1694825292,-0.0779909417,-0.3400383294,0.2042782009,-0.159998998,-0.1875737309,0.2839224935,0.2088894844,-0.1037751511,-0.0227849204,0.1018831581,0.1511971503,-0.1055432111,0.0276853014,-0.123123616,-0.1423805505,0.1725804061,0.0685132965,-0.2583670616,-0.1137779951,0.2595126629,0.024177283,0.3915384114,-0.0147941019,-0.0749591962,0.0135183893,0.3583312035,-0.4317682683,0.162547648,-0.0722043589,-0.3222065568,-0.219033882,-0.609854579,0.0809970871,-0.233429864,-0.1778639853,-0.1760995686,-0.1031846404,0.0356507711,0.0966448039,0.3250931799,0.1980949938,-0.0805992335,0.2429176867,0.0244634617,-0.0335509703,-0.0382368453,-0.1474969685,-0.3175995946,0.2270222902,-0.1953133941,0.0706109107,0.1879199296,0.3434305489,0.1390678883,-0.0944974348,0.1414386928,0.0136982501,0.095292449,0.2897173464,-0.1747563481,0.1883360296,0.2945405543,-0.2585287094,0.0214189049,0.3033981323,0.3103998601,0.0042769727,-0.1195691824,0.3716930151,0.0615129918,-0.0408308692,0.0971447751,0.029447915,0.4241673052,-0.0936098993,0.1024363488,-0.5371354222,0.168956995,0.1346676648,0.2095032185,0.0624080636,-0.5084814429,0.439779222,0.4208018482,-0.0966070369,0.0676753074,-0.1444756836,-0.1639984995,0.0293982271,0.0471223593,0.1910907626,0.2587710917,0.2780264914,0.0747869611,0.0024714572,-0.0050052307,-0.4039029181,0.34732306,-0.0582332052,0.0497225896,0.2490378618,0.1140516698,-0.0766907632,-0.3842312396,0.1208591983,-0.0701796636,0.0866600648,-0.2455856353,0.1962551326,-0.3665637672,-0.5910026431,-0.3121281564,-0.2207443565,0.1649776548,-0.2732044756,0.28374663,-0.1319197714,-0.2716562748,0.2348254919,0.1886444241,0.3562226892,0.238636449,0.2746743858,-0.310839653,-0.0177151002,-0.1842351109,0.310754627,0.1267085671,0.2490748763,0.1786777973,-0.2989786267,-0.1730295867,-0.2467732877,-0.4817385077,0.089297086,-0.032913927,0.3612640798,0.0369145088,-0.0061595929,-0.1015347168,-0.0593845882,-0.0475925393,0.1554066539,-0.321716845,-0.099974297,-0.1066943109,-0.0564780273,-0.2586625516,-0.5629889369,-0.2710545957,-0.1909809411,-0.0850012749,-0.0118034007,0.2423169762,-0.1813126653,-0.0354815312,0.1197240576,0.090256907,-0.0801457763,-0.3614607751,0.2991599441,0.361500293,-0.153798148,-0.4440942407,-0.1383514553,-0.1057202816,0.2217350155,-0.3010010421,-0.3994612396,-0.0869593993,-0.140109688,-0.2230976373,-0.0174844526,-0.073769547,0.2164241225,-0.1130887046,-0.2328870147,-0.0172807127,-0.1985118836,0.1154638603,0.130371049,0.169789657,-0.1722306162,0.5826883912,0.0520810261,0.1483297944,0.1584417224,-0.1299767792,0.171134308,-0.1682365537,0.1268296391,-0.0480326973,-0.045324862,0.3635538816,0.1533102244,-0.0188578218,0.3690085411,-0.057263691,-0.2640039027,0.1094783768,0.0063495804,-0.0653221384,-0.1079196334,0.0292380881,-0.2916397452,-0.0198185183,0.083358869,0.0959740281,0.1944290251,-0.1557240486,0.113623023,0.0154413022,0.0269193277,-0.1321808696,-0.531431973,-0.2443500906,-0.2926125526,0.3258085847,0.0728762224,0.3792927265,-0.2615154386,-0.1328953952,0.1868347973,-0.1185835674,0.5819468498,0.1304207742,0.1935622245,0.2138529122,0.1884213239,0.0790649056,0.1521455199,-0.466581583,0.3041167557,0.1202671006,0.2619284689,-0.1860757321,0.0104635172,0.1320597976,0.2706112862,-0.4115587473,0.0313057527,-0.3207559586,-0.2958230376,-0.0329888202,0.1487252116,0.1233900636,0.3150461614,-0.0871901885,-0.0636227503,-0.1172879636,-0.4636466801,0.04573467,0.2088248581,0.3036416471,-0.0808656588,-0.0717990547,-0.1067918241,0.3183946908,-0.1104695424,0.330857873,0.1301068366,-0.0969371796,0.0098023294,-0.4112183452,0.2951388359,0.0367309451,-0.1786257476,-0.0091440324,0.4109277129,0.0753104836,-0.1215833947,-0.1314014047,0.0782288015,0.154828459,-0.1019289941,-0.066371046,0.210072428,-0.2370853871,-0.1019979715,-0.0266773477,0.3568893075,-0.3343978524,0.4691640139,0.230417937,0.9524292946,0.2300878614,0.004815958,0.2026661187,-0.2724691927,0.0495895818,0.2469626069,0.3125798702,-0.3602144122,0.0808099508,-0.1237326413,-0.0259009004,0.1456841528,0.2686609328,-0.2391446084,0.0368269384,-0.0991390571,0.0187147185,0.0578457825,-0.0420080163,0.0123539818,0.0667764693,-0.3098329008,0.2407772094,0.1604472995,-0.1262230128,-0.1036741138,-0.1553587019,0.0229907036,-0.1915262491,-0.0891853645,0.0344958864,-0.4054076672,0.0044303206,0.1254346073,-0.3809590042,-0.110140115,0.1247104555,0.2879429758,0.1120830551,0.0355701856,0.1754250377,0.2798658609,0.333129853,0.2754772604,0.0300889499,0.2048370242,-0.3041776121,-0.1624447554,0.2047544718,0.11769142,-0.1924744099,-0.2655920088,0.2533445358,0.0037148867,0.0201470908,-0.1261609644,-0.0415273048,-0.0486049242,-0.2539508045,0.2601626217,-0.1233119741,-0.1218632832,0.3450837135,-0.015102515,-0.2232404798,-0.1870571226,0.2631909251,0.0915394872,0.2749405801,0.4604128301,0.0735991597,-0.2701103687,-0.467040658,-0.1441354305,-0.0737694427,-0.2279803902,0.1191655472,-0.1223850623,-0.2254580855,-0.0895111784,0.0805105567,-0.0582050197,0.1274322569,-0.1490082294,-0.61260432,-0.2825827301,-0.0735413954,0.0543799847,0.1607298106,0.1171972975,0.0226549432,-0.1883123368,0.0475361571,-0.4655611515,0.1205714047,-0.1525133252,0.068241179,-0.3576040566,0.0517567545,-0.056235522,0.1945682317,0.4125513434,-0.0648631826,-0.2503980994,-0.4053132534,-0.0830439925,0.0675584599,-0.1761305034,-0.1299041361,0.0353784822,-0.0416139252,-0.2331759036,0.1856706142,0.0996722728,-0.1140601933,0.0705626085,0.0362370946,0.1259871125,-0.0754710361,-0.0776163861,-0.0489319339,0.1121360734,-0.1620777547,0.1734581739,0.2188940048,-0.0234114621,-0.0476469025,0.3097182214,-0.1309831887,-0.0507532395,0.0357708558,0.3090589345,-0.1389124244,-0.1823407859,0.2213166058,0.0059361998,0.0923449323,-0.3267617226,-0.1344147474,0.0450517796,0.4097962677,-0.5532665849,-0.0793981999,0.0638042241,-0.0084645161,0.1509258747,0.1421750039,0.2615834475,-0.1343318373,0.126941666,0.2209291607,0.3258605301,-0.2247743607,-0.1728530079,0.050757736,-0.1218095198,0.1692665368,0.112531893,0.0999643132,0.1192644089,0.6916362047,-0.0437443219,0.2923962772,0.1101408973,0.0679828003,0.0815226734,0.1393963844,0.1451892406,0.4456792772,-0.2920979559,0.1568640769,0.5386681557,-0.1165414453,0.2051559836,0.0246807523,-0.288428098,-0.148318544,-0.291851759,-0.2522684932,-0.0286990777,-0.1616626233,-0.0294934344,-0.1111187562,0.0027753818,0.0421171524,0.5325813293,-0.007183094,-0.0443251655,-0.1829901636,-0.0525388308,-0.0752628222,-0.0456416085,-0.1999215931,0.3746516109,0.3940322697,-0.0059070871,0.2505495846,-0.0202580262,0.1869512349,0.0737290159,-0.0990660191,0.0158184003,-0.1997042149,-0.2816879153,-0.184263289,0.2646845281,0.1495168954,0.3046125472,0.2148529142,0.3038095236,-0.2171559036,0.1580083668,0.5882334113,0.151739046,-0.0693793744,0.0392128937,-0.4237434268,-0.0257013272,-0.3209612668,0.0478476658,-0.2464501858,0.077816397,0.0034496307,-0.2867472768,0.1594921201,0.1201608777,0.1503569931,0.0604840592,0.6015758514,0.1133554652,-0.0661049932,-0.1337702423,-0.2346678823,-0.5674973726,0.1958338916,0.2363114953,-0.0519492887,-0.1713323146,0.3429603279,0.1597278863,-0.0710562691,0.0902101845,0.0962439999,-0.0821874961,-0.0162224844,-0.4123512506,-0.106794402,0.1234901175,0.2328350395,-0.1281647086,-0.3708758056,0.3792416453,0.0888491347,0.1948824972,-0.2314056754,-0.2158889771,-0.1824168116,0.2466814965,0.0318037346,0.3799424767,0.1526882946,-0.3498205245,0.0361984484,0.1120776013,0.0147677688,-0.1114462987,0.2188643962,0.2435276508,0.2173151821,0.1415945441,0.0435096994,0.1522723287,0.4297482073,0.0463595279,-0.1174224317,-0.2635605335,0.5547003746,-0.1008020863,-0.0520818941,-0.1178275943,0.1396929026,0.0085183904,-0.0532908775,-0.0157468654,-0.4543573856,0.3082257807,-0.5355829597,-0.2539766133,-0.0591680072,0.1182153746,0.1231312752,-0.0180409644,-0.2814457119,0.0923632756,0.3500164449,-0.2909099758,0.1102767512,0.4204139411,0.1702205241,-0.0142796729,-0.2048442066,0.1536104232,0.4446440935,-0.1842172146,-0.048598107,-0.4758795798]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2132","title":"TydiQA dataset is mixed and is not split per language ","comments":"You can filter the languages this way:\r\n```python\r\ntydiqa_en = tydiqa_dataset.filter(lambda x: x[\"language\"] == \"english\")\r\n```\r\n\r\nOtherwise maybe we can have one configuration per language ?\r\nWhat do you think of this for example ?\r\n\r\n```python\r\nload_dataset(\"tydiqa\", \"primary_task.en\")\r\n```","body":"Hi @lhoestq \r\nCurrently TydiQA is mixed and user can only access the whole training set of all languages:\r\nhttps:\/\/www.tensorflow.org\/datasets\/catalog\/tydi_qa\r\n\r\nfor using this dataset, one need to train\/evaluate in each separate language, and having them mixed, makes it hard to use this dataset. This is much convenient for user to have them split and I appreciate your help on this. \r\n\r\nMeanwhile, till hopefully this is split per language, I greatly appreciate telling me how I can preprocess and get data per language. thanks a lot ","comment_length":39,"text":"TydiQA dataset is mixed and is not split per language \nHi @lhoestq \r\nCurrently TydiQA is mixed and user can only access the whole training set of all languages:\r\nhttps:\/\/www.tensorflow.org\/datasets\/catalog\/tydi_qa\r\n\r\nfor using this dataset, one need to train\/evaluate in each separate language, and having them mixed, makes it hard to use this dataset. This is much convenient for user to have them split and I appreciate your help on this. \r\n\r\nMeanwhile, till hopefully this is split per language, I greatly appreciate telling me how I can preprocess and get data per language. thanks a lot \nYou can filter the languages this way:\r\n```python\r\ntydiqa_en = tydiqa_dataset.filter(lambda x: x[\"language\"] == \"english\")\r\n```\r\n\r\nOtherwise maybe we can have one configuration per language ?\r\nWhat do you think of this for example ?\r\n\r\n```python\r\nload_dataset(\"tydiqa\", \"primary_task.en\")\r\n```","embeddings":[-0.2592062056,-0.23528862,-0.204411611,0.2665453851,0.2771573961,-0.025829006,0.3412586749,0.3332071304,-0.1514983028,0.077681914,-0.3343148828,-0.0462720059,-0.0167529695,0.4214190245,0.0419181399,-0.320132941,-0.1105380207,0.0581458621,-0.1372380257,-0.009822676,0.023834575,0.1549571157,-0.1817224771,0.1316394061,0.0915518329,-0.1736210138,-0.0809654817,-0.0455806702,0.0590337105,-0.0585781932,0.4464208782,0.1073087528,0.267031461,0.3341659904,-0.0001084638,0.1922285259,-0.0415637456,-0.3396548927,0.1110021845,-0.2360355109,-0.1122633591,-0.3174516559,-0.0924162939,-0.204316467,-0.5365697145,-0.1827822179,0.0295533817,-0.3825767338,0.3870956004,0.2981826663,0.1465835422,0.1663268059,-0.1849585027,0.2282588482,-0.0572569855,0.0546517596,-0.1057092249,-0.0530425459,0.5951901674,-0.0774741173,0.3307021856,0.2777237296,0.1899045557,0.0749691203,-0.4010953307,0.1116975546,0.0430990271,-0.5291082263,0.0033926866,0.5172391534,0.4501475692,0.0541579202,-0.3280532956,-0.1335923374,-0.0303184725,-0.2640711069,0.1254836172,0.3814415932,-0.1881920695,0.2486717999,0.3558285534,0.0850683898,-0.262557447,0.1491750926,-0.2938213944,0.3037575781,-0.0313543119,0.0603938252,-0.0548035987,-0.1423953772,0.2361643016,0.2435273081,0.0689521581,0.256490767,-0.4733611643,-0.2544720471,-0.1434990466,-0.3753200769,-0.1192742288,-0.3847779334,0.137922585,0.3340659738,-0.1463015974,-0.0021057045,0.2769569755,-0.0824466422,0.2340909988,0.3656898141,0.2035248131,-0.1067949831,0.0105769364,0.0757092983,-0.1920621395,-0.3306031525,-0.4287641644,0.2712708414,0.174067229,0.0394181311,-0.3159442842,-0.1879136562,-0.1263174713,-0.2216869295,0.1088807583,0.3437946141,-0.0275603309,-0.009170562,0.0795595348,0.4542489052,-0.3114559054,-0.3753986061,-0.1421310157,-0.053627532,-0.1073865518,-0.1681628376,0.0842008591,0.2148738205,-0.0248488355,0.0667255297,-0.0025112557,-0.1574177444,0.3517777622,-0.2013829499,0.117537111,0.0731887966,0.1900308579,0.0227768905,0.2191588134,-0.2975347042,-0.3487254381,0.0580839403,-0.278357178,-0.2406381369,0.0743742287,0.1674489826,0.3175940514,0.1445011795,0.0269591846,0.7229220271,0.3423041105,0.0413441323,-0.0289358981,0.0486030392,-0.2848916948,-0.0555126928,0.2881773114,0.162882939,-0.5802620053,-0.1281797439,0.2130147964,-0.322067529,0.215567112,0.2149673998,-0.1866123378,0.3718908727,0.0974325612,-0.2031816691,0.7471084595,-0.2007604241,-0.3531040549,0.039042335,0.1203382835,0.1161096543,0.0331046879,0.0693071336,0.1348564476,-0.0257895254,0.205305934,0.5336025357,-0.2365159988,0.0297370832,0.0447351895,-0.0627163351,0.3733536303,0.4188765585,-0.0572330654,-0.2378853709,-0.0126547432,0.2349274904,0.382068783,-0.1548735797,0.0942421779,0.0616843626,-0.303639859,0.1535947621,0.0307177715,-0.2157306373,-0.2534970939,0.2257353961,-0.3984822035,0.2190700918,0.0171267856,-0.1806215197,0.1356500089,-0.1984942108,-0.3050163984,-0.2456896752,0.1831280589,-0.1110467687,0.0207428113,0.1629618406,-0.1032495424,0.0985127985,-0.1514897197,-0.0800129473,-0.1922066063,0.0206570756,0.0212977212,-0.0716993809,0.0058197477,0.2450117022,-0.1134706959,-0.2013849914,-0.0574636124,-0.0935559794,0.1642187089,0.1159625202,-0.1913137138,-0.2085589916,0.3557848334,-0.0733127221,-0.0767932683,0.0469789654,0.0040056356,-0.1576796919,0.0584184006,0.5854184031,-0.0236305781,0.2771928012,0.0699686334,-0.0590587445,0.3816142678,-0.1984386891,-0.0212117136,-0.2191328704,0.2769998014,-0.1963724792,0.0008180983,-0.0419192724,-0.441222012,0.1670796573,0.4789676368,0.1411519647,0.2277944386,0.0006079229,0.305113405,0.1892092675,-0.0941263661,0.2189483792,0.1907681823,0.0479791425,0.2750668824,-0.1851740032,0.3741773963,-0.1362421513,0.0320604295,-0.158335954,0.0158972535,0.0188290253,-0.0313223079,-0.1062834114,-0.013290707,0.1975892037,0.1503423303,0.128862083,-0.022720214,0.2100903392,-0.4768428802,-0.3273683488,-0.2023428231,-0.307257086,0.0733935237,0.2087882161,0.2061761469,-0.4034985602,-0.1359766126,0.2662344277,0.2984233797,-0.0250988193,-0.1683412343,-0.1758296937,0.0169531498,-0.2763130963,-0.216508761,0.1792732626,0.2920338809,-0.0906245112,0.2981304228,-0.3847173154,-0.2344311476,0.0706419796,-0.2707393765,0.093471922,-0.2213926166,0.0547199436,-0.1595893353,0.1746494174,0.0280353669,-0.1556295156,0.3326012492,0.2598111033,-0.1141361222,0.1519023627,-0.0054758387,0.1311878562,0.0371142253,-0.6562100053,-0.7368292212,-0.1958166808,-0.2545368671,-0.1273786575,0.3844645321,-0.4101475775,0.0553293005,-0.1220950484,0.0513957217,0.1855765134,-0.0182751603,-0.157880187,0.1308661252,-0.3448569477,-0.3513146937,-0.0799213648,-0.0584392808,0.6852425337,0.0431133397,-0.1299108565,0.3011080325,-0.1743428707,0.055523973,-0.131667763,-0.0169700161,0.2986245751,-0.0941570997,0.1243958324,-0.0248790327,0.2955364585,0.1937161684,-0.2547276914,0.2401013076,0.0134681333,0.3823543489,0.1917834431,0.711796701,0.2415472418,-0.0094934255,0.2683726251,0.0604744852,-0.2061232775,-0.096832782,-0.3948800862,0.0126799615,0.2939455509,0.1647691876,0.263042748,0.0525145121,-0.4610899091,-0.381929636,-0.2104378939,-0.3223951459,-0.4282873571,0.3108154833,-0.2334423661,0.3259696066,-0.110340178,-0.1093477681,0.0415193103,0.0197684206,-0.2085597813,0.2697065473,-0.2869630754,0.1213246435,-0.1540878713,-0.3814828992,-0.1956708282,0.2603662312,0.0273123663,0.2598648965,-0.3271664977,-0.0667763352,0.257771194,0.0129684685,0.4652477801,-0.0030340692,-0.0346149467,-0.1049609184,-0.2077572197,-0.0547856539,0.0457566641,-0.1851525307,-0.0577439554,0.2859673798,0.0354526564,-0.3558894098,-0.3494700193,-0.1935219318,0.2499514222,-0.4031064808,0.1149708331,-0.2828820646,-0.418150425,-0.1601431519,0.3174799085,0.0429387353,0.0426308401,0.094890289,-0.3793869615,-0.2059390098,0.1876849532,0.0874072909,0.3865199983,0.0501492396,-0.1424170732,0.2228981256,-0.0953625143,-0.2537565231,0.3669161201,0.5436347723,-0.1135801524,-0.5056569576,0.0044328957,-0.2693913281,0.4042411745,0.1753688455,-0.0647639632,0.0708355159,-0.1406499147,-0.031750679,-0.2224081606,0.2009249777,0.454495728,0.2071083486,-0.6362085938,-0.5407907963,0.1689659059,-0.0918039307,-0.0340849273,0.4179161191,0.1066995338,-0.3516064882,0.4943614602,0.1378395855,0.8109483719,-0.290471971,0.1912714392,0.094633773,0.1352834404,0.1510024071,-0.2028165162,-0.0893720239,-0.0699798986,-0.0036595475,-0.1287712753,0.0796497688,0.204509154,0.3761956692,-0.3296663761,0.041761972,-0.1871775389,-0.5379321575,0.016105691,0.2127106339,-0.2093775868,-0.1705773771,-0.2777920961,0.0850294679,-0.1662088931,0.1448354274,-0.0725889206,-0.0730766356,-0.0807035118,0.0600315221,-0.128290087,-0.0290325973,-0.3498610556,-0.1406414211,-0.1632148027,-0.5811936259,0.4124368429,0.5101847649,0.3279884756,0.2206023633,-0.0412411802,0.2668744922,0.150714457,0.0529303662,0.0414051563,-0.2150479257,0.191854924,-0.0683289543,-0.3392774761,0.0616895184,0.1848707944,0.2014485002,-0.1583432555,0.2512345016,0.2051075995,-0.1945831478,-0.1419589818,0.1469659656,0.1715349555,-0.0241322611,0.1722888947,0.2702116668,-0.2065200061,0.1338063329,-0.2121042609,-0.2760449648,-0.1011270881,-0.3500294685,0.4322403073,0.2798494399,0.4518121183,0.2038094848,-0.0289550368,-0.1422361135,0.3866252899,0.0002726129,-0.1847754717,-0.1028970778,0.3043477833,0.050091669,-0.0555190109,0.2425949425,-0.2680446208,-0.1343913823,-0.0525049269,-0.3713813126,-0.0147940917,0.0284190942,-0.0023855926,0.0218011625,0.2537812889,0.229824543,-0.0625415966,0.6077544689,-0.2886663973,-0.0139997723,-0.0860647187,0.3429329097,-0.2048236281,0.1122022495,0.1519576013,0.1797051728,0.0378549695,-0.1501203775,0.0580248833,-0.3018404543,-0.1762670279,0.1387174428,0.0194343794,0.1557268202,0.0683930814,-0.2810458243,0.0348379761,-0.1254879683,0.1631982177,0.0618572347,0.1895332187,0.4650207162,0.2440118641,0.1847275048,-0.2461517304,0.1380319148,-0.1916904151,-0.034915287,-0.0188467894,0.0493478179,-0.1243448779,-0.2026319206,-0.1169788316,-0.0052631306,0.1201917678,-0.1395374388,-0.1064232811,0.1043248847,0.4723398685,-0.2647071779,-0.0682183653,0.2373574525,-0.0518935323,0.2495449185,0.4482098818,0.1784880161,-0.1083473563,0.1239553764,-0.2226423174,-0.1907120049,0.1714605093,0.4320271611,-0.1884322017,0.2997784317,0.4383264184,-0.2500911653,0.0253664888,-0.5639311671,0.172579512,0.3753857017,-0.156050548,0.0294167437,0.2055679411,-0.1107945666,-0.0382424928,-0.002099291,-0.2823620439,0.1047887132,-0.0546162426,-0.1241024211,0.0581861362,-0.2883281708,-0.348944217,-0.0700911358,-0.073878102,-0.0301886667,0.0473893546,0.3643285036,-0.0386862792,-0.1570389271,0.042186521,0.1285604388,0.0572405234,-0.1922876686,-0.2062021941,0.0678527728,-0.0501696281,-0.1608560085,0.0136613734,-0.1735758632,-0.1209880188,0.3514298797,0.0343758054,-0.0936689302,0.1517285407,0.002456395,-0.1705127805,-0.0356276073,0.3725081682,0.1863262355,-0.0454809591,0.0694924816,0.1436004043,0.0203575287,0.0109444074,-0.3813976347,0.3755215108,-0.140677467,-0.1401248574,-0.1014751568,0.396998167,0.1822823137,0.3812870979,0.2599556446,0.1357997358,-0.1200288013,-0.1728627384,0.2138911039,0.0051509566,-0.2243544161,0.5801805854,-0.1053805128,0.1787991524,-0.1761499047,-0.0352081209,-0.593629837,0.0816390738,0.1742512435,-0.2596294582,0.1341774762,-0.2041000873,0.0519880205,0.1876397431,0.3666480184,-0.1925254762,0.5568435192,-0.249699682,-0.0354940034,-0.2386247069,0.0994287655,-0.1303417087,0.0004957916,-0.2024938166,0.2077078968,0.3562858999,0.359028548,0.1907352805,-0.2272052318,0.0411611833,0.341391325,-0.2184964269,-0.0689904019,-0.4191856682,0.5319705009,-0.1182231158,-0.2942887545,0.2460460663,-0.0494286194,-0.0780365691,-0.0757015124,-0.0791640952,0.3901766837,0.2521813214,0.1452967972,0.2444940954,0.1770256758,-0.2114203423,0.3767094314,-0.0797894448,0.0217253231,-0.4029797018,0.2660863996,0.2743576765,0.0599229336,-0.4111655951,-0.2919216752,-0.124520801,0.3216745853,0.014834865,-0.1471544206,0.1995241046,0.0352679901,0.4938175976,-0.0977457315,0.3207582533,0.5601398945,0.0472073704,-0.1317714304,-0.3974683583,-0.0904370025,0.101198107,-0.1274487972,-0.246039927,0.1740920991,-0.0403293855,0.0771199688,0.2812560201,-0.3595997691,0.0837732628,0.3247872889,-0.1062043086,0.0207077879,0.2874739766,0.3893792927,-0.103133291,-0.287514925,0.3336113095,0.1676432341,-0.0421697833,-0.205171138,-0.2213898152]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2132","title":"TydiQA dataset is mixed and is not split per language ","comments":"Hi\nthank you very much for the great response, this will be really wonderful\nto have one configuration per language, as one need the dataset in majority\nof case per language for cross-lingual evaluations.\nThis becomes also then more close to TFDS format, which is separated per\nlanguage https:\/\/www.tensorflow.org\/datasets\/catalog\/tydi_qa which will be\nreally awesome to have.\nthanks\n\nOn Mon, Mar 29, 2021 at 6:17 PM Quentin Lhoest ***@***.***>\nwrote:\n\n> You can filter the languages this way:\n>\n> tydiqa_en = tydiqa_dataset.filter(lambda x: x[\"language\"] == \"english\")\n>\n> Otherwise maybe we can have one configuration per language ?\n> What do you think of this for example ?\n>\n> load_dataset(\"tydiqa\", \"primary_task.en\")\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","body":"Hi @lhoestq \r\nCurrently TydiQA is mixed and user can only access the whole training set of all languages:\r\nhttps:\/\/www.tensorflow.org\/datasets\/catalog\/tydi_qa\r\n\r\nfor using this dataset, one need to train\/evaluate in each separate language, and having them mixed, makes it hard to use this dataset. This is much convenient for user to have them split and I appreciate your help on this. \r\n\r\nMeanwhile, till hopefully this is split per language, I greatly appreciate telling me how I can preprocess and get data per language. thanks a lot ","comment_length":145,"text":"TydiQA dataset is mixed and is not split per language \nHi @lhoestq \r\nCurrently TydiQA is mixed and user can only access the whole training set of all languages:\r\nhttps:\/\/www.tensorflow.org\/datasets\/catalog\/tydi_qa\r\n\r\nfor using this dataset, one need to train\/evaluate in each separate language, and having them mixed, makes it hard to use this dataset. This is much convenient for user to have them split and I appreciate your help on this. \r\n\r\nMeanwhile, till hopefully this is split per language, I greatly appreciate telling me how I can preprocess and get data per language. thanks a lot \nHi\nthank you very much for the great response, this will be really wonderful\nto have one configuration per language, as one need the dataset in majority\nof case per language for cross-lingual evaluations.\nThis becomes also then more close to TFDS format, which is separated per\nlanguage https:\/\/www.tensorflow.org\/datasets\/catalog\/tydi_qa which will be\nreally awesome to have.\nthanks\n\nOn Mon, Mar 29, 2021 at 6:17 PM Quentin Lhoest ***@***.***>\nwrote:\n\n> You can filter the languages this way:\n>\n> tydiqa_en = tydiqa_dataset.filter(lambda x: x[\"language\"] == \"english\")\n>\n> Otherwise maybe we can have one configuration per language ?\n> What do you think of this for example ?\n>\n> load_dataset(\"tydiqa\", \"primary_task.en\")\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","embeddings":[-0.3552147746,-0.232304275,-0.1865600795,0.2470623255,0.3051548004,-0.0820614621,0.3825841844,0.3362080455,-0.191450879,0.1315774769,-0.4605718851,-0.0628384873,0.0637698621,0.4531888664,0.0351658799,-0.2893125117,-0.1692789942,0.0985133275,-0.2443104386,0.0038089035,0.1467833221,0.2375836372,-0.1056776047,0.0857928395,0.0535366237,-0.2327951342,-0.0800041556,-0.0570415594,0.0407805629,-0.0867752507,0.4129017889,0.2202050239,0.2382610142,0.3516655266,-0.000102664,0.2214549333,-0.0919180363,-0.2966122925,0.0862295106,-0.1834036857,-0.101301223,-0.2273289114,-0.1310020983,-0.1961812079,-0.5325117707,-0.1905819923,-0.0044926647,-0.3459679782,0.3396964371,0.2731972039,0.1980174333,0.2754566371,-0.1090018749,0.1592267305,-0.0758001879,0.0541526936,-0.1506026238,-0.0746902749,0.4563406408,-0.0522275232,0.1966013759,0.273450762,0.2479817718,0.0726507157,-0.3480714858,0.0870823935,-0.0185377765,-0.4929688573,0.0308897756,0.6588577628,0.4845636785,-0.0275688395,-0.3650477529,-0.1043434516,-0.1176581979,-0.245856598,0.1216570064,0.2802472413,-0.1396408975,0.3281907439,0.3242917359,0.1832720935,-0.2296533287,0.1559876651,-0.3659547269,0.2008359134,-0.0907996818,-0.0041711116,0.0995028615,-0.1770900637,0.0782897547,0.2008456886,0.0825700238,0.194704935,-0.5239602923,-0.3435084224,0.0030070683,-0.4299471676,0.021633314,-0.3731372952,0.1823283583,0.3100450933,-0.2238860279,-0.0205531996,0.1694973707,0.0432292968,0.3170821965,0.2966723442,0.1683046371,-0.0715260655,0.1160192788,0.0968123525,-0.1079168543,-0.2469825447,-0.5589663386,0.1996992677,0.086108759,0.061725568,-0.2949924767,-0.26252985,-0.0720699802,-0.1945292056,0.1462904215,0.315251112,-0.0117469188,-0.0516727678,0.0411455706,0.338816911,-0.3357954621,-0.3760806024,-0.2025689334,0.0154270008,-0.0782455131,-0.1298018396,0.0722822621,0.153594017,-0.0297240801,0.0980604738,-0.0132704731,-0.0353594497,0.2344280034,-0.1937999129,0.1276607811,0.0803777575,0.1863416731,0.006922184,0.1699039042,-0.3082532585,-0.2829098105,-0.0149562228,-0.1393337846,-0.1989016235,-0.0089012599,0.2219691277,0.1782563478,0.1121172681,0.0467124507,0.8153547049,0.2904577851,0.0116438242,-0.0301055796,0.030306533,-0.3385447264,-0.0141155692,0.3378117979,0.1740821898,-0.6436671019,-0.0770445913,0.3162026405,-0.341324091,0.1643741578,0.2579717338,-0.1845357269,0.2711555958,0.0388609543,0.0341985077,0.5991305709,-0.1869712472,-0.3114596307,0.068309322,0.067710191,0.0123742167,0.029267801,-0.0617258921,0.2344316989,-0.0872586668,0.0991235226,0.4348345995,-0.2404679954,0.029494239,-0.0098550096,-0.1206749305,0.2951943278,0.3530770838,-0.1855426133,-0.2928940654,0.0624755621,0.2671692371,0.3282123804,-0.2255352587,0.0736841857,0.1006432101,-0.267234534,0.0892196447,0.0731190518,-0.2335533947,-0.3995292783,0.2664976418,-0.3235734701,0.2553372085,0.0927843228,-0.2169161737,0.1256585419,-0.2715001404,-0.1884758621,-0.2257441878,0.2386907786,-0.0153274126,0.0980970263,0.1489165127,-0.1053274199,-0.0331549011,-0.2175190002,-0.1276916862,-0.2352512479,-0.0042178808,0.0115162125,-0.0245697871,0.0336272828,0.205608651,-0.1945265383,-0.2773687541,-0.0686696544,-0.0554685481,0.0843357369,0.1762463599,-0.1867500693,-0.0557047464,0.360755831,-0.0796791762,-0.1604560316,0.0065563787,-0.0439492241,-0.1347127557,-0.0142064327,0.5713962317,-0.0082733352,0.153434068,0.1384377629,-0.1113999635,0.4390425086,-0.2118981183,-0.0669085383,-0.3066830039,0.3075168133,-0.2405388951,-0.0250526965,-0.0206583682,-0.4667216241,0.293887794,0.5241797566,0.1195422187,0.2275018096,0.0260458216,0.3398803473,0.1909879148,-0.0613884032,0.2351474315,0.1668780297,0.1093578115,0.3604397178,-0.211672008,0.3253709078,-0.1727220714,0.0418736897,-0.2282512337,-0.00065654,0.1424652338,0.012185324,-0.0922971666,-0.1998010129,0.1816362441,0.153148219,0.138164863,0.0138187101,0.1529121101,-0.5816357732,-0.3761355281,-0.2104390562,-0.2378638238,0.0063176104,0.2482803017,0.2964659631,-0.3051173091,-0.1677930057,0.3162819445,0.3392682374,0.0744111836,-0.1750604659,-0.0538644753,-0.071854502,-0.1794320494,-0.2444062233,0.2388399094,0.3472923338,-0.0135217048,0.3145675957,-0.3366048634,-0.2414300144,0.0362312235,-0.3904020488,0.1810480058,-0.2720422745,0.0022113458,-0.146835506,0.1355131716,-0.0168852434,-0.1380896121,0.2735020518,0.1713932008,-0.2211097032,0.0685836151,-0.0055789077,0.1013226211,-0.1264633685,-0.6710653901,-0.5951109529,-0.2512134314,-0.1333300173,-0.245185107,0.3593440354,-0.2196395695,0.0570242219,-0.0083577083,0.0164688993,0.154420957,-0.089319922,-0.1075412259,0.1558358818,-0.3837592006,-0.3826019168,-0.1815673411,-0.0794025734,0.6028256416,-0.0109306956,-0.1639606804,0.1444200724,-0.217967838,0.0313021056,-0.1359931976,-0.0464085154,0.3886911273,-0.1523509026,0.0700265169,-0.0642261431,0.210116908,0.2466238439,-0.1676119715,0.1926054358,0.0213303845,0.3507079482,0.2853187323,0.6870272756,0.2225359082,-0.0186998509,0.2611910403,0.0376181938,-0.0973070636,-0.1640946716,-0.3305539787,0.1045776755,0.2790667713,0.249492079,0.3200264275,0.0075699165,-0.419473052,-0.3046821952,-0.0682018995,-0.2497380525,-0.4224902093,0.2797598541,-0.1485929787,0.2589622438,-0.080786556,-0.12018089,0.0658529475,0.0611247756,-0.1272088587,0.2753649354,-0.2356356233,0.1366682649,-0.1638369411,-0.2549642622,-0.2577459812,0.2432102561,0.1393216848,0.1295703501,-0.4074013233,-0.0348056667,0.2553428411,0.0101977875,0.4792658389,-0.0116427578,0.0334226564,-0.056810271,-0.249784857,-0.047551211,0.0518026501,-0.2437526584,-0.2110764235,0.2786969543,0.0483304113,-0.4830057621,-0.400410831,-0.1896265149,0.2210504264,-0.318015188,0.1045722961,-0.1502375752,-0.4936713576,-0.1855929196,0.2497286499,0.1380714923,0.0737947002,0.1140896082,-0.346945107,-0.0948423222,0.0271495469,0.0746498257,0.3943496346,0.0387782753,0.0003896187,0.1479102522,-0.0242580324,-0.3104053438,0.3375666142,0.4441432655,-0.0714132339,-0.4775902033,0.0689738989,-0.2635333538,0.2257966846,0.2028355747,-0.0787838474,0.0508447886,-0.1394345313,-0.0164052118,-0.2262815833,0.2664328814,0.4254592061,0.2397475392,-0.4981907904,-0.3799905479,0.1270402819,-0.1661979705,-0.0188151058,0.4364909232,0.2236266434,-0.3434129655,0.3876727819,0.1485225558,0.7250760198,-0.2446470559,0.1286921054,0.1899480969,0.1315005124,0.2746508718,-0.1346434802,-0.0511471964,-0.0727814063,-0.0371351764,-0.0880490616,0.1354049295,0.1841680259,0.3696353137,-0.2419295311,-0.0197665133,-0.0026524956,-0.3682945371,0.0759803504,0.2364657372,-0.1546994895,-0.2406041324,-0.2274138927,0.1521265954,-0.137799263,0.1421947777,-0.0245000571,-0.1441590339,0.0232242197,0.0637230724,-0.1509362757,-0.0486021712,-0.4481599033,-0.0678946599,-0.0977499783,-0.524166882,0.4261682928,0.487775445,0.3539227247,0.1410793513,-0.0583278127,0.2123113573,0.0495105945,0.1084081531,-0.0114246113,-0.2722733319,0.2864248753,-0.1215188056,-0.2456294,-0.0700947791,0.2272700965,0.1699273735,-0.058795698,0.1454828382,0.0873055309,-0.141028583,-0.0046755685,0.1079627573,0.1413411647,-0.0187478121,0.2128383219,0.1345560104,-0.1931188405,0.0653740019,-0.1284159422,-0.1850831211,-0.0785558596,-0.2814275026,0.375626415,0.2322543114,0.3384278119,0.3050632179,-0.0658050999,-0.1704979539,0.301807642,0.0218592342,-0.2949540019,-0.0365006626,0.2810054719,0.0227460526,-0.0240749903,0.3036064804,-0.284165442,-0.0699389949,-0.0845457613,-0.2926839292,0.0099825859,0.0420372486,-0.1245799512,0.105403848,0.2771078348,0.3391000926,-0.1311744452,0.5737305284,-0.3698702455,0.005886002,-0.0565496981,0.2760067582,-0.103648603,0.0499211513,0.2413436472,0.1679945141,0.0667528287,-0.038192302,-0.0351048596,-0.3488069177,-0.1829791069,0.1007537246,0.0071057286,0.1759009361,0.1313992143,-0.1661435515,-0.0925511718,-0.1509744525,0.2855470479,-0.0105730407,0.1109220535,0.4425671399,0.1738980412,0.1973272562,-0.1370546967,0.0630765557,-0.0370739959,-0.0110817906,-0.0142127154,0.0821503773,-0.1756988317,-0.2209682614,-0.0653028041,0.096694909,0.1255739182,-0.1968714446,-0.1459978074,0.1183369756,0.3806863129,-0.1864889115,0.0730318055,0.0195790976,-0.0736097768,0.1659247428,0.3901276588,0.2354630381,-0.0950925723,0.0966484994,-0.1705894023,-0.139021799,0.2327314913,0.4411738515,-0.1391784996,0.2975328565,0.5234043002,-0.2016002685,0.023544509,-0.5701315999,0.135840252,0.3580000103,-0.248292625,0.0147583503,0.2565579712,-0.1580221355,0.0489180535,-0.0746175051,-0.3017635345,0.0902372673,-0.1508256048,-0.0031722006,0.1290086359,-0.4052057862,-0.3348175883,-0.0826362222,-0.1171635389,-0.0509097651,0.0296207387,0.3479800224,-0.0709969252,-0.1315670609,0.0349302813,0.1996857673,-0.0084865708,-0.1756475419,-0.2219951749,0.0693949014,-0.0239416175,-0.0740996078,-0.0412874818,-0.1026531309,-0.072054185,0.2207167149,-0.0475663505,-0.1042251512,0.0458889119,-0.0511510149,-0.1356994808,0.0019163056,0.4835402369,0.2026075572,-0.0861568972,-0.0192566253,0.29334113,0.0328770727,-0.108520776,-0.2172896862,0.3303786516,-0.22424227,-0.0797753483,-0.0526933521,0.4048778713,0.1722880751,0.2607539892,0.2372877896,0.1530282646,-0.1554990709,-0.1858928502,0.1608017832,0.0239322148,-0.1835128963,0.4393494129,-0.0966398418,0.3170117736,-0.2066838443,-0.0554377325,-0.5749291182,0.1027127802,0.196230799,-0.2680198848,0.1372488737,-0.181008786,0.1022188514,0.0909454897,0.418143332,-0.1886613816,0.3962576687,-0.1188272163,-0.0018161706,-0.1924286485,0.1012445167,-0.0004212699,0.13541013,-0.2290120721,0.2973607481,0.4049984813,0.3552607596,0.1729693413,-0.2654713392,0.0930187926,0.2175473869,-0.2079216391,0.0469513088,-0.4969522953,0.4309385419,-0.096274972,-0.2155321091,0.2567342222,-0.0382610932,-0.0408003554,-0.0768252611,0.040742971,0.36103338,0.0791588575,0.0091055185,0.2970996499,0.1525090188,-0.1799484789,0.3190000951,-0.0694283694,-0.0266434103,-0.5007231832,0.2509753704,0.2622109354,0.1316749156,-0.3571842909,-0.3711140752,-0.1261772215,0.3362699747,0.0409440398,-0.1161235422,0.2225253731,0.0758973062,0.4533267915,-0.1562904418,0.362739712,0.6081723571,0.063625589,-0.1379100531,-0.399207741,-0.2314001024,0.1343281716,-0.1640171707,-0.183318615,0.2519248128,-0.0357958451,0.1729254872,0.1860670596,-0.4305514097,0.1040730625,0.2373618484,-0.1121477485,-0.0187773723,0.3154479265,0.2885210216,-0.0782677755,-0.278344512,0.2762378454,0.2643612921,-0.0876059756,-0.1760576814,-0.2279146165]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2132","title":"TydiQA dataset is mixed and is not split per language ","comments":"@lhoestq I greatly appreciate any updates on this. thanks a lot","body":"Hi @lhoestq \r\nCurrently TydiQA is mixed and user can only access the whole training set of all languages:\r\nhttps:\/\/www.tensorflow.org\/datasets\/catalog\/tydi_qa\r\n\r\nfor using this dataset, one need to train\/evaluate in each separate language, and having them mixed, makes it hard to use this dataset. This is much convenient for user to have them split and I appreciate your help on this. \r\n\r\nMeanwhile, till hopefully this is split per language, I greatly appreciate telling me how I can preprocess and get data per language. thanks a lot ","comment_length":11,"text":"TydiQA dataset is mixed and is not split per language \nHi @lhoestq \r\nCurrently TydiQA is mixed and user can only access the whole training set of all languages:\r\nhttps:\/\/www.tensorflow.org\/datasets\/catalog\/tydi_qa\r\n\r\nfor using this dataset, one need to train\/evaluate in each separate language, and having them mixed, makes it hard to use this dataset. This is much convenient for user to have them split and I appreciate your help on this. \r\n\r\nMeanwhile, till hopefully this is split per language, I greatly appreciate telling me how I can preprocess and get data per language. thanks a lot \n@lhoestq I greatly appreciate any updates on this. thanks a lot","embeddings":[-0.3341826797,-0.219795689,-0.2160544097,0.3134114146,0.2735143602,-0.063344866,0.3558137417,0.3471710682,-0.2479394227,0.1068426371,-0.3467676044,-0.0342324264,0.1134987772,0.3849471211,-0.0458300598,-0.4077254236,-0.145146504,0.1169788316,-0.2427179664,0.0096723223,0.0590936616,0.1880040169,-0.2538176477,0.1141905859,0.0407580845,-0.2891880572,-0.1372783184,-0.0962979496,0.0062398957,-0.066113703,0.3586678207,0.0637228563,0.2767430544,0.3454371691,-0.000107033,0.1514690965,-0.0559867509,-0.259452194,0.1849095672,-0.1034398153,-0.1377590299,-0.1611643136,-0.1590997428,-0.2146858424,-0.5707762837,-0.0994699001,-0.0266018305,-0.377333194,0.4266954064,0.3204797208,0.1613897085,0.0901735425,-0.083493948,0.1302596629,-0.0305768885,0.0198310371,-0.1586862803,-0.0940747857,0.5296193361,0.0149319014,0.2982111275,0.2727367282,0.2931114435,0.0497655086,-0.4348328412,0.0173865668,-0.0043154764,-0.4805805385,0.0829970464,0.5706542134,0.5673200488,0.1022830531,-0.2374247462,0.0263976231,-0.0994557515,-0.2368197292,0.1380312294,0.3702037036,-0.0598399155,0.2787835896,0.3277665377,0.1662658453,-0.2879876792,0.1017285511,-0.2917073071,0.2383697033,-0.0558549725,0.0337253548,-0.0460467711,-0.1530551165,0.2351774126,0.3327561319,-0.0359154232,0.1539257467,-0.4239603281,-0.3202764988,-0.1663822234,-0.3785956204,-0.1728868037,-0.3643614054,0.1729641557,0.3011416793,-0.1277855933,0.0015709973,0.2482796907,-0.0736487657,0.2656732202,0.389351368,0.2302724421,-0.0720902309,0.0209431648,0.0107350023,-0.2109262198,-0.3366945982,-0.5842689276,0.2606152892,0.0651226044,-0.0203891341,-0.2942717075,-0.2605697513,-0.077789709,-0.1822794676,0.1094081625,0.3357257843,-0.0115643907,-0.1001846492,0.019196555,0.3687182367,-0.2632759213,-0.475985378,-0.1439901441,0.0249915514,-0.114419423,-0.1345222443,0.0629137903,0.2963956296,-0.0611799322,0.042307537,-0.0188184194,-0.0975157619,0.2771528363,-0.270760268,0.21303913,-0.0072916527,0.1962269843,0.0031131804,0.0975463986,-0.1886541694,-0.2623991668,0.0896683559,-0.2720659077,-0.2368646413,0.0322301909,0.1798862517,0.3559785187,0.1455612183,0.1820339262,0.6407597065,0.3432513475,-0.0181704722,0.0276834499,0.0707935169,-0.1901146621,-0.0999596789,0.2893732786,0.0155329015,-0.6631097198,-0.1160600185,0.2150276899,-0.2834798098,0.3190277815,0.3184204996,-0.1294466406,0.3321465552,0.0764848515,-0.1160115153,0.7449868917,-0.1764701158,-0.4513644576,-0.0280478019,0.1341896206,-0.0809646249,0.0247878507,0.0738922879,0.2510614693,-0.0192803349,0.1240604669,0.517721355,-0.2435470074,0.0233131368,-0.035811238,-0.12494082,0.2566721737,0.3442321122,-0.1270582974,-0.2786641121,0.0671768188,0.2204769701,0.3840945661,-0.1397107393,0.1215269342,0.0785965025,-0.173194319,0.0848650858,0.0723704845,-0.1774570495,-0.1559367478,0.1932966262,-0.4467013776,0.1175358891,0.1057371721,-0.2129176855,0.1208324134,-0.2870440483,-0.2118051946,-0.224233672,0.2006084323,-0.0539895967,0.1610208005,0.1804329008,0.0056310394,-0.0918603316,-0.1631024182,-0.1371694505,-0.2076457143,0.051311288,-0.0605421215,-0.0771815404,0.0618007891,0.2999250889,-0.1164596602,-0.2530798316,-0.0627337322,-0.0396612622,0.0089333486,0.1692280471,-0.156003952,-0.2488587052,0.3501918018,-0.17268309,-0.0802313536,-0.0919748992,-0.0875357836,-0.1606047601,0.0502602682,0.5349607468,-0.014916623,0.1154595092,0.0795533359,-0.0375651941,0.3595077395,-0.2502317727,0.0795722082,-0.2492515445,0.2685214877,-0.2445413172,-0.0129758241,-0.0968061611,-0.3880863786,0.2395841479,0.6105270386,0.1944782883,0.280099839,-0.0745308399,0.2943936288,0.2021717429,-0.0609518066,0.2526619136,0.0992240906,0.0270312391,0.3604414165,-0.2349016964,0.4523713589,-0.1639474779,0.0001045742,-0.2162144482,0.01628845,0.0491706952,-0.0395625718,-0.0254436284,0.0253688488,0.2684893906,0.1568875611,0.1884227693,0.1004190072,0.1147316173,-0.5336604118,-0.3531954587,-0.1357716918,-0.3193672299,0.1438200623,0.2265342027,0.2347002327,-0.4390042424,-0.0687010437,0.288482219,0.3474974334,0.1198111102,-0.0701347589,0.0368116722,-0.0301041864,-0.2298366874,-0.165179804,0.2237248123,0.3004914522,-0.0708124563,0.2994616926,-0.3053939044,-0.1573508978,0.0095013734,-0.2742449045,0.1601751894,-0.2291736305,-0.0532391705,-0.1842290163,0.1069146693,0.0342074148,-0.119120717,0.2510519922,0.1581301987,-0.1644051969,0.1193826869,-0.062673755,0.0493070073,-0.0121718328,-0.627415657,-0.6424371004,-0.2232624292,-0.1254791766,-0.1603544354,0.3598537743,-0.5451484919,-0.0008339691,-0.0218067523,-0.0265086219,0.1445773393,-0.0292999782,-0.2068520635,0.1270244271,-0.4478138983,-0.4501766264,-0.0846405178,-0.0231984053,0.7710687518,-0.0875607505,-0.1822045594,0.3894678056,-0.2245053798,0.0385755226,-0.0618317164,-0.141659379,0.3277653754,-0.1759009212,0.0660660937,-0.042374745,0.3298002183,0.2182697654,-0.2301257998,0.2598190904,-0.0388481393,0.3564358652,0.2895714641,0.7208630443,0.2424979061,-0.0573878177,0.1715794057,0.065846689,-0.2438090593,-0.1186279431,-0.3257149756,0.1334428489,0.2396509498,0.2534110844,0.2653037012,0.0285407919,-0.447021395,-0.3439774215,-0.210875839,-0.2924391627,-0.4148337543,0.2576583326,-0.2251511812,0.267719239,-0.0693931431,-0.1156919152,0.0527055226,-0.039207641,-0.141172111,0.1826926023,-0.2571318746,0.0868143067,-0.1533453614,-0.3684000373,-0.1972416788,0.3145287037,0.0659773499,0.1984093487,-0.4402619004,-0.0355020016,0.2964382768,-0.0160648245,0.3480646014,0.018359242,-0.0058424384,-0.0643963292,-0.196383208,0.0647079498,0.1155658439,-0.2693772018,-0.0486984067,0.3292928338,-0.0511681437,-0.3023515046,-0.3346575797,-0.0572569221,0.1327728778,-0.3655069172,0.126726985,-0.1792030036,-0.3023888171,-0.1126713455,0.2212357968,0.0010639357,-0.0064654928,-0.0168045145,-0.4505140483,-0.1666927785,0.1072768122,0.0541482903,0.3336575031,0.0550703891,-0.2259818465,0.1987225264,-0.0341394879,-0.2511054575,0.4930507839,0.4068543613,-0.0379316956,-0.3640225232,0.0955864042,-0.3782552481,0.3291049302,0.1752371788,-0.0782855675,0.1048159525,-0.145490557,0.0614679381,-0.1850657612,0.1332270503,0.5149447322,0.2060796767,-0.6102374792,-0.3330621123,0.0419504084,-0.0953628868,-0.0222050883,0.4956060946,0.1412348151,-0.2657104433,0.4428017139,0.1394456029,0.7930889726,-0.2444378585,0.1250111163,0.0649893284,0.2207368314,0.1622461826,-0.1764516979,-0.0308546964,-0.1063769236,-0.0085093286,-0.1180347875,0.1115536839,0.2691453397,0.3575242758,-0.4042754173,-0.0411569104,-0.1026654169,-0.4788879752,-0.0269338265,0.1307571828,-0.1471298337,-0.1028427556,-0.1250110716,0.0823379532,-0.1207392141,0.1174729466,-0.0301985834,-0.2146517485,-0.0984267741,0.1033413485,-0.0825689957,0.0526421107,-0.3035773337,-0.082002148,-0.0907941908,-0.5577977896,0.444262743,0.4975311458,0.4504649937,0.2037342191,-0.1101311818,0.3311321437,0.0928212777,-0.0855534002,0.0114166811,-0.1798416227,0.1379630268,-0.1516302228,-0.3137019873,0.0779600739,0.173752293,0.1916474551,-0.1314036995,0.2697084248,0.2640148401,-0.1752762198,-0.0504103675,0.1271533966,0.0186739843,-0.053927388,0.2056313604,0.1911983788,-0.2079934627,0.106382072,-0.1734172255,-0.2195421755,-0.0860516056,-0.3373149633,0.4061424136,0.2861096263,0.4155091941,0.1482646316,-0.0577155277,-0.1935301125,0.402094394,0.0306597855,-0.2608371079,-0.1366213858,0.2080632448,0.0301338285,-0.0128309177,0.2771270573,-0.3000468314,-0.1507990807,-0.1967320889,-0.3376718163,-0.0087729879,0.0220239572,-0.0955606699,0.0282264259,0.2722578645,0.2682134509,-0.0821796432,0.6825973988,-0.3123399615,0.0734404474,-0.1370053142,0.3113692105,-0.2457566261,-0.0554626435,0.1387242824,0.2057183385,0.0421581157,0.0028565254,0.0768043622,-0.3136995137,-0.0971516669,0.1223925725,0.0333188958,0.2022057623,0.0274683964,-0.2317596823,-0.1000616774,-0.0342453457,0.1807026565,-0.0495013595,0.2013694644,0.4807705581,0.2438490838,0.22545214,-0.2045208812,0.1556846201,-0.126342684,0.0659703165,-0.06463442,0.0627044588,-0.1858199537,-0.2634203732,-0.1430743635,0.0494022407,0.1397846937,-0.1635315865,-0.026601119,0.1175560206,0.4828469157,-0.2593568861,-0.0056786132,0.06837859,-0.0728225112,0.2397686988,0.3747881055,0.1914761513,-0.037016917,0.0962004438,-0.2642201483,-0.1772302389,0.2046131939,0.3458667994,-0.1508930922,0.2661679685,0.4690912664,-0.2513614595,0.0496505238,-0.612142086,0.1442756504,0.4140922725,-0.170547381,-0.0038789392,0.2451522648,-0.1102439314,0.0003436618,-0.0316995308,-0.3134559095,0.1350121349,-0.0500967652,-0.1175222397,0.1954541504,-0.299048692,-0.3336952627,-0.0597808063,-0.1199838966,-0.0265645813,-0.0017582228,0.4174157679,-0.0392127149,-0.1875279099,-0.0220227335,0.1502452344,0.0194423217,-0.1387476623,-0.2578671873,0.0659648329,-0.0315253362,-0.1660295427,0.0282188468,0.0230876897,-0.1314381808,0.3070317209,-0.0024187635,-0.03338762,0.1728599072,0.0050840261,-0.1557624191,-0.022625545,0.4310380518,0.1749167144,-0.0739917532,0.0935711339,0.2067303061,-0.0054713953,-0.0252876543,-0.3753136694,0.3562538624,-0.1760766208,-0.201954022,-0.0200441238,0.4143506885,0.2356769741,0.2862789929,0.2584941983,0.1512291431,-0.145222038,-0.1752874851,0.2011549771,0.2060983032,-0.1430915743,0.4231911302,-0.2401476651,0.1360337287,-0.1739275455,-0.0433620811,-0.6170922518,0.0310267024,0.0665762052,-0.3644182682,0.145884946,-0.1952064931,0.0621065944,0.1342538595,0.4330539107,-0.2222086191,0.5335527062,-0.1469089836,-0.1772122979,-0.1543294638,0.1766443998,-0.0487656407,0.0975188613,-0.2551178932,0.2996798754,0.3221598268,0.2612881362,0.3221772015,-0.2587347627,0.0143294288,0.2929213643,-0.2763554752,0.0707175732,-0.4496517777,0.4090987146,-0.1472508907,-0.2698590159,0.3830693662,0.0370922983,-0.0650855899,0.0343311541,-0.1235174611,0.3012859821,0.307811141,0.0526233912,0.2859829366,0.1080696434,-0.1818976104,0.3374500573,-0.0495579876,-0.0585871674,-0.2757332325,0.2795338035,0.319342941,0.0410476141,-0.3684865534,-0.2978402078,-0.0808067322,0.3212357461,-0.0269998107,-0.1512754261,0.1901633888,-0.0185536332,0.4699672163,-0.0612486787,0.448472321,0.4851565659,0.0574183576,-0.2428665608,-0.3953605592,-0.0815058872,0.1279604584,-0.1125093922,-0.1617418528,0.1765657514,-0.0612519123,0.1604031473,0.2848606408,-0.3357298672,0.0317497253,0.2764459848,-0.0927020758,0.0536396913,0.2155440599,0.3506450653,0.000297292,-0.2660468519,0.4019908309,0.1996868551,-0.0453990102,-0.1307315826,-0.1653575748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2131","title":"When training with Multi-Node Multi-GPU the worker 2 has TypeError: 'NoneType' object","comments":"Hi ! Thanks for reporting\r\nI was able to reproduce this issue. This was caused by missing split infos if a worker reloads the cache of the other worker.\r\n\r\nI just opened https:\/\/github.com\/huggingface\/datasets\/pull\/2137 to fix this issue","body":"\bversion: 1.5.0\r\nmet a very strange error, I am training large scale language model, and need train on 2 machines(workers).\r\nAnd sometimes I will get this error `TypeError: 'NoneType' object is not iterable`\r\nThis is traceback\r\n```\r\n\r\n71 | \u00a0 | Traceback (most recent call last):\r\n-- | -- | --\r\n72 | \u00a0 | File \"run_gpt.py\", line 316, in \r\n73 | \u00a0 | main()\r\n74 | \u00a0 | File \"run_gpt.py\", line 222, in main\r\n75 | \u00a0 | delimiter=\"\\t\", column_names=[\"input_ids\", \"attention_mask\", \"chinese_ref\"])\r\n76 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 747, in load_dataset\r\n77 | \u00a0 | use_auth_token=use_auth_token,\r\n78 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 513, in download_and_prepare\r\n79 | \u00a0 | self.download_post_processing_resources(dl_manager)\r\n80 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 673, in download_post_processing_resources\r\n81 | \u00a0 | for split in self.info.splits:\r\n82 | \u00a0 | TypeError: 'NoneType' object is not iterable\r\n83 | \u00a0 | WARNING:datasets.builder:Reusing dataset csv (\/usr\/local\/app\/.cache\/huggingface\/datasets\/csv\/default-1c257ebd48e225e7\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2)\r\n84 | \u00a0 | Traceback (most recent call last):\r\n85 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 193, in _run_module_as_main\r\n86 | \u00a0 | \"__main__\", mod_spec)\r\n87 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 85, in _run_code\r\n88 | \u00a0 | exec(code, run_globals)\r\n89 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 340, in \r\n90 | \u00a0 | main()\r\n91 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 326, in main\r\n92 | \u00a0 | sigkill_handler(signal.SIGTERM, None) # not coming back\r\n93 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 301, in sigkill_handler\r\n94 | \u00a0 | raise subprocess.CalledProcessError(returncode=last_return_code, cmd=cmd)\r\n\r\n```\r\nOn worker 1 it loads the dataset well, however on worker 2 will get this error. \r\nAnd I will meet this error from time to time, sometimes it just goes well.","comment_length":37,"text":"When training with Multi-Node Multi-GPU the worker 2 has TypeError: 'NoneType' object\n\bversion: 1.5.0\r\nmet a very strange error, I am training large scale language model, and need train on 2 machines(workers).\r\nAnd sometimes I will get this error `TypeError: 'NoneType' object is not iterable`\r\nThis is traceback\r\n```\r\n\r\n71 | \u00a0 | Traceback (most recent call last):\r\n-- | -- | --\r\n72 | \u00a0 | File \"run_gpt.py\", line 316, in \r\n73 | \u00a0 | main()\r\n74 | \u00a0 | File \"run_gpt.py\", line 222, in main\r\n75 | \u00a0 | delimiter=\"\\t\", column_names=[\"input_ids\", \"attention_mask\", \"chinese_ref\"])\r\n76 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 747, in load_dataset\r\n77 | \u00a0 | use_auth_token=use_auth_token,\r\n78 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 513, in download_and_prepare\r\n79 | \u00a0 | self.download_post_processing_resources(dl_manager)\r\n80 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 673, in download_post_processing_resources\r\n81 | \u00a0 | for split in self.info.splits:\r\n82 | \u00a0 | TypeError: 'NoneType' object is not iterable\r\n83 | \u00a0 | WARNING:datasets.builder:Reusing dataset csv (\/usr\/local\/app\/.cache\/huggingface\/datasets\/csv\/default-1c257ebd48e225e7\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2)\r\n84 | \u00a0 | Traceback (most recent call last):\r\n85 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 193, in _run_module_as_main\r\n86 | \u00a0 | \"__main__\", mod_spec)\r\n87 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 85, in _run_code\r\n88 | \u00a0 | exec(code, run_globals)\r\n89 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 340, in \r\n90 | \u00a0 | main()\r\n91 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 326, in main\r\n92 | \u00a0 | sigkill_handler(signal.SIGTERM, None) # not coming back\r\n93 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 301, in sigkill_handler\r\n94 | \u00a0 | raise subprocess.CalledProcessError(returncode=last_return_code, cmd=cmd)\r\n\r\n```\r\nOn worker 1 it loads the dataset well, however on worker 2 will get this error. \r\nAnd I will meet this error from time to time, sometimes it just goes well.\nHi ! Thanks for reporting\r\nI was able to reproduce this issue. This was caused by missing split infos if a worker reloads the cache of the other worker.\r\n\r\nI just opened https:\/\/github.com\/huggingface\/datasets\/pull\/2137 to fix this issue","embeddings":[-0.1806102991,-0.4468090534,0.0129236523,0.5495145917,0.1124854609,-0.0091340337,0.5815147161,0.2967011034,0.1129247844,0.2367109209,0.3380283117,0.0173764732,-0.1162154898,0.1675734371,-0.042410925,-0.2821591198,-0.1764215678,0.0434676446,-0.227805078,-0.214592278,-0.3433728218,0.1396662146,-0.1313170046,0.1889922619,-0.5063354373,-0.4234542847,-0.1063818559,-0.0327261165,0.2665303946,-0.2152226418,0.2046148032,-0.3352004588,0.1856167316,0.8994958401,-0.0001186402,0.148360461,0.1208232865,0.0624540038,-0.144217357,-0.4792484939,0.083634451,-0.3793367445,0.0438947678,-0.372520566,-0.1640225798,-0.1153089702,0.2508283257,0.0683696643,0.038129054,0.2097055912,0.1652571112,0.3379758,0.1185292006,-0.0910865888,-0.1230920553,-0.078175433,-0.124581486,-0.0570696816,-0.1463324875,0.2151985615,0.1408524811,0.3001521528,0.0600437447,0.0522123091,-0.0947405323,-0.0429644659,0.5682382584,-0.4400096834,0.0403892882,0.3434166312,0.0855879858,0.0078646038,-0.2014046609,-0.1323599219,0.1542741805,-0.118224524,0.1847367287,0.3629119694,-0.0531053208,0.1500600725,-0.1300462335,-0.0676265731,-0.2308019549,0.0534002073,-0.1115095243,0.0901557282,-0.1316496879,0.4046879113,0.1569740921,0.0425376371,0.1446679085,-0.1151811853,0.1561760455,0.0373051763,-0.4768414795,-0.2667486668,0.0841662511,-0.1505966187,-0.1545930207,-0.1207949668,0.1211863682,-0.1792515069,-0.1328291744,0.3402727544,0.2215323895,-0.208968699,-0.0280376002,0.0667775273,0.1572731435,-0.1440431923,-0.3448980153,0.0656591803,-0.123423703,-0.3975611329,0.0672159195,0.3313922286,0.550983429,0.0814335942,-0.2558076084,-0.3254729509,-0.4844892621,0.0634808987,0.1591971517,0.071599789,-0.0215618368,0.7668600082,-0.1380043477,0.1425333768,-0.2015482634,-0.4433032274,-0.150910601,0.2117178291,-0.2316481024,0.0988812521,-0.0608816147,0.0209043287,-0.0537675731,0.0669301525,0.006623521,-0.0675481558,0.2549972832,-0.6412526369,0.0265079904,0.2047789544,0.1866111457,0.1211187914,0.1911760271,-0.0848483518,-0.1786593199,0.0570020229,-0.4990594983,-0.264444977,-0.0435095765,0.071315363,0.1389947832,0.2218963653,0.0635314956,0.151714012,0.5338413119,-0.0446067266,0.041408807,-0.3242377341,-0.3369706869,-0.3063197136,-0.3215333521,0.4540395737,0.0853606761,-0.0651732013,0.2063797265,-0.058066532,0.3713686466,0.4545201957,0.0005992663,0.0837388784,-0.1234111786,0.0138952779,0.2660521865,-0.2399116904,0.0215655714,0.4763100743,-0.4028788805,0.0788712651,0.3922178149,-0.0496553518,-0.2771501541,-0.0372722521,0.1895166337,0.3423955739,-0.0628310069,0.009341618,-0.2271394581,-0.0792751983,0.4815947115,-0.1569288224,0.3270474672,0.0703963637,-0.0276751965,-0.1134375632,0.2275328934,-0.3470682502,-0.0631333143,0.1020137817,0.2154045105,-0.1462944895,0.0686946362,-0.1687207073,-0.2463460565,0.0542072281,-0.1364753246,0.1516544819,0.2225850374,-0.0803485885,0.2349827886,-0.2266496718,-0.2233921885,0.0553607643,0.0752304047,0.3720036149,-0.2636727989,0.1632019877,0.0245809779,0.0021455172,-0.3904486895,-0.2759417295,-0.1010340154,0.2078496218,-0.0291660819,-0.1070364043,-0.0839110985,0.2642500401,0.4416565597,-0.2129094154,0.2406531572,0.1404104829,-0.1080904603,0.0494443811,-0.0524048693,0.1803079396,0.35804829,-0.1872611344,-0.1409421414,0.2008781582,0.134699285,-0.1484245509,0.3699091673,0.4953466356,0.090521425,0.2097523212,0.0449271016,0.078612186,0.1215231866,0.0240474232,-0.0164274201,0.0827938691,0.0056439694,0.1275957525,-0.1125174537,0.0920489132,-0.6476395726,0.0956622809,0.8060963154,0.1173804998,0.4062262475,0.0086445129,0.1291960925,0.0277588628,-0.154481113,-0.0285173319,0.4229156077,-0.0422861688,0.2465559989,-0.0392972976,-0.0626192093,-0.1413778216,0.0522282161,-0.1825522035,0.7166270614,0.2830475867,0.132720679,-0.0022643518,-0.0056484104,-0.2538818419,0.2736298442,0.450527519,-0.3334290981,0.3502019644,-0.3673314154,-0.0159904975,-0.2581938505,0.2661478817,-0.0938294679,-0.0245775245,-0.2472632974,0.1864033341,0.1503834277,0.1223219782,0.35891819,0.1457659602,0.1071996093,0.2317183763,0.2299328446,0.0111648263,-0.2177335769,-0.072553508,0.2922357321,-0.3358941078,0.0090650646,-0.1226599514,-0.5176730752,-0.351428628,-0.0638254508,0.0656745881,0.013568298,0.0029365411,0.2864256203,0.0785256252,-0.0120970001,-0.0619992651,0.1552059203,-0.4553328753,-0.1128283218,0.2476546913,-0.3855648041,0.1096370816,-0.1286381036,-0.2846000791,-0.2985165715,-0.1682437807,0.6335537434,0.036101602,0.4326667488,0.1627968699,0.1140319705,0.1608608812,0.1367166787,0.0956066176,-0.1563305557,-0.4536629319,0.1469080001,0.0463184677,-0.2639672458,-0.0813129917,0.0779389441,0.3259366453,0.1940153837,-0.301644206,-0.2473767698,0.0698908642,-0.0073892139,-0.0802942067,0.043606136,0.3106846511,-0.081747666,-0.0795649514,0.1049370244,0.1663469374,0.1308562309,0.2061449587,0.2597747743,0.215860039,0.2530972958,0.2540295124,0.535112977,0.2206321359,-0.0674460903,0.2258820832,-0.1522954851,-0.004480199,0.1564436704,-0.3594617546,0.1473370194,-0.2330273092,-0.1332301199,-0.0466546267,-0.441717416,-0.5093246102,-0.1664875448,-0.1738309413,-0.3591647446,-0.1715164185,0.2290109694,-0.2058067173,0.3189735711,-0.2008253485,-0.0663316622,-0.2482412905,0.020823529,-0.0461255014,0.2327274233,-0.0202390645,-0.0802518353,-0.413092345,-0.5657935143,-0.0441711657,0.2615533471,0.0113745593,0.4401904047,-0.1060471758,0.019314928,0.1748602092,-0.0069007189,0.8240171075,-0.2342891544,0.0243170429,0.2899026871,-0.5752906799,-0.0847982988,-0.0399576202,-0.2760018408,0.4216439724,0.5022407174,0.3049765825,-0.1920865029,-0.3170815706,0.2327643633,-0.1730203927,-0.1513353437,-0.0994562805,-0.2681988478,-0.1267526001,0.0031301854,0.0991752595,0.0423637964,0.41966483,-0.4223131835,-0.0110674258,-0.3594005704,0.1035603508,0.0992429331,0.0183773581,-0.0220204499,-0.0157828331,-0.2933521867,0.0666060001,0.4181612432,0.3456064463,0.1352541447,0.1237927079,-0.222049877,0.3866360784,-0.2193987668,0.3998539448,0.4414469898,-0.0689045191,0.2593995333,0.046429269,0.3088773489,0.0099545876,-0.2134131789,0.5013944507,0.0964831039,-0.3567266762,-0.4405709505,0.2417843342,0.2308721542,0.1443294287,0.0753162056,-0.3131642342,-0.0920787081,-0.0364313535,-0.0348871164,0.8628455997,0.1303196251,0.1319111735,0.2637454867,0.0648117885,0.7199362516,-0.4422337711,0.3917687237,-0.2568787336,-0.1663816869,-0.1640912145,-0.0817656144,0.3914418817,0.2064569592,-0.2381402701,0.3658711314,-0.1030779406,-0.1616385281,-0.158077091,0.1148330495,-0.2505121827,-0.0613890663,-0.149653703,-0.0496145748,0.0577678978,0.1704435498,-0.1470059156,-0.3725346625,0.3594648838,-0.0050592548,-0.2195948511,-0.0886685327,-0.3248027563,0.2875602543,0.4124962687,-0.4293585122,0.237770915,0.2076031417,0.3087032437,-0.2606231272,-0.1220987365,-0.2071407288,-0.2024772912,-0.0072195283,-0.2808032334,-0.1926895678,0.5347767472,-0.1014083549,-0.5287340879,0.2500193417,-0.0031347205,-0.0043283291,0.0241581351,0.0992155746,-0.4970460236,0.2518159449,-0.0458345078,0.1943402141,-0.167885825,-0.2178985178,0.0715451837,-0.0830308795,-0.6331080198,-0.0050315158,-0.0159770828,-0.406650275,0.188466534,0.3562635779,0.5661857724,0.4459779859,0.4622503519,0.3645583689,-0.0835475028,-0.3533835411,0.2155154347,0.3437524438,-0.3580479026,0.226923272,-0.0134585239,0.0545984097,0.0333130844,0.1038374528,0.0438436568,0.5941749811,-0.1677767783,-0.2888618112,0.1138362288,-0.1244143173,-0.2211531848,-0.1914723366,0.0318109803,0.4678594172,0.3696142137,0.1560432464,-0.2256106585,0.183211565,-0.6862777472,0.1564278305,-0.2976360619,-0.1875285208,0.0351730846,0.2281868011,0.011276192,0.0311042834,-0.0151206814,-0.1135714129,-0.1084555835,0.1914585084,0.0762739256,0.2077237517,0.219142586,0.0013007895,-0.2629858851,0.0861249492,0.1031461284,0.0263838526,-0.0127823241,0.3347818255,-0.1318498403,0.1470612586,-0.0516887121,0.0160815567,-0.1663547903,-0.0264431108,0.0684207007,0.2599076033,0.2681325674,0.1051005349,-0.2012492418,-0.0905272663,-0.1690200418,0.1310450733,-0.0031563214,-0.3379543722,-0.1537528336,-0.0610398687,-0.1287465096,0.0786517039,-0.2081065625,0.1642848402,-0.0082762223,0.075190343,-0.1720363051,0.0160495415,0.3984807432,0.0401428938,-0.0545231737,0.3129579723,0.1964588016,-0.014788745,-0.2255492657,0.0989417359,-0.1798563451,-0.5493915081,0.1987308264,0.0789996162,-0.172555238,0.1692458093,0.2340233773,0.1941285282,-0.0107985949,0.1986323446,0.1553979218,0.2505668998,0.0114463819,0.0647496507,0.1874623001,-0.3419452906,-0.1851764172,0.1027245522,-0.187814936,-0.0758213401,-0.2439704537,0.1061948091,-0.3426975012,-0.2329878658,-0.1344069839,0.1067933068,-0.3716104925,-0.07589636,-0.4822933078,-0.033825215,-0.2222220749,-0.0620703995,0.1131879613,0.0784955546,0.3637962639,0.2248342037,-0.1549200863,-0.3011435568,0.0974638313,0.1635616869,0.2713882327,-0.222735852,0.0118846055,0.1044878513,-0.0192578882,-0.2971428931,0.4542773366,0.3316576183,0.3410503566,-0.227893427,-0.1521065682,0.1208092198,-0.1401267946,0.0169089995,-0.0865086541,0.1232791767,0.2424122095,0.206556499,0.0007886984,-0.050825499,0.3772299886,0.1411721408,0.0152803967,-0.3947429955,0.4997736812,-0.3795516193,0.0404789969,-0.0975971743,0.3181056678,-0.5076952577,0.1147272587,0.6606998444,-0.203587383,0.2390411943,0.0054493127,0.0247680712,-0.1507310867,0.4882811606,0.0556453913,-0.0264799502,-0.2904711664,0.121670261,-0.1169741154,0.5847662687,-0.3455072045,0.2629430294,-0.0381880477,0.602992177,-0.0786876529,-0.0741019771,0.1273525506,0.1574851871,0.1948416084,0.3991425633,-0.4147590399,-0.2249621451,-0.3403022885,-0.1307871044,0.1633239239,-0.3058509529,0.1660936028,-0.1133297086,0.0163280759,-0.0395119525,0.0059385221,-0.2246301025,0.0200872887,0.2809771895,0.1872590929,0.2327943444,0.0099964524,-0.0424147658,0.0101231411,-0.1967514157,-0.2997451127,0.5487582088,0.0568799414,0.0997763723,-0.1939614415,0.2067076266,-0.1505220085,0.2600656152,-0.0193499308,0.0963500366,0.0100755356,0.2616032362,-0.0259610247,-0.0744594187,0.0825509652,0.0806281418,0.1278177947,-0.2672429383,-0.1666998714,-0.0958619118,0.3261485994,-0.3049176931,-0.3767558932,-0.4874207675,0.3538348377,0.0580240414,0.1417587101,-0.4789167345,-0.1494803429,0.3774831295,-0.0684738085,-0.3998480439,-0.0441302881,-0.0841486678,-0.1253637522,0.0760098025,-0.2112776041,-0.1244162843,-0.2668003738,0.1230953559,-0.3076861203]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2131","title":"When training with Multi-Node Multi-GPU the worker 2 has TypeError: 'NoneType' object","comments":"The PR got merged :)\r\nFeel free to try it out on the `master` branch","body":"\bversion: 1.5.0\r\nmet a very strange error, I am training large scale language model, and need train on 2 machines(workers).\r\nAnd sometimes I will get this error `TypeError: 'NoneType' object is not iterable`\r\nThis is traceback\r\n```\r\n\r\n71 | \u00a0 | Traceback (most recent call last):\r\n-- | -- | --\r\n72 | \u00a0 | File \"run_gpt.py\", line 316, in \r\n73 | \u00a0 | main()\r\n74 | \u00a0 | File \"run_gpt.py\", line 222, in main\r\n75 | \u00a0 | delimiter=\"\\t\", column_names=[\"input_ids\", \"attention_mask\", \"chinese_ref\"])\r\n76 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 747, in load_dataset\r\n77 | \u00a0 | use_auth_token=use_auth_token,\r\n78 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 513, in download_and_prepare\r\n79 | \u00a0 | self.download_post_processing_resources(dl_manager)\r\n80 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 673, in download_post_processing_resources\r\n81 | \u00a0 | for split in self.info.splits:\r\n82 | \u00a0 | TypeError: 'NoneType' object is not iterable\r\n83 | \u00a0 | WARNING:datasets.builder:Reusing dataset csv (\/usr\/local\/app\/.cache\/huggingface\/datasets\/csv\/default-1c257ebd48e225e7\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2)\r\n84 | \u00a0 | Traceback (most recent call last):\r\n85 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 193, in _run_module_as_main\r\n86 | \u00a0 | \"__main__\", mod_spec)\r\n87 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 85, in _run_code\r\n88 | \u00a0 | exec(code, run_globals)\r\n89 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 340, in \r\n90 | \u00a0 | main()\r\n91 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 326, in main\r\n92 | \u00a0 | sigkill_handler(signal.SIGTERM, None) # not coming back\r\n93 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 301, in sigkill_handler\r\n94 | \u00a0 | raise subprocess.CalledProcessError(returncode=last_return_code, cmd=cmd)\r\n\r\n```\r\nOn worker 1 it loads the dataset well, however on worker 2 will get this error. \r\nAnd I will meet this error from time to time, sometimes it just goes well.","comment_length":15,"text":"When training with Multi-Node Multi-GPU the worker 2 has TypeError: 'NoneType' object\n\bversion: 1.5.0\r\nmet a very strange error, I am training large scale language model, and need train on 2 machines(workers).\r\nAnd sometimes I will get this error `TypeError: 'NoneType' object is not iterable`\r\nThis is traceback\r\n```\r\n\r\n71 | \u00a0 | Traceback (most recent call last):\r\n-- | -- | --\r\n72 | \u00a0 | File \"run_gpt.py\", line 316, in \r\n73 | \u00a0 | main()\r\n74 | \u00a0 | File \"run_gpt.py\", line 222, in main\r\n75 | \u00a0 | delimiter=\"\\t\", column_names=[\"input_ids\", \"attention_mask\", \"chinese_ref\"])\r\n76 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 747, in load_dataset\r\n77 | \u00a0 | use_auth_token=use_auth_token,\r\n78 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 513, in download_and_prepare\r\n79 | \u00a0 | self.download_post_processing_resources(dl_manager)\r\n80 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 673, in download_post_processing_resources\r\n81 | \u00a0 | for split in self.info.splits:\r\n82 | \u00a0 | TypeError: 'NoneType' object is not iterable\r\n83 | \u00a0 | WARNING:datasets.builder:Reusing dataset csv (\/usr\/local\/app\/.cache\/huggingface\/datasets\/csv\/default-1c257ebd48e225e7\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2)\r\n84 | \u00a0 | Traceback (most recent call last):\r\n85 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 193, in _run_module_as_main\r\n86 | \u00a0 | \"__main__\", mod_spec)\r\n87 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 85, in _run_code\r\n88 | \u00a0 | exec(code, run_globals)\r\n89 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 340, in \r\n90 | \u00a0 | main()\r\n91 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 326, in main\r\n92 | \u00a0 | sigkill_handler(signal.SIGTERM, None) # not coming back\r\n93 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 301, in sigkill_handler\r\n94 | \u00a0 | raise subprocess.CalledProcessError(returncode=last_return_code, cmd=cmd)\r\n\r\n```\r\nOn worker 1 it loads the dataset well, however on worker 2 will get this error. \r\nAnd I will meet this error from time to time, sometimes it just goes well.\nThe PR got merged :)\r\nFeel free to try it out on the `master` branch","embeddings":[-0.1806102991,-0.4468090534,0.0129236523,0.5495145917,0.1124854609,-0.0091340337,0.5815147161,0.2967011034,0.1129247844,0.2367109209,0.3380283117,0.0173764732,-0.1162154898,0.1675734371,-0.042410925,-0.2821591198,-0.1764215678,0.0434676446,-0.227805078,-0.214592278,-0.3433728218,0.1396662146,-0.1313170046,0.1889922619,-0.5063354373,-0.4234542847,-0.1063818559,-0.0327261165,0.2665303946,-0.2152226418,0.2046148032,-0.3352004588,0.1856167316,0.8994958401,-0.0001186402,0.148360461,0.1208232865,0.0624540038,-0.144217357,-0.4792484939,0.083634451,-0.3793367445,0.0438947678,-0.372520566,-0.1640225798,-0.1153089702,0.2508283257,0.0683696643,0.038129054,0.2097055912,0.1652571112,0.3379758,0.1185292006,-0.0910865888,-0.1230920553,-0.078175433,-0.124581486,-0.0570696816,-0.1463324875,0.2151985615,0.1408524811,0.3001521528,0.0600437447,0.0522123091,-0.0947405323,-0.0429644659,0.5682382584,-0.4400096834,0.0403892882,0.3434166312,0.0855879858,0.0078646038,-0.2014046609,-0.1323599219,0.1542741805,-0.118224524,0.1847367287,0.3629119694,-0.0531053208,0.1500600725,-0.1300462335,-0.0676265731,-0.2308019549,0.0534002073,-0.1115095243,0.0901557282,-0.1316496879,0.4046879113,0.1569740921,0.0425376371,0.1446679085,-0.1151811853,0.1561760455,0.0373051763,-0.4768414795,-0.2667486668,0.0841662511,-0.1505966187,-0.1545930207,-0.1207949668,0.1211863682,-0.1792515069,-0.1328291744,0.3402727544,0.2215323895,-0.208968699,-0.0280376002,0.0667775273,0.1572731435,-0.1440431923,-0.3448980153,0.0656591803,-0.123423703,-0.3975611329,0.0672159195,0.3313922286,0.550983429,0.0814335942,-0.2558076084,-0.3254729509,-0.4844892621,0.0634808987,0.1591971517,0.071599789,-0.0215618368,0.7668600082,-0.1380043477,0.1425333768,-0.2015482634,-0.4433032274,-0.150910601,0.2117178291,-0.2316481024,0.0988812521,-0.0608816147,0.0209043287,-0.0537675731,0.0669301525,0.006623521,-0.0675481558,0.2549972832,-0.6412526369,0.0265079904,0.2047789544,0.1866111457,0.1211187914,0.1911760271,-0.0848483518,-0.1786593199,0.0570020229,-0.4990594983,-0.264444977,-0.0435095765,0.071315363,0.1389947832,0.2218963653,0.0635314956,0.151714012,0.5338413119,-0.0446067266,0.041408807,-0.3242377341,-0.3369706869,-0.3063197136,-0.3215333521,0.4540395737,0.0853606761,-0.0651732013,0.2063797265,-0.058066532,0.3713686466,0.4545201957,0.0005992663,0.0837388784,-0.1234111786,0.0138952779,0.2660521865,-0.2399116904,0.0215655714,0.4763100743,-0.4028788805,0.0788712651,0.3922178149,-0.0496553518,-0.2771501541,-0.0372722521,0.1895166337,0.3423955739,-0.0628310069,0.009341618,-0.2271394581,-0.0792751983,0.4815947115,-0.1569288224,0.3270474672,0.0703963637,-0.0276751965,-0.1134375632,0.2275328934,-0.3470682502,-0.0631333143,0.1020137817,0.2154045105,-0.1462944895,0.0686946362,-0.1687207073,-0.2463460565,0.0542072281,-0.1364753246,0.1516544819,0.2225850374,-0.0803485885,0.2349827886,-0.2266496718,-0.2233921885,0.0553607643,0.0752304047,0.3720036149,-0.2636727989,0.1632019877,0.0245809779,0.0021455172,-0.3904486895,-0.2759417295,-0.1010340154,0.2078496218,-0.0291660819,-0.1070364043,-0.0839110985,0.2642500401,0.4416565597,-0.2129094154,0.2406531572,0.1404104829,-0.1080904603,0.0494443811,-0.0524048693,0.1803079396,0.35804829,-0.1872611344,-0.1409421414,0.2008781582,0.134699285,-0.1484245509,0.3699091673,0.4953466356,0.090521425,0.2097523212,0.0449271016,0.078612186,0.1215231866,0.0240474232,-0.0164274201,0.0827938691,0.0056439694,0.1275957525,-0.1125174537,0.0920489132,-0.6476395726,0.0956622809,0.8060963154,0.1173804998,0.4062262475,0.0086445129,0.1291960925,0.0277588628,-0.154481113,-0.0285173319,0.4229156077,-0.0422861688,0.2465559989,-0.0392972976,-0.0626192093,-0.1413778216,0.0522282161,-0.1825522035,0.7166270614,0.2830475867,0.132720679,-0.0022643518,-0.0056484104,-0.2538818419,0.2736298442,0.450527519,-0.3334290981,0.3502019644,-0.3673314154,-0.0159904975,-0.2581938505,0.2661478817,-0.0938294679,-0.0245775245,-0.2472632974,0.1864033341,0.1503834277,0.1223219782,0.35891819,0.1457659602,0.1071996093,0.2317183763,0.2299328446,0.0111648263,-0.2177335769,-0.072553508,0.2922357321,-0.3358941078,0.0090650646,-0.1226599514,-0.5176730752,-0.351428628,-0.0638254508,0.0656745881,0.013568298,0.0029365411,0.2864256203,0.0785256252,-0.0120970001,-0.0619992651,0.1552059203,-0.4553328753,-0.1128283218,0.2476546913,-0.3855648041,0.1096370816,-0.1286381036,-0.2846000791,-0.2985165715,-0.1682437807,0.6335537434,0.036101602,0.4326667488,0.1627968699,0.1140319705,0.1608608812,0.1367166787,0.0956066176,-0.1563305557,-0.4536629319,0.1469080001,0.0463184677,-0.2639672458,-0.0813129917,0.0779389441,0.3259366453,0.1940153837,-0.301644206,-0.2473767698,0.0698908642,-0.0073892139,-0.0802942067,0.043606136,0.3106846511,-0.081747666,-0.0795649514,0.1049370244,0.1663469374,0.1308562309,0.2061449587,0.2597747743,0.215860039,0.2530972958,0.2540295124,0.535112977,0.2206321359,-0.0674460903,0.2258820832,-0.1522954851,-0.004480199,0.1564436704,-0.3594617546,0.1473370194,-0.2330273092,-0.1332301199,-0.0466546267,-0.441717416,-0.5093246102,-0.1664875448,-0.1738309413,-0.3591647446,-0.1715164185,0.2290109694,-0.2058067173,0.3189735711,-0.2008253485,-0.0663316622,-0.2482412905,0.020823529,-0.0461255014,0.2327274233,-0.0202390645,-0.0802518353,-0.413092345,-0.5657935143,-0.0441711657,0.2615533471,0.0113745593,0.4401904047,-0.1060471758,0.019314928,0.1748602092,-0.0069007189,0.8240171075,-0.2342891544,0.0243170429,0.2899026871,-0.5752906799,-0.0847982988,-0.0399576202,-0.2760018408,0.4216439724,0.5022407174,0.3049765825,-0.1920865029,-0.3170815706,0.2327643633,-0.1730203927,-0.1513353437,-0.0994562805,-0.2681988478,-0.1267526001,0.0031301854,0.0991752595,0.0423637964,0.41966483,-0.4223131835,-0.0110674258,-0.3594005704,0.1035603508,0.0992429331,0.0183773581,-0.0220204499,-0.0157828331,-0.2933521867,0.0666060001,0.4181612432,0.3456064463,0.1352541447,0.1237927079,-0.222049877,0.3866360784,-0.2193987668,0.3998539448,0.4414469898,-0.0689045191,0.2593995333,0.046429269,0.3088773489,0.0099545876,-0.2134131789,0.5013944507,0.0964831039,-0.3567266762,-0.4405709505,0.2417843342,0.2308721542,0.1443294287,0.0753162056,-0.3131642342,-0.0920787081,-0.0364313535,-0.0348871164,0.8628455997,0.1303196251,0.1319111735,0.2637454867,0.0648117885,0.7199362516,-0.4422337711,0.3917687237,-0.2568787336,-0.1663816869,-0.1640912145,-0.0817656144,0.3914418817,0.2064569592,-0.2381402701,0.3658711314,-0.1030779406,-0.1616385281,-0.158077091,0.1148330495,-0.2505121827,-0.0613890663,-0.149653703,-0.0496145748,0.0577678978,0.1704435498,-0.1470059156,-0.3725346625,0.3594648838,-0.0050592548,-0.2195948511,-0.0886685327,-0.3248027563,0.2875602543,0.4124962687,-0.4293585122,0.237770915,0.2076031417,0.3087032437,-0.2606231272,-0.1220987365,-0.2071407288,-0.2024772912,-0.0072195283,-0.2808032334,-0.1926895678,0.5347767472,-0.1014083549,-0.5287340879,0.2500193417,-0.0031347205,-0.0043283291,0.0241581351,0.0992155746,-0.4970460236,0.2518159449,-0.0458345078,0.1943402141,-0.167885825,-0.2178985178,0.0715451837,-0.0830308795,-0.6331080198,-0.0050315158,-0.0159770828,-0.406650275,0.188466534,0.3562635779,0.5661857724,0.4459779859,0.4622503519,0.3645583689,-0.0835475028,-0.3533835411,0.2155154347,0.3437524438,-0.3580479026,0.226923272,-0.0134585239,0.0545984097,0.0333130844,0.1038374528,0.0438436568,0.5941749811,-0.1677767783,-0.2888618112,0.1138362288,-0.1244143173,-0.2211531848,-0.1914723366,0.0318109803,0.4678594172,0.3696142137,0.1560432464,-0.2256106585,0.183211565,-0.6862777472,0.1564278305,-0.2976360619,-0.1875285208,0.0351730846,0.2281868011,0.011276192,0.0311042834,-0.0151206814,-0.1135714129,-0.1084555835,0.1914585084,0.0762739256,0.2077237517,0.219142586,0.0013007895,-0.2629858851,0.0861249492,0.1031461284,0.0263838526,-0.0127823241,0.3347818255,-0.1318498403,0.1470612586,-0.0516887121,0.0160815567,-0.1663547903,-0.0264431108,0.0684207007,0.2599076033,0.2681325674,0.1051005349,-0.2012492418,-0.0905272663,-0.1690200418,0.1310450733,-0.0031563214,-0.3379543722,-0.1537528336,-0.0610398687,-0.1287465096,0.0786517039,-0.2081065625,0.1642848402,-0.0082762223,0.075190343,-0.1720363051,0.0160495415,0.3984807432,0.0401428938,-0.0545231737,0.3129579723,0.1964588016,-0.014788745,-0.2255492657,0.0989417359,-0.1798563451,-0.5493915081,0.1987308264,0.0789996162,-0.172555238,0.1692458093,0.2340233773,0.1941285282,-0.0107985949,0.1986323446,0.1553979218,0.2505668998,0.0114463819,0.0647496507,0.1874623001,-0.3419452906,-0.1851764172,0.1027245522,-0.187814936,-0.0758213401,-0.2439704537,0.1061948091,-0.3426975012,-0.2329878658,-0.1344069839,0.1067933068,-0.3716104925,-0.07589636,-0.4822933078,-0.033825215,-0.2222220749,-0.0620703995,0.1131879613,0.0784955546,0.3637962639,0.2248342037,-0.1549200863,-0.3011435568,0.0974638313,0.1635616869,0.2713882327,-0.222735852,0.0118846055,0.1044878513,-0.0192578882,-0.2971428931,0.4542773366,0.3316576183,0.3410503566,-0.227893427,-0.1521065682,0.1208092198,-0.1401267946,0.0169089995,-0.0865086541,0.1232791767,0.2424122095,0.206556499,0.0007886984,-0.050825499,0.3772299886,0.1411721408,0.0152803967,-0.3947429955,0.4997736812,-0.3795516193,0.0404789969,-0.0975971743,0.3181056678,-0.5076952577,0.1147272587,0.6606998444,-0.203587383,0.2390411943,0.0054493127,0.0247680712,-0.1507310867,0.4882811606,0.0556453913,-0.0264799502,-0.2904711664,0.121670261,-0.1169741154,0.5847662687,-0.3455072045,0.2629430294,-0.0381880477,0.602992177,-0.0786876529,-0.0741019771,0.1273525506,0.1574851871,0.1948416084,0.3991425633,-0.4147590399,-0.2249621451,-0.3403022885,-0.1307871044,0.1633239239,-0.3058509529,0.1660936028,-0.1133297086,0.0163280759,-0.0395119525,0.0059385221,-0.2246301025,0.0200872887,0.2809771895,0.1872590929,0.2327943444,0.0099964524,-0.0424147658,0.0101231411,-0.1967514157,-0.2997451127,0.5487582088,0.0568799414,0.0997763723,-0.1939614415,0.2067076266,-0.1505220085,0.2600656152,-0.0193499308,0.0963500366,0.0100755356,0.2616032362,-0.0259610247,-0.0744594187,0.0825509652,0.0806281418,0.1278177947,-0.2672429383,-0.1666998714,-0.0958619118,0.3261485994,-0.3049176931,-0.3767558932,-0.4874207675,0.3538348377,0.0580240414,0.1417587101,-0.4789167345,-0.1494803429,0.3774831295,-0.0684738085,-0.3998480439,-0.0441302881,-0.0841486678,-0.1253637522,0.0760098025,-0.2112776041,-0.1244162843,-0.2668003738,0.1230953559,-0.3076861203]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2131","title":"When training with Multi-Node Multi-GPU the worker 2 has TypeError: 'NoneType' object","comments":"Sorry for the late reply. \r\nNow everything just works well XD","body":"\bversion: 1.5.0\r\nmet a very strange error, I am training large scale language model, and need train on 2 machines(workers).\r\nAnd sometimes I will get this error `TypeError: 'NoneType' object is not iterable`\r\nThis is traceback\r\n```\r\n\r\n71 | \u00a0 | Traceback (most recent call last):\r\n-- | -- | --\r\n72 | \u00a0 | File \"run_gpt.py\", line 316, in \r\n73 | \u00a0 | main()\r\n74 | \u00a0 | File \"run_gpt.py\", line 222, in main\r\n75 | \u00a0 | delimiter=\"\\t\", column_names=[\"input_ids\", \"attention_mask\", \"chinese_ref\"])\r\n76 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 747, in load_dataset\r\n77 | \u00a0 | use_auth_token=use_auth_token,\r\n78 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 513, in download_and_prepare\r\n79 | \u00a0 | self.download_post_processing_resources(dl_manager)\r\n80 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 673, in download_post_processing_resources\r\n81 | \u00a0 | for split in self.info.splits:\r\n82 | \u00a0 | TypeError: 'NoneType' object is not iterable\r\n83 | \u00a0 | WARNING:datasets.builder:Reusing dataset csv (\/usr\/local\/app\/.cache\/huggingface\/datasets\/csv\/default-1c257ebd48e225e7\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2)\r\n84 | \u00a0 | Traceback (most recent call last):\r\n85 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 193, in _run_module_as_main\r\n86 | \u00a0 | \"__main__\", mod_spec)\r\n87 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 85, in _run_code\r\n88 | \u00a0 | exec(code, run_globals)\r\n89 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 340, in \r\n90 | \u00a0 | main()\r\n91 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 326, in main\r\n92 | \u00a0 | sigkill_handler(signal.SIGTERM, None) # not coming back\r\n93 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 301, in sigkill_handler\r\n94 | \u00a0 | raise subprocess.CalledProcessError(returncode=last_return_code, cmd=cmd)\r\n\r\n```\r\nOn worker 1 it loads the dataset well, however on worker 2 will get this error. \r\nAnd I will meet this error from time to time, sometimes it just goes well.","comment_length":11,"text":"When training with Multi-Node Multi-GPU the worker 2 has TypeError: 'NoneType' object\n\bversion: 1.5.0\r\nmet a very strange error, I am training large scale language model, and need train on 2 machines(workers).\r\nAnd sometimes I will get this error `TypeError: 'NoneType' object is not iterable`\r\nThis is traceback\r\n```\r\n\r\n71 | \u00a0 | Traceback (most recent call last):\r\n-- | -- | --\r\n72 | \u00a0 | File \"run_gpt.py\", line 316, in \r\n73 | \u00a0 | main()\r\n74 | \u00a0 | File \"run_gpt.py\", line 222, in main\r\n75 | \u00a0 | delimiter=\"\\t\", column_names=[\"input_ids\", \"attention_mask\", \"chinese_ref\"])\r\n76 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 747, in load_dataset\r\n77 | \u00a0 | use_auth_token=use_auth_token,\r\n78 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 513, in download_and_prepare\r\n79 | \u00a0 | self.download_post_processing_resources(dl_manager)\r\n80 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 673, in download_post_processing_resources\r\n81 | \u00a0 | for split in self.info.splits:\r\n82 | \u00a0 | TypeError: 'NoneType' object is not iterable\r\n83 | \u00a0 | WARNING:datasets.builder:Reusing dataset csv (\/usr\/local\/app\/.cache\/huggingface\/datasets\/csv\/default-1c257ebd48e225e7\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2)\r\n84 | \u00a0 | Traceback (most recent call last):\r\n85 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 193, in _run_module_as_main\r\n86 | \u00a0 | \"__main__\", mod_spec)\r\n87 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/runpy.py\", line 85, in _run_code\r\n88 | \u00a0 | exec(code, run_globals)\r\n89 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 340, in \r\n90 | \u00a0 | main()\r\n91 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 326, in main\r\n92 | \u00a0 | sigkill_handler(signal.SIGTERM, None) # not coming back\r\n93 | \u00a0 | File \"\/data\/miniconda3\/lib\/python3.7\/site-packages\/torch\/distributed\/launch.py\", line 301, in sigkill_handler\r\n94 | \u00a0 | raise subprocess.CalledProcessError(returncode=last_return_code, cmd=cmd)\r\n\r\n```\r\nOn worker 1 it loads the dataset well, however on worker 2 will get this error. \r\nAnd I will meet this error from time to time, sometimes it just goes well.\nSorry for the late reply. \r\nNow everything just works well XD","embeddings":[-0.1806102991,-0.4468090534,0.0129236523,0.5495145917,0.1124854609,-0.0091340337,0.5815147161,0.2967011034,0.1129247844,0.2367109209,0.3380283117,0.0173764732,-0.1162154898,0.1675734371,-0.042410925,-0.2821591198,-0.1764215678,0.0434676446,-0.227805078,-0.214592278,-0.3433728218,0.1396662146,-0.1313170046,0.1889922619,-0.5063354373,-0.4234542847,-0.1063818559,-0.0327261165,0.2665303946,-0.2152226418,0.2046148032,-0.3352004588,0.1856167316,0.8994958401,-0.0001186402,0.148360461,0.1208232865,0.0624540038,-0.144217357,-0.4792484939,0.083634451,-0.3793367445,0.0438947678,-0.372520566,-0.1640225798,-0.1153089702,0.2508283257,0.0683696643,0.038129054,0.2097055912,0.1652571112,0.3379758,0.1185292006,-0.0910865888,-0.1230920553,-0.078175433,-0.124581486,-0.0570696816,-0.1463324875,0.2151985615,0.1408524811,0.3001521528,0.0600437447,0.0522123091,-0.0947405323,-0.0429644659,0.5682382584,-0.4400096834,0.0403892882,0.3434166312,0.0855879858,0.0078646038,-0.2014046609,-0.1323599219,0.1542741805,-0.118224524,0.1847367287,0.3629119694,-0.0531053208,0.1500600725,-0.1300462335,-0.0676265731,-0.2308019549,0.0534002073,-0.1115095243,0.0901557282,-0.1316496879,0.4046879113,0.1569740921,0.0425376371,0.1446679085,-0.1151811853,0.1561760455,0.0373051763,-0.4768414795,-0.2667486668,0.0841662511,-0.1505966187,-0.1545930207,-0.1207949668,0.1211863682,-0.1792515069,-0.1328291744,0.3402727544,0.2215323895,-0.208968699,-0.0280376002,0.0667775273,0.1572731435,-0.1440431923,-0.3448980153,0.0656591803,-0.123423703,-0.3975611329,0.0672159195,0.3313922286,0.550983429,0.0814335942,-0.2558076084,-0.3254729509,-0.4844892621,0.0634808987,0.1591971517,0.071599789,-0.0215618368,0.7668600082,-0.1380043477,0.1425333768,-0.2015482634,-0.4433032274,-0.150910601,0.2117178291,-0.2316481024,0.0988812521,-0.0608816147,0.0209043287,-0.0537675731,0.0669301525,0.006623521,-0.0675481558,0.2549972832,-0.6412526369,0.0265079904,0.2047789544,0.1866111457,0.1211187914,0.1911760271,-0.0848483518,-0.1786593199,0.0570020229,-0.4990594983,-0.264444977,-0.0435095765,0.071315363,0.1389947832,0.2218963653,0.0635314956,0.151714012,0.5338413119,-0.0446067266,0.041408807,-0.3242377341,-0.3369706869,-0.3063197136,-0.3215333521,0.4540395737,0.0853606761,-0.0651732013,0.2063797265,-0.058066532,0.3713686466,0.4545201957,0.0005992663,0.0837388784,-0.1234111786,0.0138952779,0.2660521865,-0.2399116904,0.0215655714,0.4763100743,-0.4028788805,0.0788712651,0.3922178149,-0.0496553518,-0.2771501541,-0.0372722521,0.1895166337,0.3423955739,-0.0628310069,0.009341618,-0.2271394581,-0.0792751983,0.4815947115,-0.1569288224,0.3270474672,0.0703963637,-0.0276751965,-0.1134375632,0.2275328934,-0.3470682502,-0.0631333143,0.1020137817,0.2154045105,-0.1462944895,0.0686946362,-0.1687207073,-0.2463460565,0.0542072281,-0.1364753246,0.1516544819,0.2225850374,-0.0803485885,0.2349827886,-0.2266496718,-0.2233921885,0.0553607643,0.0752304047,0.3720036149,-0.2636727989,0.1632019877,0.0245809779,0.0021455172,-0.3904486895,-0.2759417295,-0.1010340154,0.2078496218,-0.0291660819,-0.1070364043,-0.0839110985,0.2642500401,0.4416565597,-0.2129094154,0.2406531572,0.1404104829,-0.1080904603,0.0494443811,-0.0524048693,0.1803079396,0.35804829,-0.1872611344,-0.1409421414,0.2008781582,0.134699285,-0.1484245509,0.3699091673,0.4953466356,0.090521425,0.2097523212,0.0449271016,0.078612186,0.1215231866,0.0240474232,-0.0164274201,0.0827938691,0.0056439694,0.1275957525,-0.1125174537,0.0920489132,-0.6476395726,0.0956622809,0.8060963154,0.1173804998,0.4062262475,0.0086445129,0.1291960925,0.0277588628,-0.154481113,-0.0285173319,0.4229156077,-0.0422861688,0.2465559989,-0.0392972976,-0.0626192093,-0.1413778216,0.0522282161,-0.1825522035,0.7166270614,0.2830475867,0.132720679,-0.0022643518,-0.0056484104,-0.2538818419,0.2736298442,0.450527519,-0.3334290981,0.3502019644,-0.3673314154,-0.0159904975,-0.2581938505,0.2661478817,-0.0938294679,-0.0245775245,-0.2472632974,0.1864033341,0.1503834277,0.1223219782,0.35891819,0.1457659602,0.1071996093,0.2317183763,0.2299328446,0.0111648263,-0.2177335769,-0.072553508,0.2922357321,-0.3358941078,0.0090650646,-0.1226599514,-0.5176730752,-0.351428628,-0.0638254508,0.0656745881,0.013568298,0.0029365411,0.2864256203,0.0785256252,-0.0120970001,-0.0619992651,0.1552059203,-0.4553328753,-0.1128283218,0.2476546913,-0.3855648041,0.1096370816,-0.1286381036,-0.2846000791,-0.2985165715,-0.1682437807,0.6335537434,0.036101602,0.4326667488,0.1627968699,0.1140319705,0.1608608812,0.1367166787,0.0956066176,-0.1563305557,-0.4536629319,0.1469080001,0.0463184677,-0.2639672458,-0.0813129917,0.0779389441,0.3259366453,0.1940153837,-0.301644206,-0.2473767698,0.0698908642,-0.0073892139,-0.0802942067,0.043606136,0.3106846511,-0.081747666,-0.0795649514,0.1049370244,0.1663469374,0.1308562309,0.2061449587,0.2597747743,0.215860039,0.2530972958,0.2540295124,0.535112977,0.2206321359,-0.0674460903,0.2258820832,-0.1522954851,-0.004480199,0.1564436704,-0.3594617546,0.1473370194,-0.2330273092,-0.1332301199,-0.0466546267,-0.441717416,-0.5093246102,-0.1664875448,-0.1738309413,-0.3591647446,-0.1715164185,0.2290109694,-0.2058067173,0.3189735711,-0.2008253485,-0.0663316622,-0.2482412905,0.020823529,-0.0461255014,0.2327274233,-0.0202390645,-0.0802518353,-0.413092345,-0.5657935143,-0.0441711657,0.2615533471,0.0113745593,0.4401904047,-0.1060471758,0.019314928,0.1748602092,-0.0069007189,0.8240171075,-0.2342891544,0.0243170429,0.2899026871,-0.5752906799,-0.0847982988,-0.0399576202,-0.2760018408,0.4216439724,0.5022407174,0.3049765825,-0.1920865029,-0.3170815706,0.2327643633,-0.1730203927,-0.1513353437,-0.0994562805,-0.2681988478,-0.1267526001,0.0031301854,0.0991752595,0.0423637964,0.41966483,-0.4223131835,-0.0110674258,-0.3594005704,0.1035603508,0.0992429331,0.0183773581,-0.0220204499,-0.0157828331,-0.2933521867,0.0666060001,0.4181612432,0.3456064463,0.1352541447,0.1237927079,-0.222049877,0.3866360784,-0.2193987668,0.3998539448,0.4414469898,-0.0689045191,0.2593995333,0.046429269,0.3088773489,0.0099545876,-0.2134131789,0.5013944507,0.0964831039,-0.3567266762,-0.4405709505,0.2417843342,0.2308721542,0.1443294287,0.0753162056,-0.3131642342,-0.0920787081,-0.0364313535,-0.0348871164,0.8628455997,0.1303196251,0.1319111735,0.2637454867,0.0648117885,0.7199362516,-0.4422337711,0.3917687237,-0.2568787336,-0.1663816869,-0.1640912145,-0.0817656144,0.3914418817,0.2064569592,-0.2381402701,0.3658711314,-0.1030779406,-0.1616385281,-0.158077091,0.1148330495,-0.2505121827,-0.0613890663,-0.149653703,-0.0496145748,0.0577678978,0.1704435498,-0.1470059156,-0.3725346625,0.3594648838,-0.0050592548,-0.2195948511,-0.0886685327,-0.3248027563,0.2875602543,0.4124962687,-0.4293585122,0.237770915,0.2076031417,0.3087032437,-0.2606231272,-0.1220987365,-0.2071407288,-0.2024772912,-0.0072195283,-0.2808032334,-0.1926895678,0.5347767472,-0.1014083549,-0.5287340879,0.2500193417,-0.0031347205,-0.0043283291,0.0241581351,0.0992155746,-0.4970460236,0.2518159449,-0.0458345078,0.1943402141,-0.167885825,-0.2178985178,0.0715451837,-0.0830308795,-0.6331080198,-0.0050315158,-0.0159770828,-0.406650275,0.188466534,0.3562635779,0.5661857724,0.4459779859,0.4622503519,0.3645583689,-0.0835475028,-0.3533835411,0.2155154347,0.3437524438,-0.3580479026,0.226923272,-0.0134585239,0.0545984097,0.0333130844,0.1038374528,0.0438436568,0.5941749811,-0.1677767783,-0.2888618112,0.1138362288,-0.1244143173,-0.2211531848,-0.1914723366,0.0318109803,0.4678594172,0.3696142137,0.1560432464,-0.2256106585,0.183211565,-0.6862777472,0.1564278305,-0.2976360619,-0.1875285208,0.0351730846,0.2281868011,0.011276192,0.0311042834,-0.0151206814,-0.1135714129,-0.1084555835,0.1914585084,0.0762739256,0.2077237517,0.219142586,0.0013007895,-0.2629858851,0.0861249492,0.1031461284,0.0263838526,-0.0127823241,0.3347818255,-0.1318498403,0.1470612586,-0.0516887121,0.0160815567,-0.1663547903,-0.0264431108,0.0684207007,0.2599076033,0.2681325674,0.1051005349,-0.2012492418,-0.0905272663,-0.1690200418,0.1310450733,-0.0031563214,-0.3379543722,-0.1537528336,-0.0610398687,-0.1287465096,0.0786517039,-0.2081065625,0.1642848402,-0.0082762223,0.075190343,-0.1720363051,0.0160495415,0.3984807432,0.0401428938,-0.0545231737,0.3129579723,0.1964588016,-0.014788745,-0.2255492657,0.0989417359,-0.1798563451,-0.5493915081,0.1987308264,0.0789996162,-0.172555238,0.1692458093,0.2340233773,0.1941285282,-0.0107985949,0.1986323446,0.1553979218,0.2505668998,0.0114463819,0.0647496507,0.1874623001,-0.3419452906,-0.1851764172,0.1027245522,-0.187814936,-0.0758213401,-0.2439704537,0.1061948091,-0.3426975012,-0.2329878658,-0.1344069839,0.1067933068,-0.3716104925,-0.07589636,-0.4822933078,-0.033825215,-0.2222220749,-0.0620703995,0.1131879613,0.0784955546,0.3637962639,0.2248342037,-0.1549200863,-0.3011435568,0.0974638313,0.1635616869,0.2713882327,-0.222735852,0.0118846055,0.1044878513,-0.0192578882,-0.2971428931,0.4542773366,0.3316576183,0.3410503566,-0.227893427,-0.1521065682,0.1208092198,-0.1401267946,0.0169089995,-0.0865086541,0.1232791767,0.2424122095,0.206556499,0.0007886984,-0.050825499,0.3772299886,0.1411721408,0.0152803967,-0.3947429955,0.4997736812,-0.3795516193,0.0404789969,-0.0975971743,0.3181056678,-0.5076952577,0.1147272587,0.6606998444,-0.203587383,0.2390411943,0.0054493127,0.0247680712,-0.1507310867,0.4882811606,0.0556453913,-0.0264799502,-0.2904711664,0.121670261,-0.1169741154,0.5847662687,-0.3455072045,0.2629430294,-0.0381880477,0.602992177,-0.0786876529,-0.0741019771,0.1273525506,0.1574851871,0.1948416084,0.3991425633,-0.4147590399,-0.2249621451,-0.3403022885,-0.1307871044,0.1633239239,-0.3058509529,0.1660936028,-0.1133297086,0.0163280759,-0.0395119525,0.0059385221,-0.2246301025,0.0200872887,0.2809771895,0.1872590929,0.2327943444,0.0099964524,-0.0424147658,0.0101231411,-0.1967514157,-0.2997451127,0.5487582088,0.0568799414,0.0997763723,-0.1939614415,0.2067076266,-0.1505220085,0.2600656152,-0.0193499308,0.0963500366,0.0100755356,0.2616032362,-0.0259610247,-0.0744594187,0.0825509652,0.0806281418,0.1278177947,-0.2672429383,-0.1666998714,-0.0958619118,0.3261485994,-0.3049176931,-0.3767558932,-0.4874207675,0.3538348377,0.0580240414,0.1417587101,-0.4789167345,-0.1494803429,0.3774831295,-0.0684738085,-0.3998480439,-0.0441302881,-0.0841486678,-0.1253637522,0.0760098025,-0.2112776041,-0.1244162843,-0.2668003738,0.1230953559,-0.3076861203]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2130","title":"wikiann dataset is missing columns ","comments":"Here please find TFDS format of this dataset: https:\/\/www.tensorflow.org\/datasets\/catalog\/wikiann\r\nwhere there is a span column, this is really necessary to be able to use the data, and I appreciate your help @lhoestq ","body":"Hi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq ","comment_length":32,"text":"wikiann dataset is missing columns \nHi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq \nHere please find TFDS format of this dataset: https:\/\/www.tensorflow.org\/datasets\/catalog\/wikiann\r\nwhere there is a span column, this is really necessary to be able to use the data, and I appreciate your help @lhoestq ","embeddings":[0.0038926781,-0.4387932718,-0.0947454646,0.2763384283,0.3197283745,0.1833388954,0.3208168447,0.0807619765,0.0531267151,0.2570092082,0.0109932544,-0.2380872667,0.0703101009,0.4205832183,0.3480723798,-0.8203633428,0.1133004576,0.3876077533,-0.2301765978,-0.1201595664,-0.2421356142,0.3445684314,-0.4137241244,-0.1726709008,-0.2490336597,-0.1022257879,-0.1513452083,-0.0896822661,-0.0339641571,-0.0308071841,0.3040222526,-0.2934643924,0.3014791906,0.1141969711,-0.0001120765,0.1079274938,0.013424987,-0.1143514216,0.076958634,0.2684707046,-0.0305692889,-0.3217623234,0.0123331919,-0.396816045,-0.3583127856,-0.1827049255,-0.1768914312,0.1160641909,0.1129548922,0.217622906,0.1882001907,0.7334356904,0.3261469007,-0.0215925407,0.073895067,-0.2582473457,-0.243468374,-0.3627047837,0.2368402481,-0.0988853648,0.2397650182,0.3428517282,-0.0141156651,-0.1902943254,0.2873823047,0.152955547,-0.2721420228,-0.5023332238,0.292786628,0.7253635526,0.8114988804,0.0970162377,0.1451937556,-0.1972335875,0.0165993534,0.0472852811,0.0804864988,0.1341166198,-0.0196679346,0.2952947617,-0.0729851499,-0.0781829357,-0.2694770694,0.2836079597,-0.3395297527,0.3484340906,-0.0432369746,-0.0545714423,-0.1588187963,-0.0111133317,-0.0922963843,0.1739283949,0.0884341225,0.1710797697,-0.3939895928,-0.2244472504,-0.08043769,0.2654077113,0.1304720789,-0.0493794307,-0.2567351758,-0.26376158,0.0772433802,-0.0034747294,-0.0845331997,0.086401619,-0.0976359025,0.1400863379,0.033922147,0.1944786459,0.1251031011,-0.1460148841,-0.1888661236,0.0409027934,-0.1001603454,-0.2774039209,-0.1225066632,0.0023716278,0.1778179407,0.0947623923,0.1200572774,0.0799391717,0.0292139482,0.2910702825,-0.0344533809,0.5733726025,-0.0184937119,0.0801496059,0.072881788,-0.1036665514,-0.3151019812,0.1644501239,0.1746065021,0.066531308,0.145810321,-0.2616475821,0.449447155,0.0023808659,-0.1679136455,0.2190256417,-0.0388755314,0.0814908966,-0.0857503265,0.5386383533,0.0207409132,0.0401842035,0.1553890109,0.0984406471,-0.0684101135,0.0521576405,-0.3235920966,-0.0773139,-0.3769297004,0.1812785715,-0.002946808,-0.3248754144,-0.1078693792,0.1064671949,0.0755224377,-0.2321736515,0.2516083121,-0.0360109285,-0.122053422,-0.1593221426,0.4439876378,-0.0531386323,-0.3692472279,-0.2977433801,0.2755208611,-0.1768388301,0.2126059681,0.38605088,0.1058009863,0.1704327315,-0.0902446136,0.4356479943,0.2490220815,-0.2292732447,-0.2811501324,-0.123422116,0.188975662,0.1803060174,0.0494902506,-0.070163317,0.3523239195,0.1053998172,-0.071211867,0.2468882352,-0.154971078,0.0208387226,-0.0559573509,-0.2708706558,0.1583362967,0.1909640282,0.3186049461,0.1195719391,-0.2315596342,0.1412235498,0.2170293033,-0.3112365007,0.094575949,0.3153347075,0.1052559242,0.2595463097,0.1662837714,-0.1629929245,-0.7005294561,0.1809137613,0.1923979521,0.1310077459,0.2885630131,-0.1591424048,-0.088489674,-0.1614854336,-0.1323769093,-0.1435146034,0.1459515691,-0.1655987799,0.0570097491,0.2606689632,-0.2820781767,-0.2286586314,-0.3409322798,0.1114449874,-0.1707026064,0.2989851832,-0.0085171163,0.1448131055,0.0420191735,0.353082478,0.0917257145,0.0065469411,0.2474441826,0.0339430049,-0.3063539863,0.1176372319,-0.1642912924,0.3263201118,0.2306770533,-0.2691120207,-0.2675723433,-0.1782523245,-0.0008584204,-0.3207311034,-0.0369398966,0.062526159,0.0201673154,0.2263794392,0.0898453593,0.3009009361,0.096681878,-0.2566103339,0.0848557279,-0.2246658951,0.1342727691,-0.5700149536,0.2383306324,-0.1081248894,-0.5302510858,0.1398995519,0.5346270204,-0.1347516924,0.2633163929,-0.0109232608,-0.3155109286,0.1827801615,0.1891463548,-0.0476073734,0.1203833595,0.195845291,0.2573873103,-0.018990973,0.0233980305,-0.2002914399,0.1523865163,-0.130906567,0.243817091,0.2309399247,-0.0889771283,0.1434019208,-0.1445164531,0.3678418398,-0.1947112978,0.3326030076,-0.1929797232,-0.2458497584,-0.2023401856,-0.2728340924,0.1774611026,0.0141584771,-0.0482394844,-0.2429255396,-0.0138839437,-0.0451257378,0.077334255,0.1221307963,-0.1163477823,0.275680989,-0.0069521661,0.1026896685,-0.4535602033,-0.0234313942,-0.2051280737,0.1776485443,0.005037473,-0.1009427905,0.1653717011,-0.282486707,-0.0113938898,-0.4114501774,-0.398784101,0.1531444341,-0.2570134997,-0.1782176644,0.2222473323,0.7760813832,-0.1808810532,-0.1068493277,0.2203474343,-0.3107444942,-0.1329945922,0.0850411877,-0.2469329685,-0.0271146875,-0.1479097009,-0.0456992537,-0.2469081581,-0.0544394366,0.0063213194,-0.1434307694,-0.033137545,0.0108501948,-0.0731811896,0.2382520288,0.0096790949,-0.0223693885,-0.1764644086,-0.3380376995,0.4421738684,-0.1514675021,-0.4279817939,0.1477312446,-0.1339934021,0.4414248168,0.1018832847,-0.4725303352,-0.2244140953,-0.0765634477,0.1601535976,-0.0337576568,0.0460723266,0.5994674563,-0.2907123268,0.1441126019,-0.3438046277,0.2465127707,-0.1237938702,-0.2274716794,0.3126408756,-0.1842316836,0.2806168795,0.1196215823,0.4372527897,0.1340361536,0.2335826755,0.1642498672,-0.3273261786,0.3505504429,-0.0554360561,-0.1419948936,0.3357430398,0.0569067821,0.3127399683,0.349514097,0.0804627016,-0.1397723258,-0.1895394623,-0.1385489851,-0.3253615201,-0.0360933952,0.0056714215,0.3305633068,0.3602113724,-0.0667295083,0.1412798613,-0.0958656818,-0.2964458168,0.1164380237,0.4351152182,0.1846984178,0.0768583268,-0.2016157955,0.0425064228,-0.535374403,-0.1199572161,-0.0783547536,0.1430696845,-0.1610945612,0.2194125801,0.0153736454,0.1239083931,0.5483193994,-0.4586716294,-0.2511317432,-0.1514674574,0.0694242865,-0.2658082843,0.1500856578,-0.175961569,-0.6598418355,0.0633214191,0.5552468896,-0.1945563108,-0.0761140287,0.4672186673,0.1109074131,-0.2404938638,-0.3276116252,-0.2175736278,-0.5198202729,-0.0598056875,-0.2966896296,0.2284470052,0.1514730901,0.0812747777,0.018642867,-0.1038896516,0.072267808,-0.0559470505,0.1264347583,0.1970183253,0.2099636942,-0.1462200433,0.3026448786,0.1411452144,0.0228797663,0.1872916073,-0.1113172024,-0.266818136,-0.0893916711,-0.1148821563,0.3527824879,-0.0293570515,-0.0428982005,0.1596740484,0.1022230536,0.1611078531,-0.1556324661,0.1408911347,0.4840656519,0.0306437165,-0.3259811103,-0.21976237,0.3390568197,0.0310100745,0.2772326767,0.3194664121,0.6091133952,-0.2198730111,-0.0713709816,0.0882646143,0.6027920842,-0.080849722,0.2150681615,0.6387683153,0.1078479961,0.5061591864,-0.1738476455,0.0813757181,-0.1399829388,-0.2035210282,-0.2088134885,0.1178931817,0.1535357982,-0.1363838762,-0.1380755156,0.2302026004,0.1945407987,0.6590548754,0.2112264782,0.0206192303,0.1981081516,-0.3410155773,0.0037913555,0.0509998053,-0.1828629673,0.2289459258,-0.1815903932,0.0274271704,0.1095731258,-0.0521215722,-0.3887438476,-0.0467693284,-0.3028003573,0.0837234855,0.2496719062,-0.4280476272,0.0260982309,0.5828906298,0.2919094265,0.0749523342,-0.4588287771,0.339194417,-0.2899720073,-0.1697791815,-0.1052600592,-0.0187036786,0.1468977481,-0.3607316017,-0.2569969594,0.08878389,-0.1979540735,0.1123604178,-0.5468319654,-0.1329645813,-0.2318889499,-0.1394215822,-0.2019579411,0.028669199,0.2051310986,-0.1704581529,0.1036361977,-0.2340830266,0.0311709996,0.0023115098,-0.0145664951,0.1151164174,-0.0465862937,-0.0088065648,0.1513922513,-0.1076866686,0.5726299882,0.2279133946,-0.201962322,-0.209669441,0.158648327,0.3235658705,-0.1113513783,-0.2232872695,0.3407527804,0.0636782125,-0.2243641615,0.4084559679,0.0545293093,-0.3415661752,-0.0959716141,-0.2599239349,-0.3298791945,0.2840153575,-0.0095749581,0.0850335434,-0.0457578935,0.3730965555,0.0390129276,-0.0064162584,-0.2616870999,0.0534339435,-0.15295057,-0.0701642334,0.1162127405,-0.0758844391,0.2101151496,-0.0620270036,0.0646236688,0.0621431209,0.1470191181,-0.1799550951,-0.1612076908,0.0971434712,0.1325246543,0.0819813311,0.1157548055,0.0776537359,-0.3033305109,-0.037734177,0.2392124683,-0.0650349408,0.1242353916,0.4137268364,0.2629142404,0.1006624922,-0.4023756981,0.0708326846,0.0459402949,0.0502923876,-0.0766862258,-0.0175247416,0.3782873154,0.0034457811,0.3644511104,0.1347179264,-0.0263651218,-0.0105212135,0.4533822834,-0.0718136802,0.1905425042,-0.0129271541,0.0046088728,-0.1390444189,-0.150426954,0.213917762,0.1837683916,0.1791432053,-0.3440064192,-0.0165623967,0.6128945351,0.2153338939,0.0547349378,0.0826410577,0.0124051953,-0.2125090659,-0.0821576118,-0.0622269399,0.3721988797,-0.24949494,0.1169542596,0.2945453823,0.1688923389,-0.0569904372,-0.1070456281,0.0550780557,-0.1166194752,0.0124534341,0.1552196741,0.1546727419,-0.3471730351,0.089611575,0.3072935343,-0.2104275376,-0.1748601645,-0.1847704202,-0.1456419826,0.0071286722,-0.2843002975,0.1226411164,0.2596562207,-0.1760370731,-0.1146961302,0.2897744477,-0.4225583673,-0.0365368836,-0.5136740208,-0.0949566737,0.4643177688,-0.0116002187,0.2184687853,-0.070120573,-0.2975370288,0.0358342603,-0.1305198073,0.0632900298,-0.130907774,0.334638983,0.2443826497,0.0962918028,0.1649875194,0.5131886601,-0.0827161744,-0.065314278,0.1417787969,0.1540222317,-0.130161047,0.203856945,0.0423884764,-0.0868407339,-0.2127304524,0.1242758334,0.1656407416,0.3661209345,0.0714524537,0.2031335533,0.0916304588,-0.1607858539,0.1727110893,-0.3102754056,0.062736772,-0.2525841892,-0.431489408,-0.4790424109,-0.3664103448,-0.3255728483,0.0700912327,-0.3667893112,0.4393817186,0.3142423034,0.1522528827,0.0496208407,0.0120735327,0.0408183001,-0.0565607734,0.5336560011,0.375323534,0.0652272478,-0.0631979108,-0.5100795031,-0.4604103863,0.1406422704,-0.3310749233,-0.280960232,0.3991977274,0.0121899797,0.2246186882,-0.1120978519,0.0712404326,0.1818882972,-0.0611512512,-0.0849650279,-0.3626081645,-0.3241222501,-0.2102888077,0.1575199366,-0.0781006962,0.0994565189,0.3338730931,0.233671248,0.0444557928,-0.0549853109,0.5272193551,-0.2992135584,-0.2561642826,-0.2594646215,0.1792296618,0.4275265038,-0.201232329,0.1252470464,0.0064845867,-0.0552079193,-0.2133355141,0.2091842294,-0.0420503654,0.2689840496,-0.0863775983,-0.3895616531,-0.0416735038,0.0830232427,0.2297479212,0.1148254797,-0.2521401942,0.1560745239,0.2910846174,-0.1661329567,0.5159576535,0.2352435291,-0.3428997695,-0.1456560045,-0.1788382381,-0.4296723306,0.3856591284,-0.3516278863,-0.070929341,-0.1147501767,-0.3248476386,0.1656797528,0.1717690825,-0.6097331643,-0.0852120817,0.2544306517,-0.0300879013,-0.1170116514,0.2773386538,0.200366661,0.136727646,0.0131960437,-0.1779351234,0.0776455775,-0.1768998802,-0.1311736554,-0.142207399]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2130","title":"wikiann dataset is missing columns ","comments":"Hi !\r\nApparently you can get the spans from the NER tags using `tags_to_spans` defined here:\r\n\r\nhttps:\/\/github.com\/tensorflow\/datasets\/blob\/c7096bd38e86ed240b8b2c11ecab9893715a7d55\/tensorflow_datasets\/text\/wikiann\/wikiann.py#L81-L126\r\n\r\nIt would be nice to include the `spans` field in this dataset as in TFDS. This could be a good first issue for new contributors !\r\n\r\nThe objective is to use `tags_to_spans` in the `_generate_examples` method [here](https:\/\/github.com\/huggingface\/nlp\/blob\/c98e4b8f23e3770c401c6d9326e243e1ffd599ec\/datasets\/wikiann\/wikiann.py#L292-L316) to create he `spans` for each example.","body":"Hi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq ","comment_length":61,"text":"wikiann dataset is missing columns \nHi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq \nHi !\r\nApparently you can get the spans from the NER tags using `tags_to_spans` defined here:\r\n\r\nhttps:\/\/github.com\/tensorflow\/datasets\/blob\/c7096bd38e86ed240b8b2c11ecab9893715a7d55\/tensorflow_datasets\/text\/wikiann\/wikiann.py#L81-L126\r\n\r\nIt would be nice to include the `spans` field in this dataset as in TFDS. This could be a good first issue for new contributors !\r\n\r\nThe objective is to use `tags_to_spans` in the `_generate_examples` method [here](https:\/\/github.com\/huggingface\/nlp\/blob\/c98e4b8f23e3770c401c6d9326e243e1ffd599ec\/datasets\/wikiann\/wikiann.py#L292-L316) to create he `spans` for each example.","embeddings":[0.0096025355,-0.3400487602,-0.0416718014,0.2219899446,0.2957082987,0.1506409198,0.3469608426,0.0389459059,0.0920513347,0.2753715515,0.0485684238,-0.045065973,-0.0080623208,0.3632449508,0.3476263881,-0.7035677433,0.0311781894,0.2775577009,-0.1697649062,0.0125467908,-0.2386454344,0.4350678325,-0.3131059706,-0.1572514474,-0.4427957535,-0.0176231563,-0.2585735023,-0.0887654722,-0.0000620789,-0.2098566294,0.3393801153,-0.2286444157,0.2572216094,-0.1111909077,-0.0001127702,0.0135668544,0.0615054667,-0.0787709877,0.0453730971,-0.0841397271,0.0922175422,-0.2578555644,-0.0348374099,-0.2561403215,-0.3284046948,-0.0753948241,-0.1517588496,0.0951740965,-0.0247506388,0.2180269063,0.1504639238,0.7661395073,0.1667824239,-0.0779862031,0.3054935932,-0.1118118763,-0.3342805505,-0.2091901451,0.1571556777,-0.1502291113,0.0648353174,0.4155132771,0.1755719334,-0.210781768,0.382522136,0.3127261996,-0.3683208525,-0.4562546313,0.0013717967,0.7677554488,0.4328103364,0.0056352643,-0.0264454279,-0.4336354136,0.01052956,-0.0954040512,0.0216443855,0.1007573307,-0.1240713,0.2177965641,-0.0346526355,-0.0619390868,-0.2247723192,0.1798864007,-0.2414661199,0.2327037901,0.0405296348,-0.0485336147,0.0854124948,-0.0204809867,-0.2247584909,0.1801456213,0.1715926826,0.1254052967,-0.429428637,-0.0337392278,-0.0181058794,0.3602000475,0.2100777924,0.0503829122,-0.0640601814,-0.2628415525,0.056887351,0.1270198673,-0.1555838138,0.0523052365,0.0083033936,0.1369297951,-0.0462952778,0.2109606564,0.0836109295,-0.1855955273,-0.0599287488,0.0971124768,-0.3002229333,-0.3053974211,-0.1432881206,0.0434895977,0.0042301407,0.1945303231,0.0694304854,0.1005565748,0.0134357084,0.2938938439,-0.0275055505,0.4606292546,-0.0773592144,0.2148902118,0.0255656186,-0.0853917748,-0.3643148839,0.1905200779,0.0240063369,0.1809857786,0.2539698482,-0.3256501555,0.471021086,-0.0390254408,-0.0434678458,0.1570568681,-0.052610606,0.0993080512,-0.1652287543,0.4395032525,-0.0349980034,0.0834425688,0.1697856337,-0.1075861007,-0.0663325265,-0.0548300222,-0.2166231573,-0.0040973453,-0.2454872876,0.1436573416,-0.1251898706,-0.2313668728,-0.0372504257,0.2354412079,0.1080498323,-0.2084845752,0.0712388083,0.0248907413,-0.3170247674,-0.1683148593,0.5056077838,0.0399223454,-0.3781881928,-0.3986588418,0.2939234674,-0.1754312366,0.1552590728,0.2094493359,0.0846953318,0.1171685159,-0.1244228855,0.6250929236,0.1958319098,-0.2356544137,-0.1009486169,-0.1965232342,0.1829894483,0.2725662589,0.0235283356,-0.0120192813,0.4229515195,0.1425526589,-0.04571734,0.3367981911,-0.2082446963,-0.0657014027,-0.1000476927,-0.1101307869,0.2190688848,0.1781906635,0.184169203,0.1768701673,-0.2432332784,0.2072175443,0.1569544971,-0.3025969267,0.1999224126,0.2825665772,0.022987444,0.2774751782,0.1182470098,-0.2388511896,-0.6610696316,0.2167615891,-0.0592281856,0.4719746411,0.2986933887,-0.2056801617,-0.1835254431,-0.1310995668,0.0069148391,-0.3616106808,0.108855322,-0.0447663888,0.0705667883,0.2840746045,-0.2449754775,-0.1201994866,-0.4714967906,0.2484924793,-0.5372623205,0.3105761111,-0.0796099827,0.0625527874,-0.0304808728,0.4805233181,-0.0285154581,0.0097297132,0.1863795072,0.0043972102,-0.2089272439,0.2935680747,-0.1489443332,0.3889605105,0.2140198648,-0.3447876275,-0.1804744601,-0.0957653672,-0.0777004063,-0.2883735895,-0.0209821425,0.2886338234,0.0869845748,0.4630931914,0.1234813854,0.3386545479,-0.0108269025,-0.1775801182,-0.0193868633,-0.2788067162,0.1496568769,-0.5367226601,0.1365720183,-0.0645934567,-0.4196057022,0.2042813897,0.427975446,0.0633509979,0.0718592405,0.0658705533,-0.4009689987,0.1018072888,0.0327859856,0.0492319576,0.1975305676,0.1843510568,0.3715714514,0.0923902094,-0.0590708666,-0.2055073678,0.2382936627,-0.0001002185,0.1313720047,0.2200181633,-0.0017095954,0.0597875081,-0.1700548679,0.1733988523,-0.2498636097,0.3950412869,-0.3663571775,-0.0438934714,-0.1436811239,-0.2902813852,0.0288748071,-0.0873045996,-0.141522035,-0.4789790809,0.0346712805,-0.0348249972,0.0401363261,0.171226576,-0.1327504516,0.2457701415,-0.11718072,0.1424688846,-0.4899621308,-0.1282466203,-0.17699866,0.148713395,-0.0496284142,-0.0152700916,0.0834247991,-0.2558934987,0.060629528,-0.2867175341,-0.5384836793,0.1535821855,-0.2546754777,-0.0277708229,0.2747757733,0.6557905078,-0.2294668406,-0.2044860721,0.256631881,-0.3007084727,-0.1779071093,0.1082173362,-0.2639544606,0.1515257061,-0.251955837,-0.2615916729,-0.2821849585,-0.2204409391,0.1645997465,-0.0915771127,-0.012906163,0.236612767,-0.0878046602,0.2831338942,-0.1692287624,0.0539389476,-0.1425020844,-0.2312063873,0.4616045058,-0.2854109108,-0.2608576119,0.080700174,-0.3057962358,0.5293775797,0.0693496987,-0.3885041177,-0.1409018785,0.1390135735,0.1803540736,0.0790945366,0.0125414105,0.4569828808,-0.3402540088,0.1417889595,-0.3257477283,0.316938132,-0.0234829076,-0.2708432972,0.142745629,-0.159111172,0.1861771345,0.1096415371,0.2882342935,0.2257869989,0.0538421944,0.2514748573,-0.4117173851,0.2570097744,0.0497902818,-0.1021385714,0.3323720992,-0.0246806033,0.3655768037,0.3398316801,0.1039027199,-0.0304457285,-0.1205992326,-0.0682411492,-0.241681844,-0.1490650326,0.0589585491,0.238777563,0.3836188912,-0.0671869144,-0.0412841029,-0.0111016454,-0.3337470591,0.0258440785,0.2859095037,0.2306169719,0.0882383734,-0.2942484021,0.1608676761,-0.495525986,-0.1271621734,-0.035528522,0.1020803973,-0.0938194916,-0.0358863473,0.0103501435,0.1439751834,0.6889334321,-0.419459939,-0.4454329312,-0.1986941993,0.1826639026,-0.4102360308,0.1526100785,-0.1578062773,-0.6618097425,0.1869552284,0.5838924646,-0.2288831323,-0.0369236208,0.4610932767,0.2755669951,-0.1745368242,-0.1738995761,-0.2141486406,-0.3961091042,-0.2489791811,-0.3374049067,0.1674870998,0.2511649728,-0.105062604,-0.0483422838,0.0309323482,0.1877819151,0.055779662,0.1609740704,0.291723758,-0.0202620141,-0.0054042148,0.2812901437,0.0475018322,0.0096258167,0.3408079147,-0.1727465689,-0.3893283904,0.0165146068,-0.0480982922,0.3733415604,-0.0120249446,-0.0054974523,0.2215070277,0.180039674,0.1133569777,-0.229287073,0.2550822496,0.4812558591,-0.0678879097,-0.0369301103,-0.2190570086,0.3981318176,0.1039310992,0.2893247902,0.2170252353,0.5756605268,-0.204061985,-0.0088243643,0.0043378104,0.6888923645,0.0431720205,0.2604459226,0.6488656998,0.1300826967,0.6366934776,-0.1697514057,-0.024728125,-0.3252650201,-0.190898478,-0.1308867633,0.0493107848,0.1411261261,-0.0547626801,-0.1942636818,0.1737315357,0.2398361117,0.6364045143,0.1294537783,0.013268711,-0.0044815629,-0.4697115123,0.0351223238,0.0425626114,-0.1734475344,0.2748548687,-0.0885678232,0.0530234091,-0.0597690716,-0.1742842942,-0.303494662,0.1209153906,-0.1696868986,-0.0543456078,0.4672954679,-0.2665228546,0.088800557,0.5696345568,0.2888655663,-0.0574188381,-0.4292489886,0.3164286017,-0.3440201581,-0.1321984828,0.1689972579,-0.0002475014,0.3026667237,-0.2705130875,-0.2358099222,-0.0012535194,-0.0581090525,0.2543415129,-0.496527642,-0.2070839554,-0.2827220261,-0.1058981791,-0.07925044,0.1633304805,0.1962474138,-0.2985829413,0.110230729,-0.1691223979,-0.1328527182,-0.0437236652,-0.076951474,0.1435174644,-0.0132539282,-0.0233261529,0.0821141973,-0.2183973491,0.4074656069,0.3545238078,-0.2506162524,-0.1042833775,0.0408609621,0.3909344673,-0.425508678,-0.1830893308,0.3778044879,0.1231880561,-0.1935776174,0.4957154393,-0.0833096504,-0.261208564,-0.0919973031,-0.2706179619,-0.3745496571,0.3178093731,0.0038883432,0.1583868861,-0.1975408345,0.3579546809,0.0004277346,0.1056279615,-0.2790939212,-0.014235802,-0.1662764847,0.0083159599,0.1782567054,-0.1054553613,0.2839036286,0.0324821472,0.0307686795,0.1814560443,0.0852223486,-0.1257598251,-0.2736720741,0.0976115242,0.139945507,0.1397403777,0.0373199172,0.3310207129,-0.233046025,-0.1823267937,0.280921787,-0.0835620239,0.0644367486,0.3747586012,0.1335432827,0.2256377637,-0.1648500711,0.0778085515,0.0603142306,0.1051202118,0.0486091599,0.0451048613,0.1113063917,-0.0334514566,0.1679162681,0.2166941464,-0.0283962805,-0.0299928617,0.3547175229,0.0130243804,0.045335982,-0.1104091853,0.194545716,-0.2014467418,-0.0422599353,0.3111617267,0.1811486185,0.1537886858,-0.2879523337,0.0950942561,0.5618888736,0.0091566714,0.1660359949,0.2860006392,0.2187125981,-0.1394684315,-0.2600670457,-0.1360290945,0.3661950827,-0.2573718727,0.0548411086,0.3716140687,0.0797657296,-0.0221911557,-0.0395521671,-0.0004072602,0.0410141796,-0.028788114,0.0501169227,0.3103124797,-0.5167754292,-0.0165042225,0.4140037596,-0.2651940286,-0.2028977126,-0.1084448248,-0.0689845011,-0.0224914923,-0.3505363166,0.302315712,0.3013870418,-0.1437640786,-0.1384696662,0.3903588951,-0.3028219044,-0.1124036685,-0.5327321291,-0.175851047,0.1697009951,-0.114469789,0.2899709344,-0.0980347916,-0.1407258064,-0.0042727487,-0.0231498424,0.0454274118,-0.1444651186,0.2635335326,0.1723185331,0.04406045,0.1336942017,0.4879324734,0.0019556917,-0.1372496784,0.0469448455,0.229680568,-0.1872312725,0.2690658271,0.0805297717,-0.0810470954,-0.1450400949,0.1178016141,0.2352012992,0.3659512401,0.0038219681,0.226804316,0.0050576003,-0.160735786,0.0690551177,-0.1922186911,0.0795901567,-0.1604693532,-0.1012974083,-0.3562685847,-0.2814124823,-0.3153088093,0.1010260582,-0.4224944711,0.3207923174,0.3138857186,0.0636068657,0.001910712,-0.0168929733,0.0447771735,-0.0498585701,0.5095519423,0.2807851732,0.2484706193,-0.1566556096,-0.4722030759,-0.5128248334,0.1290957034,-0.2937474847,-0.1378597915,0.2778899074,0.0712928772,0.1493514329,0.0080829551,-0.1277738512,0.1497380883,0.0351926312,-0.0261972155,-0.3916731179,-0.3891707063,-0.2848655581,-0.0568103306,-0.0499064066,0.2799979448,0.1806738526,0.108494617,-0.0296755582,-0.1062722653,0.5135328174,-0.1903238147,-0.2412327826,-0.2446423024,0.0120436912,0.4205495417,-0.1620685607,-0.0037234002,-0.0436831713,-0.1016224772,-0.2380420268,0.165715903,-0.1277560145,0.438722074,-0.0151886651,-0.4770477116,-0.0812945962,-0.0144443782,0.3587680161,-0.0919234976,-0.2852696776,0.2698578835,0.1301117986,0.0226346664,0.4985995591,0.2061764449,-0.1455754936,-0.031354513,-0.0372555591,-0.5985196829,0.4953221679,-0.1852775365,0.0062366957,-0.1610581428,-0.2958181202,0.2354765385,0.1117395237,-0.8591642976,-0.068565838,0.2883366644,-0.00157657,0.065197207,0.2277213484,0.1733224541,0.0647350177,-0.0727829114,-0.4401552975,0.0614163168,0.0164777078,0.0301134046,-0.1097397804]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2130","title":"wikiann dataset is missing columns ","comments":"Hi @lhoestq \r\nthank you very much for the help, it would be very nice to have it included, here is the full code, one need to also convert tags to string first:\r\n\r\n```\r\nimport datasets \r\nfrom datasets import load_dataset\r\n\r\ndef tags_to_spans(tags):\r\n \"\"\"Convert tags to spans.\"\"\"\r\n spans = set()\r\n span_start = 0\r\n span_end = 0\r\n active_conll_tag = None\r\n for index, string_tag in enumerate(tags):\r\n # Actual BIO tag.\r\n bio_tag = string_tag[0]\r\n assert bio_tag in [\"B\", \"I\", \"O\"], \"Invalid Tag\"\r\n conll_tag = string_tag[2:]\r\n if bio_tag == \"O\":\r\n # The span has ended.\r\n if active_conll_tag:\r\n spans.add((active_conll_tag, (span_start, span_end)))\r\n active_conll_tag = None\r\n # We don't care about tags we are\r\n # told to ignore, so we do nothing.\r\n continue\r\n elif bio_tag == \"B\":\r\n # We are entering a new span; reset indices and active tag to new span.\r\n if active_conll_tag:\r\n spans.add((active_conll_tag, (span_start, span_end)))\r\n active_conll_tag = conll_tag\r\n span_start = index\r\n span_end = index\r\n elif bio_tag == \"I\" and conll_tag == active_conll_tag:\r\n # We're inside a span.\r\n span_end += 1\r\n else:\r\n # This is the case the bio label is an \"I\", but either:\r\n # 1) the span hasn't started - i.e. an ill formed span.\r\n # 2) We have IOB1 tagging scheme.\r\n # We'll process the previous span if it exists, but also include this\r\n # span. This is important, because otherwise, a model may get a perfect\r\n # F1 score whilst still including false positive ill-formed spans.\r\n if active_conll_tag:\r\n spans.add((active_conll_tag, (span_start, span_end)))\r\n active_conll_tag = conll_tag\r\n span_start = index\r\n span_end = index\r\n # Last token might have been a part of a valid span.\r\n if active_conll_tag:\r\n spans.add((active_conll_tag, (span_start, span_end)))\r\n # Return sorted list of spans\r\n return sorted(list(spans), key=lambda x: x[1][0])\r\n\r\ndataset = load_dataset('wikiann', 'en', split=\"train\")\r\nner_tags = {\r\n 0:\"O\",\r\n 1:\"B-PER\",\r\n 2:\"I-PER\",\r\n 3:\"B-ORG\",\r\n 4:\"I-ORG\",\r\n 5:\"B-LOC\",\r\n 6:\"I-LOC\"\r\n}\r\n\r\ndef get_spans(tokens, tags):\r\n \"\"\"Convert tags to textspans.\"\"\"\r\n spans = tags_to_spans(tags)\r\n text_spans = [\r\n x[0] + \": \" + \" \".join([tokens[i]\r\n for i in range(x[1][0], x[1][1] + 1)])\r\n for x in spans\r\n ]\r\n if not text_spans:\r\n text_spans = [\"None\"]\r\n return text_spans\r\n\r\n\r\nfor i, d in enumerate(dataset):\r\n tokens = d['tokens']\r\n tags = d['ner_tags']\r\n tags = [ner_tags[i] for i in tags]\r\n spans = get_spans(tokens, tags)\r\n print(\"spans \", spans)\r\n print(d)\r\n if i > 10:\r\n break; \r\n```\r\nI am not sure how to contribute to the repository and how things work, could you let me know how one can access the datasets to be able to contribute to the repository? Maybe I could do it then\r\nthanks \r\n","body":"Hi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq ","comment_length":402,"text":"wikiann dataset is missing columns \nHi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq \nHi @lhoestq \r\nthank you very much for the help, it would be very nice to have it included, here is the full code, one need to also convert tags to string first:\r\n\r\n```\r\nimport datasets \r\nfrom datasets import load_dataset\r\n\r\ndef tags_to_spans(tags):\r\n \"\"\"Convert tags to spans.\"\"\"\r\n spans = set()\r\n span_start = 0\r\n span_end = 0\r\n active_conll_tag = None\r\n for index, string_tag in enumerate(tags):\r\n # Actual BIO tag.\r\n bio_tag = string_tag[0]\r\n assert bio_tag in [\"B\", \"I\", \"O\"], \"Invalid Tag\"\r\n conll_tag = string_tag[2:]\r\n if bio_tag == \"O\":\r\n # The span has ended.\r\n if active_conll_tag:\r\n spans.add((active_conll_tag, (span_start, span_end)))\r\n active_conll_tag = None\r\n # We don't care about tags we are\r\n # told to ignore, so we do nothing.\r\n continue\r\n elif bio_tag == \"B\":\r\n # We are entering a new span; reset indices and active tag to new span.\r\n if active_conll_tag:\r\n spans.add((active_conll_tag, (span_start, span_end)))\r\n active_conll_tag = conll_tag\r\n span_start = index\r\n span_end = index\r\n elif bio_tag == \"I\" and conll_tag == active_conll_tag:\r\n # We're inside a span.\r\n span_end += 1\r\n else:\r\n # This is the case the bio label is an \"I\", but either:\r\n # 1) the span hasn't started - i.e. an ill formed span.\r\n # 2) We have IOB1 tagging scheme.\r\n # We'll process the previous span if it exists, but also include this\r\n # span. This is important, because otherwise, a model may get a perfect\r\n # F1 score whilst still including false positive ill-formed spans.\r\n if active_conll_tag:\r\n spans.add((active_conll_tag, (span_start, span_end)))\r\n active_conll_tag = conll_tag\r\n span_start = index\r\n span_end = index\r\n # Last token might have been a part of a valid span.\r\n if active_conll_tag:\r\n spans.add((active_conll_tag, (span_start, span_end)))\r\n # Return sorted list of spans\r\n return sorted(list(spans), key=lambda x: x[1][0])\r\n\r\ndataset = load_dataset('wikiann', 'en', split=\"train\")\r\nner_tags = {\r\n 0:\"O\",\r\n 1:\"B-PER\",\r\n 2:\"I-PER\",\r\n 3:\"B-ORG\",\r\n 4:\"I-ORG\",\r\n 5:\"B-LOC\",\r\n 6:\"I-LOC\"\r\n}\r\n\r\ndef get_spans(tokens, tags):\r\n \"\"\"Convert tags to textspans.\"\"\"\r\n spans = tags_to_spans(tags)\r\n text_spans = [\r\n x[0] + \": \" + \" \".join([tokens[i]\r\n for i in range(x[1][0], x[1][1] + 1)])\r\n for x in spans\r\n ]\r\n if not text_spans:\r\n text_spans = [\"None\"]\r\n return text_spans\r\n\r\n\r\nfor i, d in enumerate(dataset):\r\n tokens = d['tokens']\r\n tags = d['ner_tags']\r\n tags = [ner_tags[i] for i in tags]\r\n spans = get_spans(tokens, tags)\r\n print(\"spans \", spans)\r\n print(d)\r\n if i > 10:\r\n break; \r\n```\r\nI am not sure how to contribute to the repository and how things work, could you let me know how one can access the datasets to be able to contribute to the repository? Maybe I could do it then\r\nthanks \r\n","embeddings":[0.0544689596,-0.3270129263,-0.0403905883,0.1326312274,0.2911989093,0.2451720685,0.4067163765,0.1009387821,0.3828304708,0.1556164026,-0.0446803346,-0.2178338468,0.022738019,0.5038781166,0.3199461699,-0.5506708026,0.1787430644,0.2836467326,0.0272963587,-0.1717296094,-0.3373773396,0.4151010215,-0.3747096658,-0.0913586766,-0.4198635519,-0.0521617867,-0.1558331102,0.0584633909,-0.2034338564,-0.4034020305,0.329411298,-0.2170119286,0.3202043772,-0.1154671088,-0.000112541,-0.1020316333,0.1146142855,-0.074127838,-0.0352271535,0.1029127836,-0.002470843,-0.3005408645,-0.0322350711,-0.2677301466,-0.2919104695,-0.127156347,-0.3608503342,-0.0884254649,-0.1699058861,0.1173611358,0.159419924,0.4930120409,0.187277481,-0.0216508061,0.3861594796,-0.3597493768,-0.2000508308,-0.2237435728,0.2682072818,-0.1672688574,0.2621860504,0.5828677416,-0.0686727986,-0.2373126149,0.3806412518,0.1863372028,-0.0965012759,-0.3542980254,0.1478399783,0.4884112477,0.4684244394,-0.1065146923,-0.1449523121,-0.2883477807,0.0657660663,-0.3216191232,0.0381794125,0.1077186316,0.0494384505,0.2737787962,-0.0201080497,0.0279782601,-0.0424054191,0.0508821867,-0.210271731,0.5055118203,0.0122651393,-0.0037946987,0.0270377286,0.0542170182,-0.2672324479,0.0604036003,-0.000239622,0.1808323115,-0.4676630497,-0.0149633326,-0.0637844354,0.7178283334,0.3234871626,0.1278939843,-0.2642259598,-0.4074225724,0.2040248364,0.0630208775,0.0163024049,0.2194528282,0.017754212,0.1384577304,-0.1093440801,0.2411258668,0.1371101886,-0.1718849093,0.1250827461,0.0112145869,0.0479666665,-0.0958264023,-0.0406199582,-0.1761170626,-0.1931806505,0.1283470541,-0.0858135149,0.0573211499,0.1037845463,0.3675775528,0.0357893631,0.4057350755,-0.018197991,0.2109326571,0.0414993502,-0.184901908,-0.3456234634,0.172377035,-0.0278714001,0.0412089787,0.206149891,-0.2425575107,0.4308030903,-0.0141541502,-0.1144159213,0.1334284395,-0.1852980107,-0.1513625234,0.0604233481,0.4951552749,-0.1300824881,0.2043958306,0.1434557885,-0.2034413368,-0.0840799585,0.0595366172,-0.2339164466,0.0201573782,-0.4020797014,0.1664364785,-0.1108929515,-0.2205124497,-0.0506509095,0.1065469459,0.0239960384,-0.2974378765,0.1126538292,-0.0470756963,-0.2015538961,-0.030695213,0.5745875239,0.189858973,-0.1930187941,-0.4148733914,0.436081022,-0.0263606999,0.2025561035,0.2573499382,0.2017432451,0.1720208228,-0.1324974298,0.4924700558,0.0134954182,-0.40783301,-0.2425951064,-0.3458082676,0.2205595076,0.3580059409,0.1147091091,0.0673441067,0.3221626282,0.0256251562,0.0146008562,0.3263651431,-0.0976219475,-0.020421287,-0.2326065004,-0.1400739849,0.236929372,0.1449745744,0.0888518617,0.0245976467,-0.2037452757,-0.005946245,0.2381316125,-0.316313684,0.1725292206,0.2288783789,0.08282125,0.2735074461,0.232669726,-0.3583384752,-0.5895218849,0.2272499949,0.1954203993,0.255638063,0.1291144043,-0.2900441289,-0.3728681803,-0.0594472177,-0.1254445165,-0.3232939243,0.1320796162,0.0530789942,0.0084884148,0.1862318367,-0.2948439717,-0.1074061468,-0.2357054651,0.2480612695,-0.3208344877,0.2400354594,-0.1686615497,-0.0328182504,0.0763440505,0.5403302312,0.2565369308,0.1897570938,0.2228338718,0.0916352421,-0.1450525075,0.0279494468,-0.0130796339,0.3257299066,0.1200422198,-0.3045971096,-0.1943529397,-0.1095432118,0.170019567,-0.2002994716,0.0100841383,0.0964758247,-0.0876425877,0.3575112224,-0.0103016002,0.2266384512,0.1442184597,-0.317783922,-0.0587051809,-0.2331504822,0.1952025294,-0.5777381063,0.1849351972,-0.1640563011,-0.2464762926,0.1813760549,0.4493895471,-0.098596096,0.0781156868,-0.1623220593,-0.4065583348,0.1908221841,0.0960349143,0.0082345149,0.2017858922,0.2465304881,0.1954315603,-0.087123096,-0.0235711709,-0.0434327573,0.2000274062,-0.0140677243,0.0008669352,0.2576184273,0.042649895,0.083630316,-0.1842699945,0.1168388054,-0.3363680243,0.2314849198,-0.3505465984,-0.0540646389,-0.0917514116,-0.1673776954,0.0889806449,-0.4040912092,-0.0554510579,-0.3870818913,0.1263013333,-0.2790803611,0.0139314625,0.2027008981,-0.2238692641,0.3222596943,-0.0865766481,0.0373751707,-0.5950977206,-0.1842696071,-0.174240604,0.0743246302,-0.1024547964,0.1173359379,0.0428954512,-0.4117192328,-0.0802081302,-0.2833993435,-0.4494692981,0.1166326553,-0.2666516304,0.1625020057,0.3171647787,0.4674129486,-0.0920891464,-0.1513972878,0.3166525364,-0.1866428107,-0.0389574729,0.1505758762,-0.0790561214,0.0194667708,-0.1821569353,-0.1631705165,-0.3570208848,-0.0780605748,0.1331015676,-0.0158120357,-0.139848724,0.1193088666,-0.0854908153,0.1214030683,-0.1633446962,0.155112505,-0.1893157959,-0.3184945583,0.3573021889,-0.0028754638,-0.2122382373,0.0864414126,-0.2347598672,0.3610833287,-0.0930034667,-0.3724879026,-0.1330963075,0.2365102023,0.1786486953,0.0681660175,0.0832075551,0.4093667865,-0.1339916289,0.1313051432,-0.3668015599,0.0353037491,-0.1806097031,-0.4232722819,0.2840663493,-0.3843836486,0.2557320893,0.0164650436,0.1918068528,0.206547454,0.2342860401,0.2982036173,-0.2027377784,0.2933222651,-0.0803180486,-0.315286845,0.286696285,-0.1873613596,0.1270025074,0.3359539807,0.0512873121,-0.0133397402,-0.1811530292,-0.1329340041,-0.297929734,-0.1639046818,-0.0078742886,0.4573527873,0.2752798498,-0.0486616269,-0.0057233637,-0.1397505552,-0.1836078763,0.2187120616,0.4127667546,0.2451678067,0.099310182,-0.2173625231,0.0696448535,-0.4181894958,-0.0287405718,0.0339625701,0.0534690656,0.1074629799,-0.0410086475,0.0089009078,0.077793479,0.3949428797,-0.4277153313,-0.1420179009,-0.085979864,0.2854680121,-0.1988028884,0.039426513,-0.2738189697,-0.5508035421,0.280757457,0.5836244822,-0.3903928995,-0.0757064223,0.7279964089,0.3608463407,-0.076367557,-0.2026692778,-0.4303674996,-0.4182761312,-0.2724165916,-0.1223142445,0.324090451,0.2643380761,-0.0167246666,-0.046523422,-0.1571272612,0.000397471,0.0761726424,0.2137708217,0.3229747415,-0.1104358211,-0.0561896637,0.2773343921,0.000599834,0.0057455706,0.4690227807,-0.3493476212,-0.5048450232,-0.1113383397,-0.0512561537,0.477504313,-0.0490518846,0.025559606,0.1722625345,0.1563808322,0.0661832839,-0.2490689903,-0.0385927334,0.4647085369,0.1458493471,-0.0820628703,-0.1797797829,0.4675607979,0.1549694985,0.1091684029,0.26870507,0.6941232681,-0.2821185291,0.1339612305,-0.0013887654,0.8452179432,-0.0397437364,0.1042331532,0.4053997993,-0.0132363234,0.3032665551,0.0281621721,0.1390977353,-0.3000718355,-0.2921208143,-0.0908368453,0.1666675508,0.1481084079,-0.1318234652,-0.2458907217,0.275598079,0.0452361777,0.6778355241,0.0746098384,-0.1033955216,-0.0017686592,-0.4012656808,-0.2720258534,0.0822917372,-0.1310529262,0.184404701,-0.201508075,0.1920023859,0.087227121,-0.2494218051,-0.2389716804,0.0637117922,-0.2709907591,-0.1070002913,0.3383772075,-0.3196052611,-0.0449286811,0.4351572692,0.2685519457,0.0004199021,-0.278811872,0.2626274228,-0.2549905181,-0.0973985419,-0.0191494357,-0.135215655,0.3243162632,-0.2111025006,-0.3843373656,0.0903838649,-0.1274330765,0.0182264615,-0.2287192196,-0.0358418077,-0.0269988719,-0.2296764851,0.0142785301,0.1640332043,0.3964558542,-0.3822667301,0.0931846499,-0.1406217068,-0.1895360202,0.0779170021,-0.0804077089,0.0065079345,-0.0941133723,0.0342353694,0.011029101,0.0023250442,0.5822216272,0.1355502605,-0.3204783499,-0.152446717,-0.1715672761,0.3862132728,-0.2608520985,-0.0786110014,0.2744740248,-0.0822326764,-0.0807961524,0.5901309848,0.1658878326,-0.3848463595,-0.0733298585,-0.4468899667,-0.1905550957,0.4615369141,0.0596764982,0.1075812951,-0.0156427864,0.2550135553,-0.0372405015,0.1408180594,-0.3331301808,-0.1739227176,-0.3433177471,0.0917421058,0.048310332,-0.108537361,0.2005737424,0.0188449658,0.1040671691,0.0480880626,0.2977728546,-0.1665762514,-0.2269820869,0.1043345705,0.099751085,0.2559483349,-0.0327716507,0.1996017545,-0.1600095481,-0.2049890757,0.3374839723,0.0187263079,0.1683520824,0.2305443138,0.186361745,0.1055728197,-0.3049565554,-0.021705348,-0.0400156602,-0.0664833561,-0.0186127946,0.0228280351,0.1278032064,0.0514722168,0.5048713684,0.2136392742,0.0830894336,0.2258837223,0.5083485842,-0.0964582413,0.0130085777,0.1635410637,0.1499213725,-0.1923559159,-0.0679066107,0.1851631999,0.0591698736,0.231948927,-0.4967128038,0.0002578,0.4716872275,-0.0388899185,0.2034625709,0.2427162677,0.2596533298,-0.2499643117,-0.1458478272,-0.1017299369,0.4809441566,0.028289428,-0.0263646152,0.3083896339,0.0784661546,-0.0630010292,0.1078327149,0.0511320792,0.1140611395,0.0823959932,-0.0422184952,0.3785286844,-0.2654440701,0.2412343174,0.0684906989,-0.0437602438,-0.0139463479,-0.1738166809,0.0251186695,-0.1282300502,-0.3254926503,0.2293959856,0.120278798,-0.1393979341,-0.2085933983,0.3673755527,-0.2201185375,-0.0590571389,-0.5520489812,-0.1504018158,0.1835420132,-0.2491616458,0.0816262662,0.0153165217,0.0778467208,0.0939845294,0.0406756476,-0.0017693157,-0.4534667134,0.3404271901,0.3654014766,-0.0818671808,0.1732188612,0.6445074081,0.0751832202,-0.0365172699,0.3512734473,0.2951741219,0.0236942265,0.2055936903,-0.1491785496,-0.1340323538,-0.293841213,0.0907438844,0.3199096918,0.2083317637,0.0336417817,0.173656866,0.0767551586,-0.1981589198,0.0633970872,-0.2220886946,0.1091489196,-0.227944836,-0.2378139943,-0.391230762,-0.3179879785,-0.465973407,0.029981371,-0.3282496929,0.3707672954,0.1707827747,0.0695131347,0.0342908166,0.1300940663,0.0593452826,0.10577479,0.4426098466,0.2982802689,0.0838686302,-0.3084408939,-0.3819876611,-0.5238916278,-0.0934573188,-0.3235938549,-0.4068906903,0.213756904,-0.0177834779,0.1535544246,-0.0101389522,0.101721622,0.3596038818,0.1713932157,-0.0809158683,-0.4236364961,-0.2975741625,0.0174140744,-0.0518015437,-0.0253844615,0.0251307134,0.2179336399,0.1491937488,0.017383026,0.1136091575,0.2490448058,-0.1416776031,-0.2748277783,-0.0798643529,0.174821496,0.5028145909,-0.2660527825,-0.0049170065,-0.0348543487,-0.1281792521,-0.2321847379,0.1482658237,-0.0469658002,0.2544283569,0.0553977601,-0.3131453097,-0.0385000594,0.1561456174,0.3408409655,-0.0530774705,-0.4521439075,0.1979328245,0.0803285912,-0.0011352883,0.3005705774,0.2109133005,-0.1628839672,0.1323053539,-0.1846253723,-0.5597349405,0.6108741164,-0.2102544606,-0.1114958674,-0.2517187893,-0.0512028784,0.3974973559,0.1314936578,-0.8236064911,0.0154704498,0.1567982137,0.1745764464,-0.0727523044,0.3572431505,0.1974403113,0.1512195915,0.022446828,-0.4629210234,0.2020682544,0.0254843291,-0.1014105678,-0.047272753]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2130","title":"wikiann dataset is missing columns ","comments":"Cool ! Let me give you some context:\r\n\r\n#### Contribution guide\r\n\r\nYou can find the contribution guide here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/CONTRIBUTING.md\r\n\r\nIt explains how to set up your dev environment in a few steps.\r\n\r\n#### Dataset loading\r\n\r\nEach Dataset is defined by a Table that have many rows (one row = one example) and columns (one column = one feature).\r\nTo change how a dataset is constructed, you have to modify its dataset script that you can find here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wikiann\/wikiann.py\r\n\r\nIt includes everything needed to load the WikiANN dataset.\r\nYou can load locally a modified version of `wikiann.py` with `load_dataset(\"path\/to\/wikiann.py\")`.\r\n\r\n#### Define a new column\r\n\r\nEach column has a name and a type. You can see how the features of WikiANN are defined here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/c98e4b8f23e3770c401c6d9326e243e1ffd599ec\/datasets\/wikiann\/wikiann.py#L245-L263\r\n\r\nIdeally we would have one additional feature \"spans\":\r\n```python\r\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\r\n```\r\n\r\n#### Compute the content of each row\r\n\r\nTo build the WikiANN rows, the _generate_examples method from [here](https:\/\/github.com\/huggingface\/nlp\/blob\/c98e4b8f23e3770c401c6d9326e243e1ffd599ec\/datasets\/wikiann\/wikiann.py#L292-L316) is used. This function `yield` one python dictionary for each example:\r\n```python\r\nyield guid_index, {\"tokens\": tokens, \"ner_tags\": ner_tags, \"langs\": langs}\r\n```\r\n\r\nThe objective would be to return instead something like\r\n```python\r\nspans = spans = get_spans(tokens, tags)\r\nyield guid_index, {\"tokens\": tokens, \"ner_tags\": ner_tags, \"langs\": langs, \"spans\": spans}\r\n```\r\n\r\nLet me know if you have questions !","body":"Hi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq ","comment_length":208,"text":"wikiann dataset is missing columns \nHi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq \nCool ! Let me give you some context:\r\n\r\n#### Contribution guide\r\n\r\nYou can find the contribution guide here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/CONTRIBUTING.md\r\n\r\nIt explains how to set up your dev environment in a few steps.\r\n\r\n#### Dataset loading\r\n\r\nEach Dataset is defined by a Table that have many rows (one row = one example) and columns (one column = one feature).\r\nTo change how a dataset is constructed, you have to modify its dataset script that you can find here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wikiann\/wikiann.py\r\n\r\nIt includes everything needed to load the WikiANN dataset.\r\nYou can load locally a modified version of `wikiann.py` with `load_dataset(\"path\/to\/wikiann.py\")`.\r\n\r\n#### Define a new column\r\n\r\nEach column has a name and a type. You can see how the features of WikiANN are defined here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/c98e4b8f23e3770c401c6d9326e243e1ffd599ec\/datasets\/wikiann\/wikiann.py#L245-L263\r\n\r\nIdeally we would have one additional feature \"spans\":\r\n```python\r\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\r\n```\r\n\r\n#### Compute the content of each row\r\n\r\nTo build the WikiANN rows, the _generate_examples method from [here](https:\/\/github.com\/huggingface\/nlp\/blob\/c98e4b8f23e3770c401c6d9326e243e1ffd599ec\/datasets\/wikiann\/wikiann.py#L292-L316) is used. This function `yield` one python dictionary for each example:\r\n```python\r\nyield guid_index, {\"tokens\": tokens, \"ner_tags\": ner_tags, \"langs\": langs}\r\n```\r\n\r\nThe objective would be to return instead something like\r\n```python\r\nspans = spans = get_spans(tokens, tags)\r\nyield guid_index, {\"tokens\": tokens, \"ner_tags\": ner_tags, \"langs\": langs, \"spans\": spans}\r\n```\r\n\r\nLet me know if you have questions !","embeddings":[0.00569537,-0.2647256255,-0.0400679149,0.0995709822,0.3907708824,0.108981356,0.2883937061,0.1041567475,0.0132615454,0.1453700364,-0.0328590088,-0.012727851,0.0221513528,0.4286503494,0.2620583177,-0.7255477905,0.1374138445,0.2676771879,-0.2057022899,0.064583838,-0.3263520598,0.3927432895,-0.3106453419,-0.1321281493,-0.3484145701,0.0134390527,-0.2690922916,0.1206341535,-0.094850786,-0.2926417589,0.3413374126,-0.2243309915,0.1851857305,-0.0907957479,-0.0001031631,-0.0195615683,0.1880478561,-0.0933684632,-0.1129946187,0.104285799,0.0963925868,-0.3275562525,-0.0058890199,-0.3028342426,-0.3155614734,-0.1220751405,-0.314163208,-0.0753199756,0.0279854368,0.0826015249,0.258515358,0.7911584973,0.3671272695,-0.1308148801,0.1523732394,-0.1912457943,-0.2992332876,-0.1442314386,0.1775325537,-0.1432342082,0.0846133232,0.4101109803,-0.0642807782,-0.1549720764,0.3905907869,0.1574980766,-0.3582746983,-0.2313543558,0.1976914704,0.5012369752,0.4747832119,-0.1316248775,-0.1059161797,-0.3632461429,0.0720394626,0.013684826,0.1230837405,0.1361771077,0.0144161973,0.2254358381,-0.0582174473,0.0265036747,-0.0693991557,0.1189300045,-0.2114753872,0.2074497342,-0.1068044305,0.0620140657,0.0106154885,-0.0684678257,-0.2683206797,0.1802330613,-0.0499178059,0.1471552402,-0.3356831372,-0.004959424,0.0384217538,0.6451608539,0.3114064932,0.160973534,-0.1400149614,-0.1313830912,0.1271444261,0.0326535068,0.0230206288,0.1219923571,-0.0616221689,0.0901369378,-0.1017376259,0.4083092511,0.1303039044,-0.2530061603,0.0655365959,-0.0008754515,-0.2312635183,-0.2843500972,0.0263554864,-0.0653640851,0.0867279246,0.0201112106,0.2051498741,0.1149983779,0.0867084563,0.4324637055,0.0214245394,0.4151434004,-0.0994242132,0.2552137077,-0.0415621698,-0.1355647743,-0.372338891,0.1453407109,-0.0185190216,0.1431327164,0.2540236712,-0.2808769643,0.5359506011,0.0414320454,0.0351255909,0.0959896296,-0.1034573913,0.0935248286,-0.219346121,0.3652416766,-0.0746772513,0.0697429031,0.2229989767,0.0003277576,-0.1197451428,-0.1145025715,-0.1494786739,-0.1088829413,-0.2668517232,0.2236600816,-0.1043155044,-0.1133723333,-0.0867222175,0.0667773783,0.0291008446,-0.3127827644,0.1017027423,0.0485590585,-0.0921484828,-0.2088678628,0.6047569513,0.1654423475,-0.1434135288,-0.3574486077,0.2727653384,-0.0880977288,0.1259169877,0.1465769857,0.0717699602,-0.0579953566,-0.1217666268,0.3233641386,-0.033621598,-0.1268147528,-0.2349395156,-0.1781380624,0.1438904107,0.2943347096,0.1185503453,-0.0268246569,0.2524349988,0.0249335952,-0.0260239821,0.2677815557,-0.0522507429,0.0685677305,-0.1652991623,-0.2542303503,-0.1006456688,0.114961639,0.0070869769,0.0217571016,-0.1243227646,0.0791169778,0.1917811185,-0.2824204862,0.2072332799,0.2506262958,0.0283801816,0.3811194897,0.1086606234,-0.1813731343,-0.55864501,0.1994146258,0.1036452651,0.218346417,0.1075798646,-0.3085534275,-0.2711724937,-0.0610955432,-0.1522563845,-0.3668555617,0.1992108822,0.0398298502,0.2403293103,0.2699042261,-0.2106195092,0.019899223,-0.1637937129,0.1998686343,-0.5162340403,0.2422401458,-0.144382,0.1127644554,0.2207106501,0.5053961873,0.1098286957,0.095648095,0.2538810372,0.2179433852,-0.2460688949,0.2443443537,-0.0961793214,0.1683004946,0.203624621,-0.2858159542,0.0485560559,-0.0938987285,0.12423978,-0.1547206193,-0.0726822913,0.3829087913,-0.1319556385,0.3680707216,0.0671642795,0.3307341933,0.1689669788,-0.2372763753,-0.1055499837,-0.2722427845,0.1091720164,-0.4910502434,0.1773588657,-0.1322771311,-0.3371235132,0.1953537762,0.2848120928,0.025427267,0.1182907894,-0.2042376548,-0.4753547609,0.1389717013,0.1840975881,0.0435797796,0.2743403912,0.3023234904,0.2259216607,0.0136573724,-0.073548831,-0.1543085277,0.1365187764,0.0112465713,-0.0455773249,0.1225967035,0.0138392234,-0.0227478985,-0.1749592274,0.1751216054,-0.3515152335,0.2264763713,-0.2688479722,-0.1025939211,-0.1479930282,-0.1891606003,0.087436609,-0.2197569311,-0.0882473812,-0.3651939034,-0.041730389,-0.1194724292,-0.0319992974,0.1705795676,-0.2978868186,0.1058899164,-0.1315856725,-0.0018019387,-0.4364799857,-0.1045141891,-0.1432973146,0.189244628,-0.0223207641,0.0549125224,0.1327008456,-0.2095164806,0.0066544134,-0.3827700615,-0.4155872762,0.0275884233,-0.35132581,0.0488559753,0.311540097,0.5177799463,-0.1577759236,-0.2482329756,0.4025249779,-0.2532039881,-0.0400607102,0.1358628124,-0.0422104783,0.0660645738,-0.1633740366,-0.1545830518,-0.1375987232,-0.2262219787,0.2925009727,-0.0385330878,-0.1350613236,0.0658270046,0.0675549135,0.2127572298,-0.2418706715,0.033891391,-0.2137432247,-0.3011052012,0.3603340983,-0.2365947664,-0.3118934631,0.114138253,-0.2673543096,0.4330877662,0.0052811247,-0.3415950835,-0.2719323933,0.0759830549,0.1883362532,0.0712454543,0.1267052144,0.4980594814,-0.0780174583,0.0473269336,-0.3099205792,0.0694152638,-0.1727192849,-0.3517266214,0.2287162542,-0.2808218896,0.2558266521,-0.0000867576,0.2596426904,0.2716817856,0.3824597597,0.3410160542,-0.2625616193,0.4241220951,-0.0869308338,-0.2107981443,0.3139961362,-0.2213087976,0.3367098272,0.3756849766,0.2335915267,0.0158644225,-0.206597358,-0.2852897346,-0.2803173363,-0.2121930867,0.0474607907,0.2853039503,0.4106867313,-0.1406691223,-0.019201722,-0.1328485608,-0.1860246807,0.0182932187,0.4052118659,0.0905307531,0.0928514302,-0.3699908853,0.1083805934,-0.5568282008,0.0198754687,-0.1077681258,0.0017431704,0.1339832544,-0.1440435499,0.0111702085,-0.00082365,0.4261743426,-0.1615867615,-0.3184938729,-0.1017037109,0.152330339,-0.4004542828,0.1023613513,-0.1053795144,-0.3845584989,0.058222305,0.4981087446,-0.2920540571,-0.0141466036,0.5848606825,0.3693964481,-0.1100339442,-0.180022791,-0.3422320485,-0.5269818306,-0.2619642317,-0.3119789362,0.0501005799,0.2887431681,0.0004598242,0.0406164527,-0.0858594775,-0.1096921265,0.005795172,0.1932647973,0.3751416206,-0.0855220407,0.0837985426,0.3181835711,0.1286624372,0.027687069,0.4846265018,-0.1169974655,-0.2862417698,-0.2341118902,-0.0015342917,0.3049421012,-0.0503736474,-0.0121489745,0.2031054199,0.1640188992,0.1740409285,-0.2772317529,0.0653348491,0.4079397321,0.0161236152,-0.1211910918,-0.3104068637,0.4774593711,-0.0909735709,0.2345766425,0.1480847448,0.6963037848,-0.1908765286,0.0257107485,-0.0374630839,0.6728922129,-0.0156476013,0.2137645483,0.5228447318,0.1378846914,0.4920895398,0.0322048925,0.1136461794,-0.1365510523,-0.124429062,-0.0557650961,0.1319319159,0.1979288459,-0.2405105084,-0.3001885116,0.2108460218,0.1302981675,0.4815489054,-0.0115635786,-0.0840538293,-0.2299473137,-0.458912313,-0.2244805992,0.1763319075,-0.1310941577,0.1835683286,-0.0612324513,0.0835069939,-0.0620185025,-0.2151123881,-0.2962893844,0.0489887297,-0.1781813502,0.0015896331,0.2778372467,-0.3197650909,0.0429980941,0.3772464097,0.2068167776,0.0808542594,-0.3791576922,0.2942670882,-0.3318923712,-0.3013336957,-0.0808051378,0.0272310004,0.1504332125,-0.2005128264,-0.3311750889,-0.0175726637,-0.153834939,0.0748688504,-0.4822948575,-0.055079475,-0.0164402891,-0.2165943235,-0.1426627487,0.0882919505,0.2669075429,-0.3419037759,0.1730667502,-0.2297581583,-0.0707632527,0.0559880137,-0.0646732226,0.1425490826,-0.1878101379,0.1235514581,-0.0277568903,-0.1479176134,0.5038405061,0.2157179862,-0.2083934397,-0.1833307296,-0.0420924947,0.3237203956,-0.3441511393,-0.1837069839,0.3745023906,0.0978188813,-0.0480917171,0.4815382957,0.0797586069,-0.3061766326,-0.1033128425,-0.2695040107,-0.3904180825,0.4605471194,0.0443820208,0.0814252794,-0.0559591167,0.4622196257,0.0489016064,0.1277226657,-0.3721938431,-0.104990907,-0.1650709808,0.0057259845,0.3218866289,-0.130542919,0.2574725747,0.0498018004,0.1919074059,0.077309005,0.0629747957,-0.2080519497,-0.1677065492,0.062608242,0.010498696,0.1252632737,-0.0322182551,0.2193450779,-0.1892018914,-0.2381414771,0.3489740789,0.0091073653,0.0480813496,0.1124820709,0.1010193825,0.1393238157,-0.2098072916,0.0525637828,-0.0510027334,-0.0601741746,-0.0357956775,-0.0229461566,0.0929102004,-0.0398607701,0.4534706771,0.2300549597,-0.0618665814,0.1680831909,0.4504091144,0.0778943077,-0.0724385157,-0.0130135976,0.179434225,-0.0295940749,0.0005950949,0.242806077,0.209038794,0.2586454749,-0.3936882615,-0.0178882219,0.3326262832,0.0186036956,0.2908472717,0.2393268645,0.094536379,-0.2792143524,-0.1221789941,-0.0453509614,0.5128665566,-0.064764671,0.0349017382,0.3482366204,0.1782117188,-0.1240446791,0.0959935188,0.1547138244,0.2134791017,0.042370256,-0.1238437817,0.2358178496,-0.4316964746,0.1954885274,0.2507392466,-0.2044876069,-0.1044673175,-0.1691025496,-0.0405264385,-0.0144350929,-0.3837434649,0.401022464,0.0296538863,-0.0108313467,-0.1622748673,0.3016045988,-0.2187898606,-0.1499933004,-0.3768318295,-0.2381186485,0.22730802,-0.1794091314,0.1908576339,-0.0246977992,0.0561866313,-0.0710707083,-0.0403258763,-0.1101830527,-0.0587493591,0.3492384255,0.1889839768,0.0533253141,0.1089648381,0.6260311007,0.1344870329,-0.0055447496,0.0823426396,0.3404819071,-0.10857407,0.2931349576,-0.0031082327,-0.1426032782,-0.1822574735,0.0227754489,0.1597888321,0.2254237384,0.0273290146,0.1800416708,0.1658463031,-0.2506607473,-0.010702787,-0.3270643651,0.0899443179,-0.2241419554,-0.1398084313,-0.4170125127,-0.3447076976,-0.2472239435,0.0605845377,-0.5223691463,0.3943909705,0.1322869807,0.0699925721,-0.0346823186,0.0318932422,0.1035730615,0.1147833169,0.4758183062,0.3135999441,0.0263632443,-0.0811810419,-0.5944883823,-0.5643666983,-0.0804431438,-0.3559878469,-0.3042838573,0.1598599255,0.0643119514,0.2171023935,-0.0359672643,-0.0209060963,0.2259666175,-0.0324342586,-0.1610321701,-0.2761782706,-0.3786724806,0.0112641919,-0.0282949507,-0.0162164699,0.0762998462,0.1975307614,0.1973191947,0.0702492669,-0.0398603566,0.4521839321,-0.2078344375,-0.4676822722,-0.2300964296,0.1423254162,0.2826929688,-0.1948659718,-0.0114491116,-0.2692463696,-0.0511224158,-0.1538994461,0.1939672083,-0.047121793,0.4537032247,-0.0226639267,-0.4011211991,-0.0338797867,0.2450492084,0.397295773,0.037338201,-0.2113557011,0.2295653671,0.17386581,-0.0869549438,0.259858489,0.3310913146,-0.224101916,0.0792579576,-0.1126662791,-0.6037771702,0.4661970735,-0.2314511836,-0.1007008925,-0.0388116464,-0.0909451023,0.0995386243,0.0423088931,-0.7625928521,0.127469033,0.2277512103,0.166481331,0.0087281745,0.2390075028,0.0062217773,-0.0185780879,-0.0361022428,-0.2505773008,0.1227641478,-0.0009987087,0.0648507997,-0.0154444547]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2130","title":"wikiann dataset is missing columns ","comments":"The PR was merged. Issue should be closed.\r\n\r\nCC: @lhoestq ","body":"Hi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq ","comment_length":10,"text":"wikiann dataset is missing columns \nHi\r\nWikiann dataset needs to have \"spans\" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq \nThe PR was merged. Issue should be closed.\r\n\r\nCC: @lhoestq ","embeddings":[-0.0456641428,-0.2902656496,-0.0365703069,0.2266795039,0.2069964856,0.1496347934,0.2924788296,0.0287773237,0.1094219759,0.2043087035,-0.0699864551,-0.1247739121,0.2133326977,0.428469032,0.1699955761,-0.6608502269,0.2024963647,0.3118388355,-0.2481651604,-0.0876540169,-0.4071162343,0.4551123083,-0.4294943213,-0.1716876477,-0.39672333,-0.01074031,-0.1833508909,0.1742880791,-0.1701849401,-0.0485226773,0.3073181808,-0.2585666478,0.0885372311,-0.1233104318,-0.0001060665,-0.0167342834,0.1545477957,-0.0130075216,0.0680841506,0.3492953479,0.0039154906,-0.4332817197,-0.0513224266,-0.3083142638,-0.2042663246,-0.1504977643,-0.32640329,0.0080184918,-0.0158672202,0.0601182878,0.2389516085,0.6332120895,0.2080785632,-0.210040316,0.1236173809,-0.3740528226,-0.2235100269,-0.2819723189,0.2227950245,-0.2161840051,0.1507985294,0.4916646779,0.0351754613,-0.30526039,0.2301347405,0.0378107578,-0.2582044005,-0.2624982297,0.2695438266,0.3955919445,0.624463439,0.0786480382,0.1405136436,-0.2014081925,0.0971102715,0.0920143053,0.2225950211,0.109161824,0.0630548894,0.2008294463,-0.1283195615,-0.1645237803,-0.0562552176,0.0104667144,-0.3184458017,0.353951931,0.0462323241,0.0019379081,0.0282844957,-0.0296749119,-0.0307528339,0.0392412134,-0.0869310349,0.0320740677,-0.37890324,-0.1362947524,-0.1241780296,0.707267046,0.1340485662,0.0818286985,-0.2566408217,-0.3022326529,0.2038222551,0.0714660808,0.0922571197,0.127342239,-0.1234160513,0.13612625,0.0524783619,0.2977475524,0.0871332586,-0.1336001903,0.0685955808,-0.0928392559,-0.2121053636,-0.3480027914,-0.031146545,-0.1947943568,-0.0173455272,-0.0316682458,0.2946727574,0.0927961916,0.0803585202,0.3019924164,0.0543204322,0.5039098859,-0.1373522729,0.198797822,-0.0172658768,-0.1243404299,-0.2974757552,0.1278586835,0.061453864,-0.0350140482,0.1238749921,-0.1636170894,0.4241490066,0.1374425441,-0.0434030741,0.2071986198,-0.1817697138,-0.0071022129,-0.1452275217,0.4511607587,0.0227498449,-0.0055786865,0.0997617841,-0.0033321287,-0.0240630321,-0.0129435509,-0.2789736986,-0.0564791858,-0.3474133909,0.2397240102,0.1235508844,-0.2347607017,-0.0405301936,0.1382241994,-0.0041859676,-0.4215364456,0.0148859704,-0.008089019,-0.1376891434,-0.1404498219,0.3495951891,0.0033508528,-0.2096732706,-0.3991636038,0.2231338471,-0.2363518625,0.3361438811,0.5266341567,-0.017227767,-0.1257573962,-0.1867010295,0.2969889641,0.0747412965,-0.1894038916,-0.28626436,-0.2388184965,0.0761222839,0.1868797541,0.0180234592,0.0328413583,0.3419510126,0.015541642,-0.001681563,0.2652478814,-0.088603884,0.0232283454,-0.2380051017,-0.1473487616,0.1147324741,0.1392635256,0.2326247543,0.1132117286,-0.0645131022,0.0755307078,0.0547545366,-0.2667126656,0.2606785893,0.2782123387,0.0405260362,0.1365453005,0.1784157306,-0.0689868405,-0.4885246754,0.1121194512,0.0127024082,0.1966109723,0.280067116,-0.251803726,-0.2850102782,-0.1428892314,-0.1190458983,-0.2662290335,0.1993163228,-0.030469304,-0.0359650441,0.1961743683,-0.1823302358,-0.0071562259,-0.2345344126,0.1218747795,-0.2449610382,0.2825642526,-0.1784056127,0.1545465738,0.1340057105,0.5063812137,0.1568497866,0.1883104891,0.2430268377,0.0057591638,-0.3233004808,0.2204312831,0.236094296,0.0141054764,0.2181397676,-0.3501194715,-0.0655657575,-0.2322968543,0.0897562578,-0.2463583797,-0.1001430899,0.1282768399,-0.0355901644,0.2200693935,0.0930305272,0.3180950284,0.1052846164,-0.3221536875,0.0852804556,-0.1640907824,0.1239387542,-0.4761052728,0.2471910417,-0.0336400829,-0.4064642489,0.1854012012,0.4920435548,0.0035473288,-0.0066454299,-0.1018029377,-0.4116765261,0.0588461906,0.1945987791,0.0475588143,0.3914643824,0.3404217064,0.3068467379,0.0211038347,0.0588747486,-0.1393497437,0.1965773851,-0.019356573,0.0810641125,0.198563233,0.1642591059,0.0596695766,-0.3041508198,0.3683764338,-0.1893589199,0.1791540533,-0.1323878914,-0.2624650598,-0.1335688978,-0.1916732043,0.0608786047,0.0907521918,-0.201592803,-0.4137214422,0.1820250899,-0.1491125822,-0.0669531748,0.1413061321,-0.1299406737,0.4140458405,-0.1930241436,0.2486850321,-0.483722508,-0.1137053147,-0.0768031776,0.2433720231,-0.0580631495,-0.0494568013,0.2488623112,-0.2427961528,-0.0118237529,-0.4847470522,-0.4046311975,0.0871304274,-0.2323858738,-0.0403072424,0.2742837369,0.4978197515,-0.1792396903,-0.1823363453,0.2664556801,-0.280786097,-0.1809408665,-0.0119519476,-0.1329199225,-0.0686694905,-0.1173480079,-0.1703397483,-0.2972858846,-0.1684342772,0.1782195866,-0.1186512485,-0.1688735038,-0.2593436837,-0.1094089821,0.0249693189,-0.2221629769,-0.0831284821,-0.3521214724,-0.2579722404,0.2972651124,-0.107662648,-0.3733371794,0.2273407429,-0.0079986984,0.4669185281,-0.0778958425,-0.3591079712,-0.1169174761,0.0556990243,0.2812693119,0.1172344834,0.0771753266,0.6187852621,-0.153406322,0.0030642329,-0.331543684,0.046515692,-0.1923074424,-0.2517641783,0.4206949174,-0.2735407352,0.1164188683,-0.1058131456,0.3984113634,0.2302145511,0.4203751683,0.3090737462,-0.1722459495,0.4642578661,-0.1103410572,-0.2339179814,0.3316310644,-0.0825881436,0.291061759,0.3735985458,0.1797408015,-0.1536646336,-0.3363868296,-0.2876378,-0.224749729,-0.084573634,-0.1354389787,0.4449847043,0.2806254923,0.0680197254,-0.1300586462,-0.1041298956,-0.1248366684,0.0818032101,0.4370712638,0.1159670949,0.010153397,-0.3249357939,0.1478532553,-0.4678448737,0.0009838025,-0.1662235558,-0.001007957,-0.0449191108,-0.1017462462,0.0352484174,0.0245783739,0.5011236072,-0.3864817321,-0.3723323643,-0.1799845248,0.2209269255,-0.2438625842,0.1318574846,-0.1117614731,-0.5700436831,0.2217318714,0.5367744565,-0.2391563654,0.0279639456,0.6397787929,0.2030441761,-0.1324659288,-0.2695122659,-0.3396401703,-0.4653710425,-0.1824559867,-0.1838407516,0.1401269287,0.179186359,-0.2009733766,0.0172839016,-0.0658435002,0.0107001234,-0.0208930764,0.1766950041,0.319096595,-0.0721050128,-0.020402059,0.2938320637,0.3453406096,0.0873379707,0.5156215429,-0.075182043,-0.3042873442,-0.0572050773,-0.1580076069,0.4614025354,0.0596210733,0.0353042632,0.1077809855,0.1352488548,0.1899636835,-0.0734143183,-0.0613812394,0.4097476304,0.0606052391,-0.2522113919,0.0499280766,0.3089773059,0.1687723845,0.1189181656,0.1705719233,0.4658099413,-0.160061568,-0.1414457411,-0.151013881,0.8203548789,-0.009185941,0.2091209888,0.4446378946,-0.0740145966,0.3951168656,0.0644609183,0.0933846533,-0.2340831608,-0.1626955718,-0.0664607361,0.1217039302,0.1798134595,-0.3124618232,-0.2740472257,0.1866546422,0.2178708464,0.6891412139,0.0010967209,-0.1041590348,-0.0081286179,-0.3364423513,-0.2467958033,0.1547080725,0.0183297414,0.1949289441,-0.2002957165,0.042839285,0.013396061,-0.2068722248,-0.2053354234,0.041322697,-0.0796341076,-0.0043599983,0.2511912286,-0.3558033705,0.0299053732,0.3944310844,0.2820809484,0.0252750292,-0.4280293882,0.3367337286,-0.1963068694,-0.1241200939,-0.056760408,-0.1451513767,0.005070094,-0.3718340397,-0.3611502349,0.0464044586,-0.2602705956,-0.0119658867,-0.6300517917,0.0882558301,-0.061769329,-0.2377735078,-0.1053657606,0.0796356797,0.2825163305,-0.4149397016,0.1626387388,-0.3067811131,0.0070202271,0.0952964425,0.0378078148,0.0676579177,-0.1350903213,0.190089494,0.1217357218,-0.1346075982,0.4903514385,0.0668201596,-0.292734921,-0.2198024094,-0.06620083,0.2695795,-0.3048811257,-0.1122731641,0.3354774714,-0.0792302862,-0.0315377563,0.4675700665,0.1058105379,-0.3540227711,-0.0951678604,-0.24640356,-0.4144424796,0.3630357981,-0.0729639679,0.0879802182,0.0975322723,0.3114541471,0.0613562986,0.1024321169,-0.3661194444,0.0154492371,-0.300265789,-0.0165141653,0.1463135332,-0.1454800814,0.0610213131,-0.0574025773,0.1359447986,0.0057927021,0.2279230058,-0.1986791193,-0.0264102872,0.0923006907,0.0877254829,0.0483545251,-0.0435146354,0.1044401601,-0.2144277543,0.0553827137,0.301882565,-0.1224655434,0.0931583494,0.0531601124,0.2272004485,0.0147643229,-0.2749665678,0.1966319382,0.0169651154,0.0017883778,-0.1822899282,-0.0691568851,0.1798469424,-0.0326940008,0.561478436,0.1553892195,0.1379999369,0.2431952357,0.4110943675,-0.0415063873,0.039878998,0.0359995477,0.0547427833,-0.0821687356,-0.1051528677,0.1659439206,0.1149085537,0.2559486032,-0.5949689746,-0.0959019512,0.5087761283,0.1143141463,0.1320482045,0.0406692885,0.2630636096,-0.2975491285,-0.0795574188,0.0100119747,0.4661887586,-0.116213426,0.047908619,0.2418574691,0.1707941741,-0.0087930877,0.0176603496,0.0928937793,-0.0278650522,0.0894862041,-0.0218596663,0.2615591884,-0.2401667088,0.2138645798,0.2925148904,-0.1687953472,0.1138782799,-0.1221922785,-0.0705928579,-0.1212386265,-0.1837412417,0.2241652906,0.1780371815,-0.2656686902,-0.2166675627,0.4562142193,-0.3672887683,-0.034844242,-0.558380723,-0.2739365697,0.447551012,-0.0690499246,0.1294159144,0.1264788359,0.0657249689,-0.0113320798,-0.0222260766,-0.011192508,-0.0491987206,0.3765012622,0.2704458535,0.0766118392,0.0746340379,0.5730025768,0.0623905957,0.1218972579,0.2951198518,0.301133126,-0.128022328,0.2811307609,-0.1477909237,0.020793952,-0.2414903492,0.0358938314,0.2671460211,0.3401961029,-0.0658072531,0.2198199928,0.1721150428,-0.2074832022,0.1572322845,-0.2929970026,-0.0102568259,-0.2212410569,-0.2820284963,-0.4210676849,-0.4179886281,-0.3591761887,0.1212360337,-0.205279842,0.3817219138,0.1797717661,0.098873347,0.1127789319,-0.0516868122,0.1038359031,0.1000865325,0.4804516733,0.4372805953,-0.0993437469,-0.1060731933,-0.6481342912,-0.5375947356,0.1144686937,-0.39502877,-0.1902672499,0.1562085301,-0.0651892498,0.1912803203,-0.1329668909,0.1848231405,0.3275175691,0.0917105004,-0.1949383914,-0.3911990225,-0.2452260554,0.2390619665,-0.097393997,0.0142421294,-0.0061374758,0.3765430152,0.3208436966,0.0844001696,0.0378332771,0.4363313913,-0.3127728403,-0.2775098681,-0.3571380973,0.3187267482,0.442692399,-0.2771702707,0.0041873776,0.0349556282,-0.1656450331,-0.2045072019,0.2145155072,0.0329023078,0.1769306511,-0.0062727001,-0.3052831292,0.1211590767,0.1209674552,0.2977418303,0.2453752607,-0.3755245805,0.117170997,0.2783867717,-0.0093106609,0.2622888386,0.1835701913,-0.1766636074,-0.1093750522,-0.1974923611,-0.5481283665,0.4577805698,-0.077703163,-0.1573266238,-0.1552992612,-0.1648037732,0.1365679353,0.0174154323,-0.5732591748,-0.1205301434,0.2299271971,0.0259661302,0.0951876715,0.3358927071,-0.0096480353,0.0371373631,0.0085337451,-0.1852687448,0.1538310051,-0.1607857794,-0.0106046861,-0.0376703963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2129","title":"How to train BERT model with next sentence prediction?","comments":"Hi !\r\nWe're not using `TextDatasetForNextSentencePrediction` in `datasets`.\r\nAlthough you can probably use the `TextDatasetForNextSentencePrediction.create_examples_from_document` on a dataset to prepare it for next sentence prediction.","body":"Hello.\r\n\r\nI'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction \r\nlike ` TextDatasetForNextSentencePrediction` of `huggingface\/transformers` ?\r\n","comment_length":25,"text":"How to train BERT model with next sentence prediction?\nHello.\r\n\r\nI'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction \r\nlike ` TextDatasetForNextSentencePrediction` of `huggingface\/transformers` ?\r\n\nHi !\r\nWe're not using `TextDatasetForNextSentencePrediction` in `datasets`.\r\nAlthough you can probably use the `TextDatasetForNextSentencePrediction.create_examples_from_document` on a dataset to prepare it for next sentence prediction.","embeddings":[0.1917076111,-0.4076450765,-0.0147695839,-0.2003295273,0.0005590364,-0.2783757746,0.1422066987,-0.0127167599,-0.025662059,0.1647019535,0.1556011438,0.0493725613,-0.154809311,0.0617535748,0.3285053372,-0.6952239275,0.1574005932,0.2188371867,0.009745351,-0.1505313516,-0.1178788096,0.0806166455,-0.4232550859,0.3682828546,-0.2839281857,0.2008214891,-0.3334325254,-0.3163951933,0.0460604019,-0.3753812909,0.4582475722,0.4070344269,0.4587988555,0.2150741071,-0.0001232347,-0.2639684081,0.2025311142,-0.2711344659,0.1549269408,-0.397015512,-0.2493353486,-0.0373750441,0.224298805,-0.1815945953,-0.2838987112,0.5994516015,0.3069365919,-0.1343758255,0.5240787268,0.1173760816,0.0644793734,-0.2960521281,0.2634472549,-0.0121575696,0.1068005115,0.0945525542,-0.0285038985,-0.1371493936,0.5605081916,-0.3304388225,-0.214719072,0.1836271137,0.0088032726,0.0330856815,0.1647519916,0.2331249118,0.08448603,-0.3325142264,-0.2616679966,0.4053135216,0.6561758518,-0.2457776666,-0.5032140017,-0.1677221358,0.2668515742,0.0886337683,0.0660710782,-0.2455876619,-0.0742204636,0.172976777,0.0470753126,-0.3753949106,-0.4167921245,-0.0373331085,-0.1894734949,0.4580947161,-0.0336559638,-0.2123973519,0.2601470351,-0.2106651068,0.0755168945,-0.0940305889,0.1235339269,0.4406594336,-0.2064272463,-0.3648631275,-0.4254557788,0.1489193439,0.0206151064,-0.1328348368,-0.1927832663,-0.0155303888,-0.3786577582,-0.1238597631,0.3669160306,0.2585484982,-0.1486679614,0.156735003,-0.2529693544,0.0401985459,-0.2351615131,-0.187681824,-0.1457653791,0.0405989587,0.2635551393,-0.0009324501,-0.2121559978,-0.2638917267,0.0923204944,-0.0122005958,-0.2975632548,-0.0437658429,-0.0768808052,-0.0495994948,-0.2629685998,0.3475267887,-0.0341194905,0.0840811431,0.004310641,-0.0952290446,0.1146304458,-0.2163175344,-0.2744835913,0.5209566951,0.0122855129,0.0309622642,0.1323937774,-0.1471810341,-0.0324883945,0.0797838643,-0.3151405454,-0.4305484891,-0.0083031971,-0.1275520623,-0.3998718262,0.0512893051,-0.1620687842,0.4676387012,-0.0828430876,-0.0061797379,-0.2025858015,-0.0096685551,-0.0974933654,-0.039380189,-0.0925439224,-0.3396705389,0.1031367257,0.5064285994,-0.1084639728,-0.16338332,0.1413177997,0.1575701982,-0.3678383231,0.1389539242,0.1589907259,0.1643917114,-0.4927240908,-0.2593483627,0.0213332176,0.1455704421,0.1135781407,-0.0017803777,0.2418296635,0.2236623317,0.0471895337,0.1285944134,-0.121634312,-0.1538394094,0.0056323283,0.0265370309,0.1224368885,0.3793030977,-0.1333853304,-0.0062227282,0.0107168267,-0.0767609626,-0.1099487543,0.1562724113,-0.1232269108,-0.1303879321,0.2724360824,-0.1773424745,-0.1098015755,-0.0045532607,0.1601713747,0.1707350761,-0.273924619,0.5601226687,0.1560769379,-0.0312377177,0.0705626458,-0.0506209098,0.1408510208,0.3964399397,0.0488680042,-0.3296991587,-0.2331066579,0.025221467,0.0732033923,-0.1738704741,0.0573339611,0.156717062,0.0682236925,0.1022693142,-0.0619812123,-0.153783083,-0.0262000691,0.007966659,-0.1567399949,0.104198277,-0.3908598423,0.1572053879,0.1015610397,0.078009814,-0.1400816143,0.2063272297,-0.0807496905,-0.0708810985,0.0169515051,0.1848720312,0.0469684489,0.0842282698,0.1061468348,0.1018844172,0.2673927546,0.1256760955,-0.0165784508,-0.2433182001,0.0560218543,-0.3695042729,-0.0651000887,0.0415278189,-0.1241791919,-0.0720176026,-0.1598884165,0.0651435107,0.4588318765,0.0063851457,-0.0904362947,-0.021364117,-0.1394985318,0.0529068187,-0.0706688091,-0.3643370271,0.1267165989,0.0481702015,0.252730161,-0.086599566,-0.2399302125,-0.479178071,0.2228339761,-0.174281925,0.1773670614,0.6320177913,-0.0266932193,0.0455350094,-0.1115405187,-0.2695039809,-0.0333991498,0.074216187,-0.0677747875,-0.1553356498,-0.0478103161,0.0303329155,-0.0855010897,-0.0484032221,0.1988034844,-0.1507501453,-0.256457597,0.1240442544,0.208206296,-0.3337570131,-0.1325201541,-0.0418708101,-0.079084821,-0.0052340426,-0.3815504014,-0.3519111574,-0.0259710513,-0.433439374,-0.0323948599,0.2949759364,-0.0653684363,-0.4361508191,0.1682783961,-0.0546021946,0.0482203513,0.3473086357,0.1947134584,0.2138714343,-0.2305919975,-0.1608908921,-0.0018192065,0.0149620892,-0.1995003819,-0.1814264208,-0.3212387562,-0.002616164,0.0561572872,-0.1347346455,-0.2250508368,0.3262950182,-0.1361160427,0.0319761075,0.2602560222,-0.1232942268,-0.147617504,0.0506936051,-0.0717645884,-0.0051311892,-0.0586234108,-0.1770613492,0.2005548477,0.1248717755,0.268186748,-0.0115823569,-0.2322618067,0.3484328091,0.2574885786,0.1877579093,-0.0506631993,-0.0934298486,-0.0130336666,0.1530901492,-0.335332036,-0.1277411133,0.1583463997,-0.6027779579,0.1883777082,0.1695082486,0.0336447842,-0.3637563288,-0.0378945097,-0.1281221509,-0.3695198596,-0.0321028531,-0.2308461517,0.0703089014,-0.1127632186,-0.0981119722,-0.1799512953,-0.0321077332,0.1058504805,0.1223225966,-0.1785277575,-0.0347802527,0.3163150549,-0.0391908661,0.3434461951,0.059299998,0.0938187838,-0.2042819262,0.4010342062,0.1032167375,-0.29291448,0.3866353631,0.2231460512,-0.1186935306,-0.0734554529,0.0201253314,0.0247448087,0.0631081164,0.0930059329,0.4289288521,-0.1150057986,0.0435726084,-0.1596843153,-0.1794784516,0.0603451505,-0.0714286044,0.0728941113,-0.2532197833,0.2317323238,-0.23556979,-0.1013941765,0.0414578654,-0.1059637889,-0.0669277236,0.0172649566,0.7841863632,-0.1106740013,0.6015790105,-0.1010554731,-0.3486663103,-0.0605147853,0.2454375774,-0.0601606108,-0.1274957061,0.2383757532,-0.0657186061,0.1213704273,-0.0068481839,-0.0616919883,-0.0566114858,-0.0942688882,-0.3269929588,0.3027222455,0.1651551425,-0.3013947606,-0.3171466589,0.1521595567,0.2069497854,0.1803200245,-0.1959773302,0.2799467444,0.0934672058,-0.1397625804,-0.5285359621,0.0431036018,0.8037567735,-0.1165577173,0.0836314186,-0.0647991449,-0.2123404443,0.0651605427,-0.2847047746,0.2285289019,0.1364208907,-0.0540830269,0.0294560026,-0.0028115499,-0.2884871066,0.0143234581,0.3151859045,0.0962371379,0.242391929,-0.2252099514,-0.016093377,-0.2057083845,0.1425428689,0.1346212327,0.4665711522,-0.1037176028,0.1235325933,0.4220383167,0.0214608479,-0.0144513715,-0.1236788407,-0.1614875793,0.6815999746,0.3218707144,-0.2997699082,-0.2854650021,0.5634611845,-0.1164377257,-0.2301301807,-0.4375348091,0.6853522062,-0.328528136,0.2281905115,0.5515842438,0.7926334739,-0.0096535971,0.2065885216,-0.3263476491,0.1404922605,0.5958323479,-0.1239839494,-0.1427018046,0.0330235139,0.2338633984,-0.3432445526,0.2214526534,0.0697835982,0.2556044459,-0.7860977054,0.3353565633,0.1801725775,-0.1527982801,0.11677102,-0.0082780896,0.3885197937,-0.4617995322,-0.0048262854,-0.0076271645,0.1532492191,0.1993293017,-0.1854361594,-0.0329356454,0.5668469071,0.1171777099,-0.433346957,0.0316883847,-0.2583078742,-0.3672170341,0.0752776787,-0.0096676759,0.7311245799,0.421728462,0.6386290789,0.2278711647,-0.2766393721,-0.0145207923,0.102805756,-0.0718985349,-0.3596817851,-0.0008189778,0.4709010124,-0.1251448393,-0.3809984922,0.2287142873,0.0857181102,-0.2684838176,-0.1642098874,-0.2814658284,0.5577855706,0.0331194103,0.0293633286,-0.0116133587,0.173758924,-0.1102287024,0.0694146827,0.1510995924,-0.1651632041,0.0612508841,0.2021568865,-0.3422718644,-0.0281577613,0.0646882132,0.0919549987,-0.0822275802,0.5879970789,0.2623740435,0.0200398117,-0.1871555299,-0.0087565919,0.2483518869,-0.1598691195,0.1676208079,0.0919230357,0.1605503857,0.1754573584,0.6059984565,0.2905309796,-0.2866260111,-0.248162806,-0.0358545668,-0.2740088999,-0.1466109604,-0.1725732684,0.1570884138,0.2134370357,0.0523281358,0.4901928604,0.413915962,-0.1262048185,-0.2374562472,-0.2027714849,0.1641663015,0.0826697648,0.0243980382,-0.1381001472,0.042445872,-0.0397878624,0.2467683107,0.3000513315,-0.038016744,-0.1571827978,0.1104030684,0.0027842019,0.0456356518,-0.1846472323,0.1091515571,-0.008050533,-0.0272238366,-0.1992295235,0.0547469668,0.311511904,0.4729119837,0.4307446182,-0.1541298181,-0.3780584037,-0.2686293125,0.2248029262,0.0470785312,-0.0866403654,-0.0844637156,0.0088333143,-0.1457201093,0.0360588469,0.191204831,0.5139396787,0.1459176242,0.0652674213,0.0956189483,-0.099386543,-0.1315359324,0.0460898392,-0.184878692,0.1676028371,-0.3008453846,0.5770235658,0.0201424323,-0.222630024,0.0027385952,0.0728146955,0.1160575897,-0.0762046129,0.013838483,-0.1462629437,0.3463765383,0.323939234,-0.1283792853,-0.3920297027,0.2033091486,0.0426399708,0.4159962833,0.2745645642,0.1892276406,0.2468419522,0.2012286186,0.4515752792,0.5087415576,-0.1123400927,0.0888403654,-0.2122878879,0.2115273625,0.5340945721,-0.1541152149,-0.1312069893,0.0304419938,-0.0124152424,0.1880620867,-0.0117500648,0.5242729783,0.211565569,-0.1821555346,-0.301766932,0.1600116938,0.0822906718,-0.4470415711,-0.1828816831,0.1288456023,0.2121818364,-0.082818903,-0.078535147,0.0438898243,0.1241260991,0.290913254,0.3355188668,-0.1258829385,-0.0707311034,0.0368153304,0.3359080851,-0.0366836116,0.3192498982,-0.1124020964,0.1083966941,0.0926401392,0.4123351574,-0.083228223,0.1506921947,-0.1259784102,0.1461826861,-0.07325086,-0.0132193509,0.3068969548,0.0166674815,0.1719355881,-0.0788676217,0.3082796931,-0.0208631661,0.0020847959,0.315662086,0.0452641472,0.3824549615,-0.3345514834,0.1478634775,0.3477678597,0.387968868,-0.4021084607,0.3009830117,-0.4886258245,0.476559937,-0.3405267894,-0.3245186508,-0.1606909037,0.235073179,-0.0235532578,0.0613738634,0.4114911556,0.5823990703,0.213008672,-0.0491898321,-0.2652749121,-0.1193499044,-0.1415781975,0.0590277649,0.0451794006,0.0604438484,-0.2913477719,-0.0414710045,0.032930363,-0.0440396443,0.0037750911,-0.6597325802,0.2102500945,-0.2647414505,-0.0164888762,-0.0986361355,0.255206883,-0.0948322564,0.2848756313,0.0068678339,0.2559460104,-0.1820853502,0.2368067056,0.0664249882,-0.1656373888,0.1881761253,0.1926780343,-0.1980177909,0.1864918172,-0.1682225764,0.3638273776,0.0152523424,0.0624063201,-0.2151347101,-0.2940662801,0.1702277213,0.2356828302,-0.1999231726,0.2335377634,-0.0017077291,-0.1841318905,0.1036582589,-0.0436888561,-0.5682262182,0.0535769835,0.1174621657,-0.2165209204,0.1294321716,0.1187784448,-0.1587744802,-0.0105822533,-0.384224683,0.0522069447,0.3020175993,-0.0924395621,-0.1122982427,-0.3752437532,-0.3578157723,0.2979234159,-0.2171179205,-1.1201757193,0.1488175988,0.309050709,0.3505089283,-0.1097138822,0.2135538459,0.135074392,0.1876415759,-0.3854333162,0.5604220629,0.3421281874,0.098850362,-0.3323817849,-0.2206930071]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2129","title":"How to train BERT model with next sentence prediction?","comments":"Thanks.\r\n\r\nDo you mean that `TextDatasetForNextSentencePrediction.create_exapmles_from_document` can be applied to dataset object other than `TextDatasetForNextSentencePrediction` e.g. a `Dataset` object which is loaded by `datasets.load_dataset`?","body":"Hello.\r\n\r\nI'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction \r\nlike ` TextDatasetForNextSentencePrediction` of `huggingface\/transformers` ?\r\n","comment_length":24,"text":"How to train BERT model with next sentence prediction?\nHello.\r\n\r\nI'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction \r\nlike ` TextDatasetForNextSentencePrediction` of `huggingface\/transformers` ?\r\n\nThanks.\r\n\r\nDo you mean that `TextDatasetForNextSentencePrediction.create_exapmles_from_document` can be applied to dataset object other than `TextDatasetForNextSentencePrediction` e.g. a `Dataset` object which is loaded by `datasets.load_dataset`?","embeddings":[0.1056432575,-0.4607722163,0.0155634182,-0.0898428783,0.0474332199,-0.2899057567,0.1532109827,-0.0051718443,0.0640042797,0.1214186475,0.1853307784,0.1281926483,-0.15475142,0.1282212734,0.3845975101,-0.6245176792,0.1314673722,0.2932183444,-0.0514884144,-0.2433327734,-0.1696441323,0.073376745,-0.3516555727,0.3405378163,-0.2001281977,0.2702140808,-0.2841590941,-0.2478072792,0.0087696882,-0.3340138495,0.4047503173,0.3907679617,0.4258323312,0.1643717438,-0.0001291847,-0.2329467982,0.114739798,-0.327388674,0.1137860715,-0.2843592763,-0.3555471599,-0.1060970351,0.26434201,-0.2460024804,-0.2422992587,0.5384497046,0.3581981957,-0.3058199584,0.5242074728,0.2908453941,0.014402139,-0.2546208799,0.3124649823,-0.020818809,0.234868899,0.1671464443,-0.0251937862,-0.221622318,0.5485693812,-0.2986629307,-0.1138822734,0.1442739964,-0.001834902,0.0014028711,0.2522639632,0.2281707823,0.1782483011,-0.4217782915,-0.1953303516,0.3760996461,0.8002337813,-0.3221176863,-0.4739646316,-0.2419865429,0.337128818,0.0954880714,0.125540942,-0.3030329049,-0.1238534153,0.1631256491,0.0612535663,-0.5166053772,-0.4674112201,-0.0005794692,-0.1490326375,0.4557164311,0.0035350297,-0.1510754377,0.2814887762,-0.1666588485,0.0924607813,-0.0972436219,0.0533188321,0.4821875989,-0.1632685363,-0.3454045653,-0.4761663675,0.0677624866,0.052934207,-0.0809000805,-0.1527076811,-0.0065810508,-0.3215615153,-0.0709301978,0.3516069651,0.2880739272,-0.0382891856,0.1914600134,-0.2848033309,0.019716531,-0.2681785524,-0.1537960172,-0.2019642442,0.0399128795,0.2256612331,-0.0506861508,-0.2289211899,-0.2590069175,0.0584818088,0.0120561328,-0.3375009,-0.0117463842,-0.0488630719,-0.0032535426,-0.1972683519,0.365483284,0.0605770722,0.1042166278,-0.0432598777,-0.1118823513,0.1542611718,-0.1464899629,-0.3213078082,0.5611475706,0.0152224051,0.0879947469,0.1360875666,-0.257285893,-0.0913072154,0.0353684165,-0.2435012609,-0.4149817526,-0.0002142072,-0.0694689378,-0.3622319698,0.0641431585,-0.1449044347,0.4078865647,-0.1863521188,0.0402226113,-0.1744345874,-0.0844530538,-0.1815864444,-0.0750281587,-0.0413931832,-0.2907975614,0.0477495193,0.594668448,-0.0108278953,-0.2039385736,0.2527200878,0.0989334062,-0.4636226594,0.1473245323,0.1507641822,0.233998239,-0.5984503031,-0.3204489648,-0.1350890547,-0.0203734282,0.0243378524,-0.1360595226,0.2206615806,0.3552303612,0.0420456268,0.1495957226,-0.1622886956,-0.225841105,-0.0467957594,0.0574617088,0.1385168433,0.3321613967,0.0013421765,0.0143065415,0.1212003529,-0.0726614147,-0.1335150599,0.1792034209,-0.0856366977,-0.1414667517,0.3205605745,-0.1384384483,-0.2136109173,0.0385105684,0.2292429209,0.2022790164,-0.2451477349,0.5747007728,0.1156553701,-0.1170086786,0.1535997987,0.0178294871,0.0713092312,0.4100022316,0.0239023808,-0.3240316808,-0.3736844361,-0.0477182567,0.0286675133,-0.1072706431,0.1274763793,0.1054651141,0.012155191,0.1328706145,-0.0856349394,-0.121808812,-0.0996610671,0.0531059392,-0.1422735006,0.0766886249,-0.3936302066,0.1868783385,0.1223072857,0.1092066094,-0.1749433726,0.1587536484,0.0218498018,-0.0378621854,-0.0522038043,0.1195597053,-0.0094299642,0.1290657222,0.067081295,0.0998976976,0.2876978219,0.1083296165,-0.0522969216,-0.0948836952,0.0758774877,-0.3679954112,-0.0043352665,0.083757773,-0.0230586398,-0.08894182,-0.086956501,0.0916647315,0.5599464774,-0.0662254244,0.0157278627,-0.0259141997,-0.1526126862,0.0673014149,-0.0987571776,-0.477316767,0.0501062199,0.1455092132,0.3399935663,-0.1226238757,-0.1897749305,-0.4507785141,0.2656005323,-0.1263488084,0.2417237163,0.6629160643,-0.0783080757,-0.0234727152,-0.1915446222,-0.3000428677,0.0701903924,0.0617020093,-0.052704934,-0.0462213382,0.0219264328,0.0213555023,0.0432414412,-0.0637597293,0.2537965477,-0.0874788687,-0.3057594597,0.0475675277,0.136055544,-0.3302097321,-0.1117590368,-0.0098609962,-0.1607448608,0.037585672,-0.2359712571,-0.2540568411,-0.1496952325,-0.4255619943,-0.081215702,0.3417153955,-0.0035799034,-0.3079102933,0.1087584645,-0.0709210187,0.0714120269,0.4477214813,0.1611574143,0.1817821711,-0.2483633906,-0.1266447902,-0.0054879463,-0.0562205277,-0.1853322238,-0.1140364558,-0.273488462,-0.0050244643,0.0911660194,-0.1832859963,-0.2373598218,0.2947241068,-0.100845322,0.0405861288,0.3184700906,-0.230072543,-0.1163907796,-0.0433314592,-0.0805316716,-0.0763998181,-0.1201311275,-0.3259272277,0.1578608304,0.1687268019,0.2795046866,-0.0727304071,-0.2464631945,0.2916783392,0.2228225321,0.2209516019,0.0110416273,-0.1808915585,-0.0432179756,0.1967174113,-0.2912327349,-0.0663998947,0.1601238847,-0.5610364079,0.1195918098,0.0951546729,0.0408797301,-0.4551675022,-0.0261343662,-0.1833859682,-0.2468456775,-0.1003962308,-0.2364137918,0.0273860488,-0.2666942179,-0.0847500861,-0.0711287037,0.156466648,0.1362821311,0.1602434218,-0.1897275001,-0.1324950904,0.4403705895,-0.0629497617,0.2673963904,-0.0151512064,0.0811432227,-0.2939181924,0.3493477106,0.0586958528,-0.2686465681,0.48464939,0.2132169306,-0.0882057771,-0.1651723087,-0.0161853656,-0.0867068544,0.0082968734,0.0832559392,0.4489497244,-0.1140031144,-0.0129833203,-0.2246312946,-0.211699903,-0.0043846644,-0.0787050873,0.0240182206,-0.2155964673,0.2518781722,-0.2402085662,-0.0373718478,-0.0661785379,-0.0609552152,-0.0766064525,-0.0004376107,0.7021431327,-0.2040721923,0.4300746024,-0.1523265541,-0.3472485244,0.0436213613,0.2906056345,-0.0336055905,-0.1176429763,0.169507578,-0.017006997,0.1150396243,0.149136886,-0.1024861708,-0.1103024259,-0.0607236251,-0.3613397777,0.3181831539,0.1899960339,-0.274468869,-0.3287584484,0.1643841863,0.2621647716,0.0494755581,-0.1573314071,0.2379766256,0.0778618902,-0.1574308276,-0.4162982106,0.1044439971,0.6891816854,-0.1487593055,0.1067001596,-0.1123583764,-0.1162270233,0.1644563377,-0.1982583702,0.0921159163,0.1108335331,-0.0500882864,0.0315719806,-0.0258051865,-0.2509803176,0.0666440353,0.2289300859,0.184146598,0.3121169209,-0.110003747,-0.1405964643,-0.2539544702,0.076856941,0.130689472,0.5098465681,0.0191487316,0.1812544912,0.3650161624,0.0297457352,-0.0710019246,-0.1195961237,-0.0747348368,0.6595306993,0.2114694268,-0.3762809038,-0.3568400741,0.5436474085,-0.0196717139,-0.2051053941,-0.2914199531,0.6835458279,-0.4077726603,0.2534978986,0.5497311354,0.9744489789,-0.0374594331,0.2863143682,-0.264816463,0.1990032792,0.5697479248,-0.1444056034,-0.0742442682,-0.0938630626,0.1445078105,-0.3599348664,0.1986391097,0.1488520354,0.3245767951,-0.7072689533,0.2890149355,0.1858673394,-0.0824410245,0.0933630094,0.0756098405,0.3367232382,-0.5600144863,-0.1659643203,-0.0831452981,0.1920495778,0.182482779,-0.1349291652,-0.0879467949,0.582605958,-0.0210051313,-0.4990071654,0.0452838279,-0.2814256847,-0.2888810933,0.0005389012,-0.0347594693,0.6878584623,0.3567858934,0.6128691435,0.2255895734,-0.2492447942,-0.0071977917,0.1721433848,-0.1504907012,-0.265272975,-0.0764863417,0.5720114112,-0.1382039785,-0.3254819214,0.2795313895,0.1015457362,-0.3372758031,-0.0818868503,-0.3989439309,0.5842721462,-0.0144157922,-0.1239872128,0.0598439798,0.1022565365,-0.1970918626,0.0016729332,0.0871962011,-0.1682985425,0.1306882501,0.1620140076,-0.3973863721,-0.0814884901,0.1589452028,0.0932038277,-0.0849237218,0.6222240925,0.2137706578,0.0519981422,-0.216443345,0.0564528182,0.342169106,-0.3368004858,0.2119741291,0.1599097997,-0.0201007295,0.1936511248,0.5828647614,0.2529809773,-0.1969836652,-0.2264260948,-0.1262505054,-0.2213078588,-0.1293241084,-0.1437798887,0.2000528872,0.2188762724,0.0399122685,0.4216785729,0.3823932707,-0.0677022189,-0.2510085702,-0.127662167,0.1136756092,0.2548125684,0.1491671205,-0.0394453779,0.0981893167,-0.0617552288,0.2860830426,0.2141279429,0.0348787792,-0.1573131531,0.2046242505,0.0016040951,0.0094382325,-0.170386523,-0.0169744324,-0.136268422,-0.1141251624,-0.1313111484,0.0985914171,0.2123113126,0.4492029548,0.3921929896,-0.0584696904,-0.2855450213,-0.1886838973,0.2176955342,0.0627674833,-0.0851583406,-0.0332639143,-0.0541749634,-0.1285975277,-0.0343865976,0.2448503524,0.5551089048,0.1981613934,0.1298579723,0.1563272923,-0.1594551057,-0.096920602,0.0731710717,-0.0511524938,0.1640902609,-0.2559703887,0.5922891498,-0.0524083264,-0.2567975223,-0.077082634,0.1780082881,-0.0581967905,-0.081221737,0.0298096929,-0.1027333215,0.3521797061,0.2211383581,-0.1163537204,-0.4240362644,0.28134045,0.0834194198,0.5365974307,0.2504784167,0.1899952739,0.1743182689,0.304394424,0.5802149773,0.5396011472,-0.0946695507,0.1569250524,-0.1880414933,0.262765497,0.5272644758,-0.1657825559,-0.0490096919,0.0662341341,0.0509975925,0.1888044626,0.0240651574,0.4067956805,0.1684356779,-0.134640038,-0.4231340289,0.2094099522,0.1448746473,-0.4133072793,-0.1138605252,0.0826578215,0.2206772119,-0.0743773207,-0.1168400124,0.0290403329,0.1580933779,0.2191918492,0.3983952105,-0.0463253707,-0.0783950388,-0.0547137409,0.2943121791,-0.0852542073,0.2956570983,-0.2421523929,0.1789082885,0.1489425153,0.4170618057,-0.1188827753,0.2679031193,-0.1568498164,0.0877341479,-0.129255861,0.0478323959,0.1877611727,0.0604904778,0.2324232459,-0.0659116805,0.3413060308,-0.0977269188,0.0567000173,0.2536103427,0.1228322387,0.3264482617,-0.2582112253,0.2013040185,0.3605732024,0.360632658,-0.3829971552,0.3268388212,-0.442774266,0.3775325418,-0.2477124929,-0.2968658209,-0.150454849,0.2291055322,-0.0632222816,0.0447385535,0.4278230369,0.5208466649,0.231128484,-0.057415735,-0.2300031781,-0.2174286693,-0.029476909,0.1514995396,-0.118258208,0.0844169855,-0.2819283605,0.0597909465,0.016807206,0.0466318876,0.0233676434,-0.5602827072,0.1847176105,-0.1549167484,0.0030541671,-0.1352066696,0.3056132495,-0.08230979,0.0896471664,-0.0146754337,0.2100084126,-0.2170940042,0.1330136955,0.0821461156,-0.1845584363,0.225022316,0.2851876616,-0.1943356395,0.2990404367,-0.2385918647,0.3976639807,0.0766031295,0.0586292893,-0.3067618906,-0.1804482341,0.3098560572,0.1983598173,-0.2711743414,0.280654192,-0.072811611,-0.1415168047,0.1355854571,0.04005678,-0.6012182236,0.1438590139,0.0292283949,-0.2243852019,0.1326211542,0.1417180151,-0.1705304086,-0.0256200247,-0.3078814447,0.0602427199,0.3240751922,-0.158422187,-0.122210823,-0.3839448094,-0.2924028933,0.2498108894,-0.3335623145,-1.1292834282,0.1586370617,0.3405273855,0.2833714187,-0.2243623137,0.1930171549,0.0710296854,0.1255245358,-0.3565337658,0.5560401678,0.2431781143,0.1063830331,-0.3350417018,-0.2991442382]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2129","title":"How to train BERT model with next sentence prediction?","comments":"It would probably require a bit of tweaking, but you can apply it to a dataset, yes.\r\nThis should give you a new dataset with sentence pairs you can train a model on.\r\n\r\nYou can find the documentation about dataset processing here:\r\nhttps:\/\/huggingface.co\/docs\/datasets\/processing.html#processing-data-with-map","body":"Hello.\r\n\r\nI'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction \r\nlike ` TextDatasetForNextSentencePrediction` of `huggingface\/transformers` ?\r\n","comment_length":43,"text":"How to train BERT model with next sentence prediction?\nHello.\r\n\r\nI'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction \r\nlike ` TextDatasetForNextSentencePrediction` of `huggingface\/transformers` ?\r\n\nIt would probably require a bit of tweaking, but you can apply it to a dataset, yes.\r\nThis should give you a new dataset with sentence pairs you can train a model on.\r\n\r\nYou can find the documentation about dataset processing here:\r\nhttps:\/\/huggingface.co\/docs\/datasets\/processing.html#processing-data-with-map","embeddings":[0.2216068804,-0.4078709483,0.0245913751,-0.1242590845,0.0343485698,-0.1844581366,0.090413101,0.022120554,-0.0510081947,0.0038792929,0.0329256169,0.0937431753,-0.188250035,0.1736756414,0.340739429,-0.6513742805,0.1297619492,0.1553862095,-0.0457976721,-0.0744871125,-0.0303964727,0.1138738766,-0.2975053489,0.3866290748,-0.2856276035,0.167577818,-0.3479522169,-0.1773992926,0.0666725039,-0.2951296866,0.4142337739,0.4732449651,0.4460688531,0.2521967292,-0.0001230383,-0.2152750939,0.1302850842,-0.317743361,0.2301827669,-0.4261970222,-0.2667488754,0.0393679813,0.2027443647,-0.1260904372,-0.3113471568,0.5966160893,0.2558301985,-0.2117690146,0.6537800431,0.0716565251,0.0451868847,-0.1689057946,0.2814650238,-0.0095189353,0.1461712867,0.1682046354,0.051208362,-0.075174652,0.6243174672,-0.3600681424,-0.229109928,0.1985415518,0.0063695475,-0.0540356338,0.2797682583,0.2602438331,-0.0144231571,-0.3843044043,-0.2895087898,0.4200867116,0.5511063933,-0.3132858574,-0.4524233639,-0.2447676212,0.2225971371,0.1461891383,0.0692787766,-0.2382965386,-0.1358793378,0.217027247,0.0773830041,-0.4006276727,-0.3453338146,-0.031248413,-0.105140999,0.3863505423,-0.0517031439,-0.1975738704,0.3232114017,-0.1545153111,-0.038774088,-0.0306088217,0.0835926607,0.5048011541,-0.2043062598,-0.4376601279,-0.3445557356,0.2067991197,0.0553344972,-0.1207953691,-0.2273675501,0.0705550388,-0.4263335764,-0.1015718356,0.4228085577,0.2061456591,-0.0779852644,0.165561229,-0.2637424171,0.032218229,-0.2298651189,-0.1523415446,-0.0986558348,0.0267325249,0.0893845111,-0.0812669396,-0.2630622089,-0.2469582409,0.1177515984,-0.0168839917,-0.2530955076,-0.0311104003,-0.0024541211,-0.0078826984,-0.2393765599,0.3082728386,0.0032434908,0.1589819491,-0.0273883268,-0.0793853626,0.1286915392,-0.2796353102,-0.2976554036,0.521655798,-0.0006238387,0.0101393722,0.1354135424,-0.1380978823,-0.0575402379,0.0933877677,-0.3288878202,-0.3122949898,-0.1033507437,-0.1362433732,-0.4028102756,0.1046491563,-0.0766845122,0.4250044525,-0.0973897427,-0.01419181,-0.2520745397,-0.028616054,-0.1406524479,-0.0588148199,-0.1818667501,-0.315163821,0.1049203873,0.533912003,-0.0480680913,-0.2282016426,0.0822864249,0.2302295715,-0.409604907,0.1246807352,0.1977958977,0.2521820068,-0.4910140336,-0.3078060448,0.073643513,0.1320023239,0.0859720558,0.0172373969,0.1739121825,0.1844780445,0.0657802522,0.1815827936,-0.1284930259,-0.1241852939,0.0365569219,0.0787436441,0.0417461954,0.4049219191,-0.1343845278,0.0324973091,0.0313700065,-0.1705056578,-0.0235561207,0.1146536097,-0.1572643518,-0.1915020794,0.1869989634,-0.1793524176,-0.0607713647,-0.0713655651,0.0958517864,0.2079959661,-0.3159270287,0.5599302053,0.1924262047,-0.0377030969,0.0784757808,0.0126629313,0.0805470347,0.491473645,-0.0041995738,-0.346273303,-0.1972341686,0.0738095418,-0.043400038,-0.0861883387,0.0072639431,0.0966252461,0.023781158,0.096205242,-0.0591682643,-0.2067830861,-0.0411678143,-0.1293233037,-0.1154093295,0.0758733824,-0.374376595,0.1479420513,0.1461742371,0.1246055812,-0.1697932929,0.2056616843,-0.0149310539,-0.0443587676,-0.0572474971,0.1798478514,0.0025718522,0.0809474736,0.1336023957,0.1328283846,0.1827095449,0.199707374,0.0009612282,-0.1804948598,0.1244325712,-0.3619657159,-0.1211431101,0.0214708112,-0.1050131246,-0.129161045,-0.1315614134,0.1197164729,0.4141704142,0.0427773632,-0.1037675217,0.0138473157,-0.1520467401,0.0279948413,-0.0964471102,-0.3477111459,0.0861943439,0.0540065132,0.258962065,-0.0506213121,-0.1999686807,-0.3999041319,0.2041917741,-0.0999379456,0.0958664268,0.6327774525,-0.1422082782,0.0661653429,-0.130759865,-0.2718037069,-0.014897678,0.1047483012,-0.0653653517,-0.091326572,-0.0169510506,0.0341923162,-0.1671380699,0.0311099086,0.1905682385,-0.2052601129,-0.2006826103,0.1428112537,0.209374696,-0.2070282102,-0.1485776454,-0.019425666,-0.1435665786,0.0612083487,-0.3151902854,-0.4285087585,-0.1279743612,-0.4911929071,-0.1034004763,0.2582746744,-0.0149435019,-0.3871529102,0.1514899135,-0.0499115773,0.1510235369,0.2841986716,0.194639042,0.1277811825,-0.2070503831,-0.2238790989,0.0237685461,0.0494989716,-0.0592209361,-0.1773704588,-0.2605481744,-0.0906507373,-0.0399954878,-0.0956390277,-0.3120402396,0.2897357643,-0.2499188334,0.1297583282,0.2941756546,-0.0759723112,-0.2079925537,0.0435942486,-0.0186340138,-0.0096070059,-0.0767076239,-0.2290391922,0.2426138222,0.2458166778,0.1850925982,-0.0338907242,-0.2169891149,0.3477907479,0.3525303006,0.1444896609,-0.0369680151,-0.1149627641,0.0262422953,0.1613724828,-0.3884990513,-0.164185375,0.1081548929,-0.5920629501,0.1830576956,0.1576792449,-0.0129233748,-0.3662762344,0.0132029438,-0.1370257884,-0.3882237673,-0.0276667308,-0.3468875289,0.096284017,-0.0810795054,-0.0687437728,-0.115794912,0.045665957,0.115896225,0.143484056,-0.1767230183,0.0795535147,0.3418131769,-0.0882980153,0.3423989415,0.0596097447,0.1252828985,-0.1889260858,0.4627276659,0.1235032305,-0.2810504138,0.3817255497,0.1742737889,-0.0153120877,-0.1077783331,-0.0047323238,-0.057310883,-0.0584741272,0.1513936967,0.3700029552,0.0191862807,0.0558755212,-0.1386035234,-0.1250238121,0.0858004168,-0.1367910504,0.0801763237,-0.265797466,0.1514858603,-0.1907012761,-0.1260736585,0.0377775021,-0.1597577929,-0.0473721772,-0.0175876357,0.7697459459,-0.1035643294,0.5713800192,-0.0400644019,-0.4635260701,-0.0110367872,0.3162865937,-0.1676734388,-0.1123738736,0.121964775,0.0061175372,0.1712811291,0.1437922567,0.0257674735,-0.0948603302,-0.1573239565,-0.3071129322,0.2070153952,0.2726218998,-0.2748065591,-0.3259212971,0.2437614352,0.2212954462,0.1280037165,-0.1737731248,0.2057458907,0.073568143,-0.09856195,-0.4630938172,0.0339650288,0.7298446894,-0.1961898804,0.0703766793,0.0374593884,-0.1456705183,0.127628237,-0.2202310413,0.2877981663,0.088348195,0.0621092319,0.045333188,-0.0263223462,-0.2703903615,0.0158028509,0.3881349266,0.041640535,0.1126371175,-0.0956169888,-0.0024821796,-0.2261900008,0.0859417617,0.1339311898,0.4489901364,-0.0524886586,0.1700175554,0.4762187302,0.0810796544,-0.018765796,-0.2345959246,-0.0024600353,0.6682594419,0.3292628527,-0.4029551744,-0.2895724475,0.5090538263,-0.1143748686,-0.2380094677,-0.4068022072,0.6209004521,-0.3231641352,0.1940322816,0.4422944486,0.8429120779,-0.0886928886,0.3112946451,-0.354575634,0.0624652095,0.7213706374,-0.1341113746,-0.1559299231,0.0058311149,0.2178891003,-0.3336976767,0.2488537133,0.0343312137,0.2194342464,-0.657143414,0.3025606275,0.1946094036,0.0111523913,0.1594286561,-0.0642634258,0.4093576968,-0.4654178619,-0.0820050165,-0.0213495158,0.166569069,0.270544678,-0.145581454,-0.1580580324,0.4408813715,0.0572816767,-0.5037370324,-0.0270970091,-0.2786689103,-0.4037694335,0.0886403248,0.0015841835,0.6850602031,0.4200372696,0.5400064588,0.2962406576,-0.358351469,-0.0038297719,0.0752435103,-0.0949679986,-0.2911288738,-0.0247940868,0.5104466081,-0.1082377136,-0.3413871527,0.2263132781,0.079521589,-0.306178242,-0.2628553808,-0.2435480654,0.5197245479,-0.0082630506,-0.0014372009,0.0698679313,0.1954891533,-0.1484988183,0.0564693063,0.147427544,-0.2316266745,0.0120894294,0.1262312084,-0.2625083327,-0.0406861193,0.1699935049,0.074420698,-0.1156331226,0.5515834093,0.3813837171,0.0235719681,-0.1429357529,-0.0415313058,0.2874449193,-0.3117373586,0.1730520129,0.1424445361,0.1089017838,0.119654417,0.560464859,0.3167203963,-0.2096678019,-0.1709754914,-0.0789635926,-0.4202402234,-0.1521322727,-0.1382852495,0.0889523029,0.2018940747,0.226287961,0.4319854975,0.4812993705,-0.1335812658,-0.1865675002,-0.1568321139,0.1854759604,0.1668442786,0.0507207699,-0.0354780965,-0.0340177864,-0.0489807725,0.3479300439,0.2405827641,-0.031507805,-0.1657157391,0.1171114072,-0.0828163251,-0.0193191506,-0.1044514105,0.1295056641,0.0014496319,-0.0560880005,-0.1370194703,0.1237348989,0.3126097918,0.3356466889,0.4685325325,-0.1630620956,-0.3651550412,-0.235574767,0.2533028722,0.0607485212,-0.0653114244,-0.1176358536,-0.0481865741,-0.1056725383,0.0576904044,0.2247026265,0.5772359967,0.0416007712,0.0999538302,0.100092791,-0.1463627815,-0.079378441,0.1456748694,-0.1524339765,0.2749128342,-0.2594358027,0.6059514284,0.0132495994,-0.2488042712,0.121942386,0.1560313851,0.071245417,-0.0755531564,0.0563122258,-0.1663425714,0.4083591402,0.311874181,-0.1110294834,-0.3372121751,0.1957022995,0.0659356788,0.3942149878,0.246223554,0.1359488964,0.2396061122,0.1566779763,0.4447750747,0.4075518548,-0.2255685627,0.1009693965,-0.2790302932,0.2017891705,0.5044291019,-0.1668917537,-0.2003413886,0.1333300769,0.1274654269,0.2278030962,-0.0769983903,0.7284395099,0.1680295318,-0.2433644086,-0.220365569,0.3128042817,0.0742322057,-0.4544348121,-0.1270726025,-0.0152986525,0.1933587939,-0.1192289591,-0.0843903124,-0.0199441276,0.1979461759,0.2530480623,0.3101106584,-0.185070619,-0.1399118006,-0.0365399495,0.3561722338,-0.0465296283,0.2955493927,-0.1366995871,0.0691889524,0.2079573125,0.347925365,-0.078551814,0.1551565975,-0.0947600827,0.1725800484,-0.1646720469,0.0695851147,0.2889030874,0.0571300723,0.1896821409,-0.0453448594,0.299015224,-0.0403401367,-0.0271157492,0.2549426854,0.0124206385,0.4369236827,-0.3256870508,0.1549673826,0.2438472807,0.3556319177,-0.3943241835,0.3008925915,-0.4989355505,0.4902075231,-0.382558912,-0.3178111315,-0.1649131626,0.231849581,-0.0218925755,0.0985028297,0.4464309514,0.6065987349,0.1910277456,-0.1123104468,-0.2440869659,-0.2296120822,-0.1221581474,0.1418815851,0.0993728489,-0.0558671579,-0.3151007891,0.0122434981,0.0761608407,-0.0800182819,-0.0388826504,-0.7266442776,0.1229676008,-0.2297173142,0.0446914062,-0.2524687052,0.2073400319,-0.0367170461,0.3225855231,-0.0386123247,0.2700695693,-0.1861474663,0.1864187717,0.0401762575,-0.223837465,0.0726817921,0.1357154399,-0.1165212393,0.1795517206,-0.1516773552,0.3388016522,-0.0210337844,-0.0172183383,-0.2530869842,-0.3071561158,0.1226312742,0.2526157796,-0.2492694408,0.1780521721,-0.0967136845,-0.1678788513,0.1459289491,-0.0184516031,-0.6037929058,0.0245391186,0.0723755136,-0.2292449623,0.1054957882,0.2302971929,-0.1658845991,0.036146313,-0.3356179595,0.0039449031,0.3455246687,-0.2045921534,-0.0846925452,-0.2976564765,-0.366068542,0.3209731579,-0.2705556452,-1.1808876991,0.0807430297,0.268093437,0.3806757331,-0.0940280557,0.1979695708,0.0488505773,0.0961834639,-0.3995482624,0.5143039227,0.3337180614,0.1103341803,-0.2128124833,-0.2657096386]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2129","title":"How to train BERT model with next sentence prediction?","comments":"Thank you for detail information.\r\n\r\nI'll try to apply `create_examples_from_document` to `Dataset` object.\r\n","body":"Hello.\r\n\r\nI'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction \r\nlike ` TextDatasetForNextSentencePrediction` of `huggingface\/transformers` ?\r\n","comment_length":13,"text":"How to train BERT model with next sentence prediction?\nHello.\r\n\r\nI'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction \r\nlike ` TextDatasetForNextSentencePrediction` of `huggingface\/transformers` ?\r\n\nThank you for detail information.\r\n\r\nI'll try to apply `create_examples_from_document` to `Dataset` object.\r\n","embeddings":[0.1811716557,-0.4566830993,0.006565928,-0.1272096634,0.0192352962,-0.261523217,0.1080397144,-0.0018368327,-0.0366002433,0.1279347986,0.1303460598,0.0927896798,-0.1462009698,0.0135782165,0.3786928058,-0.7038691044,0.1878689826,0.2664217949,-0.0669772997,-0.1529922038,-0.1063541174,0.1032557636,-0.4242304564,0.3705087602,-0.2825053036,0.1698747873,-0.374905616,-0.2637747228,0.0586312525,-0.3363387585,0.4275523126,0.3872366548,0.4913852811,0.2281727642,-0.0001244387,-0.2452990115,0.1568442285,-0.2868937254,0.264968425,-0.4218095243,-0.2380530238,0.0226640236,0.2535345256,-0.2136957198,-0.3207916915,0.6206229329,0.2915273607,-0.1096360832,0.5922269821,0.1877553016,0.0401318409,-0.2800793946,0.3021287024,0.0067301304,0.1251006722,0.1005316153,-0.0091366852,-0.1577246487,0.5044577122,-0.3580864966,-0.2305166274,0.169808194,0.0392113812,0.0275635682,0.1667278856,0.2506494522,0.1053283587,-0.3698679209,-0.2072779387,0.387635231,0.5949965715,-0.257031709,-0.3850871623,-0.0600074753,0.2547540367,0.1218117326,0.0491459258,-0.2246625423,-0.0690213144,0.1594689041,0.1069620773,-0.376309365,-0.496118933,-0.0041682506,-0.1037259847,0.3296584487,-0.0345238261,-0.2420344353,0.3455446959,-0.2042863071,0.0518041626,-0.0710135102,0.1039599255,0.4439158738,-0.1991324276,-0.3407089412,-0.397369802,0.1649733037,0.0136928875,-0.2152696699,-0.1782313734,-0.0048641469,-0.3135538101,-0.1254834533,0.3182000816,0.2394569218,-0.1729573905,0.1767034233,-0.2891677916,0.107267119,-0.2590835392,-0.1527149528,-0.1722375005,0.0756534263,0.3117355108,-0.038268134,-0.246142149,-0.2343745381,0.1223077253,-0.0034498787,-0.3257761598,-0.0032562574,-0.0833988115,-0.0059768581,-0.295537591,0.2428187728,-0.0478255078,0.0984518901,0.0129854772,-0.0798093528,0.1180169433,-0.1389218271,-0.3051178753,0.4588390589,0.0059402823,0.0254997127,0.1593932509,-0.0965481177,-0.0857239366,0.0541138947,-0.2407462597,-0.4481382072,-0.0297978055,-0.1461060345,-0.3871707916,0.0526975803,-0.1325249821,0.5027804971,-0.088059403,0.0102905734,-0.2132119536,-0.0772768557,-0.1150285751,-0.0386952534,-0.0411203392,-0.2566654682,0.150740698,0.515501976,-0.0916548595,-0.1471963078,0.1210726202,0.1657631099,-0.3950568736,0.1449167728,0.166337803,0.1587932259,-0.4179449677,-0.2513995469,-0.0008395452,0.1610341668,0.1418705434,-0.0859826356,0.2223476171,0.2173795849,0.0618507937,0.1582924426,-0.0919184536,-0.1188711077,0.0797252432,0.0392617583,0.1321941763,0.3491226435,-0.1580113918,-0.0436114594,-0.0302462038,-0.1043355614,-0.1165890843,0.1454109401,-0.1442205012,-0.1280111074,0.2452618331,-0.1400336772,-0.1353196949,0.0104630813,0.1926227361,0.1683061719,-0.2828564346,0.6158300638,0.1120114326,-0.0280702356,0.091986157,-0.0144784767,0.1494889706,0.3142777681,0.0620196499,-0.3569203317,-0.1595443338,-0.0404665098,0.0124195786,-0.0865056366,0.0886754617,0.184088558,0.0018678815,0.08739461,-0.0622179098,-0.1891113818,-0.0338794291,-0.0138874557,-0.1278494745,0.114015311,-0.3435776532,0.1307547539,0.1251713932,0.0783802271,-0.1355419755,0.1279728711,-0.1471309066,-0.0859823525,0.049512133,0.1084504724,0.1013617292,0.0657610223,0.1294570416,0.0795742199,0.2590523362,0.0944588855,-0.0602070242,-0.2575830221,0.0268258601,-0.4199830592,-0.0258153491,0.1000658646,-0.0803217962,-0.0927757919,-0.1394110471,0.0182149485,0.4471776485,-0.0622078814,-0.0254670084,-0.0144686997,-0.1473199576,0.0292514637,-0.0161241312,-0.3602469563,0.1300712377,0.0646016151,0.2297546417,-0.1018288136,-0.2513411045,-0.4636214375,0.2033274919,-0.1258050203,0.1772621572,0.6849727035,-0.0209969785,0.0480694398,-0.1376748234,-0.3577693105,-0.0364851654,0.0414680727,-0.0705837831,-0.1554187983,-0.0460425913,0.0809853598,-0.0809293091,-0.0902592763,0.2863870859,-0.1212273613,-0.2465887815,0.1103552952,0.2517444193,-0.3320311308,-0.1168969125,-0.028169034,-0.106060259,0.0220131669,-0.2987450361,-0.3965741396,-0.0487439595,-0.4322789907,0.0162918661,0.4456003904,-0.0610799305,-0.44146806,0.2204672992,-0.0357688181,0.0825056881,0.3077788949,0.2314265221,0.2130292505,-0.2422742546,-0.0695094541,-0.0116785849,0.0097391969,-0.2114513069,-0.1401646286,-0.3604177833,-0.0367612019,0.0466334037,-0.0401206315,-0.2550583184,0.3360816538,-0.1320139915,0.1041982174,0.2653329968,-0.1513865441,-0.1810452044,0.0714609697,-0.1070022434,-0.0156514216,-0.0463652797,-0.1217714027,0.1729326695,0.1469568312,0.2660102248,-0.0257939007,-0.2217928469,0.3535870612,0.2449139953,0.1877187192,0.0003566853,-0.1103744656,-0.00126172,0.1417571753,-0.2583549619,-0.0484793074,0.2034033984,-0.623003006,0.1823342144,0.1869464666,-0.004082175,-0.3724667132,-0.0993264318,-0.1377630979,-0.3688852787,-0.0043838047,-0.2822189927,0.1467647105,-0.1392562538,-0.142733857,-0.171497494,0.0223461036,0.1376890242,0.1250106841,-0.2032371163,-0.0107109332,0.3885644376,-0.0592871718,0.3515566289,0.0533846356,0.102897875,-0.2337778509,0.4081052244,0.0819106549,-0.2416945249,0.4071000516,0.216634497,-0.1856219471,-0.0908480734,0.0020262436,-0.0333244801,0.072255753,0.0753605589,0.4074221551,-0.1137890741,0.0086060623,-0.1681064814,-0.2162116617,0.0337748341,-0.0192334298,0.1026901379,-0.2277426869,0.266988337,-0.2573697865,-0.1588297188,0.0737556294,-0.1396293193,-0.0886135995,0.0325446427,0.7358711362,-0.078512162,0.6551982164,-0.1099136174,-0.3244145811,-0.0740558952,0.2622248232,-0.095755294,-0.1537629664,0.2326010913,-0.0760250017,0.0972555503,0.0098444298,-0.0225517247,-0.0123478882,-0.0817039385,-0.3444413841,0.3208887279,0.1906016171,-0.2813338041,-0.3292371035,0.2032600939,0.1859620512,0.1781931072,-0.1862403452,0.3138991892,0.04261484,-0.0937731564,-0.4952296317,0.0366252065,0.7614908218,-0.0672432035,0.0738816634,-0.0302911289,-0.1823798418,0.0727002323,-0.2734971344,0.2497668564,0.1376467645,-0.0413976684,0.0152588822,-0.0343430303,-0.3177923262,0.059377268,0.2268594652,0.0789654925,0.272146821,-0.2784755528,-0.0470010042,-0.2166517079,0.1058298051,0.1562308967,0.414962858,-0.2040316314,0.1351421326,0.420440197,0.0003670574,-0.015313833,-0.0650005192,-0.1044174358,0.7137170434,0.3134933412,-0.3192279935,-0.2611184418,0.525154233,-0.1504902095,-0.230277881,-0.4523173571,0.5873501897,-0.3055891097,0.2241383642,0.5311316848,0.7619830966,0.0203049462,0.1979234219,-0.3473212123,0.078142345,0.5547345877,-0.1853762269,-0.1516981125,-0.0136134857,0.2008116394,-0.3418283165,0.2175174952,0.0761607215,0.2416641265,-0.8018054962,0.3001826406,0.1458635479,-0.1258593947,0.1361566782,-0.0056172195,0.4247322977,-0.4614695013,-0.0447601527,-0.0285019334,0.1387549192,0.17625314,-0.1749585867,-0.0658468008,0.5325291157,0.0954301283,-0.387834847,0.0745729804,-0.2514553964,-0.3898224235,0.083858259,0.0031449378,0.7230706811,0.4111355841,0.5449239016,0.1791594774,-0.2948954403,-0.0372258946,0.1032967865,-0.1434627175,-0.3109311163,0.0069640856,0.5280860066,-0.1833196878,-0.3602490425,0.2307525277,0.0503159314,-0.2003135532,-0.1661245376,-0.2925085425,0.5922154188,0.0326375104,-0.0163127575,-0.0219844095,0.1811494976,-0.1217991337,0.0735211372,0.1083266437,-0.2215813845,0.0673815981,0.1548199207,-0.2752067447,-0.0251827631,0.0284532942,0.0957800522,-0.0192645397,0.5287688375,0.2868987918,0.0585232154,-0.1985492408,-0.0385400355,0.271299988,-0.1541177928,0.2003815621,0.1282168627,0.1479376704,0.2026077062,0.6068410873,0.2861301601,-0.2601397634,-0.2041834146,-0.0237019751,-0.3021211028,-0.2560712993,-0.148811087,0.1279230267,0.2381844074,0.0704521239,0.5135133266,0.4863113761,-0.1156639978,-0.2066729218,-0.2112158239,0.191837579,0.0971789882,0.0400932431,-0.1305995286,0.0544069,-0.0467154495,0.2646773458,0.2783359885,-0.0250598975,-0.1727414131,0.1034421921,-0.0723844245,0.022973394,-0.1418528557,0.0979154408,-0.0896638557,0.0343551859,-0.2344052643,-0.0423008166,0.2570245266,0.4769190252,0.4425982535,-0.1961592436,-0.3698067069,-0.2818663716,0.211040467,0.0480473489,-0.0621569268,-0.079240188,-0.0082104467,-0.1964205801,0.0218582563,0.1872855574,0.4955450296,0.0943828151,0.1224743724,0.100303039,-0.098499395,-0.2097967863,0.0530039631,-0.2545681596,0.1791661382,-0.2339473814,0.5599741936,0.0094187614,-0.2441125363,0.0148609579,0.046328444,0.0471090637,-0.1058669686,0.0208949652,-0.2029547244,0.3632281125,0.3305657804,-0.0972993225,-0.3911981285,0.2316498607,0.0034216053,0.4124136269,0.2030699253,0.2333666235,0.2647662163,0.2283283323,0.4502506852,0.4563674033,-0.099526085,0.0760387108,-0.2041584253,0.208658129,0.5304998159,-0.1609071344,-0.1471225023,0.0539724194,0.0334300809,0.1513593942,-0.0308621749,0.6187849641,0.260887444,-0.1995937377,-0.304286778,0.1719257236,0.0921044201,-0.3985320032,-0.1008523405,0.1032182127,0.2234949768,-0.0848533586,-0.0341350548,0.0631725937,0.1100843102,0.3399848342,0.2729272842,-0.0571726002,-0.0836735815,0.0677673444,0.3318592906,-0.0495354421,0.3022405803,-0.1372508109,0.0884241834,0.0208009612,0.3400118649,-0.0665801167,0.234180063,-0.1245904118,0.1382945925,-0.0115445992,-0.0075617419,0.3096675575,0.0320998169,0.2338070273,-0.0086330334,0.2170799226,-0.061530605,-0.0088711428,0.335991919,0.0609672889,0.3417721391,-0.2975010574,0.1233410984,0.3559040427,0.3738325238,-0.4083286822,0.3010285199,-0.4700013995,0.5272444487,-0.3335236907,-0.309569478,-0.1603513211,0.2285316139,-0.0230275989,0.0393306427,0.418566376,0.4727056623,0.2345810682,-0.086101003,-0.3081861734,-0.0784481391,-0.1549714357,0.0940550715,0.0235689692,0.1038743705,-0.2679848671,-0.0829344168,-0.0023390078,-0.0420152023,-0.0145664839,-0.6489328742,0.2693020701,-0.2401760519,0.0077435118,-0.2109990567,0.2610618472,-0.076932393,0.2202469409,-0.0240137707,0.2507861555,-0.1824483871,0.2121575773,0.0203036908,-0.1288504452,0.1722975671,0.1894010752,-0.2252983451,0.1688458771,-0.23914285,0.4226031601,0.1150023416,0.1091992632,-0.24709557,-0.351552397,0.2214273363,0.1792999208,-0.1757946908,0.2033767253,-0.0533719286,-0.1928163469,0.141784057,-0.0420697108,-0.5694050193,0.0530698672,0.1253872365,-0.2448327094,0.1572689563,0.1471282542,-0.1768380553,-0.0545228384,-0.3798387647,0.058580365,0.2941043377,-0.0732751638,-0.1013696119,-0.3901349604,-0.3399846852,0.2296010852,-0.2471867651,-1.1346590519,0.1351636648,0.3023927808,0.3247350752,-0.0992515683,0.2861013114,0.1304363012,0.2118569165,-0.3610820472,0.5224041939,0.3709828258,0.1511571109,-0.3196524382,-0.2466977239]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2128","title":"Dialogue action slot name and value are reversed in MultiWoZ 2.2","comments":"Hi\r\nGood catch ! Thanks for reporting\r\n\r\nIf you are interested in contributing, feel free to open a PR to fix this :) ","body":"Hi @yjernite, thank you for adding MultiWoZ 2.2 in the huggingface datasets platform. It is beneficial!\r\n\r\nI spot an error that the order of Dialogue action slot names and values are reversed.\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/649b2c469779bc4221e1b6969aa2496d63eb5953\/datasets\/multi_woz_v22\/multi_woz_v22.py#L251-L262","comment_length":23,"text":"Dialogue action slot name and value are reversed in MultiWoZ 2.2\nHi @yjernite, thank you for adding MultiWoZ 2.2 in the huggingface datasets platform. It is beneficial!\r\n\r\nI spot an error that the order of Dialogue action slot names and values are reversed.\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/649b2c469779bc4221e1b6969aa2496d63eb5953\/datasets\/multi_woz_v22\/multi_woz_v22.py#L251-L262\nHi\r\nGood catch ! Thanks for reporting\r\n\r\nIf you are interested in contributing, feel free to open a PR to fix this :) ","embeddings":[0.4248688221,-0.3456932008,0.017481707,0.4803560972,-0.2102743387,-0.0071412404,0.1827978194,0.1119403541,-0.1007885039,0.2386501282,-0.3579909801,0.1305642724,0.023053579,0.392791599,-0.0238381848,-0.2758994401,0.0942913964,-0.0138568273,0.0493680164,0.1311452091,-0.4356461167,0.5747422576,-0.5933912992,0.6446860433,-0.1071594134,-0.014253404,0.0634024963,0.1111348197,-0.0194266532,-0.0573899932,0.0214845669,0.086402446,-0.0686788261,0.3353702724,-0.0001161957,-0.2951157689,0.2519213557,0.1138677374,0.0176503081,-0.0307998043,-0.264221549,0.0847944692,-0.1248991862,-0.1370581985,0.2321756035,0.1895438582,-0.0746237189,-0.0801590085,0.2754888237,-0.0784257203,0.1157124788,0.1648204774,0.0357437097,-0.0580541827,-0.021584345,0.2198867649,-0.0824432895,0.1995959878,0.0212184414,0.1268387884,0.0749811083,0.3417924643,-0.2010028362,0.0132200038,-0.1365127265,0.1374370456,-0.0064535835,0.0807196572,-0.0942605808,0.3944839537,0.1280011237,-0.2043231577,-0.2542478144,-0.3739617765,0.2252251953,0.0107932147,0.2898096442,0.3622597754,0.1570699662,0.2073945403,0.0218984596,0.2123642266,0.2015572935,-0.1239941493,0.0848845616,0.3029039204,-0.1620901823,0.087652415,-0.3311100602,-0.2417797595,0.0060684457,0.2482507527,-0.2515794635,0.1156895831,-0.4204523861,0.2017872632,0.0330352709,0.3425073624,-0.00988492,-0.119001694,0.2433529496,0.1067766324,-0.1916660964,-0.1237335205,0.2064361125,0.3674930632,0.3211267591,-0.254106462,-0.1370896995,0.2693676651,0.2884452343,0.0714496598,-0.0131935403,-0.5080451369,-0.2649772465,0.0574989282,0.3858119845,0.0727554187,-0.1866700053,-0.1129761487,-0.2294873446,0.0604380444,0.6087936759,0.1458617449,-0.3165177107,0.4559125602,0.182823047,0.2985946238,-0.0327907093,0.0739530623,-0.0395368226,-0.2116193473,-0.2245818377,0.2388994247,0.1142526194,0.4980134666,0.2482367754,0.256526649,0.6652235985,-0.0223227236,-0.049070809,0.1431645304,-0.0727885142,-0.2191022187,0.198055163,0.5540353656,-0.087507233,0.2238435894,0.0393962264,-0.0764862001,-0.5501940846,-0.1435861886,0.1170959175,0.1250051111,0.0781774819,0.1016248092,-0.0038582888,-0.1853638738,0.2374021411,-0.1618606448,0.1959500909,0.1565700322,-0.2136678249,-0.117701456,-0.1799676567,0.033535827,0.2766271532,0.0454835296,0.1398204863,0.0260129627,0.0869095623,0.2126455754,-0.161983788,-0.1217643097,-0.3524559736,0.3236255944,-0.1048871949,-0.5939317942,-0.0289269723,-0.0806844532,-0.3253728747,0.0964092091,0.2077775896,-0.0007428674,-0.0522921421,-0.2352713645,0.3298335373,0.05404092,0.257714808,0.2721787691,-0.1492700577,-0.2556462586,0.4765302241,-0.30493173,-0.2597441971,0.0622302182,-0.3183143735,0.1614961773,0.5554067492,-0.024431359,-0.1462979168,0.0587144643,-0.1342418045,0.1856333464,0.4008899629,-0.0592837818,-0.2871947587,0.1710009575,-0.7374218106,-0.1946219355,0.2939944267,0.0273175333,-0.0290292259,-0.0295994636,-0.3857558668,-0.1096083075,0.1072959602,0.0746717602,-0.1831603497,-0.3749780953,-0.207635507,0.7400954366,0.0544771738,0.1494110078,-0.0958260372,0.0385058634,-0.001846522,0.0911564082,-0.02874461,0.30574283,0.4141573012,0.1709284931,-0.043063581,0.6762667298,-0.0163328517,0.2083843648,-0.0231600385,-0.2112208307,0.0802685916,-0.0880829468,0.2206207365,0.0709040761,-0.1941143721,0.1024219245,-0.1042685583,0.6511005163,-0.2412299812,0.0029580537,-0.0778319389,0.0669838712,-0.0248160791,-0.0509845242,-0.4145665765,-0.2964108288,0.1980013847,0.0032161677,-0.1446511894,0.1842003465,-0.4677919447,0.0189950932,0.5268551111,0.0036459549,0.1161582619,-0.1344466954,0.0493617319,-0.1965816617,0.140163213,0.3095352352,0.2136596888,0.1553581655,-0.2179352343,0.0333869159,-0.1488153785,0.0075613973,0.2790917456,-0.1165453196,-0.0671329126,0.1257618219,0.1785932779,0.0813031644,-0.0708084702,0.5163924098,0.2012716979,-0.1768668741,-0.4543136358,-0.4007063806,-0.3898670673,-0.2114553601,-0.3471463323,-0.5310077667,-0.2130122185,-0.0491226614,0.0900170654,-0.1668141931,0.1789435893,0.216297552,0.2227875292,0.0556248501,0.0094747059,0.4551240504,-0.1428201497,0.183876887,0.0256185047,0.0008189881,-0.3795509934,0.2141218632,0.0089382753,-0.1296350062,-0.2861487567,-0.220164299,-0.3470645547,0.1366154253,0.0111505929,-0.233149007,0.1379038841,-0.0610333905,0.3058896363,-0.3460125327,0.2869271338,0.1975712776,-0.0733103678,0.2513197064,-0.0523500293,-0.1882897019,-0.2862944305,-0.0519899428,0.1491796225,-0.0457237214,0.5657224655,-0.0610188767,0.1535643041,0.2129755169,-0.4396947622,-0.1446075141,-0.38625741,0.3639066219,-0.3491976857,-0.3759961724,0.2740384936,-0.2382833362,-0.0076247258,-0.287404418,0.080244109,0.0812307075,-0.2473169267,0.2349969,-0.4288898408,-0.2434288412,-0.2202375084,0.2920053005,0.014921573,0.1497980207,0.1676976383,0.037194714,-0.2795480192,-0.3579645157,0.4529719055,0.4093859196,0.3168850839,0.1465575844,-0.3183181584,0.1441505998,-0.1315051615,0.1915724128,0.4237494767,-0.0403439254,-0.0546637811,-0.1575858295,0.0940183401,-0.4223298728,0.4233852625,0.0929758996,0.2533917129,0.1302949637,0.1842269301,-0.005336741,-0.0324562937,0.2600782514,-0.5889777541,-0.2407675534,-0.4462688267,0.1636603773,-0.0769674405,-0.0008846633,-0.1493411809,-0.3815714419,-0.0077782031,0.3646033406,-0.0682364553,-0.1093359143,-0.2107598633,-0.2921125889,0.0683451816,0.0984128863,0.0979407281,-0.0030865585,0.0732307136,0.2095306367,-0.0111974096,0.0909686089,-0.072290279,0.8490325212,0.3808697164,0.158961609,-0.1371571869,0.0585924834,-0.3818630874,-0.3467120528,-0.2079803497,0.2196664959,0.1029151976,0.225872457,0.0135820452,-0.3630976081,0.382904619,-0.426043123,-0.1406198889,-0.2655020058,-0.027305223,0.3439027667,-0.1893531829,-0.4021939933,-0.1628113687,0.1936210245,-0.0712733716,-0.2060952038,0.287833035,0.0196364392,0.1206153408,-0.0654525608,0.2733005881,0.3706089854,-0.0706125721,0.1210651249,0.2503660321,-0.1985100806,-0.0920987874,0.2520002127,-0.1673197746,-0.1930326223,-0.3155900836,0.0780280456,0.5454142094,-0.3548977375,0.1936652958,-0.059113171,0.3458314538,-0.189355582,-0.1768037677,0.6700114608,0.2171916962,0.0598325096,-0.2858707607,0.0723928884,0.0444633067,-0.3967036307,0.160186097,0.3718762398,-0.227028951,-0.3977589011,0.0656137541,0.9026181102,0.0739224255,-0.0204972774,0.5355348587,-0.2725102007,0.187436074,-0.0083452975,0.014412988,-0.0275004338,0.2743515074,-0.0258421432,-0.0873967856,-0.0741574541,-0.0525762476,-0.3574902415,0.1298758239,-0.0773587674,-0.2054112405,-0.3921938539,-0.2816848457,-0.0590481162,-0.4611425698,-0.1275652051,0.0406200327,0.0264171734,-0.0352798365,-0.1628308147,-0.1320158392,0.1165287197,-0.053740263,-0.2162316889,0.0781160817,-0.6180945635,-0.1173079982,0.3643268049,0.144109562,0.0489344522,0.7493516803,0.3992188871,0.1004937664,-0.1019064859,-0.0346710421,0.0380587727,0.1468992531,-0.2057351768,0.2596022189,0.0500523336,-0.0948388278,-0.2738461494,-0.3900847733,0.2031746358,0.2213356793,-0.2702949047,0.1476627588,0.1704065502,-0.1550690383,-0.13918975,0.2486215085,0.3979614973,-0.1134906784,0.0703733191,0.032525219,-0.2852581739,-0.0750052929,-0.1604664624,-0.1852905154,-0.0326347761,0.3739660978,-0.0991980359,-0.2063402534,0.2567950785,0.3225004971,-0.0970663875,0.0443217047,-0.0425853133,-0.227540344,-0.1387375593,-0.0276308209,0.2397423387,-0.3784756064,-0.1364806741,0.0608389974,0.2178943008,0.3661232293,0.0013343547,-0.1555372179,-0.1057618782,0.2193027288,0.238652274,-0.2483076304,-0.1474938393,0.022912778,0.2719471157,-0.0955517069,-0.2666562498,0.225019753,-0.0137287583,0.1234964728,0.0806684569,-0.2712791264,0.0472587384,0.0140392147,0.1686256528,-0.2445572019,-0.0775929615,-0.1566593647,-0.2303319126,0.0986044109,-0.105973348,-0.1805461049,0.2839008868,-0.3058714867,-0.0257989094,-0.1783965379,0.0310361944,0.5340154767,0.1980638802,0.3427356184,0.0737948492,-0.2695972621,-0.054884363,-0.0169344489,-0.1495076418,-0.2583514452,-0.1543706656,-0.0108810207,-0.3470697999,-0.0255424343,0.20651187,0.1599380821,-0.0566163622,0.1992516369,0.0756699592,-0.4159165323,-0.0555665232,0.2746032476,0.3202625513,-0.2126526982,0.1599392891,-0.0379318185,-0.0689276606,0.1196051762,-0.1338181198,-0.1015196443,0.1273588836,-0.1396164596,-0.1532641351,0.0211792942,-0.4035213888,-0.0985614583,0.0096171414,-0.0531235449,0.0245269164,-0.3315933049,0.1222826242,0.0853318349,-0.1164043546,0.2402065694,0.3083595335,0.116484195,0.1611391902,0.1334272623,0.3029757142,0.0237994399,0.0805646703,0.4619534612,0.4842875302,-0.1208141372,-0.4640147686,0.2537319958,-0.2379836887,0.1630959958,0.1570443213,0.6198300719,-0.1227156073,-0.0865099579,-0.393435657,0.3281670809,-0.2067819685,-0.3310202062,-0.2828932405,-0.0322409794,-0.1390275806,-0.2844771147,-0.0741893351,0.2360834479,0.4354280233,0.185791254,-0.1263843775,-0.5862572789,-0.2476723343,0.1614829302,0.1531927437,-0.2323200554,0.2458814383,0.1750435531,-0.0998743251,0.2778573334,0.4051599801,0.1902062893,0.3123472333,-0.0923781395,-0.036278557,0.080068402,0.0430137627,0.0009690753,-0.1873481572,0.2856108844,0.2895578146,-0.0567991808,0.0739884302,-0.1955512762,0.1364593953,-0.0608531833,0.2052821666,-0.1229842082,0.3780428171,-0.4240855873,-0.0320691057,-0.2460831255,0.0847502351,-0.4977919757,0.0217527412,-0.1338019967,0.0879800916,-0.0383844823,0.1069188043,0.0677048415,0.3495083153,0.2672064304,0.203305155,-0.0520670079,-0.0190252028,-0.0816775933,-0.2635883987,0.0648213178,0.5914783478,0.4455241859,0.1026989296,-0.0389801301,-0.0654779524,-0.2587243617,0.3318720758,-0.0806489289,0.0748645887,-0.0159796178,-0.3904719353,0.3514471352,0.4120182693,0.1034909263,0.2982913852,0.0427882001,0.1380146593,0.0990965366,-0.0682734028,0.1041040197,0.3699242175,-0.2538019121,-0.4620382786,-0.0537601151,0.0715959966,-0.1099158451,-0.0680898502,-0.0628550127,-0.0639532208,-0.0505758747,-0.1750563085,-0.2207097411,-0.3777935207,0.3229448795,-0.2659549415,-0.0538450032,-0.1311865151,-0.1632922292,0.0874988511,-0.1817830652,-0.4835101366,0.4868098199,0.1875004619,-0.0776985809,0.1575503498,-0.0747743621,-0.0410982184,0.243616581,-0.0356815197,-0.2809416056,0.5687158108,-0.252186209,-0.0559182502,-0.1319849342,0.4510646164,-0.3866219819,-0.1106122434,-0.2237984836,-0.0065835943,0.0902521908,0.173641026,-0.1896889806,0.2094055116,-0.1601656079,-0.2573341727,-0.1231333166,-0.009901437,0.5217808485,-0.3492459655,0.2818293869,0.1111744046]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2125","title":"Is dataset timit_asr broken?","comments":"Hi,\r\n\r\nthanks for the report, but this is a duplicate of #2052. ","body":"Using `timit_asr` dataset, I saw all records are the same.\r\n\r\n``` python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n\r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\n\r\nshow_random_elements(timit['train'].remove_columns([\"file\", \"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \r\n \"sentence_type\", \"speaker_id\"]), num_examples=20)\r\n\r\n```\r\n\r\n`output`\r\n\r\n\"Screen\r\n\r\n\r\nI double-checked it [here](https:\/\/huggingface.co\/datasets\/viewer\/), and met the same problem.\r\n\r\n\"Screen\r\n","comment_length":12,"text":"Is dataset timit_asr broken?\nUsing `timit_asr` dataset, I saw all records are the same.\r\n\r\n``` python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n\r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\n\r\nshow_random_elements(timit['train'].remove_columns([\"file\", \"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \r\n \"sentence_type\", \"speaker_id\"]), num_examples=20)\r\n\r\n```\r\n\r\n`output`\r\n\r\n\"Screen\r\n\r\n\r\nI double-checked it [here](https:\/\/huggingface.co\/datasets\/viewer\/), and met the same problem.\r\n\r\n\"Screen\r\n\nHi,\r\n\r\nthanks for the report, but this is a duplicate of #2052. ","embeddings":[-0.0864459649,-0.2443610728,-0.1574445516,0.4346131086,0.2811744809,-0.0320198238,0.3418569267,0.2223212719,-0.1953612864,0.1805195063,-0.1039599925,0.253262192,-0.2609292567,0.170999527,0.0706091151,-0.1757813096,-0.1363489628,0.0410866924,-0.2023348361,-0.3297544718,0.0239939317,0.0907824114,-0.3317570984,0.0330468565,0.0272830296,0.2073672116,0.1165766194,-0.3478392065,-0.1344901323,-0.5642359853,-0.0524484776,0.0100956922,-0.0705477372,0.5825358629,-0.0001125055,-0.0404750444,0.131135866,0.0386235788,-0.1782006323,-0.2203257084,-0.2623075545,-0.0327549428,0.0084048873,-0.0051329061,-0.0202647056,-0.1709392667,0.1297937036,-0.1705719978,0.0058129742,0.4632574916,0.2381157577,0.3627276719,-0.0282105077,-0.0081764515,0.2548128664,0.1993289888,-0.0354686268,0.1304604858,-0.0160881933,0.0898343772,0.0584133416,-0.063235119,0.0664299801,0.4422768652,0.1682106853,0.0355607755,-0.2900030315,-0.3493136764,0.1104457378,0.4285259247,0.5643061996,-0.1547569782,-0.0485515259,-0.0634842217,-0.1352199167,-0.2255717069,0.0758167356,0.4280950427,-0.3700413406,0.1879719049,-0.3344534636,0.4435712695,-0.0282766707,0.449079901,-0.1262289733,-0.2517416775,-0.117681928,0.1361628622,-0.1599470675,0.027076656,0.1574308276,0.0241248906,-0.0323786139,-0.1641057283,-0.1163189188,-0.1178775877,0.0288973898,-0.0330343992,0.3292217553,0.2917673588,0.5199828744,0.0568303317,-0.227605179,-0.0899850056,0.2013644874,-0.1034285501,-0.0476598255,0.1354500651,0.0782987103,0.0437692143,-0.2351275533,0.0412333645,0.2357629389,-0.162827462,0.2317268699,-0.3880986273,0.2661204338,-0.1144302487,-0.5204657316,0.1775119007,-0.2190032601,-0.149691999,-0.1537780315,-0.0226347949,0.0523561388,0.1964841336,-0.0300342813,0.1232212409,-0.1145664752,-0.2157569677,-0.1990878433,-0.0570910238,0.0965650901,-0.0102621419,0.028264707,-0.1216310933,0.1630016267,0.2737435102,-0.0451363213,-0.1191575527,0.1176274419,-0.217988953,0.3810287714,0.1698542535,-0.0202935133,0.3295325935,0.1251344979,-0.1745572984,-0.0669185892,0.1669266522,0.1027479321,0.1278280914,0.0612640902,0.1892178208,-0.0664387122,-0.2454702556,-0.2965540886,0.2373639792,0.0953956172,-0.0552498475,0.3006780744,-0.1960556209,-0.4821628332,-0.1077155247,-0.0011138848,0.3073503971,-0.546114862,-0.0379681401,0.195615083,0.2569815516,0.1737464964,-0.0088896416,-0.1072838753,0.2527164519,-0.385460943,-0.1786682308,-0.0548376217,0.0298563801,-0.5264158845,-0.0519389212,0.1205973923,0.0428814292,0.0571133196,-0.0018367317,0.221132502,0.3147023618,0.1224644184,0.0308612753,0.0845292211,-0.1178953797,-0.3046291471,-0.0202998891,0.2798672915,-0.0267863031,0.1082807779,0.1864746064,0.1516290307,-0.0179631412,0.3041369915,0.2192090154,-0.1059229076,0.0931760594,0.061039377,0.0719918385,-0.0760785937,-0.3043048382,0.1549986005,0.0611109883,0.3125684857,0.3011164367,0.1049447358,0.048001878,-0.2941234708,-0.2452527881,-0.2467855215,-0.152209729,0.1460535675,0.4637942016,-0.0220617168,0.0664738044,-0.1401007026,0.3726413548,-0.3047705889,-0.0282638427,-0.4319956005,0.4172754288,-0.061429996,-0.0500411615,-0.0036429183,-0.2354686558,-0.0901411995,-0.0968032554,-0.1601571441,0.312921077,0.308836937,-0.2328247577,-0.1456055939,-0.2556965053,0.3211739659,-0.6220970154,-0.1399717629,0.3905026019,0.0921237171,0.030257009,-0.1150545403,0.1637626439,-0.2727924585,0.1507478505,0.2286612988,0.2957230508,0.0333484076,0.1808082312,-0.170718208,-0.1718964577,0.1902545393,-0.0723187402,0.0715330839,0.1091762185,-0.1962137222,0.2316340506,0.4030924439,0.2020916492,-0.0273811035,0.2894678712,-0.1090172604,0.2302659452,-0.0928679705,-0.0156677663,0.3729660809,0.2088799328,-0.0953669474,-0.2414551973,0.0264161546,-0.1072487608,0.025687797,0.448348254,-0.4473807514,0.2485096008,0.2507135272,-0.0865421817,-0.4716270566,-0.228177011,0.0623264946,0.4123413861,-0.2768035233,0.0480621532,-0.3042761385,0.0706739873,-0.2762551606,-0.2442917079,-0.1131715775,-0.187987566,-0.0200080425,0.147875458,0.1640335321,0.1737643778,-0.3774854541,0.0802463517,0.142410472,0.2666999102,0.0755342767,0.355655849,-0.2649778426,0.0805977732,0.2563588917,0.2614737749,0.3012186289,-0.3790412247,-0.1666582227,-0.2933736444,-0.2548195124,0.1949618608,-0.1010442078,0.491796881,0.18626073,0.1617494673,-0.4810594916,-0.0050772685,0.1953458935,-0.2474920005,-0.2180173397,0.1798492819,0.234681949,0.1212505475,0.1843240708,-0.4743080735,-0.2182658613,-0.2020218074,0.1711692363,0.4613437057,0.3109264076,0.4793639779,0.2761179507,0.0255648028,0.3644609749,0.2563191056,-0.3652284741,-0.2790178359,0.0963020623,-0.1500901729,-0.34138906,-0.3217272162,0.0566476174,-0.0939271301,0.0215307306,-0.5035953522,0.0211521871,-0.4214098155,0.314701885,-0.172426343,0.0053863176,0.1490311921,-0.0876984745,-0.1095642447,-0.0973269194,0.1187222153,0.1006601602,-0.2743149996,0.1497716308,0.1701515168,0.1510663033,-0.0164214168,0.4120506346,0.3573698401,-0.3324873149,0.4053789973,-0.4424506128,0.3208769262,-0.2696481347,-0.4025384486,0.0798881203,0.1742584407,0.0781974792,0.0966284797,-0.130990997,0.2963245511,-0.0832739845,0.3459349573,-0.0501949824,-0.1300786287,-0.0904238373,-0.1626685113,-0.0336054601,0.0790110976,0.0641852319,0.0498266183,-0.3970893919,-0.1569623202,0.0445409045,-0.0751565173,-0.1245057806,-0.5061219335,-0.0352432318,-0.4036788046,0.2355426997,0.2805844545,0.507234931,0.0299234279,-0.2966660857,0.0922337398,0.2061196417,0.5514611006,-0.3398970366,-0.097969614,0.0970167965,0.3634591103,-0.3369245827,0.0543134548,-0.3699727654,-0.0982591435,0.0051428722,0.0715076774,0.09467756,-0.0232689921,0.3341372311,0.3008844852,-0.107996963,-0.0471415445,-0.3078944683,-0.2371719927,0.0529015921,-0.2208154351,0.1728818715,0.0012107396,-0.3578479588,-0.031216085,-0.2409801781,-0.1293711066,0.3509055674,-0.0750372261,0.4739066064,0.0410915762,0.2869811058,0.1446430385,-0.027709242,0.2665628791,0.496622026,-0.2788310945,-0.0947812721,0.400090605,-0.2362223864,0.1225083619,0.3385011256,0.0362053663,0.0384677537,-0.2309156209,0.1746630818,-0.1083072051,0.2419946492,0.3126618266,-0.014840031,-0.0183224436,-0.4851876795,0.2522706389,0.0630789548,-0.0964786932,0.501113832,0.013892035,-0.1409137398,0.2544831634,-0.1497292221,0.8568838239,-0.0308649242,0.2993584573,0.0875402763,0.0374528505,0.023583008,0.0480057709,0.1957513839,-0.3019755185,-0.4598467648,-0.2112267762,-0.2421401888,0.0035446584,0.2470919341,-0.1035337821,0.1237002537,-0.0992962271,0.3063516319,0.0404543132,0.2245170325,-0.0328920633,0.1258927137,0.057913851,0.1418959945,-0.1650194526,0.314252913,0.2185323685,0.1562515348,-0.1354680359,-0.1218523309,-0.5502171516,-0.2664139867,0.0397154205,0.1994995922,0.1325662285,-0.6272940636,-0.0505507737,0.0681160986,0.1370395571,0.1105291471,0.0722211152,0.1885420084,0.3306235969,0.1965001374,0.1171642542,0.1006599069,0.2333312333,0.1928985119,-0.2647651136,0.2729426324,0.3091043532,-0.1935537457,-0.4419547319,0.083402358,0.1633358449,-0.5871553421,0.0672345757,-0.2345443964,-0.1870048493,-0.3611566126,0.1178228483,-0.1789943129,-0.0365661643,0.1860949993,0.2183525413,-0.4501624107,-0.1031486019,0.4739904404,0.1552780122,0.2125387192,0.4821252227,-0.0667061806,-0.0557059273,-0.1676465124,0.4262166321,0.0320264585,-0.3387601376,0.1516672522,-0.3282779157,0.5155463219,-0.0208069403,0.2009588331,0.0158308633,0.1482078135,0.0279171709,-0.3668281138,-0.2123217285,-0.0030769738,0.2542794943,0.4290836751,-0.3127452135,0.1724690497,0.0343037061,-0.090548709,-0.2784164548,0.0945302695,0.0782831237,-0.0105048446,-0.1586610526,-0.1473732442,-0.286092788,-0.2582941055,0.1378343999,0.3465049565,0.1183801517,-0.2005618364,-0.2288454771,0.1335097402,0.1539605111,-0.3416458666,-0.2093997896,-0.0374006182,-0.1284738779,0.0344648361,-0.2621636689,0.2835559547,-0.0706352144,0.1456427574,0.1420406103,-0.3068485558,0.2469903231,-0.1395252049,0.0197879616,0.3735809922,0.3695531189,0.1420268267,0.1599660069,0.1781615168,-0.3937753141,0.1774546206,-0.1778348684,-0.1640672535,0.0095910598,-0.560875833,0.1087899804,0.0832521617,0.2583893239,0.583171308,-0.3170389533,-0.1572428793,0.0690837801,0.1981882602,-0.2848282456,0.1177950948,-0.0665956363,0.2775315046,0.1413096488,0.305288285,0.0121634742,-0.3364195228,0.0608418249,-0.030434534,0.149237752,-0.2038494051,0.0064882948,0.4880976379,0.0828870609,-0.146875605,0.2379800826,0.1550234407,0.0340832509,0.5830999017,-0.1760374308,0.0659421161,-0.3406701684,-0.205363214,0.2019726038,-0.2698900402,0.0303765032,0.2715725303,-0.0638381988,0.1975301355,0.2287973464,0.3758349121,-0.1375027299,-0.139683485,-0.1035013199,-0.2434432209,-0.0297406949,-0.0681831762,-0.4048813879,-0.3176363707,-0.2506077588,0.0676791891,-0.064664565,-0.2864044309,0.4035381973,0.2066809535,-0.4247351885,-0.7394703031,-0.0748798698,-0.0934240073,0.1635483503,-0.1644335985,0.2099391371,0.2600238025,0.0766125172,0.2290042043,0.4289530814,0.2771210968,0.0892440826,0.0091933981,-0.0355348252,-0.2207076102,-0.0375216044,-0.0884563625,0.3202272356,0.1627736241,-0.2989524305,0.1397024691,0.1191016883,-0.1236587018,0.0796153992,0.3096981943,0.1495012045,-0.6249405146,0.1978197992,-0.2001100034,-0.0692289248,-0.0897506997,-0.2429320812,-0.3823128939,0.1567353457,0.2767263651,0.0312635899,0.166317448,-0.1340207905,0.0826825276,-0.2544041574,0.1365362257,0.2892346084,0.1720065624,-0.1794471294,-0.1855721176,-0.5673601627,0.1855698824,-0.2054438442,-0.1933714896,-0.1240822822,0.2948230803,-0.0668658018,0.1572427303,0.2515465915,0.2369845062,-0.4340864718,0.6195858717,-0.2376706153,-0.3148562908,-0.1020161957,0.2472931296,-0.1991700977,-0.2393509895,0.2368559986,-0.146631673,0.0331435129,-0.133035019,-0.0718430951,-0.000271604,0.3387230635,-0.0349147804,0.1415819973,0.4902482033,-0.0856481567,-0.0133821927,-0.2097018659,-0.3904011846,-0.2519111931,0.5936504006,0.1548087597,0.2597260475,0.0250413623,-0.2757870555,-0.0186384376,0.36237216,0.1613412052,-0.1446024776,-0.2119924873,-0.2027758807,-0.148818776,0.1001364216,-0.0067016855,0.3987959027,0.0470879041,0.2405490428,-0.3749098182,-0.5723357201,0.2706330717,-0.5221262574,-0.412818253,-0.0644716248,-0.0465459004,0.1833209246,-0.1218518913,-0.3395608962,-0.1269856393,0.5760954618,-0.1305646896,-0.0700631887,0.2427722812,-0.1420822442,0.2959817648,-0.0730573907,0.822298646,0.1487141997,-0.2935891449,0.2348211706,-0.0542103387]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2125","title":"Is dataset timit_asr broken?","comments":"@mariosasko \r\nThank you for your quick response! Following #2052, I've fixed the problem.","body":"Using `timit_asr` dataset, I saw all records are the same.\r\n\r\n``` python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n\r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\n\r\nshow_random_elements(timit['train'].remove_columns([\"file\", \"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \r\n \"sentence_type\", \"speaker_id\"]), num_examples=20)\r\n\r\n```\r\n\r\n`output`\r\n\r\n\"Screen\r\n\r\n\r\nI double-checked it [here](https:\/\/huggingface.co\/datasets\/viewer\/), and met the same problem.\r\n\r\n\"Screen\r\n","comment_length":13,"text":"Is dataset timit_asr broken?\nUsing `timit_asr` dataset, I saw all records are the same.\r\n\r\n``` python\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\n\r\nfrom datasets import ClassLabel\r\nimport random\r\nimport pandas as pd\r\nfrom IPython.display import display, HTML\r\n\r\ndef show_random_elements(dataset, num_examples=10):\r\n assert num_examples <= len(dataset), \"Can't pick more elements than there are in the dataset.\"\r\n picks = []\r\n for _ in range(num_examples):\r\n pick = random.randint(0, len(dataset)-1)\r\n while pick in picks:\r\n pick = random.randint(0, len(dataset)-1)\r\n picks.append(pick)\r\n\r\n df = pd.DataFrame(dataset[picks])\r\n display(HTML(df.to_html()))\r\n\r\n\r\nshow_random_elements(timit['train'].remove_columns([\"file\", \"phonetic_detail\", \"word_detail\", \"dialect_region\", \"id\", \r\n \"sentence_type\", \"speaker_id\"]), num_examples=20)\r\n\r\n```\r\n\r\n`output`\r\n\r\n\"Screen\r\n\r\n\r\nI double-checked it [here](https:\/\/huggingface.co\/datasets\/viewer\/), and met the same problem.\r\n\r\n\"Screen\r\n\n@mariosasko \r\nThank you for your quick response! Following #2052, I've fixed the problem.","embeddings":[-0.0944136307,-0.2299832255,-0.15899086,0.4263156056,0.3007159233,-0.034089271,0.344070822,0.2425390035,-0.1974536031,0.2034934759,-0.1049199179,0.2628142536,-0.2683047056,0.1570613384,0.0637481287,-0.1874102056,-0.1392097622,0.0583299734,-0.2021197975,-0.3332761228,0.0206732545,0.1031607836,-0.3101027012,0.0231755022,0.0168496594,0.2146271914,0.1049122587,-0.3353047371,-0.1445068866,-0.5664838552,-0.066272527,0.0032359811,-0.0603339933,0.5745943189,-0.0001122964,-0.0225118659,0.1222439781,0.0365277156,-0.184933126,-0.226524964,-0.2687570453,-0.0250014029,0.0155640785,-0.0129869552,-0.0356445424,-0.1710772067,0.1398568898,-0.1769040525,0.0261944849,0.4765830636,0.243869707,0.3525103033,-0.0536975265,-0.0287914593,0.2567532659,0.1916886717,-0.0373355485,0.113391906,-0.0269826949,0.0845064372,0.0569577962,-0.0610196367,0.0639974326,0.4345183671,0.1751514673,0.0408190824,-0.2794828415,-0.3360896409,0.1150866076,0.4349095225,0.5536218882,-0.1592944264,-0.0299365018,-0.0579489395,-0.1335771829,-0.2379938066,0.0719560981,0.4189505577,-0.3515238166,0.1977128983,-0.3249020278,0.4548994005,-0.0189095009,0.4610393345,-0.114049457,-0.236891374,-0.1160336062,0.1303205192,-0.1702312231,0.0221773759,0.1701573879,0.0155233219,-0.0297290068,-0.1539023221,-0.0998006761,-0.111579597,0.0230733212,-0.0501786061,0.3074500561,0.284286797,0.5185361505,0.048398409,-0.2063955218,-0.0812925175,0.1807557046,-0.1003738493,-0.0331978239,0.1284053475,0.0767413527,0.0402606688,-0.2282278836,0.0481381156,0.2214837223,-0.14091295,0.2436320037,-0.3832003474,0.2666969597,-0.1033535749,-0.5151004195,0.163535744,-0.2139723599,-0.1426067799,-0.1729005277,-0.0103709269,0.0511047244,0.1875349879,-0.0298455004,0.1230809912,-0.1091653928,-0.2158935964,-0.2111433297,-0.0542004146,0.0981449261,-0.0136457728,0.0386305563,-0.1145481616,0.1666430831,0.2836228907,-0.0602694787,-0.1283297092,0.1149625629,-0.2057009041,0.3776217401,0.1563329995,-0.0176783279,0.3158635795,0.1335187405,-0.1903614253,-0.064777188,0.1713825613,0.1025048941,0.1309217364,0.0487720966,0.1945507973,-0.0564401112,-0.2527545094,-0.296760112,0.251824975,0.091397658,-0.0583350547,0.2846869826,-0.2022020221,-0.4809475839,-0.1121225953,-0.0023110562,0.3185498714,-0.5554748774,-0.0544206686,0.197700426,0.2609818578,0.1589964181,-0.0083660157,-0.1049015895,0.2564176917,-0.3818316758,-0.1732402593,-0.0477648042,0.0161769185,-0.5107367635,-0.0563076548,0.1109882668,0.0543328486,0.0458500795,0.0018961322,0.2393533736,0.3224302232,0.1289907098,0.0326229036,0.0922353417,-0.1112632826,-0.2986896932,-0.0313351899,0.2737309933,-0.0285530221,0.1150397658,0.2086395323,0.1651357114,-0.024845155,0.2878836691,0.2145033628,-0.0914638862,0.0987547189,0.0550125241,0.0845820159,-0.0848571956,-0.2950709462,0.1494978815,0.0599041879,0.3097231984,0.3030186594,0.0910419524,0.0394325666,-0.3109816313,-0.2567048073,-0.2525522113,-0.1545086056,0.1543771923,0.4801377654,-0.0219440944,0.0739435777,-0.1331045032,0.3650311232,-0.2888017893,-0.0323501714,-0.4340262711,0.4149918556,-0.0574908182,-0.0560397841,0.0012543278,-0.2250435501,-0.0683882609,-0.0880253538,-0.1584329605,0.3177549541,0.2989354432,-0.2319931984,-0.1379734278,-0.2581929564,0.2997581065,-0.6153901219,-0.1373835504,0.3865926266,0.0841274112,0.0296019614,-0.1281595975,0.177975893,-0.267467469,0.1416684389,0.224332884,0.2857886553,0.0377951562,0.176301077,-0.1604193896,-0.1877563298,0.1830431074,-0.0583489239,0.061124634,0.1120545939,-0.2140530199,0.2224882692,0.4159753025,0.2036007792,-0.0180017687,0.2874861062,-0.1126139611,0.2327652872,-0.0957601368,-0.0144061493,0.3610875309,0.2147130817,-0.0985526219,-0.2153686434,0.0179662872,-0.1081276387,0.0277002454,0.4359531105,-0.4291336238,0.2373061031,0.2552713752,-0.084281452,-0.4789946973,-0.2357929051,0.0521990471,0.41699633,-0.2735524774,0.0509257689,-0.3167632818,0.0732414424,-0.296672374,-0.2451492101,-0.1077257395,-0.1767433286,0.0053803041,0.1615447104,0.157437399,0.1844130307,-0.3741720021,0.0866069198,0.1346716881,0.2600464821,0.069708094,0.3421808779,-0.2661169171,0.084178023,0.2482769936,0.2644500434,0.3041813374,-0.3884472549,-0.1608851254,-0.2859036326,-0.2490404695,0.1910018623,-0.1057764739,0.488663733,0.169478938,0.1752690524,-0.4851479828,-0.0132378796,0.2096191049,-0.2393307835,-0.2105317861,0.1869343221,0.2258640975,0.1076973751,0.1983316094,-0.4707356691,-0.2260404676,-0.2080564499,0.1557306498,0.4561809003,0.3075820506,0.4707158804,0.2925748527,0.0238526706,0.3549118936,0.2623209655,-0.3535295427,-0.2862578928,0.0978980586,-0.1545945108,-0.318762511,-0.3196055889,0.0570274219,-0.1052343696,0.0083997771,-0.5147380829,0.0292515848,-0.4167885482,0.3288158476,-0.16373384,-0.0086587332,0.1560747772,-0.0687764734,-0.1160510108,-0.0976256281,0.1164996922,0.0959962755,-0.287642926,0.1441399306,0.1631861627,0.1440094411,-0.0120102922,0.4265808761,0.363316834,-0.3365593553,0.4059383571,-0.4415449798,0.3211946785,-0.279903084,-0.3907304704,0.0814622492,0.1708474755,0.0958441272,0.0976480767,-0.1336951852,0.2975595593,-0.0881685466,0.3679832518,-0.0700451136,-0.1262581646,-0.0909311026,-0.1559237987,-0.0418329239,0.0852145702,0.0708201155,0.0422700457,-0.3910236359,-0.1579559445,0.0451142564,-0.0965711102,-0.1080797091,-0.5057348013,-0.0241824761,-0.3895759284,0.2380630076,0.2875640094,0.4916777313,0.0211738665,-0.3042039573,0.0874171481,0.2110399306,0.548463881,-0.3414883614,-0.105823122,0.1027663723,0.3754736781,-0.3335992992,0.0391473621,-0.3677536249,-0.1123004109,0.012857561,0.0793340728,0.0907600746,-0.0272111371,0.3316936195,0.2905203998,-0.1069219559,-0.0424406528,-0.2977701128,-0.2381023467,0.032991223,-0.2289684117,0.1603752673,0.0103717791,-0.3745903969,-0.0359542556,-0.2429721951,-0.1117576659,0.3477697968,-0.0684132129,0.4770265222,0.0493980721,0.2673359811,0.1537425369,0.0040454012,0.2676102221,0.4922198951,-0.2969177067,-0.1168243736,0.3978705406,-0.2442583591,0.110498704,0.3389165401,0.0431903377,0.0313784555,-0.2195609212,0.1784447879,-0.095107086,0.2496219873,0.3162172139,-0.025283657,-0.0231621899,-0.4884686768,0.2487924248,0.0560952127,-0.08912839,0.4972658753,0.0020714272,-0.142585814,0.2568328679,-0.1635844111,0.8594865799,-0.0307138208,0.2935119867,0.0960831121,0.0333243236,0.016875308,0.0559487045,0.2002278417,-0.3100888133,-0.4505859017,-0.2109333724,-0.223021239,-0.0085130678,0.2497554123,-0.1237293556,0.118559666,-0.0994169191,0.3084924519,0.0471444577,0.2165275216,-0.0265042912,0.1237470359,0.0573642477,0.1458674818,-0.1752040237,0.3027879298,0.2199776918,0.1572905481,-0.1419967562,-0.1131931543,-0.5301106572,-0.2589643002,0.0222144369,0.2207390964,0.1105211005,-0.6277399063,-0.0739607811,0.0796890259,0.1387319863,0.107005626,0.0697783902,0.1947240531,0.3500928283,0.2039142549,0.1056173295,0.1020974517,0.2333535552,0.1859398931,-0.269821763,0.2788249552,0.3068408966,-0.1936115324,-0.4506179988,0.0654368624,0.162845403,-0.5830277205,0.0653802082,-0.2426423132,-0.1819088906,-0.3711572587,0.1221481785,-0.1919399649,-0.0287858117,0.1772329807,0.2248618901,-0.4648852944,-0.1009581089,0.4693321586,0.1480303705,0.2039613724,0.4891717434,-0.0543296747,-0.0649099499,-0.1682599634,0.426802814,0.0319434144,-0.3455279768,0.1573603451,-0.3312208056,0.5081256032,-0.0121320002,0.2068413645,0.0130855795,0.1367182136,0.0353816338,-0.3639126718,-0.2206833661,-0.0113547826,0.2586455643,0.4356366098,-0.2898158729,0.180314675,0.022594545,-0.0894709229,-0.2851370871,0.1051382497,0.0718720481,-0.0024523074,-0.1655049026,-0.1436953545,-0.2792791128,-0.2454968393,0.1449499279,0.3362704515,0.1100570485,-0.205648303,-0.2384363562,0.1296857744,0.1451862752,-0.3378193378,-0.2249560356,-0.0429136641,-0.1171729565,0.0215402395,-0.2548116744,0.2861627936,-0.0749738365,0.1503391266,0.1408236176,-0.2975181341,0.269069463,-0.1583424062,0.0409280024,0.3635620475,0.3758502007,0.1333250403,0.1450488567,0.173467502,-0.3838422298,0.1830306053,-0.1668029726,-0.1487361938,0.0030788637,-0.5740554929,0.1108029187,0.0812298581,0.2401498109,0.5738211274,-0.3300331533,-0.1632021368,0.0584445186,0.2060611397,-0.280716449,0.1205482706,-0.0547601953,0.2846924067,0.1474231184,0.2898190022,-0.0129493261,-0.3252896965,0.0576986745,-0.0283226185,0.1399359554,-0.2223534137,0.0070726196,0.492626071,0.0913106725,-0.1450946331,0.2201403826,0.159731552,0.0197491348,0.5896523595,-0.1648549587,0.0698649958,-0.3502913415,-0.2143065631,0.1930432022,-0.2712002695,0.030830428,0.2703626752,-0.080063343,0.1966592073,0.2391740233,0.3799553812,-0.1297030449,-0.1398824453,-0.1042475775,-0.2289241254,-0.0364072621,-0.0727746487,-0.4077167511,-0.3233157694,-0.2500276566,0.0669533238,-0.067364648,-0.2733787894,0.3946804404,0.2033441216,-0.4306726754,-0.7658959627,-0.0791521072,-0.0878009424,0.1573748738,-0.1591686904,0.2115531266,0.2831963599,0.077532582,0.2259530723,0.4115032554,0.2787351012,0.0918692201,0.0142962905,-0.0312947407,-0.2229076624,-0.055536285,-0.0954719186,0.3027201891,0.1857458055,-0.2893978059,0.147346139,0.1233602092,-0.1304837763,0.0833740309,0.3048797846,0.1632215083,-0.6060594916,0.2010073066,-0.1922048926,-0.0632651076,-0.096016638,-0.232004568,-0.3824713528,0.170378983,0.2728130519,0.0432866476,0.1683956683,-0.1360924691,0.0847530738,-0.2389442325,0.1203228086,0.3004477024,0.1786033362,-0.1906438768,-0.1872394383,-0.549472928,0.1819881201,-0.2047065943,-0.1838861704,-0.1320886761,0.2907494605,-0.0620051138,0.1483184546,0.2524802983,0.2384580076,-0.4138381481,0.6303966641,-0.2375379354,-0.3110553026,-0.0882276595,0.2481026351,-0.187876597,-0.2301415205,0.2155867666,-0.1486599594,0.0448096655,-0.1434415877,-0.077588059,-0.0065604295,0.3242446184,-0.0337875597,0.1323007196,0.489463985,-0.1076235026,-0.0122366473,-0.211024195,-0.3857027888,-0.2540199459,0.5880765915,0.1665131599,0.2781363428,0.0249670483,-0.2734915316,-0.0213631261,0.3611066639,0.1587791741,-0.1373381168,-0.2133617699,-0.2024482936,-0.1398268938,0.1106565222,-0.010394644,0.4003752172,0.058259774,0.2474054992,-0.366381228,-0.5741569996,0.2795921266,-0.5135557055,-0.4329737723,-0.0715691075,-0.0603781231,0.1696548611,-0.1073379368,-0.3473367393,-0.1222760007,0.5632582307,-0.1397340298,-0.0757972151,0.2474447042,-0.1439987421,0.3169555068,-0.0632695779,0.8191850185,0.1606454402,-0.2852630913,0.2215066701,-0.0743670017]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2124","title":"Adding ScaNN library to do MIPS?","comments":"I haven't played with it (yet) but it sounds really cool !\r\n","body":"@lhoestq Hi I am thinking of adding this new google library to do the MIPS similar to **add_faiss_idex**. As the paper suggests, it is really fast when it comes to retrieving the nearest neighbors. \r\n\r\nhttps:\/\/github.com\/google-research\/google-research\/tree\/master\/scann\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/112738294-78ec9800-8fc6-11eb-9a5f-3d7ee5818e76.png)\r\n","comment_length":12,"text":"Adding ScaNN library to do MIPS?\n@lhoestq Hi I am thinking of adding this new google library to do the MIPS similar to **add_faiss_idex**. As the paper suggests, it is really fast when it comes to retrieving the nearest neighbors. \r\n\r\nhttps:\/\/github.com\/google-research\/google-research\/tree\/master\/scann\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/112738294-78ec9800-8fc6-11eb-9a5f-3d7ee5818e76.png)\r\n\nI haven't played with it (yet) but it sounds really cool !\r\n","embeddings":[-0.2439276874,0.0034289842,-0.1748069972,-0.1881957799,-0.2444058359,-0.2017466426,0.4440642595,0.1649580598,0.2253530324,0.2816860676,-0.3657340705,0.0642065555,-0.0174992178,0.0287976768,0.1417429,-0.1609864682,0.290825963,0.3200619221,0.0523744486,-0.1016989574,-0.4728910327,0.1844010055,0.0737305433,-0.364643842,0.0020868806,0.0979250297,-0.388037473,0.105969578,-0.051583834,-0.3903476894,-0.1558544934,-0.0468135737,0.1032796949,0.2193197906,-0.000106481,-0.0674293041,0.2991676927,0.3646429479,-0.0447616093,0.3702233434,0.5776326656,0.1974809468,-0.2921523452,-0.2226367742,-0.1187293231,-0.230536893,0.2675301731,-0.3972280622,-0.2300039828,0.0719330981,0.2284923345,-0.1312045455,-0.2078467309,-0.0318896174,0.5179476738,-0.1198453009,-0.0284641907,0.0396961048,0.695596993,0.1496863216,0.1771717072,0.1252788603,-0.0309136864,-0.2636859715,0.0078621386,0.2986260653,0.4804643989,0.1140592843,-0.2203001529,0.0238395389,0.1815486103,0.2251856327,-0.0257380083,-0.0747278705,-0.3316415846,0.0016215093,-0.0350675844,-0.2963382602,-0.1345903873,0.0265712999,-0.1685235053,-0.2269645482,-0.111447677,0.2361243218,-0.0228041131,0.6003221869,0.1670107692,-0.1093627512,0.310633868,-0.1485329717,0.3306520581,0.0503126122,-0.007025253,0.1065405458,-0.0516949333,-0.1211316437,0.3935858607,0.0315492637,0.085860759,0.3176424205,-0.2235083729,-0.0703808665,0.1479820013,0.0444454849,0.1247496977,-0.1915000528,-0.1532999128,0.2061452419,-0.0181469917,-0.0808439478,0.1935846806,0.0671992525,0.0162718613,-0.0014427894,-0.0765833259,-0.058442682,-0.1247452646,-0.2123304904,0.0182131566,0.000630783,-0.2326547354,-0.0856645703,-0.0019698644,0.1413330436,0.1057066992,-0.0438007228,0.172886461,-0.2304132879,0.0434571691,0.2953921258,-0.176463455,0.3611565828,-0.1907304376,-0.0753907636,0.1362164468,-0.2160347104,-0.0516859293,-0.3701979518,0.4505628645,0.3460554481,-0.0633237138,0.0135687198,0.2901624143,-0.0864409879,-0.3381961584,-0.4754308462,-0.3640990853,-0.0120292259,-0.454549402,0.0171589777,0.094444491,0.1259126663,-0.1041667536,0.232661888,0.048034627,-0.2685878873,0.2923268378,0.3267959952,-0.4835846722,-0.1569115222,0.1576842964,0.2503290772,-0.446678102,-0.1068820506,0.0614431985,0.1720516235,-0.0163181722,0.0182968006,-0.215343222,-0.0476334281,0.3054264188,0.2992741764,0.2272979617,-0.2108685374,-0.1943710297,0.3263393342,0.0957269445,-0.0909270421,-0.1099829152,0.4140731394,-0.3914115429,-0.2408130914,0.3713569045,0.660115242,-0.0675149858,0.0135660246,0.232835114,0.4148376584,-0.1766024232,-0.0111276954,-0.3273112774,-0.2288618088,0.175212279,0.7545500994,0.0234158207,0.0653035492,0.4215060472,-0.394087106,0.100602895,-0.1570695341,-0.1931440681,0.0094219605,0.4853834808,0.0043811793,0.2561779022,-0.264193058,0.165704608,0.0374602154,-0.448902756,0.0731068254,0.2382297516,-0.2553140819,0.2837933004,-0.0311403554,0.2175412923,0.0618828423,0.1398262978,-0.0373215936,0.1674234122,-0.2987346649,0.1852008849,0.0877567977,0.0501877628,-0.1837649494,-0.1171224564,0.5151113868,-0.1383142322,-0.0123645989,-0.0952299908,0.1532347798,-0.1308612823,-0.1894020289,0.3319476247,0.2608869374,0.0747468919,-0.110782668,0.1832978725,0.1126738861,0.0618322566,-0.1720483303,0.2330371588,0.2173331231,-0.1156510934,0.0232397914,-0.2236055136,-0.0421536081,-0.07882642,0.1326366067,0.0184345637,-0.2012383789,-0.3159188032,-0.1284807622,0.2013850212,-0.0969331115,0.5249751806,0.4378328919,-0.2195589095,-0.0168844108,-0.2969761789,0.216933161,-0.037681222,-0.3288104236,-0.0941482782,0.0848254785,0.2931946218,0.0334172063,-0.0281751361,-0.0239579286,0.2890660465,0.2036236376,0.303290695,0.2123272568,0.1261684299,-0.1490886509,0.1723284572,0.1045479551,-0.0584341548,-0.2907503843,0.0801903531,0.1126976758,-0.2351157069,-0.3301084638,-0.1249742806,-0.0837864205,0.1088310182,0.1020435765,0.1325643808,-0.214787364,0.0221008863,0.1201526895,0.0588300712,-0.0443548895,0.2410334349,-0.1352503598,-0.1010427326,-0.3225780725,-0.3271054626,0.5430046916,-0.2122444361,-0.5714127421,-0.1312825978,-0.104858838,-0.0185714699,0.299475193,0.0029169451,0.2248697728,0.4322256744,-0.005308257,0.2047624737,-0.4008827209,-0.4326358736,-0.053003557,-0.1019085646,0.0595919937,0.4651590288,-0.1486257464,-0.2538245618,-0.1922210455,-0.1752843708,-0.4530964196,-0.2411993295,-0.3399581313,-0.1794105768,-0.2952166498,-0.3307221234,-0.3542324007,-0.0450927019,-0.2859196067,0.5962640047,0.3283173442,0.2007227093,-0.2365074456,-0.26006338,0.1647622883,0.0599608198,-0.045453649,-0.1864426285,-0.1829024702,0.3587506413,-0.0787776932,-0.0430119783,-0.2540545166,-0.0144752981,0.127087459,0.1777641624,-0.2988947332,-0.1281728297,-0.2070728689,0.4978980124,0.0278636087,0.1416925639,0.209445551,0.0281614419,-0.2733469307,0.0803082287,-0.0237224586,-0.0069016493,0.2342169583,0.0340472721,0.2494204193,-0.1367782652,-0.2551766336,0.6270580888,-0.1674866527,-0.3651384115,0.1304645836,0.2270871401,0.1637421548,0.2353668213,0.0763131902,0.7089679241,0.2634444237,-0.226732105,0.1833946258,-0.0173373409,-0.6133036017,-0.4285646677,-0.1082388386,-0.0302212089,0.1491060108,-0.1336829066,0.5029648542,0.1386191398,-0.3031700552,0.0883855373,-0.120895721,-0.124483563,0.214840889,-0.043975126,0.1240291074,-0.145969525,-0.2674268186,-0.2757005394,-0.3354333043,0.3666348159,0.4916953444,0.0079957768,-0.1112744585,-0.1248126775,0.0095164115,-0.1879950464,0.2366002202,0.0981640592,-0.3039349914,0.1938486546,0.1228739247,0.2310431898,0.1336094588,-0.0307846367,-0.4746902585,0.2855166197,-0.0098590581,0.02789695,0.0026937663,0.4456697702,0.006392506,0.1031831503,-0.3089490235,-0.1596838981,-0.2732820511,0.1156228781,0.1495825946,-0.0283898059,0.0336155146,0.1130435616,-0.1121058986,0.2172135413,0.0421584919,-0.0188347101,-0.1700963378,-0.0508027375,0.0235258378,0.1964353621,0.2750257254,0.0580168404,0.2852747142,0.4109987319,-0.4056477845,0.3159336448,0.0938972011,-0.0265045725,0.4097904861,0.2522833049,0.0153208803,0.0409232825,0.0207608938,0.1146466285,-0.039963305,-0.1535980254,-0.0931218863,-0.0244413074,0.1771236956,0.0480435304,0.7106666565,0.4801022708,-0.0928782746,0.3190356791,-0.1976749152,-0.1145162731,0.1005823836,0.4421913922,0.8556509018,-0.1760201007,0.0513558611,-0.0394106247,-0.1808353215,0.1944292635,-0.3663876653,0.184133485,-0.149486661,-0.4856396914,-0.0709300414,-0.1102544144,-0.0231913067,-0.1520302296,-0.1599058956,0.0418779366,0.014737851,-0.2533441186,0.0114364326,0.3851204216,0.1672091633,-0.1421851069,0.427639097,0.176246047,0.0072040437,-0.0106835449,-0.0714061931,0.0629348978,-0.2269870788,-0.0714777857,-0.0943691656,0.2490312606,-0.0239142198,0.0587469414,0.2304664105,0.1235417575,0.1213321984,-0.4415458441,0.3983541429,0.0063112285,-0.1647005081,0.1885759979,0.3487853408,0.0643974692,0.1721750349,-0.2203038782,0.162700519,-0.135785073,-0.0518147163,0.3476836681,-0.1064652652,-0.4565868378,-0.8413486481,-0.0335487314,0.0761672929,0.2213761806,0.5182541013,0.1134030819,-0.3143363893,0.0683096275,0.1505320668,0.315445751,0.0137866708,0.1275366247,-0.0560378619,-0.1336726695,-0.2073729485,0.4529682696,0.2898274958,-0.4834640622,-0.0407488793,0.1940705329,-0.2930107713,-0.2851007879,-0.3741272986,-0.0914751664,0.4399764538,-0.2042037994,-0.1941661239,-0.1098435149,0.3261407912,0.102483511,0.0318717398,0.0210315939,-0.1052048579,0.2126052082,-0.3911857605,-0.040703591,-0.464078337,0.4494143128,0.0207582004,-0.076572448,-0.1850759834,-0.3093580604,-0.4104768932,-0.0578132272,-0.1113997325,0.0164884217,0.3149392307,-0.1182718053,-0.0382991806,0.0046192631,0.2115458101,0.1103821397,-0.1684899926,-0.1700077206,0.0398654975,0.0464133732,0.0427779742,-0.1468289942,-0.0405079052,0.0601342134,-0.2841018736,0.0717561394,-0.0896284506,-0.1071665883,-0.1838658005,-0.1839577407,-0.1124632582,-0.5035741329,0.2538336515,0.2823440433,-0.0094011771,0.1491546631,-0.3326444626,0.3584728539,-0.1187570393,-0.285441488,0.3170676231,0.3685527742,0.1224017367,-0.0003441816,0.1165403798,0.040222045,-0.3725841939,0.0995969549,-0.1430895627,0.2047327608,-0.0654691905,-0.1820137799,0.2268034667,0.2684770226,-0.0854113251,-0.3562788069,-0.0327468328,0.0377577357,0.1165310591,0.3444569111,0.3377906382,-0.0469856672,0.0804347172,0.0415378772,-0.1408139467,0.0331154726,-0.0295480713,0.2654822767,0.2989991307,-0.2791555226,0.1002762541,0.1401390582,0.0695382878,0.3646760285,-0.2239983678,0.6093226075,-0.1576055288,0.3123035431,0.0322957374,0.4045744538,0.1674047559,0.1657651812,-0.2857898772,0.3381987214,0.3550392091,0.4322375953,0.0102194473,0.1311620027,0.0006707166,0.2492920309,0.0500238501,0.1445665956,-0.1393262893,-0.1216372401,-0.5311270356,-0.1025549173,0.1609526724,-0.1085468903,0.3158116937,0.0112695405,0.1192410588,0.2756481171,-0.2617453337,-0.0822085068,0.3062072992,-0.3480428457,-0.2753644586,0.0605836771,-0.2288981676,0.2881251872,0.0814090893,0.0127373682,0.1090810448,-0.2987891436,-0.052355174,-0.1934139729,-0.0770231187,0.0345387533,0.2351804823,-0.0527364127,-0.3772054017,0.2721776664,0.0569708794,-0.2131283283,-0.0216332413,-0.0535409227,-0.1027824357,-0.0039983587,0.0918627977,0.059254989,-0.1689216793,-0.4211100042,0.0451244377,-0.2733216286,-0.3612781465,-0.0334707424,-0.1679474413,-0.0130787296,0.0423974022,0.1182624772,0.1124394238,0.4594082832,0.4565919936,0.539488256,-0.2235670984,-0.2910104096,-0.0299587995,0.0066865049,0.0168611277,-0.0419571213,0.0183006227,-0.0557334423,0.1117935553,0.1021086872,0.0535327941,0.499307394,0.0180981029,0.1797964573,-0.1773415357,0.3049495518,0.0948608443,-0.292347908,-0.0154018588,-0.1796404868,0.0327187441,-0.4936437905,0.1500300318,-0.3161983192,-0.033388041,-0.1832899004,0.2557410002,0.3299242556,-0.1354766488,0.1537890136,-0.1375357658,-0.3644497395,0.1426319778,-0.0241711307,-0.1534529626,-0.2779179811,0.1548444182,-0.1991669536,0.037138965,-0.1942897588,-0.034010075,0.2788951099,0.0594650768,0.1085468009,-0.0447739065,0.087082848,-0.4145021439,0.2807995975,0.2728276253,0.1489718556,0.1079054549,-0.0978990793,-0.2785604894,-0.2896271944,0.4982819557,-0.0157029293,0.0025245345,-0.3799721599,-0.0458922274,0.1704336107,0.3218684793,-0.0778674558,-0.1836545765,0.134330824,0.0114131197,-0.0216514021,-0.077390492,0.1828124672,-0.045565635,-0.2180667371,-0.5602424145,0.1226546094,0.1970307827,-0.1152981967,-0.2530136108]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2123","title":"Problem downloading GEM wiki_auto_asset_turk dataset","comments":"Hi,\r\n\r\nsadly I can't replicate the problem on my Windows machine. Try to update the library to the newest version with:\r\n```bash\r\npip install git+https:\/\/github.com\/huggingface\/datasets\r\n``` ","body":"@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family","comment_length":26,"text":"Problem downloading GEM wiki_auto_asset_turk dataset\n@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family\nHi,\r\n\r\nsadly I can't replicate the problem on my Windows machine. Try to update the library to the newest version with:\r\n```bash\r\npip install git+https:\/\/github.com\/huggingface\/datasets\r\n``` ","embeddings":[-0.0816882253,-0.0280797184,-0.0464758277,0.4495636523,0.2806958258,0.1159588769,0.3161947727,0.1778609008,0.2470398098,-0.1171152517,0.1766637862,0.1722975075,0.1142347306,-0.0033531154,-0.168628484,-0.1824515611,-0.0154063329,-0.0730278492,0.05976725,-0.1803478003,-0.0259094331,0.2145039737,-0.1021221057,-0.1298859864,-0.2996245921,0.1392064393,0.0705756918,-0.0579681844,-0.2835352421,-0.0846156478,0.5160742402,0.0152688697,0.0541247912,0.3606870472,-0.0001191775,-0.0883911103,0.4347236454,-0.075555034,-0.0962895304,-0.4143653214,-0.1749691665,-0.4833079576,0.1066898927,-0.2166323215,0.1262152642,0.385717541,-0.0858343914,0.1857326776,0.0551996343,0.0067077554,0.1897132695,0.3324497044,0.1076219603,-0.0123927351,0.1707160026,-0.4481359422,-0.2214535177,0.0033134529,-0.1796579957,0.1315105408,0.0395613648,0.194448173,-0.0939045176,0.3865469992,0.1432819068,0.1656978279,-0.112732321,-0.5313156247,0.3458316624,0.3503461778,0.7223212719,-0.136218518,-0.4294935167,-0.1829381138,0.0109780161,0.0725168213,0.1456314921,0.4450258911,-0.4367175996,0.0626264587,-0.4804292917,-0.0108169718,-0.0504395589,0.1969978064,-0.2779852152,0.1538722664,-0.1802170426,0.0275478885,0.0224016737,-0.019343324,-0.5223580003,-0.2752556801,-0.1900455356,0.3314402103,-0.3189226687,0.1227754056,0.0755734369,0.2214292139,0.3313711882,0.3213436902,0.0606174991,0.2102866024,-0.2032123506,0.0687996373,0.1237717196,0.3406439722,-0.0189021304,-0.389757216,0.2435046881,0.1455231458,-0.0094314869,0.0612481236,-0.1362194121,-0.0531070419,0.2008862942,-0.3221499026,0.1496372074,-0.107272625,-0.1480018049,0.3196874261,0.1803124547,-0.0844288692,-0.1269769818,0.2217449099,-0.3037922382,-0.1206944138,-0.1747677922,0.3599667549,-0.1778768748,0.1502551883,-0.0383374691,-0.1955487579,-0.0761461928,-0.0892637148,0.293659538,-0.3465345502,0.4333966672,0.1367743015,-0.1526529342,-0.1169113517,-0.1287565827,-0.0390414819,0.098197788,0.3059737682,0.1316255033,0.8464612365,-0.0953701511,-0.1950018257,-0.0023449142,-0.0219803285,-0.2224007547,-0.1721918285,-0.1615702212,0.1161727086,-0.5253882408,-0.244270578,-0.0805345327,0.2623373866,0.1820070148,-0.133912161,-0.0498556085,0.054721836,0.0323239528,-0.2065601498,0.5081894994,0.6407996416,-0.1834061593,-0.0922173485,-0.0692972839,0.0496931374,0.4750284553,0.0278591197,0.0567070842,-0.2645201087,-0.0762461945,0.3398239017,0.1568062305,0.0418995321,-0.6642845273,0.2593211532,-0.1147141457,0.3474900126,0.045729097,0.144865647,0.079168573,0.2138340175,-0.2915182412,0.2178662419,0.0779673234,-0.1681090444,-0.3459100425,-0.5063447952,0.0326996967,0.0652405843,0.105083473,-0.035225939,0.0922684968,-0.0297134239,0.341363132,0.2768877149,0.2760130763,0.1004940793,0.169433713,-0.1927230656,-0.0391564667,-0.1571122408,-0.0986688882,0.4336703122,0.568173945,0.0755832195,-0.0403056592,0.0529332459,-0.2321750224,-0.0986358076,-0.2844692171,-0.0613790974,0.0320412479,0.1366157383,0.2758432031,0.3003488481,0.0641758516,0.0588752404,-0.0273834262,0.0769524053,-0.0364973657,0.1805667579,0.115697585,-0.1536079049,-0.1736176312,-0.0380753577,-0.0246206131,0.12010324,-0.174036175,-0.0016022959,-0.2524647415,0.1376881599,-0.0011998893,-0.0421971753,0.1814187169,0.0924868956,-0.0820007175,0.0342547521,0.1587932706,0.0240616649,-0.0482477546,0.1165841222,0.196415171,0.3760880828,-0.0156626496,0.1976144761,0.3015960753,-0.023061689,-0.0101588164,-0.0483589545,0.3486323059,-0.2917578816,-0.0928582922,-0.0529292896,-0.1222143397,0.2972127199,0.5440853834,0.144067958,0.1042793989,0.2706980109,-0.3404791057,0.0580154024,0.0074611236,0.5212213993,0.2717370391,0.1254092157,0.2412409186,-0.3357158899,0.0227841381,-0.2253905535,0.0504496656,0.3411954641,0.0310134683,0.410066098,0.1152249724,0.1589040607,-0.1292016506,-0.0007829446,0.0924939141,0.1012289748,-0.0613567792,-0.0827763826,-0.2790410221,0.2550799251,0.0191332549,-0.4520641863,-0.1873599738,-0.1690167338,-0.1112914458,0.0601358116,0.0149312541,-0.0829623044,-0.1273140758,-0.1843731552,-0.0779941753,-0.4661277533,-0.3761636615,-0.0393191241,-0.3338101506,-0.0543680377,0.1349695474,-0.1055985838,0.3403262794,-0.1259236634,0.0895538852,-0.1539132893,-0.2383004278,0.0364209078,-0.0391555652,0.4648873508,-0.2274115384,0.4167215228,-0.2183637619,-0.0946405977,0.2081886977,-0.0117890704,0.0117485458,0.1601534933,0.1084003747,-0.1841401309,-0.1214252338,-0.2491450012,-0.3370248973,-0.3670068681,0.0554951951,0.175066784,-0.0603150278,0.1239496246,0.2380945832,0.0504883081,0.0789691582,0.0597550087,-0.0546672493,-0.3152782619,0.7877443433,-0.0981072634,-0.4563619196,0.2528219521,0.0194493122,0.3236148357,0.1190218553,-0.5418384075,0.1938307583,-0.2075989395,0.2528559864,0.3119232953,0.1334060133,0.0659780949,-0.3006307483,0.0850581005,-0.1448848099,-0.0648224205,-0.0518276505,0.0345305614,0.2522470057,0.0448781885,0.2808907032,0.0730527714,0.8497200608,0.3471905887,-0.1278250664,0.4330374897,0.0222697835,0.1222641319,0.0275241397,-0.2162035108,-0.1881283075,-0.2801279724,0.1928050369,0.1971000731,0.2568853498,-0.1927207857,-0.07057935,-0.0354842581,-0.4562742412,-0.3566124439,-0.1564673483,0.1749461293,0.0198501144,0.2528767288,0.3252738118,0.0419687256,-0.3157899082,0.0709474906,0.6042260528,-0.0180898067,0.1327620149,0.0960056856,0.0963849351,-0.6682867408,0.0801902264,-0.0731346011,0.0524440371,-0.0586328506,0.0750326365,0.1558501273,-0.1677349061,0.2104662061,-0.1685733348,0.1192496568,-0.1813427806,-0.1501006633,-0.252238512,-0.1885755807,0.0301537421,-0.0170996077,0.4971936643,0.067660667,-0.3547149897,-0.0469296128,0.0319274664,0.1990066618,-0.1385647953,-0.2439801842,-0.2731687427,-0.4530652463,-0.3138135672,-0.1619619727,-0.1599067748,0.3852410913,-0.1853048652,-0.2951054573,-0.0774151981,0.2495990098,0.173164323,-0.0336133018,0.3335492313,0.1254874766,0.4604846835,0.0982635692,-0.0491276905,-0.2059216946,0.5613113046,0.2876953483,0.004021497,0.0403383188,-0.3094997406,0.0059945341,0.2227870077,-0.1943593025,0.0088876653,-0.0427962989,0.2556528747,-0.2304717749,0.0044154455,0.2676949203,-0.0439065322,-0.1657656878,-0.3439022601,0.7110812664,-0.1870897263,0.0073710917,0.1814473718,0.3193113208,0.110846214,0.1971052289,-0.1053668931,0.7823933363,0.0310283229,0.0068507111,0.0907239765,0.0066090282,0.4280931056,-0.1839735657,0.1017169952,-0.2968285978,0.1675540954,-0.1127451509,0.0654988214,-0.1194293648,0.0317773819,-0.1100663766,0.1077058017,-0.1043975651,0.2221236229,0.0014604434,0.0588449351,-0.4258778989,-0.1570958793,-0.0271796174,0.1052967533,0.0811596364,0.6855864525,0.0226783603,-0.2424750626,-0.2606494427,-0.1097197458,-0.6135619879,0.1031906605,0.050034035,-0.1013253853,-0.143542707,-0.4398378432,0.01041984,0.4039471447,0.0100884913,0.147742331,-0.3837982714,0.2592802644,-0.3560678363,-0.0807946846,0.1573893428,-0.0986835659,0.0980850533,-0.2220797092,-0.2630308867,-0.0520919114,-0.1026492491,0.2872572243,-0.1842311323,0.099395372,0.166880101,-0.0009769483,-0.3264903128,0.0899213552,-0.0364115983,-0.1096254736,0.1086657047,0.1119769439,0.0754051954,0.2290828079,0.113952443,-0.2653917372,-0.2274806798,0.4559679627,-0.2483865619,0.1248017624,0.3189177811,-0.0318280458,-0.2407342494,-0.0966906846,-0.1099534407,0.0359343924,-0.7207142115,-0.2418959737,0.0526489392,0.3125581741,-0.1153778136,-0.1167013496,0.1347854882,-0.4312645495,-0.2894403636,-0.4584363103,0.1800460219,0.1059936285,-0.0408893265,0.1017270163,-0.4367369115,-0.4332711101,0.3100473881,0.0761299357,-0.17049402,0.0007381247,-0.1080607623,-0.1516113877,0.2151227891,0.2728155553,0.2304777801,0.0991623625,-0.015070091,0.0801523253,-0.1900815666,-0.1383890957,-0.1702752709,0.195865944,0.140555501,0.0211827867,0.198804155,-0.0894553885,-0.2101031095,-0.1073058024,0.2127431631,0.0108349398,-0.04499029,0.0538580231,0.0654742345,0.4179032147,-0.4134480059,0.101199314,-0.0769585595,0.8318890929,0.0498319417,-0.0228617694,0.5013242364,0.1181736588,-0.1698737293,-0.1642744541,-0.0966287255,-0.1165246144,0.3193700016,-0.2874974012,0.479201138,0.1438090205,0.1256595105,0.3302174211,-0.0593104996,-0.0504485592,0.200355038,0.2262878269,-0.1359418035,-0.1441453546,0.0161874276,0.028667517,-0.0199429095,-0.0807395652,0.1297767907,-0.2770057321,-0.210116446,0.0414084978,0.5093349218,-0.1377753168,-0.1708633602,0.4947827756,0.0559074655,0.3179510832,-0.0497510694,-0.0534840338,-0.1506352425,0.3966694176,-0.1369431466,0.3893026412,0.0201562513,0.1261627525,-0.4025039077,-0.479663372,-0.0644201487,0.0733944252,-0.123063013,0.2888692915,0.0823750943,0.3891579211,0.0758619085,0.0047952356,-0.4773793817,-0.015960522,0.0408849381,-0.0841685981,0.0504666567,-0.3593042493,-0.180089131,0.3654739261,0.0755967274,-0.1684400737,0.3236729801,0.103958033,-0.1719410866,-0.6745162606,-0.0464282036,-0.0719369799,0.0678930059,-0.2763384879,0.0687302053,0.2137111574,-0.1565642655,-0.0731754526,0.078323096,0.3924455345,0.3857099712,0.0832012072,-0.0141658625,0.0295007676,-0.0373414196,-0.2444286495,0.3866626322,0.2492734343,-0.0398115218,0.1174168065,0.0938084275,-0.2524935901,0.1372153908,0.0313970111,-0.0366429351,-0.2242295742,0.6495134234,-0.3568862677,0.1387893409,-0.1306098551,-0.3363532722,-0.3808221519,-0.0296450667,0.3237259984,0.0874636844,0.0638081208,-0.3256819546,0.034292303,-0.2143956721,0.2552116811,0.1064575389,0.1482562423,-0.2821870446,0.0813384578,-0.7299488783,-0.0173947252,0.1375167221,0.021474015,-0.0583585836,0.2993564904,-0.2897751033,0.150722757,0.2090243697,-0.0964548588,0.2572149038,0.3506225049,-0.2775310576,-0.0317808539,-0.3194119036,-0.0132925361,-0.0751288384,-0.2688682079,0.2787603736,-0.579687655,-0.0678595454,-0.0651756451,0.0039572301,0.1840030104,-0.3104124963,0.7294005752,0.0280633066,0.3549128771,-0.0816793591,-0.1018687561,-0.3929448426,0.0522326976,-0.0009313859,0.1937034726,-0.1403040141,0.297493428,0.0782982111,0.0930256546,-0.1326321512,-0.1125838831,0.1230463609,0.2825122178,0.0001771174,-0.3003875911,-0.0445052013,0.08815597,0.2207866907,0.3169263601,-0.130628705,0.1751397848,-0.1964043081,-0.2953552008,0.3967741728,-0.3647971749,-0.1716310978,-0.0875278488,0.1622874588,0.3407032788,-0.0746821016,-0.5330119729,0.0122238668,0.2591096163,-0.1060753763,-0.2506811321,0.1562738717,0.0640466958,-0.1545905918,-0.0293871965,0.2450350523,0.0311330631,-0.3094671071,0.3645014465,-0.149378404]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2123","title":"Problem downloading GEM wiki_auto_asset_turk dataset","comments":"Thanks for the answer! I updated the library but unfortunately it didn't solve the problem.","body":"@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family","comment_length":15,"text":"Problem downloading GEM wiki_auto_asset_turk dataset\n@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family\nThanks for the answer! I updated the library but unfortunately it didn't solve the problem.","embeddings":[-0.0587865263,-0.0158044044,-0.0622117221,0.4297458529,0.2670884132,0.1228095964,0.3928534985,0.2236767858,0.2212878466,-0.1171297878,0.2396745533,0.1797964722,0.1359751374,-0.0272461642,-0.1769780666,-0.1856086999,-0.0211512819,-0.0652547777,0.115899004,-0.236497283,-0.0488287993,0.1476558596,-0.1371296942,-0.0797877982,-0.294454515,0.1422534883,0.109025903,-0.0966864675,-0.3188710809,-0.0893697515,0.5248328447,0.0426798575,0.104640007,0.2937732637,-0.0001181829,-0.130558297,0.4365433753,-0.0896688178,-0.0736012608,-0.3886043131,-0.2674910128,-0.4644680023,0.122828126,-0.2762252986,0.1624894589,0.3609888554,-0.0902500376,0.1958525926,0.0432258993,0.0429700464,0.1975768656,0.2095094621,0.0969162583,0.0195710845,0.1852484345,-0.5094687939,-0.2003230155,0.0016359573,-0.2009952962,0.14576222,0.0539336614,0.1523119062,-0.1199778169,0.4393884838,0.1151884198,0.1440544873,-0.0374431796,-0.5540362,0.3409015536,0.386691153,0.8244818449,-0.0567953102,-0.3743310571,-0.0816556364,0.023174746,0.0925430655,0.108523801,0.4829389155,-0.4015575349,0.0412620194,-0.4422616363,-0.0078642312,-0.0662971437,0.2412119657,-0.338789314,0.2022657692,-0.1309896708,0.0536977425,-0.0731245354,-0.0159542561,-0.3736304641,-0.3164438605,-0.1815016121,0.3192771077,-0.3714425862,0.0912940875,0.071041517,0.0998858362,0.3099096119,0.3254319727,0.066599153,0.2206161171,-0.2415240854,0.0864646286,0.1192681789,0.3493227661,-0.0461632498,-0.3250125051,0.2454045266,0.1013378799,-0.0560542308,0.0919387788,-0.1385764927,-0.0094024967,0.2643097043,-0.3053127527,0.1134640053,-0.1134382635,-0.1167639494,0.3414708376,0.1263350099,-0.1051742435,-0.1748102009,0.2164576501,-0.2853922248,-0.0910964683,-0.1568110883,0.3669054508,-0.1331128031,0.123859629,-0.0553607121,-0.1852345169,-0.0421255752,-0.1193307415,0.2351108789,-0.2454438955,0.4053231776,0.1227270588,-0.2185782045,-0.1740950644,-0.0384303257,-0.0646868348,0.1097805351,0.2879380286,0.1469520181,0.8530852199,-0.1075139865,-0.1463439614,0.0196203515,0.0141644198,-0.2865085304,-0.1995422095,-0.1394772828,0.1369672269,-0.4959477484,-0.2546560764,-0.0600364171,0.2467330396,0.2154707462,-0.1666345149,-0.0750488713,0.0017655204,0.0643955842,-0.225756079,0.4632741213,0.565019846,-0.2095080763,-0.0411722474,-0.0933040231,0.0805045739,0.521325767,-0.0163652115,0.0188185591,-0.1924418211,-0.0697258338,0.3364168108,0.2614325881,0.0664996952,-0.6801207662,0.2812155783,-0.1046059132,0.2712674737,0.0341684297,0.1743239015,0.0815832764,0.2257787436,-0.3009543717,0.2981799841,0.0922938064,-0.1763691604,-0.3425973654,-0.4822602272,0.094160147,0.0634890646,0.1673563272,0.001553703,0.0862060934,-0.0384349413,0.3532193601,0.2944178283,0.228586182,0.0693795681,0.1378220767,-0.2775816619,-0.0133631947,-0.1105499566,-0.0200802181,0.4023308158,0.5374243259,0.055579517,0.0149236713,0.0906879902,-0.2551805079,-0.1314100325,-0.3177365363,-0.0211110059,0.049923975,0.1065335199,0.2246473879,0.3161019385,0.0528473407,0.0484450199,-0.0513598584,0.0509932525,0.0199053343,0.2010688931,0.0789448991,-0.1400044858,-0.1819867045,-0.101835981,0.010986506,0.1244708374,-0.2407616526,-0.0103460746,-0.2319487631,0.0784646794,-0.0483773239,-0.0947633162,0.1361846477,0.0908894092,-0.0671360642,0.0400876775,0.1734783351,-0.0024339766,0.0054791798,0.0786742717,0.2768079042,0.3784133792,-0.0100467112,0.217942968,0.3308675885,-0.0423128009,0.041567672,-0.0121263145,0.319530338,-0.2579827309,-0.1154844984,-0.0667319968,-0.0942967832,0.276560843,0.5614742041,0.1357508451,0.1383794099,0.2233991623,-0.3462817967,0.0054791309,-0.0036388051,0.5835415125,0.2726847231,0.0928585082,0.2022665441,-0.3967272937,0.0825068355,-0.2150828689,0.04674742,0.3341914117,0.0592321679,0.4049158096,0.1044787765,0.1467714608,-0.0523190424,-0.046055045,0.1257326007,0.1351308823,-0.0322471224,-0.0838046595,-0.2973432839,0.2576260865,0.0568199307,-0.4226104021,-0.1313071698,-0.2422665507,-0.1299089491,-0.0199533999,0.0348878689,-0.0770690963,-0.1166323125,-0.2157237083,-0.0580217838,-0.3938120306,-0.3648595214,-0.0981741101,-0.3577642739,-0.0483874716,0.1030182019,-0.0717924163,0.2805479467,-0.1344394684,0.055059921,-0.101571016,-0.2045387477,0.0478069186,0.0094397627,0.3803987205,-0.2353735417,0.4324938655,-0.1780560315,-0.054325711,0.1591545492,-0.0016208222,0.0407123975,0.223979786,0.0861021057,-0.1978636384,-0.0915054977,-0.3067968786,-0.370911777,-0.3473373652,-0.079403162,0.197931245,-0.036192432,0.0737538412,0.2497410625,0.0411729217,0.1624372303,0.0357378908,-0.0427038595,-0.2893684804,0.8389602304,-0.08569628,-0.4200907052,0.2790359557,-0.031976454,0.3168091774,0.1015236378,-0.5083874464,0.2361007035,-0.2012137622,0.1956342906,0.264921695,0.0713021085,0.0619724914,-0.3124967813,0.0994220376,-0.1459151655,-0.0104301423,-0.0337662883,0.1178023741,0.2901920974,0.020329725,0.2782325745,0.0409570262,0.827049911,0.247069031,-0.1513774544,0.4313257337,0.0368231758,0.029959105,0.0414868444,-0.1561195552,-0.1903298944,-0.244455561,0.161022082,0.1466153264,0.1853668988,-0.2713020742,-0.0617012046,-0.0257934686,-0.4412497878,-0.3100281358,-0.1846438199,0.1634546667,0.0020573931,0.2933793366,0.3087930083,0.0424859188,-0.2891049385,0.0491886772,0.5727034211,-0.010336251,0.111225605,0.1332457364,0.1212102324,-0.566857636,0.0943867043,-0.0872442722,0.0750029534,-0.0619020239,0.1163551733,0.1455450058,-0.1554677188,0.1266708374,-0.2956506908,0.141513437,-0.1435157061,-0.1039544269,-0.2190388441,-0.2531142831,0.0247079954,-0.0410186909,0.5399696827,-0.0086724963,-0.3333729208,-0.0198059957,0.0672420263,0.1862307638,-0.1550572067,-0.2899479568,-0.2812140286,-0.4243692756,-0.3161686659,-0.2002344131,-0.2121139318,0.3753821254,-0.2123799771,-0.3259663284,-0.1233878061,0.2697036564,0.1201604083,-0.0186981037,0.3297306299,0.0965065584,0.385278821,0.0309363212,-0.1214016676,-0.16869165,0.4905938208,0.3464444578,0.0216955189,0.0508864336,-0.324775219,0.0137795387,0.150174439,-0.2419336885,-0.0284689479,-0.0875913277,0.1964325011,-0.1293678433,0.0054139476,0.2434519529,-0.0866701156,-0.1824517548,-0.4118539989,0.6932030916,-0.2110639364,-0.0062163481,0.1687229425,0.1357955039,0.1372976452,0.287476331,-0.1023722813,0.7796039581,0.0372382626,-0.0381434523,0.0805434361,0.0932249352,0.3775441349,-0.2965007126,0.1193008348,-0.3466922641,0.1467105448,-0.1388730556,0.0215667617,-0.1108393967,0.0696525872,-0.1056377143,0.1579575688,-0.1517894566,0.2277568132,0.0172687322,0.0562571995,-0.3347537518,-0.0878748521,0.0381419584,0.1160797477,0.0225976054,0.6345829964,0.0280391574,-0.2459862679,-0.2544330359,-0.0858758092,-0.5527256727,0.1301882565,0.1021351814,-0.0735156015,-0.2077894509,-0.4644698501,0.0133315409,0.4557549953,-0.0619059131,0.1771388948,-0.3543609977,0.257129401,-0.3100244105,-0.0923838317,0.167240873,-0.0797313228,0.0913969129,-0.20076029,-0.281738162,0.0226855222,-0.0844685435,0.3061605692,-0.1218604073,0.1134619564,0.2433305979,0.0009579882,-0.388240546,0.1393717229,-0.0700264126,-0.1161784828,0.1268060952,0.1283707172,0.0412335061,0.2520283461,0.127209872,-0.2948503494,-0.2102971375,0.4106867313,-0.158999458,0.2068382353,0.3349499702,-0.077544421,-0.2120683938,-0.1101394445,-0.1027654782,0.0114795445,-0.6640030742,-0.235602349,0.0180307087,0.3191394508,-0.1411568075,-0.1435934007,0.0618617274,-0.3831599653,-0.3252959847,-0.4921446443,0.1815141588,0.0537649468,-0.080495812,0.0733151138,-0.4830565751,-0.4827947617,0.3086824715,0.1305304915,-0.1744904518,-0.019850634,-0.1611507982,-0.1481425911,0.1420597136,0.313167721,0.1254163235,0.0838895291,-0.0066865408,0.0587579347,-0.1612294614,-0.1645261943,-0.1593697518,0.2044788897,0.1713001728,0.0407910086,0.1893737167,-0.1347863376,-0.2309432775,-0.0500187688,0.1172371805,-0.0642767623,-0.0355321616,0.0900804549,0.0777717903,0.3938413262,-0.4785268009,0.1110048816,-0.1235759184,0.8345305324,0.0709334314,0.010860974,0.5638353229,0.1125245914,-0.2440788001,-0.2246654332,-0.1213443428,-0.0892061591,0.3536448181,-0.3242960274,0.5404544473,0.1011900678,0.0654297993,0.3443569839,-0.1175427884,-0.0703375116,0.163810119,0.2451671511,-0.1445290595,-0.1413402855,0.0469582714,0.0828547254,-0.0541487485,-0.0983459204,0.1374313086,-0.2931088209,-0.2466543615,0.0536004305,0.4892908931,-0.166257754,-0.1517481655,0.4848675132,0.032286007,0.3356488943,-0.044602707,-0.0770756379,-0.1822619289,0.4168893695,-0.0902733728,0.40122509,0.1206920519,0.094770059,-0.4106089175,-0.47484833,-0.0379820578,0.0738020241,-0.1574121118,0.2600326836,0.1158472672,0.2457820177,0.1320403367,0.0376958288,-0.4861235619,-0.0560169704,0.0261587445,-0.0010561227,0.0589064583,-0.3410566449,-0.1498718262,0.3593848944,0.0554412752,-0.1062339768,0.3108503819,0.1621976495,-0.1519841552,-0.6646815538,0.0370663591,-0.0892356783,0.0477504581,-0.3128109574,0.0338592492,0.1903337091,-0.136755839,-0.0480170771,0.0343597159,0.3321218491,0.4374755621,0.0460859314,-0.0144744869,0.0494404249,-0.0703593343,-0.2049508542,0.3290865421,0.2814106643,-0.0061533554,0.1292070001,0.1042325497,-0.2465543151,0.1785167605,0.0851551741,-0.0126343556,-0.2061502784,0.6849403977,-0.2915056944,0.1180257127,-0.1499754041,-0.3364810944,-0.3567934632,-0.0178078152,0.3538952768,0.1580621004,0.0585231483,-0.3384107947,0.0324024558,-0.2874608338,0.2553315163,0.0412574857,0.1805279702,-0.3415789604,0.1243434548,-0.7070644498,0.0228786543,0.1163410321,-0.023552414,-0.0181369092,0.3025934994,-0.2937301695,0.1566995829,0.2449887544,-0.027168097,0.2641110718,0.4281102419,-0.2778222859,-0.0520761982,-0.3267408609,0.0171805006,-0.1209492907,-0.2831884921,0.2753616869,-0.6043049693,-0.0436366014,-0.0505574085,-0.1020500809,0.2819584906,-0.2806486785,0.7948426008,0.0071471161,0.3984848261,-0.0808877125,-0.030594686,-0.2903903127,0.0703808814,-0.0024366945,0.1802479476,-0.0999560803,0.2721870244,0.1150143817,0.1806759685,-0.1394241154,-0.1352005452,0.0655872673,0.2888424695,-0.0020830892,-0.3338968754,-0.0182829592,0.0956825688,0.2501882911,0.2500023246,-0.1201582551,0.1561428159,-0.1740109921,-0.253035754,0.3330867887,-0.2840063572,-0.2058106214,-0.1195724085,0.1391237229,0.3153036833,0.008200163,-0.4828844666,0.0033449808,0.298607856,-0.1223255545,-0.2251889259,0.123524785,0.1208891422,-0.0898073614,-0.0301627312,0.2065233439,0.0076511372,-0.2838270664,0.2807791531,-0.1461003423]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2123","title":"Problem downloading GEM wiki_auto_asset_turk dataset","comments":"Is there an error message ?\r\nWhat stacktrace do you get if you interrupt the execution of the program while downloading ?","body":"@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family","comment_length":22,"text":"Problem downloading GEM wiki_auto_asset_turk dataset\n@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family\nIs there an error message ?\r\nWhat stacktrace do you get if you interrupt the execution of the program while downloading ?","embeddings":[-0.1321628094,0.0142390393,-0.0815602541,0.4399939477,0.2992451489,0.140218094,0.3794731498,0.147788018,0.2320196778,-0.0426291004,0.2344708741,0.1440670043,0.1074127555,-0.0218766388,-0.1961918324,-0.141847834,-0.0774908215,-0.0699517354,0.1331875622,-0.1573004574,-0.0862878039,0.1746144444,-0.1601432264,-0.0402904339,-0.2822526991,0.0316642597,0.1252023727,0.0306358933,-0.262068212,-0.1533477008,0.5731248856,0.0543964766,0.0777289867,0.4127718508,-0.0001159479,-0.0370518938,0.4838816524,-0.080642283,-0.0932808295,-0.4343478084,-0.2929488719,-0.428958118,0.1403427571,-0.3042384386,0.2589861751,0.3699575067,-0.0471938103,0.1229425743,0.1223153099,0.0604695305,0.2033275515,0.2653098702,0.0952295139,0.0126955928,0.2613676488,-0.5051375031,-0.1550694555,0.0339410566,-0.2403400242,0.2069715708,0.0613532253,0.1425447017,-0.1477518827,0.4220528305,0.0742655471,0.1051860005,-0.0462688878,-0.6357185245,0.3205806911,0.4154242277,0.785974741,-0.0836304799,-0.3444286585,-0.0804104954,0.0373599268,0.0171118677,0.1368553638,0.4656033814,-0.4767647684,0.0495452657,-0.3980374336,0.0067694043,-0.040577624,0.1948126405,-0.3115484118,0.1401275843,-0.1546680778,0.0738654211,-0.0989530906,0.0863773525,-0.2981093824,-0.3427293301,-0.1150037423,0.2818672061,-0.4654174447,0.115643926,0.131799683,0.0213572085,0.2442520708,0.2795456648,0.0694085807,0.2185829431,-0.1733071506,0.1026939377,0.2011168897,0.2909263372,-0.1033421382,-0.3722575903,0.2547006309,0.1148330793,0.0021217789,0.0427103639,-0.1206836402,0.0035349797,0.2396795452,-0.2869566679,0.2340714931,-0.1200394854,-0.2040838152,0.3872778714,0.031422168,-0.0665427744,-0.1857517809,0.2559370995,-0.1681705415,-0.0881124362,-0.1446923018,0.3099090159,-0.1549852788,0.1047627032,-0.0853241384,-0.1981688738,-0.0362772048,-0.1300715506,0.2148918211,-0.2713961005,0.3452538252,0.124105826,-0.1539851576,-0.1573804617,-0.0081057372,-0.1100425795,0.1143710092,0.3390209377,0.1603041887,0.8072764874,-0.0744100437,-0.1066129878,-0.0066189403,0.0822717994,-0.2562863529,-0.2304996699,-0.0900202468,0.157515049,-0.4614281058,-0.1936807632,-0.0424200147,0.2153875083,0.1626544297,-0.1899282336,-0.0849236324,-0.0099527137,0.0580629371,-0.251806289,0.4788047373,0.5534905791,-0.3083427846,-0.0215373598,-0.1111752093,-0.0194274858,0.4509563446,0.0034095438,0.0026833799,-0.034290757,-0.0746415108,0.3815183342,0.3053486347,0.0375183113,-0.6943234801,0.2703822851,-0.1899664253,0.2698354423,-0.0633267462,0.1415518224,0.132953316,0.181542486,-0.3257965446,0.2547839582,0.0073143984,-0.1606895924,-0.3478211462,-0.5174865723,0.1066831276,0.0958976746,0.143538624,-0.0442353822,0.1080907136,-0.0246796273,0.3150688112,0.2980012894,0.2028721124,0.1548967808,0.1308111548,-0.2210355848,-0.0933985561,-0.1403895915,-0.0056310981,0.3707253635,0.5322477221,0.06015772,-0.0091592455,0.0486918725,-0.3059341609,-0.0659251139,-0.2911716104,-0.0881628916,0.0756877288,0.0847407654,0.2108381987,0.3549518585,0.0124003189,0.0866913199,-0.0655509457,0.0727596283,0.0445239395,0.2100453377,0.0401301458,-0.1469335556,-0.2157851309,-0.1121001095,-0.035180036,0.0463373177,-0.2612461746,0.1182314754,-0.232616365,0.0806273818,-0.0174806062,-0.1799469739,0.0855987445,0.0720395893,-0.077047281,0.0010036328,0.2271564752,-0.020828791,0.064842023,0.0480485633,0.2259175926,0.3803313076,-0.0249135811,0.1595377773,0.341937989,-0.0640395507,0.0453701466,0.0332495421,0.3400067091,-0.1862511784,-0.1197081655,-0.167440787,-0.1324003637,0.1907773018,0.5834674835,0.0761657432,0.0721691549,0.2146659344,-0.2984449863,0.0452684946,0.0092086466,0.5815960765,0.3240529597,0.1171023101,0.2098080367,-0.3470689952,0.0677618086,-0.2248481065,0.0345480777,0.384731859,0.0652724132,0.4256387651,0.1201915368,0.1514938176,-0.0837181211,-0.1162302792,0.1635286361,0.1416743696,-0.0068451497,-0.0428998396,-0.232544288,0.1540494859,0.059819527,-0.4542242885,-0.1608757824,-0.2700364888,-0.1119742841,-0.0115425466,0.0120171644,-0.0320087597,-0.1950570345,-0.1726177037,-0.0271611176,-0.3007733524,-0.3105888963,-0.1425126046,-0.3204442263,-0.0535465032,0.1348847449,-0.0596891083,0.3072903156,-0.1035958678,0.0955520645,-0.0590402149,-0.1923726052,0.1506723613,0.0209468324,0.3641676307,-0.1945985258,0.4104298651,-0.1434186101,-0.1419550478,0.1534850895,-0.0161835942,0.0091253445,0.237668395,0.1352460831,-0.1796499193,-0.1143886149,-0.3453753889,-0.3516607583,-0.3800868094,-0.0870520249,0.159319222,-0.0395010971,0.0151041318,0.1947156936,0.1441307217,0.0985892713,-0.0364617966,-0.0853472054,-0.3460047543,0.7302365303,-0.1546864063,-0.3846496642,0.296200335,-0.0790098533,0.2984314859,0.1366481334,-0.5572382808,0.2780165374,-0.2417080104,0.2090036273,0.1866083741,0.0398764201,0.1015410498,-0.2553366125,0.0548728704,-0.126894623,-0.0496451035,0.008050113,0.0881617814,0.313269943,0.0028204548,0.3143161237,0.021727303,0.909424305,0.280085355,-0.1164816618,0.3772396743,0.0508962311,0.0764647126,0.0019647577,-0.1339125037,-0.1527619213,-0.2287874669,0.0307079833,0.1379284114,0.1020733044,-0.3040210605,-0.1034987718,0.0082995379,-0.4625451565,-0.4070031941,-0.203289628,0.149433732,0.0558307916,0.2711265981,0.3171423078,0.0534302033,-0.1946010292,-0.013960734,0.5690128803,-0.0494956858,0.0350785367,0.0990369096,0.2003506124,-0.5515704155,0.1002979055,-0.0405991413,0.0772852004,-0.0727350637,0.0966601595,0.1613287926,-0.1698454618,0.1840687245,-0.2604211271,0.088488318,-0.0746303052,-0.0590104684,-0.3061174154,-0.2616580725,-0.0627489612,0.000179946,0.577365756,-0.087128967,-0.335807234,0.0449208394,0.0969548896,0.1366341412,-0.1719824076,-0.2932936251,-0.2531768978,-0.487562865,-0.3430597782,-0.0970969349,-0.1672285944,0.2974107862,-0.2426985949,-0.3157044947,-0.1075708643,0.3118188381,0.0927711353,0.0217722878,0.2937594652,0.04718858,0.3761806488,0.0846757814,-0.1619000882,-0.0890385956,0.5246004462,0.4296529591,0.0228990708,0.095495373,-0.3499961793,0.0399574414,0.2479242086,-0.1904250681,0.0160617195,-0.0825030804,0.2524565756,-0.1452445835,0.0414055698,0.3016324937,-0.0441925339,-0.2093584388,-0.34148103,0.6050920486,-0.24462457,-0.0457312316,0.1185271442,0.1404813379,0.1521067172,0.2632791698,-0.0781994313,0.8599671125,0.0772292167,-0.0774453431,0.0988998786,-0.0084116366,0.4027971029,-0.2912654877,0.1771308035,-0.2803229094,0.0132391425,-0.1299609542,-0.0158460867,-0.1059279591,0.0655045509,-0.1470301151,0.0647296831,-0.1514601707,0.1660868824,-0.0288571417,0.048592072,-0.378179431,-0.1238504872,0.0621935464,0.1417936981,0.0101720048,0.64499259,-0.0319577642,-0.2541676462,-0.2328561246,-0.0989849269,-0.592376709,0.1134366617,0.0727088153,-0.0870582461,-0.25816378,-0.4596537352,0.0500055961,0.380504638,-0.0211776625,0.1624511629,-0.3153133094,0.2178965062,-0.3071716428,-0.0679484159,0.1469367594,-0.0749857649,0.1409620941,-0.1598261744,-0.3013408184,0.045402199,-0.1186192855,0.2993142307,-0.0679529533,0.0814534277,0.2012623399,-0.1159828305,-0.3673446774,0.1397227943,-0.0426697396,-0.1423517913,0.1272453666,0.1044211984,0.0396638922,0.2415796518,0.142775923,-0.3163481057,-0.259321928,0.3870818317,-0.2076293528,0.2182009816,0.3941337764,-0.0709004253,-0.247913599,-0.1153725237,-0.0883026347,-0.1054180712,-0.6121993065,-0.2186025083,-0.0251923576,0.3561534882,-0.1614972651,-0.133636117,0.1048504412,-0.349083364,-0.3628377914,-0.5235509872,0.1306188554,0.0656849369,-0.0233989786,0.1781970412,-0.5028347969,-0.3716757596,0.3024107516,0.0882698298,-0.2060586214,-0.0062338002,-0.2363537997,-0.1376996785,0.2346299738,0.2215419859,0.1833380163,-0.0016463321,0.0138037745,0.0670188591,-0.1924348474,-0.1775333434,-0.1876377165,0.1868732572,0.1435885727,-0.0046314332,0.2212842405,-0.1453614682,-0.1825259775,-0.0043025375,0.1372854114,-0.0621717758,-0.0482090488,0.1047598198,0.0236889981,0.5183144212,-0.515622735,0.1428273022,-0.1296880394,0.8146842122,0.0847086087,-0.0049638725,0.5043003559,0.1296807528,-0.2271964997,-0.2399452627,-0.0662696585,-0.0619187914,0.313711971,-0.3795987964,0.5551661253,0.1559939086,0.1405550987,0.4373143911,-0.1152108908,-0.1067775041,0.1859596968,0.2554093599,-0.1378963739,-0.0159421246,0.1133424789,0.0164480079,-0.10426943,-0.0887304619,0.0588719212,-0.1715670824,-0.201430589,0.0903036147,0.4766493142,-0.1343524009,-0.1119906828,0.4683966041,-0.043400161,0.3703645766,0.0106776711,-0.0769262612,-0.1106141359,0.4471071661,-0.1143291444,0.3792330027,0.0593007989,0.1455963254,-0.3888528645,-0.5082745552,-0.0176250059,0.0917017758,-0.1510202438,0.2723236978,0.1123315766,0.241851002,0.0616753101,0.0401139222,-0.4717873037,0.0185601413,-0.0190473646,0.0652198493,0.1212086156,-0.3413903713,-0.0984334424,0.3095863163,0.0807391405,-0.0661867782,0.3439118862,0.1642588079,-0.1780664176,-0.654658854,0.0632569268,-0.116725415,0.0596520789,-0.304633528,0.0001109768,0.2576384246,-0.1148885414,-0.0486506261,0.1723043025,0.3171806633,0.4272041917,-0.0924460739,-0.0012717962,0.1156506687,-0.1001414061,-0.1569690108,0.3286314607,0.2741537392,0.0744791925,0.2132534683,0.1276147068,-0.2593202293,0.2150091678,0.1319011003,0.0374735482,-0.2750089467,0.663490057,-0.2822513282,0.1430800557,-0.1848209947,-0.3486415744,-0.3625610173,-0.0404625796,0.310590595,0.0405470431,0.0489031672,-0.2859012783,0.0431647226,-0.2818432152,0.2512695193,0.1280829012,0.2027846873,-0.3090598881,0.1365196705,-0.7200751305,0.0663313866,0.0720290244,-0.0794796869,0.0131459869,0.3335933983,-0.2754961848,0.261502713,0.3244893551,-0.1350855678,0.1313529611,0.4813817739,-0.2831988931,-0.0716874301,-0.3385829329,-0.1188428178,-0.1176453009,-0.2906943262,0.2533259988,-0.5135531425,-0.0406851098,-0.0369684733,-0.102959998,0.2287245393,-0.3376685679,0.771841526,0.0204327479,0.4181722999,-0.0625889525,-0.047400739,-0.2577982843,0.0249536633,-0.0442730747,0.1935827285,-0.0700871497,0.2800270915,0.0949934646,0.1358511001,-0.1560492665,-0.1138925627,0.0798364356,0.3366306722,-0.0748944357,-0.3074180782,-0.0118571045,0.0736624226,0.3071089387,0.2668800652,-0.0882935375,0.2186746597,-0.1876818091,-0.2498739213,0.3628955483,-0.2184950709,-0.1908559948,-0.0703238472,0.1019258872,0.2778055668,-0.0298840161,-0.4902449548,0.0206854492,0.3025434911,-0.1489333808,-0.2680254877,0.1509250402,0.063117221,-0.1221382469,-0.0223945826,0.2566050589,-0.0247098394,-0.2528830469,0.2719268501,-0.0967396423]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2123","title":"Problem downloading GEM wiki_auto_asset_turk dataset","comments":"Sorry for the long time since my last comment, I tried again and don't seem to have the problem anymore, thanks for your support!","body":"@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family","comment_length":24,"text":"Problem downloading GEM wiki_auto_asset_turk dataset\n@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family\nSorry for the long time since my last comment, I tried again and don't seem to have the problem anymore, thanks for your support!","embeddings":[-0.0690356866,-0.0099978196,-0.0715510249,0.4162681401,0.2822793126,0.1222887412,0.3851531148,0.2134170383,0.2334703952,-0.1133927181,0.2428507507,0.1628344953,0.1382036656,-0.0105028385,-0.1870101988,-0.1910438836,-0.0316886716,-0.0682897791,0.1373938471,-0.2212913483,-0.0542819127,0.1533599347,-0.1293253452,-0.0846192688,-0.2987632453,0.1394868642,0.1094797105,-0.085387446,-0.3324198425,-0.0843059495,0.5183299184,0.0374567769,0.0902131796,0.2967351973,-0.000117695,-0.1262214333,0.4400681555,-0.080070436,-0.0858739391,-0.4032157362,-0.2881535888,-0.458510071,0.1159151644,-0.2507407665,0.1574537903,0.3613402545,-0.0828236639,0.2039815336,0.0602639616,0.0538335852,0.2015053034,0.2203440219,0.1119733229,0.0012682587,0.1720883697,-0.5136726499,-0.1961065084,-0.015647037,-0.2056452334,0.1159265116,0.0759868249,0.1538500339,-0.1160210818,0.4487346113,0.1005097404,0.1362582892,-0.0333033577,-0.5621496439,0.3449827433,0.3804026544,0.8352507949,-0.0441537835,-0.3848358095,-0.0902389884,0.0297753345,0.077689372,0.101225704,0.4795999527,-0.4089401364,0.0576068014,-0.4534149766,0.0004345138,-0.0540678427,0.2307631969,-0.3505893946,0.2053674459,-0.1281363964,0.0540480874,-0.0512708426,-0.022917768,-0.3819190264,-0.3187372983,-0.1858948916,0.3288645744,-0.3765308857,0.0876472816,0.081802547,0.1013025567,0.3065651655,0.3332847059,0.0595947914,0.2026231885,-0.2427235991,0.0974041671,0.1333460063,0.3499379158,-0.0496277288,-0.3339032233,0.233652994,0.0831762031,-0.0409998335,0.0889140368,-0.1327002347,-0.0273617897,0.2532709241,-0.2916349769,0.1138650849,-0.1210354045,-0.1329976618,0.3446778953,0.1328086406,-0.103574574,-0.1729872823,0.2109665573,-0.2614233792,-0.074850291,-0.1549175531,0.3760821223,-0.1342865229,0.1245543286,-0.0551476106,-0.1946308017,-0.032064341,-0.1308616251,0.2359174937,-0.2384796739,0.3978219926,0.1198453903,-0.2069476247,-0.1694868803,-0.0482036918,-0.0695440844,0.0892040208,0.302932173,0.1519700736,0.8524777889,-0.1055618376,-0.1533081383,0.0277879871,0.0271231532,-0.2666284442,-0.2013563663,-0.1391534954,0.1449596882,-0.4829969108,-0.2651830316,-0.0555663481,0.2311142832,0.2233172655,-0.1439942271,-0.0843165517,0.0087370137,0.0678068474,-0.2231290042,0.4462706745,0.5631440282,-0.2154604197,-0.0273493417,-0.1023787484,0.082902953,0.5172902942,-0.0132263666,0.0177302659,-0.1962208897,-0.0552291609,0.3503416181,0.2345134169,0.0634287074,-0.66987288,0.2986477911,-0.1121719927,0.2659198046,0.035139747,0.1727237254,0.0882032365,0.2288590372,-0.2764081359,0.3082546294,0.074830249,-0.1855424196,-0.3515208066,-0.4687909484,0.0907577798,0.0779029503,0.1690832973,0.0095519163,0.066642113,-0.0545492247,0.3368754685,0.2913600802,0.2231932431,0.0540956371,0.1401033401,-0.2684960663,-0.0352166593,-0.1048107073,-0.0164095256,0.4072797298,0.5592125654,0.0616346374,0.0100326911,0.097896859,-0.2644705772,-0.1222736537,-0.3213399649,-0.0197784025,0.0524631739,0.0943089351,0.2169118226,0.2984938622,0.0374979712,0.0401896127,-0.0420678146,0.052352611,0.0416194201,0.1940546185,0.067439422,-0.133195743,-0.1795913577,-0.1060785726,0.0163922012,0.1452955157,-0.2291959822,-0.0028151597,-0.2187502235,0.0633663014,-0.0434074067,-0.0914289057,0.1379369795,0.0981813371,-0.0766848847,0.032803338,0.1698219776,-0.0115782395,-0.0029624163,0.0853825659,0.2800171673,0.375519067,-0.0150109995,0.2399068624,0.3355819285,-0.0406278931,0.0454286821,-0.014503235,0.3255586624,-0.2656667829,-0.1476519108,-0.0680989102,-0.0925351679,0.2774904072,0.5297239423,0.1324802637,0.1280038059,0.2106357217,-0.3184182644,0.0189467445,-0.0144940224,0.5863617659,0.2611171901,0.0961605534,0.2073136568,-0.4086148739,0.0570140183,-0.2168998271,0.0457976833,0.32647264,0.0644427836,0.3982263505,0.101974763,0.1328803003,-0.0371610709,-0.0382136144,0.1255006641,0.1326801181,-0.041086629,-0.085897699,-0.2936616242,0.2567718327,0.0599694885,-0.4399902225,-0.1302776337,-0.2698121965,-0.1121220961,-0.0198086612,0.0327046216,-0.0770640299,-0.124431923,-0.2014309764,-0.0588909648,-0.3738242686,-0.3677837849,-0.1252192706,-0.3382562697,-0.0427056365,0.1071572006,-0.0983289182,0.2873528898,-0.1204516813,0.0573703423,-0.0958856121,-0.1930486262,0.039848987,-0.0043083508,0.3839710355,-0.2377865762,0.4564390182,-0.1663805097,-0.0597173534,0.1663670242,0.0243512075,0.0365823619,0.2220618278,0.1051800102,-0.2187347859,-0.098590076,-0.2855588496,-0.3799356818,-0.3414305151,-0.0858326033,0.2054734975,-0.0430646613,0.0882582217,0.2635269463,0.034782052,0.1744258255,0.0250118729,-0.0449167714,-0.2868692577,0.8346586823,-0.0881904364,-0.4069334865,0.2835735381,-0.0147732254,0.314887464,0.1019528434,-0.5072258115,0.2176256478,-0.2097766548,0.1955391765,0.2795872688,0.0657097325,0.0610905699,-0.3151111901,0.0881062672,-0.1394304186,-0.0146766808,-0.0487380736,0.1244137511,0.2875226438,0.0122674014,0.2843146026,0.0379387587,0.8390620947,0.2417924404,-0.1325186193,0.4327153563,0.03403043,0.0223410707,0.0224244557,-0.1650138348,-0.2018046081,-0.2494834065,0.1537341028,0.1278745532,0.1948740482,-0.2716963589,-0.0661297813,0.009021204,-0.4472631812,-0.3024114072,-0.1900714636,0.155117169,0.0075884718,0.2950231433,0.3235775828,0.0450513549,-0.2776399553,0.0365815349,0.5699430108,0.0019077015,0.102377601,0.1366873533,0.1184010506,-0.5679132938,0.0903750658,-0.0908189937,0.0821970999,-0.0941473097,0.1145556569,0.1419597566,-0.159344852,0.1375416964,-0.3117362559,0.1319679469,-0.1461318582,-0.1067251712,-0.2218557745,-0.250451982,0.0402816646,-0.0445268303,0.5353951454,-0.013035981,-0.3280189633,-0.00680135,0.0755776539,0.1726426035,-0.1590393782,-0.2739726305,-0.2760903239,-0.4223755002,-0.3200936019,-0.2010694891,-0.1957078427,0.3888675272,-0.2165103108,-0.2829757929,-0.1278852373,0.2633854449,0.1165942028,-0.0192346312,0.3295143545,0.0984780565,0.3884229064,0.0532442927,-0.1283080131,-0.1797626615,0.5015987754,0.3309796453,0.0123755354,0.0413731337,-0.3306532204,0.0061595128,0.1630473733,-0.2309937924,-0.0541749932,-0.0837312192,0.1961094588,-0.1559123397,-0.00533253,0.2369254082,-0.0777207464,-0.1833666712,-0.418269068,0.6890941858,-0.2060820311,-0.0139103057,0.15960069,0.156472981,0.1464679986,0.2997373343,-0.1261126846,0.7728821039,0.0346661508,-0.0354423039,0.0785008073,0.0871515125,0.342682302,-0.2902925611,0.1275341511,-0.3353431523,0.1532087922,-0.1406503618,0.0165487509,-0.1293691099,0.0847421363,-0.0910067633,0.1499933898,-0.1677743793,0.2257798612,0.0222394075,0.0446507223,-0.3492172062,-0.0817788392,0.0447299667,0.1272879541,0.0269737951,0.6393930912,0.0258919708,-0.240983054,-0.2464602888,-0.0886286721,-0.5789681077,0.1253204793,0.1126068234,-0.0713916719,-0.2029535174,-0.4535733759,0.0131638087,0.4507681131,-0.0583236292,0.1790383905,-0.3476181924,0.2616402805,-0.2932385802,-0.0864886567,0.1592841744,-0.0832831636,0.0926180258,-0.1949463487,-0.2863840461,0.0278712045,-0.0813430175,0.3039565086,-0.1432856917,0.1170967892,0.2210619003,-0.0046928981,-0.3729781806,0.1475703865,-0.0583352409,-0.1162187904,0.1300281286,0.115107365,0.0481416807,0.2377817035,0.1229675859,-0.2875977159,-0.2023488134,0.4077857137,-0.1703431159,0.2058811486,0.3319360018,-0.0652091727,-0.2308104783,-0.1163452342,-0.1078141779,0.0193428118,-0.6593101621,-0.2220483422,0.0155071458,0.3092422485,-0.1467106342,-0.149970293,0.0767485127,-0.3893268704,-0.3052190542,-0.5143172741,0.1934581995,0.0530862957,-0.0808657929,0.101281859,-0.4944117963,-0.4612378776,0.3130162358,0.1234032065,-0.1799676865,-0.0137859099,-0.1672127247,-0.1403013766,0.1541434228,0.291566968,0.1243613288,0.0652921721,0.0008926559,0.0452474393,-0.1702832729,-0.1682329476,-0.1493078321,0.1990280598,0.1847804189,0.0080669802,0.1910301596,-0.1320530623,-0.2113071978,-0.0451429039,0.1274879277,-0.059308596,-0.0306266062,0.092267409,0.0805762783,0.4037798941,-0.4537477195,0.1145540848,-0.1015148833,0.8488927484,0.0857639611,0.0158931985,0.5661738515,0.1179383844,-0.2378250062,-0.2062019557,-0.126663357,-0.097764127,0.3483295441,-0.3401979804,0.5501438975,0.1126512215,0.081764169,0.3406642377,-0.1266597956,-0.0663280264,0.1710990518,0.254517138,-0.147562474,-0.1292316616,0.0592223443,0.0956982821,-0.051793199,-0.1130307317,0.1341887861,-0.2904935777,-0.2642148733,0.0658190548,0.4888225496,-0.1714356244,-0.155843854,0.480555594,0.0278118905,0.322576046,-0.050470043,-0.0692567378,-0.1912908554,0.4133712351,-0.1041399017,0.3994583189,0.118431583,0.0900519565,-0.4028489292,-0.487683326,-0.0391133912,0.0621094964,-0.1485470086,0.2471346408,0.1171719953,0.2374732196,0.1296532303,0.0320803262,-0.4850429595,-0.0397876464,0.021188993,-0.0166226402,0.0458935462,-0.352101177,-0.1465231627,0.364435941,0.0529252142,-0.1100594401,0.307585448,0.1532521397,-0.1521400511,-0.6718458533,0.0426189639,-0.0922215655,0.0442210808,-0.3201610148,0.0263662748,0.2077568322,-0.1201172248,-0.0395674482,0.0311254542,0.3577190638,0.4340398014,0.0124244802,-0.0142763285,0.0483524203,-0.0698326156,-0.2068817616,0.3337103128,0.255448103,0.0059758364,0.1312027574,0.1121164411,-0.2471866608,0.176115945,0.082303904,0.0013018795,-0.2061569691,0.6831783056,-0.2835208476,0.1168932989,-0.1556620151,-0.334128201,-0.3487711549,-0.0198614858,0.3473852277,0.15197815,0.0724641606,-0.3158144355,0.0337596238,-0.2807039618,0.2559372485,0.074062027,0.1723389924,-0.3593784869,0.1387785375,-0.7013583183,-0.007882297,0.0928428397,-0.0245149732,-0.0234822389,0.3137425184,-0.2782695591,0.1713257134,0.2388454378,-0.0347119123,0.2462118715,0.4393148124,-0.2718909383,-0.0482837334,-0.3086026907,0.0126586398,-0.1189718246,-0.270419091,0.2698473334,-0.5997475386,-0.0356084295,-0.0418143161,-0.0972709805,0.2781628072,-0.2821846008,0.7650135756,0.0147422375,0.3980948627,-0.0567809381,-0.038257774,-0.2995532453,0.0535393357,0.0010870744,0.1728666723,-0.1007198691,0.2720145285,0.1195932776,0.1841214597,-0.1436821073,-0.1453594565,0.0513153151,0.2917464375,0.0047726566,-0.3444966674,-0.023924334,0.0892062262,0.2369011045,0.2461632192,-0.1208509728,0.1756345481,-0.1573687792,-0.2548574209,0.3296071887,-0.2776281536,-0.2133744508,-0.103181012,0.1321400106,0.3171691298,0.0198237821,-0.4789212644,-0.0028051757,0.2805309892,-0.1406555623,-0.2352169305,0.1149384677,0.1322642267,-0.1027849019,-0.0249515101,0.1947855651,0.0044229119,-0.2891602814,0.2658874393,-0.1413049847]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2123","title":"Problem downloading GEM wiki_auto_asset_turk dataset","comments":"Great ! I'm closing the issue then. Feel free to re-open if you experience this issue again","body":"@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family","comment_length":17,"text":"Problem downloading GEM wiki_auto_asset_turk dataset\n@yjernite \r\n\r\n### Summary\r\n\r\nI am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code.\r\n\r\n### Steps to reproduce\r\nCode snippet:\r\n\r\nfrom datasets import load_dataset\r\n#dataset = load_dataset('gem', 'web_nlg_en')\r\ndataset = load_dataset('gem', 'wiki_auto_asset_turk')\r\n\r\n```\r\n\r\n**Expected behavior:**\r\n\r\nI expect the dataset to start downloading (download bar appears and progresses toward 100%)\r\n\r\n**Actual behavior:**\r\nInstead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more:\r\n\r\nDownloading: 36.6kB [00:00, 37.2MB\/s]\r\nDownloading: 41.7kB [00:00, ?B\/s]\r\nDownloading and preparing dataset gem\/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\\Users\\sfmil\\.cache\\huggingface\\datasets\\gem\\wiki_auto_asset_turk\\1.0.0\\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d...\r\n\r\n### Is this a regression?\r\nNo, it was the first time I was trying to download this dataset (same for the other ones).\r\n\r\n### Debug info\r\n- Python version: Python 3.8.2\r\n- OS version: Windows 10 Family\nGreat ! I'm closing the issue then. Feel free to re-open if you experience this issue again","embeddings":[-0.0635910705,0.0054957317,-0.0699880645,0.3975265026,0.2704818249,0.1117716953,0.4051198661,0.2125456631,0.2219942659,-0.0980705991,0.2423038036,0.1652891636,0.1437678784,-0.0094379121,-0.1849551201,-0.1691983789,-0.0209271722,-0.0780066624,0.1397191435,-0.2287252843,-0.0718829259,0.1454975754,-0.1185270548,-0.0883480906,-0.2948794365,0.146615684,0.0971317291,-0.0787394345,-0.3126651645,-0.0990360752,0.5155933499,0.056869749,0.0788812563,0.2951326072,-0.0001170882,-0.1244743913,0.4406700134,-0.0718017891,-0.1148906499,-0.3813538849,-0.2894219756,-0.4607839882,0.1017347574,-0.2550836205,0.1578007936,0.3514332175,-0.0695673302,0.1909139305,0.0411403365,0.034796942,0.2039115578,0.2345373333,0.0773691535,-0.0092191705,0.1685435772,-0.508143127,-0.2009909451,-0.0149710234,-0.1961533278,0.113062039,0.0518401638,0.1712566763,-0.1112728342,0.4450575709,0.1092763469,0.1475444883,-0.0390336178,-0.5593597293,0.3386213183,0.3914212286,0.8110873699,-0.0349254459,-0.3909861147,-0.1163943782,0.0144400205,0.0883416012,0.109456338,0.473174572,-0.4042176604,0.0580026209,-0.4430180192,0.0162063446,-0.0424059331,0.240130052,-0.3565360904,0.2099107653,-0.1218152046,0.0495125279,-0.0501043424,-0.0193964932,-0.3825842738,-0.3250644207,-0.1779280454,0.3222630024,-0.3595634401,0.0945246145,0.1082110479,0.0944507793,0.3044976592,0.3384501636,0.0650235713,0.2179397345,-0.2423131913,0.0910515338,0.125758633,0.3379720151,-0.0297967251,-0.3428627849,0.2374131829,0.0910969079,-0.0390360728,0.0894546658,-0.1421337873,-0.0136255389,0.240049243,-0.299739182,0.1066754982,-0.1061817706,-0.1355123669,0.3545142412,0.1436535269,-0.1008923575,-0.1819547713,0.2096240073,-0.2602635324,-0.079368107,-0.1583039761,0.3832471371,-0.1480674297,0.1088962555,-0.0646661595,-0.2023756504,-0.0430512875,-0.1350317597,0.2424610555,-0.2153110206,0.3975866735,0.108227618,-0.2019285262,-0.1622778475,-0.0482724346,-0.060051363,0.1039723456,0.3016758561,0.1376928836,0.8678787947,-0.0983777121,-0.148656249,0.0211304724,0.0338245668,-0.2486865222,-0.1927576214,-0.1243552044,0.1496052444,-0.5011072755,-0.2612002194,-0.047611203,0.2438677847,0.2210102677,-0.1569941491,-0.0882685781,0.0083466312,0.0572759919,-0.2188308388,0.4434719384,0.5722365975,-0.2253334671,-0.0457061864,-0.111920245,0.0897270963,0.508541584,-0.0067210724,0.0091117164,-0.1866528094,-0.0492388383,0.3607643843,0.2355579436,0.0613823123,-0.6746481061,0.2866029143,-0.126722008,0.3046183586,0.0336515307,0.1798470914,0.0886608064,0.2086722702,-0.2679759562,0.3060261905,0.0844401345,-0.1812129766,-0.3593185246,-0.4738073051,0.0832657292,0.0717399344,0.1719872355,0.0067714499,0.0662724376,-0.0591569953,0.3414898813,0.277741313,0.2246892005,0.0503941961,0.1414612681,-0.2738792896,-0.0399673544,-0.0855868533,-0.0241376627,0.4271655381,0.5506248474,0.0638864413,0.0144227007,0.0955625251,-0.246964097,-0.1217616498,-0.3224138319,-0.044006791,0.05783673,0.1223813221,0.2030841559,0.3038417101,0.0479569882,0.0396658629,-0.0411104038,0.0564445034,0.0390907601,0.1793064624,0.087945044,-0.1351893246,-0.2017707825,-0.098815985,0.0059030135,0.139186129,-0.2159442008,0.0058312127,-0.2344039381,0.0820512027,-0.0509567596,-0.0871933699,0.1415044367,0.1149088368,-0.0709475204,0.0311205629,0.1603158563,-0.0156809967,-0.0311794616,0.1139838845,0.2691166401,0.381483078,-0.0116239432,0.2220475525,0.3370935321,-0.0468466207,0.0317519717,-0.0229212679,0.3230012059,-0.2710412145,-0.139449358,-0.0685373172,-0.0948330015,0.3044227362,0.5473522544,0.1282820106,0.1361958981,0.2058171481,-0.3219820559,0.0160424374,-0.0090268776,0.5947281122,0.2657072842,0.1015208885,0.2187125534,-0.3983244002,0.0488169231,-0.2300374359,0.03764477,0.3252567947,0.0534546152,0.4030189514,0.1133934855,0.1385077983,-0.0495172776,-0.0279438421,0.1278281361,0.1014389098,-0.0418631062,-0.0687827766,-0.2825423777,0.26626122,0.0600420386,-0.4334226847,-0.1350178272,-0.2884692252,-0.1005189791,-0.0144825391,0.0014506166,-0.0770191774,-0.1364600062,-0.1926987916,-0.0802036226,-0.3923891783,-0.349755615,-0.1259525269,-0.352196008,-0.0385328606,0.1013768688,-0.0939426199,0.2879138887,-0.1168455109,0.065668799,-0.0893291533,-0.2068766207,0.044809252,-0.0092045255,0.3843480647,-0.2553695142,0.4414508939,-0.1614165008,-0.0581801832,0.1734229028,0.0283342786,0.0432912186,0.2223571241,0.108708933,-0.1992360502,-0.0968783945,-0.2927434146,-0.3652944863,-0.3451603055,-0.0720007494,0.1937093735,-0.0490679853,0.0892691463,0.2577719986,0.0275494736,0.1637058556,0.0274684522,-0.0524254963,-0.2859866023,0.8228676915,-0.0953495875,-0.4027893543,0.2715852261,-0.0302079618,0.31057024,0.1123722121,-0.50491786,0.2296973914,-0.2130983472,0.2032733411,0.2825913131,0.055016689,0.0714829266,-0.3109722733,0.0758317932,-0.1627503783,-0.0042988164,-0.0300587881,0.1257300377,0.2862550318,-0.0014703412,0.2728950679,0.0501534902,0.8432106376,0.2358387709,-0.1333760172,0.4208791852,0.0450754985,0.0392548442,0.0224322733,-0.1539186984,-0.1987058669,-0.2561722398,0.1470113248,0.1491042227,0.199015677,-0.2637202144,-0.0537103377,0.017792657,-0.4434629381,-0.3303982615,-0.1697045416,0.1518100947,0.0003313638,0.3040691614,0.3281062841,0.0403618887,-0.2667129636,0.0343166515,0.5420290232,0.0001076604,0.0878890827,0.122037679,0.1206700504,-0.5723161101,0.0829726905,-0.0859023035,0.0734557509,-0.0841358453,0.0762102231,0.1412590146,-0.1602217853,0.146875903,-0.3039337397,0.1390366554,-0.1412777752,-0.1038268507,-0.2265356481,-0.2475598156,0.0366865769,-0.0504823364,0.5440036058,-0.0132459244,-0.328494072,-0.0015201418,0.0642130226,0.181714043,-0.1481835395,-0.2562613487,-0.2763038874,-0.4239401519,-0.3175906241,-0.1952447593,-0.1939310879,0.3753404021,-0.2319893837,-0.2743623853,-0.1211855486,0.270465076,0.1200102866,-0.0199690349,0.3140406311,0.0962479413,0.3911685646,0.0729020759,-0.1346761733,-0.1737754643,0.5136118531,0.3388256133,0.0105848303,0.0428221039,-0.3369645178,-0.0059592687,0.2091008127,-0.2164410502,-0.0649941042,-0.0753395706,0.1949351728,-0.1541579664,0.0030705906,0.2379369587,-0.0963172317,-0.1846462935,-0.4083704054,0.6840828061,-0.2124335021,-0.0203926153,0.1748204231,0.1494867206,0.1398355961,0.2975002229,-0.1081483439,0.7885554433,0.030235542,-0.0180235133,0.0893872753,0.089299053,0.3508450389,-0.3062100708,0.134030208,-0.3522855639,0.1535355896,-0.1315975636,0.0213090368,-0.1281591058,0.0869432166,-0.0777020454,0.1317059845,-0.1633622199,0.2356798053,0.0297499504,0.0323855765,-0.353371352,-0.0888422728,0.0573850274,0.1310209334,0.0415995046,0.6429333091,0.0167473629,-0.2431707978,-0.2635298371,-0.0883843899,-0.5674737096,0.1350712031,0.1171558723,-0.0663839728,-0.2022009194,-0.4554262757,-0.0210023299,0.4356853962,-0.0421972536,0.1682498604,-0.3473185003,0.2733457088,-0.2823804617,-0.0707647279,0.1702465117,-0.1023084298,0.1010854766,-0.1892507225,-0.2828213573,0.0159455668,-0.0988527164,0.2994481325,-0.1585282683,0.1114630997,0.2126403749,-0.0067992718,-0.3708256483,0.1454948485,-0.065000616,-0.1168176085,0.1353500038,0.104976669,0.0566115193,0.2303312421,0.1094347686,-0.2893566191,-0.2170312256,0.4203639627,-0.1740414798,0.1937125921,0.3277284503,-0.0734880045,-0.2277023047,-0.1194424778,-0.106681332,0.0100083491,-0.6670777798,-0.2200978249,0.0030242552,0.2900621593,-0.1375809163,-0.1370420158,0.0780351907,-0.4026843905,-0.3174935579,-0.4946440756,0.2114561498,0.0681232959,-0.0702670589,0.0945587456,-0.4956181347,-0.4544706047,0.3093895912,0.1112437919,-0.1906251162,-0.020559404,-0.1461387873,-0.1312695593,0.1437648088,0.3040228784,0.1504039615,0.0683827028,0.0024418132,0.0444079824,-0.179005459,-0.176528573,-0.1509852856,0.1962991804,0.1860581338,0.022451926,0.2005327493,-0.1181419417,-0.1993985623,-0.0583080202,0.1348382533,-0.0563729778,-0.0442204513,0.0818944275,0.0700664595,0.4368832111,-0.4313287735,0.1096332744,-0.1090185493,0.8408151865,0.0798133239,0.0093668979,0.5487747192,0.1110460535,-0.2143991143,-0.2023205012,-0.1357680857,-0.0793818086,0.3421780467,-0.3432495594,0.5485955477,0.1036884338,0.0768448263,0.3453035355,-0.1219755411,-0.0908229426,0.1542128921,0.2604792714,-0.1336329579,-0.1416879594,0.0448365808,0.0819349587,-0.0425652042,-0.0972898602,0.1431804299,-0.2931314111,-0.2391667962,0.0609626509,0.4813641012,-0.178390339,-0.1318946034,0.4905301034,0.0112407822,0.3369816542,-0.0512915179,-0.0873002261,-0.178627938,0.4202342331,-0.0920411348,0.4068499207,0.1251327246,0.0825373605,-0.4138420522,-0.5052168965,-0.0240684729,0.0552172884,-0.1465782821,0.2593788803,0.1092591882,0.2299335748,0.1215054914,0.0414034799,-0.4975200593,-0.0321864411,0.0245675873,-0.0278496593,0.0452261902,-0.3645442724,-0.1428627968,0.3663728535,0.0384925641,-0.1033891514,0.3109819889,0.1431131214,-0.1473218948,-0.6717371345,0.036505755,-0.1164723188,0.0497756191,-0.3063770533,0.0240193047,0.2131405026,-0.1313631386,-0.0178576652,0.0248099882,0.34553352,0.4297313094,0.0166224428,-0.0233589653,0.0278537665,-0.0651850626,-0.2258093506,0.3390022218,0.2549939752,-0.0104676094,0.1458033025,0.1206900626,-0.2616156042,0.1722255647,0.0933589041,-0.0106580397,-0.2104581147,0.6825149655,-0.2915141582,0.1202342138,-0.1351346225,-0.3433481753,-0.3390801251,-0.0302415155,0.3402217329,0.157163173,0.0806096271,-0.3118818104,0.0384958349,-0.2671253681,0.2604099512,0.0489829518,0.1506985277,-0.3443178236,0.1353344023,-0.7035652995,0.0071806256,0.0843349993,-0.0266272165,-0.0294900648,0.3258947134,-0.2704503536,0.1794225723,0.2376694381,-0.036499545,0.2352370769,0.4280113578,-0.2742737532,-0.0506847799,-0.3023613989,0.0068900874,-0.1276033074,-0.2826354206,0.2721915543,-0.5946931243,-0.0279306062,-0.0327478722,-0.113218829,0.2861553729,-0.2838963866,0.7775373459,0.0201640669,0.3910856247,-0.054441072,-0.0642615929,-0.323481679,0.0563339926,0.0065068034,0.1816594601,-0.1124404445,0.2903109491,0.1143260822,0.1794825196,-0.1371037513,-0.1422662735,0.0706008822,0.2998566926,-0.0024716507,-0.3307428956,-0.0206975155,0.0941232815,0.2345978171,0.266461879,-0.1194427982,0.1812016964,-0.1614152789,-0.263510257,0.3250681758,-0.2873350084,-0.2141287327,-0.1121366918,0.1395836622,0.2979430556,0.0193235744,-0.4697853029,0.0013581765,0.2750638127,-0.134632811,-0.2287648171,0.1092010736,0.1285208464,-0.1194906533,-0.0283070076,0.1960527301,0.0042830952,-0.292571485,0.2687163055,-0.141082406]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2120","title":"dataset viewer does not work anymore ","comments":"Thanks for reporting :) We're looking into it","body":"Hi\r\nI normally use this link to see all datasets and how I can load them \r\n\r\n\r\nhttps:\/\/huggingface.co\/datasets\/viewer\/\r\n\r\nNow I am getting \r\n\r\n502 Bad Gateway\r\nnginx\/1.18.0 (Ubuntu)\r\n\r\ncould you bring this webpage back ? this was very helpful @lhoestq \r\nthanks for your help ","comment_length":8,"text":"dataset viewer does not work anymore \nHi\r\nI normally use this link to see all datasets and how I can load them \r\n\r\n\r\nhttps:\/\/huggingface.co\/datasets\/viewer\/\r\n\r\nNow I am getting \r\n\r\n502 Bad Gateway\r\nnginx\/1.18.0 (Ubuntu)\r\n\r\ncould you bring this webpage back ? this was very helpful @lhoestq \r\nthanks for your help \nThanks for reporting :) We're looking into it","embeddings":[-0.5371803641,-0.0297721326,-0.0728389397,0.1986725032,0.2598983347,0.0393822938,0.1914966255,0.3085088432,-0.0682066232,-0.0875277519,-0.3016587794,-0.0916364715,0.1556514204,0.402369976,0.3185894191,0.0368490219,-0.0181391127,0.171715036,-0.2725091577,0.1096207574,-0.0308225248,-0.0666808337,-0.2506363094,0.1245341301,-0.0603329204,0.2551175654,0.2462005764,0.2662482858,-0.3655655384,-0.5013542175,0.3300395906,0.0879855454,0.2842431664,0.2857335508,-0.0001097984,0.288318336,0.5774621367,0.1273643672,-0.350554347,-0.0845102519,-0.6345464587,-0.2996250093,0.1194663495,0.0086814975,-0.1788457632,-0.4047734141,0.1844080687,0.0431774333,0.2575726509,0.2531820536,0.2151810229,0.0393861644,0.3328340054,-0.0879727751,0.0324041545,0.22337915,0.0090134442,0.2472469509,0.2922057807,0.2171685845,0.1001717746,0.2087426484,-0.1890398562,-0.0003293949,0.1614656001,-0.3043292165,0.2660669088,-0.2376513779,0.05055473,0.2104593515,0.5765340328,0.1853511035,-0.0902973935,0.0596645214,-0.2016954571,-0.1178557798,-0.072537683,0.0350517295,0.0100909099,0.3407236934,-0.2537845969,-0.3512859941,-0.2485876083,0.0872057974,-0.1973415017,0.6988922358,-0.2604932785,0.1862280667,0.1387925297,-0.0953028426,0.4454909265,0.3394754231,-0.2117493302,-0.1383013874,-0.2867534757,0.115729332,0.3426729143,0.4386004508,0.1897568703,0.0565265492,0.0044085211,0.2989141941,0.0027106996,0.086719729,0.0730968639,-0.0425856784,0.1388969719,0.0686085001,0.3116225004,0.4203718901,0.0482964702,0.072667025,-0.0297554117,-0.2579121292,-0.3088901639,-0.0970901623,0.2720565796,-0.2835687995,-0.1796881258,-0.1214947924,0.1144940555,-0.1585140675,0.3239794374,0.422637254,-0.0682100803,0.4194627404,-0.1744018942,-0.2732036114,-0.0680183917,-0.665063262,-0.1434569955,-0.1083623096,-0.0649808124,-0.1389934123,0.0321412496,-0.2075840384,0.086933963,-0.1814554781,0.019147845,-0.1019224301,-0.0292071421,0.0600414611,-0.0269971415,0.1863572448,0.2140502334,0.0628814623,0.1975435615,-0.3905587494,0.1285837144,0.0860914886,-0.450260222,-0.3290115595,-0.1611091346,0.1836554259,0.0115953404,-0.1102297977,-0.3780627251,0.0912065879,-0.4803446233,-0.4448062181,-0.1899103671,0.1197197363,-0.1756612509,-0.0281511545,0.4782161713,0.4248363376,-0.3942656219,-0.1242242903,-0.0377292186,-0.2338651717,-0.1783784181,0.076908119,-0.0980990604,-0.0343175866,-0.3387134671,-0.36821118,0.6349301934,-0.3029394448,-0.3459963799,0.4475865364,-0.0134959221,-0.2986554503,-0.0144405989,-0.0614941455,-0.0132468054,0.0224011913,-0.2783986926,-0.087805301,0.1748547256,-0.080064714,-0.0819028094,-0.229403764,0.1842787564,0.103585273,0.291343987,0.2288701981,0.1451548189,-0.5407420397,0.5333086252,-0.1217244789,0.1465622932,0.4359654486,0.3594190478,0.0408800691,0.3035683036,-0.0547572635,-0.2140238881,0.1317979842,-0.0183400735,-0.3834649026,0.1193266213,0.0850336179,-0.5722771883,-0.0305839293,-0.2762067318,0.093722567,0.1163076609,-0.0368196853,-0.2126309276,0.4247258604,-0.2251100987,0.0636054501,-0.1433300227,0.0678668469,-0.277695328,0.2736561894,-0.0636721179,0.1579114348,-0.0140213948,-0.2310860306,0.316129297,-0.0811954513,-0.0537992492,0.1357168555,-0.2439499348,0.3228222132,0.3419056535,0.0935017914,0.2940759361,-0.6011407375,0.0533302315,-0.143285796,-0.0911829695,0.0444468707,-0.0159946382,-0.2612638772,-0.2135486752,0.1949196607,0.0716624632,0.2458664328,0.27815938,-0.0648373589,-0.0203499217,0.0302904099,0.2173636854,0.0164867323,0.2563385367,-0.1528248936,-0.0216552652,0.155025661,-0.2827336788,-0.0716444924,0.1502362192,0.1649716198,-0.1966702938,-0.0342246145,0.3048832119,0.2254653126,-0.034153264,0.1497398019,0.0151788639,0.1234702095,0.2955614626,-0.1048231274,0.2386410534,-0.047325857,0.1876601279,0.035483446,-0.0841298625,-0.1541083753,-0.3472172618,-0.1043880135,-0.0534680486,0.2999843359,-0.3783727884,-0.0130905956,-0.2126906216,-0.5997260809,0.2510102391,-0.0342375524,0.1160242409,-0.4540726244,-0.3559892178,0.0722093657,-0.0139519945,0.2094420791,-0.4400201142,0.2908226252,0.0673400909,0.3312675953,-0.4536323845,-0.1213871986,-0.1584454626,0.1090451032,0.1918845922,0.1617293358,-0.0785604864,-0.2957484722,0.1538092941,-0.5220479369,-0.1974847615,0.2658593357,0.1785615236,0.3475287259,0.0355514064,0.27194646,-0.2698592246,0.1098022759,0.3279714584,-0.0094070956,0.0722670853,0.09524066,-0.1986196339,0.1251165271,0.2612911165,0.0145214237,-0.0411256067,-0.3655759096,-0.0859593749,-0.0910331458,0.1455117762,-0.084208414,0.2535293996,0.0660724342,-0.0166572072,0.1360764503,-0.0648340359,-0.6612625122,0.2860377133,-0.1889007539,-0.2088367939,0.1538697779,0.2131657153,0.2989663184,-0.0694250241,-0.5617764592,-0.1590670347,0.1054020748,-0.209491536,0.1118272245,-0.089198567,0.0170513187,0.0205678921,-0.0782558098,0.0664595515,0.0355182961,-0.5173721313,-0.266996026,0.2544161379,0.0633749291,0.4205610156,0.0540152974,0.3072002232,0.3182715476,0.2305075526,0.3247582018,-0.0194427073,0.4947715104,0.0890565142,-0.5347933769,-0.009091516,-0.0071116798,-0.0048353742,-0.1211487874,0.2403224111,-0.0603567623,-0.4305557013,-0.3066557646,-0.3957547545,-0.1200801432,-0.2370252311,-0.1862248778,0.1300745755,0.091880627,0.1461644024,0.0186294448,-0.199134931,0.2126293331,0.3498780131,0.3153153062,-0.1139074117,-0.137532264,0.1812595576,-0.2553994358,0.397726655,-0.1783327758,0.1927141398,-0.1551579237,0.1749435514,0.0007033265,-0.0545513779,0.5260671973,-0.4244332314,0.1304649711,0.0824983492,0.1417017579,-0.2917228937,0.0866705254,0.2153397799,-0.0422481559,0.3053120077,0.0217568036,-0.0074001723,-0.1317292303,0.383810848,-0.1386981755,-0.210388422,-0.0249724872,-0.1402076632,-0.0536676459,-0.048663374,-0.2288091332,-0.3434969485,-0.0365085565,0.0111849112,-0.2014020979,-0.1037043482,-0.0607830808,-0.1051285416,0.122832492,0.1723204702,0.12892434,-0.10141325,0.5118534565,0.2958032489,0.001799922,0.4127191007,-0.0543511249,-0.5412913561,0.1749861389,0.3381947577,0.1543670148,0.0977413952,-0.0398444422,-0.1198185459,0.5730608702,0.1923851818,-0.2421018928,0.1376119405,0.3695064485,0.0237063225,-0.5775990486,-0.3141235709,0.4658518136,-0.0534635,-0.0282130297,0.3942063153,0.368200928,-0.0419278592,-0.2916612923,-0.0791251212,0.7227166891,0.1290140748,-0.1527914107,0.2061553597,-0.0801090151,0.4434417188,-0.1208684966,0.0076982821,-0.2017308474,-0.1906886697,-0.1328897178,-0.1511891782,0.1685111374,-0.100643456,0.0811519325,0.3741003275,0.0704253092,0.2621543407,0.1254923791,0.1889011115,-0.0958626866,0.220422104,0.0938625038,0.1612545103,-0.3703838885,-0.0140276765,-0.281121105,0.0389113501,0.1882866323,0.0756402463,-0.2842811346,0.1522122771,-0.063222371,0.4584264159,-0.0994345248,-0.6327530742,0.0975265354,0.2238085419,-0.0991727486,0.0921601653,-0.4319713414,0.369853735,-0.3723109066,0.1027697325,-0.1147896796,0.0373387747,0.06103554,-0.000124381,-0.411711663,0.176153183,0.0641904622,-0.1056999341,-0.1832902879,0.1784666479,-0.0861831158,-0.2038882673,0.2073741704,-0.0139423022,0.1836214215,-0.2171694934,0.1013998762,0.3209508061,0.0350175723,0.0530576147,0.3398805261,-0.091626063,-0.0068177399,0.2065505981,-0.1872736961,-0.2409235388,0.2507106364,0.2230480164,-0.1504989415,-0.148465097,0.3194094598,-0.0305039324,-0.0113406219,0.0390793011,-0.1371903867,0.3386596441,0.0797371566,-0.2522369921,0.0096757272,-0.0520112179,-0.0555201061,-0.360288769,-0.2011699975,0.2453867644,-0.1476845443,0.0304969586,-0.0353874713,-0.3123947382,0.1837105155,-0.1982479692,-0.2747796476,0.2617211342,-0.0562283881,-0.2920804918,0.0129236197,0.1983729601,-0.1507362872,-0.1822561175,0.0237664375,0.0693515316,-0.260288775,-0.1276701987,-0.1179589182,0.1184714362,0.157205686,-0.28561607,-0.0490494333,-0.0766614079,-0.109049052,0.0178884938,-0.2155624181,0.0202686656,0.1973877996,-0.1783215255,-0.1568998545,-0.2919310927,-0.2108020633,0.4112970233,0.2089835256,0.2915489972,0.2113429457,0.265358597,0.2811278403,-0.0278140195,-0.1806315035,0.101363875,0.31405738,-0.289372623,0.5791755915,-0.3561896384,0.1378449947,0.0450431928,0.3549205363,0.2364017963,-0.0926064923,0.1393555999,0.1620715261,0.2261953503,-0.2444528192,-0.1212830469,0.2987006009,0.2282928228,0.0241919849,-0.1916666925,-0.0348258875,-0.0775591731,0.1145265028,0.2571989894,0.2876495719,0.0240985509,0.2200945914,-0.1824380159,-0.0610973835,-0.1597083509,0.0142659862,0.220842585,0.0500420146,0.5202046037,-0.0704704821,0.0740484968,-0.3065201044,0.4604676664,-0.0479591414,-0.4286873341,0.0918311924,0.0348403789,-0.7276844978,-0.0595595017,0.0234814566,0.3928593695,0.0515703484,0.0108179413,-0.2208641917,0.2493417561,-0.1081026047,0.0972502455,-0.4487881064,-0.3262657821,-0.035761185,0.087979719,0.1106117964,-0.179283008,-0.2892310321,0.0964989141,-0.2316547632,-0.3431424797,0.2617482841,0.2544332743,0.2509241104,0.0070072319,0.0013026615,0.382914573,-0.1009935141,-0.036754109,0.5520692468,0.2996902466,-0.0304859504,0.252602011,0.0617156662,0.1308054924,-0.289172262,-0.2684225142,0.0296769943,0.2813771963,0.0646959767,0.1414396614,0.2240991294,-0.1426705271,0.167560488,-0.2069036365,0.2795941234,-0.3255272806,0.0118233375,-0.2349156141,-0.213378787,-0.2536260784,-0.3061933815,-0.4330359697,0.3427378833,0.4435842633,-0.0125741102,0.054605905,-0.2923538983,0.039624352,0.1732530743,0.4074923396,0.3955207765,0.2856706083,-0.1391874701,-0.4194353223,-0.5789977312,0.0730881244,0.0875860602,0.3396858573,0.1375006288,-0.0197251067,-0.137267381,0.0808049738,0.2733627558,0.4210898578,-0.0883618742,0.0908989683,-0.2068850398,-0.0192117319,0.0318716951,0.3715865016,-0.2734560072,-0.1593583226,0.5487717986,-0.009540217,-0.0004240621,0.0313412957,0.0805468112,-0.1852475107,-0.4751973152,0.4065959752,0.1562699527,0.3336399794,0.0249563903,0.0575706922,-0.3602832854,-0.5273328424,0.0340943895,0.2549149096,0.2723932862,0.0075669503,-0.0568641163,-0.3910847902,-0.1699366421,0.2167674899,-0.1117664576,0.2036351115,-0.3678059876,-0.2468367368,-0.0445196331,-0.0084835878,0.259850204,-0.1863291562,-0.0030816081,-0.0398823693,-0.1533515602,-0.1508235186,0.2249942422,-0.1204989702,-0.0683084652,-0.3876357973,0.1970673054,0.3536321223,-0.0026219427,-0.1071718633,-0.03300092,0.3823224008,0.0519198477,0.2368692458,0.2957038581,0.0422728471,0.1746403575,-0.1021849737,0.1630190909,0.2644794583,-0.3031124175,0.0478721,-0.0084302621]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2120","title":"dataset viewer does not work anymore ","comments":"Back up. ","body":"Hi\r\nI normally use this link to see all datasets and how I can load them \r\n\r\n\r\nhttps:\/\/huggingface.co\/datasets\/viewer\/\r\n\r\nNow I am getting \r\n\r\n502 Bad Gateway\r\nnginx\/1.18.0 (Ubuntu)\r\n\r\ncould you bring this webpage back ? this was very helpful @lhoestq \r\nthanks for your help ","comment_length":2,"text":"dataset viewer does not work anymore \nHi\r\nI normally use this link to see all datasets and how I can load them \r\n\r\n\r\nhttps:\/\/huggingface.co\/datasets\/viewer\/\r\n\r\nNow I am getting \r\n\r\n502 Bad Gateway\r\nnginx\/1.18.0 (Ubuntu)\r\n\r\ncould you bring this webpage back ? this was very helpful @lhoestq \r\nthanks for your help \nBack up. ","embeddings":[-0.5085449219,-0.0073961332,-0.0676931441,0.2109005302,0.2437383384,0.0490321629,0.2018448412,0.2702572346,-0.0885559171,-0.1051464751,-0.2714341879,-0.0763678029,0.1603515446,0.3936680853,0.3260014951,0.0114623597,-0.0254044142,0.181356132,-0.278942436,0.1035796851,-0.010450528,-0.0695015788,-0.2603881061,0.1267054826,-0.0685188398,0.2393033952,0.2504808903,0.2946494818,-0.3781282902,-0.5050235987,0.2979981899,0.1038198844,0.2971769869,0.3132668734,-0.0001107919,0.2629814148,0.5615951419,0.1258792579,-0.3019791245,-0.0913674459,-0.6642631292,-0.2925684452,0.0914593339,-0.0215152707,-0.1450494081,-0.4067386389,0.1394538134,0.0492660031,0.3380746841,0.2728051841,0.2059211582,-0.0330458917,0.3224157393,-0.0674408451,0.0194559768,0.2289574593,0.0419500731,0.256734848,0.2840365469,0.2128195018,0.1159746498,0.1914329082,-0.1803275347,-0.0254251938,0.176390022,-0.3176693022,0.284724772,-0.2088562846,0.0935063362,0.2374196351,0.5861465335,0.2119385004,-0.042426683,0.0484921411,-0.2217937708,-0.0801047236,-0.0530902669,-0.0019340506,0.0318554007,0.3146966398,-0.2367702723,-0.3898980916,-0.2873103023,0.0638970882,-0.2188806087,0.698646605,-0.2473623902,0.1903271675,0.1309764981,-0.1077237204,0.4718031883,0.3609367013,-0.2567167282,-0.1295845956,-0.2934585512,0.1046799123,0.3132137954,0.4475222528,0.1939572394,0.0394339822,0.0449842252,0.2873446345,-0.0075683589,0.1019168496,0.075642243,-0.0029729332,0.1417365223,0.0801052228,0.3226472139,0.3485426605,0.0167152621,0.06863904,-0.0760688707,-0.2732737064,-0.2565947771,-0.0740910098,0.2500242293,-0.3264962733,-0.1990024745,-0.0994307399,0.1199456006,-0.1516885161,0.325622648,0.4204390943,-0.0722870007,0.4155551195,-0.1606567502,-0.3072172403,-0.0797603205,-0.6794695258,-0.1088964194,-0.0959412232,-0.0601598844,-0.1673417091,0.0080865286,-0.1749004573,0.0777703896,-0.1936181337,-0.0008209692,-0.0977916196,-0.0206347331,0.0249748174,-0.0209961012,0.1453616619,0.2042218894,0.0649984628,0.1674001366,-0.3857432306,0.1666058451,0.1209371462,-0.4803654552,-0.3302630782,-0.2140796483,0.1760264337,0.0045429245,-0.1350966841,-0.4012063742,0.0925940797,-0.4359095991,-0.4717687368,-0.1976877749,0.1022651643,-0.1910888553,-0.0529855266,0.4783594012,0.4020072818,-0.4532059133,-0.1169339493,-0.062823981,-0.2432918102,-0.1900791377,0.087484315,-0.0890510604,0.0457358807,-0.3335569203,-0.3827990592,0.6559303403,-0.2786400616,-0.3348894715,0.4397470951,-0.0177241452,-0.3297103047,-0.0579141602,-0.0514229052,0.0429880694,0.0196457263,-0.2481887937,-0.0942125097,0.1800561547,-0.1199117526,-0.09109018,-0.1948452592,0.1917119175,0.0603362471,0.306042701,0.2280121297,0.1538669467,-0.5391329527,0.4628996253,-0.0814113766,0.1510385424,0.4669396877,0.3995165825,0.0730408132,0.3427202702,-0.0474168248,-0.1554070562,0.1538294852,-0.0370574892,-0.4114989042,0.1105433181,0.1117239892,-0.5754289627,-0.0509327389,-0.2907711267,0.1249490082,0.1110498309,-0.0716229305,-0.2102564722,0.3680945933,-0.2254135907,0.0932323486,-0.1613443345,0.0399569459,-0.2661912739,0.3139567673,-0.0855828077,0.1492522359,-0.0237151273,-0.2371347398,0.3440355957,-0.0618862472,-0.0843326598,0.1536331475,-0.2664220929,0.3242759109,0.3025161326,0.0947599486,0.3085180521,-0.5888142586,0.0612864047,-0.1148831621,-0.1081505418,0.0461286791,-0.0162983127,-0.2827736139,-0.2344044894,0.2036709636,0.0334523022,0.2446541637,0.2486764044,-0.0526927449,0.0055560563,0.0066900738,0.2171830982,0.0311944913,0.226534605,-0.1542291939,0.0200853068,0.1505678892,-0.3068956137,-0.0861444697,0.1711260229,0.1705754846,-0.1421692073,-0.0128482012,0.3065418601,0.2875142694,-0.0180534944,0.134639293,0.0131526273,0.132610783,0.2545979023,-0.0674217716,0.1928019226,-0.0309961122,0.2330683917,0.005393954,-0.0781933516,-0.1499029249,-0.3082774878,-0.1150643602,-0.025357103,0.3372808397,-0.4205353558,-0.0365109704,-0.2339254916,-0.6573860645,0.2263655514,0.0000857333,0.1363515705,-0.4362689555,-0.3714274168,0.0436742343,-0.0032501253,0.2053747922,-0.4189718664,0.317117095,0.069824703,0.2919323742,-0.3803749979,-0.1474769711,-0.1807755232,0.1025313661,0.1913426071,0.1867540628,-0.0708519295,-0.2971154153,0.1391549855,-0.4654504359,-0.2241782546,0.2686719894,0.1914421022,0.3556558788,0.023363851,0.3369657993,-0.3097658753,0.1502643675,0.3208380938,-0.0161162708,0.0542951375,0.0550994202,-0.1441308111,0.1128176749,0.2799940109,0.0373605601,-0.0100460816,-0.3429900408,-0.1005815491,-0.0789709017,0.1723052561,-0.1147328615,0.2574815154,0.0695543736,0.024171602,0.1596718431,-0.0374261625,-0.6710076928,0.2852566242,-0.1719897985,-0.2101013213,0.1851308942,0.240469113,0.2766743898,-0.0648500323,-0.5317142606,-0.1538160741,0.0962447599,-0.2258426845,0.0849684402,-0.1111419648,-0.0249933787,0.0326887406,-0.0581403263,0.0438941382,0.0439189449,-0.4580725431,-0.2036234587,0.2677279115,0.1025789157,0.4269919395,0.0636222959,0.2706581652,0.3248777688,0.1891757399,0.3047759831,-0.0014729494,0.4591695368,0.1315025985,-0.5012581348,-0.0150105758,0.0388430096,0.0033232977,-0.1505317241,0.2661364377,-0.0527854413,-0.4490674436,-0.3505753279,-0.3946218789,-0.1148946285,-0.2385521382,-0.192065835,0.1414597183,0.0621751733,0.1496611536,0.039028503,-0.1762787849,0.2305852771,0.3284770548,0.3132909834,-0.1008367836,-0.1240432188,0.1503925771,-0.2610477507,0.4058541954,-0.2062793076,0.202556178,-0.1879079789,0.2346327156,-0.0345329009,-0.0457841493,0.4925477505,-0.4593842924,0.1278018802,0.0492223874,0.1689088494,-0.250171721,0.0777522698,0.2447308898,-0.0139972065,0.2997157872,-0.0275518596,-0.0265477914,-0.1092295498,0.3534155786,-0.1421008557,-0.2253031731,-0.0563550442,-0.1080905795,-0.0558536202,-0.0330351889,-0.2412361205,-0.3360108733,-0.0517455041,0.010343872,-0.2175793499,-0.0637041032,-0.0516592376,-0.0895157084,0.1159845442,0.1706375182,0.0952265859,-0.0983131677,0.4971832037,0.2864950597,0.0223758705,0.382829845,-0.0500885658,-0.5191224217,0.1407793462,0.3389181793,0.1521614939,0.0678501129,-0.0559885018,-0.0745639876,0.5903968215,0.1805187315,-0.2255393118,0.1023264974,0.3547874093,0.0117217768,-0.5657277703,-0.3669125438,0.4618043303,-0.0175467748,-0.0299083441,0.4007410407,0.3482125401,-0.0546222515,-0.2766669989,-0.0151196895,0.8093259335,0.1407576799,-0.1999523491,0.1761953086,-0.1293547153,0.4041396976,-0.0955235958,-0.0238709096,-0.1998520046,-0.1489596814,-0.1380436271,-0.1055747792,0.1866434366,-0.1020661145,0.0559709035,0.3987259269,0.0718134865,0.2531702518,0.1440508664,0.1712327451,-0.074150376,0.2191538066,0.1084667221,0.1482237726,-0.3621441424,-0.0723986775,-0.3146643639,0.0295825936,0.2083197683,0.0927281231,-0.3140039742,0.118461147,-0.0986702368,0.4786609411,-0.1315746903,-0.6352341771,0.1509278268,0.2353173345,-0.1178581491,0.1025373638,-0.4343202114,0.3714552224,-0.3402327895,0.0701130405,-0.0975224897,0.0240083095,0.0364434607,0.0120465923,-0.4536833465,0.1764845997,0.060881272,-0.0750779584,-0.2063764781,0.1724122614,-0.0967845172,-0.1876093596,0.2163037658,-0.0290546194,0.1467938423,-0.2044246644,0.0990969315,0.3566134572,0.036674194,0.0839632154,0.325833112,-0.0836951137,-0.0211427882,0.2044097781,-0.1792538315,-0.234961465,0.2570527494,0.1993082613,-0.1195710674,-0.1340919137,0.345533967,-0.0526288263,0.0348820537,0.0208033789,-0.1458612978,0.3704907894,0.090336293,-0.2916401923,-0.0323202163,-0.0043420605,-0.0467019081,-0.3880341649,-0.1840445995,0.2408300191,-0.1489002854,0.0255631637,-0.0529739223,-0.328985244,0.1684272438,-0.2033144981,-0.2753852904,0.2601309121,-0.0623762943,-0.2590271235,-0.0110364733,0.2037004977,-0.1931335628,-0.2212703079,0.0127000222,0.0762258917,-0.2253124416,-0.1178434268,-0.0994508341,0.1202152893,0.1616534889,-0.2731470168,-0.0742721632,-0.1436118186,-0.1250264645,0.0228807274,-0.2247329056,0.0424585119,0.1819294244,-0.1694516242,-0.2052190602,-0.2965252399,-0.2707059681,0.4143109322,0.233263582,0.2770663798,0.2348657399,0.2620335221,0.270478338,-0.0096622473,-0.1934172958,0.1035515815,0.3282463849,-0.328204006,0.5958718657,-0.3876587749,0.1507351995,0.0284383297,0.367598325,0.2415519059,-0.1082636043,0.1646944284,0.1799653322,0.2116529047,-0.2251021564,-0.1042141542,0.2887105346,0.1875708699,0.0354105569,-0.1322536469,-0.0484416597,-0.0788711384,0.0771782324,0.2754490674,0.2888807952,-0.0091175688,0.208009541,-0.1701962799,-0.0433608033,-0.1878886819,0.0130118914,0.1991245151,0.0368080065,0.5282257199,-0.041353099,0.0974327996,-0.2726055682,0.483078897,-0.0436509922,-0.4423112273,0.1003543884,0.1132709906,-0.7063918114,-0.0616162531,0.0290119797,0.361948669,0.060551174,-0.0063115209,-0.2031123638,0.2417710274,-0.1203606278,0.1161529273,-0.489192307,-0.3285319209,-0.0789578557,0.060747765,0.1234826595,-0.179185912,-0.3389648497,0.1251661777,-0.2561089098,-0.3517917991,0.2650772631,0.2902053297,0.2441374063,0.0155555001,0.0211689174,0.3999296427,-0.1219992563,-0.0244115777,0.541133821,0.3130227029,0.0128323892,0.2393258363,0.0572545156,0.1540191919,-0.30190593,-0.2176032513,0.02739981,0.2427194417,0.0582166538,0.1476767957,0.2129554898,-0.1406387687,0.1674509495,-0.1699668318,0.3093257546,-0.2914375663,-0.0387530848,-0.1929852664,-0.1926041842,-0.2314767241,-0.3288179934,-0.4519776404,0.305118978,0.4430241883,0.0273374226,0.0349418335,-0.279311955,0.0336752124,0.1418659538,0.4406072795,0.4309380054,0.2944662869,-0.1515694261,-0.4553224146,-0.5925188065,0.003168674,0.0494913012,0.3299425244,0.1613584161,-0.0506105199,-0.1370432377,0.0755655989,0.3060587943,0.4464045465,-0.1017222852,0.1091496348,-0.218627274,-0.0320274569,0.0698067322,0.3731431365,-0.2694426477,-0.1070473865,0.5287269354,0.0066721938,-0.0039547221,0.0814617872,0.0567705967,-0.209205389,-0.4771966934,0.4115969539,0.1509086192,0.3465567529,0.0517255478,0.0892291218,-0.3264387846,-0.4965635836,0.0283703655,0.2604288459,0.3200004697,0.0025274756,-0.0755257085,-0.3688131571,-0.1684904993,0.2073882371,-0.1273547113,0.1767518967,-0.3771610856,-0.234579891,-0.0448531359,0.0162519328,0.2406579852,-0.209849596,0.0044511463,-0.0312014632,-0.1693240106,-0.1367119253,0.2327942401,-0.11520046,-0.024203429,-0.4318374097,0.1926410794,0.3439033926,0.0240313672,-0.1401565075,-0.0525966436,0.3696312606,0.046124801,0.2257338315,0.2696352303,0.0301649552,0.2304569036,-0.0960371494,0.1331433207,0.2440693378,-0.2954300046,0.0708270893,0.0266608167]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2117","title":"load_metric from local \"glue.py\" meet error 'NoneType' object is not callable","comments":"@Frankie123421 what was the resolution to this?","body":"actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(path='\/home\/glue.py', name=actual_task)\r\nmetric = load_metric(path='\/home\/glue.py', name=actual_task)\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 1 actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\n 2 dataset = load_dataset(path='\/home\/jcli\/glue.py', name=actual_task)\r\n----> 3 metric = load_metric(path='\/home\/jcli\/glue.py', name=actual_task)\r\n\r\n~\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs)\r\n 508 keep_in_memory=keep_in_memory,\r\n 509 experiment_id=experiment_id,\r\n--> 510 **metric_init_kwargs,\r\n 511 )\r\n 512 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\r\nPlease help","comment_length":7,"text":"load_metric from local \"glue.py\" meet error 'NoneType' object is not callable\nactual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(path='\/home\/glue.py', name=actual_task)\r\nmetric = load_metric(path='\/home\/glue.py', name=actual_task)\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 1 actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\n 2 dataset = load_dataset(path='\/home\/jcli\/glue.py', name=actual_task)\r\n----> 3 metric = load_metric(path='\/home\/jcli\/glue.py', name=actual_task)\r\n\r\n~\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs)\r\n 508 keep_in_memory=keep_in_memory,\r\n 509 experiment_id=experiment_id,\r\n--> 510 **metric_init_kwargs,\r\n 511 )\r\n 512 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\r\nPlease help\n@Frankie123421 what was the resolution to this?","embeddings":[-0.1416041702,-0.3257978261,0.0615410134,0.3568933606,0.423979491,-0.1794937849,0.4668626785,0.3729821742,0.5315228701,0.2410525978,-0.2204619795,0.226946637,-0.1156021506,0.1400528848,0.1851810068,-0.2632711828,-0.2527934611,0.0901691467,-0.3066795468,0.0083192801,-0.5935446024,0.112843506,-0.176676631,0.1445945203,-0.1705612391,0.0393202752,0.0093168141,0.1812598258,-0.1959913522,-0.4532200098,0.3768622875,-0.1716807634,0.1465152502,0.5241673589,-0.0001139738,-0.033531595,0.4504942,-0.0290766377,-0.0863232464,-0.472538054,-0.1195410416,-0.2357003391,0.3752601743,-0.4478064477,-0.062121693,0.1416456699,-0.0828403085,-0.347415328,0.2653704882,0.4695699811,0.2061395049,0.6106847525,0.0854604766,-0.2028411031,0.0254346039,0.1798776388,0.0358551145,0.822942555,0.0657702833,-0.2006156296,0.3258185983,0.1416735798,-0.4325895011,0.2143806666,0.3801801503,-0.0234442502,0.3884054124,-0.3318966329,-0.0296703465,0.1398440003,0.0953266025,-0.2009051591,-0.1420157999,0.0897941291,0.0573740825,-0.2639654875,0.1368520111,-0.0794139281,-0.1200280041,-0.0964930877,-0.1322456598,0.0848752931,-0.3178741038,0.2693817616,-0.1124229208,0.067176953,-0.1727846414,0.2685065866,0.2361777425,-0.0117009161,-0.1839256287,-0.065087527,0.0582415089,0.253734231,-0.2643872797,-0.0305878632,0.0793225914,-0.0870629549,0.0679135174,-0.1755403429,0.0548666008,0.0314429924,0.404302448,0.2978640497,0.2420685887,0.3966830373,0.3951809108,0.1466270536,0.2279794663,0.0780079365,-0.1701204777,0.0892674774,-0.2642494738,-0.1551116705,0.265155524,0.3011448085,0.7418727279,-0.1164467558,-0.4467943609,-0.1033933088,-0.2430132478,-0.0160270091,0.1944336295,0.161204204,-0.0987311751,0.2632951438,0.2813692391,0.2256501019,-0.3031885624,0.0423738956,-0.1698928028,0.209261179,-0.2760275602,-0.0130499769,0.1990453154,0.0732425004,0.0095464503,-0.0431390516,0.0698577091,-0.2045388371,0.3016422391,-0.3301384151,-0.2868652046,0.2062863261,-0.0394195579,-0.019430574,0.3747611344,-0.1785913706,-0.2175423354,0.0917097405,-0.3233513534,-0.3507694006,-0.138763383,0.1727755964,-0.3046310842,0.1278126538,-0.0877464563,-0.2583293617,0.1053189933,0.0035506941,-0.0532424785,-0.2327440083,-0.170050025,-0.3028181195,0.1276641786,0.5399086475,-0.0531655475,-0.2346816957,-0.2798365057,-0.085427165,0.2691752911,-0.1091260985,0.0090122931,0.3593357205,-0.0901094377,-0.086501576,0.7411464453,-0.7970083356,-0.1006150916,0.3202630281,-0.2047171146,-0.0835515037,0.0395816639,-0.0504144803,0.2391885519,0.1126610935,0.7353158593,0.1236042455,0.1152543575,-0.1581477076,-0.3017596602,-0.2029255778,0.0642598644,0.1311470121,0.2572923005,0.1546133012,0.0779035315,-0.2244312763,-0.0351635702,0.1184951514,-0.0987057686,-0.0340099744,0.4869561493,-0.1010515168,-0.0138705876,-0.4992186725,-0.3632444143,0.2329906523,-0.3824585378,0.3342255056,-0.0692585632,0.029322505,-0.5362445712,0.0341832675,-0.0985036269,0.1200829297,0.1544422954,0.0870235041,-0.158263579,0.071469374,-0.1991973817,0.2324831486,-0.111060746,0.0774062052,-0.0466393381,0.1501399279,0.0305832792,-0.3826861382,-0.1226081848,0.2542795837,0.4912652075,-0.0956642777,-0.1125646681,0.3919705749,0.0828628466,-0.1706883013,-0.151740849,-0.0817401037,0.1066601574,-0.1553117186,0.1468529999,-0.1352465898,0.1840054542,-0.1702711284,0.2297586054,0.4071613848,0.0510686263,0.4183044732,-0.1768908054,0.180316627,-0.1633235961,-0.1156742349,-0.2016358078,-0.1893025339,0.2789881825,0.172680378,0.3203671873,0.0309991725,0.0763988048,-0.1049282476,0.6013664007,-0.055936709,0.2778433263,-0.1277685165,-0.2225793153,0.1136692837,-0.0877265409,0.0646688268,0.6149451137,0.1072999984,-0.0771989375,0.1156614125,-0.0927344486,0.0619658977,0.0098871542,-0.0553911962,0.0446745791,0.1551363021,0.0731625333,-0.0187440608,-0.1677016616,0.15166004,-0.0597657822,0.246403262,-0.4349276125,0.149449572,-0.0814669356,0.0082474435,-0.1552704424,-0.1223816425,-0.1316928416,-0.3157084286,-0.3301060498,0.1838598549,0.2172186077,0.130218491,0.1835803241,0.1280438155,0.0393526182,-0.2308589369,-0.0097619919,-0.2150152326,-0.1266860068,-0.0600574724,0.2621898949,0.158599332,0.3661262393,-0.3676210344,0.0607126765,0.0885161757,-0.242254138,-0.0468789451,0.0363489687,0.5206822157,0.3190010786,0.1604586244,0.083889097,-0.012051519,0.4989899993,-0.2767891586,-0.0331282392,0.357834965,-0.3447873592,0.107301265,-0.2366063744,-0.2608927786,-0.484452039,-0.5023006797,0.1545526832,0.0418613739,0.1775549799,0.1650446951,0.0954410359,0.5642625093,-0.0315710641,0.0131561598,-0.1641694456,-0.240956381,0.1250115186,-0.1228733212,-0.2683581412,0.0953330323,-0.0181740578,0.4425962865,0.1290745139,-0.100712508,-0.5479288697,0.0117202336,0.195688054,-0.14777264,-0.094888486,0.2685936391,0.0042513572,-0.04207398,-0.1827785671,-0.084740229,0.2569550574,0.3965142667,-0.070366703,0.0625078753,0.3460729122,-0.2300782502,0.4245954454,-0.0505976938,-0.1695774496,0.3556144834,-0.0911772475,0.3200713694,-0.1269998997,-0.5163712502,-0.0300166141,0.0843687132,0.1703580767,-0.1470379382,-0.1182271317,-0.1594669372,-0.465421021,-0.2434878647,-0.2569479942,-0.2508065999,0.1372584701,-0.0782527328,0.1588554084,-0.10273844,0.0329230838,-0.5103862286,0.1112362444,-0.0361608416,0.2277417183,0.0744254515,-0.0220849421,-0.327786386,-0.229569301,-0.1966590881,0.451315999,0.0126341125,0.2062479258,-0.2833121419,0.024734322,0.051033441,0.0669321045,0.4555992186,-0.0770155117,0.0951746032,0.4936599731,0.0277595539,-0.6336435676,0.126455307,-0.1002101377,0.3581635356,0.1631138921,0.7447269559,-0.4240935445,-0.2017024904,0.2619288266,0.2276256233,-0.1391590834,-0.2503652275,-0.3379265666,-0.2450611293,-0.095224306,0.0448217429,-0.005028497,0.2524263561,0.0323803872,-0.1079626903,-0.1736579239,0.0886255056,0.1121646538,-0.0716406479,0.1417014897,-0.199632138,-0.1048961133,0.3063726723,0.1095516831,-0.0839731544,0.2140050381,-0.0825214908,-0.3581177592,0.2188200653,0.1391621977,0.3078300655,0.1588245183,0.0319344588,-0.1151004434,0.0843153,0.1755105853,-0.050372012,0.2726693451,0.4059753418,0.1034014896,-0.3175453246,-0.3351827562,0.2841389477,0.1657994092,-0.0450834595,0.5239727497,-0.0486147031,-0.1896082759,-0.0234618858,-0.0010964702,0.9536359906,-0.0492479056,-0.2009790987,0.5993869901,-0.1922727227,0.2644458711,0.0974148437,0.0734480992,-0.19073838,0.2422928959,-0.0011584021,-0.1867491901,0.2876497209,-0.1201942638,-0.3611539304,0.2661924064,-0.3703894317,0.0234664753,-0.187512219,0.0176507663,-0.1186599955,0.0171357524,-0.0771397725,0.1293452531,-0.3400009573,0.334567368,-0.1245868355,-0.1437534839,0.0719397962,-0.1858252883,0.0009278022,-0.0164667405,-0.3770629764,-0.0666012242,-0.0992202014,-0.1671610922,-0.2557806373,-0.0478083827,-0.0114818327,-0.1333717704,-0.2056403011,-0.0880465433,-0.0310733747,-0.0674435347,-0.0058650961,-0.0760333166,0.078049019,-0.1186791137,-0.4413619339,-0.0041558775,-0.0343838483,-0.0450187735,0.2178777605,0.1296660751,0.1290795207,-0.2593412995,-0.2323491722,-0.0378833897,0.1455767304,-0.1006582677,0.1564953923,-0.0954220816,-0.05825359,0.1538581103,-0.1971980631,-0.3079049587,0.1487835646,0.4695439637,0.3342171013,0.0332063548,0.1334926486,-0.1003749371,-0.062083073,-0.2934113145,0.1757628173,0.0185082275,-0.3282864094,0.2313659787,0.2405150533,-0.055555772,0.0801065266,0.1296909302,0.0106344819,-0.0905771926,-0.0761703551,-0.3307938576,-0.3389729857,0.4077080488,0.0217369497,0.1497350633,-0.1225169823,0.2605583668,0.1515266597,-0.000649187,-0.313981384,0.0422876105,-0.2909036577,0.0912730843,0.3638060987,-0.280436635,0.2271239012,0.1523882896,0.1368928701,-0.0537951998,-0.2413507849,-0.2207813114,-0.1533450335,0.173290655,-0.0766403601,0.0568135604,0.2119149864,-0.3122698665,-0.0635546371,-0.0217684582,-0.0832560584,0.1599960327,-0.2444625646,0.063909173,0.067888692,0.4049760997,0.2633251548,0.435020715,-0.0374888703,-0.0562751293,-0.1320069879,-0.0121453619,0.0759137422,-0.0076632299,-0.2801422477,-0.1537376344,0.157931,0.030304078,0.0705487877,-0.0641131029,-0.2606777847,-0.1755709946,-0.0674017072,0.1147454008,-0.1129584089,0.2070543617,-0.2954784334,0.2356394231,-0.0653063282,0.0387019292,0.1438335329,-0.1058878824,0.0996290296,0.0422345735,-0.2674396932,0.1484238952,0.1236101761,0.0310535971,0.434489876,0.029390173,0.1119508445,0.1785230339,-0.0212376248,-0.146921888,0.4369482696,0.0608740225,0.2153191417,-0.0119504426,-0.146965161,0.1183181033,-0.1156563461,0.1921877712,0.1859310567,-0.2745586336,-0.021348929,0.1608658582,-0.0788903683,-0.0849490687,-0.2915712297,0.4880608022,-0.6160899997,-0.1826809794,-0.1732410789,0.0879465714,-0.0890202895,0.1348123252,-0.3088459373,-0.1553795785,-0.1199519262,-0.3049582839,0.0645720288,-0.0108419908,0.4839222729,0.1019124314,-0.0518757738,-0.0765113235,-0.2285626233,0.2538504899,0.0013043582,-0.1830806881,0.2370769233,0.2687752247,0.2687553167,-0.1647896767,0.5545255542,0.3923517764,0.5213235021,-0.3281322718,-0.0751883835,0.1791754961,-0.2624017894,-0.0859699994,0.0825163573,0.1139351651,0.2068953216,0.060302414,0.2203259319,-0.2221734971,0.2207185328,-0.2041488141,-0.1452715248,-0.216159761,0.7197526693,-0.3471832275,-0.1299152076,-0.1262446344,-0.0426462814,-0.7244431973,0.0796926469,0.3343222737,0.2418286651,0.0092006028,-0.1986442208,0.0642050654,-0.2621262968,0.2768419683,-0.1209545657,0.4752021432,-0.4868346453,0.1296972483,-0.5684576035,0.4350523353,-0.0610749498,0.1090470776,0.0926751643,-0.1541391164,-0.1793299317,0.0654201508,0.0071407668,0.2414238453,0.2169900835,-0.1889825612,-0.6521363854,-0.0086720623,-0.0077499882,-0.1539293677,0.1701852977,-0.4815095663,0.0615987889,-0.2177682221,0.0198628716,-0.0644683987,-0.2736273706,0.2979908288,0.5125359297,0.5735932589,0.1107169613,0.5845791101,0.1717773527,-0.1879412085,0.1581155658,-0.3322859704,0.0339716524,0.0177451856,0.0612541251,0.088872157,-0.1166086271,0.0220457464,-0.3064666092,0.2479760498,-0.2203525901,-0.14985919,-0.0165363513,-0.0736364126,-0.1534781754,-0.1865145117,-0.1707282811,0.140608862,0.2815619111,0.1047723591,-0.1850739717,0.1632720679,0.7465104461,-0.2200563252,-0.3565498888,0.0949090868,0.3943425417,-0.0759176016,-0.0915759578,-0.5191881657,-0.0103437109,0.5544500351,0.0181632563,-0.009582568,0.3871769607,-0.1495027244,0.0503680371,-0.016953459,-0.2015019804,-0.0118222274,-0.0083396304,0.0239531808,-0.1598630846]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2117","title":"load_metric from local \"glue.py\" meet error 'NoneType' object is not callable","comments":"> @Frankie123421 what was the resolution to this?\r\n\r\nuse glue_metric.py instead of glue.py in load_metric","body":"actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(path='\/home\/glue.py', name=actual_task)\r\nmetric = load_metric(path='\/home\/glue.py', name=actual_task)\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 1 actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\n 2 dataset = load_dataset(path='\/home\/jcli\/glue.py', name=actual_task)\r\n----> 3 metric = load_metric(path='\/home\/jcli\/glue.py', name=actual_task)\r\n\r\n~\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs)\r\n 508 keep_in_memory=keep_in_memory,\r\n 509 experiment_id=experiment_id,\r\n--> 510 **metric_init_kwargs,\r\n 511 )\r\n 512 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\r\nPlease help","comment_length":15,"text":"load_metric from local \"glue.py\" meet error 'NoneType' object is not callable\nactual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(path='\/home\/glue.py', name=actual_task)\r\nmetric = load_metric(path='\/home\/glue.py', name=actual_task)\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 1 actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\n 2 dataset = load_dataset(path='\/home\/jcli\/glue.py', name=actual_task)\r\n----> 3 metric = load_metric(path='\/home\/jcli\/glue.py', name=actual_task)\r\n\r\n~\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs)\r\n 508 keep_in_memory=keep_in_memory,\r\n 509 experiment_id=experiment_id,\r\n--> 510 **metric_init_kwargs,\r\n 511 )\r\n 512 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\r\nPlease help\n> @Frankie123421 what was the resolution to this?\r\n\r\nuse glue_metric.py instead of glue.py in load_metric","embeddings":[-0.1147277355,-0.2722516656,0.0763981193,0.292337507,0.4450437725,-0.1978471279,0.5219500661,0.3526912332,0.5789061785,0.2391433269,-0.2604035437,0.2439481765,-0.0830543935,0.1400061846,0.1668067873,-0.2594834864,-0.2505444586,0.1009072736,-0.3236852288,0.0778844282,-0.5946435928,0.1236544698,-0.1473677456,0.1011811644,-0.1684289873,0.0775865912,-0.0001657927,0.2222328782,-0.1709766537,-0.4364847243,0.3846880496,-0.1120897755,0.1347538531,0.4944859743,-0.0001143825,0.0107692266,0.4764086008,-0.0302246623,-0.0472088829,-0.4380871058,-0.0905404612,-0.2524358332,0.376814425,-0.43908903,-0.089831762,0.1394781619,-0.0933021754,-0.3293699324,0.2070919573,0.4414151013,0.1994082183,0.6218678951,0.1063449159,-0.1903834343,0.0465061255,0.1120344326,0.0213838778,0.9023581743,0.0917100757,-0.2171845734,0.2652484477,0.1430645883,-0.3956254721,0.2273076326,0.4022140801,-0.0008820229,0.3864405453,-0.293559432,-0.0759777874,0.1330364645,0.0382718258,-0.2256373018,-0.1305482537,0.0649542436,0.0407658815,-0.2457783222,0.1412881911,-0.1435367316,-0.0825372189,-0.0598918311,-0.1041728556,0.0353218913,-0.2764250338,0.247007668,-0.0618248358,0.0815453753,-0.1384880841,0.2654392123,0.2542120218,-0.0078916848,-0.1830637604,-0.0751759782,0.0407549553,0.1800290644,-0.2874908149,-0.0167457759,0.0669409782,-0.0602712147,0.0819597766,-0.1676664948,0.0451421589,0.046093598,0.4492984414,0.2641171217,0.1890162677,0.4545835257,0.4238354862,0.131631732,0.2431296259,0.0746466368,-0.1625356823,0.0555331297,-0.22119084,-0.2059956342,0.2789880931,0.3105395436,0.7083941698,-0.0835751742,-0.4071558714,-0.1752631962,-0.218430385,-0.0118819177,0.1830362827,0.1466107368,-0.0950283706,0.2633502483,0.2650561035,0.2981959879,-0.3165906668,0.0522996783,-0.1651794016,0.2026759386,-0.257391125,0.020535117,0.2021723241,0.035043139,0.0580784418,-0.0601769499,0.0608925112,-0.125683248,0.2840404212,-0.268202275,-0.2947102189,0.1866375357,-0.047529608,-0.0272630416,0.3775945604,-0.2081023157,-0.18353118,0.0696283951,-0.3384780884,-0.4022039175,-0.1151560545,0.1585686058,-0.3454427719,0.1093512326,-0.1042753085,-0.2610677183,0.0929987729,-0.014431526,-0.0808564499,-0.2253642082,-0.2190732807,-0.2995373309,0.1517105997,0.5023953319,-0.037309885,-0.2554526031,-0.2942404747,-0.1086184978,0.2652874887,-0.1006081179,0.0151286563,0.3232223988,-0.0986961275,-0.0574657694,0.7182421088,-0.8587254882,-0.1694990247,0.3227204084,-0.2079727948,-0.0620670356,0.0751104206,-0.0774931014,0.2533204556,0.10418576,0.7282133698,0.1092011929,0.1185557917,-0.1488689333,-0.3413649201,-0.2476183325,0.1248655096,0.0971835032,0.2151804566,0.1897740513,0.0891456157,-0.2451860607,-0.0479088463,0.1383477151,-0.0829283223,-0.0580089055,0.4731233716,-0.1255285889,0.0052362401,-0.4615797102,-0.3824304044,0.2583970726,-0.3975253701,0.3563768268,-0.0636327863,0.0188708548,-0.5460512638,-0.0353674181,-0.0604976229,0.0980929062,0.142668888,0.0613249913,-0.110817574,0.0613284037,-0.1385058016,0.2435562462,-0.0965485647,0.1035519913,-0.0340862162,0.1627288461,0.0613204725,-0.3887608945,-0.1003228799,0.2657421827,0.4754844606,-0.1122552454,-0.1311980337,0.4564760923,0.0398435928,-0.1530097127,-0.1682820469,-0.0622174181,0.1244410798,-0.1486093849,0.1321699023,-0.1148179471,0.1580120325,-0.1733203381,0.2741289139,0.3910351098,0.0751065984,0.4857012331,-0.143501386,0.1804318875,-0.1615391821,-0.1265912503,-0.2430025935,-0.2190829217,0.2276919931,0.1126556322,0.3553391695,0.063788943,0.1454462856,-0.1160042882,0.5826870799,-0.0253400989,0.273175925,-0.1563038528,-0.2633813918,0.1146687269,-0.0759288818,0.085521169,0.6615589261,0.1321170926,-0.0771135092,0.1120729744,-0.1054127365,0.0256459583,0.0232591536,-0.0163554102,0.0268002208,0.1347464621,0.1076344028,-0.0858329758,-0.1316937208,0.1996504515,-0.0678506568,0.1688325256,-0.4403191507,0.0842392743,-0.1204175279,0.0025506907,-0.1226908043,-0.109127596,-0.1052140668,-0.3249551654,-0.3048239946,0.1488638967,0.1850621402,0.184903264,0.2280201465,0.1128739119,0.0207352638,-0.2497547567,-0.0035873656,-0.2087519318,-0.1242401674,-0.0610454194,0.2027999163,0.1918255538,0.3334771991,-0.3396788836,0.0419641025,0.0669331327,-0.2324443012,-0.0377044789,0.0498737469,0.5538481474,0.3362161815,0.1401687562,0.1508194953,-0.0111535601,0.5445737243,-0.2677583098,0.0105481669,0.3391666114,-0.3439410031,0.1727326065,-0.2585276067,-0.2639701068,-0.4683785737,-0.5656132102,0.1103354916,0.036934454,0.1408149749,0.1372751296,0.1162050888,0.5370131135,-0.02401039,0.0048096362,-0.1442282945,-0.2503283918,0.172005266,-0.1241062656,-0.2491289079,0.0610206947,-0.0690471083,0.410630703,0.1264333427,-0.0662921593,-0.5910329819,0.0042736665,0.1663122326,-0.087583147,-0.0518641658,0.2198337018,-0.0109312264,-0.0312076509,-0.1785449535,-0.1086454019,0.3081144691,0.3683545291,-0.0459925197,0.1363203675,0.3582895696,-0.1786957085,0.4112913907,-0.0850848928,-0.1891878694,0.3793115914,-0.1115222573,0.3426722884,-0.1299353093,-0.5101319551,-0.0617492385,0.0616553947,0.2778909206,-0.1482557207,-0.0625169426,-0.130837664,-0.4408022761,-0.2569326758,-0.2242600471,-0.2941062748,0.1291534156,-0.0909341574,0.1436886489,-0.0475341454,0.0250054076,-0.4909591079,0.1159061417,0.0237212796,0.1990761161,0.0722715482,-0.0225533508,-0.3116520047,-0.1933846474,-0.2816130817,0.4606595337,0.0025459665,0.194717899,-0.2118605524,0.0071834219,0.020246176,0.0848257616,0.4501301646,-0.1359494328,0.0663774088,0.4806345999,0.0405305624,-0.653850317,0.1134532988,-0.0589164943,0.3324570656,0.1958316416,0.7789819837,-0.4639755487,-0.1980190724,0.236846745,0.2162100226,-0.1412068754,-0.3087222576,-0.3348046243,-0.2271769345,-0.1316966861,-0.0196587928,-0.0080828173,0.2710976899,0.0620063581,-0.1115108058,-0.1611374021,0.0211778302,0.0835725665,-0.0895449817,0.1783205122,-0.2302525491,-0.0758128688,0.3116340935,0.1124374419,-0.1250580996,0.2006300688,-0.1128574461,-0.3388844728,0.2116495669,0.1412357092,0.3389234543,0.1161232367,0.0184453335,-0.0974975899,0.0751723275,0.1687321365,-0.0268588103,0.2864470184,0.3784522116,0.0555451438,-0.2557336986,-0.3263701797,0.2446720749,0.1801453978,-0.0476871245,0.4970990419,-0.0652567595,-0.2364863008,0.0044212509,-0.0745192692,0.9051439762,-0.0260113571,-0.2398271859,0.5772464871,-0.2122777104,0.2895710766,0.0768682584,0.0424837805,-0.1717258394,0.177105248,0.0502462126,-0.1855963618,0.2790083289,-0.0861318335,-0.3573915362,0.3310429156,-0.3398445249,0.0341432989,-0.1849597096,0.0923868716,-0.1330191344,-0.0350222625,-0.0474282876,0.1211489588,-0.2751029134,0.2809014916,-0.0997218564,-0.1354786903,0.1013950333,-0.2199016213,0.0269520804,-0.0607558787,-0.4086233377,-0.0760370195,-0.0684967935,-0.1493056864,-0.243711248,-0.0621272959,-0.0232466143,-0.1238985434,-0.2030958831,-0.078036055,0.0286590736,-0.0232856013,0.035970971,-0.054961089,0.0408570729,-0.0899726078,-0.3945325911,-0.0158464164,-0.0582024939,-0.0861487165,0.2090536803,0.0958870947,0.1116082221,-0.2635337412,-0.2416461557,-0.0553637408,0.1601193994,-0.1014174744,0.1551358998,-0.088816151,-0.0532828048,0.1237345785,-0.237873733,-0.3137809038,0.1153747737,0.4537363052,0.2936981916,0.0395494699,0.0874688178,-0.1336937696,-0.0488690883,-0.2490254641,0.1057779044,0.0536958426,-0.3350604177,0.2138357311,0.2610098422,-0.1099501029,0.0454889014,0.1041223332,0.0436132886,-0.0973302498,-0.0838835388,-0.3004059792,-0.3459156752,0.4484198093,0.0289751925,0.1328529567,-0.1717648804,0.2663173378,0.111386098,0.0067512835,-0.3057831824,0.038467098,-0.227649644,0.0898993015,0.416230768,-0.2732821703,0.2342763245,0.1714320779,0.1320744753,-0.0218834281,-0.2366810292,-0.2132603973,-0.1669036895,0.1651955992,-0.0991374329,0.0413741954,0.2509911358,-0.3114293218,-0.0661969855,-0.0231649857,-0.0336046293,0.1448616534,-0.254106164,0.020233199,0.0587951504,0.357647568,0.2667367756,0.4359711409,-0.0391970985,-0.0571494028,-0.113494195,-0.0077659274,0.0437720269,0.0239718985,-0.2320218682,-0.1910102069,0.1641863734,0.0436676294,0.0420143716,-0.0410486162,-0.2935394049,-0.1515930444,-0.0148123568,0.084367767,-0.0979734287,0.2115448564,-0.3021131456,0.2261651009,-0.0645599738,0.0282784123,0.1544769555,-0.1422552466,0.1627965271,0.0710029751,-0.2179964036,0.1954166889,0.1421339065,0.0064530321,0.4465040267,0.0308399964,0.0935424343,0.1327252239,-0.0446310453,-0.1368142664,0.453263253,0.0014511712,0.2097149193,-0.0483290926,-0.0993311033,0.1316953599,-0.0902820453,0.2098243237,0.152858302,-0.2246890366,-0.0514357686,0.1471545547,-0.1065608561,-0.0779328644,-0.2697863579,0.5497053862,-0.5956543684,-0.1855589896,-0.1398821175,0.0751105174,-0.0611178614,0.0873054489,-0.2260777652,-0.1228880286,-0.1375590414,-0.2766264379,0.0559883527,-0.028626224,0.4973982871,0.0492386706,-0.014978582,-0.0702334344,-0.2523667812,0.2445088923,-0.016294457,-0.1581234038,0.235274002,0.2126837224,0.2662226856,-0.1381310523,0.5607129931,0.3751700222,0.5017942786,-0.2857939005,-0.1380770952,0.1855448633,-0.2190442383,-0.1076862812,0.0846967697,0.1430392265,0.1701195687,0.0536506176,0.201838091,-0.2494162619,0.1747214347,-0.2252972126,-0.1250921935,-0.2295571566,0.7471730113,-0.3336187601,-0.1541941762,-0.095358409,-0.0370797999,-0.7127786875,0.0541500412,0.2993370295,0.2036420405,0.0233094338,-0.2516190112,0.0691100135,-0.2581923306,0.2637782693,-0.1534438133,0.4763951898,-0.5100595355,0.1261321157,-0.5974845886,0.3807201087,-0.0842373446,0.1471813619,0.0778679848,-0.123294577,-0.1890851855,0.1051667109,0.024047887,0.3096905351,0.232651785,-0.2559272647,-0.6295664907,0.0054344754,-0.0194124021,-0.1742747426,0.1982780546,-0.5056791306,0.0425511263,-0.1963361651,0.0259049088,-0.0479561538,-0.2661556304,0.3450884521,0.5361177325,0.5703440309,0.1294959933,0.5637575984,0.1599162519,-0.1523956656,0.1260448396,-0.360574156,0.03276591,-0.032112617,0.0453277901,0.0697444081,-0.1543253958,-0.0224322677,-0.3226998448,0.2339266092,-0.1895594299,-0.1399300843,-0.0627184138,-0.0462949723,-0.1425766498,-0.2043561041,-0.1815452129,0.1525194347,0.2617765069,0.0742976591,-0.1545443982,0.1426274776,0.7160027623,-0.2191270292,-0.3511478603,0.1273013353,0.4129334986,-0.0309052523,-0.0741253793,-0.4815787673,-0.035461992,0.5094004273,0.0533091798,0.0234334096,0.3962220252,-0.1341096014,0.0943176076,-0.0011544406,-0.2425254732,0.0017806344,-0.0359264687,0.0025508737,-0.1539929509]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2117","title":"load_metric from local \"glue.py\" meet error 'NoneType' object is not callable","comments":"thank you!","body":"actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(path='\/home\/glue.py', name=actual_task)\r\nmetric = load_metric(path='\/home\/glue.py', name=actual_task)\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 1 actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\n 2 dataset = load_dataset(path='\/home\/jcli\/glue.py', name=actual_task)\r\n----> 3 metric = load_metric(path='\/home\/jcli\/glue.py', name=actual_task)\r\n\r\n~\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs)\r\n 508 keep_in_memory=keep_in_memory,\r\n 509 experiment_id=experiment_id,\r\n--> 510 **metric_init_kwargs,\r\n 511 )\r\n 512 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\r\nPlease help","comment_length":2,"text":"load_metric from local \"glue.py\" meet error 'NoneType' object is not callable\nactual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(path='\/home\/glue.py', name=actual_task)\r\nmetric = load_metric(path='\/home\/glue.py', name=actual_task)\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in \r\n 1 actual_task = \"mnli\" if task == \"mnli-mm\" else task\r\n 2 dataset = load_dataset(path='\/home\/jcli\/glue.py', name=actual_task)\r\n----> 3 metric = load_metric(path='\/home\/jcli\/glue.py', name=actual_task)\r\n\r\n~\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs)\r\n 508 keep_in_memory=keep_in_memory,\r\n 509 experiment_id=experiment_id,\r\n--> 510 **metric_init_kwargs,\r\n 511 )\r\n 512 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\r\nPlease help\nthank you!","embeddings":[-0.1685553938,-0.34980762,0.0580280609,0.3419708014,0.3750133514,-0.110481374,0.4141913354,0.3510377705,0.6199297905,0.2367782146,-0.2564356029,0.1840084493,-0.1458742172,0.1512304693,0.1883033365,-0.2143974155,-0.2564520836,0.1548571289,-0.2997356653,0.0473166592,-0.5600094199,0.1028408706,-0.1620021313,0.1327151656,-0.1345098019,-0.0536667369,0.0704949424,0.1623747647,-0.2327752709,-0.417309165,0.3553393483,-0.1015638858,0.1909606755,0.5567763448,-0.0001131053,-0.0202314164,0.4110562205,-0.0233401321,-0.0673664063,-0.5466103554,-0.1050174981,-0.2489411235,0.3510325253,-0.4352876544,-0.0805146471,0.1060220301,-0.0789937079,-0.2841940522,0.3118450344,0.4886212647,0.1979496628,0.6239976287,0.061246261,-0.2349032015,0.077315107,0.1554490924,0.0479787849,0.7890283465,0.0029998305,-0.1611451507,0.3759127557,0.1702410877,-0.4772416949,0.1698909551,0.3649314642,-0.0829430446,0.3727242947,-0.2869613767,-0.0648366511,0.2066849023,0.1195715219,-0.3138421476,-0.0887074322,0.1235221997,0.0526251197,-0.2789075971,0.0740658641,-0.086970672,-0.0942254663,-0.0643691123,-0.1171839461,0.0063278349,-0.3216781914,0.321865052,-0.0057426156,0.0809005275,-0.1394013017,0.2443594187,0.3130799234,-0.0211520419,-0.1470500231,-0.0496805944,0.0323218144,0.2289241105,-0.2396262437,-0.044567395,0.0220701247,-0.1211019009,0.0792391151,-0.1042634323,0.1078391299,0.0225275476,0.3588025868,0.3652290702,0.1912555248,0.3294060826,0.424819082,0.1800665706,0.2196908742,0.0637136027,-0.1990335435,0.0251446124,-0.2508533597,-0.1942585409,0.3345648944,0.3097310364,0.6467584968,-0.0698507279,-0.4557101429,-0.1424943209,-0.1719486266,-0.018760195,0.2571581304,0.2100190967,-0.1107507572,0.2116214037,0.256134361,0.2828479707,-0.3074067533,-0.0246119928,-0.1839692295,0.265598923,-0.2322266698,0.025075132,0.2131676972,0.0735751763,0.0600827895,-0.0590532944,0.0392147973,-0.1472292542,0.3551396132,-0.3385255039,-0.3098196089,0.1780318618,0.0001947732,-0.0425389111,0.3914494812,-0.1759262979,-0.2485822141,0.0755736455,-0.3061844409,-0.3294753134,-0.1346795708,0.1940579414,-0.2503405809,0.0887048915,-0.1104497761,-0.2587857544,0.0641072914,-0.043826241,-0.0550493337,-0.174520269,-0.1248540804,-0.2740499377,0.1265850663,0.5533964038,0.0117020523,-0.2785338759,-0.2841161191,-0.144921422,0.2617201507,-0.1360828131,0.0007422832,0.367549777,-0.0746124834,-0.0655993074,0.8031316996,-0.8139867187,-0.0791545138,0.3298468292,-0.1916870922,-0.1929976791,0.0745472759,0.0077429102,0.2385612279,0.1618704647,0.6445534825,0.1687407494,0.1290209144,-0.1712295711,-0.2578566074,-0.2113750577,0.1039635539,0.1269871294,0.2191532105,0.1510202587,0.1055660322,-0.1874834001,-0.1062621698,0.1241052449,-0.1632494181,-0.0188787654,0.504845202,-0.1169917956,-0.0698220208,-0.6075663567,-0.4027897716,0.1456638128,-0.3798860312,0.2700133026,-0.0065209232,0.0578655228,-0.5429913402,0.0027168859,-0.0774873048,0.1676043868,0.1600174457,0.0615823269,-0.1478934586,0.0421038419,-0.1701498479,0.2122327834,-0.0889297426,0.1334410459,-0.0273561701,0.1682845503,0.0308560692,-0.3751284182,-0.0681841001,0.2566975951,0.5198993683,-0.0944874734,-0.1016743332,0.4357823133,0.0606552586,-0.1309800297,-0.1262753606,-0.0552580729,0.1120153591,-0.1677200943,0.1194887683,-0.1164356321,0.1949766725,-0.1652452499,0.225124687,0.4219171405,0.0511814728,0.3971928954,-0.1622430086,0.1542665362,-0.1776134968,-0.1342253834,-0.2034023851,-0.1770776361,0.2670669556,0.1109021157,0.3383800387,-0.0103274323,0.07327196,-0.2296907157,0.5127056837,-0.0296619367,0.3121167123,-0.1621802896,-0.2580148578,0.1671088636,-0.0790392831,0.0064939307,0.5880747437,0.1785638183,-0.1150373667,0.0652488172,-0.0733485296,0.0381834954,0.0011072926,-0.1027942821,0.0794894025,0.1265294254,0.0277313963,-0.0477866344,-0.15826796,0.1311757267,-0.0906058773,0.2440425307,-0.3869230449,0.1270682514,-0.0901006311,-0.0292759389,-0.1891007721,-0.1231891289,-0.1268986613,-0.2819415033,-0.3203087747,0.2583920062,0.2260811627,0.1673618704,0.1855582148,0.099921301,0.0356796607,-0.2956068814,-0.0614341833,-0.1561578363,-0.0838585049,-0.0588956662,0.2739300728,0.2105236053,0.4152212739,-0.3460072577,0.0842090771,0.0950999632,-0.1782737225,-0.0287448149,0.093054004,0.4316760898,0.3089473546,0.2142065912,0.1337783784,-0.0350658335,0.5338582993,-0.3593396544,-0.0020254683,0.384544611,-0.3558255732,0.119394511,-0.186661005,-0.2660482526,-0.4716691673,-0.4813078642,0.1021534353,0.058193028,0.1822191328,0.2013226748,0.1627873629,0.5363111496,-0.0039116265,0.020459637,-0.1068014875,-0.275975436,0.1784607917,-0.1416900754,-0.2389206439,0.1174266636,-0.0102969259,0.4248230457,0.08925163,-0.0671504065,-0.5035201311,0.0024323692,0.1602489352,-0.1361293197,-0.0643690228,0.2499776334,0.0120530948,-0.0399000943,-0.1415620595,-0.1215814576,0.2691751719,0.3989044726,-0.0776900873,0.0480477437,0.3521622717,-0.2583193183,0.3864822388,-0.0333078317,-0.1945776045,0.372425139,-0.126585111,0.3453998566,-0.1206981763,-0.5470819473,-0.043998722,0.1495205611,0.187073186,-0.1136195287,-0.1702493429,-0.1493228227,-0.4479903877,-0.2434474975,-0.315805614,-0.1987373829,0.1090660021,-0.0766177252,0.1175456271,-0.1108141989,0.0212757792,-0.5275787115,0.0752441809,0.0349299572,0.2356918752,0.0351738594,-0.0066928845,-0.3124481142,-0.1908185333,-0.2314839959,0.4698508382,-0.0696070567,0.1515567154,-0.250934124,0.0894053951,0.0180086531,0.077099286,0.440228343,-0.1043859422,0.0778587162,0.4173851609,0.0655661896,-0.604588151,0.1462458819,-0.1268972605,0.3384560645,0.1919890046,0.7410521507,-0.4533586204,-0.1768128276,0.2418891639,0.2159191668,-0.1422587186,-0.3212428093,-0.3432694376,-0.2297086567,-0.1449143142,0.058530204,-0.0153985359,0.3267474174,0.0983880013,-0.0996437669,-0.1585197002,0.0257953107,0.1643201411,-0.0443039164,0.1925929785,-0.1841146201,-0.1068857461,0.2872350216,0.0473013185,-0.0358494744,0.2201441377,-0.091779083,-0.3363606334,0.1352170557,0.1612511426,0.2978087664,0.0779191032,0.0712375417,-0.0753595904,0.1326232105,0.2079142928,-0.0053935703,0.3072417676,0.3632487357,0.0845581517,-0.2937021852,-0.3604342639,0.297329098,0.1853087395,-0.0018191569,0.4595264792,-0.1230680197,-0.2024610639,-0.0091202054,-0.0083418554,0.9078728557,-0.0819074735,-0.216286093,0.5706056952,-0.2141142488,0.2388515472,0.0517221056,0.0208282173,-0.2081675231,0.1617904603,0.0024999552,-0.1762482524,0.2938710451,-0.0683652163,-0.363574326,0.3157187998,-0.3501088619,0.063509509,-0.2380225658,0.1316692382,-0.0627651587,-0.00127613,-0.0338055789,0.1347332299,-0.3502756953,0.3137503266,-0.0980719477,-0.1778634191,0.1453637481,-0.1942169219,-0.0188571922,-0.0499049127,-0.4294371903,-0.0927732214,-0.0986543223,-0.1772540808,-0.2501615882,0.0102684107,0.0137038119,-0.1171617061,-0.2222367078,-0.1034558713,-0.0459316149,-0.0937660038,-0.1347726434,-0.0254443008,0.0711226836,-0.1461118609,-0.4228412509,0.0362388678,0.0377004407,-0.0720058158,0.2785125077,0.0848013386,0.1161891893,-0.2784797251,-0.2491813898,-0.0254313387,0.1461551487,-0.126206547,0.1611812562,-0.0651832893,-0.0782891586,0.0936220735,-0.1461143643,-0.2888533771,0.0820126384,0.450143218,0.2684813142,0.0608566776,0.1588042974,-0.0886842683,-0.0595882647,-0.3241751194,0.2139002979,0.0739597678,-0.2220634073,0.2628046572,0.3162993193,-0.0540257655,0.0541858599,0.1144158021,0.0360508449,-0.07634148,-0.0290097557,-0.2949714065,-0.317777127,0.3878457844,0.0244352706,0.1234029382,-0.1111139283,0.2973822355,0.1784714013,0.0726291165,-0.3265408278,0.0242256317,-0.2812577188,0.0593949072,0.3123654127,-0.2696484327,0.2015760541,0.1687636971,0.1460365504,-0.1110710204,-0.232742101,-0.2337256968,-0.165841803,0.1458962709,-0.0662452504,0.0884240493,0.2228495926,-0.284127593,-0.0809907168,-0.0942704454,-0.1327468008,0.1173133254,-0.1916665882,0.155485034,0.089561291,0.4022723138,0.2080384493,0.3825613558,-0.0536832884,-0.0765099004,-0.1424289495,-0.0096887806,0.132453382,-0.0279151667,-0.2746556401,-0.1597284377,0.2723317444,0.0143242478,0.0147231305,-0.0654639974,-0.2673206329,-0.0737505183,-0.0667322949,0.0858689174,-0.1188276932,0.2581942081,-0.2037381381,0.2547803819,-0.1152385175,0.0595048107,0.1546072066,-0.1110335365,0.1102912501,0.0849190801,-0.3354355395,0.1412386745,0.081842266,0.035222061,0.3830217719,-0.000142243,0.1095449403,0.1286327541,0.0200379398,-0.1649656743,0.4456427097,0.1731927991,0.225072816,-0.0483670942,-0.143459186,0.1076849923,-0.1808909476,0.1340731829,0.1668551862,-0.2817907631,-0.0956750661,0.1721151322,-0.0578521006,-0.1230919436,-0.3027980626,0.5518863201,-0.5692595243,-0.1927296668,-0.184882164,0.080280602,-0.0721067116,0.1203312725,-0.3314254284,-0.071428284,-0.1422833055,-0.3342520297,0.1258567423,-0.1097721383,0.4551662505,0.0825816095,-0.0353992544,-0.0486630537,-0.2366649657,0.3436912894,-0.0951974392,-0.2284216881,0.2163337469,0.215164721,0.2552423179,-0.1783218831,0.5161550045,0.3320259154,0.5135211349,-0.3464242518,-0.0568446517,0.1583365947,-0.295093894,-0.0442395695,0.0339266285,0.0782727078,0.2263283432,0.0433610305,0.2289879322,-0.2450909764,0.2140584886,-0.200831145,-0.1990321577,-0.2014454007,0.7180337906,-0.2896307409,-0.171712324,-0.1484985501,-0.0752997696,-0.7071778774,0.074640356,0.2735154927,0.197238341,0.0416189246,-0.1872563064,0.0615368523,-0.2409869134,0.2599098086,-0.1563440859,0.4647095799,-0.5037167072,0.1321616918,-0.5583600402,0.3744518161,-0.1084669232,0.0775165111,0.1211926639,-0.1616617292,-0.2069649398,0.0563469715,-0.0195016991,0.2846516371,0.1977089047,-0.174218446,-0.6174495816,-0.0221258774,-0.0621041209,-0.1155509353,0.1986081898,-0.4332520664,0.0147060398,-0.2521724105,0.0238038339,-0.0552998781,-0.2056779861,0.2907704711,0.4587490559,0.6160825491,0.0637737587,0.5545130968,0.1112187728,-0.1168937683,0.2323438376,-0.2687861025,-0.0220868811,-0.0202973522,0.1132019609,0.0897674859,-0.1285637915,0.0039956747,-0.3306355476,0.2956731021,-0.1884841323,-0.2135472596,0.0411104895,-0.0761229917,-0.1376555413,-0.1779044718,-0.176501438,0.0560270213,0.2343596667,0.1094192713,-0.160665676,0.11024867,0.7090404034,-0.1864078343,-0.3241396546,0.0919159576,0.4307519495,-0.0169513784,-0.0848304182,-0.5384306312,0.0690004677,0.479647994,0.0348160416,0.0220223274,0.378059417,-0.1048230007,0.108712405,0.0000393656,-0.2044528872,0.0387086533,-0.044727046,0.0150476564,-0.1841933727]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2116","title":"Creating custom dataset results in error while calling the map() function","comments":"Hi,\r\n\r\nthe `_data` attribute is missing due to `MyDataset.__init__` not calling the parent `__init__`. However, I don't think it's a good idea to subclass the `datasets.Dataset` class (e.g. it's kind of dangerous to override `datasets.Dataset.__getitem__`). Instead, it's better to follow the \"association over inheritance\" approach with a simple wrapper class that delegates calls to a wrapped `Dataset` (map, etc.). Btw, the library offers the `datasets.Dataset.from_pandas` class method to directly create a `datasets.Dataset` from the dataframe.","body":"calling `map()` of `datasets` library results into an error while defining a Custom dataset.\r\nReproducible example:\r\n```\r\nimport datasets\r\nclass MyDataset(datasets.Dataset):\r\n\r\n def __init__(self, sentences):\r\n \"Initialization\"\r\n self.samples = sentences\r\n\r\n def __len__(self):\r\n \"Denotes the total number of samples\"\r\n return len(self.samples)\r\n\r\n def __getitem__(self, index):\r\n \"Generates one sample of data\"\r\n # Select sample\r\n # Load data and get label\r\n samples = self.samples[index]\r\n\r\n return samples\r\n\r\ndef preprocess_function_train(examples):\r\n inputs = examples\r\n labels = [example+tokenizer.eos_token for example in examples ]\r\n inputs = tokenizer(inputs, max_length=30, padding=True, truncation=True)\r\n labels = tokenizer(labels, max_length=30, padding=True, truncation=True)\r\n model_inputs = inputs\r\n model_inputs[\"labels\"] = labels[\"input_ids\"]\r\n print(\"about to return\")\r\n return model_inputs\r\n\r\n\r\n##train[\"sentence\"] is dataframe column\r\ntrain_dataset = MyDataset(train['sentence'].values.tolist())\r\ntrain_dataset = train_dataset.map(\r\n preprocess_function,\r\n batched = True,\r\n batch_size=32\r\n )\r\n```\r\n\r\nStack trace of error:\r\n```\r\nTraceback (most recent call last):\r\n File \"dir\/train_generate.py\", line 362, in \r\n main()\r\n File \"dir\/train_generate.py\", line 245, in main\r\n train_dataset = train_dataset.map(\r\n File \"anaconda_dir\/anaconda3\/envs\/env1\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1244, in map\r\n return self._map_single(\r\n File \"anaconda_dir\/anaconda3\/envs\/env1\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 149, in wrapper\r\n unformatted_columns = set(self.column_names) - set(self._format_columns or [])\r\n File \"anaconda_dir\/anaconda3\/envs\/env1\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 526, in column_names\r\n return self._data.column_names\r\nAttributeError: 'MyDataset' object has no attribute '_data'\r\n```","comment_length":75,"text":"Creating custom dataset results in error while calling the map() function\ncalling `map()` of `datasets` library results into an error while defining a Custom dataset.\r\nReproducible example:\r\n```\r\nimport datasets\r\nclass MyDataset(datasets.Dataset):\r\n\r\n def __init__(self, sentences):\r\n \"Initialization\"\r\n self.samples = sentences\r\n\r\n def __len__(self):\r\n \"Denotes the total number of samples\"\r\n return len(self.samples)\r\n\r\n def __getitem__(self, index):\r\n \"Generates one sample of data\"\r\n # Select sample\r\n # Load data and get label\r\n samples = self.samples[index]\r\n\r\n return samples\r\n\r\ndef preprocess_function_train(examples):\r\n inputs = examples\r\n labels = [example+tokenizer.eos_token for example in examples ]\r\n inputs = tokenizer(inputs, max_length=30, padding=True, truncation=True)\r\n labels = tokenizer(labels, max_length=30, padding=True, truncation=True)\r\n model_inputs = inputs\r\n model_inputs[\"labels\"] = labels[\"input_ids\"]\r\n print(\"about to return\")\r\n return model_inputs\r\n\r\n\r\n##train[\"sentence\"] is dataframe column\r\ntrain_dataset = MyDataset(train['sentence'].values.tolist())\r\ntrain_dataset = train_dataset.map(\r\n preprocess_function,\r\n batched = True,\r\n batch_size=32\r\n )\r\n```\r\n\r\nStack trace of error:\r\n```\r\nTraceback (most recent call last):\r\n File \"dir\/train_generate.py\", line 362, in \r\n main()\r\n File \"dir\/train_generate.py\", line 245, in main\r\n train_dataset = train_dataset.map(\r\n File \"anaconda_dir\/anaconda3\/envs\/env1\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1244, in map\r\n return self._map_single(\r\n File \"anaconda_dir\/anaconda3\/envs\/env1\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 149, in wrapper\r\n unformatted_columns = set(self.column_names) - set(self._format_columns or [])\r\n File \"anaconda_dir\/anaconda3\/envs\/env1\/lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 526, in column_names\r\n return self._data.column_names\r\nAttributeError: 'MyDataset' object has no attribute '_data'\r\n```\nHi,\r\n\r\nthe `_data` attribute is missing due to `MyDataset.__init__` not calling the parent `__init__`. However, I don't think it's a good idea to subclass the `datasets.Dataset` class (e.g. it's kind of dangerous to override `datasets.Dataset.__getitem__`). Instead, it's better to follow the \"association over inheritance\" approach with a simple wrapper class that delegates calls to a wrapped `Dataset` (map, etc.). Btw, the library offers the `datasets.Dataset.from_pandas` class method to directly create a `datasets.Dataset` from the dataframe.","embeddings":[-0.3095269501,0.191983223,-0.0280814078,0.0809183568,0.232263878,0.0244533066,0.406316489,0.373221457,0.2232667953,0.0579666048,0.1070924476,0.4760219455,-0.3742123246,-0.0511050224,0.078491807,-0.008936286,0.0559965968,0.1856331676,-0.1835915893,-0.0389979072,-0.3738663197,0.1936639845,-0.1824554652,0.2054496109,-0.210869208,-0.2884715199,0.1117464006,-0.0803328604,-0.2930240929,-0.3859240413,0.3439577818,-0.0687114596,0.0053722714,0.7015194893,-0.0001045329,0.0541853383,0.1606344581,-0.0544360206,-0.0714643225,-0.2946879864,-0.1300773621,-0.0597293377,0.0653227791,-0.3553414047,-0.0708623379,-0.2266567498,-0.220487982,-0.5003935099,0.2376445383,0.5857280493,0.3092918396,0.0065489239,0.0157839805,-0.2425520867,0.2406726778,0.2789570093,-0.1627862453,0.1254620999,0.1019261479,-0.4344261885,0.0211201198,0.2214974761,-0.0519024022,0.0171142984,0.2251113802,-0.0729819164,0.0247415639,-0.3594280183,0.0332762487,0.130114764,0.4118058085,-0.5375848413,-0.3002355099,0.0617003068,0.017713232,-0.1704155803,0.256087929,0.0747603178,-0.1573248357,0.0529475473,-0.0775137767,0.1310314834,-0.0968102962,0.3694398403,0.1306232065,0.1370378733,-0.163695991,0.2666483521,0.1102192923,-0.4297789037,0.0232802052,-0.1080652922,0.079526633,0.4080006778,-0.3664742112,-0.1668650508,0.0558801256,-0.4467190802,0.1431356817,-0.1520475298,0.0226492062,-0.1222045124,0.1436309367,0.1354880631,0.2312283814,0.1579550654,-0.020108683,0.4038410485,-0.0876074061,-0.0010177116,-0.2695802748,-0.0837880298,0.0295201782,-0.226069212,0.0857030228,0.0621499643,0.1651797742,0.0209620092,0.0974718556,0.0965921432,-0.3174827993,0.1343469769,0.1440456808,0.2185225636,0.1541198641,-0.1662794203,-0.0625314713,0.1299290508,-0.1169798896,0.203919217,-0.2468061596,0.1485862881,-0.3834022582,0.0072864708,0.1762004644,0.1544377357,0.2860071361,-0.0002727853,0.0227019284,-0.1487195492,0.1852497607,-0.348821491,0.1049800441,0.1605466306,-0.0992633328,0.2123647481,0.2893146574,-0.0225479286,-0.1482504755,0.0764975175,-0.1301563233,-0.3638644218,-0.1915854365,0.2496528625,0.0830784664,0.2395719886,-0.1176466122,0.2286229581,0.4339629114,-0.1881231517,0.0269856546,-0.1968439668,-0.2700251341,-0.352468878,-0.0846340805,0.1618772894,-0.3914586306,0.1544375569,-0.1199159697,-0.0528848283,-0.0214935802,0.1791034937,-0.3845655322,0.521454215,-0.1856967807,0.3085446954,0.475150764,-0.1975558698,-0.4616967142,0.1067754626,-0.0429601111,-0.0831896737,-0.3642078936,0.0197023433,0.3551739156,-0.0662748963,0.1959029734,0.016236145,-0.0600998737,0.1576967537,0.0483646989,-0.0516046844,0.1196902916,0.0165812187,-0.013093696,0.1287436187,0.2245815694,-0.1051992998,-0.0357201546,-0.2099777907,0.0927982405,0.3010170758,0.3053904772,0.1411412805,-0.1029464677,-0.1510809064,-0.2661797106,0.1586987823,0.1626496464,0.1728419065,-0.2033346444,-0.1110057309,-0.2070168555,0.0860201716,-0.1815240979,-0.0441401824,0.2680633068,-0.0587690622,0.1048195809,0.0061802985,-0.0609137565,0.0378706604,-0.3639878333,0.1085462272,-0.1308040172,-0.0567317382,0.0285437759,-0.203426227,0.0002930372,0.1345686615,0.1134953424,0.0211040508,-0.2943387628,0.3483093679,0.2806219757,-0.1980163604,-0.0684035048,0.0162229929,-0.1527741551,-0.2838914394,0.0488851033,0.1119783223,0.1651587188,-0.1241091639,0.002693682,0.159896329,0.0852550641,0.0486392602,-0.0865831822,0.183762297,0.2777926326,0.1556942165,-0.1324805766,-0.2694280148,-0.1678481102,0.0698685795,0.3245127201,0.016435599,0.0377558321,-0.2219960243,0.438349247,0.1324268579,0.1742410362,0.1287521273,-0.0500493273,-0.0606957451,0.0421922058,0.3934545517,0.5440600514,0.0357974358,-0.1568249613,0.0857430696,-0.156833455,0.0600520298,0.0551590733,0.0921577364,0.3318344653,-0.0541455001,0.041653946,0.0244366173,-0.1346431226,-0.3162122965,0.0016258216,0.2799021006,-0.1794581413,-0.0567138083,-0.0656723306,-0.0579748079,-0.0733893514,-0.2275720686,-0.0409509949,-0.1828249693,-0.2411832511,-0.0484186821,-0.0376831666,0.266205579,0.0047602723,-0.0348585099,0.1677853316,-0.3207717836,0.321235925,-0.3943755031,-0.1202824786,0.1123848632,0.0908779502,-0.0423453748,0.3318918347,0.1308334172,0.0404118188,-0.2614420056,-0.1993706971,0.0841476321,-0.0970380679,0.1033655852,0.3699523807,0.0143367508,-0.2100001723,-0.180920884,0.3269987106,-0.1559433192,-0.0598935224,0.0247111749,-0.1132064238,-0.1555879265,-0.1792828292,-0.3093281388,-0.2378805727,-0.2599627078,0.0836842656,-0.0193547867,0.1814557165,0.1999303401,0.2797604799,0.3256443739,-0.0174704157,-0.1835035235,-0.2639644444,-0.2820212245,0.1595141888,-0.5192967057,-0.1890835762,0.1598889977,-0.1196864843,0.4478085041,0.312395066,-0.3665910363,-0.0338880382,-0.0988793522,0.1386392117,-0.0980389714,-0.1112381741,0.4257833958,0.1343908161,-0.0249691941,-0.2975027859,-0.435723722,0.2185882926,-0.0120279333,0.1079024076,0.187114343,0.6236602664,-0.0216526594,0.7203986049,0.3328924477,-0.3126322329,0.2853400409,-0.1934003085,0.2419656664,-0.2737648189,-0.368257463,-0.0074806199,0.0095230909,-0.0864240155,0.0665627941,-0.1792473942,-0.3093301356,0.0403918363,0.2701489329,-0.3075177073,-0.2517105043,0.1083584204,-0.197545439,0.3764651716,0.0067420099,0.1433413923,-0.4830636382,0.0740653351,-0.0361063965,-0.0577552393,-0.0475460738,0.0341844931,-0.5273029208,0.0996713415,-0.3856763542,0.1921501607,0.163949281,0.4370235801,-0.1329728961,0.1676364243,-0.1669909805,0.2914547622,0.3413004875,-0.5667746067,-0.3351866305,0.2215778232,0.0317936949,-0.5315785408,0.0515852533,0.0581913032,0.4367112219,0.0312546603,0.4528930187,-0.1641841978,-0.0960778967,0.1280760467,-0.0742535666,0.0372081622,-0.1391509324,0.0831234455,-0.3158683181,-0.2392450273,-0.0701328218,0.1149533242,0.1528109014,0.2315307409,-0.1824421585,-0.0474486165,0.0338759162,0.0382725224,0.1303159744,0.2764235437,0.0445924029,0.1654510498,0.0012380061,-0.1220256984,0.2366396785,0.3606572151,-0.2667715847,-0.4080092609,-0.0055063884,-0.0355089493,0.2581010461,0.0740618408,-0.1869594604,-0.0564565957,-0.296377033,0.0111720506,-0.0714506954,0.1902810782,0.4804243445,-0.0778858289,-0.3768902719,-0.6103817225,0.1613931656,0.1002551913,0.0144903511,0.5090961456,-0.1684102416,-0.2636221349,0.4866535962,0.0940571725,0.6123242974,-0.1272228211,-0.08780922,0.1426446438,0.1132467389,0.3553879857,-0.0312677212,0.0167544615,-0.4292021096,-0.0213393457,-0.066704154,-0.2331680208,0.2304184139,0.3485708535,-0.1894580275,0.1164674684,0.0065894197,0.4110319018,-0.0401142202,0.1207791939,0.2615002096,-0.4369436204,-0.212634325,0.3301955462,-0.1307959706,0.0271132961,-0.1017588973,-0.0546173379,-0.2210533917,-0.283873558,-0.2676754296,-0.0428069308,-0.5681089759,0.3580832183,-0.028971104,0.0284080077,0.283126384,0.3539125919,0.1921068877,0.2902731001,-0.0644247755,-0.0287650488,0.1674306691,0.1439158618,-0.0080144349,-0.0916815102,0.125313133,0.0417786501,-0.1290607303,0.0896147564,-0.0845729113,-0.1493299603,-0.1274369657,-0.0991619155,0.2395207882,-0.5454420447,-0.283567667,-0.2808873355,-0.0922459662,-0.1747676581,0.1964514405,0.0148830255,0.0139843347,0.3123815656,-0.2065937668,-0.2695235014,0.0617656782,0.1066493243,-0.0698231533,0.1705077589,0.4268480539,0.0121020628,-0.0465338603,-0.3360988796,0.179516986,0.119181484,-0.3774202466,0.1738290042,-0.101074107,-0.0152177569,0.0607924461,0.2285785973,0.1035488546,0.1169956252,0.2195221335,-0.218168959,-0.3500105143,-0.0041084914,-0.2088998407,0.3608576655,0.104621999,0.273337543,-0.0505607612,-0.0350845568,-0.3905482292,0.126368627,-0.1349454075,0.1696074903,0.2214268446,0.1292124689,0.1033268347,0.0384701192,0.156261757,0.3508363962,-0.2635574043,-0.2734265327,-0.1393739432,0.0489417799,0.2369983345,-0.184882462,0.1032615304,-0.1323672682,-0.3075629473,-0.2586575747,-0.0560385473,0.2463859767,-0.1726576686,0.2121623158,0.0882713124,0.0533437803,0.0931894556,0.0673698187,-0.0631224588,0.0118180141,0.0031368942,0.1528909653,0.1373866349,-0.1222598553,-0.1124789342,0.102378279,0.0119095612,0.0824260488,0.157824859,-0.2328994125,0.0744623318,0.2115170509,0.2725221813,0.1073635668,-0.4062052369,-0.065164566,0.3674581945,0.3092105091,-0.1567986012,-0.0773328096,0.4437194765,0.0717486739,0.1419773996,0.2614397407,0.1125763655,0.1157530025,-0.2528063059,-0.1444791853,0.1303106546,-0.1338519901,0.3573174179,0.4807224274,-0.1113090888,0.0738127008,0.2318690121,0.2145039439,0.1798753291,0.6407543421,-0.1027310044,0.4042498469,0.1537199616,-0.0888840556,0.2393783778,-0.429338634,0.3233010173,0.0997004211,-0.1209915057,0.1440594792,-0.1215432435,0.2268355638,-0.2595477402,0.0063336147,-0.1651709527,0.034843605,-0.2834734619,-0.0003531646,-0.0922810063,-0.0460077077,0.1485697925,-0.2673061788,-0.1060803607,-0.1838211715,0.2417485863,-0.1152058095,-0.0084537482,-0.2145257741,-0.1257295012,0.1178743392,-0.0035542124,-0.2124055028,0.0751088262,-0.151633054,0.0578426011,-0.1489549428,0.3527969718,0.3654498458,0.5458377004,-0.1088028774,0.0583247282,0.1497375816,-0.0526247025,0.1373092532,0.0994913355,0.1339302361,0.0324912854,0.4296288788,0.1708761305,-0.1730732322,-0.1064104587,0.1563857943,0.3600678742,-0.0998782218,0.060286507,-0.1364115328,-0.0625135377,-0.1864774078,0.281699419,-0.2134115845,-0.1622436345,0.6050928831,0.0269613285,0.3453220129,-0.2474826127,0.1459153295,-0.1132811531,0.4964217842,0.11372035,-0.0273950361,-0.3946330249,-0.3023868799,-0.4063789845,0.1022878066,-0.3766296208,0.0784688964,-0.0880751014,0.0301011186,-0.0370288417,0.0664107651,0.1207996309,-0.2222495228,-0.1465134025,0.2760501802,-0.2253472507,-0.2499590814,-0.3256135583,-0.0829014108,0.3331921101,-0.3161221445,-0.0858484134,0.0034053978,0.1438674033,-0.2619487047,0.0299763381,0.082983993,-0.0884641856,0.5856481791,0.1630411446,0.3099350333,0.0192623604,-0.0643443763,-0.1792136431,-0.1715918183,-0.0905713737,0.2927236259,0.0616841391,0.6076018214,-0.149155125,-0.0012975906,-0.3115739226,-0.1097503081,0.1641416699,-0.0857373476,-0.2654819191,0.2444039285,-0.0181358792,-0.0440478548,0.1009718478,0.2455942929,-0.0819721073,0.5150860548,-0.0081002451,-0.2914106548,0.5103007555,-0.5262163281,-0.350487411,0.1792096496,0.0024521651,-0.0736705363,-0.2278245389,-0.5044870973,0.0955137312,0.3978458047,-0.0757085904,-0.0964354202,0.0544619784,-0.242840156,0.1626874804,-0.0276848171,0.2933384776,-0.2039102465,-0.1067742258,0.2142430246,-0.212530598]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2106","title":"WMT19 Dataset for Kazakh-English is not formatted correctly","comments":"Hi ! Thanks for reporting\r\n\r\nBy looking at the raw `news-commentary-v14.en-kk.tsv` file, it looks like there are at least 17 lines with this issue.\r\nMoreover these issues are not always the same:\r\n- L97 is only `kk` text and must be appended at the end of the `kk` text of the **next** line\r\n- L2897 is only `kk` text and must be appended at the end of the `kk` text of the **previous** line\r\n- L1247 and L1248 are only `kk` texts and must be inserted at the **beginning** of the `kk` text of the next line\r\n- (and there are many others)\r\n\r\nIt would be nice to have a corrected version of this file ! The file is available in the `wmt\/news-commentary` repository on the Datasets Hub here:\r\nhttps:\/\/huggingface.co\/datasets\/wmt\/news-commentary\/tree\/main\/v14\/training\r\n\r\nThen maybe we can notify the WMT authors and host the corrected version somewhere","body":"In addition to the bug of languages being switched from Issue @415, there are incorrect translations in the dataset because the English-Kazakh translations have a one off formatting error.\r\n\r\nThe News Commentary v14 parallel data set for kk-en from http:\/\/www.statmt.org\/wmt19\/translation-task.html has a bug here:\r\n\r\n> Line 94. The Swiss National Bank, for its part, has been battling with the deflationary effects of the franc\u2019s dramatic appreciation over the past few years.\t\u0428\u0432\u0435\u0439\u0446\u0430\u0440\u0438\u044f\u043d\u044b\u04a3 \u04b0\u043b\u0442\u0442\u044b\u049b \u0431\u0430\u043d\u043a\u0456 \u04e9\u0437 \u0442\u0430\u0440\u0430\u043f\u044b\u043d\u0430\u043d, \u0441\u043e\u04a3\u0493\u044b \u0431\u0456\u0440\u043d\u0435\u0448\u0435 \u0436\u044b\u043b \u0456\u0448\u0456\u043d\u0434\u0435 \u0444\u0440\u0430\u043d\u043a \u049b\u04b1\u043d\u044b\u043d\u044b\u04a3 \u049b\u0430\u0442\u0442\u044b \u04e9\u0441\u0443\u0456\u043d\u0456\u04a3 \u0434\u0435\u0444\u043b\u044f\u0446\u0438\u044f\u043b\u044b\u049b \u04d9\u0441\u0435\u0440\u0456\u043c\u0435\u043d \u043a\u04af\u0440\u0435\u0441\u0456\u043f \u043a\u0435\u043b\u0435\u0434\u0456.\r\n> \r\n> Line 95. \u0414\u0435\u0444\u043b\u044f\u0446\u0438\u044f\u043b\u044b\u049b \u043a\u04af\u0448\u0442\u0435\u0440 2008 \u0436\u044b\u043b\u044b \u0442\u0435\u0440\u0435\u04a3 \u0436\u04d9\u043d\u0435 \u04b1\u0437\u0430\u049b\u049b\u0430 \u0441\u043e\u0437\u044b\u043b\u0493\u0430\u043d \u0436\u0430\u04bb\u0430\u043d\u0434\u044b\u049b \u0434\u0430\u0493\u0434\u0430\u0440\u044b\u0441\u049b\u0430 \u0431\u0430\u0439\u043b\u0430\u043d\u044b\u0441\u0442\u044b \u043e\u0440\u044b\u043d \u0430\u043b\u0493\u0430\u043d \u0456\u0440\u0456 \u044d\u043a\u043e\u043d\u043e\u043c\u0438\u043a\u0430\u043b\u044b\u049b \u0436\u04d9\u043d\u0435 \u049b\u0430\u0440\u0436\u044b\u043b\u044b\u049b \u043e\u0440\u044b\u043d \u0430\u043b\u043c\u0430\u0441\u0443\u043b\u0430\u0440\u0434\u044b\u04a3 \u0430\u0440\u049b\u0430\u0441\u044b\u043d\u0434\u0430 \u0431\u043e\u0441\u0430\u0442\u044b\u043b\u0434\u044b. \u0416\u0435\u043a\u0435 \u049b\u0430\u0440\u044b\u0437 \u049b\u0430\u0440\u0430\u0436\u0430\u0442\u044b \u04af\u043b\u0435\u0441\u0456\u043d\u0456\u04a3 \u049b\u044b\u0441\u049b\u0430\u0440\u0443\u044b \u043e\u0440\u0442\u0430\u043b\u044b\u049b \u0431\u0430\u043d\u043a\u0442\u0456\u04a3 \u0440\u0435\u0444\u043b\u044f\u0446\u0438\u044f\u0493\u0430 \u0436\u04b1\u043c\u0441\u0430\u043b\u0493\u0430\u043d \u043a\u04af\u0448-\u0436\u0456\u0433\u0435\u0440\u0456\u043d\u0435 \u0442\u04b1\u0440\u0430\u049b\u0442\u044b \u0441\u043e\u049b\u049b\u0430\u043d \u049b\u0430\u0440\u0441\u044b \u0436\u0435\u043b\u0434\u0435\u0439 \u0431\u043e\u043b\u0434\u044b.\r\n> \r\n> Line 96. The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008. Private deleveraging became a steady headwind to central bank efforts to reflate.\t2009 \u0436\u044b\u043b\u044b, \u0430\u043b\u0434\u044b\u04a3\u0493\u044b \u049b\u0430\u0442\u0430\u0440\u043b\u044b \u044d\u043a\u043e\u043d\u043e\u043c\u0438\u043a\u0430\u043b\u0430\u0440\u0434\u044b\u04a3 \u0448\u0430\u043c\u0430\u043c\u0435\u043d \u04af\u0448\u0442\u0435\u043d \u0431\u0456\u0440\u0456 \u0431\u0430\u0493\u0430\u043d\u044b\u04a3 \u0442\u04e9\u043c\u0435\u043d\u0434\u0435\u0443\u0456\u043d \u043a\u04e9\u0440\u0441\u0435\u0442\u0442\u0456, \u0431\u04b1\u043b \u0441\u043e\u0493\u044b\u0441\u0442\u0430\u043d \u043a\u0435\u0439\u0456\u043d\u0433\u0456 \u0436\u043e\u0493\u0430\u0440\u044b \u0434\u0435\u04a3\u0433\u0435\u0439 \u0431\u043e\u043b\u0434\u044b.\r\n\r\nAs you can see, line 95 has only the Kazakh translation which should be part of line 96. This causes all of the following English-Kazakh translation pairs to be one off rendering ALL of those translations incorrect. This issue was not fixed when the dataset was imported to Huggingface. By running this code \r\n\r\n```\r\nimport datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wmt19', 'kk-en')\r\nfor key in dataset['train']['translation']:\r\n if 'The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008.' in key['kk']:\r\n print(key['en'])\r\n print(key['kk'])\r\n break\r\n```\r\nwe get: \r\n> 2009 \u0436\u044b\u043b\u044b, \u0430\u043b\u0434\u044b\u04a3\u0493\u044b \u049b\u0430\u0442\u0430\u0440\u043b\u044b \u044d\u043a\u043e\u043d\u043e\u043c\u0438\u043a\u0430\u043b\u0430\u0440\u0434\u044b\u04a3 \u0448\u0430\u043c\u0430\u043c\u0435\u043d \u04af\u0448\u0442\u0435\u043d \u0431\u0456\u0440\u0456 \u0431\u0430\u0493\u0430\u043d\u044b\u04a3 \u0442\u04e9\u043c\u0435\u043d\u0434\u0435\u0443\u0456\u043d \u043a\u04e9\u0440\u0441\u0435\u0442\u0442\u0456, \u0431\u04b1\u043b \u0441\u043e\u0493\u044b\u0441\u0442\u0430\u043d \u043a\u0435\u0439\u0456\u043d\u0433\u0456 \u0436\u043e\u0493\u0430\u0440\u044b \u0434\u0435\u04a3\u0433\u0435\u0439 \u0431\u043e\u043b\u0434\u044b.\r\n> The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008. Private deleveraging became a steady headwind to central bank efforts to reflate.\r\n\r\nwhich shows that the issue still persists in the Huggingface dataset. The Kazakh sentence matches up to the next English sentence in the dataset instead of the current one.\r\n\r\nPlease let me know if there's you have any ideas to fix this one-off error from the dataset or if this can be fixed by Huggingface.","comment_length":144,"text":"WMT19 Dataset for Kazakh-English is not formatted correctly\nIn addition to the bug of languages being switched from Issue @415, there are incorrect translations in the dataset because the English-Kazakh translations have a one off formatting error.\r\n\r\nThe News Commentary v14 parallel data set for kk-en from http:\/\/www.statmt.org\/wmt19\/translation-task.html has a bug here:\r\n\r\n> Line 94. The Swiss National Bank, for its part, has been battling with the deflationary effects of the franc\u2019s dramatic appreciation over the past few years.\t\u0428\u0432\u0435\u0439\u0446\u0430\u0440\u0438\u044f\u043d\u044b\u04a3 \u04b0\u043b\u0442\u0442\u044b\u049b \u0431\u0430\u043d\u043a\u0456 \u04e9\u0437 \u0442\u0430\u0440\u0430\u043f\u044b\u043d\u0430\u043d, \u0441\u043e\u04a3\u0493\u044b \u0431\u0456\u0440\u043d\u0435\u0448\u0435 \u0436\u044b\u043b \u0456\u0448\u0456\u043d\u0434\u0435 \u0444\u0440\u0430\u043d\u043a \u049b\u04b1\u043d\u044b\u043d\u044b\u04a3 \u049b\u0430\u0442\u0442\u044b \u04e9\u0441\u0443\u0456\u043d\u0456\u04a3 \u0434\u0435\u0444\u043b\u044f\u0446\u0438\u044f\u043b\u044b\u049b \u04d9\u0441\u0435\u0440\u0456\u043c\u0435\u043d \u043a\u04af\u0440\u0435\u0441\u0456\u043f \u043a\u0435\u043b\u0435\u0434\u0456.\r\n> \r\n> Line 95. \u0414\u0435\u0444\u043b\u044f\u0446\u0438\u044f\u043b\u044b\u049b \u043a\u04af\u0448\u0442\u0435\u0440 2008 \u0436\u044b\u043b\u044b \u0442\u0435\u0440\u0435\u04a3 \u0436\u04d9\u043d\u0435 \u04b1\u0437\u0430\u049b\u049b\u0430 \u0441\u043e\u0437\u044b\u043b\u0493\u0430\u043d \u0436\u0430\u04bb\u0430\u043d\u0434\u044b\u049b \u0434\u0430\u0493\u0434\u0430\u0440\u044b\u0441\u049b\u0430 \u0431\u0430\u0439\u043b\u0430\u043d\u044b\u0441\u0442\u044b \u043e\u0440\u044b\u043d \u0430\u043b\u0493\u0430\u043d \u0456\u0440\u0456 \u044d\u043a\u043e\u043d\u043e\u043c\u0438\u043a\u0430\u043b\u044b\u049b \u0436\u04d9\u043d\u0435 \u049b\u0430\u0440\u0436\u044b\u043b\u044b\u049b \u043e\u0440\u044b\u043d \u0430\u043b\u043c\u0430\u0441\u0443\u043b\u0430\u0440\u0434\u044b\u04a3 \u0430\u0440\u049b\u0430\u0441\u044b\u043d\u0434\u0430 \u0431\u043e\u0441\u0430\u0442\u044b\u043b\u0434\u044b. \u0416\u0435\u043a\u0435 \u049b\u0430\u0440\u044b\u0437 \u049b\u0430\u0440\u0430\u0436\u0430\u0442\u044b \u04af\u043b\u0435\u0441\u0456\u043d\u0456\u04a3 \u049b\u044b\u0441\u049b\u0430\u0440\u0443\u044b \u043e\u0440\u0442\u0430\u043b\u044b\u049b \u0431\u0430\u043d\u043a\u0442\u0456\u04a3 \u0440\u0435\u0444\u043b\u044f\u0446\u0438\u044f\u0493\u0430 \u0436\u04b1\u043c\u0441\u0430\u043b\u0493\u0430\u043d \u043a\u04af\u0448-\u0436\u0456\u0433\u0435\u0440\u0456\u043d\u0435 \u0442\u04b1\u0440\u0430\u049b\u0442\u044b \u0441\u043e\u049b\u049b\u0430\u043d \u049b\u0430\u0440\u0441\u044b \u0436\u0435\u043b\u0434\u0435\u0439 \u0431\u043e\u043b\u0434\u044b.\r\n> \r\n> Line 96. The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008. Private deleveraging became a steady headwind to central bank efforts to reflate.\t2009 \u0436\u044b\u043b\u044b, \u0430\u043b\u0434\u044b\u04a3\u0493\u044b \u049b\u0430\u0442\u0430\u0440\u043b\u044b \u044d\u043a\u043e\u043d\u043e\u043c\u0438\u043a\u0430\u043b\u0430\u0440\u0434\u044b\u04a3 \u0448\u0430\u043c\u0430\u043c\u0435\u043d \u04af\u0448\u0442\u0435\u043d \u0431\u0456\u0440\u0456 \u0431\u0430\u0493\u0430\u043d\u044b\u04a3 \u0442\u04e9\u043c\u0435\u043d\u0434\u0435\u0443\u0456\u043d \u043a\u04e9\u0440\u0441\u0435\u0442\u0442\u0456, \u0431\u04b1\u043b \u0441\u043e\u0493\u044b\u0441\u0442\u0430\u043d \u043a\u0435\u0439\u0456\u043d\u0433\u0456 \u0436\u043e\u0493\u0430\u0440\u044b \u0434\u0435\u04a3\u0433\u0435\u0439 \u0431\u043e\u043b\u0434\u044b.\r\n\r\nAs you can see, line 95 has only the Kazakh translation which should be part of line 96. This causes all of the following English-Kazakh translation pairs to be one off rendering ALL of those translations incorrect. This issue was not fixed when the dataset was imported to Huggingface. By running this code \r\n\r\n```\r\nimport datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('wmt19', 'kk-en')\r\nfor key in dataset['train']['translation']:\r\n if 'The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008.' in key['kk']:\r\n print(key['en'])\r\n print(key['kk'])\r\n break\r\n```\r\nwe get: \r\n> 2009 \u0436\u044b\u043b\u044b, \u0430\u043b\u0434\u044b\u04a3\u0493\u044b \u049b\u0430\u0442\u0430\u0440\u043b\u044b \u044d\u043a\u043e\u043d\u043e\u043c\u0438\u043a\u0430\u043b\u0430\u0440\u0434\u044b\u04a3 \u0448\u0430\u043c\u0430\u043c\u0435\u043d \u04af\u0448\u0442\u0435\u043d \u0431\u0456\u0440\u0456 \u0431\u0430\u0493\u0430\u043d\u044b\u04a3 \u0442\u04e9\u043c\u0435\u043d\u0434\u0435\u0443\u0456\u043d \u043a\u04e9\u0440\u0441\u0435\u0442\u0442\u0456, \u0431\u04b1\u043b \u0441\u043e\u0493\u044b\u0441\u0442\u0430\u043d \u043a\u0435\u0439\u0456\u043d\u0433\u0456 \u0436\u043e\u0493\u0430\u0440\u044b \u0434\u0435\u04a3\u0433\u0435\u0439 \u0431\u043e\u043b\u0434\u044b.\r\n> The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008. Private deleveraging became a steady headwind to central bank efforts to reflate.\r\n\r\nwhich shows that the issue still persists in the Huggingface dataset. The Kazakh sentence matches up to the next English sentence in the dataset instead of the current one.\r\n\r\nPlease let me know if there's you have any ideas to fix this one-off error from the dataset or if this can be fixed by Huggingface.\nHi ! Thanks for reporting\r\n\r\nBy looking at the raw `news-commentary-v14.en-kk.tsv` file, it looks like there are at least 17 lines with this issue.\r\nMoreover these issues are not always the same:\r\n- L97 is only `kk` text and must be appended at the end of the `kk` text of the **next** line\r\n- L2897 is only `kk` text and must be appended at the end of the `kk` text of the **previous** line\r\n- L1247 and L1248 are only `kk` texts and must be inserted at the **beginning** of the `kk` text of the next line\r\n- (and there are many others)\r\n\r\nIt would be nice to have a corrected version of this file ! The file is available in the `wmt\/news-commentary` repository on the Datasets Hub here:\r\nhttps:\/\/huggingface.co\/datasets\/wmt\/news-commentary\/tree\/main\/v14\/training\r\n\r\nThen maybe we can notify the WMT authors and host the corrected version somewhere","embeddings":[-0.0953304172,-0.5491961241,-0.0432180017,0.3162058294,-0.0860267431,0.0107725346,0.1837910563,0.1722095013,-0.1140662581,0.2022086978,0.1289513409,0.0987930223,0.111726664,0.5228421092,0.0502476022,-0.2106284201,0.1143199801,-0.1941502392,-0.2592694759,-0.1341278553,-0.0660601184,0.1056311578,-0.4930419326,-0.0548006594,-0.4117452502,-0.1151644886,-0.1502326727,-0.0785620958,-0.0658805519,-0.1956119835,0.0703047737,0.1335619986,-0.1053127944,0.3510219157,-0.0001112676,-0.2417628765,0.021564262,-0.0297251698,-0.1629369408,0.0762894899,-0.1257774383,-0.8057081103,-0.2112897336,-0.120443292,-0.1037973389,-0.4211285114,-0.1573449224,0.0498358682,0.3065554798,0.1048358008,0.1853925437,0.2881683111,-0.0026125675,-0.1187297106,-0.1859125048,-0.0203668773,-0.1585624367,0.0198849589,0.1100702509,0.1526906639,-0.1194509044,0.6587665081,0.0118426559,-0.1108202636,-0.0299412534,-0.1155029833,-0.1476062089,0.0271584671,0.0776109397,0.4503267705,0.4413164258,0.017042812,-0.0518957898,-0.1849430799,0.0404484235,-0.3243750036,0.0702300817,0.1509529352,-0.2087913752,0.1687709689,0.0222084727,0.3100071847,0.0986339226,0.0540111065,-0.2007593811,0.467933476,-0.1476940811,0.0270631034,0.0542254187,-0.0414870754,-0.1118217632,-0.0609524846,-0.0135244671,-0.1418372095,-0.1481644362,-0.0013134108,-0.0952405259,-0.2198383361,0.1539940387,-0.5207996368,-0.1159045249,0.355996877,-0.2307448834,-0.177794233,0.192759797,0.5753376484,0.0576573908,-0.3404280543,0.1400430948,0.1638718247,0.0051132953,-0.0042761825,0.0858228281,-0.4291135669,-0.2851946354,0.1375285685,-0.0554422997,-0.0388974287,-0.0930311009,0.1223476455,0.1957934797,-0.1440604776,-0.2310918123,0.0197817087,-0.1138918549,0.4291629791,-0.1612132788,0.0644564778,-0.1776590496,-0.169715777,-0.1797194481,-0.1399149001,-0.3185177445,-0.0326874964,-0.1918263137,0.1269253343,0.2599645555,0.2328584194,0.0105895195,-0.081656985,-0.3358832598,-0.0744915381,0.4750349224,0.2353508472,-0.109406732,0.1894775331,-0.0672534183,0.0855240226,0.001310652,0.0184608791,-0.3052257597,0.2370880991,-0.0127146421,0.1222734153,-0.128480792,-0.0958907977,-0.2741920352,0.5493567586,0.0236123577,-0.0164982509,0.1953630298,-0.1072928831,-0.3907093108,-0.1250798106,0.2603446841,0.1920642853,-0.3559532166,-0.0284667704,0.1010944322,-0.0517998189,0.4939495623,0.4126811922,0.1056014895,0.1440576911,-0.0203153882,0.1787227392,-0.0787365213,0.0409197286,-0.3128335774,0.0134839611,-0.2729091942,0.4639530182,-0.0156172393,0.0288684107,0.0604276918,0.2915114164,0.1621890217,0.0457521379,0.0671483427,0.2363936752,-0.2866164744,-0.3670729399,0.6842613816,-0.1920403242,0.1620819867,-0.3465865254,0.0237741433,-0.1346558183,0.5904834867,0.2387703955,0.1202825904,-0.1334968954,-0.0776724368,0.0483449437,0.2563450634,0.152576074,-0.2615077794,0.206906423,0.199422285,0.4913658798,0.2391413897,0.1307504177,-0.1787830442,-0.1840566248,-0.2560290694,-0.3216755092,0.213231191,0.0867892504,-0.4696191549,0.3506502509,-0.0635577142,-0.0447053984,-0.0672241002,-0.026002178,-0.2125884295,-0.0297508147,-0.0089499792,-0.02854947,-0.1508794725,0.2967285514,0.2445807904,0.1812208593,-0.2409955561,0.0913004875,0.0638272762,0.6336416602,-0.1121875346,-0.2415217608,0.2419113517,-0.7164649963,0.035252668,0.118126601,-0.0593292303,-0.0951408893,0.1239964887,0.43100667,-0.0151632149,0.0634539649,-0.1755182892,0.0894917697,0.2513158023,-0.0591865927,-0.0728715137,-0.2587556541,0.1687140316,-0.2381840944,-0.2176354825,-0.0330536291,-0.143403545,0.3936029971,0.5164348483,0.2517271936,0.226430878,-0.103132844,-0.4107159078,-0.1604110152,0.3795114756,0.2175553739,-0.0869412795,0.0945360139,0.172304675,0.2995283604,0.1139071137,-0.1324412972,0.1295514703,0.0206169207,0.0954602361,0.2805342674,-0.1022091731,0.020655809,-0.2630800605,0.518856883,-0.1302155107,0.0445324443,-0.1824714243,-0.1719665676,-0.4352489412,-0.086605221,-0.5954187512,-0.3802389503,-0.1560970545,-0.4191438556,0.0219246428,-0.2171105295,-0.2821983993,0.2730413079,0.5708176494,0.1382266283,-0.3288728595,0.6834566593,-0.3291601241,-0.5311251879,-0.1904894859,0.0764606073,0.233076781,0.5193153024,0.1068043858,-0.4199673235,-0.1401586831,-0.057606373,-0.6339119077,0.1419069469,-0.369179219,-0.0652037039,-0.0582939722,0.319121182,-0.251332283,-0.1517112255,0.2821587324,-0.1380905062,-0.0785943791,0.374920845,-0.204396829,-0.1959827691,-0.1805733442,-0.440983057,-0.3408245146,-0.1562552303,0.2199539244,-0.3337595165,-0.2653497756,0.4110278487,0.235519737,0.0143299084,-0.3028132915,0.1374218166,-0.2295391858,-0.12518242,0.0368715338,-0.0791391358,-0.487685591,0.1958709657,0.1955722123,-0.1193501353,-0.0972853601,-0.1942091733,0.4939127564,-0.006794889,-0.1669482887,0.184014231,0.1353168041,0.0118209524,-0.0623589419,-0.0028104091,-0.1753420234,-0.236891821,-0.0616775267,-0.0132810017,0.1650728434,-0.2205608785,0.2217827439,0.3598546982,0.2768031061,0.1467605531,0.2321696281,-0.078348428,-0.0849826857,0.3804539442,-0.007025497,-0.1966290176,-0.0072077503,-0.2464385033,0.098762624,0.5342597961,-0.0362028144,0.3417963982,-0.0959867835,-0.4318201244,-0.3260053098,-0.3005723953,0.1634156257,0.1365791857,0.2203222513,0.1938868016,0.0029765731,-0.0121426322,-0.2541388571,0.011944632,0.1817479134,-0.323327899,0.1496138126,-0.6493254304,-0.3908016384,-0.2101157606,0.1783205122,0.1908468008,0.5266017914,0.0216933191,-0.0867909789,0.0147217428,0.0041494905,0.1225298643,0.1175092012,-0.0526790582,0.0036357676,-0.1044451818,-0.2852303386,-0.1901822537,-0.1888112277,-0.2742128074,-0.1602444351,0.3807337284,-0.0178262182,-0.0537962727,-0.0340603106,0.5762078166,-0.4829653502,0.0547399297,-0.3097638786,-0.2468164563,-0.0778804794,0.1056276932,-0.0828282386,-0.2035560161,-0.2467963099,0.1738994271,0.3307301104,0.2122389674,-0.0511813425,0.2393713295,0.427906245,0.1697327942,-0.0459734537,0.3436512351,-0.1851816028,0.5676724315,0.3259210587,-0.1027503461,-0.016861489,-0.4144355953,-0.4195944667,0.4677900076,0.4874901175,-0.1403456628,0.2592273653,0.5915626884,-0.0070070955,-0.2882346511,-0.0628014058,0.3577047586,0.4028760791,-0.2360128313,-0.0368832722,0.1228772178,-0.1377454549,0.1044120863,0.0924254805,0.1091301665,-0.162164554,0.1848955005,-0.0433544032,0.8865994811,0.2256827354,0.2615555823,0.3602926135,0.0330766812,0.0739064515,0.1385929137,-0.1617923677,-0.0941575617,0.0045830109,-0.0466916189,0.2219844162,0.1527676731,-0.0732110888,-0.3241305947,0.2243206799,0.1023738533,0.5825098753,-0.1413495392,-0.1016587839,0.2941732407,-0.0617820136,-0.2273468524,0.0887745023,-0.141603753,0.0437638313,-0.1233864352,-0.184910357,-0.1014074013,-0.150116995,-0.3251180053,-0.0241004787,-0.2945806086,-0.103450641,-0.2135144919,0.0179530643,-0.0029566404,0.0186119322,0.4101046622,0.2033897638,-0.0546247177,-0.0376557373,0.2404320389,0.2063178867,0.0148550766,-0.0065987571,0.2028627247,-0.2263789773,-0.253356725,0.0646378472,-0.1127078459,0.236069411,-0.0339526236,-0.0533278026,-0.0063041607,0.2633965313,0.0496417694,0.3528279662,0.1620622724,-0.4291361272,0.1087315679,-0.0865190029,-0.1663240045,0.3700785935,0.2234922796,-0.2619774938,-0.1754823923,0.067255713,0.2198856622,0.1549966037,0.3292730451,0.3228929639,-0.158947885,-0.2308558375,0.164191857,0.1783352047,-0.3057085872,0.0498165004,-0.1226806343,0.014501567,-0.2461696118,0.7356243134,0.2438765466,-0.3041029572,-0.1008895561,-0.222203359,-0.2246764153,0.464839071,0.2148206383,0.0223167799,0.2717097402,0.3222668469,0.0277983118,0.1568289846,-0.3618489206,0.0553009547,0.1475099921,0.0892981216,-0.2331241667,0.1235947534,0.0784094855,-0.0459873267,-0.0491054207,0.049022451,-0.1368985027,-0.17570135,-0.2522899806,0.0703393966,0.1060350984,-0.2024833709,0.1298444271,-0.1209271401,-0.0102067282,-0.2606371343,-0.0391152389,0.3032339811,0.0214054193,0.6562044621,0.5113461018,0.3748404086,-0.1676584333,0.0306372382,0.1414141953,0.0408301763,-0.021517273,-0.0742199495,0.0576404221,-0.014035794,0.0181796737,0.0019829571,-0.0701417103,0.3109591603,0.0314572044,-0.0569070652,0.1499098837,-0.0756302699,0.2567427158,0.2328177541,-0.0061737513,-0.1368312538,0.133467719,0.1260898262,-0.355332613,0.069587335,0.2315454632,-0.2497907281,0.224711746,-0.054707706,0.6976215243,-0.1712837517,0.1296428293,-0.0597572811,-0.161115095,-0.2872351408,0.2644671798,0.1709642857,-0.0930198655,0.2680008411,-0.0916336924,-0.0525607355,-0.1002386063,0.0008199283,0.1894943714,0.3614407182,0.3250834942,-0.0061286972,-0.0762201846,-0.1764838994,0.3923094571,0.2627929449,-0.5786646008,-0.0118805887,0.3152037859,0.400391072,0.0261822697,-0.2331154197,-0.1960289478,0.1454222202,-0.4816763699,0.0971689597,-0.0988106355,-0.1527324766,-0.1637048274,-0.1387459785,-0.1691412777,0.1189202443,0.2113560736,-0.0596017949,-0.2362720221,0.0646042228,-0.3638755679,-0.0041967114,0.1547768116,-0.2315104157,0.1744867116,0.2585039139,-0.0560654327,-0.0216645468,0.4135409594,0.1456403583,0.2127303779,0.0205730256,-0.1573348939,0.2102944255,0.0312328581,0.0450182632,0.2843779325,0.2946208119,-0.0731143206,0.162848264,0.1427555084,-0.245122239,-0.1093585566,0.4080251157,-0.0703597739,-0.1483275294,-0.3365430534,-0.2267832607,0.0103149554,-0.1661456078,0.0106602656,-0.062347468,0.0447607338,0.4490066171,-0.0746461079,0.0888533443,-0.1517069191,0.1223294735,0.2219154388,0.6906311512,0.2963452637,-0.164827168,-0.008998855,-0.3020532429,-0.7928386927,0.2897747457,-0.08511886,0.4469282925,-0.0866163746,-0.1200456992,0.3575143516,0.2640541196,0.2136468291,0.2724082768,-0.1845104247,-0.1393782943,-0.4809322059,0.0902025327,0.3914111555,-0.0330011882,-0.0831682384,-0.1077030674,0.3051948249,-0.1150867343,0.0880975798,-0.0039884886,0.0451831892,-0.309897989,0.2056404203,0.2238478512,0.0749139935,0.3904040754,-0.0748715028,-0.1520861089,-0.1948975176,-0.1183006316,-0.013814955,0.3537481129,0.0109863207,0.2306621671,0.1819503307,-0.0099182315,0.0150745679,-0.0097376043,0.1971980333,-0.0832658485,-0.3364747167,0.2710092366,0.1983929724,-0.2360698581,0.1953923553,0.5181953311,0.136991486,-0.0891261399,-0.2669430673,-0.2945525348,0.1005225182,-0.1692585945,-0.2758888304,-0.2625384331,-0.1398347467,0.333219856,0.0444502123,-0.1931958795,0.0043044253,0.0915735736,0.3010489643,0.0251115598,0.1065682396,-0.2855907083,-0.1350480616,0.0577145219,0.2076470852,0.3584672511,-0.2832902968,-0.1735757589,-0.0966014713]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2105","title":"Request to remove S2ORC dataset","comments":"Hello @kyleclo! Currently, we are getting the data from your bucket, so if you remove it the HF script won't work anymore :) \r\n\r\nUntil you solve things on your end, @lhoestq suggested we just return a warning message when people try to load that dataset from HF. What would you like it to say?","body":"Hi! I was wondering if it's possible to remove [S2ORC](https:\/\/huggingface.co\/datasets\/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!","comment_length":54,"text":"Request to remove S2ORC dataset\nHi! I was wondering if it's possible to remove [S2ORC](https:\/\/huggingface.co\/datasets\/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!\nHello @kyleclo! Currently, we are getting the data from your bucket, so if you remove it the HF script won't work anymore :) \r\n\r\nUntil you solve things on your end, @lhoestq suggested we just return a warning message when people try to load that dataset from HF. What would you like it to say?","embeddings":[0.2497458011,-0.3549620509,-0.0154878516,0.2169867009,0.0509789921,-0.0599792711,0.0153631736,0.1701468229,0.3005094528,0.0649961978,-0.3151451349,-0.2419163585,-0.397151947,0.367079407,-0.2082139105,0.1481842548,-0.1210899502,0.0473451391,-0.2688441575,-0.0216256119,0.1051725298,0.2462723106,-0.0533281751,0.3302411735,-0.2204811424,-0.1824883968,0.0003759059,0.2880859673,-0.168196097,-0.3303515613,0.3342998028,0.3446072936,0.1943356246,0.4782934189,-0.0001191036,-0.2259052545,-0.0449481197,0.0634203553,-0.336304158,-0.0503856689,-0.0639070496,0.0830901861,-0.0390450619,0.0706025958,-0.1767835617,-0.1328932345,0.0946238115,-0.1127751172,0.45149526,0.2237595469,0.1472906172,0.5189114809,-0.2679886222,0.235634774,0.0838877112,0.5844472647,-0.0968283489,0.3613194525,-0.1507354081,-0.1344151646,0.0206344444,0.3544667661,-0.4820709229,-0.0808606297,0.3423905373,-0.2815232277,-0.3463054299,-0.2894932926,0.1905046552,0.3609244525,0.5013285875,-0.1849517077,-0.008570591,0.032625623,0.0586836264,0.0704151616,0.093675971,0.1089070514,0.0631737262,0.453869462,-0.2132694721,-0.5096263289,-0.2386959493,-0.1701668203,-0.191897139,0.2741118371,-0.2331748307,-0.0442752056,0.0756487027,-0.0128305592,0.1244937256,-0.1404350102,-0.1270954162,0.2118843794,-0.1232205629,-0.3587535918,-0.1462815851,0.097842671,0.3382321596,-0.0054843603,-0.033905372,-0.1031496674,-0.3579530716,0.1639176607,0.2696815729,-0.106899187,0.1650609672,0.080244422,0.2007359564,0.3710959852,0.2687785029,0.1058006957,0.2009716928,0.1437756717,-0.2677750289,0.0206930619,0.4212612212,-0.4505495429,-0.4954606593,0.3764736354,-0.0396209918,-0.125484556,0.1502397209,0.2772763073,0.1018729731,-0.0195235908,0.1454119831,0.116527833,-0.1844519079,-0.3579524755,-0.1602798849,0.2095674723,-0.251229465,0.1203959957,0.1802686155,-0.1312956214,0.1287759393,0.1143668145,0.6399711967,0.2406097054,-0.1666058302,-0.1466241181,0.2603840828,0.1203874052,-0.1811207086,0.0802044198,0.1370403469,0.0961927846,-0.2488334775,0.0890592933,-0.3099267483,-0.2802248597,-0.516005218,0.0420542955,-0.1513456851,0.0481919311,-0.3567976356,-0.051567588,-0.3598996103,0.1025721878,0.2436718196,0.1525050253,-0.4501981139,-0.0034856186,-0.1981196105,0.5347214937,-0.066969648,-0.0042566424,0.3785064518,-0.3493422866,0.0128762685,0.404050231,-0.2383314222,-0.253508985,-0.4391379058,-0.1414205581,-0.5946813822,-0.1744572818,0.0992880538,0.5375522971,-0.2992700934,0.2132284194,0.0099565657,-0.2031260878,0.0113221304,-0.5564859509,-0.0973526761,-0.0895495191,0.2681993246,-0.2259251326,-0.2278825641,-0.3242280185,0.0612322576,-0.1026290804,0.1750382781,0.3118291497,0.0767643154,-0.0738639906,0.6009222269,0.0521623716,0.0797494277,0.3716990948,0.2853856087,0.2455003411,-0.2253010124,0.0560515188,-0.1960527152,-0.0647549108,-0.0250650346,-0.0749502406,0.1551264226,-0.2883827984,-0.1902813762,0.0183652174,0.027183542,0.1615642458,-0.0629756227,-0.3182079494,-0.0179348737,-0.111243628,-0.3682821989,0.6763768792,-0.1507049054,0.1577686369,-0.1715857834,0.1568666399,0.0588740036,0.280924648,0.2297971547,0.0875786468,0.2233127654,-0.0150029752,0.1429278851,0.2585816085,-0.1537299752,0.0822614133,0.2234179676,0.4290006459,0.1444981545,-0.1048828661,0.4809744954,-0.082741186,-0.0540455692,0.2035284787,0.087127015,0.2614385188,-0.3357444704,0.0526356287,-0.3771407902,-0.1790440381,0.1647438109,-0.0615302585,-0.1390989125,0.0128756389,0.1143247783,0.1538396627,-0.2277926952,-0.0845726058,-0.3620781302,-0.1090864539,-0.1644118428,-0.2193997502,-0.2350513041,0.2299925387,-0.0151846111,-0.0578251369,0.4176206291,-0.0115158409,0.2619350255,0.128081888,-0.1451253593,0.2411237061,0.3163654208,-0.0746850371,0.2493267804,0.0866191313,-0.0355136953,-0.0647578165,0.1031338647,0.0849600285,-0.3730011284,-0.2814788222,-0.3263089061,-0.0815616027,-0.3990690708,0.1705252379,-0.1877942681,-0.3982465565,0.031280417,0.4227204025,-0.6007093191,0.0725847334,0.3227221966,0.2935101986,-0.1004199386,-0.0541537106,-0.290679872,0.3760780692,-0.3606549501,0.0042782482,-0.157555148,0.155975014,-0.0230977293,0.048961062,-0.059231896,-0.0755986795,0.591833353,0.0483928882,-0.1953922212,-0.2420415431,-0.0286112633,0.2571062744,0.0222353246,-0.2735895216,0.3110335767,0.4664096534,-0.1958955973,0.0888095424,-0.230655998,-0.3152903318,-0.0752920508,-0.0221924447,-0.0101125613,0.1665061712,-0.1660592556,0.3509397209,-0.1194058433,-0.1206385791,-0.1005800813,-0.2498231232,0.1017661914,0.1561323553,-0.2423310131,-0.0377894007,-0.4389007986,-0.2412038893,-0.4783966541,-0.7775785327,0.1322237998,0.1551061273,0.0619400218,0.3001137376,0.2907038629,0.0243347287,-0.2453390807,-0.4676840603,-0.2666758597,0.1108360961,0.0299536344,-0.0711925402,-0.0387910604,0.1316688657,-0.0325359255,-0.0444163457,0.020417789,-0.1365660727,-0.2444277853,-0.0075051161,0.2517967522,0.0325369388,0.1261018366,0.0227919519,0.4973379076,0.1448776573,0.0841678232,0.4559576213,0.1717983335,0.3662770689,0.0237877294,-0.2425237447,0.1033528522,0.1416413337,-0.300575316,0.2440726906,-0.0383811072,0.3073283732,-0.0081218015,-0.2871597707,-0.0282567851,-0.3399540484,0.009104698,0.1280028224,0.2055031508,0.0248851981,-0.2929002941,0.2974544466,-0.4071058035,0.2614758909,0.2595376074,0.540958941,-0.0682669133,-0.3393146694,0.0467197411,-0.6888621449,0.2979674935,-0.2771045864,-0.2801989019,-0.2462520301,0.1725996882,0.2283547819,0.2470279336,0.4986201525,0.2680114508,0.0235360507,-0.2184677124,-0.1825603098,-0.0582053103,-0.0362150967,-0.1056762338,0.1881690472,-0.1052639186,0.050042741,-0.3504535854,-0.531547904,0.2567065954,-0.0225877892,-0.1742413193,-0.2361788601,-0.2920235991,-0.1977845877,-0.3408434391,0.1677184254,0.3076161742,0.0153891062,-0.1253123581,0.1018319502,0.4349921942,0.1473050863,0.6156095862,-0.21141994,0.0193331968,0.343941927,-0.1309429854,0.1688972414,0.3511146009,0.4098879993,0.5923307538,0.1420724243,-0.0819057971,0.0140374117,0.1726201028,0.3772753477,0.5314565301,0.1444595009,0.2339477837,0.1110389084,0.5215489268,-0.4326654375,-0.0630815625,0.1206269488,-0.3773059547,-0.2497067899,-0.098996453,0.2030668557,-0.0780609846,-0.1188460141,-0.1410823017,0.8098531961,-0.1867140979,0.1559916139,0.1509022564,1.332015872,-0.2479284555,0.2357518673,0.271295011,-0.1290401816,0.4322173297,-0.3908835948,-0.1765439957,-0.2118513435,-0.0162080266,-0.1631900817,0.1025287881,-0.0620009229,0.1435307264,0.2626087368,0.3281172812,-0.4982745647,0.1290330142,-0.2335584015,0.2766572535,-0.1408673376,0.0030516183,-0.3498929441,0.0744982734,-0.1713302583,0.4168682098,-0.2191191614,-0.067167066,0.4257998466,-0.3778281212,-0.0424639396,0.0447319373,0.0388370268,-0.1559119374,-0.1112904996,0.0425097384,0.3011127114,-0.0560352057,0.3512803614,0.0278364681,-0.0726782531,0.0513901375,-0.1709363312,-0.0841918886,-0.295290947,0.0870442986,-0.2191108316,0.0486374199,-0.2608413696,0.1357269585,0.2565948963,-0.0180955324,-0.2227424532,-0.0412247851,-0.0989614427,-0.3706776798,0.031881012,0.1167126745,0.2910928428,0.1557995826,0.0564191155,0.3169148564,-0.0138250431,-0.3145059645,0.3934530914,-0.2551909685,-0.2107522935,0.2886604667,-0.027212372,-0.2627367079,0.3052179217,0.1696693599,-0.2657026947,-0.3482819498,-0.0754254088,0.0589205213,-0.597358942,0.0254177898,0.4354972839,-0.1620435417,-0.2526147962,-0.2109771371,0.0568803065,-0.0889867172,0.384884268,-0.0449821576,-0.1926767528,0.2018601745,-0.2169387937,0.0482340008,0.2139462829,0.0933070853,-0.3710498512,-0.1320239305,-0.2644646764,0.0163206868,0.1053566858,-0.2173151523,0.1335785538,0.2935497463,0.2910032868,-0.2786352038,-0.020313058,-0.0993207172,-0.1770353764,-0.0748839155,-0.2910591066,0.2215994447,-0.0449011363,-0.1685598493,-0.1221850365,-0.1883231252,0.0235312227,-0.0048136264,0.4513418972,0.743885994,0.2988628149,0.2216386944,-0.0300446842,-0.2597450018,-0.3106904924,0.0058656801,0.2851530015,0.4316033125,0.2342333049,0.0356953703,-0.1349045187,-0.1548625827,-0.0775108635,0.4667083323,0.3720102012,0.1954264343,0.0320643634,0.0422285497,-0.1331796944,-0.0589349009,0.28704229,-0.1698836982,-0.2773601711,-0.5174636245,0.091657795,0.0997790396,-0.2205557972,0.002080709,0.1206341833,0.0996044725,-0.2272837162,0.2557863295,0.032835044,0.298979342,0.2343001813,0.0497594997,0.1036291122,0.0457914732,0.0056177913,-0.1631210148,-0.3256462812,-0.1176439226,0.3044127524,0.1418196261,0.2957925797,-0.1029091179,-0.0012434345,0.26205495,0.1100715771,0.343667984,0.0641127676,-0.4404395521,-0.1973462105,0.4140305519,0.0036721949,-0.0359741449,0.2168565691,0.1978261173,0.0704155341,-0.336260438,-0.2519579828,0.3827542961,-0.0171566028,0.1619794518,0.1494685709,0.0006607726,-0.2370759994,0.0316930935,0.0101732276,-0.1894216985,0.2951090336,0.0410579778,-0.1641430259,-0.0630541891,-0.0407240875,0.2460594177,0.1142049581,0.0809093341,-0.1802521199,0.1558296233,0.1466406137,0.4085492194,0.3081877828,0.2187095433,-0.1930949241,0.3625764847,-0.0345162898,-0.1915352196,0.1151141301,0.1295537055,0.1988428533,-0.1566426307,0.1549256295,0.0691784546,0.0344628841,0.0099707618,-0.0998148546,-0.0111804726,0.2621310949,-0.2802784443,0.3153545558,0.1543421447,0.1042039692,-0.1866542995,-0.0157015305,-0.1792559773,0.0357771926,0.2729771733,0.0198935736,0.2161611468,-0.0756981,0.0229734518,-0.1112567857,0.0774394348,0.5089768171,0.0148963528,-0.3831487894,-0.0282359179,-0.3676782548,0.1642835289,-0.1747779995,-0.0107600056,-0.0793027282,-0.1445980668,0.2598772049,0.0191258211,0.0137535501,-0.4540669024,-0.0559215806,-0.1435645968,0.0621939488,0.2941066623,-0.2662774622,0.3607119024,0.0232319776,-0.0668496341,0.0598656312,0.0297874436,-0.1144575924,-0.2641505897,0.3247030675,-0.1846611351,-0.4762776196,0.2907908261,0.1561374515,0.0620757528,-0.1317415684,-0.3518446982,0.1171885803,-0.1519410759,-0.0625785738,-0.0601479076,0.2416973412,0.1981606036,0.0042981561,-0.1956638694,-0.0065106847,0.5049579144,-0.235705331,0.3362917304,-0.4627580345,0.0505446605,-0.2155090272,0.2378557622,-0.2511630952,0.2846774459,0.2050163597,0.0858608857,-0.2218781114,-0.0377835892,0.8010756373,-0.2489505559,-0.0627345666,0.023047179,0.7974456549,-0.1255705208,-0.5456662774,-0.6729075313,0.3868610263,-0.0742243081,0.0923881531,-0.0214735642,0.3568215668,-0.2921844125,0.0004403491,0.1347307861,0.3854475617,0.1125352979,-0.1136495247,0.2185108513,-0.0870212764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2105","title":"Request to remove S2ORC dataset","comments":"Hi @kyleclo, as of today, you have not removed your bucket data yet, and therefore HuggingFace can download it from there.\r\n\r\nIs it OK? Are you planning to eventually delete it? Thank you.","body":"Hi! I was wondering if it's possible to remove [S2ORC](https:\/\/huggingface.co\/datasets\/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!","comment_length":33,"text":"Request to remove S2ORC dataset\nHi! I was wondering if it's possible to remove [S2ORC](https:\/\/huggingface.co\/datasets\/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!\nHi @kyleclo, as of today, you have not removed your bucket data yet, and therefore HuggingFace can download it from there.\r\n\r\nIs it OK? Are you planning to eventually delete it? Thank you.","embeddings":[0.4957595468,-0.3337317109,-0.0840625614,0.5156832337,0.0438568369,-0.1015334204,-0.03269355,0.0329050347,-0.0669208765,0.035270609,-0.4459505677,-0.1573467106,-0.4066594243,0.3072044253,-0.2146405876,0.1066135839,-0.0704353452,-0.0257814061,-0.2145421058,0.1067164019,0.2157327384,0.2247689664,-0.1654158831,0.053916961,-0.210423559,-0.238173902,-0.1167775095,0.1302072853,-0.3192273676,-0.0353856906,0.3552872539,0.2568558455,0.0180798657,0.5105392337,-0.0001207894,-0.249954164,-0.1819291413,0.0273429621,-0.3482283354,-0.0850831196,-0.2539300025,0.248618573,-0.0800378323,0.0494216532,-0.0200393386,-0.089449957,0.0677072182,-0.0824018642,0.4551747441,0.3425408602,0.1748303622,0.4669805467,-0.2167333961,0.3688213825,0.046474494,0.5212647915,-0.1069017053,0.4482254088,-0.0773000717,0.117787689,0.2876632214,0.2670897245,-0.4529399872,0.0094581647,0.1681184918,-0.1748993844,-0.2927419543,-0.3420933187,0.4042110443,0.2000582963,0.55678159,-0.1356282085,-0.0795357823,0.17098777,0.1103685945,0.0262254383,0.1101982296,0.336453855,0.1220411211,0.4964210391,-0.4470813274,-0.6174142957,-0.1987923831,-0.1457072645,-0.3737245202,0.4520031214,-0.4058312178,-0.0916310027,-0.1052423492,0.0417878181,0.1657261848,-0.2152705044,-0.0421459191,0.1711713225,-0.1479821205,-0.1933970302,-0.2669911087,0.3302889466,0.0193494447,0.1699258983,0.1828711182,-0.2607261837,-0.2533019781,0.1017718837,0.2911595702,-0.0237625558,0.0344303846,0.1623583287,0.1257816851,0.1631886661,0.3278103471,0.1640237719,0.1142652929,0.3418657482,-0.4561470449,-0.1095793694,0.5117805004,-0.4140556157,-0.398042053,0.2177031934,-0.1894532144,-0.056728512,0.2985417545,0.0370175764,0.0437167287,-0.1946992427,-0.0023853509,0.0292249881,-0.1095377281,-0.2861461937,-0.1512728631,0.1469767392,-0.0758948773,-0.0422932096,0.3087784648,-0.2856433988,0.2127100229,0.1209730133,0.5065145493,0.1983678192,-0.1395118386,-0.1003878266,0.1341328174,0.0941660032,-0.2758983374,0.1906488091,-0.0578917712,0.0293378606,-0.1639855057,-0.0040592146,-0.4304202199,-0.2518908381,-0.5167289376,0.0256337728,-0.167617023,0.0279525276,-0.2884708643,-0.1311133951,-0.1527274698,0.0477719828,0.2074298412,0.2324853837,-0.3048724234,-0.1078945547,-0.1457239538,0.2310075909,0.0114551643,-0.0470037721,0.4015561044,-0.1696835607,0.1348307133,0.4137633741,-0.1444092989,-0.24889265,-0.3052416742,-0.1933327317,-0.5849938393,-0.1785206497,-0.2186419517,0.3066715002,-0.2916366458,0.0653247312,0.090788722,-0.2250339538,0.1867794842,-0.3971133232,-0.1050942913,0.0150195044,0.2615575492,-0.3209237754,-0.2446400225,-0.2160146981,0.0213175751,-0.1389903873,-0.0529971421,0.253077507,-0.0388517454,0.0655227751,0.3861151934,0.1372185349,0.2675565779,0.3679469824,0.3419263363,-0.0033928149,-0.2892647386,0.0228993725,-0.3031847775,-0.0008268586,-0.1005435139,0.014012007,0.1042811647,-0.2691430748,-0.1032084376,-0.0674806386,0.1229776368,0.0552172512,0.0060696756,-0.3333392739,0.137013346,-0.2558971047,-0.2763421237,0.5580832958,-0.0587234348,0.1693245173,-0.3764621019,0.1268547326,0.0966551006,0.2309933752,0.2974235415,0.0872849524,0.051679682,-0.1487177759,0.0167446155,0.2936429977,-0.194155857,0.2961039841,0.1957146823,0.5390326381,0.2160008401,-0.0751854777,0.3775984347,-0.0670160055,0.0553693809,0.1712252051,-0.0678174421,0.2758797109,-0.2441231161,0.09979067,-0.2766646445,-0.0535205789,0.2998173833,-0.0084763365,-0.2149959505,0.0558523685,0.2065370232,0.1228862107,-0.1381969154,0.0391976498,-0.2054745555,-0.0551348887,-0.439814508,-0.0822409913,-0.2566308677,0.0857047737,-0.2197521329,-0.1574988663,0.4005419612,0.0638062134,0.35122177,0.1117026284,0.055626303,0.1014794558,0.6616276503,-0.1314482689,0.3922662735,0.1514548063,-0.0804909617,-0.2001651227,0.3394487202,0.1562802047,-0.2818534672,-0.0442734137,-0.2306788862,0.0450031906,-0.2120528221,0.1666831523,-0.3680693209,-0.4625119269,0.0466781966,0.3735579848,-0.6214097142,0.1193179637,0.1495131403,0.2071443647,0.0515228808,-0.0294180699,-0.1133590117,0.4791510701,-0.4406369328,-0.0066905147,-0.1465628445,0.0948427022,0.0196557567,0.1045660526,0.0436965153,-0.0225986559,0.5850809813,0.085253574,-0.0397208966,-0.2896198332,-0.1460235864,0.2113802284,0.0025429567,-0.2489184588,0.1970224082,0.3608907461,-0.4428425431,0.2280668467,-0.2399172634,-0.3119657934,-0.1648208201,-0.1356619298,0.0246884134,0.1858582795,-0.1512947977,0.3677878678,-0.1659554988,-0.0851375982,0.0624416471,-0.0626598299,0.0410466641,0.0085379891,-0.4504867792,0.0301185921,-0.2970254123,-0.2289239913,-0.4113452137,-0.9189471602,0.1523662657,0.116767019,0.0193303674,0.2935320437,0.4042418003,0.314075619,-0.1340131015,-0.2661589086,-0.1224794984,0.0487615317,0.0595248528,0.0118060354,0.0652003288,0.2354321927,-0.0499623604,0.0274178684,0.0571306162,-0.247469753,-0.3330615163,0.3217721283,0.3163862824,0.0432497188,-0.090486221,0.0323355719,0.5703100562,0.1193067357,0.1081985682,0.2337166071,0.058444351,0.2107961923,0.0774302259,-0.3203353286,0.2461823672,0.1141091213,-0.3820791245,0.3038387001,0.2107921839,0.2345762551,-0.0999833569,-0.1175929308,-0.3073520958,-0.2389181852,-0.0966971889,0.1623184234,0.3810125887,0.1455528289,-0.414398849,0.183823809,-0.6200313568,0.0153120514,0.099389039,0.4457888305,-0.1423617452,-0.3448931575,0.1389118433,-0.7456740737,0.3348259628,-0.2592067719,-0.4012364447,-0.0420986637,0.3280583024,0.3923137188,0.2205257267,0.4572169781,0.3931534588,-0.1158165857,-0.3311364651,-0.1643449515,-0.1489858627,0.0543987751,-0.1202849895,0.282807678,0.1660119742,0.0103463875,-0.3399403095,-0.3981840014,0.2015839815,0.0127111133,-0.2689374685,-0.1603001654,-0.199876681,-0.3115579486,-0.1354838312,-0.0702375174,0.2671819329,0.0904126912,0.0411679111,0.2658306062,0.2342758328,0.054875385,0.5304293633,-0.3897340596,0.1439941674,0.2950492799,-0.0499286167,0.2139315307,0.3759803176,0.3007421196,0.6087440252,0.1527267098,-0.1222707629,0.039960999,0.0556128323,0.4190649986,0.4938335121,0.0840897262,0.2226264626,0.0434640907,0.4279060066,-0.3727279902,-0.1342259794,0.1682679206,-0.3318158388,-0.1991730928,-0.0686134622,0.1978191584,0.1184224337,-0.1755993962,-0.1097560078,0.7189230323,-0.2271826565,0.0384958535,-0.0355748869,1.3941197395,-0.1602261513,0.1949436367,0.1941276938,-0.1110365391,0.684520185,-0.2874392271,-0.0362649485,-0.0968802869,0.0766119733,-0.1713692248,0.1745813489,0.0234018229,0.1255317032,0.3082107306,0.2151377946,-0.4961577952,0.3687135875,-0.1846528053,0.2205484062,-0.0238176975,0.021894563,-0.5131445527,0.0759435445,-0.2100075483,0.3007341027,-0.1663540751,-0.0731302202,0.1509174407,-0.2481132299,-0.039389357,-0.1208686158,0.1561521143,-0.0291194748,-0.111567609,-0.1162345782,0.1237898991,0.118069984,0.2266778797,-0.0220316909,-0.2298934758,0.1466965973,-0.1310493946,-0.1286080927,-0.1485782713,0.1618473679,-0.5426871181,-0.0890743732,-0.0815428644,0.084443666,0.2241679728,-0.0464943275,-0.1340242177,0.0344787426,-0.0058435281,-0.2914931774,-0.0973672569,0.0303255059,0.1702993214,0.0401915349,0.0708396211,0.3123110533,0.0006763602,-0.2959254384,0.286696434,-0.1537836194,-0.2394491434,0.4287185073,0.1977637559,-0.3225503862,0.2857302427,-0.0190901738,-0.1904806644,-0.2751338184,0.2016457468,-0.3425221443,-0.5874430537,0.1141286269,0.3149661422,0.0366048701,-0.291513443,-0.1975777298,0.1104004309,-0.0264898799,0.3168047667,-0.0584951602,-0.1946906,0.1652725339,-0.0926986858,-0.0414468125,0.0176058412,0.0027850289,-0.3060526848,-0.0078444416,-0.234719485,0.2258276492,0.135363847,-0.1650213301,0.2514096797,0.2690316737,0.1896714419,-0.3752946258,0.0139903324,-0.157998085,-0.0628549233,-0.1252652407,-0.2023278624,0.2092255503,0.0137126809,-0.3652531505,-0.1408753842,0.0106529891,-0.0787612796,0.0449377894,0.2707513571,0.6551380157,0.4446861148,0.0695408881,-0.1357013285,-0.1160853729,-0.209114939,0.1226799861,0.2805311084,0.3615549505,0.2780156136,0.1833037436,0.0210823752,-0.2138291299,-0.0988973752,0.1772830188,0.2302207798,0.1133605763,0.0641400889,-0.0504004434,0.016251009,-0.0165793989,0.3196050227,0.0419563241,-0.2212166339,-0.2051952332,0.0941631272,0.0966500789,-0.2390662432,0.0912446082,0.3166816831,0.2178631127,-0.2466182411,0.1899970621,0.00809028,0.3711920977,-0.0991901085,-0.0483709089,0.1039117277,0.1159602553,0.062235225,-0.1857697368,-0.0982569829,-0.1045930982,0.3679387271,0.1211199239,0.1277451217,-0.1870661974,0.0125024831,0.3199787736,0.133212626,0.3787963092,0.3281485438,-0.3077746034,-0.3909491301,0.2668871284,-0.018267151,0.1024359167,0.2293856144,0.4026900828,-0.0670615584,-0.4168702364,-0.2844020724,0.5166845322,0.0485898033,-0.0332628377,0.1105165109,-0.2009192109,0.0519627854,0.1180614382,-0.1513552219,-0.288562268,0.3093552887,0.0303902961,-0.1930820048,-0.2587171495,-0.130548507,0.25276196,0.1449282616,0.0835626274,-0.1449500024,0.0333268419,0.1658576429,0.1370889246,0.3544532061,0.06753847,-0.3148908615,0.3163480759,-0.0451440327,-0.1601140201,0.0241891704,0.0345295519,0.3621997237,-0.0635646805,0.0739471987,-0.0713593438,0.0179505944,-0.0155647891,-0.0254430734,0.083768934,0.3100956678,-0.3714370131,0.1992072761,-0.0777844861,-0.0018284766,-0.0205145404,-0.1432978213,-0.1920784414,0.0305632316,0.1091965213,0.0172963291,0.2012600899,0.0917831138,0.017233789,-0.2065968812,0.1258370131,0.4229773581,0.1077391878,-0.3011723161,-0.289493829,-0.3903499842,0.1269800663,-0.1087451875,0.0506038964,-0.2355711162,-0.0645275712,0.2830373347,0.1545535624,-0.2088692039,-0.4022492468,-0.0171848983,-0.0691530406,-0.0063841995,0.3615445197,-0.2786248028,0.3093092144,0.0446297266,-0.065840289,0.0928762108,-0.0068054367,-0.1091715321,-0.3168356717,0.3109953403,-0.1654196978,-0.3387309313,0.2403812408,0.1307757944,-0.1046118885,-0.0211965349,-0.3472576439,-0.0699116737,-0.0985999182,0.1005970687,0.0540763326,0.2209935188,0.1504027545,-0.0379116349,-0.0102781067,0.0456226021,0.5237256885,-0.2117824405,0.2887281179,-0.4321366549,-0.0973338485,-0.2158828974,0.2435979247,-0.1751489043,0.2454287112,0.0639057681,-0.1614898741,-0.2857495844,0.0038836165,0.7217822075,-0.3279657066,-0.0523691401,0.0803431347,0.7390746474,0.0482144319,-0.4714959264,-0.6054503322,0.3153680563,0.1254726797,-0.1929335594,-0.0164859015,0.3764760196,-0.2878525555,-0.0711824372,0.1421400905,0.4748225808,0.2032745183,-0.0297302417,0.2586497664,0.0450953469]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2105","title":"Request to remove S2ORC dataset","comments":"Hi! Sorry I missed @yjernite 's previous message, thanks for responding! \r\n\r\nIs there an option where we can keep our data in our bucket, but the HF script no longer pulls data from it? ","body":"Hi! I was wondering if it's possible to remove [S2ORC](https:\/\/huggingface.co\/datasets\/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!","comment_length":34,"text":"Request to remove S2ORC dataset\nHi! I was wondering if it's possible to remove [S2ORC](https:\/\/huggingface.co\/datasets\/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!\nHi! Sorry I missed @yjernite 's previous message, thanks for responding! \r\n\r\nIs there an option where we can keep our data in our bucket, but the HF script no longer pulls data from it? ","embeddings":[0.3147089779,-0.2077655643,-0.0239039473,0.3282617927,-0.0140945502,-0.1829030514,0.0175194405,0.172574982,0.1771008372,0.1621481776,-0.4038794339,-0.2786833048,-0.3127873242,0.457226783,-0.1433838755,0.1945003718,-0.0702701509,0.0524483249,-0.3202078044,0.0324457549,0.1301748008,0.2261826545,-0.0884683505,0.0661727116,-0.1380365342,-0.1128825396,-0.0782067031,0.1449866146,-0.1255226731,0.0694023669,0.2610510886,0.4705429077,0.0800844952,0.4339930117,-0.0001245844,-0.3209076226,-0.1540060341,0.0056156814,-0.4018520713,-0.1130016148,-0.0782959759,0.2641175985,-0.090856187,0.1397795975,-0.0534056127,-0.1262888759,0.0841633454,-0.1527345777,0.3187084496,0.2342592031,0.1314422786,0.575772047,-0.4303703308,0.2741442323,0.1232632101,0.6085684896,-0.1516879201,0.3492450416,-0.1199225709,-0.1097447276,0.2139885128,0.3302992582,-0.4357930124,-0.0924803093,0.249126032,-0.1535654664,-0.39856866,-0.2252966166,0.4142615497,0.2827500403,0.4330858886,-0.2104133517,-0.090478234,0.1022289321,0.0474308133,0.0494886525,0.0116874781,0.2144204527,0.0960223377,0.428937614,-0.3170785904,-0.6390438676,-0.2341393828,-0.1140615866,-0.3001761734,0.3237962425,-0.297346741,-0.064186208,-0.0738193691,0.0914316252,0.1797328889,-0.2322839797,0.012712026,0.1589960158,-0.0431557819,-0.3171835244,-0.2178788185,0.3126013875,0.2448574007,0.0445581377,0.0243876893,-0.1624384373,-0.2845410109,0.2050913423,0.3105307519,-0.0921661034,0.0419713147,0.0940728635,0.277145952,0.2518097162,0.2529099584,0.0715969577,0.1932775676,0.2897895277,-0.3893837929,-0.1564794332,0.5164502859,-0.4300911725,-0.416610688,0.3595436811,-0.055288177,-0.0619812571,0.1564075798,0.3035776615,0.1137095541,-0.0069999732,0.0698101297,0.0436593108,-0.2103589326,-0.3135282099,-0.1677012593,0.1802435517,-0.0791733786,-0.0170090403,0.2489594072,-0.2133324295,0.1953548342,0.1686762571,0.57807827,0.2631825507,-0.1107336059,-0.2234827727,0.3362011313,0.1756037772,-0.1483076066,0.1314199865,0.0225152355,0.1181295216,-0.2510817945,0.0298500694,-0.2922266126,-0.1787335873,-0.4848286808,0.0030082706,-0.1213938668,0.0565310046,-0.3228920698,-0.0896095932,-0.2625923753,0.1147754863,0.1649736762,0.0599419586,-0.3381242156,-0.0675013363,-0.1339311302,0.3447757065,-0.1291263402,-0.0032780236,0.4492904246,-0.217882663,-0.0202052947,0.389357239,-0.1809060127,-0.2190599293,-0.331469655,-0.1568554342,-0.4964418709,-0.2505916357,-0.0249097198,0.3402804732,-0.1744554937,0.0753001571,0.1196339279,-0.2354543209,0.2151425481,-0.457449317,-0.2506250739,-0.0161332916,0.188545838,-0.2797206342,-0.1101487204,-0.3190248013,-0.0062155803,-0.1179453284,0.1130622849,0.1365055293,-0.0230848268,-0.0962608829,0.4757193923,0.0390383042,0.276992023,0.3343988955,0.2557878494,0.1921966225,-0.2453435063,-0.0499248132,-0.2992227674,-0.095937632,-0.1304721832,-0.1254461706,0.1675016731,-0.3209268153,-0.0511224084,-0.0252981614,0.2593214214,0.1686456054,-0.0502269901,-0.289798528,0.005914236,-0.1839488447,-0.4556969404,0.622061491,0.0663529411,0.2031348497,-0.2550747693,0.2425263375,0.0820942149,0.1792575419,0.2667661905,0.1372851431,0.1339348108,-0.0780462846,0.1209661216,0.2498843968,-0.1223325953,0.1720343828,0.2978731096,0.5357654095,0.1379655898,0.0415779874,0.4445880055,-0.1042024791,-0.0114139263,0.1728285253,-0.044643905,0.3055931628,-0.2436221987,0.2162426114,-0.1580538303,-0.1398379803,0.2460268885,-0.1375356913,-0.1633789986,-0.1040517762,0.2351144254,0.0534564406,-0.129294008,-0.027631931,-0.3138299584,0.08439219,-0.0928236246,-0.114094764,-0.2346712351,0.1019420847,-0.2157959044,-0.157847032,0.4094680846,-0.0664004236,0.2941377461,0.1521456391,-0.1029986367,0.1443632841,0.4886132181,-0.1357595325,0.2988614738,0.1198166832,-0.1439885348,-0.0977470353,0.2450703681,0.1141307279,-0.4201129675,-0.1339052022,-0.2944412231,-0.0589345358,-0.2316478938,0.1991477609,-0.3803668022,-0.3432620764,0.0363623835,0.3859191835,-0.5940880775,-0.0211536158,0.2378588766,0.4233310521,-0.0320284627,-0.0436418802,-0.2210730314,0.4132110476,-0.5426178575,0.020808218,-0.197937429,0.0864612684,0.0401306972,0.0360178836,0.0378568955,0.0495851785,0.6460420489,0.1960494369,-0.2760607302,-0.2868872285,-0.0752945095,0.2934106588,0.0080150366,-0.2756493986,0.2361773998,0.4233388305,-0.354983896,0.2639247775,-0.2653572559,-0.4017896056,-0.115709424,-0.0299876258,-0.0388156325,0.2117659897,-0.1225452349,0.3960286379,-0.0757073984,-0.1922045201,-0.0743346885,-0.1557215452,0.1037826389,0.0387075059,-0.3879113495,-0.04676985,-0.4429834187,-0.2824510336,-0.4912324846,-0.778198719,0.1207469106,0.0699068308,0.057207875,0.332026571,0.3459808826,0.111493811,-0.1548401415,-0.3589918911,-0.2789942622,0.1449712664,0.0791931823,-0.1249559447,0.0872814357,0.3909487128,-0.0427657627,-0.0010109162,-0.0317496695,-0.1366344392,-0.2140945196,0.2423254251,0.1992513686,0.0455872454,0.0257471818,0.0106926076,0.6046019793,0.1010861024,0.1844915003,0.3397406042,0.1942436397,0.3856178522,-0.0550438724,-0.232566908,0.2094053626,0.1510695219,-0.3510402739,0.4016523957,0.1588745713,0.3327524662,-0.0486285612,-0.2903546989,-0.1625440419,-0.3546296358,0.0428744294,0.1451779306,0.3210632503,0.1124968603,-0.2528866231,0.1618879735,-0.4717038274,0.1904237568,0.1604332626,0.5285871029,-0.141127944,-0.3676763475,0.0719339848,-0.7467413545,0.3602396846,-0.3094731569,-0.3952311873,-0.1068519354,0.2692242563,0.2746241093,0.1511479914,0.5630870461,0.2909496129,-0.0957487971,-0.2798063457,-0.3013560176,-0.155266583,0.0375740752,-0.1183612347,0.1619005948,0.0003908417,0.0660540536,-0.3809040189,-0.5336230397,0.30143556,0.0043947771,-0.2740958333,-0.1703826189,-0.1240279377,-0.3155560493,-0.2255102694,-0.1006583273,0.2319577932,-0.1187920421,-0.0158961955,0.1792653799,0.2700365186,-0.0834590793,0.5575609207,-0.3295531869,0.0393242426,0.297221005,-0.1338317543,0.2257256657,0.2999958396,0.4112845659,0.5585439801,0.1418979168,-0.1239070371,-0.0102412114,0.1979884952,0.5615346432,0.5556922555,0.128477484,0.3664983511,0.0918800905,0.5001993775,-0.3880907893,-0.0956383869,0.1000047848,-0.4081890583,-0.1625106782,-0.1049481183,0.1417706758,-0.0644084811,-0.1493146122,-0.2081586272,0.743694067,-0.2282913327,0.2903295457,-0.0346304327,1.4422258139,-0.1593497992,0.3056579828,0.2118859291,-0.2551866174,0.5051166415,-0.4148739874,-0.2021965086,-0.2295224518,0.0021034621,-0.1674166024,0.172137782,-0.0132457418,0.0274638683,0.3120126724,0.3241525292,-0.5391010046,0.3610616624,-0.2570849657,0.2458502501,-0.1036136746,-0.0680968836,-0.3832627535,0.0401934385,-0.1998199224,0.2853231728,-0.2148007303,0.0085261706,0.3091729879,-0.2910861075,0.0938097909,-0.100554049,0.0519237332,-0.1713518202,-0.1621165276,-0.1322745532,0.118065089,-0.0501562394,0.3635021746,0.0336597748,-0.1031588092,0.11896687,-0.1161589846,-0.0845567584,-0.1702689379,0.1121316478,-0.3557922244,-0.0161619391,-0.1206367612,0.0903092176,0.3170487881,-0.0723592415,-0.1763640195,-0.0874789804,-0.0170360096,-0.3474556506,-0.0331217386,0.0889291093,0.247477144,0.0167270396,0.024522569,0.3489470184,0.048667226,-0.2057616264,0.2901416123,-0.0811120793,-0.3226546347,0.4245264232,0.1744001359,-0.2102300674,0.2772931457,0.0930409729,-0.2425220609,-0.303507477,0.0678163022,-0.054580275,-0.535410583,0.1562335342,0.3695347905,-0.1622982025,-0.3122070134,-0.2166408151,-0.0735200867,-0.1083327606,0.1686461717,-0.0627775192,-0.2172715515,0.28267923,-0.0591713488,0.0217171758,0.2163206339,0.1047018468,-0.4611589313,-0.0149918655,-0.2377748042,0.0556306131,0.1930533201,-0.1208489835,0.1916596889,0.2497819215,0.2629216313,-0.4145078361,-0.0145700015,-0.1985589862,-0.0779795051,-0.1024585441,-0.3338911235,0.246131748,-0.0063889208,-0.1803735644,-0.1264831573,-0.0808150619,-0.0587179102,-0.1077544168,0.3669096529,0.7037873268,0.3963145316,0.1003689691,-0.0343869925,-0.1703873724,-0.1662849039,0.1051094383,0.3018845618,0.4229618013,0.1335193664,0.09797699,-0.0256355982,-0.1721761525,-0.1169032753,0.3478453457,0.3693528175,0.144340992,0.0850353166,-0.006866463,-0.1513214558,-0.0705442205,0.3867637217,-0.0746734664,-0.1984468549,-0.3272590935,0.182355687,0.0592766814,-0.2297749817,0.093649134,0.2319950461,0.1131342575,-0.1623991877,0.2448350191,0.0431437641,0.3019334078,0.0668520778,-0.0174010359,0.0813684538,0.1112634838,-0.0155171184,-0.2162245214,-0.1460536271,-0.0467095226,0.28930825,0.200284034,0.2045659572,-0.1425302625,0.0626847744,0.2676782608,0.1584055871,0.3818683624,0.2592196167,-0.3135669529,-0.3890518844,0.3242782056,0.0431988612,0.0384174772,0.1344476044,0.3216623068,0.0329442844,-0.4210071266,-0.1812735498,0.529178977,0.0012198394,0.0575157963,0.1724272817,-0.0433884747,-0.0143761393,0.0684770942,-0.0563087463,-0.2920878828,0.3737664819,0.0783519745,-0.1012215614,-0.0771473646,0.017835049,0.0991690233,0.1210935935,0.1670678556,-0.2026459575,-0.0042343605,0.319452554,0.1413637847,0.3663030267,0.0771815032,-0.3462436795,0.2414945066,-0.0717810914,-0.1055394784,0.023887042,0.0080830632,0.1334741414,-0.1241125837,0.1568184942,-0.0017038814,-0.0203037802,0.024002213,-0.0528658591,0.0597954653,0.2323178053,-0.3492250443,0.3407617211,0.1535464078,0.0007398657,-0.0673205405,-0.181366697,-0.1566785127,-0.0461263135,0.1495536566,-0.0137555636,0.2211215943,0.0334270149,0.0079473853,-0.2061505467,0.0759961158,0.3485670388,-0.0264005661,-0.2704100907,-0.1542821825,-0.2270944268,0.0794658139,-0.2377437353,0.0104717445,-0.2141761482,-0.0689808801,0.2982633114,0.0271291472,0.0221671052,-0.5149398446,0.0391796604,-0.0340176038,0.1177335829,0.244677037,-0.3900574446,0.32407704,0.064544715,-0.1055511832,0.0917984918,0.0427722372,-0.155804798,-0.2495812178,0.3124651313,-0.008296283,-0.5492114425,0.3167098761,0.1164444461,-0.1666411757,-0.121127151,-0.361618191,0.0660668463,-0.0631120354,-0.0784439296,-0.1120164096,0.1449250281,0.1889978051,-0.0445396975,-0.1906405091,0.1306956708,0.4611991346,-0.2786438465,0.3786755502,-0.3487754762,0.0460125469,-0.2282998115,0.3409302533,-0.1601590365,0.2992237806,0.0737596899,-0.1208450943,-0.3632829189,0.0320407376,0.7371150851,-0.2060651034,-0.1711030304,0.0094223451,0.7621918321,-0.1084106341,-0.5149688125,-0.5980624557,0.2991893888,-0.0139808385,-0.0637314767,-0.0169630349,0.383295536,-0.298522085,-0.0275079533,0.1987161934,0.4187488854,0.3243615627,-0.0627263635,0.2210436463,0.0181117691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2104","title":"Trouble loading wiki_movies","comments":"Hi ! `wiki_movies` was added in `datasets==1.2.0`. However it looks like you have `datasets==1.1.2`.\r\n\r\nTo use `wiki_movies`, please update `datasets` with\r\n```\r\npip install --upgrade datasets\r\n```","body":"Hello,\r\nI am trying to load_dataset(\"wiki_movies\") and it gives me this error - \r\n\r\n`FileNotFoundError: Couldn't find file locally at wiki_movies\/wiki_movies.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/wiki_movies\/wiki_movies.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/wiki_movies\/wiki_movies.py`\r\n\r\nTrying to do `python run_mlm.py \\\r\n --model_name_or_path roberta-base \\\r\n --dataset_name wiki_movies \\` also gives the same error. \r\n\r\nIs this something on my end? From what I can tell, this dataset was re-added by @lhoestq a few months ago. \r\nThank you!","comment_length":27,"text":"Trouble loading wiki_movies\nHello,\r\nI am trying to load_dataset(\"wiki_movies\") and it gives me this error - \r\n\r\n`FileNotFoundError: Couldn't find file locally at wiki_movies\/wiki_movies.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/wiki_movies\/wiki_movies.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/wiki_movies\/wiki_movies.py`\r\n\r\nTrying to do `python run_mlm.py \\\r\n --model_name_or_path roberta-base \\\r\n --dataset_name wiki_movies \\` also gives the same error. \r\n\r\nIs this something on my end? From what I can tell, this dataset was re-added by @lhoestq a few months ago. \r\nThank you!\nHi ! `wiki_movies` was added in `datasets==1.2.0`. However it looks like you have `datasets==1.1.2`.\r\n\r\nTo use `wiki_movies`, please update `datasets` with\r\n```\r\npip install --upgrade datasets\r\n```","embeddings":[-0.2073984444,-0.0135140801,-0.0442202128,0.3873903155,0.2891761363,0.1963276118,0.1652412564,0.2904251814,0.1623347551,-0.0591131747,-0.0651294068,0.0398705192,0.0390818752,0.0095692184,0.2180415839,-0.1336186975,0.1319327056,-0.0677647516,0.0409930721,-0.1519562751,-0.2609515786,0.1905121803,-0.0947827175,0.0493639857,-0.3012830019,-0.0213264003,-0.0860880464,0.3671931624,-0.1554056555,-0.3068653345,0.4555034339,-0.1563687027,0.6124148369,0.557012558,-0.0001154108,0.0793064833,0.4557094872,-0.0438040532,-0.3138126731,-0.2339858413,0.1990428716,-0.3066937029,0.3091194332,-0.0718859509,-0.1147370338,-0.0357826762,0.1456821859,-0.1668719202,0.1003157943,0.0856170207,0.1734515131,0.2701095939,0.2892472148,-0.0911118761,0.2961936891,0.1140339896,-0.0202871021,0.4139375389,0.2665452063,-0.4807064235,0.2214951515,0.3279593587,-0.3136359453,-0.0704162121,0.5433495641,0.1534255147,0.3824139535,-0.3687499762,0.1887979656,0.3565346599,0.6607331038,-0.082890965,-0.3993507028,-0.0977442786,0.1088015437,-0.1376087964,0.2221833766,0.2584237158,-0.0588011406,0.1871618778,0.0967815742,-0.3336417079,0.0515212454,0.5347654819,0.0972616374,0.495184958,-0.0458648615,0.0977662057,0.1348397285,-0.2122950405,-0.1636189222,0.0070710653,-0.0641391799,0.4242524803,0.0684447214,0.103847377,-0.0089857206,0.4857098162,-0.1439783871,-0.0411232486,-0.1150691286,-0.0414399244,0.0586496219,0.2267344296,0.3234931529,0.0552307591,-0.0866382122,0.0497813858,0.2694105804,0.4145142436,-0.1821436435,-0.1986727118,-0.1851852089,-0.0572147965,-0.3604752421,-0.2458050251,0.3958001733,-0.1175649837,-0.2839041054,-0.0130766714,-0.2200167477,-0.2205563337,0.2549139559,0.554427743,-0.2136487961,0.1867852509,0.0356269404,0.2033433467,-0.2397906929,0.1961438656,-0.1105699241,0.191212222,-0.1960520893,0.2319417596,0.3138676286,-0.0700656101,0.4374935925,-0.1168764308,-0.001596784,-0.1212561131,-0.0064617116,0.0378491059,-0.1349067241,0.4637263417,-0.0092822723,0.2072140574,0.3247037232,-0.310964942,-0.1152983084,0.0831781477,-0.2376967818,-0.3491517007,-0.1643438041,0.0872072205,0.1324012727,-0.1991852224,-0.2350740582,0.0310069155,0.0281784106,-0.5631130338,-0.0473254807,0.1131903678,0.0118427556,-0.2234303057,0.3480658531,0.6888164282,-0.2556487918,-0.0585722364,-0.3802088201,-0.0084375693,-0.0683746338,0.1990648955,-0.2756039798,0.0593011007,-0.3056888878,-0.0649878308,0.1098074764,-0.5922428966,-0.2333725989,0.2845813334,-0.1365546733,0.1409637332,0.125161007,0.1688186377,-0.3357461393,-0.1660052687,0.1106749028,-0.051254347,0.0735550523,-0.2340956926,-0.1769212335,-0.1521123648,0.1346911639,0.1320302784,0.0286435653,0.0403858274,0.0866599083,-0.0284465738,0.2409082204,0.0489289574,-0.1485192627,0.1628978252,0.3491697609,0.4564087987,-0.0384004116,-0.0741909742,-0.4570272863,0.1895824671,-0.107014969,0.0173462518,-0.1545140743,-0.0961494073,-0.06145107,0.0410896279,-0.1808400601,-0.2253644913,0.0850974694,0.054778289,-0.0529993176,0.1655847877,-0.27090469,-0.1925833076,-0.0947699547,0.2471811026,-0.4027302563,0.251999259,-0.1455648541,-0.0060172165,-0.0319152623,0.2065127194,0.2627585232,-0.1470631659,0.1542284042,0.390866667,-0.0781015381,0.2508268654,0.2339447886,0.1645369977,0.1641920507,-0.0635677576,0.2368172109,0.1799629778,0.2339200824,0.0259684641,-0.1443175822,0.3370316625,-0.0091855526,0.2501546443,0.218484953,0.0082469713,0.1893139035,0.0529132225,-0.0398360416,-0.2208691835,0.432536304,0.2068346143,0.1125894114,-0.0895571187,-0.2570354342,-0.0425548479,0.0272043329,-0.0885862857,0.2019982934,0.0528827906,-0.3966023326,0.1287508011,-0.0062545682,0.1457238793,0.3508633375,0.1889484227,-0.1748357713,-0.0642648563,-0.0122196469,-0.2063001543,0.2619258463,0.1753786355,0.0702292025,0.1796754152,-0.0203876663,0.0231949333,-0.2697029114,0.0314484648,-0.1116849259,0.1491054446,-0.2536031604,0.1608959585,-0.2967224717,-0.0686194301,-0.2333277017,-0.4196031392,-0.3267813027,-0.2818979323,-0.0361126885,0.3151589632,0.2061970383,0.0726836622,0.0288040824,-0.018881822,-0.3148033917,-0.1647440791,-0.1038463637,-0.3200178146,-0.2616140246,-0.0129883084,0.5495477915,-0.0611363761,0.1579234898,-0.1074337065,-0.0375607982,-0.2795412838,-0.5420339704,-0.0647832081,-0.1012944356,0.5056604743,0.0170397907,0.3776248395,-0.1283347607,-0.3390266299,0.3770352006,-0.0025069732,0.1115655228,-0.1031463221,-0.0545268431,-0.3354249597,0.0182406269,-0.4875997305,-0.4276797473,-0.1854010522,0.1409364045,0.2385593951,-0.1262493134,0.1926994622,0.1057868078,0.1727087349,-0.1761257499,0.027730396,-0.0571366362,-0.3297139108,0.3998129666,-0.226129353,-0.2458426505,0.4008429348,0.0212226678,0.0199608039,-0.296489954,-0.5815371871,-0.1569053829,0.1353253424,0.067797929,0.1800185144,0.1077848375,0.1994774491,-0.0731825605,0.0334883668,0.0757279992,-0.1539250761,-0.2185354829,-0.1357319504,0.2919263542,-0.1059441492,0.486774981,-0.4204573929,0.7676123977,0.0365934484,0.0802361146,0.3636780679,-0.1858739853,0.4716041684,-0.0990008041,-0.4413351417,-0.0774862617,-0.2061032802,-0.1245743856,0.2974345982,0.1417627335,-0.268399328,-0.3975665867,-0.0912308693,-0.3086287379,-0.2211138308,0.0352220461,0.3111131787,0.0772450119,0.2566117644,0.1754665524,-0.0801588148,-0.3820620775,0.1539733559,0.7092061639,0.1075186878,0.0038920434,-0.457084924,-0.4015654922,-0.2711562216,0.3313780427,-0.0077607129,0.0697348714,-0.1062444523,-0.0869510546,0.174001649,-0.1624230444,0.6706092358,-0.0837712809,-0.1302607507,0.0203569531,0.0557792708,-0.2821876705,-0.1787357926,0.0687449872,0.2667483389,0.0251793824,0.4684501886,-0.0857275203,-0.1418095976,0.2317419499,0.1474017054,-0.2520614564,-0.1686155498,-0.4849333167,-0.2130189985,-0.3204255104,-0.2913609743,-0.1253483146,0.2390796542,0.011321242,0.1562641561,-0.0167716797,0.0761718526,0.1815519184,0.0424850024,0.3289707601,0.2473449558,0.037546318,0.2949092388,0.4270302653,0.3645468056,0.5436066389,-0.1697588712,-0.0709252432,-0.1214908734,-0.0281768162,0.2388252318,0.3939904273,-0.0232044347,-0.1704547554,0.3597781062,-0.109695822,-0.3963788152,-0.2588639855,0.3368563354,-0.093652539,-0.1389511973,-0.4274101555,0.7648703456,0.0248678662,0.0830103457,0.0545723587,0.2724692822,-0.1628162712,0.0477627739,-0.0272066016,1.0328421593,-0.1855564713,0.3174365163,0.3454584777,0.2381198555,0.315577656,-0.4082057774,0.1720009446,-0.2932628095,0.0589847602,-0.1529383361,0.0050000553,0.229162991,-0.0227830894,-0.2222685367,0.382942766,-0.0745870695,0.1983661503,-0.2022183537,0.1291624457,-0.0472921282,-0.0310841259,-0.5368956327,0.0727835894,0.1796755642,0.4815181792,-0.2882339954,0.0028446394,-0.2557936013,-0.1622874439,-0.3092894554,-0.0766073167,-0.4254750311,0.1606012285,0.1578443199,-0.1742732972,-0.1863705963,0.4830913842,-0.0314789228,-0.024061298,-0.3653795123,0.174251318,-0.1446528882,-0.5440136194,-0.2830952406,0.0278855544,0.3291461766,-0.082562536,-0.2290508151,0.0745175779,-0.2145106643,-0.0334134251,-0.2611868978,-0.048483137,0.1944985092,-0.0629848689,-0.5832720399,0.0490506664,-0.0102972323,-0.0552632734,0.0820280835,0.0675187632,-0.0686377585,-0.1999364048,0.2128704041,-0.0681738108,-0.1432762593,0.6447443366,-0.2534944713,-0.0229541156,0.614634037,0.4854166806,-0.1640946716,-0.1541500688,0.2329573482,0.1565604955,-0.2674960196,-0.1197681129,0.5612847805,0.2112629563,0.2064646184,0.172425136,0.2825322449,-0.2618932426,-0.0820551589,-0.6242221594,-0.0708412305,0.4252009392,0.0374956504,-0.0521982051,0.0249280576,-0.0716777742,0.1992085129,-0.0941833183,-0.2477563918,0.1305895448,-0.0087979613,-0.0262038931,0.501226306,0.1420454085,0.3230374455,0.1515271217,0.0972075239,-0.1200804338,-0.1471910924,-0.1446714997,0.0379568562,0.182738483,0.0398262739,-0.1624146253,0.0093399696,-0.288536489,0.0145149669,-0.1437316388,-0.1796106994,0.0179264005,0.1462146044,-0.0515186079,0.1145930141,0.0572146475,-0.2606999278,0.1546106637,-0.120540224,0.295958519,0.1922041774,-0.2004704028,0.2075010985,0.0634206161,-0.0060105571,-0.0779672116,-0.2188494354,0.1123150066,0.4374074936,-0.1960472912,0.1059693396,0.1638552397,-0.0032018274,0.4794400036,-0.1635046601,0.2560341358,0.278272748,0.1207379252,-0.4208910465,-0.0099682473,0.3926069736,-0.0322331935,-0.1157750338,0.1749530882,0.0917502642,0.1105135903,-0.1610085815,0.2474080175,0.3215138316,-0.2639528215,0.2476571351,0.2489957511,0.1460111737,0.0742936656,-0.0269911122,0.2259486914,0.0321108811,0.4658453763,-0.3481173515,0.218534112,-0.1139162928,0.2591001689,-0.1901556849,-0.3677089214,0.1250477433,0.3036642373,0.1084167734,0.0828297734,-0.2962623835,0.3766218722,-0.2566363215,0.2435160279,-0.2705171704,0.2557413578,-0.1564921588,-0.3460777402,-0.1184041426,-0.3570038974,0.0342638604,0.0363691747,0.023547383,-0.1979919076,-0.0623365454,0.2127105147,-0.1376571953,-0.3341412842,-0.0650519133,0.0843127742,0.1563048214,-0.1054155752,0.0710222125,0.3002874553,-0.0058786403,0.1646107733,0.0461681932,0.5031881332,0.4529600143,-0.3615342379,0.1513344496,-0.0766324103,-0.1688948423,0.043856997,0.2025815994,0.1634923667,0.1932382733,0.0230461545,0.1304173917,-0.1090235561,0.3621741235,-0.055103384,0.2701662183,-0.2747119367,0.0897111967,-0.4953476787,-0.0787106231,-0.4654944241,0.1308906823,-0.4352192283,0.2043396235,0.3818311095,0.0859170929,-0.0824410096,-0.1214954481,0.0348882414,-0.0609525032,0.5068445206,0.4655907154,-0.004284089,-0.3190932572,-0.0413866378,-0.84354949,-0.0332939476,-0.1007931754,-0.0957551599,-0.1398044378,0.1500568986,0.1667107195,0.0404701531,0.0403536297,-0.5100941658,0.1848676354,0.024006797,-0.1230856478,-0.2198768705,-0.2002838552,0.2074137032,-0.027175203,-0.0048105302,-0.1052840501,-0.4328902066,-0.1044766456,-0.2810299695,-0.0002233496,-0.1893548667,0.0383282565,0.2756373286,-0.1559370309,0.4702200592,0.0535785854,-0.297331959,-0.4378244579,-0.2957700193,-0.0427513272,0.2842938304,0.0013646459,0.4818758368,-0.4417456388,-0.0757945031,-0.3557061255,0.197299853,0.0120003885,0.0551024266,-0.2170319557,-0.1389431804,0.057626538,-0.0787537619,0.2426235974,0.2271084487,-0.0575167313,0.2714572251,-0.0860496014,-0.124000147,0.6063528061,-0.6618577242,0.0450970493,-0.0212196056,0.0025201014,-0.2047627568,-0.0113494201,-0.6350098252,0.1499135047,0.2958351374,0.1970753968,-0.2056305408,0.0955725983,-0.0133368857,-0.1813097149,-0.1173966378,-0.1059836894,-0.0219055973,-0.1565832198,-0.0299983677,-0.086439006]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2104","title":"Trouble loading wiki_movies","comments":"Thanks a lot! That solved it and I was able to upload a model trained on it as well :)","body":"Hello,\r\nI am trying to load_dataset(\"wiki_movies\") and it gives me this error - \r\n\r\n`FileNotFoundError: Couldn't find file locally at wiki_movies\/wiki_movies.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/wiki_movies\/wiki_movies.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/wiki_movies\/wiki_movies.py`\r\n\r\nTrying to do `python run_mlm.py \\\r\n --model_name_or_path roberta-base \\\r\n --dataset_name wiki_movies \\` also gives the same error. \r\n\r\nIs this something on my end? From what I can tell, this dataset was re-added by @lhoestq a few months ago. \r\nThank you!","comment_length":20,"text":"Trouble loading wiki_movies\nHello,\r\nI am trying to load_dataset(\"wiki_movies\") and it gives me this error - \r\n\r\n`FileNotFoundError: Couldn't find file locally at wiki_movies\/wiki_movies.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/wiki_movies\/wiki_movies.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/wiki_movies\/wiki_movies.py`\r\n\r\nTrying to do `python run_mlm.py \\\r\n --model_name_or_path roberta-base \\\r\n --dataset_name wiki_movies \\` also gives the same error. \r\n\r\nIs this something on my end? From what I can tell, this dataset was re-added by @lhoestq a few months ago. \r\nThank you!\nThanks a lot! That solved it and I was able to upload a model trained on it as well :)","embeddings":[-0.1121565476,-0.0139938025,0.0073024975,0.4026295841,0.3210339546,0.2078264952,0.2428797483,0.2546978891,0.1591462195,-0.1293400973,-0.0809662715,-0.0636573881,-0.0174834076,0.0371454656,0.2689795792,-0.1464951634,0.1440912783,-0.0804251581,0.0386452563,-0.1759168506,-0.2859064937,0.1560275704,-0.0958821923,0.0523040965,-0.3731435835,-0.0070697526,-0.1327134371,0.4059880972,-0.1415062249,-0.2147176713,0.4609825909,-0.1746772528,0.615390718,0.5836373568,-0.0001164071,0.1489440203,0.359936893,-0.1114915237,-0.2162447274,-0.1964496374,0.2064691633,-0.2007347345,0.2710044682,-0.0929706842,-0.1247370839,0.0024904485,0.1638142616,-0.1582214087,0.0603854433,0.166710943,0.1763141602,0.2521987855,0.217990756,-0.0762792006,0.2325390875,0.1530012637,-0.0561034381,0.4803478718,0.1946611851,-0.5227861404,0.1577254832,0.2466911525,-0.2417711616,-0.1198350862,0.5773373842,0.1589505821,0.3579044044,-0.3634532988,0.1996272802,0.4087169766,0.6135014892,0.0077126753,-0.370562017,-0.1067453921,0.1170774624,-0.0947812945,0.2538561225,0.3591564596,-0.0323338993,0.2213998437,0.0957057402,-0.2623707354,-0.0076847128,0.4480513036,0.0597298406,0.4561125338,-0.0188436452,0.0658945441,0.1543425322,-0.1914173961,-0.2707585394,-0.0120477099,-0.037583001,0.4495453835,0.0333939567,0.1286818385,-0.0111189717,0.4703658819,-0.1006703675,-0.0471908003,-0.0657706112,-0.0687851161,0.0245423317,0.1941131949,0.2931633592,0.0672160462,-0.1521092057,-0.0033023122,0.2521961629,0.4258029759,-0.280762732,-0.2049828321,-0.25954476,-0.048532486,-0.4100686014,-0.3608543575,0.378241986,-0.1459192783,-0.3006827831,0.0537187345,-0.2405789495,-0.2009180188,0.2333174646,0.5762203932,-0.1611888111,0.1728943437,0.0266179629,0.1784474105,-0.253541559,0.1513384879,-0.1137235165,0.2067685872,-0.2304682285,0.2217813432,0.2740677893,-0.0368960947,0.4466458261,-0.1666414291,0.0131379003,-0.1406337619,0.0264505632,-0.005662269,-0.1238665432,0.4525800347,-0.0446489863,0.1515550613,0.3736090958,-0.3536576033,-0.1087918207,0.0837685466,-0.1776937395,-0.3517285287,-0.1060338616,0.0890505984,0.0701194853,-0.193263337,-0.2120414525,0.1135182306,0.0000123633,-0.4642557204,-0.0614666604,0.1810542345,-0.0781361982,-0.2276006937,0.3975328803,0.7473238707,-0.2878182828,-0.0496014506,-0.2907510102,0.0037929702,-0.0579738282,0.2702949643,-0.2691882551,0.151519835,-0.3390530944,0.0829209238,0.0662628189,-0.4536378086,-0.2166564614,0.3275752366,-0.1736964434,0.1387213767,0.059236344,0.1650346369,-0.3982459903,-0.1746533513,0.1784097999,0.0247272737,0.0772709846,-0.2549577653,-0.2193595171,-0.1762289256,0.1955959201,0.2131317854,0.0416980609,0.0317923725,0.0563543476,-0.0642138794,0.2531671822,-0.019083675,-0.203308776,0.1463198215,0.3620998263,0.3864982128,-0.0049206833,-0.0095542707,-0.4130249619,0.2022417784,-0.1039938107,0.0395719111,-0.0981931686,-0.07035245,-0.0420112349,0.0717680156,-0.2699703574,-0.2291836739,0.0739770234,0.0106790503,-0.1434487253,0.2013347596,-0.3167356849,-0.2031098753,-0.2150320858,0.2404493541,-0.3909479082,0.2344633788,-0.1204902753,0.0242043771,-0.0409120955,0.1739754528,0.2111233771,-0.1363126636,0.1331696361,0.3344074786,-0.0879498795,0.1992669553,0.3177417517,0.2289922982,0.171147123,-0.0794308186,0.1990139633,0.2143538892,0.2429057956,-0.029391164,-0.1889046133,0.3995218277,-0.0126831708,0.2303565294,0.24644126,0.0190832987,0.1926739961,0.0369023979,-0.0532949306,-0.1471385807,0.4825780094,0.249484852,0.1442347616,-0.0129474057,-0.2999362946,-0.1178653836,-0.0263981614,-0.1227505654,0.2149269432,0.0591586344,-0.4093381763,0.0853517577,-0.0433144458,0.1072097197,0.3932321668,0.1262845546,-0.1545924991,-0.0733814016,-0.003325501,-0.2250157893,0.2733381987,0.1382026374,0.0763287544,0.2013794333,0.0129279839,0.0100938091,-0.3816268146,0.0092583671,-0.077999495,0.0810074881,-0.2232536972,0.1762468219,-0.1877779067,-0.0294050667,-0.2262083292,-0.3067170978,-0.3279341161,-0.2259712368,-0.0376599059,0.2222719938,0.1373322308,0.0279576629,-0.0798412561,0.0734919384,-0.2874613702,-0.1831968576,-0.0358819626,-0.2942383885,-0.2293563038,-0.0159935523,0.6218863726,-0.1350497454,0.1590105891,-0.0959247127,-0.0893320516,-0.213015005,-0.5159598589,-0.0486749336,-0.0142011419,0.4615418017,-0.0363291353,0.4115782678,-0.1496347636,-0.348587662,0.4001303017,0.0615613274,0.1341527253,-0.054527875,-0.0358310677,-0.2908650935,-0.0004587636,-0.4532815814,-0.3932527006,-0.1345376372,0.1750051379,0.218378216,-0.1383410841,0.2596439421,0.0735045075,0.2045798451,-0.2176378965,-0.005353678,-0.0445678681,-0.3735512495,0.4692232013,-0.1874055713,-0.3353832364,0.401188612,0.005090809,-0.0372040942,-0.3107057512,-0.593129158,-0.0952707753,0.1236285716,0.0605258867,0.2066353858,0.0428930148,0.186994493,-0.1618367285,0.0479859821,0.0944118649,-0.1230252236,-0.1039203629,-0.0901140198,0.3017452657,-0.1561519206,0.4466749132,-0.4208746552,0.714972496,0.0078862729,0.0581380278,0.3880026937,-0.1477459669,0.4323696494,-0.1417575926,-0.3912588954,-0.019208448,-0.2129064053,-0.202950716,0.376585722,0.2192340493,-0.2898501754,-0.3992303312,-0.1611683816,-0.2319954336,-0.2161508203,0.0800605491,0.3433607221,0.0476763062,0.2242882699,0.1665181518,-0.0271702781,-0.3645151854,0.1873081028,0.7527638674,0.2025979161,-0.0392807834,-0.4031485915,-0.3619454503,-0.2974769473,0.2064363807,-0.0998303667,0.0445729382,-0.1402748972,-0.1308009624,0.1976874769,-0.149513185,0.7062321901,-0.0151013844,-0.1303633302,-0.0237949006,0.0101142507,-0.2143746763,-0.1873928756,0.1263722777,0.2255820781,0.0175897162,0.5091714263,-0.0220080521,-0.0225527342,0.1574703455,0.1466571987,-0.2593086362,-0.1771282852,-0.4565154612,-0.240256831,-0.2624642849,-0.2855707109,-0.1059529483,0.2425221503,0.0959400311,0.2101483792,-0.0117289023,0.1187970117,0.2532066405,0.0592055358,0.2856893241,0.236504212,0.0610848255,0.3171516359,0.5217161775,0.2833810747,0.5718852878,-0.1491394341,-0.1094029024,-0.0825064257,-0.0690719783,0.2677215636,0.4327822924,0.0016410935,-0.180128783,0.4139865339,-0.0810645372,-0.4332716465,-0.1700492054,0.3224598467,-0.1302316487,-0.2027725875,-0.3941427767,0.7360111475,-0.0714670047,0.1236150861,0.0375176594,0.24553518,-0.166727826,0.1385148615,-0.0377053842,1.1215853691,-0.2203507572,0.3800052404,0.379666537,0.2959088683,0.2363575697,-0.4449314177,0.1420252472,-0.290137589,-0.0213325862,-0.1697994918,0.0020003421,0.1596799344,0.0362583883,-0.1308234036,0.3028576374,-0.1104075685,0.2878307998,-0.1666545272,0.0959408358,0.0523209944,-0.0351440348,-0.5819414854,0.0499434955,0.21045129,0.4309814274,-0.2776635289,-0.0368851237,-0.2028919756,-0.1022416353,-0.2853079438,-0.0471933112,-0.3949620426,0.1297575086,0.132298544,-0.1530512869,-0.1574645787,0.5224311948,-0.0201168228,-0.0069639804,-0.4100202024,0.1855685413,-0.1916394532,-0.5221064687,-0.2741031945,-0.0148120783,0.330452621,-0.0886402577,-0.1722717583,0.0328984633,-0.2435247302,-0.0217962097,-0.319501996,-0.0631714687,0.1740754396,-0.0999660268,-0.6227611303,0.0590884425,0.0145931868,-0.0334770307,0.0772041455,0.0882931799,-0.1070371196,-0.2859709263,0.2736419141,-0.0845417976,-0.144507423,0.6002654433,-0.2356881648,0.0307294838,0.6164119244,0.5096241832,-0.1603321433,-0.1509061903,0.2975326777,0.2535950243,-0.3025029302,-0.0875748768,0.5863990188,0.1962473691,0.0981076062,0.1826610118,0.3016443551,-0.3405370414,-0.0588858053,-0.6151734591,-0.1146372855,0.3677110076,-0.0197941232,-0.0412863903,0.0121511854,-0.0842735842,0.2014103979,-0.0830736682,-0.2267301381,0.1225779876,-0.0573123991,-0.0907820016,0.4527491927,0.1415142268,0.352938503,0.1707458943,0.1157718077,-0.0936380997,-0.0537631847,-0.1379546076,-0.0103543894,0.195990473,0.1063687876,-0.1523096263,0.041415181,-0.2955009937,-0.0407717489,-0.138650462,-0.1851912141,0.0187309794,0.1879850179,-0.089841485,0.1396678984,0.0272052325,-0.3407518566,0.1479416639,-0.1170674339,0.2890985608,0.2224602252,-0.2969357073,0.1697117984,0.0425906517,-0.0384227633,-0.0439409204,-0.2481618226,0.1228452548,0.4163686931,-0.2034342289,0.0929973423,0.1349324733,0.0209091753,0.4537694454,-0.2005510032,0.3117965162,0.3188026845,0.1028336808,-0.4711425006,-0.0322462916,0.4285385609,-0.097493574,-0.1177174374,0.132906422,0.0203044824,0.1275998503,-0.2521111369,0.2020098567,0.2492932975,-0.1423899829,0.1922290474,0.2257120907,0.1371825784,0.096766904,-0.0186322369,0.1762084216,0.0434417948,0.4835735261,-0.3364115059,0.1638944447,-0.1898630261,0.1845696568,-0.1459106505,-0.3627074361,0.112294741,0.3381086588,0.0842225626,0.0779021084,-0.3157329857,0.4324378669,-0.2558783591,0.2031059265,-0.2942218184,0.2279841453,-0.1795712709,-0.3268271685,-0.1394117326,-0.3754661679,0.0363077112,0.102945812,0.0061258394,-0.1819610447,-0.0557351559,0.1991254389,-0.1180707291,-0.2143520564,0.0058970228,0.0865517557,0.1842312366,-0.1216409281,0.0318917744,0.2790632844,0.0243090224,0.2111039162,0.0075357603,0.4232622981,0.4421955645,-0.3254863322,0.2190794051,-0.033355929,-0.1536405832,0.0618893243,0.2094311267,0.2220917046,0.2221371382,0.0808847398,0.0996713564,-0.084229216,0.4198633134,-0.0531385615,0.2680285275,-0.2802626789,0.0459184982,-0.4868516326,-0.0923358724,-0.5042527914,0.0925878286,-0.3706069291,0.2190076709,0.3753274977,0.1103682891,-0.0495603085,-0.0606172867,0.0213887505,-0.0646248385,0.4957806766,0.4861955643,-0.0153356707,-0.3640639484,-0.0198308527,-0.8502106071,0.0321466587,-0.0840492323,-0.0641455054,-0.1065228954,0.1282497644,0.2296151966,-0.0354772396,-0.019413596,-0.4398723841,0.158752352,0.0537591465,-0.1137024537,-0.2470296174,-0.254022181,0.2503528297,-0.0919794291,-0.0143097434,-0.0867225155,-0.4156889021,-0.0861692652,-0.2571044564,-0.0574235395,-0.2672639191,0.1049047932,0.2118829638,-0.1344621629,0.4210891128,0.0399807692,-0.267646879,-0.4160711169,-0.2614874542,-0.054323446,0.1965261847,-0.0077152173,0.4464056492,-0.3979943395,-0.0311583504,-0.3330067992,0.1670854837,-0.0031125462,0.0065231193,-0.3167001903,-0.1473855823,0.0537001081,-0.0788211152,0.2945427299,0.2342967242,-0.1015659645,0.1317185611,-0.0073740571,-0.0793456212,0.6290510893,-0.7522625327,0.0383949615,0.0358508267,-0.0148079982,-0.1958946735,0.0364187881,-0.6877139211,0.1022055596,0.264813751,0.1969151348,-0.2437797487,0.0641052127,-0.0187960658,-0.177250281,-0.083993949,-0.0926371217,-0.0633512959,-0.1978275478,-0.0351123028,-0.115508832]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2103","title":"citation, homepage, and license fields of `dataset_info.json` are duplicated many times","comments":"Thanks for reporting :)\r\nMaybe we can concatenate fields only if they are different.\r\n\r\nCurrently this is done here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/src\/datasets\/info.py#L180-L196\r\n\r\nThis can be a good first contribution to the library.\r\nPlease comment if you'd like to improve this and open a PR :)","body":"This happens after a `map` operation when `num_proc` is set to `>1`. I tested this by cleaning up the json before running the `map` op on the dataset so it's unlikely it's coming from an earlier concatenation.\r\n\r\nExample result:\r\n```\r\n\"citation\": \"@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n\r\n```\r\n\r\n@lhoestq and I believe this is happening due to the fields being concatenated `num_proc` times.","comment_length":43,"text":"citation, homepage, and license fields of `dataset_info.json` are duplicated many times\nThis happens after a `map` operation when `num_proc` is set to `>1`. I tested this by cleaning up the json before running the `map` op on the dataset so it's unlikely it's coming from an earlier concatenation.\r\n\r\nExample result:\r\n```\r\n\"citation\": \"@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n url = {https:\/\/dumps.wikimedia.org}\\n}\\n\\n@ONLINE {wikidump,\\n author = {Wikimedia Foundation},\\n title = {Wikimedia Downloads},\\n\r\n```\r\n\r\n@lhoestq and I believe this is happening due to the fields being concatenated `num_proc` times.\nThanks for reporting :)\r\nMaybe we can concatenate fields only if they are different.\r\n\r\nCurrently this is done here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/src\/datasets\/info.py#L180-L196\r\n\r\nThis can be a good first contribution to the library.\r\nPlease comment if you'd like to improve this and open a PR :)","embeddings":[0.1381235868,0.0136456965,-0.0630835369,0.3942780793,-0.0054810438,0.0727092922,0.2289254814,0.3846517801,-0.0693545863,0.0431411639,0.0773394853,0.6607301235,0.4834605753,-0.1113678068,0.0986291617,0.2540152669,0.0726539865,0.0371466614,0.1630943567,-0.0247602761,-0.1442041695,0.3476740122,-0.0284530371,0.0256110393,-0.2232681066,-0.0213754456,-0.2590197325,0.2612617314,-0.0107875718,-0.2939233184,-0.2302542329,0.3371426463,-0.3443260789,0.0409678519,-0.000113471,-0.1048205495,0.093062371,0.0950631872,-0.1159348041,-0.1097441167,-0.331269294,0.0067826402,0.1233062446,-0.3301422596,0.2031037509,-0.0964053869,-0.1437792182,0.101548247,0.3578597307,0.0314734094,0.1344466507,-0.3486812413,-0.0364137962,0.211799413,0.1670621932,0.5066618919,0.2788807154,-0.0225377176,0.0567879193,0.1372627169,-0.0372865982,0.3509612978,-0.0174391065,0.0689737424,-0.0889396742,0.0198654644,0.3704961836,-0.2638970912,0.4029415846,0.2036248744,0.1998370588,-0.2152676433,-0.064980261,-0.1508311033,0.2594078779,0.043362163,0.0116458386,0.1248427853,0.291568011,0.0952090696,-0.1523608714,-0.1289327741,0.0156486388,-0.2076678574,-0.1042188406,0.1388810277,0.0545036197,0.2547236085,0.4554472864,-0.175698638,-0.0094513241,-0.3732994497,-0.3630179763,-0.1488604546,-0.1156102791,-0.2586490512,-0.0468389168,-0.0697722435,0.1238707379,0.1188213527,0.1158270091,0.2533144355,0.0044774348,0.1921118647,0.4902763367,-0.1692830473,-0.016740907,0.1683241874,0.1627899259,-0.1561909616,-0.3621469438,0.28835392,0.2324282974,0.250054121,-0.2877439857,-0.2357924581,0.2499600947,-0.1283691972,-0.1634404063,0.164468646,-0.2285562903,-0.2335813791,-0.0743250772,-0.0127040986,0.2380670011,0.0311072487,-0.0816188157,-0.0761941373,0.0614849664,-0.1274145097,-0.1334026605,-0.0573373772,0.0249821022,0.0447401442,0.2108948976,0.195805952,0.2852685153,0.4387192726,-0.2492441833,-0.650668025,0.407578975,0.0235326048,0.0609327853,0.2766374946,-0.0646025911,0.2921264768,0.0561475046,-0.0985993445,-0.184970662,0.2049913704,-0.5609199405,-0.170074448,-0.0984943584,0.1963616312,0.069088906,0.0584161617,0.069195658,0.04308974,0.1906209588,0.0887872651,-0.0559138395,0.0977033228,-0.2419963181,-0.0013050016,-0.2179880887,0.4913587272,-0.2344311029,0.0301890373,-0.1352132708,0.1047382727,0.1158543229,0.3253320158,-0.300350666,0.4413999021,-0.1408083737,0.1248408854,0.1913560927,-0.2989719808,-0.1260942519,0.4549366236,-0.2909891903,-0.1740697026,0.1084959358,0.0151913567,0.1984761506,0.0299537629,0.2903322875,-0.1629746556,0.2847070098,0.1826170683,-0.107877031,0.0077407528,0.1597626656,0.1104981229,-0.5287647843,-0.1930352598,0.0745646432,0.1035661548,-0.0100893853,-0.1794373989,-0.1288513839,0.4532905221,-0.0080440436,-0.1855049878,0.2804617882,0.2361408919,-0.1617327482,-0.0047418061,-0.2121648937,-0.2374351174,0.0126130339,-0.3103972077,0.1363417208,0.057701651,-0.2258719057,-0.0005589699,0.1718885899,0.2167414129,-0.1872630268,-0.0663008541,0.034436278,-0.0065423255,-0.2294784039,-0.0580725074,-0.3690659106,-0.2251571864,-0.1929299533,0.2902039289,0.1001267433,0.0516589992,0.4500667751,-0.0769634619,-0.0083697289,0.1213895679,0.4090743363,0.2413639724,0.3329517841,0.2167107612,0.5112675428,0.1091349646,0.0661497861,0.0320899449,0.037671186,-0.3805708587,-0.3177200258,0.2866712511,0.0399540626,0.0024847279,-0.0771473721,0.0277876277,0.4658515155,0.1128012985,0.0328102298,-0.4600546956,-0.0184919033,0.6355430484,-0.2225127816,0.4609511197,0.0620354526,0.5856894851,0.4512516558,-0.3411588073,-0.128510341,-0.0126765436,-0.1237115413,-0.4660646021,-0.0280859452,0.3222884834,0.2815458179,0.1424501687,-0.074334234,0.1308515668,0.1566407979,-0.2218484879,0.3593808711,0.0051235263,0.0784253031,0.0761669502,0.2985602617,-0.076149419,-0.4390314519,-0.2105133981,0.5419505835,0.1828189194,-0.3398180008,-0.0741727725,-0.5753316879,0.2566515207,-0.208587423,0.0790688396,-0.2199468911,-0.3306314349,0.1915864795,0.5351461172,-0.1917527169,0.0248043165,0.2189728767,0.4054920971,-0.0478357635,0.0378665105,-0.2421166599,-0.2526839077,-0.0021718435,0.0007377458,0.1682926267,-0.1667028368,0.0985403582,-0.000967806,-0.0618157461,-0.6569470763,-0.385278821,0.2029548883,-0.2176854908,-0.0880903304,0.0604702905,-0.0229235943,-0.4931525886,-0.1360338181,-0.2491198331,0.4120309651,-0.3168639839,0.1458394229,-0.2303530276,-0.1695064455,-0.3084942997,-0.2869493365,0.0622559264,0.0428432189,0.2375161201,0.0092667229,0.2903403938,0.0147328591,-0.2407885045,-0.0194590893,-0.1292811632,0.1680930108,-0.6054062247,-0.3265309334,-0.1008286327,-0.1986559182,-0.0840509832,-0.0753735378,0.298283875,-0.0494761169,-0.1083335951,-0.5357647538,-0.1331084222,0.1364542544,0.0310637057,-0.2595835924,-0.026640201,0.322006464,0.3555791378,-0.2242719382,0.0075758635,-0.0670364723,0.220254451,0.3102791607,0.2186113447,-0.1807077676,-0.1428261697,-0.2503644526,0.4967601299,0.6802615523,-0.0872247666,-0.0812431723,0.0648198426,0.1408363879,0.0672508702,0.0875329748,-0.1258482635,-0.088083759,0.0161327422,0.4707184732,-0.0853134915,-0.0835384727,0.2374545634,0.2500020266,-0.3685618937,-0.3060157001,-0.0534653328,-0.1619236171,0.1360420436,0.1580480486,0.2378737628,-0.004199795,-0.0372933112,-0.0610609986,-0.0035315417,0.060745623,-0.2818319798,-0.3054351211,-0.3967858553,-0.3837317824,0.183228448,-0.1095387414,0.0491575859,-0.12818937,0.1125838682,0.0721788257,-0.1063116044,0.8338919282,-0.4579745829,0.1676812917,0.1676789075,-0.2081492692,-0.1239825189,-0.0899586678,-0.2645112574,0.2085199952,0.4078379273,0.4869430959,-0.1510308385,0.2392729074,0.1157794073,0.0050841006,-0.1674870104,0.0384286009,0.0659293607,-0.3831528723,-0.32291466,-0.1569055021,-0.145907402,-0.0177645404,-0.1131132245,-0.2006789297,0.2002695203,0.156857729,0.1397455186,0.23048684,0.1816809773,0.1829194725,0.1565915793,0.0130523946,-0.0504618585,0.1862033606,0.3360453844,0.1446378678,-0.0975744724,0.0652160272,-0.3386711478,0.0743773729,0.2682875991,0.2098905444,0.1782053858,0.0416063145,-0.1456061453,-0.1458850652,-0.2388183624,0.2699166834,-0.0264548976,-0.1145348698,0.0346521623,0.1618281007,0.2271596342,-0.3504941463,0.404194206,-0.2357999384,-0.1218328178,0.331246376,0.0987785235,1.0419565439,-0.2682436705,-0.2027228475,-0.2757574916,-0.3043020666,0.6576958299,-0.299523443,0.1170619726,-0.5067406893,0.0080067432,0.1176195964,-0.0710269883,-0.1293040663,0.0975941345,-0.4630860388,-0.2039552778,0.1638487577,0.3448338509,-0.1945906729,0.4422997534,0.3032673597,-0.1749410927,0.0702671185,0.134946987,-0.1902206093,-0.2142538726,0.0570903458,0.0328278057,0.1692997664,-0.2354005575,-0.3341638446,0.1725388765,0.0772554129,0.5955434442,-0.2048778236,0.006133595,-0.2287254184,0.0173462387,0.2061511725,0.2898864448,-0.2459372282,0.1247583702,0.1347133517,0.3820049465,-0.1349058896,-0.1198941767,-0.082163319,0.0021311166,-0.3452020586,-0.340783596,0.1976812035,-0.1150376722,-0.332924515,0.0917970985,-0.2368969023,-0.0227236897,0.089896284,0.2984758317,-0.4994901419,-0.2512393296,0.0576176569,-0.1796247661,-0.2918622792,0.3302533925,0.0361313783,-0.5448485017,-0.0778442398,0.3736341,0.3602990806,0.3496987522,0.1237992346,-0.0500634089,-0.0867902562,-0.3264315426,0.1801474988,-0.4874097109,0.019855259,0.1520214826,-0.0563341454,-0.1188793331,0.0048146504,0.0186504554,-0.2342497259,0.6001549959,-0.2011206448,-0.2790072262,-0.0927292258,-0.0859942064,0.0145782735,0.2944092155,0.3819886148,-0.0769348741,-0.2734094858,0.3077001274,-0.2819328606,0.3402886987,-0.1201231256,0.1973596811,-0.1631113291,-0.0544106625,-0.093654573,0.0871964246,0.1969876885,0.2647090256,-0.0703181177,-0.2248143405,-0.0555460863,0.1566103846,0.0011510482,0.13341254,-0.2993566394,-0.089667663,0.0987131894,-0.1832572818,0.1441937834,-0.1658317596,0.1817144006,-0.0507687926,-0.2867120504,-0.009659295,0.1697323024,-0.1586879939,-0.1601486802,0.3902057707,0.1244469881,0.2572150528,-0.1888333559,-0.1511443108,-0.4423004687,0.0501112603,0.0833741948,-0.0657706857,0.1648105681,0.1309522837,-0.3424821794,0.0470101126,0.2161830664,0.3447584212,-0.3605043292,-0.2973206937,-0.3167572916,0.1349867284,-0.2402786911,-0.4959181845,0.3653346896,0.0229368489,-0.1422833353,0.1555383801,0.3599428236,-0.3576847613,-0.0080193439,-0.1060036048,0.6999678612,-0.3942872882,0.2264445722,0.1856575906,-0.2435833216,-0.2395694554,0.1625288129,0.1568731517,0.1077072918,0.1365297735,0.1403701752,0.1975158453,-0.0283695366,-0.0221697092,0.4178748727,-0.1817902178,0.7519584298,0.4424462616,-0.0307774935,0.2606292069,0.404194504,-0.1422681957,-0.1419086158,-0.2161731124,-0.379558146,0.3629844785,-0.290050447,-0.0145787671,-0.0925829485,-0.1002374589,0.1920472682,0.3459124267,-0.0118032116,0.0370019153,0.4131368101,0.1536708623,-0.2489722222,-0.5125786662,0.3363232017,-0.3303619325,0.292304337,-0.4009962678,0.0602481216,0.0862205476,-0.1040332168,0.0416762829,0.1990960687,0.276317507,0.0948037729,-0.0074726795,0.18016693,-0.2811059058,-0.0544888079,0.0867213979,0.0831744745,-0.142089501,-0.3626761138,0.180274725,0.0474522933,0.0600221716,-0.1908129156,0.5934585333,0.0997420102,0.0297356006,0.2293608785,0.1729854196,-0.1588791311,0.1403053701,0.166198194,-0.2062401325,-0.3719598055,0.4716158211,0.1392489374,0.2531027198,-0.0488619879,0.1116376966,0.0022120157,0.5143994689,0.0144088268,-0.3647166491,-0.1657686085,-0.0649365261,-0.1773392111,0.129446879,-0.231402576,0.2702771425,0.2181691676,0.1608425528,-0.0447104387,0.0602462105,0.0802511573,-0.1226298213,-0.2659411728,0.0275378171,-0.3932188749,0.105107069,-0.0493949465,-0.172289744,0.014040038,-0.1084190458,0.6255809665,-0.008962119,-0.0429820791,-0.1360013634,-0.0931820348,-0.0505891405,0.0410645269,0.5731683373,0.3952974975,0.1402285397,-0.1946157813,-0.1487463117,0.0740425214,0.0988872722,-0.176059559,0.3726884723,0.2469548136,-0.1355239004,0.0200732723,-0.2488483489,-0.0950446203,0.0444405042,0.0166765023,-0.1245670766,-0.4153572917,-0.1074526906,-0.0459630862,0.1718653888,-0.2151598334,0.1535415798,0.0929630548,0.137956351,-0.0088243801,-0.3202521801,0.4107365012,-0.1049894392,-0.1741436273,-0.1306234598,0.5648161769,-0.5690383315,0.3015939593,-0.5685537457,-0.0466231257,0.3892926574,-0.2156589329,-0.3872373998,0.2250092179,-0.2500568926,-0.0817379057,-0.1571166813,-0.0951610282,0.189546749,-0.2068070918,0.1089525595,-0.1232032627]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2099","title":"load_from_disk takes a long time to load local dataset","comments":"Hi !\r\nCan you share more information about the features of your dataset ? You can get them by printing `my_dataset.features`\r\nCan you also share the code of your `map` function ?","body":"I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)","comment_length":32,"text":"load_from_disk takes a long time to load local dataset\nI have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)\nHi !\r\nCan you share more information about the features of your dataset ? You can get them by printing `my_dataset.features`\r\nCan you also share the code of your `map` function ?","embeddings":[-0.3500205278,-0.198214978,-0.0544512309,0.270072937,0.3246714175,0.0501476154,0.4278249741,0.1963976473,0.775203824,-0.0128923468,0.1478900909,0.5318043232,0.0868463516,-0.082643792,0.1153064743,0.0942703038,0.3582594991,0.2302624285,0.4533495307,-0.1887631267,-0.270419538,-0.0162078347,-0.136372,0.0524611287,-0.1678334326,-0.1337087303,-0.023513848,0.0352617837,-0.1384560019,-0.2863384187,0.0162087195,-0.056751959,0.2766021192,0.2574591637,-0.0001268705,-0.2512075901,0.0312588625,0.1771276891,-0.2025001496,0.244969964,0.3863898218,-0.46975106,0.2121178955,-0.3534870744,0.1457035094,-0.3091139197,-0.2663039565,-0.2810011804,-0.0431338586,-0.1522449851,0.0458118096,0.4439669847,-0.2780011296,0.0748594701,-0.1293984354,0.3457420468,-0.1573246717,0.0449594967,0.5400736928,-0.0728072003,-0.1051312536,-0.0802588388,-0.1870917529,-0.1056297943,0.3797875047,0.076493606,0.2742572725,0.1065707058,0.0830784366,0.1269641221,0.5899764299,-0.0041460837,0.0585617311,-0.2526357174,-0.1250291765,-0.2961908281,0.3886522651,0.0054056742,-0.2881904542,0.203135401,-0.0714381486,-0.0730997324,0.0088510998,0.2437554747,-0.1803190559,-0.0990389809,0.0195841622,0.1328806728,0.3168940544,-0.16144602,0.0230974704,-0.0082517387,0.1234517619,0.3986281157,-0.4098905623,-0.2113994658,0.0400182419,0.0835748315,-0.1435200572,0.1120383367,0.1514651477,0.2489472926,-0.1630488485,0.0016306768,0.2422059625,0.3972100616,-0.1217968985,0.1301053613,0.0731808767,-0.3419057131,-0.0699739307,-0.0560454279,-0.2259919494,-0.0106498972,0.5112823248,-0.4784032106,-0.2339437306,-0.1448340118,-0.0333806463,0.2067790329,0.4317516983,-0.0641698688,0.2521469295,0.2410562187,0.0641919672,0.5306972861,-0.1165344194,0.0117174303,-0.1896087825,-0.1619036347,-0.0651431233,-0.307161361,-0.1684478372,0.1598518491,0.0181163065,0.2619099617,-0.1174448505,0.1726414263,-0.248095274,0.0674418509,-0.1035650074,-0.1866309494,0.2897593379,0.1619751304,-0.0831962973,0.2204560041,0.1658516973,-0.2604821622,-0.4331320524,0.364184618,-0.1685276777,-0.3845943213,-0.4457713068,0.0039892276,-0.2279209942,-0.0610848553,-0.3854816258,-0.1208797842,0.1200976595,0.0442280546,-0.0343466476,-0.152447626,-0.144491598,-0.0993630439,-0.0391864553,0.1216116995,-0.3704289198,0.2121679336,-0.2364278734,0.2569226325,0.3447540402,0.6626860499,-0.2571009398,0.3425361812,-0.2578446865,0.3158119321,0.548707664,-0.0578072369,-0.3862134218,0.7333583832,-0.1003699228,-0.2248929739,0.0265344549,0.5490338802,0.2924337387,-0.0852065608,0.1003952026,0.6558295488,0.1512116641,0.210539341,-0.391095221,-0.0279238429,0.0919300616,0.4662360847,-0.4127826095,0.158823669,0.1538113952,-0.0441682301,0.1545038074,0.0907559097,-0.0659692585,0.2432232797,0.2831676602,-0.0318482891,0.0112171844,0.2562739551,-0.1681343913,0.1245542392,0.464259088,-0.2245752364,0.041568771,-0.1681670696,-0.344547987,0.3147711158,-0.1054795831,0.6033324003,-0.1821069568,0.0020355629,0.0804636702,-0.1432267725,-0.149512291,-0.1174869612,-0.2531777024,-0.1158315986,-0.0902426466,0.051841706,0.1564155817,-0.0151104908,-0.34988603,-0.0625957251,0.0574983582,0.0494381301,-0.2109956443,0.284044981,0.1521351486,0.0479029864,-0.0496525988,0.1704577059,0.1336129755,0.0028490291,0.3439742923,-0.0021683658,0.2081943601,-0.2890040278,-0.1218547896,0.4259040952,-0.1942901313,0.2047604173,-0.2710195482,-0.0709528103,0.4258694351,0.1262152195,0.1576503515,-0.0164163448,0.1865303963,0.3095088899,0.2882080674,0.2737433016,-0.2397418618,0.261765182,0.2991085947,-0.3452146351,-0.0612501949,0.1969376504,0.0367677696,-0.3007032275,0.0062167998,-0.1039675698,0.3100560308,0.1159394756,0.0958434045,-0.3176046908,-0.0347823203,-0.144626379,0.0870060548,-0.1001754478,-0.0837553814,0.1299334466,-0.0537791699,-0.1364710182,-0.3259738386,0.0318153277,0.1118117645,0.3960193098,-0.3259372711,0.1538781673,-0.090751715,0.0990272984,0.0559910685,-0.25811252,0.1266683936,-0.2027240843,-0.2237585932,0.2324294895,-0.026160188,-0.1411309391,-0.0124721266,0.3762325644,0.1917368919,-0.4314824045,-0.3557509184,-0.2873743773,-0.1534438282,-0.1426837295,0.4407370687,0.1697454602,0.2259090394,-0.0629057214,-0.1602994651,-0.2504196167,-0.1463385671,-0.0980741233,0.0538785383,0.1577214897,-0.1832669824,0.3381577134,-0.5822999477,-0.032939516,0.2617312074,-0.3245382309,-0.2410393804,0.3782650828,-0.1310499758,-0.3202217221,0.0137208216,0.113514863,-0.1331354976,-0.2363513261,0.2219308913,0.0367419086,0.1437613368,-0.2337397337,0.1127445847,0.0259840637,0.4734171033,-0.2453035265,-0.1945905983,-0.27755934,0.6613589525,0.0028795663,-0.2596795559,-0.0730992258,0.1196758822,-0.0998840705,-0.0242242701,-0.5121884942,0.2339362353,-0.4204593897,0.4694085717,-0.2292629778,0.0604310967,0.1301382184,-0.2692603469,0.0377081782,0.2235512882,-0.1082632765,0.2056974918,0.1367424875,0.083491981,0.0862882137,0.6676974893,-0.237845704,0.3274142146,-0.2057073414,-0.1607969403,0.3523729146,-0.0135515574,0.1972753406,-0.3578559458,-0.1765050739,-0.2438223809,0.1217150092,-0.3225125074,0.1974404752,-0.039084468,-0.3568358123,0.023487296,-0.0529188178,-0.0557643399,-0.0565586723,0.2541282773,-0.0024458254,0.0846787244,-0.1358803064,0.2029365748,-0.3828940094,-0.2328350991,-0.0731363371,0.0034161741,0.1737784892,-0.249621585,-0.6529747844,-0.0044412292,-0.4316208661,0.1233486012,-0.1962352246,0.4478288591,0.0173853245,-0.2779054046,0.429597497,0.007992126,0.605700016,0.2846139073,0.1263311356,0.1792880595,-0.173865214,-0.6647119522,0.0659500062,0.2427068353,-0.0304169673,0.3263396025,0.7139401436,-0.0432582833,-0.1389229298,-0.1624472737,0.3244585693,-0.2122953981,-0.487318933,-0.1144833937,-0.3401408494,-0.0956713557,0.1660429686,0.1704065353,0.4786660969,0.421318233,-0.0492857657,-0.2140697092,-0.1068122238,0.0743908808,0.0897785053,0.3559323549,-0.0836096555,0.4325259626,-0.3090108931,0.1667516083,-0.1315863431,0.4047930241,-0.1201718077,-0.0328734554,0.3473880887,-0.2058636993,0.0278259944,0.3513396084,0.2020438462,0.0419761576,-0.1166421697,0.1926029176,-0.2225783318,-0.0224438086,-0.193552196,-0.0132624833,-0.0181619283,-0.1326584816,0.2976230681,0.0367987342,-0.0336684883,0.2161800861,-0.1391735673,0.0804776773,0.7127652168,0.2099802792,1.1476422548,-0.4765789211,-0.0367588103,0.1055497229,0.1875545382,-0.0967544988,-0.1556256562,-0.3527749479,-0.3770470023,0.1047224104,0.1252688915,-0.2228828371,-0.1351712197,0.4446427524,0.1457490623,-0.127387628,-0.2085412443,0.324582696,-0.3329761922,0.4716747701,0.0304985624,-0.4644049406,-0.5360887051,-0.0256496109,0.002109315,-0.0625469536,0.0134712961,0.1504437625,0.0280224737,-0.1613013148,-0.0555919185,-0.1683797091,0.1311299056,0.1593941301,0.2132583559,-0.242970109,-0.0600926168,0.4783608317,0.3922135234,-0.0488339961,-0.0285530984,0.1756227016,0.1071218327,0.1411294788,0.0880615488,-0.2471680343,0.0647240728,-0.1228903159,0.0296013989,-0.1266628504,-0.1492752433,-0.4715046585,-0.110463351,0.1728015244,0.3435623944,-0.1678790748,-0.2620112002,-0.280431211,0.1687634289,-0.0469870344,0.0108307814,0.3402671218,0.1919054985,0.2474282533,0.000438992,-0.2470659912,0.0262290146,0.3180592656,0.0175307859,-0.2553491294,0.2725578547,0.0369648598,-0.2449594587,-0.2093434632,0.3605388403,0.2967856526,-0.4077206552,0.0861006752,-0.0421442352,0.0005220664,0.0812188983,0.0793014616,0.1944931746,0.0004830987,-0.1094093919,-0.2380563468,-0.5587601066,-0.053292267,0.1740911156,0.0981364399,0.1941781342,-0.4832066,-0.2332377732,0.0859462172,-0.1355979741,0.0622784719,-0.0595663078,0.1091916785,0.0401640907,-0.0415792055,0.0677755699,-0.1500801146,0.0465443507,-0.1525708735,-0.0100934366,-0.0801539794,0.072009027,0.2114048302,-0.0165333953,-0.3091340661,-0.2002136707,-0.4752689004,-0.1812930107,-0.3601987362,-0.1456623077,0.0568097681,-0.0676269233,-0.3485494256,0.3751345277,0.1153298318,-0.0378486924,0.3707078993,-0.0508098863,0.4025710821,-0.0554909445,0.119132407,0.2175970227,-0.1941404194,-0.3011140823,0.1783866584,0.3650397658,-0.1530926824,0.3515724838,-0.2173791826,0.0392777398,0.2798125446,-0.0359385051,0.1138463169,-0.1738660336,-0.2687506676,0.2702850103,0.0583915971,-0.0130299162,-0.2406965345,0.005447974,-0.0960762948,-0.2235891223,0.2846253514,0.4408999979,-0.1794555485,-0.1238967627,0.0501795597,0.5539265275,0.0341320746,-0.452645272,0.5829268694,-0.0826049745,0.0033456364,0.1782965958,0.214112103,-0.0035183281,0.5654761791,-0.2079436183,0.0983969346,0.3606002331,0.5861619711,-0.388065964,-0.3600417674,0.3370532691,0.2944210768,-0.2477842122,-0.1721238345,0.1744413078,0.2293726802,-0.1746055484,-0.1171965972,-0.210152939,0.290843308,-0.2680514753,0.0505828783,-0.3962656558,-0.003801381,-0.0263498332,0.1709143221,-0.2733541429,0.0057038832,-0.2857446373,0.3377734125,0.043734502,0.0874258429,0.0678672418,0.0516937338,-0.0261326935,-0.4181884229,0.0240778904,-0.3533783853,-0.0266269892,0.2478286177,-0.2746102214,0.371930927,0.5034902692,-0.0068528508,0.0553242192,-0.0133310808,-0.0688841715,0.0226926655,0.1387315094,-0.1509943455,0.2419558018,0.1769801974,-0.1034574062,-0.0439834408,-0.4177389145,-0.1573689729,0.4149636924,0.09254843,0.3043407202,0.0882519782,-0.3174013197,-0.2717196941,0.0808462426,-0.0330294743,-0.1450405568,0.3015883565,0.0379135832,0.162357688,-0.0567283146,-0.0364083312,-0.3725741506,0.5649231076,0.3192876875,-0.3951666653,-0.1880398691,-0.3771695793,-0.6647297144,0.0834212154,0.0558312759,-0.0527230799,0.4076804519,0.1982892305,0.4616668224,0.0610228665,-0.1537917107,0.0770152286,-0.1567260772,0.3889112771,-0.0518679284,0.0534201488,0.0339904651,0.2281391919,-0.2226357609,-0.4697470665,0.4194073677,0.2180146128,-0.0897895992,-0.1330071539,0.0592485778,-0.3414674997,0.4256179631,0.3861466944,0.0256975014,0.0922034755,0.3125084937,0.0879867077,0.3015055656,-0.0270412974,0.0697823539,0.242068395,-0.0561338328,-0.0393762849,-0.0404638983,0.4130625725,-0.173322916,-0.0444145165,-0.2778540254,-0.1421041191,-0.1402700543,-0.0258711223,-0.2182967067,0.2299895287,0.267699331,-0.023984395,0.2052936703,0.203226164,-0.2472174913,-0.0509744808,0.2124585509,-0.0083215749,-0.1552095413,0.0224937629,0.1243808419,0.2134562284,0.1865949184,-0.3039721847,-0.1745001674,0.2332245559,-0.1703486294,-0.0961515605,0.387635082,-0.2346216738,0.1899157017,-0.171601817,0.2427430302,-0.0772559419,-0.0695405751,-0.4443511367,-0.2711589932]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2099","title":"load_from_disk takes a long time to load local dataset","comments":"It is actually just the tokenized `wikipedia` dataset with `input_ids`, `attention_mask`, etc, with one extra column which is a list of integers. The `text` column is removed during tokenization.\r\n\r\n```\r\ndef add_len_and_seq(example):\r\n end_idx = example['input_ids'].index(SEP)\r\n example['actual_len'] = end_idx-1\r\n seq_len = len(example['input_ids'])\r\n \r\n\r\n example['seq'] = [PAD_ID] + [np.uint8(example['some_integer'])]*(end_idx-1) + [PAD_ID]*(seq_len-end_idx)\r\n \r\n return example\r\n```\r\n","body":"I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)","comment_length":51,"text":"load_from_disk takes a long time to load local dataset\nI have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)\nIt is actually just the tokenized `wikipedia` dataset with `input_ids`, `attention_mask`, etc, with one extra column which is a list of integers. The `text` column is removed during tokenization.\r\n\r\n```\r\ndef add_len_and_seq(example):\r\n end_idx = example['input_ids'].index(SEP)\r\n example['actual_len'] = end_idx-1\r\n seq_len = len(example['input_ids'])\r\n \r\n\r\n example['seq'] = [PAD_ID] + [np.uint8(example['some_integer'])]*(end_idx-1) + [PAD_ID]*(seq_len-end_idx)\r\n \r\n return example\r\n```\r\n","embeddings":[-0.2723700404,-0.1468664557,-0.0388626195,0.1874405742,0.4000378251,0.0453446023,0.4641442895,0.2294565886,0.6530973911,-0.0394133255,0.2031130493,0.5122695565,-0.0241835993,-0.1652990133,0.1699289083,0.1053498611,0.3435818553,0.1828202009,0.5284671783,-0.2227723897,-0.2088466734,-0.0600212328,-0.0657849684,-0.0276850443,-0.2205491662,-0.0452075787,-0.0313016213,0.0227651168,-0.1343157291,-0.3194649816,0.0928841531,0.0193027686,0.2287545651,0.2781813145,-0.0001281752,-0.1722031683,0.0046071191,0.1535199732,-0.1430971771,0.2643969953,0.4690803885,-0.5940458179,0.2383883148,-0.35034284,0.0618379153,-0.2827260792,-0.3401739299,-0.2147907913,-0.0075521097,-0.1195013151,0.0158577468,0.361687541,-0.333750397,0.1539883912,-0.0743322521,0.1938998699,-0.1434206069,0.1258124262,0.4729337394,-0.1927229464,0.0029671367,0.0226153564,-0.2509937584,-0.0965019763,0.3857630193,0.0741025284,0.1951773018,0.0483410358,0.0506101027,0.1477310807,0.5389365554,-0.0009465857,-0.0577600747,-0.3210890889,-0.0554644354,-0.3969578147,0.3212765455,0.0301397387,-0.2767955363,0.0759613663,-0.1086108908,-0.0788263232,0.0298109204,0.2188173085,-0.2157344967,0.0422307178,0.0866941735,0.1933327317,0.3188702464,-0.108380802,-0.098751843,-0.0084894449,0.1952891648,0.4358851314,-0.32837376,-0.2487675399,0.0326809287,-0.0019629954,-0.1444237828,0.1964237988,0.0889205411,0.2106192708,-0.1054590121,0.0051103127,0.2394284308,0.4608980417,-0.2296869457,0.1211880669,0.0778399855,-0.3174993396,-0.1782213151,-0.0916159153,-0.2083225548,0.0359108225,0.4978763759,-0.5051928759,-0.2428243905,-0.1187323183,-0.0430596881,0.2367561013,0.3178033829,0.0038013018,0.2405851185,0.2986124158,0.0302921738,0.5268796086,-0.0790837556,0.1430350542,-0.2173659503,-0.0234914441,-0.0706594512,-0.2602777779,-0.1074772254,0.1086402237,-0.0289179366,0.3335449994,-0.0739851072,0.1510807425,-0.1349657774,0.0091545433,-0.0671710148,-0.1628105342,0.2305504978,0.207768634,-0.165542528,0.2659240365,0.2583853602,-0.3331991136,-0.4613665044,0.2853639424,-0.1705411524,-0.4137839377,-0.3964413404,-0.0067317081,-0.2626550198,-0.0938732624,-0.4485415518,-0.040241316,0.1792416871,0.0432296805,-0.0430105366,-0.2162176222,-0.1451558769,-0.1301585585,0.0579306632,0.0728645623,-0.2945594192,0.1127770171,-0.2678593099,0.3557194471,0.3770870268,0.7180876136,-0.2919610441,0.4202105999,-0.2172201127,0.2997386754,0.5369105339,-0.1884544194,-0.4042335749,0.6187259555,-0.1343872249,0.0445357151,0.0361252315,0.4975116253,0.2584006786,-0.0394756347,0.1837880462,0.6960242987,0.1099874303,0.2276614457,-0.4064319432,-0.0267873649,0.219917953,0.3522608578,-0.3312678635,0.1514369547,0.0653631762,0.0905347392,0.2092308849,0.061622601,-0.1017940342,0.2813984752,0.2582951486,0.1643486321,0.0434327312,0.2720756233,-0.1440769434,0.2002818286,0.5107887983,-0.0763214156,-0.0500876233,-0.2480376959,-0.3892806172,0.2931813002,-0.1443336755,0.3294907212,-0.1960772425,0.0404949859,0.130680263,0.0196183752,-0.1077524498,-0.273044616,-0.2539159358,-0.031821534,-0.0143755488,0.0497076549,0.1867201775,-0.0047953371,-0.4482590258,0.1168485582,0.0564297922,0.0313600414,-0.2220863551,0.2688148916,0.2077786177,0.0680243,-0.0495331548,0.2058762163,0.1093481556,0.0337096527,0.2628691196,0.0712591782,0.1975321621,-0.2217781097,-0.1914296001,0.5089946389,-0.0855100527,0.248140797,-0.2729073763,-0.0396678485,0.4773667157,0.1242774874,0.0985501483,-0.025013186,0.2052785456,0.2778957188,0.2757321,0.1117201671,-0.1666743606,0.271894753,0.2947527468,-0.2841126323,-0.03043513,0.2060741037,0.0274871588,-0.3105518222,-0.0116950702,-0.1333327889,0.2517815828,0.1176275611,0.1077013463,-0.3004761338,-0.0617363006,-0.0989977047,0.17740722,-0.0865197703,-0.1013619453,0.1303023398,0.0194883365,-0.0826523304,-0.3771759868,-0.0525450744,0.0686625987,0.4366832376,-0.2205613852,0.2747827768,-0.0571823306,0.1997994334,-0.0164392851,-0.301741153,0.0251299702,-0.3157307804,-0.2038689554,0.2908666134,-0.0447209217,-0.1348048449,-0.068859227,0.2550409436,0.2128171325,-0.4541468024,-0.2794964612,-0.3061749935,-0.1113633513,-0.1972857565,0.3697757125,0.0489766374,0.3008130193,-0.0138004106,-0.1955286711,-0.2244669497,-0.341160208,-0.1097940505,-0.0475998111,0.2139697522,-0.1861353517,0.3008183837,-0.5481994748,-0.0654698014,0.3417084217,-0.2881565392,-0.1716946065,0.3768146932,-0.0064205378,-0.1605555713,0.1009441391,0.1047660187,-0.1269279718,-0.1898217648,0.2512330711,0.0706498772,0.1390314847,-0.2185506821,0.0698025078,0.021884976,0.4788020253,-0.2052250654,-0.1983879507,-0.1957680881,0.5628789067,0.0449668951,-0.183082208,-0.1186386943,0.0367398113,-0.0935700461,0.0155285075,-0.5713308454,0.1604190171,-0.3806986213,0.4317447245,-0.1833308488,0.0472297631,0.1662706286,-0.2633866966,0.0482754558,0.1891607046,0.020515671,0.1094460562,0.0715712905,0.0302696917,-0.0074167978,0.7050042152,-0.1910776794,0.5004677176,-0.2420652956,-0.0491507463,0.3664191365,-0.1120703742,0.2564029992,-0.3437128961,-0.2357252389,-0.2404978573,0.0993407369,-0.3650355041,0.180310756,-0.0105273798,-0.3311646879,0.0819814801,-0.0493068248,-0.019012861,-0.1505444199,0.347838372,-0.1515969783,0.0368440263,-0.1103061736,0.2615605593,-0.4225663245,-0.242309913,-0.1179583371,0.0207945947,0.1237534061,-0.2384304255,-0.7123931646,-0.1228405908,-0.5423576832,0.1210410967,-0.1256302744,0.4711712599,0.0812951699,-0.3076661825,0.4302410185,-0.0289578289,0.6507672668,0.1201318204,0.096233651,0.1961339861,-0.2613889873,-0.7808228731,0.0258691441,0.1718291044,0.0333679318,0.3404671848,0.7582909465,-0.079474017,-0.1321242452,-0.1942912936,0.4059050381,-0.2330976278,-0.286664784,-0.2519877553,-0.3326835036,-0.2424204797,0.2168319523,0.1198684275,0.5127633214,0.365509212,-0.0541317388,-0.2658092976,-0.0786047876,0.1119235381,0.1096714735,0.3797810972,-0.0863723308,0.3377001584,-0.2857827842,0.0629304945,-0.1840693057,0.4262422919,-0.1851895899,-0.0862930566,0.206556797,-0.0905981734,0.0726824924,0.4006850123,0.2137317657,0.0808738843,-0.1739758998,0.2801629305,-0.350102663,0.0860043168,-0.0204656478,0.0299961809,-0.0342003889,-0.2859373093,0.3274013996,0.0167085044,-0.0674652532,0.2595994473,-0.09517847,0.007013665,0.7409731746,0.202038601,1.2241344452,-0.4361752868,0.1046029031,0.1566031724,0.2955697775,0.0488455296,-0.2537310719,-0.2328999341,-0.3573130369,0.1385998875,0.1361425966,-0.1821783781,-0.1807617694,0.3662807643,0.1105945334,-0.0953259617,-0.3047887683,0.3298876286,-0.3755404651,0.4509918094,-0.038976334,-0.3950835168,-0.5488098264,-0.0458752327,0.0927856788,0.0167337358,-0.0419050865,0.2078908384,-0.0197340492,-0.2473613024,-0.156085968,-0.1223901138,0.1882690787,0.3284366727,0.0779665485,-0.3352366686,-0.202718392,0.456858784,0.3226767182,0.0175197348,-0.0467572697,0.2161892951,0.0357559063,0.0785137936,0.1623002589,-0.2645243406,0.1143185347,-0.0748121291,0.0375724956,-0.1490096748,-0.2162973434,-0.592670083,-0.0849835649,0.1830936521,0.3696853518,-0.2103634477,-0.2464728653,-0.1820779592,0.1899179667,-0.1519608945,-0.0263233781,0.2444147021,0.1326882839,0.2322719842,-0.0355084911,-0.2975770831,0.0561396331,0.3613017201,0.1219026372,-0.1763508618,0.3943614364,-0.0458821915,-0.2334232777,-0.2297955155,0.2831092477,0.3257548809,-0.3678573072,0.2020450085,-0.0597112551,-0.0261540767,0.0612433329,0.1117014959,0.1226689145,-0.0526755638,-0.2048232853,-0.2920001447,-0.5180852413,-0.0142699471,0.2212660164,0.1805761009,0.1349380463,-0.3628823757,-0.256118834,0.1121544018,-0.1135133877,0.0301676299,-0.1033240855,0.1016057283,0.0881103501,-0.0794898346,0.1899868846,-0.1210955381,0.0247975234,-0.1655266434,0.0063721654,-0.0626746491,-0.0018176099,0.23469612,-0.0313989297,-0.2769498527,-0.1427206397,-0.3840763569,-0.1196034104,-0.4405585825,-0.0466870219,0.0308999512,-0.0530009083,-0.4075684249,0.3622803092,0.1974085271,-0.0273239706,0.2847129405,-0.169237107,0.3989355862,0.0205528978,0.0156394597,0.2750974,-0.1561070234,-0.31680125,0.187391445,0.4050490856,-0.1289350688,0.3607321978,-0.1552073061,0.0698475242,0.2845456302,0.0124520166,0.2120549381,-0.1705856472,-0.3054560721,0.3490548134,0.0346519165,-0.0181310009,-0.3050706089,-0.0451132432,-0.1514684856,-0.2121546417,0.3869555593,0.4360878468,-0.1036216319,-0.2026675195,0.0185809694,0.5638002753,0.0744235888,-0.3574274778,0.5144121051,-0.1318355948,-0.0445704497,0.1361664683,0.1590793729,0.0855830535,0.5236283541,-0.1440825462,0.1383867115,0.2728661299,0.4584133029,-0.4138359725,-0.3418091834,0.3759522736,0.3029159009,-0.1723515689,-0.0599532016,0.1050039679,0.2829161286,-0.2486313879,-0.0871200562,-0.2164557725,0.4080291986,-0.2787197828,-0.0049716509,-0.2101798803,-0.0080127828,0.0091441013,0.1979605108,-0.2225671262,-0.0109473364,-0.3567083478,0.2962246537,0.038165316,0.0487785488,0.1087220386,0.0426627249,0.108242169,-0.4354820549,-0.0357793905,-0.2812235653,-0.0054105702,0.221860677,-0.2468221635,0.4374468625,0.4777419567,-0.0077329301,0.0011962686,-0.0068760468,-0.016529223,-0.0608338639,0.1732795388,-0.1783579886,0.2844756544,0.1485008448,-0.1241733879,-0.0721785873,-0.4527173042,-0.1113945991,0.367922157,0.1252459437,0.3018085659,0.1230110675,-0.2424144298,-0.3305615187,0.1317330748,-0.0461976342,-0.1522235572,0.1501564384,0.0900423154,0.1735492945,-0.0438108072,-0.0419203751,-0.2884946167,0.5451473594,0.2815956771,-0.3723652065,-0.2056159973,-0.3782559037,-0.6515106559,0.0716660395,-0.1074581295,-0.1714022607,0.3680438995,0.2453299165,0.4008659422,0.0430511311,-0.141557768,0.0269901101,-0.1825286895,0.3073928058,-0.0530224182,0.1137599647,0.0189639796,0.1804411858,-0.1774557233,-0.5137280822,0.3781629801,0.2077545375,-0.1100399271,-0.0539736971,0.0188240074,-0.4160960317,0.3677490056,0.4663529992,0.0675621331,0.1766430438,0.2604014874,0.0236685853,0.2030624002,-0.0768174231,-0.0392013863,0.2884309292,-0.1165913939,0.0287643485,-0.0811061338,0.3949267864,-0.1959096342,0.0813039616,-0.1703322828,-0.2662628889,-0.1360894144,-0.0353827737,-0.1896097809,0.3324862123,0.1649116725,0.1120360717,0.2056087852,0.2885064185,-0.2644026875,-0.1547446847,0.3250331283,-0.0914279148,-0.19286865,-0.0104169771,0.1300896108,0.1925909668,0.1720910966,-0.3801366091,-0.1749806553,0.2354858965,-0.0609318428,-0.2268810719,0.3500665724,-0.2122201324,0.0729692951,-0.1263055801,0.2208830714,-0.1411448121,-0.063737236,-0.3829962015,-0.2339343429]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2099","title":"load_from_disk takes a long time to load local dataset","comments":"Is `PAD_ID` a python integer ? You need all the integers in `example['seq']` to have the same type.\r\nDoes this work if you remove the `np.uint8` and use python integers instead ?","body":"I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)","comment_length":32,"text":"load_from_disk takes a long time to load local dataset\nI have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)\nIs `PAD_ID` a python integer ? You need all the integers in `example['seq']` to have the same type.\r\nDoes this work if you remove the `np.uint8` and use python integers instead ?","embeddings":[-0.2943402231,-0.2248792797,-0.0571161397,0.303353399,0.3156488538,0.0497011691,0.4435780048,0.2428758591,0.6799733043,-0.018701205,0.1529660374,0.5113555789,0.0885321572,-0.1085035205,0.13383618,0.0986105055,0.3429254591,0.206077233,0.4759643972,-0.1658245474,-0.2517735362,0.0001301761,-0.1306847334,0.0234819558,-0.1588277519,-0.0829920173,-0.0751809031,0.0350681543,-0.1523588151,-0.3155409992,0.0216222703,-0.0448314697,0.2884912789,0.2051530778,-0.0001264279,-0.2548462451,0.0216278266,0.1173048094,-0.2475825846,0.2874677181,0.4266411364,-0.4523167014,0.2212568372,-0.3689641058,0.1373716593,-0.3382209241,-0.3590539694,-0.2098047584,-0.018680945,-0.141103223,0.0444991328,0.4352433681,-0.2819193006,0.0345782079,-0.0528538078,0.2976847291,-0.163785845,0.0466387942,0.5985652208,-0.072557874,-0.1019449607,-0.0817549601,-0.2182448655,-0.1018448174,0.3846515119,0.074810721,0.2570776939,0.1627735496,0.0662124231,0.1427181959,0.5887855291,-0.0352816582,-0.0009134836,-0.2447482347,-0.0952465832,-0.2890897095,0.3673810363,0.0128193377,-0.2871285379,0.1850773394,-0.0602000766,-0.0723918676,-0.0501539148,0.2330685556,-0.1294519603,-0.0757532194,0.0434352942,0.1163351238,0.3337874115,-0.1512931287,0.0617027804,0.0351881124,0.1166585088,0.3841658831,-0.4037293792,-0.2078680247,-0.0331366211,0.0053695031,-0.1448202431,0.127843082,0.2320950925,0.2485832423,-0.0449300297,0.0197506305,0.239918828,0.3678704202,-0.1198359132,0.1469492614,0.0768400431,-0.3085559905,-0.028876232,-0.0406091362,-0.2590741217,0.0038079319,0.4785634875,-0.4682095945,-0.2703816295,-0.1632350236,-0.0524667203,0.2429694831,0.395481348,-0.0433485284,0.2423650175,0.275617063,0.0462449193,0.4945330322,-0.0492988527,0.0374574512,-0.1581906825,-0.1149220616,-0.0604098663,-0.2971879542,-0.1712874323,0.1489873827,0.0418655127,0.3508808017,-0.1205050945,0.189094469,-0.2329151481,0.0578361601,-0.0910502151,-0.1327035278,0.2696327269,0.1806907505,-0.0991710275,0.2365833819,0.1698706299,-0.2017394751,-0.4431205988,0.3833531737,-0.1958324909,-0.3932256699,-0.434984237,0.0218760427,-0.2119769305,-0.079288736,-0.4514289796,-0.1795584261,0.1182682887,0.0589968935,-0.01975137,-0.1308733076,-0.0821726024,-0.092337139,-0.047403086,0.0929241329,-0.3962214887,0.1668631732,-0.2080180347,0.2890934348,0.3477072418,0.622900784,-0.2996450961,0.3621346653,-0.2523355484,0.297699362,0.5223905444,-0.1004187688,-0.364366293,0.718966186,-0.0472505428,-0.1865952313,0.1076065302,0.5277528167,0.3615674078,-0.0806183591,0.1315562874,0.6244168878,0.1654666513,0.2055848539,-0.4061889648,-0.0612445734,0.0919046775,0.4283311367,-0.3837227225,0.1274905354,0.1401904672,-0.0290596485,0.1692350358,0.044532001,-0.0756354332,0.2814265788,0.2664147019,-0.0544777848,0.0677250996,0.2265750617,-0.1943380684,0.151138857,0.3955882192,-0.1846632063,0.0356449001,-0.2098936439,-0.3576269746,0.3173824251,-0.0696600825,0.5633525848,-0.1680370122,-0.016169386,0.1312579364,-0.1643915176,-0.1563707441,-0.1088241413,-0.2634977102,-0.0942779183,-0.0907814428,0.0605390705,0.1737845689,-0.0462965369,-0.3290225565,-0.0069326889,0.0490829721,0.0580228567,-0.1874863356,0.3025074601,0.2110293657,-0.0094984965,-0.0084419586,0.2171442509,0.1239149198,0.0457701124,0.2926034331,0.0471671373,0.1885788441,-0.262453109,-0.1158252656,0.4826082885,-0.2119685411,0.263625443,-0.2107875347,-0.0570614114,0.4025437236,0.1707692742,0.1318585873,-0.0161363874,0.1807725579,0.3143539429,0.2750582695,0.2520087957,-0.2294910699,0.2636447549,0.2475572824,-0.3291889429,-0.0721794665,0.2215183824,0.0327369049,-0.2925549448,0.0452846065,-0.0746015087,0.3398680687,0.1416780651,0.0331318006,-0.3377315402,-0.0807491615,-0.1799663305,0.1585643589,-0.0963617563,-0.166472137,0.079352513,-0.1049145088,-0.1405676305,-0.3333320618,0.0431477353,0.1139277518,0.4340325892,-0.3360694051,0.1819747686,-0.1397999823,0.1006547958,0.0214313176,-0.3120059669,0.0983556286,-0.2060772777,-0.1601317227,0.2585887313,-0.0568330735,-0.1273800582,-0.045708131,0.4071220458,0.2142862827,-0.4044517875,-0.380630672,-0.3044311702,-0.1384850293,-0.1443670392,0.4448207021,0.1631069034,0.2551799715,-0.0569447391,-0.1024075821,-0.268809706,-0.1337063313,-0.1084033474,0.0277535208,0.1366552711,-0.1401946545,0.3482026458,-0.5923082232,-0.068254672,0.2889491916,-0.3040878475,-0.2487547845,0.4316572547,-0.0898168162,-0.3463097513,0.0212814119,0.1075281724,-0.1065266207,-0.2346157283,0.1980972588,0.0223735832,0.1970598102,-0.2903107703,0.1200641915,0.0475447774,0.4194039106,-0.2496537417,-0.1606363505,-0.2298171371,0.6143157482,0.0052384324,-0.2826807201,-0.0352879651,0.1667377949,-0.0689897686,-0.0044366303,-0.5236109495,0.2110106647,-0.3914633393,0.3879409134,-0.2507885098,0.1295179725,0.1006241366,-0.2705986798,0.0279504862,0.2041381747,-0.0470446907,0.2030743361,0.0508885458,0.0871171281,0.0146013889,0.5647197962,-0.2276450247,0.3679971993,-0.2421871424,-0.1133374274,0.3385269642,-0.0457907356,0.2186239809,-0.3233049512,-0.1835278124,-0.2548119724,0.1148838475,-0.3281368911,0.2052339911,-0.0470522083,-0.32210958,0.0186666157,0.0380094051,-0.0206913203,-0.0753713623,0.2960327864,-0.0189603567,0.0157178324,-0.1564354002,0.2078144103,-0.3696020842,-0.2330321074,-0.0829179063,-0.0495275706,0.1429994404,-0.2620231211,-0.6429176331,-0.0781933367,-0.4179568887,0.1745575219,-0.1723270565,0.4994799197,0.0736008361,-0.341901511,0.4459114373,0.0000096377,0.610689342,0.2610674202,0.0313419625,0.1837454587,-0.1467971951,-0.716468513,0.0446783267,0.2052855939,-0.0682096183,0.2740387619,0.6911302805,-0.0954619572,-0.1232027188,-0.1753898561,0.3305182457,-0.1972630322,-0.4610158503,-0.1612582356,-0.33329162,-0.157751441,0.1362024248,0.162478894,0.4764493108,0.4036732614,-0.0795260742,-0.2010962963,-0.1514124572,0.076922372,0.0857554302,0.3958840072,-0.1649831086,0.4091664851,-0.3334085941,0.1923512071,-0.0831135288,0.3989641964,-0.1917422563,-0.0647912174,0.3708535135,-0.2073857337,0.0760349035,0.3880732059,0.1935632974,0.1367917955,-0.1403948814,0.1752069741,-0.245393768,-0.0806318969,-0.1874670535,-0.0669726655,0.0424895808,-0.1677794904,0.2720348239,0.0439279191,-0.0331758559,0.3123394251,-0.1232544705,0.0711575225,0.7600028515,0.2339065075,1.2246540785,-0.5035197735,0.0143717537,0.1216156781,0.1656195074,-0.1179600358,-0.1760447174,-0.3269803524,-0.4070284367,0.1482392997,0.12771523,-0.2269847691,-0.1235538721,0.3887034655,0.0860179365,-0.1470342875,-0.2462285757,0.3335520327,-0.3722383678,0.4040262699,-0.0381876305,-0.4741929173,-0.5782680511,-0.0209844653,0.0157225858,-0.1392736137,-0.01047784,0.1738459766,-0.0361927673,-0.222405538,-0.1036521792,-0.1861285567,0.0907135233,0.181790337,0.1882383674,-0.2340189219,-0.1706718504,0.4167111814,0.3755473793,-0.0700849444,-0.0401893072,0.182368353,0.0647353679,0.1224832684,0.0813118368,-0.2595706284,0.0828861222,-0.1189796627,0.0572677255,-0.0665015653,-0.1295836866,-0.5366300941,-0.1099401861,0.199248001,0.3654546142,-0.1988744438,-0.2844585478,-0.2924718559,0.1557070166,-0.0658675656,0.0162828714,0.3401009142,0.1590853781,0.2823824883,-0.047331512,-0.2356753349,0.0175015163,0.360090524,0.0381032191,-0.2959870398,0.2296036333,0.0347345695,-0.2232731134,-0.2273774594,0.3738379776,0.3516613841,-0.4658204913,0.1134405136,-0.0073066889,-0.0589032806,0.0803503767,0.1580491811,0.2022910863,0.0176383778,-0.1211419404,-0.2088405639,-0.5282036662,0.0122620575,0.2407434583,0.1340420842,0.1419382095,-0.5530801415,-0.3032374978,0.0276500229,-0.1465066522,0.0846247002,-0.0470661372,0.1170279682,0.0066627269,-0.0371655114,0.1006972641,-0.1650077403,0.0521322079,-0.1201461926,0.0040520392,-0.0913612694,0.0671675578,0.2125337422,-0.0326332003,-0.2659902871,-0.2516624629,-0.4543337822,-0.2050011456,-0.3530681133,-0.1167320982,0.077926375,-0.1097265035,-0.3546777368,0.3384277523,0.1532832086,-0.0454579704,0.3576588929,-0.0494337603,0.4739779532,-0.0162983965,0.057773374,0.1461661607,-0.189484328,-0.23029688,0.2283019572,0.3283313513,-0.1512211859,0.398527056,-0.2034523636,0.0327935219,0.2581099272,0.02408267,0.2060749531,-0.1661803871,-0.2903409302,0.3153257966,0.0767675862,-0.0047934148,-0.2500506639,-0.0421130657,-0.1186143532,-0.2442470193,0.3506319821,0.4611359537,-0.2066257,-0.0887789056,0.0317300744,0.5728581548,0.0141484234,-0.4393891692,0.6040983796,-0.0466110148,-0.0192425195,0.1124724448,0.2082586586,0.0788008869,0.4882698953,-0.2155218869,0.1153136194,0.3541575074,0.5426251292,-0.3740213811,-0.3426681161,0.3637278676,0.2706037462,-0.2007653415,-0.1184999198,0.1577220559,0.2770216465,-0.1784947962,-0.1285788417,-0.2322596163,0.2374083698,-0.2579989433,0.0892989039,-0.3661226034,-0.0265115295,-0.018316742,0.1205130965,-0.260654062,0.0107326275,-0.2762074769,0.3323470652,0.0759625658,0.0478878543,0.0287432503,0.077931419,-0.0161675531,-0.4089176357,-0.0030229806,-0.3486030996,-0.0520249233,0.2731178403,-0.2664619088,0.390658468,0.5201631784,-0.0037661353,0.0543629974,-0.0440450385,-0.0332824849,0.0063723214,0.2009409964,-0.1043526456,0.2317556292,0.1520817429,-0.1098469719,-0.0681061521,-0.4159817994,-0.1741552055,0.4164218605,0.1408072859,0.3191126585,0.0841156319,-0.3321055472,-0.2562817037,0.0822448507,-0.0990472063,-0.1693680137,0.2653482258,0.0121956952,0.1619280428,-0.0405135527,-0.0301769935,-0.331315577,0.5836264491,0.3414525092,-0.3722861111,-0.2083298713,-0.3158126771,-0.6785955429,0.0826001242,0.0061991019,-0.029148804,0.4067520499,0.2453690171,0.4798544347,0.0738412142,-0.0745152384,0.0058917655,-0.1493743211,0.3993548751,-0.0918728411,0.0801308975,0.0865440369,0.2133027166,-0.236836955,-0.4747449458,0.4323938191,0.2044922262,-0.0929565057,-0.0872170702,0.0433862209,-0.2884812057,0.3838074505,0.3971142471,0.00272359,0.1114627048,0.3268879354,0.0639132559,0.2400347739,-0.0643624663,0.0969405398,0.2231922746,-0.0742232651,-0.077414088,-0.0497454628,0.3446443677,-0.2318461984,0.014853687,-0.2300883979,-0.2026833743,-0.1397498101,0.0247968212,-0.2142137736,0.2600075006,0.2427759171,-0.0022024547,0.1724119633,0.269413203,-0.2461118102,-0.0634343997,0.2341076732,-0.0005448365,-0.1780952364,0.0749683157,0.1997462362,0.2011824101,0.2328178585,-0.3857895732,-0.1835776269,0.2323231101,-0.131627664,-0.0851393864,0.3760204613,-0.206654802,0.2136587501,-0.1630482674,0.2711825669,-0.0848431885,-0.0476505794,-0.4124533832,-0.3080239594]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2099","title":"load_from_disk takes a long time to load local dataset","comments":"yup I casted it to `np.uint8` outside the function where it was defined. It was originally using python integers.","body":"I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)","comment_length":19,"text":"load_from_disk takes a long time to load local dataset\nI have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)\nyup I casted it to `np.uint8` outside the function where it was defined. It was originally using python integers.","embeddings":[-0.2997468412,-0.2523109615,-0.0470442772,0.2852835953,0.3280441463,0.0210994203,0.4262614846,0.2177656293,0.7503632903,-0.0394667275,0.1555927992,0.5000484586,0.0829721913,-0.0863406584,0.1403086632,0.1166696474,0.3541413546,0.2110988945,0.4538331032,-0.1608220637,-0.2806877792,0.0042520943,-0.124198705,0.0208434816,-0.1499311477,-0.0341603979,-0.0299530849,0.0478597358,-0.1242577806,-0.3127477169,0.0386533998,-0.0707973018,0.2723852992,0.2255568206,-0.0001278378,-0.2478280813,0.0132427784,0.1549170613,-0.2342653871,0.2983807325,0.3765066266,-0.4481585324,0.2435012162,-0.3445627689,0.1320651621,-0.3008111119,-0.3529150188,-0.2031404376,-0.0605137236,-0.1430762857,0.029551262,0.4554936588,-0.3025601804,0.0823656917,-0.1058336422,0.3159891367,-0.1544771045,0.0775512978,0.5621140599,-0.048869431,-0.0532413758,-0.0539404899,-0.2200506777,-0.0803386942,0.3734757006,0.0614769869,0.2543971241,0.1495889425,0.0807372555,0.108223781,0.5584256053,-0.0173403956,0.0234142635,-0.247217387,-0.0852084756,-0.3046393991,0.3880217969,-0.0014706694,-0.2571530342,0.1764448732,-0.0475827456,-0.0687611923,-0.0145756854,0.2424103171,-0.1865227371,-0.0527717099,0.020485878,0.1147159562,0.3606211543,-0.1024469137,0.0572253503,0.05434075,0.1102679148,0.3917821944,-0.4160100818,-0.1962457746,0.0078821732,0.0762107819,-0.1992874146,0.1077012569,0.2171662599,0.2246488482,-0.1134413406,-0.0168444458,0.2522183657,0.3522306979,-0.1196840405,0.1481120586,0.0786631107,-0.2717193663,-0.0636469275,-0.0346713625,-0.2698744833,-0.0448171012,0.5096763968,-0.4679943025,-0.2419988662,-0.1596256942,-0.0828347504,0.2333762646,0.43275702,-0.0528065674,0.2506528199,0.2286755741,0.0346168578,0.568198204,-0.036098361,0.0221790168,-0.1965541989,-0.1412778497,-0.0486950949,-0.2828444242,-0.1765477955,0.1495119482,0.0459260419,0.2822552621,-0.1178844348,0.1888041347,-0.214137122,0.0234120935,-0.1138857827,-0.1401478052,0.2944344282,0.2083587646,-0.1148258448,0.2261591554,0.1771098077,-0.2330551445,-0.4272059202,0.4167661667,-0.1968292147,-0.415686667,-0.4845806658,0.0009356908,-0.2500711679,-0.0882833451,-0.4191097319,-0.1701916158,0.1220637783,0.0307379514,-0.0117618674,-0.1612859964,-0.1153206527,-0.0993768498,-0.0477486141,0.1461014599,-0.4231797755,0.1730349511,-0.2556329966,0.2764599919,0.3224262893,0.6396907568,-0.2801316977,0.3539845347,-0.2494268864,0.2317926437,0.5141884685,-0.0979819968,-0.406452924,0.7154762745,-0.0799033567,-0.1685021669,0.0801311806,0.5803714395,0.3193608522,-0.0707680434,0.1484781355,0.6385694742,0.1840558499,0.2328279465,-0.393238157,-0.0290973932,0.0782501101,0.4464017451,-0.4105349481,0.1632714868,0.1406912357,-0.0007513441,0.1959132403,0.0519312248,-0.0848506466,0.2653449178,0.241429612,-0.0164800063,0.0377391018,0.2370992601,-0.1661601514,0.1361671388,0.4677137136,-0.2021825612,0.032636825,-0.1964571029,-0.3855073452,0.326543808,-0.0973818302,0.5834560394,-0.2067902982,-0.0223459862,0.0919388086,-0.162326932,-0.1692397892,-0.0781930536,-0.283765614,-0.1043159738,-0.0629550144,0.0552179776,0.1984125525,-0.0376981311,-0.3756640255,-0.022458151,0.0238030795,0.057233721,-0.2001612037,0.2730818391,0.1836063564,-0.0312019717,-0.0294404607,0.2263236493,0.1067596003,0.068008095,0.342223227,0.0136774359,0.2121348679,-0.2573305666,-0.0757785141,0.4699652195,-0.195330888,0.2148167789,-0.2350046635,-0.0943373889,0.3795797229,0.160831213,0.1499011219,-0.0218495727,0.1916032732,0.2930188477,0.2311092913,0.3050779104,-0.2529478669,0.2679817975,0.2955803275,-0.3433099091,-0.0599351265,0.2346324176,0.0216725785,-0.294165343,-0.0241064727,-0.0860864818,0.365613699,0.1009640247,0.0232074372,-0.3252738416,-0.0509922802,-0.1901419908,0.1565678269,-0.0855932534,-0.1552980095,0.1201144829,-0.0784913301,-0.144950211,-0.3378221989,0.0529579967,0.1145993695,0.408233434,-0.3593492806,0.2051281631,-0.1127206311,0.1153935939,0.0001296447,-0.3107719719,0.1094294041,-0.2215994895,-0.203345716,0.2541099191,-0.0517486036,-0.1613624543,-0.0621436872,0.3496825695,0.1964776814,-0.4854280949,-0.3430970609,-0.2745361626,-0.1107225195,-0.1661279351,0.4014419615,0.1267112046,0.2230729908,-0.0598717742,-0.1314963847,-0.2171730846,-0.148517698,-0.138150841,0.0304663647,0.1391072422,-0.172321707,0.3341897726,-0.5773805976,-0.05464774,0.2761105597,-0.327429682,-0.2220072597,0.4210555553,-0.1017543077,-0.3189682961,0.0206263773,0.1182105318,-0.1174128354,-0.2252645642,0.193953827,0.0601112694,0.1746490449,-0.270293504,0.1287637651,0.0201709736,0.4685717821,-0.253742516,-0.1709933877,-0.2462950647,0.6392534375,0.0141658336,-0.2546547055,-0.0216951892,0.1561202854,-0.0678284168,-0.019215297,-0.5329577923,0.2137179226,-0.4107220173,0.4174284935,-0.2403510958,0.1101479083,0.1109749004,-0.2708202899,0.0401136614,0.2173806876,-0.0477832966,0.1758057773,0.0863866583,0.0455207415,0.022664126,0.5955463052,-0.2443090826,0.3329944015,-0.2710857093,-0.136176303,0.3635106683,-0.0390579328,0.1948391497,-0.3454588056,-0.1980042458,-0.2656887174,0.1232635155,-0.3189104497,0.1959451288,-0.0129415486,-0.3360695839,0.0162459016,-0.0077085542,-0.0500189587,-0.0527191162,0.2995039523,0.0096211974,0.0443261154,-0.1402286142,0.2028515935,-0.3820977509,-0.2429337353,-0.0864677802,0.0005127424,0.1392797232,-0.2710648775,-0.5346099138,-0.0809863433,-0.407841742,0.1613990366,-0.1725641042,0.4567165673,0.0902644321,-0.338349849,0.4657453001,-0.005345474,0.5948346257,0.3084180951,0.0827636197,0.1923332959,-0.2019624412,-0.6922329068,0.0536255315,0.2320828885,-0.0532434508,0.3043691218,0.6459015012,-0.025588356,-0.1351346374,-0.0998530611,0.3506252468,-0.2139658481,-0.4529130161,-0.1806497574,-0.372996062,-0.1395716816,0.1319064051,0.1839442104,0.4820202291,0.4124492705,-0.0313324444,-0.2082057297,-0.13885203,0.0631121472,0.0752195865,0.3415224552,-0.0943398401,0.4265583456,-0.3278444111,0.1668903828,-0.1639242768,0.4506520927,-0.1249106377,-0.055145178,0.3552742302,-0.1789216697,0.0373832323,0.3945655227,0.2130493075,0.1087983549,-0.129309684,0.1753536612,-0.2159733623,-0.0754474998,-0.163915351,-0.0365844667,0.0398231559,-0.2066567391,0.2804308236,0.0647027269,-0.0397064649,0.2836984992,-0.1028813198,0.0937395766,0.6956081986,0.1971291602,1.2087390423,-0.4919820428,0.0336680189,0.1298188865,0.1862668246,-0.1252052337,-0.1392670721,-0.3274510503,-0.4056124091,0.1182406992,0.1332864612,-0.2227781564,-0.1108398363,0.377936691,0.1421389431,-0.145756349,-0.2845987976,0.3121123314,-0.3484480679,0.373178035,0.0001420608,-0.466259867,-0.5828540325,-0.0365543328,0.0176297426,-0.1060936004,-0.0159202088,0.1835933775,-0.0159110557,-0.176120922,-0.0690224394,-0.1480958909,0.1852276474,0.1799697727,0.1777569056,-0.2263096869,-0.120237641,0.447863996,0.3832814097,-0.0640562028,0.0072164722,0.1981435418,0.0716865435,0.1402180195,0.1192443594,-0.2417064756,0.0906333253,-0.1088959128,0.0084319245,-0.0712699965,-0.1276837289,-0.534802556,-0.0959504768,0.2205498368,0.3436756134,-0.2152745575,-0.2737527192,-0.2669528425,0.1950725764,-0.0696175918,0.0016729807,0.3451690674,0.1746703535,0.2470204532,-0.0803083405,-0.2670123279,0.0034175378,0.3597741127,0.0523384102,-0.2441837043,0.2246299386,0.0463057645,-0.23074238,-0.2069212347,0.3407580554,0.2651689053,-0.4949905276,0.1304307729,0.0353018939,-0.0418810248,0.0673030317,0.1329654604,0.1585170478,0.0157134831,-0.1375875771,-0.2012660354,-0.5275339484,-0.0297443029,0.2316035777,0.0961119607,0.1536243707,-0.5203171372,-0.2615032792,0.0122942971,-0.1291831583,0.0679062083,-0.0211725123,0.1300717741,0.0219918862,-0.0549305715,0.1272863746,-0.1701265275,0.0446750633,-0.1215179563,0.0353415497,-0.0732466877,0.0595907196,0.2262460887,-0.0183588751,-0.2851118743,-0.2297663391,-0.4710466266,-0.2037384659,-0.3548018336,-0.0840500444,0.1096756756,-0.0829447582,-0.367442131,0.3324759305,0.1370842159,0.0097594,0.4062235653,-0.060251087,0.4979396462,-0.0586131029,0.0656366944,0.1778946072,-0.1666070968,-0.2706589997,0.2173452377,0.343008399,-0.1129642352,0.3921229243,-0.2453584522,0.0555606671,0.2345794886,0.007280814,0.1970244944,-0.1595992595,-0.2889465094,0.2770781815,0.0497235507,-0.0093571413,-0.279538542,-0.0432367064,-0.1091426983,-0.2531602383,0.3529644907,0.4270124137,-0.2079939693,-0.0807909891,0.0477476306,0.5593768954,0.0142382048,-0.4618856013,0.5625962615,-0.0283035245,-0.0195328128,0.1542936116,0.2069474012,0.0614777431,0.5321910977,-0.203332901,0.0598361194,0.3535824716,0.553202033,-0.4282675385,-0.3345476985,0.3514855802,0.2562885284,-0.2642850876,-0.1204353496,0.1763008535,0.1915052831,-0.2000450641,-0.0962897837,-0.2419995964,0.2856771648,-0.2213685066,0.0616025403,-0.3761419356,-0.0039552548,-0.0297697894,0.1692757905,-0.2597093582,0.0175362676,-0.2455750704,0.3440790176,0.0641215518,0.0444135964,0.0583960414,0.0609890781,0.0008990326,-0.4227383435,-0.0232525375,-0.3469661474,-0.0325265042,0.2949286699,-0.2910652161,0.3708652854,0.4920545816,0.0080861198,0.053411752,-0.0598783754,-0.0462807491,0.019048525,0.2101278752,-0.1083664149,0.3407147229,0.1441574097,-0.1136075333,-0.065486528,-0.4170873165,-0.1691431999,0.4294930995,0.0875380784,0.3718037605,0.0822222456,-0.3147863448,-0.2893888056,0.0842055976,-0.0928579867,-0.1836446226,0.2388719171,0.0186640844,0.1382622719,-0.0549394302,-0.0443665944,-0.3314493001,0.5940251946,0.3194931746,-0.3627039492,-0.1800250858,-0.3111115992,-0.6691558361,0.0797606334,0.0565281361,-0.0767395198,0.3885866404,0.2282045484,0.4421895742,0.0508389212,-0.1075005084,0.011397263,-0.1168243513,0.3702003062,-0.0382715426,0.0379904956,0.0463442542,0.2053325772,-0.2436439842,-0.4843665659,0.4108671546,0.1688494533,-0.1000111103,-0.1184422895,0.0033153363,-0.2772935927,0.4129746556,0.3902547956,0.0072056432,0.1270107031,0.3299395442,0.059231326,0.2613452077,-0.0972217843,0.0697034672,0.2434346676,-0.0778223425,-0.0453930646,-0.0346203558,0.3828944564,-0.1904073954,0.0101255644,-0.2644207776,-0.1641371846,-0.1226105541,-0.0124836639,-0.2296936363,0.2783011198,0.2322643548,0.0218800195,0.177547425,0.2454540879,-0.2312730551,-0.0554251298,0.2142624408,-0.0180718731,-0.1550247967,0.0602331012,0.1880708039,0.1968451887,0.1879764348,-0.3372845352,-0.1518452466,0.2316040844,-0.1202039644,-0.1169504225,0.4154011309,-0.2452114522,0.1990677714,-0.144470185,0.2633864284,-0.1030327752,-0.0655323118,-0.4062345028,-0.2756508589]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2099","title":"load_from_disk takes a long time to load local dataset","comments":"Strangely, even when I manually created `np.arrays` of specific `dtypes`, the types in the final `dataset_info.json` that gets written are still `int64`.\r\n\r\nUpdate: I tried creating lists of `int8`s and got the same result.","body":"I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)","comment_length":34,"text":"load_from_disk takes a long time to load local dataset\nI have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)\nStrangely, even when I manually created `np.arrays` of specific `dtypes`, the types in the final `dataset_info.json` that gets written are still `int64`.\r\n\r\nUpdate: I tried creating lists of `int8`s and got the same result.","embeddings":[-0.2173988074,-0.1876725256,-0.0520239398,0.336670965,0.3299890161,0.0349005423,0.4153964818,0.2823090255,0.6786063313,-0.1099481285,0.1326237321,0.5752293468,0.194253251,-0.0471497588,0.1443952918,0.1174114272,0.3587273955,0.2722390592,0.3763953149,-0.1764423251,-0.3060879707,0.0475458466,-0.0370804928,-0.0319931991,-0.0833014771,-0.0650950223,0.0567429177,0.0116351899,-0.1999146342,-0.3068855703,0.1065616608,-0.0948057771,0.2616828084,0.2380930036,-0.0001292553,-0.1857509911,0.1267102659,0.0981938243,-0.2422356755,0.324352026,0.4163019657,-0.4820283353,0.2607607245,-0.3481028974,0.1100002602,-0.4955370724,-0.271995604,-0.3215941489,-0.0316408873,-0.1341709346,0.0246905722,0.467775166,-0.2241472602,0.1327084005,-0.201158911,0.4392254055,-0.2193626016,-0.0071726125,0.4425097108,0.0582930297,-0.0485814847,-0.0865372121,-0.1406695396,-0.0158314817,0.4063599408,0.0615013614,0.2055451274,0.1228796989,0.0694571212,0.172819525,0.620713532,0.0334614404,-0.0707622766,-0.2485562265,-0.1489282101,-0.2460744828,0.4324035048,0.0362544432,-0.2357409298,0.1717061549,-0.1762757003,-0.0015281078,-0.073338151,0.2504836023,-0.1308272183,-0.1616676152,-0.0110073313,0.1591522545,0.3527541161,-0.196405679,0.077416636,-0.0155625092,0.0091212317,0.3035268784,-0.3383488953,-0.2830924392,0.0515062213,-0.0079683717,-0.0780863836,0.0634728298,0.162239328,0.2495290041,-0.2331532538,-0.047210671,0.2891207337,0.360879004,-0.0328744687,0.0068867933,0.0910084769,-0.3389004469,0.0694476366,0.0200346243,-0.1676364541,0.0830196068,0.3784140646,-0.496773541,-0.2162392288,-0.1500123739,-0.0208427552,0.240024671,0.3636525869,0.0160096996,0.1731543243,0.1814385355,0.0771478713,0.6115540862,-0.0976982936,0.0728253201,-0.1629833132,-0.1739878207,-0.019771032,-0.3106483221,-0.1181267276,0.085008882,0.1496679187,0.3727173507,-0.0459691957,0.1671038717,-0.2715291381,0.015793955,-0.0969670117,-0.113113232,0.312556833,0.2298932523,-0.0987831131,0.255743295,0.1460681409,-0.2891290188,-0.4497122467,0.3425271809,-0.1358734816,-0.2943195701,-0.4542927444,-0.0250244699,-0.2671667635,-0.0196941737,-0.4391439259,-0.1076664925,0.1381552815,0.0255165212,-0.0122429384,-0.2078366429,-0.0829693824,-0.1118213981,-0.0885614827,0.0849554241,-0.4510560334,0.1749621928,-0.2742944062,0.2847523987,0.2883858085,0.6702421308,-0.2067277133,0.2299301922,-0.1860372871,0.3189185262,0.4110121131,-0.0977737606,-0.3115602434,0.7315465808,-0.0117100887,-0.1459036469,0.1338101327,0.5266262293,0.2496895939,-0.0425945669,0.1365237534,0.6743009686,0.1748803407,0.2215825468,-0.2907084227,-0.0866582021,0.048170846,0.4137925208,-0.4890213609,0.1240891442,0.1250769049,-0.054476399,0.1471737921,0.0456148833,-0.0502899215,0.2346315384,0.2530584335,-0.0990096927,0.0564774163,0.3052899241,-0.3008289337,0.1053384021,0.3949753046,-0.180567354,0.0343267322,-0.2042903155,-0.2784738243,0.3042504489,-0.0289720707,0.5546165705,-0.1815521121,0.0827285722,0.0816990957,-0.1209331155,-0.1571292728,-0.1332882643,-0.2426025718,-0.1102127433,-0.1819319129,0.0081924116,0.2339276224,-0.0504297167,-0.30652529,-0.0196361449,0.0269085113,-0.0384555683,-0.1630454808,0.2365633547,0.2002563924,0.0035289566,-0.0308959708,0.1877446175,0.151982531,0.0807134062,0.3181470931,0.0528730489,0.1747149825,-0.2780156136,-0.1761330813,0.503749907,-0.189139083,0.2216318548,-0.2732480764,-0.078467682,0.4578700662,0.1897336245,0.0803913251,-0.0869936049,0.1122872159,0.3219029903,0.1343201846,0.2910538316,-0.2332690954,0.3427419066,0.4129641652,-0.3111322522,-0.0754261836,0.2076277435,0.0162483081,-0.3657492101,-0.0054653161,-0.0687501132,0.3628416061,0.0558839776,0.0507821552,-0.369697839,-0.0517441444,-0.1846944243,0.1200352162,-0.0936766416,-0.0346512869,0.0992259383,-0.0428062268,-0.2816371322,-0.3656868935,0.0651431158,0.1108409688,0.388858825,-0.3100927174,0.2180454433,-0.1968772113,0.0692615584,0.0168185998,-0.2861744165,0.1109502688,-0.1754750162,-0.2048086971,0.3162035346,-0.0743373856,-0.1524071544,0.064966999,0.423001796,0.2276905626,-0.4015130699,-0.3833144307,-0.2376029491,-0.1568235606,-0.1752423644,0.3804137707,0.1047094241,0.1703634411,-0.1451798677,-0.1550821364,-0.2253527939,-0.1828177124,-0.1096083447,0.0458928607,0.2592884302,-0.1274664849,0.3801020384,-0.5672895908,0.0562553778,0.2641807497,-0.2876508832,-0.285835743,0.3660439849,-0.1474811584,-0.3140183687,-0.0337165073,0.03936892,-0.0902423635,-0.2155958563,0.3713660538,0.0776870474,0.1806075871,-0.2034236491,0.1504123211,0.007472916,0.4369546175,-0.2281594872,-0.2046726346,-0.2325005084,0.6616532207,-0.061055664,-0.2485224307,-0.0557904243,0.1784061342,-0.1016363725,-0.0587272793,-0.5192955136,0.1694786549,-0.4142033458,0.3584964871,-0.1935166419,0.0395521745,0.1161186844,-0.293694526,0.0407142825,0.1232160851,-0.1552179754,0.2262769789,0.1684099436,0.1085288152,0.0971095338,0.667519629,-0.2841599584,0.2335277349,-0.1284044534,-0.2436570525,0.3138743639,0.0045588366,0.2500695586,-0.4183475971,-0.0997267738,-0.2140558362,0.1322830468,-0.2786403,0.2092730254,-0.0230737142,-0.3529036939,0.056518659,0.0554478206,-0.0645913631,-0.053827256,0.2805733681,0.0725996867,-0.0482393652,-0.0943965167,0.3200037479,-0.3532731235,-0.2117579132,-0.1610506624,0.065060474,0.1711561829,-0.1487732977,-0.6796736121,-0.167427063,-0.3332336843,0.1392716765,-0.2569188774,0.5929815769,0.0225006212,-0.2331376821,0.3818157911,-0.02671865,0.5649272799,0.2307366878,0.1245870963,0.1613277793,-0.1329400092,-0.6834199429,0.0788837299,0.2797006369,-0.0070202891,0.3291249573,0.7125704885,-0.1811696738,-0.0999959335,-0.1415992826,0.3365512788,-0.1852432787,-0.3992156982,-0.149357453,-0.4388356209,-0.070788607,0.0791190639,0.2100012153,0.4462957382,0.3405472636,-0.0486600809,-0.2184106857,-0.1173243299,0.0599522442,0.0986243635,0.3926590979,-0.0840948448,0.3604735136,-0.2897668183,0.2088491768,-0.0379227474,0.3952316344,-0.0141892275,-0.0131745785,0.3565963805,-0.1868538857,-0.0403169058,0.3948362768,0.2167873085,-0.0137316985,0.0050160564,0.1383889467,-0.2069513798,-0.0761823505,-0.1330932975,-0.061947681,-0.0004001069,-0.2413403094,0.342110157,0.0458911061,0.0366980135,0.2169332653,-0.1064546257,0.0730891302,0.700540483,0.2282200903,1.0924137831,-0.4371308982,0.0410692245,0.1504455507,0.1532889754,0.045146957,-0.2118844092,-0.2649596632,-0.4133767486,0.2359389514,0.1291562617,-0.2488388717,-0.1227727383,0.3823552728,0.1817042977,-0.1715881526,-0.1866682917,0.3809774518,-0.3225501478,0.4586809576,-0.0498508252,-0.5222053528,-0.5489799976,-0.0528285168,-0.04924367,-0.0831591561,0.0803819299,0.041178558,-0.0531216599,-0.1567960829,-0.0624625385,-0.2056075633,0.3019216061,0.1974998564,0.2539479136,-0.26063025,-0.1255116761,0.4229888022,0.4324880838,0.0126522137,-0.1240098849,0.1753550619,0.0493507273,0.1119155213,0.1297535598,-0.2778564692,0.0515651591,-0.0893167406,0.0382789597,-0.1961215287,-0.1636033654,-0.4552238584,-0.1367639899,0.1120985895,0.1615016162,-0.2072803229,-0.2247730196,-0.3366514146,0.1160389259,-0.0405575484,-0.0055044703,0.3154764473,0.1900609732,0.2958968878,0.0065445285,-0.2580960989,-0.0127004888,0.3266024888,0.1172463149,-0.2492703497,0.2166793197,0.0922861323,-0.2408176512,-0.23052755,0.3632845581,0.3062568307,-0.3923356533,0.1796457767,0.0062000025,-0.0776297748,0.0210312773,0.064481616,0.1066220924,0.0795006827,-0.1134222969,-0.2339743227,-0.5309724212,0.0488857925,0.1587391049,0.1442113072,0.1871338636,-0.455325067,-0.2841465175,-0.0301209781,-0.108377099,0.0317692086,-0.0022478607,0.1025848985,-0.0446095392,-0.0486175045,0.1031708866,-0.1538616419,0.0147899827,-0.1188466772,-0.0506375097,-0.0317351036,0.0575858839,0.2309266031,0.0243806913,-0.3043709099,-0.2005882263,-0.5046975017,-0.2750754654,-0.3541999161,-0.1433112472,0.0248264614,-0.0564157926,-0.3651974797,0.3095920384,0.1771620512,0.0478553362,0.3821151257,-0.0351383127,0.5291316509,-0.0838509947,0.0906450823,0.141985938,-0.1484673619,-0.3130100369,0.2056972831,0.2749147713,-0.0741816089,0.3640145063,-0.2556057572,0.0451694615,0.1968622804,0.0170880239,0.2804639935,-0.1707636416,-0.2876013517,0.2010520548,0.0151961995,0.0878365338,-0.2189321667,-0.0090425573,-0.0722838193,-0.275579989,0.3545841277,0.4353601336,-0.226568833,-0.1017175168,0.0076612304,0.6051027179,0.0567761362,-0.3878492415,0.5591935515,-0.1221305579,-0.0224982947,0.1900873482,0.102067031,0.0879572779,0.5890636444,-0.1564524472,0.0454869606,0.3945411742,0.5005429983,-0.4387721717,-0.2563966215,0.2966640592,0.3051196039,-0.1579086035,-0.0303891618,0.2748235762,0.2423794866,-0.2620984912,-0.0929099992,-0.3873699903,0.2268604785,-0.2815282345,-0.0116102053,-0.4437780678,-0.0770194009,0.0786994323,0.2058532089,-0.3155549765,0.0444349051,-0.2954046428,0.3063108325,-0.0046826904,-0.0141255753,0.1462667286,-0.0559757911,-0.0438110419,-0.3790052533,0.0477780476,-0.2433989048,-0.0542871207,0.2168344259,-0.3252421021,0.3133122325,0.4564042687,0.0446825735,0.0164988544,-0.0667361692,-0.0828174204,-0.0087055517,0.2081096619,-0.0590163432,0.1394257247,0.1305967569,-0.1088898927,0.0057669068,-0.4132788479,-0.1747808158,0.3841588199,0.1387759149,0.3312199116,0.0410332233,-0.2562456727,-0.2217212766,0.0317922048,-0.0179433003,-0.1998837143,0.2481618077,0.107070379,0.1406807899,-0.1053970605,-0.0489188656,-0.4211722016,0.6321418881,0.1931618899,-0.4575023353,-0.1210154146,-0.2958081067,-0.495721519,0.0898916349,0.0185373202,-0.0842300206,0.37334162,0.2994863391,0.4935496151,0.0418795832,-0.1397035569,0.0613696836,-0.0776114911,0.4077981114,-0.0841037855,0.0084019061,0.0422826409,0.2542155683,-0.2144946456,-0.4766701162,0.429518044,0.2177685946,-0.1116072312,-0.1958276927,0.1195867136,-0.2376290709,0.4204731882,0.3587202728,0.1112726852,0.0840308368,0.322786212,-0.0830638632,0.2185931951,-0.1235918403,0.0773009658,0.3300933838,-0.1025028825,0.0824519098,-0.083706215,0.4998011589,-0.2368961871,-0.0310852118,-0.3372482359,-0.0928841606,-0.1342758834,0.0473145358,-0.3381272256,0.1738967597,0.2373443693,-0.0133346785,0.2018319964,0.2496103644,-0.2224580348,-0.0728699192,0.2571419477,0.0301768798,-0.2030026168,0.1415959746,0.0672397614,0.1613492519,0.1663156003,-0.3300634921,-0.0823367909,0.1679615676,-0.1623261273,-0.1481468678,0.3159272671,-0.2683696151,0.0756740794,-0.1605041623,0.2990536094,-0.0243849494,-0.1328065991,-0.4334263206,-0.2827863097]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2099","title":"load_from_disk takes a long time to load local dataset","comments":"Yes this is a known issue: #625 \r\nWe're working on making the precision kept for numpy :)\r\nTo specify the precision of the integers, currently one needs to specify the output features with `.map(..., features=output_features)`","body":"I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)","comment_length":35,"text":"load_from_disk takes a long time to load local dataset\nI have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)\nYes this is a known issue: #625 \r\nWe're working on making the precision kept for numpy :)\r\nTo specify the precision of the integers, currently one needs to specify the output features with `.map(..., features=output_features)`","embeddings":[-0.2853317261,-0.2187879235,-0.0280981176,0.2345814109,0.3253434598,0.0638445765,0.4536391497,0.2049428672,0.6538368464,-0.0338119492,0.1011138856,0.4731715918,0.1274633706,-0.169902727,0.0634809881,0.1007766947,0.3845754564,0.170215264,0.4007259607,-0.2280999869,-0.3928019702,-0.0222928356,-0.0956596732,-0.0267841909,-0.2128368318,-0.1019784436,-0.0361412168,0.0544555523,-0.1071031392,-0.2511247993,0.021519443,-0.1343869418,0.3368290067,0.2338037044,-0.0001270739,-0.1998752505,0.0475257188,0.1346090436,-0.2297703028,0.3513155878,0.4877164662,-0.6189042926,0.216591239,-0.3899850249,0.1582983881,-0.1895425618,-0.3014677465,-0.266580075,-0.0368053429,-0.1214371175,0.0450730026,0.5076544881,-0.2321638167,0.1393104643,-0.2855154872,0.3035943508,-0.2053806335,0.0653144121,0.4982056916,0.054926306,-0.0345793851,-0.0517828688,-0.1683321148,-0.0302100033,0.3090568185,0.107944496,0.4078367651,0.1171322837,-0.0080201579,0.1199575067,0.4965890944,0.0335557647,-0.0846852586,-0.3417033255,-0.1112953573,-0.4144537449,0.3293813467,0.0911913514,-0.3490361571,0.1724072993,-0.0470476486,-0.0798033103,0.0139332637,0.2245960385,-0.1558129638,-0.0913809836,0.0197384842,0.1317343712,0.2276474237,-0.1148111597,0.0845651254,0.0103159174,0.2136578709,0.42485255,-0.4246657193,-0.2353165001,0.0208281279,0.014246136,-0.1653255224,0.0102034872,0.1822424978,0.2515753806,-0.1862856001,0.0219252128,0.3129630089,0.4280665815,-0.1733574569,0.1863619834,0.0927741602,-0.243365854,0.1412958801,0.0152633199,-0.1005442068,0.0422579832,0.5480375886,-0.4681058228,-0.1633123904,-0.132016778,-0.142205447,0.2277147174,0.4887236655,-0.0417849496,0.1561329663,0.1455444992,0.0806359574,0.5697749257,-0.0562968105,0.1194836274,-0.1807684451,-0.1578468531,-0.0843190402,-0.3356249034,-0.0562649444,0.1903012842,0.0402992442,0.3683851361,-0.0683654025,0.1541199237,-0.1981162876,0.0685257316,-0.1171715334,-0.1645004898,0.2333038598,0.1856643856,-0.1380418688,0.255790025,0.1359824538,-0.1931857169,-0.4076722562,0.4262596071,-0.1646172255,-0.340272218,-0.4790644348,0.009015671,-0.1866096854,0.0037994366,-0.3539022505,-0.1691920608,0.2202159166,0.0695998296,-0.0052030212,-0.2440938503,-0.1140649393,-0.1563969254,-0.0493967533,0.0769275203,-0.3092054427,0.1904052049,-0.1717406362,0.3061513603,0.3065825105,0.7243385315,-0.1892374307,0.216844812,-0.2830536664,0.237182647,0.4602162838,-0.0460090525,-0.3574863672,0.6271573305,-0.1066113636,-0.078496851,0.0641870126,0.5633755922,0.2529529035,-0.1901667863,0.1485000402,0.6897596717,0.10922122,0.199160859,-0.4090479612,-0.0733782426,0.0952199027,0.4398112893,-0.407392621,0.1320135444,0.0775019005,-0.0325040147,0.2144829333,0.1083314866,-0.0670487806,0.2884332538,0.2247227728,-0.0826955587,-0.0157324541,0.2709610164,-0.179816857,0.1594817191,0.4164332449,-0.1629320979,-0.0383521691,-0.1796811223,-0.3229567111,0.3180896342,-0.0788329244,0.5116457343,-0.1500653028,0.0236002747,0.1203980818,-0.1022779197,-0.07345853,-0.017423816,-0.2825340629,-0.0966027305,-0.1130622104,0.0154976984,0.1525216252,-0.0754139274,-0.3999119997,0.0115298182,0.0648076534,0.0145667084,-0.2211715579,0.3539468348,0.1323072165,-0.0865555406,-0.1187740117,0.266340524,0.1861677915,0.0791090503,0.3964467347,0.0986074582,0.2052937597,-0.2505775988,-0.1120992154,0.4844498038,-0.1962548494,0.1562964618,-0.2818818986,-0.0745494142,0.4282444119,0.1892055869,0.149519816,-0.0909317881,0.1781179309,0.3216135204,0.2176568657,0.2503772378,-0.2243136317,0.2397245318,0.2951150537,-0.3496774435,0.0404902436,0.2224862128,0.0427819677,-0.2908237875,-0.0233579762,-0.1991864443,0.357840091,0.0737689435,0.065910399,-0.2941621542,-0.0975086167,-0.1970733702,0.0982190073,-0.0139931822,-0.0982626826,0.0808211043,-0.0495457463,-0.1687764525,-0.2680055499,0.1539381295,0.0446598753,0.3528857529,-0.3059108555,0.0973189697,-0.0127622513,0.2521743774,-0.0604741015,-0.3014185727,0.0552328452,-0.2555603087,-0.2118463367,0.2482574731,-0.047424823,-0.1437081099,-0.0814494416,0.3157687485,0.2143785655,-0.3621745706,-0.3067261875,-0.2365983725,-0.1183376387,-0.1699169874,0.3378154635,0.0872206688,0.1963418722,-0.0735865384,-0.1557127684,-0.1935285479,-0.0977740362,-0.1680152863,0.0816521496,0.1938207895,-0.1836411655,0.3560246825,-0.5311698914,0.0350948758,0.2957322896,-0.2823811769,-0.164279148,0.3820283711,-0.0902661309,-0.3041514158,0.0562240109,0.1490205973,-0.2280776203,-0.226206407,0.1455477625,0.0563633256,0.1158922613,-0.280510664,0.090152964,-0.0288886279,0.46661219,-0.1933833063,-0.2513419092,-0.18547602,0.6473256946,0.0334665217,-0.2793290317,-0.1435446143,0.0893830135,-0.1167957261,-0.0809702426,-0.4640571177,0.1544277817,-0.3944261968,0.5134144425,-0.2122781277,0.1099839881,0.1466535926,-0.3103833199,0.0481796786,0.2114723176,-0.0594053343,0.1511201859,0.1554528773,0.1205791458,0.0782571137,0.6237229705,-0.2136434317,0.2885495424,-0.1777602434,-0.1728207916,0.3498382866,0.0695884153,0.241609022,-0.3530932367,-0.1464719772,-0.2309068888,0.0744832978,-0.3759373426,0.1659772694,-0.0140436729,-0.2986568213,0.030100707,-0.0075991619,-0.1319840997,-0.0708852783,0.2679990828,0.044614803,0.0690062717,-0.0856294408,0.274379164,-0.361964494,-0.2205962092,-0.0780423358,-0.0102083487,0.1685323566,-0.2418090552,-0.5344122052,-0.1225279123,-0.4297573864,0.1782601923,-0.1769814342,0.5807365775,0.0980259031,-0.3366993368,0.3795421124,0.062400084,0.6375271678,0.2469396591,0.060842488,0.1415727139,-0.1365589052,-0.7212398052,0.0286244564,0.2283572853,-0.1198741421,0.3681832552,0.6995567083,-0.106727466,-0.0936523378,-0.1264533401,0.3635672629,-0.2390449494,-0.429412365,-0.1454217583,-0.3395488262,-0.0486600138,0.1412876397,0.2077574283,0.5102298856,0.3965984285,0.0031084442,-0.2096402943,-0.0473886058,-0.0402110368,0.0379883461,0.3384119868,-0.0584836155,0.3670160472,-0.2865017056,0.2127586454,-0.1602149755,0.4126995504,-0.2022039145,-0.0719748735,0.333532393,-0.2052238733,-0.0112007717,0.3617828488,0.2375926673,-0.0724401921,-0.1386765987,0.2013556212,-0.1970278025,-0.1319131106,-0.161780104,0.0107374005,-0.0266841576,-0.1130774841,0.3312250376,0.0599075742,0.0554097705,0.1973131448,-0.2578000128,0.1007328555,0.6957800388,0.2066813856,1.0700024366,-0.4297008514,0.0261727199,0.1830733865,0.133229509,-0.0664959401,-0.1622080207,-0.295810014,-0.3466523588,0.1041414738,0.1430855095,-0.2020399421,-0.1270277947,0.3857885897,0.1655088663,-0.0757700428,-0.3193400502,0.2646409571,-0.3175810277,0.3741774261,0.0279612709,-0.4371841848,-0.5361632705,-0.0206226297,-0.002340571,-0.045386523,0.0146705331,0.1655902714,-0.0171103533,-0.2421849668,-0.0378909372,-0.2156960815,0.1315557957,0.2387965471,0.229111582,-0.3089773655,-0.1866568476,0.4068339169,0.4967942238,-0.1496448666,-0.0450803153,0.1493596286,0.1311110258,0.1146297976,0.1725499928,-0.2026840001,0.1210019439,-0.0790865868,0.1139864028,-0.1197757497,-0.2074473351,-0.4152695835,-0.1680642217,0.0990769342,0.2944338918,-0.2348843515,-0.2459638715,-0.4113587439,0.2255997211,-0.0879934281,0.0171282049,0.3532120883,0.1283498704,0.2408347428,-0.1167029515,-0.205785051,-0.036938455,0.3736330867,0.0963334888,-0.3826689422,0.2273896188,0.0478901453,-0.2305285484,-0.1743086278,0.3632317185,0.2470691353,-0.4045790732,0.1640166044,0.0160556566,-0.0228646901,-0.0384315848,0.1288540065,0.2164090127,0.0222297553,-0.1649475247,-0.2441233993,-0.5539364219,-0.0741589218,0.2339703292,0.0861405134,0.1739705801,-0.5396533012,-0.2221939266,0.0597382076,-0.1364130676,-0.0341856033,-0.0190554131,0.104413189,0.125719294,-0.0074808216,0.1417950541,-0.0443211682,0.0587475412,-0.1436631829,-0.0112721147,-0.0658592135,0.1035137624,0.1987862289,-0.0456754081,-0.3167658448,-0.1872192621,-0.4499413073,-0.1662233323,-0.3510998189,-0.1727164388,0.0560750514,-0.0659364089,-0.2894792557,0.3735766411,0.1247936115,0.0399030745,0.3756256998,-0.0926093683,0.4059261084,-0.0596292466,0.1381234378,0.2167510241,-0.1698891968,-0.3051192462,0.1427729875,0.2466362566,-0.1016236618,0.3579626083,-0.2795626819,0.034239132,0.2548421025,0.0009393012,0.1559845805,-0.2034269124,-0.339635402,0.2007690072,0.0653720945,-0.0298558287,-0.2698592246,-0.00488541,-0.0707809776,-0.2545434237,0.3708546758,0.439925164,-0.2043177336,-0.124781549,0.0630115345,0.5689046979,-0.0542681962,-0.4450153708,0.7241832018,-0.0098483944,-0.0151664214,0.1700899005,0.0891747102,0.0543679185,0.5115529895,-0.2473190725,0.1056129709,0.3380523324,0.4634325206,-0.4425362647,-0.2838799357,0.3434019983,0.2554889023,-0.2900569141,-0.1349399835,0.1769851595,0.242483601,-0.2656917274,-0.077292949,-0.2843950093,0.2597663999,-0.3065816462,0.0730447397,-0.3176290691,-0.0146050761,0.0316000432,0.0886099562,-0.2566404343,-0.0176419523,-0.215032801,0.2963841856,0.0061716055,0.1056660041,0.0663891658,0.0169938859,-0.0523176417,-0.4044031799,-0.0340560637,-0.285150677,-0.0472715609,0.2069104314,-0.3241387904,0.4127492309,0.5296902061,-0.0159821976,0.1172258183,0.0412643179,-0.0809832513,-0.0211193059,0.1505697817,-0.1537804157,0.2145948708,0.1504526585,-0.0943382531,-0.0771882907,-0.3264719546,-0.1722001433,0.4279322028,0.0687267333,0.4512882829,0.006875739,-0.3435263038,-0.270804137,-0.0174020659,0.0194845777,-0.2078591138,0.2200274616,0.036758922,0.1026140749,-0.0131484801,-0.0386601947,-0.3887904584,0.5161964297,0.305285275,-0.405207783,-0.2370075434,-0.2756316662,-0.6192650795,0.1022996455,0.0852302313,-0.15629825,0.4201318026,0.2299827337,0.4341727197,0.1145569608,-0.1235430911,0.1211630777,-0.1024006605,0.4559914172,-0.0728729591,0.0316683576,-0.0157787669,0.1788159758,-0.1881528199,-0.493119359,0.4214515388,0.2105397582,-0.0832052901,-0.1440434307,0.1135315299,-0.3183138371,0.4916071296,0.4417491555,0.0854352415,0.0012170757,0.3090946376,0.0343610756,0.2861704528,-0.0496502891,0.0325209834,0.2648922801,-0.1444528848,-0.006360143,-0.0998491421,0.3538447618,-0.244594112,0.0030524987,-0.236217469,-0.1950999051,-0.076506719,0.0317683816,-0.2294916213,0.2126773745,0.3267274797,-0.0665907711,0.2403953075,0.282975018,-0.2431827337,-0.1056497395,0.2305815369,0.0404738747,-0.1472240686,0.0045779129,0.1573981941,0.0249260422,0.1547422558,-0.3673827946,-0.1348811835,0.2557360232,-0.1095803902,-0.1240494326,0.3603811264,-0.2274812311,0.13320297,-0.1306196153,0.2775725722,-0.1001715437,-0.0750298947,-0.4015408754,-0.2252212316]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2099","title":"load_from_disk takes a long time to load local dataset","comments":"Do you know what step is taking forever in the code ?\r\nWhat happens if you interrupt the execution of the dataset loading ?","body":"I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)","comment_length":24,"text":"load_from_disk takes a long time to load local dataset\nI have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)\nDo you know what step is taking forever in the code ?\r\nWhat happens if you interrupt the execution of the dataset loading ?","embeddings":[-0.357709229,-0.1911456436,-0.0542891771,0.2522047162,0.3178031147,0.0741005987,0.4124632478,0.2019445449,0.6933247447,0.0321035273,0.1932916492,0.4987606406,0.1199500635,-0.0854377076,0.078265138,0.1477905661,0.3398175836,0.2383556068,0.4402107,-0.1339953244,-0.3000718951,-0.0153765325,-0.1427359879,0.0422020368,-0.1994234174,-0.1508752704,-0.0522637442,0.0502910502,-0.1458295584,-0.2906326354,0.0190012082,-0.0184636898,0.2739640474,0.2816301286,-0.0001262479,-0.2672604918,0.0696517602,0.1940072477,-0.2356897742,0.3061158955,0.3798666894,-0.4438560903,0.2077921778,-0.3478523493,0.1490973085,-0.2778462768,-0.2573793828,-0.3251244724,-0.0373226218,-0.1680577695,0.045437146,0.4464398324,-0.3012800813,0.0890896022,-0.1744759232,0.3106313646,-0.1221053749,0.0339281447,0.5823899508,-0.0583966412,-0.0453900956,-0.0751692504,-0.1739264131,-0.1305466592,0.4071652889,0.0562382899,0.2787828743,0.0998648778,0.0840034708,0.1261284649,0.6219447255,-0.0203819778,0.0500371382,-0.2161280066,-0.0790367424,-0.3577191234,0.4000557661,0.003064764,-0.3239591122,0.1873583347,-0.0897821486,-0.0954797268,-0.0011594529,0.2260670811,-0.1863961369,-0.1411526501,0.0256379247,0.1344275326,0.2987030745,-0.1170691326,0.0555922054,-0.0563122779,0.1497805715,0.4445002675,-0.417232424,-0.191189155,0.0042201998,0.1244223416,-0.1765966713,0.1036820933,0.1432673037,0.214681834,-0.1279767752,-0.0154735511,0.2563345432,0.3761742711,-0.1128492653,0.09652026,0.1039664671,-0.3212371469,-0.0395908877,-0.0354100689,-0.2497454733,-0.0189914219,0.4814788699,-0.4471083283,-0.200149402,-0.132312119,-0.0559440106,0.2049459964,0.4125044048,-0.0235757884,0.2181765288,0.2371752858,0.0863840953,0.4874051809,-0.0957277715,-0.0333782248,-0.2088415474,-0.1472740769,-0.0681420043,-0.2910850942,-0.1855145991,0.1931467503,0.0448805355,0.2586331069,-0.1528833807,0.2015702128,-0.2130119652,0.0637814105,-0.1322868764,-0.155823037,0.254134655,0.1592623591,-0.1326764524,0.2026861757,0.1348429918,-0.3092466295,-0.405815959,0.4017782211,-0.1136583388,-0.3545094728,-0.4384785593,0.013739787,-0.2659834325,-0.0327593312,-0.3687290847,-0.0985771567,0.0929255858,0.0433784761,-0.0409642234,-0.1871865094,-0.0954578146,-0.0469432361,0.0055030761,0.1939486116,-0.4462917149,0.2134213448,-0.2279573083,0.2241958082,0.2977640927,0.641292274,-0.2757306695,0.3365720809,-0.2637676299,0.2700733542,0.4910587966,-0.0406596847,-0.3431300819,0.7155482769,-0.1504546255,-0.1908347011,0.0214608889,0.5547387004,0.2817492187,-0.0781407282,0.0928895995,0.6829154491,0.0984284952,0.1808472872,-0.3708581924,-0.0509107038,0.0724437535,0.4867711961,-0.4286133945,0.1816755533,0.1529930085,-0.0166825186,0.1958255768,0.057199575,-0.0741308257,0.2589164674,0.231020391,0.0226330291,0.0371011607,0.2178665698,-0.2101085633,0.1258466542,0.527991116,-0.2242015153,-0.0082037067,-0.1891955137,-0.3130001128,0.3344785273,-0.1324252784,0.6150864959,-0.1887344867,0.0111453244,0.0530636646,-0.1579793096,-0.2006955445,-0.0270678494,-0.2793725133,-0.0978188068,-0.121064581,0.0068969922,0.1456569582,-0.0332153738,-0.4086243212,-0.028823968,-0.0118711898,0.0069377348,-0.1949533373,0.3073679209,0.1759618521,0.049708467,0.0049637621,0.1439480633,0.0891773999,0.0135565326,0.3173201382,-0.0389472879,0.2369768173,-0.2791323364,-0.0835857391,0.4249715805,-0.2261408567,0.2390493751,-0.2577944994,-0.1333370656,0.3874959648,0.0980762988,0.1315007359,0.0421084613,0.2405784279,0.3197242022,0.284244746,0.2978320718,-0.293325454,0.2334150076,0.281560719,-0.3991186023,-0.1306752861,0.1706736535,0.0838242322,-0.2651279569,-0.0376284048,-0.1081735119,0.292342335,0.12961483,0.044421263,-0.3026949465,-0.0511693098,-0.1725466698,0.1334510893,-0.0285677481,-0.0655025467,0.0821654499,-0.0790793151,-0.1506348103,-0.335841924,0.0008955905,0.1101568267,0.4040704072,-0.3022896051,0.2084807903,-0.1015755013,0.0667805225,-0.0115138814,-0.2169559598,0.1370065659,-0.1746398956,-0.2419863492,0.2149674147,-0.0095614661,-0.1359001547,-0.0398203433,0.4075907767,0.1740256548,-0.4136068225,-0.3883339167,-0.2927636206,-0.1515942216,-0.1457689553,0.4902339578,0.1144228652,0.2129697204,-0.0206116792,-0.1446669251,-0.2113123983,-0.1823829263,-0.1195882559,0.0484520197,0.1704612076,-0.1940094233,0.3333062232,-0.544780314,-0.0393551663,0.2153694779,-0.2854782641,-0.266269207,0.3533629179,-0.1288697422,-0.331546694,0.0084265741,0.0862202495,-0.1368310004,-0.2057427615,0.2223112881,-0.0022178309,0.1104766503,-0.2818082571,0.0997146294,0.0512808561,0.4467150569,-0.2579179108,-0.1771113724,-0.2561301887,0.6002789736,0.0168119725,-0.2479852885,-0.0719557405,0.10865134,-0.1167465374,-0.045904845,-0.5427120924,0.2715320587,-0.4648799002,0.4532858431,-0.263738066,0.039822679,0.0943992287,-0.2558395565,0.0357978009,0.234397158,-0.1071772575,0.1661345363,0.0742164031,0.1016753688,0.054877568,0.6813262701,-0.222206369,0.3173042238,-0.2216823548,-0.1420523971,0.3213466406,0.0146830436,0.2703052163,-0.3492729366,-0.1600702554,-0.2527992725,0.1392287314,-0.3603532016,0.1603133827,-0.0825729594,-0.3704121113,0.0258405618,0.0244620685,-0.0270040594,-0.037663769,0.2160638571,-0.0352763757,0.0675722733,-0.0972187966,0.1948964149,-0.361354053,-0.235332787,-0.1472590417,0.0244838595,0.1677017659,-0.2305628806,-0.6383590102,-0.0635867417,-0.4090248644,0.1462840289,-0.1857921183,0.4481492639,0.0315240845,-0.2720533609,0.4575543702,-0.0128171574,0.6455437541,0.3132842481,0.0661056861,0.1790159792,-0.166753307,-0.6522533298,0.051006753,0.2501664162,-0.0701482072,0.3510232568,0.6979834437,-0.0248679779,-0.1324616671,-0.0856419206,0.3518333435,-0.2339815199,-0.5040017962,-0.1591250002,-0.3802190721,-0.1185237244,0.2002323419,0.1421143115,0.5277095437,0.3856259584,-0.0689020902,-0.1792732924,-0.0730502605,0.0671484023,0.0946525931,0.2961640954,-0.0750326887,0.4328200519,-0.2789561152,0.1389795244,-0.1538670957,0.4000518322,-0.1149559468,-0.0283026583,0.3496232033,-0.2214353234,0.091939345,0.431624651,0.2227868438,0.0650333837,-0.112034604,0.2019871622,-0.237281695,-0.021790646,-0.1702501476,0.0353050046,-0.0436676554,-0.1517342925,0.269977808,0.0380760729,-0.0135925841,0.2379070818,-0.0789268315,0.1283086687,0.681194067,0.1408226788,1.191838026,-0.4987128377,-0.0082949074,0.096124053,0.1430038661,-0.1005699188,-0.1519063562,-0.3217864931,-0.3388417065,0.0811910704,0.1482400596,-0.2489883602,-0.1232921258,0.4238460362,0.1637028754,-0.1260513216,-0.2401729822,0.3148695827,-0.3412213922,0.4762751162,0.0087071341,-0.4426178932,-0.4776338339,-0.0112774307,-0.0151017811,-0.0698857903,-0.0370835848,0.1763518155,0.0186255872,-0.1760469973,-0.0399731249,-0.1425021589,0.1810689867,0.1528136432,0.1570997089,-0.2341354787,-0.0617597178,0.4883335829,0.4086743891,-0.0087386305,-0.0027235872,0.2221084833,0.070540458,0.1289339215,0.102866143,-0.2500023246,0.0571480505,-0.1275206953,0.0429312177,-0.1080657989,-0.1555736661,-0.4990603924,-0.1131400913,0.1605971903,0.3514451683,-0.1212956756,-0.2276404202,-0.2395575494,0.2034029812,-0.0467719883,0.0184318721,0.3179139793,0.1737434864,0.2757465243,-0.021259075,-0.2673325837,0.0162612591,0.302054137,0.0170834605,-0.3022372723,0.3309478164,0.0865601897,-0.2429788113,-0.2007220238,0.377785027,0.3468036652,-0.4271838665,0.1152065992,-0.0215389431,0.0429571532,0.0788996667,0.0687574074,0.1683828235,-0.0139961028,-0.1238839999,-0.2372082174,-0.5546423197,-0.0506821424,0.1872413754,0.1470966339,0.2197025865,-0.4283142984,-0.2838523984,0.0944497213,-0.1446280926,0.0714248791,-0.1175397187,0.1290342361,0.0972546563,-0.0604094416,0.0707973912,-0.161698997,0.0573554263,-0.1087406799,0.027835818,-0.0904188156,0.0773825645,0.2150378823,-0.0356082357,-0.2600277662,-0.1834864467,-0.4841841757,-0.1761963964,-0.31145823,-0.1365018189,0.0177764986,-0.048217766,-0.4042716622,0.33324036,0.1770690233,0.0136793824,0.3767220378,-0.0358713344,0.428891778,-0.0598752946,0.123934716,0.2064660639,-0.1862601191,-0.2938393652,0.1813883483,0.341072768,-0.1296020746,0.3239104152,-0.2449260056,0.0658398643,0.2200422287,-0.0047926214,0.1478656679,-0.1227512434,-0.276717484,0.2896311283,0.0481098257,0.0177412834,-0.224945724,-0.0619258843,-0.1368287802,-0.2501651943,0.2854056954,0.356718272,-0.123779282,-0.10790959,0.0449002609,0.5746719837,0.0202325694,-0.4783284068,0.5880820751,-0.0703881532,-0.0385696366,0.2276902348,0.2314764261,-0.0237397086,0.6256372333,-0.2442821413,0.0691806674,0.3443623483,0.6154091358,-0.3641617,-0.3480578065,0.3312956691,0.2959241569,-0.2161132693,-0.1658429801,0.1747482419,0.1850936562,-0.2219771445,-0.132007122,-0.2278194278,0.3173690736,-0.2714668512,0.1096020788,-0.377047509,0.008414289,0.0155454744,0.1957464367,-0.2598244548,0.0120411348,-0.2839374542,0.2939406931,0.0450807363,0.0761367455,0.0730996504,0.0068180226,-0.0483460613,-0.4155160785,-0.0145709105,-0.3586179018,-0.0320335291,0.2837156057,-0.293530643,0.3421538174,0.4888449907,-0.0014262202,0.071187906,-0.0501821265,-0.0979098603,0.0505334549,0.1878017932,-0.119886227,0.2970226407,0.1994648725,-0.1026525199,-0.0517393462,-0.4618648291,-0.1480226368,0.431840688,0.0595424883,0.3271251321,0.1006138623,-0.2961321473,-0.3213715851,0.12290328,-0.0750292018,-0.1075348035,0.2633378208,0.0025236874,0.0956815407,-0.0516838878,-0.0350605883,-0.3884761333,0.5839490891,0.3617029488,-0.3527215421,-0.2170392573,-0.383613497,-0.6794734001,0.0733423904,0.0394686349,-0.0809108615,0.4049057662,0.2302096486,0.4591118097,0.0503456779,-0.1319795549,0.1001781598,-0.1748912483,0.3998596072,-0.0295775123,0.0278941356,0.0828517526,0.1551289558,-0.2668427825,-0.4360180199,0.4621651173,0.207424745,-0.104706198,-0.1270342469,0.0779104605,-0.3120552897,0.3903242946,0.359515518,0.0061716433,0.1256822944,0.342846483,0.1056971699,0.2859561145,-0.0386305265,0.024041947,0.2289869338,-0.0896629691,-0.0777299851,-0.0537523553,0.4169670939,-0.1711568832,-0.0300592054,-0.2994278669,-0.1029417962,-0.1097694114,-0.0798362121,-0.2138459682,0.2689225376,0.3082722425,-0.0050267419,0.2078597993,0.2081071138,-0.2628808022,-0.02121488,0.2440171242,0.0182844475,-0.1810207665,0.0215865877,0.1300424933,0.256203711,0.1501440406,-0.2618189454,-0.1183138639,0.2275496721,-0.1547401696,-0.1117919087,0.451146096,-0.2433207184,0.1806146204,-0.164907068,0.1835401505,-0.0875930339,-0.0238008555,-0.4564522505,-0.2792279124]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2099","title":"load_from_disk takes a long time to load local dataset","comments":"After a synchronous discussion, we found that the cache file sizes have an enormous effect on the loading speed: smaller cache files result in faster load times. `num_proc` controls the number of cache files that are being written and is inversely proportional to the individual file size. In other words, increase `num_proc` for smaller cache files :)\r\n\r\nMaybe this can be highlighted somewhere in the docs.","body":"I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)","comment_length":66,"text":"load_from_disk takes a long time to load local dataset\nI have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).\r\n\r\nDoes anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?\r\n\r\nTagging @lhoestq since you seem to be working on these issues and PRs :)\nAfter a synchronous discussion, we found that the cache file sizes have an enormous effect on the loading speed: smaller cache files result in faster load times. `num_proc` controls the number of cache files that are being written and is inversely proportional to the individual file size. In other words, increase `num_proc` for smaller cache files :)\r\n\r\nMaybe this can be highlighted somewhere in the docs.","embeddings":[-0.3226234615,-0.1608172357,-0.074590832,0.2495029122,0.298553437,0.0387583971,0.422141403,0.282289356,0.7425697446,-0.0142084211,0.1286982894,0.5218667984,0.1710792929,-0.0604943521,0.1026293635,0.1128980443,0.3352093399,0.2497863024,0.4394035637,-0.1052012146,-0.2460135221,0.0447133258,-0.0732722878,0.0021351634,-0.2016454041,-0.0709652081,0.0490918644,0.0603730306,-0.1569585502,-0.3110060096,0.0576225892,-0.0301358979,0.2484993786,0.1560704559,-0.0001240786,-0.25024876,0.0976851359,0.1681643128,-0.2215498686,0.2777531445,0.4692359269,-0.4544668496,0.2095421851,-0.3286706209,0.1093200073,-0.2615490556,-0.2773521245,-0.234493345,-0.1445318758,-0.1392451525,0.0578051284,0.4935595393,-0.3358249366,0.0392284468,-0.1054476723,0.2524732351,-0.2108767182,0.0278607272,0.5619883537,-0.0642874613,-0.1517484784,-0.0723269731,-0.1665252149,-0.0186403897,0.4654624462,0.0913548246,0.3007298708,0.1481363475,0.0586196668,0.198486805,0.6188426614,0.0647190362,0.0020857456,-0.251701504,-0.1446768641,-0.3261279762,0.4937447309,-0.0850370824,-0.2823262215,0.1968884021,-0.0919127986,0.0032242469,-0.0007410151,0.1887155771,-0.1272675246,-0.1371296197,0.0356843844,0.0739296302,0.4173921049,-0.0824339688,0.0908488482,0.0813887864,0.1108416989,0.3806838989,-0.4470407367,-0.2022477388,0.0474867932,0.1267513484,-0.1574438661,0.1289217472,0.250787437,0.2840546668,-0.1623958349,-0.0145501504,0.1943362951,0.4872796237,-0.0886643827,0.0122230453,0.1112783998,-0.284334451,-0.0522237867,-0.0256252363,-0.2161005288,-0.0837232023,0.4585703611,-0.4411866367,-0.2838889956,-0.1548160166,-0.0263332501,0.1521281302,0.5146113038,-0.0873988569,0.2107797861,0.2231996506,0.0778515115,0.6204023361,-0.1289455891,-0.0406199768,-0.2698877752,-0.1998659372,-0.0555292144,-0.3001203835,-0.1892191768,0.2059864402,0.0779303089,0.2282763869,-0.0683338046,0.1639872938,-0.134283036,0.0825811699,-0.112588644,-0.1415144503,0.2363202274,0.2305313349,-0.1208691821,0.178139478,0.1677362621,-0.2345184833,-0.4652573466,0.3355948031,-0.2128035575,-0.4199279845,-0.4273709953,0.0072537852,-0.2377495915,-0.0651967227,-0.3955855966,-0.0914321989,0.1173492074,0.005325845,-0.0422434621,-0.204665482,-0.1664453894,-0.1167344749,-0.0435947105,0.1261857152,-0.378498286,0.2095520943,-0.1697654575,0.2436336428,0.3041909039,0.7334475517,-0.284147203,0.2297366261,-0.2150828093,0.305883795,0.4823344648,-0.0905484706,-0.3905163407,0.6984345317,-0.0315406024,-0.1900236011,0.1222814247,0.5592471361,0.3014048636,-0.1081742421,0.0844724476,0.6842624545,0.1519716233,0.1931607723,-0.4408969283,-0.0889709592,0.0863688216,0.448525548,-0.4481652081,0.2106366903,0.1492393911,-0.0531964451,0.1388078183,0.061864119,-0.08243341,0.2182705253,0.1804098636,-0.0536857285,0.0835708752,0.2880495191,-0.2089420855,0.1803558469,0.3568312824,-0.2001403719,0.1745316088,-0.1622901708,-0.2808536291,0.2578678131,-0.0638185441,0.5850840807,-0.1655812711,0.0002597683,0.1232822016,-0.18954584,-0.1822953522,0.0056731813,-0.2607091069,-0.1338571906,-0.118710041,-0.0538710356,0.1495468616,-0.027684249,-0.3481990397,-0.1139468178,0.0274145566,0.0504438952,-0.230782941,0.3136663735,0.2384157777,0.0728301331,-0.0540893935,0.2536273301,0.1177997589,0.0799255148,0.3854442239,-0.0348918959,0.15779607,-0.2845805883,-0.1133620217,0.474383235,-0.194581449,0.2044995874,-0.2393619418,-0.1376259029,0.4117901623,0.1533876806,0.1523716003,0.0122169489,0.2239199728,0.3204170465,0.3103183806,0.3557278216,-0.2361183167,0.299849987,0.2733623683,-0.3111574352,-0.1637509763,0.1904313266,0.0677721575,-0.3896792829,-0.0465774164,-0.0718714148,0.3178163767,0.1108624786,0.0967897847,-0.3655157983,-0.1263010502,-0.2689044178,0.103578791,-0.0913731009,-0.0813279748,0.126849696,-0.0838834047,-0.1407702267,-0.343701601,0.0768223479,0.0583926924,0.3713030815,-0.3226996064,0.1442296803,-0.1177649572,0.1012655199,0.0854585096,-0.2584552169,0.0992416814,-0.1990335584,-0.1569407135,0.238261506,-0.0595441423,-0.2230942994,0.0155618982,0.3806137145,0.1620953232,-0.3483735621,-0.4424181283,-0.2981950641,-0.1366091669,-0.1440680772,0.40893507,0.1800209433,0.1240882203,-0.0514699854,-0.1754902899,-0.1541039795,-0.14170219,-0.1339687556,0.0971917808,0.163480103,-0.2072190344,0.2906833887,-0.5274925828,-0.0480186194,0.2445218712,-0.3920345008,-0.2549420595,0.3617928624,-0.1478643715,-0.3017109334,-0.0606981777,0.0505279079,-0.1136074066,-0.2817176282,0.2005344629,0.026804233,0.1629438549,-0.1670721471,0.084068574,0.024105655,0.4369485974,-0.2090643346,-0.20050852,-0.3366442621,0.665576756,-0.0674730986,-0.2943997681,-0.1106713116,0.1626203507,-0.0999386311,0.0176796075,-0.4852070212,0.1754689664,-0.3991000354,0.4361282885,-0.1872492135,0.0231979657,0.0902207643,-0.2943100035,-0.0100117195,0.1943797171,-0.0735586882,0.2152554691,0.1402682513,0.0760771185,0.0566534661,0.5815755129,-0.2033928484,0.2937358022,-0.1928665191,-0.144156158,0.3797111213,0.0508842468,0.2769877613,-0.3881303072,-0.038531892,-0.2365735024,0.0751916543,-0.3433196843,0.197209388,-0.0419467017,-0.3988678753,0.0004406806,0.0113648996,0.029694654,-0.0258544181,0.2768124342,0.0031795749,0.0571153983,-0.1580576599,0.1955288798,-0.3689616323,-0.220746696,-0.0347217247,0.0054608728,0.1675373763,-0.226294145,-0.5818343163,-0.0450042337,-0.4573364258,0.1102723032,-0.1511334628,0.4317114949,0.0326875485,-0.3267956674,0.4497900903,0.0165359601,0.5992977023,0.2426706553,0.0653817058,0.1691059321,-0.1939439774,-0.6289978623,0.0009363479,0.2882154286,-0.0558008887,0.3640823066,0.6817572117,-0.0207284279,-0.1427894086,-0.1771029681,0.2964544296,-0.1825656742,-0.4764574766,-0.1564636827,-0.3197120428,-0.0298837181,0.1361918002,0.1223990247,0.4730424881,0.3970288932,-0.0510308035,-0.1397122145,-0.0794909,0.0569183491,0.1025461853,0.3265589178,-0.0831367448,0.4017071724,-0.2665261328,0.1724072099,-0.1436823756,0.3578255773,-0.1594264805,0.0329037718,0.3169455826,-0.1690717041,-0.0378931947,0.4042511582,0.2179452032,0.0230433159,-0.0922689959,0.1755314022,-0.1289856732,-0.0814666748,-0.2395368963,-0.0894291624,0.0621004477,-0.1828180104,0.3073565364,0.033796154,0.0229254048,0.2029400319,-0.1777444631,0.180343613,0.694470942,0.240253374,1.1032589674,-0.5289762616,-0.0497177094,0.0803679377,0.1587390751,-0.0673929825,-0.2195410728,-0.3331917822,-0.3431916833,0.0512360968,0.1407550722,-0.2378757745,-0.0988540277,0.3970498145,0.1244291365,-0.135952726,-0.1707547307,0.2494417429,-0.3618461788,0.4492153525,0.0081473989,-0.4845143557,-0.5036971569,-0.0180476103,-0.0226320438,-0.0701756999,0.0573068634,0.1429814249,-0.0137580084,-0.1082959026,-0.0467247926,-0.1943830848,0.1463376582,0.1657030582,0.2384936213,-0.2413228601,-0.1573362797,0.4588550627,0.4183011353,-0.1278501004,-0.0574063696,0.2050831467,0.0940762237,0.156606257,0.1413860023,-0.3012240529,0.1008494645,-0.1678157896,0.0478643216,-0.1583500206,-0.1808530092,-0.5726488829,-0.0293080024,0.1275885999,0.2592934668,-0.0812684298,-0.2170277983,-0.2679758966,0.1614524126,-0.0560318865,0.0301198959,0.3356842399,0.1979068071,0.2878731787,-0.0600540303,-0.272539705,0.029441379,0.3319970071,0.0695776045,-0.2590417564,0.1596335322,0.0203190837,-0.2458759546,-0.1667045504,0.2685497403,0.2566318512,-0.3890849352,0.1351532489,-0.0054590055,-0.071138598,0.1105740517,0.0917997658,0.2113634944,0.083394587,-0.1534252912,-0.2042268813,-0.5619236827,-0.0693942532,0.1208486483,0.1197357997,0.2027741224,-0.5310309529,-0.2314298004,0.10634505,-0.1486593783,0.0223781522,-0.0514783077,0.0978180319,0.0250890385,-0.0955784023,0.084709011,-0.1104153767,0.024597697,-0.1290598959,-0.0582062006,-0.0986009985,0.0661910251,0.2022105455,-0.0261722207,-0.3134868741,-0.214947179,-0.444640398,-0.2229793221,-0.407894671,-0.1421503276,0.0821997225,-0.1103836745,-0.3081531227,0.3577410877,0.1771010011,0.0394447148,0.4326915145,-0.033210896,0.4170434475,-0.0510408729,0.129196614,0.0842801481,-0.1912926733,-0.2347051352,0.1777189076,0.286891073,-0.1441696733,0.3072478771,-0.1540198028,0.0407077819,0.2540165782,-0.0518829413,0.1273432076,-0.0752437264,-0.2777110934,0.1884882897,0.0585009865,0.031669993,-0.2745830417,-0.0376398936,-0.1089215279,-0.2237673998,0.3037154377,0.589148283,-0.2170902193,0.0305275843,0.0220114235,0.5801923275,-0.0736669526,-0.4540026188,0.5093343854,-0.0273504071,-0.0193563178,0.2056340873,0.1050567701,0.0155768003,0.5493296981,-0.1837938279,0.072378628,0.3812023401,0.6312505007,-0.4592194557,-0.2652798295,0.3273638189,0.3456683159,-0.2631082535,-0.1615939736,0.2169776112,0.2208641618,-0.1987282485,-0.0772714391,-0.2528799474,0.2544390559,-0.2474384457,0.0137341032,-0.3680056334,-0.0047220113,-0.003245082,0.2030934244,-0.2940654457,0.0595923178,-0.2064142525,0.3588228524,0.0225917678,0.0779864863,0.0846003965,0.0032434366,0.0085201478,-0.3517470658,-0.0209236331,-0.3850735724,-0.044468917,0.3163784742,-0.2234604508,0.3441755772,0.470610112,0.0197637379,0.0042301612,-0.0007336148,-0.0443941914,0.0451006964,0.1677299738,-0.1006323174,0.2130613178,0.1323037595,-0.1080196798,-0.0720207021,-0.4006671607,-0.2658493817,0.4388965964,0.0698800832,0.3287579715,0.1410548836,-0.3015426099,-0.2264058292,0.1161170453,-0.0401536673,-0.1296411902,0.2020007968,0.0305200089,0.1341317892,-0.0184480511,-0.0197693314,-0.3499533534,0.6466864944,0.3268618286,-0.3764690459,-0.1538328826,-0.342990011,-0.659437418,0.0901030377,0.1378397197,0.0102191688,0.4335321784,0.2976467013,0.4147016704,0.0565112606,-0.0912942514,0.0680739954,-0.0804568976,0.3307954967,-0.0441674255,0.0997030362,0.0676347092,0.1545778811,-0.2172723562,-0.4932133257,0.4301346242,0.2421676517,-0.0855952501,-0.1617184728,0.0767552853,-0.2224020213,0.5364189148,0.3434116542,0.0456859544,0.0447002612,0.3310735822,0.0513145104,0.1851570159,-0.0491818525,0.0688122138,0.160445258,-0.1198964491,-0.0720121339,-0.1018097699,0.351531893,-0.1771846414,0.0738598183,-0.3217801452,-0.1348325461,-0.1010626107,-0.0538108721,-0.253195256,0.2246579528,0.2578519285,-0.007413249,0.2533705831,0.161356926,-0.2900331914,-0.0014642135,0.1636457592,-0.0152296191,-0.1333995759,-0.0005012888,0.1547702104,0.2593905032,0.2119565755,-0.3160808086,-0.1988023371,0.174123317,-0.1279717684,-0.1273602992,0.4363152087,-0.2384277433,0.2239916027,-0.1776050478,0.172525987,-0.0055292686,-0.0610178858,-0.5113272071,-0.2912894487]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2098","title":"SQuAD version ","comments":"Hi ! This is 1.1 as specified by the download urls here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/datasets\/squad\/squad.py#L50-L55","body":"Hi~ \r\nI want train on squad dataset. What's the version of the squad? Is it 1.1 or 1.0? I'm new in QA, I don't find some descriptions about it. ","comment_length":13,"text":"SQuAD version \nHi~ \r\nI want train on squad dataset. What's the version of the squad? Is it 1.1 or 1.0? I'm new in QA, I don't find some descriptions about it. \nHi ! This is 1.1 as specified by the download urls here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/datasets\/squad\/squad.py#L50-L55","embeddings":[-0.0604724661,-0.5005539656,-0.1113543883,-0.2531746924,0.1078832597,0.0757490695,0.4927109182,0.2018381059,-0.3321818411,0.1766837388,-0.1202347353,0.2396534234,-0.203777656,-0.0142663307,0.2275711745,-0.1260699183,0.2643026412,-0.0178150665,-0.0621315166,-0.2468939275,0.1244260818,0.1808848232,-0.4038811922,0.3772419691,-0.2603084147,-0.1802249551,-0.3102293313,0.0355396867,-0.4338434935,-0.2745416164,0.4582298696,0.2267114222,0.5414133072,0.4453876019,-0.0001107172,-0.2476429641,0.1264922768,-0.0645884201,0.0011971648,-0.3692577481,0.0501793884,-0.2730044425,0.1349951178,0.004150454,-0.0457882769,0.1555507779,0.5160467029,0.0329222642,0.4147241414,0.0608778149,0.1975602359,0.0755674616,-0.0231449176,-0.0101675475,-0.1238896474,-0.1790872812,-0.1439724267,0.0720417053,0.6169973612,0.0176966358,0.312764138,0.0682802498,0.1100431234,-0.393999815,0.2163251638,-0.0185573101,0.1599590778,-0.5074597001,0.0321490206,0.2443792969,0.4826879203,-0.2027590871,-0.4878724515,0.2473998964,0.301202476,0.0714597031,0.116130814,0.3434726,0.0603666864,0.1525558978,-0.3582869172,0.004857487,-0.3315418661,0.1272823811,0.1082721353,0.0368821807,-0.0541833043,0.1033969,0.139861986,-0.0036582579,-0.1931750327,0.0566064678,0.0704858005,0.6720510721,-0.5896235108,-0.3965839744,-0.2825645804,-0.0334492698,0.1003138423,-0.0155720105,-0.1087440923,-0.2207333595,0.0954695046,0.1306281388,0.4844123721,-0.0547729544,0.6630285382,0.1081697866,0.0299204532,-0.0430450626,-0.0602652729,0.2785119116,-0.5207515359,-0.182373032,0.0150943073,0.1009019613,-0.0180764329,-0.3224002421,0.236353144,-0.0500236824,-0.3046161532,-0.1342633218,-0.091306664,0.076814644,0.204405129,0.1669698358,-0.0377823152,-0.1550873816,0.0005208693,-0.2245546728,-0.0964398831,0.1066121832,-0.0729127303,-0.0286777187,-0.0249737054,-0.3783900142,0.085946627,-0.1349927932,0.032957729,-0.1838262081,0.2755532265,-0.1375363171,-0.2517420948,-0.1604972631,-0.0698522031,0.0304595772,-0.1291236132,-0.0982333571,-0.0355510674,0.1320831776,-0.2861220539,0.0421966277,-0.220705539,0.1380918622,-0.0924938023,-0.1291389465,0.0180063229,0.5646390915,-0.1134838685,-0.0551957823,0.0279937182,0.082219027,-0.2973814905,-0.2253068388,0.0687772706,-0.0354451835,-0.7420763373,-0.2849624753,0.4531505108,-0.4215435088,-0.188939333,0.1605757922,-0.0520962328,-0.1387757659,0.1614702493,-0.0348095074,0.7505707741,-0.5568032265,-0.4271907508,0.2451104373,-0.2239862978,-0.6825326681,-0.0616949163,0.1946398169,0.3666146398,0.0152833927,-0.144416362,0.2963422239,-0.1579452455,-0.2114263326,-0.189354524,-0.3943149745,-0.1572930962,0.1219405606,-0.0001907834,0.0519359596,0.0522155762,0.1778580397,0.3875611126,0.2161297351,-0.1114566103,0.1696806401,0.3218328059,-0.0921081752,0.0732910708,-0.2041306943,-0.396163702,-0.0504424237,-0.4237092733,0.0470685363,0.3704512715,-0.1808873862,-0.1910664141,-0.0823048651,0.0850572586,-0.1463008374,0.0854929313,-0.4067727923,0.0662172809,-0.2602149546,-0.4900669456,0.1474868953,-0.142045483,0.0817623287,-0.1132441685,0.2171086967,-0.24693124,-0.1381515115,0.0902544782,0.1726662517,0.0934391692,-0.1572674364,0.2053254694,0.5166586637,-0.1499262899,0.1497205049,0.5995023251,-0.0639196932,0.0713356361,-0.0563653298,0.0805249959,-0.1124596298,-0.074114494,0.100548178,-0.3070874214,0.2903752029,0.1427933276,-0.2065396309,0.0159025881,0.1261915714,-0.0942418054,0.1829054058,-0.2112744004,0.0076797321,-0.0724262297,0.1975722015,0.3583517373,0.0025670051,-0.028386144,0.0715176389,0.4071611762,-0.3727051318,-0.0227079634,0.0086071407,0.0170659851,-0.0620179437,-0.0365042612,-0.1410409212,0.0747909248,0.0067119412,-0.0174533762,-0.0146445204,-0.0817076862,0.04715183,0.0734976009,-0.0602795966,0.1515006125,-0.3605563939,0.2086027712,0.0106547158,-0.0427849852,-0.0485754497,0.0289142765,0.1890627891,0.1163928434,-0.0253824275,-0.1774740815,-0.0563534088,-0.3995397389,0.1482508034,-0.4936044216,0.0401574373,0.0200316794,-0.1972613633,0.6922691464,0.2939473391,-0.0335600749,0.3918246925,0.1358150393,0.2241220027,0.3551130891,0.0566787682,-0.3530104756,0.0939865261,-0.0448667444,-0.3811526,0.0408442728,-0.0316711627,-0.0945518613,0.1562310904,-0.3388020694,0.0907582864,-0.0624302141,0.0612140633,0.1749307513,0.1946676373,-0.5116884112,0.181592986,-0.0039552748,-0.4206486344,-0.3080036044,-0.3722048104,-0.1926862299,0.1191516444,0.0026455766,-0.4208528697,-0.573461175,0.2840429246,0.187936157,0.0836472362,0.2935256362,0.0856368169,-0.1481181085,0.3127865195,-0.2134325355,-0.1985609829,-0.0499730222,-0.5394474268,0.1535696387,0.1311529428,-0.3442728817,0.1530231535,0.199617967,-0.0533059165,-0.3038859665,-0.0736142844,-0.4201351404,0.3028987646,0.0585109293,0.2862521708,-0.0139650945,0.3663473129,-0.3289885521,0.0125921024,-0.0682660565,-0.3335057497,0.0953909457,0.2710407972,0.3975188434,-0.0055530183,0.2795490324,0.3161558211,0.708853662,0.4372851849,-0.0577653237,-0.3681850135,-0.0446289405,-0.1607796252,0.1098805368,-0.2003583014,0.4002752602,0.123064585,-0.1008396894,0.1653744429,0.4074633121,-0.2134417295,-0.2938202024,0.0026979181,-0.1020308882,0.090736635,-0.3995859027,0.2129637301,0.0970633104,-0.1339818984,0.0880589262,-0.2223588824,-0.1587817222,-0.1005864143,0.5886604786,0.0423766337,0.1950947195,-0.5706486106,0.1988033503,-0.3469282091,0.1723811328,-0.0663921982,0.4820694327,0.2072028071,0.1028664783,0.3199673593,-0.0567052253,0.5454041362,0.0732605085,-0.0809019804,-0.0902110264,0.1535128504,-0.2145118862,-0.1749156266,-0.037485186,0.1441420466,0.487067014,0.2909119129,0.1764205843,-0.5303234458,0.429780066,-0.2236730903,-0.0685088113,0.0126375156,-0.3396907449,0.0798800662,0.1776165962,0.074425526,-0.4443281889,-0.2665457428,0.0948800221,0.0609638542,0.386584729,-0.2668601573,0.033959683,0.0428257138,0.0108196111,0.0735151917,-0.1461829841,0.4102562964,-0.1442699134,-0.3158854246,-0.0351240039,-0.1061661094,-0.2361128628,0.3097695708,-0.3020908237,0.4117127955,0.2726761997,-0.2288061976,0.0729606152,-0.2060049474,-0.063817136,0.0218378641,-0.3854987323,0.4973915517,0.2166264951,-0.2300628424,-0.3079767525,0.3631807566,0.1779246479,-0.1107733622,0.0598730147,-0.0480299406,0.0436805412,-0.0075788358,0.0517690554,0.9592048526,-0.30614236,0.1039992049,0.0354341269,-0.1431864947,0.4146847725,-0.1516351551,0.1562862843,-0.4506365955,-0.045740325,-0.1509741247,-0.0957534388,0.2920434177,0.5563539863,-0.3170226514,0.2857271433,-0.0737692416,0.2894061506,0.0784910619,0.2327453941,0.2043851465,-0.0007566347,-0.4698919952,0.16560781,0.06872271,0.1788500845,-0.1883130074,0.1626438349,-0.0264461264,-0.041638948,-0.0937205255,-0.1257178932,-0.0746224746,-0.078438051,0.3415365517,-0.3441480994,0.0262171272,0.5350235105,0.3355273902,0.0644404516,-0.0959001258,0.0829460248,0.0788656697,0.219903633,-0.0589142852,0.260830909,0.2456692904,-0.1754500419,-0.2147518098,0.1978820711,0.159184292,-0.0545234792,-0.3687667847,0.1211358905,-0.0508408807,-0.0547934249,0.1769806892,0.0121883173,-0.0153821697,-0.085386537,0.1707690358,0.3071931601,-0.278967917,-0.0325641483,0.4317191243,-0.259867698,0.193863675,0.5831207037,0.1436671168,-0.4028575718,0.3135014474,0.1310107559,-0.0323764384,-0.2997565866,0.440373987,0.0361823738,-0.0364800878,-0.1354497075,-0.0846361518,0.0729987994,0.2431489974,0.3108285069,0.0951439962,0.1006870866,-0.1681343019,-0.201105684,-0.5263999701,0.0357227363,-0.150444895,-0.0449563749,0.182804659,0.080820553,0.4722096026,0.2067954689,-0.2717354,0.1639797986,-0.2110464424,0.1643192321,-0.1813667566,-0.1159133613,0.0824464932,0.0277165305,0.1009824201,-0.2445933521,-0.4144873321,-0.1730864346,0.0538356453,0.1291409731,-0.1060418561,-0.0481670909,0.0120419934,-0.1089472324,-0.2387551218,0.2363579869,0.0409895182,0.2659821808,0.3076216877,0.3728092611,0.0257739965,0.0242311694,-0.0736459792,-0.174315244,0.2524178922,0.0622518137,0.2003593147,0.2619144917,0.1875299513,-0.1634182334,-0.2380188555,0.3844319284,-0.0285769142,0.3260649145,-0.0365561061,-0.3141112626,0.2010638565,0.0714894682,0.1433418989,0.0951343626,-0.1911011785,0.070818305,0.0355352648,0.1501321495,-0.0355288237,0.1136204749,0.6113937497,0.101858601,-0.1725233197,0.1814536452,-0.0572288521,0.332234025,0.2119911611,0.0072400295,0.1977330744,-0.2837781012,0.3790709972,0.1515924484,0.2920316458,-0.0655525923,0.1719231009,-0.0864864811,-0.2518186569,-0.0075767864,-0.0872733817,0.0912943929,-0.0108024105,0.10657873,0.3674458265,-0.0245535318,-0.1822178513,0.3278936446,0.0375949815,-0.0532659814,-0.0178605411,1.0530567169,-0.2542074323,-0.3908022344,0.0507369377,-0.0707630813,0.112746492,-0.2862549722,-0.3267157674,-0.0138079915,-0.1788100898,-0.0917093456,-0.0422084592,-0.1492134184,0.0318279453,0.1725757718,0.2963972092,-0.081835866,-0.1774877459,0.281730473,0.4079298377,0.0430997312,0.1715651304,-0.5016174316,0.3788875639,0.3945910037,0.0188307874,-0.1128300056,0.115353547,0.0869412571,0.108420901,-0.1222062334,-0.1210893095,0.4018638432,-0.068324849,-0.0545753799,-0.2136568278,0.1968787313,0.0519699529,0.0838856548,0.1141487956,0.1425134093,-0.13680242,-0.4029943049,0.1708858311,-0.3986904323,-0.130026117,-0.1756217182,0.2188029438,-0.2260891646,-0.0502287485,-0.0521228872,0.1876015812,0.0433190018,-0.0505823828,0.0904101506,-0.0961798206,0.2217498273,0.2951085567,0.0415495113,0.1856013536,-0.146798566,-0.1809261441,0.087469697,0.2560961246,-0.0000584814,-0.0843439698,0.2960462868,-0.1819216162,0.2040473372,0.2597268224,-0.3279156387,0.0743372515,0.2223664075,-0.4015161991,0.0096070338,-0.1889804602,0.4410136342,-0.0823227912,-0.0564594716,0.0529348999,-0.1709976345,-0.1236880049,0.1245725676,-0.2162141353,-0.0345906615,-0.2936959565,0.0193859302,-0.2012465447,0.1969224215,0.2457057685,0.0090375887,-0.3789976239,0.1012589186,-0.4191335142,0.0850061998,-0.0169961322,0.1890716851,-0.1598741114,-0.010638007,0.0839643255,0.3623060286,-0.0496716835,0.0147314621,-0.3146959245,-0.1579574347,0.2083927244,-0.0871321186,-0.0394606218,0.3834626973,-0.0895294547,0.0633947924,-0.3692092001,0.1598598659,0.5752273798,-0.2479431927,0.0722916946,-0.4558905363,0.1699466854,0.3161945343,-0.1397292614,-0.5007919073,-0.0761982203,0.2575037777,0.0262941327,-0.1704020351,0.058037363,0.0961531922,0.2460616827,-0.1447620988,0.0379123203,-0.124296315,0.1825109273,-0.0617342256,-0.1933396757]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2098","title":"SQuAD version ","comments":"Got it. Thank you~","body":"Hi~ \r\nI want train on squad dataset. What's the version of the squad? Is it 1.1 or 1.0? I'm new in QA, I don't find some descriptions about it. ","comment_length":4,"text":"SQuAD version \nHi~ \r\nI want train on squad dataset. What's the version of the squad? Is it 1.1 or 1.0? I'm new in QA, I don't find some descriptions about it. \nGot it. Thank you~","embeddings":[-0.1714234948,-0.515360415,-0.1852476001,-0.2383303046,0.0884607956,-0.0129926838,0.5424319506,0.116594635,-0.3603753448,0.2608959079,-0.0646641999,0.2786116004,-0.1885498464,0.0209455956,0.1161852032,-0.0584342703,0.2282135636,0.0222366918,-0.1146634072,-0.2129392773,0.0999723077,0.0379907675,-0.3529395163,0.3193666041,-0.3149598837,-0.2424562275,-0.2066773027,0.0148566766,-0.369700402,-0.240610972,0.3372548819,0.2089513838,0.4845718741,0.4628250003,-0.000104421,-0.2370910197,0.1464423686,-0.0677604601,-0.050247658,-0.2075698376,-0.1065290943,-0.1608659029,0.1176997349,-0.0065859826,-0.1238058954,0.1689900011,0.4667293727,-0.1120547503,0.3734356463,0.0820054412,0.2614960968,0.0296942722,-0.0640115887,-0.1683346927,-0.1251656711,-0.2872431874,-0.2115925848,-0.0211937968,0.6374988556,0.1268024594,0.1762776673,0.0198749322,0.1388837546,-0.3146718442,0.1982913017,-0.0999397486,0.2464962006,-0.4561897516,0.1052529663,0.1958416551,0.6229358315,-0.1254565567,-0.3412466645,0.3464104235,0.2560540736,0.0762374625,0.1049442813,0.250343442,0.0699697137,0.2042286396,-0.4547204971,0.0125869224,-0.3874140382,0.0330562703,0.0727448165,-0.0474397279,-0.0843543038,0.0594703108,0.0874273106,-0.0097987913,-0.0211552922,0.0654827133,-0.0295471251,0.6029179692,-0.5182099342,-0.3541201651,-0.3001096547,-0.0387406088,0.0884136632,-0.0509962924,-0.0545395538,-0.1654292494,0.202159971,0.1375938952,0.4549564719,-0.0535004623,0.6626728177,0.1037234589,-0.1026172414,-0.1666956395,-0.1791886389,0.182067886,-0.4571767449,-0.0989423469,0.1386832446,0.0874435902,-0.1061775312,-0.3737578094,0.2490991354,-0.0015519012,-0.3112854064,-0.1405179799,-0.0950976163,0.1025947779,0.2253739983,0.1115620956,-0.1058472693,-0.2240349203,0.0874696225,-0.2551193833,-0.1656317264,0.0833071768,-0.0774829835,0.0530080721,-0.0597581826,-0.3041484356,0.0949845165,-0.1522014737,0.0611645058,-0.078795068,0.3586582243,-0.1835463941,-0.1328050345,-0.1385039389,-0.019368751,-0.0190452375,-0.2044804096,-0.1003341451,0.038320668,0.1629018486,-0.2262185216,-0.0531199947,-0.2639763653,0.2021293193,0.0362463035,-0.0864667073,0.0630558655,0.5407665372,-0.0998801738,-0.0643514842,-0.0158446971,0.1860176027,-0.2705161273,-0.2066270411,0.0147679094,-0.2327215374,-0.852080822,-0.1360359192,0.34472844,-0.4130936563,-0.0956104025,0.1088803858,-0.0361034311,-0.1303999871,0.2118008584,-0.1113511473,0.7444841266,-0.4506480992,-0.3886448145,0.2432378978,-0.1785425842,-0.7243402004,0.0631781742,0.1784792542,0.4303041399,0.0469117016,-0.1348855942,0.1826411188,-0.268014878,-0.2188429534,-0.2101479322,-0.340521574,-0.1675263196,0.1685580313,0.0218873005,0.0805512071,0.091692403,0.0875545144,0.3154952228,0.1322449297,-0.0975660309,0.127196759,0.3246459365,-0.1444002241,0.0803845674,-0.245249033,-0.3111546636,-0.0692235902,-0.4179562926,0.0332167968,0.583886683,-0.1497004777,-0.2063901871,-0.1736004949,0.152455762,-0.1185747311,0.1779927611,-0.2487881184,-0.0496619865,-0.3022180796,-0.5016999841,0.1341532469,-0.1036704257,-0.0906440318,0.0114734834,0.1671350449,-0.2774255276,-0.1831575036,0.0731937662,0.1134702489,0.0193802267,-0.0808688849,0.1404613554,0.4229572117,-0.1012845933,0.1102973148,0.6376450658,0.0070139309,0.0178934056,-0.0368497185,0.1066468135,-0.155999139,-0.0979810432,0.1072963178,-0.2849812508,0.2198887616,0.1730900407,-0.299097091,-0.0771694779,0.0331895463,-0.0173522551,0.1277958751,-0.1793309599,-0.0097435443,-0.115021348,0.1783090532,0.2629019618,0.0587194711,-0.0451209471,0.0933613032,0.3839700818,-0.2770760357,0.0284894835,-0.1614230722,-0.0205998272,-0.0261256285,0.0427123345,-0.2185187638,0.0623247288,0.0586443506,-0.0232685544,-0.0595391542,-0.0740626529,0.0695311353,0.0979359299,-0.0766934305,0.2846027613,-0.3421801031,0.1466488689,-0.0367652141,0.0070468117,-0.094150953,0.0228145607,0.1838062704,0.0957136378,-0.0891993493,-0.1426150948,0.0362368561,-0.2888900042,0.2374832034,-0.332485646,0.1378794909,0.0969904214,-0.177962333,0.6446098685,0.3295192719,-0.0141349426,0.479950577,0.1988283694,0.3078750372,0.3495235741,-0.0071531939,-0.3481632471,0.170306012,-0.0283797383,-0.3934655786,0.0590568334,0.0426058173,-0.0342531838,0.0563431606,-0.171451956,0.0851569548,0.0399074629,0.0791458488,0.0660279766,0.235010609,-0.5185770392,0.1217958331,-0.0774106383,-0.3218410015,-0.357927233,-0.297372371,-0.237609759,0.0580532923,0.0188650656,-0.4684638083,-0.4601971209,0.2419168204,0.1174012721,0.1056557521,0.2481649369,0.0496431254,-0.1029303074,0.2786078453,-0.2356699556,-0.1401587576,-0.1132727936,-0.5934365392,0.1576788723,0.1258229911,-0.3125293553,0.1086087599,0.2340373546,-0.0036272747,-0.2257950753,-0.0176086836,-0.3208906353,0.3237960041,0.0325689018,0.2206660807,-0.1110534295,0.4283551276,-0.2540647388,-0.0684342608,-0.0665681958,-0.25950104,0.123866424,0.1754167676,0.4057615697,-0.0858163908,0.2219165862,0.2538681626,0.7843418121,0.309835583,-0.1433063149,-0.4000651538,-0.0370393656,-0.2568949759,0.0904726088,-0.2200380862,0.4252043962,0.1005433127,-0.1750196815,0.1763601601,0.2897087634,-0.3356650472,-0.2932011187,0.0290167704,-0.1110203266,0.1230752245,-0.3354628086,0.1490933746,0.1548569202,-0.169345811,0.1181832179,-0.2022531629,-0.1951263696,-0.069854565,0.4263343215,0.0060360609,0.0456249639,-0.6336096525,0.2309988141,-0.1551162899,0.2097461671,-0.0435559489,0.4374475777,0.1169562936,0.1043332443,0.2876336873,-0.056266997,0.4858569801,-0.0416485742,-0.0945996121,-0.0401120223,0.1433017403,-0.109305568,-0.1516136825,-0.1428320855,0.1494316161,0.4162275195,0.3025323451,0.2823834717,-0.4878827333,0.4061513543,-0.2712241113,-0.0545667559,0.0077743232,-0.3203661144,0.2171017826,0.2217337787,0.1314929277,-0.4245090485,-0.2480593175,0.080717057,-0.0842691809,0.3559777141,-0.2366940081,-0.0441769809,0.0703837499,-0.0436443649,0.0340670384,-0.0511960313,0.394548744,-0.0784420893,-0.1809613705,-0.1768200994,-0.0170317292,-0.2084486932,0.3731680512,-0.3617933989,0.4088259637,0.225755915,-0.1925108135,0.0219515339,-0.1458453089,-0.0453781337,0.0505584665,-0.4198434949,0.4224511683,0.0836629793,-0.148741886,-0.2535159588,0.3379790485,0.2240514606,-0.1224400923,0.0815334171,-0.129832238,-0.0088714901,0.0272595137,0.1310092211,0.9608196616,-0.2634598613,-0.0012209301,-0.0917922556,-0.1825740933,0.1662733555,-0.214593634,0.1119847596,-0.4683315456,-0.1872598529,-0.135562703,-0.1195638105,0.3139266074,0.6222885847,-0.3805499375,0.2246611267,-0.1153025776,0.3261021078,0.075203225,0.1890877932,0.1975927204,-0.0263644401,-0.3398930132,0.2196127772,0.153298229,-0.0613716207,-0.1641232371,0.1886736453,-0.0208342783,0.0734556094,-0.0754043832,-0.1709629148,-0.0750265345,-0.0720474944,0.2972912192,-0.3231336176,0.0491749085,0.4743679166,0.3435612917,0.0526937917,-0.0838319585,0.0880582258,0.143313393,0.2203981876,-0.0723084658,0.1579211354,0.2457441241,-0.2456362247,-0.2643068731,0.2704195678,0.0471663363,-0.0815958902,-0.3865659237,0.0922846198,0.0987498015,-0.137210384,0.2705215812,0.0267884918,0.0335369706,-0.0485844277,0.2395222932,0.2621620893,-0.2888078988,0.0753736496,0.5045003295,-0.2171352357,0.1655793339,0.5644489527,0.1865550429,-0.3181385994,0.2368661761,-0.0026420194,-0.108599171,-0.3872972131,0.5186133981,0.1374242008,0.0363850594,-0.1595486999,-0.0890104175,-0.0628306419,0.3202327192,0.3734608591,0.1434310824,0.1121238023,-0.2542220652,-0.1132358536,-0.5211709142,0.0804105774,-0.2353359461,-0.0270661246,0.1459252983,0.016907353,0.3739332855,0.2130668759,-0.3655725718,0.0777895302,-0.2544480264,0.2437032461,-0.2791649699,-0.1939009577,-0.0138213001,0.0146012204,0.1793810725,-0.2287051082,-0.3939008415,-0.2659384608,0.1269683391,0.0941084549,-0.0789844841,0.0531732999,-0.0481770262,-0.0202014502,-0.3425104618,0.3073035479,0.0357332267,0.2016645819,0.3389319479,0.3496212661,0.1437006593,-0.0583138987,-0.190186277,-0.1482762396,0.1773959845,0.0276857633,0.0169538856,0.2441101968,0.1367349029,-0.2202112228,-0.0668240264,0.3997052908,0.0116369138,0.3874320984,-0.0535349175,-0.2534686923,0.2299669832,0.2258441895,0.0398966484,0.0480691046,-0.204673782,0.0291890763,0.0506579876,0.2301028669,0.0288927555,0.0680720657,0.5700932741,0.027243698,-0.1115650013,0.1958630234,-0.0161286462,0.2369382679,0.3314217925,0.0566440225,0.1301088035,-0.3406976163,0.3161947429,0.1158203036,0.234073922,-0.0829931796,0.15963462,-0.0896941423,-0.2120927721,0.0628598109,0.0338217691,0.1318411976,0.1767567545,0.1173140034,0.3610673547,0.0141935069,-0.0797587335,0.2911619246,-0.0665062666,-0.0487135276,0.0760098323,0.849732697,-0.1916416883,-0.4103937149,0.0466994494,-0.1556260884,0.0306623206,-0.1686744541,-0.3954004049,-0.0168653131,-0.1669023484,-0.0990029871,-0.0216988381,-0.0131407864,0.0522024147,0.1410286725,0.2744200528,0.132161364,-0.1108423099,0.3183368444,0.3663989007,0.0715595186,0.1944115311,-0.5904944539,0.3838161826,0.3701199293,0.0920953155,-0.1616401076,0.084704034,-0.0041814707,0.0697556287,-0.0056687612,-0.2041247636,0.3841087222,-0.1233107001,-0.0281528905,-0.2529359162,0.2083767056,0.1251594275,0.0426432043,0.1213947833,0.1521710157,-0.1948195398,-0.413766712,-0.032569252,-0.2447050214,-0.1539545059,-0.0369638056,0.1206081584,-0.1568693668,-0.0542593226,-0.0975768268,0.0528603829,0.0545118302,0.0320172682,0.135773629,-0.087034069,0.235457927,0.1829829067,0.0312993899,0.1885230839,-0.2079709768,-0.0651985034,0.1439738721,0.2786975205,0.0629917085,-0.1673344225,0.3767098188,-0.0939901546,0.2515974641,0.3812958002,-0.3862840533,0.1039398536,0.324944526,-0.4895946681,0.0032963241,-0.0183302015,0.396107465,-0.0652726516,-0.0566048622,0.1018347815,-0.0924760476,-0.0070320475,0.2009110302,-0.3031056821,0.1265840977,-0.2223070413,0.0392791629,-0.3087501824,0.2258552313,0.21238859,0.0904947296,-0.2051653266,0.0956419557,-0.2982998192,0.0625623167,0.1242035776,0.1422182024,-0.1565885544,-0.0109778354,0.1707016826,0.338894099,-0.0615794063,0.0749800652,-0.3519890308,-0.1170929745,0.2585780621,-0.0877475515,-0.0635718331,0.2170905322,-0.1385075748,0.07733915,-0.350949198,0.1904935539,0.5290573239,-0.0716158003,0.0504106507,-0.5144851804,0.154297933,0.3053902984,-0.0951424912,-0.3824816644,-0.1517298818,0.3274199367,-0.0749699697,-0.0745549724,0.0531631224,0.1594296396,0.2648669183,-0.1746937335,-0.0186046958,-0.0871806443,0.202639848,-0.1737577319,-0.1735702455]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2092","title":"How to disable making arrow tables in load_dataset ?","comments":"Hi ! We plan to add streaming features in the future.\r\n\r\nThis should allow to load a dataset instantaneously without generating the arrow table. The trade-off is that accessing examples from a streaming dataset must be done in an iterative way, and with an additional (but hopefully minor) overhead.\r\nWhat do you think about this ?\r\n\r\nIf you have ideas or suggestions of what you expect from such features as a user, feel free to share them, this is really valuable to us !","body":"Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?","comment_length":84,"text":"How to disable making arrow tables in load_dataset ?\nIs there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?\nHi ! We plan to add streaming features in the future.\r\n\r\nThis should allow to load a dataset instantaneously without generating the arrow table. The trade-off is that accessing examples from a streaming dataset must be done in an iterative way, and with an additional (but hopefully minor) overhead.\r\nWhat do you think about this ?\r\n\r\nIf you have ideas or suggestions of what you expect from such features as a user, feel free to share them, this is really valuable to us !","embeddings":[-0.3108787835,-0.3046272099,-0.1025219932,-0.0479800925,0.0687685907,-0.0700554103,0.1910864115,-0.0508346595,0.1361465901,0.1569507569,0.3037062883,0.1940681338,-0.1848988086,0.3730876446,0.3175566792,-0.0077214697,0.1201919317,0.4260940254,-0.230573684,0.0468868725,-0.0075155105,-0.3914999962,-0.102672033,-0.0120189423,-0.023875501,-0.1791452467,0.2561783791,-0.2268669009,-0.3328372836,-0.2920048237,-0.0948429257,0.3715645373,-0.1481718868,0.1622747779,-0.0001207907,-0.0428891517,0.1374643594,0.0590490215,-0.584915638,-0.1030332372,-0.1841097176,0.0425006896,0.4369480312,0.0075864997,0.2101271749,-0.3526919484,-0.1396361142,-0.4950588942,0.1005162671,-0.2342170477,0.1155825034,0.1027221382,-0.3094597757,0.233692646,0.3298903406,0.4536732435,-0.5429467559,0.0720201582,0.2324804217,0.0431366414,-0.1210567653,0.2974909842,-0.2007706463,-0.1528941989,0.5350207686,-0.0998743996,-0.2107012868,-0.4075063169,-0.0728181526,0.3070437908,0.872205019,-0.2314073592,-0.2588221729,-0.3002792597,0.295943886,-0.2492587268,0.0199244469,0.2672727704,-0.1843286902,0.0638791099,-0.411372453,-0.5367267132,-0.3401836753,-0.0841031745,0.5507418513,-0.175000459,0.0372192971,-0.0060784062,0.1979063898,0.0224980544,0.6170967817,-0.2264441699,-0.202296257,0.1820368469,-0.192279458,-0.2122168541,-0.0563994907,0.2362249494,0.1903985143,0.4510350227,0.737575233,-0.0329863988,0.2755361199,-0.103270404,-0.1177371815,0.0825221166,0.1956573427,-0.1914610863,0.234956339,0.174017325,0.2198196203,0.0179999061,0.1257082373,-0.0469559729,-0.1022861525,-0.1341961026,0.1262620538,-0.1515747011,0.0811666623,-0.0285396948,-0.1162555516,-0.0750428811,0.2993290126,0.0555921905,0.1886073202,0.2044251114,-0.1871051341,0.3508723974,-0.0804979801,-0.257153362,0.0088628987,-0.2000311315,-0.1664548963,0.4750286341,0.3624806404,-0.0876034349,-0.1268801391,0.3686203063,-0.061152298,0.2437937707,0.3272447288,-0.2562045455,0.0929731578,0.3177512586,-0.0812246427,-0.2186977565,-0.1118941978,0.2880063653,-0.0830698684,0.3657184839,0.0053953482,-0.5483026505,0.1708895713,0.0928531736,-0.1274678707,-0.024637064,-0.7149603367,0.4606952369,-0.257935971,0.0401662104,0.2783625126,0.1570739448,-0.0480402298,-0.1101604775,0.0858072191,0.37892434,-0.2608045936,-0.0232439507,0.0542487912,-0.1358050108,0.1218103617,-0.0106066307,-0.4702012539,0.2000154704,0.046233844,-0.1331994683,0.5498135686,0.2194538414,-0.1146796271,0.3090109527,-0.0080374833,0.1739728898,0.4368072748,0.1638070494,0.0332268327,-0.0741099864,-0.3550978899,0.1997154057,0.0999734923,-0.1582485139,0.0386440046,-0.2926287651,0.2996139228,-0.0504014269,-0.330348134,0.3204112947,0.1587290317,-0.2219254822,0.0707157254,-0.3053122759,-0.1652635634,-0.1405466348,0.2203620225,0.1397408843,-0.0833456665,0.1656155884,-0.5774102807,0.2231894881,0.1686066389,-0.0816205963,0.0517520197,-0.2786859274,0.2936156392,0.2566470206,0.0069583808,0.2985518575,-0.0226737857,-0.4991947711,-0.0524846353,-0.2828563452,-0.4251045883,0.1305734962,-0.4586176574,0.1097255424,-0.3391103148,0.1836680621,0.3947967291,0.07195732,-0.1009670049,0.0197074916,-0.0954394266,0.1250946224,-0.1820270717,0.1980029643,0.1273067743,0.4344156981,0.0719590932,0.5853416324,0.1416966617,-0.379953742,0.3378413618,-0.2595508695,0.2171443403,-0.0517186411,-0.0917369276,0.218434155,-0.383518368,0.4063928723,0.2279461473,-0.1913320869,-0.004030948,0.3014157116,-0.2829813957,-0.1211414859,0.1992322206,0.0372680277,0.0940754861,0.1732017249,-0.2697510123,-0.0403756909,-0.156816408,-0.0427142791,-0.0762667283,0.1561242044,-0.010928208,-0.3204746544,0.2867370248,-0.0927798003,0.3964141011,0.1984130442,0.0820917413,-0.065176785,0.08171729,-0.2824463546,0.4200906456,0.120518975,0.3258138001,0.0309969522,-0.0969612896,-0.0039797118,-0.4386236966,-0.1053166837,-0.1223819628,-0.1773044169,-0.3413564265,0.0089966953,-0.1540701091,-0.1955712438,0.1467411071,-0.1188895106,-0.1021894589,-0.0126229292,0.2394536883,0.1834588796,-0.10758394,-0.103643328,-0.217333585,0.5270037055,-0.1992479563,-0.0251716599,-0.5137592554,-0.0452125855,0.198899135,0.0472594239,0.0032080088,0.0661687031,0.4163732529,0.1682758182,0.1379339993,-0.3983553648,-0.1744869053,0.1206521466,-0.0368908085,-0.0963435173,0.087049529,0.1624215394,-0.0842902958,-0.1047167927,-0.0965829864,-0.3372974098,-0.0799453855,0.104952164,-0.3326285481,0.2669529617,-0.0691864341,0.0566678159,-0.0044627176,-0.2767613232,0.2775223255,0.0485984571,0.0606902614,0.1481176913,0.0797357261,0.0233636647,0.1676543802,0.0961704403,-0.0521218777,-0.3047828972,0.2845378518,-0.1672820747,-0.1685497165,0.092368789,0.0648944154,0.0349513367,0.563757956,-0.698859632,0.3132865727,-0.1534546763,-0.0188971739,-0.3195340037,-0.1869521737,0.138636753,0.1478307694,0.0079044495,-0.1186840385,0.1993915886,0.0207718965,0.011829799,-0.0454185456,0.1649000645,0.4724437892,-0.2799657583,0.7499019504,-0.2328367084,0.2019324303,0.2562409639,0.0631855726,0.2966492474,-0.2636047304,0.1922911853,-0.2881789804,-0.183678329,0.0131842112,0.0328525044,0.045457907,-0.3248323202,-0.3126922846,0.1917994618,0.127061069,0.0417600311,0.0821791589,-0.8073264956,0.1014949456,-0.0185745899,0.0763842389,-0.1245129928,-0.2342164516,-0.0198268853,-0.1275018454,0.5562236905,-0.1463593096,-0.4232421219,-0.0645832494,-0.2479700148,0.0382358879,-0.0302796252,0.1276204139,0.0367242247,-0.1858836412,0.2239858359,0.2786925733,0.1654665619,-0.0336551033,-0.2804149091,-0.1969596595,-0.1634918302,-0.3167749047,0.1226158366,0.0242010783,0.1295231283,-0.4652240276,0.1284015179,-0.0064168954,-0.213562265,0.0068530953,0.3506622314,-0.0871527046,-0.1996817142,-0.2058728486,0.1956337988,-0.0508984849,-0.0524374396,-0.037870273,-0.2354493737,-0.4645160139,-0.2731612623,-0.2341298908,0.0858574212,0.1269871444,0.0659001023,0.2296323627,-0.0395507663,0.0203174353,0.4007854164,0.0371181741,0.184732303,0.5451728106,0.2208453417,-0.3756820858,0.1030652225,0.0247747488,0.3193539083,0.4243552685,-0.1623535603,0.3445610106,0.0268173274,0.0296845343,-0.2936312556,-0.2795344591,0.0181758814,-0.3120169342,-0.1259585321,-0.6251106262,0.4853060842,0.0108013116,-0.111654155,0.763006568,0.0746065751,-0.176955536,0.3175776601,0.278411746,0.9641879797,0.1181245297,0.1516842097,0.3379052579,-0.1131616905,0.1988106221,-0.2497759759,0.0938189402,-0.0258221589,-0.5687533021,-0.1515208632,-0.2918737829,0.300747484,0.4270749986,-0.0435459949,0.0606684498,-0.2622155845,0.2980580628,-0.0733743161,0.0167884305,-0.2591687441,-0.39747563,-0.4063355029,0.0142313689,0.106700778,-0.2799774706,-0.0834434628,0.147334829,0.2530609369,-0.2305136919,0.0695230141,-0.050354626,-0.3953573406,0.1217973307,-0.2768029273,-0.4214160442,0.1907436997,0.2504668236,-0.2979573607,-0.0829799771,-0.1388768256,0.0645044148,0.4400084615,0.3049492538,-0.4049280882,0.1805903912,0.0650019199,-0.1086579338,-0.1329960078,0.2398073822,-0.1366392523,-0.2205265462,0.0143776145,-0.1716659367,0.3094859719,0.0045681256,0.1812605262,0.1024334133,0.2157661468,0.0518090688,0.0570429787,0.081173107,0.0695529953,-0.0609835312,0.123036027,-0.2393790931,-0.1617502272,0.1064377204,0.2069577277,-0.1210745201,0.3754949868,-0.0528481863,-0.3789214492,-0.2747313976,0.2893602252,0.1397413015,-0.7035720348,-0.0738162026,0.1489484161,0.0486304462,0.0033603311,-0.0001007464,0.0190453064,0.0578898564,-0.1634994149,0.1071622521,-0.2024370432,0.185176298,-0.0340727568,0.2404623181,-0.0695147887,-0.7264934778,0.1870399117,0.1360985637,-0.1877744794,-0.2514925897,0.4108295143,-0.0702515319,-0.0630376488,0.3864993155,-0.1346450597,-0.0181869883,-0.030204216,-0.172129631,-0.1750432104,-0.0975412205,-0.0544017777,0.2513286173,0.0059325667,-0.0712464154,-0.2988918126,-0.2655926347,-0.3327393234,-0.116099827,0.2361611426,-0.1152721047,0.014365199,-0.217025727,0.4655834436,0.0696041882,-0.4628807604,0.2213396728,0.0723901689,0.3732800186,0.1005705744,-0.0579679646,0.2426191121,-0.200786829,-0.033478573,0.1931098551,0.1166947111,-0.0207371395,0.0865891501,-0.0320582688,0.2666085362,-0.2654092908,-0.066825904,0.4697323442,0.1004367992,-0.4031161666,0.3078159988,0.0776486099,-0.1227740273,-0.2271872014,-0.0982635841,0.0789744332,-0.1092831939,0.2004919499,0.3673202991,0.0407642946,-0.1634551287,-0.0276837554,-0.3696680963,-0.0030445878,0.2932275832,0.4805445969,-0.0604712181,-0.0904304385,-0.08803197,0.3710192144,-0.0625513494,-0.1868816763,0.1075199619,0.4149567187,0.3388376832,0.2124043554,-0.0217889193,-0.2971410155,-0.4956049323,0.019676635,-0.3421946466,0.0325548276,0.1744559854,0.1863129139,0.4558748007,-0.2427102029,-0.2187995166,-0.206807375,0.139927417,0.0624872744,-0.177453503,-0.1007472277,-0.1063495576,0.0197084341,-0.0756954178,0.3763749003,0.2480235845,0.0486603007,0.2859090269,-0.079163149,0.5971517563,0.2496227175,0.329076767,0.1091702506,-0.0760495812,-0.059475109,0.2214472294,0.0817488879,0.1819231212,-0.2918079495,0.2225530744,-0.0516973138,0.0093403878,0.3133381009,-0.0394246355,0.1230923682,0.3916559815,0.4307102561,0.0155740017,0.2503002286,-0.0300436076,0.0618177578,0.2081743181,0.1110254452,0.2306712866,-0.2183971405,0.0934884325,0.3168645203,-0.2441274226,0.156606406,0.016038252,0.0042060916,0.0520986356,0.5282915831,-0.3212019205,-0.2864100635,-0.1384579837,0.0164806694,-0.0730280727,0.7810717225,0.0189245697,0.2590923607,-0.083041653,-0.212101981,-0.2754499614,-0.2016729563,0.0507686511,0.3989797235,0.4041068256,0.0457767732,0.2310047001,0.4444927573,-0.0495433956,-0.4349990785,-0.0837579519,0.1848332137,-0.0879783854,-0.1295755953,0.0880278721,0.4434053302,-0.5368415117,-0.1422273219,0.0875879601,0.0825187489,-0.104125306,-0.0113941599,0.1107940152,0.1597875357,0.354370445,0.4221051037,-0.052865766,0.0254524872,0.0959211364,-0.005467834,-0.0167448577,-0.4153028429,-0.0394845866,0.3166199625,0.1445341408,-0.1955688894,-0.0833493471,0.2049835771,-0.0062324703,-0.0104393447,0.090666309,-0.0033311136,-0.3901956081,0.0468915142,-0.0875860453,0.1657284945,-0.0493021943,-0.0677297115,-0.0582719892,0.0354527757,0.0463262126,0.034199141,0.4161968827,0.0515393391,0.0287154987,0.1138913408,0.2947539985,0.5818417668,0.0063396422,-0.4682577252,-0.0402445905,0.2554460764,-0.4561417699,0.2449321002,0.5373560786,0.0233583394,0.0914055184,-0.194437921,0.3159258962,-0.0879743919,0.1385931671,-0.2600429654,-0.4006445408]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2092","title":"How to disable making arrow tables in load_dataset ?","comments":"People mainly want this feature either because it takes too much time too make arrow tables, or they occupy too much memory on the disk. I think both the problem can be solved if we provide arrow tables themselves on datasets hub. Can we do this currently @lhoestq ? \r\n","body":"Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?","comment_length":49,"text":"How to disable making arrow tables in load_dataset ?\nIs there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?\nPeople mainly want this feature either because it takes too much time too make arrow tables, or they occupy too much memory on the disk. I think both the problem can be solved if we provide arrow tables themselves on datasets hub. Can we do this currently @lhoestq ? \r\n","embeddings":[-0.2959380448,-0.2490315884,-0.1036210507,0.2489057779,-0.0450339466,0.1605256647,0.2928253114,-0.0211192481,0.3396778107,0.2951470315,0.0326353945,0.4835270643,0.0262511838,0.0174946953,0.2436440885,0.1563296914,-0.0142451841,0.5176954269,-0.3697742522,-0.0229949504,-0.0617189966,-0.3304600716,-0.0252592657,0.1437551677,-0.2534678578,-0.1641295403,0.1912166774,-0.2878650427,-0.404660821,-0.2799391448,0.3027493656,0.3371615708,-0.1649561822,0.1134256497,-0.0001224674,0.0932132602,0.1214200929,0.2131106704,-0.4440474808,0.0841318667,-0.2974019051,-0.0556284562,0.4793878496,-0.0023389403,0.2543599308,-0.2447424382,-0.2770701349,-0.3397575617,0.1714115143,-0.1323938221,0.1215376779,0.112274453,-0.1933120042,0.0961996689,0.454631269,0.3035830259,-0.5730573535,0.1401541531,-0.038998995,0.0041109356,-0.2913177013,-0.1048902199,-0.0712384135,-0.130702123,0.3765466511,-0.10007447,-0.2564789355,-0.3019478619,0.0996722877,0.2157416642,0.7043825984,-0.5096978545,-0.1869640201,-0.296082139,0.3445961475,0.1056356579,0.2261914611,0.2963042259,-0.1142078191,0.1142809018,-0.2742856145,-0.673818469,-0.2642534375,-0.1397593468,0.3414140046,-0.1845974624,0.0385366976,0.0644926801,0.1961012483,-0.0430389866,0.7297346592,-0.2825857401,-0.3581550419,0.1335631609,-0.2341220379,-0.2815178037,-0.2018962651,0.4096239805,0.290076673,0.2482919395,0.4759663343,-0.0250483006,0.0451112874,-0.0968000963,-0.1799183935,0.1840306669,0.3575602174,0.0966903418,0.2082593292,-0.0555774756,0.2637524903,0.052723296,0.0418464057,-0.3272048235,-0.0190531276,0.0066924822,-0.0198303405,0.0030468088,-0.007712564,-0.0944463387,0.1107688099,0.1129938141,0.3573161364,-0.0435708389,0.2807615101,0.0208206773,-0.1791737676,0.3096635938,-0.1984789371,-0.2092533857,-0.0486997217,-0.0780898333,-0.1887566298,0.4118489325,0.3620302975,0.083716251,-0.2157825083,0.4662021995,-0.1323503703,0.0752661526,0.2060665637,-0.2312607169,0.0327525847,0.4121606648,0.0306291562,-0.1248455942,-0.1179491505,0.1336194873,-0.1240875646,0.4790473878,-0.2371688932,-0.3656114042,0.0523756742,0.0470530055,-0.092440933,0.0502534062,-0.8543148041,0.1679319441,-0.1389351785,0.3042337,0.3795031607,0.0627918616,-0.1190134808,-0.1159868836,0.0628859475,0.4255441725,-0.2484682351,0.090476267,-0.0120127928,-0.2226300836,-0.2038796246,-0.0825044364,-0.4281786978,0.287289232,-0.0954106823,-0.0719880536,0.4668298066,-0.024769485,-0.1460669786,-0.0146440296,0.0894162953,-0.1040178314,0.1017408445,0.0415880531,-0.0115556549,-0.0248694289,-0.2259537578,-0.0121472236,0.1868849099,-0.2042139024,0.1919275522,-0.3240942955,0.059901882,-0.0337832235,-0.2282293141,0.2141553611,0.2180252075,-0.1546841264,-0.0180260744,-0.1164072081,-0.0072901039,0.0910575092,0.0413847864,-0.0866690725,-0.0808133706,0.0877005234,-0.6414068341,0.2503461242,-0.1015685126,-0.0174977668,0.0042331922,-0.2572941184,0.1323525161,0.2630307078,-0.1018625572,0.5300439596,-0.0298506264,-0.7860901356,0.0059396685,-0.2782944739,-0.2895979285,0.0892977566,-0.4312280118,0.2539480329,-0.3105936348,0.2821172476,0.316422224,0.0983440205,0.0581024401,0.1112550497,-0.0316416584,0.3322236836,-0.1496320963,0.1671593785,0.1781448573,0.1764755994,0.1676577181,0.3872046173,0.1097668856,-0.3977028728,0.4344574511,-0.4769638777,-0.0125671364,0.0138704302,0.0165406018,0.1664523482,-0.3843301535,0.3188412488,-0.0439417958,-0.0056265057,-0.0875653252,0.2539410293,-0.1913870871,-0.2431895435,0.0255522653,0.3106911778,0.3294464946,0.3639852405,0.1657052934,-0.1471230537,-0.0390839428,0.0423742421,0.1576631069,0.1884934306,0.0106430501,-0.1029715091,0.1735065728,0.0350404494,0.241239816,0.1518500894,0.0659835339,0.0166904256,-0.1517919749,-0.2184060365,0.2604802549,0.0894291848,0.3960984945,0.0850111246,-0.0699949786,-0.0542901345,-0.4305777252,-0.0012596905,-0.2288176864,-0.1033506691,-0.3287900686,-0.1780311316,-0.2584324181,0.0095932512,0.0861177444,0.0227590278,-0.4254294336,0.0443890058,0.2645111084,-0.1380120665,0.0785575062,-0.0195887592,-0.2097582966,0.4630798697,-0.1646036953,-0.0274950657,-0.5118434429,-0.024354402,0.243795976,0.0163122471,-0.0275832489,0.1487127393,0.2704418302,-0.0155711165,0.2265708596,-0.3180449605,0.1024793983,0.1528233439,-0.0561017171,-0.1438629329,0.2807646692,0.1074606329,-0.1540966034,0.2022063881,-0.1312025338,-0.1807152331,-0.0840286463,0.0891090631,-0.3830259442,0.2847309411,-0.0689710751,0.1236855239,0.0513851829,-0.1602576077,0.2237327546,0.0258761458,0.1076872125,0.147577107,0.1839425117,-0.0231058877,-0.0198115297,0.0666629896,0.0129089206,-0.3925453424,0.4017270803,-0.2407953739,-0.0834747702,0.2055182159,0.1197904944,0.2135044187,0.3010386825,-0.5540850759,0.3444393277,-0.3526318371,0.0488325916,-0.2269196957,0.0578687191,0.1938934475,0.2327788323,0.1075320542,-0.2064687461,0.0343337543,0.2128235698,-0.0164831784,-0.209298417,0.0607978925,0.2286076397,-0.3945918381,0.3263339698,-0.1310033649,0.1730514467,0.1884269118,0.0671583116,0.4778157175,-0.2602343857,0.2625123262,-0.4720084667,0.0513146147,0.1655120701,0.0642235056,0.1445962042,-0.091688551,-0.1687663943,-0.0959242657,0.1636490524,-0.1391070336,-0.1692537516,-0.8409453034,-0.0066028,0.0214685164,-0.0270963125,-0.1616044194,-0.2039076239,0.0421282947,-0.1729838252,0.4614591002,-0.1603586674,-0.4641271234,-0.2199057788,-0.1615909189,-0.0332779735,-0.1495441794,0.1606397927,0.0686803237,0.2121348828,0.2860306203,0.3172592223,0.1961809993,-0.1700156182,-0.1128938124,-0.2017814964,0.03975223,-0.5741522312,0.2088984549,0.1850617081,0.203901872,-0.6787978411,0.1030543298,-0.1165687218,-0.2855761051,-0.0235273223,0.4652271569,-0.2474166304,-0.2797547579,-0.2131309509,0.1208001152,0.0432896689,-0.1333009005,0.0193479843,-0.2369774282,-0.3839558959,-0.3204476833,-0.2770613134,-0.2282411307,0.164760381,0.1295913458,0.2308636755,-0.008403603,0.0389993824,0.2662968636,-0.0761282966,0.3563057184,0.6240321398,0.0110756494,-0.3349001408,0.0865763128,-0.1905127019,0.23693569,0.0228434,-0.1593130678,0.348282516,-0.2124980241,-0.1461246312,-0.2357016206,-0.3540178835,-0.1109113842,-0.258676976,-0.2403573245,-0.3882800043,0.4413147867,0.0731975287,-0.0762721077,0.5744847059,0.394292146,-0.3450450599,0.3246238232,0.4225629866,0.7420523167,-0.1131470129,0.0360067226,-0.0178501904,-0.0810802206,0.1877836734,-0.1822287291,-0.0467945635,-0.2022935599,-0.3159269989,-0.163257122,-0.2784869671,0.344740659,0.6192927957,0.2166482955,-0.0175142158,-0.3532562554,0.1148602143,-0.1342864484,0.082370095,-0.4092534781,-0.3848981857,-0.3651781678,0.038949836,0.1036935449,-0.2788300216,-0.0092845028,0.3134475946,0.1003666595,-0.1381445378,0.0320032574,-0.1144309118,-0.2251799405,0.1295299083,-0.0997898281,-0.2469804436,0.3995575011,0.37689206,0.0459101573,0.0989437401,-0.0909538716,0.0764317885,0.183638528,0.134440437,-0.3134602606,0.2698400319,-0.0851521045,-0.0682081208,-0.2274216712,0.1610963941,0.0598653182,-0.2559136748,0.1703696698,-0.082758598,0.0910622403,-0.0907437503,0.2708331048,-0.0857185721,0.3001690805,0.0072661927,0.0426113158,0.165231809,0.2661094964,0.1643645912,0.0870302394,-0.3796955347,-0.0499842837,0.0777693465,0.2479712069,-0.0401925743,0.2925239503,-0.056991417,-0.3208400905,-0.3159715533,0.390805155,0.3319867253,-0.8370425701,-0.1128989235,0.1087167934,0.0598225705,-0.0825120956,-0.0057955887,-0.0543951355,0.1286134869,-0.0054598288,0.081138216,-0.1605966538,0.2510657012,0.0607836954,0.3145678341,0.0369502716,-0.639361918,0.0692764744,0.0838358104,-0.1256213188,-0.0904996917,0.2313766479,-0.1532007456,-0.4022028744,0.4075799882,-0.1101997867,-0.1579757929,-0.0587483756,-0.0998657122,-0.0359668396,-0.0815389082,-0.0636892244,0.1724896133,-0.0877542421,0.1102741137,-0.30566746,-0.37975806,-0.2704481781,-0.1719855517,0.2031867057,0.0923337117,0.0228987616,-0.0289879926,0.7183502316,0.0173951071,-0.4473558664,0.0990376621,0.1887394786,0.334504962,0.0438572951,-0.0104972841,0.1380794793,-0.1125099957,-0.0017856614,0.2892638147,0.118403621,-0.0798297301,-0.0856045783,0.0595137477,0.2994785309,-0.1719048321,-0.0529755987,0.3663614094,0.0731238574,-0.3403516114,0.1992560923,0.049078498,-0.0033569776,-0.1472725123,0.1413505524,0.0453236401,-0.0976066068,0.27928406,0.1008361727,0.1894990206,-0.3241706789,0.0541471206,-0.1422611177,0.070430845,0.0356506258,0.4357763827,0.0052406429,-0.0888255984,0.2888085246,0.2855095863,-0.0374678522,0.0688914359,0.1648097485,0.5121223927,0.4532750547,0.1403604299,0.1436410695,-0.1542581171,-0.3748068511,-0.1003770381,-0.1214207411,-0.3154938817,0.1127622426,0.135370478,0.4430636466,-0.207692489,-0.2558024824,-0.1439390928,0.0078850295,0.0213687941,-0.21267654,-0.0534501523,0.0525829494,-0.1347836554,-0.0791001618,0.1382482499,0.1554531604,0.0106589468,0.2289693356,0.0903938711,0.65400666,0.3387284577,0.0953362063,0.2242257297,-0.1946869791,-0.252174437,0.2352359891,0.0463108718,0.2704990208,-0.3665053546,0.0616069436,0.0022661549,-0.1412227899,0.3014585674,-0.0397314467,0.0251538642,0.3092958927,0.1441100538,0.1605692357,0.2176132053,0.0162470564,0.0413007662,-0.1807843894,0.0200513955,0.4210734665,-0.1039447486,0.2098526359,0.2836740017,-0.2959513962,0.2372809052,0.1531402916,0.1266164482,0.0172361284,0.6544238925,-0.3165145814,-0.3685070872,-0.4101751447,-0.0002380799,-0.1131397635,0.7101011276,0.0155703751,0.3783337176,0.0592983514,-0.2251211852,-0.1065513864,-0.1425494701,-0.0313019566,0.2766107619,0.3308057487,-0.1430417001,0.0999976993,0.3642890155,0.1951114684,-0.1251500249,0.1098169386,0.0106137041,0.0004843683,-0.0219854061,0.0068768524,0.2887131572,-0.4466703534,-0.2600390315,0.0182090309,0.2909234464,-0.1568896174,-0.0491016731,0.1367696673,0.3282163739,0.2682012022,0.435329318,-0.0235375464,0.1507951617,0.1068361551,0.0892709866,-0.0473043844,-0.5445352793,0.0005119736,0.2407065034,-0.0126558943,-0.3778368831,-0.1757321805,0.1783486456,0.1030278131,-0.1377818584,-0.0067405696,0.4482305646,-0.2071400881,0.0878022388,0.0554711185,0.0166326296,0.0253275577,-0.1492112428,0.1007248908,0.1356732845,0.1847242862,0.0047007301,0.2120039165,0.1417589635,-0.0096699689,0.2286661118,0.3601230383,0.2433695346,-0.3316451311,-0.4114707708,-0.0757463202,0.3266933262,-0.3170722127,0.0737631321,0.3929124177,-0.0987640843,0.2164562792,-0.0882342309,0.6065359712,-0.143599391,0.2302938104,-0.2816718817,-0.3595016599]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2092","title":"How to disable making arrow tables in load_dataset ?","comments":"@lhoestq I think the ```try_from_hf_gcs``` provide the same functionality. What all datasets are available on HF GCS? Are all the datasets on huggingFace datasets hub are made available on GCS, automatically?","body":"Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?","comment_length":31,"text":"How to disable making arrow tables in load_dataset ?\nIs there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?\n@lhoestq I think the ```try_from_hf_gcs``` provide the same functionality. What all datasets are available on HF GCS? Are all the datasets on huggingFace datasets hub are made available on GCS, automatically?","embeddings":[-0.0699815601,-0.5376306772,-0.0606945418,0.2699545324,0.0311884806,0.1857694387,0.204721123,-0.0738599524,0.5251614451,0.1915540397,-0.2083578557,0.2142614573,0.1429023594,0.0653284416,0.2521748543,0.0714230388,0.0697012991,0.3191800117,-0.4851082563,-0.2686434388,-0.1109206453,-0.03626002,-0.1380767375,0.1288930029,-0.2874598503,-0.2463818341,-0.0120837195,-0.0107512623,-0.2599879205,-0.2313734889,0.3290418684,0.3906885684,-0.0985376686,0.2785464227,-0.0001202594,0.0522382483,-0.0483481288,0.0375811793,-0.4667463899,-0.0595739074,-0.0433042981,0.0232827868,0.5407465696,0.1272079498,0.0793076903,-0.2443219572,-0.2729644477,-0.3654932976,0.3282220662,-0.0181431919,0.1577931494,0.3644492924,-0.2300797403,-0.131116733,0.2805165648,0.1233369708,-0.4778993428,0.2537664473,0.0540299267,0.0632021576,-0.161352694,-0.0243520178,-0.1481767595,-0.0582384244,0.3744199872,-0.1106753796,-0.2596371472,-0.4817552865,0.1657986343,0.4390707314,0.781811893,-0.3694945574,-0.0524643138,-0.2837283015,0.1551827043,0.1223525554,0.1964025348,0.3899795413,-0.0861202106,0.1196514368,-0.1751600057,-0.515737474,-0.2267815024,-0.2266264409,0.3075189292,-0.1318724006,-0.1225428358,0.0585406087,0.4478563666,0.1124921814,0.3323293626,-0.1756272614,-0.4155189395,0.3108769059,-0.413418144,-0.2898392677,-0.1153087616,0.4277424812,0.3985426128,0.2306174189,0.317887336,0.1235973686,0.0537714101,-0.0411233418,-0.0586042218,0.1582379788,0.1617207527,0.0768049657,0.0983530208,-0.036856506,0.1641530693,-0.0357238688,0.0670006722,-0.3181801736,-0.2331126928,-0.0434866585,-0.0704027712,-0.1852671802,-0.073716417,0.0096376333,0.3312657475,0.0439499468,0.3386299908,0.271124959,0.3304340243,-0.0882129744,-0.1544590443,0.3299454153,-0.2381676435,-0.2385379225,-0.1456321627,-0.1315277219,-0.2117994279,0.4393602908,0.2759017348,-0.0501812212,-0.1849318743,0.4688886106,-0.0440242849,0.1317591965,-0.0492257662,-0.303842783,0.2460277081,0.4969675541,-0.0509323329,-0.0564777702,0.1598425508,0.2020442188,-0.2483125627,0.1833151579,-0.3329205215,-0.3900697231,-0.0035929398,0.032222718,0.0412290543,-0.0404381827,-0.8469280601,0.3460439742,-0.3431281447,0.3712806106,0.4440469444,0.1190370098,-0.0005574742,-0.1731143743,0.0333195813,0.5034125447,-0.1700749546,-0.0085305916,0.1776849329,-0.3054769635,-0.1037433594,0.0439892076,-0.3630769551,0.0695024729,-0.1272325665,-0.0857857466,0.3792748153,0.0228773598,-0.052542299,0.0845265687,0.1702493131,-0.0148179904,0.0711665228,-0.1773425341,-0.0218242295,0.0199649539,-0.4476978481,0.1567911506,0.2190752327,-0.1982150972,0.0974611491,-0.2478875816,0.0965178534,-0.0324610323,-0.1310712397,0.2336816341,0.2506481409,-0.5099535584,0.135522902,-0.1414466202,-0.1844237447,0.1786037683,0.0470607132,-0.0133712832,-0.0411119349,0.17425403,-0.7623881102,0.3165916204,-0.1788858622,0.1900544316,0.0594747365,-0.1816143692,0.1207243353,0.2537692785,-0.102853626,0.3356228769,-0.0349904932,-0.5231173038,0.0846730992,-0.3067087829,-0.4564953744,0.3038044572,-0.3575839102,0.3665586114,-0.1958850771,0.315084964,0.3337168992,0.0993265584,0.0923652723,-0.0198599678,0.106133312,0.0853918791,-0.1777103841,0.2503392696,0.1215639114,0.3644580245,0.2029536664,0.3083251417,0.0576008298,-0.4684823155,0.323420614,-0.3931081295,0.0375391282,0.1042507961,0.0386784077,0.3395082355,-0.2522573471,0.3089871407,0.1465860456,-0.066014573,-0.0955755487,0.1360981166,-0.1097238511,-0.1193592921,0.1630503088,0.1692170948,0.3703482449,0.364467144,-0.0822132975,0.0035113879,0.0637509376,-0.1441782415,-0.0282288883,0.1420616806,-0.098956883,-0.2244120836,0.3118466437,-0.2024132013,0.0964982882,0.193342939,0.1441769898,0.0530510619,-0.2020272613,-0.2763196826,0.2760865986,0.0468645208,0.3984256387,-0.0255295113,-0.0146948826,-0.0302966777,-0.6330464482,0.1473524868,-0.2354974449,-0.1206886396,-0.3019895554,0.0783719271,-0.3228674233,0.0418340079,0.1446887255,0.0071426434,-0.567517519,0.0022964124,0.2356795669,-0.040466506,0.0701785386,-0.0669188276,-0.3137353063,0.6049495935,-0.2892213166,-0.0028113348,-0.5583313704,0.0597380102,0.0854417458,0.0307134129,0.0690543875,0.2096919268,0.341091603,0.0348372832,0.0712579191,-0.2587764263,0.1375205517,0.3671420217,-0.1972143054,-0.0893801972,0.0204221793,0.1433391422,-0.2664320171,0.2388625592,-0.1036956236,-0.4493598044,-0.0378563739,0.1693824083,-0.4139093161,0.172244221,-0.032829769,0.0564120971,-0.086694181,-0.1806505471,0.2011324614,-0.0130633609,0.2047112137,0.3151851892,0.0498370714,-0.0383823663,-0.1923555136,0.1486049592,-0.1679401547,-0.4587257504,0.3519231081,-0.1713169068,-0.1477936953,0.2070529163,0.0551617034,0.1059929654,0.1161226481,-0.609770298,0.2893597484,-0.2824665606,0.0281240512,-0.0854872093,0.0346316695,0.3091964722,-0.0576828122,0.0163572039,-0.2320910543,0.0540945902,0.1564190984,0.0065894662,-0.0831755027,0.0299194176,0.0104064625,-0.3959492445,0.4441452622,-0.038936194,0.1681032479,0.3979781568,0.1006596386,0.4846961796,-0.3332364559,0.1686082333,-0.3442380726,-0.0842414275,0.1325339079,0.1842341274,0.1962065101,-0.053755112,-0.1722316742,0.0105929114,0.0044036312,-0.221876055,-0.0611205772,-0.5079457164,0.0127447974,0.0026007621,0.0114912242,-0.1859084666,-0.1508384347,0.1693353504,0.1334264725,0.5717545748,-0.1340471953,-0.4638620615,-0.0279990807,-0.3130036294,0.2442297339,-0.2751083672,-0.0359079614,-0.1184326559,-0.0128294528,0.3601954877,0.2560434043,0.3104082048,0.0107170697,-0.2367504835,-0.0191601384,-0.0265234988,-0.5357001424,0.1173598394,0.1510366499,-0.0022137756,-0.4368471801,0.1654154062,-0.1191512868,-0.5349696279,-0.0801898688,0.2858907282,-0.1413327307,-0.2020368427,-0.2497860789,0.0934698209,0.0612705126,-0.1175012141,-0.0457975231,-0.3165153563,-0.4211547971,-0.0554826595,-0.0705459043,-0.1286263615,0.2033261359,0.1038831174,0.2141490132,0.1060634404,-0.0934167355,0.5110619664,0.1661100537,0.3581534028,0.9107125998,-0.0446407124,-0.2656352222,0.1637275815,0.0344621465,0.2740888596,0.3277547657,-0.0091777937,0.3827396929,-0.0922224969,0.0102124736,-0.2542273104,-0.3056763113,-0.136502713,-0.1869210005,-0.3803131878,-0.316940099,0.5907815099,0.0083095161,-0.231743291,0.4854333997,0.5626702905,-0.2878555655,0.3857735395,0.2358469069,0.9838715196,-0.1772717535,0.0812229812,0.1337614954,-0.0976820067,0.1905500144,-0.2969303131,-0.0419341698,-0.2447943985,-0.3998810351,-0.2160449177,-0.2909178734,0.3766040504,0.3890935779,0.1521084607,0.1069706678,-0.3283910751,0.1717345268,-0.1524013579,0.0609430559,-0.358985424,-0.4290205836,-0.4078108966,0.0921133012,0.1940683722,0.0279372018,-0.0268977042,0.1983849704,0.1167579517,-0.1925660074,-0.0727301911,-0.0872986093,-0.3718625903,-0.0995277688,0.2164990753,-0.3137662709,0.3319266438,0.0594871566,0.2201473415,-0.0738487169,-0.0867242068,0.1555140465,0.1360012144,0.1661960036,-0.4536583424,0.2146769166,0.200830698,0.0410224423,-0.2308662087,0.1811324358,0.0421470292,-0.2422100902,-0.0079100067,-0.2315262109,-0.0648975074,-0.0798667148,0.2335766107,-0.0316583179,0.3374997079,-0.0101270741,0.0579020567,-0.0249845181,0.1941373795,-0.0251550823,0.2263510674,-0.3452513814,-0.1922566891,0.0684185028,0.1450202316,-0.2188827991,0.2154673785,0.2166032195,-0.3611484766,-0.3672829866,0.1145815104,0.5732604861,-0.8200086951,-0.0353960618,0.1304689199,-0.020137243,-0.1083753183,0.2939601541,0.0920678079,0.0358148254,-0.106462568,0.0567696728,-0.2458814085,0.2419513166,-0.1274214834,0.3390811384,0.0505640879,-0.5303604603,0.1395341754,0.1860040128,-0.2086116374,-0.0222781207,0.1690416485,-0.1958283335,-0.0635480955,0.2094288468,-0.0045031901,-0.2351339906,-0.0456194244,-0.1633234769,-0.2075599283,-0.1332043409,-0.0886218473,0.200215593,-0.0326209813,-0.0785959214,-0.3392745554,-0.4105184674,-0.1677013189,-0.2065607905,0.2757324278,0.1025791168,0.0300328061,-0.1021803543,0.5886275768,-0.0973772705,-0.4074095488,0.0635471418,0.231092602,0.4092469513,0.1268726289,-0.2562837303,-0.0187429804,-0.0799733624,-0.0105267717,0.485570401,0.271260947,-0.1843242496,-0.058483392,0.0346151963,0.1839804202,-0.277197212,0.0525438152,0.1220447794,0.1618273705,-0.244751364,0.2144116908,0.0584876686,-0.0481413193,-0.1942317635,0.1113038883,0.0692720115,-0.0840400681,0.3382104039,0.381034404,0.1406192482,-0.0843543038,0.0275523178,0.0296821091,0.2904462814,0.1392172277,0.2014310658,-0.0209841356,0.0038979233,0.4665327072,0.397429198,-0.0417923629,-0.1978858262,-0.0551535375,0.3230111301,0.2193620652,0.1007833704,0.1601347178,-0.4338372946,-0.2487107664,-0.0212122519,-0.1007903367,-0.183948189,0.0439903736,0.3703761101,0.3934574425,-0.5257890821,-0.2690373957,0.0910345912,-0.0085922023,-0.1007651687,-0.0235459115,-0.161611706,-0.025903713,0.0410804562,-0.0912399217,-0.1119716465,0.3176028132,0.0411706828,0.2570929825,0.0519143306,0.4101999402,0.3513185084,0.0475453548,0.2566618323,0.0302044488,-0.2460518628,0.19108513,0.1403933316,0.1635117084,-0.1430163383,-0.0353591591,-0.1710088998,-0.1283731908,0.3038410246,-0.0325829908,0.1178204864,0.264389962,0.0730953813,0.0128317028,0.1634646952,0.0399564914,0.0088367406,-0.0845971555,-0.1734227091,0.4944341183,-0.2157095969,0.247106865,0.1537955999,-0.1295028478,0.1028585956,0.0866306201,0.0296535306,0.1488037258,0.5596119165,-0.3162333071,-0.1543396115,-0.2396563143,0.0132858707,0.0208273027,0.6919667125,0.1588595212,0.4518476427,-0.1023001671,-0.2958510518,-0.1313503534,-0.1365274787,0.0269480515,0.3040649295,0.3247371912,-0.2221427113,0.3693711758,0.2857192457,0.3169396222,0.0334942155,0.2022929192,0.0263128541,0.0317356698,0.0865407661,-0.2278984934,0.312741816,-0.4761163294,-0.2303731441,0.0193676148,0.1770935655,-0.1575981826,-0.1239933223,0.2766904235,0.2105663866,0.0544075631,0.3564323783,0.0460316092,0.3243827522,0.0022357423,-0.0187230743,0.0145059731,-0.5402626991,-0.0021693115,0.0811913162,-0.1059499159,-0.2795075178,-0.0309944768,-0.0478407145,0.0178755745,-0.1392123997,0.1770959795,0.4349824488,-0.3496014178,0.0698190182,-0.106413424,-0.1567607522,0.2592369318,0.0271981992,0.1648727059,0.022875296,0.1580034941,0.1100541875,0.36882177,0.0301724598,-0.0409582257,0.0816957057,0.3904668689,0.3605585694,-0.1757492274,-0.5782330036,-0.0241748709,0.1788348556,-0.3058331907,0.008183158,0.3805195391,-0.0424586907,0.1322424561,-0.1377236247,0.5173623562,0.0197629519,0.1178389043,-0.4071608186,-0.3186650574]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2092","title":"How to disable making arrow tables in load_dataset ?","comments":"Only datasets like wikipedia, wiki40b, wiki_dpr and natural questions are available already processed on the HF google storage. This is used to download directly the arrow file instead of building it from the original data files.","body":"Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?","comment_length":36,"text":"How to disable making arrow tables in load_dataset ?\nIs there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?\nOnly datasets like wikipedia, wiki40b, wiki_dpr and natural questions are available already processed on the HF google storage. This is used to download directly the arrow file instead of building it from the original data files.","embeddings":[-0.1666766554,-0.0997227728,-0.0602355674,0.251595974,-0.046882987,0.1143143401,0.2708075941,0.0048356182,0.4697786868,0.1632679105,-0.0370749906,0.1923091859,0.1889043152,0.0169300884,0.2863861322,0.1164424717,-0.0252678543,0.4686415792,-0.3328589797,-0.1243106127,-0.1999843121,-0.2028210759,-0.0752780363,0.0748854652,-0.0713145509,-0.2044893652,0.1787602305,-0.1432494372,-0.3065359592,-0.3528852165,0.1729183197,0.3800322115,-0.1642081887,0.1244354025,-0.0001261445,0.094289951,0.0329206325,0.0948778763,-0.5260359049,0.006027014,-0.1746383309,0.0031732447,0.4379302263,-0.0186206643,0.2819115818,-0.4082455337,-0.1863089353,-0.3450863361,0.3139982522,0.0053225714,0.0905807093,-0.0716364905,-0.0604900382,0.2094389796,0.5697895288,0.2733406425,-0.4536009133,0.2838582098,-0.160365209,-0.0454603545,-0.233036831,-0.019829955,-0.1649195403,-0.0176153965,0.3427163661,-0.0658057034,-0.2182617188,-0.4648956954,0.3253163993,0.4471846819,0.8690113425,-0.4102916121,-0.0729228109,-0.1844525039,0.2027647644,0.0973021239,0.2856376767,0.3990594149,-0.1066478416,0.0758944452,-0.2127400041,-0.6784489155,-0.25742957,-0.0711820871,0.2603639662,-0.1153429598,0.0183343701,0.050158713,0.3017888069,0.0081159929,0.3686649203,-0.223840639,-0.3648709059,0.272334367,-0.1842785031,-0.2159967422,-0.3739869595,0.1744879186,0.4409631193,0.1009864882,0.4295786917,0.0503217205,0.2170754522,-0.071750775,-0.1252764016,0.1128390431,0.1504207999,0.0533880554,0.2999628484,0.0077718552,0.1263809949,-0.0047516157,0.1571422666,-0.2876321971,-0.0752578303,-0.1376900822,-0.0675592944,-0.1553011984,0.0428053699,0.0308059659,-0.0968296155,-0.0888143331,0.1404576004,0.1764500588,0.231339559,-0.0716202036,-0.0928623602,0.3906411231,-0.2245062888,-0.1648521125,0.065603666,-0.1498602033,-0.1348668039,0.4797789156,0.4003249407,0.0539176986,-0.1964550912,0.47218135,-0.2374716252,0.0166779142,0.3698478937,-0.344773978,0.1794475317,0.5079669952,0.1643345654,-0.1345516741,-0.0287734866,0.0166882053,-0.1465750039,0.379989922,-0.3486206234,-0.2767961323,-0.0916240662,0.0192897059,0.1560793966,0.1055180803,-0.8544896841,0.2341753542,-0.0627288297,0.192218557,0.3423011899,0.1384256929,-0.147481814,-0.1924449801,0.1250506341,0.4684499502,-0.3885135949,0.0869876742,0.0318108052,-0.1592184603,-0.1226311997,0.0867344886,-0.3584826887,0.256198436,-0.1413198113,-0.0886124223,0.4231479168,0.1055300534,-0.19324103,0.2495382875,0.1716930568,-0.0479312763,0.1314125806,-0.0283833425,0.0007526541,0.0692844614,-0.3884218335,0.1332307309,0.3127720952,-0.148754105,0.1245481968,-0.2645056248,0.1171085909,-0.1291519403,-0.1417371631,0.3576332033,0.3157827854,-0.2228238136,0.1151502952,-0.1869778186,-0.0566203855,0.2460997701,0.0988696292,0.0628359094,-0.0105710886,0.2677644491,-0.7945275903,0.1788479835,-0.3360014558,0.0576645248,-0.0180710051,-0.2592436671,0.0695411116,0.1732881069,-0.0863894597,0.2867399454,-0.0377667658,-0.5865476131,0.022451492,-0.0337138176,-0.133320421,0.0592200272,-0.4370775223,0.2424857765,-0.2334551513,0.3700748682,0.2047596872,0.1187113523,0.0710401684,-0.0198318381,0.0175126679,0.2195648104,-0.1529205143,0.1186263859,0.1463334411,0.2047602236,0.1609807163,0.3847929239,0.0475857146,-0.5845116973,0.3888181746,-0.3807119727,0.0493073575,-0.0686550662,0.0091466028,0.2706066668,-0.0396314114,0.3539585471,0.1418545991,-0.0684293434,-0.0405804627,0.1539785266,-0.0198241435,-0.3033726215,0.0530690961,0.2990089953,0.4224357605,0.3638781905,-0.0066055553,-0.1219093576,0.1077943668,-0.0360347591,0.2045954913,0.3079280853,-0.044127509,-0.3254676461,0.0753500983,-0.1600909531,0.1006897688,0.1284569651,0.1205157787,0.0344251357,-0.1513556838,-0.2663257122,0.329082489,-0.0247675776,0.4524427056,0.1972550601,0.0525163971,-0.0429168604,-0.5053237081,0.1169964001,-0.2084615827,-0.0675217137,-0.3271872103,0.0067364187,-0.23215051,-0.0429422073,0.0692347512,0.197429508,-0.5449183583,-0.0646310076,0.2237851471,-0.0641596019,0.1264277846,-0.1384565234,-0.2107720226,0.3652667701,-0.0992822796,-0.1764004529,-0.6031213403,-0.0247923918,0.1110709682,-0.0146484962,0.0877328217,0.1183421239,0.2117766887,0.0654745623,-0.0797165409,-0.4650186002,-0.1034691408,0.2821801007,-0.0597591959,-0.1261900067,0.1421791911,0.2849062085,-0.1957388818,0.1819579452,-0.1306826174,-0.2314673364,-0.1515339464,0.2393718362,-0.3711264729,0.4133032262,-0.0417639241,-0.0285002366,0.1013329104,-0.0524804294,0.1647360772,0.140779227,0.2111324966,0.1314134151,0.169273898,-0.0328297131,0.0723700151,0.0704223588,-0.1246540397,-0.2096573561,0.4650502801,-0.2878260016,-0.1776955724,0.2696314454,0.0248889253,0.0481998138,0.3943352699,-0.7647732496,0.4628793895,-0.2994994223,0.0385507531,-0.3009708524,-0.0657135546,0.177926302,0.1902365685,0.0707082599,-0.2031704187,0.1442662925,0.3033227623,-0.0519685261,-0.1278569102,0.2401959449,0.3711486161,-0.4483270049,0.6514479518,-0.2132665366,0.2048831135,0.3713408113,0.0653752908,0.4804442525,-0.3940525651,0.3682645559,-0.3533893824,0.0440669544,0.2167505771,0.2128106058,0.023932891,-0.281735003,-0.1636437029,0.0663921237,-0.0486384816,-0.1369582415,-0.0671559647,-0.5520762205,0.1195465848,0.0210163798,-0.0450575761,-0.002931589,-0.2983916998,0.0829188153,-0.0675418228,0.4848745763,-0.0068184924,-0.2792008221,-0.1240624115,-0.3302901685,0.0892267004,-0.1196119264,0.1181204095,-0.1329015642,0.2243361473,0.3356899321,0.1744421571,0.1654407233,-0.2095064223,-0.0267093536,-0.0790950283,0.0380938686,-0.3711466491,0.2080153823,0.0716089904,-0.0657842234,-0.4943979979,0.1716094464,-0.1547258645,-0.3168120384,0.0115835527,0.4513144195,-0.2096797526,-0.1479445696,-0.3307107985,-0.0254508909,-0.0559079833,-0.1177633852,-0.0423424728,-0.3654842079,-0.4115668535,-0.1679454297,-0.2127004266,-0.1215470731,0.0806567147,0.0565326065,0.2683748007,0.0463493355,-0.1071775556,0.3505336642,-0.0325152762,0.5019224286,0.6358849406,0.0561863072,-0.1729899794,0.2044605911,-0.1537905037,0.1589084566,0.1165414602,-0.1232469976,0.3661242127,-0.0137458229,-0.0890680179,-0.1629221439,-0.3557829857,-0.0765865445,-0.3086439967,-0.3322865367,-0.4312345088,0.5538908243,0.0869347006,-0.1268823445,0.545566082,0.439669162,-0.4222631752,0.3849903345,0.4214794338,0.8774513006,-0.1859513074,0.1373995095,0.0420105532,-0.1602692753,0.0516840592,-0.3475292325,-0.0553091243,-0.2762123644,-0.2480943799,-0.2208876163,-0.2604963481,0.3207469881,0.5598617196,0.0758311674,-0.0168097261,-0.3847162724,0.0434471741,-0.0894948989,0.2776198089,-0.2985426784,-0.3013461828,-0.329484731,-0.0017132628,0.080520995,-0.18589212,-0.0120102931,0.1934579462,0.1445691735,-0.208453998,0.0309502222,-0.0524717681,-0.3397980034,0.176210463,-0.2059482336,-0.3090371192,0.3525288105,0.3789273798,-0.0324096568,0.1247446015,-0.1507905126,0.0342162549,0.1641656756,0.0364575759,-0.4183193147,0.2958885133,-0.0074213212,-0.032589227,-0.2283522934,0.2319410294,-0.0671027899,-0.2458571345,0.0267583244,-0.202127561,-0.0008233984,-0.0305230059,0.295430541,-0.1471578032,0.1176411808,0.0350883529,-0.0040230313,-0.0434718616,0.3414011896,-0.1429952979,0.1262927502,-0.4455615878,-0.1689261198,0.2117101699,0.287364006,0.0582482219,0.4512853026,0.114897795,-0.3726489544,-0.346250385,0.238901794,0.2295860052,-0.6402404904,-0.1173894182,0.1829472631,0.0842127427,-0.1442791075,0.085901998,-0.1411006302,0.1200365946,0.0319073983,-0.059322428,-0.1330087334,0.2434148937,-0.1136931852,0.2819097638,0.1148826033,-0.6818988323,0.2275173217,0.2323222309,-0.111889191,-0.0881017298,0.2363083363,-0.2088969797,-0.1703198701,0.3237912357,-0.0378168635,-0.1207607687,-0.077024281,-0.1613738686,-0.0819709301,-0.0679908171,-0.0786916539,0.2069288045,-0.0629197136,-0.0170486402,-0.3042809069,-0.4994447231,-0.2313915193,-0.17665115,0.3165335655,0.0037678652,0.0465870798,-0.0950726867,0.5397336483,-0.0374463126,-0.5059788823,0.0949018523,0.0681319907,0.3120401204,0.0345160924,-0.1459198296,0.1798985004,-0.0784862041,-0.136937812,0.2776891589,0.1636315137,-0.1743123382,-0.0390989482,-0.0886482894,0.1683629304,-0.2907938063,-0.0498126447,0.325946629,0.0578666814,-0.4079630077,0.2748281956,-0.0258511845,-0.0484588407,-0.1697887629,0.1993431598,-0.1215256453,-0.1191868857,0.298969537,0.2730692625,0.0533032902,-0.2868098915,0.053457506,-0.2210312933,0.0159477443,0.0093288058,0.3286040425,0.0121222008,-0.0208329707,0.1727491617,0.3737918735,-0.0068508973,-0.1617454737,0.1635169387,0.2933345139,0.2526713014,0.0588131547,0.05663817,-0.2255409956,-0.293430835,0.2154268622,-0.322465688,-0.1408231705,0.1355365962,0.2070710212,0.6086524129,-0.2856192589,-0.2322924584,-0.0215821303,0.0368518755,-0.0152966678,-0.158585459,-0.1238402128,0.0329355448,0.0500573926,-0.0123464735,-0.0004823177,0.2223769128,0.1357267201,0.1743844301,-0.0047154226,0.7997297645,0.2598021328,0.0542510934,0.126211375,-0.0841554701,-0.268335402,0.2477657199,0.0414632298,0.255484134,-0.2757996321,0.112037912,-0.1659208983,-0.2564076781,0.3634276092,-0.0557472967,0.1121235266,0.2614706159,0.2424995005,-0.0933834612,0.3008231223,-0.0379583463,0.0355848335,-0.2434578985,0.1682687253,0.4942243695,0.0066179489,-0.087939404,0.2070055604,-0.2129439265,0.2865079939,-0.0173454154,0.0251475759,0.1361804903,0.5175824165,-0.1937057376,-0.2439182997,-0.2746246457,-0.0198901501,-0.0908430964,0.7207257152,0.0993879586,0.3605440855,-0.0419986434,-0.2870460153,-0.1729601771,-0.1709921211,-0.2053613216,0.24526833,0.342568934,0.0046527954,0.1894655228,0.3650675118,0.2575707138,-0.1026759446,0.127069965,0.1555002183,-0.0229718424,-0.014990828,-0.0577145815,0.3434262276,-0.6017377973,-0.3483431637,0.0381361693,0.1895769238,-0.1263940185,-0.1201359853,0.1993088275,0.2455294132,0.3667787611,0.4701032043,0.0451769121,0.2308686674,-0.0662701055,-0.0260925274,0.1762669384,-0.4044029415,-0.0097008469,0.2698429525,0.0217320826,-0.2480506897,-0.0770383999,0.1335393339,0.0351827778,0.045413889,0.0607021675,0.4718453884,-0.3078616858,0.071121484,0.0982076749,-0.0478930399,0.0559051633,-0.1070276499,0.1613435745,0.0645470023,0.0472759753,-0.0593670979,0.1997018009,-0.0551064089,-0.0160728917,0.280706048,0.4276411235,0.2663159668,-0.1641878784,-0.4257063866,0.0837852582,0.3360573053,-0.300160557,-0.0754560381,0.2571478486,-0.1475955397,0.2642498314,-0.1879017502,0.5215159059,-0.0773454532,0.098689191,-0.5550127625,-0.377808243]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2092","title":"How to disable making arrow tables in load_dataset ?","comments":"@lhoestq How can we make sure that the data we upload on HuggingFace hub is available in form of preprocessed arrow files ?","body":"Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?","comment_length":23,"text":"How to disable making arrow tables in load_dataset ?\nIs there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?\n@lhoestq How can we make sure that the data we upload on HuggingFace hub is available in form of preprocessed arrow files ?","embeddings":[-0.1100975871,-0.4545698762,-0.1245342419,0.3275228739,0.138504684,0.0666502416,0.1080085412,0.0051873387,0.2907291651,0.2163161486,-0.0762575269,0.3311557472,0.0758255497,0.2674562335,0.3114099205,0.0753525868,-0.0126066515,0.3374112546,-0.4707742035,-0.0729587302,-0.0226511937,-0.1312700659,-0.0280798543,0.1429725587,-0.3697292507,-0.1202766299,0.1659023911,0.1318922639,-0.4114736021,-0.3004578054,0.2799259424,0.3918724656,-0.2110333443,0.2840105295,-0.0001185283,0.0347914472,0.0148414318,0.0469427444,-0.42345348,-0.0630001873,-0.1052796096,-0.0900127366,0.552696228,0.1363579482,0.1049978584,-0.4011049569,-0.2245007008,-0.1469555646,0.4965695739,-0.1221488342,0.1708035469,0.400201261,-0.0367949791,0.0012566323,0.2102968395,0.3903617859,-0.490552336,0.2777503729,0.1035267413,0.042191416,-0.2085119486,-0.0491254888,-0.0048570805,-0.1373191476,0.3756073117,-0.0360606983,-0.1985394508,-0.4639984667,0.0708648711,0.3484270573,0.5639098287,-0.3297149539,-0.1446353048,-0.3594789207,0.2402592599,-0.0216076244,0.3293654323,0.225537464,-0.1578767896,0.0673475042,-0.2663391531,-0.5470802188,-0.3333836794,-0.218575567,0.5047805309,-0.076915063,-0.1086540893,-0.0129904263,0.2491814792,-0.0777414218,0.2917518914,-0.0805463344,-0.2887803316,0.2161269933,-0.1914640665,-0.3093226552,-0.281237334,0.4735992253,0.3961885571,0.400200367,0.5239634514,0.0959406942,0.072010994,-0.1174758673,-0.0317182876,0.1587304324,0.193554014,-0.1155328155,0.2119951695,0.1582829058,0.1720876396,0.009821374,0.0601810478,-0.3296580017,-0.1955471039,-0.0224679429,0.1544831693,-0.1553844362,0.0939445496,-0.1374528259,0.2613368034,0.0638690218,0.3560952544,0.0569743812,0.180691883,-0.0295460783,-0.1791046262,0.379006207,-0.248083204,-0.2792924047,-0.1047427654,-0.1694387048,-0.1806078106,0.4586046338,0.4090617895,-0.1122873053,-0.0887114778,0.4487025738,0.0572253019,-0.0153427804,0.0484907404,-0.1698323339,0.1490182281,0.4378680885,-0.0290969741,-0.0705789328,-0.0994414911,0.1250169873,-0.1465258449,0.2650176883,-0.3646680713,-0.4418751597,0.0444759578,0.0543172583,-0.0444458537,0.0182588138,-0.7932122946,0.1292459667,-0.2953799963,0.3496223092,0.2732408345,0.1834332496,-0.0424465351,-0.2014659047,0.1859554499,0.352290839,-0.1842111498,0.0959440693,0.1387863308,-0.3303410709,-0.1780183613,0.1179890856,-0.3667709529,0.242792204,-0.1544751376,0.0416446663,0.4212193191,-0.0907354727,-0.0077271443,-0.0763670057,-0.0343230106,0.1033494473,0.1471956819,-0.1076357365,-0.0380745903,-0.1216002479,-0.2622415721,0.1754321456,0.1581322998,-0.1894269288,0.089052707,-0.2200370729,-0.0470202528,0.0155239971,-0.3854585588,0.1361887455,0.0694885924,-0.3777515888,0.0612149648,-0.1447183937,-0.106374152,0.07156609,-0.0054898555,0.0905893818,-0.0275947712,0.2464648336,-0.8245880008,0.261398077,-0.1207316667,-0.0144083798,-0.1286077648,-0.1991339326,0.0065702735,0.196771428,-0.1613360196,0.4198437035,0.0400856398,-0.6630686522,0.123702541,-0.2078924626,-0.4388538003,0.1259002537,-0.3588649333,0.4169155061,-0.5347799063,0.1409382224,0.2798046768,0.002394771,0.10170836,0.0857685208,-0.005962065,0.1202664897,-0.0833942071,0.208268553,0.2024602443,0.4575892687,0.3043973446,0.4158650339,0.1502268165,-0.4250659347,0.2253432125,-0.4385340214,-0.0400626548,-0.0068691913,-0.0001241759,0.2602646947,-0.4774878025,0.2487118244,0.071763739,-0.0799846649,-0.0640240014,0.1609707922,-0.2656231523,-0.2145772427,0.1559644043,0.1631298959,0.3208202124,0.2711210847,0.0731402189,-0.0346236266,0.0183139648,-0.105354175,-0.0562522002,0.1356912255,0.0144982468,-0.0940969512,0.1921983659,-0.1028650925,0.1302126795,0.1553015262,0.1930710524,0.0749020875,-0.1782563627,-0.3413229883,0.2432800531,0.1592385769,0.340529263,-0.0173164885,-0.0734461918,-0.0706315041,-0.5441478491,0.1775170416,-0.3389026225,-0.1754332185,-0.3400018513,-0.0805465057,-0.2896670997,0.0255501829,0.0797001049,-0.2292761207,-0.5078178048,0.0496624038,0.322514981,-0.0190664548,-0.0947642475,-0.0793125555,-0.058397565,0.5088121295,-0.227946803,0.0185197163,-0.5136077404,0.1025357172,0.1898605227,0.0411525331,0.1310687214,0.115196839,0.3711613715,-0.0504336767,0.1480102986,-0.2743229568,0.0571415238,0.2185662389,-0.1748706698,-0.0808990002,0.0650817528,0.2000825256,-0.0712435842,0.1652932465,0.0119652422,-0.2896600962,-0.1578966379,0.098006919,-0.2179299444,0.2611680329,-0.1466182917,0.078511335,0.084781535,-0.2372642457,0.3560937643,0.0727703571,0.1659401357,0.1821744889,0.1853740215,0.0225035213,-0.2359180599,0.133266151,-0.0792166218,-0.4334353805,0.4475567937,-0.233407557,-0.2612316906,0.3169552684,0.129692778,0.2379167527,0.1163240746,-0.5254436135,0.1258656383,-0.1705875099,0.0319621451,-0.1600883752,0.0561917908,0.3184776008,0.0864020437,-0.0027333333,-0.1439156085,0.1006568968,0.3076089025,-0.0458948351,-0.0421247296,0.0426631048,0.0149013232,-0.4100801647,0.4363572598,-0.0640627742,0.2218597829,0.3528681397,0.1127993464,0.396458149,-0.2960090041,0.2807895243,-0.5286573172,-0.0807221904,0.2156338394,0.1306065321,0.4294693768,0.1301455498,-0.2506133318,0.0370223895,0.0122810826,-0.2276196182,0.084249191,-0.4871127903,-0.1096766442,-0.0329922549,-0.0747422352,-0.0610882267,-0.2015198022,0.1468239129,-0.085472852,0.4783972502,-0.1560278535,-0.446125567,-0.1290576905,-0.3959272206,0.1136702225,-0.1200310364,-0.1101222709,-0.0130362716,-0.0511963814,0.3627043664,0.2612914443,0.3381401002,-0.0175811686,-0.2269538194,-0.1811385304,-0.1522102952,-0.5588306785,0.2730967999,-0.0250024665,0.0530471578,-0.432138741,0.2322058082,-0.1590506881,-0.4867697358,-0.0347318351,0.342204541,-0.1691904366,-0.156561017,-0.2592453361,0.0075490861,0.0096302694,-0.1962583512,0.0238092989,-0.3379902542,-0.3770409822,-0.2375996411,-0.1367364973,-0.0702639148,0.0693276897,0.1305486113,0.3056545258,-0.0491025001,-0.0501070544,0.6013351679,0.1318714023,0.3350690603,0.7355259061,0.0471780822,-0.3311446011,-0.0033870018,-0.0484329797,0.3059706092,0.3620956838,-0.0146431029,0.2749542892,-0.1106643379,-0.034234982,-0.4976652563,-0.2990252972,0.0701588988,-0.2617882192,-0.2227217257,-0.2190456837,0.5680629611,0.0189966653,-0.1520609558,0.5435661674,0.7026253939,-0.2865083516,0.3620840907,0.250754863,1.0189739466,-0.1258427948,0.1765022427,0.1644608229,-0.159613058,0.221593529,-0.1565438509,-0.0728096738,-0.110818781,-0.181283325,-0.1127720401,-0.202447772,0.234939903,0.3957484961,0.0797575489,0.0002264623,-0.3505664766,0.0246779583,-0.277859509,0.1437403262,-0.3000412881,-0.5094458461,-0.521559,0.09045434,0.1492137015,-0.0668033659,-0.0609569214,0.171968922,0.0723226294,-0.1597865671,0.0820637271,-0.1538577378,-0.2308883518,-0.0509246103,0.1718118042,-0.2952211797,0.1772787571,0.307436794,0.124035649,-0.066750437,-0.193637535,0.0937847272,0.019895779,0.0419722944,-0.4512567818,0.1863234341,0.0436513089,-0.1145981774,-0.1931340992,-0.0149624264,0.0689750537,-0.1675017774,0.0595378764,-0.1282438785,-0.0260729156,-0.1771702021,0.250400573,-0.1011194661,0.2999172211,0.0453323498,0.0786273926,0.1116082445,0.1935479492,-0.005318867,0.3507255018,-0.2922370732,-0.100529626,0.170935601,0.1227687076,-0.311318934,0.363663733,0.0727191046,-0.3071179986,-0.3510376513,0.3143333793,0.3919265866,-0.842854321,-0.0073088408,0.2356801182,-0.0706527829,-0.1913158298,0.1781543046,0.1387346536,-0.0710748211,0.06333372,0.0630550459,-0.3358898461,0.3309183717,0.115262717,0.225965187,-0.0569997504,-0.7230476737,0.1906858087,0.2346635163,-0.2108670175,-0.011026429,0.1433056146,-0.1370856613,-0.1973810196,0.2182661593,0.0825318024,-0.2296374142,-0.0310004335,-0.1259457171,-0.2582264245,-0.1548725665,-0.1498700678,0.163393721,-0.0731901005,-0.1037511379,-0.3107844889,-0.3471992016,-0.0970347077,-0.1701603532,0.2492843121,0.0597427003,0.12586689,-0.268315345,0.6817316413,0.1024312973,-0.4241314828,0.1237007454,0.2305525541,0.4535217583,0.1163683832,-0.3404822052,-0.1207611635,-0.1914929301,0.2119704336,0.3374477327,0.3766676486,-0.1355810612,-0.1138689741,0.0466498435,0.1565054953,-0.3454875052,0.0947084278,0.3999513686,0.054244902,-0.2553439736,0.2509409487,0.0872724727,0.0465143174,-0.0608861968,0.25091061,0.0469506979,-0.1093305126,0.2647152245,0.2352207005,0.2208656371,-0.102205731,0.0490365885,-0.1480417848,0.1447732896,0.030470008,0.1158181652,-0.2176097929,-0.085872747,0.3323891759,0.4036595225,-0.0065657101,-0.0435485058,-0.026687149,0.4827599525,0.1713672131,0.1913904548,0.1979712844,-0.2921998203,-0.3643681407,-0.0274741109,-0.1032728776,-0.1535485983,0.1526007056,0.4873872101,0.3485182226,-0.4362683892,-0.2760209739,-0.0571555197,-0.0304017942,-0.0383339711,-0.1758431345,-0.1453133076,-0.0165432282,-0.0856231898,-0.0803146064,-0.0160517786,0.2937105,0.0897771418,0.2287437171,0.063960731,0.3939080238,0.2061113566,0.1510290354,0.3409267962,0.0631595477,0.005954802,0.3230861127,0.0947478712,0.2646947503,-0.1124231815,-0.0500251278,-0.1727470309,-0.0082481056,0.3032369614,-0.0925093666,0.0496074408,0.3993669748,0.1748822927,0.1327618062,0.1311957985,0.0755117759,-0.037275739,-0.1949138492,-0.08900024,0.4191934764,-0.0188638009,0.1095526963,0.0970699191,-0.2161709219,-0.0350928567,-0.0308076162,-0.0494230315,0.0970060378,0.4881445467,-0.2978170514,-0.2739186585,-0.2831573784,0.0307260957,0.0582158267,0.8948065639,0.2400965095,0.2643409073,0.0113205807,-0.35473755,-0.2399243563,-0.0330052078,-0.0751702562,0.2713891864,0.1888168007,-0.1306457669,0.2097887844,0.4919033945,0.2400507182,-0.1573126465,0.1594818234,-0.002805273,-0.0664312541,0.0140243778,0.0303643979,0.1939195693,-0.4093895555,-0.1508574933,0.1641547233,0.2551954985,-0.1288655698,0.0609907247,0.2979824245,0.0469960719,0.2677919567,0.3522815704,0.1014788151,0.1359042525,0.0656998083,-0.0964440778,-0.0508313179,-0.4565628767,0.001053848,0.2205652893,-0.0074120653,-0.3085287213,-0.1222024858,0.0345908515,0.0705062002,-0.0084912097,0.0030951973,0.2791023254,-0.283687681,0.0735960454,0.1059361324,-0.055023659,0.1909136176,0.0099867368,-0.0054557445,0.0161598567,0.049666211,0.0464628339,0.4059775472,0.0641093999,-0.0398205034,0.1455324888,0.3363781571,0.3294084966,-0.1950829178,-0.5431334972,0.0223273784,0.0680249557,-0.320487529,-0.0192941949,0.5207051635,-0.1929074526,0.0761844218,-0.1330620199,0.6729959846,0.0351433605,0.2069901824,-0.279656142,-0.2860947847]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2092","title":"How to disable making arrow tables in load_dataset ?","comments":"We're still working on this :) This will be available soon\r\nUsers will be able to put their processed arrow files on the Hub","body":"Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?","comment_length":24,"text":"How to disable making arrow tables in load_dataset ?\nIs there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?\nWe're still working on this :) This will be available soon\r\nUsers will be able to put their processed arrow files on the Hub","embeddings":[-0.3103685677,-0.2093830109,-0.1464321464,0.2652353942,0.0345696248,0.0666943863,0.2737096548,0.1132889166,0.3428294659,0.2253294438,0.0225788374,0.5258630514,0.0204506256,0.1348215342,0.2148839384,0.2104867846,-0.0567545183,0.3835603595,-0.370223552,-0.0217931662,-0.0132307783,-0.242534712,-0.0369567052,0.1654260159,-0.352792412,-0.1516540945,0.2050920129,-0.0799770728,-0.3733310401,-0.3553424776,0.2808741033,0.3713862002,-0.2569585443,0.2140728086,-0.0001159839,0.0817523003,0.1164755374,0.1294047385,-0.3511275947,0.0271335766,-0.2709644437,-0.0506317429,0.5330007672,0.048189234,0.1740213484,-0.3140257001,-0.326105535,-0.2373576313,0.3075525463,-0.1084174141,0.1901614219,0.1338144392,-0.1258650869,0.0214221459,0.3372371793,0.2644141614,-0.5166210532,0.1766559333,0.0029492094,0.0186856538,-0.2683089375,-0.0692397505,-0.0806634501,-0.1670561135,0.2791842818,-0.1535366327,-0.1141778752,-0.4275369346,0.0803016573,0.2599158585,0.7489739656,-0.3426682949,-0.1013403311,-0.2271207571,0.2399844527,0.046319887,0.2915193141,0.2484015524,-0.1438123435,0.0492912047,-0.3220917583,-0.580715239,-0.3312959075,-0.2287511379,0.3389687538,-0.101069957,-0.0031777222,-0.0043580965,0.2739900947,-0.032246463,0.6454591155,-0.0957823172,-0.4226688147,0.1684318185,-0.2579934299,-0.345099777,-0.2567856312,0.3940750957,0.2697523832,0.3345091939,0.5258702636,0.0830111206,0.1021843106,-0.0601601824,-0.0829810873,0.1526259184,0.3308385909,-0.0539739765,0.2798469663,-0.029062232,0.0941072702,0.030074412,0.0683990493,-0.457054168,-0.1736430973,0.0321198478,-0.0016698214,-0.1858097911,0.0548646078,-0.1981121302,0.1823126525,0.0743919834,0.3349526525,0.0447428524,0.2945014238,0.0740870833,-0.1831943095,0.2993355393,-0.2095319033,-0.1715037823,-0.0472710803,-0.2076269537,-0.2439629138,0.3211621642,0.3855163157,0.0219374523,-0.1519086957,0.4505899847,-0.063513115,0.0171689428,0.2573388219,-0.1818991452,0.1704493314,0.5423360467,0.0090393126,-0.1620635241,-0.1104332358,0.0282398332,-0.0701754913,0.4509628415,-0.417442441,-0.3716728389,0.0600933656,0.1057705209,-0.0788251087,-0.0629551485,-0.771576345,0.0448148362,-0.1329560578,0.3094379008,0.2648726702,0.1242686138,-0.0766548067,-0.2044762373,0.0895226151,0.2795028687,-0.3330899775,0.1844024807,-0.0560634099,-0.2037720829,-0.17547144,-0.0012652776,-0.4258303642,0.2005177587,-0.1304040402,0.0118652526,0.5192417502,-0.0667947084,-0.0070453533,0.0726575479,0.0436769091,-0.0900366902,0.1525606513,-0.0635976642,-0.0286896247,-0.1447896361,-0.2948603332,0.0517011844,0.237039566,-0.1687286794,0.1082211956,-0.2682799995,0.094784379,-0.091040425,-0.2780421078,0.121010989,0.230222553,-0.2342467159,0.1505142897,-0.1324554533,-0.1091565043,0.1529869139,0.0997963771,-0.096534647,-0.002666326,0.1729294509,-0.64591223,0.2446274161,-0.1348222941,-0.0035569465,-0.0143185882,-0.2581925988,0.0553368852,0.1292538792,-0.1676240861,0.3968893588,0.0644265562,-0.759450078,-0.0210416224,-0.2730541825,-0.3172904551,0.0633029118,-0.3434702456,0.2090676874,-0.3996973634,0.1945395172,0.2741888165,0.0175567735,0.0969060883,0.0465892106,-0.0056083496,0.1921326816,-0.1821298599,0.2162335664,0.2495952696,0.3024962544,0.2750135064,0.3957771957,0.0631694719,-0.4625014663,0.4302769601,-0.4933479428,-0.0508707277,-0.0299595408,0.0575819127,0.2281270474,-0.438893646,0.2713938653,0.0143624665,-0.064069286,-0.0025123826,0.1378728896,-0.1525087059,-0.1967370063,0.069967024,0.2555822134,0.2578304112,0.2911311388,0.1528061181,-0.0244936161,0.0216730591,0.0796902329,-0.0018178378,0.1429063529,0.0162566788,-0.1354296207,0.1111126095,0.0136045255,0.2229940146,0.1735732108,0.0848785788,-0.0212465655,-0.1328105032,-0.3586786687,0.3063268363,0.1005614549,0.4162053466,0.0529248305,-0.0220041536,-0.1683461368,-0.5321386456,0.102372326,-0.200342223,-0.1142188236,-0.3129298091,-0.105699636,-0.3163559735,-0.0081692068,0.0785483122,-0.1067342833,-0.4908870757,-0.0413646847,0.344756484,-0.0384743065,-0.0207313877,-0.0190737359,-0.2208305597,0.3961042762,-0.1212590188,0.0660755485,-0.6435341835,-0.0223674718,0.2465204448,0.1031156704,0.0695400611,0.1105958,0.2626501918,0.0976542607,0.1982641667,-0.2104328275,0.0442119911,0.1528565884,-0.0880778879,-0.1403047889,0.1137235165,0.1261760592,-0.0997866467,0.1234928444,-0.1222463474,-0.2544909716,-0.120075278,0.168158412,-0.2586185038,0.2979022861,-0.1329230666,-0.0157442782,0.1073456928,-0.2231793404,0.1770680249,0.1241577491,0.1861981899,0.1440699846,0.2328467518,-0.0193372183,-0.0556620844,-0.0799492672,-0.0262091141,-0.3828001022,0.4329149127,-0.3228119612,-0.1916915327,0.34599787,0.2078003883,0.3050369024,0.1848863661,-0.4744650722,0.3793217838,-0.2185674608,0.0129554383,-0.2385337055,-0.0208479501,0.2446891665,0.1760331392,-0.0418691449,-0.1761095077,0.2295515239,0.2544288337,0.0437615439,-0.0548274145,0.0251075812,0.0771376565,-0.3237439692,0.356377393,-0.1245106086,0.1490378678,0.2294834852,0.0269236248,0.3668178618,-0.306694299,0.3087341189,-0.4477266669,-0.0415461659,0.1445001811,0.0641132072,0.2390729785,-0.1201753169,-0.1601918936,0.0096521275,0.1319193691,-0.161261633,-0.0346196219,-0.7436398864,-0.1173464581,-0.0558333434,-0.0480463319,-0.0292785931,-0.2988164425,0.1522434652,-0.1356478631,0.4277974367,-0.2315807492,-0.3994406164,-0.0872971117,-0.2561006844,0.0836074874,-0.1827530414,0.0100355977,-0.0118827568,0.1019517854,0.2888185382,0.2473774552,0.316110909,-0.2173167169,-0.090402104,-0.0167898573,0.0268067084,-0.5356786847,0.2516008914,0.100354135,0.1265264452,-0.5037956834,0.1540519595,0.0333894938,-0.4428293705,-0.0943353996,0.3720583618,-0.2775084972,-0.233915478,-0.2597722709,0.1157044321,-0.0050635817,-0.1438821703,0.0610014796,-0.3092712462,-0.5351692438,-0.3712533414,-0.2194992751,-0.2014858425,0.0451499484,0.1697090417,0.2755418718,-0.0576478988,-0.0901062861,0.464502871,0.0100135272,0.3138617277,0.5680856109,0.1147116423,-0.2208566219,0.1053473353,-0.2555282414,0.3461354971,0.183343038,-0.1466768086,0.2589337826,-0.1299972832,-0.1473017037,-0.2657667696,-0.3406081796,-0.0761996657,-0.2886853516,-0.1925839931,-0.4018352628,0.4588519633,0.0464713313,-0.1920524389,0.6431564093,0.3190080822,-0.3394657671,0.4343406856,0.3548600674,0.8494632244,-0.1127496436,0.0916340947,0.0217832215,-0.079370141,0.0984683484,-0.2582957447,-0.0308728088,-0.2190079093,-0.2396124154,-0.1552689224,-0.2967155874,0.3031743467,0.614132762,0.1015933678,-0.0674711987,-0.4352980256,-0.0642315596,-0.2260585278,0.1462736875,-0.2950591445,-0.3401350975,-0.3785334229,0.1125151813,0.1348388195,-0.2577759624,0.0197710432,0.1866468042,0.0442516506,-0.1310450137,0.1150083691,-0.0062731644,-0.1419145614,0.1525866836,-0.118970409,-0.2175102383,0.2758843303,0.3597983122,0.0238153208,0.052527979,-0.1432892233,0.202545464,0.1643689573,0.1549889296,-0.3488694429,0.1979309022,0.0289441943,-0.1203055009,-0.2554783523,0.064689137,0.0931319967,-0.2896289527,0.2149234861,-0.0353779197,0.0891196951,-0.1722824574,0.2817962468,-0.0768026114,0.221368283,0.0420479849,0.1181833819,0.1777941734,0.162217617,0.1248469949,0.1956323385,-0.3417986929,0.0051922211,0.1840080768,0.2707306147,-0.2205857038,0.304852277,-0.0095326211,-0.3066757917,-0.4091405571,0.4217349291,0.3146269321,-0.814915061,-0.0572425984,0.2092867047,-0.0014706096,-0.0540957488,0.0822001547,-0.0592984296,0.122071065,-0.0522361621,-0.0146663236,-0.2908222675,0.2115477026,0.1122519821,0.3109154105,-0.0897130072,-0.7034543753,0.1804858148,0.1536323577,-0.2199938148,-0.0599917099,0.2268671989,-0.102559559,-0.3732074499,0.2789998949,-0.0280101765,-0.2221342325,-0.0012285648,-0.0953467041,-0.1374152005,-0.1694260091,-0.141427055,0.1486065239,-0.0452480353,-0.0276263282,-0.3236452639,-0.3971851468,-0.2167584151,-0.0718020424,0.2295221388,0.0195029266,0.0725385547,-0.1225272417,0.6832389832,-0.0225254688,-0.3817603886,0.1587914824,0.1356395185,0.3880698085,0.1000987589,-0.1560023427,0.0220113695,-0.1547561288,0.1818911433,0.2245164663,0.0930377916,-0.0119636161,-0.09836898,0.0272612683,0.1843195558,-0.315669328,0.027468551,0.397152096,0.097307682,-0.3781383038,0.258587122,0.1290524006,0.0012020803,-0.1575753391,0.222888723,0.0566256493,-0.1367578059,0.1579708606,0.2353886962,0.0697353631,-0.2357997298,0.0800212622,-0.1176943108,-0.0643169656,0.0803122967,0.2674285173,-0.1702557504,-0.1227638274,0.2504902482,0.2941710651,-0.0121871326,0.1303354502,0.1312681735,0.4220908582,0.4682114124,0.0644309521,0.0893926397,-0.1402354091,-0.3395057917,-0.0389651209,-0.1989297122,-0.2140468657,0.247110188,0.2983278334,0.4900263846,-0.3115791678,-0.195256725,-0.0586151853,0.0022104206,0.0599646121,-0.2511165142,-0.1193673983,0.0283702128,-0.1294394881,-0.0732685179,0.1970208585,0.2804987133,0.1395811439,0.2168690264,0.1357159764,0.6554526687,0.2773421109,0.1334074587,0.2608630061,-0.0865958929,-0.1507759094,0.3010555506,0.0982848108,0.2140385956,-0.2161132097,0.0128244553,-0.1068076268,-0.0141812544,0.3038367927,-0.1228850335,-0.0021904791,0.3791877627,0.1692741364,0.1502942443,0.1970709264,0.0862880498,-0.0042296988,-0.2414921224,0.0157359876,0.4444578886,0.0776924714,0.1533194035,0.331096828,-0.3567512631,0.1181738377,0.0079829916,0.064665705,0.0585175082,0.6564621925,-0.3163788915,-0.2799340785,-0.3861468136,0.03665727,-0.0452194102,0.8656638265,-0.0083552022,0.2765959501,0.0494278818,-0.2842508554,-0.229454875,-0.0662249699,-0.0475872681,0.3431770504,0.2284072042,-0.1373748481,0.2349410951,0.4011735916,0.2569224834,-0.1622017473,0.2006437182,0.0988554731,-0.1066654623,0.0774331987,0.0570642799,0.3371536136,-0.5004722476,-0.2039076835,0.1073568687,0.2701493502,-0.0749056786,-0.0041895304,0.1313371062,0.3006615937,0.4070839286,0.4192461371,0.0541680306,0.2003678977,0.0273272507,0.0839662254,-0.0259346012,-0.4960831404,-0.0011241753,0.2126317918,0.0638607442,-0.3875549138,-0.0927988738,0.1831552833,0.0277684852,-0.1626914889,-0.0997522622,0.3595461547,-0.2504120767,0.0240564421,0.1289916337,-0.0077830004,0.0447151028,-0.0787102953,0.0542051718,0.0227095746,0.1066959873,0.0210290514,0.2980059087,0.0833556056,-0.1187910363,0.2398762256,0.3707804978,0.2511375844,-0.1286710948,-0.3679949343,-0.0019752097,0.2696501911,-0.2883358598,0.0905909911,0.4025953412,-0.0758805797,0.1464559287,-0.1101843566,0.5680816174,-0.0156494845,0.1962841898,-0.4282812178,-0.3301557302]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2089","title":"Add documentaton for dataset README.md files","comments":"Hi ! We are using the [datasets-tagging app](https:\/\/github.com\/huggingface\/datasets-tagging) to select the tags to add.\r\n\r\nWe are also adding the full list of tags in #2107 \r\nThis covers multilinguality, language_creators, licenses, size_categories and task_categories.\r\n\r\nIn general if you want to add a tag that doesn't exist (for example for a custom license) you must make it start with `other-` and then a custom tag name.\r\n\r\nedit (@theo-m) if you ever find yourself resorting to adding an `other-*` tag, please do ping us somewhere so we can think about adding it to the \"official\" list :)","body":"Hi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip","comment_length":94,"text":"Add documentaton for dataset README.md files\nHi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip\nHi ! We are using the [datasets-tagging app](https:\/\/github.com\/huggingface\/datasets-tagging) to select the tags to add.\r\n\r\nWe are also adding the full list of tags in #2107 \r\nThis covers multilinguality, language_creators, licenses, size_categories and task_categories.\r\n\r\nIn general if you want to add a tag that doesn't exist (for example for a custom license) you must make it start with `other-` and then a custom tag name.\r\n\r\nedit (@theo-m) if you ever find yourself resorting to adding an `other-*` tag, please do ping us somewhere so we can think about adding it to the \"official\" list :)","embeddings":[-0.2009669244,0.252024591,-0.1304934472,0.0798938423,0.1809331179,0.3571422994,0.3119640648,0.1849832982,0.1644606739,-0.016346449,-0.0213761125,0.3338293135,-0.0963025689,0.1983852983,0.1711067706,0.0399823152,0.1146339625,-0.0630787686,0.3094624579,-0.1812989861,-0.2011339366,-0.0329795852,-0.057861574,-0.2001816034,-0.3019789159,-0.0922157615,-0.2026932538,0.0341942944,-0.3977000415,-0.607642889,0.0248835068,0.3334518075,0.1742963344,0.0735345557,-0.000107235,-0.1596610099,0.1788149178,-0.239566654,-0.0191534813,-0.1502264142,-0.2650319934,-0.4221037328,-0.1440393329,-0.0700992867,0.0782822296,-0.2990449965,-0.0470196195,-0.1426072866,0.0189226512,0.4477118552,0.2050608993,-0.2106184363,0.3308573961,-0.2609883249,0.4178492427,0.2113643438,-0.1037406772,0.3762668073,0.452617228,0.0936253443,0.0395131744,0.1774764657,-0.1239313483,0.179583773,0.3749469519,-0.045122914,0.3577607572,-0.5386645198,0.2289720774,0.3996181786,1.153159976,-0.2931374013,-0.310759902,-0.4575535357,0.1056198329,-0.1025070623,0.1988972574,0.3195011318,0.0421364456,0.1530991048,0.1238834932,-0.34485659,-0.2380984873,0.1258865893,0.1208824664,0.4693534076,-0.0495393053,-0.0461795665,0.2738636136,-0.0765385404,0.0914462432,-0.00860688,-0.2317645997,0.1363527626,0.1594582349,-0.0646353066,-0.2846120894,-0.105354555,0.2494053841,-0.0171731189,-0.0379194431,0.0125727346,-0.1057558358,0.3409814835,0.1243555397,-0.1768715531,0.165177092,0.2110916972,0.1014955044,-0.2915710509,0.3328114152,-0.1448082477,-0.159210965,-0.1529749483,0.0148062557,-0.0020573905,-0.1087305993,-0.2182008326,0.0389909446,-0.02366082,0.0868689343,-0.1800802499,0.0146845635,0.162394017,0.1409135461,-0.2257521302,-0.2854691744,0.1749260426,0.0696320757,-0.1272712797,-0.0241921861,0.2270143479,-0.1285205483,0.1915755123,0.301363349,-0.0326029696,0.2947966158,0.2256808728,-0.1369580477,-0.1198657006,0.2090932876,-0.0301327631,0.2764083445,0.3786623478,-0.0219721943,0.0751835331,-0.0591326617,-0.4303908944,-0.2476119399,0.1332641542,-0.4022579193,-0.2395637184,-0.4142950773,0.2273344398,-0.1181674078,-0.0451684557,-0.2261714041,0.2805584669,0.0936193168,-0.1459254622,0.2434444577,0.2106729597,-0.3568451703,-0.2805345356,0.0597083196,0.436062783,-0.4021865726,-0.0059407167,-0.0924031287,-0.0220785905,-0.1453588158,-0.1148246452,-0.2623159289,0.4330624044,-0.3410653472,0.2678823769,0.4367043674,-0.4631901383,-0.2488020062,0.1513609141,0.0247621946,-0.2738645375,0.1768344641,0.0510142669,0.5288446546,-0.1293236464,-0.2133076787,-0.0302911382,-0.1984529942,-0.1061873287,-0.1198968068,0.0688978583,0.0974699929,0.092744723,0.093556799,-0.0396653973,0.4201057553,0.1274108291,-0.0038211634,-0.07197579,0.1842219979,0.3900684714,0.2178639024,0.2241284251,-0.03491541,-0.0805127546,-0.4012998343,0.1436464041,0.1019424722,0.2396580577,0.0087045608,-0.2792170942,-0.3434494734,-0.1810734421,-0.3264996111,-0.3358734548,0.1371124834,0.0958833694,-0.2006122172,0.1670179516,-0.1382451653,0.2745993137,0.0647682399,0.1290216297,-0.212790072,-0.0356389806,-0.2517776787,-0.0058522774,0.226232335,0.0662213042,0.2643677592,0.0570805259,-0.0368257388,0.2464054525,0.1964491904,0.2589336932,0.2621414661,0.540453732,0.410502255,-0.2865335941,0.3115128577,-0.3376363814,0.1178432927,-0.2898557484,-0.3992738426,0.0765101612,-0.1542877853,-0.1191549525,-0.1171760187,-0.050810542,0.1809277833,0.0152329504,-0.1221300662,-0.4397419691,-0.1711908132,0.0106513444,0.0177947488,0.000757023,0.200720951,0.1873478293,0.6095921993,-0.1133483797,0.1476428509,0.0923634395,-0.1258233786,-0.0289077107,0.1834839582,0.2011351883,0.1320294738,0.2399761528,0.2614587843,0.0891629234,-0.0447775535,-0.2854482532,0.1424058974,-0.0416227803,-0.0293980613,0.2343936116,-0.2252787352,-0.0814030692,-0.2844702899,-0.1078715771,0.1900336295,0.3333036304,-0.2733850181,-0.2253767103,-0.2778774798,-0.0811996087,-0.2023283094,-0.2605310977,-0.4937746227,-0.3111183345,0.0932645053,0.2138171345,0.0015652273,-0.00290539,-0.2456213981,-0.0235087667,-0.2265373617,0.1571208835,-0.2096743882,-0.2361392081,-0.1060385182,0.0858783573,0.4665773809,0.4596540332,0.5857223272,-0.2945281565,0.2909271419,-0.3250324726,-0.4925007522,0.1065503433,-0.1111655533,0.3061677217,0.0278481282,0.0064205467,-0.033051338,-0.0431219079,0.0325089544,0.0170207079,-0.0860795304,-0.1726716608,-0.0578184798,-0.1768981367,-0.0446688682,-0.5428530574,-0.1846921444,-0.3722652793,0.0926657468,-0.1413791776,0.1481264681,0.108059302,0.0608188137,0.2076724619,-0.1641061604,0.2128446847,-0.2528631985,-0.0471615717,0.483152777,-0.3762440383,-0.2860786617,0.5221699476,0.0618429184,0.3283829689,-0.22356686,-0.2752349675,0.1938800961,0.2008592933,0.0614607818,-0.0204016268,-0.0826197192,0.2799642682,0.2153713554,0.0446460471,-0.2033184618,-0.2075478435,-0.1618234217,-0.0651315823,-0.012049268,0.0391638502,0.1518172324,-0.0432340801,0.4195065796,0.0300061014,0.184286207,0.3928280175,-0.0128651299,0.3791911602,0.047049623,-0.3830392659,0.197731629,0.0401779152,0.1801675856,0.2621788085,0.0809963793,0.1486769468,-0.0885647237,-0.1598252058,-0.3769283891,-0.2243569791,0.1970245242,-0.1406930536,0.0961304531,-0.194437623,0.0117480308,-0.1141679287,-0.2063038945,0.2016546726,0.8246311545,-0.1158847213,0.1593085527,-0.384501785,0.2631762922,-0.027377449,0.0666645244,0.1726115644,0.1781697124,-0.3565507233,0.214683935,-0.0305372085,0.2248924524,0.3928493261,-0.740640521,-0.057441406,0.2093599588,0.1787932813,0.1652439386,-0.0393430777,-0.0803127959,0.1039108187,0.1280385703,0.2370574623,-0.205042854,-0.1506266445,-0.1625539958,0.0157614574,0.1063736081,-0.2557649612,-0.0409351848,-0.2579031289,-0.5642453432,-0.078264229,0.1524727941,-0.2003750056,-0.2615467906,-0.2212112993,0.1958745122,0.1203876063,0.2747458816,0.0823178589,0.0588776805,0.2790190578,0.0877268761,0.2392885089,0.0709757879,0.7544961572,0.3973069191,-0.2398381829,-0.3075879812,0.060138274,-0.8074307442,0.3809865415,0.0540101044,0.2422094345,0.0558554605,0.2597579062,-0.4497740865,-0.1920723766,0.0220504049,0.0874503702,0.0080108568,-0.3625230789,-0.6086066365,0.3206201494,0.2070430964,-0.0592226833,0.2183651626,-0.1070625931,-0.3987124562,-0.1290472001,0.231145367,0.9148773551,0.0308576655,-0.1712233424,-0.3451571465,-0.4033821225,0.1548526585,-0.3034472764,-0.0884220526,-0.2218056023,0.1109474599,-0.0989173055,-0.0335407741,0.4710514843,0.2911988795,-0.2942738533,0.0700265989,0.2238087207,0.2063951492,-0.0518061295,0.7011324167,-0.3197645545,-0.3344275951,0.3089041412,0.1733638942,-0.0158108827,0.0151838539,-0.0882291049,-0.4142897427,-0.1256167889,-0.0210395064,-0.433616221,-0.29973948,-0.0721497387,-0.2690705955,-0.2709858119,-0.0649745539,0.491509676,0.2541736662,0.5265597105,0.1324874908,-0.4326536953,0.0483922288,0.0332511701,-0.2202478647,-0.1651024818,-0.1230398417,0.4145585299,-0.1207267195,-0.2863664627,0.0005600223,-0.0733183324,-0.0285639837,-0.2847560048,-0.0850122273,0.207232967,-0.2811368406,0.1283921301,-0.1535327435,-0.0904169306,-0.200610429,0.1820025444,0.2120548189,-0.0714697838,0.0506916903,0.0995878503,-0.0841392055,-0.1222504601,0.2139973491,-0.2150967121,-0.0058414708,0.0704578012,0.1020489037,-0.235319525,-0.3233818114,0.0702529699,0.0886449069,-0.2820439637,-0.3095580041,0.0614126734,-0.0352438204,0.0950573385,0.1328797638,0.2042713314,-0.0185091347,0.0005058963,-0.3790468574,-0.3702391088,0.0342240185,-0.2673853338,0.1535260081,-0.2898293138,-0.2604185343,0.2245225757,0.1190140769,-0.3590902984,-0.0504847169,-0.0110530173,0.1322031617,0.0799838677,0.1367359012,-0.2101821601,-0.3123732805,0.0132989138,0.0580882095,-0.2618293762,-0.1395930201,-0.159297362,0.1427099407,0.057481423,0.0153549332,-0.1745447665,-0.2434616536,-0.2724674642,-0.0781410262,0.0927745625,-0.078505002,-0.0112109343,0.2574883103,0.1386410147,0.2418695092,0.1323314458,0.0255112033,-0.0375771075,0.1836921573,0.021550335,-0.1092889532,-0.1287727952,0.0987934098,0.2268556207,0.2050512731,0.0678640977,-0.0804052949,0.3333837688,-0.1049419492,0.3814509213,0.5024676919,-0.0427524373,0.1786157787,-0.2168629766,0.0869578272,0.3599404395,0.2423609793,-0.175774321,-0.1600550711,0.4079779685,-0.10844706,0.0340233184,0.2446231693,0.676751852,0.0957913175,-0.2147061527,0.1233725622,0.3482161164,-0.1065704226,0.1027776897,0.3946240544,-0.3101719618,0.0968421623,0.3800020516,0.3323697448,-0.0092316344,0.2448052913,0.07276351,0.2660258114,0.3410690427,0.2237115502,0.3017820418,-0.3540908396,0.3159093261,0.2309855819,0.2680965066,0.1732629836,0.1194523349,0.0512712933,0.2235285938,-0.2030753493,-0.0023635903,0.0438612811,-0.1717479825,0.2513712049,-0.1015837938,-0.1802262068,-0.0769733489,-0.1295685917,0.0470907427,-0.2133479863,-0.0673356578,-0.1005152091,-0.1876310855,-0.0508354865,0.102148287,-0.057199005,-0.0003613117,-0.2093274742,0.0606581941,-0.0896802098,-0.2043671161,0.2844094038,0.1165175661,0.3416538835,0.0779170021,0.0125776865,0.1170876771,-0.0561452806,-0.2006210834,0.1499140859,0.5387471318,0.0797772557,0.1669759154,0.1917854249,0.1372159272,-0.0568487495,-0.0080062589,-0.0328726396,-0.1629106998,0.1068563387,-0.1778678149,0.1730200052,-0.1983295679,-0.1582744122,0.0420843437,-0.2208035588,-0.3285656273,0.7464051843,0.0114616035,0.0166644771,-0.2216933817,0.0808436796,0.2176806629,0.2572939694,0.2359849215,0.0511279292,-0.0788502991,0.044208888,-0.7780945301,-0.0200137943,-0.2156759501,-0.0004589423,0.1460186094,-0.2096678764,0.3006260097,0.2180907279,0.2118472308,0.015851045,0.1395213157,0.1853697598,-0.0782911405,-0.2735221684,0.07687857,0.1883656532,-0.0298586357,0.0039624106,0.2491553426,0.1665064543,0.0615227856,-0.3908043206,0.2039913237,0.0418253653,0.1013740674,0.5442777872,-0.0910800695,0.5890606046,-0.2108475119,0.0496782176,-0.008277677,0.3336321414,-0.0740555227,0.4311032295,0.0965911597,0.1571973264,-0.1267198026,-0.0782226771,-0.2770254314,0.1978088319,-0.0521992184,0.0593078993,-0.5329076648,0.0613827221,0.0637991279,-0.0955714956,0.104689002,-0.1525367647,0.0139467157,-0.0539960377,-0.1509700865,-0.2541344166,0.4322204292,-0.0579260178,-0.0995429084,0.0221746713,0.020163646,0.4060533941,0.20046857,-0.8780812025,0.0528768264,0.4730555117,-0.0019283269,-0.1871085465,0.0964555889,-0.015067107,0.05717979,-0.164305836,0.1252115518,0.0883672982,-0.1450264752,-0.127170369,-0.1554459184]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2089","title":"Add documentaton for dataset README.md files","comments":"@lhoestq hmm - ok thanks for the answer.\r\nTo be honest I am not sure if this issue can be closed now.\r\nI just wanted to point out that this should either be documented or linked in the documentation.\r\nIf you feel like it is (will be) please just close this.","body":"Hi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip","comment_length":51,"text":"Add documentaton for dataset README.md files\nHi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip\n@lhoestq hmm - ok thanks for the answer.\r\nTo be honest I am not sure if this issue can be closed now.\r\nI just wanted to point out that this should either be documented or linked in the documentation.\r\nIf you feel like it is (will be) please just close this.","embeddings":[-0.1834505647,0.251365006,-0.108965151,0.1974366605,0.1107171848,0.3312231898,0.3745901287,0.1719959974,0.0320847295,0.0473188944,0.0283438452,0.2382667959,-0.06312792,0.0927071273,0.0448256098,0.0325517729,0.0677399784,-0.0201820545,0.2282102108,-0.2717784047,-0.1881774515,-0.0120634753,-0.1926741004,-0.2600643933,-0.2601344883,-0.1061909795,-0.1686320007,0.1679512262,-0.4175153971,-0.5190830827,0.0170215424,0.2592081428,0.1794570684,0.069258146,-0.000108684,-0.2377832085,0.2274885327,-0.2222405076,0.0124146575,-0.0039696,-0.256370604,-0.392083168,-0.2011699975,-0.1843143404,0.0813640505,-0.1434149742,-0.1106911227,-0.2170637399,0.1881171465,0.5700099468,0.2062799335,-0.2507901192,0.3001177907,-0.2143563777,0.4646461606,0.2964088321,-0.1445172429,0.3467390537,0.5063992143,0.116429463,0.0515992343,0.1678493321,-0.1461413056,0.1065743566,0.311891824,-0.1332371533,0.4417340159,-0.5099037886,0.2496395558,0.3756162524,1.1862678528,-0.1897372007,-0.1440565884,-0.3056923747,0.0748555884,-0.0499141552,0.1617724895,0.3897077143,0.0141748274,0.144161731,0.138372466,-0.3804753423,-0.3181429207,0.0608467199,0.1116281077,0.419542253,-0.0779646263,-0.075421229,0.1478106678,-0.0849926621,0.2784859538,0.0570362695,-0.2141937315,0.225955084,0.189255476,-0.1122400016,-0.3635398746,-0.1978414059,0.2118561715,-0.1735544801,-0.0536954589,0.0306305848,-0.0166816674,0.2824446261,0.1840374321,-0.098953709,0.1554633826,0.2676413357,0.1074184477,-0.2280970961,0.2706859708,-0.0682267174,-0.2281811833,-0.1300102174,-0.0250221174,0.041349709,-0.0668682903,-0.345576942,0.0765684023,-0.0438406542,0.0707652122,-0.148353368,0.0131454086,0.2365357429,0.0193656068,-0.1580260992,-0.2276223749,0.109149605,0.0131610446,-0.1187187806,0.0190704297,0.2393930852,-0.1226379648,0.190498367,0.267563045,0.0570891798,0.206442818,0.2571180761,-0.0690384209,-0.2132615745,0.2207864374,-0.0350578055,0.3134301901,0.3436415792,-0.0439306572,0.0869112089,-0.0894113556,-0.3784480393,-0.1885696352,0.1912267357,-0.5235507488,-0.2824635208,-0.3953041732,0.205935955,-0.0752783492,0.0405865051,-0.1238837391,0.2208991647,0.1222835556,-0.1311614513,0.1706708819,0.1315923631,-0.379452616,-0.2931358516,-0.0147849862,0.4112755954,-0.3530879319,0.0920699909,-0.1462333351,0.0608904026,-0.0973646492,0.0338928849,-0.3202523887,0.3123343289,-0.3375391662,0.1994763911,0.4865379035,-0.6444005966,-0.2490532696,0.1680991501,0.0996946618,-0.337928623,0.1998403221,-0.076159656,0.5245919228,-0.2472617328,-0.1961391121,-0.1009456366,-0.1206154674,-0.044988554,-0.2056626827,0.0024414766,-0.0279927906,0.1433908641,0.0930473357,-0.0085915076,0.5141079426,0.111548692,0.11130023,-0.1130470037,0.2350051701,0.5121799707,0.278008312,0.0733763576,-0.0036751251,0.0601929538,-0.3458781242,0.1016534045,0.0381818712,0.2979979217,0.0225258414,-0.2697308958,-0.3078657687,-0.2098493576,-0.3070385754,-0.3466337323,0.1509182602,0.072135374,-0.2899815142,0.1742564142,-0.0583876967,0.258552283,-0.022553971,-0.0229609944,-0.1868529618,0.0147313429,-0.1795284599,0.0714114159,0.3635661602,0.1179975122,0.2934602201,0.0673181117,-0.0512963347,0.1727052331,0.1681962311,0.1867890209,0.1970174909,0.5089138746,0.348587513,-0.3183594644,0.3216242194,-0.3319413066,0.1059602946,-0.2947618067,-0.3823413551,0.093760021,-0.0875143111,-0.1562456191,-0.096652925,-0.0550427102,0.1559996605,-0.0430997014,-0.1213439107,-0.418656379,-0.2139490098,0.1382642686,0.0492486134,0.0433533676,0.1525081694,0.2591985166,0.6951568127,-0.089364782,0.1233370453,0.0477645323,-0.044406157,-0.0458785295,0.2343299687,0.1793375909,0.2504524589,0.1950840801,0.2688598633,0.1685174257,-0.025638463,-0.2154762894,0.1143739894,-0.1564902067,0.0548340157,0.1859874725,-0.2092626542,-0.0657521039,-0.2408761233,0.0019681288,0.1768082976,0.3061312735,-0.2752851844,-0.3088805974,-0.2675393224,-0.010409792,-0.2429445237,-0.1491997838,-0.5167268515,-0.174337998,0.0638108775,0.1178996637,0.0370092876,-0.0429781489,-0.1899590045,0.1676542461,-0.2003879398,0.2505880892,-0.1858588606,-0.2149976641,-0.2299165875,0.0794585422,0.4636722505,0.5456963181,0.5315335393,-0.2386776805,0.2786902189,-0.3564065397,-0.440682143,0.0858712047,-0.0464326851,0.3145121932,-0.0619887598,0.1028822437,-0.0427765138,0.0008728108,-0.0199087579,0.1111815944,-0.1103436574,-0.2667787075,0.0028240881,-0.3291390836,-0.0390195288,-0.594435513,-0.1559260488,-0.3857374489,0.0730233863,-0.1942426115,0.2018292248,0.0708701611,-0.0117161777,0.1006474048,-0.1449375451,0.1740794182,-0.3098455966,-0.1001522988,0.4769842029,-0.4734630883,-0.4342747331,0.5254812837,0.0934798867,0.3927616477,-0.1414426416,-0.3136377335,0.2251268327,0.1927728355,-0.2009271532,-0.1818776131,-0.0200747941,0.3267469704,0.2113699168,0.0663215294,-0.2078844458,-0.2461215407,-0.0968589932,-0.0086201597,0.0727901161,-0.0197774451,0.0684674531,-0.0048043313,0.3320693374,-0.0051236353,0.2867011428,0.3341181278,-0.0068125003,0.3818561733,0.1176935434,-0.3456970453,0.2654700279,0.0527135134,0.142698735,0.3200286925,0.1204754487,-0.0220799837,-0.1031062156,-0.2067573965,-0.4100351036,-0.2712362111,0.1620358974,-0.1131092533,0.1874025017,-0.13714616,0.0095383804,-0.1100553349,-0.1893767267,0.1881662607,0.8119858503,-0.1975711584,0.1577806324,-0.4499599934,0.2427947521,-0.0247811545,0.0462701656,0.2061759531,0.2806915641,-0.4020807147,0.2396637797,0.0146312537,0.2007170022,0.3344191909,-0.6941183805,-0.0313096605,0.2469896972,0.2598072886,0.1858717948,-0.0627180263,-0.0573009737,0.1486288309,0.0910740942,0.2207508385,-0.1442534178,-0.205688417,-0.1584260166,-0.0439018607,0.0089819161,-0.3161910772,-0.0455073044,-0.2997612655,-0.4207756817,-0.1131021529,0.2319419533,-0.2269568741,-0.3190456629,-0.3087130189,0.211005196,0.0780145377,0.2206957936,0.0608880743,0.080651693,0.2518090606,0.0275456067,0.2140597105,0.0965084136,0.7797985077,0.3743667901,-0.2350525856,-0.2111730576,0.0804942995,-0.8616188169,0.3713947237,0.201215297,0.1970300525,-0.0521993823,0.1709300429,-0.4224600792,-0.0202349927,0.0221521016,0.1396090984,-0.0443696119,-0.5320263505,-0.6122278571,0.2845601737,0.0895267278,-0.0566586144,0.2384880781,-0.028061945,-0.3622268438,-0.13422288,0.3018082082,1.0777900219,0.01652156,-0.1200805232,-0.3285025358,-0.3679195642,0.1422358304,-0.2336852849,-0.100995943,-0.3288811147,0.1367958188,-0.0901475251,-0.0349895917,0.4635131061,0.3538165987,-0.3111799359,0.0179573316,0.3268076777,0.0849093944,-0.0209026616,0.6265957952,-0.2827164531,-0.2374101281,0.2187497616,0.1473239809,0.0629228801,-0.0408319533,-0.1525386125,-0.4462299049,-0.2223451287,-0.0729608387,-0.3919940591,-0.2424043715,-0.1159103587,-0.2178744376,-0.2231838852,-0.0555174127,0.5816289186,0.2121003419,0.4717833102,0.2234110683,-0.4294231832,0.1494556516,0.1007508487,-0.2574824989,-0.0995198935,-0.0714194775,0.3697448969,-0.2065969855,-0.2942830622,-0.0452797301,-0.221146524,-0.0304519404,-0.356975168,0.0559318811,0.2022374868,-0.2618910074,0.0648248345,-0.176684916,-0.135613516,-0.0611033849,0.1754431874,0.1777977794,0.0257324576,0.1413969696,0.1436350793,-0.1016764641,-0.020080097,0.2583363354,-0.0717073753,-0.1364506483,0.1451640725,0.0553957708,-0.1846818328,-0.3555240929,0.0928392112,0.0359540805,-0.1644406617,-0.3201909959,0.0094084442,-0.0408768207,0.1302224249,0.1905411333,0.1497412026,-0.0666366667,-0.008763439,-0.3896551132,-0.3948333561,0.0196484625,-0.3371899724,0.1265928,-0.3014386296,-0.2930763662,0.2061997652,0.0458548069,-0.3518182337,0.0463753529,-0.0195582341,0.1141611934,0.1615622044,0.1669155955,-0.208194688,-0.2656026185,-0.0043309145,0.0928427726,-0.2421131581,-0.1493668556,-0.102693148,0.1623726785,0.0701809749,0.0178050157,-0.1415640116,-0.3763379157,-0.2985941768,0.0253548883,0.0363175608,-0.0169916935,-0.0258339755,0.2100508362,0.1485459358,0.2503185868,0.0477388613,0.181237191,-0.0379158705,0.1779117435,-0.0292303786,-0.022249626,-0.1017936841,0.0560355559,0.3677481413,0.1331212968,-0.000875372,-0.0102632549,0.382704556,-0.0980423167,0.3971006572,0.3668281138,-0.1173237115,0.2400014699,-0.2710825503,0.1094223782,0.3641287684,0.2230263054,-0.1283956319,-0.176691398,0.2905825078,-0.1728892028,-0.0529514104,0.2369321883,0.6858570576,0.0375597477,-0.1410785615,0.1213962436,0.3364261985,-0.0668805391,-0.0066330638,0.3637579978,-0.3210895956,0.0258628614,0.3457219005,0.2561428249,0.0270670932,0.211230129,0.2078403831,0.157172665,0.4067305624,0.2854524851,0.2139687985,-0.3492390513,0.3475604951,0.2237096727,0.2600417137,0.2057801485,0.123928763,0.0404876247,0.3550107181,-0.2074645013,-0.0038652446,0.0089760618,-0.1628413945,0.2597954571,0.0170353912,-0.1725346297,-0.107103385,-0.1465835422,-0.0477203168,-0.1930259019,-0.1966565549,-0.038809523,-0.1325841695,0.0795845762,0.1474021524,0.0681676567,-0.0064935233,-0.1807656288,0.0624091178,-0.1190536916,-0.1388164312,0.3226592839,0.0985401273,0.3425760865,0.1869330853,0.0227838531,0.0996563733,0.0919608101,-0.1991127133,0.1540684402,0.4935736656,0.1587813795,0.2255327553,0.1589426249,0.116051361,-0.0620893389,-0.0103721526,-0.1367608309,-0.2212295979,0.1200184375,-0.1878756881,0.1110759601,-0.2756795287,-0.1833453327,-0.1039051414,-0.2305717468,-0.3599985838,0.740899086,-0.031478323,0.0653128996,-0.2456745058,0.093450062,0.1306949407,0.3288664222,0.272908628,0.0109779453,-0.0049298671,-0.1226842701,-0.8196237683,0.0528867245,-0.1650004685,0.010589621,0.1823818386,-0.1700480729,0.2991136611,0.151156038,0.2017651945,-0.0778430402,0.1487456113,0.2435832024,-0.1664606482,-0.238110885,0.1285687834,0.1655010581,-0.0566773117,-0.0922164693,0.294996202,0.2356586009,0.0685037896,-0.3996609449,0.2636984587,0.0268791001,0.1268487126,0.4562400877,-0.0580274127,0.5467060804,-0.2063592821,0.0199130867,-0.0018069807,0.392599076,0.0396026224,0.42384094,0.1667914093,0.1110908464,-0.1063587368,-0.0735869929,-0.2389905453,0.227880314,-0.0945502669,0.115051724,-0.5860778689,0.0501819141,0.0824644417,-0.0306458045,0.0794337317,-0.116082184,-0.0748553425,-0.0965918303,-0.2133959383,-0.182393387,0.4661231935,-0.1344535798,-0.1463748962,0.109036006,0.1153692007,0.2609986961,0.1699074209,-0.8375416994,0.0024473926,0.4686846733,-0.0455614962,-0.1749612987,0.0643828511,-0.0900104046,0.0679630488,-0.1342227757,0.1363300085,0.0832288936,-0.118756935,-0.1518054157,-0.1722614616]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2089","title":"Add documentaton for dataset README.md files","comments":"We're still working on the validation+documentation in this.\r\nFeel free to keep this issue open till we've added them","body":"Hi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip","comment_length":19,"text":"Add documentaton for dataset README.md files\nHi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip\nWe're still working on the validation+documentation in this.\r\nFeel free to keep this issue open till we've added them","embeddings":[-0.2152332067,0.2649821639,-0.1258544773,0.1683345139,0.153197214,0.2933406234,0.2924147546,0.1951218694,0.031965781,0.020962853,0.0500942841,0.2049097121,-0.0792760625,0.1006224751,-0.0031886012,0.0226522125,0.0487911105,-0.0505709611,0.209790498,-0.2532515228,-0.1110622212,-0.0721068233,-0.1680813283,-0.2169418484,-0.279617697,-0.1498723179,-0.121067971,0.1117399558,-0.5551061034,-0.5625538826,-0.0569595918,0.3236434758,0.2056470215,0.0757121667,-0.0001088561,-0.2372320294,0.3131054044,-0.1969533563,0.0418356284,-0.0134433964,-0.3186578751,-0.3487761617,-0.1418905705,-0.1790460944,0.0734013245,-0.1668996066,-0.0970635936,-0.1242037937,0.2178792953,0.6139640212,0.192051664,-0.176824674,0.3337362409,-0.1620907336,0.3486065865,0.2870318294,-0.1387577951,0.3151708543,0.4389590025,0.1328053623,-0.0011085592,0.172968775,-0.0888033286,0.0742474571,0.354427278,-0.1740669608,0.4313996136,-0.5642960072,0.1373408437,0.4447307587,1.1757175922,-0.2466039658,-0.147630468,-0.2829697132,0.0595339239,0.0118933395,0.2021713555,0.3936532438,0.0201638862,0.1482875943,0.0523763299,-0.2962444723,-0.3464110196,0.027329538,0.1638356596,0.444943279,-0.0405867808,-0.1621433347,0.2406282723,-0.0741209388,0.2830766141,-0.00571978,-0.2207933366,0.2731741071,0.1728255004,-0.1457782239,-0.3448441029,-0.1132450327,0.2412940264,-0.1358947456,-0.0598471463,0.0647012591,-0.0455052815,0.3457221389,0.199695304,-0.0847962797,0.184991315,0.2489361912,0.113955982,-0.192776233,0.2991896272,-0.0795426071,-0.149065882,-0.1413848996,0.0583407022,0.1044747978,-0.1232852116,-0.3599949479,0.0938883796,0.0048555108,0.1318528652,-0.2013231069,0.0663122311,0.1965319812,0.0225523729,-0.1681499779,-0.3468319476,0.0708110183,0.0511130691,-0.1352705657,0.0246487856,0.2435915619,-0.1412160397,0.1586894989,0.2875902355,-0.0040694489,0.2766801417,0.1819750816,-0.062001776,-0.2075318396,0.2298635095,-0.0740521848,0.3065574765,0.3483856022,-0.0645222738,-0.0704309493,-0.1015463248,-0.3342269659,-0.2269043475,0.1176794767,-0.3748679459,-0.3299014866,-0.3605083525,0.2099002898,-0.0848126188,0.0780488551,-0.0824628845,0.2596985996,0.1449100226,-0.0913847238,0.1895816177,0.1342038512,-0.2440786362,-0.2446256876,-0.0487653241,0.3932786286,-0.3049049675,0.1307745278,-0.1182967275,-0.0363224708,-0.1880007982,0.0069952076,-0.3300700784,0.3603716195,-0.2602694035,0.2314064205,0.5233443975,-0.6178893447,-0.190892309,0.2019526362,0.1188919842,-0.3653545976,0.1996655613,-0.1225311011,0.5625225306,-0.3021692634,-0.2760796547,-0.0837178305,-0.2096390724,-0.0862350911,-0.2159940302,0.065836899,0.0004535391,0.0925549418,0.087124981,-0.1019032076,0.4954625666,0.149652034,0.0810179487,-0.1325366795,0.1871227175,0.4300885499,0.3101040721,0.0051652817,-0.0187908895,0.0297723208,-0.3159400821,0.0068745865,0.0792017356,0.3121462464,0.0602492094,-0.2735976875,-0.309137702,-0.2166574746,-0.3294516206,-0.3489198983,0.1617864966,-0.0128998496,-0.1812311262,0.1981271356,-0.146534577,0.2803251147,-0.0659631118,0.046155367,-0.1705372483,-0.0243125334,-0.2345082313,0.0765738711,0.2834572196,0.1128891408,0.2357717752,0.059747085,-0.0370491147,0.2769049108,0.2859248221,0.1337736696,0.1228956655,0.5454466939,0.406909287,-0.3052136898,0.3498113751,-0.3611955941,0.0050803572,-0.2370350063,-0.3916619122,0.0973380879,-0.2199732363,-0.1376938224,-0.091186896,-0.0909057707,0.1451142579,-0.0685637593,-0.2191044688,-0.4312099218,-0.2434509695,0.0687407479,0.0123265199,0.0651024878,0.0938151777,0.15814282,0.6090510488,-0.1135696471,0.133507818,0.073922351,0.0357706137,-0.1065810472,0.2230447829,0.2026791722,0.1815531105,0.137719512,0.2089840025,0.1837079674,-0.1069417372,-0.2314970493,0.1740404367,-0.1743568182,0.0433527939,0.1960575879,-0.3303851485,-0.0107408874,-0.2964419723,-0.0297243763,0.1803777963,0.3155517578,-0.2707270086,-0.2726546228,-0.2162563354,-0.0204582606,-0.1964751929,-0.1118424535,-0.4751445949,-0.2135225832,0.1043400839,0.1421674192,0.0377768651,-0.0222376026,-0.2111313641,0.0961314291,-0.174366191,0.219769612,-0.2013042569,-0.229246676,-0.1875761449,0.098502405,0.4579515457,0.5772249699,0.4862652719,-0.24764736,0.2650541067,-0.2878770232,-0.5216217041,0.073495537,-0.0353743546,0.299456507,-0.0513310991,0.0952728689,-0.0173348971,0.0269205794,-0.0814487487,0.0998441353,-0.1886216104,-0.2019718736,-0.0158067979,-0.3107983172,-0.028933933,-0.5464140177,-0.1762472689,-0.3826165199,0.0350002348,-0.2176812738,0.1791503876,0.0694593266,0.0290386882,0.089963913,-0.2129889578,0.2034207135,-0.2511434853,-0.0605329312,0.4775668979,-0.4183160663,-0.4114857316,0.4373288155,0.0720372126,0.363257587,-0.1737964302,-0.2805134654,0.1150470078,0.1619207412,-0.1244618818,-0.0841160268,-0.0052799494,0.3449583352,0.2725786567,0.0719382241,-0.1632326096,-0.2096580118,-0.1451490223,0.02962064,0.1314594299,-0.0244175438,0.067877017,0.0100661647,0.3052534759,0.0540943593,0.1991354972,0.3418984413,0.0613150336,0.4321850538,0.1601080745,-0.3594634533,0.3082561493,0.048932787,0.1297257841,0.1927376986,0.0335303321,0.0281047318,-0.0965796411,-0.2087707371,-0.4129326344,-0.2539665699,0.1316394508,-0.1952831,0.2315448672,-0.1327530146,0.0222197529,0.0530538559,-0.1492286325,0.2017635405,0.819164753,-0.129289344,0.1271850318,-0.453255266,0.228445828,-0.0516444147,-0.0158983897,0.1734146327,0.2282457501,-0.3218296468,0.2446126789,-0.0393139832,0.1863501668,0.2971590161,-0.7153296471,-0.0542860962,0.1525995433,0.2156134993,0.2075253576,-0.0387250632,-0.0837579742,0.0629183203,0.1047913656,0.2673693299,-0.2166026533,-0.245544374,-0.129427582,-0.0640398115,0.0481542535,-0.3154783547,-0.0323937461,-0.3209096491,-0.4009484649,-0.0511271767,0.1986699402,-0.2276548892,-0.3399577141,-0.1832659692,0.1877876967,0.1165814027,0.2261722088,0.1751224548,0.1291006505,0.1373228729,0.011145642,0.2052291483,0.0505239926,0.744935751,0.2976525128,-0.2528681457,-0.2054596543,-0.0036787165,-0.8008688688,0.3690815568,0.0956031233,0.2240999788,-0.1240482852,0.1854742467,-0.387586236,-0.0491374619,0.1045400947,0.0718085617,-0.0467887521,-0.5273841619,-0.6166073084,0.2538611293,0.1003488228,-0.0705331787,0.1761015505,-0.0395003855,-0.372125864,-0.1266324967,0.4468619525,0.9914255142,0.1262829006,-0.2045708895,-0.3064756095,-0.3717343509,0.052141916,-0.3224625289,-0.1487305462,-0.3163974285,0.1355244964,-0.10691984,-0.0483630709,0.4740169048,0.3395563364,-0.3923214078,0.0449133441,0.3716097176,0.0468901768,-0.045536682,0.778431356,-0.3186798692,-0.2367444783,0.2047445029,0.1485614479,0.0571701489,-0.0882895514,-0.114243634,-0.4399167597,-0.0712076649,-0.1273901016,-0.4061338305,-0.2920452952,-0.1056122482,-0.2347762883,-0.2092839032,-0.0353603736,0.5741903782,0.2189936638,0.5228844285,0.2558458447,-0.4140743017,0.1219184026,0.1434279084,-0.2566819489,-0.1321093589,-0.1000360698,0.4088859558,-0.1934925467,-0.2750222981,-0.1079204306,-0.20196338,0.0491352901,-0.2805183232,0.0071555818,0.123335138,-0.2930015922,0.100442186,-0.2139693052,-0.1978071928,-0.0495634563,0.1837702692,0.1370212883,-0.0501509383,0.2450447381,0.1658842117,-0.0298105814,-0.0288016256,0.1847514212,-0.159929812,-0.1423684359,0.0201721378,0.0121787041,-0.1760814339,-0.3845846057,0.1371570081,0.0841647536,-0.1237304211,-0.3440847099,-0.0246255677,-0.1328614652,0.1308331192,0.1908018887,0.2627682686,-0.0222688969,0.0386842452,-0.3699916899,-0.3957141042,0.0528622046,-0.278219372,0.1491161436,-0.3459352851,-0.3236505985,0.169055596,0.1183301732,-0.3498450518,-0.0151389707,0.0446687415,0.1642171592,0.1244405955,0.1052194387,-0.2040437758,-0.2846588194,-0.003975939,0.1282729805,-0.237268284,-0.1594514847,-0.0915248021,0.1603152007,0.1045829058,0.0281223767,-0.2157405764,-0.2551920414,-0.3554727137,0.0011782108,0.0073995688,-0.0960798115,-0.0068954285,0.314222455,0.1840312034,0.175030902,0.0671591014,0.0826026872,-0.0483656526,0.1907180995,0.0435034372,-0.0402087681,-0.0470530167,0.0312540568,0.3196408153,0.2204013616,0.0649542809,-0.0196187478,0.3327846825,-0.0325327814,0.3535744548,0.2818246782,-0.0220320057,0.3322063982,-0.1898075789,0.0579260699,0.3802912235,0.2299609333,-0.1175769046,-0.1709210873,0.2942973971,-0.155504033,-0.0043481644,0.2126320153,0.6201548576,0.1270665675,-0.1094940901,0.126518026,0.3498152494,-0.0843791515,0.0455260761,0.3721078038,-0.3127887845,0.0199328065,0.3568803966,0.2236070037,-0.0063331928,0.1896829903,0.2401071936,0.188901186,0.3412061334,0.2725245655,0.230956465,-0.226235792,0.3366862833,0.2455381006,0.2224627435,0.1677641273,0.1381456554,-0.0108236363,0.362106204,-0.2713707685,-0.0867323875,0.0033020133,-0.1210903227,0.2783879936,-0.0242258273,-0.2079207748,-0.0468395688,-0.133020848,0.0422125719,-0.1667911857,-0.1251584888,-0.0157970134,-0.1101193354,0.0106614185,0.2138074785,0.1266691089,0.0282290373,-0.1807094514,0.1252239943,-0.0621807985,-0.1493133307,0.2648481429,0.1393435597,0.306221962,0.1757593602,-0.0334936194,0.1038624719,0.0300800595,-0.2353146821,0.2081428468,0.5127877593,0.2062882036,0.2292843014,0.1889365613,0.1285614371,-0.0238433238,0.0586725958,-0.1586120427,-0.2941415012,0.115585953,-0.1173787192,0.277680546,-0.2918071449,-0.1561423391,-0.0569998845,-0.2494588643,-0.3193227351,0.7226656079,-0.0276611857,0.0874958038,-0.3053849041,0.0853744522,0.2062303722,0.3919088542,0.2959234715,0.0102295345,-0.0385658145,-0.1136142761,-0.8005523086,0.0155005492,-0.1430623382,0.0058486029,0.2073902637,-0.209223032,0.2151771486,0.2566597462,0.175708726,-0.0690447539,0.099763833,0.259283632,-0.1479414999,-0.1535761356,0.1440171748,0.2405224144,-0.0750675648,-0.0351719409,0.2898609936,0.1919389069,0.0664849579,-0.3775225878,0.2747054696,0.0822802708,0.1815143377,0.4061072171,0.0140217077,0.5782948732,-0.1807848364,0.0282423813,0.0103864241,0.4699586928,-0.0213535056,0.3975284696,0.173147276,0.1162141934,-0.0753777027,-0.0973030701,-0.2672234178,0.2640320659,-0.0881977081,0.0404768698,-0.537892282,-0.0544229038,0.163933754,-0.1024201512,0.0689290538,-0.1300317496,-0.0429740995,-0.0417467728,-0.1964355111,-0.2542347312,0.4476870894,-0.0841647983,-0.1262207776,0.0838516206,0.0271256603,0.2643665075,0.1328545213,-0.9024062157,0.0439189486,0.4450147152,-0.0812542886,-0.1915602535,0.0787962303,0.0548191257,0.0558449291,-0.1466098428,0.1313599944,0.0618939288,-0.1494855434,-0.2070047557,-0.1194759011]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2089","title":"Add documentaton for dataset README.md files","comments":"@lhoestq what is the status on this? Did you add documentation?","body":"Hi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip","comment_length":11,"text":"Add documentaton for dataset README.md files\nHi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip\n@lhoestq what is the status on this? Did you add documentation?","embeddings":[-0.2382017821,0.2307251394,-0.12107943,0.2242703289,0.0777842328,0.285769105,0.357168287,0.2082532346,0.0231781956,0.003575051,0.0194034074,0.2430146784,-0.0810823441,0.1499089897,0.0315339677,0.0419154353,0.0424560197,-0.0595662147,0.1584259719,-0.2877525687,-0.1939023286,-0.0034312014,-0.2022227496,-0.2588932514,-0.2402795255,-0.1145347208,-0.1969533712,0.1336056441,-0.4516215026,-0.5080741048,-0.078761749,0.3219215572,0.2283244878,0.1567933857,-0.0001107983,-0.2528265715,0.2467262149,-0.2200267315,0.023897808,-0.0343634672,-0.2735436559,-0.3961105049,-0.1296986192,-0.2020410001,0.0987432152,-0.1950823814,-0.1598177403,-0.1798450798,0.1975742579,0.5709463954,0.1979782432,-0.2741444111,0.3411794305,-0.123206079,0.468990773,0.3005565405,-0.1423635781,0.3145506084,0.5015829206,0.1502025127,0.0531735122,0.1116694808,-0.1083510965,0.1322141588,0.2941990495,-0.1196355447,0.4581794143,-0.5745273829,0.1844039112,0.3873667121,1.2237083912,-0.2271257192,-0.1482422352,-0.3146815002,0.1004288569,-0.0377368666,0.1716087013,0.3275838792,0.0230393689,0.1064472422,0.0465183668,-0.3872238398,-0.3484821022,0.0304730553,0.1511501074,0.4315900505,-0.1152539775,-0.0837862566,0.1705452502,-0.0948333219,0.327195406,-0.0168162193,-0.1761334985,0.2535949945,0.1611305475,-0.134071365,-0.3483766317,-0.1095171198,0.2270707786,-0.1855770499,-0.1000994742,0.0622532032,-0.0331867784,0.2725318372,0.1599832326,-0.0980608016,0.1437219977,0.2207628191,0.1378806978,-0.2418580949,0.2272423059,-0.0991121233,-0.2452053726,-0.0903480724,-0.0306579433,0.076536499,-0.0853418037,-0.3645092845,0.1453969479,-0.0635715798,0.1039573029,-0.2101813853,0.0349516086,0.2222395539,0.0494414754,-0.1550487429,-0.3052637279,0.0843680277,-0.0129708704,-0.1352057755,0.0063666566,0.1838555783,-0.1002186909,0.1784711033,0.2696610987,0.0600286536,0.1894936711,0.1955830157,-0.0167743936,-0.1936303228,0.226068154,-0.035807211,0.2463222146,0.3049765229,-0.0201430637,0.0613066033,-0.0945795178,-0.3468339741,-0.2515133321,0.1600264907,-0.4446657002,-0.318644166,-0.3217230737,0.2000313103,-0.0616638027,0.0675044805,-0.1230831444,0.2567409277,0.1724913716,-0.0946908742,0.1675537676,0.1209061295,-0.3025279045,-0.306877315,-0.0177271664,0.4363150001,-0.356626004,0.1213078797,-0.1483422816,0.0130998483,-0.1502872556,0.0057985117,-0.3053542078,0.3262139857,-0.278478533,0.157468155,0.5449951887,-0.6003803611,-0.2398066074,0.221661374,0.0704741701,-0.3534964919,0.1734768748,-0.0832111537,0.564189136,-0.2616329193,-0.2730780542,-0.0564036779,-0.167321831,-0.0736944377,-0.2166149318,0.0006402799,0.0040781768,0.1651628911,0.0891690552,-0.0301033519,0.5168114305,0.0806958005,0.0860267356,-0.0787265599,0.2207030803,0.3858751655,0.3125066161,0.1027598679,0.0176893547,0.1053028703,-0.3504444063,0.0321711153,0.0723712295,0.2682656944,0.0578225143,-0.2944746017,-0.2499279976,-0.2519793808,-0.3532543778,-0.3060566187,0.1546335369,0.028365612,-0.267593205,0.1969803423,-0.1469881833,0.2964144051,-0.0377334878,-0.0071758037,-0.2015296072,0.0031553018,-0.2073200047,0.0522334687,0.2823230326,0.1108751744,0.2597791851,0.0104890717,-0.060960263,0.2342997342,0.1782829314,0.2049443275,0.2082875967,0.4826801419,0.4023012817,-0.3261708617,0.4281273186,-0.4049421549,0.0373246819,-0.2863466442,-0.3341762722,0.0693129674,-0.2022868097,-0.1347599626,-0.0939670876,-0.1192478836,0.0950722396,-0.0107496437,-0.1302951723,-0.3814884424,-0.2588594258,0.119348675,0.042167455,0.0144002801,0.1218737364,0.2183638662,0.6343774796,-0.1168134511,0.0651189759,0.0844819173,-0.0045685079,-0.0798681453,0.182779029,0.2188086063,0.2263330966,0.1738400012,0.274286747,0.165068239,0.0416703373,-0.228757605,0.1384708881,-0.1610216647,0.03479423,0.2139618695,-0.3033936024,0.0052351397,-0.2582336962,-0.0644012392,0.1825442463,0.3306224644,-0.2619177699,-0.3329291046,-0.201536119,-0.0361574441,-0.1863036156,-0.1398624629,-0.4571513832,-0.1647140682,0.0460740738,0.090967007,0.0890410766,-0.0145526044,-0.1752408892,0.159747079,-0.1409135312,0.209213227,-0.1864109337,-0.1975218505,-0.1850772947,0.0829271823,0.4582439065,0.5748545527,0.5003085136,-0.2450153828,0.3081699014,-0.2758024037,-0.4390899241,0.0492139198,-0.0654588863,0.2865407169,-0.0577931739,0.0399369635,-0.0438095406,0.0632626712,-0.0165203195,0.0865346044,-0.1636763811,-0.2716554105,0.0127265342,-0.2613998353,-0.0066004186,-0.5313936472,-0.251401931,-0.419367522,0.0611031316,-0.1473015845,0.2091094404,-0.0127881328,0.0031441334,0.1487190872,-0.1317075342,0.1985785216,-0.2548788786,-0.1189383417,0.4998610914,-0.4561309218,-0.4031468332,0.4994611442,0.1284713745,0.3512805104,-0.1366043091,-0.3613204062,0.2071905136,0.1761266291,-0.100490585,-0.0921192616,0.0715648979,0.335488975,0.2523616254,0.065702863,-0.1564962268,-0.2450284958,-0.1376985312,0.0000510375,0.1073704511,0.034270063,0.1570424139,-0.034858834,0.3351401687,0.0980879068,0.2328974903,0.3561880887,0.0721960068,0.3788128793,0.12919043,-0.352083236,0.3067707717,0.1015610173,0.0971399173,0.2806913555,0.1122598276,0.0144504616,-0.1441479623,-0.2491993755,-0.4029028118,-0.2586832345,0.1353016645,-0.1777350008,0.2150707841,-0.2188319862,-0.0314015411,-0.0079452433,-0.160999313,0.1751051694,0.8299337029,-0.1490165889,0.1176219806,-0.4854890406,0.2191637605,-0.0382376499,0.0350186005,0.147222653,0.2502410114,-0.3638915718,0.2342741489,0.0194120184,0.1980069429,0.3397782743,-0.694288373,-0.0694337636,0.2027306259,0.174333483,0.1901370883,-0.0050942674,-0.0739479661,0.1391361505,0.0768235177,0.2346041501,-0.2092064172,-0.2126745433,-0.1357084513,-0.0521975085,0.0429634303,-0.3694679141,-0.0064463634,-0.3270926476,-0.391507864,-0.0938077718,0.1549540013,-0.2849597335,-0.3738642335,-0.2731730342,0.2052716166,0.067247659,0.2338292003,0.0776828676,0.0653762221,0.1822101474,0.0649858192,0.2881248593,0.0153493518,0.7638105154,0.3473385274,-0.2996017039,-0.1567117721,0.05044429,-0.8324533701,0.3898203969,0.1477258056,0.2027561069,-0.0912162513,0.1379865259,-0.3686730564,0.0026208581,0.0049273814,0.0677920505,-0.0493827835,-0.5439712405,-0.669467628,0.2955725193,0.058971554,-0.0917917341,0.2525810003,-0.0472646914,-0.3702811301,-0.0728510544,0.3281899691,1.067065239,0.0344536416,-0.1668466032,-0.358459264,-0.3431890309,0.0894242749,-0.2671737373,-0.1855907291,-0.323754549,0.1818853021,-0.0842370391,-0.0634055808,0.4777651131,0.3672617376,-0.4085952044,0.0097614145,0.3270622492,0.0013934285,-0.0386817493,0.7176378369,-0.3095363081,-0.2387333661,0.2416205555,0.1509019434,0.0451666601,-0.0776779652,-0.1133972332,-0.4386181831,-0.1481463164,-0.0503458455,-0.3170398772,-0.2887897491,-0.1091233566,-0.2237955779,-0.2422574311,-0.1131319404,0.6448535323,0.2213968039,0.4948666692,0.2503511608,-0.4473156035,0.1082950607,0.1303570122,-0.1760199368,-0.1142384484,-0.079467319,0.31345281,-0.1619536132,-0.2900297344,-0.0612089671,-0.1611083597,0.0244895741,-0.3481976688,-0.0135521563,0.2361944914,-0.2912136614,0.0921499655,-0.1741644889,-0.1775014997,-0.0142044174,0.1767190993,0.2195820808,-0.0362453125,0.24150455,0.1687852144,-0.1007038727,-0.0387205444,0.2047067136,-0.1378791183,-0.1729701012,0.0504764169,0.0390650108,-0.227791816,-0.341429621,0.1677935421,0.0461251736,-0.1246552244,-0.3886179924,-0.0228885226,-0.0564539172,0.1423715651,0.1723636538,0.1894030422,-0.016139904,0.0253562219,-0.3533585668,-0.4129772484,0.0379112996,-0.2572102249,0.1801859289,-0.3242875338,-0.3658299744,0.1855311096,0.0937517062,-0.3394749463,0.0716197416,0.0357467718,0.1260523796,0.1996835917,0.1301543415,-0.2337152511,-0.2107922137,0.0029231159,0.1080126166,-0.2767837048,-0.1277999729,-0.0667589158,0.1748126745,0.089380607,-0.0583929121,-0.2286777347,-0.3893969059,-0.3422329724,0.0216786303,0.029594481,-0.0617081299,-0.0297520105,0.2436679602,0.1758689731,0.1475295722,0.0248649176,0.1996092647,-0.0080664931,0.1530717015,0.0412142016,-0.0363901444,-0.0589520968,0.0387086347,0.2955311835,0.1784841418,0.0585312396,-0.0158959478,0.3987704813,-0.0958602056,0.4400488734,0.3519692719,-0.0067887511,0.2975907922,-0.2528961599,0.0626204759,0.4218917191,0.2212225944,-0.073289454,-0.1345282346,0.3153685629,-0.1611553133,-0.0505208671,0.2560867369,0.6267188787,0.0987601653,-0.1649105251,0.0966253802,0.3983308673,-0.0221291073,-0.0325459428,0.3452893198,-0.3686929941,0.0322197899,0.3588172495,0.3029911518,-0.0489354618,0.3099589944,0.2056019604,0.1549178064,0.3644079268,0.2757620215,0.2170075625,-0.3287484646,0.3994604349,0.2665036023,0.2322468907,0.0951228514,0.1313143671,0.0132463612,0.357896477,-0.2666037381,0.0345180556,0.0201504566,-0.1418247223,0.3033528328,0.047088366,-0.2008247674,-0.1047958732,-0.1621444225,-0.0473186411,-0.1860138178,-0.1269052178,0.0250379071,-0.0836069584,0.0085592614,0.1963878572,0.0086578662,0.039320197,-0.1794988662,0.1373905241,-0.0451025888,-0.1183148623,0.266841799,0.0590348952,0.2715172768,0.2061873525,-0.0483386293,0.1316832602,0.106911391,-0.202021867,0.1556459516,0.5027684569,0.1999706477,0.2460289448,0.144171536,0.1107133403,-0.0432855152,0.0483451746,-0.0806608424,-0.2726669908,0.0932090357,-0.1369392425,0.1648145914,-0.3271408379,-0.1242531538,-0.0951970667,-0.2802283466,-0.2964979112,0.755948782,-0.0602639504,0.0511773303,-0.279760778,0.0724504665,0.1928808987,0.3218851686,0.2689842284,0.099474147,-0.0300121624,-0.1658107191,-0.8214002252,0.0253460445,-0.0916960761,0.0373008028,0.1534888744,-0.1416035891,0.2602717578,0.1900973767,0.265081197,-0.1230621859,0.1495441049,0.2551912665,-0.1193157956,-0.1708683372,0.0895573273,0.2353970706,-0.0586904958,-0.0242921002,0.2851044536,0.2299483865,0.0670786351,-0.4307863712,0.2563816011,0.0528068468,0.1266962737,0.4103432,-0.0170377605,0.5939036608,-0.1806780994,0.0358672664,0.0390293896,0.3899903297,0.0560594946,0.3941634297,0.1257208139,0.0965612605,-0.0914364085,-0.0824892446,-0.2665773928,0.1959240139,-0.1153452471,0.0681763962,-0.5511751771,0.002365848,0.1328924745,-0.1198992729,0.1180201843,-0.0771593153,-0.0416470505,-0.0666080415,-0.2249855101,-0.1241889745,0.4870794117,-0.0953745916,-0.1596913636,0.086470075,0.0571621656,0.2501241267,0.1711013615,-0.8708137274,0.00986866,0.4884990752,-0.0964554027,-0.1505250484,0.1542257518,0.0136241913,0.0464439951,-0.146461159,0.1431165934,0.0402833857,-0.0844948813,-0.1621236354,-0.1734789908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2089","title":"Add documentaton for dataset README.md files","comments":"Hi ! There's the tagging app at https:\/\/huggingface.co\/datasets\/tagging\/ that you can use.\r\nIt shows the list of all the tags you can use.\r\n\r\nIt is based on all the tag sets defined in this folder:\r\nhttps:\/\/github.com\/huggingface\/datasets\/tree\/master\/src\/datasets\/utils\/resources","body":"Hi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip","comment_length":36,"text":"Add documentaton for dataset README.md files\nHi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip\nHi ! There's the tagging app at https:\/\/huggingface.co\/datasets\/tagging\/ that you can use.\r\nIt shows the list of all the tags you can use.\r\n\r\nIt is based on all the tag sets defined in this folder:\r\nhttps:\/\/github.com\/huggingface\/datasets\/tree\/master\/src\/datasets\/utils\/resources","embeddings":[-0.286616981,0.0945327431,-0.1588486731,0.1930643469,0.2768509686,0.315613687,0.2661206424,0.1931910217,0.171658054,0.0233468805,-0.1272895783,0.2412937731,-0.1264384389,0.359706521,0.0914905146,0.0267632809,0.0860838294,-0.1028402001,0.1592577845,-0.1845795959,-0.1691451818,0.0174296387,-0.044866249,-0.1641505063,-0.3286385238,-0.1695875823,-0.1266560256,0.0994561017,-0.4146160781,-0.5925245285,-0.0502110906,0.3110641241,0.2041491121,0.1158077419,-0.0001010503,-0.2147537619,0.2052169591,-0.2352301031,-0.013936406,-0.1126761436,-0.1768777668,-0.3279271722,-0.0601527914,-0.1067197546,-0.0597188883,-0.2925773561,-0.0841798484,-0.1904508024,0.1979566365,0.4205386937,0.2686114609,-0.1103500277,0.340637058,-0.2250534147,0.2706674039,0.1677974612,-0.1239818856,0.4374400079,0.4900670052,0.1472700536,-0.0278129727,0.250887543,-0.0373196527,0.1873492599,0.4723614454,-0.0820315927,0.2423785776,-0.5768037438,0.2497014999,0.3870515823,1.039131403,-0.2780817747,-0.2529759705,-0.3692144752,0.0836822018,-0.1449427456,0.1967424452,0.3172575235,0.0246470422,0.1599288136,-0.0215232354,-0.3635124862,-0.2490009964,0.030807713,0.1805025488,0.4041182697,-0.1161129922,-0.1012218967,0.2309455723,-0.1453734934,-0.0280741658,0.1179569662,-0.1853637695,0.1704542786,0.125797376,-0.0564186536,-0.324847132,-0.0684511289,0.2850894034,-0.0259751752,-0.0902743563,0.0960413814,-0.0556011684,0.2861751914,0.1512073129,-0.1017969623,0.1857317537,0.176115647,0.1215500161,-0.1716192961,0.2133102119,-0.1562376469,-0.153006956,-0.0906531587,-0.0275148321,-0.0823393092,-0.1527571231,-0.2826740742,0.0200317372,-0.11338792,0.0969016626,-0.1256388575,0.0243133307,0.2854116261,0.0740623772,-0.2125341147,-0.2671316862,0.1896766275,-0.0061659305,-0.1419605017,-0.0466414765,0.1665447801,-0.1256685704,0.207286194,0.2524135411,-0.1257540733,0.3280339837,0.1352378428,-0.0758281052,-0.1853524148,0.1354757547,0.0137874149,0.2163782716,0.2859224081,-0.0209440663,0.0056272959,-0.0394663066,-0.4077000618,-0.2394081503,-0.0369075313,-0.4110156298,-0.2267906815,-0.3655214608,0.2676398754,-0.124319613,-0.0459319018,-0.2027338445,0.3219774961,0.0812157765,0.0450323857,0.166280359,0.2537733912,-0.3481631279,-0.2576192319,0.0302518811,0.4434049428,-0.2620563209,-0.1156657487,0.0625194162,-0.0872343853,-0.1335525066,0.0185211655,-0.257881403,0.2984237671,-0.2528005838,0.2648495734,0.3774803281,-0.4607772231,-0.1682922393,0.0759339035,0.0210004933,-0.3429535329,0.1542671919,-0.0457942262,0.4741377234,-0.1783985943,-0.1736360937,0.0215551723,-0.1356056333,0.038315203,-0.1707610786,-0.0267440602,0.042084042,0.1477921456,0.0901245028,-0.0927910879,0.3923928142,0.0840129331,-0.0065961652,-0.1246610805,0.1666934341,0.3540766239,0.1807696372,0.1878831238,-0.0094583705,-0.1003722772,-0.4468117654,0.1082814559,0.078050144,0.2980458438,-0.0181669407,-0.3064984083,-0.3663927019,-0.2408527881,-0.3977145255,-0.3615729809,0.1890882999,0.0303558763,-0.1249261945,0.1889827847,-0.1747094691,0.2096574157,-0.0263159256,0.125870347,-0.2828810215,-0.054106608,-0.1897651404,-0.0121795414,0.3533417583,0.0889376253,0.2029140145,-0.0041824896,-0.0547653288,0.3221335709,0.1733797193,0.2728250623,0.2341203839,0.5424048305,0.3943711519,-0.3049960136,0.2806555331,-0.3187286556,0.0982841775,-0.241307959,-0.4253418148,0.1217970327,-0.2108707279,-0.0855719894,-0.087822713,-0.0431888029,0.2023260891,0.0257894844,-0.1415568739,-0.4092408419,-0.1996077299,-0.0929376036,0.1340217888,-0.0142621389,0.0702899694,0.1593573391,0.6423447132,-0.0103725214,0.1053344235,0.1113668531,-0.116008088,0.0188492369,0.2301196009,0.1541118771,0.0834338441,0.2618065774,0.2813137174,0.1955121905,0.0121777561,-0.2820138931,0.1582181454,-0.0850909278,-0.008879411,0.1671547741,-0.2094474286,-0.1137016043,-0.2982536256,-0.0130098648,0.1025971174,0.2986978889,-0.297252208,-0.2102067471,-0.3044349849,-0.1905401647,-0.2461327612,-0.2486096472,-0.5396837592,-0.2466231585,0.0836437494,0.0687134564,0.018367378,0.057609871,-0.1129017919,0.0185980666,-0.2259149253,0.2494397908,-0.1671553552,-0.218542859,-0.1116598696,0.1529919058,0.5170536637,0.4568577409,0.641466558,-0.3583681285,0.2536119521,-0.2470322996,-0.4476350546,0.1278040558,-0.0985560715,0.2760167718,0.0628417954,0.0869512632,-0.0324171707,0.0335605703,0.1297655404,-0.0260875169,-0.1214939356,-0.2069978714,-0.0215773471,-0.2263179123,-0.0622157939,-0.5958678722,-0.211692065,-0.4526795745,0.1733249873,-0.1179511845,0.1648702621,0.1894071698,-0.0055302889,0.2414371222,-0.2219726145,0.2568546832,-0.2677108049,-0.1366797835,0.3523485065,-0.4366490543,-0.3181603551,0.5248668194,0.2030135989,0.407084018,-0.1904558986,-0.2272369862,0.0399037413,0.1460729837,-0.005459764,0.0456531234,0.0450852737,0.2071583271,0.1842418611,0.0126624154,-0.1487964541,-0.1339661032,-0.1094261259,-0.1532999575,-0.0452630222,0.000242649,0.1260112971,-0.0166641101,0.4119056165,0.2007798254,0.1465070993,0.4331977069,-0.0386988483,0.3499621451,0.0074201785,-0.3237253428,0.2413958013,0.0319864005,0.2089139372,0.352170974,0.181169942,0.1111798063,-0.1400909275,-0.1847595125,-0.4298864603,-0.2474628836,0.1573478729,-0.0736720562,0.1495696604,-0.2470838726,0.0599072054,0.066197969,-0.2099087238,0.2634676397,0.7211679816,-0.168671608,0.1718892008,-0.4008209407,0.1314113289,-0.0836732909,0.1011056006,0.1651478261,0.0881719738,-0.3792565167,0.1564689875,0.0412229374,0.1743075997,0.3637605906,-0.6297622323,-0.0566454008,0.1625447571,0.1533397287,0.1671082526,0.0625703111,-0.0757090598,0.059778031,0.1686846465,0.3057149947,-0.3238658905,-0.2382699996,-0.1911197454,0.0012688936,0.0694754794,-0.1967354566,-0.0358059965,-0.2466350049,-0.5505596399,-0.064287737,0.1555037946,-0.1778168231,-0.2148697972,-0.1831828058,0.1584177315,0.1110777482,0.2155773789,0.2287570387,0.1362678558,0.1819660366,0.0507552475,0.3108938336,0.0780816078,0.6399559975,0.4952017367,-0.3066950142,-0.2850310802,-0.0071270037,-0.6396297216,0.3948835731,0.0834927335,0.2394914031,-0.0576920435,0.2203536332,-0.3084156811,-0.185070917,0.0404487476,0.0153223863,0.0767164007,-0.4257958829,-0.5817590952,0.3822163045,0.1445125043,-0.0373900048,0.1783940941,0.0191677324,-0.3615923226,-0.1124398559,0.2835023403,0.9116527438,-0.0347929709,-0.1959257126,-0.254016012,-0.3699360192,0.1958469152,-0.2573854625,-0.184287712,-0.241070658,0.1239646077,-0.0534163192,0.0122654531,0.4137495458,0.3091444373,-0.3899720907,0.1745363474,0.2809738517,0.0706808418,-0.1250815243,0.732436955,-0.3097575307,-0.3987361491,0.0499379896,0.2309393734,-0.0987997502,0.0267546587,-0.0631668121,-0.4066277742,-0.0312035456,0.0307153165,-0.3434127867,-0.2780145407,-0.1524257958,-0.2990489602,-0.1413584352,0.0004013104,0.499779433,0.221560955,0.4800951481,0.1198946685,-0.4500291348,0.0706318542,-0.0343946815,-0.1911528707,-0.1531998366,-0.0326549336,0.4229026735,-0.1708020121,-0.1803760827,-0.0846183375,0.0167172831,0.1134261712,-0.2783231735,-0.0745028406,0.1415951103,-0.3183542788,0.1876696348,-0.2271133363,-0.1257998943,-0.1471267492,0.2240530998,0.1488534212,-0.0457382575,0.0788742676,0.2366682142,-0.0627141818,-0.1444889754,0.1992690861,-0.1947140098,-0.0680141225,-0.0035513244,0.1839550883,-0.2444028109,-0.3652846515,0.0715261772,0.097090438,-0.291097492,-0.2242043465,0.1269226223,-0.0833303034,0.0854663104,0.2186808288,0.1883475929,-0.1238784119,0.0134759219,-0.313144356,-0.3636693358,0.0712501183,-0.1343916953,0.1780825555,-0.3083217144,-0.226096198,0.2204605937,0.1632943749,-0.4225837886,-0.0844783112,-0.0263737217,0.1050240323,0.1314013004,0.0804963708,-0.0760189295,-0.2077512741,0.0321268924,0.1538887471,-0.3212023675,-0.2053708136,-0.1563445777,0.104481563,-0.0265800692,-0.0187580362,-0.1343594491,-0.1624271423,-0.2730727494,-0.1060789451,0.1421529204,-0.0777093619,-0.030681517,0.2276830226,0.2036581933,0.1546230465,0.1125504225,-0.0232956726,0.0342300795,0.2108054757,0.0369202718,-0.1707936078,-0.1542032659,0.0156043293,0.2406664044,0.2459498942,0.1718397886,-0.1532493979,0.3232672811,-0.0717486739,0.3182328343,0.4493522942,0.0756755695,0.2300282568,-0.160261184,0.1481696367,0.3929872215,0.3036416471,-0.1357552856,-0.1440729648,0.2680753171,-0.1421367079,0.0591873117,0.3028164804,0.58684057,0.1136305407,-0.1470515579,0.0602575094,0.3613986373,-0.1302964538,0.102585271,0.2724301219,-0.2673549354,0.0426760204,0.4261931777,0.3482817113,0.0647917017,0.183628872,0.0837780014,0.3153514862,0.2759131193,0.291980803,0.2143126577,-0.2817307115,0.192290172,0.2428547889,0.2343322635,0.1910162121,0.0642859414,0.2352612913,0.2019614875,-0.349503696,-0.0240816604,0.1841568649,-0.1216901466,0.2093120664,-0.0417979285,-0.173892051,-0.1552828401,-0.0585137233,-0.0290588904,-0.3398755193,-0.0005944032,-0.0907354504,-0.1026560515,-0.1430023611,-0.0001046445,0.0046435385,0.1416653395,-0.2170203477,0.1782305837,-0.0363959409,-0.1286474764,0.2489005923,0.1796155274,0.3387133479,0.1064288095,0.0049876468,0.1239872351,-0.0810935795,-0.1823848337,0.0675776079,0.4848442078,0.0636665449,0.1557958126,0.2096106559,0.1593335718,-0.1425123513,-0.0425816327,-0.2261159867,-0.0634315386,0.0810691789,-0.2003158927,0.0650965124,-0.1656087637,-0.2010240108,-0.0708221421,-0.3651632071,-0.1357478946,0.5536826849,-0.056329716,0.0573450066,-0.2745270729,0.1198181883,0.2667573094,0.3728229105,0.1885754913,0.1164147705,-0.030619368,-0.084050715,-0.7492578626,-0.0726459175,-0.0913963392,0.0137569802,0.1357330978,-0.1710389107,0.2055644393,0.2695925832,0.1236689389,0.0102315051,0.1659901738,0.1899448186,-0.1142811477,-0.2107823491,-0.0115236584,0.154395476,0.0576941483,0.0333153196,0.2340153456,0.0524598658,0.0689008534,-0.3286010921,0.2945065498,-0.0065186908,0.0737774596,0.4180580974,-0.0850041136,0.4787151515,-0.1953876466,0.0174061339,-0.0229144953,0.1985984445,-0.1394623965,0.4079094827,0.1480745226,0.2008756548,-0.0928502232,-0.1245314106,-0.2761179805,0.203565374,-0.0565916784,0.0693400204,-0.4167175591,0.0129359728,0.1236258745,-0.1546286792,0.0853092968,0.0508726612,-0.0075763962,-0.1288853586,-0.1927996874,-0.2794760466,0.4735989273,-0.0692359805,-0.115505375,0.121336408,0.0093208868,0.4011226594,0.062269967,-0.9307847619,0.1676558405,0.4140986204,0.0344725735,-0.2256744206,0.1757377237,-0.0129394829,-0.0155297285,-0.1914819777,0.1828604937,0.0711721554,-0.1126632169,-0.0645718127,-0.1252688318]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2089","title":"Add documentaton for dataset README.md files","comments":"@lhoestq is there something like this form Models?","body":"Hi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip","comment_length":8,"text":"Add documentaton for dataset README.md files\nHi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip\n@lhoestq is there something like this form Models?","embeddings":[-0.1892131865,0.1374738365,-0.1079932526,0.2555692792,0.0939081535,0.2768142819,0.3183864653,0.2300292999,0.1645849347,-0.1127747968,0.0745507851,0.0342930071,-0.0814396739,0.2743811905,0.1169215664,-0.1123545915,0.0842320323,-0.039210137,0.1340635121,-0.3693170547,-0.1382398456,-0.1064185202,-0.1125326231,-0.1623550206,-0.1940815598,-0.1530481726,-0.2530029118,0.154425025,-0.4536896646,-0.4092657268,-0.1563839614,0.3343506455,0.2502833307,0.0810746849,-0.0001162479,-0.2053538114,0.104530856,-0.2975952625,0.0215009712,-0.0090972744,-0.2934570014,-0.2886509001,-0.1155399457,-0.2452708185,0.1124637723,-0.0846302658,-0.1111483574,-0.1347021163,0.2573862374,0.480997771,0.1344562024,-0.1449739635,0.4320973456,-0.1261445135,0.4246874452,0.444484055,-0.0951251462,0.2714169025,0.4715499282,0.1207257137,0.1775285751,0.1390412897,-0.1933168024,-0.0740661845,0.4517748952,-0.1032128558,0.6197342873,-0.5987223983,0.0737339109,0.4165807366,1.1477199793,-0.2185470909,-0.1705532223,-0.3784971237,0.0600288361,0.1633538604,0.0665950105,0.3191079199,0.0675154924,0.1344113052,0.1839361787,-0.4062265456,-0.3453059793,0.0127970818,0.2390885949,0.383018285,-0.0634346604,-0.1223107278,0.1149022207,-0.0514288023,0.1035462469,-0.161098659,-0.1212519035,0.2782517076,0.2145283669,-0.1776943952,-0.3989804685,-0.0778853819,0.3203375638,-0.287769556,0.1265299469,-0.0202603005,0.0336569585,0.2777013481,0.2568237484,-0.1489581466,0.0729935691,0.1186129227,0.0216185953,-0.2297594398,0.1992921382,-0.128287673,-0.3023124337,-0.009224412,-0.0877899528,0.0374719054,-0.0169248804,-0.2701617479,0.1349975765,-0.0588489436,0.2212803513,-0.1593736708,0.1877566427,0.1930721104,-0.0958030894,-0.0640480071,-0.2569177747,0.0864461288,0.0827749744,-0.27458933,0.0459688492,0.1602547318,-0.1171032488,0.1420560777,0.1685256511,-0.1099598035,0.1526618451,0.2468887568,-0.0759645104,-0.2444542795,0.3459559977,-0.0210514981,0.2700673342,0.334323585,-0.07821358,-0.0162612647,-0.0593752041,-0.5116698146,-0.253482759,0.1355548352,-0.4304493666,-0.335023433,-0.3629218936,0.1511071771,0.0140583757,0.0566419847,-0.1084751189,0.3653182983,0.2101233453,-0.0216921959,0.1662020981,0.1333694309,-0.180910036,-0.2174251378,-0.065194279,0.528327167,-0.4366075099,0.0707030743,-0.1870044619,-0.0185594968,-0.1382756233,-0.1187208518,-0.191573754,0.4723303914,-0.1588302851,0.2398563921,0.620108068,-0.6063182354,-0.2082887441,0.1673395038,0.1434897929,-0.4222839475,0.2637714744,-0.0061416677,0.49020648,-0.2072938681,-0.3717418909,-0.1517240554,-0.1497493684,-0.1014453694,-0.130609706,0.0407835282,0.065479897,0.1901593059,0.0905476138,0.0413705558,0.4848511517,0.2509725988,0.1322815716,-0.1344645768,0.1000591293,0.2828455567,0.3206346035,-0.0517489351,0.007824285,0.0029191191,-0.3821248114,-0.014451433,0.005937933,0.1456044316,0.2763911784,-0.2670413554,-0.2504827976,-0.1905211955,-0.2876441181,-0.209469676,0.1196377426,-0.0566184558,-0.346424371,0.0829881504,-0.2414966971,0.1945409924,-0.0041330168,0.0113494229,-0.2184317708,-0.1211530194,-0.1695393473,0.1360557824,0.2451477349,0.1067260206,0.325971365,0.119772397,0.0753588751,0.1262142658,0.1661564112,0.0316691846,0.1331104636,0.5415959954,0.3843406737,-0.0765710697,0.3589949608,-0.4546002746,0.024616722,-0.2564453781,-0.345913887,0.1344826221,-0.0737737641,-0.1118151024,-0.0603275336,-0.0118477307,0.1058455929,-0.0115277329,-0.1039805412,-0.4485040903,-0.301876843,0.0742636919,0.1224374399,-0.0419310182,0.0600713454,0.2685750723,0.6090638041,-0.1046281159,0.2115733176,0.0923780799,-0.0723237395,0.0395944193,0.136946395,0.2114700824,0.1702990383,0.1273440421,0.183841452,0.1588413566,0.0158208311,-0.1585069597,0.2106633186,-0.1411124319,-0.0764195174,0.2258550674,-0.4097333252,-0.0364375338,-0.2152242959,-0.1368401498,0.129925102,0.2222624123,-0.2828366756,-0.2946797013,-0.0637263581,0.1476428956,-0.1038571298,-0.1623350531,-0.4165496528,-0.0012929362,0.128599748,0.0874966532,0.0100773619,-0.1578063369,-0.3078169525,0.3782791197,-0.1311701983,0.2513727248,-0.2256935537,-0.0745098367,-0.1446803212,0.0524628237,0.4915921688,0.533593297,0.4040523469,-0.2889174521,0.3322999477,-0.1517459303,-0.5205305815,0.1615602523,-0.0161453094,0.3424582779,0.0173859522,0.069673501,0.0164823513,0.0350993574,-0.0295709111,0.2510409057,-0.2386514992,-0.189370349,-0.0319822282,-0.3734861314,-0.0677445531,-0.4019959271,-0.2503369153,-0.292488426,0.1226796657,-0.0684964284,0.2726713121,0.1564430147,0.0779757649,0.2437521219,-0.1809014231,0.3198706508,-0.1898368299,-0.1240767315,0.5545032024,-0.3386681974,-0.3829134703,0.4119906723,0.0163215436,0.280919224,-0.1904956847,-0.2899836004,0.1454818994,0.1913117319,-0.2888531387,-0.0414490849,-0.0388706811,0.3066635132,0.2685942948,0.1033292338,-0.2274049819,-0.1176447272,-0.0104333498,0.0571026616,0.1228503212,-0.0744746923,0.1280175745,-0.1049505025,0.2513609529,0.1206094325,0.1662162244,0.2881846428,0.1978102028,0.293002367,0.1822907031,-0.3342968524,0.1902850866,-0.0093318904,0.0836249068,0.2529734671,0.114382714,0.0640440732,-0.1729288399,-0.3472167552,-0.2775484025,-0.2363649309,0.1032924876,-0.0525021292,0.2877365351,-0.2052501291,-0.0651847348,-0.0966551378,-0.1021853685,0.2307762504,0.9404412508,-0.1174850985,0.0731686205,-0.393658936,0.2091520429,-0.0122997053,0.0485011302,0.0922839642,0.1110663489,-0.3337133527,0.1863428354,-0.0441672131,0.1886458546,0.3490284383,-0.6521158218,-0.0897436813,0.0828607976,0.1362740695,0.2682285607,0.0315503925,-0.1097652912,-0.0405021161,0.1684574485,0.3733771145,-0.2985587716,-0.2474035323,-0.1632599384,-0.1097335592,-0.0521878004,-0.3385280669,-0.0451560356,-0.2664290071,-0.3336053193,-0.0784752443,0.2424217016,-0.407112509,-0.2606169581,-0.29098472,0.0911969021,0.1602161378,0.3362230062,0.1826062053,0.1097419411,0.1648151129,-0.0071508288,0.3154138625,0.1494585425,0.7110338807,0.292568773,-0.2428147644,-0.1017557606,0.1301749051,-0.8477705717,0.4049889445,0.1382693648,0.2497065067,0.0050390181,0.3976887763,-0.3832319975,-0.0289140902,0.1183266267,0.076665096,-0.1123534143,-0.5377398133,-0.7101687193,0.2296769619,0.0409443006,-0.1524796635,0.0502424054,-0.0770164952,-0.3576981723,-0.0347085595,0.4671364725,1.1773312092,0.0424181595,-0.1263383329,-0.4454863667,-0.3501135111,-0.0512799546,-0.3390200734,-0.2591575086,-0.339568764,0.0816607922,-0.1593977213,-0.0992532894,0.4435895085,0.5010032058,-0.4616156816,-0.1055489779,0.298423022,0.1032807156,0.0218182486,0.6970396638,-0.2272858918,-0.2275352031,0.1736018956,0.0805965289,0.1102271155,-0.1874934137,-0.0733901635,-0.5448701382,-0.0055986596,-0.0423444323,-0.2838095427,-0.3230549097,-0.1700941771,-0.3540831208,-0.1020520777,-0.091778852,0.6814969778,0.3135823309,0.5261582136,0.2729029655,-0.477471143,0.0472006127,0.0082075447,-0.1538494974,-0.1426047534,-0.1301736981,0.4568017125,-0.1952655613,-0.2153355479,-0.0762903616,-0.1845936626,0.064525485,-0.481535852,0.0050986712,0.2115015984,-0.3218396902,0.0893787667,-0.2040638328,-0.1909985691,-0.0359860361,0.1310126483,0.1233632416,0.1034432277,0.2031531185,0.2207787931,-0.0469207577,0.0318007581,0.0997029692,-0.1222554967,-0.0988921747,0.1103744283,-0.0616487302,-0.2505646348,-0.3356704414,0.1078935936,0.0551571585,-0.1956308335,-0.3082476258,-0.0531093515,-0.1542869806,0.0042357733,0.1490458697,0.2780423462,-0.1424552649,0.0176719651,-0.3562105894,-0.4555741847,-0.0057257093,-0.265987128,0.1324766576,-0.2660617828,-0.349375844,0.1374839842,0.1090160757,-0.2898060977,-0.0864036158,0.1256826371,0.0740175992,0.165507406,0.1552730203,-0.3110131025,-0.2205822468,-0.0640473068,0.0596888326,-0.1786031574,-0.1073741168,-0.2482940108,0.2074777037,0.1285375208,-0.0402480289,-0.1777065694,-0.2915664911,-0.3184789419,0.0438254066,0.0323709436,-0.1660754532,0.075507313,0.2733933628,0.1611204445,0.1913261265,-0.0621933639,0.1040566638,0.0302613694,0.1389533281,0.1188666672,-0.0847125128,-0.1443087608,0.0103974603,0.3016714752,0.2901953757,0.2415293455,-0.0280307867,0.4383666813,-0.0147835687,0.3985525966,0.3143335581,-0.0970232785,0.2661940455,-0.1834867299,0.1592306644,0.3743915558,0.1858091354,-0.1445150971,-0.1342585236,0.3412231803,-0.223821193,-0.032004647,0.2655311525,0.5713142157,0.1137654856,-0.0510605462,0.0669437796,0.2519155145,0.1476421952,-0.08528696,0.3896257877,-0.3139173687,0.080391489,0.4374387264,0.3345302641,-0.0554722771,0.1996976733,0.2611010373,-0.0533404164,0.3104866743,0.2606636286,0.2740537822,-0.2492319345,0.409037143,0.2496211529,0.3365511894,0.1795206666,0.1450453848,0.0084850024,0.5307109356,-0.2408809662,0.013588788,0.0021371047,-0.1060406119,0.1992889643,-0.1569819003,-0.2004438937,-0.0948796943,-0.1669911891,-0.0156488884,-0.1731768847,-0.2392129302,0.0177590009,0.0606973208,0.0695950314,0.1414793432,0.1694580764,-0.018348109,-0.2216408998,0.084803924,0.1480597556,-0.0744934082,0.2579431534,0.0572520196,0.1771721244,0.22623083,-0.0924839228,0.10882359,0.1265794635,-0.2244118452,0.1188046485,0.4267133176,0.2664747834,0.2706908584,0.1474289596,0.0493008159,0.0167000648,0.0729112774,-0.1955402195,-0.2403552085,0.0328126848,-0.1179178283,0.2668775916,-0.3217728436,-0.1445289403,-0.0922431499,-0.3296331167,-0.263850987,0.9350246787,-0.044583112,0.0843879357,-0.2251927406,0.0421466865,0.2806237042,0.302755028,0.1570699513,0.0038558512,-0.0181399249,0.0368590765,-0.6634763479,-0.024631504,-0.0204875655,0.0163948368,0.2265131772,-0.2622849941,0.2710732818,0.1750900745,0.3165645003,-0.2178888023,-0.0167963114,0.3330146968,-0.082403332,-0.1596010625,-0.0407243036,0.3839495778,-0.005365165,-0.0299368203,0.2548992336,0.3557260931,0.0187958758,-0.5260224938,0.2502631545,0.0366817564,0.200324297,0.2919612527,-0.0277544055,0.6241118312,-0.2833373547,0.0015269675,0.0965967551,0.502628386,-0.0172506478,0.2892138362,0.1545428038,0.0092282994,-0.1154234335,-0.1820160747,-0.2417771965,0.2354895622,-0.0439478941,-0.1459023803,-0.6238760352,0.0087131467,0.1090922281,-0.1535331458,0.1467058063,-0.0579249002,-0.0638544858,-0.0427451804,-0.1927126497,-0.0368695445,0.4777636528,-0.1020268425,-0.1507814974,0.1574825197,0.0362437367,0.2500579357,0.1982196867,-0.8684783578,-0.0873547047,0.4482676387,-0.1863953769,-0.130280897,0.1085750908,0.1340779215,0.0672273859,-0.1632074565,0.184337005,0.0208898839,-0.1881575137,-0.3692519367,-0.204114005]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2089","title":"Add documentaton for dataset README.md files","comments":"I don't think so. Feel free to take a look at the tags of other models (example [here](https:\/\/huggingface.co\/bert-base-uncased\/blob\/main\/README.md)). But we should definitely have some docs or an app to write the tags. Feel free to open an issue in the `transformers` repo or in the `huggingface_hub` repo so we can discuss this","body":"Hi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip","comment_length":52,"text":"Add documentaton for dataset README.md files\nHi,\r\nthe dataset README files have special headers.\r\nSomehow a documenation of the allowed values and tags is missing.\r\nCould you add that?\r\n\r\nJust to give some concrete questions that should be answered imo:\r\n- which values can be passted to multilinguality?\r\n- what should be passed to language_creators?\r\n- which values should licenses have? What do I say when it is a custom license? Should I add a link?\r\n- how should I choose size_categories ? What are valid ranges?\r\n- what are valid task_categories?\r\n\r\nThanks\r\nPhilip\nI don't think so. Feel free to take a look at the tags of other models (example [here](https:\/\/huggingface.co\/bert-base-uncased\/blob\/main\/README.md)). But we should definitely have some docs or an app to write the tags. Feel free to open an issue in the `transformers` repo or in the `huggingface_hub` repo so we can discuss this","embeddings":[-0.1301450878,-0.0352435075,-0.0532184169,0.2471045256,0.2942425311,0.2559415102,0.4152002931,0.1293420494,0.0928785503,-0.0484237932,-0.1729551107,0.0700884908,-0.1718711555,0.3322659135,0.1102287322,-0.0799867585,0.0745821446,-0.064012073,0.0829297826,-0.3425408304,-0.1205999777,0.0473311469,0.0086057996,-0.1590167731,-0.3415005803,-0.0647564381,-0.1081090271,0.087882258,-0.3661057651,-0.4418179095,0.0343780592,0.36452806,0.2753328681,0.2000294924,-0.0001122496,-0.0820678696,0.201798141,-0.2937220037,0.0805850103,-0.0695166364,-0.1306844801,-0.2235603034,0.0002540752,-0.1659983844,-0.0443721972,-0.1104560047,-0.0838512406,0.0729084164,0.4370497465,0.37518695,0.148133561,0.1568712592,0.3579421341,-0.0753630251,0.3906325996,0.5329559445,-0.1721142977,0.3704756498,0.4002628922,0.1344771534,-0.0086024338,0.3429555893,0.0102651846,0.0815118849,0.5320785642,0.0614764616,0.4577286839,-0.4999290407,0.0746051297,0.4968686402,0.8872467279,-0.2778629363,-0.3920515776,-0.4509535432,0.0511733517,-0.0022098804,0.1275162399,0.2292139977,-0.0223254487,0.2023370564,-0.0359625295,-0.3970157206,-0.2950106859,-0.0164486282,0.2132585198,0.5303966999,-0.0830988958,-0.0483889282,0.2028928697,-0.0750717819,0.0379065275,-0.0209300052,-0.133031413,0.180884257,0.18294397,-0.2720577121,-0.4023817778,-0.0091049857,0.2835808098,-0.1288718879,-0.0929570571,0.0104748737,-0.0570272692,0.2156037092,0.2461791337,-0.0706653595,-0.1041215435,0.2844152153,0.1489333659,-0.0632366166,0.350486666,-0.1537154317,-0.1432397068,0.0320519693,-0.1625291705,-0.0585426688,-0.1197234988,-0.2173473984,0.1657661945,-0.1894990802,0.2245896459,-0.1191413254,0.0865387097,0.3314216733,0.0674898252,-0.1350482404,-0.2403621227,0.0849950463,0.0512669496,-0.1340712458,-0.0295302644,0.0856456608,-0.1304166615,0.2566341758,0.2308344841,-0.045377247,0.2571469545,0.0018116845,0.1141917408,-0.0951160043,0.0790747926,0.0553013124,0.1696841568,0.3754540682,-0.219862327,0.0422180071,-0.0593294054,-0.3838603199,-0.2995629907,0.0680775791,-0.3602797985,-0.2845461965,-0.3764235675,0.1332672536,-0.1712684035,0.0672326908,-0.2517314553,0.292639941,0.1982022375,-0.0982505456,0.2711409032,0.2577008009,-0.2646192312,-0.1525433958,0.0543294661,0.5442987084,-0.1550184637,-0.2414232641,-0.0121600637,-0.0153322872,-0.2538617551,0.1582325399,-0.3104732335,0.288032949,-0.2125798464,0.1954707801,0.3075235188,-0.5969015956,-0.2878676057,0.1899215728,0.04410135,-0.243017897,0.2752088904,-0.0098639047,0.4363812506,-0.2221133113,-0.2216964513,-0.1164819226,-0.1454751641,0.0300425999,-0.1788652986,-0.1824240685,0.180166468,0.1770681292,0.1427143812,-0.0056527299,0.2813097835,0.3177366853,0.1807597876,-0.0894367993,0.1050366908,0.3970932066,0.3476620018,0.0344510376,0.0576082245,-0.0956659093,-0.6226606965,0.0369949825,-0.0407796726,0.3914708495,-0.0753991529,-0.3238521218,-0.264056772,-0.2537076473,-0.3115502596,-0.4422731102,0.0766931549,-0.0490222834,-0.2323368639,0.1121200547,-0.2760218084,0.3250449002,0.084157981,0.176702261,-0.4231535792,0.0879205242,-0.3023082614,0.0981676504,0.2412701547,0.1098993495,0.127358675,-0.0951983631,0.0094879605,0.2334367931,-0.0210516434,0.2175308615,0.0982459635,0.6624143124,0.5612775087,-0.3053041995,0.1880624741,-0.2835327983,0.0984657779,-0.1810097694,-0.5105965734,0.0683180094,-0.1871127784,-0.057043232,-0.0027717671,-0.1004650444,0.1907692105,0.0320719443,-0.2853167951,-0.3556480706,-0.2345433235,0.0964967459,0.23932302,0.0928759575,0.1901282668,0.0627860054,0.750420332,-0.1258005053,0.0917745605,0.1957152486,-0.2276316434,-0.0212666411,0.1258779466,0.122194767,0.1449053735,0.2134449035,0.0604290143,0.1615928411,0.0986356139,-0.1737308502,0.1242914572,0.0372124575,-0.0700257421,0.242862463,-0.2385572791,0.011652953,-0.2925586998,-0.0238551814,0.1365642846,0.2298060954,-0.3926616013,-0.168823719,0.0505814403,0.0385569893,-0.2868267894,-0.2530362606,-0.6583180428,-0.1110902429,0.0961049572,0.0289084744,0.0561192855,0.0624089018,0.0627806485,0.1970192492,-0.1762809157,0.2620811462,-0.3391105831,-0.2260823995,-0.0866138414,0.0383844711,0.4431889653,0.3795891702,0.3404743075,-0.3237490654,0.2406792492,-0.1147800907,-0.6005818248,0.2257779539,-0.1384498924,0.3139522076,0.1184751168,0.0575810559,0.1074348912,-0.1111849323,0.1358054876,-0.0334656984,-0.1298608184,-0.2549876869,-0.0332795344,-0.2619003356,-0.1209807396,-0.4445782006,-0.2484944314,-0.3826474845,0.3122243583,-0.1325089335,0.0555029251,0.307759285,-0.0077625881,0.1976738274,-0.2474431694,0.1154809222,-0.1699935347,-0.090065755,0.4744247496,-0.3464311361,-0.2834531367,0.2812028229,0.0528321378,0.25548172,-0.2843570411,-0.2434411198,0.0428086631,0.093488425,-0.2509506941,0.0276176929,0.0766631067,0.2747790217,0.2590764165,0.1170005798,-0.1130938604,-0.1745864153,-0.0401857384,0.0621444434,-0.0020896541,-0.104607448,0.1465646178,-0.0757131651,0.3623982966,0.0772003904,0.0615587085,0.33162269,0.0571857095,0.4331573546,0.0660251081,-0.3546023369,0.2062720358,0.0067272028,0.0800777972,0.3720385134,0.0802140608,0.1718423516,-0.125184983,-0.312266469,-0.3093342483,-0.3860101402,0.1021918133,0.2228010446,0.0934711173,-0.1705839038,0.0037210207,-0.0533825122,-0.1663524508,0.354372263,0.8319618106,-0.1410356164,0.0883160233,-0.2800044417,0.1888351738,-0.2273827195,0.0215548407,0.1896581948,0.0126957176,-0.2334344983,0.2113342434,-0.0076567037,0.156587407,0.5497598052,-0.5247405767,-0.0631487817,0.0129602058,0.0176207051,0.1108446047,0.0663734525,-0.1372967064,-0.0921607167,0.1316772401,0.494962275,-0.4311374128,-0.2816446424,-0.084401086,-0.076588206,-0.0694201812,-0.2530347705,-0.0952584669,-0.1593744904,-0.5372489095,-0.127892226,0.2674381435,-0.2953924239,-0.1729172617,-0.2430229783,0.1832462251,0.0455965102,0.2772335112,0.2463165224,0.1555468142,0.1407820731,-0.0212297756,0.3834705353,0.0792544633,0.6136366725,0.4085799158,-0.3780501485,-0.324305594,0.0322221555,-0.7105805874,0.2947055101,0.2757335305,0.2516627312,-0.0687002018,0.4258865118,-0.2119211853,-0.1852735579,0.0589732975,0.115647614,0.019267736,-0.5070009828,-0.5510529876,0.4138425887,-0.0060827215,-0.0376355425,0.0127169024,0.2641605139,-0.3906282485,0.1536782086,0.3243082762,1.2151699066,-0.1024533063,0.056402266,-0.3050798774,-0.2416739911,0.3678264618,-0.4156566858,-0.12343207,-0.2671492398,0.0144282449,-0.1264306158,0.0145882685,0.3629639447,0.3674153984,-0.380130738,0.0606568642,0.342132926,0.2065848261,-0.1084157228,0.7030028701,-0.1309195012,-0.4052950442,0.0217955001,0.076462388,0.0294231009,0.1367142648,-0.0836408213,-0.5859959722,-0.0780554265,-0.1361371726,-0.3741081059,-0.3525400162,-0.2044964731,-0.2110218108,0.0239239316,0.0246664733,0.5534372926,0.2366509885,0.5483725667,0.1621498317,-0.6404772997,0.1355976611,-0.123923555,-0.1956999302,-0.2247792482,0.0478748493,0.3694460988,-0.1299369931,-0.2023034096,-0.1330410689,-0.0682686344,-0.0005881621,-0.2465473562,-0.1496175081,0.1054779887,-0.3484675884,0.1442780793,-0.1986942291,-0.2411316931,-0.1479476541,0.1371401697,0.0323929861,-0.115126729,0.0351002887,0.2759845853,-0.1552312225,-0.0878978595,0.2337965816,-0.1964874864,-0.0990456268,0.105518274,0.1004095301,-0.2399874628,-0.2959397733,-0.0156736784,0.2685163021,-0.6028063297,-0.2454488128,0.0627225265,-0.1808457375,0.0122802444,0.3699559569,0.3838859499,-0.1363750994,-0.0521357283,-0.2257379144,-0.5842853189,0.1538548023,-0.1406340003,0.169954747,-0.1889738292,-0.1157269403,0.1967945695,0.2167566419,-0.3300186992,-0.139452368,-0.0166445244,0.0411742292,0.1615885347,-0.0045227562,0.0675561428,-0.3016514778,-0.0523988195,0.2579047084,-0.2703534961,-0.0999564752,-0.2450138032,0.1775098145,0.0253216848,0.0142252864,-0.2025500089,-0.023404751,-0.2965800166,-0.127816245,0.2104473412,0.041906748,0.1458959579,0.2594268024,0.2857012153,0.1632176936,-0.1503937989,0.0279436409,0.128509596,0.1703544855,0.1481310874,-0.1565219611,-0.1611434817,0.1048820391,0.3086868823,0.2369429022,0.2171382457,-0.0049652993,0.3991962969,-0.0075297807,0.2695425749,0.34346506,0.1190741435,0.1901553124,-0.129544124,0.1385047436,0.3368333876,0.1772239059,-0.2271178812,-0.1055096239,0.3189112544,-0.1777693629,0.102086857,0.4221236706,0.57029742,0.2688582242,-0.0172718819,0.1258379966,0.4669812024,0.0257781968,0.0944088697,0.4013044238,-0.2711415589,0.1281040758,0.4452599585,0.3683162332,0.1730044037,0.1913362145,0.0918563753,0.1147198975,0.0034723324,0.2881265581,0.3339245915,-0.155417949,0.1877932549,0.1563862264,0.4228897393,0.2024159729,0.0772230327,0.0768751055,0.3115520477,-0.2270364016,-0.1206152439,0.1191147566,-0.0648452193,0.066463463,0.2424653918,-0.1248513386,-0.2164675146,-0.1007560268,-0.0400179476,-0.3261401951,-0.0629059374,0.0003625495,-0.0012205328,0.0954758301,-0.0573175959,0.0682996809,0.1424608678,-0.2303170562,0.1369354725,0.1743843853,-0.1468616575,0.1629304737,0.2221173942,0.2091760933,0.0719911382,-0.0252238326,0.2093201578,0.0419278629,-0.1346253008,0.0898773223,0.4182916284,0.1899677068,0.2705292106,0.0956108794,0.0613956936,-0.114889428,-0.0388118811,-0.2847309113,-0.0863256007,-0.0099290954,0.0628956184,-0.0068667219,-0.2412634194,-0.2367473245,-0.1236436367,-0.4529113173,-0.0892211869,0.5199387074,-0.143440634,0.0544258431,-0.181055516,0.0502703227,0.2086042613,0.4289119244,0.2435464412,0.0278237741,-0.0888556466,0.0625941306,-0.7101934552,-0.0124144843,-0.0191065855,0.1245448589,0.1892187744,-0.214103654,0.2409208268,0.1618461609,0.1212539524,-0.0745744184,-0.0671567917,0.0800964534,-0.0710477084,-0.1740487665,-0.2068470567,0.1768306345,0.1298550963,0.0131967431,0.1974115223,0.3441025019,-0.0479952358,-0.4523759186,0.3501650095,-0.0330708772,0.0621090271,0.314914465,-0.0859730542,0.4451007545,-0.2029885352,-0.1355413049,0.0671531186,0.2498446405,-0.1163872555,0.3499435186,0.1667208523,0.246656239,-0.2440741211,-0.2746078372,-0.4681322277,0.0580258444,-0.0034032001,-0.1327598989,-0.6510510445,0.0006269382,0.0464831851,-0.227935344,0.2111953199,0.0542859361,-0.0371281691,-0.118222937,-0.188624531,-0.1355215013,0.621186316,-0.1726373434,-0.1415697932,0.1333980262,-0.0454601347,0.4083578885,-0.0170500278,-1.161651969,0.1322577149,0.3177457154,0.0258129146,-0.2481320202,0.0275044665,0.0788401067,-0.0259972308,-0.2317184508,0.2802074254,0.0976850688,-0.1532402039,-0.1820608228,-0.1760691255]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2084","title":"CUAD - Contract Understanding Atticus Dataset","comments":"+1 on this request","body":"## Adding a Dataset\r\n- **Name:** CUAD - Contract Understanding Atticus Dataset\r\n- **Description:** As one of the only large, specialized NLP benchmarks annotated by experts, CUAD can serve as a challenging research benchmark for the broader NLP community.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2103.06268\r\n- **Data:** https:\/\/github.com\/TheAtticusProject\/cuad\/\r\n- **Motivation:** good domain specific datasets are valuable\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":4,"text":"CUAD - Contract Understanding Atticus Dataset\n## Adding a Dataset\r\n- **Name:** CUAD - Contract Understanding Atticus Dataset\r\n- **Description:** As one of the only large, specialized NLP benchmarks annotated by experts, CUAD can serve as a challenging research benchmark for the broader NLP community.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2103.06268\r\n- **Data:** https:\/\/github.com\/TheAtticusProject\/cuad\/\r\n- **Motivation:** good domain specific datasets are valuable\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n+1 on this request","embeddings":[-0.4965970218,0.0208293721,-0.233336553,-0.0203883424,-0.0079529099,-0.1036102176,0.2067573369,0.0460304059,0.2794986963,0.1158508807,-0.2583628297,-0.0390605703,0.1218197569,0.1817899495,0.3810776174,0.0720431805,0.0444419086,0.152594775,0.2333089262,-0.1025141478,-0.1309536844,0.0490208864,-0.028234804,-0.152535677,-0.1820746511,-0.3704994023,-0.1978627145,0.0686235577,-0.0153591279,-0.3936802745,0.0823517963,0.34418571,-0.0151596498,0.3367461264,-0.0001025624,-0.2175514251,0.0373849533,0.0170588996,-0.3063836992,0.0844527036,-0.2846211195,-0.2644030452,-0.0957518965,-0.3989253342,-0.1970429569,0.2674052119,-0.0870608613,-0.3221615255,0.2796079218,0.1935322583,0.320810914,-0.1603705883,-0.083312504,-0.1926111877,-0.1051480398,-0.1137779802,-0.2263291627,0.2289686203,0.5243505239,-0.0986055806,-0.3384030759,-0.0406478941,-0.0952830762,-0.0960704759,0.0551988147,-0.2548935711,0.1724293083,-0.0490789786,0.1143641919,0.3872315884,0.3384855092,-0.1673197001,-0.2578010261,-0.0772863328,0.0608056895,-0.412251085,-0.178907752,0.2539698482,0.0176335461,0.2115363479,-0.1582386643,0.0497885607,-0.1146902293,0.1934850812,-0.2328156382,0.2575761378,-0.1937003136,-0.1638332456,0.0961102769,-0.1140481234,-0.3636328578,-0.0152420728,0.1110479608,0.0167167168,-0.587518394,-0.3990604877,0.2299249023,0.1615538895,0.2506324947,0.3154518604,0.4877947867,0.0443213694,0.0977036282,0.145362258,0.1340793073,0.2190856189,0.0826046094,-0.0943706781,0.2557077706,0.2319061756,-0.1853383929,0.259329319,-0.2509947121,0.3481095433,-0.0941377878,-0.0742857456,-0.1638440639,-0.4229709506,-0.2869313359,0.2655291557,-0.1743927598,0.0275757276,0.0214117207,0.1547445655,0.0301263779,-0.2232960314,-0.2816762328,0.2731695175,-0.1623442918,0.166756466,-0.2059942037,0.0145516628,-0.2383691221,-0.0727324113,0.3681968451,0.0047595114,0.3417093456,-0.1967728287,0.1183116436,0.0101332869,0.1984405071,-0.1521917135,0.0943790525,0.0183884595,-0.1044948474,-0.1930255592,-0.0839966089,-0.3935000896,-0.3311685026,-0.1361510009,0.1594425738,-0.3162936568,-0.1059883013,0.2689113617,-0.0911247954,-0.3008293808,0.069040142,0.5064145327,-0.1428143084,0.1838221252,0.1068612263,0.1213718876,-0.3899358511,-0.1481891274,0.0976841748,0.1540111899,-0.2896846831,-0.0485208929,-0.1411653012,0.0868079439,-0.1891569495,0.2749276161,-0.0404150039,-0.2260694504,-0.0135437986,0.2638514042,-0.1422206163,-0.5083970428,-0.3960338533,0.1116964594,0.0600361936,-0.0768420547,0.1688004136,0.1562254727,0.688873589,-0.2662532628,0.1261958927,0.4661051035,0.1391987205,-0.1743205637,-0.0554956719,-0.1299219131,-0.0520174503,0.1393163651,-0.2577243447,-0.133489579,0.172987029,-0.1817975938,0.0974307358,0.107453391,-0.0952691287,0.1773878485,0.2405951321,-0.3796293139,0.033798378,-0.1026724726,-0.2905294597,-0.0584224537,0.0958489552,0.3401944339,0.1605198234,-0.2456324399,-0.3762195408,0.0719162077,0.1529633403,-0.2499807179,0.2334551811,-0.1043168902,0.1421578526,0.0711632445,-0.1824686676,0.0807271451,-0.2538830042,-0.0176553112,0.0509895161,0.0303018708,0.0639697835,-0.0033100315,0.2119340003,0.6679909229,-0.0994321257,0.0891521275,0.0498088114,0.0483433083,-0.1479644179,0.0045686895,0.2002121359,0.7751000524,0.1707381606,-0.3066225946,0.2015017122,0.0065234392,0.1033849269,-0.1012723967,-0.3287756741,0.4182461798,0.5771569014,0.2326866239,0.2398910224,0.110016942,0.3413057625,-0.2893384695,-0.0392140076,-0.0146073652,0.3038915992,0.3841083944,0.8453088403,0.0083623659,-0.1949848086,0.38531968,-0.0242069401,-0.2450220734,0.2931092083,0.0676716641,-0.2532572746,-0.1045024246,0.1996695548,0.0630501583,-0.0576733239,0.3876649737,0.2304320931,-0.1353090852,-0.2635174692,-0.1688485891,0.1936854124,0.0611378402,-0.2581332922,-0.0337053537,0.210427165,-0.0122613749,-0.4401892126,-0.3548946381,-0.2888381183,0.0985448733,-0.1464900821,0.1161352545,-0.0728080571,-0.3098972142,-0.0432824492,-0.0132963536,-0.0420291945,-0.2631287575,0.1121718585,-0.059249796,-0.034071371,0.231620729,-0.0845515281,0.611839056,-0.2055685222,0.0045932336,0.0211991575,-0.527097106,-0.0432462357,0.3168726563,0.2018110454,-0.0169826187,0.4022506773,0.0400504619,0.30664891,0.0421002284,-0.5469748378,0.0492531843,-0.2601768076,0.2111669928,-0.1475678384,-0.0729002953,-0.210161671,-0.1000057682,-0.2995276153,-0.1503613889,-0.1890665889,-0.1230539382,-0.2070941329,-0.2706622779,-0.0474488363,-0.3372122645,-0.0489038713,-0.5605352521,0.2238143981,0.1123862416,0.0231167879,0.2178563029,-0.1667850763,0.2833046019,0.1303727627,0.177052334,-0.2421730757,-0.1803266853,0.2087550461,-0.2781074941,-0.1376707107,-0.2201994509,-0.3597940505,0.273322314,0.0730271339,-0.3149775863,0.1399327815,-0.1516210884,0.2150787115,0.1382821649,-0.0747028664,0.0811987668,-0.1877327114,-0.2577446699,0.0090414416,-0.501809299,0.2261540592,0.1440347284,-0.0085462406,-0.1970181167,0.1155687273,-0.0757733434,0.4231408536,0.0007139428,-0.1347273141,0.2329782844,-0.0972239673,0.1510143578,-0.0412761904,-0.1598037779,0.3136182427,0.0132986801,-0.0105055626,0.3404256105,0.2334908694,-0.4819126725,0.112294361,-0.1025752649,-0.2488517165,-0.0649173483,0.0300538968,-0.1400034428,0.2820761204,-0.069071658,-0.2081816941,-0.0394437872,-0.1912290752,-0.0270015094,0.0006106214,-0.1367837191,-0.1532902718,-0.0765879676,0.4278486371,-0.4327189624,0.3038777709,-0.0161036011,-0.0472565517,-0.2945230007,-0.0649260804,-0.0074583227,0.1585816741,0.472067982,-0.0051484853,0.0166152641,0.3959294558,-0.0646888018,-0.1273967922,0.2851311564,-0.3942846358,-0.0500934646,0.2418890148,0.1992907375,-0.3441027403,0.1076029763,0.2004610151,0.097461693,0.0193019304,-0.0783169195,0.0062066838,-0.0893583447,-0.2224637568,0.4236643016,-0.0110329203,0.0313376635,0.0688762814,-0.0707203001,-0.1144827679,-0.2381530106,0.4025571942,0.2241166383,0.0179573875,0.1208280027,0.2021340728,-0.098494187,-0.0407623798,-0.3470959663,0.5185818076,0.0415944755,-0.0223724749,0.1024168655,-0.190326944,0.395993799,0.2308735698,0.2234707326,0.1690370888,-0.0959916264,-0.0607069395,-0.2706047893,0.0493802764,0.0179150328,0.332816869,-0.4567455947,-0.2868648469,0.471736908,0.0382879153,-0.2158206254,-0.0507319681,0.3543282449,-0.0378752761,0.3463664651,0.0097628077,1.0340200663,-0.0429067239,0.2087913305,0.3699320853,-0.3179438412,0.1007053852,-0.0814070255,-0.1506948173,-0.127883628,0.0218858216,-0.039237693,-0.0787706524,0.0120658977,0.4288262725,-0.2462495863,0.1256568879,-0.1203297451,-0.179524675,0.0721882954,0.1346822977,0.0633995757,-0.2286005467,-0.3506273031,0.2896688581,0.0141738867,0.0404161215,0.0652424097,0.1398962289,-0.0774607956,0.0878294334,-0.3923447728,-0.0716333836,0.0552491397,0.1202135012,0.1555299759,-0.5176730752,0.2494943887,-0.4378528595,0.2725130618,0.1153015792,-0.2204663604,0.01456396,0.0042025549,0.0086402465,0.4955891371,-0.2128269523,0.3186938167,-0.0711279362,-0.0870092884,0.2078814656,0.223538667,0.0212741084,-0.0532361381,0.2049522549,-0.0147911124,-0.2307043672,0.015899267,0.3315166831,-0.1421822459,-0.2106697112,0.2217265368,0.1603980362,0.1821352839,0.1169832274,0.0655662939,-0.2044888586,-0.2239236832,0.0858246237,0.2255832404,-0.1220096946,0.2170660645,-0.1656143516,-0.1498422176,-0.4178696275,-0.3233384192,0.0372756794,-0.2974383235,0.1992124468,-0.0163675472,-0.421389848,-0.0740691945,0.0343650207,0.1217721179,-0.1057855859,-0.074029915,-0.308855623,-0.0645196065,0.210163638,-0.2423413247,0.3780554831,-0.0971373841,0.0996658728,-0.1850442141,0.2720391452,-0.4597626626,0.1588378102,-0.2308292538,-0.0231041256,-0.0210545193,-0.1214517877,-0.0032287035,-0.0514779538,0.1981776059,-0.1648102552,-0.3105805814,-0.3956192434,-0.3741519451,0.083368957,-0.136477977,0.0981898382,0.3163701892,0.185457319,-0.1298381984,-0.1289100498,0.143339321,-0.1553377509,0.116885446,0.1116677895,-0.1184898764,0.1595136523,0.0437696017,0.2297572494,0.1588643938,0.0624907091,0.2197850496,0.1190112904,-0.18606475,-0.0817647949,-0.0189508181,0.3399614692,0.0368341208,-0.053259287,0.313926518,0.2866506577,0.048349984,0.3716550767,0.0161098484,0.0007949458,-0.0734477714,-0.0388318077,0.0450031646,0.3677760363,-0.3932712674,0.1037382782,0.3044113815,-0.1737763584,0.3948438168,0.3252357244,0.4229724705,-0.0688594729,0.1174554154,-0.1729154587,0.3712846041,-0.1252161711,0.0320204794,-0.2266666591,0.2815023065,0.0061894553,0.0040227687,-0.0145922154,0.4520942867,0.1025214493,0.2333649546,0.4167588651,0.2909471691,0.256341517,0.2657599449,-0.0618981421,0.4525990784,-0.0560594387,-0.1554635763,0.16860874,0.1719439179,0.5134887099,-0.1634740233,0.113682501,-0.304920435,0.0489035547,-0.0741871148,-0.124451831,0.0735484883,-0.0656518266,0.0001780736,-0.1817506105,-0.1853094697,-0.1901857257,0.2244735062,-0.1990799457,0.1965166032,-0.4829082191,0.0856152922,0.1291151494,0.4296555221,-0.0910244584,0.3184482455,0.1505437642,0.050712321,0.0418259799,0.634803772,0.5352048278,-0.2288816571,-0.0382534713,0.1204628497,0.013648646,-0.0249665026,-0.2032335848,0.1199001372,0.0024867591,-0.006518411,0.3161707222,0.2355193198,-0.2208194733,0.1295009255,0.1915676445,-0.1931168735,-0.1639721245,-0.2089206874,-0.1038833112,-0.0149615034,-0.3016755581,0.1877062917,-0.1402787566,0.086699076,0.3235279322,-0.144010067,0.0412712432,-0.090806067,0.1588253677,-0.1266366243,0.5540981889,0.3878645301,0.3791642785,-0.2192458957,-0.0528869219,-0.5427497029,0.2954648733,-0.0725850835,-0.2209935635,-0.1009818092,-0.0000632914,0.2585310936,0.248785466,-0.294754684,-0.0346338078,-0.1363819093,-0.3342483044,-0.4317246079,0.0334278978,0.1239256188,-0.0938493535,-0.0968317538,-0.0432841405,-0.0475071371,0.2306159139,0.0994545743,-0.1719502509,-0.062347386,0.0834817737,0.0983310938,-0.0231442079,0.1069741175,0.2078289092,0.0566936173,-0.0501745641,-0.0990115106,-0.173309654,0.0125706038,0.2085122764,0.3343246877,0.0631656274,-0.0756716281,-0.0571908467,0.4224995375,0.1578278095,0.0511480197,-0.0414934158,-0.0663163885,0.0607776567,0.0053767916,0.1209978238,-0.1437504888,0.4906178415,-0.0023903633,0.0267019123,0.1544553488,-0.4406585693,0.1211514622,0.1499329358,-0.1492207199,0.0177937895,0.3000476658,-0.0385965891,0.0934264958,-0.4432423115,-0.1145975813,0.3524498343,-0.1875695139,-0.2045790851,0.48066926,0.0429910384,-0.0958433896,-0.1897006631,0.2680135369,0.0385384522,-0.2053641826,0.0235921852,-0.2483790964]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2083","title":"`concatenate_datasets` throws error when changing the order of datasets to concatenate","comments":"Hi,\r\n\r\nthis bug is related to `Dataset.{remove_columns, rename_column, flatten}` not propagating the change to the schema metadata when the info features are updated, so this line is the culprit:\r\n```python\r\ncommon_voice_train = common_voice_train.remove_columns(['client_id', 'up_votes', 'down_votes', 'age', 'gender', 'accent', 'locale', 'segment'])\r\n\r\n``` \r\nThe order is important because the resulting dataset inherits the schema metadata of the first dataset passed to the `concatenate_datasets(...)` function (`pa.concat_tables` [docs](https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.concat_tables.html)). I'll try to fix this ASAP.","body":"Hey, \r\n\r\nI played around with the `concatenate_datasets(...)` function: https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?highlight=concatenate_datasets#datasets.concatenate_datasets\r\n\r\nand noticed that when the order in which the datasets are concatenated changes an error is thrown where it should not IMO.\r\n\r\nHere is a google colab to reproduce the error: https:\/\/colab.research.google.com\/drive\/17VTFU4KQ735-waWZJjeOHS6yDTfV5ekK?usp=sharing","comment_length":70,"text":"`concatenate_datasets` throws error when changing the order of datasets to concatenate\nHey, \r\n\r\nI played around with the `concatenate_datasets(...)` function: https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?highlight=concatenate_datasets#datasets.concatenate_datasets\r\n\r\nand noticed that when the order in which the datasets are concatenated changes an error is thrown where it should not IMO.\r\n\r\nHere is a google colab to reproduce the error: https:\/\/colab.research.google.com\/drive\/17VTFU4KQ735-waWZJjeOHS6yDTfV5ekK?usp=sharing\nHi,\r\n\r\nthis bug is related to `Dataset.{remove_columns, rename_column, flatten}` not propagating the change to the schema metadata when the info features are updated, so this line is the culprit:\r\n```python\r\ncommon_voice_train = common_voice_train.remove_columns(['client_id', 'up_votes', 'down_votes', 'age', 'gender', 'accent', 'locale', 'segment'])\r\n\r\n``` \r\nThe order is important because the resulting dataset inherits the schema metadata of the first dataset passed to the `concatenate_datasets(...)` function (`pa.concat_tables` [docs](https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.concat_tables.html)). I'll try to fix this ASAP.","embeddings":[-0.0796854198,-0.0035570161,0.0493244082,0.0962445959,0.4001432955,0.2031105757,0.1476463228,0.1685884148,-0.4686336219,0.0390344746,-0.11334382,0.2235859036,0.1865087897,0.2279952466,-0.1524720341,-0.3764883876,0.1803440005,-0.0335420966,-0.0318521857,0.1589729786,-0.3306152225,0.2782516181,-0.4217684269,0.1194007322,-0.2770402431,0.1655754745,-0.2876584828,0.2142007649,-0.131747961,-0.308303684,0.221818909,0.2089252919,0.0268243812,0.4889546037,-0.0001056023,-0.0131241502,0.1394525766,0.0269066729,-0.2604879141,-0.2272506654,-0.225127086,-0.1086767837,-0.1623241156,0.0379752778,0.1469448805,-0.159701854,-0.2429727167,-0.1502541155,0.2837979496,0.2774026096,0.2379924804,0.1755374074,0.2426477224,-0.3864704967,0.1160053238,-0.1271754503,-0.0323286094,0.0827695578,-0.157835111,-0.2114221156,0.1623141021,0.2935610712,-0.2205399424,-0.3700946271,0.0482722595,0.1449314654,-0.0490996353,-0.135188058,-0.0003450025,0.072688289,0.4189139605,-0.4847950041,-0.3530631959,-0.2878331244,0.3797289729,-0.3848426044,0.2335145026,0.1079680622,0.1808571815,0.1601840109,-0.1117425859,-0.0052864044,0.0029117123,0.0073696179,-0.0543385968,0.0794908032,0.0577444658,0.0121798953,0.1555245072,-0.3029306531,0.0764301568,-0.1139460281,-0.2723210752,0.1129819676,-0.4819879532,0.0012335911,0.0431282371,-0.252556473,0.3593414724,0.1304217428,-0.0217060558,-0.0283695441,-0.0975560844,0.1175317317,0.253303647,0.3586049676,0.0956618488,0.2279179692,0.0909649283,0.2217277437,0.1781282425,0.022759283,0.122976847,-0.293256104,-0.039441742,0.1017413363,0.305955708,0.0601080246,-0.1682273895,0.1103920862,-0.1277956516,0.0942963511,0.045383051,0.1752948463,0.0419019386,0.3478609622,0.2925291657,0.3554245532,0.1251151711,0.0136122657,-0.1590159833,-0.2540304959,-0.1026407182,0.29680264,-0.037919201,-0.11697115,-0.020783022,0.3439497948,0.1724196076,0.1211991161,-0.2369533777,0.0161234085,-0.1319067776,0.2203481644,0.0439352281,-0.0071234931,0.0624270923,-0.007491855,-0.1447689682,-0.0189630631,-0.2858016491,-0.5416709781,-0.4975184798,0.248180896,0.1760655344,-0.0887838677,0.0011430816,0.2109565288,0.3790776432,-0.2144869119,-0.0100913104,-0.1311472058,-0.3337848186,-0.1783127338,-0.0306516513,0.2728123367,-0.0783873945,0.0389477611,0.1670252681,0.0284016803,-0.135316208,0.4119367599,-0.1325293928,-0.119503662,-0.1640453488,0.2925847769,0.2148224711,-0.3234607577,-0.1954002827,0.015626844,-0.1306155324,0.0493506454,0.0488129668,-0.1344640404,-0.193211019,-0.1346464753,0.3389673233,0.3616016805,0.017334627,0.0530788675,-0.0639780313,-0.0736650974,0.2254837155,-0.1554727852,-0.4454149008,0.2084553838,-0.2820552289,-0.1806510389,0.3597200513,-0.2028734535,-0.0249193776,0.3321020007,0.1902955174,0.1028186008,0.2281092852,0.0330319628,-0.4944085479,0.0692005306,0.0779701546,0.1120494828,-0.2732640207,-0.3051612675,-0.1832832396,-0.0261381604,-0.3229482174,0.1075261682,0.1899009496,0.0190829579,0.1013582349,-0.1390779912,-0.1406923831,0.3601429164,-0.0401884206,0.1133174747,-0.4093932509,0.294631362,-0.2757107615,-0.0057807006,0.0960211456,0.2850590646,0.1517305523,0.0952609628,-0.2058046609,0.6714096069,-0.2294558883,-0.2981728613,0.0129514113,-0.0409915037,0.0063927756,0.0604904629,-0.101959236,-0.0107201189,-0.1760645807,-0.0578090511,-0.2104922235,0.3796176016,0.0027714288,0.2878172398,-0.0005501803,0.1273317039,0.2977004349,0.0236182511,-0.052281227,-0.3162916303,0.1110669598,0.0720413923,0.1083640754,0.0913331062,-0.5487784147,0.1539600044,0.6155299544,-0.0723493472,-0.0204856396,0.0756000876,-0.0469926298,-0.0043916823,0.0246623587,-0.034298718,0.5151794553,0.1922239363,-0.1451512128,0.3800708354,-0.1407290846,-0.1236649305,0.1264637858,0.0085585769,0.1879565418,0.3748966157,0.4192982614,0.2545351386,-0.2852300704,-0.2111072242,0.0020570953,0.0120813008,-0.2482442856,0.0661968216,-0.2989223599,-0.1473240107,-0.2143440694,-0.4559225738,-0.534398973,-0.36468032,0.1637233794,0.0951002836,-0.2306734174,0.320912689,-0.0733216628,0.0126334261,0.3799387217,-0.080790706,-0.2083370984,-0.030446792,0.0225907359,-0.0015236454,0.0968098938,-0.1379451752,0.0062935026,-0.0825955048,-0.0205822196,-0.271705091,-0.2621474564,0.0251964014,-0.3797262311,-0.0711877495,0.2013302743,-0.0451544262,0.1132562235,-0.4360528588,0.4303462803,0.1572064757,-0.2147203684,0.6075904369,0.1684282124,-0.1878268272,-0.1536263376,-0.407448113,-0.1293826252,-0.4146422744,0.2266934216,-0.2033040971,-0.0041574929,0.5315075517,0.1389883757,-0.2513957918,-0.1662716866,0.1982160062,-0.3319247365,-0.22686176,0.1182051748,-0.2356241345,-0.3592353165,-0.0796587616,0.1398949027,0.0986564979,0.063547574,-0.1372637153,-0.1302847415,-0.1049028412,0.2873049974,-0.3098780513,0.0758379847,0.3729993701,0.3206865489,-0.1161200628,-0.1279148608,-0.1826058179,-0.0058533968,0.2448066026,0.2611472607,-0.1807264835,0.3100435436,-0.0188209228,0.3491805494,0.2283605039,-0.0057375743,0.1428749263,0.040973112,0.3439367414,-0.2610262334,-0.303193748,-0.1533000022,-0.2464200407,0.2255495489,0.1236191019,0.0041146223,0.1353749782,-0.0419904105,0.3345293999,0.0660418347,-0.2268213183,-0.0082920035,-0.1154847816,0.277418673,0.1853944957,-0.046812579,-0.2113550901,-0.0027039154,0.0934270099,0.1489716768,-0.1669705659,-0.1312146932,-0.4330023825,-0.1807778776,-0.3780790865,0.3607479334,0.1474868804,0.3197355568,0.1055634171,-0.1144767329,-0.1143714115,0.0573050044,0.5016084909,-0.0727626607,-0.4048734903,0.4655727446,0.1853970289,-0.2289371341,-0.3301275671,-0.1556292325,0.1920014471,0.0992613882,0.3979488313,-0.3422978818,-0.1660372466,0.3565297127,0.2136690021,0.110990718,-0.0912760347,-0.3642391562,-0.2468781471,-0.30641675,-0.1125770584,-0.0565446503,0.4393942058,-0.4744166732,0.1360107213,0.2629765272,-0.2029903978,-0.1302559972,-0.0940752551,0.270575583,0.0067048916,0.1725480109,0.0501535684,0.2290183604,0.3115620315,0.5891232491,0.056888368,-0.7132000327,-0.1493978798,0.0947641879,0.1179932281,0.1717372686,-0.0863205865,0.3397352695,0.0198446251,0.1931262761,0.0650932342,-0.1433603466,0.6809390187,-0.0762922913,-0.6452266574,-0.314309746,0.5154020786,0.2080062628,0.012077922,0.1749998182,0.0367933922,-0.2994349897,0.287583977,-0.18846488,0.8607555032,0.0481342189,0.3077599704,0.3695949912,0.0936544836,0.3031533062,-0.0661678165,0.3982154429,-0.3754053414,-0.2533069551,-0.0042337254,-0.1032130867,0.0357057899,-0.0705222115,-0.5698779821,0.1437730789,-0.399630487,0.1986439377,-0.1967106164,0.1399356276,-0.2798862159,-0.173888728,-0.5225824118,0.1542306989,0.2264679372,-0.1210525408,-0.086457178,-0.0520522036,-0.0778321251,-0.2702745497,-0.2036771476,0.303085357,-0.2647879124,0.3141849935,0.1632529944,-0.1228603199,0.0996114314,0.0565376915,0.2340859771,-0.0648672953,-0.112313509,0.0260384362,0.1305633634,0.0637315661,-0.0409147739,-0.1266128272,0.3307842612,0.1718061566,-0.2644076049,0.0615200996,0.1078256443,-0.2041437477,-0.3926246464,-0.0122633632,-0.0215868149,-0.4964892268,0.1905368567,0.1542435288,0.1922668219,-0.2586192191,0.1730123162,0.0989521816,-0.1877945662,0.0471728407,-0.1320606917,-0.1740372926,-0.0082020229,0.3596715629,-0.0174257513,0.0402567573,0.4046959579,0.1664826125,-0.0942216665,-0.0620395243,-0.0262371097,-0.0034962802,-0.2976414859,0.1244309247,0.0895570964,0.0000789897,0.0600552596,0.4102918506,0.0109281344,-0.0974497795,0.0142664351,-0.068390809,-0.2066235393,0.2139131874,0.1953101158,0.0667900667,-0.0840723664,0.4412161708,-0.0367631316,0.0116877984,-0.357722342,-0.0698248222,-0.1693753749,0.2305311263,0.1366795003,-0.1088692918,-0.0193266142,-0.0654060245,0.190039143,0.1487475336,-0.0606444553,-0.2406534404,-0.0759333521,0.0728920922,0.2958261967,-0.0050349133,-0.0025417896,-0.1038459241,-0.1390255392,-0.250774771,0.058556322,0.2111117393,-0.0264701806,-0.0114154862,-0.0151091237,0.1182797551,0.3251684904,0.1095738634,0.0756919309,-0.1658451259,-0.1762661189,0.2312308401,-0.4591180086,-0.0251224879,0.044968091,0.2719659805,0.1442511082,0.0505010821,0.0981495827,-0.4305644035,-0.1101253778,-0.030727135,0.5766248107,0.4599285126,-0.2510532737,-0.0947915018,0.1912534386,0.2583906651,-0.433308363,-0.297131747,0.1146485656,-0.009314823,-0.07562875,-0.1008461639,0.1585254669,0.0354204364,-0.002257609,0.1471885145,0.1625920385,0.0039576022,0.363578856,0.3141833842,0.0872927383,0.1776060313,0.3767493665,0.1845766008,0.2560097873,0.2790784836,0.0073419334,0.1740577668,0.172059238,-0.2379532903,0.2201375067,-0.3379882574,-0.1055328697,0.0283958577,-0.016760163,-0.0702319145,-0.0515054017,0.5235236883,-0.3884119987,-0.2773981392,-0.142958954,0.3619551063,-0.0724961236,-0.0734698996,0.1042920873,0.0501956642,-0.1011230275,-0.1904467195,0.0580789186,-0.0687995255,0.6510472298,-0.1681030989,-0.0054694889,-0.5817869902,0.1904340684,0.2763527334,0.2139446735,-0.155439347,0.0498784818,0.2739631236,-0.1657039672,0.3347497284,0.2308692336,0.6353477836,0.2914364338,0.0203436948,0.0469661802,-0.1969207972,0.0724481046,0.2094468474,0.1526715904,0.120088838,-0.2670752108,0.2294176817,0.1359442919,-0.1342739463,-0.0999947786,0.2606732249,0.3138768077,-0.3910777867,0.1773741692,-0.2687882781,-0.266405195,-0.0935905725,0.1029702201,-0.2182296664,0.08945968,0.5371326804,0.3144436479,0.1687238514,-0.1476963758,0.1568285078,0.372618705,0.3128845394,0.5908777714,0.0228300355,-0.1525826752,-0.1991578937,-0.444182843,0.1017584205,0.0636743009,0.2398935407,-0.0085150851,0.0807565227,-0.0171229821,-0.0271069892,0.1293005198,-0.143470034,-0.1768602729,-0.0366691165,-0.4124816358,-0.1587408036,0.0380613208,-0.2000683844,0.1036180481,-0.2433555126,0.064430058,0.0986111313,0.0784641579,-0.1197605133,0.1893948913,0.1105799824,-0.0421844907,0.2163260281,0.2280187458,0.2743910551,-0.0294956751,-0.117654264,-0.2608792782,-0.6040965319,-0.414090395,0.149684906,-0.0973921046,0.0983686224,-0.0659843758,-0.0665804669,-0.1639699787,0.1544443071,0.1515691131,-0.1218537614,-0.4928680062,0.2210885286,0.0852267593,0.0022881364,-0.1137427911,0.200694263,0.0190556943,0.5071313381,-0.2035804987,-0.6065548062,0.6595958471,-0.0683464408,-0.3848757744,-0.0330868512,0.0976776406,0.080231756,-0.1977588087,-0.5424656272,0.0594670214,0.1686284691,0.039524015,-0.2772478461,0.2007496059,-0.0864769816,-0.0491296984,-0.11125011,0.3544621468,0.0607270896,-0.4413898289,0.2329764962,-0.1056705639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2080","title":"Multidimensional arrays in a Dataset","comments":"Hi !\r\n\r\nThis is actually supported ! but not yet in `from_pandas`.\r\nYou can use `from_dict` for now instead:\r\n```python\r\nfrom datasets import Dataset, Array2D, Features, Value\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = {\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n}\r\ndataset = Dataset.from_dict(dataset)\r\n```\r\n\r\nThis will work but to use it with the torch formatter you must specify the `Array2D` feature type in order to tell the shape:\r\n```python\r\nfrom datasets import Dataset, Array2D, Features, Value\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = {\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n}\r\ndataset = Dataset.from_dict(dataset, features=Features({\r\n \"bbox\": Array2D(shape=(3, 4), dtype=\"int64\"),\r\n \"input_ids\": Value(\"int64\")\r\n}))\r\ndataset.set_format(\"torch\")\r\nprint(dataset[0]['bbox'])\r\n# tensor([[1, 2, 3, 4],\r\n# [1, 2, 3, 4],\r\n# [1, 2, 3, 4]])\r\n```\r\nIf you don't specify the `Array2D` feature type, then the inferred type will be Sequence(Sequence(Value(\"int64\"))) and therefore the torch formatter will return list of tensors","body":"Hi,\r\n\r\nI'm trying to put together a `datasets.Dataset` to be used with LayoutLM which is available in `transformers`. This model requires as input the bounding boxes of each of the token of a sequence. This is when I realized that `Dataset` does not support multi-dimensional arrays as a value for a column in a row.\r\n\r\nThe following code results in conversion error in pyarrow (`pyarrow.lib.ArrowInvalid: ('Can only convert 1-dimensional array values', 'Conversion failed for column bbox with type object')`)\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = pd.DataFrame({\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n})\r\ndataset = Dataset.from_pandas(dataset)\r\n```\r\n\r\nSince I wanted to use pytorch for the downstream training task, I also tried a few ways to directly put in a column of 2-D pytorch tensor in a formatted dataset, but I can only have a list of 1-D tensors, or a list of arrays, or a list of lists.\r\n\r\n```\r\nimport torch\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\n\r\ndataset = pd.DataFrame({\r\n 'bbox': [\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]]\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n})\r\ndataset = Dataset.from_pandas(dataset)\r\n\r\ndef test(examples):\r\n return {'bbbox': torch.Tensor(examples['bbox'])}\r\ndataset = dataset.map(test)\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n\r\ndataset.set_format(type='torch', columns=['input_ids', 'bbox'], output_all_columns=True)\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n\r\ndef test2(examples):\r\n return {'bbbox': torch.stack(examples['bbox'])}\r\ndataset = dataset.map(test2)\r\n\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n```\r\n\r\nIs is possible to support n-D arrays\/tensors in datasets? \r\nIt seems that it can also be useful for this [feature request](https:\/\/github.com\/huggingface\/datasets\/issues\/263).","comment_length":165,"text":"Multidimensional arrays in a Dataset\nHi,\r\n\r\nI'm trying to put together a `datasets.Dataset` to be used with LayoutLM which is available in `transformers`. This model requires as input the bounding boxes of each of the token of a sequence. This is when I realized that `Dataset` does not support multi-dimensional arrays as a value for a column in a row.\r\n\r\nThe following code results in conversion error in pyarrow (`pyarrow.lib.ArrowInvalid: ('Can only convert 1-dimensional array values', 'Conversion failed for column bbox with type object')`)\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = pd.DataFrame({\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n})\r\ndataset = Dataset.from_pandas(dataset)\r\n```\r\n\r\nSince I wanted to use pytorch for the downstream training task, I also tried a few ways to directly put in a column of 2-D pytorch tensor in a formatted dataset, but I can only have a list of 1-D tensors, or a list of arrays, or a list of lists.\r\n\r\n```\r\nimport torch\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\n\r\ndataset = pd.DataFrame({\r\n 'bbox': [\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]]\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n})\r\ndataset = Dataset.from_pandas(dataset)\r\n\r\ndef test(examples):\r\n return {'bbbox': torch.Tensor(examples['bbox'])}\r\ndataset = dataset.map(test)\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n\r\ndataset.set_format(type='torch', columns=['input_ids', 'bbox'], output_all_columns=True)\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n\r\ndef test2(examples):\r\n return {'bbbox': torch.stack(examples['bbox'])}\r\ndataset = dataset.map(test2)\r\n\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n```\r\n\r\nIs is possible to support n-D arrays\/tensors in datasets? \r\nIt seems that it can also be useful for this [feature request](https:\/\/github.com\/huggingface\/datasets\/issues\/263).\nHi !\r\n\r\nThis is actually supported ! but not yet in `from_pandas`.\r\nYou can use `from_dict` for now instead:\r\n```python\r\nfrom datasets import Dataset, Array2D, Features, Value\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = {\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n}\r\ndataset = Dataset.from_dict(dataset)\r\n```\r\n\r\nThis will work but to use it with the torch formatter you must specify the `Array2D` feature type in order to tell the shape:\r\n```python\r\nfrom datasets import Dataset, Array2D, Features, Value\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = {\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n}\r\ndataset = Dataset.from_dict(dataset, features=Features({\r\n \"bbox\": Array2D(shape=(3, 4), dtype=\"int64\"),\r\n \"input_ids\": Value(\"int64\")\r\n}))\r\ndataset.set_format(\"torch\")\r\nprint(dataset[0]['bbox'])\r\n# tensor([[1, 2, 3, 4],\r\n# [1, 2, 3, 4],\r\n# [1, 2, 3, 4]])\r\n```\r\nIf you don't specify the `Array2D` feature type, then the inferred type will be Sequence(Sequence(Value(\"int64\"))) and therefore the torch formatter will return list of tensors","embeddings":[0.0149022313,-0.2567944229,-0.079935424,0.1915314645,0.4495704174,0.054927934,0.8991123438,0.1010944247,0.180424422,0.0678299218,-0.110976994,0.3204252124,-0.2819293737,0.0522116907,0.1784114242,-0.4505749643,0.163114652,0.2940376401,-0.3680661023,0.0973285586,-0.0565509163,-0.0563466921,-0.2456144243,-0.0419221669,-0.3301548064,-0.2008834332,-0.1476366818,-0.1640363932,-0.0638760179,-0.1978779584,0.2409163713,-0.1706257612,0.2700603604,0.2606250346,-0.0001195509,0.2267989665,0.024773784,-0.1292835027,-0.0359679013,-0.0393488258,0.2725146115,-0.6480606198,-0.0806007534,-0.432475239,0.0080717802,-0.4712079763,-0.2638435662,-0.5005581975,0.0513902642,0.3928325176,0.1141964942,0.3314775527,0.2363567799,0.1579979956,0.1462766677,0.0132805388,-0.1706013232,-0.2005452514,0.239397049,-0.207225889,0.0300292391,0.0932103172,-0.2798197865,0.1661209762,0.2681391537,0.2049356699,0.0058009871,0.0611363277,-0.0313577019,0.1278439164,0.4276586771,-0.3980729878,-0.3639862239,-0.2793341279,0.0686928406,-0.1197742745,-0.109529987,0.3258959353,-0.1265999377,0.0910808444,-0.4061063826,0.1408004314,-0.318246603,0.3343325257,-0.404346168,0.2282194644,0.1337881237,0.1612996906,0.0395999178,-0.1324870884,0.7339075804,-0.0246707387,0.3878382444,0.0780725256,-0.3060498238,-0.0292106401,-0.1911205053,-0.5545958281,-0.0388238579,-0.4604115784,0.3082380891,0.1663304567,0.1043882743,0.1976925582,0.0353528671,0.2383281887,-0.4363076389,0.4121215045,-0.0672477633,-0.2368495762,0.481027782,0.099248156,-0.1561656743,-0.1502647549,0.1455623209,0.0061738966,-0.0272424854,0.3684129715,-0.2227881998,0.0070690843,-0.5160890222,-0.0086483778,0.0363631845,0.1473617405,0.1086722165,0.3957225382,0.0416682325,0.3720445931,0.1662516147,-0.119526498,-0.0711396337,0.3224996924,-0.3391715884,-0.0316563956,0.1639197618,0.5651168823,-0.0971913114,0.0204912834,0.20878084,-0.076432839,0.0289029647,-0.08741685,0.2978046238,0.290881604,-0.1416594237,-0.0486805998,0.0412730426,0.2416127324,-0.0130374432,0.4392251372,-0.0711847767,-0.0471878722,-0.3636685312,-0.0374877378,-0.0945892558,-0.0264146719,-0.1960165352,0.1866909117,0.4951613545,-0.0282086171,0.1075465605,-0.5084322095,0.0889307186,-0.2072445154,0.1400897652,-0.3502917588,-0.3905958235,0.2010028064,0.2407013625,0.3463239074,0.0975594893,0.2834509909,-0.3449477255,0.0540575795,0.0329005793,0.2473342866,0.338768661,-0.0420008563,-0.0871174335,-0.1622388661,-0.0173253324,-0.000717895,0.1112007126,0.3168395162,0.1397200674,0.1958103329,0.1861190051,0.1972095817,-0.0389661156,0.3523800671,-0.0916294903,-0.2154469192,0.4633560181,0.2579582334,0.3060801327,-0.1548969448,0.0176662505,0.0563277788,0.0940437391,-0.5431793928,0.1949042529,0.3491542637,-0.018713491,-0.0986341834,0.0564414375,-0.4492571056,-0.7101914883,0.1608666629,0.0483023599,0.1065320298,-0.1837565303,-0.0663251951,-0.1079390198,0.1118320078,0.072069563,0.1178974584,0.0417497531,0.0352335237,-0.0735230744,-0.397126019,-0.2047450691,0.0776522681,0.3375150561,-0.0829180181,-0.3645120859,0.592897296,-0.1039523855,-0.2862936556,0.0101346485,0.4462893009,0.0611800328,-0.0496978685,0.197759077,0.2274239659,0.1230929941,-0.3832360804,-0.4160242081,0.4727387428,0.2562884688,-0.272180438,0.2036811858,0.0511426404,0.01573167,-0.0084934952,-0.3207271695,0.568791151,-0.001434369,-0.0261085052,-0.272628665,0.1914387941,0.0833539218,0.0909292623,-0.1171600968,-0.0677580982,-0.2300039828,-0.2115735263,0.0922158435,-0.0045593404,-0.389783442,0.0558723211,0.267754972,-0.189251557,0.2653499544,-0.1347501725,-0.0375846289,0.0150864795,0.1793589741,-0.0371525697,0.1663657874,0.2058114409,0.1150868014,-0.1248200312,-0.4518621266,-0.1397283375,0.0517302863,-0.1079203933,0.181242317,0.0518616252,0.2955330908,-0.1676353514,-0.0637023523,0.0914284363,0.3197639287,0.2288444191,-0.406082809,0.1624588519,-0.3874128163,-0.0664661229,-0.2304770052,-0.260181427,0.0049729138,0.0868771896,-0.088019453,0.234202981,0.0812050253,0.1689583063,0.0277667344,-0.0450002551,0.2548814416,-0.2477962226,-0.062487632,-0.1313533485,-0.0641593635,-0.0014591572,0.1030865833,0.1419320852,0.1337921619,0.4435394108,-0.0571619235,-0.4592995942,-0.0347506143,-0.1463270187,-0.2035893202,-0.1697765142,0.0904792026,0.0106052151,-0.1125862151,-0.0403797366,0.3096112907,0.2927798927,0.0355215482,0.378958106,0.2087776512,-0.1153353527,-0.1274021864,-0.1328262538,-0.1498316526,-0.2656078935,0.0212591682,-0.3921453655,0.1053935587,-0.2647850215,0.3787118196,0.1925883591,0.1587032676,-0.2593991458,0.0740580037,0.344668448,0.6661354899,-0.0351950526,-0.3907384872,-0.3848359585,-0.1644585729,0.0585448891,0.701117754,-0.2124484181,-0.2320621461,-0.1337760389,-0.0626931563,-0.1328706145,-0.059933722,0.6345390677,0.0861986876,0.0013020391,-0.0903871134,-0.1663161516,0.06630335,0.1509484798,0.0939808339,-0.0813990384,0.3435147107,0.2753554881,0.3017744124,0.0231571905,-0.2025800496,0.2516397238,-0.1907729506,0.1549074352,-0.0294431727,-0.2573098242,-0.0765906349,-0.3783545792,-0.1627089977,-0.0877784193,-0.2194223255,-0.5588403344,-0.0134953903,0.3541659713,-0.112861149,-0.0620546862,0.4988287091,-0.386607796,0.3288064301,-0.0005443048,0.0491639338,-0.3488150537,-0.1622702926,-0.1883964837,-0.3264348209,-0.1317931116,0.0207401123,-0.5361710191,-0.2696361244,0.1808828115,0.1261781752,0.3767213523,0.7153761983,0.1138262749,0.0400265455,0.21624358,-0.0295358226,0.5638820529,-0.0700070187,-0.3026607633,0.3299584091,0.0555714704,-0.521756053,-0.2325491905,-0.3563009202,-0.1439822465,-0.0971958712,0.4788838923,-0.3085343838,-0.2306003869,0.1910068542,0.185374096,-0.1017973796,-0.0734611228,-0.2473593354,-0.2527901828,-0.1889770776,-0.1428678483,0.4407948256,0.3197946846,0.211081937,-0.2233940065,-0.1055804789,-0.3295325339,0.0886833519,-0.03684102,0.0896116793,0.1358523816,-0.1247346997,0.1726720482,0.2221591622,0.4825526476,0.0961530209,-0.1107679084,-0.0135595156,0.1709893644,-0.1279083043,0.3795845807,0.1518592685,-0.1087806895,-0.2579708993,-0.0883317664,0.0015154268,-0.1207925379,-0.0297300015,0.2756111622,-0.0356957801,-0.2073619366,-0.277988404,0.2343326956,-0.0254750997,0.1139989197,0.1716229171,0.2811611295,-0.1681227535,0.288346529,0.3334234059,0.677400887,-0.505526185,0.3920155168,0.4707365632,-0.1107947081,0.2089315355,-0.1366589814,0.3178485334,-0.2395371348,0.0199291408,-0.2089606076,-0.185018152,0.2092852145,0.418643862,-0.0276028123,0.0813101158,0.0485318229,0.210947752,0.1919577569,-0.087106809,0.1290740073,-0.2450117469,-0.0733691528,0.0722153485,-0.1949899346,-0.3241555691,-0.0126737989,-0.2041720003,-0.3422338068,-0.0203461293,-0.3445939124,-0.1296945959,-0.3006146848,0.4228592813,0.1384602636,-0.4948453605,0.0487053208,0.5728102922,0.0256157983,-0.3815951645,-0.1450159699,-0.3738671839,-0.2349481732,-0.2208291441,-0.0488315299,-0.2558015883,0.3040538132,0.015650576,-0.0739732534,0.1643846035,-0.0304168481,-0.3962722123,-0.0108928662,0.2503052056,0.097207509,-0.2461638302,-0.0253751241,-0.1884801239,0.1446112245,-0.0500300489,0.0604319349,-0.0965674371,-0.0386197977,0.2899439037,-0.0460204408,0.0774850026,0.1257364601,0.2239428163,0.1179819182,-0.3031094074,0.1812613606,0.109249182,-0.3035005927,-0.1538247913,0.2004556954,0.5446794033,-0.0663514733,0.0234458279,-0.3518611491,-0.3031123579,-0.0683953092,0.148407042,0.3131949008,0.1276943684,-0.1933960319,-0.2029373795,-0.263492316,0.4287037551,0.3794879615,0.2747296393,-0.0351037905,0.2906456292,-0.1598078161,0.0051986584,-0.231417641,-0.0259404089,-0.0979493484,0.2771215141,-0.0215536505,-0.3222759366,-0.1066991389,-0.1097234264,0.0815521479,-0.0875355974,-0.1162485555,-0.1111934781,0.078116037,0.1195414662,0.118383415,-0.0039284574,0.0257312991,-0.4714922607,-0.3264598846,-0.2363394946,0.008007939,0.4533220232,-0.0041555106,0.1654735655,0.2998788357,0.1157933846,-0.0761755928,-0.1688297093,-0.1501292437,-0.1383606941,0.0050386144,0.2183294296,0.248464644,0.0066202679,-0.2967952192,0.1318215877,0.0704101101,-0.0740277916,0.3137459755,-0.3049809635,-0.0527507961,-0.1453633308,0.4793773293,-0.0544440784,-0.3318729699,0.0426694378,0.1517041773,0.0878730491,-0.2129775435,-0.3487560153,0.1269572079,0.0348963849,0.1410359144,0.5094712377,0.3671321273,0.1969045401,0.1261757612,-0.1247610748,0.3551172316,-0.1337393373,-0.1359168142,0.3305079639,-0.0477365032,-0.0159560032,0.2496727258,0.0032824131,0.2069059312,0.5683193207,-0.0471288599,0.2942135036,0.4007830322,0.1045905426,-0.1734313965,0.1579915583,-0.0968309194,-0.0942730606,0.2389733642,-0.1445460171,-0.1482386887,0.2099242657,-0.1258606315,-0.0203978047,-0.1106933281,0.0989400893,-0.2336272448,0.1443283409,0.0261035878,0.0543620735,-0.0992336795,-0.1871893853,-0.0987330675,-0.0773622021,0.3327878714,-0.0060645468,0.1644575745,0.0103617897,0.179524079,0.4636001289,0.0343802571,-0.1801714897,0.2851693034,0.317489177,-0.1626901925,-0.066927202,0.0285953674,0.2932972312,0.2905929387,0.0055147386,0.0178469568,-0.158621788,-0.1852167845,-0.0713802427,0.0422844924,0.0392187946,0.2712653875,0.2794286907,0.028664548,-0.0653822869,0.178452909,-0.0235885754,0.0513595194,-0.1432749927,0.2960940301,-0.1465903521,-0.5235889554,0.1866844147,-0.3249715269,-0.3480745852,0.0296944175,0.3392666578,-0.0802819729,0.2214090526,0.2769951224,0.0203873236,0.1460019499,0.4759078026,0.0976115316,0.3561215997,-0.0767005682,-0.0689726695,-0.5012392998,0.1398172379,-0.4139970839,0.0408438742,0.2894043028,0.1724706739,0.322507143,0.2895634472,0.0376227014,0.1403890252,-0.096874401,0.3843570948,-0.2270287573,-0.0645690262,0.1950525343,0.2118771374,0.1090166569,-0.2388573736,0.3231708705,0.2930503786,0.0205491371,-0.2603202462,0.3992149234,-0.1971804053,0.307482481,-0.0728120208,-0.1389238685,0.5194199085,0.2293463796,0.2214035243,-0.1439764053,0.0300036855,-0.0908123478,0.3211365044,0.1461285055,0.4942115247,-0.431876868,-0.2108418494,-0.0542695075,0.1428682655,0.0246400442,-0.0425273143,-0.1893388629,0.2584880292,-0.2203613669,-0.0148228314,0.0579342619,0.0337438472,0.0794112459,0.219566226,-0.2821283638,-0.4187386036,0.4235398769,-0.1683394015,-0.5067198277,0.1799818873,-0.2447522879,0.155995518,0.1381157935,-0.5674667358,-0.1494414061,0.2118353099,-0.1582861543,-0.2038371563,-0.0375708379,-0.0058429874,-0.1672918648,-0.1489332765,0.3547782898,-0.1051538885,-0.1671753228,-0.4202934504,-0.3254515827]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2080","title":"Multidimensional arrays in a Dataset","comments":"Thanks for the explanation. \r\nWith my original DataFrame, I did\r\n```\r\ndataset = dataset.to_dict(\"list\")\r\n```\r\nand then the rest of the transformation from dictionary works just fine.","body":"Hi,\r\n\r\nI'm trying to put together a `datasets.Dataset` to be used with LayoutLM which is available in `transformers`. This model requires as input the bounding boxes of each of the token of a sequence. This is when I realized that `Dataset` does not support multi-dimensional arrays as a value for a column in a row.\r\n\r\nThe following code results in conversion error in pyarrow (`pyarrow.lib.ArrowInvalid: ('Can only convert 1-dimensional array values', 'Conversion failed for column bbox with type object')`)\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = pd.DataFrame({\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n})\r\ndataset = Dataset.from_pandas(dataset)\r\n```\r\n\r\nSince I wanted to use pytorch for the downstream training task, I also tried a few ways to directly put in a column of 2-D pytorch tensor in a formatted dataset, but I can only have a list of 1-D tensors, or a list of arrays, or a list of lists.\r\n\r\n```\r\nimport torch\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\n\r\ndataset = pd.DataFrame({\r\n 'bbox': [\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]]\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n})\r\ndataset = Dataset.from_pandas(dataset)\r\n\r\ndef test(examples):\r\n return {'bbbox': torch.Tensor(examples['bbox'])}\r\ndataset = dataset.map(test)\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n\r\ndataset.set_format(type='torch', columns=['input_ids', 'bbox'], output_all_columns=True)\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n\r\ndef test2(examples):\r\n return {'bbbox': torch.stack(examples['bbox'])}\r\ndataset = dataset.map(test2)\r\n\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n```\r\n\r\nIs is possible to support n-D arrays\/tensors in datasets? \r\nIt seems that it can also be useful for this [feature request](https:\/\/github.com\/huggingface\/datasets\/issues\/263).","comment_length":27,"text":"Multidimensional arrays in a Dataset\nHi,\r\n\r\nI'm trying to put together a `datasets.Dataset` to be used with LayoutLM which is available in `transformers`. This model requires as input the bounding boxes of each of the token of a sequence. This is when I realized that `Dataset` does not support multi-dimensional arrays as a value for a column in a row.\r\n\r\nThe following code results in conversion error in pyarrow (`pyarrow.lib.ArrowInvalid: ('Can only convert 1-dimensional array values', 'Conversion failed for column bbox with type object')`)\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = pd.DataFrame({\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n})\r\ndataset = Dataset.from_pandas(dataset)\r\n```\r\n\r\nSince I wanted to use pytorch for the downstream training task, I also tried a few ways to directly put in a column of 2-D pytorch tensor in a formatted dataset, but I can only have a list of 1-D tensors, or a list of arrays, or a list of lists.\r\n\r\n```\r\nimport torch\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\n\r\ndataset = pd.DataFrame({\r\n 'bbox': [\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]],\r\n [[1,2,3,4],[1,2,3,4],[1,2,3,4]]\r\n ],\r\n 'input_ids': [1, 2, 3, 4]\r\n})\r\ndataset = Dataset.from_pandas(dataset)\r\n\r\ndef test(examples):\r\n return {'bbbox': torch.Tensor(examples['bbox'])}\r\ndataset = dataset.map(test)\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n\r\ndataset.set_format(type='torch', columns=['input_ids', 'bbox'], output_all_columns=True)\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n\r\ndef test2(examples):\r\n return {'bbbox': torch.stack(examples['bbox'])}\r\ndataset = dataset.map(test2)\r\n\r\nprint(dataset[0]['bbox'])\r\nprint(dataset[0]['bbbox'])\r\n```\r\n\r\nIs is possible to support n-D arrays\/tensors in datasets? \r\nIt seems that it can also be useful for this [feature request](https:\/\/github.com\/huggingface\/datasets\/issues\/263).\nThanks for the explanation. \r\nWith my original DataFrame, I did\r\n```\r\ndataset = dataset.to_dict(\"list\")\r\n```\r\nand then the rest of the transformation from dictionary works just fine.","embeddings":[0.0149022313,-0.2567944229,-0.079935424,0.1915314645,0.4495704174,0.054927934,0.8991123438,0.1010944247,0.180424422,0.0678299218,-0.110976994,0.3204252124,-0.2819293737,0.0522116907,0.1784114242,-0.4505749643,0.163114652,0.2940376401,-0.3680661023,0.0973285586,-0.0565509163,-0.0563466921,-0.2456144243,-0.0419221669,-0.3301548064,-0.2008834332,-0.1476366818,-0.1640363932,-0.0638760179,-0.1978779584,0.2409163713,-0.1706257612,0.2700603604,0.2606250346,-0.0001195509,0.2267989665,0.024773784,-0.1292835027,-0.0359679013,-0.0393488258,0.2725146115,-0.6480606198,-0.0806007534,-0.432475239,0.0080717802,-0.4712079763,-0.2638435662,-0.5005581975,0.0513902642,0.3928325176,0.1141964942,0.3314775527,0.2363567799,0.1579979956,0.1462766677,0.0132805388,-0.1706013232,-0.2005452514,0.239397049,-0.207225889,0.0300292391,0.0932103172,-0.2798197865,0.1661209762,0.2681391537,0.2049356699,0.0058009871,0.0611363277,-0.0313577019,0.1278439164,0.4276586771,-0.3980729878,-0.3639862239,-0.2793341279,0.0686928406,-0.1197742745,-0.109529987,0.3258959353,-0.1265999377,0.0910808444,-0.4061063826,0.1408004314,-0.318246603,0.3343325257,-0.404346168,0.2282194644,0.1337881237,0.1612996906,0.0395999178,-0.1324870884,0.7339075804,-0.0246707387,0.3878382444,0.0780725256,-0.3060498238,-0.0292106401,-0.1911205053,-0.5545958281,-0.0388238579,-0.4604115784,0.3082380891,0.1663304567,0.1043882743,0.1976925582,0.0353528671,0.2383281887,-0.4363076389,0.4121215045,-0.0672477633,-0.2368495762,0.481027782,0.099248156,-0.1561656743,-0.1502647549,0.1455623209,0.0061738966,-0.0272424854,0.3684129715,-0.2227881998,0.0070690843,-0.5160890222,-0.0086483778,0.0363631845,0.1473617405,0.1086722165,0.3957225382,0.0416682325,0.3720445931,0.1662516147,-0.119526498,-0.0711396337,0.3224996924,-0.3391715884,-0.0316563956,0.1639197618,0.5651168823,-0.0971913114,0.0204912834,0.20878084,-0.076432839,0.0289029647,-0.08741685,0.2978046238,0.290881604,-0.1416594237,-0.0486805998,0.0412730426,0.2416127324,-0.0130374432,0.4392251372,-0.0711847767,-0.0471878722,-0.3636685312,-0.0374877378,-0.0945892558,-0.0264146719,-0.1960165352,0.1866909117,0.4951613545,-0.0282086171,0.1075465605,-0.5084322095,0.0889307186,-0.2072445154,0.1400897652,-0.3502917588,-0.3905958235,0.2010028064,0.2407013625,0.3463239074,0.0975594893,0.2834509909,-0.3449477255,0.0540575795,0.0329005793,0.2473342866,0.338768661,-0.0420008563,-0.0871174335,-0.1622388661,-0.0173253324,-0.000717895,0.1112007126,0.3168395162,0.1397200674,0.1958103329,0.1861190051,0.1972095817,-0.0389661156,0.3523800671,-0.0916294903,-0.2154469192,0.4633560181,0.2579582334,0.3060801327,-0.1548969448,0.0176662505,0.0563277788,0.0940437391,-0.5431793928,0.1949042529,0.3491542637,-0.018713491,-0.0986341834,0.0564414375,-0.4492571056,-0.7101914883,0.1608666629,0.0483023599,0.1065320298,-0.1837565303,-0.0663251951,-0.1079390198,0.1118320078,0.072069563,0.1178974584,0.0417497531,0.0352335237,-0.0735230744,-0.397126019,-0.2047450691,0.0776522681,0.3375150561,-0.0829180181,-0.3645120859,0.592897296,-0.1039523855,-0.2862936556,0.0101346485,0.4462893009,0.0611800328,-0.0496978685,0.197759077,0.2274239659,0.1230929941,-0.3832360804,-0.4160242081,0.4727387428,0.2562884688,-0.272180438,0.2036811858,0.0511426404,0.01573167,-0.0084934952,-0.3207271695,0.568791151,-0.001434369,-0.0261085052,-0.272628665,0.1914387941,0.0833539218,0.0909292623,-0.1171600968,-0.0677580982,-0.2300039828,-0.2115735263,0.0922158435,-0.0045593404,-0.389783442,0.0558723211,0.267754972,-0.189251557,0.2653499544,-0.1347501725,-0.0375846289,0.0150864795,0.1793589741,-0.0371525697,0.1663657874,0.2058114409,0.1150868014,-0.1248200312,-0.4518621266,-0.1397283375,0.0517302863,-0.1079203933,0.181242317,0.0518616252,0.2955330908,-0.1676353514,-0.0637023523,0.0914284363,0.3197639287,0.2288444191,-0.406082809,0.1624588519,-0.3874128163,-0.0664661229,-0.2304770052,-0.260181427,0.0049729138,0.0868771896,-0.088019453,0.234202981,0.0812050253,0.1689583063,0.0277667344,-0.0450002551,0.2548814416,-0.2477962226,-0.062487632,-0.1313533485,-0.0641593635,-0.0014591572,0.1030865833,0.1419320852,0.1337921619,0.4435394108,-0.0571619235,-0.4592995942,-0.0347506143,-0.1463270187,-0.2035893202,-0.1697765142,0.0904792026,0.0106052151,-0.1125862151,-0.0403797366,0.3096112907,0.2927798927,0.0355215482,0.378958106,0.2087776512,-0.1153353527,-0.1274021864,-0.1328262538,-0.1498316526,-0.2656078935,0.0212591682,-0.3921453655,0.1053935587,-0.2647850215,0.3787118196,0.1925883591,0.1587032676,-0.2593991458,0.0740580037,0.344668448,0.6661354899,-0.0351950526,-0.3907384872,-0.3848359585,-0.1644585729,0.0585448891,0.701117754,-0.2124484181,-0.2320621461,-0.1337760389,-0.0626931563,-0.1328706145,-0.059933722,0.6345390677,0.0861986876,0.0013020391,-0.0903871134,-0.1663161516,0.06630335,0.1509484798,0.0939808339,-0.0813990384,0.3435147107,0.2753554881,0.3017744124,0.0231571905,-0.2025800496,0.2516397238,-0.1907729506,0.1549074352,-0.0294431727,-0.2573098242,-0.0765906349,-0.3783545792,-0.1627089977,-0.0877784193,-0.2194223255,-0.5588403344,-0.0134953903,0.3541659713,-0.112861149,-0.0620546862,0.4988287091,-0.386607796,0.3288064301,-0.0005443048,0.0491639338,-0.3488150537,-0.1622702926,-0.1883964837,-0.3264348209,-0.1317931116,0.0207401123,-0.5361710191,-0.2696361244,0.1808828115,0.1261781752,0.3767213523,0.7153761983,0.1138262749,0.0400265455,0.21624358,-0.0295358226,0.5638820529,-0.0700070187,-0.3026607633,0.3299584091,0.0555714704,-0.521756053,-0.2325491905,-0.3563009202,-0.1439822465,-0.0971958712,0.4788838923,-0.3085343838,-0.2306003869,0.1910068542,0.185374096,-0.1017973796,-0.0734611228,-0.2473593354,-0.2527901828,-0.1889770776,-0.1428678483,0.4407948256,0.3197946846,0.211081937,-0.2233940065,-0.1055804789,-0.3295325339,0.0886833519,-0.03684102,0.0896116793,0.1358523816,-0.1247346997,0.1726720482,0.2221591622,0.4825526476,0.0961530209,-0.1107679084,-0.0135595156,0.1709893644,-0.1279083043,0.3795845807,0.1518592685,-0.1087806895,-0.2579708993,-0.0883317664,0.0015154268,-0.1207925379,-0.0297300015,0.2756111622,-0.0356957801,-0.2073619366,-0.277988404,0.2343326956,-0.0254750997,0.1139989197,0.1716229171,0.2811611295,-0.1681227535,0.288346529,0.3334234059,0.677400887,-0.505526185,0.3920155168,0.4707365632,-0.1107947081,0.2089315355,-0.1366589814,0.3178485334,-0.2395371348,0.0199291408,-0.2089606076,-0.185018152,0.2092852145,0.418643862,-0.0276028123,0.0813101158,0.0485318229,0.210947752,0.1919577569,-0.087106809,0.1290740073,-0.2450117469,-0.0733691528,0.0722153485,-0.1949899346,-0.3241555691,-0.0126737989,-0.2041720003,-0.3422338068,-0.0203461293,-0.3445939124,-0.1296945959,-0.3006146848,0.4228592813,0.1384602636,-0.4948453605,0.0487053208,0.5728102922,0.0256157983,-0.3815951645,-0.1450159699,-0.3738671839,-0.2349481732,-0.2208291441,-0.0488315299,-0.2558015883,0.3040538132,0.015650576,-0.0739732534,0.1643846035,-0.0304168481,-0.3962722123,-0.0108928662,0.2503052056,0.097207509,-0.2461638302,-0.0253751241,-0.1884801239,0.1446112245,-0.0500300489,0.0604319349,-0.0965674371,-0.0386197977,0.2899439037,-0.0460204408,0.0774850026,0.1257364601,0.2239428163,0.1179819182,-0.3031094074,0.1812613606,0.109249182,-0.3035005927,-0.1538247913,0.2004556954,0.5446794033,-0.0663514733,0.0234458279,-0.3518611491,-0.3031123579,-0.0683953092,0.148407042,0.3131949008,0.1276943684,-0.1933960319,-0.2029373795,-0.263492316,0.4287037551,0.3794879615,0.2747296393,-0.0351037905,0.2906456292,-0.1598078161,0.0051986584,-0.231417641,-0.0259404089,-0.0979493484,0.2771215141,-0.0215536505,-0.3222759366,-0.1066991389,-0.1097234264,0.0815521479,-0.0875355974,-0.1162485555,-0.1111934781,0.078116037,0.1195414662,0.118383415,-0.0039284574,0.0257312991,-0.4714922607,-0.3264598846,-0.2363394946,0.008007939,0.4533220232,-0.0041555106,0.1654735655,0.2998788357,0.1157933846,-0.0761755928,-0.1688297093,-0.1501292437,-0.1383606941,0.0050386144,0.2183294296,0.248464644,0.0066202679,-0.2967952192,0.1318215877,0.0704101101,-0.0740277916,0.3137459755,-0.3049809635,-0.0527507961,-0.1453633308,0.4793773293,-0.0544440784,-0.3318729699,0.0426694378,0.1517041773,0.0878730491,-0.2129775435,-0.3487560153,0.1269572079,0.0348963849,0.1410359144,0.5094712377,0.3671321273,0.1969045401,0.1261757612,-0.1247610748,0.3551172316,-0.1337393373,-0.1359168142,0.3305079639,-0.0477365032,-0.0159560032,0.2496727258,0.0032824131,0.2069059312,0.5683193207,-0.0471288599,0.2942135036,0.4007830322,0.1045905426,-0.1734313965,0.1579915583,-0.0968309194,-0.0942730606,0.2389733642,-0.1445460171,-0.1482386887,0.2099242657,-0.1258606315,-0.0203978047,-0.1106933281,0.0989400893,-0.2336272448,0.1443283409,0.0261035878,0.0543620735,-0.0992336795,-0.1871893853,-0.0987330675,-0.0773622021,0.3327878714,-0.0060645468,0.1644575745,0.0103617897,0.179524079,0.4636001289,0.0343802571,-0.1801714897,0.2851693034,0.317489177,-0.1626901925,-0.066927202,0.0285953674,0.2932972312,0.2905929387,0.0055147386,0.0178469568,-0.158621788,-0.1852167845,-0.0713802427,0.0422844924,0.0392187946,0.2712653875,0.2794286907,0.028664548,-0.0653822869,0.178452909,-0.0235885754,0.0513595194,-0.1432749927,0.2960940301,-0.1465903521,-0.5235889554,0.1866844147,-0.3249715269,-0.3480745852,0.0296944175,0.3392666578,-0.0802819729,0.2214090526,0.2769951224,0.0203873236,0.1460019499,0.4759078026,0.0976115316,0.3561215997,-0.0767005682,-0.0689726695,-0.5012392998,0.1398172379,-0.4139970839,0.0408438742,0.2894043028,0.1724706739,0.322507143,0.2895634472,0.0376227014,0.1403890252,-0.096874401,0.3843570948,-0.2270287573,-0.0645690262,0.1950525343,0.2118771374,0.1090166569,-0.2388573736,0.3231708705,0.2930503786,0.0205491371,-0.2603202462,0.3992149234,-0.1971804053,0.307482481,-0.0728120208,-0.1389238685,0.5194199085,0.2293463796,0.2214035243,-0.1439764053,0.0300036855,-0.0908123478,0.3211365044,0.1461285055,0.4942115247,-0.431876868,-0.2108418494,-0.0542695075,0.1428682655,0.0246400442,-0.0425273143,-0.1893388629,0.2584880292,-0.2203613669,-0.0148228314,0.0579342619,0.0337438472,0.0794112459,0.219566226,-0.2821283638,-0.4187386036,0.4235398769,-0.1683394015,-0.5067198277,0.1799818873,-0.2447522879,0.155995518,0.1381157935,-0.5674667358,-0.1494414061,0.2118353099,-0.1582861543,-0.2038371563,-0.0375708379,-0.0058429874,-0.1672918648,-0.1489332765,0.3547782898,-0.1051538885,-0.1671753228,-0.4202934504,-0.3254515827]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"Hi ! Thanks for reporting.\r\nWe're indeed using `jiwer` to compute the WER.\r\n\r\nMaybe instead of calling `jiwer.wer` once for all the preditions\/references we can compute the WER iteratively to avoid memory issues ? I'm not too familial with `jiwer` but this must be possible.\r\n\r\nCurrently the code to compute the WER is defined here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/metrics\/wer\/wer.py#L93-L94","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":56,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nHi ! Thanks for reporting.\r\nWe're indeed using `jiwer` to compute the WER.\r\n\r\nMaybe instead of calling `jiwer.wer` once for all the preditions\/references we can compute the WER iteratively to avoid memory issues ? I'm not too familial with `jiwer` but this must be possible.\r\n\r\nCurrently the code to compute the WER is defined here:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/metrics\/wer\/wer.py#L93-L94","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"Hi,\r\n\r\nI've just pushed a pull request that is related to this issue https:\/\/github.com\/huggingface\/datasets\/pull\/2169. It's not iterative, but it should avoid memory errors. It's based on the editdistance python library. An iterative implementation should be as easy as storing scores and words stepwise and dividing at the end. ","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":48,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nHi,\r\n\r\nI've just pushed a pull request that is related to this issue https:\/\/github.com\/huggingface\/datasets\/pull\/2169. It's not iterative, but it should avoid memory errors. It's based on the editdistance python library. An iterative implementation should be as easy as storing scores and words stepwise and dividing at the end. ","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"I see, this was solved by other thread. Ok, let me know if you want to switch the implementation for any reason :)","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":23,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nI see, this was solved by other thread. Ok, let me know if you want to switch the implementation for any reason :)","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"Thanks for diving into this anyway ^^'\r\nAs you said this actually got solved a few days ago","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":18,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nThanks for diving into this anyway ^^'\r\nAs you said this actually got solved a few days ago","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"Someone created an issue https:\/\/github.com\/jitsi\/jiwer\/issues\/40 at jiwer which shows that this is still a problem in the current version. Would be curious to figure out how this can be fixed by jiwer... :) I assume that it runs of out memory because it's trying to compute the WER over (too many) test samples?","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":53,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nSomeone created an issue https:\/\/github.com\/jitsi\/jiwer\/issues\/40 at jiwer which shows that this is still a problem in the current version. Would be curious to figure out how this can be fixed by jiwer... :) I assume that it runs of out memory because it's trying to compute the WER over (too many) test samples?","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"Hi !\r\n\r\nIt's computed iteratively so not sure what could go wrong\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8afd0ba8c27800a55ea69d9fcd702dc97d9c16d8\/metrics\/wer\/wer.py#L100-L106\r\n\r\n@NiklasHoltmeyer what version of `datasets` are you running ?\r\n","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":22,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nHi !\r\n\r\nIt's computed iteratively so not sure what could go wrong\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8afd0ba8c27800a55ea69d9fcd702dc97d9c16d8\/metrics\/wer\/wer.py#L100-L106\r\n\r\n@NiklasHoltmeyer what version of `datasets` are you running ?\r\n","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"One possible explanation might be that it is the user who is passing all the sentences in a single element to `wer.compute`?\r\n\r\nAs current implementation iterates over the elements of `predictions` and `references`, this can be problematic if `predictions` and `references` contain a single huge element each. \r\n\r\nThis could be the case, for example, with a single string with all sentences:\r\n```python\r\nresult[\"predicted\"] = \"One sentence. Other sentence.\"\r\n```\r\nor with a __double__ nested list of sentence lists\r\n```python\r\nresult[\"predicted\"] = [[ [\"One sentence.\"], [\"Other sentence\"] ]]\r\n```\r\n\r\nThe user should check the dimensions of the data structure passed to `predictions` and `references`.","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":103,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nOne possible explanation might be that it is the user who is passing all the sentences in a single element to `wer.compute`?\r\n\r\nAs current implementation iterates over the elements of `predictions` and `references`, this can be problematic if `predictions` and `references` contain a single huge element each. \r\n\r\nThis could be the case, for example, with a single string with all sentences:\r\n```python\r\nresult[\"predicted\"] = \"One sentence. Other sentence.\"\r\n```\r\nor with a __double__ nested list of sentence lists\r\n```python\r\nresult[\"predicted\"] = [[ [\"One sentence.\"], [\"Other sentence\"] ]]\r\n```\r\n\r\nThe user should check the dimensions of the data structure passed to `predictions` and `references`.","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"Hi all,\r\n\r\nin my case I was using and older version of datasets and, as @albertvillanova points out, passing the full list of sentences for the metric calculation. The problem was in the way jiwer implements WER, as it tries to compute WER for the full list at once instead of doing it element-wise. I think that with the latest implementation of datasets, or by using the alternative WER function that I've contributed on this [pull request](https:\/\/github.com\/huggingface\/datasets\/pull\/2169) there shouldn't be memory errors.","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":82,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nHi all,\r\n\r\nin my case I was using and older version of datasets and, as @albertvillanova points out, passing the full list of sentences for the metric calculation. The problem was in the way jiwer implements WER, as it tries to compute WER for the full list at once instead of doing it element-wise. I think that with the latest implementation of datasets, or by using the alternative WER function that I've contributed on this [pull request](https:\/\/github.com\/huggingface\/datasets\/pull\/2169) there shouldn't be memory errors.","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"@lhoestq i was using Datasets==1.5.0 with 1.6.1 it worked (atleast the first run) but 1.5.0 is not compatible with my preprocessing. i cant save my dataset to a parquet file while using the latest datasets version\r\n\r\n-> \r\n```\r\n File \"..\/preprocess_dataset.py\", line 132, in \r\n pq.write_table(train_dataset.data, f'{resampled_data_dir}\/{data_args.dataset_config_name}.train.parquet')\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/pyarrow\/parquet.py\", line 1674, in write_table\r\n writer.write_table(table, row_group_size=row_group_size)\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/pyarrow\/parquet.py\", line 588, in write_table\r\n self.writer.write_table(table, row_group_size=row_group_size)\r\nTypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got ConcatenationTable)\r\n``` \r\n\r\nif i do \r\n```\r\nimport pyarrow.parquet as pq\r\n...\r\n...\r\npq.write_table(train_dataset.data, 'train.parquet')\r\npq.write_table(eval_dataset.data, 'eval.parquet')\r\n```\r\n\r\nwhile using 1.6.1. and its working with 1.5.0\r\n","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":96,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\n@lhoestq i was using Datasets==1.5.0 with 1.6.1 it worked (atleast the first run) but 1.5.0 is not compatible with my preprocessing. i cant save my dataset to a parquet file while using the latest datasets version\r\n\r\n-> \r\n```\r\n File \"..\/preprocess_dataset.py\", line 132, in \r\n pq.write_table(train_dataset.data, f'{resampled_data_dir}\/{data_args.dataset_config_name}.train.parquet')\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/pyarrow\/parquet.py\", line 1674, in write_table\r\n writer.write_table(table, row_group_size=row_group_size)\r\n File \"\/usr\/local\/lib\/python3.8\/dist-packages\/pyarrow\/parquet.py\", line 588, in write_table\r\n self.writer.write_table(table, row_group_size=row_group_size)\r\nTypeError: Argument 'table' has incorrect type (expected pyarrow.lib.Table, got ConcatenationTable)\r\n``` \r\n\r\nif i do \r\n```\r\nimport pyarrow.parquet as pq\r\n...\r\n...\r\npq.write_table(train_dataset.data, 'train.parquet')\r\npq.write_table(eval_dataset.data, 'eval.parquet')\r\n```\r\n\r\nwhile using 1.6.1. and its working with 1.5.0\r\n","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"Hi ! You can pass dataset.data.table instead of dataset.data to pq.write_table","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":11,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nHi ! You can pass dataset.data.table instead of dataset.data to pq.write_table","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2078","title":"MemoryError when computing WER metric","comments":"This seems to be working so far! Thanks!","body":"Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n","comment_length":8,"text":"MemoryError when computing WER metric\nHi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:\r\n\r\n```\r\nwer = load_metric(\"wer\")\r\nprint(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n```\r\n\r\nHowever, I receive the following exception:\r\n\r\n`Traceback (most recent call last):\r\n File \"\/home\/diego\/IpGlobal\/wav2vec\/test_wav2vec.py\", line 51, in \r\n print(wer.compute(predictions=result[\"predicted\"], references=result[\"target\"]))\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 403, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/diego\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/wer\/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281\/wer.py\", line 94, in _compute\r\n return wer(references, predictions)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 81, in wer\r\n truth, hypothesis, truth_transform, hypothesis_transform, **kwargs\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 192, in compute_measures\r\n H, S, D, I = _get_operation_counts(truth, hypothesis)\r\n File \"\/home\/diego\/miniconda3\/envs\/wav2vec3.6\/lib\/python3.6\/site-packages\/jiwer\/measures.py\", line 273, in _get_operation_counts\r\n editops = Levenshtein.editops(source_string, destination_string)\r\nMemoryError`\r\n\r\nMy system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.\r\n\r\n\r\n\r\n\nThis seems to be working so far! Thanks!","embeddings":[0.1249791905,-0.2007298619,0.0587182119,0.3256509304,0.433393985,0.0737894103,-0.2180431932,0.2950879037,0.0288314931,0.4309869111,0.0965305641,-0.0946346819,-0.3058875501,-0.5752207041,-0.1799835265,-0.3347723186,-0.1305915564,-0.0603135042,-0.0939678624,0.267347157,-0.2414007336,0.2181451321,-0.3488659561,0.122934103,-0.0717325732,-0.0110546872,-0.1759239286,0.1331792176,-0.2212716192,-0.5667486191,0.3025929928,-0.3079388142,0.1904908121,0.3594332933,-0.000114279,0.0282555725,0.19010216,-0.178447023,-0.0109599112,0.1174560264,-0.1258537024,-0.1865432262,-0.1319401711,-0.0092576966,-0.0329881646,0.0375669524,0.1080688611,-0.2880086005,0.0632087141,0.2449464798,0.1992428005,-0.0609460808,0.2961238921,-0.2295078933,-0.1241091788,-0.1905775517,0.1977030337,0.3163279891,-0.2703140378,-0.1776358634,-0.2714266181,0.2580875754,-0.0427771099,0.2143146247,0.0936119035,0.1157601327,0.0853944719,0.0110054342,0.0847661495,-0.2555573285,0.2671687901,-0.3263044655,-0.0301647987,0.2048876584,-0.1614335924,-0.4275716245,0.1660306156,0.1187261567,-0.5047883391,0.0975702256,-0.2040266842,-0.1204812452,-0.394231081,0.1254809946,-0.0058269789,0.0910312086,-0.1965167075,0.1057423353,0.3368060887,0.1566442996,0.0604361072,0.2069692314,-0.1233543605,0.0125260502,-0.3743089437,-0.0881424993,-0.0584868453,-0.232619822,0.1580233574,-0.569531858,0.582660079,-0.0841271803,0.3524422348,0.4458707869,-0.2218324989,0.8173162937,-0.0972185805,0.0790163055,0.041408278,0.0784390643,-0.1152928844,-0.2438932359,-0.0898666158,-0.3839600384,0.2813354135,0.1884756237,-0.1235135794,-0.1992423385,-0.4400852621,0.0639662519,-0.0409351289,0.1174827218,0.1261956394,0.318808198,0.1651479155,0.489371866,0.3132435679,0.0720733255,-0.1783270389,0.0621534623,-0.2334481925,0.2319099456,-0.2242501229,0.0308982152,0.1252744943,0.2297951728,0.3616980016,-0.2322330773,0.2106768191,-0.2574244142,0.2766292393,-0.1106605157,0.0838679075,0.2026334405,-0.0995929539,-0.0087277889,0.229771018,0.2736266553,0.0345627628,0.1165985167,-0.212802887,-0.2432083488,0.3005083799,0.1580395401,-0.058649797,0.0679567158,-0.064557381,-0.0057821609,0.2805483043,-0.0548926443,-0.3401754797,0.0868661329,-0.7389823794,-0.2600598931,0.1528928429,0.0060362252,-0.1756756008,0.0054638353,0.2558612823,-0.0120836552,0.2422092408,0.0330554321,0.0680655092,0.1206671,-0.2044021338,-0.1056600213,0.3498846292,-0.8443564773,-0.1511173397,-0.0499686413,-0.1328816861,-0.0256584417,0.2559601665,-0.2146238387,0.1497683972,0.1131592616,0.3785301447,-0.2864328027,-0.1795535982,0.0241758507,-0.1898522526,-0.1139877066,0.179043591,-0.1413308531,0.3943657875,0.0892044976,0.1655692011,0.1407639533,0.2155860364,-0.027472429,0.1473459303,0.2670481503,0.2488005161,-0.6484080553,0.3060636222,-0.0628642514,0.1538829207,0.1872498542,-0.242123276,0.0660027862,0.4159049988,0.429313302,-0.1300992519,0.016114207,0.0612439737,-0.1678278446,0.2202239931,-0.0615801029,0.306425035,-0.2538882494,-0.0041135349,0.0725794286,-0.3613118827,-0.0806768984,-0.0067587825,-0.2327689081,-0.1088713482,-0.2220413536,0.2936443686,0.1907931715,0.2006856948,0.1718473136,-0.2917720675,0.350572139,0.1063389406,-0.0170480106,0.1874941736,0.1884122491,-0.1544378251,-0.4010769427,0.2174313664,-0.0282920189,0.2030489743,-0.118592605,0.2363728136,-0.125006333,0.0056329728,0.3543412387,0.0527254231,0.3139891028,0.1135732532,0.2762680948,0.1866981685,0.1431760937,0.3140142262,-0.4340707362,-0.2543807924,-0.389414072,0.3616285622,-0.3148328066,0.3210273087,0.1509761661,-0.143353641,0.4951443672,-0.1373810619,-0.1780490279,-0.0775060281,-0.3291467428,0.4842241108,0.0247815214,0.0627912804,0.1173731685,-0.0368315242,0.1472533494,0.1195161045,0.2113156915,-0.1094565839,-0.0626852959,0.2112149298,-0.4670906663,0.1594258845,0.1257491559,0.1019963622,0.3387947083,-0.2200201005,-0.1152894199,0.267416209,0.2180289924,-0.1320870072,-0.0435180217,-0.025865024,-0.2954534888,-0.204206109,0.3511810899,0.1040427685,0.1294223517,-0.0459422618,0.2327170968,0.2416962981,0.4276043475,0.2114255726,-0.0255014356,-0.181180343,0.0590603203,0.0218610112,-0.049709592,0.1463518292,0.0862486511,-0.0498690605,-0.0435718745,-0.1804246157,-0.0071563218,-0.0013990207,0.4104417562,0.2205947489,0.0727371052,-0.2366647571,-0.0322218463,0.1743239164,-0.2374216616,-0.038494654,0.5531134009,-0.0587114953,0.0892808437,-0.0565565489,-0.1715920269,-0.093634598,-0.396466732,-0.0207021777,0.0413863994,0.1471429765,-0.1181743443,0.1409462243,0.2246731222,0.1708787829,0.0079454714,-0.2062764615,-0.3394733369,-0.0268059596,-0.2306408435,-0.1545539051,0.0618929826,0.0176213887,0.6852975488,0.1380208135,-0.1405830383,-0.4560725093,0.0017372953,-0.2098908722,-0.030375788,-0.01608417,0.2644729018,-0.0626578182,-0.1569625586,-0.284212321,-0.0542895645,0.2124326974,0.1187750623,-0.0273834895,0.0317650139,0.0322173722,0.2830829322,0.5209565759,0.3953298926,0.1029953957,0.2838952839,0.192132026,0.0036794737,0.0900348872,-0.1075095385,0.1794606596,-0.1067251489,-0.1737584621,0.116028741,0.0703591853,-0.135174647,-0.1973445714,-0.0087810243,0.0846388116,-0.1840717643,-0.0214703195,-0.0188062899,0.5045292974,0.0923848003,0.0470348448,-0.1606875956,-0.3939412832,0.1025609076,-0.0459622853,0.067070663,0.0099263815,-0.0599402972,-0.0090090446,-0.7305296659,0.2685512602,0.112360552,0.432012707,-0.1020116284,0.0522880256,-0.1586075872,0.114267461,0.4286967814,-0.0644570366,-0.1765283048,0.0596214943,0.0524648316,-0.1683597118,0.1471431255,-0.1711662412,-0.0384163894,-0.2613116503,-0.1168813705,-0.0820594057,0.0472736172,-0.0062202667,0.31311813,-0.1601988524,-0.2500498891,-0.3472495973,-0.1220534667,0.0043929592,-0.2169685811,-0.122640349,0.089552626,-0.2176413387,0.2523514032,0.0227337591,-0.2952415049,0.3482629061,-0.1124061719,0.1965744197,-0.4322452247,0.1332802176,0.4644437432,-0.2072872519,-0.0170487668,0.1137063056,0.0695510805,-0.2817786634,0.2407572269,0.2017527968,0.3595784903,0.0675774738,-0.2912520468,0.2509188652,-0.0361825377,0.0853486732,-0.5771868229,0.1801706553,0.0886632651,0.061338678,0.1381756812,-0.0624361262,0.2326971143,0.0552651621,-0.1369003057,0.2892397344,-0.0195216201,-0.2696186602,0.2825178504,0.2487511635,0.7941431403,-0.1485889703,-0.0610655062,0.4458687901,0.0083947219,-0.0881056264,-0.2118185312,0.5530893207,-0.1516976953,-0.0146362381,0.0871364474,-0.2238284796,-0.0913898945,-0.2526904941,-0.260543555,0.6565314531,0.038912598,-0.1428082585,0.0865545571,-0.047439348,-0.1187873185,-0.2465846986,0.2222437114,0.1842269599,-0.1877066791,-0.1484005898,-0.0387452766,0.0764233768,0.0927681699,-0.3213992119,-0.2134746909,0.0094131725,-0.2341890186,0.1229038388,-0.287666589,0.1384741962,-0.0338191502,0.4129698873,0.3505011499,0.134784773,-0.0439514145,0.0642215833,-0.1409435868,0.0673821121,0.0816783533,0.1936235875,-0.0772474036,-0.0423675291,-0.2362020761,0.32398054,0.0318992175,-0.3035768867,0.0209789127,0.0041499967,-0.0676421225,-0.0927414224,-0.0991548374,-0.1226348728,-0.4054276943,-0.1016839817,0.1665976942,0.3358542025,-0.6273187995,0.3717970848,0.0599586554,-0.058107309,0.2039711773,0.3034738302,0.240200907,-0.3711039424,0.5228544474,-0.070308879,-0.06644997,-0.0849606693,-0.3364023268,-0.1057752147,0.0251229797,0.187585026,0.2309452891,0.0863209888,-0.2148944288,0.1295984238,0.0714637116,0.2860014737,0.1478189528,-0.3211325109,-0.5705017447,0.1984281689,-0.1343357861,0.0697695762,-0.4777343869,-0.2928742468,-0.543330133,0.0197104607,-0.2649724185,0.130941242,-0.4698491395,0.1528991461,-0.0560230874,-0.1072276384,0.2412145138,0.1662405729,0.1316618025,0.4861496985,-0.1609184742,-0.1148330122,0.0130821848,0.0763763413,-0.0238652155,-0.1037725359,0.2343620062,0.076625742,-0.1278423816,-0.1201295257,-0.1803638786,0.2723797262,-0.2416994125,0.0868961141,-0.3836637437,0.073159188,0.6001414061,0.2037330568,-0.0219237544,0.2448608577,0.061568249,0.2490887642,-0.1119033992,-0.0583461188,0.2134888768,0.1071450412,-0.1107831895,-0.2432857156,0.194062978,-0.2006400377,-0.2309135199,-0.1189450547,0.2250618041,-0.1795278043,-0.2697106004,0.0512615368,0.1457482874,0.2108913362,0.0513860211,0.178945452,-0.1412657499,0.0825064108,0.0582628436,0.2282669395,0.1724423617,-0.2710610926,-0.0541157089,-0.0430439189,0.2277816385,-0.3114750385,-0.1649622768,-0.248133108,-0.0783553347,-0.1232031584,-0.1370621324,-0.0019347653,0.3038545251,0.5290905237,-0.2573443949,0.4064506292,-0.4600455761,0.0767795295,-0.0782730356,-0.3895433843,-0.3259205222,0.3104047179,0.0393802598,0.3134308755,-0.3366586864,0.1221518219,0.1313456446,0.1683418006,-0.1511766165,-0.2036623955,-0.1160733849,0.0636430457,0.1578115076,-0.3073452413,0.0246311985,-0.0502912924,0.3668289781,-0.0342523418,0.3742722869,0.0016255801,-0.2236442864,0.2601143122,-0.01801279,0.5469641685,0.010053426,-0.283266902,0.1146143526,0.4181115031,0.1426426917,0.1386210769,0.284604013,0.4829080999,0.3588483334,-0.2767453492,0.2184025049,-0.1127006188,-0.1293811202,-0.0103219012,0.2846423388,0.2454683334,-0.0023273858,0.1850485057,0.0348168463,-0.0351406895,-0.1497602761,0.0099583296,0.0568038896,-0.2226731777,0.0570198521,-0.2153234184,0.0255853012,-0.2253831327,-0.0923115239,-0.3778279722,0.2275391817,0.0091682253,-0.1401443481,0.2049317956,0.0773804858,0.0560774691,-0.2662176788,0.3642232716,-0.0509135015,0.215163514,-0.616761744,-0.2342028469,-0.1764986515,0.1513807774,-0.1593509465,0.4129835069,-0.1397840381,0.0939858556,-0.1402538866,-0.0286629219,-0.0677687302,0.1230890229,-0.1035885215,0.2747887969,-0.3578975201,-0.2949352562,0.2261710316,-0.1421803236,0.2543468773,-0.4365845025,-0.1602471918,-0.233558774,0.010778361,-0.0389846973,0.4080353081,-0.0949812531,0.5220329762,0.0629382059,0.3354325294,0.5877146125,-0.0117938593,-0.2971838713,-0.1095950007,-0.1789064407,-0.2137201428,-0.4510128498,-0.0793379247,0.5294624567,-0.1782436669,-0.2955734134,-0.2922706902,0.3045113385,-0.0391488597,-0.0277270209,-0.1796133965,0.3239210844,-0.2037035972,0.2875931561,0.2503643334,0.0669568256,0.1562663615,0.412556082,-0.3301568329,-0.0906283706,0.584391892,-0.1609608233,-0.3511351049,0.0949686244,0.2776319385,0.197629407,-0.363720119,-0.8314566016,-0.0270131249,0.2877660692,0.2366388291,-0.102310881,0.0471825711,0.0729522929,-0.0134900166,0.2048307359,-0.0837528184,-0.10401804,-0.1719589531,0.1146996841,-0.1332051605]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2076","title":"Issue: Dataset download error","comments":"Hi @XuhuiZhou, thanks for reporting this issue. \r\n\r\nIndeed, the old links are no longer valid (404 Not Found error), and the script must be updated with the new links to Google Drive.","body":"The download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?","comment_length":32,"text":"Issue: Dataset download error\nThe download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?\nHi @XuhuiZhou, thanks for reporting this issue. \r\n\r\nIndeed, the old links are no longer valid (404 Not Found error), and the script must be updated with the new links to Google Drive.","embeddings":[-0.2337625921,0.1458084732,0.0028066062,-0.0309512708,-0.0274386685,0.0824488178,0.3759451807,0.2059576809,0.1537598073,0.0144035025,0.1848477721,-0.1723960787,0.233960107,0.2969171107,0.1645151675,-0.0703357011,0.112443462,0.039021574,-0.2078589052,0.0084466971,-0.3404218256,0.2083994001,-0.3360630274,0.0776542872,-0.1339364201,0.1082167551,-0.2342879176,0.204613924,0.0053033242,-0.5016649365,0.1856248826,0.2496987134,0.2551788092,0.2303705812,-0.0001261118,-0.0265719853,0.4359161556,-0.1366342902,-0.2173595279,0.099079974,-0.4538171589,-0.223021552,-0.1618050784,-0.330869019,0.2659609914,-0.1933172345,0.1802410781,0.3238401711,0.1627966315,0.5249042511,0.0954555199,0.2440396845,0.383268863,-0.017859133,0.0295927692,-0.1398885399,0.0056165019,0.2431164831,0.16717875,-0.1312993467,0.1934746653,0.22652632,-0.0200180989,0.0947979912,0.2447730601,0.0861292556,0.2219021469,-0.4831433594,0.2752977014,0.0740563795,0.9506683946,-0.3367395699,-0.395308286,0.0342546217,0.1537052989,-0.0347731821,0.2187415063,-0.0993440971,-0.0766191855,0.1927044094,-0.0719738603,-0.4454037249,-0.2641430497,0.2948839962,-0.0705403015,0.6482892632,0.2004719526,0.1531991214,0.1943139136,-0.0314076692,0.3032341003,0.1202326268,-0.1972546726,-0.0020126104,0.2240179926,-0.2257652134,-0.1425874531,0.1321183443,-0.0379861146,0.0900656357,-0.0158376172,-0.0804638788,-0.1313908845,0.0318026841,0.2977488637,-0.0259650797,-0.0639488399,-0.135734573,0.5959560871,0.4367472529,0.1359974444,-0.0484655611,0.2467383593,-0.5086395144,-0.0841770098,0.1293987483,0.3410683274,-0.0649032518,-0.240212664,-0.3356491625,0.0078357346,-0.1144976616,-0.1429990828,0.2463043928,-0.0248055141,0.0981618837,0.0490982644,0.1763306558,-0.0579827689,-0.2141543925,0.0657005459,0.1828427166,0.0734454319,0.0571352467,0.1977758557,0.2320502847,-0.0572202317,-0.0442006141,-0.3850712776,-0.213274017,-0.2779664695,0.0655522272,0.0705886409,-0.005324062,-0.0922143757,0.117328763,-0.0756690428,-0.1958306432,-0.1742605716,0.300824374,-0.4043316245,-0.1922464222,-0.1143398136,-0.0011621449,-0.2243248671,-0.3667369783,0.1075007841,0.020169178,-0.3128478825,-0.3899010122,-0.0386796892,-0.0872283652,-0.3737506866,-0.1814860851,0.1254688054,0.7528218031,-0.5478048325,-0.2115307301,-0.2793195844,-0.2170637697,0.0229437053,0.2886830568,-0.132589668,0.3553591073,-0.4054445326,0.1295455247,0.5153886676,-0.417206198,-0.7371981144,0.144679144,-0.0059493012,-0.1772410125,0.1121795177,0.185042128,0.1248529702,-0.2103365809,-0.4866012335,0.1159009188,-0.0161943007,0.0633499995,-0.1106215268,-0.5575144291,0.3404843807,0.0558741838,0.2634509802,0.23671031,0.461234808,-0.0207831375,0.4065787792,-0.0609179251,0.1000960618,0.3868402541,0.4776051342,0.0447238311,0.2310122401,0.0881084874,-0.413458854,0.1487097889,0.1035044715,0.0112760682,-0.3565673232,-0.0582169667,-0.4483054876,-0.1214280277,-0.2669527829,0.0149108171,-0.0321032815,0.0315877348,0.0605464466,0.0501000136,-0.1429137588,0.0495217741,-0.1035647839,-0.0192484818,0.0885358974,0.5486960411,-0.1757840812,-0.0942082703,0.1593255997,0.1271205097,0.2328132838,-0.2303649634,-0.1200081632,0.4031567574,-0.2489540577,0.4772265255,0.3295185268,0.307207197,0.1701678187,-0.4089676738,0.2946333289,0.4525774121,0.0123852631,0.0576634742,-0.1450582743,-0.1638094187,0.0704360977,0.2487749904,0.0597915947,0.0034618124,0.1770109385,-0.2544127405,0.0570459329,-0.1840036511,-0.3074975908,-0.1961553991,-0.0975589827,-0.2141680121,-0.0272730198,0.1944456697,0.548995018,-0.0981293172,0.2854127288,0.3537480831,-0.0131097604,-0.105992727,-0.1862697303,0.635764122,0.347435385,-0.0694110394,-0.1281581074,0.1623578817,0.2249321193,-0.1026633754,0.2864802778,0.1101239994,-0.2034563124,0.1419662237,-0.0807323828,-0.1086298153,-0.0259588137,0.0884374306,0.148977533,0.3354448676,-0.3748236597,-0.0133951707,-0.3511465788,-0.4899004996,-0.3558903933,0.4411506355,-0.2762396336,-0.306037426,-0.0208473057,0.3651280403,-0.1554075927,0.120349355,0.1482290775,0.3038351536,0.0550497174,-0.0216290802,-0.3250444531,-0.0932601988,-0.1645984352,-0.0629142374,0.367408365,0.0479612648,0.0950696021,-0.5916244388,-0.0839566216,-1.030079484,-0.0702777207,0.0633477494,0.1601599604,0.1411256939,0.2158608586,0.3985548615,0.1617077589,-0.041331239,0.1102820709,-0.2546848059,0.149924174,-0.2434190363,-0.1488886178,0.1329917908,0.0021189197,-0.1931252033,-0.5508337021,-0.1271964759,0.1204171851,-0.066043064,0.0839750245,-0.1132252291,0.2934363484,-0.2033239156,0.2035319656,-0.311720103,0.1289834827,0.3097099662,0.4342479408,-0.0575692765,-0.3968016803,0.2718307972,0.1163857132,-0.0004378126,-0.0436475314,-0.5142811537,-0.1939069182,0.0119605642,0.2472350001,-0.0632775873,0.3359144032,0.1272754073,0.1365230978,0.0328033231,0.007582542,-0.2230256647,-0.1242373288,0.3777466118,0.520142436,0.5347796082,0.4276016355,-0.1535900384,0.2914156318,-0.1811151057,0.0489291362,0.3592682481,-0.3136166632,0.2962454259,0.1019999981,-0.2116543204,0.159325853,-0.1649796963,0.0663417205,0.2345590144,-0.0905938372,-0.0390401408,-0.0410172381,-0.2618568242,-0.3001762629,-0.4504909217,-0.2143325955,0.5330833793,0.2515181601,0.2891176045,0.0485665612,-0.2502557933,-0.196738705,0.4842903614,0.3863710761,0.0330068879,0.1109360158,0.2658003271,-0.1529666036,-0.4952217042,0.293727994,0.1064501256,0.3704387546,-0.0387929901,0.1575803906,0.0535714142,-0.1242312863,0.5772590041,-0.2539204657,0.2622443736,0.0684625953,0.4507426322,0.23488985,-0.1190130189,-0.1213348582,0.2323836088,0.2770289183,0.0421762764,-0.0232233163,-0.279735893,0.463760823,0.3966563344,-0.2198312581,-0.166815117,0.0358488634,-0.4945954084,-0.5470474958,-0.1587898731,-0.0464199036,-0.1277840436,-0.2064092308,0.1471783817,0.2195531428,-0.4050371945,-0.1216311902,-0.0795788318,0.3472925723,0.3004160225,-0.1756514758,0.0963393897,-0.1135500297,0.0484544002,0.2051048726,0.0558822677,-0.0908671618,-0.0405820012,-0.2574214339,-0.098509267,0.3127676845,-0.097044982,0.0623672418,0.0962246805,0.0905657262,-0.0373131186,-0.0823157728,0.4640595019,-0.1007034704,-0.3637642264,-0.3087791204,0.3928531706,-0.1518965214,-0.1130243614,0.3743221164,0.2573768198,-0.2176190913,-0.121286124,-0.1908537298,0.7290430069,0.1189942211,0.0608185902,0.2324613035,-0.1031749174,0.3928862512,0.0442273431,0.1866857708,-0.3072916865,0.0135657126,-0.1269171834,0.0699345097,-0.2178967893,0.335835278,-0.139681831,0.5958755612,0.0264242236,0.0196546372,0.1223934069,0.4716734588,-0.1609157324,0.1347672194,0.0217925087,0.0155310268,0.0058717369,0.3723073602,-0.2040599883,-0.1112762913,0.1404046863,-0.214795664,-0.5518468022,0.0312192421,-0.2196572125,0.258117646,-0.1295246482,-0.2145423591,0.2832951546,0.2206097096,0.2376624197,0.0511691049,-0.3972946107,0.3764607012,-0.1678469926,-0.1342575252,-0.0164703187,0.4251862764,-0.0932804495,-0.1290362924,-0.5057976246,-0.0665257946,-0.0633844659,-0.4335307479,0.0353138633,-0.1436430812,0.2073921561,0.0223262273,-0.2601278424,-0.1609869152,-0.2571820319,-0.3202683628,0.0321980678,-0.1532745808,0.2593232691,0.2261019498,-0.3166197538,-0.099711515,0.022772979,0.3828011453,-0.3625801206,-0.1783670336,0.7151678205,0.2753103673,0.0341886394,-0.2545825839,-0.1098653004,-0.2667811215,-0.5026059151,0.2774845064,0.0283320546,0.2976675928,-0.1477482021,0.3914981484,-0.0281556733,0.3012159765,-0.3247286379,-0.4623773396,-0.1138147861,0.1630941182,0.3029043078,0.2098998576,-0.1154576093,-0.120895192,0.2744278014,-0.7727934122,-0.1710737497,0.1700761914,0.1075908691,-0.1491922587,0.6299343109,0.2635003626,0.1223271564,0.0482027456,-0.0152934501,-0.0889600366,0.1578667611,0.0107665509,-0.1816475838,0.1998369098,-0.1018725559,-0.3029541075,-0.04351262,-0.3821090758,-0.0914798677,0.0892080143,-0.0704935789,0.1908294857,-0.0954103395,0.1595570147,0.2407404482,-0.2178866565,-0.15271312,0.1699919105,0.15817523,0.6770749092,0.0224990807,0.5906766653,0.0715453774,0.2081637979,0.1268262565,-0.02721099,-0.0118105346,-0.0902829915,0.2888884544,-0.1499811709,0.0181558356,0.1129631624,0.0499144308,0.5606864691,-0.3244295716,-0.138892144,0.1753303707,0.0103010554,-0.3160126507,-0.2641210258,0.0585895851,-0.1067053601,-0.0889444873,0.0581992976,0.2003952116,-0.1789781451,0.2719350159,0.1372804791,0.5460570455,-0.2183980197,-0.002535871,0.1831000447,0.0839236304,0.0638807938,-0.1288947165,-0.0195838287,0.1315480769,0.2045430839,-0.2031111568,-0.1803331822,0.0513298251,0.2913492918,-0.1182782724,-0.3024699986,0.0350259021,-0.175696373,-0.0054637394,-0.0331741534,0.1636323929,-0.0354048088,0.0091173043,0.0176289082,-0.2669948041,0.3506841063,0.3031772673,-0.1372591853,-0.0752670914,-0.2641394436,-0.1847486645,0.152606979,-0.1008404568,-0.1865114421,-0.1494007409,0.1171884313,-0.0281910766,0.0020575945,0.039711602,-0.1866705865,0.3073962331,-0.2139272243,0.0651060045,0.1143329144,0.0419491827,0.2495668828,0.3084132671,0.3976809084,0.2047304213,0.1932731271,0.0353641137,-0.0516659617,-0.1282121539,0.1300673038,-0.0342163481,0.0622274652,0.0144375591,0.1851664484,0.0652069896,0.0129532581,-0.2733806372,0.0699006543,0.0067428024,-0.0596716516,0.1353982538,-0.1359250396,-0.175293982,-0.3913387954,0.1247338653,-0.2187499255,-0.2254521251,0.2953235209,-0.0370717794,0.0295266621,0.0159367137,-0.0137184765,-0.0635186881,0.3564924598,0.3876615167,0.2662048936,-0.2750079334,-0.31626755,-0.6508069634,0.0528044291,-0.0115180528,-0.0463806354,-0.0717248395,0.034612529,-0.47316733,-0.2058355212,0.6622774005,0.3213905096,0.0657564178,0.0941253155,-0.0088430112,-0.2726204097,-0.0666335076,0.4215420485,-0.0503575429,-0.3091487885,0.1369160861,-0.2878558338,-0.1207707822,0.0265603364,-0.1604083329,-0.0896119103,-0.1639439315,0.4298087656,-0.0328950547,0.3747119308,-0.1414086372,-0.2511027753,-0.475607574,-0.305354923,-0.2396209389,0.2357245237,-0.064679347,0.0401484855,0.0480124205,0.3212934136,-0.4206643105,0.1166842431,-0.1158884764,0.2996494174,-0.1811911315,0.0739011094,-0.0550208129,0.127778247,0.3625876009,-0.1197771505,-0.1424405724,0.0054815561,-0.3037088215,-0.0694786236,0.3491157591,0.0033644184,-0.1348211616,-0.1744971722,0.3634095192,-0.0651312321,-0.0328346081,-0.2893856168,0.0206731595,0.1345507503,0.290976733,0.0081329579,-0.010698352,-0.0847701505,0.0984795019,0.1847158819,-0.0131015386,0.2195804268,-0.3208527565,-0.077685684,-0.0505721793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2076","title":"Issue: Dataset download error","comments":"It would be nice to update the urls indeed !\r\n\r\nTo do this, you just need to replace the urls in `iwslt2017.py` and then update the dataset_infos.json file with\r\n```\r\ndatasets-cli test .\/datasets\/iwslt2017 --all_configs --save_infos --ignore_verifications\r\n```","body":"The download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?","comment_length":37,"text":"Issue: Dataset download error\nThe download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?\nIt would be nice to update the urls indeed !\r\n\r\nTo do this, you just need to replace the urls in `iwslt2017.py` and then update the dataset_infos.json file with\r\n```\r\ndatasets-cli test .\/datasets\/iwslt2017 --all_configs --save_infos --ignore_verifications\r\n```","embeddings":[-0.2658804059,0.1200055629,-0.0836954564,-0.1060316935,-0.0561511554,0.0131992185,0.1135270521,0.3138656914,0.1883039474,-0.066574432,0.0962729752,0.0473739319,0.2630379796,0.2957364917,0.0410313308,0.0087639913,0.0977707878,-0.012059615,-0.283359617,0.0477566421,-0.4010069668,0.185447976,-0.3126529157,-0.0935631618,-0.0601683185,0.1259154081,-0.3016757965,0.1154684275,-0.028102709,-0.6263319254,0.272007972,0.3913546503,0.316203177,0.1880385131,-0.0001158188,-0.1218974292,0.3769318461,-0.136982739,-0.2984445393,0.0435839929,-0.6287513971,-0.3450363576,-0.1360737532,-0.2700529695,0.22668235,-0.3958939314,0.0692494288,0.2054876685,0.1902896464,0.5130621195,0.1835779846,0.3192787766,0.2641099989,-0.0636398941,-0.059832979,-0.0225952119,0.0009521026,0.0705709308,0.2730740309,-0.1185739487,0.2492121458,0.2715386748,-0.0152353542,0.010092359,0.2788783312,-0.0789464712,0.2402883321,-0.4742940366,0.181489706,-0.0224352125,0.8869271278,-0.3021595776,-0.4625476003,0.1310071498,0.0809325352,-0.0163424946,0.0607640631,-0.079737246,-0.1218894497,0.237042442,-0.172701776,-0.4722522497,-0.2464633584,0.3339924812,0.1414072663,0.6231669188,0.1615864336,0.1888065934,0.1574316919,-0.1041223854,0.2951930165,0.0304806884,-0.2320180684,0.0254935026,-0.007402264,-0.2957542837,-0.0889699832,0.0468946323,0.0084476089,0.2330870628,-0.0686246678,0.0362836011,-0.1587981135,-0.0047766604,0.2110501528,0.0406269357,0.1961813718,-0.0088387365,0.4371440411,0.6046849489,0.1306504011,-0.0418736488,0.2024896741,-0.3794595599,-0.2645076215,0.1908553988,0.2263668478,-0.1044933274,-0.1037598774,-0.2227149606,0.0441980623,-0.1268343776,-0.0476008616,0.2771447599,-0.1958618313,0.2930210531,0.1614924967,0.1268990338,-0.022056086,-0.2848353088,-0.0018294775,0.0769977197,0.0430149138,0.1160066724,0.1961963475,0.3151423633,0.1433610618,-0.0962003246,-0.3094979525,-0.1295165867,-0.1571160406,0.2209687084,0.1312751323,0.0491552278,-0.1215953007,0.2386569381,-0.0439963043,-0.0227940101,-0.1235121638,0.2856395245,-0.2518401444,-0.2053989768,-0.1939584911,0.0960125476,-0.1884029508,-0.3877756894,0.0536033362,-0.0562488288,-0.2581742704,-0.3503612876,0.014182834,-0.0462336726,-0.3027746379,-0.1621436626,0.1841401756,0.5909867287,-0.456173867,-0.1229094043,-0.1829642206,-0.2061738223,0.0710778534,0.1980037689,-0.1920287311,0.1715581268,-0.1458818614,0.0955499709,0.5259150863,-0.4413202703,-0.5953449607,0.2981191576,0.0802383572,-0.1355934739,0.0588275492,0.2166193724,0.0263025165,-0.2233210504,-0.5292845368,0.1446314603,-0.0034890403,0.1147148758,-0.2648130655,-0.5547558069,0.2628459632,0.0557124242,0.0934379548,0.1474309415,0.2381829172,0.1280943602,0.5918655396,-0.088500008,0.1206206903,0.2447381616,0.4602724612,0.218875289,0.2180595845,0.1307167709,-0.3153165877,0.1767422557,0.10156174,0.0456397422,-0.3460792005,-0.1567668319,-0.4196773767,-0.0620038807,-0.274181664,-0.2161063701,0.0659201071,0.1583601385,0.1244073957,-0.1213255227,-0.2550233006,0.169154942,-0.2095880359,-0.0120715657,-0.0166058764,0.4373683929,-0.1722989827,0.0374265909,0.0834373087,0.0616313852,0.180836767,-0.2504228652,-0.2579286098,0.6097425818,-0.2146878242,0.2997120023,0.4298552871,0.1492438912,0.1207442582,-0.3250736296,0.2012681216,0.4072022438,-0.00894388,0.1707870364,-0.0672066137,-0.1860921532,-0.0213031191,0.4060423672,0.0932738408,-0.0004359893,0.1702741832,-0.2249174416,-0.0809182078,-0.2327186316,-0.2935729325,-0.1963057667,-0.258728534,-0.1721239239,0.0489456989,0.2477435172,0.4090799391,-0.0953668058,0.0576708876,0.3085025251,0.0654459223,-0.0970147103,-0.1172316298,0.6027164459,0.3352910876,-0.0526720621,-0.2476411462,0.1453818977,0.216224879,-0.0910260081,0.2975569069,0.13904351,-0.0856266543,0.0401939154,-0.048189871,-0.0874652043,0.0182938781,-0.0237715133,0.1500088274,0.3771407902,-0.287627548,-0.0323111676,-0.3990878165,-0.5441100001,-0.2596680522,0.2256843001,-0.1787061989,-0.3382890522,0.0340355709,0.2766568363,-0.1662930101,0.2912934124,0.0830879062,0.2072326839,-0.0298417266,0.1257413179,-0.2831582427,-0.1041707322,-0.1987837255,-0.0051610041,0.3439479768,0.0879269093,0.1258616,-0.5627509952,-0.1183171272,-0.8395566344,-0.0598167367,-0.1080491394,0.1796332449,0.1123823896,0.2555810213,0.4747670889,0.2410743684,-0.1111321226,0.225404188,-0.1684920192,0.231316179,-0.3595496416,-0.0888079032,0.052810587,-0.031606216,-0.3374100924,-0.4721329808,-0.2592510581,0.099806428,-0.088309899,0.1336279213,-0.0836938545,0.3151398897,-0.2561671734,0.0332767144,-0.2438776344,0.0764196813,0.1267900318,0.3804925382,-0.1830556542,-0.4039416015,0.1734005958,0.1466891617,0.1630404294,-0.1210628375,-0.4460037649,-0.2127233148,0.0524972901,0.1790835857,-0.0879116654,0.3145512342,0.0866703689,0.0773892254,-0.0600989312,-0.0604950711,-0.1283682138,-0.142270565,0.2520179451,0.6086521149,0.3990364373,0.3653602898,-0.014084368,0.275329113,-0.0019119915,-0.0331550017,0.2885213792,-0.2751230896,0.2634461522,0.1812659353,-0.3473004699,0.0599865802,-0.1347781569,0.1323149204,0.1809362024,-0.0634683967,0.0471940078,-0.0834674165,-0.2222916037,-0.206291616,-0.3979876041,-0.1004896238,0.4765738547,0.2636371553,0.3525409698,0.0868991092,-0.3535329998,-0.1067998633,0.4027781785,0.3336628377,-0.0035476268,0.0683814138,0.0228772778,-0.1878342628,-0.3891737759,0.3738915622,0.1154217348,0.3910482824,0.1219120026,0.0870748311,0.0079187248,-0.1195583045,0.6106877923,-0.4253914654,0.1043016687,0.0488901101,0.4268633127,-0.1118371487,-0.141884461,-0.0411707982,0.2548780739,0.2616054714,0.0971639752,0.0374436416,-0.3399624527,0.3474550843,0.3804398775,-0.2724185884,-0.1804115325,0.1298931241,-0.4039521813,-0.6693725586,-0.2616961896,-0.1312411427,-0.118831493,-0.1276838332,0.1464577168,0.334253788,-0.1344384849,-0.1007191762,0.1044344753,0.1917995065,0.5774642229,-0.084269926,-0.0353934914,-0.1307768822,0.1626767814,0.1719180793,-0.0485261083,-0.1981222332,-0.2173077613,-0.121255748,0.031681709,0.2801321447,-0.0337842144,0.0714719743,0.1753660887,0.181717962,0.0264157318,0.1469887495,0.5127022266,-0.1981015354,-0.2774859071,-0.5861621499,0.3841053247,-0.1813444495,-0.0958093926,0.3464315534,0.0981626287,-0.1492699236,-0.012950046,0.0048457063,0.6299356818,0.2925108671,0.0522353984,0.2048249841,-0.0285992883,0.5659810901,0.0059439344,0.1267868876,-0.3390862346,-0.0320802927,-0.0287112799,0.1228825822,-0.1193818823,0.2378741503,-0.1434058249,0.5463705063,-0.0078887828,-0.0930928588,0.1464381069,0.4880785942,-0.1445132047,-0.023282297,-0.1062289104,0.1316646188,-0.0481892526,0.3345883191,-0.0951205716,-0.1221299991,0.0724357739,-0.1510442346,-0.4289672077,0.1074736491,-0.3221225441,0.2852962017,-0.1269459426,0.1787621379,0.2097796351,0.2306780666,0.2053614259,0.1746482104,-0.2870624959,0.3699127138,-0.0707248151,-0.1091891304,0.0266302656,0.3230379224,-0.0475368537,-0.2259249389,-0.5937906504,-0.0951295942,0.0163208544,-0.290142715,0.1625813991,-0.2702561021,0.370438993,-0.1117243916,-0.2816072106,-0.0553289764,-0.2641737461,-0.2904844582,0.0957321823,-0.2068888694,0.0748639852,0.342956692,-0.2037488371,-0.0831960216,0.0226183999,0.3929964006,-0.3618119657,-0.1502799541,0.651270628,0.2463868856,0.1058226079,-0.3233752549,-0.0739864632,-0.2134781778,-0.5588194132,0.1342839003,0.1284735799,0.2565954328,-0.1076334044,0.5815694332,0.0690356418,0.3135823011,-0.2030031085,-0.440582335,-0.177262947,0.160017103,0.3316250145,0.2019181848,-0.2039653063,-0.1100357473,0.1043673232,-0.6226142645,-0.2635371387,0.0713726953,0.0797323287,-0.0830932334,0.5312469006,0.2811519802,0.2359324992,0.024613535,0.0306046754,-0.1111804023,0.0029793677,-0.0818335116,-0.295236975,0.1433678865,-0.00780536,-0.370046705,-0.0679021031,-0.3382742405,-0.1110147312,0.0122495024,-0.2359367162,0.1323663741,-0.0991359949,0.3274615407,0.1289871037,-0.2455887645,-0.0798086151,0.2245870978,0.2456071079,0.6195127964,0.089369446,0.512522161,0.0343835577,0.0344536081,0.190352574,0.0126583558,-0.1068076715,0.0401704535,0.2409957498,-0.0635921434,0.1247478873,0.2878764272,0.1464390606,0.723521471,-0.1657293439,-0.1074032485,0.1783136278,0.1075240672,-0.2422147393,-0.1480084658,0.1144230738,-0.0603429303,-0.1510429233,0.0507802702,0.1459880322,-0.2648043931,0.3208700418,0.1363300383,0.4105603993,-0.2091365457,0.1390137821,0.2326114923,-0.0402198695,0.0106300795,-0.2607223094,-0.0165249947,0.1434255391,0.0729609504,-0.25908494,0.0021662659,0.1359721273,0.2214528769,-0.1256228834,-0.2092448771,0.1244166195,-0.3079873621,0.0531322323,-0.0780405998,0.1538690776,0.0930997953,0.0630085841,-0.0837696046,-0.1868067831,0.4469357729,0.3637336791,-0.228459388,-0.3311855197,-0.2089336514,-0.2458161861,-0.0108435126,-0.1313483715,-0.2673428655,-0.0420363136,0.0208687466,-0.065123558,-0.2304664552,-0.079535529,-0.1641403586,0.3442992866,-0.1986873448,-0.0340694822,0.3539015055,0.019296905,0.2935782671,0.166208297,0.389013052,0.3031672239,0.2825599015,0.1341412812,-0.0747472197,-0.0764380172,0.1327649206,-0.0515594408,0.0485227928,-0.0168992002,0.2257482409,0.1243711263,-0.0633891225,-0.2191685289,-0.0188533459,0.1339443177,0.0066306512,0.2588457763,0.0274210609,-0.0982638523,-0.379743576,0.1497978419,-0.2567159534,-0.2685228586,0.2529696822,-0.0581978075,0.0754309967,-0.0542057753,0.0456067733,-0.2022539526,0.5885548592,0.2964634895,-0.0175194535,-0.2498063892,-0.3237139881,-0.669228971,-0.0296268966,0.110742785,-0.1353233904,-0.158965975,0.040637482,-0.3685365617,-0.1781421751,0.4535003901,0.0676384941,0.0767980814,0.2088019848,-0.0742544457,-0.1603699923,-0.142530039,0.5685743093,0.1327405721,-0.2423026413,0.0843377858,-0.375014782,-0.0309346858,0.0248885863,-0.2898807824,0.0856143758,-0.2652178705,0.3958436251,0.0156826712,0.3200854957,-0.2106956989,-0.2838002741,-0.5346634984,-0.3248725832,-0.250556618,0.2824266255,-0.0468877815,0.1986866891,0.0551878773,0.3310994804,-0.4635562897,-0.0262530297,-0.1626522541,0.0588616356,-0.1297528446,0.0431794152,0.0607058592,0.0616324283,0.3075373471,-0.0734092742,-0.1837181896,0.0640982091,-0.2904723287,-0.1000133008,0.6150925159,-0.0545787737,-0.1901325434,-0.1768676788,0.2843328416,-0.1528607607,-0.1110519767,-0.2704383433,0.0030383815,0.1568035483,0.3058632314,-0.024373712,0.1428263187,-0.1586352587,0.1243697926,0.0833775029,-0.0151752634,0.1502655447,-0.3417594731,-0.1232422814,-0.0235450026]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2076","title":"Issue: Dataset download error","comments":"Is this a command to update my local files or fix the file Github repo in general? (I am not so familiar with the datasets-cli command here)\r\n\r\nI also took a brief look at the **Sharing your dataset** section, looks like I could fix that locally and push it to the repo? I guess we are \"canonical\" category?","body":"The download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?","comment_length":58,"text":"Issue: Dataset download error\nThe download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?\nIs this a command to update my local files or fix the file Github repo in general? (I am not so familiar with the datasets-cli command here)\r\n\r\nI also took a brief look at the **Sharing your dataset** section, looks like I could fix that locally and push it to the repo? I guess we are \"canonical\" category?","embeddings":[-0.328745544,0.2278453708,0.017168263,-0.1485101581,-0.0278213564,0.1295360625,0.2159447074,0.4029399455,0.1252267361,-0.0533371381,0.1292182505,-0.1483832747,0.3661812544,0.1626039147,0.0930708423,0.0371344499,0.0882588997,0.1164590269,-0.1444546878,-0.1393209249,-0.2735645771,0.0519362874,-0.1136716679,0.0818874985,-0.0942528695,0.0077920412,-0.2059002221,0.1530778706,-0.1278963983,-0.4773188829,0.3734855652,0.3872337639,0.2581133246,0.1217223108,-0.0001211941,0.0353706032,0.4054268897,-0.0881666243,-0.3705869913,0.0765684843,-0.5739338994,-0.4313016832,-0.0611455031,-0.2580585182,0.1887960732,-0.1298538148,0.2098872662,0.0811502114,0.1162736267,0.4262779951,0.1302455664,0.2230622321,0.3262282014,-0.0737565309,0.1489093006,0.1681558192,-0.0132277319,0.3721017838,0.4372789264,-0.1092656925,0.0860342085,0.1534002274,-0.0840912983,0.2124538869,0.3541404605,0.049540963,0.4574348629,-0.404850632,0.2234413922,0.032490734,0.8967596889,-0.4227740765,-0.3420685828,0.087257579,0.0095371474,-0.0898228288,0.1586155295,0.0859672278,-0.2002848536,0.2778534293,-0.1804689765,-0.3604149222,-0.1591273397,0.3106373549,-0.0873087049,0.5873485804,0.1010470092,0.2082229257,0.1031948328,-0.057155557,-0.127402395,0.0459690653,-0.2225729376,0.0741773844,0.1182245016,-0.2077798098,-0.1212854013,-0.1025085747,0.1066649109,0.1532485634,-0.1984604448,-0.0369431339,-0.1734641045,0.0718769804,0.403215915,-0.0381648429,0.0457206406,0.1548082232,0.5666858554,0.4230882525,0.0112662762,-0.0179061182,0.1470717341,-0.2979692519,-0.0977271646,0.1550770104,0.4336481392,-0.3121049702,-0.2366575152,-0.1028786153,0.1271014214,-0.1144597977,-0.0768235996,0.1278249919,-0.1613834947,0.2357040942,-0.1247183383,0.0475263856,-0.1026306227,-0.3002571762,0.0427250899,-0.0334792435,0.1329357326,0.1183083653,0.0120964209,0.0708040446,0.0517747104,-0.0831964761,-0.0121604884,-0.179164499,-0.2284108847,0.1644670367,-0.1286881268,0.2666129172,-0.2245610952,0.2676091492,0.1448450983,-0.3232682049,-0.2828286588,0.2915723324,-0.288251549,-0.3567067087,-0.3281010687,0.0365198739,-0.0712554827,-0.1758402437,-0.0324564986,-0.2105956227,-0.1574168056,-0.2907342017,0.0741475895,0.0098834308,-0.1292327195,-0.1871705204,0.1953063756,0.5042909384,-0.6066138744,-0.1614392996,-0.2313724458,-0.081650801,0.2087162137,0.3378128409,-0.2769352496,0.1699398756,-0.1460845619,0.0277961008,0.5927787423,-0.5152544379,-0.6871725321,0.2041898519,-0.0130712017,-0.2299176455,-0.0214091223,0.2520213723,0.1231497452,-0.3452990651,-0.4743498862,0.0846300796,-0.1290565133,0.0995862931,-0.2955790758,-0.5416894555,0.1625074148,0.1816478223,0.0469318554,0.1821604222,0.3992836177,0.1766407937,0.3272173405,0.0588194765,0.2011160702,0.2115599811,0.5126587749,0.1314488947,0.1533075869,0.2232039869,-0.2943217158,0.146966368,0.2489130944,0.0949624106,-0.3220586181,-0.2608608007,-0.4123025239,-0.1020810902,-0.3485299647,-0.0909804702,-0.0179825611,0.1807579845,0.1164785549,-0.1103476584,-0.3459634185,0.3093793094,-0.476324141,0.0745256543,0.2337314487,0.5755878091,-0.1695272624,-0.019681856,0.0520269014,0.1301290393,0.0639668852,-0.2785018384,-0.2143508196,0.5593360662,-0.3127242923,0.234146744,0.4399033189,0.3117654622,0.1940223128,-0.3202221096,0.1119112819,0.2882171273,-0.022600593,0.1490286142,-0.1741977036,-0.1989568919,0.0576519929,0.1708124727,-0.0655332804,0.0031271211,0.2749859393,-0.0832780227,-0.1410722733,-0.1622516066,-0.2921606898,-0.0250303149,-0.0415077321,-0.045363903,-0.0604689419,0.1785668433,0.5440226793,-0.1238421276,0.1165739074,0.3003477454,0.0952580795,-0.0596884862,-0.1099886373,0.5564720035,0.3104226589,-0.1512852013,-0.1697158962,0.1790992916,0.0578599684,-0.0373099893,0.0465996191,0.1925277859,-0.1200713366,0.0109616425,-0.046825774,-0.0372149833,0.1695201844,0.1583189517,0.1474125087,0.4839606285,-0.2462186962,-0.0133018671,-0.3664867878,-0.5353518724,-0.3251509666,0.3189966977,-0.379027307,-0.1859087646,-0.042106688,0.1787430197,-0.210716486,0.19015719,0.2043239027,0.3172184527,-0.0788311958,0.00094445,-0.1719424427,-0.2316943556,-0.3306145072,-0.0519320183,0.4443805516,-0.0417089462,0.1766371727,-0.6509737372,-0.1097551808,-1.0360774994,-0.016830612,-0.0461155251,0.2316710204,0.1153671294,0.1593004912,0.4537458718,-0.0163831022,-0.170742467,0.2607427239,-0.0401942916,0.2149461806,-0.3174853027,-0.1729994863,0.0167532824,0.04561349,-0.1350577325,-0.5641142726,-0.3411818147,0.1600964665,0.0241009574,0.1449770927,-0.0509621575,0.043548774,-0.1906140596,0.1358193606,-0.2904105186,0.1560662836,-0.010705228,0.2633077204,-0.1443635225,-0.4014081657,0.1329776198,0.3028902113,0.0468480401,0.0664788932,-0.5709301829,-0.2596457303,0.0302447546,0.4149636924,0.1201115176,0.2456101626,0.1748134196,-0.0011636629,0.0383690484,0.0454277806,-0.1914303899,-0.0953515768,0.5123045444,0.3417057693,0.4293990731,0.3754684925,-0.1085308641,0.5380147696,0.062819846,-0.1203672141,0.2591626644,-0.1626125574,0.3061215878,-0.0002962946,-0.2931819856,0.1956028342,-0.1550560445,-0.0657005534,0.4099546969,0.089260295,-0.0365754105,-0.0739256591,-0.1810676605,-0.2698777616,-0.4921988547,-0.1627943963,0.459923327,0.2576901019,0.2941862643,-0.1377737671,-0.2924903333,-0.1697051674,0.402410537,0.6239661574,-0.123548314,0.2065495253,0.1147285998,-0.1757795811,-0.1004421115,0.2647216022,0.1980748475,0.2662751973,0.0342544019,0.079440169,0.100839287,-0.1124488562,0.4005718827,-0.2048437297,0.1391569525,0.1369843781,0.4236280024,0.1360696554,-0.2744187415,0.0207302626,0.4552538097,0.2049041688,0.0759631544,-0.0221567024,-0.3152551353,0.3411841691,0.3714493215,-0.2002499849,0.0219797641,0.1467153728,-0.402273953,-0.4977632463,-0.2266691774,-0.1195310429,-0.2320986092,-0.0722359344,0.1765212119,0.2332923412,-0.1576328427,-0.1536704451,0.0030541779,0.2076951712,0.4947226346,-0.0198297538,-0.0089819329,0.0129314493,0.090997979,0.3013310134,0.070053421,-0.3510884345,-0.1641141176,-0.3205774724,-0.0949708521,0.2082443833,0.1176925525,-0.082730934,0.1448273361,0.0616363809,-0.0875911266,0.1192067116,0.3584291935,-0.084957771,-0.5300603509,-0.5256214142,0.4085562527,-0.1199645028,-0.1654942036,0.4003600776,0.15636383,-0.0323150456,-0.2735866904,0.0830515549,0.7465671301,-0.0229257774,0.0327060595,-0.0075633847,-0.091631785,0.6845269203,-0.3556339443,0.0803623423,-0.2357023507,0.0443740897,-0.1400306523,0.0937241614,-0.0345785841,0.1248346716,-0.252768904,0.3953990936,0.0111263413,0.128126502,0.167803809,0.1267899871,-0.0315408893,0.0058416454,-0.0743877441,0.083097972,-0.16223903,0.4866451025,-0.16987966,0.0068917116,0.0124568976,-0.1154112145,-0.5681362152,-0.0628740788,-0.1664634347,0.2920018137,-0.0299787074,0.0210064072,0.0415383168,0.0992110744,0.4292318225,0.1321552098,-0.2016576976,0.312969178,-0.0059430469,0.0601131096,0.0004794372,0.3337996006,0.1579629928,-0.2422535419,-0.3779408038,-0.1441570222,-0.134275496,-0.383428812,0.0235967506,-0.2298964858,0.6278172135,0.0004668927,-0.0304869302,-0.0596395768,-0.1215836704,-0.2231877297,0.0184048526,-0.1535338163,0.1416375786,0.0816714764,-0.2914767563,-0.2244675905,0.019166423,0.1895216554,-0.1399022937,-0.2665092051,0.4547466338,0.1219027713,0.0639430061,-0.1995392144,-0.0725483894,-0.1438465565,-0.6568563581,0.1191087812,-0.0967697352,0.1828152984,-0.1090995669,0.5960338712,0.2512530386,0.167328164,-0.2896704078,-0.4213556647,-0.2715561092,-0.0532644875,0.3364647329,0.1554660946,-0.0176154356,-0.1649508625,0.1936541647,-0.6747892499,-0.1614928544,0.0090757329,-0.1692629158,0.0296201333,0.6704883575,0.4356498718,0.0869972706,0.0188518222,-0.0612277612,-0.1750660241,-0.081286341,-0.0256924983,-0.132248655,0.200679794,-0.0557944737,-0.4399529696,0.050034225,-0.3374831975,0.0406050645,0.0572873801,-0.1088784337,0.0315803476,-0.0589899905,0.2244446278,0.3297688365,-0.0960424617,-0.1266285926,0.2370405048,0.2353165299,0.6963256001,0.0479421355,0.5046027899,0.1246773601,0.1058540046,0.097295627,0.0786666572,-0.0898667127,-0.0216330811,0.1773710549,-0.1876710206,0.2144841105,0.281578064,0.2141180784,0.5254456997,-0.1734647751,-0.1531732678,0.37305215,0.0736244768,-0.1744775623,-0.1968688816,0.3419990242,0.1091987044,-0.0577136092,0.0470552817,0.300906837,-0.1780049503,0.3103750348,0.0893283859,0.6003996134,-0.2275829315,0.1473661661,0.4256241322,0.185140729,0.0766025707,-0.5151530504,-0.0088374466,0.0859317705,0.0910278633,-0.3155486286,-0.0076458496,0.3904762864,0.1882211566,0.0195119698,-0.2847193778,0.4502235353,-0.0868169144,-0.006413294,-0.1602883488,0.2936162055,0.1938022971,0.1545355469,-0.2132195085,-0.1055912822,0.3199578226,0.2193830311,-0.0499651767,-0.2981783152,-0.1853589863,-0.2174353004,0.041320011,-0.1967570037,-0.0639087632,0.101966694,-0.0000181255,-0.0209763106,-0.099597618,0.0869245753,-0.225450784,0.2697065175,-0.1731997728,0.0348745435,0.2699767947,-0.0449175462,0.3573747873,0.3032530546,0.3511096835,0.3506349325,0.295642525,0.2215099037,0.0462822467,-0.0275798962,-0.07052771,-0.1413396001,-0.237899214,0.0085933022,0.0753656253,0.1146334559,0.0222330187,-0.0880036429,-0.2414205968,-0.0909090862,-0.1766739637,0.2309433967,-0.1248028278,-0.2502404451,-0.1732284278,0.1643973291,-0.1407435387,-0.2305453569,0.3772288859,0.0556099042,0.0493014194,0.1084255427,0.0184885021,-0.1183818355,0.4438583553,0.4053958654,0.0393586047,-0.050696522,-0.3549845219,-0.661313951,0.2525293231,0.0431112312,-0.3434359729,0.0202995092,-0.0365521647,-0.2431378365,-0.0149772819,0.3215935826,-0.1956984103,-0.0718833283,0.0016671035,-0.0122565562,-0.3665527701,-0.1038550287,0.2837750316,0.0206282865,-0.3084766269,0.1789597571,-0.1688573956,-0.0721252859,0.0031012588,-0.1626665145,0.2184788585,-0.447480619,0.2214148045,0.0423127785,0.3322560489,0.0157247894,-0.3238595128,-0.452031672,-0.2165611684,-0.2448043674,0.2828503549,-0.3794670105,0.1671109796,-0.1489569694,0.2827146947,-0.5202071667,-0.0893792287,-0.1439721435,0.3547424376,-0.1120031923,-0.056617558,-0.0558045134,-0.1355902255,0.3962640762,-0.1083717197,-0.0078290999,-0.0498403013,-0.3642436862,-0.0454759635,0.4113410115,-0.2912547886,-0.1576842964,-0.1980943084,0.2725577056,-0.0934819058,-0.1871381998,-0.2544820607,-0.0794430152,0.2933000326,0.2510578334,-0.1649762839,0.0285106357,0.015741352,0.0646502376,0.0683122873,-0.0831301138,0.0284456406,-0.3681693375,0.0268787742,-0.1657812595]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2076","title":"Issue: Dataset download error","comments":"This command will update your local file. Then you can open a Pull Request to push your fix to the github repo :)\r\nAnd yes you are right, it is a \"canonical\" dataset, i.e. a dataset script defined in this github repo (as opposed to dataset repositories of users on the huggingface hub)","body":"The download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?","comment_length":53,"text":"Issue: Dataset download error\nThe download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?\nThis command will update your local file. Then you can open a Pull Request to push your fix to the github repo :)\r\nAnd yes you are right, it is a \"canonical\" dataset, i.e. a dataset script defined in this github repo (as opposed to dataset repositories of users on the huggingface hub)","embeddings":[-0.2335668504,-0.1188599244,0.0189633165,-0.1395551413,0.0161116607,-0.005983714,0.1757637858,0.3258337975,0.2067858279,-0.0491480418,0.1204306707,-0.0448961146,0.2096992731,0.4580940306,0.1885541528,-0.0856994539,0.07323239,0.004608932,-0.3236835599,-0.0383639783,-0.2702765465,0.3223570287,-0.1850292981,0.1354490221,-0.2389193773,0.1353986859,-0.1885152608,0.1983837634,-0.1805908531,-0.4625661671,0.5045471191,0.2400262505,0.1559289396,0.1923993081,-0.0001092647,-0.0245884806,0.3951172531,-0.1717740744,-0.1087753251,-0.0114078615,-0.3153648376,-0.3022530079,0.0101289153,-0.2516593337,0.0398821309,-0.1055716202,0.1627428532,0.2556882501,0.2135183364,0.4070446789,0.2583003044,0.4219934642,0.4020578265,-0.1656321436,-0.09692467,0.0451835729,-0.1839631051,0.0846094936,0.1392523199,-0.1658406854,-0.0079474412,0.2628327608,0.0380655825,0.1774068475,0.3683722913,0.0332367308,0.2770965099,-0.4773376286,0.1777479351,0.2426344305,0.7822639942,-0.3499058485,-0.2700644732,0.0836930573,0.0555038229,-0.0859580487,0.1442579627,0.0836822465,-0.1838019937,0.2145691812,-0.0270690061,-0.1974052936,-0.1474059224,0.1989337206,-0.1147592813,0.4739906788,-0.0418312289,0.1100740507,0.218718797,-0.1721976846,-0.0383893065,0.2162341475,-0.1291910708,0.0995903909,0.0271943081,-0.1612711847,-0.039502915,0.1277526915,0.2365034521,0.1215084642,-0.1918831617,0.0103196688,-0.1954933703,-0.0749077797,0.3661405444,-0.0246856436,0.0377587527,-0.2054487765,0.6392711401,0.4878350198,0.1346796453,0.0405800529,0.1530724019,-0.311966449,-0.2728495598,-0.0209970269,0.4147314429,-0.235813722,-0.1946694851,-0.2294030637,0.2842105329,0.0485092849,-0.1242199019,0.2408177108,-0.06433101,0.2211938202,-0.0133276777,0.0414089859,-0.1770207137,-0.2442883253,-0.1393059492,0.0375979915,0.1275270581,0.026943218,0.1333570182,-0.0171654243,0.2050207108,-0.0505627915,-0.0681749955,-0.0954978541,-0.3666022718,0.2655593157,-0.0367649235,0.2211721539,-0.1885928959,0.2225810289,0.1322175115,-0.1558461338,-0.2449506819,0.1167248785,-0.2422583848,-0.3335548937,-0.3620155454,0.1275477707,-0.3240920007,-0.2706543803,0.0901475251,0.028584715,-0.2450877875,-0.1080999672,0.0043523996,-0.0430625863,-0.3726972044,-0.1773117781,0.2803132236,0.5215454102,-0.416731149,-0.2620277703,-0.1202760562,-0.2433588952,0.0799776986,0.345772177,-0.1807958931,0.1630963683,-0.2910603881,0.2494163811,0.2408754826,-0.356361419,-0.6736316681,-0.024149796,-0.0234209839,-0.1064473614,-0.0478005894,0.1957676709,0.0992479548,-0.2680750489,-0.3319655061,0.1364902109,0.0773571059,0.1863158643,-0.2293026149,-0.5873643756,0.1502218395,0.0775780752,0.0790321082,-0.0348252915,0.2630714476,0.2338417321,0.3538371623,-0.0224072766,0.0953424126,0.2815830708,0.4012528956,0.1791230738,0.2152360529,0.0869182572,-0.2827004492,0.1946865767,0.1992518157,0.263484776,-0.2397794127,-0.212899074,-0.5051876307,-0.0997637138,-0.3482337296,-0.2533221245,0.1373305619,0.0607018694,0.0376957692,-0.0402494632,-0.3664126694,0.3527960777,-0.2735735178,-0.0182049051,-0.1620049626,0.4413309097,-0.1737802625,-0.0358938128,0.2160104811,0.2136527896,0.11474704,-0.2147005945,-0.1133193001,0.3855250478,-0.2043389082,0.2543190122,0.2713074088,0.2959888875,0.1429627985,-0.4405148327,0.0782660767,0.2604020834,0.0026073689,0.1207899675,-0.1412735581,-0.0389373526,0.0335713923,0.1048457995,-0.0179276429,0.1726455688,0.4351980686,-0.1481435448,-0.1388987154,-0.2231837809,-0.202028051,-0.279030323,-0.1909111589,-0.1220412105,-0.0943429023,0.1953204423,0.6526303887,-0.081134215,0.1934058368,0.3059241772,-0.0880405605,-0.0834919214,-0.0656971186,0.2408764064,0.2677809894,0.0272595603,-0.1545904875,0.1892936826,0.0176765937,-0.073094666,0.1090161726,0.0896600485,-0.1755732894,0.0301620252,-0.0732033774,-0.08097101,-0.2738899887,0.2919788659,-0.0533328727,0.3131765723,-0.2786055803,-0.1210931838,-0.3761145473,-0.4734442532,-0.2801196873,0.1665584743,-0.3854687512,-0.3420085609,0.0037457533,0.2556138337,-0.1119044721,0.2181049883,0.1516188085,0.3672245145,-0.100423947,0.1979857683,-0.2305040956,-0.231559962,-0.2434086353,0.0562481172,0.303778857,0.1399769634,0.3024443388,-0.6965792775,-0.0006317458,-0.9145677686,-0.2800679803,0.1137308925,0.0784535185,0.2336857915,0.3152012527,0.4718111753,0.0362509415,-0.092990987,0.1916047335,-0.3334913254,0.0445958674,-0.2865602374,-0.202850014,0.0358545929,-0.0149767259,-0.1837147176,-0.4896522462,-0.3156398237,0.2744005322,-0.0282386187,0.0308955275,0.0294818673,0.0008582678,-0.0627072379,0.0112187136,-0.2304227203,-0.1335806251,0.1109158695,0.4065054059,-0.1313940436,-0.5269101262,0.2307729423,0.1508883685,0.1147906408,-0.2585181892,-0.5553159118,-0.3104572296,-0.0462882854,0.4160473943,0.1070752367,0.3148148954,0.2358893752,-0.0479348749,-0.0593311265,-0.0556549393,-0.3613809049,-0.0467392765,0.3273712993,0.3528597951,0.3838159442,0.2802631259,-0.0130923865,0.4280107021,0.1091012806,-0.0421830937,0.232493788,-0.3416447639,0.3144183159,-0.10110531,-0.312794596,0.2138321996,-0.2130126953,0.2422893643,0.369972676,0.1163371205,0.1208552495,-0.0625771433,-0.2283205688,-0.1877693981,-0.4996314943,-0.1915959269,0.5330437422,0.086441502,0.2457365692,0.0275906418,-0.3073690236,-0.2219127268,0.5058276057,0.5238962173,-0.1514753699,0.1856842935,0.1707983911,-0.0125574376,-0.4755599201,0.1710109115,0.0400471315,0.1965430528,0.0049423249,0.0275596846,0.1732417196,-0.0784424692,0.6037297845,-0.0472833402,0.1495149434,0.0507923141,0.1644227654,-0.0753689706,-0.1166736186,0.0149492081,0.295668453,0.0863799453,0.1583255678,-0.1283066422,-0.3624587953,0.4438434541,0.3782400191,-0.246840328,-0.1108985394,-0.0170871262,-0.6500023603,-0.506205976,-0.1907252222,0.0373515114,-0.1412652135,-0.061703898,0.2488444149,0.1306789964,-0.2411844432,-0.0958008617,0.1394643188,0.2625066936,0.4379006028,-0.0097437408,0.2182332426,-0.1364615858,-0.1653878242,0.4312852919,0.0731007084,-0.4082481563,-0.1175405011,-0.2884830236,-0.1285909861,0.2834292352,0.0284764003,-0.0203417074,0.0824792013,0.0702318475,-0.1980328411,0.1282812208,0.4534525871,-0.0799646676,-0.3448009789,-0.3542917669,0.4003818631,-0.114788048,-0.0775211453,0.3779807985,0.3538376391,-0.0498624742,-0.0894848928,-0.0248594508,0.652633667,0.114683941,0.0795668885,0.2420532405,-0.1019820049,0.6518208385,-0.0410398506,0.1222633123,-0.3257214129,0.0006454148,-0.0996314809,0.1505215317,-0.2412483245,-0.0039451406,-0.0847467482,0.4775623977,0.1514837444,0.1945160925,0.1278756112,0.3077500463,-0.0544245094,-0.1355787516,-0.2700302601,0.2111504525,-0.0885102749,0.6037105918,-0.0892815366,-0.1621542275,0.2068763226,-0.197557494,-0.5399237275,0.0476854593,-0.1331095695,0.3173460066,0.0773151368,0.1214797571,0.0688775778,0.1182058603,0.4526980817,0.0669935569,-0.372861892,0.2600724399,-0.0131286653,-0.1014176831,-0.0101234475,0.3554030657,0.0644311011,-0.3104497194,-0.4030135572,-0.1997881532,-0.0247619841,-0.2261962593,0.0545977205,-0.1979746521,0.3774759471,0.0033662615,-0.0794596747,-0.0526558571,-0.0053540054,-0.2600459754,0.1364441067,-0.1225391254,0.0210397057,0.1288243681,-0.1615355015,-0.1095585451,-0.0184487998,0.2816782296,-0.2700921595,-0.1404268295,0.4918085635,0.2861503363,-0.0161390584,-0.243947193,-0.0703868568,-0.0397299528,-0.7922115922,0.1132473201,0.0618330799,0.097764954,-0.1027815863,0.6565722227,0.1902372092,-0.0865861326,-0.2809303701,-0.4150904119,-0.3234255612,0.098098062,0.2170947343,0.2083801925,0.0841273665,0.0156482607,0.201613754,-0.5624887943,-0.3321127594,0.0650147721,-0.0897626355,-0.1126650497,0.5341254473,0.1613993496,0.1736094803,0.0590881296,0.0796298683,-0.0411021113,-0.2069408894,-0.1628685743,-0.3124522865,0.1237830445,-0.1371285617,-0.3970655799,0.0843207985,-0.2241980284,-0.1113071367,0.0092433402,0.1131364331,0.2931981385,-0.00702429,0.2214222848,0.3557651937,-0.1219320297,0.0378737524,0.1440412998,0.3490453959,0.5711564422,0.0751873553,0.3948974013,-0.0918709859,0.1050569564,0.3819204569,0.1560140699,-0.0424252152,-0.0519230254,0.1917815208,-0.1765855253,0.0707428902,0.1867443174,0.1620068699,0.5038245916,-0.2613509297,-0.1105778515,0.2460002303,0.1506229341,-0.3346230984,-0.1828799993,0.2637704015,0.0990466997,-0.0441014692,0.0831305608,0.3345210552,-0.1747628897,0.2895328701,0.0971054807,0.581063211,-0.2587447762,0.0536668636,0.2354916036,-0.0437964685,0.0007868237,-0.2322138995,0.0009578543,0.2113996446,0.1507844627,-0.201385811,-0.0427670665,-0.0092601143,0.2668437362,0.0916843712,-0.2197218984,0.2324923873,-0.1332146525,0.0301511828,-0.113677457,0.1947420985,0.2423432767,-0.032895498,-0.1802568585,-0.341166079,0.4657826126,0.1578105241,-0.1460884064,-0.1140911207,-0.1445540935,-0.1462970972,0.1082798988,-0.170943141,-0.1186125576,0.0937976763,-0.0343469568,-0.078890264,-0.1461081803,-0.1543906778,-0.1232948154,0.3736056983,-0.1328389496,0.116294302,0.2969819605,0.0382524356,0.2212441266,0.3534228802,0.4745325744,0.1789368689,0.1725443453,0.1585730314,0.0046515083,-0.0992411822,-0.0175817255,-0.0124579631,-0.0896219239,0.1290501058,0.2305071801,0.1899619997,-0.0855161548,-0.2006911486,-0.1554967761,0.0044149817,-0.1103282273,0.0721085221,-0.2311008722,-0.0567194484,-0.3328525424,0.1490005404,-0.1740016788,-0.1544375122,0.2945476174,0.0350462347,0.1269104928,0.0174955465,0.0980270505,-0.1290970594,0.5321985483,0.3212457895,0.0379091091,-0.0222432483,-0.313305825,-0.7508372068,0.1343322098,0.0117834574,-0.2444421351,0.0656240061,-0.0677617267,-0.1610686332,0.0162588079,0.2551141381,-0.0839381665,0.0151442913,-0.0634626001,-0.0591656603,-0.2827605009,-0.0881805122,0.255261451,0.1163080782,-0.2039343417,0.1172659621,-0.123466149,0.0184981618,0.0706111714,-0.0858750418,-0.0224957727,-0.3802748322,0.0649752393,0.1110791564,0.281839937,-0.1864163131,-0.2438083589,-0.4563576579,-0.2451844066,-0.3523602784,0.2957237363,-0.2310489416,0.1400176436,0.0714991018,0.1189493164,-0.4806298316,0.0499630049,0.0205127075,0.2524638772,-0.2548214197,0.1510595828,-0.1215061992,-0.0219836384,0.312618643,0.04718519,-0.1636137962,-0.0844619498,-0.1241704524,-0.1733531952,0.6147268414,-0.2698357999,-0.1295505315,-0.1446399391,0.1861272901,-0.0862272084,-0.2793920338,-0.3792189956,0.1572494209,0.0819156468,0.2910230458,-0.1266591549,0.136877954,-0.1512439251,0.0280845501,0.0543352813,0.0638352484,0.2239320725,-0.3455142081,0.0839790702,-0.1800156981]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2076","title":"Issue: Dataset download error","comments":"Hi, thanks for the answer. \r\n\r\nI gave a try to the problem today. But I encountered an upload error: \r\n\r\n```\r\ngit push -u origin fix_link_iwslt\r\nEnter passphrase for key '\/home2\/xuhuizh\/.ssh\/id_rsa': \r\nERROR: Permission to huggingface\/datasets.git denied to XuhuiZhou.\r\nfatal: Could not read from remote repository.\r\n\r\nPlease make sure you have the correct access rights\r\nand the repository exists.\r\n```\r\n\r\nAny insight here? \r\n\r\nBy the way, when I run the datasets-cli command, it shows the following error, but does not seem to be the error coming from `iwslt.py`\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home2\/xuhuizh\/anaconda3\/envs\/UMT\/bin\/datasets-cli\", line 33, in \r\n sys.exit(load_entry_point('datasets', 'console_scripts', 'datasets-cli')())\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/commands\/datasets_cli.py\", line 35, in main\r\n service.run()\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/commands\/test.py\", line 141, in run\r\n try_from_hf_gcs=False,\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/builder.py\", line 639, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/utils\/info_utils.py\", line 32, in verify_checksums\r\n raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums)))\r\ndatasets.utils.info_utils.ExpectedMoreDownloadedFiles: {'https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz'}\r\n```","body":"The download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?","comment_length":148,"text":"Issue: Dataset download error\nThe download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?\nHi, thanks for the answer. \r\n\r\nI gave a try to the problem today. But I encountered an upload error: \r\n\r\n```\r\ngit push -u origin fix_link_iwslt\r\nEnter passphrase for key '\/home2\/xuhuizh\/.ssh\/id_rsa': \r\nERROR: Permission to huggingface\/datasets.git denied to XuhuiZhou.\r\nfatal: Could not read from remote repository.\r\n\r\nPlease make sure you have the correct access rights\r\nand the repository exists.\r\n```\r\n\r\nAny insight here? \r\n\r\nBy the way, when I run the datasets-cli command, it shows the following error, but does not seem to be the error coming from `iwslt.py`\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home2\/xuhuizh\/anaconda3\/envs\/UMT\/bin\/datasets-cli\", line 33, in \r\n sys.exit(load_entry_point('datasets', 'console_scripts', 'datasets-cli')())\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/commands\/datasets_cli.py\", line 35, in main\r\n service.run()\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/commands\/test.py\", line 141, in run\r\n try_from_hf_gcs=False,\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/builder.py\", line 639, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/xuhuizh\/projects\/datasets\/src\/datasets\/utils\/info_utils.py\", line 32, in verify_checksums\r\n raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums)))\r\ndatasets.utils.info_utils.ExpectedMoreDownloadedFiles: {'https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz'}\r\n```","embeddings":[-0.1731770486,-0.0103908228,0.0573096722,-0.0393557921,0.0538161732,-0.0124162808,0.2562997043,0.3127626777,0.1133866832,0.0201736651,0.031347841,0.0189829245,0.3240020573,0.2409658432,0.1655391902,0.0832946524,0.0935480818,0.0070209429,-0.2977551818,-0.0181837156,-0.2464222908,0.2069954872,-0.1871381104,0.0984047502,-0.0664139017,0.0939154625,-0.1437421888,0.132330969,0.0711327344,-0.5863393545,0.2837122679,0.1516173184,0.2967727184,0.2796620727,-0.0001216275,-0.0298803672,0.2723470628,-0.133409366,-0.3635321259,-0.0276102722,-0.4098815024,-0.4128105938,-0.2503136992,-0.2697912157,0.1252630949,-0.0470432676,0.052626878,0.1782296002,0.2551392615,0.4470214844,0.1154646128,0.3542398214,0.4006434083,-0.0996041298,-0.0763279796,0.0624176264,-0.0808924809,0.1945292354,0.1765618473,0.004309325,0.319073379,0.1249282211,0.0774350762,0.0930948779,0.2438849509,-0.0995643437,0.4032182395,-0.391949445,0.2226319909,0.0467114076,0.757717669,-0.2985826135,-0.5799194574,-0.0442384295,0.1599040031,0.0499051549,0.1854637265,0.0469910651,-0.2484175116,0.3061733842,-0.1514507979,-0.4920485318,-0.2695293725,0.3524078131,0.0779909343,0.4447513223,0.0607923344,0.1950392127,0.2497365922,-0.1434856057,0.0307610016,0.1108072475,-0.2411747128,0.1607265174,0.0114447465,-0.101246573,-0.1041230708,0.1892344803,0.1351170391,0.0979463086,-0.229634434,-0.0681871325,-0.0995049998,-0.0101151504,0.2546986639,0.0938820541,-0.0022864302,0.0234885979,0.4250120819,0.4284123778,0.205971241,-0.0637580827,0.0849416628,-0.3227720857,-0.3013033867,0.0541344583,0.2840809226,-0.1511059105,-0.2094755173,-0.2937165499,0.0399769582,-0.0105951065,-0.0702962652,0.220625788,-0.1020833254,0.1546022892,0.0166574493,0.1494405866,-0.0292690303,-0.3486822844,0.0202518329,-0.005921334,-0.0080564497,0.1573093385,0.1737334579,-0.0985193104,0.0062291725,-0.0080840718,-0.2337122411,-0.1763401628,-0.2984265983,0.1085456312,0.1047854647,0.1262549013,-0.135873273,0.2610064745,0.034558475,-0.1750120968,-0.0659154877,0.1226122901,-0.3552239239,-0.2412035614,-0.247345522,0.0090103764,-0.2274768203,-0.1725842357,-0.151414454,-0.0946407542,-0.126471132,-0.3580670059,-0.0076441127,0.0192548428,-0.2257637233,-0.1544501334,0.2990473807,0.5962198377,-0.5039988756,-0.3450402021,-0.2770167589,-0.2029024065,0.0789278671,0.319047451,-0.2443762571,0.2259961218,-0.2795182168,0.128381446,0.3324233592,-0.5260116458,-0.7665931582,0.2068858594,0.0146962292,-0.0245410148,0.0172463097,0.0817185864,0.0900460705,-0.2895280719,-0.6715387106,0.0571655147,0.0781765506,0.1101767048,-0.174841702,-0.5463120937,0.1403319836,-0.0209613889,0.0476417989,0.2466131002,0.452853024,0.162594974,0.484161675,-0.106832318,0.2486445159,0.1819880009,0.5103002787,0.1815202534,0.1283855289,0.0267958213,-0.3259159923,0.259906143,0.2652192116,0.1192675605,-0.4695916474,-0.0714321285,-0.383805126,-0.0778345019,-0.3433720767,-0.1025640815,-0.0172441565,0.2139872462,0.0695284158,-0.0191354621,-0.1408530772,0.2331992835,-0.1202415079,-0.0135404021,-0.0989577547,0.3463067412,-0.1621232182,-0.0224540047,-0.0325786434,0.1884357184,0.2230120599,-0.2281770557,-0.141573146,0.4947054088,-0.4156080186,0.322922051,0.2500270009,0.1136658341,0.1753347516,-0.3674814105,0.1324027032,0.2234054357,0.023047762,0.1755135059,-0.1077930257,-0.2079482675,0.0823868066,0.2169461399,0.0122453496,-0.0019036907,0.2869502604,-0.1655271649,-0.0504661202,-0.1969952881,-0.1162137166,-0.1363363564,-0.1499486119,-0.1951585263,-0.0224002544,0.1898770034,0.4447519183,-0.0751990378,0.1647149622,0.3001323044,0.0432810113,0.0643934384,-0.0550909862,0.6107308865,0.2715999782,-0.1495793611,-0.2571927607,0.0710750371,0.2623735964,-0.1553240865,0.191364482,0.0627828464,-0.0459873453,0.0877594873,-0.0473504849,-0.0395073555,0.0910843536,0.2150114328,0.1679970324,0.4387922287,-0.4453455806,-0.0061937873,-0.429361552,-0.6389654279,-0.2003344744,0.250191927,-0.2366262525,-0.1858728379,-0.1833362132,0.2103633136,-0.0788792372,0.2053449303,0.0379035957,0.3174077868,-0.0843602195,0.1659031957,-0.2603982687,-0.0483416431,-0.206170246,-0.0613751225,0.2968954444,-0.0703743771,0.2665001154,-0.5192921162,-0.1063686758,-0.938561976,-0.0400097258,0.0445348062,0.0885354728,0.2453931123,0.1265654713,0.4747102857,0.1492189318,-0.059593495,0.2381046712,-0.1594686359,0.1200468764,-0.1929836571,-0.0851439834,0.0652761981,-0.1378386766,-0.1064785644,-0.5346980691,-0.2964923978,0.358810842,-0.0660284981,0.1081246287,0.0974272117,0.2043614835,-0.0727187395,-0.0373493396,-0.170776248,0.0235799383,0.0401086658,0.2624898851,-0.0981030911,-0.4878210425,0.1330920905,0.3627158403,0.2413807064,-0.0405476578,-0.4037993252,-0.0293230526,0.02823947,0.5150356293,-0.0052890922,0.2320146263,0.0692331716,-0.0297776442,0.073668994,-0.0076382211,-0.158682555,-0.0632468909,0.2388671339,0.4697849452,0.5023777485,0.5303525329,-0.0160536561,0.3157682717,0.0231780373,0.0330219492,0.4492385089,-0.3102992177,0.3521688879,0.187814638,-0.47141698,0.1839674711,-0.0733270198,0.2329026014,0.277969867,-0.0017412696,0.065965049,-0.1709498763,-0.2562528253,-0.2403241396,-0.4881181717,-0.236458987,0.4585447907,0.2765527964,0.2874034345,0.0683452412,-0.4240667522,-0.2522800863,0.4481953382,0.3350012302,0.0827876627,0.1686724424,0.0481572561,-0.2130631655,-0.4461756647,0.2900398374,0.1415863335,0.4409944117,-0.1199093312,0.1655836552,0.139020443,-0.0851578116,0.5852239728,-0.1715991348,0.1993331015,-0.1369415671,0.3749858141,-0.1045500413,-0.1160969064,0.089206472,0.2740935981,0.4025423527,0.291890949,-0.1657921225,-0.3208502829,0.2214213312,0.3830725551,-0.2241475284,-0.2301244438,0.000843759,-0.3451741636,-0.693325758,-0.328115344,-0.0069824695,-0.0880711153,0.0512094125,0.2241978049,0.2142446041,-0.067785956,-0.1070833281,-0.0017372563,0.2221690416,0.4739635885,-0.0393842235,0.1846898049,0.0023138637,0.1761827171,0.3277204335,-0.0107346317,-0.256616056,-0.1608529836,-0.2513644397,-0.1139626727,0.2904008031,-0.0327516645,0.192662701,0.1949869096,0.0892309025,-0.0346815102,-0.0578938276,0.4571619928,-0.1164326146,-0.3752474785,-0.4556741118,0.3361280262,-0.2495938092,-0.0428851917,0.1845281571,0.3937385976,-0.11252442,-0.2946995497,-0.1475628316,0.7340781093,0.2102341056,0.0129900761,0.2700057924,0.091685921,0.6499993205,0.0459565111,0.0865688622,-0.2105216086,0.0752877519,-0.0589319579,0.102904059,-0.2461962253,0.0928064957,-0.0610204786,0.4542872012,-0.0321819335,0.0278315265,0.1643616557,0.4223260581,-0.1249925867,0.0000639836,-0.0898267627,0.0960050151,-0.0762339905,0.3992200792,-0.0595668815,-0.1704473346,-0.0311486181,0.0264308155,-0.5818719864,0.1942543983,-0.27101174,0.2384095639,-0.1024199054,-0.0012000588,0.3616998792,0.1504682899,0.3591023684,0.0494045056,-0.3497042656,0.2235165685,-0.1748246551,-0.1081772447,-0.1018287763,0.3630264103,0.0883836225,-0.2549300492,-0.5241520405,-0.0086621959,-0.0361234434,-0.2364127636,0.1048422456,-0.1667547524,0.4092418849,-0.098962523,-0.2461811602,-0.2079121768,-0.1235489994,-0.2296192944,0.0210539717,-0.1208385378,0.1367802173,0.1103190556,-0.2699893415,-0.1645362228,0.0462829806,0.4265493453,-0.4591656327,-0.1281836778,0.7964831591,0.2921908796,-0.0672190711,-0.1716514379,-0.1413630694,-0.1442639679,-0.9090535045,0.0603618436,0.1543779224,0.2676517367,-0.0645497218,0.4446279109,0.122508496,0.0626251623,-0.2182368189,-0.4203623831,-0.104992196,0.0045819734,0.2443049997,0.1058994308,-0.0736550465,-0.0744286627,0.1432595849,-0.5616599321,-0.1851237118,0.0277122725,0.0049255798,-0.0934065953,0.6767575741,0.2900308073,0.260938257,-0.0907273516,-0.0271651223,-0.037719477,0.002826734,-0.0262703896,-0.2235167176,0.1833406091,0.0595036782,-0.4573156536,-0.0453925803,-0.3796971142,0.0052258577,0.0252172947,-0.0769211203,0.1197480559,0.0208436772,0.210832566,0.2600322366,-0.1026636288,-0.0479612462,0.3185764253,0.2582704425,0.4589096308,-0.0411697663,0.5171275139,-0.0219034553,0.1272802055,-0.141289115,0.0447319783,0.0138183478,-0.0031941428,0.3017369509,-0.2399451286,0.1938872784,0.2793976963,0.0257165022,0.3771682084,-0.1780409217,-0.1760485768,0.2011232376,0.0425304696,-0.2472284138,-0.1249372438,0.1232035011,-0.1259376407,-0.1002598479,0.0616432875,0.1325864643,-0.3096863329,0.2227418274,0.0760903433,0.4969694614,-0.127991423,0.0035038413,0.4426576495,0.1113885269,0.0824367553,-0.2332008928,0.1057075858,0.1084556431,0.1235668138,-0.433486402,-0.1788526624,0.1383217275,0.2426339984,0.0103891008,-0.3338914514,0.1450665444,-0.1855845302,-0.113968499,-0.1571413279,0.11353793,0.2669583559,0.0549767129,-0.2165636569,-0.1758531779,0.5872370005,0.2028712034,-0.0473489054,-0.2904474139,-0.2332371175,-0.1876093149,0.0005462788,-0.1972236484,-0.2776622176,-0.1044545099,0.179691568,-0.1478834003,-0.2340168357,-0.1056610569,-0.1362780333,0.3717666864,-0.1313414872,0.0957517326,0.190945819,-0.0500356108,0.2088549286,0.3342048824,0.4915676117,0.286650598,0.5088070631,0.2947314978,0.0730183348,-0.0274139959,0.1301703602,-0.0285185985,0.0101048928,0.1617390513,0.1069726646,0.0660343543,-0.005749533,-0.3965212405,-0.088792406,0.090799287,-0.1141300574,0.2506958544,-0.344479382,-0.2569884956,-0.3263885081,0.1786250025,-0.1417440921,0.0649629012,0.3046619296,-0.024597127,0.1086314321,-0.1176665127,0.021419771,-0.1723730415,0.4463753402,0.4036257565,0.0717265308,-0.199913159,-0.2780932188,-0.6719225645,0.0896465033,0.0252116714,-0.1961073875,-0.0607285164,-0.0723312199,-0.2990927398,-0.2401502132,0.387195617,-0.014077777,0.1944933981,0.1509940773,0.0384239517,-0.2599557042,-0.0353370421,0.399189353,0.0820802823,-0.2298554778,0.175003469,-0.3563649058,-0.0808258578,0.0815266147,-0.1314022094,-0.0405390039,-0.3032632768,0.1969623417,0.0286377147,0.2868204117,-0.1024687439,-0.331605196,-0.4810997844,-0.2985094488,-0.2880403697,0.3028406501,-0.1538033187,0.0884626359,0.0518565364,0.2830046713,-0.5554578304,0.0205871705,-0.1121595874,0.200289771,-0.0868080705,0.0174596943,-0.0110800508,-0.0363526531,0.4912633896,-0.154374361,-0.1163395345,0.1405214816,-0.356621027,-0.073570706,0.5362978578,-0.3583723009,-0.0659786016,-0.0804208666,0.3946353197,-0.0404716842,-0.2846677899,-0.3354431391,0.0593702458,0.2304164469,0.285646379,-0.0827299803,0.082880415,-0.2585714161,0.1965327561,0.146917671,0.0961877927,0.2246536762,-0.3955002725,0.1924878061,0.0614372231]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2076","title":"Issue: Dataset download error","comments":"Hi ! To create a PR on this repo your must fork it and create a branch on your fork. See how to fork the repo [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md#start-by-preparing-your-environment).\r\nAnd to make the command work without the `ExpectedMoreDownloadedFiles` error, you just need to use the `--ignore_verifications` flag.","body":"The download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?","comment_length":45,"text":"Issue: Dataset download error\nThe download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?\nHi ! To create a PR on this repo your must fork it and create a branch on your fork. See how to fork the repo [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md#start-by-preparing-your-environment).\r\nAnd to make the command work without the `ExpectedMoreDownloadedFiles` error, you just need to use the `--ignore_verifications` flag.","embeddings":[-0.2560881972,-0.0967695266,0.0178538598,-0.0659613013,0.0944884494,0.049594406,0.1920959204,0.2611156106,0.1636176109,0.0419375785,0.0050670132,-0.0133628454,0.2525415123,0.3941254914,0.3025112152,-0.1754232049,0.152458638,0.0719700158,-0.2172753215,0.0543560348,-0.2055191845,0.2804819345,-0.2303613126,0.0209608916,-0.1193833053,0.1293283403,-0.3126569688,0.1732152849,-0.1787016839,-0.6029583812,0.3105776906,0.3934975564,0.2743170857,0.2377393097,-0.0001235105,-0.0232407935,0.3910424709,-0.1918268353,-0.2590373456,-0.0267316513,-0.3742640615,-0.4201498628,-0.1484021991,-0.1779622287,0.0797267258,-0.2863840461,0.1079144999,0.3710989058,0.2191396654,0.3876252472,0.1255221516,0.4391294718,0.5360848308,-0.1845331043,-0.0708419606,0.1065436676,-0.1627684236,0.2750478685,0.178085506,-0.080706276,0.2044587582,0.173540324,-0.0209426358,0.0642212927,0.3660503328,0.1444919556,0.3109253347,-0.4145330489,0.1908102334,0.2390266955,0.7586012483,-0.338901639,-0.4380357563,-0.0868122056,0.16902937,-0.270378083,0.1488966495,-0.036151588,-0.1700431556,0.2979440391,-0.16484496,-0.4038984478,-0.2858358324,0.3131330013,0.0762056261,0.4441383183,0.0129923113,0.1372981668,0.4096088111,-0.1014984697,-0.0474601611,0.1721398234,-0.2839389443,0.0506905243,0.1691611707,-0.2852407992,-0.1591252536,0.0787700936,0.300927341,0.1759157628,-0.3208157718,-0.1262310594,-0.1952875853,-0.0971674025,0.2797228694,0.088927716,0.0287298113,-0.0519122779,0.4055681825,0.6368993521,0.2258510143,-0.0737120435,0.151320681,-0.3014290929,-0.2974278927,-0.0874256492,0.2769135237,-0.1878219247,-0.0803302974,-0.2514245212,0.1449325234,-0.1607227921,-0.086113669,0.3290611804,-0.1138228402,0.1629012227,0.0208368599,0.1700348705,-0.1408061832,-0.2607530653,-0.0238272939,0.0364172868,0.0059565594,0.1546955854,0.1064969152,0.0981587544,0.180533424,-0.0965210497,-0.0971262306,-0.1832295507,-0.3297604024,0.1840197444,0.0529415272,0.0353848338,-0.1530470252,0.1169521585,0.0272814836,-0.1550368071,-0.2040176541,0.2259788811,-0.1315045208,-0.3452872336,-0.1903003156,0.0012859319,-0.2778307199,-0.4454151988,-0.0313395187,0.0521886945,-0.2458992153,-0.2138424069,0.0153320953,0.076546602,-0.3249950111,-0.1181462556,0.2416535616,0.7265816331,-0.5139055252,-0.2789271772,-0.0764444843,-0.3277512491,0.224521786,0.251598984,-0.1615429223,0.2488374412,-0.3000688255,0.0010975152,0.3311366141,-0.3804425299,-0.6101977825,0.1879918128,-0.0540633611,-0.0505580269,0.112440519,0.1294066608,0.1397242546,-0.3284981251,-0.5146924257,-0.0157630797,-0.0232061427,0.030654937,-0.1765310317,-0.6106555462,0.2020420879,-0.0040537803,0.0648262873,0.2028805614,0.2194609642,0.2528126836,0.4670443237,-0.0675518289,0.0977601483,0.2156460285,0.4710688889,0.2730847299,0.0670088977,0.0601324365,-0.1762896627,0.1627109051,0.1490887552,0.2279654741,-0.2123599499,-0.061800994,-0.5251754522,-0.1285540313,-0.3235305548,-0.2717590332,-0.0287415627,0.0901924446,0.2446545511,-0.1121621206,-0.2586700916,0.2478964031,-0.2294490486,0.1092184335,-0.1207645014,0.4380150437,-0.2130049467,-0.0119133778,-0.0282821488,0.1945421696,0.2381058931,-0.321310252,-0.0859922245,0.4823783338,-0.3050897121,0.2762858272,0.3618417084,0.2052758634,0.2600658238,-0.4229335785,0.2289719582,0.123264052,-0.0746655166,0.1577857435,-0.1887511164,-0.0927077085,-0.0088550942,0.2620516121,-0.0418127105,-0.0275799446,0.3348996639,-0.1963994205,-0.0883994475,-0.2597973347,-0.2072836459,-0.2496730536,-0.2829404771,-0.2605138421,0.0171166938,0.2126549631,0.4123401046,-0.1385304332,0.0295032654,0.3590830863,0.0673985854,-0.0065640658,-0.2185413986,0.4872529209,0.3576416671,0.0202488285,-0.1737889498,0.1996552944,0.1690061092,-0.0750790164,0.2454595119,0.1035990864,-0.1709520072,0.058602009,-0.0797947794,-0.0475222506,-0.0519351587,0.1176466644,0.1086500436,0.3397515118,-0.2460732609,-0.0674613863,-0.3342811167,-0.6024009585,-0.3408093154,0.1691476405,-0.3219607174,-0.1643619686,-0.0301357284,0.2658126354,-0.0377430804,0.2160297632,0.1099042818,0.2593465447,-0.0848574564,0.1422858089,-0.238899067,0.0188314728,-0.1592909396,-0.0287546888,0.3545594215,0.0627913177,0.2899083495,-0.5861014128,-0.2462389469,-1.0009710789,-0.1878997684,0.1199490428,0.0032070307,0.1540724933,0.4157357812,0.3913275599,0.2890338004,0.0603233017,0.278177321,-0.2399691492,0.1530865133,-0.2830749154,-0.152023837,0.0930634364,-0.059289556,-0.2569691837,-0.4589641392,-0.2208829075,0.2960768044,0.1360130608,0.1107346714,0.0468280315,0.2436873317,-0.0483041219,0.0023817073,-0.2082563192,0.0096366294,0.0344335288,0.2957965136,-0.1325061172,-0.4118843377,0.2555496991,0.2001850307,0.0281182639,-0.2047903687,-0.5103381276,-0.3684630096,-0.0034458237,0.4062130451,0.0548579693,0.46388188,0.1891883016,0.062508747,-0.0122242803,-0.0108426567,-0.367806524,-0.0895677879,0.1972306073,0.5582021475,0.4164224267,0.4434440136,-0.1451857984,0.5577757955,0.087642312,0.009520161,0.2989994287,-0.3636427522,0.4048640132,-0.0420439765,-0.3322061896,0.1826124936,-0.1945263445,0.3318013847,0.3361841738,0.1469948143,0.1040747538,-0.0948179066,-0.170291096,-0.1329937279,-0.4028578103,-0.1485631466,0.5888517499,0.1436891854,0.2323760539,0.0228491407,-0.3772394359,-0.1427170932,0.4252016246,0.3784027994,0.030697098,0.0962839201,0.1603626758,0.0097387889,-0.5299116373,0.2029698342,0.0565260537,0.1712635905,0.022036748,0.0857075229,0.1511494964,-0.0697492734,0.7038940787,-0.1408514977,0.0877512917,-0.0066865161,0.1629096866,0.1104422584,-0.0515302494,-0.0384973101,0.1824557632,0.3331947327,0.1445677578,-0.2590572238,-0.3296528161,0.5500895381,0.4340200126,-0.1855707914,-0.2135308385,-0.0067792069,-0.4758854806,-0.6366705894,-0.1136641875,-0.0540178455,-0.0740406066,-0.154366985,0.1138616875,0.2564132214,-0.1707112342,-0.1556663364,-0.0892773718,0.1769658029,0.4189612269,-0.0561009794,0.1763493717,-0.0518058352,0.1137762889,0.3967272937,-0.0779781789,-0.3872693479,-0.1244266555,-0.2225025296,-0.1031104475,0.3866261244,0.091037102,0.103468664,0.2654752731,0.1771851033,-0.0770069882,0.056506224,0.4585667551,-0.0764741823,-0.3969675303,-0.454128325,0.330631882,-0.122872144,0.0162306335,0.2529269159,0.4226772487,-0.1093464717,-0.2192014903,-0.1736412048,0.9077013731,0.0827924311,0.0347732678,0.2712682486,-0.1025482565,0.6737480164,-0.131731838,0.1328425705,-0.2610991001,-0.0492309146,-0.0813681409,0.101221554,-0.1558448225,0.0635670573,-0.1261626482,0.4567217231,0.059892159,-0.0140631925,0.1917673051,0.4202847481,-0.1178311557,-0.1718258411,-0.3096849322,0.07607802,0.0758548006,0.42379722,-0.1220674962,-0.1469487548,0.1439669132,-0.2193653435,-0.6025186777,0.0687767714,-0.2006504536,0.0874664709,0.1361119449,0.0731362179,0.2486495227,0.1761371344,0.4182500541,0.0958505273,-0.3794705272,0.2238376141,-0.1555961072,-0.178517729,-0.0702659637,0.3249340653,-0.0614270084,-0.2574082613,-0.4357274175,-0.205508098,-0.0111800255,-0.2999996543,0.1022256538,-0.2520038188,0.5220275521,-0.1184472442,-0.0968032852,-0.0729909614,-0.1612082124,-0.3448168933,0.018516548,-0.0442700982,-0.1120540872,0.0390927643,-0.1782472879,-0.0570420921,-0.0896246657,0.2339620739,-0.3715412021,-0.181703344,0.5752364397,0.3446544409,-0.0155805219,-0.2305596471,-0.2190037221,-0.2222649902,-0.8353117704,0.1835679263,0.2915771008,0.1671943218,-0.0672511756,0.5296762586,0.1522462666,0.0785764754,-0.1942711323,-0.3521624207,-0.2618650198,-0.0023979102,0.2342166752,0.1948966533,-0.0063252058,-0.000367858,0.1384163499,-0.4774488509,-0.193897292,0.0092139244,0.0637509152,-0.0911250487,0.6803836823,0.171993345,0.2699715197,0.0258042701,0.0068254997,0.0369966961,-0.0101184119,-0.0110711744,-0.2875038683,0.1807176322,-0.1370130926,-0.4130552709,0.0267751385,-0.1023889333,-0.0760752708,-0.0343736745,0.0697011128,0.1879000813,-0.0636162981,0.070868209,0.4111145437,-0.1488436759,-0.1303729266,0.1419173777,0.3417853713,0.5433329344,0.1460548341,0.3977508545,-0.0944912359,0.0844845921,0.1740996391,0.1218340397,0.2144066095,0.0460660011,0.3353434205,-0.1372398287,0.0646516159,0.3166057169,0.2093932033,0.4750918746,-0.1000737548,-0.0756379217,0.3047186434,0.0343143716,-0.3281185627,-0.15996328,0.2656958103,-0.044131849,-0.0775051489,0.2122294307,0.1708355546,-0.156844914,0.3383660913,0.1190065593,0.4063675702,-0.2076783925,0.1075006798,0.3448408246,-0.0184340086,0.0205212925,-0.0840305015,0.0543313622,0.1479153335,-0.0069144629,-0.332580775,0.0007728359,-0.0441074818,0.3287574947,-0.0699130148,-0.3360921443,0.1808307916,-0.1922996193,0.1239924878,-0.1625453681,0.0326791108,0.3966884017,0.0730532706,-0.123724483,-0.3008503318,0.4938406646,0.2963202596,-0.1725647599,-0.1268198788,-0.1906029582,-0.1363848001,0.066862449,-0.1083364636,-0.3233785927,0.0920701846,0.0047575547,-0.0773351043,-0.4025586843,-0.1364331245,-0.2281276733,0.3287055194,-0.1347250789,0.1191616803,0.2238925099,-0.0048156288,0.2363278121,0.2858164907,0.4661472738,0.2152849436,0.4449116886,0.2052199394,-0.0639488921,-0.032474868,0.1189123765,0.0438313335,-0.0515888147,0.1417502314,0.0561558269,0.0292951651,-0.0215989724,-0.3440901637,-0.0517355278,-0.0665440932,-0.1367293447,0.3235183358,-0.0757746845,-0.1349941045,-0.497431159,0.1811789572,-0.2072880119,-0.0507324226,0.2020741105,-0.0605684519,0.03585504,0.0579457618,0.0071439277,-0.08926595,0.3542531729,0.396135658,0.0571390539,-0.2331694961,-0.1373505145,-0.608014524,0.0722813904,0.0295273345,-0.2606552839,-0.1607082933,-0.050468497,-0.2422070056,-0.0067011067,0.3835419416,-0.1139454767,0.268103838,0.1205851585,-0.0389942974,-0.1914321035,-0.1258768588,0.4343543053,0.1069079936,-0.1347616613,0.055138018,-0.1180198193,-0.1362476051,0.1671075374,-0.0345275514,-0.128647998,-0.3072927296,0.1429600865,0.0770579278,0.3717910647,-0.1529374868,-0.3054271042,-0.4834336042,-0.222626105,-0.4171458185,0.2702682316,-0.0821720064,0.0717913881,-0.0331223235,0.1794090122,-0.5594061017,0.0533704348,0.0479888543,0.2227483243,-0.2027693838,0.1464823931,-0.0488196686,-0.0336135961,0.4108432829,-0.0875505358,-0.1198268756,-0.0490775779,-0.240954712,-0.0097331777,0.6197211146,-0.218415305,-0.1049643904,-0.0750179291,0.2192890346,-0.0359685719,-0.305416584,-0.4327182472,0.1702481806,0.1123792753,0.2900764942,-0.0736745223,0.2134283334,-0.3135068715,0.0939059556,0.0636195019,0.2041409612,0.1747090667,-0.3216533363,0.0339301564,-0.0607590415]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2076","title":"Issue: Dataset download error","comments":"Hi @XuhuiZhou,\r\n\r\nAs @lhoestq has well explained, you need to fork HF's repository, create a feature branch in your fork, push your changes to it and then open a Pull Request to HF's upstream repository. This is so because at HuggingFace Datasets we follow a development model called \"Fork and Pull Model\". You can find more information here:\r\n- [Understanding the GitHub flow](https:\/\/guides.github.com\/introduction\/flow\/)\r\n- [Forking Projects](https:\/\/guides.github.com\/activities\/forking\/)\r\n\r\nAlternatively, if you find all these steps too complicated, you can use the GitHub official command line tool: [GitHub CLI](https:\/\/cli.github.com\/). Once installed, in order to create a Pull Request, you only need to use this command:\r\n```shell\r\ngh pr create --web\r\n```\r\nThis utility will automatically create the fork, push your changes and open a Pull Request, under the hood.","body":"The download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?","comment_length":126,"text":"Issue: Dataset download error\nThe download link in `iwslt2017.py` file does not seem to work anymore.\r\n\r\nFor example, `FileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnted\/texts\/zh\/en\/zh-en.tgz`\r\n\r\nWould be nice if we could modify it script and use the new downloadable link?\nHi @XuhuiZhou,\r\n\r\nAs @lhoestq has well explained, you need to fork HF's repository, create a feature branch in your fork, push your changes to it and then open a Pull Request to HF's upstream repository. This is so because at HuggingFace Datasets we follow a development model called \"Fork and Pull Model\". You can find more information here:\r\n- [Understanding the GitHub flow](https:\/\/guides.github.com\/introduction\/flow\/)\r\n- [Forking Projects](https:\/\/guides.github.com\/activities\/forking\/)\r\n\r\nAlternatively, if you find all these steps too complicated, you can use the GitHub official command line tool: [GitHub CLI](https:\/\/cli.github.com\/). Once installed, in order to create a Pull Request, you only need to use this command:\r\n```shell\r\ngh pr create --web\r\n```\r\nThis utility will automatically create the fork, push your changes and open a Pull Request, under the hood.","embeddings":[-0.2110829651,-0.2354261875,0.0539522693,-0.120353587,-0.0160378627,0.0573175848,0.0736910924,0.3318694532,0.2753102779,0.0360611007,-0.1463584155,-0.1400417089,0.3278672099,0.3584648371,0.2050697505,-0.164364472,0.1016285419,0.0192579348,-0.3486397266,-0.0259078201,-0.2046322525,0.1603539288,-0.0323321708,0.0914419442,-0.0280473791,0.019089058,-0.293988049,0.3911278546,-0.0363712646,-0.472815603,0.4034941792,0.3701826334,0.2042797953,0.3489311934,-0.0001184653,0.0142525993,0.2367955893,-0.0410921276,-0.2115555257,-0.1077309176,-0.2645911574,-0.4390313923,-0.1316624433,-0.1381315887,0.034405373,-0.0912179872,0.1589486897,0.3369330764,0.3802199364,0.2255799621,0.1258405149,0.4230258167,0.4710550904,-0.2185387313,0.0862675235,0.0614392348,-0.1282148063,0.2936409712,0.0898409784,-0.0591585785,0.036488492,0.2152909189,0.0781656206,0.2497442812,0.4219130576,0.0737006068,0.1608163565,-0.3957279325,0.1415680498,0.2385166436,0.6311095953,-0.3547596037,-0.3796705306,-0.0760816336,0.0333204195,-0.1144907176,0.1499064118,0.0264801923,-0.1659982949,0.3294635713,0.0786603838,-0.5474177003,-0.2395004928,0.231194675,0.1810732931,0.2853419185,0.0017947436,0.0789149925,0.379914701,-0.132351324,-0.1323436052,0.1575139761,-0.2554055154,0.1031677052,0.1107840315,-0.2356530577,-0.110346429,-0.0550629981,0.4048175812,0.0613307059,-0.3543204963,-0.04447468,-0.2573795617,0.0111754378,0.174470976,-0.0149969887,-0.1784735471,-0.0599963218,0.4562728703,0.4776796997,0.1987167001,-0.1451387405,0.2203816921,-0.2268173546,-0.3670242727,-0.2010921836,0.4068787694,-0.2196868807,-0.0714218169,-0.304386586,0.1304107904,-0.0501970984,-0.1943059415,0.4044190347,-0.121839188,0.0369838066,0.0639911965,0.2810359597,-0.1574193239,-0.2196881771,-0.1134860739,-0.0748006925,0.1971755177,0.1814006269,0.0033577275,-0.0108487029,0.0622827858,0.1343208253,-0.1932731718,-0.1267478317,-0.2192617208,0.2691918314,-0.0305070169,0.2095509768,-0.0249657277,0.2104760557,0.1229593828,-0.11837174,-0.1643955708,-0.025780892,-0.2253971547,-0.2972949743,-0.3480557799,0.0438442752,-0.0791406706,-0.3015474677,-0.0541525371,0.0778256878,-0.3137895167,-0.1640633941,0.0472615808,0.1616930962,-0.2903907001,-0.1199439019,0.3145063221,0.6965306997,-0.2458405495,-0.3773680925,-0.1598867178,-0.1359545439,0.0721240714,0.3017199039,-0.1499699354,0.1195338294,-0.2713197172,0.1623042077,0.2666556835,-0.4356101751,-0.5098182559,0.2222330719,-0.0900063813,-0.0561342798,0.0360271595,0.0909888521,0.1352198273,-0.2080696225,-0.5479161739,0.0284518823,0.0575360544,0.0603860132,-0.2171857506,-0.5456910133,0.0474578962,0.0427253917,0.1120493263,0.348154068,0.4164614975,0.1177427024,0.3645390868,-0.1829752922,0.1338173151,0.2897298336,0.5038339496,0.4074414074,0.0071039628,0.0301901512,-0.2982338071,0.1431921422,0.0423161201,0.254090935,-0.1266067922,-0.1244278774,-0.4942826331,-0.0951791778,-0.2824363112,-0.1889820695,0.0047092726,0.067904003,0.0674922094,-0.0355721414,-0.1386614889,0.3957072198,-0.2146778256,0.1661385596,-0.1394138038,0.5787001848,-0.160461396,0.0581317246,0.1075868532,0.1332895905,0.2444794476,-0.2259410769,-0.1514311284,0.4178691208,-0.421497345,0.2410955578,0.5301553607,0.2047100812,0.2471736968,-0.5035048723,0.3256246746,0.1582865864,-0.052460853,0.1165192947,-0.0072110454,-0.0388913155,0.0617169142,0.224024564,0.0515305176,0.063894242,0.4288527966,-0.2204475105,-0.024872506,-0.1123149991,-0.2210939229,-0.4116818607,-0.0536614023,-0.312276423,-0.1100313067,0.2835019827,0.5371803045,-0.1335303336,0.1254914105,0.4350464046,-0.0412507504,0.0413592607,0.0192504972,0.2674684823,0.0542728119,0.0737496838,-0.1289021671,0.2507378757,0.0435523689,0.0266730916,0.1505316496,0.1345540583,-0.0677167922,-0.0425433218,-0.0737454891,-0.1119700521,-0.0348199792,0.0337275825,-0.0966105834,0.2112507522,-0.3681170046,-0.0346296318,-0.3610076606,-0.6220450401,-0.3655466437,0.1438074112,-0.4196515679,-0.2589598596,-0.0270658415,0.3142710328,-0.1000511646,0.1556216925,0.1236823946,0.1750087142,-0.1223044842,0.2079287022,-0.2587126493,-0.0562148727,-0.1604387611,-0.0437365137,0.267429173,0.2079910338,0.3205575943,-0.5224847198,-0.1317167729,-0.9016187191,-0.114495188,0.2193331718,0.0459594578,0.1146957725,0.3226823211,0.4915757775,0.1684923619,0.0363832712,0.2897359431,-0.1280543506,0.0844142139,-0.271086067,-0.1807758361,0.1145590991,0.0175795574,-0.0776854381,-0.5018087029,-0.3658218682,0.3381377757,0.0024151437,0.0508712232,0.2050856948,0.1885677725,0.0024425085,0.1367033124,-0.1255963445,0.0093972124,0.056289494,0.3124879003,-0.2125174254,-0.3424602151,0.3461087048,0.2088090777,-0.062254671,-0.245833382,-0.5297685266,-0.1404603124,-0.0356180146,0.5182726979,0.0156707373,0.4950365126,0.1784035861,-0.0216535311,-0.0080255382,-0.0368843488,-0.223123759,0.0537452884,0.2179126143,0.3834924102,0.4108876288,0.4933067858,-0.1012102738,0.6489822268,0.1351569891,0.0992670879,0.3816055059,-0.2987954617,0.3904784322,-0.1469760239,-0.2986433804,-0.0016428845,-0.1519547552,0.2234380841,0.3047828078,0.1389473081,0.1592380553,-0.0936533362,-0.3841909468,-0.1008109003,-0.4200043082,-0.1524870843,0.3653809726,0.1339368373,0.2260614187,-0.0965342894,-0.387535274,-0.1447794288,0.4546304345,0.4856206477,0.078836076,0.1901089549,0.2211530656,0.0364671014,-0.446913451,0.2903983295,0.1320078373,-0.0227104556,-0.0453885198,0.0428981557,0.1085168198,-0.1409751475,0.5532923341,0.0112177711,0.2167460173,-0.098550044,0.0229880661,0.0994178429,0.0107957982,0.0437108502,0.1837855726,0.3269138336,0.0586394519,-0.4327537715,-0.4608087242,0.4464084208,0.2359813601,-0.1968592852,-0.0085386978,-0.0412100405,-0.5688753724,-0.7045999169,-0.1678563505,-0.0078561744,-0.1879734993,-0.1209304109,0.0776064843,0.1903192699,-0.1568928957,-0.1972069889,0.0078230454,0.180177629,0.2754974961,-0.0363739617,0.1566656381,-0.0247719921,0.1078745797,0.4978237152,0.0052092373,-0.2094107717,-0.1368772089,-0.0539466254,-0.1945204437,0.3270237446,0.2149183601,0.279404521,0.5513019562,0.2594561577,-0.2275274992,0.0053491923,0.3613304496,-0.0919298157,-0.3410687447,-0.3720963001,0.5128759146,-0.2919176817,0.0253153592,0.3009582162,0.61417979,-0.0424414761,-0.1070446372,-0.2116154283,0.8171716332,0.109803766,0.023318097,0.3205430508,-0.3169647753,0.4993389249,-0.1834737509,-0.0700716823,-0.2237212062,-0.1022569388,-0.1298687756,0.1635369062,-0.147566691,-0.0211655665,-0.0248915479,0.4875751138,0.2194264829,0.1664318293,0.1990476996,0.400198251,-0.1482166052,-0.1841365844,-0.3412180543,0.1323917508,-0.1016489565,0.520121336,-0.042682752,-0.3606900871,0.0399969593,-0.1636303663,-0.4959304035,0.0977746993,-0.2116215527,0.1316365898,0.0644881725,0.1285904497,0.189642027,-0.1768202484,0.4203735888,0.089101769,-0.3705147505,0.2290134579,-0.0821927637,-0.0882164985,-0.0938065946,0.3628078699,0.1292979121,-0.293597728,-0.4153259695,-0.1198916882,-0.0288055018,-0.3968341053,0.1516192704,-0.3641491234,0.0677460954,-0.0485799313,-0.0987287611,0.0722704753,-0.0735659003,-0.3161447346,0.0389035046,-0.1556701213,-0.0997206867,-0.1470519304,-0.0951972008,-0.0044186404,-0.1197341904,0.2152938992,-0.4728781581,0.0231328718,0.4791007936,0.431712538,-0.0830757767,-0.1388288438,-0.2956309319,0.0571263246,-0.802334249,0.2080302536,-0.0347241722,0.1268723905,-0.2369491607,0.510419488,0.1861512959,0.0303394496,-0.1408406794,-0.3329785466,-0.1892120242,0.1432460696,0.1213506982,0.1389914453,-0.0154818259,0.1325033009,0.1311821938,-0.3261322975,-0.2453271449,-0.0592157952,0.008890464,-0.1425931007,0.5837344527,0.067737259,0.3327388167,0.0385996774,-0.0028793807,-0.0786760077,-0.1512919366,-0.0153798182,-0.3094735146,0.1270186454,-0.1938003451,-0.3463674486,-0.0661300942,-0.2369443774,-0.0994160771,0.024589546,0.1164675355,0.067376107,-0.0998676643,0.1007109731,0.3244679868,-0.232887581,-0.1834412962,0.1127819195,0.4962287545,0.5475104451,-0.1123683229,0.352732569,-0.2155138552,0.0361016691,0.0072236066,0.1517478377,0.2413236648,-0.1230806336,0.4092426896,-0.096215114,0.0337092765,0.1993798316,0.2724160552,0.2172192037,-0.0301291291,0.0002514149,0.1970577389,0.0334303491,-0.2407176644,0.0043878057,0.2773887813,-0.3165876865,0.0001485683,0.2445091605,0.2434202731,0.0326868035,0.2097510099,0.1546023637,0.3843667805,-0.1354887187,-0.0059501389,0.2811081707,-0.0566391014,0.0723370239,-0.1455577016,0.1488766968,0.2666094005,0.0812095925,-0.2176889479,-0.0191493779,-0.0000560277,0.264621079,0.0688580871,-0.4345844388,-0.0124289896,-0.1814429462,0.1980892271,-0.12556687,0.0234949179,0.2571657896,0.0722443834,-0.3908340037,-0.1862327158,0.6674995422,0.272969842,-0.0067738737,-0.1617906839,-0.1766159981,-0.2412552088,0.0472540781,-0.1085466221,-0.3009735048,0.1466254443,0.0342855826,0.1258672923,-0.2188074142,-0.0567172877,-0.1533828378,0.35543558,-0.1098856628,0.1494349539,0.3063849807,-0.1160859615,0.099416703,0.2421589941,0.4903849065,0.088044323,0.111821875,0.1783021539,0.1303446144,-0.1638832092,0.1486399025,-0.1357665956,-0.151882574,0.1162733585,0.1003921032,0.0547589883,-0.0963516831,-0.2757140398,-0.2807415426,0.0327413268,-0.125804022,0.431625843,-0.186540097,-0.1493148208,-0.289416641,0.0295164939,-0.1919047534,0.0120491609,0.1910400838,-0.2029003501,0.0164059736,0.0376013331,0.0303533282,-0.1848806292,0.3830525279,0.3688980341,0.1084756851,-0.0404469594,-0.1493473053,-0.5221028924,-0.091198951,-0.0646284595,-0.3110883236,0.0818893686,0.0216680784,-0.1355516165,-0.0517254658,0.4342083931,-0.3080391884,0.3608540297,0.1353235543,0.1233096644,-0.2188129276,-0.3192842901,0.3344915509,0.1019834056,-0.1856915802,0.100576058,0.0284424014,-0.1047628373,0.2286315858,0.0314073563,0.0685322881,-0.7217945457,0.0577107929,0.010811517,0.2015742362,-0.2599645555,-0.4110692143,-0.2814151943,-0.340572685,-0.3281993568,0.1083416417,-0.2560905814,0.3493722677,0.0975079164,0.0146316746,-0.5679985285,0.0393325463,0.1083654165,0.1224358007,-0.1601153463,0.2394272387,-0.2191828191,-0.024632195,0.5274119377,0.0141111873,-0.0824992433,0.0562009215,-0.3022766411,-0.1346304715,0.5225575566,-0.1165839732,-0.083286263,-0.0401237644,0.2780090272,0.1697811037,-0.3663477004,-0.4627892673,0.1363793314,0.0041427407,0.2846258581,-0.0590315498,0.2200993448,-0.1446905285,0.0008831011,0.1061491519,0.1828067601,0.1683296114,-0.3448531032,0.0616682135,-0.0189395174]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2075","title":"ConnectionError: Couldn't reach common_voice.py","comments":"Hi @LifaSun, thanks for reporting this issue.\r\n\r\nSometimes, GitHub has some connectivity problems. Could you confirm that the problem persists?","body":"When I run: \r\nfrom datasets import load_dataset, load_metric\r\n\r\ncommon_voice_train = load_dataset(\"common_voice\", \"zh-CN\", split=\"train+validation\")\r\ncommon_voice_test = load_dataset(\"common_voice\", \"zh-CN\", split=\"test\")\r\n\r\nGot:\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/common_voice\/common_voice.py\r\n\r\nVersion:\r\n1.4.1\r\n\r\nThanks! @lhoestq @LysandreJik @thomwolf ","comment_length":20,"text":"ConnectionError: Couldn't reach common_voice.py\nWhen I run: \r\nfrom datasets import load_dataset, load_metric\r\n\r\ncommon_voice_train = load_dataset(\"common_voice\", \"zh-CN\", split=\"train+validation\")\r\ncommon_voice_test = load_dataset(\"common_voice\", \"zh-CN\", split=\"test\")\r\n\r\nGot:\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/common_voice\/common_voice.py\r\n\r\nVersion:\r\n1.4.1\r\n\r\nThanks! @lhoestq @LysandreJik @thomwolf \nHi @LifaSun, thanks for reporting this issue.\r\n\r\nSometimes, GitHub has some connectivity problems. Could you confirm that the problem persists?","embeddings":[-0.4876464605,-0.0846986622,-0.0791777521,-0.0048034391,0.3903559744,0.0129868183,0.2704308629,0.3016326427,-0.063188158,0.3089751303,-0.2126752585,-0.1347054243,0.1226182878,0.0170099232,-0.0094278194,-0.0931782797,-0.0044788229,0.00298292,0.192533344,0.0247693788,-0.0387781002,0.0858430639,-0.1760057956,0.3123496771,-0.1892662793,-0.0507797785,0.0589113794,0.0770447105,-0.2132197171,-0.3560753465,0.4676538706,0.0277777296,-0.041264724,0.8190293312,-0.0001130046,0.1451916546,0.4414004385,-0.068325147,-0.3076119423,-0.4945730567,-0.0904799625,-0.0165243447,0.1303874403,0.1618437022,-0.1614775807,0.0518424325,0.0687628835,-0.2308126092,0.3591995537,0.3560591638,0.2153366059,0.2006513178,0.3548300266,-0.1868588775,-0.052108977,-0.1892620027,-0.0218943432,0.8155437112,0.0865337104,-0.22358796,0.2252188474,0.1362261027,0.1147135571,-0.0424558148,-0.0971720517,-0.0446065627,0.1151468083,-0.3021116853,0.0465990752,0.2485738397,0.4080207646,-0.4108532369,-0.3393287957,-0.0096714012,0.0688005537,-0.1919004619,0.1794370115,0.0755963624,-0.3747929335,0.08968997,-0.4395814538,-0.3170163929,-0.1687384248,0.2219960839,-0.0388029628,0.0705616921,-0.111549288,0.0763533041,0.178044945,0.0002032817,-0.3434875906,0.1713796258,0.0020537416,0.405492574,-0.5099363327,-0.0270293746,-0.2446025461,-0.3834081292,0.3409876525,-0.0012129704,-0.0660155118,-0.0444582701,-0.2184315771,0.1957034618,0.3238674998,0.2319249064,-0.0145540545,0.227080971,0.4110629559,0.2189212292,-0.0950866938,-0.0442090929,-0.3507368565,-0.3648319244,0.1944305003,-0.013700786,0.380974561,-0.302343756,-0.2657470405,0.095067665,-0.2045124918,-0.0274013709,-0.0486893654,0.3798324764,-0.1350037754,0.0382551141,-0.1236829087,0.4221954346,-0.118559815,0.0815408006,-0.2166409343,-0.0903094262,-0.1128263101,0.097669512,0.0493132956,-0.3984252512,0.2074008137,0.0379754826,0.3572260141,-0.3074057102,-0.0891105384,-0.0856576413,-0.5720186234,0.1642294973,0.0533686914,0.2146516442,0.2993554473,-0.1587567925,-0.1170123294,-0.1049671918,-0.3046244383,-0.471655488,0.0482198186,0.2257326394,-0.0102702016,0.0569618605,-0.4090845883,-0.1771777272,0.1486499459,-0.2936156392,-0.0880606547,-0.301613301,0.0597571023,-0.1316051185,0.4870406091,0.0713867545,0.0527719595,-0.1694486588,-0.0351202115,-0.1018608809,0.0548310429,0.526974082,-0.2484565973,0.0601397119,-0.2221782804,0.156096682,0.521710515,-0.652610898,-0.4879722595,0.4022220075,-0.235869348,0.0671707466,-0.0939288139,-0.2874566615,0.0392743908,-0.0323440395,0.5161759853,0.1174638942,-0.0979949012,-0.1968774498,0.0407909006,-0.2047234774,0.1706810594,0.0211102925,-0.0097485371,0.0787399635,0.1623610556,-0.0269786585,0.416403383,0.10372857,0.009008836,0.2248878032,0.3108515441,-0.2277763933,-0.1555938274,-0.1596635133,0.1518708616,0.1177366078,0.2797800601,0.3692600131,-0.1291464716,-0.1252070367,-0.4446752965,-0.0257855784,-0.2310210466,0.0622798614,0.056808766,0.0320310742,0.0646314397,0.219354555,-0.1840764135,0.1622597575,-0.2816715837,0.1097330898,-0.1320050806,0.4692251086,-0.2606515586,-0.0348062143,0.2211777568,0.2767660618,0.0456460901,-0.0788386613,-0.2147000879,0.3152984679,-0.3178600073,0.2351754904,0.1205977276,0.283167541,0.1175950915,-0.591106534,0.0591872707,0.1485545635,-0.0232406184,0.0424575247,0.3778440356,0.2386407107,0.1728033572,0.4496591091,0.0726335719,0.08587493,0.290910095,0.24765715,-0.134275645,0.1453574449,0.1616977602,-0.1719187647,0.3192304075,-0.1830841154,-0.340532124,-0.1651495099,0.3829872012,-0.0832233503,0.107893385,0.1136161461,-0.1390830427,-0.1129106358,-0.0132415881,-0.0086587472,0.2383928448,0.0911889374,0.0106042596,0.3226593733,-0.2620108128,0.0036352507,0.0549353063,0.0583234318,0.1743072122,0.2589446008,0.02267267,-0.0217205696,-0.0456663743,-0.2580975294,-0.209335506,0.1478906423,-0.1808155328,0.1480846554,-0.086663872,0.0132794818,-0.0805373564,-0.3049543798,-0.5015799999,-0.2510821521,-0.1252114028,0.1089010015,0.3503382802,0.2614883184,-0.0398212261,-0.1008044556,0.1907769144,-0.2764284909,0.0604965203,-0.2182856798,-0.2243892103,0.070605509,0.0844244957,-0.312320143,0.0233451128,-0.2295905203,-0.1374640465,-0.0782050639,-0.1175296307,0.0083509637,-0.2455486953,0.3056176305,0.1816340238,0.3986811042,0.2330412269,-0.0065053413,0.4217598736,-0.2188811749,-0.0219951048,0.0442473516,-0.0634039417,0.0226563644,0.1767362803,-0.2832997441,-0.4118851721,-0.3967009187,0.2375123203,-0.1698881388,0.050104972,0.2355161756,0.123425521,-0.0473635271,0.0705346987,0.3045687377,-0.1642432511,-0.3654202819,0.0484303348,-0.0856717601,-0.1601704806,0.1718711555,0.2391327024,0.5762704611,-0.0725358129,-0.4448835552,-0.0168504268,-0.0534744635,0.3242331147,0.1021109298,-0.0072238008,-0.0170469321,-0.0905972868,-0.057672061,0.140351668,-0.2352853864,-0.0629813671,0.2043684125,0.3051237464,0.1238506883,0.5398365259,-0.1348452121,0.6648232937,0.2927260399,0.1564437598,0.187933147,-0.2086391002,0.341358602,-0.2027694732,-0.4419165254,0.1032795236,-0.17082946,-0.0757673755,0.0826868042,0.1749288887,0.0357116945,-0.3792596459,-0.4515118003,-0.1404634565,-0.2120096982,-0.2305962294,-0.2494820654,0.0973121449,0.0886196792,-0.0255278293,-0.1131446213,-0.0858877599,-0.0161934476,0.4476886094,-0.1347055882,0.1376412213,-0.2322180271,-0.2518394589,-0.3093474209,0.1862034351,0.0572023392,0.4124842584,-0.3871543407,0.1881733388,-0.1086357981,0.0231670737,0.1757865846,-0.1260646433,0.0896162763,0.0296851434,0.0075163576,-0.181434229,-0.0671502799,0.1948509365,0.3333878517,0.3082720041,0.2268792093,-0.1649070531,0.0436123647,0.1503461897,0.4491624534,0.0521083139,0.0356226973,-0.3161221743,-0.2028121203,-0.3019143641,0.0451673716,0.0064080968,0.3169747591,-0.1787019074,0.0400927886,0.3380233943,-0.2488067597,0.235093087,0.045886755,-0.0795388669,-0.3574636579,0.1848311126,0.2858290672,0.0418102853,0.0671863556,0.7470647097,0.1048019752,-0.5249845982,0.0151386308,0.2105799764,0.3927266002,0.3051048517,0.014523617,0.2129776776,0.2763592005,0.0847397894,-0.027172735,0.1968183517,0.3358708322,0.3391829431,-0.6512300372,-0.5451788902,0.1240202859,-0.1648375094,-0.0133868754,0.3071778417,0.155250147,-0.1331917942,0.0506518781,-0.5677531362,1.0728435516,0.1851099133,0.0811420456,0.0369697548,-0.0408568904,-0.0429544412,-0.315456748,0.0794291422,-0.1125622839,-0.1045742258,-0.0135762375,-0.2059077173,0.0688670874,-0.1962565035,-0.0762907341,0.4810021222,-0.0870863721,0.0726328045,0.1857952923,0.2683596611,0.0136697991,0.2445689589,-0.2734532058,0.1776363552,-0.0487892963,0.5176771283,0.0022685332,-0.0941772833,-0.2435061634,-0.3722907603,-0.3688099682,0.2839083076,-0.1193805709,0.0671749189,0.1856503487,-0.2672021389,0.2050154358,0.1718738377,0.3522239923,0.2981653214,-0.2847108245,0.1762105227,-0.0071891812,-0.0778330266,0.1546156108,0.1250278354,0.0521410741,-0.1679114401,-0.5022818446,0.3732266426,-0.0156885479,-0.3964534402,-0.1300135255,-0.2442804724,0.0088551659,-0.1638147086,0.2053537667,-0.3347201943,0.172163412,-0.323649019,0.1263958663,0.1891030818,-0.1880552918,-0.1597912461,0.2088700384,-0.1925046742,-0.0563026033,0.2488451898,0.0497435294,-0.138581425,0.4595989883,0.1396304667,-0.0454388522,-0.244351998,-0.0585724711,0.5368093848,-0.4719840884,-0.0285093617,-0.1112322137,0.1944278032,-0.1658237278,0.4409452975,0.2032209188,-0.0784861296,0.1024987474,-0.4698055685,-0.3129362166,0.1936557293,-0.0615784042,0.2066180855,-0.0816062912,0.1082931161,-0.0067187259,-0.0445609167,-0.2897928357,-0.107799612,-0.3431855738,0.020967206,0.3339733779,-0.191400677,0.3850352466,0.0402169041,0.2257783711,-0.2455058992,-0.1808192432,-0.1469322145,0.0213194564,0.1212763041,0.0197617225,-0.253361702,0.145655483,0.0564526618,-0.1473715007,-0.0086093983,0.1858487427,0.0480350964,0.0552308187,-0.1347244531,-0.0723068118,-0.1145794317,0.0543710627,0.1188735813,0.0379201025,0.4012290239,-0.0092393728,0.1457679272,0.0158031732,-0.2288104445,-0.4103943408,0.0634874701,0.1104698777,0.0188781656,0.3517830372,-0.3496560156,0.0694843754,0.0487370379,0.6509897709,0.1022275686,-0.3533657193,0.2629568875,0.2502115071,0.1545456946,-0.2229312658,-0.024043452,0.1632434428,-0.1636535972,-0.0173640903,0.1162429601,0.1822680533,0.0582860857,-0.0675276369,-0.0943046287,0.4997424185,-0.0856313556,0.2728968561,0.1634151191,0.119103767,-0.0082134074,0.1074966788,0.1330451667,0.0430902801,0.3482432961,-0.5331619382,0.0900310054,0.0672301129,-0.0937197283,0.162071228,-0.1775238216,0.1025068834,0.0001444996,-0.3571614325,-0.2571325302,-0.1978606135,0.4001483917,-0.3733700812,0.1429115385,-0.2302741259,0.1050118357,-0.055316586,-0.1685611308,0.0117503908,-0.2862842679,-0.2023342252,-0.1951297671,0.0582066029,-0.0672026351,0.2015394419,-0.0534078926,0.0870338753,-0.301774174,-0.1426230371,0.220174551,-0.0625626668,-0.1032322943,0.3250924945,0.1004205123,-0.0980515406,0.2333034873,0.2825142443,0.758891046,0.172331363,-0.0794540718,0.2104049474,0.1219781414,-0.1308842748,0.2585339248,0.0486163571,0.2171356976,0.1646729708,0.2685927749,0.2097799927,-0.1216758266,0.2220005691,-0.1417058706,0.0521771796,-0.331489563,0.4694357216,-0.4028268158,-0.1831299663,-0.4903372526,0.1656341255,-0.4790592194,-0.0280084312,0.4250092804,0.1137313694,0.0196254402,-0.4083964527,0.0869360715,-0.0538710058,0.3900737762,0.3639168143,0.4856379628,-0.1029978991,-0.1019115448,-0.4100945592,0.3474363685,-0.1695381999,-0.0634160191,-0.105685994,-0.1028937101,0.0326903164,0.2012596279,0.2235747427,-0.1411512345,-0.1996715814,0.012638093,-0.0142848995,-0.2394862324,-0.0034862948,0.1372597218,-0.0184228756,-0.2118452787,0.1580017805,-0.4327087998,0.025136888,0.0708989426,0.120217666,0.0879019573,-0.3852389753,0.2949137986,0.1643312126,0.7388056517,0.0270251576,-0.194335103,-0.0884250104,-0.3569730222,-0.1504110098,-0.009831639,-0.2185794562,0.4324523509,-0.1128251106,0.0285941269,-0.3783828318,-0.1291139424,-0.0339355692,-0.1541231573,0.0211208537,-0.0682815835,0.0382540897,0.0142834047,0.220672369,-0.1315783262,0.0165925734,0.0929292217,-0.1390490681,-0.1497680545,0.4321288168,-0.1975663453,-0.2928907871,-0.335683316,0.2677125633,0.013755938,-0.1037763655,-0.4707442522,0.2704683244,0.4233447015,0.3171430826,-0.3321478069,0.1437587291,-0.0832421407,0.1260186732,0.1612603962,0.2068045884,-0.3250118196,-0.2565755844,0.0909332484,-0.1341870129]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2075","title":"ConnectionError: Couldn't reach common_voice.py","comments":"@albertvillanova Thanks! It works well now. ","body":"When I run: \r\nfrom datasets import load_dataset, load_metric\r\n\r\ncommon_voice_train = load_dataset(\"common_voice\", \"zh-CN\", split=\"train+validation\")\r\ncommon_voice_test = load_dataset(\"common_voice\", \"zh-CN\", split=\"test\")\r\n\r\nGot:\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/common_voice\/common_voice.py\r\n\r\nVersion:\r\n1.4.1\r\n\r\nThanks! @lhoestq @LysandreJik @thomwolf ","comment_length":6,"text":"ConnectionError: Couldn't reach common_voice.py\nWhen I run: \r\nfrom datasets import load_dataset, load_metric\r\n\r\ncommon_voice_train = load_dataset(\"common_voice\", \"zh-CN\", split=\"train+validation\")\r\ncommon_voice_test = load_dataset(\"common_voice\", \"zh-CN\", split=\"test\")\r\n\r\nGot:\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/common_voice\/common_voice.py\r\n\r\nVersion:\r\n1.4.1\r\n\r\nThanks! @lhoestq @LysandreJik @thomwolf \n@albertvillanova Thanks! It works well now. ","embeddings":[-0.438498646,-0.2266579568,-0.0452615619,0.009236523,0.3884683251,0.0106349234,0.2323131412,0.2929386795,-0.0663182139,0.2616654336,-0.3805659413,-0.1440358907,0.075945504,0.0241745301,0.0246070623,-0.1632363796,-0.0137436157,0.1311538219,0.0792128742,-0.0234047249,-0.0853023753,0.0997953638,-0.2327105701,0.3559115231,-0.2548287213,-0.0750457942,-0.0081272162,0.0701377392,-0.257779032,-0.3404733837,0.4029130638,0.0253287535,0.0362400822,0.7781924605,-0.0001103523,0.0980273187,0.3813535571,-0.1037140042,-0.158028543,-0.4182621539,-0.0999600664,0.0082475208,0.081271939,0.1373164207,-0.0566391461,-0.1867436469,0.0429933928,-0.2036052942,0.3919630945,0.4508522749,0.2505361736,0.2542674541,0.2065331936,-0.2217870355,-0.2403818816,-0.2194590718,-0.0388731733,0.6860201359,0.0053741718,-0.1986454129,0.2055176198,0.14621602,0.0758624673,-0.1532789469,-0.0585294589,-0.0060370318,0.0218541306,-0.2971402705,0.0486398488,0.3641022146,0.4184179604,-0.4129585624,-0.3394877613,0.0377059095,0.0354408622,-0.2767094374,0.1723294705,-0.0767976344,-0.2390036583,0.1435801536,-0.3518535197,-0.2229428887,-0.1674263179,0.1795930564,-0.1055347919,0.0659756064,-0.0797114596,0.017785972,0.2292769551,0.0211089347,-0.1006491557,0.2084257156,0.0720879659,0.3753770888,-0.5787682533,-0.0307315923,-0.1427930892,-0.1970821768,0.3019533753,-0.0428591333,0.051434245,-0.0455117747,-0.1537582129,0.1790141314,0.2289915532,0.2924597561,0.0756106675,0.0991993025,0.3418069482,0.2139629722,-0.0366601348,-0.0909934863,-0.2272461206,-0.411057651,0.2610018849,0.0443827622,0.2667959332,-0.2850747406,-0.2843031585,0.0938717946,-0.1525906771,0.0053820373,0.0916826054,0.5348829031,-0.146566987,0.185410589,-0.0949731618,0.3732242286,-0.1832669526,0.0471825525,-0.2475320995,0.0378435254,-0.1580114365,0.1217852607,0.1517246515,-0.2745088637,0.2504462004,-0.0031982057,0.3690672219,-0.2203176767,-0.0387569591,-0.0985490754,-0.478076607,0.1165549085,0.0368048102,0.1310695857,0.299357146,0.0588497072,-0.1080306843,-0.1422370821,-0.2721568644,-0.4326581955,0.144835636,0.2444500178,-0.0209006928,0.035557121,-0.2026255578,-0.0214136615,0.096348539,-0.4408998787,-0.1163082719,-0.2170477509,-0.109177202,-0.1303443611,0.5414108038,0.1193028763,-0.1780973822,-0.1227180064,0.0624206215,-0.1230562925,0.0654191896,0.5125998259,-0.2160577774,0.0868313387,-0.1987619698,0.2737097442,0.5761854649,-0.6269261241,-0.4821455479,0.3178721368,-0.2263844907,0.015174848,-0.045991268,-0.2680706978,0.0165668949,0.081968911,0.4295030832,0.1380676329,-0.0752796382,-0.2110837847,0.056251023,-0.0963313952,0.2016203851,-0.0300312433,-0.0070932652,0.0566294901,0.0783753991,0.1033427119,0.4234933555,0.0732412562,0.0711943805,0.1767469794,0.2945969999,-0.2200978249,-0.119992964,-0.2180217505,0.1492814273,0.1163967103,0.103859812,0.3270003796,-0.013845088,-0.1695354432,-0.4868610799,-0.0839024335,-0.2739098072,0.0503218956,0.0958105847,0.0324796587,0.1228637919,0.2004476637,-0.2209336758,0.1788318604,-0.130670175,0.1376868337,-0.1562128961,0.3660401404,-0.2904378772,-0.0261882152,0.1252603531,0.2352154255,0.0730512068,-0.0644028336,-0.2124024034,0.4009394348,-0.3150076568,0.2715595961,-0.1104452088,0.1734147668,0.127357319,-0.560741961,0.0750966296,0.0728962421,-0.0191418491,-0.0092767784,0.3473390639,0.3000169694,0.156189546,0.352306664,0.1206477284,0.0131147206,0.1786147803,0.234867394,-0.1244225949,0.1378135979,0.2083551288,-0.3323960006,0.3165815771,-0.1490925103,-0.3209622502,-0.2034544647,0.4428797364,-0.0579563789,0.1314612776,-0.0038656523,-0.1255061328,-0.1011735573,0.0726249442,-0.0519562215,0.3961217701,0.14438124,0.0488530844,0.3679144084,-0.2587763369,-0.0412066542,0.0626566708,0.0014906474,0.1796446741,0.3573134243,-0.0711226836,-0.0577834472,-0.2067415863,-0.2675932646,-0.1865456849,0.121403046,-0.2125662118,0.115033254,-0.1270429045,0.1080007777,-0.0891405493,-0.2750678658,-0.3762362897,-0.2881620526,-0.0784956366,0.0950948372,0.3496240973,0.2956336737,-0.2785613835,-0.082799457,0.196729809,-0.2529861331,-0.0302363988,-0.05042978,-0.2145926803,0.0985595584,-0.0545816272,-0.145219624,-0.0111914948,-0.1641635597,-0.0566981882,0.0488971025,-0.204488337,0.0202095248,-0.1735681593,0.2301898748,0.2279120982,0.3396685719,0.1084662974,-0.0353630781,0.4018682837,-0.3214784563,-0.0749910027,0.1111713201,-0.0364722162,0.1012294963,0.1197772846,-0.4861486852,-0.436175257,-0.4732662737,0.2381197065,-0.3121365011,0.0105336644,0.2292150855,0.1117039025,0.0255072471,-0.0277288817,0.3817858696,-0.2081816494,-0.2464045584,0.1325988472,-0.1971360892,-0.2054867744,0.1031440273,0.1179006472,0.6502360106,-0.1383737177,-0.3371617794,-0.1413957924,-0.0572705381,0.2094275653,0.0434731953,0.0160657708,0.1319656223,-0.0818068311,-0.0772648454,0.1410474777,-0.1997102201,-0.1262044013,0.1207736284,0.3436855078,0.0741795748,0.5403285623,-0.1112968028,0.5098187923,0.2418302447,0.1961600035,0.2814741433,-0.2610464096,0.3632824421,-0.2679130733,-0.5452454686,0.1193651855,-0.167296946,-0.0246980749,0.1294128895,0.1413742155,0.1670038849,-0.4022504985,-0.3995702863,-0.1069084853,-0.191215083,-0.1765284538,-0.1322623789,0.0805158988,0.0900947377,0.0669178143,-0.0690527111,-0.0302113034,0.0856531486,0.3709770739,-0.2107679844,0.1462485939,-0.3400544226,-0.1836720556,-0.3878915012,0.1878745109,-0.0044836323,0.4295628965,-0.3706818521,0.0903719068,-0.106091395,0.0534263216,0.4226703346,-0.1216512173,-0.037448056,-0.0250958875,0.0507571138,-0.252949059,0.0208412912,0.1379290372,0.154004395,0.2387271672,0.2586797774,-0.103097707,0.0852543563,0.1260532886,0.4178279638,0.0079422081,-0.1165488735,-0.3257519305,-0.2110200822,-0.2840752006,0.018371094,-0.0551798493,0.3717788458,-0.1514441371,0.0629625618,0.3285206556,-0.2900640965,0.3837748468,0.1353391707,-0.0376635008,-0.3084029555,0.1960791796,0.3677443862,0.1111534089,-0.0209839046,0.6968169808,-0.048011858,-0.5971409678,-0.01097559,0.2475276142,0.5068572164,0.2300203145,-0.0542792045,0.1934650242,0.2887422144,0.0305666327,-0.0172287431,0.2766015232,0.3526761234,0.3211508095,-0.5120968223,-0.4621590078,0.0902839974,-0.005973944,-0.0441957675,0.1983999908,0.1783393621,-0.1897846758,0.0872225985,-0.6198353767,0.9910055399,0.2102651745,0.1172581315,0.1181183904,0.1503959745,0.0214601122,-0.2532121539,0.1059492305,-0.1481149197,-0.2292429358,-0.0321269669,-0.1343704015,0.1180962473,-0.110913977,-0.1800442934,0.5288247466,-0.0958058983,0.017436197,0.0744714066,0.2819085419,-0.0796244219,0.0722314566,-0.2505562305,0.1896792501,0.0735089406,0.490465194,0.0061670896,-0.056118384,-0.1572067142,-0.4204303026,-0.1033068076,0.248393476,-0.1978233159,0.0511511751,0.1949947476,-0.1728982627,0.2273744494,0.2733281851,0.2307356,0.3207000196,-0.2532390058,0.0887630209,0.0492625684,-0.1186147854,0.0947181657,-0.0184756964,0.1313877106,-0.128509447,-0.5727841854,0.3886054456,-0.0128614493,-0.2597396076,-0.1023028344,-0.2231040299,0.098029308,-0.2958098352,0.1090354174,-0.3507307768,0.1216204241,-0.2942973375,0.146093592,0.1890052557,-0.2820309699,-0.0837312639,0.2014028728,-0.1240404621,-0.0657528341,0.330065459,-0.0442874692,-0.2417904735,0.5062351227,0.0958271772,-0.0414935611,-0.2589404583,-0.0570080057,0.5460156202,-0.416433394,-0.100547716,0.0710687116,0.0635499656,-0.1303596497,0.4476863444,0.2620866001,-0.1124864966,0.0790090486,-0.393230468,-0.4977287054,0.2864905298,0.0470297188,0.2333477288,-0.041562777,0.1602092832,0.0119806007,-0.0190994963,-0.3302940726,-0.1743084043,-0.2098165601,-0.0352078117,0.2342799008,-0.2687643766,0.3611693382,0.0963156968,0.2580718398,-0.0474474803,-0.1755803823,-0.1804725826,-0.0155247748,0.1258653551,0.0862180591,-0.2064177394,0.1627457738,0.0366549045,-0.2497913539,-0.0828623846,0.069632642,0.114282988,0.055666402,-0.0950880125,-0.0777524263,-0.1443021297,0.082994841,0.0845454261,0.099147521,0.2434699833,-0.0559607483,0.1185789406,-0.0800636709,-0.2367780656,-0.2864160538,0.1507026702,0.198621735,-0.1217740402,0.2297977656,-0.3159283996,0.0103380028,0.0153633384,0.797783792,0.2040564716,-0.315032661,0.295302093,0.2739720345,0.1742625833,-0.292081356,0.0150852557,0.1197002381,-0.0937592387,0.048240684,0.1289141923,0.1516715884,0.0330065042,-0.0915447772,-0.1098548397,0.363871783,-0.1487619728,0.2459331602,0.1909465045,-0.0258897617,0.0826721564,0.3865770102,0.1568354815,0.0836964548,0.4010392427,-0.5951234102,0.1555020362,-0.1762747467,0.0218559224,0.0957257673,-0.2415126115,0.0238165539,-0.0743595883,-0.3620156944,-0.2035161555,-0.2608537674,0.6024509072,-0.4546140432,0.0709024891,-0.2980299294,0.1137510464,-0.114069663,-0.2575645745,0.0024085937,-0.2774196863,-0.1090536937,-0.2571306527,0.1361221075,-0.1444997787,0.2113354057,-0.0922729522,0.1316315979,-0.3072908819,-0.2774302065,0.2591454089,-0.0914023742,-0.0660492033,0.2788108587,0.0834247395,-0.0150344223,0.2247149795,0.2987574339,0.6470181942,0.170561552,0.0102705928,0.2250743061,0.0174141787,-0.1055953428,0.2763547897,0.1211407855,0.340497762,0.1611368209,0.3118077517,0.200834915,-0.1621732861,0.1993587762,-0.1008179113,0.1662516445,-0.3400015831,0.4337337315,-0.3279770017,-0.0672618672,-0.5286592245,0.2252018601,-0.6148167253,-0.0100849485,0.4739037752,0.0740453675,0.0797796845,-0.424459964,0.0855449215,-0.0384384915,0.2538048327,0.3030362129,0.5129186511,-0.1664479673,0.0169461761,-0.5550439954,0.3072921932,-0.0674612746,0.0136290155,-0.2314846963,-0.1046773642,-0.0357881896,0.1680673808,0.1512701809,-0.0422936007,-0.1134610623,-0.0745792165,-0.0050985473,-0.2642060518,-0.0030457468,0.1529938579,0.0059541157,-0.2285815328,0.1699103564,-0.4056281149,0.0334166102,-0.0369420573,0.0832271129,-0.0022931008,-0.1300114989,0.2771964073,0.2280357033,0.8279120326,-0.0513462387,-0.0117110666,-0.123432897,-0.2896063626,-0.1829050034,-0.0439000875,-0.0577869043,0.3475358486,-0.1248061508,-0.0269360039,-0.3445206583,0.0050720866,0.0601747185,-0.3262712061,-0.0877965316,0.0363953337,0.0224200189,0.0416516885,0.2247126848,-0.0526578054,-0.0321278945,0.0629705638,-0.0819967538,-0.3185665905,0.5844731331,-0.2284275442,-0.2508083582,-0.3156682849,0.2765185833,-0.010048232,-0.1769316792,-0.5306062698,0.2381880581,0.3302100599,0.2172377557,-0.2606558204,0.2087139785,-0.1814045459,0.0901542604,0.0744450763,0.2706765831,-0.1820732802,-0.2351362258,0.09134496,-0.1198875457]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2071","title":"Multiprocessing is slower than single process","comments":"dupe of #1992","body":"```python\r\n# benchmark_filter.py\r\nimport logging\r\nimport sys\r\nimport time\r\n\r\nfrom datasets import load_dataset, set_caching_enabled\r\n\r\n\r\nif __name__ == \"__main__\":\r\n set_caching_enabled(False)\r\n logging.basicConfig(level=logging.DEBUG)\r\n\r\n bc = load_dataset(\"bookcorpus\")\r\n\r\n now = time.time()\r\n try:\r\n bc[\"train\"].filter(lambda x: len(x[\"text\"]) < 64, num_proc=int(sys.argv[1]))\r\n except Exception as e:\r\n print(f\"cancelled: {e}\")\r\n elapsed = time.time() - now\r\n\r\n print(elapsed)\r\n```\r\n\r\nRunning `python benchmark_filter.py 1` (20min+) is faster than `python benchmark_filter.py 2` (2hrs+)","comment_length":3,"text":"Multiprocessing is slower than single process\n```python\r\n# benchmark_filter.py\r\nimport logging\r\nimport sys\r\nimport time\r\n\r\nfrom datasets import load_dataset, set_caching_enabled\r\n\r\n\r\nif __name__ == \"__main__\":\r\n set_caching_enabled(False)\r\n logging.basicConfig(level=logging.DEBUG)\r\n\r\n bc = load_dataset(\"bookcorpus\")\r\n\r\n now = time.time()\r\n try:\r\n bc[\"train\"].filter(lambda x: len(x[\"text\"]) < 64, num_proc=int(sys.argv[1]))\r\n except Exception as e:\r\n print(f\"cancelled: {e}\")\r\n elapsed = time.time() - now\r\n\r\n print(elapsed)\r\n```\r\n\r\nRunning `python benchmark_filter.py 1` (20min+) is faster than `python benchmark_filter.py 2` (2hrs+)\ndupe of #1992","embeddings":[-0.3541204929,-0.2729476094,-0.1845570356,0.0483277403,-0.0526438914,-0.0369342044,0.2611060143,0.2556811273,0.1619309336,-0.0271215755,0.0453823544,0.2863454819,-0.1358024031,0.2569368482,-0.2213481069,-0.0988899395,0.0234129801,-0.1742612422,0.2862676978,0.1069842502,-0.0914358571,0.0297238305,-0.1338146776,0.0263806377,-0.16958794,0.0381208248,0.1647137254,0.0501870476,-0.032465376,-0.2761442363,0.0294045378,0.086157158,-0.1570052058,0.7852604985,-0.0001080167,0.0136689041,0.1760716736,0.1456939876,-0.1060609519,0.10757754,0.2271268368,0.1655072272,0.198773101,-0.2053270042,0.0137587115,0.09550751,-0.1507163495,-0.4789972603,0.3586630821,0.1948708892,0.2567206919,0.3614610732,-0.2628139257,-0.1140683889,-0.088449128,-0.1553430408,-0.1256602556,0.0956059992,0.3734899461,-0.2850173116,-0.2208559066,0.4710993171,-0.2126052231,0.1906985193,-0.1027764231,-0.2353551537,0.2497159392,-0.2241068333,0.1920548677,0.1006830111,-0.1306595653,-0.3424929678,-0.2642036974,-0.1494506001,-0.2975366116,-0.4280293882,-0.129915908,0.0930568948,-0.1607154161,0.0077350917,-0.2119678557,0.1761474162,0.1228951812,-0.2883688211,0.0188410971,0.4286074042,0.0951959789,0.237973839,0.2803358734,0.2741017342,0.1921549737,-0.1838712543,0.42376405,0.0120285638,-0.3741009533,0.0452714078,0.1511151493,-0.0374723859,-0.1820900887,-0.1264330298,-0.0213255864,0.1311429143,0.5176249743,0.1346111596,0.0522633679,0.1266257167,-0.0699436888,0.151325509,0.0418289863,-0.0961197317,-0.5638108253,0.0576252639,0.3370461464,-0.2034296989,0.2287013978,-0.060331203,-0.1100637764,-0.0168804545,-0.3354436159,0.0561450794,-0.2400196344,-0.362569958,0.0606173947,0.2814366817,-0.1674857885,0.3311245143,-0.064846918,-0.0809895396,-0.4041422904,-0.1071732938,-0.3643832505,0.1707649082,-0.2238684148,-0.0225318447,-0.0276285727,-0.2990765274,0.1511617452,0.1863901317,-0.0211965032,-0.2113683075,0.4998669326,-0.3650711775,0.0190585945,0.2473016679,0.18362315,0.0645934865,0.238464877,-0.1192310154,-0.1718569398,0.0247950368,-0.1481731683,-0.2084772736,0.2935516536,0.2863411903,0.0700005889,0.0808007717,-0.0611215793,0.2531580925,0.2906790376,-0.1546072513,0.0137630887,-0.1659232527,-0.51577878,-0.2510307431,0.160472244,0.0920813903,-0.4152024686,-0.0039837849,-0.4379164875,0.0072038383,0.2997612655,0.3446548581,-0.1697324663,0.3606170416,0.0305589326,-0.0760374963,0.3521024883,-0.489236325,-0.1287090331,0.2404033691,-0.224209711,0.2888077497,0.237923488,0.0986120179,0.1812317967,-0.081372261,0.3108644485,-0.0741982386,-0.1402482241,0.0963311195,-0.2779759467,-0.0284555983,0.1745410562,0.0941732526,0.1768311262,-0.119111307,-0.20031178,-0.2174275815,0.3339415491,0.2191274613,-0.0749894604,0.1993047595,0.0609345883,0.0894334242,-0.0545012169,-0.4355373979,0.4204946458,0.217796728,-0.0035732286,-0.1188954711,0.490025878,-0.2131233066,-0.2280516922,0.1654280424,-0.1135434955,-0.3546546698,0.1870260686,0.0623004064,0.0841547996,-0.0654124767,-0.0182095785,0.4827525914,-0.222895056,-0.1618216038,0.0046090451,-0.1188710034,-0.0768597275,-0.2431431562,-0.188845709,-0.1292761564,0.0891400501,0.2274917364,-0.175512746,0.1898771375,0.1601392031,0.2017446756,0.2419621795,0.1069729775,-0.0517400466,-0.2112785727,-0.0409550443,0.4210873246,0.2231942415,-0.1430658847,-0.1019934043,0.4490645826,0.4078298211,0.2823445201,0.3364052474,-0.2149522603,-0.1468485296,0.0414532609,-0.0270479489,0.1572143584,0.3753861785,-0.1877102405,0.2038224488,0.1939772069,-0.2468114644,0.1521835923,0.5203676224,0.2015427053,-0.1335825622,0.179697901,-0.0946387053,-0.2470528036,-0.1296667308,0.082780309,0.364065975,0.2315972,0.1321682036,0.0419133604,-0.2350587547,-0.0090360455,0.2924793661,0.1011662781,0.114742063,0.0724982247,0.175853774,0.0268060286,-0.1594108641,-0.2262538821,-0.08992026,0.0656397566,-0.1785320044,0.2197624445,-0.0430209935,-0.1545688808,-0.1032814905,0.2017729282,0.1352043599,-0.1817783117,-0.0161875784,0.3327328861,0.1133791804,-0.0432271734,-0.0324150249,0.0937365443,-0.0058561363,-0.1147495732,0.0266118292,-0.1347233653,-0.1860307455,0.1979287118,0.0123174377,0.1549684405,0.4180132449,0.1599150896,-0.4260617793,-0.1023152396,-0.2994273901,-0.1184412166,-0.0411509126,0.3917999566,-0.0696886107,-0.0502058677,-0.2201803029,-0.3898442984,0.2141406983,-0.2776815295,-0.1835372895,-0.3034183383,-0.0282834452,0.1719668359,-0.0446822159,-0.2687133551,-0.3161673546,-0.2029335797,0.1687142551,-0.0167264901,0.1442136467,-0.1664030403,-0.121399343,-0.0633058697,0.1094888002,0.37180233,-0.2880044281,-0.1543732882,-0.2787105143,0.0313845053,-0.1079935431,-0.2688002586,-0.1057410985,0.0908426121,-0.070367381,-0.271022886,-0.0138979796,-0.2327632904,-0.1744315475,0.1096575111,0.0769628659,0.2127632499,0.1035173386,-0.2836925387,0.08268179,-0.2692316175,-0.2220862955,-0.2370165437,0.0682165772,-0.1733853966,0.4828255177,0.091060549,0.7160078287,0.4428505003,0.1457655728,0.0405845568,-0.1013761088,-0.106821306,-0.3627451956,-0.4419036508,-0.0116032856,-0.3123840094,-0.1468833238,0.2492904216,0.0615767725,-0.2763911188,0.0334707014,0.033063937,-0.4055764973,-0.0535578392,-0.082709223,0.2081034184,0.1608862281,0.0529254861,-0.0887272581,-0.1638806909,0.1413200349,-0.1959382594,0.1026792601,-0.2093604356,-0.1335900128,-0.2519333661,0.0598265417,-0.3048490286,0.1674191952,0.0229527056,0.1736568063,-0.0076057324,-0.1181410775,0.068348743,-0.1165577322,0.5064350963,-0.057689555,-0.0134580182,0.3359455168,-0.3830055296,-0.5119382143,-0.0148465978,-0.32138291,0.3655691445,0.4952308536,0.1596135199,0.0717503875,-0.151924789,0.1742211282,0.0769868493,-0.010169833,-0.0870204568,-0.522577703,-0.2604798079,-0.0685331523,0.3230938017,-0.0403109454,0.2531684339,-0.1671895683,-0.0900871456,0.017003594,-0.1824450642,0.409830153,-0.0010019407,0.1590157449,-0.1875927448,0.321798116,-0.2821376324,0.0738387257,-0.2387906313,0.1388521045,-0.2605443597,-0.0038212882,0.1447043419,0.2313895524,0.2033807039,0.3434059322,-0.0389645323,0.357791841,0.1581112891,0.19954063,-0.0919849575,-0.0149513232,0.3387946188,0.2913885117,-0.3603441417,-0.3383019865,0.236809507,0.5287767053,-0.156197533,0.6388680935,-0.47150594,-0.1058280617,0.1740389168,-0.0893825665,0.7218432426,-0.151102528,0.114540413,-0.0477996767,-0.0799836591,-0.164543733,-0.0496445708,0.282753408,-0.0466386974,-0.3187205493,0.1144865528,-0.1477605551,0.162731424,-0.0901566595,-0.060058713,0.3341633379,-0.2901880741,-0.1791115701,0.0768795237,0.5286809802,-0.0136492178,0.1141801029,0.021254709,0.1803944111,0.1529195309,-0.0331838913,-0.1239415631,0.1547225863,0.1730773896,-0.1714735329,-0.3432446718,-0.202794522,-0.2252465934,0.2729432285,-0.2570167482,-0.1231299043,-0.0284451488,0.0414773114,0.0516158566,0.076448001,0.0877549276,0.1950524449,0.1668304354,0.3949015141,0.2292666733,-0.2988648415,0.1881989986,-0.0632355735,-0.3310909569,0.0680575073,-0.0444641188,-0.4540430009,0.0604156703,0.2876528502,0.1042679697,-0.0710848719,-0.1523382962,0.0739472955,-0.3433868885,-0.4228083789,0.1924097687,-0.0239736717,-0.1224285215,0.5455439091,-0.1383499205,-0.5384855866,-0.1701194197,0.3690162897,0.4312333465,0.1662700921,0.138693437,0.0982336327,-0.2912175655,-0.2915677428,-0.1460373104,0.1575229466,0.1159909442,0.5468220115,-0.1520170569,-0.2105822861,0.3614597023,0.1088096946,0.1414925754,0.3442054689,-0.2248914689,-0.155143261,-0.1145680398,0.0348827392,-0.0273167882,-0.0065908344,0.001352824,0.2078740299,0.0267981477,0.1439680606,-0.4122164249,-0.0362603888,-0.1573773026,0.0424514264,-0.0476505645,-0.1395010352,0.1388792992,0.2504679263,0.29838866,0.1474726796,-0.2418826669,-0.3817972839,0.0438447967,0.0596397109,-0.3993674517,-0.1015072539,0.0844372287,-0.0929308236,-0.1048511043,-0.0737564713,-0.0506578349,0.322376579,0.0439011194,-0.1493697315,-0.0528180487,-0.1203635707,0.0826773196,-0.0615567118,-0.3012116849,-0.0447155125,0.0353026129,0.1791663319,0.4173271358,-0.0540245734,-0.0980848968,-0.2278139591,-0.0230983533,0.0880470499,0.0006674979,0.0304236468,-0.2171002924,0.1844252795,0.5604222417,0.3828015029,-0.0121288709,0.0318018198,-0.0611390807,0.2695297003,-0.2831556797,-0.1041295901,0.3444371223,-0.2012411803,0.3371264637,0.0793575346,0.1869932115,0.4522714317,0.3199201822,0.1435378939,0.2103163749,-0.3220168054,0.2615261078,-0.1618447751,0.3352008164,0.0756546631,0.1511347443,-0.0399388634,0.0853096694,0.1706988066,-0.0092242835,0.5128727555,-0.0457618423,0.2001372725,-0.1020325869,0.0814716741,0.1933491379,0.1761447638,-0.4858047366,0.0609779879,0.0479485393,0.1975879073,-0.2576858103,-0.1467925459,-0.1789151579,0.180886209,-0.0589712337,-0.2559525371,0.2616121471,-0.1302040517,0.0335883722,0.0556735769,0.1323437244,0.1473919898,0.2753302455,0.0031762726,0.0970407799,-0.1277589202,-0.2506684363,-0.1332454085,0.2785938978,-0.0871291161,0.0647188723,0.2428395748,0.0697446167,-0.1259705573,0.4104453325,0.4862681031,0.0302184634,-0.2276409268,0.0477973931,-0.0376645438,-0.1175307631,-0.0698460117,0.0138428239,0.1196753383,-0.0593750104,0.1633243859,0.2446502596,-0.2941167653,-0.2053907216,0.1617450565,-0.0946318954,-0.5544945002,0.3059913814,-0.0533478446,0.0474889129,-0.09782058,0.4003549516,-0.4853323102,0.0292671025,0.5993376374,-0.1639459282,0.0554468706,-0.0866309628,0.1203508675,-0.0346126705,0.3137259483,0.104646951,0.327203542,-0.2453809679,-0.0755114779,-0.4701820314,0.0068911598,-0.378143698,-0.002300346,0.0611737631,0.3830406666,-0.2165941149,0.0948095173,0.4464898407,-0.2461355627,0.0373850688,0.0295923986,-0.329104811,0.191416204,0.0178405866,-0.2102586478,0.1854660511,-0.4230689406,0.1024009213,-0.3402990103,0.2648972273,-0.0876418054,-0.2131944895,0.4837400019,0.1372122765,0.1596737206,-0.0276837908,0.3778598309,-0.2144955397,-0.0171399508,0.1701922864,0.2738904059,-0.1418510675,0.0095751565,0.0985787213,0.1840244681,-0.218397662,-0.1317768544,0.0154640386,0.3990646899,0.0715170056,-0.4324676991,-0.1318103969,-0.004571727,0.1930414289,0.1941159666,-0.06747365,0.5390097499,0.0501549914,0.2044937313,-0.3155217171,-0.3606788814,0.3682219088,-0.3801805377,-0.1630280018,-0.5545136333,0.2899544239,0.0679580271,0.2484755963,-0.2688362896,-0.278665036,0.3606102169,0.0865398496,-0.2144652903,0.4349571466,-0.2593582869,0.084221743,0.1108950227,-0.3822179437,-0.1003942117,-0.1719749719,0.1110254154,-0.2570669055]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2070","title":"ArrowInvalid issue for squad v2 dataset","comments":"Hi ! This error happens when you use `map` in batched mode and then your function doesn't return the same number of values per column.\r\n\r\nIndeed since you're using `map` in batched mode, `prepare_validation_features` must take a batch as input (i.e. a dictionary of multiple rows of the dataset), and return a batch.\r\n\r\nHowever it seems like `tokenized_examples` doesn't have the same number of elements in each field. One field seems to have `1180` elements while `candidate_attention_mask` only has `1178`.","body":"Hello, I am using the huggingface official question answering example notebook (https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb). \r\n\r\nIn the prepare_validation_features function, I made some modifications to tokenize a new set of quesions with the original contexts and save them in three different list called candidate_input_dis, candidate_attetion_mask and candidate_token_type_ids. When I try to run the next cell for dataset.map, I got the following error:\r\n\r\n`ArrowInvalid: Column 1 named candidate_attention_mask expected length 1180 but got length 1178`\r\n\r\nMy code is as follows:\r\n\r\n```\r\ndef generate_candidate_questions(examples):\r\n val_questions = examples[\"question\"]\r\n candididate_questions = random.sample(datasets[\"train\"][\"question\"], len(val_questions))\r\n candididate_questions = [x[:max_length] for x in candididate_questions]\r\n return candididate_questions\r\n\r\ndef prepare_validation_features(examples, use_mixing=False):\r\n pad_on_right = tokenizer.padding_side == \"right\"\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n if use_mixing:\r\n candidate_questions = generate_candidate_questions(examples)\r\n tokenized_candidates = tokenizer(\r\n candidate_questions if pad_on_right else examples[\"context\"],\r\n examples[\"context\"] if pad_on_right else candidate_questions,\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n if use_mixing:\r\n tokenized_examples[\"candidate_input_ids\"] = tokenized_candidates[\"input_ids\"]\r\n tokenized_examples[\"candidate_attention_mask\"] = tokenized_candidates[\"attention_mask\"]\r\n tokenized_examples[\"candidate_token_type_ids\"] = tokenized_candidates[\"token_type_ids\"]\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (o if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n\r\n\r\nvalidation_features = datasets[\"validation\"].map(\r\n lambda xs: prepare_validation_features(xs, True),\r\n batched=True,\r\n remove_columns=datasets[\"validation\"].column_names\r\n)\r\n```\r\n\r\nI guess this might happen because of the batched=True. I see similar issues in this repo related to arrow table length mismatch error, but in their cases, the numbers vary a lot. In my case, this error always happens when the expected length and unexpected length are very close. Thanks for the help!","comment_length":80,"text":"ArrowInvalid issue for squad v2 dataset\nHello, I am using the huggingface official question answering example notebook (https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb). \r\n\r\nIn the prepare_validation_features function, I made some modifications to tokenize a new set of quesions with the original contexts and save them in three different list called candidate_input_dis, candidate_attetion_mask and candidate_token_type_ids. When I try to run the next cell for dataset.map, I got the following error:\r\n\r\n`ArrowInvalid: Column 1 named candidate_attention_mask expected length 1180 but got length 1178`\r\n\r\nMy code is as follows:\r\n\r\n```\r\ndef generate_candidate_questions(examples):\r\n val_questions = examples[\"question\"]\r\n candididate_questions = random.sample(datasets[\"train\"][\"question\"], len(val_questions))\r\n candididate_questions = [x[:max_length] for x in candididate_questions]\r\n return candididate_questions\r\n\r\ndef prepare_validation_features(examples, use_mixing=False):\r\n pad_on_right = tokenizer.padding_side == \"right\"\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n if use_mixing:\r\n candidate_questions = generate_candidate_questions(examples)\r\n tokenized_candidates = tokenizer(\r\n candidate_questions if pad_on_right else examples[\"context\"],\r\n examples[\"context\"] if pad_on_right else candidate_questions,\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n if use_mixing:\r\n tokenized_examples[\"candidate_input_ids\"] = tokenized_candidates[\"input_ids\"]\r\n tokenized_examples[\"candidate_attention_mask\"] = tokenized_candidates[\"attention_mask\"]\r\n tokenized_examples[\"candidate_token_type_ids\"] = tokenized_candidates[\"token_type_ids\"]\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (o if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n\r\n\r\nvalidation_features = datasets[\"validation\"].map(\r\n lambda xs: prepare_validation_features(xs, True),\r\n batched=True,\r\n remove_columns=datasets[\"validation\"].column_names\r\n)\r\n```\r\n\r\nI guess this might happen because of the batched=True. I see similar issues in this repo related to arrow table length mismatch error, but in their cases, the numbers vary a lot. In my case, this error always happens when the expected length and unexpected length are very close. Thanks for the help!\nHi ! This error happens when you use `map` in batched mode and then your function doesn't return the same number of values per column.\r\n\r\nIndeed since you're using `map` in batched mode, `prepare_validation_features` must take a batch as input (i.e. a dictionary of multiple rows of the dataset), and return a batch.\r\n\r\nHowever it seems like `tokenized_examples` doesn't have the same number of elements in each field. One field seems to have `1180` elements while `candidate_attention_mask` only has `1178`.","embeddings":[0.195086509,-0.5012049079,0.0041139694,0.2735462487,0.2830952704,-0.0630852804,0.2918868959,0.12445537,-0.3849657178,0.1932278574,0.0684070587,0.2712973952,0.1175046489,-0.2701620162,-0.0708334744,-0.1942008138,0.043876484,-0.0185051747,0.1663554609,-0.0736735165,-0.1644646674,0.2084969878,-0.4730278254,0.3151913881,-0.3681923151,-0.2792748511,-0.0652289987,-0.1413509101,-0.2085033357,-0.7724220157,0.189392671,-0.2526650727,0.2286590785,0.2269305289,-0.0001293044,-0.0708507299,0.0546478927,-0.1082279235,-0.2221769392,0.0503468625,0.2006007582,-0.1157533154,0.0235010628,-0.1506015658,0.0862615928,-0.2532844245,-0.2258586138,-0.2837268412,0.5048593879,0.2183131427,0.0398420133,0.3087803721,0.13261576,-0.0180970859,0.0254218541,0.0207152404,-0.1003998369,-0.0611900054,0.2177817076,-0.016631728,0.110686101,0.6158602238,0.2363804728,0.1203650907,0.3287784159,-0.0844715908,0.0512515418,-0.2424855679,0.1621574163,0.1987718046,0.120321475,-0.080656372,-0.2992402017,-0.3082882464,0.2235186994,-0.0175583772,0.1416762769,-0.1600324064,0.2267239243,0.0799518377,-0.0211595092,0.276373595,-0.2309037298,0.1246564016,0.1555744708,0.0581968427,-0.1273341775,0.3357895613,0.0843684226,-0.4302313924,-0.3450859487,0.1799298525,-0.1760932505,0.384899199,-0.4720772803,-0.1346978843,-0.178399533,-0.2347410917,0.3301493824,-0.1297595501,0.2012562305,-0.0882847235,0.3551928401,0.0218711421,0.0719589442,0.3336797357,0.308562547,0.4262860119,-0.2180073559,-0.3491986692,0.120548971,-0.1206212938,-0.2547729611,-0.3665161133,0.2209726423,0.1167232394,0.384670794,-0.0516374893,-0.3939847946,0.4682804048,-0.7649683952,0.081198208,-0.0058187293,0.178189069,0.2198238224,-0.0031698421,-0.1204943955,0.2113592923,-0.0259241145,-0.2856942415,-0.020511305,-0.07508035,-0.1858078688,0.0906661525,0.057143975,0.1531473994,0.3249381185,-0.1028737649,0.3799371123,-0.3787389398,0.1879871935,-0.1844678521,0.3455336094,-0.0727417469,-0.3098551929,0.1424631774,0.0584285632,-0.1654934138,-0.0539139695,-0.1027214974,-0.1874339432,-0.1261960715,-0.2425878048,-0.0208732765,0.0811207071,0.1629774123,0.104338415,0.1319513619,0.2459521294,-0.0968066007,0.3539926112,-0.3905323744,0.120577842,-0.0145222489,-0.0205828305,0.1708712578,-0.3565444648,-0.1542259753,0.5501691699,0.2950061858,-0.2008676827,0.3871436417,-0.1549215466,0.0752360523,-0.1555736959,0.5609897971,0.2675245106,-0.7531769872,-0.3551869392,-0.0503066555,-0.2858443856,-0.0522272214,0.0802324414,-0.1308106929,0.3515367806,0.0559620075,0.050204888,-0.0221526362,-0.0017574559,0.0674414933,-0.1703104973,-0.0660700426,0.0601951964,-0.056051977,-0.2504400015,-0.2083390653,-0.1153802574,-0.1653351635,0.0465347208,-0.0787110776,0.2557053864,0.3074446619,0.3064408898,0.0371078067,0.2982430458,-0.0424509309,-0.5195571184,-0.0673793852,-0.086148344,-0.0555269644,-0.0934732854,-0.3999192715,-0.483473599,0.1228377223,-0.2551138699,-0.3626616597,-0.0108572375,0.0558655784,0.1518115997,-0.3138305545,-0.064778775,0.0877950713,-0.0671913549,0.2688457966,-0.1233368739,-0.016010385,-0.0595269911,-0.1805099845,0.045282837,0.1954571307,0.3556151688,-0.0670107231,-0.1062885895,0.3517386019,0.2318141907,-0.1074745283,0.1140647829,-0.1387194693,0.1696459502,-0.2511654496,-0.0463707261,0.1581393927,0.0831789002,0.0153888222,-0.1450790912,0.4966517091,0.0857706517,0.3637129366,-0.3584550321,0.0083901044,0.0568754524,0.1508820653,-0.4269770682,-0.0256515313,-0.1253533959,0.1449754983,0.4017035365,-0.1017795429,-0.0515564755,0.0492357202,0.3358851075,0.0354063287,-0.0763477013,-0.024279274,-0.0925268009,0.1731872112,0.263124764,-0.0448205955,0.2035182714,0.1293767393,-0.2184268087,-0.0190938301,-0.0147033511,0.111966379,0.055592116,0.0647543073,0.11833217,0.2495093197,0.3302026987,-0.0305676535,-0.1890271753,0.0396373719,0.1306311786,0.0940906778,-0.3066777289,0.2154724896,-0.1222217977,0.1913228929,-0.0042545311,-0.2016801387,-0.0127335032,-0.1387368143,-0.110694252,0.1354757994,0.050706517,0.3411769569,0.2539761662,0.3478592634,0.4629756212,0.025127247,-0.0785275102,-0.0609102733,-0.1276552975,-0.1625013798,0.0096559869,-0.1742608249,0.0441797562,0.2778018713,-0.0506546758,-0.460667938,-0.6184928417,0.1887655109,-0.1437838376,0.1473682672,0.2346592396,0.1375778764,-0.552395761,-0.1461130083,0.2825928032,0.1565440595,-0.3298376203,0.0078809075,-0.1075720564,-0.2112471759,0.0480222479,-0.2935303152,-0.1629592478,-0.0616582334,0.579547286,-0.3441344798,-0.1095499247,0.2096663713,0.2039573491,0.0286259688,-0.3302955925,-0.1491159499,-0.272418201,0.1424410641,0.3052252829,-0.3198955059,-0.0532762818,-0.0678489506,-0.1944743246,0.2294441611,0.0854576379,-0.0846457332,-0.3280028999,0.0363932662,0.3106918931,-0.2562968433,0.1329609752,0.4908977747,-0.1166251227,0.206622526,-0.0954710543,-0.2646959126,0.3713108003,0.1710574478,0.2116571069,-0.1192346439,0.5886774659,0.0093836384,0.8307531476,0.3969449103,0.0631243512,0.0407918356,-0.1729646176,0.008927281,-0.1118541881,-0.1739912778,0.2700004578,-0.1890034974,-0.0634996369,-0.0095267054,-0.0304046422,-0.0942551047,0.3692889214,0.1485637724,-0.306299299,-0.2004457265,-0.0791700035,0.2220516652,0.2272362411,0.2375491112,0.1232857555,-0.4511265457,0.1228211522,-0.018236313,-0.250937283,0.1650424302,0.0517890938,-0.8704366088,-0.1390441507,-0.1494329125,0.4401815236,0.3663049042,0.4408902228,0.1457426548,-0.3024488389,0.0741486177,0.0059390115,0.5031289458,0.1411448419,-0.2307069302,0.0029416883,-0.119679153,-0.3873910308,-0.2190164179,-0.049690254,0.3620971441,0.4047525525,0.3142863214,-0.3531351388,-0.2044757307,0.402562499,-0.1466727108,0.0333176255,0.0963160992,-0.2732217014,-0.1293743104,-0.4035228491,0.2679013014,0.2889303863,0.1370851994,0.3320335746,-0.0933264419,-0.1025267616,-0.1522099376,-0.0332021303,0.2846736908,0.4153455794,0.1601263285,0.0348914117,0.3511471748,0.0053562126,0.2843350768,0.1784129143,-0.0418754779,-0.681448102,0.2047102898,-0.3417012095,0.4301432669,0.6046118736,-0.0288575701,0.2293278873,-0.174183473,0.224283278,-0.0551013164,0.1478212476,0.5106641054,0.3219779432,0.2062795609,0.0965005308,0.0533223636,-0.1170240045,-0.1083358899,0.6720846295,0.4695625901,-0.1918014437,0.084682323,0.1610690355,0.9873974323,0.4777239263,0.0841025487,0.4931908548,-0.1852522194,-0.121881403,0.0448407531,0.1339326799,-0.457437098,-0.0083539197,0.0889013261,-0.2034325451,0.058374349,0.139141947,-0.0358730294,-0.1335714906,0.0207326617,0.3618429899,-0.3089623451,0.1112883314,-0.247887671,-0.1883855909,-0.2098683715,-0.0012324661,-0.04582287,0.0695340931,-0.1330588609,-0.0758012831,-0.4445632696,-0.4326427281,-0.1195753217,-0.1556158811,-0.6754673123,0.1890507489,0.6171082258,-0.1232606024,0.1216162816,0.2080591172,0.4921890497,-0.0150120156,-0.0753893554,0.2431184351,0.1800323278,-0.0202209409,-0.0825481564,-0.1039878875,-0.0103830136,-0.085309945,-0.3040125668,0.179185912,0.0415869169,-0.2450327873,-0.2042886913,-0.0527845547,0.2268567383,-0.3041245937,-0.0235776491,-0.0600932874,-0.095356442,-0.1726208627,-0.0551084504,-0.1253155917,-0.2628068626,-0.0821693987,0.4602863491,-0.2076923102,0.2340300381,0.2675139606,-0.2994674444,-0.1334337145,0.737672925,-0.0894116312,-0.078272067,-0.1300967932,0.4939996302,0.3498529792,-0.4605635405,-0.0229655001,-0.099859789,-0.2655719221,-0.1948382407,0.5396847129,0.4198506176,0.149111852,0.0435410179,-0.1314018518,-0.4555134773,0.2382653952,-0.2345763594,0.1953025609,0.0412917621,0.0127272913,-0.2794876695,0.21704413,-0.1651521921,-0.0105240652,-0.1417562366,0.1864139885,-0.182552889,-0.3321322501,0.1994382739,0.0276310015,-0.1235888526,0.4358768463,-0.0907111093,-0.0450682864,-0.2240803093,0.1943070292,0.0553535223,0.114220649,-0.2491777986,-0.1408614367,-0.1086735651,-0.133313939,0.0239942204,-0.0468278117,0.3059262633,0.1365163475,-0.0672816187,-0.0881896466,0.1248905659,-0.1457302719,0.3126369119,-0.0867113099,0.2010820657,0.1968142986,-0.1101377606,0.0904987678,-0.154469341,0.3345206082,0.1623488367,0.0165601596,0.0076676034,-0.3720282018,-0.2001096606,0.076880008,-0.2525678277,0.3860231936,-0.2191200554,-0.0479563288,0.0093513299,0.1044491827,-0.1225907728,0.1598887295,0.4138648212,-0.126506269,-0.009075298,0.2409758866,-0.0483123325,0.1263561696,0.2902407348,0.1767107248,0.3259357214,-0.2194678336,-0.0503144599,0.1213119775,-0.1504019946,0.0291122366,0.629226625,0.0923185647,0.3154925406,-0.0725593865,-0.223178491,0.227340281,-0.0814773366,0.3294860721,0.1937238425,0.1792413443,-0.3260829151,0.1299867183,0.0999940559,-0.2000181079,0.1468999386,0.5979399681,0.1495615244,-0.6717576385,-0.1625268459,0.0986655578,-0.1824252605,-0.1483792663,-0.7064804435,0.0255723484,-0.3008074462,-0.1062486172,-0.1684420258,0.1586557478,0.4834394455,0.0503592752,0.0037668913,-0.3265334368,-0.253143549,0.3497816622,0.2839297056,-0.2735243738,0.2680564523,0.0657741129,0.1180291772,0.095935829,0.4089803696,0.2865683734,0.1455129087,-0.1512249559,-0.1514043957,-0.1721804142,0.0007870636,0.2196319848,0.1611724645,0.2842543423,-0.1051215604,0.1125270948,-0.0957735777,-0.0303734113,-0.1010292098,0.1379579306,-0.0324237794,-0.1870809644,-0.349101156,-0.4216123223,-0.0241041686,-0.2664977014,0.0799739733,-0.2798425555,0.2792111933,0.1507414579,0.0445608646,0.0716165155,-0.0171584878,-0.0229842998,-0.1626792103,0.2408898622,0.3803328872,-0.1053189486,-0.2994276583,-0.2107981741,-0.5407531261,0.0478677414,-0.0895122886,-0.0352906696,0.290520519,0.2943843603,0.2752113342,0.1920381337,0.3236572146,-0.2021862566,-0.1859278232,0.2705162764,-0.4119948447,0.0504981205,0.3089594841,0.1470327526,0.2046711594,-0.2144494504,0.088646777,0.3067517281,-0.2205975205,-0.0475093797,-0.1636003554,-0.3189337254,-0.1844428927,0.294700712,0.1161172539,0.1754457206,0.2240548134,-0.2572011054,-0.1365262866,-0.1764378548,-0.1393818855,0.1451854855,0.1106580868,0.6116533875,-0.0473841988,-0.1078157946,-0.0267386045,0.0444118418,0.1038679704,-0.3887690902,-0.29202649,0.3862883151,-0.2014176846,0.264010638,-0.1644918919,0.1330406517,0.1761882156,0.4649398923,-0.247687608,-0.3671609461,0.7873598933,-0.3886917233,-0.3085346818,0.2183826566,0.1803080142,0.6121312976,-0.3203872442,-0.5169121027,-0.0446362346,0.2340022773,-0.0643147305,-0.3073170781,0.1178998649,0.0049137468,0.0817521587,-0.1735568643,0.3225365877,0.1291798651,0.1259115636,-0.117574051,-0.2294536829]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2068","title":"PyTorch not available error on SageMaker GPU docker though it is installed ","comments":"cc @philschmid ","body":"I get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n","comment_length":2,"text":"PyTorch not available error on SageMaker GPU docker though it is installed \nI get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n\ncc @philschmid ","embeddings":[-0.3578641713,0.0624302998,-0.0035595952,0.092395246,0.2011187971,0.0346482657,0.5519385934,0.3234077096,0.3856204748,0.0929586664,0.0008824305,0.396581918,-0.0354305878,0.1952465475,0.0736841932,-0.2641968429,-0.0560292676,0.228468731,-0.3105959892,0.0267087314,0.171954751,0.1197907999,-0.2741039693,0.1857319325,-0.3524909317,-0.093010217,0.250656724,-0.0640398636,-0.178896144,-0.4069884419,0.2793971002,-0.0613616109,0.2397314608,0.4969630539,-0.0001140436,0.303216666,0.1815663874,-0.0520816073,-0.2339606434,-0.4946050942,0.2693472505,-0.2135497779,0.2534041703,-0.3587452173,-0.2566689551,-0.6176720858,-0.1617843211,-0.1064133495,0.1897276342,0.4830078483,0.1778384894,0.2304899693,0.3704248965,-0.1068839654,-0.0213627666,-0.1286899149,-0.1998843551,0.193384558,0.060613174,-0.0477185659,-0.2215194106,0.3306259215,0.0532887354,0.222189188,0.2746862173,-0.0277601741,0.3027733862,-0.3284738958,0.0128042428,0.0186378229,0.3783850074,-0.1673075259,-0.2615454197,-0.1401580125,0.2406661659,-0.185671851,0.2258491665,0.2011611909,-0.2385996133,-0.2307838053,-0.0924862772,-0.2448174804,-0.4942284524,0.0012739522,-0.1486772597,0.6484239101,-0.009730015,0.2958637774,-0.1088481694,0.03162992,0.108798191,-0.0174539723,0.1367830634,-0.1035540998,-0.4412149191,0.0121147214,-0.184800595,-0.3475169241,-0.1545310766,0.1763308048,0.0244534295,0.0385519378,0.4613890648,0.2319640964,-0.0648741797,0.2133830935,-0.1676820517,0.2672363818,0.3156029582,0.1783833504,0.0582823865,-0.0110221663,-0.2719790041,-0.2618050277,0.0919806734,-0.0409463309,0.1839132905,-0.2599915266,-0.1896698624,-0.0647259578,-0.2254171371,-0.0048059095,0.3427653015,0.3795002401,-0.3616908193,0.0215544086,0.3905214369,0.0970223024,-0.1571545601,0.1086176783,-0.0689175576,0.36923787,-0.3053956032,-0.1699036956,0.2570928931,-0.2832776308,0.319039315,0.0521265119,0.4382136464,0.0194357745,0.1652491987,-0.0362840407,-0.0305198543,0.3495331705,0.1684820801,-0.0993223712,0.0673144832,-0.1310516298,0.0435603149,0.1640109867,-0.3027131259,-0.1731041074,-0.2777281404,0.2180583179,0.0239804499,-0.0828475431,-0.0956024304,-0.2363593727,0.1384330541,-0.2462698221,-0.1988418698,-0.5774737597,-0.1518825144,0.1009746268,0.3505586982,-0.1060075387,-0.2552392185,-0.242928192,0.21696347,-0.044582881,0.1921145767,0.167269066,-0.3213567138,0.364137888,-0.2241795808,-0.1162741482,0.4368959367,-0.1582633406,-0.4274958074,-0.0526534356,-0.131755814,-0.1814903766,0.2424345016,-0.0210299063,0.1027198359,0.0694861561,0.1587241441,0.2212405205,0.055063583,-0.0641467497,0.0711620227,-0.0301146377,0.0693062246,0.3470240831,0.4696928263,0.0405193046,-0.2203280181,0.1052827984,0.0011496237,-0.3349923491,-0.0729082674,0.1936165094,0.0726030841,0.1973171979,0.234205395,-0.0241505578,-0.284635216,0.437253207,-0.4078389406,0.2201975584,-0.4012741446,0.2799266577,0.0739980042,0.2309345454,-0.4405454695,-0.0290490314,0.0740211755,0.2747321129,0.1433665305,-0.071650967,-0.0431889445,0.1642151922,-0.2015472054,0.2570536733,-0.0297232922,0.1648442894,-0.3522476256,-0.2846183181,-0.026778385,0.0832818747,0.0867885128,-0.133834824,-0.1851904988,-0.0223054718,0.0563509651,-0.1856317818,-0.4077399969,0.0107585881,0.2134346366,-0.0264883619,0.1546373963,0.2261991501,0.0535074882,0.0659499317,-0.0596124232,0.2794193625,0.0768524185,0.3316195607,0.1563586891,0.1243275106,0.3506565392,0.1796323806,-0.0370965451,0.0312511362,0.3116858006,-0.1815500855,0.3884941638,-0.2803410292,-0.0670936331,-0.2269870043,0.0582206845,0.096817784,0.2094515562,-0.0584184565,-0.2641685605,-0.0197155587,0.1655121893,0.0201513059,0.1763112247,0.0855540112,-0.2054276317,-0.0731580183,-0.2630326152,-0.285345912,0.0199485365,0.2054544538,0.1803489923,0.4772658944,-0.024977481,-0.1353630126,-0.2053964734,-0.3153861165,0.0442342833,0.2464950234,-0.0167008191,0.1939774603,0.0213793144,-0.2587883472,-0.1571341008,0.0688675344,-0.0830654427,-0.0705593601,-0.056711264,0.4370977283,0.1046389937,0.3015836179,0.1242610514,0.0017419218,0.090929538,-0.0915199444,-0.101004757,-0.1997738779,-0.1029308438,0.0908151269,0.0139182741,-0.1522049904,0.4530825913,-0.0226925332,0.2397299111,-0.4752225578,-0.4546351135,0.263915062,0.0946712494,0.0703559294,0.0831279755,-0.1004693732,0.23054263,-0.1151755601,0.2874084115,-0.3667429388,-0.0099853296,0.171131447,-0.3040320277,-0.0911609456,-0.0429216512,-0.2373358309,-0.2131170332,-0.2420698404,-0.1625316441,0.1925634146,0.0692957714,0.0710771754,-0.0482115149,0.2399497479,0.0612403527,0.420825839,-0.0505228527,-0.0142032336,0.3403245509,-0.144912675,-0.4585410953,0.1990785897,-0.0710253119,0.4873732924,0.1876758784,-0.3572619855,-0.1432136744,-0.0876857936,-0.010406387,-0.2309728861,-0.3154932559,0.4347070456,0.0858675763,-0.0017584566,-0.018087849,-0.2098874748,-0.0058370498,-0.1161855832,0.1107516959,0.0180559624,0.5218024254,0.0161350742,0.5054344535,-0.1492939889,-0.1849232465,0.0787607208,-0.0914802626,0.1306820661,-0.077742286,-0.3543592393,0.0909198746,-0.0784771517,0.0423089303,-0.1143648326,-0.2375506312,-0.2545681596,-0.2243156433,0.1367321461,0.2056726664,-0.0749225467,0.1470566839,-0.0218521375,0.234773621,-0.1005828083,0.2911744118,-0.0582701378,-0.0206460617,-0.0068821111,0.40099141,0.2471125424,0.0405494012,-0.3504697084,-0.4177759588,-0.3089730442,0.1535237283,0.1111554801,0.4341416359,-0.2535318136,0.1353392899,0.2518220842,-0.0731290355,0.1305496842,0.201589033,-0.0769401938,0.1304474622,-0.5629460812,-0.2140490711,-0.4025836885,-0.2041153461,0.3084562123,-0.1378363967,0.2591085434,-0.052187264,-0.2361831814,0.377224803,-0.0014561275,-0.3549470007,-0.1654876024,-0.3093603849,-0.0071021901,-0.3891410828,0.1418218762,-0.1630358249,0.1648108363,0.1033505574,0.1802528203,-0.0478269979,-0.0673618242,-0.2418477237,0.1251702905,0.1508538276,-0.1796664447,0.0206297506,0.2583088875,0.4035355747,0.0246620458,0.1251704544,0.1658738106,-0.6877144575,0.2279514372,0.143776834,0.0337538272,-0.0006095635,-0.2813970745,-0.015133095,0.0288642962,0.2616087496,-0.2789119482,0.2849281132,0.4318098724,0.0435546674,-0.3361482918,-0.5102328062,-0.0285126679,0.1190533563,0.130708918,0.1072735712,-0.3197688758,-0.2387962937,0.1009448618,0.2339786887,0.3998080194,-0.1230270714,0.1631871611,0.3690173328,-0.0562773384,0.669048965,-0.0701484829,0.5010374784,-0.2964494824,-0.0236951038,-0.0219667424,-0.3330355585,0.1249234751,-0.0894318596,0.0196071472,0.4615965784,0.0227423031,0.1879124045,0.0824428424,0.5432862043,-0.056193281,-0.3151729703,0.0893430561,0.0824313834,-0.4715276957,0.4554201663,-0.1119189188,0.1296059489,0.2646174729,-0.0853899047,-0.2024868429,0.3169839978,-0.1035186201,0.5135717988,-0.260309875,-0.7178820372,0.2217282057,0.1935005784,0.0475799516,-0.0163907688,-0.0840439871,0.2869564891,-0.5260500908,-0.264888078,-0.1186811849,0.1446561515,0.41260463,-0.1766412705,-0.1750865728,0.0456954911,-0.0207092706,-0.3480331004,-0.1475058943,0.0229152385,0.523981154,0.0506121628,0.1382592916,-0.2436705977,-0.1985277981,-0.1187596694,0.1392894685,0.1864422858,-0.1920204759,0.0092545468,0.112670213,-0.2367699146,0.0267793797,0.4035091996,0.2485736609,-0.0754849836,0.7604902983,0.2701005936,-0.2153382152,-0.1895247102,-0.5757737756,0.1880673319,-0.4746957123,0.0904080123,0.0883788392,-0.0010317379,-0.0252464861,0.0832523406,0.240493089,-0.3185773492,-0.2064391375,-0.4979458451,-0.2446030527,-0.0121778939,-0.0970734507,0.1446900964,0.4060983658,0.0213105734,-0.0374526568,-0.1176230013,-0.2486613244,-0.0812461898,-0.0857515708,-0.0178909376,0.0288438704,-0.2464593798,0.0180503707,0.1339568049,0.116580613,-0.0161814895,-0.0523238592,-0.1377302259,-0.0593801923,0.0638023913,0.1501113325,-0.1841318607,-0.1567177474,-0.3472022116,-0.1686961502,0.0619377308,0.3511345983,0.073198162,-0.0336022787,0.1000552177,-0.1815095842,0.0305698644,-0.2393947244,0.1999550611,-0.1581184864,-0.0344601683,0.2619611919,0.2125540972,0.0491624214,0.1751203686,-0.296749711,0.0122222546,-0.2247535437,0.0795712546,0.3878742754,-0.5469018221,0.3201330304,-0.114045985,0.3263936639,0.2481028885,-0.1112366244,0.0069457488,0.2077629119,0.192999661,-0.3124342859,-0.1750279665,-0.2047646791,0.0255535133,0.1951705664,-0.0336907506,0.0335218683,0.2034249008,0.2418436557,0.2176680118,0.3072521389,-0.2194328308,0.1608728319,0.2555875778,-0.0315138809,0.1866933107,0.4408334792,0.2895790637,0.0344299637,0.2087427378,-0.0501233935,0.268355161,-0.0355201736,0.0493080206,0.2784291506,-0.1406375766,-0.1351004839,0.1598717272,-0.0666537955,-0.0975729823,-0.2356363982,0.1379940212,-0.5082181096,-0.0016119493,-0.3276998401,-0.0616720878,-0.3606294394,0.1335805058,-0.25792858,-0.0233722795,0.0797757208,0.1772313267,0.1117885932,0.1014342308,-0.4469520748,0.1259376109,-0.4083172977,0.0444617681,-0.2394631505,0.3101462126,-0.1211525947,-0.0841582343,0.1898356527,0.2903619409,-0.1219391599,-0.1150254458,0.1751779616,0.4762858152,0.1457411945,0.0434701666,-0.0806150883,-0.0292948931,-0.0321252085,-0.0190826487,0.2248752862,0.2004404217,-0.0397763513,0.1562915593,0.1541848183,-0.1182024628,0.1891729832,0.1260015219,0.0241076648,0.1159706339,0.2661477029,0.0735388547,-0.154797852,-0.1336262524,-0.0261749905,-0.3192779124,-0.1522497237,0.3894211948,0.2127337903,-0.2353196889,-0.1131786257,0.0737563595,0.0157800689,0.4464921057,0.1291050613,0.004601663,-0.1204414517,0.0396351293,-0.2076597214,0.2196232378,-0.023899002,0.0166291501,0.1226764172,0.0947790444,-0.0264727958,-0.0253282469,0.3033050895,-0.1253412366,0.1207330897,-0.1434566081,-0.2756644189,-0.0575891696,0.3088068068,0.0814981163,0.0755050182,-0.2945995927,0.0161185488,0.0222241674,0.1112492085,-0.2723045647,0.0581626073,0.0998518392,-0.3974215984,0.2679631114,-0.0830507725,0.7890290618,0.0927614272,-0.1029083282,-0.0044994657,-0.2056708336,-0.3405399323,0.2708170116,0.0972320586,0.5719087124,0.1613240987,-0.1286329329,-0.4244962335,0.3670101762,-0.1697560996,-0.6661804914,-0.1641177237,0.1922504455,-0.1173405647,0.2502705753,0.2245024294,0.1029542163,0.306530565,-0.2895623446,-0.3585673571,-0.1202796698,0.5804312229,-0.2649388611,-0.2065294981,-0.2863225341,0.1459724754,-0.1629391909,0.1301871687,-0.1651324928,0.1129610091,0.0261454377,-0.0039779241,-0.0153508894,0.2036131471,0.2484441996,0.1664725542,-0.1400396973,0.4426274598,-0.1489300579,-0.2323542684,-0.3859888911,-0.2470760643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2068","title":"PyTorch not available error on SageMaker GPU docker though it is installed ","comments":"Hey @sivakhno,\r\n\r\nhow does your `requirements.txt` look like to install the `datasets` library and which version of it are you running? Can you try to install `datasets>=1.4.0`","body":"I get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n","comment_length":27,"text":"PyTorch not available error on SageMaker GPU docker though it is installed \nI get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n\nHey @sivakhno,\r\n\r\nhow does your `requirements.txt` look like to install the `datasets` library and which version of it are you running? Can you try to install `datasets>=1.4.0`","embeddings":[-0.3578641713,0.0624302998,-0.0035595952,0.092395246,0.2011187971,0.0346482657,0.5519385934,0.3234077096,0.3856204748,0.0929586664,0.0008824305,0.396581918,-0.0354305878,0.1952465475,0.0736841932,-0.2641968429,-0.0560292676,0.228468731,-0.3105959892,0.0267087314,0.171954751,0.1197907999,-0.2741039693,0.1857319325,-0.3524909317,-0.093010217,0.250656724,-0.0640398636,-0.178896144,-0.4069884419,0.2793971002,-0.0613616109,0.2397314608,0.4969630539,-0.0001140436,0.303216666,0.1815663874,-0.0520816073,-0.2339606434,-0.4946050942,0.2693472505,-0.2135497779,0.2534041703,-0.3587452173,-0.2566689551,-0.6176720858,-0.1617843211,-0.1064133495,0.1897276342,0.4830078483,0.1778384894,0.2304899693,0.3704248965,-0.1068839654,-0.0213627666,-0.1286899149,-0.1998843551,0.193384558,0.060613174,-0.0477185659,-0.2215194106,0.3306259215,0.0532887354,0.222189188,0.2746862173,-0.0277601741,0.3027733862,-0.3284738958,0.0128042428,0.0186378229,0.3783850074,-0.1673075259,-0.2615454197,-0.1401580125,0.2406661659,-0.185671851,0.2258491665,0.2011611909,-0.2385996133,-0.2307838053,-0.0924862772,-0.2448174804,-0.4942284524,0.0012739522,-0.1486772597,0.6484239101,-0.009730015,0.2958637774,-0.1088481694,0.03162992,0.108798191,-0.0174539723,0.1367830634,-0.1035540998,-0.4412149191,0.0121147214,-0.184800595,-0.3475169241,-0.1545310766,0.1763308048,0.0244534295,0.0385519378,0.4613890648,0.2319640964,-0.0648741797,0.2133830935,-0.1676820517,0.2672363818,0.3156029582,0.1783833504,0.0582823865,-0.0110221663,-0.2719790041,-0.2618050277,0.0919806734,-0.0409463309,0.1839132905,-0.2599915266,-0.1896698624,-0.0647259578,-0.2254171371,-0.0048059095,0.3427653015,0.3795002401,-0.3616908193,0.0215544086,0.3905214369,0.0970223024,-0.1571545601,0.1086176783,-0.0689175576,0.36923787,-0.3053956032,-0.1699036956,0.2570928931,-0.2832776308,0.319039315,0.0521265119,0.4382136464,0.0194357745,0.1652491987,-0.0362840407,-0.0305198543,0.3495331705,0.1684820801,-0.0993223712,0.0673144832,-0.1310516298,0.0435603149,0.1640109867,-0.3027131259,-0.1731041074,-0.2777281404,0.2180583179,0.0239804499,-0.0828475431,-0.0956024304,-0.2363593727,0.1384330541,-0.2462698221,-0.1988418698,-0.5774737597,-0.1518825144,0.1009746268,0.3505586982,-0.1060075387,-0.2552392185,-0.242928192,0.21696347,-0.044582881,0.1921145767,0.167269066,-0.3213567138,0.364137888,-0.2241795808,-0.1162741482,0.4368959367,-0.1582633406,-0.4274958074,-0.0526534356,-0.131755814,-0.1814903766,0.2424345016,-0.0210299063,0.1027198359,0.0694861561,0.1587241441,0.2212405205,0.055063583,-0.0641467497,0.0711620227,-0.0301146377,0.0693062246,0.3470240831,0.4696928263,0.0405193046,-0.2203280181,0.1052827984,0.0011496237,-0.3349923491,-0.0729082674,0.1936165094,0.0726030841,0.1973171979,0.234205395,-0.0241505578,-0.284635216,0.437253207,-0.4078389406,0.2201975584,-0.4012741446,0.2799266577,0.0739980042,0.2309345454,-0.4405454695,-0.0290490314,0.0740211755,0.2747321129,0.1433665305,-0.071650967,-0.0431889445,0.1642151922,-0.2015472054,0.2570536733,-0.0297232922,0.1648442894,-0.3522476256,-0.2846183181,-0.026778385,0.0832818747,0.0867885128,-0.133834824,-0.1851904988,-0.0223054718,0.0563509651,-0.1856317818,-0.4077399969,0.0107585881,0.2134346366,-0.0264883619,0.1546373963,0.2261991501,0.0535074882,0.0659499317,-0.0596124232,0.2794193625,0.0768524185,0.3316195607,0.1563586891,0.1243275106,0.3506565392,0.1796323806,-0.0370965451,0.0312511362,0.3116858006,-0.1815500855,0.3884941638,-0.2803410292,-0.0670936331,-0.2269870043,0.0582206845,0.096817784,0.2094515562,-0.0584184565,-0.2641685605,-0.0197155587,0.1655121893,0.0201513059,0.1763112247,0.0855540112,-0.2054276317,-0.0731580183,-0.2630326152,-0.285345912,0.0199485365,0.2054544538,0.1803489923,0.4772658944,-0.024977481,-0.1353630126,-0.2053964734,-0.3153861165,0.0442342833,0.2464950234,-0.0167008191,0.1939774603,0.0213793144,-0.2587883472,-0.1571341008,0.0688675344,-0.0830654427,-0.0705593601,-0.056711264,0.4370977283,0.1046389937,0.3015836179,0.1242610514,0.0017419218,0.090929538,-0.0915199444,-0.101004757,-0.1997738779,-0.1029308438,0.0908151269,0.0139182741,-0.1522049904,0.4530825913,-0.0226925332,0.2397299111,-0.4752225578,-0.4546351135,0.263915062,0.0946712494,0.0703559294,0.0831279755,-0.1004693732,0.23054263,-0.1151755601,0.2874084115,-0.3667429388,-0.0099853296,0.171131447,-0.3040320277,-0.0911609456,-0.0429216512,-0.2373358309,-0.2131170332,-0.2420698404,-0.1625316441,0.1925634146,0.0692957714,0.0710771754,-0.0482115149,0.2399497479,0.0612403527,0.420825839,-0.0505228527,-0.0142032336,0.3403245509,-0.144912675,-0.4585410953,0.1990785897,-0.0710253119,0.4873732924,0.1876758784,-0.3572619855,-0.1432136744,-0.0876857936,-0.010406387,-0.2309728861,-0.3154932559,0.4347070456,0.0858675763,-0.0017584566,-0.018087849,-0.2098874748,-0.0058370498,-0.1161855832,0.1107516959,0.0180559624,0.5218024254,0.0161350742,0.5054344535,-0.1492939889,-0.1849232465,0.0787607208,-0.0914802626,0.1306820661,-0.077742286,-0.3543592393,0.0909198746,-0.0784771517,0.0423089303,-0.1143648326,-0.2375506312,-0.2545681596,-0.2243156433,0.1367321461,0.2056726664,-0.0749225467,0.1470566839,-0.0218521375,0.234773621,-0.1005828083,0.2911744118,-0.0582701378,-0.0206460617,-0.0068821111,0.40099141,0.2471125424,0.0405494012,-0.3504697084,-0.4177759588,-0.3089730442,0.1535237283,0.1111554801,0.4341416359,-0.2535318136,0.1353392899,0.2518220842,-0.0731290355,0.1305496842,0.201589033,-0.0769401938,0.1304474622,-0.5629460812,-0.2140490711,-0.4025836885,-0.2041153461,0.3084562123,-0.1378363967,0.2591085434,-0.052187264,-0.2361831814,0.377224803,-0.0014561275,-0.3549470007,-0.1654876024,-0.3093603849,-0.0071021901,-0.3891410828,0.1418218762,-0.1630358249,0.1648108363,0.1033505574,0.1802528203,-0.0478269979,-0.0673618242,-0.2418477237,0.1251702905,0.1508538276,-0.1796664447,0.0206297506,0.2583088875,0.4035355747,0.0246620458,0.1251704544,0.1658738106,-0.6877144575,0.2279514372,0.143776834,0.0337538272,-0.0006095635,-0.2813970745,-0.015133095,0.0288642962,0.2616087496,-0.2789119482,0.2849281132,0.4318098724,0.0435546674,-0.3361482918,-0.5102328062,-0.0285126679,0.1190533563,0.130708918,0.1072735712,-0.3197688758,-0.2387962937,0.1009448618,0.2339786887,0.3998080194,-0.1230270714,0.1631871611,0.3690173328,-0.0562773384,0.669048965,-0.0701484829,0.5010374784,-0.2964494824,-0.0236951038,-0.0219667424,-0.3330355585,0.1249234751,-0.0894318596,0.0196071472,0.4615965784,0.0227423031,0.1879124045,0.0824428424,0.5432862043,-0.056193281,-0.3151729703,0.0893430561,0.0824313834,-0.4715276957,0.4554201663,-0.1119189188,0.1296059489,0.2646174729,-0.0853899047,-0.2024868429,0.3169839978,-0.1035186201,0.5135717988,-0.260309875,-0.7178820372,0.2217282057,0.1935005784,0.0475799516,-0.0163907688,-0.0840439871,0.2869564891,-0.5260500908,-0.264888078,-0.1186811849,0.1446561515,0.41260463,-0.1766412705,-0.1750865728,0.0456954911,-0.0207092706,-0.3480331004,-0.1475058943,0.0229152385,0.523981154,0.0506121628,0.1382592916,-0.2436705977,-0.1985277981,-0.1187596694,0.1392894685,0.1864422858,-0.1920204759,0.0092545468,0.112670213,-0.2367699146,0.0267793797,0.4035091996,0.2485736609,-0.0754849836,0.7604902983,0.2701005936,-0.2153382152,-0.1895247102,-0.5757737756,0.1880673319,-0.4746957123,0.0904080123,0.0883788392,-0.0010317379,-0.0252464861,0.0832523406,0.240493089,-0.3185773492,-0.2064391375,-0.4979458451,-0.2446030527,-0.0121778939,-0.0970734507,0.1446900964,0.4060983658,0.0213105734,-0.0374526568,-0.1176230013,-0.2486613244,-0.0812461898,-0.0857515708,-0.0178909376,0.0288438704,-0.2464593798,0.0180503707,0.1339568049,0.116580613,-0.0161814895,-0.0523238592,-0.1377302259,-0.0593801923,0.0638023913,0.1501113325,-0.1841318607,-0.1567177474,-0.3472022116,-0.1686961502,0.0619377308,0.3511345983,0.073198162,-0.0336022787,0.1000552177,-0.1815095842,0.0305698644,-0.2393947244,0.1999550611,-0.1581184864,-0.0344601683,0.2619611919,0.2125540972,0.0491624214,0.1751203686,-0.296749711,0.0122222546,-0.2247535437,0.0795712546,0.3878742754,-0.5469018221,0.3201330304,-0.114045985,0.3263936639,0.2481028885,-0.1112366244,0.0069457488,0.2077629119,0.192999661,-0.3124342859,-0.1750279665,-0.2047646791,0.0255535133,0.1951705664,-0.0336907506,0.0335218683,0.2034249008,0.2418436557,0.2176680118,0.3072521389,-0.2194328308,0.1608728319,0.2555875778,-0.0315138809,0.1866933107,0.4408334792,0.2895790637,0.0344299637,0.2087427378,-0.0501233935,0.268355161,-0.0355201736,0.0493080206,0.2784291506,-0.1406375766,-0.1351004839,0.1598717272,-0.0666537955,-0.0975729823,-0.2356363982,0.1379940212,-0.5082181096,-0.0016119493,-0.3276998401,-0.0616720878,-0.3606294394,0.1335805058,-0.25792858,-0.0233722795,0.0797757208,0.1772313267,0.1117885932,0.1014342308,-0.4469520748,0.1259376109,-0.4083172977,0.0444617681,-0.2394631505,0.3101462126,-0.1211525947,-0.0841582343,0.1898356527,0.2903619409,-0.1219391599,-0.1150254458,0.1751779616,0.4762858152,0.1457411945,0.0434701666,-0.0806150883,-0.0292948931,-0.0321252085,-0.0190826487,0.2248752862,0.2004404217,-0.0397763513,0.1562915593,0.1541848183,-0.1182024628,0.1891729832,0.1260015219,0.0241076648,0.1159706339,0.2661477029,0.0735388547,-0.154797852,-0.1336262524,-0.0261749905,-0.3192779124,-0.1522497237,0.3894211948,0.2127337903,-0.2353196889,-0.1131786257,0.0737563595,0.0157800689,0.4464921057,0.1291050613,0.004601663,-0.1204414517,0.0396351293,-0.2076597214,0.2196232378,-0.023899002,0.0166291501,0.1226764172,0.0947790444,-0.0264727958,-0.0253282469,0.3033050895,-0.1253412366,0.1207330897,-0.1434566081,-0.2756644189,-0.0575891696,0.3088068068,0.0814981163,0.0755050182,-0.2945995927,0.0161185488,0.0222241674,0.1112492085,-0.2723045647,0.0581626073,0.0998518392,-0.3974215984,0.2679631114,-0.0830507725,0.7890290618,0.0927614272,-0.1029083282,-0.0044994657,-0.2056708336,-0.3405399323,0.2708170116,0.0972320586,0.5719087124,0.1613240987,-0.1286329329,-0.4244962335,0.3670101762,-0.1697560996,-0.6661804914,-0.1641177237,0.1922504455,-0.1173405647,0.2502705753,0.2245024294,0.1029542163,0.306530565,-0.2895623446,-0.3585673571,-0.1202796698,0.5804312229,-0.2649388611,-0.2065294981,-0.2863225341,0.1459724754,-0.1629391909,0.1301871687,-0.1651324928,0.1129610091,0.0261454377,-0.0039779241,-0.0153508894,0.2036131471,0.2484441996,0.1664725542,-0.1400396973,0.4426274598,-0.1489300579,-0.2323542684,-0.3859888911,-0.2470760643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2068","title":"PyTorch not available error on SageMaker GPU docker though it is installed ","comments":"Hi @philschmid - thanks for suggestion. I am using `datasets==1.4.1`. \r\nI have also tried using `torch=1.6.0` (docker `763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.6.0-gpu-py3 `), but the error is the same. ","body":"I get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n","comment_length":25,"text":"PyTorch not available error on SageMaker GPU docker though it is installed \nI get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n\nHi @philschmid - thanks for suggestion. I am using `datasets==1.4.1`. \r\nI have also tried using `torch=1.6.0` (docker `763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.6.0-gpu-py3 `), but the error is the same. ","embeddings":[-0.3578641713,0.0624302998,-0.0035595952,0.092395246,0.2011187971,0.0346482657,0.5519385934,0.3234077096,0.3856204748,0.0929586664,0.0008824305,0.396581918,-0.0354305878,0.1952465475,0.0736841932,-0.2641968429,-0.0560292676,0.228468731,-0.3105959892,0.0267087314,0.171954751,0.1197907999,-0.2741039693,0.1857319325,-0.3524909317,-0.093010217,0.250656724,-0.0640398636,-0.178896144,-0.4069884419,0.2793971002,-0.0613616109,0.2397314608,0.4969630539,-0.0001140436,0.303216666,0.1815663874,-0.0520816073,-0.2339606434,-0.4946050942,0.2693472505,-0.2135497779,0.2534041703,-0.3587452173,-0.2566689551,-0.6176720858,-0.1617843211,-0.1064133495,0.1897276342,0.4830078483,0.1778384894,0.2304899693,0.3704248965,-0.1068839654,-0.0213627666,-0.1286899149,-0.1998843551,0.193384558,0.060613174,-0.0477185659,-0.2215194106,0.3306259215,0.0532887354,0.222189188,0.2746862173,-0.0277601741,0.3027733862,-0.3284738958,0.0128042428,0.0186378229,0.3783850074,-0.1673075259,-0.2615454197,-0.1401580125,0.2406661659,-0.185671851,0.2258491665,0.2011611909,-0.2385996133,-0.2307838053,-0.0924862772,-0.2448174804,-0.4942284524,0.0012739522,-0.1486772597,0.6484239101,-0.009730015,0.2958637774,-0.1088481694,0.03162992,0.108798191,-0.0174539723,0.1367830634,-0.1035540998,-0.4412149191,0.0121147214,-0.184800595,-0.3475169241,-0.1545310766,0.1763308048,0.0244534295,0.0385519378,0.4613890648,0.2319640964,-0.0648741797,0.2133830935,-0.1676820517,0.2672363818,0.3156029582,0.1783833504,0.0582823865,-0.0110221663,-0.2719790041,-0.2618050277,0.0919806734,-0.0409463309,0.1839132905,-0.2599915266,-0.1896698624,-0.0647259578,-0.2254171371,-0.0048059095,0.3427653015,0.3795002401,-0.3616908193,0.0215544086,0.3905214369,0.0970223024,-0.1571545601,0.1086176783,-0.0689175576,0.36923787,-0.3053956032,-0.1699036956,0.2570928931,-0.2832776308,0.319039315,0.0521265119,0.4382136464,0.0194357745,0.1652491987,-0.0362840407,-0.0305198543,0.3495331705,0.1684820801,-0.0993223712,0.0673144832,-0.1310516298,0.0435603149,0.1640109867,-0.3027131259,-0.1731041074,-0.2777281404,0.2180583179,0.0239804499,-0.0828475431,-0.0956024304,-0.2363593727,0.1384330541,-0.2462698221,-0.1988418698,-0.5774737597,-0.1518825144,0.1009746268,0.3505586982,-0.1060075387,-0.2552392185,-0.242928192,0.21696347,-0.044582881,0.1921145767,0.167269066,-0.3213567138,0.364137888,-0.2241795808,-0.1162741482,0.4368959367,-0.1582633406,-0.4274958074,-0.0526534356,-0.131755814,-0.1814903766,0.2424345016,-0.0210299063,0.1027198359,0.0694861561,0.1587241441,0.2212405205,0.055063583,-0.0641467497,0.0711620227,-0.0301146377,0.0693062246,0.3470240831,0.4696928263,0.0405193046,-0.2203280181,0.1052827984,0.0011496237,-0.3349923491,-0.0729082674,0.1936165094,0.0726030841,0.1973171979,0.234205395,-0.0241505578,-0.284635216,0.437253207,-0.4078389406,0.2201975584,-0.4012741446,0.2799266577,0.0739980042,0.2309345454,-0.4405454695,-0.0290490314,0.0740211755,0.2747321129,0.1433665305,-0.071650967,-0.0431889445,0.1642151922,-0.2015472054,0.2570536733,-0.0297232922,0.1648442894,-0.3522476256,-0.2846183181,-0.026778385,0.0832818747,0.0867885128,-0.133834824,-0.1851904988,-0.0223054718,0.0563509651,-0.1856317818,-0.4077399969,0.0107585881,0.2134346366,-0.0264883619,0.1546373963,0.2261991501,0.0535074882,0.0659499317,-0.0596124232,0.2794193625,0.0768524185,0.3316195607,0.1563586891,0.1243275106,0.3506565392,0.1796323806,-0.0370965451,0.0312511362,0.3116858006,-0.1815500855,0.3884941638,-0.2803410292,-0.0670936331,-0.2269870043,0.0582206845,0.096817784,0.2094515562,-0.0584184565,-0.2641685605,-0.0197155587,0.1655121893,0.0201513059,0.1763112247,0.0855540112,-0.2054276317,-0.0731580183,-0.2630326152,-0.285345912,0.0199485365,0.2054544538,0.1803489923,0.4772658944,-0.024977481,-0.1353630126,-0.2053964734,-0.3153861165,0.0442342833,0.2464950234,-0.0167008191,0.1939774603,0.0213793144,-0.2587883472,-0.1571341008,0.0688675344,-0.0830654427,-0.0705593601,-0.056711264,0.4370977283,0.1046389937,0.3015836179,0.1242610514,0.0017419218,0.090929538,-0.0915199444,-0.101004757,-0.1997738779,-0.1029308438,0.0908151269,0.0139182741,-0.1522049904,0.4530825913,-0.0226925332,0.2397299111,-0.4752225578,-0.4546351135,0.263915062,0.0946712494,0.0703559294,0.0831279755,-0.1004693732,0.23054263,-0.1151755601,0.2874084115,-0.3667429388,-0.0099853296,0.171131447,-0.3040320277,-0.0911609456,-0.0429216512,-0.2373358309,-0.2131170332,-0.2420698404,-0.1625316441,0.1925634146,0.0692957714,0.0710771754,-0.0482115149,0.2399497479,0.0612403527,0.420825839,-0.0505228527,-0.0142032336,0.3403245509,-0.144912675,-0.4585410953,0.1990785897,-0.0710253119,0.4873732924,0.1876758784,-0.3572619855,-0.1432136744,-0.0876857936,-0.010406387,-0.2309728861,-0.3154932559,0.4347070456,0.0858675763,-0.0017584566,-0.018087849,-0.2098874748,-0.0058370498,-0.1161855832,0.1107516959,0.0180559624,0.5218024254,0.0161350742,0.5054344535,-0.1492939889,-0.1849232465,0.0787607208,-0.0914802626,0.1306820661,-0.077742286,-0.3543592393,0.0909198746,-0.0784771517,0.0423089303,-0.1143648326,-0.2375506312,-0.2545681596,-0.2243156433,0.1367321461,0.2056726664,-0.0749225467,0.1470566839,-0.0218521375,0.234773621,-0.1005828083,0.2911744118,-0.0582701378,-0.0206460617,-0.0068821111,0.40099141,0.2471125424,0.0405494012,-0.3504697084,-0.4177759588,-0.3089730442,0.1535237283,0.1111554801,0.4341416359,-0.2535318136,0.1353392899,0.2518220842,-0.0731290355,0.1305496842,0.201589033,-0.0769401938,0.1304474622,-0.5629460812,-0.2140490711,-0.4025836885,-0.2041153461,0.3084562123,-0.1378363967,0.2591085434,-0.052187264,-0.2361831814,0.377224803,-0.0014561275,-0.3549470007,-0.1654876024,-0.3093603849,-0.0071021901,-0.3891410828,0.1418218762,-0.1630358249,0.1648108363,0.1033505574,0.1802528203,-0.0478269979,-0.0673618242,-0.2418477237,0.1251702905,0.1508538276,-0.1796664447,0.0206297506,0.2583088875,0.4035355747,0.0246620458,0.1251704544,0.1658738106,-0.6877144575,0.2279514372,0.143776834,0.0337538272,-0.0006095635,-0.2813970745,-0.015133095,0.0288642962,0.2616087496,-0.2789119482,0.2849281132,0.4318098724,0.0435546674,-0.3361482918,-0.5102328062,-0.0285126679,0.1190533563,0.130708918,0.1072735712,-0.3197688758,-0.2387962937,0.1009448618,0.2339786887,0.3998080194,-0.1230270714,0.1631871611,0.3690173328,-0.0562773384,0.669048965,-0.0701484829,0.5010374784,-0.2964494824,-0.0236951038,-0.0219667424,-0.3330355585,0.1249234751,-0.0894318596,0.0196071472,0.4615965784,0.0227423031,0.1879124045,0.0824428424,0.5432862043,-0.056193281,-0.3151729703,0.0893430561,0.0824313834,-0.4715276957,0.4554201663,-0.1119189188,0.1296059489,0.2646174729,-0.0853899047,-0.2024868429,0.3169839978,-0.1035186201,0.5135717988,-0.260309875,-0.7178820372,0.2217282057,0.1935005784,0.0475799516,-0.0163907688,-0.0840439871,0.2869564891,-0.5260500908,-0.264888078,-0.1186811849,0.1446561515,0.41260463,-0.1766412705,-0.1750865728,0.0456954911,-0.0207092706,-0.3480331004,-0.1475058943,0.0229152385,0.523981154,0.0506121628,0.1382592916,-0.2436705977,-0.1985277981,-0.1187596694,0.1392894685,0.1864422858,-0.1920204759,0.0092545468,0.112670213,-0.2367699146,0.0267793797,0.4035091996,0.2485736609,-0.0754849836,0.7604902983,0.2701005936,-0.2153382152,-0.1895247102,-0.5757737756,0.1880673319,-0.4746957123,0.0904080123,0.0883788392,-0.0010317379,-0.0252464861,0.0832523406,0.240493089,-0.3185773492,-0.2064391375,-0.4979458451,-0.2446030527,-0.0121778939,-0.0970734507,0.1446900964,0.4060983658,0.0213105734,-0.0374526568,-0.1176230013,-0.2486613244,-0.0812461898,-0.0857515708,-0.0178909376,0.0288438704,-0.2464593798,0.0180503707,0.1339568049,0.116580613,-0.0161814895,-0.0523238592,-0.1377302259,-0.0593801923,0.0638023913,0.1501113325,-0.1841318607,-0.1567177474,-0.3472022116,-0.1686961502,0.0619377308,0.3511345983,0.073198162,-0.0336022787,0.1000552177,-0.1815095842,0.0305698644,-0.2393947244,0.1999550611,-0.1581184864,-0.0344601683,0.2619611919,0.2125540972,0.0491624214,0.1751203686,-0.296749711,0.0122222546,-0.2247535437,0.0795712546,0.3878742754,-0.5469018221,0.3201330304,-0.114045985,0.3263936639,0.2481028885,-0.1112366244,0.0069457488,0.2077629119,0.192999661,-0.3124342859,-0.1750279665,-0.2047646791,0.0255535133,0.1951705664,-0.0336907506,0.0335218683,0.2034249008,0.2418436557,0.2176680118,0.3072521389,-0.2194328308,0.1608728319,0.2555875778,-0.0315138809,0.1866933107,0.4408334792,0.2895790637,0.0344299637,0.2087427378,-0.0501233935,0.268355161,-0.0355201736,0.0493080206,0.2784291506,-0.1406375766,-0.1351004839,0.1598717272,-0.0666537955,-0.0975729823,-0.2356363982,0.1379940212,-0.5082181096,-0.0016119493,-0.3276998401,-0.0616720878,-0.3606294394,0.1335805058,-0.25792858,-0.0233722795,0.0797757208,0.1772313267,0.1117885932,0.1014342308,-0.4469520748,0.1259376109,-0.4083172977,0.0444617681,-0.2394631505,0.3101462126,-0.1211525947,-0.0841582343,0.1898356527,0.2903619409,-0.1219391599,-0.1150254458,0.1751779616,0.4762858152,0.1457411945,0.0434701666,-0.0806150883,-0.0292948931,-0.0321252085,-0.0190826487,0.2248752862,0.2004404217,-0.0397763513,0.1562915593,0.1541848183,-0.1182024628,0.1891729832,0.1260015219,0.0241076648,0.1159706339,0.2661477029,0.0735388547,-0.154797852,-0.1336262524,-0.0261749905,-0.3192779124,-0.1522497237,0.3894211948,0.2127337903,-0.2353196889,-0.1131786257,0.0737563595,0.0157800689,0.4464921057,0.1291050613,0.004601663,-0.1204414517,0.0396351293,-0.2076597214,0.2196232378,-0.023899002,0.0166291501,0.1226764172,0.0947790444,-0.0264727958,-0.0253282469,0.3033050895,-0.1253412366,0.1207330897,-0.1434566081,-0.2756644189,-0.0575891696,0.3088068068,0.0814981163,0.0755050182,-0.2945995927,0.0161185488,0.0222241674,0.1112492085,-0.2723045647,0.0581626073,0.0998518392,-0.3974215984,0.2679631114,-0.0830507725,0.7890290618,0.0927614272,-0.1029083282,-0.0044994657,-0.2056708336,-0.3405399323,0.2708170116,0.0972320586,0.5719087124,0.1613240987,-0.1286329329,-0.4244962335,0.3670101762,-0.1697560996,-0.6661804914,-0.1641177237,0.1922504455,-0.1173405647,0.2502705753,0.2245024294,0.1029542163,0.306530565,-0.2895623446,-0.3585673571,-0.1202796698,0.5804312229,-0.2649388611,-0.2065294981,-0.2863225341,0.1459724754,-0.1629391909,0.1301871687,-0.1651324928,0.1129610091,0.0261454377,-0.0039779241,-0.0153508894,0.2036131471,0.2484441996,0.1664725542,-0.1400396973,0.4426274598,-0.1489300579,-0.2323542684,-0.3859888911,-0.2470760643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2068","title":"PyTorch not available error on SageMaker GPU docker though it is installed ","comments":"Could paste the code you use the start your training job and the fine-tuning script you run? ","body":"I get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n","comment_length":17,"text":"PyTorch not available error on SageMaker GPU docker though it is installed \nI get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n\nCould paste the code you use the start your training job and the fine-tuning script you run? ","embeddings":[-0.3578641713,0.0624302998,-0.0035595952,0.092395246,0.2011187971,0.0346482657,0.5519385934,0.3234077096,0.3856204748,0.0929586664,0.0008824305,0.396581918,-0.0354305878,0.1952465475,0.0736841932,-0.2641968429,-0.0560292676,0.228468731,-0.3105959892,0.0267087314,0.171954751,0.1197907999,-0.2741039693,0.1857319325,-0.3524909317,-0.093010217,0.250656724,-0.0640398636,-0.178896144,-0.4069884419,0.2793971002,-0.0613616109,0.2397314608,0.4969630539,-0.0001140436,0.303216666,0.1815663874,-0.0520816073,-0.2339606434,-0.4946050942,0.2693472505,-0.2135497779,0.2534041703,-0.3587452173,-0.2566689551,-0.6176720858,-0.1617843211,-0.1064133495,0.1897276342,0.4830078483,0.1778384894,0.2304899693,0.3704248965,-0.1068839654,-0.0213627666,-0.1286899149,-0.1998843551,0.193384558,0.060613174,-0.0477185659,-0.2215194106,0.3306259215,0.0532887354,0.222189188,0.2746862173,-0.0277601741,0.3027733862,-0.3284738958,0.0128042428,0.0186378229,0.3783850074,-0.1673075259,-0.2615454197,-0.1401580125,0.2406661659,-0.185671851,0.2258491665,0.2011611909,-0.2385996133,-0.2307838053,-0.0924862772,-0.2448174804,-0.4942284524,0.0012739522,-0.1486772597,0.6484239101,-0.009730015,0.2958637774,-0.1088481694,0.03162992,0.108798191,-0.0174539723,0.1367830634,-0.1035540998,-0.4412149191,0.0121147214,-0.184800595,-0.3475169241,-0.1545310766,0.1763308048,0.0244534295,0.0385519378,0.4613890648,0.2319640964,-0.0648741797,0.2133830935,-0.1676820517,0.2672363818,0.3156029582,0.1783833504,0.0582823865,-0.0110221663,-0.2719790041,-0.2618050277,0.0919806734,-0.0409463309,0.1839132905,-0.2599915266,-0.1896698624,-0.0647259578,-0.2254171371,-0.0048059095,0.3427653015,0.3795002401,-0.3616908193,0.0215544086,0.3905214369,0.0970223024,-0.1571545601,0.1086176783,-0.0689175576,0.36923787,-0.3053956032,-0.1699036956,0.2570928931,-0.2832776308,0.319039315,0.0521265119,0.4382136464,0.0194357745,0.1652491987,-0.0362840407,-0.0305198543,0.3495331705,0.1684820801,-0.0993223712,0.0673144832,-0.1310516298,0.0435603149,0.1640109867,-0.3027131259,-0.1731041074,-0.2777281404,0.2180583179,0.0239804499,-0.0828475431,-0.0956024304,-0.2363593727,0.1384330541,-0.2462698221,-0.1988418698,-0.5774737597,-0.1518825144,0.1009746268,0.3505586982,-0.1060075387,-0.2552392185,-0.242928192,0.21696347,-0.044582881,0.1921145767,0.167269066,-0.3213567138,0.364137888,-0.2241795808,-0.1162741482,0.4368959367,-0.1582633406,-0.4274958074,-0.0526534356,-0.131755814,-0.1814903766,0.2424345016,-0.0210299063,0.1027198359,0.0694861561,0.1587241441,0.2212405205,0.055063583,-0.0641467497,0.0711620227,-0.0301146377,0.0693062246,0.3470240831,0.4696928263,0.0405193046,-0.2203280181,0.1052827984,0.0011496237,-0.3349923491,-0.0729082674,0.1936165094,0.0726030841,0.1973171979,0.234205395,-0.0241505578,-0.284635216,0.437253207,-0.4078389406,0.2201975584,-0.4012741446,0.2799266577,0.0739980042,0.2309345454,-0.4405454695,-0.0290490314,0.0740211755,0.2747321129,0.1433665305,-0.071650967,-0.0431889445,0.1642151922,-0.2015472054,0.2570536733,-0.0297232922,0.1648442894,-0.3522476256,-0.2846183181,-0.026778385,0.0832818747,0.0867885128,-0.133834824,-0.1851904988,-0.0223054718,0.0563509651,-0.1856317818,-0.4077399969,0.0107585881,0.2134346366,-0.0264883619,0.1546373963,0.2261991501,0.0535074882,0.0659499317,-0.0596124232,0.2794193625,0.0768524185,0.3316195607,0.1563586891,0.1243275106,0.3506565392,0.1796323806,-0.0370965451,0.0312511362,0.3116858006,-0.1815500855,0.3884941638,-0.2803410292,-0.0670936331,-0.2269870043,0.0582206845,0.096817784,0.2094515562,-0.0584184565,-0.2641685605,-0.0197155587,0.1655121893,0.0201513059,0.1763112247,0.0855540112,-0.2054276317,-0.0731580183,-0.2630326152,-0.285345912,0.0199485365,0.2054544538,0.1803489923,0.4772658944,-0.024977481,-0.1353630126,-0.2053964734,-0.3153861165,0.0442342833,0.2464950234,-0.0167008191,0.1939774603,0.0213793144,-0.2587883472,-0.1571341008,0.0688675344,-0.0830654427,-0.0705593601,-0.056711264,0.4370977283,0.1046389937,0.3015836179,0.1242610514,0.0017419218,0.090929538,-0.0915199444,-0.101004757,-0.1997738779,-0.1029308438,0.0908151269,0.0139182741,-0.1522049904,0.4530825913,-0.0226925332,0.2397299111,-0.4752225578,-0.4546351135,0.263915062,0.0946712494,0.0703559294,0.0831279755,-0.1004693732,0.23054263,-0.1151755601,0.2874084115,-0.3667429388,-0.0099853296,0.171131447,-0.3040320277,-0.0911609456,-0.0429216512,-0.2373358309,-0.2131170332,-0.2420698404,-0.1625316441,0.1925634146,0.0692957714,0.0710771754,-0.0482115149,0.2399497479,0.0612403527,0.420825839,-0.0505228527,-0.0142032336,0.3403245509,-0.144912675,-0.4585410953,0.1990785897,-0.0710253119,0.4873732924,0.1876758784,-0.3572619855,-0.1432136744,-0.0876857936,-0.010406387,-0.2309728861,-0.3154932559,0.4347070456,0.0858675763,-0.0017584566,-0.018087849,-0.2098874748,-0.0058370498,-0.1161855832,0.1107516959,0.0180559624,0.5218024254,0.0161350742,0.5054344535,-0.1492939889,-0.1849232465,0.0787607208,-0.0914802626,0.1306820661,-0.077742286,-0.3543592393,0.0909198746,-0.0784771517,0.0423089303,-0.1143648326,-0.2375506312,-0.2545681596,-0.2243156433,0.1367321461,0.2056726664,-0.0749225467,0.1470566839,-0.0218521375,0.234773621,-0.1005828083,0.2911744118,-0.0582701378,-0.0206460617,-0.0068821111,0.40099141,0.2471125424,0.0405494012,-0.3504697084,-0.4177759588,-0.3089730442,0.1535237283,0.1111554801,0.4341416359,-0.2535318136,0.1353392899,0.2518220842,-0.0731290355,0.1305496842,0.201589033,-0.0769401938,0.1304474622,-0.5629460812,-0.2140490711,-0.4025836885,-0.2041153461,0.3084562123,-0.1378363967,0.2591085434,-0.052187264,-0.2361831814,0.377224803,-0.0014561275,-0.3549470007,-0.1654876024,-0.3093603849,-0.0071021901,-0.3891410828,0.1418218762,-0.1630358249,0.1648108363,0.1033505574,0.1802528203,-0.0478269979,-0.0673618242,-0.2418477237,0.1251702905,0.1508538276,-0.1796664447,0.0206297506,0.2583088875,0.4035355747,0.0246620458,0.1251704544,0.1658738106,-0.6877144575,0.2279514372,0.143776834,0.0337538272,-0.0006095635,-0.2813970745,-0.015133095,0.0288642962,0.2616087496,-0.2789119482,0.2849281132,0.4318098724,0.0435546674,-0.3361482918,-0.5102328062,-0.0285126679,0.1190533563,0.130708918,0.1072735712,-0.3197688758,-0.2387962937,0.1009448618,0.2339786887,0.3998080194,-0.1230270714,0.1631871611,0.3690173328,-0.0562773384,0.669048965,-0.0701484829,0.5010374784,-0.2964494824,-0.0236951038,-0.0219667424,-0.3330355585,0.1249234751,-0.0894318596,0.0196071472,0.4615965784,0.0227423031,0.1879124045,0.0824428424,0.5432862043,-0.056193281,-0.3151729703,0.0893430561,0.0824313834,-0.4715276957,0.4554201663,-0.1119189188,0.1296059489,0.2646174729,-0.0853899047,-0.2024868429,0.3169839978,-0.1035186201,0.5135717988,-0.260309875,-0.7178820372,0.2217282057,0.1935005784,0.0475799516,-0.0163907688,-0.0840439871,0.2869564891,-0.5260500908,-0.264888078,-0.1186811849,0.1446561515,0.41260463,-0.1766412705,-0.1750865728,0.0456954911,-0.0207092706,-0.3480331004,-0.1475058943,0.0229152385,0.523981154,0.0506121628,0.1382592916,-0.2436705977,-0.1985277981,-0.1187596694,0.1392894685,0.1864422858,-0.1920204759,0.0092545468,0.112670213,-0.2367699146,0.0267793797,0.4035091996,0.2485736609,-0.0754849836,0.7604902983,0.2701005936,-0.2153382152,-0.1895247102,-0.5757737756,0.1880673319,-0.4746957123,0.0904080123,0.0883788392,-0.0010317379,-0.0252464861,0.0832523406,0.240493089,-0.3185773492,-0.2064391375,-0.4979458451,-0.2446030527,-0.0121778939,-0.0970734507,0.1446900964,0.4060983658,0.0213105734,-0.0374526568,-0.1176230013,-0.2486613244,-0.0812461898,-0.0857515708,-0.0178909376,0.0288438704,-0.2464593798,0.0180503707,0.1339568049,0.116580613,-0.0161814895,-0.0523238592,-0.1377302259,-0.0593801923,0.0638023913,0.1501113325,-0.1841318607,-0.1567177474,-0.3472022116,-0.1686961502,0.0619377308,0.3511345983,0.073198162,-0.0336022787,0.1000552177,-0.1815095842,0.0305698644,-0.2393947244,0.1999550611,-0.1581184864,-0.0344601683,0.2619611919,0.2125540972,0.0491624214,0.1751203686,-0.296749711,0.0122222546,-0.2247535437,0.0795712546,0.3878742754,-0.5469018221,0.3201330304,-0.114045985,0.3263936639,0.2481028885,-0.1112366244,0.0069457488,0.2077629119,0.192999661,-0.3124342859,-0.1750279665,-0.2047646791,0.0255535133,0.1951705664,-0.0336907506,0.0335218683,0.2034249008,0.2418436557,0.2176680118,0.3072521389,-0.2194328308,0.1608728319,0.2555875778,-0.0315138809,0.1866933107,0.4408334792,0.2895790637,0.0344299637,0.2087427378,-0.0501233935,0.268355161,-0.0355201736,0.0493080206,0.2784291506,-0.1406375766,-0.1351004839,0.1598717272,-0.0666537955,-0.0975729823,-0.2356363982,0.1379940212,-0.5082181096,-0.0016119493,-0.3276998401,-0.0616720878,-0.3606294394,0.1335805058,-0.25792858,-0.0233722795,0.0797757208,0.1772313267,0.1117885932,0.1014342308,-0.4469520748,0.1259376109,-0.4083172977,0.0444617681,-0.2394631505,0.3101462126,-0.1211525947,-0.0841582343,0.1898356527,0.2903619409,-0.1219391599,-0.1150254458,0.1751779616,0.4762858152,0.1457411945,0.0434701666,-0.0806150883,-0.0292948931,-0.0321252085,-0.0190826487,0.2248752862,0.2004404217,-0.0397763513,0.1562915593,0.1541848183,-0.1182024628,0.1891729832,0.1260015219,0.0241076648,0.1159706339,0.2661477029,0.0735388547,-0.154797852,-0.1336262524,-0.0261749905,-0.3192779124,-0.1522497237,0.3894211948,0.2127337903,-0.2353196889,-0.1131786257,0.0737563595,0.0157800689,0.4464921057,0.1291050613,0.004601663,-0.1204414517,0.0396351293,-0.2076597214,0.2196232378,-0.023899002,0.0166291501,0.1226764172,0.0947790444,-0.0264727958,-0.0253282469,0.3033050895,-0.1253412366,0.1207330897,-0.1434566081,-0.2756644189,-0.0575891696,0.3088068068,0.0814981163,0.0755050182,-0.2945995927,0.0161185488,0.0222241674,0.1112492085,-0.2723045647,0.0581626073,0.0998518392,-0.3974215984,0.2679631114,-0.0830507725,0.7890290618,0.0927614272,-0.1029083282,-0.0044994657,-0.2056708336,-0.3405399323,0.2708170116,0.0972320586,0.5719087124,0.1613240987,-0.1286329329,-0.4244962335,0.3670101762,-0.1697560996,-0.6661804914,-0.1641177237,0.1922504455,-0.1173405647,0.2502705753,0.2245024294,0.1029542163,0.306530565,-0.2895623446,-0.3585673571,-0.1202796698,0.5804312229,-0.2649388611,-0.2065294981,-0.2863225341,0.1459724754,-0.1629391909,0.1301871687,-0.1651324928,0.1129610091,0.0261454377,-0.0039779241,-0.0153508894,0.2036131471,0.2484441996,0.1664725542,-0.1400396973,0.4426274598,-0.1489300579,-0.2323542684,-0.3859888911,-0.2470760643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2068","title":"PyTorch not available error on SageMaker GPU docker though it is installed ","comments":"@sivakhno this should be now fixed in `datasets>=1.5.0`. ","body":"I get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n","comment_length":8,"text":"PyTorch not available error on SageMaker GPU docker though it is installed \nI get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n\n@sivakhno this should be now fixed in `datasets>=1.5.0`. ","embeddings":[-0.3578641713,0.0624302998,-0.0035595952,0.092395246,0.2011187971,0.0346482657,0.5519385934,0.3234077096,0.3856204748,0.0929586664,0.0008824305,0.396581918,-0.0354305878,0.1952465475,0.0736841932,-0.2641968429,-0.0560292676,0.228468731,-0.3105959892,0.0267087314,0.171954751,0.1197907999,-0.2741039693,0.1857319325,-0.3524909317,-0.093010217,0.250656724,-0.0640398636,-0.178896144,-0.4069884419,0.2793971002,-0.0613616109,0.2397314608,0.4969630539,-0.0001140436,0.303216666,0.1815663874,-0.0520816073,-0.2339606434,-0.4946050942,0.2693472505,-0.2135497779,0.2534041703,-0.3587452173,-0.2566689551,-0.6176720858,-0.1617843211,-0.1064133495,0.1897276342,0.4830078483,0.1778384894,0.2304899693,0.3704248965,-0.1068839654,-0.0213627666,-0.1286899149,-0.1998843551,0.193384558,0.060613174,-0.0477185659,-0.2215194106,0.3306259215,0.0532887354,0.222189188,0.2746862173,-0.0277601741,0.3027733862,-0.3284738958,0.0128042428,0.0186378229,0.3783850074,-0.1673075259,-0.2615454197,-0.1401580125,0.2406661659,-0.185671851,0.2258491665,0.2011611909,-0.2385996133,-0.2307838053,-0.0924862772,-0.2448174804,-0.4942284524,0.0012739522,-0.1486772597,0.6484239101,-0.009730015,0.2958637774,-0.1088481694,0.03162992,0.108798191,-0.0174539723,0.1367830634,-0.1035540998,-0.4412149191,0.0121147214,-0.184800595,-0.3475169241,-0.1545310766,0.1763308048,0.0244534295,0.0385519378,0.4613890648,0.2319640964,-0.0648741797,0.2133830935,-0.1676820517,0.2672363818,0.3156029582,0.1783833504,0.0582823865,-0.0110221663,-0.2719790041,-0.2618050277,0.0919806734,-0.0409463309,0.1839132905,-0.2599915266,-0.1896698624,-0.0647259578,-0.2254171371,-0.0048059095,0.3427653015,0.3795002401,-0.3616908193,0.0215544086,0.3905214369,0.0970223024,-0.1571545601,0.1086176783,-0.0689175576,0.36923787,-0.3053956032,-0.1699036956,0.2570928931,-0.2832776308,0.319039315,0.0521265119,0.4382136464,0.0194357745,0.1652491987,-0.0362840407,-0.0305198543,0.3495331705,0.1684820801,-0.0993223712,0.0673144832,-0.1310516298,0.0435603149,0.1640109867,-0.3027131259,-0.1731041074,-0.2777281404,0.2180583179,0.0239804499,-0.0828475431,-0.0956024304,-0.2363593727,0.1384330541,-0.2462698221,-0.1988418698,-0.5774737597,-0.1518825144,0.1009746268,0.3505586982,-0.1060075387,-0.2552392185,-0.242928192,0.21696347,-0.044582881,0.1921145767,0.167269066,-0.3213567138,0.364137888,-0.2241795808,-0.1162741482,0.4368959367,-0.1582633406,-0.4274958074,-0.0526534356,-0.131755814,-0.1814903766,0.2424345016,-0.0210299063,0.1027198359,0.0694861561,0.1587241441,0.2212405205,0.055063583,-0.0641467497,0.0711620227,-0.0301146377,0.0693062246,0.3470240831,0.4696928263,0.0405193046,-0.2203280181,0.1052827984,0.0011496237,-0.3349923491,-0.0729082674,0.1936165094,0.0726030841,0.1973171979,0.234205395,-0.0241505578,-0.284635216,0.437253207,-0.4078389406,0.2201975584,-0.4012741446,0.2799266577,0.0739980042,0.2309345454,-0.4405454695,-0.0290490314,0.0740211755,0.2747321129,0.1433665305,-0.071650967,-0.0431889445,0.1642151922,-0.2015472054,0.2570536733,-0.0297232922,0.1648442894,-0.3522476256,-0.2846183181,-0.026778385,0.0832818747,0.0867885128,-0.133834824,-0.1851904988,-0.0223054718,0.0563509651,-0.1856317818,-0.4077399969,0.0107585881,0.2134346366,-0.0264883619,0.1546373963,0.2261991501,0.0535074882,0.0659499317,-0.0596124232,0.2794193625,0.0768524185,0.3316195607,0.1563586891,0.1243275106,0.3506565392,0.1796323806,-0.0370965451,0.0312511362,0.3116858006,-0.1815500855,0.3884941638,-0.2803410292,-0.0670936331,-0.2269870043,0.0582206845,0.096817784,0.2094515562,-0.0584184565,-0.2641685605,-0.0197155587,0.1655121893,0.0201513059,0.1763112247,0.0855540112,-0.2054276317,-0.0731580183,-0.2630326152,-0.285345912,0.0199485365,0.2054544538,0.1803489923,0.4772658944,-0.024977481,-0.1353630126,-0.2053964734,-0.3153861165,0.0442342833,0.2464950234,-0.0167008191,0.1939774603,0.0213793144,-0.2587883472,-0.1571341008,0.0688675344,-0.0830654427,-0.0705593601,-0.056711264,0.4370977283,0.1046389937,0.3015836179,0.1242610514,0.0017419218,0.090929538,-0.0915199444,-0.101004757,-0.1997738779,-0.1029308438,0.0908151269,0.0139182741,-0.1522049904,0.4530825913,-0.0226925332,0.2397299111,-0.4752225578,-0.4546351135,0.263915062,0.0946712494,0.0703559294,0.0831279755,-0.1004693732,0.23054263,-0.1151755601,0.2874084115,-0.3667429388,-0.0099853296,0.171131447,-0.3040320277,-0.0911609456,-0.0429216512,-0.2373358309,-0.2131170332,-0.2420698404,-0.1625316441,0.1925634146,0.0692957714,0.0710771754,-0.0482115149,0.2399497479,0.0612403527,0.420825839,-0.0505228527,-0.0142032336,0.3403245509,-0.144912675,-0.4585410953,0.1990785897,-0.0710253119,0.4873732924,0.1876758784,-0.3572619855,-0.1432136744,-0.0876857936,-0.010406387,-0.2309728861,-0.3154932559,0.4347070456,0.0858675763,-0.0017584566,-0.018087849,-0.2098874748,-0.0058370498,-0.1161855832,0.1107516959,0.0180559624,0.5218024254,0.0161350742,0.5054344535,-0.1492939889,-0.1849232465,0.0787607208,-0.0914802626,0.1306820661,-0.077742286,-0.3543592393,0.0909198746,-0.0784771517,0.0423089303,-0.1143648326,-0.2375506312,-0.2545681596,-0.2243156433,0.1367321461,0.2056726664,-0.0749225467,0.1470566839,-0.0218521375,0.234773621,-0.1005828083,0.2911744118,-0.0582701378,-0.0206460617,-0.0068821111,0.40099141,0.2471125424,0.0405494012,-0.3504697084,-0.4177759588,-0.3089730442,0.1535237283,0.1111554801,0.4341416359,-0.2535318136,0.1353392899,0.2518220842,-0.0731290355,0.1305496842,0.201589033,-0.0769401938,0.1304474622,-0.5629460812,-0.2140490711,-0.4025836885,-0.2041153461,0.3084562123,-0.1378363967,0.2591085434,-0.052187264,-0.2361831814,0.377224803,-0.0014561275,-0.3549470007,-0.1654876024,-0.3093603849,-0.0071021901,-0.3891410828,0.1418218762,-0.1630358249,0.1648108363,0.1033505574,0.1802528203,-0.0478269979,-0.0673618242,-0.2418477237,0.1251702905,0.1508538276,-0.1796664447,0.0206297506,0.2583088875,0.4035355747,0.0246620458,0.1251704544,0.1658738106,-0.6877144575,0.2279514372,0.143776834,0.0337538272,-0.0006095635,-0.2813970745,-0.015133095,0.0288642962,0.2616087496,-0.2789119482,0.2849281132,0.4318098724,0.0435546674,-0.3361482918,-0.5102328062,-0.0285126679,0.1190533563,0.130708918,0.1072735712,-0.3197688758,-0.2387962937,0.1009448618,0.2339786887,0.3998080194,-0.1230270714,0.1631871611,0.3690173328,-0.0562773384,0.669048965,-0.0701484829,0.5010374784,-0.2964494824,-0.0236951038,-0.0219667424,-0.3330355585,0.1249234751,-0.0894318596,0.0196071472,0.4615965784,0.0227423031,0.1879124045,0.0824428424,0.5432862043,-0.056193281,-0.3151729703,0.0893430561,0.0824313834,-0.4715276957,0.4554201663,-0.1119189188,0.1296059489,0.2646174729,-0.0853899047,-0.2024868429,0.3169839978,-0.1035186201,0.5135717988,-0.260309875,-0.7178820372,0.2217282057,0.1935005784,0.0475799516,-0.0163907688,-0.0840439871,0.2869564891,-0.5260500908,-0.264888078,-0.1186811849,0.1446561515,0.41260463,-0.1766412705,-0.1750865728,0.0456954911,-0.0207092706,-0.3480331004,-0.1475058943,0.0229152385,0.523981154,0.0506121628,0.1382592916,-0.2436705977,-0.1985277981,-0.1187596694,0.1392894685,0.1864422858,-0.1920204759,0.0092545468,0.112670213,-0.2367699146,0.0267793797,0.4035091996,0.2485736609,-0.0754849836,0.7604902983,0.2701005936,-0.2153382152,-0.1895247102,-0.5757737756,0.1880673319,-0.4746957123,0.0904080123,0.0883788392,-0.0010317379,-0.0252464861,0.0832523406,0.240493089,-0.3185773492,-0.2064391375,-0.4979458451,-0.2446030527,-0.0121778939,-0.0970734507,0.1446900964,0.4060983658,0.0213105734,-0.0374526568,-0.1176230013,-0.2486613244,-0.0812461898,-0.0857515708,-0.0178909376,0.0288438704,-0.2464593798,0.0180503707,0.1339568049,0.116580613,-0.0161814895,-0.0523238592,-0.1377302259,-0.0593801923,0.0638023913,0.1501113325,-0.1841318607,-0.1567177474,-0.3472022116,-0.1686961502,0.0619377308,0.3511345983,0.073198162,-0.0336022787,0.1000552177,-0.1815095842,0.0305698644,-0.2393947244,0.1999550611,-0.1581184864,-0.0344601683,0.2619611919,0.2125540972,0.0491624214,0.1751203686,-0.296749711,0.0122222546,-0.2247535437,0.0795712546,0.3878742754,-0.5469018221,0.3201330304,-0.114045985,0.3263936639,0.2481028885,-0.1112366244,0.0069457488,0.2077629119,0.192999661,-0.3124342859,-0.1750279665,-0.2047646791,0.0255535133,0.1951705664,-0.0336907506,0.0335218683,0.2034249008,0.2418436557,0.2176680118,0.3072521389,-0.2194328308,0.1608728319,0.2555875778,-0.0315138809,0.1866933107,0.4408334792,0.2895790637,0.0344299637,0.2087427378,-0.0501233935,0.268355161,-0.0355201736,0.0493080206,0.2784291506,-0.1406375766,-0.1351004839,0.1598717272,-0.0666537955,-0.0975729823,-0.2356363982,0.1379940212,-0.5082181096,-0.0016119493,-0.3276998401,-0.0616720878,-0.3606294394,0.1335805058,-0.25792858,-0.0233722795,0.0797757208,0.1772313267,0.1117885932,0.1014342308,-0.4469520748,0.1259376109,-0.4083172977,0.0444617681,-0.2394631505,0.3101462126,-0.1211525947,-0.0841582343,0.1898356527,0.2903619409,-0.1219391599,-0.1150254458,0.1751779616,0.4762858152,0.1457411945,0.0434701666,-0.0806150883,-0.0292948931,-0.0321252085,-0.0190826487,0.2248752862,0.2004404217,-0.0397763513,0.1562915593,0.1541848183,-0.1182024628,0.1891729832,0.1260015219,0.0241076648,0.1159706339,0.2661477029,0.0735388547,-0.154797852,-0.1336262524,-0.0261749905,-0.3192779124,-0.1522497237,0.3894211948,0.2127337903,-0.2353196889,-0.1131786257,0.0737563595,0.0157800689,0.4464921057,0.1291050613,0.004601663,-0.1204414517,0.0396351293,-0.2076597214,0.2196232378,-0.023899002,0.0166291501,0.1226764172,0.0947790444,-0.0264727958,-0.0253282469,0.3033050895,-0.1253412366,0.1207330897,-0.1434566081,-0.2756644189,-0.0575891696,0.3088068068,0.0814981163,0.0755050182,-0.2945995927,0.0161185488,0.0222241674,0.1112492085,-0.2723045647,0.0581626073,0.0998518392,-0.3974215984,0.2679631114,-0.0830507725,0.7890290618,0.0927614272,-0.1029083282,-0.0044994657,-0.2056708336,-0.3405399323,0.2708170116,0.0972320586,0.5719087124,0.1613240987,-0.1286329329,-0.4244962335,0.3670101762,-0.1697560996,-0.6661804914,-0.1641177237,0.1922504455,-0.1173405647,0.2502705753,0.2245024294,0.1029542163,0.306530565,-0.2895623446,-0.3585673571,-0.1202796698,0.5804312229,-0.2649388611,-0.2065294981,-0.2863225341,0.1459724754,-0.1629391909,0.1301871687,-0.1651324928,0.1129610091,0.0261454377,-0.0039779241,-0.0153508894,0.2036131471,0.2484441996,0.1664725542,-0.1400396973,0.4426274598,-0.1489300579,-0.2323542684,-0.3859888911,-0.2470760643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2068","title":"PyTorch not available error on SageMaker GPU docker though it is installed ","comments":"@philschmid Recently released tensorflow-macos seems to be missing. ","body":"I get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n","comment_length":8,"text":"PyTorch not available error on SageMaker GPU docker though it is installed \nI get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n\n@philschmid Recently released tensorflow-macos seems to be missing. ","embeddings":[-0.3578641713,0.0624302998,-0.0035595952,0.092395246,0.2011187971,0.0346482657,0.5519385934,0.3234077096,0.3856204748,0.0929586664,0.0008824305,0.396581918,-0.0354305878,0.1952465475,0.0736841932,-0.2641968429,-0.0560292676,0.228468731,-0.3105959892,0.0267087314,0.171954751,0.1197907999,-0.2741039693,0.1857319325,-0.3524909317,-0.093010217,0.250656724,-0.0640398636,-0.178896144,-0.4069884419,0.2793971002,-0.0613616109,0.2397314608,0.4969630539,-0.0001140436,0.303216666,0.1815663874,-0.0520816073,-0.2339606434,-0.4946050942,0.2693472505,-0.2135497779,0.2534041703,-0.3587452173,-0.2566689551,-0.6176720858,-0.1617843211,-0.1064133495,0.1897276342,0.4830078483,0.1778384894,0.2304899693,0.3704248965,-0.1068839654,-0.0213627666,-0.1286899149,-0.1998843551,0.193384558,0.060613174,-0.0477185659,-0.2215194106,0.3306259215,0.0532887354,0.222189188,0.2746862173,-0.0277601741,0.3027733862,-0.3284738958,0.0128042428,0.0186378229,0.3783850074,-0.1673075259,-0.2615454197,-0.1401580125,0.2406661659,-0.185671851,0.2258491665,0.2011611909,-0.2385996133,-0.2307838053,-0.0924862772,-0.2448174804,-0.4942284524,0.0012739522,-0.1486772597,0.6484239101,-0.009730015,0.2958637774,-0.1088481694,0.03162992,0.108798191,-0.0174539723,0.1367830634,-0.1035540998,-0.4412149191,0.0121147214,-0.184800595,-0.3475169241,-0.1545310766,0.1763308048,0.0244534295,0.0385519378,0.4613890648,0.2319640964,-0.0648741797,0.2133830935,-0.1676820517,0.2672363818,0.3156029582,0.1783833504,0.0582823865,-0.0110221663,-0.2719790041,-0.2618050277,0.0919806734,-0.0409463309,0.1839132905,-0.2599915266,-0.1896698624,-0.0647259578,-0.2254171371,-0.0048059095,0.3427653015,0.3795002401,-0.3616908193,0.0215544086,0.3905214369,0.0970223024,-0.1571545601,0.1086176783,-0.0689175576,0.36923787,-0.3053956032,-0.1699036956,0.2570928931,-0.2832776308,0.319039315,0.0521265119,0.4382136464,0.0194357745,0.1652491987,-0.0362840407,-0.0305198543,0.3495331705,0.1684820801,-0.0993223712,0.0673144832,-0.1310516298,0.0435603149,0.1640109867,-0.3027131259,-0.1731041074,-0.2777281404,0.2180583179,0.0239804499,-0.0828475431,-0.0956024304,-0.2363593727,0.1384330541,-0.2462698221,-0.1988418698,-0.5774737597,-0.1518825144,0.1009746268,0.3505586982,-0.1060075387,-0.2552392185,-0.242928192,0.21696347,-0.044582881,0.1921145767,0.167269066,-0.3213567138,0.364137888,-0.2241795808,-0.1162741482,0.4368959367,-0.1582633406,-0.4274958074,-0.0526534356,-0.131755814,-0.1814903766,0.2424345016,-0.0210299063,0.1027198359,0.0694861561,0.1587241441,0.2212405205,0.055063583,-0.0641467497,0.0711620227,-0.0301146377,0.0693062246,0.3470240831,0.4696928263,0.0405193046,-0.2203280181,0.1052827984,0.0011496237,-0.3349923491,-0.0729082674,0.1936165094,0.0726030841,0.1973171979,0.234205395,-0.0241505578,-0.284635216,0.437253207,-0.4078389406,0.2201975584,-0.4012741446,0.2799266577,0.0739980042,0.2309345454,-0.4405454695,-0.0290490314,0.0740211755,0.2747321129,0.1433665305,-0.071650967,-0.0431889445,0.1642151922,-0.2015472054,0.2570536733,-0.0297232922,0.1648442894,-0.3522476256,-0.2846183181,-0.026778385,0.0832818747,0.0867885128,-0.133834824,-0.1851904988,-0.0223054718,0.0563509651,-0.1856317818,-0.4077399969,0.0107585881,0.2134346366,-0.0264883619,0.1546373963,0.2261991501,0.0535074882,0.0659499317,-0.0596124232,0.2794193625,0.0768524185,0.3316195607,0.1563586891,0.1243275106,0.3506565392,0.1796323806,-0.0370965451,0.0312511362,0.3116858006,-0.1815500855,0.3884941638,-0.2803410292,-0.0670936331,-0.2269870043,0.0582206845,0.096817784,0.2094515562,-0.0584184565,-0.2641685605,-0.0197155587,0.1655121893,0.0201513059,0.1763112247,0.0855540112,-0.2054276317,-0.0731580183,-0.2630326152,-0.285345912,0.0199485365,0.2054544538,0.1803489923,0.4772658944,-0.024977481,-0.1353630126,-0.2053964734,-0.3153861165,0.0442342833,0.2464950234,-0.0167008191,0.1939774603,0.0213793144,-0.2587883472,-0.1571341008,0.0688675344,-0.0830654427,-0.0705593601,-0.056711264,0.4370977283,0.1046389937,0.3015836179,0.1242610514,0.0017419218,0.090929538,-0.0915199444,-0.101004757,-0.1997738779,-0.1029308438,0.0908151269,0.0139182741,-0.1522049904,0.4530825913,-0.0226925332,0.2397299111,-0.4752225578,-0.4546351135,0.263915062,0.0946712494,0.0703559294,0.0831279755,-0.1004693732,0.23054263,-0.1151755601,0.2874084115,-0.3667429388,-0.0099853296,0.171131447,-0.3040320277,-0.0911609456,-0.0429216512,-0.2373358309,-0.2131170332,-0.2420698404,-0.1625316441,0.1925634146,0.0692957714,0.0710771754,-0.0482115149,0.2399497479,0.0612403527,0.420825839,-0.0505228527,-0.0142032336,0.3403245509,-0.144912675,-0.4585410953,0.1990785897,-0.0710253119,0.4873732924,0.1876758784,-0.3572619855,-0.1432136744,-0.0876857936,-0.010406387,-0.2309728861,-0.3154932559,0.4347070456,0.0858675763,-0.0017584566,-0.018087849,-0.2098874748,-0.0058370498,-0.1161855832,0.1107516959,0.0180559624,0.5218024254,0.0161350742,0.5054344535,-0.1492939889,-0.1849232465,0.0787607208,-0.0914802626,0.1306820661,-0.077742286,-0.3543592393,0.0909198746,-0.0784771517,0.0423089303,-0.1143648326,-0.2375506312,-0.2545681596,-0.2243156433,0.1367321461,0.2056726664,-0.0749225467,0.1470566839,-0.0218521375,0.234773621,-0.1005828083,0.2911744118,-0.0582701378,-0.0206460617,-0.0068821111,0.40099141,0.2471125424,0.0405494012,-0.3504697084,-0.4177759588,-0.3089730442,0.1535237283,0.1111554801,0.4341416359,-0.2535318136,0.1353392899,0.2518220842,-0.0731290355,0.1305496842,0.201589033,-0.0769401938,0.1304474622,-0.5629460812,-0.2140490711,-0.4025836885,-0.2041153461,0.3084562123,-0.1378363967,0.2591085434,-0.052187264,-0.2361831814,0.377224803,-0.0014561275,-0.3549470007,-0.1654876024,-0.3093603849,-0.0071021901,-0.3891410828,0.1418218762,-0.1630358249,0.1648108363,0.1033505574,0.1802528203,-0.0478269979,-0.0673618242,-0.2418477237,0.1251702905,0.1508538276,-0.1796664447,0.0206297506,0.2583088875,0.4035355747,0.0246620458,0.1251704544,0.1658738106,-0.6877144575,0.2279514372,0.143776834,0.0337538272,-0.0006095635,-0.2813970745,-0.015133095,0.0288642962,0.2616087496,-0.2789119482,0.2849281132,0.4318098724,0.0435546674,-0.3361482918,-0.5102328062,-0.0285126679,0.1190533563,0.130708918,0.1072735712,-0.3197688758,-0.2387962937,0.1009448618,0.2339786887,0.3998080194,-0.1230270714,0.1631871611,0.3690173328,-0.0562773384,0.669048965,-0.0701484829,0.5010374784,-0.2964494824,-0.0236951038,-0.0219667424,-0.3330355585,0.1249234751,-0.0894318596,0.0196071472,0.4615965784,0.0227423031,0.1879124045,0.0824428424,0.5432862043,-0.056193281,-0.3151729703,0.0893430561,0.0824313834,-0.4715276957,0.4554201663,-0.1119189188,0.1296059489,0.2646174729,-0.0853899047,-0.2024868429,0.3169839978,-0.1035186201,0.5135717988,-0.260309875,-0.7178820372,0.2217282057,0.1935005784,0.0475799516,-0.0163907688,-0.0840439871,0.2869564891,-0.5260500908,-0.264888078,-0.1186811849,0.1446561515,0.41260463,-0.1766412705,-0.1750865728,0.0456954911,-0.0207092706,-0.3480331004,-0.1475058943,0.0229152385,0.523981154,0.0506121628,0.1382592916,-0.2436705977,-0.1985277981,-0.1187596694,0.1392894685,0.1864422858,-0.1920204759,0.0092545468,0.112670213,-0.2367699146,0.0267793797,0.4035091996,0.2485736609,-0.0754849836,0.7604902983,0.2701005936,-0.2153382152,-0.1895247102,-0.5757737756,0.1880673319,-0.4746957123,0.0904080123,0.0883788392,-0.0010317379,-0.0252464861,0.0832523406,0.240493089,-0.3185773492,-0.2064391375,-0.4979458451,-0.2446030527,-0.0121778939,-0.0970734507,0.1446900964,0.4060983658,0.0213105734,-0.0374526568,-0.1176230013,-0.2486613244,-0.0812461898,-0.0857515708,-0.0178909376,0.0288438704,-0.2464593798,0.0180503707,0.1339568049,0.116580613,-0.0161814895,-0.0523238592,-0.1377302259,-0.0593801923,0.0638023913,0.1501113325,-0.1841318607,-0.1567177474,-0.3472022116,-0.1686961502,0.0619377308,0.3511345983,0.073198162,-0.0336022787,0.1000552177,-0.1815095842,0.0305698644,-0.2393947244,0.1999550611,-0.1581184864,-0.0344601683,0.2619611919,0.2125540972,0.0491624214,0.1751203686,-0.296749711,0.0122222546,-0.2247535437,0.0795712546,0.3878742754,-0.5469018221,0.3201330304,-0.114045985,0.3263936639,0.2481028885,-0.1112366244,0.0069457488,0.2077629119,0.192999661,-0.3124342859,-0.1750279665,-0.2047646791,0.0255535133,0.1951705664,-0.0336907506,0.0335218683,0.2034249008,0.2418436557,0.2176680118,0.3072521389,-0.2194328308,0.1608728319,0.2555875778,-0.0315138809,0.1866933107,0.4408334792,0.2895790637,0.0344299637,0.2087427378,-0.0501233935,0.268355161,-0.0355201736,0.0493080206,0.2784291506,-0.1406375766,-0.1351004839,0.1598717272,-0.0666537955,-0.0975729823,-0.2356363982,0.1379940212,-0.5082181096,-0.0016119493,-0.3276998401,-0.0616720878,-0.3606294394,0.1335805058,-0.25792858,-0.0233722795,0.0797757208,0.1772313267,0.1117885932,0.1014342308,-0.4469520748,0.1259376109,-0.4083172977,0.0444617681,-0.2394631505,0.3101462126,-0.1211525947,-0.0841582343,0.1898356527,0.2903619409,-0.1219391599,-0.1150254458,0.1751779616,0.4762858152,0.1457411945,0.0434701666,-0.0806150883,-0.0292948931,-0.0321252085,-0.0190826487,0.2248752862,0.2004404217,-0.0397763513,0.1562915593,0.1541848183,-0.1182024628,0.1891729832,0.1260015219,0.0241076648,0.1159706339,0.2661477029,0.0735388547,-0.154797852,-0.1336262524,-0.0261749905,-0.3192779124,-0.1522497237,0.3894211948,0.2127337903,-0.2353196889,-0.1131786257,0.0737563595,0.0157800689,0.4464921057,0.1291050613,0.004601663,-0.1204414517,0.0396351293,-0.2076597214,0.2196232378,-0.023899002,0.0166291501,0.1226764172,0.0947790444,-0.0264727958,-0.0253282469,0.3033050895,-0.1253412366,0.1207330897,-0.1434566081,-0.2756644189,-0.0575891696,0.3088068068,0.0814981163,0.0755050182,-0.2945995927,0.0161185488,0.0222241674,0.1112492085,-0.2723045647,0.0581626073,0.0998518392,-0.3974215984,0.2679631114,-0.0830507725,0.7890290618,0.0927614272,-0.1029083282,-0.0044994657,-0.2056708336,-0.3405399323,0.2708170116,0.0972320586,0.5719087124,0.1613240987,-0.1286329329,-0.4244962335,0.3670101762,-0.1697560996,-0.6661804914,-0.1641177237,0.1922504455,-0.1173405647,0.2502705753,0.2245024294,0.1029542163,0.306530565,-0.2895623446,-0.3585673571,-0.1202796698,0.5804312229,-0.2649388611,-0.2065294981,-0.2863225341,0.1459724754,-0.1629391909,0.1301871687,-0.1651324928,0.1129610091,0.0261454377,-0.0039779241,-0.0153508894,0.2036131471,0.2484441996,0.1664725542,-0.1400396973,0.4426274598,-0.1489300579,-0.2323542684,-0.3859888911,-0.2470760643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2068","title":"PyTorch not available error on SageMaker GPU docker though it is installed ","comments":"I've created a PR to add this. ","body":"I get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n","comment_length":7,"text":"PyTorch not available error on SageMaker GPU docker though it is installed \nI get en error when running data loading using SageMaker SDK\r\n\r\n```\r\n File \"main.py\", line 34, in \r\n run_training()\r\n File \"main.py\", line 25, in run_training\r\n dm.setup('fit')\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/pytorch_lightning\/core\/datamodule.py\", line 92, in wrapped_fn\r\n return fn(*args, **kwargs)\r\n File \"\/opt\/ml\/code\/data_module.py\", line 103, in setup\r\n self.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 995, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\nFile \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/formatting\/__init__.py\", line 114, in get_formatter\r\n raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]\r\nValueError: PyTorch needs to be installed to be able to return PyTorch tensors.\r\n```\r\n\r\nwhen trying to execute dataset loading using this notebook https:\/\/github.com\/PyTorchLightning\/pytorch-lightning\/blob\/master\/notebooks\/04-transformers-text-classification.ipynb, specifically lines \r\n\r\n```\r\nself.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]\r\nself.dataset[split].set_format(type=\"torch\", columns=self.columns)\r\n```\r\n\r\nThe SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com\/pytorch-training:1.4.0-gpu-py3 .\r\n\r\nBy running container interactively I have checked that torch loading completes successfully by executing `https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/config.py#L39`. \r\n\r\nAlso as a first line in the data loading module I have \r\n\r\n```\r\nimport os\r\nos.environ[\"USE_TF\"] = \"0\" \r\nos.environ[\"USE_TORCH\"] = \"1\" \r\n````\r\n\r\nBut unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.\r\nMany Thanks! \r\n\r\n\nI've created a PR to add this. ","embeddings":[-0.3578641713,0.0624302998,-0.0035595952,0.092395246,0.2011187971,0.0346482657,0.5519385934,0.3234077096,0.3856204748,0.0929586664,0.0008824305,0.396581918,-0.0354305878,0.1952465475,0.0736841932,-0.2641968429,-0.0560292676,0.228468731,-0.3105959892,0.0267087314,0.171954751,0.1197907999,-0.2741039693,0.1857319325,-0.3524909317,-0.093010217,0.250656724,-0.0640398636,-0.178896144,-0.4069884419,0.2793971002,-0.0613616109,0.2397314608,0.4969630539,-0.0001140436,0.303216666,0.1815663874,-0.0520816073,-0.2339606434,-0.4946050942,0.2693472505,-0.2135497779,0.2534041703,-0.3587452173,-0.2566689551,-0.6176720858,-0.1617843211,-0.1064133495,0.1897276342,0.4830078483,0.1778384894,0.2304899693,0.3704248965,-0.1068839654,-0.0213627666,-0.1286899149,-0.1998843551,0.193384558,0.060613174,-0.0477185659,-0.2215194106,0.3306259215,0.0532887354,0.222189188,0.2746862173,-0.0277601741,0.3027733862,-0.3284738958,0.0128042428,0.0186378229,0.3783850074,-0.1673075259,-0.2615454197,-0.1401580125,0.2406661659,-0.185671851,0.2258491665,0.2011611909,-0.2385996133,-0.2307838053,-0.0924862772,-0.2448174804,-0.4942284524,0.0012739522,-0.1486772597,0.6484239101,-0.009730015,0.2958637774,-0.1088481694,0.03162992,0.108798191,-0.0174539723,0.1367830634,-0.1035540998,-0.4412149191,0.0121147214,-0.184800595,-0.3475169241,-0.1545310766,0.1763308048,0.0244534295,0.0385519378,0.4613890648,0.2319640964,-0.0648741797,0.2133830935,-0.1676820517,0.2672363818,0.3156029582,0.1783833504,0.0582823865,-0.0110221663,-0.2719790041,-0.2618050277,0.0919806734,-0.0409463309,0.1839132905,-0.2599915266,-0.1896698624,-0.0647259578,-0.2254171371,-0.0048059095,0.3427653015,0.3795002401,-0.3616908193,0.0215544086,0.3905214369,0.0970223024,-0.1571545601,0.1086176783,-0.0689175576,0.36923787,-0.3053956032,-0.1699036956,0.2570928931,-0.2832776308,0.319039315,0.0521265119,0.4382136464,0.0194357745,0.1652491987,-0.0362840407,-0.0305198543,0.3495331705,0.1684820801,-0.0993223712,0.0673144832,-0.1310516298,0.0435603149,0.1640109867,-0.3027131259,-0.1731041074,-0.2777281404,0.2180583179,0.0239804499,-0.0828475431,-0.0956024304,-0.2363593727,0.1384330541,-0.2462698221,-0.1988418698,-0.5774737597,-0.1518825144,0.1009746268,0.3505586982,-0.1060075387,-0.2552392185,-0.242928192,0.21696347,-0.044582881,0.1921145767,0.167269066,-0.3213567138,0.364137888,-0.2241795808,-0.1162741482,0.4368959367,-0.1582633406,-0.4274958074,-0.0526534356,-0.131755814,-0.1814903766,0.2424345016,-0.0210299063,0.1027198359,0.0694861561,0.1587241441,0.2212405205,0.055063583,-0.0641467497,0.0711620227,-0.0301146377,0.0693062246,0.3470240831,0.4696928263,0.0405193046,-0.2203280181,0.1052827984,0.0011496237,-0.3349923491,-0.0729082674,0.1936165094,0.0726030841,0.1973171979,0.234205395,-0.0241505578,-0.284635216,0.437253207,-0.4078389406,0.2201975584,-0.4012741446,0.2799266577,0.0739980042,0.2309345454,-0.4405454695,-0.0290490314,0.0740211755,0.2747321129,0.1433665305,-0.071650967,-0.0431889445,0.1642151922,-0.2015472054,0.2570536733,-0.0297232922,0.1648442894,-0.3522476256,-0.2846183181,-0.026778385,0.0832818747,0.0867885128,-0.133834824,-0.1851904988,-0.0223054718,0.0563509651,-0.1856317818,-0.4077399969,0.0107585881,0.2134346366,-0.0264883619,0.1546373963,0.2261991501,0.0535074882,0.0659499317,-0.0596124232,0.2794193625,0.0768524185,0.3316195607,0.1563586891,0.1243275106,0.3506565392,0.1796323806,-0.0370965451,0.0312511362,0.3116858006,-0.1815500855,0.3884941638,-0.2803410292,-0.0670936331,-0.2269870043,0.0582206845,0.096817784,0.2094515562,-0.0584184565,-0.2641685605,-0.0197155587,0.1655121893,0.0201513059,0.1763112247,0.0855540112,-0.2054276317,-0.0731580183,-0.2630326152,-0.285345912,0.0199485365,0.2054544538,0.1803489923,0.4772658944,-0.024977481,-0.1353630126,-0.2053964734,-0.3153861165,0.0442342833,0.2464950234,-0.0167008191,0.1939774603,0.0213793144,-0.2587883472,-0.1571341008,0.0688675344,-0.0830654427,-0.0705593601,-0.056711264,0.4370977283,0.1046389937,0.3015836179,0.1242610514,0.0017419218,0.090929538,-0.0915199444,-0.101004757,-0.1997738779,-0.1029308438,0.0908151269,0.0139182741,-0.1522049904,0.4530825913,-0.0226925332,0.2397299111,-0.4752225578,-0.4546351135,0.263915062,0.0946712494,0.0703559294,0.0831279755,-0.1004693732,0.23054263,-0.1151755601,0.2874084115,-0.3667429388,-0.0099853296,0.171131447,-0.3040320277,-0.0911609456,-0.0429216512,-0.2373358309,-0.2131170332,-0.2420698404,-0.1625316441,0.1925634146,0.0692957714,0.0710771754,-0.0482115149,0.2399497479,0.0612403527,0.420825839,-0.0505228527,-0.0142032336,0.3403245509,-0.144912675,-0.4585410953,0.1990785897,-0.0710253119,0.4873732924,0.1876758784,-0.3572619855,-0.1432136744,-0.0876857936,-0.010406387,-0.2309728861,-0.3154932559,0.4347070456,0.0858675763,-0.0017584566,-0.018087849,-0.2098874748,-0.0058370498,-0.1161855832,0.1107516959,0.0180559624,0.5218024254,0.0161350742,0.5054344535,-0.1492939889,-0.1849232465,0.0787607208,-0.0914802626,0.1306820661,-0.077742286,-0.3543592393,0.0909198746,-0.0784771517,0.0423089303,-0.1143648326,-0.2375506312,-0.2545681596,-0.2243156433,0.1367321461,0.2056726664,-0.0749225467,0.1470566839,-0.0218521375,0.234773621,-0.1005828083,0.2911744118,-0.0582701378,-0.0206460617,-0.0068821111,0.40099141,0.2471125424,0.0405494012,-0.3504697084,-0.4177759588,-0.3089730442,0.1535237283,0.1111554801,0.4341416359,-0.2535318136,0.1353392899,0.2518220842,-0.0731290355,0.1305496842,0.201589033,-0.0769401938,0.1304474622,-0.5629460812,-0.2140490711,-0.4025836885,-0.2041153461,0.3084562123,-0.1378363967,0.2591085434,-0.052187264,-0.2361831814,0.377224803,-0.0014561275,-0.3549470007,-0.1654876024,-0.3093603849,-0.0071021901,-0.3891410828,0.1418218762,-0.1630358249,0.1648108363,0.1033505574,0.1802528203,-0.0478269979,-0.0673618242,-0.2418477237,0.1251702905,0.1508538276,-0.1796664447,0.0206297506,0.2583088875,0.4035355747,0.0246620458,0.1251704544,0.1658738106,-0.6877144575,0.2279514372,0.143776834,0.0337538272,-0.0006095635,-0.2813970745,-0.015133095,0.0288642962,0.2616087496,-0.2789119482,0.2849281132,0.4318098724,0.0435546674,-0.3361482918,-0.5102328062,-0.0285126679,0.1190533563,0.130708918,0.1072735712,-0.3197688758,-0.2387962937,0.1009448618,0.2339786887,0.3998080194,-0.1230270714,0.1631871611,0.3690173328,-0.0562773384,0.669048965,-0.0701484829,0.5010374784,-0.2964494824,-0.0236951038,-0.0219667424,-0.3330355585,0.1249234751,-0.0894318596,0.0196071472,0.4615965784,0.0227423031,0.1879124045,0.0824428424,0.5432862043,-0.056193281,-0.3151729703,0.0893430561,0.0824313834,-0.4715276957,0.4554201663,-0.1119189188,0.1296059489,0.2646174729,-0.0853899047,-0.2024868429,0.3169839978,-0.1035186201,0.5135717988,-0.260309875,-0.7178820372,0.2217282057,0.1935005784,0.0475799516,-0.0163907688,-0.0840439871,0.2869564891,-0.5260500908,-0.264888078,-0.1186811849,0.1446561515,0.41260463,-0.1766412705,-0.1750865728,0.0456954911,-0.0207092706,-0.3480331004,-0.1475058943,0.0229152385,0.523981154,0.0506121628,0.1382592916,-0.2436705977,-0.1985277981,-0.1187596694,0.1392894685,0.1864422858,-0.1920204759,0.0092545468,0.112670213,-0.2367699146,0.0267793797,0.4035091996,0.2485736609,-0.0754849836,0.7604902983,0.2701005936,-0.2153382152,-0.1895247102,-0.5757737756,0.1880673319,-0.4746957123,0.0904080123,0.0883788392,-0.0010317379,-0.0252464861,0.0832523406,0.240493089,-0.3185773492,-0.2064391375,-0.4979458451,-0.2446030527,-0.0121778939,-0.0970734507,0.1446900964,0.4060983658,0.0213105734,-0.0374526568,-0.1176230013,-0.2486613244,-0.0812461898,-0.0857515708,-0.0178909376,0.0288438704,-0.2464593798,0.0180503707,0.1339568049,0.116580613,-0.0161814895,-0.0523238592,-0.1377302259,-0.0593801923,0.0638023913,0.1501113325,-0.1841318607,-0.1567177474,-0.3472022116,-0.1686961502,0.0619377308,0.3511345983,0.073198162,-0.0336022787,0.1000552177,-0.1815095842,0.0305698644,-0.2393947244,0.1999550611,-0.1581184864,-0.0344601683,0.2619611919,0.2125540972,0.0491624214,0.1751203686,-0.296749711,0.0122222546,-0.2247535437,0.0795712546,0.3878742754,-0.5469018221,0.3201330304,-0.114045985,0.3263936639,0.2481028885,-0.1112366244,0.0069457488,0.2077629119,0.192999661,-0.3124342859,-0.1750279665,-0.2047646791,0.0255535133,0.1951705664,-0.0336907506,0.0335218683,0.2034249008,0.2418436557,0.2176680118,0.3072521389,-0.2194328308,0.1608728319,0.2555875778,-0.0315138809,0.1866933107,0.4408334792,0.2895790637,0.0344299637,0.2087427378,-0.0501233935,0.268355161,-0.0355201736,0.0493080206,0.2784291506,-0.1406375766,-0.1351004839,0.1598717272,-0.0666537955,-0.0975729823,-0.2356363982,0.1379940212,-0.5082181096,-0.0016119493,-0.3276998401,-0.0616720878,-0.3606294394,0.1335805058,-0.25792858,-0.0233722795,0.0797757208,0.1772313267,0.1117885932,0.1014342308,-0.4469520748,0.1259376109,-0.4083172977,0.0444617681,-0.2394631505,0.3101462126,-0.1211525947,-0.0841582343,0.1898356527,0.2903619409,-0.1219391599,-0.1150254458,0.1751779616,0.4762858152,0.1457411945,0.0434701666,-0.0806150883,-0.0292948931,-0.0321252085,-0.0190826487,0.2248752862,0.2004404217,-0.0397763513,0.1562915593,0.1541848183,-0.1182024628,0.1891729832,0.1260015219,0.0241076648,0.1159706339,0.2661477029,0.0735388547,-0.154797852,-0.1336262524,-0.0261749905,-0.3192779124,-0.1522497237,0.3894211948,0.2127337903,-0.2353196889,-0.1131786257,0.0737563595,0.0157800689,0.4464921057,0.1291050613,0.004601663,-0.1204414517,0.0396351293,-0.2076597214,0.2196232378,-0.023899002,0.0166291501,0.1226764172,0.0947790444,-0.0264727958,-0.0253282469,0.3033050895,-0.1253412366,0.1207330897,-0.1434566081,-0.2756644189,-0.0575891696,0.3088068068,0.0814981163,0.0755050182,-0.2945995927,0.0161185488,0.0222241674,0.1112492085,-0.2723045647,0.0581626073,0.0998518392,-0.3974215984,0.2679631114,-0.0830507725,0.7890290618,0.0927614272,-0.1029083282,-0.0044994657,-0.2056708336,-0.3405399323,0.2708170116,0.0972320586,0.5719087124,0.1613240987,-0.1286329329,-0.4244962335,0.3670101762,-0.1697560996,-0.6661804914,-0.1641177237,0.1922504455,-0.1173405647,0.2502705753,0.2245024294,0.1029542163,0.306530565,-0.2895623446,-0.3585673571,-0.1202796698,0.5804312229,-0.2649388611,-0.2065294981,-0.2863225341,0.1459724754,-0.1629391909,0.1301871687,-0.1651324928,0.1129610091,0.0261454377,-0.0039779241,-0.0153508894,0.2036131471,0.2484441996,0.1664725542,-0.1400396973,0.4426274598,-0.1489300579,-0.2323542684,-0.3859888911,-0.2470760643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2067","title":"Multiprocessing windows error","comments":"Hi ! Thanks for reporting.\r\nThis looks like a bug, could you try to provide a minimal code example that reproduces the issue ? This would be very helpful !\r\n\r\nOtherwise I can try to run the wav2vec2 code above on my side but probably not this week..","body":"As described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop ","comment_length":48,"text":"Multiprocessing windows error\nAs described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop \nHi ! Thanks for reporting.\r\nThis looks like a bug, could you try to provide a minimal code example that reproduces the issue ? This would be very helpful !\r\n\r\nOtherwise I can try to run the wav2vec2 code above on my side but probably not this week..","embeddings":[-0.0581421144,-0.4536787868,-0.0355622135,0.2569919527,0.0860946625,-0.0037736383,0.115531303,-0.0629601926,0.0408037528,0.3191868663,0.2538127005,0.0701293051,-0.3072563708,-0.0425456725,-0.2219943404,-0.2463472039,0.1892075837,-0.1294742376,0.3075393438,0.3353129625,-0.425796777,0.5416621566,-0.4026214778,0.1973485947,-0.414326489,0.0059316708,-0.258384347,0.4944570363,0.3894542158,-0.3567263484,0.0712473169,-0.3931779563,0.1509012282,0.1528387666,-0.0001224974,-0.1242067292,0.0399249382,-0.0401795469,0.410685569,0.2325061411,0.2327505797,0.1160329655,-0.0896714181,-0.0001363825,0.1803089231,0.317689836,0.0230255742,-0.0436639488,0.3788876235,0.2494235188,0.1086351946,0.2697453499,0.0584818125,-0.0172719918,-0.3373928964,-0.4087221026,-0.0356067978,0.1879889816,0.4800273478,-0.3880698979,0.2081042826,0.229860425,-0.2476220727,0.0222452357,-0.3043947816,0.0047380412,0.3776167333,-0.5156354904,0.0704574734,-0.08932776,-0.1066430882,-0.1918105632,0.3175110221,0.0956236348,0.1241043136,-0.8489111066,0.1194221899,0.2649233639,-0.3856474757,0.0965462103,-0.0331685618,0.0699983388,-0.060277734,-0.1735407412,0.1558316201,-0.05035761,0.0599201359,0.165252164,0.1297031045,0.3123299479,-0.2853071392,0.2992669642,-0.0785920322,0.0563823245,-0.4582304358,-0.0107892044,0.0245604292,0.2138339579,-0.3155740499,-0.2193693519,-0.0722816586,0.0577091873,0.210329175,0.1877655536,0.2185833901,-0.1882567406,0.0442293398,-0.5686359406,0.1675219238,-0.5050089359,-0.1424953043,-0.0811683536,-0.0080247419,-0.5607304573,0.3573493659,0.3995010257,0.2302154154,-0.1080729291,-0.3386083841,0.2101215869,-0.04086392,0.2901071012,0.273465097,0.2971855104,0.3146837354,0.7152168751,-0.0177826546,0.2513940036,-0.5292391181,0.0686728582,0.0788345113,0.1238428429,-0.2765547633,0.1074580178,0.2124366015,0.0187721364,0.1902923584,0.1390351951,0.2973748744,-0.368814528,-0.0184524059,-0.2097483873,-0.1706739515,0.2196747065,0.2305232435,0.1968380958,-0.0093255816,0.0610076077,0.0238536336,0.2254341245,-0.3823579848,-0.3544293344,-0.1269885302,0.0252814814,0.397954613,0.4406096637,0.3450522423,-0.1068729162,0.3525731564,-0.3299072385,-0.124025993,-0.4204719961,-0.3616985977,-0.34269014,-0.0801263303,0.2953493595,0.030153187,0.3120937049,0.3005546927,-0.0583335087,0.1898305267,0.3420273364,0.2321760505,-0.1541697383,-0.1427909285,-0.0512975156,-0.2075472027,-0.2447627336,0.0287267622,0.1045083106,-0.0876488164,-0.0770462602,0.2426429093,0.0912932903,-0.1405437142,-0.3693986237,0.4593514204,-0.2429709136,-0.1423679143,0.2521110475,-0.1832213998,0.0826770887,-0.0361781567,0.0155518427,0.0554710738,-0.0393529236,-0.0699979812,-0.0750164688,0.2342378497,0.1212345883,0.0071897889,0.2943185866,0.2643739283,-0.0255832206,0.047279533,-0.3173382282,0.02500326,0.0944315717,-0.1837330759,-0.0593166575,0.4418401718,0.1487971097,0.1914493144,0.2850689292,-0.1577903032,-0.2866767049,0.0075830286,-0.0888098776,0.0314149782,-0.2019582391,0.0416018069,0.5415998101,0.0924480632,-0.2050208151,-0.0775554553,-0.0432454906,-0.3279131949,-0.3441634178,-0.0958048925,0.3061342537,0.2324241102,0.2131822407,0.0931091383,0.2249252349,0.178348273,-0.1250432134,-0.1788385957,-0.0189846084,0.1657315642,-0.3093577623,0.0945758149,0.0233976655,0.1686377674,-0.2740843892,0.4795754552,0.1112647653,-0.0392507128,0.24645257,-0.1647881866,0.1857769489,0.1239148453,0.0868051723,0.250082314,0.2957986593,0.3540994227,0.0737764016,-0.0943075716,-0.3773719072,-0.3285286129,0.063435398,0.1519417465,0.2393576354,-0.0259907767,0.058411438,-0.1990371644,0.201797083,-0.0167576745,-0.4524724483,0.736939311,0.0708688125,-0.1586373597,0.3013827205,-0.3818297684,-0.1029828191,0.2564709783,0.0558761545,0.0142505569,0.1514872015,-0.0793216005,-0.0701853782,0.0156658385,-0.0630172342,-0.1121387333,0.0245234929,-0.3366858065,0.1437121034,-0.0703526139,0.5119808912,-0.3930772543,-0.4185869694,-0.0158641767,-0.1448655725,0.0737106055,0.4412177503,-0.1272929609,-0.0961341336,0.0165484846,0.0042984234,0.1229744181,0.2859156728,0.3148186505,-0.0421822332,-0.1014562547,-0.0997362509,0.2983363867,-0.1647490561,0.2087989748,0.3532303572,-0.0372239389,0.0849438533,0.1645260006,-0.0095089255,0.0576119013,0.2622989118,0.0440347157,0.0375002697,0.176767081,-0.4125063121,0.0998600572,-0.3187246919,0.1915216148,0.2610322237,0.1713414192,-0.3974964321,-0.2762928605,-0.0283086598,-0.4276906848,-0.1810949445,0.055199068,-0.1846828759,0.2561765015,-0.2547455132,0.0900386646,0.0847056285,0.2642419636,-0.2037704289,-0.1433807611,0.1716799438,-0.08444231,-0.2580580413,-0.1759235412,0.250746727,0.1663268954,0.0340598263,-0.2300753742,-0.1844059527,-0.3176006377,0.0196092445,0.0322731808,-0.0572806187,0.2107921094,0.3846412003,0.110975489,-0.0915515348,0.0241961833,-0.1118701622,0.1226555035,0.0250537358,0.1420998275,0.2201239318,0.0712735504,-0.169597134,0.5418444872,0.5851689577,0.1231955811,0.3024074733,0.1154460981,0.0167698357,0.0343535766,-0.1813274175,0.0432581604,-0.4538233876,-0.1846060753,0.1852776557,0.1367889047,-0.3691618145,-0.2682437897,0.2060450613,-0.3449836075,-0.1176549792,0.0889454857,0.0945942998,0.2125629932,-0.2040405273,-0.0336106531,0.098182261,0.0672464594,-0.1922684461,0.1643065959,-0.1859418154,-0.2114783525,-0.0058400813,-0.5571390986,-0.3848232925,0.3280596435,0.1219232529,0.5821322799,-0.0677651539,-0.1720140874,-0.0761444643,-0.308447957,0.8612225056,-0.1240653247,-0.1728722751,0.2027513981,0.2131693065,0.0480328351,-0.1954341531,-0.2025592029,0.37669155,0.058495272,-0.0827403739,0.3347232342,0.0146596851,0.0577530563,-0.0758640617,-0.1576246321,-0.113902837,-0.3956267238,-0.2647199929,0.0798702762,0.2954231799,0.0608333573,0.0183233246,-0.4867065847,0.0588874482,0.126533106,-0.0476332903,0.279925704,-0.1075734571,0.1051885411,-0.1954662949,0.2081285864,-0.3091813624,0.3162392676,-0.2209812403,-0.0304387286,-0.1167408898,0.2173627466,0.3433983326,-0.1115778461,0.4617955983,0.4835232198,0.1073469296,0.0466568917,-0.302339673,0.3954612911,-0.2997344136,-0.2929951549,0.3063975275,0.2252731472,0.1064766645,0.3608357608,0.2658946216,0.1251864433,-0.0970824212,0.5405082107,-0.2946852744,-0.0681202635,-0.0578516386,0.0979821086,0.9706500769,0.1672171652,0.1279153377,0.0666872412,-0.4959299266,-0.2037558556,-0.1253502369,0.3211533129,-0.2502980232,-0.2016640007,0.0483656786,-0.1047213301,0.1514296681,-0.1148620918,-0.3159129024,0.4344840944,0.0856264457,-0.2407682389,-0.1079418212,-0.3393341303,-0.0638839006,-0.1676814556,0.0529177934,0.0333369002,0.1299217939,-0.078595832,-0.2901931405,0.1603250206,0.4377808273,-0.3864811659,-0.5973951817,0.0585260205,-0.2218556404,0.1296209246,0.0247004181,0.2141951323,0.1133395508,0.345756501,0.1883313507,-0.063974455,-0.2510094941,0.0370402783,0.4117544293,-0.0252167284,-0.1266994178,0.0186136253,-0.1215969548,-0.1454340816,-0.2624289095,0.3467602432,-0.178740114,-0.0581285991,-0.3553861082,0.3969114125,0.0664770007,0.2234817594,-0.178885743,0.259115994,-0.097214587,-0.1863847822,0.0709263682,0.0898761898,-0.5308449864,0.4942597151,-0.5991371274,-0.0874867067,0.0404233709,0.089947246,-0.0044938931,-0.1427792758,0.1440734267,0.2740443051,-0.0375337601,-0.1893779039,-0.0832269713,-0.5538095236,-0.0639177263,-0.0381033532,0.3845428526,-0.1365790963,0.1988971382,-0.2631217539,0.0343524367,0.0859669745,0.0000766066,-0.4166377187,-0.1524058729,-0.0818365142,-0.1089171171,0.0761634782,-0.0680734739,0.0493103489,0.1297267377,0.0307054073,-0.2212044299,0.0496320724,-0.1451285034,0.1046878174,0.0524083711,-0.1959429979,-0.1174309924,0.1733908802,-0.0139855258,0.3372428119,-0.2768659592,-0.0766106397,0.0311620925,0.125235945,-0.1028774455,-0.012093096,0.1443313956,0.1166210771,0.0461651944,0.0722221285,0.1374606341,0.0705850869,-0.2863956094,0.1049665436,-0.0368608534,0.2669781446,0.1401186436,0.299238652,-0.1195691228,0.4209020436,0.0716726854,0.0121931788,0.1021102071,-0.0031000916,-0.0210774597,0.0072855577,-0.1587189883,-0.1828370839,-0.1783351153,-0.1451350003,-0.2222425491,-0.1602770239,0.1800923944,0.2320832163,-0.1891254485,0.3596039116,0.3616650105,0.1855931133,-0.1767401546,0.1009493098,-0.0847952366,-0.1132143363,-0.2028796822,0.3118568957,0.1951147169,-0.4828325212,0.3124465644,0.0768458173,0.2162156105,-0.399715662,-0.3270638585,0.0999630094,0.2116516829,0.1568398625,-0.1175066382,-0.1622988731,0.0289107859,0.5074876547,-0.0102633378,0.1445217133,-0.1656726748,0.0603520609,-0.0937686563,-0.2105655074,-0.2816865742,0.1144585162,-0.3063831627,0.4082044065,-0.2697686851,0.4192980528,-0.590056777,-0.3131828606,-0.3440953791,0.3409806788,-0.288580358,-0.1261041164,0.137936607,0.0138244499,0.0292382538,0.1049954295,0.3350335658,0.1035922766,0.501178205,0.1839785725,-0.1275366545,-0.0247081649,-0.2411272973,0.248379603,0.0147428485,-0.2052842975,0.1989247203,0.4318730235,-0.0715487301,-0.1616415977,-0.0173576642,0.5543359518,0.3776864409,-0.7202739716,-0.0549807027,0.133795172,0.1619910747,0.0390955843,0.1831469685,0.1867974997,0.2772275209,-0.0301895682,-0.035929855,-0.0572247505,0.0335894153,0.3075830638,0.2061607093,-0.021919664,0.2436186522,-0.1795903891,-0.1927601695,0.0978357941,0.1497306079,-0.5974994302,0.0335827954,0.2408358902,-0.3422866762,0.0626598895,0.1090904102,0.0239202157,-0.0453770608,0.3704228699,-0.0711149499,0.3302574456,-0.2325226814,-0.2024016082,-0.3596875668,0.0660740063,0.0306289624,0.083276771,-0.1472661793,-0.1648820639,-0.0934376866,0.155721128,0.0648449585,0.1428685188,0.2611782253,0.7247493863,-0.6224196553,0.2886497974,0.3918014169,-0.0296211913,0.2471479774,-0.4001511633,0.1693498045,0.010367007,0.0022284242,-0.1773337424,0.1539892852,-0.169414863,0.4468988776,0.1549283415,0.064364627,0.2695637047,0.0336142294,-0.1173482239,0.0193396695,0.2997874618,0.300620079,-0.5452706814,0.2135144025,0.0773072243,-0.4484580457,-0.0199572854,0.009789669,-0.0015112794,0.1182383522,0.1396420598,-0.1543735713,0.0191283431,-0.0932218209,0.1422597766,0.0317066424,0.375374496,-0.0367317013,0.1789433062,-0.2378179431,0.0514129177,0.0894557834,-0.1766910702,-0.3162965477,-0.3609884679,0.5703046322,-0.482365787,0.1083318293,-0.0566998385,-0.1274678409,-0.0527689122,0.0712158307,-0.621537149,0.2518406212,-0.3608244359,-0.6841000915,0.1833940744,-0.1227835491,0.1138790995,-0.0478707999,0.291803509,-0.188757211]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2067","title":"Multiprocessing windows error","comments":"```\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\n\r\n\r\nupdated_dataset = dataset.map(lambda example: {'sentence1': 'My sentence: ' + example['sentence1']}, num_proc=4)\r\n\r\n```","body":"As described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop ","comment_length":22,"text":"Multiprocessing windows error\nAs described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop \n```\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\n\r\n\r\nupdated_dataset = dataset.map(lambda example: {'sentence1': 'My sentence: ' + example['sentence1']}, num_proc=4)\r\n\r\n```","embeddings":[-0.2164214551,-0.269531697,-0.0307765733,0.1850229353,0.1784022748,0.1376417726,0.0575687103,-0.0025023981,0.0087179644,0.2766421437,0.1489748359,0.1610600799,-0.2854522765,-0.1023178324,-0.1237755269,-0.1828478575,0.1945973486,-0.1632987708,0.263887167,0.3403266966,-0.4150704443,0.4792796075,-0.3459638655,0.0734494776,-0.5422175527,-0.1041582078,-0.2976150811,0.4396467209,0.197288841,-0.4517998695,-0.013877756,-0.1838330626,0.2181247473,0.1776028126,-0.0001206921,-0.1065645739,0.0575244762,-0.0216337312,0.3331870735,0.2294842601,0.0989151821,0.0286370553,-0.0890319198,-0.0855964497,0.1740061492,0.2978885472,0.096328333,-0.1282736808,0.5274698138,0.3825234473,0.0758575127,0.2272967398,0.0876476318,0.007820514,-0.1878419071,-0.2888141274,0.0203558411,0.2024262398,0.5406011939,-0.3069016933,0.0876099318,0.2608896792,-0.3369736075,0.0058557997,-0.3788694441,-0.0015098701,0.3409759402,-0.4972144961,0.1355955601,-0.0681021661,-0.0130132353,-0.321980238,0.1528369337,-0.021674145,0.1863062382,-0.8767055869,-0.0115937479,0.1989576966,-0.314476192,0.0741171166,-0.1529564261,-0.0594550893,-0.0408123471,-0.1191436052,0.183025524,0.0230448544,0.0427067503,0.1673840731,0.2412383705,0.2289174944,-0.1954458654,0.2736682594,-0.1021791324,0.13654311,-0.5750361681,-0.0034187443,-0.0358424,0.1198296323,-0.2478046715,-0.3653281629,-0.0101330867,0.066942513,0.2934780419,0.1694190651,0.2630363405,-0.154176116,0.0505522303,-0.5396240354,0.233746469,-0.497073561,-0.1793611497,-0.0397479534,-0.039352864,-0.5704105496,0.3287129104,0.4325676262,0.2949965894,-0.0771086738,-0.3808996379,0.1982300878,-0.0343120545,0.1575937867,0.1864148825,0.2853849232,0.3009817004,0.6022768021,-0.1089114398,0.243097499,-0.5615463257,-0.024918871,0.0065897619,0.0263321586,-0.3680235744,0.1488280743,0.2685838044,0.0536361784,0.1249185503,0.1504223496,0.1372532994,-0.3701722026,-0.1397931427,-0.2361291796,-0.2652803361,0.2284741998,0.2142871171,0.2383062541,-0.0474647135,-0.0238052923,0.0262911543,0.3022709191,-0.3516969085,-0.3986416459,-0.1266380697,0.0315956064,0.3112643957,0.2649718225,0.0864211544,0.1411970854,0.4162356555,-0.2684273422,-0.1504580826,-0.3968690038,-0.2766779661,-0.3078669608,-0.0656391382,0.4170382023,-0.1365273446,0.290032804,0.2960905731,0.0314433426,0.1945922971,0.3655435443,0.1552812606,-0.0813851133,-0.0505696312,0.0832357258,-0.0205600634,-0.2881668806,-0.1992242336,0.176038146,-0.0738317147,-0.0978386626,0.3155908883,0.1629916131,0.0818682835,-0.2875578105,0.4907436669,-0.1638596207,-0.1147925854,0.219432503,-0.1701681465,0.1143033877,0.0020510694,0.0597540326,0.0055045085,-0.100123167,-0.0747044161,-0.0821861401,0.3725978732,0.0493526608,0.0329319201,0.4288325608,0.1251131445,0.042679064,-0.0380034633,-0.4412008226,-0.1410391778,0.0364274196,-0.1454827189,0.1202213094,0.215491876,0.0123201339,0.1436545104,0.2799672484,-0.1707317233,-0.3652132154,0.0262440089,-0.0855276436,0.0470418036,-0.2536361516,-0.0740335807,0.5049322844,-0.0410295837,-0.0658523813,-0.118306987,-0.035705775,-0.2768333554,-0.4343212843,-0.0377741158,0.3438779712,0.2160584927,0.1489761472,0.0480393767,0.3426544666,0.1863054037,0.0081386408,-0.1519266367,-0.0605586432,0.2119040787,-0.3094341457,0.0922355205,0.0435259901,0.1680229157,-0.1974246055,0.3308628798,0.0512520261,0.0455577262,0.3031246662,-0.1308645457,0.1274566352,0.13500835,0.1218287945,0.1568627208,0.1545986831,0.2793993652,0.0749835894,-0.0271780044,-0.2646307051,-0.3050138354,0.0248108041,0.2788725495,0.1574839652,0.0075597251,0.0476800613,-0.3225958943,0.1222753748,-0.0220541358,-0.3254919946,0.7709748745,0.1396939605,-0.0917925388,0.2921946347,-0.43721506,-0.1092860624,0.2887381017,0.0958150402,0.018291682,0.121315375,-0.0725876763,-0.0066961017,-0.1947628707,-0.0800314918,0.0012708978,0.1616831422,-0.3943112493,0.0567448884,-0.1457272321,0.3488927186,-0.3866231441,-0.4130340815,-0.0436930172,-0.2089831978,-0.0044999304,0.3486006856,-0.008753649,0.0981045291,0.0509685092,-0.1500322372,0.1428934932,0.1455201507,0.2662494481,-0.0906187147,-0.1254091561,-0.1048674658,0.3953677416,-0.1487078816,0.2262792736,0.253669858,0.0533809736,-0.0227702446,0.1229572371,-0.1032215282,0.0013769236,0.3677129447,0.0344865471,0.0714260712,0.0728884935,-0.3751188815,0.1372556537,-0.3081077933,0.1805383712,0.1420270801,0.1769648045,-0.428165257,-0.1545358449,-0.2390204221,-0.4491527379,-0.2015045434,0.0472735427,-0.2694447339,0.2583689988,-0.260084033,0.1403761804,0.1228197441,0.3062795401,-0.1006057486,-0.1949616075,0.2093833685,-0.0533713996,-0.2659558058,-0.2152358145,0.0953186676,0.149229303,-0.0089385174,-0.0756138712,-0.2012024671,-0.300093174,-0.1328931153,-0.0041212174,-0.0031261484,0.3379204571,0.5044857264,0.1537652314,-0.0895865709,0.0280563515,-0.1914943606,0.0816868246,0.1449466348,0.1434945613,0.1560181379,0.1827497333,-0.1249647066,0.4865584373,0.6543411016,-0.0066320677,0.3675019145,0.0008984012,0.0191048346,0.0103926109,-0.2558258474,0.0708779246,-0.4434515536,-0.0943344682,0.1967519075,0.1401740909,-0.3902766705,-0.2533397079,0.1971941292,-0.3684377968,-0.1661480367,0.1676751077,0.0287356395,0.2189529389,-0.0908990651,-0.0863769352,0.0174857862,0.023345219,-0.1418199539,0.142405659,-0.1911932528,-0.200051561,-0.1622465104,-0.4725292623,-0.5464188457,0.4420810342,0.1902506799,0.5572493672,-0.0004186116,-0.1033632085,-0.0077233189,-0.2483178228,0.9616155028,-0.2557925582,-0.2321639359,0.2623564005,0.3185871243,-0.1450106353,-0.0631815493,-0.3186502457,0.3344966769,0.1541581601,0.0722035691,0.2450813651,-0.1232593507,0.1116056591,-0.0964407325,-0.1693779081,-0.1299035102,-0.3090285957,-0.2947702408,-0.1030345187,0.2547100186,0.0725449771,0.0647234544,-0.2876007855,-0.0304696225,0.0557192452,-0.0892974436,0.3242617548,-0.0514049903,0.1778756678,-0.0087674288,0.1290720552,-0.3243986666,0.3358855844,-0.1060933918,0.0669693276,-0.2574530542,0.122683607,0.2483448535,-0.0843656734,0.5241105556,0.4933411479,0.0557753593,-0.0079197371,-0.3167228997,0.4300366342,-0.3185167313,-0.0990868956,0.4587610662,0.4184074104,-0.0552098639,0.0967352837,0.3463324308,0.2058133632,-0.1845186651,0.3874590397,-0.1901326776,-0.1893787384,0.0983556658,0.1140466258,1.0056911707,0.0209041238,0.196763292,-0.03634765,-0.4171991348,-0.087803632,-0.1630501002,0.3706013858,-0.1769148409,-0.1923165768,0.0185918212,-0.0688556507,0.2651750445,0.0302693974,-0.3834737241,0.5674396753,0.1610334367,-0.22660999,-0.0926407576,-0.1582125872,-0.1406064034,-0.282017678,-0.0994322971,0.0154652037,0.0581098236,-0.1524652094,-0.2407708913,0.1642874032,0.396204114,-0.3548473716,-0.5855566263,-0.0224851612,-0.3942810297,0.1594375521,-0.1440721452,0.1530199796,0.0704852119,0.5233256817,0.115535751,-0.0268229283,-0.1901291758,0.0285178646,0.4766495228,-0.0435875505,0.0016899272,0.0364158563,0.109239988,-0.1305309981,-0.3378484845,0.2988404334,-0.150852412,-0.0966782793,-0.2932015657,0.1758432388,0.1369691938,0.050684385,-0.2225040942,0.2744401991,-0.0636811107,-0.2475290447,0.0486398898,0.036909923,-0.4689870477,0.6542810202,-0.5486715436,-0.1686793864,0.0071556969,0.1973308325,0.0030326562,-0.1446202546,0.2444828451,0.2999676466,-0.0237352382,-0.1971713752,-0.1212167069,-0.4443232715,-0.0850217417,-0.0212600939,0.3350668252,-0.1141959429,0.2037623972,-0.1973080635,0.1267278939,0.1122926176,0.0136117954,-0.4640013576,-0.3002684116,0.0259882323,0.019407345,0.1499795914,0.0022457936,0.1353275925,0.1626277864,0.0826300234,-0.2359639853,0.0879278705,-0.0734422505,0.1371508688,0.2372467369,-0.2229144424,-0.0619268119,0.1409917176,0.0177729521,0.3506112993,-0.3006501794,-0.1103439629,0.0441642702,0.0941389576,-0.1622806638,0.0090969484,0.0983473286,0.0576677732,0.0028966616,-0.0053525255,0.1868871748,0.167265594,-0.317589879,0.1289644092,0.0337083191,0.2249787748,0.1565757692,0.2435608208,-0.1932958663,0.2854995728,0.1126382127,0.0922877789,0.1190698743,-0.0250295736,-0.0114191342,-0.0004562111,0.0224739537,-0.2408926189,-0.1208977625,-0.2099997997,-0.2908900678,-0.0929395482,0.3533419669,0.3665493131,-0.1636881381,0.2602789998,0.3331291676,0.1632746011,-0.2968601584,0.089497745,0.0852279812,-0.1177003309,-0.117340073,0.3039800823,0.2468138039,-0.3369474113,0.128022626,0.068053782,0.264403075,-0.3204004467,-0.3003346324,0.2181894183,0.2322850376,0.1148705855,-0.0566619821,0.0166876707,0.1003807411,0.5405622125,-0.1263541132,0.2218711525,-0.1442343891,0.1424398869,-0.1031871215,-0.1945936084,-0.1662879586,0.1563696563,-0.2756224275,0.451944083,-0.2314283997,0.6587278247,-0.3951895535,-0.4038072228,-0.270919174,0.4613659382,-0.2893258929,-0.3082677424,0.0806009918,0.033760827,0.0050714198,0.0675616935,0.1498345584,0.0158942174,0.5080648065,0.1054995805,-0.0446049161,-0.1713000238,-0.2480780333,0.1482364982,0.032927461,-0.2041103691,0.3403045237,0.3946774006,-0.0350885876,-0.1126644388,0.0304206461,0.6076420546,0.3532488644,-0.6928082108,0.0780414939,0.1227816269,0.1445378661,0.0442241393,0.1437959969,0.2057185471,0.3289967179,0.098851338,-0.0218014382,-0.0685778633,0.0414157324,0.1923013777,0.2604732513,0.0378937386,0.1881123781,-0.0340932682,-0.2330759019,0.0635789558,0.043609228,-0.6348347664,-0.0292900242,0.4114500284,-0.49036026,0.0888524652,0.0827015787,0.0420090035,-0.0604564399,0.3990845382,0.0115333917,0.301609695,-0.2673699558,-0.1688440442,-0.5573812723,0.0129775712,0.1003673673,-0.0201993529,-0.182935372,-0.1710310429,-0.0433934666,0.2719523311,0.1712253243,0.0828569829,0.1784492433,0.7048636079,-0.6334088445,0.3171179295,0.2510333657,0.1043806076,0.2963539362,-0.3799831569,0.1781571656,0.0762072951,-0.0277444124,-0.2803596854,0.196539104,-0.1178417429,0.3687140048,0.2799830139,0.0656691864,0.3750950694,-0.0018231323,-0.0076421876,-0.0902410373,0.3053018153,0.1900222152,-0.4532952309,0.2929504514,0.1708452106,-0.5065389872,0.0706262663,0.0487745143,-0.0507214218,0.1662175804,-0.0040193107,-0.2534012198,0.0329207145,0.0087969499,0.0715351924,0.072966598,0.4877997935,-0.0563627779,0.2883818448,-0.2061873078,-0.1364272535,0.156664893,-0.2487031817,-0.3022477925,-0.3512842953,0.4295423627,-0.3348473012,0.1113521382,-0.2448457628,-0.0816367418,0.0238032714,0.0570960939,-0.6880539656,0.3799354732,-0.30705598,-0.6153408289,0.1081439406,-0.0409421287,0.0843589082,-0.158270359,0.1082268283,-0.334759742]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2067","title":"Multiprocessing windows error","comments":"\r\n\r\n\r\n\r\n\r\nI was able to copy some of the shell \r\nThis is repeating every half second\r\nWin 10, Anaconda with python 3.8, datasets installed from main branche\r\n```\r\n\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 287, in _fixup_main_from_path\r\n _check_not_importing_main()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 116, in spawn_main\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 134, in _check_not_importing_main\r\n main_content = runpy.run_path(main_path,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 265, in run_path\r\n exitcode = _main(fd, parent_sentinel)\r\n raise RuntimeError('''\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 125, in _main\r\nRuntimeError:\r\n An attempt has been made to start a new process before the\r\n current process has finished its bootstrapping phase.\r\n\r\n This probably means that you are not using fork to start your\r\n child processes and you have forgotten to use the proper idiom\r\n in the main module:\r\n\r\n if __name__ == '__main__':\r\n freeze_support()\r\n ...\r\n\r\n The \"freeze_support()\" line can be omitted if the program\r\n is not going to be frozen to produce an executable. return _run_module_code(code, init_globals, run_name,\r\n prepare(preparation_data)\r\n\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 97, in _run_module_code\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 236, in prepare\r\n _run_code(code, mod_globals, init_globals,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 87, in _run_code\r\n _fixup_main_from_path(data['init_main_from_path'])\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 287, in _fixup_main_from_path\r\n exec(code, run_globals)\r\n File \"F:\\Codes\\Python Apps\\asr\\test.py\", line 6, in \r\n updated_dataset = dataset.map(lambda example: {'sentence1': 'My sentence: ' + example['sentence1']}, num_proc=4)\r\n main_content = runpy.run_path(main_path,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 1370, in map\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 265, in run_path\r\n with Pool(num_proc, initargs=(RLock(),), initializer=tqdm.set_lock) as pool:\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\context.py\", line 119, in Pool\r\n return _run_module_code(code, init_globals, run_name,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 97, in _run_module_code\r\n _run_code(code, mod_globals, init_globals,\r\n return Pool(processes, initializer, initargs, maxtasksperchild,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 87, in _run_code\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 212, in __init__\r\n exec(code, run_globals)\r\n File \"F:\\Codes\\Python Apps\\asr\\test.py\", line 6, in \r\n self._repopulate_pool()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 303, in _repopulate_pool\r\n updated_dataset = dataset.map(lambda example: {'sentence1': 'My sentence: ' + example['sentence1']}, num_proc=4)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 1370, in map\r\n return self._repopulate_pool_static(self._ctx, self.Process,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 326, in _repopulate_pool_static\r\n with Pool(num_proc, initargs=(RLock(),), initializer=tqdm.set_lock) as pool:\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\context.py\", line 119, in Pool\r\n w.start()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\process.py\", line 121, in start\r\n return Pool(processes, initializer, initargs, maxtasksperchild,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 212, in __init__\r\n self._popen = self._Popen(self)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\context.py\", line 327, in _Popen\r\n self._repopulate_pool()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 303, in _repopulate_pool\r\n return Popen(process_obj)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\popen_spawn_win32.py\", line 45, in __init__\r\n return self._repopulate_pool_static(self._ctx, self.Process,\r\n prep_data = spawn.get_preparation_data(process_obj._name)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 326, in _repopulate_pool_static\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 154, in get_preparation_data\r\n _check_not_importing_main()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 134, in _check_not_importing_main\r\n w.start()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\process.py\", line 121, in start\r\n raise RuntimeError('''\r\nRuntimeError:\r\n An attempt has been made to start a new process before the\r\n current process has finished its bootstrapping phase.\r\n\r\n This probably means that you are not using fork to start your\r\n child processes and you have forgotten to use the proper idiom\r\n in the main module:\r\n\r\n if __name__ == '__main__':\r\n freeze_support()\r\n ...\r\n```","body":"As described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop ","comment_length":433,"text":"Multiprocessing windows error\nAs described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop \n\r\n\r\n\r\n\r\n\r\nI was able to copy some of the shell \r\nThis is repeating every half second\r\nWin 10, Anaconda with python 3.8, datasets installed from main branche\r\n```\r\n\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 287, in _fixup_main_from_path\r\n _check_not_importing_main()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 116, in spawn_main\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 134, in _check_not_importing_main\r\n main_content = runpy.run_path(main_path,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 265, in run_path\r\n exitcode = _main(fd, parent_sentinel)\r\n raise RuntimeError('''\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 125, in _main\r\nRuntimeError:\r\n An attempt has been made to start a new process before the\r\n current process has finished its bootstrapping phase.\r\n\r\n This probably means that you are not using fork to start your\r\n child processes and you have forgotten to use the proper idiom\r\n in the main module:\r\n\r\n if __name__ == '__main__':\r\n freeze_support()\r\n ...\r\n\r\n The \"freeze_support()\" line can be omitted if the program\r\n is not going to be frozen to produce an executable. return _run_module_code(code, init_globals, run_name,\r\n prepare(preparation_data)\r\n\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 97, in _run_module_code\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 236, in prepare\r\n _run_code(code, mod_globals, init_globals,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 87, in _run_code\r\n _fixup_main_from_path(data['init_main_from_path'])\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 287, in _fixup_main_from_path\r\n exec(code, run_globals)\r\n File \"F:\\Codes\\Python Apps\\asr\\test.py\", line 6, in \r\n updated_dataset = dataset.map(lambda example: {'sentence1': 'My sentence: ' + example['sentence1']}, num_proc=4)\r\n main_content = runpy.run_path(main_path,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 1370, in map\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 265, in run_path\r\n with Pool(num_proc, initargs=(RLock(),), initializer=tqdm.set_lock) as pool:\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\context.py\", line 119, in Pool\r\n return _run_module_code(code, init_globals, run_name,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 97, in _run_module_code\r\n _run_code(code, mod_globals, init_globals,\r\n return Pool(processes, initializer, initargs, maxtasksperchild,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 87, in _run_code\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 212, in __init__\r\n exec(code, run_globals)\r\n File \"F:\\Codes\\Python Apps\\asr\\test.py\", line 6, in \r\n self._repopulate_pool()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 303, in _repopulate_pool\r\n updated_dataset = dataset.map(lambda example: {'sentence1': 'My sentence: ' + example['sentence1']}, num_proc=4)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 1370, in map\r\n return self._repopulate_pool_static(self._ctx, self.Process,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 326, in _repopulate_pool_static\r\n with Pool(num_proc, initargs=(RLock(),), initializer=tqdm.set_lock) as pool:\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\context.py\", line 119, in Pool\r\n w.start()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\process.py\", line 121, in start\r\n return Pool(processes, initializer, initargs, maxtasksperchild,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 212, in __init__\r\n self._popen = self._Popen(self)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\context.py\", line 327, in _Popen\r\n self._repopulate_pool()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 303, in _repopulate_pool\r\n return Popen(process_obj)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\popen_spawn_win32.py\", line 45, in __init__\r\n return self._repopulate_pool_static(self._ctx, self.Process,\r\n prep_data = spawn.get_preparation_data(process_obj._name)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\pool.py\", line 326, in _repopulate_pool_static\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 154, in get_preparation_data\r\n _check_not_importing_main()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 134, in _check_not_importing_main\r\n w.start()\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\process.py\", line 121, in start\r\n raise RuntimeError('''\r\nRuntimeError:\r\n An attempt has been made to start a new process before the\r\n current process has finished its bootstrapping phase.\r\n\r\n This probably means that you are not using fork to start your\r\n child processes and you have forgotten to use the proper idiom\r\n in the main module:\r\n\r\n if __name__ == '__main__':\r\n freeze_support()\r\n ...\r\n```","embeddings":[-0.1881467849,-0.3144050539,-0.0391355641,0.1985245347,0.1112000048,0.004422768,0.1473799646,0.0522424132,0.007206195,0.1542068422,0.0380617343,-0.0369985774,-0.2301331311,0.0049480656,-0.1641259491,-0.2017735839,0.1229714751,-0.1477968991,0.108270444,0.3176008463,-0.4681925774,0.4709626138,-0.2818864286,0.1857171208,-0.198204428,-0.0535663292,-0.2812374532,0.5518501401,0.2289519757,-0.5014718771,0.1033398733,-0.3695967793,0.2053667456,0.192868039,-0.0001191628,-0.0957157239,0.1240097806,0.0172286984,0.2794227004,0.2258308828,0.2599862516,0.0434033126,0.0091181993,-0.022887405,0.2111760527,0.2333109677,0.0288525671,-0.1462767869,0.475712508,0.1996723413,0.0683800951,0.1587361097,0.1223477423,-0.0179201271,-0.3447393179,-0.2424110025,-0.0074872836,0.2551131546,0.4874388874,-0.3183261156,0.2209866792,0.1477992982,-0.2789092064,0.0836960822,-0.4995912611,0.0385394022,0.6404622197,-0.5573272109,0.0123135289,0.0039335908,-0.1371411383,-0.1722593457,0.1412511766,-0.0714389086,0.2048701644,-0.8496664762,-0.0566585362,0.2238097042,-0.4533593059,0.0700570643,0.0376135893,0.1382519752,0.0211229026,-0.1229700223,0.1787132472,0.0480568931,0.0241068453,0.2137415856,0.2611123621,0.2568986416,-0.2901437283,0.3060054183,-0.1656566411,0.1110760197,-0.559320569,0.0435666554,-0.0875910893,0.1395756155,-0.26038903,-0.0911267251,-0.1450668722,-0.0010928861,0.2437376082,0.1224617213,0.16437985,-0.2249252349,0.016275337,-0.5754626393,0.0762996003,-0.3193160892,-0.0192709044,-0.0631263629,-0.0396311693,-0.5433811545,0.1638689786,0.4754507244,0.3127381802,-0.0908295736,-0.4087652862,0.0046493872,0.054897096,0.1591065675,0.3263029754,0.3785496056,0.2375260741,0.649328053,0.116466403,0.3008677363,-0.4983567894,0.2498873323,0.105526261,-0.009915052,-0.2646825314,0.2840289176,0.201582551,-0.1050174385,0.1919124722,0.0913679674,0.2648686171,-0.1933916658,-0.0548236333,-0.1775243133,-0.1875754297,0.2483474612,0.1039774194,0.2727549374,0.0782229826,0.0614007749,-0.0588395223,0.1217043996,-0.2332015634,-0.3323019743,-0.2908734977,0.0487952456,0.3455986083,0.4107144475,0.3590714931,-0.1826985478,0.3697884083,-0.1141108572,-0.1149107367,-0.3369792104,-0.3605934381,-0.2536899149,0.0559893548,0.3577569723,-0.0117116673,0.2625139952,0.1336178035,-0.023401374,0.1730084717,0.2555250823,0.1529283077,-0.0441390127,-0.1599845588,-0.1845404953,-0.2531806231,-0.4692437053,0.2813512981,-0.0531997643,-0.0947239995,0.0144146662,0.2635256648,-0.0924054384,0.0841419771,-0.3598734736,0.4103763103,-0.1397632509,-0.0000294809,0.156400606,-0.2134256065,-0.0189389195,-0.2104984671,0.1580802798,-0.1415789723,-0.1717151105,-0.1469049156,-0.0887722746,0.2672272325,0.13230443,0.002933617,0.2626314759,0.3290075064,-0.0504727885,0.070511654,-0.2354250848,-0.1410974711,0.1449600756,-0.0926912799,0.0614887998,0.2999998927,0.2150601894,0.184471339,0.2884706557,-0.2682780325,-0.3507547081,0.0538341627,-0.0101501597,0.1759853065,-0.1773452163,0.013145634,0.5986759067,-0.1233620197,0.0136165908,-0.2358869612,-0.1150441468,-0.3619886339,-0.2789753377,-0.010339723,0.3257572055,0.1760945618,0.0544384234,0.1365503967,0.2462071925,0.1490508914,-0.1596979201,-0.1565377712,-0.1665803939,0.1694262624,-0.2053367943,0.1545628011,0.0625823438,0.0926790237,-0.2570307553,0.4300040007,0.0516269915,-0.0339030623,0.3756663203,-0.0890954062,0.1873241514,0.1525298506,-0.0126634585,0.2188524008,0.1479418576,0.3459525406,0.1692016423,0.1437898129,-0.4511528611,-0.1519183069,0.1961424947,0.2103529871,0.0790235922,-0.1222661883,0.1141883358,-0.1815057397,0.1832985282,-0.0153749026,-0.1870618463,0.7362310886,0.0474283323,-0.1104163975,0.2805086076,-0.4284941554,-0.0939547271,0.3900580108,0.0612931587,0.2331193388,0.1693288684,-0.1249577552,-0.0459342524,-0.0291668698,-0.2392556816,-0.0362743028,0.0800577775,-0.3402797282,0.061263185,-0.0924087912,0.4969219267,-0.4114156961,-0.4017821848,0.0227532499,-0.1363874376,-0.0535736457,0.4672952294,-0.1314401478,0.0236187,-0.0362695865,-0.0113865444,0.0388440676,0.3347727656,0.156779781,-0.0217868797,-0.0881472006,-0.0752070248,0.3987684548,-0.0621618815,0.2075349689,0.2854368985,-0.0574674681,0.2356499135,0.0421252362,-0.0356818102,0.0262664538,0.3805266321,0.0301337205,0.1732631922,0.1848597229,-0.5782219172,0.1626747847,-0.163552314,0.1181733534,0.1189880297,0.0275810603,-0.4057868719,-0.1938906163,-0.0397473946,-0.4883069992,-0.296564281,0.0322543196,-0.0325806886,0.1934853643,-0.0064605447,0.1003118232,0.1495274305,0.0736577809,-0.1381520033,-0.1220574453,0.0768527016,-0.0614772812,-0.2239359319,-0.048726093,0.0344485268,0.0997959003,0.0538104735,-0.2458163649,-0.1084734052,-0.2988756597,-0.1147123873,0.0434826389,0.0423186161,0.21405904,0.4646483362,0.245115608,-0.1603371054,0.11727079,0.0354307257,0.1138501689,0.0141474232,0.1571184695,0.1290081888,0.0388406776,-0.2746428251,0.5560711026,0.7195617557,0.0079728663,0.2117627263,0.1725878119,0.1281331182,-0.0888722166,-0.2836923003,0.0189995691,-0.3320361376,-0.047518421,0.1152629405,0.1190858036,-0.1943729371,-0.2258570492,0.2984496653,-0.360991925,-0.2144330293,0.0913126245,-0.0031267332,0.106849283,-0.2871644199,-0.0014671338,0.0885587484,0.0902059898,-0.1638025045,0.2348212898,-0.2531650364,-0.3102205992,-0.2806431949,-0.4974411428,-0.4011013508,0.3255932927,0.1121691391,0.5273079276,0.0384407453,-0.1809523255,-0.0801848695,-0.2546062768,0.9073628783,-0.2285796106,-0.2365725338,0.2589763999,0.3240393102,-0.0839830115,-0.1596128047,-0.1857515574,0.2760500014,0.1724248528,0.003794166,0.1194249615,-0.187232241,0.1222395897,0.0239307825,-0.1323913187,-0.0503413714,-0.3534944654,-0.3537780344,-0.1054334044,0.1952892393,-0.0393823162,0.0809376985,-0.545144558,-0.0410296768,0.1071589142,-0.0651645213,0.1058607921,0.0425302684,0.0935709625,-0.0367010869,0.418415904,-0.2705120444,0.2269545048,-0.1325387061,0.1020635366,-0.2310526073,-0.0111666787,0.1763537526,-0.059362825,0.3624847233,0.4911215603,0.1173728555,-0.0381036289,-0.3317421675,0.4632002115,-0.2424946874,-0.2325073034,0.3120750189,0.2526546121,0.0963162929,0.303753823,0.2620723248,0.0120330229,-0.1321219653,0.6690809727,-0.2865629494,-0.1309688836,0.091562368,0.2670032978,0.9803853631,0.0002797422,0.1036657766,0.1353646219,-0.3450932205,0.0168711692,0.1485295445,0.2275093645,-0.214737758,-0.0921183079,0.1534200162,-0.0692243129,0.2798226476,-0.1486623883,-0.3306346238,0.277828753,0.1979100704,-0.298365593,-0.1088461354,-0.1559790224,-0.0305137765,-0.38702178,0.1118133217,0.0050123925,0.0803058594,-0.0226951838,-0.2418227047,0.1473931074,0.4167760909,-0.3222223222,-0.5853480697,0.0799791068,-0.2500568926,0.0987987295,-0.106463246,0.0789994225,-0.0206889529,0.2738597095,0.2924273014,-0.158672601,-0.1383257955,0.0596915483,0.317700386,-0.1067116559,-0.0287849829,-0.18386136,-0.0780223906,-0.0452293828,-0.3501313925,0.2171633989,-0.1274762899,0.0402390324,-0.236063838,0.3128050268,-0.0060308911,0.2374044955,-0.1692706496,0.3972187936,-0.12809439,-0.1625915915,0.0792383999,-0.0392376333,-0.4698782563,0.5655829906,-0.6352729797,-0.1246339902,0.0684989542,0.043192666,0.0512160771,0.0985023677,0.3221236467,0.2966786921,-0.0124750556,-0.210961774,-0.1122625917,-0.519885242,-0.1286488771,-0.1232335195,0.417942524,-0.1153107733,0.2454816997,-0.3513736725,-0.0550738499,-0.0755419135,-0.0676290914,-0.4756988883,-0.0865643919,0.0446376242,-0.2182350755,0.0746468604,-0.1440294087,0.200020358,0.0788339972,0.1162854731,-0.245302394,-0.0192251317,-0.1059378535,0.1291177571,0.065760307,-0.2747898698,0.0525428653,0.1484904438,0.0116025601,0.2926319242,-0.2524476349,-0.1338891387,0.0612612292,0.0983059555,-0.1529425681,-0.1397013813,0.202444002,0.0376425125,0.0320359059,-0.0035297316,0.0384985618,-0.0282254312,-0.1952826232,0.0285112523,-0.0443940908,0.3265428245,0.1590437144,0.2606514096,-0.1510521621,0.3736740053,0.1364553124,0.0730975792,0.0131398365,-0.0461714454,-0.1199251711,-0.0568601601,-0.1078415141,-0.0480936877,-0.0949113593,0.0458896607,-0.3931430876,-0.3150415421,0.1322902441,0.5207763314,-0.0906363726,0.2684803307,0.3860579431,0.1517119557,-0.1998000443,0.1071867421,-0.0923590884,-0.2328294218,-0.1368254274,0.3245009482,0.0644398555,-0.5508656502,0.4770566225,0.0695959553,0.3834375441,-0.3326542974,-0.1584963053,0.3198783994,0.3208474219,0.0378487855,-0.0150568495,-0.09156394,0.2071587294,0.4593250453,-0.1311327368,0.2014142722,-0.3238770068,0.1578860134,-0.0896207094,-0.2398294806,-0.110228844,0.1657060832,-0.1913155615,0.293651104,-0.1899563223,0.4689161181,-0.5960272551,-0.3746602535,-0.3592116237,0.3930900693,-0.1336373836,-0.1862935573,-0.107536532,0.088694796,0.1282711625,0.0879685208,0.3019336462,-0.1085814536,0.6149493456,0.1113223806,-0.219085753,-0.1736536473,-0.3851723671,0.2245020717,-0.0093772998,-0.1311016083,0.3541278541,0.6124849916,-0.0654933304,-0.2977392375,0.1973142773,0.5848490596,0.4469044507,-0.6962715387,0.1449216008,0.1227997988,0.1126574054,-0.0131667918,0.1467432231,0.1020413935,0.1914798468,0.040567603,0.025395263,-0.0716055855,-0.1015080661,0.3518776894,-0.0269626696,0.1627119035,0.1804663092,-0.1023851037,-0.1577706635,0.1136825308,0.1008244082,-0.6489607096,0.1941168755,0.2340757549,-0.5234286189,0.0135804275,0.1275424659,0.048893448,-0.1051836312,0.44099316,-0.0385627784,0.2765523493,-0.204859376,-0.2142634541,-0.3133778274,-0.1079993621,-0.1254138798,-0.0537104644,-0.1895312518,-0.0621597581,0.0386297591,0.2744994462,0.0273522492,0.1042200327,0.225425899,0.6995902061,-0.6562343836,0.1831998378,0.3876357377,0.006753027,0.2149626911,-0.4983593225,0.1673810482,0.1196631193,-0.002962061,-0.0181070883,0.2306739539,-0.1568345875,0.1118390486,0.2074946463,0.1057576835,0.0867726505,0.0140857166,-0.2020221055,-0.1171711832,0.2525694966,0.2592410445,-0.322548151,0.3469716012,0.0999834016,-0.4285967648,-0.1827362627,-0.0377777293,-0.0870675743,0.1813717186,0.1660645306,-0.1305914819,0.0330276415,0.0607886314,0.1322130114,0.0492528304,0.3977868855,-0.0349069163,0.2151015252,-0.2833608091,-0.045622889,0.1644546092,-0.1422145963,-0.2171929181,-0.2530259192,0.3797122836,-0.4996713996,0.2078316361,-0.2136391401,0.0040057488,-0.0160127748,-0.0012642838,-0.5598614216,0.3290337324,-0.1534708589,-0.711019814,0.0667739809,-0.0996058211,0.1139259487,-0.1034514755,0.2528049052,-0.0494636036]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2067","title":"Multiprocessing windows error","comments":"Thanks this is really helpful !\r\nI'll try to reproduce on my side and come back to you","body":"As described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop ","comment_length":18,"text":"Multiprocessing windows error\nAs described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop \nThanks this is really helpful !\r\nI'll try to reproduce on my side and come back to you","embeddings":[-0.1506284475,-0.4665443301,-0.085092485,0.231307283,0.0865275562,0.0101603009,0.0336137787,-0.0307324585,0.0230884105,0.2842899561,0.1215858534,0.0667377338,-0.3062893748,0.0160491895,-0.1941297948,-0.1727220565,0.0984337926,-0.1465362012,0.2517451346,0.3242737353,-0.4602844119,0.486584723,-0.3101291656,0.1724273115,-0.4106292725,-0.0460810959,-0.2003931105,0.4842282832,0.2549553514,-0.3877053261,-0.0000801438,-0.3113578856,0.1509665549,0.1404424161,-0.0001155468,-0.1447120905,0.0140570253,-0.0155945476,0.3914166689,0.2356814146,0.1747708768,0.0154116098,-0.0930722207,-0.027862424,0.132762447,0.2751011848,0.0475338399,0.0036513577,0.403698951,0.3019194007,0.1590741128,0.2588334978,0.1020681933,0.0435885414,-0.3204174638,-0.389621377,-0.0509017669,0.2563922703,0.5162356496,-0.2866722345,0.1748640537,0.2624700963,-0.2407959998,0.0373887308,-0.316942066,-0.0094364276,0.431633532,-0.5044822693,0.0530499406,-0.0857313499,-0.0860016197,-0.1723192185,0.3172115386,0.1498873085,0.1057220623,-0.8734208941,0.0373722538,0.2236952931,-0.3608475327,0.0636084974,-0.0768172219,0.000486433,-0.0733843148,-0.2057019621,0.106475234,-0.0420243628,0.0588759296,0.162166357,0.1567879617,0.2581148148,-0.2660805583,0.3123065829,-0.0895460024,0.0824717656,-0.5067335367,-0.0130569059,0.0244877208,0.2107514292,-0.281976223,-0.2238519341,-0.0398795083,0.1608051062,0.2515883148,0.181238845,0.2087535411,-0.1637963802,0.0577949211,-0.5973331332,0.2277550101,-0.4697573483,-0.1418823451,-0.0908495039,0.0590582006,-0.5761107206,0.3017511666,0.4158447385,0.1878656745,-0.1029671282,-0.3051148355,0.2279174775,0.0387915261,0.244416669,0.2676872015,0.3037410676,0.3295402229,0.67036587,0.0046987813,0.1355355531,-0.5587133169,0.1353852004,0.0535593033,0.0962220579,-0.3082456291,0.1239422709,0.2694671154,0.0077629364,0.164147377,0.1423988342,0.2745949924,-0.3172977865,-0.0431899056,-0.2245990038,-0.1705032289,0.2402047515,0.2170298845,0.1954387873,-0.0491465889,0.0771684945,0.0033321951,0.2177007794,-0.3714387119,-0.3710877895,-0.0780069754,0.0910405517,0.4564350545,0.4124155641,0.2775190175,-0.0414841846,0.2997341156,-0.2898950875,-0.1395444274,-0.4417795241,-0.2753013372,-0.3028675616,-0.0770642534,0.2549555302,-0.013891736,0.3129976094,0.3275535107,-0.0539206825,0.1966925859,0.3646476269,0.2357976139,-0.2295782119,-0.076528579,-0.01541859,-0.1692055315,-0.2164342701,-0.0256465394,0.1248811483,-0.0502498113,-0.1582021564,0.2369344383,0.1396849602,-0.0421049297,-0.325414449,0.4733939469,-0.1873985678,-0.138039127,0.2417535782,-0.2508323193,0.1379858553,-0.0901914462,0.074640967,0.078596428,-0.1308210194,-0.0758600682,-0.0913820043,0.2532756925,0.1211822405,-0.0288552847,0.3372828662,0.1698221266,-0.0188710131,0.0452157408,-0.3747882247,0.0022795128,0.0269336868,-0.2805185616,0.0384882465,0.3604878485,0.1296922415,0.1333177537,0.2604677081,-0.1930642724,-0.2732436955,0.0716792122,-0.0973805264,0.0637126565,-0.201838389,-0.0118380021,0.4987510443,0.0421448089,-0.1541185826,-0.1427766979,-0.0749818385,-0.3986949027,-0.3253808618,-0.0215554293,0.3518841863,0.2070241123,0.1487410814,0.1229823679,0.2170087397,0.0977165103,-0.1447271258,-0.1434539706,-0.0267221518,0.2210915238,-0.2565096021,0.091993168,0.0154685685,0.1554853916,-0.2237295657,0.4834907055,0.0780260339,-0.0517863892,0.2152548432,-0.1356084943,0.1757023036,0.1661110371,0.0268721301,0.2828970253,0.2699865103,0.3601727188,0.1210275888,-0.0984187275,-0.3358142972,-0.3411460519,0.0787207037,0.2047636211,0.1856312752,-0.0494339094,0.0344514064,-0.1975595951,0.1480878592,-0.0296658818,-0.4081707597,0.6608490944,0.1290354431,-0.1238795221,0.2897005975,-0.416824162,-0.1286766082,0.2691907287,0.0189799685,-0.0485235676,0.1837819517,-0.1021908894,-0.1305869967,-0.0627082437,-0.0959941,-0.0835921988,0.0905502811,-0.3112027049,0.0754139945,-0.0683130026,0.4818834662,-0.3881742954,-0.4849456847,-0.0363850668,-0.201094985,0.0933721662,0.3769926727,-0.0660397336,-0.0369546078,0.063978903,-0.0753771067,0.1353933662,0.3412621617,0.2241446227,-0.0824958757,-0.0757005289,-0.0427121148,0.3543700874,-0.1073642969,0.2451773137,0.3000584543,-0.008039441,0.1237863973,0.1547088772,-0.0708740205,0.0497369952,0.254612118,-0.0004117277,0.1000178233,0.1898938566,-0.4014510214,0.0981199816,-0.3271220922,0.2163941264,0.2593770325,0.1389705986,-0.3894452155,-0.2654180527,-0.1531652808,-0.4650588632,-0.1933446229,0.0158451255,-0.1655590534,0.2833057344,-0.2524057925,0.0624522232,0.1090325117,0.3357043862,-0.1297424138,-0.1685179323,0.2200897932,-0.032310348,-0.3071311414,-0.2082278132,0.2030312866,0.2279906869,0.0628904179,-0.247068435,-0.1562498361,-0.3293305337,-0.0154413581,-0.018792307,-0.0382837914,0.2496671826,0.3877030313,0.1706475317,-0.1693151891,0.1369193047,-0.0877079442,0.03437078,0.0354771949,0.1656462699,0.2234891057,0.0942002535,-0.1743258387,0.4798260927,0.5712582469,0.0682216212,0.2840141654,0.1166627631,-0.0279757325,0.0197366383,-0.1986924112,0.104167968,-0.4029532671,-0.1182696894,0.1622435451,0.1221752763,-0.3131992519,-0.260248363,0.231244877,-0.2988224626,-0.1110460535,0.0893051028,0.0715152398,0.1665275246,-0.1911999136,-0.1086433157,0.0748671368,0.0724368021,-0.1262527108,0.1831806004,-0.2587577999,-0.2337905169,-0.0494499393,-0.5941668749,-0.4766453505,0.4101198316,0.1082497165,0.5226665139,-0.0082315989,-0.1149527654,-0.0176295657,-0.2964184284,0.8070997596,-0.1431352496,-0.1830292046,0.2115636915,0.2090340406,-0.0098497188,-0.1221254021,-0.2269454449,0.2996343076,0.1128154323,-0.0601762794,0.2513709366,-0.0667590052,0.0676047131,-0.1073215306,-0.1688712984,-0.1165623888,-0.3311615586,-0.288402617,0.0485692695,0.2625979483,-0.0126281455,0.0332488269,-0.4210146964,-0.0191498417,0.1351577491,-0.068568483,0.2947754264,-0.0726623908,0.117569305,-0.1421975642,0.142837286,-0.2732450068,0.3675431609,-0.1351262629,-0.0144852903,-0.1890772432,0.1965274364,0.2784040272,-0.110242106,0.3729200661,0.3814228773,0.1360358745,0.0135936886,-0.3704007566,0.4939796031,-0.2452186793,-0.2091410756,0.3325050175,0.2664666176,0.0814805329,0.2649816275,0.317045182,0.1260368228,-0.0922468975,0.4775740206,-0.3215054274,-0.0968962833,-0.0289775375,0.1434209794,0.9238151312,0.0846898034,0.1080940887,0.0120657329,-0.4905235469,-0.2325480133,-0.1703151762,0.3275733888,-0.2693823874,-0.1481327564,0.0304516591,-0.0382659659,0.1322237998,-0.1053469554,-0.3409645855,0.4729427695,0.0977984071,-0.291333586,-0.1794241816,-0.2355453819,-0.074017711,-0.1805983335,0.0021824662,0.0786175653,0.0783556402,-0.1448467672,-0.2679341435,0.1346343309,0.4491715431,-0.3302658796,-0.4855588675,-0.0171640571,-0.1990101188,0.1551013738,-0.0552148223,0.1623103768,0.0668271482,0.3810667098,0.1539157033,-0.0829822496,-0.2485067993,0.0191290118,0.3927862048,-0.0478609838,-0.1105954126,0.0463194773,-0.0188987199,-0.1503199339,-0.2896609306,0.3926055431,-0.1228491291,-0.0140360435,-0.3258322477,0.3156884015,0.0910597295,0.1713142395,-0.1278265417,0.2440666258,-0.124429509,-0.1677267402,0.1221420839,0.0498554111,-0.5294662118,0.652123034,-0.5616600513,-0.0932952687,0.0055157882,0.0684648231,0.0146985753,-0.0557597801,0.1593241394,0.2714974284,-0.1028343588,-0.2303350121,-0.1057639271,-0.4670142531,-0.0551307462,-0.0359919481,0.4547637701,-0.1424971819,0.2489058673,-0.2128921598,0.0583950803,0.0445281155,-0.0247205235,-0.3333129883,-0.225267306,-0.0753391832,-0.0825656354,0.1531943232,-0.0569255091,0.0920917913,0.1099377424,0.0790007412,-0.2915714383,0.0473889038,-0.2277822793,0.10021355,0.0793201253,-0.2663443983,-0.0975297913,0.2109725922,0.0640609041,0.3459066451,-0.3506193161,-0.1485283077,0.0790709406,0.0717254579,-0.0964205414,-0.0437678918,0.0798891485,0.0856898502,0.0290153567,0.075072512,0.1441150904,0.0900889933,-0.2854035497,0.1194539741,-0.050244987,0.2043404728,0.1899690777,0.2846988738,-0.1666853279,0.3932851255,0.1064694896,0.0063478216,0.0523516759,-0.0321252719,-0.0687173158,0.0561645441,-0.0845440403,-0.2393322438,-0.1015347168,-0.1130252853,-0.2827159464,-0.1089400202,0.2826990187,0.2960799336,-0.1647988111,0.2869280279,0.2995323539,0.2278229892,-0.2148747593,0.0480519943,-0.1478598118,-0.1248630732,-0.1360427588,0.2684840262,0.1823744327,-0.4664898217,0.3124746084,0.0706092045,0.1849167496,-0.4060252607,-0.3674868941,0.078256011,0.2264864743,0.1041754186,0.0287756957,-0.0863822252,0.0567869656,0.4952467382,-0.0511143506,0.1885724217,-0.2259835899,0.0775546953,-0.117027916,-0.1709330827,-0.2531660199,0.0921365693,-0.3694235682,0.4250567257,-0.2324751765,0.4461484253,-0.5816573501,-0.3129300475,-0.3135754168,0.4110392928,-0.2617023885,-0.1249559894,0.1012554616,0.0115746632,0.1034589112,0.0849343389,0.3099183142,0.0889204219,0.5263205767,0.2262043357,-0.0824888498,-0.0340430066,-0.2684048712,0.2271325588,-0.0195571985,-0.1974030733,0.1882003397,0.4959092438,-0.0470260084,-0.1721446067,-0.0020356663,0.5398030281,0.3772576451,-0.7378937602,0.0080808066,0.1601217687,0.1459166855,0.0435480773,0.1561661363,0.1998277903,0.2554309964,0.0483389869,0.0203681011,-0.1242482662,-0.0045728455,0.2641785741,0.224132821,0.1128834635,0.2446004003,-0.0508268215,-0.2304031104,0.0460793488,0.0731757581,-0.6401625276,-0.0241288673,0.2348130643,-0.3941693604,0.050793957,0.0399318524,0.066941902,-0.0753958449,0.410522908,-0.0549460016,0.2512814105,-0.2637573481,-0.1587159634,-0.3679072261,0.0130185122,0.1524792612,-0.0102243535,-0.193802923,-0.1669088453,-0.0689599589,0.1117754579,0.0863858014,0.0976389349,0.3191111982,0.6845180392,-0.6055827141,0.3746795058,0.3027550876,-0.0253777783,0.2710236907,-0.3583440781,0.2116037756,0.021200195,0.0636303872,-0.2255444676,0.2101653367,-0.1030671746,0.4734421074,0.1819057465,0.0485289134,0.2563311458,0.0007866362,-0.0625474975,0.1206356809,0.3517194688,0.2666041553,-0.5822997689,0.2764896154,0.0833062604,-0.3627561331,-0.0619934127,-0.0267348792,-0.0161337387,0.0667220429,0.1025722176,-0.1087251231,0.0047641029,-0.0050738282,0.0412030667,0.0413466468,0.4595139325,0.0127106905,0.2037229389,-0.2745131552,-0.0162326004,0.0923330188,-0.0616568774,-0.316155225,-0.3100215495,0.5059921741,-0.5154704452,0.1332807094,-0.0898828655,-0.0787975714,-0.0806523189,0.0168786887,-0.6610684395,0.2708574533,-0.3067221045,-0.6747819185,0.1371156424,-0.1630323231,0.1835549176,-0.0199975371,0.2053032666,-0.1706856489]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2067","title":"Multiprocessing windows error","comments":"if __name__ == '__main__':\r\n\r\n\r\nThis line before calling the map function stops the error but the script still repeats endless","body":"As described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop ","comment_length":20,"text":"Multiprocessing windows error\nAs described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop \nif __name__ == '__main__':\r\n\r\n\r\nThis line before calling the map function stops the error but the script still repeats endless","embeddings":[-0.1108912677,-0.4392870367,-0.0885387287,0.2090211362,0.0622631758,0.0388560519,0.0055700028,0.03431236,0.1305547059,0.3507117927,0.1845920682,0.1130338311,-0.2703493237,-0.0625594631,-0.1630701721,-0.0991831198,0.095834516,-0.0558623001,0.2845197916,0.3060237467,-0.5200814605,0.4720445573,-0.3290190399,0.1505489945,-0.3321788013,-0.0268693548,-0.1273716986,0.4326402247,0.277725637,-0.4478860795,-0.0330367759,-0.309070617,0.1182406545,0.1131975502,-0.0001193809,-0.1579200476,-0.0142006651,-0.0612700544,0.4552111626,0.1623017192,0.1796985269,0.0962326601,-0.1407280117,-0.1364838779,0.213292256,0.3244244158,0.118041724,-0.0536647364,0.4028441608,0.2655664682,0.1118752733,0.2055491209,0.0939192697,0.0754237399,-0.2695282996,-0.3726022243,-0.0564221442,0.2890223861,0.5033324361,-0.3820000887,0.1414164007,0.3143700957,-0.3030370176,0.0375135019,-0.2931313515,-0.064066574,0.5305408239,-0.5780696273,0.0976736173,-0.1105160043,-0.170950681,-0.2201832086,0.3771393299,0.1610389054,0.0507414714,-0.8099094629,0.0275954138,0.1364703476,-0.4156236649,0.0305882376,-0.1491321474,0.0135831144,-0.0268383697,-0.1467842609,0.0975081772,0.0018120224,0.0674111247,0.222169742,0.2512000203,0.2676575184,-0.3172388971,0.3002196252,-0.0685031116,0.116116181,-0.4423087239,-0.0357514955,0.0773455799,0.1887768209,-0.2333673686,-0.2691735029,0.0109956246,0.1386380941,0.2498514652,0.213896811,0.2139116228,-0.1912622303,0.1267627478,-0.6191260219,0.2186712474,-0.4347941577,-0.1045144573,-0.0480078869,0.0685980618,-0.5401844382,0.3463864326,0.4573223293,0.2797254324,-0.1079146639,-0.2896052599,0.295633167,0.0119032739,0.2019830793,0.257234782,0.3027789295,0.3510761857,0.6872230172,-0.0302588232,0.0749516934,-0.5382590294,0.1709015071,0.1013110653,0.1404786259,-0.3260935545,0.1511106044,0.2556617558,0.0663330182,0.1399945617,0.0715279505,0.2646604478,-0.3194342554,-0.1016467512,-0.2185526639,-0.0239290278,0.2193690985,0.1559507251,0.2279539108,-0.0514835007,0.075115785,-0.0280017797,0.2288478315,-0.3791945875,-0.3146420419,0.0056251031,0.0608852096,0.474270761,0.4800147712,0.3062348068,-0.0977224931,0.2995284498,-0.3837092221,-0.0615888424,-0.440713644,-0.3253913522,-0.3349105418,-0.1095616743,0.3342874646,0.0165768471,0.3336035907,0.3500510156,-0.0059903609,0.1059108749,0.2597663403,0.2196130902,-0.2469085008,-0.1390855163,-0.0079351738,-0.168292582,-0.1990410984,-0.0274671968,0.1558074504,-0.0895251781,-0.1395351291,0.2032585442,0.0898367316,0.0090118442,-0.3770282269,0.4743012786,-0.2610336542,-0.1601289511,0.2233112007,-0.1887081712,0.1628859192,-0.0973653793,-0.0018645076,0.047680404,-0.0739371479,-0.0460991897,-0.0943077579,0.2538553774,0.1462079287,-0.0373493657,0.2930855453,0.1426821798,-0.0440980606,0.0412538275,-0.3803648651,-0.0369581431,0.0419527739,-0.2586917281,-0.0318134427,0.2738764286,0.1100662202,0.1533838063,0.2823152542,-0.1785801053,-0.1833353639,0.0399044976,-0.080184482,0.057030566,-0.2264947146,-0.0417506993,0.370016247,0.0136441039,-0.1833051294,-0.1689893007,-0.1147308126,-0.3820264339,-0.3083172739,-0.0627655685,0.3502779007,0.2317863703,0.097002767,0.0890680775,0.234830305,0.1195787042,-0.147467643,-0.1531673819,0.0274254475,0.1904621869,-0.2588332593,0.0580979772,0.0713124573,0.184904173,-0.2433962226,0.4806243479,0.063830331,-0.0317932963,0.2943502963,-0.1405332088,0.1632416248,0.1101213619,-0.0063846633,0.2178630531,0.1491163671,0.3250453472,0.1516055465,-0.0632991567,-0.3150221705,-0.2845452726,0.0527499691,0.2794163823,0.1230376959,-0.0862115994,-0.0113490885,-0.1456184238,0.1935890317,-0.0329517536,-0.4224088192,0.6393153071,0.0958269164,-0.1360435188,0.2696197927,-0.4196719825,-0.1150873452,0.241999805,0.0387356803,0.0164453946,0.1887708604,-0.1815696061,-0.1324128062,-0.0871769264,-0.197066322,-0.0816427469,0.0346397236,-0.2595418394,0.0612711944,-0.0489790514,0.4968196452,-0.3601051569,-0.4657548964,-0.0427572653,-0.1918397695,0.0447609946,0.4225370288,-0.0910363048,0.0003684022,0.048364386,-0.0752312392,0.1253618598,0.2083099633,0.1887212098,-0.1300217807,-0.0176637694,-0.0903121531,0.4238033593,-0.2149709314,0.2365716547,0.2388014346,-0.036035765,0.0988385081,0.1630550921,-0.0131028984,0.024331525,0.1573209167,0.054088667,0.172326535,0.1641872972,-0.3335534334,0.094628416,-0.425177753,0.2039519548,0.249865666,0.1348346174,-0.4796755016,-0.3031045198,-0.1587507129,-0.3429673612,-0.2128161788,0.0812863335,-0.1794169247,0.3070078194,-0.3042918444,0.1231354997,0.0839877576,0.2612291873,-0.1630327404,-0.2192539573,0.2409317642,0.0028167977,-0.2471591383,-0.1566565037,0.2460523993,0.197476849,0.0351051688,-0.1808267981,-0.1518914253,-0.33172068,0.0068318718,0.0114257615,-0.008176974,0.2596694529,0.4391315281,0.1996572465,-0.1285470724,0.1151487827,-0.1072459146,0.0388023481,0.0945961848,0.1217528582,0.3355996013,0.1242963001,-0.1919824928,0.3964086771,0.585052073,0.0669824779,0.2981480956,0.0712781623,0.0723011494,-0.0084353928,-0.1640690714,0.0700911209,-0.3994195461,-0.1546176821,0.1536272615,0.0361060873,-0.3688913286,-0.2827152908,0.2706746459,-0.3177093267,-0.1678566784,0.1270991564,0.0690938532,0.130702585,-0.1740142703,-0.0349136926,0.059197586,0.1001402959,-0.1112741083,0.1810824126,-0.1693285853,-0.2402610779,-0.0678072199,-0.5501942039,-0.4681795835,0.4201438725,0.1505704373,0.5695855618,-0.0023044241,-0.1535202265,-0.0597583279,-0.2935432792,0.8496723771,-0.1981758922,-0.1934449077,0.2181835324,0.2658567727,-0.002650972,-0.0798922032,-0.1824527383,0.3223561645,0.1083092093,-0.0536811054,0.2510607839,-0.0756942555,0.1179926097,-0.0927480459,-0.1728261262,-0.0914245397,-0.3367777765,-0.2960079312,-0.0545704961,0.2718888521,0.0257831011,-0.0629815087,-0.3283909261,0.0095566502,0.2025770247,-0.101072371,0.3030549586,-0.01099269,0.144242987,-0.1002156511,0.0600254349,-0.2607697845,0.2963283956,-0.1633312702,-0.0030007556,-0.1922780871,0.2404852957,0.3096624613,-0.0918825567,0.3929225504,0.4457353354,0.1424077302,-0.1233945787,-0.3963153958,0.4940875471,-0.2796323299,-0.219122231,0.2729912102,0.2718494833,0.0476513058,0.2530600429,0.2464250922,0.1082381085,-0.1077957749,0.4929204881,-0.3655336201,-0.0749297887,0.035628207,0.1304314286,0.9967202544,0.0433197282,0.1189117059,-0.059444394,-0.4330742359,-0.2991344929,-0.1547603458,0.3376123607,-0.2475492358,-0.1441724747,-0.0107802087,-0.0327461436,0.1365705281,-0.0481979027,-0.3108739853,0.5423759818,0.2191966623,-0.1552074403,-0.1597220898,-0.244906351,-0.0447623506,-0.2088465989,0.12534751,0.0260196123,0.0824686661,-0.1033287346,-0.2708802819,0.1608608812,0.4782745838,-0.4204173982,-0.447070688,0.0222201608,-0.2924539149,0.1517005116,-0.1510333568,0.2836556435,0.0082212277,0.3933490813,0.0862138867,-0.0339075848,-0.2003100961,-0.051248949,0.5149103403,-0.0734036788,-0.1548949629,0.0469514765,-0.0645003021,-0.0882798135,-0.2245568931,0.3278431594,-0.1464331746,-0.0167198945,-0.4602338076,0.3510598838,0.1427760422,0.2157928646,-0.1538727582,0.2385237515,-0.2042168379,-0.1341703087,0.087611109,-0.0520025901,-0.455840826,0.6648210287,-0.5725058317,-0.1652441472,0.0157149062,0.0385558493,0.0256590657,-0.0823846683,0.0913122818,0.1807395071,-0.0798329115,-0.2497154623,-0.1001928598,-0.4476292133,-0.0590607449,-0.014096587,0.4131954908,-0.1556125283,0.1912052631,-0.2220641971,0.0293213017,0.1280600876,-0.0034407987,-0.3608266413,-0.2059660405,0.0041356138,-0.0384432748,0.1840740293,0.0003697115,0.0903396606,0.1145679727,0.0433794968,-0.2699514925,0.0237757862,-0.20484595,0.152005896,0.0822428465,-0.2354890704,-0.1031664759,0.1423013806,0.0385233946,0.4069353342,-0.4111886322,-0.1057432592,0.1133376583,0.0924327224,-0.0820116773,0.0198561158,0.0701738372,0.0386818238,0.0940268412,-0.0306103397,0.1176602393,0.0848436281,-0.29912588,0.0746333599,-0.028800156,0.2066002935,0.1639944166,0.2659658492,-0.1864542216,0.4394560158,0.0981494039,0.0235138945,0.0044680219,-0.0100964615,-0.0554418936,0.0710715204,-0.0920922086,-0.242591396,-0.1049908325,-0.0773674324,-0.3147372901,-0.1228149757,0.2354398519,0.3333853483,-0.2171507031,0.2746225595,0.3043156266,0.1782012433,-0.2159332633,0.0613208115,-0.0242810436,-0.1104072258,-0.070286721,0.3360287845,0.2171678096,-0.5155625939,0.3308371007,0.0507117994,0.181134969,-0.4108891785,-0.3599187136,0.1516511291,0.2669435143,0.1257068664,0.0498971269,-0.031967245,0.0922816545,0.4622264206,-0.0737492144,0.2301367521,-0.2403811216,0.147182703,-0.098857142,-0.1708921045,-0.1792935878,0.1233836561,-0.3252989948,0.494176507,-0.1151761934,0.3551940024,-0.5752065778,-0.3193903267,-0.3000149429,0.3833559155,-0.2761049569,-0.0942611396,0.0731210932,-0.0046292543,0.0848834217,0.1095994711,0.3375642002,0.073058188,0.6472619176,0.2219406813,-0.096556142,0.0616377443,-0.3227698505,0.1396819502,0.0483468622,-0.2488616854,0.22720927,0.4413455427,-0.0150429178,-0.2559155524,-0.0786895007,0.5184316039,0.3973650634,-0.7148273587,0.0073228483,0.1765239686,0.1348551363,0.0246888865,0.1171334013,0.1264953464,0.2712630928,0.0323041677,0.0082615595,-0.0902380347,-0.0176294614,0.314745754,0.2156029791,0.233762756,0.16078794,-0.0054430789,-0.2068796009,0.0349383242,0.1091857776,-0.6795328856,-0.0306232907,0.1634546965,-0.4358773828,0.0823000595,0.034672834,0.0564471968,-0.0317574181,0.3202776909,-0.060823068,0.2135938555,-0.2925104201,-0.1811573654,-0.3769399226,0.0130105373,0.1031439453,0.0137510756,-0.1198756993,-0.1581248641,-0.0381522961,0.1678033471,0.1059700325,0.0903670266,0.2510844767,0.7179806232,-0.540871799,0.3911240101,0.3814263046,0.0051036398,0.2969394326,-0.4028024077,0.2393359393,-0.0329811014,0.0491695888,-0.2540533841,0.2756895423,-0.1599727124,0.4374837577,0.2795375288,0.0273051076,0.1966376305,-0.096616134,-0.0949591324,0.1099543273,0.4139923155,0.300555259,-0.5189866424,0.3097030818,0.0198680628,-0.405382663,-0.096416831,-0.0510488749,-0.0667770207,0.060622789,0.1748028845,-0.0961037651,0.0391782224,-0.0059003429,0.1120398045,-0.0357935764,0.4521823525,0.0073086792,0.2061778158,-0.2073802203,-0.0599182062,0.1864174902,-0.0306250136,-0.3157809675,-0.2376834452,0.4764109254,-0.5554817915,0.0987056047,-0.1281593442,-0.061851833,-0.1075516939,0.0303570293,-0.7285946012,0.2602621317,-0.4314205348,-0.6321413517,0.1089848652,-0.0953818113,0.1943977922,-0.0924220458,0.2282528728,-0.196215421]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2067","title":"Multiprocessing windows error","comments":"Indeed you needed `if __name__ == '__main__'` since accoding to [this stackoverflow post](https:\/\/stackoverflow.com\/a\/18205006):\r\n\r\n> On Windows the subprocesses will import (i.e. execute) the main module at start. You need to insert an if __name__ == '__main__': guard in the main module to avoid creating subprocesses recursively.\r\n\r\nRegarding the hanging issue, can you try to update `dill` and `multiprocess` ?","body":"As described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop ","comment_length":59,"text":"Multiprocessing windows error\nAs described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop \nIndeed you needed `if __name__ == '__main__'` since accoding to [this stackoverflow post](https:\/\/stackoverflow.com\/a\/18205006):\r\n\r\n> On Windows the subprocesses will import (i.e. execute) the main module at start. You need to insert an if __name__ == '__main__': guard in the main module to avoid creating subprocesses recursively.\r\n\r\nRegarding the hanging issue, can you try to update `dill` and `multiprocess` ?","embeddings":[-0.1351817697,-0.4226135314,-0.1194850281,-0.0176143181,0.1080636531,-0.0960128754,0.1320640445,0.0308473576,-0.1013921574,0.3602203131,-0.0551886074,0.0946900994,-0.3736594021,-0.1421171576,-0.1579376906,-0.0196444057,0.0329828709,-0.1128270403,0.0478783175,0.0904548019,-0.3394441903,0.4024178982,-0.2713930011,0.0268041342,-0.401966542,0.0796891004,-0.0617421083,0.5093019009,0.019131951,-0.3639236689,-0.1021172404,-0.2195554376,0.2386392504,0.1321731061,-0.0001034106,-0.1818677485,0.0205885172,0.0046904827,0.1855369657,0.0788146481,0.362146616,-0.2269083709,0.0714554638,-0.0414677821,0.0445211679,0.2184216082,0.005528091,0.0762189701,0.4799880087,0.2293442637,0.263673991,0.0895196125,0.1886885166,0.1408014446,-0.1524417549,-0.2853358388,-0.1197071671,0.4174712598,0.3994277716,-0.1610610336,0.0792940259,0.2322094887,-0.1770654768,0.0859544873,-0.2321711183,-0.000966591,0.4042548835,-0.4059965312,0.0156400278,0.0012499293,-0.09211573,-0.1074041873,0.2277501673,0.1052797064,0.2289254218,-0.8124302626,0.0166804157,0.0472573861,-0.3250298202,0.080849044,0.1410831511,-0.1242963299,-0.0220248494,-0.2145822495,-0.0428274125,0.100537844,0.1382362247,0.0571483411,0.1464475691,0.0077999323,-0.2372009009,0.3869692683,-0.0115062576,0.085008584,-0.4147729576,0.077057384,0.1552436501,0.0051671388,-0.2343978584,-0.2477191985,-0.0266916826,0.1380921006,0.2967424393,0.1935619116,0.1310121119,-0.0973500684,0.1096473336,-0.5587246418,0.3567598462,-0.4095540345,-0.0658391863,0.1163014695,-0.0664674267,-0.4812475145,0.2111390829,0.3640954792,0.1387633979,-0.1785726696,-0.2412886322,0.2611695826,0.0761051923,0.1619967818,0.1748350412,0.3731553853,0.2703039944,0.6323959827,0.1211332902,0.0395691097,-0.5594881773,0.3015013039,0.0585801527,0.0765402988,-0.1999404132,-0.0367091857,0.2190437019,0.0480022505,0.1093272939,0.0806696862,0.1975702196,-0.1294086725,-0.0299188029,-0.1190906018,-0.0193335675,0.1967867166,0.1208981425,0.2237978429,-0.0420531817,0.0469095223,-0.1270228773,0.2816563249,-0.1704918593,-0.2169582993,-0.0966705903,0.2064185292,0.3183296323,0.3302809894,0.4069041312,0.0202212799,0.0787565857,-0.3521757424,-0.212965861,-0.4262780547,-0.1532217413,-0.2204384953,0.1979792714,0.2015244067,-0.2252460271,0.2377127111,0.1215576977,-0.1078297123,0.2043155879,0.254185617,0.2908347547,-0.2798776627,0.0065354672,-0.0297299959,-0.0913808048,-0.2077922672,-0.0343788154,0.0867681205,-0.0955197811,-0.1293618232,0.0654301494,0.1055273861,0.1556668133,-0.1546861231,0.3835849762,0.0252294131,0.062547259,0.1670588851,-0.3813206553,0.1540871114,-0.1082117036,0.1325490475,-0.0153022446,-0.3098196983,-0.1858509481,-0.0440183766,0.3126030564,0.1123026311,-0.0583191514,0.3177084625,0.2139800936,-0.0429484509,0.0330179296,-0.2787873745,-0.0473936722,-0.0326261669,-0.2836299241,0.1109803393,0.2854698896,0.1276585013,0.083592169,0.2665581107,-0.1575900018,-0.2924698889,0.2016206682,0.0617833026,0.0584755093,-0.0908382758,-0.1190522909,0.2270743549,0.1003558263,-0.1943213344,-0.1213511899,-0.1139202788,-0.4869252741,-0.2864394188,0.0616920441,0.404365927,0.0870122164,0.0715784803,0.0119792195,0.2805937529,0.1202455461,-0.1503616422,-0.0562342592,-0.1117319986,0.1633470953,-0.166598618,0.0865739584,0.1966273487,0.1060523614,-0.1940512657,0.3674074709,0.0948762819,0.1114334911,0.2885196805,-0.0115702366,0.165648222,0.1555183828,0.098912783,0.1268939972,0.0978220478,0.4008514881,0.1141287312,0.1021068692,-0.2333516181,-0.2867882848,0.1027944684,0.2684311271,0.0992106199,-0.0008464996,0.1202717423,-0.1331630945,0.0946435481,-0.0441561006,-0.4050086141,0.3301173449,0.1766394973,0.0809020922,0.2493042052,-0.4532828033,-0.1189093217,0.2047943473,-0.1069699898,-0.1498229504,0.2540160716,-0.1136460304,-0.1780838966,-0.1086232215,-0.1908424199,-0.1463781744,0.1059681475,-0.1225523576,-0.1124032363,-0.2060617059,0.5869614482,-0.3379522264,-0.5343268514,-0.2172742486,-0.2400305569,0.2002085447,0.295699954,-0.1551198959,0.1767859161,0.1703995466,-0.1477064937,0.1539740562,0.2446304411,0.1054776385,-0.196619302,-0.0911579654,0.0645607635,0.3895747662,0.0536355153,0.2719883621,0.2185650766,-0.0839993656,0.1159341112,0.0179363005,-0.1176390424,0.006587489,0.2527399063,0.0523972698,0.0870604962,0.2333325297,-0.4007569551,0.1690364182,-0.414524883,0.1989921033,0.2155409604,0.1596940607,-0.2973793745,-0.3184919655,-0.3417880833,-0.4226312339,-0.4237927794,-0.004604165,-0.1088418216,0.1892410517,-0.305377841,0.0938827246,0.1626902968,0.4764080644,0.1099185124,-0.109791182,0.3985269964,0.1149214655,-0.423284322,-0.3036082685,0.073973611,0.0862476006,0.0005403945,-0.3487333953,-0.0938106179,-0.2605395019,-0.0402700678,-0.0411944725,0.0168380681,0.1922750175,0.320517689,0.2199819982,-0.1974911094,0.1253982931,-0.1087216958,0.0102563053,0.0217464101,0.0642498359,0.1324173957,-0.0082423287,-0.3252888024,0.5187214613,0.2988110781,0.0430265293,0.2415017337,0.1495644003,0.0738498792,-0.1125651523,-0.1149780378,0.2159844488,-0.1812377572,-0.1233823076,0.3385454416,0.0637296513,-0.3440635204,-0.3852456212,0.3334489167,-0.2241055816,-0.2168640494,0.1037668586,-0.0125348,-0.0054077087,-0.0672693402,-0.0053002248,0.0592482574,0.1513289809,-0.0157176927,0.1620750576,-0.1981828511,-0.2261509895,-0.0853509158,-0.5560272336,-0.5261833072,0.3601529598,0.1612607092,0.4045396149,-0.0279571284,-0.3122996986,0.1104669496,-0.3153473139,0.6050417423,-0.1040820032,-0.2310165465,0.2957623303,0.1977659464,-0.1213179156,-0.1223640218,-0.2776659727,0.0166129284,0.2755773067,-0.2339741439,-0.0187754389,-0.0773268119,0.0408605859,-0.0962731615,-0.1876423806,-0.1230346709,-0.3188303709,-0.2852318585,-0.1298854351,0.1400084794,-0.0876075923,0.1171674952,-0.4547159672,0.0090689203,0.0878729746,-0.0814699456,0.1576920003,0.1193169579,0.1675173938,-0.1490399837,0.2297841609,-0.1533106118,0.5115940571,-0.0425256453,-0.0099611115,-0.2837653756,0.1625421494,0.4291081727,-0.1789736003,0.2091447711,0.3356902897,0.1679200828,0.0008668621,-0.4147988558,0.4918942153,-0.2210251242,-0.0365475044,0.3737960458,0.3164682984,0.1306276023,0.0721177161,0.2689923644,0.0503374673,-0.1190357134,0.2975538373,-0.3439263701,0.0228017885,0.0979520753,0.1122825816,0.9095650911,0.1273317635,0.0503468253,0.1770868748,-0.4576578736,-0.2464916706,-0.2452308536,0.2150414884,-0.345248878,-0.0076946872,0.1278997809,-0.016124852,-0.0364165977,-0.1383354366,-0.3275016546,0.331716001,0.0144618629,-0.2145612538,-0.2669146061,-0.0233464316,-0.0667419657,-0.1558232009,-0.004671664,0.1918946803,0.2317627668,-0.0981317461,-0.2949542105,0.0883867964,0.3090928197,-0.3657175601,-0.2679744065,-0.0044261878,-0.0495306887,0.1375202388,-0.1322584301,-0.126316607,-0.1111652777,0.408523649,0.0641725138,0.0352638029,-0.2122752815,0.0762815997,0.2285334617,0.0259711388,-0.0888744071,0.1009900272,0.0062169475,-0.1928398162,-0.2835327685,0.2746960223,-0.0933469608,0.0528204627,-0.1957861334,0.3317499459,0.11355903,0.1840255409,-0.0895460472,0.3004458845,-0.2343102992,-0.1687006205,0.2205449939,-0.0026314135,-0.5343681574,0.6597442627,-0.3583517671,-0.0428008288,-0.0323080458,0.048882965,0.0207638275,0.1505816728,0.177813977,0.2599483728,-0.2313389033,-0.2610792518,-0.1300228387,-0.3829298615,0.0228498038,-0.031339746,0.5056536794,-0.2653230429,0.1749011129,-0.2343427092,-0.0262949206,-0.0053460244,-0.1759109795,-0.1204579547,-0.1555927843,0.0213116352,-0.050384406,0.232901603,-0.1126276925,0.166522339,0.0070859473,0.1055919677,-0.4100174606,0.0122729111,-0.2944075763,0.0815689862,-0.0398198105,-0.4213510454,0.0431633443,0.2566620708,0.2082250267,0.3661239743,-0.306091398,-0.2721685469,-0.0066807396,0.0311002005,-0.1669248492,0.0260316133,0.0875767842,0.0500216968,-0.0431532189,-0.0394676998,0.0717340931,-0.0295923017,-0.1960064322,0.1399923265,-0.071232304,0.2636042535,0.3090511262,0.4009552002,-0.1266826242,0.4257573485,-0.0223604888,0.0120972041,0.0141727692,-0.0842870995,-0.042848371,0.0461578183,-0.0682964101,-0.1955296993,0.0145976506,0.1875316352,-0.2391766608,-0.1456568539,0.5542581677,0.4825503826,-0.048680719,0.1391605139,0.4199675322,0.3092742264,-0.2318922281,0.0846430808,-0.1809268445,-0.1913694888,0.093186453,0.1590421647,0.0616302937,-0.378490895,0.2029735148,0.1472109705,-0.0433647782,-0.2426796705,-0.5065719485,0.1888123304,0.1677510142,0.0570429377,0.0523792952,-0.107890442,0.1624528319,0.4125843644,-0.1772691756,0.2198757678,-0.3080129623,0.1113029942,-0.0615301542,-0.2300929129,-0.1026434824,0.1007719934,-0.2014252096,0.2718037963,-0.1607922465,0.4502228498,-0.5096836686,-0.3224596679,-0.2438452244,0.4145302474,-0.2001018524,-0.0969569385,0.0895558372,0.0270817094,0.0742648467,0.0635609701,0.2658009827,0.1540313512,0.4127882421,0.301164031,0.0663849339,-0.1835933924,-0.2612593472,-0.0124266474,0.045187667,-0.1576484889,0.1958630383,0.564645946,0.06718117,-0.1934890896,0.0277950019,0.4950584769,0.4522051215,-0.7212633491,0.0556065291,0.077163212,0.0833076388,-0.1817294955,0.1873534471,0.0858508795,0.1851811707,0.2049118727,0.144572854,-0.2454440743,-0.0605238006,0.1923827231,0.1505837291,0.2461463362,0.1484671086,0.024957316,-0.1038148552,-0.1354483515,-0.0955180898,-0.6611912847,-0.0493910983,0.0582037158,-0.2890035808,0.0885872319,-0.158786878,0.1201858744,-0.1778049171,0.4030880928,0.0889582857,0.4124264121,-0.1165221557,-0.1426828504,-0.5166683793,0.090314582,0.0872476101,-0.1255686879,-0.3189953864,-0.1061312109,0.0490008928,0.1152090579,0.057926055,0.0430919193,0.2217225283,0.6491606236,-0.6396930814,0.3707626462,0.121919021,-0.1066777855,0.2320800424,-0.3724495769,0.1926014423,-0.0187840965,0.1491537392,-0.1150114089,0.188808769,-0.1274182647,0.5095852613,0.2310601771,0.1437452286,0.2511382699,-0.1507162601,-0.135863319,0.1007360443,0.3031366765,0.1630814373,-0.5176221728,0.2995135188,0.0890445486,-0.2253535986,-0.2122370452,-0.0719156265,0.2695272267,0.0286270846,0.1764403433,-0.1058529764,-0.0242520552,0.1053629816,0.0221929569,-0.0265655778,0.4688787162,0.0108470069,0.2176401913,-0.3726311326,-0.1097332165,0.2490936071,0.1428989917,-0.195576787,-0.3436609805,0.3892367184,-0.3814237118,0.0986457467,-0.1977394074,-0.0275967102,-0.1511519849,0.0076970379,-0.5447806716,0.2385076582,-0.228387922,-0.5854173303,-0.0143674519,-0.1419529766,0.2517256141,0.0379593186,0.132112056,-0.2587891817]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2067","title":"Multiprocessing windows error","comments":"It's already on the newest version","body":"As described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop ","comment_length":6,"text":"Multiprocessing windows error\nAs described here https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2\r\n\r\nWhen using the num_proc argument on windows the whole Python environment crashes and hanging in loop.\r\nFor example at the map_to_array part.\r\nAn error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop \nIt's already on the newest version","embeddings":[-0.1470112354,-0.4598246515,-0.098134473,0.195708409,0.0235011317,0.0182157122,0.0210335329,0.0150502343,0.0011364494,0.2937539816,0.1118116826,0.0253581367,-0.3181667626,-0.0178501215,-0.2106733173,-0.1375428587,0.1114064753,-0.1293423623,0.2496031374,0.3085015416,-0.4563819766,0.4901825786,-0.283716917,0.1693427861,-0.4059241414,-0.0594482981,-0.2049338371,0.4911695421,0.2082477957,-0.4488157928,0.0325275436,-0.293407023,0.168697089,0.1342573017,-0.0001148234,-0.1288393885,0.020090675,0.0100268517,0.3554390073,0.226509288,0.2500836551,-0.0082993964,-0.0757422224,-0.0352966413,0.1461862624,0.2326803654,0.0663136989,0.0261753406,0.4473269284,0.2970738113,0.1551674753,0.24612616,0.0922253579,0.0045153447,-0.3034842312,-0.4103448987,-0.0532904603,0.2671407461,0.5082380772,-0.2907488048,0.1515360922,0.2966929674,-0.218740955,0.0582419075,-0.3532700241,-0.0200053416,0.4740324318,-0.478682965,0.0569702126,-0.0674408451,-0.1244322434,-0.1712801158,0.3040812314,0.1232052743,0.1009927318,-0.9145490527,0.0448958874,0.1712746769,-0.3666515648,0.0557470545,-0.0665745363,-0.0276697278,-0.0317843519,-0.1999340802,0.0884605795,-0.0112628974,0.0677101761,0.1463854909,0.1663657278,0.2866080701,-0.2240591198,0.3578910232,-0.0920371562,0.0552671403,-0.4950811565,-0.0126307877,0.0604108088,0.2007015944,-0.2855954766,-0.204433769,-0.0501682945,0.1242200509,0.2444463819,0.1478214562,0.2127529234,-0.1750814915,0.0741083324,-0.6318687201,0.2306944132,-0.4335404038,-0.1181603521,-0.0693702102,0.0745586157,-0.5753516555,0.3053528666,0.4294237494,0.2262350768,-0.0798361152,-0.3273139596,0.2176989764,0.0480168946,0.1994196326,0.2315791696,0.2676903605,0.339047879,0.6821905375,-0.0515428558,0.142777741,-0.5954546332,0.1211284995,0.0350761749,0.0968117788,-0.2849585414,0.1352418065,0.2617924809,-0.039270699,0.1523281038,0.1389423311,0.2618159354,-0.3040523827,-0.1048570871,-0.1955327541,-0.1651125848,0.2426718175,0.2088986188,0.2113162726,-0.0545401908,0.0690641329,-0.0304464605,0.2005785704,-0.3519704044,-0.3887860775,-0.1034217849,0.0973835364,0.4691507518,0.4268140793,0.3057056665,0.0027848741,0.2835047543,-0.3198097646,-0.1439372152,-0.4706510901,-0.3024942577,-0.2919241488,-0.0951119363,0.2973691225,-0.0044548684,0.2853105664,0.3140617311,-0.0515510626,0.1699762046,0.3550128639,0.2188690007,-0.2503147423,-0.0856242701,-0.0460296199,-0.1577486545,-0.2252583504,-0.0649216473,0.0899937898,-0.0562110692,-0.115186967,0.2696543932,0.1211127564,-0.0120847244,-0.3536699712,0.4566870928,-0.239656046,-0.1260048449,0.1861274689,-0.2492340952,0.1251110733,-0.0903931037,0.0405130647,0.0842546076,-0.1491047889,-0.0528489351,-0.0604138635,0.2615601122,0.1260359436,-0.0569600202,0.2892727852,0.1757917255,-0.0548230484,0.0518657975,-0.4055272043,-0.0569051653,0.0417499766,-0.2612916827,0.0185322613,0.373206228,0.1133816689,0.1269827783,0.2554255724,-0.1810826957,-0.2802907228,0.0666107535,-0.0655216873,0.1027075723,-0.2107228339,-0.0398380123,0.4921443462,0.0057044984,-0.1609346867,-0.1453109682,-0.0871372372,-0.419326961,-0.3308399022,-0.0106385453,0.3531771302,0.1999643594,0.1231449768,0.1218544394,0.2198801637,0.0772422254,-0.1479680389,-0.1874117404,-0.0154533107,0.2249727994,-0.2310840636,0.0875393078,0.0257167872,0.115991421,-0.20197092,0.4855372608,0.1144150645,0.0088923322,0.2118048519,-0.1013672277,0.1152060181,0.1570578218,0.0443003364,0.2630828321,0.2179477811,0.3673399389,0.1484789997,-0.1053150818,-0.350102514,-0.3466369808,0.0875113085,0.2346928418,0.186216116,-0.0467611477,0.0487236716,-0.1498637348,0.128766045,-0.0473608933,-0.4027210474,0.6512473226,0.1455041617,-0.1063285097,0.3118601441,-0.4597226679,-0.1481981575,0.2632541358,0.0308389273,-0.0304451119,0.1898847222,-0.0629523396,-0.1062905788,-0.110757485,-0.1032398269,-0.0911729857,0.072235696,-0.2952122986,0.0793626383,-0.097921446,0.486019522,-0.4318524599,-0.4729999304,-0.0097159892,-0.1944988072,0.128987655,0.4141593277,-0.0652972981,-0.0303301513,0.0624693185,-0.0378902517,0.1235274374,0.3879701793,0.2149743736,-0.0560178533,-0.1076677069,-0.0407683775,0.3430657089,-0.1595221162,0.2469024509,0.2830944061,0.0049718455,0.1169176176,0.1319651157,-0.0528288595,0.0481341034,0.2385084033,0.0147255687,0.0871120095,0.1784282923,-0.4111972153,0.1046803892,-0.3886035681,0.1887014806,0.1922444552,0.1077913493,-0.3739150763,-0.2542684972,-0.191607669,-0.4613941014,-0.1927898377,0.0190807674,-0.1174287349,0.3171318769,-0.2492703944,0.0680901036,0.1014063656,0.3346304297,-0.1071487293,-0.1781199425,0.2035069019,-0.0583426394,-0.2641663849,-0.1700255424,0.1663596034,0.2139270008,0.050841298,-0.2382321954,-0.1723600626,-0.3711561561,-0.0119617311,-0.0076152328,-0.042756442,0.274274081,0.4325153232,0.1744091809,-0.1820848584,0.1615243852,-0.0951685011,0.0119105391,0.0394734591,0.1459180713,0.2159732431,0.1244317368,-0.1638201624,0.5024106503,0.5853393078,0.0443879254,0.2695492804,0.1100000963,0.0283618979,-0.0068995254,-0.1807430536,0.1016137004,-0.3968820274,-0.1230503321,0.1551330537,0.1118571907,-0.3179630935,-0.2604637146,0.2305342853,-0.2906076908,-0.1081010103,0.0651057661,0.0840154067,0.1348668188,-0.158357352,-0.0630924776,0.0931487903,0.1017890722,-0.1042639017,0.191878438,-0.2866155803,-0.2100088,-0.052357778,-0.5948957205,-0.4821405113,0.4274253845,0.1487260461,0.5266762972,0.0551702268,-0.1471850425,-0.0239899792,-0.3150900602,0.7597226501,-0.1324266344,-0.1797020584,0.2366455793,0.2095224559,0.0049613952,-0.1426147223,-0.2107353508,0.3032571673,0.1316261292,-0.0709688962,0.27800107,-0.054188028,0.0947249308,-0.0750147477,-0.1814965159,-0.0863068178,-0.3491305113,-0.314440757,0.0285313856,0.2567334473,-0.0197352171,0.0228632744,-0.4232239127,0.02211928,0.1250812709,-0.1208658889,0.2660653293,-0.0509600602,0.1182888597,-0.1211699247,0.1162620187,-0.2728251815,0.3739458323,-0.1477684826,0.0179835483,-0.2101266086,0.1917238832,0.2880157232,-0.0477378368,0.3391045332,0.418885082,0.1319903731,-0.0249010045,-0.3289114535,0.4727149904,-0.2315450609,-0.2001162171,0.3241170645,0.2321806401,0.0701397583,0.2588511109,0.3359827995,0.1365028322,-0.0995366201,0.4898032546,-0.3295747638,-0.0860401392,-0.0426769666,0.121533297,0.9191321731,0.097755678,0.1022146195,0.0254258811,-0.5016573668,-0.1980990916,-0.1537685096,0.3429461718,-0.2533531487,-0.1485435814,0.0313661471,-0.0065089767,0.1544480026,-0.1586741805,-0.3109474182,0.4999437034,0.1488056183,-0.2398976684,-0.1789102107,-0.201960519,-0.0961258933,-0.1700949669,0.0000727372,0.0731061772,0.0872833133,-0.1342588067,-0.2742702663,0.1501696259,0.4967750609,-0.3882322609,-0.4797723591,-0.0055332589,-0.2071376592,0.1861408502,-0.0527716912,0.1332581639,0.0282346252,0.3386255205,0.1650202274,-0.1060371771,-0.2099058181,0.0157803874,0.4105873406,-0.045768138,-0.1060388684,0.0400682576,0.0194894075,-0.1595455855,-0.2663602829,0.3908213973,-0.1003779471,-0.0361853354,-0.3542098701,0.3331389725,0.0891607627,0.2047449797,-0.0904038101,0.2502554059,-0.1177242175,-0.2135710418,0.1179823279,0.023577705,-0.4966156185,0.5932769179,-0.5668897033,-0.1128286272,-0.0067208754,0.0946821943,0.0710753128,-0.0588521808,0.1835203022,0.2587498426,-0.1115447134,-0.2307255864,-0.1396129727,-0.468349278,-0.0696760044,-0.0165652763,0.4037169218,-0.1470126212,0.260902971,-0.1564845145,0.1110850945,0.0783907101,-0.0152138155,-0.3078763187,-0.2087115794,-0.0669984967,-0.0681939051,0.1597523242,-0.0140644405,0.127865836,0.0982215405,0.0952176899,-0.2987969816,0.0324001722,-0.1385433674,0.0888370052,0.0250538178,-0.2708188891,-0.0974803343,0.2364629507,0.0724490881,0.3582081497,-0.3873918653,-0.1579744369,0.0654318184,0.0704368278,-0.1098410264,-0.0604104139,0.0570210926,0.1207987815,0.0539918616,0.0870007426,0.1592305899,0.0741682351,-0.2711723149,0.1186738759,-0.0312391426,0.1943428516,0.2189780027,0.2901547253,-0.1725918055,0.4277845919,0.1120711118,0.0316596776,0.0460075177,-0.0193494018,-0.0817662254,0.0581092685,-0.0793272182,-0.2154501677,-0.1505439579,-0.1204727367,-0.3179673254,-0.0961108804,0.2954895496,0.3488516808,-0.1456304193,0.2288564146,0.3252750039,0.2260995656,-0.2585379779,0.0405044109,-0.1237016693,-0.1081092879,-0.1210197136,0.2877282798,0.2103772759,-0.4824697077,0.3463277817,0.07005018,0.1727939397,-0.4697990716,-0.3136231303,0.0902142897,0.2846082151,0.0795837641,0.0387877524,-0.0887193456,0.0669911727,0.5037372708,-0.0846512318,0.1777022928,-0.2204081714,0.0596687272,-0.1348769814,-0.1713755578,-0.2105768472,0.0984808579,-0.3697226942,0.422704041,-0.1869651526,0.4300532043,-0.5738232136,-0.3019706905,-0.3565172255,0.388058722,-0.2606543005,-0.1709741056,0.1060052812,0.0062725926,0.130229786,0.1161564514,0.2900328934,0.0934843943,0.5081411004,0.2037745565,-0.1138663888,-0.0665679649,-0.272770524,0.1585333049,-0.0143219633,-0.1964063793,0.2317933291,0.5241129398,-0.0588919818,-0.1761346459,0.0145498095,0.544082284,0.3519469798,-0.7084515095,0.002187157,0.1215379387,0.1416290849,0.0238321181,0.1357809305,0.1838688552,0.2677450478,0.0336716138,0.0317024775,-0.1329991072,0.0696771517,0.2652842402,0.1750365496,0.0748417899,0.2494674176,-0.0632198751,-0.2025806159,0.0665572137,0.1197799072,-0.6514270902,-0.0261439867,0.2199442536,-0.3901629746,0.0666712746,0.0275690295,0.0719346702,-0.0651705712,0.3854533732,-0.0350825004,0.2194461673,-0.2210822105,-0.1508053243,-0.3602021337,0.0201833341,0.1119715571,0.0033467268,-0.1798759699,-0.1112011075,-0.0674330294,0.1332281977,0.0848520398,0.1017170325,0.2966872454,0.6516944766,-0.6099458337,0.3992673457,0.3078179061,-0.0335439369,0.2716999054,-0.3522412777,0.2086438537,-0.0108551746,0.0724921376,-0.2307866812,0.2142491192,-0.0783884078,0.4642305672,0.20224379,0.0564735532,0.2176684141,-0.0123056369,-0.0979412347,0.0539324358,0.3582793474,0.2449377179,-0.5376458168,0.2820643485,0.0945077315,-0.3884758949,-0.0558623523,-0.0283809397,0.0294171851,0.0775818676,0.1024303138,-0.1247152016,0.0267611463,-0.0202903096,0.0385810509,0.0461178608,0.4650012553,0.0407982804,0.2234863788,-0.2948276997,-0.0501333922,0.1047163382,-0.0819447562,-0.2927937508,-0.3564984202,0.520560205,-0.4690045416,0.1540401578,-0.0875386074,-0.0464819372,-0.0634218156,0.0286296941,-0.6570464373,0.2914811969,-0.3136895597,-0.6943062544,0.1476496905,-0.1398388892,0.1760209799,-0.0790360048,0.197909236,-0.191296488]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2067","title":"Multiprocessing windows error","comments":"```\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\shutil.py\", line 791, in move\r\n os.rename(src, real_dst)\r\nFileExistsError: [WinError 183] Eine Datei kann nicht erstellt werden, wenn sie bereits vorhanden ist: 'D:\\\\huggingfacecache\\\\common_voice\\\\de\\\\6.1.0\\\\0041e06ab061b91d0a23234a2221e87970a19cf3a81b20901474cffffeb7869f\\\\tmpx9fl_jg8' -> 'D:\\\\huggingfacecache\\\\common_voice\\\\de\\\\6.1.0\\\\0041e06ab061b91d0a23234a2221e87970a19cf3a81b20901474cffffeb7869f\\\\cache-9b4f203a63742dfc.arrow'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 116, in spawn_main\r\n exitcode = _main(fd, parent_sentinel)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 125, in _main\r\n prepare(preparation_data)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 236, in prepare\r\n _fixup_main_from_path(data['init_main_from_path'])\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 287, in _fixup_main_from_path\r\n main_content = runpy.run_path(main_path,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 265, in run_path\r\n return _run_module_code(code, init_globals, run_name,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 97, in _run_module_code\r\n _run_code(code, mod_globals, init_globals,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 87, in _run_code\r\n exec(code, run_globals)\r\n File \"F:\\Codes\\Python Apps\\asr\\cvtrain.py\", line 243, in \r\n common_voice_train = common_voice_train.map(remove_special_characters, remove_columns=[\"sentence\"])\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 1339, in map\r\n return self._map_single(\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 203, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 1646, in _map_single\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\shutil.py\", line 805, in move\r\n copy_function(src, real_dst)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\shutil.py\", line 435, in copy2\r\n copyfile(src, dst, follow_symlinks=follow_symlinks)\r\n 0%| | 0\/27771 [00:00 'D:\\\\huggingfacecache\\\\common_voice\\\\de\\\\6.1.0\\\\0041e06ab061b91d0a23234a2221e87970a19cf3a81b20901474cffffeb7869f\\\\cache-9b4f203a63742dfc.arrow'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 116, in spawn_main\r\n exitcode = _main(fd, parent_sentinel)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 125, in _main\r\n prepare(preparation_data)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 236, in prepare\r\n _fixup_main_from_path(data['init_main_from_path'])\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\multiprocess\\spawn.py\", line 287, in _fixup_main_from_path\r\n main_content = runpy.run_path(main_path,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 265, in run_path\r\n return _run_module_code(code, init_globals, run_name,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 97, in _run_module_code\r\n _run_code(code, mod_globals, init_globals,\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\runpy.py\", line 87, in _run_code\r\n exec(code, run_globals)\r\n File \"F:\\Codes\\Python Apps\\asr\\cvtrain.py\", line 243, in \r\n common_voice_train = common_voice_train.map(remove_special_characters, remove_columns=[\"sentence\"])\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 1339, in map\r\n return self._map_single(\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 203, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\site-packages\\datasets\\arrow_dataset.py\", line 1646, in _map_single\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\shutil.py\", line 805, in move\r\n copy_function(src, real_dst)\r\n File \"C:\\Users\\flozi\\anaconda3\\envs\\wav2vec\\lib\\shutil.py\", line 435, in copy2\r\n copyfile(src, dst, follow_symlinks=follow_symlinks)\r\n 0%| | 0\/27771 [00:00 This way you just need to set the permissions once after doing load_dataset for example, and then all the new transformed cached files will have the same permissions.\r\n\r\nI was referring to this. Ensuring that newly generated `cache_files` have the same permissions","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":43,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\n> This way you just need to set the permissions once after doing load_dataset for example, and then all the new transformed cached files will have the same permissions.\r\n\r\nI was referring to this. Ensuring that newly generated `cache_files` have the same permissions","embeddings":[-0.1649939716,0.2504474223,-0.1311805099,0.0018534489,0.0164031461,0.069685705,0.3859352171,0.2012920529,-0.243175894,-0.0268525351,-0.0832427293,0.0317552835,0.0088323886,-0.2032843828,-0.0566226169,0.1244930476,0.2251376659,0.1058571711,-0.0469120219,-0.0759035721,-0.1004975885,-0.1561995745,0.0020295053,-0.1382747442,-0.2803068757,-0.1393663138,0.250901103,0.1854483038,0.0243440717,-0.1551291049,0.2401857376,0.5586883426,0.1195890307,0.1278124452,-0.000108509,-0.1533883512,-0.055117704,-0.0804115534,0.0271396842,0.1929069906,-0.0747349039,-0.1969085336,-0.1917303503,-0.1530253291,-0.1357583404,0.1554865688,0.100497894,-0.6505795717,0.5274168849,0.1168165579,0.2193976641,-0.209947139,-0.3844199479,0.0381097049,0.2251027673,0.0479847714,-0.1684113145,0.1200009957,0.2247462869,-0.1187588498,-0.0439551845,0.1332410127,-0.3048710525,0.1042669639,0.2058614343,-0.1382554322,-0.2443340719,-0.5563910604,0.046993196,0.0875775963,0.6264379621,-0.0997860432,-0.5162678361,-0.2456747442,0.1034486592,0.1492293626,0.2983026206,0.3047206998,0.0544698052,0.2399952114,-0.5317008495,-0.0625439957,-0.1593406498,-0.0585576221,0.2357821465,-0.1048958525,0.0082516326,0.0236698687,0.0207114555,-0.0796308666,0.1798670888,0.1895848215,-0.4009330273,0.3293815851,-0.2059290409,0.0583290905,-0.2278135568,0.4405922294,0.2491425574,0.3186540604,0.1585353911,0.2033318281,0.0782409832,0.138376072,-0.1031730548,0.25676018,0.3083620071,0.3042286038,0.5483266115,-0.373014003,-0.2216498405,0.0326330364,0.0325833224,-0.3678637743,0.0904005691,0.3038929403,0.0947401226,-0.1761719137,0.2136986703,0.1021986604,-0.0048243995,-0.2351537645,-0.0421869718,0.2320284396,0.15459764,-0.0539756417,0.0452964231,0.1512020975,-0.0480053052,-0.3640183508,-0.0102021722,-0.3659335077,-0.3953460753,0.4914129972,0.4919768572,-0.1580433846,-0.0312158372,0.2863592505,0.1892482638,-0.1733047068,0.2945377827,0.009412948,0.4202327132,0.2606206536,0.1965859681,0.297606498,0.0516441613,0.0267492365,-0.0783434138,0.3222174048,-0.3108763099,-0.1091556996,0.2779992223,0.093814522,-0.0476204678,0.2909790277,-0.3671762347,0.0775539577,0.5478352904,-0.1628122181,0.3683684468,0.2572859228,-0.0955010355,-0.3553991914,-0.1884086132,0.3907552361,-0.0281885117,-0.0231803469,0.0068803262,-0.0012331039,-0.1035846323,0.2831072211,-0.0647817925,0.1235592961,-0.1371367723,0.1607354879,0.4129779041,0.0167185403,-0.4896352291,0.1158401445,0.1546936035,-0.1152524054,0.2207864225,0.0094192261,0.1001000702,-0.1719110906,-0.1612109691,0.3214648962,0.2198984027,0.2970988452,0.0704512745,0.0779512376,0.0667019188,-0.1811559796,0.0924618393,0.4182465374,0.2940609157,-0.2295071781,0.1312010437,-0.3995591402,0.2576647699,0.3549225032,0.056103576,-0.1388367414,-0.0227873698,0.0151936188,-0.465315789,0.2769156098,-0.2263011783,-0.3846693933,0.1097714528,-0.4639790356,0.0545065328,-0.357956171,-0.1813165545,-0.0714326426,0.1170308292,-0.0007731954,0.0085544297,-0.3308463991,0.1423318684,0.475835681,0.2152563483,0.010602708,-0.1697903126,0.0889068022,-0.0840100795,-0.0551139712,-0.2312562168,-0.0318296291,0.341377914,-0.0363165475,-0.2875853777,0.2420509309,0.0710614175,0.1021632701,-0.0845881253,0.2678926289,0.0363490619,0.126285091,0.116907008,-0.3277269602,-0.1060258076,-0.1014627218,-0.1738274693,0.3757101893,-0.0945548564,-0.0556380637,-0.2235654444,0.1533218473,0.1444268823,-0.0640033036,-0.1271198392,-0.1657225639,-0.0189023614,0.1657900959,0.327701658,0.2167519033,0.1122761443,0.1225076765,0.2859781981,0.1796833277,0.0327649675,-0.0168508645,-0.0244936328,0.1001113653,0.1447474509,0.4372525811,0.2620206773,0.0575689003,0.2727523148,-0.1311240494,0.3154355884,-0.1805612892,-0.0081166187,-0.1601167321,0.0520053692,-0.0672347024,-0.1780880988,-0.0127130188,-0.0466869473,0.2545972168,0.1635733247,0.0634753853,-0.3407387137,-0.0851239786,-0.220859766,-0.1682960391,0.2248603106,-0.0677297488,-0.5411041379,-0.0004616114,-0.1686425656,0.1575246304,-0.0167004056,0.2290778011,-0.2502934039,0.230942741,-0.0517680682,-0.1068639457,-0.2852395177,0.1904079318,0.0318781622,0.0391563587,0.2648197412,-0.6819695234,0.108217001,-0.049818676,0.3214016855,-0.3362497389,-0.00658532,-0.1964795291,-0.0312058963,0.0932099745,-0.5233697891,0.110667862,-0.0999888703,-0.0006877767,-0.1943660229,-0.2517957091,-0.2066126317,-0.0055068177,-0.0349692963,-0.0534206629,-0.2905823588,-0.1511854231,-0.0614063665,-0.1431301534,0.4335410297,0.0851116404,0.2298046649,0.0896163732,-0.2578107119,-0.481002599,-0.161527887,0.329818964,-0.4195415676,-0.5174462199,0.3248214126,-0.5282965899,-0.3527770936,0.0235857219,0.1081352681,0.0974141732,0.5158916116,0.0116876289,-0.0420574248,-0.1728509516,0.2145496309,-0.2988652885,0.0996037349,0.0261344928,0.0606896952,0.0976809114,-0.0486990474,-0.1222708896,0.0462080315,-0.0357472785,-0.2109108269,0.0705680549,0.015196871,0.1574347317,0.6721852422,0.395839572,0.271814853,0.2819131017,0.2598074675,0.6446481943,0.0928580835,-0.1120511666,0.0937767923,-0.0434907377,-0.1148508266,-0.0076510161,-0.013474158,-0.2159669995,0.0148530444,0.0615101457,-0.2859520316,-0.4133946896,-0.1890836507,-0.393058002,0.2043073326,0.1477512568,0.1117432117,-0.2024671584,-0.1891775131,0.0177695341,0.3212450445,0.2525002062,0.2085790187,-0.3571209013,-0.032528773,-0.2535929382,0.385959506,-0.0413270891,0.471311599,-0.1928551793,0.0616709553,0.1350405961,0.3420900404,0.433167249,-0.3725552559,0.1792226583,-0.0567197092,0.2247841954,0.3942975998,-0.3113518357,0.4675217271,0.3880842924,-0.0016575743,0.1709219068,0.1048133969,-0.1330661774,-0.4171785414,-0.2021065652,-0.1294090748,-0.1466636211,0.1025315076,0.2365315259,-0.1127215475,-0.0477675125,0.2065170407,-0.336530298,-0.2709621787,0.0050461558,-0.2208545357,0.162790224,0.1010997966,0.017234372,0.334277004,0.1163222194,0.0395709909,0.3923638761,0.0280250981,0.0475220941,0.3474339843,-0.2491829842,0.3688551188,-0.2201862335,-0.2018560022,0.2742723525,0.0419488102,-0.3801099956,-0.1346422881,0.0952149183,-0.1462850124,-0.31521222,-0.0926018059,-0.1412156224,-0.0508099943,-0.2401287556,-0.6445422769,0.2152569592,0.5688731074,-0.3429946899,0.2212749422,-0.33525455,-0.3594666719,0.1160676032,0.1292529702,0.7586529255,-0.1700624377,0.3338815272,-0.3670007288,0.1270795614,0.0235381853,-0.4897478819,0.0970184579,-0.4280510247,-0.352004081,-0.1349547803,-0.1423650086,0.0306568388,0.4354758859,-0.0121358735,0.106662944,-0.1066157371,-0.0949271992,-0.2107930034,-0.1699941605,-0.2021683902,-0.3152822852,0.285990864,0.1521002352,-0.3186113834,0.0700334311,-0.0612332374,-0.0316593349,-0.135367468,0.2244103253,-0.2181379497,0.0759987161,-0.5058142543,0.095240809,-0.1806235015,-0.0695649981,0.1136438698,0.2644086182,0.2961452007,0.2816789448,0.0190243386,0.040675506,0.3031314611,0.220173493,-0.2503170967,-0.025692407,0.3489367664,0.1967999935,0.0993245766,-0.3126575649,-0.1262638718,-0.2124952674,-0.3996905386,0.1870090365,0.0007015962,-0.2862558961,0.0942010805,0.0512909815,0.0647279993,0.0765366703,0.1084605753,0.3929049969,-0.0862961337,0.591555953,-0.2861962914,-0.3366979659,0.0760435089,0.515617013,-0.0015924396,-0.2648315728,0.1439149827,-0.375662446,-0.2430399656,-0.1684212834,0.1851774156,0.0896777958,-0.193616733,-0.0058992025,0.0644206479,-0.0277561545,0.2419086844,-0.4868358374,0.0766687915,0.1860869974,0.2071248591,-0.3204559684,-0.3578536808,-0.2923529744,-0.3840437829,0.2802538276,0.2492889166,-0.4645656049,0.1048678756,0.1395434886,-0.2441070527,-0.1916093081,0.1240930855,0.1581029594,0.1949627995,-0.0215587784,-0.1759290397,-0.2068544179,-0.0686160699,0.0287300125,-0.0646962002,-0.1990014315,-0.1046936512,0.1134178489,0.1913428605,-0.2448037565,0.034024328,-0.3747049868,-0.0902366117,-0.4159332812,0.0488302819,0.2445970178,0.1655190438,0.3717602193,0.1695109904,0.0040591219,-0.1480575502,0.1289695948,-0.1361818761,-0.1748696715,-0.1619594395,0.246004492,-0.3091619313,-0.0388600342,-0.0312950686,-0.089454852,0.1638978273,0.1182409748,0.083917819,0.0065051569,-0.0093806544,0.1649020463,0.2747564912,0.090751566,-0.1394848675,0.0161914583,0.3732375503,0.134622097,0.0957644358,-0.2577314079,0.3211343884,-0.1577557474,0.0095060095,0.366600275,0.3107985556,-0.2781637013,0.1229009107,-0.0818173662,0.078799516,-0.3840858936,0.488989532,-0.0606765896,0.0978192464,0.028835969,0.4561582208,0.2342559099,0.0128776003,0.3948357999,0.1913920045,0.2803450525,0.3155324757,-0.295694232,0.0932704359,-0.6603910923,0.1270816922,-0.0133499485,0.035137292,0.3175449371,0.0470762998,0.0308494102,0.0010085002,-0.3721475601,-0.0357873812,-0.0288170427,-0.129378438,-0.1112948284,0.1116017401,0.094968535,0.0634832084,-0.2576341629,-0.1162931249,-0.1733274013,0.2095679194,0.3084501922,0.0606205761,0.2429190576,0.0343625061,0.1661372334,0.0629638433,-0.1984054595,-0.058695633,-0.4229193032,-0.1399531364,0.2167055905,0.0652495325,0.0771240816,0.1780663729,0.1526355743,0.2093239278,0.2992311716,0.1038565859,-0.2305787802,0.3860934973,-0.0636697561,-0.2987654507,-0.0091264136,0.0410542674,-0.0676222518,-0.0138775231,-0.2073088586,0.0249307361,0.0993677527,0.0009104472,-0.0046150521,-0.3906281292,0.0279034302,-0.0259139575,0.1253708303,0.0821533576,0.6104055643,0.030668579,0.1854172647,-0.5288829207,0.068934314,0.503528595,0.3143566251,0.3292217255,-0.1399837583,-0.0738605633,0.057714507,-0.3813987076,0.1041041389,0.2445229739,0.4890331924,-0.0709786788,-0.119734101,0.0386080965,-0.1364952475,0.5329033136,-0.3225668967,0.2944774628,0.3345531225,-0.2373430878,-0.2263546735,0.1648193747,-0.041113805,0.1047648266,-0.2933596373,0.2788981795,0.2183083892,0.0790699795,-0.1036352441,-0.1770000905,0.2364227027,0.4915590286,0.5217715502,0.2991144061,-0.064364396,0.0274173524,0.4217230976,-0.4742162526,0.05947797,-0.1954739541,-0.1081849635,-0.1217607483,-0.0582186878,-0.5593055487,-0.14126046,-0.438093096,0.0096385023,-0.0817699432,0.2550711334,-0.01171837,0.023389712,-0.0902685672,0.0334536582,0.2685429752,0.3378418684,-0.1183919832,0.4349156022,-0.4485858083,-0.1820120066,0.3286102414,-0.2922587991,-0.2384978533,-0.0066258018,0.4036854804,0.112254329,-0.2181578577,0.0144234402,0.1285201758,0.35363096,-0.3826270998,-0.0202641729,0.1170403957,-0.3663059175,0.030218849,-0.1388580501,0.4482254982,0.0068106032,-0.2219231725,0.0491896495,-0.2370274961]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"Yes exactly\r\n\r\nI imagine users can first do `load_dataset`, then chmod on the arrow files. After that all the new cache files could have the same permissions as the first arrow files. Opinions on this ?","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":36,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nYes exactly\r\n\r\nI imagine users can first do `load_dataset`, then chmod on the arrow files. After that all the new cache files could have the same permissions as the first arrow files. Opinions on this ?","embeddings":[-0.0982318446,0.2663797736,-0.1286433786,0.0853287876,-0.1320201755,0.0446737185,0.3324171603,0.118530646,-0.2356887311,-0.0621478073,-0.0989212915,0.0370316058,0.0195826124,-0.2461390197,-0.0805585906,0.0872824639,0.1227616966,0.1072847396,-0.0332508013,-0.0952983871,-0.0240278449,-0.1715578735,-0.0351817384,-0.0332676768,-0.1896342188,-0.2290271521,0.2739097178,0.2133919001,0.0309158172,-0.2010471672,0.2349330038,0.5016699433,0.1013555303,0.2010825276,-0.0001147357,-0.2658905089,0.0338738635,-0.0575639419,0.0078436742,0.2669466734,-0.1161731705,-0.2100120932,-0.0704279616,-0.1335719526,0.0365238823,0.0545726605,0.091035977,-0.7105931044,0.5204786658,0.0122761745,0.1791727096,-0.1938915104,-0.4023864865,0.083286427,0.2860995233,0.0430654585,-0.1606021523,0.1730280221,0.3103099167,-0.1361086369,-0.0524351932,0.1056952253,-0.2200270742,0.0986945555,0.2257885933,-0.1704335958,-0.2636166513,-0.4594815671,0.0759378374,0.1365863681,0.8266537786,-0.1198400781,-0.4820735455,-0.2638094723,0.0733138099,0.1983567327,0.2936440408,0.3246242404,0.0016637691,0.2001912594,-0.3582293093,-0.1222776547,-0.1994502544,-0.0233813897,0.2468711138,-0.1406525075,-0.0500772372,0.0535528772,0.0249775331,-0.0849356428,0.2674473822,0.0949541032,-0.3414618969,0.3800834417,-0.1853994131,0.0996082351,-0.2771460712,0.4416387975,0.1848989874,0.367099911,0.3062287569,0.261547029,0.0208358727,0.1910760105,0.014087867,0.2300602198,0.3232108951,0.3352376819,0.4978227019,-0.4137336612,-0.127027601,0.0657259524,-0.0633114055,-0.4195164442,0.0475073755,0.2454414964,-0.0278086159,-0.2097613066,0.222966671,0.1323930621,-0.0605982095,-0.2762775719,0.0271370858,0.3039213717,0.0922554731,0.0124805961,-0.0268436596,0.153289035,0.0331942029,-0.3798217475,0.0337653942,-0.3871083856,-0.4200714529,0.4929867983,0.4044268727,-0.1765277535,-0.033105135,0.2085524648,0.2425050139,-0.1636571139,0.3366356492,-0.0425914712,0.379229635,0.2863856852,0.1798949838,0.363432616,0.0198319405,0.0766787007,-0.1069900841,0.3586049974,-0.3342375159,-0.1467377394,0.2806135416,0.0304297805,-0.0477084741,0.2889646292,-0.4568911493,0.0635594353,0.4904056191,-0.1834812313,0.4105270505,0.2586520612,-0.0667117164,-0.4665744901,-0.203553617,0.4040682316,-0.1568007618,-0.043655809,-0.0511367545,0.0621061735,-0.1280885935,0.3088611662,-0.1244794205,0.114096649,-0.1328294575,0.1697129458,0.6214051247,-0.0554199554,-0.5478649735,0.1586862952,0.1892972887,-0.1027346328,0.2174239904,-0.0242303237,0.0259667523,-0.1259121299,-0.2156150937,0.2760068476,0.2407997102,0.1559189409,0.1018859372,0.0270283762,-0.0300635956,-0.0831490383,0.0225224774,0.2714837492,0.4265298843,-0.4358933568,0.2547696829,-0.296164602,0.243110463,0.4442699552,0.085216172,-0.0715317428,-0.0012013061,-0.0202656426,-0.5487953424,0.2893896103,-0.1768640429,-0.5273402333,0.0809141025,-0.4074394405,0.0392736867,-0.2897551358,-0.1810069829,0.0466867089,0.0411264412,0.0338834487,-0.0887254253,-0.4143950939,0.1162468866,0.3877355754,0.2359033525,0.031515725,-0.0698557869,0.0841894448,-0.071255438,-0.0374090448,-0.2687806189,-0.04930567,0.1768609732,0.0349990427,-0.2530092895,0.2679135203,0.0601470843,0.1441573054,-0.0802506134,0.2444082946,0.0304318555,-0.0141182709,0.1684853435,-0.2584220171,-0.0873372331,-0.1304801553,-0.2503659427,0.3609991074,-0.1785717607,-0.0334620401,-0.1110743806,0.1398726702,0.0843084604,0.0162244067,-0.1475671977,-0.1299829036,0.0527595319,0.2098969668,0.4378208816,0.2361806631,0.0789472386,0.1187857091,0.2683714926,0.1318301409,0.1181724072,-0.073609218,-0.1377052218,0.1421776861,0.247188136,0.36562711,0.2177818716,0.0397991464,0.3672138453,-0.2019691765,0.3553775251,-0.1610699892,-0.046028547,-0.1480808258,0.1355393976,-0.0648089275,-0.1560852081,-0.0596935451,0.0181031991,0.2383271307,0.1332809776,0.0081427488,-0.3435696363,-0.1004173458,-0.1508623809,-0.1511526555,0.1755269915,-0.1299466491,-0.4723000824,0.0583783053,-0.2219809443,0.1132966131,-0.0233516917,0.1973372549,-0.3970792592,0.2917776108,-0.128702566,-0.1223498359,-0.2481480837,0.230194509,0.0850458816,-0.001760689,0.3133168817,-0.609719336,0.0888756588,0.0885057524,0.3788987398,-0.2676569521,0.042918358,-0.1726459563,0.0873144269,-0.0265730564,-0.5216035247,0.0375708565,-0.2170250714,-0.0558556765,-0.1870139837,-0.2340089679,-0.2217162997,-0.0772934034,-0.0286511667,-0.0113656241,-0.2143590897,-0.0855834559,-0.1366393864,-0.1832230389,0.5226250887,0.1414613575,0.2560490966,-0.0023058013,-0.2259359956,-0.4002258182,-0.2243372947,0.3215575516,-0.3546554148,-0.4346234798,0.3281576633,-0.5130246282,-0.3416810632,0.0300765783,0.1240739971,0.1058700457,0.5141780972,-0.0260713454,0.08564087,-0.2186697721,0.201466918,-0.3087972999,0.1535314769,-0.0030912021,0.1196375713,0.1466685832,-0.0533184335,-0.0471319593,0.0780136362,-0.08128272,-0.2787156105,0.1123440564,-0.0189688616,0.1544212699,0.8165989518,0.4180293977,0.3073640168,0.2578543723,0.355474174,0.6551366448,0.1428945512,-0.0472472981,0.1153945178,-0.0610897467,-0.0731156543,0.0584624633,0.0761319324,-0.2318724543,0.019900687,0.1093602404,-0.2403735816,-0.3566123247,-0.1428791583,-0.452537775,0.1425889432,0.1079090089,0.0464174263,-0.2667253315,-0.2012426406,0.0368890353,0.2855019569,0.2145124823,0.2000102103,-0.4393431842,0.0232195649,-0.2913892567,0.3825502098,0.0240377951,0.376793772,-0.1595309228,0.0878664479,0.1422178298,0.3085395396,0.1991611123,-0.3688870072,0.217595309,-0.0722918361,0.3133903742,0.1935796589,-0.2300415784,0.4239263535,0.4167358577,-0.0649997741,0.0880103558,0.1043233946,-0.162448287,-0.4611425996,-0.1629195511,-0.1400511861,-0.1490297019,0.1277849078,0.3009722233,-0.059190128,-0.0507067069,0.2749481499,-0.3887252212,-0.1966952682,-0.0569889396,-0.2175360322,0.1835900396,0.058552403,0.0031904785,0.3556634188,0.0618902929,-0.0241897907,0.421960175,0.1054663211,0.1905620545,0.4501100481,-0.2394204885,0.2526356876,-0.1963379085,-0.2536073327,0.3440071642,0.1280103326,-0.3881405294,-0.0913410932,0.0948837921,-0.2409778535,-0.1785273999,-0.1467931718,-0.1393214017,-0.0003508169,-0.1793936491,-0.5722702146,0.2993484139,0.5455884933,-0.387396276,0.4094864726,-0.1822432131,-0.2915193439,0.1244032905,0.0855534077,0.8844551444,-0.1866296679,0.3774630427,-0.3722923994,-0.0078610834,0.0343177691,-0.3573721647,0.0480113924,-0.481505543,-0.3130868971,-0.1221838593,-0.1820074469,-0.00265422,0.4709053934,0.0955754742,0.0888458341,-0.1663901359,-0.0967439264,-0.1240568236,-0.1856989712,-0.1745550036,-0.2546985149,0.3172065616,0.0995548964,-0.2195305377,0.0381074697,-0.1016562581,-0.0048117442,-0.2958861291,0.2237450331,-0.263799876,0.0663347095,-0.6536076069,0.0883999094,-0.1841793656,-0.1916700453,0.1924735755,0.3233481646,0.2500437796,0.1908156127,0.010780286,0.084897466,0.3455126286,0.1513488889,-0.2724229991,0.0424260497,0.2580708265,0.1514903009,0.102369763,-0.1992269605,-0.1965437382,-0.3692899048,-0.304992646,0.2606559992,0.0203785226,-0.1966232657,0.0167658217,0.0096559739,0.0193984285,0.1338638365,0.0641382039,0.4409195483,0.011419924,0.5084742904,-0.1870291084,-0.2663937211,0.1212542132,0.4861816764,-0.051714614,-0.2246517688,0.2001373917,-0.3566205502,-0.2840156257,-0.1574020982,0.2058403194,0.2287734449,-0.2739263177,-0.0978605077,-0.0500142947,-0.1207570806,0.2087789625,-0.4400067925,0.1652747542,0.222519666,0.1445554495,-0.3259631395,-0.3702551425,-0.1557091773,-0.3802726567,0.169402346,0.0981938839,-0.6323500276,0.0848052502,0.158921659,-0.1927313954,-0.0859458297,0.1824995875,0.0555276461,0.1917813271,0.0183820855,-0.1213937253,-0.1498743892,-0.1152230799,-0.0271479841,-0.0567795299,-0.1641157866,-0.0398013778,0.1679484397,0.1820958704,-0.3579389751,0.0340808965,-0.478461504,-0.0929455236,-0.3725866377,-0.0069160927,0.1440308094,0.1984070987,0.2655472755,0.1583365053,0.0265441965,-0.2508019805,0.2939379215,-0.0832642466,-0.1457856297,-0.178561002,0.2487280965,-0.2560299337,-0.0689091831,0.1211491302,-0.0534530506,0.1599540561,0.0876603872,0.0329497643,-0.039718654,0.0136438934,0.0903760865,0.3623434603,0.1479850709,-0.0748307854,0.0384249762,0.3330177069,0.0894021317,0.0964139327,-0.2548564672,0.255743593,-0.1211253479,-0.0138891963,0.3589084148,0.2485765219,-0.2838557959,0.1082213968,-0.1117147058,-0.0246061683,-0.3429886997,0.418474853,0.0051704315,0.0947767645,0.0316189006,0.4915165901,0.279782027,0.1336300969,0.3028979003,0.1247790158,0.158916682,0.4848209918,-0.3369288445,0.1364086419,-0.7019543052,0.0402296595,0.0643207356,0.1500836164,0.2370772362,0.0920818076,0.1744425893,0.0702622011,-0.3857193887,0.0537546948,-0.2150057256,-0.1069023758,-0.173550576,0.0525690205,0.0101662828,0.0235932097,-0.2575086057,-0.2358400375,-0.2523638904,0.2380785346,0.2800002992,0.1252836734,0.1473443955,0.1435046792,0.2382258326,0.0518713295,-0.1395544112,-0.0491847806,-0.4437826574,-0.1235464364,0.222360149,0.070716165,0.0082142791,0.1415562481,0.0843228474,0.1917478591,0.3114384115,0.1031476781,-0.2605128288,0.2802602351,0.0313886069,-0.4263990521,0.0495663173,-0.0118828462,-0.0661640912,0.0395997353,-0.165369451,-0.0143568125,0.0257082302,-0.0964867026,-0.2372016311,-0.485943228,0.1218275428,-0.0287643038,0.1725472063,0.0998133272,0.6371156573,-0.0350283161,0.135784179,-0.5039880872,0.0236024614,0.422478497,0.30497244,0.4329595864,0.086129576,0.0417962782,-0.004052463,-0.427985549,0.1783240288,0.1930170208,0.4480085373,-0.0499122553,-0.0718685612,0.0979916751,-0.099458091,0.6948085427,-0.1792328656,0.4205053449,0.3750423789,-0.2903120518,-0.1950784326,0.3442022502,-0.0036933613,0.0794747472,-0.3978531361,0.2591148317,0.2087323815,0.0137087181,-0.0616119914,-0.2549702525,0.1834773272,0.5491983294,0.596067071,0.1380422711,-0.0373201892,0.0778951719,0.2939432263,-0.4272295535,0.0046468582,-0.1335633546,-0.0110724447,-0.1224510223,-0.0802226663,-0.5871816874,-0.0836935192,-0.3447754383,-0.0714138821,-0.0832641646,0.2952828109,0.0031480335,0.0124896951,-0.1258894205,0.0484202467,0.2322921753,0.2587404251,-0.177636072,0.3560259938,-0.4420720339,-0.0428557098,0.3386681974,-0.268160522,-0.1325930357,0.0472122654,0.3721211553,0.3301633894,-0.1808748841,0.0167279392,0.0868162736,0.3808145523,-0.3920348287,-0.0991075188,-0.0171486083,-0.4355275631,-0.0082124332,-0.1383624077,0.6257168651,-0.0122777689,-0.1993691027,-0.0361487009,-0.2800708115]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"Sounds nice but I feel this is a sub-part of the approach mentioned by @siddk. Instead of letting the user set new permissions by itself first and then making sure newly generated files have same permissions why don't we ask the user initially only what they want? What are your thoughts?","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":51,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nSounds nice but I feel this is a sub-part of the approach mentioned by @siddk. Instead of letting the user set new permissions by itself first and then making sure newly generated files have same permissions why don't we ask the user initially only what they want? What are your thoughts?","embeddings":[-0.0215067733,0.3034674823,-0.1613317132,-0.0614497289,-0.0943001211,0.0116148414,0.359359175,0.2102664709,-0.2881747484,0.0673978999,0.0218295306,-0.1237791181,0.0104278307,-0.2228574008,-0.14883551,0.1429995596,0.1225395575,0.1753851026,0.0643005967,-0.085233666,-0.0920050368,-0.25709185,0.0005783064,-0.0861270055,-0.3042495549,-0.1460853517,0.1376358569,0.1973124892,-0.0369805433,-0.1767524332,0.1454704255,0.5835268497,0.0604470633,0.1700236797,-0.000113813,-0.2299204469,0.0330866985,-0.0436646901,0.0165126044,0.2122360617,-0.073548831,-0.1644747406,-0.1943162978,-0.2433009744,-0.0462049842,0.1997121722,0.1979092062,-0.6541274786,0.5026413202,-0.036366798,0.1695914268,-0.2603195608,-0.4581024647,0.1149789691,0.2927057147,0.15541327,-0.1459863037,0.1088739336,0.4278711677,-0.2223556638,-0.0490458384,0.1052149683,-0.3066333234,0.1089935899,0.1439869106,-0.1189896241,-0.2343177795,-0.5785056949,0.0060195769,0.1186374128,0.7419643998,-0.041439373,-0.500002563,-0.3280445933,0.1157064736,0.2075068206,0.2222390324,0.3163419366,-0.1446903646,0.263974309,-0.4496850073,-0.093929708,-0.1393081397,-0.049711898,0.2192329317,-0.1179769263,-0.0371810645,0.0456820801,-0.035000667,0.017656872,0.2272628993,0.1144418865,-0.3335204422,0.3299391866,-0.1112532988,-0.0360920504,-0.2102335095,0.4718672931,0.2461711168,0.3557560146,0.2872901857,0.2251413018,0.0457360595,0.1456659436,0.0792336091,0.1384641081,0.3433903158,0.1938130111,0.5769106746,-0.3486528993,-0.0621921681,0.0978375822,0.0803735331,-0.309866786,0.1034736335,0.3815611601,-0.0208681747,-0.1177322268,0.1676379442,0.1382079273,-0.1400250793,-0.4070581198,-0.0386475138,0.2055515051,0.0458725207,-0.078172043,-0.005032972,0.1522562951,0.0363857076,-0.3549081385,0.0399276763,-0.3292353153,-0.3953119218,0.4818220139,0.5079624057,-0.210538581,-0.0186098106,0.2744808197,0.1505053788,-0.1056959406,0.3306500316,-0.0605823994,0.3992083967,0.2370217294,0.165595293,0.3706308901,-0.0237126276,-0.0111304512,-0.1217286736,0.3696787953,-0.3321189582,-0.1372836083,0.3728598356,0.054056108,-0.0034280377,0.3053071499,-0.3055223525,0.0975191444,0.4749915004,-0.1878505945,0.2645231187,0.3264573216,-0.1508256793,-0.4486918449,-0.1300678998,0.4278678894,-0.0520535856,0.0365230031,-0.0196552537,0.0565940328,-0.1223031953,0.2681515217,-0.0935709327,0.3671370745,-0.1649925113,0.1947631538,0.4293022752,-0.0106864683,-0.4389024675,0.1946795881,0.1696154028,-0.035028778,0.2052502036,0.0265082084,0.0568236969,-0.2115881741,-0.1429784149,0.2417724133,0.0068370011,0.2508014739,0.0217115916,-0.0139128566,0.0808936432,-0.1215712577,0.0580402538,0.3386098444,0.4151690602,-0.3177106977,0.2092598528,-0.3718044758,0.2121628076,0.2398813665,0.1287965626,-0.1351441592,-0.1405808926,0.0094823986,-0.4876556396,0.3215155005,-0.1577846259,-0.3858752251,0.145815447,-0.3815568984,0.1196946427,-0.3553343713,-0.0348144807,-0.0107425814,0.0641673431,0.011809337,-0.1238074079,-0.3839871883,0.0663139001,0.4213301837,0.1884029955,0.010893099,-0.0326096453,0.0793724954,-0.1622633934,0.0366548821,-0.3233988285,-0.1027956232,0.1893297732,-0.0289433897,-0.1541258991,0.2733415961,0.0848938525,0.0598437302,-0.0764382258,0.283999294,0.0712548941,0.1836898476,0.0505548939,-0.2241450548,-0.0589557067,-0.2423047572,-0.3031650186,0.3828921914,-0.0749819502,-0.0996832997,-0.1992581189,0.0980178714,-0.0098716058,-0.1272235513,-0.1304626912,-0.099872984,-0.0674129054,0.2601042688,0.2791475356,0.2810331881,0.0791931972,0.0599673539,0.3185290396,0.1296101063,0.0791056901,-0.0605271421,0.0566118956,0.0924589634,0.1974955946,0.3689728379,0.2394682914,0.0554052033,0.2686427832,-0.1833655983,0.3721331656,-0.1642819792,-0.0501156189,-0.0609723777,0.0270876717,-0.0607792512,-0.1656572372,0.0488088503,0.0602557845,0.2263970077,0.0412445255,-0.0455979444,-0.2876110375,-0.1677026451,-0.1197688505,-0.2105101496,0.2882496715,-0.0389637016,-0.4841929078,0.0286956802,-0.1127680317,0.1295776814,-0.0789717212,0.1243952364,-0.2907033861,0.343616128,-0.1635691375,-0.0660114139,-0.1650728583,0.2653517127,0.0996403098,0.0267295688,0.3398630619,-0.6474868655,0.2354015261,0.0241679084,0.3293393552,-0.2896651626,-0.0847147554,-0.1201487482,-0.0619772561,0.1082119122,-0.4897520542,0.0002001086,-0.1599302441,-0.024095282,-0.272926569,-0.1890534014,-0.169342339,-0.1547238082,0.0760243684,0.0209451504,-0.3150222898,-0.115810059,-0.0677887201,-0.1333333552,0.4151948392,0.1257451326,0.2871500552,0.0431081839,-0.2571903169,-0.4677893519,-0.2548931539,0.4332953095,-0.2978619933,-0.573590219,0.2667667568,-0.366728276,-0.2950362563,-0.0547699742,0.0534234121,0.0496604182,0.5260625482,-0.0280429497,-0.0710068494,-0.1727811247,0.2378222793,-0.3201941848,0.1760142893,0.0507704318,0.0443122573,0.1043993905,-0.005994197,-0.063017644,0.0995541736,-0.0724110752,-0.3326378763,0.210569337,0.054405693,0.2172205001,0.8299294114,0.3929793239,0.3720266521,0.2671792805,0.3564840853,0.6644089222,0.1048224345,-0.0725023672,0.1425319463,-0.0364810452,-0.0717449412,0.0694955587,0.081197463,-0.2659108639,-0.0207885373,0.1108503491,-0.3122138977,-0.4389321804,-0.1026622951,-0.3760763109,0.1761967093,0.1192985773,0.1065726727,-0.199528113,-0.1544144452,-0.0210666265,0.2632148862,0.3557827175,0.1747672856,-0.4637182951,0.0369300134,-0.3498628438,0.2690496147,0.0891756713,0.418587774,-0.1488755494,0.0581607334,0.175257802,0.2864212394,0.3135198355,-0.3702077568,0.0821559876,-0.1242390051,0.1768730283,0.3828261793,-0.1949596703,0.4329000413,0.4933440387,0.0772912428,0.1491273046,0.0295708831,-0.1438003927,-0.4299341142,-0.1553681493,-0.0387888439,-0.1753016263,0.2569079995,0.2764847875,-0.0702058971,-0.0548049361,0.3186102509,-0.4268202186,-0.1772151738,0.0125553971,-0.1444712877,0.3334314823,0.1070250347,-0.0956252441,0.1886481494,0.0868557245,0.0401330851,0.3317952454,0.082937628,0.0104299318,0.4379093945,-0.2634051144,0.4645081162,-0.1202840507,-0.3070551753,0.3799802363,0.2001865804,-0.2821282744,-0.1460819244,0.1002367064,-0.1474112123,-0.3917487562,-0.1683851033,-0.0840121955,-0.1172895655,-0.2973452806,-0.6107850671,0.2031315565,0.5352294445,-0.3493175507,0.2802352607,-0.197672978,-0.3199747801,0.1157681718,0.089970544,0.8514484763,-0.1189205721,0.4581424594,-0.3852830529,0.0906193629,0.0300648231,-0.4793527126,0.0184784327,-0.4417715371,-0.3964937627,-0.1573924571,-0.1264575571,0.0611548685,0.5075002313,0.0699487776,0.131933853,-0.0312287249,-0.0157000776,-0.0805767402,-0.0955837965,-0.2434277833,-0.2530539632,0.4673323929,0.1035224572,-0.2109701335,0.0416739807,-0.1058871299,-0.0191945992,-0.2302455306,0.1174574569,-0.1970194131,-0.0147033278,-0.5546617508,0.0146597689,-0.1509053707,-0.0532316454,0.1978967488,0.2467458695,0.2595117688,0.3690892458,0.0198924039,-0.0037854251,0.3759623468,0.2865215838,-0.1338464767,0.0350457765,0.2950537205,0.1952957064,0.1418515444,-0.2592481971,-0.1916336864,-0.3534186482,-0.5514914393,0.2923975885,0.0362899825,-0.2155066878,-0.0296292529,0.0761448294,0.0758176371,0.0935083777,0.0711848512,0.4465313256,0.0325407833,0.5419017673,-0.3480488956,-0.2729810774,0.0171713959,0.4562426805,-0.0087913563,-0.2799369395,0.1219349429,-0.4247003198,-0.2871575952,-0.1838988364,0.2165851891,0.0992644429,-0.1942404956,-0.1453796625,0.0039367247,-0.053670533,0.2282117903,-0.5018802285,0.1364123821,0.0631130338,0.2404439747,-0.2232862711,-0.3366977274,-0.2222482562,-0.2816527784,0.2878143191,0.1748156697,-0.5316985846,0.0342518277,0.1116781235,-0.2204289883,-0.161744684,0.2376939952,0.1767830104,0.237423867,0.0645030737,-0.2117530107,-0.1975309998,-0.0868834406,-0.0958414525,-0.0733445138,-0.1839970052,-0.0220428705,0.1630773097,0.1925931424,-0.3057267368,0.1155593023,-0.4096888006,-0.1216664538,-0.4183574915,0.0601981394,0.2559384406,0.2084679455,0.2805687785,0.2027067244,-0.0340545252,-0.1311457008,0.271774143,-0.1490413249,-0.1615099162,-0.2404677868,0.2573897243,-0.2780954838,-0.0271795131,0.0881201103,-0.0225398988,0.0841833055,0.1487890333,0.063941367,0.0238669943,0.0719450861,0.17729792,0.3038207591,-0.0395749696,-0.1294690669,0.0597624481,0.4733662009,0.1325888485,0.1642730385,-0.2211527079,0.296395272,-0.1967694908,0.0397145338,0.4383155704,0.261691153,-0.1876129657,0.1947375089,-0.1431871206,0.0286020096,-0.421778053,0.4083346426,-0.0260620285,0.0835901871,-0.0092685884,0.4539856911,0.1205734983,0.0059177103,0.2759369314,0.2359694093,0.2251792997,0.3980239332,-0.3255467415,0.0371676013,-0.6843452454,0.1587275714,0.0663741454,0.139927581,0.264914453,0.0653444529,0.0981892049,-0.0088553028,-0.3272551894,0.0779380724,-0.2223503441,-0.1073073894,-0.0925635621,0.092549324,-0.0064459261,0.0414032638,-0.2435081452,-0.2393182218,-0.2039109915,0.1722235978,0.2994827926,0.0268010087,0.2390525043,-0.0057677263,0.0227450561,0.1052185446,-0.2305179387,-0.0886004865,-0.549012959,-0.1587802321,0.1963799,0.157689631,-0.002471775,0.2507821321,0.115507029,0.1143042743,0.2105147839,0.1410673857,-0.3057365716,0.3160779774,-0.0784320757,-0.3726974428,0.0205800459,0.0133964606,-0.055695571,0.0213041492,-0.1417459846,-0.0697399378,-0.1042276993,-0.0591589287,-0.0012651468,-0.3647179604,0.1928966492,-0.0246814545,0.0996285304,0.0929882154,0.5304291844,-0.080103524,0.213189885,-0.5350333452,0.0406225137,0.4460453093,0.0969034955,0.4369759858,-0.0602700263,-0.1579893231,0.0823486894,-0.3973454535,0.067230843,0.2058847696,0.4810692966,-0.0473967493,-0.1647843271,0.160333842,-0.0201027133,0.5293853283,-0.3117315471,0.2309979349,0.3042064011,-0.2668891549,-0.1646341383,0.2002536207,-0.0559968464,0.0255701728,-0.3361581862,0.2876137793,0.2087489665,0.0566418804,-0.1186754331,-0.2009537965,0.1894121319,0.5180318356,0.4471519589,0.2234645188,-0.1281674802,-0.0144973202,0.3214601874,-0.4525119662,0.1204556078,-0.1858819127,-0.1433234513,-0.1372973472,0.0232628547,-0.5915025473,-0.1409793347,-0.3900015056,-0.0541636087,-0.0959737226,0.3669762611,0.0574753396,-0.0047727278,-0.0992674753,0.0349664055,0.2579489946,0.3559637666,-0.1373485029,0.4368424714,-0.426482439,-0.0077001676,0.2884618342,-0.2015532553,-0.146673888,-0.017401617,0.3491700292,0.1621322483,-0.1071659774,0.0584281571,-0.0640763715,0.3837525845,-0.4046954513,-0.0600084327,0.1296227276,-0.4057175517,-0.1622667909,-0.0585070699,0.5083826184,-0.0459665805,-0.2583738565,0.0135502685,-0.2333886176]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"Yes sounds good. Should this be a parameter in `load_dataset` ? Or an env variable ? Or use the value of `os.umask` ?","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":23,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nYes sounds good. Should this be a parameter in `load_dataset` ? Or an env variable ? Or use the value of `os.umask` ?","embeddings":[-0.1157243773,0.1915663481,-0.1204443127,0.0102188159,-0.0046569323,0.0738378987,0.3548937142,0.0830475613,-0.0977408066,0.0234236512,-0.1051575541,0.1137064472,0.0743694901,-0.2714213133,-0.0607254766,0.1525210589,0.2136559784,0.0242683012,0.0314018764,-0.1150255427,-0.1682758629,-0.0577769279,-0.1090057641,-0.0970223472,-0.2315439284,-0.1235303134,0.2449167967,0.2954755425,-0.0123770125,-0.1583914012,0.2343732268,0.5587266684,0.1839400828,0.0664559156,-0.0001146072,-0.1373330206,0.0245990865,-0.130072847,0.0329921618,0.2129813582,-0.0198336262,-0.1354002208,-0.1063277572,-0.2107834071,-0.1455820501,0.1845481992,0.0864102468,-0.669739604,0.4485447109,0.2498931885,0.184911862,-0.1789575666,-0.3419660628,0.1327785999,0.2498422414,0.1016633734,-0.1213095784,0.0644466281,0.2906067073,-0.1677010655,-0.0334176458,-0.0065966598,-0.2996161878,0.0673053935,0.2002077848,-0.0560005531,-0.262217164,-0.5426024199,0.0775001347,0.0678996667,0.7661519051,-0.0833125412,-0.51856637,-0.1043560579,0.0909009725,0.2254609168,0.3825606406,0.3455035985,-0.0217853263,0.3150945306,-0.5068026781,0.0105497101,-0.2292101681,0.0499693714,0.1538022608,0.0141860349,-0.001282188,0.1121540293,0.0697188452,-0.0866594464,0.1956954449,0.1137941554,-0.3244980276,0.3942143619,-0.2211425751,0.0370483957,-0.2199562788,0.4640308619,0.2384044975,0.2879548371,0.2111440897,0.1848994493,0.0075349989,0.2291600555,-0.0334106572,0.1429675817,0.3944539428,0.2765599191,0.5267387033,-0.3781735003,-0.110384129,0.031559065,-0.051687263,-0.3931072354,0.0737509131,0.2395581156,0.0403588451,-0.2036320865,0.1282831728,0.082135424,-0.1116086617,-0.2773318589,0.0313262269,0.3696843386,0.1135343835,-0.1150821596,0.039891094,0.2530939579,-0.0455173403,-0.3753743768,0.0114078391,-0.3871253133,-0.2792084813,0.4033256173,0.4752327502,-0.1266807765,0.0374808311,0.2577849329,0.1295548677,-0.2477270514,0.3621706069,-0.0441975184,0.39860636,0.3669503629,0.2635758519,0.2514336705,0.0983390361,0.02832569,-0.107914947,0.4005104303,-0.3557933271,-0.1548413187,0.2978987396,0.056241367,-0.0700574443,0.2229293585,-0.3771528602,0.100808844,0.5208300948,-0.0400726162,0.3602368534,0.2457832098,-0.2040585428,-0.4434044361,-0.0713331699,0.2709903419,-0.1343948692,-0.0585171096,-0.0877236873,-0.065246813,0.0261634514,0.3507983983,-0.0272545628,0.2278839648,-0.0913008526,0.163222909,0.4935807586,-0.0270813461,-0.5619918704,0.1231576353,0.2241749465,-0.1495332867,0.1917965859,0.0252751652,0.0326284543,-0.0548465289,-0.0813222677,0.3807575405,0.2587207258,0.2704668641,0.1384900957,0.1101884618,0.0516449511,-0.0187087525,0.0409566388,0.3495685756,0.384989053,-0.3570520878,0.2508898675,-0.2888385355,0.2376910746,0.4559271038,0.0941976309,-0.1270498186,0.0241289437,0.0441784486,-0.3931060135,0.250417918,-0.0999295861,-0.4025928974,0.0985727385,-0.4768047035,0.0507055409,-0.2261070162,-0.1504827291,-0.0876726881,0.0498791784,0.0069951494,-0.0551148281,-0.3226664066,0.1372427344,0.378921479,0.2070661783,-0.0071923095,-0.1255985349,0.1369618773,-0.0797372609,-0.0754805207,-0.2361454517,-0.0997979045,0.2494993359,0.0149250543,-0.2919357121,0.3010894656,0.2039221674,0.0037091915,-0.1793101728,0.15792422,-0.0316228084,0.0792514831,0.0922478735,-0.1883662939,-0.0780678466,-0.171876058,-0.2997308075,0.3796446621,-0.1070984155,-0.0961287543,-0.2705844045,0.1875978857,0.1234655604,0.0009863049,-0.1368592232,-0.1022856161,0.0375525877,0.1670027226,0.3814094961,0.2300312668,0.0853392407,-0.0208789576,0.3448313177,0.1670389175,0.0647967011,0.0143180341,-0.0658063218,0.0711300969,0.2488198876,0.2583678663,0.3715967834,0.0479277149,0.2792146206,-0.168510586,0.3713317811,-0.1328787804,0.0269255508,-0.0919242874,-0.0579607002,-0.2032177448,-0.212809518,-0.0640657544,-0.0018405407,0.2651608586,0.1012498811,0.0838894993,-0.4108765125,-0.1133110449,-0.1967409104,-0.1337637305,0.3249425292,0.0046006143,-0.5532537103,-0.0559352674,-0.1994499266,0.2022299618,-0.0330775864,0.2091742456,-0.2046139389,0.2034395486,-0.0310076401,-0.1547106653,-0.1921057105,0.2037311196,0.0917624161,0.012129914,0.172627151,-0.572696507,0.1163188443,-0.0581245758,0.2865843177,-0.2865249515,0.0354089662,-0.1741320044,-0.0007085068,0.1235657558,-0.438562125,0.0518097021,-0.2357430607,0.0070025581,-0.1307341754,-0.2610491812,-0.2307140678,-0.0324152783,0.1210667491,-0.0235937536,-0.2637910247,-0.0945184678,-0.1961842626,-0.1584451497,0.472394377,0.1527373642,0.2521524131,-0.0147098089,-0.2428223938,-0.4192585647,-0.1423542202,0.282173425,-0.3960949779,-0.5978460908,0.3621937335,-0.4948404431,-0.4088158011,0.0648189485,0.1902253926,0.0873688236,0.4965362251,-0.0206678938,-0.0327208899,-0.1303581595,0.2110580355,-0.3021684587,0.149070859,-0.018323062,-0.0033314307,0.1987172216,-0.0616129376,-0.2049269974,0.0157702249,-0.0018585279,-0.2435672134,0.1733835489,0.1804015487,0.0368578285,0.6638664007,0.4006105363,0.3711779118,0.3447342217,0.2460648566,0.6566482782,0.1277868301,-0.0983544216,0.1328796297,0.0307422653,-0.1344203651,0.1476204395,0.0507361218,-0.2245369554,0.0086268596,-0.0305153336,-0.2527899444,-0.4145503938,-0.1356714815,-0.2857632935,0.0906813145,0.0120840399,0.2275615335,-0.2544756532,-0.199630037,0.0157246478,0.3498134613,0.1901768893,0.2888481021,-0.5011578202,0.0322066732,-0.3375477195,0.4071542621,-0.1128841266,0.4579398632,-0.2347192019,0.1226631328,0.2193189412,0.3533581197,0.420682162,-0.4260231853,0.219175607,0.024685597,0.1591666639,0.2435566485,-0.1874754429,0.4024350047,0.4582351744,0.0386999324,0.1010611653,0.1391744614,-0.1269146949,-0.4196950793,-0.2121928483,-0.0887371004,-0.3115251958,0.0865531191,0.1946702451,-0.1184443682,-0.0838521421,0.2580322921,-0.3196932077,-0.1928002983,0.0135650365,-0.2359616905,0.2144568712,0.0882400572,-0.023359593,0.3749974966,0.029077895,-0.013873456,0.3818997443,0.0074323742,0.1824667752,0.4704225361,-0.2757956982,0.2875595093,-0.1840997487,-0.2768234611,0.4023550749,-0.020773178,-0.3571340144,-0.1568695456,0.1743985862,-0.1930555403,-0.267413497,-0.0111442972,-0.1028448343,0.0749542564,-0.280110985,-0.6301468611,0.202337876,0.5458945632,-0.3691300154,0.2700288296,-0.3061341643,-0.3580110371,0.21224612,0.1005997881,0.7887321115,-0.246486932,0.3536963165,-0.4175288975,0.1269745529,0.0216056891,-0.4390439689,0.0716993138,-0.4734859467,-0.3681210577,-0.1136456206,-0.1103551611,0.0601077974,0.5340064764,-0.0449446924,0.097828716,-0.0706809238,-0.1043381467,-0.1764693111,-0.1874247491,-0.1608287692,-0.3270985484,0.2392128557,0.1140990406,-0.2312960625,0.0558354221,-0.0365257189,0.0121988924,-0.1867185831,0.1956736445,-0.1977063119,0.0684634075,-0.6160359383,0.0682230145,-0.1547408849,-0.0878569484,0.1309370697,0.415876776,0.2986324131,0.3483609259,-0.0136350468,0.0292691458,0.3476778567,0.2069432884,-0.2035660595,0.0342273265,0.2627515495,0.1758734584,0.0214601215,-0.250577718,-0.1745347977,-0.3691315353,-0.269172281,0.2384556234,0.0540877171,-0.2971132398,-0.0545926429,-0.1923466176,0.0965938643,0.1019243449,0.0662560314,0.5287200809,-0.0330315083,0.5756495595,-0.3444020748,-0.3904916644,0.0317518488,0.5647519231,-0.0669802129,-0.2368167788,0.1545802206,-0.3134378791,-0.2485006005,-0.1570892483,0.2500908375,0.1942848265,-0.1230061203,-0.0877433866,0.0684736893,-0.0695993155,0.1941007823,-0.4771358371,0.0461783186,0.1764192283,0.1300098598,-0.3780312538,-0.4907993376,-0.3061333895,-0.3063032329,0.2266625166,0.3198889792,-0.6408579946,0.1010532826,-0.016023973,-0.1997380257,-0.2347139269,0.07675495,0.0967479795,0.1776961684,0.0467801057,-0.0919493288,-0.202885747,-0.0703608096,-0.1117065027,-0.0533863157,-0.1622434705,-0.0469911173,0.1502444148,0.1753890812,-0.3635746539,0.1325892806,-0.4973804653,-0.1254298389,-0.4760818481,0.0602506399,0.3147101402,0.1937381774,0.3190412223,0.1800258458,0.0004236259,-0.2174244821,0.1761341244,-0.1847559363,-0.133519575,-0.2566711903,0.249626711,-0.2918269634,-0.0442203954,0.0959338918,-0.0525663868,0.1684160382,0.0968466476,0.10871429,0.0077954023,0.0210091639,0.0825725645,0.2972578406,0.0636649579,-0.1853008121,0.0696976408,0.3839071691,0.0971770138,0.1314060837,-0.2571958303,0.3448217809,-0.1068820208,0.0000315114,0.3978754878,0.1617184281,-0.2894000113,0.1561603248,-0.1030877829,0.1565568596,-0.2774780989,0.3243854046,-0.0405135751,0.0415859744,0.1016231179,0.3367091715,0.2523678243,0.0584526621,0.3339830935,0.1865362823,0.1996393055,0.3867455721,-0.3236469924,0.1058689207,-0.7008692622,0.0736417323,-0.0000704297,-0.0661127344,0.3008562922,0.1088775471,0.0256843604,0.0033407013,-0.288598299,-0.0088926032,-0.127847001,-0.1557437778,-0.1520419866,0.0488087349,0.1350798905,0.0352474675,-0.234405458,-0.1990815401,-0.2742220163,0.1608579904,0.258552283,0.0971265659,0.1773428023,0.0228627864,0.1148761883,-0.0264974888,-0.2395642251,-0.0577448681,-0.513641119,-0.0990762711,0.345736742,-0.0201044697,0.0873502791,0.151155144,0.1820172518,0.1513983458,0.3052732348,0.1070546806,-0.2939303815,0.2839918733,-0.0807648599,-0.2318909466,0.1009294391,-0.0029200364,-0.0153888101,-0.1272793561,-0.1417843401,-0.0105079142,0.0553012304,-0.0667922497,-0.1364723295,-0.439178139,0.1145023406,-0.0423414707,0.1652352661,0.0423086546,0.4903941751,0.0976867676,0.2387545854,-0.4316917062,0.0201406553,0.346418798,0.2648150921,0.3824138641,-0.0085255746,-0.0242716521,0.0564008914,-0.4088501632,0.137702778,0.2087855488,0.3114922345,0.019945154,-0.2332526445,0.1249171421,-0.0681980327,0.5730346441,-0.247541368,0.2576634586,0.3815858662,-0.2472929806,-0.3133666217,0.200939402,-0.0212376416,0.075151518,-0.3698140383,0.3097246885,0.061447341,0.0272919945,-0.1679833829,-0.1955737174,0.1988150179,0.5369480848,0.4967286289,0.1937572807,-0.0033436278,0.0288407914,0.4734125137,-0.3762740791,0.1190755591,-0.2065241933,-0.030942278,-0.1803783625,0.0247025918,-0.5975059271,-0.2022420168,-0.3634684086,0.0467547737,-0.1567855924,0.1914337724,0.0811972544,0.1009824127,-0.1009099633,0.0274032727,0.3133312166,0.2561118603,-0.1445178688,0.3793402612,-0.3996711671,-0.1239276305,0.2655777931,-0.2716206908,-0.1795497984,-0.0824190602,0.357558459,0.1435213089,-0.1771571338,-0.0216587037,0.0905536786,0.3699934483,-0.4360894263,-0.1380515546,0.0460883193,-0.3514821231,0.0730111152,-0.1463425756,0.4525746107,-0.0052706501,-0.2306556851,0.0680167377,-0.287799418]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"Ideally it should be a parameter in `load_dataset` but I'm not sure how important it is for the users (considering only important things should go into `load_dataset` parameters)","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":28,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nIdeally it should be a parameter in `load_dataset` but I'm not sure how important it is for the users (considering only important things should go into `load_dataset` parameters)","embeddings":[-0.1360274553,0.2357687652,-0.1311270148,0.0613821335,-0.0553654358,0.0940652415,0.3185192049,0.1379301399,-0.1731739938,0.0212598611,-0.0261565167,0.0268607009,0.0603254214,-0.258431375,-0.040404167,0.0852189437,0.1406024247,0.04694812,0.041757457,-0.0636764914,-0.1245545372,-0.1261472106,-0.0492923558,-0.0956183597,-0.2034237981,-0.1365754455,0.3221455812,0.2506717741,0.0373609588,-0.1135130897,0.2764498889,0.5581905842,0.1219631061,0.0878672078,-0.0001150559,-0.1816449165,0.0381917693,-0.132101208,0.0325187519,0.1941934079,-0.1203788891,-0.1345284879,-0.121626541,-0.1587391049,-0.1057848781,0.1650481522,0.1044892743,-0.6469731927,0.446066469,0.1828893572,0.1859185249,-0.2256558686,-0.3946871459,0.0816635936,0.2488645464,0.1016989201,-0.146209836,0.1121084243,0.2739817202,-0.1514118165,-0.0044227303,0.0000063964,-0.3130864203,0.0512259938,0.1830156296,-0.1587805748,-0.2716639042,-0.5145754218,0.1429129094,0.0730130374,0.7643355131,-0.1047453284,-0.4819129705,-0.1847703159,0.022310175,0.2133846879,0.3478791416,0.297359556,0.0136778569,0.3095957637,-0.4908583164,-0.0657006428,-0.1992785782,-0.0027754772,0.1880514473,-0.0514002778,-0.0381643884,0.1153752133,-0.0052998574,-0.0531306751,0.140549764,0.1195211336,-0.3659527302,0.3780120015,-0.1778173,0.0718212277,-0.1930512041,0.4293599427,0.233318612,0.3262350559,0.2667450607,0.2092345506,0.082654424,0.2248352766,0.0094386162,0.1519308239,0.4054580331,0.2742932737,0.5419009924,-0.3882970512,-0.0662724227,0.0290870741,-0.0284643751,-0.392372787,0.1260011345,0.2452686727,0.0128252367,-0.2103177756,0.1821021885,0.0922434926,-0.0493150987,-0.2737068236,0.030683862,0.3166087568,0.0591574609,-0.0872472674,0.007686303,0.1134036705,0.0246478841,-0.404242903,-0.0034629304,-0.375095129,-0.358987838,0.464140743,0.4577978253,-0.2448268384,0.0782258585,0.2190753371,0.1637288928,-0.2404778898,0.3377048671,-0.049982138,0.3577246964,0.4029445648,0.2828168273,0.3107583225,0.038994424,0.0398122482,-0.0994031802,0.3312436044,-0.3699700236,-0.1543810666,0.2854832709,0.0420134701,-0.0352834612,0.2946747243,-0.4751756489,0.1014180183,0.5583773851,-0.1506776959,0.3145248294,0.2650932074,-0.221911639,-0.450276494,-0.0772985369,0.3010312915,-0.1203651875,-0.047724802,-0.0348954834,-0.0149976118,-0.0078381952,0.2800571322,-0.0775766149,0.2429604232,-0.0706208944,0.1884447485,0.5337375998,-0.0477489345,-0.5750143528,0.1883643866,0.18891491,-0.1798315048,0.1891258955,-0.0152316205,0.1123888269,-0.0776226074,-0.066791825,0.3543637693,0.2272964269,0.2602181435,0.1782481372,0.1146014035,0.0456520505,-0.0540024154,0.0377732851,0.3933876157,0.3831637204,-0.3089499176,0.1788147837,-0.2511121333,0.2620522082,0.4649666548,0.0307009146,-0.1328852326,-0.007730193,-0.0465990156,-0.4959963262,0.2730861902,-0.1101468205,-0.4933682382,0.0916128308,-0.4283303022,-0.0183198694,-0.2726962566,-0.144802928,0.0064639733,0.0430833288,-0.0094614532,-0.0796331167,-0.3605632484,0.1758659482,0.4297905862,0.2050468177,-0.0376983695,-0.1288122982,0.0865142345,-0.0465494543,-0.1004960909,-0.2174246162,-0.1228889525,0.2693600357,0.0302325543,-0.2796184421,0.2872935832,0.1608500928,0.0764014423,-0.1081747934,0.2295518667,-0.0259981789,0.0941333771,0.0347965993,-0.2608456314,-0.0801375732,-0.1637900621,-0.2425535619,0.3675415218,-0.1197711751,-0.0849228129,-0.2220300734,0.1864243746,0.1616752595,0.0087417401,-0.1625157297,-0.0869642273,-0.0125953937,0.1924944371,0.4017890096,0.2435413152,0.0101195909,-0.0242001936,0.3577238321,0.1539075077,0.0820484459,-0.0283465497,-0.0804598853,0.0578168258,0.2434148937,0.2833217382,0.3093886673,0.0471605845,0.2991295755,-0.2422782183,0.3165649474,-0.1635980755,-0.0105387066,-0.1063483804,0.0025095716,-0.1118765846,-0.1996194124,-0.0505079739,-0.0126420828,0.2895021737,0.1080947444,0.0564978309,-0.3899313509,-0.0709529445,-0.2065171748,-0.1147947088,0.2945739329,-0.0922339857,-0.5483307242,-0.0536429621,-0.2354534119,0.229098767,-0.0110396519,0.1633169949,-0.2583620548,0.2286836952,-0.0755812228,-0.1648586839,-0.250680536,0.2484352887,0.0460659824,-0.0024244073,0.1846737862,-0.5759427547,0.1644121408,-0.0296029188,0.3323052824,-0.251352787,0.0949379206,-0.1534491479,0.0210907143,0.1508962661,-0.4861028492,0.1240450516,-0.1586667597,0.0218377467,-0.1734283417,-0.1874536276,-0.1731722504,-0.0213988591,0.0278458856,-0.0310983285,-0.3103387654,-0.1054449528,-0.1327475458,-0.1433522105,0.3912353516,0.1155995205,0.2324867845,0.0079756351,-0.2187782228,-0.4117312133,-0.2179618627,0.3084759712,-0.3918885887,-0.6182736754,0.3513796329,-0.5044491887,-0.4009488225,0.047519844,0.2151391804,0.1240807474,0.514362812,0.0238785278,0.0336519517,-0.1428749561,0.180702135,-0.3144257069,0.0916597396,0.0156757534,0.0490067266,0.196096316,-0.0667604357,-0.1699825525,0.054724358,-0.0005291421,-0.2220746577,0.1749123633,0.1327409446,0.0634817556,0.7849994898,0.3883143961,0.3106006384,0.3129928112,0.2840475738,0.6152814031,0.134921506,-0.0927729607,0.1380268633,0.0194206722,-0.0925643444,0.1346922815,0.02323981,-0.221025601,-0.0166213289,-0.0366627797,-0.3092912734,-0.3717314899,-0.2026934028,-0.3159177005,0.1572172046,0.0642259866,0.1235330179,-0.2987570465,-0.1908789277,-0.0169987269,0.3514697552,0.2301059514,0.2705020308,-0.3379973769,-0.0295762029,-0.3026228547,0.3595712781,-0.0352892838,0.4659916162,-0.2662653327,0.210681811,0.1673325747,0.3745906353,0.4072617292,-0.426346153,0.2426746339,-0.0541127101,0.2029170692,0.2937229872,-0.2115208656,0.4624210298,0.4730280936,-0.0399512239,0.0854400024,0.1279440224,-0.174868077,-0.4338357151,-0.2216638327,-0.1441207975,-0.2356077731,0.0851387605,0.1846866459,-0.0517425612,-0.1039298475,0.3252375126,-0.3153504133,-0.1791091859,-0.0511318743,-0.1827656925,0.1800732017,0.0917650983,0.0356340893,0.3379076123,0.0901944041,-0.0477944985,0.4214046299,0.0411990844,0.1237611845,0.3518939018,-0.1940187961,0.3321698904,-0.1683047116,-0.2559118569,0.4673245549,-0.0353711545,-0.432164073,-0.1290151328,0.1249464825,-0.233409524,-0.2891232073,-0.0210473221,-0.0685048625,0.005854588,-0.2907117307,-0.637455821,0.2226540148,0.5671569705,-0.3972543478,0.3289411068,-0.3693507016,-0.3797777295,0.1075255498,0.1092871204,0.7499828935,-0.2078108191,0.3130450547,-0.4576827586,0.1495295465,-0.0146856075,-0.3989052176,0.0419782661,-0.41422382,-0.3809765279,-0.1449602097,-0.1103521958,0.0486164168,0.4917732477,0.0318699405,0.1424922943,-0.1179925576,-0.0948824584,-0.1461008638,-0.1570739001,-0.0780510008,-0.3369984031,0.2686450481,0.1036930904,-0.2924657464,0.0383192115,-0.0258427244,0.00235913,-0.1695407182,0.2391934246,-0.2151148766,0.0428009406,-0.5830400586,0.0276745446,-0.1792512238,0.0034086679,0.1712263525,0.384940207,0.2264019996,0.3392247856,-0.0134629458,0.0257634185,0.2369792908,0.1699475646,-0.2879479825,0.1346048266,0.2897391915,0.1768381,0.0375643224,-0.1905825585,-0.1705692261,-0.296084106,-0.3060022593,0.204854995,0.0204059035,-0.2292106897,-0.041102279,-0.1447500288,0.1306892484,0.089327544,0.0652522594,0.508628726,0.0084815128,0.5539218783,-0.2621820867,-0.3646312356,0.050299976,0.5522649288,-0.0477820672,-0.2210428119,0.1484201849,-0.3808464706,-0.2166649848,-0.1491111964,0.2213981748,0.1041220725,-0.1133882552,-0.0702696815,0.0528913476,-0.0642354786,0.17087771,-0.4494755864,0.0465703532,0.1083830073,0.1382250786,-0.3907188177,-0.4287422895,-0.2864274681,-0.3578280807,0.2063931376,0.3053347468,-0.5955711603,0.1072509512,0.0413543321,-0.1910178065,-0.2428407222,0.066580154,0.0913419947,0.2466428876,0.1041267142,-0.1178682297,-0.1624155939,-0.1032345444,-0.0816647783,-0.0287347045,-0.1598639786,-0.09342926,0.157121256,0.1877707839,-0.2460125387,0.0820715055,-0.4838794172,-0.1269897074,-0.4443708658,0.0597611368,0.240121752,0.2697981,0.3521519899,0.2059805244,0.0279868077,-0.1935165972,0.2300647199,-0.1304266453,-0.167560786,-0.25440377,0.2528738678,-0.2788145542,-0.1030370519,0.0131670088,-0.0807305798,0.1769057661,0.0376387239,0.0823836625,0.0170303434,0.0768730789,0.1447359473,0.2402615249,0.0273305662,-0.134866327,0.0675533861,0.3256235421,0.0825691447,0.1661927402,-0.2373446077,0.3468434811,-0.1327026337,0.0176859256,0.3570058048,0.210067302,-0.2102587968,0.1317542195,-0.1489014477,0.0954338238,-0.3153632283,0.3432016373,-0.0471954718,0.113983959,0.114441976,0.4225999117,0.2717164159,0.0656474903,0.3585306108,0.2171221226,0.2283774912,0.3618634641,-0.3181797564,0.0836165622,-0.698356986,0.1138347983,0.0265255645,-0.0405231379,0.3049774766,0.0828140974,0.0733668879,-0.0011718252,-0.2486261129,-0.0329704359,-0.1235062778,-0.130683735,-0.2032732368,0.0106207998,0.0806914717,0.031384103,-0.2732295096,-0.1574331224,-0.2429824024,0.2137712687,0.2132885605,0.0428232662,0.2503815889,0.0251998547,0.150156185,-0.0311449394,-0.1700297743,-0.0418196917,-0.5771830082,-0.1135220081,0.2809130847,-0.0252453219,0.0186533555,0.1169376224,0.1724523604,0.1115639955,0.3579921126,0.0717485994,-0.2909567058,0.3410485387,-0.0630935803,-0.2622015476,0.0715565234,0.0119651128,-0.0319208875,-0.1076718196,-0.1233387068,0.0585291907,0.0732382461,-0.0332730487,-0.1695599556,-0.4606755972,0.0883854628,-0.0230554156,0.1516735107,0.0533864461,0.5730985403,0.0757677108,0.2085248828,-0.4813012183,0.0242604148,0.3927841485,0.3333714902,0.3398482203,-0.0345548242,0.0203795061,0.0488824472,-0.4188895226,0.1595375687,0.2577327192,0.4444583952,0.048987519,-0.1877273321,0.0786524862,-0.0804445967,0.5692911148,-0.2544927895,0.3134144247,0.3377063274,-0.1995334029,-0.3216018677,0.1196496338,0.0005236734,0.0797114,-0.3342708349,0.2815948129,0.1148243919,0.0373959988,-0.2328619361,-0.2309465259,0.204328537,0.5594170094,0.4742109179,0.1728249639,-0.0184819959,0.0043155737,0.4483694434,-0.3441611528,0.1336248815,-0.1462295502,-0.1085249782,-0.2192159891,-0.0158677306,-0.5714678168,-0.1290591061,-0.382949084,0.0065428857,-0.1274741441,0.2369356602,0.023130253,0.0051634382,-0.1058145463,-0.0039867521,0.299020648,0.2356524169,-0.177802667,0.3501921594,-0.3655160069,-0.0733126253,0.290748179,-0.2873263955,-0.2061223686,-0.0213183165,0.3530815542,0.1035550609,-0.162448734,0.0034141566,0.1304533035,0.3935117126,-0.4170918465,-0.0632619932,0.0756741837,-0.4133792222,0.1053446382,-0.1192170903,0.4856008589,-0.000268186,-0.1820430458,0.0550932325,-0.233718887]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"I think it's fairly important; for context, our team uses a shared file-system where many folks run experiments based on datasets that are cached by other users.\r\n\r\nFor example, I might start a training run, downloading a dataset. Then, a couple of days later, a collaborator using the same repository might want to use the same dataset on the same shared filesystem, but won't be able to under the default permissions.\r\n\r\nBeing able to specify directly in the top-level `load_dataset()` call seems important, but an equally valid option would be to just inherit from the running user's `umask` (this should probably be the default anyway).\r\n\r\nSo basically, argument that takes a custom set of permissions, and by default, use the running user's umask!","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":123,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nI think it's fairly important; for context, our team uses a shared file-system where many folks run experiments based on datasets that are cached by other users.\r\n\r\nFor example, I might start a training run, downloading a dataset. Then, a couple of days later, a collaborator using the same repository might want to use the same dataset on the same shared filesystem, but won't be able to under the default permissions.\r\n\r\nBeing able to specify directly in the top-level `load_dataset()` call seems important, but an equally valid option would be to just inherit from the running user's `umask` (this should probably be the default anyway).\r\n\r\nSo basically, argument that takes a custom set of permissions, and by default, use the running user's umask!","embeddings":[-0.1335193068,0.2638700008,-0.0887797624,-0.0136506241,-0.0483198538,0.1047770232,0.3552216589,0.1191613078,-0.1658600122,0.0378849208,-0.106523864,-0.0248673074,-0.0191920474,-0.3137134314,-0.1179799512,0.1778150201,0.1263326705,0.0858631954,0.0602515973,-0.108570978,-0.1009776145,-0.1389106661,0.0054339189,-0.1248307154,-0.2158824205,-0.1946879774,0.2352733314,0.2853993177,0.0203850195,-0.0933629945,0.3288930357,0.5935889482,0.108140178,0.1202231944,-0.0001126702,-0.1727904677,-0.0196036771,-0.1369560361,-0.0687135532,0.1222948506,-0.0050359312,-0.1052324176,-0.080206342,-0.1873364002,-0.0884159729,0.1841935962,0.1439650208,-0.7255420685,0.4894111156,0.095662944,0.1734694839,-0.1860495359,-0.3902597427,0.0528155975,0.1976868361,0.0250731856,-0.1162096336,0.1660548002,0.3335852027,-0.1733828038,-0.0796371624,-0.0047566737,-0.3949322701,0.0934961587,0.1791542917,-0.168472603,-0.2764598429,-0.4620153606,0.091608122,0.159108296,0.6701866984,-0.1889321804,-0.4884358943,-0.2769802809,0.0945892707,0.2630771101,0.3322439194,0.2956201732,-0.070246771,0.294301033,-0.4962939024,0.0249761268,-0.1143071577,0.0711589605,0.2440434396,-0.0776517168,-0.0221331306,0.0635323748,0.0589097142,-0.0742979869,0.1298757792,0.0513516665,-0.2814569473,0.3380032778,-0.144187659,0.0683853924,-0.1546314061,0.3203953505,0.2561030388,0.3386801183,0.2594955266,0.2057443708,0.0260845218,0.2732971311,0.0364703201,0.1471388489,0.3354699612,0.2981513143,0.5131576657,-0.3563105166,-0.0789603442,0.0803651586,0.0015175763,-0.3296959102,0.0532517582,0.2754203975,-0.0348209627,-0.1684566885,0.1972961426,0.043387752,-0.1375850886,-0.2175005972,0.0006882965,0.2218636721,0.143580094,-0.0419343412,0.050697498,0.1568997353,-0.0298380684,-0.3474836349,-0.0023438695,-0.3836150765,-0.3437673748,0.4538429081,0.502584815,-0.1503937244,0.0259815305,0.2286189497,0.1914967,-0.176214993,0.3555212319,-0.0703046024,0.369527638,0.2953026295,0.310430944,0.3790474832,0.1347994059,-0.0384794399,-0.1857957095,0.2925340235,-0.3189618886,-0.1988264769,0.3306159377,0.0252362601,-0.129405424,0.3491237462,-0.435300976,0.0539279692,0.4892047346,-0.1339069307,0.3949003816,0.2284995466,-0.1604258567,-0.4301351607,-0.097062245,0.3187615871,-0.0068246857,-0.0861124545,-0.0970914438,-0.0044264956,-0.1046250388,0.334502399,-0.0875778571,0.1819878221,-0.0739462227,0.1962854713,0.4308474958,-0.1192560419,-0.5625243187,0.085646227,0.2081537545,0.0200194307,0.1798261851,0.0332902782,0.0928128958,-0.1795361489,-0.013073747,0.3748737872,0.1747640222,0.2336510718,0.1346783638,-0.0211188998,-0.0107621932,-0.0737503245,0.016223602,0.2922054827,0.3899533749,-0.2811438739,0.233219713,-0.3469516337,0.2264998257,0.3875041902,0.0886686966,-0.0321269333,-0.0775921419,-0.0125407958,-0.4649496973,0.3139138222,-0.019517919,-0.3743110895,0.0781528056,-0.4427528679,0.1527733952,-0.3058959544,-0.078541331,-0.0590090863,0.0684402958,-0.0169347655,-0.0626724809,-0.2871268094,0.1386516839,0.397677809,0.1569042951,-0.0005136139,-0.1833364666,0.1158879325,-0.0207362417,-0.0901093408,-0.2096593976,-0.0049833176,0.2107448727,0.0538748764,-0.2475261539,0.3181375861,0.1360021085,0.0878940597,-0.2036231458,0.3133203089,-0.0595914982,0.143978402,0.0800896585,-0.2660570145,-0.0818465874,-0.1320270598,-0.2821069062,0.4214651585,-0.1982382387,-0.0984292552,-0.2274123132,0.1985201389,0.1358624548,-0.0531105921,-0.1689100564,-0.1018535942,0.0592443943,0.2333414853,0.4352440834,0.2416633219,0.0863659605,-0.0734826997,0.2969388664,0.1667918116,0.1391642541,-0.0248610619,-0.1173865423,0.071324572,0.2680752277,0.3193115592,0.2627224326,0.0815075636,0.3029861748,-0.1794038266,0.3617700934,-0.1280270666,-0.0533135124,-0.063828893,-0.0096192937,-0.1516234279,-0.2004453838,0.0383140184,-0.0102994144,0.2832241654,0.0139391981,0.0310584754,-0.3533719182,-0.0851112679,-0.1569666117,-0.1439567208,0.3133104742,-0.0409684479,-0.5186011195,-0.0337253846,-0.2119198143,0.2901796699,-0.0679176375,0.2043979615,-0.2976181209,0.272285521,-0.1464714557,-0.210093379,-0.1833209246,0.1580687612,0.049779065,0.0190461595,0.3156298101,-0.599206388,0.1461791694,-0.0014800217,0.258728683,-0.1778861582,0.071714893,-0.1856780797,0.0394682437,0.0848226175,-0.4404526949,0.0827055499,-0.1666044593,-0.0552184246,-0.1495183706,-0.2587300539,-0.1490297318,0.0165517759,0.0491509512,0.0126490491,-0.2662484646,-0.0743446648,-0.1646046489,-0.1609423608,0.506388247,0.1258137375,0.2180757672,0.0177706312,-0.2047579288,-0.443353653,-0.2076893151,0.3096786439,-0.3285373151,-0.5994923711,0.3129225075,-0.439291507,-0.2962847054,-0.0024911431,0.117717512,0.077165626,0.508112669,0.0141370408,-0.099913016,-0.1493712813,0.2239403427,-0.226396963,0.1315018386,0.0092313997,0.080677405,0.1833283007,-0.0216768179,-0.1074555442,0.0599435456,-0.0806834474,-0.3560809195,0.1849569231,0.044839479,0.2210550457,0.688346684,0.4353192449,0.3087536693,0.1947027594,0.3311053216,0.6988666654,0.0714703426,-0.096798569,0.1601403505,-0.0599365234,-0.1391473413,0.1530293375,0.0268654078,-0.1599033475,0.0362283215,0.0209258869,-0.2974901497,-0.5224307775,-0.1170669124,-0.3631197214,0.0746197924,0.0841091946,0.0950958654,-0.3080649078,-0.229477644,0.0215059984,0.3599597514,0.1409596801,0.353849411,-0.4579848051,0.0931677967,-0.2501485646,0.3669549525,-0.0359292142,0.4248611629,-0.1644567102,0.1730359346,0.1208910048,0.3693688512,0.2832073867,-0.3664070666,0.2412290871,-0.0656469539,0.1311780065,0.30794397,-0.2037863731,0.426363349,0.3683973253,0.0024619426,0.0910819322,0.0815369263,-0.129903689,-0.4899196625,-0.1961275488,-0.0934181511,-0.1836797893,0.179394111,0.1602242589,-0.1438174248,-0.1086396724,0.2946444154,-0.3695813119,-0.2043013424,0.0676582307,-0.2385042757,0.2018124908,0.1579188704,-0.0005507361,0.2536374927,0.0071781571,-0.0459017567,0.5038197637,-0.0325040407,0.2236606926,0.3921177983,-0.3262414038,0.3009204566,-0.2534970641,-0.2326678634,0.3481919169,0.0109600425,-0.3378420472,-0.1636282355,0.1595122963,-0.2159738541,-0.3587737381,-0.0627601221,-0.1337773055,0.024638053,-0.2204288244,-0.7126753926,0.192399919,0.6057905555,-0.39554739,0.280661881,-0.2725962102,-0.3773383796,0.1679943651,0.1162307113,0.7774208188,-0.2349472493,0.4384993017,-0.4347285628,0.1221261099,0.0675324798,-0.5041333437,0.0852131322,-0.4016420543,-0.3306939304,-0.1211037785,-0.1156786233,0.0600077175,0.4919955134,0.0954435989,0.1253556758,-0.0263798777,-0.0819223225,-0.113021642,-0.2094170153,-0.157378599,-0.3598347008,0.2407252938,0.1178229526,-0.2239521742,0.0005098097,0.0178837441,0.0252515245,-0.1955073029,0.1852475107,-0.2485401481,0.0230550002,-0.6235012412,0.0564118661,-0.1469193995,0.0111744832,0.1415770501,0.3558231592,0.2799707055,0.3346356452,0.0078630373,0.006832046,0.2641917467,0.2242465764,-0.2496599108,0.0437003374,0.3597234488,0.2290325761,0.1369515061,-0.1850254238,-0.1416846663,-0.328404367,-0.3017079234,0.1812123507,0.083269842,-0.250002414,-0.0394959077,-0.1044537872,0.1008087173,0.1058467925,0.0466520861,0.4602094591,0.0222135372,0.5224806666,-0.383736968,-0.3255946934,0.054460451,0.5755968094,-0.0124589456,-0.2675655782,0.1389062554,-0.3579819202,-0.2112160623,-0.1237872243,0.2073870152,0.1584110558,-0.1332009882,-0.1082580164,0.0072026574,-0.1046754867,0.2162418365,-0.4664973021,0.1011894047,0.1262042373,0.2242450565,-0.2977729738,-0.3513051867,-0.2399708033,-0.2813026011,0.2754709721,0.3540195823,-0.5409365296,0.0350809991,0.051145941,-0.2095492929,-0.1660881042,0.0907016993,0.1038901061,0.2778420448,0.1022330672,-0.0826604366,-0.2200219333,-0.1028364971,-0.0692286566,-0.0679975599,-0.1563985795,-0.1039413884,0.1390435696,0.190046519,-0.3333748877,0.1476512998,-0.4402852654,-0.0977087766,-0.5249465108,0.0726314932,0.3600779176,0.2148393989,0.303367734,0.1483828723,0.1028189808,-0.2563403249,0.2047588825,-0.1927185953,-0.1573558152,-0.2231704146,0.2588319778,-0.2905417085,-0.0344234928,0.0884247199,-0.031647861,0.2469404042,0.0783759505,0.1219610572,0.0382645652,0.0328223482,0.1121715829,0.2635773122,-0.037371479,-0.1741434187,0.0290920716,0.3594093025,0.0863357931,0.1663937718,-0.221291855,0.3022618592,-0.1214046106,0.0444687083,0.4241456091,0.200565204,-0.2796059549,0.0410664678,-0.1848446578,0.1399807632,-0.2930706143,0.4141856134,-0.044921238,0.1124625057,0.0260012355,0.2568044662,0.2728781104,0.1806203276,0.3987114727,0.1678881347,0.2269147933,0.324650377,-0.3882226646,0.109043628,-0.7499385476,0.1160010695,0.0643169582,0.0552677251,0.292470783,0.004791656,0.0324178599,0.0339511745,-0.2421956211,0.0731698051,-0.1951654404,-0.10713052,-0.1231279895,0.0984254107,0.0617454723,-0.0580442548,-0.2758308053,-0.2101801187,-0.2737853229,0.222896561,0.2543891072,0.107631363,0.1795518845,-0.0354109183,0.1680822223,-0.0305830725,-0.2526371181,-0.1355951577,-0.4439837635,-0.1797692925,0.2402365506,0.0428735055,0.1080204546,0.0699903592,0.2003017217,0.1465199888,0.3104406297,0.1460891515,-0.3191908002,0.2480013818,-0.063952677,-0.2943786681,0.0317078382,0.0325664654,-0.0388120376,-0.0634450167,-0.1798167676,-0.0688497052,-0.0215035975,-0.0553789996,-0.0878477022,-0.3917623758,0.1457397491,0.0017111637,0.2160775512,0.0725500658,0.5099499822,0.1127278209,0.2034332305,-0.4516061246,0.0393084884,0.477519691,0.2224102169,0.3928313851,-0.1754790395,0.0194799267,0.0481335782,-0.4098705947,0.1693513393,0.1614996344,0.3672335148,-0.0502838418,-0.2409628779,0.213797763,-0.0342042521,0.4339053333,-0.2005824,0.263514787,0.3118850589,-0.1893591583,-0.3202570081,0.1547316909,-0.1253811121,0.1053415984,-0.3782761693,0.1674528271,0.0702300966,0.0624998771,-0.1400252432,-0.2013217509,0.2447036952,0.5015896559,0.4458535314,0.2797785401,-0.0513317138,0.0455646031,0.3394953907,-0.336224407,0.1756870598,-0.192013368,-0.0833615214,-0.2430045009,0.009313154,-0.6937606335,-0.1220569387,-0.3894574046,-0.024660876,-0.1187719405,0.1842211932,-0.0180332121,0.0751315355,-0.1444794983,0.0782491937,0.3316343427,0.3493755162,-0.1586183012,0.4514397681,-0.4411894083,-0.0188557412,0.2554280162,-0.3255093992,-0.1499676257,-0.0959344506,0.3641744554,0.1132619902,-0.1313416809,0.0271323677,0.130466193,0.3607616425,-0.3851851225,-0.1099031791,-0.0116720796,-0.3629735112,-0.0249140076,-0.11872641,0.3934583068,0.0186428856,-0.2428622544,0.1357965022,-0.2689588964]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"Maybe let's start by defaulting to the user's umask !\r\nDo you want to give it a try @bhavitvyamalik ?","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":20,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nMaybe let's start by defaulting to the user's umask !\r\nDo you want to give it a try @bhavitvyamalik ?","embeddings":[-0.180294916,0.1989278346,-0.105565615,-0.0858242586,0.0109792566,0.0972137526,0.309679538,0.0719593242,-0.1649410129,0.087335676,-0.0806741789,0.0011490682,0.065246217,-0.265019089,-0.0587538555,0.1390801221,0.1544880271,-0.008077438,-0.0103487819,-0.0530851297,-0.1247114167,-0.0811096281,-0.0939109325,-0.0244817473,-0.3202103674,-0.2104613334,0.2709257603,0.2456599474,-0.1097432449,-0.1701467335,0.2531243861,0.5244038105,0.0950799882,0.1362226009,-0.000110292,-0.2203178853,0.0572102033,-0.098393552,0.0717844069,0.1836276799,-0.0081606368,-0.0578650422,-0.1526501924,-0.1333853602,-0.1902170628,0.2176801711,0.1347751766,-0.6735301614,0.3757870793,0.1801066101,0.2284765393,-0.1668886989,-0.2878263593,0.0412441492,0.2450849563,0.0536144041,-0.1640152931,0.0123872831,0.3500931859,-0.2160444111,-0.0395845622,0.033809375,-0.2767978609,0.0918716267,0.111241214,-0.0900154039,-0.2701267302,-0.5084251761,0.1664372385,0.0999982506,0.7544164062,-0.1295631379,-0.4987752736,-0.0465831421,0.1072057709,0.3296329081,0.3315171599,0.3367734551,-0.0377411358,0.3057036996,-0.5110291839,0.0716741085,-0.1549808383,0.0514568612,0.2104095519,0.0108153317,-0.0395894796,0.0690243095,0.0063470216,-0.1661574394,0.1708116531,0.1088744402,-0.3357993066,0.3499305844,-0.13664563,0.0234397966,-0.1685891002,0.399399519,0.2073740065,0.3033181727,0.2516773343,0.2496393174,0.1149909571,0.2281814367,-0.0468490906,0.1980661005,0.4191332459,0.2782429159,0.6045436859,-0.2715248764,-0.1065614894,0.014116386,0.0908902809,-0.3845100999,0.0430960469,0.3155744374,0.0347556733,-0.2312795371,0.2005331665,0.0753685832,-0.1098826006,-0.1989782602,0.0171883088,0.3396971226,0.1132361591,-0.0913886204,0.0945198312,0.1875855476,-0.025398057,-0.3473285437,-0.028197024,-0.3111818433,-0.3306752145,0.4159270227,0.4743276238,-0.1079856902,0.0150205903,0.2490666509,0.129945159,-0.1909890771,0.3414691091,-0.0159487408,0.4194329083,0.307403028,0.2932574451,0.2807782292,0.0849469826,-0.025000656,-0.1067662016,0.3163594306,-0.2679615915,-0.1825825572,0.3279302418,0.0949654803,0.0148497326,0.2611830533,-0.3177102506,0.0577969067,0.4880384207,-0.0669542924,0.4080171287,0.3107382655,-0.1341863573,-0.4094082117,-0.0629705489,0.2717763484,-0.0865582898,-0.008627655,-0.1093441993,-0.0392959155,-0.0511427484,0.3505493999,-0.0363037549,0.1679966748,-0.1379119307,0.2179732323,0.4348886311,-0.0785186812,-0.5814692974,0.0801319927,0.1961143017,-0.1239168346,0.1525450349,0.0180690605,-0.0493479185,-0.1270513684,-0.0112659466,0.3844030201,0.2174694538,0.2480631471,0.0529647544,0.0200448986,-0.0610317402,-0.0057231491,0.0757830292,0.3318834603,0.3329181671,-0.372689724,0.1883392632,-0.3001170158,0.2619217038,0.4566673338,0.1620623022,-0.0956024379,-0.0418221243,-0.0447195955,-0.374174118,0.250623405,-0.1500740051,-0.3351980448,0.0283068232,-0.5107872486,0.0638449118,-0.2787100077,-0.1423332393,-0.0906037539,0.1143662632,-0.0131134167,-0.0166821536,-0.2777712643,0.2108010054,0.3133116364,0.2674199343,-0.0196904503,-0.1041235477,0.1147962436,-0.1456022263,-0.1046613753,-0.1880643219,-0.057057485,0.2682104111,0.0339883827,-0.3065500557,0.3243804276,0.1456222236,0.0264429636,-0.1777470261,0.1663871408,-0.065391928,0.1011266187,0.038499862,-0.1685673296,-0.0843007192,-0.1399026662,-0.293335408,0.4097215235,-0.071134381,-0.1476326287,-0.2306483388,0.2377378494,0.1563982964,-0.0517028384,-0.0797408894,-0.1849278957,0.0297500379,0.1552098244,0.3590900898,0.1520941556,0.0944396034,0.0355229974,0.3280803263,0.1983678341,0.1446081102,-0.0338478051,-0.0787688568,0.1132239178,0.2217765003,0.3202686906,0.282559067,0.0687993839,0.2878521383,-0.1768565029,0.3473739922,-0.1926754117,-0.0030032764,-0.071838811,-0.0644502342,-0.1540563405,-0.1414109021,-0.0584209897,-0.0871267095,0.3273696899,0.0399385951,0.114734605,-0.3470172584,-0.1586537659,-0.1861213446,-0.2396656722,0.3650272191,-0.0057387589,-0.4932681322,-0.0867698193,-0.0944306254,0.1654313356,-0.1129116714,0.2546655834,-0.1431786865,0.191202417,-0.089322269,-0.0425658412,-0.2824662328,0.1512673497,0.0966491997,0.0732803643,0.2500422895,-0.5685628653,0.1663985699,0.0087674065,0.3142611384,-0.3341976702,-0.015175431,-0.1984585375,0.0717621222,0.0794694126,-0.4089492261,0.0436938331,-0.168653667,-0.0708964989,-0.059497308,-0.3048062623,-0.2399098575,0.0328295715,0.0480148643,-0.1145294234,-0.258240968,-0.0789977089,-0.1354624778,-0.1996629536,0.4575275779,0.126586616,0.179821074,-0.1231067181,-0.2027316093,-0.4167491794,-0.1443784237,0.2437559068,-0.3575887084,-0.6189984083,0.3860661685,-0.5305783153,-0.4261040986,0.0611994974,0.2009063959,0.111974813,0.4502743483,0.0272937976,-0.0719336197,-0.2201864719,0.169935137,-0.1490112841,0.1046857387,0.021952454,0.0574289933,0.0926088616,0.0207038708,-0.1879633367,0.0426694714,-0.0683081076,-0.2640701234,0.1049343124,0.1235696524,0.2320937216,0.6044672728,0.4137212932,0.3512723148,0.2618797123,0.323172152,0.5492525101,-0.0017727775,-0.1161274239,0.1286386997,-0.0301050916,-0.0633835122,0.1106949374,0.0715712979,-0.2163660675,0.0260057431,-0.0334104113,-0.2192905396,-0.5093737245,-0.1887930334,-0.3285442293,0.0789114982,0.0744798109,0.1274672002,-0.237672776,-0.1932530105,0.0826273263,0.317646116,0.1346713603,0.3178752065,-0.5666816235,0.1274270713,-0.3356295526,0.4427454174,-0.0463518538,0.4909642041,-0.2584385872,0.1571986973,0.2063754052,0.3330774009,0.3566008508,-0.5286798477,0.2724564075,-0.0511136353,0.11264541,0.2861071229,-0.2046379894,0.4098228514,0.4641554952,0.1529790312,0.1525813341,0.1273508966,-0.1262139976,-0.4393823445,-0.3390440941,-0.0751291141,-0.2391100526,0.1486616433,0.1858578324,-0.1596784741,-0.0376076177,0.3153556287,-0.3611548543,-0.1233719438,-0.0250221491,-0.281865567,0.2162919343,0.0639569685,-0.0931850597,0.2871546149,-0.0401855446,-0.0098045003,0.3608036339,0.0214157254,0.2246811986,0.3886467516,-0.2325308919,0.3059424758,-0.1744608581,-0.310908556,0.3567636907,0.0151983732,-0.3405252993,-0.152734682,0.1999453902,-0.1115405113,-0.2377289683,-0.1410805583,-0.1133169904,0.0559860542,-0.2763209343,-0.5705082417,0.1906406134,0.5891581178,-0.3231378198,0.2600392103,-0.23346892,-0.353117764,0.202607125,0.103410393,0.8043721318,-0.1996926963,0.3696449995,-0.4210781753,0.0299932249,0.0329407379,-0.4663682282,0.0739386901,-0.4748346508,-0.3609425724,-0.1139467806,-0.1361667514,0.074821651,0.5008012652,-0.0509088039,0.0705021992,-0.044662606,-0.1175938547,-0.124056831,-0.1433368176,-0.2436745018,-0.2804812789,0.2834151387,0.1546863168,-0.220759958,0.0418036021,-0.0105150407,-0.0082709845,-0.1930461228,0.2263318151,-0.2536785007,0.0976541638,-0.6266340613,0.1544733196,-0.1700222641,-0.0480081961,0.158940509,0.3637183309,0.3779087663,0.3805488646,-0.0493059754,0.1027276069,0.3861988187,0.1945714802,-0.2120798677,0.0180715546,0.2800189257,0.1179423481,0.0594936647,-0.293174237,-0.1358923614,-0.3070363998,-0.3386482894,0.1563060582,0.0740319118,-0.2711719871,-0.0631194338,-0.1831537634,0.0566377118,0.029595226,0.1210525632,0.4231938124,0.0002312948,0.5362323523,-0.3272691667,-0.3025859594,0.0578076243,0.6893190742,-0.0827621892,-0.1979092807,0.1402797401,-0.2300558537,-0.3200796545,-0.2083719224,0.1428576112,0.1729991436,-0.1417028159,-0.0427537225,0.0286193844,-0.1275181919,0.299700737,-0.4429406822,0.0804360434,0.104656212,0.1733718067,-0.3032200933,-0.4135520756,-0.2953553498,-0.2933428586,0.2246070057,0.2591468096,-0.5468912125,0.1559989899,0.1168904975,-0.2534261942,-0.1681212038,-0.0050445199,0.1027851552,0.1480888575,0.00083993,-0.1242502034,-0.1824801266,-0.0637661964,-0.0312931165,-0.0848485306,-0.2207484096,-0.0021496089,0.10942325,0.2109240144,-0.4119981527,0.1522824615,-0.3853232861,-0.1325986683,-0.4285097718,0.0475912206,0.3590772748,0.1528203934,0.3160887659,0.1928800493,-0.0775801316,-0.2047160715,0.1588991135,-0.1796808094,-0.2339083552,-0.2131631672,0.1666121185,-0.322132647,-0.0782366246,0.1922112852,-0.0624021925,0.1268669963,0.10941457,0.1152406111,-0.0662821382,-0.0342526771,0.0675665736,0.2566539347,-0.0771353915,-0.139917478,0.0049634501,0.3268656135,0.1283869594,0.1042007729,-0.3107629716,0.3365099728,-0.1364433765,0.0646822304,0.2991357446,0.1867587566,-0.3084298074,0.1326437294,-0.0844072923,0.1173594072,-0.3579932153,0.4131159484,-0.1350716054,0.0431742147,0.0543685593,0.3032656312,0.2324018031,0.0289036389,0.3037873805,0.2282189727,0.1596159935,0.307166338,-0.3268916607,0.1433589906,-0.6421842575,0.1418134421,0.0488688201,-0.0195025411,0.2297239006,0.1978284717,0.0500346012,0.0529764518,-0.3077777624,0.0020789392,-0.1514123827,-0.1449217349,-0.0856482759,0.0482414961,0.1191725954,0.0386532545,-0.2337671071,-0.2303375453,-0.1831656992,0.1543704122,0.2487840056,0.1142754853,0.1897891313,0.0508770123,0.0736705884,0.0393196046,-0.2107701451,-0.0765493959,-0.4362026453,-0.1638540775,0.259552002,0.0762883052,0.0896849558,0.1124024764,0.1307534575,0.2092396766,0.2867303193,0.027161153,-0.3119152486,0.306692481,-0.0135366498,-0.3385615945,0.1068586558,0.0755265206,-0.076826632,-0.0926587656,-0.1684389412,0.028938707,0.1210223362,-0.064037174,-0.1738271117,-0.4742428064,0.1610174328,-0.0720793158,0.1171806231,0.0234367941,0.4542984962,0.0079748854,0.2162086368,-0.5413377285,0.0501318388,0.4407142997,0.2609605789,0.4043426216,-0.1206955388,-0.0195330661,-0.070427388,-0.3925542235,0.1042101532,0.2170504332,0.425180763,-0.0046340963,-0.2288281173,0.1126690656,-0.0804098248,0.5458396673,-0.1988474727,0.206576243,0.404327035,-0.2793478668,-0.2090274543,0.2195939124,-0.0404683948,0.1201099008,-0.359300226,0.2756617963,0.1133943498,0.1277196556,-0.1560972631,-0.131435141,0.1653334796,0.5161165595,0.423601985,0.2357954234,-0.0478289723,-0.0321175456,0.4266489744,-0.3159963489,0.1327674091,-0.1734093875,-0.0589500852,-0.108484067,0.0971880332,-0.5752668381,-0.2216766924,-0.3395993412,0.0396896265,-0.1851831526,0.264490366,0.0461133532,0.1106877849,-0.1216788962,0.0893349051,0.2738634348,0.2784029543,-0.178467378,0.3664804697,-0.4719304442,-0.1327183694,0.2813080847,-0.1820606738,-0.1882600039,-0.0503824688,0.372369945,0.1919568777,-0.1998262852,0.104513973,0.081495218,0.2974182367,-0.3783482909,-0.0964141414,0.1138468012,-0.3751120865,0.043580927,-0.210797742,0.4439977109,0.0698247999,-0.2243980765,0.0163831245,-0.2619254291]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"Yeah sure! Instead of using default `0o644` should I first extract umask of current user and then use `os.umask` on it? We can do it inside `Dataset` class so that all folders\/files created during the call use running user's umask\r\n\r\n","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":40,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nYeah sure! Instead of using default `0o644` should I first extract umask of current user and then use `os.umask` on it? We can do it inside `Dataset` class so that all folders\/files created during the call use running user's umask\r\n\r\n","embeddings":[-0.2242726237,0.2261134684,-0.0817525238,0.0094758645,0.0471359864,0.0149757005,0.2634158432,0.0719240457,-0.1914955974,0.0407302789,-0.1341527104,0.0863360539,0.0493934341,-0.3142187595,-0.116745472,0.1358978152,0.1529258937,-0.0015896432,-0.0006147567,-0.0963858441,-0.2239349782,-0.0862397924,-0.1034872457,-0.0336775966,-0.3453719318,-0.2179553211,0.2119508833,0.2314933091,-0.088236481,-0.1905014664,0.2426152229,0.6062232256,0.180053249,0.1983390749,-0.0001155414,-0.2169194967,0.0884057879,-0.0958579928,0.0502289869,0.2252657861,0.013323619,-0.0652768463,-0.1900797188,-0.1833731681,-0.2362517565,0.1567798853,0.1003571898,-0.7182676792,0.3611292243,0.2002629042,0.1723333001,-0.2232726365,-0.2373960316,0.0888243765,0.2409683764,0.1156111211,-0.0989544019,0.0126231266,0.3320398033,-0.1668721884,-0.0690127909,0.0343173519,-0.2808453739,0.0031792175,0.1546429545,-0.0779458731,-0.2983679771,-0.5014804602,0.1238657981,0.1186875179,0.7641479969,-0.1171290502,-0.5640271306,-0.0496227294,0.1295876354,0.2901589274,0.3258309066,0.305441618,-0.0033054377,0.3387885094,-0.556910336,0.0916791111,-0.2294591665,0.1012438983,0.1967002898,0.0366982259,-0.0670412257,0.1275733858,0.0062289634,-0.1861501485,0.2320861965,0.0242003761,-0.3480373025,0.3791960776,-0.0957939103,-0.0125216506,-0.1537094861,0.5009945631,0.2390560806,0.2913291156,0.1713189781,0.274330616,0.050613381,0.1730332375,-0.0316947699,0.1815221906,0.4626689255,0.2795345783,0.5615982413,-0.2898006737,-0.1274800003,0.0201109517,0.0461159162,-0.3072869778,0.0547322407,0.2786273658,0.0398364998,-0.174011156,0.2315386385,0.0633214638,-0.1945544779,-0.197217226,0.0998361036,0.3442124426,0.0919306949,-0.0721913651,0.0562526621,0.1736780405,-0.0946675763,-0.356892854,-0.0090116914,-0.3280628324,-0.2786850035,0.3977695405,0.4268735349,-0.0590404123,0.048290953,0.2228619903,0.1398226917,-0.161517486,0.4451647699,-0.104907997,0.4017587006,0.3126306534,0.2725572884,0.2442822158,0.0619794242,0.0393465497,-0.1138395071,0.336044848,-0.3299659491,-0.1983573437,0.3492939174,0.0604472309,-0.0421798714,0.2782869935,-0.2858790755,0.0682794452,0.4992926717,-0.0701061413,0.3984587491,0.195801869,-0.2189560235,-0.4167913795,-0.0710380599,0.2780780196,-0.0480387099,-0.0400205776,-0.1446887553,-0.0054468689,0.0121445721,0.330481708,-0.0306329019,0.2323587388,-0.1623285711,0.1956809014,0.4786145389,-0.1494109333,-0.6601572037,0.0933339521,0.1913601309,-0.1107339188,0.1814989895,0.0880109221,-0.0511975065,-0.1090390608,0.0170315672,0.2962633073,0.1734892279,0.2679624557,0.0782138109,0.0145116327,-0.1027067155,0.0111187836,0.0351286866,0.2638350129,0.3305090368,-0.3714771569,0.2493853122,-0.3268905878,0.2354728878,0.4526799023,0.2369099408,-0.1237699613,0.0088803181,0.0077164839,-0.3304775655,0.2219384611,-0.1120770723,-0.3366325796,0.0353856198,-0.5332360268,0.1161451787,-0.2586063445,-0.1835006028,-0.1028007269,0.0589737743,-0.0131905703,0.0431489311,-0.2665457428,0.1984104961,0.2996811867,0.2209105939,0.0467694215,-0.1186486408,0.0867330134,-0.0954204723,-0.1309914589,-0.2401005775,-0.0371330455,0.2725090086,-0.0368210115,-0.2458095402,0.3138571382,0.2480554283,-0.0437143184,-0.2244311124,0.2249997556,-0.0020516347,0.079135552,0.0931482315,-0.1511032581,-0.0838271007,-0.1197993234,-0.2827210724,0.5314025879,-0.091330409,-0.1963916123,-0.2848781645,0.2199092954,0.1736144274,-0.0392812863,-0.1375671029,-0.1251328737,0.0259200186,0.2324845195,0.387634933,0.1938304454,0.1323322803,0.0504521653,0.3756450713,0.1494712234,0.1309051216,-0.0374019146,-0.0467793383,0.0487797484,0.2571569979,0.2776413858,0.379956305,0.0739180669,0.2769946754,-0.1387668848,0.3776808679,-0.1041438803,-0.0102290418,-0.0151650468,-0.0896696225,-0.1636366695,-0.219245851,-0.0727818832,-0.022523215,0.2986067235,0.0683608279,0.1479007453,-0.3589763641,-0.1564719528,-0.1277645528,-0.2656018734,0.3322516978,0.0047389707,-0.4301035702,-0.0467858799,-0.1731005311,0.2626761496,-0.0556140803,0.2959812284,-0.1558345556,0.2158390582,-0.0157291833,-0.0303070098,-0.1898551434,0.1892218292,0.0821965858,0.0232586712,0.2347505093,-0.6215863824,0.1095520407,-0.0536568686,0.2752892673,-0.388805747,-0.0542002879,-0.1635511816,0.0965677276,0.0827460662,-0.3341892064,-0.0747675076,-0.2251954824,-0.0046242233,-0.0955513343,-0.3325315714,-0.2238927782,0.0089835487,0.0110308044,-0.0332578756,-0.2029950619,-0.0610420071,-0.1555436999,-0.1267533749,0.4859426618,0.1549123973,0.2311938405,-0.2037691623,-0.2212675959,-0.3293208182,-0.1553503722,0.1997089237,-0.346811384,-0.6803945899,0.3059538603,-0.5067756176,-0.3965727985,0.0117284013,0.1094526798,0.1193613037,0.5079334378,0.0697839335,-0.2112737894,-0.1897013038,0.2861461937,-0.2172395885,0.1488739252,-0.0102196513,0.0184298791,0.1693856865,-0.0138924997,-0.1542496085,0.0458573811,-0.0391508415,-0.2779386044,0.1855635941,0.1891299635,0.1423265189,0.5497563481,0.4713593721,0.3665600419,0.2696023583,0.2861098051,0.6051007509,0.078364104,-0.1019397676,0.1641366333,0.078506127,-0.0938903689,0.1588465422,0.0909419954,-0.1556552351,0.0585833266,-0.1055934578,-0.2563833296,-0.4986239076,-0.1460413635,-0.2895292044,0.0465915538,0.0067771263,0.2124719769,-0.2476478219,-0.2021618336,0.0607077926,0.296166867,0.1324817091,0.2858157754,-0.6697198749,0.1126682088,-0.3476349115,0.4643508196,-0.0250690151,0.4630535245,-0.1830045879,0.1925448477,0.200042516,0.3045369685,0.4123745263,-0.5408961773,0.3032761216,-0.0348493792,0.0447061881,0.2485870421,-0.150705412,0.3441003859,0.5285432339,0.1163998619,0.1270213425,0.130396381,-0.1236462221,-0.4273568988,-0.2987099588,-0.083702527,-0.2671335936,0.1920259297,0.1025150791,-0.0579169728,-0.0449829213,0.2800468504,-0.310229063,-0.1342485696,-0.0060632671,-0.3104099929,0.165164113,0.0825788602,-0.1498119831,0.3525903523,-0.0445756577,-0.0529518388,0.3878894448,-0.009541899,0.3630973399,0.4198490977,-0.2888559103,0.2272986025,-0.155498758,-0.3446565568,0.3685617447,0.0035257142,-0.2869550884,-0.1578528732,0.183310613,-0.0863534734,-0.2582291961,-0.107128188,-0.1099065021,0.1533146203,-0.2869219184,-0.6173260808,0.1457749009,0.6142501831,-0.3409030437,0.2515214682,-0.1457551718,-0.4050543904,0.2226870358,0.1497237682,0.8280984163,-0.1241632253,0.413251549,-0.39656353,0.0532721914,0.1600639075,-0.4766876698,0.0904370546,-0.5270587802,-0.3178489506,-0.1073178276,-0.1308599263,0.1058214456,0.4674871564,-0.0953694507,0.1056402549,-0.0690561235,-0.1904557794,-0.1384858787,-0.1392826885,-0.2735178173,-0.2676415741,0.2913942635,0.1141194031,-0.1776751131,0.0626792535,-0.0040102089,0.0608895011,-0.2183615565,0.2051762193,-0.2044147253,0.0277181584,-0.6304150224,0.1333508939,-0.1440234333,-0.038788408,0.1345829964,0.4161984921,0.3846624494,0.3360762596,-0.0493359342,0.062047258,0.3749551475,0.2493367195,-0.1366826445,-0.0668106005,0.3273591101,0.1755493283,0.0858665332,-0.2430320382,-0.1404042691,-0.2752978206,-0.2996563017,0.0528528392,0.0927358642,-0.3621003032,-0.0694540963,-0.2669883966,0.0355918482,0.0817804337,0.0584820099,0.4753650427,-0.0615737699,0.5999324322,-0.3380889595,-0.3707462251,0.0232545082,0.6508812308,-0.0748978704,-0.1919886023,0.1610781848,-0.2853004336,-0.2563264072,-0.1611593068,0.2667996287,0.3067183197,-0.108432129,-0.1093942225,0.0621910319,-0.0928495154,0.2604233921,-0.5158432126,0.0780071616,0.1946417987,0.0990702361,-0.2784630358,-0.4767000377,-0.2796086371,-0.1793207824,0.1743407249,0.2712527514,-0.5941870809,0.134515211,0.0228963494,-0.2046209723,-0.1761206537,0.0375181213,0.0708982795,0.1485579908,0.0247309618,-0.1150831357,-0.189116016,-0.081625469,-0.1084240824,-0.1095855907,-0.1518594027,-0.0716804117,0.1274793446,0.1931977123,-0.4675964117,0.1774009317,-0.4139453769,-0.1233777627,-0.4254440963,0.0573093407,0.3909038901,0.1731759757,0.3108251095,0.2425544262,-0.0621311441,-0.2998076379,0.1931417584,-0.2430414408,-0.2243489325,-0.2455173433,0.2626123428,-0.3039444983,-0.0883889571,0.2513172626,-0.0878844112,0.1320441216,0.0770525336,0.0945399255,-0.0262544658,0.0228617955,0.1102326065,0.2923847139,-0.0552265123,-0.1660205722,-0.0323023461,0.3039821982,0.0958500952,0.1196826026,-0.3297826946,0.2782692909,-0.1658328325,0.0262979399,0.3702871501,0.086722523,-0.2481269091,0.0856607705,-0.1553162336,0.1704511642,-0.2783286273,0.3902421296,-0.1039145514,0.0281766448,0.072722286,0.3354695141,0.2010548562,0.0278960355,0.298098892,0.1519480199,0.2082490474,0.3124473691,-0.2757878304,0.1694065779,-0.664400816,0.0975837559,0.0966763273,-0.0346900448,0.2268094569,0.1662105024,0.0640404373,-0.0367963724,-0.2975319922,0.0596018098,-0.1610552222,-0.1909186542,-0.1152938455,0.0046006641,0.0813321844,0.0455231443,-0.2006155401,-0.2622351348,-0.1746820807,0.0928321034,0.2766190469,0.0742262825,0.1658158153,-0.008980806,0.0760786757,0.0218041576,-0.2504441142,-0.0410892554,-0.4445384443,-0.1119054109,0.2693799734,0.0846511126,0.1198275387,0.126466915,0.1806228757,0.171916604,0.3239747286,0.0822094306,-0.2945541739,0.2347816229,0.0127739804,-0.3522435427,0.0530846342,0.0332129262,-0.0139331548,-0.1016917601,-0.1442115903,-0.0461672917,0.0976248831,-0.0147445444,-0.1317888796,-0.4960697293,0.1650494784,-0.0090070553,0.1675388813,0.0311256163,0.4447352886,0.0160280038,0.2173564285,-0.4712947905,0.0246176403,0.3944928944,0.2075607628,0.3714955747,-0.1532778442,-0.0514095053,-0.0501778647,-0.3450664878,0.1659596264,0.2271414548,0.2972990274,-0.0110217845,-0.2264147103,0.0516584069,-0.1125049144,0.4493536651,-0.1799069494,0.2619473338,0.361106813,-0.3241623044,-0.215873763,0.2193960547,-0.0090490468,0.100637652,-0.4241232276,0.3100613058,0.0644676462,0.0631120279,-0.1299867034,-0.1061271578,0.1513738781,0.5122240186,0.3870996833,0.2622054815,-0.005773522,0.0156651661,0.4608681202,-0.2425303012,0.1981351078,-0.2715996504,-0.0081661642,-0.1149508581,0.1225867197,-0.5734186769,-0.2123507857,-0.3278991878,0.0249956064,-0.1814757586,0.2328229398,0.0309174582,0.1897112131,-0.1288593709,0.0232264027,0.2941923141,0.301998198,-0.1238590032,0.386236757,-0.4685359597,-0.0766074136,0.3101243973,-0.2580967247,-0.118823275,-0.131292671,0.354113698,0.1892221719,-0.2718223333,0.136686042,0.0733439773,0.3832210898,-0.3898983002,-0.1043068245,0.0537061356,-0.3147775233,0.0502255037,-0.1929299235,0.4530383646,0.0467530936,-0.2346209139,0.0810261294,-0.2331266999]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"You can get the umask using `os.umask` and then I guess you can just use `os.chmod` as in your previous PR, but with the right permissions depending on the umask.","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":30,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nYou can get the umask using `os.umask` and then I guess you can just use `os.chmod` as in your previous PR, but with the right permissions depending on the umask.","embeddings":[-0.1857263744,0.1737921834,-0.1075274125,-0.1851043403,-0.0081702275,0.0839048028,0.3565381169,0.0757277459,-0.1108319834,0.1052552685,-0.2269385457,0.1543242037,0.1039311737,-0.2934491336,-0.0649807751,0.2179357857,0.0900316015,-0.0351480022,-0.0519519933,-0.008974527,-0.139015764,-0.1062896103,-0.0973075032,-0.022077186,-0.2167216241,-0.1977164894,0.1695958227,0.3388963938,-0.0508378111,-0.2708571553,0.1947233826,0.5540147424,0.0244192667,0.1041323543,-0.0001054582,-0.1182964668,0.0259445105,-0.0766808167,0.0837031528,0.2544776201,0.0616562963,-0.0599539205,-0.2419357747,-0.1721672565,-0.2474578321,0.1950842589,0.1271020174,-0.5841297507,0.3706359565,0.2619087398,0.2632213831,0.0041867811,-0.2159233093,0.0827071518,0.2322053611,0.0394572988,-0.1844671965,0.004054938,0.3705589175,-0.0817732215,-0.029460717,0.0447111353,-0.1966822892,0.07993792,0.1410067081,-0.0492858887,-0.2201645672,-0.4966886044,0.171047762,0.0686404631,0.737410903,-0.0505575724,-0.5258098245,0.0515309945,0.1633486897,0.268779397,0.3548974693,0.3101957142,-0.0543667637,0.3213352561,-0.4402199388,0.1428738385,-0.2138585299,0.0231555253,0.2160202712,0.1043673605,-0.0138586164,0.1185606271,0.1312624663,-0.1723819524,0.1273884475,0.2220983505,-0.2841736376,0.4184722304,-0.0803543478,0.0213506687,-0.179577142,0.4277115166,0.1930829436,0.1829823107,0.2210577428,0.2631050944,0.0399355628,0.2481009811,-0.1107326448,0.087615408,0.3066399097,0.3309390843,0.600415349,-0.2753514349,-0.1773978174,0.03613941,-0.0040830355,-0.4007755816,-0.0075756093,0.2617320418,0.1033153832,-0.2411533147,0.2197194248,-0.0143741863,-0.0503947176,-0.1352396458,0.0554776005,0.4130481184,0.2159236968,-0.2010885477,0.1329979897,0.1610405743,0.0002650758,-0.2362119555,-0.009621718,-0.2826401293,-0.3536853194,0.3084272146,0.4239220321,0.0281935409,-0.0059308391,0.2135394067,0.1640729159,-0.2162478566,0.3558654785,0.0013643519,0.4887394607,0.3780561388,0.2298045456,0.1242796332,0.0713280812,0.0070001739,-0.1356545389,0.375425607,-0.1685845703,-0.1559192836,0.2985610366,0.1440183669,0.0517648198,0.2353738695,-0.1760039479,0.0034030655,0.4222274721,0.019279385,0.4730708301,0.263438344,-0.1811551601,-0.3831002712,-0.0662586316,0.1607260108,-0.1245954037,-0.0031760992,-0.2033124715,-0.1225336045,-0.0638828427,0.3225443661,-0.0349490419,0.1460774988,-0.1530025601,0.1799973994,0.4943962097,-0.105420202,-0.5366950631,0.0179225653,0.1420943141,-0.1575104743,0.2553617656,0.0699409097,-0.0862172246,-0.1205488741,-0.0840222538,0.3386438191,0.2401405275,0.1952783763,0.0759221613,0.0618522577,-0.1156338155,0.0749076232,0.0716744363,0.3632016182,0.3213571906,-0.3289202154,0.2820542455,-0.3049741089,0.2371451855,0.4326992035,0.2295791209,-0.0457228646,0.0391492844,-0.0006195487,-0.2805831134,0.2985621691,-0.2028404325,-0.2408683002,0.0562757328,-0.5901844501,0.0555274002,-0.2122088671,-0.1397847235,-0.1616668701,0.158700794,0.0242923088,0.1034163311,-0.3360012472,0.2234393954,0.274657011,0.2683152556,-0.0121355113,-0.0398344696,0.0267179031,-0.2164893746,-0.1205914617,-0.1934047192,0.0145761492,0.18436566,-0.0178955551,-0.2853236794,0.353510052,0.1852293611,0.0203632955,-0.2001335323,0.1085040718,-0.1056202352,0.1050096005,0.0359733067,-0.186685428,-0.1000032648,-0.0848841146,-0.2744738758,0.3847475946,-0.0844669119,-0.1805409193,-0.2337636054,0.2074716985,0.1745804399,-0.0609582812,-0.085184738,-0.1573993266,-0.0265763216,0.1765721738,0.318298012,0.144673124,0.1370994747,0.0576272644,0.2981226444,0.1863227487,0.0654523596,-0.0301004052,-0.013126622,0.1375167668,0.2089313567,0.2619199753,0.3403193653,0.1240618601,0.3443061709,-0.0757716522,0.3184883595,-0.0739094466,0.0338750035,-0.111021325,-0.125604108,-0.1951065212,-0.1830316037,-0.090452686,-0.136238113,0.2723574936,0.0790062174,0.1283605695,-0.3007153571,-0.170316115,-0.1353120208,-0.2051401287,0.351392597,0.0678389221,-0.5054529905,-0.0896584615,-0.051605396,0.2065724134,-0.1915788352,0.3344317973,-0.1828148216,0.1299615949,0.0263543371,-0.0370130427,-0.2518053055,0.1204160973,0.1502980739,0.1103513986,0.1937009394,-0.5752273798,0.1569736749,-0.0448526442,0.225627169,-0.3547329009,0.0017738494,-0.1311611831,-0.0715568066,-0.0430169702,-0.3177586496,-0.0399482287,-0.1978435218,-0.09189073,-0.0097182319,-0.3843921423,-0.3132214844,-0.0307987873,0.1463947743,-0.0677189454,-0.3414189517,-0.0593514219,-0.1389716715,-0.2776058018,0.4889319241,0.1193414181,0.2346739471,-0.0731223822,-0.330290705,-0.2986045778,-0.0375004373,0.1694009453,-0.4259344637,-0.5024003386,0.3395012021,-0.5373857021,-0.4263425767,0.1487544179,0.2579756975,0.1109817699,0.4555156827,0.0572565794,-0.1329280585,-0.1570355445,0.211027652,-0.2334004343,0.1151978225,0.0440225862,0.0399837866,0.0330163501,-0.0285793152,-0.1677332073,0.0434171669,-0.0440920554,-0.2562850416,0.029993454,0.1130884588,0.185866341,0.5306411386,0.2862323225,0.3638160527,0.2622100413,0.217113167,0.64016819,0.0089031262,-0.0704265684,0.1620161682,0.0473679267,-0.0726466253,0.2092021406,0.0582449399,-0.2456905097,0.0901000127,-0.0096494611,-0.1603545696,-0.5348592401,-0.1364097446,-0.2738773227,0.0883172899,-0.0140957115,0.1447206736,-0.2262011617,-0.1098507047,0.1252885163,0.2826583087,0.0938107967,0.2822671831,-0.5877381563,0.1400512755,-0.4001904428,0.4587441683,-0.108924143,0.4167456329,-0.2380623221,0.0556820892,0.2281851918,0.2692908347,0.3717995882,-0.5058965087,0.3724425137,0.0563733168,0.1023863554,0.2538180649,-0.1848615557,0.337785393,0.4753628671,0.1743141264,0.2008108795,0.1267883331,-0.0715224668,-0.377574265,-0.319856137,-0.0379983187,-0.2707748711,0.0847480148,0.1233548,-0.2259889543,0.0749147236,0.2653068006,-0.3210915625,-0.1286652386,0.0373397321,-0.2165566534,0.2111465037,-0.0313396566,-0.0530680977,0.2663189173,0.0370253064,-0.0229362901,0.2862891257,0.027783433,0.222078532,0.4201602042,-0.2369470894,0.3318751752,-0.1702903956,-0.336882472,0.2943888903,0.0241818801,-0.2756510377,-0.214330256,0.2107574791,-0.111005269,-0.1910564601,-0.030451335,-0.1182826161,0.0806012228,-0.2883733511,-0.4726788402,0.097919859,0.5448185802,-0.3314675391,0.2728619277,-0.1408364177,-0.3869784474,0.2660946846,0.1094389558,0.727016449,-0.2749713361,0.3488993347,-0.4370073676,0.0407633893,0.0323103629,-0.444514513,0.0848009437,-0.401492238,-0.4133333862,-0.0595447272,-0.1893133819,0.0402193107,0.4990395904,-0.1465308517,-0.0222712457,-0.0178483967,-0.0763649717,-0.163012892,-0.1716429889,-0.3598255813,-0.2808669209,0.2287325114,0.2174030691,-0.188920036,0.0951367989,-0.0304712467,0.0031736514,-0.1627361029,0.2038644701,-0.2175656408,0.0892556608,-0.594380796,0.13763614,-0.2181157023,0.0248077586,0.0880711973,0.3157127202,0.4517297745,0.4367559254,0.0026658282,0.119159095,0.4417306781,0.1977397352,-0.190671131,-0.0773941278,0.2027617991,0.1221280172,0.079196997,-0.2834093273,-0.1204707101,-0.4084975421,-0.2066504955,0.1360190362,0.0426761545,-0.2759870887,-0.0068433811,-0.1643593162,0.0046577249,0.0218640883,0.1564436406,0.3987327218,-0.2099607587,0.4780230224,-0.4339782298,-0.3344350159,0.086293608,0.6166929007,-0.0899298787,-0.1760612279,0.1232265085,-0.2064266503,-0.2042393535,-0.2376882136,0.1317479461,0.2023619562,-0.118687354,-0.0029150844,0.1298507303,-0.2015672475,0.2526235282,-0.4055039883,0.0515465401,0.2109722197,0.0356707871,-0.2672998011,-0.5579944849,-0.2802609801,-0.2669665217,0.2188715786,0.2362939417,-0.4425866306,0.1286575794,0.003392007,-0.3072440624,-0.1627130955,-0.0362462178,0.1224087402,0.1837943047,-0.0766969323,-0.0325233787,-0.1841505468,-0.0003416852,-0.0184252858,-0.0941526964,-0.256641835,-0.0072461595,0.0827958807,0.192666471,-0.4056092799,0.1834541261,-0.4027767777,-0.1051968709,-0.3707084954,0.1169970036,0.3733254969,0.1056439728,0.2204140127,0.1463718861,-0.0642917901,-0.089133732,0.1361055672,-0.2207280248,-0.2519513667,-0.2602116168,0.2186689377,-0.3524957001,-0.0847535878,0.269778043,-0.0807428211,0.1106368825,0.1064037755,0.1349806041,0.0692940578,-0.0848778039,-0.0026868822,0.3689288199,-0.0710172951,-0.0355717875,-0.0149403168,0.3070230782,0.1688727736,0.0794644579,-0.3034918606,0.2815037668,-0.2608578205,0.1138424277,0.3206259012,0.1516405642,-0.3100244701,0.1678882837,-0.0553319082,0.1427605003,-0.2854298055,0.3725979626,-0.0370527953,-0.0242446661,0.1064664274,0.3017580211,0.2455573976,0.0067350776,0.1786158383,0.1392190009,0.2059953064,0.316149354,-0.377920419,0.2491225451,-0.5594152808,0.0484786965,0.0082318466,0.0325462222,0.3114193976,0.1086302847,-0.0149196032,-0.0677377358,-0.3752583265,-0.0246973075,-0.0611316599,-0.2230540067,-0.0955307037,0.0446852632,0.1792000234,-0.0076636681,-0.1667665094,-0.1541598588,-0.181434527,0.0794107914,0.1991389245,0.1623674333,0.2552837133,0.0390834883,0.0343711227,-0.0218187273,-0.2530502081,-0.0737129673,-0.4660960734,-0.1147785783,0.3790410161,0.0014307948,0.1322906911,0.0263407249,0.0815609097,0.2408283204,0.2220010459,0.0782455951,-0.4063874781,0.3051211238,-0.0724514872,-0.3299732804,0.1291553825,0.0806963518,-0.1295105517,-0.1849515736,-0.2925862372,-0.0761936754,0.1665280014,-0.0796483159,-0.1766750216,-0.5326669812,0.1927374005,-0.0364255458,0.1345303208,0.0583582073,0.3334976733,0.0147842076,0.1852365881,-0.4901877344,0.0773783252,0.3475421071,0.2903096974,0.4328676462,-0.1794337034,0.0067896992,-0.0588163361,-0.3755061626,0.0790170878,0.2403367907,0.2012201846,-0.0435046256,-0.2368942648,0.113991797,-0.1731886119,0.6218096614,-0.2811032832,0.1527004391,0.3637656569,-0.21879749,-0.1694778502,0.2325690389,-0.0128173279,0.1089786142,-0.4467694163,0.235107094,0.1090224236,0.144573763,-0.0753417686,-0.1449445039,0.179943189,0.4712297916,0.4616510272,0.2216768861,-0.0527184606,0.0086718062,0.3859558702,-0.3913455307,0.0796373114,-0.2222420424,-0.0789974108,-0.0957826599,0.0782982558,-0.5545232892,-0.2305322438,-0.2888013721,0.038664747,-0.1613692939,0.2081177086,0.1303797662,0.1568286717,-0.1873615235,0.1222098991,0.3609289527,0.2784610987,-0.0892702192,0.2829227448,-0.4997110069,-0.2178987265,0.2488850802,-0.1825582385,-0.232347548,-0.0642231256,0.3706893027,0.1894252151,-0.2345671803,0.101349093,0.1262744963,0.2508757412,-0.3542096615,-0.1510207653,0.0408623964,-0.3003990948,0.0062546036,-0.2502034903,0.3976642191,0.0391334333,-0.2391091436,-0.0354226232,-0.2917317152]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"FWIW, we have this issue with other caches - e.g. `transformers` model files. So probably will need to backport this into `transformers` as well.\r\n\r\nthanks @thomwolf for the pointer.","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":29,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nFWIW, we have this issue with other caches - e.g. `transformers` model files. So probably will need to backport this into `transformers` as well.\r\n\r\nthanks @thomwolf for the pointer.","embeddings":[-0.1207443476,0.2110619843,-0.0841049924,0.0468187705,-0.0028637,0.1459375322,0.3854053617,0.1391602904,-0.2011681795,-0.0585675202,-0.1354780197,-0.0176231824,0.1156202182,-0.2905223966,-0.0399506688,0.1175884604,0.1337425113,0.0035101185,-0.0366873182,-0.0839240626,-0.0998415574,-0.0752499774,-0.0850521475,-0.0726410747,-0.3615028858,-0.1329764724,0.2650974691,0.2255142927,0.0090469653,-0.1246050224,0.3047429919,0.4333624244,0.0784689039,0.1299982369,-0.0001106508,-0.1399454772,0.0487940088,-0.0642532706,0.0905670524,0.2714790702,-0.1321593374,-0.0761748701,-0.1686177701,-0.1069586277,-0.1106696054,0.2972267866,0.1226569265,-0.5791894197,0.4974912405,0.1974837929,0.2274942696,-0.0538505316,-0.3413310051,0.0540440641,0.292683512,0.0809753537,-0.2243646085,0.0552525185,0.2692474425,-0.1958608627,-0.0183214154,0.0551859252,-0.2950374782,0.1153828874,0.1548876911,-0.121912688,-0.2390939146,-0.5055351257,0.1950537115,0.0620003827,0.7401837707,-0.0504377522,-0.4978570938,-0.1324620992,0.0188213438,0.2531050146,0.4132610261,0.339654386,-0.0014014319,0.2719789445,-0.5397583246,0.0072858073,-0.1382894665,-0.0447920263,0.1522361636,-0.0007977514,-0.0132668167,0.0848157778,-0.0201911423,-0.0922665223,0.090726532,0.0848192573,-0.3772098124,0.3344877064,-0.1749615073,0.08187107,-0.177461639,0.4131331146,0.1411212683,0.2401570976,0.2163678855,0.2027886808,0.0941766724,0.2004716247,-0.0032466499,0.1715729535,0.3708486855,0.3696019351,0.5744590759,-0.3903861344,-0.1783677936,0.0079541774,0.0443068855,-0.32673648,0.1036925837,0.2564816773,0.018813489,-0.2269121408,0.1517628729,0.1165376753,-0.0508614704,-0.2100030929,0.0428451113,0.2555075586,0.1683595628,-0.1430663168,0.0302899703,0.1445005834,0.033597365,-0.4030351341,-0.0297992416,-0.342314899,-0.4116488695,0.3594991267,0.4779577553,-0.1639427245,-0.0407774895,0.222055167,0.106182754,-0.247854054,0.2698282003,-0.038396921,0.4771815538,0.3560424149,0.1411651522,0.3438854814,0.0771397278,-0.0027546473,-0.0662941262,0.3606114089,-0.2789525092,-0.0623124391,0.3091772497,0.0600036979,-0.0259723179,0.2341597825,-0.3002974391,0.0653690845,0.6256068349,-0.1201736033,0.4150093198,0.3077589273,-0.1959592104,-0.3834025562,-0.1492616832,0.3160617948,-0.1199639887,-0.0305181965,-0.0605188683,-0.0397669449,-0.0460440218,0.2801170945,-0.0383947827,0.1791476458,-0.0919910967,0.1846491396,0.4149959981,-0.0262843464,-0.5820008516,0.1069188416,0.1509170234,-0.162860617,0.131841585,-0.0172963869,0.0163426846,-0.0730525032,-0.1057644933,0.3387136161,0.1975997537,0.2817466259,0.0935706571,0.1124939024,0.0462467633,-0.0897384584,0.1349194348,0.4027235806,0.2795726955,-0.2379645109,0.173605606,-0.2297310829,0.300124824,0.4829814732,0.0882581174,-0.1560760438,0.0094194477,-0.0881859511,-0.4463413358,0.3059563637,-0.1769651771,-0.3649234474,0.0936228931,-0.386436373,0.0234633759,-0.2765288055,-0.2104593515,-0.0824699402,0.0991922617,0.0023034241,-0.0160406232,-0.3887529969,0.2274432927,0.3831131756,0.3045082986,-0.0300990697,-0.0952215046,0.1234670952,-0.1274682879,-0.0749267191,-0.2616179883,-0.0406631567,0.292234391,0.0081635015,-0.3218149245,0.2308075279,0.1162156835,0.0567136817,-0.0873666108,0.2493588179,-0.0352321453,0.0807025135,0.04281931,-0.1840926707,-0.0539492406,-0.1325837523,-0.2853808701,0.2740264237,-0.0370531045,-0.065147616,-0.2375504524,0.2129855603,0.1421752423,-0.0171996634,-0.0711244345,-0.1782414019,-0.0635961741,0.1644781828,0.3153994083,0.2228892446,0.1195312962,0.0457767583,0.4026435614,0.2261164486,0.1156601459,-0.0225620512,-0.1593435258,0.0662745684,0.1717029065,0.3607619107,0.2226926237,0.0145552959,0.3567924798,-0.2121688128,0.2847136259,-0.1805050224,0.0212035645,-0.0329974815,-0.0282044876,-0.150320068,-0.0930583924,0.0097617758,-0.1050756276,0.3722705543,0.167938292,0.0914387628,-0.3406541646,-0.1041124314,-0.2076177597,-0.1616516113,0.2695324123,-0.0192013569,-0.6123397946,-0.0851016343,-0.1596945822,0.1629980206,-0.0465835556,0.1571589708,-0.2691048682,0.2040271908,-0.0234534331,-0.1180556118,-0.3213480413,0.1613807082,0.0332060531,0.0537653603,0.2560738325,-0.6495655179,0.0917300135,0.0269913189,0.3185979426,-0.2326247543,0.0595846288,-0.146228075,-0.0715482086,0.039310541,-0.5172543526,0.1141471863,-0.1451077312,0.0379493348,-0.0977479443,-0.2473940104,-0.2224614471,0.0385755375,0.0551554561,-0.1070951447,-0.3245078623,-0.0422688909,-0.0262890328,-0.1569161266,0.434792906,0.0526706651,0.1734728366,0.1191670671,-0.3077769876,-0.3521289229,-0.2285962403,0.2458472103,-0.4041653275,-0.5424076319,0.3612726033,-0.4819200039,-0.4319618046,0.0786874145,0.2185782641,0.0797881484,0.4152204692,0.0437998883,0.08299537,-0.2696126699,0.1614716202,-0.1929752082,0.0766387656,0.0515767932,0.0490494967,0.1112638786,-0.0858579054,-0.1944163442,0.101161994,0.0200057719,-0.2241318375,0.0323726051,-0.0041683228,0.1705233306,0.6693534851,0.3900607228,0.2593128383,0.266666919,0.3642522693,0.5709922314,0.0490256846,-0.1328871846,0.1190948635,-0.0469398797,-0.0885546282,0.0905668512,0.0463615283,-0.2562758327,0.0302772243,0.0650870278,-0.2295381725,-0.4630319178,-0.2507559657,-0.3004631102,0.1445189267,0.1581032723,0.1250147671,-0.236182645,-0.1631697714,0.0586484857,0.2605683506,0.2015032917,0.2332779169,-0.3842125833,-0.0104054948,-0.2581269741,0.3752669394,-0.0234400015,0.5265198946,-0.2363996506,0.1108480245,0.1992043108,0.3797487617,0.4119975865,-0.4367414415,0.3121191859,-0.0290367231,0.1729656905,0.3494873047,-0.2548164427,0.4342045784,0.4451816082,0.09644261,0.1208937466,0.1503413618,-0.1660154462,-0.4023326933,-0.2793961167,-0.0991647318,-0.1978135556,0.1059619188,0.3193877935,-0.0468382202,-0.0331529081,0.3301813006,-0.3447495997,-0.1954693049,-0.010511646,-0.1184765697,0.1609705091,0.0516065173,-0.0585898198,0.2994741499,0.0600552186,0.056608256,0.4806612134,0.0394468755,0.025030816,0.3715111911,-0.1879866123,0.3247098327,-0.1371188909,-0.2970531583,0.3379131556,0.040755108,-0.3488020599,-0.1530902386,0.1046259478,-0.2581830919,-0.2618082464,-0.1563231796,-0.1067790687,0.0492528751,-0.2082594633,-0.4934577644,0.2734438777,0.4977818429,-0.3501679301,0.2535853982,-0.313516438,-0.3273327649,0.1499074847,0.0324317776,0.7552210093,-0.2470967174,0.316973716,-0.4600428641,0.0671679229,0.0081999563,-0.3994583786,0.1234069392,-0.4250825644,-0.3580511808,-0.1449174434,-0.1411130428,0.0188522991,0.4696406424,0.0408214852,0.0398944691,-0.1510104835,0.021585729,-0.0914280564,-0.2567688823,-0.0955884606,-0.3043406308,0.312451601,0.1515035778,-0.2343553454,0.099085182,-0.0298338998,-0.0536014959,-0.2279219627,0.2225684524,-0.3397302926,0.1291856021,-0.51916188,0.1301492453,-0.1092741489,0.0138142202,0.2091045678,0.2999201715,0.3164104819,0.3617789447,-0.0635984391,0.1356314272,0.3436041474,0.2180134654,-0.2584341764,0.1038955599,0.2954527438,0.1615244001,0.0394166633,-0.2853001952,-0.1933385581,-0.2738273442,-0.3235791624,0.2683514655,0.0029195573,-0.198918879,0.0242783669,-0.1108779758,0.0610023066,0.0155594917,0.1157344431,0.4238666892,-0.0209254939,0.524325788,-0.3454038501,-0.3492100537,0.0638426244,0.6376696825,-0.0707529187,-0.1771105826,0.193582952,-0.2559060156,-0.2988440096,-0.1964825094,0.1373319179,0.0616040044,-0.1859369278,-0.0401697531,0.0337331779,-0.1001523957,0.2225742489,-0.4149352014,0.0018225514,0.0836656019,0.0938208476,-0.3515955508,-0.4222249687,-0.3398015201,-0.4331282675,0.1447082013,0.2051096708,-0.4737236798,0.2301024795,0.0923268944,-0.2330985963,-0.2326958477,0.0179570671,0.1307041049,0.2646757662,-0.0388002992,-0.0975536853,-0.2362687588,-0.0860925317,0.069035843,-0.046472311,-0.2174898386,-0.0203860812,0.1334382743,0.2556906939,-0.3523866832,0.0947977975,-0.3680514693,-0.131171152,-0.4231099188,0.0421783626,0.3070077002,0.1483245641,0.3742547333,0.1887093633,-0.0083825048,-0.1244743839,0.148142904,-0.1677154154,-0.2091808468,-0.2064441442,0.1940438896,-0.2729392648,-0.0500816964,-0.0063156891,-0.0616191328,0.0653186589,0.0550925396,0.1459687054,-0.0709647387,-0.0067333612,0.138604477,0.237282902,-0.0020307661,-0.0502694026,-0.0103899566,0.2375590205,0.12212798,0.1217859611,-0.2690328956,0.3627375364,-0.1349646002,0.0576028898,0.2876047492,0.2320633531,-0.2794457972,0.1833702028,-0.0390092917,0.1056975424,-0.360968411,0.4730057418,-0.0594379567,0.1213218346,0.1009886563,0.4183328152,0.2128695548,-0.0196254719,0.3656490445,0.3269251585,0.1012692228,0.2761337459,-0.3488725722,0.1409060657,-0.6575961113,0.0969430059,-0.0386149883,0.016814271,0.3265792727,0.0776372701,0.0090592699,-0.0321244225,-0.2968240976,-0.1123801246,-0.0842023492,-0.1260809749,-0.1141327545,0.1189675182,0.1017139629,0.0481321849,-0.2513362765,-0.1284192801,-0.1908626109,0.2366961241,0.1894706637,0.0225434992,0.2310845256,0.0273274127,0.070668146,-0.0191343855,-0.2209811211,-0.0542261824,-0.513969481,-0.1181327328,0.2671553493,0.0219152682,0.0516738854,0.1115734652,0.0948626399,0.1615170091,0.2507715225,0.0109942611,-0.3054790497,0.4592891037,-0.1178688034,-0.3537321687,0.0937659964,0.0210205708,-0.0784111097,-0.1089820787,-0.2171056867,0.1453212798,0.1375289708,-0.0040109078,-0.2561235726,-0.5012044311,0.0770843625,-0.0805306137,0.1889166236,0.0708738267,0.5467100143,0.0331711546,0.2029747516,-0.5516821742,0.0514182076,0.4038561285,0.3221231997,0.4221486151,-0.1119187325,-0.0375205092,0.0698338896,-0.3635555804,0.1241149902,0.2484867573,0.506573379,0.0020049999,-0.1873555928,0.1179129779,-0.2155749798,0.6327934861,-0.254635483,0.2672393918,0.3509069681,-0.2299675494,-0.2000048012,0.1187212393,-0.0796076208,0.1035526618,-0.3836973906,0.3240478635,0.2162733227,0.0956885517,-0.2002370954,-0.2255232036,0.1838062257,0.472713083,0.4594870508,0.2286716402,0.0339514092,0.0351210423,0.3650698364,-0.4535739422,0.0547579899,-0.1416039169,-0.0921010301,-0.2004774958,0.0867912844,-0.5620113015,-0.1265317649,-0.372298032,0.0174254626,-0.1991787553,0.269778043,0.0173303392,0.0229685418,-0.2298799008,0.0992417186,0.3050507903,0.2027697563,-0.1115300059,0.3502509296,-0.4840323031,-0.151173085,0.3006496429,-0.2600475252,-0.3285694122,-0.013344273,0.3732891679,0.1472627819,-0.2172841877,0.0106930248,0.0626192316,0.3244998157,-0.3747304976,-0.0860439241,-0.0225111246,-0.3646335304,0.0916026682,-0.1782128662,0.5489760041,-0.0094658528,-0.2097586691,0.0720796958,-0.2157584429]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"Hi @stas00,\r\nFor this should we use the same umask code in the respective model directory inside `TRANSFORMERS_CACHE`?","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":18,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nHi @stas00,\r\nFor this should we use the same umask code in the respective model directory inside `TRANSFORMERS_CACHE`?","embeddings":[-0.2081566155,0.1117953211,-0.0612362847,0.0407365672,-0.0061203358,0.1279416531,0.3894425333,0.0571595915,-0.2127809525,-0.0419883057,-0.1921856999,-0.123159267,0.1095725074,-0.3343217373,-0.0680713058,0.1472350657,0.1823889464,-0.0031142761,-0.0365220085,-0.1263749003,-0.1284440309,-0.0820216462,-0.0372605622,-0.0927570686,-0.4490078986,-0.1368952245,0.2483736128,0.1558468193,-0.0073176953,-0.0654333159,0.2889894247,0.5932964087,0.2336144149,0.1448652744,-0.0001169936,-0.1180792525,0.0391271897,-0.1211671084,0.0457089245,0.2500948608,0.0250783712,0.0906369463,-0.1391555667,-0.0956835151,-0.1904146224,0.3318998516,0.1194253191,-0.5785183907,0.5322913527,0.1401566565,0.130934,-0.0141243609,-0.3408102095,0.10062664,0.2094195932,0.1773287505,-0.1400357336,0.0186299924,0.3194201589,-0.2108982801,-0.0711339787,0.0976831466,-0.2137307972,0.0867346451,0.1340913922,-0.0724966973,-0.1330477446,-0.5008397102,0.1076873764,0.0051327101,0.7109176517,-0.1423660368,-0.612354219,-0.0560642667,-0.0337141119,0.384621948,0.3383271694,0.2856696844,-0.0450287275,0.3755660057,-0.5367846489,0.0684916824,-0.2287129462,-0.0421604998,0.1386980861,0.2050118148,-0.0096681016,0.0889965668,-0.0827852711,-0.0961588174,0.2455018461,-0.0053520827,-0.3007267118,0.3485356867,-0.1820778102,-0.0426302142,-0.2202080637,0.3356885612,0.1667661667,0.3436241448,0.1889274865,0.1737207174,0.1336089373,0.2200419158,0.0078025744,0.234630838,0.3791342378,0.5106111169,0.5242033005,-0.30716452,-0.2008763999,-0.0352666862,0.0575917177,-0.2570674419,0.0307384301,0.3571114838,-0.0152630061,-0.188887924,0.2196753174,-0.034931723,-0.066640012,-0.2599709928,0.0711568817,0.3590391278,0.1452923268,-0.0390979014,0.0123907421,0.1415419281,-0.0037328633,-0.3380565047,-0.0130785946,-0.3680552244,-0.3254832923,0.4648858905,0.4442012906,0.08915288,0.0396748893,0.1346452236,0.2065699846,-0.2064386606,0.3750051856,0.0013635495,0.4003739655,0.4353657663,0.2459950745,0.336381644,0.0151282893,-0.0809630826,-0.1440072656,0.3134813607,-0.2428768277,-0.1931257248,0.4719362259,0.0360613652,-0.043691285,0.2789411545,-0.2668830156,0.0263302345,0.6654939055,-0.0902685896,0.4075945318,0.3725253344,-0.1105863452,-0.4050094783,-0.0672455877,0.3489086032,-0.0102784438,-0.1288803965,-0.0141267609,-0.0060507483,-0.0985238478,0.463062048,-0.0963642597,0.1696590483,-0.0943508521,0.0977954343,0.5256636143,-0.1732740849,-0.6490622163,0.1500491053,0.2032510191,-0.1029816791,0.2669632137,0.0139924325,-0.0650034547,-0.1346264035,-0.0769748762,0.3790937066,0.1322385818,0.2626502812,0.0344484486,0.0248263404,0.0166534409,-0.0300011933,0.1266818941,0.3690103889,0.2432256788,-0.193096146,0.3039462566,-0.229432568,0.2638662457,0.4840605855,0.1777195781,-0.1601688862,0.0862397403,-0.0189624559,-0.4171614349,0.2549883723,-0.3125730455,-0.3193022609,-0.0076533998,-0.5207284689,0.0762457103,-0.244481802,-0.206301704,-0.2149771005,0.0228075814,-0.0447027534,0.1049694568,-0.411161691,0.1110285223,0.3464643359,0.311540693,0.0495175533,-0.193200022,0.1817408949,-0.1896053404,-0.0689440891,-0.337669462,-0.1110593602,0.2050257325,-0.0192736387,-0.2887729108,0.2957594991,0.1430360079,0.0078316079,-0.1420380324,0.3261239529,0.0783728734,0.0470811464,0.0594101362,-0.1094659939,-0.12912862,-0.1017894223,-0.3648785949,0.3784070909,-0.0848443881,-0.0770208091,-0.1734049618,0.2051728368,0.1564763188,0.0075013209,-0.1363317221,-0.1656583101,-0.0612607673,0.1625006646,0.4034815729,0.2103193402,0.2471065968,-0.0259754378,0.3793479502,0.1497155577,0.0658812821,-0.065851137,-0.1965869516,0.0703303218,0.1888887286,0.3214994371,0.2710371017,0.0138437944,0.2941034734,-0.2110759318,0.4143128991,-0.1599639803,0.010028325,0.0432761908,-0.0896273553,-0.131138429,-0.170338273,-0.0357452147,-0.0115990462,0.2849645317,0.0954615474,0.0764548853,-0.3082891703,-0.0691173673,-0.0060016289,-0.1206072494,0.2113471776,0.0088370368,-0.5343105793,0.0134322289,-0.1519223005,0.1806573272,-0.045663856,0.2692695558,-0.0677061602,0.1702588797,-0.0167088509,0.0004538158,-0.3200584948,0.138548553,0.1185337976,0.0039669154,0.2043606043,-0.6563039422,-0.0275075492,0.046050474,0.3301205039,-0.2155572623,-0.0556217991,-0.1608067155,0.0905814245,0.0782435313,-0.4964037538,0.0002072454,-0.108616285,-0.065435417,-0.0726546645,-0.2998086214,-0.1591831595,0.0295738354,-0.0222705938,-0.0988710895,-0.222018227,-0.0480080284,-0.1937407702,-0.1297093183,0.4764782786,0.1165347323,0.1581726521,-0.0076772133,-0.2429285645,-0.4943922758,-0.1973128766,0.2132521272,-0.247784391,-0.5221064091,0.4046732783,-0.5165218115,-0.4042021632,-0.0378913768,0.1206363961,0.1027639732,0.4209198952,0.1223703325,0.0109216245,-0.2580068409,0.0667786002,-0.1446611732,0.0787593052,-0.0300868824,0.0193842519,0.19900617,0.0284438841,-0.099014394,0.0424443819,-0.0189030934,-0.1615502685,-0.0023801303,0.1194666922,0.1578582376,0.5748952627,0.515283227,0.3354964554,0.1676900536,0.4023922384,0.4540117979,0.1008544788,-0.0585556477,0.040979851,0.020249771,-0.2165099978,0.1230836436,0.0902070925,-0.1655688733,0.0152077572,-0.0641173348,-0.1857699603,-0.5319417119,-0.1408689916,-0.2721810639,0.1616265774,0.1576242596,0.1266051829,-0.2879586816,-0.1087512597,0.1160159558,0.3581826985,0.1002248153,0.3036808372,-0.5236626863,-0.0116658946,-0.2940407693,0.3712227643,-0.0526747629,0.4781831205,-0.1404579431,0.1359277666,0.1758012027,0.2980241776,0.4374085963,-0.5280734301,0.2366176099,-0.0874200761,0.0176161248,0.2904190123,-0.2159897685,0.3345871568,0.3485830426,0.1534444243,0.232731998,0.1184235215,-0.1943306327,-0.4178481996,-0.3418982625,-0.1998859346,-0.2810037136,0.0884602666,0.2675311565,-0.1563452333,-0.1238456294,0.3386442065,-0.4243189991,-0.1691780388,-0.0309505798,-0.2532270253,0.1516336948,0.12433853,0.0380246527,0.3162674904,-0.0869844556,-0.0493335426,0.4028517008,-0.0670197979,0.1903662682,0.3316862583,-0.216317296,0.2620120645,-0.1415671557,-0.3252259791,0.3678044081,-0.0137583436,-0.2971774936,-0.1293168813,0.2408998907,-0.1591078341,-0.2108198851,-0.1638898253,-0.1980969906,0.1150968671,-0.244539693,-0.628033936,0.2098322213,0.5318979621,-0.3056167662,0.0931617618,-0.3135227561,-0.3025102615,0.4424118102,0.0626895875,0.8635259271,-0.2236696333,0.3869574368,-0.5408222675,0.0700362548,0.092800267,-0.692737937,0.16626513,-0.4834659696,-0.3403416574,-0.1500464529,-0.12808837,0.0517080091,0.5709947944,-0.0222601108,0.1149771884,-0.036574427,-0.0271955337,-0.1076061651,-0.190433532,-0.1143006459,-0.2419590354,0.3502944708,0.0627257898,-0.1259156764,0.1566949636,-0.0438902937,-0.0800254941,-0.2501616478,0.172967881,-0.2542056143,0.0647724345,-0.6468902826,0.2180819064,-0.0837719142,-0.0250673238,0.1623344272,0.4216913283,0.4039413035,0.3650012016,-0.0314708799,0.1082861871,0.3321976662,0.1446735412,-0.3237605393,0.0723176003,0.2941040695,0.1543795019,0.0152103193,-0.2700211704,-0.1584867835,-0.3013113141,-0.3081579208,0.0867515355,0.106804356,-0.3021416664,0.0785163343,-0.1490351856,-0.089051187,0.0587596856,0.0504109636,0.3910552561,-0.0951103568,0.5340352654,-0.3091874719,-0.3310893178,0.0359249711,0.6514391303,-0.0453703329,-0.1939592063,0.1575784087,-0.2966133654,-0.2723999023,-0.1678064018,0.0682685524,0.2438824773,-0.3963705301,-0.0685688779,-0.0029538532,-0.2336229086,0.2451342195,-0.3847491741,0.1689639986,0.1639367044,0.0246436223,-0.2654284835,-0.5594639778,-0.2834984362,-0.2164628208,0.2027434558,0.2369089127,-0.4887759089,0.1068663001,0.1887075305,-0.1841538846,-0.315910548,-0.0630158558,0.1014623567,0.0915858671,-0.0882614851,-0.0840086639,-0.137898609,-0.1152757853,0.0491511263,-0.0711610094,-0.1431347728,-0.0230297372,0.1677424312,0.218734324,-0.3538273871,0.0371720381,-0.2598275542,-0.0885274336,-0.4373986721,0.1055152193,0.2622198761,0.2569029033,0.3842912614,0.2679114342,-0.0695101619,-0.3964473903,0.1665250063,-0.1236992851,-0.2651290596,-0.1265081912,0.2297110856,-0.3186562359,-0.061018616,0.0764735863,-0.1375403255,0.2135011256,0.0946732461,0.1467306316,-0.0578687936,-0.03367294,0.0651393533,0.1982593089,-0.0259977076,-0.0449464396,0.0132999839,0.1737016886,0.0841284096,0.1291260421,-0.2564397156,0.2524062991,-0.1515730619,0.0428198129,0.4236337841,0.2772771716,-0.2070760131,0.3170792758,-0.0381498337,0.1826370507,-0.2552196085,0.3662482202,-0.0064939004,0.1493500471,0.1899190545,0.3061316907,0.2371293902,0.0702329949,0.2771945,0.2158375531,0.1419621855,0.1707586199,-0.3348016143,0.1745646447,-0.6335683465,0.0674178228,-0.078757152,0.0458232053,0.2807557285,0.1773887724,-0.1100368649,0.115252912,-0.1991170496,-0.0117488755,-0.1957808435,-0.098508954,-0.2010090053,0.1680101603,0.1225193962,-0.0301287007,-0.2715129554,-0.178638503,-0.1218053102,0.1585366726,0.2640938163,0.1430709511,0.3340466917,0.0551333055,0.1570996195,-0.0177938323,-0.2439787537,-0.0739672408,-0.3206779957,-0.1343293637,0.2049992234,0.000067359,0.0579926036,0.0511214137,0.1186683327,0.3105690479,0.3402494788,0.0178907719,-0.2919165492,0.2963426411,-0.0827790573,-0.3041669726,-0.0184302125,0.002115153,-0.0544859432,-0.1834938228,-0.241205126,0.0429906137,0.0939483419,0.0925781429,-0.1781886071,-0.5503027439,0.0945282206,-0.0658452138,0.0702487454,0.0940991193,0.4745093286,-0.0591113307,0.1857349724,-0.4849863052,0.0060690916,0.4230893254,0.3215770423,0.4120588601,-0.0990029573,-0.0443120934,-0.0291150678,-0.3054766059,0.1631394625,0.2837891579,0.5087825656,-0.0238904729,-0.1826167107,0.0606078245,-0.1670143157,0.5149814487,-0.2089087218,0.13681674,0.4491739571,-0.2177427262,-0.2192540169,0.1072033644,0.0406980775,0.1862445325,-0.3907720149,0.2738783956,0.2794932723,0.0458572954,-0.2287319601,-0.1771534532,0.2512519062,0.5285198092,0.406250596,0.2499743849,0.0079821469,-0.0523310453,0.3134745061,-0.3260475099,0.1364386231,-0.1692623496,-0.1373758018,-0.1882678717,0.1372042149,-0.5929562449,-0.2624829113,-0.3854259849,-0.096544303,-0.1314247847,0.0482253395,0.0445374548,0.0756680667,-0.2000483423,0.1109349653,0.4055042863,0.3100291193,-0.1148724109,0.3489429057,-0.4045487344,-0.0610268302,0.3257163167,-0.2306214124,-0.2178132385,-0.0727310628,0.338294208,0.1306987405,-0.216746822,-0.0988923982,0.1064758897,0.3080398738,-0.3644308746,-0.1264482737,0.0499947071,-0.141332522,0.0234393962,-0.2185951769,0.5168433189,0.0258426052,-0.2161868513,0.012243934,-0.2569963932]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2065","title":"Only user permission of saved cache files, not group","comments":"That sounds very right to me, @bhavitvyamalik ","body":"Hello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?","comment_length":7,"text":"Only user permission of saved cache files, not group\nHello,\r\n\r\nIt seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?\nThat sounds very right to me, @bhavitvyamalik ","embeddings":[-0.077850841,0.1945603788,-0.1292284131,0.0264360029,-0.0338685513,0.0946465954,0.3274182081,0.1153583676,-0.2828425765,0.0262177214,-0.0821715295,-0.0187454112,0.0873853564,-0.2467288971,-0.0453172065,0.1151472181,0.1780335009,0.0487025566,-0.0159747545,-0.0991487429,-0.0823791772,-0.1133960336,-0.1194860414,-0.0470999889,-0.2884196341,-0.1761447489,0.2531003356,0.2225664407,-0.0323379114,-0.1077931225,0.2468822002,0.4627785087,0.0527875088,0.1555407047,-0.0001133149,-0.2020952851,0.0452580228,-0.0988934413,0.1118607819,0.235458672,-0.1313233972,-0.1396265179,-0.1512697786,-0.1874095649,-0.1279632449,0.1940650791,0.1454287022,-0.6279552579,0.4828821421,0.1859230101,0.2172092199,-0.2187978923,-0.334472537,0.1007299274,0.2463333607,0.0856950805,-0.1789076477,0.0865201578,0.308897227,-0.1798691154,0.0447360426,-0.0043426957,-0.2947018743,0.1093075871,0.1493914276,-0.1214689836,-0.2964897454,-0.5701627731,0.1725436598,0.0878843442,0.7565999031,-0.0710790306,-0.4421907067,-0.1178059205,-0.0070086182,0.2394378334,0.3420344293,0.3599202633,-0.0080125649,0.3022674322,-0.5501104593,-0.0070526996,-0.1685642302,-0.0327100381,0.1736531705,-0.0474882871,-0.0714283958,0.0774406642,-0.0433725715,-0.0512578227,0.1130525023,0.1053653061,-0.3588461578,0.3447251916,-0.1632204652,0.0486851335,-0.2424871475,0.4523541331,0.2012560219,0.2875212431,0.2373650074,0.1883926839,0.1066204607,0.2009605914,-0.0022984277,0.1599903852,0.3922254741,0.2794615924,0.5639121532,-0.3956024945,-0.1239968166,0.031631127,0.0887379274,-0.3854694963,0.0783657879,0.2934468091,0.020045856,-0.2394179255,0.1554283649,0.1124606207,-0.118817389,-0.272672832,-0.01751028,0.2735336721,0.0952352956,-0.1594373733,-0.0117101744,0.1441085339,0.0152256852,-0.3786134124,-0.0089397971,-0.3210310042,-0.3839820027,0.4100488722,0.4567628503,-0.2406007051,0.0312449597,0.2720077336,0.1368939281,-0.2285005003,0.2984774113,-0.0782457441,0.4530949295,0.3220610619,0.2461236119,0.3197130859,0.0571357496,0.0334937349,-0.0789458379,0.3397924304,-0.3481224179,-0.1139799431,0.3107299507,0.0632071272,-0.0404012315,0.2396291345,-0.3680608571,0.0809382647,0.5749333501,-0.1179365218,0.3949382305,0.315841347,-0.1771802306,-0.4328792393,-0.1003816724,0.2516900003,-0.1475661397,0.0115988692,-0.0194355957,-0.0358851105,0.0101287439,0.2701225579,-0.0284244381,0.2501036525,-0.131080091,0.2066724151,0.4657340646,0.0008683493,-0.5675798655,0.1416958421,0.2172107399,-0.217787534,0.1630982906,-0.0465063453,0.0823085159,-0.0818732381,-0.1223698333,0.3039622009,0.1903675497,0.2472683042,0.0909670666,0.1089275852,0.0414825864,-0.1230239198,0.0792809054,0.363848418,0.3990249634,-0.3651306629,0.2048448026,-0.2547181845,0.2688563466,0.4586365223,0.069640547,-0.13428545,-0.0394759551,-0.0514419675,-0.4270368516,0.2756611109,-0.1703832299,-0.3962517977,0.1114787757,-0.3977277279,0.0174828377,-0.3331592381,-0.199418202,-0.0532497875,0.0910096243,-0.0002045859,-0.0091269314,-0.3462994993,0.19000718,0.4061692953,0.2568560243,-0.0537569858,-0.0650390312,0.1307583004,-0.1303946972,-0.0773027688,-0.2626877725,-0.0748970434,0.2733687758,-0.0115552694,-0.2809652984,0.2517764568,0.1214170456,0.0617850311,-0.0591496304,0.2375347614,-0.033019226,0.0605015904,0.0641864017,-0.2169941962,-0.0733028278,-0.1650895178,-0.2724393606,0.3187416494,-0.0660021976,-0.0792162716,-0.2327267081,0.1829765588,0.1435815245,-0.0265037064,-0.1087024957,-0.1407757252,-0.0445999317,0.1990101635,0.3484247923,0.2205441147,0.0578364879,0.0216558818,0.3769901991,0.1736387014,0.1176276729,-0.0807975009,-0.0707196817,0.0855902806,0.1970076263,0.3025594652,0.2666495144,0.0447665416,0.3202512562,-0.1761974543,0.3278062344,-0.1966613531,0.0206487495,-0.077248469,-0.0153632974,-0.1469224691,-0.1561074406,-0.0379708521,-0.054894872,0.3305529356,0.1080549657,0.0812009573,-0.3506090343,-0.1337744147,-0.2111362219,-0.213193655,0.2848464549,-0.0379067436,-0.5447072387,-0.059082482,-0.1810652167,0.13707304,-0.0075990674,0.1759767085,-0.2534243464,0.2190607339,-0.0794411227,-0.094411768,-0.2333145142,0.2412389368,0.0317479968,0.0536436252,0.2344001234,-0.5763405561,0.1876414269,-0.0316408202,0.3001267016,-0.2904071808,0.0556422137,-0.1521506906,-0.0369915888,0.0911744609,-0.512167573,0.0918262303,-0.1773890853,0.0646204725,-0.1498512328,-0.2051414549,-0.2681816518,-0.0543950461,0.0442150719,-0.0436248854,-0.3007307053,-0.1157791689,-0.1214882582,-0.125619784,0.4222796261,0.0957448632,0.2262447625,-0.0105193891,-0.2931213081,-0.3875506818,-0.2200314552,0.2893750072,-0.4056942761,-0.596744895,0.3342914879,-0.4838905334,-0.4180379212,0.0883178636,0.2176100016,0.1116991118,0.4083548486,0.0046635661,0.0192680452,-0.2196452618,0.1822637916,-0.2186732441,0.1385066956,0.0544300526,0.0343210921,0.12994425,-0.0364065841,-0.1751301438,0.0517079011,-0.0187347624,-0.2606351078,0.1737815738,0.0666251257,0.1488120109,0.7675432563,0.456418395,0.3149317205,0.2984256148,0.3146898746,0.6100285649,0.0401061065,-0.1399187446,0.1404712647,-0.0029258479,-0.0849311277,0.1119089648,0.0328086838,-0.2537790537,-0.0178726651,0.0251274221,-0.3068608642,-0.4149442315,-0.2449522913,-0.3320506811,0.1484372318,0.086943835,0.113115795,-0.1921394765,-0.1918792725,0.009506505,0.3157039285,0.2203317881,0.2552585006,-0.4090786874,-0.0138497576,-0.3178271651,0.3680672944,-0.0166543592,0.4647294581,-0.2723039389,0.1788159758,0.2304436415,0.3849028349,0.3558349609,-0.4315281808,0.2551910877,-0.0449922867,0.193858549,0.3841296732,-0.2309416234,0.39791435,0.4955318272,0.0727692023,0.1544339359,0.1550781727,-0.1741650105,-0.4339565039,-0.2610613704,-0.0989723653,-0.2234742045,0.1283208281,0.2552474439,-0.0545524284,-0.0631120354,0.3576787114,-0.3474756479,-0.177917257,-0.0307595469,-0.1399606764,0.2102750093,0.0807347447,-0.0835584924,0.2859620154,0.0654862821,-0.0101381866,0.4172437787,0.1040855497,0.0608603172,0.3833703101,-0.197781086,0.3825981915,-0.1346659064,-0.3164895475,0.4220315814,0.0159902629,-0.3371483386,-0.1519464403,0.1048096865,-0.2008634359,-0.2695474923,-0.1520680934,-0.0880255178,0.0424753018,-0.2393114865,-0.5491760969,0.2317564785,0.578792274,-0.3515723944,0.2864050269,-0.27340734,-0.3435636759,0.0492256992,0.0824009255,0.8245052695,-0.1570135057,0.3399865329,-0.4449856877,0.0615816712,-0.0006294519,-0.4175870419,0.0551954806,-0.4454827309,-0.3959822059,-0.1612460315,-0.1325594187,0.0439851731,0.4727652967,0.0112569472,0.1554724425,-0.1338386387,-0.0563316867,-0.108994849,-0.2401520163,-0.094650954,-0.266266942,0.3211959898,0.1213325933,-0.2361840308,0.0753098205,-0.0442159325,-0.0239133593,-0.2295075953,0.270532757,-0.2532867193,0.0633199364,-0.5600627065,0.0804981291,-0.1705276966,-0.0347039551,0.1901958734,0.3388801813,0.3313163519,0.3677704632,-0.0305669345,0.0163499136,0.335501343,0.2248622477,-0.2331716269,0.090479739,0.2441729754,0.1301910579,0.0394665077,-0.2786605656,-0.1976122558,-0.270770818,-0.3841599822,0.2655350268,0.0055322289,-0.2187041193,-0.0292454083,-0.1014343277,0.0849132091,0.0656941384,0.0971331894,0.4472293854,0.0057016965,0.5958167911,-0.3069600165,-0.3455336392,0.0636834502,0.5843608379,-0.0234861802,-0.2064985633,0.1651492864,-0.3270200789,-0.3019123673,-0.1845724881,0.2274655551,0.087837033,-0.0857013687,-0.0838642046,0.0144805713,-0.0518331677,0.2013657689,-0.4190284014,0.0592180677,0.1349335015,0.119151175,-0.3782561123,-0.3959508836,-0.3127212524,-0.3778339624,0.2120594382,0.2451861054,-0.5309824347,0.2176299393,0.0575032085,-0.2149486691,-0.1813488752,0.0606082566,0.1324714273,0.2562167943,0.0105037959,-0.1692505926,-0.1798910201,-0.1011626124,-0.0052823815,-0.0501799881,-0.2037784457,0.0062615871,0.1503558606,0.2390056998,-0.3489636481,0.1107806042,-0.4261183143,-0.1575119942,-0.3733848035,0.0183751341,0.2796146572,0.2221855074,0.3559332192,0.1717709899,-0.0505776107,-0.1615602523,0.1799246669,-0.1400864869,-0.2125051171,-0.2536296546,0.193122074,-0.2708525062,-0.0752658844,0.0500249118,-0.0535174198,0.117826663,0.0978867859,0.1157162338,-0.0514070354,0.0553387851,0.1485608965,0.2598297298,-0.009996335,-0.1281270981,0.0315552913,0.3280070126,0.1107165515,0.1365115047,-0.2748074234,0.3794934154,-0.0974141732,0.0376169086,0.3225703239,0.2122430056,-0.2396617085,0.1190098748,-0.1187464669,0.0735101253,-0.3696058095,0.4044213891,-0.0978710279,0.0616262183,0.0823138207,0.4420959353,0.2040981948,-0.0174558386,0.3497549295,0.3010709584,0.1624065489,0.3885084987,-0.3277785778,0.1459352821,-0.6631071568,0.1182905585,0.0254892539,-0.0093758721,0.2789162695,0.1154918075,0.0619186759,0.0028432491,-0.3135056496,-0.0107066967,-0.1270486563,-0.1391761452,-0.0814194605,0.0415175036,0.0752453431,0.0842119977,-0.2727113962,-0.1672590822,-0.2246659249,0.2352373153,0.191482842,0.034631595,0.2649727166,0.0632936805,0.1109194756,0.0157302935,-0.1759743392,-0.0381732471,-0.542776227,-0.1189671606,0.2092645317,0.077762641,0.0435708985,0.1508892626,0.1441983581,0.1081217751,0.30235973,0.0327097736,-0.2866697609,0.3757750094,-0.0762383342,-0.3837065995,0.0950137153,0.0285079088,-0.0551988594,-0.0453544036,-0.1456598639,0.0919228122,0.093888402,-0.0919917822,-0.2144003361,-0.4797991514,0.1291687489,-0.0527601093,0.1168146208,0.0630269796,0.5354109406,0.0119308243,0.2287889272,-0.5202322602,0.0408497378,0.4016137719,0.2752881944,0.3932060599,-0.0210444573,-0.0365946256,0.007023498,-0.3666660488,0.1248177141,0.2670272589,0.4878149927,-0.0149682453,-0.1860357523,0.1132207438,-0.1116492227,0.6541266441,-0.2681244612,0.2810773551,0.3759559691,-0.2325328737,-0.2354937643,0.13346228,0.0004526077,0.0816905499,-0.3390656412,0.3469357193,0.1778548956,0.0747392997,-0.1564865708,-0.1837963164,0.1751953363,0.5109377503,0.4153042734,0.2217810005,0.0105212349,-0.0107666096,0.4294727445,-0.3230730593,0.1443816572,-0.148446098,-0.0348047614,-0.190640226,-0.0045428099,-0.5637736917,-0.1479047239,-0.3426666558,0.0024920329,-0.162657842,0.3042175174,0.0169887934,0.0003928628,-0.1439859718,0.0324618369,0.2553045452,0.2248649299,-0.1624774039,0.3524665534,-0.4517415166,-0.0936703086,0.2820447385,-0.2153718323,-0.2549439371,-0.048770722,0.3724588454,0.1851778924,-0.1634171009,0.0564100519,0.064816758,0.3787714243,-0.4221010208,-0.0667258948,0.0511902794,-0.4351349473,0.0552484542,-0.1271036714,0.5182294846,0.005256684,-0.1787897199,0.0367546231,-0.2394028902]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2061","title":"Cannot load udpos subsets from xtreme dataset using load_dataset()","comments":"@lhoestq Adding \"_\" to the class labels in the dataset script will fix the issue.\r\n\r\nThe bigger issue IMO is that the data files are in conll format, but the examples are tokens, not sentences.","body":"Hello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n","comment_length":35,"text":"Cannot load udpos subsets from xtreme dataset using load_dataset()\nHello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n\n@lhoestq Adding \"_\" to the class labels in the dataset script will fix the issue.\r\n\r\nThe bigger issue IMO is that the data files are in conll format, but the examples are tokens, not sentences.","embeddings":[-0.3464027941,-0.0752857924,0.002613897,0.4443909824,0.3713300824,0.069119066,0.2405683845,0.0394754633,0.5101771355,0.2061247975,-0.3589054644,0.0527050309,0.0860960111,0.0808662176,0.1144125015,-0.3151043355,-0.0602493994,0.0173558537,-0.2503674328,-0.065504171,-0.4105169475,-0.0623261929,-0.3046581447,0.1095427871,-0.1711033434,0.2064580172,0.0141455419,0.1647338569,-0.3150216639,-0.4950218201,0.5633596182,0.0180328377,0.5800744891,0.221322,-0.0001253432,0.1060385182,0.4861793816,-0.0613055825,-0.3452345729,-0.4049583972,-0.4189888239,-0.1315970272,-0.0690613464,-0.0964650735,0.1045276895,-0.0043987469,-0.0823340118,-0.2585622072,0.3321028948,0.3678756952,0.0941703171,0.3720282614,0.061857868,-0.2211653441,0.4757196605,-0.0422705263,-0.1489076465,0.1563735753,0.4271000326,-0.3246812224,0.1801449507,0.1246253327,-0.209139958,0.2249173075,-0.1637178361,0.0560058244,0.1821350753,-0.2583509088,0.2391248494,0.1349655539,0.8412337303,-0.1512385607,-0.4147819579,0.0182388742,-0.1014063731,-0.3554817438,0.1830140948,0.0008575391,-0.0578870513,0.0788415372,-0.0767996833,-0.0602417439,0.0503758267,0.289073348,-0.1434578896,0.2910360396,0.0541204214,0.1338610649,0.3239519,-0.1089203432,0.4901404083,0.0979419872,-0.0730714723,0.2410603017,-0.3231164515,0.2865316868,-0.0756045282,-0.3037876189,0.2389643937,-0.1045979336,-0.1012069955,0.0039554867,-0.1443649828,0.3295553327,0.4467023611,0.258592695,0.3614312708,0.2899973094,0.1072534993,0.1587685496,-0.0565737747,-0.0633865967,-0.4474610388,-0.2160627842,0.0857923925,-0.0271296799,0.1571703404,-0.4506189227,-0.5827982426,0.0448348261,-0.1629013866,-0.217883721,0.05834952,0.2688898146,0.1545642018,0.4297795296,0.0458623022,0.4458411336,-0.4962795973,-0.2659801245,-0.0544239134,0.1687626541,-0.4189635813,0.0585133731,0.1439427137,0.0052871024,0.2148580551,0.0167755205,0.2656741142,-0.2452537715,0.1288290769,-0.2358862758,-0.244421348,0.3540515602,0.3105243742,0.1218978986,0.2711307406,-0.3746761978,-0.0785986483,0.2891557217,-0.2225109637,-0.2350116372,0.0909441113,0.0328545943,-0.1796912402,0.0550205037,-1.0357831717,-0.0764572024,0.0851031467,0.1069326326,-0.0155084543,-0.275529623,-0.1712114811,0.0025955918,0.2848057449,0.5355801582,-0.2868715227,-0.0464064963,-0.1408394873,-0.0400896706,0.3355934024,0.3613086641,-0.3357942998,0.0737545416,-0.1279850602,0.1057119593,0.1982139647,-0.5150758028,-0.4803656042,0.1640977561,0.0509281084,-0.0161880385,-0.0857606456,-0.3274516165,0.2393174767,0.0357203931,0.1953124404,0.2589243054,0.0580973849,-0.0908872336,-0.1751016676,-0.0985951722,0.136113286,0.4271915853,0.2911248505,0.1453995854,0.2614878714,0.4392369092,0.0436218232,0.0165753514,0.0131826354,0.3216635585,0.0333599299,0.0902014673,0.0362407826,-0.0593443327,-0.4222900867,0.2021975517,-0.0930997282,-0.0453706384,0.0975539014,0.1812404096,-0.3897111714,0.1230222583,-0.3247447312,-0.3197046518,-0.0942896307,0.0866035298,-0.0964135006,-0.0818081424,-0.1698002815,0.0806680992,-0.0715484098,0.1103202999,-0.3754514456,0.4191193879,0.049578242,-0.0648903623,0.0807236657,0.2063002884,0.2299803942,-0.1260796189,-0.1753814518,0.1877421886,0.6032566428,-0.1791830808,0.1357364804,-0.0441033766,0.3042578995,-0.3885697424,0.1962021738,-0.0273815524,0.1187217534,0.1130647287,0.0598304868,0.254188925,-0.0359747075,0.1840643734,-0.1281909943,0.1224722341,0.3874266744,0.0022266286,0.0613593124,-0.2924954891,0.5102493167,0.3250283897,0.3184416592,0.0524797477,-0.1459858865,-0.2242728472,-0.1109304726,-0.1516300291,0.0816720501,0.1515698433,-0.2236988097,0.0444757342,0.2037227601,0.2373357564,0.32943362,0.0214671995,0.0758472681,-0.0468033701,0.0237630997,0.0701770708,-0.0097976169,-0.0718164816,-0.1015756652,0.0644405559,0.0008363276,-0.2057575434,-0.2248116881,0.0729394481,0.2377972454,0.4230693877,-0.300316304,0.0473215431,-0.4231781065,-0.1595294625,0.1389964968,-0.0802792013,-0.0387310572,-0.1483372748,-0.3189133108,0.2778151631,0.3382665813,0.1288223416,0.1079095602,-0.239862293,-0.0247332957,-0.2022278309,0.1674776077,-0.1916826367,-0.1363240331,-0.1142772362,0.2343533635,0.1731068641,0.0974104404,-0.3408343494,-0.0454138368,-0.2319974899,-0.0555951707,0.2320308834,-0.0026652913,0.30958727,-0.0770000219,0.14246355,-0.3586760759,0.0364723913,0.3661396801,0.0212173313,-0.1678794324,0.2677277923,-0.0123514775,-0.1251045763,0.0236814786,-0.14731206,-0.299810648,-0.2093851715,0.0171216857,0.0482656993,0.0239049792,-0.0471153446,0.0395985246,0.0380500816,0.1774878949,-0.1749292463,-0.340343833,-0.4319575131,0.4691960514,-0.2710248232,-0.2475002557,0.0816968828,0.0473560803,0.2014110088,0.3418824375,-0.4603409767,-0.0737863109,-0.1173928976,0.1646256745,0.0758609176,0.1864943802,0.0469675176,-0.1311487854,0.1585203111,-0.0770411864,-0.3395670652,-0.1058110222,0.2425631136,0.2578392923,-0.1052164957,0.4483567774,-0.2508384287,0.2393931001,-0.0129460888,0.2891106904,0.5426401496,-0.0893333852,0.1529127359,-0.3217939138,-0.3143526018,-0.099711448,-0.1948753744,-0.1389905959,0.0483570509,0.0269657038,-0.1491460651,-0.2706747651,-0.0529081263,-0.0879614204,-0.1746995449,-0.0553434081,-0.0693777204,0.3293852508,-0.0504932068,0.2791500986,0.0136173237,-0.0404188186,-0.1539395452,0.2246424258,0.1235996708,0.2225517929,-0.3244818151,0.1071843952,-0.2705765367,0.3529714644,-0.0407876782,0.7153809667,-0.1447271109,0.3261329234,0.1128169373,0.0381268747,0.1055565476,-0.2915969193,0.1107803509,0.0259097721,-0.4394381642,-0.0571818091,-0.0327192396,-0.0487926118,-0.0640700683,-0.029191507,0.2615399957,-0.2473840266,-0.1046056673,-0.0212867279,0.3048670292,-0.0421134904,-0.2426180691,-0.3949687183,-0.5412188768,-0.4392507076,-0.1890492439,0.1465301365,0.1577130854,-0.0798719376,0.2045397013,-0.1588291377,0.1053305566,0.2242294401,0.084091939,0.2725592852,-0.0330140777,0.1534871459,0.2682338357,-0.1665609181,0.7140191793,0.7792875171,-0.1654731035,-0.3519219756,0.253123343,-0.0765697509,0.0217691138,-0.0563563295,-0.0262792483,0.0473596342,-0.199334532,-0.1701831073,-0.0464213379,0.198322013,0.1967192143,0.1544240564,-0.292883724,-0.7532656789,0.5231232643,0.0733509213,0.1194912866,0.2948041558,-0.0994655937,-0.1877111197,0.2691773474,0.1097404063,0.7309182882,-0.4024430215,0.0663996488,0.1255254596,0.0390540808,0.0969472975,-0.0389399156,-0.2549379468,-0.1899633855,-0.2134175152,-0.1464372128,-0.1630649865,0.2815146148,0.1523412913,-0.0743444934,0.2819465995,-0.1734498143,-0.0285564661,0.2377475947,0.1179537997,-0.0852734521,-0.3346666694,-0.5701636672,0.0695263073,-0.1744922549,0.2196643502,-0.0229284316,0.0446263291,-0.0770891085,-0.0558210611,-0.3917295933,0.3648408055,-0.2748184204,0.3384894431,-0.1301485598,-0.3494686484,0.189233914,0.4507753253,0.2284569144,0.1401094347,-0.1527668387,0.0742595568,0.2642703056,-0.2523836493,-0.0304523762,-0.1002178192,-0.0693553537,-0.0230432879,-0.2731992006,0.1419645548,-0.0702483132,-0.2247540653,-0.0742935315,0.1391913295,0.0287422985,-0.0994305909,-0.3413546085,-0.071890451,0.0351187959,-0.1643306762,-0.0067225127,0.209125787,-0.0939807147,0.0110958582,0.0571492016,-0.3362408876,0.0208268687,0.4598697722,-0.0671367273,-0.1035425216,0.5126202703,0.2207635939,-0.2053335756,-0.1291891932,0.2668746114,0.343277216,-0.2435242385,0.0247940086,-0.1015284583,0.1226382703,0.0169650894,0.0190838203,0.2321757823,-0.3506089747,0.3566080034,-0.59125036,-0.3739056289,0.1904153675,-0.1842523664,0.0209109206,-0.0230076816,0.0114088757,-0.1538707614,-0.0855408907,-0.1131536886,0.1104641557,-0.3555292189,0.0531755239,0.2333114892,-0.0038374737,0.026617974,-0.1702212244,0.0084072324,0.1382467002,0.0358453728,-0.0385317653,-0.1823156774,0.2142098844,0.1291017532,-0.346626848,0.0095422855,-0.2173159271,-0.0406782925,-0.1456348747,0.175234288,0.4642310739,-0.0508853458,0.1162913889,0.2020449042,0.1192648113,-0.3610048294,0.377338171,-0.0597466715,0.2192406356,0.1052805781,0.1996854097,0.2261882126,0.1284537315,-0.298638612,-0.0918607861,0.254362464,-0.0207492411,0.2784873545,-0.424654901,0.3846978843,0.093402341,0.1392694265,0.4423628151,-0.2708707154,0.0668304861,0.3226239681,0.0277718697,-0.239678219,-0.1068023145,0.1143150181,0.0567229837,-0.1326247156,0.3174369335,0.3981636763,0.0658837408,-0.1405836493,0.0366226286,0.0629203171,-0.1436378211,0.0519923568,0.6180887222,0.018257726,0.2121533453,0.0627515018,0.2251056135,-0.0261530075,0.3331249356,-0.1577174217,0.4461428821,0.2134561688,0.1908597201,-0.3277816176,-0.5346593857,-0.1699689627,0.3204804361,0.0944019482,-0.0257126112,-0.208639726,0.2422643751,-0.1822718382,0.1587971896,-0.0697560012,0.1305374503,-0.034036275,-0.1838852763,-0.1329226047,-0.1520823389,-0.5502743721,0.1870924532,-0.2470207661,-0.1265384704,0.1939038336,0.1568964869,-0.1329802722,-0.3558672369,-0.0509607121,0.3773249984,-0.3437860012,-0.3400719464,0.3964489698,0.0177266039,0.0172705334,0.2495322376,0.1393802017,0.3306991458,0.420499593,-0.3608880341,-0.1154865772,0.0613388866,0.0774012357,0.0765652359,0.1979313791,0.0195528045,0.2065385729,0.3709867597,-0.0114097325,-0.0139406091,-0.0316576846,0.1572707593,-0.2297257632,-0.0879770666,0.2299744934,-0.0481610782,0.199876681,-0.117637977,0.0245101266,-0.2423740178,0.021785982,0.5751782656,0.0103393085,0.1068291664,0.0592920072,-0.0527461395,-0.1844878495,0.3957369924,0.6384893656,0.3486227691,-0.4021956027,-0.1908638626,-0.7238464952,0.0598110072,-0.4783252478,-0.2568792999,0.5058496594,-0.0741570145,0.101831302,0.3912000656,-0.0939410925,0.2525469065,0.1857650876,0.2951029241,-0.2620002925,-0.2589090765,0.4163457751,0.2090041637,-0.1786013097,-0.4936587512,-0.0105341868,0.042624142,-0.0922619253,-0.1824915558,0.0417558402,0.1412569433,0.3790921271,0.1933657229,0.0584114566,0.7760936022,-0.0257243253,-0.064801462,-0.0789829865,-0.6323983669,-0.054236386,0.0605209023,0.0649564788,0.1296199411,-0.1022136584,-0.1126848385,-0.3824729323,0.3061400652,-0.1255756468,-0.1483518481,-0.2318839282,-0.0043637305,-0.0508504957,-0.0628491864,-0.0281334668,0.0610886924,0.1224483997,0.3552176952,-0.125965029,-0.2262785584,0.3989257812,-0.0534627326,-0.4505837858,0.0725823417,-0.1192312092,0.3350661695,-0.2118573189,-0.461424768,0.1921404898,0.2282822132,-0.1450064182,-0.3645923734,0.0354042202,0.1440638602,0.1055092961,-0.1191350445,0.6080466509,0.0068141595,-0.158861652,0.1072711051,-0.1721357256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2061","title":"Cannot load udpos subsets from xtreme dataset using load_dataset()","comments":"Hi ! Thanks for reporting @adzcodez \r\n\r\n\r\n> @lhoestq Adding \"_\" to the class labels in the dataset script will fix the issue.\r\n> \r\n> The bigger issue IMO is that the data files are in conll format, but the examples are tokens, not sentences.\r\n\r\nYou're right: \"_\" should be added to the list of labels, and the examples must be sequences of tokens, not singles tokens.\r\n","body":"Hello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n","comment_length":66,"text":"Cannot load udpos subsets from xtreme dataset using load_dataset()\nHello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n\nHi ! Thanks for reporting @adzcodez \r\n\r\n\r\n> @lhoestq Adding \"_\" to the class labels in the dataset script will fix the issue.\r\n> \r\n> The bigger issue IMO is that the data files are in conll format, but the examples are tokens, not sentences.\r\n\r\nYou're right: \"_\" should be added to the list of labels, and the examples must be sequences of tokens, not singles tokens.\r\n","embeddings":[-0.3464027941,-0.0752857924,0.002613897,0.4443909824,0.3713300824,0.069119066,0.2405683845,0.0394754633,0.5101771355,0.2061247975,-0.3589054644,0.0527050309,0.0860960111,0.0808662176,0.1144125015,-0.3151043355,-0.0602493994,0.0173558537,-0.2503674328,-0.065504171,-0.4105169475,-0.0623261929,-0.3046581447,0.1095427871,-0.1711033434,0.2064580172,0.0141455419,0.1647338569,-0.3150216639,-0.4950218201,0.5633596182,0.0180328377,0.5800744891,0.221322,-0.0001253432,0.1060385182,0.4861793816,-0.0613055825,-0.3452345729,-0.4049583972,-0.4189888239,-0.1315970272,-0.0690613464,-0.0964650735,0.1045276895,-0.0043987469,-0.0823340118,-0.2585622072,0.3321028948,0.3678756952,0.0941703171,0.3720282614,0.061857868,-0.2211653441,0.4757196605,-0.0422705263,-0.1489076465,0.1563735753,0.4271000326,-0.3246812224,0.1801449507,0.1246253327,-0.209139958,0.2249173075,-0.1637178361,0.0560058244,0.1821350753,-0.2583509088,0.2391248494,0.1349655539,0.8412337303,-0.1512385607,-0.4147819579,0.0182388742,-0.1014063731,-0.3554817438,0.1830140948,0.0008575391,-0.0578870513,0.0788415372,-0.0767996833,-0.0602417439,0.0503758267,0.289073348,-0.1434578896,0.2910360396,0.0541204214,0.1338610649,0.3239519,-0.1089203432,0.4901404083,0.0979419872,-0.0730714723,0.2410603017,-0.3231164515,0.2865316868,-0.0756045282,-0.3037876189,0.2389643937,-0.1045979336,-0.1012069955,0.0039554867,-0.1443649828,0.3295553327,0.4467023611,0.258592695,0.3614312708,0.2899973094,0.1072534993,0.1587685496,-0.0565737747,-0.0633865967,-0.4474610388,-0.2160627842,0.0857923925,-0.0271296799,0.1571703404,-0.4506189227,-0.5827982426,0.0448348261,-0.1629013866,-0.217883721,0.05834952,0.2688898146,0.1545642018,0.4297795296,0.0458623022,0.4458411336,-0.4962795973,-0.2659801245,-0.0544239134,0.1687626541,-0.4189635813,0.0585133731,0.1439427137,0.0052871024,0.2148580551,0.0167755205,0.2656741142,-0.2452537715,0.1288290769,-0.2358862758,-0.244421348,0.3540515602,0.3105243742,0.1218978986,0.2711307406,-0.3746761978,-0.0785986483,0.2891557217,-0.2225109637,-0.2350116372,0.0909441113,0.0328545943,-0.1796912402,0.0550205037,-1.0357831717,-0.0764572024,0.0851031467,0.1069326326,-0.0155084543,-0.275529623,-0.1712114811,0.0025955918,0.2848057449,0.5355801582,-0.2868715227,-0.0464064963,-0.1408394873,-0.0400896706,0.3355934024,0.3613086641,-0.3357942998,0.0737545416,-0.1279850602,0.1057119593,0.1982139647,-0.5150758028,-0.4803656042,0.1640977561,0.0509281084,-0.0161880385,-0.0857606456,-0.3274516165,0.2393174767,0.0357203931,0.1953124404,0.2589243054,0.0580973849,-0.0908872336,-0.1751016676,-0.0985951722,0.136113286,0.4271915853,0.2911248505,0.1453995854,0.2614878714,0.4392369092,0.0436218232,0.0165753514,0.0131826354,0.3216635585,0.0333599299,0.0902014673,0.0362407826,-0.0593443327,-0.4222900867,0.2021975517,-0.0930997282,-0.0453706384,0.0975539014,0.1812404096,-0.3897111714,0.1230222583,-0.3247447312,-0.3197046518,-0.0942896307,0.0866035298,-0.0964135006,-0.0818081424,-0.1698002815,0.0806680992,-0.0715484098,0.1103202999,-0.3754514456,0.4191193879,0.049578242,-0.0648903623,0.0807236657,0.2063002884,0.2299803942,-0.1260796189,-0.1753814518,0.1877421886,0.6032566428,-0.1791830808,0.1357364804,-0.0441033766,0.3042578995,-0.3885697424,0.1962021738,-0.0273815524,0.1187217534,0.1130647287,0.0598304868,0.254188925,-0.0359747075,0.1840643734,-0.1281909943,0.1224722341,0.3874266744,0.0022266286,0.0613593124,-0.2924954891,0.5102493167,0.3250283897,0.3184416592,0.0524797477,-0.1459858865,-0.2242728472,-0.1109304726,-0.1516300291,0.0816720501,0.1515698433,-0.2236988097,0.0444757342,0.2037227601,0.2373357564,0.32943362,0.0214671995,0.0758472681,-0.0468033701,0.0237630997,0.0701770708,-0.0097976169,-0.0718164816,-0.1015756652,0.0644405559,0.0008363276,-0.2057575434,-0.2248116881,0.0729394481,0.2377972454,0.4230693877,-0.300316304,0.0473215431,-0.4231781065,-0.1595294625,0.1389964968,-0.0802792013,-0.0387310572,-0.1483372748,-0.3189133108,0.2778151631,0.3382665813,0.1288223416,0.1079095602,-0.239862293,-0.0247332957,-0.2022278309,0.1674776077,-0.1916826367,-0.1363240331,-0.1142772362,0.2343533635,0.1731068641,0.0974104404,-0.3408343494,-0.0454138368,-0.2319974899,-0.0555951707,0.2320308834,-0.0026652913,0.30958727,-0.0770000219,0.14246355,-0.3586760759,0.0364723913,0.3661396801,0.0212173313,-0.1678794324,0.2677277923,-0.0123514775,-0.1251045763,0.0236814786,-0.14731206,-0.299810648,-0.2093851715,0.0171216857,0.0482656993,0.0239049792,-0.0471153446,0.0395985246,0.0380500816,0.1774878949,-0.1749292463,-0.340343833,-0.4319575131,0.4691960514,-0.2710248232,-0.2475002557,0.0816968828,0.0473560803,0.2014110088,0.3418824375,-0.4603409767,-0.0737863109,-0.1173928976,0.1646256745,0.0758609176,0.1864943802,0.0469675176,-0.1311487854,0.1585203111,-0.0770411864,-0.3395670652,-0.1058110222,0.2425631136,0.2578392923,-0.1052164957,0.4483567774,-0.2508384287,0.2393931001,-0.0129460888,0.2891106904,0.5426401496,-0.0893333852,0.1529127359,-0.3217939138,-0.3143526018,-0.099711448,-0.1948753744,-0.1389905959,0.0483570509,0.0269657038,-0.1491460651,-0.2706747651,-0.0529081263,-0.0879614204,-0.1746995449,-0.0553434081,-0.0693777204,0.3293852508,-0.0504932068,0.2791500986,0.0136173237,-0.0404188186,-0.1539395452,0.2246424258,0.1235996708,0.2225517929,-0.3244818151,0.1071843952,-0.2705765367,0.3529714644,-0.0407876782,0.7153809667,-0.1447271109,0.3261329234,0.1128169373,0.0381268747,0.1055565476,-0.2915969193,0.1107803509,0.0259097721,-0.4394381642,-0.0571818091,-0.0327192396,-0.0487926118,-0.0640700683,-0.029191507,0.2615399957,-0.2473840266,-0.1046056673,-0.0212867279,0.3048670292,-0.0421134904,-0.2426180691,-0.3949687183,-0.5412188768,-0.4392507076,-0.1890492439,0.1465301365,0.1577130854,-0.0798719376,0.2045397013,-0.1588291377,0.1053305566,0.2242294401,0.084091939,0.2725592852,-0.0330140777,0.1534871459,0.2682338357,-0.1665609181,0.7140191793,0.7792875171,-0.1654731035,-0.3519219756,0.253123343,-0.0765697509,0.0217691138,-0.0563563295,-0.0262792483,0.0473596342,-0.199334532,-0.1701831073,-0.0464213379,0.198322013,0.1967192143,0.1544240564,-0.292883724,-0.7532656789,0.5231232643,0.0733509213,0.1194912866,0.2948041558,-0.0994655937,-0.1877111197,0.2691773474,0.1097404063,0.7309182882,-0.4024430215,0.0663996488,0.1255254596,0.0390540808,0.0969472975,-0.0389399156,-0.2549379468,-0.1899633855,-0.2134175152,-0.1464372128,-0.1630649865,0.2815146148,0.1523412913,-0.0743444934,0.2819465995,-0.1734498143,-0.0285564661,0.2377475947,0.1179537997,-0.0852734521,-0.3346666694,-0.5701636672,0.0695263073,-0.1744922549,0.2196643502,-0.0229284316,0.0446263291,-0.0770891085,-0.0558210611,-0.3917295933,0.3648408055,-0.2748184204,0.3384894431,-0.1301485598,-0.3494686484,0.189233914,0.4507753253,0.2284569144,0.1401094347,-0.1527668387,0.0742595568,0.2642703056,-0.2523836493,-0.0304523762,-0.1002178192,-0.0693553537,-0.0230432879,-0.2731992006,0.1419645548,-0.0702483132,-0.2247540653,-0.0742935315,0.1391913295,0.0287422985,-0.0994305909,-0.3413546085,-0.071890451,0.0351187959,-0.1643306762,-0.0067225127,0.209125787,-0.0939807147,0.0110958582,0.0571492016,-0.3362408876,0.0208268687,0.4598697722,-0.0671367273,-0.1035425216,0.5126202703,0.2207635939,-0.2053335756,-0.1291891932,0.2668746114,0.343277216,-0.2435242385,0.0247940086,-0.1015284583,0.1226382703,0.0169650894,0.0190838203,0.2321757823,-0.3506089747,0.3566080034,-0.59125036,-0.3739056289,0.1904153675,-0.1842523664,0.0209109206,-0.0230076816,0.0114088757,-0.1538707614,-0.0855408907,-0.1131536886,0.1104641557,-0.3555292189,0.0531755239,0.2333114892,-0.0038374737,0.026617974,-0.1702212244,0.0084072324,0.1382467002,0.0358453728,-0.0385317653,-0.1823156774,0.2142098844,0.1291017532,-0.346626848,0.0095422855,-0.2173159271,-0.0406782925,-0.1456348747,0.175234288,0.4642310739,-0.0508853458,0.1162913889,0.2020449042,0.1192648113,-0.3610048294,0.377338171,-0.0597466715,0.2192406356,0.1052805781,0.1996854097,0.2261882126,0.1284537315,-0.298638612,-0.0918607861,0.254362464,-0.0207492411,0.2784873545,-0.424654901,0.3846978843,0.093402341,0.1392694265,0.4423628151,-0.2708707154,0.0668304861,0.3226239681,0.0277718697,-0.239678219,-0.1068023145,0.1143150181,0.0567229837,-0.1326247156,0.3174369335,0.3981636763,0.0658837408,-0.1405836493,0.0366226286,0.0629203171,-0.1436378211,0.0519923568,0.6180887222,0.018257726,0.2121533453,0.0627515018,0.2251056135,-0.0261530075,0.3331249356,-0.1577174217,0.4461428821,0.2134561688,0.1908597201,-0.3277816176,-0.5346593857,-0.1699689627,0.3204804361,0.0944019482,-0.0257126112,-0.208639726,0.2422643751,-0.1822718382,0.1587971896,-0.0697560012,0.1305374503,-0.034036275,-0.1838852763,-0.1329226047,-0.1520823389,-0.5502743721,0.1870924532,-0.2470207661,-0.1265384704,0.1939038336,0.1568964869,-0.1329802722,-0.3558672369,-0.0509607121,0.3773249984,-0.3437860012,-0.3400719464,0.3964489698,0.0177266039,0.0172705334,0.2495322376,0.1393802017,0.3306991458,0.420499593,-0.3608880341,-0.1154865772,0.0613388866,0.0774012357,0.0765652359,0.1979313791,0.0195528045,0.2065385729,0.3709867597,-0.0114097325,-0.0139406091,-0.0316576846,0.1572707593,-0.2297257632,-0.0879770666,0.2299744934,-0.0481610782,0.199876681,-0.117637977,0.0245101266,-0.2423740178,0.021785982,0.5751782656,0.0103393085,0.1068291664,0.0592920072,-0.0527461395,-0.1844878495,0.3957369924,0.6384893656,0.3486227691,-0.4021956027,-0.1908638626,-0.7238464952,0.0598110072,-0.4783252478,-0.2568792999,0.5058496594,-0.0741570145,0.101831302,0.3912000656,-0.0939410925,0.2525469065,0.1857650876,0.2951029241,-0.2620002925,-0.2589090765,0.4163457751,0.2090041637,-0.1786013097,-0.4936587512,-0.0105341868,0.042624142,-0.0922619253,-0.1824915558,0.0417558402,0.1412569433,0.3790921271,0.1933657229,0.0584114566,0.7760936022,-0.0257243253,-0.064801462,-0.0789829865,-0.6323983669,-0.054236386,0.0605209023,0.0649564788,0.1296199411,-0.1022136584,-0.1126848385,-0.3824729323,0.3061400652,-0.1255756468,-0.1483518481,-0.2318839282,-0.0043637305,-0.0508504957,-0.0628491864,-0.0281334668,0.0610886924,0.1224483997,0.3552176952,-0.125965029,-0.2262785584,0.3989257812,-0.0534627326,-0.4505837858,0.0725823417,-0.1192312092,0.3350661695,-0.2118573189,-0.461424768,0.1921404898,0.2282822132,-0.1450064182,-0.3645923734,0.0354042202,0.1440638602,0.1055092961,-0.1191350445,0.6080466509,0.0068141595,-0.158861652,0.1072711051,-0.1721357256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2061","title":"Cannot load udpos subsets from xtreme dataset using load_dataset()","comments":"@lhoestq Can you please label this issue with the \"good first issue\" label? I'm not sure I'll find time to fix this.\r\n\r\nTo resolve it, the user should:\r\n1. add `\"_\"` to the list of labels\r\n2. transform the udpos subset to the conll format (I think the preprocessing logic can be borrowed from [the original repo](https:\/\/github.com\/google-research\/xtreme\/blob\/58a76a0d02458c4b3b6a742d3fd4ffaca80ff0de\/utils_preprocess.py#L187-L204))\r\n3. update the dummy data\r\n4. update the dataset info\r\n5. [optional] add info about the data fields structure of the udpos subset to the dataset readme","body":"Hello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n","comment_length":84,"text":"Cannot load udpos subsets from xtreme dataset using load_dataset()\nHello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n\n@lhoestq Can you please label this issue with the \"good first issue\" label? I'm not sure I'll find time to fix this.\r\n\r\nTo resolve it, the user should:\r\n1. add `\"_\"` to the list of labels\r\n2. transform the udpos subset to the conll format (I think the preprocessing logic can be borrowed from [the original repo](https:\/\/github.com\/google-research\/xtreme\/blob\/58a76a0d02458c4b3b6a742d3fd4ffaca80ff0de\/utils_preprocess.py#L187-L204))\r\n3. update the dummy data\r\n4. update the dataset info\r\n5. [optional] add info about the data fields structure of the udpos subset to the dataset readme","embeddings":[-0.3464027941,-0.0752857924,0.002613897,0.4443909824,0.3713300824,0.069119066,0.2405683845,0.0394754633,0.5101771355,0.2061247975,-0.3589054644,0.0527050309,0.0860960111,0.0808662176,0.1144125015,-0.3151043355,-0.0602493994,0.0173558537,-0.2503674328,-0.065504171,-0.4105169475,-0.0623261929,-0.3046581447,0.1095427871,-0.1711033434,0.2064580172,0.0141455419,0.1647338569,-0.3150216639,-0.4950218201,0.5633596182,0.0180328377,0.5800744891,0.221322,-0.0001253432,0.1060385182,0.4861793816,-0.0613055825,-0.3452345729,-0.4049583972,-0.4189888239,-0.1315970272,-0.0690613464,-0.0964650735,0.1045276895,-0.0043987469,-0.0823340118,-0.2585622072,0.3321028948,0.3678756952,0.0941703171,0.3720282614,0.061857868,-0.2211653441,0.4757196605,-0.0422705263,-0.1489076465,0.1563735753,0.4271000326,-0.3246812224,0.1801449507,0.1246253327,-0.209139958,0.2249173075,-0.1637178361,0.0560058244,0.1821350753,-0.2583509088,0.2391248494,0.1349655539,0.8412337303,-0.1512385607,-0.4147819579,0.0182388742,-0.1014063731,-0.3554817438,0.1830140948,0.0008575391,-0.0578870513,0.0788415372,-0.0767996833,-0.0602417439,0.0503758267,0.289073348,-0.1434578896,0.2910360396,0.0541204214,0.1338610649,0.3239519,-0.1089203432,0.4901404083,0.0979419872,-0.0730714723,0.2410603017,-0.3231164515,0.2865316868,-0.0756045282,-0.3037876189,0.2389643937,-0.1045979336,-0.1012069955,0.0039554867,-0.1443649828,0.3295553327,0.4467023611,0.258592695,0.3614312708,0.2899973094,0.1072534993,0.1587685496,-0.0565737747,-0.0633865967,-0.4474610388,-0.2160627842,0.0857923925,-0.0271296799,0.1571703404,-0.4506189227,-0.5827982426,0.0448348261,-0.1629013866,-0.217883721,0.05834952,0.2688898146,0.1545642018,0.4297795296,0.0458623022,0.4458411336,-0.4962795973,-0.2659801245,-0.0544239134,0.1687626541,-0.4189635813,0.0585133731,0.1439427137,0.0052871024,0.2148580551,0.0167755205,0.2656741142,-0.2452537715,0.1288290769,-0.2358862758,-0.244421348,0.3540515602,0.3105243742,0.1218978986,0.2711307406,-0.3746761978,-0.0785986483,0.2891557217,-0.2225109637,-0.2350116372,0.0909441113,0.0328545943,-0.1796912402,0.0550205037,-1.0357831717,-0.0764572024,0.0851031467,0.1069326326,-0.0155084543,-0.275529623,-0.1712114811,0.0025955918,0.2848057449,0.5355801582,-0.2868715227,-0.0464064963,-0.1408394873,-0.0400896706,0.3355934024,0.3613086641,-0.3357942998,0.0737545416,-0.1279850602,0.1057119593,0.1982139647,-0.5150758028,-0.4803656042,0.1640977561,0.0509281084,-0.0161880385,-0.0857606456,-0.3274516165,0.2393174767,0.0357203931,0.1953124404,0.2589243054,0.0580973849,-0.0908872336,-0.1751016676,-0.0985951722,0.136113286,0.4271915853,0.2911248505,0.1453995854,0.2614878714,0.4392369092,0.0436218232,0.0165753514,0.0131826354,0.3216635585,0.0333599299,0.0902014673,0.0362407826,-0.0593443327,-0.4222900867,0.2021975517,-0.0930997282,-0.0453706384,0.0975539014,0.1812404096,-0.3897111714,0.1230222583,-0.3247447312,-0.3197046518,-0.0942896307,0.0866035298,-0.0964135006,-0.0818081424,-0.1698002815,0.0806680992,-0.0715484098,0.1103202999,-0.3754514456,0.4191193879,0.049578242,-0.0648903623,0.0807236657,0.2063002884,0.2299803942,-0.1260796189,-0.1753814518,0.1877421886,0.6032566428,-0.1791830808,0.1357364804,-0.0441033766,0.3042578995,-0.3885697424,0.1962021738,-0.0273815524,0.1187217534,0.1130647287,0.0598304868,0.254188925,-0.0359747075,0.1840643734,-0.1281909943,0.1224722341,0.3874266744,0.0022266286,0.0613593124,-0.2924954891,0.5102493167,0.3250283897,0.3184416592,0.0524797477,-0.1459858865,-0.2242728472,-0.1109304726,-0.1516300291,0.0816720501,0.1515698433,-0.2236988097,0.0444757342,0.2037227601,0.2373357564,0.32943362,0.0214671995,0.0758472681,-0.0468033701,0.0237630997,0.0701770708,-0.0097976169,-0.0718164816,-0.1015756652,0.0644405559,0.0008363276,-0.2057575434,-0.2248116881,0.0729394481,0.2377972454,0.4230693877,-0.300316304,0.0473215431,-0.4231781065,-0.1595294625,0.1389964968,-0.0802792013,-0.0387310572,-0.1483372748,-0.3189133108,0.2778151631,0.3382665813,0.1288223416,0.1079095602,-0.239862293,-0.0247332957,-0.2022278309,0.1674776077,-0.1916826367,-0.1363240331,-0.1142772362,0.2343533635,0.1731068641,0.0974104404,-0.3408343494,-0.0454138368,-0.2319974899,-0.0555951707,0.2320308834,-0.0026652913,0.30958727,-0.0770000219,0.14246355,-0.3586760759,0.0364723913,0.3661396801,0.0212173313,-0.1678794324,0.2677277923,-0.0123514775,-0.1251045763,0.0236814786,-0.14731206,-0.299810648,-0.2093851715,0.0171216857,0.0482656993,0.0239049792,-0.0471153446,0.0395985246,0.0380500816,0.1774878949,-0.1749292463,-0.340343833,-0.4319575131,0.4691960514,-0.2710248232,-0.2475002557,0.0816968828,0.0473560803,0.2014110088,0.3418824375,-0.4603409767,-0.0737863109,-0.1173928976,0.1646256745,0.0758609176,0.1864943802,0.0469675176,-0.1311487854,0.1585203111,-0.0770411864,-0.3395670652,-0.1058110222,0.2425631136,0.2578392923,-0.1052164957,0.4483567774,-0.2508384287,0.2393931001,-0.0129460888,0.2891106904,0.5426401496,-0.0893333852,0.1529127359,-0.3217939138,-0.3143526018,-0.099711448,-0.1948753744,-0.1389905959,0.0483570509,0.0269657038,-0.1491460651,-0.2706747651,-0.0529081263,-0.0879614204,-0.1746995449,-0.0553434081,-0.0693777204,0.3293852508,-0.0504932068,0.2791500986,0.0136173237,-0.0404188186,-0.1539395452,0.2246424258,0.1235996708,0.2225517929,-0.3244818151,0.1071843952,-0.2705765367,0.3529714644,-0.0407876782,0.7153809667,-0.1447271109,0.3261329234,0.1128169373,0.0381268747,0.1055565476,-0.2915969193,0.1107803509,0.0259097721,-0.4394381642,-0.0571818091,-0.0327192396,-0.0487926118,-0.0640700683,-0.029191507,0.2615399957,-0.2473840266,-0.1046056673,-0.0212867279,0.3048670292,-0.0421134904,-0.2426180691,-0.3949687183,-0.5412188768,-0.4392507076,-0.1890492439,0.1465301365,0.1577130854,-0.0798719376,0.2045397013,-0.1588291377,0.1053305566,0.2242294401,0.084091939,0.2725592852,-0.0330140777,0.1534871459,0.2682338357,-0.1665609181,0.7140191793,0.7792875171,-0.1654731035,-0.3519219756,0.253123343,-0.0765697509,0.0217691138,-0.0563563295,-0.0262792483,0.0473596342,-0.199334532,-0.1701831073,-0.0464213379,0.198322013,0.1967192143,0.1544240564,-0.292883724,-0.7532656789,0.5231232643,0.0733509213,0.1194912866,0.2948041558,-0.0994655937,-0.1877111197,0.2691773474,0.1097404063,0.7309182882,-0.4024430215,0.0663996488,0.1255254596,0.0390540808,0.0969472975,-0.0389399156,-0.2549379468,-0.1899633855,-0.2134175152,-0.1464372128,-0.1630649865,0.2815146148,0.1523412913,-0.0743444934,0.2819465995,-0.1734498143,-0.0285564661,0.2377475947,0.1179537997,-0.0852734521,-0.3346666694,-0.5701636672,0.0695263073,-0.1744922549,0.2196643502,-0.0229284316,0.0446263291,-0.0770891085,-0.0558210611,-0.3917295933,0.3648408055,-0.2748184204,0.3384894431,-0.1301485598,-0.3494686484,0.189233914,0.4507753253,0.2284569144,0.1401094347,-0.1527668387,0.0742595568,0.2642703056,-0.2523836493,-0.0304523762,-0.1002178192,-0.0693553537,-0.0230432879,-0.2731992006,0.1419645548,-0.0702483132,-0.2247540653,-0.0742935315,0.1391913295,0.0287422985,-0.0994305909,-0.3413546085,-0.071890451,0.0351187959,-0.1643306762,-0.0067225127,0.209125787,-0.0939807147,0.0110958582,0.0571492016,-0.3362408876,0.0208268687,0.4598697722,-0.0671367273,-0.1035425216,0.5126202703,0.2207635939,-0.2053335756,-0.1291891932,0.2668746114,0.343277216,-0.2435242385,0.0247940086,-0.1015284583,0.1226382703,0.0169650894,0.0190838203,0.2321757823,-0.3506089747,0.3566080034,-0.59125036,-0.3739056289,0.1904153675,-0.1842523664,0.0209109206,-0.0230076816,0.0114088757,-0.1538707614,-0.0855408907,-0.1131536886,0.1104641557,-0.3555292189,0.0531755239,0.2333114892,-0.0038374737,0.026617974,-0.1702212244,0.0084072324,0.1382467002,0.0358453728,-0.0385317653,-0.1823156774,0.2142098844,0.1291017532,-0.346626848,0.0095422855,-0.2173159271,-0.0406782925,-0.1456348747,0.175234288,0.4642310739,-0.0508853458,0.1162913889,0.2020449042,0.1192648113,-0.3610048294,0.377338171,-0.0597466715,0.2192406356,0.1052805781,0.1996854097,0.2261882126,0.1284537315,-0.298638612,-0.0918607861,0.254362464,-0.0207492411,0.2784873545,-0.424654901,0.3846978843,0.093402341,0.1392694265,0.4423628151,-0.2708707154,0.0668304861,0.3226239681,0.0277718697,-0.239678219,-0.1068023145,0.1143150181,0.0567229837,-0.1326247156,0.3174369335,0.3981636763,0.0658837408,-0.1405836493,0.0366226286,0.0629203171,-0.1436378211,0.0519923568,0.6180887222,0.018257726,0.2121533453,0.0627515018,0.2251056135,-0.0261530075,0.3331249356,-0.1577174217,0.4461428821,0.2134561688,0.1908597201,-0.3277816176,-0.5346593857,-0.1699689627,0.3204804361,0.0944019482,-0.0257126112,-0.208639726,0.2422643751,-0.1822718382,0.1587971896,-0.0697560012,0.1305374503,-0.034036275,-0.1838852763,-0.1329226047,-0.1520823389,-0.5502743721,0.1870924532,-0.2470207661,-0.1265384704,0.1939038336,0.1568964869,-0.1329802722,-0.3558672369,-0.0509607121,0.3773249984,-0.3437860012,-0.3400719464,0.3964489698,0.0177266039,0.0172705334,0.2495322376,0.1393802017,0.3306991458,0.420499593,-0.3608880341,-0.1154865772,0.0613388866,0.0774012357,0.0765652359,0.1979313791,0.0195528045,0.2065385729,0.3709867597,-0.0114097325,-0.0139406091,-0.0316576846,0.1572707593,-0.2297257632,-0.0879770666,0.2299744934,-0.0481610782,0.199876681,-0.117637977,0.0245101266,-0.2423740178,0.021785982,0.5751782656,0.0103393085,0.1068291664,0.0592920072,-0.0527461395,-0.1844878495,0.3957369924,0.6384893656,0.3486227691,-0.4021956027,-0.1908638626,-0.7238464952,0.0598110072,-0.4783252478,-0.2568792999,0.5058496594,-0.0741570145,0.101831302,0.3912000656,-0.0939410925,0.2525469065,0.1857650876,0.2951029241,-0.2620002925,-0.2589090765,0.4163457751,0.2090041637,-0.1786013097,-0.4936587512,-0.0105341868,0.042624142,-0.0922619253,-0.1824915558,0.0417558402,0.1412569433,0.3790921271,0.1933657229,0.0584114566,0.7760936022,-0.0257243253,-0.064801462,-0.0789829865,-0.6323983669,-0.054236386,0.0605209023,0.0649564788,0.1296199411,-0.1022136584,-0.1126848385,-0.3824729323,0.3061400652,-0.1255756468,-0.1483518481,-0.2318839282,-0.0043637305,-0.0508504957,-0.0628491864,-0.0281334668,0.0610886924,0.1224483997,0.3552176952,-0.125965029,-0.2262785584,0.3989257812,-0.0534627326,-0.4505837858,0.0725823417,-0.1192312092,0.3350661695,-0.2118573189,-0.461424768,0.1921404898,0.2282822132,-0.1450064182,-0.3645923734,0.0354042202,0.1440638602,0.1055092961,-0.1191350445,0.6080466509,0.0068141595,-0.158861652,0.1072711051,-0.1721357256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2061","title":"Cannot load udpos subsets from xtreme dataset using load_dataset()","comments":"I tried fixing this issue, but its working fine in the dev version : \"1.6.2.dev0\"\r\n\r\nI think somebody already fixed it. ","body":"Hello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n","comment_length":21,"text":"Cannot load udpos subsets from xtreme dataset using load_dataset()\nHello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n\nI tried fixing this issue, but its working fine in the dev version : \"1.6.2.dev0\"\r\n\r\nI think somebody already fixed it. ","embeddings":[-0.3464027941,-0.0752857924,0.002613897,0.4443909824,0.3713300824,0.069119066,0.2405683845,0.0394754633,0.5101771355,0.2061247975,-0.3589054644,0.0527050309,0.0860960111,0.0808662176,0.1144125015,-0.3151043355,-0.0602493994,0.0173558537,-0.2503674328,-0.065504171,-0.4105169475,-0.0623261929,-0.3046581447,0.1095427871,-0.1711033434,0.2064580172,0.0141455419,0.1647338569,-0.3150216639,-0.4950218201,0.5633596182,0.0180328377,0.5800744891,0.221322,-0.0001253432,0.1060385182,0.4861793816,-0.0613055825,-0.3452345729,-0.4049583972,-0.4189888239,-0.1315970272,-0.0690613464,-0.0964650735,0.1045276895,-0.0043987469,-0.0823340118,-0.2585622072,0.3321028948,0.3678756952,0.0941703171,0.3720282614,0.061857868,-0.2211653441,0.4757196605,-0.0422705263,-0.1489076465,0.1563735753,0.4271000326,-0.3246812224,0.1801449507,0.1246253327,-0.209139958,0.2249173075,-0.1637178361,0.0560058244,0.1821350753,-0.2583509088,0.2391248494,0.1349655539,0.8412337303,-0.1512385607,-0.4147819579,0.0182388742,-0.1014063731,-0.3554817438,0.1830140948,0.0008575391,-0.0578870513,0.0788415372,-0.0767996833,-0.0602417439,0.0503758267,0.289073348,-0.1434578896,0.2910360396,0.0541204214,0.1338610649,0.3239519,-0.1089203432,0.4901404083,0.0979419872,-0.0730714723,0.2410603017,-0.3231164515,0.2865316868,-0.0756045282,-0.3037876189,0.2389643937,-0.1045979336,-0.1012069955,0.0039554867,-0.1443649828,0.3295553327,0.4467023611,0.258592695,0.3614312708,0.2899973094,0.1072534993,0.1587685496,-0.0565737747,-0.0633865967,-0.4474610388,-0.2160627842,0.0857923925,-0.0271296799,0.1571703404,-0.4506189227,-0.5827982426,0.0448348261,-0.1629013866,-0.217883721,0.05834952,0.2688898146,0.1545642018,0.4297795296,0.0458623022,0.4458411336,-0.4962795973,-0.2659801245,-0.0544239134,0.1687626541,-0.4189635813,0.0585133731,0.1439427137,0.0052871024,0.2148580551,0.0167755205,0.2656741142,-0.2452537715,0.1288290769,-0.2358862758,-0.244421348,0.3540515602,0.3105243742,0.1218978986,0.2711307406,-0.3746761978,-0.0785986483,0.2891557217,-0.2225109637,-0.2350116372,0.0909441113,0.0328545943,-0.1796912402,0.0550205037,-1.0357831717,-0.0764572024,0.0851031467,0.1069326326,-0.0155084543,-0.275529623,-0.1712114811,0.0025955918,0.2848057449,0.5355801582,-0.2868715227,-0.0464064963,-0.1408394873,-0.0400896706,0.3355934024,0.3613086641,-0.3357942998,0.0737545416,-0.1279850602,0.1057119593,0.1982139647,-0.5150758028,-0.4803656042,0.1640977561,0.0509281084,-0.0161880385,-0.0857606456,-0.3274516165,0.2393174767,0.0357203931,0.1953124404,0.2589243054,0.0580973849,-0.0908872336,-0.1751016676,-0.0985951722,0.136113286,0.4271915853,0.2911248505,0.1453995854,0.2614878714,0.4392369092,0.0436218232,0.0165753514,0.0131826354,0.3216635585,0.0333599299,0.0902014673,0.0362407826,-0.0593443327,-0.4222900867,0.2021975517,-0.0930997282,-0.0453706384,0.0975539014,0.1812404096,-0.3897111714,0.1230222583,-0.3247447312,-0.3197046518,-0.0942896307,0.0866035298,-0.0964135006,-0.0818081424,-0.1698002815,0.0806680992,-0.0715484098,0.1103202999,-0.3754514456,0.4191193879,0.049578242,-0.0648903623,0.0807236657,0.2063002884,0.2299803942,-0.1260796189,-0.1753814518,0.1877421886,0.6032566428,-0.1791830808,0.1357364804,-0.0441033766,0.3042578995,-0.3885697424,0.1962021738,-0.0273815524,0.1187217534,0.1130647287,0.0598304868,0.254188925,-0.0359747075,0.1840643734,-0.1281909943,0.1224722341,0.3874266744,0.0022266286,0.0613593124,-0.2924954891,0.5102493167,0.3250283897,0.3184416592,0.0524797477,-0.1459858865,-0.2242728472,-0.1109304726,-0.1516300291,0.0816720501,0.1515698433,-0.2236988097,0.0444757342,0.2037227601,0.2373357564,0.32943362,0.0214671995,0.0758472681,-0.0468033701,0.0237630997,0.0701770708,-0.0097976169,-0.0718164816,-0.1015756652,0.0644405559,0.0008363276,-0.2057575434,-0.2248116881,0.0729394481,0.2377972454,0.4230693877,-0.300316304,0.0473215431,-0.4231781065,-0.1595294625,0.1389964968,-0.0802792013,-0.0387310572,-0.1483372748,-0.3189133108,0.2778151631,0.3382665813,0.1288223416,0.1079095602,-0.239862293,-0.0247332957,-0.2022278309,0.1674776077,-0.1916826367,-0.1363240331,-0.1142772362,0.2343533635,0.1731068641,0.0974104404,-0.3408343494,-0.0454138368,-0.2319974899,-0.0555951707,0.2320308834,-0.0026652913,0.30958727,-0.0770000219,0.14246355,-0.3586760759,0.0364723913,0.3661396801,0.0212173313,-0.1678794324,0.2677277923,-0.0123514775,-0.1251045763,0.0236814786,-0.14731206,-0.299810648,-0.2093851715,0.0171216857,0.0482656993,0.0239049792,-0.0471153446,0.0395985246,0.0380500816,0.1774878949,-0.1749292463,-0.340343833,-0.4319575131,0.4691960514,-0.2710248232,-0.2475002557,0.0816968828,0.0473560803,0.2014110088,0.3418824375,-0.4603409767,-0.0737863109,-0.1173928976,0.1646256745,0.0758609176,0.1864943802,0.0469675176,-0.1311487854,0.1585203111,-0.0770411864,-0.3395670652,-0.1058110222,0.2425631136,0.2578392923,-0.1052164957,0.4483567774,-0.2508384287,0.2393931001,-0.0129460888,0.2891106904,0.5426401496,-0.0893333852,0.1529127359,-0.3217939138,-0.3143526018,-0.099711448,-0.1948753744,-0.1389905959,0.0483570509,0.0269657038,-0.1491460651,-0.2706747651,-0.0529081263,-0.0879614204,-0.1746995449,-0.0553434081,-0.0693777204,0.3293852508,-0.0504932068,0.2791500986,0.0136173237,-0.0404188186,-0.1539395452,0.2246424258,0.1235996708,0.2225517929,-0.3244818151,0.1071843952,-0.2705765367,0.3529714644,-0.0407876782,0.7153809667,-0.1447271109,0.3261329234,0.1128169373,0.0381268747,0.1055565476,-0.2915969193,0.1107803509,0.0259097721,-0.4394381642,-0.0571818091,-0.0327192396,-0.0487926118,-0.0640700683,-0.029191507,0.2615399957,-0.2473840266,-0.1046056673,-0.0212867279,0.3048670292,-0.0421134904,-0.2426180691,-0.3949687183,-0.5412188768,-0.4392507076,-0.1890492439,0.1465301365,0.1577130854,-0.0798719376,0.2045397013,-0.1588291377,0.1053305566,0.2242294401,0.084091939,0.2725592852,-0.0330140777,0.1534871459,0.2682338357,-0.1665609181,0.7140191793,0.7792875171,-0.1654731035,-0.3519219756,0.253123343,-0.0765697509,0.0217691138,-0.0563563295,-0.0262792483,0.0473596342,-0.199334532,-0.1701831073,-0.0464213379,0.198322013,0.1967192143,0.1544240564,-0.292883724,-0.7532656789,0.5231232643,0.0733509213,0.1194912866,0.2948041558,-0.0994655937,-0.1877111197,0.2691773474,0.1097404063,0.7309182882,-0.4024430215,0.0663996488,0.1255254596,0.0390540808,0.0969472975,-0.0389399156,-0.2549379468,-0.1899633855,-0.2134175152,-0.1464372128,-0.1630649865,0.2815146148,0.1523412913,-0.0743444934,0.2819465995,-0.1734498143,-0.0285564661,0.2377475947,0.1179537997,-0.0852734521,-0.3346666694,-0.5701636672,0.0695263073,-0.1744922549,0.2196643502,-0.0229284316,0.0446263291,-0.0770891085,-0.0558210611,-0.3917295933,0.3648408055,-0.2748184204,0.3384894431,-0.1301485598,-0.3494686484,0.189233914,0.4507753253,0.2284569144,0.1401094347,-0.1527668387,0.0742595568,0.2642703056,-0.2523836493,-0.0304523762,-0.1002178192,-0.0693553537,-0.0230432879,-0.2731992006,0.1419645548,-0.0702483132,-0.2247540653,-0.0742935315,0.1391913295,0.0287422985,-0.0994305909,-0.3413546085,-0.071890451,0.0351187959,-0.1643306762,-0.0067225127,0.209125787,-0.0939807147,0.0110958582,0.0571492016,-0.3362408876,0.0208268687,0.4598697722,-0.0671367273,-0.1035425216,0.5126202703,0.2207635939,-0.2053335756,-0.1291891932,0.2668746114,0.343277216,-0.2435242385,0.0247940086,-0.1015284583,0.1226382703,0.0169650894,0.0190838203,0.2321757823,-0.3506089747,0.3566080034,-0.59125036,-0.3739056289,0.1904153675,-0.1842523664,0.0209109206,-0.0230076816,0.0114088757,-0.1538707614,-0.0855408907,-0.1131536886,0.1104641557,-0.3555292189,0.0531755239,0.2333114892,-0.0038374737,0.026617974,-0.1702212244,0.0084072324,0.1382467002,0.0358453728,-0.0385317653,-0.1823156774,0.2142098844,0.1291017532,-0.346626848,0.0095422855,-0.2173159271,-0.0406782925,-0.1456348747,0.175234288,0.4642310739,-0.0508853458,0.1162913889,0.2020449042,0.1192648113,-0.3610048294,0.377338171,-0.0597466715,0.2192406356,0.1052805781,0.1996854097,0.2261882126,0.1284537315,-0.298638612,-0.0918607861,0.254362464,-0.0207492411,0.2784873545,-0.424654901,0.3846978843,0.093402341,0.1392694265,0.4423628151,-0.2708707154,0.0668304861,0.3226239681,0.0277718697,-0.239678219,-0.1068023145,0.1143150181,0.0567229837,-0.1326247156,0.3174369335,0.3981636763,0.0658837408,-0.1405836493,0.0366226286,0.0629203171,-0.1436378211,0.0519923568,0.6180887222,0.018257726,0.2121533453,0.0627515018,0.2251056135,-0.0261530075,0.3331249356,-0.1577174217,0.4461428821,0.2134561688,0.1908597201,-0.3277816176,-0.5346593857,-0.1699689627,0.3204804361,0.0944019482,-0.0257126112,-0.208639726,0.2422643751,-0.1822718382,0.1587971896,-0.0697560012,0.1305374503,-0.034036275,-0.1838852763,-0.1329226047,-0.1520823389,-0.5502743721,0.1870924532,-0.2470207661,-0.1265384704,0.1939038336,0.1568964869,-0.1329802722,-0.3558672369,-0.0509607121,0.3773249984,-0.3437860012,-0.3400719464,0.3964489698,0.0177266039,0.0172705334,0.2495322376,0.1393802017,0.3306991458,0.420499593,-0.3608880341,-0.1154865772,0.0613388866,0.0774012357,0.0765652359,0.1979313791,0.0195528045,0.2065385729,0.3709867597,-0.0114097325,-0.0139406091,-0.0316576846,0.1572707593,-0.2297257632,-0.0879770666,0.2299744934,-0.0481610782,0.199876681,-0.117637977,0.0245101266,-0.2423740178,0.021785982,0.5751782656,0.0103393085,0.1068291664,0.0592920072,-0.0527461395,-0.1844878495,0.3957369924,0.6384893656,0.3486227691,-0.4021956027,-0.1908638626,-0.7238464952,0.0598110072,-0.4783252478,-0.2568792999,0.5058496594,-0.0741570145,0.101831302,0.3912000656,-0.0939410925,0.2525469065,0.1857650876,0.2951029241,-0.2620002925,-0.2589090765,0.4163457751,0.2090041637,-0.1786013097,-0.4936587512,-0.0105341868,0.042624142,-0.0922619253,-0.1824915558,0.0417558402,0.1412569433,0.3790921271,0.1933657229,0.0584114566,0.7760936022,-0.0257243253,-0.064801462,-0.0789829865,-0.6323983669,-0.054236386,0.0605209023,0.0649564788,0.1296199411,-0.1022136584,-0.1126848385,-0.3824729323,0.3061400652,-0.1255756468,-0.1483518481,-0.2318839282,-0.0043637305,-0.0508504957,-0.0628491864,-0.0281334668,0.0610886924,0.1224483997,0.3552176952,-0.125965029,-0.2262785584,0.3989257812,-0.0534627326,-0.4505837858,0.0725823417,-0.1192312092,0.3350661695,-0.2118573189,-0.461424768,0.1921404898,0.2282822132,-0.1450064182,-0.3645923734,0.0354042202,0.1440638602,0.1055092961,-0.1191350445,0.6080466509,0.0068141595,-0.158861652,0.1072711051,-0.1721357256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2061","title":"Cannot load udpos subsets from xtreme dataset using load_dataset()","comments":"Hi,\r\n\r\nafter #2326, the lines with pos tags equal to `\"_\"` are filtered out when generating the dataset, so this fixes the KeyError described above. However, the udpos subset should be in the conll format i.e. it should yield sequences of tokens and not single tokens, so it would be great to see this fixed (feel free to borrow the logic from [here](https:\/\/github.com\/google-research\/xtreme\/blob\/58a76a0d02458c4b3b6a742d3fd4ffaca80ff0de\/utils_preprocess.py#L187-L204) if you decide to work on this). ","body":"Hello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n","comment_length":70,"text":"Cannot load udpos subsets from xtreme dataset using load_dataset()\nHello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n\nHi,\r\n\r\nafter #2326, the lines with pos tags equal to `\"_\"` are filtered out when generating the dataset, so this fixes the KeyError described above. However, the udpos subset should be in the conll format i.e. it should yield sequences of tokens and not single tokens, so it would be great to see this fixed (feel free to borrow the logic from [here](https:\/\/github.com\/google-research\/xtreme\/blob\/58a76a0d02458c4b3b6a742d3fd4ffaca80ff0de\/utils_preprocess.py#L187-L204) if you decide to work on this). ","embeddings":[-0.3464027941,-0.0752857924,0.002613897,0.4443909824,0.3713300824,0.069119066,0.2405683845,0.0394754633,0.5101771355,0.2061247975,-0.3589054644,0.0527050309,0.0860960111,0.0808662176,0.1144125015,-0.3151043355,-0.0602493994,0.0173558537,-0.2503674328,-0.065504171,-0.4105169475,-0.0623261929,-0.3046581447,0.1095427871,-0.1711033434,0.2064580172,0.0141455419,0.1647338569,-0.3150216639,-0.4950218201,0.5633596182,0.0180328377,0.5800744891,0.221322,-0.0001253432,0.1060385182,0.4861793816,-0.0613055825,-0.3452345729,-0.4049583972,-0.4189888239,-0.1315970272,-0.0690613464,-0.0964650735,0.1045276895,-0.0043987469,-0.0823340118,-0.2585622072,0.3321028948,0.3678756952,0.0941703171,0.3720282614,0.061857868,-0.2211653441,0.4757196605,-0.0422705263,-0.1489076465,0.1563735753,0.4271000326,-0.3246812224,0.1801449507,0.1246253327,-0.209139958,0.2249173075,-0.1637178361,0.0560058244,0.1821350753,-0.2583509088,0.2391248494,0.1349655539,0.8412337303,-0.1512385607,-0.4147819579,0.0182388742,-0.1014063731,-0.3554817438,0.1830140948,0.0008575391,-0.0578870513,0.0788415372,-0.0767996833,-0.0602417439,0.0503758267,0.289073348,-0.1434578896,0.2910360396,0.0541204214,0.1338610649,0.3239519,-0.1089203432,0.4901404083,0.0979419872,-0.0730714723,0.2410603017,-0.3231164515,0.2865316868,-0.0756045282,-0.3037876189,0.2389643937,-0.1045979336,-0.1012069955,0.0039554867,-0.1443649828,0.3295553327,0.4467023611,0.258592695,0.3614312708,0.2899973094,0.1072534993,0.1587685496,-0.0565737747,-0.0633865967,-0.4474610388,-0.2160627842,0.0857923925,-0.0271296799,0.1571703404,-0.4506189227,-0.5827982426,0.0448348261,-0.1629013866,-0.217883721,0.05834952,0.2688898146,0.1545642018,0.4297795296,0.0458623022,0.4458411336,-0.4962795973,-0.2659801245,-0.0544239134,0.1687626541,-0.4189635813,0.0585133731,0.1439427137,0.0052871024,0.2148580551,0.0167755205,0.2656741142,-0.2452537715,0.1288290769,-0.2358862758,-0.244421348,0.3540515602,0.3105243742,0.1218978986,0.2711307406,-0.3746761978,-0.0785986483,0.2891557217,-0.2225109637,-0.2350116372,0.0909441113,0.0328545943,-0.1796912402,0.0550205037,-1.0357831717,-0.0764572024,0.0851031467,0.1069326326,-0.0155084543,-0.275529623,-0.1712114811,0.0025955918,0.2848057449,0.5355801582,-0.2868715227,-0.0464064963,-0.1408394873,-0.0400896706,0.3355934024,0.3613086641,-0.3357942998,0.0737545416,-0.1279850602,0.1057119593,0.1982139647,-0.5150758028,-0.4803656042,0.1640977561,0.0509281084,-0.0161880385,-0.0857606456,-0.3274516165,0.2393174767,0.0357203931,0.1953124404,0.2589243054,0.0580973849,-0.0908872336,-0.1751016676,-0.0985951722,0.136113286,0.4271915853,0.2911248505,0.1453995854,0.2614878714,0.4392369092,0.0436218232,0.0165753514,0.0131826354,0.3216635585,0.0333599299,0.0902014673,0.0362407826,-0.0593443327,-0.4222900867,0.2021975517,-0.0930997282,-0.0453706384,0.0975539014,0.1812404096,-0.3897111714,0.1230222583,-0.3247447312,-0.3197046518,-0.0942896307,0.0866035298,-0.0964135006,-0.0818081424,-0.1698002815,0.0806680992,-0.0715484098,0.1103202999,-0.3754514456,0.4191193879,0.049578242,-0.0648903623,0.0807236657,0.2063002884,0.2299803942,-0.1260796189,-0.1753814518,0.1877421886,0.6032566428,-0.1791830808,0.1357364804,-0.0441033766,0.3042578995,-0.3885697424,0.1962021738,-0.0273815524,0.1187217534,0.1130647287,0.0598304868,0.254188925,-0.0359747075,0.1840643734,-0.1281909943,0.1224722341,0.3874266744,0.0022266286,0.0613593124,-0.2924954891,0.5102493167,0.3250283897,0.3184416592,0.0524797477,-0.1459858865,-0.2242728472,-0.1109304726,-0.1516300291,0.0816720501,0.1515698433,-0.2236988097,0.0444757342,0.2037227601,0.2373357564,0.32943362,0.0214671995,0.0758472681,-0.0468033701,0.0237630997,0.0701770708,-0.0097976169,-0.0718164816,-0.1015756652,0.0644405559,0.0008363276,-0.2057575434,-0.2248116881,0.0729394481,0.2377972454,0.4230693877,-0.300316304,0.0473215431,-0.4231781065,-0.1595294625,0.1389964968,-0.0802792013,-0.0387310572,-0.1483372748,-0.3189133108,0.2778151631,0.3382665813,0.1288223416,0.1079095602,-0.239862293,-0.0247332957,-0.2022278309,0.1674776077,-0.1916826367,-0.1363240331,-0.1142772362,0.2343533635,0.1731068641,0.0974104404,-0.3408343494,-0.0454138368,-0.2319974899,-0.0555951707,0.2320308834,-0.0026652913,0.30958727,-0.0770000219,0.14246355,-0.3586760759,0.0364723913,0.3661396801,0.0212173313,-0.1678794324,0.2677277923,-0.0123514775,-0.1251045763,0.0236814786,-0.14731206,-0.299810648,-0.2093851715,0.0171216857,0.0482656993,0.0239049792,-0.0471153446,0.0395985246,0.0380500816,0.1774878949,-0.1749292463,-0.340343833,-0.4319575131,0.4691960514,-0.2710248232,-0.2475002557,0.0816968828,0.0473560803,0.2014110088,0.3418824375,-0.4603409767,-0.0737863109,-0.1173928976,0.1646256745,0.0758609176,0.1864943802,0.0469675176,-0.1311487854,0.1585203111,-0.0770411864,-0.3395670652,-0.1058110222,0.2425631136,0.2578392923,-0.1052164957,0.4483567774,-0.2508384287,0.2393931001,-0.0129460888,0.2891106904,0.5426401496,-0.0893333852,0.1529127359,-0.3217939138,-0.3143526018,-0.099711448,-0.1948753744,-0.1389905959,0.0483570509,0.0269657038,-0.1491460651,-0.2706747651,-0.0529081263,-0.0879614204,-0.1746995449,-0.0553434081,-0.0693777204,0.3293852508,-0.0504932068,0.2791500986,0.0136173237,-0.0404188186,-0.1539395452,0.2246424258,0.1235996708,0.2225517929,-0.3244818151,0.1071843952,-0.2705765367,0.3529714644,-0.0407876782,0.7153809667,-0.1447271109,0.3261329234,0.1128169373,0.0381268747,0.1055565476,-0.2915969193,0.1107803509,0.0259097721,-0.4394381642,-0.0571818091,-0.0327192396,-0.0487926118,-0.0640700683,-0.029191507,0.2615399957,-0.2473840266,-0.1046056673,-0.0212867279,0.3048670292,-0.0421134904,-0.2426180691,-0.3949687183,-0.5412188768,-0.4392507076,-0.1890492439,0.1465301365,0.1577130854,-0.0798719376,0.2045397013,-0.1588291377,0.1053305566,0.2242294401,0.084091939,0.2725592852,-0.0330140777,0.1534871459,0.2682338357,-0.1665609181,0.7140191793,0.7792875171,-0.1654731035,-0.3519219756,0.253123343,-0.0765697509,0.0217691138,-0.0563563295,-0.0262792483,0.0473596342,-0.199334532,-0.1701831073,-0.0464213379,0.198322013,0.1967192143,0.1544240564,-0.292883724,-0.7532656789,0.5231232643,0.0733509213,0.1194912866,0.2948041558,-0.0994655937,-0.1877111197,0.2691773474,0.1097404063,0.7309182882,-0.4024430215,0.0663996488,0.1255254596,0.0390540808,0.0969472975,-0.0389399156,-0.2549379468,-0.1899633855,-0.2134175152,-0.1464372128,-0.1630649865,0.2815146148,0.1523412913,-0.0743444934,0.2819465995,-0.1734498143,-0.0285564661,0.2377475947,0.1179537997,-0.0852734521,-0.3346666694,-0.5701636672,0.0695263073,-0.1744922549,0.2196643502,-0.0229284316,0.0446263291,-0.0770891085,-0.0558210611,-0.3917295933,0.3648408055,-0.2748184204,0.3384894431,-0.1301485598,-0.3494686484,0.189233914,0.4507753253,0.2284569144,0.1401094347,-0.1527668387,0.0742595568,0.2642703056,-0.2523836493,-0.0304523762,-0.1002178192,-0.0693553537,-0.0230432879,-0.2731992006,0.1419645548,-0.0702483132,-0.2247540653,-0.0742935315,0.1391913295,0.0287422985,-0.0994305909,-0.3413546085,-0.071890451,0.0351187959,-0.1643306762,-0.0067225127,0.209125787,-0.0939807147,0.0110958582,0.0571492016,-0.3362408876,0.0208268687,0.4598697722,-0.0671367273,-0.1035425216,0.5126202703,0.2207635939,-0.2053335756,-0.1291891932,0.2668746114,0.343277216,-0.2435242385,0.0247940086,-0.1015284583,0.1226382703,0.0169650894,0.0190838203,0.2321757823,-0.3506089747,0.3566080034,-0.59125036,-0.3739056289,0.1904153675,-0.1842523664,0.0209109206,-0.0230076816,0.0114088757,-0.1538707614,-0.0855408907,-0.1131536886,0.1104641557,-0.3555292189,0.0531755239,0.2333114892,-0.0038374737,0.026617974,-0.1702212244,0.0084072324,0.1382467002,0.0358453728,-0.0385317653,-0.1823156774,0.2142098844,0.1291017532,-0.346626848,0.0095422855,-0.2173159271,-0.0406782925,-0.1456348747,0.175234288,0.4642310739,-0.0508853458,0.1162913889,0.2020449042,0.1192648113,-0.3610048294,0.377338171,-0.0597466715,0.2192406356,0.1052805781,0.1996854097,0.2261882126,0.1284537315,-0.298638612,-0.0918607861,0.254362464,-0.0207492411,0.2784873545,-0.424654901,0.3846978843,0.093402341,0.1392694265,0.4423628151,-0.2708707154,0.0668304861,0.3226239681,0.0277718697,-0.239678219,-0.1068023145,0.1143150181,0.0567229837,-0.1326247156,0.3174369335,0.3981636763,0.0658837408,-0.1405836493,0.0366226286,0.0629203171,-0.1436378211,0.0519923568,0.6180887222,0.018257726,0.2121533453,0.0627515018,0.2251056135,-0.0261530075,0.3331249356,-0.1577174217,0.4461428821,0.2134561688,0.1908597201,-0.3277816176,-0.5346593857,-0.1699689627,0.3204804361,0.0944019482,-0.0257126112,-0.208639726,0.2422643751,-0.1822718382,0.1587971896,-0.0697560012,0.1305374503,-0.034036275,-0.1838852763,-0.1329226047,-0.1520823389,-0.5502743721,0.1870924532,-0.2470207661,-0.1265384704,0.1939038336,0.1568964869,-0.1329802722,-0.3558672369,-0.0509607121,0.3773249984,-0.3437860012,-0.3400719464,0.3964489698,0.0177266039,0.0172705334,0.2495322376,0.1393802017,0.3306991458,0.420499593,-0.3608880341,-0.1154865772,0.0613388866,0.0774012357,0.0765652359,0.1979313791,0.0195528045,0.2065385729,0.3709867597,-0.0114097325,-0.0139406091,-0.0316576846,0.1572707593,-0.2297257632,-0.0879770666,0.2299744934,-0.0481610782,0.199876681,-0.117637977,0.0245101266,-0.2423740178,0.021785982,0.5751782656,0.0103393085,0.1068291664,0.0592920072,-0.0527461395,-0.1844878495,0.3957369924,0.6384893656,0.3486227691,-0.4021956027,-0.1908638626,-0.7238464952,0.0598110072,-0.4783252478,-0.2568792999,0.5058496594,-0.0741570145,0.101831302,0.3912000656,-0.0939410925,0.2525469065,0.1857650876,0.2951029241,-0.2620002925,-0.2589090765,0.4163457751,0.2090041637,-0.1786013097,-0.4936587512,-0.0105341868,0.042624142,-0.0922619253,-0.1824915558,0.0417558402,0.1412569433,0.3790921271,0.1933657229,0.0584114566,0.7760936022,-0.0257243253,-0.064801462,-0.0789829865,-0.6323983669,-0.054236386,0.0605209023,0.0649564788,0.1296199411,-0.1022136584,-0.1126848385,-0.3824729323,0.3061400652,-0.1255756468,-0.1483518481,-0.2318839282,-0.0043637305,-0.0508504957,-0.0628491864,-0.0281334668,0.0610886924,0.1224483997,0.3552176952,-0.125965029,-0.2262785584,0.3989257812,-0.0534627326,-0.4505837858,0.0725823417,-0.1192312092,0.3350661695,-0.2118573189,-0.461424768,0.1921404898,0.2282822132,-0.1450064182,-0.3645923734,0.0354042202,0.1440638602,0.1055092961,-0.1191350445,0.6080466509,0.0068141595,-0.158861652,0.1072711051,-0.1721357256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2061","title":"Cannot load udpos subsets from xtreme dataset using load_dataset()","comments":"Closed by #2466.","body":"Hello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n","comment_length":3,"text":"Cannot load udpos subsets from xtreme dataset using load_dataset()\nHello, \r\n\r\nI am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error. \r\n\r\nReprex is: \r\n\r\n`from datasets import load_dataset `\r\n`dataset = load_dataset('xtreme', 'udpos.English')`\r\n\r\nThe error is: \r\n`KeyError: '_'`\r\n\r\nThe full traceback is: \r\nKeyError Traceback (most recent call last)\r\n in \r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('xtreme', 'udpos.English')\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 738 \r\n 739 # Download and prepare data\r\n--> 740 builder_instance.download_and_prepare(\r\n 741 download_config=download_config,\r\n 742 download_mode=download_mode,\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)\r\n 576 logger.warning(\"HF google storage unreachable. Downloading and preparing it from source\")\r\n 577 if not downloaded_from_gcs:\r\n--> 578 self._download_and_prepare(\r\n 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 580 )\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 654 try:\r\n 655 # Prepare split will record examples associated to the split\r\n--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 657 except OSError as e:\r\n 658 raise OSError(\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\builder.py in _prepare_split(self, split_generator)\r\n 977 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 978 ):\r\n--> 979 example = self.info.features.encode_example(record)\r\n 980 writer.write(example)\r\n 981 finally:\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example)\r\n 946 def encode_example(self, example):\r\n 947 example = cast_to_python_objects(example)\r\n--> 948 return encode_nested_example(self, example)\r\n 949 \r\n 950 def encode_batch(self, batch):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 840 # Nested structures: we allow dict, list\/tuples, sequences\r\n 841 if isinstance(schema, dict):\r\n--> 842 return {\r\n 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in (.0)\r\n 841 if isinstance(schema, dict):\r\n 842 return {\r\n--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n 844 }\r\n 845 elif isinstance(schema, (list, tuple)):\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_nested_example(schema, obj)\r\n 868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks\r\n 869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):\r\n--> 870 return schema.encode_example(obj)\r\n 871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)\r\n 872 return obj\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in encode_example(self, example_data)\r\n 647 # If a string is given, convert to associated integer\r\n 648 if isinstance(example_data, str):\r\n--> 649 example_data = self.str2int(example_data)\r\n 650 \r\n 651 # Allowing -1 to mean no label.\r\n\r\n~\\Anaconda3\\envs\\mlenv\\lib\\site-packages\\datasets\\features.py in str2int(self, values)\r\n 605 if value not in self._str2int:\r\n 606 value = value.strip()\r\n--> 607 output.append(self._str2int[str(value)])\r\n 608 else:\r\n 609 # No names provided, try to integerize\r\n\r\nKeyError: '_'\r\n\r\n\nClosed by #2466.","embeddings":[-0.3464027941,-0.0752857924,0.002613897,0.4443909824,0.3713300824,0.069119066,0.2405683845,0.0394754633,0.5101771355,0.2061247975,-0.3589054644,0.0527050309,0.0860960111,0.0808662176,0.1144125015,-0.3151043355,-0.0602493994,0.0173558537,-0.2503674328,-0.065504171,-0.4105169475,-0.0623261929,-0.3046581447,0.1095427871,-0.1711033434,0.2064580172,0.0141455419,0.1647338569,-0.3150216639,-0.4950218201,0.5633596182,0.0180328377,0.5800744891,0.221322,-0.0001253432,0.1060385182,0.4861793816,-0.0613055825,-0.3452345729,-0.4049583972,-0.4189888239,-0.1315970272,-0.0690613464,-0.0964650735,0.1045276895,-0.0043987469,-0.0823340118,-0.2585622072,0.3321028948,0.3678756952,0.0941703171,0.3720282614,0.061857868,-0.2211653441,0.4757196605,-0.0422705263,-0.1489076465,0.1563735753,0.4271000326,-0.3246812224,0.1801449507,0.1246253327,-0.209139958,0.2249173075,-0.1637178361,0.0560058244,0.1821350753,-0.2583509088,0.2391248494,0.1349655539,0.8412337303,-0.1512385607,-0.4147819579,0.0182388742,-0.1014063731,-0.3554817438,0.1830140948,0.0008575391,-0.0578870513,0.0788415372,-0.0767996833,-0.0602417439,0.0503758267,0.289073348,-0.1434578896,0.2910360396,0.0541204214,0.1338610649,0.3239519,-0.1089203432,0.4901404083,0.0979419872,-0.0730714723,0.2410603017,-0.3231164515,0.2865316868,-0.0756045282,-0.3037876189,0.2389643937,-0.1045979336,-0.1012069955,0.0039554867,-0.1443649828,0.3295553327,0.4467023611,0.258592695,0.3614312708,0.2899973094,0.1072534993,0.1587685496,-0.0565737747,-0.0633865967,-0.4474610388,-0.2160627842,0.0857923925,-0.0271296799,0.1571703404,-0.4506189227,-0.5827982426,0.0448348261,-0.1629013866,-0.217883721,0.05834952,0.2688898146,0.1545642018,0.4297795296,0.0458623022,0.4458411336,-0.4962795973,-0.2659801245,-0.0544239134,0.1687626541,-0.4189635813,0.0585133731,0.1439427137,0.0052871024,0.2148580551,0.0167755205,0.2656741142,-0.2452537715,0.1288290769,-0.2358862758,-0.244421348,0.3540515602,0.3105243742,0.1218978986,0.2711307406,-0.3746761978,-0.0785986483,0.2891557217,-0.2225109637,-0.2350116372,0.0909441113,0.0328545943,-0.1796912402,0.0550205037,-1.0357831717,-0.0764572024,0.0851031467,0.1069326326,-0.0155084543,-0.275529623,-0.1712114811,0.0025955918,0.2848057449,0.5355801582,-0.2868715227,-0.0464064963,-0.1408394873,-0.0400896706,0.3355934024,0.3613086641,-0.3357942998,0.0737545416,-0.1279850602,0.1057119593,0.1982139647,-0.5150758028,-0.4803656042,0.1640977561,0.0509281084,-0.0161880385,-0.0857606456,-0.3274516165,0.2393174767,0.0357203931,0.1953124404,0.2589243054,0.0580973849,-0.0908872336,-0.1751016676,-0.0985951722,0.136113286,0.4271915853,0.2911248505,0.1453995854,0.2614878714,0.4392369092,0.0436218232,0.0165753514,0.0131826354,0.3216635585,0.0333599299,0.0902014673,0.0362407826,-0.0593443327,-0.4222900867,0.2021975517,-0.0930997282,-0.0453706384,0.0975539014,0.1812404096,-0.3897111714,0.1230222583,-0.3247447312,-0.3197046518,-0.0942896307,0.0866035298,-0.0964135006,-0.0818081424,-0.1698002815,0.0806680992,-0.0715484098,0.1103202999,-0.3754514456,0.4191193879,0.049578242,-0.0648903623,0.0807236657,0.2063002884,0.2299803942,-0.1260796189,-0.1753814518,0.1877421886,0.6032566428,-0.1791830808,0.1357364804,-0.0441033766,0.3042578995,-0.3885697424,0.1962021738,-0.0273815524,0.1187217534,0.1130647287,0.0598304868,0.254188925,-0.0359747075,0.1840643734,-0.1281909943,0.1224722341,0.3874266744,0.0022266286,0.0613593124,-0.2924954891,0.5102493167,0.3250283897,0.3184416592,0.0524797477,-0.1459858865,-0.2242728472,-0.1109304726,-0.1516300291,0.0816720501,0.1515698433,-0.2236988097,0.0444757342,0.2037227601,0.2373357564,0.32943362,0.0214671995,0.0758472681,-0.0468033701,0.0237630997,0.0701770708,-0.0097976169,-0.0718164816,-0.1015756652,0.0644405559,0.0008363276,-0.2057575434,-0.2248116881,0.0729394481,0.2377972454,0.4230693877,-0.300316304,0.0473215431,-0.4231781065,-0.1595294625,0.1389964968,-0.0802792013,-0.0387310572,-0.1483372748,-0.3189133108,0.2778151631,0.3382665813,0.1288223416,0.1079095602,-0.239862293,-0.0247332957,-0.2022278309,0.1674776077,-0.1916826367,-0.1363240331,-0.1142772362,0.2343533635,0.1731068641,0.0974104404,-0.3408343494,-0.0454138368,-0.2319974899,-0.0555951707,0.2320308834,-0.0026652913,0.30958727,-0.0770000219,0.14246355,-0.3586760759,0.0364723913,0.3661396801,0.0212173313,-0.1678794324,0.2677277923,-0.0123514775,-0.1251045763,0.0236814786,-0.14731206,-0.299810648,-0.2093851715,0.0171216857,0.0482656993,0.0239049792,-0.0471153446,0.0395985246,0.0380500816,0.1774878949,-0.1749292463,-0.340343833,-0.4319575131,0.4691960514,-0.2710248232,-0.2475002557,0.0816968828,0.0473560803,0.2014110088,0.3418824375,-0.4603409767,-0.0737863109,-0.1173928976,0.1646256745,0.0758609176,0.1864943802,0.0469675176,-0.1311487854,0.1585203111,-0.0770411864,-0.3395670652,-0.1058110222,0.2425631136,0.2578392923,-0.1052164957,0.4483567774,-0.2508384287,0.2393931001,-0.0129460888,0.2891106904,0.5426401496,-0.0893333852,0.1529127359,-0.3217939138,-0.3143526018,-0.099711448,-0.1948753744,-0.1389905959,0.0483570509,0.0269657038,-0.1491460651,-0.2706747651,-0.0529081263,-0.0879614204,-0.1746995449,-0.0553434081,-0.0693777204,0.3293852508,-0.0504932068,0.2791500986,0.0136173237,-0.0404188186,-0.1539395452,0.2246424258,0.1235996708,0.2225517929,-0.3244818151,0.1071843952,-0.2705765367,0.3529714644,-0.0407876782,0.7153809667,-0.1447271109,0.3261329234,0.1128169373,0.0381268747,0.1055565476,-0.2915969193,0.1107803509,0.0259097721,-0.4394381642,-0.0571818091,-0.0327192396,-0.0487926118,-0.0640700683,-0.029191507,0.2615399957,-0.2473840266,-0.1046056673,-0.0212867279,0.3048670292,-0.0421134904,-0.2426180691,-0.3949687183,-0.5412188768,-0.4392507076,-0.1890492439,0.1465301365,0.1577130854,-0.0798719376,0.2045397013,-0.1588291377,0.1053305566,0.2242294401,0.084091939,0.2725592852,-0.0330140777,0.1534871459,0.2682338357,-0.1665609181,0.7140191793,0.7792875171,-0.1654731035,-0.3519219756,0.253123343,-0.0765697509,0.0217691138,-0.0563563295,-0.0262792483,0.0473596342,-0.199334532,-0.1701831073,-0.0464213379,0.198322013,0.1967192143,0.1544240564,-0.292883724,-0.7532656789,0.5231232643,0.0733509213,0.1194912866,0.2948041558,-0.0994655937,-0.1877111197,0.2691773474,0.1097404063,0.7309182882,-0.4024430215,0.0663996488,0.1255254596,0.0390540808,0.0969472975,-0.0389399156,-0.2549379468,-0.1899633855,-0.2134175152,-0.1464372128,-0.1630649865,0.2815146148,0.1523412913,-0.0743444934,0.2819465995,-0.1734498143,-0.0285564661,0.2377475947,0.1179537997,-0.0852734521,-0.3346666694,-0.5701636672,0.0695263073,-0.1744922549,0.2196643502,-0.0229284316,0.0446263291,-0.0770891085,-0.0558210611,-0.3917295933,0.3648408055,-0.2748184204,0.3384894431,-0.1301485598,-0.3494686484,0.189233914,0.4507753253,0.2284569144,0.1401094347,-0.1527668387,0.0742595568,0.2642703056,-0.2523836493,-0.0304523762,-0.1002178192,-0.0693553537,-0.0230432879,-0.2731992006,0.1419645548,-0.0702483132,-0.2247540653,-0.0742935315,0.1391913295,0.0287422985,-0.0994305909,-0.3413546085,-0.071890451,0.0351187959,-0.1643306762,-0.0067225127,0.209125787,-0.0939807147,0.0110958582,0.0571492016,-0.3362408876,0.0208268687,0.4598697722,-0.0671367273,-0.1035425216,0.5126202703,0.2207635939,-0.2053335756,-0.1291891932,0.2668746114,0.343277216,-0.2435242385,0.0247940086,-0.1015284583,0.1226382703,0.0169650894,0.0190838203,0.2321757823,-0.3506089747,0.3566080034,-0.59125036,-0.3739056289,0.1904153675,-0.1842523664,0.0209109206,-0.0230076816,0.0114088757,-0.1538707614,-0.0855408907,-0.1131536886,0.1104641557,-0.3555292189,0.0531755239,0.2333114892,-0.0038374737,0.026617974,-0.1702212244,0.0084072324,0.1382467002,0.0358453728,-0.0385317653,-0.1823156774,0.2142098844,0.1291017532,-0.346626848,0.0095422855,-0.2173159271,-0.0406782925,-0.1456348747,0.175234288,0.4642310739,-0.0508853458,0.1162913889,0.2020449042,0.1192648113,-0.3610048294,0.377338171,-0.0597466715,0.2192406356,0.1052805781,0.1996854097,0.2261882126,0.1284537315,-0.298638612,-0.0918607861,0.254362464,-0.0207492411,0.2784873545,-0.424654901,0.3846978843,0.093402341,0.1392694265,0.4423628151,-0.2708707154,0.0668304861,0.3226239681,0.0277718697,-0.239678219,-0.1068023145,0.1143150181,0.0567229837,-0.1326247156,0.3174369335,0.3981636763,0.0658837408,-0.1405836493,0.0366226286,0.0629203171,-0.1436378211,0.0519923568,0.6180887222,0.018257726,0.2121533453,0.0627515018,0.2251056135,-0.0261530075,0.3331249356,-0.1577174217,0.4461428821,0.2134561688,0.1908597201,-0.3277816176,-0.5346593857,-0.1699689627,0.3204804361,0.0944019482,-0.0257126112,-0.208639726,0.2422643751,-0.1822718382,0.1587971896,-0.0697560012,0.1305374503,-0.034036275,-0.1838852763,-0.1329226047,-0.1520823389,-0.5502743721,0.1870924532,-0.2470207661,-0.1265384704,0.1939038336,0.1568964869,-0.1329802722,-0.3558672369,-0.0509607121,0.3773249984,-0.3437860012,-0.3400719464,0.3964489698,0.0177266039,0.0172705334,0.2495322376,0.1393802017,0.3306991458,0.420499593,-0.3608880341,-0.1154865772,0.0613388866,0.0774012357,0.0765652359,0.1979313791,0.0195528045,0.2065385729,0.3709867597,-0.0114097325,-0.0139406091,-0.0316576846,0.1572707593,-0.2297257632,-0.0879770666,0.2299744934,-0.0481610782,0.199876681,-0.117637977,0.0245101266,-0.2423740178,0.021785982,0.5751782656,0.0103393085,0.1068291664,0.0592920072,-0.0527461395,-0.1844878495,0.3957369924,0.6384893656,0.3486227691,-0.4021956027,-0.1908638626,-0.7238464952,0.0598110072,-0.4783252478,-0.2568792999,0.5058496594,-0.0741570145,0.101831302,0.3912000656,-0.0939410925,0.2525469065,0.1857650876,0.2951029241,-0.2620002925,-0.2589090765,0.4163457751,0.2090041637,-0.1786013097,-0.4936587512,-0.0105341868,0.042624142,-0.0922619253,-0.1824915558,0.0417558402,0.1412569433,0.3790921271,0.1933657229,0.0584114566,0.7760936022,-0.0257243253,-0.064801462,-0.0789829865,-0.6323983669,-0.054236386,0.0605209023,0.0649564788,0.1296199411,-0.1022136584,-0.1126848385,-0.3824729323,0.3061400652,-0.1255756468,-0.1483518481,-0.2318839282,-0.0043637305,-0.0508504957,-0.0628491864,-0.0281334668,0.0610886924,0.1224483997,0.3552176952,-0.125965029,-0.2262785584,0.3989257812,-0.0534627326,-0.4505837858,0.0725823417,-0.1192312092,0.3350661695,-0.2118573189,-0.461424768,0.1921404898,0.2282822132,-0.1450064182,-0.3645923734,0.0354042202,0.1440638602,0.1055092961,-0.1191350445,0.6080466509,0.0068141595,-0.158861652,0.1072711051,-0.1721357256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2059","title":"Error while following docs to load the `ted_talks_iwslt` dataset","comments":"@skyprince999 as you authored the PR for this dataset, any comments?","body":"I am currently trying to load the `ted_talks_iwslt` dataset into google colab.\r\n\r\nThe [docs](https:\/\/huggingface.co\/datasets\/ted_talks_iwslt) mention the following way of doing so.\r\n\r\n```python\r\ndataset = load_dataset(\"ted_talks_iwslt\", language_pair=(\"it\", \"pl\"), year=\"2014\")\r\n```\r\n\r\nExecuting it results in the error attached below.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"ted_talks_iwslt\", language_pair=(\"it\", \"pl\"), year=\"2014\")\r\n\r\n4 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 730 hash=hash,\r\n 731 features=features,\r\n--> 732 **config_kwargs,\r\n 733 )\r\n 734 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in __init__(self, writer_batch_size, *args, **kwargs)\r\n 927 \r\n 928 def __init__(self, *args, writer_batch_size=None, **kwargs):\r\n--> 929 super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n 930 # Batch size used by the ArrowWriter\r\n 931 # It defines the number of samples that are kept in memory before writing them\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)\r\n 241 name,\r\n 242 custom_features=features,\r\n--> 243 **config_kwargs,\r\n 244 )\r\n 245 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)\r\n 337 if \"version\" not in config_kwargs and hasattr(self, \"VERSION\") and self.VERSION:\r\n 338 config_kwargs[\"version\"] = self.VERSION\r\n--> 339 builder_config = self.BUILDER_CONFIG_CLASS(**config_kwargs)\r\n 340 \r\n 341 # otherwise use the config_kwargs to overwrite the attributes\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/ted_talks_iwslt\/024d06b1376b361e59245c5878ab8acf9a7576d765f2d0077f61751158e60914\/ted_talks_iwslt.py in __init__(self, language_pair, year, **kwargs)\r\n 219 description=description,\r\n 220 version=datasets.Version(\"1.1.0\", \"\"),\r\n--> 221 **kwargs,\r\n 222 )\r\n 223 \r\n\r\nTypeError: __init__() got multiple values for keyword argument 'version'\r\n```\r\n\r\nHow to resolve this? \r\n\r\nPS: Thanks a lot @huggingface team for creating this great library!","comment_length":11,"text":"Error while following docs to load the `ted_talks_iwslt` dataset\nI am currently trying to load the `ted_talks_iwslt` dataset into google colab.\r\n\r\nThe [docs](https:\/\/huggingface.co\/datasets\/ted_talks_iwslt) mention the following way of doing so.\r\n\r\n```python\r\ndataset = load_dataset(\"ted_talks_iwslt\", language_pair=(\"it\", \"pl\"), year=\"2014\")\r\n```\r\n\r\nExecuting it results in the error attached below.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"ted_talks_iwslt\", language_pair=(\"it\", \"pl\"), year=\"2014\")\r\n\r\n4 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 730 hash=hash,\r\n 731 features=features,\r\n--> 732 **config_kwargs,\r\n 733 )\r\n 734 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in __init__(self, writer_batch_size, *args, **kwargs)\r\n 927 \r\n 928 def __init__(self, *args, writer_batch_size=None, **kwargs):\r\n--> 929 super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n 930 # Batch size used by the ArrowWriter\r\n 931 # It defines the number of samples that are kept in memory before writing them\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)\r\n 241 name,\r\n 242 custom_features=features,\r\n--> 243 **config_kwargs,\r\n 244 )\r\n 245 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)\r\n 337 if \"version\" not in config_kwargs and hasattr(self, \"VERSION\") and self.VERSION:\r\n 338 config_kwargs[\"version\"] = self.VERSION\r\n--> 339 builder_config = self.BUILDER_CONFIG_CLASS(**config_kwargs)\r\n 340 \r\n 341 # otherwise use the config_kwargs to overwrite the attributes\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/ted_talks_iwslt\/024d06b1376b361e59245c5878ab8acf9a7576d765f2d0077f61751158e60914\/ted_talks_iwslt.py in __init__(self, language_pair, year, **kwargs)\r\n 219 description=description,\r\n 220 version=datasets.Version(\"1.1.0\", \"\"),\r\n--> 221 **kwargs,\r\n 222 )\r\n 223 \r\n\r\nTypeError: __init__() got multiple values for keyword argument 'version'\r\n```\r\n\r\nHow to resolve this? \r\n\r\nPS: Thanks a lot @huggingface team for creating this great library!\n@skyprince999 as you authored the PR for this dataset, any comments?","embeddings":[-0.2140100002,0.1284443438,0.0485115461,0.2001313567,0.0312912539,0.0933724046,0.6712156534,0.1944082826,0.1241025403,-0.0619180724,-0.0495757982,0.3381051421,-0.291054666,0.4190407395,0.077388525,-0.2890882194,0.0997464731,0.0731823891,-0.0559581928,0.206050396,-0.3254210353,0.2124046534,-0.4152063429,0.0127494466,-0.2546612024,-0.2034210265,0.0468139276,0.2763831913,-0.2661168575,-0.3130564094,0.388687551,-0.0194335897,0.4227880836,0.4655911326,-0.0001143111,0.1071715355,0.51889956,-0.1366393417,-0.5111317039,-0.3482480347,-0.42827788,0.0564358458,0.192046836,-0.1413850635,0.0406109989,0.1922591478,0.1550243199,0.0707044527,0.2410453558,0.401366204,0.1814022809,0.3804128468,0.2313805819,-0.2351676375,0.1545953453,0.0248502288,-0.0978888422,0.5647636056,-0.0470754243,-0.1141025722,0.1286965609,0.3489504457,-0.1531593651,0.1483006477,0.2271824032,0.0371534787,-0.206946224,-0.2533878982,0.1699537337,0.1438006461,0.6555977464,-0.2468865812,-0.2574800253,-0.0335899629,0.0099552674,-0.257073462,0.2009013444,0.1602668911,-0.0814757422,0.0440074243,-0.0826552436,-0.0206362251,-0.2317146063,0.2486288399,-0.1192691177,0.3788589239,-0.1296508908,0.2292333394,0.1909642965,-0.1613063663,0.4198672473,0.0673462003,-0.00037431,0.1236649305,-0.172006011,0.0920586661,0.0743063763,0.2679724097,0.0050417609,-0.1803118438,-0.0461912714,-0.1731351018,0.0700950995,0.3650493324,0.4140588045,0.1718422025,0.0364552327,0.2923616469,0.2413597256,0.3500970602,-0.0186878219,0.0057673766,-0.2672667205,-0.1643571705,0.1300109178,0.0049816333,0.4590969682,-0.0423844941,-0.4038477242,-0.0003768379,-0.3010182083,0.0824715495,0.0890631452,0.3704461157,0.0278564673,-0.0001119386,0.2095947862,0.1817014366,-0.1293827146,-0.0431667268,-0.1425672174,0.4191194475,-0.1319151819,0.1497374773,0.2187040746,0.0767599866,0.1410082281,-0.1379396319,0.2220699191,0.0446452945,-0.0268260315,-0.1082442328,-0.1772271544,0.2855977118,0.1986786872,0.1157576293,0.271702975,-0.4462215602,-0.2021680623,0.1938804984,-0.3336772621,-0.438097626,-0.2477481961,0.1399913877,-0.1422355026,-0.080782041,-0.6888030171,0.0970536843,0.3203808963,-0.2366490066,-0.1176822484,-0.3462983072,-0.4640255272,-0.1770429164,0.1038890705,0.6105265617,-0.4028030932,-0.103244558,-0.2107320875,0.1306753308,0.1703797132,0.2542798221,-0.3040802181,0.4380046427,-0.1264621168,0.1686676145,0.0272127017,-0.2167752981,-0.2442434728,0.0688678175,0.1441135108,0.1390661299,-0.1006631553,-0.1111619696,0.1963772774,-0.1156174466,0.4870840609,0.2835547328,-0.1532050222,0.0493033864,-0.1728231311,-0.1591173261,0.2652175725,0.2218355387,0.1602453738,0.1063035578,0.0204085279,0.1632985771,-0.0760866478,-0.1615734845,-0.1027200222,0.10231518,0.3372314274,0.0457199253,-0.0576756485,0.070242539,-0.6290019155,0.1177144274,-0.0899862126,0.4876350462,-0.2378153205,-0.0781501308,-0.375954181,0.2315167785,-0.2586623132,0.0211970247,0.1027925834,0.1375456005,-0.3405932784,0.1095857993,0.0024863463,0.3937729895,-0.1800981611,-0.0502063073,-0.5546491146,0.3506638408,-0.2476859093,-0.2609458864,0.2360201627,0.0128415721,0.1512277275,-0.0852371752,-0.1042425856,0.220842585,0.0838627964,0.1782544851,-0.0415431447,0.0498291962,0.0866228566,-0.2470028996,-0.0502538495,0.2800141275,0.30624336,-0.02226023,0.0072829183,0.3388239443,0.2145283222,0.0617963001,-0.1064285338,0.0869064629,0.1529307961,-0.0321946591,-0.01544616,-0.109802112,0.0940818042,0.3077631593,-0.217350319,-0.006536785,-0.3673067391,0.1065693945,0.3472861946,-0.0406758636,0.1063068956,0.2004859149,-0.0458071828,-0.2589944899,0.1528945863,-0.0988076031,0.3416545689,-0.0652151853,-0.1908445209,-0.020092275,0.0983544439,0.0000412357,0.1111444309,0.1305607706,0.2045361847,-0.021802593,0.1338376552,0.0469109789,-0.1636053473,-0.1726353616,0.0007486735,0.489151746,-0.4590349793,0.3445876539,-0.0838145465,-0.3980570436,-0.1177790537,-0.021256553,-0.09370967,-0.3758860826,-0.1497988552,0.476954937,-0.016330108,0.0647049546,0.3318704069,0.3093937635,0.3434869051,-0.1801780015,-0.091370672,-0.366915822,-0.2936855257,0.0322301835,0.4665715098,0.1150845885,0.2103034109,-0.2565609515,-0.0744037926,-0.1555275321,-0.3767433167,0.1977673918,-0.2820949256,0.5574057698,0.1927935779,0.4036362469,-0.4021496177,-0.417080164,0.2165425271,-0.2159273028,0.0228553899,0.0600701869,-0.2575754821,-0.1695455909,-0.0780540779,-0.2154356837,-0.388065964,-0.3334845304,0.0411791652,0.0234914459,0.1087777913,0.4082966149,0.1039955169,0.0826147273,0.500186801,-0.1201792136,-0.2204819471,-0.1896117479,0.3713559806,-0.2679738402,-0.2996506691,0.0574339963,-0.1118457392,-0.0345388837,-0.12770015,-0.3979893625,-0.1679548174,-0.0796003416,0.0343476385,0.0870103762,0.0046829493,0.0903062001,0.0409919173,0.0165092535,-0.1444610506,-0.3971961141,-0.0447067805,0.286280632,0.1896523237,0.3199491799,0.1527796835,-0.2465851158,0.4933204651,0.1270400584,-0.1835066378,0.3492406607,0.0250619091,0.1754009724,-0.3715402186,-0.4368455708,0.2376172245,-0.1879225224,-0.2280799001,0.2979317009,-0.2045451254,-0.3474058807,-0.2248645425,0.043816112,-0.1779914349,-0.437746942,0.2583796978,0.3308208883,0.1570836306,0.0771662146,-0.2223608047,-0.2364102304,-0.1451247334,-0.0093168002,0.406742543,0.0352082923,0.023753399,0.2685171962,-0.01963971,-0.6131858826,0.2619727254,0.3598898351,0.2567749321,-0.1841001809,-0.0063451338,0.0039632614,0.0835695118,0.1082003042,-0.1093386486,-0.0108380774,0.1519353688,0.039184045,-0.0263426173,-0.079027012,-0.3449522853,0.3172373772,0.1861913055,0.2938253284,-0.3232548237,0.102165021,0.2999716401,0.1534431726,-0.1450287849,-0.1421936303,-0.3056723475,-0.5755369663,-0.4369577169,-0.1568490714,0.2291861326,0.3046579957,-0.0445028543,0.2770195007,0.1160707399,-0.0249738786,0.2497914135,0.1906315386,0.2642820776,0.2793500721,-0.003047446,-0.1023888439,-0.0883053988,0.3251291513,0.5740010142,0.0765839741,-0.3027872443,0.1534443647,-0.0007592123,-0.0351323299,0.1420032382,-0.223596096,-0.1190416142,-0.0791736543,0.119191125,0.0921079442,0.3412065208,0.112110123,0.0040060924,-0.6925353408,-0.498118192,0.409096837,-0.0072400025,0.0192797296,0.4975596368,0.186741814,-0.2983270288,0.3348350823,0.1221375391,0.8771522641,-0.1109948829,0.1701840758,0.7048407793,0.3051095605,0.3482864201,0.1293726563,0.1986810714,-0.2439749986,-0.0101689417,-0.1779771447,-0.0347320214,-0.1059867591,-0.178979829,-0.5133098364,0.1631120294,-0.1395597607,0.1532981694,0.1962642223,0.205508858,-0.0852645561,-0.2222532928,-0.4470224679,0.135743469,-0.1891505271,0.1525286436,-0.1698508859,-0.1999875903,-0.0422524959,-0.2237295806,-0.4431297779,-0.0101407804,-0.3017944098,0.2878568172,0.1285059154,-0.3540192842,0.38213557,0.0873135254,0.2093686759,-0.0861255527,-0.2630381882,0.0796146095,-0.0422880054,-0.1822861731,-0.2927992344,-0.0839489549,0.3890060484,0.0104223052,-0.0555330291,-0.1049366593,-0.021770237,-0.3225515783,-0.0210439861,0.106969215,-0.1146417409,-0.3590726256,-0.1863963604,-0.0565642565,-0.0802444518,-0.142415449,0.1051646173,-0.0230294466,-0.0452695191,-0.0582768507,0.0906590894,-0.2889914811,-0.0133597171,0.2108227164,-0.0169843938,0.2019533366,0.5059750676,0.323105365,-0.1487448364,-0.1556158066,0.0903205872,0.0370923765,-0.1766443551,0.3680168688,0.0764473453,0.1777411848,-0.1176417172,0.7146791816,0.0049760677,0.1166734472,0.2042620033,-0.6232945323,-0.1541107148,0.274636358,-0.2994156778,0.1232724786,0.1035278067,0.1669118404,0.0297569539,-0.2967009842,-0.2817041874,0.0297346544,-0.3422638178,0.1165573597,0.5153737664,-0.1448722184,0.2689292133,0.0518569425,0.1047898903,0.0106331427,0.0130031267,-0.2168715596,-0.1562123001,0.13669689,-0.0015312206,-0.2776115239,-0.1031044945,-0.0232947692,-0.2041480094,-0.125789836,0.2397995591,0.2675269544,-0.2015846968,0.0886180773,-0.058473859,0.0901300833,0.0289789978,0.2691264749,0.0576194301,0.3624676764,0.0758867711,0.2564904392,0.121913746,0.0619145632,0.1604729891,0.3062053621,0.1793515235,0.0806679502,0.1938150525,-0.4666214883,0.084897086,-0.3038907349,0.1471400112,0.3480239809,-0.3421799839,0.0292507149,0.1527010947,0.0747291744,-0.1406992674,0.0847325176,0.2097515464,-0.1218280196,-0.0363923311,0.3370079696,0.0952704921,0.1607520878,-0.1302410811,-0.00916922,0.2860715687,0.2742423713,0.1155956089,-0.0315118469,-0.1654588431,0.1139841899,0.0179614518,0.1874173284,0.1831042469,0.0614976995,-0.122942917,0.1994289607,-0.2232550234,0.001412074,0.2800662518,-0.5291490555,0.0418860577,0.0880155936,0.1026348248,0.0713383108,0.0034238612,0.409915477,-0.2465043366,-0.0876880288,-0.2012912929,0.1213345975,-0.1715997159,-0.0220152475,-0.1088467464,-0.085129939,-0.3159866333,0.1381996125,-0.0426836349,-0.0950907618,0.20096156,0.3130761087,-0.0781462565,-0.2880213559,0.0612916015,0.2677075565,-0.1242277622,-0.146975562,0.2357491851,0.0108402213,-0.1058395803,-0.0509677641,0.2660687566,0.7285702825,0.3192522526,-0.2551065385,0.0773122087,-0.0690372959,-0.1405919939,0.027850518,0.0032028682,0.1834236681,0.1844145358,0.3254415393,0.1618837565,-0.0949521735,0.1528599709,0.0883176103,-0.076748617,-0.044845894,-0.06969098,-0.3189104497,0.1108427718,-0.2443201095,0.0555678606,-0.3604250848,0.0488565639,0.3898122311,0.0613457263,0.14079009,0.0984555259,0.0432326831,-0.0891610831,0.7480602264,0.424529314,0.5701190829,-0.4606824219,-0.2080962509,-0.4583554268,0.145920366,-0.0689231753,-0.04683844,-0.0326221958,0.0107746851,0.0743057877,0.2133212984,-0.1900985986,-0.1546644419,-0.1080703586,-0.0489009768,-0.1239415184,-0.3723287582,0.0555763058,-0.1665824652,-0.0945496261,-0.3566226363,-0.0013356621,-0.3693269491,0.0391545407,-0.3989432156,-0.2728853822,-0.0115521234,-0.0481827371,0.1678233594,0.180310905,0.6532689333,-0.1587125659,-0.1381350011,-0.0625675023,-0.3694784939,-0.1216888279,0.1717711687,0.026565209,0.4272362292,-0.0047245272,0.2010180205,-0.4344115257,0.3339697421,0.0141455438,-0.1143100485,-0.4061487615,0.0375447497,-0.2808113098,0.0591574721,-0.0056760763,0.0953870416,-0.1021751538,0.1446498632,-0.1778619289,-0.3812479675,0.7138036489,-0.6547108293,-0.5138835907,-0.0173225515,-0.1966511011,-0.1308061182,0.0336085446,-0.7398446798,0.1849341989,0.0491946824,-0.0236793887,-0.367248565,0.1721331179,0.1016584188,0.0948799625,-0.1482052505,0.2744233608,0.0389564708,-0.1822542995,-0.1232797205,-0.2610343099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2059","title":"Error while following docs to load the `ted_talks_iwslt` dataset","comments":"This has been fixed in #2064 by @mariosasko (thanks again !)\r\n\r\nThe fix is available on the master branch and we'll do a new release very soon :)","body":"I am currently trying to load the `ted_talks_iwslt` dataset into google colab.\r\n\r\nThe [docs](https:\/\/huggingface.co\/datasets\/ted_talks_iwslt) mention the following way of doing so.\r\n\r\n```python\r\ndataset = load_dataset(\"ted_talks_iwslt\", language_pair=(\"it\", \"pl\"), year=\"2014\")\r\n```\r\n\r\nExecuting it results in the error attached below.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"ted_talks_iwslt\", language_pair=(\"it\", \"pl\"), year=\"2014\")\r\n\r\n4 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 730 hash=hash,\r\n 731 features=features,\r\n--> 732 **config_kwargs,\r\n 733 )\r\n 734 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in __init__(self, writer_batch_size, *args, **kwargs)\r\n 927 \r\n 928 def __init__(self, *args, writer_batch_size=None, **kwargs):\r\n--> 929 super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n 930 # Batch size used by the ArrowWriter\r\n 931 # It defines the number of samples that are kept in memory before writing them\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)\r\n 241 name,\r\n 242 custom_features=features,\r\n--> 243 **config_kwargs,\r\n 244 )\r\n 245 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)\r\n 337 if \"version\" not in config_kwargs and hasattr(self, \"VERSION\") and self.VERSION:\r\n 338 config_kwargs[\"version\"] = self.VERSION\r\n--> 339 builder_config = self.BUILDER_CONFIG_CLASS(**config_kwargs)\r\n 340 \r\n 341 # otherwise use the config_kwargs to overwrite the attributes\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/ted_talks_iwslt\/024d06b1376b361e59245c5878ab8acf9a7576d765f2d0077f61751158e60914\/ted_talks_iwslt.py in __init__(self, language_pair, year, **kwargs)\r\n 219 description=description,\r\n 220 version=datasets.Version(\"1.1.0\", \"\"),\r\n--> 221 **kwargs,\r\n 222 )\r\n 223 \r\n\r\nTypeError: __init__() got multiple values for keyword argument 'version'\r\n```\r\n\r\nHow to resolve this? \r\n\r\nPS: Thanks a lot @huggingface team for creating this great library!","comment_length":28,"text":"Error while following docs to load the `ted_talks_iwslt` dataset\nI am currently trying to load the `ted_talks_iwslt` dataset into google colab.\r\n\r\nThe [docs](https:\/\/huggingface.co\/datasets\/ted_talks_iwslt) mention the following way of doing so.\r\n\r\n```python\r\ndataset = load_dataset(\"ted_talks_iwslt\", language_pair=(\"it\", \"pl\"), year=\"2014\")\r\n```\r\n\r\nExecuting it results in the error attached below.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"ted_talks_iwslt\", language_pair=(\"it\", \"pl\"), year=\"2014\")\r\n\r\n4 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)\r\n 730 hash=hash,\r\n 731 features=features,\r\n--> 732 **config_kwargs,\r\n 733 )\r\n 734 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in __init__(self, writer_batch_size, *args, **kwargs)\r\n 927 \r\n 928 def __init__(self, *args, writer_batch_size=None, **kwargs):\r\n--> 929 super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n 930 # Batch size used by the ArrowWriter\r\n 931 # It defines the number of samples that are kept in memory before writing them\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)\r\n 241 name,\r\n 242 custom_features=features,\r\n--> 243 **config_kwargs,\r\n 244 )\r\n 245 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)\r\n 337 if \"version\" not in config_kwargs and hasattr(self, \"VERSION\") and self.VERSION:\r\n 338 config_kwargs[\"version\"] = self.VERSION\r\n--> 339 builder_config = self.BUILDER_CONFIG_CLASS(**config_kwargs)\r\n 340 \r\n 341 # otherwise use the config_kwargs to overwrite the attributes\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/ted_talks_iwslt\/024d06b1376b361e59245c5878ab8acf9a7576d765f2d0077f61751158e60914\/ted_talks_iwslt.py in __init__(self, language_pair, year, **kwargs)\r\n 219 description=description,\r\n 220 version=datasets.Version(\"1.1.0\", \"\"),\r\n--> 221 **kwargs,\r\n 222 )\r\n 223 \r\n\r\nTypeError: __init__() got multiple values for keyword argument 'version'\r\n```\r\n\r\nHow to resolve this? \r\n\r\nPS: Thanks a lot @huggingface team for creating this great library!\nThis has been fixed in #2064 by @mariosasko (thanks again !)\r\n\r\nThe fix is available on the master branch and we'll do a new release very soon :)","embeddings":[-0.2140100002,0.1284443438,0.0485115461,0.2001313567,0.0312912539,0.0933724046,0.6712156534,0.1944082826,0.1241025403,-0.0619180724,-0.0495757982,0.3381051421,-0.291054666,0.4190407395,0.077388525,-0.2890882194,0.0997464731,0.0731823891,-0.0559581928,0.206050396,-0.3254210353,0.2124046534,-0.4152063429,0.0127494466,-0.2546612024,-0.2034210265,0.0468139276,0.2763831913,-0.2661168575,-0.3130564094,0.388687551,-0.0194335897,0.4227880836,0.4655911326,-0.0001143111,0.1071715355,0.51889956,-0.1366393417,-0.5111317039,-0.3482480347,-0.42827788,0.0564358458,0.192046836,-0.1413850635,0.0406109989,0.1922591478,0.1550243199,0.0707044527,0.2410453558,0.401366204,0.1814022809,0.3804128468,0.2313805819,-0.2351676375,0.1545953453,0.0248502288,-0.0978888422,0.5647636056,-0.0470754243,-0.1141025722,0.1286965609,0.3489504457,-0.1531593651,0.1483006477,0.2271824032,0.0371534787,-0.206946224,-0.2533878982,0.1699537337,0.1438006461,0.6555977464,-0.2468865812,-0.2574800253,-0.0335899629,0.0099552674,-0.257073462,0.2009013444,0.1602668911,-0.0814757422,0.0440074243,-0.0826552436,-0.0206362251,-0.2317146063,0.2486288399,-0.1192691177,0.3788589239,-0.1296508908,0.2292333394,0.1909642965,-0.1613063663,0.4198672473,0.0673462003,-0.00037431,0.1236649305,-0.172006011,0.0920586661,0.0743063763,0.2679724097,0.0050417609,-0.1803118438,-0.0461912714,-0.1731351018,0.0700950995,0.3650493324,0.4140588045,0.1718422025,0.0364552327,0.2923616469,0.2413597256,0.3500970602,-0.0186878219,0.0057673766,-0.2672667205,-0.1643571705,0.1300109178,0.0049816333,0.4590969682,-0.0423844941,-0.4038477242,-0.0003768379,-0.3010182083,0.0824715495,0.0890631452,0.3704461157,0.0278564673,-0.0001119386,0.2095947862,0.1817014366,-0.1293827146,-0.0431667268,-0.1425672174,0.4191194475,-0.1319151819,0.1497374773,0.2187040746,0.0767599866,0.1410082281,-0.1379396319,0.2220699191,0.0446452945,-0.0268260315,-0.1082442328,-0.1772271544,0.2855977118,0.1986786872,0.1157576293,0.271702975,-0.4462215602,-0.2021680623,0.1938804984,-0.3336772621,-0.438097626,-0.2477481961,0.1399913877,-0.1422355026,-0.080782041,-0.6888030171,0.0970536843,0.3203808963,-0.2366490066,-0.1176822484,-0.3462983072,-0.4640255272,-0.1770429164,0.1038890705,0.6105265617,-0.4028030932,-0.103244558,-0.2107320875,0.1306753308,0.1703797132,0.2542798221,-0.3040802181,0.4380046427,-0.1264621168,0.1686676145,0.0272127017,-0.2167752981,-0.2442434728,0.0688678175,0.1441135108,0.1390661299,-0.1006631553,-0.1111619696,0.1963772774,-0.1156174466,0.4870840609,0.2835547328,-0.1532050222,0.0493033864,-0.1728231311,-0.1591173261,0.2652175725,0.2218355387,0.1602453738,0.1063035578,0.0204085279,0.1632985771,-0.0760866478,-0.1615734845,-0.1027200222,0.10231518,0.3372314274,0.0457199253,-0.0576756485,0.070242539,-0.6290019155,0.1177144274,-0.0899862126,0.4876350462,-0.2378153205,-0.0781501308,-0.375954181,0.2315167785,-0.2586623132,0.0211970247,0.1027925834,0.1375456005,-0.3405932784,0.1095857993,0.0024863463,0.3937729895,-0.1800981611,-0.0502063073,-0.5546491146,0.3506638408,-0.2476859093,-0.2609458864,0.2360201627,0.0128415721,0.1512277275,-0.0852371752,-0.1042425856,0.220842585,0.0838627964,0.1782544851,-0.0415431447,0.0498291962,0.0866228566,-0.2470028996,-0.0502538495,0.2800141275,0.30624336,-0.02226023,0.0072829183,0.3388239443,0.2145283222,0.0617963001,-0.1064285338,0.0869064629,0.1529307961,-0.0321946591,-0.01544616,-0.109802112,0.0940818042,0.3077631593,-0.217350319,-0.006536785,-0.3673067391,0.1065693945,0.3472861946,-0.0406758636,0.1063068956,0.2004859149,-0.0458071828,-0.2589944899,0.1528945863,-0.0988076031,0.3416545689,-0.0652151853,-0.1908445209,-0.020092275,0.0983544439,0.0000412357,0.1111444309,0.1305607706,0.2045361847,-0.021802593,0.1338376552,0.0469109789,-0.1636053473,-0.1726353616,0.0007486735,0.489151746,-0.4590349793,0.3445876539,-0.0838145465,-0.3980570436,-0.1177790537,-0.021256553,-0.09370967,-0.3758860826,-0.1497988552,0.476954937,-0.016330108,0.0647049546,0.3318704069,0.3093937635,0.3434869051,-0.1801780015,-0.091370672,-0.366915822,-0.2936855257,0.0322301835,0.4665715098,0.1150845885,0.2103034109,-0.2565609515,-0.0744037926,-0.1555275321,-0.3767433167,0.1977673918,-0.2820949256,0.5574057698,0.1927935779,0.4036362469,-0.4021496177,-0.417080164,0.2165425271,-0.2159273028,0.0228553899,0.0600701869,-0.2575754821,-0.1695455909,-0.0780540779,-0.2154356837,-0.388065964,-0.3334845304,0.0411791652,0.0234914459,0.1087777913,0.4082966149,0.1039955169,0.0826147273,0.500186801,-0.1201792136,-0.2204819471,-0.1896117479,0.3713559806,-0.2679738402,-0.2996506691,0.0574339963,-0.1118457392,-0.0345388837,-0.12770015,-0.3979893625,-0.1679548174,-0.0796003416,0.0343476385,0.0870103762,0.0046829493,0.0903062001,0.0409919173,0.0165092535,-0.1444610506,-0.3971961141,-0.0447067805,0.286280632,0.1896523237,0.3199491799,0.1527796835,-0.2465851158,0.4933204651,0.1270400584,-0.1835066378,0.3492406607,0.0250619091,0.1754009724,-0.3715402186,-0.4368455708,0.2376172245,-0.1879225224,-0.2280799001,0.2979317009,-0.2045451254,-0.3474058807,-0.2248645425,0.043816112,-0.1779914349,-0.437746942,0.2583796978,0.3308208883,0.1570836306,0.0771662146,-0.2223608047,-0.2364102304,-0.1451247334,-0.0093168002,0.406742543,0.0352082923,0.023753399,0.2685171962,-0.01963971,-0.6131858826,0.2619727254,0.3598898351,0.2567749321,-0.1841001809,-0.0063451338,0.0039632614,0.0835695118,0.1082003042,-0.1093386486,-0.0108380774,0.1519353688,0.039184045,-0.0263426173,-0.079027012,-0.3449522853,0.3172373772,0.1861913055,0.2938253284,-0.3232548237,0.102165021,0.2999716401,0.1534431726,-0.1450287849,-0.1421936303,-0.3056723475,-0.5755369663,-0.4369577169,-0.1568490714,0.2291861326,0.3046579957,-0.0445028543,0.2770195007,0.1160707399,-0.0249738786,0.2497914135,0.1906315386,0.2642820776,0.2793500721,-0.003047446,-0.1023888439,-0.0883053988,0.3251291513,0.5740010142,0.0765839741,-0.3027872443,0.1534443647,-0.0007592123,-0.0351323299,0.1420032382,-0.223596096,-0.1190416142,-0.0791736543,0.119191125,0.0921079442,0.3412065208,0.112110123,0.0040060924,-0.6925353408,-0.498118192,0.409096837,-0.0072400025,0.0192797296,0.4975596368,0.186741814,-0.2983270288,0.3348350823,0.1221375391,0.8771522641,-0.1109948829,0.1701840758,0.7048407793,0.3051095605,0.3482864201,0.1293726563,0.1986810714,-0.2439749986,-0.0101689417,-0.1779771447,-0.0347320214,-0.1059867591,-0.178979829,-0.5133098364,0.1631120294,-0.1395597607,0.1532981694,0.1962642223,0.205508858,-0.0852645561,-0.2222532928,-0.4470224679,0.135743469,-0.1891505271,0.1525286436,-0.1698508859,-0.1999875903,-0.0422524959,-0.2237295806,-0.4431297779,-0.0101407804,-0.3017944098,0.2878568172,0.1285059154,-0.3540192842,0.38213557,0.0873135254,0.2093686759,-0.0861255527,-0.2630381882,0.0796146095,-0.0422880054,-0.1822861731,-0.2927992344,-0.0839489549,0.3890060484,0.0104223052,-0.0555330291,-0.1049366593,-0.021770237,-0.3225515783,-0.0210439861,0.106969215,-0.1146417409,-0.3590726256,-0.1863963604,-0.0565642565,-0.0802444518,-0.142415449,0.1051646173,-0.0230294466,-0.0452695191,-0.0582768507,0.0906590894,-0.2889914811,-0.0133597171,0.2108227164,-0.0169843938,0.2019533366,0.5059750676,0.323105365,-0.1487448364,-0.1556158066,0.0903205872,0.0370923765,-0.1766443551,0.3680168688,0.0764473453,0.1777411848,-0.1176417172,0.7146791816,0.0049760677,0.1166734472,0.2042620033,-0.6232945323,-0.1541107148,0.274636358,-0.2994156778,0.1232724786,0.1035278067,0.1669118404,0.0297569539,-0.2967009842,-0.2817041874,0.0297346544,-0.3422638178,0.1165573597,0.5153737664,-0.1448722184,0.2689292133,0.0518569425,0.1047898903,0.0106331427,0.0130031267,-0.2168715596,-0.1562123001,0.13669689,-0.0015312206,-0.2776115239,-0.1031044945,-0.0232947692,-0.2041480094,-0.125789836,0.2397995591,0.2675269544,-0.2015846968,0.0886180773,-0.058473859,0.0901300833,0.0289789978,0.2691264749,0.0576194301,0.3624676764,0.0758867711,0.2564904392,0.121913746,0.0619145632,0.1604729891,0.3062053621,0.1793515235,0.0806679502,0.1938150525,-0.4666214883,0.084897086,-0.3038907349,0.1471400112,0.3480239809,-0.3421799839,0.0292507149,0.1527010947,0.0747291744,-0.1406992674,0.0847325176,0.2097515464,-0.1218280196,-0.0363923311,0.3370079696,0.0952704921,0.1607520878,-0.1302410811,-0.00916922,0.2860715687,0.2742423713,0.1155956089,-0.0315118469,-0.1654588431,0.1139841899,0.0179614518,0.1874173284,0.1831042469,0.0614976995,-0.122942917,0.1994289607,-0.2232550234,0.001412074,0.2800662518,-0.5291490555,0.0418860577,0.0880155936,0.1026348248,0.0713383108,0.0034238612,0.409915477,-0.2465043366,-0.0876880288,-0.2012912929,0.1213345975,-0.1715997159,-0.0220152475,-0.1088467464,-0.085129939,-0.3159866333,0.1381996125,-0.0426836349,-0.0950907618,0.20096156,0.3130761087,-0.0781462565,-0.2880213559,0.0612916015,0.2677075565,-0.1242277622,-0.146975562,0.2357491851,0.0108402213,-0.1058395803,-0.0509677641,0.2660687566,0.7285702825,0.3192522526,-0.2551065385,0.0773122087,-0.0690372959,-0.1405919939,0.027850518,0.0032028682,0.1834236681,0.1844145358,0.3254415393,0.1618837565,-0.0949521735,0.1528599709,0.0883176103,-0.076748617,-0.044845894,-0.06969098,-0.3189104497,0.1108427718,-0.2443201095,0.0555678606,-0.3604250848,0.0488565639,0.3898122311,0.0613457263,0.14079009,0.0984555259,0.0432326831,-0.0891610831,0.7480602264,0.424529314,0.5701190829,-0.4606824219,-0.2080962509,-0.4583554268,0.145920366,-0.0689231753,-0.04683844,-0.0326221958,0.0107746851,0.0743057877,0.2133212984,-0.1900985986,-0.1546644419,-0.1080703586,-0.0489009768,-0.1239415184,-0.3723287582,0.0555763058,-0.1665824652,-0.0945496261,-0.3566226363,-0.0013356621,-0.3693269491,0.0391545407,-0.3989432156,-0.2728853822,-0.0115521234,-0.0481827371,0.1678233594,0.180310905,0.6532689333,-0.1587125659,-0.1381350011,-0.0625675023,-0.3694784939,-0.1216888279,0.1717711687,0.026565209,0.4272362292,-0.0047245272,0.2010180205,-0.4344115257,0.3339697421,0.0141455438,-0.1143100485,-0.4061487615,0.0375447497,-0.2808113098,0.0591574721,-0.0056760763,0.0953870416,-0.1021751538,0.1446498632,-0.1778619289,-0.3812479675,0.7138036489,-0.6547108293,-0.5138835907,-0.0173225515,-0.1966511011,-0.1308061182,0.0336085446,-0.7398446798,0.1849341989,0.0491946824,-0.0236793887,-0.367248565,0.1721331179,0.1016584188,0.0948799625,-0.1482052505,0.2744233608,0.0389564708,-0.1822542995,-0.1232797205,-0.2610343099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2056","title":"issue with opus100\/en-fr dataset ","comments":"@lhoestq I also deleted the cache and redownload the file and still the same issue, I appreciate any help on this. thanks ","body":"Hi\r\nI am running run_mlm.py code of huggingface repo with opus100\/fr-en pair, I am getting this error, note that this error occurs for only this pairs and not the other pairs. Any idea why this is occurring? and how I can solve this? \r\n\r\nThanks a lot @lhoestq for your help in advance.\r\n\r\n`\r\nthread '' panicked at 'index out of bounds: the len is 617 but the index is 617', \/__w\/tokenizers\/tokenizers\/tokenizers\/src\/tokenizer\/normalizer.rs:382:21\r\nnote: run with `RUST_BACKTRACE=1` environment variable to display a backtrace\r\n 63%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 626\/1000 [00:27<00:16, 22.69ba\/s]\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm.py\", line 550, in \r\n main()\r\n File \"run_mlm.py\", line 412, in main\r\n in zip(data_args.dataset_name, data_args.dataset_config_name)]\r\n File \"run_mlm.py\", line 411, in \r\n logger) for dataset_name, dataset_config_name\\\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 96, in get_tokenized_dataset\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in map\r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in \r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1309, in map\r\n update_data=update_data,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 204, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1574, in _map_single\r\n batch, indices, check_same_num_examples=len(self.list_indexes()) > 0, offset=offset\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1490, in apply_function_on_filtered_inputs\r\n function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 89, in tokenize_function\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2347, in __call__\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2532, in batch_encode_plus\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_fast.py\", line 384, in _batch_encode_plus\r\n is_pretokenized=is_split_into_words,\r\npyo3_runtime.PanicException: index out of bounds: the len is 617 but the index is 617\r\n\r\n`","comment_length":22,"text":"issue with opus100\/en-fr dataset \nHi\r\nI am running run_mlm.py code of huggingface repo with opus100\/fr-en pair, I am getting this error, note that this error occurs for only this pairs and not the other pairs. Any idea why this is occurring? and how I can solve this? \r\n\r\nThanks a lot @lhoestq for your help in advance.\r\n\r\n`\r\nthread '' panicked at 'index out of bounds: the len is 617 but the index is 617', \/__w\/tokenizers\/tokenizers\/tokenizers\/src\/tokenizer\/normalizer.rs:382:21\r\nnote: run with `RUST_BACKTRACE=1` environment variable to display a backtrace\r\n 63%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 626\/1000 [00:27<00:16, 22.69ba\/s]\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm.py\", line 550, in \r\n main()\r\n File \"run_mlm.py\", line 412, in main\r\n in zip(data_args.dataset_name, data_args.dataset_config_name)]\r\n File \"run_mlm.py\", line 411, in \r\n logger) for dataset_name, dataset_config_name\\\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 96, in get_tokenized_dataset\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in map\r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in \r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1309, in map\r\n update_data=update_data,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 204, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1574, in _map_single\r\n batch, indices, check_same_num_examples=len(self.list_indexes()) > 0, offset=offset\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1490, in apply_function_on_filtered_inputs\r\n function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 89, in tokenize_function\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2347, in __call__\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2532, in batch_encode_plus\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_fast.py\", line 384, in _batch_encode_plus\r\n is_pretokenized=is_split_into_words,\r\npyo3_runtime.PanicException: index out of bounds: the len is 617 but the index is 617\r\n\r\n`\n@lhoestq I also deleted the cache and redownload the file and still the same issue, I appreciate any help on this. thanks ","embeddings":[-0.4117364585,-0.1673511267,-0.0156450551,0.3972483575,0.1182573065,0.0823763385,0.1922526211,0.2819981277,-0.1431498528,0.2579930723,-0.2424349785,-0.0248958599,0.1322234273,0.3834213614,-0.2399135977,-0.1725855172,-0.0782720894,0.1208091751,-0.4165076911,-0.1202893779,-0.4559483826,0.2238832414,-0.1563926637,0.2297671288,0.3186606765,-0.0702556372,0.1518447101,0.057933744,-0.2905824184,-0.3938674927,-0.1512016654,-0.2992527187,0.3983563483,0.2486596107,-0.0001240568,0.0139625799,0.1579867601,-0.0384621657,0.0809596032,-0.0592175759,0.1000378951,0.1473319381,-0.0287124142,-0.3051408231,0.0371817499,-0.2251016945,-0.3708876371,-0.2680359483,0.2880217433,0.3753131926,0.0867760256,0.5228812695,0.3373613358,-0.1462075114,0.1830143183,-0.0422982462,0.0586619824,-0.0693615079,0.323479563,-0.2280785143,0.3684239984,0.2993505597,-0.1102045476,-0.136514172,0.2203771174,-0.2563655972,0.6804661155,-0.5535476804,0.011115835,0.3440338969,0.073990345,0.0230665654,-0.0722477585,0.0284326747,-0.0147417244,-0.8122330308,-0.0916823149,0.1136502177,-0.0426637977,0.1667119116,-0.3281251788,0.056635309,0.0846511796,0.180134967,0.0343027003,0.3327208459,0.0061036893,0.2988551259,0.2502726614,-0.229519859,-0.409422785,0.2705802023,-0.108532019,0.2871657014,-0.4670937657,-0.0439887866,0.0700741261,-0.1362367272,0.4946710467,0.2132567763,-0.5131159425,-0.162680462,0.4100387394,-0.0067464188,0.2849927545,0.269594878,-0.0111154756,0.023031868,0.202965647,0.0215273388,0.1223784089,-0.0572648272,-0.0405006334,-0.2940114141,-0.2561855912,-0.0725328997,0.1384642124,-0.489279747,-0.2409026027,0.3881003857,-0.3146696985,-0.0595293082,0.5197177529,0.3339307308,0.196629256,-0.1151291132,-0.0928354487,0.2790851295,-0.3406359553,0.0021832027,-0.0180696789,-0.0317981578,-0.0366431437,0.0159903839,-0.0337793902,-0.6434463263,0.2684477866,0.1619419903,0.5190970302,-0.4269457459,0.1198232174,-0.2199243009,-0.0168505087,0.2775986493,-0.0571954623,0.1540126652,0.3632331789,0.0032506017,-0.0590387098,0.0531924404,-0.4613539279,-0.4059747458,-0.3444899619,0.0495101698,-0.1387339532,0.4477998018,0.1756897718,-0.0166503843,0.4423442781,-0.1872699559,0.2043519616,-0.182628423,-0.3020606339,0.0518628433,0.3899740875,0.1399208009,0.0265579876,0.1484417319,0.2194242626,0.2127541155,0.3668286502,0.5186986923,-0.2727029324,-0.0015196375,-0.4783477783,0.2312484533,0.0005177835,-0.4624033868,-0.2426678091,-0.1247510314,-0.2872537076,-0.2944654524,0.0553656332,-0.1781386286,0.6758924723,-0.0452677011,0.2189907879,0.0732169747,0.0459526889,-0.1863196194,-0.2554853857,0.0077219098,0.3513787687,-0.1988129765,0.0082609039,-0.0232906919,-0.0493819155,-0.1482340842,0.1101337969,-0.060543824,0.1323255599,0.2601273656,0.2107945085,0.0742594004,0.2487873733,-0.0611242391,0.1535781622,-0.0508140847,-0.4693951607,0.251824528,-0.0783604905,0.0079228571,-0.3655864894,0.0975491628,-0.1030970439,-0.2655682862,-0.0031700695,-0.2116549015,0.1113093868,0.1097497866,0.1644181311,-0.1983311623,-0.0019175176,0.2205998451,-0.3111978769,0.3991585374,-0.2442423105,-0.0184909273,0.0290290676,0.3571317792,0.3908420801,-0.2759011984,-0.1644442379,0.2347953171,0.0223703906,-0.2906487584,-0.2894754708,0.3155043125,0.1678450257,-0.3867717683,0.0024176578,-0.282569766,0.0304201841,-0.0328209475,0.1935184002,0.527335763,-0.1704324037,0.3984831274,-0.2610982358,0.1100407392,0.0254600924,0.1451858878,0.0004749658,-0.1778210849,0.3125909269,0.3107431531,0.1120001227,0.0675271749,-0.1241336092,-0.0501553342,0.2982763052,0.0116719827,0.1062054485,0.0498672538,-0.0823500603,-0.0183439478,0.1508110166,-0.2998584211,0.1713975817,0.1358766407,-0.0547967032,0.0122387363,-0.0049690604,-0.0121079898,0.2557174265,0.0212531257,0.3840557337,0.192031607,0.0013486212,-0.0890289843,-0.1562478244,-0.105488345,0.1233065203,0.4478662312,-0.2748400867,0.0627049878,-0.0854327828,0.0197419319,-0.1227129921,-0.3330768645,-0.6193659306,-0.1888411939,0.1114291176,0.0365404487,0.0309601035,0.0561556891,0.1487271488,-0.070354946,0.2286869437,0.1457809955,0.169361338,-0.0629823133,-0.1823603064,-0.1482490748,0.1953943819,-0.0163381845,0.0471094511,0.1365008801,-0.0967713669,-0.0325959511,-0.5894052386,0.310652107,-0.0568139292,0.3301675618,0.0491235182,0.1147355363,-0.0633864552,-0.1362303346,0.1913006306,0.099512808,-0.2120435089,0.0822472572,-0.0835589767,0.2379554212,-0.1326675117,-0.095053643,-0.2327533215,-0.4111004472,0.2914522886,-0.1913591027,0.0587136559,0.2794517279,-0.1564147621,-0.0205895007,0.0797978863,-0.1391308457,-0.3224676251,0.1024730131,0.347124368,-0.2841886282,-0.2615577579,-0.2328037471,0.1744868457,0.2109635621,-0.1036836728,-0.4139459133,-0.021609962,-0.1247051358,0.0797610208,-0.2371442914,0.1470896006,0.3746415079,-0.1469144076,-0.0013900505,-0.1065231264,-0.1248332486,0.1823372692,-0.0900530741,0.0056254733,-0.0408166088,0.744967103,-0.1414622217,0.5462422371,0.1166381612,0.046352122,0.3334827721,-0.185360834,0.3638515472,-0.092579186,-0.3035498559,0.2539817989,0.2515974045,-0.2397891879,0.1635255367,-0.1627686918,0.0914804041,-0.0657879934,-0.1014053598,0.0549396351,-0.2810799778,-0.1190145016,0.4435954094,0.3401423693,0.0735057816,0.41277349,-0.0851292908,-0.157809943,0.1060775742,-0.0302284099,-0.0383120961,0.1415879577,-0.6263697743,-0.4313113689,-0.5289623141,0.4155212641,0.0741583556,0.748724699,-0.187167272,0.1006233096,0.0712511837,0.063962765,0.7198229432,0.2071083933,-0.0213005617,-0.0399165452,-0.1477170885,-0.2423105687,-0.1408007592,-0.0278389994,0.5661061406,0.0240683462,0.6595690846,-0.2567321658,-0.2979437113,0.3916319609,0.1320486963,-0.1679752469,-0.0606126711,-0.1955357492,-0.3098434508,-0.2639372945,0.3379350305,0.1940551847,0.5438346863,-0.1234486997,-0.0158896577,0.1631438881,-0.1943352222,0.1099143252,0.1643001735,0.4007405937,0.0797052085,0.0898569897,0.2112823874,-0.0558952838,0.5950174928,0.1040690988,-0.0604194663,-0.0915406346,0.1939139068,0.0131677492,0.3254158497,0.1391021162,0.0967573076,0.1080204174,0.1463077366,0.3830463886,-0.2549988329,-0.0272539631,0.2192225605,0.2564139366,-0.3760876358,0.0446603857,0.0586689338,0.0565799884,-0.0172104146,0.2615416944,0.3256276846,-0.1692548543,0.1965737045,0.4372459054,1.125349164,-0.2956349552,-0.078039892,-0.1270174831,-0.3361191452,0.6042557955,-0.2630318403,-0.0462497771,-0.3473007679,-0.1285597831,0.0290812161,-0.1867011786,-0.0097719468,-0.2190949172,-0.099764891,0.1527134031,-0.2251032293,-0.1580952257,-0.218285352,0.226946041,0.3224573731,-0.2332212925,-0.1530539542,-0.0327682011,-0.0646058768,0.1714268178,-0.2102534473,-0.1722311825,0.1970991492,-0.2368182093,-0.2299950868,-0.1415082216,-0.4525684714,0.2946381271,0.0208070837,-0.3227231205,0.3300321102,0.1336202323,0.1228149831,0.4325302541,-0.2202580869,-0.0268769749,-0.0063686059,-0.1093931869,-0.0867890269,0.0674539655,0.0446674637,-0.1134758964,-0.4193345904,0.1008559316,-0.0415861048,-0.037232589,-0.1182167381,0.1514539123,0.1590322554,-0.2331365496,0.3106337488,0.0292764306,-0.1071387455,-0.2391301095,-0.0605494678,0.1102115586,-0.4307785928,0.0195449647,0.0935486481,-0.4853807688,0.0089948885,0.4414263964,-0.0217245277,-0.1256331056,0.5604746938,0.3870260119,-0.0833009779,-0.0891310275,0.0665605441,0.3794840872,-0.3428511322,-0.0206781495,-0.0268811062,0.0366385616,-0.0318849385,0.0771935806,0.1465908736,-0.0055677518,-0.1001252905,-0.4927748442,-0.4502181709,0.1050361916,-0.1411198974,0.0919762179,0.1732307374,0.0533688739,-0.2124432176,-0.0619384088,-0.1585219353,-0.0634850785,-0.420733124,0.2324081659,0.100873895,-0.4312159717,0.0396448337,-0.0233106241,-0.0935195684,0.1902826726,-0.3172387779,-0.0028025273,-0.155354172,0.2393624485,-0.0955499709,-0.5256474614,0.0028382766,-0.1402626336,-0.1286252439,-0.1450250298,-0.0684643611,0.1721591651,0.0660321191,-0.2844120562,0.2785306275,-0.2525462806,-0.1802474558,0.0735497847,0.2667729855,0.1805518717,-0.0312233157,0.4272300005,0.0469693057,-0.179416284,-0.4443587661,0.0844650939,0.2403591126,-0.308485806,0.0818073303,-0.2272557318,-0.1002404541,0.4624747038,0.4395340085,0.395262152,0.0054785269,-0.0640100166,0.2593303621,0.0696913451,-0.1671456099,-0.1368118972,0.251573801,-0.00493652,-0.0690292567,0.4297878146,0.2808301747,0.0913784578,0.0630438253,0.0423072539,0.2633594275,0.1425088048,0.08284612,0.1284723282,-0.0913555473,-0.0380900465,0.2750041783,-0.0623211786,0.4039423764,0.3254387379,-0.1907827556,0.1284182072,0.4210965037,0.1274306476,0.1366223544,-0.262463361,-0.112323992,0.2244531214,0.2346693873,0.1128061041,-0.1113445535,0.7335345745,-0.0874511972,-0.3345427215,-0.0257382207,0.4591722488,-0.2747836411,-0.0890216753,-0.1561333835,-0.1324678957,-0.0322470851,-0.068789348,0.1778804064,0.0589245968,-0.022229692,0.0233925693,-0.222255826,-0.3103192151,-0.5450291634,0.5357750654,-0.0293432325,-0.3221189976,0.3328133523,0.386228919,-0.1073277593,0.1693546176,0.3321025074,0.5232444406,0.2327000052,-0.1036638245,-0.0796245709,0.0639358088,0.1587669402,-0.0327845067,-0.0839481801,0.0879547298,0.1356130987,0.0877456293,-0.0667454153,-0.0521095432,-0.0449392349,0.2036179751,0.059542004,0.0997648016,-0.5444077253,-0.1548729837,-0.1310877502,-0.3298236728,0.0288382452,-0.1454810649,0.2842448354,0.3009548783,-0.0438529812,0.0339539237,0.0367655456,0.0023883199,-0.1856237799,0.4092003405,0.4066260457,0.0112806344,-0.2685199678,-0.1344890743,-0.4066004753,0.2426142544,-0.3182495236,-0.1087302491,0.416859746,0.1510163844,0.0382546,0.0451904312,0.4998104572,-0.2465702444,-0.1361933351,0.2110142559,-0.3237796426,-0.0380775221,-0.2183679342,-0.0580942333,0.1589758843,0.0137183722,0.3272307813,-0.0571000203,-0.0797360167,-0.309233129,0.2265756577,-0.3428740799,-0.0971133262,0.1831239164,0.4092595875,0.2603732049,0.1465851367,0.0863461643,0.1715337485,-0.2983761132,-0.0559379421,0.2614252567,0.0259766113,0.0188197214,-0.1280235201,-0.7292591333,-0.4894476235,0.1618310958,0.0186074711,-0.2161207497,-0.2587124407,0.322077781,-0.0869769156,0.0181753226,0.1464764774,0.3136515021,0.3915123045,0.242584765,-0.2641705275,-0.4147348106,0.6981275082,-0.5851553679,-0.202804029,-0.202739343,0.2233406752,0.5235945582,-0.3709131777,-0.3115878999,-0.0513185337,0.2126802206,0.2346115261,-0.4300559163,-0.0479076914,-0.2759914994,-0.087454766,0.026260104,-0.0255265851,0.1617361456,-0.2601888776,0.2698343098,-0.0476782434]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2056","title":"issue with opus100\/en-fr dataset ","comments":"Here please find the minimal code to reproduce the issue @lhoestq note this only happens with MT5TokenizerFast\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import MT5TokenizerFast\r\n\r\ndef get_tokenized_dataset(dataset_name, dataset_config_name, tokenizer):\r\n datasets = load_dataset(dataset_name, dataset_config_name, script_version=\"master\")\r\n column_names = datasets[\"train\"].column_names\r\n text_column_name = \"translation\"\r\n def process_dataset(datasets):\r\n def process_function(examples):\r\n lang = \"fr\"\r\n return {\"src_texts\": [example[lang] for example in examples[text_column_name]]}\r\n datasets = datasets.map(\r\n process_function,\r\n batched=True,\r\n num_proc=None,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n )\r\n return datasets\r\n datasets = process_dataset(datasets)\r\n text_column_name = \"src_texts\"\r\n column_names = [\"src_texts\"]\r\n def tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n tokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=None,\r\n remove_columns=column_names,\r\n load_from_cache_file=True\r\n )\r\n\r\nif __name__ == \"__main__\":\r\n tokenizer_kwargs = {\r\n \"cache_dir\": None,\r\n \"use_fast\": True,\r\n \"revision\": \"main\",\r\n \"use_auth_token\": None\r\n }\r\n tokenizer = MT5TokenizerFast.from_pretrained(\"google\/mt5-small\", **tokenizer_kwargs)\r\n get_tokenized_dataset(dataset_name=\"opus100\", dataset_config_name=\"en-fr\", tokenizer=tokenizer)\r\n~ \r\n```","body":"Hi\r\nI am running run_mlm.py code of huggingface repo with opus100\/fr-en pair, I am getting this error, note that this error occurs for only this pairs and not the other pairs. Any idea why this is occurring? and how I can solve this? \r\n\r\nThanks a lot @lhoestq for your help in advance.\r\n\r\n`\r\nthread '' panicked at 'index out of bounds: the len is 617 but the index is 617', \/__w\/tokenizers\/tokenizers\/tokenizers\/src\/tokenizer\/normalizer.rs:382:21\r\nnote: run with `RUST_BACKTRACE=1` environment variable to display a backtrace\r\n 63%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 626\/1000 [00:27<00:16, 22.69ba\/s]\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm.py\", line 550, in \r\n main()\r\n File \"run_mlm.py\", line 412, in main\r\n in zip(data_args.dataset_name, data_args.dataset_config_name)]\r\n File \"run_mlm.py\", line 411, in \r\n logger) for dataset_name, dataset_config_name\\\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 96, in get_tokenized_dataset\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in map\r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in \r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1309, in map\r\n update_data=update_data,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 204, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1574, in _map_single\r\n batch, indices, check_same_num_examples=len(self.list_indexes()) > 0, offset=offset\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1490, in apply_function_on_filtered_inputs\r\n function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 89, in tokenize_function\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2347, in __call__\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2532, in batch_encode_plus\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_fast.py\", line 384, in _batch_encode_plus\r\n is_pretokenized=is_split_into_words,\r\npyo3_runtime.PanicException: index out of bounds: the len is 617 but the index is 617\r\n\r\n`","comment_length":114,"text":"issue with opus100\/en-fr dataset \nHi\r\nI am running run_mlm.py code of huggingface repo with opus100\/fr-en pair, I am getting this error, note that this error occurs for only this pairs and not the other pairs. Any idea why this is occurring? and how I can solve this? \r\n\r\nThanks a lot @lhoestq for your help in advance.\r\n\r\n`\r\nthread '' panicked at 'index out of bounds: the len is 617 but the index is 617', \/__w\/tokenizers\/tokenizers\/tokenizers\/src\/tokenizer\/normalizer.rs:382:21\r\nnote: run with `RUST_BACKTRACE=1` environment variable to display a backtrace\r\n 63%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 626\/1000 [00:27<00:16, 22.69ba\/s]\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm.py\", line 550, in \r\n main()\r\n File \"run_mlm.py\", line 412, in main\r\n in zip(data_args.dataset_name, data_args.dataset_config_name)]\r\n File \"run_mlm.py\", line 411, in \r\n logger) for dataset_name, dataset_config_name\\\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 96, in get_tokenized_dataset\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in map\r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in \r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1309, in map\r\n update_data=update_data,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 204, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1574, in _map_single\r\n batch, indices, check_same_num_examples=len(self.list_indexes()) > 0, offset=offset\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1490, in apply_function_on_filtered_inputs\r\n function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 89, in tokenize_function\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2347, in __call__\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2532, in batch_encode_plus\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_fast.py\", line 384, in _batch_encode_plus\r\n is_pretokenized=is_split_into_words,\r\npyo3_runtime.PanicException: index out of bounds: the len is 617 but the index is 617\r\n\r\n`\nHere please find the minimal code to reproduce the issue @lhoestq note this only happens with MT5TokenizerFast\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import MT5TokenizerFast\r\n\r\ndef get_tokenized_dataset(dataset_name, dataset_config_name, tokenizer):\r\n datasets = load_dataset(dataset_name, dataset_config_name, script_version=\"master\")\r\n column_names = datasets[\"train\"].column_names\r\n text_column_name = \"translation\"\r\n def process_dataset(datasets):\r\n def process_function(examples):\r\n lang = \"fr\"\r\n return {\"src_texts\": [example[lang] for example in examples[text_column_name]]}\r\n datasets = datasets.map(\r\n process_function,\r\n batched=True,\r\n num_proc=None,\r\n remove_columns=column_names,\r\n load_from_cache_file=True,\r\n )\r\n return datasets\r\n datasets = process_dataset(datasets)\r\n text_column_name = \"src_texts\"\r\n column_names = [\"src_texts\"]\r\n def tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n tokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=None,\r\n remove_columns=column_names,\r\n load_from_cache_file=True\r\n )\r\n\r\nif __name__ == \"__main__\":\r\n tokenizer_kwargs = {\r\n \"cache_dir\": None,\r\n \"use_fast\": True,\r\n \"revision\": \"main\",\r\n \"use_auth_token\": None\r\n }\r\n tokenizer = MT5TokenizerFast.from_pretrained(\"google\/mt5-small\", **tokenizer_kwargs)\r\n get_tokenized_dataset(dataset_name=\"opus100\", dataset_config_name=\"en-fr\", tokenizer=tokenizer)\r\n~ \r\n```","embeddings":[-0.4117364585,-0.1673511267,-0.0156450551,0.3972483575,0.1182573065,0.0823763385,0.1922526211,0.2819981277,-0.1431498528,0.2579930723,-0.2424349785,-0.0248958599,0.1322234273,0.3834213614,-0.2399135977,-0.1725855172,-0.0782720894,0.1208091751,-0.4165076911,-0.1202893779,-0.4559483826,0.2238832414,-0.1563926637,0.2297671288,0.3186606765,-0.0702556372,0.1518447101,0.057933744,-0.2905824184,-0.3938674927,-0.1512016654,-0.2992527187,0.3983563483,0.2486596107,-0.0001240568,0.0139625799,0.1579867601,-0.0384621657,0.0809596032,-0.0592175759,0.1000378951,0.1473319381,-0.0287124142,-0.3051408231,0.0371817499,-0.2251016945,-0.3708876371,-0.2680359483,0.2880217433,0.3753131926,0.0867760256,0.5228812695,0.3373613358,-0.1462075114,0.1830143183,-0.0422982462,0.0586619824,-0.0693615079,0.323479563,-0.2280785143,0.3684239984,0.2993505597,-0.1102045476,-0.136514172,0.2203771174,-0.2563655972,0.6804661155,-0.5535476804,0.011115835,0.3440338969,0.073990345,0.0230665654,-0.0722477585,0.0284326747,-0.0147417244,-0.8122330308,-0.0916823149,0.1136502177,-0.0426637977,0.1667119116,-0.3281251788,0.056635309,0.0846511796,0.180134967,0.0343027003,0.3327208459,0.0061036893,0.2988551259,0.2502726614,-0.229519859,-0.409422785,0.2705802023,-0.108532019,0.2871657014,-0.4670937657,-0.0439887866,0.0700741261,-0.1362367272,0.4946710467,0.2132567763,-0.5131159425,-0.162680462,0.4100387394,-0.0067464188,0.2849927545,0.269594878,-0.0111154756,0.023031868,0.202965647,0.0215273388,0.1223784089,-0.0572648272,-0.0405006334,-0.2940114141,-0.2561855912,-0.0725328997,0.1384642124,-0.489279747,-0.2409026027,0.3881003857,-0.3146696985,-0.0595293082,0.5197177529,0.3339307308,0.196629256,-0.1151291132,-0.0928354487,0.2790851295,-0.3406359553,0.0021832027,-0.0180696789,-0.0317981578,-0.0366431437,0.0159903839,-0.0337793902,-0.6434463263,0.2684477866,0.1619419903,0.5190970302,-0.4269457459,0.1198232174,-0.2199243009,-0.0168505087,0.2775986493,-0.0571954623,0.1540126652,0.3632331789,0.0032506017,-0.0590387098,0.0531924404,-0.4613539279,-0.4059747458,-0.3444899619,0.0495101698,-0.1387339532,0.4477998018,0.1756897718,-0.0166503843,0.4423442781,-0.1872699559,0.2043519616,-0.182628423,-0.3020606339,0.0518628433,0.3899740875,0.1399208009,0.0265579876,0.1484417319,0.2194242626,0.2127541155,0.3668286502,0.5186986923,-0.2727029324,-0.0015196375,-0.4783477783,0.2312484533,0.0005177835,-0.4624033868,-0.2426678091,-0.1247510314,-0.2872537076,-0.2944654524,0.0553656332,-0.1781386286,0.6758924723,-0.0452677011,0.2189907879,0.0732169747,0.0459526889,-0.1863196194,-0.2554853857,0.0077219098,0.3513787687,-0.1988129765,0.0082609039,-0.0232906919,-0.0493819155,-0.1482340842,0.1101337969,-0.060543824,0.1323255599,0.2601273656,0.2107945085,0.0742594004,0.2487873733,-0.0611242391,0.1535781622,-0.0508140847,-0.4693951607,0.251824528,-0.0783604905,0.0079228571,-0.3655864894,0.0975491628,-0.1030970439,-0.2655682862,-0.0031700695,-0.2116549015,0.1113093868,0.1097497866,0.1644181311,-0.1983311623,-0.0019175176,0.2205998451,-0.3111978769,0.3991585374,-0.2442423105,-0.0184909273,0.0290290676,0.3571317792,0.3908420801,-0.2759011984,-0.1644442379,0.2347953171,0.0223703906,-0.2906487584,-0.2894754708,0.3155043125,0.1678450257,-0.3867717683,0.0024176578,-0.282569766,0.0304201841,-0.0328209475,0.1935184002,0.527335763,-0.1704324037,0.3984831274,-0.2610982358,0.1100407392,0.0254600924,0.1451858878,0.0004749658,-0.1778210849,0.3125909269,0.3107431531,0.1120001227,0.0675271749,-0.1241336092,-0.0501553342,0.2982763052,0.0116719827,0.1062054485,0.0498672538,-0.0823500603,-0.0183439478,0.1508110166,-0.2998584211,0.1713975817,0.1358766407,-0.0547967032,0.0122387363,-0.0049690604,-0.0121079898,0.2557174265,0.0212531257,0.3840557337,0.192031607,0.0013486212,-0.0890289843,-0.1562478244,-0.105488345,0.1233065203,0.4478662312,-0.2748400867,0.0627049878,-0.0854327828,0.0197419319,-0.1227129921,-0.3330768645,-0.6193659306,-0.1888411939,0.1114291176,0.0365404487,0.0309601035,0.0561556891,0.1487271488,-0.070354946,0.2286869437,0.1457809955,0.169361338,-0.0629823133,-0.1823603064,-0.1482490748,0.1953943819,-0.0163381845,0.0471094511,0.1365008801,-0.0967713669,-0.0325959511,-0.5894052386,0.310652107,-0.0568139292,0.3301675618,0.0491235182,0.1147355363,-0.0633864552,-0.1362303346,0.1913006306,0.099512808,-0.2120435089,0.0822472572,-0.0835589767,0.2379554212,-0.1326675117,-0.095053643,-0.2327533215,-0.4111004472,0.2914522886,-0.1913591027,0.0587136559,0.2794517279,-0.1564147621,-0.0205895007,0.0797978863,-0.1391308457,-0.3224676251,0.1024730131,0.347124368,-0.2841886282,-0.2615577579,-0.2328037471,0.1744868457,0.2109635621,-0.1036836728,-0.4139459133,-0.021609962,-0.1247051358,0.0797610208,-0.2371442914,0.1470896006,0.3746415079,-0.1469144076,-0.0013900505,-0.1065231264,-0.1248332486,0.1823372692,-0.0900530741,0.0056254733,-0.0408166088,0.744967103,-0.1414622217,0.5462422371,0.1166381612,0.046352122,0.3334827721,-0.185360834,0.3638515472,-0.092579186,-0.3035498559,0.2539817989,0.2515974045,-0.2397891879,0.1635255367,-0.1627686918,0.0914804041,-0.0657879934,-0.1014053598,0.0549396351,-0.2810799778,-0.1190145016,0.4435954094,0.3401423693,0.0735057816,0.41277349,-0.0851292908,-0.157809943,0.1060775742,-0.0302284099,-0.0383120961,0.1415879577,-0.6263697743,-0.4313113689,-0.5289623141,0.4155212641,0.0741583556,0.748724699,-0.187167272,0.1006233096,0.0712511837,0.063962765,0.7198229432,0.2071083933,-0.0213005617,-0.0399165452,-0.1477170885,-0.2423105687,-0.1408007592,-0.0278389994,0.5661061406,0.0240683462,0.6595690846,-0.2567321658,-0.2979437113,0.3916319609,0.1320486963,-0.1679752469,-0.0606126711,-0.1955357492,-0.3098434508,-0.2639372945,0.3379350305,0.1940551847,0.5438346863,-0.1234486997,-0.0158896577,0.1631438881,-0.1943352222,0.1099143252,0.1643001735,0.4007405937,0.0797052085,0.0898569897,0.2112823874,-0.0558952838,0.5950174928,0.1040690988,-0.0604194663,-0.0915406346,0.1939139068,0.0131677492,0.3254158497,0.1391021162,0.0967573076,0.1080204174,0.1463077366,0.3830463886,-0.2549988329,-0.0272539631,0.2192225605,0.2564139366,-0.3760876358,0.0446603857,0.0586689338,0.0565799884,-0.0172104146,0.2615416944,0.3256276846,-0.1692548543,0.1965737045,0.4372459054,1.125349164,-0.2956349552,-0.078039892,-0.1270174831,-0.3361191452,0.6042557955,-0.2630318403,-0.0462497771,-0.3473007679,-0.1285597831,0.0290812161,-0.1867011786,-0.0097719468,-0.2190949172,-0.099764891,0.1527134031,-0.2251032293,-0.1580952257,-0.218285352,0.226946041,0.3224573731,-0.2332212925,-0.1530539542,-0.0327682011,-0.0646058768,0.1714268178,-0.2102534473,-0.1722311825,0.1970991492,-0.2368182093,-0.2299950868,-0.1415082216,-0.4525684714,0.2946381271,0.0208070837,-0.3227231205,0.3300321102,0.1336202323,0.1228149831,0.4325302541,-0.2202580869,-0.0268769749,-0.0063686059,-0.1093931869,-0.0867890269,0.0674539655,0.0446674637,-0.1134758964,-0.4193345904,0.1008559316,-0.0415861048,-0.037232589,-0.1182167381,0.1514539123,0.1590322554,-0.2331365496,0.3106337488,0.0292764306,-0.1071387455,-0.2391301095,-0.0605494678,0.1102115586,-0.4307785928,0.0195449647,0.0935486481,-0.4853807688,0.0089948885,0.4414263964,-0.0217245277,-0.1256331056,0.5604746938,0.3870260119,-0.0833009779,-0.0891310275,0.0665605441,0.3794840872,-0.3428511322,-0.0206781495,-0.0268811062,0.0366385616,-0.0318849385,0.0771935806,0.1465908736,-0.0055677518,-0.1001252905,-0.4927748442,-0.4502181709,0.1050361916,-0.1411198974,0.0919762179,0.1732307374,0.0533688739,-0.2124432176,-0.0619384088,-0.1585219353,-0.0634850785,-0.420733124,0.2324081659,0.100873895,-0.4312159717,0.0396448337,-0.0233106241,-0.0935195684,0.1902826726,-0.3172387779,-0.0028025273,-0.155354172,0.2393624485,-0.0955499709,-0.5256474614,0.0028382766,-0.1402626336,-0.1286252439,-0.1450250298,-0.0684643611,0.1721591651,0.0660321191,-0.2844120562,0.2785306275,-0.2525462806,-0.1802474558,0.0735497847,0.2667729855,0.1805518717,-0.0312233157,0.4272300005,0.0469693057,-0.179416284,-0.4443587661,0.0844650939,0.2403591126,-0.308485806,0.0818073303,-0.2272557318,-0.1002404541,0.4624747038,0.4395340085,0.395262152,0.0054785269,-0.0640100166,0.2593303621,0.0696913451,-0.1671456099,-0.1368118972,0.251573801,-0.00493652,-0.0690292567,0.4297878146,0.2808301747,0.0913784578,0.0630438253,0.0423072539,0.2633594275,0.1425088048,0.08284612,0.1284723282,-0.0913555473,-0.0380900465,0.2750041783,-0.0623211786,0.4039423764,0.3254387379,-0.1907827556,0.1284182072,0.4210965037,0.1274306476,0.1366223544,-0.262463361,-0.112323992,0.2244531214,0.2346693873,0.1128061041,-0.1113445535,0.7335345745,-0.0874511972,-0.3345427215,-0.0257382207,0.4591722488,-0.2747836411,-0.0890216753,-0.1561333835,-0.1324678957,-0.0322470851,-0.068789348,0.1778804064,0.0589245968,-0.022229692,0.0233925693,-0.222255826,-0.3103192151,-0.5450291634,0.5357750654,-0.0293432325,-0.3221189976,0.3328133523,0.386228919,-0.1073277593,0.1693546176,0.3321025074,0.5232444406,0.2327000052,-0.1036638245,-0.0796245709,0.0639358088,0.1587669402,-0.0327845067,-0.0839481801,0.0879547298,0.1356130987,0.0877456293,-0.0667454153,-0.0521095432,-0.0449392349,0.2036179751,0.059542004,0.0997648016,-0.5444077253,-0.1548729837,-0.1310877502,-0.3298236728,0.0288382452,-0.1454810649,0.2842448354,0.3009548783,-0.0438529812,0.0339539237,0.0367655456,0.0023883199,-0.1856237799,0.4092003405,0.4066260457,0.0112806344,-0.2685199678,-0.1344890743,-0.4066004753,0.2426142544,-0.3182495236,-0.1087302491,0.416859746,0.1510163844,0.0382546,0.0451904312,0.4998104572,-0.2465702444,-0.1361933351,0.2110142559,-0.3237796426,-0.0380775221,-0.2183679342,-0.0580942333,0.1589758843,0.0137183722,0.3272307813,-0.0571000203,-0.0797360167,-0.309233129,0.2265756577,-0.3428740799,-0.0971133262,0.1831239164,0.4092595875,0.2603732049,0.1465851367,0.0863461643,0.1715337485,-0.2983761132,-0.0559379421,0.2614252567,0.0259766113,0.0188197214,-0.1280235201,-0.7292591333,-0.4894476235,0.1618310958,0.0186074711,-0.2161207497,-0.2587124407,0.322077781,-0.0869769156,0.0181753226,0.1464764774,0.3136515021,0.3915123045,0.242584765,-0.2641705275,-0.4147348106,0.6981275082,-0.5851553679,-0.202804029,-0.202739343,0.2233406752,0.5235945582,-0.3709131777,-0.3115878999,-0.0513185337,0.2126802206,0.2346115261,-0.4300559163,-0.0479076914,-0.2759914994,-0.087454766,0.026260104,-0.0255265851,0.1617361456,-0.2601888776,0.2698343098,-0.0476782434]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2056","title":"issue with opus100\/en-fr dataset ","comments":"as per https:\/\/github.com\/huggingface\/tokenizers\/issues\/626 this looks like to be the tokenizer bug, I therefore, reported it there https:\/\/github.com\/huggingface\/tokenizers\/issues\/626 and I am closing this one.","body":"Hi\r\nI am running run_mlm.py code of huggingface repo with opus100\/fr-en pair, I am getting this error, note that this error occurs for only this pairs and not the other pairs. Any idea why this is occurring? and how I can solve this? \r\n\r\nThanks a lot @lhoestq for your help in advance.\r\n\r\n`\r\nthread '' panicked at 'index out of bounds: the len is 617 but the index is 617', \/__w\/tokenizers\/tokenizers\/tokenizers\/src\/tokenizer\/normalizer.rs:382:21\r\nnote: run with `RUST_BACKTRACE=1` environment variable to display a backtrace\r\n 63%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 626\/1000 [00:27<00:16, 22.69ba\/s]\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm.py\", line 550, in \r\n main()\r\n File \"run_mlm.py\", line 412, in main\r\n in zip(data_args.dataset_name, data_args.dataset_config_name)]\r\n File \"run_mlm.py\", line 411, in \r\n logger) for dataset_name, dataset_config_name\\\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 96, in get_tokenized_dataset\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in map\r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in \r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1309, in map\r\n update_data=update_data,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 204, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1574, in _map_single\r\n batch, indices, check_same_num_examples=len(self.list_indexes()) > 0, offset=offset\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1490, in apply_function_on_filtered_inputs\r\n function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 89, in tokenize_function\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2347, in __call__\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2532, in batch_encode_plus\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_fast.py\", line 384, in _batch_encode_plus\r\n is_pretokenized=is_split_into_words,\r\npyo3_runtime.PanicException: index out of bounds: the len is 617 but the index is 617\r\n\r\n`","comment_length":23,"text":"issue with opus100\/en-fr dataset \nHi\r\nI am running run_mlm.py code of huggingface repo with opus100\/fr-en pair, I am getting this error, note that this error occurs for only this pairs and not the other pairs. Any idea why this is occurring? and how I can solve this? \r\n\r\nThanks a lot @lhoestq for your help in advance.\r\n\r\n`\r\nthread '' panicked at 'index out of bounds: the len is 617 but the index is 617', \/__w\/tokenizers\/tokenizers\/tokenizers\/src\/tokenizer\/normalizer.rs:382:21\r\nnote: run with `RUST_BACKTRACE=1` environment variable to display a backtrace\r\n 63%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258a | 626\/1000 [00:27<00:16, 22.69ba\/s]\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm.py\", line 550, in \r\n main()\r\n File \"run_mlm.py\", line 412, in main\r\n in zip(data_args.dataset_name, data_args.dataset_config_name)]\r\n File \"run_mlm.py\", line 411, in \r\n logger) for dataset_name, dataset_config_name\\\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 96, in get_tokenized_dataset\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in map\r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py\", line 448, in \r\n for k, dataset in self.items()\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1309, in map\r\n update_data=update_data,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 204, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py\", line 337, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1574, in _map_single\r\n batch, indices, check_same_num_examples=len(self.list_indexes()) > 0, offset=offset\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1490, in apply_function_on_filtered_inputs\r\n function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/user\/dara\/dev\/codes\/seq2seq\/data\/tokenize_datasets.py\", line 89, in tokenize_function\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2347, in __call__\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_base.py\", line 2532, in batch_encode_plus\r\n **kwargs,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/transformers\/tokenization_utils_fast.py\", line 384, in _batch_encode_plus\r\n is_pretokenized=is_split_into_words,\r\npyo3_runtime.PanicException: index out of bounds: the len is 617 but the index is 617\r\n\r\n`\nas per https:\/\/github.com\/huggingface\/tokenizers\/issues\/626 this looks like to be the tokenizer bug, I therefore, reported it there https:\/\/github.com\/huggingface\/tokenizers\/issues\/626 and I am closing this one.","embeddings":[-0.4117364585,-0.1673511267,-0.0156450551,0.3972483575,0.1182573065,0.0823763385,0.1922526211,0.2819981277,-0.1431498528,0.2579930723,-0.2424349785,-0.0248958599,0.1322234273,0.3834213614,-0.2399135977,-0.1725855172,-0.0782720894,0.1208091751,-0.4165076911,-0.1202893779,-0.4559483826,0.2238832414,-0.1563926637,0.2297671288,0.3186606765,-0.0702556372,0.1518447101,0.057933744,-0.2905824184,-0.3938674927,-0.1512016654,-0.2992527187,0.3983563483,0.2486596107,-0.0001240568,0.0139625799,0.1579867601,-0.0384621657,0.0809596032,-0.0592175759,0.1000378951,0.1473319381,-0.0287124142,-0.3051408231,0.0371817499,-0.2251016945,-0.3708876371,-0.2680359483,0.2880217433,0.3753131926,0.0867760256,0.5228812695,0.3373613358,-0.1462075114,0.1830143183,-0.0422982462,0.0586619824,-0.0693615079,0.323479563,-0.2280785143,0.3684239984,0.2993505597,-0.1102045476,-0.136514172,0.2203771174,-0.2563655972,0.6804661155,-0.5535476804,0.011115835,0.3440338969,0.073990345,0.0230665654,-0.0722477585,0.0284326747,-0.0147417244,-0.8122330308,-0.0916823149,0.1136502177,-0.0426637977,0.1667119116,-0.3281251788,0.056635309,0.0846511796,0.180134967,0.0343027003,0.3327208459,0.0061036893,0.2988551259,0.2502726614,-0.229519859,-0.409422785,0.2705802023,-0.108532019,0.2871657014,-0.4670937657,-0.0439887866,0.0700741261,-0.1362367272,0.4946710467,0.2132567763,-0.5131159425,-0.162680462,0.4100387394,-0.0067464188,0.2849927545,0.269594878,-0.0111154756,0.023031868,0.202965647,0.0215273388,0.1223784089,-0.0572648272,-0.0405006334,-0.2940114141,-0.2561855912,-0.0725328997,0.1384642124,-0.489279747,-0.2409026027,0.3881003857,-0.3146696985,-0.0595293082,0.5197177529,0.3339307308,0.196629256,-0.1151291132,-0.0928354487,0.2790851295,-0.3406359553,0.0021832027,-0.0180696789,-0.0317981578,-0.0366431437,0.0159903839,-0.0337793902,-0.6434463263,0.2684477866,0.1619419903,0.5190970302,-0.4269457459,0.1198232174,-0.2199243009,-0.0168505087,0.2775986493,-0.0571954623,0.1540126652,0.3632331789,0.0032506017,-0.0590387098,0.0531924404,-0.4613539279,-0.4059747458,-0.3444899619,0.0495101698,-0.1387339532,0.4477998018,0.1756897718,-0.0166503843,0.4423442781,-0.1872699559,0.2043519616,-0.182628423,-0.3020606339,0.0518628433,0.3899740875,0.1399208009,0.0265579876,0.1484417319,0.2194242626,0.2127541155,0.3668286502,0.5186986923,-0.2727029324,-0.0015196375,-0.4783477783,0.2312484533,0.0005177835,-0.4624033868,-0.2426678091,-0.1247510314,-0.2872537076,-0.2944654524,0.0553656332,-0.1781386286,0.6758924723,-0.0452677011,0.2189907879,0.0732169747,0.0459526889,-0.1863196194,-0.2554853857,0.0077219098,0.3513787687,-0.1988129765,0.0082609039,-0.0232906919,-0.0493819155,-0.1482340842,0.1101337969,-0.060543824,0.1323255599,0.2601273656,0.2107945085,0.0742594004,0.2487873733,-0.0611242391,0.1535781622,-0.0508140847,-0.4693951607,0.251824528,-0.0783604905,0.0079228571,-0.3655864894,0.0975491628,-0.1030970439,-0.2655682862,-0.0031700695,-0.2116549015,0.1113093868,0.1097497866,0.1644181311,-0.1983311623,-0.0019175176,0.2205998451,-0.3111978769,0.3991585374,-0.2442423105,-0.0184909273,0.0290290676,0.3571317792,0.3908420801,-0.2759011984,-0.1644442379,0.2347953171,0.0223703906,-0.2906487584,-0.2894754708,0.3155043125,0.1678450257,-0.3867717683,0.0024176578,-0.282569766,0.0304201841,-0.0328209475,0.1935184002,0.527335763,-0.1704324037,0.3984831274,-0.2610982358,0.1100407392,0.0254600924,0.1451858878,0.0004749658,-0.1778210849,0.3125909269,0.3107431531,0.1120001227,0.0675271749,-0.1241336092,-0.0501553342,0.2982763052,0.0116719827,0.1062054485,0.0498672538,-0.0823500603,-0.0183439478,0.1508110166,-0.2998584211,0.1713975817,0.1358766407,-0.0547967032,0.0122387363,-0.0049690604,-0.0121079898,0.2557174265,0.0212531257,0.3840557337,0.192031607,0.0013486212,-0.0890289843,-0.1562478244,-0.105488345,0.1233065203,0.4478662312,-0.2748400867,0.0627049878,-0.0854327828,0.0197419319,-0.1227129921,-0.3330768645,-0.6193659306,-0.1888411939,0.1114291176,0.0365404487,0.0309601035,0.0561556891,0.1487271488,-0.070354946,0.2286869437,0.1457809955,0.169361338,-0.0629823133,-0.1823603064,-0.1482490748,0.1953943819,-0.0163381845,0.0471094511,0.1365008801,-0.0967713669,-0.0325959511,-0.5894052386,0.310652107,-0.0568139292,0.3301675618,0.0491235182,0.1147355363,-0.0633864552,-0.1362303346,0.1913006306,0.099512808,-0.2120435089,0.0822472572,-0.0835589767,0.2379554212,-0.1326675117,-0.095053643,-0.2327533215,-0.4111004472,0.2914522886,-0.1913591027,0.0587136559,0.2794517279,-0.1564147621,-0.0205895007,0.0797978863,-0.1391308457,-0.3224676251,0.1024730131,0.347124368,-0.2841886282,-0.2615577579,-0.2328037471,0.1744868457,0.2109635621,-0.1036836728,-0.4139459133,-0.021609962,-0.1247051358,0.0797610208,-0.2371442914,0.1470896006,0.3746415079,-0.1469144076,-0.0013900505,-0.1065231264,-0.1248332486,0.1823372692,-0.0900530741,0.0056254733,-0.0408166088,0.744967103,-0.1414622217,0.5462422371,0.1166381612,0.046352122,0.3334827721,-0.185360834,0.3638515472,-0.092579186,-0.3035498559,0.2539817989,0.2515974045,-0.2397891879,0.1635255367,-0.1627686918,0.0914804041,-0.0657879934,-0.1014053598,0.0549396351,-0.2810799778,-0.1190145016,0.4435954094,0.3401423693,0.0735057816,0.41277349,-0.0851292908,-0.157809943,0.1060775742,-0.0302284099,-0.0383120961,0.1415879577,-0.6263697743,-0.4313113689,-0.5289623141,0.4155212641,0.0741583556,0.748724699,-0.187167272,0.1006233096,0.0712511837,0.063962765,0.7198229432,0.2071083933,-0.0213005617,-0.0399165452,-0.1477170885,-0.2423105687,-0.1408007592,-0.0278389994,0.5661061406,0.0240683462,0.6595690846,-0.2567321658,-0.2979437113,0.3916319609,0.1320486963,-0.1679752469,-0.0606126711,-0.1955357492,-0.3098434508,-0.2639372945,0.3379350305,0.1940551847,0.5438346863,-0.1234486997,-0.0158896577,0.1631438881,-0.1943352222,0.1099143252,0.1643001735,0.4007405937,0.0797052085,0.0898569897,0.2112823874,-0.0558952838,0.5950174928,0.1040690988,-0.0604194663,-0.0915406346,0.1939139068,0.0131677492,0.3254158497,0.1391021162,0.0967573076,0.1080204174,0.1463077366,0.3830463886,-0.2549988329,-0.0272539631,0.2192225605,0.2564139366,-0.3760876358,0.0446603857,0.0586689338,0.0565799884,-0.0172104146,0.2615416944,0.3256276846,-0.1692548543,0.1965737045,0.4372459054,1.125349164,-0.2956349552,-0.078039892,-0.1270174831,-0.3361191452,0.6042557955,-0.2630318403,-0.0462497771,-0.3473007679,-0.1285597831,0.0290812161,-0.1867011786,-0.0097719468,-0.2190949172,-0.099764891,0.1527134031,-0.2251032293,-0.1580952257,-0.218285352,0.226946041,0.3224573731,-0.2332212925,-0.1530539542,-0.0327682011,-0.0646058768,0.1714268178,-0.2102534473,-0.1722311825,0.1970991492,-0.2368182093,-0.2299950868,-0.1415082216,-0.4525684714,0.2946381271,0.0208070837,-0.3227231205,0.3300321102,0.1336202323,0.1228149831,0.4325302541,-0.2202580869,-0.0268769749,-0.0063686059,-0.1093931869,-0.0867890269,0.0674539655,0.0446674637,-0.1134758964,-0.4193345904,0.1008559316,-0.0415861048,-0.037232589,-0.1182167381,0.1514539123,0.1590322554,-0.2331365496,0.3106337488,0.0292764306,-0.1071387455,-0.2391301095,-0.0605494678,0.1102115586,-0.4307785928,0.0195449647,0.0935486481,-0.4853807688,0.0089948885,0.4414263964,-0.0217245277,-0.1256331056,0.5604746938,0.3870260119,-0.0833009779,-0.0891310275,0.0665605441,0.3794840872,-0.3428511322,-0.0206781495,-0.0268811062,0.0366385616,-0.0318849385,0.0771935806,0.1465908736,-0.0055677518,-0.1001252905,-0.4927748442,-0.4502181709,0.1050361916,-0.1411198974,0.0919762179,0.1732307374,0.0533688739,-0.2124432176,-0.0619384088,-0.1585219353,-0.0634850785,-0.420733124,0.2324081659,0.100873895,-0.4312159717,0.0396448337,-0.0233106241,-0.0935195684,0.1902826726,-0.3172387779,-0.0028025273,-0.155354172,0.2393624485,-0.0955499709,-0.5256474614,0.0028382766,-0.1402626336,-0.1286252439,-0.1450250298,-0.0684643611,0.1721591651,0.0660321191,-0.2844120562,0.2785306275,-0.2525462806,-0.1802474558,0.0735497847,0.2667729855,0.1805518717,-0.0312233157,0.4272300005,0.0469693057,-0.179416284,-0.4443587661,0.0844650939,0.2403591126,-0.308485806,0.0818073303,-0.2272557318,-0.1002404541,0.4624747038,0.4395340085,0.395262152,0.0054785269,-0.0640100166,0.2593303621,0.0696913451,-0.1671456099,-0.1368118972,0.251573801,-0.00493652,-0.0690292567,0.4297878146,0.2808301747,0.0913784578,0.0630438253,0.0423072539,0.2633594275,0.1425088048,0.08284612,0.1284723282,-0.0913555473,-0.0380900465,0.2750041783,-0.0623211786,0.4039423764,0.3254387379,-0.1907827556,0.1284182072,0.4210965037,0.1274306476,0.1366223544,-0.262463361,-0.112323992,0.2244531214,0.2346693873,0.1128061041,-0.1113445535,0.7335345745,-0.0874511972,-0.3345427215,-0.0257382207,0.4591722488,-0.2747836411,-0.0890216753,-0.1561333835,-0.1324678957,-0.0322470851,-0.068789348,0.1778804064,0.0589245968,-0.022229692,0.0233925693,-0.222255826,-0.3103192151,-0.5450291634,0.5357750654,-0.0293432325,-0.3221189976,0.3328133523,0.386228919,-0.1073277593,0.1693546176,0.3321025074,0.5232444406,0.2327000052,-0.1036638245,-0.0796245709,0.0639358088,0.1587669402,-0.0327845067,-0.0839481801,0.0879547298,0.1356130987,0.0877456293,-0.0667454153,-0.0521095432,-0.0449392349,0.2036179751,0.059542004,0.0997648016,-0.5444077253,-0.1548729837,-0.1310877502,-0.3298236728,0.0288382452,-0.1454810649,0.2842448354,0.3009548783,-0.0438529812,0.0339539237,0.0367655456,0.0023883199,-0.1856237799,0.4092003405,0.4066260457,0.0112806344,-0.2685199678,-0.1344890743,-0.4066004753,0.2426142544,-0.3182495236,-0.1087302491,0.416859746,0.1510163844,0.0382546,0.0451904312,0.4998104572,-0.2465702444,-0.1361933351,0.2110142559,-0.3237796426,-0.0380775221,-0.2183679342,-0.0580942333,0.1589758843,0.0137183722,0.3272307813,-0.0571000203,-0.0797360167,-0.309233129,0.2265756577,-0.3428740799,-0.0971133262,0.1831239164,0.4092595875,0.2603732049,0.1465851367,0.0863461643,0.1715337485,-0.2983761132,-0.0559379421,0.2614252567,0.0259766113,0.0188197214,-0.1280235201,-0.7292591333,-0.4894476235,0.1618310958,0.0186074711,-0.2161207497,-0.2587124407,0.322077781,-0.0869769156,0.0181753226,0.1464764774,0.3136515021,0.3915123045,0.242584765,-0.2641705275,-0.4147348106,0.6981275082,-0.5851553679,-0.202804029,-0.202739343,0.2233406752,0.5235945582,-0.3709131777,-0.3115878999,-0.0513185337,0.2126802206,0.2346115261,-0.4300559163,-0.0479076914,-0.2759914994,-0.087454766,0.026260104,-0.0255265851,0.1617361456,-0.2601888776,0.2698343098,-0.0476782434]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2055","title":"is there a way to override a dataset object saved with save_to_disk?","comments":"Hi\r\nYou can rename the arrow file and update the name in `state.json`","body":"At the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object? ","comment_length":13,"text":"is there a way to override a dataset object saved with save_to_disk?\nAt the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object? \nHi\r\nYou can rename the arrow file and update the name in `state.json`","embeddings":[0.0290824901,-0.0669385567,-0.1109161451,0.0110404678,0.297128737,0.1145982593,0.0862810612,0.1684683561,-0.0366806313,0.1778945625,0.3181955218,0.5976210833,-0.1179353818,0.117323108,0.2590530813,-0.004592753,0.4122774601,-0.0318114944,-0.1144979745,0.201816678,-0.389906317,0.1658174694,0.102802217,0.0652164221,-0.1087922305,-0.021918945,0.0443024524,-0.1522608101,0.0768251121,-0.4691328108,0.2196728736,0.2376427352,-0.0211878587,0.0780239403,-0.0001201354,-0.1255272329,0.1620586514,-0.0674505159,-0.232074365,0.2319110036,-0.3821344674,-0.334648937,0.0507733226,-0.4449907541,0.3736407161,-0.2400531769,0.155268833,-0.2317253351,0.129359439,-0.1640929431,0.1052748635,-0.0394303948,0.2383006513,0.276639998,-0.0255456138,0.3497878909,-0.1356008351,0.3613725007,0.1114880815,0.2871632278,0.145024389,-0.0392704494,-0.0986597836,-0.1135112569,0.5661987662,-0.0524487048,0.071011439,0.1016326845,0.1451319903,0.0862656161,0.752417922,-0.445489198,-0.0584942177,-0.0084846877,0.0242855009,-0.0545631237,0.3127009273,-0.2887464464,0.0267860387,0.1331492811,-0.1134403422,-0.7609710097,-0.2147166729,0.2170454115,0.1896302998,-0.2571585476,-0.1575543433,0.2476945966,0.1259164214,-0.1129888371,0.4618984759,-0.4031203687,-0.3698241413,-0.2871871591,-0.1512638479,-0.2856685221,-0.4118880928,0.0531867221,0.0667927861,0.2599148154,0.4170829058,0.2022216916,-0.0699661598,0.2390246242,0.2848590612,0.0473786406,0.1300095916,0.0400888659,0.2266876996,-0.2781829536,-0.1903050989,-0.0823016837,0.2452842593,-0.3144748509,0.5298185349,-0.2523054481,0.2437642068,-0.2539272606,0.106559284,-0.0950793773,-0.0034800549,0.0607712157,-0.0243500434,-0.1558921188,0.1961905509,0.0610318892,-0.0343190059,0.1067039222,0.200573653,0.1831292808,0.0492347926,-0.2769542933,0.3125210404,0.3710670769,0.0425575599,0.1037907973,-0.0442901291,-0.2469238043,-0.040880397,-0.1207182035,0.3319947422,0.0600199141,0.2957596183,-0.0033138993,-0.1686688811,0.1642791033,-0.1006455943,-0.2992245853,-0.1989631355,0.3963280022,-0.2676696181,-0.3020299077,-0.020183932,0.0615913086,-0.3808911443,-0.1690968871,-0.7177521586,0.0844821408,-0.0997368246,-0.1498227417,0.2548542321,0.245305866,-0.0374228358,-0.2999063134,-0.0667878464,0.0506545268,-0.4749055803,-0.0439055152,-0.2439725995,-0.1922059953,-0.3146085143,-0.1835587472,-0.1996402889,-0.0103700319,-0.2293478101,-0.0496715121,0.377484262,-0.2410544604,-0.5433995128,-0.0995490402,-0.1206321418,-0.2913807631,-0.0210714079,0.4069001079,0.0101901665,0.0324670747,-0.4412511885,0.1430630535,0.1519557834,-0.274125129,-0.15277493,-0.0737239495,-0.1998746693,0.0011387629,-0.4312528074,0.2588478625,0.0071301553,-0.0250984393,0.0349325165,0.2196632922,0.2274094969,0.5643891096,0.5418727398,-0.212308079,-0.01393924,-0.1912164837,-0.389308989,-0.0707242563,0.5634804368,-0.3940131962,0.2090540677,-0.1628065705,-0.0089854579,0.1640586257,-0.1410789937,0.3554646671,-0.0399111547,0.2848323584,0.0461760946,-0.1394424736,-0.2814504802,0.0334778167,0.0477443449,0.0716225058,-0.0060219741,0.1657306701,0.0637458339,-0.0780330375,-0.3442320824,-0.3392095864,0.0996021479,-0.1624126732,-0.1523749679,0.5221378803,-0.1396122575,0.1011630595,0.2908811271,0.3236258924,-0.0433599055,-0.4458353221,0.2582396269,0.2441400439,0.0273389351,0.0045894734,-0.3043874204,0.1308549196,-0.0665124431,0.118428342,-0.1692990363,0.1760061234,0.2048779279,0.023611242,-0.2822107375,-0.2561579049,-0.4309224784,0.1442116946,0.2510586977,-0.0558755547,-0.2947147191,0.1864539236,0.6170754433,-0.1444340646,0.1783150136,0.0325228646,-0.3956991434,-0.0681257844,-0.1258204132,0.7047411203,0.4692658186,0.1177531034,0.009963911,-0.213684991,0.1690686792,-0.0879919827,0.1957046539,0.2556256056,0.3052059412,0.0415167771,0.3370558918,-0.2920261621,-0.1460894942,0.043099802,0.0722187161,0.0933783948,-0.0219361335,0.0593290441,-0.195883587,-0.0212290175,-0.3647371531,0.0940381736,-0.400949806,-0.3060769141,0.0384267643,0.3992936909,0.0505511276,0.1373799145,0.1651247442,0.4330902398,-0.2336854637,-0.3855931461,-0.2035268098,-0.0001023264,-0.2182395756,-0.0781503618,0.1760454625,-0.1680297256,0.2269697487,0.1133611649,0.1872547865,-0.5202222466,0.0806106478,-0.1432268023,0.2445131987,0.1871163249,-0.0173230749,0.1961875558,-0.2711485326,0.1844316721,-0.0164413638,0.0624958389,-0.0613760091,-0.0111817457,0.1029594019,0.1911108792,0.1017780155,0.2066723257,-0.1549317241,-0.1079751477,0.4887972474,-0.0678082258,0.0099254269,-0.3017837405,0.1004355624,0.1337541342,-0.0164625309,-0.4623020589,-0.1416591257,-0.5856049657,0.5107581019,-0.16722399,-0.0411232561,0.0865840912,0.1080156416,0.0427767262,0.0063413866,-0.3641648591,-0.1153290048,-0.0590736903,0.1000240892,-0.1342115998,-0.2845230401,0.1953382641,0.5945359468,-0.031161055,-0.0720400438,-0.5326192975,0.2499675751,0.3281763792,0.4306745827,0.2552344203,0.1460701823,-0.0796321109,0.1701421887,-0.2780136168,-0.000340199,0.3048796952,0.1217500269,0.1935799718,0.0966188014,-0.010677672,-0.1960144788,-0.088985607,-0.0361211114,-0.145998314,0.0066382233,0.1517252028,0.2609936595,-0.0837110281,-0.0258252323,-0.3155072629,-0.0879202038,-0.1165833995,0.087140426,-0.0230363738,-0.2030093074,-0.2643209696,-0.1335124522,0.1249362081,0.4396831095,0.0289706569,-0.0888493434,-0.7454845905,-0.1419745088,-0.2499866635,0.3597760201,-0.3070669174,0.0614770241,0.0586859137,0.095296599,-0.0165499151,-0.1471580565,0.7538661957,-0.0972329676,0.2131872475,-0.0057177674,0.2353410125,-0.2610651851,0.144500196,0.2348044068,0.3513081074,-0.1716157198,0.4301171601,0.1886257827,-0.2693012953,-0.0137178153,0.2918458581,-0.48774001,-0.1964940429,0.1009440497,-0.2170743048,-0.0993293896,-0.3106736243,0.082747601,0.0179571211,0.0786489099,-0.1387915164,-0.3701258898,-0.0341941305,-0.0043319752,-0.2703355253,0.4833751023,-0.2149157971,-0.0187592693,-0.0977406278,0.1842317879,0.614035666,0.188400507,-0.0784540549,-0.2790254653,-0.0207726359,-0.2073467821,0.2958264351,0.0129321478,0.0990576297,0.2249243855,0.2713085711,-0.2006138563,-0.0454213209,0.1081405133,-0.0877801552,0.2875762284,-0.1103388518,-0.4240391254,0.4368521273,0.250015676,0.0269886721,0.2084657997,0.3884197772,-0.2716828585,0.4755642712,0.1014195681,0.755209446,-0.0074858153,0.3052990735,-0.2329822332,-0.5667369366,0.1364646554,-0.0478653535,0.0066398284,-0.1517811716,0.0428306423,0.0337784439,-0.0285743475,0.3358472884,0.5603522062,-0.0493734665,0.1053327695,-0.1483107209,-0.1109031886,-0.3616367877,0.2909418941,-0.0573154017,-0.0550332777,0.0916220322,0.0759030208,-0.152282536,-0.191505,0.1505873799,-0.1448829025,-0.033398509,0.0883762762,-0.1821546108,-0.3437512517,0.0289865229,-0.2063203603,-0.1138186902,-0.4113215208,0.1767830998,0.3219112456,0.2812591791,0.1062320247,-0.0888070539,0.1967458874,0.1019174308,0.0956749097,-0.1265209466,-0.040836338,0.1657615602,-0.02824682,-0.1447343528,0.0086870231,0.1913620383,-0.4807872474,0.0618212037,-0.1549662799,0.0231459886,-0.0384138264,-0.2852809727,-0.1299421489,-0.2674476206,-0.3484968245,0.0961202979,0.1228709519,-0.0424971804,0.3161394,-0.0750699267,-0.0701030642,0.1758438349,0.4452525377,-0.0477126464,-0.0114418659,0.6408709884,-0.3480842412,0.081864588,-0.2041751593,0.374148339,0.1339058131,-0.0708971769,0.2477232218,-0.2579486668,0.0942860022,-0.041300144,-0.091920726,-0.033024177,0.2621027827,-0.2888781428,-0.126081109,-0.6444514394,0.0440161377,0.3089289367,0.2850215137,-0.094940342,-0.3505547643,-0.2382342666,-0.1243175492,-0.1499541104,0.0454239361,0.0425499044,-0.0665838942,0.0831764564,0.4690559804,0.0228693094,-0.2365098596,-0.0386081524,-0.1043951064,0.0844458863,-0.1475324333,-0.2246354222,0.1777027696,-0.0517768227,-0.2722081244,-0.2015628666,-0.4173530638,0.1813202053,0.1127904952,-0.3148560822,0.0965544879,0.0988762677,0.3424230516,0.0645455122,0.0296050422,0.1837991327,-0.1949179918,0.2364148647,0.2168076187,0.0957372338,0.2437461764,0.115267016,-0.0904911757,-0.1846913695,-0.0727789402,0.0104076723,0.0992325321,-0.2214678675,-0.1243915856,-0.0802088082,0.2927674055,0.2218065411,0.1856421381,-0.251537323,-0.0926882029,0.3577319384,0.0540595725,0.1310709864,-0.409584403,0.4628585875,-0.2472867221,-0.1322832555,0.0201052707,-0.1416823119,0.1920672357,-0.4662486315,-0.01677108,0.1114990041,-0.2027468532,0.1156817302,0.2634921968,-0.0330526344,-0.0106119448,0.4564465284,0.4160779715,0.3829723001,0.7170675397,0.0947211832,-0.0266385563,0.4433262646,-0.0387773961,0.0982696712,0.0748567134,0.0512479804,-0.1241474599,0.0158540905,0.3326816857,0.5359646082,0.068987228,0.0911074653,-0.3595743477,0.1858490407,-0.0825455263,0.0173168387,-0.3327590525,-0.0834919512,0.0463569015,0.2678429484,0.2543881238,-0.2210082859,-0.3618994057,0.3464702666,-0.1263299733,0.0903273076,0.1433230788,0.315145582,-0.0680628121,0.0674439222,0.0023731091,-0.2823308706,-0.1507751048,-0.0559959337,0.1576706916,-0.0111342557,-0.2723992467,-0.2319913805,-0.0598047115,0.132829383,-0.2204555273,-0.1445578635,-0.0826268792,-0.2605776787,0.0636811554,-0.2475251853,0.2030412108,0.0719681382,0.0910797343,0.0059938603,0.3620734215,0.3813378513,-0.235213697,0.0053514317,0.1446432173,-0.5274692774,0.0251739435,-0.0338447504,0.0276431963,-0.5903801918,0.5948520899,0.1444467753,0.0653395206,-0.0022549035,0.0297405701,-0.2192582041,0.4458364248,-0.1469017118,-0.1265872568,-0.0777726173,-0.2012470663,-0.2906328738,0.0687579066,0.2800124288,-0.0004207446,0.0284658279,-0.0887903124,0.0508008599,-0.0373322666,0.1651456654,0.0966690406,0.045388177,0.6008269787,-0.2657188475,-0.3128871918,0.2576597929,0.314650923,0.2519697249,-0.2195741981,0.0636841506,0.4484404325,-0.0848196968,-0.09799999,0.1075752005,-0.0907739103,0.1093703434,0.5033141971,-0.106587328,0.0152194174,-0.0965926051,-0.0367571227,0.3132163286,0.1083215252,-0.442776829,0.267206043,-0.0694008842,0.1407467872,-0.3290606737,0.4435735047,-0.0094891647,-0.2512437105,-0.2555652261,-0.0928757712,-0.3300835192,0.1714189202,0.2314831913,-0.120286122,0.3378194869,0.2010096461,0.0420820192,0.3639495075,-0.4103559554,0.5006648898,0.4353063405,0.1364458799,-0.0873571709,0.3364314735,-0.0369656868,0.3378377259,-0.1917542517,-0.4197158515,-0.2669377029,0.4117030799,-0.1799109876,-0.1397609562,0.3466663063,-0.4194225371,0.1942957789,0.0694921762,0.896739006,0.1470733136,0.0028970109,-0.2585141063,0.016520109]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2055","title":"is there a way to override a dataset object saved with save_to_disk?","comments":"I tried this way, but when there is a mapping process to the dataset, it again uses a random cache name. atm, I am trying to use the following method by setting an exact cache file,\r\n\r\n```\r\n dataset_with_embedding =csv_dataset.map(\r\n partial(self.embed, ctx_encoder=ctx_encoder, ctx_tokenizer=self.context_tokenizer),\r\n batched=True,\r\n batch_size=1,\r\n features=new_features,\r\n cache_file_name=cache_arrow_path,\r\n load_from_cache_file=False\r\n )\r\n```\r\nSo here we set a cache_file_name , after this it uses the same file name when saving again and again. ","body":"At the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object? ","comment_length":69,"text":"is there a way to override a dataset object saved with save_to_disk?\nAt the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object? \nI tried this way, but when there is a mapping process to the dataset, it again uses a random cache name. atm, I am trying to use the following method by setting an exact cache file,\r\n\r\n```\r\n dataset_with_embedding =csv_dataset.map(\r\n partial(self.embed, ctx_encoder=ctx_encoder, ctx_tokenizer=self.context_tokenizer),\r\n batched=True,\r\n batch_size=1,\r\n features=new_features,\r\n cache_file_name=cache_arrow_path,\r\n load_from_cache_file=False\r\n )\r\n```\r\nSo here we set a cache_file_name , after this it uses the same file name when saving again and again. ","embeddings":[0.0187769011,-0.1548894942,-0.0146417003,-0.039853137,0.244917348,0.2703030109,0.1437665075,0.1278606057,0.1100359708,0.1771552861,0.3504517972,0.4532105923,-0.2038121372,-0.0362367816,0.2456676215,0.1528116912,0.4748806357,-0.0202535633,-0.0256069303,0.2713933587,-0.4494588673,0.2261940986,0.2889965773,-0.0424482822,-0.1577510387,-0.1020553112,0.0510712862,-0.1924010962,0.1730673611,-0.4009984732,0.287232846,0.172415778,0.0001321342,0.1486887336,-0.0001201857,-0.1788837165,-0.0300848149,-0.1175805777,-0.2467471063,0.1999171823,-0.2080583572,-0.3076746464,0.0557780489,-0.5268291235,0.2334907353,0.0689518228,0.2301573306,-0.3817866147,0.191548869,-0.1445045322,0.1007296965,-0.1398894042,0.0179280937,0.2429853827,0.0684334114,-0.0258235205,-0.2400593609,0.2939268649,0.025716627,0.3028870523,0.1099615246,0.0542864278,-0.1694210172,0.0480514728,0.685513854,-0.0258398112,-0.2212391496,-0.0490679108,0.2112247497,0.0301928502,0.6607089639,-0.4477230906,-0.1442882717,-0.0601028204,-0.0591602027,-0.4760395288,0.2344955057,-0.2521856427,0.091259107,0.1244170517,-0.1475167274,-0.567283988,-0.1388987005,0.0413065106,0.2410663813,-0.4322021306,-0.1471636593,0.2276435494,0.2035791129,-0.0040798206,0.4306220412,-0.3934549093,-0.3842329383,-0.203083083,-0.2061029226,-0.1462682486,-0.385091722,0.2905322909,0.0256242938,0.1989117563,0.4617836475,0.2264640033,-0.0000423613,0.2483998835,0.0427427962,0.286901921,0.0507781915,0.0147927627,0.0957549587,-0.2234209776,-0.2146575004,-0.1435916126,0.2676704824,-0.3017749786,0.6434220076,-0.1205687001,0.0917803124,-0.1733267009,0.1807958186,-0.0465304218,0.2298899144,0.2185389102,-0.1140381843,-0.0172941312,0.016586734,-0.0285297409,-0.1195977628,0.0099899285,0.1471544653,0.1509025991,0.0224689245,-0.1985741556,0.1876461506,0.426202327,0.1172284707,0.0619218536,0.0814464912,-0.1745312065,0.0754381046,-0.1534605473,0.2329119146,0.0419621468,0.4006009996,-0.1149874628,-0.1312373728,0.2729934156,0.0239582006,-0.2295759171,-0.1466408074,0.2271616608,-0.367708981,-0.3400831223,0.1010109335,0.023598386,-0.3579705656,-0.0778505132,-0.564493835,0.0533482321,0.0624166206,-0.22273013,0.2458759695,0.2875828743,-0.1862507313,-0.3175987601,-0.0898554027,0.2636732161,-0.3861024082,-0.0023469662,-0.0945786014,0.064221926,-0.3352030814,-0.0623731017,-0.1391376257,-0.2302371562,-0.3009421229,-0.0663430318,0.3127155304,-0.2570265234,-0.649944663,-0.0419114865,-0.0457674526,-0.0369239263,0.1332404613,0.3977469206,0.2084973156,-0.0175151583,-0.3062950373,0.0642192662,0.0359246843,-0.2491101027,-0.1534541994,-0.1586133689,-0.1160491779,-0.2702701986,-0.2579095364,0.381293416,-0.0460059047,-0.1806786209,0.0544846766,0.0745689347,0.2152922153,0.458814919,0.3959977925,-0.1752629578,-0.0587200224,-0.1027324572,-0.4525175691,0.059537679,0.4358718395,-0.4237766564,0.102919884,-0.2893728316,0.0303792041,0.0588048287,-0.1857581288,0.0930105746,-0.0494703762,0.3206633925,0.0682427287,-0.1817137301,-0.2254603654,0.3056722879,0.0819492266,0.0807876661,-0.1969593614,-0.0839454159,0.0476201624,-0.0904087797,-0.4644240141,-0.2012928277,0.1987289488,-0.1882765591,-0.1613456011,0.4910293519,-0.1910765916,0.2047098428,0.1569013745,0.4119258523,-0.0411734022,-0.3291231096,0.2022393793,0.0987962261,0.0716338828,0.0659280941,-0.1734007001,0.2870254219,0.0487030596,0.078516677,-0.0599838234,0.1326052397,0.225709796,-0.0373648442,-0.1995096058,-0.2379368097,-0.3404450715,0.1995523423,0.3960496485,-0.1004426554,-0.1189195514,0.1776395291,0.5097330213,0.0374561585,0.1199823096,-0.0382120237,-0.3550250232,-0.1304751933,-0.0980043411,0.606318891,0.3569963872,0.1172183454,0.0578324907,-0.2137328088,0.112681888,0.0051050931,0.2453827709,0.1002432629,0.2005971968,-0.0218010414,0.2279223502,-0.1744115651,-0.2117969543,0.1703075618,0.0876713768,-0.0393557921,0.0572534278,0.1778856814,-0.140629068,-0.0025293666,-0.3594079018,0.1417996138,-0.3370628357,-0.2788515389,-0.070038788,0.534375608,0.0857299492,0.1341638118,-0.0817061514,0.4854326546,-0.1819152981,-0.4802110791,-0.0635461807,0.0765655339,-0.1230189428,-0.097504057,0.0190388225,-0.4540122747,0.2778124213,0.1107966527,0.2113085836,-0.637880981,0.1011233702,-0.1067345962,0.3585079312,0.2471572906,-0.1479629427,0.1134592593,-0.2218858749,-0.0172922481,-0.0711410418,-0.0229312368,0.0792189687,-0.0204045158,0.035268683,0.1745040119,0.0540647805,0.1256746799,-0.2316399217,-0.1592323035,0.1551640332,-0.1161960959,-0.005929274,-0.1207597479,0.0031600527,-0.0589698367,0.1094887331,-0.4539576173,-0.2308218777,-0.6659761667,0.5767184496,-0.2546722591,-0.0310062543,0.1016660556,-0.0943055376,-0.0100531857,0.2011738271,-0.3397991955,-0.103417784,-0.1008762568,0.223428756,0.0948505253,-0.3411518931,0.2415471971,0.5115864277,-0.054048989,-0.0983425453,-0.5508167744,0.2543379068,0.2440887243,0.4389616549,0.3565298915,0.1832325608,0.042617321,0.4608428776,-0.2074300498,0.0254488885,0.4877781868,0.1006794423,0.3907102644,0.0514013171,-0.0589574352,-0.2013409436,-0.3123382628,-0.1426025182,-0.0268095452,-0.0115307504,-0.0156695154,0.2334087193,0.0504277982,-0.0298050418,-0.334670037,0.007738993,-0.2926947474,0.1200675741,0.0433672555,-0.1981489956,-0.4714189172,-0.2149313837,0.0872355774,0.391233325,0.1272730827,-0.082675375,-0.5049142241,-0.0117386598,-0.4805624187,0.4296471775,-0.2495910525,0.1166212782,0.0442131273,0.0584543273,-0.1080800071,-0.0846033767,0.8759832978,-0.1353387386,0.1624020934,-0.0896903425,0.1907726824,-0.0841150805,0.0878337994,0.4205208421,0.4367337823,-0.2517591715,0.2692509592,0.257145375,-0.0806375891,-0.1169363558,0.3535146713,-0.5934962034,-0.1206401959,0.2151967734,-0.0826923996,-0.0579545833,-0.3540091217,0.1801159531,0.0230675172,0.1058657616,-0.0819563344,-0.4227649271,-0.034343604,0.0452647805,-0.2890200913,0.4680197835,-0.1545275003,0.0069528837,-0.0850611925,0.2909590304,0.3853334486,0.28279531,-0.3384393752,-0.3447610736,-0.0637485981,-0.1411978006,0.1837272048,0.1123549864,-0.0605233237,0.242113471,0.1979688257,-0.1880516559,-0.0057778987,0.1027549207,-0.092440322,0.3334353864,-0.1502129734,-0.4014714956,0.374170512,0.2887127101,-0.0066097332,0.2434092164,0.1107720584,-0.4094184339,0.4683394432,0.0238110982,0.6857135892,0.0328948162,0.3769575059,-0.2315592021,-0.3167063594,0.1525568217,-0.0891744122,0.0909287483,-0.1217383593,-0.2212104499,-0.0291227847,0.0953065529,0.3272645175,0.4822123051,-0.05027356,0.3661999106,-0.1519141197,0.0964507461,-0.2594840229,0.0771626309,-0.1766588092,-0.1769029349,0.1472624987,0.0571525097,0.0117421271,-0.1370297521,0.1000077128,-0.0760274604,0.0786839798,-0.0002040824,-0.1984481663,-0.2250747234,-0.1740598679,-0.2131843418,-0.2469963878,-0.4627684653,0.0121716345,0.2771262228,0.3272089064,0.0756052807,-0.0139241517,0.1802669019,0.22201702,0.0546761155,-0.2464651167,-0.0767864957,0.1695948243,0.0448298343,0.0436455123,0.0350598842,0.072672382,-0.4616362751,0.3340903521,-0.1473330259,0.1642683297,-0.0476404838,-0.540594101,-0.1617492884,-0.2307675928,-0.3211683035,0.0524357595,-0.1010854766,-0.054256063,0.3596061766,-0.237888068,-0.0876786411,0.1408545524,0.3142167926,-0.115786165,-0.1948649138,0.6367977858,-0.4938899577,0.0482371785,-0.1050162613,0.2423691452,0.1008897945,-0.0780598745,0.1757743359,-0.1397124976,0.114077881,-0.0665978119,-0.1368860751,0.1169704348,0.3202656209,-0.2614651322,-0.1922057569,-0.4788091183,-0.105626069,0.2239226699,0.3366801143,-0.0704574883,-0.4245383143,-0.1424576044,-0.047673434,-0.1446592659,0.0385076068,0.0199610833,-0.0327929333,0.2165947855,0.4628361762,0.0585293286,-0.0902767107,-0.0717538744,-0.0228658523,-0.0630903915,-0.1342514604,-0.2262952477,0.188858673,-0.1266929954,-0.2323740721,-0.1736101657,-0.3062692583,0.2362321019,-0.0703763887,-0.2308465987,0.1891656369,0.1071476415,0.2860004008,-0.0335355029,0.111500144,0.0147682838,-0.1662110537,0.2184540182,0.3839322031,0.179078117,0.2159478962,-0.0518779904,-0.0688613132,-0.1427633166,-0.1199695617,-0.051873941,-0.0187679529,-0.1914045215,-0.1052303463,-0.072073929,0.4625981152,0.1093427315,0.0106477644,-0.2217101753,-0.1483030468,0.3819298148,0.0556833521,0.0622013696,-0.3556944728,0.4837986529,-0.3391482234,0.0261016749,0.1197485924,0.1509521604,0.0948638618,-0.2867463529,-0.1076825038,0.3201233447,-0.4162610769,0.2764832675,0.1164498925,0.1428447813,0.1220634207,0.4623321891,0.3871821463,0.3841492236,0.7751978636,-0.0112574203,0.0066853841,0.4595455527,-0.0849630311,-0.0102804722,-0.1649308354,0.1495009363,-0.1043557152,-0.0656151548,0.545907259,0.4420746267,0.0076582297,0.0097148074,-0.3250493705,0.0948142931,-0.0091078812,0.0600405857,-0.3881416619,0.2045374662,0.0207889602,0.175720647,0.2321554273,-0.2062299401,-0.3199569583,0.3652498722,-0.148203969,0.0497946702,0.2316642851,0.2691744566,-0.1120106503,0.1832073033,-0.0372555517,-0.1261143982,-0.3010330796,-0.0776104257,0.0920063257,-0.0261494163,-0.1929473728,-0.1082378104,0.0387248173,0.1681859642,-0.0358160809,-0.0256359968,-0.1180936247,-0.3194728196,0.0108244969,-0.3204699755,0.1494805515,0.0559626333,0.0598623678,0.1930320263,0.2173729092,0.4506444633,-0.2284136415,-0.1394959688,0.2904867828,-0.4147695899,-0.0107200295,0.1111338735,-0.0737023279,-0.3833247423,0.7265056372,0.0043065189,0.0901440158,0.0839709193,0.0455225594,-0.1325954348,0.3578693867,-0.1725378633,-0.042536173,-0.265402317,-0.0307880566,-0.3812366426,0.1356670111,0.3085462749,0.0837810934,-0.1038390696,-0.070707269,-0.0179448631,-0.0395766869,0.1623757184,-0.0570540838,0.1817876697,0.5798847079,-0.2913216949,-0.2843711376,0.2068998367,0.3773007691,0.2617412508,-0.4093430042,-0.1080823317,0.3444977701,-0.0864498317,0.0478462838,-0.0441098586,-0.0265317839,0.051597178,0.5468366146,0.0009369376,0.0241113678,-0.1304380149,0.0329682678,0.281388253,0.1347506493,-0.384239912,0.0073212627,0.0967051312,0.1371987462,-0.4002992213,0.3226004243,-0.1326918453,-0.0920270309,-0.0128560672,-0.1971436441,-0.1830771118,0.2361608297,0.2111166418,-0.1121970117,0.3844278157,0.40358904,-0.007087444,0.4863263071,-0.3266405165,0.4129480124,0.4259540141,-0.0659817085,-0.0850185975,0.2840052843,0.1194752678,0.2796078026,-0.2027405947,-0.241662994,-0.162198469,0.3066489398,-0.2225730121,-0.1236333251,0.2972763181,-0.439438194,0.2138994634,0.0830428153,0.9734278321,0.2376054078,0.0048444774,-0.2613398731,-0.1835116446]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2055","title":"is there a way to override a dataset object saved with save_to_disk?","comments":"I'm not sure I understand your issue, can you elaborate ?\r\n\r\n`cache_file_name` is indeed an argument you can set to specify the cache file that will be used for the processed dataset. By default the file is named with something like `cache-.arrow` where the fingerprint is a hash.","body":"At the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object? ","comment_length":48,"text":"is there a way to override a dataset object saved with save_to_disk?\nAt the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object? \nI'm not sure I understand your issue, can you elaborate ?\r\n\r\n`cache_file_name` is indeed an argument you can set to specify the cache file that will be used for the processed dataset. By default the file is named with something like `cache-.arrow` where the fingerprint is a hash.","embeddings":[0.0058808769,-0.0709295273,-0.045395419,0.0091859847,0.2773501575,0.2527240515,0.2612765133,0.2426063865,-0.0535695441,0.1069626808,0.3020359278,0.4447039962,-0.1616931856,-0.1279573888,0.205916509,0.0063203103,0.3146958649,-0.1188129112,-0.0984395593,0.1764316857,-0.3626472354,0.2778185904,0.2146423757,0.0628884211,-0.3578145802,-0.1051508114,0.1591539532,0.0235732347,0.0712339804,-0.4981004,0.2488244027,0.2438838482,0.0018548085,0.044608634,-0.000113177,-0.1921939403,0.175977543,-0.1021563783,-0.3182258606,0.1989146024,-0.2904453278,-0.2070807666,-0.0157763828,-0.5379847884,0.3022142351,0.0479414761,0.1390975118,-0.3058054745,0.1568408608,-0.0851744339,0.1883069426,-0.1913198382,0.0957808569,0.1804773062,0.0681001022,0.0802315101,-0.2438443601,0.2512896359,0.021615129,0.2512300014,0.0488665588,0.1368550509,-0.0925930738,-0.0484558679,0.5817062855,-0.0571092926,-0.1138632149,0.0380058773,0.3123865426,0.1523032337,0.7274105549,-0.4069977999,-0.0880730152,-0.0935112536,-0.1064908206,-0.2785566449,0.2970902324,-0.2870459855,0.031411279,0.1581908315,-0.1443065107,-0.5655369759,-0.1578780562,0.0454114228,0.2925548553,-0.3315795064,-0.184719786,0.2547592521,0.2245904207,0.0062038698,0.4588221014,-0.278788954,-0.4055844545,-0.2095421851,-0.22318694,-0.0995256305,-0.3349794149,0.4045699835,0.0675696954,0.2259330451,0.4345136881,0.3098502457,0.0374989882,0.1940719187,0.1098939627,0.2558328211,0.0914492607,-0.0145353973,0.075344786,-0.2580299079,-0.3399423063,-0.1969913244,0.210166201,-0.3805344701,0.7020931244,-0.1838436276,0.1940517277,-0.3050473928,0.1908460259,-0.0842874125,0.2190545946,0.0675936565,0.0781935081,0.0128200417,0.0667795539,0.0648913756,-0.0888025984,0.0561332963,0.0615240596,0.0973784253,-0.0642034784,-0.2585119903,0.1058640629,0.3715211749,0.0700621754,0.1134168282,0.0457305461,-0.1482766122,0.1364432275,-0.0534764677,0.3880754113,-0.0520716794,0.4737306237,-0.061555054,-0.1686994433,0.255069375,0.00813556,-0.156441927,-0.1375379711,0.3093259931,-0.3865028322,-0.4239960909,0.0999027789,0.0781676024,-0.3999314904,-0.0620709658,-0.525015831,-0.038405057,0.1165150031,-0.1437495202,0.2381729037,0.2040822953,-0.1525623947,-0.3873444498,-0.0526700206,0.2291747183,-0.417843312,0.0511242449,-0.2441598475,-0.0211358797,-0.2414739728,-0.1609629095,-0.2630755603,-0.1054811776,-0.3466656506,0.0644727796,0.3076820076,-0.3473539054,-0.6475951076,-0.059336599,-0.06429708,0.0283018313,0.0475253128,0.3307835162,0.0566511154,-0.1022358686,-0.4012673199,0.101296626,-0.0200254451,-0.3065453768,-0.2780852318,-0.0080476673,-0.105739519,-0.1461751908,-0.2202514559,0.3657616973,-0.0380264111,-0.0284255017,0.0723902434,0.2011933625,0.2046846449,0.4713914096,0.484555006,-0.1623984873,0.0280855689,-0.0522791296,-0.3321339488,0.1471669525,0.3700075746,-0.3939211965,0.0009514749,-0.2499645799,0.0163442474,0.0734223053,-0.1349820942,0.154103741,0.0280377977,0.3183521032,0.0730620697,-0.2711367011,-0.1399842352,0.4156740904,-0.0353840217,-0.035253454,-0.1052668244,-0.0459287241,0.0036338344,-0.0581223965,-0.4752197862,-0.2656543553,0.1838871688,-0.1295546144,-0.2298534065,0.5186124444,-0.0912202969,0.0534739867,0.2685379684,0.4209367931,-0.1508518457,-0.442193985,0.2504714727,-0.0025525822,0.0273712426,-0.0055460529,-0.1252529472,0.1777490079,-0.0786021203,0.058030989,-0.1963467598,0.138000682,0.2874149978,-0.0780566484,-0.1735179573,-0.2393267155,-0.3135153651,0.0750133917,0.4613127708,-0.0820760727,-0.2110939473,0.3025094867,0.5622271299,0.058873374,0.1013874412,0.0364010483,-0.264415741,-0.1285411865,-0.1024732888,0.6723667383,0.5421388149,0.1920127422,0.1405190974,-0.2133501023,0.0958545655,-0.0383532718,0.1324896514,0.1543866247,0.2147182673,-0.1020127833,0.1443184167,-0.1821781248,-0.1338694543,0.1596543789,0.1118857041,0.0110336011,-0.0063969037,0.1888252646,-0.1927698106,-0.0278133117,-0.2950424552,-0.0319356658,-0.4077279568,-0.3630105257,0.0669340342,0.3269263506,0.0898088291,0.1490609795,-0.1156611219,0.5042241216,-0.1734948903,-0.4204646051,-0.1750683337,-0.0311101675,-0.2325831205,-0.022364879,0.0742438957,-0.2729974091,0.2291760296,-0.0166326165,0.2555164397,-0.5720787644,0.0293579642,-0.0567883924,0.2725668252,0.3326309323,-0.0666651949,0.1852223277,-0.228722766,0.0670945495,-0.0157873258,-0.0791906565,0.0158427898,-0.0162483361,0.1165607199,0.0800131038,0.0378385745,0.2091630548,-0.1519974172,-0.2444863766,0.1520084143,-0.1092127264,0.0612642504,-0.0512665547,0.015881991,-0.0117451353,0.0459355004,-0.36172539,-0.3731367886,-0.8067584038,0.5466674566,-0.3449462056,-0.0970713198,0.1465294808,0.0542480536,0.1421537548,0.2120144069,-0.3342684209,-0.2800634205,-0.0528380722,0.2175366282,-0.0190757308,-0.2813302875,0.2258331925,0.4095545709,-0.0971838832,-0.248385787,-0.5289547443,0.3987857103,0.1717531979,0.5748049617,0.2610895336,0.0920389742,0.0547520407,0.505979538,-0.2828648388,0.0113985296,0.4043925107,0.1901494116,0.3782545328,0.0162271634,-0.1309809387,-0.1413495094,-0.2167472243,-0.0760765672,-0.0497826859,0.087872453,0.0190347042,0.2673096955,0.0565378331,-0.1562475562,-0.3716948926,0.0874496996,-0.2596992552,0.115963541,0.0422529206,-0.1429021508,-0.5049213171,-0.144814074,0.1870449185,0.3815188706,0.1140297949,-0.0851540118,-0.546657145,0.1567207575,-0.357234478,0.4369426966,-0.2001793534,0.1197870225,-0.1039198786,-0.0498086885,-0.0212270245,-0.0503555052,0.8804520369,-0.1738556474,0.158569783,0.0629766434,0.1106299385,-0.1314806044,0.0982224792,0.4506983161,0.4493229389,-0.1458299905,0.1064607278,0.0941272676,-0.2235233188,-0.173330918,0.336966306,-0.5265506506,-0.2175543904,0.152284503,-0.1450504065,-0.0076508643,-0.2675729394,0.2054926306,0.0131744053,0.0813187286,0.0208212398,-0.4381811321,-0.1331949681,-0.0034737843,-0.195069313,0.5359504223,-0.228697747,0.0376808271,-0.0225112289,0.2528629005,0.389048785,0.2947440743,-0.3565094173,-0.20212318,-0.0354507305,-0.2481733412,0.0178414434,0.2599074543,-0.0762654468,0.1664107591,0.2039911598,-0.1703591049,-0.0140159074,0.1486971825,-0.1441234946,0.2527445853,-0.1916001141,-0.3499116302,0.3759484291,0.2600643337,-0.0648220107,0.3666514158,0.1746676117,-0.3780328631,0.4197975397,0.0901346803,0.9501265883,-0.0370361432,0.2972920537,-0.3037034273,-0.3726537526,0.0298512895,-0.0703395531,-0.0271230228,-0.1310967207,-0.2019097656,0.0092435637,0.0699818879,0.2341230661,0.5468813181,-0.0399582461,0.0949292332,-0.2273949981,-0.1861843318,-0.2574064434,0.2381446064,-0.1397060156,-0.1130989641,0.0804449841,0.1651393473,-0.0612482987,-0.1538336128,0.0608010292,-0.2177507132,0.0214239527,-0.0302506629,-0.1726944894,-0.1897626072,-0.1683496535,-0.2046888173,-0.070413202,-0.4095523357,0.0185398627,0.2018807083,0.4421583116,-0.0273573529,-0.0565068834,0.3041874468,0.1745548099,0.0735192299,-0.2249201387,-0.1495260596,0.258333683,-0.0437389687,-0.0192859117,0.0022182832,0.0770842358,-0.6202127337,0.2067082077,-0.0997246429,0.130041942,-0.0917503238,-0.4944214523,-0.1646654457,-0.2461181134,-0.3005343676,0.1419224888,0.2287475616,-0.1075784564,0.414141655,-0.1770246774,-0.0451599211,0.1554747224,0.2797190845,-0.2176186293,-0.1377959698,0.5094897747,-0.3504292071,0.0148184076,-0.1385893524,0.06893415,0.0487168245,-0.1485421807,0.1494805068,-0.1786648929,0.1388297379,-0.0322537869,-0.2361465991,0.1513173282,0.3105769753,-0.2591745853,-0.1748177856,-0.5221589804,-0.0817385167,0.0659392178,0.3063133061,-0.0762282237,-0.4527073205,-0.0630410165,0.0887997597,-0.253058672,0.0121619245,-0.0453254804,0.0271609966,0.1892688721,0.2453793734,0.1555829495,-0.1864813119,-0.0187083576,0.0017403637,-0.1725643128,-0.1992296576,-0.2664844394,0.1718348861,-0.0936260745,-0.2779681981,-0.1310477555,-0.3101246953,0.2115918696,-0.0475881398,-0.1959195435,0.1781339347,0.0373373702,0.2285753936,0.0721873641,0.0019591504,0.0486024655,-0.1798073351,0.1992194653,0.3495193124,0.1664069295,0.125814721,-0.0140221957,-0.066203095,-0.0015538601,-0.1024212465,-0.1055414379,-0.0304204486,-0.1901530921,-0.0638885796,-0.037877772,0.4332277179,0.1673293561,0.0405725539,-0.2100210935,-0.1713738441,0.3683485091,0.1341478974,0.1061888784,-0.4036818743,0.4270189106,-0.2677614689,-0.0381476879,0.0870680586,0.1465191692,0.1951242387,-0.1684445143,-0.0055486229,0.2567140758,-0.4714050293,0.1689612269,0.052709043,-0.0133025432,0.0811087787,0.5374557972,0.3864861429,0.3091652095,0.7966071963,0.014929465,0.1586894095,0.4057993889,-0.0334069394,-0.0127482396,-0.1542294472,0.0290581584,-0.0464884341,-0.0431821942,0.4586323798,0.2863246799,0.1327946782,0.1238334626,-0.3554238677,0.1732826531,-0.0289881155,0.0208512768,-0.3102228045,0.0797470137,0.02053128,0.1260371357,0.1096810177,-0.1941624284,-0.2527129352,0.4634740651,-0.13702178,0.0431457497,0.1683923602,0.1710461378,-0.0685694441,0.128636837,-0.039517384,-0.0998140797,-0.2691250443,-0.1345331222,0.2595928609,0.0021043469,-0.1687575579,-0.0542042218,0.0429620035,0.0815422013,-0.0411023907,-0.0792398453,-0.2262188345,-0.2022945136,0.0682455376,-0.3262187839,0.2005546838,0.1387023926,0.0121766552,0.13348943,0.1436517388,0.3790521324,-0.1472217888,-0.1271031499,0.3025744855,-0.363319397,-0.121689491,0.1171797514,-0.0705770478,-0.3746970892,0.7522788644,0.0425844789,0.0491754189,-0.0540876612,0.067596741,-0.17592071,0.4489005804,-0.0187774654,-0.0136209847,-0.2923688591,-0.1183670834,-0.3946759403,0.1372793764,0.1089551374,0.0853914768,0.0672109053,-0.0531308204,0.0318031348,0.0138657773,0.2937546968,0.0810594782,0.1560963541,0.5610340834,-0.3906924725,-0.1711736023,0.3270383179,0.2972218692,0.2671146989,-0.3106239438,-0.117632255,0.4650087953,0.021207273,0.0037845492,-0.099389486,0.0067931055,0.2326111048,0.5678701401,-0.018731324,0.0685637444,-0.0676010475,0.0152100194,0.1250105798,0.100457035,-0.3279696405,-0.0648333132,0.0788087025,0.1504220217,-0.3691146076,0.3105677366,-0.2134571373,0.0019574922,-0.123385556,-0.0738363639,-0.20916228,0.217723161,0.1516338438,-0.0270657111,0.4272915423,0.4429099262,0.0462191515,0.4174529314,-0.4453116357,0.3533111811,0.5391594768,-0.0078870635,-0.1656595916,0.2888102233,0.1509992331,0.2713836432,-0.2632124424,-0.347270757,-0.2103031874,0.2743267715,-0.1712188721,-0.0767598301,0.2778632343,-0.3333678842,0.3009935617,0.1058188528,0.8167826533,0.171621263,0.088463299,-0.3020500243,-0.0966657773]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2055","title":"is there a way to override a dataset object saved with save_to_disk?","comments":"Let's say I am updating a set of embedding in a dataset that is around 40GB inside a training loop every 500 steps (Ex: calculating the embeddings in updated ctx_encoder in RAG and saving it to the passage path). So when we use **dataset_object.save_to_disk('passage_path_directory')** it will save the new dataset object every time with a random file name, especially when we do some transformations to dataset objects such as map or shards. This way, we keep collecting unwanted files that will eventually eat up all the disk space. \r\n\r\nBut if we can save the dataset object every time by a single name like **data_shard_1.arrow**, it will automatically remove the previous file and save the new one in the same directory. I found the above-mentioned code snippet useful to complete this task. \r\n\r\nIs this clear?","body":"At the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object? ","comment_length":134,"text":"is there a way to override a dataset object saved with save_to_disk?\nAt the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object? \nLet's say I am updating a set of embedding in a dataset that is around 40GB inside a training loop every 500 steps (Ex: calculating the embeddings in updated ctx_encoder in RAG and saving it to the passage path). So when we use **dataset_object.save_to_disk('passage_path_directory')** it will save the new dataset object every time with a random file name, especially when we do some transformations to dataset objects such as map or shards. This way, we keep collecting unwanted files that will eventually eat up all the disk space. \r\n\r\nBut if we can save the dataset object every time by a single name like **data_shard_1.arrow**, it will automatically remove the previous file and save the new one in the same directory. I found the above-mentioned code snippet useful to complete this task. \r\n\r\nIs this clear?","embeddings":[0.0457346216,-0.0365211032,-0.0065601123,-0.0663145259,0.2652109563,0.1467555016,0.0534220524,0.1336881816,-0.1323261559,0.3330439031,0.3283967674,0.4801915288,-0.3888702691,0.0933786184,0.2319840342,0.0744830072,0.373470217,-0.0670178831,-0.1254188418,0.1810304821,-0.4179126322,0.0441506915,0.2109159231,0.0159829482,-0.0820500329,-0.0264369417,0.0004691452,-0.2508091927,0.0576554537,-0.4308080077,0.2283994853,0.2565983236,0.1199135557,0.1386187524,-0.0001212558,-0.0695586503,0.0095337871,-0.0692437291,-0.416901648,0.0994594991,-0.0609362386,-0.4046192169,-0.0431455038,-0.4321476221,0.3878163993,-0.2496526539,0.2778483927,-0.1676830202,0.1013099253,-0.1304500103,0.0757567957,-0.0730952471,0.1910556108,0.2092887312,0.1304677725,0.2837990224,-0.1073035821,0.2409989238,0.0536172539,0.2485469431,0.0044527454,0.104701452,-0.0398981124,-0.0903648436,0.7107020617,0.0397537798,-0.0830013826,-0.0689411908,0.063255161,0.1232702956,0.5087640882,-0.533945322,-0.1315351725,-0.1749764681,0.1265869588,-0.3090947866,0.2692341506,-0.2114600688,-0.0695016161,0.0209068414,-0.0852169693,-0.6508269906,-0.198544085,0.2556485534,0.2591593564,-0.5391643643,-0.0664123595,0.2288216203,0.316231966,0.0076818229,0.3161955178,-0.3927267194,-0.2238267362,-0.3384256661,-0.1280718744,-0.1810763776,-0.3935975134,0.2291760147,0.0554537289,0.3497772813,0.4361371398,0.1356812268,-0.0677442178,0.2203283459,0.1018314883,0.0233845469,-0.0405252948,0.1268221289,0.0779919401,-0.0195702258,-0.169791624,-0.0301521216,0.2458382398,-0.3112148345,0.3921233118,-0.2665874958,0.257268995,0.058126498,0.0073820548,-0.0774414167,-0.0419784375,0.1095786989,-0.0056771566,-0.1309606433,0.235304296,-0.1221491098,-0.1005398855,0.1382056624,0.1775675714,0.1451976746,0.0803287625,-0.1761173606,0.104832679,0.4112702906,0.0389890447,0.172429949,-0.0452261455,-0.2201778442,0.0412365906,-0.0717496648,0.1835631281,0.0025440357,0.4037235081,-0.098399803,-0.1946253181,0.1945430189,-0.1654605716,-0.1861855239,-0.2288786173,0.2491870373,-0.239801228,-0.279717207,-0.0215783864,0.0262750406,-0.2249813229,-0.1048920155,-0.5177105069,0.1623329073,-0.1310679317,-0.2600252628,0.2126580477,0.2772542238,-0.2532300949,-0.2626474202,-0.000172325,0.1168229133,-0.338301003,-0.0473852791,-0.1328247637,-0.0268523376,-0.3620640635,-0.1325851083,-0.2609158754,-0.0106197828,-0.295156002,-0.0211456101,0.3409298062,-0.1320462674,-0.5908213258,-0.2278280854,-0.1508217603,-0.005291455,-0.0067438367,0.5044074059,0.1441428661,-0.0576039106,-0.5045726895,0.053438548,0.0619736947,-0.350050658,-0.0606923327,-0.1441928595,-0.197633028,-0.1717526466,-0.2981700301,0.2197761685,-0.0556042753,0.0251110438,0.2167973518,0.0732120574,0.2401830405,0.4485553801,0.4341867566,-0.1154959351,-0.1688873321,-0.1425728649,-0.3905547261,-0.0319552533,0.5924943686,-0.4588957131,0.1132578477,-0.2073935866,0.174704805,0.123046875,-0.1968377233,0.2352232784,-0.0548228249,0.3122776449,-0.0037066585,-0.1566853523,-0.3619379103,0.0674391091,-0.172998786,0.1804399937,-0.1348822415,0.1464636177,0.1023183763,-0.0458523631,-0.3633918464,-0.1533870846,-0.1064044684,-0.096509032,-0.0751480162,0.4044415355,-0.246415019,0.1191925928,0.1726580709,0.4314657152,-0.0619406402,-0.3676759899,0.2172438353,0.2357064933,0.0893714577,0.1228102148,-0.3586910963,0.1210450679,-0.0270180535,0.1744958609,-0.061239928,0.1384723485,0.0513607152,-0.0509707928,-0.3274543285,-0.2260819077,-0.3494554758,0.184198454,0.4219741821,-0.0683638528,-0.1296262443,0.1065360755,0.6030349731,-0.101148963,0.0626432374,0.1921167225,-0.2957580686,-0.0801921189,-0.2568456233,0.8007151484,0.3792867064,0.0745852366,0.1722947806,-0.2163551897,0.0766077787,-0.0620703399,0.1834325045,0.3528274894,0.2313379943,0.0689827651,0.2171487808,-0.0842011049,-0.2327295542,0.0229159687,0.0310181547,-0.004507483,-0.0311258566,0.0636650845,-0.0542579889,0.0626263544,-0.4437845647,0.0974419639,-0.2583097219,-0.1972875297,0.0239639804,0.3298319578,-0.0633206964,0.180943951,0.2926631868,0.4197939634,-0.264410466,-0.3293955326,0.0235056039,0.2038994282,-0.1538181454,-0.1272312403,0.1778415293,-0.3315117061,0.3227912188,0.134487927,0.1641263515,-0.7020509839,0.0724505633,-0.1511764824,0.2363016009,0.1226458549,-0.0113487924,0.0670083091,-0.2285551131,0.068781279,-0.0212822631,0.020753067,0.0384541973,-0.0363367721,0.0566614121,0.2914115191,0.1204237416,0.2116761059,-0.1690319031,-0.1292986423,0.189424634,-0.0924684107,-0.0127827143,-0.1523361206,0.2509376407,0.0716323406,-0.0205147211,-0.3927164078,0.0094660949,-0.6643105149,0.3599241078,-0.0630418807,0.0037833641,-0.0244873203,-0.0377796106,-0.0436001047,0.2727342248,-0.3989458382,-0.2452392876,-0.039836105,0.1406462044,-0.076299116,-0.1881086528,0.1733016074,0.4887489676,-0.020689059,0.0275857393,-0.4749938846,0.2654614449,0.1800307184,0.3496613503,0.2071923167,0.1719932705,0.1153355539,0.3271223307,-0.0671866015,-0.1364271194,0.4031299353,0.1584939659,0.3615847528,0.0697800592,-0.0579105429,-0.1681996137,-0.2103423476,-0.1022322252,-0.0086561022,-0.0691740438,0.2302610129,0.3729231656,-0.0315751545,0.0750683919,-0.3564731479,0.1301707923,-0.1990267187,0.0818053558,-0.0245358068,-0.1395408958,-0.2211086452,-0.1581019163,0.0965667069,0.3935674429,0.2068521976,-0.0604201891,-0.6695663929,-0.1548299044,-0.350404501,0.2457050979,-0.3112828434,-0.0530458055,0.1481726468,0.0319638476,0.0137646096,-0.0669058412,0.646992147,0.0000054415,0.0835074931,-0.19046399,0.1843107194,-0.1899842322,0.0944081098,0.3197165728,0.3267342448,-0.3434049785,0.4133321345,0.108060509,-0.2338376343,0.090972513,0.4599337876,-0.5098069906,-0.0579445101,0.2214580327,-0.2996602654,-0.1480363309,-0.3749023378,0.0997122899,0.0366373919,0.2042602152,-0.2132804394,-0.3660641313,-0.1237860322,0.1425287724,-0.3959293962,0.4069080949,-0.065429166,-0.0009560375,-0.0664051995,0.3982369602,0.6073397994,0.2417913526,-0.4718919098,-0.3955793083,-0.1128654554,-0.1703763902,0.2326044291,0.2393698543,0.0122222463,0.1966258287,0.1238601059,-0.133985281,-0.2612435818,0.1820873618,0.1194561347,0.3022388816,-0.1734209955,-0.3173740506,0.4509798884,0.2749408185,0.0477714837,0.1978363544,0.353849262,-0.3776675463,0.3673600256,0.1181275696,0.666908443,-0.018206533,0.4968538582,-0.0985921174,-0.3663337827,-0.0350099504,0.0183201805,0.0802297145,-0.0711577311,-0.1028734744,-0.0167901069,0.0819389597,0.2323680073,0.518414557,-0.0479845628,0.21587874,-0.1498367637,0.0852448791,-0.3729451597,0.2046392858,-0.1910217702,-0.1594382077,0.041063033,0.0237622038,0.0888336226,-0.1331807077,0.149634406,-0.0856838971,0.0917766392,-0.0596954338,-0.1656382531,-0.407268852,-0.1481129676,-0.2073445618,-0.1955056489,-0.5290107131,0.0385956317,0.4284878075,0.1972430944,0.0606711395,0.0835260153,0.2208250463,0.0484206975,0.1239434928,-0.1668683738,-0.1823074222,0.1879824698,0.075245738,0.0273724701,0.0250488427,0.1128558069,-0.4212214947,0.0675832406,-0.1193266511,0.2632932961,-0.0307076816,-0.336596489,-0.1666620523,-0.3178779185,-0.4264823496,0.0370022133,-0.0681352466,-0.00918906,0.415502578,-0.197499603,-0.1237899214,0.1959585249,0.1716915071,-0.1490484327,0.0487342402,0.7534675598,-0.4408060014,0.1156603843,-0.1474776715,0.4632131159,0.25515306,-0.1543340534,0.3137852252,-0.1544508934,0.2359162569,-0.0555396266,-0.1251839548,0.0500303991,0.1026498824,-0.3332237303,-0.0957837254,-0.3727158308,0.0459640399,0.3631151021,0.4548832774,-0.1544362605,-0.3133370876,-0.1745501012,-0.0623906925,-0.1290360987,0.1928516626,0.1115650013,-0.0337864384,0.0892683491,0.5190905929,0.0927497819,-0.0439772047,-0.0874967724,-0.0811779052,0.0490686744,-0.0677700043,-0.2429113984,0.2171798944,-0.1623784006,-0.0961467177,-0.126495108,-0.2579058409,0.1673526913,-0.1099618003,-0.3237514794,0.1433318704,0.1516302973,0.1925877184,0.0393255353,0.1373102069,0.2304431051,-0.1271109879,0.264839679,0.3415459692,0.1894309372,0.2597721815,0.1179965287,-0.1669148952,-0.2726455331,0.0306753982,0.1163638607,-0.1207784638,-0.2451503873,0.0025416866,-0.1187879071,0.3289445341,0.1872168928,0.281493783,-0.293967396,-0.2809571028,0.3967543542,0.0327843204,0.0694624633,-0.3122707009,0.4320258498,-0.2886408567,-0.0244981237,0.0649065599,-0.1104807779,0.1635616571,-0.6306898594,-0.1315702051,-0.0050382009,-0.1757171154,0.1836410016,0.3621574044,0.0351954997,-0.0113811921,0.2998208702,0.3326317668,0.4769551754,0.7569009066,-0.1946564764,0.3213788271,0.3584157228,-0.1121045724,0.1665575802,-0.1336346865,0.0062579075,0.019079661,0.0839050114,0.3280992806,0.4132535756,0.0703577921,0.167775616,-0.3083192408,0.077165097,-0.0210549254,0.1172891185,-0.4065390825,0.0521078035,-0.0315603316,0.1487139463,0.2962121069,-0.186200276,-0.4566995203,0.3414573371,-0.1709360927,0.0925915986,0.2572569549,0.2887911499,-0.1281081587,-0.0105041536,-0.0594284311,-0.2913373709,-0.1945294589,0.0401326828,0.1520558596,-0.0660124943,-0.2403428257,-0.1592816561,0.0114735588,0.1467015892,-0.1973903328,0.0433328301,-0.1993125677,-0.3347667754,0.0072897202,-0.1861634701,0.3576769531,0.006786298,0.1036223322,0.0400266536,0.2810911834,0.4203669131,-0.4444261193,0.0576521233,0.1505892575,-0.3244647086,-0.0227694605,0.0239829887,-0.0229800921,-0.4238347411,0.7393186092,0.0473704375,0.0094849318,0.1237339973,0.0314199477,-0.2672820687,0.3729068637,-0.0912545025,-0.2442427576,-0.2016926855,-0.1476555765,-0.4075842202,0.0889035016,0.1213942245,-0.0014501472,-0.0897428915,-0.0808511302,0.074439384,-0.1016436294,0.2257521152,-0.1425706744,0.0313429199,0.5929239392,-0.2693684399,-0.2329321951,0.1044757217,0.4249645472,0.2707130015,-0.3766430318,-0.0587717071,0.2983462512,-0.0946875662,0.0555637516,0.1400135159,-0.1116012409,0.0986749679,0.6502913237,-0.1444051564,-0.1245682389,-0.1674843878,-0.146026969,0.1860037893,0.2573998868,-0.4718904495,0.245792523,-0.045047529,0.2827143371,-0.4167078137,0.5806436539,-0.0772674903,-0.3462807238,-0.1111065224,-0.1564860493,-0.1729536802,0.2177678347,0.3041037917,-0.0099668857,0.3712159991,0.3615013063,0.0517330915,0.395437181,-0.414478302,0.3705530167,0.509077549,0.0948972702,-0.0532972589,0.2723087072,-0.0018008735,0.2699899971,-0.124106288,-0.489656955,-0.2182664275,0.347981602,-0.1333281994,-0.1373224556,0.2824610174,-0.2615499794,0.1249848679,0.0714559257,0.8420636654,0.1822080165,-0.118595086,-0.1620347947,-0.2001540661]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2054","title":"Could not find file for ZEST dataset","comments":"The zest dataset url was changed (allenai\/zest#3) and #2057 should resolve this.","body":"I am trying to use zest dataset from Allen AI using below code in colab,\r\n```\r\n!pip install -q datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"zest\")\r\n```\r\n\r\nI am getting the following error,\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset zest\/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to \/root\/.cache\/huggingface\/datasets\/zest\/default\/0.0.0\/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"zest\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 612 )\r\n 613 elif response is not None and response.status_code == 404:\r\n--> 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/zest\/zest.zip\r\n```","comment_length":12,"text":"Could not find file for ZEST dataset\nI am trying to use zest dataset from Allen AI using below code in colab,\r\n```\r\n!pip install -q datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"zest\")\r\n```\r\n\r\nI am getting the following error,\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset zest\/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to \/root\/.cache\/huggingface\/datasets\/zest\/default\/0.0.0\/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"zest\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 612 )\r\n 613 elif response is not None and response.status_code == 404:\r\n--> 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/zest\/zest.zip\r\n```\nThe zest dataset url was changed (allenai\/zest#3) and #2057 should resolve this.","embeddings":[-0.5492244363,-0.1562176496,-0.0973048881,0.3641074598,0.3012053668,0.0931077302,0.0318725109,0.2230132669,0.2383637726,0.4310354888,-0.2034218013,-0.0097504938,-0.1162817851,0.1218978167,-0.0843498036,0.2447763383,-0.2308540195,0.4342668355,0.4354298413,0.0809467658,-0.1463077515,0.3227610588,-0.0329578854,0.0134566315,-0.2296866775,-0.0266707186,0.064969182,0.2269898206,-0.2469374985,-0.450075388,0.5318314433,-0.2040969729,0.382145524,0.4862642884,-0.0001192676,0.076782912,0.2923074663,-0.1723966599,-0.277204752,-0.6352200508,-0.5189527273,-0.1069162339,0.0227837767,-0.1745985448,-0.4987584949,0.1419637799,0.0905801207,0.0844901279,0.1318611056,0.4710632861,0.1323474646,-0.0255525988,0.4266806841,-0.2614115179,0.6230580807,0.0675981194,-0.1357329488,0.0985978022,0.2885711491,0.2700979114,0.4061024189,0.1312184632,-0.0209040325,0.1605392992,0.1259364933,0.0051018451,0.2155117989,-0.3926919401,0.2901311219,0.2027785182,0.7724493742,-0.2937853634,-0.3040547073,0.1115150675,0.0811822861,-0.4664172828,0.2170495391,0.1507623792,-0.2802621722,0.2095118612,0.1901822388,-0.0583263561,-0.1123306453,0.3451277912,0.0879818723,0.1269708425,-0.1504745185,0.005552602,-0.0237789806,-0.1501269788,0.2711737752,-0.0389018208,-0.0841390938,0.1122513935,-0.2754395604,-0.1604266167,0.0166549496,0.2436979711,-0.0223717783,0.1306643784,0.0592908077,-0.2001150995,0.1021633819,0.1904091984,0.0622776076,0.1541700661,-0.0759224519,-0.3409619927,0.2246865183,0.1661467701,-0.2793195844,-0.2687019706,-0.1667175591,-0.2494900674,-0.1447491348,-0.0926700681,0.1913476735,-0.350455761,-0.4722189307,-0.2292265892,-0.5416941047,0.087120004,0.1921856552,0.3942376375,0.0877431333,-0.2509027421,-0.099341929,-0.0406720936,-0.2872309387,-0.0326353163,-0.1425810456,0.2985761166,-0.1021685675,0.0024052823,0.2607377768,0.0807706863,0.5792786479,-0.2843334973,0.0084049255,-0.0233256873,0.2975989282,-0.2074875236,0.0521377176,0.4288754463,0.380862385,0.0161638167,-0.0982530788,-0.444539398,-0.3056356609,0.3296527863,-0.439458847,-0.4657526314,-0.1354969889,0.0743892714,-0.4352447391,0.003523994,0.0006512358,-0.2347811162,0.1604177505,-0.3929885924,-0.0217237957,-0.3515498638,-0.043433331,-0.2112887949,0.2502272725,0.3778166175,-0.6014804244,0.0128424959,-0.3805820644,0.1550828367,0.1199296936,0.1989078075,-0.1685970575,0.5266979933,-0.2713928223,-0.0915017128,0.6319549084,-0.6368858814,-0.8415307999,-0.1662150472,-0.0623471439,-0.0491294153,0.2330861986,0.2953896523,0.1769868433,-0.2069680542,0.0053665191,0.0280054733,-0.1691209078,-0.113520205,-0.1612265706,-0.1857514232,-0.3377084732,0.145174399,-0.0424749926,-0.1901915222,0.3948430717,-0.0299163144,-0.0283179618,0.0357074253,0.0041886875,0.133307308,0.5193724036,0.1891492605,-0.0042397128,0.0231180303,-0.308227241,0.0448477343,-0.1094072685,0.0647506788,-0.3085492551,-0.0404104739,-0.2715916932,-0.3036621809,-0.4424655139,-0.2636521459,0.0211440884,-0.0341807157,0.3390513957,0.0827218741,-0.040954344,0.0463739112,-0.0391621701,0.037134897,-0.1816708893,0.1381236762,-0.1515363753,-0.3019536138,-0.0527058132,0.2655154169,-0.0792637393,0.005150808,-0.1573840231,0.1158519909,0.1705407947,-0.1826977134,0.2029107809,0.2453830838,0.2108663321,-0.2708515823,0.1889864504,-0.0082362387,0.1406192929,0.0224563219,-0.026514288,0.1814837456,0.0642991066,-0.0926171616,0.0142035084,-0.1576364487,0.3949759603,-0.0413950495,0.0071036331,-0.0691644773,0.102370508,0.2601845264,0.2686906457,0.0029978487,-0.1557581872,0.0182827525,0.5315015316,0.1142820269,0.0069914851,0.0530432053,-0.1327366233,0.0963900611,0.2903290093,0.4729301333,0.604775846,0.3320865035,-0.2014874071,-0.1512369961,0.0997540131,-0.1310831904,0.0955172479,0.1010877788,0.2135856599,0.2487532645,0.0912274346,0.0835678726,-0.0502954423,-0.5824819207,-0.0593077838,0.3502303362,-0.1561137587,-0.1299252361,-0.16298531,-0.4231378138,0.0121324444,-0.2939905822,0.1531610489,-0.0411801115,-0.0124833537,0.279330492,0.3121997118,0.1506228,-0.1241767779,0.1046569273,0.0333119668,-0.5073771477,-0.1087347716,-0.0898247883,-0.1168642789,0.0260922201,0.429665029,0.0990311503,0.2851288617,-0.4056546986,0.1447036564,-0.2727582157,-0.1465431601,0.2022347003,0.049581822,0.3368256986,0.1884136349,0.1252973676,-0.2552151382,-0.0274087675,0.2641089261,-0.0612297095,-0.1408555359,-0.0980735049,-0.2689775229,-0.0554559156,0.0529270358,-0.3438365757,-0.56391114,-0.2585717142,-0.1811029166,0.1891873181,0.1829727888,0.0982800871,0.1670468599,0.1794397831,0.0341231078,-0.0652257949,0.0789482594,-0.5655246973,0.4193027318,-0.5766721964,-0.1247311234,0.4705350399,-0.1488472223,0.5684301853,0.0205997638,-0.2560326755,-0.0891146585,-0.1598949581,0.445884198,-0.1885035634,0.0396770313,0.3315299451,-0.0147490259,0.043048095,0.0007880731,-0.0024127138,-0.0621490963,0.0407078937,0.6619494557,0.192100808,0.4910691381,-0.239226386,0.5676385164,0.0756407902,-0.044208888,0.4262981713,-0.1783228964,0.221961543,-0.5061756968,-0.1453301609,0.164827913,0.185512051,-0.1260237843,0.1272166669,0.2610474825,-0.0493679568,-0.2332868129,0.0922019258,-0.3493933678,-0.0400582738,0.0150930276,0.2960338891,0.1670237035,0.0034790193,0.0046507297,-0.1122640818,-0.4151189625,0.0135920439,0.4828692079,-0.0788001493,0.2738875747,0.2162005901,-0.4309868813,-0.3567707539,0.2008530349,0.2648075819,-0.1865750551,-0.1755002141,0.1863674074,0.1618403047,0.1547298729,0.444773227,-0.0323359482,0.0782782584,0.1802159697,-0.0247225799,-0.02521169,-0.0299438424,-0.1443329155,0.2997746766,0.2518151104,0.1430124044,-0.0863924772,-0.178986311,0.1666567028,0.1022855118,-0.1228082404,0.0257532075,-0.100467585,-0.5379290581,-0.3752611279,-0.157457307,0.0406527817,0.2452535629,0.1521888971,-0.087018989,0.0569656305,0.0963928476,0.0727749094,0.2335206866,0.5060460567,-0.0360475928,0.025024496,-0.0150614157,-0.3126757145,0.6949639916,0.4729435146,0.0567890443,-0.2865799963,0.0340606123,-0.1437415034,0.1473527104,-0.0102484012,-0.0963008851,-0.2435187101,-0.4935561121,0.0672565922,0.2556268275,0.1614801139,0.1735777855,0.0213428866,-0.3130961061,-0.2729228735,0.18261154,0.0327072889,0.215635255,0.3947996795,0.2493322939,-0.1905591488,0.4591023624,-0.0358183496,0.8300575614,0.0358709097,0.1106391847,0.1712232083,0.116165787,0.4242929816,-0.1524285674,0.3545801044,-0.1772007346,-0.0623140894,-0.1570384651,-0.1664742678,0.0843952671,0.2573538423,-0.2287192792,0.3569446504,0.3331587911,0.1048731282,-0.0323954634,0.1906478852,-0.2452900261,0.0818854868,-0.157023266,0.0791838914,-0.1549096256,0.6620941162,-0.0841335878,-0.1675601453,-0.5331509709,-0.0649937093,-0.1665574759,-0.1421957165,-0.5519368052,0.274459213,-0.1833067387,-0.8469384909,0.0923221037,0.291670084,-0.1129484475,0.1048729569,-0.337862134,0.2216861844,0.3370819092,-0.238738969,0.183471337,0.1530790925,0.1034666374,-0.2592184544,-0.2260493189,0.0414019786,0.1249218807,-0.3734330833,0.1193469465,-0.1821624488,0.0780832097,-0.1210064441,-0.4089339972,-0.1684284806,-0.1427181661,-0.054125715,0.0881960839,0.0645788088,0.0624320097,0.2328215688,-0.2019101977,-0.2659597099,-0.1875076145,0.0992815867,-0.2765612304,0.411406368,0.5215316415,0.0224758945,-0.1664819568,-0.1105084941,0.2171848714,-0.16085504,-0.3135446906,0.025799688,0.1030323952,0.2682360709,0.0022015639,0.3596257269,0.3383302391,-0.1976757348,-0.0665353909,-0.5111649036,-0.1794642508,0.2216684073,0.2026061118,0.3132026494,-0.180256784,0.2094542086,0.1714221835,-0.4007724226,-0.2214138806,0.1799665689,-0.0862286761,0.0507299118,-0.0341355167,0.1443575323,0.3577928245,0.0980709121,-0.0068740747,-0.1318576485,-0.1616234034,-0.2157357484,0.1419702321,0.142949298,-0.0185366329,0.0105162673,-0.0688952804,-0.0605122112,-0.0823873803,-0.0300132222,0.1519717723,0.2290616035,-0.0094976537,0.1990972161,0.0601868592,-0.0760746822,-0.1169416383,0.3815583587,0.1356505007,0.1704593301,0.1690393239,0.2944316268,0.0512570105,-0.0311673153,-0.0093207974,-0.1481214166,-0.3203995228,0.021011766,0.194728151,-0.1855778098,0.0162720513,0.2852484882,0.0702980235,0.0583953969,-0.0802151114,0.1453116834,0.2804788947,0.0619568303,-0.212262705,0.1061518714,0.3297356665,0.2141732574,0.0580622703,0.2490781844,-0.0242490582,0.05695986,0.0340284295,0.1954718381,0.7585300207,-0.2084520906,-0.095143728,0.2974690199,-0.0486343019,0.296654284,0.0151992822,0.032012511,0.0889321715,0.4520904422,-0.1312082261,0.3690885603,-0.1816398203,0.103618741,-0.0168701336,-0.4776887,-0.3266048729,-0.2345927656,-0.1599963009,-0.1680214852,-0.0013954275,0.5693010092,0.0206115991,0.1268594265,-0.2767931819,0.0355461687,-0.3136084974,0.2228032202,-0.2753284276,-0.2170565724,0.2489824593,0.0473271608,0.028721014,0.0838117078,-0.1233946979,0.2148335576,-0.2976414859,0.1965078413,0.0569757372,0.0758291706,-0.050015308,-0.0799545199,0.3521754444,0.0548958853,-0.0995636582,-0.1438276321,0.2892383635,0.4752375185,0.3081264198,-0.0057245623,-0.1063618138,-0.0198746845,-0.2170747072,-0.1112358421,0.0235575847,0.1022666395,-0.0205627736,-0.0117277475,0.1469002068,-0.0678027719,0.0423887968,0.0682884529,-0.0984206423,-0.0591916107,0.0013837522,-0.0875109509,-0.2262156606,-0.1455897689,0.0645142123,-0.3305794895,0.0700751469,-0.035104204,0.0266138818,-0.0092228102,-0.0550507382,0.0207758117,-0.2382714748,0.0930659994,0.2795323133,0.3545050323,-0.435983181,-0.1675844491,-0.5266874433,0.0868719593,-0.1102219,-0.163673535,-0.1949089468,0.1538091898,-0.4237272143,0.082098648,0.3141462505,0.0027330159,0.2736623585,0.3161560595,-0.1667107195,-0.2694206834,-0.0941869318,-0.0749528557,-0.0045557171,-0.2586075068,0.0318249986,-0.1493289322,-0.0617902055,-0.1887201518,-0.2497360855,0.1525996923,-0.0256480891,0.4946905971,-0.1738687307,0.5443026423,-0.142663762,0.1813519597,0.0970860124,-0.0135706402,-0.0842951834,0.0893274322,0.1416112483,0.3505218327,-0.2550515532,-0.0157500058,-0.2995866239,0.1975644529,-0.0592747144,-0.0341165625,-0.042828273,-0.1127494723,-0.2846974432,0.2559111416,0.1605642736,0.0230689067,-0.1698351502,0.1207969189,-0.3196050525,-0.1180877537,0.5395075083,-0.1812970936,0.1485289782,-0.1643863171,0.0286928527,-0.0790736079,-0.2623628676,0.0592337437,0.0853297859,0.2187630087,0.1401187927,-0.1786881685,0.2941478789,0.1196623892,0.035121046,0.1572332382,0.5685510039,0.0433790907,-0.0230195336,-0.2887485921,-0.1851941049]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2054","title":"Could not find file for ZEST dataset","comments":"This has been fixed in #2057 by @matt-peters (thanks again !)\r\n\r\nThe fix is available on the master branch and we'll do a new release very soon :)","body":"I am trying to use zest dataset from Allen AI using below code in colab,\r\n```\r\n!pip install -q datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"zest\")\r\n```\r\n\r\nI am getting the following error,\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset zest\/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to \/root\/.cache\/huggingface\/datasets\/zest\/default\/0.0.0\/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"zest\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 612 )\r\n 613 elif response is not None and response.status_code == 404:\r\n--> 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/zest\/zest.zip\r\n```","comment_length":28,"text":"Could not find file for ZEST dataset\nI am trying to use zest dataset from Allen AI using below code in colab,\r\n```\r\n!pip install -q datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"zest\")\r\n```\r\n\r\nI am getting the following error,\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset zest\/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to \/root\/.cache\/huggingface\/datasets\/zest\/default\/0.0.0\/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"zest\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 612 )\r\n 613 elif response is not None and response.status_code == 404:\r\n--> 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/zest\/zest.zip\r\n```\nThis has been fixed in #2057 by @matt-peters (thanks again !)\r\n\r\nThe fix is available on the master branch and we'll do a new release very soon :)","embeddings":[-0.5492244363,-0.1562176496,-0.0973048881,0.3641074598,0.3012053668,0.0931077302,0.0318725109,0.2230132669,0.2383637726,0.4310354888,-0.2034218013,-0.0097504938,-0.1162817851,0.1218978167,-0.0843498036,0.2447763383,-0.2308540195,0.4342668355,0.4354298413,0.0809467658,-0.1463077515,0.3227610588,-0.0329578854,0.0134566315,-0.2296866775,-0.0266707186,0.064969182,0.2269898206,-0.2469374985,-0.450075388,0.5318314433,-0.2040969729,0.382145524,0.4862642884,-0.0001192676,0.076782912,0.2923074663,-0.1723966599,-0.277204752,-0.6352200508,-0.5189527273,-0.1069162339,0.0227837767,-0.1745985448,-0.4987584949,0.1419637799,0.0905801207,0.0844901279,0.1318611056,0.4710632861,0.1323474646,-0.0255525988,0.4266806841,-0.2614115179,0.6230580807,0.0675981194,-0.1357329488,0.0985978022,0.2885711491,0.2700979114,0.4061024189,0.1312184632,-0.0209040325,0.1605392992,0.1259364933,0.0051018451,0.2155117989,-0.3926919401,0.2901311219,0.2027785182,0.7724493742,-0.2937853634,-0.3040547073,0.1115150675,0.0811822861,-0.4664172828,0.2170495391,0.1507623792,-0.2802621722,0.2095118612,0.1901822388,-0.0583263561,-0.1123306453,0.3451277912,0.0879818723,0.1269708425,-0.1504745185,0.005552602,-0.0237789806,-0.1501269788,0.2711737752,-0.0389018208,-0.0841390938,0.1122513935,-0.2754395604,-0.1604266167,0.0166549496,0.2436979711,-0.0223717783,0.1306643784,0.0592908077,-0.2001150995,0.1021633819,0.1904091984,0.0622776076,0.1541700661,-0.0759224519,-0.3409619927,0.2246865183,0.1661467701,-0.2793195844,-0.2687019706,-0.1667175591,-0.2494900674,-0.1447491348,-0.0926700681,0.1913476735,-0.350455761,-0.4722189307,-0.2292265892,-0.5416941047,0.087120004,0.1921856552,0.3942376375,0.0877431333,-0.2509027421,-0.099341929,-0.0406720936,-0.2872309387,-0.0326353163,-0.1425810456,0.2985761166,-0.1021685675,0.0024052823,0.2607377768,0.0807706863,0.5792786479,-0.2843334973,0.0084049255,-0.0233256873,0.2975989282,-0.2074875236,0.0521377176,0.4288754463,0.380862385,0.0161638167,-0.0982530788,-0.444539398,-0.3056356609,0.3296527863,-0.439458847,-0.4657526314,-0.1354969889,0.0743892714,-0.4352447391,0.003523994,0.0006512358,-0.2347811162,0.1604177505,-0.3929885924,-0.0217237957,-0.3515498638,-0.043433331,-0.2112887949,0.2502272725,0.3778166175,-0.6014804244,0.0128424959,-0.3805820644,0.1550828367,0.1199296936,0.1989078075,-0.1685970575,0.5266979933,-0.2713928223,-0.0915017128,0.6319549084,-0.6368858814,-0.8415307999,-0.1662150472,-0.0623471439,-0.0491294153,0.2330861986,0.2953896523,0.1769868433,-0.2069680542,0.0053665191,0.0280054733,-0.1691209078,-0.113520205,-0.1612265706,-0.1857514232,-0.3377084732,0.145174399,-0.0424749926,-0.1901915222,0.3948430717,-0.0299163144,-0.0283179618,0.0357074253,0.0041886875,0.133307308,0.5193724036,0.1891492605,-0.0042397128,0.0231180303,-0.308227241,0.0448477343,-0.1094072685,0.0647506788,-0.3085492551,-0.0404104739,-0.2715916932,-0.3036621809,-0.4424655139,-0.2636521459,0.0211440884,-0.0341807157,0.3390513957,0.0827218741,-0.040954344,0.0463739112,-0.0391621701,0.037134897,-0.1816708893,0.1381236762,-0.1515363753,-0.3019536138,-0.0527058132,0.2655154169,-0.0792637393,0.005150808,-0.1573840231,0.1158519909,0.1705407947,-0.1826977134,0.2029107809,0.2453830838,0.2108663321,-0.2708515823,0.1889864504,-0.0082362387,0.1406192929,0.0224563219,-0.026514288,0.1814837456,0.0642991066,-0.0926171616,0.0142035084,-0.1576364487,0.3949759603,-0.0413950495,0.0071036331,-0.0691644773,0.102370508,0.2601845264,0.2686906457,0.0029978487,-0.1557581872,0.0182827525,0.5315015316,0.1142820269,0.0069914851,0.0530432053,-0.1327366233,0.0963900611,0.2903290093,0.4729301333,0.604775846,0.3320865035,-0.2014874071,-0.1512369961,0.0997540131,-0.1310831904,0.0955172479,0.1010877788,0.2135856599,0.2487532645,0.0912274346,0.0835678726,-0.0502954423,-0.5824819207,-0.0593077838,0.3502303362,-0.1561137587,-0.1299252361,-0.16298531,-0.4231378138,0.0121324444,-0.2939905822,0.1531610489,-0.0411801115,-0.0124833537,0.279330492,0.3121997118,0.1506228,-0.1241767779,0.1046569273,0.0333119668,-0.5073771477,-0.1087347716,-0.0898247883,-0.1168642789,0.0260922201,0.429665029,0.0990311503,0.2851288617,-0.4056546986,0.1447036564,-0.2727582157,-0.1465431601,0.2022347003,0.049581822,0.3368256986,0.1884136349,0.1252973676,-0.2552151382,-0.0274087675,0.2641089261,-0.0612297095,-0.1408555359,-0.0980735049,-0.2689775229,-0.0554559156,0.0529270358,-0.3438365757,-0.56391114,-0.2585717142,-0.1811029166,0.1891873181,0.1829727888,0.0982800871,0.1670468599,0.1794397831,0.0341231078,-0.0652257949,0.0789482594,-0.5655246973,0.4193027318,-0.5766721964,-0.1247311234,0.4705350399,-0.1488472223,0.5684301853,0.0205997638,-0.2560326755,-0.0891146585,-0.1598949581,0.445884198,-0.1885035634,0.0396770313,0.3315299451,-0.0147490259,0.043048095,0.0007880731,-0.0024127138,-0.0621490963,0.0407078937,0.6619494557,0.192100808,0.4910691381,-0.239226386,0.5676385164,0.0756407902,-0.044208888,0.4262981713,-0.1783228964,0.221961543,-0.5061756968,-0.1453301609,0.164827913,0.185512051,-0.1260237843,0.1272166669,0.2610474825,-0.0493679568,-0.2332868129,0.0922019258,-0.3493933678,-0.0400582738,0.0150930276,0.2960338891,0.1670237035,0.0034790193,0.0046507297,-0.1122640818,-0.4151189625,0.0135920439,0.4828692079,-0.0788001493,0.2738875747,0.2162005901,-0.4309868813,-0.3567707539,0.2008530349,0.2648075819,-0.1865750551,-0.1755002141,0.1863674074,0.1618403047,0.1547298729,0.444773227,-0.0323359482,0.0782782584,0.1802159697,-0.0247225799,-0.02521169,-0.0299438424,-0.1443329155,0.2997746766,0.2518151104,0.1430124044,-0.0863924772,-0.178986311,0.1666567028,0.1022855118,-0.1228082404,0.0257532075,-0.100467585,-0.5379290581,-0.3752611279,-0.157457307,0.0406527817,0.2452535629,0.1521888971,-0.087018989,0.0569656305,0.0963928476,0.0727749094,0.2335206866,0.5060460567,-0.0360475928,0.025024496,-0.0150614157,-0.3126757145,0.6949639916,0.4729435146,0.0567890443,-0.2865799963,0.0340606123,-0.1437415034,0.1473527104,-0.0102484012,-0.0963008851,-0.2435187101,-0.4935561121,0.0672565922,0.2556268275,0.1614801139,0.1735777855,0.0213428866,-0.3130961061,-0.2729228735,0.18261154,0.0327072889,0.215635255,0.3947996795,0.2493322939,-0.1905591488,0.4591023624,-0.0358183496,0.8300575614,0.0358709097,0.1106391847,0.1712232083,0.116165787,0.4242929816,-0.1524285674,0.3545801044,-0.1772007346,-0.0623140894,-0.1570384651,-0.1664742678,0.0843952671,0.2573538423,-0.2287192792,0.3569446504,0.3331587911,0.1048731282,-0.0323954634,0.1906478852,-0.2452900261,0.0818854868,-0.157023266,0.0791838914,-0.1549096256,0.6620941162,-0.0841335878,-0.1675601453,-0.5331509709,-0.0649937093,-0.1665574759,-0.1421957165,-0.5519368052,0.274459213,-0.1833067387,-0.8469384909,0.0923221037,0.291670084,-0.1129484475,0.1048729569,-0.337862134,0.2216861844,0.3370819092,-0.238738969,0.183471337,0.1530790925,0.1034666374,-0.2592184544,-0.2260493189,0.0414019786,0.1249218807,-0.3734330833,0.1193469465,-0.1821624488,0.0780832097,-0.1210064441,-0.4089339972,-0.1684284806,-0.1427181661,-0.054125715,0.0881960839,0.0645788088,0.0624320097,0.2328215688,-0.2019101977,-0.2659597099,-0.1875076145,0.0992815867,-0.2765612304,0.411406368,0.5215316415,0.0224758945,-0.1664819568,-0.1105084941,0.2171848714,-0.16085504,-0.3135446906,0.025799688,0.1030323952,0.2682360709,0.0022015639,0.3596257269,0.3383302391,-0.1976757348,-0.0665353909,-0.5111649036,-0.1794642508,0.2216684073,0.2026061118,0.3132026494,-0.180256784,0.2094542086,0.1714221835,-0.4007724226,-0.2214138806,0.1799665689,-0.0862286761,0.0507299118,-0.0341355167,0.1443575323,0.3577928245,0.0980709121,-0.0068740747,-0.1318576485,-0.1616234034,-0.2157357484,0.1419702321,0.142949298,-0.0185366329,0.0105162673,-0.0688952804,-0.0605122112,-0.0823873803,-0.0300132222,0.1519717723,0.2290616035,-0.0094976537,0.1990972161,0.0601868592,-0.0760746822,-0.1169416383,0.3815583587,0.1356505007,0.1704593301,0.1690393239,0.2944316268,0.0512570105,-0.0311673153,-0.0093207974,-0.1481214166,-0.3203995228,0.021011766,0.194728151,-0.1855778098,0.0162720513,0.2852484882,0.0702980235,0.0583953969,-0.0802151114,0.1453116834,0.2804788947,0.0619568303,-0.212262705,0.1061518714,0.3297356665,0.2141732574,0.0580622703,0.2490781844,-0.0242490582,0.05695986,0.0340284295,0.1954718381,0.7585300207,-0.2084520906,-0.095143728,0.2974690199,-0.0486343019,0.296654284,0.0151992822,0.032012511,0.0889321715,0.4520904422,-0.1312082261,0.3690885603,-0.1816398203,0.103618741,-0.0168701336,-0.4776887,-0.3266048729,-0.2345927656,-0.1599963009,-0.1680214852,-0.0013954275,0.5693010092,0.0206115991,0.1268594265,-0.2767931819,0.0355461687,-0.3136084974,0.2228032202,-0.2753284276,-0.2170565724,0.2489824593,0.0473271608,0.028721014,0.0838117078,-0.1233946979,0.2148335576,-0.2976414859,0.1965078413,0.0569757372,0.0758291706,-0.050015308,-0.0799545199,0.3521754444,0.0548958853,-0.0995636582,-0.1438276321,0.2892383635,0.4752375185,0.3081264198,-0.0057245623,-0.1063618138,-0.0198746845,-0.2170747072,-0.1112358421,0.0235575847,0.1022666395,-0.0205627736,-0.0117277475,0.1469002068,-0.0678027719,0.0423887968,0.0682884529,-0.0984206423,-0.0591916107,0.0013837522,-0.0875109509,-0.2262156606,-0.1455897689,0.0645142123,-0.3305794895,0.0700751469,-0.035104204,0.0266138818,-0.0092228102,-0.0550507382,0.0207758117,-0.2382714748,0.0930659994,0.2795323133,0.3545050323,-0.435983181,-0.1675844491,-0.5266874433,0.0868719593,-0.1102219,-0.163673535,-0.1949089468,0.1538091898,-0.4237272143,0.082098648,0.3141462505,0.0027330159,0.2736623585,0.3161560595,-0.1667107195,-0.2694206834,-0.0941869318,-0.0749528557,-0.0045557171,-0.2586075068,0.0318249986,-0.1493289322,-0.0617902055,-0.1887201518,-0.2497360855,0.1525996923,-0.0256480891,0.4946905971,-0.1738687307,0.5443026423,-0.142663762,0.1813519597,0.0970860124,-0.0135706402,-0.0842951834,0.0893274322,0.1416112483,0.3505218327,-0.2550515532,-0.0157500058,-0.2995866239,0.1975644529,-0.0592747144,-0.0341165625,-0.042828273,-0.1127494723,-0.2846974432,0.2559111416,0.1605642736,0.0230689067,-0.1698351502,0.1207969189,-0.3196050525,-0.1180877537,0.5395075083,-0.1812970936,0.1485289782,-0.1643863171,0.0286928527,-0.0790736079,-0.2623628676,0.0592337437,0.0853297859,0.2187630087,0.1401187927,-0.1786881685,0.2941478789,0.1196623892,0.035121046,0.1572332382,0.5685510039,0.0433790907,-0.0230195336,-0.2887485921,-0.1851941049]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2054","title":"Could not find file for ZEST dataset","comments":"Thanks @lhoestq and @matt-peters ","body":"I am trying to use zest dataset from Allen AI using below code in colab,\r\n```\r\n!pip install -q datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"zest\")\r\n```\r\n\r\nI am getting the following error,\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset zest\/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to \/root\/.cache\/huggingface\/datasets\/zest\/default\/0.0.0\/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"zest\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 612 )\r\n 613 elif response is not None and response.status_code == 404:\r\n--> 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/zest\/zest.zip\r\n```","comment_length":4,"text":"Could not find file for ZEST dataset\nI am trying to use zest dataset from Allen AI using below code in colab,\r\n```\r\n!pip install -q datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"zest\")\r\n```\r\n\r\nI am getting the following error,\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset zest\/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to \/root\/.cache\/huggingface\/datasets\/zest\/default\/0.0.0\/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"zest\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 612 )\r\n 613 elif response is not None and response.status_code == 404:\r\n--> 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/zest\/zest.zip\r\n```\nThanks @lhoestq and @matt-peters ","embeddings":[-0.5492244363,-0.1562176496,-0.0973048881,0.3641074598,0.3012053668,0.0931077302,0.0318725109,0.2230132669,0.2383637726,0.4310354888,-0.2034218013,-0.0097504938,-0.1162817851,0.1218978167,-0.0843498036,0.2447763383,-0.2308540195,0.4342668355,0.4354298413,0.0809467658,-0.1463077515,0.3227610588,-0.0329578854,0.0134566315,-0.2296866775,-0.0266707186,0.064969182,0.2269898206,-0.2469374985,-0.450075388,0.5318314433,-0.2040969729,0.382145524,0.4862642884,-0.0001192676,0.076782912,0.2923074663,-0.1723966599,-0.277204752,-0.6352200508,-0.5189527273,-0.1069162339,0.0227837767,-0.1745985448,-0.4987584949,0.1419637799,0.0905801207,0.0844901279,0.1318611056,0.4710632861,0.1323474646,-0.0255525988,0.4266806841,-0.2614115179,0.6230580807,0.0675981194,-0.1357329488,0.0985978022,0.2885711491,0.2700979114,0.4061024189,0.1312184632,-0.0209040325,0.1605392992,0.1259364933,0.0051018451,0.2155117989,-0.3926919401,0.2901311219,0.2027785182,0.7724493742,-0.2937853634,-0.3040547073,0.1115150675,0.0811822861,-0.4664172828,0.2170495391,0.1507623792,-0.2802621722,0.2095118612,0.1901822388,-0.0583263561,-0.1123306453,0.3451277912,0.0879818723,0.1269708425,-0.1504745185,0.005552602,-0.0237789806,-0.1501269788,0.2711737752,-0.0389018208,-0.0841390938,0.1122513935,-0.2754395604,-0.1604266167,0.0166549496,0.2436979711,-0.0223717783,0.1306643784,0.0592908077,-0.2001150995,0.1021633819,0.1904091984,0.0622776076,0.1541700661,-0.0759224519,-0.3409619927,0.2246865183,0.1661467701,-0.2793195844,-0.2687019706,-0.1667175591,-0.2494900674,-0.1447491348,-0.0926700681,0.1913476735,-0.350455761,-0.4722189307,-0.2292265892,-0.5416941047,0.087120004,0.1921856552,0.3942376375,0.0877431333,-0.2509027421,-0.099341929,-0.0406720936,-0.2872309387,-0.0326353163,-0.1425810456,0.2985761166,-0.1021685675,0.0024052823,0.2607377768,0.0807706863,0.5792786479,-0.2843334973,0.0084049255,-0.0233256873,0.2975989282,-0.2074875236,0.0521377176,0.4288754463,0.380862385,0.0161638167,-0.0982530788,-0.444539398,-0.3056356609,0.3296527863,-0.439458847,-0.4657526314,-0.1354969889,0.0743892714,-0.4352447391,0.003523994,0.0006512358,-0.2347811162,0.1604177505,-0.3929885924,-0.0217237957,-0.3515498638,-0.043433331,-0.2112887949,0.2502272725,0.3778166175,-0.6014804244,0.0128424959,-0.3805820644,0.1550828367,0.1199296936,0.1989078075,-0.1685970575,0.5266979933,-0.2713928223,-0.0915017128,0.6319549084,-0.6368858814,-0.8415307999,-0.1662150472,-0.0623471439,-0.0491294153,0.2330861986,0.2953896523,0.1769868433,-0.2069680542,0.0053665191,0.0280054733,-0.1691209078,-0.113520205,-0.1612265706,-0.1857514232,-0.3377084732,0.145174399,-0.0424749926,-0.1901915222,0.3948430717,-0.0299163144,-0.0283179618,0.0357074253,0.0041886875,0.133307308,0.5193724036,0.1891492605,-0.0042397128,0.0231180303,-0.308227241,0.0448477343,-0.1094072685,0.0647506788,-0.3085492551,-0.0404104739,-0.2715916932,-0.3036621809,-0.4424655139,-0.2636521459,0.0211440884,-0.0341807157,0.3390513957,0.0827218741,-0.040954344,0.0463739112,-0.0391621701,0.037134897,-0.1816708893,0.1381236762,-0.1515363753,-0.3019536138,-0.0527058132,0.2655154169,-0.0792637393,0.005150808,-0.1573840231,0.1158519909,0.1705407947,-0.1826977134,0.2029107809,0.2453830838,0.2108663321,-0.2708515823,0.1889864504,-0.0082362387,0.1406192929,0.0224563219,-0.026514288,0.1814837456,0.0642991066,-0.0926171616,0.0142035084,-0.1576364487,0.3949759603,-0.0413950495,0.0071036331,-0.0691644773,0.102370508,0.2601845264,0.2686906457,0.0029978487,-0.1557581872,0.0182827525,0.5315015316,0.1142820269,0.0069914851,0.0530432053,-0.1327366233,0.0963900611,0.2903290093,0.4729301333,0.604775846,0.3320865035,-0.2014874071,-0.1512369961,0.0997540131,-0.1310831904,0.0955172479,0.1010877788,0.2135856599,0.2487532645,0.0912274346,0.0835678726,-0.0502954423,-0.5824819207,-0.0593077838,0.3502303362,-0.1561137587,-0.1299252361,-0.16298531,-0.4231378138,0.0121324444,-0.2939905822,0.1531610489,-0.0411801115,-0.0124833537,0.279330492,0.3121997118,0.1506228,-0.1241767779,0.1046569273,0.0333119668,-0.5073771477,-0.1087347716,-0.0898247883,-0.1168642789,0.0260922201,0.429665029,0.0990311503,0.2851288617,-0.4056546986,0.1447036564,-0.2727582157,-0.1465431601,0.2022347003,0.049581822,0.3368256986,0.1884136349,0.1252973676,-0.2552151382,-0.0274087675,0.2641089261,-0.0612297095,-0.1408555359,-0.0980735049,-0.2689775229,-0.0554559156,0.0529270358,-0.3438365757,-0.56391114,-0.2585717142,-0.1811029166,0.1891873181,0.1829727888,0.0982800871,0.1670468599,0.1794397831,0.0341231078,-0.0652257949,0.0789482594,-0.5655246973,0.4193027318,-0.5766721964,-0.1247311234,0.4705350399,-0.1488472223,0.5684301853,0.0205997638,-0.2560326755,-0.0891146585,-0.1598949581,0.445884198,-0.1885035634,0.0396770313,0.3315299451,-0.0147490259,0.043048095,0.0007880731,-0.0024127138,-0.0621490963,0.0407078937,0.6619494557,0.192100808,0.4910691381,-0.239226386,0.5676385164,0.0756407902,-0.044208888,0.4262981713,-0.1783228964,0.221961543,-0.5061756968,-0.1453301609,0.164827913,0.185512051,-0.1260237843,0.1272166669,0.2610474825,-0.0493679568,-0.2332868129,0.0922019258,-0.3493933678,-0.0400582738,0.0150930276,0.2960338891,0.1670237035,0.0034790193,0.0046507297,-0.1122640818,-0.4151189625,0.0135920439,0.4828692079,-0.0788001493,0.2738875747,0.2162005901,-0.4309868813,-0.3567707539,0.2008530349,0.2648075819,-0.1865750551,-0.1755002141,0.1863674074,0.1618403047,0.1547298729,0.444773227,-0.0323359482,0.0782782584,0.1802159697,-0.0247225799,-0.02521169,-0.0299438424,-0.1443329155,0.2997746766,0.2518151104,0.1430124044,-0.0863924772,-0.178986311,0.1666567028,0.1022855118,-0.1228082404,0.0257532075,-0.100467585,-0.5379290581,-0.3752611279,-0.157457307,0.0406527817,0.2452535629,0.1521888971,-0.087018989,0.0569656305,0.0963928476,0.0727749094,0.2335206866,0.5060460567,-0.0360475928,0.025024496,-0.0150614157,-0.3126757145,0.6949639916,0.4729435146,0.0567890443,-0.2865799963,0.0340606123,-0.1437415034,0.1473527104,-0.0102484012,-0.0963008851,-0.2435187101,-0.4935561121,0.0672565922,0.2556268275,0.1614801139,0.1735777855,0.0213428866,-0.3130961061,-0.2729228735,0.18261154,0.0327072889,0.215635255,0.3947996795,0.2493322939,-0.1905591488,0.4591023624,-0.0358183496,0.8300575614,0.0358709097,0.1106391847,0.1712232083,0.116165787,0.4242929816,-0.1524285674,0.3545801044,-0.1772007346,-0.0623140894,-0.1570384651,-0.1664742678,0.0843952671,0.2573538423,-0.2287192792,0.3569446504,0.3331587911,0.1048731282,-0.0323954634,0.1906478852,-0.2452900261,0.0818854868,-0.157023266,0.0791838914,-0.1549096256,0.6620941162,-0.0841335878,-0.1675601453,-0.5331509709,-0.0649937093,-0.1665574759,-0.1421957165,-0.5519368052,0.274459213,-0.1833067387,-0.8469384909,0.0923221037,0.291670084,-0.1129484475,0.1048729569,-0.337862134,0.2216861844,0.3370819092,-0.238738969,0.183471337,0.1530790925,0.1034666374,-0.2592184544,-0.2260493189,0.0414019786,0.1249218807,-0.3734330833,0.1193469465,-0.1821624488,0.0780832097,-0.1210064441,-0.4089339972,-0.1684284806,-0.1427181661,-0.054125715,0.0881960839,0.0645788088,0.0624320097,0.2328215688,-0.2019101977,-0.2659597099,-0.1875076145,0.0992815867,-0.2765612304,0.411406368,0.5215316415,0.0224758945,-0.1664819568,-0.1105084941,0.2171848714,-0.16085504,-0.3135446906,0.025799688,0.1030323952,0.2682360709,0.0022015639,0.3596257269,0.3383302391,-0.1976757348,-0.0665353909,-0.5111649036,-0.1794642508,0.2216684073,0.2026061118,0.3132026494,-0.180256784,0.2094542086,0.1714221835,-0.4007724226,-0.2214138806,0.1799665689,-0.0862286761,0.0507299118,-0.0341355167,0.1443575323,0.3577928245,0.0980709121,-0.0068740747,-0.1318576485,-0.1616234034,-0.2157357484,0.1419702321,0.142949298,-0.0185366329,0.0105162673,-0.0688952804,-0.0605122112,-0.0823873803,-0.0300132222,0.1519717723,0.2290616035,-0.0094976537,0.1990972161,0.0601868592,-0.0760746822,-0.1169416383,0.3815583587,0.1356505007,0.1704593301,0.1690393239,0.2944316268,0.0512570105,-0.0311673153,-0.0093207974,-0.1481214166,-0.3203995228,0.021011766,0.194728151,-0.1855778098,0.0162720513,0.2852484882,0.0702980235,0.0583953969,-0.0802151114,0.1453116834,0.2804788947,0.0619568303,-0.212262705,0.1061518714,0.3297356665,0.2141732574,0.0580622703,0.2490781844,-0.0242490582,0.05695986,0.0340284295,0.1954718381,0.7585300207,-0.2084520906,-0.095143728,0.2974690199,-0.0486343019,0.296654284,0.0151992822,0.032012511,0.0889321715,0.4520904422,-0.1312082261,0.3690885603,-0.1816398203,0.103618741,-0.0168701336,-0.4776887,-0.3266048729,-0.2345927656,-0.1599963009,-0.1680214852,-0.0013954275,0.5693010092,0.0206115991,0.1268594265,-0.2767931819,0.0355461687,-0.3136084974,0.2228032202,-0.2753284276,-0.2170565724,0.2489824593,0.0473271608,0.028721014,0.0838117078,-0.1233946979,0.2148335576,-0.2976414859,0.1965078413,0.0569757372,0.0758291706,-0.050015308,-0.0799545199,0.3521754444,0.0548958853,-0.0995636582,-0.1438276321,0.2892383635,0.4752375185,0.3081264198,-0.0057245623,-0.1063618138,-0.0198746845,-0.2170747072,-0.1112358421,0.0235575847,0.1022666395,-0.0205627736,-0.0117277475,0.1469002068,-0.0678027719,0.0423887968,0.0682884529,-0.0984206423,-0.0591916107,0.0013837522,-0.0875109509,-0.2262156606,-0.1455897689,0.0645142123,-0.3305794895,0.0700751469,-0.035104204,0.0266138818,-0.0092228102,-0.0550507382,0.0207758117,-0.2382714748,0.0930659994,0.2795323133,0.3545050323,-0.435983181,-0.1675844491,-0.5266874433,0.0868719593,-0.1102219,-0.163673535,-0.1949089468,0.1538091898,-0.4237272143,0.082098648,0.3141462505,0.0027330159,0.2736623585,0.3161560595,-0.1667107195,-0.2694206834,-0.0941869318,-0.0749528557,-0.0045557171,-0.2586075068,0.0318249986,-0.1493289322,-0.0617902055,-0.1887201518,-0.2497360855,0.1525996923,-0.0256480891,0.4946905971,-0.1738687307,0.5443026423,-0.142663762,0.1813519597,0.0970860124,-0.0135706402,-0.0842951834,0.0893274322,0.1416112483,0.3505218327,-0.2550515532,-0.0157500058,-0.2995866239,0.1975644529,-0.0592747144,-0.0341165625,-0.042828273,-0.1127494723,-0.2846974432,0.2559111416,0.1605642736,0.0230689067,-0.1698351502,0.1207969189,-0.3196050525,-0.1180877537,0.5395075083,-0.1812970936,0.1485289782,-0.1643863171,0.0286928527,-0.0790736079,-0.2623628676,0.0592337437,0.0853297859,0.2187630087,0.1401187927,-0.1786881685,0.2941478789,0.1196623892,0.035121046,0.1572332382,0.5685510039,0.0433790907,-0.0230195336,-0.2887485921,-0.1851941049]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2054","title":"Could not find file for ZEST dataset","comments":"I am closing this issue since its fixed!","body":"I am trying to use zest dataset from Allen AI using below code in colab,\r\n```\r\n!pip install -q datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"zest\")\r\n```\r\n\r\nI am getting the following error,\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset zest\/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to \/root\/.cache\/huggingface\/datasets\/zest\/default\/0.0.0\/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"zest\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 612 )\r\n 613 elif response is not None and response.status_code == 404:\r\n--> 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/zest\/zest.zip\r\n```","comment_length":8,"text":"Could not find file for ZEST dataset\nI am trying to use zest dataset from Allen AI using below code in colab,\r\n```\r\n!pip install -q datasets\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"zest\")\r\n```\r\n\r\nI am getting the following error,\r\n```\r\nUsing custom data configuration default\r\n\r\nDownloading and preparing dataset zest\/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to \/root\/.cache\/huggingface\/datasets\/zest\/default\/0.0.0\/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n in ()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 dataset = load_dataset(\"zest\")\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)\r\n 612 )\r\n 613 elif response is not None and response.status_code == 404:\r\n--> 614 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 615 _raise_if_offline_mode_is_enabled(f\"Tried to reach {url}\")\r\n 616 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/zest\/zest.zip\r\n```\nI am closing this issue since its fixed!","embeddings":[-0.5492244363,-0.1562176496,-0.0973048881,0.3641074598,0.3012053668,0.0931077302,0.0318725109,0.2230132669,0.2383637726,0.4310354888,-0.2034218013,-0.0097504938,-0.1162817851,0.1218978167,-0.0843498036,0.2447763383,-0.2308540195,0.4342668355,0.4354298413,0.0809467658,-0.1463077515,0.3227610588,-0.0329578854,0.0134566315,-0.2296866775,-0.0266707186,0.064969182,0.2269898206,-0.2469374985,-0.450075388,0.5318314433,-0.2040969729,0.382145524,0.4862642884,-0.0001192676,0.076782912,0.2923074663,-0.1723966599,-0.277204752,-0.6352200508,-0.5189527273,-0.1069162339,0.0227837767,-0.1745985448,-0.4987584949,0.1419637799,0.0905801207,0.0844901279,0.1318611056,0.4710632861,0.1323474646,-0.0255525988,0.4266806841,-0.2614115179,0.6230580807,0.0675981194,-0.1357329488,0.0985978022,0.2885711491,0.2700979114,0.4061024189,0.1312184632,-0.0209040325,0.1605392992,0.1259364933,0.0051018451,0.2155117989,-0.3926919401,0.2901311219,0.2027785182,0.7724493742,-0.2937853634,-0.3040547073,0.1115150675,0.0811822861,-0.4664172828,0.2170495391,0.1507623792,-0.2802621722,0.2095118612,0.1901822388,-0.0583263561,-0.1123306453,0.3451277912,0.0879818723,0.1269708425,-0.1504745185,0.005552602,-0.0237789806,-0.1501269788,0.2711737752,-0.0389018208,-0.0841390938,0.1122513935,-0.2754395604,-0.1604266167,0.0166549496,0.2436979711,-0.0223717783,0.1306643784,0.0592908077,-0.2001150995,0.1021633819,0.1904091984,0.0622776076,0.1541700661,-0.0759224519,-0.3409619927,0.2246865183,0.1661467701,-0.2793195844,-0.2687019706,-0.1667175591,-0.2494900674,-0.1447491348,-0.0926700681,0.1913476735,-0.350455761,-0.4722189307,-0.2292265892,-0.5416941047,0.087120004,0.1921856552,0.3942376375,0.0877431333,-0.2509027421,-0.099341929,-0.0406720936,-0.2872309387,-0.0326353163,-0.1425810456,0.2985761166,-0.1021685675,0.0024052823,0.2607377768,0.0807706863,0.5792786479,-0.2843334973,0.0084049255,-0.0233256873,0.2975989282,-0.2074875236,0.0521377176,0.4288754463,0.380862385,0.0161638167,-0.0982530788,-0.444539398,-0.3056356609,0.3296527863,-0.439458847,-0.4657526314,-0.1354969889,0.0743892714,-0.4352447391,0.003523994,0.0006512358,-0.2347811162,0.1604177505,-0.3929885924,-0.0217237957,-0.3515498638,-0.043433331,-0.2112887949,0.2502272725,0.3778166175,-0.6014804244,0.0128424959,-0.3805820644,0.1550828367,0.1199296936,0.1989078075,-0.1685970575,0.5266979933,-0.2713928223,-0.0915017128,0.6319549084,-0.6368858814,-0.8415307999,-0.1662150472,-0.0623471439,-0.0491294153,0.2330861986,0.2953896523,0.1769868433,-0.2069680542,0.0053665191,0.0280054733,-0.1691209078,-0.113520205,-0.1612265706,-0.1857514232,-0.3377084732,0.145174399,-0.0424749926,-0.1901915222,0.3948430717,-0.0299163144,-0.0283179618,0.0357074253,0.0041886875,0.133307308,0.5193724036,0.1891492605,-0.0042397128,0.0231180303,-0.308227241,0.0448477343,-0.1094072685,0.0647506788,-0.3085492551,-0.0404104739,-0.2715916932,-0.3036621809,-0.4424655139,-0.2636521459,0.0211440884,-0.0341807157,0.3390513957,0.0827218741,-0.040954344,0.0463739112,-0.0391621701,0.037134897,-0.1816708893,0.1381236762,-0.1515363753,-0.3019536138,-0.0527058132,0.2655154169,-0.0792637393,0.005150808,-0.1573840231,0.1158519909,0.1705407947,-0.1826977134,0.2029107809,0.2453830838,0.2108663321,-0.2708515823,0.1889864504,-0.0082362387,0.1406192929,0.0224563219,-0.026514288,0.1814837456,0.0642991066,-0.0926171616,0.0142035084,-0.1576364487,0.3949759603,-0.0413950495,0.0071036331,-0.0691644773,0.102370508,0.2601845264,0.2686906457,0.0029978487,-0.1557581872,0.0182827525,0.5315015316,0.1142820269,0.0069914851,0.0530432053,-0.1327366233,0.0963900611,0.2903290093,0.4729301333,0.604775846,0.3320865035,-0.2014874071,-0.1512369961,0.0997540131,-0.1310831904,0.0955172479,0.1010877788,0.2135856599,0.2487532645,0.0912274346,0.0835678726,-0.0502954423,-0.5824819207,-0.0593077838,0.3502303362,-0.1561137587,-0.1299252361,-0.16298531,-0.4231378138,0.0121324444,-0.2939905822,0.1531610489,-0.0411801115,-0.0124833537,0.279330492,0.3121997118,0.1506228,-0.1241767779,0.1046569273,0.0333119668,-0.5073771477,-0.1087347716,-0.0898247883,-0.1168642789,0.0260922201,0.429665029,0.0990311503,0.2851288617,-0.4056546986,0.1447036564,-0.2727582157,-0.1465431601,0.2022347003,0.049581822,0.3368256986,0.1884136349,0.1252973676,-0.2552151382,-0.0274087675,0.2641089261,-0.0612297095,-0.1408555359,-0.0980735049,-0.2689775229,-0.0554559156,0.0529270358,-0.3438365757,-0.56391114,-0.2585717142,-0.1811029166,0.1891873181,0.1829727888,0.0982800871,0.1670468599,0.1794397831,0.0341231078,-0.0652257949,0.0789482594,-0.5655246973,0.4193027318,-0.5766721964,-0.1247311234,0.4705350399,-0.1488472223,0.5684301853,0.0205997638,-0.2560326755,-0.0891146585,-0.1598949581,0.445884198,-0.1885035634,0.0396770313,0.3315299451,-0.0147490259,0.043048095,0.0007880731,-0.0024127138,-0.0621490963,0.0407078937,0.6619494557,0.192100808,0.4910691381,-0.239226386,0.5676385164,0.0756407902,-0.044208888,0.4262981713,-0.1783228964,0.221961543,-0.5061756968,-0.1453301609,0.164827913,0.185512051,-0.1260237843,0.1272166669,0.2610474825,-0.0493679568,-0.2332868129,0.0922019258,-0.3493933678,-0.0400582738,0.0150930276,0.2960338891,0.1670237035,0.0034790193,0.0046507297,-0.1122640818,-0.4151189625,0.0135920439,0.4828692079,-0.0788001493,0.2738875747,0.2162005901,-0.4309868813,-0.3567707539,0.2008530349,0.2648075819,-0.1865750551,-0.1755002141,0.1863674074,0.1618403047,0.1547298729,0.444773227,-0.0323359482,0.0782782584,0.1802159697,-0.0247225799,-0.02521169,-0.0299438424,-0.1443329155,0.2997746766,0.2518151104,0.1430124044,-0.0863924772,-0.178986311,0.1666567028,0.1022855118,-0.1228082404,0.0257532075,-0.100467585,-0.5379290581,-0.3752611279,-0.157457307,0.0406527817,0.2452535629,0.1521888971,-0.087018989,0.0569656305,0.0963928476,0.0727749094,0.2335206866,0.5060460567,-0.0360475928,0.025024496,-0.0150614157,-0.3126757145,0.6949639916,0.4729435146,0.0567890443,-0.2865799963,0.0340606123,-0.1437415034,0.1473527104,-0.0102484012,-0.0963008851,-0.2435187101,-0.4935561121,0.0672565922,0.2556268275,0.1614801139,0.1735777855,0.0213428866,-0.3130961061,-0.2729228735,0.18261154,0.0327072889,0.215635255,0.3947996795,0.2493322939,-0.1905591488,0.4591023624,-0.0358183496,0.8300575614,0.0358709097,0.1106391847,0.1712232083,0.116165787,0.4242929816,-0.1524285674,0.3545801044,-0.1772007346,-0.0623140894,-0.1570384651,-0.1664742678,0.0843952671,0.2573538423,-0.2287192792,0.3569446504,0.3331587911,0.1048731282,-0.0323954634,0.1906478852,-0.2452900261,0.0818854868,-0.157023266,0.0791838914,-0.1549096256,0.6620941162,-0.0841335878,-0.1675601453,-0.5331509709,-0.0649937093,-0.1665574759,-0.1421957165,-0.5519368052,0.274459213,-0.1833067387,-0.8469384909,0.0923221037,0.291670084,-0.1129484475,0.1048729569,-0.337862134,0.2216861844,0.3370819092,-0.238738969,0.183471337,0.1530790925,0.1034666374,-0.2592184544,-0.2260493189,0.0414019786,0.1249218807,-0.3734330833,0.1193469465,-0.1821624488,0.0780832097,-0.1210064441,-0.4089339972,-0.1684284806,-0.1427181661,-0.054125715,0.0881960839,0.0645788088,0.0624320097,0.2328215688,-0.2019101977,-0.2659597099,-0.1875076145,0.0992815867,-0.2765612304,0.411406368,0.5215316415,0.0224758945,-0.1664819568,-0.1105084941,0.2171848714,-0.16085504,-0.3135446906,0.025799688,0.1030323952,0.2682360709,0.0022015639,0.3596257269,0.3383302391,-0.1976757348,-0.0665353909,-0.5111649036,-0.1794642508,0.2216684073,0.2026061118,0.3132026494,-0.180256784,0.2094542086,0.1714221835,-0.4007724226,-0.2214138806,0.1799665689,-0.0862286761,0.0507299118,-0.0341355167,0.1443575323,0.3577928245,0.0980709121,-0.0068740747,-0.1318576485,-0.1616234034,-0.2157357484,0.1419702321,0.142949298,-0.0185366329,0.0105162673,-0.0688952804,-0.0605122112,-0.0823873803,-0.0300132222,0.1519717723,0.2290616035,-0.0094976537,0.1990972161,0.0601868592,-0.0760746822,-0.1169416383,0.3815583587,0.1356505007,0.1704593301,0.1690393239,0.2944316268,0.0512570105,-0.0311673153,-0.0093207974,-0.1481214166,-0.3203995228,0.021011766,0.194728151,-0.1855778098,0.0162720513,0.2852484882,0.0702980235,0.0583953969,-0.0802151114,0.1453116834,0.2804788947,0.0619568303,-0.212262705,0.1061518714,0.3297356665,0.2141732574,0.0580622703,0.2490781844,-0.0242490582,0.05695986,0.0340284295,0.1954718381,0.7585300207,-0.2084520906,-0.095143728,0.2974690199,-0.0486343019,0.296654284,0.0151992822,0.032012511,0.0889321715,0.4520904422,-0.1312082261,0.3690885603,-0.1816398203,0.103618741,-0.0168701336,-0.4776887,-0.3266048729,-0.2345927656,-0.1599963009,-0.1680214852,-0.0013954275,0.5693010092,0.0206115991,0.1268594265,-0.2767931819,0.0355461687,-0.3136084974,0.2228032202,-0.2753284276,-0.2170565724,0.2489824593,0.0473271608,0.028721014,0.0838117078,-0.1233946979,0.2148335576,-0.2976414859,0.1965078413,0.0569757372,0.0758291706,-0.050015308,-0.0799545199,0.3521754444,0.0548958853,-0.0995636582,-0.1438276321,0.2892383635,0.4752375185,0.3081264198,-0.0057245623,-0.1063618138,-0.0198746845,-0.2170747072,-0.1112358421,0.0235575847,0.1022666395,-0.0205627736,-0.0117277475,0.1469002068,-0.0678027719,0.0423887968,0.0682884529,-0.0984206423,-0.0591916107,0.0013837522,-0.0875109509,-0.2262156606,-0.1455897689,0.0645142123,-0.3305794895,0.0700751469,-0.035104204,0.0266138818,-0.0092228102,-0.0550507382,0.0207758117,-0.2382714748,0.0930659994,0.2795323133,0.3545050323,-0.435983181,-0.1675844491,-0.5266874433,0.0868719593,-0.1102219,-0.163673535,-0.1949089468,0.1538091898,-0.4237272143,0.082098648,0.3141462505,0.0027330159,0.2736623585,0.3161560595,-0.1667107195,-0.2694206834,-0.0941869318,-0.0749528557,-0.0045557171,-0.2586075068,0.0318249986,-0.1493289322,-0.0617902055,-0.1887201518,-0.2497360855,0.1525996923,-0.0256480891,0.4946905971,-0.1738687307,0.5443026423,-0.142663762,0.1813519597,0.0970860124,-0.0135706402,-0.0842951834,0.0893274322,0.1416112483,0.3505218327,-0.2550515532,-0.0157500058,-0.2995866239,0.1975644529,-0.0592747144,-0.0341165625,-0.042828273,-0.1127494723,-0.2846974432,0.2559111416,0.1605642736,0.0230689067,-0.1698351502,0.1207969189,-0.3196050525,-0.1180877537,0.5395075083,-0.1812970936,0.1485289782,-0.1643863171,0.0286928527,-0.0790736079,-0.2623628676,0.0592337437,0.0853297859,0.2187630087,0.1401187927,-0.1786881685,0.2941478789,0.1196623892,0.035121046,0.1572332382,0.5685510039,0.0433790907,-0.0230195336,-0.2887485921,-0.1851941049]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2052","title":"Timit_asr dataset repeats examples","comments":"Hi,\r\n\r\nthis was fixed by #1995, so you can wait for the next release or install the package directly from the master branch with the following command: \r\n```bash\r\npip install git+https:\/\/github.com\/huggingface\/datasets\r\n```","body":"Summary\r\n\r\nWhen loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same\r\nSteps to reproduce\r\n\r\nAs an example, on this code there is the text from the training part:\r\n\r\nCode snippet:\r\n```\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\ntimit['train']['text']\r\n#['Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n```\r\nThe same behavior happens for other columns\r\n\r\nExpected behavior:\r\n\r\nDifferent info on the actual timit_asr dataset\r\n\r\nActual behavior:\r\n\r\nWhen loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same. I've checked datasets 1.3 and the rows are different\r\nDebug info\r\n\r\n Streamlit version: (get it with $ streamlit version)\r\n Python version: Python 3.6.12\r\n Using Conda? PipEnv? PyEnv? Pex? Using pip\r\n OS version: Centos-release-7-9.2009.1.el7.centos.x86_64\r\n\r\nAdditional information\r\n\r\nYou can check the same behavior on https:\/\/huggingface.co\/datasets\/viewer\/?dataset=timit_asr","comment_length":32,"text":"Timit_asr dataset repeats examples\nSummary\r\n\r\nWhen loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same\r\nSteps to reproduce\r\n\r\nAs an example, on this code there is the text from the training part:\r\n\r\nCode snippet:\r\n```\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\ntimit['train']['text']\r\n#['Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n```\r\nThe same behavior happens for other columns\r\n\r\nExpected behavior:\r\n\r\nDifferent info on the actual timit_asr dataset\r\n\r\nActual behavior:\r\n\r\nWhen loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same. I've checked datasets 1.3 and the rows are different\r\nDebug info\r\n\r\n Streamlit version: (get it with $ streamlit version)\r\n Python version: Python 3.6.12\r\n Using Conda? PipEnv? PyEnv? Pex? Using pip\r\n OS version: Centos-release-7-9.2009.1.el7.centos.x86_64\r\n\r\nAdditional information\r\n\r\nYou can check the same behavior on https:\/\/huggingface.co\/datasets\/viewer\/?dataset=timit_asr\nHi,\r\n\r\nthis was fixed by #1995, so you can wait for the next release or install the package directly from the master branch with the following command: \r\n```bash\r\npip install git+https:\/\/github.com\/huggingface\/datasets\r\n```","embeddings":[0.0693238303,-0.2360813767,0.0143211801,0.3875417411,0.2564877272,-0.052774936,0.3894462585,0.1509235948,-0.3530937731,0.2145144641,0.0362559035,0.3838868439,-0.1914107949,0.3149206638,0.1428200752,0.0897241086,-0.0645085797,0.1746298671,-0.5109459758,-0.207854107,0.1559592038,0.107349582,-0.0441997871,-0.0219411999,-0.1925637275,0.2516241968,0.034533523,-0.2753400803,0.0288738497,-0.6021385193,0.2152353972,0.0859404504,0.1602355838,0.5223380923,-0.0001203386,0.085062556,-0.1006255597,-0.0585892051,-0.4437145889,-0.3315133452,-0.129373461,-0.0098590283,0.4994364381,-0.1294307709,-0.353879869,-0.2658421397,0.0228594225,0.1678988934,0.1387133151,0.1977487206,0.1442998946,0.2076290697,-0.4852289855,-0.0598838851,0.1789754033,0.0927969217,-0.0897974819,0.1834605187,-0.1145897433,0.34348014,-0.0861930922,0.2969873846,0.0472152121,0.3001303971,-0.0325891189,0.0343293697,0.0803429931,-0.1385795176,0.0530866496,0.3074523211,0.6112798452,-0.1911728531,-0.006164331,-0.2379683256,0.2809107602,-0.0089171771,0.1315224618,0.2282369733,-0.1578891575,0.1716043204,-0.4439931214,0.1640815884,-0.2220607996,0.1022906825,-0.0723651275,-0.4805187881,-0.2431519479,0.0890064687,-0.1993243694,-0.1923929602,0.2014489323,-0.2141873837,0.1594441235,0.0345593467,-0.2622407675,0.1092169285,-0.0398223922,0.2607025504,0.0437209159,0.2176585197,0.7037342191,0.195019111,-0.1651889831,-0.1209520847,0.1111265793,0.214263007,0.1403140426,-0.1036118343,0.0006801457,0.1406616271,-0.3247894347,0.0161448438,-0.0948205218,0.3506808877,0.2983540297,-0.4594764709,0.2485127151,-0.2413159907,-0.4751442373,0.1753272116,-0.4329754412,-0.1862203479,-0.0301720351,0.1282126009,-0.032817442,0.2887885273,-0.0671522915,-0.1436191946,0.033126764,-0.142973125,-0.1733697951,-0.2447229028,-0.2997201979,0.3173596859,0.1830929518,-0.5795408487,0.1755428463,0.5316569805,-0.0979854912,-0.0599521585,0.1499044001,0.0484337136,0.0605391487,0.0527878292,-0.0916916057,0.2154544741,0.0453242026,0.1959704161,-0.0045511969,0.0992062241,0.0060083386,0.0848728716,0.5086863041,0.1952257603,-0.325463444,-0.1416299641,-0.1361305118,0.1503886133,-0.1309700757,-0.1011544988,0.0790347829,-0.2211146057,-0.1790559888,-0.1774930656,0.0675057545,0.3432037234,-0.3481624424,-0.1797856688,0.3411158323,0.067322433,0.1209600568,0.0170693621,-0.0397707447,0.2547914088,-0.1584244668,-0.0625496134,0.0198324639,-0.1838140935,-0.2675595284,0.1368087679,-0.0317496583,0.4677166343,0.0180712417,-0.0457966253,0.325183183,0.3720587194,0.0871371999,-0.0469075069,0.4046121836,-0.0370944291,-0.4191308022,-0.0683913678,0.1653861851,0.1985431314,-0.173589766,0.2753488421,0.1103103012,-0.0650227144,0.3484919071,-0.1165023744,-0.1479747742,-0.1333269775,0.0910538957,0.155309394,0.1657625288,-0.3785127699,-0.1361991912,-0.0131421974,0.2551571131,0.19182688,0.3550348878,-0.0333180465,-0.3664958477,-0.2809584439,-0.39504987,-0.0378554352,0.102124989,0.5243572593,-0.0206522215,0.2404852659,0.0254026055,0.654519558,-0.4834011793,0.1554483622,-0.1168727428,0.1829820573,0.0746080354,-0.0083817737,0.0238861367,0.0853050053,0.181991592,-0.1432335526,0.119768247,0.3222537637,0.0608440898,-0.0229393691,-0.2227678597,-0.1932323426,0.2595642805,-0.1983759105,-0.1534409821,0.1877188385,0.1179831624,0.0357431434,-0.0955214202,0.3096632957,0.0103773084,0.148306787,0.2213652581,0.0738518089,0.021175174,-0.0808992386,-0.3685730994,-0.4152880311,0.4115943313,-0.0445591882,0.0660966039,0.173598513,-0.413744837,0.179787755,0.0957723185,-0.1290572584,-0.1319271475,0.1978547126,-0.1585223675,0.2480879724,0.0489423648,0.3009958267,0.2230157256,0.1953908056,-0.0019308525,0.0747675821,-0.2692965269,-0.1935736835,0.1433221847,0.0193449222,-0.2603229284,0.4037386477,0.2191151977,-0.0176813658,-0.6143211722,-0.0884906352,0.2171187997,0.2974383533,-0.2661319673,0.0853498876,-0.4890751839,-0.0450068302,-0.3011750877,-0.1606367528,0.1040866077,0.0812572762,0.039384041,0.208057642,0.1990811676,0.2026453763,-0.3972111642,0.2981900573,-0.0682340711,0.4391578436,-0.0972708389,0.2479812652,-0.3348937333,0.0424616002,-0.085272558,0.1215661615,0.2688836455,-0.4862579703,0.0993539914,-0.3178024888,-0.329951942,0.3918776512,-0.1489201188,0.4757384062,0.2216872126,0.0783465505,-0.1970855892,-0.2506314814,0.0102321412,-0.1400298178,-0.2091223747,0.1264417022,-0.0564196296,0.0704466626,-0.1707552373,-0.5320759416,0.1326203495,-0.1759246141,0.4266978502,0.192959249,0.1590533257,0.5182472467,0.0276148356,0.3334203064,0.1154283062,0.2547902167,-0.4598690867,-0.2551384568,0.2225894332,-0.0588355176,-0.1310429275,-0.1529302299,0.0204576701,0.0693665817,0.2738563716,-0.6682428122,-0.0594674423,-0.2347604781,0.0669735521,0.0098983468,0.0657857433,0.1793263555,-0.0859024301,-0.0352972932,-0.1392109543,-0.0137345372,0.2026993185,-0.0850526467,-0.0505669229,0.0524007902,0.1572902948,0.0933310315,0.1990322769,0.4609867632,-0.1091711521,0.2948640883,-0.0948017314,0.5107052922,-0.0797246322,-0.1776985526,-0.2111649066,-0.1123107299,0.0125113744,0.2650344968,-0.1442334801,0.1855433285,-0.052801948,-0.007715655,0.0556769669,-0.2703033388,-0.2283927798,-0.4031398594,0.3369254172,0.1973817945,0.1473370492,0.0239828136,-0.1598916948,-0.2391502559,0.2751942575,0.1058551073,-0.0546759665,-0.7649233341,-0.2049749345,-0.4185813665,0.2463745028,0.3461543024,0.3524974585,0.0982511863,-0.2850190997,0.4231262207,0.089397192,0.5272787809,-0.426173687,-0.1225829348,0.0781606063,0.070810847,-0.5122358203,-0.1821724325,-0.3769619763,-0.1074797809,0.1066753566,0.3323648572,-0.1447001696,-0.2569247782,0.2897038162,0.2483832389,-0.1374631077,-0.2136364132,-0.0776474699,-0.0747209266,0.0646530315,-0.0018665456,-0.2464236617,-0.1212768629,-0.3947780728,0.012344812,-0.2985564768,0.1659851968,0.5115221143,0.2698182762,0.4900927544,0.0091707092,0.3090979457,0.3732331097,-0.0511603057,0.1870750785,0.571359694,-0.3596920967,-0.004707769,0.1924051195,-0.26983881,0.1046855673,0.4873094261,-0.0608644113,0.2108043283,-0.2117762566,-0.0184895545,-0.1353256404,0.2354843616,0.2572872043,0.0083178254,0.089765422,-0.6204378605,0.2661387622,0.0086182337,-0.2467441708,0.2705598176,0.0744711906,-0.0955314264,0.1597130746,-0.2795657218,0.8087245822,0.1241441667,0.5083410144,0.3728880286,0.0276526324,0.0537173003,-0.1236926988,0.0781790912,-0.2046161592,-0.2007231265,-0.1222515479,-0.1652892381,-0.0426901691,0.119620651,-0.3727223575,0.0550795086,-0.0400887877,0.2199938893,0.1189480945,0.0309292544,-0.2442181557,0.0977584869,-0.0443655141,0.0591698326,0.064988263,0.1120805368,0.1047761962,0.1555496901,0.3324554861,-0.2676259577,-0.2883507311,0.025152918,-0.118530035,-0.1233941168,0.0150882751,-0.5537112355,-0.0543640442,0.0227169227,0.0376645103,0.0825629532,0.0196774937,-0.1897841096,0.1444820613,0.2891579866,-0.2672913671,-0.0402197167,0.5374183655,0.231379658,-0.3256859779,0.1473491788,0.2718449831,-0.0018393787,-0.3763468266,0.1328825355,-0.2728020251,-0.6834982634,0.0993262231,0.0188119505,-0.0548777208,-0.2263853252,0.0769661963,-0.3152380288,-0.1691734642,0.2465950698,0.3125596046,-0.4311827421,-0.1518791914,0.4057873189,0.389921844,-0.0322274379,0.4833839834,-0.0553574562,-0.0931493267,-0.2447719276,0.0896772891,0.1940064281,-0.2671209574,0.0831904709,-0.2182865292,0.235419482,0.2110475153,0.2475493848,0.0582140535,-0.0375554822,-0.3182398379,-0.2167702764,0.0550098978,0.1399965435,0.2336660475,0.3746979833,-0.0493179485,0.2886081934,-0.0128019964,-0.1217281893,-0.2259507328,0.3694863319,0.0717831552,0.2320923954,0.0476728827,-0.2835389972,0.2777878344,-0.0037462835,0.0894872248,0.2811301649,-0.1567191929,-0.1079946309,-0.375407815,0.1233824566,0.0319673344,0.0622643642,-0.2143042982,0.0490686148,0.1448476911,-0.0080120815,0.2008197606,0.216795668,0.0344922021,0.1800308824,-0.063382335,0.1611385196,0.0755770132,-0.1940605789,-0.0879659057,0.0265815351,0.3356522918,-0.1145701334,-0.3686912954,0.1255862415,-0.2908984125,0.1372589171,0.2278490514,-0.0607039407,0.0060289064,-0.225022614,-0.0057339901,0.0907274559,0.2555993795,0.5138317347,-0.2099701911,-0.3390116692,0.0577032976,0.1151940376,-0.1063014939,-0.1141085178,-0.1480736434,0.1249520779,0.1658431441,0.3090191185,0.0514546037,-0.4245748818,0.4315136075,-0.1258101612,0.2499516457,-0.2317195982,0.2440086007,0.4143305421,-0.0603278577,-0.1476066709,0.2601909637,0.1235292032,-0.2130868137,0.5466448069,-0.0603725798,0.0290232897,-0.3541246057,-0.0248370934,0.089893654,-0.1018580943,-0.0802809149,0.2303501368,-0.2128078938,0.3305294514,0.1146321297,0.4066492915,-0.0408456884,-0.0810562447,-0.334146142,-0.2729619741,-0.2580868006,-0.1305443347,-0.3098895848,-0.4051511586,-0.0625761226,0.0575430728,-0.1961643994,0.0544602051,0.4997911453,0.1460045576,-0.0082598776,-0.7142173052,-0.2702354193,0.110790126,0.2933400571,0.0491911434,0.4703094065,0.1340152919,0.0178152788,0.0853116438,0.1957297623,0.2397433221,0.18891792,0.1225126982,0.1102057323,-0.3764492869,0.0745750219,0.063703537,0.307059288,0.0765786543,-0.2732903659,0.1072877049,0.0333979204,-0.0789425597,0.2780892253,0.0680696443,-0.0264700353,-0.8032155037,0.3519619703,0.0192745849,0.2420410067,0.0527270921,-0.0352879129,-0.457416743,0.4931133091,0.3092861176,0.0321063288,0.0530531183,-0.2545377314,0.0737999156,-0.2497573495,0.2542212605,0.1347325146,-0.0190863032,-0.124003768,-0.1625430584,-0.6735637188,0.0717220008,-0.2230775654,-0.0124966782,0.1540582627,0.0600397177,-0.0518550202,0.0883440524,0.1788545251,0.0864390209,-0.4328099787,0.047537744,-0.401068598,-0.132304728,0.1205495,0.0634412318,-0.0710199624,-0.1109477058,0.0470497496,-0.1067716554,-0.0390103795,-0.0458408222,0.0019417286,-0.0850799233,0.18566145,-0.01132771,-0.0458152816,0.3190466464,0.1376358122,-0.1362697035,-0.3025194705,-0.3942960799,-0.1327105612,0.4103940725,0.1652321368,0.1351452023,-0.1027042121,-0.5430188775,0.02905339,0.0596098006,0.3160614967,0.106830433,0.0354208723,0.002879529,-0.2154764831,0.1601435542,0.0273871943,0.6485620141,0.1067196131,0.1607167125,-0.1940805614,-0.6553233266,0.4583485425,-0.6851570606,-0.3094448149,-0.1223158091,0.176112324,0.1354708672,-0.2074052691,-0.7160696387,-0.0106192995,0.5319483876,-0.1739033312,-0.0766740143,0.3915344775,-0.0429109409,0.0192743633,-0.2680866122,0.6225160956,0.4484646618,-0.348552525,0.1075245813,-0.3683794439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2052","title":"Timit_asr dataset repeats examples","comments":"Ty!","body":"Summary\r\n\r\nWhen loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same\r\nSteps to reproduce\r\n\r\nAs an example, on this code there is the text from the training part:\r\n\r\nCode snippet:\r\n```\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\ntimit['train']['text']\r\n#['Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n```\r\nThe same behavior happens for other columns\r\n\r\nExpected behavior:\r\n\r\nDifferent info on the actual timit_asr dataset\r\n\r\nActual behavior:\r\n\r\nWhen loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same. I've checked datasets 1.3 and the rows are different\r\nDebug info\r\n\r\n Streamlit version: (get it with $ streamlit version)\r\n Python version: Python 3.6.12\r\n Using Conda? PipEnv? PyEnv? Pex? Using pip\r\n OS version: Centos-release-7-9.2009.1.el7.centos.x86_64\r\n\r\nAdditional information\r\n\r\nYou can check the same behavior on https:\/\/huggingface.co\/datasets\/viewer\/?dataset=timit_asr","comment_length":1,"text":"Timit_asr dataset repeats examples\nSummary\r\n\r\nWhen loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same\r\nSteps to reproduce\r\n\r\nAs an example, on this code there is the text from the training part:\r\n\r\nCode snippet:\r\n```\r\nfrom datasets import load_dataset, load_metric\r\n\r\ntimit = load_dataset(\"timit_asr\")\r\ntimit['train']['text']\r\n#['Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n# 'Would such an act of refusal be useful?',\r\n```\r\nThe same behavior happens for other columns\r\n\r\nExpected behavior:\r\n\r\nDifferent info on the actual timit_asr dataset\r\n\r\nActual behavior:\r\n\r\nWhen loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same. I've checked datasets 1.3 and the rows are different\r\nDebug info\r\n\r\n Streamlit version: (get it with $ streamlit version)\r\n Python version: Python 3.6.12\r\n Using Conda? PipEnv? PyEnv? Pex? Using pip\r\n OS version: Centos-release-7-9.2009.1.el7.centos.x86_64\r\n\r\nAdditional information\r\n\r\nYou can check the same behavior on https:\/\/huggingface.co\/datasets\/viewer\/?dataset=timit_asr\nTy!","embeddings":[0.0989102423,-0.2225528359,-0.0077919294,0.4152596891,0.2339853048,-0.0709775239,0.4454134405,0.1640912741,-0.3221930861,0.1786646843,0.0749746636,0.3964119852,-0.1484193057,0.2935279608,0.1611962914,0.0794432536,-0.0078590345,0.1196177676,-0.4891848862,-0.2469549179,0.1509256512,0.0292406958,-0.0927153304,-0.0054012346,-0.1635167003,0.3077848554,0.0001795032,-0.2649904788,0.0756134242,-0.5593196154,0.2212052941,0.0765711218,0.2061577141,0.4773226976,-0.0001191902,0.0572748482,-0.088580519,-0.0878749564,-0.4351503551,-0.3336096406,-0.2282954454,-0.0018591578,0.4544062912,-0.1512349993,-0.3395612836,-0.2512342334,0.0287082624,0.1518220007,0.1025856212,0.2404614091,0.1510352045,0.2045672089,-0.5535507798,-0.0390319116,0.2077764422,0.0575214848,-0.0956579894,0.1412107199,-0.0615998283,0.3370451629,-0.0436336026,0.3062425554,0.0098433634,0.2862468064,-0.0754556954,0.0575153455,0.137378186,-0.1157972589,0.0900406688,0.3028086126,0.6127233505,-0.1755514145,0.0001597601,-0.205015108,0.2454696894,0.0452232547,0.1403374523,0.2333649844,-0.1640398502,0.1755173057,-0.4493053257,0.2011569738,-0.2176719755,0.0975787193,-0.090035744,-0.4185143113,-0.1893828213,0.1090499908,-0.2871719301,-0.1850624233,0.2887518108,-0.2921885252,0.1885451525,0.0676705018,-0.2829754353,0.1494015306,-0.0358249322,0.2496208996,0.0124437474,0.2042994052,0.7109580636,0.2377070785,-0.130228743,-0.1135727316,0.1076797247,0.2039637417,0.1340145469,-0.0466267429,-0.0739576146,0.070787169,-0.3941363394,0.0274523236,-0.1635030508,0.342387557,0.3735173047,-0.4171498418,0.2156481147,-0.240857482,-0.4846591949,0.198764056,-0.4843815267,-0.2124012709,-0.0764559433,0.1077340767,-0.0186058395,0.2642625868,-0.1238154024,-0.0979389921,0.052294787,-0.1384135634,-0.1931199282,-0.2623009682,-0.2720706463,0.2743638754,0.1755682379,-0.5277848244,0.1612015665,0.4869962037,-0.1236506924,-0.0473747626,0.2312754393,0.0293478593,0.1281853169,0.0485776737,-0.0830232129,0.2257317305,0.067559883,0.2030798793,0.0167736467,0.2044563591,0.0252585933,0.0809128582,0.558957696,0.2177877277,-0.3195416033,-0.1249117851,-0.1390608698,0.1336652488,-0.0640136153,-0.1081591323,0.1181163713,-0.2824503183,-0.1365134418,-0.1952818781,0.0364736542,0.3270919025,-0.3575615287,-0.155170545,0.3366031647,0.0815942362,0.2026403844,-0.0166164041,-0.0358653925,0.2803322077,-0.1282074153,-0.0008115674,0.0686848313,-0.2312913388,-0.2643617392,0.1933108866,-0.0410825834,0.4655153751,0.0438708104,-0.0481862873,0.3766214848,0.3783048987,0.1365660876,0.0103711477,0.3914209008,-0.038897004,-0.4284938574,-0.0250852872,0.226893574,0.1737152785,-0.1589303762,0.2444460541,0.0696491301,-0.0345707461,0.3851392865,-0.0988942161,-0.1669492275,-0.1770093739,0.0899295956,0.0953073427,0.1621816456,-0.3506909609,-0.0949921533,0.004777737,0.2483480573,0.177991569,0.4129677713,0.0180270541,-0.3658150434,-0.2651265562,-0.4224319458,-0.0809061527,0.1220720038,0.517272234,-0.0123230275,0.2241945863,0.0371391363,0.6469510794,-0.48562482,0.1167731062,-0.0534782819,0.2008742988,0.0789700374,-0.0021116894,-0.0342812687,0.0703206807,0.1733400822,-0.1211784855,0.1000722423,0.254283309,0.0814893618,-0.0504701175,-0.2694038153,-0.2033138275,0.2039894611,-0.1976954788,-0.1077783406,0.2105886042,0.1562798917,0.0126118036,-0.0823544487,0.307595104,0.0569566041,0.1766377687,0.2195506543,0.0256764404,-0.0049767569,-0.0548202768,-0.3573397398,-0.4050461054,0.3902083933,-0.0015297043,0.1111840531,0.1847780943,-0.3534191847,0.2469746768,0.1177940741,-0.1868613213,-0.1525240242,0.1231411919,-0.1690305024,0.1976415813,0.0661403313,0.3433302045,0.2653207183,0.186935395,0.0779193714,0.0303519499,-0.2613981068,-0.2293583602,0.1644851267,-0.0093328506,-0.281353265,0.4067205489,0.2671443224,-0.032505054,-0.5773596764,-0.0789780319,0.2472221404,0.2488528788,-0.2368168533,0.1227146089,-0.4884318113,-0.0248017982,-0.3066206276,-0.1918983907,0.164310053,0.0456015132,0.0185904838,0.131919533,0.2243712097,0.1767510921,-0.3980655968,0.2499869317,-0.0378539599,0.3858734965,-0.0362369083,0.254566133,-0.4002576172,0.0595241711,-0.1190915033,0.1099402532,0.2069377154,-0.5000792742,0.0614441857,-0.2703506052,-0.2949748933,0.4115564227,-0.1248464212,0.4184917808,0.1727770716,0.05239629,-0.1363333166,-0.2187476009,-0.0222437195,-0.0902913362,-0.2241424173,0.1192703024,-0.0338497087,0.0742322579,-0.1728588194,-0.5758616924,0.078215979,-0.1269825697,0.3518572748,0.1941635162,0.1768839061,0.4906210899,0.0106065273,0.3438171744,0.1771775186,0.2350356579,-0.4492761791,-0.2054259032,0.2069892734,-0.0229329839,-0.1319351792,-0.1705038249,0.0153770763,0.0505377762,0.3169457614,-0.6579898,-0.0109850504,-0.2257585078,0.0923368707,0.032107193,0.0537613481,0.2059143186,-0.0928353965,-0.0315529592,-0.1436258107,0.0383677036,0.2231892496,-0.0494642518,-0.0522198752,0.0252683349,0.1436446309,0.0979187265,0.208357513,0.3998866379,-0.0991097018,0.2671526968,-0.1014767364,0.4371002316,-0.1137849912,-0.1344069988,-0.1619121879,-0.0969095379,-0.0592032447,0.2511379719,-0.1137691289,0.1284635663,-0.0234517865,-0.0390670672,0.0751952901,-0.2449523658,-0.1833536923,-0.4072388411,0.330765605,0.2468751222,0.1608939618,-0.0286245663,-0.1569230556,-0.3084007204,0.2602107227,0.1162038296,-0.0744653344,-0.7141475677,-0.1906746179,-0.413327992,0.2560719848,0.3304316998,0.378285408,0.1281196177,-0.3584359288,0.4207448363,0.0445677303,0.4862847328,-0.478543222,-0.0561886579,0.0976721495,0.0730924159,-0.5091207027,-0.2434050441,-0.4101661742,-0.0820504725,0.1360016912,0.3206499517,-0.1147220731,-0.2352579534,0.2853874862,0.2641364634,-0.1053103656,-0.2099864632,-0.0916340351,-0.0555807613,0.1091987267,-0.0389140882,-0.2558635771,-0.1527946144,-0.4663057029,0.0184533689,-0.3065110743,0.2326108813,0.4977618456,0.2560831606,0.4027290642,-0.0160086397,0.2788285911,0.2585754991,-0.0349566229,0.1576760113,0.5241253972,-0.361472249,0.0439837389,0.1972373128,-0.2992482185,0.1092974544,0.4790363908,-0.0477608629,0.2122824192,-0.203897655,-0.0588562638,-0.0940220729,0.2247219533,0.270649761,0.0104719456,0.101708293,-0.6981610656,0.2175437659,0.0123314792,-0.2895797491,0.2782367468,0.0002484362,-0.1160616651,0.1829815209,-0.2896981239,0.8095844388,0.1427437067,0.527807951,0.4136286676,0.0363412946,-0.0149061838,-0.1198458076,0.0764826164,-0.1932017207,-0.2132991105,-0.131424278,-0.1944511384,-0.0489592627,0.1546222717,-0.3623226881,0.0220419019,-0.0660019964,0.2353173941,0.1359341592,-0.0264599826,-0.2268809825,0.1499749869,-0.068032667,0.0686676204,0.1199299097,0.081077531,0.1006339565,0.1691439301,0.2840334475,-0.2867975533,-0.3098422289,0.0562279224,-0.1185266972,-0.1240215451,0.0340433531,-0.5667598844,-0.0996623263,0.0209938567,0.0165710095,0.0749429017,0.0622274354,-0.2317612171,0.2040733099,0.3222239912,-0.2157096863,-0.1128893271,0.5602844357,0.2369358242,-0.3686026931,0.1829148531,0.2361863256,0.0147397602,-0.40084216,0.1892077178,-0.237557739,-0.6892206073,0.0405782163,0.0556914359,-0.0836302787,-0.2031555027,0.0897929668,-0.3696023226,-0.1696904153,0.2380331308,0.2827782333,-0.4386645555,-0.1538320035,0.3941916823,0.419591099,0.0232370365,0.4874532223,-0.0748060122,-0.0696111396,-0.2655162215,0.1134376898,0.2071311474,-0.1971130073,0.0511918664,-0.2652720213,0.2075986415,0.2318050563,0.2292132527,0.0377188511,-0.0495323129,-0.3496434987,-0.2421807945,0.0544518344,0.1099661663,0.2450587004,0.3274364471,-0.0954241231,0.2693485022,0.012948025,-0.1217917949,-0.2322159261,0.3120190799,0.0859223008,0.2491950691,0.0120492512,-0.2695143521,0.2401782423,-0.0035014318,0.0997845456,0.203269124,-0.1520634294,-0.1329098642,-0.3474862278,0.1211761311,0.0492925718,0.0725010708,-0.2247762084,0.02839409,0.1514493376,-0.0059539122,0.2308398187,0.2081442773,0.0242882073,0.2081385404,-0.0714544281,0.2072908878,0.0594607741,-0.1902016699,-0.1332113743,0.0332512781,0.3320817053,-0.1389742196,-0.2847731113,0.1399036348,-0.3334291577,0.0989910066,0.2167953849,0.0089112697,0.0147330379,-0.2345357239,0.0277557634,0.0203106534,0.2228631973,0.5198347569,-0.1981296092,-0.3495951593,0.0087255426,0.1226777956,-0.088503316,-0.1142819226,-0.1556088775,0.0569224171,0.1670891643,0.3259461522,0.0162431598,-0.4734797776,0.3968091607,-0.1219186187,0.2920052707,-0.2440853417,0.1959059387,0.4127357602,-0.0413652137,-0.0966035351,0.2172791064,0.0726800114,-0.2451894879,0.5598279238,-0.0131778074,0.0917783827,-0.2586664855,-0.0469019935,0.0822780058,-0.1040256172,-0.0732567757,0.2454630286,-0.2353817075,0.3328186572,0.141903311,0.2974191308,-0.0073497528,-0.0226254296,-0.3484489322,-0.2988160253,-0.2394945771,-0.1696675867,-0.2909618914,-0.4145784378,-0.025740914,0.026929792,-0.2175788432,0.1000406072,0.5003553033,0.158965826,0.0418757685,-0.6586124301,-0.2175093442,0.075004749,0.2171880603,0.0302091185,0.4967205524,0.1225694194,0.0460319333,0.1053960621,0.1365808696,0.2748981714,0.2569794357,0.0420100391,0.074989602,-0.3617217541,0.041999612,0.0290400162,0.3255832791,0.0486902967,-0.2769916952,0.1208695173,0.0429728031,-0.0781551749,0.3282752931,0.1404009461,-0.0888405293,-0.8085055351,0.3766825795,0.0167517401,0.2309182286,0.0724171624,-0.0024892166,-0.4760936499,0.4893349111,0.3303557932,0.0595098697,0.1220439747,-0.184856683,0.083704561,-0.2804265916,0.2545168102,0.1019514352,0.0080569843,-0.1584173739,-0.1007545665,-0.6602063179,0.0862901136,-0.267347306,-0.0576224588,0.1119692698,0.0709862188,-0.024695985,0.0547964647,0.1556333452,0.1264126301,-0.4332690239,0.0720109642,-0.4466657937,-0.1716458201,0.1123188883,0.1215180084,-0.1177549362,-0.178637296,0.0440857373,-0.1012476906,-0.0162522681,-0.0400270037,-0.061877016,0.0090026353,0.2595134377,-0.033575464,-0.0384142734,0.354945451,0.1310971975,-0.1109541878,-0.3006770909,-0.3420492709,-0.0941353738,0.4512909055,0.1183778718,0.0775769427,-0.0791893452,-0.535911262,0.0877194256,0.0200177506,0.3185458183,0.0874574333,0.0338928998,-0.0119595909,-0.24069646,0.1622124463,0.0022928191,0.6019895673,0.1000052169,0.1067515165,-0.1704578549,-0.6369557381,0.4100932479,-0.7224936485,-0.3515584767,-0.2280663699,0.1739075184,0.0768777356,-0.1217112467,-0.6946240067,0.0135344435,0.5767679214,-0.2107257992,-0.0747160092,0.3796476126,0.0407356545,0.067220144,-0.2712350488,0.6195218563,0.3928466737,-0.3472814262,0.0148643516,-0.3914866447]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2050","title":"Build custom dataset to fine-tune Wav2Vec2","comments":"@lhoestq - We could simply use the \"general\" json dataset for this no? ","body":"Thank you for your recent tutorial on how to finetune Wav2Vec2 on a custom dataset. The example you gave here (https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2) was on the CommonVoice dataset. However, what if I want to load my own dataset? I have a manifest (transcript and their audio files) in a JSON file. \r\n","comment_length":13,"text":"Build custom dataset to fine-tune Wav2Vec2\nThank you for your recent tutorial on how to finetune Wav2Vec2 on a custom dataset. The example you gave here (https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2) was on the CommonVoice dataset. However, what if I want to load my own dataset? I have a manifest (transcript and their audio files) in a JSON file. \r\n\n@lhoestq - We could simply use the \"general\" json dataset for this no? ","embeddings":[-0.0622440875,0.0348938331,0.0137573285,0.1445438415,-0.1110931709,0.1218437478,0.0940706432,0.0563694164,0.1314444244,-0.0642109886,-0.1219049841,0.2266791314,-0.0514387228,0.2130797654,0.0829685479,0.1270682067,-0.1625839323,0.2323298007,0.2772756219,-0.0528926738,-0.2511222959,0.1114695594,-0.0345702246,-0.0587337501,-0.2246275097,0.1605609655,0.2075202614,0.4153412282,-0.2021602839,-0.3214997947,0.5517761111,0.3358494043,0.2289199382,0.1611868441,-0.0001115951,-0.1021808013,0.0438133813,-0.3512279987,0.1741255373,0.2562049031,-0.5536302328,-0.0819852427,-0.1176473573,0.0017799068,-0.339825958,-0.1086229682,-0.2092565298,-0.2648890913,0.6307888031,0.2754027545,0.1267050952,-0.4373144209,-0.073522009,-0.0770784393,0.0216000061,0.0839470178,0.049499698,0.4411941171,0.3102349341,0.1237346828,0.1407768279,-0.1033120379,-0.235068351,0.0683089942,0.4571740031,-0.0240661222,-0.4745515287,-0.1568822563,-0.0371970423,0.5689551234,0.6908248067,-0.2889418304,-0.0930363014,-0.1065931618,0.2058357745,-0.4316125214,-0.0701020211,0.0247985274,-0.1270194948,0.391461283,-0.0334293321,-0.5124310851,-0.2786418498,-0.0642879978,-0.0654524118,0.045004718,-0.1668723375,0.0572302751,0.0265865251,0.0004292626,-0.1039089188,-0.0561822243,-0.3513596654,0.1819574088,-0.4081173837,-0.4492120147,-0.1753094345,-0.3259967268,0.1090601608,0.0710522309,0.3910751939,0.1233142093,-0.1460694373,0.0555158928,0.1452013552,-0.1166857779,0.0779340416,-0.1966766119,0.0800614059,-0.5448824763,-0.0293127745,0.0011295242,-0.1765591502,0.1012766361,-0.391640991,-0.0769940987,0.1183329374,-0.4141617119,-0.2937143147,-0.0362717472,-0.0552935377,-0.0269142669,0.0046670926,0.3715262711,0.2542221248,-0.0660165772,-0.2318949252,0.4409132004,0.2935177684,-0.115709424,0.058518976,0.0736122429,0.3282890916,0.2488210052,0.3763130307,-0.0949520394,0.4115787446,-0.2405443639,0.2811505198,-0.1440054774,-0.0704793483,0.0513712987,-0.3219588101,-0.0250865892,-0.0982505828,0.1443459392,-0.0844772607,0.0896202549,-0.3642826676,0.0388284065,-0.3370604813,-0.4700230658,0.0813912675,0.1147800609,-0.2617973685,-0.2272800803,-0.171438396,0.1435775012,-0.1606439203,-0.2710344195,-0.3427544832,0.0888163671,-0.2954405248,-0.189347446,-0.0471798964,-0.1530841589,-0.8702062368,0.1764576286,-0.3130331933,0.2046673596,-0.0904310718,0.325887382,-0.1906788945,0.5594411492,-0.0591556542,0.3236888945,0.2820644677,-0.3932394385,-0.3187333047,0.3751401603,-0.2023470998,0.0127085922,0.1722374856,0.0860643461,0.3811219931,-0.3684342504,0.1641567945,0.2392026782,0.1475247145,-0.1203168631,0.2582568228,-0.2207168043,0.2021354437,-0.0036473451,-0.3254508078,0.053275276,-0.043230325,-0.0410565808,0.1577410698,-0.1638382971,0.0237075649,0.1018317714,0.2625572681,-0.2639671266,0.073049739,0.1371962726,-0.2139312774,0.213296473,0.0254509076,-0.1044961959,0.590323925,-0.0330249555,-0.3159267306,-0.2544132471,-0.13036336,-0.2772693038,0.0724314973,0.1420301497,0.219718352,-0.6761963367,-0.5398189425,0.1300341189,-0.2265326381,-0.1379926354,-0.1368346363,-0.037011262,0.4340130091,0.1023076698,0.3576132953,0.2732828259,0.076285027,0.1172831878,0.0658076704,0.3764742911,0.1438181251,0.3986373246,0.2814199626,0.2750840783,-0.1979035437,-0.5793613195,0.4434407055,0.2860706449,0.1679250598,-0.1349300593,-0.3361740112,0.3128340244,0.2373643667,0.3019483089,-0.2731260061,-0.2158028483,0.3199341595,0.0693350136,-0.2557486892,-0.0282825492,-0.1152599379,-0.0084751146,0.4955792129,-0.096575357,-0.1997408569,0.0293704588,0.2016980797,-0.0680193007,0.0599929467,0.194260627,-0.398099333,-0.2089958936,-0.0915945396,0.1554207951,0.482803911,0.2838301957,-0.126572147,-0.0724340379,0.0083229896,-0.2132435888,0.1696165949,-0.1818493605,-0.045319546,0.2405308634,0.044819776,-0.1867560595,-0.1614674181,0.0519280881,-0.2623183429,0.0361020789,-0.3185086548,-0.4315363467,-0.0771742463,0.3814121485,-0.136012271,-0.2498396635,-0.0836029202,0.238837108,0.1787195504,0.0119619053,0.0992206782,-0.0609161332,-0.1042661145,0.654455483,-0.0919092596,-0.3881019354,0.1036615372,-0.0712919459,0.04333736,0.192459926,0.0833218917,0.1847088635,0.1367884725,0.129388988,0.1154852435,-0.0329451673,0.2478087395,0.093387343,-0.1548098773,0.1764987409,0.1266334057,0.2903391719,0.3003922999,0.0699932426,-0.2194239795,0.266815871,0.0397422165,-0.1797085106,-0.0008618116,0.2149640322,0.2517831624,-0.414952457,-0.0702808723,-0.2755764127,0.3077764809,0.2238498181,0.048767481,-0.2955520153,-0.0250034425,0.0661455691,0.0383924134,0.226358369,-0.3296994269,-0.3670618236,0.3647490144,-0.131808877,-0.0342096016,0.147053808,-0.0537902527,0.3338879645,-0.0884330496,-0.2237657309,0.3075970113,0.1795708537,-0.0635742024,0.1408860832,-0.1421521753,0.2542923391,-0.0874245539,0.047035642,-0.2961481214,-0.3052161932,0.1143367067,0.4581757188,0.1024644673,0.3942191899,0.0159990117,-0.6823664904,0.6610146165,0.1823385805,0.0492970273,-0.0116568431,0.0214157477,0.2038580477,-0.1298983246,0.1313479692,0.2279751748,0.2235337198,-0.155727461,0.4360446632,0.2634658217,-0.0164682064,-0.1682202369,0.2613985538,-0.3682889342,0.0149940951,0.1195868552,0.0864165053,0.0953177437,0.0629568994,0.3327549994,0.1474848837,-0.1976315379,0.0999559611,-0.0095401369,0.3675438166,-0.2454436123,-0.2599685788,-0.1079252362,-0.3146162927,0.2712981105,-0.3245955706,0.338685751,0.0229675304,-0.022758754,-0.1014799252,-0.1461506039,0.280287385,-0.4390376806,-0.0030876598,0.0131691433,0.3138356209,0.0455427058,0.2395610809,-0.0955081955,-0.0783158615,-0.3915884793,-0.0026793426,-0.071193479,-0.079074055,0.0402003527,0.2717010379,-0.2562013268,-0.3223455846,0.2077310532,-0.007121047,-0.1591716856,-0.3196695149,-0.2795785666,0.0264992584,-0.2183170915,0.0822089612,-0.1039249897,-0.0875858516,0.4358543158,0.1155502498,0.3398869932,0.2098990381,0.1874721199,0.1525720656,-0.0761663541,0.1328156143,0.4705693722,0.3234003484,-0.3085997105,0.0047933324,-0.0604211465,0.3731178045,0.1793788224,-0.0393166728,0.0544137657,-0.1140868068,-0.2686268985,-0.437197417,0.0309839826,0.1368802339,-0.154069826,-0.2207272947,-0.4605407715,0.6691031456,-0.0083559584,-0.1676024944,-0.1396014988,-0.1645568311,-0.4388115704,0.237591818,0.2870748341,0.7743760943,0.0856454,0.064801842,0.2668553889,-0.3107101321,0.2440307587,-0.773185432,-0.1547762901,-0.308249414,0.1378377378,-0.2642838657,0.0478445105,0.043278832,0.5231707692,-0.0033907811,0.2848709226,0.282853514,-0.149943158,0.0798148811,0.0983596593,0.16319561,-0.4263272285,-0.1817080677,0.0938970894,-0.1275310814,0.0158132333,0.0073504634,0.0703866556,-0.0905912817,-0.337901324,-0.3221264184,-0.1495656818,0.0715887919,-0.3642689288,0.0095081758,-0.3359204829,0.2523986101,0.070785597,-0.0901584253,0.2883352637,-0.5456092358,0.1920564324,0.118761912,0.0221812744,-0.0381616466,0.0155257965,-0.227780968,0.0022963814,-0.0000864451,0.3819089532,-0.0773354098,-0.2689729035,-0.0400069393,-0.0933185518,0.1024105921,-0.1545493603,-0.5847405195,0.2917191982,0.2206373662,-0.1031211168,0.0930254683,0.3974840045,-0.2939286232,0.2432529628,-0.0279309154,-0.267275095,-0.1009155437,-0.0541147925,0.2685643733,-0.2596722543,0.3028131723,-0.0871045217,0.2019279599,-0.3515237272,0.4162255526,0.2948209047,0.0058591617,-0.1979033947,0.4478189051,-0.1559180766,-0.1201280951,0.1577201784,0.137797147,-0.0116866082,-0.0459843986,-0.484233588,-0.4056857228,0.1098498628,0.1812372059,0.3497460186,-0.2256917804,-0.1258194,-0.295130372,0.1168172434,-0.2997504771,0.0132723833,0.231422931,-0.0534125753,0.317538172,0.6285011172,-0.0909154117,0.1093205288,0.0786798671,-0.0939413533,-0.3790251911,-0.1434861422,-0.0389690883,0.2045470923,0.1247084886,0.0446030609,0.0176092405,-0.0122351404,-0.2549827993,-0.0633187741,-0.0287121125,-0.1388171315,0.0733094662,0.1349899471,-0.0170281436,-0.1497570425,-0.3035832047,0.1161439344,0.1484255195,0.2585057914,0.1550180763,0.1761817932,-0.1164068058,-0.0264815502,0.3834628761,0.4018870592,0.107033588,-0.0042466484,0.1659982353,-0.0460695587,0.2107677609,0.1925647706,0.1824074388,0.5378133655,-0.2504273653,0.0145957293,0.4011613131,0.194269985,-0.191045925,0.1603448987,0.1411920637,-0.1393624991,0.0728636831,0.296197176,-0.0713527277,-0.1101366952,-0.3351210356,-0.1433361471,0.1520894617,-0.1098548472,-0.0253222119,0.2999508977,-0.1137220263,-0.1653693467,-0.1145576462,0.1533523947,0.2077518702,0.3254540861,-0.1907507181,0.0576058589,0.1745031625,0.042619016,0.0193732288,-0.4339895546,0.1377586126,0.4450004697,-0.0863045976,0.1247905791,0.2217563093,0.1306287795,-0.0299198404,-0.0454234071,-0.1887218207,-0.0026844929,0.1505831778,-0.3173789382,-0.2321218401,-0.2290908843,0.0848131776,0.1203436777,-0.2306001186,-0.1457431763,-0.2323751301,-0.306659013,0.0556036532,-0.0770387352,0.2853402495,-0.1671963483,-0.3297420144,-0.1182327345,0.3357230425,-0.136029318,0.0954639614,0.6305387616,0.2327859551,0.0797630996,0.0112612871,-0.162111938,0.0978514403,-0.0420623235,-0.0891949907,0.028768247,0.2993565202,-0.069160223,-0.0138495853,0.3349553943,0.1624187529,-0.0278548021,0.1292009801,0.1678327024,-0.0669388175,-0.2974166274,0.0841191113,0.2122238129,-0.1245621592,0.1719578356,0.0331498608,-0.2018283308,-0.311940372,0.2635442615,0.2810539901,0.1167590171,0.2859294415,0.0967845544,-0.2964444458,0.4196895957,0.0194610972,0.1429789811,-0.1346838772,-0.1288335323,-0.2677492201,0.1004765481,-0.0320313349,-0.1709441245,-0.2985166013,0.0186018273,0.6413117647,0.5423722267,-0.050393071,-0.0104589239,-0.0702777058,0.3513939977,-0.0215184838,-0.0708295926,0.2639291584,0.6188367009,-0.0354541317,-0.3405914009,-0.2729951143,0.3316335976,-0.072405614,0.1155157015,0.1910111159,-0.0261661895,-0.2818938792,0.4617296457,-0.2121486962,0.5100897551,-0.1785549372,0.0596062839,-0.1096858233,0.0792141408,-0.1681080014,0.0509322844,0.016661955,0.3457073569,-0.2420514971,0.2226743102,-0.1398672014,-0.3308929503,0.0361929461,-0.0680915862,-0.3529877961,0.1466756463,-0.4273989797,-0.0539233573,0.0185080189,-0.023330804,-0.0146631813,0.0569617227,0.0394232571,0.1463472098,0.6421450973,0.1769835353,-0.0508721322,0.2210002095,0.4158798456,-0.1280393451,0.0627465919,-0.3103792369,0.2994474173,0.3640410602,-0.0458087139,0.1147541627,-0.0027868722,-0.1670216173,-0.2081725895,-0.0788631141,0.2228658348,0.0574361235,-0.1231062189,-0.3782861829,-0.3550027609]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2050","title":"Build custom dataset to fine-tune Wav2Vec2","comments":"Sure you can use the json loader\r\n```python\r\ndata_files = {\"train\": \"path\/to\/your\/train_data.json\", \"test\": \"path\/to\/your\/test_data.json\"}\r\ntrain_dataset = load_dataset(\"json\", data_files=data_files, split=\"train\")\r\ntest_dataset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n```\r\n\r\nYou just need to make sure that the data contain the paths to the audio files.\r\nIf not, feel free to use `.map()` to add them.","body":"Thank you for your recent tutorial on how to finetune Wav2Vec2 on a custom dataset. The example you gave here (https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2) was on the CommonVoice dataset. However, what if I want to load my own dataset? I have a manifest (transcript and their audio files) in a JSON file. \r\n","comment_length":51,"text":"Build custom dataset to fine-tune Wav2Vec2\nThank you for your recent tutorial on how to finetune Wav2Vec2 on a custom dataset. The example you gave here (https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2) was on the CommonVoice dataset. However, what if I want to load my own dataset? I have a manifest (transcript and their audio files) in a JSON file. \r\n\nSure you can use the json loader\r\n```python\r\ndata_files = {\"train\": \"path\/to\/your\/train_data.json\", \"test\": \"path\/to\/your\/test_data.json\"}\r\ntrain_dataset = load_dataset(\"json\", data_files=data_files, split=\"train\")\r\ntest_dataset = load_dataset(\"json\", data_files=data_files, split=\"test\")\r\n```\r\n\r\nYou just need to make sure that the data contain the paths to the audio files.\r\nIf not, feel free to use `.map()` to add them.","embeddings":[-0.0861028954,0.0252731182,0.0131383371,0.0565956719,-0.0154160764,-0.0016975594,0.040185865,0.1672806442,0.1113969013,-0.0323929042,-0.2139553726,0.4168869257,-0.2158389539,0.1142557338,-0.0266242605,0.0259716772,-0.1856243163,0.225793615,0.1270139068,-0.0297930297,-0.2303095013,0.150539577,-0.0891870037,-0.0707921013,-0.1212103069,0.1165201217,0.177677691,0.4099347293,-0.1358955055,-0.3398261666,0.407466501,0.3265336156,0.1645614952,0.3329025209,-0.0000988554,-0.1569031477,-0.0146500124,-0.2888993621,0.0687238723,0.1292687804,-0.4952058494,-0.1590586156,-0.1031423062,-0.0598484576,-0.2974746227,-0.1860620081,-0.298794359,-0.2663181126,0.7216199636,0.3141840994,0.2168492973,-0.3120064139,0.0287783369,-0.0162665453,0.0748966634,0.0478535108,0.066221267,0.4144843817,0.2814677954,0.0652268752,0.2580654919,0.043904271,-0.3099857569,0.1127453893,0.265550971,-0.0027987545,-0.4111703336,-0.2162590772,-0.0957747549,0.577278614,0.4842572808,-0.5012128353,-0.1985921413,-0.189050734,0.1999863088,-0.4806697965,-0.0414562412,0.0458245836,-0.1702143699,0.3052021563,-0.0256954655,-0.3619791865,-0.2076643556,-0.0906062201,0.0454348624,0.0527963378,-0.1664787829,-0.0589325055,0.042361144,-0.051307749,-0.1957069039,0.0393166207,-0.2192359418,0.3246355355,-0.4265645444,-0.3958183825,-0.095357798,-0.4423437715,0.0504589006,0.0338856839,0.3694809675,0.143151626,-0.1174625009,0.1512011141,0.0218686815,0.0106418319,0.0032325401,-0.1732013226,0.0986957997,-0.432667017,-0.0350537896,0.0264260657,-0.2180433124,-0.0432572737,-0.3427793086,0.0566503368,0.148542583,-0.3140284121,-0.3048003018,-0.0890865549,-0.1365513057,0.0110783745,0.0665753931,0.4089030921,0.1010563821,0.0541140921,-0.056967169,0.3980821073,0.1612128168,0.0188260544,0.0439640395,0.2589755356,0.3334704638,0.1825697869,0.3748436272,0.0000860344,0.4036771059,-0.1982594281,0.2077446282,-0.1387869567,-0.0327320136,0.133154735,-0.2826277316,-0.0561921336,-0.0048104748,0.1839497387,-0.1019115523,0.0315981098,-0.3335483372,-0.0936334953,-0.3694884479,-0.3046224117,0.1110639796,0.2222231477,-0.2240485251,-0.1587861031,-0.1389620304,0.0635609925,-0.1052223668,-0.2892745733,-0.2854251862,-0.0336191878,-0.26550439,-0.1619248539,0.1819533408,-0.0987349972,-0.6757566929,0.0897569954,-0.1216203794,0.1931489855,0.0001066574,0.2764821649,-0.0473640747,0.5055429339,-0.0960018933,0.3563430011,0.2618568242,-0.4331692457,-0.2342766672,0.3278563023,-0.1937007606,0.1395905763,0.1119498685,-0.0538505726,0.4180216789,-0.3144515455,0.2169784904,0.4147839248,0.1870000362,0.0101187285,0.1476110518,-0.0495184697,0.1854101121,0.0495967306,-0.3701095283,-0.0695197806,-0.1135778651,-0.0335274599,0.2687246799,-0.1552427709,0.0629094094,0.18053855,0.2452691048,-0.2102152556,0.0133582,0.1091393232,0.0447729416,0.1026234776,-0.1104324013,0.005972906,0.34248209,0.0717757791,-0.2754268646,-0.2076176256,-0.3448870778,-0.3608686328,0.185911119,0.1474171728,0.2212572396,-0.4752925038,-0.4122073054,0.0807687119,-0.0943108499,-0.1427309811,-0.2458337843,-0.0367690213,0.3501383066,0.007016798,0.3941415846,0.3947726488,0.1325919777,0.1408108473,-0.0319033936,0.4642842114,0.2063316107,0.3702318966,0.1435026079,0.2380604744,-0.1249080598,-0.4643808603,0.3657501638,0.3537258804,0.2512151301,-0.076651141,-0.3008253574,0.3731018305,0.2522552609,0.3634337187,-0.0481414311,-0.1168700159,0.3635115325,0.1261614561,-0.2628050745,0.1092672423,-0.091446951,-0.0808728561,0.4064539373,-0.2065781206,-0.0683862567,0.0845374316,0.2053621113,-0.0750657693,0.0809962749,0.1536124796,-0.4257684648,-0.1188572049,-0.1134339124,0.1789563149,0.4203054011,0.3598633707,-0.0708565563,-0.0662555024,-0.0526872054,-0.1265423447,0.1080329344,-0.1157122329,-0.0939975083,0.3089731336,0.1058522537,-0.1834134459,-0.1501322389,-0.0604770221,-0.3132826686,0.0747417882,-0.3948696256,-0.4067824483,-0.1029475257,0.3815223277,-0.1875569969,-0.354523629,0.0532795675,0.1806257963,0.2497491241,0.0836810991,-0.0780070722,0.0876738206,0.0497489795,0.4142825902,0.0394403823,-0.3689314127,0.2387041748,-0.0507123694,0.078801319,0.2575671971,0.1206863075,0.2905208766,0.113808386,0.2241352201,0.091978021,0.040780127,0.1982498318,0.0474724695,-0.1298720986,0.1506040394,0.0200970359,0.2760104239,0.3697803617,-0.1345028281,-0.1777005047,0.1584922373,-0.0070836954,-0.0365135446,0.0327608846,0.0971892178,0.1346780211,-0.5730341077,-0.2228899002,-0.401925534,0.3425261378,0.1432172358,0.0825379714,-0.2697078288,0.0670059398,0.059599556,0.0927966088,0.316563338,-0.2320633829,-0.2677368522,0.477547735,-0.2591694891,-0.0345867649,0.1361819655,-0.105293721,0.3692614138,-0.1506731361,-0.0477362685,0.2083643675,0.0922414288,-0.1593067646,0.1981938481,0.0033950168,0.299620986,0.0201970898,-0.0316830464,-0.2138721496,-0.1847884208,0.139983505,0.3487140834,0.0934473425,0.2376148254,0.0113632325,-0.6527879238,0.5451339483,0.2003678381,0.1047428548,0.0255910363,-0.1037690639,0.0688263476,-0.114126429,0.0598314591,0.1120026708,0.063936159,-0.1524635255,0.4134088755,0.3461427093,-0.0516759418,-0.1652644426,0.1413024217,-0.4106563628,-0.0305017345,0.1508205235,0.0608626381,0.0484407209,-0.0202791505,0.3535822034,0.1242536008,-0.1363382339,0.1239679232,-0.0720518455,0.3260033131,-0.2398374379,-0.2358723134,-0.198716417,-0.3337763846,0.1443874538,-0.1739084423,0.2854415774,0.0727062076,-0.105820246,-0.1804712564,-0.1442708522,0.287953794,-0.3175036609,-0.0472839139,0.0281100776,0.1946291178,-0.0486725084,0.1891381741,-0.2258087397,-0.1626952142,-0.3231976628,-0.0086010406,-0.1487406939,-0.0403890125,-0.035936676,0.3147672713,-0.1958138794,-0.2717557549,0.1249785274,-0.0154530536,-0.2908768058,-0.2386449873,-0.2708975077,0.0667561516,-0.3513464034,0.0070519946,-0.0032194061,-0.0408312008,0.4304586053,0.1467397958,0.4306105375,0.162155062,0.2497423589,0.1420546919,-0.1393047124,0.0920069069,0.4196506441,0.134537369,-0.2270633131,0.0326280817,-0.050733529,0.291241765,0.064162381,-0.092099987,0.1154699028,-0.1916182488,-0.1043755636,-0.4369693696,0.1850703508,0.2909200788,-0.1390722096,-0.09829247,-0.5546112061,0.5026655197,-0.0573655404,-0.1340867728,-0.1058608592,-0.1697183698,-0.369712919,0.3407267034,0.0301523246,0.6115977168,0.1831113547,0.0765376091,0.4470015168,-0.2707526684,0.1296691149,-0.7494617701,-0.0601621158,-0.2570714056,0.2156748027,-0.1564163715,0.0875767842,-0.1179152951,0.3387053311,-0.1610487252,0.2530444264,0.2300137877,-0.2307197154,-0.0164558887,0.1258823425,0.1463291794,-0.4770156443,-0.332991302,0.2037638724,-0.0281247254,-0.0517737903,0.1280700862,-0.0601573288,-0.1179044172,-0.3157678843,-0.2466215938,0.021975087,-0.0180837233,-0.2330807149,0.0067215511,-0.2160553336,0.1298190057,0.152584374,-0.2598616779,0.341121614,-0.4805778265,0.1543211788,0.0876806602,0.1264079362,-0.0937596634,0.0119871292,-0.0926353708,0.0309848636,-0.0102693895,0.3972669244,-0.0559796318,-0.220926255,0.1096413285,-0.1084759608,0.0888477936,-0.111310564,-0.4517171085,0.2794417739,0.0820836052,-0.1447937191,0.1789726615,0.3935125768,-0.4975090921,0.2964437008,-0.0663139671,-0.2111898363,-0.0223858748,0.0383559018,0.1171052083,-0.1720434725,0.4663717151,-0.0123501569,0.1118691042,-0.4094527364,0.2964843512,0.2564250529,-0.0312091522,-0.1114209443,0.4296889305,-0.1672208458,-0.0986037105,0.0839874819,0.0439674072,-0.0571219474,-0.0933861509,-0.5307201147,-0.4021138549,0.0915457904,0.2293607146,0.3262863755,-0.262968719,-0.0243994594,-0.2712800503,0.0742354691,-0.4020578563,-0.0253429096,0.1834711879,-0.1052995622,0.2061378509,0.5022348166,0.0588144436,0.112675108,0.2139057815,-0.0593331195,-0.3235196769,-0.2381598651,-0.1208349913,0.1364920139,0.0521110296,0.0750076473,0.0321718045,0.0613974817,-0.2957636416,-0.2528942227,0.0021810716,-0.1184125096,0.028022958,0.1504420936,-0.0798131749,-0.1458129138,-0.1736918241,0.0743582025,0.0843824446,0.074846752,0.270375222,0.1307301372,-0.1491893977,-0.1101871282,0.3103753328,0.3500757217,0.1600982696,-0.0866961032,0.2211908698,-0.0560307689,0.1250950545,0.1910205185,0.3222618401,0.4756333232,-0.2523064315,0.0999067053,0.3864905238,0.3116902709,-0.1433416158,0.1414837986,0.0606214367,-0.1722407788,0.1818788052,0.1739681661,-0.1242596805,-0.1034300774,-0.3624983728,-0.0380299538,0.1487063169,-0.18539612,-0.1252017468,0.2422108501,-0.1227990463,-0.1672617495,-0.2144638747,0.0211596787,0.2113020271,0.2708630562,-0.2979032099,0.15759556,0.0637577102,-0.0509712808,0.0109268604,-0.4018692076,-0.0099684894,0.2817002535,-0.0491189621,0.1850633174,0.1528871506,0.267470032,-0.0995557234,-0.0807873383,-0.1803278923,0.0866116956,0.2035046369,-0.3923350871,-0.1149985716,-0.2227960229,-0.0547104552,-0.0287101734,-0.0006919379,-0.1335812062,-0.2146769762,-0.2621063292,0.110352017,-0.1122842357,0.2066514045,-0.0289751664,-0.2205846906,-0.1823903322,0.2438370734,0.0532717854,0.0331175812,0.436803937,0.1147642583,0.2617253959,0.156660825,-0.3266106546,0.1419264078,-0.0725843087,-0.0590925962,0.0747030377,0.2985953689,-0.1415741742,0.1084065139,0.2894104719,0.2608926594,-0.1358896792,0.0838238671,0.1716749966,-0.0309382174,-0.1729512811,0.0944245607,0.1896504611,-0.0510361977,0.0038928338,-0.0637777075,-0.3798277378,-0.2290681899,0.2155451626,0.2613808215,0.1403049827,0.222584039,0.1633605659,-0.253549248,0.406627655,0.0116542475,0.2219263166,-0.1308041066,-0.0523893796,-0.4105676413,-0.0238276459,-0.0015747922,-0.2007561475,-0.3632597625,0.0339233987,0.5755015612,0.4340139925,-0.1519070119,-0.017475605,-0.0307909083,0.2664519548,-0.147030741,-0.0322138593,0.2022717744,0.546646595,0.1108960435,-0.3682801127,-0.339710325,0.1616309434,0.0610878542,0.1865168661,0.1779488325,-0.1060706154,-0.3083718717,0.3497233689,-0.1479789019,0.4285054207,-0.3451287448,0.0563310347,-0.158446461,0.1195360497,-0.169272393,-0.1563918442,0.0275500808,0.5028765798,-0.1112711355,0.1065117195,-0.1819577217,-0.189649716,0.0831209421,-0.2165195197,-0.4016497433,0.2057697922,-0.2983716428,0.05188841,-0.0800708979,-0.1055039614,-0.0849212036,0.1245814264,0.0509297512,0.0649692863,0.631929934,0.0092692366,-0.0461351499,0.1577686965,0.4638565183,-0.1795598269,0.1404388547,-0.3925161362,0.3227731586,0.2682934999,0.0447464064,0.0198243763,0.0708589032,-0.1254260391,-0.155136615,-0.0289503578,0.2028579265,0.085664399,-0.0417919643,-0.2312366664,-0.4446925819]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2050","title":"Build custom dataset to fine-tune Wav2Vec2","comments":"Many thanks! that was what I was looking for. ","body":"Thank you for your recent tutorial on how to finetune Wav2Vec2 on a custom dataset. The example you gave here (https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2) was on the CommonVoice dataset. However, what if I want to load my own dataset? I have a manifest (transcript and their audio files) in a JSON file. \r\n","comment_length":9,"text":"Build custom dataset to fine-tune Wav2Vec2\nThank you for your recent tutorial on how to finetune Wav2Vec2 on a custom dataset. The example you gave here (https:\/\/huggingface.co\/blog\/fine-tune-xlsr-wav2vec2) was on the CommonVoice dataset. However, what if I want to load my own dataset? I have a manifest (transcript and their audio files) in a JSON file. \r\n\nMany thanks! that was what I was looking for. ","embeddings":[-0.1311176419,-0.0219140444,0.0276128929,0.1244109198,-0.1120389476,0.1131009087,0.0107534304,0.0098957587,0.1362034529,-0.042633418,-0.1810429245,0.2923782468,-0.0878038853,0.2396270335,0.1277544349,0.0982615575,-0.1905371249,0.2785508335,0.1800097972,-0.0648362637,-0.1826207042,0.1573688984,-0.0878943801,-0.0408129767,-0.2211071104,0.0856775492,0.1976130307,0.4190138578,-0.2000145316,-0.3390532732,0.5667071342,0.3598968983,0.2059651166,0.2323544323,-0.000110025,-0.1727849096,0.0020980823,-0.2994367182,0.1923188418,0.254417479,-0.4829745293,-0.0639972016,-0.0562798642,0.0626706704,-0.2993437648,-0.1172577217,-0.2281741947,-0.2685032785,0.6499584317,0.287750423,0.1287252903,-0.4400818348,-0.0267854258,-0.1018939689,-0.0211601797,0.110922806,0.0352039896,0.4440509081,0.2972204089,0.0296613798,0.2508305609,-0.0690362081,-0.2513581812,0.0034039458,0.4103522599,-0.0814945623,-0.468521148,-0.2326299995,-0.0433884859,0.5760381222,0.6853452325,-0.3528547883,-0.0505195484,-0.0456937924,0.2641236782,-0.4669241905,-0.0199174769,0.052884113,-0.1343454719,0.4071944654,-0.0615764372,-0.4456338882,-0.3256994188,-0.1199257001,-0.0438148491,-0.0047621303,-0.2345706671,-0.0400467813,0.0383174755,-0.0507505685,-0.1237186342,0.0039462084,-0.3813065886,0.2877165675,-0.4767254591,-0.4398052394,-0.129692331,-0.3349096477,0.1092893034,0.0332515165,0.4096532464,0.142172724,-0.1600676775,0.0435072966,0.0487690158,-0.0717924237,0.1217535287,-0.255405426,0.0584063642,-0.5459836125,-0.0423583016,0.0226868652,-0.2619847059,0.0774080008,-0.4220342934,-0.0761074424,0.1447574198,-0.4342285097,-0.3354520202,-0.0497709364,-0.0252149031,0.0222796202,0.0379527695,0.4244723916,0.1837073714,-0.0937333927,-0.1715826988,0.3703307211,0.21970357,-0.1063561514,0.0678529367,0.092854552,0.3314465582,0.2130458206,0.3683996499,-0.1358840168,0.3506934047,-0.2018977255,0.2128852308,-0.1838152111,-0.0898385942,0.1326024979,-0.3107835352,0.0090765357,-0.045045577,0.0934339389,-0.0670869574,0.0885552466,-0.3393354714,0.006911472,-0.3596331477,-0.4335782528,0.0000497673,0.1336196065,-0.2455039769,-0.1984924674,-0.1254943311,0.1563447416,-0.1574732661,-0.264277041,-0.3791072369,0.0699886307,-0.3138044477,-0.1704280227,0.0912701041,-0.1396386176,-0.9070277214,0.1831873208,-0.2768664658,0.1050503924,-0.0493628085,0.2135892212,-0.0671705902,0.5327351689,-0.0783950612,0.3597759902,0.3181098104,-0.3019596636,-0.2831253111,0.4276229143,-0.2323810607,-0.0258409008,0.1163117588,0.0587745197,0.4729611278,-0.3008342087,0.1220357344,0.3416975439,0.1861348897,-0.1286004335,0.2394817919,-0.1484794021,0.1279454231,0.0119882086,-0.411852926,0.0253585801,0.0573584586,-0.1039974242,0.2102117389,-0.1332789063,0.022170715,0.1656506062,0.2872276306,-0.2010725886,0.0855862051,0.0368290581,-0.1234042421,0.1405924112,-0.011865967,-0.145941183,0.5361707807,0.0428783856,-0.3310780227,-0.2020611018,-0.2413443476,-0.2678895295,0.0808324143,0.1652666479,0.2065101713,-0.6212902069,-0.4915654957,0.1410379261,-0.2421019077,-0.1408395022,-0.1605573297,-0.0550627187,0.4413066208,0.0196887441,0.4232440889,0.2419693172,0.1345499009,0.140915513,0.0137784602,0.4066531062,0.1077549011,0.5067462325,0.2279725522,0.2109337896,-0.193315953,-0.6236417294,0.4890676439,0.2043759078,0.1636250913,-0.1056471393,-0.336881429,0.2931999266,0.2029936761,0.2683027685,-0.2250952721,-0.1612810045,0.3432513773,0.0372772999,-0.2140915245,0.0218127593,-0.0286149662,-0.0329753049,0.4561294317,-0.1831470728,-0.1410104334,0.1327504367,0.1552731097,-0.0762709826,0.1214939132,0.1625693738,-0.3932464719,-0.1991149187,-0.1518826038,0.1117567942,0.4104435146,0.2863570154,-0.1089132503,-0.0796909854,-0.0422702841,-0.2029042542,0.186974749,-0.2950645387,-0.0743584782,0.3175665438,0.0586934946,-0.17859447,-0.1997977644,-0.0169866849,-0.319858551,0.0061437208,-0.3362183869,-0.4482158124,-0.1236769035,0.3906091154,-0.1344137788,-0.2669600844,0.0055526663,0.2599190772,0.1707124263,0.049089618,0.1146890968,-0.0210565794,-0.039671503,0.6205705404,-0.0091853039,-0.2753220499,0.2066388726,-0.034858726,0.0547028594,0.2019744217,0.0504209362,0.2462283224,0.1557762325,0.186775744,0.1698237062,-0.0032279461,0.2269793749,0.0849773064,-0.1121438146,0.1583898813,0.0791526958,0.3029594421,0.340741396,0.0384857059,-0.2422705293,0.2711529732,0.0259245504,-0.1265649348,0.0115681197,0.1464823484,0.2183628529,-0.4437328279,-0.1126381382,-0.2720981538,0.3229070008,0.1821849793,0.0832828507,-0.3374363482,0.0327789076,0.1187443659,0.0003523327,0.253865391,-0.2896342576,-0.3609889746,0.4009687304,-0.1982981712,-0.0632661656,0.2054640651,-0.005132243,0.327281177,-0.1323198527,-0.2013374418,0.3739027679,0.128488794,-0.131028235,0.0957842842,-0.1024077758,0.2903605998,-0.0614142753,0.0396522768,-0.3385054469,-0.2835969925,0.0742879882,0.4068523943,0.1083337292,0.3293242753,-0.0025683895,-0.7417115569,0.6604666114,0.1895305663,0.0538493209,-0.0378276035,0.0219149832,0.1782189757,-0.0443988293,0.0826812536,0.207990706,0.1723489165,-0.1042812318,0.4808909893,0.3075762689,0.0092257354,-0.1544027776,0.1916943789,-0.3328796327,0.1351943463,0.0951097757,0.0297134891,0.0983925238,-0.0015817701,0.3377763927,0.1584325284,-0.1742211133,0.1218781769,-0.0220605172,0.4137531817,-0.2725585401,-0.1497018933,-0.1690255255,-0.2646279037,0.2124409974,-0.3103001416,0.2213338017,-0.0235693734,0.0136866504,-0.1377410293,-0.1096275598,0.3041925728,-0.432302624,0.0443977416,0.0313914791,0.3173485994,0.0642690063,0.2167655975,-0.110957101,-0.1162645444,-0.439540863,0.0118705118,-0.0114849834,-0.0546129867,0.056703642,0.2726527154,-0.2713692188,-0.3398472965,0.2208768725,0.0059610098,-0.1852706671,-0.2527118623,-0.3459210396,0.008804353,-0.3367186487,-0.0130526526,-0.1138221174,-0.0171880554,0.3887637556,0.1074456424,0.3556492627,0.2277381718,0.2224822342,0.1441638172,-0.1232926548,0.1169449687,0.4039838314,0.2937915027,-0.2852511108,0.0412691943,-0.0725725293,0.3704729974,0.1362920702,-0.0607478656,0.0154357795,-0.1046169475,-0.225538969,-0.446600765,0.0367136598,0.1656826586,-0.1258587688,-0.1899804622,-0.4585807323,0.7112780213,-0.0455807038,-0.1778044105,-0.0656606853,-0.2121642232,-0.3837401569,0.2224501371,0.2290952951,0.8211355209,0.1403382123,0.013831038,0.2940026522,-0.2987604439,0.2052488625,-0.7979861498,-0.1632274091,-0.3190098703,0.123750411,-0.2626346052,0.0478395335,-0.0209732801,0.5001817346,-0.1108278781,0.2250265181,0.2541864216,-0.1969888657,0.1002388895,0.1051974371,0.1830620617,-0.4536912143,-0.2057946175,0.1162376925,-0.0850249976,0.0152863245,-0.0120154349,0.042817425,0.0174727011,-0.2829570472,-0.3264710903,-0.0018831546,0.026104616,-0.3366992474,0.0011488473,-0.3163114488,0.2956165075,0.1053108424,-0.1700490415,0.3430676758,-0.5407760143,0.2077240795,0.1880635023,-0.0078059291,-0.0993204042,0.032019373,-0.1980016381,-0.0138598355,-0.0205864906,0.3708782792,-0.0564668775,-0.1542304605,0.0924090818,-0.1211960018,0.1543631703,-0.12823084,-0.5310716629,0.2704443634,0.2170688063,-0.1139392257,0.1059166864,0.4591341615,-0.3517726958,0.21574907,0.0227155108,-0.2322004586,-0.0670369864,-0.023141915,0.1880235821,-0.2928949296,0.3514951169,0.0149986558,0.1613944769,-0.3837572038,0.3605015874,0.1878651828,-0.0009554034,-0.1859170496,0.4257439077,-0.155754596,-0.0770182535,0.1471772939,0.0867193341,-0.033905711,-0.0936345682,-0.5270662904,-0.3360167742,0.0271713026,0.1106061786,0.3457417488,-0.2353309691,-0.1186860427,-0.2585025728,0.1842938215,-0.3159094155,0.0267049223,0.2298123688,-0.1070681885,0.3150395155,0.6320040822,-0.0954247862,0.1346543133,0.0971232206,-0.0675760284,-0.3556584716,-0.1414915621,-0.0676372349,0.1929017603,0.0701018125,0.0659397542,-0.0249703899,-0.0104415035,-0.2888188362,-0.0769710019,0.0074717863,-0.2009524256,0.0716175213,0.0938371718,0.0011857229,-0.1509313434,-0.3047553301,0.0908705816,0.1831632853,0.1953002363,0.1112669483,0.1255805939,-0.1533748806,-0.1039610505,0.3656131327,0.4120562971,0.1654625237,-0.0796994492,0.2267250121,-0.0716284141,0.178270787,0.161351338,0.2056624442,0.436167419,-0.2258826494,0.0604557507,0.4084309936,0.2077419907,-0.1364863217,0.1658937782,0.0936146602,-0.1610638499,0.0492797978,0.1694182456,-0.1554241627,-0.1249726042,-0.3474122584,-0.0754712448,0.1429356337,-0.1538891196,-0.0866784602,0.2436867356,-0.1149554178,-0.1701363623,-0.1230582222,0.1277365237,0.1960514784,0.3247928321,-0.2050922513,-0.0035365319,0.1415783316,0.0482683554,0.0294723194,-0.4661990106,0.117412746,0.400197804,-0.1661041528,0.1451875567,0.1733531058,0.2299599797,-0.0582306311,-0.0753689036,-0.2471583486,0.1414097548,0.1641402245,-0.2957159579,-0.2334101796,-0.2473320216,0.1051322296,0.075806506,-0.1149770543,-0.1328061223,-0.1986828148,-0.2753731012,0.0851039886,-0.0785027295,0.278186053,-0.0815890208,-0.340154916,-0.1473162174,0.3355051577,-0.0908492953,0.0832813308,0.5074509978,0.2064167112,0.1658245921,0.0506605543,-0.2189704627,0.1865641773,-0.0396676548,-0.1220223308,0.0941175818,0.3238725066,-0.1291094124,0.0168529768,0.306475848,0.1774503291,-0.0327721648,0.0847170129,0.2055724412,-0.0251770895,-0.2417053729,0.0577997081,0.1739514619,-0.0952054858,0.0562662296,0.014893434,-0.2872819006,-0.2113499194,0.3360798657,0.2998085916,0.09452039,0.3023610115,0.1133713573,-0.2933159173,0.4404741228,0.1000734568,0.1558806598,-0.0866167545,-0.1483723223,-0.3129458427,0.1152537838,0.0747620612,-0.2220387012,-0.2917435169,0.0385879576,0.638284862,0.49399212,-0.0620137826,-0.0241107997,-0.0236049555,0.3207548857,-0.0490945019,-0.0210403707,0.2260084152,0.6426194906,-0.0549643934,-0.3411759436,-0.2604997754,0.2949799001,-0.054789409,0.1556093842,0.1906349361,-0.0764399916,-0.3093334138,0.4185398221,-0.2067509443,0.4909340739,-0.2031441629,0.0951858535,-0.0984497368,0.1020103619,-0.0518989712,0.0458477139,0.0881415606,0.3604233265,-0.2146313787,0.1932332367,-0.1467022598,-0.2563399374,0.0897630975,-0.1110077202,-0.3253333271,0.1141882986,-0.3892456591,-0.0705857053,0.0871146098,-0.0670811608,-0.0354647674,0.0854651853,0.0932853445,0.1029274166,0.5775825977,0.1115853488,-0.0853946432,0.194803223,0.4766004682,-0.1361408383,0.076508306,-0.2808019817,0.3623526096,0.3311735392,-0.0714842528,0.1279804558,0.0882856995,-0.1655316651,-0.1157664731,-0.0702993274,0.2323382497,0.1461945921,-0.0633185357,-0.3337101936,-0.4147855639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"I think faiss automatically sets the number of threads to use to build the index.\r\nCan you check how many CPU cores are being used when you build the index in `use_own_knowleldge_dataset` as compared to this script ? Are there other programs running (maybe for rank>0) ?","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":47,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\nI think faiss automatically sets the number of threads to use to build the index.\r\nCan you check how many CPU cores are being used when you build the index in `use_own_knowleldge_dataset` as compared to this script ? Are there other programs running (maybe for rank>0) ?","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"Hi,\r\n I am running the add_faiss_index during the training process of the RAG from the master process (rank 0). But at the exact moment, I do not run any other process since I do it in every 5000 training steps. \r\n \r\n I think what you say is correct. It depends on the number of CPU cores. I did an experiment to compare the time taken to finish the add_faiss_index process on use_own_knowleldge_dataset.py vs the training loop thing. The training loop thing takes 40 mins more. It might be natural right? \r\n \r\n \r\n at the moment it uses around 40 cores of a 96 core machine (I am fine-tuning the entire process). ","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":108,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\nHi,\r\n I am running the add_faiss_index during the training process of the RAG from the master process (rank 0). But at the exact moment, I do not run any other process since I do it in every 5000 training steps. \r\n \r\n I think what you say is correct. It depends on the number of CPU cores. I did an experiment to compare the time taken to finish the add_faiss_index process on use_own_knowleldge_dataset.py vs the training loop thing. The training loop thing takes 40 mins more. It might be natural right? \r\n \r\n \r\n at the moment it uses around 40 cores of a 96 core machine (I am fine-tuning the entire process). ","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"Can you try to set the number of threads manually ?\r\nIf you set the same number of threads for both the `use_own_knowledge_dataset.py` and RAG training, it should take the same amount of time.\r\nYou can see how to set the number of thread in the faiss wiki: https:\/\/github.com\/facebookresearch\/faiss\/wiki\/Threads-and-asynchronous-calls","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":49,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\nCan you try to set the number of threads manually ?\r\nIf you set the same number of threads for both the `use_own_knowledge_dataset.py` and RAG training, it should take the same amount of time.\r\nYou can see how to set the number of thread in the faiss wiki: https:\/\/github.com\/facebookresearch\/faiss\/wiki\/Threads-and-asynchronous-calls","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"Ok, I will report the details too soon. I am the first one on the list and currently add_index being computed for the 3rd time in the loop. Actually seems like the time is taken to complete each interaction is the same, but around 1 hour more compared to running it without the training loop. A the moment this takes 5hrs and 30 mins. If there is any way to faster the process, an end-to-end rag will be perfect. So I will also try out with different thread numbers too. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/111453464-798c5f80-8778-11eb-86d0-19d212f58e38.png)\r\n","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":91,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\nOk, I will report the details too soon. I am the first one on the list and currently add_index being computed for the 3rd time in the loop. Actually seems like the time is taken to complete each interaction is the same, but around 1 hour more compared to running it without the training loop. A the moment this takes 5hrs and 30 mins. If there is any way to faster the process, an end-to-end rag will be perfect. So I will also try out with different thread numbers too. \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/111453464-798c5f80-8778-11eb-86d0-19d212f58e38.png)\r\n","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"@lhoestq on a different note, I read about using Faiss-GPU, but the documentation says we should use it when the dataset has the ability to fit into the GPU memory. Although this might work, in the long-term this is not that practical for me.\r\n\r\nhttps:\/\/github.com\/matsui528\/faiss_tips","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":45,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\n@lhoestq on a different note, I read about using Faiss-GPU, but the documentation says we should use it when the dataset has the ability to fit into the GPU memory. Although this might work, in the long-term this is not that practical for me.\r\n\r\nhttps:\/\/github.com\/matsui528\/faiss_tips","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"@lhoestq \r\n\r\nHi, I executed the **use_own_dataset.py** script independently and ask a few of my friends to run their programs in the HPC machine at the same time. \r\n\r\n Once there are so many other processes are running the add_index function gets slows down naturally. So basically the speed of the add_index depends entirely on the number of CPU processes. Then I set the number of threads as you have mentioned and got actually the same time for RAG training and independat running. So you are correct! :) \r\n\r\n \r\n Then I added this [issue in Faiss repostiary](https:\/\/github.com\/facebookresearch\/faiss\/issues\/1767). I got an answer saying our current **IndexHNSWFlat** can get slow for 30 million vectors and it would be better to use alternatives. What do you think?","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":121,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\n@lhoestq \r\n\r\nHi, I executed the **use_own_dataset.py** script independently and ask a few of my friends to run their programs in the HPC machine at the same time. \r\n\r\n Once there are so many other processes are running the add_index function gets slows down naturally. So basically the speed of the add_index depends entirely on the number of CPU processes. Then I set the number of threads as you have mentioned and got actually the same time for RAG training and independat running. So you are correct! :) \r\n\r\n \r\n Then I added this [issue in Faiss repostiary](https:\/\/github.com\/facebookresearch\/faiss\/issues\/1767). I got an answer saying our current **IndexHNSWFlat** can get slow for 30 million vectors and it would be better to use alternatives. What do you think?","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"It's a matter of tradeoffs.\r\nHSNW is fast at query time but takes some time to build.\r\nA flat index is flat to build but is \"slow\" at query time.\r\nAn IVF index is probably a good choice for you: fast building and fast queries (but still slower queries than HSNW).\r\n\r\nNote that for an IVF index you would need to have an `nprobe` parameter (number of cells to visit for one query, there are `nlist` in total) that is not too small in order to have good retrieval accuracy, but not too big otherwise the queries will take too much time. From the faiss documentation:\r\n> The nprobe parameter is always a way of adjusting the tradeoff between speed and accuracy of the result. Setting nprobe = nlist gives the same result as the brute-force search (but slower).\r\n\r\nFrom my experience with indexes on DPR embeddings, setting nprobe around 1\/4 of nlist gives really good retrieval accuracy and there's no need to have a value higher than that (or you would need to brute-force in order to see a difference).","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":181,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\nIt's a matter of tradeoffs.\r\nHSNW is fast at query time but takes some time to build.\r\nA flat index is flat to build but is \"slow\" at query time.\r\nAn IVF index is probably a good choice for you: fast building and fast queries (but still slower queries than HSNW).\r\n\r\nNote that for an IVF index you would need to have an `nprobe` parameter (number of cells to visit for one query, there are `nlist` in total) that is not too small in order to have good retrieval accuracy, but not too big otherwise the queries will take too much time. From the faiss documentation:\r\n> The nprobe parameter is always a way of adjusting the tradeoff between speed and accuracy of the result. Setting nprobe = nlist gives the same result as the brute-force search (but slower).\r\n\r\nFrom my experience with indexes on DPR embeddings, setting nprobe around 1\/4 of nlist gives really good retrieval accuracy and there's no need to have a value higher than that (or you would need to brute-force in order to see a difference).","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"@lhoestq \r\n\r\nThanks a lot for sharing all this prior knowledge. \r\n\r\nJust asking what would be a good nlist of parameters for 30 million embeddings?","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":24,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\n@lhoestq \r\n\r\nThanks a lot for sharing all this prior knowledge. \r\n\r\nJust asking what would be a good nlist of parameters for 30 million embeddings?","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"When IVF is used alone, nlist should be between `4*sqrt(n)` and `16*sqrt(n)`.\r\nFor more details take a look at [this section of the Faiss wiki](https:\/\/github.com\/facebookresearch\/faiss\/wiki\/Guidelines-to-choose-an-index#how-big-is-the-dataset)","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":25,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\nWhen IVF is used alone, nlist should be between `4*sqrt(n)` and `16*sqrt(n)`.\r\nFor more details take a look at [this section of the Faiss wiki](https:\/\/github.com\/facebookresearch\/faiss\/wiki\/Guidelines-to-choose-an-index#how-big-is-the-dataset)","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"Thanks a lot. I was lost with calling the index from class and using faiss_index_factory. ","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":15,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\nThanks a lot. I was lost with calling the index from class and using faiss_index_factory. ","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2046","title":"add_faisis_index gets very slow when doing it interatively ","comments":"@lhoestq Thanks a lot for the help you have given to solve this issue. As per my experiments, IVF index suits well for my case and it is a lot faster. The use of this can make the entire RAG end-to-end trainable lot faster. So I will close this issue. Will do the final PR soon. ","body":"As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n","comment_length":56,"text":"add_faisis_index gets very slow when doing it interatively \nAs the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag\/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster? \r\n\r\n@lhoestq \r\n\r\n```\r\n def training_step(self, batch, batch_idx) -> Dict:\r\n\r\n \r\n if (not batch_idx==0) and (batch_idx%5==0):\r\n\r\n print(\"******************************************************\")\r\n ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder\r\n model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU\r\n model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff\r\n\r\n\r\n list_of_gpus = ['cuda:2','cuda:3']\r\n c_dir='\/custom\/cache\/dir'\r\n\r\n kb_dataset = load_dataset(\"csv\", data_files=[self.custom_config.csv_path], split=\"train\", delimiter=\"\\t\", column_names=[\"title\", \"text\"],cache_dir=c_dir) \r\n\r\n print(kb_dataset)\r\n\r\n \r\n n=len(list_of_gpus) #nunber of dedicated GPUs\r\n kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]\r\n\r\n #kb_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/haha-dir')\r\n\r\n\r\n print(self.trainer.global_rank)\r\n dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])\r\n output = [None for _ in list_of_gpus]\r\n\r\n #self.trainer.accelerator_connector.accelerator.barrier(\"embedding_process\")\r\n dist.all_gather_object(output, dataset_shards)\r\n \r\n\r\n #This creation and re-initlaization of the new index\r\n if (self.trainer.global_rank==0): #saving will be done in the main process \r\n \r\n combined_dataset = concatenate_datasets(output)\r\n \r\n passages_path =self.config.passages_path\r\n\r\n logger.info(\"saving the dataset with \")\r\n #combined_dataset.save_to_disk('\/hpc\/gsir059\/MY-Test\/RAY\/transformers\/examples\/research_projects\/rag\/MY-Passage')\r\n combined_dataset.save_to_disk(passages_path)\r\n logger.info(\"Add faiss index to the dataset that consist of embeddings\") \r\n\r\n \r\n embedding_dataset=combined_dataset\r\n index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)\r\n embedding_dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n\r\n embedding_dataset.get_index(\"embeddings\").save(self.config.index_path)\r\n\r\n\n@lhoestq Thanks a lot for the help you have given to solve this issue. As per my experiments, IVF index suits well for my case and it is a lot faster. The use of this can make the entire RAG end-to-end trainable lot faster. So I will close this issue. Will do the final PR soon. ","embeddings":[-0.4988238811,-0.2682781518,-0.0245007072,0.1282283813,0.0594396777,0.2077840716,0.1214630976,0.4241297543,0.291741699,0.2924969196,-0.1188249737,0.1883148551,0.1372130662,0.085820809,-0.1362560391,0.1750429869,0.2370326221,0.0853344873,0.2607980072,-0.1344140023,-0.3274796903,-0.3090646267,-0.1589119136,-0.2578627467,-0.4006919265,-0.0958186537,-0.2436220646,-0.0820504501,0.1982473135,-0.4729171097,-0.137532711,0.1398431212,0.3402583897,0.2582926452,-0.0001220497,0.0556767285,0.2280078083,0.0654628575,-0.2229975015,0.6411964297,0.3400724828,-0.0694884807,-0.1751104295,-0.3107712865,-0.1455549747,-0.1110416427,0.0810796022,-0.0469295718,-0.166365087,0.1943010539,0.0266239122,-0.0582869202,0.0522551611,-0.0154001648,0.4025111794,-0.0755912513,-0.0691769868,0.3174717128,0.1635129154,-0.0722376183,0.0997427031,0.4481079876,0.12064147,-0.2267771959,0.1089045852,0.2051612586,0.7838166952,0.0224396084,0.0652679056,0.152669698,0.0153403804,-0.0905528963,-0.4430085719,-0.1029038355,0.4257141352,-0.6180593967,-0.0674013048,-0.0837419853,-0.1014744639,0.0538599454,0.1745852083,-0.030230185,0.0363178104,0.0474507846,0.1748995632,0.1653574407,0.3874710798,-0.0599023178,0.5584340096,0.0424474142,-0.1168545187,0.0448825061,-0.0920769498,0.0121963527,-0.6050299406,0.0880849659,-0.0309161283,-0.099331513,-0.0926872119,0.0178016424,-0.3694860637,0.0061654514,0.2365601063,0.0033363802,-0.183065027,-0.0902583003,-0.5061043501,0.055881232,0.083901681,0.106286943,-0.3780879974,0.0992346108,-0.0033380785,-0.0043100738,-0.3259801865,-0.2981352806,-0.2288905531,0.2228072584,-0.1811424345,0.0771879777,-0.403968364,-0.25227651,0.2050724775,0.416221559,-0.1177439839,-0.0761364475,0.0108397733,-0.0346156918,-0.2344279587,0.1761687398,-0.2220168859,0.2306603193,-0.0319639184,0.5045982003,-0.0291060489,-0.2213940918,-0.0890693069,-0.0217307452,0.083963491,-0.0078422474,-0.044163648,-0.2998936474,0.2466678023,0.0749230385,-0.2632145882,-0.0676850677,-0.0939525291,0.1211105511,-0.2278231531,0.2215083539,-0.2173948735,-0.3651023209,0.0271579791,0.0193311907,0.3086776733,-0.0630925894,0.420822531,0.1143972278,0.1730186641,0.1613348126,0.159176752,-0.1314434856,-0.2680214942,0.0831866488,0.2267838418,0.3627386689,0.0937216207,0.0347714052,0.2980125546,0.3901303113,0.2124055773,0.3999934494,-0.1159487739,0.6762604713,-0.320004195,0.0842921436,0.2451203614,-0.2553785443,-0.2924265265,0.1869049817,-0.1663640887,-0.0773261935,0.3296633363,0.518270731,0.3120296896,0.0870154649,0.3410930037,0.3062556088,0.0090267127,-0.0312473625,-0.4050686955,-0.0582714975,0.2639195919,0.380407393,-0.3354525268,0.10174077,-0.0401814356,-0.3298804164,0.6009740233,-0.4761057198,-0.1217618585,0.0680584759,0.3615735471,0.3393865526,0.2328698933,0.2761101127,0.1495419145,0.1626259685,0.0960143954,0.0987884179,0.1701551676,-0.4183505476,0.2156298608,-0.0257499516,-0.2181926966,0.1078814194,-0.0609779805,0.1500251442,-0.0456768312,-0.1998734474,-0.1568868458,0.5427840352,-0.6157974005,0.0279460251,-0.2488968521,-0.1560835391,0.0070153689,-0.2134255767,-0.0342439711,0.142958805,0.052496288,-0.3868663311,-0.0538331419,0.1028612107,0.1266249865,-0.187676087,0.5111077428,-0.1136960238,-0.0370367952,0.2962219417,0.044952888,-0.1512648761,0.1090166271,-0.1937616616,-0.0018785683,0.4852037132,0.3487839997,0.2458340824,-0.0214935225,-0.2332086414,0.0470289811,-0.0406502709,-0.0793870017,0.2275258452,0.0799438357,0.1302532405,0.409172982,-0.0333352648,-0.3332441449,0.2324638963,0.1857769042,-0.2031303048,-0.4314085543,0.2710421979,0.3516993821,-0.0368976481,-0.3497084081,-0.5220082998,0.2360264212,0.3185990751,0.1883933693,-0.071653828,-0.0276507791,-0.2472812831,0.1691094786,0.0565532297,-0.048372224,0.1000481471,0.1766420156,-0.2145446539,-0.2067899555,-0.2552964985,-0.1363964677,0.4286457598,-0.0443389192,0.3094405532,0.1777624935,0.2771865427,-0.1247245297,-0.1249763295,0.2887524962,0.0088405078,0.0416105874,-0.1089353785,-0.0362386964,0.1931093037,-0.0469494499,0.3649187982,0.1957939118,-0.3250905871,0.0198445637,-0.0984703004,-0.121494934,-0.0789786056,0.1210055202,-0.1505499333,0.1174001396,0.0439472869,-0.0525806546,-0.1851366758,-0.0862713903,-0.1575880647,-0.0794765055,-0.2974514365,-0.1662347317,-0.2460688353,-0.1250380278,-0.119644545,0.0043304516,0.0005504017,-0.0106870206,-0.0404876322,-0.2826742828,-0.1588367969,0.060794808,-0.2936334312,-0.2167021781,-0.1739186198,0.2281414121,-0.0243999027,0.0930592716,-0.1156688854,0.0667516664,0.1222407445,-0.0957010388,-0.1543302834,-0.0534614287,-0.2331429124,0.2917095423,0.1491725743,0.0468316413,-0.0794368014,-0.4122965932,-0.0143970978,0.4969037771,-0.5218712687,-0.0407988019,-0.3583908975,0.2859594226,-0.315094173,0.4582990408,0.0567070097,-0.2462490499,-0.1380138248,0.3644826114,0.1745104492,0.3875301182,-0.0617133379,0.0549775884,-0.087917462,0.2675451636,-0.1149751395,0.5162414312,0.327231437,-0.2136980891,-0.0478191786,-0.2623966038,0.1159517244,0.0527700186,-0.1398820877,0.2965247333,0.0834778845,-0.3149955571,0.166527763,-0.2887444198,-0.4117940664,0.0320426561,-0.052759815,0.1982454956,-0.2443107069,0.3169922233,0.0489506237,0.4910995066,-0.1780503839,0.2910963297,-0.238360092,-0.1767041981,0.1975145787,0.2727532089,0.2449010462,-0.1675575376,0.0264966935,-0.5409094691,-0.421677947,0.2834408283,0.1674793512,0.2352832109,0.27911973,-0.1008348912,0.5258054733,-0.1562118083,0.7999860644,0.0206833687,-0.3363376558,0.3844905198,-0.1618198901,0.0528762192,-0.217792958,-0.1827090234,-0.0659896061,0.2337735593,0.4407466054,-0.0931776688,-0.2479643226,0.6446050406,0.0670878068,-0.1472481787,-0.3441523612,-0.1861819625,-0.3543058634,-0.0412193835,0.7678548098,-0.1216907948,0.1556899697,0.1822201014,-0.0512349457,-0.3818880618,0.1598657519,0.2386750281,-0.0051203095,0.0505829975,0.4557841718,0.434810549,0.0714174435,0.0007315621,0.1192743033,-0.0369113721,-0.304396987,0.0731521398,-0.1888458878,-0.2556853592,0.3799425066,0.7006738186,0.1868769079,0.2484076768,-0.2102308869,0.3775886893,-0.1914457679,0.0245311894,0.4383130968,-0.0020623421,-0.1787728816,-0.5383264422,0.3112787902,0.2034226507,0.0733027011,0.0608612858,-0.0757305995,-0.0163804982,0.3021707833,0.1249995977,0.9838235974,-0.0269193128,0.1929833293,0.5063086748,0.0048403963,-0.2439047098,-0.6487340927,0.3692772686,-0.3329212964,-0.2241527438,0.19311665,-0.1011415944,-0.0847687051,0.0735340491,-0.0393425748,0.0991450474,-0.0472755358,-0.0862561688,-0.2586017251,0.3895545304,0.3151937425,-0.427839756,0.0248834006,0.0013510936,0.2714567184,-0.0987162814,0.1278156638,-0.1634003222,-0.1108861864,0.0184409097,0.1251738518,-0.079113625,-0.1283788085,0.0130584268,-0.048773136,-0.0873115808,-0.2329261005,-0.3244183362,0.4604183435,-0.0299460199,0.099533841,0.0815839842,-0.0991102606,0.0721638948,-0.0867692456,-0.4970832765,0.2265409529,0.0808123723,-0.1736778021,0.1493705064,0.0787266716,-0.3227716684,-0.1460989416,0.1744306087,0.0205767713,0.0075482796,-0.0367376842,0.3179990053,-0.2626375854,-0.233565405,0.0200596862,0.0920771807,-0.4317522943,0.5605168343,-0.3023820817,-0.3800685406,-0.0546555109,0.0166955758,0.3058082461,0.0560379177,0.2623713315,-0.0891195536,-0.3048813343,-0.154676944,-0.3024288416,0.0612835661,0.038625434,0.0938879102,0.302420646,-0.006131018,0.0253992043,-0.2326641232,-0.3227176368,-0.1417698264,-0.3823908567,0.1145998389,-0.3576403558,-0.1850678772,0.2572282255,0.3836848438,0.0360177681,0.2342240661,-0.4448954761,0.2590474188,-0.2176698297,-0.0653392002,-0.0312390663,0.2600193918,-0.0327091888,-0.2574275136,0.1754399091,0.216269955,0.1199890003,0.1412038803,0.0373655409,-0.1117988154,0.1029488668,0.1834249794,-0.1107470617,0.1684509367,0.0811557174,-0.1483059675,-0.3009882569,-0.2652982175,0.1343657374,-0.1107031256,-0.0258548204,-0.130359605,-0.1217091233,-0.1124874875,0.3323770165,0.2738248408,0.2040049136,0.2833091319,-0.0734992623,0.0767018721,-0.2428326607,-0.2653926313,-0.0510375872,0.1903247386,0.0616441816,0.0863678828,-0.0918656886,-0.1731881648,-0.3799477518,-0.038117867,0.0317691378,0.449894011,0.0424936824,-0.2883665562,0.1852958649,0.0379373133,-0.2797203958,-0.1952762306,0.3624266684,-0.450701654,0.2002288103,0.4017177224,0.2645750344,0.1657235473,-0.48327443,0.0937636197,-0.2345171422,0.0963051319,-0.4337077141,-0.0904511884,0.3514743447,-0.1948996931,0.2041253448,0.305214107,0.2311200649,0.6089174747,0.0314238705,0.2752903998,0.1878775954,0.2133270055,-0.3217031658,-0.1798365116,0.2906689942,0.2322690785,-0.031772878,-0.2078448087,0.0090950485,0.234734416,0.244035542,-0.5169654489,-0.2611279488,0.1630211473,-0.1220198125,-0.1281811148,0.4117490351,0.0567706563,-0.2780815363,0.2270267457,0.0776639879,0.1923735142,-0.1272480339,0.2008739859,0.293453455,0.2212003469,-0.0231135879,0.0910241529,0.4337710738,-0.3367401361,-0.2079150379,-0.0111506963,-0.0027164146,0.0235897284,-0.1512230933,0.2809455693,0.2448252439,-0.0555025823,0.2687214911,-0.0504418314,-0.0088708708,0.0438044965,-0.0344011858,-0.1084126681,-0.1003935561,0.100617148,-0.0908763632,-0.1285336912,-0.1207231656,-0.2426326573,0.2026132047,-0.1804250181,0.0230685603,0.3459430039,-0.097540006,-0.394805491,0.0952416509,-0.2833034098,0.0691806376,0.1062706709,-0.3566012383,-0.0891359225,0.260307014,-0.001634132,-0.4257482886,0.5468898416,0.1724049896,0.1286819428,-0.3299412429,-0.0014343098,-0.3207880855,-0.03450967,-0.5890921354,0.1126354709,0.1176836565,0.4403544068,0.0739260614,-0.0703547075,0.403108716,-0.4079712927,0.0992396548,0.3735492826,-0.1734416634,-0.0306224171,-0.4397849739,-0.0738472417,0.0422833189,-0.3743228316,0.2847961485,-0.128206864,-0.014769651,0.1923716366,-0.0202694833,0.1226510704,0.0626824573,0.5298789144,-0.0979442969,0.0000549222,-0.0964540243,0.0086346753,-0.2321906537,0.1219591051,-0.2070786059,0.0085758734,-0.0218882244,0.1733755469,-0.2073738724,-0.3919604123,-0.1815342754,0.4176576436,0.3126299679,-0.4186961055,0.1980270892,0.1526593566,-0.1558282375,0.4795940518,0.1503624618,0.228284359,0.1467327774,-0.0340511203,-0.3561269045,-0.1054201797,0.6236690283,-0.1882943511,-0.0293618869,-0.3494385481,0.2549015284,0.4437921047,0.5510439277,-0.2645769119,0.0862376094,0.1219433472,0.0535738133,-0.2103946507,0.5777599216,0.1868614554,0.0147712212,-0.0737837702,-0.2648080289,-0.0630890578,-0.0029669453,-0.3532103598,-0.2814310193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2040","title":"ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk","comments":"Hi ! To help me understand the situation, can you print the values of `load_from_disk(PATH_DATA_CLS_A)['train']._indices_data_files` and `load_from_disk(PATH_DATA_CLS_B)['train']._indices_data_files` ?\r\nThey should both have a path to an arrow file\r\n\r\nAlso note that from #2025 concatenating datasets will no longer have such restrictions.","body":"Hi there,\r\n\r\nI am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):\r\n```python\r\nconcatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])\r\n```\r\nYielding the following error:\r\n```python\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\nBeen trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...\r\n\r\n`load_from_disk(PATH_DATA_CLS_A)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 785\r\n})\r\n```\r\n`load_from_disk(PATH_DATA_CLS_B)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 3341\r\n})\r\n```","comment_length":41,"text":"ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk\nHi there,\r\n\r\nI am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):\r\n```python\r\nconcatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])\r\n```\r\nYielding the following error:\r\n```python\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\nBeen trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...\r\n\r\n`load_from_disk(PATH_DATA_CLS_A)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 785\r\n})\r\n```\r\n`load_from_disk(PATH_DATA_CLS_B)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 3341\r\n})\r\n```\nHi ! To help me understand the situation, can you print the values of `load_from_disk(PATH_DATA_CLS_A)['train']._indices_data_files` and `load_from_disk(PATH_DATA_CLS_B)['train']._indices_data_files` ?\r\nThey should both have a path to an arrow file\r\n\r\nAlso note that from #2025 concatenating datasets will no longer have such restrictions.","embeddings":[-0.0112980641,-0.0258976668,-0.0449595004,0.5074425936,0.1715030521,0.1827167124,0.0579556674,0.1286477149,-0.062890619,0.1336227506,0.0703736618,0.2955010831,-0.0843805,-0.1025009677,-0.325756371,-0.1238994524,0.2124682367,-0.0417845137,-0.1874171495,-0.1163776815,-0.433898896,0.3035929203,-0.0941289663,0.0441321991,-0.3788501918,-0.1216486916,-0.3139139712,0.3639658391,-0.1314383149,-0.1400639564,0.3840758502,-0.0434939414,0.1431432962,0.4993310869,-0.000123913,-0.0518494174,-0.0255158506,-0.1264899373,-0.3373901546,-0.3795106709,-0.2954404652,-0.2493540496,0.1174825877,-0.3515410125,0.3774068356,-0.1962019503,-0.1029743999,-0.72650069,0.2809568048,0.2295499593,0.0761976615,0.1972938329,0.1317069829,-0.0478940718,0.0319104157,0.0735818446,0.1537163258,0.2029614002,-0.0550396815,0.023870077,0.1942694187,-0.0005485428,-0.2328488678,-0.118594557,0.1362632364,0.4979306757,0.0554191284,-0.5291479826,-0.0690567121,0.0863210261,0.4802346826,-0.3763854802,-0.3781012595,-0.1759452671,0.0070042545,-0.3268713355,0.1010299474,0.3545840085,0.0001945458,0.111808531,-0.0617500842,-0.0342379436,-0.2941899002,0.072594814,-0.0851020068,0.0358606912,-0.085884653,0.1781402677,0.4493532777,0.0479511246,0.3216394186,-0.4869163632,0.0636960641,0.1922716051,-0.1096954197,0.1452182978,-0.1640024483,-0.5908123851,0.0066751731,-0.0109491264,0.1836389899,-0.1348267496,-0.175842151,0.302200824,0.2534691989,0.3053895533,-0.0141317127,0.4288300574,-0.0103183277,-0.1330448687,0.0501374006,-0.0258551724,-0.0293559134,-0.422658354,0.0796900839,-0.0718142763,0.1062273234,0.0129516535,-0.3877182007,0.1803555936,-0.2840075195,-0.1573118418,0.097162649,-0.0178047977,0.0777068138,0.3530093431,0.3876343668,0.3053706586,0.0820669904,0.2978774905,-0.1446357071,0.1967317462,0.0749064609,0.083897993,0.1068450958,0.0016405154,0.0119908359,0.1067731008,0.0797290355,-0.1220689043,0.0271952283,-0.3416927755,0.2821960449,0.1566369385,0.0981630832,0.3480808735,0.2592440546,-0.2748633921,-0.1870557368,0.3995105326,-0.357647866,-0.1869686842,-0.1376323253,0.0494735204,-0.1045715734,0.1410053968,-0.4229938686,-0.0059976471,0.3018890619,0.3051957488,-0.1290160865,-0.0568313226,-0.1296139359,-0.4591296613,0.341227144,0.0073847673,-0.4581022561,0.0097240172,0.2903791368,0.1048898995,0.1965613365,0.4308505058,-0.1441585571,0.3636823595,-0.3045046329,0.0102365799,0.4004561603,-0.4094220102,-0.1038569584,0.1716967374,-0.1796122342,-0.0410915762,0.3645077944,0.027034644,0.0063065663,0.209405601,0.7220623493,0.242559135,0.0746658593,-0.1657408774,-0.0952380598,-0.2326027751,0.3555584252,-0.2086943984,-0.133979246,-0.0259477682,-0.2588166296,-0.5167700052,0.4240078032,-0.2052067667,0.0640659034,0.4364206791,0.3568811715,-0.1233663261,-0.1038491875,0.1537686735,-0.1934182346,0.1686595827,-0.0148310009,-0.222709313,-0.2647582591,-0.0220493209,-0.2106941342,0.1269541979,-0.0709804296,0.0814448446,-0.0153741334,0.0325087346,0.0056665661,-0.1054476202,-0.3481857479,0.5563704371,-0.163145721,0.0219019782,-0.2541760206,0.4569391906,-0.1578050554,-0.0263907127,-0.0453710034,0.0369337983,0.2729869187,0.1677791625,0.0227815844,0.4717352688,-0.0430073328,0.0950465053,0.0893382505,0.0808452666,0.1716275215,0.0057399063,0.1627613157,-0.0579361059,0.1558342874,-0.1728034914,-0.0909185782,0.4870926142,-0.3344883025,0.4485846162,0.1146494597,0.0512755178,0.1062385887,0.0439708121,-0.0784747005,-0.2369313389,0.0169220511,0.0849660113,0.2192247063,0.2170421481,-0.3525520861,0.0700837746,0.1829123944,0.0754989311,-0.0589676462,-0.0108339787,-0.1080609784,-0.0159977395,-0.1095021665,0.3902995288,0.5840013623,0.1719545275,0.1649895161,-0.1167301685,-0.13713108,-0.1706053317,0.3417837322,0.1398366243,0.2364146113,0.1728817374,-0.0857251808,-0.0756203309,-0.0717299432,-0.0175315533,-0.1918958873,0.0328854062,-0.5536921024,-0.0458903313,-0.2176486552,-0.1098990589,-0.4481921196,-0.3460378349,-0.0974932611,-0.5419760346,-0.2592168748,0.4547411799,-0.1169457287,-0.0527836606,-0.2844993174,0.1456625015,0.0052279974,0.067862317,0.1234738752,0.0104368366,-0.078316763,-0.0318523794,-0.1368813068,0.0180019811,-0.0136504984,-0.0726327598,0.2563133538,-0.2595613599,0.066964522,-0.1467585117,-0.1533612609,0.0849603042,0.0322470106,0.315906316,-0.2281421572,-0.6266650558,0.0580363348,0.5674738884,-0.1389603317,0.0900544003,0.2620217204,-0.18418549,0.0017389727,-0.2506844997,-0.4523783922,-0.2502115369,0.0298633371,-0.1747046113,0.2491324097,0.1712756604,0.2696705759,0.1090817451,0.3577369452,0.1941004395,-0.1135379747,-0.2557307482,0.4422734082,-0.0536296591,-0.1642774045,0.0017761585,0.008876658,0.0754111558,0.1504588127,-0.3661916852,0.018025646,-0.2361494601,0.1363372058,-0.1147485897,0.3011988699,0.1340018809,0.4595842361,-0.0462786779,-0.258654505,-0.273722291,0.2683783472,0.1589276046,0.3044026494,-0.0123012429,0.2090227306,0.0731703639,0.3439159095,0.3640868664,0.1217375472,0.2549828589,0.1836441308,0.5346710086,0.0144409453,-0.4406895638,-0.4077027738,-0.001152494,-0.2065867186,-0.13948825,0.0483936556,0.0872390419,-0.0935864896,0.2118113637,-0.2357505858,-0.3036771715,0.0938029513,-0.3211862445,0.1249999776,-0.2163255513,-0.0679508224,-0.2828488052,0.3194229305,-0.1286117285,0.0600240454,0.2188934833,-0.1355677396,-0.3572336733,-0.3390055895,-0.3777068257,0.304944694,0.0543895811,0.3831124902,0.0539906174,-0.2907821238,-0.2016689777,0.044948753,0.7679034472,0.2945116162,-0.3783245683,0.4131022692,0.0676539764,-0.5299774408,-0.038080968,0.0837753862,0.2424825877,-0.2675106525,0.4023001492,-0.2402883172,-0.1255395412,-0.1106792241,0.4439459145,-0.1429796815,0.0849760324,-0.0313197449,-0.1303118318,-0.1526848525,-0.3208156228,0.0542216152,0.3823355138,-0.3698827326,-0.0626258254,-0.000834224,-0.2915801704,0.3242531717,-0.3542226851,0.601585269,-0.0703038499,0.1090659648,0.2168456465,0.1730602086,0.4914812744,0.467076689,-0.2443522662,-0.1175284609,0.0126572279,0.0911095515,0.5109778643,0.3323250413,-0.2225600779,-0.03019494,0.0194206852,-0.0991326049,-0.2740884125,-0.1163052768,0.2955112755,-0.0105290608,-0.4240542948,-0.3991217911,0.2900102139,0.3583661616,-0.1670621336,0.441819638,0.0083779991,-0.4729835391,0.3470169604,-0.1212992743,1.075656414,0.0404348187,0.2828830481,0.1625709534,-0.2942945957,0.0904012471,0.2924495935,-0.0318743996,-0.1988106221,-0.1252881885,-0.1203400493,-0.3440369368,0.0014600576,0.2321632951,-0.3387193382,0.1795612127,-0.5210160613,-0.1161916479,-0.2029711306,0.0473358408,-0.1271822453,-0.2273617983,-0.3534165025,-0.0209864825,0.2143520713,-0.1160492375,-0.0184171684,-0.1729265004,-0.525723815,-0.1964709014,-0.1285266429,0.040109098,-0.3198243976,0.6732954979,-0.2941882312,-0.4334102869,-0.4329429865,0.3981278837,0.408562839,-0.0319073312,0.026282372,-0.1287884414,0.0475807488,0.0084446771,-0.0751561373,-0.2583680153,0.1939774752,0.1476748288,-0.0745912716,0.1135495231,-0.0817414001,-0.3694067597,-0.0846799985,0.100899905,0.0339482464,-0.2673567235,-0.3413455188,0.1735767573,-0.0646068677,-0.0645267293,0.0239265338,0.3677726984,-0.0626553148,0.2074286938,-0.2126592845,-0.3376839161,-0.05606924,0.4843715131,0.3131256998,-0.0178009309,0.5047691464,-0.0361333564,-0.1005940661,-0.0298029911,0.3686383367,0.2418309599,0.1030880436,0.3431331217,0.0801178738,0.0010804662,-0.1808097959,0.3765243292,0.3969090879,0.2314935029,-0.0998983458,-0.2665150166,-0.5434687138,0.403437227,0.0367556773,0.4078207016,-0.0284728166,-0.4364436269,-0.0923396498,-0.0955633596,-0.1985139847,0.0572784208,-0.1068870872,0.0202854052,0.2047138959,0.0198701117,0.1778991669,-0.0135628683,0.0580114424,-0.1984192282,-0.0634627789,-0.078251332,0.0476515032,0.1854497045,0.1121996269,-0.0773319751,-0.1123948321,-0.4440903664,0.0286660604,-0.2024708837,-0.1707081646,0.5613838434,0.0619375408,0.1161672622,0.0000499555,0.2100527138,0.016176587,0.1269423068,0.0819807351,0.2238188237,0.045877628,0.2438220233,-0.3202713132,-0.2187683135,-0.3172678649,0.0508361943,-0.1387191266,-0.033538498,0.1269544363,-0.1474207491,-0.222672984,0.0340960994,0.3946391642,0.4278351068,-0.1175069958,0.1583574265,-0.0283109453,0.0834254846,0.1216277182,-0.2142873406,-0.1380100399,-0.1968071312,-0.2094020247,0.1497976333,0.0626813918,-0.0821174756,-0.3396139443,0.0256713629,0.5485681295,-0.1606158912,0.4001900554,0.0932295322,0.1508625597,0.0077852253,-0.0089490078,0.2421151251,0.4247482717,-0.0585648902,-0.2144656628,0.0903658122,0.481608808,-0.1609710157,0.2162282616,-0.230855301,-0.0546205901,0.1499199271,-0.1688839942,-0.0110081155,0.0624805428,0.3620595932,-0.3919822574,-0.2237451822,0.0981596932,0.3435716927,-0.1242571846,-0.1706208438,0.0088113947,-0.2805442512,0.036251206,-0.1801573485,0.1347273141,-0.3597926199,0.1951784492,-0.0274089221,0.0288217533,-0.11451713,0.3473956883,0.4074490666,0.1072439328,-0.2372736931,-0.0107777165,0.1962230057,-0.0470069647,0.1492601782,0.1895022988,0.5935686827,0.5518577695,-0.0385352224,0.1228287965,0.0862643644,-0.2647485137,0.0007321407,-0.0269172359,-0.2013376057,-0.3894877434,0.4084110558,0.1807920337,0.0093471967,0.0648568943,0.2460978478,0.5886136293,-0.1220221594,0.1053811908,-0.1613057107,-0.2382538766,0.1525796503,-0.0851618797,-0.3969837725,-0.2639622688,0.4410467148,-0.0716996938,0.1276208162,0.2629283071,0.0381974503,0.0462067612,0.2446825951,0.2988674045,0.1290016323,-0.46796453,-0.0109863337,-0.3217752874,-0.2128876448,0.3092116117,0.1748177111,-0.1901315004,0.366796881,0.163945809,-0.0158231985,0.2488795221,0.0814523995,-0.1605889648,0.1002152786,-0.4152258337,-0.139394477,0.3665907085,0.0396438092,-0.0583290122,-0.2068106085,0.1909677684,0.1240238473,0.012798395,0.0291861277,0.0827609822,0.1202375442,0.4333843887,0.2898519039,0.1749394983,0.2608098686,-0.1220642328,0.0228217989,0.0182279646,-0.3067571819,-0.0446161442,0.1826509088,-0.0878289565,0.4609199762,-0.5516213179,0.0427908264,-0.1169721559,0.0292599071,-0.0026024608,-0.4689415991,-0.0616028011,0.0664885044,-0.0075736465,-0.0252796542,0.1431801468,0.2222637832,0.0159162544,0.5831586719,0.070473738,-0.4461385608,0.4071060717,0.0208621267,-0.1125457287,0.1859321445,-0.0085480437,0.3054602146,0.425175339,-0.6048345566,-0.1813082993,0.4282561839,-0.2483880967,-0.2672109306,0.0092466613,-0.0783350766,-0.0299415588,0.0153823402,0.4939104617,-0.10249383,-0.3544383645,-0.0302168019,-0.2761022151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2040","title":"ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk","comments":"Sure, thanks for the fast reply!\r\n\r\nFor dataset A: `[{'filename': 'drive\/MyDrive\/data_target_task\/dataset_a\/train\/cache-4797266bf4db1eb7.arrow'}]`\r\nFor dataset B: `[]`\r\n\r\nNo clue why for B it returns nothing. `PATH_DATA_CLS_B` is exactly the same in `save_to_disk` and `load_from_disk`... Also I can verify that the folder physically exists under 'drive\/MyDrive\/data_target_task\/dataset_b\/'","body":"Hi there,\r\n\r\nI am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):\r\n```python\r\nconcatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])\r\n```\r\nYielding the following error:\r\n```python\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\nBeen trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...\r\n\r\n`load_from_disk(PATH_DATA_CLS_A)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 785\r\n})\r\n```\r\n`load_from_disk(PATH_DATA_CLS_B)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 3341\r\n})\r\n```","comment_length":43,"text":"ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk\nHi there,\r\n\r\nI am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):\r\n```python\r\nconcatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])\r\n```\r\nYielding the following error:\r\n```python\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\nBeen trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...\r\n\r\n`load_from_disk(PATH_DATA_CLS_A)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 785\r\n})\r\n```\r\n`load_from_disk(PATH_DATA_CLS_B)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 3341\r\n})\r\n```\nSure, thanks for the fast reply!\r\n\r\nFor dataset A: `[{'filename': 'drive\/MyDrive\/data_target_task\/dataset_a\/train\/cache-4797266bf4db1eb7.arrow'}]`\r\nFor dataset B: `[]`\r\n\r\nNo clue why for B it returns nothing. `PATH_DATA_CLS_B` is exactly the same in `save_to_disk` and `load_from_disk`... Also I can verify that the folder physically exists under 'drive\/MyDrive\/data_target_task\/dataset_b\/'","embeddings":[-0.0077575627,-0.0073159016,-0.0188266747,0.5390487909,0.2280756831,0.1842921823,0.0531079397,0.130966872,-0.0564404465,0.1484978646,0.062885955,0.2342555374,-0.0325047113,-0.1666058749,-0.3268118501,-0.0597667173,0.2327091545,0.0071995803,-0.0789073482,-0.1194586605,-0.422042191,0.2971969545,-0.1387291849,-0.0585789718,-0.3954503536,-0.0667551979,-0.4116065502,0.4386834204,-0.1577688307,-0.1015263945,0.5081785321,-0.0862840116,0.1224367991,0.6370720267,-0.0001226115,-0.0308201071,-0.032479845,-0.2020162344,-0.2702163756,-0.4532000124,-0.2112329751,-0.2356748879,-0.0447952561,-0.3365784883,0.2640665174,-0.0735307634,-0.1124737114,-0.7267848849,0.1495839357,0.2615768313,0.0758540258,0.2023937106,0.0412549563,-0.011546365,-0.0034611092,0.1429173648,0.1469982415,0.1667424291,-0.0944767669,-0.0606658533,0.2073462307,-0.0751989707,-0.3042544127,0.0131352898,0.1142961979,0.4793724716,-0.0599415526,-0.4340187013,-0.0781594962,-0.0228613894,0.4604327679,-0.369572401,-0.3713535666,-0.1061630473,0.0483915769,-0.2898160219,0.1351525187,0.3806920946,-0.0266378783,0.1082230732,-0.2135585099,0.0090284878,-0.252833426,-0.0439741537,-0.1308439374,-0.0517339781,-0.068910569,0.1527519077,0.42635566,0.1246833652,0.3369889557,-0.5939996839,0.0247724019,0.2136669159,-0.2035794705,0.0758955702,-0.0231927112,-0.4214418828,0.0178550277,-0.0495417453,0.2194831818,-0.1294461191,-0.1592091769,0.2647929192,0.1793381274,0.4705998302,-0.0306805484,0.4814172685,0.0561166964,-0.1248716637,-0.0164853912,0.0220423788,-0.0107339174,-0.2809270322,0.17250669,-0.0376351178,0.0512935445,0.0172459036,-0.4902620018,0.172274068,-0.2627789676,-0.1579497606,0.1375524551,-0.0358217433,0.0247514136,0.3958359361,0.3578871191,0.3385949135,0.0773075074,0.3362582922,-0.175747022,0.2379208058,0.0614542402,0.1129006669,0.2048764229,-0.0268174615,-0.0324295498,0.0669087991,0.0196863823,-0.0630932227,0.007189549,-0.4055666625,0.2116440386,0.1468408108,0.1655864418,0.3385901451,0.3134979904,-0.3476326466,-0.1363859773,0.3720194697,-0.390168786,-0.1624980271,-0.0608660802,0.0683365762,-0.1990797222,0.0822419822,-0.4669093788,0.0059260186,0.2804733515,0.3146547973,-0.2105138451,-0.0311501529,-0.2271650136,-0.4472225904,0.3581000268,0.0410957895,-0.4393667579,0.0318099149,0.265956223,0.1168274358,0.2134186924,0.4657851458,-0.0183206759,0.4720543623,-0.3376360536,-0.0125897452,0.3418802023,-0.4551288486,-0.1726371199,0.2068525404,-0.171176374,-0.0108986041,0.3100948036,0.0450184084,-0.0124492357,0.2387021035,0.7105070353,0.3106862605,-0.036810305,-0.0591208488,-0.1282616407,-0.1791668236,0.3601859212,-0.1305136085,-0.071739614,0.1258464158,-0.2418709695,-0.4316968918,0.3587989807,-0.2525398135,0.0563765168,0.4634277523,0.3992080986,-0.1521523148,-0.1093364879,0.0702472553,-0.1755597442,0.2409662008,0.0380816013,-0.141144231,-0.2892437577,0.0492358245,-0.1995348483,0.0720840171,-0.1083417609,0.0309189446,0.0110549619,0.1900498271,0.0202374849,-0.128693521,-0.2456143945,0.641156733,-0.0821437463,-0.0126171969,-0.1710728258,0.562954247,-0.2242329568,-0.0949676037,-0.0774531215,-0.0047572316,0.2522273958,0.1173365116,0.0192381889,0.4581305087,-0.0293524768,0.0652101263,0.1646009833,-0.0015374377,0.1587028652,0.0926408395,0.1691076308,-0.0322373174,0.2578332722,-0.2327931523,-0.1162926629,0.3702292144,-0.3233274519,0.4082520306,0.1217195764,-0.0517719053,0.1363519728,0.0588621609,0.0029764769,-0.1529113352,0.0995619074,0.1531849355,0.2225988209,0.2118729651,-0.3046705425,-0.0127987899,0.2571274936,0.0371879302,0.0053212987,0.0151513517,-0.0527290627,-0.1384252459,-0.2007956654,0.3541591763,0.7381997108,0.1167203337,0.1836723089,-0.0615902022,-0.156595096,-0.1144140214,0.2958693802,0.138366282,0.188490063,0.2836687863,0.0051927832,-0.0268551148,-0.1313907206,0.0052786744,-0.1450089365,0.0549790338,-0.5187076926,-0.0354224518,-0.1969224513,-0.0688647479,-0.3841024637,-0.3000256419,-0.0957658738,-0.5155733228,-0.2982620597,0.4832041562,0.034070272,-0.1017188281,-0.216841206,0.1896806955,0.0425127409,-0.0658033416,0.0900249109,-0.0020249798,-0.1174458414,-0.0193233937,-0.1849119067,-0.0644801632,0.0272129346,-0.1759793311,0.2759288847,-0.3246739805,0.0726150647,-0.1426634938,-0.2029588223,0.173454985,0.077978313,0.2803866863,-0.2035324872,-0.5924379826,0.1457236707,0.5013505816,-0.186918214,0.1658065021,0.2946106195,-0.2110977173,-0.0700031146,-0.2169883549,-0.3874444067,-0.2615768015,-0.0783747956,-0.196580708,0.175164789,0.0926072896,0.1931917667,0.1270978153,0.3630620837,0.1643963158,-0.1537372172,-0.367639035,0.44455567,0.0054613114,-0.1883690804,-0.0298494119,0.0075802496,0.1189812571,0.1703369021,-0.3710669279,-0.0209495798,-0.2660009265,0.1324578971,-0.0986753255,0.3205017149,0.1483410448,0.3702240288,-0.0279010814,-0.2777628005,-0.3389345407,0.1872110963,0.2447299063,0.3925762773,-0.0807114989,0.2742601335,0.1444842368,0.3211856186,0.3186561167,0.0706983656,0.3618190289,0.2049423158,0.5604659319,-0.0410338752,-0.5407702327,-0.3021301329,-0.0544505753,-0.2745864689,-0.0456636623,0.0395170152,0.0254857186,-0.1716779172,0.1394793093,-0.2076471746,-0.2167317271,0.0467615537,-0.2657444477,0.1350404322,-0.1639406234,-0.0409942605,-0.2967512608,0.2964035273,-0.1615643948,0.127874583,0.2145499289,-0.0778273419,-0.3957501054,-0.34474051,-0.4116416872,0.3273962736,0.1121813953,0.3608625531,0.053526774,-0.2476403117,-0.2672664225,0.0001896014,0.8296086192,0.2859653533,-0.3504905105,0.4240784049,-0.0052016666,-0.4869101644,-0.0699313059,0.0651897341,0.195327282,-0.3118861616,0.4359986186,-0.2213545591,-0.0214835536,-0.1237546727,0.5002299547,-0.1325320303,0.0295929424,-0.1533931345,-0.2113338411,-0.1027412638,-0.2540705204,-0.013713235,0.4512710571,-0.3278296292,-0.0148410872,0.0140141016,-0.282543391,0.3552677333,-0.3786018789,0.5898547173,-0.0837832689,0.1461852193,0.0838903934,0.1358466148,0.3902834654,0.4980373085,-0.267896831,-0.0898630843,0.0580673255,0.0452432483,0.4692693353,0.2317836434,-0.2059552222,-0.1364581436,-0.0980528817,-0.0807299763,-0.3447636068,-0.0482692309,0.3455119431,0.0466461256,-0.3898504376,-0.3483811319,0.2334392816,0.3130782545,-0.1007195413,0.3356810212,0.0064239041,-0.4863872826,0.3060374856,-0.1692826152,0.9695227146,0.069070667,0.285823524,0.2072306275,-0.2142317146,0.0614585355,0.253698945,0.081127584,-0.154038921,-0.2758868337,-0.1287540495,-0.31115821,0.0397877209,0.2840384841,-0.3629772961,0.3248740137,-0.4089421034,-0.1213365197,-0.1199786142,0.0567536168,-0.1175730228,-0.1675236672,-0.3534539342,0.0193218589,0.173236981,0.0113885999,-0.0091675371,-0.041998703,-0.4983077645,-0.2119645923,-0.129663229,0.1295048296,-0.2722189426,0.661272645,-0.3607723713,-0.4087208211,-0.4596438706,0.353454113,0.4468973279,-0.0406765901,0.0203794446,-0.0941185877,-0.004010241,0.1347249448,0.0148610026,-0.2953517437,0.1934388876,0.1635867059,-0.0587041639,-0.0635411143,-0.0726569965,-0.3442114294,-0.0935278535,0.0591619015,-0.0022294039,-0.2073688358,-0.3623820841,0.1780844182,0.0290744882,-0.0666610897,0.0257020872,0.3393504024,-0.0507446788,0.3158523738,-0.3541829288,-0.3773806989,-0.0739707276,0.5312651992,0.224258393,-0.0562743843,0.412694633,-0.0774751902,-0.1203335747,-0.0022459861,0.3146800995,0.1522158235,0.2107281834,0.3843662739,-0.0186378248,-0.0036370526,-0.0969064161,0.3600209057,0.3548289835,0.1859159768,-0.0842136219,-0.2710140646,-0.4922533929,0.3682181835,0.1036622003,0.3379027545,0.0438439809,-0.3522379398,0.0508228429,-0.1391114146,-0.1926677078,0.0128736282,-0.2097492069,0.0421811491,0.2238101214,-0.0165762641,0.185569182,0.0448212028,0.0728660151,-0.1860189438,-0.005239428,-0.0883277729,0.093125537,0.1873927265,0.111086376,0.0119039658,-0.005746101,-0.4661687016,0.025892619,-0.2465316802,-0.0944527313,0.579357028,0.0175719503,0.1690877527,-0.01962726,0.2136960626,0.0616988242,0.076389119,-0.0025381034,0.1215912774,0.0036704598,0.2305388749,-0.2393221706,-0.1587927341,-0.383444339,-0.0233675651,-0.1116823852,0.0098162675,0.2292733341,-0.1504137814,-0.1484252363,-0.0124735637,0.3728413284,0.353838563,-0.1588054001,0.1364240199,-0.0716918409,0.0865088478,0.0492604636,-0.2904603779,-0.194779858,-0.2292595357,-0.2107806951,0.1381398737,-0.0088702487,-0.0149322385,-0.4017885923,-0.0100989677,0.7563078403,-0.1419937313,0.2795636058,0.0461813994,0.0902781487,0.0505699143,0.0326937176,0.1408356726,0.3030175269,0.0640622526,-0.1083740443,0.1676545739,0.3823309541,-0.1280300617,0.2609055936,-0.1799259037,-0.0088662179,0.1120494902,-0.1318459362,-0.0275410786,0.0511896536,0.1771026701,-0.5174019933,-0.1464341283,0.0079960926,0.3008130193,-0.1359507889,-0.1351794153,0.0749864504,-0.2732968032,0.0489555933,-0.1438182741,0.1196195334,-0.2969895601,0.1513877362,-0.0544957258,-0.0423917249,-0.0471481271,0.2694114149,0.387237817,0.0599572845,-0.2542601824,0.0099732159,0.211407885,-0.054467909,0.1391242743,0.2713523805,0.5723224282,0.5754204392,0.0224327389,0.0843969285,0.042722784,-0.2839963138,-0.0437580347,0.0650865957,-0.3334129751,-0.3743065298,0.3329984546,0.1922887713,0.0142770316,0.0668353364,0.2160837501,0.507180512,-0.2229152173,0.2355360985,-0.1977863908,-0.2077814639,0.1037065536,0.0138595281,-0.3823937774,-0.2068970948,0.3432179689,-0.0100476034,0.1653649658,0.196069032,0.0451611802,0.061803408,0.1946893036,0.2713280022,0.0085707651,-0.4885701239,-0.014843246,-0.4134443104,-0.1725321561,0.3790316582,0.1948206276,-0.2688859701,0.4094080329,0.1388222128,-0.0694585145,0.1629866213,0.1621203125,-0.1625233591,0.0518666543,-0.4363815188,-0.1906347573,0.2093261331,-0.0331981592,-0.0653619319,-0.3189267814,0.1869718879,0.1655371487,0.0406783372,-0.0085661151,0.0848994106,0.1778649986,0.439478606,0.2291012853,0.198749274,0.3351501822,-0.1279801279,-0.0079934886,-0.0180618502,-0.2785053849,-0.0973230675,0.0944773629,-0.1241475493,0.4446907043,-0.512281239,0.0918894634,-0.1592813879,0.0154882232,0.0713982284,-0.3703894019,-0.1453998238,0.0396948531,-0.0367695726,0.0127521893,0.1586250067,0.2262328714,0.0452177711,0.5562224388,0.0172788762,-0.4441852868,0.367926687,0.1131617129,-0.1632418931,0.0926662013,-0.0541829839,0.1721479297,0.3824457526,-0.5713046789,-0.1840988547,0.4492428601,-0.2179424018,-0.3280405104,0.0755852386,-0.1019627005,0.0461992174,-0.0079678595,0.3582443595,-0.132250309,-0.3258394003,0.0553666241,-0.1872070134]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2040","title":"ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk","comments":"In the next release you'll be able to concatenate any kinds of dataset (either from memory or from disk).\r\n\r\nFor now I'd suggest you to flatten the indices of the A and B datasets. This will remove the indices mapping and you will be able to concatenate them. You can flatten the indices with\r\n```python\r\ndataset = dataset.flatten_indices()\r\n```","body":"Hi there,\r\n\r\nI am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):\r\n```python\r\nconcatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])\r\n```\r\nYielding the following error:\r\n```python\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\nBeen trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...\r\n\r\n`load_from_disk(PATH_DATA_CLS_A)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 785\r\n})\r\n```\r\n`load_from_disk(PATH_DATA_CLS_B)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 3341\r\n})\r\n```","comment_length":59,"text":"ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk\nHi there,\r\n\r\nI am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):\r\n```python\r\nconcatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])\r\n```\r\nYielding the following error:\r\n```python\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\nBeen trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...\r\n\r\n`load_from_disk(PATH_DATA_CLS_A)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 785\r\n})\r\n```\r\n`load_from_disk(PATH_DATA_CLS_B)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 3341\r\n})\r\n```\nIn the next release you'll be able to concatenate any kinds of dataset (either from memory or from disk).\r\n\r\nFor now I'd suggest you to flatten the indices of the A and B datasets. This will remove the indices mapping and you will be able to concatenate them. You can flatten the indices with\r\n```python\r\ndataset = dataset.flatten_indices()\r\n```","embeddings":[-0.0903852209,-0.0803103447,-0.0217147637,0.4296606481,0.2122907937,0.2667319477,0.0407163389,0.1795164198,-0.0718934461,0.174695462,-0.0009916488,0.1995578706,-0.0881771892,-0.0326921009,-0.3439054191,-0.0558386669,0.1930517405,0.0187105779,-0.1041580364,0.0098615009,-0.4442354441,0.243652001,-0.155220598,-0.0359293222,-0.311473012,-0.0647420511,-0.3580109477,0.3418426216,-0.1153114215,-0.1206123084,0.5508387685,-0.0477537625,0.0869863778,0.5565166473,-0.0001167422,-0.0058714054,-0.0398115367,-0.159683913,-0.2412981391,-0.4698682725,-0.2993323505,-0.2401568294,-0.0213559829,-0.2968689799,0.208661586,-0.1860080361,-0.1097184047,-0.7185208797,0.2358155102,0.2194238901,0.131616354,0.1606049538,0.1590324193,-0.0537228845,0.0060946266,0.0604138039,0.1772609204,0.1795203984,-0.1299273819,-0.0358634815,0.2642757595,-0.0068642069,-0.3868632019,-0.0542209633,0.1337354779,0.4457705021,0.0942109674,-0.4561156034,-0.1477254331,0.0752305239,0.4200157225,-0.3262958825,-0.3272228241,-0.091366224,0.0971419364,-0.3858567774,0.1183445603,0.3693805337,0.0639820844,0.1274228245,-0.2128302157,-0.0889138207,-0.1895785928,0.0065884353,-0.115011625,0.0218899827,-0.0210954417,0.1646567136,0.5099681616,-0.0325126573,0.4155707657,-0.5363280773,0.0476031303,0.1897673905,-0.199121505,0.0359905288,-0.0857117847,-0.546063602,0.0706862956,-0.1810738593,0.1162746027,-0.1160381213,-0.1602855325,0.3259332776,0.1867987663,0.4233537316,0.0459591895,0.5026447773,0.0342941098,-0.0802171454,0.0644280985,0.0524058603,0.0719278976,-0.3668627441,0.0957351327,0.0153415883,0.1143159419,0.0008162334,-0.4245216846,0.1277352124,-0.2855323851,-0.0848076418,0.0569595098,-0.0795739964,0.0466366075,0.3995075226,0.3896420598,0.3132678866,0.0783832967,0.2504098117,-0.1776385903,0.176717788,0.1420002282,0.1240313724,0.152634263,-0.0639882907,-0.0817171186,0.1919679046,-0.0037851273,0.0142077599,0.0226849988,-0.359480828,0.1906329244,0.1651127189,0.1070034131,0.2670373321,0.3068104684,-0.3512372673,-0.1848161668,0.3993564546,-0.4019362032,-0.1600684375,-0.1483420432,0.1177690029,-0.1106535122,0.0929520577,-0.4305930138,0.1036647558,0.2892502248,0.2214403301,-0.2550332248,-0.1051570699,-0.2563718557,-0.3879457414,0.3073952496,-0.0501087978,-0.431759268,-0.0085132178,0.2673155665,0.1099048629,0.2351316512,0.4662023783,-0.0782097131,0.3340336978,-0.2627853453,-0.0979297161,0.1698978543,-0.3320308328,-0.1988217384,0.1342197955,-0.1535136253,0.03503225,0.1987936944,0.0435760245,0.0285198595,0.2498156577,0.7739896774,0.3166610599,-0.0380970016,-0.1191873774,-0.1531618983,-0.2666220963,0.4661263227,-0.1315011382,-0.1416645795,0.0988497511,-0.3725159168,-0.397574544,0.3472753465,-0.2832825184,0.0636298358,0.4213288426,0.3263689578,-0.1817166209,-0.1018906087,0.024623448,-0.2054289579,0.1895036548,-0.0335666239,-0.1018203795,-0.2266622931,-0.0587005801,-0.1673020273,0.0187402908,-0.0538976155,0.141298905,0.0392029732,0.0141676487,-0.0718826056,-0.1022611409,-0.2673949003,0.5654734969,-0.090332523,0.0383026525,-0.2548471689,0.4797309637,-0.1687238365,-0.0423460342,-0.0257890597,-0.007185135,0.2499664873,0.098609969,-0.0060409666,0.4919343293,-0.0936736017,0.0119200563,0.0779165924,0.0159203801,0.227192685,0.0178498216,0.0769848973,0.0056219194,0.1291652322,-0.1870341152,-0.0795801282,0.4245119989,-0.3667976558,0.3369196951,0.0314631686,0.0869895667,0.2149218321,-0.0200444143,-0.0327523798,-0.2353004366,0.0180129632,0.0077934843,0.0598469116,0.1597383618,-0.3271050751,-0.0502549335,0.1370064467,0.0435917601,-0.008159495,0.0476348028,-0.0792988464,-0.0775588676,-0.1422454268,0.2626615167,0.595733881,0.205141291,0.0933410451,0.0301323123,-0.1562049389,-0.0462851711,0.3124066889,0.1212166324,0.1940273345,0.2194419503,-0.0138920229,-0.0226647984,-0.0724640936,0.0575230457,-0.1133960783,0.1180530414,-0.5014936328,-0.0962722301,-0.3015748262,-0.1021734476,-0.3828296959,-0.3004683256,-0.1973970681,-0.4838761389,-0.2615421116,0.4586673677,-0.0590238944,0.038206771,-0.1270175278,0.1683130711,0.0720048472,0.0265385341,0.1558901817,-0.0216734577,-0.0583191738,0.0444997177,-0.0778574124,-0.0622468777,0.0591107383,-0.1217503026,0.2177443206,-0.2410200834,0.0385862365,-0.150095433,-0.1827576756,0.0993428677,0.0969478637,0.2907016277,-0.2663249373,-0.5855751038,0.1232848167,0.5215249062,-0.1266570985,0.1730052233,0.2626032531,-0.182980448,-0.0798323452,-0.2237952054,-0.3680772781,-0.2371441722,-0.0798078775,-0.2592636049,0.2472724766,0.0990316272,0.2649503946,0.0820237175,0.3654113412,0.1007734388,-0.1446047574,-0.287836194,0.3886061311,-0.0084413933,-0.1790605932,0.0065309382,-0.0050052297,0.086275056,0.1664551497,-0.3701839447,-0.0095428824,-0.1804279238,0.096353583,-0.1326226592,0.2878193855,0.1116801575,0.4846867621,-0.053226389,-0.1380140185,-0.3144214749,0.1558962464,0.1922058761,0.3816485405,-0.0430138633,0.2722981274,0.0688322559,0.2513549626,0.3474906385,0.1060320064,0.2202700973,0.268048048,0.467143923,-0.0679680854,-0.5600995421,-0.3960489333,-0.0207857993,-0.20792678,-0.0920800641,-0.0302984901,0.0046056244,-0.1437469721,0.1522496492,-0.1051038578,-0.2345129699,0.088960968,-0.2979528606,0.1742399037,-0.1462112218,-0.1123135909,-0.3172487617,0.2338033915,-0.1621076763,0.00723049,0.1826259196,-0.1477534026,-0.394855231,-0.3019553721,-0.4231874943,0.4284077883,0.094011806,0.360036999,0.1029912233,-0.211602971,-0.3140146434,0.0644621626,0.7793235779,0.2428842485,-0.4793565869,0.4762117565,0.0486611277,-0.5282108188,-0.0892103091,0.0776744783,0.1703041941,-0.2709617913,0.5176336765,-0.2395984232,0.0078170672,-0.0545135066,0.5147518516,-0.1700829118,0.0628878102,-0.0855596438,-0.237239331,-0.1355617046,-0.2421569973,-0.0071065784,0.5239639878,-0.2932553291,-0.0072202967,-0.0405998677,-0.2700951099,0.4007339478,-0.3850299716,0.5835527182,-0.0071734251,0.1568863988,0.1716206223,0.1505173892,0.34743312,0.485283941,-0.2380062342,-0.1381063312,-0.0185343064,0.1223378032,0.4762547016,0.2324672788,-0.120175451,-0.00480195,-0.0342460759,0.017818043,-0.3697012663,-0.0657593459,0.2507535219,0.0280167982,-0.4654098749,-0.4313584268,0.2042631209,0.3102700412,-0.1245822385,0.3593559563,-0.0323423706,-0.470258981,0.3982971013,-0.1162720695,0.9791230559,-0.0679161176,0.1978092492,0.1263821572,-0.1889854074,0.1685775518,0.2858894169,0.1136950701,-0.2019079626,-0.2907350957,-0.1255354285,-0.2693724334,0.0491759032,0.2388661653,-0.4425667822,0.2669600844,-0.4332692623,-0.0990274921,-0.1775871962,0.0657880455,-0.0341670066,-0.2173333615,-0.3752557039,0.0614647828,0.0879903808,-0.1109582484,0.0080184601,-0.0633541495,-0.3858080804,-0.1335165352,-0.110168986,0.0976584777,-0.1991309077,0.7657146454,-0.3843827844,-0.3101261854,-0.3907933235,0.258900106,0.4397601783,0.0024202901,0.0508848876,-0.1327612549,0.0165405814,0.042283792,-0.0125446999,-0.245363906,0.1864331365,0.1857319772,-0.1204650849,0.0721949935,0.024608966,-0.3490176499,-0.1142052934,0.0029146983,-0.0094681717,-0.2997998595,-0.3640498221,0.1790685803,0.1427507102,-0.0608195253,0.0668348223,0.329407692,-0.0538921356,0.2757905126,-0.3172311783,-0.3427124023,-0.0965484828,0.5646895766,0.3627413213,-0.02367099,0.4135935307,-0.0395331457,-0.0701676309,-0.0161433928,0.3499980867,0.1511656344,0.2037112415,0.4430519342,0.0101718716,-0.0109118968,-0.1815682501,0.3628650606,0.3714293242,0.2277030796,-0.0420892462,-0.164664045,-0.5605289936,0.3672233522,0.1599869877,0.3651226461,0.0818182975,-0.3125768602,-0.1021243855,-0.0895860419,-0.2543860078,0.0288697407,-0.1790504754,0.0381929129,0.1982562095,-0.040777538,0.131346941,-0.0527273901,0.137776047,-0.2134715915,0.0124422051,-0.1549429446,0.0620771796,0.1388073713,0.0865528956,-0.1200791821,-0.0210390892,-0.4546807408,-0.0239886083,-0.2092135698,-0.094084695,0.6052910686,0.0361986086,0.1405508667,-0.0611778051,0.1395591646,0.0812328011,0.0595501624,0.0576574057,0.0565805361,-0.0629660785,0.2205439806,-0.283511579,-0.1896280199,-0.3457252681,0.0105383079,-0.1799463183,-0.0584227778,0.2393825352,-0.1386640966,-0.2037059814,0.1689529568,0.3811874092,0.3318369687,-0.1980982274,0.1788566113,-0.0079739131,0.141233027,-0.0037085523,-0.2662277818,-0.0902473181,-0.1298976541,-0.2266336679,0.0693537518,-0.0230853446,0.024143545,-0.3969614804,0.0425187275,0.6044135094,-0.172918573,0.3787428141,0.0949078277,0.1064308882,-0.0563941188,0.0330811851,0.1187555045,0.4463941157,-0.047361739,-0.1196499839,0.0988508537,0.3390355706,-0.1144941151,0.2200872451,-0.2309978604,0.0215810947,0.2077857703,-0.1855534762,0.0140400408,0.1407551616,0.2351659089,-0.5215612054,-0.2858961225,0.0506993197,0.4173052311,-0.1760130674,-0.1280167997,0.1223991513,-0.1778519452,0.0480354875,-0.1471846104,0.1240582243,-0.3219695091,0.1912062913,-0.0857975036,-0.0282375775,-0.034460444,0.244611308,0.4152513444,0.1087239087,-0.2196062952,-0.1193915233,0.2977906466,-0.0735249743,0.1069012284,0.3158182204,0.554058373,0.5522788763,-0.0312055312,0.1308662891,0.0689574033,-0.204862386,0.0429761522,0.0134647796,-0.2174228132,-0.2798497677,0.4076700509,0.1971636862,-0.030816013,0.0530910827,0.1769339591,0.5629740953,-0.176153779,0.1420214772,-0.0943724439,-0.1839530468,0.1142069176,0.0349502899,-0.3102140725,-0.2542949617,0.3667085469,-0.0110502187,0.143546015,0.2448738366,0.0854234099,0.063178651,0.1999257058,0.3337058425,-0.0418820269,-0.5096418262,-0.0054250415,-0.4245630205,-0.1814159751,0.2851063907,0.2403236926,-0.1437005252,0.3626502156,0.1679267138,-0.0371577777,0.0659625903,0.0472927205,-0.190937981,-0.0524985604,-0.3705115914,-0.1487302184,0.2106342614,-0.0604470186,-0.0587445609,-0.2856874466,0.2719460428,0.1523067355,0.0535925888,0.0091661876,0.1640532017,0.1470794082,0.3854424059,0.194820255,0.2095012814,0.3070428371,-0.1152507365,0.0316455811,0.0195721425,-0.3304347098,-0.1423636228,0.1332145482,-0.086014457,0.5072710514,-0.529245913,0.035310097,-0.1213726848,0.1196242794,0.0159203298,-0.4145380557,-0.2292204946,0.1031463593,-0.0113652907,-0.0491565131,0.0773865506,0.2998973131,0.0154600013,0.5862027407,0.0981467664,-0.4228115082,0.404863745,0.113264598,-0.1494789869,0.0822545588,-0.0872515514,0.204318732,0.3383574188,-0.5801483989,-0.2121742368,0.4233350158,-0.2602575123,-0.3111741245,0.1098243743,-0.1028584391,0.0029291385,0.0121321185,0.4039597511,-0.064184241,-0.4016014338,0.0237527173,-0.2212913185]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2040","title":"ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk","comments":"Indeed this works. Not the most elegant solution, but it does the trick. Thanks a lot! ","body":"Hi there,\r\n\r\nI am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):\r\n```python\r\nconcatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])\r\n```\r\nYielding the following error:\r\n```python\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\nBeen trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...\r\n\r\n`load_from_disk(PATH_DATA_CLS_A)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 785\r\n})\r\n```\r\n`load_from_disk(PATH_DATA_CLS_B)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 3341\r\n})\r\n```","comment_length":16,"text":"ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk\nHi there,\r\n\r\nI am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):\r\n```python\r\nconcatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])\r\n```\r\nYielding the following error:\r\n```python\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\nBeen trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...\r\n\r\n`load_from_disk(PATH_DATA_CLS_A)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 785\r\n})\r\n```\r\n`load_from_disk(PATH_DATA_CLS_B)['train']` yields:\r\n```python\r\nDataset({\r\n features: ['labels', 'text'],\r\n num_rows: 3341\r\n})\r\n```\nIndeed this works. Not the most elegant solution, but it does the trick. Thanks a lot! ","embeddings":[-0.0483823046,-0.0796561316,-0.0216175411,0.4603260756,0.1872414052,0.265417695,0.0407092683,0.1193434298,-0.0479684733,0.1485673636,0.0525876097,0.1789602637,-0.0630663633,-0.0712663308,-0.2987713516,-0.06330055,0.2559678555,-0.0143304858,-0.1243883595,-0.0335024521,-0.4365563393,0.2530574501,-0.1242119372,-0.0446646698,-0.3036776781,-0.0668847337,-0.4069908559,0.355433315,-0.0752542615,-0.1231013164,0.5189289451,-0.020126231,0.0941633731,0.5816605687,-0.0001213742,-0.0043541668,-0.1039503515,-0.1759781688,-0.2422228158,-0.4407136738,-0.2366058677,-0.2508820593,-0.0364469998,-0.2931851149,0.2539064288,-0.1341369748,-0.1175958812,-0.7231462002,0.2457882315,0.1986134201,0.0922247693,0.1944197714,0.065648973,-0.0409409329,-0.0409295559,0.1304789186,0.199579373,0.1953564584,-0.1021021456,-0.0277385227,0.2431663722,-0.0192548074,-0.3580238223,-0.0478648692,0.1482318491,0.4531691968,-0.0053289556,-0.436386317,-0.1360747218,0.0079884473,0.4547097087,-0.3950954974,-0.3489311039,-0.117168285,0.0837752298,-0.3639191389,0.1145964861,0.3478938043,0.007878786,0.1255422384,-0.2352766544,-0.1032309085,-0.2160908431,-0.0336089022,-0.1303815842,-0.023333637,-0.0100513063,0.1940192133,0.4782025516,0.0310361367,0.3325331509,-0.5216777325,0.0564999692,0.2090531737,-0.1852713227,0.0392924212,-0.0924744084,-0.5081064105,0.0348231904,-0.1193098724,0.1242330447,-0.1469303817,-0.1655237377,0.3356410861,0.2434247881,0.3892985284,0.010719114,0.4672593474,0.0521024577,-0.1690679193,0.0195897762,0.0288330633,0.0493382998,-0.318564862,0.139673546,-0.0642719492,0.0394785851,0.0019535716,-0.4597357512,0.2338107228,-0.2746943831,-0.1261484176,0.1121941805,-0.0276267305,0.0581164844,0.398830235,0.4030522704,0.3217599392,0.0658205152,0.2085612118,-0.1650107056,0.2475845665,0.1256020069,0.1402796507,0.1536212116,-0.0288812853,-0.0635674819,0.1218630672,0.0053790906,-0.0764348656,-0.0092227235,-0.3630711734,0.1362043172,0.1413986683,0.1619116217,0.3276949823,0.2981003225,-0.3494321108,-0.1738559753,0.4263890684,-0.3527106047,-0.1570555419,-0.1083386615,0.0766615868,-0.1636047661,0.080162622,-0.4262652695,0.0944418833,0.3520645201,0.2512075007,-0.2572674155,-0.054867126,-0.2776347995,-0.4151387811,0.3119912148,0.0014741232,-0.4745102525,-0.0321234763,0.2938989103,0.0989334956,0.1729961336,0.4579010904,-0.0473457612,0.4347834289,-0.2784132063,-0.0760162249,0.2648218274,-0.3892309666,-0.1594605595,0.149980098,-0.1660080254,0.0081460131,0.2545245886,0.0666961074,-0.0031891796,0.2272637486,0.7549238205,0.3503751755,-0.046582669,-0.0890982896,-0.1083980575,-0.2029781938,0.4351123869,-0.1151676327,-0.0740921944,0.0758219287,-0.3071580231,-0.4076962173,0.3597152233,-0.2330857515,0.0511181466,0.450995326,0.3342948854,-0.1403176188,-0.122274965,0.0558319539,-0.1654948592,0.1752090454,0.0425067358,-0.1681343615,-0.2135870457,-0.0231748912,-0.2340223938,0.0914140269,-0.0404028706,0.1261012405,-0.0060419003,0.0984412059,-0.0675806254,-0.0924200416,-0.2483515292,0.6367930174,-0.1156131551,0.0506182648,-0.2480134964,0.5165681243,-0.1867624074,-0.0749849826,-0.0716303736,-0.0034891213,0.2698657811,0.1431924254,0.0232766606,0.5011243224,-0.0864258185,0.0465389751,0.1062584594,-0.028330043,0.2086668611,0.0641518459,0.1096678078,-0.0449994393,0.18428877,-0.2220497131,-0.1296049505,0.4378694296,-0.3333467841,0.3640969098,0.0791068748,0.0575429425,0.1631729454,0.0447131731,0.0018771293,-0.2020855397,0.0737311915,0.101402849,0.1597453803,0.1903139651,-0.3452679217,-0.040562287,0.1253634542,0.0527346209,0.013652334,0.0142112365,-0.0673974454,-0.0898621976,-0.130638808,0.3200011551,0.6371323466,0.132965669,0.0911492929,-0.0353958495,-0.2003272921,-0.0530101769,0.2857239246,0.1153876632,0.1698904634,0.1835750788,-0.0004646585,-0.0553121045,-0.0652937517,0.0293046031,-0.1215849295,0.0884269476,-0.5587563515,-0.0408842415,-0.2592019141,-0.0891951248,-0.4326558411,-0.3007506728,-0.178582266,-0.5216221809,-0.2752329409,0.5282120705,-0.0444639958,-0.0342006795,-0.1715994179,0.1994298846,0.0160332117,0.0158332027,0.2233744264,-0.0164536368,-0.0683609843,-0.0021434813,-0.1374316663,-0.0709894672,0.0867682248,-0.1476976126,0.210637629,-0.2377523184,0.0644207671,-0.1513043344,-0.1804700792,0.1254289597,0.0607750416,0.2815028429,-0.282474488,-0.6274942756,0.1269658953,0.5603994131,-0.1214346066,0.1552202851,0.2652809322,-0.2035024464,-0.0397576801,-0.1804067492,-0.3837980926,-0.2179225087,-0.0579305477,-0.2351757139,0.2247659117,0.1102717146,0.2297351211,0.0725850835,0.3910377622,0.1411735415,-0.1505566984,-0.3173216283,0.4028314948,-0.0106664188,-0.1393475533,0.0115737915,0.024893418,0.0791704804,0.1679452807,-0.3714910746,-0.0036198429,-0.1924351007,0.1260193437,-0.1266137511,0.2927983701,0.0760923475,0.4290399849,-0.0072637009,-0.215341568,-0.3689443767,0.1965474933,0.1868611574,0.3983032107,-0.0778275356,0.2956408262,0.1176374331,0.2697820961,0.301225692,0.0938747302,0.284866035,0.2522512376,0.5076844096,-0.0031347922,-0.5560578108,-0.3561556935,-0.0455510914,-0.2627851069,-0.1099143624,-0.0035718877,0.0282401051,-0.1696554869,0.1573153585,-0.1575006843,-0.2193347067,0.0123280017,-0.2399486005,0.1458227783,-0.1426558197,-0.0676001608,-0.3264603615,0.2706416547,-0.1785920858,0.0445348583,0.2091866434,-0.1165117919,-0.4453098476,-0.3459737897,-0.4078504145,0.3821882904,0.0803226158,0.3729103208,0.0760275871,-0.2197804451,-0.2744779587,0.0357371531,0.8388231993,0.2913853824,-0.4263603389,0.3884181082,0.0263928901,-0.5281162262,-0.0850612968,0.1199571192,0.1918411404,-0.2607576847,0.3966219127,-0.2330010235,0.0217178743,-0.049145788,0.5023084283,-0.1987692714,0.0102226948,-0.0815036297,-0.2123834789,-0.1326718479,-0.21986489,0.0007041501,0.4571669996,-0.3032931089,0.0313693769,0.0264570955,-0.2583820224,0.409430027,-0.353621304,0.6024320126,-0.0385146439,0.1886684746,0.128686592,0.130033344,0.3836228848,0.5007213354,-0.2303155214,-0.1833070964,0.0078261904,0.1063215733,0.4872787297,0.2422619164,-0.1522426903,-0.0383133776,-0.0520326756,-0.0673809275,-0.3485023081,-0.0687056407,0.2765888572,0.0112873577,-0.4571103752,-0.4127927721,0.2228419632,0.3240772784,-0.118054226,0.3458712399,-0.03905439,-0.4746818841,0.3770340383,-0.1490742564,1.0115308762,-0.0302794352,0.2235889882,0.1392248422,-0.2224709988,0.1609476954,0.2818366885,0.0586570501,-0.1898101866,-0.2558055222,-0.1155659854,-0.2795865834,0.0443630442,0.2588219047,-0.39289698,0.3099092543,-0.4455198348,-0.1018280238,-0.164174363,0.0484611765,-0.0464083888,-0.2220676988,-0.352137357,0.0185462367,0.1050286144,-0.0609446354,-0.0374624245,-0.0209582634,-0.4371869266,-0.158318758,-0.1735188961,0.0985001326,-0.1609060913,0.7278354764,-0.3450314403,-0.3661651015,-0.4455928206,0.331671834,0.4819719791,0.0272447336,0.039823968,-0.1074399352,-0.0060060085,0.0863997936,-0.0281226784,-0.2376881242,0.1929191053,0.178489536,-0.0919377208,0.0154875759,0.0067068865,-0.3319256306,-0.1330366433,0.0451705791,0.000028787,-0.273517698,-0.3578084409,0.1885162294,0.1014325321,-0.118404597,0.0265281871,0.3572260439,-0.0651484951,0.2823158205,-0.3681316376,-0.3245562315,-0.0924424157,0.5428097248,0.3373905122,-0.070923686,0.3994652033,-0.0069440617,-0.1154888719,0.0111194886,0.3332043588,0.1729922742,0.1851282865,0.4267687798,0.006929717,-0.0070109838,-0.1956735551,0.3540216088,0.3503943682,0.1956197172,0.0313796923,-0.2186260074,-0.5530782342,0.3396945596,0.1706299186,0.327675432,0.104965046,-0.3088966906,-0.0684683397,-0.0234604608,-0.2058250755,0.0000222643,-0.1870582998,0.0434532166,0.2184444964,0.0287768766,0.0832636654,-0.010145531,0.098774828,-0.1689889878,0.0233571287,-0.0975207165,0.0628541037,0.1782514453,0.0738009214,-0.0904068574,0.0051430743,-0.4593447149,0.0140078655,-0.2294316739,-0.1217431501,0.6715194583,0.0510695688,0.1214961782,0.0158857554,0.2285824269,0.0680389032,0.0878005177,0.048475083,0.0963512957,0.0100165596,0.2222524285,-0.2736035883,-0.1764546186,-0.3879478574,-0.0089697735,-0.1200424805,-0.0471004508,0.2031809688,-0.1601869613,-0.2005193383,0.1108040661,0.4110427201,0.3873534799,-0.1979260296,0.1900835633,-0.0304580145,0.0874345973,0.016890822,-0.2827236056,-0.098320365,-0.1849257201,-0.275777638,0.1254781634,0.007580563,0.0357762501,-0.3907299042,0.0127613926,0.6847847104,-0.1335172802,0.3405169845,0.060774751,0.1156374291,-0.0105971433,-0.0250713229,0.1619508117,0.422331214,-0.0091833761,-0.1420733333,0.0973931178,0.3798938692,-0.1379367113,0.2299740314,-0.2404863089,0.0408821814,0.1885399818,-0.165081352,-0.0241939407,0.0771467015,0.1839596033,-0.5160553455,-0.2480302453,0.0021000374,0.42480281,-0.138577953,-0.1343550831,0.0907642022,-0.2246541828,0.002783933,-0.1104549021,0.1289765984,-0.3996534348,0.1721117049,-0.0819720179,-0.0096011003,-0.0884574726,0.2355580628,0.3815259039,0.0721736029,-0.2652971745,-0.0507384874,0.2551621795,-0.0694527254,0.1916284561,0.2259948701,0.5554209948,0.556127727,0.0255504698,0.1406385452,0.0032016756,-0.2122265399,-0.0053526843,-0.0025255396,-0.2935774028,-0.3012850285,0.4094851613,0.1750785708,0.0024357964,0.0674763843,0.2661316991,0.5477368236,-0.229787454,0.156458661,-0.1596251279,-0.2005819976,0.0917253122,0.0024659804,-0.3389194906,-0.2537000477,0.3616604805,-0.0187023431,0.1398175061,0.2566127181,0.0528876446,0.0547341667,0.1891521662,0.2871243954,-0.0130728511,-0.5009988546,-0.0356708094,-0.4185153842,-0.1472506672,0.3142979443,0.1910641193,-0.1895875186,0.3773264289,0.1416850388,-0.0410860889,0.1177175418,0.1273605227,-0.1753463596,0.0234478805,-0.393068552,-0.1863916665,0.2176950723,-0.0762482136,-0.0608633533,-0.2366347015,0.2287172824,0.1713470221,0.0196565483,0.0119344462,0.0836449862,0.1588283181,0.414206177,0.2572519779,0.1691017896,0.3638537526,-0.100317508,0.0047958135,0.0438849106,-0.3505228162,-0.158304438,0.1300311685,-0.1268360466,0.5008471012,-0.5447794795,0.0813271552,-0.1266759038,0.0849690288,0.0344079137,-0.4037499726,-0.1758655459,0.055065874,-0.0133664031,0.018315874,0.1382131279,0.2336795628,0.0256984234,0.6657631397,0.0608548261,-0.4091666043,0.400999099,0.1377356648,-0.1514993757,0.0492458791,-0.0779657289,0.182699725,0.3516889811,-0.6099100113,-0.215733096,0.4358257651,-0.2188643217,-0.2990232408,0.0933738351,-0.0947813988,-0.0106872087,0.0053799557,0.4136137366,-0.1291244924,-0.3908876777,0.0667116195,-0.2082781941]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2038","title":"outdated dataset_infos.json might fail verifications","comments":"Hi ! Thanks for reporting.\r\n\r\nTo update the dataset_infos.json you can run:\r\n```\r\ndatasets-cli test .\/datasets\/doc2dial --all_configs --save_infos --ignore_verifications\r\n```","body":"The [doc2dial\/dataset_infos.json](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/doc2dial\/dataset_infos.json) is outdated. It would fail data_loader when verifying download checksum etc..\r\n\r\nCould you please update this file or point me how to update this file?\r\n\r\nThank you.","comment_length":20,"text":"outdated dataset_infos.json might fail verifications\nThe [doc2dial\/dataset_infos.json](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/doc2dial\/dataset_infos.json) is outdated. It would fail data_loader when verifying download checksum etc..\r\n\r\nCould you please update this file or point me how to update this file?\r\n\r\nThank you.\nHi ! Thanks for reporting.\r\n\r\nTo update the dataset_infos.json you can run:\r\n```\r\ndatasets-cli test .\/datasets\/doc2dial --all_configs --save_infos --ignore_verifications\r\n```","embeddings":[-0.120092757,0.1984136701,-0.111874871,0.1850017458,0.1130970418,0.2163039148,0.1035631448,0.4946838021,0.2069722265,-0.0703191087,0.0729325563,0.0472202599,0.1981547624,0.2417432368,-0.068925254,-0.0911237523,-0.0267710928,0.2664070427,0.0681155771,0.0885860845,-0.0266231392,0.1407285482,-0.1923577338,-0.1282282919,-0.2190597206,-0.0406580307,0.0387163423,0.1655928046,-0.5414218307,-0.4992533326,0.4304624498,0.4793901742,0.0819710493,0.3855613768,-0.0001081328,-0.0376769081,0.5466258526,-0.029058706,-0.4618392289,-0.0519545451,-0.3342080414,-0.1855138689,-0.141147241,-0.0835442767,-0.0152585106,-0.3255074024,-0.1239994019,0.0789585784,0.2135807127,0.1726304293,0.2469211668,0.4211307168,0.5330948234,-0.1848715693,0.1094438359,-0.0424621068,0.1855426133,0.4435175955,0.3252074718,0.0316635482,0.2927803397,0.3482064903,-0.0668817461,-0.1015633717,0.1929742694,-0.1665298045,0.1315116435,-0.1064366922,0.1657461971,0.1777226031,0.6516272426,-0.3433517218,-0.3737469912,-0.0835310742,-0.1550707519,-0.1994565874,0.391864717,-0.2893846333,0.2753103673,0.2195625752,-0.1972355843,-0.3501672745,-0.0167721212,0.055279512,-0.0843022764,-0.1005989462,-0.1842762381,-0.1455242485,0.1457502395,-0.1432099789,-0.1317905039,-0.0046239425,-0.3207756579,0.0597509742,-0.0772146285,-0.299791187,-0.2040339261,0.1048173681,0.3750511706,0.2149078548,0.0941932574,0.0497511104,-0.3412804008,0.1715231836,0.1495141983,0.1182381883,0.4304278493,-0.0208122265,0.3718360066,0.6826687455,0.1144597232,-0.0479870811,-0.0042079687,-0.2481483668,0.0243405681,0.0238256902,0.3393543065,-0.4905768633,-0.0651530549,0.1865612864,-0.0159931201,-0.055738017,0.2356281877,0.3187642992,-0.3612932265,0.1625730842,0.1662807018,-0.0296565257,0.044611685,-0.3122760952,-0.0763921142,0.0261264704,-0.0284057055,0.215308249,0.3751752377,-0.3757915497,0.4739308059,0.0353367962,-0.1602192819,0.0439420529,-0.074450843,-0.1207977086,-0.2017621845,0.2391945273,-0.0254514012,0.0504769906,-0.0973938107,-0.0127743306,0.0273161884,-0.1816865206,-0.3045568764,-0.3762091398,-0.2405503392,0.2019519657,-0.023463089,-0.1118202433,-0.2365120947,-0.1028553918,0.1793441027,-0.4997892678,-0.0455499254,-0.0730096623,-0.155411467,-0.0220387727,0.192650035,0.3321076632,-0.0709209517,0.0900682062,0.0667346194,-0.3115293384,-0.2241204828,0.2014637589,-0.0882985443,0.1940482706,-0.2010300308,0.0603739433,-0.0404514968,-0.6165745854,-0.3859308362,0.1060364693,0.223071456,-0.093243517,-0.1347037405,-0.2367650419,0.2366414517,-0.295226872,-0.1073620096,0.0489997827,0.105088301,-0.0239576884,-0.3267243803,-0.2293679863,-0.0333115645,0.1030370221,-0.0429917984,0.0195009224,0.0344419479,0.4565271735,0.1423131675,-0.0409456752,0.0469592325,0.2812905312,0.4166460633,0.0335949063,-0.0447669327,-0.0408999324,-0.5700752139,0.2099355757,0.1407439113,-0.1254222691,-0.1250814945,-0.1638533324,-0.2609325349,-0.0597878806,-0.065704219,-0.0178748127,0.1555882245,0.1592798233,0.3321816325,0.1094373241,-0.1191442758,0.2130118906,-0.5373790264,0.1782788485,-0.2622483075,0.1779333204,-0.0304578319,0.1253200918,0.3239291906,-0.0993104726,0.0879661664,-0.1290887594,-0.0831435993,0.5027294755,0.0093866559,0.1357946843,0.2471980602,0.3728539348,0.0543974303,-0.0852174759,0.0808365643,-0.165267691,-0.0419656858,0.0343083777,-0.26733163,0.0326821022,-0.0849874839,-0.0673847646,0.2248898298,-0.0741193071,0.3140821457,-0.0297689978,-0.0388831608,-0.2751832306,-0.1777119637,-0.0751736611,0.3607105315,-0.1430140436,-0.0626577437,0.1559810489,0.4095159471,-0.0206291452,-0.0906753987,0.1984057724,0.0815259218,-0.2291123271,-0.0816606358,0.2641119063,0.2345742136,0.2276795805,-0.014012428,0.1894861758,0.0711768642,-0.0561925173,0.1673857868,-0.1691849679,-0.0097601917,0.4036227465,0.1256836802,0.0108988751,-0.265037179,0.0454338603,0.2030114681,0.516420424,-0.3069397211,-0.0780188665,-0.2033129483,-0.0223651901,-0.0027797318,-0.3437184393,-0.3578514755,-0.1349753141,0.0679993033,0.4726887941,0.1156928837,0.1248631254,-0.0538580865,0.2224537879,0.0925403833,-0.1651447713,-0.187139973,0.0658758953,-0.0979108512,0.0556772798,0.3926015794,-0.0635674596,0.3531554639,-0.524425745,-0.1744995117,-0.5462660193,-0.4836139679,0.2196639627,-0.2057814747,0.2577606738,0.2761024833,0.317279458,0.1261567175,-0.0000820585,0.155358687,-0.2760593295,-0.354603678,-0.2557950914,-0.0191665124,-0.1092161834,-0.1351924241,-0.3698330522,0.1012540907,-0.2100868821,0.1020710245,0.0008240548,0.0267242696,0.2857055366,0.081451714,0.1861861497,-0.1705827713,0.206465587,-0.2983559966,-0.7436742187,0.0950695127,-0.1993618757,-0.228820622,0.0111966496,0.2518720627,0.2966251969,-0.0640645251,-0.4512532353,-0.3920636177,-0.0903876573,0.1593727618,0.0165096372,-0.1554335654,0.395457536,-0.0997767374,-0.1225270405,-0.192751348,-0.3537536561,0.1185755879,-0.0744083971,0.3598248363,-0.0353286564,0.2783931494,-0.0069323485,0.5527372956,0.0637156218,-0.3731845021,0.2554654777,0.036172159,0.5047958493,-0.1203782111,-0.1876147091,-0.088133797,-0.0018534587,0.1378801763,0.1637531817,0.0805529952,0.1937589198,-0.2014654428,-0.3641486466,-0.4616741836,-0.1876395345,-0.1725013405,0.2224400938,0.3064265847,-0.036734622,0.0382233672,-0.1044301242,-0.0385642722,0.2388698608,0.6212214231,0.0579106547,0.1147977114,-0.2876327634,0.2032408267,-0.1122885421,0.4098138213,-0.0514042787,0.2883841097,-0.0588749163,0.0835441202,-0.087576963,-0.2908474505,0.3274446726,-0.1836207658,0.3218438327,0.0458329394,-0.1038445234,0.0235117953,-0.1497751921,-0.1295833439,0.1956035942,0.2656336427,0.4955189228,-0.4432013333,-0.3598159254,0.3211878836,0.0673751161,-0.0559841655,-0.1024619564,-0.1567430943,-0.1922605038,-0.276361227,-0.0230155885,-0.2217038721,0.0376867615,-0.1569881886,0.0427965336,0.1348345578,-0.0168492496,0.0792320967,0.2699025273,0.2004428655,0.159264788,0.2687500715,0.2128037661,0.2716595829,0.6806364059,0.3026641607,-0.0993549079,-0.1245268211,-0.201150924,-0.0846628249,0.0673402175,0.1649761647,-0.0475492589,0.1596611738,0.3370102644,-0.1489340663,-0.0169889163,0.0854194537,0.4840280414,0.0210708324,-0.2948287129,-0.2754924595,0.2434901595,0.1808879972,-0.0636723936,0.1274702251,0.103014186,-0.1275834888,-0.2348642349,0.1591079384,0.7583656907,0.0867966339,-0.3215256929,0.1678169072,-0.2078514248,0.3344697654,-0.307687521,-0.0956158787,-0.3612344861,-0.0629245266,0.0182163138,0.051825203,0.2038840204,-0.2713729739,-0.4220634401,0.2160598785,-0.2913263142,0.0490428954,-0.1281304359,0.4355084002,-0.1518864185,-0.0533982627,-0.4187335968,0.226239711,0.037785951,0.1134105027,-0.2614865601,-0.1508184075,0.0885041058,-0.4387350678,-0.2300230861,-0.0084301196,0.0856359303,-0.0768019035,0.1855610907,0.2683863342,0.1197185442,0.031668961,0.4012101889,0.2393573374,-0.3454666734,0.0814239457,-0.3129982352,-0.1795913875,-0.0088283001,-0.079256855,0.3597064018,-0.2641718984,-0.3239214122,-0.0188622661,-0.0921397805,-0.3278953731,-0.0180030186,-0.2467570007,-0.2671568096,-0.2533731461,-0.0436260514,0.0642351508,-0.1675362885,-0.2751354873,0.2132402807,-0.025732033,-0.008235544,0.0102759898,0.2115840465,-0.1840821207,-0.0640985444,0.5821517706,-0.2382154614,0.0018265438,0.444983542,0.0454298556,-0.0873587132,-0.3670882285,-0.2623551786,-0.0204441156,-0.2235721052,0.0712911114,0.1381031722,0.2234358639,0.0132099381,0.2815973759,0.2363381833,-0.0803342313,0.2098326981,-0.5782491565,-0.3554052114,0.2415507883,-0.2080074698,0.1350494772,-0.1409765482,0.1401205063,-0.0375347957,-0.0232212991,-0.3887763023,0.2520696819,-0.1794509292,0.1692464352,0.3353224993,0.0583297201,0.27808249,-0.1733772755,0.0913906768,-0.1016336679,-0.3689282238,-0.1969299167,-0.1718817651,0.1263724267,0.0359874032,0.0617918074,0.1393479854,-0.0402232669,-0.2844001651,0.0475091301,-0.0473834872,-0.0463262089,0.1311321706,0.1752661318,0.0194596723,0.2657259703,0.0524950176,-0.0032036223,0.1948533505,0.1250243932,-0.0769355372,0.1401854604,-0.2970594764,-0.1372710615,0.1731645465,0.2029390782,0.1774637848,-0.1687950492,0.1919316202,-0.2113257796,0.1138476282,0.2148859203,0.3608193398,0.5215636492,-0.1709477603,0.0289684664,0.1324632466,0.2687722445,-0.1644560844,0.0366511419,0.2581138611,-0.0537411459,-0.1529437155,0.0027491522,0.348257333,-0.2134980112,0.2003032863,0.2779766619,0.326761663,-0.1841547042,0.0532511137,0.2237270772,-0.2742609084,-0.0029423072,0.2484386116,0.0194265544,0.0884083584,0.2186546773,-0.1588412076,0.0872126892,-0.1164119765,0.3795849681,-0.0588845089,-0.2583576441,0.2092460394,0.4876078367,0.2635889947,-0.0730532631,0.0753361881,0.4748782218,0.0822157413,-0.0485141762,-0.5303442478,0.154235065,-0.1189720854,-0.1307224631,-0.2884327173,-0.3437520564,0.0124431523,0.0092099896,-0.0430592299,0.025643643,0.068822369,-0.0043091443,-0.3971511722,-0.5800516009,-0.1741507649,0.2551028132,0.1090601385,-0.1722018421,0.2795851231,0.2570461929,-0.1520544738,0.3070884645,0.7304964662,0.4203173518,0.2456785738,0.2398928702,-0.1011305377,-0.3405097723,0.0049239402,0.0588030852,0.4122327268,0.0230985582,-0.3534127772,0.2803244591,0.1893059909,-0.1752357334,0.0789258406,-0.2120951712,0.0533216223,-0.116332233,-0.0487634055,0.1541918069,0.1408227384,-0.3030187786,0.166428715,-0.3497418165,-0.0581533164,0.2189167887,0.2301871777,0.188895151,0.0403144099,0.1150403544,-0.0352208577,0.5214557052,0.3677416742,-0.1461050659,0.0032998936,-0.2817929089,-0.607765615,-0.0128704263,0.0303865988,0.1156281531,-0.0765309408,-0.0257928595,0.1060030162,0.1566839069,0.3102231324,-0.2241885811,-0.1152935699,-0.3192282021,-0.3017616868,0.2308926582,-0.1066340357,-0.0629995614,0.1245408952,-0.1716695726,0.0954400375,-0.0398961827,0.0247018132,-0.0732380077,0.3276511729,-0.2639654577,0.0717818365,0.5029471517,0.2613282502,0.4752799869,-0.2293337733,-0.2174243033,-0.4161618948,-0.0664637089,-0.2767219245,0.4328137636,0.1306508929,0.129475832,-0.0277893934,-0.1076736599,-0.2423769981,0.5078527927,0.0326120034,-0.2268681675,-0.4183292985,-0.0554708466,0.06937062,-0.0651141182,-0.0651555285,0.0281289648,0.0025854893,0.0613471791,-0.5129874945,-0.0875921622,0.5630141497,-0.0581568852,0.0108214337,-0.0710178912,0.2790162563,0.1725585312,-0.353672415,-0.6980358958,0.2993614078,0.2710722983,0.0164055675,-0.2873925865,0.3345274627,-0.0833239183,0.2824677527,-0.0970922559,0.2147715688,0.0305270869,-0.2756707668,-0.1090888605,0.0299155843]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2038","title":"outdated dataset_infos.json might fail verifications","comments":"Fixed by #2041, thanks again @songfeng !","body":"The [doc2dial\/dataset_infos.json](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/doc2dial\/dataset_infos.json) is outdated. It would fail data_loader when verifying download checksum etc..\r\n\r\nCould you please update this file or point me how to update this file?\r\n\r\nThank you.","comment_length":7,"text":"outdated dataset_infos.json might fail verifications\nThe [doc2dial\/dataset_infos.json](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/doc2dial\/dataset_infos.json) is outdated. It would fail data_loader when verifying download checksum etc..\r\n\r\nCould you please update this file or point me how to update this file?\r\n\r\nThank you.\nFixed by #2041, thanks again @songfeng !","embeddings":[-0.0857268721,0.1044914499,-0.1394559592,0.2325322181,0.1875333339,0.1512674689,0.071318388,0.4500180781,0.1751806438,-0.0317864195,0.1085600555,0.0681417063,0.2350581288,0.312574625,-0.0413378142,-0.1277010441,0.0066963434,0.2233191878,0.0351115465,0.1056710258,-0.0786317065,0.2068634927,-0.210232541,-0.1579403281,-0.171534583,0.0077096652,0.0275200792,0.1978197247,-0.5413463116,-0.5012401342,0.3992556632,0.3788878918,0.0202661864,0.354344368,-0.0001052504,-0.0446893312,0.5696935058,-0.0416184887,-0.3886764646,0.0118216677,-0.2822033167,-0.1768706441,-0.1285970211,-0.0415245332,-0.0039361566,-0.312022388,-0.080406338,0.1275375634,0.2788227797,0.1188315675,0.2883096635,0.4280460477,0.5445426106,-0.2173356861,0.110948965,0.0121380799,0.1840841323,0.4315685332,0.3044531941,0.0369584598,0.1877786964,0.4541942179,-0.0388612002,-0.0919689238,0.2579936087,-0.1853752583,0.0463249609,-0.0815444514,0.1566417962,0.1047496572,0.6073865294,-0.3033322096,-0.3559180498,-0.0898057073,-0.1854233295,-0.211418286,0.4257402718,-0.2804466784,0.2549777329,0.1945779026,-0.1740779281,-0.3787658811,0.0124036409,0.0565589927,-0.0480675213,-0.1133723482,-0.2346587181,-0.1485685706,0.1381814778,-0.1796471179,-0.0854637697,0.0360506587,-0.35254246,0.0042831949,-0.0963514447,-0.2799724638,-0.1980806738,0.0584893823,0.3153969347,0.2418911308,0.0904291198,0.0864150971,-0.3296873271,0.1677580178,0.164905414,0.119330816,0.3950082362,-0.0906450376,0.37542364,0.6012405157,0.1056767702,-0.0116009265,0.007486254,-0.3032489717,-0.0259315372,0.0586222447,0.2916142344,-0.47038275,-0.0565630049,0.1673550457,0.0245673526,-0.0746420696,0.1372628808,0.2977614403,-0.3223849237,0.1130588502,0.160257861,0.0057925996,0.0329642519,-0.3144440353,-0.1083633006,0.0406364128,-0.0075237383,0.1477899253,0.3733167946,-0.3625813127,0.4891429543,0.0000448696,-0.1311520189,-0.0023475636,-0.1282005012,-0.0876029804,-0.226651594,0.2673437893,-0.0626103207,0.0388953164,-0.0915930122,0.0735574737,0.056975577,-0.1592897475,-0.3390559852,-0.3347389102,-0.2467830926,0.2307251692,-0.0497631207,-0.115590252,-0.2619443834,-0.0831243098,0.1342260242,-0.501109004,0.0071686148,-0.0397413,-0.1422120035,-0.0698445514,0.1510945857,0.2774855196,-0.0161003768,0.0446667746,0.1161651313,-0.3395305276,-0.2015574723,0.259339571,-0.069652997,0.0895810574,-0.1894998848,0.1100166216,-0.0341514684,-0.5247470737,-0.4081273675,0.1752500683,0.1859287918,-0.1092510074,-0.1279680133,-0.1522427499,0.1946623325,-0.2576690316,-0.0294073075,0.0990639925,0.1730167419,-0.0362704583,-0.3446158767,-0.2268319726,-0.0606659874,0.1769757569,-0.0442230627,0.0362305641,0.0795409605,0.3674478531,0.1300021261,-0.0545415357,0.0300061796,0.3562529385,0.4141275287,-0.0000773683,-0.0395767093,-0.0318559967,-0.5719261765,0.1694633961,0.0508159325,-0.1062118411,-0.1133591086,-0.1411291808,-0.2546457648,-0.0779709145,-0.0450924113,-0.0623477362,0.187547788,0.120737493,0.2329923958,0.0785598606,-0.1249947622,0.236975044,-0.4429727793,0.1645688415,-0.3280670345,0.2244873196,-0.07279329,0.0777952895,0.3571224511,-0.1016227752,0.1318474859,-0.1490375549,-0.1367939562,0.4920094311,-0.0342858844,0.1593009084,0.2432354987,0.3110990226,0.0756986514,-0.0365001708,0.0567063652,-0.1203427464,-0.0641232803,-0.0075017926,-0.2024268359,0.0732878074,-0.1416478753,-0.1127191931,0.1372610927,-0.0706155151,0.32207039,-0.0068846461,-0.0229112059,-0.1962287426,-0.1181997061,-0.0463209338,0.2286891192,-0.0793172568,-0.1190646142,0.1233826205,0.4538817704,-0.0048209582,-0.0790026858,0.2463441491,0.0947225094,-0.2789717317,-0.0093782414,0.2432694584,0.2355402261,0.198435232,-0.007321835,0.2384935319,0.113589026,-0.1258766502,0.1700114459,-0.2082782686,0.070431374,0.3977612853,0.12451832,0.0042015184,-0.312666744,0.0504111312,0.1602742821,0.5041103363,-0.284337163,-0.117480807,-0.195208475,-0.0846701041,-0.0156606622,-0.3108170629,-0.445040822,-0.2003310025,0.0834115967,0.4364151657,0.0369068272,0.1526349187,0.0604125746,0.1537462473,0.1232911795,-0.0739083141,-0.1852822453,0.0508616269,-0.0365219712,0.0668715388,0.3578797877,-0.0556179546,0.2884843349,-0.5230861902,-0.1156506389,-0.5066557527,-0.4701025486,0.1771076322,-0.2434888929,0.2063983232,0.2852979302,0.3453367651,0.0853942484,0.016644815,0.1876368374,-0.2909503877,-0.3853655457,-0.2362888753,0.0202493388,-0.0987058729,-0.1884863377,-0.2773067057,0.0663908198,-0.2346473485,0.2027070373,-0.0451794825,0.0381708592,0.3008177876,0.1180897057,0.142667681,-0.2176997364,0.1401198208,-0.3408536017,-0.8032144904,0.1165456623,-0.2175061256,-0.2753388584,0.0334260054,0.3145299554,0.2690524459,-0.0862895474,-0.4916515052,-0.4815646112,-0.0502631143,0.1115959063,-0.0520309508,-0.085846439,0.3131735325,-0.115236111,-0.1618236154,-0.2248121649,-0.3696803153,0.0858747885,-0.0710309297,0.2880928814,-0.0507137291,0.2898249328,0.0350003876,0.4975779951,0.0497702137,-0.3406611383,0.2349116206,0.0870380402,0.502835989,-0.1201901585,-0.1735134721,-0.0290171467,-0.0749817863,0.1830485314,0.2019392401,0.0666027218,0.2343546599,-0.1795960814,-0.3761564791,-0.4432550371,-0.1809521914,-0.1519084871,0.2326864004,0.2582484782,-0.0118890582,0.0826849937,-0.1184165999,-0.0247938968,0.2890644968,0.5808851123,0.1143822521,0.11650455,-0.3093804717,0.1264024377,-0.1134716421,0.4254815876,-0.1058291942,0.2956213653,-0.0707588792,0.1061825231,-0.0734329298,-0.2445176691,0.3870094717,-0.1323428899,0.2310820967,-0.0370431393,-0.1570010036,0.0471887849,-0.0819774643,-0.0590758994,0.2221333086,0.2458956242,0.4593394995,-0.4197362661,-0.3258239627,0.3473775387,0.0274276137,-0.0546618663,-0.1036282331,-0.2066548467,-0.2233431935,-0.2983759642,-0.0423343927,-0.1828248054,0.0414614268,-0.1609468311,0.0987631083,0.0985471606,0.0093365069,0.0625055656,0.241264537,0.2207420468,0.1506770253,0.1920169592,0.2081608623,0.2933064997,0.6157563925,0.3259396553,-0.0844914764,-0.1526333243,-0.2200955003,-0.0935882702,0.076916717,0.2272888422,-0.0382387452,0.1942356527,0.3077540398,-0.1109454259,-0.0344398357,0.0151219387,0.456066221,0.0224397294,-0.2457534075,-0.2374703437,0.3054959774,0.1069982797,-0.0142818475,0.1488596648,0.1249996349,-0.0835075304,-0.2086605281,0.0756630823,0.7571489215,0.1154305637,-0.2564593852,0.163364172,-0.2155642956,0.2870562375,-0.2390077859,-0.0435625613,-0.3719157577,-0.1366297752,0.0173346344,0.0705045983,0.2069431543,-0.2751024961,-0.3866590559,0.1767390221,-0.1542913914,0.1276845485,-0.1500226259,0.406085521,-0.1204334497,-0.0204885546,-0.4955421984,0.2375459224,-0.005207398,0.1984514296,-0.2777522802,-0.2224818468,0.1204710752,-0.4220785201,-0.1898532659,0.0046363389,0.097353816,-0.0552396514,0.2513027489,0.1956269592,0.0837510824,0.0283696223,0.3956626654,0.2205123007,-0.33170259,0.1149644703,-0.2772340477,-0.1430600733,-0.0945336148,-0.0059699849,0.2583813369,-0.2865524888,-0.3032086194,0.0063771829,-0.1189883798,-0.3908973336,-0.0207682811,-0.2016387731,-0.3541719317,-0.2221975774,-0.0388777256,0.0033385654,-0.1450394392,-0.3388101757,0.2209027559,0.0553912409,0.0103460867,-0.0457546078,0.2296443135,-0.1507496834,-0.0755859762,0.6048678756,-0.2730655074,-0.045895312,0.4467056096,0.0884537026,-0.1323518455,-0.3556785583,-0.2287998796,-0.0147367055,-0.2570452094,0.1229241937,0.0557165332,0.1709830314,-0.0286028348,0.2884240746,0.2503726184,-0.0683768541,0.2611654997,-0.5942689776,-0.3970952928,0.2931919396,-0.2172250301,0.1339838803,-0.1335011572,0.1060830131,0.0215663798,0.0022467822,-0.4168638289,0.324131757,-0.1417961717,0.152562812,0.2349677384,-0.0148071414,0.2760246992,-0.1452235729,0.1075126901,-0.0486321338,-0.4070465565,-0.2047161758,-0.0923671871,0.1079568714,-0.0082236426,0.0363587849,0.1135029793,-0.0469815172,-0.2890306115,0.0681724846,-0.0405898653,-0.0162142776,0.1320535988,0.2156442851,0.0355268344,0.3202003539,0.1303132027,-0.0181020182,0.2259680331,0.1914212406,-0.1189341918,0.0881474167,-0.2905039787,-0.1255598962,0.1655315012,0.205060482,0.1934411675,-0.1190127805,0.2172875702,-0.2848485112,0.0876177996,0.2013476342,0.3840884566,0.5029956698,-0.1881061047,0.0761310384,0.1080144495,0.2819471955,-0.1898917109,0.0813264772,0.283546418,-0.0044893022,-0.1652697474,-0.0396988764,0.3772835732,-0.1312256157,0.2380292416,0.3620763719,0.303373307,-0.1219569668,0.0727457106,0.1685713232,-0.2489296347,-0.0549641661,0.2338484228,0.0671141222,0.0973972157,0.2295092493,-0.1188579649,0.0857031941,-0.1140335202,0.3391358852,-0.0715231895,-0.2610103488,0.1278283149,0.4546133578,0.2511268258,-0.0693245456,0.1099777147,0.5222744346,0.0190881975,-0.0816483572,-0.5117464066,0.1824726313,-0.140557304,-0.1973858178,-0.2351136655,-0.3342036307,-0.0125320777,0.052414611,-0.0644476786,0.0344243646,0.1969603747,-0.0435720533,-0.3683120012,-0.5924250484,-0.1569314003,0.2273051739,0.0858577862,-0.1218269989,0.2826162577,0.2781518996,-0.1386149228,0.3562749624,0.6598345637,0.4140587747,0.1946381778,0.1449317634,-0.1226504892,-0.353761524,-0.0055740084,0.0349135771,0.4112181067,0.0008520027,-0.2888383567,0.294898212,0.200537011,-0.1928182691,0.0974546671,-0.1595926583,0.0687388033,-0.1245113313,0.0247482397,0.043544963,0.083880268,-0.295611769,0.0837557539,-0.3328899145,-0.1367941797,0.1506269723,0.2361779213,0.1390326619,0.0087535176,0.1366646141,-0.0975156203,0.5990440249,0.4314471185,-0.1183937714,0.0068102991,-0.2862880528,-0.6351528168,0.0400249623,0.0421272628,0.1158431396,-0.0390567407,-0.0471782349,0.093139939,0.1612215936,0.2838636339,-0.2534851134,-0.0568662137,-0.2978146672,-0.3684127033,0.2670494616,-0.0604398176,-0.0621896647,0.1521381289,-0.137767598,0.1221760288,-0.088809438,0.0525331013,-0.0819211155,0.3129155338,-0.2794638276,0.0623921044,0.484151423,0.2182160616,0.3679857254,-0.2015001774,-0.3083273768,-0.3716007769,-0.180575639,-0.2003195137,0.4573623836,0.1340679675,0.2414015979,-0.0015641308,-0.0949336663,-0.2568140328,0.536655724,-0.0518552363,-0.2091262043,-0.4673114121,0.0755395591,0.0677415133,-0.0396087728,-0.0304834805,0.0331689417,-0.008009322,0.0795805678,-0.5288831592,-0.1199071035,0.6047830582,-0.1174354404,-0.0393455885,-0.0511536002,0.2046602964,0.2164425105,-0.346935302,-0.7498266101,0.3317582011,0.2261270881,0.0666717067,-0.2390806377,0.3029561341,-0.1761597544,0.1988654584,-0.0772834867,0.2687012255,0.0805802941,-0.3111481667,-0.0445392206,0.0183114484]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2036","title":"Cannot load wikitext","comments":"Solved!","body":"when I execute these codes\r\n```\r\n>>> from datasets import load_dataset\r\n>>> test_dataset = load_dataset(\"wikitext\")\r\n```\r\n\r\nI got an error,any help?\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/xxx\/anaconda3\/envs\/transformer\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/home\/xxx\/anaconda3\/envs\/transformer\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/xxx\/anaconda3\/envs\/transformer\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/xxx\/anaconda3\/envs\/transformer\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/wikitext\/wikitext.py\r\n```","comment_length":1,"text":"Cannot load wikitext\nwhen I execute these codes\r\n```\r\n>>> from datasets import load_dataset\r\n>>> test_dataset = load_dataset(\"wikitext\")\r\n```\r\n\r\nI got an error,any help?\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/home\/xxx\/anaconda3\/envs\/transformer\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/home\/xxx\/anaconda3\/envs\/transformer\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/xxx\/anaconda3\/envs\/transformer\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/xxx\/anaconda3\/envs\/transformer\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/wikitext\/wikitext.py\r\n```\nSolved!","embeddings":[-0.3367408812,-0.23027578,-0.1397335827,0.1838162839,0.3017690778,0.0841983408,0.1699698269,0.2658137679,0.3607211113,-0.1185667217,0.0495107099,0.1573416442,0.0248359591,0.1399417818,0.0457672179,-0.1526674032,-0.015670212,0.2236160338,-0.2360098362,-0.0181014072,-0.3638026416,0.141430527,-0.402097404,0.3108009696,-0.1711391062,-0.0065301447,0.0113600222,0.1210449934,-0.2165504098,-0.4598167241,0.4323220253,-0.0264297947,0.2413971126,0.3558642268,-0.0001063627,0.1924295723,0.5467564464,0.0190203004,-0.326079011,-0.185828656,-0.4803674519,-0.271456778,0.3118495345,-0.2677704394,-0.083795324,-0.2626655102,-0.0697994009,-0.3780320585,0.3280992508,0.5041875243,0.2590939701,0.2291842103,0.3073313534,0.0117985504,-0.0951367244,-0.2129946947,-0.0582993366,-0.0683291182,0.1809799224,-0.1146186739,0.3007126153,0.0732223839,0.0726456642,0.0358040556,0.3709361553,-0.0175872967,0.1025686041,-0.3876442313,0.173807621,0.1018332541,0.7477242351,-0.1726830453,-0.4016180336,-0.0955811664,-0.1260935217,-0.0426344462,0.3729508519,-0.1333252788,-0.1358018667,0.1096631661,-0.2758546472,-0.2264588475,-0.1740466058,0.2741719186,0.0352079049,0.4328197241,-0.0783929154,0.2166642547,-0.0042749597,-0.1427832544,0.3562891483,-0.2170193195,-0.0867623612,0.4375528991,-0.3770934045,0.1797866523,-0.0918514654,0.187612772,0.0370060094,0.1042982861,-0.1728191823,-0.0493723415,0.0541353039,0.2103219181,0.195305109,0.2901900113,0.2317935377,-0.0808880851,0.0161563903,0.1255065501,-0.1226320118,-0.1719958484,-0.2100158036,-0.3754652739,0.0164335035,0.0000486015,0.2186291963,-0.1498177201,-0.1382944584,-0.0779774636,0.2700715065,-0.0373544842,0.1296626925,0.3706628382,-0.130000338,0.3312480748,0.0580967925,0.2263744473,-0.1135654747,0.0925188512,-0.1698773503,0.101901412,-0.2192988545,0.000064079,0.2578030527,0.0370917469,0.3453154862,-0.1356375068,0.0790968165,-0.1230505705,0.1415797174,-0.1039999798,-0.1067585126,0.2406819016,-0.0150285969,0.3889046907,0.2559725642,-0.0103638358,-0.1401321143,0.0759749264,-0.4048928916,-0.245498091,0.1289956272,0.3090299368,-0.141064316,-0.1207785904,-0.3374784291,-0.2613454759,0.1430168152,-0.2568772435,-0.1723741442,-0.28677845,-0.1426932514,-0.2226628512,0.5136910081,0.3490233123,-0.1525130272,-0.0082457447,-0.2186620831,-0.2087220401,-0.1372521371,0.0252883732,-0.3989379406,0.4467042685,-0.2541057765,-0.0060853525,0.6596385837,-0.5441517234,-0.3231917918,0.087035507,-0.1859038472,0.2002677768,-0.1264960915,-0.1362926066,-0.0398896821,-0.1717664748,0.0296643749,0.4009484649,0.1694048345,0.1856176853,-0.0539686307,-0.3904011846,0.3086673021,0.1999133378,0.091523096,0.0192825384,0.0747812763,0.1881748438,0.1517933905,-0.047866419,0.0307562146,0.1486403644,0.0267457627,0.0333496034,0.0806149468,0.0409075208,-0.3802528977,0.3236694634,0.0872170478,-0.2359613627,-0.091235593,0.0665974319,-0.3228420913,0.0699288398,-0.4147218764,-0.1000008062,0.2135595381,0.1608667374,0.2945221663,0.1725721061,-0.2552895546,0.4353122711,-0.0732849538,0.0456564277,-0.1883710027,0.1033089384,0.0345010534,0.0091629848,0.1265733838,0.1368939579,0.1225206554,-0.1903113574,-0.2385952026,0.2036130279,-0.0268763471,0.2233936042,0.0674183294,-0.1428951025,-0.1128891557,-0.3632498682,0.0548764355,0.1585877538,0.278254509,-0.0904210359,0.1497354805,-0.0851149932,-0.0104083773,0.4202741981,0.106271714,0.0247632302,0.0483728945,-0.0218874197,0.2209818512,0.0027979747,0.1411916763,0.1171364188,0.0784015656,-0.1033449471,0.3514585197,0.0416427776,0.4883224964,0.1475695223,0.1176094115,0.1048737913,-0.2099523246,-0.0333326012,-0.0883695334,0.0734384507,0.3943816423,0.0735889524,-0.2152396142,0.02554412,0.1403293163,0.014844127,0.4122641981,0.1941721439,0.129891187,0.1658740044,-0.1445367336,-0.1530309916,-0.0271744039,-0.2263303995,0.1854083687,0.3627988398,-0.2639921606,0.0977620557,-0.153862536,0.051053647,0.0544983074,-0.2079430819,-0.1450559944,-0.4938233793,-0.1517465115,0.1689139903,0.2029057592,0.0696088001,-0.3753259778,0.0609942302,0.0974113122,-0.2594437897,-0.2470017374,-0.277463913,-0.3010046184,0.0957604125,0.3334614038,0.1730907261,0.3951117992,-0.2816528976,0.156280145,-0.0599535517,0.0560981557,-0.0068099708,-0.1854914725,0.0895661265,-0.0302694738,0.3221381903,0.0576803163,-0.3504441977,0.4370700121,-0.2962601483,-0.009397543,-0.007269342,0.1560153514,-0.0975328237,0.0062044654,-0.4541410506,-0.4098352492,-0.3291491866,0.0016392233,0.036102131,0.1108197346,0.1200042516,0.1529750377,0.1200311631,-0.0299842991,0.0603094697,-0.023477383,-0.425745219,0.3725128472,-0.2753011882,-0.3711420894,0.1486932337,0.2826427817,0.3305844665,0.0525110438,-0.2940741777,-0.1955361217,-0.1167877689,0.157686621,-0.0994003639,0.172832787,0.0686621815,0.0344931297,-0.0371087343,-0.1067855954,0.0581999943,0.0271454602,-0.332901001,0.2175712287,0.2745802402,0.3658528626,-0.3726649582,0.6092254519,0.3624806404,-0.1030539051,0.4046730697,-0.0643302649,0.1745516062,-0.2141288817,-0.4548097253,0.0562273227,-0.0200632866,0.2848120928,0.2508373559,-0.0250795856,-0.0668584406,-0.3675087988,-0.0508808494,-0.4069004953,-0.3060418665,-0.0385405794,0.0336289443,0.0967677832,-0.0540937558,0.0689466819,-0.3290653229,-0.1467446536,0.2126190662,0.3051057458,-0.1372660398,0.0866485089,-0.155722037,-0.3006304502,-0.30952847,0.2337561399,0.161271736,0.535107553,-0.0795465484,0.0882411152,0.0799660832,0.0567050949,0.4538629055,-0.4872840345,0.1689286232,0.2993975282,0.1483651102,-0.4050452709,-0.03134479,-0.0883317515,0.3675279319,-0.0298453774,-0.0133149661,-0.1835073829,-0.1930592805,-0.0591070019,0.2952820361,-0.2519690394,-0.1707223356,-0.2400615215,-0.2633197904,-0.3659632206,-0.3143581152,-0.1253948361,0.4171612263,-0.1468813568,0.0368241742,-0.3325974941,0.0398796424,0.0206533168,0.1445650458,0.3441622555,0.0192307569,0.1283019632,0.2033283859,0.0064952248,-0.0599399395,0.5627630353,0.0370029323,-0.3410872221,-0.0817575604,0.014553356,0.1808905005,-0.0782798305,-0.2105623037,-0.0622146018,-0.2981804311,0.1267915517,0.0575538725,0.2871847153,0.3981765211,0.1195451021,-0.3885909319,-0.4799150527,0.4277274609,-0.0155380964,-0.0575481132,0.5144018531,0.2076826543,-0.3488051593,0.0160921235,-0.0927931815,0.6467411518,-0.0607166588,-0.0195946489,0.1357352436,0.17479375,0.4947400689,-0.2175121307,0.0731936619,-0.2801511586,-0.2447715253,0.0255902763,-0.160170719,0.1678454727,0.0030607719,-0.2363362163,0.3183510602,-0.2471477389,0.1624567211,0.0999430791,0.2366993874,-0.0872670487,-0.1529429257,-0.278795898,0.2439016402,-0.0849104598,0.4001356065,-0.139532432,-0.0355102941,-0.0610119067,-0.2897410393,-0.174872905,0.2348323762,-0.2458771467,0.4941704869,-0.2618135214,-0.0447351821,0.1758972853,0.4582683146,0.4964675903,0.1549499184,-0.3839634359,0.1219532713,-0.3176619411,-0.2536861598,-0.1877799779,0.1043673456,-0.0141956694,-0.1089091226,-0.4621698558,0.4079231322,-0.1470319331,-0.1597764194,0.1399773806,-0.1155702397,0.2841660082,-0.2072222084,-0.3300986588,-0.0865567923,-0.0166786984,0.0698550642,0.1618818641,0.070602797,-0.0909537598,0.3006505072,0.0590264983,-0.208151862,0.1606001258,0.3588599861,-0.0931361541,0.0238848701,0.7217999101,0.0842141435,-0.2271583676,-0.2806792855,-0.0704248101,0.1475030929,-0.4006390572,-0.2509387136,0.3251067698,0.2036440223,-0.0664731935,0.3641472459,0.0173844006,-0.0068067815,0.0435954556,-0.6172276735,-0.1666513085,0.1075021401,0.0098204147,0.1712378711,-0.0052707288,-0.1753909588,0.1337720752,-0.1426579058,-0.3141853809,0.0090166358,-0.146493271,-0.1258580834,0.1733521521,-0.1465707421,0.3231536746,-0.034043923,0.1565434188,-0.0747523829,-0.2774266303,-0.2652937472,0.0413937941,0.1448760927,0.1649057269,-0.2221177071,-0.1244264022,-0.2396118343,0.0619802475,-0.1662486345,0.1843664348,-0.0777705833,-0.0285879876,-0.0598304793,0.2764567733,0.0281217694,-0.2981856465,0.3120129406,-0.1651208103,0.0991862416,0.0318672955,0.0654607117,0.176933527,-0.0152839907,0.0045796046,0.0126122776,0.1089585125,0.1015548185,0.3485559821,-0.3015218377,0.2593790293,0.0857469216,-0.0059951884,0.6035738587,-0.282238096,-0.0064482461,0.2595029175,0.2438310534,-0.1389178038,-0.3047925532,0.2390183359,0.0106010623,-0.1022512838,0.0626971945,-0.0483242609,-0.456987381,0.1497095376,-0.0287382044,0.523899138,0.0957116336,0.1216852292,0.3886884153,-0.0292506851,-0.0615916587,-0.0987187326,0.2006284893,0.0473872237,0.3048479855,-0.5333997011,0.2371261567,-0.1644490957,0.1488182396,-0.180395782,-0.4060031772,0.2638046741,-0.1137740612,-0.0254012775,0.0828988105,-0.1021073759,0.0458543077,-0.1450718343,0.0639972314,-0.40864411,0.2758887112,0.0353820287,-0.1110078841,-0.0562315881,-0.1424094439,-0.0945022181,-0.209281534,0.0081842802,-0.3941867948,0.088887997,0.1453356892,-0.2184594423,-0.2324629724,0.0482148863,0.218068108,0.0652110949,-0.1214957237,0.2653007507,0.1984923482,0.0445116945,0.0704572871,0.2671015561,0.3147932887,0.4626870453,-0.2137852907,0.2466282696,0.0822898597,-0.1122324243,0.0099818204,0.2098151594,-0.0742007643,0.1539242715,0.3578349948,0.2583840489,-0.1714670658,-0.1894368678,-0.1168342084,0.1170322075,-0.0529666394,0.1372185796,0.1079160571,0.1070985943,-0.3633163869,0.2782748342,-0.4677443504,0.0069143786,0.2590411603,0.0154725527,0.0605288967,-0.1565762758,0.0852581114,-0.1049754173,0.72368294,0.3289071321,0.1922010034,-0.23585549,-0.2940261364,-0.6231970191,0.1477846354,-0.1110865325,-0.3319758773,0.2593551278,-0.2436511666,-0.1674314588,0.2448843271,0.3246091306,0.3264445662,-0.0203061718,0.0900230333,-0.4020781815,-0.276635766,-0.0152785843,0.1853681654,-0.0739966407,-0.3661749363,0.2818694115,-0.3410940468,0.1256805062,-0.0752091929,0.1614650339,0.0216534436,-0.5074633956,0.5169349909,0.0693245307,0.5676766634,0.0929307193,0.0756079033,-0.3626318276,-0.2353037894,-0.0988221839,0.277038157,0.0398469642,0.0157252066,-0.2022733092,-0.0822439268,-0.241965279,0.1866291314,0.0683882907,0.1709020585,0.0251314808,-0.2090106755,0.089557372,0.1154510975,0.2673041821,-0.0106326723,-0.1139105856,0.1889781356,-0.1675655544,-0.2815136909,0.5208865404,-0.3782648146,-0.3060480654,-0.0052373093,0.1844539344,-0.2839393318,-0.0749108493,-0.3890549541,0.1937181503,0.4181625843,0.0318211094,-0.1538008749,0.2372706234,-0.0613766909,0.0363209955,-0.0788982362,0.1766899079,-0.2009954304,0.0292678513,-0.1853794456,-0.0984205902]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"Dear @lhoestq for wikipedia dataset I also get the same error, I greatly appreciate if you could have a look into this dataset as well. Below please find the command to reproduce the error:\r\n\r\n```\r\ndataset = load_dataset(\"wikipedia\", \"20200501.bg\")\r\nprint(dataset)\r\n```\r\n\r\nYour library is my only chance to be able training the models at scale and I am grateful for your help.\r\n\r\n","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":62,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nDear @lhoestq for wikipedia dataset I also get the same error, I greatly appreciate if you could have a look into this dataset as well. Below please find the command to reproduce the error:\r\n\r\n```\r\ndataset = load_dataset(\"wikipedia\", \"20200501.bg\")\r\nprint(dataset)\r\n```\r\n\r\nYour library is my only chance to be able training the models at scale and I am grateful for your help.\r\n\r\n","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"Hi @dorost1234,\r\nTry installing this library first, `pip install 'apache-beam[gcp]' --use-feature=2020-resolver` followed by loading dataset like this using beam runner.\r\n\r\n`dataset = load_dataset(\"wiki40b\", \"cs\", beam_runner='DirectRunner')`\r\n\r\n I also read in error stack trace that:\r\n\r\n> Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc.\r\n\r\nWorked perfectly fine after this (Ignore these warnings)\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/19718818\/110908410-c7e2ce00-8334-11eb-8d10-7354359e9ec3.png)\r\n\r\n","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":83,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nHi @dorost1234,\r\nTry installing this library first, `pip install 'apache-beam[gcp]' --use-feature=2020-resolver` followed by loading dataset like this using beam runner.\r\n\r\n`dataset = load_dataset(\"wiki40b\", \"cs\", beam_runner='DirectRunner')`\r\n\r\n I also read in error stack trace that:\r\n\r\n> Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc.\r\n\r\nWorked perfectly fine after this (Ignore these warnings)\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/19718818\/110908410-c7e2ce00-8334-11eb-8d10-7354359e9ec3.png)\r\n\r\n","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"For wikipedia dataset, looks like the files it's looking for are no longer available. For `bg`, I checked [here](https:\/\/dumps.wikimedia.org\/bgwiki\/). For this I think `dataset_infos.json` for this dataset has to made again? You'll have to load this dataset also using beam runner.\r\n\r\n","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":41,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nFor wikipedia dataset, looks like the files it's looking for are no longer available. For `bg`, I checked [here](https:\/\/dumps.wikimedia.org\/bgwiki\/). For this I think `dataset_infos.json` for this dataset has to made again? You'll have to load this dataset also using beam runner.\r\n\r\n","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"Hello @dorost1234,\r\n\r\nIndeed, Wikipedia datasets need a lot of preprocessing and this is done using Apache Beam. That is the reason why it is required that you install Apache Beam in order to preform this preprocessing.\r\n\r\nFor some specific default parameters (English Wikipedia), Hugging Face has already preprocessed the dataset for you (and it is stored in the cloud). That is the reason why you do not get the error for English: the preprocessing is already done by HF and you just get the preprocessed dataset; Apache Beam is not required in that case.","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":94,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nHello @dorost1234,\r\n\r\nIndeed, Wikipedia datasets need a lot of preprocessing and this is done using Apache Beam. That is the reason why it is required that you install Apache Beam in order to preform this preprocessing.\r\n\r\nFor some specific default parameters (English Wikipedia), Hugging Face has already preprocessed the dataset for you (and it is stored in the cloud). That is the reason why you do not get the error for English: the preprocessing is already done by HF and you just get the preprocessed dataset; Apache Beam is not required in that case.","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"Hi\nI really appreciate if huggingface can kindly provide preprocessed\ndatasets, processing these datasets require sufficiently large resources\nand I do not have unfortunately access to, and perhaps many others too.\nthanks\n\nOn Fri, Mar 12, 2021 at 9:04 AM Albert Villanova del Moral <\n***@***.***> wrote:\n\n> Hello @dorost1234 ,\n>\n> Indeed, Wikipedia datasets need a lot of preprocessing and this is done\n> using Apache Beam. That is the reason why it is required that you install\n> Apache Beam in order to preform this preprocessing.\n>\n> For some specific default parameters (English Wikipedia), Hugging Face has\n> already preprocessed the dataset for you (and it is stored in the cloud).\n> That is the reason why you do not get the error for English: the\n> preprocessing is already done by HF and you just get the preprocessed\n> dataset; Apache Beam is not required in that case.\n>\n> \u2014\n> You are receiving this because you were mentioned.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":185,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nHi\nI really appreciate if huggingface can kindly provide preprocessed\ndatasets, processing these datasets require sufficiently large resources\nand I do not have unfortunately access to, and perhaps many others too.\nthanks\n\nOn Fri, Mar 12, 2021 at 9:04 AM Albert Villanova del Moral <\n***@***.***> wrote:\n\n> Hello @dorost1234 ,\n>\n> Indeed, Wikipedia datasets need a lot of preprocessing and this is done\n> using Apache Beam. That is the reason why it is required that you install\n> Apache Beam in order to preform this preprocessing.\n>\n> For some specific default parameters (English Wikipedia), Hugging Face has\n> already preprocessed the dataset for you (and it is stored in the cloud).\n> That is the reason why you do not get the error for English: the\n> preprocessing is already done by HF and you just get the preprocessed\n> dataset; Apache Beam is not required in that case.\n>\n> \u2014\n> You are receiving this because you were mentioned.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"Hi everyone\r\nthanks for the helpful pointers, I did it as @bhavitvyamalik suggested, for me this freezes on this command for several hours, \r\n\r\n`Downloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/users\/dara\/cache\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\n`\r\n\r\nDo you know how long this takes? Any specific requirements the machine should have? like very large memory or so? @lhoestq \r\n\r\nthanks \r\n\r\n\r\n","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":65,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nHi everyone\r\nthanks for the helpful pointers, I did it as @bhavitvyamalik suggested, for me this freezes on this command for several hours, \r\n\r\n`Downloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/users\/dara\/cache\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\n`\r\n\r\nDo you know how long this takes? Any specific requirements the machine should have? like very large memory or so? @lhoestq \r\n\r\nthanks \r\n\r\n\r\n","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"HI @dorost1234, \r\nThe dataset size is 631.84 MiB so depending on your internet speed it'll take some time. You can monitor your internet speed meanwhile to see if it's downloading the dataset or not (use `nload` if you're using linux\/mac to monitor the same). In my case it took around 3-4 mins. Since they haven't used `download_and_extract` here that's why there's no download progress bar.","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":65,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nHI @dorost1234, \r\nThe dataset size is 631.84 MiB so depending on your internet speed it'll take some time. You can monitor your internet speed meanwhile to see if it's downloading the dataset or not (use `nload` if you're using linux\/mac to monitor the same). In my case it took around 3-4 mins. Since they haven't used `download_and_extract` here that's why there's no download progress bar.","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"Hi\r\nthanks, my internet speed should be good, but this really freezes for me, this is how I try to get this dataset:\r\n\r\n`from datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\", beam_runner='DirectRunner')`\r\n\r\nthe output I see if different also from what you see after writing this command:\r\n\r\n`Downloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/users\/dara\/cache\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...`\r\n\r\ndo you have any idea why it might get freezed? anything I am missing @lhoestq @bhavitvyamalik. Do I need maybe to set anything special for apache-beam? \r\n\r\nthanks a lot \r\n\r\nOn Tue, Mar 16, 2021 at 9:03 AM Bhavitvya Malik ***@***.***>\r\nwrote:\r\n\r\n> HI @dorost1234 ,\r\n> The dataset size is 631.84 MiB so depending on your internet speed it'll\r\n> take some time. You can monitor your internet speed meanwhile to see if\r\n> it's downloading the dataset or not (use nload if you're using linux\/mac\r\n> to monitor the same). In my case it took around 3-4 mins. Since they\r\n> haven't used download_and_extract here that's why there's no download\r\n> progress bar.\r\n>\r\n> \u2014\r\n> You are receiving this because you were mentioned.\r\n> Reply to this email directly, view it on GitHub\r\n> ,\r\n> or unsubscribe\r\n> \r\n> .\r\n>\r\n","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":212,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nHi\r\nthanks, my internet speed should be good, but this really freezes for me, this is how I try to get this dataset:\r\n\r\n`from datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\", beam_runner='DirectRunner')`\r\n\r\nthe output I see if different also from what you see after writing this command:\r\n\r\n`Downloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/users\/dara\/cache\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...`\r\n\r\ndo you have any idea why it might get freezed? anything I am missing @lhoestq @bhavitvyamalik. Do I need maybe to set anything special for apache-beam? \r\n\r\nthanks a lot \r\n\r\nOn Tue, Mar 16, 2021 at 9:03 AM Bhavitvya Malik ***@***.***>\r\nwrote:\r\n\r\n> HI @dorost1234 ,\r\n> The dataset size is 631.84 MiB so depending on your internet speed it'll\r\n> take some time. You can monitor your internet speed meanwhile to see if\r\n> it's downloading the dataset or not (use nload if you're using linux\/mac\r\n> to monitor the same). In my case it took around 3-4 mins. Since they\r\n> haven't used download_and_extract here that's why there's no download\r\n> progress bar.\r\n>\r\n> \u2014\r\n> You are receiving this because you were mentioned.\r\n> Reply to this email directly, view it on GitHub\r\n> ,\r\n> or unsubscribe\r\n> \r\n> .\r\n>\r\n","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"I tried this on another machine (followed the same procedure I've mentioned above). This is what it shows (during the freeze period) for me:\r\n```\r\n>>> dataset = load_dataset(\"wiki40b\", \"cs\", beam_runner='DirectRunner')\r\nDownloading: 5.26kB [00:00, 1.23MB\/s] \r\nDownloading: 1.40kB [00:00, 327kB\/s] \r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bhavitvya\/.cache\/huggingface\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nWARNING:apache_beam.internal.gcp.auth:Unable to find default credentials to use: The Application Default Credentials are not available. They are available if running in Google Compute Engine. Otherwise, the environment variable GOOGLE_APPLICATION_CREDENTIALS must be defined pointing to a file defining the credentials. See https:\/\/developers.google.com\/accounts\/docs\/application-default-credentials for more information.\r\nConnecting anonymously.\r\nWARNING:apache_beam.io.tfrecordio:Couldn't find python-snappy so the implementation of _TFRecordUtil._masked_crc32c is not as fast as it could be.\r\n```\r\nAfter around 10 minutes, here's the loading of dataset:\r\n```\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:16<00:00, 16.42s\/sources]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.12sources\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.14sources\/s]\r\nDataset wiki40b downloaded and prepared to \/home\/bhavitvya\/.cache\/huggingface\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f. Subsequent calls will reuse this data.\r\n```","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":156,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nI tried this on another machine (followed the same procedure I've mentioned above). This is what it shows (during the freeze period) for me:\r\n```\r\n>>> dataset = load_dataset(\"wiki40b\", \"cs\", beam_runner='DirectRunner')\r\nDownloading: 5.26kB [00:00, 1.23MB\/s] \r\nDownloading: 1.40kB [00:00, 327kB\/s] \r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bhavitvya\/.cache\/huggingface\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nWARNING:apache_beam.internal.gcp.auth:Unable to find default credentials to use: The Application Default Credentials are not available. They are available if running in Google Compute Engine. Otherwise, the environment variable GOOGLE_APPLICATION_CREDENTIALS must be defined pointing to a file defining the credentials. See https:\/\/developers.google.com\/accounts\/docs\/application-default-credentials for more information.\r\nConnecting anonymously.\r\nWARNING:apache_beam.io.tfrecordio:Couldn't find python-snappy so the implementation of _TFRecordUtil._masked_crc32c is not as fast as it could be.\r\n```\r\nAfter around 10 minutes, here's the loading of dataset:\r\n```\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:16<00:00, 16.42s\/sources]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.12sources\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.14sources\/s]\r\nDataset wiki40b downloaded and prepared to \/home\/bhavitvya\/.cache\/huggingface\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f. Subsequent calls will reuse this data.\r\n```","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2035","title":"wiki40b\/wikipedia for almost all languages cannot be downloaded","comments":"Hi\r\nI honestly also now tried on another machine and nothing shows up after\r\nhours of waiting. Are you sure you have not set any specific setting? maybe\r\ngoogle cloud which seems it is used here, needs some credential setting?\r\nthanks for any suggestions on this\r\n\r\nOn Tue, Mar 16, 2021 at 10:02 AM Bhavitvya Malik ***@***.***>\r\nwrote:\r\n\r\n> I tried this on another machine (followed the same procedure I've\r\n> mentioned above). This is what it shows (during the freeze period) for me:\r\n>\r\n> >>> dataset = load_dataset(\"wiki40b\", \"cs\", beam_runner='DirectRunner')\r\n> Downloading: 5.26kB [00:00, 1.23MB\/s]\r\n> Downloading: 1.40kB [00:00, 327kB\/s]\r\n> Downloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bhavitvya\/.cache\/huggingface\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\n> WARNING:apache_beam.internal.gcp.auth:Unable to find default credentials to use: The Application Default Credentials are not available. They are available if running in Google Compute Engine. Otherwise, the environment variable GOOGLE_APPLICATION_CREDENTIALS must be defined pointing to a file defining the credentials. See https:\/\/developers.google.com\/accounts\/docs\/application-default-credentials for more information.\r\n> Connecting anonymously.\r\n> WARNING:apache_beam.io.tfrecordio:Couldn't find python-snappy so the implementation of _TFRecordUtil._masked_crc32c is not as fast as it could be.\r\n>\r\n> After around 10 minutes, here's the loading of dataset:\r\n>\r\n> 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:16<00:00, 16.42s\/sources]\r\n> 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.12sources\/s]\r\n> 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.14sources\/s]\r\n> Dataset wiki40b downloaded and prepared to \/home\/bhavitvya\/.cache\/huggingface\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f. Subsequent calls will reuse this data.\r\n>\r\n> \u2014\r\n> You are receiving this because you were mentioned.\r\n> Reply to this email directly, view it on GitHub\r\n> ,\r\n> or unsubscribe\r\n> \r\n> .\r\n>\r\n","body":"Hi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```","comment_length":259,"text":"wiki40b\/wikipedia for almost all languages cannot be downloaded\nHi\r\nI am trying to download the data as below:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"wiki40b\", \"cs\")\r\nprint(dataset)\r\n```\r\n\r\nI am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.\r\n\r\nI really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources. \r\n\r\nthank you very much.\r\n\r\n```\r\n(fast) dara@vgne046:\/user\/dara\/dev\/codes\/seq2seq$ python test_data.py\r\nDownloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp\/dara\/cache_home_2\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\nTraceback (most recent call last):\r\n File \"test_data.py\", line 3, in \r\n dataset = load_dataset(\"wiki40b\", \"cs\")\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 579, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1105, in _download_and_prepare\r\n import apache_beam as beam\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/__init__.py\", line 96, in \r\n from apache_beam import io\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/__init__.py\", line 23, in \r\n from apache_beam.io.avroio import *\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/apache_beam-2.28.0-py3.7-linux-x86_64.egg\/apache_beam\/io\/avroio.py\", line 55, in \r\n import avro\r\n File \"\", line 983, in _find_and_load\r\n File \"\", line 967, in _find_and_load_unlocked\r\n File \"\", line 668, in _load_unlocked\r\n File \"\", line 638, in _load_backward_compatible\r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 34, in \r\n File \"\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/__init__.py\", line 30, in LoadResource\r\nNotADirectoryError: [Errno 20] Not a directory: '\/user\/dara\/libs\/anaconda3\/envs\/fast\/lib\/python3.7\/site-packages\/avro_python3-1.9.2.1-py3.7.egg\/avro\/VERSION.txt'\r\n```\nHi\r\nI honestly also now tried on another machine and nothing shows up after\r\nhours of waiting. Are you sure you have not set any specific setting? maybe\r\ngoogle cloud which seems it is used here, needs some credential setting?\r\nthanks for any suggestions on this\r\n\r\nOn Tue, Mar 16, 2021 at 10:02 AM Bhavitvya Malik ***@***.***>\r\nwrote:\r\n\r\n> I tried this on another machine (followed the same procedure I've\r\n> mentioned above). This is what it shows (during the freeze period) for me:\r\n>\r\n> >>> dataset = load_dataset(\"wiki40b\", \"cs\", beam_runner='DirectRunner')\r\n> Downloading: 5.26kB [00:00, 1.23MB\/s]\r\n> Downloading: 1.40kB [00:00, 327kB\/s]\r\n> Downloading and preparing dataset wiki40b\/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bhavitvya\/.cache\/huggingface\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...\r\n> WARNING:apache_beam.internal.gcp.auth:Unable to find default credentials to use: The Application Default Credentials are not available. They are available if running in Google Compute Engine. Otherwise, the environment variable GOOGLE_APPLICATION_CREDENTIALS must be defined pointing to a file defining the credentials. See https:\/\/developers.google.com\/accounts\/docs\/application-default-credentials for more information.\r\n> Connecting anonymously.\r\n> WARNING:apache_beam.io.tfrecordio:Couldn't find python-snappy so the implementation of _TFRecordUtil._masked_crc32c is not as fast as it could be.\r\n>\r\n> After around 10 minutes, here's the loading of dataset:\r\n>\r\n> 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:16<00:00, 16.42s\/sources]\r\n> 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.12sources\/s]\r\n> 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.14sources\/s]\r\n> Dataset wiki40b downloaded and prepared to \/home\/bhavitvya\/.cache\/huggingface\/datasets\/wiki40b\/cs\/1.1.0\/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f. Subsequent calls will reuse this data.\r\n>\r\n> \u2014\r\n> You are receiving this because you were mentioned.\r\n> Reply to this email directly, view it on GitHub\r\n> ,\r\n> or unsubscribe\r\n> \r\n> .\r\n>\r\n","embeddings":[-0.2544980347,-0.0779204071,-0.1537582129,0.4308663607,0.3997030556,0.3504618704,0.1368506402,0.5331582427,0.1893276423,0.0235845186,-0.1773036569,-0.0942173079,0.0944036767,0.0111841084,-0.0297050718,-0.4591732025,-0.0672903061,0.0295370072,-0.1352714151,-0.1168936566,-0.2519630194,0.1256146282,-0.0791749284,-0.0490285978,-0.0387454666,-0.1828399897,-0.095835425,-0.2073058486,-0.3507466614,-0.240864411,0.3545782268,0.0192575995,0.3101119697,0.152997002,-0.0001114368,-0.0509645268,0.313116461,-0.2357945591,-0.2936119437,-0.2888573706,-0.0720896423,-0.3261856437,-0.0904391855,-0.420031637,0.1274462193,0.0466828533,0.080671154,-0.1318009049,0.0757578686,0.299225539,0.2468712181,-0.1426034421,0.313434124,-0.1457377374,0.3651962578,-0.3777080178,-0.0280337762,0.3395861685,0.2187208384,0.0713426769,0.2154791504,0.2220428586,-0.2341407239,0.0134992879,0.0482247174,-0.291223824,0.222205475,-0.6297016144,0.5206420422,0.5344284177,0.8290324211,-0.1283516139,-0.0874669701,0.0723921135,-0.0260593966,-0.031374447,-0.0557402261,0.4819797575,-0.1893264949,0.0975982994,0.1261651665,-0.1254302114,-0.1567192525,0.4228464365,-0.1241363883,0.4435888529,-0.0527197532,0.2112598121,-0.0216665938,-0.102701664,-0.2636416852,-0.0696592405,-0.0961601064,0.2933215201,-0.2793983221,0.0196358114,-0.1615019292,0.0556470677,0.3511354923,-0.1503831744,-0.2346212864,-0.0450777449,0.0960106328,0.1575745195,0.1801504046,0.1637832075,0.0620483644,-0.200006485,0.2782812417,0.3079036772,-0.0842119008,-0.0761709884,-0.2100430429,-0.2230561525,-0.2947570086,0.0944294706,0.0875293911,-0.2031713277,-0.1779466271,0.0071135005,-0.1443609297,-0.0194389597,-0.202394858,0.3628205657,-0.1501425803,0.4312385917,0.1779012084,0.3288222253,-0.1035727784,-0.2613099515,-0.0173523482,0.3065485656,-0.1393195242,-0.1036603302,0.2187829465,-0.1099155918,0.2648515999,-0.089371942,-0.1031381935,-0.1764115244,-0.1536911428,-0.3695675731,-0.1791034937,0.230408892,0.1619764566,0.2839277089,0.1476140618,-0.244747147,-0.2305807471,0.1207451969,-0.3596149981,-0.1681548953,-0.2051547766,0.1944684684,0.0610982925,-0.0546504743,-0.3160067499,0.2876641154,0.3030638993,-0.1423988342,0.1180728897,-0.0075459145,-0.0232539773,-0.225190118,0.1790115982,0.4908005595,-0.399959296,0.091897428,-0.0238504577,-0.0337120444,0.1897819787,0.1842971891,-0.1779527366,0.2565417886,-0.0978842974,0.0381600037,0.2101310194,-0.3214419484,-0.5776352286,0.186053887,0.113410078,-0.1670030951,0.0994797125,0.0160005782,0.2750306129,-0.0493875705,0.052454181,0.3956748247,0.1368381381,-0.1056447476,-0.3274506032,-0.5078234673,0.3264816701,0.1633179337,0.4547148943,-0.1322700828,0.2759852409,0.2917218208,0.3191742897,-0.1915477812,0.1246789992,0.2333606035,-0.0089483308,-0.0445753597,-0.002562576,-0.3310278654,-0.3097580373,0.1477600932,0.1029136032,0.1654645205,-0.1609553099,0.029683331,-0.3161896169,-0.0888073742,-0.1953119785,0.10326875,0.1995639354,-0.1052765027,0.1195931062,0.4854590893,0.1265164316,-0.0994593799,-0.4156903923,-0.0948526934,-0.17893444,0.0923319757,-0.2174285501,0.2044921666,0.0336807854,0.0258089639,0.1796839386,0.0730525404,-0.1607236117,-0.0110613015,-0.0784521401,-0.1035443768,0.1354628503,0.116963841,0.1331282407,-0.417375356,0.1920297742,0.1084797084,0.3297463953,-0.1357724667,0.1178161129,0.1222370416,0.0399733894,0.3655338883,0.1234308109,0.1498314887,0.4555997252,0.0305262543,0.1263544261,-0.0049747,0.3271189332,0.1556606591,0.0127488235,-0.1329593062,-0.02927793,-0.1135142446,0.3166727722,-0.0845338032,0.2671483159,0.2927592099,-0.0765190423,-0.004859813,-0.0978270099,0.1533981711,0.0522076748,0.0699049756,0.1065892354,-0.0998440012,0.359649688,-0.1293927282,0.2643093467,0.1755426973,0.1863330156,0.209135592,0.1350800842,-0.0111756697,-0.1882829219,-0.1296795756,0.0699935928,0.3717179298,0.0691837072,0.165945366,-0.0484345779,-0.1418599933,0.0841909274,-0.0523921847,-0.2509426475,-0.2163173407,-0.1256383955,-0.0857288986,0.091583252,-0.0001049751,-0.3266668916,-0.0721021518,0.0621662624,-0.3203505874,-0.1605682075,-0.2981233299,-0.3684496582,0.0767874196,0.3871294558,0.1112902835,0.2952094376,-0.3326290548,-0.1918789148,-0.0899458826,-0.2681111097,-0.0143140424,-0.1569705009,0.2413951159,-0.0241171494,0.7152194381,-0.1799962372,-0.251421392,0.1846487224,-0.0070200525,0.0658744872,0.2497714758,-0.1534786075,-0.0429239906,0.0845356733,-0.4856536686,-0.5155764222,-0.2696566582,0.0125194602,0.048517894,0.0878061801,0.1686406732,0.1924842894,0.0680263713,0.121187754,0.1809071153,-0.1900433302,-0.1518690884,0.4496890306,-0.0274131726,-0.4299626648,0.1881868094,0.0109940656,0.3149316907,0.0079274345,-0.6572358608,0.1441125572,-0.2763589025,0.2499436736,-0.0219936892,0.183476612,0.114343822,-0.2136970013,0.0479722433,0.0449956916,0.1059101671,-0.0406209044,-0.2095253319,0.2556217909,0.0299457368,0.2961854935,0.1644209623,0.6888408661,0.0548558161,0.1658407599,0.4367816746,0.2385823429,0.2310316861,-0.2314289063,-0.3645718992,0.0605744347,-0.2394915372,-0.0310032498,0.2670128345,-0.1623201221,-0.4674171805,-0.3477128148,-0.2426850945,-0.3674669564,-0.2107714713,0.1539980173,-0.0501817837,0.0829403698,0.0575965643,0.1203985438,0.0215738118,-0.2823496759,0.2137899101,0.5317094326,-0.2486018538,0.1685896665,-0.2782668173,-0.3421222866,-0.4126341641,0.2524550557,0.0777070075,0.4147613049,-0.0938022807,0.3499184847,0.2011797726,-0.0892689303,0.290623039,-0.1702442914,0.1230505109,0.1972466558,0.1049934402,-0.2152172327,-0.0624398403,-0.2943330109,-0.0166878775,0.305416882,-0.044455111,-0.6709661484,0.0843290389,-0.0425857045,0.5003297925,-0.0571766347,-0.0173498485,-0.1618567407,-0.5653654337,-0.4854040146,-0.0276873074,0.1895754188,0.4046091437,-0.1972095817,0.2628991306,-0.1020783857,0.2205626667,0.0213163923,0.0077347509,0.2132563591,0.2327862382,0.0974328592,-0.1967884898,-0.1378474534,0.1503265202,0.3620435894,-0.0635325536,-0.2343129963,-0.1212477908,-0.0430041067,0.0434832536,0.1415467858,-0.1012905389,0.047707092,0.1486631334,-0.1379353851,0.1658210456,-0.0128418375,0.2519462705,-0.1063374579,-0.4962349236,-0.5676657557,0.6620466709,0.0233823843,0.0701422319,0.254171133,0.1754049212,-0.195861131,0.3753736615,0.1058031619,1.1372184753,-0.2281340957,0.1104206368,-0.1039845571,0.1094662994,0.3614169061,-0.3578394651,0.2084074765,-0.3226415515,-0.1089071035,-0.0195194241,0.0736745819,-0.0461390689,0.0998932719,-0.3072267473,0.2564415932,-0.3662326932,-0.086954087,0.1056496054,0.2329857051,-0.4058896601,-0.0680367425,-0.3213793039,0.1196285784,-0.2146603614,0.2546931207,-0.2110011131,-0.1254787296,-0.0614245199,-0.2672702968,-0.505050838,0.0706436113,-0.2883504033,0.3298689425,-0.4893512428,-0.1897748858,0.3386204541,0.4448707998,0.0414795317,0.403154701,-0.2879836857,0.1116522104,-0.2638610601,-0.4236196876,0.0944909304,0.0112072332,0.0871337727,-0.1816289127,-0.370990485,0.1669728905,-0.0575728603,0.0528493039,-0.095987767,0.1043777987,0.2016545683,-0.0458974019,-0.3565553129,-0.0222827923,-0.0943860784,-0.0894329175,0.1241667569,0.0821967423,-0.1119039133,0.1409135014,0.1852949709,-0.3349809647,0.0217915084,0.3491058648,0.183449164,0.0239540208,0.6245266795,0.3155454993,-0.3965409398,-0.1898151934,0.0439142361,-0.2713772655,-0.4585742354,-0.0896052718,-0.0421900302,0.3383434117,-0.4149283469,0.3502537608,0.1749068648,-0.2348314971,0.20646438,-0.7065883875,-0.124021627,0.2160943598,-0.0789091438,0.0065746442,0.0840711519,-0.3095994294,0.1218283251,0.2352631092,-0.2610890269,0.2092312276,-0.3233409524,0.1007633656,0.0994939059,0.0633322895,0.0223014448,0.199102357,0.0650481731,0.0691613108,-0.0496639535,-0.2599102557,-0.070227094,0.1948125809,0.2051140368,-0.0579005219,0.2278748602,-0.5071967244,-0.168670103,0.0822387636,-0.058475513,0.1155845076,0.0038598548,0.1811006963,0.2013250291,0.2885712981,-0.3500797451,0.2458777577,-0.0869051144,0.2614985704,0.1408281624,0.0759360194,0.4579918683,0.0213007908,-0.3352098763,-0.0335408524,-0.0300773904,0.0866357163,0.2558811009,-0.2333586812,0.30034706,0.1989205778,-0.1174951568,0.5623862147,-0.1496048272,0.2065964341,0.4130835235,0.1872234792,-0.3698635697,0.0016462884,0.2932014763,-0.0966379121,-0.0709824041,0.0759611428,0.2589094341,-0.3223283887,0.1137206927,-0.0657388866,0.1647827327,-0.0565516204,0.1922584623,0.670076251,0.2270726562,0.0499260277,-0.0694646165,0.1699932665,0.0953947902,0.3628066778,-0.2234565765,0.1927634627,-0.0520039126,0.024634745,-0.1621611118,-0.3400563598,0.298611939,0.1632015258,-0.0247749183,0.0986764655,-0.1722574979,0.2616477609,0.1543592066,0.0302232467,-0.2611553967,0.1914432198,-0.1224777922,-0.0129824765,-0.044210311,-0.268007338,0.0015667842,0.1556260139,-0.0046621701,-0.3565731645,0.2555734813,0.3037431538,-0.2440855801,-0.4411394298,0.1348094344,0.2124924064,0.031170642,-0.300221324,0.1577611119,0.4137805402,0.0780922696,-0.1270555258,0.2019162774,0.443346113,0.4205470085,-0.3373264372,0.0424413644,0.0753402933,-0.0656807274,-0.0233784299,0.2276101559,0.4672250748,0.2502681911,0.3970101476,0.1184260622,-0.1415998489,-0.0319754519,0.2360505462,-0.0639189929,-0.0566706173,-0.0451859049,-0.059035439,0.1407208294,-0.4205531776,-0.0135848057,-0.5078696609,0.3400221467,0.1610560417,0.0025873203,0.1775293648,-0.0925812796,0.047985848,-0.1569759697,0.3962932229,0.3195855916,0.1991413683,-0.3908326924,-0.0089951809,-0.4645188153,-0.0186358485,-0.2982345819,0.1121009961,-0.0814326257,0.1286558956,0.0755084306,0.2132720202,0.0366321988,0.0450984463,0.0668031126,0.4949170053,-0.2315317541,-0.2447035462,-0.1948164254,0.2191630602,-0.1583237201,-0.3909311891,0.0575938039,-0.1385606527,0.0340732373,-0.3200198114,-0.0665002689,0.095572859,0.0059600249,0.332428515,0.1523778141,0.5925299525,-0.1540948153,-0.2021768838,-0.208671689,-0.0110697085,-0.1599347144,0.3116776645,0.1967292577,0.0989065692,-0.1607936919,-0.1119716093,-0.0295638815,0.2984875739,-0.0448162481,0.1949228942,-0.3669620156,-0.3863566816,0.0579175428,0.1138557866,0.1190269515,0.0469262749,-0.1308560967,-0.0533182286,-0.3421497941,-0.3086865544,0.386654228,-0.3773033321,-0.5697819591,-0.0921551287,0.0079290252,0.0470561683,0.1902980208,-0.3660884202,0.2775549293,0.38977319,0.0513331741,-0.1509680152,0.1664218903,0.0286483783,-0.0162155982,-0.0940432772,0.1160700768,-0.0082844039,-0.3563147783,-0.2205843776,-0.391258508]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2031","title":"wikipedia.py generator that extracts XML doesn't release memory","comments":"Hi @miyamonz \r\nThanks for investigating this issue, good job !\r\nIt would be awesome to integrate your fix in the library, could you open a pull request ?","body":"I tried downloading Japanese wikipedia, but it always failed because of out of memory maybe.\r\n\r\nI found that the generator function that extracts XML data in wikipedia.py doesn't release memory in the loop.\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L464-L502\r\n\r\n`root.clear()` intend to clear memory, but it doesn't.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L490\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L494\r\nI replaced them with `elem.clear()`, then it seems to work correctly.\r\n\r\nhere is the notebook to reproduce it.\r\nhttps:\/\/gist.github.com\/miyamonz\/dc06117302b6e85fa51cbf46dde6bb51#file-xtract_content-ipynb","comment_length":28,"text":"wikipedia.py generator that extracts XML doesn't release memory\nI tried downloading Japanese wikipedia, but it always failed because of out of memory maybe.\r\n\r\nI found that the generator function that extracts XML data in wikipedia.py doesn't release memory in the loop.\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L464-L502\r\n\r\n`root.clear()` intend to clear memory, but it doesn't.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L490\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L494\r\nI replaced them with `elem.clear()`, then it seems to work correctly.\r\n\r\nhere is the notebook to reproduce it.\r\nhttps:\/\/gist.github.com\/miyamonz\/dc06117302b6e85fa51cbf46dde6bb51#file-xtract_content-ipynb\nHi @miyamonz \r\nThanks for investigating this issue, good job !\r\nIt would be awesome to integrate your fix in the library, could you open a pull request ?","embeddings":[0.218652904,-0.1109624952,-0.0523667894,0.6246852279,0.3160352409,0.0942655355,-0.2117367834,0.3332336545,0.2093814313,0.2579506338,0.0060537024,0.1517625153,0.2327899486,-0.1265527904,-0.2175300568,-0.2103440166,0.0437293872,0.0743473619,0.0371215492,-0.2225494236,-0.2259665877,0.1241694018,-0.2827935219,-0.12147104,-0.1757008135,0.1507509947,0.0834618583,-0.1925340891,-0.1768444926,-0.3868060112,0.142742753,0.0021264469,-0.1252807826,0.2270927429,-0.0001113045,-0.1786878556,0.5870485306,-0.0175461024,-0.1438791752,0.1073924974,-0.2037125081,-0.0739126429,-0.1006892473,-0.2878377438,0.1415049881,-0.0300183408,0.1694545895,-0.2743773758,0.3588359654,-0.0555356257,0.2296685129,-0.0995117724,0.4134443104,-0.0468071289,0.7186948657,0.2662892342,-0.0335875936,0.029499257,0.1483324766,-0.1202112958,0.0261039324,0.404242605,-0.1569174826,0.0165875256,0.2526498139,-0.1493367255,0.0439883471,-0.5066655278,0.462076813,0.3427121341,0.5755248666,-0.149921447,0.1062823087,0.0310698599,-0.103971675,-0.1315815896,0.2247127742,0.4001595676,-0.6533283591,-0.0683356151,0.310906291,-0.4820676148,-0.1790922135,0.2063481659,-0.0007484592,0.3183763027,-0.0073742974,0.0634122789,0.2590226829,-0.0018349971,-0.1167000756,-0.0434845239,-0.0742314756,0.2187654525,0.0305326488,0.1633389443,0.1844055355,-0.1822960973,0.2779972553,-0.4200806022,-0.4094339311,0.0066289743,0.2780645192,-0.0407775752,0.3894085586,0.0270341001,-0.0170564633,0.1133421734,0.3600495756,0.2115766406,-0.349632591,-0.0846013352,0.130389154,0.0520469174,0.0831234306,-0.1065111905,0.0191854648,-0.0676592961,0.1701021045,0.2386854142,-0.2658776343,-0.0708776414,-0.2080523819,0.4120154381,-0.0438275561,0.372759521,0.1869971752,0.0853901058,-0.4100779891,-0.311619401,-0.1378388852,0.2522204816,-0.2821925879,0.0472872071,0.0010046506,-0.1443132162,0.2116989046,0.1731064916,-0.1932032406,-0.4035503566,0.0540767089,-0.1906531304,0.2073850334,0.2648632824,0.0696727708,0.2217064202,-0.1199448183,-0.0872631669,-0.0428388752,0.2829099,-0.0179263931,-0.0642531589,0.000170208,0.2416815013,0.0659806058,0.1953769028,0.0082763946,-0.1624775529,0.5027115345,0.1204016134,0.0157041252,0.0061148582,-0.0120037021,-0.2563004196,0.1782667786,0.4500468075,0.115413323,0.1249703392,-0.1744931787,0.133442983,0.4826317132,0.1633789986,0.0041837408,0.4243835807,-0.1227871999,0.2591704428,-0.0239813644,-0.2639952004,-0.3197956979,-0.0183573831,0.1209944561,0.2170036286,-0.0394850187,0.027487237,0.1240423471,0.3383145034,0.0136304786,0.2789423168,0.0746846274,0.1292003542,-0.4539269507,-0.227173686,0.2193467021,-0.3085442483,0.3155327439,-0.0395956486,0.1276439577,0.4033349752,0.6220804453,-0.2563405931,0.3680494428,0.1581970751,0.3149207532,0.0550059788,0.0325528122,-0.1054977551,-0.121082373,0.0815544575,-0.2072693706,0.3825250268,-0.0645792186,-0.1309960932,-0.1074616387,0.0873796344,-0.2546545267,-0.3456029594,0.2049910873,0.1016071886,0.3502486348,0.2771036327,0.131781891,-0.4485380948,-0.1533707082,0.1642220616,-0.7328389287,0.0174813066,-0.164858073,-0.085667558,-0.1884648651,0.0739084408,-0.074687399,0.0146213006,-0.1174387038,-0.1440369487,0.2105429918,0.1954229027,-0.2769904435,-0.0807574838,0.2259412706,-0.4512328207,0.1940934062,0.4587644041,0.1753469408,-0.0962863714,0.1605661511,-0.1155879945,-0.0212012045,0.045373708,0.1233118102,0.0445288047,0.2474220991,0.1230115145,0.159300819,-0.1362757236,0.1121829897,0.5882779956,-0.0906394497,0.0201329291,-0.1109917685,-0.052671887,0.448946476,0.2313284576,0.1596060991,0.0479361117,-0.3074405491,-0.3058307171,0.0191244688,0.0920182243,-0.1586149633,-0.0036709667,0.1324584484,-0.0507996678,0.3378245831,-0.2958723903,0.4203760922,-0.0014915981,0.1611881405,0.2169839293,0.1296709478,-0.0276828781,-0.3015906215,0.2943036258,-0.0072424943,0.3758419454,0.1096902117,-0.2038994581,0.0588484295,-0.4482650757,0.0137910014,-0.1235477105,-0.2024699301,-0.3298225403,0.1109552011,0.2990667522,-0.2745769322,-0.1786494851,0.0398271568,-0.2857567668,0.0882399678,-0.0641841069,-0.1608580053,-0.3678180277,-0.1747340411,0.0267358031,0.2860378325,-0.0619838387,0.0603972748,0.0625112429,-0.2241583019,-0.2190746814,-0.1240182966,0.2660339773,-0.0900258049,0.0177640691,-0.2295581549,0.4114502966,0.1463361233,-0.1430453509,0.2513380051,-0.3257459104,0.015834108,0.2487401962,0.1499990225,0.1744736135,-0.1233965084,-0.4725928009,0.095373489,-0.4844042063,0.216541484,-0.071549505,0.1014145464,0.6841795444,0.0560019203,-0.0651491731,-0.0548735559,0.1761159748,-0.3081732094,-0.1929737628,0.2051042914,-0.1673519909,-0.3377017677,0.0364821218,-0.0153450109,0.3164588213,0.3823059201,-0.2814798951,0.281981945,-0.1632010192,0.3446781039,0.0712391883,0.2762082517,0.1284642816,-0.0342609622,-0.0964972302,0.0976973027,0.4960945249,-0.0526958145,-0.0706965923,0.0776841268,0.1368727386,0.0494218767,0.0495294519,0.4761870503,0.1436742991,0.5873883367,0.3695218265,0.117123574,0.0994822308,-0.0463386811,-0.1920556873,0.007617882,-0.2097374946,-0.2204042077,0.1437850595,0.0391851664,-0.1361601353,-0.1121597886,-0.0174565967,-0.1355139166,-0.4964639544,0.2188590765,0.2334727198,0.3340344131,-0.0010710737,0.2686026394,0.1284434348,-0.3472023606,0.0751789808,-0.2073795348,0.2089843154,-0.0089348117,0.0095372312,-0.3298080564,-0.5612035394,0.0703181848,0.0946088806,-0.1601820141,0.136947006,-0.2673763037,0.1915389746,-0.0156702902,0.2761349082,0.1296467632,0.0107779047,0.0797568858,-0.0402338095,-0.0932375714,0.2106632739,-0.1364921331,-0.0208932664,0.2161007971,-0.0451990813,-0.4547621906,-0.1504540592,0.0004929095,0.3498814106,-0.171781227,0.0372958481,-0.1772454977,-0.3322714269,-0.5068716407,-0.1046344042,0.1379766315,0.4159510136,0.0982623249,-0.134845227,0.28210181,0.0894633383,-0.0784473643,0.0764087364,0.302516222,-0.1999039948,-0.0024878238,-0.0331882052,-0.2916375697,0.0972167999,0.1391355693,0.0937781557,0.0013570432,-0.3511095047,-0.1426324248,-0.1238651052,0.0209074747,-0.0997716859,0.2029341459,-0.0969599113,0.100651212,-0.2624906003,-0.1832824796,0.1236005723,-0.1975307018,0.1065525711,-0.2093930542,0.3150336444,-0.0883085057,0.097024329,0.5454167724,0.0013303269,-0.3294945657,-0.0218490399,0.2135730833,0.8503885269,-0.1525959373,0.2035499811,0.3837127388,0.2043781728,0.4438613951,-0.0951671973,0.4557697475,-0.446080476,0.1965766698,0.1880086809,0.2044650018,-0.1236271486,-0.3261246085,-0.0760761201,0.0926753357,-0.1996020824,0.0572188571,-0.0746131837,0.338676542,-0.0680009201,-0.1012846455,-0.1137883961,0.0897194892,-0.0381051637,0.3747819662,-0.1996064931,0.1092801765,0.1518605947,-0.0086168507,-0.3571421504,0.1347111166,-0.1278684884,0.3138391376,-0.1692618728,-0.2985732257,-0.1522773057,0.2091228366,-0.3088812232,0.2668367922,-0.1183214113,0.009040026,-0.1343429685,-0.3013686836,-0.0981239825,0.2041148394,0.2514209449,-0.0793510526,-0.3915471435,0.3663363159,-0.1248372346,-0.2694856226,-0.3880557716,0.006317548,-0.2515437901,-0.0010343521,-0.019907536,-0.038866166,-0.4225254357,-0.0467490181,0.1766571254,0.2087343782,-0.3466226459,0.1273144484,-0.2620408535,-0.1558553576,-0.0108480118,-0.0530654788,0.4631014168,0.1669984162,0.4985492826,0.3981515467,-0.1369816214,-0.2070303261,-0.2525363564,-0.2442411929,-0.2741562724,0.3001496792,-0.0792824477,-0.0194021054,-0.2108329237,0.1423033029,-0.1996143013,-0.2362213284,0.0427225418,-0.1615916044,-0.29339993,0.2349670678,-0.0004801908,0.0744672269,-0.0216564313,0.0198189281,0.1276602149,0.131080538,-0.2810684144,-0.0354295634,-0.0050444715,0.1445888281,0.1383127123,-0.1445304453,-0.062524043,-0.029495297,0.0763277858,0.1032244787,-0.1214114204,-0.2354750633,0.1276704818,0.0888685882,0.0758989975,-0.1377899796,0.1375190467,-0.258872807,0.0512371548,-0.243494913,0.14515917,0.0246921964,-0.1223565638,0.2717655897,0.0594721697,0.0915278047,-0.1313969195,0.2522022724,0.0426177271,0.3345507085,0.0080394121,0.1259252131,0.3993858993,-0.1260975152,-0.7414156199,-0.0434415117,-0.123516053,-0.120308049,0.3450509012,-0.0021813931,-0.2966969311,0.3581522107,0.1455202401,0.363809824,0.0317699797,0.0791696981,0.0330610424,0.2661410868,-0.1345178038,0.1233594343,0.0110244174,-0.1922677606,-0.1088291407,0.0627266616,-0.0308701023,-0.2388404161,0.0254228245,0.1598429233,-0.0791817307,-0.1824392527,0.4027221203,0.051598195,0.1644679755,-0.0325859711,0.0224655438,0.1362171471,0.1064551175,0.273283422,0.1820954978,0.1362058371,0.09505786,0.0025314845,-0.1988527924,0.0505050756,-0.116030775,0.4192367494,-0.3653098941,0.2405392677,-0.1250740439,0.0292006154,0.2645374835,0.0168230254,-0.28561005,0.2863125801,-0.2693559825,-0.1376813203,0.0908491835,-0.3488929868,0.0829970837,0.1361267567,0.2529473603,-0.1999788135,0.2733370662,0.0997959003,-0.4058679044,-0.3179627955,0.6245741248,-0.0224793721,0.2209910154,-0.4029295146,-0.1296217442,0.155476436,-0.1812809855,-0.4392130971,0.4111103714,0.2919528186,0.4301742017,-0.5130152702,0.1023003608,-0.1698184907,-0.0790034086,-0.2631183565,0.241145879,0.1912723631,0.1533513218,-0.0848816633,0.0882202908,-0.2010697275,-0.3969244957,0.5125550628,0.2918366492,-0.0400139168,-0.2914663255,0.0826321617,-0.2327853143,-0.1004539728,-0.1515384614,-0.4054411948,0.5398058295,0.1902051419,0.0122688739,-0.0175531767,-0.2326767594,0.0941384211,-0.1395607144,0.203352347,0.1744119525,0.0770211443,-0.3007369637,-0.2969961166,-0.4083119333,0.4256489873,-0.3481346667,-0.1179603785,0.0067750546,0.4143896997,-0.2831629813,-0.0244602021,-0.1137403548,-0.405359447,0.0567634739,0.2254750729,-0.1330938339,0.0131690474,0.0438831188,-0.290220052,-0.0594618991,-0.4011066556,0.3153992593,-0.3039936423,0.032792937,-0.0892022401,0.201095432,0.244383961,0.1934106201,0.3121204376,0.1702417731,-0.1130365729,-0.0869575441,-0.5135855079,-0.2686930001,-0.0153141972,0.0984290615,0.1275863051,0.0767477006,0.3675940037,-0.2123593837,-0.0594635159,-0.5560925007,0.4299428761,-0.2201288342,0.0037100876,-0.2014356405,-0.1379737407,-0.0401602648,0.0747390687,0.1364296675,0.366969347,-0.2426819801,0.0443307422,-0.5643947124,-0.457070142,0.0585050918,-0.4836917222,-0.5538260937,-0.12026342,0.250752449,-0.0282402523,-0.1778517365,-0.3676704168,0.2605850101,0.0732462853,0.0482600443,-0.2900910378,-0.1871774942,-0.0593820922,-0.0868129581,0.102200821,0.2677107155,0.3074436188,-0.332114011,0.2800374925,-0.1016466469]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2031","title":"wikipedia.py generator that extracts XML doesn't release memory","comments":"OK! I'll send it later.","body":"I tried downloading Japanese wikipedia, but it always failed because of out of memory maybe.\r\n\r\nI found that the generator function that extracts XML data in wikipedia.py doesn't release memory in the loop.\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L464-L502\r\n\r\n`root.clear()` intend to clear memory, but it doesn't.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L490\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L494\r\nI replaced them with `elem.clear()`, then it seems to work correctly.\r\n\r\nhere is the notebook to reproduce it.\r\nhttps:\/\/gist.github.com\/miyamonz\/dc06117302b6e85fa51cbf46dde6bb51#file-xtract_content-ipynb","comment_length":5,"text":"wikipedia.py generator that extracts XML doesn't release memory\nI tried downloading Japanese wikipedia, but it always failed because of out of memory maybe.\r\n\r\nI found that the generator function that extracts XML data in wikipedia.py doesn't release memory in the loop.\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L464-L502\r\n\r\n`root.clear()` intend to clear memory, but it doesn't.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L490\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L494\r\nI replaced them with `elem.clear()`, then it seems to work correctly.\r\n\r\nhere is the notebook to reproduce it.\r\nhttps:\/\/gist.github.com\/miyamonz\/dc06117302b6e85fa51cbf46dde6bb51#file-xtract_content-ipynb\nOK! I'll send it later.","embeddings":[0.2158349752,-0.1369753778,-0.0751149952,0.6292429566,0.2823996842,0.1216343343,-0.2138463706,0.3074065745,0.2163672596,0.2581599951,-0.0161915012,0.0686329976,0.2661634982,-0.2086939067,-0.2403720021,-0.2032506019,0.0633780062,0.0676672161,0.1054105014,-0.2325051725,-0.2528131306,0.0822209418,-0.2931561768,-0.0598032773,-0.1600023061,0.1352131516,0.0478544347,-0.1636054367,-0.1926824003,-0.3672473133,0.1330125481,0.004866844,-0.1329094172,0.2526717782,-0.0001079737,-0.1155835763,0.558042407,0.0044671753,-0.1495727599,0.1526376158,-0.2392689139,-0.0780476183,-0.1389864832,-0.3345431685,0.1944391429,-0.0353079438,0.2427640259,-0.2529093027,0.4104289412,0.0464484096,0.2503244579,-0.1724791229,0.3579753339,-0.0701295063,0.6777843237,0.197313875,0.0013392722,0.0032173288,0.1296724975,-0.0917803571,0.055004552,0.3496797681,-0.2074803859,0.0967564657,0.1529202312,-0.1346193999,0.1378275454,-0.5233784318,0.5264941454,0.3576197922,0.5429794192,-0.0519957095,0.0990832895,0.0860021859,-0.1492464989,-0.0942551792,0.1740495861,0.4373601377,-0.6394232512,-0.0306665394,0.3317213953,-0.4371909201,-0.2004711777,0.1992294341,-0.0843696222,0.3451048434,-0.0120809972,0.0693189278,0.1655682921,-0.0122835152,-0.0838595331,-0.0946310982,-0.0836681426,0.230920136,0.0973629802,0.2114774436,0.1666530371,-0.1935658455,0.2298356891,-0.4608115852,-0.3901153803,0.0311449133,0.3433533311,-0.0331876352,0.2848020196,-0.0499070212,-0.0216571055,0.1940096468,0.351249963,0.2011832893,-0.3329522014,-0.0461509228,0.0436184891,0.0462528616,0.0721842796,-0.1648235023,0.0370919481,-0.0621378943,0.1311497688,0.2598977983,-0.2726944983,-0.0475046039,-0.2579089403,0.4383779168,-0.051184129,0.3087646961,0.1662275195,0.0794224218,-0.3880963326,-0.2915152311,-0.1592338085,0.2298775166,-0.2451315224,0.0275536701,-0.0132058663,-0.0584564283,0.2217471153,0.1041256711,-0.2486633509,-0.3810890317,0.0879303887,-0.2145116031,0.2050263584,0.2636179626,0.1114806011,0.2321072221,-0.1076993495,-0.1284864545,0.0203145873,0.2721976638,0.0117595028,-0.0108311465,0.0871680081,0.2803074718,0.0620645732,0.1927654445,0.0599317551,-0.152308926,0.4960599244,0.1264657378,-0.0007867201,0.0198805854,-0.0392347462,-0.2574646175,0.1899013072,0.4254548848,0.1054156944,0.1113458201,-0.1624340564,0.1126982048,0.466568023,0.180078432,0.0110831447,0.4139524996,-0.1271940619,0.2413067222,-0.0269646198,-0.2288848311,-0.3176385462,0.0308336392,0.0893415809,0.2379782945,0.0304078087,-0.0030947332,0.1411515325,0.3175809085,0.0635481104,0.2578467131,0.0774677098,0.0595604852,-0.4397712946,-0.1846203655,0.1921966225,-0.2772417665,0.3908475935,-0.0579296947,0.1497924328,0.3963705301,0.5963266492,-0.2192007899,0.3427790701,0.1747478992,0.228233397,0.0448015891,0.0111986184,-0.0643938035,-0.0805764347,0.136311844,-0.213589862,0.3308815956,-0.0734973326,-0.1048049554,-0.1490776837,0.0996099189,-0.265637368,-0.3756019175,0.2406111807,0.1607122123,0.3659499586,0.2443017364,0.1553455442,-0.4350113273,-0.1897163093,0.1262608469,-0.6109926105,-0.0041946881,-0.2433442771,-0.0467459261,-0.2092917114,-0.0052879057,-0.0782327801,0.0493751019,-0.0701051354,-0.125676915,0.2239345312,0.191682443,-0.234987855,-0.1165395826,0.184896335,-0.4659763277,0.1765286177,0.478664875,0.1942842305,-0.0871648043,0.1069535464,-0.1432007104,0.0645554364,0.0693601966,0.1518578082,0.0292566232,0.2952629328,0.1344446689,0.153275758,-0.0965467468,0.130162254,0.6366350651,-0.1526687741,0.0027818405,-0.1186397672,-0.0443787575,0.4339186251,0.1683220118,0.1931665093,0.013177719,-0.3161838055,-0.2908439636,-0.0058203358,0.0668733269,-0.1865340471,0.0073924926,0.1654313356,-0.0470482223,0.3062999249,-0.3109093308,0.3907746971,0.0170615036,0.1543491632,0.1979542673,0.0678742453,-0.0515913218,-0.2662218511,0.2834482491,0.0346538797,0.3559324145,0.101037994,-0.1817295402,0.0735233799,-0.4145115018,0.0346898437,-0.1115361154,-0.1560193896,-0.3400929272,0.0767323375,0.2744958103,-0.2601095438,-0.1900546998,-0.0217119586,-0.2775219381,0.1211473718,-0.0322341993,-0.1213632375,-0.3608002663,-0.2089332789,0.0806400478,0.2885161638,-0.045444686,0.0577427149,0.0018981649,-0.1829972118,-0.1901790649,-0.0785688758,0.2843331695,-0.0465374514,0.0035794335,-0.2878242433,0.4657643437,0.1142904535,-0.0525132194,0.2262684852,-0.2775018215,0.0330453739,0.3065686226,0.1920882612,0.0833772272,-0.0734756961,-0.4759866595,0.0439973995,-0.4765582681,0.1499234438,-0.0130581614,0.1203970835,0.6290409565,0.1225100085,0.0280424356,-0.0009593377,0.2082349211,-0.2697204649,-0.1406001598,0.2571451962,-0.1571314186,-0.3456115425,0.067509301,-0.0186463278,0.2488391399,0.3809982836,-0.255764842,0.3032865524,-0.1820967346,0.3210250437,0.1324587911,0.2160001099,0.1375291198,-0.0466521978,-0.1358308345,0.1157023758,0.5716738701,-0.0500157028,-0.1181557328,0.0288743768,0.087034896,0.0202060025,0.0623630732,0.4602249861,0.100918971,0.6075725555,0.3145149052,0.1060747653,0.0643850416,-0.0199760552,-0.1505504102,0.0007113472,-0.1930157393,-0.2625409663,0.1800676137,0.0223676804,-0.1738435179,-0.1260260344,0.0365756564,-0.1565298736,-0.4426422417,0.2121716887,0.2325849235,0.3388128877,0.0160799809,0.2886724174,0.138164416,-0.3062340319,0.0141240098,-0.2014666647,0.1335475594,-0.0021777507,-0.024326941,-0.3123292923,-0.5426830053,0.1340603679,0.0861776248,-0.1384168565,0.087488398,-0.260001868,0.1435364932,-0.108705163,0.2154995054,0.0847231448,0.0749068111,0.1337309033,-0.0091558592,-0.12551938,0.1765097231,-0.2332523763,-0.008931837,0.19248043,-0.0352047235,-0.4376105368,-0.1106676608,-0.0431596227,0.3304582238,-0.1965502799,0.0979879498,-0.2300691754,-0.3356262743,-0.5042068362,-0.1140069589,0.1279083937,0.378146261,0.0595377386,-0.0902252644,0.2353288084,0.1477892995,-0.0917846411,0.1106798947,0.2080611736,-0.1964752376,-0.0978617966,-0.044974234,-0.2665425539,0.0971990824,0.1655766666,0.10603787,-0.0076676351,-0.3158636391,-0.1953428686,-0.1484566629,0.0385889336,-0.0786762834,0.0974871069,-0.0234323628,0.1119647846,-0.2412762642,-0.1401754469,0.1552370489,-0.1904844493,0.0437604971,-0.2737182975,0.2985530794,-0.0892819762,0.0834234878,0.4767294824,0.0033276964,-0.3228062987,-0.016029533,0.1953747869,0.8174183965,-0.1128353253,0.1811814457,0.3971939981,0.1842700243,0.3672097325,-0.1004601941,0.417850256,-0.3242821693,0.200301975,0.1550057381,0.2087133527,-0.0848019123,-0.2472199649,-0.1160029024,0.0980992839,-0.1839704961,0.062194135,-0.0269649774,0.3528735638,-0.0409207158,-0.0470515899,-0.0684714317,0.1340453178,-0.0271025915,0.3319363892,-0.2111084312,0.1507387608,0.0998307914,0.0043853065,-0.3272662759,0.1514105201,-0.1350732744,0.3370357752,-0.2833485901,-0.2497534901,-0.2287993878,0.2023250163,-0.3272545636,0.2612752616,-0.1079231128,0.0178140327,-0.1259153783,-0.3707506061,-0.1022351831,0.1292860061,0.2487343848,-0.0946751609,-0.3780904114,0.4387860894,-0.1209071353,-0.2283458859,-0.366764605,-0.045112893,-0.256031841,0.0486361831,-0.0469619632,-0.0746453702,-0.5152693987,-0.0562241636,0.2016043514,0.1320368052,-0.3306322694,0.0927391499,-0.2136823684,-0.1950319558,-0.0230219383,-0.053170938,0.4060647488,0.1952251345,0.5134326816,0.3931502104,-0.1501265019,-0.2494011521,-0.2607581913,-0.2954850793,-0.1898863465,0.2792170942,-0.1323646009,-0.0095403409,-0.2072062492,0.1139446124,-0.1370292753,-0.2486861497,0.0677508265,-0.1886686087,-0.2662290335,0.2092092484,0.0896669328,0.0652457699,-0.0520619676,-0.0708747953,0.1886032969,0.1192496791,-0.3147353232,-0.041037865,-0.0497989841,0.1462720335,0.117951639,-0.1398849189,-0.0666060746,-0.0129704066,0.1196641698,0.0680592209,-0.0983712226,-0.2813386023,0.1572321653,0.0724923238,0.0907289386,-0.1704663783,0.1622582227,-0.2980371416,0.0856304467,-0.1914435178,0.0966661125,0.0329670198,-0.1560000032,0.2644455433,0.0696851611,0.15476197,-0.1392099261,0.2220221311,-0.0452206917,0.3500640094,-0.0282657091,0.1178739518,0.4682978392,-0.1601721495,-0.698065877,-0.0366021581,-0.0998923182,-0.1383635402,0.3623509407,-0.0534911975,-0.2065788954,0.3107617199,0.1019723192,0.4059424996,0.0048126276,0.0548192337,0.0387870595,0.3177114427,-0.1325269639,0.1394551545,-0.0137358122,-0.2529683709,-0.0703393221,0.0560205579,-0.0867717564,-0.2230414599,-0.0281715356,0.206105426,-0.0840318277,-0.1604317874,0.3487671614,0.0706781745,0.1897751987,0.0072721969,0.0043588225,0.1346706748,0.1098331586,0.2904845178,0.1830527186,0.1304128468,0.1246399507,0.0253186505,-0.1853075176,0.0293230806,-0.0743287057,0.4108400345,-0.3968816996,0.2713361382,-0.1209966764,-0.0507183857,0.2588759661,0.0707534105,-0.3629613817,0.2759001851,-0.2374019176,-0.1480952948,0.0559554957,-0.3931978345,0.065132685,0.1409322172,0.2603569031,-0.2272590399,0.2101437002,0.1188499853,-0.4053248763,-0.2639064193,0.6612070203,-0.0745034888,0.1843363494,-0.4403138459,-0.1484626234,0.1694589406,-0.1650131941,-0.3876648843,0.3738335073,0.2839654088,0.4148187637,-0.5674369931,0.14299348,-0.1419646889,-0.1438913792,-0.3115709424,0.2482184768,0.1655297279,0.1100194603,-0.0197208785,0.1207034439,-0.2558164001,-0.3616748154,0.5092564225,0.2501886487,0.0102694044,-0.310259223,0.0577800572,-0.2230667621,-0.0919040516,-0.1387246251,-0.4623877108,0.4992473125,0.1522053927,-0.0221000202,0.0176000781,-0.2017215192,0.1040232778,-0.1240179092,0.2395274937,0.1670815498,0.1375832707,-0.3113651872,-0.2753790319,-0.3885924518,0.4034686089,-0.4044094384,-0.1867394149,-0.0192143731,0.4237848818,-0.2960512936,0.0302431006,-0.170336619,-0.3335257173,0.0171087161,0.3526278436,-0.1065591648,0.0349970572,0.0943119526,-0.2125261575,-0.0820670724,-0.4304197431,0.3457657993,-0.3272180557,0.0996514335,-0.0460915007,0.1258660555,0.290017128,0.2071767598,0.3289859891,0.1387987137,-0.0778758898,-0.0791931003,-0.4812913537,-0.2450189441,0.0338514224,0.0875172988,0.142426759,0.1022322029,0.3082788289,-0.1951502115,-0.0624132864,-0.4630456865,0.3853928149,-0.2425408214,-0.0196728557,-0.1848293692,-0.1752944887,-0.0381177627,0.0882598311,0.1240867227,0.3478791416,-0.2160550505,-0.0022804816,-0.5521243215,-0.4354024529,0.0125320861,-0.4672850966,-0.546379745,-0.1532113999,0.2249580622,-0.0545903146,-0.101271227,-0.2618284523,0.2778609395,0.1171852872,0.012088744,-0.3050743043,-0.2020770162,-0.0330057591,-0.1170556694,0.0670178309,0.2718422115,0.2311572433,-0.3560227752,0.1737933308,-0.1251563728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2029","title":"Loading a faiss index KeyError","comments":"In your code `dataset2` doesn't contain the \"embeddings\" column, since it is created from the pandas DataFrame with columns \"text\" and \"label\".\r\n\r\nTherefore when you call `dataset2[embeddings_name]`, you get a `KeyError`.\r\n\r\nIf you want the \"embeddings\" column back, you can create `dataset2` with\r\n```python\r\ndataset2 = load_from_disk(dataset_filename)\r\n```\r\nwhere `dataset_filename` is the place where you saved you dataset with the embeddings in the first place.","body":"I've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.\r\n\r\nThe basic steps are:\r\n\r\n1. Create a dataset (dataset1)\r\n2. Create an embeddings column using DPR\r\n3. Add a faiss index to the dataset\r\n4. Save faiss index to a file\r\n5. Create a new dataset (dataset2) with the same text and label information as dataset1\r\n6. Try to load the faiss index from file to dataset2\r\n7. Get `KeyError: \"Column embeddings not in the dataset\"`\r\n\r\nI've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing\r\n\r\nUbuntu Version\r\nVERSION=\"18.04.5 LTS (Bionic Beaver)\"\r\n\r\ndatasets==1.4.1\r\nfaiss==1.5.3\r\nfaiss-gpu==1.7.0\r\ntorch==1.8.0+cu101\r\ntransformers==4.3.3\r\n\r\nNVIDIA-SMI 460.56\r\nDriver Version: 460.32.03\r\nCUDA Version: 11.2 \r\nTesla K80 \r\n\r\nI was basically following the steps here: https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index\r\n\r\nI included the exact code from the documentation at the end of the notebook to show that they don't work either.\r\n","comment_length":65,"text":"Loading a faiss index KeyError\nI've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.\r\n\r\nThe basic steps are:\r\n\r\n1. Create a dataset (dataset1)\r\n2. Create an embeddings column using DPR\r\n3. Add a faiss index to the dataset\r\n4. Save faiss index to a file\r\n5. Create a new dataset (dataset2) with the same text and label information as dataset1\r\n6. Try to load the faiss index from file to dataset2\r\n7. Get `KeyError: \"Column embeddings not in the dataset\"`\r\n\r\nI've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing\r\n\r\nUbuntu Version\r\nVERSION=\"18.04.5 LTS (Bionic Beaver)\"\r\n\r\ndatasets==1.4.1\r\nfaiss==1.5.3\r\nfaiss-gpu==1.7.0\r\ntorch==1.8.0+cu101\r\ntransformers==4.3.3\r\n\r\nNVIDIA-SMI 460.56\r\nDriver Version: 460.32.03\r\nCUDA Version: 11.2 \r\nTesla K80 \r\n\r\nI was basically following the steps here: https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index\r\n\r\nI included the exact code from the documentation at the end of the notebook to show that they don't work either.\r\n\nIn your code `dataset2` doesn't contain the \"embeddings\" column, since it is created from the pandas DataFrame with columns \"text\" and \"label\".\r\n\r\nTherefore when you call `dataset2[embeddings_name]`, you get a `KeyError`.\r\n\r\nIf you want the \"embeddings\" column back, you can create `dataset2` with\r\n```python\r\ndataset2 = load_from_disk(dataset_filename)\r\n```\r\nwhere `dataset_filename` is the place where you saved you dataset with the embeddings in the first place.","embeddings":[0.0550739542,-0.6190827489,0.0702500269,0.3613130152,0.1554000527,0.2710704207,0.3453339934,0.0510576218,0.5391324162,0.2844024301,-0.0778134242,0.1564425826,0.4091768861,-0.0636667162,-0.0600452609,-0.0365858525,0.2299646586,0.2709702551,0.2771186829,-0.1172366887,-0.2132829577,0.1254137605,-0.2415112406,0.239650771,-0.2419934869,0.1532631963,-0.2360691279,0.020749405,-0.1963155121,-0.4184430838,0.321844846,-0.2817219794,0.6558502913,0.1962961704,-0.0001232087,0.0766335726,0.3582345247,-0.1224535927,-0.2154051512,0.0617960207,0.1387543678,0.0691665038,0.2078349441,-0.0417269841,-0.0147014875,-0.6515750289,0.0266365558,-0.1597283334,0.2974438667,-0.0055642687,0.0943733007,0.0098311165,0.0809608251,-0.3012819588,0.0691149384,-0.3261920512,-0.0540437624,0.2427324355,0.0452901497,0.0076503479,0.0668617189,0.29880476,-0.0401020646,0.0219106451,0.4202647209,0.2766120732,0.3658125699,-0.1659394056,0.1298123002,0.1871755868,0.3916547894,-0.1082114428,-0.4308947623,-0.1739238501,0.2771785855,-0.1161144003,0.2110203952,-0.3297830224,-0.0157120544,0.2792980671,0.4298000336,-0.243448332,0.1159293428,0.0830414221,-0.0895649642,-0.0029288968,-0.1924940199,0.0390617587,0.1796053648,-0.1281749308,-0.307997942,-0.1144421324,0.1000975594,0.1209597439,-0.3051662743,0.1096482873,0.1938894242,0.0081687309,-0.1355560273,0.0080531584,-0.2509580255,-0.0373694599,0.0572971366,0.379947722,-0.164004311,0.0147695262,0.090032205,0.0611759201,0.0604663938,0.2328324318,0.0334695764,-0.2108221501,-0.1764001697,-0.1549174488,-0.4953107834,-0.3186217248,-0.0450263172,-0.3158306479,-0.4681292772,0.3127001822,-0.4021078646,-0.1081995666,-0.0031759485,0.575340271,0.2059751302,-0.1731270552,0.2817462683,0.3410767615,-0.2425622195,0.2558035553,-0.1618468463,0.1206276342,-0.1269558668,0.374479413,0.0527705029,-0.4389058053,0.2052887082,0.0444829464,0.2107004672,0.102874659,-0.4216088951,0.0001691421,0.2830238938,0.2812859416,-0.0830779523,0.0920130461,0.3403278589,-0.3423228264,-0.1909656823,0.0451865792,-0.1148211733,-0.0747320205,-0.178488344,0.0911493599,-0.0704007819,-0.089724198,0.174400419,0.0561904423,0.1586268246,-0.2581044436,0.0801398531,0.025198523,-0.1384505183,-0.1799932271,0.4784166813,0.3956215382,-0.7722764015,-0.3988278806,0.1143974215,0.16897735,-0.1344381273,0.2855871916,-0.0892437398,0.0623679198,-0.2319801748,0.4219353795,0.1115486994,-0.1278712451,-0.1927447021,0.0685717985,0.0679323524,0.0048264419,0.2271308601,-0.1617614329,0.3247577846,0.1531335115,0.3888324201,0.071337685,0.0399592631,-0.3257410824,-0.1789891124,-0.3758673668,0.0991174653,0.1026760042,-0.0961196348,-0.0742361397,-0.1836107671,-0.3227061033,0.0883146152,-0.2380028069,-0.0485619232,0.0400439128,0.3023085296,0.4282283783,0.4427762628,0.0368306786,-0.0115634566,0.2662013471,0.0324873514,0.0951553211,-0.2985555232,-0.0519876368,-0.2356908619,0.0464012995,-0.0661984682,-0.20133847,-0.0659089163,-0.009110312,-0.0246781707,0.0475275218,-0.3220712543,0.0777285993,-0.1947692037,0.2195265889,-0.4618268013,0.4788576961,-0.2620447278,-0.2415874451,0.0462658219,0.3040081859,0.186638236,-0.1725841314,0.1663331091,0.094939366,-0.0937456936,-0.1992226392,0.1190167964,-0.1917669028,0.159357354,-0.1571003348,0.0574439988,-0.0669224784,0.192354843,0.1220769957,-0.1792988032,0.4870289862,0.0392018817,0.2792657614,-0.1853858978,-0.1539820135,0.1131258532,-0.0026398764,0.0101032462,-0.3382643461,-0.0889248773,-0.0046553123,-0.1199177504,-0.0626572818,0.0268098097,0.1231286377,0.0174136776,0.0285699107,-0.1038769111,0.053535033,-0.3347164094,0.0676233172,0.0471458174,-0.3681941032,0.3542415798,0.289367795,-0.1669884175,-0.246678561,-0.2466314435,-0.2522492409,0.1304650158,-0.0527902134,-0.2380274534,0.0619876496,-0.1329466254,0.0930444598,-0.2039635777,-0.0393128879,0.007351818,0.1990103424,-0.5885115862,0.0874701738,-0.3268994689,-0.0848519132,-0.0089322384,-0.1183215305,-0.0350329727,-0.3714100122,0.1818681508,-0.1842870414,-0.0339048915,0.2288198322,0.0108241513,0.1986064464,-0.0158729255,0.0349431261,-0.1994726807,-0.1922690272,-0.2766529322,-0.0750846788,-0.0261261296,0.2121435702,0.2464039624,-0.0174932294,-0.1035188213,-0.1036008671,-0.2901809812,0.2608532906,-0.2184689045,0.0960317403,-0.1443014145,0.1334666908,-0.1150142848,-0.2030875385,0.3162400126,-0.0712855309,-0.2233364731,-0.0748690814,0.1445922405,-0.0172571,-0.1171412542,-0.3439745307,0.0865703896,-0.2017683089,-0.0519000329,-0.1715556085,-0.0757853314,-0.3403173089,0.388887614,0.1098029464,-0.1206581891,0.0835879892,-0.2783555686,0.0760336444,0.3620343804,-0.0495057367,-0.229421407,0.16086182,0.0534705631,-0.2114243507,-0.073970668,-0.3018965423,-0.5255084038,-0.0137849767,0.2082932889,0.2294152379,0.4104632437,0.1461436599,-0.0166889336,-0.0821893066,-0.0818065554,-0.3491121233,0.0339904241,-0.1274582595,0.1863522828,-0.1580067128,0.5168377757,-0.2627623081,0.4241446257,0.4182263613,-0.3344832957,0.401863575,-0.1026427299,0.4811010957,-0.1938903332,-0.1900715828,0.1305034757,0.0521099679,0.0414344966,0.2256229371,-0.0783845633,0.1515789032,0.0500757471,0.274684608,-0.1358632147,-0.2487671077,0.0696826652,-0.0265750699,-0.0034900566,-0.1062735319,0.2850164771,-0.1292747259,0.150362134,0.2282374054,0.4088185728,0.369728744,-0.1107958704,0.3304606378,-0.3246169388,-0.2479223907,0.5417066813,0.0769141614,0.2590820193,-0.195161581,-0.0886173919,0.4172723591,0.1633312851,0.5660610199,-0.3769990802,-0.3231419027,0.2717078924,0.2671845853,-0.3712124527,-0.1177443862,-0.0206980594,-0.0435317829,-0.0180328935,0.6748079062,-0.4430410266,-0.0473423377,0.2658809721,0.0650958866,-0.1818184406,-0.4256838858,-0.4909735322,-0.1080659702,-0.4735231996,-0.0377673358,-0.0643602684,0.3547733724,0.0586999729,0.2814427912,-0.072004877,0.021322988,0.2308121771,0.3325805962,0.0880558863,0.0596565902,0.1927494854,0.4122957587,0.4405751824,0.2876455486,0.3852203488,-0.0886230171,-0.5030117035,0.1189950928,0.0795299038,0.1915130317,0.306651175,-0.0577340126,0.2583630979,-0.1517118514,0.0912912041,-0.2620264292,-0.3267897666,0.1230381653,0.1136146262,0.0320000164,-0.2792311013,0.4609561861,-0.0027985128,-0.1177051589,-0.1131939739,0.7454811335,-0.0445637256,0.9066886306,-0.0224905498,0.9038154483,-0.1837662458,0.0054945988,0.4124697149,-0.1430424005,0.5982552767,-0.1397474557,0.0056404513,-0.4372067451,-0.1494947672,-0.0543744192,0.0136123085,0.2145189494,-0.0728711858,-0.3070855439,0.122802183,-0.0723005831,-0.1465180367,-0.0893571973,-0.0270682201,0.0611018725,-0.4498348832,-0.3294598162,0.0078240773,0.4473239779,0.3218044043,0.0595408455,-0.079830803,-0.318334192,0.0049136328,-0.358839184,0.0712882653,0.2994551659,0.1605610102,0.3601663411,-0.2049202323,0.266972065,0.1277434975,0.716063261,-0.3889604807,-0.2528815866,0.1204800084,-0.366764456,-0.242717728,-0.128036201,-0.3013758063,0.3244974613,0.2448948473,-0.4326147139,0.2945291102,0.1806962639,-0.1646522284,-0.1618563533,0.145504728,-0.067706652,-0.3563479185,-0.1175533012,-0.0905720368,0.3375670612,-0.1444010437,0.065976575,0.2046833336,-0.0820481554,0.1091364026,0.0838302821,-0.1231091619,-0.0055338126,0.5033709407,-0.2285594493,0.1146046668,0.3499397337,0.1286236048,-0.4498925805,0.000388312,-0.4983139336,0.5014635921,-0.3836678267,-0.1845146865,0.3491197526,-0.2349703908,-0.0226772651,-0.0013709555,0.3730081022,-0.2326380908,-0.2815312147,-0.2601917684,-0.1082756892,0.3613974154,-0.1328597367,0.1616758406,-0.0860442668,0.1158279479,-0.1279035211,0.0128334723,-0.2126484066,0.1579232067,-0.1372287273,0.0753514618,-0.0734352395,-0.2143034488,-0.0054207034,-0.2459960282,0.0198299922,0.1660509557,0.0280935671,-0.0167215746,-0.0102717625,0.1721183509,0.0617215224,0.1258159131,-0.4139373899,-0.0717706531,-0.032081008,-0.3967060149,0.3543541431,0.2007510513,-0.0468919501,-0.0718142912,0.0021629347,0.0731871426,0.0412617475,0.2462120801,0.1241777614,0.5196718574,0.2802632153,-0.1163906604,-0.482732743,0.1483894289,0.0242191199,0.409227252,0.1506153643,-0.0409019664,0.3504683673,-0.2940050364,-0.0680305362,0.004269863,0.2235937417,0.3811903596,-0.0977509692,0.0681168586,0.2136992961,0.0474054925,-0.250356704,0.0347795039,0.3809844255,-0.1474710107,-0.0341240689,0.2338687479,0.1813001931,0.1227883622,-0.1979247481,0.069260411,0.2064262033,0.1313117594,-0.3917587698,0.0485444106,0.1434887797,0.2531651258,0.4207601547,0.2260639668,0.361533463,0.1074410602,-0.1683152318,0.0304047894,-0.0428696349,0.3624765873,-0.1702777147,-0.5442960262,0.4591153264,0.1990813464,0.4446201026,-0.3005710542,0.0110941259,0.2246434987,-0.0075122607,0.0029912821,-0.1264295131,0.0194798373,-0.2073527873,-0.4814678729,0.2469709069,0.0070394143,-0.1617958099,-0.0748416036,0.0904632732,-0.1671434194,-0.1127616093,-0.1288221478,-0.3127630353,-0.1633299738,-0.3798960745,0.2976242304,0.2746294439,0.0172058251,-0.0924302638,-0.0820374861,-0.0129512018,0.3628854752,0.0710031837,0.5582457781,0.092833288,-0.1161194593,0.1136865541,0.0108317435,0.0522577465,-0.0147902584,0.1180372834,-0.160405457,0.2515517771,0.2814060152,0.0793314874,-0.1262225062,-0.3866676688,-0.0260220002,0.6515324712,0.0928577036,-0.1293682903,-0.4074525237,0.0724367127,-0.1108803377,0.0383629091,-0.3113244772,0.2983470559,0.0368692912,-0.09624926,-0.012902745,0.2667491734,0.0236879569,-0.1871161759,0.2547547817,0.8629194498,-0.0874772444,-0.2432618439,0.1577601284,-0.5931995511,0.2370046675,-0.0749402791,-0.2206315249,0.2276031077,0.6163492203,0.4311290383,-0.0319095887,-0.0105085438,-0.0449877828,0.268978864,0.4219914973,-0.0000861897,-0.3120776415,-0.1198638901,-0.2052935362,-0.1085835472,0.0160013679,0.1113153696,-0.1017316505,-0.043883279,0.0389880501,0.1228803322,-0.1738748699,-0.3629499972,0.6114473939,0.0369078703,0.3470844924,-0.1360600889,-0.005969211,-0.1750845164,-0.1696828455,0.0329453051,0.2829148471,-0.1806606054,0.2205716074,-0.1766614914,-0.288369149,-0.2644261718,0.251652211,0.0712012425,-0.27015239,-0.2092000246,0.1826221645,-0.1758237928,0.0448129848,-0.0361687578,0.052193474,-0.0462342948,0.296638757,0.1056770459,-0.2640367448,0.6632553339,-0.3797849119,0.0318564475,0.1826309413,0.1951089501,0.4907484353,0.0071047232,-0.890553534,0.0415130481,0.0382940248,-0.0612080805,-0.3232002854,-0.0907065719,0.2005214542,0.0787764266,-0.2311304957,-0.1733236164,0.0168193355,0.1216910779,0.0177735742,-0.2673077881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2029","title":"Loading a faiss index KeyError","comments":"Ok in that case HF should fix their misleading example at https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index \r\n\r\nI copy-pasted it here.\r\n\r\n> When you are done with your queries you can save your index on disk:\r\n> \r\n> ```python\r\n> ds_with_embeddings.save_faiss_index('embeddings', 'my_index.faiss')\r\n> ```\r\n> Then reload it later:\r\n> \r\n> ```python\r\n> ds = load_dataset('crime_and_punish', split='train[:100]')\r\n> ds.load_faiss_index('embeddings', 'my_index.faiss')\r\n> ```","body":"I've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.\r\n\r\nThe basic steps are:\r\n\r\n1. Create a dataset (dataset1)\r\n2. Create an embeddings column using DPR\r\n3. Add a faiss index to the dataset\r\n4. Save faiss index to a file\r\n5. Create a new dataset (dataset2) with the same text and label information as dataset1\r\n6. Try to load the faiss index from file to dataset2\r\n7. Get `KeyError: \"Column embeddings not in the dataset\"`\r\n\r\nI've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing\r\n\r\nUbuntu Version\r\nVERSION=\"18.04.5 LTS (Bionic Beaver)\"\r\n\r\ndatasets==1.4.1\r\nfaiss==1.5.3\r\nfaiss-gpu==1.7.0\r\ntorch==1.8.0+cu101\r\ntransformers==4.3.3\r\n\r\nNVIDIA-SMI 460.56\r\nDriver Version: 460.32.03\r\nCUDA Version: 11.2 \r\nTesla K80 \r\n\r\nI was basically following the steps here: https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index\r\n\r\nI included the exact code from the documentation at the end of the notebook to show that they don't work either.\r\n","comment_length":57,"text":"Loading a faiss index KeyError\nI've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.\r\n\r\nThe basic steps are:\r\n\r\n1. Create a dataset (dataset1)\r\n2. Create an embeddings column using DPR\r\n3. Add a faiss index to the dataset\r\n4. Save faiss index to a file\r\n5. Create a new dataset (dataset2) with the same text and label information as dataset1\r\n6. Try to load the faiss index from file to dataset2\r\n7. Get `KeyError: \"Column embeddings not in the dataset\"`\r\n\r\nI've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing\r\n\r\nUbuntu Version\r\nVERSION=\"18.04.5 LTS (Bionic Beaver)\"\r\n\r\ndatasets==1.4.1\r\nfaiss==1.5.3\r\nfaiss-gpu==1.7.0\r\ntorch==1.8.0+cu101\r\ntransformers==4.3.3\r\n\r\nNVIDIA-SMI 460.56\r\nDriver Version: 460.32.03\r\nCUDA Version: 11.2 \r\nTesla K80 \r\n\r\nI was basically following the steps here: https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index\r\n\r\nI included the exact code from the documentation at the end of the notebook to show that they don't work either.\r\n\nOk in that case HF should fix their misleading example at https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index \r\n\r\nI copy-pasted it here.\r\n\r\n> When you are done with your queries you can save your index on disk:\r\n> \r\n> ```python\r\n> ds_with_embeddings.save_faiss_index('embeddings', 'my_index.faiss')\r\n> ```\r\n> Then reload it later:\r\n> \r\n> ```python\r\n> ds = load_dataset('crime_and_punish', split='train[:100]')\r\n> ds.load_faiss_index('embeddings', 'my_index.faiss')\r\n> ```","embeddings":[0.0172792971,-0.5772305727,0.0631230474,0.279512614,0.0801492929,0.2722066641,0.3179306686,0.1032484248,0.5430009365,0.2702875733,-0.1154732332,0.1119506285,0.4107095003,-0.1057186723,-0.0582559854,-0.0193880815,0.2288171202,0.2564585805,0.2251714319,-0.133942306,-0.2051008195,0.0973475128,-0.2735082805,0.2104548216,-0.2178766876,0.1518886089,-0.2278769314,0.0568944402,-0.1768641174,-0.4319815338,0.3245584965,-0.2239339501,0.6598060727,0.1977226138,-0.0001215702,0.1218788475,0.3807426989,-0.1595098823,-0.2513687015,0.1151010543,0.0764162019,0.0272239614,0.2132130265,-0.0726772174,0.0061008078,-0.6095624566,0.0533289872,-0.1794200838,0.2412755638,0.0245589074,0.1126011312,0.0576791242,0.0422044955,-0.3755269051,0.0992318392,-0.3127266467,-0.0681922734,0.2848208547,-0.0071604131,0.0023000252,0.0063478462,0.3573625684,-0.0200696494,0.019340964,0.40046224,0.2284757048,0.3043691218,-0.104426235,0.1567376405,0.1700628698,0.3159224689,-0.154877454,-0.3962084949,-0.1216680035,0.1943065971,-0.126398623,0.1970370412,-0.3063472509,0.0235137101,0.3119217753,0.3952243328,-0.1900372207,0.1177399307,0.0820095688,-0.0645475909,-0.0513839945,-0.210855037,0.0240792986,0.1902793646,-0.0865309462,-0.282009393,-0.1559734195,0.0747636557,0.082327269,-0.3504092395,0.1680515409,0.2065456957,0.0209433213,-0.0723240972,-0.0248584077,-0.2064311802,0.0152599523,0.1569410264,0.3611029983,-0.2030455619,-0.017811818,0.0448900573,0.0191625301,0.1014064252,0.2769632638,-0.0010159569,-0.2169072479,-0.1131690145,-0.1501648128,-0.4902278185,-0.3367262781,-0.0414405912,-0.348993659,-0.4786964655,0.3067212999,-0.4552730024,-0.130434528,0.018645715,0.6528446674,0.1860147119,-0.2571901679,0.252235204,0.3296818435,-0.3448681235,0.2203894705,-0.1981097162,0.1131346524,-0.1160796732,0.4127857387,0.1097563803,-0.4600390792,0.1754471064,0.0120851984,0.2221800536,0.0842839256,-0.3430346251,-0.0307907071,0.3162079155,0.2977688015,-0.0658048019,0.0719670281,0.3783288896,-0.3262055516,-0.2240067273,0.0465687253,-0.126117155,-0.1155303195,-0.1683064103,0.114999406,-0.0447194502,-0.0695156902,0.235982269,0.0991610661,0.1431575119,-0.2634614706,0.0815166011,0.0393043198,-0.1817218661,-0.1781526506,0.4668316245,0.4261373878,-0.7533151507,-0.3583073318,0.1044595167,0.224639833,-0.1304668635,0.2801527381,-0.1336553246,0.0512032956,-0.2663244009,0.4776728451,0.160535723,-0.1500926614,-0.2101146579,0.1152872294,0.1037184149,0.0163553115,0.1906678379,-0.1538437605,0.3397403955,0.1171113402,0.3372926116,-0.0242131148,0.0364694707,-0.3472739458,-0.2551965415,-0.3373095393,0.0823269412,0.0959324986,-0.0580283813,-0.1127974764,-0.1097772121,-0.3149940968,0.1116435379,-0.2194283158,-0.0689909086,0.1341983676,0.2931415141,0.39780882,0.4509916306,0.0653036982,-0.0793156028,0.2604404688,0.0285848882,0.0608433485,-0.1846120656,-0.1017718986,-0.2828640342,0.0477874689,-0.1088200212,-0.2191662639,-0.0282452758,-0.0028138554,-0.0579029992,0.0574611612,-0.3144201934,0.1071905494,-0.2037627101,0.1530472189,-0.4385951161,0.4663409591,-0.3173539937,-0.2310287952,0.0460032001,0.2543338537,0.1745008379,-0.1636387557,0.1295262575,0.1026977822,-0.0566196516,-0.2618732154,0.1746410429,-0.1655095518,0.1427622139,-0.2291909605,0.0832302794,-0.0419446938,0.2045514137,0.1157635376,-0.2281857133,0.4575839937,0.052060809,0.2677704096,-0.147359997,-0.2205034792,0.1386884302,0.0298014004,-0.0034481036,-0.3238343894,-0.1025569364,0.0377517603,-0.1164529398,-0.0653850213,0.0181211494,0.1922570169,0.0198896974,0.0316699035,-0.1030548885,0.0811387077,-0.2599281073,0.0253882762,0.0923055783,-0.4249454141,0.3604758382,0.2861751616,-0.1327142715,-0.2303947806,-0.2657188177,-0.2874986827,0.067374818,-0.0804419145,-0.2273133397,0.1274018884,-0.0742557272,0.0862707794,-0.2223926038,-0.0440521315,0.0122665884,0.14903754,-0.5887740254,0.1396591216,-0.3880037963,-0.0942457318,0.029119445,-0.1383638084,-0.0869042873,-0.38944754,0.2312373668,-0.1868146658,-0.0542590059,0.2351478785,-0.0808565244,0.182911247,0.0158675853,-0.0098379692,-0.1690037996,-0.1615710407,-0.3648664951,-0.0854422674,-0.0241525564,0.2970384359,0.2358442545,-0.0316454992,-0.1209298298,-0.1339874715,-0.3524387479,0.2762563825,-0.143265456,0.1291477531,-0.1065488905,0.0773718283,-0.1703564078,-0.1709821969,0.3147136867,-0.0222135,-0.1975483745,-0.0149076078,0.1242613047,0.0355618633,-0.0651528388,-0.3328302205,0.1047071815,-0.2776853442,-0.1136709005,-0.1886593103,-0.0535951145,-0.3046745062,0.3891735375,0.1195937842,-0.1083313227,0.0964702517,-0.2930319011,0.1472296715,0.3678099513,-0.1009715348,-0.2994407117,0.1863138974,0.0289020743,-0.2174456269,-0.0439000092,-0.3964253068,-0.4838353395,-0.0262634344,0.223999843,0.2083174735,0.3773666024,0.1585734785,-0.0878758579,-0.0857687965,-0.0988790467,-0.3516177237,0.0867616683,-0.1840332896,0.1607315242,-0.1917382479,0.5027089119,-0.2363250405,0.4415203929,0.4734849036,-0.3138492107,0.4359805882,-0.0560271367,0.502671957,-0.2558838427,-0.1605994105,0.1289680749,0.0415621363,-0.0153032886,0.2319121957,-0.0699337944,0.1313170791,0.0795842931,0.248021096,-0.0846442804,-0.2826348543,0.1131232232,0.0279455334,0.0049015312,-0.0829245597,0.2915319204,-0.0784453675,0.1364877969,0.2619969547,0.4994945824,0.374681592,-0.0595980994,0.3922794461,-0.2716040909,-0.2988324463,0.5406864882,0.0766688287,0.2704482675,-0.2316275835,-0.1329107434,0.3817055225,0.1187173873,0.4669493139,-0.3202130795,-0.276958257,0.3623041511,0.215747714,-0.3539117575,-0.1602832526,-0.0421993583,-0.0649764016,-0.045661252,0.6810308099,-0.4488099217,-0.0919196904,0.2719634175,0.0484947711,-0.191775471,-0.4296601415,-0.4913111925,-0.2099976689,-0.4370381534,-0.0712673962,-0.0832089484,0.2882447541,0.0675765127,0.2692575753,-0.0639866218,0.0211343765,0.1759268045,0.2988153398,0.0854445696,0.0554003641,0.222578302,0.3516510725,0.4493592381,0.3105473816,0.3761473596,-0.0923000053,-0.3685176671,0.1949786693,0.1215486825,0.1347626001,0.3488344848,-0.0598926172,0.3128324747,-0.1410940886,0.1342670023,-0.2591910362,-0.3105552793,0.0768892169,0.0582714826,0.0437598191,-0.2096960098,0.5280877352,-0.0176830217,-0.108842127,-0.1019317582,0.7750861645,-0.0008788293,0.8857607841,0.0565028377,0.9090359211,-0.1713172048,0.0234891213,0.4463816583,-0.1876297146,0.4825900495,-0.1951458454,0.0426127762,-0.4373918176,-0.2519880831,-0.0566142984,0.0048723649,0.1658104062,-0.0841307417,-0.3571432233,0.1029772535,-0.0496154241,-0.1644603312,-0.0859971717,-0.0637874454,-0.0010358626,-0.3792616427,-0.3039005101,0.0290283374,0.4822971523,0.3695409894,0.0220443122,-0.0790324658,-0.291441232,0.0064428342,-0.3014764786,0.0409133099,0.2598647475,0.1597630382,0.3091951609,-0.1646123081,0.2454795241,0.035959255,0.6067768931,-0.3963741958,-0.2278278917,0.0660856217,-0.2978229821,-0.2367136329,-0.0910730064,-0.3210307956,0.3530059457,0.2164136618,-0.3930769563,0.2455592602,0.1277557462,-0.1699525416,-0.1908746213,0.1798235625,-0.1001081988,-0.3621915877,-0.0792510435,-0.0853156,0.2435476929,-0.0945236385,0.0748394802,0.1713629216,-0.0192909148,0.1398162246,0.1292548776,-0.1324254721,0.006368631,0.5463545918,-0.260517627,0.1305040866,0.3212526441,0.1546414793,-0.4956640303,-0.0322358385,-0.5321473479,0.4935287237,-0.3097320795,-0.1765691191,0.228667289,-0.1850276887,-0.0002237296,0.035654936,0.4112836123,-0.3013073504,-0.2987973392,-0.2700887322,-0.095146209,0.4340412319,-0.0944088697,0.1565034539,-0.0819235891,0.079901658,-0.1040564179,0.0419618562,-0.2391067296,0.1509058028,-0.1227875948,0.0378885306,-0.1788295656,-0.2138812095,0.0205805767,-0.1755611002,0.0168479551,0.1716372818,-0.0522175804,-0.0424936488,0.0114651937,0.1759465188,0.0268441569,0.0967809707,-0.3826612234,-0.0819117278,-0.0648974851,-0.3780142963,0.3031316996,0.1970614791,-0.0779175013,-0.1188662872,0.0153416758,0.0307868309,0.0624903478,0.2531218827,0.184230566,0.528167069,0.2460182905,-0.1327631623,-0.5212600231,0.1234878972,0.0725357756,0.4531140625,0.1454980373,-0.0640456378,0.3715769351,-0.3348658979,-0.0354462191,-0.0635817945,0.2820008397,0.3686344028,-0.1174698919,0.0312433448,0.1743299961,0.0696528777,-0.2446893007,0.0674639717,0.37839818,-0.127639249,0.021246532,0.2595114708,0.1960684806,0.1245091408,-0.17767784,0.0934912339,0.1437723041,0.1801262498,-0.4203804433,0.0146657899,0.0593919121,0.2737781405,0.3931542039,0.1846985966,0.4034123123,0.0729817003,-0.1266162843,0.0767262578,0.0048418855,0.3915552795,-0.2059018463,-0.6314306855,0.4730730355,0.2434369475,0.5203196406,-0.3003265262,-0.0273208395,0.2113475502,0.0158675648,0.0019209201,-0.1404470652,0.008723882,-0.1797018349,-0.4228444695,0.2603083849,-0.0135752056,-0.2500580549,-0.1014534011,0.0556352623,-0.1375440359,-0.1221994087,-0.0983669907,-0.2738232613,-0.1351662427,-0.2959284782,0.3009161949,0.2911666632,0.0407974236,-0.0253563486,-0.0420554206,0.0072704763,0.3467206657,0.0485006645,0.532879293,0.142139405,-0.1048809588,0.0760595649,0.0631731525,0.0589937568,0.0000358203,0.1344168633,-0.1673480868,0.2181902379,0.3409720063,0.0637019947,-0.1314150095,-0.3197767437,-0.0504972748,0.5998354554,0.1000026986,-0.1618773192,-0.3587870598,0.104611285,-0.0621893406,0.0215112157,-0.2632642686,0.2826019228,0.1195051447,-0.140535444,0.0846822858,0.2874371111,0.0385906361,-0.1772707403,0.3113382757,0.8044199347,-0.0780334547,-0.2670656145,0.1288082451,-0.542734623,0.2680163383,-0.1829782128,-0.1850353628,0.2906189561,0.6496264935,0.4190321863,-0.0323069617,0.0721314549,-0.1104706302,0.3419901431,0.4437326789,0.0212640353,-0.3350872397,-0.1664215326,-0.2112395763,-0.1398438662,-0.0645549074,0.1304071248,-0.0612431131,-0.012106631,0.0348459408,0.106087409,-0.0949874371,-0.343710959,0.6024445295,0.0380871333,0.350869149,-0.1602428555,-0.0448032394,-0.1796191782,-0.1176821142,0.0382440127,0.2302426398,-0.2109965086,0.1636526436,-0.1435657442,-0.3141819835,-0.2684743404,0.293322742,0.0934773162,-0.2817205787,-0.1717187166,0.1432599127,-0.2002487779,0.0334747098,-0.0270992313,0.0788930804,-0.0335665792,0.3309988976,0.07544294,-0.2962703109,0.69662714,-0.3728350401,0.0705316812,0.2038527727,0.2360535264,0.50114429,0.0328812189,-0.8710232973,-0.0064153816,0.0906481668,-0.0719543993,-0.2740331292,-0.0554657578,0.1980943233,0.0384441204,-0.2326389402,-0.0869512856,-0.0006064575,0.1136784628,-0.0463073365,-0.2873271704]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2029","title":"Loading a faiss index KeyError","comments":"Hi !\r\n\r\nThe code of the example is valid.\r\nAn index is a search engine, it's not considered a column of a dataset.\r\nWhen you do `ds.load_faiss_index(\"embeddings\", 'my_index.faiss')`, it attaches an index named \"embeddings\" to the dataset but it doesn't re-add the \"embeddings\" column. You can list the indexes of a dataset by using `ds.list_indexes()`.\r\n\r\nIf I understand correctly by reading this example you thought that it was re-adding the \"embeddings\" column.\r\nThis looks misleading indeed, and we should add a note to make it more explicit that it doesn't store the column that was used to build the index.\r\n\r\nFeel free to open a PR to suggest an improvement on the documentation if you want to contribute :)","body":"I've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.\r\n\r\nThe basic steps are:\r\n\r\n1. Create a dataset (dataset1)\r\n2. Create an embeddings column using DPR\r\n3. Add a faiss index to the dataset\r\n4. Save faiss index to a file\r\n5. Create a new dataset (dataset2) with the same text and label information as dataset1\r\n6. Try to load the faiss index from file to dataset2\r\n7. Get `KeyError: \"Column embeddings not in the dataset\"`\r\n\r\nI've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing\r\n\r\nUbuntu Version\r\nVERSION=\"18.04.5 LTS (Bionic Beaver)\"\r\n\r\ndatasets==1.4.1\r\nfaiss==1.5.3\r\nfaiss-gpu==1.7.0\r\ntorch==1.8.0+cu101\r\ntransformers==4.3.3\r\n\r\nNVIDIA-SMI 460.56\r\nDriver Version: 460.32.03\r\nCUDA Version: 11.2 \r\nTesla K80 \r\n\r\nI was basically following the steps here: https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index\r\n\r\nI included the exact code from the documentation at the end of the notebook to show that they don't work either.\r\n","comment_length":119,"text":"Loading a faiss index KeyError\nI've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.\r\n\r\nThe basic steps are:\r\n\r\n1. Create a dataset (dataset1)\r\n2. Create an embeddings column using DPR\r\n3. Add a faiss index to the dataset\r\n4. Save faiss index to a file\r\n5. Create a new dataset (dataset2) with the same text and label information as dataset1\r\n6. Try to load the faiss index from file to dataset2\r\n7. Get `KeyError: \"Column embeddings not in the dataset\"`\r\n\r\nI've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing\r\n\r\nUbuntu Version\r\nVERSION=\"18.04.5 LTS (Bionic Beaver)\"\r\n\r\ndatasets==1.4.1\r\nfaiss==1.5.3\r\nfaiss-gpu==1.7.0\r\ntorch==1.8.0+cu101\r\ntransformers==4.3.3\r\n\r\nNVIDIA-SMI 460.56\r\nDriver Version: 460.32.03\r\nCUDA Version: 11.2 \r\nTesla K80 \r\n\r\nI was basically following the steps here: https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index\r\n\r\nI included the exact code from the documentation at the end of the notebook to show that they don't work either.\r\n\nHi !\r\n\r\nThe code of the example is valid.\r\nAn index is a search engine, it's not considered a column of a dataset.\r\nWhen you do `ds.load_faiss_index(\"embeddings\", 'my_index.faiss')`, it attaches an index named \"embeddings\" to the dataset but it doesn't re-add the \"embeddings\" column. You can list the indexes of a dataset by using `ds.list_indexes()`.\r\n\r\nIf I understand correctly by reading this example you thought that it was re-adding the \"embeddings\" column.\r\nThis looks misleading indeed, and we should add a note to make it more explicit that it doesn't store the column that was used to build the index.\r\n\r\nFeel free to open a PR to suggest an improvement on the documentation if you want to contribute :)","embeddings":[0.139708057,-0.5335465074,0.0514714457,0.3196974695,0.1264748573,0.2750567496,0.4119959772,-0.0404785536,0.6586285233,0.2073548138,-0.0624137037,0.153858155,0.3866050839,-0.0489014611,0.0130298138,-0.0173613261,0.3034028709,0.2208625972,0.2414616048,-0.1565983891,-0.1813698411,0.0717174187,-0.2770503163,0.1112730652,-0.1968314648,0.1754428148,-0.2499136031,-0.0327954814,-0.2255384624,-0.3927151859,0.3298938274,-0.2455042154,0.6396087408,0.1909313798,-0.0001244103,0.1136664003,0.3821778893,-0.1372379363,-0.2595387101,0.1243394017,-0.0407922491,0.0283891633,0.2259729803,-0.0741742477,0.0443998501,-0.6045323014,0.0306181759,-0.2664104402,0.2041671723,0.0784025118,0.0811231956,0.0266717058,0.0594932884,-0.3501586318,0.1453064531,-0.29447034,-0.1072397977,0.2723374963,0.0363665484,0.0591787286,0.0967705175,0.2234830856,-0.0297007617,-0.0044607348,0.3816767037,0.3206753731,0.3233527839,-0.1499414593,0.1832972765,0.1396442652,0.4778426588,-0.0867141932,-0.3786843717,-0.1274588257,0.2133221924,-0.1693466753,0.1781800389,-0.3018350303,0.0494873561,0.2833381593,0.3947203457,-0.2305176258,0.1729187667,0.1110191867,-0.1058646441,0.0296114553,-0.2333698124,0.0812688395,0.1254786402,-0.0815552846,-0.2309127748,-0.1110212654,0.1203840896,0.1144591793,-0.3478138745,0.2244480401,0.210499987,0.0192310531,-0.1170734689,0.0355053283,-0.1175126806,0.0451515131,0.0825698897,0.3602949083,-0.161278978,0.0425950997,0.0905499384,-0.0014759696,0.0315709524,0.2060628831,0.0346700773,-0.2354734838,-0.1273182929,-0.0511647947,-0.4601175785,-0.3972397447,-0.0220848061,-0.293967247,-0.4682563245,0.3269699216,-0.3949016035,-0.1495504081,0.0359078683,0.6100538373,0.2423648089,-0.2410033941,0.2574202418,0.3329061866,-0.2494074851,0.2505556643,-0.2155791223,0.100160785,-0.179163456,0.4160974026,0.0883165598,-0.4362990558,0.2058899403,-0.0156852864,0.1153184846,0.0160861574,-0.4422804117,-0.0107994024,0.3289083242,0.3100945652,-0.11815501,0.06469886,0.2906962931,-0.3939503729,-0.1781438142,0.0867635831,-0.1132606044,-0.0243502967,-0.1908863187,0.0711474419,-0.1113985702,-0.1381538957,0.2029106319,0.1236761436,0.1540800482,-0.2610228956,0.1042369753,0.0324439295,-0.1694502234,-0.1880511492,0.5441301465,0.390566051,-0.754373908,-0.3546563387,0.0210219808,0.1953683048,-0.0898331255,0.2088247091,-0.0703034028,0.1384410709,-0.2801352143,0.4364034235,0.2500101924,-0.0299273785,-0.1367249489,-0.0278813988,0.0822835788,0.011289943,0.1462400109,-0.04641141,0.3089675605,0.2303739935,0.2861495614,0.0061393036,0.0597964413,-0.3615947366,-0.1948387325,-0.2727174759,0.1160120293,0.0804308429,-0.0372421294,-0.0280089863,-0.0962934196,-0.3618800044,0.0644499883,-0.1609582454,-0.0479053222,0.0518079177,0.319955796,0.42494753,0.4439717829,0.0635534972,-0.1428378075,0.2731236815,0.1668573618,0.0024764196,-0.1946393996,-0.0410471819,-0.3104975522,0.1002364233,-0.0861708149,-0.1901261508,-0.0936026499,0.0434049889,-0.0393041335,0.0348566957,-0.3543615341,0.0917701945,-0.1389931142,0.1869313419,-0.4618312716,0.4687990248,-0.244671911,-0.2399786562,-0.0660559684,0.2928885221,0.0953243077,-0.23050116,0.1591469049,0.1184905693,0.0150072752,-0.2839363813,0.1743018031,-0.175525561,0.1507220715,-0.212949276,0.1087340266,-0.0652895123,0.256676048,0.0690950677,-0.239273876,0.4237160385,0.0181186292,0.2957782745,-0.1987942308,-0.21147798,0.1270289868,0.041851487,-0.0545729175,-0.301696986,-0.1867242754,-0.0259514283,-0.1224479675,-0.0458274446,0.0013797299,0.1479832083,-0.0317740999,0.0533249378,-0.0762059763,0.0467843376,-0.3594025671,0.0175261647,0.0633010194,-0.4014998972,0.4099878669,0.267688185,-0.042484086,-0.2847337723,-0.158041507,-0.292766422,0.0362489186,-0.0372612216,-0.2339008451,0.1330188066,-0.0564494058,0.0249525327,-0.2709622383,0.0384881869,0.1010742933,0.1226191595,-0.5553826094,0.0918593928,-0.3217408955,-0.0228082351,0.0247970279,-0.1072799042,-0.0440434925,-0.3912600279,0.1531036347,-0.1146237552,-0.0169351585,0.2143235952,-0.0619997643,0.218171373,-0.0247203466,-0.0401436351,-0.187653169,-0.2124548405,-0.3034582734,-0.0988075808,-0.0032459972,0.295403868,0.2241867334,-0.0202666204,-0.07088162,-0.1617212147,-0.3294917643,0.2559995651,-0.1673352122,0.1325889826,-0.1250998527,0.1196718812,-0.1969907135,-0.1260314584,0.2635339797,-0.0487390421,-0.1794922203,-0.0537561104,0.1822658777,0.0249118358,-0.1008231267,-0.4248007536,0.1342135072,-0.1854077578,-0.0751737133,-0.1683256179,-0.0700722784,-0.2950831652,0.3934319615,0.1232633367,-0.0822989047,0.0814593434,-0.3621648252,0.1432950199,0.3559267223,-0.0511646233,-0.30620417,0.2269723713,-0.023869615,-0.2126529813,-0.0232424326,-0.300165087,-0.4479035735,0.0247274693,0.148649171,0.3076381087,0.4144200683,0.1959233135,-0.1259261519,-0.0382230207,-0.1248020902,-0.420760721,0.1242525429,-0.0117373336,0.1973400861,-0.0983321816,0.5074976683,-0.3051088452,0.4946763515,0.3491691351,-0.4038631916,0.449875176,-0.1538039297,0.5353226066,-0.2270278633,-0.187646836,0.097731173,0.0977133512,-0.0742567182,0.2291563898,-0.077146709,0.1077797636,0.0786255524,0.3998549283,-0.1534987539,-0.2350585908,0.0933814645,0.0416990295,-0.0043322886,-0.0967477784,0.2947267592,-0.0990584195,0.1402982473,0.1448268741,0.4661459029,0.4105900526,-0.1144731492,0.3640741706,-0.2789015174,-0.3221739531,0.5076093078,-0.0094893705,0.3112463057,-0.1909679323,-0.2257632762,0.4270910025,0.1622133553,0.5094698668,-0.3713767827,-0.3883592188,0.3326300979,0.2954976559,-0.3631890714,-0.1525405794,-0.0273867249,-0.0215931162,-0.0370033421,0.6526851058,-0.4228726923,-0.0011689346,0.2416596562,0.081842415,-0.2017993033,-0.3822226822,-0.5365284085,-0.2004128247,-0.3691923022,-0.0537201948,-0.1174363121,0.3132626116,0.1693927646,0.1766942292,-0.0936179832,-0.0618894808,0.1937378645,0.3094623983,0.084909007,0.1815102398,0.2277319282,0.3715901375,0.4876290858,0.2552193403,0.3997570276,-0.190037936,-0.2439514101,0.186018154,0.0257179085,0.211774081,0.2899109721,-0.0330047309,0.203644067,-0.1958997399,0.0649260432,-0.220166117,-0.2590919435,0.1462918222,0.0934910774,0.0041815951,-0.2914550006,0.5448599458,-0.0175987594,-0.2207261324,-0.1630166471,0.7354040742,-0.0529471897,0.8978901505,-0.0028410179,0.9030044079,-0.2579962015,-0.0560950786,0.4146472812,-0.0845782459,0.5710940957,-0.1455394775,0.0741493106,-0.4189853668,-0.23358953,-0.0828632265,-0.052309785,0.1802420616,-0.1072507948,-0.290047586,0.0874736309,-0.0836214423,-0.134256199,-0.0875007436,-0.073203519,0.1548388898,-0.3399547935,-0.2096614689,0.0111379139,0.475135386,0.3716224134,0.0004492543,-0.0049822945,-0.3148491085,-0.0208307877,-0.294821173,0.025920758,0.3043396175,0.1437457949,0.3011051714,-0.2468359023,0.2028660923,0.136037305,0.5763691664,-0.4888342917,-0.1904988736,0.1182536185,-0.3001485765,-0.2296494246,-0.158682093,-0.3349235058,0.2959028184,0.3055291176,-0.3957147896,0.2116837949,0.1347509772,-0.2160882801,-0.1810226142,0.2022085786,-0.034885671,-0.4023381174,-0.1147627011,0.0080678733,0.2858684957,-0.0976247117,0.0407560728,0.1661278903,-0.0370492674,0.2013924867,0.147473678,-0.1411550194,-0.0054668109,0.5710303187,-0.2033818662,0.1725045592,0.2977305353,0.1402340084,-0.435829699,-0.0353426076,-0.5515818596,0.5049035549,-0.2644136548,-0.2070164233,0.2630949914,-0.1769471616,0.0523212291,0.0210847147,0.3146510422,-0.2509732842,-0.3915979862,-0.2711707354,-0.047689382,0.4246589839,-0.0891589746,0.1883840412,-0.1719497889,0.0001761243,-0.0920156762,0.0131457495,-0.1961168945,0.1657242775,-0.0758267269,0.1075026318,-0.0297344998,-0.1764875203,0.0076232734,-0.3154183626,0.0121846069,0.0821583197,0.0200673211,-0.0173694752,0.0558543541,0.1841937453,0.0815486759,0.1543059498,-0.4690368474,-0.036917083,-0.0566254556,-0.4456321001,0.3262780607,0.1444025785,-0.1596531719,-0.1317916214,0.0373504572,0.141336605,0.0474607982,0.212556988,0.1360689551,0.5164855719,0.2325788587,-0.1356124878,-0.4268552661,0.1551964283,0.0897998661,0.4254201651,0.1661588401,-0.1481363177,0.3840103447,-0.3068833053,0.0112675987,-0.0135346614,0.213460952,0.4560756981,-0.0628297031,0.0791653767,0.1910355091,0.0411778763,-0.1962540597,0.0836328417,0.395475328,-0.1042647287,0.0151064126,0.2821960747,0.2161112577,0.1574378759,-0.2313300222,0.0467251875,0.1695625484,0.2741905451,-0.3947870731,-0.0031681219,0.0739474446,0.2783548832,0.4403621256,0.170648545,0.3151627779,0.1510327607,-0.1643122435,0.1035961732,0.0389908962,0.4055016041,-0.185333088,-0.5798711777,0.5499575138,0.2564828694,0.4967413545,-0.2785110176,0.0276356246,0.2131630033,0.0077694119,-0.0112876426,-0.2104987055,0.0370273516,-0.1464020759,-0.4431442916,0.2175291181,-0.0443866551,-0.2370198071,-0.1008000225,0.0298514254,-0.1973166913,-0.1029090583,-0.1672842056,-0.3034962118,-0.1612385511,-0.3351871967,0.2984619439,0.2237494588,0.0694232285,-0.0510128923,-0.172666043,-0.0147902006,0.3651046753,0.0982870311,0.4093309343,0.0922429338,-0.05016651,0.1371534914,0.0476711914,0.0098275794,-0.0680904761,0.1368502825,-0.2060672343,0.1784705818,0.2098882049,0.0624125078,-0.0611174032,-0.3300631642,-0.0272065941,0.6360585093,0.0477027334,-0.1378114671,-0.3980540335,0.0392396115,-0.083263278,-0.0116745485,-0.2653491497,0.2788213193,0.0598351695,-0.1342647076,0.0567127913,0.3222052157,0.0116631342,-0.2166980654,0.265363425,0.7818457484,-0.0992651433,-0.1718611866,0.1167359129,-0.5884333253,0.2082718611,-0.1338330209,-0.2299387604,0.2374172807,0.6179019213,0.4279896915,-0.0581782758,0.0317359604,-0.0683934465,0.2986148298,0.3998093307,0.0247500744,-0.3153722882,-0.162643984,-0.1644559354,-0.1396199316,0.0125087155,0.1287264973,-0.1496262997,-0.061278861,0.0495574661,0.1129623875,-0.1767143607,-0.3649273813,0.7231637239,0.0115464674,0.371481508,-0.1130697876,0.0516385175,-0.2628451586,-0.2042689621,0.0472769029,0.3148731589,-0.192939803,0.1571576893,-0.1494903713,-0.2276141495,-0.18848598,0.2476057708,0.0908745527,-0.2714473009,-0.230993703,0.128522262,-0.2478741705,0.0102836676,-0.0149865383,0.0248446707,-0.0483203121,0.2772885561,0.1054853722,-0.2724974751,0.5993978977,-0.4096085727,0.1013203636,0.0972105488,0.221180439,0.5730558038,0.0223233905,-0.865645051,-0.0058098659,0.0837960765,-0.1317081004,-0.2458977401,-0.098705098,0.1803891063,0.0558538213,-0.3165011108,-0.1591804773,-0.0125577021,0.1369659901,-0.0901832506,-0.2896229029]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2029","title":"Loading a faiss index KeyError","comments":"> If I understand correctly by reading this example you thought that it was re-adding the \"embeddings\" column.\r\nYes. I was trying to use the dataset in RAG and it complained that the dataset didn't have the right columns. No problems when loading the dataset with `load_from_disk` and then doing `load_faiss_index`\r\n\r\nWhat I learned was\r\n1. column and index are different\r\n2. loading the index does not create a column\r\n3. the column is not needed to be able to use the index\r\n4. RAG needs both the embeddings column and the index\r\n\r\nIf I can come up with a way to articulate this in the right spot in the docs, I'll open a PR","body":"I've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.\r\n\r\nThe basic steps are:\r\n\r\n1. Create a dataset (dataset1)\r\n2. Create an embeddings column using DPR\r\n3. Add a faiss index to the dataset\r\n4. Save faiss index to a file\r\n5. Create a new dataset (dataset2) with the same text and label information as dataset1\r\n6. Try to load the faiss index from file to dataset2\r\n7. Get `KeyError: \"Column embeddings not in the dataset\"`\r\n\r\nI've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing\r\n\r\nUbuntu Version\r\nVERSION=\"18.04.5 LTS (Bionic Beaver)\"\r\n\r\ndatasets==1.4.1\r\nfaiss==1.5.3\r\nfaiss-gpu==1.7.0\r\ntorch==1.8.0+cu101\r\ntransformers==4.3.3\r\n\r\nNVIDIA-SMI 460.56\r\nDriver Version: 460.32.03\r\nCUDA Version: 11.2 \r\nTesla K80 \r\n\r\nI was basically following the steps here: https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index\r\n\r\nI included the exact code from the documentation at the end of the notebook to show that they don't work either.\r\n","comment_length":115,"text":"Loading a faiss index KeyError\nI've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.\r\n\r\nThe basic steps are:\r\n\r\n1. Create a dataset (dataset1)\r\n2. Create an embeddings column using DPR\r\n3. Add a faiss index to the dataset\r\n4. Save faiss index to a file\r\n5. Create a new dataset (dataset2) with the same text and label information as dataset1\r\n6. Try to load the faiss index from file to dataset2\r\n7. Get `KeyError: \"Column embeddings not in the dataset\"`\r\n\r\nI've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing\r\n\r\nUbuntu Version\r\nVERSION=\"18.04.5 LTS (Bionic Beaver)\"\r\n\r\ndatasets==1.4.1\r\nfaiss==1.5.3\r\nfaiss-gpu==1.7.0\r\ntorch==1.8.0+cu101\r\ntransformers==4.3.3\r\n\r\nNVIDIA-SMI 460.56\r\nDriver Version: 460.32.03\r\nCUDA Version: 11.2 \r\nTesla K80 \r\n\r\nI was basically following the steps here: https:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html#adding-a-faiss-index\r\n\r\nI included the exact code from the documentation at the end of the notebook to show that they don't work either.\r\n\n> If I understand correctly by reading this example you thought that it was re-adding the \"embeddings\" column.\r\nYes. I was trying to use the dataset in RAG and it complained that the dataset didn't have the right columns. No problems when loading the dataset with `load_from_disk` and then doing `load_faiss_index`\r\n\r\nWhat I learned was\r\n1. column and index are different\r\n2. loading the index does not create a column\r\n3. the column is not needed to be able to use the index\r\n4. RAG needs both the embeddings column and the index\r\n\r\nIf I can come up with a way to articulate this in the right spot in the docs, I'll open a PR","embeddings":[0.1141662821,-0.6093646288,0.0622365586,0.3725110888,0.1373017132,0.2862834334,0.4066281021,0.0279657878,0.5614460111,0.2074248642,-0.0576396435,0.1514220387,0.4737912714,-0.0607365854,-0.052989319,-0.0416151509,0.2702905834,0.2431628108,0.2737343609,-0.1626692861,-0.2025646716,0.1247658506,-0.2621251643,0.1718708277,-0.248210907,0.2487478256,-0.2778123617,0.0308961533,-0.2404202968,-0.3826129138,0.3425858915,-0.2952020168,0.6600481868,0.1492620856,-0.0001231916,0.1020782292,0.3273308277,-0.1871940792,-0.2122855037,0.0915357992,0.1122795045,0.0143894553,0.2014682144,-0.0870565027,-0.0057215164,-0.6249670982,-0.0121625978,-0.1644823253,0.1824403107,0.0332699306,0.1068647653,-0.0065372856,0.0351543799,-0.378389746,0.205112651,-0.333899796,-0.1093248278,0.3227841258,0.091817826,0.0659408346,0.054192394,0.1877554655,-0.0682540908,0.1206275746,0.4152822495,0.2805346549,0.3271452487,-0.094155848,0.1916977316,0.1926267892,0.4542701542,-0.0401832089,-0.3339141309,-0.1143841222,0.2563905418,-0.1585264355,0.2307870388,-0.2589789927,0.0183759909,0.3088291585,0.3913377821,-0.2083024085,0.14921166,0.1125609502,-0.0701969862,-0.0499437191,-0.2336436063,0.1062823236,0.1187618524,-0.0412242934,-0.2293090522,-0.1664850116,0.069802843,0.0867944583,-0.3313302994,0.1300160289,0.1641416997,0.0661594048,-0.1091429666,0.0547850318,-0.1893745512,-0.0264874715,0.1374494433,0.3345857263,-0.1803767979,0.0246970616,0.044420246,-0.0005222475,-0.0113884984,0.2225162685,0.0441291928,-0.2677798569,-0.0962475091,-0.1068619564,-0.5135743618,-0.3039865196,-0.1129447892,-0.3393259943,-0.4211431146,0.2222638577,-0.4232880473,-0.1412899494,-0.0298155472,0.5755088925,0.275249064,-0.1636072695,0.2179998308,0.3787249327,-0.2972567379,0.1781837791,-0.1884079576,0.087363176,-0.179817751,0.4388460517,0.1041527614,-0.4247675538,0.2150660455,0.0994253233,0.0897854567,0.0637221336,-0.3770472109,-0.0550754666,0.3408403993,0.2761136889,-0.0871317983,0.0847101957,0.3560982049,-0.2921209633,-0.2050967067,0.059205655,-0.1236091405,-0.1076332703,-0.1479506642,0.0917600542,-0.0937497392,-0.058144372,0.199640885,0.0405034423,0.1504989862,-0.3299962282,0.0449836962,-0.0093830032,-0.1618902683,-0.2245410234,0.447907865,0.3724362254,-0.7759336829,-0.3368355334,0.0550395437,0.1556860805,-0.0823373571,0.2427782863,-0.1244228259,0.1349112391,-0.270999074,0.4014906883,0.1220537797,-0.0901549235,-0.1708913743,0.0448942482,0.0930247009,0.0285857841,0.2380019128,-0.0986707658,0.3254885972,0.1497478038,0.3354112506,-0.0001988137,0.0043552821,-0.3671826124,-0.1630478948,-0.3921634853,0.0697727874,0.0956226438,-0.1325723231,-0.0251712631,-0.1350774169,-0.4143718481,0.0810390562,-0.2481820732,-0.048055999,0.0244530439,0.2674900591,0.3455544114,0.4671978056,0.0824592113,-0.1673379987,0.2564873695,0.1291600615,0.0423053987,-0.1225517914,-0.0157379657,-0.2794288397,-0.0127059286,-0.0496821292,-0.1818123311,-0.0790601149,0.0145893861,-0.0118765989,-0.0757844374,-0.33210814,0.1303437799,-0.257668376,0.2069558501,-0.5300257206,0.4520913363,-0.2802708149,-0.2228663564,0.0329993255,0.2488915771,0.14724347,-0.2128631026,0.1438020319,0.1014066562,-0.0003674161,-0.2585800588,0.1111174598,-0.1829434633,0.1620029062,-0.1446011513,0.0851277485,-0.1179838702,0.2549681664,0.0650633052,-0.2244727015,0.4591362774,-0.0234917179,0.3052208424,-0.202988565,-0.2119348198,0.1173565015,-0.0178667922,-0.0267293882,-0.3300513327,-0.1173002943,0.0439026356,-0.1680182666,-0.0541885011,0.0857990757,0.1069037989,-0.0282339547,0.0522947609,-0.1552650481,0.0112996986,-0.2901861072,0.0189379416,0.0887711123,-0.3720996976,0.4891066849,0.2894639075,-0.1515061259,-0.2688131332,-0.2140420377,-0.3313252926,0.0595771186,-0.0840370655,-0.2100378275,0.1587175876,-0.1157231256,0.0546810143,-0.224160701,0.0081050992,0.0759624913,0.157697618,-0.5991175175,0.0572370812,-0.3049148917,-0.036601983,0.0308595318,-0.1012304276,0.0415180661,-0.3487308025,0.1815759391,-0.1190846041,-0.0377943367,0.1511288434,-0.0626414046,0.2203113884,-0.0816774219,0.0488701053,-0.1951137781,-0.2207871825,-0.334548831,-0.0690884665,-0.0555741191,0.301159054,0.1825916618,-0.0411623158,-0.1298975945,-0.1396417469,-0.2877671123,0.3002024293,-0.1688946486,0.1524758488,-0.1127929464,0.1133968979,-0.1385278404,-0.1531698406,0.3262204826,-0.0357740298,-0.1949459612,-0.0704273805,0.0943008736,-0.0298892073,-0.0883331448,-0.4785193503,0.1902282983,-0.2006299049,-0.141656056,-0.1947053671,-0.0839786455,-0.4174040556,0.4110864103,0.0925738513,-0.0698832646,0.0942998752,-0.3801019788,0.069809489,0.3063396513,-0.0815023631,-0.2894349396,0.1465222538,0.0096548581,-0.2156970799,0.0390084982,-0.3196030557,-0.5543977618,0.0217474271,0.159510687,0.2703954875,0.4036834836,0.156766817,-0.1057035625,-0.0533162802,-0.1179203019,-0.4052333236,0.0328005999,-0.0535211489,0.2113653421,-0.1147602126,0.486948967,-0.3206720352,0.522379458,0.381057322,-0.3383180499,0.4276521206,-0.1127688363,0.5202954412,-0.2422707379,-0.1422119439,0.1987451464,0.0586947873,0.001760066,0.2437190711,-0.0958641395,0.1190660223,0.0689601153,0.3653145134,-0.0896220431,-0.2674026191,0.1314044446,0.0305519681,-0.0160448756,-0.0716383532,0.3272547722,-0.0920909792,0.1545764208,0.184921205,0.3969687521,0.3998638391,-0.0730620921,0.2712453902,-0.3327034116,-0.208106041,0.5384803414,0.0814219564,0.2804942727,-0.1965026259,-0.162313357,0.4398677647,0.1852532178,0.5423484445,-0.4256882668,-0.3764909804,0.3058630228,0.3216744661,-0.4337252676,-0.1476146132,-0.0224916711,-0.0453610644,-0.0149933649,0.7402037382,-0.4742542505,-0.0177627895,0.2606261671,0.1066690534,-0.1701066196,-0.4776918888,-0.5002132654,-0.144437477,-0.4203992486,-0.076269798,-0.2062743604,0.3078433573,0.1021929905,0.2367188185,-0.0812158808,0.0129864039,0.1501802206,0.3022860289,0.0446066484,0.0525975227,0.2101492435,0.4190895557,0.435888499,0.2720904946,0.4205085933,-0.1399782598,-0.3320228755,0.2088844031,-0.0006845412,0.1612280309,0.3638099432,0.029952392,0.2056963444,-0.1669259369,0.0103233308,-0.2124233246,-0.3417513669,0.0968279317,0.0627069622,0.071824193,-0.2635461092,0.4906757176,0.0296074599,-0.1465217918,-0.102447927,0.7082784772,-0.0038828347,0.9643495083,-0.0068482943,0.9512917995,-0.1551676542,-0.0376357213,0.440862745,-0.1820308715,0.4978231788,-0.1612353623,0.1090130359,-0.4382877946,-0.2174162567,-0.0792143121,-0.0454561971,0.1918792874,-0.0233830381,-0.2681635618,0.155336678,-0.102279231,-0.09495233,-0.0957573354,-0.0553626046,0.0645109415,-0.4010760784,-0.192283079,0.018740274,0.4626621306,0.3580811322,0.0256027617,0.0297266431,-0.3238193095,-0.0115194572,-0.3098014891,0.0463716164,0.3120693862,0.1080762818,0.3823556602,-0.2416798472,0.2194236368,0.1066802591,0.6348162889,-0.4909479618,-0.1986659616,0.1314476579,-0.2250174284,-0.2495675534,-0.141205743,-0.3402716219,0.3036342263,0.2455486804,-0.415851146,0.2610159218,0.1105015427,-0.1313995272,-0.1638645977,0.1526336819,0.0098769879,-0.4759309292,-0.1079159677,-0.0234594904,0.3508502543,-0.0961570293,0.0572825,0.1041795164,-0.0218206719,0.1986280531,0.0722786561,-0.1242037863,-0.0319024473,0.4855908453,-0.1633221656,0.1417820752,0.3344900906,0.098691158,-0.4170288146,-0.0284618028,-0.5361010432,0.5242940187,-0.3646396697,-0.1897098273,0.3145854473,-0.1721122265,0.0639139116,0.0145346383,0.3819753528,-0.2045632601,-0.2862965763,-0.2601949275,-0.0206423458,0.3878620863,-0.0625405237,0.1380119771,-0.1181082278,0.0041945935,-0.1206752434,0.0701413378,-0.2063860595,0.1769719571,-0.069777295,0.1305287331,-0.0349912643,-0.2045271248,-0.036428526,-0.2391808033,0.0181268286,0.0927694663,-0.0183125343,-0.0059972,0.0139023671,0.1815475971,0.0901795998,0.111163266,-0.4262259305,-0.0474817045,-0.0676278099,-0.3948301077,0.3541818857,0.1756357551,-0.1166958138,-0.1651634127,0.0495953187,0.0697007403,-0.0563950986,0.2638918757,0.1273393184,0.5054332018,0.2831273079,-0.1307985187,-0.4669467509,0.1515465379,0.0683479905,0.399682343,0.1805761307,-0.0922706798,0.3608169258,-0.2189326137,0.0289396103,-0.0035550918,0.1313319057,0.4595068097,-0.0530051738,0.0577949174,0.1733681262,0.0563398674,-0.1533013731,0.0975931287,0.3312352002,-0.0935293213,-0.0380810611,0.2321978211,0.2101738751,0.0921623334,-0.1772083044,0.0337090567,0.2256146371,0.187129125,-0.3880457878,-0.0120700337,0.1033218428,0.261035502,0.490157187,0.1079503968,0.3249964714,0.0446728133,-0.108773388,0.0432824679,0.0471098721,0.4198775887,-0.1686375588,-0.613433063,0.5157346129,0.2900369763,0.4744550288,-0.3256697655,0.036297854,0.2324466556,0.0131480377,-0.0064875465,-0.1716701686,-0.0322825387,-0.1626160443,-0.4494574368,0.1296900958,-0.0375101864,-0.1559707373,-0.0305806957,0.0236073658,-0.0976384506,-0.0056416085,-0.0976480544,-0.3179889023,-0.1804525703,-0.3293999135,0.2765711248,0.2327003181,0.0657433346,-0.0451880954,-0.113355644,-0.0320089124,0.4509784579,0.0384962112,0.502271235,0.1259166747,0.0164805688,0.0336955562,0.0694676116,0.0463824756,-0.0118554747,0.2121435702,-0.1274953038,0.24947837,0.2408227175,0.0902942121,-0.103824839,-0.3525876403,-0.0541390032,0.5762381554,0.0636182949,-0.1043975055,-0.3979019523,0.052312769,-0.0834136009,0.013452963,-0.1997368634,0.3871743083,-0.0032992275,-0.1049189568,0.0381019562,0.3390271366,0.0234479345,-0.1851810813,0.2432608455,0.8182145357,-0.2200938016,-0.2327183634,0.1255899072,-0.5277732611,0.2583816946,-0.1381704807,-0.1659809649,0.2066939622,0.6418519616,0.492626667,-0.0448903851,-0.0263894629,-0.040765468,0.2749048471,0.4152514637,0.0423217453,-0.3399311602,-0.0948812515,-0.2120658457,-0.1627363414,-0.0597314388,0.1098054722,-0.0464062952,-0.0607466288,0.0680687875,0.0699819252,-0.1350307316,-0.3679941595,0.6232010126,0.0587596372,0.3626820147,-0.1151081622,0.0254577026,-0.2804363668,-0.1538053751,-0.0043473928,0.2639605105,-0.1937406659,0.1482799649,-0.159425199,-0.1516765058,-0.2438272238,0.2302541733,0.0425608307,-0.1830163598,-0.2047594935,0.1614382267,-0.2503418922,0.0224260427,-0.0029815896,0.0231916513,-0.0504194461,0.2751443982,0.0746593103,-0.2265358418,0.617156446,-0.3164285421,0.090938352,0.1647376567,0.2101889998,0.4974494874,0.0436607823,-0.8163808584,-0.0133223897,0.0778879672,-0.1394903213,-0.2482402176,-0.0961590931,0.2267888784,0.0229928456,-0.284794867,-0.2223566175,0.0500374362,0.0527016595,-0.0666384771,-0.2481245846]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2026","title":"KeyError on using map after renaming a column","comments":"Hi,\r\n\r\nActually, the error occurs due to these two lines:\r\n```python\r\nraw_dataset.set_format('torch',columns=['img','label'])\r\nraw_dataset = raw_dataset.rename_column('img','image')\r\n```\r\n`Dataset.rename_column` doesn't update the `_format_columns` attribute, previously defined by `Dataset.set_format`, with a new column name which is why this new column is missing in the output.","body":"Hi,\r\n\r\nI'm trying to use `cifar10` dataset. I want to rename the `img` feature to `image` in order to make it consistent with `mnist`, which I'm also planning to use. By doing this, I was trying to avoid modifying `prepare_train_features` function.\r\n\r\nHere is what I try:\r\n\r\n```python\r\ntransform = Compose([ToPILImage(),ToTensor(),Normalize([0.0,0.0,0.0],[1.0,1.0,1.0])])\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n print(examples)\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n else:\r\n images.append(examples[\"image\"][example_idx].permute(2,0,1))\r\n labels.append(examples[\"label\"][example_idx])\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('cifar10')\r\nraw_dataset.set_format('torch',columns=['img','label'])\r\nraw_dataset = raw_dataset.rename_column('img','image')\r\n\r\nfeatures = datasets.Features({\r\n \"image\": datasets.Array3D(shape=(3,32,32),dtype=\"float32\"),\r\n \"label\": datasets.features.ClassLabel(names=[\r\n \"airplane\",\r\n \"automobile\",\r\n \"bird\",\r\n \"cat\",\r\n \"deer\",\r\n \"dog\",\r\n \"frog\",\r\n \"horse\",\r\n \"ship\",\r\n \"truck\",\r\n ]),\r\n })\r\ntrain_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n```\r\nThe error:\r\n```python\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n in ()\r\n 14 ]),\r\n 15 })\r\n---> 16 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1287 test_inputs = self[:2] if batched else self[0]\r\n 1288 test_indices = [0, 1] if batched else 0\r\n-> 1289 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1290 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1291 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1258 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1259 processed_inputs = (\r\n-> 1260 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1261 )\r\n 1262 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n in prepare_features(examples)\r\n 3 labels = []\r\n 4 print(examples)\r\n----> 5 for example_idx, example in enumerate(examples[\"image\"]):\r\n 6 if transform is not None:\r\n 7 images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n\r\nKeyError: 'image'\r\n```\r\n\r\nThe print statement inside returns this:\r\n```python\r\n{'label': tensor([6, 9])}\r\n```\r\nApparently, both `img` and `image` do not exist after renaming. \r\n\r\nNote that this code works fine with `img` everywhere.\r\n\r\nNotebook: https:\/\/colab.research.google.com\/drive\/1SzESAlz3BnVYrgQeJ838vbMp1OsukiA2?usp=sharing\r\n\r\n","comment_length":42,"text":"KeyError on using map after renaming a column\nHi,\r\n\r\nI'm trying to use `cifar10` dataset. I want to rename the `img` feature to `image` in order to make it consistent with `mnist`, which I'm also planning to use. By doing this, I was trying to avoid modifying `prepare_train_features` function.\r\n\r\nHere is what I try:\r\n\r\n```python\r\ntransform = Compose([ToPILImage(),ToTensor(),Normalize([0.0,0.0,0.0],[1.0,1.0,1.0])])\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n print(examples)\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n else:\r\n images.append(examples[\"image\"][example_idx].permute(2,0,1))\r\n labels.append(examples[\"label\"][example_idx])\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('cifar10')\r\nraw_dataset.set_format('torch',columns=['img','label'])\r\nraw_dataset = raw_dataset.rename_column('img','image')\r\n\r\nfeatures = datasets.Features({\r\n \"image\": datasets.Array3D(shape=(3,32,32),dtype=\"float32\"),\r\n \"label\": datasets.features.ClassLabel(names=[\r\n \"airplane\",\r\n \"automobile\",\r\n \"bird\",\r\n \"cat\",\r\n \"deer\",\r\n \"dog\",\r\n \"frog\",\r\n \"horse\",\r\n \"ship\",\r\n \"truck\",\r\n ]),\r\n })\r\ntrain_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n```\r\nThe error:\r\n```python\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n in ()\r\n 14 ]),\r\n 15 })\r\n---> 16 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1287 test_inputs = self[:2] if batched else self[0]\r\n 1288 test_indices = [0, 1] if batched else 0\r\n-> 1289 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1290 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1291 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1258 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1259 processed_inputs = (\r\n-> 1260 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1261 )\r\n 1262 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n in prepare_features(examples)\r\n 3 labels = []\r\n 4 print(examples)\r\n----> 5 for example_idx, example in enumerate(examples[\"image\"]):\r\n 6 if transform is not None:\r\n 7 images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n\r\nKeyError: 'image'\r\n```\r\n\r\nThe print statement inside returns this:\r\n```python\r\n{'label': tensor([6, 9])}\r\n```\r\nApparently, both `img` and `image` do not exist after renaming. \r\n\r\nNote that this code works fine with `img` everywhere.\r\n\r\nNotebook: https:\/\/colab.research.google.com\/drive\/1SzESAlz3BnVYrgQeJ838vbMp1OsukiA2?usp=sharing\r\n\r\n\nHi,\r\n\r\nActually, the error occurs due to these two lines:\r\n```python\r\nraw_dataset.set_format('torch',columns=['img','label'])\r\nraw_dataset = raw_dataset.rename_column('img','image')\r\n```\r\n`Dataset.rename_column` doesn't update the `_format_columns` attribute, previously defined by `Dataset.set_format`, with a new column name which is why this new column is missing in the output.","embeddings":[0.0343904048,0.0061999829,-0.0635800064,-0.334228158,0.4808440804,0.2616392076,0.6042662859,0.237331897,0.1370912194,0.0865909159,0.054035522,0.5270302892,-0.1558949649,0.2994630933,-0.1945005655,-0.0881165639,0.4133820236,0.0935034528,-0.1156333089,0.1972025782,-0.4435485899,0.0497060306,-0.3014574051,0.2807840407,-0.3147513866,-0.1778963804,0.054533463,-0.1938546598,-0.2473179102,-0.274669081,-0.0492155105,0.0971030518,-0.0223661978,0.6005281806,-0.0001189421,0.0826987103,0.188707307,-0.0865515545,0.082985498,-0.2960751951,-0.1819153279,-0.0115416646,-0.2230614275,-0.5142738819,0.0406857617,-0.0636158809,-0.077172786,-0.2671773434,0.1024493352,0.2466998845,0.1638297886,-0.1138828397,0.2209231108,-0.0220486131,0.0439468622,0.2382541597,-0.1926510036,0.0769321546,0.0491049998,-0.3582004011,0.2290986776,0.693173945,-0.2142525613,-0.0826261491,0.2492942363,0.1485666484,0.4132107496,-0.3595764339,0.3614650071,-0.1366176903,0.2379055619,-0.1512385756,-0.1043806374,-0.054232765,-0.0018236447,-0.2537844181,-0.0232361704,-0.3196378648,0.0776960477,-0.2296992242,-0.3305271864,-0.1850074828,0.1285365671,0.2078439891,0.1231135502,-0.097198315,-0.0296082962,0.4492817819,0.1049034894,-0.322670877,-0.1246059462,0.0269131735,0.098084107,0.2211732715,-0.2241782248,-0.0697751865,0.105466947,-0.0227153953,-0.2865684628,-0.6728146672,0.017100919,-0.1154988408,0.1804634184,0.0785013214,0.106362313,0.2072906345,-0.087925978,0.3839033544,0.0437416993,0.0499913432,-0.3903314173,-0.2533300221,0.1843829304,-0.3990157247,0.3597591519,0.2472612262,0.3776059747,0.0936776847,0.4118515253,-0.0127613107,0.1394363791,-0.0965100452,-0.0235801451,0.3582357764,0.2444261461,-0.0504651479,0.1044974774,0.1474120915,-0.0377255827,0.2744560242,-0.0399699099,0.1520807594,-0.5921468139,-0.10537678,-0.1591845006,0.0644563138,0.0273747109,-0.1856066436,0.1200411022,-0.1029791012,-0.017414419,-0.062561661,0.4089117348,0.1914608777,-0.3518828154,0.1143235788,0.2287517935,-0.2352549285,-0.1231607944,0.2178232223,-0.6010237336,-0.1866197884,-0.1935406029,0.0911867172,0.2030768991,0.1996438056,-0.2752969265,-0.1346438378,0.5062403679,-0.4040708244,0.082678318,-0.2427093983,-0.2949000895,-0.2364552915,-0.1690340787,-0.0339538306,-0.0462761037,-0.0256844573,-0.1759432405,0.3576537967,0.0155100189,-0.0486620031,-0.1009822637,0.2132537961,-0.0342827775,0.1602661163,0.4271647036,-0.3145627677,-0.2063587606,-0.0152603928,-0.2203991115,-0.1616605967,-0.0920220613,0.093030706,0.2039043605,-0.0727139935,0.4399832785,-0.154413715,-0.0898718163,0.178547591,0.1311589926,-0.0462149791,0.1775305122,-0.0791230053,0.1281124204,0.208673045,0.1111345366,0.0116959661,-0.0227944031,-0.2068924159,0.0939341784,0.0991031379,0.2192526907,-0.0245531406,0.0360569097,0.0803689882,-0.3255301416,0.167881459,0.1803426594,0.141612947,-0.2925769389,-0.1196023375,-0.2550840676,-0.1336407512,-0.176066801,0.0316427574,0.0552044027,-0.054000631,-0.0344205536,-0.1491251588,-0.1648832858,-0.1527555883,0.0497404523,0.1093318611,-0.3486706614,-0.0184663273,-0.2110998482,-0.1441620439,-0.3850237131,0.0132771088,0.1881341487,-0.0540964752,-0.2381078899,0.3495651782,0.1393090487,0.0440969206,-0.3154114485,0.1756070554,0.0911011919,-0.0334517248,-0.0013047188,0.1042696312,0.1814650595,-0.0348044224,0.1123570204,0.2557912469,-0.2581357658,0.3806613088,-0.2395886481,0.1288557351,-0.0841588005,-0.1766913235,-0.2373664528,-0.393781811,-0.143782109,-0.2950774133,-0.2323311865,-0.0712196752,0.3053756952,-0.2079800516,0.5871576667,-0.0409827009,0.0240916926,-0.0057099815,0.0615631565,0.289617002,0.0114168618,0.1064895242,0.297829479,-0.009614123,-0.2233593762,0.2672612965,0.1708517224,0.2767239213,0.2040939629,0.2167057693,0.244703874,0.0480267294,0.1485542208,-0.0613893084,0.0164967366,-0.2999012768,0.3099987805,0.2844228446,-0.2336353511,0.0998230577,-0.0660420731,0.0700871125,0.2823909223,-0.1527176499,0.0562228933,-0.2700430751,-0.094888635,0.3865874112,0.0920764729,0.4332132339,-0.1842424124,-0.3789016604,0.2417967469,-0.35465464,-0.0365538523,-0.4494333863,0.030626379,-0.0325060226,0.0856916159,-0.4606234431,0.0817554891,-0.1647819877,-0.1355617791,-0.1211330071,-0.5227620006,0.1218558326,-0.4426466227,0.0148037989,0.2468884289,0.2771109343,-0.2926217616,-0.1771504432,0.4542585015,-0.2178657502,-0.1494309604,0.125867188,0.0121634537,-0.039333865,-0.1180340871,-0.1904156059,-0.0265278462,-0.1234948337,0.1169085056,-0.2646969259,0.0163139403,0.5068838596,0.2759175897,-0.2442613244,0.1430361718,-0.0489320122,-0.1584293842,-0.3822782338,0.208315134,-0.0868597105,0.0838408098,0.0047291089,-0.1143030971,-0.0037284025,0.0410409383,-0.1813217849,-0.292372793,-0.0811148286,0.252065897,0.1551319957,0.1015874445,0.2861323655,0.4163930118,-0.1233831346,0.0155468863,-0.2297620475,-0.0407741778,0.4496167004,0.2750443816,0.2839065492,0.4489243031,-0.0161301196,0.2931136191,-0.1320103854,-0.1058140993,0.4552730322,-0.3281630576,0.097139366,-0.427151382,-0.4983298779,0.021225499,-0.3456521928,-0.2692156136,-0.252612561,-0.1974028945,-0.2074296474,-0.1250825375,0.2784183919,-0.4862356186,-0.3314012587,0.1603580266,0.0705963597,0.30742383,0.0487649925,-0.0192793254,-0.4495798647,-0.1688798666,0.0980947465,0.0244836956,0.055906225,-0.3355075121,-0.6632701159,-0.2365220785,-0.3934516907,0.4276641011,0.1131599844,0.1529776901,-0.1410124451,-0.0606670156,-0.0556474812,0.0688659251,0.7746023536,-0.4415769279,-0.1148998216,0.3010852635,-0.0476227738,-0.0988456905,-0.2432980686,-0.0672431141,0.2733609378,0.0472895429,0.8124082088,0.0474174544,0.0843919963,0.214545846,0.0629557222,-0.0658334941,0.0286184251,-0.406054467,-0.2894008756,-0.5907236934,-0.0206708051,0.0572946109,0.2173590213,0.3088783324,0.1889963001,0.1960792691,-0.1945684999,0.0154501591,0.2664080858,0.1986875087,0.3779762089,0.0560513809,-0.0483673327,0.1246331483,0.0030192293,0.1406399608,-0.2207594961,-0.2776945531,-0.0573592074,0.0422726683,0.3460413516,0.2213218659,0.0628878772,0.0959237814,0.1461020708,0.2581580877,-0.2759142816,0.2963343561,0.4933303595,0.1403307617,-0.4810526073,-0.3202238679,0.2370316833,0.2185061872,-0.1690102518,0.38164258,0.0929525048,-0.3298314214,0.5882653594,0.0854706466,0.8382591009,-0.3112327456,0.0554842912,-0.0850671604,0.0784885734,0.2867376208,0.0625677705,0.2233817428,-0.1989947259,-0.4336817265,0.0294411965,-0.0587247387,0.347617805,0.0871918276,-0.1870900691,0.2699613571,-0.2224864066,0.584843874,-0.2539536059,-0.2409918308,0.2116029114,-0.3743119836,0.1243179068,0.056649372,-0.0686615184,0.0904513001,0.0100760283,0.1401854753,-0.1021842211,-0.3583537638,0.0675272495,-0.0539399981,-0.2007298023,0.2325115949,-0.2137070447,-0.2194651663,-0.0675614402,0.3379043937,0.3049491644,0.2256917655,0.0769784302,0.2345013022,0.4431695342,0.2091673464,-0.0408834182,-0.1617862284,-0.0470286757,0.1401213855,-0.2852287292,0.1369479746,-0.0141862789,-0.2416308075,-0.364243418,0.0415634848,0.4323446155,-0.3393744826,-0.2499048561,-0.1570414305,-0.0112912552,-0.1000931934,0.0678598732,-0.3800917268,-0.1897207946,0.4002565742,-0.1145981029,-0.4112877548,0.1257440895,0.363469094,0.0822052211,0.0131317768,0.5031767488,-0.1995394677,0.0113480333,-0.08634267,0.0982044041,0.2158655226,-0.0867505968,0.0297973417,0.3238884807,-0.0309048165,-0.1596030295,0.4304632843,0.056714043,0.2015253901,0.0784534812,-0.2432671636,-0.6696433425,0.4877867103,0.1031506881,0.2492165565,0.1051922888,0.3467668593,-0.294855684,-0.011246155,-0.2399551123,0.0649206117,-0.0438210703,0.3154606819,0.4421718121,0.0253377911,-0.0896078646,-0.2898142636,0.2092776299,0.264129132,-0.0850462019,-0.213859275,-0.0167676937,0.1183085814,0.1053171828,-0.1505651921,-0.0243897997,-0.3231567144,-0.0274435021,-0.2703390718,-0.0828690007,0.3389027417,0.0339184143,0.1923668236,-0.5534645915,-0.0861907825,0.3065199554,0.1393099278,-0.1251810491,-0.0567402318,0.2653055489,0.1776153892,-0.050407812,0.0970639437,-0.2915789485,-0.068407774,-0.1543523222,0.1156141385,0.3472285569,-0.1888154894,0.1339828819,0.0561051331,0.2796280384,0.1310018599,-0.2598958313,-0.1948141307,0.3021291494,0.0988662317,-0.3267253339,-0.1829232723,0.5215923786,-0.0156410933,0.2411123663,0.2352423817,-0.0485072844,0.1316879392,-0.2601729333,-0.0177907143,0.1529346704,-0.0737773851,0.2469815612,0.6254125237,-0.1150746867,0.2715166807,0.2926054597,0.373080641,-0.1249766499,0.6644849181,0.0635503903,0.2042205036,0.1586616784,0.2688653469,-0.1144806743,-0.0899507031,0.5314686894,0.1346583664,-0.2235026807,0.2969570458,0.1907867789,-0.2590565085,-0.0018368585,-0.225257501,-0.0775448382,0.2807589173,-0.2177532315,-0.2554299235,-0.1515018791,-0.036063239,0.0017208962,0.0805628821,-0.0572090372,0.0685864314,0.6512500644,0.026028268,0.0510043502,0.0960447788,-0.5209681988,0.1352971196,0.3933674097,-0.1334355921,-0.017999975,0.0571532957,0.1642050743,0.2200044841,0.0155200986,0.3567886353,0.1692755967,-0.2766989768,0.1629797667,-0.0195907224,-0.1051777452,0.2641951144,0.0439048819,0.3267834783,0.008745905,0.1000831351,0.0357460454,-0.1855373085,-0.1050797775,0.148779586,0.1994472444,0.1967056245,-0.066673778,0.0776693895,-0.1384180486,-0.0612892099,0.1662924588,-0.2106761187,-0.2168651372,0.4741016328,-0.1182754636,0.1831553429,0.1314244568,0.0622664802,0.1147598624,0.1901857108,0.1068619639,-0.1823746413,-0.5617814064,-0.0506489016,-0.3618877232,-0.1454607099,-0.0483252704,0.0958933979,0.0001837238,-0.2958847284,0.2807190716,0.1011902243,0.4350568354,-0.4830625653,-0.0566600636,0.5606931448,0.0539841652,-0.0601437464,0.0202838592,0.2473428994,0.1078898311,-0.3471124768,0.0970269293,0.0586616993,-0.0088258684,-0.339279294,-0.1747089773,-0.2541410327,-0.2222277373,0.4715786278,0.1747033447,0.4249306321,-0.1517661214,-0.072775431,0.1637441069,-0.1676822305,-0.016834598,0.2095226794,0.2309502363,0.3883229494,-0.0711031109,0.128017351,-0.3970904052,0.0074610743,0.0302021448,-0.1853101104,-0.3765938282,0.186921224,0.2549928129,0.0408338793,-0.1853446364,0.3607902229,0.0388966613,0.5180019736,-0.0680310279,-0.3690660894,0.7985036969,-0.3579202294,-0.6600227952,0.0823596269,-0.0900159925,-0.1427418143,0.0412078798,-0.2388151884,-0.0630643219,0.073640883,-0.0334531926,-0.3466158807,0.0901280493,-0.1645772755,0.079836078,0.0364476554,0.4860982895,-0.3197670579,-0.0945576355,-0.2467267662,-0.0632384643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2026","title":"KeyError on using map after renaming a column","comments":"Hi @mariosasko,\n\nThanks for opening a PR on this :)\nWhy does the old name also disappear?","body":"Hi,\r\n\r\nI'm trying to use `cifar10` dataset. I want to rename the `img` feature to `image` in order to make it consistent with `mnist`, which I'm also planning to use. By doing this, I was trying to avoid modifying `prepare_train_features` function.\r\n\r\nHere is what I try:\r\n\r\n```python\r\ntransform = Compose([ToPILImage(),ToTensor(),Normalize([0.0,0.0,0.0],[1.0,1.0,1.0])])\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n print(examples)\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n else:\r\n images.append(examples[\"image\"][example_idx].permute(2,0,1))\r\n labels.append(examples[\"label\"][example_idx])\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('cifar10')\r\nraw_dataset.set_format('torch',columns=['img','label'])\r\nraw_dataset = raw_dataset.rename_column('img','image')\r\n\r\nfeatures = datasets.Features({\r\n \"image\": datasets.Array3D(shape=(3,32,32),dtype=\"float32\"),\r\n \"label\": datasets.features.ClassLabel(names=[\r\n \"airplane\",\r\n \"automobile\",\r\n \"bird\",\r\n \"cat\",\r\n \"deer\",\r\n \"dog\",\r\n \"frog\",\r\n \"horse\",\r\n \"ship\",\r\n \"truck\",\r\n ]),\r\n })\r\ntrain_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n```\r\nThe error:\r\n```python\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n in ()\r\n 14 ]),\r\n 15 })\r\n---> 16 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1287 test_inputs = self[:2] if batched else self[0]\r\n 1288 test_indices = [0, 1] if batched else 0\r\n-> 1289 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1290 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1291 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1258 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1259 processed_inputs = (\r\n-> 1260 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1261 )\r\n 1262 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n in prepare_features(examples)\r\n 3 labels = []\r\n 4 print(examples)\r\n----> 5 for example_idx, example in enumerate(examples[\"image\"]):\r\n 6 if transform is not None:\r\n 7 images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n\r\nKeyError: 'image'\r\n```\r\n\r\nThe print statement inside returns this:\r\n```python\r\n{'label': tensor([6, 9])}\r\n```\r\nApparently, both `img` and `image` do not exist after renaming. \r\n\r\nNote that this code works fine with `img` everywhere.\r\n\r\nNotebook: https:\/\/colab.research.google.com\/drive\/1SzESAlz3BnVYrgQeJ838vbMp1OsukiA2?usp=sharing\r\n\r\n","comment_length":17,"text":"KeyError on using map after renaming a column\nHi,\r\n\r\nI'm trying to use `cifar10` dataset. I want to rename the `img` feature to `image` in order to make it consistent with `mnist`, which I'm also planning to use. By doing this, I was trying to avoid modifying `prepare_train_features` function.\r\n\r\nHere is what I try:\r\n\r\n```python\r\ntransform = Compose([ToPILImage(),ToTensor(),Normalize([0.0,0.0,0.0],[1.0,1.0,1.0])])\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n print(examples)\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n else:\r\n images.append(examples[\"image\"][example_idx].permute(2,0,1))\r\n labels.append(examples[\"label\"][example_idx])\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('cifar10')\r\nraw_dataset.set_format('torch',columns=['img','label'])\r\nraw_dataset = raw_dataset.rename_column('img','image')\r\n\r\nfeatures = datasets.Features({\r\n \"image\": datasets.Array3D(shape=(3,32,32),dtype=\"float32\"),\r\n \"label\": datasets.features.ClassLabel(names=[\r\n \"airplane\",\r\n \"automobile\",\r\n \"bird\",\r\n \"cat\",\r\n \"deer\",\r\n \"dog\",\r\n \"frog\",\r\n \"horse\",\r\n \"ship\",\r\n \"truck\",\r\n ]),\r\n })\r\ntrain_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n```\r\nThe error:\r\n```python\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n in ()\r\n 14 ]),\r\n 15 })\r\n---> 16 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1287 test_inputs = self[:2] if batched else self[0]\r\n 1288 test_indices = [0, 1] if batched else 0\r\n-> 1289 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1290 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1291 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1258 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1259 processed_inputs = (\r\n-> 1260 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1261 )\r\n 1262 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n in prepare_features(examples)\r\n 3 labels = []\r\n 4 print(examples)\r\n----> 5 for example_idx, example in enumerate(examples[\"image\"]):\r\n 6 if transform is not None:\r\n 7 images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n\r\nKeyError: 'image'\r\n```\r\n\r\nThe print statement inside returns this:\r\n```python\r\n{'label': tensor([6, 9])}\r\n```\r\nApparently, both `img` and `image` do not exist after renaming. \r\n\r\nNote that this code works fine with `img` everywhere.\r\n\r\nNotebook: https:\/\/colab.research.google.com\/drive\/1SzESAlz3BnVYrgQeJ838vbMp1OsukiA2?usp=sharing\r\n\r\n\nHi @mariosasko,\n\nThanks for opening a PR on this :)\nWhy does the old name also disappear?","embeddings":[0.0343904048,0.0061999829,-0.0635800064,-0.334228158,0.4808440804,0.2616392076,0.6042662859,0.237331897,0.1370912194,0.0865909159,0.054035522,0.5270302892,-0.1558949649,0.2994630933,-0.1945005655,-0.0881165639,0.4133820236,0.0935034528,-0.1156333089,0.1972025782,-0.4435485899,0.0497060306,-0.3014574051,0.2807840407,-0.3147513866,-0.1778963804,0.054533463,-0.1938546598,-0.2473179102,-0.274669081,-0.0492155105,0.0971030518,-0.0223661978,0.6005281806,-0.0001189421,0.0826987103,0.188707307,-0.0865515545,0.082985498,-0.2960751951,-0.1819153279,-0.0115416646,-0.2230614275,-0.5142738819,0.0406857617,-0.0636158809,-0.077172786,-0.2671773434,0.1024493352,0.2466998845,0.1638297886,-0.1138828397,0.2209231108,-0.0220486131,0.0439468622,0.2382541597,-0.1926510036,0.0769321546,0.0491049998,-0.3582004011,0.2290986776,0.693173945,-0.2142525613,-0.0826261491,0.2492942363,0.1485666484,0.4132107496,-0.3595764339,0.3614650071,-0.1366176903,0.2379055619,-0.1512385756,-0.1043806374,-0.054232765,-0.0018236447,-0.2537844181,-0.0232361704,-0.3196378648,0.0776960477,-0.2296992242,-0.3305271864,-0.1850074828,0.1285365671,0.2078439891,0.1231135502,-0.097198315,-0.0296082962,0.4492817819,0.1049034894,-0.322670877,-0.1246059462,0.0269131735,0.098084107,0.2211732715,-0.2241782248,-0.0697751865,0.105466947,-0.0227153953,-0.2865684628,-0.6728146672,0.017100919,-0.1154988408,0.1804634184,0.0785013214,0.106362313,0.2072906345,-0.087925978,0.3839033544,0.0437416993,0.0499913432,-0.3903314173,-0.2533300221,0.1843829304,-0.3990157247,0.3597591519,0.2472612262,0.3776059747,0.0936776847,0.4118515253,-0.0127613107,0.1394363791,-0.0965100452,-0.0235801451,0.3582357764,0.2444261461,-0.0504651479,0.1044974774,0.1474120915,-0.0377255827,0.2744560242,-0.0399699099,0.1520807594,-0.5921468139,-0.10537678,-0.1591845006,0.0644563138,0.0273747109,-0.1856066436,0.1200411022,-0.1029791012,-0.017414419,-0.062561661,0.4089117348,0.1914608777,-0.3518828154,0.1143235788,0.2287517935,-0.2352549285,-0.1231607944,0.2178232223,-0.6010237336,-0.1866197884,-0.1935406029,0.0911867172,0.2030768991,0.1996438056,-0.2752969265,-0.1346438378,0.5062403679,-0.4040708244,0.082678318,-0.2427093983,-0.2949000895,-0.2364552915,-0.1690340787,-0.0339538306,-0.0462761037,-0.0256844573,-0.1759432405,0.3576537967,0.0155100189,-0.0486620031,-0.1009822637,0.2132537961,-0.0342827775,0.1602661163,0.4271647036,-0.3145627677,-0.2063587606,-0.0152603928,-0.2203991115,-0.1616605967,-0.0920220613,0.093030706,0.2039043605,-0.0727139935,0.4399832785,-0.154413715,-0.0898718163,0.178547591,0.1311589926,-0.0462149791,0.1775305122,-0.0791230053,0.1281124204,0.208673045,0.1111345366,0.0116959661,-0.0227944031,-0.2068924159,0.0939341784,0.0991031379,0.2192526907,-0.0245531406,0.0360569097,0.0803689882,-0.3255301416,0.167881459,0.1803426594,0.141612947,-0.2925769389,-0.1196023375,-0.2550840676,-0.1336407512,-0.176066801,0.0316427574,0.0552044027,-0.054000631,-0.0344205536,-0.1491251588,-0.1648832858,-0.1527555883,0.0497404523,0.1093318611,-0.3486706614,-0.0184663273,-0.2110998482,-0.1441620439,-0.3850237131,0.0132771088,0.1881341487,-0.0540964752,-0.2381078899,0.3495651782,0.1393090487,0.0440969206,-0.3154114485,0.1756070554,0.0911011919,-0.0334517248,-0.0013047188,0.1042696312,0.1814650595,-0.0348044224,0.1123570204,0.2557912469,-0.2581357658,0.3806613088,-0.2395886481,0.1288557351,-0.0841588005,-0.1766913235,-0.2373664528,-0.393781811,-0.143782109,-0.2950774133,-0.2323311865,-0.0712196752,0.3053756952,-0.2079800516,0.5871576667,-0.0409827009,0.0240916926,-0.0057099815,0.0615631565,0.289617002,0.0114168618,0.1064895242,0.297829479,-0.009614123,-0.2233593762,0.2672612965,0.1708517224,0.2767239213,0.2040939629,0.2167057693,0.244703874,0.0480267294,0.1485542208,-0.0613893084,0.0164967366,-0.2999012768,0.3099987805,0.2844228446,-0.2336353511,0.0998230577,-0.0660420731,0.0700871125,0.2823909223,-0.1527176499,0.0562228933,-0.2700430751,-0.094888635,0.3865874112,0.0920764729,0.4332132339,-0.1842424124,-0.3789016604,0.2417967469,-0.35465464,-0.0365538523,-0.4494333863,0.030626379,-0.0325060226,0.0856916159,-0.4606234431,0.0817554891,-0.1647819877,-0.1355617791,-0.1211330071,-0.5227620006,0.1218558326,-0.4426466227,0.0148037989,0.2468884289,0.2771109343,-0.2926217616,-0.1771504432,0.4542585015,-0.2178657502,-0.1494309604,0.125867188,0.0121634537,-0.039333865,-0.1180340871,-0.1904156059,-0.0265278462,-0.1234948337,0.1169085056,-0.2646969259,0.0163139403,0.5068838596,0.2759175897,-0.2442613244,0.1430361718,-0.0489320122,-0.1584293842,-0.3822782338,0.208315134,-0.0868597105,0.0838408098,0.0047291089,-0.1143030971,-0.0037284025,0.0410409383,-0.1813217849,-0.292372793,-0.0811148286,0.252065897,0.1551319957,0.1015874445,0.2861323655,0.4163930118,-0.1233831346,0.0155468863,-0.2297620475,-0.0407741778,0.4496167004,0.2750443816,0.2839065492,0.4489243031,-0.0161301196,0.2931136191,-0.1320103854,-0.1058140993,0.4552730322,-0.3281630576,0.097139366,-0.427151382,-0.4983298779,0.021225499,-0.3456521928,-0.2692156136,-0.252612561,-0.1974028945,-0.2074296474,-0.1250825375,0.2784183919,-0.4862356186,-0.3314012587,0.1603580266,0.0705963597,0.30742383,0.0487649925,-0.0192793254,-0.4495798647,-0.1688798666,0.0980947465,0.0244836956,0.055906225,-0.3355075121,-0.6632701159,-0.2365220785,-0.3934516907,0.4276641011,0.1131599844,0.1529776901,-0.1410124451,-0.0606670156,-0.0556474812,0.0688659251,0.7746023536,-0.4415769279,-0.1148998216,0.3010852635,-0.0476227738,-0.0988456905,-0.2432980686,-0.0672431141,0.2733609378,0.0472895429,0.8124082088,0.0474174544,0.0843919963,0.214545846,0.0629557222,-0.0658334941,0.0286184251,-0.406054467,-0.2894008756,-0.5907236934,-0.0206708051,0.0572946109,0.2173590213,0.3088783324,0.1889963001,0.1960792691,-0.1945684999,0.0154501591,0.2664080858,0.1986875087,0.3779762089,0.0560513809,-0.0483673327,0.1246331483,0.0030192293,0.1406399608,-0.2207594961,-0.2776945531,-0.0573592074,0.0422726683,0.3460413516,0.2213218659,0.0628878772,0.0959237814,0.1461020708,0.2581580877,-0.2759142816,0.2963343561,0.4933303595,0.1403307617,-0.4810526073,-0.3202238679,0.2370316833,0.2185061872,-0.1690102518,0.38164258,0.0929525048,-0.3298314214,0.5882653594,0.0854706466,0.8382591009,-0.3112327456,0.0554842912,-0.0850671604,0.0784885734,0.2867376208,0.0625677705,0.2233817428,-0.1989947259,-0.4336817265,0.0294411965,-0.0587247387,0.347617805,0.0871918276,-0.1870900691,0.2699613571,-0.2224864066,0.584843874,-0.2539536059,-0.2409918308,0.2116029114,-0.3743119836,0.1243179068,0.056649372,-0.0686615184,0.0904513001,0.0100760283,0.1401854753,-0.1021842211,-0.3583537638,0.0675272495,-0.0539399981,-0.2007298023,0.2325115949,-0.2137070447,-0.2194651663,-0.0675614402,0.3379043937,0.3049491644,0.2256917655,0.0769784302,0.2345013022,0.4431695342,0.2091673464,-0.0408834182,-0.1617862284,-0.0470286757,0.1401213855,-0.2852287292,0.1369479746,-0.0141862789,-0.2416308075,-0.364243418,0.0415634848,0.4323446155,-0.3393744826,-0.2499048561,-0.1570414305,-0.0112912552,-0.1000931934,0.0678598732,-0.3800917268,-0.1897207946,0.4002565742,-0.1145981029,-0.4112877548,0.1257440895,0.363469094,0.0822052211,0.0131317768,0.5031767488,-0.1995394677,0.0113480333,-0.08634267,0.0982044041,0.2158655226,-0.0867505968,0.0297973417,0.3238884807,-0.0309048165,-0.1596030295,0.4304632843,0.056714043,0.2015253901,0.0784534812,-0.2432671636,-0.6696433425,0.4877867103,0.1031506881,0.2492165565,0.1051922888,0.3467668593,-0.294855684,-0.011246155,-0.2399551123,0.0649206117,-0.0438210703,0.3154606819,0.4421718121,0.0253377911,-0.0896078646,-0.2898142636,0.2092776299,0.264129132,-0.0850462019,-0.213859275,-0.0167676937,0.1183085814,0.1053171828,-0.1505651921,-0.0243897997,-0.3231567144,-0.0274435021,-0.2703390718,-0.0828690007,0.3389027417,0.0339184143,0.1923668236,-0.5534645915,-0.0861907825,0.3065199554,0.1393099278,-0.1251810491,-0.0567402318,0.2653055489,0.1776153892,-0.050407812,0.0970639437,-0.2915789485,-0.068407774,-0.1543523222,0.1156141385,0.3472285569,-0.1888154894,0.1339828819,0.0561051331,0.2796280384,0.1310018599,-0.2598958313,-0.1948141307,0.3021291494,0.0988662317,-0.3267253339,-0.1829232723,0.5215923786,-0.0156410933,0.2411123663,0.2352423817,-0.0485072844,0.1316879392,-0.2601729333,-0.0177907143,0.1529346704,-0.0737773851,0.2469815612,0.6254125237,-0.1150746867,0.2715166807,0.2926054597,0.373080641,-0.1249766499,0.6644849181,0.0635503903,0.2042205036,0.1586616784,0.2688653469,-0.1144806743,-0.0899507031,0.5314686894,0.1346583664,-0.2235026807,0.2969570458,0.1907867789,-0.2590565085,-0.0018368585,-0.225257501,-0.0775448382,0.2807589173,-0.2177532315,-0.2554299235,-0.1515018791,-0.036063239,0.0017208962,0.0805628821,-0.0572090372,0.0685864314,0.6512500644,0.026028268,0.0510043502,0.0960447788,-0.5209681988,0.1352971196,0.3933674097,-0.1334355921,-0.017999975,0.0571532957,0.1642050743,0.2200044841,0.0155200986,0.3567886353,0.1692755967,-0.2766989768,0.1629797667,-0.0195907224,-0.1051777452,0.2641951144,0.0439048819,0.3267834783,0.008745905,0.1000831351,0.0357460454,-0.1855373085,-0.1050797775,0.148779586,0.1994472444,0.1967056245,-0.066673778,0.0776693895,-0.1384180486,-0.0612892099,0.1662924588,-0.2106761187,-0.2168651372,0.4741016328,-0.1182754636,0.1831553429,0.1314244568,0.0622664802,0.1147598624,0.1901857108,0.1068619639,-0.1823746413,-0.5617814064,-0.0506489016,-0.3618877232,-0.1454607099,-0.0483252704,0.0958933979,0.0001837238,-0.2958847284,0.2807190716,0.1011902243,0.4350568354,-0.4830625653,-0.0566600636,0.5606931448,0.0539841652,-0.0601437464,0.0202838592,0.2473428994,0.1078898311,-0.3471124768,0.0970269293,0.0586616993,-0.0088258684,-0.339279294,-0.1747089773,-0.2541410327,-0.2222277373,0.4715786278,0.1747033447,0.4249306321,-0.1517661214,-0.072775431,0.1637441069,-0.1676822305,-0.016834598,0.2095226794,0.2309502363,0.3883229494,-0.0711031109,0.128017351,-0.3970904052,0.0074610743,0.0302021448,-0.1853101104,-0.3765938282,0.186921224,0.2549928129,0.0408338793,-0.1853446364,0.3607902229,0.0388966613,0.5180019736,-0.0680310279,-0.3690660894,0.7985036969,-0.3579202294,-0.6600227952,0.0823596269,-0.0900159925,-0.1427418143,0.0412078798,-0.2388151884,-0.0630643219,0.073640883,-0.0334531926,-0.3466158807,0.0901280493,-0.1645772755,0.079836078,0.0364476554,0.4860982895,-0.3197670579,-0.0945576355,-0.2467267662,-0.0632384643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2026","title":"KeyError on using map after renaming a column","comments":"I just merged a @mariosasko 's PR that fixes this issue.\r\nIf it happens again, feel free to re-open :)","body":"Hi,\r\n\r\nI'm trying to use `cifar10` dataset. I want to rename the `img` feature to `image` in order to make it consistent with `mnist`, which I'm also planning to use. By doing this, I was trying to avoid modifying `prepare_train_features` function.\r\n\r\nHere is what I try:\r\n\r\n```python\r\ntransform = Compose([ToPILImage(),ToTensor(),Normalize([0.0,0.0,0.0],[1.0,1.0,1.0])])\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n print(examples)\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n else:\r\n images.append(examples[\"image\"][example_idx].permute(2,0,1))\r\n labels.append(examples[\"label\"][example_idx])\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('cifar10')\r\nraw_dataset.set_format('torch',columns=['img','label'])\r\nraw_dataset = raw_dataset.rename_column('img','image')\r\n\r\nfeatures = datasets.Features({\r\n \"image\": datasets.Array3D(shape=(3,32,32),dtype=\"float32\"),\r\n \"label\": datasets.features.ClassLabel(names=[\r\n \"airplane\",\r\n \"automobile\",\r\n \"bird\",\r\n \"cat\",\r\n \"deer\",\r\n \"dog\",\r\n \"frog\",\r\n \"horse\",\r\n \"ship\",\r\n \"truck\",\r\n ]),\r\n })\r\ntrain_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n```\r\nThe error:\r\n```python\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n in ()\r\n 14 ]),\r\n 15 })\r\n---> 16 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1287 test_inputs = self[:2] if batched else self[0]\r\n 1288 test_indices = [0, 1] if batched else 0\r\n-> 1289 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1290 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1291 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1258 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1259 processed_inputs = (\r\n-> 1260 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1261 )\r\n 1262 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n in prepare_features(examples)\r\n 3 labels = []\r\n 4 print(examples)\r\n----> 5 for example_idx, example in enumerate(examples[\"image\"]):\r\n 6 if transform is not None:\r\n 7 images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n\r\nKeyError: 'image'\r\n```\r\n\r\nThe print statement inside returns this:\r\n```python\r\n{'label': tensor([6, 9])}\r\n```\r\nApparently, both `img` and `image` do not exist after renaming. \r\n\r\nNote that this code works fine with `img` everywhere.\r\n\r\nNotebook: https:\/\/colab.research.google.com\/drive\/1SzESAlz3BnVYrgQeJ838vbMp1OsukiA2?usp=sharing\r\n\r\n","comment_length":20,"text":"KeyError on using map after renaming a column\nHi,\r\n\r\nI'm trying to use `cifar10` dataset. I want to rename the `img` feature to `image` in order to make it consistent with `mnist`, which I'm also planning to use. By doing this, I was trying to avoid modifying `prepare_train_features` function.\r\n\r\nHere is what I try:\r\n\r\n```python\r\ntransform = Compose([ToPILImage(),ToTensor(),Normalize([0.0,0.0,0.0],[1.0,1.0,1.0])])\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n print(examples)\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n else:\r\n images.append(examples[\"image\"][example_idx].permute(2,0,1))\r\n labels.append(examples[\"label\"][example_idx])\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('cifar10')\r\nraw_dataset.set_format('torch',columns=['img','label'])\r\nraw_dataset = raw_dataset.rename_column('img','image')\r\n\r\nfeatures = datasets.Features({\r\n \"image\": datasets.Array3D(shape=(3,32,32),dtype=\"float32\"),\r\n \"label\": datasets.features.ClassLabel(names=[\r\n \"airplane\",\r\n \"automobile\",\r\n \"bird\",\r\n \"cat\",\r\n \"deer\",\r\n \"dog\",\r\n \"frog\",\r\n \"horse\",\r\n \"ship\",\r\n \"truck\",\r\n ]),\r\n })\r\ntrain_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n```\r\nThe error:\r\n```python\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n in ()\r\n 14 ]),\r\n 15 })\r\n---> 16 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1287 test_inputs = self[:2] if batched else self[0]\r\n 1288 test_indices = [0, 1] if batched else 0\r\n-> 1289 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1290 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1291 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1258 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1259 processed_inputs = (\r\n-> 1260 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1261 )\r\n 1262 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n in prepare_features(examples)\r\n 3 labels = []\r\n 4 print(examples)\r\n----> 5 for example_idx, example in enumerate(examples[\"image\"]):\r\n 6 if transform is not None:\r\n 7 images.append(transform(examples[\"image\"][example_idx].permute(2,0,1)))\r\n\r\nKeyError: 'image'\r\n```\r\n\r\nThe print statement inside returns this:\r\n```python\r\n{'label': tensor([6, 9])}\r\n```\r\nApparently, both `img` and `image` do not exist after renaming. \r\n\r\nNote that this code works fine with `img` everywhere.\r\n\r\nNotebook: https:\/\/colab.research.google.com\/drive\/1SzESAlz3BnVYrgQeJ838vbMp1OsukiA2?usp=sharing\r\n\r\n\nI just merged a @mariosasko 's PR that fixes this issue.\r\nIf it happens again, feel free to re-open :)","embeddings":[0.0343904048,0.0061999829,-0.0635800064,-0.334228158,0.4808440804,0.2616392076,0.6042662859,0.237331897,0.1370912194,0.0865909159,0.054035522,0.5270302892,-0.1558949649,0.2994630933,-0.1945005655,-0.0881165639,0.4133820236,0.0935034528,-0.1156333089,0.1972025782,-0.4435485899,0.0497060306,-0.3014574051,0.2807840407,-0.3147513866,-0.1778963804,0.054533463,-0.1938546598,-0.2473179102,-0.274669081,-0.0492155105,0.0971030518,-0.0223661978,0.6005281806,-0.0001189421,0.0826987103,0.188707307,-0.0865515545,0.082985498,-0.2960751951,-0.1819153279,-0.0115416646,-0.2230614275,-0.5142738819,0.0406857617,-0.0636158809,-0.077172786,-0.2671773434,0.1024493352,0.2466998845,0.1638297886,-0.1138828397,0.2209231108,-0.0220486131,0.0439468622,0.2382541597,-0.1926510036,0.0769321546,0.0491049998,-0.3582004011,0.2290986776,0.693173945,-0.2142525613,-0.0826261491,0.2492942363,0.1485666484,0.4132107496,-0.3595764339,0.3614650071,-0.1366176903,0.2379055619,-0.1512385756,-0.1043806374,-0.054232765,-0.0018236447,-0.2537844181,-0.0232361704,-0.3196378648,0.0776960477,-0.2296992242,-0.3305271864,-0.1850074828,0.1285365671,0.2078439891,0.1231135502,-0.097198315,-0.0296082962,0.4492817819,0.1049034894,-0.322670877,-0.1246059462,0.0269131735,0.098084107,0.2211732715,-0.2241782248,-0.0697751865,0.105466947,-0.0227153953,-0.2865684628,-0.6728146672,0.017100919,-0.1154988408,0.1804634184,0.0785013214,0.106362313,0.2072906345,-0.087925978,0.3839033544,0.0437416993,0.0499913432,-0.3903314173,-0.2533300221,0.1843829304,-0.3990157247,0.3597591519,0.2472612262,0.3776059747,0.0936776847,0.4118515253,-0.0127613107,0.1394363791,-0.0965100452,-0.0235801451,0.3582357764,0.2444261461,-0.0504651479,0.1044974774,0.1474120915,-0.0377255827,0.2744560242,-0.0399699099,0.1520807594,-0.5921468139,-0.10537678,-0.1591845006,0.0644563138,0.0273747109,-0.1856066436,0.1200411022,-0.1029791012,-0.017414419,-0.062561661,0.4089117348,0.1914608777,-0.3518828154,0.1143235788,0.2287517935,-0.2352549285,-0.1231607944,0.2178232223,-0.6010237336,-0.1866197884,-0.1935406029,0.0911867172,0.2030768991,0.1996438056,-0.2752969265,-0.1346438378,0.5062403679,-0.4040708244,0.082678318,-0.2427093983,-0.2949000895,-0.2364552915,-0.1690340787,-0.0339538306,-0.0462761037,-0.0256844573,-0.1759432405,0.3576537967,0.0155100189,-0.0486620031,-0.1009822637,0.2132537961,-0.0342827775,0.1602661163,0.4271647036,-0.3145627677,-0.2063587606,-0.0152603928,-0.2203991115,-0.1616605967,-0.0920220613,0.093030706,0.2039043605,-0.0727139935,0.4399832785,-0.154413715,-0.0898718163,0.178547591,0.1311589926,-0.0462149791,0.1775305122,-0.0791230053,0.1281124204,0.208673045,0.1111345366,0.0116959661,-0.0227944031,-0.2068924159,0.0939341784,0.0991031379,0.2192526907,-0.0245531406,0.0360569097,0.0803689882,-0.3255301416,0.167881459,0.1803426594,0.141612947,-0.2925769389,-0.1196023375,-0.2550840676,-0.1336407512,-0.176066801,0.0316427574,0.0552044027,-0.054000631,-0.0344205536,-0.1491251588,-0.1648832858,-0.1527555883,0.0497404523,0.1093318611,-0.3486706614,-0.0184663273,-0.2110998482,-0.1441620439,-0.3850237131,0.0132771088,0.1881341487,-0.0540964752,-0.2381078899,0.3495651782,0.1393090487,0.0440969206,-0.3154114485,0.1756070554,0.0911011919,-0.0334517248,-0.0013047188,0.1042696312,0.1814650595,-0.0348044224,0.1123570204,0.2557912469,-0.2581357658,0.3806613088,-0.2395886481,0.1288557351,-0.0841588005,-0.1766913235,-0.2373664528,-0.393781811,-0.143782109,-0.2950774133,-0.2323311865,-0.0712196752,0.3053756952,-0.2079800516,0.5871576667,-0.0409827009,0.0240916926,-0.0057099815,0.0615631565,0.289617002,0.0114168618,0.1064895242,0.297829479,-0.009614123,-0.2233593762,0.2672612965,0.1708517224,0.2767239213,0.2040939629,0.2167057693,0.244703874,0.0480267294,0.1485542208,-0.0613893084,0.0164967366,-0.2999012768,0.3099987805,0.2844228446,-0.2336353511,0.0998230577,-0.0660420731,0.0700871125,0.2823909223,-0.1527176499,0.0562228933,-0.2700430751,-0.094888635,0.3865874112,0.0920764729,0.4332132339,-0.1842424124,-0.3789016604,0.2417967469,-0.35465464,-0.0365538523,-0.4494333863,0.030626379,-0.0325060226,0.0856916159,-0.4606234431,0.0817554891,-0.1647819877,-0.1355617791,-0.1211330071,-0.5227620006,0.1218558326,-0.4426466227,0.0148037989,0.2468884289,0.2771109343,-0.2926217616,-0.1771504432,0.4542585015,-0.2178657502,-0.1494309604,0.125867188,0.0121634537,-0.039333865,-0.1180340871,-0.1904156059,-0.0265278462,-0.1234948337,0.1169085056,-0.2646969259,0.0163139403,0.5068838596,0.2759175897,-0.2442613244,0.1430361718,-0.0489320122,-0.1584293842,-0.3822782338,0.208315134,-0.0868597105,0.0838408098,0.0047291089,-0.1143030971,-0.0037284025,0.0410409383,-0.1813217849,-0.292372793,-0.0811148286,0.252065897,0.1551319957,0.1015874445,0.2861323655,0.4163930118,-0.1233831346,0.0155468863,-0.2297620475,-0.0407741778,0.4496167004,0.2750443816,0.2839065492,0.4489243031,-0.0161301196,0.2931136191,-0.1320103854,-0.1058140993,0.4552730322,-0.3281630576,0.097139366,-0.427151382,-0.4983298779,0.021225499,-0.3456521928,-0.2692156136,-0.252612561,-0.1974028945,-0.2074296474,-0.1250825375,0.2784183919,-0.4862356186,-0.3314012587,0.1603580266,0.0705963597,0.30742383,0.0487649925,-0.0192793254,-0.4495798647,-0.1688798666,0.0980947465,0.0244836956,0.055906225,-0.3355075121,-0.6632701159,-0.2365220785,-0.3934516907,0.4276641011,0.1131599844,0.1529776901,-0.1410124451,-0.0606670156,-0.0556474812,0.0688659251,0.7746023536,-0.4415769279,-0.1148998216,0.3010852635,-0.0476227738,-0.0988456905,-0.2432980686,-0.0672431141,0.2733609378,0.0472895429,0.8124082088,0.0474174544,0.0843919963,0.214545846,0.0629557222,-0.0658334941,0.0286184251,-0.406054467,-0.2894008756,-0.5907236934,-0.0206708051,0.0572946109,0.2173590213,0.3088783324,0.1889963001,0.1960792691,-0.1945684999,0.0154501591,0.2664080858,0.1986875087,0.3779762089,0.0560513809,-0.0483673327,0.1246331483,0.0030192293,0.1406399608,-0.2207594961,-0.2776945531,-0.0573592074,0.0422726683,0.3460413516,0.2213218659,0.0628878772,0.0959237814,0.1461020708,0.2581580877,-0.2759142816,0.2963343561,0.4933303595,0.1403307617,-0.4810526073,-0.3202238679,0.2370316833,0.2185061872,-0.1690102518,0.38164258,0.0929525048,-0.3298314214,0.5882653594,0.0854706466,0.8382591009,-0.3112327456,0.0554842912,-0.0850671604,0.0784885734,0.2867376208,0.0625677705,0.2233817428,-0.1989947259,-0.4336817265,0.0294411965,-0.0587247387,0.347617805,0.0871918276,-0.1870900691,0.2699613571,-0.2224864066,0.584843874,-0.2539536059,-0.2409918308,0.2116029114,-0.3743119836,0.1243179068,0.056649372,-0.0686615184,0.0904513001,0.0100760283,0.1401854753,-0.1021842211,-0.3583537638,0.0675272495,-0.0539399981,-0.2007298023,0.2325115949,-0.2137070447,-0.2194651663,-0.0675614402,0.3379043937,0.3049491644,0.2256917655,0.0769784302,0.2345013022,0.4431695342,0.2091673464,-0.0408834182,-0.1617862284,-0.0470286757,0.1401213855,-0.2852287292,0.1369479746,-0.0141862789,-0.2416308075,-0.364243418,0.0415634848,0.4323446155,-0.3393744826,-0.2499048561,-0.1570414305,-0.0112912552,-0.1000931934,0.0678598732,-0.3800917268,-0.1897207946,0.4002565742,-0.1145981029,-0.4112877548,0.1257440895,0.363469094,0.0822052211,0.0131317768,0.5031767488,-0.1995394677,0.0113480333,-0.08634267,0.0982044041,0.2158655226,-0.0867505968,0.0297973417,0.3238884807,-0.0309048165,-0.1596030295,0.4304632843,0.056714043,0.2015253901,0.0784534812,-0.2432671636,-0.6696433425,0.4877867103,0.1031506881,0.2492165565,0.1051922888,0.3467668593,-0.294855684,-0.011246155,-0.2399551123,0.0649206117,-0.0438210703,0.3154606819,0.4421718121,0.0253377911,-0.0896078646,-0.2898142636,0.2092776299,0.264129132,-0.0850462019,-0.213859275,-0.0167676937,0.1183085814,0.1053171828,-0.1505651921,-0.0243897997,-0.3231567144,-0.0274435021,-0.2703390718,-0.0828690007,0.3389027417,0.0339184143,0.1923668236,-0.5534645915,-0.0861907825,0.3065199554,0.1393099278,-0.1251810491,-0.0567402318,0.2653055489,0.1776153892,-0.050407812,0.0970639437,-0.2915789485,-0.068407774,-0.1543523222,0.1156141385,0.3472285569,-0.1888154894,0.1339828819,0.0561051331,0.2796280384,0.1310018599,-0.2598958313,-0.1948141307,0.3021291494,0.0988662317,-0.3267253339,-0.1829232723,0.5215923786,-0.0156410933,0.2411123663,0.2352423817,-0.0485072844,0.1316879392,-0.2601729333,-0.0177907143,0.1529346704,-0.0737773851,0.2469815612,0.6254125237,-0.1150746867,0.2715166807,0.2926054597,0.373080641,-0.1249766499,0.6644849181,0.0635503903,0.2042205036,0.1586616784,0.2688653469,-0.1144806743,-0.0899507031,0.5314686894,0.1346583664,-0.2235026807,0.2969570458,0.1907867789,-0.2590565085,-0.0018368585,-0.225257501,-0.0775448382,0.2807589173,-0.2177532315,-0.2554299235,-0.1515018791,-0.036063239,0.0017208962,0.0805628821,-0.0572090372,0.0685864314,0.6512500644,0.026028268,0.0510043502,0.0960447788,-0.5209681988,0.1352971196,0.3933674097,-0.1334355921,-0.017999975,0.0571532957,0.1642050743,0.2200044841,0.0155200986,0.3567886353,0.1692755967,-0.2766989768,0.1629797667,-0.0195907224,-0.1051777452,0.2641951144,0.0439048819,0.3267834783,0.008745905,0.1000831351,0.0357460454,-0.1855373085,-0.1050797775,0.148779586,0.1994472444,0.1967056245,-0.066673778,0.0776693895,-0.1384180486,-0.0612892099,0.1662924588,-0.2106761187,-0.2168651372,0.4741016328,-0.1182754636,0.1831553429,0.1314244568,0.0622664802,0.1147598624,0.1901857108,0.1068619639,-0.1823746413,-0.5617814064,-0.0506489016,-0.3618877232,-0.1454607099,-0.0483252704,0.0958933979,0.0001837238,-0.2958847284,0.2807190716,0.1011902243,0.4350568354,-0.4830625653,-0.0566600636,0.5606931448,0.0539841652,-0.0601437464,0.0202838592,0.2473428994,0.1078898311,-0.3471124768,0.0970269293,0.0586616993,-0.0088258684,-0.339279294,-0.1747089773,-0.2541410327,-0.2222277373,0.4715786278,0.1747033447,0.4249306321,-0.1517661214,-0.072775431,0.1637441069,-0.1676822305,-0.016834598,0.2095226794,0.2309502363,0.3883229494,-0.0711031109,0.128017351,-0.3970904052,0.0074610743,0.0302021448,-0.1853101104,-0.3765938282,0.186921224,0.2549928129,0.0408338793,-0.1853446364,0.3607902229,0.0388966613,0.5180019736,-0.0680310279,-0.3690660894,0.7985036969,-0.3579202294,-0.6600227952,0.0823596269,-0.0900159925,-0.1427418143,0.0412078798,-0.2388151884,-0.0630643219,0.073640883,-0.0334531926,-0.3466158807,0.0901280493,-0.1645772755,0.079836078,0.0364476554,0.4860982895,-0.3197670579,-0.0945576355,-0.2467267662,-0.0632384643]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2022","title":"ValueError when rename_column on splitted dataset","comments":"Hi,\r\n\r\nThis is a bug so thanks for reporting it. `Dataset.__setstate__` is the problem, which is called when `Dataset.rename_column` tries to copy the dataset with `copy.deepcopy(self)`. This only happens if the `split` arg in `load_dataset` was defined as `ReadInstruction`.\r\n\r\nTo overcome this issue, use the named splits API (for now):\r\n```python\r\ntrain_ds, test_ds = load_dataset(\r\n path='csv', \r\n delimiter='\\t', \r\n data_files=text_files, \r\n split=['train[:90%]', 'train[-10%:]'],\r\n)\r\n\r\ntrain_ds = train_ds.rename_column('sentence', 'text')\r\n```","body":"Hi there,\r\nI am loading `.tsv` file via `load_dataset` and subsequently split the rows into training and test set via the `ReadInstruction` API like so:\r\n\r\n```python\r\nsplit = {\r\n 'train': ReadInstruction('train', to=90, unit='%'),\r\n 'test': ReadInstruction('train', from_=-10, unit='%')\r\n}\r\n\r\ndataset = load_dataset(\r\n path='csv', # use 'text' loading script to load from local txt-files\r\n delimiter='\\t', # xxx\r\n data_files=text_files, # list of paths to local text files\r\n split=split, # xxx\r\n)\r\n\r\ndataset\r\n```\r\n\r\nPart of output:\r\n```python\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'sentiment'],\r\n num_rows: 900\r\n })\r\n test: Dataset({\r\n features: ['sentence', 'sentiment'],\r\n num_rows: 100\r\n })\r\n})\r\n```\r\nAfterwards I'd like to rename the 'sentence' column to 'text' in order to be compatible with my modelin pipeline. If I run the following code I experience a `ValueError` however:\r\n```python\r\ndataset['train'].rename_column('sentence', 'text')\r\n```\r\n```python\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/splits.py in __init__(self, name)\r\n 353 for split_name in split_names_from_instruction:\r\n 354 if not re.match(_split_re, split_name):\r\n--> 355 raise ValueError(f\"Split name should match '{_split_re}'' but got '{split_name}'.\")\r\n 356 \r\n 357 def __str__(self):\r\n\r\nValueError: Split name should match '^\\w+(\\.\\w+)*$'' but got 'ReadInstruction('.\r\n```\r\nIn particular, these behavior does not arise if I use the deprecated `rename_column_` method. Any idea what causes the error? Would assume something in the way I defined the split.\r\n\r\nThanks in advance! :)","comment_length":66,"text":"ValueError when rename_column on splitted dataset\nHi there,\r\nI am loading `.tsv` file via `load_dataset` and subsequently split the rows into training and test set via the `ReadInstruction` API like so:\r\n\r\n```python\r\nsplit = {\r\n 'train': ReadInstruction('train', to=90, unit='%'),\r\n 'test': ReadInstruction('train', from_=-10, unit='%')\r\n}\r\n\r\ndataset = load_dataset(\r\n path='csv', # use 'text' loading script to load from local txt-files\r\n delimiter='\\t', # xxx\r\n data_files=text_files, # list of paths to local text files\r\n split=split, # xxx\r\n)\r\n\r\ndataset\r\n```\r\n\r\nPart of output:\r\n```python\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'sentiment'],\r\n num_rows: 900\r\n })\r\n test: Dataset({\r\n features: ['sentence', 'sentiment'],\r\n num_rows: 100\r\n })\r\n})\r\n```\r\nAfterwards I'd like to rename the 'sentence' column to 'text' in order to be compatible with my modelin pipeline. If I run the following code I experience a `ValueError` however:\r\n```python\r\ndataset['train'].rename_column('sentence', 'text')\r\n```\r\n```python\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/splits.py in __init__(self, name)\r\n 353 for split_name in split_names_from_instruction:\r\n 354 if not re.match(_split_re, split_name):\r\n--> 355 raise ValueError(f\"Split name should match '{_split_re}'' but got '{split_name}'.\")\r\n 356 \r\n 357 def __str__(self):\r\n\r\nValueError: Split name should match '^\\w+(\\.\\w+)*$'' but got 'ReadInstruction('.\r\n```\r\nIn particular, these behavior does not arise if I use the deprecated `rename_column_` method. Any idea what causes the error? Would assume something in the way I defined the split.\r\n\r\nThanks in advance! :)\nHi,\r\n\r\nThis is a bug so thanks for reporting it. `Dataset.__setstate__` is the problem, which is called when `Dataset.rename_column` tries to copy the dataset with `copy.deepcopy(self)`. This only happens if the `split` arg in `load_dataset` was defined as `ReadInstruction`.\r\n\r\nTo overcome this issue, use the named splits API (for now):\r\n```python\r\ntrain_ds, test_ds = load_dataset(\r\n path='csv', \r\n delimiter='\\t', \r\n data_files=text_files, \r\n split=['train[:90%]', 'train[-10%:]'],\r\n)\r\n\r\ntrain_ds = train_ds.rename_column('sentence', 'text')\r\n```","embeddings":[-0.0835750848,0.2197992057,-0.0348290578,-0.0423846953,0.4208770096,0.0730622262,0.6438042521,0.417984277,-0.0225529727,0.3396532834,-0.0866872147,0.3978035748,-0.0521849729,0.3644353151,-0.2562588155,-0.2477057576,0.1044141725,-0.0795424059,0.1267364323,0.2697291374,-0.3300528526,0.0778078511,-0.3932777643,0.126068756,-0.0781336427,0.0623800941,0.0325193703,0.095376946,0.0282755606,-0.3610774279,0.1441161931,-0.0355102234,0.1899649948,0.5556836724,-0.0001162362,0.123281084,0.2225487083,-0.1280173063,-0.1891635507,-0.4528475702,-0.0052785026,-0.1680362076,0.0017363135,-0.350676477,0.1707240045,0.2284773737,-0.2309949547,-0.2584788203,0.1006070673,0.3693120182,0.1076899767,0.0965298414,-0.0107203871,0.1191240698,0.1085148677,0.02124759,-0.3047722876,0.0899827629,-0.0761238411,-0.1792219281,-0.1137903929,0.2317518592,-0.0711766481,0.1480419189,-0.0235446021,0.3106743395,0.0345892012,-0.046964284,0.0888525695,0.0671797469,0.2144692689,-0.2401781976,-0.4639912844,-0.4442960322,0.1637722403,-0.3250167668,0.2570495903,-0.1877467483,0.1232365146,0.1840747148,-0.1398658454,0.0523615777,-0.1373051703,-0.1406392157,-0.0923041031,0.0893618017,-0.0323616415,0.4094693363,-0.1002934575,-0.1114295572,0.1749086529,0.0513332188,-0.0785202086,-0.1311808378,-0.5019623637,0.0278138947,-0.2302477956,-0.2595581114,-0.320128262,-0.1200023293,0.015836617,-0.2517315745,0.2726792395,0.2164970785,-0.0362976082,0.1996526569,0.1300997585,0.4851779044,0.2359965295,0.1402986795,-0.2038473189,-0.0915264264,0.0607342832,-0.2988728285,0.1496177316,0.1908584237,0.1388384253,0.1012246609,-0.422322154,-0.0214158427,-0.299931258,-0.1886160672,-0.030533433,0.2214560509,0.2032395601,0.4568883777,0.1693875492,0.0574802384,-0.0006995207,-0.1406515241,-0.0371253602,0.1793324798,-0.301969558,0.1505992264,0.0399702266,0.1964562237,0.0814438388,-0.2706573308,-0.0208249222,-0.299153924,-0.0203204229,-0.1991917193,0.0952113345,0.2918279171,-0.0835521519,-0.107908383,0.116414018,-0.4457986653,0.0352158584,0.2579235137,-0.2536048889,-0.320792675,0.1397257596,0.1286002249,0.0399009176,-0.0527537614,0.0482414179,-0.0229599494,0.6056432724,-0.1394455284,0.1351458579,-0.3096872866,-0.2267056406,-0.0160398446,-0.0332796611,0.1000444666,-0.5333275795,-0.1184836403,0.0636194199,0.1941967756,0.1769787073,0.1755306721,0.1830014735,0.1583334655,0.1335396022,0.266903609,0.2967285514,-0.3958649337,0.0288950782,0.2807647586,-0.3185531199,-0.1601911634,0.3707808554,-0.2173383236,0.1207509637,-0.1076937914,0.1341798902,0.0113364346,-0.0657366589,0.0960765854,-0.0378147103,-0.0701453909,0.3307214677,-0.018021524,0.0188636724,-0.1157904789,-0.1568621844,-0.2933054864,0.3243606389,-0.0264651477,0.1164182797,0.1300329119,0.1057565063,0.5420783758,0.1200013235,-0.0832299888,-0.2253147513,0.0283834282,0.3180817366,-0.1683246642,-0.3165953457,-0.1709285378,-0.432600975,0.0489438772,-0.0752222762,0.047204908,0.0313136242,0.2542936206,-0.1051330194,-0.1217908859,-0.4117276967,0.2951657772,-0.5749081969,0.2464329451,-0.3180003464,0.0775999576,0.0061799488,-0.3219508827,-0.2502307892,0.1552675813,0.1714720279,-0.0112890517,-0.2374314368,0.6657509208,0.34775123,0.0463436544,-0.3765339553,-0.2769818604,-0.1085001081,0.2417751104,-0.2358049601,-0.1330698282,0.2457468659,-0.0412518121,-0.0628126115,0.2938631177,-0.4202154875,0.5500609875,-0.0109366123,-0.1532364488,0.002068257,-0.2575076222,-0.1037362888,-0.3930054307,-0.1505555212,-0.3311613202,0.1140188873,-0.196557343,-0.1869239062,-0.2932314277,0.6310689449,-0.0536831878,-0.1708038598,0.0077965646,0.1357934773,-0.0466479398,-0.0799038783,0.6231101751,0.4926960766,0.0294919666,-0.1018566489,0.2511084676,0.1871850491,-0.2857350409,0.2048228085,0.2681159675,0.1419734806,0.4111416042,0.2456434667,0.005155154,-0.2268374413,-0.1838648766,0.3968054354,0.1195372418,-0.4919120967,0.0235623717,-0.164911896,0.2419187725,-0.2057008147,-0.226712659,0.1853089333,-0.544261992,-0.2349176556,0.3472461104,-0.3372885883,0.3051025867,-0.1589921862,-0.3950490952,0.0922147259,-0.1661286056,-0.0719222948,-0.1782116592,-0.0885189027,-0.0458369814,-0.126146242,-0.0972286686,0.0410267226,-0.1662219167,-0.0943965763,-0.1365423203,-0.2923829257,-0.0274039879,-0.4529346526,-0.0093845455,0.400508523,0.0366891734,0.1800103635,-0.4637690783,0.2288455218,0.0775950626,-0.0209740959,0.0346015543,0.2929466963,0.0061281142,-0.1640412956,-0.669023633,-0.2725984752,-0.3042469621,0.1338291615,-0.2729791403,-0.1675990075,0.0596722402,-0.1618382335,0.0142350011,0.1734488904,0.1514361352,-0.0215064287,0.0119573008,0.150385648,0.0146297365,0.1776951849,0.1265307963,0.1827682853,-0.1011585221,0.0592221878,-0.1364729106,-0.0564785674,-0.0916979313,0.3380533755,0.0930299312,-0.0017823929,0.2639890909,0.2979951799,-0.0363329686,-0.1717988253,-0.2675010264,0.246560514,0.0576315522,0.1198230162,0.0803304985,0.6318507195,0.2055416703,0.1752791852,-0.150716871,-0.0701785162,0.0658943951,-0.0693151802,0.0188845173,-0.190381512,-0.0143460678,-0.0719314367,-0.3799416125,-0.2754399478,0.0968471766,-0.1404148936,-0.0746427327,-0.0370618328,0.0086293966,0.0436806567,-0.2274768203,-0.0524387397,-0.0882400125,0.4275676012,0.1327891052,0.139078632,-0.0748876482,-0.0587312877,0.1260323972,0.1799938828,0.1219586879,-0.3082248867,-0.2348091751,0.0278907008,-0.2218738943,0.2311760783,0.3654352725,0.5237157941,0.1095001772,-0.4189083874,0.0391259268,0.1244078428,0.5819937587,-0.2301923931,-0.0308651738,0.1820007116,-0.158254832,0.0998644531,-0.168094635,-0.2026173323,-0.029728381,0.2626563609,0.4312624633,-0.0818835497,-0.0455454998,0.5021079183,0.2213865966,-0.0002616506,-0.2003735304,-0.1444071531,-0.3349117935,-0.2491312474,-0.0207623672,0.0515861511,0.1574812829,0.0514002778,0.3127942979,-0.0301315226,-0.3142569661,0.1404965967,0.2325978875,-0.0428324528,-0.0776599124,0.1895497143,0.0559758134,0.1787735671,0.0804946572,0.2318656892,-0.4986412227,-0.6792055368,-0.0838990584,-0.1733059734,0.4837495685,0.2373588234,-0.2355024964,0.2386443764,-0.0622689351,0.2222546637,-0.1055742428,-0.0054897927,0.3826481402,0.0841321126,-0.3165595829,-0.5406337976,0.3414804935,0.0795637742,-0.0830515847,0.0900762975,0.3665972054,-0.4676115811,0.4190027118,-0.1272712648,0.539450407,0.1711793393,-0.0342090391,0.4204019308,-0.1584545076,-0.0678915158,-0.0687830672,0.1058086008,-0.3747582436,-0.2032316327,-0.0629145205,0.0981257334,0.445771724,0.0439912751,-0.023987921,0.3131878078,-0.0377460159,0.3181696236,-0.0129596712,-0.0236804876,-0.1248432323,-0.2890895903,0.031416703,0.0170258172,0.0585119203,-0.3307894468,0.3468234837,0.1850599647,-0.0445544347,-0.0068801297,-0.1796743274,-0.0920961052,-0.0396208726,0.2621534169,-0.070273675,-0.2068292499,0.2561681271,0.2360015213,0.3455247581,0.1204388887,0.2218238413,0.2901317179,0.0536015704,0.3150189221,0.0510928147,-0.0520142913,0.3549725413,0.2559574544,-0.1892668158,-0.1287592649,-0.0970653892,-0.1592885107,-0.1369861662,0.004530733,0.3254045248,-0.712518096,-0.2792513967,-0.1463442296,0.0076214471,-0.2057145387,0.0702228099,-0.4360507131,-0.321126461,0.0423506647,-0.2473160177,-0.5409811139,-0.0410532616,0.2723687887,0.0928785577,0.4236688912,0.6302575469,0.074452199,-0.118960768,-0.0769284368,0.2815288007,-0.0017072947,-0.237287432,0.3587244451,0.2491544485,0.012526351,0.1421516538,0.0894049779,0.0139233461,0.1374352425,-0.1794731617,-0.2946612239,-0.2364348024,0.0364805385,0.2371767908,0.0985815376,0.182802856,0.2921203077,-0.4138016701,-0.1671654135,-0.1980048865,-0.0679664761,-0.0445374101,0.3395097256,0.2307618111,0.0757984892,-0.0225777905,-0.169207111,0.2093006372,0.3660119474,0.3369351923,-0.1701856405,-0.3115433455,0.1045251191,0.0879617333,0.3199940622,-0.2685281038,-0.0645283014,-0.1654905379,-0.2432912886,0.1397914737,0.4403597713,0.009370368,0.6279401183,-0.2326393574,0.0618086569,-0.0221137647,0.197350949,-0.3701749444,0.1364082247,-0.177895233,0.3094801903,-0.0463115312,0.1947893798,-0.3871267736,-0.1407456994,-0.1017042398,0.0919170305,0.162463069,-0.2329881787,0.0275496356,-0.2276391536,0.3421429396,0.5766685605,-0.0366016701,-0.3406235278,0.260145992,0.0938546285,-0.2149893343,0.2604703605,0.1313986778,-0.3491407931,0.2399219275,0.2198221982,0.02382572,0.2739397883,-0.4158225358,0.0962316617,0.4593583643,-0.2312578559,0.0933077112,0.747166872,-0.0716879517,0.3275974691,0.2411202639,0.1174242198,-0.1474771351,0.5122222304,0.2956359982,0.3313479424,0.1835952997,0.0873710364,-0.1988353878,-0.0440695621,0.0718020946,0.0233148132,-0.2806161344,-0.0010325464,0.0523829348,0.0993889272,-0.1351288557,0.0063865436,0.0764461532,0.0787790418,-0.0853217691,-0.3726897538,-0.1807307601,-0.0773208365,-0.4351949096,0.0100043993,-0.0010135098,0.2239075005,0.5572535396,-0.0630996972,0.2330803126,-0.4022642076,-0.1216403991,-0.0671678632,0.3376493752,-0.4362981319,0.0450070947,0.196776256,0.0978504643,0.3461167812,-0.273642391,0.367490381,0.0158707518,-0.0313476436,-0.3278319836,-0.2703055739,-0.1479447186,0.258428216,0.4770182073,-0.2473756522,0.1601275951,0.3426007032,0.0708493665,-0.0214186683,0.3009721637,0.2921059132,0.1484791785,-0.3536863327,0.627040565,0.0342792906,-0.1725035012,-0.0649004653,0.0291266907,-0.2207119912,0.0153460484,0.4927920103,-0.0857059211,0.2625669837,0.0981514975,0.0294042714,0.2018343657,0.3894335926,0.2932920754,-0.4533614516,-0.0536827073,0.165880844,-0.3433842361,-0.1413582265,0.1090332642,-0.1523897201,-0.104317762,-0.2610399127,0.100369066,-0.0362396799,0.0448415726,0.1843632162,-0.2561531663,0.238036558,-0.0057349671,-0.3758705854,-0.0841977596,0.0209682155,0.0370429195,-0.3455042839,-0.0190412477,0.4193592668,-0.1062776744,-0.2612063885,-0.1117050424,0.3965735137,0.0212082788,-0.0702146515,0.2049697191,0.5175044537,-0.2131402194,0.1191594452,0.0022142441,-0.3588138819,-0.4568068981,0.4172988236,-0.0768352598,0.3570566773,0.037399061,0.0126483692,-0.0493573882,0.103703931,-0.2392890453,0.0344574749,-0.4673057795,0.0310813338,0.2342381328,0.0649655536,0.1397208273,0.2943555415,0.1055681333,0.4483832717,-0.124229677,-0.3101114631,0.5607996583,-0.3047522008,-0.6528941393,0.1209915355,-0.1744092405,0.3874828219,0.0959179625,-0.6975602508,-0.1059781834,0.3279694021,-0.0964661986,-0.2087817937,0.177573964,0.2100795805,-0.0227706879,0.02478984,0.1218380257,0.0436747037,-0.1834020466,-0.0161779039,-0.0734406784]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2022","title":"ValueError when rename_column on splitted dataset","comments":"This has been fixed in #2043 , thanks @mariosasko \r\nThe fix is available on master and we'll do a new release soon :)\r\n\r\nfeel free to re-open if you still have issues","body":"Hi there,\r\nI am loading `.tsv` file via `load_dataset` and subsequently split the rows into training and test set via the `ReadInstruction` API like so:\r\n\r\n```python\r\nsplit = {\r\n 'train': ReadInstruction('train', to=90, unit='%'),\r\n 'test': ReadInstruction('train', from_=-10, unit='%')\r\n}\r\n\r\ndataset = load_dataset(\r\n path='csv', # use 'text' loading script to load from local txt-files\r\n delimiter='\\t', # xxx\r\n data_files=text_files, # list of paths to local text files\r\n split=split, # xxx\r\n)\r\n\r\ndataset\r\n```\r\n\r\nPart of output:\r\n```python\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'sentiment'],\r\n num_rows: 900\r\n })\r\n test: Dataset({\r\n features: ['sentence', 'sentiment'],\r\n num_rows: 100\r\n })\r\n})\r\n```\r\nAfterwards I'd like to rename the 'sentence' column to 'text' in order to be compatible with my modelin pipeline. If I run the following code I experience a `ValueError` however:\r\n```python\r\ndataset['train'].rename_column('sentence', 'text')\r\n```\r\n```python\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/splits.py in __init__(self, name)\r\n 353 for split_name in split_names_from_instruction:\r\n 354 if not re.match(_split_re, split_name):\r\n--> 355 raise ValueError(f\"Split name should match '{_split_re}'' but got '{split_name}'.\")\r\n 356 \r\n 357 def __str__(self):\r\n\r\nValueError: Split name should match '^\\w+(\\.\\w+)*$'' but got 'ReadInstruction('.\r\n```\r\nIn particular, these behavior does not arise if I use the deprecated `rename_column_` method. Any idea what causes the error? Would assume something in the way I defined the split.\r\n\r\nThanks in advance! :)","comment_length":32,"text":"ValueError when rename_column on splitted dataset\nHi there,\r\nI am loading `.tsv` file via `load_dataset` and subsequently split the rows into training and test set via the `ReadInstruction` API like so:\r\n\r\n```python\r\nsplit = {\r\n 'train': ReadInstruction('train', to=90, unit='%'),\r\n 'test': ReadInstruction('train', from_=-10, unit='%')\r\n}\r\n\r\ndataset = load_dataset(\r\n path='csv', # use 'text' loading script to load from local txt-files\r\n delimiter='\\t', # xxx\r\n data_files=text_files, # list of paths to local text files\r\n split=split, # xxx\r\n)\r\n\r\ndataset\r\n```\r\n\r\nPart of output:\r\n```python\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'sentiment'],\r\n num_rows: 900\r\n })\r\n test: Dataset({\r\n features: ['sentence', 'sentiment'],\r\n num_rows: 100\r\n })\r\n})\r\n```\r\nAfterwards I'd like to rename the 'sentence' column to 'text' in order to be compatible with my modelin pipeline. If I run the following code I experience a `ValueError` however:\r\n```python\r\ndataset['train'].rename_column('sentence', 'text')\r\n```\r\n```python\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/splits.py in __init__(self, name)\r\n 353 for split_name in split_names_from_instruction:\r\n 354 if not re.match(_split_re, split_name):\r\n--> 355 raise ValueError(f\"Split name should match '{_split_re}'' but got '{split_name}'.\")\r\n 356 \r\n 357 def __str__(self):\r\n\r\nValueError: Split name should match '^\\w+(\\.\\w+)*$'' but got 'ReadInstruction('.\r\n```\r\nIn particular, these behavior does not arise if I use the deprecated `rename_column_` method. Any idea what causes the error? Would assume something in the way I defined the split.\r\n\r\nThanks in advance! :)\nThis has been fixed in #2043 , thanks @mariosasko \r\nThe fix is available on master and we'll do a new release soon :)\r\n\r\nfeel free to re-open if you still have issues","embeddings":[-0.0835750848,0.2197992057,-0.0348290578,-0.0423846953,0.4208770096,0.0730622262,0.6438042521,0.417984277,-0.0225529727,0.3396532834,-0.0866872147,0.3978035748,-0.0521849729,0.3644353151,-0.2562588155,-0.2477057576,0.1044141725,-0.0795424059,0.1267364323,0.2697291374,-0.3300528526,0.0778078511,-0.3932777643,0.126068756,-0.0781336427,0.0623800941,0.0325193703,0.095376946,0.0282755606,-0.3610774279,0.1441161931,-0.0355102234,0.1899649948,0.5556836724,-0.0001162362,0.123281084,0.2225487083,-0.1280173063,-0.1891635507,-0.4528475702,-0.0052785026,-0.1680362076,0.0017363135,-0.350676477,0.1707240045,0.2284773737,-0.2309949547,-0.2584788203,0.1006070673,0.3693120182,0.1076899767,0.0965298414,-0.0107203871,0.1191240698,0.1085148677,0.02124759,-0.3047722876,0.0899827629,-0.0761238411,-0.1792219281,-0.1137903929,0.2317518592,-0.0711766481,0.1480419189,-0.0235446021,0.3106743395,0.0345892012,-0.046964284,0.0888525695,0.0671797469,0.2144692689,-0.2401781976,-0.4639912844,-0.4442960322,0.1637722403,-0.3250167668,0.2570495903,-0.1877467483,0.1232365146,0.1840747148,-0.1398658454,0.0523615777,-0.1373051703,-0.1406392157,-0.0923041031,0.0893618017,-0.0323616415,0.4094693363,-0.1002934575,-0.1114295572,0.1749086529,0.0513332188,-0.0785202086,-0.1311808378,-0.5019623637,0.0278138947,-0.2302477956,-0.2595581114,-0.320128262,-0.1200023293,0.015836617,-0.2517315745,0.2726792395,0.2164970785,-0.0362976082,0.1996526569,0.1300997585,0.4851779044,0.2359965295,0.1402986795,-0.2038473189,-0.0915264264,0.0607342832,-0.2988728285,0.1496177316,0.1908584237,0.1388384253,0.1012246609,-0.422322154,-0.0214158427,-0.299931258,-0.1886160672,-0.030533433,0.2214560509,0.2032395601,0.4568883777,0.1693875492,0.0574802384,-0.0006995207,-0.1406515241,-0.0371253602,0.1793324798,-0.301969558,0.1505992264,0.0399702266,0.1964562237,0.0814438388,-0.2706573308,-0.0208249222,-0.299153924,-0.0203204229,-0.1991917193,0.0952113345,0.2918279171,-0.0835521519,-0.107908383,0.116414018,-0.4457986653,0.0352158584,0.2579235137,-0.2536048889,-0.320792675,0.1397257596,0.1286002249,0.0399009176,-0.0527537614,0.0482414179,-0.0229599494,0.6056432724,-0.1394455284,0.1351458579,-0.3096872866,-0.2267056406,-0.0160398446,-0.0332796611,0.1000444666,-0.5333275795,-0.1184836403,0.0636194199,0.1941967756,0.1769787073,0.1755306721,0.1830014735,0.1583334655,0.1335396022,0.266903609,0.2967285514,-0.3958649337,0.0288950782,0.2807647586,-0.3185531199,-0.1601911634,0.3707808554,-0.2173383236,0.1207509637,-0.1076937914,0.1341798902,0.0113364346,-0.0657366589,0.0960765854,-0.0378147103,-0.0701453909,0.3307214677,-0.018021524,0.0188636724,-0.1157904789,-0.1568621844,-0.2933054864,0.3243606389,-0.0264651477,0.1164182797,0.1300329119,0.1057565063,0.5420783758,0.1200013235,-0.0832299888,-0.2253147513,0.0283834282,0.3180817366,-0.1683246642,-0.3165953457,-0.1709285378,-0.432600975,0.0489438772,-0.0752222762,0.047204908,0.0313136242,0.2542936206,-0.1051330194,-0.1217908859,-0.4117276967,0.2951657772,-0.5749081969,0.2464329451,-0.3180003464,0.0775999576,0.0061799488,-0.3219508827,-0.2502307892,0.1552675813,0.1714720279,-0.0112890517,-0.2374314368,0.6657509208,0.34775123,0.0463436544,-0.3765339553,-0.2769818604,-0.1085001081,0.2417751104,-0.2358049601,-0.1330698282,0.2457468659,-0.0412518121,-0.0628126115,0.2938631177,-0.4202154875,0.5500609875,-0.0109366123,-0.1532364488,0.002068257,-0.2575076222,-0.1037362888,-0.3930054307,-0.1505555212,-0.3311613202,0.1140188873,-0.196557343,-0.1869239062,-0.2932314277,0.6310689449,-0.0536831878,-0.1708038598,0.0077965646,0.1357934773,-0.0466479398,-0.0799038783,0.6231101751,0.4926960766,0.0294919666,-0.1018566489,0.2511084676,0.1871850491,-0.2857350409,0.2048228085,0.2681159675,0.1419734806,0.4111416042,0.2456434667,0.005155154,-0.2268374413,-0.1838648766,0.3968054354,0.1195372418,-0.4919120967,0.0235623717,-0.164911896,0.2419187725,-0.2057008147,-0.226712659,0.1853089333,-0.544261992,-0.2349176556,0.3472461104,-0.3372885883,0.3051025867,-0.1589921862,-0.3950490952,0.0922147259,-0.1661286056,-0.0719222948,-0.1782116592,-0.0885189027,-0.0458369814,-0.126146242,-0.0972286686,0.0410267226,-0.1662219167,-0.0943965763,-0.1365423203,-0.2923829257,-0.0274039879,-0.4529346526,-0.0093845455,0.400508523,0.0366891734,0.1800103635,-0.4637690783,0.2288455218,0.0775950626,-0.0209740959,0.0346015543,0.2929466963,0.0061281142,-0.1640412956,-0.669023633,-0.2725984752,-0.3042469621,0.1338291615,-0.2729791403,-0.1675990075,0.0596722402,-0.1618382335,0.0142350011,0.1734488904,0.1514361352,-0.0215064287,0.0119573008,0.150385648,0.0146297365,0.1776951849,0.1265307963,0.1827682853,-0.1011585221,0.0592221878,-0.1364729106,-0.0564785674,-0.0916979313,0.3380533755,0.0930299312,-0.0017823929,0.2639890909,0.2979951799,-0.0363329686,-0.1717988253,-0.2675010264,0.246560514,0.0576315522,0.1198230162,0.0803304985,0.6318507195,0.2055416703,0.1752791852,-0.150716871,-0.0701785162,0.0658943951,-0.0693151802,0.0188845173,-0.190381512,-0.0143460678,-0.0719314367,-0.3799416125,-0.2754399478,0.0968471766,-0.1404148936,-0.0746427327,-0.0370618328,0.0086293966,0.0436806567,-0.2274768203,-0.0524387397,-0.0882400125,0.4275676012,0.1327891052,0.139078632,-0.0748876482,-0.0587312877,0.1260323972,0.1799938828,0.1219586879,-0.3082248867,-0.2348091751,0.0278907008,-0.2218738943,0.2311760783,0.3654352725,0.5237157941,0.1095001772,-0.4189083874,0.0391259268,0.1244078428,0.5819937587,-0.2301923931,-0.0308651738,0.1820007116,-0.158254832,0.0998644531,-0.168094635,-0.2026173323,-0.029728381,0.2626563609,0.4312624633,-0.0818835497,-0.0455454998,0.5021079183,0.2213865966,-0.0002616506,-0.2003735304,-0.1444071531,-0.3349117935,-0.2491312474,-0.0207623672,0.0515861511,0.1574812829,0.0514002778,0.3127942979,-0.0301315226,-0.3142569661,0.1404965967,0.2325978875,-0.0428324528,-0.0776599124,0.1895497143,0.0559758134,0.1787735671,0.0804946572,0.2318656892,-0.4986412227,-0.6792055368,-0.0838990584,-0.1733059734,0.4837495685,0.2373588234,-0.2355024964,0.2386443764,-0.0622689351,0.2222546637,-0.1055742428,-0.0054897927,0.3826481402,0.0841321126,-0.3165595829,-0.5406337976,0.3414804935,0.0795637742,-0.0830515847,0.0900762975,0.3665972054,-0.4676115811,0.4190027118,-0.1272712648,0.539450407,0.1711793393,-0.0342090391,0.4204019308,-0.1584545076,-0.0678915158,-0.0687830672,0.1058086008,-0.3747582436,-0.2032316327,-0.0629145205,0.0981257334,0.445771724,0.0439912751,-0.023987921,0.3131878078,-0.0377460159,0.3181696236,-0.0129596712,-0.0236804876,-0.1248432323,-0.2890895903,0.031416703,0.0170258172,0.0585119203,-0.3307894468,0.3468234837,0.1850599647,-0.0445544347,-0.0068801297,-0.1796743274,-0.0920961052,-0.0396208726,0.2621534169,-0.070273675,-0.2068292499,0.2561681271,0.2360015213,0.3455247581,0.1204388887,0.2218238413,0.2901317179,0.0536015704,0.3150189221,0.0510928147,-0.0520142913,0.3549725413,0.2559574544,-0.1892668158,-0.1287592649,-0.0970653892,-0.1592885107,-0.1369861662,0.004530733,0.3254045248,-0.712518096,-0.2792513967,-0.1463442296,0.0076214471,-0.2057145387,0.0702228099,-0.4360507131,-0.321126461,0.0423506647,-0.2473160177,-0.5409811139,-0.0410532616,0.2723687887,0.0928785577,0.4236688912,0.6302575469,0.074452199,-0.118960768,-0.0769284368,0.2815288007,-0.0017072947,-0.237287432,0.3587244451,0.2491544485,0.012526351,0.1421516538,0.0894049779,0.0139233461,0.1374352425,-0.1794731617,-0.2946612239,-0.2364348024,0.0364805385,0.2371767908,0.0985815376,0.182802856,0.2921203077,-0.4138016701,-0.1671654135,-0.1980048865,-0.0679664761,-0.0445374101,0.3395097256,0.2307618111,0.0757984892,-0.0225777905,-0.169207111,0.2093006372,0.3660119474,0.3369351923,-0.1701856405,-0.3115433455,0.1045251191,0.0879617333,0.3199940622,-0.2685281038,-0.0645283014,-0.1654905379,-0.2432912886,0.1397914737,0.4403597713,0.009370368,0.6279401183,-0.2326393574,0.0618086569,-0.0221137647,0.197350949,-0.3701749444,0.1364082247,-0.177895233,0.3094801903,-0.0463115312,0.1947893798,-0.3871267736,-0.1407456994,-0.1017042398,0.0919170305,0.162463069,-0.2329881787,0.0275496356,-0.2276391536,0.3421429396,0.5766685605,-0.0366016701,-0.3406235278,0.260145992,0.0938546285,-0.2149893343,0.2604703605,0.1313986778,-0.3491407931,0.2399219275,0.2198221982,0.02382572,0.2739397883,-0.4158225358,0.0962316617,0.4593583643,-0.2312578559,0.0933077112,0.747166872,-0.0716879517,0.3275974691,0.2411202639,0.1174242198,-0.1474771351,0.5122222304,0.2956359982,0.3313479424,0.1835952997,0.0873710364,-0.1988353878,-0.0440695621,0.0718020946,0.0233148132,-0.2806161344,-0.0010325464,0.0523829348,0.0993889272,-0.1351288557,0.0063865436,0.0764461532,0.0787790418,-0.0853217691,-0.3726897538,-0.1807307601,-0.0773208365,-0.4351949096,0.0100043993,-0.0010135098,0.2239075005,0.5572535396,-0.0630996972,0.2330803126,-0.4022642076,-0.1216403991,-0.0671678632,0.3376493752,-0.4362981319,0.0450070947,0.196776256,0.0978504643,0.3461167812,-0.273642391,0.367490381,0.0158707518,-0.0313476436,-0.3278319836,-0.2703055739,-0.1479447186,0.258428216,0.4770182073,-0.2473756522,0.1601275951,0.3426007032,0.0708493665,-0.0214186683,0.3009721637,0.2921059132,0.1484791785,-0.3536863327,0.627040565,0.0342792906,-0.1725035012,-0.0649004653,0.0291266907,-0.2207119912,0.0153460484,0.4927920103,-0.0857059211,0.2625669837,0.0981514975,0.0294042714,0.2018343657,0.3894335926,0.2932920754,-0.4533614516,-0.0536827073,0.165880844,-0.3433842361,-0.1413582265,0.1090332642,-0.1523897201,-0.104317762,-0.2610399127,0.100369066,-0.0362396799,0.0448415726,0.1843632162,-0.2561531663,0.238036558,-0.0057349671,-0.3758705854,-0.0841977596,0.0209682155,0.0370429195,-0.3455042839,-0.0190412477,0.4193592668,-0.1062776744,-0.2612063885,-0.1117050424,0.3965735137,0.0212082788,-0.0702146515,0.2049697191,0.5175044537,-0.2131402194,0.1191594452,0.0022142441,-0.3588138819,-0.4568068981,0.4172988236,-0.0768352598,0.3570566773,0.037399061,0.0126483692,-0.0493573882,0.103703931,-0.2392890453,0.0344574749,-0.4673057795,0.0310813338,0.2342381328,0.0649655536,0.1397208273,0.2943555415,0.1055681333,0.4483832717,-0.124229677,-0.3101114631,0.5607996583,-0.3047522008,-0.6528941393,0.1209915355,-0.1744092405,0.3874828219,0.0959179625,-0.6975602508,-0.1059781834,0.3279694021,-0.0964661986,-0.2087817937,0.177573964,0.2100795805,-0.0227706879,0.02478984,0.1218380257,0.0436747037,-0.1834020466,-0.0161779039,-0.0734406784]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2021","title":"Interactively doing save_to_disk and load_from_disk corrupts the datasets object?","comments":"Hi,\r\n\r\nCan you give us a minimal reproducible example? This [part](https:\/\/huggingface.co\/docs\/datasets\/master\/processing.html#controling-the-cache-behavior) of the docs explains how to control caching.","body":" dataset_info.json file saved after using save_to_disk gets corrupted as follows. \r\n \r\n \r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/110568474-ed969880-81b7-11eb-832f-2e5129656016.png)\r\n\r\nIs there a way to disable the cache that will save to \/tmp\/huggiface\/datastes ? \r\nI have a feeling there is a serious issue with cashing.","comment_length":19,"text":"Interactively doing save_to_disk and load_from_disk corrupts the datasets object?\n dataset_info.json file saved after using save_to_disk gets corrupted as follows. \r\n \r\n \r\n![image](https:\/\/user-images.githubusercontent.com\/16892570\/110568474-ed969880-81b7-11eb-832f-2e5129656016.png)\r\n\r\nIs there a way to disable the cache that will save to \/tmp\/huggiface\/datastes ? \r\nI have a feeling there is a serious issue with cashing.\nHi,\r\n\r\nCan you give us a minimal reproducible example? This [part](https:\/\/huggingface.co\/docs\/datasets\/master\/processing.html#controling-the-cache-behavior) of the docs explains how to control caching.","embeddings":[-0.0748554915,-0.1469714046,0.0561726168,0.775801003,0.2905217409,0.3213843107,-0.2317185104,0.1311905235,0.1942945272,0.1316702813,-0.150269568,0.066835992,0.2389940321,0.2033745795,0.1379984468,0.2394488454,0.4092518687,-0.102217637,-0.2969146073,-0.0024894476,-0.0571397915,0.201320827,0.2392439842,-0.0475083664,-0.370827347,-0.1665406972,0.0865426436,0.3370094299,-0.0457871743,-0.3307515681,0.1703151166,0.2500181794,-0.2806940675,0.3997516632,-0.0001322348,-0.278096348,0.0654913113,-0.1120894477,-0.4807301462,0.3449938595,-0.2146836668,0.1056199148,0.1832559109,-0.051474601,0.018076079,0.0305722896,-0.0448686555,-0.2771779299,0.4946137369,-0.141809091,0.0099722333,0.2939893603,0.0540788174,0.170543164,-0.0891119391,0.352006942,-0.1730331928,0.247376591,-0.0567087792,-0.0422368832,-0.1335352063,0.3271785676,-0.1942922324,-0.221531406,0.4750732183,0.1670784801,-0.2261087745,-0.1015610918,0.1416803747,0.0300045144,0.2404650599,-0.4413738251,-0.4901424646,-0.409457624,-0.1626343131,-0.2622802854,0.4829567075,0.1149294302,0.0237637423,0.2698065937,-0.6267800331,-0.3822417259,-0.0752123967,0.1046850234,0.2587689459,-0.6426105499,-0.2962203026,-0.0202892181,0.0960970968,-0.0269153994,-0.2518183887,-0.2842938304,-0.4094386399,0.2043643147,-0.2804290354,-0.1706220955,-0.0365125388,0.6992857456,0.3096770644,0.0266730599,0.1539710015,-0.0268790461,-0.3828503788,0.0570360385,0.4450926185,0.1841095984,-0.1510336846,-0.3038631976,0.1937832385,0.2487015128,0.1934034079,-0.1279371232,0.6235292554,0.0654036105,0.3158103824,-0.2794091403,0.1027026176,-0.426775068,-0.158187449,0.2238548249,0.1136852205,0.1270822883,0.1518256515,0.2743179798,-0.2032476068,0.2388757765,-0.1741434634,0.2308161557,-0.1112838238,-0.1644691378,-0.2411801964,0.0200549606,-0.1028403789,0.4697910845,0.3572986126,-0.2158576995,0.1555408388,0.1488798708,-0.2194773406,-0.2177105844,-0.1413962096,-0.1730891615,0.2110491693,0.2001749575,-0.0784013942,0.261449188,0.0442139916,0.1813435405,-0.1044088379,-0.016748948,-0.3025808632,-0.1653333157,0.2371432632,-0.0093313269,-0.3137509823,-0.0204300024,-0.3833115101,-0.0406767949,0.373403877,-0.0606607869,0.2270976603,0.3761088252,-0.4323396385,-0.2126269937,0.1186989844,0.7272849083,-0.0777010769,-0.1809046268,0.3380313516,-0.0677055046,-0.0976461768,0.4955670834,-0.0461681187,0.1240394264,-0.1751255095,-0.0000537494,0.0921350494,-0.4036758244,-0.5119621754,-0.0031919228,0.2435677648,0.1814943403,0.0818844736,0.170397684,0.0276783817,-0.2080336809,-0.3021430373,0.2457524091,0.0571978651,0.1007413268,-0.2472885996,-0.2150123417,0.0834205523,-0.0444382764,-0.3568450809,0.4231317639,-0.0431624055,-0.0345389694,0.0292218924,-0.004833607,0.2021774203,0.3284560442,0.2372113466,0.1297816485,0.0238769706,0.0728912726,-0.4457373619,0.1206498072,0.0823695213,-0.2579562664,-0.0335220397,-0.2511535585,-0.0211948436,0.1351121068,-0.078266114,0.1193073615,-0.0983167291,-0.0323054828,0.2460309565,0.171704784,-0.2977082431,0.8645343781,0.06487602,0.2302663475,-0.5460054874,-0.0653682947,0.1276176572,-0.0912412032,-0.3207089603,-0.145103693,0.1751980335,-0.212664187,-0.0814768299,0.6701141,-0.0175038632,0.4155429006,0.0389863402,0.350451082,0.2606140971,0.1764128357,-0.0385825448,-0.011305158,0.1953107566,0.0030963675,-0.2079170346,0.2323743105,0.0401252396,-0.0596024506,-0.05956278,-0.1137470901,0.1619164348,-0.1846334636,0.0205509271,-0.3181634247,-0.0380241275,-0.2064877599,0.2917069495,-0.0238008369,-0.3807862103,0.1157953665,0.2447306663,0.0247230623,0.0320941396,-0.0001367864,-0.0198976006,-0.3366125524,0.0410211496,0.2540482879,0.5070416927,-0.0893044025,0.0645957738,0.0287317503,0.0835928917,-0.0630104914,0.1469515264,0.0372381955,0.07628382,0.1165311262,0.0270344745,-0.0039836271,-0.4554965198,0.4673797786,-0.1203905419,0.1310746521,-0.4288453162,0.3027190268,-0.0754572377,-0.1827237159,-0.0489479601,0.0194734093,-0.3586302698,-0.2553177178,0.1200568601,0.7634333372,0.0096856477,-0.0343590416,-0.0973856151,0.6595603228,-0.1374249607,0.1637319922,-0.5104157925,0.0852933675,-0.1150370985,-0.2006239742,0.1317938268,-0.2283476293,0.3360621631,-0.2017661035,-0.1169518903,-0.3274325728,0.1613985896,0.1672955453,0.152726993,0.1604951024,-0.030452121,0.1365007907,-0.0491777658,0.2691923082,0.0262319092,-0.2377795428,-0.2840476334,0.0405837484,0.1341718137,0.038881477,-0.1139045432,0.3498009145,-0.0329399891,-0.0541743338,0.2562649548,-0.1604389846,0.0446980335,0.1442906708,0.0514641702,-0.1174157262,-0.2412376255,0.09536919,-0.3460167646,-0.8206087351,0.3271396756,0.0187050141,-0.2429233193,0.1773413271,0.4334367812,-0.0686252937,0.2557433546,-0.4300456047,-0.3236347139,-0.1416156739,-0.1264070123,-0.2315827906,0.0207725018,0.111052677,0.107680954,0.0291786287,-0.0460481755,-0.2006755471,0.2317097783,-0.0316279344,0.4652905166,-0.1587823778,0.2857140899,0.1445428133,0.1652269661,0.0474158451,0.0871285126,0.5161062479,0.1988771558,0.7607293129,-0.0900925919,-0.1815169305,-0.341968298,-0.287255615,-0.1473704278,-0.0324973762,0.2255665362,0.2973240912,-0.0842176452,0.0380383022,-0.1443546116,-0.2730962038,-0.127672255,0.0462880135,0.1949003041,-0.0110854516,0.0630232543,0.0096932799,-0.2521271408,0.1720574498,0.3098351359,0.3786351085,0.0557169579,-0.3670054674,0.0577390566,-0.2918662727,0.0747630075,-0.3715940714,0.2076220661,-0.1661169082,0.0529626384,0.1380390227,-0.1151386872,1.0467971563,0.0593487211,0.2043457031,-0.1869152486,-0.3095416725,-0.0777510628,0.06915842,0.016184194,0.2611722052,0.0595148392,0.616232574,-0.150195688,-0.4283818305,-0.0670605302,0.2121829391,-0.3520036042,-0.1192596257,-0.0290309507,-0.0092236325,-0.1950492263,-0.1264072657,-0.0532119051,-0.0326082706,0.0709035695,0.1760476828,-0.1867092252,0.1301402003,0.0816661492,-0.0664822459,0.4074722826,-0.0674883723,0.1242029518,0.0730616152,0.2456011921,0.1384438574,0.2380992919,0.199892506,-0.1974647194,-0.1380946189,0.221522674,-0.0196408965,0.3996655643,-0.1574790925,-0.0589963347,0.1405600309,-0.1138249487,-0.3557964563,0.1030612588,0.0660834312,0.0492118225,-0.3060366213,-0.0036101236,0.2606483996,0.1160129905,-0.0404053181,-0.1814965904,0.1962471008,-0.3767836392,0.2850159407,0.2217984051,1.1114574671,-0.0965870246,0.4467279911,-0.0190250166,-0.1858020425,0.4888450503,-0.4004942179,0.0697797835,-0.3076687753,0.070544444,-0.238184616,-0.0994823053,0.0114016663,0.0542053804,-0.0264039803,0.3500372469,-0.3125452697,0.3876651525,-0.2943257391,-0.0302391537,-0.2995473146,-0.4565078616,-0.154295668,0.0095066987,-0.1227840558,0.0694232583,0.0924300849,-0.1352423429,-0.0731394142,0.0310234763,-0.075594902,0.0435001962,-0.1127257124,0.0468673632,0.1002518609,-0.1901458949,-0.180704236,0.4376677871,0.5001234412,0.0209765024,-0.2449200451,0.0116410563,-0.2824941576,0.0253807921,-0.150049895,0.1342827976,0.2004396915,0.0859990716,0.1513374895,-0.1590439379,-0.0797578618,-0.1197037622,-0.0414700955,-0.0937318727,-0.1113537028,-0.4145846367,-0.4675276279,-0.0541545562,0.2668151259,-0.1406589299,-0.0024489278,0.3070694804,-0.0184366778,0.0786455423,-0.3082565665,-0.1136912704,-0.1069941223,0.6052975655,-0.0238780659,-0.174639672,0.2837362587,-0.2377168387,-0.0794075727,-0.1719006151,-0.1477000117,0.192336455,-0.4884540141,0.0662492663,0.0631099045,-0.0760516301,-0.3701545596,-0.1271535903,0.0942049623,-0.1351930648,0.0251705889,-0.185065791,-0.4053344131,0.0911817551,0.4046017826,0.2675650716,0.1915950179,-0.2586055398,-0.3346121907,0.0083083343,-0.1136094257,0.0293384679,-0.0723339394,-0.1758124083,0.4042057991,0.1595693082,0.3628733754,-0.3410280645,-0.1344439536,-0.0828421786,-0.2898585796,-0.0524391197,-0.0671124384,0.2208081633,0.1286533475,-0.2424934655,-0.0785709471,-0.0933793187,0.1115616187,-0.2209191769,0.1932240576,0.3436004221,0.0891949013,0.0913301557,-0.0406274945,0.3802583516,0.056800738,-0.1113144457,0.1512752771,0.1528163701,-0.0399671793,0.0099178348,-0.3080691099,-0.2685752809,-0.2312733531,0.1376548856,0.3123863339,0.0286119822,0.0696548596,-0.2291850448,-0.1985617131,0.1794015467,0.4456974268,0.3498789668,-0.1561807841,-0.1637626141,0.2775178254,-0.0343011208,0.2283119112,-0.2759074271,0.3110443354,-0.1442942619,-0.2181528658,0.219491303,0.1457749158,0.0894240364,0.1508738697,-0.0159936361,0.5966019034,0.1635359228,0.1562797874,-0.0331899859,-0.175557524,0.2858833969,0.2703327537,0.2711171508,0.0543092415,0.2297401875,0.1829445213,0.2589280903,-0.0046060216,0.1549368054,-0.2851569057,-0.3397421241,0.1565559059,0.1272260994,-0.1917471886,0.2534337938,0.0077063786,0.5218217969,-0.03363958,-0.5268293619,-0.2592496276,0.3149910867,-0.2393060029,-0.1920985729,0.15738976,-0.0409391038,0.3146927953,0.3009059727,-0.1844608635,-0.2195094973,0.3485593796,-0.036061734,-0.0218561608,-0.1828766465,-0.0929949731,0.1665177792,0.1911558509,-0.0352355652,-0.0829480737,-0.1525166482,0.0117238499,-0.2291654497,0.3340147138,0.0944073051,0.1614550799,0.2619745135,0.1246223152,0.0411862582,0.0491621308,0.1089008301,0.1243213117,-0.1317133158,-0.2439180911,0.1301480532,-0.1043872386,0.0813058987,-0.0305601452,0.0009524059,0.5091177821,-0.186443001,-0.1353569627,-0.1505175531,-0.0499468111,-0.1239094287,0.0432748869,-0.2306683362,-0.0394361354,0.353020072,-0.0909419209,0.0934866592,-0.0840717629,-0.0211620741,0.138538599,0.6743152142,0.311059773,-0.263690412,-0.178188026,0.0558290854,-0.3784157634,0.1127924547,0.1490972191,0.2689667046,-0.2375582904,0.0925046206,-0.2027710229,0.1909042448,0.0163916498,-0.2266663909,-0.0408781134,0.2388919294,-0.0339324512,-0.1032305136,-0.0601591095,0.0945415124,0.0901064873,-0.1053362936,0.4479230642,0.2539941072,-0.0884993598,0.1033507138,0.2387168258,-0.6483798623,-0.0095176343,0.310865134,0.3298609257,-0.1611239165,0.0723904297,-0.1324669272,-0.0474595055,-0.1372129619,-0.155209735,-0.0879194066,-0.0672582686,0.2129001319,-0.3041414022,-0.0716676712,-0.2815505266,0.2846009135,0.1085600406,-0.1737365276,-0.1152777597,0.197837323,0.1536780596,-0.0098530194,0.3255874217,0.2880610824,-0.1077119336,0.4771923423,-0.5032348633,0.1247443408,0.3944431543,-0.1870740205,-0.0229148027,0.1646422148,0.2516717017,0.4809964001,-0.5691509843,-0.6756888032,0.0115081724,-0.0385030247,-0.1943022311,-0.2786025703,0.5799127221,-0.4410836101,-0.1232316643,0.0045277523,0.6662654281,0.3753519356,-0.1191736758,0.3322765529,-0.0927177519]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2012","title":"No upstream branch","comments":"What's the issue exactly ?\r\n\r\nGiven an `upstream` remote repository with url `https:\/\/github.com\/huggingface\/datasets.git`, you can totally rebase from `upstream\/master`.\r\n\r\nIt's mentioned at the beginning how to add the `upstream` remote repository\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/987df6b4e9e20fc0c92bc9df48137d170756fd7b\/ADD_NEW_DATASET.md#L10-L14","body":"Feels like the documentation on adding a new dataset is outdated?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/987df6b4e9e20fc0c92bc9df48137d170756fd7b\/ADD_NEW_DATASET.md#L49-L54\r\n\r\nThere is no upstream branch on remote. ","comment_length":32,"text":"No upstream branch\nFeels like the documentation on adding a new dataset is outdated?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/987df6b4e9e20fc0c92bc9df48137d170756fd7b\/ADD_NEW_DATASET.md#L49-L54\r\n\r\nThere is no upstream branch on remote. \nWhat's the issue exactly ?\r\n\r\nGiven an `upstream` remote repository with url `https:\/\/github.com\/huggingface\/datasets.git`, you can totally rebase from `upstream\/master`.\r\n\r\nIt's mentioned at the beginning how to add the `upstream` remote repository\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/987df6b4e9e20fc0c92bc9df48137d170756fd7b\/ADD_NEW_DATASET.md#L10-L14","embeddings":[-0.0541970842,-0.342902869,-0.0709576011,-0.2234128565,0.1322251409,0.0014155612,0.1212346554,0.0149749089,-0.4329250455,0.1663282961,0.0103836032,-0.0504560061,0.1446346939,0.2027385086,0.0767325982,-0.0622126684,0.2159233242,-0.1112824976,0.0094948607,-0.3877199292,-0.0959463716,0.0407857597,0.0525698029,-0.4642565846,-0.0445150174,0.0981033295,-0.1298272461,0.0079100905,-0.4901851416,-0.4163923562,0.3651945889,-0.0329049341,-0.0429880321,0.476279676,-0.0001129406,-0.0150297079,0.3386003077,0.0055787019,-0.1953437775,-0.2580168545,-0.431886524,-0.4034749866,-0.0033092254,0.2223095149,-0.1875630021,0.0652669817,0.1741567552,0.0976460204,0.1949372739,0.1445427239,0.2238371819,0.2049379796,0.2905164659,-0.4031758904,0.1369868666,0.3619071245,-0.2284917831,0.6680122018,0.2545554936,0.4873504639,0.0701367334,0.14298594,0.3143943548,-0.1501743644,0.2716187537,0.0007992763,0.0826397389,-0.1987708509,0.0030304641,0.1926567405,0.6094884872,0.122026749,-0.5760812759,-0.1504706144,-0.0158565044,-0.2908196449,0.0060971137,0.1517653316,-0.0127051836,0.1186981946,-0.279786855,-0.5576728582,-0.1758411378,0.3202894926,0.2198292464,0.283773452,-0.158690542,-0.0596121699,0.1148347929,0.158334896,-0.304923892,0.2595874071,-0.1861921698,0.0499552265,0.0524859428,-0.1112526357,0.017377466,0.1793416291,0.1147997528,0.4388834834,-0.2054879665,-0.2199703455,-0.2314783484,0.0753109008,0.2165307552,0.0756217539,-0.0312717557,-0.079086341,0.1986977309,0.0602077693,0.3548210561,-0.2259039581,0.0818692148,-0.0484484732,-0.3386690617,-0.1433110684,0.2231083512,-0.3168765008,0.1877762079,-0.2154599726,-0.0260353778,-0.0105401417,-0.0132227642,-0.0294341352,0.0110923816,-0.0049992367,-0.1991396695,0.2391584814,-0.1727277935,-0.0976906419,-0.2451280355,0.0471845642,-0.3421871364,-0.0774903446,-0.0103982724,-0.5276147127,0.4298349321,0.2547621727,0.2393214256,0.2254777998,-0.4373151064,0.1930025369,-0.0207110047,0.2717285752,-0.0286174472,-0.0238103941,-0.0957150683,0.0014506789,-0.0622847714,-0.3066673875,-0.2370884866,-0.3535678983,-0.3002385795,0.1684078276,-0.2077749521,-0.0839427635,-0.4613353312,0.2800756693,-0.2034398466,-0.1949636489,0.2160711884,0.311763078,-0.1525012851,-0.2019260973,0.3219577074,0.3567749262,0.1260176301,-0.2448609173,-0.1671650261,0.0270011984,0.0185945444,0.231768325,-0.196821481,0.0661111996,-0.1416386366,-0.3127050102,0.1076084748,-0.3901762068,-0.2116607726,0.1747912019,-0.1828833818,-0.2197279781,-0.031187484,0.1089667529,0.0628408641,-0.120844841,-0.152158618,-0.0259998683,-0.2195287794,-0.3192836642,-0.148110047,-0.1362634301,-0.4844519496,-0.084295705,0.0573401675,0.1630784124,0.3549596071,-0.0026099316,-0.1355063021,0.0839673728,0.0962489396,0.4097358286,0.6342180371,0.2227144986,-0.148292616,0.0710422769,-0.4192636013,0.1008315533,0.1169343814,0.3873185813,0.1521229893,-0.0850232095,-0.044359196,-0.1968233734,-0.0004862121,-0.4259001911,0.0244929101,0.1884888411,0.213853538,0.2234975249,-0.2566771209,0.095883742,-0.0294915326,0.1528879702,-0.6297127008,0.1088110358,-0.1664846092,0.0939432904,0.0237173997,0.2124968767,0.0070326482,-0.2777433097,0.0966232345,0.3237044215,-0.2774696052,0.2817159891,0.3697200418,0.1944118738,0.3049992621,-0.1333852112,-0.0943499804,-0.1449738592,-0.0889540166,0.1931201816,-0.2182801962,0.0946670026,0.0110948728,-0.0022163887,0.2311751992,0.1143167168,0.2847479582,0.0765318498,-0.051417511,-0.3194243312,0.1716019809,-0.0226421971,0.244502753,-0.0863726661,-0.1692863256,0.2999427319,0.11817801,-0.1917345077,-0.2504562438,0.2126444876,0.0553066768,-0.034940321,0.1509420872,0.3782607019,0.2094811201,0.3140182197,0.2590820789,0.2057402134,0.2297553718,-0.2771490812,0.0332849883,-0.0133661879,-0.0976354778,0.0102001075,0.2610309124,0.0893155113,-0.2038084418,0.261734277,-0.1053256765,0.2662078142,0.039454136,0.1391785294,-0.2544880807,-0.1987805665,-0.2773379087,-0.4256397486,-0.4906124175,-0.2860661149,0.050332889,0.2227463573,0.0436121225,0.0481662527,0.180276975,0.0652741268,-0.3773147464,0.074153997,-0.0229747146,-0.2716953158,-0.0949644372,0.1281139851,0.0593441203,-0.1130554304,0.537050724,-0.321662873,-0.0148123214,-0.551178515,-0.7157526612,0.1464820504,-0.1888093501,0.2367566079,0.3718587458,-0.0906212851,-0.0535561927,-0.121469602,0.1372005194,-0.3795280755,-0.2920962572,-0.2372146398,-0.1862315983,-0.0549148843,-0.1697185636,-0.5078220367,0.0070132078,-0.3114489913,0.6385708451,-0.1654120982,-0.059781149,0.3390489817,0.0503287353,-0.1282752007,-0.2300447226,0.0783430934,-0.1004794165,-0.3781207502,0.1260284483,-0.3046660423,-0.3719869554,0.3863863051,0.1393262148,0.4151358902,-0.1058664173,-0.5360229015,0.0873725712,-0.1337156445,0.1259727925,0.2594906986,0.3358611763,0.4246498942,-0.297468096,-0.0998068675,-0.1367073655,-0.1937779188,0.025569981,0.0290752817,0.3333816528,-0.1351306438,0.6006179452,-0.0499070957,0.6402546763,-0.0982800499,-0.0154752545,0.1143738925,-0.1493915617,0.6323479414,-0.3316772282,-0.3233627379,0.1364963353,0.03455244,0.006887034,0.1125793383,0.4248794615,0.0923247263,-0.2468123883,-0.3051476479,-0.1002643779,-0.1460124105,-0.1558082402,-0.138960287,0.0273379907,0.1567372829,-0.1699365973,-0.1294645965,0.0594990738,-0.190482378,0.4983603954,0.3119199276,0.1762958765,-0.1139164716,-0.0200239122,-0.328776896,0.0223799665,0.1333619356,0.1110154316,-0.0989659727,0.0227772184,0.0639506057,0.0045187706,0.0821684524,-0.2684128881,-0.0876869559,-0.1755539179,-0.3973679841,-0.0734746754,0.0761880353,0.1294651181,0.2193843573,0.5229238868,0.1918776631,-0.437140584,-0.2494710684,0.4528583288,-0.1036303937,0.0242378153,0.0774803981,-0.3639128506,-0.1927822083,-0.1417672634,-0.1561278105,-0.1499822289,-0.2423300147,0.2002395689,0.209741056,0.0804550126,0.1307310313,-0.0218578298,-0.110086441,-0.0781899244,0.1754275262,0.1059264168,0.1379511505,0.2865147591,0.5018665791,0.6798103452,-0.1456864327,-0.0953032076,0.1463418305,-0.102347821,0.3624456823,0.2534082532,0.3015636504,0.3917582035,0.3839931488,0.0469166934,-0.137991339,0.0368472151,0.1982233822,0.0981242955,-0.1525485367,-0.5134531856,0.3481762111,-0.2456048429,0.1051675007,0.2561163902,0.6245746017,0.1192331612,-0.1702460051,-0.1379127502,0.7449843287,0.271157831,0.1839510798,0.2813600004,-0.2067982554,0.6587557793,-0.0252559092,0.1206002682,0.0192948096,0.1490716189,-0.0277019348,-0.0383049846,-0.1677702665,-0.3307249546,0.1468951553,0.3333962858,0.20542036,0.8445014358,0.1578492373,0.2077130079,0.1023798212,0.101057671,-0.3259067535,0.1762962341,0.1065439284,0.1627651304,-0.2111749947,-0.3545359671,-0.2533884645,-0.2073817849,-0.2653510571,-0.1439163536,0.1521336287,0.1510456204,0.1530211121,-0.4044550955,0.1483870298,-0.1124776378,0.6150485277,0.0065370882,-0.4030731916,0.0843151584,-0.0826877356,-0.251729846,-0.0360669307,0.0323889554,0.044734031,-0.2364155501,-0.3017362058,-0.1546248198,-0.0294619426,-0.6098447442,-0.1334955543,0.0929760709,-0.4606899917,0.0419835262,0.1105692014,0.3218734562,-0.0663348734,-0.2480003536,0.0948794782,0.0347454846,-0.1556085348,-0.4615311623,0.2726500034,-0.0990650877,-0.3219062686,0.4335672557,-0.0901327655,-0.1981205195,0.0350437239,0.1176131442,-0.1752160788,-0.0787028968,-0.25997141,0.4198158979,-0.6655976772,-0.07201989,-0.0602664649,0.1404733658,0.2696124911,0.0201429855,0.1931986511,-0.1554909348,-0.0399052016,-0.0657762587,-0.4366452992,0.4004332423,-0.4324063361,-0.0296746027,-0.0345092304,-0.0262647755,0.260899663,0.0994910374,-0.3136251271,0.1021840572,0.0710777417,0.1782092899,0.1231846362,-0.1891827881,0.366171211,-0.3087970316,-0.0154167069,-0.2810227573,-0.3800630867,-0.1121125147,-0.0706323758,0.105683133,-0.0388632827,0.0025967187,-0.0108118048,0.0262662992,-0.2635314167,0.06493745,0.1543997079,0.2039617002,0.0449462272,-0.080502063,0.2132820338,0.1636173874,0.0024138277,-0.0155849159,0.2798169851,0.2435316145,-0.0116383908,-0.0579942614,-0.1243355498,-0.0605255887,-0.0442779809,0.1273370683,-0.0094818119,0.0360433348,0.2648443878,-0.0183523111,0.2034748346,-0.140474081,0.4461747706,-0.0409633182,0.2287865281,0.2913943827,0.1161270738,0.1003620848,-0.1521701664,-0.0122982655,0.1328453422,0.1795765609,0.1124673858,0.3269707561,0.51926893,-0.0869516283,0.147543624,0.2571843863,-0.0131181078,-0.025328923,0.3326092064,0.3951198757,0.0955789834,0.1072608307,0.0986685902,0.2187076211,0.1012298167,0.0415246226,-0.0990245566,0.2544952035,0.0086003132,-0.0534024723,0.2118666768,-0.0729229003,-0.1415666789,0.0891044661,0.3963585198,-0.1169433892,-0.0006468737,0.4232811034,0.081231609,-0.1144535691,-0.1230816394,0.2472333461,0.0531452931,-0.1036687642,-0.1387124211,-0.0699665919,0.0223778598,0.0309686549,0.0588695183,-0.1924860924,0.2392429113,-0.0461436994,-0.1177368313,-0.1222555935,0.0664885566,-0.020142518,0.2105909735,0.1161599532,0.2627842128,0.3309721053,-0.3042774796,0.1512636095,0.4403746724,0.3446225524,0.0783987716,0.300809741,0.2424588352,0.1001098827,-0.1045982093,0.0281086639,0.4148676991,-0.0387382135,0.1191426218,-0.0958628282,0.1689063311,-0.0897561088,0.5723360181,-0.1623557806,0.1140350252,-0.0797849074,0.4220880568,-0.1311379224,0.0350482017,-0.205966115,0.3387073576,-0.3323359489,0.1330703795,0.1188881025,-0.0054402314,-0.1003242582,-0.084421508,0.1077454463,0.1567731947,0.4633106887,0.5227739811,0.0973715186,-0.0121583324,-0.3283228576,-0.3530429304,-0.0891450271,-0.1584777981,0.1687248945,-0.1021850556,0.1534795165,0.1190977246,0.046624355,0.1040747911,-0.1226316988,-0.0831453353,-0.100363709,-0.0197657235,-0.1404772848,-0.0286275242,-0.0737159029,0.0457006507,0.2124543786,0.0754463896,-0.1354753673,0.0448204167,0.2117477655,0.348275423,-0.0373944268,-0.1236431301,-0.1607389897,0.1037760675,0.1800482869,-0.1206173301,-0.5101879239,-0.4385741651,-0.2309196591,-0.2537136078,0.333817929,-0.2255733311,0.4225814044,-0.2648857236,-0.2386316806,-0.3333469033,0.1711944342,0.1513841748,0.0108096078,-0.2621189058,0.2301332504,-0.1468754709,-0.0217889119,0.2312790006,0.1507332325,-0.099286437,-0.0273185428,-0.3974137306,-0.1334389001,0.5408654213,-0.3637286127,0.2908587456,-0.1150251329,0.0339612924,0.6897444725,-0.2622001171,-0.7218092084,0.1165167019,0.1182702929,0.1573733389,-0.2257873118,0.3257895708,-0.2251301557,-0.3165768683,0.0178391468,0.4571698606,0.1328653246,-0.3172670901,0.1752414107,-0.14566046]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2012","title":"No upstream branch","comments":"~~What difference is there with the default `origin` remote that is set when you clone the repo?~~ I've just understood that this applies to **forks** of the repo \ud83e\udd21 ","body":"Feels like the documentation on adding a new dataset is outdated?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/987df6b4e9e20fc0c92bc9df48137d170756fd7b\/ADD_NEW_DATASET.md#L49-L54\r\n\r\nThere is no upstream branch on remote. ","comment_length":29,"text":"No upstream branch\nFeels like the documentation on adding a new dataset is outdated?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/987df6b4e9e20fc0c92bc9df48137d170756fd7b\/ADD_NEW_DATASET.md#L49-L54\r\n\r\nThere is no upstream branch on remote. \n~~What difference is there with the default `origin` remote that is set when you clone the repo?~~ I've just understood that this applies to **forks** of the repo \ud83e\udd21 ","embeddings":[-0.1526438892,-0.3908179402,-0.056868434,-0.3803965747,-0.0535282604,-0.1234395429,0.3151813745,-0.0022337413,-0.3539056182,0.2694294155,-0.0092931604,-0.0466281921,0.3955129087,0.235126704,0.2467384189,-0.0680186525,0.3274163902,-0.1903973073,0.2409152389,-0.7290988564,-0.2180368006,-0.0492812134,0.1212639958,-0.2099992931,-0.1465394646,0.2308082134,0.0482762642,0.1814907789,-0.2902173698,-0.3672446012,0.4538211524,0.0617859401,0.0159863252,0.453630805,-0.000121437,-0.0352861173,0.4274776578,0.0357965268,-0.2610278726,-0.1529745609,-0.5940438509,-0.2314037532,0.0071179071,0.0715084374,-0.0290648807,0.1804949194,0.2290717065,-0.1103390157,0.0469161086,0.129533723,0.1420780569,0.191397205,0.0362158604,-0.3778528273,0.2821260095,0.4759959877,-0.2266931087,0.875222683,0.2417099923,0.5030210614,0.1716814339,-0.0652809516,0.1075375155,0.0430668779,-0.0373703577,0.1029734015,0.1750000566,-0.1549441665,0.162227869,0.4301557541,0.770275116,0.2360711694,-0.4754008949,0.0212695822,-0.0125339497,-0.3896008134,-0.0593696386,0.3835000396,0.0646640137,0.0951499492,-0.197941795,-0.3189469576,-0.3143386543,0.2806482613,-0.0658181906,0.414660573,-0.1366889775,0.0840012729,-0.0252129678,0.2557754517,-0.1660642475,0.0732245743,-0.215294838,-0.1190663949,0.015089591,-0.1759963036,0.0897787511,0.0872920007,0.071812354,0.5886428952,-0.3382132649,-0.1244905666,-0.2546327114,0.0834155902,0.1628780812,-0.0049647046,0.0689398721,0.1233234257,0.3434395194,-0.1844355911,0.3132756054,-0.1580080986,0.0624264888,0.1461664587,-0.2167900205,-0.1568918675,0.3482620418,-0.4060587585,0.2222378105,-0.1124550551,-0.1277005225,-0.2261945158,-0.1769113839,-0.0030167596,0.0541891828,0.1421895176,-0.4768745601,0.2858996987,-0.181034565,-0.1608571559,-0.2593808472,-0.1099007949,-0.1611193269,-0.0340505168,-0.0575041324,-0.4412372708,0.4918140173,0.1535722762,0.2645317316,0.3689874709,-0.117259115,0.0552305244,0.0956316367,0.3831594586,-0.1271001101,0.0321269631,-0.1134603843,0.1048264652,-0.1427420527,-0.02694376,-0.1989972591,-0.4819029272,-0.2253121734,0.0954170898,-0.2828347981,-0.2053881735,-0.6288796067,0.1489581019,-0.3456094861,-0.0665438026,0.3376171887,0.1509412527,-0.0143865338,-0.1780399829,0.0957168043,0.4086882472,-0.0168425534,-0.2962556183,-0.1248484701,-0.1975780427,-0.0173986685,0.2092630714,-0.3593125343,-0.0457934663,0.05381082,-0.098974064,0.0449748486,-0.4179222882,-0.154373914,0.0906203166,-0.2368871123,-0.2399967313,0.0398535579,0.081840232,0.2099033147,-0.1554738879,-0.1764334142,0.0391689762,-0.2171976864,-0.2206084877,-0.1783546358,-0.0482932329,-0.3847338259,0.1111883521,0.0142521476,0.1173519343,0.3190358579,-0.139579758,-0.2348851264,0.0622808672,-0.1315763742,0.1852912903,0.7095392942,-0.0590321422,-0.1003813222,0.2234230489,-0.5277835131,0.1685790271,0.1947854757,0.2524726093,0.3422012329,-0.0094505297,0.075596571,-0.0866315439,-0.1810861826,-0.1649609357,-0.0894197896,0.246516645,0.2977789342,0.1671322286,-0.5274182558,0.0090602934,-0.2239992619,0.1525299251,-0.4646700025,0.2709396482,-0.0408055633,0.149014622,-0.2467189282,0.0117413802,-0.0781420618,-0.2670853734,0.0479298122,0.2713980079,-0.1804299057,0.4189496934,0.4851288199,0.2931009233,0.3441980779,-0.24791044,0.1567442268,-0.0440133065,-0.1111493185,0.1980708987,-0.2691898346,0.09703587,0.1809622943,-0.0314294808,0.1452760249,-0.1158271804,0.2320694774,0.0150042111,-0.104176417,-0.3393951058,0.0519040562,0.217008099,0.2867290676,-0.0579628721,-0.1593402475,0.3818152249,0.0962785259,-0.572160244,-0.2647396922,0.227770865,0.0936385244,-0.2030418366,0.1316748112,0.5656302571,0.3525888026,0.1717997193,0.273319006,0.1977961212,0.09862829,-0.4801003337,0.1898612231,-0.0167470761,-0.1267503053,-0.0097042723,0.1698029786,-0.0199272987,-0.1559547335,0.0612970106,0.0982962549,0.2183147967,0.1269356906,0.096589759,-0.3688340485,-0.1766267717,-0.3448282182,-0.2950824797,-0.5318112969,-0.334975034,-0.0198540967,0.0183206033,-0.0965118632,-0.0817903355,0.1770522743,0.1709259003,-0.4175526798,0.0343733653,-0.030464638,-0.3974835873,-0.1017925814,0.0360544845,-0.1647347957,-0.0609871596,0.5170274973,-0.4658833742,-0.1691503674,-0.4035306573,-0.5375563502,0.1256886572,0.0598176308,0.1879414618,0.1697303802,0.1899940073,0.0890597031,0.0315885991,0.1639567167,-0.2339728177,-0.2296453565,-0.26270926,-0.2379183769,-0.1141394153,-0.0046677124,-0.4012688398,-0.1181753054,-0.200720489,0.4642678499,-0.1780476123,0.0379760936,0.2009680122,-0.1617183089,-0.0656944364,-0.2549103498,0.0839422792,-0.1440756321,-0.5398855805,0.0550481007,-0.2148223072,-0.1005781367,0.301027,0.0545756221,0.2800721824,0.0477071106,-0.5282685161,-0.010066594,-0.169036299,0.3068681359,0.5019596219,0.3753704131,0.5677351952,-0.4530457556,-0.0075602653,0.0147912567,-0.300311625,0.037048094,0.2082605064,0.1650034636,-0.048484914,0.4850202501,-0.2723739743,0.7536585331,-0.1300549805,0.0822992846,-0.0404143482,0.0084378608,0.5139191747,-0.407975167,-0.0811411291,0.230147332,0.1241429895,-0.0654326677,0.0256669186,0.3337323368,-0.0201790985,-0.172877565,-0.5013913512,0.0714676529,-0.1127291694,-0.0697223246,-0.0687469095,0.1283317357,0.1590294391,-0.1638230681,-0.0739360079,0.0784243718,-0.2701803148,0.4493615031,0.1999807358,-0.0080485763,-0.5328277946,-0.1079100445,-0.1444196999,0.0893221945,0.1260150671,-0.0015050224,-0.0078388648,-0.1675852388,0.1276485473,-0.0504658073,0.0177272167,-0.3045566678,-0.107779488,-0.2122401148,-0.4268776178,-0.0055092354,-0.1225965694,0.0550623089,0.1536028385,0.422129184,0.205597952,-0.2968251705,-0.2599084973,0.4703335762,0.0919473767,-0.0268705692,0.1649599969,-0.3371912837,-0.0162044428,-0.1141689196,-0.0482295528,-0.3885562122,-0.3450277746,-0.0113600409,0.2983252108,0.0268261395,0.1325591952,-0.2346941233,0.0812704414,-0.1878333241,-0.0034686504,0.0483656675,0.2260609418,0.1647786051,0.5362097025,0.6573815942,-0.1838724166,-0.2224896699,0.0636850372,-0.0694224834,0.2016168237,0.4448964,0.4241682887,0.2504865229,0.3098895848,-0.0499736853,0.0536903106,-0.0254458189,0.07952407,0.1270528585,-0.219223693,-0.3969988823,0.4532273114,-0.248435244,0.0308037363,0.2647315264,0.6452983022,0.2309067398,-0.161471799,-0.1040887162,0.8261780739,0.4181175232,0.2513293028,0.3437062502,-0.298166424,0.4382250607,-0.0093539152,0.0064925421,-0.0554458499,0.1779248714,-0.0186354816,-0.1022797301,-0.075701125,-0.1465042084,0.2845981419,0.185384661,0.0716180801,0.7963956594,0.1774626225,0.1313206553,0.2925502062,0.1426706761,-0.2062411159,0.0773949847,-0.0784640536,0.146626696,-0.1534097046,-0.119725436,-0.1770538986,-0.1667891443,-0.2664333284,-0.054639928,0.3031916916,0.0117743397,0.2971507311,-0.5078319907,-0.1206813976,-0.2457906157,0.6949744821,-0.0030186628,-0.1307381988,0.0745630264,0.1353504509,-0.0212771017,0.2675535381,-0.1156915426,0.0138353361,-0.2264951617,-0.546622932,-0.060535904,-0.0539434105,-0.5696589947,-0.1581389308,0.0199408848,-0.362567693,0.026528677,0.0525791198,0.4546355009,-0.0310121905,-0.1170661226,0.0229262505,0.1107155457,-0.1056679264,-0.3512170613,0.0517788418,-0.1526311636,-0.399830997,0.3213113546,-0.0364835635,-0.2209537029,0.0265421085,-0.0800675675,-0.1310902536,-0.081128642,0.045173876,0.3683877885,-0.7441094518,0.0638630241,-0.1749021858,0.0116595607,0.2676517665,-0.0153748691,0.0099121556,0.052878961,-0.1558082551,-0.012355783,-0.1575998515,0.2336529493,-0.3643558621,0.0906210989,0.1775608957,-0.3076362908,0.1613839567,-0.0623827837,-0.226440236,0.1388544738,0.0980685502,0.2257533222,0.1407101005,-0.0913136452,0.2690132558,-0.3536029756,-0.0267342478,-0.3960966468,-0.272575289,-0.1162827536,-0.2229146361,0.1896712333,-0.0910483301,-0.2563007176,0.1080815494,-0.1417185217,-0.1377775669,0.2155240625,0.2188667506,0.2175673544,0.0644463524,0.0149094844,0.2243240774,0.036061123,-0.1561956257,0.1475901604,0.2517767251,0.3375065029,-0.0272805411,0.0907971933,-0.1516242176,-0.1514412463,-0.2205789983,0.1333149076,-0.1322076917,0.160887152,0.1646528244,-0.0180427805,0.2100692093,-0.2077447772,0.2365709841,0.1006640494,0.0847128108,0.0704976693,0.1816988587,0.111027956,-0.0214974806,-0.1847308725,0.3753157556,0.2301634699,0.046896752,0.3382747173,0.4278505147,0.0108202221,0.460439533,0.1821481586,0.0776316226,-0.0883552507,0.644982934,0.5471997261,0.0260918736,0.2726807296,0.0170382187,0.2292101383,0.0071516279,-0.0280053839,-0.0036744841,0.1594765782,0.2348799855,0.002236279,0.3667227328,-0.027277628,-0.0535841323,-0.0611047074,0.1047352403,-0.141879186,0.2086808085,0.2416000217,-0.0101769138,-0.0855592266,-0.2205315381,0.1717692465,0.0124134542,-0.2053410858,-0.3689616621,0.0545773208,0.2000612617,0.2004273832,-0.1807766855,-0.2027132362,0.1707306802,0.0611332543,-0.1006785184,-0.0421070755,0.268745929,-0.2398380637,0.1068274528,0.2622804344,0.2268816829,0.4346130788,-0.3286220431,0.2640412152,0.4957016706,0.179323554,0.0315489955,0.3371671736,0.3460944593,0.3019033074,-0.1499713808,0.1991121471,0.2072295398,0.0444979258,0.1712990105,-0.1347510815,0.1457039714,0.0161226429,0.5959374309,-0.0484442338,0.0492675453,-0.1122324988,0.4181317687,-0.0545025468,0.1122143269,-0.2390368283,0.3924507797,-0.2974774539,0.2598190308,0.1190328449,0.0362642743,-0.2335348427,-0.0534400679,0.0490713082,0.011198448,0.5582712293,0.4555211067,0.0861450657,-0.0428805538,-0.1485629678,0.0097037889,-0.0047383877,-0.2231496274,0.0895552114,-0.160451889,0.075075753,0.094800204,0.2105299681,-0.0350965783,-0.1206276566,-0.146315977,0.0992914215,0.1164443046,-0.2136070281,0.0344924666,-0.0069063981,-0.1713659912,0.0482468493,0.0996371359,0.0567171164,-0.1007902175,0.170119971,0.1624995321,0.1479487717,-0.1819944382,-0.2239581496,0.1941969991,0.1853152514,0.1380940229,-0.3846302032,-0.2508637011,-0.1912682354,-0.4125622213,0.2045009434,-0.2136582434,0.3190742135,-0.088317126,-0.1272102296,-0.160396412,0.0490942113,0.0788586587,0.2051484138,-0.1112901941,-0.0165228266,-0.0758931041,-0.3586298227,0.2597531676,0.031591516,0.01492434,-0.1502949148,-0.3193619251,0.0743467361,0.410939604,-0.3343662322,0.3913282454,-0.3347210586,0.0273355432,0.5880268216,-0.1575365663,-0.4989510477,0.1742940843,0.1900559962,-0.1772028655,-0.3052076697,0.3734638393,-0.0330675803,-0.3052958548,-0.0091743683,0.4189300537,0.1355575919,-0.4480019212,0.018124979,-0.0638759285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2010","title":"Local testing fails","comments":"I'm not able to reproduce on my side.\r\nCan you provide the full stacktrace please ?\r\nWhat version of `python` and `dill` do you have ? Which OS are you using ?","body":"I'm following the CI setup as described in \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8eee4fa9e133fe873a7993ba746d32ca2b687551\/.circleci\/config.yml#L16-L19\r\n\r\nin a new conda environment, at commit https:\/\/github.com\/huggingface\/datasets\/commit\/4de6dbf84e93dad97e1000120d6628c88954e5d4\r\n\r\nand getting\r\n\r\n```\r\nFAILED tests\/test_caching.py::RecurseDumpTest::test_dump_ipython_function - TypeError: an integer is required (got type bytes)\r\n1 failed, 2321 passed, 5109 skipped, 10 warnings in 124.32s (0:02:04)\r\n```\r\n\r\nSeems like a discrepancy with CI, perhaps a lib version that's not controlled? \r\nTried with `pyarrow=={1.0.0,0.17.1,2.0.0}`","comment_length":32,"text":"Local testing fails\nI'm following the CI setup as described in \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8eee4fa9e133fe873a7993ba746d32ca2b687551\/.circleci\/config.yml#L16-L19\r\n\r\nin a new conda environment, at commit https:\/\/github.com\/huggingface\/datasets\/commit\/4de6dbf84e93dad97e1000120d6628c88954e5d4\r\n\r\nand getting\r\n\r\n```\r\nFAILED tests\/test_caching.py::RecurseDumpTest::test_dump_ipython_function - TypeError: an integer is required (got type bytes)\r\n1 failed, 2321 passed, 5109 skipped, 10 warnings in 124.32s (0:02:04)\r\n```\r\n\r\nSeems like a discrepancy with CI, perhaps a lib version that's not controlled? \r\nTried with `pyarrow=={1.0.0,0.17.1,2.0.0}`\nI'm not able to reproduce on my side.\r\nCan you provide the full stacktrace please ?\r\nWhat version of `python` and `dill` do you have ? Which OS are you using ?","embeddings":[-0.1566001624,0.0945401862,0.0029507412,0.0519755632,-0.1364501864,-0.259108901,0.4086020291,0.2240652144,-0.1059585363,0.2676886618,-0.0058867261,0.0788329914,-0.153766185,0.5032173395,-0.2350831777,0.1064348221,0.029259108,0.151907742,-0.2006684542,0.0791208073,-0.0376023613,0.2799378037,-0.2265824378,-0.0447771065,-0.3562315404,-0.034193404,-0.2187789977,0.1896177977,-0.1608967483,-0.3254478276,0.6411889791,0.0324766263,0.0277306344,0.7641415,-0.0001287131,0.0393472388,0.4674041271,-0.1851536632,-0.3570637405,-0.2890094817,0.352075994,-0.0747951642,0.4789135456,-0.1217302904,-0.0535692163,0.3681681156,-0.3571668565,-0.2829234302,0.337782383,0.2963617146,0.0819578841,0.5929397941,-0.1577762514,-0.0510140397,-0.0935970247,0.6183962226,-0.2803871334,0.6484384537,0.4351875484,-0.1534273028,-0.1185101122,-0.0808705688,0.1125618145,0.1081115827,-0.1265399009,0.0491170213,-0.1212580875,-0.2817679346,0.1882838756,-0.0434961244,0.427869767,-0.334849149,-0.3669345677,0.2918598652,-0.1322660148,-0.3039441109,0.4314469695,-0.0946971774,-0.4797498882,0.2464614213,-0.5746465921,0.1803811342,-0.095984593,0.1889845878,-0.3290427327,0.0139242448,-0.0037794746,0.0756979659,-0.0833096504,-0.0759193376,0.0775504485,0.218075335,0.1124901474,0.2105472386,0.0192041602,-0.1825955212,0.0042889011,0.2666103542,0.2472054362,0.4240624905,-0.182477057,-0.122575298,0.0941482261,0.1986427158,0.0297927018,0.414953351,0.1767910719,0.3287169933,0.4472732842,-0.0314272344,-0.0349344239,0.21398063,0.2701272964,-0.2194721401,0.0874325112,-0.0599674396,0.1236558408,-0.4884465933,-0.2430199832,0.3342878819,-0.2255838662,0.0130314277,0.2921376526,0.237573728,0.0871120691,0.1962379962,0.1621374488,0.4079699218,-0.1178594828,0.3706064224,-0.1039627865,-0.1124639362,-0.1322607696,0.190343529,0.5385653377,-0.0471802279,0.4167149365,-0.0070279688,0.5418527126,-0.0736422986,0.0300187226,-0.0224932767,0.1987535506,0.5158859491,-0.2314249128,-0.0430944599,0.2318384051,-0.1680154502,-0.2934478521,0.1876366138,-0.036553476,-0.360126704,-0.1811524779,-0.0328332819,-0.2402826995,0.0049480121,0.4045207202,-0.4238608778,0.1729899198,0.015730679,0.0126330955,-0.2405943274,0.0043097562,0.0955928937,0.325524956,0.5624173284,-0.1283846796,-0.040545214,0.0909309685,0.0560348853,0.3275008798,0.322997272,-0.2174919546,-0.079765968,-0.1228102669,-0.1518611461,0.3863773048,-0.5328124762,-0.2762471735,0.3100045621,0.1289285719,-0.1696591526,-0.1427436322,-0.208109349,0.1203560159,0.1268339306,0.3403026164,0.0339691974,-0.1404906511,-0.0252439156,-0.3712669313,-0.2300162911,0.0015914919,0.0133343441,0.0256101005,0.1600136161,0.1667650193,-0.4279124141,0.1311002374,-0.0518474355,-0.1038099676,0.0527333058,0.4886046052,-0.4241816401,0.0636517629,0.2276821285,-0.1836992353,0.3035302758,-0.0784523413,0.313644439,-0.1373235136,-0.0588411503,-0.5334228873,0.4065414369,-0.0188787188,-0.1267283708,-0.1220251769,0.0448220596,0.3527234793,0.1135549918,-0.1742255986,0.0604515374,-0.1182366535,0.104523845,0.3406917155,-0.0194027256,-0.4063432217,-0.4013935924,-0.1491001397,-0.0286082122,0.0115148136,-0.2519133389,-0.3854123652,0.2479814291,0.0569710359,-0.2244205177,0.2305828482,0.0816854611,0.2954339981,-0.0162090566,0.0367697999,0.0003966364,-0.1849222481,-0.0339136943,0.0098309023,0.3970498443,-0.0316920765,0.169982478,-0.1750234216,0.1441203803,0.4002425969,0.0310115479,-0.1585004479,-0.2736608684,0.3045320809,-0.0049121412,0.5264109969,0.0516441762,-0.1131985411,-0.019890476,-0.237845093,-0.1569035947,0.0667518228,0.3204009533,0.2629225552,-0.013487855,0.2638959587,0.0033946426,0.503872633,-0.1240240857,0.1816550493,0.1384320855,-0.2771652043,-0.3122504652,0.0401483029,-0.1818047613,-0.096017167,0.2377836555,0.2507990301,-0.0702297986,-0.2175699323,-0.2124915868,-0.1106227636,0.3328876495,-0.2270733416,0.2456769198,-0.1492201388,0.1999557167,-0.1773367077,-0.2198433727,-0.1061055735,-0.2637876868,-0.0378574058,0.2395730466,-0.0407100022,0.2350499034,-0.3171925545,0.4599715471,0.1577052921,-0.6352958083,-0.2629722655,-0.1444387585,-0.3395789862,-0.0776634365,0.3256201744,-0.3727446198,0.1843812317,-0.4821203649,0.012367812,0.1615027338,-0.4731438756,0.2949295938,0.135427624,0.5952253342,-0.0459183864,0.1508339494,-0.1922258884,-0.3344004452,0.4108138978,-0.319222331,-0.3365814984,0.0547794253,0.007266148,-0.16813609,-0.5058836341,-0.3063298166,-0.3148087859,-0.0533497594,0.1918007135,0.0501124673,0.1443279386,0.3339980245,0.4101510048,-0.0044366512,0.2073384672,0.0506999753,-0.0100285169,-0.4861396253,0.2371548414,-0.3061432242,-0.3615093827,-0.1987366676,-0.0529225543,0.3845674992,0.184077844,-0.5582749844,-0.2909577489,-0.1777923852,0.454624474,0.1030155793,-0.0468172766,0.2920683026,0.0422036573,0.0365013853,-0.0270703416,-0.3199698031,0.2881191969,0.0237831213,0.3060825765,-0.0239138398,0.1803099811,-0.0305550229,0.7351323962,0.1958531886,-0.240850091,0.3919067085,0.1805097163,0.3743961453,-0.4311841428,-0.2628128529,0.2902439833,0.0368955061,-0.4468597472,0.028092891,0.0327522643,-0.7290301919,-0.2692887187,0.0021749432,-0.1888222992,-0.2501921356,0.2722430825,-0.2999300659,0.3638345599,-0.0839752033,0.0755899921,-0.0710916519,-0.0229981132,0.0046499409,0.2835949361,-0.0393829867,0.0536701903,-0.1316987127,-0.4243867397,-0.4543828964,0.2604158819,0.1815054864,0.7051234841,-0.346678555,-0.2844624817,0.082671456,-0.0127998842,0.154745698,0.1888163239,-0.2712615132,0.2439539284,-0.335961163,-0.4482176602,-0.2516494691,0.0370886363,0.3414697349,0.1930963099,0.440770179,0.0702780485,-0.0986780077,0.165722087,0.1349698305,0.076014936,-0.0748338923,-0.386775583,-0.4530236125,-0.3725787103,0.3067518473,-0.0564760081,0.118548207,-0.1779237986,0.0036992624,0.17224738,-0.0218833722,-0.0572844297,0.3491224647,0.0232074112,-0.1563138515,0.0927621424,-0.0239244755,0.1609332561,0.0487536266,0.1074488387,-0.0694214106,0.2719088197,0.2726915777,-0.0642965138,0.0035998726,0.4597555995,-0.0078347111,-0.2611933053,-0.0098290322,-0.0911174566,-0.1698314697,0.2874849737,-0.0188481174,-0.0519399755,-0.0542290397,0.2032802999,-0.0349700972,0.0481021293,-0.085920766,0.4281371236,-0.0671040416,0.0126871029,0.2742472887,0.1784006655,0.9760027528,-0.043448884,0.0564655364,0.1795065403,0.0215456933,-0.2822630405,0.2678416073,-0.0515394844,-0.1963824332,0.2389208376,0.0208764523,-0.3305761516,0.2563527226,-0.1427588463,-0.2644149065,0.072164461,-0.4716963172,0.2073220909,0.0974567756,-0.0784889609,-0.322383225,0.0074124229,0.0452967957,0.0926403999,0.115573898,0.6465581656,-0.1674456894,0.104268074,-0.4589010775,-0.3181274533,-0.3372455537,-0.0750408843,-0.3058594465,0.170250237,0.4338040054,-0.6257207394,-0.1260721982,0.2802771032,0.5041136742,0.2558199167,-0.3557374775,-0.0234596934,-0.2054350823,0.0840063095,0.4423813522,-0.1081309766,0.121227406,-0.1389624327,-0.371650964,-0.2003159523,-0.1685028672,-0.3224394917,0.386313647,0.1003361419,-0.0086438693,-0.210942626,0.2304939479,-0.1333600134,-0.0340410545,-0.0159156751,0.0460458249,0.0832815766,0.060430523,0.3773466349,-0.0958931074,-0.1846657693,-0.136563912,0.0956185833,-0.038309712,-0.185235709,0.312359184,-0.0822093114,-0.0399587974,-0.095523119,0.1017749608,-0.3029921055,-0.1660038382,0.0567771681,-0.1600360125,-0.3582680523,-0.2082920223,0.5189055204,0.3025403321,0.2123356014,0.0068275491,-0.1937332153,-0.6883788109,0.1580587476,0.1393257529,0.2334798276,-0.0501988977,0.3976534009,0.095595181,-0.0819226652,-0.1154661179,-0.0777927563,-0.4233609736,0.0948056802,0.0910695344,-0.1875590533,0.3472159803,-0.1100250408,-0.0526953377,0.0446533076,0.0246979259,-0.0928222463,-0.0453760661,0.1755247563,0.0206200089,-0.3655266166,0.1706085652,-0.0415365696,0.1199657023,-0.1534401029,-0.097507298,-0.0371737033,-0.0049487855,-0.0805908814,-0.2490549535,0.168092981,-0.0559433252,0.1664381325,0.02138591,0.3151865005,-0.0494110473,0.1104306132,0.014967219,-0.1250571162,0.1689389795,0.2227516472,-0.2050159723,0.0805274621,0.0157679748,-0.1730946004,-0.0178681258,-0.0220928155,0.440037936,0.0800031349,-0.1780273765,0.1011368632,0.4403356016,0.0641917288,-0.0180315115,-0.0614109822,0.1983546168,0.1625598669,-0.0008159125,0.4802027643,0.064189598,-0.3462342918,0.4527828693,-0.0038468505,0.1812499613,-0.0662150532,-0.0012350854,-0.0291389227,-0.0532597564,0.3174806237,0.1411473602,-0.1872005016,0.1661782861,-0.0876652747,-0.0936512128,-0.0932968333,-0.2477650046,0.2596846223,0.0540902354,-0.1713436842,0.2499300241,-0.3742425442,0.1550906003,-0.0024953694,-0.1267191917,0.6321114898,-0.4966115952,-0.3118758202,-0.3651078939,-0.0407326445,-0.1091231257,-0.1233011484,-0.2257380784,-0.0633400008,0.1475431323,0.0406720936,0.013159032,0.1704794019,0.2608259916,0.11692442,-0.2754315138,-0.4348377585,-0.0534836464,0.1039655805,0.0875292867,-0.0200047679,0.5283891559,0.2062072754,-0.2827413976,-0.2568258047,0.4794691503,0.5924417377,0.391443342,0.0883463919,-0.2070693225,-0.2208623886,0.0311344713,-0.3435667455,0.2105430216,0.13870354,-0.0917729288,0.0225523971,0.0210673772,-0.0827917159,0.0328726247,-0.1348568499,0.139083907,-0.0759691298,0.3264746368,-0.0253339373,0.0435011089,0.0243980102,0.133956939,-0.248752594,0.3929110169,0.0833409876,0.0035572161,0.1769812554,0.0438139029,-0.0130078625,-0.0683165565,0.4770763218,0.3794436455,0.3073860109,-0.2329626828,-0.2339456081,-0.1714249253,0.1797963083,0.1334394366,0.3775700927,-0.0777674764,0.1723692864,0.1474080533,0.1493644267,0.0007748036,-0.2288633287,0.0070788185,-0.0403423496,-0.2509398162,0.0798590779,-0.5340796709,-0.1518705785,0.0936042815,-0.3492444456,0.3774879277,-0.25180462,-0.0802711919,0.0140894931,-0.256649375,0.1946002394,-0.0269120801,0.2261867374,0.3685008287,0.4666382968,0.189694643,-0.2133913636,-0.1679934561,-0.3907411695,0.1059961468,0.04576388,-0.2344073504,0.2104592472,-0.2749539018,-0.0344559178,-0.183785215,0.4403218627,0.1964014918,-0.2830712795,0.3341466486,-0.1689451784,-0.2094667405,-0.1900050193,0.1234180555,0.4776777327,-0.0616325252,0.2371267825,-0.282721132,-0.1846233308,0.5225761533,-0.2278445214,-0.0152757615,-0.0748453736,-0.0244677123,-0.123194389,-0.0727070495,-0.3173838556,-0.1097046733,-0.0619611442,-0.1095768437,-0.4085150063,0.3155002296,0.2680289447,-0.0570438281,-0.0413174145,0.594587028,-0.1424866617,0.0122231776,-0.0042418973,-0.0761355683]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2010","title":"Local testing fails","comments":"```\r\nco_filename = '', returned_obj = [0]\r\n \r\n def create_ipython_func(co_filename, returned_obj):\r\n def func():\r\n return returned_obj\r\n \r\n code = func.__code__\r\n> code = CodeType(*[getattr(code, k) if k != \"co_filename\" else co_filename for k in code_args])\r\nE TypeError: an integer is required (got type bytes)\r\n\r\ntests\/test_caching.py:152: TypeError\r\n```\r\n\r\nPython 3.8.8 \r\ndill==0.3.1.1\r\n","body":"I'm following the CI setup as described in \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8eee4fa9e133fe873a7993ba746d32ca2b687551\/.circleci\/config.yml#L16-L19\r\n\r\nin a new conda environment, at commit https:\/\/github.com\/huggingface\/datasets\/commit\/4de6dbf84e93dad97e1000120d6628c88954e5d4\r\n\r\nand getting\r\n\r\n```\r\nFAILED tests\/test_caching.py::RecurseDumpTest::test_dump_ipython_function - TypeError: an integer is required (got type bytes)\r\n1 failed, 2321 passed, 5109 skipped, 10 warnings in 124.32s (0:02:04)\r\n```\r\n\r\nSeems like a discrepancy with CI, perhaps a lib version that's not controlled? \r\nTried with `pyarrow=={1.0.0,0.17.1,2.0.0}`","comment_length":47,"text":"Local testing fails\nI'm following the CI setup as described in \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8eee4fa9e133fe873a7993ba746d32ca2b687551\/.circleci\/config.yml#L16-L19\r\n\r\nin a new conda environment, at commit https:\/\/github.com\/huggingface\/datasets\/commit\/4de6dbf84e93dad97e1000120d6628c88954e5d4\r\n\r\nand getting\r\n\r\n```\r\nFAILED tests\/test_caching.py::RecurseDumpTest::test_dump_ipython_function - TypeError: an integer is required (got type bytes)\r\n1 failed, 2321 passed, 5109 skipped, 10 warnings in 124.32s (0:02:04)\r\n```\r\n\r\nSeems like a discrepancy with CI, perhaps a lib version that's not controlled? \r\nTried with `pyarrow=={1.0.0,0.17.1,2.0.0}`\n```\r\nco_filename = '', returned_obj = [0]\r\n \r\n def create_ipython_func(co_filename, returned_obj):\r\n def func():\r\n return returned_obj\r\n \r\n code = func.__code__\r\n> code = CodeType(*[getattr(code, k) if k != \"co_filename\" else co_filename for k in code_args])\r\nE TypeError: an integer is required (got type bytes)\r\n\r\ntests\/test_caching.py:152: TypeError\r\n```\r\n\r\nPython 3.8.8 \r\ndill==0.3.1.1\r\n","embeddings":[-0.1595425755,0.0836868957,0.0080741448,0.0541945584,-0.0605914593,-0.2532820702,0.4322317541,0.3210858405,0.0665906072,0.2115702927,-0.0241569411,0.1039160192,-0.1923717856,0.4952004254,-0.2158735842,0.1359183192,0.0552796796,0.1934275776,-0.1318421066,0.0838604644,-0.095966205,0.2499605566,-0.2564118505,-0.04104276,-0.3022986948,-0.0245991461,-0.148721233,0.1830328107,-0.1826664209,-0.3603295684,0.6797194481,0.0199894588,-0.0009315464,0.6896492839,-0.00012576,0.0270796269,0.5052758455,-0.1809416413,-0.4039542675,-0.3320510089,0.1519881189,0.0037361854,0.3448631763,-0.1304060519,-0.0594432652,0.4255079925,-0.269272536,-0.3886942267,0.2537177503,0.3457771242,0.0861626714,0.6625935435,-0.1529203802,-0.0738562047,-0.0913394094,0.7564921379,-0.2964876592,0.6160543561,0.3982497454,-0.1529932469,-0.0890129432,-0.0370634012,0.1238623559,0.2078801244,-0.1936955452,0.1046222895,-0.1856215894,-0.3124096692,0.127361849,-0.0407238081,0.3077149093,-0.2930814624,-0.4210140109,0.2640115619,-0.1962429136,-0.4216315448,0.4477261603,-0.0321058594,-0.5543996096,0.1441136599,-0.6164290905,0.3599109054,-0.0956348106,0.1937013716,-0.2613513172,-0.0860241354,-0.1219527274,0.1039446294,-0.068969138,-0.1067535803,-0.0950204208,0.1158439144,0.0893242508,0.2673538625,0.0417058803,-0.2120662332,0.074565962,0.2886244953,0.1956859827,0.4594943821,-0.2210506797,-0.0970078409,0.0751854256,0.1858525425,0.0422291197,0.3725644946,0.1567064077,0.3691079617,0.40904212,0.0981971771,-0.1236698702,0.1423254311,0.2234022915,-0.2243603319,0.1463966519,0.0168841872,0.1898580343,-0.4522972107,-0.281761229,0.305057317,-0.1572923958,0.0335101858,0.3083818853,0.2298793048,0.0924803242,0.1790280342,0.1544673294,0.316252321,-0.217780754,0.2494838536,-0.0718299225,-0.0928375646,-0.1575595587,0.2292731553,0.5492322445,0.0396385491,0.405988425,0.0170017499,0.5969008803,-0.0530274436,0.084572956,-0.0820266455,0.2296279371,0.461274296,-0.2623299658,0.0007759786,0.2538804114,-0.1570122689,-0.309691608,0.0716572851,0.0254918672,-0.3389891982,-0.1903701574,-0.0041225697,-0.236270383,0.0440888926,0.3062964976,-0.4115735888,0.2350123227,-0.0339255668,0.066669412,-0.297963053,0.0334088169,0.0907244608,0.294031769,0.5816891789,0.0320189446,-0.0587128475,0.0872225538,0.1703658998,0.3222624958,0.3197820783,-0.2711986601,-0.080106847,-0.2009895593,-0.2122240365,0.4604290128,-0.5281284451,-0.2205269933,0.3383586407,0.1630102247,-0.0469934419,-0.024037499,-0.177936852,0.1230152026,0.0250453204,0.3560113013,0.02011583,-0.1439755708,-0.0507279299,-0.3050992787,-0.2096912265,-0.0326529033,-0.052001711,-0.0125007462,0.1478217989,0.1682515889,-0.3553977609,0.1316869557,-0.0926364511,-0.1003679484,-0.0265650917,0.4984470904,-0.436465919,-0.0033532269,0.1779424399,-0.1466665268,0.3157603443,-0.0652014092,0.3325464427,-0.1267909557,-0.0786102414,-0.4938027263,0.3182767332,-0.0920939893,-0.1416630447,-0.0879943743,0.1350090951,0.3590696454,0.1158088967,-0.2030635178,0.1976779103,-0.1674304605,0.0674079508,0.1798962653,-0.1173199266,-0.331875056,-0.3938655853,-0.1756162345,-0.0437940694,-0.0576129481,-0.2317662388,-0.3719640672,0.2077530324,0.0961100161,-0.1659799218,0.1020836383,0.0658373535,0.2748291492,0.0711638778,0.0056704115,0.0138983317,-0.1340230554,0.0324418806,-0.0586631671,0.5032771826,0.0161019936,0.1436363161,-0.1191534922,0.0993645117,0.3589843512,0.0462640114,-0.2223187238,-0.2488673031,0.3026212454,0.0815824345,0.4269850552,0.1282005608,-0.1495231241,0.0208277684,0.0208050832,-0.16235511,0.0347993001,0.2654781342,0.2746232152,-0.0285593122,0.2743022442,-0.0487582088,0.6123738885,-0.1288354844,0.1622859985,0.0833441317,-0.2782640755,-0.2611792386,0.0942928717,-0.1272592396,-0.1881013811,0.1684157401,0.2399555296,-0.0283272844,-0.2438034415,-0.2251516283,-0.1123573557,0.2600666583,-0.3167208731,0.3092109561,-0.1657240987,0.1626483798,-0.2243668437,-0.1949146241,0.0339774676,-0.2156327367,-0.1114063933,0.3362267911,-0.0429508388,0.2492925078,-0.3598666489,0.3789109588,0.1270722896,-0.6568129063,-0.2761151493,-0.0493920036,-0.358027786,-0.0783266053,0.3884337544,-0.3923020363,0.2230754346,-0.4847135246,-0.0083078006,0.1642359793,-0.4976350963,0.2448467165,0.1052220985,0.5596272945,0.0088488068,0.0886570513,-0.2389024943,-0.4123895466,0.4063134193,-0.2631421983,-0.3811514676,0.0973349065,-0.0332760774,-0.2352034003,-0.5069959164,-0.3434551358,-0.2640545368,-0.0690315664,0.3037874699,0.0882631391,0.1101689935,0.4585253894,0.408048898,-0.0020045231,0.1282039881,0.0467161201,-0.0791560635,-0.5806210637,0.2987908721,-0.3213188052,-0.306871295,-0.239809677,-0.1218192056,0.4034667313,0.2295228094,-0.4456574023,-0.1923150271,-0.191101715,0.4425347149,0.1964614242,0.0090472912,0.2945613563,0.0079896888,0.0029951094,-0.0171324536,-0.2513278127,0.2703384757,0.1097452715,0.316788733,0.0037349085,0.2698606253,0.0379693881,0.6234934926,0.3126415014,-0.2371502519,0.3645984828,0.1278417557,0.4174308479,-0.5137651563,-0.3205949366,0.3046328425,-0.0243830644,-0.4611196816,-0.0523291938,0.0129071046,-0.6639070511,-0.1939108372,0.0162188895,-0.1264440566,-0.306594491,0.2799482644,-0.2277898788,0.3235549927,-0.0602760911,0.0650712028,-0.1772476733,-0.01825477,-0.0232354645,0.2895423174,0.0458128303,0.070125334,-0.0966774225,-0.3918623626,-0.4547607005,0.2472249866,0.2436315268,0.6850363612,-0.2653803825,-0.2727941275,-0.0125285657,0.0156398565,0.008544852,0.2388784587,-0.2327783853,0.2690615356,-0.3548619151,-0.4105350673,-0.2229921967,-0.0113190422,0.4066485465,0.1472586989,0.4967995882,0.0819327757,-0.1409923285,0.2631053925,0.1353672743,0.0555800349,0.0363296121,-0.4470657706,-0.3837277591,-0.3355274498,0.2988200188,0.0592851117,0.0493523069,-0.1205607876,-0.005728988,0.1324140579,-0.0191627033,0.0592209399,0.3717755377,-0.031931147,-0.095816493,0.0943270773,-0.1094131693,0.0194689315,-0.0191997066,0.1649529636,0.0110499868,0.0687098354,0.2090958208,-0.041629687,0.0180174056,0.4032599926,-0.0849667042,-0.3252248168,-0.0046239556,-0.0674571544,-0.260946393,0.3393782377,0.0590507388,-0.0289673172,-0.1317611486,0.1066780016,0.0054074088,-0.0309332404,-0.0670477077,0.5506902933,-0.0220204201,-0.0417805463,0.2268805355,0.16521734,0.8252404928,0.0644442663,0.1021418795,0.2496579438,0.0548897795,-0.1910450757,0.3437012434,0.0162840746,-0.1306669712,0.2628117204,0.0059522148,-0.3169222772,0.3387096822,-0.1293205321,-0.2476847917,0.1091311201,-0.375210762,0.3533781767,0.0685973167,-0.0344468802,-0.3180185258,-0.0550896153,-0.0110920453,0.1156556159,0.1104644984,0.5969853997,-0.0907769054,0.0594348721,-0.4712801874,-0.3245825171,-0.3380560279,-0.0460062064,-0.3487530351,0.1952670962,0.3555602133,-0.5530384779,-0.0638664216,0.1838650256,0.4580511451,0.2221153677,-0.3360338211,-0.0448382758,-0.0872653946,0.127704829,0.4157726467,-0.1899158806,0.1803596914,-0.0780538395,-0.2937190235,-0.2260476649,-0.1601119787,-0.3429050744,0.399027586,0.0802356973,0.0668484271,-0.2352944911,0.1735438555,-0.1413566619,0.0017312756,-0.0245899186,0.0601652227,0.0110426098,0.050463926,0.3054960668,-0.1586059183,-0.1889247149,-0.1347510815,0.0767960548,-0.0377657227,-0.1541689485,0.3368021548,-0.2317029089,-0.0590792634,-0.1062084809,0.1889463067,-0.2515741885,-0.2681213915,0.0837739781,-0.1097014248,-0.3355261683,-0.1889190674,0.5555794835,0.2536102831,0.14496921,0.0675851703,-0.219927147,-0.6909095645,0.1423329413,0.0476176478,0.2239873856,-0.0385322422,0.3672914207,0.0404757708,-0.1701398939,-0.1303741932,-0.1746755838,-0.3640626967,0.0801698342,0.1796462387,-0.1369644701,0.3592987061,-0.0891140327,-0.0150517384,0.0554107763,0.0056980201,-0.1087834165,-0.1023395285,0.1610949188,-0.0209927205,-0.3436934352,0.2249809206,0.0064686281,0.0898526534,-0.2450373322,-0.0761692524,0.0171049107,-0.01760846,-0.0669981763,-0.3150114119,0.1212767139,0.0127036916,0.1613851637,-0.011731904,0.2935883999,-0.0230018776,0.1198140308,0.0001847995,-0.0750523582,0.1143960282,0.2372387648,-0.2427397519,0.1428281963,0.0291054901,-0.1974077821,-0.0549752638,-0.0433401987,0.3361026049,0.1295180619,-0.1475603431,0.0766785964,0.3650749624,0.0751589611,-0.0195954572,-0.0257223826,0.1018336937,0.1505981535,0.0228371583,0.5174657702,0.0231225695,-0.3435842097,0.5378297567,-0.0749443546,0.2276362926,-0.1015102416,0.1055783629,-0.0402405001,-0.0303136054,0.2926293314,0.1018690392,-0.1841751784,0.2063316852,-0.0679171905,-0.1282240301,-0.1412269026,-0.1888823956,0.2897832394,-0.0143539151,-0.139730081,0.1568306088,-0.3757257462,0.0937148035,0.1500715613,-0.0578574762,0.5642964244,-0.585581243,-0.2829949558,-0.3944428861,-0.1439433098,-0.0790508017,-0.1622623205,-0.2222795188,-0.0152671747,0.0844991356,0.0522095002,0.0321792699,0.201047495,0.3205499947,0.1275767535,-0.2623807192,-0.3341464698,-0.0527927503,0.0878486559,0.0949488506,-0.0202660747,0.5610430241,0.2345622033,-0.2423794419,-0.2002236843,0.4721873999,0.5895764232,0.3786364794,0.122732535,-0.195104003,-0.2294461429,0.0129828881,-0.3257735372,0.2213953137,0.1242721677,-0.0131640732,-0.0532540306,0.0574236698,-0.0716467872,0.0075355265,-0.1439877003,0.0037623339,-0.1829033494,0.4468245804,-0.0985436589,0.0938501135,0.0514191575,0.1554649174,-0.241360411,0.28232494,0.0888577104,0.0573535301,0.2049228847,0.1141103208,0.0028201421,-0.0617848672,0.4479651451,0.2357168496,0.2131757885,-0.2273445427,-0.1807401478,-0.1836016327,0.114117831,0.1576030552,0.2857304215,-0.1136940122,0.2241499275,0.1839517653,0.0712651759,-0.0494570397,-0.3080796301,0.0026940412,-0.0428029001,-0.2135598361,-0.0127306627,-0.4230502844,-0.1899386048,0.091598168,-0.4213809669,0.3242907226,-0.2475095242,-0.0484454706,0.0406978428,-0.2419206947,0.1870486587,-0.104673177,0.2583915591,0.4208637178,0.4123576283,0.1219663098,-0.2801037431,-0.2073035538,-0.3272327483,0.1092548668,0.1512139738,-0.2796409428,0.3367854655,-0.2444268167,0.0386393368,-0.246767953,0.3907268047,0.1972016543,-0.3832684457,0.3099580407,-0.1385192722,-0.287047565,-0.1150681525,0.044684615,0.504112184,-0.117757462,0.2682074606,-0.252507925,-0.1620399654,0.504615128,-0.4364575446,-0.0934123397,-0.0293934382,0.0501447096,-0.1696913838,-0.0489449278,-0.3520064652,-0.0549144074,-0.0120237721,-0.1073765531,-0.385491997,0.3357231617,0.2234318554,-0.0321989506,-0.0404456146,0.7254137397,-0.203180179,-0.025390001,0.0430914611,-0.0735731199]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2010","title":"Local testing fails","comments":"I managed to reproduce. This comes from the CodeType init signature that is different in python 3.8.8\r\nI opened a PR to fix this test\r\nThanks !","body":"I'm following the CI setup as described in \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8eee4fa9e133fe873a7993ba746d32ca2b687551\/.circleci\/config.yml#L16-L19\r\n\r\nin a new conda environment, at commit https:\/\/github.com\/huggingface\/datasets\/commit\/4de6dbf84e93dad97e1000120d6628c88954e5d4\r\n\r\nand getting\r\n\r\n```\r\nFAILED tests\/test_caching.py::RecurseDumpTest::test_dump_ipython_function - TypeError: an integer is required (got type bytes)\r\n1 failed, 2321 passed, 5109 skipped, 10 warnings in 124.32s (0:02:04)\r\n```\r\n\r\nSeems like a discrepancy with CI, perhaps a lib version that's not controlled? \r\nTried with `pyarrow=={1.0.0,0.17.1,2.0.0}`","comment_length":27,"text":"Local testing fails\nI'm following the CI setup as described in \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/8eee4fa9e133fe873a7993ba746d32ca2b687551\/.circleci\/config.yml#L16-L19\r\n\r\nin a new conda environment, at commit https:\/\/github.com\/huggingface\/datasets\/commit\/4de6dbf84e93dad97e1000120d6628c88954e5d4\r\n\r\nand getting\r\n\r\n```\r\nFAILED tests\/test_caching.py::RecurseDumpTest::test_dump_ipython_function - TypeError: an integer is required (got type bytes)\r\n1 failed, 2321 passed, 5109 skipped, 10 warnings in 124.32s (0:02:04)\r\n```\r\n\r\nSeems like a discrepancy with CI, perhaps a lib version that's not controlled? \r\nTried with `pyarrow=={1.0.0,0.17.1,2.0.0}`\nI managed to reproduce. This comes from the CodeType init signature that is different in python 3.8.8\r\nI opened a PR to fix this test\r\nThanks !","embeddings":[-0.2129509598,0.0823646188,0.0238413643,0.0946619436,0.0496156551,-0.1908244491,0.3812552691,0.3104431927,0.0505841784,0.2253711671,0.1375240833,0.1101146713,-0.1860496998,0.6575943232,-0.1123831272,0.1562343091,0.0674784482,0.2533706129,-0.0266020261,0.131237939,-0.1200583279,0.2680254579,-0.2432248741,-0.0653258264,-0.2336260676,-0.0076859454,-0.1573687196,0.2269874364,-0.1869493574,-0.5005517006,0.7258083224,-0.0596951246,-0.01177736,0.6850973368,-0.0001276803,0.0460287258,0.5900765061,-0.1311247945,-0.2848140895,-0.2090544403,0.1225332916,-0.009903796,0.4425723255,-0.0584858656,-0.0636177137,0.4232131541,-0.3844325244,-0.450411737,0.3081777394,0.3412259221,0.0856561214,0.6891777515,-0.1833498776,-0.0647268742,-0.1557070017,0.7216380835,-0.2735095322,0.5735644698,0.4503455758,-0.1201026291,-0.0678781569,-0.1077148542,0.1148411334,0.1880001724,-0.1224890649,0.0224849395,-0.0936857313,-0.1882404387,0.1383006722,-0.0737014711,0.3249828219,-0.2547633648,-0.3429161608,0.1858745962,-0.1495690942,-0.5330841541,0.5146340728,-0.1113534197,-0.476841867,0.1915977299,-0.5451625586,0.2967489064,-0.0748954862,0.1553913951,-0.3125766516,-0.083272025,-0.1046864837,0.110724628,-0.0891967937,-0.0674922615,0.0989183336,0.1787144542,-0.0076315664,0.2872788608,0.1070478484,-0.2111832947,0.014469808,0.3994742036,0.1787104905,0.4556684494,-0.2514183819,-0.0727530271,0.122231327,0.1218675002,-0.0088822292,0.3863516748,0.2146190107,0.230663076,0.3758414984,0.0232177433,-0.0810358673,0.1415196359,0.2208197266,-0.2816311717,0.2822886109,-0.0321859568,0.1462036818,-0.4389745295,-0.0922711268,0.2229402214,-0.0682022572,0.0167919975,0.2327680588,0.2272420973,0.1824134588,0.1523486078,0.0944300443,0.4298304319,-0.1778352708,0.2726736963,-0.0876457244,-0.1687568575,-0.1537641287,0.1956510395,0.4986808002,-0.1537046432,0.4561303854,-0.0057980446,0.6063702703,-0.0310987569,0.0206634328,-0.027587628,0.2237470895,0.4527343512,-0.2669318318,-0.0231731255,0.3080402613,-0.2273074985,-0.3217340708,0.0499487519,0.1005844027,-0.4410426617,-0.2197590023,-0.0188295841,-0.283090204,-0.0106937038,0.272988081,-0.4864016771,0.2079430223,0.0007704758,0.0649352446,-0.2876134515,0.1068172902,0.1075926796,0.2090523392,0.5335060358,0.0195477121,-0.0479588173,0.0798538178,0.0793181211,0.285722822,0.3176197708,-0.267046392,-0.0958705693,-0.199275896,-0.1947425008,0.4257667959,-0.5156852007,-0.1774297804,0.3740007281,0.1563418657,-0.1985573322,-0.0494492166,-0.2684140801,0.0657720938,0.098622635,0.240212962,0.0567436665,-0.1665887833,0.060193494,-0.3022293746,-0.2904709876,-0.0210461076,0.0338267833,0.0249781422,0.176968202,0.199806124,-0.4017611742,0.1179298162,-0.0926960185,-0.1164355874,0.0205856804,0.5448556542,-0.3864399791,-0.0036130815,0.1999371499,-0.0327697322,0.3577795327,-0.0780031607,0.2045546025,-0.1400953829,-0.0209076274,-0.5408804417,0.3800971806,-0.0991014838,-0.0528328083,-0.1179859787,0.0609114096,0.3830114007,0.1435604393,-0.2402081192,0.0653643534,-0.0774748623,0.094665207,0.2909987569,-0.0861129239,-0.3287919462,-0.4251502156,-0.15633443,-0.0233165193,0.0048968778,-0.2242097408,-0.4006009698,0.2171485126,0.088674739,-0.1416569501,0.074613139,0.0953619629,0.2319166362,0.0282800142,-0.0058214879,0.0119786272,-0.1869169474,-0.0353843607,0.0147935646,0.4722611308,0.0298198331,0.0833989531,-0.1591881812,0.1093278974,0.353066206,0.0234856755,-0.1824235171,-0.3159005642,0.2667004168,0.0317160487,0.588888526,0.1291597039,-0.1410611421,-0.1075278744,-0.0508211851,-0.1599721909,-0.0017898654,0.1870503128,0.2271011174,-0.0158884823,0.2266928256,-0.1101670861,0.553195715,-0.1585794538,0.074934043,0.1816506088,-0.2237858772,-0.3028706908,0.1822277009,-0.1517623514,-0.0119571825,0.1467459798,0.2263767421,-0.0592967048,-0.2383322716,-0.2199627608,-0.0692370012,0.1834665239,-0.3558005095,0.240716666,-0.1439508647,0.0936436579,-0.3681105375,-0.2125537544,0.0038098046,-0.253749311,-0.1256662607,0.2570477128,0.0167053044,0.2124892622,-0.3055720925,0.4417043328,0.1103474349,-0.6165693402,-0.230748862,-0.0421803705,-0.2874609232,-0.1102833077,0.3440142572,-0.4251949191,0.1442624032,-0.5436748266,-0.0106434077,0.2099768966,-0.5716857314,0.3610416055,0.1101282462,0.5630653501,0.0908894315,0.1359783709,-0.0833356082,-0.32854563,0.3907449543,-0.2811042964,-0.320284456,0.1116458178,-0.0592874885,-0.2386706769,-0.4834084809,-0.224175185,-0.1732220352,-0.0486699156,0.2961143851,0.0813507736,0.0692149922,0.4372489452,0.3239246309,-0.0071029454,0.0879161656,0.0492147356,-0.0323043391,-0.6007462144,0.2518290579,-0.2372844368,-0.2854530513,-0.1597947329,-0.0864265561,0.4862286747,0.0981935784,-0.3998621404,-0.1802986115,-0.1212024838,0.4978191555,0.1771852523,-0.0901314989,0.3163525462,0.0246620309,0.0493826084,-0.0573718548,-0.2262752652,0.1844124347,-0.004375041,0.3145067394,0.017887475,0.2293648869,-0.0737022236,0.6722061038,0.2346595973,-0.2608801723,0.3040290177,0.0912792236,0.3293498755,-0.4512397051,-0.2017839253,0.2601115406,0.024302505,-0.4762885571,-0.0437384546,0.0247655362,-0.6657323241,-0.1394970566,0.1744857728,-0.1548580974,-0.1806101799,0.2503747344,-0.2951577902,0.3791856468,-0.0946983546,0.113441579,-0.135285452,-0.0312159415,-0.0109150289,0.3398591578,-0.1045146361,0.0421219468,-0.1297154427,-0.4960711896,-0.3819331527,0.2305472046,0.218238771,0.6499724984,-0.210214898,-0.3641653359,0.0249854829,0.0465437062,0.0963647664,0.1762783378,-0.2380157262,0.2721733749,-0.3735032678,-0.4916273654,-0.2925334573,0.0255824756,0.4043450654,0.0784006342,0.5375287533,0.1264640093,-0.1271789819,0.1342837811,0.1102669239,0.0929217786,-0.0375756063,-0.4543280303,-0.4238272905,-0.3833255768,0.3148334324,0.0197407492,0.0344986282,-0.2098483592,0.0249782763,0.0312269311,0.0165390577,-0.072083734,0.3944136202,0.0223523602,-0.1345159113,0.0904434249,-0.0883588344,0.0729425028,-0.0201753732,0.1470246911,0.1128853783,0.0944927782,0.1423823535,-0.1168444976,-0.0356834903,0.45736745,-0.0383813977,-0.1947221309,0.1069134921,-0.1083418205,-0.2660885751,0.2129255682,0.0121903867,-0.0684786513,-0.0959031209,0.1370833814,0.0226377118,-0.0034940359,-0.0994243398,0.4119966328,-0.055170171,-0.0298769884,0.36515221,0.1775611639,0.815760076,-0.008666073,0.0692775249,0.2503994107,-0.0209907293,-0.345080018,0.5113013983,-0.0401657373,-0.2692530453,0.2452153862,-0.0141847041,-0.3603081405,0.319277972,-0.1270448565,-0.2522837222,0.0036985581,-0.4036070704,0.3278361857,0.0643663034,-0.1236021295,-0.359387219,-0.0345865861,-0.1425473541,0.0836255699,0.1283957064,0.6133269072,-0.1106794998,0.076792486,-0.3790447712,-0.3198170066,-0.3265846372,-0.0653562471,-0.2735286057,0.2031435221,0.4923526049,-0.4473854303,-0.0076335594,0.1897126883,0.5910066366,0.2151007801,-0.3710512817,-0.0241356976,0.0959766358,0.0187282357,0.391875267,-0.1623622626,0.1390871406,-0.0467726476,-0.3181010783,-0.1999569833,-0.2435209006,-0.3389804363,0.3941819966,0.0037320661,-0.0933826044,-0.3095394373,0.1387613416,-0.1527013481,0.0707067475,-0.0347961076,0.045121491,0.007147376,0.0995954946,0.3269562423,-0.0798412561,-0.2007296234,-0.1152878553,0.0390938967,-0.0577860363,-0.1012968868,0.3804948926,-0.110172525,-0.00550442,-0.1118166149,0.1505404264,-0.2089612633,-0.2460995615,0.0759301484,-0.1104968712,-0.3170503676,-0.0975174904,0.5581837893,0.2994672358,0.1594979167,0.0512622036,-0.1069932431,-0.6706625819,0.1032543927,0.0669698045,0.2133810967,0.0146429464,0.4575365782,-0.0103570838,-0.2144230008,-0.1220070869,-0.0885186642,-0.3779917657,0.0571436994,0.0681567565,-0.0752217174,0.3518764973,-0.1678056866,-0.0305306017,0.0250393394,-0.0496744663,-0.0900232196,-0.0462163799,0.1742309928,0.0484291203,-0.322148025,0.3157858849,0.0064819762,0.118368119,-0.1151713431,0.0057879575,-0.0856154412,0.0142186312,-0.0809696391,-0.2023141682,0.0976982936,-0.1170178801,0.0960860401,-0.0926927105,0.3523995876,-0.08898592,0.1439667791,-0.0554910526,-0.0477977805,0.2524555624,0.2699755132,-0.2041554451,0.168750897,0.0303086732,-0.2895763814,0.017219495,0.0655822456,0.304215014,0.0805393457,-0.2911088765,0.1201175153,0.2948726714,0.0722345784,0.0454737432,-0.0120311547,-0.1217375025,0.1613795608,-0.015614788,0.4948710203,0.2002597153,-0.3090687692,0.4006702006,-0.0324272923,0.3457435966,-0.030457275,0.0559674092,-0.0055869259,-0.0296824072,0.2638401389,0.1978754848,-0.1267087907,0.1102813259,-0.1203780621,-0.0820246339,-0.044404272,-0.2603597045,0.3358689547,-0.0584292375,-0.1032193303,0.2719579041,-0.3337039351,0.1102375686,0.0633105934,-0.0020893919,0.6600573063,-0.6151897311,-0.3002285659,-0.3712926805,-0.1774308383,-0.0651963875,-0.1714660078,-0.3542012572,0.0501394011,0.0443811417,0.0622213557,-0.0641276017,0.2548420727,0.3229344189,0.1223639846,-0.2019755542,-0.4548483193,-0.0604462177,0.0894837081,0.0883827358,-0.0564639829,0.5304108858,0.1447068751,-0.3150200844,-0.1222660393,0.5399008393,0.661506474,0.2859234214,0.0971381888,-0.2560989559,-0.2292697281,0.009942065,-0.2649981976,0.1255940795,0.1515680403,0.108160831,-0.0581224374,0.0584351607,-0.0417929627,0.042836722,-0.147020489,0.0676144138,-0.1739849001,0.3367727697,-0.0356601477,0.1551565379,0.0026472048,0.1935731322,-0.2381478399,0.3953502774,0.1332026422,0.0825400501,0.1771988571,0.1677247137,-0.0010882423,0.0668807924,0.5126700401,0.2918177545,0.1503744274,-0.185412243,-0.2878711522,-0.1938413829,0.1387003511,0.1663697511,0.2949790955,-0.0488843657,0.1859448701,0.1205979288,0.089424409,-0.0119657861,-0.1441293508,-0.1215384454,-0.1335771382,-0.1835174412,-0.0250042863,-0.3701706231,-0.1906794906,0.0149998218,-0.3371478617,0.3524735868,-0.1409901381,-0.0637197569,-0.0543728434,-0.2187353969,0.1781956106,-0.2250043005,0.2600577474,0.3480683863,0.3272466958,0.1648110449,-0.3105279207,-0.1107887849,-0.4063545167,0.0237594768,0.1805666983,-0.3148145974,0.3108045757,-0.2133536935,-0.0274856258,-0.233799994,0.265732199,0.1744324267,-0.3407872021,0.1513764411,-0.139425531,-0.2918334603,-0.0840960741,0.166721642,0.4874753654,-0.1143287271,0.2222248912,-0.161990881,-0.2502763569,0.4320521653,-0.3146128356,-0.0356788412,0.0171358548,-0.030336421,-0.2345733494,-0.0523699708,-0.3275541961,-0.0547793023,-0.0755200088,-0.1453135759,-0.4074226022,0.3302369714,0.2467478812,-0.029241303,-0.0792271271,0.6571492553,-0.1429576129,0.0041963686,-0.0779266655,-0.033836063]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2009","title":"Ambiguous documentation","comments":"Hi @theo-m !\r\n\r\nA few lines above this line, you'll find that the `_split_generators` method returns a list of `SplitGenerator`s objects:\r\n\r\n```python\r\ndatasets.SplitGenerator(\r\n name=datasets.Split.VALIDATION,\r\n # These kwargs will be passed to _generate_examples\r\n gen_kwargs={\r\n \"filepath\": os.path.join(data_dir, \"dev.jsonl\"),\r\n \"split\": \"dev\",\r\n },\r\n),\r\n```\r\n\r\nNotice the `gen_kwargs` argument passed to the constructor of `SplitGenerator`: this dict will be unpacked as keyword arguments to pass to the `_generat_examples` method (in this case the `filepath` and `split` arguments).\r\n\r\nLet me know if that helps!","body":"https:\/\/github.com\/huggingface\/datasets\/blob\/2ac9a0d24a091989f869af55f9f6411b37ff5188\/templates\/new_dataset_script.py#L156-L158\r\n\r\nLooking at the template, I find this documentation line to be confusing, the method parameters don't include the `gen_kwargs` so I'm unclear where they're coming from.\r\n\r\nHappy to push a PR with a clearer statement when I understand the meaning.","comment_length":79,"text":"Ambiguous documentation\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/2ac9a0d24a091989f869af55f9f6411b37ff5188\/templates\/new_dataset_script.py#L156-L158\r\n\r\nLooking at the template, I find this documentation line to be confusing, the method parameters don't include the `gen_kwargs` so I'm unclear where they're coming from.\r\n\r\nHappy to push a PR with a clearer statement when I understand the meaning.\nHi @theo-m !\r\n\r\nA few lines above this line, you'll find that the `_split_generators` method returns a list of `SplitGenerator`s objects:\r\n\r\n```python\r\ndatasets.SplitGenerator(\r\n name=datasets.Split.VALIDATION,\r\n # These kwargs will be passed to _generate_examples\r\n gen_kwargs={\r\n \"filepath\": os.path.join(data_dir, \"dev.jsonl\"),\r\n \"split\": \"dev\",\r\n },\r\n),\r\n```\r\n\r\nNotice the `gen_kwargs` argument passed to the constructor of `SplitGenerator`: this dict will be unpacked as keyword arguments to pass to the `_generat_examples` method (in this case the `filepath` and `split` arguments).\r\n\r\nLet me know if that helps!","embeddings":[0.0630255416,-0.0559889339,-0.0537045784,0.1070300266,0.0235678162,0.1693243831,0.3407925665,0.1025089547,-0.1445971131,-0.1650757343,0.0825752988,0.3592206538,0.0475475378,0.0229658056,0.1957679987,-0.226720497,0.0967702419,0.1418141127,-0.0160175748,-0.1553122252,-0.3404794335,0.260173142,-0.3286114335,0.3037527204,-0.2405357361,-0.2140244842,-0.1296740472,0.3539068997,-0.4259344935,-0.2638630569,0.2149847448,0.1732774228,-0.1561237723,0.1987597942,-0.0001124477,0.0061698449,0.294578284,-0.1892530769,-0.1790917218,0.0210828166,-0.3679366112,-0.177504614,0.0289063752,-0.3533598781,0.1910025924,-0.3274487853,0.0365873501,-0.1565573066,0.5122282505,0.1903434098,0.1555772275,0.0230273176,0.1395213306,-0.1815897673,-0.1611030996,0.5714946985,-0.2454582304,-0.1139202565,0.0408839472,0.1185335889,0.0628223196,0.2183078676,0.0905427784,-0.1116909012,0.4655440748,-0.0411249548,-0.2104089409,-0.5389917493,0.097147584,0.2854478657,0.5018543601,-0.2640081048,-0.2595461011,-0.2530981004,-0.1830708086,-0.1483088136,-0.1307505369,0.347219348,-0.3724952936,-0.0102221146,-0.3185660541,-0.2991984189,-0.26646173,-0.3025457561,0.1488544345,0.3738004565,-0.1506491452,0.1400251687,0.0473176725,0.1209765524,-0.0652601048,-0.227573663,0.0284532961,0.4920312762,0.0334739983,-0.1478964984,0.2251163274,0.0053131371,0.4009979963,0.2251070142,-0.0026085726,-0.2663418651,0.0731081888,0.1978611946,0.358804673,-0.2031565905,0.3699664474,0.0002882908,0.0691666305,0.0147317173,0.226358518,-0.1669269949,0.1684111953,-0.3830891252,0.0449414216,0.0379318297,0.3937673271,-0.1123331413,-0.0958527774,0.0414182171,-0.1373912543,-0.3935705125,0.0558243245,0.0493346043,0.2171518803,-0.0029187831,-0.089061588,-0.1021661758,-0.1244849116,-0.2616605163,-0.2409771383,0.2836477757,-0.1866672635,0.1743055582,-0.1942203939,-0.2025459409,0.3773424625,0.2009484172,0.2653453946,0.1425868273,0.1070011929,0.1838041246,0.1337120235,0.2818725109,-0.0353092216,0.1737410277,0.220831573,0.0379706286,-0.302809,0.0211159531,-0.4155864716,-0.3842844665,0.045171719,0.105499588,-0.2765238583,0.1617227197,-0.2013297826,0.3348644376,0.0546006337,-0.1171721071,0.0920717791,0.0104708672,-0.0390683413,-0.5046525598,0.312372297,0.4569101334,-0.4068639278,-0.1470859945,-0.1976564378,-0.2394910455,-0.0142202331,0.0178308152,-0.3050229847,0.4071798325,-0.2098042667,0.4111057222,0.5502249599,-0.0264273565,0.2617900968,0.4143196642,0.2440982014,0.0400957167,0.1967811733,-0.3653977811,0.1281254292,-0.1726318449,-0.2287610918,0.1094249412,-0.120420076,-0.2198890299,0.0926813558,0.0025510988,-0.1228934303,-0.1769018918,-0.115951471,0.0135553312,0.0102812005,0.1956638843,0.4094674885,-0.2382920235,0.1461001784,0.3196187913,0.3091386259,-0.1995504797,0.222786501,0.0575036407,-0.6098825932,-0.0888738409,0.0193511471,0.1056629047,-0.1684192419,-0.2631372213,-0.3937975764,0.0714038163,-0.2420980334,-0.2368248403,0.1145801023,-0.0982949883,0.2783874869,-0.1827614307,-0.3041557968,-0.0179167967,-0.0206736829,0.0775979906,-0.6391506195,0.1906316578,-0.0363016762,0.1559221745,-0.0530466586,0.0459443703,0.0631185621,-0.2324765623,0.1714751869,0.4587950706,0.4186858833,-0.0333929174,-0.0633999854,0.3129434288,-0.0341993682,-0.0261276308,0.3035819829,-0.0439402387,0.1650856137,0.0573666841,-0.3358438015,0.4323942661,-0.2407272309,0.3330581188,0.2003154308,0.0207136869,0.1567843109,0.1918630004,-0.1606838405,-0.113787815,-0.2559685409,0.206464842,0.3144348562,0.0809791163,-0.2916631997,0.4554891586,0.5633227825,-0.3260163963,-0.2243342996,0.0192822777,0.0195236579,-0.3728529811,-0.0673384592,0.3394281268,0.5309390426,0.024987286,0.1743120849,0.0886852294,-0.267105937,-0.226940006,0.4196886718,0.0794181749,0.0610033311,0.065126434,-0.5002262592,-0.265781343,-0.3489812315,-0.1541882008,0.2562901676,0.2185079604,-0.367577225,0.2802488208,0.0644766912,0.1209354028,-0.262529254,-0.1141392887,-0.2750662565,-0.5612815619,0.042691756,-0.0299975779,-0.0898886174,0.0450083874,0.0352852643,0.2285885364,-0.0247040503,0.1770927012,-0.3000349402,-0.1925522536,-0.0262379795,0.1083843559,-0.0044304915,0.0198434778,0.2194134891,-0.1660426855,0.1192978993,-0.266951263,-0.3701042235,0.2968113124,-0.4268538654,0.2854997516,0.7126973271,0.3422035873,0.0494979918,-0.2332897037,0.0504644662,-0.1468913406,-0.2825329006,-0.2010700405,-0.1071046591,0.072487399,-0.3547919393,-0.5272268653,0.0341721885,-0.2080563158,0.2492889166,-0.0573888198,0.2517394722,0.2855532169,0.2062866837,0.2444442958,0.0874435753,0.1648581028,-0.1609761119,-0.3166238368,0.2856165767,-0.4865406454,-0.1536630541,0.1032462344,-0.0922755897,0.4634689391,-0.0369105339,-0.106406264,-0.3701996505,-0.1186551675,0.0424704887,0.0186975393,0.3841692209,0.1690154821,-0.0273057893,0.0256089401,-0.3647847772,-0.1000140235,0.0551919639,-0.3473075926,0.1823521256,0.0608928725,0.4155501425,-0.0704485551,0.6626817584,0.3334254026,0.0407076627,-0.0579508692,-0.1177552864,0.4817964137,-0.019930752,-0.1661497205,0.0748713687,0.1015366763,-0.1329635978,0.0953426436,0.0903722569,0.0937505737,0.1487014294,-0.0274143666,-0.2631871402,-0.3521752954,0.1254463643,0.2102634758,0.2774250805,-0.0712267086,0.2989457846,-0.1903245747,0.1821182221,-0.2548038363,0.4908554852,0.614569068,0.0827763304,-0.3357668817,-0.1644798964,-0.3440629244,0.1310023963,-0.0074124504,-0.0851845145,-0.0927434638,-0.1233957261,0.141481325,-0.1461156607,0.413836956,-0.3477959633,-0.1935280859,0.152582407,-0.0163217634,-0.1738789082,0.0659221113,-0.1577224135,0.2936332822,0.1764202714,0.0440504476,-0.5996695161,-0.4980016351,0.1934526116,0.0108284913,-0.204181388,0.1298812926,-0.1511049718,-0.6448942423,-0.086166434,-0.1322562397,0.1849901527,0.0545251742,-0.179307431,0.3947924972,0.1312598437,0.1218222603,0.1520989835,0.4695465863,0.306617707,-0.1120469794,-0.2064917088,0.0765416697,0.1102181152,0.0202548113,0.3572825193,-0.1449250877,-0.2403621972,-0.0022131344,-0.2800565958,0.5578088164,0.2927026153,-0.1902349144,0.034793444,-0.1490893662,-0.1357687116,-0.2394869328,0.0017741587,0.3852307796,-0.3035707772,-0.3060829341,-0.8698117137,0.4335879982,0.3480955958,-0.20586209,0.3483261764,0.3761308789,-0.2476671934,0.0583863929,0.3835319579,0.7816467881,0.4372035861,0.2335245162,0.4435303807,-0.2188877314,0.5801462531,0.0565150306,-0.0785417035,-0.4775753021,0.004559346,-0.2190327644,-0.1659436673,0.2511464357,0.3887832463,-0.2179741561,0.1936330944,-0.1872211099,0.0677072182,-0.1618505716,0.2255186439,-0.1253626943,-0.0913020372,-0.2029479146,0.1133400947,0.0569847561,-0.003909823,-0.0460459031,-0.2679027915,0.0644950718,-0.4196642041,-0.3684495986,-0.0609957129,0.3079555929,0.0672421679,0.2474593222,-0.2270311415,0.388653636,0.1202156097,0.2785611153,-0.0669822097,-0.1209646687,0.1790650785,-0.1707371175,0.0155601781,-0.1539044231,0.1677234471,0.4168996513,0.0969812572,-0.0540144295,0.0565552115,-0.1589327455,-0.0883345157,-0.4750368893,0.1810630262,-0.2370461673,-0.5176566839,-0.1148443669,0.0828726143,-0.2824341953,-0.0762306005,0.0554892458,0.2443851382,-0.2320571393,0.2936023772,-0.3023997247,-0.0224810578,0.0489074998,-0.2302009612,0.0862597227,0.036872521,0.2618043125,0.0112962713,-0.1381825954,-0.3272832632,0.2000401169,0.572620213,-0.6263508797,0.2564403713,0.2468701601,0.1769984961,0.1548410654,0.1777608544,-0.1841894537,0.3437292874,-0.2117950022,-0.3027583063,0.0321090445,0.3065128624,-0.2223132551,0.2031894028,0.0234012399,-0.0157859251,0.1977748275,0.1179457456,-0.2993242741,0.1151680052,0.1789742857,0.0359147452,0.2088566124,0.0154797528,0.1003277451,0.075209789,0.00387806,-0.0526536256,-0.3928084671,-0.1168734133,-0.2186903954,0.1830581874,-0.0288584139,-0.1877184659,0.1499773562,0.1668448001,-0.4248850346,-0.0539350286,0.3341398239,-0.0373083353,0.0551678725,0.2743198574,0.132389158,0.1528703719,-0.1730083674,-0.1509543061,0.1343629956,0.424654305,0.2246306837,0.2432980984,-0.2240974605,-0.0863253325,-0.1576269269,0.4860997498,0.2494944334,-0.0847699344,0.3280572295,0.1216636822,0.0873245895,-0.4062583745,0.4440463185,0.4752523601,-0.017964622,0.2364118397,-0.0518491827,0.0944416001,-0.0688727126,-0.1788314581,0.1537085474,0.0425412729,0.1513472944,0.3643048704,0.1570621133,0.2181692421,-0.05423785,-0.2468826324,0.5586373806,-0.0006456364,0.3511783481,0.7551176548,-0.0354601964,0.1235647127,0.2820723951,0.0567203872,0.1068036407,0.1088804752,0.0239949021,0.239390254,0.0669956207,-0.0123410923,0.2862534821,-0.1619566977,-0.0269901603,0.2715644836,0.048706878,0.1325716823,-0.0262092724,0.271527797,-0.0600212403,0.1020961255,-0.2493344992,0.2315691262,0.0015934512,-0.2097304463,-0.4564832449,-0.2576331794,-0.1872611642,-0.0849312246,0.0605744496,-0.4115421176,0.3470729589,-0.3500343859,0.2807442546,0.0230730474,0.4015862346,0.0115210088,0.0562061071,-0.0486792289,0.0392043367,0.1939410567,-0.0702420101,0.233079493,0.4043902755,0.446485132,0.2524724901,0.0869973674,-0.1364696622,-0.1309299469,-0.0494083799,0.0036351937,0.153779313,-0.3043949902,-0.1800197959,0.220889166,0.1544551402,-0.002964322,-0.0388854183,0.1989026219,0.0792414472,-0.1073250994,0.5643259287,-0.2565623224,-0.1210433617,-0.0079535702,0.0975218639,-0.350130856,-0.3489460051,0.2783760428,0.2449503541,0.0525118224,-0.3614912331,0.0828199834,-0.0364129916,0.5925099254,0.0213758722,0.042278707,-0.1893229783,-0.2608793378,-0.4184709787,-0.1952403933,-0.0120758042,-0.3624277115,0.1239404604,0.2639778554,0.0511849597,0.0453204215,-0.1107774526,0.0066592176,-0.0647141188,0.3001202047,-0.2498517185,-0.1399869025,0.0974658281,0.1280575991,-0.1126329973,-0.0070935055,0.1019563228,-0.010543338,-0.0253552217,-0.0875792652,0.1269019991,0.3288878202,0.3355790675,-0.1888049841,0.2449076921,0.2353849858,-0.0825102106,-0.291090399,-0.2382965237,0.16945678,-0.3728818595,0.4929943979,-0.0161866732,0.1797351837,-0.0702203512,-0.2198335379,-0.2998581231,0.2033131421,0.0825977549,-0.0941113383,-0.237261042,0.2453476638,-0.4747853875,-0.0209367946,0.2546732426,0.3233121037,-0.1294375062,0.0668988973,-0.0858339593,-0.3203435242,0.4812019765,-0.2044383585,-0.3090689182,-0.1241493747,0.0593836159,0.0217447374,-0.3169500828,-0.5852509737,-0.0643530264,0.2532244027,-0.111368008,-0.1404604763,0.1280755401,0.0493841022,-0.298943311,-0.1453455389,0.1923640072,0.1709427536,-0.3089118004,0.1824216843,-0.1736833751]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2009","title":"Ambiguous documentation","comments":"Oh ok I hadn't made the connection between those two, will offer a tweak to the comment and the template then - thanks!","body":"https:\/\/github.com\/huggingface\/datasets\/blob\/2ac9a0d24a091989f869af55f9f6411b37ff5188\/templates\/new_dataset_script.py#L156-L158\r\n\r\nLooking at the template, I find this documentation line to be confusing, the method parameters don't include the `gen_kwargs` so I'm unclear where they're coming from.\r\n\r\nHappy to push a PR with a clearer statement when I understand the meaning.","comment_length":23,"text":"Ambiguous documentation\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/2ac9a0d24a091989f869af55f9f6411b37ff5188\/templates\/new_dataset_script.py#L156-L158\r\n\r\nLooking at the template, I find this documentation line to be confusing, the method parameters don't include the `gen_kwargs` so I'm unclear where they're coming from.\r\n\r\nHappy to push a PR with a clearer statement when I understand the meaning.\nOh ok I hadn't made the connection between those two, will offer a tweak to the comment and the template then - thanks!","embeddings":[0.0907537043,-0.1784818918,-0.0821813196,-0.1347765923,0.2026104182,0.1043046713,0.4321020842,0.0742146522,-0.0925442725,-0.1297511011,0.1296102107,0.2071546614,0.0149755012,0.1009712443,0.3126719594,0.0319476649,0.2020220309,0.0814612806,-0.1270157397,-0.1860020608,-0.3509488404,0.1204457358,-0.2026405036,0.2474186122,-0.3214331269,-0.0715385154,-0.0957001448,0.3153885305,-0.380150795,-0.3472014368,0.3936366141,0.2135883868,-0.1184881628,0.1507874429,-0.0001064841,0.0313349441,0.4308875203,-0.1848325431,-0.1925987154,0.2388115823,-0.2063976526,-0.2435216606,0.0402131043,-0.3196239471,0.0919546187,-0.2830057442,0.1082213372,-0.0474153124,0.3947280943,0.2065674663,0.2524823844,0.0181180071,0.0514019579,-0.2937716544,-0.0214617737,0.3528165221,-0.2551085949,0.1114547998,0.1892898679,-0.0178190805,-0.2224143296,0.3152564764,0.0603582934,-0.2014649063,0.5262850523,-0.0914600566,0.1253803372,-0.3355511725,0.2306743711,0.1955230981,0.6530069709,-0.2880720198,-0.3390621841,-0.1587803215,-0.0159975514,-0.1725635082,-0.1106833965,0.1456159502,-0.4438188076,0.0985820666,-0.3290769756,-0.2995276451,-0.3381352425,-0.1893777251,0.1041536853,0.3985577524,-0.1467222124,0.0287834778,-0.079810366,-0.0637106225,0.0606430843,-0.0514689386,-0.0620313957,0.4795161188,0.0475221723,-0.0571064651,0.3913280666,0.3072203398,0.4849086404,0.3319509327,0.0647529662,-0.1249570772,0.0482135676,0.1219810545,0.1688570827,-0.0952121541,0.2625269592,-0.1981397867,0.1932673156,0.0716884732,0.219352603,-0.1851039827,0.1651634425,-0.4140134454,0.096027039,-0.1382640302,0.3235099018,-0.2193565518,-0.0478596203,0.0212953109,0.1541705281,-0.2329805791,0.0033720983,0.1460006982,0.0977548808,-0.0294088554,-0.0267413203,-0.0824776515,-0.1639438123,-0.0177769605,-0.33675161,0.2282175273,-0.2819629014,0.2159061134,-0.2383446097,-0.1050835401,0.3639842868,0.2932413518,0.2600688338,0.3762291372,-0.1156295836,0.2062501162,0.1441384554,0.2639551163,-0.30725196,0.1014636084,0.15148139,0.097640872,-0.21210967,0.0222522859,-0.3375568688,-0.4068917036,0.033155825,0.1558642685,-0.2776870728,-0.0517162718,-0.0528339557,0.4435343444,0.1255674064,-0.1862482131,0.043944709,0.0011063695,-0.0643742234,-0.4972919822,0.3103567958,0.4556650817,-0.300209254,-0.201537326,-0.0556298494,-0.258897543,0.1801887602,0.1800517738,-0.2338027507,0.115781568,-0.2990383506,0.2452955097,0.4009967446,-0.0305048618,0.1852113158,0.2886868417,0.2408436984,-0.0456033088,0.2281843424,-0.1671770811,-0.1202177703,-0.1961425096,-0.1287868172,-0.0771344081,0.0493371077,-0.1397655308,-0.1034402624,-0.0711304918,-0.0705191121,-0.1175256222,-0.1683060378,0.0547597893,0.0953352898,0.101554729,0.393232435,-0.2345935851,0.1285251677,0.3467071354,0.4189898372,-0.1288628429,0.0794210732,0.0565179512,-0.6715270877,-0.0466049649,0.0131097082,0.2423960418,-0.1882204562,-0.2485075593,-0.4149152637,0.0028945433,-0.0466381311,-0.2650403976,0.149155423,-0.0113790995,0.1515475214,-0.1366390139,-0.1981600374,-0.0785095468,-0.0282414369,0.011772505,-0.4073155522,0.126412645,-0.0308604874,0.0549484864,0.0660938174,0.1211642995,0.1455517113,0.0127390949,0.2302755415,0.3799113333,0.2193588614,0.0544941276,0.1166277677,0.367143631,-0.1192397103,0.0965395421,0.2640690506,-0.0831126943,0.1438301951,-0.0361757167,-0.3096537292,0.2935355604,0.1455044299,0.1645067036,0.1982659847,-0.0875326321,0.1234214157,0.1111753657,-0.1437384039,-0.1031609476,-0.2327030599,0.1511912048,0.4141843617,0.094940789,-0.3783423901,0.4364842176,0.2460189909,-0.2534933388,-0.1884205341,0.1434516311,0.0920640752,-0.4229130745,-0.0424715467,-0.0089846468,0.4792378545,0.1818936318,0.2370540947,0.1093299687,-0.2840774655,-0.3740889728,0.2590414286,-0.0067262975,-0.0551690087,0.02403065,-0.1259056628,-0.2680931985,-0.2931367159,-0.0103911245,0.0646330044,0.2418514639,-0.4246813059,0.1512791514,0.0138725247,-0.0488583334,-0.2073214501,-0.132053405,-0.367715776,-0.5768344402,0.0648537576,0.0012631109,-0.2023102194,0.0999623239,-0.1299651265,0.3149814904,0.0695011169,0.2884173691,-0.3500846326,-0.2509894371,0.0356564596,0.1521712393,-0.0046584504,-0.1069710329,0.3328808546,-0.1114894897,0.0681734979,-0.438057214,-0.5030035973,0.1767327338,-0.3395259678,0.3017337918,0.5701349378,0.4036961198,0.0565251261,-0.0830231905,0.0947981849,-0.1777533293,-0.2436972857,-0.2531808615,-0.1625373214,0.0825258419,-0.3432886004,-0.3473718464,-0.0341793224,-0.2782125473,0.2322392166,0.0811353996,0.1493464857,0.4569232464,0.1185975596,0.1602661461,0.0463690646,0.1483184695,-0.2061444074,-0.3614016771,0.3522810936,-0.3719358444,-0.2803904116,0.0186990555,0.0436760001,0.4063167274,-0.0261209644,-0.2084701061,-0.4375233948,-0.0988153741,0.2255115807,0.0795594454,0.463264823,0.1754227877,-0.1288135052,-0.1690650284,-0.3697242439,-0.1438215226,0.0056055249,-0.5320705771,0.1350747496,0.116126202,0.2016551346,-0.1373723596,0.6183996201,0.1834114045,-0.0393332243,0.0863244534,-0.235701263,0.6174334884,-0.1425574422,-0.0519838259,0.1715782732,0.0455174148,-0.0295404308,0.201607585,0.1906784624,0.0252240729,0.0489925221,-0.0591112971,-0.0439601466,-0.3756275177,0.0385443047,0.2568562925,0.1877179444,-0.0145306736,0.1676348299,-0.2716947198,0.0075926031,-0.1225030646,0.4069201052,0.4516637623,0.1684549004,-0.2381351888,0.0187007543,-0.4803547859,0.2803091109,-0.0482406802,-0.0525970273,-0.1732232869,-0.2888436913,0.1235279739,-0.0884486735,0.2902471423,-0.2935277522,-0.2524642944,0.1743180007,0.1187512055,-0.3780834675,-0.0156439822,-0.1283370852,0.264836967,0.2669674456,0.0926482379,-0.5838429928,-0.4488579929,0.0871472135,0.0339722894,-0.0944151133,0.0418097116,-0.2512210011,-0.6246233582,-0.0316525213,-0.132545799,0.0676612258,-0.0896012262,-0.2021629512,0.3516410887,0.2178480327,0.1741161942,0.1369447857,0.3481551409,0.1863053739,0.0585366674,-0.1171579584,0.0357602686,0.2629088163,0.1564124227,0.3872846961,-0.2450017929,-0.1556182057,-0.0913923159,-0.2430649698,0.5037791729,0.4062492847,-0.190799728,0.1011988893,-0.068062678,-0.0576713346,-0.2223553658,-0.0498010851,0.2901691794,-0.1759475917,-0.3967424929,-0.8526316881,0.5501234531,0.2221503705,-0.3509260416,0.123376824,0.386431545,-0.2540650964,-0.104256615,0.2390331924,0.9269068241,0.4665981531,0.3813878298,0.5670026541,-0.0193574764,0.3943982124,0.2751504779,-0.0854454488,-0.5506343246,-0.1121298671,-0.1418149173,-0.1199333891,0.1193922609,0.3322307169,-0.0728058368,0.2612597942,-0.2402741462,0.1152075753,-0.128375262,-0.0113827866,-0.1996078789,-0.087793313,-0.3388989568,0.2031575441,0.0803393424,-0.0278080553,-0.1115986109,-0.2137956172,0.0383401774,-0.3241895139,-0.3365114331,-0.0112345619,0.1663067043,0.1232651174,0.1898578405,-0.1164369732,0.1951514632,0.0620344356,0.3750985861,-0.0663105845,-0.2711077034,0.1484338939,-0.1516664177,0.0322642773,-0.0726146623,0.2096664906,0.5035306215,-0.0155192623,-0.1664477587,0.0706080049,-0.037519455,-0.0495993458,-0.5324882865,0.0037848817,-0.3056996167,-0.5237842798,-0.2686621249,-0.0552678406,-0.1828413308,-0.1046867594,0.1207876951,0.2544153929,0.0414539054,0.2985797822,-0.1615923494,-0.0102480873,0.0130823832,0.0546991751,-0.0894687325,0.0576743074,0.3235442638,0.0554719828,-0.3141049147,-0.3113207221,-0.0027128465,0.5453394055,-0.5239105821,0.125055477,0.4224225283,0.185546428,0.0352303535,0.2699256241,-0.1150508896,0.1045576558,-0.1019812003,-0.2470863611,-0.2766984701,0.3957057595,-0.215226531,0.2896230221,-0.042039115,-0.0012257382,0.2128677964,-0.065003708,-0.4069843888,-0.0639204755,0.2035140395,-0.0313216634,0.3573608696,0.028350288,0.1642309427,0.0997427255,0.0617151409,-0.1078157648,-0.4556204975,-0.1732759625,-0.162059769,0.1470470428,-0.0821726546,-0.1966527104,0.2791030109,0.0727943182,-0.3431844711,-0.0373159908,0.4923642576,0.0939225331,-0.0151705779,0.2618499398,0.0440334082,0.1581651121,0.1447545588,-0.0478176735,0.2307638377,0.2744841874,0.1432128251,0.0781847015,-0.3892549276,-0.1141836867,0.3225949109,0.5643659234,0.2307515889,-0.0074930489,0.2125400305,0.113669008,-0.0127276275,-0.1772308201,0.3757170737,0.3366527259,-0.1439671367,0.1291634142,0.0873715058,0.1960113347,-0.2959816158,-0.1362830549,-0.0016902944,0.001386413,0.1935544759,0.2963780165,0.1171802953,0.0232157186,-0.0299497787,-0.163060382,0.4087218046,-0.1278822124,0.0309389923,0.4737963378,-0.0006789952,0.0457710624,0.3533280194,0.2377202213,0.0006729952,0.1627083272,0.0224441029,0.1393559426,0.0614362769,-0.0373765267,-0.0052464809,-0.2839750051,0.0158074908,0.3037021458,0.0354409628,0.0836520866,0.055017978,0.3576849997,0.2673793137,0.0782653019,-0.1751665026,0.2874327004,-0.0290824287,-0.3568983972,-0.2810841203,-0.2341429442,-0.056882333,-0.203129828,-0.056024611,-0.3469703496,0.3962552547,-0.2842955589,0.3317706585,-0.0262897611,0.2526832521,0.0308219809,0.213883549,0.0191823635,0.1056705415,0.3438788354,-0.0920378715,0.3848397136,0.4586820006,0.6834434867,0.2331957966,0.2456308752,-0.1129719391,-0.1820154786,-0.0338710137,-0.0387642682,0.1852032393,-0.1110441685,-0.1605175883,0.2998470962,0.1813001931,-0.0675622821,0.1590365618,0.1981128454,0.149173528,-0.120693624,0.4673968554,-0.1642167121,-0.2568279505,-0.1632745713,0.200333029,-0.3464716375,-0.1483396739,0.2493664622,0.2265753597,0.0990869328,-0.2746402621,0.1339313388,0.1253910512,0.703727901,0.1380413175,0.0544935837,-0.0764225274,-0.1834016889,-0.5739237666,-0.1766334325,0.111469321,-0.4577989876,0.0147802159,-0.0284786988,0.1749310791,-0.0576868728,-0.022850791,-0.0576025359,0.0830661729,0.0520929843,-0.3459965587,-0.0429475829,-0.0165246669,0.0558005609,-0.0455079749,-0.0307319704,0.0758171454,-0.0435290039,0.0906403139,-0.0488402508,0.2820123434,0.1464838982,0.3140290082,-0.048904188,0.2010912597,0.4074599445,-0.1519566476,-0.323446095,-0.3913858533,-0.0676136687,-0.3365273774,0.4239864647,-0.0204095151,0.0350309461,-0.0516266786,-0.4418330491,-0.193015784,0.3930758238,0.0923602283,-0.1800706983,-0.1732862294,0.249571383,-0.2952673137,-0.2162275612,0.0389366671,0.2992434502,-0.0870982781,-0.0171059836,-0.0439769588,-0.5386336446,0.7789410949,-0.2815496027,-0.2707871795,-0.1414672583,0.178096652,-0.0699508786,-0.2371910661,-0.6841182709,0.1179747209,0.1855073571,0.0775523707,-0.2221452147,0.2276582271,-0.1803615391,-0.2749416232,-0.1646128893,0.2249433249,0.1434904337,-0.2990719676,-0.0570985638,-0.2372339815]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2007","title":"How to not load huggingface datasets into memory ","comments":"So maybe a summary here: \r\nIf I could fit a large model with batch_size = X into memory, is there a way I could train this model for huge datasets with keeping setting the same? thanks ","body":"Hi\r\nI am running this example from transformers library version 4.3.3:\r\n(Here is the full documentation https:\/\/github.com\/huggingface\/transformers\/issues\/8771 but the running command should work out of the box)\r\n\r\n USE_TF=0 deepspeed run_seq2seq.py --model_name_or_path google\/mt5-base --dataset_name wmt16 --dataset_config_name ro-en --source_prefix \"translate English to Romanian: \" --task translation_en_to_ro --output_dir \/test\/test_large --do_train --do_eval --predict_with_generate --max_train_samples 500 --max_val_samples 500 --max_source_length 128 --max_target_length 128 --sortish_sampler --per_device_train_batch_size 8 --val_max_target_length 128 --deepspeed ds_config.json --num_train_epochs 1 --eval_steps 25000 --warmup_steps 500 --overwrite_output_dir\r\n\r\n(Here please find the script: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/run_seq2seq.py)\r\n\r\nIf you do not pass max_train_samples in above command to load the full dataset, then I get memory issue on a gpu with 24 GigBytes of memory.\r\n \r\nI need to train large-scale mt5 model on large-scale datasets of wikipedia (multiple of them concatenated or other datasets in multiple languages like OPUS), could you help me how I can avoid loading the full data into memory? to make the scripts not related to data size? \r\n\r\nIn above example, I was hoping the script could work without relying on dataset size, so I can still train the model without subsampling training set.\r\n\r\nthank you so much @lhoestq for your great help in advance\r\n\r\n\r\n","comment_length":36,"text":"How to not load huggingface datasets into memory \nHi\r\nI am running this example from transformers library version 4.3.3:\r\n(Here is the full documentation https:\/\/github.com\/huggingface\/transformers\/issues\/8771 but the running command should work out of the box)\r\n\r\n USE_TF=0 deepspeed run_seq2seq.py --model_name_or_path google\/mt5-base --dataset_name wmt16 --dataset_config_name ro-en --source_prefix \"translate English to Romanian: \" --task translation_en_to_ro --output_dir \/test\/test_large --do_train --do_eval --predict_with_generate --max_train_samples 500 --max_val_samples 500 --max_source_length 128 --max_target_length 128 --sortish_sampler --per_device_train_batch_size 8 --val_max_target_length 128 --deepspeed ds_config.json --num_train_epochs 1 --eval_steps 25000 --warmup_steps 500 --overwrite_output_dir\r\n\r\n(Here please find the script: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/run_seq2seq.py)\r\n\r\nIf you do not pass max_train_samples in above command to load the full dataset, then I get memory issue on a gpu with 24 GigBytes of memory.\r\n \r\nI need to train large-scale mt5 model on large-scale datasets of wikipedia (multiple of them concatenated or other datasets in multiple languages like OPUS), could you help me how I can avoid loading the full data into memory? to make the scripts not related to data size? \r\n\r\nIn above example, I was hoping the script could work without relying on dataset size, so I can still train the model without subsampling training set.\r\n\r\nthank you so much @lhoestq for your great help in advance\r\n\r\n\r\n\nSo maybe a summary here: \r\nIf I could fit a large model with batch_size = X into memory, is there a way I could train this model for huge datasets with keeping setting the same? thanks ","embeddings":[-0.1572741419,-0.5049949884,0.0060410541,0.5300241113,0.558508575,0.0471684635,0.0790678412,0.2601298988,0.3925331831,0.1179506406,-0.0250124354,-0.2572232187,-0.3355440795,0.3502956033,0.0798836127,-0.1671071649,0.0480975658,-0.0093382346,-0.4906504154,0.0814422742,-0.2567529678,-0.058337044,0.023544712,-0.3496483266,-0.3608575165,0.0781597048,0.0802220479,0.0628451556,0.0150360418,-0.000408594,0.0273320936,-0.0705826581,0.1996215731,0.4450730085,-0.0001240301,0.0955719501,0.1058853641,-0.3430503905,-0.0113493083,-0.0202338677,-0.0401825197,-0.1584909409,0.1759109944,-0.096055828,-0.1249656156,0.1642036289,0.0570753813,-0.1192683429,0.6155748367,0.0377041698,0.0811976343,0.1557729691,-0.1867898107,0.1391717345,0.0997060239,0.1715507805,0.0613706894,0.1924704015,-0.1117654741,0.0804736465,-0.0205961969,0.3296141624,-0.155104056,-0.160643369,0.3712903559,-0.029681623,0.0977140665,-0.4963965714,0.0698852167,0.2278073281,0.3327106535,-0.1846735328,-0.2815987468,-0.2376737297,-0.2664910555,-0.2981759012,0.035622444,0.3874962926,-0.1627117395,0.195293054,-0.5647282004,-0.3583250344,-0.1173302382,-0.0793804303,0.0170941427,-0.1479891241,-0.1117734462,0.0931011289,0.3886768222,-0.1748268157,-0.3857375979,0.0535067841,0.1244651005,0.2021829039,-0.4465097785,-0.2754250467,-0.0981329158,-0.0126574812,0.519705236,0.0696831793,-0.0840645134,0.1435506791,0.3678051829,-0.0124563091,0.255928874,0.2244566828,-0.3579405546,-0.0226911176,0.3058372438,0.1548227966,0.0666357279,-0.098081179,-0.0483796895,-0.1039824113,-0.1187516674,-0.2178400904,0.1938429177,-0.2472089976,0.1151949912,-0.1014178842,-0.0034157014,0.145704478,-0.0707054362,0.572339952,-0.1075777933,0.2447873652,-0.0547052473,0.1319083273,-0.4047715664,0.0199621227,-0.1732313633,0.1808792502,-0.2549785376,0.3128548861,0.3761709332,0.107735604,0.3884443641,-0.1295557618,0.0667767823,-0.0720064566,0.0200995356,-0.3184742033,0.0743142813,0.4120983183,0.0514300279,0.1046668813,0.0525585301,0.2978248894,-0.2142996192,0.304782331,-0.1877875477,-0.3000202179,0.4647508562,0.0260905921,-0.3193106651,0.2963601053,-0.5062684417,0.2329648733,0.1344292611,0.2149205804,0.0263023563,0.0652706623,-0.3889720142,-0.0968876854,0.4445872605,0.7276234031,-0.0871028155,-0.3735770285,0.2625146508,-0.0919227377,0.0998731926,0.6178457141,-0.0144170113,0.0265060142,0.0063410676,-0.3365337849,0.0979604498,-0.1402827352,-0.3693682253,0.2471103221,-0.2083163708,0.1737016737,0.2634863257,0.1960675567,-0.1570897698,0.1254315078,-0.0142418714,0.4411009252,0.0268202163,0.2472616136,-0.1398960054,-0.531604588,0.1208485067,0.2106387019,0.2256649435,0.0572575927,-0.1509833783,0.5526144505,0.0886667967,-0.1365785748,0.1413095146,0.4671052098,-0.0440808572,-0.1722710431,-0.0133903157,0.0327295996,-0.5094329119,0.2002456039,-0.2553119659,0.0325731896,-0.1079561561,-0.0101286117,0.0486869588,-0.0930219591,0.1028227136,-0.1937768906,-0.0928160921,-0.3027759492,0.1868076026,-0.0771659762,-0.0579810031,0.1618708968,0.053369429,0.280700326,-0.629330039,-0.1864142716,0.0876695886,-0.0823396966,0.050921727,0.0869900435,-0.0571699217,-0.0752620697,0.1155380532,0.1845800579,0.0681403503,0.0221212376,-0.2876875401,0.5894073844,0.264546454,-0.215177685,0.1873768568,0.0568239279,0.0807592347,-0.0167980064,-0.2281123698,0.2756064832,0.0014033815,0.3533557057,0.0675171167,-0.2433751971,-0.0153757986,-0.2704691887,-0.0966262445,-0.3189077377,0.1948177367,0.3701253533,0.2904979885,0.3180954158,-0.1097561568,-0.6400544047,0.4141535163,0.0285229627,0.0119818039,0.1240218878,-0.4441150427,-0.0441276319,-0.1043968573,-0.429769218,0.2082951069,0.2234414667,-0.0286325254,0.1574373841,0.4532974362,-0.0383554958,0.1597278118,0.3495739996,0.2293863446,-0.0217218716,0.0718977749,-0.1523364782,-0.1377293617,0.1474387944,0.0416091681,-0.2868335545,-0.3908244371,0.0496986583,-0.1045076549,-0.1814420223,-0.221285969,0.2457416505,-0.4682894647,0.1405825913,-0.0243129507,0.1828608662,-0.0790765882,0.1298356652,0.1344469488,0.4491246045,0.2074004412,-0.3838298619,0.0398870334,0.0210683066,0.0268770084,-0.0198071972,0.1581047624,-0.1872714907,0.2134438306,0.1074101627,-0.2879982293,0.0881509483,-0.1072518155,0.2509253025,-0.1020286605,0.340228796,-0.114542231,0.172845304,0.1816582978,0.0474162735,-0.1080527455,-0.1735255271,-0.0394937284,0.096226804,0.0592295937,0.3814143538,-0.2071724087,-0.1616590619,-0.0630643293,-0.4113830626,0.5235686898,0.1085386351,-0.0144699048,0.3826697469,0.2527718842,0.0553608239,-0.2323665768,0.2439751923,-0.045403704,-0.1866938621,0.4161210656,-0.0248939004,-0.2029857635,0.1008592024,-0.1128142476,0.1596003622,0.2732194662,-0.589743793,-0.0266855899,-0.2009775788,-0.3303202987,-0.0016956278,0.1910510361,0.2152054161,-0.3849983811,0.0567036793,0.1027528867,-0.0541191697,0.1080615744,-0.0217945501,0.3186091483,0.133681342,0.5374333858,-0.0703099594,0.8954889774,0.087374635,0.2348245382,0.0591824278,0.0127950748,0.1876633763,-0.1213639006,0.0318818688,-0.102543816,-0.115897648,-0.1761571765,0.2353009433,-0.0091061546,0.1402727962,-0.2111800611,-0.116071105,0.2083765864,-0.3471627831,0.3769301772,0.0272887759,0.0852973685,-0.0675217062,-0.0631651357,0.0236641802,-0.2326041758,0.0497847125,0.0912872404,0.2584247887,0.1641099751,-0.1399655938,-0.0377403833,-0.6823832393,0.0472901762,-0.1628506631,0.0705491379,-0.0118344799,-0.3628765643,-0.1011050493,0.1495700181,0.7003593445,0.2093114555,-0.0992958918,-0.1513701975,-0.6746287346,-0.3832984865,0.008753987,-0.1994628459,-0.1459478438,0.1336379796,0.3363019526,-0.2980456054,-0.1955032051,0.1403114498,0.2218353003,-0.2850463092,-0.1815048903,-0.4925501943,-0.006113044,-0.437033385,0.1137482077,0.3473318815,0.088776052,-0.1772831082,0.0635333136,-0.0417181738,-0.041481629,0.3954689801,-0.0079489425,0.2531649172,0.0058498052,-0.0506543219,0.1109766066,0.0996195525,-0.0075106537,0.6166421771,0.2945856154,-0.2460392267,0.1950681359,0.2299350053,0.4992121756,0.218608737,-0.0731270984,0.0601154007,0.1847862452,0.3394141197,-0.4147107899,0.3623566329,-0.0575157218,-0.1000472307,-0.2485262007,-0.2642414272,0.5508927107,0.2732253075,0.1602786034,-0.0658206642,-0.0929712653,-0.3220253885,0.2822587788,0.1427023113,0.9183750153,-0.2307801098,0.4209938943,0.0939180478,0.3291245103,0.491990298,-0.3210009038,0.2628663778,-0.1850237399,-0.3845914304,0.0541864038,-0.2425593287,-0.0895833075,0.2086640894,-0.1083811224,0.5269124508,0.0039228615,0.0559044704,-0.0358922109,0.2178502828,0.048001159,-0.5682682395,-0.3984393775,-0.0315095149,0.0182681512,0.0549058206,-0.0910824314,0.0196431857,0.3000813127,-0.2966687083,0.0915757045,0.0470600761,-0.6259862781,-0.0620192699,-0.1373541504,0.0400002636,0.2674304247,0.1379145533,-0.0757274777,0.3967711627,-0.1971822083,-0.1352535337,-0.2323102951,-0.0285188667,-0.2319064736,0.136667043,0.2972686887,0.0881888568,-0.0716520771,0.0905366614,-0.0424038023,0.0052065449,-0.0677170828,-0.329854399,-0.135856688,-0.2493791282,-0.2000972927,-0.0352988616,-0.0061112754,0.1681891829,0.0240746904,-0.0144768739,-0.2260143608,0.1662599444,0.1755396724,-0.4944987595,-0.1915171146,0.6306126714,0.058146026,-0.1318102181,0.3846061528,0.268207252,0.0156310163,-0.1055140272,-0.0801856518,0.4531804621,-0.402428031,0.2040705234,0.133473888,-0.1669792533,-0.2300867736,0.176057145,0.0783464685,-0.2996075749,0.0395554528,-0.0767427385,-0.6522591114,0.5526909828,-0.1600833237,0.1457242966,0.2210736424,0.0098228818,-0.1001027375,0.3121766448,-0.1771322638,-0.1239199415,-0.2733043432,0.0264998637,0.1547726989,-0.1664403081,0.2666933537,0.1607721597,-0.0716533288,0.1832024157,-0.2298360318,-0.1442578286,-0.0083224438,0.1810944825,-0.0127280969,0.0924256667,-0.0273558721,-0.0063771377,0.0125058172,-0.173469305,0.6875865459,0.0699365586,0.1886780262,0.0807864144,0.0989773795,0.025180921,-0.3838288784,0.1272808909,0.0643460825,0.1245058402,0.3593574762,0.0459397249,0.0412662774,-0.0036158629,-0.4344602525,-0.0119325528,0.448880434,-0.1360449046,-0.0200086776,-0.0839243457,-0.2861672938,0.2059268504,0.0695195869,0.1605725586,-0.2592917681,-0.0067391549,0.1479416341,0.0630949661,-0.0758257508,0.0062517282,0.126900509,-0.341454953,0.0104466714,0.6033747196,0.1643061638,0.0432601348,0.0644857734,0.1293430775,0.2971199751,0.0488731638,0.0711292401,0.0600033849,0.0165282283,0.0223468319,0.3840658665,0.4498668909,0.4290983975,0.1533199251,-0.0152008543,0.2711749077,-0.6284002066,0.2273329049,-0.1653504372,-0.5089746714,-0.324837923,0.4284764528,-0.2122273743,0.1208923385,-0.4700190723,0.0007581296,0.3411900997,-0.1815993041,-0.2354870588,0.2056088746,-0.2333950847,-0.1625028104,0.2038331032,-0.041026406,-0.2441984117,0.3038098514,0.0528432913,-0.3602623045,0.2383437902,0.2009576857,-0.0919383839,0.1840369254,-0.0409347117,0.3867111802,-0.0109364958,-0.3442848027,0.0856594369,0.1658646017,0.136633262,-0.038855955,0.0982627496,0.0975973234,0.037500076,-0.0968802348,-0.0132484818,0.2822808921,0.0023819371,-0.193320632,0.189728871,0.1748276055,0.5260715485,0.0271383356,0.0026788162,-0.0231288504,-0.0301114134,0.007849534,0.1332119703,-0.4083077013,0.110594593,0.3215964437,0.012033443,-0.2242650688,0.1360995471,-0.5450908542,0.1193032414,0.3821794391,-0.3217529655,0.0950526074,0.1002370119,-0.0032947571,-0.2248201668,0.6864812374,0.1166609526,0.0842610002,-0.5442641973,0.189744696,-0.1960073113,0.0287146922,-0.0612842776,0.3867122531,0.14481242,0.2428366095,-0.0364580825,0.1784197241,-0.4886824191,-0.0889457539,-0.3430784643,-0.1544513255,-0.1451508254,-0.1504172683,-0.4410767555,0.3365375102,0.1452553123,-0.1759848595,0.0739889145,-0.1276941597,-0.2232112586,-0.1174627915,-0.0170446504,-0.0372445546,0.3146273196,0.2162883878,0.1040515527,0.1622670889,-0.1483354419,-0.1856825352,-0.1942181736,-0.036262244,-0.240635097,-0.2403424084,-0.0347794108,0.2431439757,-0.2745268941,0.0526010692,-0.2590027153,0.0730430037,-0.0146215735,-0.3016211689,0.0945670977,-0.1222135276,0.0813927576,0.2626322508,0.1797933578,0.385515511,-0.2010323107,-0.0622668639,-0.1728283763,-0.0252196565,0.4078325629,-0.2210169733,-0.1623037308,-0.2203921676,0.2522371113,-0.1805445552,-0.0796525329,-0.5692534447,0.2579885721,0.0481676906,0.1522705555,-0.3700564504,0.4148771763,-0.0936177894,-0.1156240627,-0.2093270123,0.1317287087,-0.014186766,-0.1515181363,-0.2511226535,-0.3791803122]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2007","title":"How to not load huggingface datasets into memory ","comments":"The `datastets` library doesn't load datasets into memory. Therefore you can load a dataset that is terabytes big without filling up your RAM.\r\n\r\nThe only thing that's loaded into memory during training is the batch used in the training step.\r\nSo as long as your model works with batch_size = X, then you can load an even bigger dataset and it will work as well with the same batch_size.\r\n\r\nNote that you still have to take into account that some batches take more memory than others, depending on the texts lengths. If it works for a batch with batch_size = X and with texts of maximum length, then it will work for all batches.\r\n\r\nIn your case I guess that there are a few long sentences in the dataset. For those long sentences you get a memory error on your GPU because they're too long. By passing `max_train_samples` you may have taken a subset of the dataset that only contain short sentences. That's probably why in your case it worked only when you set `max_train_samples`.\r\nI'd suggest you to reduce the batch size so that the batches with long sentences can be loaded on the GPU.\r\n\r\nLet me know if that helps or if you have other questions","body":"Hi\r\nI am running this example from transformers library version 4.3.3:\r\n(Here is the full documentation https:\/\/github.com\/huggingface\/transformers\/issues\/8771 but the running command should work out of the box)\r\n\r\n USE_TF=0 deepspeed run_seq2seq.py --model_name_or_path google\/mt5-base --dataset_name wmt16 --dataset_config_name ro-en --source_prefix \"translate English to Romanian: \" --task translation_en_to_ro --output_dir \/test\/test_large --do_train --do_eval --predict_with_generate --max_train_samples 500 --max_val_samples 500 --max_source_length 128 --max_target_length 128 --sortish_sampler --per_device_train_batch_size 8 --val_max_target_length 128 --deepspeed ds_config.json --num_train_epochs 1 --eval_steps 25000 --warmup_steps 500 --overwrite_output_dir\r\n\r\n(Here please find the script: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/run_seq2seq.py)\r\n\r\nIf you do not pass max_train_samples in above command to load the full dataset, then I get memory issue on a gpu with 24 GigBytes of memory.\r\n \r\nI need to train large-scale mt5 model on large-scale datasets of wikipedia (multiple of them concatenated or other datasets in multiple languages like OPUS), could you help me how I can avoid loading the full data into memory? to make the scripts not related to data size? \r\n\r\nIn above example, I was hoping the script could work without relying on dataset size, so I can still train the model without subsampling training set.\r\n\r\nthank you so much @lhoestq for your great help in advance\r\n\r\n\r\n","comment_length":208,"text":"How to not load huggingface datasets into memory \nHi\r\nI am running this example from transformers library version 4.3.3:\r\n(Here is the full documentation https:\/\/github.com\/huggingface\/transformers\/issues\/8771 but the running command should work out of the box)\r\n\r\n USE_TF=0 deepspeed run_seq2seq.py --model_name_or_path google\/mt5-base --dataset_name wmt16 --dataset_config_name ro-en --source_prefix \"translate English to Romanian: \" --task translation_en_to_ro --output_dir \/test\/test_large --do_train --do_eval --predict_with_generate --max_train_samples 500 --max_val_samples 500 --max_source_length 128 --max_target_length 128 --sortish_sampler --per_device_train_batch_size 8 --val_max_target_length 128 --deepspeed ds_config.json --num_train_epochs 1 --eval_steps 25000 --warmup_steps 500 --overwrite_output_dir\r\n\r\n(Here please find the script: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/run_seq2seq.py)\r\n\r\nIf you do not pass max_train_samples in above command to load the full dataset, then I get memory issue on a gpu with 24 GigBytes of memory.\r\n \r\nI need to train large-scale mt5 model on large-scale datasets of wikipedia (multiple of them concatenated or other datasets in multiple languages like OPUS), could you help me how I can avoid loading the full data into memory? to make the scripts not related to data size? \r\n\r\nIn above example, I was hoping the script could work without relying on dataset size, so I can still train the model without subsampling training set.\r\n\r\nthank you so much @lhoestq for your great help in advance\r\n\r\n\r\n\nThe `datastets` library doesn't load datasets into memory. Therefore you can load a dataset that is terabytes big without filling up your RAM.\r\n\r\nThe only thing that's loaded into memory during training is the batch used in the training step.\r\nSo as long as your model works with batch_size = X, then you can load an even bigger dataset and it will work as well with the same batch_size.\r\n\r\nNote that you still have to take into account that some batches take more memory than others, depending on the texts lengths. If it works for a batch with batch_size = X and with texts of maximum length, then it will work for all batches.\r\n\r\nIn your case I guess that there are a few long sentences in the dataset. For those long sentences you get a memory error on your GPU because they're too long. By passing `max_train_samples` you may have taken a subset of the dataset that only contain short sentences. That's probably why in your case it worked only when you set `max_train_samples`.\r\nI'd suggest you to reduce the batch size so that the batches with long sentences can be loaded on the GPU.\r\n\r\nLet me know if that helps or if you have other questions","embeddings":[-0.1518766731,-0.5147151947,0.0332163535,0.4974522889,0.5210765004,0.0139152762,0.1180271581,0.2357333153,0.4079456329,0.1879961044,0.0076194578,-0.202884227,-0.2850468755,0.3407030404,0.0886425823,-0.103249006,0.0532818362,0.086624898,-0.5714523196,0.052350603,-0.2737412453,-0.0161269214,0.0159723274,-0.3641580939,-0.3886540532,0.1600887328,0.0624867231,0.1186249182,-0.0098638106,-0.0714208186,0.1428127885,-0.0393951349,0.1770171374,0.4372489452,-0.0001269822,0.1079645008,0.1380944401,-0.3148227036,-0.0306409597,-0.0322296172,-0.00419241,-0.2059285641,0.2809613943,-0.1789403409,-0.150394097,0.1292396337,-0.0043970579,-0.1768301129,0.6732686162,0.0866890699,0.0589532889,0.1535408795,-0.2452387959,0.1667440385,0.157634601,0.1976086199,0.0586231574,0.1164000556,-0.104916431,0.1034323648,0.0304610971,0.3367860019,-0.248097226,-0.1271037906,0.412345171,0.0033178255,0.0176058318,-0.4697835445,0.1020778939,0.2507305443,0.4073416591,-0.2239231765,-0.2138442248,-0.2794689834,-0.3462489545,-0.2446087301,0.1370269954,0.3081847727,-0.143369481,0.211253196,-0.5356808305,-0.3474308848,-0.1170952469,-0.0391612761,0.0610946529,-0.1411525458,-0.1147686169,0.118608132,0.4361445904,-0.1669552028,-0.3358583152,-0.0003861996,0.1707512885,0.2230681926,-0.5019465685,-0.3163248301,-0.0657733232,-0.1057263538,0.4721098542,0.1215846986,-0.0796092525,0.0936994851,0.3299092054,0.0091587044,0.2849020362,0.1574461758,-0.3003250659,-0.0072965627,0.2969490886,0.1490688771,0.1434296668,-0.1105317473,-0.0333796069,-0.0507383198,-0.2047497332,-0.259154737,0.2845853567,-0.2782020271,0.0794593021,-0.0892999843,-0.0343100466,0.170126304,-0.0699241608,0.5180566907,-0.0827048868,0.3102358878,-0.0071106786,0.131597057,-0.4338736534,0.0145770172,-0.1941797286,0.2515988946,-0.2240199149,0.3322900236,0.411667794,0.0499283858,0.379703939,-0.0956689641,0.0674586296,-0.0763127133,0.0796634704,-0.3235810399,0.1239142492,0.4826919138,0.0355792567,0.1951072961,0.0725165308,0.1766030341,-0.2136323899,0.396887064,-0.230730623,-0.2883965969,0.363097012,0.0053726593,-0.4209285676,0.2904184163,-0.5321798325,0.2664295435,0.1223289296,0.2021841109,0.0557638854,0.078967303,-0.3984458148,-0.0862805545,0.4873942137,0.8010213971,-0.1476057023,-0.3931111991,0.1329436004,-0.1004578918,0.1959773898,0.5725108385,-0.0142801609,0.0148478178,-0.0144559685,-0.2140351981,0.0709257945,-0.1415775716,-0.354434818,0.2842701674,-0.170895949,0.1742833555,0.2594866455,0.245840773,-0.0926848352,0.1357045472,-0.1018791869,0.3744672239,0.0681137666,0.2252355516,-0.1271794438,-0.5379964113,0.2040785849,0.2169702202,0.2022631615,0.0872878507,-0.1350940913,0.6384108663,0.0238448363,-0.1233225539,0.150255397,0.5214672089,-0.029772684,-0.0962148979,-0.0108517697,-0.0647933111,-0.5006047487,0.1989262551,-0.0586789697,0.0444423407,-0.1048355103,0.0006818472,0.0573625341,-0.1273258924,0.0957001373,-0.1771862358,-0.1211929917,-0.2916912436,0.1514546126,-0.1347030401,-0.0903468654,0.1488622576,-0.0584070943,0.2972849607,-0.6091175079,-0.1088676304,0.1102921367,-0.1234868392,0.0259958319,0.1248041317,-0.0882156715,-0.0845849365,0.1067788377,0.2415024489,0.1124199778,0.0178392511,-0.2222641259,0.6307321787,0.2434305549,-0.3127706051,0.120512858,0.0697668567,0.1251268089,0.0252608154,-0.1785954386,0.1950393319,-0.1331377178,0.3484589159,0.0058566616,-0.2301434278,-0.0358800404,-0.2495764047,-0.1238652468,-0.3175210655,0.2297707349,0.3747276664,0.3159516454,0.3822626472,-0.1301639974,-0.6248890162,0.4318443835,-0.0019924701,0.0073690708,0.1488064826,-0.4452287853,-0.0800923407,-0.0596907996,-0.374764502,0.2052786648,0.2200757861,-0.0135828219,0.1306917071,0.4082516134,-0.051337,0.1627495438,0.3493756354,0.1488740146,-0.0548449643,-0.0312599577,-0.1541599631,-0.1904212981,0.2100085169,0.027010059,-0.2555870712,-0.4284456074,0.054600928,-0.1863786578,-0.1533458978,-0.19714351,0.2054142356,-0.4946786165,0.1187362373,-0.0151108671,0.1751280576,-0.0662437603,0.1509770453,0.1996802539,0.4412361085,0.1918815821,-0.4602290988,-0.0057205306,0.0038936462,0.0292233676,-0.0567871071,0.156275332,-0.1057111621,0.2035379261,0.0814274028,-0.2594136894,0.0631288737,-0.1128867716,0.2416180372,-0.1343260258,0.3047715425,-0.0908702239,0.2014537007,0.1846033186,0.0711597651,-0.1057425439,-0.1495919079,-0.0326353051,0.1421228945,0.0993307829,0.3976770639,-0.3058061004,-0.206270352,-0.0010525942,-0.3736867607,0.4215454757,0.1011799723,-0.0064921095,0.3571302295,0.2769917846,0.0848625675,-0.1952385604,0.295612067,-0.0905101672,-0.2000130266,0.4355374277,0.0201416388,-0.2268256694,0.2055979371,-0.0464503802,0.195723027,0.2565194964,-0.6173763275,-0.0449292623,-0.245066151,-0.3239588737,-0.0527185872,0.2297547013,0.2915318608,-0.3330810666,0.061943084,0.041418232,-0.1222939864,0.0963732973,-0.04288489,0.2026363611,0.2002973109,0.5743064284,-0.0909149274,0.9312599301,0.1241061613,0.1804883778,0.0755283907,-0.0374021158,0.2525525987,-0.1314361542,0.0153975999,-0.1186559275,-0.005834965,-0.1430701613,0.2657731771,-0.0255037062,0.1975001395,-0.2431700975,0.0442302413,0.1530369073,-0.2962439954,0.3551985025,-0.0067105084,0.1253744662,-0.0403264053,-0.0258750208,0.01080928,-0.2412958294,0.0123064369,0.1530932486,0.2456722856,0.1957315058,-0.1906873733,-0.0760546401,-0.6258146167,0.00058075,-0.1969399601,0.0993860513,-0.0586039759,-0.3489128947,-0.0277031455,0.1523522884,0.6768870354,0.253205955,-0.1729893684,-0.1487431377,-0.7294568419,-0.4025163352,-0.0095961746,-0.2291607708,-0.1238157228,0.0738100186,0.3609037399,-0.3914929628,-0.2916172743,0.0733061954,0.2556804717,-0.2727658749,-0.1538665593,-0.5318119526,-0.09207239,-0.4376599789,0.0552046113,0.3393112719,0.087072432,-0.1991936415,0.0410296619,-0.0185950939,-0.094006978,0.413297832,-0.043346405,0.2744604647,-0.0019244759,-0.0136764813,0.1732084155,0.1116560921,-0.0193675105,0.6520434618,0.314073354,-0.1921579391,0.1988084465,0.2186891735,0.4826212823,0.2193189114,-0.1357713491,0.1231534481,0.1372123361,0.2870837748,-0.4672865868,0.3837448657,0.0091922069,-0.0902185887,-0.1464430839,-0.313822329,0.5189355612,0.3223374486,0.154501617,-0.099445425,0.0033903155,-0.2873597443,0.1786357164,0.1093033478,0.9667987823,-0.2780707479,0.4263842702,0.0963516906,0.3156810403,0.6530103087,-0.2819084227,0.2129286677,-0.1729738116,-0.3770285547,0.0020607463,-0.2352292091,-0.026804816,0.2167588323,-0.0766807497,0.4903028011,0.0245775841,0.2099646777,-0.0258128997,0.2198755741,0.1225688457,-0.5316357017,-0.3666524291,-0.0411537811,-0.0330703445,0.0237536784,-0.0148276109,0.0282777045,0.2234194726,-0.3717512488,0.095389165,0.0572575629,-0.4754978418,-0.0344231836,-0.1170828789,0.0182389989,0.3291778862,0.2111402005,-0.0804235935,0.3448342681,-0.2216811776,-0.2169554979,-0.327776581,-0.0226133689,-0.2501816154,0.1860280335,0.2770850658,0.0731896535,-0.0908855051,0.1434045285,-0.029025916,-0.0290762391,-0.0541504882,-0.2843359113,-0.0897052661,-0.2674399018,-0.2032888681,0.0334842056,0.0284731034,0.1863417327,0.0086723557,0.053234525,-0.2010976374,0.1519869566,0.1226778775,-0.497064054,-0.1683167964,0.6233395934,0.0827039778,-0.0550160781,0.4000042677,0.2361922115,-0.0242860094,-0.1113119721,-0.1081455797,0.3746094704,-0.398139745,0.1701200157,0.1448520273,-0.1961875111,-0.2891498208,0.1320085227,0.0303233843,-0.2681396902,-0.0012037213,-0.0147275142,-0.5990898013,0.572332859,-0.1569628417,0.2224693149,0.2193519324,0.0136701511,-0.1739604771,0.2325651795,-0.1629931182,-0.0749226511,-0.3024333119,0.0509450398,0.1844855398,-0.1453210115,0.2424788624,0.0973962322,-0.0883140787,0.1387359351,-0.2038185745,-0.1181400418,0.0385507457,0.205616042,0.0101965247,0.0345210731,-0.0567131788,-0.0300882906,-0.0445922017,-0.2133755982,0.6652671695,0.0559383854,0.151523307,0.2013332248,0.1088072434,0.0293841735,-0.4303299189,0.1400620043,0.0157280397,0.1691092849,0.2504603863,0.0828849375,0.1003697067,-0.0310391709,-0.454180479,0.0629141554,0.4600507021,-0.1720419526,-0.0807558373,-0.0880520046,-0.2382043451,0.2162968069,0.0730879307,0.2468606532,-0.2693391442,-0.0418196656,0.1638191491,0.0239935163,-0.0863865912,-0.0251838882,0.0991352573,-0.3105214834,0.0191547591,0.6123440266,0.1744194776,0.0566273443,0.0871606395,0.1088968888,0.3070230186,0.1072300225,0.0915318504,0.1431955248,-0.0377501249,0.0073172403,0.2631452978,0.4207814634,0.3498210907,0.2286448777,-0.0336090177,0.3077254593,-0.5739456415,0.2216343284,-0.2341077179,-0.4970018268,-0.3280585706,0.4078508317,-0.1767874807,0.045470655,-0.4919362664,0.096931763,0.2620433867,-0.15572837,-0.2288416624,0.1897572875,-0.199320823,-0.1496073753,0.1461617053,-0.0007470816,-0.3564024866,0.2511369884,0.0528331362,-0.3649171889,0.3265804648,0.1220587417,-0.1009769812,0.1410528868,-0.052825395,0.4061150253,-0.0852793381,-0.2893257737,0.0562759936,0.1439653933,0.1041059121,0.0213185232,0.117744118,0.0637095496,0.0304473117,-0.0880029947,-0.0177032482,0.2320955545,-0.0181748345,-0.1179958805,0.2010704875,0.0862520039,0.4885615706,0.0414116085,-0.0412044749,-0.0073161419,-0.1223622859,-0.0063678101,0.1090927646,-0.459777087,0.1422394514,0.255818516,0.0264405571,-0.218346864,0.1529573351,-0.5634790659,-0.0025339429,0.3973648548,-0.2394969314,0.0816483274,0.0570386462,-0.0124045648,-0.2247115821,0.7063226104,0.1493255049,0.1009957716,-0.4904619753,0.0878694355,-0.2523314655,0.0116971452,-0.0564800501,0.3043066263,0.1508430988,0.2124760747,-0.0562986322,0.2239233553,-0.5081438422,-0.0757123902,-0.3126007318,-0.1842348874,-0.0984203741,-0.1548344344,-0.4467178583,0.2984538674,0.1048926786,-0.1687078178,0.0945810154,-0.1009206101,-0.2375288904,-0.0783402622,0.0572415777,-0.0377932414,0.3225003779,0.2368447483,0.1222198233,0.2072189301,-0.1605268866,-0.155576542,-0.1473958641,-0.0927625746,-0.2573646307,-0.211567685,-0.0741405338,0.2012782693,-0.3087124527,0.0707918331,-0.2480508387,0.0743223354,-0.0339072943,-0.2258388996,0.0630182549,-0.1184448451,-0.0050125336,0.2903841138,0.1295926273,0.3474197388,-0.1978291124,0.0190579444,-0.1749930233,-0.0600227043,0.4449537694,-0.2349969149,-0.1213464886,-0.2829075754,0.2263375819,-0.098606512,-0.1131581366,-0.5062593818,0.1941122562,0.0572322682,0.1676397622,-0.4304614663,0.4660091102,-0.1316936761,-0.114258796,-0.2120940387,0.1924820095,0.0097149191,-0.0852976665,-0.1749822497,-0.4122220576]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"Adding to the previous information, I think `torch.utils.data.DataLoader` is doing some conversion. \r\nWhat I tried:\r\n```python\r\ntrain_dataset = load_dataset('mnist')\r\n```\r\nI don't use any `map` or `set_format` or any `transform`. I use this directly, and try to load batches using the `DataLoader` with batch size 2, I get an output like this for the `image`:\r\n\r\n```\r\n[[tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor...\r\n```\r\nFor `label`, it works fine:\r\n```\r\ntensor([7, 6])\r\n```\r\nNote that I didn't specify conversion to torch tensors anywhere.\r\n\r\nBasically, there are two problems here:\r\n1. `dataset.map` doesn't return tensor type objects, even though it uses the transforms, the grayscale conversion in transform was done, but the output was lists only.\r\n2. The `DataLoader` performs its own conversion, which may be not desired.\r\n\r\nI understand that we can't change `DataLoader` because it is a torch functionality, however, is there a way we can handle image data to allow using it with torch `DataLoader` and `torchvision` properly?\r\n\r\nI think if the `image` was a torch tensor (N,H,W,C), or a list of torch tensors (H,W,C), before it is passed to `DataLoader`, then we might not face this issue. ","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":202,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\nAdding to the previous information, I think `torch.utils.data.DataLoader` is doing some conversion. \r\nWhat I tried:\r\n```python\r\ntrain_dataset = load_dataset('mnist')\r\n```\r\nI don't use any `map` or `set_format` or any `transform`. I use this directly, and try to load batches using the `DataLoader` with batch size 2, I get an output like this for the `image`:\r\n\r\n```\r\n[[tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor([0, 0]), tensor...\r\n```\r\nFor `label`, it works fine:\r\n```\r\ntensor([7, 6])\r\n```\r\nNote that I didn't specify conversion to torch tensors anywhere.\r\n\r\nBasically, there are two problems here:\r\n1. `dataset.map` doesn't return tensor type objects, even though it uses the transforms, the grayscale conversion in transform was done, but the output was lists only.\r\n2. The `DataLoader` performs its own conversion, which may be not desired.\r\n\r\nI understand that we can't change `DataLoader` because it is a torch functionality, however, is there a way we can handle image data to allow using it with torch `DataLoader` and `torchvision` properly?\r\n\r\nI think if the `image` was a torch tensor (N,H,W,C), or a list of torch tensors (H,W,C), before it is passed to `DataLoader`, then we might not face this issue. ","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"What's the feature types of your new dataset after `.map` ?\r\n\r\nCan you try with adding `features=` in the `.map` call in order to set the \"image\" feature type to `Array2D` ?\r\nThe default feature type is lists of lists, we've not implemented shape verification to use ArrayXD instead of nested lists yet","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":53,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\nWhat's the feature types of your new dataset after `.map` ?\r\n\r\nCan you try with adding `features=` in the `.map` call in order to set the \"image\" feature type to `Array2D` ?\r\nThe default feature type is lists of lists, we've not implemented shape verification to use ArrayXD instead of nested lists yet","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"Hi @lhoestq\r\n\r\nRaw feature types are like this:\r\n```\r\nImage:\r\n 60000 #(type, len)\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\nInside the `prepare_feature` method with batch size 100000 , after processing, they are like this:\r\n\r\nInside Prepare Train Features\r\n```\r\nImage:\r\n 10000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 10000\r\n\r\n```\r\n\r\nAfter map, the feature type are like this:\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\n\r\nAfter dataloader with batch size 2, the batch features are like this:\r\n```\r\nImage:\r\n 1\r\n 28\r\n 28\r\n 2\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n
\r\n\r\nWhen I was setting the format of `train_dataset` to 'torch' after mapping - \r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\n\r\nCorresponding DataLoader batch:\r\n```\r\nFrom DataLoader batch features\r\nImage:\r\n 1\r\n 28\r\n 2\r\n 28\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n\r\nI will check with features and get back.\r\n\r\n\r\n\r\n","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":213,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\nHi @lhoestq\r\n\r\nRaw feature types are like this:\r\n```\r\nImage:\r\n 60000 #(type, len)\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\nInside the `prepare_feature` method with batch size 100000 , after processing, they are like this:\r\n\r\nInside Prepare Train Features\r\n```\r\nImage:\r\n 10000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 10000\r\n\r\n```\r\n\r\nAfter map, the feature type are like this:\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\n\r\nAfter dataloader with batch size 2, the batch features are like this:\r\n```\r\nImage:\r\n 1\r\n 28\r\n 28\r\n 2\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n
\r\n\r\nWhen I was setting the format of `train_dataset` to 'torch' after mapping - \r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\n\r\nCorresponding DataLoader batch:\r\n```\r\nFrom DataLoader batch features\r\nImage:\r\n 1\r\n 28\r\n 2\r\n 28\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n\r\nI will check with features and get back.\r\n\r\n\r\n\r\n","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"Hi @lhoestq\r\n\r\n# Using Array3D\r\nI tried this:\r\n```python\r\nfeatures = datasets.Features({\r\n \"image\": datasets.Array3D(shape=(1,28,28),dtype=\"float32\"),\r\n \"label\": datasets.features.ClassLabel(names=[\"0\", \"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\"]),\r\n })\r\ntrain_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n```\r\nand it didn't fix the issue.\r\n\r\nDuring the `prepare_train_features:\r\n```\r\nImage:\r\n 10000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 10000\r\n\r\n```\r\n\r\nAfter the `map`:\r\n\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\nFrom the DataLoader batch:\r\n```\r\nImage:\r\n 1\r\n 28\r\n 28\r\n 2\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\nIt is the same as before.\r\n\r\n---\r\n\r\nUsing `datasets.Sequence(datasets.Array2D(shape=(28,28),dtype=\"float32\"))` gave an error during `map`:\r\n\r\n```python\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n 3 \"label\": datasets.features.ClassLabel(names=[\"0\", \"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\"]),\r\n 4 })\r\n----> 5 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n\r\n15 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 446 num_proc=num_proc,\r\n 447 )\r\n--> 448 for k, dataset in self.items()\r\n 449 }\r\n 450 )\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/dataset_dict.py in (.0)\r\n 446 num_proc=num_proc,\r\n 447 )\r\n--> 448 for k, dataset in self.items()\r\n 449 }\r\n 450 )\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1307 fn_kwargs=fn_kwargs,\r\n 1308 new_fingerprint=new_fingerprint,\r\n-> 1309 update_data=update_data,\r\n 1310 )\r\n 1311 else:\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 202 }\r\n 203 # apply actual function\r\n--> 204 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 205 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 206 # re-apply format to the output\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 335 # Call actual function\r\n 336 \r\n--> 337 out = func(self, *args, **kwargs)\r\n 338 \r\n 339 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1580 if update_data:\r\n 1581 batch = cast_to_python_objects(batch)\r\n-> 1582 writer.write_batch(batch)\r\n 1583 if update_data:\r\n 1584 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 274 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 275 typed_sequence_examples[col] = typed_sequence\r\n--> 276 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 277 self.write_table(pa_table, writer_batch_size)\r\n 278 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._handle_arrow_array_protocol()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in __arrow_array__(self, type)\r\n 95 out = pa.ExtensionArray.from_storage(type, pa.array(self.data, type.storage_dtype))\r\n 96 else:\r\n---> 97 out = pa.array(self.data, type=type)\r\n 98 if trying_type and out[0].as_py() != self.data[0]:\r\n 99 raise TypeError(\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: extension\r\n```","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":447,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\nHi @lhoestq\r\n\r\n# Using Array3D\r\nI tried this:\r\n```python\r\nfeatures = datasets.Features({\r\n \"image\": datasets.Array3D(shape=(1,28,28),dtype=\"float32\"),\r\n \"label\": datasets.features.ClassLabel(names=[\"0\", \"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\"]),\r\n })\r\ntrain_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n```\r\nand it didn't fix the issue.\r\n\r\nDuring the `prepare_train_features:\r\n```\r\nImage:\r\n 10000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 10000\r\n\r\n```\r\n\r\nAfter the `map`:\r\n\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\nFrom the DataLoader batch:\r\n```\r\nImage:\r\n 1\r\n 28\r\n 28\r\n 2\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\nIt is the same as before.\r\n\r\n---\r\n\r\nUsing `datasets.Sequence(datasets.Array2D(shape=(28,28),dtype=\"float32\"))` gave an error during `map`:\r\n\r\n```python\r\nArrowNotImplementedError Traceback (most recent call last)\r\n in ()\r\n 3 \"label\": datasets.features.ClassLabel(names=[\"0\", \"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\"]),\r\n 4 })\r\n----> 5 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)\r\n\r\n15 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 446 num_proc=num_proc,\r\n 447 )\r\n--> 448 for k, dataset in self.items()\r\n 449 }\r\n 450 )\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/dataset_dict.py in (.0)\r\n 446 num_proc=num_proc,\r\n 447 )\r\n--> 448 for k, dataset in self.items()\r\n 449 }\r\n 450 )\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1307 fn_kwargs=fn_kwargs,\r\n 1308 new_fingerprint=new_fingerprint,\r\n-> 1309 update_data=update_data,\r\n 1310 )\r\n 1311 else:\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 202 }\r\n 203 # apply actual function\r\n--> 204 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 205 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 206 # re-apply format to the output\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 335 # Call actual function\r\n 336 \r\n--> 337 out = func(self, *args, **kwargs)\r\n 338 \r\n 339 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1580 if update_data:\r\n 1581 batch = cast_to_python_objects(batch)\r\n-> 1582 writer.write_batch(batch)\r\n 1583 if update_data:\r\n 1584 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 274 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 275 typed_sequence_examples[col] = typed_sequence\r\n--> 276 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 277 self.write_table(pa_table, writer_batch_size)\r\n 278 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._handle_arrow_array_protocol()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/datasets\/arrow_writer.py in __arrow_array__(self, type)\r\n 95 out = pa.ExtensionArray.from_storage(type, pa.array(self.data, type.storage_dtype))\r\n 96 else:\r\n---> 97 out = pa.array(self.data, type=type)\r\n 98 if trying_type and out[0].as_py() != self.data[0]:\r\n 99 raise TypeError(\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowNotImplementedError: extension\r\n```","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"# Convert raw tensors to torch format\r\nStrangely, converting to torch tensors works perfectly on `raw_dataset`:\r\n```python\r\nraw_dataset.set_format('torch',columns=['image','label'])\r\n```\r\nTypes:\r\n```\r\nImage:\r\n 60000\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\n\r\nUsing this for transforms:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n examples[\"image\"][example_idx].numpy()\r\n ))\r\n else:\r\n images.append(examples[\"image\"][example_idx].numpy())\r\n labels.append(examples[\"label\"][example_idx])\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n```\r\n\r\nInside `prepare_train_features`:\r\n```\r\nImage:\r\n 10000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 10000\r\n\r\n```\r\n\r\nAfter `map`:\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\nDataLoader batch:\r\n\r\n```\r\nImage:\r\n 1\r\n 28\r\n 2\r\n 28\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n\r\n---\r\n\r\n## Using `torch` format:\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\nDataLoader batches:\r\n\r\n```\r\nImage:\r\n 1\r\n 28\r\n 2\r\n 28\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n\r\n---\r\n## Using the features - `Array3D`:\r\n\r\n```\r\nImage:\r\n 10000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 10000\r\n\r\n```\r\n\r\nAfter `map`:\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\n\r\nAfter DataLoader `batch`:\r\n```\r\nImage:\r\n 2\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n\r\nThe last one works perfectly.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29076344\/110491452-4cf09c00-8117-11eb-8a47-73bf3fc0c3dc.png)\r\n\r\nI wonder why this worked, and others didn't.\r\n\r\n\r\n\r\n\r\n\r\n\r\n","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":299,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\n# Convert raw tensors to torch format\r\nStrangely, converting to torch tensors works perfectly on `raw_dataset`:\r\n```python\r\nraw_dataset.set_format('torch',columns=['image','label'])\r\n```\r\nTypes:\r\n```\r\nImage:\r\n 60000\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\n\r\nUsing this for transforms:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n examples[\"image\"][example_idx].numpy()\r\n ))\r\n else:\r\n images.append(examples[\"image\"][example_idx].numpy())\r\n labels.append(examples[\"label\"][example_idx])\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n```\r\n\r\nInside `prepare_train_features`:\r\n```\r\nImage:\r\n 10000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 10000\r\n\r\n```\r\n\r\nAfter `map`:\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\nDataLoader batch:\r\n\r\n```\r\nImage:\r\n 1\r\n 28\r\n 2\r\n 28\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n\r\n---\r\n\r\n## Using `torch` format:\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\nDataLoader batches:\r\n\r\n```\r\nImage:\r\n 1\r\n 28\r\n 2\r\n 28\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n\r\n---\r\n## Using the features - `Array3D`:\r\n\r\n```\r\nImage:\r\n 10000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 10000\r\n\r\n```\r\n\r\nAfter `map`:\r\n```\r\nImage:\r\n 60000\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 60000\r\n\r\n```\r\n\r\nAfter DataLoader `batch`:\r\n```\r\nImage:\r\n 2\r\n 1\r\n 28\r\n 28\r\n\r\nLabel:\r\n 2\r\n\r\n```\r\n\r\nThe last one works perfectly.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29076344\/110491452-4cf09c00-8117-11eb-8a47-73bf3fc0c3dc.png)\r\n\r\nI wonder why this worked, and others didn't.\r\n\r\n\r\n\r\n\r\n\r\n\r\n","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"Concluding, the way it works right now is:\r\n\r\n1. Converting raw dataset to `torch` format.\r\n2. Use the transform and apply using `map`, ensure the returned values are tensors. \r\n3. When mapping, use `features` with `image` being `Array3D` type.","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":39,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\nConcluding, the way it works right now is:\r\n\r\n1. Converting raw dataset to `torch` format.\r\n2. Use the transform and apply using `map`, ensure the returned values are tensors. \r\n3. When mapping, use `features` with `image` being `Array3D` type.","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"What the dataset returns depends on the feature type.\r\nFor a feature type that is Sequence(Sequence(Sequence(Value(\"uint8\")))), a dataset formatted as \"torch\" return lists of lists of tensors. This is because the lists lengths may vary.\r\nFor a feature type that is Array3D on the other hand it returns one tensor. This is because the size of the tensor is fixed and defined bu the Array3D type.","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":66,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\nWhat the dataset returns depends on the feature type.\r\nFor a feature type that is Sequence(Sequence(Sequence(Value(\"uint8\")))), a dataset formatted as \"torch\" return lists of lists of tensors. This is because the lists lengths may vary.\r\nFor a feature type that is Array3D on the other hand it returns one tensor. This is because the size of the tensor is fixed and defined bu the Array3D type.","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"Okay, that makes sense.\r\nRaw images are list of Array2D, hence we get a single tensor when `set_format` is used. But, why should I need to convert the raw images to `torch` format when `map` does this internally?\r\n\r\nUsing `Array3D` did not work with `map` when raw images weren't `set_format`ted to torch type.","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":53,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\nOkay, that makes sense.\r\nRaw images are list of Array2D, hence we get a single tensor when `set_format` is used. But, why should I need to convert the raw images to `torch` format when `map` does this internally?\r\n\r\nUsing `Array3D` did not work with `map` when raw images weren't `set_format`ted to torch type.","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2005","title":"Setting to torch format not working with torchvision and MNIST","comments":"I understand that `map` needs to know what kind of output tensors are expected, and thus converting the raw dataset to `torch` format is necessary. Closing the issue since it is resolved.","body":"Hi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.","comment_length":32,"text":"Setting to torch format not working with torchvision and MNIST\nHi\r\n\r\nI am trying to use `torchvision.transforms` to handle the transformation of the image data in the `mnist` dataset. Assume I have a `transform` variable which contains the `torchvision.transforms` object.\r\n\r\nA snippet of what I am trying to do:\r\n```python\r\ndef prepare_features(examples):\r\n images = []\r\n labels = []\r\n for example_idx, example in enumerate(examples[\"image\"]):\r\n if transform is not None:\r\n images.append(transform(\r\n np.array(examples[\"image\"][example_idx], dtype=np.uint8)\r\n ))\r\n else:\r\n images.append(torch.tensor(np.array(examples[\"image\"][example_idx], dtype=np.uint8)))\r\n labels.append(torch.tensor(examples[\"label\"][example_idx]))\r\n output = {\"label\":labels, \"image\":images}\r\n return output\r\n\r\nraw_dataset = load_dataset('mnist')\r\ntrain_dataset = raw_dataset.map(prepare_features, batched=True, batch_size=10000)\r\ntrain_dataset.set_format(\"torch\",columns=[\"image\",\"label\"])\r\n```\r\n\r\nAfter this, I check the type of the following:\r\n```python\r\nprint(type(train_dataset[\"train\"][\"label\"]))\r\nprint(type(train_dataset[\"train\"][\"image\"][0]))\r\n```\r\nThis leads to the following output:\r\n\r\n```python\r\n\r\n\r\n```\r\nI use `torch.utils.DataLoader` for batches, the type of `batch[\"train\"][\"image\"]` is also ``.\r\n\r\nI don't understand why only the `label` is converted to a torch tensor, why does the image not get converted? How can I fix this issue?\r\n\r\nThanks,\r\nGunjan\r\n\r\nEDIT:\r\nI just checked the shapes, and the types, `batch[image]` is a actually a list of list of tensors. Shape is (1,28,2,28), where `batch_size` is 2. I don't understand why this is happening. Ideally it should be a tensor of shape (2,1,28,28).\r\n\r\nEDIT 2:\r\nInside `prepare_train_features`, the shape of `images[0]` is `torch.Size([1,28,28])`, the conversion is working. However, the output of the `map` is a list of list of list of list.\nI understand that `map` needs to know what kind of output tensors are expected, and thus converting the raw dataset to `torch` format is necessary. Closing the issue since it is resolved.","embeddings":[-0.12737526,-0.3511648178,-0.017932741,0.3536089063,0.4760023057,0.0887141973,0.7330293059,0.3804347813,0.0655255616,-0.0380423479,-0.1111534312,0.3819508553,-0.2329687476,-0.3339346349,0.0351093672,-0.56149441,0.2225798965,-0.0998275802,-0.2807984054,-0.0613472909,-0.243133992,0.0354893506,-0.2184970975,-0.1084961668,-0.547935307,0.0440959707,-0.0195557699,-0.0765685439,-0.06229949,-0.1391803324,0.1223469973,-0.1744965166,0.4444530904,0.7012936473,-0.0001277011,0.0703018233,0.3642728925,-0.1737781912,0.0499992296,-0.1610135287,0.0400467664,-0.2573829293,0.0727764145,-0.0946834609,-0.2234939188,-0.1708452553,0.0959103853,-0.17902942,-0.0281423349,0.4191688895,0.077129975,0.2790103853,0.2466612458,0.1386305541,0.2278117537,0.5182864666,-0.2734378576,-0.0561165325,0.2261500657,0.37547791,0.1240043715,0.5647351146,-0.2913345993,-0.0362333432,0.266944766,0.1290054172,-0.1014554724,-0.3140054941,0.0491575636,0.089462921,0.4219441712,-0.1494393945,-0.0807789043,-0.1436640322,-0.143978551,-0.1273805052,0.0405908599,0.2803191841,-0.0718406066,-0.1872386485,-0.5392424464,0.3052915633,-0.1486232579,0.2131595016,-0.2805242538,-0.0034206891,-0.1381631345,0.2928087115,-0.0471190289,-0.112712875,0.1930810809,-0.1234148592,0.1300757527,-0.0098258834,-0.1436316669,-0.0534907505,-0.3895993233,-0.3150125444,-0.2001479119,-0.1602650434,0.1664855033,0.1392757297,-0.1163957268,0.1280712485,0.1826325953,0.2680338025,0.105782643,0.3207025826,-0.0858922377,-0.2895810604,-0.0469189361,0.0877334401,-0.2962270379,-0.435752362,0.1545339972,0.3252966106,0.2452212572,0.0687427595,0.1298554242,-0.3239412606,-0.1089225262,-0.0095866034,0.075225018,0.285038203,-0.1186793,0.24390468,0.4804830253,0.3972030282,-0.1538181752,0.0695624724,0.0250724964,0.0315209553,-0.440495193,-0.2856986523,0.1534098238,-0.1177733615,0.100070186,0.0374316685,0.0464155935,0.1006969139,0.0734966248,-0.1858698428,0.6246291399,0.147647351,-0.1944740117,0.2767927051,0.2709063888,0.5297806263,-0.3042984307,0.2565316558,-0.5095701218,0.0652529001,-0.0230914187,-0.0145437596,0.2186551988,-0.020721972,-0.2121624202,0.0158288926,0.5923122168,-0.0836780667,0.2734875381,-0.57916224,0.0949639827,-0.1913348734,0.034152355,0.0636695474,-0.2047151774,-0.0523584187,0.3863441348,0.0919497311,0.3154585958,0.1602817774,0.0510833338,0.0798699334,-0.2253530473,-0.2350505739,0.3363234699,-0.3768565059,0.0920383558,-0.0176931173,0.0938790813,-0.0638038367,0.1120190248,0.3361779451,0.0326004587,-0.128001526,0.0294485986,0.1963148117,-0.1661822349,0.1402871758,0.0773598477,0.1913028061,0.1933745444,0.0350063629,-0.0156836379,0.2076539844,-0.2216719985,0.1775767207,-0.0542802997,-0.0747133866,0.0711769462,0.0403783172,-0.3698907197,-0.0121967075,-0.0474696793,0.1864667684,-0.23310031,-0.048721496,0.0236351732,0.0822172239,0.0582793653,0.047925096,0.0049781557,-0.1919564456,-0.1822061539,-0.1390272975,-0.0032419825,0.0922108144,-0.2094767392,-0.2617242932,-0.0523435548,-0.1232490763,0.1783037037,0.0588379279,-0.3590371013,0.0510254167,0.019273432,-0.3653303087,-0.2013600469,0.1733924001,0.2676690519,-0.0996206775,-0.0422345139,0.1916035116,0.0632886291,0.1146331578,-0.5482739806,0.0834660977,0.1778419167,-0.2469699979,0.2043911219,0.2162791938,0.1485340595,-0.2755032778,-0.1960345507,0.4387866855,0.1604105234,0.315912962,-0.3300254345,0.116101861,-0.0726469457,-0.0394018255,-0.1906339079,-0.0408082679,-0.277281642,0.0730238482,-0.1318593472,0.0129822334,-0.4569184184,-0.0449360423,0.2962621152,0.0887641758,0.1758723557,-0.0012164807,-0.399846226,0.2133122981,0.1771268845,-0.6013535261,0.2978383601,-0.0043616337,0.1892076284,-0.1050189212,0.0955833048,-0.1625430584,0.2484440655,0.1140429005,0.090748094,-0.1994652301,0.1418524235,-0.1079259515,0.0268462822,0.0263701156,-0.1406289488,-0.2336931378,-0.3589504957,0.2713825405,-0.1661832929,-0.1345185637,-0.4018543959,-0.2621856332,0.0360565595,-0.0740050003,-0.0823338255,0.1507581174,0.1855677366,0.190434888,0.26341784,-0.1818942726,0.3743162453,-0.4130297601,-0.0353545919,-0.0750095248,-0.2867960632,-0.0991648287,0.0896557719,-0.5234113932,0.0914847627,-0.2289899141,-0.0931907445,-0.2501759529,0.0135480622,0.1418226808,-0.2678500712,-0.2850030065,0.3085908592,0.1939281374,-0.142482236,0.040662054,0.2663511038,0.0103072627,0.0469011813,0.1537833363,-0.0896200389,0.1165602952,-0.063298367,-0.1767405272,-0.0966183841,-0.0750613436,-0.0704168677,-0.0217595734,0.0633770004,0.3462581038,0.3721836209,-0.0898904651,-0.052042447,0.0356094316,0.06296473,-0.3976880014,0.5225032568,-0.2261542529,-0.3608259261,0.1192016751,0.0238555167,0.0651473776,0.4862691164,0.0605229028,0.0497316308,0.0117824115,-0.0789448172,0.2525928319,0.2211969346,0.3510203063,-0.091124773,0.0191658437,-0.1647072583,0.0154163484,-0.1018712595,-0.0632551014,0.608886838,0.185083881,0.2399424464,0.0992583558,0.6996648312,-0.0611373,-0.7380955219,0.2168067545,-0.2425099015,0.1921664476,-0.1181217954,-0.3539586067,0.3988561332,-0.2094196975,0.0643741786,-0.0859671235,-0.0249070711,-0.239157781,-0.0902243406,0.3573613763,-0.3172071576,-0.1585742086,0.4044301808,-0.1447026283,0.2763725817,-0.231342569,0.3245894909,-0.2609263659,-0.2137927711,0.0600571744,0.448771745,0.1385134459,0.0417248718,-0.3342839479,-0.2166968733,-0.1432266682,0.3568273485,0.1195773929,0.5672699809,-0.0980361179,-0.0744652227,0.101556398,0.2413696051,0.8307591081,0.0455576442,-0.0904590189,0.1514420211,-0.3478348553,-0.4549778998,-0.1022292972,-0.0514178276,0.0906811357,0.1716903597,0.1203863919,-0.0910534263,-0.1833337396,-0.0058324421,0.1955546588,-0.0551248267,0.0347188525,-0.2269114107,0.0587578602,-0.1296497434,-0.1552830487,0.2211823612,0.0296752974,0.0051616579,-0.3081279397,-0.326230526,-0.1266999394,0.1824821085,-0.133012116,0.4110670984,0.0992513001,-0.1780857146,0.0769067258,0.5561061502,0.4184045792,-0.0389390662,-0.1187736243,0.0806439966,0.2931110859,0.1064325869,0.3002041876,0.3208246231,-0.0618727207,-0.439663738,-0.2268486619,-0.1938151866,-0.3305326402,0.3972934484,0.4080631137,0.0989354551,-0.4953551888,-0.4921707213,0.386190027,0.2937195897,0.1958299726,0.2045914978,-0.5417591333,-0.400967598,0.0781787708,0.4662419856,0.8556224108,-0.0370788649,0.370282948,0.0315654092,0.0108070988,0.1721023768,-0.0018885463,0.4117940664,-0.3542813659,0.1392895579,-0.1648536623,-0.3125498295,0.2525228858,-0.0602033772,-0.2439697087,0.0425365344,-0.3494711518,0.193619892,-0.1256936789,-0.0828530416,-0.0498262756,-0.4714181721,0.1878739595,0.0154518224,-0.0825132653,0.2617392242,0.1111612171,-0.1636207253,-0.3798606396,0.061696101,-0.0762220845,-0.0150116216,-0.245968923,-0.163243264,-0.0243103318,-0.7298656106,0.058679942,0.4961579442,0.3345543146,-0.1605663151,-0.0478364117,0.1396413296,0.4438471198,0.1629004478,-0.0990170017,-0.3401841521,0.285340637,0.0973757356,0.1843322515,0.1086381972,0.0681486949,0.2105204314,-0.5777894855,0.0372510701,0.3060289919,-0.3449506462,-0.3229849041,-0.1984501928,0.0491848439,0.0381636098,0.0286185313,0.1244887859,-0.0389032289,0.207243219,-0.0415121652,-0.2264615446,0.1328899264,0.4523311853,0.1129837558,-0.3500465751,0.3907567859,0.3549585044,-0.0447776243,-0.0437438712,0.4166939259,0.6353192329,-0.3456966877,0.0994974226,-0.1027163863,-0.1569872051,-0.0110736163,0.5430936813,-0.2343171388,-0.1309780926,-0.2389575988,-0.0408169143,-0.4308940172,0.010813714,-0.3162984848,0.4566304684,-0.0432225056,0.3627102077,0.2105059177,-0.1873495281,-0.1441359371,-0.2586488426,0.1723507196,0.3679333925,-0.2471826226,0.2187610567,-0.1269586384,-0.1220227554,-0.0362524614,0.026031455,-0.2497396916,-0.0333908275,-0.0576778054,0.1773060709,0.0666141063,-0.1135403067,-0.0735471323,-0.1736694723,-0.1385808885,-0.2681717873,-0.0377181508,0.1109209061,-0.1159995794,0.353161186,-0.1783660948,0.2312256545,0.3186767399,-0.2700004578,-0.1026227921,0.0358714946,0.2229631692,0.0600506775,0.0045318487,-0.0615568347,-0.2772363126,0.1125913784,0.1652033031,0.2549748421,0.2046177238,-0.099996686,0.1091783717,-0.0438067093,0.2041839957,0.1012553498,-0.1718702465,-0.0508288406,0.06037293,-0.0076234695,0.100621663,0.1449992061,0.0388693474,0.125281468,-0.0834642798,0.2755780816,0.0665590689,0.3113946617,-0.3705079556,0.0451420546,0.4178563058,-0.2410733551,0.2350863814,0.4640146792,0.0147997756,0.1930498034,0.2907047868,0.1093436927,-0.1187017336,0.3226965666,0.3945164979,0.8204309344,0.5895952582,0.0837697685,0.2746028006,0.0992571115,-0.1536079347,-0.2072512656,0.2358240783,0.3919684589,0.0188666359,0.1856933236,-0.2407987714,-0.4871240854,0.2258820236,0.121774748,-0.1946179271,-0.0822237059,-0.6648061275,0.0118753593,0.1358264983,-0.1866598725,-0.1797155589,-0.0501378328,0.2205909193,0.259753406,-0.1032203063,-0.1609126478,-0.1724745929,0.069101274,0.037528865,-0.0073296069,0.1697574705,-0.1368460953,0.1617610157,0.2448244989,0.1969081163,0.2849530876,0.4870797694,-0.1681713909,-0.15989393,-0.0105532985,-0.1923659146,0.0899014473,0.0667119697,0.0315340199,-0.1532076448,-0.0149565199,0.0084626377,0.0386277772,0.0314607546,-0.0940809324,-0.2182593197,0.1676253527,0.3055747151,-0.3637773991,-0.1490929127,-0.0998983607,-0.2466557473,-0.3483100235,-0.2655349374,0.5327388048,0.0278957784,0.1280097067,0.1421678662,-0.0421595052,0.0487859771,0.1574120075,-0.1215568632,0.0775766745,-0.1540935338,0.1359660625,-0.0858048126,0.0319104679,0.1184748635,-0.3834480345,-0.0188495293,-0.0391642191,0.2425420731,0.0646747798,0.2020902336,-0.0983475819,0.498744905,0.6146565676,-0.1336159557,-0.141480729,-0.1304193288,0.3706511557,0.0733084828,-0.0774655864,0.0219212025,0.1024273708,0.006506992,-0.3420650363,-0.2554905117,-0.0199822076,0.3041453063,0.3512827158,0.0932885408,0.3203133345,-0.0192205366,0.0484099202,-0.2411338687,0.0504152812,-0.1276013106,0.0096704597,-0.0228028372,0.5810167193,-0.2244020551,-0.0593425669,-0.413649857,0.4290263057,0.0905310959,-0.3388551772,-0.0805196688,0.3584000766,-0.3173207343,-0.2010882497,0.2261953056,0.4100218415,0.1513600647,0.194357127,-0.5254404545,-0.406911701,0.4472021759,-0.4417076409,-0.5069124103,0.0290745515,-0.3133939803,0.0679038763,0.4964230657,-0.4710501134,-0.1491340697,0.0015054079,-0.0808276311,-0.0916040763,-0.09819372,0.2610400021,0.0151107572,-0.2518461943,0.1790075153,0.176450491,-0.0228161905,-0.4204162061,-0.2753599882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2003","title":"Messages are being printed to the `stdout`","comments":"This is expected to show this message to the user via stdout.\r\nThis way the users see it directly and can cancel the downloading if they want to.\r\nCould you elaborate why it would be better to have it in stderr instead of stdout ?","body":"In this code segment, we can see some messages are being printed to the `stdout`.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/7e60bb509b595e8edc60a87f32b2bacfc065d607\/src\/datasets\/builder.py#L545-L554\r\nAccording to the comment, it is done intentionally, but I don't really understand why don't we log it with a higher level or print it directly to the `stderr`.\r\nIn my opinion, this kind of messages should never printed to the stdout. At least some configuration\/flag should make it possible to provide in order to explicitly prevent the package to contaminate the stdout.\r\n","comment_length":45,"text":"Messages are being printed to the `stdout`\nIn this code segment, we can see some messages are being printed to the `stdout`.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/7e60bb509b595e8edc60a87f32b2bacfc065d607\/src\/datasets\/builder.py#L545-L554\r\nAccording to the comment, it is done intentionally, but I don't really understand why don't we log it with a higher level or print it directly to the `stderr`.\r\nIn my opinion, this kind of messages should never printed to the stdout. At least some configuration\/flag should make it possible to provide in order to explicitly prevent the package to contaminate the stdout.\r\n\nThis is expected to show this message to the user via stdout.\r\nThis way the users see it directly and can cancel the downloading if they want to.\r\nCould you elaborate why it would be better to have it in stderr instead of stdout ?","embeddings":[-0.0470962487,-0.370927155,-0.0396360606,0.2706121504,0.2318519801,-0.0572964177,0.2466085553,0.1149493307,-0.0417824052,0.198417902,0.1732622683,0.1642702371,-0.1199179217,0.3688814044,0.1807003617,0.1382270604,-0.0852121785,-0.1296653897,-0.3972431123,0.2983495295,0.1317119002,0.1694648117,0.3729337454,0.4860750437,-0.5674242973,-0.0275846142,0.2408996224,-0.0820059404,-0.2319761813,-0.5787138939,0.0741454437,0.1343484968,0.1837686896,0.1799414456,-0.00011806,0.1449533105,0.4684007764,0.2060822248,-0.681963861,-0.0856236443,-0.1658631861,-0.127990365,0.3208130002,-0.0690777823,0.1317871213,-0.2993815243,0.2840439379,-0.3241891563,0.1400134116,0.4661122262,0.1768932641,0.2232476175,0.0355274938,0.2963497043,0.2246194631,0.3057037592,-0.0642746761,0.0225167722,0.2490636408,0.5164734721,-0.0466824397,0.3044881225,-0.109712705,-0.3461034298,0.1512940526,0.1833879948,0.1994549781,-0.2152294368,0.1340060383,0.0134589383,0.5597186685,-0.4376442432,-0.1601393819,-0.2399809808,-0.1339333504,-0.1432037652,0.2294410169,0.1206058264,-0.2808728516,0.4999408126,-0.3562501669,-0.3182742894,-0.2685394585,-0.0708177909,-0.0102504212,-0.1528872102,-0.2506071627,0.1657451391,-0.182757169,0.2157797515,-0.083234787,-0.2934994996,-0.0698479041,-0.1787753701,0.0691016242,-0.0906247348,-0.0050333366,-0.1460138559,-0.0950104147,-0.0288153011,-0.0023840182,-0.0100431712,0.345340997,0.0766516551,0.4637843072,0.0726141632,0.3933348358,0.1709566265,-0.0247005634,0.1267792135,0.3316883147,-0.3063760996,0.3164304495,0.2413713038,0.1670803279,-0.3005855083,0.1296257228,-0.3068225086,-0.0729350522,0.0900440514,-0.0182394534,0.0932418555,0.0036982386,0.1688583642,0.1365753859,-0.0851947293,0.1672441512,0.0279339328,0.2186060101,0.0639156252,-0.1007551327,0.0687317252,-0.1920704246,-0.0202607978,-0.0244908929,-0.1460309923,0.1155524328,0.0659441426,0.2069480121,-0.1051694676,-0.1531255692,-0.0523859374,0.1600597054,0.5171230435,-0.4261567891,0.2639125586,0.0571803041,-0.079849191,-0.1211639717,0.3827722967,-0.0130477687,-0.4193401337,0.1440353096,0.0797205046,-0.1925650686,0.1210361272,-0.3839595616,-0.1830466986,0.022852717,-0.1718223542,0.2358080745,-0.2428783923,0.0101258429,-0.2272456288,-0.0457868874,0.1608248949,-0.1947693527,-0.1069454178,-0.148947522,-0.4786727428,0.7286791205,-0.128542766,0.0494991876,0.4333671033,-0.2160493433,-0.3962885141,0.4247333407,-0.2746763825,-0.260584563,0.3615231216,-0.4956054986,0.463319391,0.3245587051,0.1980674714,0.0893636122,-0.0297740195,0.1201320961,-0.3002189398,0.2200370282,0.2071027458,-0.2394646555,-0.1902949363,-0.0840352252,0.1256890297,-0.0978205726,0.054635331,-0.0002775001,0.0439175032,0.3048693538,0.1535117924,0.0284015313,-0.2191172391,0.1361135542,0.2374198586,-0.192817077,-0.0950850919,-0.1431473047,-0.0327363759,-0.2658724189,-0.0368485227,-0.4485417306,-0.2092805803,0.0502898023,0.2111274004,0.0256071035,-0.2873522937,0.0026614587,0.2300602496,0.1467435807,0.2893278599,-0.2135193646,0.4315739274,-0.492436558,0.1644489914,0.1382875293,-0.1706264168,0.1656695753,-0.0245365147,-0.0424523912,0.0380517244,-0.2710111141,0.1686792672,-0.036004521,0.288677901,-0.0269568618,0.1526560336,-0.242581144,0.227211833,0.0727074146,0.0035592469,-0.0096563092,0.1620015055,-0.0982812047,0.1219337583,0.0100621255,-0.0682206005,0.0612161122,0.1239883006,-0.0029775125,-0.0025338749,-0.2458423674,0.3253138959,-0.2929027081,-0.2020265907,-0.0993836299,-0.2407432497,0.1680091172,0.2397518605,-0.2445978373,0.1981148422,0.5945554972,-0.1142946035,-0.0673989207,0.3340904415,0.1119411588,-0.0784529895,0.1370971352,0.0903344899,0.1073063463,0.164195478,0.1526741385,0.01734473,0.1243095994,-0.006975221,0.2460539937,0.1179323941,-0.3268172443,0.0067924424,-0.2291564792,-0.1677503735,-0.1305981725,0.1232898757,-0.4932422042,-0.0216217544,-0.3262925744,0.031834498,-0.2363896817,-0.3483000994,-0.6054049134,-0.0629226342,-0.1207658947,-0.3933293223,0.2839479446,0.0013739818,-0.2738998532,-0.1414479166,0.1626858115,0.3985839486,-0.0244086236,0.181313917,-0.5563161373,0.3898977339,-0.3940701783,0.0558720566,-0.1074178442,0.1637454778,0.3210862875,-0.3589545488,-0.068653442,-0.2984561026,-0.0329050831,0.1503029913,-0.1243849397,0.3846013844,0.183629185,0.1446369737,0.1930282414,-0.1698112488,-0.1140002385,-0.2072126269,-0.0516148359,-0.2044169456,0.2338192165,0.1644957513,-0.344499737,-0.3177576363,-0.2811035514,-0.2852053046,0.13959755,-0.2266306877,0.3257314861,-0.3604027033,-0.0258162003,0.2454823703,-0.3515597284,0.5465459228,0.1042624339,-0.7475808859,-0.2174991369,0.0302954167,-0.05975933,-0.2551209033,0.4536778033,-0.2607750595,-0.076736778,-0.7462233305,0.0660572574,-0.2379437834,-0.1159117073,0.3059663475,-0.0538375676,0.1463974565,0.1931360811,-0.0612522289,0.053965304,0.1244837642,0.0279822815,-0.4099282324,0.3065110445,0.2965654433,0.1893914342,0.0822210014,0.2465160638,0.2810038328,0.0852204114,-0.2378197163,-0.086736016,0.7142142057,-0.3417423666,-0.1789518893,0.1240456253,0.0556167252,-0.3776142299,0.1220375001,-0.1001810208,0.3989617527,0.1636797935,0.1444735676,0.049114015,-0.3692813218,0.0547103845,0.0996540114,-0.0200872291,-0.0651734322,0.0506227463,0.2321880162,0.0743867755,0.0726089776,0.6089343429,0.2598015666,0.3592443168,-0.3870895803,-0.3552199602,-0.5714660883,-0.1087388098,0.1692549139,0.0436364971,-0.4236883819,-0.0780555606,0.3002098501,0.0457754396,0.4229607582,0.1064226478,-0.0676104873,-0.2087324113,0.0002408027,0.1899527311,0.0841218978,-0.2580262125,0.2130460441,0.4400639832,-0.2385075688,-0.342969656,-0.2365400493,0.3087860346,-0.2244276553,-0.0891733989,-0.2934219241,-0.3690193594,-0.0788289532,-0.2005903274,-0.0087905787,0.1715461314,0.1133402586,-0.5838228464,-0.3357173204,0.3812627494,0.3573177457,0.0346925929,-0.2963628471,0.077324219,-0.2547169924,-0.2744425237,0.1503992081,0.6171966195,0.4491704106,0.2272644043,0.3225549161,-0.0342020877,0.0128655788,-0.2113777101,0.2160160094,0.2802975476,-0.2634666562,0.0105418507,-0.0466082208,0.2632806301,-0.1785398424,0.1241727918,0.4459402263,0.1868680865,-0.2318391204,-0.3095212877,0.0230621938,0.0784425363,0.0268973261,0.3604999781,0.1483644545,-0.1881163865,-0.0462070741,0.0034316389,1.1015583277,-0.1373364925,0.2609904408,0.2877703011,-0.3302653432,0.3510971665,0.1281828433,0.0136729628,-0.131152764,-0.0354512595,0.0643991083,-0.2890268564,0.1884390712,-0.0490271859,-0.1270768493,-0.0428867638,0.021585457,-0.0415108204,0.0800622404,0.1023904905,-0.2160449475,0.1060844958,-0.2881341577,0.0192071535,-0.0265412796,-0.0139209535,-0.1174461916,0.0361354537,0.1522872448,-0.3208289444,-0.1644942164,-0.3527446687,0.5632613897,-0.110978961,0.026469551,0.1226056069,-0.0527873859,0.3851517737,0.0218639486,0.2955076993,0.0929023474,-0.4084978402,-0.0887902305,0.1000248194,0.1308554858,0.0152938599,0.1741431803,-0.0800395757,-0.3821748495,0.1858168691,-0.0296149626,-0.3801407516,0.2591865063,-0.2220320553,0.1948300451,-0.3954454064,0.0477756821,0.0720040649,-0.1736709774,0.0508769788,0.0975689441,0.3021850884,-0.2064762264,0.2860212922,-0.3368216753,-0.6666278243,-0.2260286957,0.179553628,0.1741956919,-0.0506715477,0.2291257977,-0.0015561674,-0.4376499057,-0.1027027145,0.1720990539,-0.004859929,-0.1171882451,0.1323738396,0.2386433631,-0.0581315979,-0.300134629,0.324600935,-0.0627133697,0.0210994706,-0.4857229292,-0.03629677,-0.4227102697,0.2273532152,-0.1513472199,0.4059913158,0.1191440374,-0.032222189,0.0361272655,-0.2973768413,-0.2570326626,-0.3470446765,-0.2968780994,-0.0136987846,0.1911920607,0.0763879344,0.3403457403,0.1696931124,0.0262363162,0.0371681303,-0.2174250633,-0.1059134603,0.1565163434,0.1749454439,-0.1161083207,0.3147472143,-0.078636311,-0.0577073917,0.1212151572,0.1780866832,0.6493924856,0.398489058,0.0585941933,0.0507427044,0.1323220283,0.1496216953,-0.0532165915,0.1916860491,0.0050909976,0.0918293372,-0.1048529595,-0.0315282419,-0.3787106276,-0.1021003947,0.4027521014,0.2205823213,0.3114818931,0.431188494,0.4334314167,-0.0558763035,0.0868912712,-0.0006720541,0.2163829058,0.1681903899,-0.0863658339,-0.3955235183,-0.1428178251,0.0818750784,0.2249925882,0.0101318769,-0.344075799,-0.2673560083,-0.2052139938,0.3394453526,-0.0800558627,-0.066098161,0.5081413388,0.0618029051,0.2550147176,-0.1932756603,-0.1969998032,0.0877113566,-0.3469744027,0.2466040254,0.3824838102,0.2974501848,-0.1563679427,-0.1022447422,-0.0636552945,0.1238490567,0.1860054731,-0.0362980776,-0.1381464005,0.0802374482,0.0349016823,0.4141359031,-0.162323162,-0.0629572868,0.1144630685,-0.2482882142,-0.0862406641,0.1089327484,0.0779775083,-0.1425808072,-0.2707242966,0.1697156727,-0.1490852684,-0.0106324079,-0.0939104185,-0.0192942005,0.030403113,0.0467401333,-0.158387512,-0.0355551466,0.0707538202,-0.2346909493,-0.2739594877,-0.2087140828,0.0998001695,-0.0873740315,-0.013681883,-0.040372245,0.4138678014,0.4520373046,0.3165356815,0.4676989615,0.245083347,0.5179719329,0.0220568962,-0.1365854442,0.0280262101,-0.2476313859,0.43323493,-0.0208045039,-0.1089089811,0.314086616,0.0920555443,-0.036075186,0.1127450764,-0.0555473827,0.2130413949,-0.3242710829,0.7055280209,0.0282981731,-0.2022131383,-0.2135059834,0.0174818486,-0.5186851621,-0.0205287989,0.1742928177,-0.1370126158,0.0333149359,-0.4020622969,0.0729152411,0.1446977407,0.4893354774,0.6453385353,0.3660604358,-0.1532873809,-0.0984083116,-0.1929564923,0.2629054487,0.1571944654,0.0401859693,-0.0300215799,-0.1838284135,-0.0041461159,0.3262839913,0.1542159766,-0.244558081,-0.5066193342,-0.0445624962,0.0106722312,0.0687692612,0.3976640999,0.2572368979,-0.1642965972,-0.0032929236,0.5116845369,-0.2611887753,-0.0154583743,0.1191211417,0.3817170858,-0.1467202455,-0.2076235116,0.0127187679,0.2066595107,-0.0640811548,0.0247612428,-0.3828255236,-0.0304998271,0.0328764766,-0.1318734586,-0.0021378917,-0.260481447,0.4260798395,-0.0194156468,-0.0970739722,-0.3136404753,0.1424210966,0.1574994922,0.1751764268,0.2302175313,-0.1871317923,-0.2503665984,-0.0390238538,0.2734516859,0.2703126967,0.3504107594,-0.0574521758,-0.1396736205,-0.1904397905,0.3595772982,-0.1550253481,-0.2110392898,-0.118463099,0.2866286337,0.1683883816,-0.1935353577,-0.5391814113,-0.0296887327,0.219997257,-0.2367369831,0.1419858634,0.0874837339,-0.3980550468,0.2825977206,-0.0191190876,0.4385917485,-0.2042137533,0.3405649662,-0.1298887581,-0.5378353]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2003","title":"Messages are being printed to the `stdout`","comments":"@lhoestq, sorry for the late reply\r\n\r\nI completely understand why you decided to output a message that is always shown. The only problem is that the message is printed to the `stdout`. For example, if the user runs `python run_glue.py > log_file`, it will redirect `stdout` to the file named `log_file`, and the message will not be shown to the user.\r\n\r\nInstead, we should print this message to `stderr`. Even in the case of `python run_glue.py > log_file` only `stdout` is being redirected and so the message is always shown.","body":"In this code segment, we can see some messages are being printed to the `stdout`.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/7e60bb509b595e8edc60a87f32b2bacfc065d607\/src\/datasets\/builder.py#L545-L554\r\nAccording to the comment, it is done intentionally, but I don't really understand why don't we log it with a higher level or print it directly to the `stderr`.\r\nIn my opinion, this kind of messages should never printed to the stdout. At least some configuration\/flag should make it possible to provide in order to explicitly prevent the package to contaminate the stdout.\r\n","comment_length":90,"text":"Messages are being printed to the `stdout`\nIn this code segment, we can see some messages are being printed to the `stdout`.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/7e60bb509b595e8edc60a87f32b2bacfc065d607\/src\/datasets\/builder.py#L545-L554\r\nAccording to the comment, it is done intentionally, but I don't really understand why don't we log it with a higher level or print it directly to the `stderr`.\r\nIn my opinion, this kind of messages should never printed to the stdout. At least some configuration\/flag should make it possible to provide in order to explicitly prevent the package to contaminate the stdout.\r\n\n@lhoestq, sorry for the late reply\r\n\r\nI completely understand why you decided to output a message that is always shown. The only problem is that the message is printed to the `stdout`. For example, if the user runs `python run_glue.py > log_file`, it will redirect `stdout` to the file named `log_file`, and the message will not be shown to the user.\r\n\r\nInstead, we should print this message to `stderr`. Even in the case of `python run_glue.py > log_file` only `stdout` is being redirected and so the message is always shown.","embeddings":[0.054125268,-0.4283034801,-0.0178409368,0.1863200217,0.1776289493,-0.1547809094,0.3838965297,0.1615999639,0.0672421008,0.2444254011,0.190782994,0.3053060472,-0.1425594836,0.2849757373,0.2610240579,0.1417885572,-0.0724066272,0.0228824783,-0.4255489409,0.107640557,-0.0089996178,0.1110258028,0.1996388882,0.4757939875,-0.6451058388,-0.0969342664,0.2199249417,0.0682550073,-0.1504664123,-0.5180291533,-0.0038762686,0.1959688216,-0.0711483508,0.2187255025,-0.0001107846,0.1114448011,0.558391273,0.1802080423,-0.5341908932,-0.1494897157,-0.0687197596,-0.2411948144,0.3232271075,-0.1385637671,-0.1318151653,-0.263232559,0.2714948058,-0.3419891894,0.3885952532,0.2979872823,0.2185657918,0.2383882999,0.0107336165,0.2430755496,0.1254059374,0.2416490316,0.0376767814,-0.0428685956,-0.0429108776,0.3080660999,-0.2036425173,0.4286541939,-0.0826231539,-0.2804137468,0.0262908377,0.1417250037,0.1627256423,-0.3487530351,0.0097641358,0.0993574187,0.2654751539,-0.4483452737,-0.1097153425,-0.3162628114,-0.0963749588,-0.1171589941,0.1848396957,0.1446038634,-0.1712843478,0.3809222281,-0.190245837,-0.2524473071,-0.2295149714,-0.073409155,-0.124781318,0.019961467,-0.320853591,0.2482610345,-0.2218214869,0.2160465717,-0.1857299507,-0.2165313214,-0.1245649159,-0.0516973436,-0.0614772849,0.0057034413,0.115155071,-0.0911643878,-0.0116026597,-0.1248358861,-0.2013799846,0.0620362461,0.2232733965,0.1918914467,0.4884245694,0.0879197717,0.5125946403,0.2814555168,0.1907999218,-0.04213164,0.0905307159,-0.2021928281,0.3088915944,0.1708641648,0.2888863087,-0.2328936458,0.2866511941,-0.0900234655,-0.0649691075,0.1431056261,-0.061183352,0.0710895211,-0.0234851856,0.2419447452,0.0153705021,-0.1141114309,0.3781521916,-0.0502083115,0.0626935288,-0.0070137731,-0.1976067871,0.0261096414,-0.3317315578,0.0355130285,-0.1024647728,-0.049990233,0.067420274,0.0976911858,0.153589204,-0.1075139344,-0.1262116134,-0.0576669611,0.4185282886,0.4810254872,-0.2644611895,0.29549402,0.1943539679,-0.2578784525,-0.1323119849,0.1952917576,0.0771409944,-0.2959942222,0.1112999618,0.1830358356,-0.3067573309,0.3234333694,-0.2131379992,0.0265495963,-0.0012505222,-0.1195104942,0.235718742,-0.1698851883,-0.0174311064,-0.217944026,0.1805113405,0.3643777966,-0.2229035795,-0.1052753404,-0.0771993324,-0.4211502373,0.3274549544,-0.2325004041,0.1644014567,0.3366909623,-0.3936917186,-0.3250600696,0.4573607147,-0.3337568641,-0.1708879024,0.2365000099,-0.4048209786,0.4652937353,0.2076699138,0.0043006013,-0.0059952196,0.0376351476,0.1742016226,-0.1800449938,0.3051739931,0.1302233487,-0.2291006446,-0.0451015867,-0.1515160203,-0.00341606,-0.0958907083,-0.0259335395,-0.0169936493,-0.1171620861,0.4299755394,0.2623735666,0.0772481635,-0.2361664027,0.2408175766,0.2613838911,-0.0685643181,-0.0481198914,-0.0634819716,0.0062270737,-0.1658295989,0.0544401035,-0.3115944862,-0.2219983935,0.1507137269,0.1317087561,-0.1692024469,-0.3026206493,0.1607024223,0.2380982488,0.1132995039,0.302678138,-0.1198463142,0.5206785798,-0.3534439504,0.0884945691,-0.1319672763,-0.0621947758,0.257437855,-0.1281383336,-0.075649552,0.2081562132,-0.093410708,0.167134732,0.0381667428,0.4232925177,-0.0043456359,0.1183209568,-0.3124951422,0.0500708185,0.0097992299,-0.0047034323,0.0895714983,0.1420443952,-0.1465670317,0.1471806169,0.0384487025,-0.0887246281,0.172126621,0.150696978,-0.0438549928,0.0488436855,-0.1227513924,0.1988383234,-0.4711333811,-0.3036490977,-0.2785580456,-0.1713651419,0.303311497,0.1184073612,-0.3035778403,0.2379687428,0.7682510614,-0.0301368106,0.0205507204,0.0657910481,-0.2529127598,0.0181573071,0.1102729589,0.2093629688,0.1760621667,0.2615461648,0.2001331747,0.0508067124,0.006275434,0.0126902675,0.2241543382,0.0651092231,-0.2292538136,0.0869500786,-0.0851280615,-0.226238057,-0.2331650257,0.2796397209,-0.5140915513,-0.1945192218,-0.3919097185,0.0182173494,-0.2353746146,-0.4033135772,-0.5682982802,-0.0971945673,-0.0938994214,-0.3950068057,0.3544939458,-0.0493450053,-0.2513554096,0.0566962585,0.3190596402,0.276720196,0.0284163244,0.1718168855,-0.4514873326,0.2630797327,-0.4042059481,0.0542790033,-0.2848892212,0.1247040033,0.3621223271,-0.2116229683,-0.0718007088,-0.2785495222,-0.0587097704,0.2219152004,-0.2349594533,0.5601465702,0.3041003048,0.0496970862,0.2447494566,-0.1081943884,0.0988120511,-0.3349266946,-0.0422325283,-0.2306946963,0.1663376391,0.1885206699,-0.3057380319,-0.2492800206,-0.3592651486,-0.339918375,0.2599515617,-0.2132177651,0.2335208356,-0.1511961669,-0.112375237,0.1376269609,-0.3935935199,0.4972854555,-0.0881851092,-0.6302135587,-0.1551142335,-0.1244047806,-0.0908231884,-0.1654584408,0.1136178747,-0.2327751964,-0.2250757068,-0.5604439974,-0.2937108278,-0.2211570442,-0.1356341988,0.2647372186,0.3401813209,0.1737589836,0.115520142,-0.1217627823,-0.0775534511,-0.0946765617,0.0891970992,-0.3017315567,0.2018404305,0.2314448059,0.1525161266,0.0308498573,0.1316176653,0.3769789934,0.123182714,-0.2599914372,-0.1977273077,0.9344843626,-0.307679683,-0.1384760588,-0.0149093661,0.1217967272,-0.3135747612,0.1371774822,-0.0295712668,0.4288767874,0.1746635884,0.0135243889,-0.1403273195,-0.4702349305,-0.0877080634,0.0079168966,-0.1061840877,-0.0382950082,0.0241903253,0.2270083129,0.1013319269,-0.00780894,0.4861997366,0.3321846426,0.3255206048,-0.2572869956,-0.3790225089,-0.4488613307,-0.1084599718,-0.1151634306,-0.017060563,-0.2848509848,-0.3079065681,0.3015505672,0.1109932289,0.4875513315,-0.079973124,0.073655352,-0.0477083251,0.0399414524,-0.059490405,-0.0044629616,-0.2306834012,0.2239956409,0.3898353279,-0.066558741,-0.2347535342,-0.2896490693,0.3323343992,-0.1647358984,-0.2118754387,-0.3469479084,-0.4939076304,0.0554590523,-0.0708007365,0.163312003,0.1989546269,-0.0298537724,-0.4865504801,-0.4637316167,0.2801906466,0.1129297242,0.0338786356,-0.0750611946,0.2129462212,-0.2921409607,-0.2156690061,0.3172504306,0.5881595612,0.0688567162,0.2073207051,0.0794393122,0.0881092101,0.0209725481,-0.2785405815,0.3618935943,0.2727494836,-0.2944973111,0.025618555,-0.2036672831,0.26322335,-0.2290896922,0.0429350436,0.3444985449,0.211792171,-0.1790154129,-0.3018656075,0.0961390808,-0.0411252938,-0.1296864152,0.3108427823,0.2002142072,-0.3194753826,0.0456749983,-0.1375168711,0.9276247025,-0.0491979346,0.2219365239,0.2238164544,-0.2233250439,0.395976752,0.0447032861,0.1350615174,-0.2658403814,-0.1578319818,0.1012426242,-0.237329796,-0.009509854,0.034193635,-0.1483629793,0.039288681,0.0145296576,-0.047416769,-0.0873329043,0.1859939992,-0.1557833403,0.0478831157,-0.4868294299,0.0810601115,0.1670165062,-0.0533475392,-0.0447651483,0.0376888737,0.4169229269,-0.3255182803,-0.0336936675,-0.2197085619,0.2313967645,-0.1718175113,0.1313592643,0.0232822783,-0.1901381165,0.2200405151,-0.0098535968,0.5312997103,0.2272986025,-0.3040031791,0.2454185933,0.1460796744,-0.0669892356,0.1092391759,0.3509557247,0.0825152323,-0.4282553792,0.2340435237,-0.0461784638,-0.2830713093,0.075105615,-0.0555660836,0.1881588399,-0.3063421845,0.1682015806,0.1003822908,-0.0473899841,0.0465488769,0.1458730847,0.3736649752,-0.3556673229,0.3597012758,-0.2477943003,-0.642005384,-0.0725754127,-0.0022319236,0.141022861,0.132180348,0.1211481318,0.0531031005,-0.3038240075,-0.1832686812,0.0849186331,0.2430808395,0.0785034895,0.1308674216,0.3694796562,-0.0232785456,-0.092341423,0.1055511236,-0.1175148562,-0.0438433401,-0.5810171962,-0.1720307171,-0.2917281985,0.2343805432,-0.1726208925,0.3928425014,0.3371572495,-0.0697138533,0.1000314876,-0.0485731177,-0.3665171862,-0.3736622632,-0.1796021312,0.1179320142,0.2822411954,0.0690285265,0.3685359657,0.01983748,0.1314409077,-0.0938631147,-0.2340662032,-0.2157686204,0.1162944883,0.114927724,-0.1505063176,0.3382869363,-0.1192692071,-0.1658317447,0.2831532359,0.0107908174,0.830278039,0.3805103302,-0.0503564626,0.0797237307,0.0365441442,0.0940266699,-0.0143634072,0.1291987598,0.0394773223,-0.1125794873,-0.002794856,-0.0728810653,-0.4296863377,-0.061837256,0.4453523755,0.2021913081,0.1739319414,0.532674551,0.19151555,-0.0546605214,-0.125648424,-0.2378994524,0.404691875,0.0986345783,-0.076989457,-0.4025780559,-0.1153149083,0.1952556074,0.0609416366,0.0851305574,-0.1067547798,-0.3498922884,0.0727883726,0.3925569355,0.0188669041,-0.0715337768,0.4490739703,0.1405650377,0.1830877066,-0.352250427,-0.1274860352,0.0602809526,-0.3531312048,0.1583711207,0.4669979513,0.1810620874,-0.1216159388,0.0250403434,0.1749839783,0.1975757778,0.2322557867,-0.1198534518,-0.0334317647,0.1030743644,0.0993200094,0.2516612709,-0.2458589971,-0.0670951381,-0.0377258807,0.0330495797,-0.2389986664,0.1359064728,0.0128854588,-0.1712244004,-0.1869764328,0.1009695604,0.0068519246,0.0185961127,-0.1414887011,-0.1694042832,0.2064240277,-0.0798680559,-0.0593565442,-0.0337697864,0.1082217395,-0.1675757766,-0.2668060064,-0.17477265,0.0726372004,-0.0165566001,0.1455656737,-0.1115003303,0.4508191943,0.3521215916,0.2865580618,0.3099737763,0.1852804422,0.335639596,0.0134086935,-0.1203370839,-0.076482743,-0.335362643,0.3160897195,-0.0129126618,-0.2447958291,0.1927164644,0.2330977619,-0.1097833365,0.0718364641,-0.0239260066,0.0817841366,-0.2276338488,0.6042565107,0.1579875499,-0.2104892135,-0.2526172996,-0.0216332469,-0.6160032153,-0.113065131,0.4817127585,-0.0899934545,-0.0288039893,-0.5377426744,0.1322622001,0.1168787703,0.2316972017,0.4534968138,0.3030885756,0.0223234724,0.0053229923,-0.2670538425,0.3397335708,0.0974476114,-0.1241304278,-0.1588825732,-0.1152591482,0.0200446527,0.1332339197,0.2141840309,-0.0767310113,-0.4619903266,-0.2051266581,-0.0448314436,0.0367866196,0.3576841354,0.1559441537,-0.0222836807,-0.0010504022,0.2269449234,-0.2366591245,0.0921262354,0.3040310442,0.3641937077,-0.0291611198,-0.1873098761,0.2222921848,0.2810076177,0.0603166148,-0.0935685337,-0.1979129612,-0.1583644599,0.0454434939,-0.3171055019,0.0281604026,-0.1240469068,0.433672756,-0.0226715188,-0.2668972611,-0.2442461401,0.0407407731,0.2229068875,0.043587625,0.3966628015,-0.0439386182,-0.1420144588,-0.0084749917,0.1539933532,0.346327424,0.2588292062,-0.0758059919,-0.0459534079,-0.218385756,0.2991540432,-0.3736026883,-0.3177954555,-0.1893786937,0.3814629614,0.0958505198,-0.1871172041,-0.5491605401,-0.103808701,0.2456925511,-0.1885929108,0.1300996095,0.1423919946,-0.3548879027,0.2615862787,-0.0409369431,0.2532685995,-0.0565302707,0.2077296078,-0.0719627589,-0.5686923265]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2000","title":"Windows Permission Error (most recent version of datasets)","comments":"Hi @itsLuisa !\r\n\r\nCould you give us more information about the error you're getting, please?\r\nA copy-paste of the Traceback would be nice to get a better understanding of what is wrong :) ","body":"Hi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n","comment_length":33,"text":"Windows Permission Error (most recent version of datasets)\nHi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n\nHi @itsLuisa !\r\n\r\nCould you give us more information about the error you're getting, please?\r\nA copy-paste of the Traceback would be nice to get a better understanding of what is wrong :) ","embeddings":[-0.1825569272,0.1823850721,-0.0400314778,0.2590482235,0.0813782141,0.1336378604,0.4580342472,0.0122542866,0.1864634454,0.0741284713,-0.0634557605,-0.0112269307,-0.1087958515,0.1572928578,-0.0323617235,-0.0282511655,0.1453506798,0.0480685383,0.0319974497,0.1396457404,-0.4052302539,-0.0071208877,-0.2379359007,0.1073663011,-0.2231619507,0.278344065,-0.0157263745,0.3908289373,-0.0338031724,-0.2350183874,0.2101650983,0.3624857664,0.662093997,0.245353058,-0.0001126285,-0.048450049,0.0420433767,-0.0201318935,0.2068192661,0.2423872203,0.1341389418,-0.0451870672,0.0041253492,-0.1066693291,-0.1698371172,0.2000023574,-0.0654302686,-0.3538345098,0.1139979437,0.4263345599,0.1871957481,0.2474624962,-0.1806875318,-0.0106535321,0.5910394788,0.047598172,-0.1261592358,0.1927935481,0.2900241613,-0.4428050816,0.214141503,0.1609825641,-0.2268584669,0.0855402425,0.3818270862,-0.163264066,-0.1299571693,-0.2830388844,0.1584062725,0.00005305,0.727063477,-0.1692433953,-0.3009923995,-0.0341043584,0.1393089741,-0.2671689689,0.2275048792,0.3365985155,-0.0505269095,0.0753582716,-0.2347000837,-0.167150721,-0.4434008598,0.1358958036,0.1775047481,-0.2292983681,-0.084932968,0.2297086567,0.0906637236,-0.1301721931,0.3544327915,0.1294433624,0.0419623218,0.2517595291,-0.4671482146,0.3360512853,0.2092746347,0.2980217934,-0.0320270881,0.2779500782,-0.1361709833,-0.2106272429,0.2018981278,0.1250198483,0.1772745848,-0.0542738736,0.1519606858,0.2411280125,0.1375094652,-0.0225111879,0.0402428471,-0.117440857,-0.4900486767,-0.6061118841,0.1152087823,0.2454818487,0.3073368371,-0.0892671719,-0.3613361418,0.1164757684,0.1181355715,0.0680992082,0.2068185359,0.4097331166,0.2663407922,0.0119451405,0.3739327192,0.2714998126,-0.0628641695,-0.0258001015,-0.0862123594,0.0042463713,-0.2458985001,-0.0498456098,0.4257872105,-0.344258368,-0.1134994924,0.1173631623,0.0353439488,-0.1213498861,0.0494542569,-0.0187720023,0.1164598688,0.4388581216,0.2584182322,0.034997154,0.1450432539,-0.1851189882,-0.0282773599,0.2364417911,-0.132414341,-0.3404768407,-0.1883570999,0.1011299416,0.0170740597,0.0499140508,-0.218305707,-0.133036688,0.1228498816,-0.0184608568,0.1442888379,-0.1007024646,-0.2660652995,-0.3397949338,-0.1015146449,0.8314390182,-0.5272637606,0.1402395666,-0.1147450805,-0.3047668636,0.1072497219,0.3311969638,-0.0764998347,0.1113562062,-0.214482978,0.1516456157,-0.1357124448,-0.3680853248,-0.1788965762,0.3780059218,-0.017753955,-0.103679657,0.2437345237,-0.0390078723,0.0444860794,-0.2033561915,-0.0898012519,0.134128198,0.1384616196,0.2605628967,0.2249749005,0.0075680758,0.2322096527,0.3124230802,0.0553449914,-0.0627514571,0.2918676436,-0.2186992466,0.189552173,-0.3010883629,0.2245462686,0.4166680276,0.1158194914,0.3091973662,0.0332210027,-0.2322232127,-0.4455744326,0.2652016878,0.3054819703,-0.1239581481,0.0469031185,-0.2565841079,-0.1502389461,0.2109269798,-0.3883987963,0.1167708114,0.0899116099,0.297970891,-0.1774025708,-0.320898056,0.034541335,0.0563448928,-0.2444298863,-0.1111344919,-0.153433159,0.0384588875,-0.3066704273,0.0180423688,0.2049439102,-0.0148219531,0.415769726,-0.074380219,0.0026424972,0.3591325581,0.2803249657,-0.0367785916,0.036796052,0.0983095542,-0.2261272371,0.0771442056,0.1272577792,-0.0095818341,0.0916867405,-0.1025229618,-0.0255540945,0.1265098751,-0.2259499133,0.0041122376,-0.2879075408,0.1314649135,-0.0096176816,0.0248836838,-0.024606308,-0.2199265212,0.3593811691,0.5096359849,-0.3285354376,-0.0453583524,-0.102629602,-0.0515072905,0.2793931961,-0.1505599469,-0.1328611523,0.068707265,0.0991909578,0.2852436006,0.1979034096,0.2263821661,0.4385739267,0.0644598529,-0.1653025299,0.0529506505,0.2032765597,-0.0528064743,0.1483865529,-0.3353100717,0.0793047324,0.2603348792,-0.2992096245,0.1275261194,-0.3775518239,-0.1403336674,0.1440867037,0.3369528651,-0.5138478875,-0.1326299012,-0.0149593158,-0.1334604025,0.1105089188,-0.3155808747,-0.2831487656,0.0130276047,-0.201353848,0.1626410633,-0.2732628286,-0.1256408542,-0.4015872478,0.0573596619,0.2376947105,-0.0031728048,-0.0217249822,0.0865617096,-0.023845451,0.0146014299,0.49329561,-0.2829744816,0.4370732903,-0.1266709417,0.3814791143,-0.2898307145,0.106090866,-0.2475935519,-0.3344739974,0.3203744292,0.246234566,0.0983890146,-0.0262450539,-0.2058806717,0.1195861474,-0.1224583089,-0.0811637938,0.1755108088,0.3042165041,-0.5918186307,-0.25063923,-0.1534152478,-0.2042932659,-0.3533948064,0.2209898382,0.135534808,0.3110541701,0.2072859555,-0.1218600199,0.0530254319,0.0434561335,0.3111526668,-0.3197958469,-0.0964614972,0.013731421,-0.1502657682,-0.5285490155,0.220506236,0.3835156858,-0.1217221916,0.1091838852,-0.1158387959,0.1251094043,-0.2140979767,0.2546712458,-0.2282879651,0.1662585884,0.1928665638,0.0734918788,0.0805564895,-0.1694658101,-0.0990942419,-0.3450900614,-0.2438754141,-0.264610678,0.154290542,0.3330400884,-0.131513983,0.2149935812,0.3426207602,0.2264254987,0.4461936355,-0.1276842952,0.4316768348,0.2818519771,-0.6437909603,0.1521578729,-0.1861309707,-0.053072609,0.3601593673,-0.0777115822,0.0294729602,-0.1574553996,0.0583065599,-0.3378964961,-0.2107926905,-0.0072617149,-0.2567565143,0.2782015204,-0.1479895264,0.0072967722,-0.2808846235,-0.0628952831,0.0054723453,0.568931222,0.1167742088,-0.061601568,-0.3760638535,0.0078706741,-0.4856024384,0.26964733,0.0624525361,0.6091942191,-0.1555267274,0.0315033421,0.1437783986,0.1085258052,0.9010464549,-0.4391025007,0.2763887942,0.2481005341,0.1166944727,-0.263177067,-0.211541608,-0.2682219446,0.067434594,-0.0131370779,0.3030076325,0.0118963942,-0.3937038779,0.0339972079,-0.1530512869,-0.0794086903,-0.4198121428,-0.253552407,-0.1815235615,-0.3771493137,-0.1041810736,0.2397942096,0.1634427458,-0.2563313544,0.2497168332,-0.0224103611,0.2882218957,0.1148120835,-0.0406747237,0.1016073227,0.3072133064,0.0786264092,-0.0809458047,0.1236321405,0.2123339176,0.5588169098,-0.026124591,-0.3647992313,-0.1107676178,-0.0270074625,0.4848950505,0.0845985115,-0.153303653,0.1014053077,-0.2211105675,-0.058644563,-0.3480154276,0.0739530697,0.2731055617,-0.0515423417,0.0595365576,-0.2157613635,0.3164066672,0.1723264158,-0.1943370253,-0.0005688795,0.1422991753,-0.2947748601,-0.0935064554,0.0510758013,0.5023060441,-0.2940065563,0.2797781527,0.1955997795,-0.1659105569,-0.071143426,0.213571772,-0.0846962035,-0.4017235041,-0.3005031049,-0.0134299649,-0.1052668542,-0.0054742121,0.1585410982,0.0849552229,0.0400238074,-0.0924142972,0.3835968375,0.0317604132,0.0596932992,-0.202075392,-0.2441049665,0.1019104496,0.151968509,-0.1361004859,-0.0235723518,-0.1311588436,0.0135436291,0.1534249038,0.0239182711,-0.3978714943,0.2855559587,-0.3279962242,0.0107293762,-0.1360235959,0.0947399661,0.5240969658,0.4402703941,-0.0064931316,0.1454547793,-0.2531163096,-0.2666570544,0.1540938467,-0.2917832136,-0.291580379,-0.0236652046,0.2547366619,0.0212811362,-0.3019715846,-0.1922039539,-0.0331776142,-0.2432691008,-0.0630597994,0.4675462842,0.1187870353,-0.507037282,-0.1682941318,-0.2320503443,-0.0584004521,0.0715503842,0.1030866802,0.1528616995,-0.1085439026,0.0619823374,-0.2745189965,-0.2699647248,0.0699824989,0.4720028639,-0.2750083506,0.1021838114,0.5445803404,0.4029547274,-0.1602605581,-0.2423876524,0.0846469402,0.0800300464,-0.2135888487,0.0007553757,0.312019676,0.0452671908,0.4206763506,0.0679824278,0.2228612602,-0.4020850658,0.1897781193,-0.4625521302,-0.110298954,-0.0878793746,-0.1858239919,0.3120950758,0.0934090316,-0.1442581862,0.0303128902,-0.2334226519,-0.29327479,0.1127857938,-0.0025985991,0.1187579483,0.4522465765,-0.229709357,-0.0856508985,-0.0219585318,0.0636571944,-0.1481769979,-0.1781420857,-0.2058385909,0.0039554443,0.1113952175,0.187864542,-0.3387500346,0.0858408138,-0.5403545499,-0.2182604074,-0.1620232761,0.2066287249,0.2949583232,-0.1138813123,0.4738371372,0.0064511998,-0.0148760919,0.0212341156,0.191027984,-0.1915093064,0.0859473795,-0.3789152205,0.4132497311,0.0258933119,-0.0854482129,-0.3049198687,0.2492909878,0.2421408892,0.1775051206,0.136467576,-0.2938653827,0.0452842601,0.1257751435,0.5298547745,0.0940537453,-0.3113827109,0.0305275526,0.1404470205,0.203754127,-0.3257121444,-0.1571353674,-0.2650839388,-0.1184605137,-0.0624468252,0.115416728,0.0112674851,-0.3411978185,-0.0892772675,-0.1154887825,0.2970600128,0.2271778584,-0.0939651132,0.5441298485,0.0009060174,-0.0297029205,0.3708809316,0.5498604774,0.0500368364,0.424381882,0.194929719,-0.1030110866,-0.1730057895,0.0265199021,-0.0050954581,-0.7224727273,0.1903757304,-0.0895253345,-0.0153279537,0.0070986184,-0.1846191287,0.4962287545,-0.1142299399,-0.0175317507,-0.3881719708,0.1512207389,-0.2028812766,-0.1655408293,-0.4238615632,0.0989473909,-0.0877483711,-0.0205381736,-0.1252162606,-0.2460019886,0.0037858104,0.228457734,-0.0325791202,-0.2135648876,0.349271208,0.2529511154,-0.1378227025,-0.3874364793,-0.0780463591,-0.0577866174,0.1203663275,0.1329323798,0.0222340655,0.6245731711,0.3318480551,-0.0558945425,-0.012673567,0.0129633136,-0.0314593799,-0.035633672,0.3866260946,-0.2138810754,0.3700492382,0.3404141068,0.0974966511,-0.058864668,-0.0613856651,0.0397349261,-0.1048275977,-0.0849886537,-0.0384141132,-0.3763240874,-0.2585601211,-0.1624862403,0.1023473442,-0.0869153216,-0.0223293044,0.6900111437,0.1503067762,0.341988951,-0.3378275931,0.0661960021,0.0295750573,0.4036607146,0.470158726,0.0760127977,-0.4037770629,-0.0799946636,-0.4968746305,0.0621396154,0.1211791858,-0.1317172647,0.1792279482,-0.136156112,-0.013322914,-0.1985217631,-0.0573430285,0.0297257993,0.0242389757,0.3217083514,-0.0086836405,-0.1077249572,0.2655580938,-0.0672235638,0.1426306814,-0.5638715625,0.0671390519,-0.2215017527,0.0575661026,-0.4719406068,0.0898197666,0.1238537803,0.1396856904,0.2431280315,0.0155660342,0.3771497309,-0.0043464405,0.1913803816,-0.5739651918,0.0678090081,-0.1726519465,0.1434074193,-0.1353252679,-0.0251373108,-0.2696663737,0.0728011355,-0.0983978063,0.1188255697,-0.0468874201,0.1527613699,-0.051272545,0.1282429248,-0.1833435446,-0.0416468568,0.2216355652,-0.1437097043,-0.1895443648,0.3802645802,-0.3233775496,-0.3723746538,0.4683177173,-0.2650411427,-0.2512885928,0.0131054819,0.3566480577,-0.1201093346,-0.1665672511,-0.1439184397,0.3625415266,0.1724920869,-0.004814927,-0.3380492628,0.4713961482,-0.2036845535,-0.1848139763,0.0364141874,0.4234890342,-0.0084362719,-0.2079798728,0.0341635793,-0.115126878]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2000","title":"Windows Permission Error (most recent version of datasets)","comments":"Hello @SBrandeis , this is it:\r\n```\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 537, in incomplete_dir\r\n yield tmp_dir\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 656, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 982, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\arrow_writer.py\", line 297, in finalize\r\n self.write_on_file()\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\arrow_writer.py\", line 230, in write_on_file\r\n pa_array = pa.array(typed_sequence)\r\n File \"pyarrow\\array.pxi\", line 222, in pyarrow.lib.array\r\n File \"pyarrow\\array.pxi\", line 110, in pyarrow.lib._handle_arrow_array_protocol\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\arrow_writer.py\", line 97, in __arrow_array__\r\n out = pa.array(self.data, type=type)\r\n File \"pyarrow\\array.pxi\", line 305, in pyarrow.lib.array\r\n File \"pyarrow\\array.pxi\", line 39, in pyarrow.lib._sequence_to_array\r\n File \"pyarrow\\error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\\error.pxi\", line 107, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowTypeError: Expected bytes, got a 'list' object\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"C:\/Users\/Luisa\/Documents\/Uni\/WS 2020,21\/Neural Networks\/Final_Project\/NN_Project\/data_loading.py\", line 122, in \r\n main()\r\n File \"C:\/Users\/Luisa\/Documents\/Uni\/WS 2020,21\/Neural Networks\/Final_Project\/NN_Project\/data_loading.py\", line 111, in main\r\n dataset = datasets.load_dataset(\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 586, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\contextlib.py\", line 131, in __exit__\r\n self.gen.throw(type, value, traceback)\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 543, in incomplete_dir\r\n shutil.rmtree(tmp_dir)\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\shutil.py\", line 740, in rmtree\r\n return _rmtree_unsafe(path, onerror)\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\shutil.py\", line 618, in _rmtree_unsafe\r\n onerror(os.unlink, fullname, sys.exc_info())\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\shutil.py\", line 616, in _rmtree_unsafe\r\n os.unlink(fullname)\r\nPermissionError: [WinError 32] Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird: 'C:\\\\Users\\\\Luisa\\\\.cache\\\\huggingface\\\\datasets\\\\sample\\\\default-20ee7d51a6a9454f\\\\0.0.0\\\\5fc4c3a355ea77ab446bd31fca5082437600b8364d29b2b95264048bd1f398b1.incomplete\\\\sample-train.arrow'\r\n\r\nProcess finished with exit code 1\r\n```","body":"Hi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n","comment_length":230,"text":"Windows Permission Error (most recent version of datasets)\nHi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n\nHello @SBrandeis , this is it:\r\n```\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 537, in incomplete_dir\r\n yield tmp_dir\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 656, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 982, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\arrow_writer.py\", line 297, in finalize\r\n self.write_on_file()\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\arrow_writer.py\", line 230, in write_on_file\r\n pa_array = pa.array(typed_sequence)\r\n File \"pyarrow\\array.pxi\", line 222, in pyarrow.lib.array\r\n File \"pyarrow\\array.pxi\", line 110, in pyarrow.lib._handle_arrow_array_protocol\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\arrow_writer.py\", line 97, in __arrow_array__\r\n out = pa.array(self.data, type=type)\r\n File \"pyarrow\\array.pxi\", line 305, in pyarrow.lib.array\r\n File \"pyarrow\\array.pxi\", line 39, in pyarrow.lib._sequence_to_array\r\n File \"pyarrow\\error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\\error.pxi\", line 107, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowTypeError: Expected bytes, got a 'list' object\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"C:\/Users\/Luisa\/Documents\/Uni\/WS 2020,21\/Neural Networks\/Final_Project\/NN_Project\/data_loading.py\", line 122, in \r\n main()\r\n File \"C:\/Users\/Luisa\/Documents\/Uni\/WS 2020,21\/Neural Networks\/Final_Project\/NN_Project\/data_loading.py\", line 111, in main\r\n dataset = datasets.load_dataset(\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 586, in download_and_prepare\r\n self._save_info()\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\contextlib.py\", line 131, in __exit__\r\n self.gen.throw(type, value, traceback)\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 543, in incomplete_dir\r\n shutil.rmtree(tmp_dir)\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\shutil.py\", line 740, in rmtree\r\n return _rmtree_unsafe(path, onerror)\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\shutil.py\", line 618, in _rmtree_unsafe\r\n onerror(os.unlink, fullname, sys.exc_info())\r\n File \"C:\\Users\\Luisa\\AppData\\Local\\Programs\\Python\\Python38\\lib\\shutil.py\", line 616, in _rmtree_unsafe\r\n os.unlink(fullname)\r\nPermissionError: [WinError 32] Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird: 'C:\\\\Users\\\\Luisa\\\\.cache\\\\huggingface\\\\datasets\\\\sample\\\\default-20ee7d51a6a9454f\\\\0.0.0\\\\5fc4c3a355ea77ab446bd31fca5082437600b8364d29b2b95264048bd1f398b1.incomplete\\\\sample-train.arrow'\r\n\r\nProcess finished with exit code 1\r\n```","embeddings":[-0.1825569272,0.1823850721,-0.0400314778,0.2590482235,0.0813782141,0.1336378604,0.4580342472,0.0122542866,0.1864634454,0.0741284713,-0.0634557605,-0.0112269307,-0.1087958515,0.1572928578,-0.0323617235,-0.0282511655,0.1453506798,0.0480685383,0.0319974497,0.1396457404,-0.4052302539,-0.0071208877,-0.2379359007,0.1073663011,-0.2231619507,0.278344065,-0.0157263745,0.3908289373,-0.0338031724,-0.2350183874,0.2101650983,0.3624857664,0.662093997,0.245353058,-0.0001126285,-0.048450049,0.0420433767,-0.0201318935,0.2068192661,0.2423872203,0.1341389418,-0.0451870672,0.0041253492,-0.1066693291,-0.1698371172,0.2000023574,-0.0654302686,-0.3538345098,0.1139979437,0.4263345599,0.1871957481,0.2474624962,-0.1806875318,-0.0106535321,0.5910394788,0.047598172,-0.1261592358,0.1927935481,0.2900241613,-0.4428050816,0.214141503,0.1609825641,-0.2268584669,0.0855402425,0.3818270862,-0.163264066,-0.1299571693,-0.2830388844,0.1584062725,0.00005305,0.727063477,-0.1692433953,-0.3009923995,-0.0341043584,0.1393089741,-0.2671689689,0.2275048792,0.3365985155,-0.0505269095,0.0753582716,-0.2347000837,-0.167150721,-0.4434008598,0.1358958036,0.1775047481,-0.2292983681,-0.084932968,0.2297086567,0.0906637236,-0.1301721931,0.3544327915,0.1294433624,0.0419623218,0.2517595291,-0.4671482146,0.3360512853,0.2092746347,0.2980217934,-0.0320270881,0.2779500782,-0.1361709833,-0.2106272429,0.2018981278,0.1250198483,0.1772745848,-0.0542738736,0.1519606858,0.2411280125,0.1375094652,-0.0225111879,0.0402428471,-0.117440857,-0.4900486767,-0.6061118841,0.1152087823,0.2454818487,0.3073368371,-0.0892671719,-0.3613361418,0.1164757684,0.1181355715,0.0680992082,0.2068185359,0.4097331166,0.2663407922,0.0119451405,0.3739327192,0.2714998126,-0.0628641695,-0.0258001015,-0.0862123594,0.0042463713,-0.2458985001,-0.0498456098,0.4257872105,-0.344258368,-0.1134994924,0.1173631623,0.0353439488,-0.1213498861,0.0494542569,-0.0187720023,0.1164598688,0.4388581216,0.2584182322,0.034997154,0.1450432539,-0.1851189882,-0.0282773599,0.2364417911,-0.132414341,-0.3404768407,-0.1883570999,0.1011299416,0.0170740597,0.0499140508,-0.218305707,-0.133036688,0.1228498816,-0.0184608568,0.1442888379,-0.1007024646,-0.2660652995,-0.3397949338,-0.1015146449,0.8314390182,-0.5272637606,0.1402395666,-0.1147450805,-0.3047668636,0.1072497219,0.3311969638,-0.0764998347,0.1113562062,-0.214482978,0.1516456157,-0.1357124448,-0.3680853248,-0.1788965762,0.3780059218,-0.017753955,-0.103679657,0.2437345237,-0.0390078723,0.0444860794,-0.2033561915,-0.0898012519,0.134128198,0.1384616196,0.2605628967,0.2249749005,0.0075680758,0.2322096527,0.3124230802,0.0553449914,-0.0627514571,0.2918676436,-0.2186992466,0.189552173,-0.3010883629,0.2245462686,0.4166680276,0.1158194914,0.3091973662,0.0332210027,-0.2322232127,-0.4455744326,0.2652016878,0.3054819703,-0.1239581481,0.0469031185,-0.2565841079,-0.1502389461,0.2109269798,-0.3883987963,0.1167708114,0.0899116099,0.297970891,-0.1774025708,-0.320898056,0.034541335,0.0563448928,-0.2444298863,-0.1111344919,-0.153433159,0.0384588875,-0.3066704273,0.0180423688,0.2049439102,-0.0148219531,0.415769726,-0.074380219,0.0026424972,0.3591325581,0.2803249657,-0.0367785916,0.036796052,0.0983095542,-0.2261272371,0.0771442056,0.1272577792,-0.0095818341,0.0916867405,-0.1025229618,-0.0255540945,0.1265098751,-0.2259499133,0.0041122376,-0.2879075408,0.1314649135,-0.0096176816,0.0248836838,-0.024606308,-0.2199265212,0.3593811691,0.5096359849,-0.3285354376,-0.0453583524,-0.102629602,-0.0515072905,0.2793931961,-0.1505599469,-0.1328611523,0.068707265,0.0991909578,0.2852436006,0.1979034096,0.2263821661,0.4385739267,0.0644598529,-0.1653025299,0.0529506505,0.2032765597,-0.0528064743,0.1483865529,-0.3353100717,0.0793047324,0.2603348792,-0.2992096245,0.1275261194,-0.3775518239,-0.1403336674,0.1440867037,0.3369528651,-0.5138478875,-0.1326299012,-0.0149593158,-0.1334604025,0.1105089188,-0.3155808747,-0.2831487656,0.0130276047,-0.201353848,0.1626410633,-0.2732628286,-0.1256408542,-0.4015872478,0.0573596619,0.2376947105,-0.0031728048,-0.0217249822,0.0865617096,-0.023845451,0.0146014299,0.49329561,-0.2829744816,0.4370732903,-0.1266709417,0.3814791143,-0.2898307145,0.106090866,-0.2475935519,-0.3344739974,0.3203744292,0.246234566,0.0983890146,-0.0262450539,-0.2058806717,0.1195861474,-0.1224583089,-0.0811637938,0.1755108088,0.3042165041,-0.5918186307,-0.25063923,-0.1534152478,-0.2042932659,-0.3533948064,0.2209898382,0.135534808,0.3110541701,0.2072859555,-0.1218600199,0.0530254319,0.0434561335,0.3111526668,-0.3197958469,-0.0964614972,0.013731421,-0.1502657682,-0.5285490155,0.220506236,0.3835156858,-0.1217221916,0.1091838852,-0.1158387959,0.1251094043,-0.2140979767,0.2546712458,-0.2282879651,0.1662585884,0.1928665638,0.0734918788,0.0805564895,-0.1694658101,-0.0990942419,-0.3450900614,-0.2438754141,-0.264610678,0.154290542,0.3330400884,-0.131513983,0.2149935812,0.3426207602,0.2264254987,0.4461936355,-0.1276842952,0.4316768348,0.2818519771,-0.6437909603,0.1521578729,-0.1861309707,-0.053072609,0.3601593673,-0.0777115822,0.0294729602,-0.1574553996,0.0583065599,-0.3378964961,-0.2107926905,-0.0072617149,-0.2567565143,0.2782015204,-0.1479895264,0.0072967722,-0.2808846235,-0.0628952831,0.0054723453,0.568931222,0.1167742088,-0.061601568,-0.3760638535,0.0078706741,-0.4856024384,0.26964733,0.0624525361,0.6091942191,-0.1555267274,0.0315033421,0.1437783986,0.1085258052,0.9010464549,-0.4391025007,0.2763887942,0.2481005341,0.1166944727,-0.263177067,-0.211541608,-0.2682219446,0.067434594,-0.0131370779,0.3030076325,0.0118963942,-0.3937038779,0.0339972079,-0.1530512869,-0.0794086903,-0.4198121428,-0.253552407,-0.1815235615,-0.3771493137,-0.1041810736,0.2397942096,0.1634427458,-0.2563313544,0.2497168332,-0.0224103611,0.2882218957,0.1148120835,-0.0406747237,0.1016073227,0.3072133064,0.0786264092,-0.0809458047,0.1236321405,0.2123339176,0.5588169098,-0.026124591,-0.3647992313,-0.1107676178,-0.0270074625,0.4848950505,0.0845985115,-0.153303653,0.1014053077,-0.2211105675,-0.058644563,-0.3480154276,0.0739530697,0.2731055617,-0.0515423417,0.0595365576,-0.2157613635,0.3164066672,0.1723264158,-0.1943370253,-0.0005688795,0.1422991753,-0.2947748601,-0.0935064554,0.0510758013,0.5023060441,-0.2940065563,0.2797781527,0.1955997795,-0.1659105569,-0.071143426,0.213571772,-0.0846962035,-0.4017235041,-0.3005031049,-0.0134299649,-0.1052668542,-0.0054742121,0.1585410982,0.0849552229,0.0400238074,-0.0924142972,0.3835968375,0.0317604132,0.0596932992,-0.202075392,-0.2441049665,0.1019104496,0.151968509,-0.1361004859,-0.0235723518,-0.1311588436,0.0135436291,0.1534249038,0.0239182711,-0.3978714943,0.2855559587,-0.3279962242,0.0107293762,-0.1360235959,0.0947399661,0.5240969658,0.4402703941,-0.0064931316,0.1454547793,-0.2531163096,-0.2666570544,0.1540938467,-0.2917832136,-0.291580379,-0.0236652046,0.2547366619,0.0212811362,-0.3019715846,-0.1922039539,-0.0331776142,-0.2432691008,-0.0630597994,0.4675462842,0.1187870353,-0.507037282,-0.1682941318,-0.2320503443,-0.0584004521,0.0715503842,0.1030866802,0.1528616995,-0.1085439026,0.0619823374,-0.2745189965,-0.2699647248,0.0699824989,0.4720028639,-0.2750083506,0.1021838114,0.5445803404,0.4029547274,-0.1602605581,-0.2423876524,0.0846469402,0.0800300464,-0.2135888487,0.0007553757,0.312019676,0.0452671908,0.4206763506,0.0679824278,0.2228612602,-0.4020850658,0.1897781193,-0.4625521302,-0.110298954,-0.0878793746,-0.1858239919,0.3120950758,0.0934090316,-0.1442581862,0.0303128902,-0.2334226519,-0.29327479,0.1127857938,-0.0025985991,0.1187579483,0.4522465765,-0.229709357,-0.0856508985,-0.0219585318,0.0636571944,-0.1481769979,-0.1781420857,-0.2058385909,0.0039554443,0.1113952175,0.187864542,-0.3387500346,0.0858408138,-0.5403545499,-0.2182604074,-0.1620232761,0.2066287249,0.2949583232,-0.1138813123,0.4738371372,0.0064511998,-0.0148760919,0.0212341156,0.191027984,-0.1915093064,0.0859473795,-0.3789152205,0.4132497311,0.0258933119,-0.0854482129,-0.3049198687,0.2492909878,0.2421408892,0.1775051206,0.136467576,-0.2938653827,0.0452842601,0.1257751435,0.5298547745,0.0940537453,-0.3113827109,0.0305275526,0.1404470205,0.203754127,-0.3257121444,-0.1571353674,-0.2650839388,-0.1184605137,-0.0624468252,0.115416728,0.0112674851,-0.3411978185,-0.0892772675,-0.1154887825,0.2970600128,0.2271778584,-0.0939651132,0.5441298485,0.0009060174,-0.0297029205,0.3708809316,0.5498604774,0.0500368364,0.424381882,0.194929719,-0.1030110866,-0.1730057895,0.0265199021,-0.0050954581,-0.7224727273,0.1903757304,-0.0895253345,-0.0153279537,0.0070986184,-0.1846191287,0.4962287545,-0.1142299399,-0.0175317507,-0.3881719708,0.1512207389,-0.2028812766,-0.1655408293,-0.4238615632,0.0989473909,-0.0877483711,-0.0205381736,-0.1252162606,-0.2460019886,0.0037858104,0.228457734,-0.0325791202,-0.2135648876,0.349271208,0.2529511154,-0.1378227025,-0.3874364793,-0.0780463591,-0.0577866174,0.1203663275,0.1329323798,0.0222340655,0.6245731711,0.3318480551,-0.0558945425,-0.012673567,0.0129633136,-0.0314593799,-0.035633672,0.3866260946,-0.2138810754,0.3700492382,0.3404141068,0.0974966511,-0.058864668,-0.0613856651,0.0397349261,-0.1048275977,-0.0849886537,-0.0384141132,-0.3763240874,-0.2585601211,-0.1624862403,0.1023473442,-0.0869153216,-0.0223293044,0.6900111437,0.1503067762,0.341988951,-0.3378275931,0.0661960021,0.0295750573,0.4036607146,0.470158726,0.0760127977,-0.4037770629,-0.0799946636,-0.4968746305,0.0621396154,0.1211791858,-0.1317172647,0.1792279482,-0.136156112,-0.013322914,-0.1985217631,-0.0573430285,0.0297257993,0.0242389757,0.3217083514,-0.0086836405,-0.1077249572,0.2655580938,-0.0672235638,0.1426306814,-0.5638715625,0.0671390519,-0.2215017527,0.0575661026,-0.4719406068,0.0898197666,0.1238537803,0.1396856904,0.2431280315,0.0155660342,0.3771497309,-0.0043464405,0.1913803816,-0.5739651918,0.0678090081,-0.1726519465,0.1434074193,-0.1353252679,-0.0251373108,-0.2696663737,0.0728011355,-0.0983978063,0.1188255697,-0.0468874201,0.1527613699,-0.051272545,0.1282429248,-0.1833435446,-0.0416468568,0.2216355652,-0.1437097043,-0.1895443648,0.3802645802,-0.3233775496,-0.3723746538,0.4683177173,-0.2650411427,-0.2512885928,0.0131054819,0.3566480577,-0.1201093346,-0.1665672511,-0.1439184397,0.3625415266,0.1724920869,-0.004814927,-0.3380492628,0.4713961482,-0.2036845535,-0.1848139763,0.0364141874,0.4234890342,-0.0084362719,-0.2079798728,0.0341635793,-0.115126878]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2000","title":"Windows Permission Error (most recent version of datasets)","comments":"Hi @itsLuisa, thanks for sharing the Traceback.\r\n\r\nYou are defining the \"id\" field as a `string` feature:\r\n```python\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n ...\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n # ^^ here\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(datasets.features.ClassLabel(names=[...])),\r\n[...]\r\n```\r\n\r\nBut in the `_generate_examples`, the \"id\" field is a list:\r\n```python\r\nids = list()\r\n```\r\n\r\nChanging:\r\n```python\r\n\"id\": datasets.Value(\"string\"),\r\n```\r\nInto:\r\n```python\r\n\"id\": datasets.Sequence(datasets.Value(\"string\")),\r\n```\r\n\r\nShould fix your issue.\r\n\r\nLet me know if this helps!","body":"Hi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n","comment_length":73,"text":"Windows Permission Error (most recent version of datasets)\nHi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n\nHi @itsLuisa, thanks for sharing the Traceback.\r\n\r\nYou are defining the \"id\" field as a `string` feature:\r\n```python\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n ...\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n # ^^ here\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(datasets.features.ClassLabel(names=[...])),\r\n[...]\r\n```\r\n\r\nBut in the `_generate_examples`, the \"id\" field is a list:\r\n```python\r\nids = list()\r\n```\r\n\r\nChanging:\r\n```python\r\n\"id\": datasets.Value(\"string\"),\r\n```\r\nInto:\r\n```python\r\n\"id\": datasets.Sequence(datasets.Value(\"string\")),\r\n```\r\n\r\nShould fix your issue.\r\n\r\nLet me know if this helps!","embeddings":[-0.1825569272,0.1823850721,-0.0400314778,0.2590482235,0.0813782141,0.1336378604,0.4580342472,0.0122542866,0.1864634454,0.0741284713,-0.0634557605,-0.0112269307,-0.1087958515,0.1572928578,-0.0323617235,-0.0282511655,0.1453506798,0.0480685383,0.0319974497,0.1396457404,-0.4052302539,-0.0071208877,-0.2379359007,0.1073663011,-0.2231619507,0.278344065,-0.0157263745,0.3908289373,-0.0338031724,-0.2350183874,0.2101650983,0.3624857664,0.662093997,0.245353058,-0.0001126285,-0.048450049,0.0420433767,-0.0201318935,0.2068192661,0.2423872203,0.1341389418,-0.0451870672,0.0041253492,-0.1066693291,-0.1698371172,0.2000023574,-0.0654302686,-0.3538345098,0.1139979437,0.4263345599,0.1871957481,0.2474624962,-0.1806875318,-0.0106535321,0.5910394788,0.047598172,-0.1261592358,0.1927935481,0.2900241613,-0.4428050816,0.214141503,0.1609825641,-0.2268584669,0.0855402425,0.3818270862,-0.163264066,-0.1299571693,-0.2830388844,0.1584062725,0.00005305,0.727063477,-0.1692433953,-0.3009923995,-0.0341043584,0.1393089741,-0.2671689689,0.2275048792,0.3365985155,-0.0505269095,0.0753582716,-0.2347000837,-0.167150721,-0.4434008598,0.1358958036,0.1775047481,-0.2292983681,-0.084932968,0.2297086567,0.0906637236,-0.1301721931,0.3544327915,0.1294433624,0.0419623218,0.2517595291,-0.4671482146,0.3360512853,0.2092746347,0.2980217934,-0.0320270881,0.2779500782,-0.1361709833,-0.2106272429,0.2018981278,0.1250198483,0.1772745848,-0.0542738736,0.1519606858,0.2411280125,0.1375094652,-0.0225111879,0.0402428471,-0.117440857,-0.4900486767,-0.6061118841,0.1152087823,0.2454818487,0.3073368371,-0.0892671719,-0.3613361418,0.1164757684,0.1181355715,0.0680992082,0.2068185359,0.4097331166,0.2663407922,0.0119451405,0.3739327192,0.2714998126,-0.0628641695,-0.0258001015,-0.0862123594,0.0042463713,-0.2458985001,-0.0498456098,0.4257872105,-0.344258368,-0.1134994924,0.1173631623,0.0353439488,-0.1213498861,0.0494542569,-0.0187720023,0.1164598688,0.4388581216,0.2584182322,0.034997154,0.1450432539,-0.1851189882,-0.0282773599,0.2364417911,-0.132414341,-0.3404768407,-0.1883570999,0.1011299416,0.0170740597,0.0499140508,-0.218305707,-0.133036688,0.1228498816,-0.0184608568,0.1442888379,-0.1007024646,-0.2660652995,-0.3397949338,-0.1015146449,0.8314390182,-0.5272637606,0.1402395666,-0.1147450805,-0.3047668636,0.1072497219,0.3311969638,-0.0764998347,0.1113562062,-0.214482978,0.1516456157,-0.1357124448,-0.3680853248,-0.1788965762,0.3780059218,-0.017753955,-0.103679657,0.2437345237,-0.0390078723,0.0444860794,-0.2033561915,-0.0898012519,0.134128198,0.1384616196,0.2605628967,0.2249749005,0.0075680758,0.2322096527,0.3124230802,0.0553449914,-0.0627514571,0.2918676436,-0.2186992466,0.189552173,-0.3010883629,0.2245462686,0.4166680276,0.1158194914,0.3091973662,0.0332210027,-0.2322232127,-0.4455744326,0.2652016878,0.3054819703,-0.1239581481,0.0469031185,-0.2565841079,-0.1502389461,0.2109269798,-0.3883987963,0.1167708114,0.0899116099,0.297970891,-0.1774025708,-0.320898056,0.034541335,0.0563448928,-0.2444298863,-0.1111344919,-0.153433159,0.0384588875,-0.3066704273,0.0180423688,0.2049439102,-0.0148219531,0.415769726,-0.074380219,0.0026424972,0.3591325581,0.2803249657,-0.0367785916,0.036796052,0.0983095542,-0.2261272371,0.0771442056,0.1272577792,-0.0095818341,0.0916867405,-0.1025229618,-0.0255540945,0.1265098751,-0.2259499133,0.0041122376,-0.2879075408,0.1314649135,-0.0096176816,0.0248836838,-0.024606308,-0.2199265212,0.3593811691,0.5096359849,-0.3285354376,-0.0453583524,-0.102629602,-0.0515072905,0.2793931961,-0.1505599469,-0.1328611523,0.068707265,0.0991909578,0.2852436006,0.1979034096,0.2263821661,0.4385739267,0.0644598529,-0.1653025299,0.0529506505,0.2032765597,-0.0528064743,0.1483865529,-0.3353100717,0.0793047324,0.2603348792,-0.2992096245,0.1275261194,-0.3775518239,-0.1403336674,0.1440867037,0.3369528651,-0.5138478875,-0.1326299012,-0.0149593158,-0.1334604025,0.1105089188,-0.3155808747,-0.2831487656,0.0130276047,-0.201353848,0.1626410633,-0.2732628286,-0.1256408542,-0.4015872478,0.0573596619,0.2376947105,-0.0031728048,-0.0217249822,0.0865617096,-0.023845451,0.0146014299,0.49329561,-0.2829744816,0.4370732903,-0.1266709417,0.3814791143,-0.2898307145,0.106090866,-0.2475935519,-0.3344739974,0.3203744292,0.246234566,0.0983890146,-0.0262450539,-0.2058806717,0.1195861474,-0.1224583089,-0.0811637938,0.1755108088,0.3042165041,-0.5918186307,-0.25063923,-0.1534152478,-0.2042932659,-0.3533948064,0.2209898382,0.135534808,0.3110541701,0.2072859555,-0.1218600199,0.0530254319,0.0434561335,0.3111526668,-0.3197958469,-0.0964614972,0.013731421,-0.1502657682,-0.5285490155,0.220506236,0.3835156858,-0.1217221916,0.1091838852,-0.1158387959,0.1251094043,-0.2140979767,0.2546712458,-0.2282879651,0.1662585884,0.1928665638,0.0734918788,0.0805564895,-0.1694658101,-0.0990942419,-0.3450900614,-0.2438754141,-0.264610678,0.154290542,0.3330400884,-0.131513983,0.2149935812,0.3426207602,0.2264254987,0.4461936355,-0.1276842952,0.4316768348,0.2818519771,-0.6437909603,0.1521578729,-0.1861309707,-0.053072609,0.3601593673,-0.0777115822,0.0294729602,-0.1574553996,0.0583065599,-0.3378964961,-0.2107926905,-0.0072617149,-0.2567565143,0.2782015204,-0.1479895264,0.0072967722,-0.2808846235,-0.0628952831,0.0054723453,0.568931222,0.1167742088,-0.061601568,-0.3760638535,0.0078706741,-0.4856024384,0.26964733,0.0624525361,0.6091942191,-0.1555267274,0.0315033421,0.1437783986,0.1085258052,0.9010464549,-0.4391025007,0.2763887942,0.2481005341,0.1166944727,-0.263177067,-0.211541608,-0.2682219446,0.067434594,-0.0131370779,0.3030076325,0.0118963942,-0.3937038779,0.0339972079,-0.1530512869,-0.0794086903,-0.4198121428,-0.253552407,-0.1815235615,-0.3771493137,-0.1041810736,0.2397942096,0.1634427458,-0.2563313544,0.2497168332,-0.0224103611,0.2882218957,0.1148120835,-0.0406747237,0.1016073227,0.3072133064,0.0786264092,-0.0809458047,0.1236321405,0.2123339176,0.5588169098,-0.026124591,-0.3647992313,-0.1107676178,-0.0270074625,0.4848950505,0.0845985115,-0.153303653,0.1014053077,-0.2211105675,-0.058644563,-0.3480154276,0.0739530697,0.2731055617,-0.0515423417,0.0595365576,-0.2157613635,0.3164066672,0.1723264158,-0.1943370253,-0.0005688795,0.1422991753,-0.2947748601,-0.0935064554,0.0510758013,0.5023060441,-0.2940065563,0.2797781527,0.1955997795,-0.1659105569,-0.071143426,0.213571772,-0.0846962035,-0.4017235041,-0.3005031049,-0.0134299649,-0.1052668542,-0.0054742121,0.1585410982,0.0849552229,0.0400238074,-0.0924142972,0.3835968375,0.0317604132,0.0596932992,-0.202075392,-0.2441049665,0.1019104496,0.151968509,-0.1361004859,-0.0235723518,-0.1311588436,0.0135436291,0.1534249038,0.0239182711,-0.3978714943,0.2855559587,-0.3279962242,0.0107293762,-0.1360235959,0.0947399661,0.5240969658,0.4402703941,-0.0064931316,0.1454547793,-0.2531163096,-0.2666570544,0.1540938467,-0.2917832136,-0.291580379,-0.0236652046,0.2547366619,0.0212811362,-0.3019715846,-0.1922039539,-0.0331776142,-0.2432691008,-0.0630597994,0.4675462842,0.1187870353,-0.507037282,-0.1682941318,-0.2320503443,-0.0584004521,0.0715503842,0.1030866802,0.1528616995,-0.1085439026,0.0619823374,-0.2745189965,-0.2699647248,0.0699824989,0.4720028639,-0.2750083506,0.1021838114,0.5445803404,0.4029547274,-0.1602605581,-0.2423876524,0.0846469402,0.0800300464,-0.2135888487,0.0007553757,0.312019676,0.0452671908,0.4206763506,0.0679824278,0.2228612602,-0.4020850658,0.1897781193,-0.4625521302,-0.110298954,-0.0878793746,-0.1858239919,0.3120950758,0.0934090316,-0.1442581862,0.0303128902,-0.2334226519,-0.29327479,0.1127857938,-0.0025985991,0.1187579483,0.4522465765,-0.229709357,-0.0856508985,-0.0219585318,0.0636571944,-0.1481769979,-0.1781420857,-0.2058385909,0.0039554443,0.1113952175,0.187864542,-0.3387500346,0.0858408138,-0.5403545499,-0.2182604074,-0.1620232761,0.2066287249,0.2949583232,-0.1138813123,0.4738371372,0.0064511998,-0.0148760919,0.0212341156,0.191027984,-0.1915093064,0.0859473795,-0.3789152205,0.4132497311,0.0258933119,-0.0854482129,-0.3049198687,0.2492909878,0.2421408892,0.1775051206,0.136467576,-0.2938653827,0.0452842601,0.1257751435,0.5298547745,0.0940537453,-0.3113827109,0.0305275526,0.1404470205,0.203754127,-0.3257121444,-0.1571353674,-0.2650839388,-0.1184605137,-0.0624468252,0.115416728,0.0112674851,-0.3411978185,-0.0892772675,-0.1154887825,0.2970600128,0.2271778584,-0.0939651132,0.5441298485,0.0009060174,-0.0297029205,0.3708809316,0.5498604774,0.0500368364,0.424381882,0.194929719,-0.1030110866,-0.1730057895,0.0265199021,-0.0050954581,-0.7224727273,0.1903757304,-0.0895253345,-0.0153279537,0.0070986184,-0.1846191287,0.4962287545,-0.1142299399,-0.0175317507,-0.3881719708,0.1512207389,-0.2028812766,-0.1655408293,-0.4238615632,0.0989473909,-0.0877483711,-0.0205381736,-0.1252162606,-0.2460019886,0.0037858104,0.228457734,-0.0325791202,-0.2135648876,0.349271208,0.2529511154,-0.1378227025,-0.3874364793,-0.0780463591,-0.0577866174,0.1203663275,0.1329323798,0.0222340655,0.6245731711,0.3318480551,-0.0558945425,-0.012673567,0.0129633136,-0.0314593799,-0.035633672,0.3866260946,-0.2138810754,0.3700492382,0.3404141068,0.0974966511,-0.058864668,-0.0613856651,0.0397349261,-0.1048275977,-0.0849886537,-0.0384141132,-0.3763240874,-0.2585601211,-0.1624862403,0.1023473442,-0.0869153216,-0.0223293044,0.6900111437,0.1503067762,0.341988951,-0.3378275931,0.0661960021,0.0295750573,0.4036607146,0.470158726,0.0760127977,-0.4037770629,-0.0799946636,-0.4968746305,0.0621396154,0.1211791858,-0.1317172647,0.1792279482,-0.136156112,-0.013322914,-0.1985217631,-0.0573430285,0.0297257993,0.0242389757,0.3217083514,-0.0086836405,-0.1077249572,0.2655580938,-0.0672235638,0.1426306814,-0.5638715625,0.0671390519,-0.2215017527,0.0575661026,-0.4719406068,0.0898197666,0.1238537803,0.1396856904,0.2431280315,0.0155660342,0.3771497309,-0.0043464405,0.1913803816,-0.5739651918,0.0678090081,-0.1726519465,0.1434074193,-0.1353252679,-0.0251373108,-0.2696663737,0.0728011355,-0.0983978063,0.1188255697,-0.0468874201,0.1527613699,-0.051272545,0.1282429248,-0.1833435446,-0.0416468568,0.2216355652,-0.1437097043,-0.1895443648,0.3802645802,-0.3233775496,-0.3723746538,0.4683177173,-0.2650411427,-0.2512885928,0.0131054819,0.3566480577,-0.1201093346,-0.1665672511,-0.1439184397,0.3625415266,0.1724920869,-0.004814927,-0.3380492628,0.4713961482,-0.2036845535,-0.1848139763,0.0364141874,0.4234890342,-0.0084362719,-0.2079798728,0.0341635793,-0.115126878]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2000","title":"Windows Permission Error (most recent version of datasets)","comments":"It seems to be working now, thanks a lot for the help, @SBrandeis !","body":"Hi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n","comment_length":14,"text":"Windows Permission Error (most recent version of datasets)\nHi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n\nIt seems to be working now, thanks a lot for the help, @SBrandeis !","embeddings":[-0.1825569272,0.1823850721,-0.0400314778,0.2590482235,0.0813782141,0.1336378604,0.4580342472,0.0122542866,0.1864634454,0.0741284713,-0.0634557605,-0.0112269307,-0.1087958515,0.1572928578,-0.0323617235,-0.0282511655,0.1453506798,0.0480685383,0.0319974497,0.1396457404,-0.4052302539,-0.0071208877,-0.2379359007,0.1073663011,-0.2231619507,0.278344065,-0.0157263745,0.3908289373,-0.0338031724,-0.2350183874,0.2101650983,0.3624857664,0.662093997,0.245353058,-0.0001126285,-0.048450049,0.0420433767,-0.0201318935,0.2068192661,0.2423872203,0.1341389418,-0.0451870672,0.0041253492,-0.1066693291,-0.1698371172,0.2000023574,-0.0654302686,-0.3538345098,0.1139979437,0.4263345599,0.1871957481,0.2474624962,-0.1806875318,-0.0106535321,0.5910394788,0.047598172,-0.1261592358,0.1927935481,0.2900241613,-0.4428050816,0.214141503,0.1609825641,-0.2268584669,0.0855402425,0.3818270862,-0.163264066,-0.1299571693,-0.2830388844,0.1584062725,0.00005305,0.727063477,-0.1692433953,-0.3009923995,-0.0341043584,0.1393089741,-0.2671689689,0.2275048792,0.3365985155,-0.0505269095,0.0753582716,-0.2347000837,-0.167150721,-0.4434008598,0.1358958036,0.1775047481,-0.2292983681,-0.084932968,0.2297086567,0.0906637236,-0.1301721931,0.3544327915,0.1294433624,0.0419623218,0.2517595291,-0.4671482146,0.3360512853,0.2092746347,0.2980217934,-0.0320270881,0.2779500782,-0.1361709833,-0.2106272429,0.2018981278,0.1250198483,0.1772745848,-0.0542738736,0.1519606858,0.2411280125,0.1375094652,-0.0225111879,0.0402428471,-0.117440857,-0.4900486767,-0.6061118841,0.1152087823,0.2454818487,0.3073368371,-0.0892671719,-0.3613361418,0.1164757684,0.1181355715,0.0680992082,0.2068185359,0.4097331166,0.2663407922,0.0119451405,0.3739327192,0.2714998126,-0.0628641695,-0.0258001015,-0.0862123594,0.0042463713,-0.2458985001,-0.0498456098,0.4257872105,-0.344258368,-0.1134994924,0.1173631623,0.0353439488,-0.1213498861,0.0494542569,-0.0187720023,0.1164598688,0.4388581216,0.2584182322,0.034997154,0.1450432539,-0.1851189882,-0.0282773599,0.2364417911,-0.132414341,-0.3404768407,-0.1883570999,0.1011299416,0.0170740597,0.0499140508,-0.218305707,-0.133036688,0.1228498816,-0.0184608568,0.1442888379,-0.1007024646,-0.2660652995,-0.3397949338,-0.1015146449,0.8314390182,-0.5272637606,0.1402395666,-0.1147450805,-0.3047668636,0.1072497219,0.3311969638,-0.0764998347,0.1113562062,-0.214482978,0.1516456157,-0.1357124448,-0.3680853248,-0.1788965762,0.3780059218,-0.017753955,-0.103679657,0.2437345237,-0.0390078723,0.0444860794,-0.2033561915,-0.0898012519,0.134128198,0.1384616196,0.2605628967,0.2249749005,0.0075680758,0.2322096527,0.3124230802,0.0553449914,-0.0627514571,0.2918676436,-0.2186992466,0.189552173,-0.3010883629,0.2245462686,0.4166680276,0.1158194914,0.3091973662,0.0332210027,-0.2322232127,-0.4455744326,0.2652016878,0.3054819703,-0.1239581481,0.0469031185,-0.2565841079,-0.1502389461,0.2109269798,-0.3883987963,0.1167708114,0.0899116099,0.297970891,-0.1774025708,-0.320898056,0.034541335,0.0563448928,-0.2444298863,-0.1111344919,-0.153433159,0.0384588875,-0.3066704273,0.0180423688,0.2049439102,-0.0148219531,0.415769726,-0.074380219,0.0026424972,0.3591325581,0.2803249657,-0.0367785916,0.036796052,0.0983095542,-0.2261272371,0.0771442056,0.1272577792,-0.0095818341,0.0916867405,-0.1025229618,-0.0255540945,0.1265098751,-0.2259499133,0.0041122376,-0.2879075408,0.1314649135,-0.0096176816,0.0248836838,-0.024606308,-0.2199265212,0.3593811691,0.5096359849,-0.3285354376,-0.0453583524,-0.102629602,-0.0515072905,0.2793931961,-0.1505599469,-0.1328611523,0.068707265,0.0991909578,0.2852436006,0.1979034096,0.2263821661,0.4385739267,0.0644598529,-0.1653025299,0.0529506505,0.2032765597,-0.0528064743,0.1483865529,-0.3353100717,0.0793047324,0.2603348792,-0.2992096245,0.1275261194,-0.3775518239,-0.1403336674,0.1440867037,0.3369528651,-0.5138478875,-0.1326299012,-0.0149593158,-0.1334604025,0.1105089188,-0.3155808747,-0.2831487656,0.0130276047,-0.201353848,0.1626410633,-0.2732628286,-0.1256408542,-0.4015872478,0.0573596619,0.2376947105,-0.0031728048,-0.0217249822,0.0865617096,-0.023845451,0.0146014299,0.49329561,-0.2829744816,0.4370732903,-0.1266709417,0.3814791143,-0.2898307145,0.106090866,-0.2475935519,-0.3344739974,0.3203744292,0.246234566,0.0983890146,-0.0262450539,-0.2058806717,0.1195861474,-0.1224583089,-0.0811637938,0.1755108088,0.3042165041,-0.5918186307,-0.25063923,-0.1534152478,-0.2042932659,-0.3533948064,0.2209898382,0.135534808,0.3110541701,0.2072859555,-0.1218600199,0.0530254319,0.0434561335,0.3111526668,-0.3197958469,-0.0964614972,0.013731421,-0.1502657682,-0.5285490155,0.220506236,0.3835156858,-0.1217221916,0.1091838852,-0.1158387959,0.1251094043,-0.2140979767,0.2546712458,-0.2282879651,0.1662585884,0.1928665638,0.0734918788,0.0805564895,-0.1694658101,-0.0990942419,-0.3450900614,-0.2438754141,-0.264610678,0.154290542,0.3330400884,-0.131513983,0.2149935812,0.3426207602,0.2264254987,0.4461936355,-0.1276842952,0.4316768348,0.2818519771,-0.6437909603,0.1521578729,-0.1861309707,-0.053072609,0.3601593673,-0.0777115822,0.0294729602,-0.1574553996,0.0583065599,-0.3378964961,-0.2107926905,-0.0072617149,-0.2567565143,0.2782015204,-0.1479895264,0.0072967722,-0.2808846235,-0.0628952831,0.0054723453,0.568931222,0.1167742088,-0.061601568,-0.3760638535,0.0078706741,-0.4856024384,0.26964733,0.0624525361,0.6091942191,-0.1555267274,0.0315033421,0.1437783986,0.1085258052,0.9010464549,-0.4391025007,0.2763887942,0.2481005341,0.1166944727,-0.263177067,-0.211541608,-0.2682219446,0.067434594,-0.0131370779,0.3030076325,0.0118963942,-0.3937038779,0.0339972079,-0.1530512869,-0.0794086903,-0.4198121428,-0.253552407,-0.1815235615,-0.3771493137,-0.1041810736,0.2397942096,0.1634427458,-0.2563313544,0.2497168332,-0.0224103611,0.2882218957,0.1148120835,-0.0406747237,0.1016073227,0.3072133064,0.0786264092,-0.0809458047,0.1236321405,0.2123339176,0.5588169098,-0.026124591,-0.3647992313,-0.1107676178,-0.0270074625,0.4848950505,0.0845985115,-0.153303653,0.1014053077,-0.2211105675,-0.058644563,-0.3480154276,0.0739530697,0.2731055617,-0.0515423417,0.0595365576,-0.2157613635,0.3164066672,0.1723264158,-0.1943370253,-0.0005688795,0.1422991753,-0.2947748601,-0.0935064554,0.0510758013,0.5023060441,-0.2940065563,0.2797781527,0.1955997795,-0.1659105569,-0.071143426,0.213571772,-0.0846962035,-0.4017235041,-0.3005031049,-0.0134299649,-0.1052668542,-0.0054742121,0.1585410982,0.0849552229,0.0400238074,-0.0924142972,0.3835968375,0.0317604132,0.0596932992,-0.202075392,-0.2441049665,0.1019104496,0.151968509,-0.1361004859,-0.0235723518,-0.1311588436,0.0135436291,0.1534249038,0.0239182711,-0.3978714943,0.2855559587,-0.3279962242,0.0107293762,-0.1360235959,0.0947399661,0.5240969658,0.4402703941,-0.0064931316,0.1454547793,-0.2531163096,-0.2666570544,0.1540938467,-0.2917832136,-0.291580379,-0.0236652046,0.2547366619,0.0212811362,-0.3019715846,-0.1922039539,-0.0331776142,-0.2432691008,-0.0630597994,0.4675462842,0.1187870353,-0.507037282,-0.1682941318,-0.2320503443,-0.0584004521,0.0715503842,0.1030866802,0.1528616995,-0.1085439026,0.0619823374,-0.2745189965,-0.2699647248,0.0699824989,0.4720028639,-0.2750083506,0.1021838114,0.5445803404,0.4029547274,-0.1602605581,-0.2423876524,0.0846469402,0.0800300464,-0.2135888487,0.0007553757,0.312019676,0.0452671908,0.4206763506,0.0679824278,0.2228612602,-0.4020850658,0.1897781193,-0.4625521302,-0.110298954,-0.0878793746,-0.1858239919,0.3120950758,0.0934090316,-0.1442581862,0.0303128902,-0.2334226519,-0.29327479,0.1127857938,-0.0025985991,0.1187579483,0.4522465765,-0.229709357,-0.0856508985,-0.0219585318,0.0636571944,-0.1481769979,-0.1781420857,-0.2058385909,0.0039554443,0.1113952175,0.187864542,-0.3387500346,0.0858408138,-0.5403545499,-0.2182604074,-0.1620232761,0.2066287249,0.2949583232,-0.1138813123,0.4738371372,0.0064511998,-0.0148760919,0.0212341156,0.191027984,-0.1915093064,0.0859473795,-0.3789152205,0.4132497311,0.0258933119,-0.0854482129,-0.3049198687,0.2492909878,0.2421408892,0.1775051206,0.136467576,-0.2938653827,0.0452842601,0.1257751435,0.5298547745,0.0940537453,-0.3113827109,0.0305275526,0.1404470205,0.203754127,-0.3257121444,-0.1571353674,-0.2650839388,-0.1184605137,-0.0624468252,0.115416728,0.0112674851,-0.3411978185,-0.0892772675,-0.1154887825,0.2970600128,0.2271778584,-0.0939651132,0.5441298485,0.0009060174,-0.0297029205,0.3708809316,0.5498604774,0.0500368364,0.424381882,0.194929719,-0.1030110866,-0.1730057895,0.0265199021,-0.0050954581,-0.7224727273,0.1903757304,-0.0895253345,-0.0153279537,0.0070986184,-0.1846191287,0.4962287545,-0.1142299399,-0.0175317507,-0.3881719708,0.1512207389,-0.2028812766,-0.1655408293,-0.4238615632,0.0989473909,-0.0877483711,-0.0205381736,-0.1252162606,-0.2460019886,0.0037858104,0.228457734,-0.0325791202,-0.2135648876,0.349271208,0.2529511154,-0.1378227025,-0.3874364793,-0.0780463591,-0.0577866174,0.1203663275,0.1329323798,0.0222340655,0.6245731711,0.3318480551,-0.0558945425,-0.012673567,0.0129633136,-0.0314593799,-0.035633672,0.3866260946,-0.2138810754,0.3700492382,0.3404141068,0.0974966511,-0.058864668,-0.0613856651,0.0397349261,-0.1048275977,-0.0849886537,-0.0384141132,-0.3763240874,-0.2585601211,-0.1624862403,0.1023473442,-0.0869153216,-0.0223293044,0.6900111437,0.1503067762,0.341988951,-0.3378275931,0.0661960021,0.0295750573,0.4036607146,0.470158726,0.0760127977,-0.4037770629,-0.0799946636,-0.4968746305,0.0621396154,0.1211791858,-0.1317172647,0.1792279482,-0.136156112,-0.013322914,-0.1985217631,-0.0573430285,0.0297257993,0.0242389757,0.3217083514,-0.0086836405,-0.1077249572,0.2655580938,-0.0672235638,0.1426306814,-0.5638715625,0.0671390519,-0.2215017527,0.0575661026,-0.4719406068,0.0898197666,0.1238537803,0.1396856904,0.2431280315,0.0155660342,0.3771497309,-0.0043464405,0.1913803816,-0.5739651918,0.0678090081,-0.1726519465,0.1434074193,-0.1353252679,-0.0251373108,-0.2696663737,0.0728011355,-0.0983978063,0.1188255697,-0.0468874201,0.1527613699,-0.051272545,0.1282429248,-0.1833435446,-0.0416468568,0.2216355652,-0.1437097043,-0.1895443648,0.3802645802,-0.3233775496,-0.3723746538,0.4683177173,-0.2650411427,-0.2512885928,0.0131054819,0.3566480577,-0.1201093346,-0.1665672511,-0.1439184397,0.3625415266,0.1724920869,-0.004814927,-0.3380492628,0.4713961482,-0.2036845535,-0.1848139763,0.0364141874,0.4234890342,-0.0084362719,-0.2079798728,0.0341635793,-0.115126878]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2000","title":"Windows Permission Error (most recent version of datasets)","comments":"Glad to hear it!\r\nI'm closing the issue","body":"Hi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n","comment_length":8,"text":"Windows Permission Error (most recent version of datasets)\nHi everyone,\r\nCan anyone help me with why the dataset loading script below raises a Windows Permission Error? I stuck quite closely to https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py , only I want to load the data from three local three-column tsv-files (id\\ttokens\\tpos_tags\\n). I am using the most recent version of datasets. Thank you in advance!\r\nLuisa\r\n\r\nMy script:\r\n```\r\nimport datasets\r\nimport csv\r\n\r\nlogger = datasets.logging.get_logger(__name__)\r\n\r\n\r\nclass SampleConfig(datasets.BuilderConfig):\r\n\r\n def __init__(self, **kwargs):\r\n super(SampleConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Sample(datasets.GeneratorBasedBuilder):\r\n BUILDER_CONFIGS = [\r\n SampleConfig(name=\"conll2003\", version=datasets.Version(\"1.0.0\"), description=\"Conll2003 dataset\"),\r\n ]\r\n\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=\"Dataset with words and their POS-Tags\",\r\n features=datasets.Features(\r\n {\r\n \"id\": datasets.Value(\"string\"),\r\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\r\n \"pos_tags\": datasets.Sequence(\r\n datasets.features.ClassLabel(\r\n names=[\r\n \"''\",\r\n \",\",\r\n \"-LRB-\",\r\n \"-RRB-\",\r\n \".\",\r\n \":\",\r\n \"CC\",\r\n \"CD\",\r\n \"DT\",\r\n \"EX\",\r\n \"FW\",\r\n \"HYPH\",\r\n \"IN\",\r\n \"JJ\",\r\n \"JJR\",\r\n \"JJS\",\r\n \"MD\",\r\n \"NN\",\r\n \"NNP\",\r\n \"NNPS\",\r\n \"NNS\",\r\n \"PDT\",\r\n \"POS\",\r\n \"PRP\",\r\n \"PRP$\",\r\n \"RB\",\r\n \"RBR\",\r\n \"RBS\",\r\n \"RP\",\r\n \"TO\",\r\n \"UH\",\r\n \"VB\",\r\n \"VBD\",\r\n \"VBG\",\r\n \"VBN\",\r\n \"VBP\",\r\n \"VBZ\",\r\n \"WDT\",\r\n \"WP\",\r\n \"WRB\",\r\n \"``\"\r\n ]\r\n )\r\n ),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/catalog.ldc.upenn.edu\/LDC2011T03\",\r\n citation=\"Weischedel, Ralph, et al. OntoNotes Release 4.0 LDC2011T03. Web Download. Philadelphia: Linguistic Data Consortium, 2011.\",\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n loaded_files = dl_manager.download_and_extract(self.config.data_files)\r\n return [\r\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\"filepath\": loaded_files[\"train\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={\"filepath\": loaded_files[\"test\"]}),\r\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\"filepath\": loaded_files[\"val\"]})\r\n ]\r\n\r\n def _generate_examples(self, filepath):\r\n logger.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"cp1252\") as f:\r\n data = csv.reader(f, delimiter=\"\\t\")\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n for id_, line in enumerate(data):\r\n #print(line)\r\n if len(line) == 1:\r\n if tokens:\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n ids = list()\r\n tokens = list()\r\n pos_tags = list()\r\n else:\r\n ids.append(line[0])\r\n tokens.append(line[1])\r\n pos_tags.append(line[2])\r\n # last example\r\n yield id_, {\"id\": ids, \"tokens\": tokens, \"pos_tags\": pos_tags}\r\n\r\n\r\ndef main():\r\n dataset = datasets.load_dataset(\r\n \"data_loading.py\", data_files={\r\n \"train\": \"train.tsv\",\r\n \"test\": \"test.tsv\",\r\n \"val\": \"val.tsv\"\r\n }\r\n )\r\n\r\n #print(dataset)\r\n\r\nif __name__==\"__main__\":\r\n main()\r\n```\r\n\nGlad to hear it!\r\nI'm closing the issue","embeddings":[-0.1825569272,0.1823850721,-0.0400314778,0.2590482235,0.0813782141,0.1336378604,0.4580342472,0.0122542866,0.1864634454,0.0741284713,-0.0634557605,-0.0112269307,-0.1087958515,0.1572928578,-0.0323617235,-0.0282511655,0.1453506798,0.0480685383,0.0319974497,0.1396457404,-0.4052302539,-0.0071208877,-0.2379359007,0.1073663011,-0.2231619507,0.278344065,-0.0157263745,0.3908289373,-0.0338031724,-0.2350183874,0.2101650983,0.3624857664,0.662093997,0.245353058,-0.0001126285,-0.048450049,0.0420433767,-0.0201318935,0.2068192661,0.2423872203,0.1341389418,-0.0451870672,0.0041253492,-0.1066693291,-0.1698371172,0.2000023574,-0.0654302686,-0.3538345098,0.1139979437,0.4263345599,0.1871957481,0.2474624962,-0.1806875318,-0.0106535321,0.5910394788,0.047598172,-0.1261592358,0.1927935481,0.2900241613,-0.4428050816,0.214141503,0.1609825641,-0.2268584669,0.0855402425,0.3818270862,-0.163264066,-0.1299571693,-0.2830388844,0.1584062725,0.00005305,0.727063477,-0.1692433953,-0.3009923995,-0.0341043584,0.1393089741,-0.2671689689,0.2275048792,0.3365985155,-0.0505269095,0.0753582716,-0.2347000837,-0.167150721,-0.4434008598,0.1358958036,0.1775047481,-0.2292983681,-0.084932968,0.2297086567,0.0906637236,-0.1301721931,0.3544327915,0.1294433624,0.0419623218,0.2517595291,-0.4671482146,0.3360512853,0.2092746347,0.2980217934,-0.0320270881,0.2779500782,-0.1361709833,-0.2106272429,0.2018981278,0.1250198483,0.1772745848,-0.0542738736,0.1519606858,0.2411280125,0.1375094652,-0.0225111879,0.0402428471,-0.117440857,-0.4900486767,-0.6061118841,0.1152087823,0.2454818487,0.3073368371,-0.0892671719,-0.3613361418,0.1164757684,0.1181355715,0.0680992082,0.2068185359,0.4097331166,0.2663407922,0.0119451405,0.3739327192,0.2714998126,-0.0628641695,-0.0258001015,-0.0862123594,0.0042463713,-0.2458985001,-0.0498456098,0.4257872105,-0.344258368,-0.1134994924,0.1173631623,0.0353439488,-0.1213498861,0.0494542569,-0.0187720023,0.1164598688,0.4388581216,0.2584182322,0.034997154,0.1450432539,-0.1851189882,-0.0282773599,0.2364417911,-0.132414341,-0.3404768407,-0.1883570999,0.1011299416,0.0170740597,0.0499140508,-0.218305707,-0.133036688,0.1228498816,-0.0184608568,0.1442888379,-0.1007024646,-0.2660652995,-0.3397949338,-0.1015146449,0.8314390182,-0.5272637606,0.1402395666,-0.1147450805,-0.3047668636,0.1072497219,0.3311969638,-0.0764998347,0.1113562062,-0.214482978,0.1516456157,-0.1357124448,-0.3680853248,-0.1788965762,0.3780059218,-0.017753955,-0.103679657,0.2437345237,-0.0390078723,0.0444860794,-0.2033561915,-0.0898012519,0.134128198,0.1384616196,0.2605628967,0.2249749005,0.0075680758,0.2322096527,0.3124230802,0.0553449914,-0.0627514571,0.2918676436,-0.2186992466,0.189552173,-0.3010883629,0.2245462686,0.4166680276,0.1158194914,0.3091973662,0.0332210027,-0.2322232127,-0.4455744326,0.2652016878,0.3054819703,-0.1239581481,0.0469031185,-0.2565841079,-0.1502389461,0.2109269798,-0.3883987963,0.1167708114,0.0899116099,0.297970891,-0.1774025708,-0.320898056,0.034541335,0.0563448928,-0.2444298863,-0.1111344919,-0.153433159,0.0384588875,-0.3066704273,0.0180423688,0.2049439102,-0.0148219531,0.415769726,-0.074380219,0.0026424972,0.3591325581,0.2803249657,-0.0367785916,0.036796052,0.0983095542,-0.2261272371,0.0771442056,0.1272577792,-0.0095818341,0.0916867405,-0.1025229618,-0.0255540945,0.1265098751,-0.2259499133,0.0041122376,-0.2879075408,0.1314649135,-0.0096176816,0.0248836838,-0.024606308,-0.2199265212,0.3593811691,0.5096359849,-0.3285354376,-0.0453583524,-0.102629602,-0.0515072905,0.2793931961,-0.1505599469,-0.1328611523,0.068707265,0.0991909578,0.2852436006,0.1979034096,0.2263821661,0.4385739267,0.0644598529,-0.1653025299,0.0529506505,0.2032765597,-0.0528064743,0.1483865529,-0.3353100717,0.0793047324,0.2603348792,-0.2992096245,0.1275261194,-0.3775518239,-0.1403336674,0.1440867037,0.3369528651,-0.5138478875,-0.1326299012,-0.0149593158,-0.1334604025,0.1105089188,-0.3155808747,-0.2831487656,0.0130276047,-0.201353848,0.1626410633,-0.2732628286,-0.1256408542,-0.4015872478,0.0573596619,0.2376947105,-0.0031728048,-0.0217249822,0.0865617096,-0.023845451,0.0146014299,0.49329561,-0.2829744816,0.4370732903,-0.1266709417,0.3814791143,-0.2898307145,0.106090866,-0.2475935519,-0.3344739974,0.3203744292,0.246234566,0.0983890146,-0.0262450539,-0.2058806717,0.1195861474,-0.1224583089,-0.0811637938,0.1755108088,0.3042165041,-0.5918186307,-0.25063923,-0.1534152478,-0.2042932659,-0.3533948064,0.2209898382,0.135534808,0.3110541701,0.2072859555,-0.1218600199,0.0530254319,0.0434561335,0.3111526668,-0.3197958469,-0.0964614972,0.013731421,-0.1502657682,-0.5285490155,0.220506236,0.3835156858,-0.1217221916,0.1091838852,-0.1158387959,0.1251094043,-0.2140979767,0.2546712458,-0.2282879651,0.1662585884,0.1928665638,0.0734918788,0.0805564895,-0.1694658101,-0.0990942419,-0.3450900614,-0.2438754141,-0.264610678,0.154290542,0.3330400884,-0.131513983,0.2149935812,0.3426207602,0.2264254987,0.4461936355,-0.1276842952,0.4316768348,0.2818519771,-0.6437909603,0.1521578729,-0.1861309707,-0.053072609,0.3601593673,-0.0777115822,0.0294729602,-0.1574553996,0.0583065599,-0.3378964961,-0.2107926905,-0.0072617149,-0.2567565143,0.2782015204,-0.1479895264,0.0072967722,-0.2808846235,-0.0628952831,0.0054723453,0.568931222,0.1167742088,-0.061601568,-0.3760638535,0.0078706741,-0.4856024384,0.26964733,0.0624525361,0.6091942191,-0.1555267274,0.0315033421,0.1437783986,0.1085258052,0.9010464549,-0.4391025007,0.2763887942,0.2481005341,0.1166944727,-0.263177067,-0.211541608,-0.2682219446,0.067434594,-0.0131370779,0.3030076325,0.0118963942,-0.3937038779,0.0339972079,-0.1530512869,-0.0794086903,-0.4198121428,-0.253552407,-0.1815235615,-0.3771493137,-0.1041810736,0.2397942096,0.1634427458,-0.2563313544,0.2497168332,-0.0224103611,0.2882218957,0.1148120835,-0.0406747237,0.1016073227,0.3072133064,0.0786264092,-0.0809458047,0.1236321405,0.2123339176,0.5588169098,-0.026124591,-0.3647992313,-0.1107676178,-0.0270074625,0.4848950505,0.0845985115,-0.153303653,0.1014053077,-0.2211105675,-0.058644563,-0.3480154276,0.0739530697,0.2731055617,-0.0515423417,0.0595365576,-0.2157613635,0.3164066672,0.1723264158,-0.1943370253,-0.0005688795,0.1422991753,-0.2947748601,-0.0935064554,0.0510758013,0.5023060441,-0.2940065563,0.2797781527,0.1955997795,-0.1659105569,-0.071143426,0.213571772,-0.0846962035,-0.4017235041,-0.3005031049,-0.0134299649,-0.1052668542,-0.0054742121,0.1585410982,0.0849552229,0.0400238074,-0.0924142972,0.3835968375,0.0317604132,0.0596932992,-0.202075392,-0.2441049665,0.1019104496,0.151968509,-0.1361004859,-0.0235723518,-0.1311588436,0.0135436291,0.1534249038,0.0239182711,-0.3978714943,0.2855559587,-0.3279962242,0.0107293762,-0.1360235959,0.0947399661,0.5240969658,0.4402703941,-0.0064931316,0.1454547793,-0.2531163096,-0.2666570544,0.1540938467,-0.2917832136,-0.291580379,-0.0236652046,0.2547366619,0.0212811362,-0.3019715846,-0.1922039539,-0.0331776142,-0.2432691008,-0.0630597994,0.4675462842,0.1187870353,-0.507037282,-0.1682941318,-0.2320503443,-0.0584004521,0.0715503842,0.1030866802,0.1528616995,-0.1085439026,0.0619823374,-0.2745189965,-0.2699647248,0.0699824989,0.4720028639,-0.2750083506,0.1021838114,0.5445803404,0.4029547274,-0.1602605581,-0.2423876524,0.0846469402,0.0800300464,-0.2135888487,0.0007553757,0.312019676,0.0452671908,0.4206763506,0.0679824278,0.2228612602,-0.4020850658,0.1897781193,-0.4625521302,-0.110298954,-0.0878793746,-0.1858239919,0.3120950758,0.0934090316,-0.1442581862,0.0303128902,-0.2334226519,-0.29327479,0.1127857938,-0.0025985991,0.1187579483,0.4522465765,-0.229709357,-0.0856508985,-0.0219585318,0.0636571944,-0.1481769979,-0.1781420857,-0.2058385909,0.0039554443,0.1113952175,0.187864542,-0.3387500346,0.0858408138,-0.5403545499,-0.2182604074,-0.1620232761,0.2066287249,0.2949583232,-0.1138813123,0.4738371372,0.0064511998,-0.0148760919,0.0212341156,0.191027984,-0.1915093064,0.0859473795,-0.3789152205,0.4132497311,0.0258933119,-0.0854482129,-0.3049198687,0.2492909878,0.2421408892,0.1775051206,0.136467576,-0.2938653827,0.0452842601,0.1257751435,0.5298547745,0.0940537453,-0.3113827109,0.0305275526,0.1404470205,0.203754127,-0.3257121444,-0.1571353674,-0.2650839388,-0.1184605137,-0.0624468252,0.115416728,0.0112674851,-0.3411978185,-0.0892772675,-0.1154887825,0.2970600128,0.2271778584,-0.0939651132,0.5441298485,0.0009060174,-0.0297029205,0.3708809316,0.5498604774,0.0500368364,0.424381882,0.194929719,-0.1030110866,-0.1730057895,0.0265199021,-0.0050954581,-0.7224727273,0.1903757304,-0.0895253345,-0.0153279537,0.0070986184,-0.1846191287,0.4962287545,-0.1142299399,-0.0175317507,-0.3881719708,0.1512207389,-0.2028812766,-0.1655408293,-0.4238615632,0.0989473909,-0.0877483711,-0.0205381736,-0.1252162606,-0.2460019886,0.0037858104,0.228457734,-0.0325791202,-0.2135648876,0.349271208,0.2529511154,-0.1378227025,-0.3874364793,-0.0780463591,-0.0577866174,0.1203663275,0.1329323798,0.0222340655,0.6245731711,0.3318480551,-0.0558945425,-0.012673567,0.0129633136,-0.0314593799,-0.035633672,0.3866260946,-0.2138810754,0.3700492382,0.3404141068,0.0974966511,-0.058864668,-0.0613856651,0.0397349261,-0.1048275977,-0.0849886537,-0.0384141132,-0.3763240874,-0.2585601211,-0.1624862403,0.1023473442,-0.0869153216,-0.0223293044,0.6900111437,0.1503067762,0.341988951,-0.3378275931,0.0661960021,0.0295750573,0.4036607146,0.470158726,0.0760127977,-0.4037770629,-0.0799946636,-0.4968746305,0.0621396154,0.1211791858,-0.1317172647,0.1792279482,-0.136156112,-0.013322914,-0.1985217631,-0.0573430285,0.0297257993,0.0242389757,0.3217083514,-0.0086836405,-0.1077249572,0.2655580938,-0.0672235638,0.1426306814,-0.5638715625,0.0671390519,-0.2215017527,0.0575661026,-0.4719406068,0.0898197666,0.1238537803,0.1396856904,0.2431280315,0.0155660342,0.3771497309,-0.0043464405,0.1913803816,-0.5739651918,0.0678090081,-0.1726519465,0.1434074193,-0.1353252679,-0.0251373108,-0.2696663737,0.0728011355,-0.0983978063,0.1188255697,-0.0468874201,0.1527613699,-0.051272545,0.1282429248,-0.1833435446,-0.0416468568,0.2216355652,-0.1437097043,-0.1895443648,0.3802645802,-0.3233775496,-0.3723746538,0.4683177173,-0.2650411427,-0.2512885928,0.0131054819,0.3566480577,-0.1201093346,-0.1665672511,-0.1439184397,0.3625415266,0.1724920869,-0.004814927,-0.3380492628,0.4713961482,-0.2036845535,-0.1848139763,0.0364141874,0.4234890342,-0.0084362719,-0.2079798728,0.0341635793,-0.115126878]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1996","title":"Error when exploring `arabic_speech_corpus`","comments":"Thanks for reporting! We'll fix that as soon as possible","body":"Navigate to https:\/\/huggingface.co\/datasets\/viewer\/?dataset=arabic_speech_corpus\r\n\r\nError:\r\n```\r\nImportError: To be able to use this dataset, you need to install the following dependencies['soundfile'] using 'pip install soundfile' for instance'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/script_runner.py\", line 332, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 233, in \r\n configs = get_confs(option)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 604, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 588, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 145, in get_confs\r\n module_path = nlp.load.prepare_module(path, dataset=True\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 342, in prepare_module\r\n f\"To be able to use this {module_type}, you need to install the following dependencies\"\r\n```","comment_length":10,"text":"Error when exploring `arabic_speech_corpus`\nNavigate to https:\/\/huggingface.co\/datasets\/viewer\/?dataset=arabic_speech_corpus\r\n\r\nError:\r\n```\r\nImportError: To be able to use this dataset, you need to install the following dependencies['soundfile'] using 'pip install soundfile' for instance'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/script_runner.py\", line 332, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 233, in \r\n configs = get_confs(option)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 604, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 588, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 145, in get_confs\r\n module_path = nlp.load.prepare_module(path, dataset=True\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 342, in prepare_module\r\n f\"To be able to use this {module_type}, you need to install the following dependencies\"\r\n```\nThanks for reporting! We'll fix that as soon as possible","embeddings":[-0.2647770643,-0.1406262517,-0.1047626808,0.1659750342,0.0484970398,0.0187585838,0.0180519335,0.3557018042,-0.1736771315,0.1097977534,-0.2727349699,0.0144260935,-0.1204698905,0.069011502,0.0556852929,-0.4141934514,0.0205909759,0.1821699291,0.0634328946,-0.0388768539,-0.0982397497,0.3467887044,-0.367123723,0.0249966886,0.0222650692,-0.2762486935,-0.0160324387,-0.1413342655,-0.3296339512,-0.4560682774,-0.0061979471,-0.1350026876,0.1645259708,0.4052603245,-0.0001078306,-0.0901877433,0.3818080425,-0.1193058416,-0.2389035374,-0.3968277872,0.1958684325,-0.1154383719,-0.1058383286,-0.1073874608,-0.1151578277,-0.2009516805,0.0327067077,-0.2125924379,0.297137022,0.4106985629,0.3067451119,0.1321010441,0.1731583923,-0.0706191212,0.1614953578,-0.0349468477,-0.1427047849,0.1528975219,0.2208325416,0.0547837727,-0.1303085685,0.4795190692,-0.3637337089,-0.0123670138,-0.0388261601,-0.1991412044,0.0189395603,-0.4695085585,0.2406820655,-0.1063255072,0.4441643357,-0.2392779887,-0.1333870888,-0.1355413347,0.1991636157,-0.2646439672,0.2066099644,0.3633556962,-0.328838855,0.0860720724,0.0829828605,-0.2417379171,-0.0434331633,0.2043682635,0.0927154347,0.3236615062,-0.2734171152,-0.1509078443,0.2811872363,-0.2527626157,-0.1634566933,0.1047044471,-0.1116545051,0.2632725537,-0.0792132467,0.0336183272,0.16880925,0.1220468208,-0.0657668039,-0.0330570191,-0.0079581235,0.269931078,-0.106539607,0.2289962173,-0.0424607918,0.164982751,0.2403814048,0.0721287355,0.1728423238,0.3926004469,0.0023054071,0.0084934467,-0.1289272606,-0.2238308042,-0.4769639969,-0.0499766469,0.380461216,-0.2624093294,-0.0854418874,-0.0040904642,-0.0821079388,-0.2446199954,0.049963858,0.4604204893,-0.1244553477,0.0236061495,0.110829547,0.33172521,-0.2499329448,-0.1933601946,0.0166869108,0.2607304752,-0.1447466463,0.1265717298,0.2937808931,-0.4275924861,0.4628612995,-0.0891423076,0.1675929129,0.1987249404,-0.0288703796,-0.1015613377,-0.066005148,0.2685878277,0.0834726766,0.1761125773,0.1545976996,-0.0276912656,-0.1763593405,0.0319922008,0.0031905237,-0.2118691802,-0.0669172481,0.2590911686,-0.165569976,-0.0797726065,-0.1784221828,0.251837194,0.1563615203,-0.4835758209,-0.0865484998,0.0853819847,-0.4418664873,-0.0467355996,0.0436123721,0.2249166071,-0.0117174685,-0.1861529052,-0.0910202414,0.1004466787,0.1615720838,0.1810456067,-0.1064435914,0.0148693556,-0.2650836706,0.3452013135,0.407959342,-0.5480015874,-0.31644997,0.1258461624,0.1269969642,0.1261341274,0.3108423054,-0.2783137858,0.3713175356,-0.1717862785,0.1240484789,0.3689344823,0.1944082379,-0.039817106,-0.2393045872,-0.0864515826,0.0784153864,0.1081740707,-0.0881460235,0.1216119379,0.0460062288,-0.0683435351,0.2124584764,0.0432519987,0.0412101336,0.1832434386,0.0854034573,0.0268232375,0.0125402045,-0.2984178066,0.2535354495,-0.0696656406,-0.3383366764,0.4695850015,-0.3493668735,-0.0981953219,-0.2635187507,-0.160210982,-0.2829509377,-0.1377125531,0.249029085,0.1869929582,0.1921047121,0.3711824417,-0.1921373159,-0.0765714869,0.2046335191,-0.0511340238,-0.5637167096,0.0963110402,-0.2870705724,-0.143820703,0.3668653667,0.2116125524,0.181635201,-0.1025372744,-0.0822386444,0.1406834275,-0.1764976829,0.3381091058,-0.2775963247,0.0826022923,0.1328433454,-0.3464300334,0.3901206851,0.1550879478,0.1762218028,-0.0968008488,0.3981122971,0.150066033,0.2168174833,0.1147204787,0.2016976923,0.0723261833,0.4555347264,0.1035146341,-0.1288552731,-0.2171871215,0.421195358,-0.1869081259,0.3490540087,-0.0706448331,-0.3176205754,-0.0518032573,0.4970329404,-0.1215777323,0.1572321355,0.1135795191,-0.3207268119,-0.1885945201,0.1696972847,0.0762472749,0.3288859427,0.151788041,0.062957935,0.1460950375,0.0834065601,-0.3502940238,0.2956216037,-0.0278114155,0.1115899235,0.2987816632,0.0995068401,-0.0112167941,-0.5312961936,-0.0463962294,-0.0693681613,0.2184875458,-0.0773013607,-0.0438785516,-0.2064111233,-0.5863729119,-0.100257732,-0.3250047565,-0.2945673764,-0.215709269,0.0707623586,0.2373447716,0.0789356902,0.2986311615,0.2304368913,0.0680895746,-0.0829086304,-0.0664284229,-0.223331809,-0.0887668058,-0.1240076795,0.1581112444,0.2881472707,0.2647997141,0.214910835,-0.1713727415,0.0205936972,-0.1771332324,-0.136429742,0.1822233647,-0.1996455491,0.0669105127,-0.0087470282,0.2607525587,0.0287767183,-0.3666214049,0.2020437568,-0.0196201596,-0.1017676815,0.2357333601,-0.0982104167,-0.0096116243,-0.112823993,-0.3187180161,-0.3328931332,-0.5846563578,-0.0727261603,0.0810314119,0.2357703,0.5017256737,-0.0085092308,0.2079911381,-0.1317996383,0.4584853053,-0.200250566,-0.2499569803,0.3932137489,-0.4018228054,-0.3797583282,0.0852371976,0.0656598434,0.5130385756,-0.2382516414,-0.4224380851,-0.0073340726,-0.188276276,-0.1218123958,0.0372602269,0.0947230458,0.3159774244,0.1045533195,-0.0879580528,-0.0671045929,-0.1385822296,-0.2521793246,0.1087351739,0.1702072024,0.0976112261,0.3672985137,0.0655111969,0.4590443671,0.0846558064,0.084049426,0.4737160206,0.2001265883,0.3686895967,-0.2308712751,-0.4915699065,0.003191882,-0.0070181158,0.0352572389,0.2876551151,0.0085073197,-0.2958385944,-0.4569379985,-0.3464999199,-0.2373202294,-0.1974304467,-0.020490801,-0.2049037516,0.3031080067,0.1181852967,0.0037893537,0.0306787845,-0.04006676,0.2480336577,0.1280059665,-0.0142483627,0.0211184472,-0.3824164867,-0.2707668841,-0.6098897457,0.4273880422,0.0417089537,0.2269166112,-0.2018888295,0.0482090488,0.0909230411,-0.0193464942,0.3386313617,-0.1662507802,0.0197947305,0.3010656238,0.0586272664,-0.3092999458,0.0308697242,-0.2920117378,0.1320397854,-0.0340613574,0.0862411186,-0.1726428866,-0.0221966226,0.1698696017,0.2494883835,0.1013160124,-0.0130660394,-0.356030345,-0.4239629805,-0.42540133,-0.1127539352,0.1030720919,0.2711144686,0.2606952488,0.2043023854,0.2808110714,-0.0075348127,0.1951279789,0.0645831674,0.0405634269,0.2604700923,-0.0356786624,0.1683723032,0.0480666049,-0.1511381865,0.5268608332,0.0988171101,-0.2979227602,-0.1332915127,0.0798559114,0.1206755191,0.2356595546,-0.0304412097,0.2901094258,0.084829241,-0.0035442088,-0.0811002403,-0.0784568414,0.3486787677,0.2332342118,-0.2319896221,-0.4710700214,0.1666249633,-0.1449351907,0.0074472888,0.1506209821,0.0120096626,-0.2120891213,0.6520795822,-0.3518099785,0.9770261645,0.1397426575,0.1687323153,0.4786772728,0.2956974208,0.1800136715,-0.0318741053,0.3160134852,0.0561900623,-0.1314911246,-0.0648047253,-0.0204503164,0.3903802931,-0.1523286104,-0.4528801143,0.2550067008,0.0851945803,-0.2312515825,-0.1556728333,0.1983759999,-0.1999443918,-0.3213458657,-0.7144451737,0.2300630659,-0.1815147996,0.3918433189,0.0168757215,-0.0991478637,0.0263581444,-0.3732077777,-0.3035427332,0.2110049129,-0.1927026808,0.0886272117,-0.183223173,-0.0042378013,0.3131559789,0.2994610369,0.0670882314,0.2481057644,-0.2020278126,0.0786114559,-0.1160669103,-0.4727212489,0.0520700589,0.1944475919,0.2155802101,-0.308535248,-0.2547793984,0.3766150177,0.0574005358,-0.0176377781,-0.0065145679,0.001260862,0.0630848333,-0.1007633284,-0.5495936275,0.0595003068,-0.1587446034,-0.095028609,0.1858499199,-0.0107840998,-0.4529657364,0.0494737402,0.0034093386,-0.0933972299,-0.0336643308,0.4373268485,0.186305359,-0.2016571462,0.4039295614,0.3517253697,-0.2276720405,-0.2530688941,-0.1420620978,-0.4000782967,-0.2833403647,-0.0925976336,0.3696297705,-0.1416189075,-0.1536395997,0.299049288,-0.0407505594,0.0117829265,-0.0938842297,-0.3271164,-0.1523017883,0.3797287941,0.0181775484,-0.1856630594,0.1090341583,-0.0554992817,0.1577365994,0.1567482054,-0.3849673271,0.1104392037,-0.147869736,0.238805607,0.2805691957,-0.0871180519,0.1437693238,0.123741135,0.2125836462,0.1233474091,-0.2566297352,-0.27528283,0.1530153155,0.0655044466,0.1286610365,-0.0747930035,0.1098912805,-0.0975863189,-0.1732886136,-0.1007242575,-0.0573276021,0.3307023048,-0.0031985936,0.0318900496,-0.247336939,-0.0387943722,0.2210645378,0.0337535143,-0.1974915266,-0.0231384728,0.1314007789,-0.0580348894,-0.0582876578,-0.1034732237,-0.0600603521,0.046545215,0.0156406108,-0.0560309291,0.2912419438,-0.1382356882,0.1967399567,0.1888625175,0.2428280711,0.1611290723,-0.2117114365,-0.0716245174,0.3995046616,0.220940277,-0.5028002858,-0.0699506849,0.0905669555,-0.1755448282,0.1682738215,0.1383533776,0.1834814996,-0.0120678823,0.2761122584,0.0699292868,0.376491338,-0.0778701454,0.2685905695,-0.0508871786,0.0489611216,-0.1533808559,0.0774425417,0.1724157333,0.1495759338,0.3588538766,-0.4883660078,0.2701590657,0.1536500901,0.038753029,0.1131588221,-0.2871171534,0.1023574397,0.4357617795,0.1245350167,0.1146021038,-0.0774186403,0.2609979808,-0.1145726815,-0.0807459578,0.047166571,0.2666114867,0.026778426,0.1247860938,-0.0106659094,-0.1316577643,-0.113797985,-0.1242272854,0.1807096004,-0.0539705306,-0.0056681377,-0.0515231006,-0.1872655898,-0.1176184863,0.0452277213,-0.0089906184,0.0291774664,-0.1250736862,0.4329302013,0.0503127426,0.0450017005,0.0538463555,0.4644492567,0.6024806499,0.266990304,-0.0246226378,0.3057027161,-0.3444402218,-0.0817290694,0.0231776815,0.2484193593,0.31757918,0.3270504475,0.2173752785,0.1583997905,-0.2388501912,-0.0385172889,0.3172759414,0.0697454587,-0.1039826572,0.0280899182,-0.4795947969,-0.0227692388,-0.3527711034,-0.0310374033,-0.8419408798,0.0702846348,0.461004287,-0.168916449,0.0305586644,-0.2793902755,0.1048266664,-0.0557179078,0.3970556855,0.424618125,0.267680198,0.0152048143,-0.3114179671,-0.6501590014,0.1693025678,0.095505923,0.0882557482,-0.2490250617,-0.0983227491,0.3455832899,0.3441427648,-0.0920214579,0.1048038378,-0.2152733207,-0.2301523834,-0.1844753325,0.1275899559,-0.0389169492,0.2720361054,-0.1688975543,-0.2960501909,-0.0978706256,-0.4150532484,0.1793086231,-0.0858662128,-0.0319454074,-0.3042278886,-0.0444010831,0.0609880872,0.1916637272,0.3226231933,-0.2406675369,-0.0741653517,-0.0638248622,-0.2165373713,-0.0154591901,0.2548857033,0.1438571215,0.519012332,-0.1811359376,0.136189431,-0.2725774646,0.4580572248,0.0545805916,0.0224092212,-0.1779645383,0.2791248858,-0.0281327441,0.1130800396,0.0138159469,0.0169661622,-0.0279966276,-0.0033440629,-0.4646090567,-0.2608162165,0.6433056593,-0.5007050633,-0.3400933743,0.1120998263,0.2912752032,0.1196996197,-0.2733761668,-0.6002378464,0.118084617,0.0994462818,0.1283966303,-0.0785420761,0.2289501429,-0.3475034833,-0.2081159353,-0.0622860678,0.1466837078,0.3043296933,-0.1502313167,0.0458045937,-0.247022301]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1996","title":"Error when exploring `arabic_speech_corpus`","comments":"Actually soundfile is not a dependency of this dataset.\r\nThe error comes from a bug that was fixed in this commit: https:\/\/github.com\/huggingface\/datasets\/pull\/1767\/commits\/c304e63629f4453367de2fd42883a78768055532\r\nBasically the library used to consider the `import soundfile` in the docstring as a dependency, while it's just here as a code example.\r\n\r\nUpdating the viewer to the latest version of `datasets` should fix this issue\r\n","body":"Navigate to https:\/\/huggingface.co\/datasets\/viewer\/?dataset=arabic_speech_corpus\r\n\r\nError:\r\n```\r\nImportError: To be able to use this dataset, you need to install the following dependencies['soundfile'] using 'pip install soundfile' for instance'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/script_runner.py\", line 332, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 233, in \r\n configs = get_confs(option)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 604, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 588, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 145, in get_confs\r\n module_path = nlp.load.prepare_module(path, dataset=True\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 342, in prepare_module\r\n f\"To be able to use this {module_type}, you need to install the following dependencies\"\r\n```","comment_length":58,"text":"Error when exploring `arabic_speech_corpus`\nNavigate to https:\/\/huggingface.co\/datasets\/viewer\/?dataset=arabic_speech_corpus\r\n\r\nError:\r\n```\r\nImportError: To be able to use this dataset, you need to install the following dependencies['soundfile'] using 'pip install soundfile' for instance'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/script_runner.py\", line 332, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 233, in \r\n configs = get_confs(option)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 604, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 588, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 145, in get_confs\r\n module_path = nlp.load.prepare_module(path, dataset=True\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 342, in prepare_module\r\n f\"To be able to use this {module_type}, you need to install the following dependencies\"\r\n```\nActually soundfile is not a dependency of this dataset.\r\nThe error comes from a bug that was fixed in this commit: https:\/\/github.com\/huggingface\/datasets\/pull\/1767\/commits\/c304e63629f4453367de2fd42883a78768055532\r\nBasically the library used to consider the `import soundfile` in the docstring as a dependency, while it's just here as a code example.\r\n\r\nUpdating the viewer to the latest version of `datasets` should fix this issue\r\n","embeddings":[-0.2591242194,-0.1135890707,-0.0363894068,0.2228231728,0.0417897142,0.0414845236,0.0535821207,0.3140556812,-0.1731435061,0.0504592061,-0.2981741428,0.096908465,-0.1112824306,-0.04064418,0.106541194,-0.3713718355,0.0633603632,0.2032412887,0.005492107,-0.013290599,-0.069745712,0.4542385638,-0.3286394179,-0.0139067518,-0.0968818739,-0.2608732879,-0.0601527542,-0.1154938638,-0.2918647826,-0.4660024047,0.1471929848,-0.076424405,0.0929136053,0.3895767331,-0.0001143709,-0.0865352377,0.3681040704,-0.1396298409,-0.2088981122,-0.4318641722,0.1541046649,-0.1369885504,-0.0995633006,-0.1122164428,-0.0662241131,-0.1979067475,0.0398826748,-0.3098419011,0.291513294,0.3186619878,0.2341214418,0.1870319992,0.2347778231,-0.0418161787,0.2049755454,-0.001658614,-0.1434957534,0.2094648927,0.1850297153,0.0741269141,-0.1772945672,0.5136584044,-0.3165772259,-0.0839223936,0.0919756144,-0.1553822309,-0.0313259773,-0.402174741,0.1798986644,-0.0613103509,0.4357353151,-0.361638546,-0.2422319204,-0.2717438042,0.1527153701,-0.23318322,0.2647298574,0.3108150065,-0.2944638133,0.1316215843,0.0926432088,-0.2117515504,-0.0141713871,0.1655197144,0.0573513359,0.2916793227,-0.2994027436,-0.1460861266,0.3502129912,-0.1967080832,-0.248562932,0.0474212132,-0.0852020159,0.2425550222,-0.0651767328,-0.0221833549,0.1632066518,0.1580073833,-0.038769275,0.0165190455,-0.1195856184,0.2261201441,-0.179515183,0.2216259688,-0.014147697,0.1789913625,0.2417352796,-0.0200956222,0.2245754004,0.4123170674,0.0120387496,0.0000338793,-0.0760709643,-0.1948828399,-0.4148888886,-0.0548379421,0.5135284066,-0.2739714384,-0.1555497646,0.0554313846,-0.0786811337,-0.2209295183,0.0962209627,0.4137216508,-0.1246656626,-0.0105801523,0.1273990124,0.3244036734,-0.2293329239,-0.2025601268,0.0115854368,0.2179549485,-0.0469130166,0.1443846077,0.2509663999,-0.463083297,0.5074711442,0.0049599991,0.170227915,0.1916871667,-0.0764003545,-0.0905369222,-0.0938392431,0.3047252297,-0.0437420234,0.1594018787,0.2130078077,-0.0378161259,-0.2543689013,0.0117811859,0.0311254319,-0.2173605859,-0.0682519376,0.1823175699,-0.3190977275,-0.0554397292,-0.1916362494,0.2330007702,0.1791097969,-0.496638447,-0.0456907302,0.0339263417,-0.531935215,-0.0135487672,0.0597186871,0.4067554772,-0.0727342069,-0.2927673757,-0.0841919407,0.0253014583,0.1555551142,0.1164548472,-0.0388877057,0.0291495975,-0.2782105803,0.3276979029,0.3454294503,-0.5265828967,-0.4309079945,0.1337174028,0.1185409948,0.2625019252,0.331253171,-0.2331041694,0.2469700873,-0.2004732341,0.1713262349,0.3000825942,0.1903238297,-0.1005732864,-0.1891236007,-0.1464096457,-0.0314402953,0.0946110785,-0.1408831626,0.1675128043,-0.0015275398,-0.0004478886,0.1974836141,0.0762266964,0.0679527372,0.2516246438,0.2334304005,0.0558910705,0.0565039031,-0.4098466039,0.0838468522,0.0377535149,-0.3253620863,0.425362289,-0.4172182977,-0.1359889805,-0.2352312803,-0.0995791852,-0.2411767989,-0.1433660686,0.1581530273,0.2674131095,0.1796350032,0.3149423897,-0.2450620979,-0.0018464492,0.1778332293,-0.0270125307,-0.6278404593,0.08613047,-0.213807106,-0.111375235,0.3707222939,0.2790821195,0.1563702971,-0.1181210577,-0.0805430114,0.2213092595,-0.0635786131,0.3369646966,-0.4050099254,0.1184294224,0.1225035116,-0.3153544664,0.38484326,0.1509910375,0.124534443,-0.055124674,0.3454687595,0.190918833,0.2136061043,0.1452511549,0.1701558977,0.0841465667,0.4480590522,0.1295617819,-0.2246260196,-0.3109557629,0.4079634249,-0.2685557008,0.5072845221,0.0207666755,-0.3180295825,0.0073704054,0.5209943652,-0.0545143411,0.1844329238,0.1243077293,-0.3222737908,-0.1683200747,0.2562192082,0.1419637054,0.4471825361,0.1286046356,0.0779226944,0.196065858,0.0627725869,-0.3435527384,0.2704863548,0.0305357389,0.0706947371,0.2770344317,0.0956616327,0.0486453585,-0.5301122665,-0.0327140242,-0.1379942298,0.1653428078,-0.1218276396,0.016171189,-0.3099750876,-0.5472095013,-0.1764027178,-0.2263454944,-0.3613773584,-0.1843630821,0.0760281458,0.2584050596,0.0806570798,0.3466959596,0.1766165793,0.0651235208,-0.1550276279,-0.183698222,-0.2633356154,-0.0257067997,-0.1293578595,0.0739425123,0.3326383531,0.2319948673,0.219444856,-0.124803029,-0.0389512628,-0.2586819828,-0.178327024,0.2145289332,-0.2075853795,0.1726353765,0.1471442282,0.1403462887,-0.0251215193,-0.3683167696,0.2150206566,0.0055194492,-0.0820420235,0.1785067469,-0.1032170579,0.0165980794,-0.1582727134,-0.2487445176,-0.2342053056,-0.5632388592,0.0237041507,0.0359938331,0.1542959362,0.6405217648,-0.0233093947,0.2233794779,-0.1866188198,0.4464220703,-0.1898086518,-0.2654583752,0.3701419234,-0.3133110106,-0.3851189911,0.0832386091,0.0741175562,0.5423164368,-0.1641801596,-0.4779160917,-0.0479016714,-0.2519604862,-0.0908003151,-0.02739962,0.1235230565,0.3727428913,0.1244206727,-0.0558640882,-0.1544632614,-0.112755768,-0.2120387852,0.0303627662,0.1103438288,0.1173396856,0.3938210011,0.0364331193,0.5581880212,0.1893629134,0.078609027,0.5083714724,0.2056812793,0.4620286822,-0.1761783212,-0.500648737,-0.064554818,-0.0106145479,0.0782230645,0.3012274206,0.0582546219,-0.2225743085,-0.4180655777,-0.345164001,-0.1373446286,-0.2126168311,-0.0264404882,-0.2075889856,0.3192279935,0.162668094,0.0413433723,0.0573910512,-0.0178873613,0.2042737603,0.0971604288,0.0731191859,-0.0063096834,-0.3482369781,-0.2147995383,-0.6190564632,0.3982363343,0.059107542,0.2502116859,-0.1125882417,0.0264824145,0.0450295955,-0.0094931722,0.3792364299,-0.0885430351,0.0296731759,0.2719982862,0.0443006828,-0.435283035,0.0519292392,-0.2449258566,0.1143578067,-0.0659605712,0.2073743939,-0.1998904347,-0.0538147949,0.1953683197,0.2936965227,0.1380239129,-0.0931454897,-0.4292460382,-0.4469646811,-0.4218345582,-0.0641522482,0.0973074436,0.2505412102,0.2945302725,0.2912116349,0.2619181573,-0.029073691,0.1567693949,-0.0027871188,0.0268260166,0.2493472844,0.0158035737,0.1950422525,-0.0052004503,-0.1470480114,0.5904424191,0.0382179655,-0.3190025389,-0.122130096,0.1556513011,0.0909768641,0.3510743976,-0.1618376076,0.2359801233,0.1832559258,-0.0882861689,-0.1249307394,-0.1221575812,0.35414204,0.2038032115,-0.1567082107,-0.4649757147,0.2240638137,-0.0867288113,0.0144095998,0.2086192071,0.1458695382,-0.2154883444,0.5865086913,-0.3359625936,0.9864442348,0.190093115,0.1509925872,0.5796772242,0.2852431536,0.2736801505,-0.0253481977,0.2704006135,0.1270617098,-0.238360405,-0.0637592003,-0.0524395406,0.4439462125,-0.1628191769,-0.447186023,0.1975349039,0.1380912215,-0.1930736303,-0.0888213962,0.150246188,-0.3072443306,-0.3726422489,-0.7613896132,0.1629418731,-0.0892553553,0.4359186888,-0.032936614,-0.0648277402,0.0107070953,-0.4352568388,-0.3513130248,0.1951482892,-0.2214563191,0.0813255981,-0.2043264359,0.0401032157,0.3659397364,0.2889255881,0.1460969448,0.2889964879,-0.2387267053,0.1064589769,-0.1322721988,-0.3824173212,0.1719483733,0.1419314295,0.2849148512,-0.2990046442,-0.1611959785,0.2049422562,-0.0569149628,0.0150962304,0.0441516042,0.0061556352,-0.0307205487,-0.11245846,-0.5855567455,0.0706819892,-0.055304952,-0.1381873339,0.1358395815,0.0243080333,-0.4325526357,0.0252658874,-0.1699226797,-0.1066305339,-0.0793187246,0.5607010722,0.188002035,-0.1698703766,0.3974300921,0.3162394762,-0.2394101769,-0.1771647483,-0.0958883241,-0.3306083977,-0.4093735218,-0.0427022167,0.3858549893,-0.1172357872,-0.2525817156,0.3071599007,-0.0619057752,-0.0096783722,-0.0838892683,-0.2928608656,-0.1446076632,0.4782183766,0.0091462722,-0.1794577539,0.1440698653,-0.0686885267,0.0756362081,0.2157730311,-0.3266898096,0.1565910876,-0.0723539516,0.2035768181,0.310582608,0.0021459099,0.2390638739,0.0502424724,0.1470933408,0.1637432873,-0.2215701491,-0.2032015473,0.0223944653,0.1006707102,0.1046754345,-0.0168410596,0.1903942674,-0.0843024179,-0.2097076923,-0.1461955756,-0.0074352361,0.3898077905,-0.0087416545,0.035072878,-0.1786101907,0.0307351705,0.179019019,0.0001560032,-0.1717806906,-0.1406946778,0.1362287849,0.0179869402,-0.1118769646,-0.0856126398,-0.0397433154,0.1052897722,0.0032103818,-0.0333004929,0.2414391041,-0.1479864269,0.1153252199,0.2241200954,0.2735120952,0.2041810006,-0.1730301678,-0.0259724613,0.3789480031,0.1346418262,-0.5750605464,-0.1294369251,0.223694846,-0.1634568423,0.1694572121,0.1812381148,0.2479253709,0.004999904,0.2487309724,0.0701472908,0.4369534254,-0.0982136354,0.3608632088,0.0550136641,0.173820138,-0.0835207477,0.1091649607,0.1809803396,0.2475746274,0.358109951,-0.5064556599,0.2399813831,0.1221566498,-0.031216763,0.1307129115,-0.2954682112,0.117845647,0.5222585201,0.0929436758,0.0833183303,-0.1304382831,0.2733746767,-0.1937644035,-0.0702187493,0.0682204813,0.2420196831,0.0126387617,0.1243281141,0.0829734579,-0.1413499564,-0.0892260447,-0.1605895758,0.1149321347,-0.110643886,0.0178373288,-0.0927027389,-0.13889274,-0.1294720918,0.0077093933,-0.0036960028,0.0070393737,-0.1362717748,0.3819719851,0.0402797051,0.0134988399,0.0545710549,0.52889961,0.6157764196,0.2225263864,0.1539784521,0.229183808,-0.3454715014,-0.0407248177,-0.0221723486,0.2495703101,0.258128643,0.3174584508,0.1913505048,0.102173306,-0.1908325851,-0.1247801334,0.3226773739,0.1148628518,-0.2692886591,0.1078656465,-0.6182484627,-0.0192068238,-0.2506207526,0.0213130917,-0.7599904537,0.0891006812,0.4706355929,-0.088784568,0.032754235,-0.2252693027,0.0818886533,-0.0447083935,0.3434863389,0.3529574871,0.2142374963,0.0181449242,-0.3278158307,-0.7539111376,0.1326278597,0.0954454839,0.1085136086,-0.1816552281,-0.0798448324,0.3282215297,0.3236334324,-0.1485588253,0.0536653697,-0.2219112366,-0.3084178567,-0.1348038167,0.0154723572,-0.0775049925,0.21508573,-0.1514213383,-0.3206720054,-0.1670661569,-0.3572373688,0.0848185942,-0.0512857623,-0.0297918972,-0.3333074749,-0.097757414,0.0851007029,0.2058525383,0.3144366145,-0.255230695,-0.1501885653,-0.1613139957,-0.2632245719,-0.0887675583,0.303211987,0.035533458,0.568321228,-0.2828616202,0.0654952228,-0.3003047407,0.5356615186,0.0982845947,0.004089674,-0.2274110764,0.3459278047,-0.1644188315,0.1881946325,-0.0067118257,0.0318458118,-0.0616733171,0.1145704761,-0.4813273251,-0.3261032999,0.6847342849,-0.5025780201,-0.2505757809,0.1268018633,0.3253098726,0.1369988024,-0.3709216416,-0.6134672761,0.1257463098,0.1614841372,0.1585963964,-0.1042658836,0.1900280267,-0.371320039,-0.2188589573,-0.0757652894,0.2658396065,0.2960599959,-0.1947958022,0.2011632323,-0.291243434]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1994","title":"not being able to get wikipedia es language","comments":"@lhoestq I really appreciate if you could help me providiing processed datasets, I do not really have access to enough resources to run the apache-beam and need to run the codes on these datasets. Only en\/de\/fr currently works, but I need all the languages more or less. thanks ","body":"Hi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help ","comment_length":48,"text":"not being able to get wikipedia es language\nHi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help \n@lhoestq I really appreciate if you could help me providiing processed datasets, I do not really have access to enough resources to run the apache-beam and need to run the codes on these datasets. Only en\/de\/fr currently works, but I need all the languages more or less. thanks ","embeddings":[-0.342106998,0.0471930392,-0.1062649339,0.0620172322,0.200454846,0.1777444929,0.1768746823,0.3401965797,0.1373334974,0.0921178386,0.4105240405,0.3346937597,0.0160005912,0.3076457381,0.1113204286,-0.3370872438,0.0556506142,0.0752215236,-0.1947185397,-0.2001760751,-0.1320842654,0.1037431955,-0.2725183666,-0.0701157525,-0.1488354355,0.1070382744,0.1227778196,-0.0196700636,-0.192601651,-0.1862009615,0.0680222437,-0.036213655,0.2280335724,0.1079433188,-0.0001084146,0.0806313604,0.5717446804,-0.1583800614,-0.4590220749,-0.130322203,-0.081363827,-0.3462582827,0.2564504743,-0.3715744317,-0.3347848058,-0.0256394912,0.3187549412,-0.5857167244,0.2412877679,0.1479617506,0.2182685584,-0.074210979,0.3515349925,-0.0048845191,0.4105975032,0.218606472,-0.0360494033,-0.1243073717,0.0106633129,0.0249649677,0.0606173314,0.4618479013,0.0472859293,-0.1142473742,0.3039729893,-0.3094903827,0.2656840682,-0.3731164932,0.4229684472,0.2541317344,1.0161632299,-0.1197014451,0.0995212495,0.1050666645,0.0040745316,0.1287817061,0.2428365648,0.2005350441,-0.3642728925,-0.0796793103,0.1134936661,-0.2830438912,-0.335678637,0.3455581665,-0.1017287895,0.5368958116,0.1090375558,0.1416433603,-0.1509752721,-0.1852888316,0.0773017183,-0.0707799047,0.1030773669,0.3541413844,-0.0930447504,0.1351349652,0.2457793206,0.1751329154,0.0443397351,-0.0947965086,-0.2797648609,0.1917012036,0.368519485,0.0461870804,0.0220659487,0.0323531926,0.3556627333,-0.1849304587,0.2050388753,-0.030573098,0.0079394793,0.06664785,-0.0586212762,-0.3365755379,-0.6453832984,0.2325167805,-0.0009588756,-0.1367516071,0.192342177,0.1587269008,-0.3540838063,-0.2998664379,-0.0057489178,0.2017250508,0.1522044241,0.2051189095,0.2833256125,-0.0867259279,-0.3417105973,-0.3844312429,-0.0466403812,0.2829635143,-0.4593487978,0.0780525506,0.2093625367,0.1067860499,0.3750687838,-0.0355147943,-0.0439997055,0.0425268859,0.1307713985,0.0788350478,-0.1385989934,0.1594749987,0.1490190625,0.2470012158,0.199904263,-0.222467348,-0.052217897,0.0016842578,-0.0903736725,-0.1905186325,-0.18565543,0.1717608422,0.1227803901,0.1018858626,0.0458261073,0.3413800597,0.1713995934,-0.2258878946,0.1436647028,-0.0075095682,-0.1345748454,-0.0918872654,0.3624586463,0.372572124,-0.6563616395,0.1040466204,-0.0495637171,0.1877678782,0.143699944,-0.2795007229,-0.281387955,0.4237986207,-0.093520321,-0.1007723063,0.3089175224,-0.2708930373,0.0815908834,0.1968420744,0.067334421,-0.2822613418,0.2371014357,-0.1123532802,0.0422963984,0.1624549776,0.1141090319,0.1663445979,0.116460517,-0.0702252239,-0.2196758091,-0.1222946942,-0.0431540161,0.1172707975,0.2665446103,-0.1195803061,0.1249019429,0.5266057849,0.3261623085,-0.2632548511,0.0764246359,0.5432159305,-0.3122918904,0.0806591958,0.1918928772,-0.0554747656,-0.0498997234,0.1809134334,-0.239194259,0.4039117098,0.0790574178,-0.0053859027,-0.2668075264,-0.0530283935,-0.2360287756,-0.4283643067,0.2114921659,0.0466564186,-0.0355188176,0.2530859709,0.0529352501,-0.0900353342,-0.278223455,-0.1389373243,-0.8122121692,0.2934658229,-0.2860761285,-0.0800810456,-0.0875601247,0.0202554129,0.0639516488,-0.1236343607,-0.2867787182,0.1009152383,0.1873278469,0.0372007079,-0.011949122,0.16449368,0.1220298633,-0.2648282647,0.3166707158,0.3054229021,0.1199404374,0.0179688167,-0.2019145489,0.1108080298,0.0477327667,0.3355243504,0.0727865323,0.2685893774,0.185183093,0.1680476964,-0.1562812775,-0.1591639668,0.302156806,0.4226678014,0.1818974614,-0.145703271,0.0370389149,0.0772163868,0.451574564,0.1048435643,-0.0922244564,-0.1263718456,-0.3054101467,-0.0594792664,0.4833448827,-0.1791246086,-0.0160230976,0.2791057527,-0.0161974188,0.0771187916,-0.1511853486,-0.1175531745,0.4446229041,0.0723261684,0.5178593993,-0.0257208422,-0.1244166493,-0.1868189722,-0.1908410639,-0.2879425287,-0.1234138981,0.2847407758,-0.312468946,0.2768905461,-0.4178833961,-0.5642531514,-0.2084293067,0.4299700558,-0.3480241597,-0.3430698216,-0.0680083781,0.0778862834,-0.0229108091,0.2500188053,0.107210204,-0.035542842,0.2320447713,0.0074737356,-0.2216777354,-0.5956392884,-0.4356403053,0.0326633267,0.2823260725,0.0373752452,0.1272295564,-0.1569059193,-0.0672418326,-0.1452952772,-0.4021958709,0.3414442241,-0.0869849101,-0.0084638475,0.1166531071,0.4272188842,-0.19598113,-0.0950585008,0.3098010421,0.1152016073,-0.1476498395,-0.0692656934,-0.084672384,-0.0623619594,-0.0308230799,-0.4383021891,-0.1505855471,-0.3273377717,-0.1820004731,0.1689832062,0.1332353055,0.3523068726,0.2645772398,0.2091489136,0.3283595741,0.0954612941,-0.0965576619,-0.15933685,0.2170184106,-0.3185723126,-0.3167462349,0.1040425673,-0.1413134485,0.2305069566,0.1135405153,-0.146329999,0.0680313334,0.1159731671,0.0200549383,-0.1013204157,0.3257164955,0.7469013929,-0.0126242423,-0.0016661583,-0.0988114774,0.0478255115,0.0307381041,-0.4368053079,0.347318083,0.1777924001,0.5277687311,-0.0701474324,0.7515996695,0.1148355156,0.1166792214,0.1321653128,-0.0752609968,0.066186063,-0.272472769,-0.1884595752,0.292780757,0.0403270945,-0.3030905426,0.3898431957,-0.1277024895,-0.2598915696,-0.4387769699,0.0750479698,0.0015996028,-0.2447390854,-0.0376059711,-0.3750056326,0.3839118779,-0.0953029543,0.3992151022,-0.0802452266,-0.0731422156,0.0592001341,0.2196860462,-0.0982458889,0.1342238635,-0.1730971187,-0.2497496158,-0.3993260562,0.2461431772,0.0012104877,0.1436176598,-0.2927529216,0.0079169897,0.2831737101,0.0087449783,0.520098865,-0.647043407,-0.0973827019,0.3353388906,0.1921800226,-0.6571789384,-0.1413870305,-0.1876775622,0.2771171629,0.328502208,-0.0478469282,-0.3322197795,-0.1513579339,0.2680896223,0.0485692993,-0.0990989432,0.0515208356,-0.2417580783,-0.1433175504,-0.3353155851,-0.0657824352,-0.0996938273,0.1351074874,0.1002394855,0.2149961293,0.0813034996,0.0262187272,-0.1448780149,0.2050136924,0.1226052493,0.1383641809,-0.2560669482,0.1037125513,0.4806472957,-0.0265891049,-0.1080244482,0.1851131022,-0.2079540342,0.1726381034,-0.2150923759,0.1455085427,0.1604099125,-0.1206459701,-0.2393637598,0.0364794657,-0.2542999983,-0.0464653,0.0528427027,0.2087159753,0.0810022354,-0.1274007857,-0.4575788677,0.5609183908,0.1885291785,-0.0729709864,0.2990112603,-0.0156652872,-0.4306363761,0.4052890837,0.4329436719,0.8635653257,0.0299162623,0.0139987627,0.1714195311,0.1671361923,0.6254860759,-0.5466772914,0.209164381,-0.2812898755,0.3198871613,-0.0952548385,0.0653773472,0.3957326412,0.1431853473,-0.2221702784,0.3536064923,-0.0173848756,0.0741586462,0.056596484,0.3767177761,0.1265334934,-0.3078525662,-0.1175289601,0.1213964745,-0.1823625863,0.2824687064,-0.2498215139,-0.0125390952,-0.0145078488,-0.2449239939,-0.4107005596,0.0911933407,-0.2241836935,0.460855484,-0.3423030674,-0.4593057334,0.2506963611,0.2268087268,0.2354613245,0.3696662784,-0.3735500574,0.2001679838,0.01500485,-0.2361067981,-0.0761883631,0.0410975851,0.3866562247,-0.1922802478,-0.392267704,0.2118118405,-0.1738403141,-0.0978837684,-0.2912859023,-0.3177188933,0.3188324571,0.2502019703,0.1240342259,0.1725913435,-0.0860563293,-0.1338620186,0.1198045909,-0.284142524,0.0101039587,0.0086884415,0.0882760808,-0.0868183151,0.030908918,0.2571513057,0.1876173764,0.0172616523,0.5952085853,0.3785053194,-0.2194748968,-0.2659060359,0.1786083579,-0.2292437404,-0.0606894419,-0.2634168267,0.1508063525,0.0923178792,-0.0571265072,0.1029199064,-0.0157441776,-0.1409106106,-0.0071760365,-0.268037647,-0.0209581573,0.0222998895,-0.1638095528,-0.0015579133,0.1883054525,0.1332456172,0.2853651345,-0.2360227406,-0.2791166902,-0.238835454,-0.0223192759,0.0453209914,0.413713932,-0.0166400243,-0.2360471487,-0.038738668,0.1452317834,-0.2301054597,-0.2452716976,-0.1725503653,-0.0468346812,0.1290072203,-0.0397849046,-0.1874575168,0.135853678,-0.3403603733,-0.2925044596,-0.2041295767,-0.1316379607,-0.1293069422,-0.0242854469,0.2222887725,-0.1042095274,0.3449909389,-0.2877995074,0.0879830047,0.0348752849,0.2938313782,0.0511713922,0.2495486438,0.4168614745,-0.1764672995,-0.5602076054,0.1830450594,-0.2211462706,0.174384281,0.1633221805,-0.0828019083,-0.0480107516,0.0299655739,-0.0214739311,0.026401367,-0.0819890797,-0.0302336216,0.135588184,0.2045512199,-0.2827223837,0.1239569485,0.4220980108,-0.0330974609,0.0961894318,0.026279293,0.2615511715,-0.0534218065,0.1746627986,0.0410566963,0.0168639217,-0.0434479639,0.2639892101,0.0946073607,0.0811476633,0.1160609722,-0.120071359,0.046419654,0.1310725361,0.3746131957,0.1071262732,0.0660357401,0.1008156091,0.21656923,-0.2253761888,-0.1409864128,0.3062992394,-0.1027750447,-0.0209084284,0.0228774995,-0.064572908,0.0748565868,0.1480170488,-0.0637497306,-0.318022579,0.138672933,-0.0777273774,-0.1830140501,-0.6274288893,-0.2322814167,0.0546508804,-0.0176585633,0.0890892744,-0.0836972222,-0.1092474014,0.2026977241,-0.0350260288,-0.2306366116,0.6797729731,-0.0616487004,0.0446974188,-0.0551761761,0.2520872355,-0.2132354975,-0.0712455139,-0.1705531627,-0.0018399615,0.1578250974,0.210928753,-0.4701041579,-0.143992573,-0.0009005668,0.0189441014,-0.05917757,0.0006444135,0.0630011186,-0.0377971195,0.2940508723,0.1096709967,-0.0898015797,-0.1984426975,0.2169168144,-0.0071816775,-0.0200378653,-0.2484554499,0.0045184782,0.1042121798,-0.1066088304,0.1146246046,-0.5152721405,0.0413271375,0.2593996525,0.3884597719,-0.1085285619,-0.2229959518,0.0836435929,-0.3275434077,0.5394175053,0.0841736719,0.1905831844,-0.118047744,-0.2930147946,-0.2950969338,0.0054531754,-0.1695942879,-0.368024677,-0.1203953549,0.3059180677,0.3928041756,0.2445942461,0.1168810576,-0.2535528839,-0.1298053265,0.2650598884,-0.1438629478,-0.1589305401,-0.0732825994,0.0492410101,-0.2251886576,-0.1773677468,0.0832061097,-0.0194310006,0.0752256215,-0.2276214212,-0.1473570764,0.3186527193,-0.4347778261,0.2595769763,0.0258953478,0.5472237468,0.0291320775,-0.0422725081,-0.1655568033,0.222838372,-0.0450001284,0.4914217591,0.0616205968,0.2285088748,-0.215500474,-0.3188503683,-0.3452579677,0.631701231,0.0320965722,-0.0550996438,-0.2490802556,-0.0586701371,-0.2294994295,0.2923282981,-0.1192456484,0.0662930906,-0.0366520099,0.1867338717,-0.3581700027,-0.2367307842,0.3384183347,-0.4211325049,-0.3356799185,-0.189714089,0.0682086274,-0.1494337618,0.2929883599,-0.1057338789,-0.0153541863,0.2634702921,-0.1352269202,-0.0167236272,0.0001432241,0.0598571114,0.1573980451,-0.2458989173,-0.4044720829,-0.1943290979,-0.0603254028,-0.3799315095,-0.2492922992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1994","title":"not being able to get wikipedia es language","comments":"Hi @dorost1234, I think I can help you a little. I\u2019ve processed some Wikipedia datasets (Spanish inclusive) using the HF\/datasets library during recent research.\r\n\r\n@lhoestq Could you help me to upload these preprocessed datasets to Huggingface's repositories? To be more precise, I've built datasets from the following languages using the 20201201 dumps: Spanish, Portuguese, Russian, French, Japanese, Chinese, and Turkish. Process these datasets have high costs that most of the community can't afford. I think these preprocessed datasets I have could be helpful for someone without access to high-resource machines to process Wikipedia's dumps like @dorost1234\r\n\r\n","body":"Hi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help ","comment_length":96,"text":"not being able to get wikipedia es language\nHi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help \nHi @dorost1234, I think I can help you a little. I\u2019ve processed some Wikipedia datasets (Spanish inclusive) using the HF\/datasets library during recent research.\r\n\r\n@lhoestq Could you help me to upload these preprocessed datasets to Huggingface's repositories? To be more precise, I've built datasets from the following languages using the 20201201 dumps: Spanish, Portuguese, Russian, French, Japanese, Chinese, and Turkish. Process these datasets have high costs that most of the community can't afford. I think these preprocessed datasets I have could be helpful for someone without access to high-resource machines to process Wikipedia's dumps like @dorost1234\r\n\r\n","embeddings":[-0.342106998,0.0471930392,-0.1062649339,0.0620172322,0.200454846,0.1777444929,0.1768746823,0.3401965797,0.1373334974,0.0921178386,0.4105240405,0.3346937597,0.0160005912,0.3076457381,0.1113204286,-0.3370872438,0.0556506142,0.0752215236,-0.1947185397,-0.2001760751,-0.1320842654,0.1037431955,-0.2725183666,-0.0701157525,-0.1488354355,0.1070382744,0.1227778196,-0.0196700636,-0.192601651,-0.1862009615,0.0680222437,-0.036213655,0.2280335724,0.1079433188,-0.0001084146,0.0806313604,0.5717446804,-0.1583800614,-0.4590220749,-0.130322203,-0.081363827,-0.3462582827,0.2564504743,-0.3715744317,-0.3347848058,-0.0256394912,0.3187549412,-0.5857167244,0.2412877679,0.1479617506,0.2182685584,-0.074210979,0.3515349925,-0.0048845191,0.4105975032,0.218606472,-0.0360494033,-0.1243073717,0.0106633129,0.0249649677,0.0606173314,0.4618479013,0.0472859293,-0.1142473742,0.3039729893,-0.3094903827,0.2656840682,-0.3731164932,0.4229684472,0.2541317344,1.0161632299,-0.1197014451,0.0995212495,0.1050666645,0.0040745316,0.1287817061,0.2428365648,0.2005350441,-0.3642728925,-0.0796793103,0.1134936661,-0.2830438912,-0.335678637,0.3455581665,-0.1017287895,0.5368958116,0.1090375558,0.1416433603,-0.1509752721,-0.1852888316,0.0773017183,-0.0707799047,0.1030773669,0.3541413844,-0.0930447504,0.1351349652,0.2457793206,0.1751329154,0.0443397351,-0.0947965086,-0.2797648609,0.1917012036,0.368519485,0.0461870804,0.0220659487,0.0323531926,0.3556627333,-0.1849304587,0.2050388753,-0.030573098,0.0079394793,0.06664785,-0.0586212762,-0.3365755379,-0.6453832984,0.2325167805,-0.0009588756,-0.1367516071,0.192342177,0.1587269008,-0.3540838063,-0.2998664379,-0.0057489178,0.2017250508,0.1522044241,0.2051189095,0.2833256125,-0.0867259279,-0.3417105973,-0.3844312429,-0.0466403812,0.2829635143,-0.4593487978,0.0780525506,0.2093625367,0.1067860499,0.3750687838,-0.0355147943,-0.0439997055,0.0425268859,0.1307713985,0.0788350478,-0.1385989934,0.1594749987,0.1490190625,0.2470012158,0.199904263,-0.222467348,-0.052217897,0.0016842578,-0.0903736725,-0.1905186325,-0.18565543,0.1717608422,0.1227803901,0.1018858626,0.0458261073,0.3413800597,0.1713995934,-0.2258878946,0.1436647028,-0.0075095682,-0.1345748454,-0.0918872654,0.3624586463,0.372572124,-0.6563616395,0.1040466204,-0.0495637171,0.1877678782,0.143699944,-0.2795007229,-0.281387955,0.4237986207,-0.093520321,-0.1007723063,0.3089175224,-0.2708930373,0.0815908834,0.1968420744,0.067334421,-0.2822613418,0.2371014357,-0.1123532802,0.0422963984,0.1624549776,0.1141090319,0.1663445979,0.116460517,-0.0702252239,-0.2196758091,-0.1222946942,-0.0431540161,0.1172707975,0.2665446103,-0.1195803061,0.1249019429,0.5266057849,0.3261623085,-0.2632548511,0.0764246359,0.5432159305,-0.3122918904,0.0806591958,0.1918928772,-0.0554747656,-0.0498997234,0.1809134334,-0.239194259,0.4039117098,0.0790574178,-0.0053859027,-0.2668075264,-0.0530283935,-0.2360287756,-0.4283643067,0.2114921659,0.0466564186,-0.0355188176,0.2530859709,0.0529352501,-0.0900353342,-0.278223455,-0.1389373243,-0.8122121692,0.2934658229,-0.2860761285,-0.0800810456,-0.0875601247,0.0202554129,0.0639516488,-0.1236343607,-0.2867787182,0.1009152383,0.1873278469,0.0372007079,-0.011949122,0.16449368,0.1220298633,-0.2648282647,0.3166707158,0.3054229021,0.1199404374,0.0179688167,-0.2019145489,0.1108080298,0.0477327667,0.3355243504,0.0727865323,0.2685893774,0.185183093,0.1680476964,-0.1562812775,-0.1591639668,0.302156806,0.4226678014,0.1818974614,-0.145703271,0.0370389149,0.0772163868,0.451574564,0.1048435643,-0.0922244564,-0.1263718456,-0.3054101467,-0.0594792664,0.4833448827,-0.1791246086,-0.0160230976,0.2791057527,-0.0161974188,0.0771187916,-0.1511853486,-0.1175531745,0.4446229041,0.0723261684,0.5178593993,-0.0257208422,-0.1244166493,-0.1868189722,-0.1908410639,-0.2879425287,-0.1234138981,0.2847407758,-0.312468946,0.2768905461,-0.4178833961,-0.5642531514,-0.2084293067,0.4299700558,-0.3480241597,-0.3430698216,-0.0680083781,0.0778862834,-0.0229108091,0.2500188053,0.107210204,-0.035542842,0.2320447713,0.0074737356,-0.2216777354,-0.5956392884,-0.4356403053,0.0326633267,0.2823260725,0.0373752452,0.1272295564,-0.1569059193,-0.0672418326,-0.1452952772,-0.4021958709,0.3414442241,-0.0869849101,-0.0084638475,0.1166531071,0.4272188842,-0.19598113,-0.0950585008,0.3098010421,0.1152016073,-0.1476498395,-0.0692656934,-0.084672384,-0.0623619594,-0.0308230799,-0.4383021891,-0.1505855471,-0.3273377717,-0.1820004731,0.1689832062,0.1332353055,0.3523068726,0.2645772398,0.2091489136,0.3283595741,0.0954612941,-0.0965576619,-0.15933685,0.2170184106,-0.3185723126,-0.3167462349,0.1040425673,-0.1413134485,0.2305069566,0.1135405153,-0.146329999,0.0680313334,0.1159731671,0.0200549383,-0.1013204157,0.3257164955,0.7469013929,-0.0126242423,-0.0016661583,-0.0988114774,0.0478255115,0.0307381041,-0.4368053079,0.347318083,0.1777924001,0.5277687311,-0.0701474324,0.7515996695,0.1148355156,0.1166792214,0.1321653128,-0.0752609968,0.066186063,-0.272472769,-0.1884595752,0.292780757,0.0403270945,-0.3030905426,0.3898431957,-0.1277024895,-0.2598915696,-0.4387769699,0.0750479698,0.0015996028,-0.2447390854,-0.0376059711,-0.3750056326,0.3839118779,-0.0953029543,0.3992151022,-0.0802452266,-0.0731422156,0.0592001341,0.2196860462,-0.0982458889,0.1342238635,-0.1730971187,-0.2497496158,-0.3993260562,0.2461431772,0.0012104877,0.1436176598,-0.2927529216,0.0079169897,0.2831737101,0.0087449783,0.520098865,-0.647043407,-0.0973827019,0.3353388906,0.1921800226,-0.6571789384,-0.1413870305,-0.1876775622,0.2771171629,0.328502208,-0.0478469282,-0.3322197795,-0.1513579339,0.2680896223,0.0485692993,-0.0990989432,0.0515208356,-0.2417580783,-0.1433175504,-0.3353155851,-0.0657824352,-0.0996938273,0.1351074874,0.1002394855,0.2149961293,0.0813034996,0.0262187272,-0.1448780149,0.2050136924,0.1226052493,0.1383641809,-0.2560669482,0.1037125513,0.4806472957,-0.0265891049,-0.1080244482,0.1851131022,-0.2079540342,0.1726381034,-0.2150923759,0.1455085427,0.1604099125,-0.1206459701,-0.2393637598,0.0364794657,-0.2542999983,-0.0464653,0.0528427027,0.2087159753,0.0810022354,-0.1274007857,-0.4575788677,0.5609183908,0.1885291785,-0.0729709864,0.2990112603,-0.0156652872,-0.4306363761,0.4052890837,0.4329436719,0.8635653257,0.0299162623,0.0139987627,0.1714195311,0.1671361923,0.6254860759,-0.5466772914,0.209164381,-0.2812898755,0.3198871613,-0.0952548385,0.0653773472,0.3957326412,0.1431853473,-0.2221702784,0.3536064923,-0.0173848756,0.0741586462,0.056596484,0.3767177761,0.1265334934,-0.3078525662,-0.1175289601,0.1213964745,-0.1823625863,0.2824687064,-0.2498215139,-0.0125390952,-0.0145078488,-0.2449239939,-0.4107005596,0.0911933407,-0.2241836935,0.460855484,-0.3423030674,-0.4593057334,0.2506963611,0.2268087268,0.2354613245,0.3696662784,-0.3735500574,0.2001679838,0.01500485,-0.2361067981,-0.0761883631,0.0410975851,0.3866562247,-0.1922802478,-0.392267704,0.2118118405,-0.1738403141,-0.0978837684,-0.2912859023,-0.3177188933,0.3188324571,0.2502019703,0.1240342259,0.1725913435,-0.0860563293,-0.1338620186,0.1198045909,-0.284142524,0.0101039587,0.0086884415,0.0882760808,-0.0868183151,0.030908918,0.2571513057,0.1876173764,0.0172616523,0.5952085853,0.3785053194,-0.2194748968,-0.2659060359,0.1786083579,-0.2292437404,-0.0606894419,-0.2634168267,0.1508063525,0.0923178792,-0.0571265072,0.1029199064,-0.0157441776,-0.1409106106,-0.0071760365,-0.268037647,-0.0209581573,0.0222998895,-0.1638095528,-0.0015579133,0.1883054525,0.1332456172,0.2853651345,-0.2360227406,-0.2791166902,-0.238835454,-0.0223192759,0.0453209914,0.413713932,-0.0166400243,-0.2360471487,-0.038738668,0.1452317834,-0.2301054597,-0.2452716976,-0.1725503653,-0.0468346812,0.1290072203,-0.0397849046,-0.1874575168,0.135853678,-0.3403603733,-0.2925044596,-0.2041295767,-0.1316379607,-0.1293069422,-0.0242854469,0.2222887725,-0.1042095274,0.3449909389,-0.2877995074,0.0879830047,0.0348752849,0.2938313782,0.0511713922,0.2495486438,0.4168614745,-0.1764672995,-0.5602076054,0.1830450594,-0.2211462706,0.174384281,0.1633221805,-0.0828019083,-0.0480107516,0.0299655739,-0.0214739311,0.026401367,-0.0819890797,-0.0302336216,0.135588184,0.2045512199,-0.2827223837,0.1239569485,0.4220980108,-0.0330974609,0.0961894318,0.026279293,0.2615511715,-0.0534218065,0.1746627986,0.0410566963,0.0168639217,-0.0434479639,0.2639892101,0.0946073607,0.0811476633,0.1160609722,-0.120071359,0.046419654,0.1310725361,0.3746131957,0.1071262732,0.0660357401,0.1008156091,0.21656923,-0.2253761888,-0.1409864128,0.3062992394,-0.1027750447,-0.0209084284,0.0228774995,-0.064572908,0.0748565868,0.1480170488,-0.0637497306,-0.318022579,0.138672933,-0.0777273774,-0.1830140501,-0.6274288893,-0.2322814167,0.0546508804,-0.0176585633,0.0890892744,-0.0836972222,-0.1092474014,0.2026977241,-0.0350260288,-0.2306366116,0.6797729731,-0.0616487004,0.0446974188,-0.0551761761,0.2520872355,-0.2132354975,-0.0712455139,-0.1705531627,-0.0018399615,0.1578250974,0.210928753,-0.4701041579,-0.143992573,-0.0009005668,0.0189441014,-0.05917757,0.0006444135,0.0630011186,-0.0377971195,0.2940508723,0.1096709967,-0.0898015797,-0.1984426975,0.2169168144,-0.0071816775,-0.0200378653,-0.2484554499,0.0045184782,0.1042121798,-0.1066088304,0.1146246046,-0.5152721405,0.0413271375,0.2593996525,0.3884597719,-0.1085285619,-0.2229959518,0.0836435929,-0.3275434077,0.5394175053,0.0841736719,0.1905831844,-0.118047744,-0.2930147946,-0.2950969338,0.0054531754,-0.1695942879,-0.368024677,-0.1203953549,0.3059180677,0.3928041756,0.2445942461,0.1168810576,-0.2535528839,-0.1298053265,0.2650598884,-0.1438629478,-0.1589305401,-0.0732825994,0.0492410101,-0.2251886576,-0.1773677468,0.0832061097,-0.0194310006,0.0752256215,-0.2276214212,-0.1473570764,0.3186527193,-0.4347778261,0.2595769763,0.0258953478,0.5472237468,0.0291320775,-0.0422725081,-0.1655568033,0.222838372,-0.0450001284,0.4914217591,0.0616205968,0.2285088748,-0.215500474,-0.3188503683,-0.3452579677,0.631701231,0.0320965722,-0.0550996438,-0.2490802556,-0.0586701371,-0.2294994295,0.2923282981,-0.1192456484,0.0662930906,-0.0366520099,0.1867338717,-0.3581700027,-0.2367307842,0.3384183347,-0.4211325049,-0.3356799185,-0.189714089,0.0682086274,-0.1494337618,0.2929883599,-0.1057338789,-0.0153541863,0.2634702921,-0.1352269202,-0.0167236272,0.0001432241,0.0598571114,0.1573980451,-0.2458989173,-0.4044720829,-0.1943290979,-0.0603254028,-0.3799315095,-0.2492922992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1994","title":"not being able to get wikipedia es language","comments":"Thank you so much @jonatasgrosman , I greatly appreciate your help with them. \r\nYes, I unfortunately does not have access to a good resource and need it for my\r\nresearch. I greatly appreciate @lhoestq your help with uploading the processed datasets in huggingface datasets. This would be really helpful for some users like me with not access to high-memory GPU resources.\r\n\r\nthank you both so much again.\r\n\r\nOn Sat, Mar 6, 2021 at 12:55 AM Jonatas Grosman \r\nwrote:\r\n\r\n> Hi @dorost1234 , I think I can help you a\r\n> little. I\u2019ve processed some Wikipedia datasets (Spanish inclusive) using\r\n> the HF\/datasets library during recent research.\r\n>\r\n> @lhoestq Could you help me to upload these\r\n> preprocessed datasets to Huggingface's repositories? To be more precise,\r\n> I've built datasets from the following languages using the 20201201 dumps:\r\n> Spanish, Portuguese, Russian, French, Japanese, Chinese, and Turkish.\r\n> Process these datasets have high costs that most of the community can't\r\n> afford. I think these preprocessed datasets I have could be helpful for\r\n> someone without access to high-resource machines to process Wikipedia's\r\n> dumps like @dorost1234 \r\n>\r\n> \u2014\r\n> You are receiving this because you were mentioned.\r\n> Reply to this email directly, view it on GitHub\r\n> ,\r\n> or unsubscribe\r\n> \r\n> .\r\n>\r\n","body":"Hi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help ","comment_length":222,"text":"not being able to get wikipedia es language\nHi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help \nThank you so much @jonatasgrosman , I greatly appreciate your help with them. \r\nYes, I unfortunately does not have access to a good resource and need it for my\r\nresearch. I greatly appreciate @lhoestq your help with uploading the processed datasets in huggingface datasets. This would be really helpful for some users like me with not access to high-memory GPU resources.\r\n\r\nthank you both so much again.\r\n\r\nOn Sat, Mar 6, 2021 at 12:55 AM Jonatas Grosman \r\nwrote:\r\n\r\n> Hi @dorost1234 , I think I can help you a\r\n> little. I\u2019ve processed some Wikipedia datasets (Spanish inclusive) using\r\n> the HF\/datasets library during recent research.\r\n>\r\n> @lhoestq Could you help me to upload these\r\n> preprocessed datasets to Huggingface's repositories? To be more precise,\r\n> I've built datasets from the following languages using the 20201201 dumps:\r\n> Spanish, Portuguese, Russian, French, Japanese, Chinese, and Turkish.\r\n> Process these datasets have high costs that most of the community can't\r\n> afford. I think these preprocessed datasets I have could be helpful for\r\n> someone without access to high-resource machines to process Wikipedia's\r\n> dumps like @dorost1234 \r\n>\r\n> \u2014\r\n> You are receiving this because you were mentioned.\r\n> Reply to this email directly, view it on GitHub\r\n> ,\r\n> or unsubscribe\r\n> \r\n> .\r\n>\r\n","embeddings":[-0.342106998,0.0471930392,-0.1062649339,0.0620172322,0.200454846,0.1777444929,0.1768746823,0.3401965797,0.1373334974,0.0921178386,0.4105240405,0.3346937597,0.0160005912,0.3076457381,0.1113204286,-0.3370872438,0.0556506142,0.0752215236,-0.1947185397,-0.2001760751,-0.1320842654,0.1037431955,-0.2725183666,-0.0701157525,-0.1488354355,0.1070382744,0.1227778196,-0.0196700636,-0.192601651,-0.1862009615,0.0680222437,-0.036213655,0.2280335724,0.1079433188,-0.0001084146,0.0806313604,0.5717446804,-0.1583800614,-0.4590220749,-0.130322203,-0.081363827,-0.3462582827,0.2564504743,-0.3715744317,-0.3347848058,-0.0256394912,0.3187549412,-0.5857167244,0.2412877679,0.1479617506,0.2182685584,-0.074210979,0.3515349925,-0.0048845191,0.4105975032,0.218606472,-0.0360494033,-0.1243073717,0.0106633129,0.0249649677,0.0606173314,0.4618479013,0.0472859293,-0.1142473742,0.3039729893,-0.3094903827,0.2656840682,-0.3731164932,0.4229684472,0.2541317344,1.0161632299,-0.1197014451,0.0995212495,0.1050666645,0.0040745316,0.1287817061,0.2428365648,0.2005350441,-0.3642728925,-0.0796793103,0.1134936661,-0.2830438912,-0.335678637,0.3455581665,-0.1017287895,0.5368958116,0.1090375558,0.1416433603,-0.1509752721,-0.1852888316,0.0773017183,-0.0707799047,0.1030773669,0.3541413844,-0.0930447504,0.1351349652,0.2457793206,0.1751329154,0.0443397351,-0.0947965086,-0.2797648609,0.1917012036,0.368519485,0.0461870804,0.0220659487,0.0323531926,0.3556627333,-0.1849304587,0.2050388753,-0.030573098,0.0079394793,0.06664785,-0.0586212762,-0.3365755379,-0.6453832984,0.2325167805,-0.0009588756,-0.1367516071,0.192342177,0.1587269008,-0.3540838063,-0.2998664379,-0.0057489178,0.2017250508,0.1522044241,0.2051189095,0.2833256125,-0.0867259279,-0.3417105973,-0.3844312429,-0.0466403812,0.2829635143,-0.4593487978,0.0780525506,0.2093625367,0.1067860499,0.3750687838,-0.0355147943,-0.0439997055,0.0425268859,0.1307713985,0.0788350478,-0.1385989934,0.1594749987,0.1490190625,0.2470012158,0.199904263,-0.222467348,-0.052217897,0.0016842578,-0.0903736725,-0.1905186325,-0.18565543,0.1717608422,0.1227803901,0.1018858626,0.0458261073,0.3413800597,0.1713995934,-0.2258878946,0.1436647028,-0.0075095682,-0.1345748454,-0.0918872654,0.3624586463,0.372572124,-0.6563616395,0.1040466204,-0.0495637171,0.1877678782,0.143699944,-0.2795007229,-0.281387955,0.4237986207,-0.093520321,-0.1007723063,0.3089175224,-0.2708930373,0.0815908834,0.1968420744,0.067334421,-0.2822613418,0.2371014357,-0.1123532802,0.0422963984,0.1624549776,0.1141090319,0.1663445979,0.116460517,-0.0702252239,-0.2196758091,-0.1222946942,-0.0431540161,0.1172707975,0.2665446103,-0.1195803061,0.1249019429,0.5266057849,0.3261623085,-0.2632548511,0.0764246359,0.5432159305,-0.3122918904,0.0806591958,0.1918928772,-0.0554747656,-0.0498997234,0.1809134334,-0.239194259,0.4039117098,0.0790574178,-0.0053859027,-0.2668075264,-0.0530283935,-0.2360287756,-0.4283643067,0.2114921659,0.0466564186,-0.0355188176,0.2530859709,0.0529352501,-0.0900353342,-0.278223455,-0.1389373243,-0.8122121692,0.2934658229,-0.2860761285,-0.0800810456,-0.0875601247,0.0202554129,0.0639516488,-0.1236343607,-0.2867787182,0.1009152383,0.1873278469,0.0372007079,-0.011949122,0.16449368,0.1220298633,-0.2648282647,0.3166707158,0.3054229021,0.1199404374,0.0179688167,-0.2019145489,0.1108080298,0.0477327667,0.3355243504,0.0727865323,0.2685893774,0.185183093,0.1680476964,-0.1562812775,-0.1591639668,0.302156806,0.4226678014,0.1818974614,-0.145703271,0.0370389149,0.0772163868,0.451574564,0.1048435643,-0.0922244564,-0.1263718456,-0.3054101467,-0.0594792664,0.4833448827,-0.1791246086,-0.0160230976,0.2791057527,-0.0161974188,0.0771187916,-0.1511853486,-0.1175531745,0.4446229041,0.0723261684,0.5178593993,-0.0257208422,-0.1244166493,-0.1868189722,-0.1908410639,-0.2879425287,-0.1234138981,0.2847407758,-0.312468946,0.2768905461,-0.4178833961,-0.5642531514,-0.2084293067,0.4299700558,-0.3480241597,-0.3430698216,-0.0680083781,0.0778862834,-0.0229108091,0.2500188053,0.107210204,-0.035542842,0.2320447713,0.0074737356,-0.2216777354,-0.5956392884,-0.4356403053,0.0326633267,0.2823260725,0.0373752452,0.1272295564,-0.1569059193,-0.0672418326,-0.1452952772,-0.4021958709,0.3414442241,-0.0869849101,-0.0084638475,0.1166531071,0.4272188842,-0.19598113,-0.0950585008,0.3098010421,0.1152016073,-0.1476498395,-0.0692656934,-0.084672384,-0.0623619594,-0.0308230799,-0.4383021891,-0.1505855471,-0.3273377717,-0.1820004731,0.1689832062,0.1332353055,0.3523068726,0.2645772398,0.2091489136,0.3283595741,0.0954612941,-0.0965576619,-0.15933685,0.2170184106,-0.3185723126,-0.3167462349,0.1040425673,-0.1413134485,0.2305069566,0.1135405153,-0.146329999,0.0680313334,0.1159731671,0.0200549383,-0.1013204157,0.3257164955,0.7469013929,-0.0126242423,-0.0016661583,-0.0988114774,0.0478255115,0.0307381041,-0.4368053079,0.347318083,0.1777924001,0.5277687311,-0.0701474324,0.7515996695,0.1148355156,0.1166792214,0.1321653128,-0.0752609968,0.066186063,-0.272472769,-0.1884595752,0.292780757,0.0403270945,-0.3030905426,0.3898431957,-0.1277024895,-0.2598915696,-0.4387769699,0.0750479698,0.0015996028,-0.2447390854,-0.0376059711,-0.3750056326,0.3839118779,-0.0953029543,0.3992151022,-0.0802452266,-0.0731422156,0.0592001341,0.2196860462,-0.0982458889,0.1342238635,-0.1730971187,-0.2497496158,-0.3993260562,0.2461431772,0.0012104877,0.1436176598,-0.2927529216,0.0079169897,0.2831737101,0.0087449783,0.520098865,-0.647043407,-0.0973827019,0.3353388906,0.1921800226,-0.6571789384,-0.1413870305,-0.1876775622,0.2771171629,0.328502208,-0.0478469282,-0.3322197795,-0.1513579339,0.2680896223,0.0485692993,-0.0990989432,0.0515208356,-0.2417580783,-0.1433175504,-0.3353155851,-0.0657824352,-0.0996938273,0.1351074874,0.1002394855,0.2149961293,0.0813034996,0.0262187272,-0.1448780149,0.2050136924,0.1226052493,0.1383641809,-0.2560669482,0.1037125513,0.4806472957,-0.0265891049,-0.1080244482,0.1851131022,-0.2079540342,0.1726381034,-0.2150923759,0.1455085427,0.1604099125,-0.1206459701,-0.2393637598,0.0364794657,-0.2542999983,-0.0464653,0.0528427027,0.2087159753,0.0810022354,-0.1274007857,-0.4575788677,0.5609183908,0.1885291785,-0.0729709864,0.2990112603,-0.0156652872,-0.4306363761,0.4052890837,0.4329436719,0.8635653257,0.0299162623,0.0139987627,0.1714195311,0.1671361923,0.6254860759,-0.5466772914,0.209164381,-0.2812898755,0.3198871613,-0.0952548385,0.0653773472,0.3957326412,0.1431853473,-0.2221702784,0.3536064923,-0.0173848756,0.0741586462,0.056596484,0.3767177761,0.1265334934,-0.3078525662,-0.1175289601,0.1213964745,-0.1823625863,0.2824687064,-0.2498215139,-0.0125390952,-0.0145078488,-0.2449239939,-0.4107005596,0.0911933407,-0.2241836935,0.460855484,-0.3423030674,-0.4593057334,0.2506963611,0.2268087268,0.2354613245,0.3696662784,-0.3735500574,0.2001679838,0.01500485,-0.2361067981,-0.0761883631,0.0410975851,0.3866562247,-0.1922802478,-0.392267704,0.2118118405,-0.1738403141,-0.0978837684,-0.2912859023,-0.3177188933,0.3188324571,0.2502019703,0.1240342259,0.1725913435,-0.0860563293,-0.1338620186,0.1198045909,-0.284142524,0.0101039587,0.0086884415,0.0882760808,-0.0868183151,0.030908918,0.2571513057,0.1876173764,0.0172616523,0.5952085853,0.3785053194,-0.2194748968,-0.2659060359,0.1786083579,-0.2292437404,-0.0606894419,-0.2634168267,0.1508063525,0.0923178792,-0.0571265072,0.1029199064,-0.0157441776,-0.1409106106,-0.0071760365,-0.268037647,-0.0209581573,0.0222998895,-0.1638095528,-0.0015579133,0.1883054525,0.1332456172,0.2853651345,-0.2360227406,-0.2791166902,-0.238835454,-0.0223192759,0.0453209914,0.413713932,-0.0166400243,-0.2360471487,-0.038738668,0.1452317834,-0.2301054597,-0.2452716976,-0.1725503653,-0.0468346812,0.1290072203,-0.0397849046,-0.1874575168,0.135853678,-0.3403603733,-0.2925044596,-0.2041295767,-0.1316379607,-0.1293069422,-0.0242854469,0.2222887725,-0.1042095274,0.3449909389,-0.2877995074,0.0879830047,0.0348752849,0.2938313782,0.0511713922,0.2495486438,0.4168614745,-0.1764672995,-0.5602076054,0.1830450594,-0.2211462706,0.174384281,0.1633221805,-0.0828019083,-0.0480107516,0.0299655739,-0.0214739311,0.026401367,-0.0819890797,-0.0302336216,0.135588184,0.2045512199,-0.2827223837,0.1239569485,0.4220980108,-0.0330974609,0.0961894318,0.026279293,0.2615511715,-0.0534218065,0.1746627986,0.0410566963,0.0168639217,-0.0434479639,0.2639892101,0.0946073607,0.0811476633,0.1160609722,-0.120071359,0.046419654,0.1310725361,0.3746131957,0.1071262732,0.0660357401,0.1008156091,0.21656923,-0.2253761888,-0.1409864128,0.3062992394,-0.1027750447,-0.0209084284,0.0228774995,-0.064572908,0.0748565868,0.1480170488,-0.0637497306,-0.318022579,0.138672933,-0.0777273774,-0.1830140501,-0.6274288893,-0.2322814167,0.0546508804,-0.0176585633,0.0890892744,-0.0836972222,-0.1092474014,0.2026977241,-0.0350260288,-0.2306366116,0.6797729731,-0.0616487004,0.0446974188,-0.0551761761,0.2520872355,-0.2132354975,-0.0712455139,-0.1705531627,-0.0018399615,0.1578250974,0.210928753,-0.4701041579,-0.143992573,-0.0009005668,0.0189441014,-0.05917757,0.0006444135,0.0630011186,-0.0377971195,0.2940508723,0.1096709967,-0.0898015797,-0.1984426975,0.2169168144,-0.0071816775,-0.0200378653,-0.2484554499,0.0045184782,0.1042121798,-0.1066088304,0.1146246046,-0.5152721405,0.0413271375,0.2593996525,0.3884597719,-0.1085285619,-0.2229959518,0.0836435929,-0.3275434077,0.5394175053,0.0841736719,0.1905831844,-0.118047744,-0.2930147946,-0.2950969338,0.0054531754,-0.1695942879,-0.368024677,-0.1203953549,0.3059180677,0.3928041756,0.2445942461,0.1168810576,-0.2535528839,-0.1298053265,0.2650598884,-0.1438629478,-0.1589305401,-0.0732825994,0.0492410101,-0.2251886576,-0.1773677468,0.0832061097,-0.0194310006,0.0752256215,-0.2276214212,-0.1473570764,0.3186527193,-0.4347778261,0.2595769763,0.0258953478,0.5472237468,0.0291320775,-0.0422725081,-0.1655568033,0.222838372,-0.0450001284,0.4914217591,0.0616205968,0.2285088748,-0.215500474,-0.3188503683,-0.3452579677,0.631701231,0.0320965722,-0.0550996438,-0.2490802556,-0.0586701371,-0.2294994295,0.2923282981,-0.1192456484,0.0662930906,-0.0366520099,0.1867338717,-0.3581700027,-0.2367307842,0.3384183347,-0.4211325049,-0.3356799185,-0.189714089,0.0682086274,-0.1494337618,0.2929883599,-0.1057338789,-0.0153541863,0.2634702921,-0.1352269202,-0.0167236272,0.0001432241,0.0598571114,0.1573980451,-0.2458989173,-0.4044720829,-0.1943290979,-0.0603254028,-0.3799315095,-0.2492922992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1994","title":"not being able to get wikipedia es language","comments":"Hi @dorost1234, so sorry, but looking at my files here, I figure out that I've preprocessed files using the HF\/datasets for all the languages previously listed by me (Portuguese, Russian, French, Japanese, Chinese, and Turkish) except the Spanish (on my tests I've used the [wikicorpus](https:\/\/www.cs.upc.edu\/~nlp\/wikicorpus\/) instead).\r\n\r\nOnly with the Spanish Wikipedia's dump, I had the same `KeyError: '000nbsp'` problem already reported here https:\/\/github.com\/huggingface\/datasets\/issues\/577\r\n\r\nSo nowadays, even with access to a high resource machine, you couldn't be able to get Wikipedia's Spanish data using the HF\/datasets :(\r\n\r\n\r\n\r\n\r\n","body":"Hi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help ","comment_length":86,"text":"not being able to get wikipedia es language\nHi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help \nHi @dorost1234, so sorry, but looking at my files here, I figure out that I've preprocessed files using the HF\/datasets for all the languages previously listed by me (Portuguese, Russian, French, Japanese, Chinese, and Turkish) except the Spanish (on my tests I've used the [wikicorpus](https:\/\/www.cs.upc.edu\/~nlp\/wikicorpus\/) instead).\r\n\r\nOnly with the Spanish Wikipedia's dump, I had the same `KeyError: '000nbsp'` problem already reported here https:\/\/github.com\/huggingface\/datasets\/issues\/577\r\n\r\nSo nowadays, even with access to a high resource machine, you couldn't be able to get Wikipedia's Spanish data using the HF\/datasets :(\r\n\r\n\r\n\r\n\r\n","embeddings":[-0.342106998,0.0471930392,-0.1062649339,0.0620172322,0.200454846,0.1777444929,0.1768746823,0.3401965797,0.1373334974,0.0921178386,0.4105240405,0.3346937597,0.0160005912,0.3076457381,0.1113204286,-0.3370872438,0.0556506142,0.0752215236,-0.1947185397,-0.2001760751,-0.1320842654,0.1037431955,-0.2725183666,-0.0701157525,-0.1488354355,0.1070382744,0.1227778196,-0.0196700636,-0.192601651,-0.1862009615,0.0680222437,-0.036213655,0.2280335724,0.1079433188,-0.0001084146,0.0806313604,0.5717446804,-0.1583800614,-0.4590220749,-0.130322203,-0.081363827,-0.3462582827,0.2564504743,-0.3715744317,-0.3347848058,-0.0256394912,0.3187549412,-0.5857167244,0.2412877679,0.1479617506,0.2182685584,-0.074210979,0.3515349925,-0.0048845191,0.4105975032,0.218606472,-0.0360494033,-0.1243073717,0.0106633129,0.0249649677,0.0606173314,0.4618479013,0.0472859293,-0.1142473742,0.3039729893,-0.3094903827,0.2656840682,-0.3731164932,0.4229684472,0.2541317344,1.0161632299,-0.1197014451,0.0995212495,0.1050666645,0.0040745316,0.1287817061,0.2428365648,0.2005350441,-0.3642728925,-0.0796793103,0.1134936661,-0.2830438912,-0.335678637,0.3455581665,-0.1017287895,0.5368958116,0.1090375558,0.1416433603,-0.1509752721,-0.1852888316,0.0773017183,-0.0707799047,0.1030773669,0.3541413844,-0.0930447504,0.1351349652,0.2457793206,0.1751329154,0.0443397351,-0.0947965086,-0.2797648609,0.1917012036,0.368519485,0.0461870804,0.0220659487,0.0323531926,0.3556627333,-0.1849304587,0.2050388753,-0.030573098,0.0079394793,0.06664785,-0.0586212762,-0.3365755379,-0.6453832984,0.2325167805,-0.0009588756,-0.1367516071,0.192342177,0.1587269008,-0.3540838063,-0.2998664379,-0.0057489178,0.2017250508,0.1522044241,0.2051189095,0.2833256125,-0.0867259279,-0.3417105973,-0.3844312429,-0.0466403812,0.2829635143,-0.4593487978,0.0780525506,0.2093625367,0.1067860499,0.3750687838,-0.0355147943,-0.0439997055,0.0425268859,0.1307713985,0.0788350478,-0.1385989934,0.1594749987,0.1490190625,0.2470012158,0.199904263,-0.222467348,-0.052217897,0.0016842578,-0.0903736725,-0.1905186325,-0.18565543,0.1717608422,0.1227803901,0.1018858626,0.0458261073,0.3413800597,0.1713995934,-0.2258878946,0.1436647028,-0.0075095682,-0.1345748454,-0.0918872654,0.3624586463,0.372572124,-0.6563616395,0.1040466204,-0.0495637171,0.1877678782,0.143699944,-0.2795007229,-0.281387955,0.4237986207,-0.093520321,-0.1007723063,0.3089175224,-0.2708930373,0.0815908834,0.1968420744,0.067334421,-0.2822613418,0.2371014357,-0.1123532802,0.0422963984,0.1624549776,0.1141090319,0.1663445979,0.116460517,-0.0702252239,-0.2196758091,-0.1222946942,-0.0431540161,0.1172707975,0.2665446103,-0.1195803061,0.1249019429,0.5266057849,0.3261623085,-0.2632548511,0.0764246359,0.5432159305,-0.3122918904,0.0806591958,0.1918928772,-0.0554747656,-0.0498997234,0.1809134334,-0.239194259,0.4039117098,0.0790574178,-0.0053859027,-0.2668075264,-0.0530283935,-0.2360287756,-0.4283643067,0.2114921659,0.0466564186,-0.0355188176,0.2530859709,0.0529352501,-0.0900353342,-0.278223455,-0.1389373243,-0.8122121692,0.2934658229,-0.2860761285,-0.0800810456,-0.0875601247,0.0202554129,0.0639516488,-0.1236343607,-0.2867787182,0.1009152383,0.1873278469,0.0372007079,-0.011949122,0.16449368,0.1220298633,-0.2648282647,0.3166707158,0.3054229021,0.1199404374,0.0179688167,-0.2019145489,0.1108080298,0.0477327667,0.3355243504,0.0727865323,0.2685893774,0.185183093,0.1680476964,-0.1562812775,-0.1591639668,0.302156806,0.4226678014,0.1818974614,-0.145703271,0.0370389149,0.0772163868,0.451574564,0.1048435643,-0.0922244564,-0.1263718456,-0.3054101467,-0.0594792664,0.4833448827,-0.1791246086,-0.0160230976,0.2791057527,-0.0161974188,0.0771187916,-0.1511853486,-0.1175531745,0.4446229041,0.0723261684,0.5178593993,-0.0257208422,-0.1244166493,-0.1868189722,-0.1908410639,-0.2879425287,-0.1234138981,0.2847407758,-0.312468946,0.2768905461,-0.4178833961,-0.5642531514,-0.2084293067,0.4299700558,-0.3480241597,-0.3430698216,-0.0680083781,0.0778862834,-0.0229108091,0.2500188053,0.107210204,-0.035542842,0.2320447713,0.0074737356,-0.2216777354,-0.5956392884,-0.4356403053,0.0326633267,0.2823260725,0.0373752452,0.1272295564,-0.1569059193,-0.0672418326,-0.1452952772,-0.4021958709,0.3414442241,-0.0869849101,-0.0084638475,0.1166531071,0.4272188842,-0.19598113,-0.0950585008,0.3098010421,0.1152016073,-0.1476498395,-0.0692656934,-0.084672384,-0.0623619594,-0.0308230799,-0.4383021891,-0.1505855471,-0.3273377717,-0.1820004731,0.1689832062,0.1332353055,0.3523068726,0.2645772398,0.2091489136,0.3283595741,0.0954612941,-0.0965576619,-0.15933685,0.2170184106,-0.3185723126,-0.3167462349,0.1040425673,-0.1413134485,0.2305069566,0.1135405153,-0.146329999,0.0680313334,0.1159731671,0.0200549383,-0.1013204157,0.3257164955,0.7469013929,-0.0126242423,-0.0016661583,-0.0988114774,0.0478255115,0.0307381041,-0.4368053079,0.347318083,0.1777924001,0.5277687311,-0.0701474324,0.7515996695,0.1148355156,0.1166792214,0.1321653128,-0.0752609968,0.066186063,-0.272472769,-0.1884595752,0.292780757,0.0403270945,-0.3030905426,0.3898431957,-0.1277024895,-0.2598915696,-0.4387769699,0.0750479698,0.0015996028,-0.2447390854,-0.0376059711,-0.3750056326,0.3839118779,-0.0953029543,0.3992151022,-0.0802452266,-0.0731422156,0.0592001341,0.2196860462,-0.0982458889,0.1342238635,-0.1730971187,-0.2497496158,-0.3993260562,0.2461431772,0.0012104877,0.1436176598,-0.2927529216,0.0079169897,0.2831737101,0.0087449783,0.520098865,-0.647043407,-0.0973827019,0.3353388906,0.1921800226,-0.6571789384,-0.1413870305,-0.1876775622,0.2771171629,0.328502208,-0.0478469282,-0.3322197795,-0.1513579339,0.2680896223,0.0485692993,-0.0990989432,0.0515208356,-0.2417580783,-0.1433175504,-0.3353155851,-0.0657824352,-0.0996938273,0.1351074874,0.1002394855,0.2149961293,0.0813034996,0.0262187272,-0.1448780149,0.2050136924,0.1226052493,0.1383641809,-0.2560669482,0.1037125513,0.4806472957,-0.0265891049,-0.1080244482,0.1851131022,-0.2079540342,0.1726381034,-0.2150923759,0.1455085427,0.1604099125,-0.1206459701,-0.2393637598,0.0364794657,-0.2542999983,-0.0464653,0.0528427027,0.2087159753,0.0810022354,-0.1274007857,-0.4575788677,0.5609183908,0.1885291785,-0.0729709864,0.2990112603,-0.0156652872,-0.4306363761,0.4052890837,0.4329436719,0.8635653257,0.0299162623,0.0139987627,0.1714195311,0.1671361923,0.6254860759,-0.5466772914,0.209164381,-0.2812898755,0.3198871613,-0.0952548385,0.0653773472,0.3957326412,0.1431853473,-0.2221702784,0.3536064923,-0.0173848756,0.0741586462,0.056596484,0.3767177761,0.1265334934,-0.3078525662,-0.1175289601,0.1213964745,-0.1823625863,0.2824687064,-0.2498215139,-0.0125390952,-0.0145078488,-0.2449239939,-0.4107005596,0.0911933407,-0.2241836935,0.460855484,-0.3423030674,-0.4593057334,0.2506963611,0.2268087268,0.2354613245,0.3696662784,-0.3735500574,0.2001679838,0.01500485,-0.2361067981,-0.0761883631,0.0410975851,0.3866562247,-0.1922802478,-0.392267704,0.2118118405,-0.1738403141,-0.0978837684,-0.2912859023,-0.3177188933,0.3188324571,0.2502019703,0.1240342259,0.1725913435,-0.0860563293,-0.1338620186,0.1198045909,-0.284142524,0.0101039587,0.0086884415,0.0882760808,-0.0868183151,0.030908918,0.2571513057,0.1876173764,0.0172616523,0.5952085853,0.3785053194,-0.2194748968,-0.2659060359,0.1786083579,-0.2292437404,-0.0606894419,-0.2634168267,0.1508063525,0.0923178792,-0.0571265072,0.1029199064,-0.0157441776,-0.1409106106,-0.0071760365,-0.268037647,-0.0209581573,0.0222998895,-0.1638095528,-0.0015579133,0.1883054525,0.1332456172,0.2853651345,-0.2360227406,-0.2791166902,-0.238835454,-0.0223192759,0.0453209914,0.413713932,-0.0166400243,-0.2360471487,-0.038738668,0.1452317834,-0.2301054597,-0.2452716976,-0.1725503653,-0.0468346812,0.1290072203,-0.0397849046,-0.1874575168,0.135853678,-0.3403603733,-0.2925044596,-0.2041295767,-0.1316379607,-0.1293069422,-0.0242854469,0.2222887725,-0.1042095274,0.3449909389,-0.2877995074,0.0879830047,0.0348752849,0.2938313782,0.0511713922,0.2495486438,0.4168614745,-0.1764672995,-0.5602076054,0.1830450594,-0.2211462706,0.174384281,0.1633221805,-0.0828019083,-0.0480107516,0.0299655739,-0.0214739311,0.026401367,-0.0819890797,-0.0302336216,0.135588184,0.2045512199,-0.2827223837,0.1239569485,0.4220980108,-0.0330974609,0.0961894318,0.026279293,0.2615511715,-0.0534218065,0.1746627986,0.0410566963,0.0168639217,-0.0434479639,0.2639892101,0.0946073607,0.0811476633,0.1160609722,-0.120071359,0.046419654,0.1310725361,0.3746131957,0.1071262732,0.0660357401,0.1008156091,0.21656923,-0.2253761888,-0.1409864128,0.3062992394,-0.1027750447,-0.0209084284,0.0228774995,-0.064572908,0.0748565868,0.1480170488,-0.0637497306,-0.318022579,0.138672933,-0.0777273774,-0.1830140501,-0.6274288893,-0.2322814167,0.0546508804,-0.0176585633,0.0890892744,-0.0836972222,-0.1092474014,0.2026977241,-0.0350260288,-0.2306366116,0.6797729731,-0.0616487004,0.0446974188,-0.0551761761,0.2520872355,-0.2132354975,-0.0712455139,-0.1705531627,-0.0018399615,0.1578250974,0.210928753,-0.4701041579,-0.143992573,-0.0009005668,0.0189441014,-0.05917757,0.0006444135,0.0630011186,-0.0377971195,0.2940508723,0.1096709967,-0.0898015797,-0.1984426975,0.2169168144,-0.0071816775,-0.0200378653,-0.2484554499,0.0045184782,0.1042121798,-0.1066088304,0.1146246046,-0.5152721405,0.0413271375,0.2593996525,0.3884597719,-0.1085285619,-0.2229959518,0.0836435929,-0.3275434077,0.5394175053,0.0841736719,0.1905831844,-0.118047744,-0.2930147946,-0.2950969338,0.0054531754,-0.1695942879,-0.368024677,-0.1203953549,0.3059180677,0.3928041756,0.2445942461,0.1168810576,-0.2535528839,-0.1298053265,0.2650598884,-0.1438629478,-0.1589305401,-0.0732825994,0.0492410101,-0.2251886576,-0.1773677468,0.0832061097,-0.0194310006,0.0752256215,-0.2276214212,-0.1473570764,0.3186527193,-0.4347778261,0.2595769763,0.0258953478,0.5472237468,0.0291320775,-0.0422725081,-0.1655568033,0.222838372,-0.0450001284,0.4914217591,0.0616205968,0.2285088748,-0.215500474,-0.3188503683,-0.3452579677,0.631701231,0.0320965722,-0.0550996438,-0.2490802556,-0.0586701371,-0.2294994295,0.2923282981,-0.1192456484,0.0662930906,-0.0366520099,0.1867338717,-0.3581700027,-0.2367307842,0.3384183347,-0.4211325049,-0.3356799185,-0.189714089,0.0682086274,-0.1494337618,0.2929883599,-0.1057338789,-0.0153541863,0.2634702921,-0.1352269202,-0.0167236272,0.0001432241,0.0598571114,0.1573980451,-0.2458989173,-0.4044720829,-0.1943290979,-0.0603254028,-0.3799315095,-0.2492922992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1994","title":"not being able to get wikipedia es language","comments":"Thanks a lot for the information and help. This would be great to have\nthese datasets.\n@lhoestq Do you know a way I could get\nsmaller amount of these data like 1 GBtype of each language to deal with\ncomputatioanl requirements? thanks\n\nOn Sat, Mar 6, 2021 at 5:36 PM Jonatas Grosman \nwrote:\n\n> Hi @dorost1234 , so sorry, but looking at\n> my files here, I figure out that I've preprocessed files using the\n> HF\/datasets for all the languages previously listed by me (Portuguese,\n> Russian, French, Japanese, Chinese, and Turkish) except the Spanish (on my\n> tests I've used the wikicorpus \n> instead).\n>\n> Only with the Spanish Wikipedia's dump, I had the same KeyError: '000nbsp'\n> problem already reported here #577\n> \n>\n> So nowadays, even with access to a high resource machine, you couldn't be\n> able to get Wikipedia's Spanish data using the HF\/datasets :(\n>\n> \u2014\n> You are receiving this because you were mentioned.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","body":"Hi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help ","comment_length":189,"text":"not being able to get wikipedia es language\nHi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help \nThanks a lot for the information and help. This would be great to have\nthese datasets.\n@lhoestq Do you know a way I could get\nsmaller amount of these data like 1 GBtype of each language to deal with\ncomputatioanl requirements? thanks\n\nOn Sat, Mar 6, 2021 at 5:36 PM Jonatas Grosman \nwrote:\n\n> Hi @dorost1234 , so sorry, but looking at\n> my files here, I figure out that I've preprocessed files using the\n> HF\/datasets for all the languages previously listed by me (Portuguese,\n> Russian, French, Japanese, Chinese, and Turkish) except the Spanish (on my\n> tests I've used the wikicorpus \n> instead).\n>\n> Only with the Spanish Wikipedia's dump, I had the same KeyError: '000nbsp'\n> problem already reported here #577\n> \n>\n> So nowadays, even with access to a high resource machine, you couldn't be\n> able to get Wikipedia's Spanish data using the HF\/datasets :(\n>\n> \u2014\n> You are receiving this because you were mentioned.\n> Reply to this email directly, view it on GitHub\n> ,\n> or unsubscribe\n> \n> .\n>\n","embeddings":[-0.342106998,0.0471930392,-0.1062649339,0.0620172322,0.200454846,0.1777444929,0.1768746823,0.3401965797,0.1373334974,0.0921178386,0.4105240405,0.3346937597,0.0160005912,0.3076457381,0.1113204286,-0.3370872438,0.0556506142,0.0752215236,-0.1947185397,-0.2001760751,-0.1320842654,0.1037431955,-0.2725183666,-0.0701157525,-0.1488354355,0.1070382744,0.1227778196,-0.0196700636,-0.192601651,-0.1862009615,0.0680222437,-0.036213655,0.2280335724,0.1079433188,-0.0001084146,0.0806313604,0.5717446804,-0.1583800614,-0.4590220749,-0.130322203,-0.081363827,-0.3462582827,0.2564504743,-0.3715744317,-0.3347848058,-0.0256394912,0.3187549412,-0.5857167244,0.2412877679,0.1479617506,0.2182685584,-0.074210979,0.3515349925,-0.0048845191,0.4105975032,0.218606472,-0.0360494033,-0.1243073717,0.0106633129,0.0249649677,0.0606173314,0.4618479013,0.0472859293,-0.1142473742,0.3039729893,-0.3094903827,0.2656840682,-0.3731164932,0.4229684472,0.2541317344,1.0161632299,-0.1197014451,0.0995212495,0.1050666645,0.0040745316,0.1287817061,0.2428365648,0.2005350441,-0.3642728925,-0.0796793103,0.1134936661,-0.2830438912,-0.335678637,0.3455581665,-0.1017287895,0.5368958116,0.1090375558,0.1416433603,-0.1509752721,-0.1852888316,0.0773017183,-0.0707799047,0.1030773669,0.3541413844,-0.0930447504,0.1351349652,0.2457793206,0.1751329154,0.0443397351,-0.0947965086,-0.2797648609,0.1917012036,0.368519485,0.0461870804,0.0220659487,0.0323531926,0.3556627333,-0.1849304587,0.2050388753,-0.030573098,0.0079394793,0.06664785,-0.0586212762,-0.3365755379,-0.6453832984,0.2325167805,-0.0009588756,-0.1367516071,0.192342177,0.1587269008,-0.3540838063,-0.2998664379,-0.0057489178,0.2017250508,0.1522044241,0.2051189095,0.2833256125,-0.0867259279,-0.3417105973,-0.3844312429,-0.0466403812,0.2829635143,-0.4593487978,0.0780525506,0.2093625367,0.1067860499,0.3750687838,-0.0355147943,-0.0439997055,0.0425268859,0.1307713985,0.0788350478,-0.1385989934,0.1594749987,0.1490190625,0.2470012158,0.199904263,-0.222467348,-0.052217897,0.0016842578,-0.0903736725,-0.1905186325,-0.18565543,0.1717608422,0.1227803901,0.1018858626,0.0458261073,0.3413800597,0.1713995934,-0.2258878946,0.1436647028,-0.0075095682,-0.1345748454,-0.0918872654,0.3624586463,0.372572124,-0.6563616395,0.1040466204,-0.0495637171,0.1877678782,0.143699944,-0.2795007229,-0.281387955,0.4237986207,-0.093520321,-0.1007723063,0.3089175224,-0.2708930373,0.0815908834,0.1968420744,0.067334421,-0.2822613418,0.2371014357,-0.1123532802,0.0422963984,0.1624549776,0.1141090319,0.1663445979,0.116460517,-0.0702252239,-0.2196758091,-0.1222946942,-0.0431540161,0.1172707975,0.2665446103,-0.1195803061,0.1249019429,0.5266057849,0.3261623085,-0.2632548511,0.0764246359,0.5432159305,-0.3122918904,0.0806591958,0.1918928772,-0.0554747656,-0.0498997234,0.1809134334,-0.239194259,0.4039117098,0.0790574178,-0.0053859027,-0.2668075264,-0.0530283935,-0.2360287756,-0.4283643067,0.2114921659,0.0466564186,-0.0355188176,0.2530859709,0.0529352501,-0.0900353342,-0.278223455,-0.1389373243,-0.8122121692,0.2934658229,-0.2860761285,-0.0800810456,-0.0875601247,0.0202554129,0.0639516488,-0.1236343607,-0.2867787182,0.1009152383,0.1873278469,0.0372007079,-0.011949122,0.16449368,0.1220298633,-0.2648282647,0.3166707158,0.3054229021,0.1199404374,0.0179688167,-0.2019145489,0.1108080298,0.0477327667,0.3355243504,0.0727865323,0.2685893774,0.185183093,0.1680476964,-0.1562812775,-0.1591639668,0.302156806,0.4226678014,0.1818974614,-0.145703271,0.0370389149,0.0772163868,0.451574564,0.1048435643,-0.0922244564,-0.1263718456,-0.3054101467,-0.0594792664,0.4833448827,-0.1791246086,-0.0160230976,0.2791057527,-0.0161974188,0.0771187916,-0.1511853486,-0.1175531745,0.4446229041,0.0723261684,0.5178593993,-0.0257208422,-0.1244166493,-0.1868189722,-0.1908410639,-0.2879425287,-0.1234138981,0.2847407758,-0.312468946,0.2768905461,-0.4178833961,-0.5642531514,-0.2084293067,0.4299700558,-0.3480241597,-0.3430698216,-0.0680083781,0.0778862834,-0.0229108091,0.2500188053,0.107210204,-0.035542842,0.2320447713,0.0074737356,-0.2216777354,-0.5956392884,-0.4356403053,0.0326633267,0.2823260725,0.0373752452,0.1272295564,-0.1569059193,-0.0672418326,-0.1452952772,-0.4021958709,0.3414442241,-0.0869849101,-0.0084638475,0.1166531071,0.4272188842,-0.19598113,-0.0950585008,0.3098010421,0.1152016073,-0.1476498395,-0.0692656934,-0.084672384,-0.0623619594,-0.0308230799,-0.4383021891,-0.1505855471,-0.3273377717,-0.1820004731,0.1689832062,0.1332353055,0.3523068726,0.2645772398,0.2091489136,0.3283595741,0.0954612941,-0.0965576619,-0.15933685,0.2170184106,-0.3185723126,-0.3167462349,0.1040425673,-0.1413134485,0.2305069566,0.1135405153,-0.146329999,0.0680313334,0.1159731671,0.0200549383,-0.1013204157,0.3257164955,0.7469013929,-0.0126242423,-0.0016661583,-0.0988114774,0.0478255115,0.0307381041,-0.4368053079,0.347318083,0.1777924001,0.5277687311,-0.0701474324,0.7515996695,0.1148355156,0.1166792214,0.1321653128,-0.0752609968,0.066186063,-0.272472769,-0.1884595752,0.292780757,0.0403270945,-0.3030905426,0.3898431957,-0.1277024895,-0.2598915696,-0.4387769699,0.0750479698,0.0015996028,-0.2447390854,-0.0376059711,-0.3750056326,0.3839118779,-0.0953029543,0.3992151022,-0.0802452266,-0.0731422156,0.0592001341,0.2196860462,-0.0982458889,0.1342238635,-0.1730971187,-0.2497496158,-0.3993260562,0.2461431772,0.0012104877,0.1436176598,-0.2927529216,0.0079169897,0.2831737101,0.0087449783,0.520098865,-0.647043407,-0.0973827019,0.3353388906,0.1921800226,-0.6571789384,-0.1413870305,-0.1876775622,0.2771171629,0.328502208,-0.0478469282,-0.3322197795,-0.1513579339,0.2680896223,0.0485692993,-0.0990989432,0.0515208356,-0.2417580783,-0.1433175504,-0.3353155851,-0.0657824352,-0.0996938273,0.1351074874,0.1002394855,0.2149961293,0.0813034996,0.0262187272,-0.1448780149,0.2050136924,0.1226052493,0.1383641809,-0.2560669482,0.1037125513,0.4806472957,-0.0265891049,-0.1080244482,0.1851131022,-0.2079540342,0.1726381034,-0.2150923759,0.1455085427,0.1604099125,-0.1206459701,-0.2393637598,0.0364794657,-0.2542999983,-0.0464653,0.0528427027,0.2087159753,0.0810022354,-0.1274007857,-0.4575788677,0.5609183908,0.1885291785,-0.0729709864,0.2990112603,-0.0156652872,-0.4306363761,0.4052890837,0.4329436719,0.8635653257,0.0299162623,0.0139987627,0.1714195311,0.1671361923,0.6254860759,-0.5466772914,0.209164381,-0.2812898755,0.3198871613,-0.0952548385,0.0653773472,0.3957326412,0.1431853473,-0.2221702784,0.3536064923,-0.0173848756,0.0741586462,0.056596484,0.3767177761,0.1265334934,-0.3078525662,-0.1175289601,0.1213964745,-0.1823625863,0.2824687064,-0.2498215139,-0.0125390952,-0.0145078488,-0.2449239939,-0.4107005596,0.0911933407,-0.2241836935,0.460855484,-0.3423030674,-0.4593057334,0.2506963611,0.2268087268,0.2354613245,0.3696662784,-0.3735500574,0.2001679838,0.01500485,-0.2361067981,-0.0761883631,0.0410975851,0.3866562247,-0.1922802478,-0.392267704,0.2118118405,-0.1738403141,-0.0978837684,-0.2912859023,-0.3177188933,0.3188324571,0.2502019703,0.1240342259,0.1725913435,-0.0860563293,-0.1338620186,0.1198045909,-0.284142524,0.0101039587,0.0086884415,0.0882760808,-0.0868183151,0.030908918,0.2571513057,0.1876173764,0.0172616523,0.5952085853,0.3785053194,-0.2194748968,-0.2659060359,0.1786083579,-0.2292437404,-0.0606894419,-0.2634168267,0.1508063525,0.0923178792,-0.0571265072,0.1029199064,-0.0157441776,-0.1409106106,-0.0071760365,-0.268037647,-0.0209581573,0.0222998895,-0.1638095528,-0.0015579133,0.1883054525,0.1332456172,0.2853651345,-0.2360227406,-0.2791166902,-0.238835454,-0.0223192759,0.0453209914,0.413713932,-0.0166400243,-0.2360471487,-0.038738668,0.1452317834,-0.2301054597,-0.2452716976,-0.1725503653,-0.0468346812,0.1290072203,-0.0397849046,-0.1874575168,0.135853678,-0.3403603733,-0.2925044596,-0.2041295767,-0.1316379607,-0.1293069422,-0.0242854469,0.2222887725,-0.1042095274,0.3449909389,-0.2877995074,0.0879830047,0.0348752849,0.2938313782,0.0511713922,0.2495486438,0.4168614745,-0.1764672995,-0.5602076054,0.1830450594,-0.2211462706,0.174384281,0.1633221805,-0.0828019083,-0.0480107516,0.0299655739,-0.0214739311,0.026401367,-0.0819890797,-0.0302336216,0.135588184,0.2045512199,-0.2827223837,0.1239569485,0.4220980108,-0.0330974609,0.0961894318,0.026279293,0.2615511715,-0.0534218065,0.1746627986,0.0410566963,0.0168639217,-0.0434479639,0.2639892101,0.0946073607,0.0811476633,0.1160609722,-0.120071359,0.046419654,0.1310725361,0.3746131957,0.1071262732,0.0660357401,0.1008156091,0.21656923,-0.2253761888,-0.1409864128,0.3062992394,-0.1027750447,-0.0209084284,0.0228774995,-0.064572908,0.0748565868,0.1480170488,-0.0637497306,-0.318022579,0.138672933,-0.0777273774,-0.1830140501,-0.6274288893,-0.2322814167,0.0546508804,-0.0176585633,0.0890892744,-0.0836972222,-0.1092474014,0.2026977241,-0.0350260288,-0.2306366116,0.6797729731,-0.0616487004,0.0446974188,-0.0551761761,0.2520872355,-0.2132354975,-0.0712455139,-0.1705531627,-0.0018399615,0.1578250974,0.210928753,-0.4701041579,-0.143992573,-0.0009005668,0.0189441014,-0.05917757,0.0006444135,0.0630011186,-0.0377971195,0.2940508723,0.1096709967,-0.0898015797,-0.1984426975,0.2169168144,-0.0071816775,-0.0200378653,-0.2484554499,0.0045184782,0.1042121798,-0.1066088304,0.1146246046,-0.5152721405,0.0413271375,0.2593996525,0.3884597719,-0.1085285619,-0.2229959518,0.0836435929,-0.3275434077,0.5394175053,0.0841736719,0.1905831844,-0.118047744,-0.2930147946,-0.2950969338,0.0054531754,-0.1695942879,-0.368024677,-0.1203953549,0.3059180677,0.3928041756,0.2445942461,0.1168810576,-0.2535528839,-0.1298053265,0.2650598884,-0.1438629478,-0.1589305401,-0.0732825994,0.0492410101,-0.2251886576,-0.1773677468,0.0832061097,-0.0194310006,0.0752256215,-0.2276214212,-0.1473570764,0.3186527193,-0.4347778261,0.2595769763,0.0258953478,0.5472237468,0.0291320775,-0.0422725081,-0.1655568033,0.222838372,-0.0450001284,0.4914217591,0.0616205968,0.2285088748,-0.215500474,-0.3188503683,-0.3452579677,0.631701231,0.0320965722,-0.0550996438,-0.2490802556,-0.0586701371,-0.2294994295,0.2923282981,-0.1192456484,0.0662930906,-0.0366520099,0.1867338717,-0.3581700027,-0.2367307842,0.3384183347,-0.4211325049,-0.3356799185,-0.189714089,0.0682086274,-0.1494337618,0.2929883599,-0.1057338789,-0.0153541863,0.2634702921,-0.1352269202,-0.0167236272,0.0001432241,0.0598571114,0.1573980451,-0.2458989173,-0.4044720829,-0.1943290979,-0.0603254028,-0.3799315095,-0.2492922992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1994","title":"not being able to get wikipedia es language","comments":"Hi ! As mentioned above the Spanish configuration have parsing issues from `mwparserfromhell`. I haven't tested with the latest `mwparserfromhell` >=0.6 though. Which version of `mwparserfromhell` are you using ?\r\n\r\n> @lhoestq Could you help me to upload these preprocessed datasets to Huggingface's repositories? To be more precise, I've built datasets from the following languages using the 20201201 dumps: Spanish, Portuguese, Russian, French, Japanese, Chinese, and Turkish. Process these datasets have high costs that most of the community can't afford. I think these preprocessed datasets I have could be helpful for someone without access to high-resource machines to process Wikipedia's dumps like @dorost1234\r\n\r\nThat would be awesome ! Feel free to ping me on slack so we can put the processed wikipedia files on google storage with the other ones we've already preprocessed.\r\n\r\n> Do you know a way I could get smaller amount of these data like 1 GBtype of each language to deal with computatioanl requirements? thanks\r\n\r\nI'd suggest to copy the [wikipedia.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wikipedia\/wikipedia.py) to a new script `custom_wikipedia.py` and modify it to only download and process only a subset of the raw data files.\r\nYou can for example replace [this line](https:\/\/github.com\/huggingface\/datasets\/blob\/64e59fc45ca2134218b3e42e83fddddbe840ff74\/datasets\/wikipedia\/wikipedia.py#L446) by:\r\n```python\r\n if total_bytes >= (1 << 30): # stop if the total amount of data is >= 1GB\r\n break\r\n else:\r\n xml_urls.append(_base_url(lang) + fname)\r\n```\r\n\r\nThen you can load your custom wikipedia dataset with\r\n```python\r\nload_dataset(\"path\/to\/my\/custom_wikipedia.py\", f\"{date}.{language}\")\r\n```","body":"Hi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help ","comment_length":231,"text":"not being able to get wikipedia es language\nHi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help \nHi ! As mentioned above the Spanish configuration have parsing issues from `mwparserfromhell`. I haven't tested with the latest `mwparserfromhell` >=0.6 though. Which version of `mwparserfromhell` are you using ?\r\n\r\n> @lhoestq Could you help me to upload these preprocessed datasets to Huggingface's repositories? To be more precise, I've built datasets from the following languages using the 20201201 dumps: Spanish, Portuguese, Russian, French, Japanese, Chinese, and Turkish. Process these datasets have high costs that most of the community can't afford. I think these preprocessed datasets I have could be helpful for someone without access to high-resource machines to process Wikipedia's dumps like @dorost1234\r\n\r\nThat would be awesome ! Feel free to ping me on slack so we can put the processed wikipedia files on google storage with the other ones we've already preprocessed.\r\n\r\n> Do you know a way I could get smaller amount of these data like 1 GBtype of each language to deal with computatioanl requirements? thanks\r\n\r\nI'd suggest to copy the [wikipedia.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wikipedia\/wikipedia.py) to a new script `custom_wikipedia.py` and modify it to only download and process only a subset of the raw data files.\r\nYou can for example replace [this line](https:\/\/github.com\/huggingface\/datasets\/blob\/64e59fc45ca2134218b3e42e83fddddbe840ff74\/datasets\/wikipedia\/wikipedia.py#L446) by:\r\n```python\r\n if total_bytes >= (1 << 30): # stop if the total amount of data is >= 1GB\r\n break\r\n else:\r\n xml_urls.append(_base_url(lang) + fname)\r\n```\r\n\r\nThen you can load your custom wikipedia dataset with\r\n```python\r\nload_dataset(\"path\/to\/my\/custom_wikipedia.py\", f\"{date}.{language}\")\r\n```","embeddings":[-0.342106998,0.0471930392,-0.1062649339,0.0620172322,0.200454846,0.1777444929,0.1768746823,0.3401965797,0.1373334974,0.0921178386,0.4105240405,0.3346937597,0.0160005912,0.3076457381,0.1113204286,-0.3370872438,0.0556506142,0.0752215236,-0.1947185397,-0.2001760751,-0.1320842654,0.1037431955,-0.2725183666,-0.0701157525,-0.1488354355,0.1070382744,0.1227778196,-0.0196700636,-0.192601651,-0.1862009615,0.0680222437,-0.036213655,0.2280335724,0.1079433188,-0.0001084146,0.0806313604,0.5717446804,-0.1583800614,-0.4590220749,-0.130322203,-0.081363827,-0.3462582827,0.2564504743,-0.3715744317,-0.3347848058,-0.0256394912,0.3187549412,-0.5857167244,0.2412877679,0.1479617506,0.2182685584,-0.074210979,0.3515349925,-0.0048845191,0.4105975032,0.218606472,-0.0360494033,-0.1243073717,0.0106633129,0.0249649677,0.0606173314,0.4618479013,0.0472859293,-0.1142473742,0.3039729893,-0.3094903827,0.2656840682,-0.3731164932,0.4229684472,0.2541317344,1.0161632299,-0.1197014451,0.0995212495,0.1050666645,0.0040745316,0.1287817061,0.2428365648,0.2005350441,-0.3642728925,-0.0796793103,0.1134936661,-0.2830438912,-0.335678637,0.3455581665,-0.1017287895,0.5368958116,0.1090375558,0.1416433603,-0.1509752721,-0.1852888316,0.0773017183,-0.0707799047,0.1030773669,0.3541413844,-0.0930447504,0.1351349652,0.2457793206,0.1751329154,0.0443397351,-0.0947965086,-0.2797648609,0.1917012036,0.368519485,0.0461870804,0.0220659487,0.0323531926,0.3556627333,-0.1849304587,0.2050388753,-0.030573098,0.0079394793,0.06664785,-0.0586212762,-0.3365755379,-0.6453832984,0.2325167805,-0.0009588756,-0.1367516071,0.192342177,0.1587269008,-0.3540838063,-0.2998664379,-0.0057489178,0.2017250508,0.1522044241,0.2051189095,0.2833256125,-0.0867259279,-0.3417105973,-0.3844312429,-0.0466403812,0.2829635143,-0.4593487978,0.0780525506,0.2093625367,0.1067860499,0.3750687838,-0.0355147943,-0.0439997055,0.0425268859,0.1307713985,0.0788350478,-0.1385989934,0.1594749987,0.1490190625,0.2470012158,0.199904263,-0.222467348,-0.052217897,0.0016842578,-0.0903736725,-0.1905186325,-0.18565543,0.1717608422,0.1227803901,0.1018858626,0.0458261073,0.3413800597,0.1713995934,-0.2258878946,0.1436647028,-0.0075095682,-0.1345748454,-0.0918872654,0.3624586463,0.372572124,-0.6563616395,0.1040466204,-0.0495637171,0.1877678782,0.143699944,-0.2795007229,-0.281387955,0.4237986207,-0.093520321,-0.1007723063,0.3089175224,-0.2708930373,0.0815908834,0.1968420744,0.067334421,-0.2822613418,0.2371014357,-0.1123532802,0.0422963984,0.1624549776,0.1141090319,0.1663445979,0.116460517,-0.0702252239,-0.2196758091,-0.1222946942,-0.0431540161,0.1172707975,0.2665446103,-0.1195803061,0.1249019429,0.5266057849,0.3261623085,-0.2632548511,0.0764246359,0.5432159305,-0.3122918904,0.0806591958,0.1918928772,-0.0554747656,-0.0498997234,0.1809134334,-0.239194259,0.4039117098,0.0790574178,-0.0053859027,-0.2668075264,-0.0530283935,-0.2360287756,-0.4283643067,0.2114921659,0.0466564186,-0.0355188176,0.2530859709,0.0529352501,-0.0900353342,-0.278223455,-0.1389373243,-0.8122121692,0.2934658229,-0.2860761285,-0.0800810456,-0.0875601247,0.0202554129,0.0639516488,-0.1236343607,-0.2867787182,0.1009152383,0.1873278469,0.0372007079,-0.011949122,0.16449368,0.1220298633,-0.2648282647,0.3166707158,0.3054229021,0.1199404374,0.0179688167,-0.2019145489,0.1108080298,0.0477327667,0.3355243504,0.0727865323,0.2685893774,0.185183093,0.1680476964,-0.1562812775,-0.1591639668,0.302156806,0.4226678014,0.1818974614,-0.145703271,0.0370389149,0.0772163868,0.451574564,0.1048435643,-0.0922244564,-0.1263718456,-0.3054101467,-0.0594792664,0.4833448827,-0.1791246086,-0.0160230976,0.2791057527,-0.0161974188,0.0771187916,-0.1511853486,-0.1175531745,0.4446229041,0.0723261684,0.5178593993,-0.0257208422,-0.1244166493,-0.1868189722,-0.1908410639,-0.2879425287,-0.1234138981,0.2847407758,-0.312468946,0.2768905461,-0.4178833961,-0.5642531514,-0.2084293067,0.4299700558,-0.3480241597,-0.3430698216,-0.0680083781,0.0778862834,-0.0229108091,0.2500188053,0.107210204,-0.035542842,0.2320447713,0.0074737356,-0.2216777354,-0.5956392884,-0.4356403053,0.0326633267,0.2823260725,0.0373752452,0.1272295564,-0.1569059193,-0.0672418326,-0.1452952772,-0.4021958709,0.3414442241,-0.0869849101,-0.0084638475,0.1166531071,0.4272188842,-0.19598113,-0.0950585008,0.3098010421,0.1152016073,-0.1476498395,-0.0692656934,-0.084672384,-0.0623619594,-0.0308230799,-0.4383021891,-0.1505855471,-0.3273377717,-0.1820004731,0.1689832062,0.1332353055,0.3523068726,0.2645772398,0.2091489136,0.3283595741,0.0954612941,-0.0965576619,-0.15933685,0.2170184106,-0.3185723126,-0.3167462349,0.1040425673,-0.1413134485,0.2305069566,0.1135405153,-0.146329999,0.0680313334,0.1159731671,0.0200549383,-0.1013204157,0.3257164955,0.7469013929,-0.0126242423,-0.0016661583,-0.0988114774,0.0478255115,0.0307381041,-0.4368053079,0.347318083,0.1777924001,0.5277687311,-0.0701474324,0.7515996695,0.1148355156,0.1166792214,0.1321653128,-0.0752609968,0.066186063,-0.272472769,-0.1884595752,0.292780757,0.0403270945,-0.3030905426,0.3898431957,-0.1277024895,-0.2598915696,-0.4387769699,0.0750479698,0.0015996028,-0.2447390854,-0.0376059711,-0.3750056326,0.3839118779,-0.0953029543,0.3992151022,-0.0802452266,-0.0731422156,0.0592001341,0.2196860462,-0.0982458889,0.1342238635,-0.1730971187,-0.2497496158,-0.3993260562,0.2461431772,0.0012104877,0.1436176598,-0.2927529216,0.0079169897,0.2831737101,0.0087449783,0.520098865,-0.647043407,-0.0973827019,0.3353388906,0.1921800226,-0.6571789384,-0.1413870305,-0.1876775622,0.2771171629,0.328502208,-0.0478469282,-0.3322197795,-0.1513579339,0.2680896223,0.0485692993,-0.0990989432,0.0515208356,-0.2417580783,-0.1433175504,-0.3353155851,-0.0657824352,-0.0996938273,0.1351074874,0.1002394855,0.2149961293,0.0813034996,0.0262187272,-0.1448780149,0.2050136924,0.1226052493,0.1383641809,-0.2560669482,0.1037125513,0.4806472957,-0.0265891049,-0.1080244482,0.1851131022,-0.2079540342,0.1726381034,-0.2150923759,0.1455085427,0.1604099125,-0.1206459701,-0.2393637598,0.0364794657,-0.2542999983,-0.0464653,0.0528427027,0.2087159753,0.0810022354,-0.1274007857,-0.4575788677,0.5609183908,0.1885291785,-0.0729709864,0.2990112603,-0.0156652872,-0.4306363761,0.4052890837,0.4329436719,0.8635653257,0.0299162623,0.0139987627,0.1714195311,0.1671361923,0.6254860759,-0.5466772914,0.209164381,-0.2812898755,0.3198871613,-0.0952548385,0.0653773472,0.3957326412,0.1431853473,-0.2221702784,0.3536064923,-0.0173848756,0.0741586462,0.056596484,0.3767177761,0.1265334934,-0.3078525662,-0.1175289601,0.1213964745,-0.1823625863,0.2824687064,-0.2498215139,-0.0125390952,-0.0145078488,-0.2449239939,-0.4107005596,0.0911933407,-0.2241836935,0.460855484,-0.3423030674,-0.4593057334,0.2506963611,0.2268087268,0.2354613245,0.3696662784,-0.3735500574,0.2001679838,0.01500485,-0.2361067981,-0.0761883631,0.0410975851,0.3866562247,-0.1922802478,-0.392267704,0.2118118405,-0.1738403141,-0.0978837684,-0.2912859023,-0.3177188933,0.3188324571,0.2502019703,0.1240342259,0.1725913435,-0.0860563293,-0.1338620186,0.1198045909,-0.284142524,0.0101039587,0.0086884415,0.0882760808,-0.0868183151,0.030908918,0.2571513057,0.1876173764,0.0172616523,0.5952085853,0.3785053194,-0.2194748968,-0.2659060359,0.1786083579,-0.2292437404,-0.0606894419,-0.2634168267,0.1508063525,0.0923178792,-0.0571265072,0.1029199064,-0.0157441776,-0.1409106106,-0.0071760365,-0.268037647,-0.0209581573,0.0222998895,-0.1638095528,-0.0015579133,0.1883054525,0.1332456172,0.2853651345,-0.2360227406,-0.2791166902,-0.238835454,-0.0223192759,0.0453209914,0.413713932,-0.0166400243,-0.2360471487,-0.038738668,0.1452317834,-0.2301054597,-0.2452716976,-0.1725503653,-0.0468346812,0.1290072203,-0.0397849046,-0.1874575168,0.135853678,-0.3403603733,-0.2925044596,-0.2041295767,-0.1316379607,-0.1293069422,-0.0242854469,0.2222887725,-0.1042095274,0.3449909389,-0.2877995074,0.0879830047,0.0348752849,0.2938313782,0.0511713922,0.2495486438,0.4168614745,-0.1764672995,-0.5602076054,0.1830450594,-0.2211462706,0.174384281,0.1633221805,-0.0828019083,-0.0480107516,0.0299655739,-0.0214739311,0.026401367,-0.0819890797,-0.0302336216,0.135588184,0.2045512199,-0.2827223837,0.1239569485,0.4220980108,-0.0330974609,0.0961894318,0.026279293,0.2615511715,-0.0534218065,0.1746627986,0.0410566963,0.0168639217,-0.0434479639,0.2639892101,0.0946073607,0.0811476633,0.1160609722,-0.120071359,0.046419654,0.1310725361,0.3746131957,0.1071262732,0.0660357401,0.1008156091,0.21656923,-0.2253761888,-0.1409864128,0.3062992394,-0.1027750447,-0.0209084284,0.0228774995,-0.064572908,0.0748565868,0.1480170488,-0.0637497306,-0.318022579,0.138672933,-0.0777273774,-0.1830140501,-0.6274288893,-0.2322814167,0.0546508804,-0.0176585633,0.0890892744,-0.0836972222,-0.1092474014,0.2026977241,-0.0350260288,-0.2306366116,0.6797729731,-0.0616487004,0.0446974188,-0.0551761761,0.2520872355,-0.2132354975,-0.0712455139,-0.1705531627,-0.0018399615,0.1578250974,0.210928753,-0.4701041579,-0.143992573,-0.0009005668,0.0189441014,-0.05917757,0.0006444135,0.0630011186,-0.0377971195,0.2940508723,0.1096709967,-0.0898015797,-0.1984426975,0.2169168144,-0.0071816775,-0.0200378653,-0.2484554499,0.0045184782,0.1042121798,-0.1066088304,0.1146246046,-0.5152721405,0.0413271375,0.2593996525,0.3884597719,-0.1085285619,-0.2229959518,0.0836435929,-0.3275434077,0.5394175053,0.0841736719,0.1905831844,-0.118047744,-0.2930147946,-0.2950969338,0.0054531754,-0.1695942879,-0.368024677,-0.1203953549,0.3059180677,0.3928041756,0.2445942461,0.1168810576,-0.2535528839,-0.1298053265,0.2650598884,-0.1438629478,-0.1589305401,-0.0732825994,0.0492410101,-0.2251886576,-0.1773677468,0.0832061097,-0.0194310006,0.0752256215,-0.2276214212,-0.1473570764,0.3186527193,-0.4347778261,0.2595769763,0.0258953478,0.5472237468,0.0291320775,-0.0422725081,-0.1655568033,0.222838372,-0.0450001284,0.4914217591,0.0616205968,0.2285088748,-0.215500474,-0.3188503683,-0.3452579677,0.631701231,0.0320965722,-0.0550996438,-0.2490802556,-0.0586701371,-0.2294994295,0.2923282981,-0.1192456484,0.0662930906,-0.0366520099,0.1867338717,-0.3581700027,-0.2367307842,0.3384183347,-0.4211325049,-0.3356799185,-0.189714089,0.0682086274,-0.1494337618,0.2929883599,-0.1057338789,-0.0153541863,0.2634702921,-0.1352269202,-0.0167236272,0.0001432241,0.0598571114,0.1573980451,-0.2458989173,-0.4044720829,-0.1943290979,-0.0603254028,-0.3799315095,-0.2492922992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1994","title":"not being able to get wikipedia es language","comments":"Hi @lhoestq!\r\n\r\n> Hi ! As mentioned above the Spanish configuration have parsing issues from mwparserfromhell. I haven't tested with the latest mwparserfromhell >=0.6 though. Which version of mwparserfromhell are you using ?\r\n\r\nI'm using the latest mwparserfromhell version (0.6)\r\n\r\n> That would be awesome ! Feel free to ping me on slack so we can put the processed wikipedia files on google storage with the other ones we've already preprocessed.\r\n\r\nI'll ping you there \ud83d\udc4d ","body":"Hi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help ","comment_length":76,"text":"not being able to get wikipedia es language\nHi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help \nHi @lhoestq!\r\n\r\n> Hi ! As mentioned above the Spanish configuration have parsing issues from mwparserfromhell. I haven't tested with the latest mwparserfromhell >=0.6 though. Which version of mwparserfromhell are you using ?\r\n\r\nI'm using the latest mwparserfromhell version (0.6)\r\n\r\n> That would be awesome ! Feel free to ping me on slack so we can put the processed wikipedia files on google storage with the other ones we've already preprocessed.\r\n\r\nI'll ping you there \ud83d\udc4d ","embeddings":[-0.342106998,0.0471930392,-0.1062649339,0.0620172322,0.200454846,0.1777444929,0.1768746823,0.3401965797,0.1373334974,0.0921178386,0.4105240405,0.3346937597,0.0160005912,0.3076457381,0.1113204286,-0.3370872438,0.0556506142,0.0752215236,-0.1947185397,-0.2001760751,-0.1320842654,0.1037431955,-0.2725183666,-0.0701157525,-0.1488354355,0.1070382744,0.1227778196,-0.0196700636,-0.192601651,-0.1862009615,0.0680222437,-0.036213655,0.2280335724,0.1079433188,-0.0001084146,0.0806313604,0.5717446804,-0.1583800614,-0.4590220749,-0.130322203,-0.081363827,-0.3462582827,0.2564504743,-0.3715744317,-0.3347848058,-0.0256394912,0.3187549412,-0.5857167244,0.2412877679,0.1479617506,0.2182685584,-0.074210979,0.3515349925,-0.0048845191,0.4105975032,0.218606472,-0.0360494033,-0.1243073717,0.0106633129,0.0249649677,0.0606173314,0.4618479013,0.0472859293,-0.1142473742,0.3039729893,-0.3094903827,0.2656840682,-0.3731164932,0.4229684472,0.2541317344,1.0161632299,-0.1197014451,0.0995212495,0.1050666645,0.0040745316,0.1287817061,0.2428365648,0.2005350441,-0.3642728925,-0.0796793103,0.1134936661,-0.2830438912,-0.335678637,0.3455581665,-0.1017287895,0.5368958116,0.1090375558,0.1416433603,-0.1509752721,-0.1852888316,0.0773017183,-0.0707799047,0.1030773669,0.3541413844,-0.0930447504,0.1351349652,0.2457793206,0.1751329154,0.0443397351,-0.0947965086,-0.2797648609,0.1917012036,0.368519485,0.0461870804,0.0220659487,0.0323531926,0.3556627333,-0.1849304587,0.2050388753,-0.030573098,0.0079394793,0.06664785,-0.0586212762,-0.3365755379,-0.6453832984,0.2325167805,-0.0009588756,-0.1367516071,0.192342177,0.1587269008,-0.3540838063,-0.2998664379,-0.0057489178,0.2017250508,0.1522044241,0.2051189095,0.2833256125,-0.0867259279,-0.3417105973,-0.3844312429,-0.0466403812,0.2829635143,-0.4593487978,0.0780525506,0.2093625367,0.1067860499,0.3750687838,-0.0355147943,-0.0439997055,0.0425268859,0.1307713985,0.0788350478,-0.1385989934,0.1594749987,0.1490190625,0.2470012158,0.199904263,-0.222467348,-0.052217897,0.0016842578,-0.0903736725,-0.1905186325,-0.18565543,0.1717608422,0.1227803901,0.1018858626,0.0458261073,0.3413800597,0.1713995934,-0.2258878946,0.1436647028,-0.0075095682,-0.1345748454,-0.0918872654,0.3624586463,0.372572124,-0.6563616395,0.1040466204,-0.0495637171,0.1877678782,0.143699944,-0.2795007229,-0.281387955,0.4237986207,-0.093520321,-0.1007723063,0.3089175224,-0.2708930373,0.0815908834,0.1968420744,0.067334421,-0.2822613418,0.2371014357,-0.1123532802,0.0422963984,0.1624549776,0.1141090319,0.1663445979,0.116460517,-0.0702252239,-0.2196758091,-0.1222946942,-0.0431540161,0.1172707975,0.2665446103,-0.1195803061,0.1249019429,0.5266057849,0.3261623085,-0.2632548511,0.0764246359,0.5432159305,-0.3122918904,0.0806591958,0.1918928772,-0.0554747656,-0.0498997234,0.1809134334,-0.239194259,0.4039117098,0.0790574178,-0.0053859027,-0.2668075264,-0.0530283935,-0.2360287756,-0.4283643067,0.2114921659,0.0466564186,-0.0355188176,0.2530859709,0.0529352501,-0.0900353342,-0.278223455,-0.1389373243,-0.8122121692,0.2934658229,-0.2860761285,-0.0800810456,-0.0875601247,0.0202554129,0.0639516488,-0.1236343607,-0.2867787182,0.1009152383,0.1873278469,0.0372007079,-0.011949122,0.16449368,0.1220298633,-0.2648282647,0.3166707158,0.3054229021,0.1199404374,0.0179688167,-0.2019145489,0.1108080298,0.0477327667,0.3355243504,0.0727865323,0.2685893774,0.185183093,0.1680476964,-0.1562812775,-0.1591639668,0.302156806,0.4226678014,0.1818974614,-0.145703271,0.0370389149,0.0772163868,0.451574564,0.1048435643,-0.0922244564,-0.1263718456,-0.3054101467,-0.0594792664,0.4833448827,-0.1791246086,-0.0160230976,0.2791057527,-0.0161974188,0.0771187916,-0.1511853486,-0.1175531745,0.4446229041,0.0723261684,0.5178593993,-0.0257208422,-0.1244166493,-0.1868189722,-0.1908410639,-0.2879425287,-0.1234138981,0.2847407758,-0.312468946,0.2768905461,-0.4178833961,-0.5642531514,-0.2084293067,0.4299700558,-0.3480241597,-0.3430698216,-0.0680083781,0.0778862834,-0.0229108091,0.2500188053,0.107210204,-0.035542842,0.2320447713,0.0074737356,-0.2216777354,-0.5956392884,-0.4356403053,0.0326633267,0.2823260725,0.0373752452,0.1272295564,-0.1569059193,-0.0672418326,-0.1452952772,-0.4021958709,0.3414442241,-0.0869849101,-0.0084638475,0.1166531071,0.4272188842,-0.19598113,-0.0950585008,0.3098010421,0.1152016073,-0.1476498395,-0.0692656934,-0.084672384,-0.0623619594,-0.0308230799,-0.4383021891,-0.1505855471,-0.3273377717,-0.1820004731,0.1689832062,0.1332353055,0.3523068726,0.2645772398,0.2091489136,0.3283595741,0.0954612941,-0.0965576619,-0.15933685,0.2170184106,-0.3185723126,-0.3167462349,0.1040425673,-0.1413134485,0.2305069566,0.1135405153,-0.146329999,0.0680313334,0.1159731671,0.0200549383,-0.1013204157,0.3257164955,0.7469013929,-0.0126242423,-0.0016661583,-0.0988114774,0.0478255115,0.0307381041,-0.4368053079,0.347318083,0.1777924001,0.5277687311,-0.0701474324,0.7515996695,0.1148355156,0.1166792214,0.1321653128,-0.0752609968,0.066186063,-0.272472769,-0.1884595752,0.292780757,0.0403270945,-0.3030905426,0.3898431957,-0.1277024895,-0.2598915696,-0.4387769699,0.0750479698,0.0015996028,-0.2447390854,-0.0376059711,-0.3750056326,0.3839118779,-0.0953029543,0.3992151022,-0.0802452266,-0.0731422156,0.0592001341,0.2196860462,-0.0982458889,0.1342238635,-0.1730971187,-0.2497496158,-0.3993260562,0.2461431772,0.0012104877,0.1436176598,-0.2927529216,0.0079169897,0.2831737101,0.0087449783,0.520098865,-0.647043407,-0.0973827019,0.3353388906,0.1921800226,-0.6571789384,-0.1413870305,-0.1876775622,0.2771171629,0.328502208,-0.0478469282,-0.3322197795,-0.1513579339,0.2680896223,0.0485692993,-0.0990989432,0.0515208356,-0.2417580783,-0.1433175504,-0.3353155851,-0.0657824352,-0.0996938273,0.1351074874,0.1002394855,0.2149961293,0.0813034996,0.0262187272,-0.1448780149,0.2050136924,0.1226052493,0.1383641809,-0.2560669482,0.1037125513,0.4806472957,-0.0265891049,-0.1080244482,0.1851131022,-0.2079540342,0.1726381034,-0.2150923759,0.1455085427,0.1604099125,-0.1206459701,-0.2393637598,0.0364794657,-0.2542999983,-0.0464653,0.0528427027,0.2087159753,0.0810022354,-0.1274007857,-0.4575788677,0.5609183908,0.1885291785,-0.0729709864,0.2990112603,-0.0156652872,-0.4306363761,0.4052890837,0.4329436719,0.8635653257,0.0299162623,0.0139987627,0.1714195311,0.1671361923,0.6254860759,-0.5466772914,0.209164381,-0.2812898755,0.3198871613,-0.0952548385,0.0653773472,0.3957326412,0.1431853473,-0.2221702784,0.3536064923,-0.0173848756,0.0741586462,0.056596484,0.3767177761,0.1265334934,-0.3078525662,-0.1175289601,0.1213964745,-0.1823625863,0.2824687064,-0.2498215139,-0.0125390952,-0.0145078488,-0.2449239939,-0.4107005596,0.0911933407,-0.2241836935,0.460855484,-0.3423030674,-0.4593057334,0.2506963611,0.2268087268,0.2354613245,0.3696662784,-0.3735500574,0.2001679838,0.01500485,-0.2361067981,-0.0761883631,0.0410975851,0.3866562247,-0.1922802478,-0.392267704,0.2118118405,-0.1738403141,-0.0978837684,-0.2912859023,-0.3177188933,0.3188324571,0.2502019703,0.1240342259,0.1725913435,-0.0860563293,-0.1338620186,0.1198045909,-0.284142524,0.0101039587,0.0086884415,0.0882760808,-0.0868183151,0.030908918,0.2571513057,0.1876173764,0.0172616523,0.5952085853,0.3785053194,-0.2194748968,-0.2659060359,0.1786083579,-0.2292437404,-0.0606894419,-0.2634168267,0.1508063525,0.0923178792,-0.0571265072,0.1029199064,-0.0157441776,-0.1409106106,-0.0071760365,-0.268037647,-0.0209581573,0.0222998895,-0.1638095528,-0.0015579133,0.1883054525,0.1332456172,0.2853651345,-0.2360227406,-0.2791166902,-0.238835454,-0.0223192759,0.0453209914,0.413713932,-0.0166400243,-0.2360471487,-0.038738668,0.1452317834,-0.2301054597,-0.2452716976,-0.1725503653,-0.0468346812,0.1290072203,-0.0397849046,-0.1874575168,0.135853678,-0.3403603733,-0.2925044596,-0.2041295767,-0.1316379607,-0.1293069422,-0.0242854469,0.2222887725,-0.1042095274,0.3449909389,-0.2877995074,0.0879830047,0.0348752849,0.2938313782,0.0511713922,0.2495486438,0.4168614745,-0.1764672995,-0.5602076054,0.1830450594,-0.2211462706,0.174384281,0.1633221805,-0.0828019083,-0.0480107516,0.0299655739,-0.0214739311,0.026401367,-0.0819890797,-0.0302336216,0.135588184,0.2045512199,-0.2827223837,0.1239569485,0.4220980108,-0.0330974609,0.0961894318,0.026279293,0.2615511715,-0.0534218065,0.1746627986,0.0410566963,0.0168639217,-0.0434479639,0.2639892101,0.0946073607,0.0811476633,0.1160609722,-0.120071359,0.046419654,0.1310725361,0.3746131957,0.1071262732,0.0660357401,0.1008156091,0.21656923,-0.2253761888,-0.1409864128,0.3062992394,-0.1027750447,-0.0209084284,0.0228774995,-0.064572908,0.0748565868,0.1480170488,-0.0637497306,-0.318022579,0.138672933,-0.0777273774,-0.1830140501,-0.6274288893,-0.2322814167,0.0546508804,-0.0176585633,0.0890892744,-0.0836972222,-0.1092474014,0.2026977241,-0.0350260288,-0.2306366116,0.6797729731,-0.0616487004,0.0446974188,-0.0551761761,0.2520872355,-0.2132354975,-0.0712455139,-0.1705531627,-0.0018399615,0.1578250974,0.210928753,-0.4701041579,-0.143992573,-0.0009005668,0.0189441014,-0.05917757,0.0006444135,0.0630011186,-0.0377971195,0.2940508723,0.1096709967,-0.0898015797,-0.1984426975,0.2169168144,-0.0071816775,-0.0200378653,-0.2484554499,0.0045184782,0.1042121798,-0.1066088304,0.1146246046,-0.5152721405,0.0413271375,0.2593996525,0.3884597719,-0.1085285619,-0.2229959518,0.0836435929,-0.3275434077,0.5394175053,0.0841736719,0.1905831844,-0.118047744,-0.2930147946,-0.2950969338,0.0054531754,-0.1695942879,-0.368024677,-0.1203953549,0.3059180677,0.3928041756,0.2445942461,0.1168810576,-0.2535528839,-0.1298053265,0.2650598884,-0.1438629478,-0.1589305401,-0.0732825994,0.0492410101,-0.2251886576,-0.1773677468,0.0832061097,-0.0194310006,0.0752256215,-0.2276214212,-0.1473570764,0.3186527193,-0.4347778261,0.2595769763,0.0258953478,0.5472237468,0.0291320775,-0.0422725081,-0.1655568033,0.222838372,-0.0450001284,0.4914217591,0.0616205968,0.2285088748,-0.215500474,-0.3188503683,-0.3452579677,0.631701231,0.0320965722,-0.0550996438,-0.2490802556,-0.0586701371,-0.2294994295,0.2923282981,-0.1192456484,0.0662930906,-0.0366520099,0.1867338717,-0.3581700027,-0.2367307842,0.3384183347,-0.4211325049,-0.3356799185,-0.189714089,0.0682086274,-0.1494337618,0.2929883599,-0.1057338789,-0.0153541863,0.2634702921,-0.1352269202,-0.0167236272,0.0001432241,0.0598571114,0.1573980451,-0.2458989173,-0.4044720829,-0.1943290979,-0.0603254028,-0.3799315095,-0.2492922992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1994","title":"not being able to get wikipedia es language","comments":"Thank you so much @jonatasgrosman and @lhoestq this would be a great help. I am really thankful to you both and to wonderful Huggingface dataset library allowing us to train models at scale.","body":"Hi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help ","comment_length":33,"text":"not being able to get wikipedia es language\nHi\r\nI am trying to run a code with wikipedia of config 20200501.es, getting:\r\n\r\nTraceback (most recent call last):\r\n File \"run_mlm_t5.py\", line 608, in \r\n main()\r\n File \"run_mlm_t5.py\", line 359, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/libs\/anaconda3\/envs\/success432\/lib\/python3.7\/site-packages\/datasets-1.2.1-py3.7.egg\/datasets\/builder.py\", line 1050, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\ndatasets.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\nthanks @lhoestq for any suggestion\/help \nThank you so much @jonatasgrosman and @lhoestq this would be a great help. I am really thankful to you both and to wonderful Huggingface dataset library allowing us to train models at scale.","embeddings":[-0.342106998,0.0471930392,-0.1062649339,0.0620172322,0.200454846,0.1777444929,0.1768746823,0.3401965797,0.1373334974,0.0921178386,0.4105240405,0.3346937597,0.0160005912,0.3076457381,0.1113204286,-0.3370872438,0.0556506142,0.0752215236,-0.1947185397,-0.2001760751,-0.1320842654,0.1037431955,-0.2725183666,-0.0701157525,-0.1488354355,0.1070382744,0.1227778196,-0.0196700636,-0.192601651,-0.1862009615,0.0680222437,-0.036213655,0.2280335724,0.1079433188,-0.0001084146,0.0806313604,0.5717446804,-0.1583800614,-0.4590220749,-0.130322203,-0.081363827,-0.3462582827,0.2564504743,-0.3715744317,-0.3347848058,-0.0256394912,0.3187549412,-0.5857167244,0.2412877679,0.1479617506,0.2182685584,-0.074210979,0.3515349925,-0.0048845191,0.4105975032,0.218606472,-0.0360494033,-0.1243073717,0.0106633129,0.0249649677,0.0606173314,0.4618479013,0.0472859293,-0.1142473742,0.3039729893,-0.3094903827,0.2656840682,-0.3731164932,0.4229684472,0.2541317344,1.0161632299,-0.1197014451,0.0995212495,0.1050666645,0.0040745316,0.1287817061,0.2428365648,0.2005350441,-0.3642728925,-0.0796793103,0.1134936661,-0.2830438912,-0.335678637,0.3455581665,-0.1017287895,0.5368958116,0.1090375558,0.1416433603,-0.1509752721,-0.1852888316,0.0773017183,-0.0707799047,0.1030773669,0.3541413844,-0.0930447504,0.1351349652,0.2457793206,0.1751329154,0.0443397351,-0.0947965086,-0.2797648609,0.1917012036,0.368519485,0.0461870804,0.0220659487,0.0323531926,0.3556627333,-0.1849304587,0.2050388753,-0.030573098,0.0079394793,0.06664785,-0.0586212762,-0.3365755379,-0.6453832984,0.2325167805,-0.0009588756,-0.1367516071,0.192342177,0.1587269008,-0.3540838063,-0.2998664379,-0.0057489178,0.2017250508,0.1522044241,0.2051189095,0.2833256125,-0.0867259279,-0.3417105973,-0.3844312429,-0.0466403812,0.2829635143,-0.4593487978,0.0780525506,0.2093625367,0.1067860499,0.3750687838,-0.0355147943,-0.0439997055,0.0425268859,0.1307713985,0.0788350478,-0.1385989934,0.1594749987,0.1490190625,0.2470012158,0.199904263,-0.222467348,-0.052217897,0.0016842578,-0.0903736725,-0.1905186325,-0.18565543,0.1717608422,0.1227803901,0.1018858626,0.0458261073,0.3413800597,0.1713995934,-0.2258878946,0.1436647028,-0.0075095682,-0.1345748454,-0.0918872654,0.3624586463,0.372572124,-0.6563616395,0.1040466204,-0.0495637171,0.1877678782,0.143699944,-0.2795007229,-0.281387955,0.4237986207,-0.093520321,-0.1007723063,0.3089175224,-0.2708930373,0.0815908834,0.1968420744,0.067334421,-0.2822613418,0.2371014357,-0.1123532802,0.0422963984,0.1624549776,0.1141090319,0.1663445979,0.116460517,-0.0702252239,-0.2196758091,-0.1222946942,-0.0431540161,0.1172707975,0.2665446103,-0.1195803061,0.1249019429,0.5266057849,0.3261623085,-0.2632548511,0.0764246359,0.5432159305,-0.3122918904,0.0806591958,0.1918928772,-0.0554747656,-0.0498997234,0.1809134334,-0.239194259,0.4039117098,0.0790574178,-0.0053859027,-0.2668075264,-0.0530283935,-0.2360287756,-0.4283643067,0.2114921659,0.0466564186,-0.0355188176,0.2530859709,0.0529352501,-0.0900353342,-0.278223455,-0.1389373243,-0.8122121692,0.2934658229,-0.2860761285,-0.0800810456,-0.0875601247,0.0202554129,0.0639516488,-0.1236343607,-0.2867787182,0.1009152383,0.1873278469,0.0372007079,-0.011949122,0.16449368,0.1220298633,-0.2648282647,0.3166707158,0.3054229021,0.1199404374,0.0179688167,-0.2019145489,0.1108080298,0.0477327667,0.3355243504,0.0727865323,0.2685893774,0.185183093,0.1680476964,-0.1562812775,-0.1591639668,0.302156806,0.4226678014,0.1818974614,-0.145703271,0.0370389149,0.0772163868,0.451574564,0.1048435643,-0.0922244564,-0.1263718456,-0.3054101467,-0.0594792664,0.4833448827,-0.1791246086,-0.0160230976,0.2791057527,-0.0161974188,0.0771187916,-0.1511853486,-0.1175531745,0.4446229041,0.0723261684,0.5178593993,-0.0257208422,-0.1244166493,-0.1868189722,-0.1908410639,-0.2879425287,-0.1234138981,0.2847407758,-0.312468946,0.2768905461,-0.4178833961,-0.5642531514,-0.2084293067,0.4299700558,-0.3480241597,-0.3430698216,-0.0680083781,0.0778862834,-0.0229108091,0.2500188053,0.107210204,-0.035542842,0.2320447713,0.0074737356,-0.2216777354,-0.5956392884,-0.4356403053,0.0326633267,0.2823260725,0.0373752452,0.1272295564,-0.1569059193,-0.0672418326,-0.1452952772,-0.4021958709,0.3414442241,-0.0869849101,-0.0084638475,0.1166531071,0.4272188842,-0.19598113,-0.0950585008,0.3098010421,0.1152016073,-0.1476498395,-0.0692656934,-0.084672384,-0.0623619594,-0.0308230799,-0.4383021891,-0.1505855471,-0.3273377717,-0.1820004731,0.1689832062,0.1332353055,0.3523068726,0.2645772398,0.2091489136,0.3283595741,0.0954612941,-0.0965576619,-0.15933685,0.2170184106,-0.3185723126,-0.3167462349,0.1040425673,-0.1413134485,0.2305069566,0.1135405153,-0.146329999,0.0680313334,0.1159731671,0.0200549383,-0.1013204157,0.3257164955,0.7469013929,-0.0126242423,-0.0016661583,-0.0988114774,0.0478255115,0.0307381041,-0.4368053079,0.347318083,0.1777924001,0.5277687311,-0.0701474324,0.7515996695,0.1148355156,0.1166792214,0.1321653128,-0.0752609968,0.066186063,-0.272472769,-0.1884595752,0.292780757,0.0403270945,-0.3030905426,0.3898431957,-0.1277024895,-0.2598915696,-0.4387769699,0.0750479698,0.0015996028,-0.2447390854,-0.0376059711,-0.3750056326,0.3839118779,-0.0953029543,0.3992151022,-0.0802452266,-0.0731422156,0.0592001341,0.2196860462,-0.0982458889,0.1342238635,-0.1730971187,-0.2497496158,-0.3993260562,0.2461431772,0.0012104877,0.1436176598,-0.2927529216,0.0079169897,0.2831737101,0.0087449783,0.520098865,-0.647043407,-0.0973827019,0.3353388906,0.1921800226,-0.6571789384,-0.1413870305,-0.1876775622,0.2771171629,0.328502208,-0.0478469282,-0.3322197795,-0.1513579339,0.2680896223,0.0485692993,-0.0990989432,0.0515208356,-0.2417580783,-0.1433175504,-0.3353155851,-0.0657824352,-0.0996938273,0.1351074874,0.1002394855,0.2149961293,0.0813034996,0.0262187272,-0.1448780149,0.2050136924,0.1226052493,0.1383641809,-0.2560669482,0.1037125513,0.4806472957,-0.0265891049,-0.1080244482,0.1851131022,-0.2079540342,0.1726381034,-0.2150923759,0.1455085427,0.1604099125,-0.1206459701,-0.2393637598,0.0364794657,-0.2542999983,-0.0464653,0.0528427027,0.2087159753,0.0810022354,-0.1274007857,-0.4575788677,0.5609183908,0.1885291785,-0.0729709864,0.2990112603,-0.0156652872,-0.4306363761,0.4052890837,0.4329436719,0.8635653257,0.0299162623,0.0139987627,0.1714195311,0.1671361923,0.6254860759,-0.5466772914,0.209164381,-0.2812898755,0.3198871613,-0.0952548385,0.0653773472,0.3957326412,0.1431853473,-0.2221702784,0.3536064923,-0.0173848756,0.0741586462,0.056596484,0.3767177761,0.1265334934,-0.3078525662,-0.1175289601,0.1213964745,-0.1823625863,0.2824687064,-0.2498215139,-0.0125390952,-0.0145078488,-0.2449239939,-0.4107005596,0.0911933407,-0.2241836935,0.460855484,-0.3423030674,-0.4593057334,0.2506963611,0.2268087268,0.2354613245,0.3696662784,-0.3735500574,0.2001679838,0.01500485,-0.2361067981,-0.0761883631,0.0410975851,0.3866562247,-0.1922802478,-0.392267704,0.2118118405,-0.1738403141,-0.0978837684,-0.2912859023,-0.3177188933,0.3188324571,0.2502019703,0.1240342259,0.1725913435,-0.0860563293,-0.1338620186,0.1198045909,-0.284142524,0.0101039587,0.0086884415,0.0882760808,-0.0868183151,0.030908918,0.2571513057,0.1876173764,0.0172616523,0.5952085853,0.3785053194,-0.2194748968,-0.2659060359,0.1786083579,-0.2292437404,-0.0606894419,-0.2634168267,0.1508063525,0.0923178792,-0.0571265072,0.1029199064,-0.0157441776,-0.1409106106,-0.0071760365,-0.268037647,-0.0209581573,0.0222998895,-0.1638095528,-0.0015579133,0.1883054525,0.1332456172,0.2853651345,-0.2360227406,-0.2791166902,-0.238835454,-0.0223192759,0.0453209914,0.413713932,-0.0166400243,-0.2360471487,-0.038738668,0.1452317834,-0.2301054597,-0.2452716976,-0.1725503653,-0.0468346812,0.1290072203,-0.0397849046,-0.1874575168,0.135853678,-0.3403603733,-0.2925044596,-0.2041295767,-0.1316379607,-0.1293069422,-0.0242854469,0.2222887725,-0.1042095274,0.3449909389,-0.2877995074,0.0879830047,0.0348752849,0.2938313782,0.0511713922,0.2495486438,0.4168614745,-0.1764672995,-0.5602076054,0.1830450594,-0.2211462706,0.174384281,0.1633221805,-0.0828019083,-0.0480107516,0.0299655739,-0.0214739311,0.026401367,-0.0819890797,-0.0302336216,0.135588184,0.2045512199,-0.2827223837,0.1239569485,0.4220980108,-0.0330974609,0.0961894318,0.026279293,0.2615511715,-0.0534218065,0.1746627986,0.0410566963,0.0168639217,-0.0434479639,0.2639892101,0.0946073607,0.0811476633,0.1160609722,-0.120071359,0.046419654,0.1310725361,0.3746131957,0.1071262732,0.0660357401,0.1008156091,0.21656923,-0.2253761888,-0.1409864128,0.3062992394,-0.1027750447,-0.0209084284,0.0228774995,-0.064572908,0.0748565868,0.1480170488,-0.0637497306,-0.318022579,0.138672933,-0.0777273774,-0.1830140501,-0.6274288893,-0.2322814167,0.0546508804,-0.0176585633,0.0890892744,-0.0836972222,-0.1092474014,0.2026977241,-0.0350260288,-0.2306366116,0.6797729731,-0.0616487004,0.0446974188,-0.0551761761,0.2520872355,-0.2132354975,-0.0712455139,-0.1705531627,-0.0018399615,0.1578250974,0.210928753,-0.4701041579,-0.143992573,-0.0009005668,0.0189441014,-0.05917757,0.0006444135,0.0630011186,-0.0377971195,0.2940508723,0.1096709967,-0.0898015797,-0.1984426975,0.2169168144,-0.0071816775,-0.0200378653,-0.2484554499,0.0045184782,0.1042121798,-0.1066088304,0.1146246046,-0.5152721405,0.0413271375,0.2593996525,0.3884597719,-0.1085285619,-0.2229959518,0.0836435929,-0.3275434077,0.5394175053,0.0841736719,0.1905831844,-0.118047744,-0.2930147946,-0.2950969338,0.0054531754,-0.1695942879,-0.368024677,-0.1203953549,0.3059180677,0.3928041756,0.2445942461,0.1168810576,-0.2535528839,-0.1298053265,0.2650598884,-0.1438629478,-0.1589305401,-0.0732825994,0.0492410101,-0.2251886576,-0.1773677468,0.0832061097,-0.0194310006,0.0752256215,-0.2276214212,-0.1473570764,0.3186527193,-0.4347778261,0.2595769763,0.0258953478,0.5472237468,0.0291320775,-0.0422725081,-0.1655568033,0.222838372,-0.0450001284,0.4914217591,0.0616205968,0.2285088748,-0.215500474,-0.3188503683,-0.3452579677,0.631701231,0.0320965722,-0.0550996438,-0.2490802556,-0.0586701371,-0.2294994295,0.2923282981,-0.1192456484,0.0662930906,-0.0366520099,0.1867338717,-0.3581700027,-0.2367307842,0.3384183347,-0.4211325049,-0.3356799185,-0.189714089,0.0682086274,-0.1494337618,0.2929883599,-0.1057338789,-0.0153541863,0.2634702921,-0.1352269202,-0.0167236272,0.0001432241,0.0598571114,0.1573980451,-0.2458989173,-0.4044720829,-0.1943290979,-0.0603254028,-0.3799315095,-0.2492922992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1993","title":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? ","comments":"Hi ! That looks like a bug, can you provide some code so that we can reproduce ?\r\nIt's not supposed to update the original dataset","body":"I am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n","comment_length":26,"text":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? \nI am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n\nHi ! That looks like a bug, can you provide some code so that we can reproduce ?\r\nIt's not supposed to update the original dataset","embeddings":[-0.2773600221,-0.1424060762,-0.0097123682,0.2224389315,0.234768346,0.1286469102,-0.0314267799,-0.0836327448,-0.0445274338,0.0425948873,0.0756771415,0.3508275449,0.0656664595,0.2150926739,0.1894967705,0.2755098641,0.30281201,0.2871473432,-0.3395662904,-0.1187531799,-0.2247807235,-0.1271229088,0.0034543565,-0.2915343642,-0.1811131388,-0.2847261727,-0.1051794589,0.0112861879,0.119211264,0.1282495111,0.1472534239,0.1170899719,0.2598932385,0.3208320737,-0.0001200398,0.0065589393,-0.2820394039,-0.0981556326,-0.3171302676,-0.148311317,0.0040694536,-0.1686504632,0.0125795854,-0.3187098205,0.0167110711,-0.0577575341,-0.1532112658,-0.2047884464,0.5439517498,-0.2619331181,0.107594505,-0.0745562315,-0.1875942945,0.0535860099,-0.2637374401,0.3224839568,0.1635244787,0.0492271073,-0.0166150704,0.3577507436,0.0076645566,0.2626109421,-0.215463087,-0.257628262,0.3756664693,-0.0119469641,0.1762693822,-0.3532520235,0.1566132456,-0.0655312389,0.7306799293,-0.4183009267,-0.2183706611,-0.2186258584,0.1446022838,-0.1294228882,0.1498940885,0.2500557303,-0.0096549811,0.1328632683,-0.501339376,-0.7865973115,-0.0961471647,0.0907199755,0.0513803437,-0.5484752655,0.0074869199,0.1316185743,0.1669691652,0.3084483445,0.4444078207,-0.2608914077,-0.2427158803,0.1032831594,-0.2903403044,-0.3080198169,-0.3247923255,0.1850420833,-0.0413478985,0.2232471555,0.0851408988,-0.0896956027,-0.2581596971,0.140354529,0.1931944191,0.2976649702,0.1033943594,0.2071948647,0.0640979707,0.0867727399,-0.2276673764,-0.0555465035,0.2099663466,0.111686863,0.6322048306,-0.3062034249,0.2605336905,0.0181004014,0.0575665571,0.1173090264,0.1301251799,-0.0381411798,-0.1821733564,0.3218336105,0.10125117,0.107065022,0.1898788363,0.2406492531,-0.0559981689,-0.1312303096,-0.1708797216,-0.0795024782,-0.3020848334,0.3862603605,0.0738743469,0.0053746332,0.1103697419,0.4839027524,-0.4521849155,-0.1569833606,0.0093465056,-0.1823616028,0.3648672998,0.1807547212,0.1329004467,0.1929761022,-0.0767004862,-0.025980575,-0.0506211706,0.7035810351,-0.3622858822,-0.1167384014,0.0317438245,0.128875643,-0.0997810215,0.1189190745,-0.6574715376,0.1957009584,0.1090387627,-0.229483515,0.2326243818,0.0644701794,-0.3933758736,-0.1955350786,-0.0600752197,0.2446706295,-0.3517104983,-0.1330956817,0.1671473235,-0.2163995355,-0.0562942438,0.1603805125,-0.2772349715,0.2107772827,-0.3416086137,-0.3182988167,0.5734582543,0.1237904876,-0.4386877716,0.1376541108,-0.0473181829,-0.0854982883,-0.0338457376,0.5759631991,0.0887061656,-0.1218519211,-0.4468078017,0.2835900486,0.0914309174,0.0993331671,-0.0000944741,-0.1719219536,0.2446641624,-0.3138799071,-0.1098827347,0.4869318604,0.2913513482,0.4426918924,0.3346465826,0.0546466149,0.1226766333,0.4205866456,-0.1444275081,0.0363451801,-0.1889337897,0.1604404896,-0.6050271392,-0.0283062197,0.1515417248,-0.5714516044,0.4265675843,-0.096628435,-0.059457887,-0.2061415613,-0.0825063586,0.0565844364,-0.0742596239,0.0540457033,-0.0702320561,-0.1691200435,-0.2333442271,0.4565918446,-0.0745811686,0.1019803956,-0.6434781551,0.4120730758,0.1448139399,-0.1876072884,-0.2897032797,-0.0770050138,0.2022823542,0.0158426929,-0.1604731083,0.3794608116,0.1969830245,0.3693661988,-0.0724361613,0.1969150007,0.0275861826,-0.036629796,0.0015106014,-0.1320199221,0.1783733815,-0.0849591792,-0.1314108521,-0.0251700878,-0.1762411743,0.0567238703,0.069346033,-0.1900815219,0.0065394868,-0.3226765394,-0.0129018389,-0.2311117649,-0.2205683291,0.194451198,0.3617491722,0.1074925661,-0.1966824383,-0.1728535444,0.2807909548,-0.2587944269,-0.0464546494,0.0443932414,-0.2743505239,0.0095037175,-0.1260944903,0.5707713366,0.189101249,0.1061206013,0.0777842626,0.1742013842,0.1364916712,-0.0232198015,0.0245520435,0.0353943706,0.5409010649,0.2572120428,-0.0064349682,0.0857307538,-0.1593874991,0.4665906131,0.1098956019,-0.1207867116,-0.30450508,0.1523127854,-0.1394732744,0.0934548825,-0.4907522798,-0.085855253,-0.2307418883,0.1629548967,-0.2530109882,0.7551643252,0.206769675,0.1947494149,0.1520796716,0.0233585779,-0.0885055289,-0.4377730787,-0.0893533155,0.1271415651,-0.0993080139,-0.0616395026,0.0731258243,-0.0676445961,0.3914980888,-0.0824567303,-0.0544498526,-0.4722400308,0.067814678,-0.0471551158,0.0970695764,0.117943354,-0.1721117944,0.1620834172,-0.4026908576,-0.015497786,-0.1130088717,-0.4427279234,-0.2376411706,0.0469758548,0.1358089149,0.2856687307,-0.0207176786,-0.176030457,0.031320788,-0.0240333863,-0.164477542,-0.179878816,0.0381252915,0.0707304925,0.1785485744,-0.249313131,0.2252117842,0.0366107188,-0.2901349962,-0.5891330838,0.3810376525,0.1769789606,-0.2777751386,0.2593784332,-0.1830547303,-0.0385802872,0.3359547853,-0.6108106971,-0.0890948251,0.0473068841,0.0712167919,-0.4067156315,0.3266804516,0.3233913779,-0.110593535,0.0090249265,-0.1059059873,-0.2592627406,0.0587809496,0.0994874462,0.4200102687,-0.2023087293,0.2221485525,-0.081459567,0.2542732358,-0.0070602391,-0.1169082448,0.3640681207,0.1134620085,0.5133032799,-0.244579643,-0.1410063654,-0.5643542409,-0.1771263927,-0.2818339467,-0.0114473673,-0.0227353889,-0.0442445949,-0.1466464698,-0.2983010113,-0.2498422563,-0.1560127884,-0.0004163511,-0.2736831307,0.3052914441,0.1021622643,0.2262859643,0.1891043037,-0.2032574266,-0.0671015382,0.2009994686,0.3148538768,0.0471924022,-0.3226694763,-0.030612817,-0.1401808858,0.1583727747,-0.208939597,-0.0048104455,0.1612966955,-0.064239122,0.0978768021,0.209305957,0.8259220719,0.2282086164,0.1617141664,0.0667234883,-0.387262851,-0.2855741084,-0.0050819525,0.0198310036,-0.2780990303,-0.0880926251,0.6831374168,0.1787525117,-0.0507892556,-0.2957937419,0.4659903646,-0.3318034708,-0.3347007036,-0.2603489459,-0.1991749555,-0.5212270617,-0.1314112097,-0.2619122267,-0.0284724608,-0.1773988008,0.1137886047,-0.1871513575,-0.0758932307,0.0513755381,-0.1798712015,0.3351958394,0.3329252005,0.0540519357,0.0823942274,0.1873522848,0.171730563,0.3534220159,-0.0996237844,-0.0959055945,-0.0483041778,0.0681137368,0.1413040757,0.0803858265,0.0522795469,0.0050397064,-0.0257186946,-0.2154927403,-0.3657272756,-0.085214518,-0.0878766328,-0.1379946917,-0.2146567404,-0.4565048218,0.5085780621,-0.0864706561,-0.0846951678,-0.0616627708,-0.0977539942,-0.2706187665,0.1074461788,0.118212007,0.8301039934,0.166253984,0.2188905925,0.18120341,-0.246858418,0.036708001,0.0289868359,-0.0480845422,-0.3911790848,-0.34594661,-0.151596114,-0.1716965586,0.0086912597,0.3013535738,-0.346003145,0.3065209091,-0.2786551416,-0.0542787462,-0.1268519014,-0.023213027,-0.16166614,-0.1308979541,-0.1627981067,0.0030435522,-0.0186123233,0.0688089132,-0.0729737803,0.0433277115,0.1778715998,-0.1023885161,0.3078854978,0.1381435543,-0.0280983765,0.2491618097,-0.1203390136,-0.4506568313,-0.0506093651,0.5732136965,0.4336514175,0.0530103892,0.0687489361,-0.0989170969,-0.0141947055,0.3902993202,-0.1978017986,-0.1549128592,0.4320685565,0.1253273785,-0.0773228183,0.0669242442,-0.0534172058,-0.1615175754,-0.0730297565,-0.0121712238,0.0610182881,-0.3188638985,0.0215416849,0.2385422587,0.3357427418,-0.1063526198,0.0380113833,0.0047739153,-0.0055090785,0.2677955925,0.0356959514,-0.2979891002,0.0134071177,0.5746378899,0.2181512862,-0.0201751478,0.4256415367,-0.3335492611,0.124863334,-0.2270946801,0.5075287223,0.1834751219,-0.2038392872,0.2045847178,0.038367141,0.4741452932,-0.0272540078,-0.0167224593,0.0738130286,-0.007936894,0.1543698162,-0.0486600287,-0.3517208993,0.0910364464,0.228633821,0.3515547514,0.34589535,-0.1389693022,0.0248692892,0.3707372546,-0.2032517791,0.1789879054,0.2704664469,0.2227460593,0.2163436711,0.2249105573,-0.3069205582,-0.3214671016,-0.0353057235,-0.193394959,-0.2613629401,-0.1477644891,-0.0689215213,0.1421384364,-0.0735849962,0.2744117379,-0.2740536034,-0.428692162,0.1984905005,-0.1260866225,0.1044978276,0.1898788512,0.2151736617,0.2136215866,0.1381159276,0.4634352922,-0.271399796,-0.1352324933,-0.2096988261,0.1733743548,-0.0220643319,-0.0064741611,0.1258606166,-0.1178697422,-0.624211967,-0.0673956424,0.5685315132,0.1167619303,-0.0426341817,-0.2280732393,0.0550832525,0.1555532664,0.2226059139,0.11807473,-0.4023568034,-0.3365201652,0.31684196,0.1549126208,-0.079256326,-0.191256538,0.2755452991,-0.1716838926,-0.3012566566,0.1850246638,0.1411788464,-0.0526258908,-0.3393751383,0.0854596123,0.2024499625,-0.1482756436,0.4022917747,0.6510437131,-0.2748965025,0.0234390311,0.1474159509,0.3377691209,-0.1483326554,0.4305054545,-0.0011163068,0.4770657122,-0.005231956,0.0808773413,-0.1037396491,-0.4036500454,0.0769736394,0.1380400211,-0.2588012218,0.0405354276,-0.0300862137,-0.1536857486,0.1886304617,-0.3479964733,-0.1480426788,0.3714606464,-0.241190657,0.1074004248,-0.0134340487,-0.1359696984,0.2627212405,0.2455383241,-0.0099924998,-0.2409632802,0.3795967996,0.0855144411,-0.1817884594,0.1468252838,0.0894483328,0.2030430883,-0.1944673657,-0.1617188007,0.0693570748,-0.123747997,0.0871067494,0.087450549,0.0389341675,-0.1845514923,0.1138078421,0.2477380633,0.2041704208,0.2740829289,-0.0373133346,0.1291216463,0.1125145033,-0.0035213539,0.0093158316,0.2051873654,-0.000748948,0.0038972194,0.304861486,0.0947670713,0.3823273778,-0.4848358929,0.2595471442,0.2937276363,-0.0746517479,-0.2495839894,0.042300947,0.1689215004,0.0155459605,0.5528270006,-0.2686411142,0.0056018415,0.0508064441,0.0381750092,0.0514433458,0.4367139339,0.2386621982,-0.4496119618,-0.2709541023,-0.2319019735,-0.5596634746,0.018425094,0.3477916121,0.3131184578,-0.2493397295,0.1025708169,0.0763872489,0.165216431,0.0482027903,-0.1767886281,0.0744339898,0.3412543237,0.1767937243,-0.2534859776,-0.0234439503,0.2920185328,-0.1093696058,-0.2514945567,0.1846382767,0.0725314096,-0.0073188818,0.1785518378,0.170284614,-0.2235817462,0.4020435512,0.6338937879,0.0853700638,0.0235492848,-0.1998801082,0.2453515232,0.3206823766,-0.0403755791,-0.1977472156,-0.0849019736,0.0871070176,0.1833114922,-0.3971991539,0.3609264493,-0.2117165029,-0.1129459739,0.1277288496,0.0651453808,-0.5358834267,0.049110353,0.0694136396,0.1766885519,0.1368408352,0.3822607398,0.0120541407,0.3451385498,-0.1944079399,-0.1288726777,0.3826500773,-0.1838705689,-0.3120285571,-0.182034716,0.3298672438,-0.0515807681,-0.1072218195,-0.3974069953,0.1122050211,0.4550141096,-0.2239272743,0.012750919,0.5110885501,-0.2035415322,0.1770402044,0.1363914758,0.5885228515,0.1271755546,-0.3351356983,-0.0040458902,-0.3007531762]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1993","title":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? ","comments":"Hi, I experimented with RAG. \r\n\r\nActually, you can run the [use_own_knowldge_dataset.py](https:\/\/github.com\/shamanez\/transformers\/blob\/rag-end-to-end-retrieval\/examples\/research_projects\/rag\/use_own_knowledge_dataset.py#L80). In the 80 you can save the dataset object to the disk with save_to_disk. Then in order to compute the embeddings in this use **load_from_disk**. \r\n\r\nThen finally save it. You can see the original dataset object (CSV after splitting also will be changed)\r\n\r\nOne more thing- when I save the dataset object with **save_to_disk** it name the arrow file with cache.... rather than using dataset. arrow. Can you add a variable that we can feed a name to save_to_disk function?","body":"I am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n","comment_length":91,"text":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? \nI am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n\nHi, I experimented with RAG. \r\n\r\nActually, you can run the [use_own_knowldge_dataset.py](https:\/\/github.com\/shamanez\/transformers\/blob\/rag-end-to-end-retrieval\/examples\/research_projects\/rag\/use_own_knowledge_dataset.py#L80). In the 80 you can save the dataset object to the disk with save_to_disk. Then in order to compute the embeddings in this use **load_from_disk**. \r\n\r\nThen finally save it. You can see the original dataset object (CSV after splitting also will be changed)\r\n\r\nOne more thing- when I save the dataset object with **save_to_disk** it name the arrow file with cache.... rather than using dataset. arrow. Can you add a variable that we can feed a name to save_to_disk function?","embeddings":[-0.2981736064,-0.0828652084,0.0267273188,0.1493998915,0.3067896664,0.1228007376,-0.0562063009,-0.0614770278,0.0445269383,0.0541939586,-0.0652618036,0.373824507,0.0215632059,0.1011979878,0.1474010199,0.2532070875,0.1987998039,0.2516977489,-0.2688197494,-0.0971940234,-0.2197251469,-0.2309300452,0.0529332533,-0.3456068635,-0.2162120938,-0.2678647637,-0.1700974256,0.0145703144,0.1032246277,-0.0398215204,0.1639603972,0.1300251782,0.3564484715,0.2765147686,-0.0001222565,0.0449472219,-0.3267861307,-0.1661964506,-0.3974144757,-0.1439903378,0.1779921055,-0.2638602853,0.0647072569,-0.3373181522,-0.0446428657,-0.1765089631,-0.107605584,-0.2636471987,0.6995682716,-0.3083574772,0.0656189248,-0.054051958,-0.2422589958,0.0614366159,-0.0760419965,0.2449961752,0.1191361323,0.1519682109,-0.0484969094,0.2226030082,-0.1404744536,0.2548884451,-0.2061859518,-0.1893691421,0.3967930377,-0.0019805473,0.0755400509,-0.3313621581,0.0782873556,-0.0002817264,0.7876380682,-0.4693832099,-0.3323039114,-0.2304031104,0.2201095819,-0.0707611889,0.043340873,0.206916675,0.0135489209,0.2221802473,-0.4472397566,-0.8105933666,-0.1399762779,0.1488463879,0.0819213465,-0.4945161641,0.0238880273,0.1483843029,0.2077623159,0.289745003,0.372831434,-0.254982233,-0.174267292,0.1039844602,-0.3044407666,-0.2441856712,-0.2922518253,0.0569520146,-0.0547048226,0.3812344372,0.1349292845,-0.0329647623,-0.2412497103,0.1737683564,0.1737943292,0.2897120714,0.1109258682,0.1706047058,0.1379611641,0.0049738353,-0.2106245607,-0.1055162549,0.1187181026,0.1315823942,0.5387070775,-0.2161186934,0.2105357051,0.04565515,0.1234001964,0.0126175126,0.0549343489,-0.0481300615,-0.1775926948,0.3871252239,0.0820179284,0.1812818646,0.1506492347,0.2366101146,0.0228765868,-0.1150496453,-0.1261828989,-0.034810666,-0.2900650203,0.3913136125,0.20096232,-0.0995993018,0.1868816465,0.4064928889,-0.434612304,-0.1599060297,0.0141043393,-0.2374427021,0.5390252471,0.1750427932,0.0522038862,0.2025613487,0.0051055625,-0.0455207899,-0.1667996198,0.519553721,-0.2719989121,-0.1811145395,0.1025095806,0.0902323723,-0.1660820544,0.2113942653,-0.4340537488,0.1076642126,0.0573433675,-0.1293478906,0.1663846523,0.0311017428,-0.3366870284,-0.278573066,0.0628452227,0.3072368205,-0.4218387306,-0.1641575843,0.0618918389,-0.0940397754,-0.2057061642,0.2003053725,-0.2843594849,0.2975263,-0.2760423422,-0.2422124594,0.5749604106,-0.0287337955,-0.481277436,0.1532562077,0.0253595226,0.0056969882,0.0016571193,0.6175714731,0.3191406429,-0.1399003863,-0.4540526867,0.3999417424,0.1578346789,0.1049262434,0.0218707919,-0.2957147658,0.2183943689,-0.2868964076,-0.1871246696,0.3673587143,0.2825940847,0.2646553814,0.3615919352,0.0234334562,0.1554546952,0.3602969646,0.0220470112,0.0340864584,-0.1074186936,0.052772861,-0.6752665043,-0.0008339359,0.1554773301,-0.5374922752,0.1736091971,-0.1934855133,-0.0924271867,-0.165345639,-0.0885926932,-0.0135686425,-0.0897683352,-0.0308930166,-0.0441634394,-0.2119160444,-0.2746407092,0.4201807082,-0.0376425385,0.1762324125,-0.6207967401,0.4170429111,0.1973149329,-0.2141876221,-0.1617540121,-0.0837895796,0.2295793295,0.0185931753,-0.0512569696,0.3639732897,0.1689340323,0.372076273,0.0582000352,0.1233947948,0.0281126741,-0.1681797951,0.2201291025,-0.0435152948,0.1642446816,-0.1232943609,-0.2530221641,0.0886408612,-0.1890542954,0.1372634321,0.086750105,-0.1663254499,-0.0118455375,-0.3778166175,-0.0842535794,-0.2564434409,-0.2976899743,0.1848404706,0.445897907,0.0632535964,0.0249878038,-0.1737427562,0.316331625,-0.1756477356,0.0119170323,0.0152700162,-0.3919644356,-0.0695018098,-0.1718242615,0.4888994694,0.1010778844,0.0891606286,0.0631394237,0.1061022803,0.1201624349,0.0404632613,0.0341845043,0.0189376473,0.5688123703,0.3238120377,0.0262624957,0.0392558388,-0.1586344689,0.4841207862,0.1311215311,-0.1580862105,-0.2029995918,0.0602509975,-0.1202484593,0.2065158486,-0.4876646399,-0.1865930557,-0.3327080011,0.1092467234,-0.2565141618,0.6586427093,0.2108451575,0.2717781067,0.1587678641,0.0456164107,-0.1248292401,-0.4218960404,-0.0536406897,-0.031254977,0.0374605246,-0.0909010321,-0.018166393,0.0350252353,0.4715485275,0.1138012186,-0.0306689758,-0.5500047803,0.0920290872,-0.0775489062,0.2022336721,0.0827576369,-0.1859748811,0.1372915953,-0.2366597801,-0.0745643452,-0.1108954623,-0.4580129087,-0.1761930138,-0.0046438426,0.0675838739,0.3397305608,0.0321200006,-0.3095202744,-0.0657437518,-0.0772382021,-0.1141663566,-0.1819245815,-0.0122263776,-0.0531609021,0.165982753,-0.2359533757,0.2322050631,0.0053282385,-0.2006025761,-0.4661620557,0.4412536323,0.1245172247,-0.1237185821,0.264113903,-0.2815726995,-0.1656348705,0.478548497,-0.5246481895,-0.1548048854,0.0558007322,0.1381165832,-0.3403528631,0.4552120864,0.302120924,0.0476573855,0.0140215773,-0.0724772513,-0.2228309512,0.0957019553,0.0429991521,0.403321296,-0.1240115091,0.1896506846,-0.0526103452,0.4766650498,0.0482534617,-0.1304227561,0.2802925706,0.1950444132,0.4728262424,-0.1371495724,-0.120395042,-0.4867364764,-0.2577689886,-0.2264502943,-0.0197656266,0.0007065187,0.0517286509,-0.0943002924,-0.2863279581,-0.1904703975,-0.2041421086,0.0761577338,-0.2113741189,0.3135751188,0.057280235,0.1835840642,0.1221519411,-0.1776895523,0.0991699696,0.1698618978,0.4022675753,0.0150110852,-0.3136452734,0.0082833301,-0.150718227,0.1457839459,-0.200870946,-0.1088743135,0.2638161778,-0.0960791707,0.0826123506,0.2346355915,0.8441630602,0.0945911258,0.0936687514,0.0452637933,-0.2911583781,-0.2852752507,0.0557534397,0.0138809318,-0.2369900346,-0.1452254057,0.6316406727,0.0028979988,-0.0583520643,-0.163096875,0.5997638702,-0.4410640299,-0.3710057735,-0.1498325914,-0.1423989832,-0.5282940865,-0.2274989039,-0.2783735693,-0.1651137471,-0.1657167971,0.1216922849,-0.3377402723,-0.0797508433,0.1186226308,-0.1386127025,0.3896143734,0.2384508848,0.055058945,0.1810155362,0.3151736557,0.2113418728,0.4770441353,-0.2193970829,-0.3072481155,-0.1956392527,0.1561985016,0.1901343614,0.1089453697,0.0555665307,0.0437804386,-0.0478346646,-0.1942570806,-0.4450428188,-0.1347373575,-0.1656711698,-0.0990399048,-0.1916393638,-0.4690860808,0.6322375536,-0.1739894152,-0.1316201538,-0.1235203966,-0.0665303543,-0.2540270388,0.3347468674,0.1608252227,0.9694903493,0.0472935587,0.1549809426,0.1151882634,-0.3309181929,0.1146973968,-0.0952972695,-0.0556075163,-0.4986627996,-0.2365258038,-0.1501405984,-0.1499619335,0.0369533971,0.4310852289,-0.3788287938,0.356651783,-0.3201094568,-0.0296788514,-0.1564435065,0.17677477,-0.1408024132,-0.2388236821,-0.1166882664,-0.0169758685,0.009540963,0.1132638827,0.0370531939,0.0269385818,0.1899421811,-0.019800555,0.2504062653,0.1620726734,-0.0853404924,0.3151869774,-0.1668691784,-0.5314210653,-0.0107575618,0.5266741514,0.5019896626,0.0503388271,0.0927917361,-0.0129766446,0.0295860711,0.2790659964,-0.1464172304,-0.1045441329,0.4724716544,0.175294891,-0.2453791648,0.2079087049,-0.0959707648,-0.1067540124,0.077022329,-0.0453327298,0.2414695024,-0.3650997877,0.0446828082,0.2673708498,0.2766034901,-0.085038729,-0.0106883943,0.0355995484,0.0274106842,0.2962166071,0.0216507446,-0.2268109173,0.0739027858,0.4579524696,0.2298096716,0.0274650455,0.452978611,-0.3027492762,0.0437544659,-0.2348895222,0.3578221798,0.1558244973,-0.2728921771,0.1640939713,0.1499307305,0.352101326,-0.0766545162,-0.1872918457,0.096082896,0.0095184445,-0.0317107998,-0.092792362,-0.3228533268,0.1501917839,0.3122368455,0.3856064081,0.2359106392,-0.2864782214,-0.0425245613,0.4829116166,-0.2021658719,0.2236270308,0.2858696282,0.2984408438,0.0155731272,0.1222397462,-0.2630719543,-0.3910295963,-0.0439396352,-0.2537714839,-0.2928431928,-0.1167143211,-0.0936503261,0.1613434553,-0.1719062179,0.2861396074,-0.3794550002,-0.4428187013,0.2977697253,-0.2330527902,0.1779194474,0.2634782791,0.2481666803,0.2294872999,0.1904302686,0.3597935438,-0.3228962719,-0.0372621045,-0.2004564106,0.1767511964,0.0193684492,0.0437839925,-0.0080933059,-0.1049268991,-0.5006384254,-0.062363416,0.5708612204,0.0377837531,0.0752921328,-0.1121585071,-0.07330814,0.0764947459,0.2585806549,0.1424664706,-0.3166401386,-0.3115602732,0.3152821958,0.1083468646,-0.1660612077,-0.1455033123,0.2591181099,-0.3066038191,-0.2280294448,0.1627318114,0.0407925546,0.0030276249,-0.3377047181,0.0188239384,0.3602542877,-0.1565861106,0.3047995567,0.4423649311,-0.0967154056,-0.0677859783,0.1052057743,0.3428689241,0.0583968535,0.2408365607,-0.0606638305,0.3778834939,-0.0158734974,0.0611626431,-0.0311948601,-0.4003718495,0.0675602257,0.2207546979,-0.2321504653,0.0245229173,0.0097706849,0.0145138707,0.2708436847,-0.3510390222,-0.1519544125,0.3862373829,-0.0674808621,0.064906992,0.0501350872,-0.1028002873,0.2203303874,0.2604193389,-0.0059285024,-0.3765155673,0.3823733032,0.1547002196,-0.0371921174,0.1426365227,0.0961540863,0.1613837332,-0.0567566045,-0.2217998356,-0.0699081644,0.015247779,0.1601680666,0.0454157777,-0.0462476201,-0.1148170009,0.0529400557,0.1100773513,0.3344461918,0.307713747,-0.0333198458,0.0886465386,-0.0147228613,-0.1016329527,0.01018818,0.2018380165,0.000448895,0.0180066619,0.1105642244,0.018414462,0.3374853134,-0.3615120947,0.2868885398,0.3628122509,-0.1820760965,-0.1669215411,0.0322595574,0.1311510056,-0.0288601257,0.6571774483,-0.3129000962,-0.0014360712,0.1126716956,0.0295944586,-0.1300785244,0.2919332087,0.3886236548,-0.4222636521,-0.2638187706,-0.1415784955,-0.5935504436,-0.0703666136,0.2431910634,0.1278576553,-0.2338165045,0.1324103624,0.207968235,0.2505531311,-0.0467729904,-0.2302667052,0.1695554554,0.3301645517,0.0832536146,-0.2390666902,-0.0924552083,0.4059368372,-0.0799019784,-0.3620835841,0.0415074117,0.1896494776,-0.0573388301,0.245522216,0.1919705868,0.0094557675,0.3001295328,0.61625278,-0.0190494284,-0.0020515304,-0.2047892958,0.2092382759,0.4031697512,0.022274401,-0.1401592493,-0.0994511694,0.1238329932,0.2696532011,-0.4371399581,0.3272208273,-0.2485329062,-0.170999065,0.0642756745,0.0851403326,-0.431586355,0.0509648733,0.1431972086,0.2185950279,0.1820223778,0.3784556389,-0.0685855374,0.3610672057,-0.1881100833,-0.1009036899,0.4753727913,-0.1566340774,-0.2016954273,-0.0870269611,0.4287046194,-0.0821658149,0.0284779221,-0.5293526053,0.0917174965,0.4609177709,-0.230727002,-0.0510534011,0.4078030586,-0.0736484379,0.1740234345,0.047764577,0.4229542315,0.1043643504,-0.3273547292,-0.1745105386,-0.3433350027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1993","title":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? ","comments":"@lhoestq I also found that cache in tmp directory gets updated after transformations. This is really problematic when using datasets interactively. Let's say we use the shards function to a dataset loaded with csv, atm when we do transformations to shards and combine them it updates the original csv cache. ","body":"I am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n","comment_length":50,"text":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? \nI am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n\n@lhoestq I also found that cache in tmp directory gets updated after transformations. This is really problematic when using datasets interactively. Let's say we use the shards function to a dataset loaded with csv, atm when we do transformations to shards and combine them it updates the original csv cache. ","embeddings":[-0.3775416613,-0.1095244214,0.0013312791,0.2857868671,0.2583917081,0.1318965405,-0.0723689497,-0.0919271037,-0.0414132737,-0.0703074262,-0.0545117334,0.2699582875,0.0232355762,0.0732333884,0.1965217441,0.247630313,0.2620951831,0.2115170956,-0.4262385368,-0.0511157475,-0.2185805887,-0.1405919343,-0.0047834129,-0.2865779102,-0.2252596468,-0.356764257,-0.0927880406,0.0716755912,0.2163686603,0.0895133018,0.197044611,0.1367615163,0.2189597934,0.3962489069,-0.0001211941,0.0404283516,-0.3316937387,-0.1156464741,-0.315446049,-0.0537972711,0.0001290025,-0.2120242119,0.0525789894,-0.2816846073,-0.0468880758,0.0064463159,-0.1075415611,-0.3362385929,0.56011796,-0.2909089029,0.0684703588,-0.1258438081,-0.3435747325,0.147519961,-0.2772858441,0.3350464106,0.1380095035,-0.1039549187,-0.0801337212,0.3443596959,0.0155066093,0.2434511781,-0.2749492228,-0.1867043078,0.363091141,-0.0399348363,0.0585354753,-0.2551787496,0.2135364562,-0.0502433553,0.7311311364,-0.4320097268,-0.2579140663,-0.2524380684,0.0672600642,-0.0608980209,0.1881186813,0.2648538351,0.1020963117,0.1773752123,-0.6674378514,-0.7337232232,-0.0253950693,0.1337731332,-0.0163036641,-0.5745205283,0.0006099565,0.1487910599,0.0624434613,0.2391048372,0.4866171777,-0.2371118069,-0.3087271154,0.1965992451,-0.3022064865,-0.1864703745,-0.2256144136,0.2395218611,-0.1068113521,0.1615797132,0.0734243393,-0.0433905348,-0.3852365613,0.1298671514,0.1477092057,0.3210479319,0.150812149,0.2113255858,0.1089457721,-0.1171292141,-0.221382305,-0.0770055205,0.2559005618,0.0755556226,0.6239591241,-0.273537308,0.1777050346,0.0221854653,0.1790609211,0.0160738621,0.1482888311,-0.0388496183,-0.156610176,0.3401579857,0.0454195216,0.1725093126,0.1815041155,0.1542683095,-0.0600077063,-0.1481684893,-0.1722118109,-0.1459460407,-0.2952183187,0.4529227018,0.1697227359,-0.0070349863,0.0887540951,0.4451136589,-0.4322243035,-0.1200901046,-0.0233448241,-0.1521674842,0.4138379693,0.2323964089,0.135147199,0.2169096023,-0.0307578314,0.0287322979,-0.0869795009,0.6420631409,-0.3159511983,-0.1536168456,0.2198780328,0.0855471641,-0.2015261501,0.1320282221,-0.6194599271,0.1727582216,0.271661073,-0.1368692219,0.1567055136,0.0461915843,-0.3810898364,-0.2422449738,-0.0971097127,0.340615809,-0.42462942,-0.2163517326,0.2547135949,-0.1289497018,-0.2241920829,0.2647132277,-0.2679210305,0.2254056036,-0.2347070277,-0.3467530012,0.5462701917,0.0552276261,-0.3988825381,0.0837451071,0.0544948429,-0.0453623421,-0.0010164796,0.5464828014,0.1674204618,-0.1307812929,-0.5164803863,0.3973061144,0.1192881018,0.1762742251,0.039111685,-0.1514856219,0.2610336244,-0.2685460448,-0.0825335905,0.4303967357,0.2913797796,0.2922714353,0.2142867297,0.0226373412,0.1894330233,0.3739198148,-0.1842550188,0.1289743185,-0.1220929548,0.1327108741,-0.6422823071,0.0001149304,0.1669109166,-0.6285805702,0.2326988876,-0.0832135752,0.0245007295,-0.1535638273,-0.0242388975,0.0609049313,-0.1061299816,0.0199475568,-0.0505514666,-0.2283900827,-0.2481737733,0.6387430429,-0.0299804155,0.1081735045,-0.6304489374,0.3527702987,0.2514104843,-0.2475289851,-0.3831181228,-0.1212913916,0.200368464,-0.0978032351,-0.2113641351,0.4761021435,0.225103721,0.399803102,0.006494415,0.2575464845,0.0505138524,0.0444833711,0.2104929984,-0.1847687215,0.087361455,-0.0954155475,-0.1755372882,0.0191601273,-0.2041374147,0.0782023817,0.1050141007,-0.2045371085,0.1093192548,-0.2950010598,0.0116043435,-0.233898133,-0.3099147081,0.1454846561,0.3009774089,0.1118343621,-0.0831262842,-0.1830652952,0.2358202189,-0.2309436202,-0.0270286873,0.0012549966,-0.3378604054,-0.0525653325,-0.1386366934,0.5303159356,0.1911537498,0.0668933094,-0.0152239343,0.1222703457,0.131340459,-0.0362393036,-0.0084724724,0.0071459915,0.4104354978,0.2127825767,-0.0609661676,0.1045977995,-0.1168995053,0.4943213761,0.1455186456,-0.1304632723,-0.3501494527,0.100255765,-0.1553595066,0.0072967149,-0.318831414,-0.0977089033,-0.2095637023,0.1084101945,-0.2698780894,0.7121255398,0.2677802444,0.2180776447,0.1329807192,0.1029780731,-0.071062237,-0.4081368148,-0.0947037414,-0.075343661,-0.0411883853,-0.0888822973,0.0353815593,-0.0306444839,0.3885793388,-0.1165231243,0.1027390212,-0.458065629,0.0715828836,-0.0804569125,0.1002640799,0.1204332337,-0.2551056445,0.2448674589,-0.4125901163,0.0325438865,-0.1087881476,-0.4531968236,-0.1931724101,0.0459400117,0.0540232509,0.1399481893,0.0338582806,-0.1722104251,-0.0154421171,-0.0375103056,-0.1620900333,-0.1627314538,0.0563961305,0.1628095955,0.0724283978,-0.3136128485,0.1613569558,0.1359173059,-0.3148086667,-0.618999064,0.3667176664,0.1239371076,-0.1368625462,0.2416620404,-0.1194218397,-0.0519928709,0.5253454447,-0.6610502005,-0.0536314808,0.0278386604,0.1480025649,-0.3420058489,0.3685707748,0.3670852184,0.0587698072,0.0175577383,-0.0795671865,-0.2193681747,0.0355430022,0.1981669962,0.3416231871,-0.148890242,0.1682678461,-0.012552117,0.3555630445,-0.0052123354,-0.187668398,0.349634409,0.2760183811,0.5141162872,-0.18929708,-0.1055221856,-0.5744802952,-0.1770068258,-0.1977824569,0.0307643227,0.1037045717,-0.0454978459,-0.1361054778,-0.172848925,-0.3196498454,-0.1677065939,0.0215868857,-0.3802634776,0.3766799569,0.1258678138,0.2262209058,0.0465897992,-0.1971980184,-0.0069374572,0.2207194269,0.3263590336,0.0762136504,-0.1963417083,0.056027133,0.0341770314,0.2427733094,-0.1556178182,-0.0294063333,0.1411680579,-0.1030972078,0.0389259383,0.331690371,0.7368652821,0.0918534547,0.0934501961,0.0727234855,-0.3702737987,-0.2972922921,-0.0185835976,0.0517490357,-0.2108961642,-0.1065102667,0.6030531526,0.1897177547,-0.0930278897,-0.3880621195,0.4908544719,-0.3521109223,-0.4243793488,-0.1121429503,-0.0784945711,-0.492872268,-0.1960447729,-0.3165834844,-0.1781642288,-0.1199575141,0.0184137207,-0.3197627366,-0.1178011373,0.0356115699,-0.1558897495,0.333553046,0.3068920672,0.0824248344,0.1043403894,0.1859696656,0.0567427836,0.3127235174,-0.1555327177,-0.0530600548,-0.1701943427,0.2026268989,0.0648548156,0.0702834427,-0.0058349459,-0.0307656787,-0.1382617503,-0.2153833508,-0.2612842917,-0.0104752667,-0.2210043371,-0.0227105338,-0.1755328923,-0.4874750078,0.4563183188,-0.1010736972,-0.1883671731,-0.1476007402,-0.0701861605,-0.3030474782,0.1635485142,0.1351780295,0.8293338418,-0.0081271473,0.1564800441,0.0635269955,-0.1072181463,0.0256146174,-0.1322309673,-0.0666994825,-0.3328029513,-0.2906267643,-0.182567656,-0.1834264845,-0.0186124798,0.340477854,-0.3613784611,0.3638460338,-0.252361834,-0.0129183372,-0.0684406385,-0.0061908676,-0.0827891529,-0.2392550558,-0.1205003783,0.0178153776,-0.1263612211,0.0206519756,-0.0170574468,0.1329820901,0.1653869599,0.0883446112,0.3602747917,0.0859850943,-0.0614523888,0.3906381428,-0.1673132479,-0.4016444087,-0.1059151962,0.6008502841,0.4341290593,0.0510478243,0.0976215228,-0.0349936001,0.0490859896,0.2996875942,-0.1671101302,-0.0298140589,0.4338347316,0.2377783209,-0.0086455718,0.0113460906,-0.1001394689,-0.1390534341,-0.0175677706,-0.0650363639,0.0446571596,-0.3660856783,0.0471717194,0.2655333579,0.3460141718,0.0343473926,0.0075051393,0.0799464583,0.0904643834,0.3925037384,0.0361486115,-0.3290790319,0.0518718958,0.5086259842,0.1846586615,-0.0978844911,0.2865205109,-0.275865376,0.0242179167,-0.1840946674,0.413282007,0.0370373391,-0.2001437992,0.1599382609,0.0937231854,0.4383989573,0.0149290739,-0.1348427534,0.0453366153,0.0022428655,0.0129872803,-0.0573017262,-0.3163817227,0.1181789041,0.2467707098,0.3534682691,0.3582066596,-0.2377302498,-0.0091207987,0.4705203474,-0.1857517362,0.1302448213,0.3519063294,0.3294466734,0.2843283713,0.1291950941,-0.2915623188,-0.349634707,-0.0622041412,-0.2297858149,-0.317373395,-0.1254490167,0.0059904377,0.1760306954,-0.0534444675,0.1571178883,-0.3403251171,-0.4503931701,0.2765614986,-0.2510400712,0.1360728741,0.170859158,0.084204793,0.1272505075,0.1955584437,0.4348476231,-0.2793298364,-0.0205031931,-0.2099926174,0.082277447,-0.137168169,-0.0476504192,0.0680936947,-0.1127315685,-0.5614906549,0.0071773981,0.7099939585,0.1469898969,0.084573701,-0.2261671275,0.0306620263,0.1549828053,0.2425213456,0.026595125,-0.2630420327,-0.3262077272,0.3222832978,0.1072231755,0.0007812451,-0.1187379882,0.2667143643,-0.1083325967,-0.2915621102,0.1660125554,0.1182722598,0.0491989255,-0.1892528385,0.0828060731,0.4146974683,-0.2138024271,0.3286812901,0.5098997355,-0.2169280648,0.0500127412,0.1384067684,0.3438334465,-0.1801340431,0.4385668933,0.0227652807,0.4802241027,-0.0042788316,0.0963330343,-0.0411803201,-0.4797659814,0.0866856873,0.0685357526,-0.2614332139,0.090110153,0.0387507975,0.0042929389,0.171146661,-0.3964384496,-0.1746979058,0.4800515771,-0.1526473761,0.0949188396,0.1018081307,-0.1112465411,0.2824836969,0.2111829519,-0.0595505908,-0.2754020393,0.4276595414,0.1697075963,-0.0429920889,0.06445916,0.0970565826,0.1154236495,-0.1176576391,-0.1695513129,0.0883056447,-0.1025509089,0.0809189081,0.0268968791,-0.0096894605,-0.2110137939,0.0987922624,0.0845461413,0.3834230304,0.305773288,-0.0233612731,0.1660701185,0.1643571109,-0.1570235938,-0.0190252941,0.1402451992,-0.0021696137,0.0448974185,0.217905432,-0.0300386716,0.4400350749,-0.379780978,0.2381946594,0.3017783761,-0.1042671874,-0.1633945256,0.0579111204,0.1083805338,0.0334220417,0.5928037763,-0.3671610355,-0.0240957029,0.0383854918,0.0305531137,0.0665908903,0.4713372588,0.2755865157,-0.3269456029,-0.2479602695,-0.149293229,-0.5072792172,-0.0564656518,0.3451339602,0.3302523494,-0.3488498926,0.09018334,0.0922692567,0.2255484462,-0.0054463083,-0.1682596654,0.1758976132,0.322655499,0.0473422781,-0.2322089672,-0.0225877892,0.2931557298,-0.1141253263,-0.1815204918,0.1177115962,0.1478688419,-0.0170852765,0.1777738482,0.1491085887,-0.0692705214,0.3727116585,0.5638625026,0.0122294873,0.0836882442,-0.1030982211,0.3902325034,0.3161498606,-0.1495496333,-0.0536084138,-0.1487935334,0.0682308823,0.1575758159,-0.5072302222,0.2509377301,-0.1679805815,-0.2084991336,0.0972754136,0.1540718973,-0.4413489103,-0.0845655799,0.0100885369,0.1932955384,0.0599622168,0.3546241522,0.0319328122,0.4073848426,-0.1656405032,-0.1417236477,0.3911108971,-0.1909821779,-0.2934282422,-0.1084177718,0.2496730387,-0.1243701428,-0.1343964338,-0.3410141468,0.1087990031,0.4196589291,-0.2753848433,-0.0127331894,0.438187778,-0.1423455477,0.134046331,0.0069051627,0.5526450872,0.10680563,-0.2677078843,-0.0825621039,-0.2169014961]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1993","title":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? ","comments":"I plan to update the save_to_disk method in #2025 so I can make sure the new save_to_disk doesn't corrupt your cache files.\r\nBut from your last message it looks like save_to_disk isn't the root cause right ?","body":"I am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n","comment_length":37,"text":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? \nI am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n\nI plan to update the save_to_disk method in #2025 so I can make sure the new save_to_disk doesn't corrupt your cache files.\r\nBut from your last message it looks like save_to_disk isn't the root cause right ?","embeddings":[-0.3105080128,-0.0680618212,0.0046882723,0.213742435,0.3093694746,0.0905763432,-0.0354265086,0.0527892485,-0.0732766688,0.006163144,0.0222795159,0.3242502511,0.0898376629,0.1039876118,0.1767604351,0.2547088563,0.2500333786,0.1937623173,-0.3340500295,-0.0268145371,-0.2293639928,-0.1129300967,0.1087111235,-0.2835570276,-0.330468595,-0.3502230942,-0.061735969,0.0674631819,0.1010746285,0.002596098,0.16506356,0.1604154557,0.2431449294,0.2890101075,-0.0001189022,0.003664962,-0.2793877125,-0.0914112702,-0.3182113469,-0.062790595,0.0237718113,-0.1687760204,0.0643625483,-0.2784352303,-0.0268728323,0.0422446094,-0.1302841306,-0.2041808516,0.6294941902,-0.2092859745,0.1225881949,-0.1459359527,-0.1816229522,0.0268533751,-0.2328369319,0.2694652677,0.0554410666,0.0393736362,0.0392527133,0.3304750621,-0.0896674395,0.2254993767,-0.3003205359,-0.2334360778,0.41138345,-0.061823599,0.1770342141,-0.352799058,0.1796052903,-0.0407533422,0.8042439222,-0.3945490718,-0.2202159911,-0.2704587579,0.0771651119,-0.126911059,0.1884325445,0.0927256495,0.0741950646,0.1356611997,-0.5429519415,-0.8058210015,-0.079282321,0.0439732037,0.0849198475,-0.5327285528,-0.028630957,0.1143597662,0.2698810399,0.280354321,0.51173383,-0.1818695813,-0.2572776973,0.1177860498,-0.2956849933,-0.2979613543,-0.2101858705,0.2007384449,-0.0370490476,0.2396800071,0.1317304671,-0.0441242456,-0.2862701714,0.073063381,0.1336907148,0.334923327,0.0764580145,0.111513488,0.0928423256,0.0828307644,-0.1914321631,-0.08878804,0.3011989594,-0.0253592879,0.6220217347,-0.2496522218,0.3028310835,-0.0809742883,0.1319755465,0.0198019072,0.1669227332,-0.030759193,-0.1410048902,0.2388069779,0.1260763407,0.2384852767,0.1625214666,0.2277155071,-0.0890761018,-0.0952495337,-0.1974941194,-0.0420266464,-0.2991293371,0.4488770962,0.1797114462,-0.0550356284,0.0623022504,0.444839865,-0.4268714786,-0.1361715198,-0.0558103733,-0.2085800767,0.33602795,0.2015105486,0.1050624326,0.20187217,-0.0205758177,-0.0619851537,-0.080488652,0.6188934445,-0.4360120595,-0.2547796369,0.0720278099,0.1330969334,-0.1292629987,0.1620911658,-0.6500713229,0.1481628269,0.1823387146,-0.1603377759,0.2286168933,0.0785053745,-0.3790595233,-0.2501143813,-0.0768620595,0.2681693137,-0.3776040375,-0.0919734538,0.1421821713,-0.1908731163,-0.078642644,0.2338188887,-0.3230122328,0.2565499544,-0.2728489339,-0.2951149344,0.530872941,0.1095688939,-0.4193350673,0.1607092619,-0.0415808894,-0.0676332787,0.0187847894,0.5070464015,0.1284545362,-0.2125234902,-0.3108474612,0.2478173226,0.1098065823,0.0558847487,-0.0957123861,-0.1768736988,0.2356046587,-0.2986460924,-0.0614739247,0.4517145455,0.2861812413,0.4201847315,0.2112652659,0.0092199091,0.0990410745,0.3704631031,-0.074640356,-0.0521767922,-0.2253950983,0.2060769796,-0.6755854487,0.0409301892,0.1792390049,-0.6058937907,0.3495005071,-0.1251087338,-0.0100670494,-0.2055464089,-0.068482697,0.028655747,-0.0467946827,0.0033021199,-0.0353210457,-0.1637226939,-0.2710678577,0.533338666,-0.0845813602,0.0758440867,-0.620757401,0.3459838331,0.1316459775,-0.1627105623,-0.3209216893,-0.0708475113,0.2944808006,0.0222208984,-0.1368241757,0.388717562,0.1961941868,0.4301020503,0.0012601739,0.2552749217,0.0939521492,-0.0422644056,0.0979755148,-0.199407652,0.111207284,-0.1304408908,-0.1259728223,-0.056881424,-0.2236297131,0.0619360767,0.0439170226,-0.2145471871,0.0435210504,-0.3417886794,0.0263666082,-0.2048581392,-0.2158522308,0.1839472353,0.3429169357,0.1599927396,-0.1657072604,-0.1038518772,0.2246282846,-0.2585544288,-0.0839284286,0.0609280728,-0.1497137994,-0.096774675,-0.1595600098,0.5698390603,0.1799433529,0.0722898543,0.0952895284,0.1322449148,0.0780074671,-0.075699985,0.0063114837,-0.0098382747,0.5663112402,0.2622880936,-0.0351031497,0.1173703372,-0.1512570232,0.436498642,0.158765927,-0.0320275798,-0.333922267,0.0973033905,-0.12079864,0.0436149649,-0.3793645799,-0.115243651,-0.2534185648,0.0617879443,-0.1707229167,0.7471997142,0.2428154051,0.1477638632,0.0102008535,0.107787177,-0.129888624,-0.3767109811,-0.1586440355,0.1013624296,-0.0191389024,-0.0515207089,0.1256586313,-0.0810722709,0.3847855031,-0.0449679457,-0.0184505098,-0.5178281665,0.070510067,-0.067127496,0.0722048208,0.0966846272,-0.1959410906,0.1818971038,-0.3614756763,-0.0663395822,-0.126332745,-0.4590773284,-0.1805945337,0.0563868172,0.0017029888,0.2645662427,-0.0602228716,-0.2389340103,0.0197035652,-0.0604273193,-0.1862342358,-0.2104735672,0.0663478002,0.0624297,0.1308450997,-0.2201138586,0.2857636213,0.0032416596,-0.3236318529,-0.6576063633,0.4279029369,0.1779450178,-0.2001407146,0.3122074902,-0.110426791,-0.0269734878,0.4068779349,-0.5890811682,-0.1938370913,0.0083895,0.104708612,-0.4142176509,0.2215412706,0.3652362227,-0.0205243528,-0.0421430618,-0.0373719111,-0.2552675009,0.0953872278,0.1543393731,0.4772271216,-0.2085597962,0.2322051674,0.0358263813,0.3859612942,-0.0511958785,-0.0470176712,0.3821455538,0.2632104158,0.5747004151,-0.2582440078,-0.1334565133,-0.5062974095,-0.2519566119,-0.3096196651,-0.0341153145,-0.0407003053,-0.0676548183,-0.1260348111,-0.2426861376,-0.2178800553,-0.1950591058,0.0797049552,-0.2889825702,0.3272450566,0.0038291092,0.1908492595,0.1868156791,-0.2061285079,-0.0138957249,0.2137178481,0.3998015821,0.0800073743,-0.3275633752,0.0140291024,-0.2869059741,0.2075462639,-0.2099593729,0.0030590752,0.0532290861,-0.0138413543,0.0894816071,0.203609705,0.7742400765,0.1549123973,0.1421173811,0.1629953533,-0.4092797339,-0.2115705758,0.0028166687,0.0995186642,-0.2312848717,-0.1408337057,0.6834016442,0.1378550828,-0.1164821386,-0.407292515,0.461717248,-0.2899160385,-0.2482812554,-0.1648038179,-0.2642420828,-0.4946279526,-0.1494385898,-0.2242548466,-0.1094842032,-0.1916747689,0.1236998141,-0.1940564513,-0.1066654921,0.0562275909,-0.1835999936,0.3257081509,0.3064141572,0.0279383603,0.1640270054,0.184833169,0.1314572394,0.3027799428,-0.1124108359,-0.0372946002,-0.0790411383,0.1409601271,0.0719048381,0.169445768,0.0061108209,0.017624028,-0.0119182849,-0.1746780276,-0.2852495313,-0.0722135082,-0.1454936862,-0.1713914424,-0.2936673462,-0.4344517291,0.5158485174,-0.0422941372,-0.1093990356,-0.1100788042,-0.107598573,-0.2939422727,0.1297604591,0.2262260914,0.9243401289,0.0883862078,0.1814991087,0.0876085535,-0.1721745282,0.0167084038,-0.0423726663,-0.046322342,-0.3026043177,-0.3657317162,-0.176059261,-0.144395858,0.047770083,0.2639947832,-0.3115513027,0.3357236981,-0.2909373641,-0.1034817994,-0.1046893522,-0.0075613996,-0.2122483402,-0.2458733022,-0.2050827146,0.0313251652,-0.151548028,-0.0273836982,-0.031810578,-0.0045625153,0.2649879158,-0.0017585804,0.3558607399,0.156926617,-0.08030691,0.2989886105,-0.169228211,-0.4347157478,-0.0530970693,0.5133090615,0.4914640486,0.0112366555,0.0094525022,-0.055323001,-0.0235152338,0.3599002957,-0.1810908616,-0.1940719038,0.4862323403,0.0869534686,-0.0590406694,0.0497814901,-0.0420021079,-0.291921556,-0.047430899,-0.128767699,0.1325778365,-0.2795438468,-0.04591313,0.2894847989,0.273804009,-0.0742639303,0.0540255606,0.0834544972,0.0090321284,0.3221890926,-0.0452659242,-0.3151636422,0.0543161668,0.5431888103,0.180405587,-0.0464722924,0.3360171318,-0.3092155159,0.0523002744,-0.2045600861,0.398933351,0.1732898951,-0.143210113,0.1380074471,0.020450579,0.4439578354,-0.0311348587,-0.0242852513,0.0992704704,-0.0084157502,0.1196399108,-0.0480960794,-0.3914588988,0.1300692558,0.2458519936,0.3818084896,0.3204666376,-0.1763056815,0.0028133055,0.3804045618,-0.218831256,0.1583355069,0.2809144557,0.2071995735,0.2206033021,0.1162377,-0.2657069862,-0.3280172944,-0.0514550656,-0.1594754308,-0.3051246703,-0.1652151048,-0.0302931909,0.136443451,-0.0375773832,0.1899552792,-0.2565503418,-0.450643152,0.2677793801,-0.129545629,0.1429374218,0.2473307103,0.2051231861,0.1521451473,0.0653467923,0.4487791657,-0.2328964621,-0.1185056791,-0.2105093449,0.1233182922,-0.0207088292,-0.0022413959,0.0580328368,-0.1482613385,-0.5246192813,-0.0672190934,0.556946218,0.1424945295,-0.0765211806,-0.2013106644,0.006891069,0.2139699161,0.2115524113,0.0510492325,-0.3432705104,-0.2787159085,0.3432043195,0.1442660391,-0.0160444789,-0.2254811972,0.2461643666,-0.180695951,-0.3061958849,0.1442252249,0.1767639816,-0.058568716,-0.2754856944,0.158903271,0.2719600797,-0.3255973756,0.3828686178,0.5380005836,-0.2591253519,0.0096257431,0.225722298,0.281971097,-0.1335764974,0.4269311428,0.0098501565,0.5303928256,-0.039582327,0.1332318932,-0.0633739084,-0.4151782393,0.0702961013,0.1649889201,-0.2766591907,0.0600557365,0.024320988,-0.097980015,0.1432588398,-0.4174813032,-0.1507119685,0.4192144275,-0.2484193891,0.107425645,0.0579757616,-0.1120400503,0.2826018333,0.2744211853,-0.0536198914,-0.226840958,0.4427161515,0.125539571,-0.1913319528,0.2044383138,0.1249766201,0.158897385,-0.0993127748,-0.1169832349,0.1147721186,-0.0848859996,-0.0037676345,0.0291740242,0.0298790541,-0.1809425205,0.1118033603,0.1676972955,0.2783719897,0.3059472144,-0.0228865426,0.1679927707,0.0827448219,-0.0256147627,-0.0337706804,0.2122331113,0.0137036452,0.0140267555,0.3118373752,-0.0362616591,0.4630796015,-0.3419649601,0.2062887698,0.4129410982,-0.107400164,-0.1485196054,0.0904574469,0.1278402656,-0.0378109887,0.5705618262,-0.2957132757,-0.0160776153,0.0124919377,0.0395254977,0.1230504066,0.433227092,0.212803185,-0.4697756171,-0.2823480666,-0.1664899141,-0.6273463964,-0.0509000681,0.3253649175,0.3701803684,-0.2258560359,0.1118433774,0.0707369298,0.209854275,0.036916595,-0.2351111174,0.1503262967,0.2999602556,0.123491846,-0.1595014185,0.050346449,0.277072221,-0.134976089,-0.2388400733,0.1724776775,0.1970548332,-0.0143192848,0.2137266248,0.1389212757,-0.1556822509,0.4832547903,0.539032042,0.1382324845,0.0534417778,-0.1822286695,0.3323269486,0.2797116041,-0.0581519753,-0.1694723219,-0.1795464307,0.0746415555,0.1931521893,-0.3903319538,0.3852839172,-0.2698163092,-0.0080449646,0.073005721,0.1210964322,-0.5197293162,0.0209048744,0.1311253756,0.1718316972,0.1720623374,0.4176313579,0.045043949,0.3535895646,-0.2674767077,-0.1531490535,0.4640339017,-0.1586665362,-0.3031260073,-0.1587207913,0.309746027,0.0269460529,-0.0641721785,-0.3660287857,0.0210532453,0.3587178588,-0.2671586871,-0.0269523151,0.5555734038,-0.2037760764,0.1576040387,0.0878340006,0.5945928693,0.0969752073,-0.3053180575,-0.0882790908,-0.3290880024]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1993","title":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? ","comments":"ok, one more thing. When we use save_to_disk there are two files other than .arrow. dataset_info.json and state.json. Sometimes most of the fields in the dataset_infor.json are null, especially when saving dataset objects. Anyways I think load_from_disk uses the arrow files mentioned in state.json right? ","body":"I am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n","comment_length":45,"text":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? \nI am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n\nok, one more thing. When we use save_to_disk there are two files other than .arrow. dataset_info.json and state.json. Sometimes most of the fields in the dataset_infor.json are null, especially when saving dataset objects. Anyways I think load_from_disk uses the arrow files mentioned in state.json right? ","embeddings":[-0.3007218838,-0.0692068636,-0.0054887459,0.2683240175,0.18687585,0.0893005505,-0.0907420963,-0.0502164327,-0.0936112627,-0.0296892803,0.0845306218,0.3922941685,0.0921069384,0.2010050416,0.1915561855,0.2176252455,0.2755788863,0.254154861,-0.2981148362,-0.0849448964,-0.1826583743,-0.2011492848,0.0351460502,-0.2450300306,-0.2057038397,-0.3581506312,-0.0677712783,0.015393083,0.1225500628,0.0245489161,0.1340078264,0.0892242864,0.2528147995,0.2696574628,-0.000121676,0.0475671478,-0.3047076762,-0.1001055315,-0.3526186049,-0.1412885487,-0.0697196573,-0.2212417126,0.0719519481,-0.3070665002,0.0589956082,-0.2824417353,-0.1260732114,-0.2450783253,0.6672537327,-0.2722013593,0.0699747428,-0.1284598857,-0.1617945284,0.0976968184,-0.2888140082,0.3928675056,0.1804583818,0.0834343582,-0.0248981174,0.3822441995,-0.0219692346,0.1612482369,-0.2000204921,-0.2527583241,0.3833414912,-0.0076433229,0.2212197632,-0.2953317761,0.1467781067,0.013578278,0.8064634204,-0.4050227404,-0.2574875951,-0.2882359028,0.1536288112,-0.0580589697,0.1496305019,0.2156998217,0.0218875743,0.1402100474,-0.5136809349,-0.8684421778,-0.1266848445,0.1264935583,0.1287239492,-0.5532692075,-0.0229649302,0.1637459993,0.2024900466,0.2185248584,0.4860473871,-0.3468028009,-0.2665126622,0.1081005484,-0.2416039109,-0.2579500675,-0.354360193,0.101179935,-0.0114108305,0.2894110084,0.1406663656,-0.0836784169,-0.3567636311,0.160385564,0.2414058,0.2809044123,0.2183113843,0.1693065166,0.0739967525,-0.0631906241,-0.186182797,-0.0970506147,0.1862160712,0.0415194817,0.560059607,-0.3111509085,0.2295670807,-0.0122573748,0.1502870321,0.0228205808,0.0895476416,-0.1179456413,-0.1425845474,0.2670946419,0.1297624707,0.2194279879,0.1593542695,0.2407317609,0.0423475951,-0.0376061946,-0.1410646886,-0.0918838382,-0.2097061723,0.4661527574,0.0937507749,-0.0388412066,0.1977848113,0.4313408136,-0.4969690144,-0.1085939556,0.0501839109,-0.1930902153,0.3274519145,0.1736688167,0.1454110593,0.2489248365,-0.0879915655,-0.0936397612,-0.1191328317,0.5767201185,-0.3575347662,-0.1520616859,0.0701175109,0.1043709368,-0.1598282307,0.0983543098,-0.8150238395,0.1427524984,0.0247533061,-0.123168692,0.1728976518,0.0534343235,-0.2563692629,-0.2302288115,-0.0512655973,0.1930936426,-0.4120173752,-0.1680333316,0.1258455366,-0.2549983561,-0.1799414456,0.210870102,-0.3636855483,0.2950892448,-0.2587227821,-0.302321732,0.6723531485,0.0204605758,-0.3387461305,0.1760644913,-0.0282904599,-0.1184442192,0.0147047229,0.52508986,0.1372104287,-0.0953126997,-0.4145095646,0.2945404947,0.1407813132,0.0868426487,0.0795525312,-0.2047659755,0.1286368519,-0.2869850993,-0.2456305325,0.3859400451,0.2542551458,0.3583353758,0.3105891645,0.0150293978,0.1410465837,0.4581476748,-0.1164281517,0.0385382064,-0.2127675265,0.1784372926,-0.7484533787,-0.0065718978,0.2338047028,-0.6842118502,0.3345311582,-0.1203256696,-0.0277768821,-0.1687025726,-0.0664194971,0.1298466623,-0.0762263536,0.0424030274,-0.0529155731,-0.1547596455,-0.2675639689,0.3820064962,-0.0523812845,0.1227223128,-0.6055701375,0.3676992357,0.2111807168,-0.2002485693,-0.2397426963,-0.1931338906,0.205521062,0.0092982724,-0.0893531218,0.378677845,0.1809562445,0.4551723897,0.0918436497,0.2322871536,0.0534293242,-0.0460140593,0.1447259635,-0.1604724228,0.1452556252,-0.1221863627,-0.263800472,0.044219058,-0.2015414238,0.178759262,0.1345491409,-0.1926780045,0.0233341306,-0.2902032435,-0.0180919431,-0.2220669985,-0.2539126873,0.1853529811,0.3977223635,0.1267140359,-0.1721722633,-0.1254348159,0.295930475,-0.2575980127,-0.0214607492,-0.0102388263,-0.3651685119,0.0082921209,-0.0941087157,0.6046720147,0.1611140966,0.0820363984,0.0866392925,0.0911179706,0.151191622,-0.0454357713,0.0442335419,-0.0101497471,0.5799141526,0.2443162352,-0.0383164622,-0.0365616903,-0.1149246618,0.3536050618,0.1652428508,-0.0946870297,-0.2773093879,0.1180244088,-0.2011263967,0.0713499561,-0.4884609878,-0.1273127496,-0.2521910965,0.1512754411,-0.271494627,0.7723132968,0.2325681597,0.1263019145,0.1219079271,0.0154319406,-0.1470237374,-0.4388962686,-0.0966933221,-0.0195642691,-0.0103024421,-0.0622432157,0.1154476255,0.049392689,0.3781045079,-0.0810580254,-0.0131591242,-0.439604342,0.1891238689,-0.0656802952,0.164598316,0.0659910142,-0.2215655893,0.2795462906,-0.3811887205,-0.0384196267,-0.0594334416,-0.3607291579,-0.2850465178,-0.0065293768,0.0134899691,0.2896597087,0.0172295738,-0.1902122051,0.0335531235,-0.0402981117,0.0173520166,-0.1225783825,0.0557491481,-0.0140358889,0.2444540709,-0.1130922288,0.180433169,0.0764265805,-0.2114164829,-0.5961003304,0.4128051102,0.1633280069,-0.1741424948,0.3108078837,-0.174173072,-0.0210463982,0.3101853132,-0.6087191701,-0.1141924858,0.0600616932,0.1037474275,-0.3919464648,0.3361907899,0.3062698841,-0.0278478414,0.0144448588,-0.0334713459,-0.2389530391,0.0784413666,0.1437770277,0.4244503975,-0.1612543762,0.2261483669,-0.0966194198,0.3579064608,0.0317309126,-0.1503335536,0.2577560246,0.1534895599,0.4387212098,-0.1750784069,-0.0522773303,-0.5761281252,-0.161820814,-0.2217735946,-0.1148253158,-0.0057710223,0.0323960073,-0.1075939536,-0.2496583909,-0.2805704176,-0.1933779418,0.0845356658,-0.2493488938,0.2655926943,0.0192373246,0.1759132743,0.1493865401,-0.1331045181,-0.0371735618,0.2424814999,0.356749922,0.0340544805,-0.4226408899,-0.0830405876,-0.0818447098,0.1717509478,-0.2144640535,-0.0884095058,0.1600837111,0.0011428865,0.0677132607,0.1965410858,0.736178875,0.1874350756,0.0637004822,0.0926948786,-0.3455916643,-0.3881474435,0.0843278393,0.050614085,-0.2877066731,-0.1174880415,0.724560082,0.0803995058,-0.1206825525,-0.3576766253,0.5562203526,-0.3704287708,-0.2955069542,-0.1457949579,-0.1664602757,-0.547655046,-0.2133507282,-0.3082762063,-0.158236742,-0.1307761669,0.0646412745,-0.2534776926,-0.0915706754,0.0220779572,-0.1182216704,0.3918710053,0.26667431,0.0193663146,0.1644690037,0.2219794691,0.3284146786,0.3666436672,-0.0153753338,-0.0974271223,-0.1326036155,0.0867798105,0.1648629755,0.0936449617,0.0686152503,0.047984682,0.0268370956,-0.2086209804,-0.3092784286,-0.0560037307,-0.1340935379,-0.136720255,-0.144084245,-0.5415611863,0.5702608824,-0.065022625,-0.1408147365,-0.0851876885,-0.0459545478,-0.2711398005,0.1703906804,0.1498789191,0.8716302514,0.1089723334,0.1778829843,0.1109232455,-0.2702061832,0.087459892,0.0335815847,-0.1201459393,-0.3685467541,-0.1829183549,-0.1566405594,-0.2075347304,0.0405540541,0.4153393209,-0.2801257074,0.3261986971,-0.223812893,-0.150596559,-0.199074477,0.0961721018,-0.1657651812,-0.1743447483,-0.1681726128,-0.0114012463,-0.0881939307,-0.0306611769,0.0053103175,0.0413889438,0.0553506725,-0.0187188257,0.3721584082,0.0846239179,-0.0624594949,0.277123332,-0.1577947438,-0.4986288249,-0.0826174468,0.5926850438,0.3960545957,-0.0334480032,0.0526533909,-0.0723138675,0.0268860888,0.3063696921,-0.2422439754,-0.17192702,0.4481704831,0.1376743615,-0.0855185986,0.1409664303,-0.0945118517,-0.1698204875,-0.0376815759,-0.0860807747,0.0551038869,-0.2550101578,0.0474657416,0.2712349892,0.253031671,-0.0210115146,0.009576655,0.1501528323,0.0579037443,0.2980197966,0.0394545197,-0.2374412715,0.0459560975,0.5188029408,0.2067332864,0.0214130152,0.4381721616,-0.3275930882,0.1210740358,-0.2402662635,0.5366315246,0.2017429173,-0.1746361852,0.1869060397,0.0694174543,0.400788337,-0.0473455787,-0.0213023182,0.0853855908,0.0111679211,0.1445219368,-0.0958094224,-0.4380567074,0.2343147546,0.2624767423,0.3535428345,0.2583889067,-0.23243393,-0.0680489317,0.3585879803,-0.1744461507,0.2054147124,0.3005222678,0.2823808491,0.1684187204,0.204443872,-0.2087183446,-0.3440043032,-0.0438150316,-0.2555530071,-0.2365293056,-0.1251322329,-0.0598074719,0.1539491564,-0.0436912887,0.1677895039,-0.3266103566,-0.4847747087,0.2891402841,-0.1375051737,0.0824396834,0.0819535479,0.219714269,0.1840926856,0.093729578,0.4481506348,-0.3184418082,-0.1187559739,-0.192102015,0.178301543,-0.0328108743,0.0121082598,0.0642218739,-0.1316829026,-0.6189084053,-0.0990728214,0.5383808613,0.1635977328,-0.0226289704,-0.167563051,-0.0290072355,0.1075523719,0.2544230521,0.2053806782,-0.3422957361,-0.2245641798,0.3293198049,0.1255977899,0.0269108526,-0.1925330311,0.2880922556,-0.1918520033,-0.2873286903,0.1465846002,0.0319879353,-0.0512793511,-0.4122623503,0.0254708398,0.1751537174,-0.1580880284,0.3159530163,0.5974047184,-0.2331489325,-0.0708440244,0.14665775,0.3994494081,-0.0056820922,0.3305520117,-0.0007100745,0.3764371574,0.0240038242,0.0916100517,-0.0240890123,-0.3671692014,0.0081163878,0.138963595,-0.2195097357,0.0497018807,0.0751757398,-0.0880869031,0.2082580626,-0.3309559524,-0.0835982189,0.3510943651,-0.1629564613,0.0826335028,-0.1298543811,-0.1498635113,0.3026362956,0.2705848217,-0.0788764283,-0.3361346126,0.328160435,0.1423033327,-0.0971574709,0.1063301116,0.1872467697,0.1876194775,-0.1655017287,-0.1503584087,0.0128684193,-0.0289312117,0.012515543,0.0165960528,0.0012614295,-0.2274467051,0.0419403464,0.1657588929,0.316542089,0.319621712,-0.0838937685,0.1499597877,0.0178740472,-0.0008007619,-0.0735295936,0.2127766609,0.0151791228,0.0294536892,0.2632454634,0.0044013718,0.3795756102,-0.3571351469,0.2479705215,0.3106712401,-0.1404623687,-0.1665807068,0.0103722224,0.2103223652,-0.1065976396,0.543430984,-0.2923462391,-0.0012199946,0.0241246484,0.0142698456,-0.0414992571,0.4267641306,0.2331003249,-0.4311102629,-0.1866676807,-0.1895916015,-0.5319392681,-0.0989796445,0.2813744545,0.2668113708,-0.2489684373,0.1820268631,0.1321471781,0.2249126434,0.0598294213,-0.1167491376,0.1265683323,0.3670851886,0.118841067,-0.2378845066,0.0502788126,0.3357857764,-0.0974639207,-0.214331314,0.1917597651,0.132594198,-0.0512679666,0.2214893103,0.1896936148,-0.1321160942,0.4478760064,0.6462352872,0.0189492851,0.0312899202,-0.1774075776,0.2362672836,0.3744183481,-0.0216822904,-0.1565989107,-0.016700536,0.0922045931,0.2499128431,-0.4566820562,0.444871217,-0.2041887045,-0.186814189,0.0137965474,-0.0477889888,-0.4581757188,0.0463743135,0.0651584268,0.1430877596,0.1621837318,0.3478299081,-0.0187160093,0.318493247,-0.1899687797,-0.024314208,0.4145450592,-0.0772262961,-0.1907040626,-0.0427960828,0.2818667293,0.0554945618,0.0050824471,-0.4016096294,0.1224801987,0.4839176536,-0.2663434446,-0.0787809864,0.3923036754,-0.1582095772,0.1256153584,0.0675276816,0.6033189893,0.0422040783,-0.3597431481,-0.092121467,-0.2827445567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1993","title":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? ","comments":"> Anyways I think load_from_disk uses the arrow files mentioned in state.json right?\r\n\r\nYes exactly","body":"I am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n","comment_length":15,"text":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? \nI am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n\n> Anyways I think load_from_disk uses the arrow files mentioned in state.json right?\r\n\r\nYes exactly","embeddings":[-0.2718343139,-0.1161887571,-0.0352353342,0.2897481322,0.173390314,0.0210333075,-0.0633087233,-0.0184237435,0.0392631516,-0.0410164855,0.0878003612,0.3908457756,0.1072007939,0.2664932609,0.2687342763,0.211744383,0.2464091033,0.2824453413,-0.3475175202,-0.1113381907,-0.1542824507,-0.2085547596,-0.0193535276,-0.2514696419,-0.1910442561,-0.3410944939,-0.1070777848,0.0550615005,0.0983739793,-0.0121028097,0.1449960321,0.1127832681,0.2192594856,0.2422087193,-0.0001218375,0.0475495383,-0.3169622719,-0.1216885075,-0.2325944752,-0.1641324461,-0.0777008533,-0.1760741472,0.0788733661,-0.3348271847,0.0588452555,-0.2760922611,-0.1388326436,-0.2848832607,0.7580259442,-0.2513796091,0.0563447028,-0.0585998297,-0.1987206042,0.1199141592,-0.3091675937,0.4043109715,0.2187694162,0.1050173119,0.0503210835,0.36747244,-0.093694292,0.1117038429,-0.2461740226,-0.207431674,0.3435823619,-0.0634660646,0.3340567946,-0.269988358,0.1193195358,0.0136377849,0.7991198897,-0.3812215328,-0.1796463281,-0.2513909638,0.1178055927,-0.0368309356,0.1927764565,0.2262728661,-0.0248626806,0.1718224734,-0.4951219559,-0.8388234973,-0.1489239335,0.1190174967,0.1009291112,-0.5302556753,-0.0082670115,0.1453509182,0.1877666563,0.2782385051,0.5109779835,-0.3100205958,-0.2236690372,0.1894250065,-0.2346062213,-0.2521696985,-0.3452332616,0.0221534148,0.000602995,0.2887879312,0.1636209935,-0.0569021031,-0.4284398854,0.1628808677,0.2259630561,0.3298417628,0.1668211371,0.1453311592,0.0563375354,-0.044476755,-0.2209341377,-0.0619554855,0.1491920203,-0.0276772399,0.4737729132,-0.2588220239,0.163620308,0.0016253302,0.1400839388,0.0495912693,0.0357529819,-0.1088385358,-0.1309854388,0.2539918423,0.1987355053,0.2483858168,0.2178698033,0.1705151498,0.0197667461,-0.0209412146,-0.1334893703,-0.0617986806,-0.2309666127,0.4641858935,0.069070369,-0.0887325928,0.2188076824,0.4066779613,-0.4791705012,-0.0771632865,0.0510129929,-0.1394921839,0.4023084641,0.2048662305,0.1272922009,0.1724875122,-0.041722931,-0.1416489631,-0.1382184327,0.5596106052,-0.3508445024,-0.1457885504,0.0306995474,0.090330407,-0.2036074102,0.1293104589,-0.81440413,0.1050892994,0.0007868717,-0.0853074268,0.2230197489,0.0610768385,-0.2575233579,-0.2021159381,-0.0737195611,0.2209573984,-0.4382852614,-0.2139052898,0.1710943431,-0.2812177539,-0.1509940326,0.261719048,-0.4043684602,0.326295495,-0.2754833698,-0.2434694022,0.6665347219,0.0070560472,-0.3046990037,0.2079366595,-0.031089697,-0.1648170799,0.0657574385,0.4862804413,0.1443699747,-0.0609247983,-0.3693814576,0.2963362336,0.215858683,0.0800246298,0.0464105271,-0.2589761019,0.131042257,-0.2477623522,-0.211529538,0.378969878,0.2782936394,0.3037509918,0.3548749387,0.0152990418,0.1266255528,0.4119395614,-0.1069112271,0.0315830819,-0.1939610094,0.1638491452,-0.6857842207,-0.0042512799,0.2076853067,-0.7038931847,0.3453542888,-0.086630784,-0.0253844168,-0.1518404037,-0.0710496306,0.104336746,-0.0901051983,-0.05074903,-0.0171570629,-0.1545765549,-0.3196608722,0.3492329121,-0.003164947,0.0888069496,-0.5791190267,0.4158156812,0.2212296277,-0.1650812328,-0.2256940752,-0.2001319677,0.1829229444,-0.0074457042,-0.0901779756,0.3612886667,0.2404170632,0.4049839377,0.0921566561,0.2112299055,0.0593833737,0.0135781486,0.150851354,-0.1268196255,0.1192975789,-0.1198687926,-0.1908881962,0.0652803108,-0.2321770042,0.1572923809,0.0765929222,-0.1667610705,0.0335816666,-0.3088069856,-0.0480581298,-0.2288487703,-0.2117875218,0.1583105773,0.4406870902,0.1159684137,-0.1591977924,-0.1882124841,0.2952923477,-0.2909323871,-0.0262471791,-0.0537171066,-0.3521838486,0.0089244274,-0.093455635,0.5716808438,0.1758241653,0.0644302666,0.0659166053,0.1340860128,0.1895989627,-0.0933209211,0.0836411044,-0.041623313,0.6041048765,0.2483198941,-0.0612275153,-0.0569608025,-0.1651546061,0.3208298981,0.1617819965,-0.1070077345,-0.3039286733,0.0892878398,-0.2373192459,0.068352975,-0.5137993693,-0.0907843709,-0.2132945508,0.1997136176,-0.2643348873,0.7351220846,0.2415581048,0.1234899536,0.1471776664,0.0484481268,-0.1123831347,-0.4336647987,-0.1235333234,-0.0797572881,0.0217445511,-0.0706892312,0.1573005021,0.0970508382,0.3531443477,-0.0618230738,-0.0247243382,-0.3828346729,0.2122982889,-0.0122122653,0.1914204657,0.0478795804,-0.1822677106,0.2861307561,-0.4074433148,0.0005159195,-0.1111803576,-0.4047201276,-0.3357618749,-0.0125855133,-0.0166208651,0.2885576189,-0.0127109233,-0.1355067194,0.0438457429,-0.0531556308,0.0625010207,-0.1334987879,0.0965997726,-0.0791736171,0.2284816206,-0.0237409398,0.1923693269,0.0633351877,-0.2285635322,-0.595017314,0.4055571854,0.16817303,-0.1907067001,0.3224246204,-0.1593238413,0.0080415048,0.3026723266,-0.6242733598,-0.126975283,0.0876560509,0.0979647189,-0.4367262423,0.3287938833,0.3484051526,-0.0549789071,0.0121673243,-0.0437284447,-0.155940637,0.0839103386,0.1567784548,0.3712688088,-0.1370222718,0.2379801124,-0.0419514067,0.4168044329,0.0455463417,-0.1582577676,0.255551219,0.1512086987,0.3843092024,-0.1916628927,0.0029946556,-0.6070473194,-0.1174782664,-0.2734977007,-0.1383567601,-0.0366954505,0.0050278436,-0.0930868238,-0.2602549195,-0.2978343368,-0.1838683337,0.0744128451,-0.2486144602,0.2656247616,0.0099636829,0.189233765,0.1631921083,-0.1182047352,-0.0223084334,0.2692699134,0.3769269288,0.0224297587,-0.3797899187,-0.1179046035,-0.0901240185,0.1852074713,-0.1861287802,-0.1409107745,0.1490107775,-0.0084957713,0.1097094789,0.1873061657,0.8571017385,0.2298957556,0.1024272367,0.1174280047,-0.3584455848,-0.3792292774,0.119878009,0.0484821387,-0.3112246692,-0.1034067571,0.7384950519,0.1157798395,-0.1241323948,-0.4459344447,0.5131250024,-0.3287904859,-0.3560480475,-0.1932100058,-0.1531292945,-0.6175579429,-0.1648265719,-0.2992198765,-0.1718769968,-0.1339315027,0.041087918,-0.2182900608,-0.0384295098,-0.0045501608,-0.131248042,0.3865060806,0.2553060949,-0.0206534341,0.1911816895,0.1961998791,0.2972922325,0.3267876506,0.0544974022,-0.0128217218,-0.1107505336,0.0589022636,0.1450710893,0.1571551561,0.0703206733,0.0520314239,0.0676189065,-0.1939865798,-0.2812299728,-0.0230429657,-0.2166571319,-0.1420742422,-0.1258673817,-0.5997289419,0.5338152647,-0.0672864914,-0.1901238263,-0.0784502476,-0.0442128181,-0.2895660996,0.2332212776,0.1667561382,0.890385747,0.0526470877,0.2017977536,0.1241298616,-0.2553425133,0.0394476168,0.0662134513,-0.1410413235,-0.3871965706,-0.1446400583,-0.217621848,-0.2792264819,0.0614954717,0.4207955897,-0.2487496585,0.3165193796,-0.1748750955,-0.1440926939,-0.2096947581,0.0603321269,-0.1172027513,-0.1681214571,-0.1699972004,-0.0213871039,-0.1236260533,-0.020981418,-0.0011770085,0.0230336059,0.085429728,0.0010389432,0.3451480865,0.0560773015,-0.0831320956,0.3321541548,-0.1623865217,-0.4522760808,-0.054744374,0.5875549912,0.3452653587,-0.0441728532,0.0292595942,-0.0739375129,0.0948239639,0.3109366,-0.2370249778,-0.1441314667,0.4764250815,0.1349079758,-0.0889207944,0.1434366554,-0.0577835664,-0.2296339571,-0.0459273867,-0.0453845412,0.0363203958,-0.213347435,0.052167438,0.2864359319,0.2828983963,0.0115315169,0.0111031784,0.1509435922,0.0376832336,0.2617278695,0.0662290305,-0.231271565,0.081133388,0.580947876,0.2229579687,0.0512588732,0.3897204399,-0.2858748734,0.1089943051,-0.2599898875,0.5443816781,0.1697918773,-0.1922824234,0.1979249865,0.0437701531,0.3796108365,-0.05466526,-0.0060143895,0.0673477128,0.0287794862,0.1426058114,-0.0758740604,-0.4620836675,0.222335875,0.2451365888,0.3159635663,0.2951500416,-0.1403969973,-0.1127021611,0.3135163188,-0.1744306833,0.2544686794,0.3054512441,0.2604134083,0.1035924703,0.1888873726,-0.2063466012,-0.3922478259,-0.062478289,-0.2128615379,-0.2844782472,-0.1424082965,-0.1012736484,0.1670764983,-0.050453268,0.1956088841,-0.3279799819,-0.5155194998,0.260212183,-0.1302507669,0.1250554323,0.0721867979,0.2125461698,0.2162415236,0.1530762911,0.4170274436,-0.3205250502,-0.1202609167,-0.2072765231,0.1700719446,-0.0368493423,-0.012890772,0.0030831397,-0.1040381119,-0.6347949505,-0.0721644163,0.5802595615,0.228553921,0.0028292465,-0.1216733158,-0.0890728384,0.0781302899,0.2832781971,0.1795536876,-0.3574940562,-0.2113585621,0.3531332016,0.1090220138,0.0320583731,-0.2015241385,0.257967025,-0.1630499363,-0.2899367511,0.1263776571,0.0541124605,-0.1016248241,-0.3947136998,0.0499629267,0.1948272735,-0.1795295179,0.2764655054,0.5871967673,-0.2222953588,-0.0410539918,0.1395605505,0.4255005419,0.0143888379,0.3339314759,0.0543658249,0.3362127841,0.0444846228,0.1634128839,-0.0770832375,-0.3645168543,-0.0481443405,0.1652766615,-0.2980026901,0.0461642742,0.1244694293,-0.0286287442,0.2054723799,-0.3259216547,-0.1062838882,0.3587942123,-0.1409890354,0.088929899,-0.1632903367,-0.1057851389,0.2865636051,0.2876821756,-0.0562374331,-0.2394174635,0.3689659536,0.1718235165,-0.067410402,0.1448661387,0.1560452282,0.1954452544,-0.1644708514,-0.1904947311,-0.060328614,-0.0041764555,0.0154811293,0.0102711208,-0.0681501925,-0.2158234417,0.0196701363,0.1003510356,0.2574125826,0.2494761497,-0.0933900848,0.0989125669,0.0373016559,-0.0037809776,-0.069940336,0.175513804,0.0142176328,0.042904716,0.2642636895,0.04912211,0.4381881654,-0.3555495143,0.2491775602,0.2990668714,-0.1117345989,-0.1690796912,-0.0314141437,0.1784495562,-0.0467247777,0.4658777714,-0.3290247321,0.0173482783,0.0306533016,0.0099924756,-0.047554858,0.4196546972,0.203335613,-0.4373653829,-0.1728261709,-0.1898131669,-0.529168725,-0.1232097,0.3429064751,0.2254955918,-0.250320822,0.1693479121,0.1571043432,0.2181128263,-0.0132866101,-0.1186592281,0.1006056517,0.3395012021,0.1501172781,-0.2381059676,0.0286989491,0.3330336511,-0.1056633145,-0.2288174778,0.210751906,0.1350077242,-0.0558573678,0.2134573609,0.1828323901,-0.1485542357,0.4979897738,0.6290314794,0.0061495891,-0.0157596599,-0.1853654981,0.3123157322,0.3738644421,-0.0125719784,-0.1405423433,-0.0493751243,0.0467535071,0.2047130913,-0.4784535468,0.4090868235,-0.1684514582,-0.180833146,-0.0156435613,-0.0482163206,-0.5606992245,0.0552172661,0.0289150551,0.152665168,0.1520670056,0.3861209452,-0.0366068371,0.2796961367,-0.0986855999,0.0191101227,0.4286305308,-0.1547548473,-0.1903103143,0.0153500875,0.273360312,0.1002365947,-0.0246258322,-0.3998605013,0.0855823234,0.4397142231,-0.2660641074,-0.1272836775,0.4620869458,-0.1776352972,0.1517995447,0.036894571,0.5529010892,0.0806337744,-0.4112178087,-0.1290168315,-0.3285099268]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1993","title":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? ","comments":"Perfect. For now, I am loading the dataset from CSV in my interactive process and will wait until you make the PR!","body":"I am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n","comment_length":22,"text":"How to load a dataset with load_from disk and save it again after doing transformations without changing the original? \nI am using the latest datasets library. In my work, I first use **load_from_disk** to load a data set that contains 3.8Gb information. Then during my training process, I update that dataset object and add new elements and save it in a different place. \r\n\r\nWhen I save the dataset with **save_to_disk**, the original dataset which is already in the disk also gets updated. I do not want to update it. How to prevent from this?\r\n\nPerfect. For now, I am loading the dataset from CSV in my interactive process and will wait until you make the PR!","embeddings":[-0.3435497284,-0.1861001253,-0.0237765033,0.1979099065,0.235196054,0.0903734267,-0.0758549795,-0.0891268402,-0.0024480005,0.0613528565,0.0818682164,0.288820982,0.0089173801,0.2742179036,0.211016655,0.2311643064,0.2571016252,0.3034494519,-0.3161807954,-0.0310900286,-0.1955247968,-0.1905546933,0.0157797616,-0.2773298025,-0.1335471421,-0.3582568467,-0.0840454772,-0.0099302558,0.0723503977,0.0670775324,0.1237914488,0.1421202421,0.3197945058,0.3073055744,-0.0001188234,0.0112640066,-0.3935620785,-0.1174851358,-0.2952136695,-0.1752285957,0.0198726803,-0.1928234398,0.0426131189,-0.282327652,-0.0355925113,-0.0789245367,-0.1382195801,-0.1842427701,0.6355670094,-0.2757766247,0.0882279724,-0.1013195962,-0.2604755163,0.1003407985,-0.3609260023,0.3109507263,0.1697656363,0.0327253751,-0.0230117068,0.3523843884,-0.0059232889,0.1834039837,-0.2202827185,-0.2364939749,0.368504107,-0.0557336733,0.1962572187,-0.3266142011,0.1366037875,-0.050418485,0.7573289275,-0.3973721862,-0.2224439383,-0.2200314552,0.2223112136,-0.1780084819,0.1028199121,0.2710522115,-0.0004193895,0.1309111714,-0.5915709138,-0.8006381392,-0.1265453547,0.0599240996,0.0174780302,-0.6005370617,-0.0098792957,0.1061759517,0.2177744657,0.2815048099,0.4598244429,-0.2353613228,-0.2161782384,0.1422934383,-0.294994086,-0.2745671868,-0.2943738103,0.1907633543,-0.0581794083,0.1961970776,0.0701111555,-0.0923652202,-0.2840215266,0.1482525617,0.1528834552,0.3160923719,0.1148205847,0.180868119,0.0832407027,0.0590583868,-0.2331293374,-0.0536452942,0.1533101946,0.0603940263,0.5767165422,-0.2273498625,0.1721357107,0.0239444021,0.0714495629,0.0731745958,0.1662715375,-0.0902282894,-0.117442593,0.3433509171,0.1202501431,0.2047808617,0.1899908036,0.2010244429,-0.0489036813,-0.1141749471,-0.1551879346,-0.0595462397,-0.3120179772,0.41968894,0.1253365278,0.0080648456,0.1199120879,0.5349015594,-0.4715520144,-0.1450412124,0.0144295339,-0.1974763125,0.3453066945,0.1837601215,0.1647830755,0.1672118604,-0.0709511787,-0.0298968963,-0.1003757268,0.6497283578,-0.3161403537,-0.1092399731,0.1016733795,0.1262910515,-0.1379485428,0.1107459962,-0.7127209306,0.1979913712,0.0198915936,-0.1377772242,0.160496667,0.0238329396,-0.3551610112,-0.1828225255,-0.0532662161,0.1804096848,-0.38580212,-0.1600945145,0.1927725524,-0.2199116349,-0.1124547571,0.2244847864,-0.3141235113,0.2418796122,-0.2732774913,-0.357734561,0.5291256309,0.1001327857,-0.3341226578,0.1677851826,-0.0348184444,-0.0790609345,0.0231738556,0.5614807606,0.1681259871,-0.1092801839,-0.4543691874,0.3388874233,0.0994205475,0.1335136443,0.0695958138,-0.1695295423,0.2710565031,-0.311650902,-0.1427363902,0.480778873,0.2460567653,0.3417785764,0.3239541948,-0.024512453,0.0995803922,0.407705605,-0.1512080431,0.0790723637,-0.2095583975,0.1262258887,-0.6138045788,-0.0311102159,0.1936899126,-0.5951254368,0.3615722954,-0.1226511449,-0.0642847121,-0.2044179887,-0.0127755674,0.0749249533,-0.0688732266,-0.002293925,-0.1109882668,-0.1776316017,-0.2606809735,0.4837686718,-0.0590841062,0.0931875706,-0.6138254404,0.3396213949,0.1712860465,-0.2045485526,-0.2430604994,-0.0967049971,0.2103912383,0.0319461636,-0.0962717235,0.3377920091,0.1939775348,0.4088332057,0.0129824532,0.1993591636,0.0663033053,-0.0132702179,0.0733956248,-0.2320130467,0.136041373,-0.0811839551,-0.1756194234,0.0617034361,-0.2363677323,0.0892639831,0.1581325233,-0.204508841,0.0544074439,-0.3416891694,0.0190343335,-0.2046752274,-0.1858683228,0.2192384452,0.3088561893,0.1363543272,-0.226831913,-0.2074437737,0.2330199033,-0.2737059891,-0.0912187025,0.0353360139,-0.2595542967,0.0168843996,-0.1204081327,0.5437600017,0.118866384,0.1275318563,0.0053325994,0.119435586,0.1251199394,0.0129794646,0.0339159742,-0.0343918912,0.5523133278,0.2647176087,-0.0509434678,0.0724444613,-0.1624070555,0.4198661447,0.1504427195,-0.1196854413,-0.3011409938,0.0994732305,-0.1615885943,0.0660834089,-0.4680760503,-0.0841383561,-0.2472438812,0.2366381884,-0.2964737117,0.7885241508,0.259465605,0.1625138223,0.1108180657,0.0459692441,-0.1312014014,-0.4180258214,-0.0573595986,0.0374375694,0.0125371478,-0.0483345091,0.075237073,-0.003702749,0.4305587113,-0.0291063748,-0.0665599629,-0.4269747138,0.1278461218,-0.0823732764,0.1029738337,0.1104518101,-0.2120501697,0.2974178791,-0.4168384969,-0.0541332215,-0.1244075149,-0.4201669693,-0.2594532669,0.0294520706,0.0546952635,0.261502713,0.0107821692,-0.2051879019,-0.0094756298,-0.0083108591,-0.1173740625,-0.1713139862,0.0468005538,0.0203509293,0.1569827646,-0.2946605682,0.2539893985,0.0822965875,-0.2376686633,-0.5970135927,0.3417876363,0.1599008292,-0.1892437637,0.2897326052,-0.1801794171,-0.0554160178,0.3803261817,-0.5858959556,-0.0945967212,0.0443135947,0.141354993,-0.4069656134,0.3347259164,0.308191359,-0.0225071684,-0.0081292214,-0.0677946508,-0.2454162687,0.0576772951,0.1356204152,0.3997930884,-0.2072359622,0.2203688174,-0.05733677,0.3448049128,-0.0246945042,-0.1513893008,0.2897460163,0.1356435865,0.424543947,-0.1676989496,-0.1270466447,-0.5524557233,-0.17350021,-0.235457018,-0.0060488256,-0.0379285254,0.0072150845,-0.1281888783,-0.2558845282,-0.2843938172,-0.130805999,0.0504188836,-0.24523139,0.3822573125,0.0587057583,0.1576600373,0.1527377069,-0.1986500025,-0.0689637586,0.2094496638,0.3548832536,0.0272180643,-0.2600480616,-0.0348764956,-0.1532876939,0.1102515757,-0.2091602385,-0.0276745036,0.1221666485,-0.0030596075,0.0609425381,0.2545267344,0.8182939887,0.2156579196,0.1374484599,0.0889697745,-0.4251458347,-0.2478490174,0.0456906222,0.0322485454,-0.3078769445,-0.0940757841,0.6723995209,0.1526483446,-0.09020897,-0.3045021892,0.5188004375,-0.3563797772,-0.3737732768,-0.2266186029,-0.1663163304,-0.5012161732,-0.1288881004,-0.2598293424,-0.043999102,-0.1832844466,0.1061317548,-0.2196145356,-0.0676039532,0.1095384508,-0.1310002208,0.3300730586,0.3421667218,0.0496449284,0.1242490634,0.2176742703,0.1669383943,0.3173028231,-0.1053371131,-0.1126391068,-0.1280662566,0.1298919022,0.1712062806,0.0756075904,0.0527421273,0.0732389763,-0.0068590827,-0.1699391752,-0.3829432726,-0.0453667864,-0.1490216851,-0.0730891377,-0.2039042413,-0.4751245677,0.4746285975,-0.092576094,-0.1258590966,-0.1472214162,-0.1061394364,-0.2921000421,0.1399554759,0.1306450814,0.863239646,0.0326740146,0.1525593102,0.1198745966,-0.1705316454,0.0168492813,-0.0012273288,-0.0839210972,-0.4019954801,-0.3279488981,-0.1733226329,-0.1934763044,0.0378247239,0.3253153265,-0.3283300102,0.3507508636,-0.2487777174,-0.1089887097,-0.1078147292,0.0443412848,-0.1169934422,-0.1649635136,-0.1856908798,-0.006143176,-0.0798516124,-0.0044733244,-0.0543574244,0.0740772188,0.2115780711,0.0019400406,0.3221925795,0.1227276772,-0.0707127526,0.2805036902,-0.1962251663,-0.4649519622,-0.0051138531,0.6101840138,0.4340718389,0.0248929802,0.0813790485,-0.0706977025,-0.026658101,0.331471175,-0.2104783654,-0.1034401506,0.4504612088,0.1366672516,-0.1030479074,0.0925333649,-0.0474378355,-0.1213083565,0.019843597,-0.0646405369,0.1135451943,-0.3509952426,0.0355062075,0.2403179705,0.3644151092,-0.0586991049,0.0314691402,0.0625162497,0.0130531089,0.3171579242,0.0708596855,-0.2808800638,0.0376206525,0.5294992328,0.2024803907,-0.0426782258,0.3699873984,-0.2707158625,0.0899938866,-0.267544508,0.4640224278,0.1178866401,-0.1746014357,0.1856891364,0.1257218122,0.4507123232,-0.0159635972,-0.0439136177,0.0883902609,-0.062978901,0.1653642803,-0.0940570459,-0.3822163343,0.1210728213,0.2372953147,0.3630728424,0.3535116613,-0.1228337809,-0.0329008587,0.3971119821,-0.2104936391,0.1599497348,0.2804693878,0.2849237025,0.2143125832,0.1544959694,-0.3230305314,-0.3599324226,-0.0096668908,-0.2506662011,-0.2394524515,-0.1587037891,-0.0525229387,0.1439282447,-0.1139955521,0.2321478575,-0.2854385078,-0.4382475615,0.2356998175,-0.1452858895,0.0925420597,0.1704600453,0.2144286931,0.2009414583,0.0785360038,0.426656276,-0.3415497839,-0.1125037298,-0.2172945589,0.1889526397,-0.0322937369,0.005121504,0.0702451915,-0.1169623658,-0.6251552105,-0.0948788077,0.6250933409,0.1212151125,-0.0073651513,-0.1670639366,0.0100223515,0.153445825,0.2904006243,0.0841554627,-0.3773589432,-0.2461010516,0.377358526,0.1504216939,-0.095623076,-0.1158768013,0.2852276266,-0.2092484236,-0.2813018262,0.1848055869,0.1058103591,-0.0454825945,-0.3390641212,0.0373214222,0.255398035,-0.168448329,0.3619692326,0.5913919806,-0.2182156593,-0.0562761687,0.139661938,0.4150112867,-0.1396206915,0.3794614375,-0.0036846767,0.381894201,-0.057031706,0.1477711052,-0.109039627,-0.4267403185,0.1011379063,0.165410459,-0.2610379159,0.0225983057,-0.0011140646,-0.1030327305,0.1897853613,-0.3766215444,-0.1189317778,0.4679526091,-0.2189384252,0.1519278139,-0.0048864181,-0.1219151318,0.2515875399,0.3047000468,-0.0060800333,-0.2621264756,0.3338620663,0.1696125716,-0.1097286344,0.0838387012,0.0973997712,0.2091323733,-0.1747378707,-0.1729257256,0.04247142,-0.0594643429,0.0830611736,0.0510298386,0.0068148361,-0.2014681399,0.068335481,0.1713885069,0.2928634286,0.2508704662,-0.0251089614,0.129016906,0.0307842735,-0.0182990171,0.0967060775,0.1933039278,0.0061853509,0.0254682787,0.263940841,0.0461047813,0.3334491849,-0.3797313571,0.206874609,0.4354177415,-0.0902820453,-0.2419116199,0.0280021355,0.2125843763,0.0150209153,0.5192960501,-0.2999186814,0.0126510644,0.0953990296,0.0301374234,0.0329528041,0.3783248663,0.2253105789,-0.3717052937,-0.300845623,-0.2156757116,-0.5639156103,-0.083594203,0.36889413,0.2425530255,-0.238333419,0.0979686305,0.0944174677,0.1586007476,0.0331653357,-0.1553775519,0.1296761334,0.383538425,0.1775435209,-0.2048424929,-0.003446331,0.3353901207,-0.1584968716,-0.192084372,0.1638853997,0.0912283361,-0.0265973862,0.1800369024,0.1969441921,-0.1996730566,0.4232634604,0.5758475065,0.0331577063,0.0936796963,-0.2577305734,0.326477617,0.4559466243,-0.0390745625,-0.1805486381,-0.1083184257,0.1445316225,0.1627392024,-0.4258211553,0.3830116987,-0.2226230204,-0.1333855093,0.0844240636,0.0172237381,-0.5296295285,0.0026823692,0.1237283722,0.1853373349,0.1231872588,0.4071831703,-0.020855194,0.3346155584,-0.1735910177,-0.1036755666,0.3725511432,-0.1448468566,-0.2540253401,-0.1261511147,0.3549654782,-0.0646672696,-0.019083472,-0.3931910396,0.1495354921,0.4303222895,-0.2442901134,-0.0405435264,0.5292304158,-0.1951893568,0.1532835811,0.1000508741,0.5462207794,0.1079475805,-0.3573218286,-0.0870739743,-0.2885380387]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1992","title":"`datasets.map` multi processing much slower than single processing ","comments":"Hi @hwijeen, you might want to look at issues #1796 and #1949. I think it could be something related to the I\/O operations being performed.","body":"Hi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n","comment_length":25,"text":"`datasets.map` multi processing much slower than single processing \nHi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n\nHi @hwijeen, you might want to look at issues #1796 and #1949. I think it could be something related to the I\/O operations being performed.","embeddings":[-0.4131574631,-0.3192059696,-0.0879166499,0.3574213982,-0.1027497426,0.0203935038,0.3424057961,0.1157955751,0.0579245947,-0.0040248358,0.065698877,0.414229691,0.186818257,0.2118217349,-0.1675523669,0.0080591897,0.1905092746,-0.0455829203,0.1740608811,-0.0060444372,-0.1642414182,0.1392722279,-0.4981355667,-0.0023563658,-0.416852802,-0.1493239552,0.1035198569,0.1356756389,-0.1574589014,-0.2285336107,-0.2626928687,0.1599624604,-0.0759368762,0.5732403994,-0.0001228117,-0.1976781934,0.0128008407,0.2621522248,0.0754679441,0.0612592809,-0.1325607896,-0.2765828371,-0.049440071,-0.1092961952,0.1161134019,0.2015769333,-0.0607613437,-0.4477421343,0.0259235501,0.0857624561,0.0549793281,0.3612498343,-0.3099785149,0.0393089876,-0.3776819408,0.1687246561,-0.1538441032,0.050548397,0.2789547443,-0.1809555143,-0.1093195379,0.3070887327,-0.0712712407,0.2966097295,0.0082043353,-0.082981579,0.0533868819,-0.450948894,0.18975465,0.3490732908,0.0353319421,-0.0632039234,-0.1647021174,-0.2374163568,-0.3268585801,-0.1349690557,0.202265054,0.1401260793,0.1821072549,-0.0051691858,-0.5538257957,0.1473838985,0.3451812565,-0.0721948594,-0.1189694703,0.0112252105,0.1275280416,0.3288134634,0.3404774368,0.1203624234,-0.050015267,-0.2203397751,0.319372952,0.2178269476,-0.7854886651,-0.0352138765,0.2125976831,-0.1020861268,-0.0854895785,-0.1859987378,-0.1023483351,0.3468084037,-0.3305373192,0.0954948738,0.3070702255,-0.0764189139,0.0073991055,0.1116778031,-0.0347225517,-0.1712293178,-0.3314741254,0.0631494969,0.0929078758,-0.3245918453,-0.0648111552,0.1292852163,-0.3052825928,-0.0903897285,-0.1549672335,0.0295762606,-0.2378287911,-0.1023422629,0.1326784045,0.0770512819,-0.0129047334,0.7939462066,-0.2598311007,0.1157042608,-0.472381264,-0.5671402812,-0.0263562519,-0.1367470026,-0.372112751,0.2238743752,0.1279514879,-0.0222434774,0.140732348,0.3258404434,0.050365489,-0.2243557423,0.3272614479,-0.5305342674,0.1972386986,0.0816774815,0.1585823596,0.5495990515,-0.1144087911,0.3264811635,-0.1295257509,0.2438604385,-0.5274049044,-0.2451353967,0.1805239618,-0.0436880216,0.1031669304,0.1064126492,-0.4637281895,0.4980573356,0.3198492229,-0.218412444,-0.1693590134,-0.199179545,-0.6358721852,-0.1516774148,-0.0613874942,0.167331323,-0.330935359,0.2380714715,-0.3223769963,0.1468189508,0.4350244403,0.6122487783,-0.1885802299,0.3844541609,0.0008836694,0.1984124482,-0.0064408001,-0.076507166,-0.3745238185,0.5169079304,-0.1592252403,-0.0011449816,-0.1906943768,0.1531788558,0.2403192073,-0.0547442548,0.3251417875,0.293664664,0.0304956846,0.3874751031,-0.2230064124,-0.1107543111,0.1460555792,0.0392807163,-0.0841827542,-0.0912581235,-0.013679794,-0.29399997,0.3163222075,-0.0127959643,0.0252214484,0.4302806556,-0.3920022845,-0.151980117,-0.147237286,-0.2302908748,-0.0732991621,0.3663209975,-0.0278735459,0.0501863472,0.4308092594,0.0351639725,0.1925522983,0.1138258949,-0.0405154489,-0.0642077401,-0.1118852496,-0.080577679,-0.1543217748,-0.1789610237,-0.0090868864,0.4500130415,0.1817546785,-0.1041791365,-0.0312027484,-0.1325511336,0.0812653601,0.0562892333,-0.1986998469,-0.0696268603,0.0854894966,0.0502698235,-0.0791463181,0.2618642449,0.4633000791,-0.0539056547,0.1709706336,0.1155771986,0.3728671074,-0.0540255979,0.0484472886,-0.1458726823,0.1021566167,-0.2390238941,-0.0239049681,0.312073946,0.0839073509,0.4422298372,0.0316376686,-0.1361154616,-0.0101459734,0.2965461612,0.1542539299,0.0568739846,0.3458305299,0.3308038712,0.2301621139,0.3350547254,-0.1322713941,0.3230473697,0.6291481256,0.0487079881,-0.3480742872,0.0437146612,-0.1839314401,-0.3695401549,0.0226528216,-0.1412669867,0.5598792434,0.0200398825,0.2318464816,-0.0362140015,-0.1006811336,-0.0587041229,-0.0230610687,0.1029917896,0.2264130712,-0.1050920933,0.295919925,-0.0159352459,-0.0601109825,-0.2421924025,0.3389692605,0.2527018189,-0.2600570619,0.1031630635,-0.2926941216,0.1555970013,0.1763422191,-0.100340791,-0.2901907265,-0.2128409296,-0.1558169425,-0.0844460279,0.2200267166,-0.0549233854,0.2770826519,-0.1361938566,-0.105542168,-0.0475386642,0.0010066793,-0.1382571906,-0.1498213112,-0.0109747723,0.2863363922,0.226347819,0.0418869965,-0.0677929968,-0.2715706229,-0.0647876933,-0.2076481581,-0.1236517504,0.1410822719,0.0555924624,-0.1938761622,-0.0573871136,-0.2245586514,0.153676793,0.1814560443,-0.2490199953,-0.189970836,0.0516333245,-0.1066285148,-0.2332419008,0.0450814515,-0.1557418853,-0.0870592445,-0.1077109352,0.2935895324,-0.1721993983,0.341969043,-0.2260829508,0.0634779185,-0.0797818005,-0.1079568714,0.015877638,-0.3331030905,-0.3667624295,-0.0119899632,-0.018404644,-0.192117095,-0.1344933957,0.0897425786,0.1568120271,0.1717814654,-0.1599652022,0.1699120104,-0.3907409012,0.1194331646,0.0705737844,0.0674815774,0.4676005542,0.0276590642,0.0626560152,0.0156416148,-0.4355051219,-0.0635614619,0.3192830384,0.0368367285,0.0172181092,0.3445580304,0.0313459262,0.6403149962,0.4947387576,-0.1645864695,0.2020886242,0.0096630668,-0.1716504693,-0.4027168453,-0.218011409,0.1240068749,-0.2614226639,0.1684180647,0.4269162118,0.0036421001,-0.429479599,-0.0166892968,0.2439737022,-0.2445280105,-0.0141028259,0.1481050402,-0.1232679486,0.1966449022,0.2461677045,-0.1976921409,-0.3519734144,-0.0353395939,-0.0661896691,-0.1060343459,-0.098615475,-0.2421052903,-0.4882799685,-0.0338981263,-0.3716017306,0.2611175776,0.0420062803,0.2969749272,-0.0439387448,0.1057715118,0.3316419125,0.0295935869,0.5544948578,-0.4164328873,-0.1365627497,0.1310188472,-0.3980678022,-0.2873519361,0.0707289428,-0.1707302481,0.2828143835,0.5884535313,0.5143984556,-0.0563279018,-0.205391705,0.0086564953,0.1212090552,0.1553758383,-0.266105175,-0.3546744585,0.1345768124,-0.1294958591,0.1268242449,-0.0423868448,0.1320046484,0.1232842505,-0.1420472115,0.019436555,0.1154948622,0.238198936,0.1704018414,0.1206151992,0.0930257142,0.3004200161,0.2673074901,-0.0723713413,0.0622349828,0.329382509,-0.1249066219,-0.0998571515,0.2132408172,0.0241130739,0.3280750215,0.319684118,0.1498575509,0.0306495111,0.1590183675,0.2774959505,-0.2244988531,0.2721706927,0.3960230052,0.3916812539,-0.4936520159,-0.5062215328,-0.0205668379,0.4649193883,-0.146079123,0.2714835405,-0.7679567337,0.0800636262,0.0503082201,0.0999305397,0.7852854133,-0.4731326699,0.0379717909,-0.3119589388,0.2154320031,-0.0481756367,-0.5948623419,0.1334774047,-0.1847289205,-0.3472980559,0.0275629815,-0.1090762392,0.2031953931,0.5102791786,0.1583227515,0.2661416829,0.3404269814,0.0784626901,-0.0262437332,0.3504178524,0.5003811717,-0.3187172711,0.1645903438,0.0327547193,0.1939356923,-0.207008794,0.0298594553,0.0704516321,-0.0611505546,0.0293347053,-0.3183437288,-0.2279880196,-0.39779827,0.1963718981,-0.0764122233,0.2826351523,0.3900783658,0.2677480578,-0.0713075623,0.0901603922,0.0085475463,0.0534890406,0.263794452,0.2334204018,0.3824106157,-0.4639774263,-0.0287042018,-0.1009592339,-0.2506047785,-0.2602870464,-0.1532426476,-0.1185093969,-0.1225385815,0.4491189718,0.1657283157,0.1570968479,-0.2264566272,0.2328821272,0.0790415108,-0.1099075004,-0.0125143826,0.0914813951,0.0223492179,0.6711222529,-0.2834818661,-0.5138155818,-0.034297511,0.3942315578,0.2499610633,-0.0952273905,0.0476648323,0.2129702121,-0.2575139701,-0.1355023682,0.1535299867,-0.1033581644,-0.0991862267,0.105922088,0.0208500307,-0.224525407,0.254693985,-0.1348883361,-0.1512796581,-0.0532309376,-0.1917673498,-0.2759402394,-0.2528984547,-0.180083558,-0.351654917,-0.2163468152,0.2313819081,0.0247284491,0.1297436208,0.4729893506,-0.1329039335,0.0583199151,0.0901111811,0.21108675,-0.0189659465,-0.2514662147,0.0779228583,-0.0258274451,-0.0658428147,0.0795580372,-0.0867979974,-0.1474020481,-0.0117546832,0.1851729304,-0.0224858411,-0.2631607056,0.329619348,-0.1456150413,-0.2043025941,-0.4302749038,-0.0622506663,0.0521588847,-0.0218544044,-0.1042599753,0.2791091204,-0.2983775139,-0.2712824345,0.4427229762,-0.2003646195,-0.0969033763,-0.0010559623,0.2870850265,0.4510101974,-0.0946663246,-0.100541912,0.0290755294,0.2644409835,0.0692050532,0.1374590397,-0.0965791047,0.1816248447,0.2666929364,0.3389314413,0.1272505969,0.1346454769,-0.2531648278,0.0023173788,0.001643437,-0.240311563,-0.1764900386,0.5670022368,0.0328633934,0.1380591542,0.051613234,0.3361942768,0.2363513857,-0.0093842614,-0.1060245261,0.081082359,-0.2906021476,0.0611946732,0.1536391228,0.2585486174,0.2591066658,0.4466488063,0.1197482124,-0.0015337064,0.4460707009,0.2742266953,0.2714106739,0.4872277677,0.2653108239,-0.0395575166,-0.3783236742,0.2097000927,0.503713727,-0.3329574168,-0.0574922487,-0.0578089394,0.2234261334,-0.2494561374,-0.3842220604,-0.2609777451,0.3868533075,-0.1655793786,-0.2582350075,0.2212099135,-0.0687714815,-0.0023638823,0.1776569188,-0.2204602063,0.1424396485,0.7928742766,0.130153209,0.0441229455,-0.3891386688,-0.3411386907,-0.1499896944,0.1697582155,-0.1381987184,0.182852596,-0.3203031719,-0.0100657903,-0.0007450023,0.3008362353,0.3764934242,0.2772647738,-0.1428851932,0.1006296575,-0.0221898481,0.1612599492,0.0074004908,0.182881102,0.0322884545,0.1748710573,0.165418312,-0.1053682864,-0.0901593789,-0.4533826113,0.0302179214,0.3537489772,-0.14939408,0.1578140706,-0.3038739562,-0.1554052383,0.0109660095,0.2941293716,-0.2735160589,-0.0898395628,0.4953567982,-0.4910948277,0.0730638951,-0.1513822079,0.0253128223,-0.0798000395,0.5650858283,0.3381140828,0.0105567249,-0.4115411043,-0.1151144058,-0.5973332524,-0.0711948723,-0.32060498,0.2698095441,0.0732985064,0.2909341156,0.0744075775,0.1212652996,0.4038819075,-0.0933897495,-0.0835429057,0.3543788791,-0.3133423328,0.4769584239,-0.2434433699,-0.1946042776,-0.0849098787,-0.4186908007,0.3781765103,0.0753604621,-0.1486230046,-0.1737609357,-0.0027932953,0.1544392407,0.1478449404,0.1885033846,0.3548042178,0.3634973466,0.0412941873,0.0078075365,-0.1190250218,0.1936971396,-0.0414553955,0.3311108053,-0.0583353229,0.0689003542,-0.3055815697,0.1666466296,0.1130410656,-0.1093769521,0.0038674194,-0.2370978594,-0.3271932602,-0.0505090691,-0.0389764123,0.3171595931,0.0952741802,0.4033234417,0.1399891526,-0.0934452862,-0.1850916147,-0.0801377594,-0.0288043879,-0.2668254972,-0.2905770838,-0.4305563867,0.2277179211,-0.2528906465,0.1613218039,-0.2331474423,-0.3237475455,0.3040854633,-0.0249335151,-0.368162185,0.3654679954,-0.3870623112,-0.0856687576,-0.1473059952,0.0290635452,-0.0981446728,-0.4448868334,0.1913250536,-0.1197809279]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1992","title":"`datasets.map` multi processing much slower than single processing ","comments":"I see that many people are experiencing the same issue. Is this problem considered an \"official\" bug that is worth a closer look? @lhoestq","body":"Hi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n","comment_length":24,"text":"`datasets.map` multi processing much slower than single processing \nHi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n\nI see that many people are experiencing the same issue. Is this problem considered an \"official\" bug that is worth a closer look? @lhoestq","embeddings":[-0.3866343796,-0.2725809813,-0.0787934735,0.3493690193,-0.1060922891,-0.0051658023,0.3705376983,0.1268497705,0.0547803566,-0.0164260231,0.0762916431,0.4323446155,0.1645940393,0.1775638312,-0.1425016224,0.0690406859,0.2205799669,-0.0627327189,0.1990935057,-0.0131204696,-0.1687246114,0.1928426474,-0.4963891208,-0.0021656686,-0.4106545448,-0.1248212904,0.1040452495,0.104869172,-0.135728687,-0.2046503723,-0.2140667439,0.1370120496,-0.1355689913,0.5738759041,-0.0001229709,-0.1984731257,0.0549875014,0.2633066475,0.0764158443,0.0830866322,-0.1329540014,-0.2601753473,-0.0424126312,-0.0829153508,0.1244733259,0.2161737382,-0.0585636832,-0.4646043777,-0.0096629243,0.0843353644,0.0631441474,0.3764816523,-0.2817437351,0.0143860718,-0.3346056044,0.1581185609,-0.1740894318,0.0507874526,0.2852863073,-0.1864520311,-0.1175598055,0.3048100173,-0.0474985726,0.3209557533,0.0011856457,-0.032223843,0.0270203575,-0.4442287683,0.1620766073,0.3706927299,0.0055631162,-0.0727401003,-0.1854880899,-0.2690021992,-0.3177765608,-0.1561565101,0.2311840802,0.128877297,0.1493102908,-0.0132253533,-0.5340886116,0.137452364,0.3866601288,-0.0713502467,-0.1523336172,0.0201121643,0.1103035137,0.3094667494,0.3142451644,0.1228403747,-0.0455443561,-0.1715080887,0.2885457575,0.1829549372,-0.7955273986,-0.0417103432,0.2489290684,-0.1129598469,-0.0615523718,-0.1814656854,-0.0985894799,0.3503522575,-0.334012717,0.1081213728,0.3353157341,-0.0971493945,0.0022064412,0.0807370842,0.0160701834,-0.1644104421,-0.3038092554,0.1016722172,0.1055112407,-0.3424444795,-0.0712615848,0.1095166728,-0.2824561596,-0.0873280242,-0.1478875875,0.0589937903,-0.215959996,-0.0632729158,0.1097481698,0.0793705136,-0.0192513689,0.8041564822,-0.2690217495,0.0870174095,-0.4744417071,-0.5820878148,-0.0185751952,-0.1468966454,-0.3591332734,0.1967960298,0.1042194664,-0.0287775639,0.1002882794,0.3114521801,0.0865714997,-0.227888003,0.3016511798,-0.4975840747,0.2251563221,0.1189085171,0.147158891,0.5624080896,-0.1207757592,0.3020247519,-0.1247043014,0.2404897362,-0.53864187,-0.226262033,0.139010787,-0.0469704196,0.0906334221,0.1150415838,-0.4588497281,0.4970785081,0.3395263851,-0.2471324801,-0.1647280008,-0.1944390237,-0.6674047112,-0.1534185112,-0.0376286544,0.1841220856,-0.3353411257,0.2206213027,-0.3466385603,0.1560323089,0.4348999858,0.6155382991,-0.1899303943,0.357547313,-0.0506498404,0.2021543533,-0.0163856633,-0.039771121,-0.3898433447,0.5365133882,-0.1695716828,0.0112684853,-0.2257880867,0.1606897265,0.2166389525,-0.0608123802,0.3597379327,0.2520622909,0.0037233033,0.4067167938,-0.2268607616,-0.1484753937,0.1806099713,0.0356511325,-0.0509552695,-0.1092451662,-0.0207675043,-0.2912405431,0.3273311257,0.0346136056,0.0190522056,0.4201274812,-0.3767901063,-0.1781177968,-0.1492089331,-0.2470980287,-0.0693643987,0.3475503325,-0.0082942108,0.0345355272,0.4530823231,0.0276290234,0.2079565078,0.1241186559,-0.0410281569,-0.0788135156,-0.119667992,-0.0874338523,-0.1914291978,-0.1583937407,-0.0411385261,0.451487571,0.2057485431,-0.1229939982,-0.0540051162,-0.1302855015,0.0977361202,0.076828219,-0.1990164667,-0.082281217,0.0642756745,0.0279882308,-0.0868979022,0.268009603,0.4869925976,-0.0700863525,0.1350846589,0.087985158,0.4050623477,-0.0231117141,-0.017242007,-0.1115205511,0.1151866391,-0.2409275323,-0.033332102,0.3163405061,0.115886271,0.3995565772,0.0307296328,-0.1051464677,0.0267663766,0.285171032,0.1078848243,0.0460056402,0.3108842373,0.3136247694,0.2151776105,0.3205104172,-0.1061691269,0.3456167579,0.6657009125,0.0833152011,-0.3104912043,0.049639795,-0.1518193334,-0.3581573963,0.0141462162,-0.1226074845,0.5757823586,0.0022508441,0.2472881228,-0.0382558778,-0.1148135141,-0.0809825361,-0.0141083207,0.1080095023,0.2107912749,-0.0710323378,0.3324417472,-0.0037922249,-0.1000336036,-0.2148354799,0.3497828245,0.2076937854,-0.2833828628,0.1074738652,-0.3034688532,0.2022354752,0.159931466,-0.1036078706,-0.301363945,-0.2368092686,-0.1221833304,-0.0886198133,0.185317412,-0.0142413387,0.2501130104,-0.0963020176,-0.1153759509,-0.0365741216,0.0350124799,-0.1128342673,-0.1560254097,-0.0114225792,0.2804483175,0.1901462525,0.0729656518,-0.0563285835,-0.2777389586,-0.124676235,-0.2436757535,-0.093754828,0.1492806971,0.0766111836,-0.1636891663,-0.1074149162,-0.2078325152,0.1690101177,0.1589678377,-0.3020038605,-0.2018781304,0.0461159386,-0.118549116,-0.2314856648,-0.0106227342,-0.1588877589,-0.0949640647,-0.109947525,0.2879401445,-0.1672471613,0.319992125,-0.1932982057,0.0527212024,-0.0889679864,-0.1160119101,-0.0425557047,-0.3604641557,-0.3457473516,-0.0074071651,-0.0419457816,-0.225216791,-0.1405963302,0.0771877766,0.1546166092,0.1581604183,-0.1823362708,0.1749543995,-0.3803266883,0.1397097111,0.086909838,0.0220161304,0.4934562147,-0.0037517208,0.0688597411,0.0072936276,-0.4403322637,-0.0629743114,0.2977604866,0.0230549797,0.0195931364,0.3518145382,0.0513639562,0.6315261722,0.48890993,-0.1507582814,0.1987278908,-0.0477831364,-0.1115790382,-0.399653703,-0.2119543254,0.1117125526,-0.2826229632,0.1254270971,0.4168576002,-0.0117396079,-0.4363338649,-0.0093134958,0.2886060476,-0.2262658477,-0.0335349143,0.1407323331,-0.1364978999,0.1556614339,0.2492784411,-0.1519773751,-0.3358052671,-0.0443015806,-0.0655479133,-0.1016883999,-0.1039938629,-0.2754542828,-0.4782886207,-0.011619634,-0.3779994547,0.2300857157,0.0478082262,0.3498315811,-0.0442245342,0.0869569182,0.3212327659,-0.0168989636,0.5688456297,-0.391073525,-0.1088192463,0.1516809016,-0.3594423234,-0.2883349061,0.0306875408,-0.1507093012,0.2883810103,0.6098175049,0.5065738559,-0.0416915566,-0.1537956893,0.0076695345,0.0976752713,0.1281900108,-0.2506240606,-0.3729524314,0.1217523217,-0.0959062576,0.1306332201,0.014318862,0.1115724444,0.1250599623,-0.1616070569,0.0231329035,0.0852763951,0.2221221924,0.1544780731,0.1185791716,0.0978058875,0.3397481143,0.253183037,-0.0906116739,0.0212656111,0.3452213109,-0.1188256443,-0.1019111574,0.275282532,0.0147789288,0.3062773943,0.3589968383,0.1124797389,0.0289847869,0.1859190911,0.2604249716,-0.2158243954,0.2447054833,0.4137657285,0.3523693681,-0.4705941379,-0.4870205522,-0.0180302225,0.4810301065,-0.1415031999,0.3186241388,-0.7869355083,0.0805432573,0.0308834296,0.068520315,0.7341234684,-0.4244520962,0.0844886228,-0.2732584178,0.2042396665,-0.0453797057,-0.5877487063,0.1424937695,-0.199372679,-0.3385683894,0.0265437644,-0.1084180176,0.1729334891,0.4824619889,0.1901157349,0.2567997873,0.3360639811,0.1500342488,-0.0127167609,0.3075162768,0.4955806434,-0.3149344623,0.1551859528,0.0297931768,0.2141148001,-0.191116184,0.0349998996,0.0768583938,-0.0452108681,-0.0057108714,-0.3479850888,-0.1982782036,-0.3737335205,0.2166278362,-0.0152999908,0.2827970684,0.3574751616,0.2437443882,-0.0598132536,0.0877166614,-0.0267882179,0.0855963677,0.2774118483,0.2729418874,0.3718234599,-0.448490262,-0.0329684317,-0.0818323791,-0.1895794272,-0.2662759721,-0.1657214016,-0.1383874118,-0.1347093135,0.5105390549,0.1755458415,0.1854097843,-0.198313266,0.2382971197,0.1368567646,-0.1418376565,-0.0154677853,0.091069527,0.0367255658,0.6409508586,-0.3012275398,-0.4983736575,-0.0441618301,0.4551079869,0.2364105284,-0.0826209933,0.037869487,0.2227643132,-0.2593609095,-0.1275498718,0.1750154197,-0.1204332858,-0.1331339926,0.1422906965,-0.0169949178,-0.2429295927,0.2397280484,-0.1239747107,-0.1485935599,-0.0695606917,-0.1981938332,-0.274543047,-0.2284975052,-0.1922996342,-0.3533962071,-0.2301900685,0.2255146801,0.0270397924,0.1312210262,0.4111307859,-0.1242666617,0.0634136871,0.1151673123,0.1511102617,-0.047739584,-0.2192236483,0.0848914459,-0.0454674102,-0.0886407569,0.1240959615,-0.0863494426,-0.1349318326,-0.0359486639,0.1907496452,-0.0293504894,-0.2500733137,0.363828212,-0.1229363456,-0.2473060042,-0.4124507606,-0.0434984267,0.0781401396,-0.0084611271,-0.1043872386,0.2735491991,-0.3038494289,-0.2333781868,0.4588110447,-0.1831393242,-0.0493289754,-0.0164843723,0.2902966142,0.4413701296,-0.0644909963,-0.1011531577,0.0276401676,0.2080297172,0.0845939219,0.1611505896,-0.1326494664,0.1946250349,0.2788701355,0.3437087834,0.163782075,0.0735671967,-0.2542277575,-0.0023204465,-0.0041941009,-0.2574895918,-0.208783567,0.5838805437,0.09643583,0.1197781563,0.041137062,0.3612366319,0.1861500442,-0.0093719168,-0.0711825341,0.0869954601,-0.3279539049,0.0698115528,0.1804551929,0.2434641868,0.2993232608,0.427968502,0.0917912573,0.0042312355,0.4806732833,0.2743299007,0.25569731,0.5029736161,0.2437455058,-0.0248373449,-0.3844453394,0.2235188186,0.4877788723,-0.3467884064,-0.0386704877,-0.035630241,0.239819482,-0.2759623528,-0.3680637181,-0.311170578,0.3639385402,-0.1686622202,-0.2518333793,0.250821054,-0.0536625907,0.011008732,0.188388586,-0.2071026266,0.1562429667,0.8072531819,0.1221691146,0.0070223706,-0.3940955698,-0.3178661466,-0.1933319271,0.1779829562,-0.1600327641,0.1832060814,-0.3481118381,-0.0136615047,-0.0037103198,0.3205560744,0.3718766272,0.2885127068,-0.1304165274,0.0477966666,-0.0299581122,0.1783384234,-0.0260137469,0.2189337909,0.0283291712,0.177646488,0.1494833678,-0.1087211147,-0.0802764669,-0.3959601223,0.0510725789,0.3922521174,-0.1891978234,0.1715061218,-0.3878603578,-0.1462846398,0.0034628271,0.2993427813,-0.2532048523,-0.1017914414,0.5167986751,-0.4817173779,0.088193208,-0.1758198291,0.0260638446,-0.0718232766,0.5631066561,0.2815414667,-0.0157232843,-0.3569284678,-0.0953782573,-0.619818151,-0.0195465144,-0.3267097473,0.3234165013,0.0649846569,0.309630096,0.0509843901,0.1159684658,0.3893101215,-0.1246113405,-0.0964443982,0.3271763325,-0.3246978819,0.4432681799,-0.2233607918,-0.1983085573,-0.072717391,-0.4446120262,0.3880726695,0.0543565713,-0.1546519548,-0.1481702924,-0.0597661994,0.1554165483,0.1724309176,0.1892540455,0.350300163,0.3331796825,0.038078215,-0.0212678649,-0.1654650122,0.1553388387,-0.0420160294,0.3519107997,-0.0852085501,0.0682388991,-0.2896396816,0.1777038276,0.0935718715,-0.1146294549,0.0163655914,-0.2013131827,-0.3573488891,-0.0273927134,-0.0389281921,0.3227158487,0.0799018517,0.3904703856,0.07981278,-0.0845014006,-0.1620467007,-0.1161692664,-0.0229110569,-0.3266756237,-0.2992129922,-0.4464145601,0.2472543418,-0.2571519911,0.1094821021,-0.2349529117,-0.3330851197,0.302557528,0.0018811249,-0.3634395301,0.3403205276,-0.426315397,-0.0976772308,-0.1567452848,0.0731856525,-0.0885370225,-0.4450535774,0.2454167157,-0.1163510084]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1992","title":"`datasets.map` multi processing much slower than single processing ","comments":"Yes this is an official bug. On my side I haven't managed to reproduce it but @theo-m has. We'll investigate this !","body":"Hi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n","comment_length":22,"text":"`datasets.map` multi processing much slower than single processing \nHi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n\nYes this is an official bug. On my side I haven't managed to reproduce it but @theo-m has. We'll investigate this !","embeddings":[-0.4096490145,-0.304094851,-0.0847013071,0.3398919106,-0.089700602,0.0127338329,0.3527610004,0.1181460395,0.0582124367,0.0094833281,0.0727031678,0.4454630613,0.1630734205,0.1893298775,-0.1400846541,0.0302210581,0.2145635486,-0.0554908104,0.1930122077,-0.0090904068,-0.162576586,0.1556021571,-0.502361834,0.0061127511,-0.4096408486,-0.1224155501,0.1134260371,0.1184026748,-0.1367546767,-0.2299796939,-0.2356911898,0.1276423633,-0.0973548442,0.5718792081,-0.0001234951,-0.1887065619,0.039246738,0.275313586,0.0935694575,0.0743591338,-0.1399452388,-0.267549932,-0.0483155064,-0.0967941359,0.1221139655,0.2130559534,-0.0675207898,-0.4581288397,-0.0038366562,0.0869868025,0.0545999445,0.3697355986,-0.3081156015,0.0285564456,-0.3477623165,0.1482301354,-0.165335685,0.0483662859,0.2727541924,-0.1806854904,-0.1124003753,0.3173137307,-0.0748661608,0.3036655784,-0.0166217722,-0.0566674583,0.0528347939,-0.4617325664,0.1841140687,0.3488671184,0.004646197,-0.0733980536,-0.1476436257,-0.2400643229,-0.3316350877,-0.1445308179,0.2060097307,0.1393589079,0.1707959175,-0.0030378841,-0.5398437977,0.1457110643,0.356788367,-0.0727699474,-0.1326304674,0.0157895591,0.121410206,0.3247838914,0.3295474052,0.1282672286,-0.0393188782,-0.1861974001,0.3062698543,0.2130707502,-0.7936059237,-0.0484880321,0.2307216674,-0.1161449552,-0.0666975006,-0.169345215,-0.1044374928,0.3401468992,-0.3380639255,0.10364829,0.2965787351,-0.0940639079,0.0115040448,0.1055324078,-0.0219706185,-0.1720061302,-0.3138705194,0.0885760486,0.1108507439,-0.3309875727,-0.0520168133,0.126475811,-0.292132467,-0.0832153484,-0.1314231455,0.0433552377,-0.2141340226,-0.065792352,0.1219370291,0.0876592174,-0.0205289386,0.8091928959,-0.2801376283,0.1047492623,-0.4791386724,-0.5906727314,-0.018312715,-0.1310725808,-0.3497509658,0.215964213,0.1177537814,-0.0320881195,0.1305220425,0.3339615166,0.0465606414,-0.2266206741,0.3357838392,-0.5207974315,0.2082923353,0.1060209498,0.1533152759,0.5560808182,-0.1143068969,0.3106354773,-0.1215611845,0.264598012,-0.5323991179,-0.2379590422,0.1448973715,-0.0522088818,0.1052640155,0.1218633652,-0.4693053365,0.4962759018,0.3239688873,-0.241521284,-0.1572944224,-0.1935511678,-0.6807100773,-0.1566354334,-0.0482701249,0.1644489467,-0.3335207105,0.2270424664,-0.3317820728,0.1459442377,0.446377635,0.5977518559,-0.1945764571,0.3576439321,-0.0237722658,0.1998110414,-0.0121902153,-0.0591245107,-0.3943870068,0.521302402,-0.1812318712,-0.006930348,-0.2187439203,0.1681263298,0.2308537513,-0.0585272126,0.3284294307,0.274037838,0.0182395671,0.4006508589,-0.2312323153,-0.1404776722,0.1745206714,0.0220593773,-0.0664573088,-0.0973705351,-0.0106211863,-0.2823180556,0.3429299593,0.0160740335,0.0190716367,0.4239673316,-0.3839474022,-0.1766140908,-0.1408476382,-0.2507996261,-0.0771214515,0.3548492789,-0.0162000526,0.0330526866,0.436034292,0.0295186676,0.1947910041,0.1215821877,-0.0341638736,-0.0828239173,-0.1245478168,-0.0914572999,-0.1777034402,-0.1682785898,-0.0154164759,0.4407097995,0.2054557502,-0.1136585698,-0.044865191,-0.1442036033,0.0908382908,0.0958688334,-0.2234875858,-0.096726127,0.0980231315,0.0393655561,-0.0858789384,0.2614976168,0.4735931158,-0.0647708029,0.1389831007,0.088604033,0.3819343448,-0.051806163,0.0125038223,-0.1227142587,0.1179005653,-0.2445401996,-0.0177454352,0.3030709624,0.1027351022,0.4237642288,0.0093257474,-0.1245474666,0.0180587508,0.2912706733,0.1467743367,0.0439320877,0.3255986273,0.3314091265,0.2144910395,0.3231862187,-0.1254148334,0.343501389,0.636177063,0.0632761568,-0.3252539635,0.0318580419,-0.178333059,-0.3582537174,0.0271824468,-0.1357810199,0.5621811152,0.0008782812,0.2314520627,-0.0275592264,-0.1016284302,-0.0770715773,-0.0270144995,0.1100118309,0.2270747125,-0.1055658236,0.3235618174,-0.0128858238,-0.045421686,-0.2321382016,0.3384540975,0.225801006,-0.2655066848,0.1165241599,-0.3123525381,0.192536667,0.1699938029,-0.1106924117,-0.2881991267,-0.2154274285,-0.1364985555,-0.0801257566,0.1979137212,-0.0297544003,0.2590248585,-0.1393475384,-0.1104804501,-0.0400148444,0.0211554971,-0.1237185448,-0.1452609748,-0.0168919303,0.2741608322,0.2089086026,0.0379456952,-0.0567800179,-0.2843343019,-0.0929552242,-0.2236779183,-0.1036633924,0.1599895358,0.0655342788,-0.1882638186,-0.0869254544,-0.2204577029,0.1812093854,0.1581970602,-0.265048027,-0.1830227226,0.0601653159,-0.119067207,-0.2122448236,0.0102412989,-0.1542371958,-0.0883423015,-0.0925781727,0.2850442529,-0.1693919003,0.3432775438,-0.214786306,0.0628611073,-0.0864122286,-0.1067367718,-0.0121780084,-0.3401830792,-0.3741228878,-0.0052905926,-0.0206397008,-0.211735338,-0.143867299,0.0744261071,0.153165862,0.147621721,-0.1737326086,0.1924836785,-0.3921591043,0.1361328363,0.0613372996,0.0266093183,0.4794920385,0.0184000414,0.0740958676,0.0087161502,-0.4381608069,-0.0841046795,0.2906069756,0.0134121599,0.0183809716,0.3479084074,0.0413982533,0.6347267628,0.4861660004,-0.1431156844,0.192154035,-0.0242669154,-0.1544582248,-0.3937441111,-0.2161056697,0.1225521564,-0.262909323,0.1614927202,0.4107545614,-0.0043057846,-0.4227926731,-0.0008046668,0.2321252078,-0.2147948891,-0.0154874008,0.1425653845,-0.1290990263,0.1569970995,0.2508654892,-0.1738856286,-0.347938776,-0.0462884828,-0.0596801005,-0.1070540473,-0.1073276028,-0.2466079295,-0.4860386252,-0.0262889713,-0.3734413683,0.2421615422,0.0494488329,0.3200060129,-0.0182066597,0.11256928,0.319367826,0.0135952951,0.5667484999,-0.3824685812,-0.1051650569,0.1393255889,-0.3690501451,-0.2961438,0.0472287275,-0.1584999412,0.2740091085,0.5899748802,0.5148357153,-0.0092025241,-0.1744453609,-0.0026540973,0.1248297319,0.1343102902,-0.2558061779,-0.3702329993,0.1224808618,-0.1186213791,0.1311181933,-0.0073254928,0.1288424432,0.1130923182,-0.1465186775,0.0253687687,0.0986628681,0.2052632123,0.1549219489,0.1151603311,0.0833990052,0.310826093,0.232906118,-0.0853871182,0.029309025,0.3309386373,-0.1140495688,-0.0981434956,0.2457464188,0.0296434313,0.3158788383,0.3062427938,0.1195377111,0.050381951,0.1765029281,0.2736959159,-0.2165954709,0.2651928663,0.4047731161,0.3798927963,-0.4751810431,-0.4996895194,-0.0021378079,0.4812622368,-0.1455397606,0.2885548472,-0.8008277416,0.094181627,0.0404596888,0.0750565827,0.75130862,-0.4283931851,0.0547336005,-0.2906386554,0.1864209324,-0.0401577763,-0.5946896076,0.1480188519,-0.193869248,-0.359297365,0.0260499325,-0.1081058979,0.1894506365,0.4942734838,0.1704988927,0.264792949,0.3171112537,0.1156066507,-0.0115674017,0.3212321699,0.5124132633,-0.3162496984,0.1704147309,0.0245614052,0.2073138207,-0.2166738659,0.0332603604,0.0919925123,-0.039499294,0.0051118312,-0.3249228597,-0.1912873685,-0.3838329613,0.2179381847,-0.0508306511,0.292031467,0.3759081662,0.2679609358,-0.0715066418,0.0968253464,0.0025584628,0.0619813986,0.2923753262,0.2626044154,0.3842230141,-0.4712410569,-0.027831547,-0.1048299819,-0.240194127,-0.2531824112,-0.1586779356,-0.1203282028,-0.1395844966,0.4965134859,0.1751244515,0.1903962493,-0.208500132,0.2218897939,0.1242451891,-0.1349464059,-0.0184657983,0.0999105647,0.0267607495,0.6535869241,-0.3103854954,-0.5013615489,-0.0315314792,0.4279741347,0.2679144442,-0.0882951245,0.031192882,0.2301479429,-0.2587218583,-0.1286641359,0.1715878844,-0.1222177073,-0.1143973991,0.1223205328,-0.0090480233,-0.2285378128,0.2377553731,-0.1473871917,-0.1642845273,-0.0562658869,-0.1955032945,-0.2799701393,-0.2492468208,-0.1993587315,-0.3673148155,-0.2489990592,0.2481578887,0.0355371237,0.1180194318,0.4483625293,-0.1203230768,0.0711580366,0.1135157049,0.1802235842,-0.0537471399,-0.2146884501,0.0826701745,-0.0330266766,-0.0839465186,0.1185893938,-0.0826951638,-0.1370535493,-0.0379954204,0.1925502121,-0.0232914183,-0.2450541109,0.341016382,-0.1211563945,-0.2160396725,-0.4086188078,-0.0402072072,0.0703652278,0.0017752836,-0.1067223549,0.2822424173,-0.3028358519,-0.2639414668,0.4480737746,-0.1902100593,-0.0775017291,-0.015557901,0.2981882989,0.4519784153,-0.0809328556,-0.0976521373,0.0195760019,0.2365054339,0.0848745331,0.1398590505,-0.1199764311,0.1827155501,0.2626988888,0.3164890707,0.1239284724,0.0985502526,-0.2516510487,0.0024659927,-0.0108009176,-0.2553520799,-0.2139285505,0.5657914877,0.0606568977,0.1268813759,0.0375493653,0.3244931698,0.2057412714,-0.0158220865,-0.0779164806,0.0725623667,-0.311545819,0.0609437749,0.1631505191,0.2488363683,0.2783691287,0.4251829684,0.1101088747,0.0030399133,0.4774037302,0.2828223407,0.2613268495,0.5220367908,0.2338133156,-0.0255850218,-0.3653490543,0.200115487,0.506908834,-0.3683279157,-0.0568224974,-0.0439149924,0.2323500514,-0.2388916761,-0.3795402944,-0.263964653,0.3889547884,-0.1608819216,-0.2568490207,0.243508935,-0.0473555736,0.0063971933,0.1846168339,-0.2035260201,0.1530353874,0.8075755239,0.1228650361,-0.0009974042,-0.3829832375,-0.3443283737,-0.1443084031,0.1634515375,-0.1493178606,0.1704648733,-0.3239407539,-0.0275484286,-0.0192410015,0.2792573273,0.3590279818,0.2796686292,-0.1201950982,0.0742152929,-0.009749583,0.1716918349,0.0075825788,0.1788501889,0.0395966619,0.1739648879,0.1511008739,-0.1174862012,-0.0779490098,-0.4345599115,0.065447472,0.3959390521,-0.1649538279,0.1745495945,-0.3532505333,-0.1534028798,-0.0058769612,0.2983115017,-0.2523602545,-0.0957887918,0.5177614093,-0.4819993675,0.0849937648,-0.1764226556,0.0201749764,-0.0753381401,0.5530558228,0.2912755311,-0.0139784198,-0.3776095212,-0.1037469879,-0.5993496776,-0.0270024035,-0.3273997307,0.2998808324,0.0710294098,0.2913455367,0.0507971235,0.1112925857,0.4081784189,-0.0945438668,-0.0640786663,0.3450731933,-0.3213582933,0.4485043883,-0.2388506234,-0.192446366,-0.0728886276,-0.4286518693,0.393148005,0.0691620111,-0.150486812,-0.1682016999,-0.0282133166,0.1636442393,0.1525285095,0.1952544749,0.3520746529,0.3359541297,0.0250959937,-0.0006472326,-0.121674709,0.2023510337,-0.0488462374,0.3206845522,-0.0580119304,0.0556001663,-0.2994981706,0.1774427295,0.1082615852,-0.1121884584,0.0085697537,-0.2228364795,-0.3369599581,-0.0309144873,-0.0465862639,0.3373244107,0.068680495,0.3729414642,0.1067379192,-0.0717504695,-0.1552366465,-0.106936723,-0.0228319149,-0.2881878614,-0.3033364713,-0.451374203,0.248239398,-0.26328215,0.1336895376,-0.2243403345,-0.3340465724,0.3018876016,-0.002103739,-0.3622487783,0.3494460881,-0.4049951136,-0.0751620829,-0.1555635035,0.0664815679,-0.074797079,-0.4459969103,0.2267221957,-0.1301652789]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1992","title":"`datasets.map` multi processing much slower than single processing ","comments":"Thank you for the reply! I would be happy to follow the discussions related to the issue.\r\nIf you do not mind, could you also give a little more explanation on my p.s.2? I am having a hard time figuring out why the single processing `map` uses all of my cores.\r\n@lhoestq @theo-m ","body":"Hi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n","comment_length":53,"text":"`datasets.map` multi processing much slower than single processing \nHi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n\nThank you for the reply! I would be happy to follow the discussions related to the issue.\r\nIf you do not mind, could you also give a little more explanation on my p.s.2? I am having a hard time figuring out why the single processing `map` uses all of my cores.\r\n@lhoestq @theo-m ","embeddings":[-0.3987980485,-0.3398563266,-0.0919537544,0.347813189,-0.1069308296,0.0302635599,0.3706741631,0.1026252285,0.0623089932,0.0149745839,0.0955683365,0.4472282231,0.1740685552,0.1985437125,-0.1503281742,0.0324672796,0.2079118043,-0.0139351441,0.2080077231,-0.0224978235,-0.1496804655,0.1676014662,-0.4635993242,-0.0011216999,-0.427606374,-0.1505295932,0.1084919721,0.1583688259,-0.1540078521,-0.2310477346,-0.2729551196,0.1561855227,-0.0772200376,0.5987077355,-0.0001209586,-0.1784320623,-0.0154779339,0.2657155097,0.0826333761,0.0226127245,-0.1489183158,-0.2439104021,-0.0508770943,-0.1065188497,0.1241560355,0.2492055148,-0.0815335363,-0.4744797647,0.0348683111,0.0918458477,0.0700884014,0.3698405623,-0.3573285043,0.0058346856,-0.4008417726,0.1153549477,-0.1398288906,0.0462569483,0.30848822,-0.1706033796,-0.1413282603,0.3538624644,-0.051077567,0.3019242287,-0.0005348569,-0.0917705521,0.0509418696,-0.4716900885,0.2142590433,0.3441271782,0.0050044786,-0.0449159406,-0.1534067094,-0.2339178026,-0.3566298485,-0.1425357014,0.1976323873,0.119156152,0.1468938589,0.0236070249,-0.5594055653,0.1424560845,0.3238302171,-0.0603595935,-0.1207488403,0.0340404846,0.102419354,0.3249111772,0.3525274396,0.1588870287,-0.0753305927,-0.221325025,0.3150898218,0.1927129924,-0.7903565168,-0.0546728112,0.2171829492,-0.0919128954,-0.0570161603,-0.1511578262,-0.1151762232,0.3496940732,-0.347454071,0.0873438865,0.317974478,-0.0674332157,0.0333352499,0.0778571069,-0.0398387872,-0.1619227827,-0.2981231809,0.0592453554,0.0821144208,-0.3091109395,-0.0677434057,0.103674151,-0.3128350973,-0.0841979161,-0.1708450466,0.0062154941,-0.2175653577,-0.0867868215,0.1496063024,0.0932789668,-0.0010675321,0.840298295,-0.2644474208,0.093375124,-0.443873316,-0.6254184246,-0.0214878377,-0.0977298319,-0.378390491,0.2289134115,0.1202447489,-0.0522700511,0.1299927533,0.3237181902,0.0419285074,-0.1985152066,0.3197094798,-0.5415113568,0.1895833313,0.0605318025,0.1435223669,0.5663229823,-0.0963238552,0.2923194766,-0.1278509498,0.2347625792,-0.5155062675,-0.2475563288,0.1758901179,-0.0223864168,0.1043807268,0.1009875908,-0.453175813,0.5067190528,0.292876035,-0.1959371716,-0.1508533359,-0.1562158316,-0.6555333734,-0.1608223021,-0.0710844174,0.1317817569,-0.3583333194,0.2057326734,-0.3523509502,0.1483931988,0.4070552886,0.6043972373,-0.2138480395,0.3782386184,0.0245644338,0.2347583622,0.0038246829,-0.0866079181,-0.3828225732,0.5263620615,-0.2283221483,-0.0003068696,-0.1912580281,0.175038293,0.2384458333,-0.0272581726,0.3657393157,0.2632677257,0.009416352,0.3835636377,-0.2240213305,-0.1533070654,0.1631630957,0.0522787794,-0.0803728849,-0.1023805812,-0.0463994704,-0.2913084328,0.3220100403,-0.0063353805,0.0290314797,0.4283193052,-0.3774930537,-0.190831989,-0.1373764575,-0.2449973822,-0.0953904018,0.3824571371,0.0239032898,0.0706383735,0.4634763598,0.0386811346,0.1495977044,0.1042516008,-0.0175235029,-0.0793729797,-0.0917160138,-0.1016785428,-0.180559516,-0.2226997018,-0.0659697205,0.4419493973,0.2227581143,-0.098893337,-0.0392994136,-0.1586316973,0.0430578552,0.0839976668,-0.2125139087,-0.0789049044,0.0700058118,0.0641730428,-0.0729280934,0.2602718771,0.4253133833,-0.0739465728,0.1746495217,0.1128149182,0.3552783132,-0.0458750539,0.0203691777,-0.1413392276,0.1449813843,-0.2591532469,-0.052130729,0.2841012478,0.083830066,0.4506697059,0.0235608686,-0.1425478011,-0.0140936971,0.2643783092,0.1708122194,0.0492692925,0.3555659354,0.3383421302,0.2349534631,0.3360346854,-0.1631879658,0.3445366323,0.612660408,0.0417517982,-0.3295563757,0.0292304177,-0.2006421387,-0.3668963015,0.0839066878,-0.1144676805,0.5725329518,0.0204459317,0.2464558631,-0.0426016785,-0.0617185682,-0.0788922757,-0.0321179926,0.1035840958,0.2068733573,-0.1335346848,0.2963628471,-0.0233534034,-0.0464381054,-0.2598059773,0.3200156987,0.2167464942,-0.2419921607,0.1305710524,-0.3196627796,0.1452299505,0.2040527463,-0.1186888888,-0.3031505644,-0.2074022591,-0.1162160113,-0.0619461574,0.2765515745,-0.0591169856,0.2854481041,-0.092461966,-0.1519616544,-0.0172398984,-0.0202096663,-0.1655507982,-0.1726305932,0.0085463403,0.2644037008,0.2206041515,0.0475902334,-0.0641812161,-0.2819513679,-0.0524871126,-0.1731323153,-0.1106104553,0.1379324645,0.0642573461,-0.1711055636,0.0039563258,-0.2347945124,0.189412117,0.1470809132,-0.2577964664,-0.163847506,0.0355857089,-0.1214443445,-0.2122675627,0.0128884548,-0.1658165157,-0.0854584798,-0.0955391079,0.3060115576,-0.1661301851,0.323209703,-0.194445312,0.0642021894,-0.0389953032,-0.0717668384,0.0312101003,-0.3452689946,-0.371363312,-0.0210943501,0.0280967597,-0.200767681,-0.1226667911,0.0851430371,0.2075583339,0.1145337299,-0.1510454565,0.1748339534,-0.3728004992,0.1301907301,0.0589384846,0.0784902275,0.4467203617,0.0360531248,0.0617042854,0.0150992125,-0.41154477,-0.0789791197,0.2864608169,0.02886739,0.0285069048,0.3272490203,0.0790639743,0.6817412376,0.4720701575,-0.1880872101,0.1948011518,0.0043178932,-0.1779785007,-0.3762973547,-0.221464932,0.1091218293,-0.2775560617,0.1900647581,0.4069162011,0.0025335229,-0.4038463831,-0.0298229009,0.2467475533,-0.2389826775,-0.0058508129,0.1471872032,-0.0795921236,0.1282855123,0.2544082999,-0.1898119301,-0.3439298868,-0.0338944234,-0.0574088283,-0.0707049221,-0.1244392097,-0.2597436309,-0.5114725828,-0.0330200493,-0.381346494,0.2475262433,0.0228804462,0.2694840431,-0.0378903486,0.1243406534,0.3278180063,-0.006477538,0.5314375758,-0.3730396032,-0.1724523306,0.1234857365,-0.412956208,-0.2700184584,0.1126654372,-0.1809013039,0.2845947742,0.5743815899,0.51442945,-0.0455423743,-0.2244948,-0.0374761596,0.1309106201,0.1329891831,-0.2703753114,-0.3282872736,0.1430770904,-0.1507306397,0.1006364003,-0.0207584724,0.136767298,0.128336966,-0.1023597941,0.0580426492,0.0914529935,0.2539273202,0.1980279833,0.0963080674,0.0602107868,0.258325249,0.265718013,-0.0980602503,0.0382840969,0.3328794241,-0.1548348963,-0.0922503248,0.233023569,0.0603707358,0.3217382431,0.3481735885,0.1211917698,0.0601415969,0.1878466457,0.2428903729,-0.2255191654,0.2873825133,0.3583303988,0.3687764704,-0.4462328553,-0.4971280396,-0.0129834199,0.4985198379,-0.1683432311,0.3029264212,-0.7764919996,0.073723793,0.0215329584,0.074474521,0.7982966304,-0.4352927804,0.0622138493,-0.2951289415,0.2099918723,-0.0567078516,-0.5967226624,0.1344829351,-0.1911982149,-0.378200233,0.0069172028,-0.1047018319,0.1857600957,0.5643200278,0.1523874849,0.2857613266,0.3366233408,0.0805623904,-0.0095910178,0.3668844104,0.5028193593,-0.3357380331,0.1540343314,0.0463160016,0.1937491149,-0.1932332069,0.0020395243,0.0996448621,-0.0719177499,-0.0027787115,-0.3143114746,-0.2183893025,-0.4510262311,0.2037439197,-0.0680122599,0.3175693154,0.3931640089,0.2769828141,-0.0670501068,0.0411295965,0.0065714517,0.0356714614,0.3010514677,0.2570824623,0.3631042838,-0.4733554125,-0.0057051787,-0.1186290011,-0.245458439,-0.2174409479,-0.1700531989,-0.0586603992,-0.1510441899,0.487614125,0.1578538418,0.1609752625,-0.1905021369,0.2473519146,0.1159876511,-0.1148685068,-0.0019466956,0.1373942494,0.0247041583,0.6386431456,-0.2702286839,-0.4856572449,-0.0345760398,0.3869275451,0.2485567331,-0.1072566584,0.0116949705,0.2025443614,-0.2665699422,-0.1496108025,0.1637888253,-0.0857774019,-0.0729986653,0.127365768,0.0348298177,-0.2494173795,0.2480981946,-0.1001478806,-0.1314290166,-0.0336117595,-0.2042182386,-0.2643249631,-0.2544718087,-0.1720092148,-0.3750605881,-0.207009092,0.2421014458,-0.0172371399,0.1284514517,0.4695535302,-0.1497863233,0.0607125089,0.1021557823,0.1906091124,-0.0384997241,-0.2601143122,0.0941978991,-0.0170744099,-0.0643610582,0.0783964917,-0.0853478238,-0.1668204963,-0.0115808947,0.1847727448,-0.0130351735,-0.2575166821,0.3359077573,-0.1025240868,-0.1739261001,-0.3928409517,-0.0501778573,0.0468402207,0.0097020362,-0.1023318321,0.3172993958,-0.2931720614,-0.3031933308,0.406866163,-0.2091906965,-0.0665455014,-0.0005234958,0.269798398,0.4595621228,-0.0918447077,-0.1277807057,0.0299210679,0.2942765355,0.0688005984,0.1120858192,-0.1015246734,0.1829239875,0.269906342,0.2781372964,0.1202000529,0.1141079515,-0.2037565857,0.0053946991,0.0241427422,-0.2130512744,-0.1793759316,0.5706425905,0.0138295852,0.1518812627,0.0689559951,0.3217915595,0.2287770212,-0.0012451352,-0.1071082652,0.0849716887,-0.3123015165,0.0684472471,0.1110094413,0.2458712459,0.2573178411,0.436222285,0.1256205291,0.0101210009,0.4173806906,0.242287457,0.3094064891,0.5200669169,0.2401477247,0.0135991732,-0.3639056981,0.1917427331,0.490093559,-0.3441900313,-0.066210717,-0.0487065613,0.2394540757,-0.2684269845,-0.3709075153,-0.2196556032,0.3906934261,-0.1724593788,-0.2432681173,0.2041936815,-0.038143836,-0.045964431,0.1604242921,-0.2366207093,0.1313254237,0.8348470926,0.1233759969,0.0309466049,-0.3607608974,-0.3917119205,-0.1342182308,0.148079291,-0.1345745772,0.190297097,-0.3134203255,-0.0068785045,-0.0136601087,0.2765809596,0.3565859795,0.2774432003,-0.1138226688,0.0915465727,-0.0268786717,0.1539753228,-0.0184285715,0.1534866244,0.0066649434,0.1805169582,0.1535555571,-0.0864600688,-0.0981989726,-0.4284718633,0.0447323434,0.3398983479,-0.1634690911,0.1326927245,-0.3101339638,-0.1590601504,-0.0120583763,0.2858177423,-0.249538675,-0.0862631351,0.4984052479,-0.4902805686,0.072650522,-0.1841207743,0.0356182642,-0.0405586399,0.5247302055,0.2865372598,-0.0129882563,-0.4199312329,-0.1138789505,-0.5841233134,-0.0655887797,-0.3243660033,0.2615009546,0.0450748019,0.2684961259,0.0611372329,0.1447686553,0.4047620595,-0.1011889428,-0.0866829231,0.3594867885,-0.3306727707,0.5003919601,-0.2142937779,-0.1856480986,-0.0553078875,-0.4004398882,0.3668005466,0.096898824,-0.1297347248,-0.1715724319,-0.0162613373,0.1787750572,0.1407820582,0.1806530803,0.3315006793,0.326703459,0.0269237794,-0.0089413021,-0.1012832299,0.2193169445,-0.0792323872,0.3145402968,-0.0376671441,0.0590494871,-0.3016564846,0.1776498854,0.1021894738,-0.1452145129,0.0031901,-0.2515270412,-0.3215590417,-0.0398885198,-0.0551042743,0.3557627797,0.0897440985,0.4125417769,0.1387647986,-0.0893788263,-0.2004950047,-0.0998046845,-0.0184925124,-0.2824784517,-0.3231762648,-0.4460316598,0.2157075107,-0.2852642536,0.1410285532,-0.2329756171,-0.3740328848,0.3034420311,-0.022932034,-0.4084106088,0.3051842451,-0.3485661745,-0.0823975727,-0.1502872854,0.0432494767,-0.0781462491,-0.448741585,0.2001753151,-0.1336032301]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1992","title":"`datasets.map` multi processing much slower than single processing ","comments":"Regarding your ps2: It depends what function you pass to `map`.\r\nFor example, fast tokenizers from `transformers` in Rust tokenize texts and parallelize the tokenization over all the cores.","body":"Hi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n","comment_length":29,"text":"`datasets.map` multi processing much slower than single processing \nHi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n\nRegarding your ps2: It depends what function you pass to `map`.\r\nFor example, fast tokenizers from `transformers` in Rust tokenize texts and parallelize the tokenization over all the cores.","embeddings":[-0.4439504147,-0.2767181396,-0.0788298249,0.3751109242,-0.0908027664,-0.0050794897,0.3211759031,0.0643401295,-0.0458027311,-0.0092515377,0.0477654375,0.4086365104,0.2179492712,0.1524443328,-0.1457520574,-0.0019726125,0.2254880369,-0.0213480853,0.1467742175,0.017389413,-0.2218794972,0.1027975529,-0.5076216459,0.0070183063,-0.43289572,-0.1630928069,0.0852133259,0.0920830145,-0.1628730446,-0.2309202701,-0.2925642431,0.2094140053,-0.0740564167,0.5870193243,-0.0001180967,-0.1688386947,-0.0100619318,0.263231039,0.1063561291,0.0013800593,-0.0580675416,-0.2970840931,-0.0959245935,-0.0974916518,0.1089042947,0.1052417606,-0.0769347697,-0.4238997102,0.1259805709,0.0137929441,0.1001286283,0.3818843067,-0.2749175429,-0.034933567,-0.3510386348,0.1526193619,-0.1374157369,-0.0240369998,0.241744712,-0.1240661666,-0.1272667646,0.3543689549,-0.1285182983,0.2587579787,0.0789577588,-0.0478510372,0.0509219579,-0.4571819603,0.1748204231,0.4000486732,0.0125549957,-0.0977234021,-0.1687846333,-0.2309564352,-0.2791661918,-0.1573265046,0.2020871639,0.104083024,0.1339534968,-0.0070664226,-0.6016136408,0.2399581224,0.3542020917,-0.1202202514,-0.1160987765,0.0890455022,0.119870998,0.3441294134,0.3367708027,0.077030234,-0.08753106,-0.2567175329,0.2177376896,0.2279644012,-0.8232127428,-0.0756709501,0.2212411165,-0.1676312685,-0.0521672927,-0.1305063516,-0.1465456486,0.3569334149,-0.3285430372,0.1192994565,0.2981947958,-0.0345787741,0.0140539873,0.1581130326,-0.010236118,-0.2339845896,-0.312699914,0.0593383312,0.0346820764,-0.262565434,-0.0355956405,0.0848609731,-0.3532977104,-0.0575250313,-0.2106844038,0.0534194596,-0.2790293396,-0.0660476759,0.1619876027,0.1405691653,0.0106093036,0.7866417766,-0.3018142581,0.0486737937,-0.3648938239,-0.5528380275,-0.0429502241,-0.1930681467,-0.3997564018,0.2632676959,0.1055118889,0.0493484512,0.1494627297,0.2821126282,0.0765934512,-0.1413563937,0.3494647145,-0.5429766774,0.1496422738,0.0726197287,0.154560104,0.5177941918,-0.138474986,0.314902097,-0.1971454918,0.2203018218,-0.475266695,-0.2318782359,0.2151331753,0.013539426,0.0885788277,0.0796750262,-0.5455653667,0.5509293675,0.2919282317,-0.172665447,-0.1275849491,-0.1853656173,-0.6103130579,-0.1781904399,-0.0018492461,0.1226013973,-0.2417798936,0.1732539535,-0.2710019946,0.1222158447,0.4497342408,0.6742501855,-0.2196296602,0.4246218801,0.0158408955,0.3035623729,0.0449148156,-0.1139074489,-0.3109883964,0.4910986125,-0.1610904932,-0.0346942134,-0.1811563224,0.1404015571,0.3190973103,-0.0290230568,0.2742301822,0.3147366345,0.0265183486,0.3793387413,-0.1575172096,-0.1305783689,0.1962330639,0.0451965034,-0.1150431037,-0.1356102675,-0.0503521562,-0.2750127614,0.224800542,-0.0303576495,0.0487311482,0.3911471069,-0.3012832105,-0.1758214831,-0.1415572315,-0.2277328521,-0.0599743053,0.3407979012,0.0430903807,0.0639524162,0.460814625,0.0015955908,0.2276877761,0.1355304271,-0.0380949266,-0.1376335174,-0.0313482508,-0.0491703339,-0.1346637458,-0.2262031883,-0.0678611919,0.3672987521,0.2238070518,-0.0643264651,0.0883311555,-0.1257754117,0.0670795441,0.0064629288,-0.2424031049,-0.0341825634,0.0654881671,0.1089310721,-0.0454825759,0.2792190015,0.4492981434,-0.0380155668,0.1182730347,0.1675735861,0.3643091619,-0.0421776362,0.0124646351,-0.1279924363,0.0809552521,-0.229295373,-0.0607795864,0.2941325307,0.1038633659,0.4727625847,0.0048596333,-0.176206395,0.0205464344,0.2855567634,0.1121584997,0.1042972803,0.2605012655,0.3831946254,0.2929072082,0.3807497919,-0.1670197695,0.2910594642,0.5920889974,-0.0072922744,-0.3589833379,0.0816680118,-0.2233637869,-0.4092293978,0.0779984668,-0.1123960838,0.5411706567,0.0466136485,0.2133837789,-0.0555458404,-0.1274143606,-0.0524572879,-0.006476284,0.0726008639,0.1973280609,-0.0815938562,0.2709230781,0.0563166067,-0.0644898787,-0.2889431417,0.3361002803,0.2415702641,-0.2043419331,0.0760013685,-0.2641548812,0.112465322,0.1709750891,-0.0607179031,-0.21242553,-0.1904428899,-0.121468313,-0.1239397004,0.1627361178,-0.0067014256,0.2737019658,-0.0764560699,-0.0927326679,-0.0275587533,-0.0323879272,-0.1445223689,-0.1797593385,0.0104280403,0.2918093204,0.2076402009,0.1169967428,-0.0755054876,-0.232487008,-0.0665045828,-0.2315778434,-0.0556138493,0.079541795,0.0135089662,-0.1537777334,-0.0750604197,-0.2716275752,0.0461053438,0.1981592625,-0.3055951595,-0.2410659492,0.0426123813,-0.1434981525,-0.2572675049,0.0451567695,-0.1229170039,-0.0881868526,-0.1083155796,0.3268550038,-0.1563324779,0.2969707847,-0.2059313208,0.0423742197,-0.0350836627,-0.1536312848,0.0389460288,-0.3043987155,-0.3341442049,0.0511069298,-0.0429351628,-0.1612553298,-0.131100744,0.1183316559,0.1708219945,0.1495732218,-0.1346123219,0.1372227222,-0.3764241636,0.1445037723,0.0494961403,0.1159208119,0.4347307682,0.0688747242,0.0020963789,0.0667256936,-0.419267118,-0.0635158271,0.2809041142,0.0510030612,0.0156693384,0.3901453316,0.0600493215,0.6509938836,0.506364882,-0.1862688661,0.1578833163,0.0148957791,-0.1145068929,-0.4058781862,-0.1702889204,0.1095638052,-0.2204565108,0.1181804016,0.4101580679,-0.005214531,-0.4526441693,0.0020453932,0.2613644898,-0.2650698125,-0.0189336967,0.1734432429,-0.1055791155,0.195678398,0.189076528,-0.1457652748,-0.31885764,-0.0576312877,-0.0498399101,-0.1312036812,-0.1142136455,-0.2327936292,-0.5820653439,0.0001046544,-0.3677584827,0.2439447939,0.0063359751,0.2492839992,0.0271377861,0.0890652835,0.327380538,-0.0116161527,0.4494200349,-0.3170929253,-0.191743806,0.1078528538,-0.4520253837,-0.2145804912,0.0766330212,-0.1998126209,0.2188529372,0.6070114374,0.5290973186,-0.1593936384,-0.2424485534,-0.0145309446,0.0675070211,0.2063866854,-0.2496495545,-0.2906808555,0.1218738183,-0.1460334659,0.1630269885,-0.0644271374,0.1508836597,0.0983762592,-0.1814582348,0.036500819,0.0545814112,0.2548568249,0.2476999015,0.1457834393,0.0650622323,0.310420692,0.3232932091,-0.1338669956,0.1037744731,0.2178996801,-0.0925103724,-0.113243416,0.1909930259,0.0520568714,0.3045853376,0.3385318518,0.0856506824,0.0982141271,0.0943279639,0.3066310585,-0.2347388715,0.3093484044,0.4002850354,0.4037322998,-0.466727972,-0.5559285283,-0.0690212995,0.4330460727,-0.1820667684,0.2211446911,-0.7285043001,0.0773606747,0.1097152531,0.1712530255,0.8330796957,-0.4548249841,0.0848217979,-0.3236432374,0.253736645,0.0580802523,-0.7181652784,0.096222952,-0.2373639643,-0.2546758056,0.024620194,-0.1100668833,0.1980310977,0.5148921013,0.115482226,0.2528627217,0.424382031,0.2372803241,-0.0995558277,0.3643096983,0.4965112805,-0.4183228612,0.1333128512,0.0802545473,0.1976226568,-0.238912344,0.0270951204,0.0834894031,-0.079570435,0.0184416249,-0.3758701682,-0.2915439308,-0.4238469005,0.2038731724,-0.0037291031,0.2687775493,0.3288625479,0.279758811,-0.0635965616,0.1441981643,-0.0059177135,0.0863516927,0.2842148244,0.2989465296,0.3509505689,-0.5309547186,-0.067119725,-0.0798979402,-0.2635248005,-0.1854159981,-0.1425006837,-0.1600122601,-0.1232456043,0.3970688283,0.1707141846,0.1685151607,-0.171010226,0.2841367126,0.0973168761,-0.1147535294,0.0305125378,0.1024273857,-0.0632519051,0.6600760818,-0.2263176441,-0.4700722992,-0.0673076063,0.3055168688,0.2026191205,-0.1621508002,0.0514472835,0.1235845089,-0.2654483318,-0.1657324284,0.2679197788,-0.0332038328,-0.0632684529,0.0777310506,-0.0185653046,-0.2477255464,0.2914330661,0.0157897938,-0.0839612633,-0.0706937686,-0.1701770276,-0.2372152656,-0.2801140547,-0.146753028,-0.2820578516,-0.1556294411,0.2202247828,0.0711505413,0.0766150802,0.4940489829,-0.1891953796,-0.0150634712,-0.0130067831,0.2362150103,-0.051591374,-0.2214390934,0.0374448411,-0.0237098653,-0.0277682208,0.0990924761,-0.1186020821,-0.1766845882,-0.0388225839,0.1717950106,-0.0014159676,-0.3216920793,0.2939102352,-0.1387358904,-0.2362678945,-0.4579728544,-0.0253314078,0.0440421663,-0.0457148403,-0.1652331501,0.3154085279,-0.2161611915,-0.2762578428,0.4109307826,-0.1830657423,-0.078049615,0.0185394101,0.2842990458,0.4484986365,-0.0958242938,-0.0226852093,0.0760356635,0.3730266988,0.0505047403,0.1415210515,-0.0844870955,0.159935087,0.2427989691,0.3568882048,0.1891925931,0.1293494701,-0.2801517248,0.0930409059,0.0573855229,-0.1989368498,-0.1231892779,0.4903810024,0.0182056595,0.1448251754,0.0927430168,0.3649458289,0.3037870526,0.0161980912,-0.1273450702,0.1338044852,-0.2889988124,0.0724494159,0.0860034376,0.1859660596,0.2598794699,0.4262659252,0.0697908327,0.0363907143,0.4503248334,0.2709546089,0.3376744092,0.4775066674,0.3344381154,-0.0255360156,-0.3478305936,0.2448361069,0.4501988292,-0.3315539956,-0.1329411268,-0.0553730205,0.193781212,-0.2377573401,-0.3416785002,-0.2749478221,0.4440901875,-0.2080302387,-0.3045457006,0.178730756,-0.0995024517,-0.0396524891,0.1984784156,-0.2173382342,0.1424004585,0.7284575105,0.073427394,0.0488693267,-0.390539825,-0.2922240198,-0.1570208222,0.1194631383,-0.1403240263,0.2661330104,-0.2927122712,0.0441741608,0.0352609567,0.3550224602,0.3926257491,0.330347985,-0.1617084146,0.0841675922,-0.057186015,0.1578370482,-0.0234033559,0.251714915,0.0219425932,0.0856103376,0.1510846317,-0.0631432459,-0.1325363368,-0.5011906624,-0.056989748,0.2782666683,-0.1450001597,0.2023577541,-0.289016813,-0.1213661134,0.0647266209,0.303177774,-0.2443993241,-0.0538231358,0.4503957331,-0.4847294688,0.0268958025,-0.1731549203,0.0600576513,-0.0299803987,0.5599967241,0.279804945,-0.0255501289,-0.4537350535,-0.1623744369,-0.6362349987,-0.0669201985,-0.2945365608,0.2655956149,0.0150774103,0.2924502492,0.0856208056,0.2242643535,0.330614835,0.0055515589,-0.1317665726,0.37577039,-0.2399875522,0.5442321301,-0.2748645246,-0.1883133203,-0.0854738355,-0.4263988733,0.3769555986,0.1100715175,-0.1097065508,-0.2021090388,0.0195115693,0.2056244314,0.1438642591,0.2136922479,0.371399045,0.3717583716,0.0488746949,0.0633167848,-0.1134325862,0.1278325319,-0.0979392156,0.25566715,-0.1105476394,0.0603189096,-0.2797147036,0.1180019379,0.0593304709,-0.1554958522,-0.0002601092,-0.2810496986,-0.3650560379,-0.0507835113,-0.1279818267,0.2740593553,0.0406712256,0.3598727882,0.1757293493,-0.1003123224,-0.2375797331,-0.0797622651,-0.0203341451,-0.3013189435,-0.267773509,-0.5111196637,0.1080258265,-0.171879217,0.1276894808,-0.2887659669,-0.3083531559,0.3282149434,-0.0181480125,-0.4100753665,0.3328624964,-0.3763077855,-0.139082104,-0.1773988008,0.0250442699,-0.1261186302,-0.4447464049,0.147843346,-0.0966654718]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1992","title":"`datasets.map` multi processing much slower than single processing ","comments":"I am still experiencing this issue with datasets 1.9.0..\r\nHas there been a further investigation? \r\n\"image\"\r\n","body":"Hi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n","comment_length":19,"text":"`datasets.map` multi processing much slower than single processing \nHi, thank you for the great library.\r\n\r\nI've been using datasets to pretrain language models, and it often involves datasets as large as ~70G.\r\nMy data preparation step is roughly two steps: `load_dataset` which splits corpora into a table of sentences, and `map` converts a sentence into a list of integers, using a tokenizer.\r\n\r\nI noticed that `map` function with `num_proc=mp.cpu_count() \/\/2` takes more than 20 hours to finish the job where as `num_proc=1` gets the job done in about 5 hours. The machine I used has 40 cores, with 126G of RAM. There were no other jobs when `map` function was running.\r\n\r\nWhat could be the reason? I would be happy to provide information necessary to spot the reason.\r\n\r\np.s. I was experiencing the imbalance issue mentioned in [here](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-705177036) when I was using multi processing.\r\np.s.2 When I run `map` with `num_proc=1`, I see one tqdm bar but all the cores are working. When `num_proc=20`, only 20 cores work. \r\n![Screen Shot 2021-03-05 at 11 04 59](https:\/\/user-images.githubusercontent.com\/29157715\/110056895-ef6cf000-7da2-11eb-8307-6698e9fb1ad4.png)\r\n\nI am still experiencing this issue with datasets 1.9.0..\r\nHas there been a further investigation? \r\n\"image\"\r\n","embeddings":[-0.4408694506,-0.2572331727,-0.1001466885,0.3424948156,-0.1136195362,0.0254862215,0.335966289,0.1372743845,0.0238077715,-0.0266041029,0.0728430375,0.426761657,0.1649507433,0.1844829768,-0.1783974469,0.0735184923,0.1910468787,-0.0237313621,0.1997010559,-0.0175593514,-0.172121197,0.1267192066,-0.4863221943,-0.0251146331,-0.418707788,-0.1551897079,0.1128484309,0.1078351289,-0.1613447219,-0.2424036711,-0.2254241258,0.1812607646,-0.0989795253,0.5712867379,-0.0001213278,-0.1757612377,0.0262497552,0.26557675,0.0566576533,0.0632180348,-0.1749771684,-0.26562953,-0.0603124537,-0.1064622477,0.1302662641,0.1959712207,-0.0412265956,-0.4592748284,0.0279739648,0.0956204832,0.0683356822,0.3693211377,-0.3358564377,0.0190168042,-0.3772294223,0.1757356524,-0.1743134111,0.0439391397,0.2703153491,-0.189501822,-0.1327500939,0.2942250967,-0.0594553053,0.2955110371,-0.0293099452,-0.0800075904,0.0435416587,-0.4581356645,0.2109427005,0.3413070142,0.0336227156,-0.0598060898,-0.2014713734,-0.2505912185,-0.3135108352,-0.0960752293,0.1881505698,0.1586238891,0.143742457,0.000274898,-0.5647354126,0.1250050813,0.3329138756,-0.0845777914,-0.1606686711,0.0009775285,0.1214764491,0.3006643951,0.3285801113,0.1026582047,-0.0240339898,-0.2341752946,0.3186653852,0.2144552469,-0.7959993482,-0.0313890278,0.2321540564,-0.1166578382,-0.0702337921,-0.2011479288,-0.0824341848,0.3807694614,-0.3303593695,0.0957021937,0.3113458455,-0.0816923156,0.0368814319,0.1326556355,-0.0200128127,-0.1649896055,-0.3204186261,0.0855832472,0.0375055708,-0.3145132959,-0.0588686801,0.1090919524,-0.2920226157,-0.0776981115,-0.1529199481,0.0578000583,-0.2044586688,-0.0747551695,0.1519744843,0.0931527093,-0.0251007862,0.7979210615,-0.2888891995,0.0677869245,-0.4647265077,-0.5510899425,-0.0266884044,-0.1435279697,-0.3716950715,0.2059828192,0.1467278749,-0.0441316739,0.1381739825,0.2995373011,0.0675618723,-0.1942556351,0.3298442066,-0.5215236545,0.1908664405,0.089770101,0.1518644542,0.5952267051,-0.1308276206,0.3086303473,-0.114612408,0.2425247729,-0.5119680762,-0.2261331826,0.1801623553,-0.0164731424,0.0661159083,0.108579129,-0.4507105947,0.4928016961,0.310757786,-0.2005254179,-0.1917562038,-0.2191664875,-0.6590129733,-0.1549583077,-0.0378507301,0.1781466007,-0.3581593335,0.2357657999,-0.3264361322,0.1465032846,0.4354635179,0.5881886482,-0.1908946335,0.3742271364,-0.0167533867,0.2168684006,-0.005193877,-0.0612444021,-0.409089148,0.5504702926,-0.1573844999,-0.0077369595,-0.1996926665,0.132253468,0.2520763874,-0.0667857826,0.2841029167,0.2935760915,0.0093998676,0.3959366977,-0.2198728025,-0.1206632107,0.1546422243,0.07583303,-0.052067332,-0.1174090207,-0.0040198578,-0.2566078305,0.3107918203,0.0013408916,0.0485638939,0.4202738404,-0.376527369,-0.181443274,-0.1402570158,-0.2091579288,-0.089415051,0.3687587082,-0.0112051889,0.036914859,0.3978402317,0.025036376,0.1763718128,0.1276012063,-0.0591956191,-0.083436206,-0.0933097526,-0.070499301,-0.1872852594,-0.1743258983,-0.022125937,0.4432820082,0.2146649063,-0.1109524071,-0.0165356025,-0.1274927109,0.0751287788,0.0695862398,-0.186287567,-0.0800083727,0.0683454126,0.044272162,-0.0770368278,0.2573568821,0.457980603,-0.0575390272,0.1573877633,0.0783579201,0.3972081244,-0.0391156785,0.0224031191,-0.143343389,0.1175412908,-0.2493546307,-0.0373166203,0.3127611876,0.1098824888,0.4193719327,0.0704618618,-0.1443873942,0.0071856701,0.3014125526,0.1643483043,0.0150564266,0.3293915689,0.3466155529,0.2566779554,0.3388873041,-0.115452908,0.3463789821,0.6238071918,0.0678065345,-0.3361598849,0.0286232531,-0.1701198816,-0.3801422417,0.0395852625,-0.1131452098,0.5269431472,0.0170898419,0.2577416003,-0.0255265627,-0.1162775308,-0.0751271471,-0.0387657471,0.1148215756,0.2368961126,-0.0769186318,0.3169465065,-0.035439834,-0.0635434389,-0.2647989988,0.3754664958,0.2200479507,-0.2382181734,0.0989622325,-0.3365543187,0.1643039137,0.1856621504,-0.123356171,-0.2650382519,-0.2415278107,-0.1381919086,-0.0913198218,0.2174679488,-0.0295518301,0.2439494133,-0.1340822577,-0.0943095237,-0.0267706104,0.0108898859,-0.1664037853,-0.1559698433,0.0049409857,0.2704434991,0.2264301628,0.0572019555,-0.0552712008,-0.2538318932,-0.0788089633,-0.2205043435,-0.099053055,0.1591480523,0.0593895093,-0.1872386932,-0.0787216723,-0.2335619628,0.1749194413,0.1698775887,-0.2436436862,-0.1829372495,0.0668812394,-0.1159514412,-0.2466948181,0.0244951714,-0.149846226,-0.0929289088,-0.1155197471,0.2580092251,-0.1713645607,0.3255140781,-0.1932542473,0.0785440803,-0.0488123782,-0.1118045449,-0.0171387941,-0.3181794882,-0.3501659036,0.026869826,-0.0320225693,-0.2036227137,-0.1449833363,0.0673169643,0.180342719,0.1953355521,-0.1574309468,0.1874525249,-0.4001614153,0.1639227569,0.0731303617,0.0248603169,0.4335301518,0.0243834369,0.0598816089,-0.0029196024,-0.4167127311,-0.072429046,0.2753514349,0.0008795081,-0.0096099852,0.3409216404,0.0701248199,0.6287553906,0.5038680434,-0.1901980191,0.1679047197,-0.0031842974,-0.1370066702,-0.3670369685,-0.227397427,0.1343926936,-0.2523145974,0.1493886113,0.4214679003,-0.008276443,-0.4179485738,-0.0153305763,0.2546685636,-0.2011983842,-0.0177169852,0.1325407624,-0.1200014353,0.2148343921,0.2546669245,-0.1645083576,-0.3432222307,-0.0456095301,-0.0871300101,-0.1086413488,-0.1062070206,-0.2582450509,-0.4892933667,-0.0094412593,-0.3960962892,0.2227958888,0.0528319962,0.3650021851,-0.0304194391,0.1201309264,0.3037044108,0.0031186431,0.5263445973,-0.4160194099,-0.1148830727,0.1485777348,-0.3820876181,-0.3232316971,0.0533494465,-0.1624045521,0.3027198911,0.5608937144,0.517560482,-0.0368241891,-0.2009236664,-0.0143956477,0.1242436916,0.1565524638,-0.2552374303,-0.3296492398,0.1101543978,-0.1132433712,0.1082992926,-0.018909255,0.1200617477,0.1201654598,-0.1573020071,0.0094523979,0.1104448587,0.2438207418,0.2051901668,0.1092441976,0.0710771009,0.3489991128,0.218920216,-0.107520923,0.0731957629,0.327277869,-0.1145711541,-0.1128835827,0.2355375588,0.0178016108,0.3012571037,0.2909519374,0.1026960835,0.0352194123,0.1488352567,0.2517143488,-0.2204656601,0.3065986037,0.37360847,0.3654960096,-0.4923394322,-0.5268409848,-0.019487394,0.4596676826,-0.1460443437,0.2637678683,-0.7810152173,0.0906474739,0.0490969345,0.0831259042,0.7045721412,-0.4334190488,0.0311700702,-0.3011369109,0.2061882913,-0.0606679879,-0.5662207603,0.1323686391,-0.189680323,-0.349000603,0.0372541361,-0.1087566987,0.2165382355,0.5372943878,0.1653522104,0.276542902,0.3553307354,0.1084923223,-0.0140684778,0.3652673364,0.5064214468,-0.3291374445,0.1828556806,0.0533329546,0.1820658296,-0.2416639477,0.0324814171,0.0963764936,-0.0506631918,0.033735998,-0.3181338608,-0.1936038733,-0.4162679911,0.2353662252,-0.0543479361,0.280702889,0.398090899,0.2447252572,-0.0802798569,0.0983422548,-0.0073444112,0.074307099,0.2642246485,0.2403722405,0.3874528408,-0.456779182,-0.0083359247,-0.0959096402,-0.2261184007,-0.2581935227,-0.1545786709,-0.0594801307,-0.1083791703,0.4497854412,0.1455170512,0.1635873765,-0.1944186091,0.2312683165,0.0901709273,-0.1244404316,-0.0003969568,0.0900223777,0.0621659979,0.6624906659,-0.2686817646,-0.4959302843,-0.024275722,0.4147465229,0.2246547192,-0.0770734027,0.0569662452,0.2068658471,-0.2856074274,-0.1542339176,0.1575707048,-0.1063693017,-0.1054044962,0.110289596,-0.0285245404,-0.2333064377,0.2675739229,-0.1570523083,-0.1613530219,-0.091973044,-0.1774332672,-0.2806318104,-0.2413361669,-0.1948486269,-0.3812475502,-0.2053018361,0.2162438035,0.0149360513,0.1217982993,0.4578033984,-0.1445572674,0.0478685386,0.0947161987,0.1962702274,-0.0460679345,-0.2353435904,0.0793067068,-0.0385895707,-0.0526579283,0.1063538641,-0.0901943371,-0.157567963,-0.0040826849,0.1757276803,-0.0131400675,-0.2532690465,0.3499311507,-0.1179877147,-0.2294914573,-0.4110829234,-0.0305959303,0.0631312355,-0.0195532013,-0.083675079,0.2614877522,-0.3048563898,-0.2629296184,0.4484550953,-0.1784745902,-0.0706910565,-0.0015810376,0.2966397703,0.4717692733,-0.0923740268,-0.0766384527,0.0200872682,0.2456790954,0.051516626,0.1467117667,-0.1209753603,0.208591342,0.2829654813,0.3256470859,0.1422912925,0.1022858173,-0.2498798966,0.0113925459,0.0165359266,-0.2410697639,-0.1997709423,0.5652070045,0.0460899584,0.1719613969,0.0491668768,0.3464100063,0.1937997937,-0.0170080177,-0.0793520957,0.0941891074,-0.2903536558,0.0594080053,0.1839843094,0.2628834546,0.268840909,0.4361988008,0.1250479221,-0.0083826752,0.4837760925,0.2577881813,0.3150918782,0.5179544687,0.2492692918,-0.0326479152,-0.3772100806,0.1907109171,0.4949614108,-0.3534457684,-0.0469899923,-0.0648747385,0.2101315409,-0.2522915006,-0.3471132219,-0.2580218315,0.3463008702,-0.1636479646,-0.2520795166,0.2229753435,-0.0683909208,-0.0241794065,0.1769171804,-0.2295461595,0.1455718726,0.7704079151,0.1165910065,0.0410268083,-0.3815882504,-0.3096581101,-0.1625858843,0.1533405483,-0.1430544704,0.1957416385,-0.3325944245,-0.0119431438,-0.0246119518,0.3084972799,0.3674661815,0.3030283153,-0.1277540922,0.0828337222,-0.0116253607,0.1468795389,-0.0056867716,0.1899861991,-0.0187839139,0.1753653735,0.1828661114,-0.0948648155,-0.0923044905,-0.4366371036,0.0524755754,0.3449982405,-0.1648990363,0.1421356499,-0.3117671907,-0.1528320462,0.0407487936,0.2989790142,-0.2570863664,-0.097102575,0.53951478,-0.4918875694,0.0776443705,-0.1672454029,0.0359826759,-0.0960044935,0.5680377483,0.3039910793,-0.0151546439,-0.3837935925,-0.1259976923,-0.6268149018,-0.060552381,-0.3254488111,0.2521608174,0.0694541708,0.2867886722,0.0498369597,0.1253742278,0.3945254982,-0.0954661742,-0.0745222718,0.3330543339,-0.3459301889,0.4537988901,-0.2335766554,-0.2067918777,-0.0811020508,-0.4442897439,0.3849733472,0.0730768144,-0.1252094656,-0.1655890793,-0.0261920672,0.1744764149,0.1303198189,0.1954288632,0.3366410136,0.3453725278,0.0390878431,0.0184384286,-0.1458937526,0.1976263225,-0.03892323,0.3265101016,-0.0632227734,0.0547957793,-0.2816343606,0.1722019315,0.1057350487,-0.1668170691,0.0084096333,-0.2033714652,-0.3303463757,-0.0389226042,-0.0294360593,0.3369326293,0.0936920792,0.385831207,0.1202970594,-0.0968832821,-0.1704243124,-0.1283489615,-0.0215361305,-0.2728497386,-0.3036374748,-0.4393435419,0.2135729343,-0.2672087848,0.1529793292,-0.2265084535,-0.3191963732,0.3211677969,-0.0462014452,-0.3679229021,0.3406694829,-0.3830785155,-0.0795063451,-0.163463071,0.0433351472,-0.0836593658,-0.4640290141,0.1866618991,-0.1077286676]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1990","title":"OSError: Memory mapping file failed: Cannot allocate memory","comments":"Do you think this is trying to bring the dataset into memory and if I can avoid it to save on memory so it only brings a batch into memory? @lhoestq thank you","body":"Hi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n","comment_length":33,"text":"OSError: Memory mapping file failed: Cannot allocate memory\nHi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n\nDo you think this is trying to bring the dataset into memory and if I can avoid it to save on memory so it only brings a batch into memory? @lhoestq thank you","embeddings":[-0.2615419328,-0.0372641645,0.0520401485,0.6045719385,0.4538374543,0.2834013402,0.1456435472,0.2724665403,0.1785508692,0.1056612283,-0.0581841916,0.2273861468,-0.1771185547,-0.1459168196,-0.0370903499,-0.1933317184,0.084020853,0.0134716695,-0.5429398417,0.1604688615,-0.35815534,0.1157062352,-0.185632959,-0.1171597093,-0.2980377078,-0.1035162807,-0.0006043313,-0.0976312384,0.0188779235,-0.3274089992,0.2548563778,-0.1293895692,0.0864005461,0.5501937866,-0.0001231607,-0.0137761598,0.2835690081,-0.1672083884,-0.2533998191,-0.1284161955,-0.0416564196,-0.0297981855,0.0984970108,-0.2252508551,-0.0119242286,-0.052592624,0.2317306399,-0.34451437,0.4429748058,0.2243882269,0.1404946446,0.0699657425,0.4524520934,-0.0441127867,0.173906967,0.3341218829,0.0713807344,0.3320225477,-0.2908716202,-0.4356866479,-0.0304129403,0.3930506408,-0.039941147,0.0052876757,0.5653295517,-0.1445832551,0.0253530722,-0.3368177116,0.1123084202,-0.0072496962,0.5341086388,-0.490260005,-0.0283727888,-0.0147870975,-0.0533835031,-0.0952400044,0.2970977426,0.2689523995,-0.2722630501,-0.1447603852,-0.1304171085,-0.3249985576,-0.3197005689,0.4428686202,-0.0909217522,0.1910010129,-0.0247160532,0.3037818074,0.5653706789,-0.2050897777,-0.2340620607,-0.0758121908,0.1365436316,0.3287970722,-0.2677206099,-0.1660289764,-0.3029488623,-0.2047091722,0.3511367142,-0.4987058938,-0.3861933053,-0.1260007769,0.151669845,0.0212391727,0.4036456943,0.2817353308,-0.2963656187,0.3228372633,0.3122969866,0.1742716432,-0.2495381683,-0.1589899361,0.0767950714,-0.0373786837,-0.1110499874,-0.1914832443,0.0837740824,-0.0016539318,-0.0137460371,-0.0218034852,-0.2128316164,-0.1232094243,-0.0538497157,0.5055233836,-0.1040841639,-0.0846191123,0.373827666,0.1444597691,-0.0126846684,-0.0080771344,-0.0563751087,0.4104228318,-0.3436276913,0.284306854,0.0502635874,-0.0011586546,0.3263411224,-0.0727059469,-0.0331281535,-0.0566817522,0.1620447189,-0.3580230474,0.0087110661,0.1846879423,0.1788559556,0.2493289113,0.2503515184,-0.1479860991,-0.1526842713,0.1969967932,-0.163468048,-0.2433037609,0.038736742,0.0109290453,0.0980136842,0.2277469039,-0.277513355,0.1828102767,0.6528620124,-0.0775153115,-0.0572649427,-0.0801966265,-0.2296683937,-0.1687685549,0.2708779573,0.519313395,-0.0646530613,-0.1054837331,-0.1169700846,0.1677043438,0.2510127127,0.4413834214,-0.1638228297,0.2519159317,-0.097221449,0.0167985056,0.4141839147,-0.3105092943,-0.341904074,-0.005623993,-0.0252293199,-0.134322226,0.0232777502,0.1321309656,0.0774000436,0.1156655923,0.1119258851,0.3104183376,0.0561037064,0.2800839543,-0.3041983247,-0.2975762486,0.2528690994,0.0747464374,0.0653967336,-0.204082936,-0.0608494729,0.810767591,0.2972948551,-0.2623122931,0.1555345356,0.3147752583,0.1061372161,0.0064363666,0.0871187598,-0.2199127823,-0.2514474392,-0.0734247714,-0.0566988736,0.3305937052,-0.1451888531,-0.0414808579,0.1537224948,-0.1186705902,-0.1722497493,-0.3908631504,0.0622953624,0.0130732879,0.1235822812,0.0900844857,0.081060715,0.0560085624,-0.0939699262,0.2478925139,-0.5998023152,0.171503067,-0.2896074653,-0.107694298,-0.0214403253,-0.0624916069,0.0274150521,-0.0474654585,-0.068555817,0.1519924104,0.0771347657,-0.1671152562,-0.0906850323,0.0034200561,0.3101095259,-0.3221141994,0.1657055467,0.2421721667,0.1955043375,-0.0940227509,-0.1976960003,-0.1558783054,0.1051341444,0.3647213876,0.0791877508,0.1687752157,0.0262405463,0.0895476341,0.1797861308,-0.2070017755,0.218163535,0.01235501,0.2408843338,0.095006451,-0.0293943677,-0.3097331524,0.4583239555,0.2681971192,0.2772995234,0.2424818277,-0.4937297702,-0.0740422904,-0.065651603,-0.0718549192,0.3221566379,0.0500783995,-0.1195640638,0.1201550812,0.196337834,0.0097032078,0.3097595274,0.1743885428,0.4940783083,-0.0187082645,0.1057226509,-0.0806349367,-0.099587217,-0.1468482018,0.0500457101,0.5077639222,-0.175380379,-0.0157852471,-0.1864265651,-0.4006144702,-0.2753976882,0.164314881,-0.4642424881,-0.1369274259,-0.3849302828,0.3247582912,0.0695629567,0.2584396601,0.4160758257,-0.0507141761,0.3513425589,-0.1406714618,0.0687720701,-0.230412662,-0.1190755963,-0.0627371892,0.4409618676,-0.2204726189,0.0868323967,0.1567287594,-0.3195576966,-0.2135092765,-0.1457150131,0.1469898671,-0.0808212906,0.1255745143,0.0596423596,0.5084351897,-0.1233984753,-0.2323765159,0.1721098423,0.089907065,-0.0588723905,-0.0373000093,0.0232937448,0.1553850919,0.0540889502,-0.2931139171,-0.179757297,-0.508184433,0.4014800787,-0.033890903,0.1513629258,0.3279185891,0.1617626399,0.1156149805,-0.1213662103,0.0894483551,-0.1621506959,0.0160200186,0.3207914829,-0.1488257647,-0.2114350796,-0.0290636532,0.0627378225,0.306303829,0.2301850766,-0.6157578826,-0.0209107678,-0.1136818826,-0.0108944643,-0.0588507131,0.286311388,0.3913796842,0.078921251,0.107800208,0.1375889033,-0.1677574217,0.094906956,-0.0409829542,0.3248968422,0.2200790793,0.5234260559,0.0637268201,0.8384488225,0.341964066,0.1871405095,0.2288259119,-0.0220659021,0.0835645646,0.0049702087,-0.3481871486,-0.0082089063,-0.0368972011,0.0619993955,0.1661427468,0.0087548671,-0.3374429345,-0.1676723063,-0.3505290449,0.0459905639,-0.3771670461,0.2698533237,0.1751251519,0.3674235642,-0.0838355944,-0.0824133605,0.0231614523,-0.3515779972,0.1945331246,0.2135322541,-0.0175284855,0.0211241636,-0.1137084141,-0.3958891332,-0.6405459046,0.1182040572,-0.1544018984,0.0629801229,-0.1038483456,0.0933976695,0.0600519739,0.1583072543,0.7553266883,0.0059234039,-0.3072926104,-0.0461269431,-0.2609421611,-0.5063745379,0.2137841731,-0.1841569394,0.3441569507,0.143389374,0.5205535889,-0.3616023064,-0.0981338844,0.321287632,0.374714762,-0.2170827985,-0.2046922445,-0.1343318671,-0.2551260591,-0.6043947935,0.0041025551,0.0545691848,0.1934488118,0.4964534044,0.2044566721,0.0526693799,-0.0508972369,0.1371936649,-0.0533758029,0.3105924428,0.110654138,0.1008514464,0.1345039904,-0.1580678225,0.3764185309,0.2939079404,0.0716900676,-0.379755497,-0.0997854844,0.0576216578,0.2523964047,-0.0627555698,-0.0139149139,0.0038341719,0.0268995017,0.1780955493,-0.0997749195,0.1322197765,0.1875683814,0.0874076858,-0.4511457086,-0.4366309047,0.3414416313,0.1997043192,0.1596941948,0.3301558793,-0.1424441189,-0.5566364527,0.3253412545,0.3589582145,0.960832715,-0.3551480174,0.3937279582,0.0962183997,0.2117647976,0.6567651629,-0.4751399457,0.3529615998,-0.3699406683,-0.0815761462,0.0431752093,-0.1330996603,0.1379617155,-0.0172327515,-0.4360865951,0.2891325355,0.0913326368,-0.0398276299,-0.1840585619,0.3883704841,0.1124473214,-0.4811996818,-0.1841866523,0.0319509618,-0.1269843727,0.243967101,-0.0456034765,-0.0274401437,0.115675576,-0.1150276288,-0.3924042284,-0.0374513902,-0.4836362898,0.3076913059,-0.2517828047,0.0015952894,0.4047460854,0.3928067088,0.0101812379,0.3698486388,-0.1340497881,0.0105267046,-0.3600815237,-0.2767649889,-0.185056895,0.0676643252,0.2424043119,-0.1937424839,-0.1071275026,0.0763597563,-0.0399471037,-0.1982413977,-0.2103530616,-0.0318763927,-0.1298194379,-0.2045654804,-0.2123933136,0.0132841198,-0.3584757745,-0.0186204016,0.0458632261,0.0108470637,-0.3378366232,0.3286833465,0.1423627734,-0.4008429646,0.0307404082,0.4164584875,0.348543793,0.1605854779,0.7200255394,0.3232616484,-0.2456386834,-0.1618594825,-0.0824732929,-0.025335988,-0.304834336,0.2100929171,0.083754994,0.106053941,-0.3066006005,-0.0003773624,0.1285434067,0.0452359319,0.0363892019,-0.3043982387,-0.5602001548,0.2082497925,-0.010477718,0.036499247,0.236894533,0.0221908819,0.0295315944,0.2480150908,-0.156630531,0.020054128,-0.1691422611,0.2507917583,0.2438191026,-0.1116867065,0.1032438129,-0.0129493577,-0.0043390617,0.1593108326,-0.0150237987,-0.1269526333,0.0454815626,0.1532624811,-0.026851235,-0.2777851522,-0.0963580236,-0.5430374146,-0.0158967581,-0.3039938509,0.0111275995,0.0212793741,-0.1362702698,0.110977523,0.2019030601,-0.2731659412,-0.063025862,0.2792618573,-0.0221649799,-0.050752569,0.1658708602,0.2017495036,0.0937991664,-0.1570737511,-0.5362299681,0.055834543,-0.0843362063,-0.0266278703,0.1019938439,-0.1142851859,-0.3993756771,0.3194217682,0.0904670879,0.1510138661,-0.1423276961,0.20684129,0.2004518658,0.0089688553,-0.1937268674,-0.0418563969,0.3301984668,-0.2920083702,-0.0505978391,0.3277898431,-0.1311677396,0.0453453697,-0.0299580656,0.1311582178,0.3588027954,-0.0734308213,0.3226182461,0.0061582224,-0.0834842399,-0.0167209767,0.1939019412,0.4057275653,0.2587025166,0.204386279,-0.1556384861,0.0376391597,-0.1058716699,-0.2548227906,0.0379435755,-0.0836717263,0.1414890438,0.4189734757,-0.1775912941,0.1592200398,-0.193913281,0.1780178845,0.1845969856,-0.2745580077,0.078776978,0.2984717786,-0.2116876245,0.0464751199,0.1499889493,-0.115781568,-0.0657879114,0.141785726,0.0771138668,-0.6087394357,0.4591955841,0.1467865705,-0.2206508368,-0.1409267187,0.2773652971,0.4522538185,0.0710785612,-0.3053991199,0.0575940534,0.0269881077,0.136588797,-0.3318274915,0.309446007,0.4621617496,0.2915849388,-0.1956237853,0.2535728216,-0.1581641287,-0.012619175,0.0154259196,-0.0236886721,0.116465725,0.2165056765,0.1130549312,-0.0109328497,-0.0586273782,-0.151095286,0.1630205214,0.1425201148,-0.227916345,-0.3811423779,-0.0991922468,-0.171089232,-0.0140399672,-0.0115701789,-0.5901709795,0.2826703489,0.5286229849,-0.1380668133,0.081210427,0.0202327408,0.0137296338,-0.031966269,0.4846240878,0.2030230612,-0.0330804884,-0.481416434,-0.3177812397,-0.4150764942,0.0241078436,-0.3297815621,0.0040629571,-0.075407438,0.1153876483,-0.0473557673,0.0465940349,0.0486276001,-0.3642242849,-0.0435181372,0.1776145101,-0.1560722739,-0.2553372085,-0.1273898184,0.17875956,0.1793888956,-0.1981143653,0.2631115615,0.0087835463,-0.1398376226,-0.3518662453,0.1338686347,0.0501965582,0.177921772,0.1576354504,0.0610057227,0.3895047605,0.0040854425,-0.2422718108,0.0258157793,0.0002693687,-0.0998139083,0.1413407028,0.1294570565,0.3691285551,-0.4097653031,-0.3827797174,-0.4327935874,0.2897745073,-0.0637847111,-0.1197685599,-0.0350803994,-0.0376015715,0.2526537478,0.0688290522,0.0636018813,0.1586660147,-0.0656890944,-0.0003347614,-0.4509042203,-0.2615719438,0.459885478,-0.4583003223,-0.3856023252,-0.086100094,-0.0950353891,-0.1459356546,-0.1394127011,-0.7795305252,0.0910593942,0.2944117486,0.0955306068,-0.4224690795,0.0314957574,-0.1058145389,0.1201661602,-0.1256255955,0.3365498185,-0.0900946259,-0.432505995,0.1651308239,-0.2077717483]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1990","title":"OSError: Memory mapping file failed: Cannot allocate memory","comments":"It's not trying to bring the dataset into memory.\r\n\r\nActually, it's trying to memory map the dataset file, which is different. It allows to load large dataset files without filling up memory.\r\n\r\nWhat dataset did you use to get this error ?\r\nOn what OS are you running ? What's your python and pyarrow version ?","body":"Hi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n","comment_length":56,"text":"OSError: Memory mapping file failed: Cannot allocate memory\nHi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n\nIt's not trying to bring the dataset into memory.\r\n\r\nActually, it's trying to memory map the dataset file, which is different. It allows to load large dataset files without filling up memory.\r\n\r\nWhat dataset did you use to get this error ?\r\nOn what OS are you running ? What's your python and pyarrow version ?","embeddings":[-0.2615419328,-0.0372641645,0.0520401485,0.6045719385,0.4538374543,0.2834013402,0.1456435472,0.2724665403,0.1785508692,0.1056612283,-0.0581841916,0.2273861468,-0.1771185547,-0.1459168196,-0.0370903499,-0.1933317184,0.084020853,0.0134716695,-0.5429398417,0.1604688615,-0.35815534,0.1157062352,-0.185632959,-0.1171597093,-0.2980377078,-0.1035162807,-0.0006043313,-0.0976312384,0.0188779235,-0.3274089992,0.2548563778,-0.1293895692,0.0864005461,0.5501937866,-0.0001231607,-0.0137761598,0.2835690081,-0.1672083884,-0.2533998191,-0.1284161955,-0.0416564196,-0.0297981855,0.0984970108,-0.2252508551,-0.0119242286,-0.052592624,0.2317306399,-0.34451437,0.4429748058,0.2243882269,0.1404946446,0.0699657425,0.4524520934,-0.0441127867,0.173906967,0.3341218829,0.0713807344,0.3320225477,-0.2908716202,-0.4356866479,-0.0304129403,0.3930506408,-0.039941147,0.0052876757,0.5653295517,-0.1445832551,0.0253530722,-0.3368177116,0.1123084202,-0.0072496962,0.5341086388,-0.490260005,-0.0283727888,-0.0147870975,-0.0533835031,-0.0952400044,0.2970977426,0.2689523995,-0.2722630501,-0.1447603852,-0.1304171085,-0.3249985576,-0.3197005689,0.4428686202,-0.0909217522,0.1910010129,-0.0247160532,0.3037818074,0.5653706789,-0.2050897777,-0.2340620607,-0.0758121908,0.1365436316,0.3287970722,-0.2677206099,-0.1660289764,-0.3029488623,-0.2047091722,0.3511367142,-0.4987058938,-0.3861933053,-0.1260007769,0.151669845,0.0212391727,0.4036456943,0.2817353308,-0.2963656187,0.3228372633,0.3122969866,0.1742716432,-0.2495381683,-0.1589899361,0.0767950714,-0.0373786837,-0.1110499874,-0.1914832443,0.0837740824,-0.0016539318,-0.0137460371,-0.0218034852,-0.2128316164,-0.1232094243,-0.0538497157,0.5055233836,-0.1040841639,-0.0846191123,0.373827666,0.1444597691,-0.0126846684,-0.0080771344,-0.0563751087,0.4104228318,-0.3436276913,0.284306854,0.0502635874,-0.0011586546,0.3263411224,-0.0727059469,-0.0331281535,-0.0566817522,0.1620447189,-0.3580230474,0.0087110661,0.1846879423,0.1788559556,0.2493289113,0.2503515184,-0.1479860991,-0.1526842713,0.1969967932,-0.163468048,-0.2433037609,0.038736742,0.0109290453,0.0980136842,0.2277469039,-0.277513355,0.1828102767,0.6528620124,-0.0775153115,-0.0572649427,-0.0801966265,-0.2296683937,-0.1687685549,0.2708779573,0.519313395,-0.0646530613,-0.1054837331,-0.1169700846,0.1677043438,0.2510127127,0.4413834214,-0.1638228297,0.2519159317,-0.097221449,0.0167985056,0.4141839147,-0.3105092943,-0.341904074,-0.005623993,-0.0252293199,-0.134322226,0.0232777502,0.1321309656,0.0774000436,0.1156655923,0.1119258851,0.3104183376,0.0561037064,0.2800839543,-0.3041983247,-0.2975762486,0.2528690994,0.0747464374,0.0653967336,-0.204082936,-0.0608494729,0.810767591,0.2972948551,-0.2623122931,0.1555345356,0.3147752583,0.1061372161,0.0064363666,0.0871187598,-0.2199127823,-0.2514474392,-0.0734247714,-0.0566988736,0.3305937052,-0.1451888531,-0.0414808579,0.1537224948,-0.1186705902,-0.1722497493,-0.3908631504,0.0622953624,0.0130732879,0.1235822812,0.0900844857,0.081060715,0.0560085624,-0.0939699262,0.2478925139,-0.5998023152,0.171503067,-0.2896074653,-0.107694298,-0.0214403253,-0.0624916069,0.0274150521,-0.0474654585,-0.068555817,0.1519924104,0.0771347657,-0.1671152562,-0.0906850323,0.0034200561,0.3101095259,-0.3221141994,0.1657055467,0.2421721667,0.1955043375,-0.0940227509,-0.1976960003,-0.1558783054,0.1051341444,0.3647213876,0.0791877508,0.1687752157,0.0262405463,0.0895476341,0.1797861308,-0.2070017755,0.218163535,0.01235501,0.2408843338,0.095006451,-0.0293943677,-0.3097331524,0.4583239555,0.2681971192,0.2772995234,0.2424818277,-0.4937297702,-0.0740422904,-0.065651603,-0.0718549192,0.3221566379,0.0500783995,-0.1195640638,0.1201550812,0.196337834,0.0097032078,0.3097595274,0.1743885428,0.4940783083,-0.0187082645,0.1057226509,-0.0806349367,-0.099587217,-0.1468482018,0.0500457101,0.5077639222,-0.175380379,-0.0157852471,-0.1864265651,-0.4006144702,-0.2753976882,0.164314881,-0.4642424881,-0.1369274259,-0.3849302828,0.3247582912,0.0695629567,0.2584396601,0.4160758257,-0.0507141761,0.3513425589,-0.1406714618,0.0687720701,-0.230412662,-0.1190755963,-0.0627371892,0.4409618676,-0.2204726189,0.0868323967,0.1567287594,-0.3195576966,-0.2135092765,-0.1457150131,0.1469898671,-0.0808212906,0.1255745143,0.0596423596,0.5084351897,-0.1233984753,-0.2323765159,0.1721098423,0.089907065,-0.0588723905,-0.0373000093,0.0232937448,0.1553850919,0.0540889502,-0.2931139171,-0.179757297,-0.508184433,0.4014800787,-0.033890903,0.1513629258,0.3279185891,0.1617626399,0.1156149805,-0.1213662103,0.0894483551,-0.1621506959,0.0160200186,0.3207914829,-0.1488257647,-0.2114350796,-0.0290636532,0.0627378225,0.306303829,0.2301850766,-0.6157578826,-0.0209107678,-0.1136818826,-0.0108944643,-0.0588507131,0.286311388,0.3913796842,0.078921251,0.107800208,0.1375889033,-0.1677574217,0.094906956,-0.0409829542,0.3248968422,0.2200790793,0.5234260559,0.0637268201,0.8384488225,0.341964066,0.1871405095,0.2288259119,-0.0220659021,0.0835645646,0.0049702087,-0.3481871486,-0.0082089063,-0.0368972011,0.0619993955,0.1661427468,0.0087548671,-0.3374429345,-0.1676723063,-0.3505290449,0.0459905639,-0.3771670461,0.2698533237,0.1751251519,0.3674235642,-0.0838355944,-0.0824133605,0.0231614523,-0.3515779972,0.1945331246,0.2135322541,-0.0175284855,0.0211241636,-0.1137084141,-0.3958891332,-0.6405459046,0.1182040572,-0.1544018984,0.0629801229,-0.1038483456,0.0933976695,0.0600519739,0.1583072543,0.7553266883,0.0059234039,-0.3072926104,-0.0461269431,-0.2609421611,-0.5063745379,0.2137841731,-0.1841569394,0.3441569507,0.143389374,0.5205535889,-0.3616023064,-0.0981338844,0.321287632,0.374714762,-0.2170827985,-0.2046922445,-0.1343318671,-0.2551260591,-0.6043947935,0.0041025551,0.0545691848,0.1934488118,0.4964534044,0.2044566721,0.0526693799,-0.0508972369,0.1371936649,-0.0533758029,0.3105924428,0.110654138,0.1008514464,0.1345039904,-0.1580678225,0.3764185309,0.2939079404,0.0716900676,-0.379755497,-0.0997854844,0.0576216578,0.2523964047,-0.0627555698,-0.0139149139,0.0038341719,0.0268995017,0.1780955493,-0.0997749195,0.1322197765,0.1875683814,0.0874076858,-0.4511457086,-0.4366309047,0.3414416313,0.1997043192,0.1596941948,0.3301558793,-0.1424441189,-0.5566364527,0.3253412545,0.3589582145,0.960832715,-0.3551480174,0.3937279582,0.0962183997,0.2117647976,0.6567651629,-0.4751399457,0.3529615998,-0.3699406683,-0.0815761462,0.0431752093,-0.1330996603,0.1379617155,-0.0172327515,-0.4360865951,0.2891325355,0.0913326368,-0.0398276299,-0.1840585619,0.3883704841,0.1124473214,-0.4811996818,-0.1841866523,0.0319509618,-0.1269843727,0.243967101,-0.0456034765,-0.0274401437,0.115675576,-0.1150276288,-0.3924042284,-0.0374513902,-0.4836362898,0.3076913059,-0.2517828047,0.0015952894,0.4047460854,0.3928067088,0.0101812379,0.3698486388,-0.1340497881,0.0105267046,-0.3600815237,-0.2767649889,-0.185056895,0.0676643252,0.2424043119,-0.1937424839,-0.1071275026,0.0763597563,-0.0399471037,-0.1982413977,-0.2103530616,-0.0318763927,-0.1298194379,-0.2045654804,-0.2123933136,0.0132841198,-0.3584757745,-0.0186204016,0.0458632261,0.0108470637,-0.3378366232,0.3286833465,0.1423627734,-0.4008429646,0.0307404082,0.4164584875,0.348543793,0.1605854779,0.7200255394,0.3232616484,-0.2456386834,-0.1618594825,-0.0824732929,-0.025335988,-0.304834336,0.2100929171,0.083754994,0.106053941,-0.3066006005,-0.0003773624,0.1285434067,0.0452359319,0.0363892019,-0.3043982387,-0.5602001548,0.2082497925,-0.010477718,0.036499247,0.236894533,0.0221908819,0.0295315944,0.2480150908,-0.156630531,0.020054128,-0.1691422611,0.2507917583,0.2438191026,-0.1116867065,0.1032438129,-0.0129493577,-0.0043390617,0.1593108326,-0.0150237987,-0.1269526333,0.0454815626,0.1532624811,-0.026851235,-0.2777851522,-0.0963580236,-0.5430374146,-0.0158967581,-0.3039938509,0.0111275995,0.0212793741,-0.1362702698,0.110977523,0.2019030601,-0.2731659412,-0.063025862,0.2792618573,-0.0221649799,-0.050752569,0.1658708602,0.2017495036,0.0937991664,-0.1570737511,-0.5362299681,0.055834543,-0.0843362063,-0.0266278703,0.1019938439,-0.1142851859,-0.3993756771,0.3194217682,0.0904670879,0.1510138661,-0.1423276961,0.20684129,0.2004518658,0.0089688553,-0.1937268674,-0.0418563969,0.3301984668,-0.2920083702,-0.0505978391,0.3277898431,-0.1311677396,0.0453453697,-0.0299580656,0.1311582178,0.3588027954,-0.0734308213,0.3226182461,0.0061582224,-0.0834842399,-0.0167209767,0.1939019412,0.4057275653,0.2587025166,0.204386279,-0.1556384861,0.0376391597,-0.1058716699,-0.2548227906,0.0379435755,-0.0836717263,0.1414890438,0.4189734757,-0.1775912941,0.1592200398,-0.193913281,0.1780178845,0.1845969856,-0.2745580077,0.078776978,0.2984717786,-0.2116876245,0.0464751199,0.1499889493,-0.115781568,-0.0657879114,0.141785726,0.0771138668,-0.6087394357,0.4591955841,0.1467865705,-0.2206508368,-0.1409267187,0.2773652971,0.4522538185,0.0710785612,-0.3053991199,0.0575940534,0.0269881077,0.136588797,-0.3318274915,0.309446007,0.4621617496,0.2915849388,-0.1956237853,0.2535728216,-0.1581641287,-0.012619175,0.0154259196,-0.0236886721,0.116465725,0.2165056765,0.1130549312,-0.0109328497,-0.0586273782,-0.151095286,0.1630205214,0.1425201148,-0.227916345,-0.3811423779,-0.0991922468,-0.171089232,-0.0140399672,-0.0115701789,-0.5901709795,0.2826703489,0.5286229849,-0.1380668133,0.081210427,0.0202327408,0.0137296338,-0.031966269,0.4846240878,0.2030230612,-0.0330804884,-0.481416434,-0.3177812397,-0.4150764942,0.0241078436,-0.3297815621,0.0040629571,-0.075407438,0.1153876483,-0.0473557673,0.0465940349,0.0486276001,-0.3642242849,-0.0435181372,0.1776145101,-0.1560722739,-0.2553372085,-0.1273898184,0.17875956,0.1793888956,-0.1981143653,0.2631115615,0.0087835463,-0.1398376226,-0.3518662453,0.1338686347,0.0501965582,0.177921772,0.1576354504,0.0610057227,0.3895047605,0.0040854425,-0.2422718108,0.0258157793,0.0002693687,-0.0998139083,0.1413407028,0.1294570565,0.3691285551,-0.4097653031,-0.3827797174,-0.4327935874,0.2897745073,-0.0637847111,-0.1197685599,-0.0350803994,-0.0376015715,0.2526537478,0.0688290522,0.0636018813,0.1586660147,-0.0656890944,-0.0003347614,-0.4509042203,-0.2615719438,0.459885478,-0.4583003223,-0.3856023252,-0.086100094,-0.0950353891,-0.1459356546,-0.1394127011,-0.7795305252,0.0910593942,0.2944117486,0.0955306068,-0.4224690795,0.0314957574,-0.1058145389,0.1201661602,-0.1256255955,0.3365498185,-0.0900946259,-0.432505995,0.1651308239,-0.2077717483]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1990","title":"OSError: Memory mapping file failed: Cannot allocate memory","comments":"Dear @lhoestq \r\nthank you so much for coming back to me. Please find info below:\r\n1) Dataset name: I used wikipedia with config 20200501.en\r\n2) I got these pyarrow in my environment:\r\npyarrow 2.0.0 \r\npyarrow 3.0.0 \r\n\r\n3) python version 3.7.10\r\n4) OS version \r\n\r\nlsb_release -a\r\nNo LSB modules are available.\r\nDistributor ID:\tDebian\r\nDescription:\tDebian GNU\/Linux 10 (buster)\r\nRelease:\t10\r\nCodename:\tbuster\r\n\r\n\r\nIs there a way I could solve the memory issue and if I could run this model, I am using GeForce GTX 108, \r\nthanks \r\n","body":"Hi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n","comment_length":88,"text":"OSError: Memory mapping file failed: Cannot allocate memory\nHi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n\nDear @lhoestq \r\nthank you so much for coming back to me. Please find info below:\r\n1) Dataset name: I used wikipedia with config 20200501.en\r\n2) I got these pyarrow in my environment:\r\npyarrow 2.0.0 \r\npyarrow 3.0.0 \r\n\r\n3) python version 3.7.10\r\n4) OS version \r\n\r\nlsb_release -a\r\nNo LSB modules are available.\r\nDistributor ID:\tDebian\r\nDescription:\tDebian GNU\/Linux 10 (buster)\r\nRelease:\t10\r\nCodename:\tbuster\r\n\r\n\r\nIs there a way I could solve the memory issue and if I could run this model, I am using GeForce GTX 108, \r\nthanks \r\n","embeddings":[-0.2615419328,-0.0372641645,0.0520401485,0.6045719385,0.4538374543,0.2834013402,0.1456435472,0.2724665403,0.1785508692,0.1056612283,-0.0581841916,0.2273861468,-0.1771185547,-0.1459168196,-0.0370903499,-0.1933317184,0.084020853,0.0134716695,-0.5429398417,0.1604688615,-0.35815534,0.1157062352,-0.185632959,-0.1171597093,-0.2980377078,-0.1035162807,-0.0006043313,-0.0976312384,0.0188779235,-0.3274089992,0.2548563778,-0.1293895692,0.0864005461,0.5501937866,-0.0001231607,-0.0137761598,0.2835690081,-0.1672083884,-0.2533998191,-0.1284161955,-0.0416564196,-0.0297981855,0.0984970108,-0.2252508551,-0.0119242286,-0.052592624,0.2317306399,-0.34451437,0.4429748058,0.2243882269,0.1404946446,0.0699657425,0.4524520934,-0.0441127867,0.173906967,0.3341218829,0.0713807344,0.3320225477,-0.2908716202,-0.4356866479,-0.0304129403,0.3930506408,-0.039941147,0.0052876757,0.5653295517,-0.1445832551,0.0253530722,-0.3368177116,0.1123084202,-0.0072496962,0.5341086388,-0.490260005,-0.0283727888,-0.0147870975,-0.0533835031,-0.0952400044,0.2970977426,0.2689523995,-0.2722630501,-0.1447603852,-0.1304171085,-0.3249985576,-0.3197005689,0.4428686202,-0.0909217522,0.1910010129,-0.0247160532,0.3037818074,0.5653706789,-0.2050897777,-0.2340620607,-0.0758121908,0.1365436316,0.3287970722,-0.2677206099,-0.1660289764,-0.3029488623,-0.2047091722,0.3511367142,-0.4987058938,-0.3861933053,-0.1260007769,0.151669845,0.0212391727,0.4036456943,0.2817353308,-0.2963656187,0.3228372633,0.3122969866,0.1742716432,-0.2495381683,-0.1589899361,0.0767950714,-0.0373786837,-0.1110499874,-0.1914832443,0.0837740824,-0.0016539318,-0.0137460371,-0.0218034852,-0.2128316164,-0.1232094243,-0.0538497157,0.5055233836,-0.1040841639,-0.0846191123,0.373827666,0.1444597691,-0.0126846684,-0.0080771344,-0.0563751087,0.4104228318,-0.3436276913,0.284306854,0.0502635874,-0.0011586546,0.3263411224,-0.0727059469,-0.0331281535,-0.0566817522,0.1620447189,-0.3580230474,0.0087110661,0.1846879423,0.1788559556,0.2493289113,0.2503515184,-0.1479860991,-0.1526842713,0.1969967932,-0.163468048,-0.2433037609,0.038736742,0.0109290453,0.0980136842,0.2277469039,-0.277513355,0.1828102767,0.6528620124,-0.0775153115,-0.0572649427,-0.0801966265,-0.2296683937,-0.1687685549,0.2708779573,0.519313395,-0.0646530613,-0.1054837331,-0.1169700846,0.1677043438,0.2510127127,0.4413834214,-0.1638228297,0.2519159317,-0.097221449,0.0167985056,0.4141839147,-0.3105092943,-0.341904074,-0.005623993,-0.0252293199,-0.134322226,0.0232777502,0.1321309656,0.0774000436,0.1156655923,0.1119258851,0.3104183376,0.0561037064,0.2800839543,-0.3041983247,-0.2975762486,0.2528690994,0.0747464374,0.0653967336,-0.204082936,-0.0608494729,0.810767591,0.2972948551,-0.2623122931,0.1555345356,0.3147752583,0.1061372161,0.0064363666,0.0871187598,-0.2199127823,-0.2514474392,-0.0734247714,-0.0566988736,0.3305937052,-0.1451888531,-0.0414808579,0.1537224948,-0.1186705902,-0.1722497493,-0.3908631504,0.0622953624,0.0130732879,0.1235822812,0.0900844857,0.081060715,0.0560085624,-0.0939699262,0.2478925139,-0.5998023152,0.171503067,-0.2896074653,-0.107694298,-0.0214403253,-0.0624916069,0.0274150521,-0.0474654585,-0.068555817,0.1519924104,0.0771347657,-0.1671152562,-0.0906850323,0.0034200561,0.3101095259,-0.3221141994,0.1657055467,0.2421721667,0.1955043375,-0.0940227509,-0.1976960003,-0.1558783054,0.1051341444,0.3647213876,0.0791877508,0.1687752157,0.0262405463,0.0895476341,0.1797861308,-0.2070017755,0.218163535,0.01235501,0.2408843338,0.095006451,-0.0293943677,-0.3097331524,0.4583239555,0.2681971192,0.2772995234,0.2424818277,-0.4937297702,-0.0740422904,-0.065651603,-0.0718549192,0.3221566379,0.0500783995,-0.1195640638,0.1201550812,0.196337834,0.0097032078,0.3097595274,0.1743885428,0.4940783083,-0.0187082645,0.1057226509,-0.0806349367,-0.099587217,-0.1468482018,0.0500457101,0.5077639222,-0.175380379,-0.0157852471,-0.1864265651,-0.4006144702,-0.2753976882,0.164314881,-0.4642424881,-0.1369274259,-0.3849302828,0.3247582912,0.0695629567,0.2584396601,0.4160758257,-0.0507141761,0.3513425589,-0.1406714618,0.0687720701,-0.230412662,-0.1190755963,-0.0627371892,0.4409618676,-0.2204726189,0.0868323967,0.1567287594,-0.3195576966,-0.2135092765,-0.1457150131,0.1469898671,-0.0808212906,0.1255745143,0.0596423596,0.5084351897,-0.1233984753,-0.2323765159,0.1721098423,0.089907065,-0.0588723905,-0.0373000093,0.0232937448,0.1553850919,0.0540889502,-0.2931139171,-0.179757297,-0.508184433,0.4014800787,-0.033890903,0.1513629258,0.3279185891,0.1617626399,0.1156149805,-0.1213662103,0.0894483551,-0.1621506959,0.0160200186,0.3207914829,-0.1488257647,-0.2114350796,-0.0290636532,0.0627378225,0.306303829,0.2301850766,-0.6157578826,-0.0209107678,-0.1136818826,-0.0108944643,-0.0588507131,0.286311388,0.3913796842,0.078921251,0.107800208,0.1375889033,-0.1677574217,0.094906956,-0.0409829542,0.3248968422,0.2200790793,0.5234260559,0.0637268201,0.8384488225,0.341964066,0.1871405095,0.2288259119,-0.0220659021,0.0835645646,0.0049702087,-0.3481871486,-0.0082089063,-0.0368972011,0.0619993955,0.1661427468,0.0087548671,-0.3374429345,-0.1676723063,-0.3505290449,0.0459905639,-0.3771670461,0.2698533237,0.1751251519,0.3674235642,-0.0838355944,-0.0824133605,0.0231614523,-0.3515779972,0.1945331246,0.2135322541,-0.0175284855,0.0211241636,-0.1137084141,-0.3958891332,-0.6405459046,0.1182040572,-0.1544018984,0.0629801229,-0.1038483456,0.0933976695,0.0600519739,0.1583072543,0.7553266883,0.0059234039,-0.3072926104,-0.0461269431,-0.2609421611,-0.5063745379,0.2137841731,-0.1841569394,0.3441569507,0.143389374,0.5205535889,-0.3616023064,-0.0981338844,0.321287632,0.374714762,-0.2170827985,-0.2046922445,-0.1343318671,-0.2551260591,-0.6043947935,0.0041025551,0.0545691848,0.1934488118,0.4964534044,0.2044566721,0.0526693799,-0.0508972369,0.1371936649,-0.0533758029,0.3105924428,0.110654138,0.1008514464,0.1345039904,-0.1580678225,0.3764185309,0.2939079404,0.0716900676,-0.379755497,-0.0997854844,0.0576216578,0.2523964047,-0.0627555698,-0.0139149139,0.0038341719,0.0268995017,0.1780955493,-0.0997749195,0.1322197765,0.1875683814,0.0874076858,-0.4511457086,-0.4366309047,0.3414416313,0.1997043192,0.1596941948,0.3301558793,-0.1424441189,-0.5566364527,0.3253412545,0.3589582145,0.960832715,-0.3551480174,0.3937279582,0.0962183997,0.2117647976,0.6567651629,-0.4751399457,0.3529615998,-0.3699406683,-0.0815761462,0.0431752093,-0.1330996603,0.1379617155,-0.0172327515,-0.4360865951,0.2891325355,0.0913326368,-0.0398276299,-0.1840585619,0.3883704841,0.1124473214,-0.4811996818,-0.1841866523,0.0319509618,-0.1269843727,0.243967101,-0.0456034765,-0.0274401437,0.115675576,-0.1150276288,-0.3924042284,-0.0374513902,-0.4836362898,0.3076913059,-0.2517828047,0.0015952894,0.4047460854,0.3928067088,0.0101812379,0.3698486388,-0.1340497881,0.0105267046,-0.3600815237,-0.2767649889,-0.185056895,0.0676643252,0.2424043119,-0.1937424839,-0.1071275026,0.0763597563,-0.0399471037,-0.1982413977,-0.2103530616,-0.0318763927,-0.1298194379,-0.2045654804,-0.2123933136,0.0132841198,-0.3584757745,-0.0186204016,0.0458632261,0.0108470637,-0.3378366232,0.3286833465,0.1423627734,-0.4008429646,0.0307404082,0.4164584875,0.348543793,0.1605854779,0.7200255394,0.3232616484,-0.2456386834,-0.1618594825,-0.0824732929,-0.025335988,-0.304834336,0.2100929171,0.083754994,0.106053941,-0.3066006005,-0.0003773624,0.1285434067,0.0452359319,0.0363892019,-0.3043982387,-0.5602001548,0.2082497925,-0.010477718,0.036499247,0.236894533,0.0221908819,0.0295315944,0.2480150908,-0.156630531,0.020054128,-0.1691422611,0.2507917583,0.2438191026,-0.1116867065,0.1032438129,-0.0129493577,-0.0043390617,0.1593108326,-0.0150237987,-0.1269526333,0.0454815626,0.1532624811,-0.026851235,-0.2777851522,-0.0963580236,-0.5430374146,-0.0158967581,-0.3039938509,0.0111275995,0.0212793741,-0.1362702698,0.110977523,0.2019030601,-0.2731659412,-0.063025862,0.2792618573,-0.0221649799,-0.050752569,0.1658708602,0.2017495036,0.0937991664,-0.1570737511,-0.5362299681,0.055834543,-0.0843362063,-0.0266278703,0.1019938439,-0.1142851859,-0.3993756771,0.3194217682,0.0904670879,0.1510138661,-0.1423276961,0.20684129,0.2004518658,0.0089688553,-0.1937268674,-0.0418563969,0.3301984668,-0.2920083702,-0.0505978391,0.3277898431,-0.1311677396,0.0453453697,-0.0299580656,0.1311582178,0.3588027954,-0.0734308213,0.3226182461,0.0061582224,-0.0834842399,-0.0167209767,0.1939019412,0.4057275653,0.2587025166,0.204386279,-0.1556384861,0.0376391597,-0.1058716699,-0.2548227906,0.0379435755,-0.0836717263,0.1414890438,0.4189734757,-0.1775912941,0.1592200398,-0.193913281,0.1780178845,0.1845969856,-0.2745580077,0.078776978,0.2984717786,-0.2116876245,0.0464751199,0.1499889493,-0.115781568,-0.0657879114,0.141785726,0.0771138668,-0.6087394357,0.4591955841,0.1467865705,-0.2206508368,-0.1409267187,0.2773652971,0.4522538185,0.0710785612,-0.3053991199,0.0575940534,0.0269881077,0.136588797,-0.3318274915,0.309446007,0.4621617496,0.2915849388,-0.1956237853,0.2535728216,-0.1581641287,-0.012619175,0.0154259196,-0.0236886721,0.116465725,0.2165056765,0.1130549312,-0.0109328497,-0.0586273782,-0.151095286,0.1630205214,0.1425201148,-0.227916345,-0.3811423779,-0.0991922468,-0.171089232,-0.0140399672,-0.0115701789,-0.5901709795,0.2826703489,0.5286229849,-0.1380668133,0.081210427,0.0202327408,0.0137296338,-0.031966269,0.4846240878,0.2030230612,-0.0330804884,-0.481416434,-0.3177812397,-0.4150764942,0.0241078436,-0.3297815621,0.0040629571,-0.075407438,0.1153876483,-0.0473557673,0.0465940349,0.0486276001,-0.3642242849,-0.0435181372,0.1776145101,-0.1560722739,-0.2553372085,-0.1273898184,0.17875956,0.1793888956,-0.1981143653,0.2631115615,0.0087835463,-0.1398376226,-0.3518662453,0.1338686347,0.0501965582,0.177921772,0.1576354504,0.0610057227,0.3895047605,0.0040854425,-0.2422718108,0.0258157793,0.0002693687,-0.0998139083,0.1413407028,0.1294570565,0.3691285551,-0.4097653031,-0.3827797174,-0.4327935874,0.2897745073,-0.0637847111,-0.1197685599,-0.0350803994,-0.0376015715,0.2526537478,0.0688290522,0.0636018813,0.1586660147,-0.0656890944,-0.0003347614,-0.4509042203,-0.2615719438,0.459885478,-0.4583003223,-0.3856023252,-0.086100094,-0.0950353891,-0.1459356546,-0.1394127011,-0.7795305252,0.0910593942,0.2944117486,0.0955306068,-0.4224690795,0.0314957574,-0.1058145389,0.1201661602,-0.1256255955,0.3365498185,-0.0900946259,-0.432505995,0.1651308239,-0.2077717483]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1990","title":"OSError: Memory mapping file failed: Cannot allocate memory","comments":"I noticed that the error happens when loading the validation dataset.\r\nWhat value of `data_args.validation_split_percentage` did you use ?","body":"Hi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n","comment_length":19,"text":"OSError: Memory mapping file failed: Cannot allocate memory\nHi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n\nI noticed that the error happens when loading the validation dataset.\r\nWhat value of `data_args.validation_split_percentage` did you use ?","embeddings":[-0.2615419328,-0.0372641645,0.0520401485,0.6045719385,0.4538374543,0.2834013402,0.1456435472,0.2724665403,0.1785508692,0.1056612283,-0.0581841916,0.2273861468,-0.1771185547,-0.1459168196,-0.0370903499,-0.1933317184,0.084020853,0.0134716695,-0.5429398417,0.1604688615,-0.35815534,0.1157062352,-0.185632959,-0.1171597093,-0.2980377078,-0.1035162807,-0.0006043313,-0.0976312384,0.0188779235,-0.3274089992,0.2548563778,-0.1293895692,0.0864005461,0.5501937866,-0.0001231607,-0.0137761598,0.2835690081,-0.1672083884,-0.2533998191,-0.1284161955,-0.0416564196,-0.0297981855,0.0984970108,-0.2252508551,-0.0119242286,-0.052592624,0.2317306399,-0.34451437,0.4429748058,0.2243882269,0.1404946446,0.0699657425,0.4524520934,-0.0441127867,0.173906967,0.3341218829,0.0713807344,0.3320225477,-0.2908716202,-0.4356866479,-0.0304129403,0.3930506408,-0.039941147,0.0052876757,0.5653295517,-0.1445832551,0.0253530722,-0.3368177116,0.1123084202,-0.0072496962,0.5341086388,-0.490260005,-0.0283727888,-0.0147870975,-0.0533835031,-0.0952400044,0.2970977426,0.2689523995,-0.2722630501,-0.1447603852,-0.1304171085,-0.3249985576,-0.3197005689,0.4428686202,-0.0909217522,0.1910010129,-0.0247160532,0.3037818074,0.5653706789,-0.2050897777,-0.2340620607,-0.0758121908,0.1365436316,0.3287970722,-0.2677206099,-0.1660289764,-0.3029488623,-0.2047091722,0.3511367142,-0.4987058938,-0.3861933053,-0.1260007769,0.151669845,0.0212391727,0.4036456943,0.2817353308,-0.2963656187,0.3228372633,0.3122969866,0.1742716432,-0.2495381683,-0.1589899361,0.0767950714,-0.0373786837,-0.1110499874,-0.1914832443,0.0837740824,-0.0016539318,-0.0137460371,-0.0218034852,-0.2128316164,-0.1232094243,-0.0538497157,0.5055233836,-0.1040841639,-0.0846191123,0.373827666,0.1444597691,-0.0126846684,-0.0080771344,-0.0563751087,0.4104228318,-0.3436276913,0.284306854,0.0502635874,-0.0011586546,0.3263411224,-0.0727059469,-0.0331281535,-0.0566817522,0.1620447189,-0.3580230474,0.0087110661,0.1846879423,0.1788559556,0.2493289113,0.2503515184,-0.1479860991,-0.1526842713,0.1969967932,-0.163468048,-0.2433037609,0.038736742,0.0109290453,0.0980136842,0.2277469039,-0.277513355,0.1828102767,0.6528620124,-0.0775153115,-0.0572649427,-0.0801966265,-0.2296683937,-0.1687685549,0.2708779573,0.519313395,-0.0646530613,-0.1054837331,-0.1169700846,0.1677043438,0.2510127127,0.4413834214,-0.1638228297,0.2519159317,-0.097221449,0.0167985056,0.4141839147,-0.3105092943,-0.341904074,-0.005623993,-0.0252293199,-0.134322226,0.0232777502,0.1321309656,0.0774000436,0.1156655923,0.1119258851,0.3104183376,0.0561037064,0.2800839543,-0.3041983247,-0.2975762486,0.2528690994,0.0747464374,0.0653967336,-0.204082936,-0.0608494729,0.810767591,0.2972948551,-0.2623122931,0.1555345356,0.3147752583,0.1061372161,0.0064363666,0.0871187598,-0.2199127823,-0.2514474392,-0.0734247714,-0.0566988736,0.3305937052,-0.1451888531,-0.0414808579,0.1537224948,-0.1186705902,-0.1722497493,-0.3908631504,0.0622953624,0.0130732879,0.1235822812,0.0900844857,0.081060715,0.0560085624,-0.0939699262,0.2478925139,-0.5998023152,0.171503067,-0.2896074653,-0.107694298,-0.0214403253,-0.0624916069,0.0274150521,-0.0474654585,-0.068555817,0.1519924104,0.0771347657,-0.1671152562,-0.0906850323,0.0034200561,0.3101095259,-0.3221141994,0.1657055467,0.2421721667,0.1955043375,-0.0940227509,-0.1976960003,-0.1558783054,0.1051341444,0.3647213876,0.0791877508,0.1687752157,0.0262405463,0.0895476341,0.1797861308,-0.2070017755,0.218163535,0.01235501,0.2408843338,0.095006451,-0.0293943677,-0.3097331524,0.4583239555,0.2681971192,0.2772995234,0.2424818277,-0.4937297702,-0.0740422904,-0.065651603,-0.0718549192,0.3221566379,0.0500783995,-0.1195640638,0.1201550812,0.196337834,0.0097032078,0.3097595274,0.1743885428,0.4940783083,-0.0187082645,0.1057226509,-0.0806349367,-0.099587217,-0.1468482018,0.0500457101,0.5077639222,-0.175380379,-0.0157852471,-0.1864265651,-0.4006144702,-0.2753976882,0.164314881,-0.4642424881,-0.1369274259,-0.3849302828,0.3247582912,0.0695629567,0.2584396601,0.4160758257,-0.0507141761,0.3513425589,-0.1406714618,0.0687720701,-0.230412662,-0.1190755963,-0.0627371892,0.4409618676,-0.2204726189,0.0868323967,0.1567287594,-0.3195576966,-0.2135092765,-0.1457150131,0.1469898671,-0.0808212906,0.1255745143,0.0596423596,0.5084351897,-0.1233984753,-0.2323765159,0.1721098423,0.089907065,-0.0588723905,-0.0373000093,0.0232937448,0.1553850919,0.0540889502,-0.2931139171,-0.179757297,-0.508184433,0.4014800787,-0.033890903,0.1513629258,0.3279185891,0.1617626399,0.1156149805,-0.1213662103,0.0894483551,-0.1621506959,0.0160200186,0.3207914829,-0.1488257647,-0.2114350796,-0.0290636532,0.0627378225,0.306303829,0.2301850766,-0.6157578826,-0.0209107678,-0.1136818826,-0.0108944643,-0.0588507131,0.286311388,0.3913796842,0.078921251,0.107800208,0.1375889033,-0.1677574217,0.094906956,-0.0409829542,0.3248968422,0.2200790793,0.5234260559,0.0637268201,0.8384488225,0.341964066,0.1871405095,0.2288259119,-0.0220659021,0.0835645646,0.0049702087,-0.3481871486,-0.0082089063,-0.0368972011,0.0619993955,0.1661427468,0.0087548671,-0.3374429345,-0.1676723063,-0.3505290449,0.0459905639,-0.3771670461,0.2698533237,0.1751251519,0.3674235642,-0.0838355944,-0.0824133605,0.0231614523,-0.3515779972,0.1945331246,0.2135322541,-0.0175284855,0.0211241636,-0.1137084141,-0.3958891332,-0.6405459046,0.1182040572,-0.1544018984,0.0629801229,-0.1038483456,0.0933976695,0.0600519739,0.1583072543,0.7553266883,0.0059234039,-0.3072926104,-0.0461269431,-0.2609421611,-0.5063745379,0.2137841731,-0.1841569394,0.3441569507,0.143389374,0.5205535889,-0.3616023064,-0.0981338844,0.321287632,0.374714762,-0.2170827985,-0.2046922445,-0.1343318671,-0.2551260591,-0.6043947935,0.0041025551,0.0545691848,0.1934488118,0.4964534044,0.2044566721,0.0526693799,-0.0508972369,0.1371936649,-0.0533758029,0.3105924428,0.110654138,0.1008514464,0.1345039904,-0.1580678225,0.3764185309,0.2939079404,0.0716900676,-0.379755497,-0.0997854844,0.0576216578,0.2523964047,-0.0627555698,-0.0139149139,0.0038341719,0.0268995017,0.1780955493,-0.0997749195,0.1322197765,0.1875683814,0.0874076858,-0.4511457086,-0.4366309047,0.3414416313,0.1997043192,0.1596941948,0.3301558793,-0.1424441189,-0.5566364527,0.3253412545,0.3589582145,0.960832715,-0.3551480174,0.3937279582,0.0962183997,0.2117647976,0.6567651629,-0.4751399457,0.3529615998,-0.3699406683,-0.0815761462,0.0431752093,-0.1330996603,0.1379617155,-0.0172327515,-0.4360865951,0.2891325355,0.0913326368,-0.0398276299,-0.1840585619,0.3883704841,0.1124473214,-0.4811996818,-0.1841866523,0.0319509618,-0.1269843727,0.243967101,-0.0456034765,-0.0274401437,0.115675576,-0.1150276288,-0.3924042284,-0.0374513902,-0.4836362898,0.3076913059,-0.2517828047,0.0015952894,0.4047460854,0.3928067088,0.0101812379,0.3698486388,-0.1340497881,0.0105267046,-0.3600815237,-0.2767649889,-0.185056895,0.0676643252,0.2424043119,-0.1937424839,-0.1071275026,0.0763597563,-0.0399471037,-0.1982413977,-0.2103530616,-0.0318763927,-0.1298194379,-0.2045654804,-0.2123933136,0.0132841198,-0.3584757745,-0.0186204016,0.0458632261,0.0108470637,-0.3378366232,0.3286833465,0.1423627734,-0.4008429646,0.0307404082,0.4164584875,0.348543793,0.1605854779,0.7200255394,0.3232616484,-0.2456386834,-0.1618594825,-0.0824732929,-0.025335988,-0.304834336,0.2100929171,0.083754994,0.106053941,-0.3066006005,-0.0003773624,0.1285434067,0.0452359319,0.0363892019,-0.3043982387,-0.5602001548,0.2082497925,-0.010477718,0.036499247,0.236894533,0.0221908819,0.0295315944,0.2480150908,-0.156630531,0.020054128,-0.1691422611,0.2507917583,0.2438191026,-0.1116867065,0.1032438129,-0.0129493577,-0.0043390617,0.1593108326,-0.0150237987,-0.1269526333,0.0454815626,0.1532624811,-0.026851235,-0.2777851522,-0.0963580236,-0.5430374146,-0.0158967581,-0.3039938509,0.0111275995,0.0212793741,-0.1362702698,0.110977523,0.2019030601,-0.2731659412,-0.063025862,0.2792618573,-0.0221649799,-0.050752569,0.1658708602,0.2017495036,0.0937991664,-0.1570737511,-0.5362299681,0.055834543,-0.0843362063,-0.0266278703,0.1019938439,-0.1142851859,-0.3993756771,0.3194217682,0.0904670879,0.1510138661,-0.1423276961,0.20684129,0.2004518658,0.0089688553,-0.1937268674,-0.0418563969,0.3301984668,-0.2920083702,-0.0505978391,0.3277898431,-0.1311677396,0.0453453697,-0.0299580656,0.1311582178,0.3588027954,-0.0734308213,0.3226182461,0.0061582224,-0.0834842399,-0.0167209767,0.1939019412,0.4057275653,0.2587025166,0.204386279,-0.1556384861,0.0376391597,-0.1058716699,-0.2548227906,0.0379435755,-0.0836717263,0.1414890438,0.4189734757,-0.1775912941,0.1592200398,-0.193913281,0.1780178845,0.1845969856,-0.2745580077,0.078776978,0.2984717786,-0.2116876245,0.0464751199,0.1499889493,-0.115781568,-0.0657879114,0.141785726,0.0771138668,-0.6087394357,0.4591955841,0.1467865705,-0.2206508368,-0.1409267187,0.2773652971,0.4522538185,0.0710785612,-0.3053991199,0.0575940534,0.0269881077,0.136588797,-0.3318274915,0.309446007,0.4621617496,0.2915849388,-0.1956237853,0.2535728216,-0.1581641287,-0.012619175,0.0154259196,-0.0236886721,0.116465725,0.2165056765,0.1130549312,-0.0109328497,-0.0586273782,-0.151095286,0.1630205214,0.1425201148,-0.227916345,-0.3811423779,-0.0991922468,-0.171089232,-0.0140399672,-0.0115701789,-0.5901709795,0.2826703489,0.5286229849,-0.1380668133,0.081210427,0.0202327408,0.0137296338,-0.031966269,0.4846240878,0.2030230612,-0.0330804884,-0.481416434,-0.3177812397,-0.4150764942,0.0241078436,-0.3297815621,0.0040629571,-0.075407438,0.1153876483,-0.0473557673,0.0465940349,0.0486276001,-0.3642242849,-0.0435181372,0.1776145101,-0.1560722739,-0.2553372085,-0.1273898184,0.17875956,0.1793888956,-0.1981143653,0.2631115615,0.0087835463,-0.1398376226,-0.3518662453,0.1338686347,0.0501965582,0.177921772,0.1576354504,0.0610057227,0.3895047605,0.0040854425,-0.2422718108,0.0258157793,0.0002693687,-0.0998139083,0.1413407028,0.1294570565,0.3691285551,-0.4097653031,-0.3827797174,-0.4327935874,0.2897745073,-0.0637847111,-0.1197685599,-0.0350803994,-0.0376015715,0.2526537478,0.0688290522,0.0636018813,0.1586660147,-0.0656890944,-0.0003347614,-0.4509042203,-0.2615719438,0.459885478,-0.4583003223,-0.3856023252,-0.086100094,-0.0950353891,-0.1459356546,-0.1394127011,-0.7795305252,0.0910593942,0.2944117486,0.0955306068,-0.4224690795,0.0314957574,-0.1058145389,0.1201661602,-0.1256255955,0.3365498185,-0.0900946259,-0.432505995,0.1651308239,-0.2077717483]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1990","title":"OSError: Memory mapping file failed: Cannot allocate memory","comments":"Dear @lhoestq \r\n\r\nthank you very much for the very sharp observation, indeed, this happens there, I use the default value of 5, I basically plan to subsample a part of the large dataset and choose it as validation set. Do you think this is bringing the data into memory during subsampling? Is there a way I could avoid this?\r\n\r\nThank you very much for the great help.\r\n\r\n\r\nOn Mon, Mar 8, 2021 at 11:28 AM Quentin Lhoest ***@***.***>\r\nwrote:\r\n\r\n> I noticed that the error happens when loading the validation dataset.\r\n> What value of data_args.validation_split_percentage did you use ?\r\n>\r\n> \u2014\r\n> You are receiving this because you authored the thread.\r\n> Reply to this email directly, view it on GitHub\r\n> ,\r\n> or unsubscribe\r\n> \r\n> .\r\n>\r\n","body":"Hi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n","comment_length":133,"text":"OSError: Memory mapping file failed: Cannot allocate memory\nHi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n\nDear @lhoestq \r\n\r\nthank you very much for the very sharp observation, indeed, this happens there, I use the default value of 5, I basically plan to subsample a part of the large dataset and choose it as validation set. Do you think this is bringing the data into memory during subsampling? Is there a way I could avoid this?\r\n\r\nThank you very much for the great help.\r\n\r\n\r\nOn Mon, Mar 8, 2021 at 11:28 AM Quentin Lhoest ***@***.***>\r\nwrote:\r\n\r\n> I noticed that the error happens when loading the validation dataset.\r\n> What value of data_args.validation_split_percentage did you use ?\r\n>\r\n> \u2014\r\n> You are receiving this because you authored the thread.\r\n> Reply to this email directly, view it on GitHub\r\n> ,\r\n> or unsubscribe\r\n> \r\n> .\r\n>\r\n","embeddings":[-0.2615419328,-0.0372641645,0.0520401485,0.6045719385,0.4538374543,0.2834013402,0.1456435472,0.2724665403,0.1785508692,0.1056612283,-0.0581841916,0.2273861468,-0.1771185547,-0.1459168196,-0.0370903499,-0.1933317184,0.084020853,0.0134716695,-0.5429398417,0.1604688615,-0.35815534,0.1157062352,-0.185632959,-0.1171597093,-0.2980377078,-0.1035162807,-0.0006043313,-0.0976312384,0.0188779235,-0.3274089992,0.2548563778,-0.1293895692,0.0864005461,0.5501937866,-0.0001231607,-0.0137761598,0.2835690081,-0.1672083884,-0.2533998191,-0.1284161955,-0.0416564196,-0.0297981855,0.0984970108,-0.2252508551,-0.0119242286,-0.052592624,0.2317306399,-0.34451437,0.4429748058,0.2243882269,0.1404946446,0.0699657425,0.4524520934,-0.0441127867,0.173906967,0.3341218829,0.0713807344,0.3320225477,-0.2908716202,-0.4356866479,-0.0304129403,0.3930506408,-0.039941147,0.0052876757,0.5653295517,-0.1445832551,0.0253530722,-0.3368177116,0.1123084202,-0.0072496962,0.5341086388,-0.490260005,-0.0283727888,-0.0147870975,-0.0533835031,-0.0952400044,0.2970977426,0.2689523995,-0.2722630501,-0.1447603852,-0.1304171085,-0.3249985576,-0.3197005689,0.4428686202,-0.0909217522,0.1910010129,-0.0247160532,0.3037818074,0.5653706789,-0.2050897777,-0.2340620607,-0.0758121908,0.1365436316,0.3287970722,-0.2677206099,-0.1660289764,-0.3029488623,-0.2047091722,0.3511367142,-0.4987058938,-0.3861933053,-0.1260007769,0.151669845,0.0212391727,0.4036456943,0.2817353308,-0.2963656187,0.3228372633,0.3122969866,0.1742716432,-0.2495381683,-0.1589899361,0.0767950714,-0.0373786837,-0.1110499874,-0.1914832443,0.0837740824,-0.0016539318,-0.0137460371,-0.0218034852,-0.2128316164,-0.1232094243,-0.0538497157,0.5055233836,-0.1040841639,-0.0846191123,0.373827666,0.1444597691,-0.0126846684,-0.0080771344,-0.0563751087,0.4104228318,-0.3436276913,0.284306854,0.0502635874,-0.0011586546,0.3263411224,-0.0727059469,-0.0331281535,-0.0566817522,0.1620447189,-0.3580230474,0.0087110661,0.1846879423,0.1788559556,0.2493289113,0.2503515184,-0.1479860991,-0.1526842713,0.1969967932,-0.163468048,-0.2433037609,0.038736742,0.0109290453,0.0980136842,0.2277469039,-0.277513355,0.1828102767,0.6528620124,-0.0775153115,-0.0572649427,-0.0801966265,-0.2296683937,-0.1687685549,0.2708779573,0.519313395,-0.0646530613,-0.1054837331,-0.1169700846,0.1677043438,0.2510127127,0.4413834214,-0.1638228297,0.2519159317,-0.097221449,0.0167985056,0.4141839147,-0.3105092943,-0.341904074,-0.005623993,-0.0252293199,-0.134322226,0.0232777502,0.1321309656,0.0774000436,0.1156655923,0.1119258851,0.3104183376,0.0561037064,0.2800839543,-0.3041983247,-0.2975762486,0.2528690994,0.0747464374,0.0653967336,-0.204082936,-0.0608494729,0.810767591,0.2972948551,-0.2623122931,0.1555345356,0.3147752583,0.1061372161,0.0064363666,0.0871187598,-0.2199127823,-0.2514474392,-0.0734247714,-0.0566988736,0.3305937052,-0.1451888531,-0.0414808579,0.1537224948,-0.1186705902,-0.1722497493,-0.3908631504,0.0622953624,0.0130732879,0.1235822812,0.0900844857,0.081060715,0.0560085624,-0.0939699262,0.2478925139,-0.5998023152,0.171503067,-0.2896074653,-0.107694298,-0.0214403253,-0.0624916069,0.0274150521,-0.0474654585,-0.068555817,0.1519924104,0.0771347657,-0.1671152562,-0.0906850323,0.0034200561,0.3101095259,-0.3221141994,0.1657055467,0.2421721667,0.1955043375,-0.0940227509,-0.1976960003,-0.1558783054,0.1051341444,0.3647213876,0.0791877508,0.1687752157,0.0262405463,0.0895476341,0.1797861308,-0.2070017755,0.218163535,0.01235501,0.2408843338,0.095006451,-0.0293943677,-0.3097331524,0.4583239555,0.2681971192,0.2772995234,0.2424818277,-0.4937297702,-0.0740422904,-0.065651603,-0.0718549192,0.3221566379,0.0500783995,-0.1195640638,0.1201550812,0.196337834,0.0097032078,0.3097595274,0.1743885428,0.4940783083,-0.0187082645,0.1057226509,-0.0806349367,-0.099587217,-0.1468482018,0.0500457101,0.5077639222,-0.175380379,-0.0157852471,-0.1864265651,-0.4006144702,-0.2753976882,0.164314881,-0.4642424881,-0.1369274259,-0.3849302828,0.3247582912,0.0695629567,0.2584396601,0.4160758257,-0.0507141761,0.3513425589,-0.1406714618,0.0687720701,-0.230412662,-0.1190755963,-0.0627371892,0.4409618676,-0.2204726189,0.0868323967,0.1567287594,-0.3195576966,-0.2135092765,-0.1457150131,0.1469898671,-0.0808212906,0.1255745143,0.0596423596,0.5084351897,-0.1233984753,-0.2323765159,0.1721098423,0.089907065,-0.0588723905,-0.0373000093,0.0232937448,0.1553850919,0.0540889502,-0.2931139171,-0.179757297,-0.508184433,0.4014800787,-0.033890903,0.1513629258,0.3279185891,0.1617626399,0.1156149805,-0.1213662103,0.0894483551,-0.1621506959,0.0160200186,0.3207914829,-0.1488257647,-0.2114350796,-0.0290636532,0.0627378225,0.306303829,0.2301850766,-0.6157578826,-0.0209107678,-0.1136818826,-0.0108944643,-0.0588507131,0.286311388,0.3913796842,0.078921251,0.107800208,0.1375889033,-0.1677574217,0.094906956,-0.0409829542,0.3248968422,0.2200790793,0.5234260559,0.0637268201,0.8384488225,0.341964066,0.1871405095,0.2288259119,-0.0220659021,0.0835645646,0.0049702087,-0.3481871486,-0.0082089063,-0.0368972011,0.0619993955,0.1661427468,0.0087548671,-0.3374429345,-0.1676723063,-0.3505290449,0.0459905639,-0.3771670461,0.2698533237,0.1751251519,0.3674235642,-0.0838355944,-0.0824133605,0.0231614523,-0.3515779972,0.1945331246,0.2135322541,-0.0175284855,0.0211241636,-0.1137084141,-0.3958891332,-0.6405459046,0.1182040572,-0.1544018984,0.0629801229,-0.1038483456,0.0933976695,0.0600519739,0.1583072543,0.7553266883,0.0059234039,-0.3072926104,-0.0461269431,-0.2609421611,-0.5063745379,0.2137841731,-0.1841569394,0.3441569507,0.143389374,0.5205535889,-0.3616023064,-0.0981338844,0.321287632,0.374714762,-0.2170827985,-0.2046922445,-0.1343318671,-0.2551260591,-0.6043947935,0.0041025551,0.0545691848,0.1934488118,0.4964534044,0.2044566721,0.0526693799,-0.0508972369,0.1371936649,-0.0533758029,0.3105924428,0.110654138,0.1008514464,0.1345039904,-0.1580678225,0.3764185309,0.2939079404,0.0716900676,-0.379755497,-0.0997854844,0.0576216578,0.2523964047,-0.0627555698,-0.0139149139,0.0038341719,0.0268995017,0.1780955493,-0.0997749195,0.1322197765,0.1875683814,0.0874076858,-0.4511457086,-0.4366309047,0.3414416313,0.1997043192,0.1596941948,0.3301558793,-0.1424441189,-0.5566364527,0.3253412545,0.3589582145,0.960832715,-0.3551480174,0.3937279582,0.0962183997,0.2117647976,0.6567651629,-0.4751399457,0.3529615998,-0.3699406683,-0.0815761462,0.0431752093,-0.1330996603,0.1379617155,-0.0172327515,-0.4360865951,0.2891325355,0.0913326368,-0.0398276299,-0.1840585619,0.3883704841,0.1124473214,-0.4811996818,-0.1841866523,0.0319509618,-0.1269843727,0.243967101,-0.0456034765,-0.0274401437,0.115675576,-0.1150276288,-0.3924042284,-0.0374513902,-0.4836362898,0.3076913059,-0.2517828047,0.0015952894,0.4047460854,0.3928067088,0.0101812379,0.3698486388,-0.1340497881,0.0105267046,-0.3600815237,-0.2767649889,-0.185056895,0.0676643252,0.2424043119,-0.1937424839,-0.1071275026,0.0763597563,-0.0399471037,-0.1982413977,-0.2103530616,-0.0318763927,-0.1298194379,-0.2045654804,-0.2123933136,0.0132841198,-0.3584757745,-0.0186204016,0.0458632261,0.0108470637,-0.3378366232,0.3286833465,0.1423627734,-0.4008429646,0.0307404082,0.4164584875,0.348543793,0.1605854779,0.7200255394,0.3232616484,-0.2456386834,-0.1618594825,-0.0824732929,-0.025335988,-0.304834336,0.2100929171,0.083754994,0.106053941,-0.3066006005,-0.0003773624,0.1285434067,0.0452359319,0.0363892019,-0.3043982387,-0.5602001548,0.2082497925,-0.010477718,0.036499247,0.236894533,0.0221908819,0.0295315944,0.2480150908,-0.156630531,0.020054128,-0.1691422611,0.2507917583,0.2438191026,-0.1116867065,0.1032438129,-0.0129493577,-0.0043390617,0.1593108326,-0.0150237987,-0.1269526333,0.0454815626,0.1532624811,-0.026851235,-0.2777851522,-0.0963580236,-0.5430374146,-0.0158967581,-0.3039938509,0.0111275995,0.0212793741,-0.1362702698,0.110977523,0.2019030601,-0.2731659412,-0.063025862,0.2792618573,-0.0221649799,-0.050752569,0.1658708602,0.2017495036,0.0937991664,-0.1570737511,-0.5362299681,0.055834543,-0.0843362063,-0.0266278703,0.1019938439,-0.1142851859,-0.3993756771,0.3194217682,0.0904670879,0.1510138661,-0.1423276961,0.20684129,0.2004518658,0.0089688553,-0.1937268674,-0.0418563969,0.3301984668,-0.2920083702,-0.0505978391,0.3277898431,-0.1311677396,0.0453453697,-0.0299580656,0.1311582178,0.3588027954,-0.0734308213,0.3226182461,0.0061582224,-0.0834842399,-0.0167209767,0.1939019412,0.4057275653,0.2587025166,0.204386279,-0.1556384861,0.0376391597,-0.1058716699,-0.2548227906,0.0379435755,-0.0836717263,0.1414890438,0.4189734757,-0.1775912941,0.1592200398,-0.193913281,0.1780178845,0.1845969856,-0.2745580077,0.078776978,0.2984717786,-0.2116876245,0.0464751199,0.1499889493,-0.115781568,-0.0657879114,0.141785726,0.0771138668,-0.6087394357,0.4591955841,0.1467865705,-0.2206508368,-0.1409267187,0.2773652971,0.4522538185,0.0710785612,-0.3053991199,0.0575940534,0.0269881077,0.136588797,-0.3318274915,0.309446007,0.4621617496,0.2915849388,-0.1956237853,0.2535728216,-0.1581641287,-0.012619175,0.0154259196,-0.0236886721,0.116465725,0.2165056765,0.1130549312,-0.0109328497,-0.0586273782,-0.151095286,0.1630205214,0.1425201148,-0.227916345,-0.3811423779,-0.0991922468,-0.171089232,-0.0140399672,-0.0115701789,-0.5901709795,0.2826703489,0.5286229849,-0.1380668133,0.081210427,0.0202327408,0.0137296338,-0.031966269,0.4846240878,0.2030230612,-0.0330804884,-0.481416434,-0.3177812397,-0.4150764942,0.0241078436,-0.3297815621,0.0040629571,-0.075407438,0.1153876483,-0.0473557673,0.0465940349,0.0486276001,-0.3642242849,-0.0435181372,0.1776145101,-0.1560722739,-0.2553372085,-0.1273898184,0.17875956,0.1793888956,-0.1981143653,0.2631115615,0.0087835463,-0.1398376226,-0.3518662453,0.1338686347,0.0501965582,0.177921772,0.1576354504,0.0610057227,0.3895047605,0.0040854425,-0.2422718108,0.0258157793,0.0002693687,-0.0998139083,0.1413407028,0.1294570565,0.3691285551,-0.4097653031,-0.3827797174,-0.4327935874,0.2897745073,-0.0637847111,-0.1197685599,-0.0350803994,-0.0376015715,0.2526537478,0.0688290522,0.0636018813,0.1586660147,-0.0656890944,-0.0003347614,-0.4509042203,-0.2615719438,0.459885478,-0.4583003223,-0.3856023252,-0.086100094,-0.0950353891,-0.1459356546,-0.1394127011,-0.7795305252,0.0910593942,0.2944117486,0.0955306068,-0.4224690795,0.0314957574,-0.1058145389,0.1201661602,-0.1256255955,0.3365498185,-0.0900946259,-0.432505995,0.1651308239,-0.2077717483]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1990","title":"OSError: Memory mapping file failed: Cannot allocate memory","comments":"Methods like `dataset.shard`, `dataset.train_test_split`, `dataset.select` etc. don't bring the dataset in memory. \r\nThe only time when samples are brought to memory is when you access elements via `dataset[0]`, `dataset[:10]`, `dataset[\"my_column_names\"]`.\r\n\r\nBut it's possible that trying to use those methods to build your validation set doesn't fix the issue since, if I understand correctly, the error happens when when the dataset arrow file is opened (just before the 5% percentage is applied).\r\n\r\nDid you try to reproduce this issue in a google colab ? This would be super helpful to investigate why this happened.\r\n\r\nAlso maybe you can try clearing your cache at `~\/.cache\/huggingface\/datasets` and try again. If the arrow file was corrupted somehow, removing it and rebuilding may fix the issue.","body":"Hi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n","comment_length":121,"text":"OSError: Memory mapping file failed: Cannot allocate memory\nHi,\r\nI am trying to run a code with a wikipedia dataset, here is the command to reproduce the error. You can find the codes for run_mlm.py in huggingface repo here: https:\/\/github.com\/huggingface\/transformers\/blob\/v4.3.2\/examples\/language-modeling\/run_mlm.py \r\n```\r\npython run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.en --do_train --do_eval --output_dir \/dara\/test --max_seq_length 128\r\n```\r\n\r\nI am using transformer version: 4.3.2 \r\n\r\nBut I got memory erorr using this dataset, is there a way I could save on memory with dataset library with wikipedia dataset?\r\nSpecially I need to train a model with multiple of wikipedia datasets concatenated. thank you very much @lhoestq for your help and suggestions:\r\n\r\n```\r\n File \"run_mlm.py\", line 441, in \r\n main()\r\n File \"run_mlm.py\", line 233, in main\r\n split=f\"train[{data_args.validation_split_percentage}%:]\",\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 740, in as_dataset\r\n map_tuple=True,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 757, in _build_single_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 829, in _as_dataset\r\n in_memory=in_memory,\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 215, in read\r\n return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\n\r\n\r\n\nMethods like `dataset.shard`, `dataset.train_test_split`, `dataset.select` etc. don't bring the dataset in memory. \r\nThe only time when samples are brought to memory is when you access elements via `dataset[0]`, `dataset[:10]`, `dataset[\"my_column_names\"]`.\r\n\r\nBut it's possible that trying to use those methods to build your validation set doesn't fix the issue since, if I understand correctly, the error happens when when the dataset arrow file is opened (just before the 5% percentage is applied).\r\n\r\nDid you try to reproduce this issue in a google colab ? This would be super helpful to investigate why this happened.\r\n\r\nAlso maybe you can try clearing your cache at `~\/.cache\/huggingface\/datasets` and try again. If the arrow file was corrupted somehow, removing it and rebuilding may fix the issue.","embeddings":[-0.2615419328,-0.0372641645,0.0520401485,0.6045719385,0.4538374543,0.2834013402,0.1456435472,0.2724665403,0.1785508692,0.1056612283,-0.0581841916,0.2273861468,-0.1771185547,-0.1459168196,-0.0370903499,-0.1933317184,0.084020853,0.0134716695,-0.5429398417,0.1604688615,-0.35815534,0.1157062352,-0.185632959,-0.1171597093,-0.2980377078,-0.1035162807,-0.0006043313,-0.0976312384,0.0188779235,-0.3274089992,0.2548563778,-0.1293895692,0.0864005461,0.5501937866,-0.0001231607,-0.0137761598,0.2835690081,-0.1672083884,-0.2533998191,-0.1284161955,-0.0416564196,-0.0297981855,0.0984970108,-0.2252508551,-0.0119242286,-0.052592624,0.2317306399,-0.34451437,0.4429748058,0.2243882269,0.1404946446,0.0699657425,0.4524520934,-0.0441127867,0.173906967,0.3341218829,0.0713807344,0.3320225477,-0.2908716202,-0.4356866479,-0.0304129403,0.3930506408,-0.039941147,0.0052876757,0.5653295517,-0.1445832551,0.0253530722,-0.3368177116,0.1123084202,-0.0072496962,0.5341086388,-0.490260005,-0.0283727888,-0.0147870975,-0.0533835031,-0.0952400044,0.2970977426,0.2689523995,-0.2722630501,-0.1447603852,-0.1304171085,-0.3249985576,-0.3197005689,0.4428686202,-0.0909217522,0.1910010129,-0.0247160532,0.3037818074,0.5653706789,-0.2050897777,-0.2340620607,-0.0758121908,0.1365436316,0.3287970722,-0.2677206099,-0.1660289764,-0.3029488623,-0.2047091722,0.3511367142,-0.4987058938,-0.3861933053,-0.1260007769,0.151669845,0.0212391727,0.4036456943,0.2817353308,-0.2963656187,0.3228372633,0.3122969866,0.1742716432,-0.2495381683,-0.1589899361,0.0767950714,-0.0373786837,-0.1110499874,-0.1914832443,0.0837740824,-0.0016539318,-0.0137460371,-0.0218034852,-0.2128316164,-0.1232094243,-0.0538497157,0.5055233836,-0.1040841639,-0.0846191123,0.373827666,0.1444597691,-0.0126846684,-0.0080771344,-0.0563751087,0.4104228318,-0.3436276913,0.284306854,0.0502635874,-0.0011586546,0.3263411224,-0.0727059469,-0.0331281535,-0.0566817522,0.1620447189,-0.3580230474,0.0087110661,0.1846879423,0.1788559556,0.2493289113,0.2503515184,-0.1479860991,-0.1526842713,0.1969967932,-0.163468048,-0.2433037609,0.038736742,0.0109290453,0.0980136842,0.2277469039,-0.277513355,0.1828102767,0.6528620124,-0.0775153115,-0.0572649427,-0.0801966265,-0.2296683937,-0.1687685549,0.2708779573,0.519313395,-0.0646530613,-0.1054837331,-0.1169700846,0.1677043438,0.2510127127,0.4413834214,-0.1638228297,0.2519159317,-0.097221449,0.0167985056,0.4141839147,-0.3105092943,-0.341904074,-0.005623993,-0.0252293199,-0.134322226,0.0232777502,0.1321309656,0.0774000436,0.1156655923,0.1119258851,0.3104183376,0.0561037064,0.2800839543,-0.3041983247,-0.2975762486,0.2528690994,0.0747464374,0.0653967336,-0.204082936,-0.0608494729,0.810767591,0.2972948551,-0.2623122931,0.1555345356,0.3147752583,0.1061372161,0.0064363666,0.0871187598,-0.2199127823,-0.2514474392,-0.0734247714,-0.0566988736,0.3305937052,-0.1451888531,-0.0414808579,0.1537224948,-0.1186705902,-0.1722497493,-0.3908631504,0.0622953624,0.0130732879,0.1235822812,0.0900844857,0.081060715,0.0560085624,-0.0939699262,0.2478925139,-0.5998023152,0.171503067,-0.2896074653,-0.107694298,-0.0214403253,-0.0624916069,0.0274150521,-0.0474654585,-0.068555817,0.1519924104,0.0771347657,-0.1671152562,-0.0906850323,0.0034200561,0.3101095259,-0.3221141994,0.1657055467,0.2421721667,0.1955043375,-0.0940227509,-0.1976960003,-0.1558783054,0.1051341444,0.3647213876,0.0791877508,0.1687752157,0.0262405463,0.0895476341,0.1797861308,-0.2070017755,0.218163535,0.01235501,0.2408843338,0.095006451,-0.0293943677,-0.3097331524,0.4583239555,0.2681971192,0.2772995234,0.2424818277,-0.4937297702,-0.0740422904,-0.065651603,-0.0718549192,0.3221566379,0.0500783995,-0.1195640638,0.1201550812,0.196337834,0.0097032078,0.3097595274,0.1743885428,0.4940783083,-0.0187082645,0.1057226509,-0.0806349367,-0.099587217,-0.1468482018,0.0500457101,0.5077639222,-0.175380379,-0.0157852471,-0.1864265651,-0.4006144702,-0.2753976882,0.164314881,-0.4642424881,-0.1369274259,-0.3849302828,0.3247582912,0.0695629567,0.2584396601,0.4160758257,-0.0507141761,0.3513425589,-0.1406714618,0.0687720701,-0.230412662,-0.1190755963,-0.0627371892,0.4409618676,-0.2204726189,0.0868323967,0.1567287594,-0.3195576966,-0.2135092765,-0.1457150131,0.1469898671,-0.0808212906,0.1255745143,0.0596423596,0.5084351897,-0.1233984753,-0.2323765159,0.1721098423,0.089907065,-0.0588723905,-0.0373000093,0.0232937448,0.1553850919,0.0540889502,-0.2931139171,-0.179757297,-0.508184433,0.4014800787,-0.033890903,0.1513629258,0.3279185891,0.1617626399,0.1156149805,-0.1213662103,0.0894483551,-0.1621506959,0.0160200186,0.3207914829,-0.1488257647,-0.2114350796,-0.0290636532,0.0627378225,0.306303829,0.2301850766,-0.6157578826,-0.0209107678,-0.1136818826,-0.0108944643,-0.0588507131,0.286311388,0.3913796842,0.078921251,0.107800208,0.1375889033,-0.1677574217,0.094906956,-0.0409829542,0.3248968422,0.2200790793,0.5234260559,0.0637268201,0.8384488225,0.341964066,0.1871405095,0.2288259119,-0.0220659021,0.0835645646,0.0049702087,-0.3481871486,-0.0082089063,-0.0368972011,0.0619993955,0.1661427468,0.0087548671,-0.3374429345,-0.1676723063,-0.3505290449,0.0459905639,-0.3771670461,0.2698533237,0.1751251519,0.3674235642,-0.0838355944,-0.0824133605,0.0231614523,-0.3515779972,0.1945331246,0.2135322541,-0.0175284855,0.0211241636,-0.1137084141,-0.3958891332,-0.6405459046,0.1182040572,-0.1544018984,0.0629801229,-0.1038483456,0.0933976695,0.0600519739,0.1583072543,0.7553266883,0.0059234039,-0.3072926104,-0.0461269431,-0.2609421611,-0.5063745379,0.2137841731,-0.1841569394,0.3441569507,0.143389374,0.5205535889,-0.3616023064,-0.0981338844,0.321287632,0.374714762,-0.2170827985,-0.2046922445,-0.1343318671,-0.2551260591,-0.6043947935,0.0041025551,0.0545691848,0.1934488118,0.4964534044,0.2044566721,0.0526693799,-0.0508972369,0.1371936649,-0.0533758029,0.3105924428,0.110654138,0.1008514464,0.1345039904,-0.1580678225,0.3764185309,0.2939079404,0.0716900676,-0.379755497,-0.0997854844,0.0576216578,0.2523964047,-0.0627555698,-0.0139149139,0.0038341719,0.0268995017,0.1780955493,-0.0997749195,0.1322197765,0.1875683814,0.0874076858,-0.4511457086,-0.4366309047,0.3414416313,0.1997043192,0.1596941948,0.3301558793,-0.1424441189,-0.5566364527,0.3253412545,0.3589582145,0.960832715,-0.3551480174,0.3937279582,0.0962183997,0.2117647976,0.6567651629,-0.4751399457,0.3529615998,-0.3699406683,-0.0815761462,0.0431752093,-0.1330996603,0.1379617155,-0.0172327515,-0.4360865951,0.2891325355,0.0913326368,-0.0398276299,-0.1840585619,0.3883704841,0.1124473214,-0.4811996818,-0.1841866523,0.0319509618,-0.1269843727,0.243967101,-0.0456034765,-0.0274401437,0.115675576,-0.1150276288,-0.3924042284,-0.0374513902,-0.4836362898,0.3076913059,-0.2517828047,0.0015952894,0.4047460854,0.3928067088,0.0101812379,0.3698486388,-0.1340497881,0.0105267046,-0.3600815237,-0.2767649889,-0.185056895,0.0676643252,0.2424043119,-0.1937424839,-0.1071275026,0.0763597563,-0.0399471037,-0.1982413977,-0.2103530616,-0.0318763927,-0.1298194379,-0.2045654804,-0.2123933136,0.0132841198,-0.3584757745,-0.0186204016,0.0458632261,0.0108470637,-0.3378366232,0.3286833465,0.1423627734,-0.4008429646,0.0307404082,0.4164584875,0.348543793,0.1605854779,0.7200255394,0.3232616484,-0.2456386834,-0.1618594825,-0.0824732929,-0.025335988,-0.304834336,0.2100929171,0.083754994,0.106053941,-0.3066006005,-0.0003773624,0.1285434067,0.0452359319,0.0363892019,-0.3043982387,-0.5602001548,0.2082497925,-0.010477718,0.036499247,0.236894533,0.0221908819,0.0295315944,0.2480150908,-0.156630531,0.020054128,-0.1691422611,0.2507917583,0.2438191026,-0.1116867065,0.1032438129,-0.0129493577,-0.0043390617,0.1593108326,-0.0150237987,-0.1269526333,0.0454815626,0.1532624811,-0.026851235,-0.2777851522,-0.0963580236,-0.5430374146,-0.0158967581,-0.3039938509,0.0111275995,0.0212793741,-0.1362702698,0.110977523,0.2019030601,-0.2731659412,-0.063025862,0.2792618573,-0.0221649799,-0.050752569,0.1658708602,0.2017495036,0.0937991664,-0.1570737511,-0.5362299681,0.055834543,-0.0843362063,-0.0266278703,0.1019938439,-0.1142851859,-0.3993756771,0.3194217682,0.0904670879,0.1510138661,-0.1423276961,0.20684129,0.2004518658,0.0089688553,-0.1937268674,-0.0418563969,0.3301984668,-0.2920083702,-0.0505978391,0.3277898431,-0.1311677396,0.0453453697,-0.0299580656,0.1311582178,0.3588027954,-0.0734308213,0.3226182461,0.0061582224,-0.0834842399,-0.0167209767,0.1939019412,0.4057275653,0.2587025166,0.204386279,-0.1556384861,0.0376391597,-0.1058716699,-0.2548227906,0.0379435755,-0.0836717263,0.1414890438,0.4189734757,-0.1775912941,0.1592200398,-0.193913281,0.1780178845,0.1845969856,-0.2745580077,0.078776978,0.2984717786,-0.2116876245,0.0464751199,0.1499889493,-0.115781568,-0.0657879114,0.141785726,0.0771138668,-0.6087394357,0.4591955841,0.1467865705,-0.2206508368,-0.1409267187,0.2773652971,0.4522538185,0.0710785612,-0.3053991199,0.0575940534,0.0269881077,0.136588797,-0.3318274915,0.309446007,0.4621617496,0.2915849388,-0.1956237853,0.2535728216,-0.1581641287,-0.012619175,0.0154259196,-0.0236886721,0.116465725,0.2165056765,0.1130549312,-0.0109328497,-0.0586273782,-0.151095286,0.1630205214,0.1425201148,-0.227916345,-0.3811423779,-0.0991922468,-0.171089232,-0.0140399672,-0.0115701789,-0.5901709795,0.2826703489,0.5286229849,-0.1380668133,0.081210427,0.0202327408,0.0137296338,-0.031966269,0.4846240878,0.2030230612,-0.0330804884,-0.481416434,-0.3177812397,-0.4150764942,0.0241078436,-0.3297815621,0.0040629571,-0.075407438,0.1153876483,-0.0473557673,0.0465940349,0.0486276001,-0.3642242849,-0.0435181372,0.1776145101,-0.1560722739,-0.2553372085,-0.1273898184,0.17875956,0.1793888956,-0.1981143653,0.2631115615,0.0087835463,-0.1398376226,-0.3518662453,0.1338686347,0.0501965582,0.177921772,0.1576354504,0.0610057227,0.3895047605,0.0040854425,-0.2422718108,0.0258157793,0.0002693687,-0.0998139083,0.1413407028,0.1294570565,0.3691285551,-0.4097653031,-0.3827797174,-0.4327935874,0.2897745073,-0.0637847111,-0.1197685599,-0.0350803994,-0.0376015715,0.2526537478,0.0688290522,0.0636018813,0.1586660147,-0.0656890944,-0.0003347614,-0.4509042203,-0.2615719438,0.459885478,-0.4583003223,-0.3856023252,-0.086100094,-0.0950353891,-0.1459356546,-0.1394127011,-0.7795305252,0.0910593942,0.2944117486,0.0955306068,-0.4224690795,0.0314957574,-0.1058145389,0.1201661602,-0.1256255955,0.3365498185,-0.0900946259,-0.432505995,0.1651308239,-0.2077717483]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"It seems that I get parsing errors for various fields in my data. For example now I get this:\r\n```\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: https:\/\/www.netgalley.com\/catalog\/book\/121872\r\n```","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":128,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nIt seems that I get parsing errors for various fields in my data. For example now I get this:\r\n```\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: https:\/\/www.netgalley.com\/catalog\/book\/121872\r\n```","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"Not sure if this helps, this is how I load my files (as in the sample scripts on transformers):\r\n\r\n```\r\n if data_args.train_file.endswith(\".csv\"):\r\n # Loading a dataset from local csv files\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n```","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":35,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nNot sure if this helps, this is how I load my files (as in the sample scripts on transformers):\r\n\r\n```\r\n if data_args.train_file.endswith(\".csv\"):\r\n # Loading a dataset from local csv files\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n```","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"Since this worked out of the box in a few examples before, I wonder if it's some quoting issue or something else. ","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":22,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nSince this worked out of the box in a few examples before, I wonder if it's some quoting issue or something else. ","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"Hi @ioana-blue,\r\nCan you share a sample from your .csv? A dummy where you get this error will also help.\r\n\r\nI tried this csv:\r\n```csv\r\nfeature,label\r\n1.2,not nurse\r\n1.3,nurse\r\n1.5,surgeon\r\n```\r\nand the following snippet:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nd = load_dataset(\"csv\",data_files=['test.csv'])\r\n\r\nprint(d)\r\nprint(d['train']['label'])\r\n```\r\nand this works perfectly fine for me:\r\n```sh\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['feature', 'label'],\r\n num_rows: 3\r\n })\r\n})\r\n['not nurse', 'nurse', 'surgeon']\r\n```\r\nI'm sure your csv is more complicated than this one. But it is hard to tell where the issue might be without looking at a sample.","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":95,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nHi @ioana-blue,\r\nCan you share a sample from your .csv? A dummy where you get this error will also help.\r\n\r\nI tried this csv:\r\n```csv\r\nfeature,label\r\n1.2,not nurse\r\n1.3,nurse\r\n1.5,surgeon\r\n```\r\nand the following snippet:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nd = load_dataset(\"csv\",data_files=['test.csv'])\r\n\r\nprint(d)\r\nprint(d['train']['label'])\r\n```\r\nand this works perfectly fine for me:\r\n```sh\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['feature', 'label'],\r\n num_rows: 3\r\n })\r\n})\r\n['not nurse', 'nurse', 'surgeon']\r\n```\r\nI'm sure your csv is more complicated than this one. But it is hard to tell where the issue might be without looking at a sample.","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"I've had versions where it worked fain. For this dataset, I had all kind of parsing issues that I couldn't understand. What I ended up doing is strip all the columns that I didn't need and also make the label 0\/1. \r\n\r\nI think one line that may have caused a problem was the csv version of this:\r\n\r\n```crawl-data\/CC-MAIN-2017-47\/segments\/1510934806225.78\/wet\/CC-MAIN-20171120203833-20171120223833-00571.warc.wet.gz Rose Blakey is an aspiring journalist. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job. ^M ('Rose', '', 'Blakey') journalist F 38 journalist https:\/\/www.netgalley.com\/catalog\/book\/121872 _ is desperate to escape the from the small Australian town in which _ lives. Rejection after rejection mean _ is stuck in what _ sees as a dead-end waitressing job. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job.```\r\n\r\nThe error I got in this case is this one: https:\/\/github.com\/huggingface\/datasets\/issues\/1989#issuecomment-790842771\r\n\r\nNote, this line was part of a much larger file and until this line I guess it was working fine. ","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":197,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nI've had versions where it worked fain. For this dataset, I had all kind of parsing issues that I couldn't understand. What I ended up doing is strip all the columns that I didn't need and also make the label 0\/1. \r\n\r\nI think one line that may have caused a problem was the csv version of this:\r\n\r\n```crawl-data\/CC-MAIN-2017-47\/segments\/1510934806225.78\/wet\/CC-MAIN-20171120203833-20171120223833-00571.warc.wet.gz Rose Blakey is an aspiring journalist. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job. ^M ('Rose', '', 'Blakey') journalist F 38 journalist https:\/\/www.netgalley.com\/catalog\/book\/121872 _ is desperate to escape the from the small Australian town in which _ lives. Rejection after rejection mean _ is stuck in what _ sees as a dead-end waitressing job. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job.```\r\n\r\nThe error I got in this case is this one: https:\/\/github.com\/huggingface\/datasets\/issues\/1989#issuecomment-790842771\r\n\r\nNote, this line was part of a much larger file and until this line I guess it was working fine. ","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"Hi @ioana-blue,\r\n\r\nWhat is the separator you're using for the csv? I see there are only two commas in the given line, but they don't seem like appropriate points. Also, is this a string part of one line, or an entire line? There should also be a label, right?","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":49,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nHi @ioana-blue,\r\n\r\nWhat is the separator you're using for the csv? I see there are only two commas in the given line, but they don't seem like appropriate points. Also, is this a string part of one line, or an entire line? There should also be a label, right?","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"Sorry for the confusion, the sample above was from a tsv that was used to derive the csv. Let me construct the csv again (I had remove it). \r\n\r\nThis is the line in the csv - this is the whole line:\r\n```crawl-data\/CC-MAIN-2017-47\/segments\/1510934806225.78\/wet\/CC-MAIN-20171120203833-20171120223833-00571.warc.wet.gz,Rose Blakey is an aspiring journalist. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead,\"('Rose', '', 'Blakey')\",journalist,F,38,journalist,https:\/\/www.netgalley.com\/catalog\/book\/121872,_ is desperate to escape the from the small Australian town in which _ lives. Rejection after rejection mean _ is stuck in what _ sees as a dead-end waitressing job., She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job.```","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":139,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nSorry for the confusion, the sample above was from a tsv that was used to derive the csv. Let me construct the csv again (I had remove it). \r\n\r\nThis is the line in the csv - this is the whole line:\r\n```crawl-data\/CC-MAIN-2017-47\/segments\/1510934806225.78\/wet\/CC-MAIN-20171120203833-20171120223833-00571.warc.wet.gz,Rose Blakey is an aspiring journalist. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead,\"('Rose', '', 'Blakey')\",journalist,F,38,journalist,https:\/\/www.netgalley.com\/catalog\/book\/121872,_ is desperate to escape the from the small Australian town in which _ lives. Rejection after rejection mean _ is stuck in what _ sees as a dead-end waitressing job., She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job.```","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"Hi,\r\nJust in case you want to use tsv directly, you can use the separator argument while loading the dataset.\r\n```python\r\nd = load_dataset(\"csv\",data_files=['test.csv'],sep=\"\\t\")\r\n```\r\n\r\nAdditionally, I don't face the issues with the following csv (same as the one you provided):\r\n\r\n```sh\r\nlink1,text1,info1,info2,info3,info4,info5,link2,text2,text3\r\ncrawl-data\/CC-MAIN-2017-47\/segments\/1510934806225.78\/wet\/CC-MAIN-20171120203833-20171120223833-00571.warc.wet.gz,Rose Blakey is an aspiring journalist. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead,\"('Rose', '', 'Blakey')\",journalist,F,38,journalist,https:\/\/www.netgalley.com\/catalog\/book\/121872,_ is desperate to escape the from the small Australian town in which _ lives. Rejection after rejection mean _ is stuck in what _ sees as a dead-end waitressing job., She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job.\r\n```\r\nOutput after loading:\r\n```sh\r\n{'link1': 'crawl-data\/CC-MAIN-2017-47\/segments\/1510934806225.78\/wet\/CC-MAIN-20171120203833-20171120223833-00571.warc.wet.gz', 'text1': 'Rose Blakey is an aspiring journalist. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead', 'info1': \"('Rose', '', 'Blakey')\", 'info2': 'journalist', 'info3': 'F', 'info4': 38, 'info5': 'journalist', 'link2': 'https:\/\/www.netgalley.com\/catalog\/book\/121872', 'text2': '_ is desperate to escape the from the small Australian town in which _ lives. Rejection after rejection mean _ is stuck in what _ sees as a dead-end waitressing job.', 'text3': ' She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job.'}\r\n```\r\nCan you check once if the tsv works for you directly using the separator argument? The conversion from tsv to csv could create issues, I'm only guessing though.","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":292,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nHi,\r\nJust in case you want to use tsv directly, you can use the separator argument while loading the dataset.\r\n```python\r\nd = load_dataset(\"csv\",data_files=['test.csv'],sep=\"\\t\")\r\n```\r\n\r\nAdditionally, I don't face the issues with the following csv (same as the one you provided):\r\n\r\n```sh\r\nlink1,text1,info1,info2,info3,info4,info5,link2,text2,text3\r\ncrawl-data\/CC-MAIN-2017-47\/segments\/1510934806225.78\/wet\/CC-MAIN-20171120203833-20171120223833-00571.warc.wet.gz,Rose Blakey is an aspiring journalist. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead,\"('Rose', '', 'Blakey')\",journalist,F,38,journalist,https:\/\/www.netgalley.com\/catalog\/book\/121872,_ is desperate to escape the from the small Australian town in which _ lives. Rejection after rejection mean _ is stuck in what _ sees as a dead-end waitressing job., She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job.\r\n```\r\nOutput after loading:\r\n```sh\r\n{'link1': 'crawl-data\/CC-MAIN-2017-47\/segments\/1510934806225.78\/wet\/CC-MAIN-20171120203833-20171120223833-00571.warc.wet.gz', 'text1': 'Rose Blakey is an aspiring journalist. She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead', 'info1': \"('Rose', '', 'Blakey')\", 'info2': 'journalist', 'info3': 'F', 'info4': 38, 'info5': 'journalist', 'link2': 'https:\/\/www.netgalley.com\/catalog\/book\/121872', 'text2': '_ is desperate to escape the from the small Australian town in which _ lives. Rejection after rejection mean _ is stuck in what _ sees as a dead-end waitressing job.', 'text3': ' She is desperate to escape the from the small Australian town in which she lives. Rejection after rejection mean she is stuck in what she sees as a dead-end waitressing job.'}\r\n```\r\nCan you check once if the tsv works for you directly using the separator argument? The conversion from tsv to csv could create issues, I'm only guessing though.","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"thanks for the tip. very strange :\/ I'll check my datasets version as well. \r\n\r\nI will have more similar experiments soon so I'll let you know if I manage to get rid of this. ","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":34,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nthanks for the tip. very strange :\/ I'll check my datasets version as well. \r\n\r\nI will have more similar experiments soon so I'll let you know if I manage to get rid of this. ","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1989","title":"Question\/problem with dataset labels","comments":"No problem at all. I thought I'd be able to solve this but I'm unable to replicate the issue :\/","body":"Hi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. ","comment_length":20,"text":"Question\/problem with dataset labels\nHi, I'm using a dataset with two labels \"nurse\" and \"not nurse\". For whatever reason (that I don't understand), I get an error that I think comes from the datasets package (using csv). Everything works fine if the labels are \"nurse\" and \"surgeon\". \r\n\r\nThis is the trace I get:\r\n\r\n```\r\nFile \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 523, in \r\n main()\r\n File \"..\/..\/..\/models\/tr-4.3.2\/run_puppets.py\", line 249, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 650, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1028, in _prepare_split\r\n writer.write_table(table)\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/datasets\/arrow_writer.py\", line 292, in write_table\r\n pa_table = pa_table.cast(self._schema)\r\n File \"pyarrow\/table.pxi\", line 1311, in pyarrow.lib.Table.cast\r\n File \"pyarrow\/table.pxi\", line 265, in pyarrow.lib.ChunkedArray.cast\r\n File \"\/dccstor\/redrug_ier\/envs\/last-tr\/lib\/python3.8\/site-packages\/pyarrow\/compute.py\", line 87, in cast\r\n return call_function(\"cast\", [arr], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Failed to parse string: not nurse\r\n```\r\n\r\nAny ideas how to fix this? For now, I'll probably make them numeric. \nNo problem at all. I thought I'd be able to solve this but I'm unable to replicate the issue :\/","embeddings":[0.1789145917,-0.057323277,0.0051793605,0.1343670487,0.4113538861,0.3502125442,0.6419943571,0.1637074947,-0.1613174826,0.072349824,0.1788543761,0.063875109,-0.0607732385,0.0305257887,-0.1302312016,-0.1193707809,0.0890839919,0.1734792441,0.1816422939,-0.0444667041,-0.2867427766,-0.0398255885,-0.1566803157,0.3662883341,-0.2972057462,-0.3766840994,0.210543111,-0.1205281168,-0.1131336242,-0.4633953273,0.2718096077,0.0861199424,0.0026804437,0.403427273,-0.0001030963,0.0587566718,0.2967477739,0.1559889019,-0.3984076679,-0.4321283996,-0.1893598735,-0.0234080609,0.0345639475,-0.304477036,0.097983405,-0.2846682072,-0.0905411765,-0.2756657004,0.1046864018,0.4386056066,0.3266401291,0.2373397052,-0.2064999342,-0.2293033898,0.2311526239,0.1250885427,-0.030509105,0.0494754016,-0.1344388425,-0.0064582322,0.451570183,0.4609141946,-0.2723953128,0.2647711635,0.1353649944,0.1463634074,0.1377247274,-0.3910478354,0.27630952,0.1961430758,0.7452605367,-0.2964998186,-0.2943909466,-0.1697050184,0.1373761445,-0.4167192578,0.1813386828,0.0700835735,0.0063757841,-0.0470148064,-0.0331626348,0.0610272661,0.0807420164,0.0881801546,-0.0580747053,0.0383287892,-0.0986413881,0.3288169205,0.1125301197,-0.0747505948,0.2975304723,-0.0059944061,-0.0885701403,0.0415447913,-0.4325243533,-0.0296081789,-0.03688097,0.0516761839,-0.1181376427,-0.0387405939,0.0183867961,-0.3749561906,-0.147435531,0.3403448462,0.2151345611,0.2662760913,0.2706991434,0.5616859198,0.0911723003,-0.0761628821,0.0308042038,-0.0558076203,-0.0015736214,-0.4626432359,0.2433211356,0.0989312381,0.2934700251,-0.2398587912,-0.5455211401,0.1924418211,-0.0281014573,0.0940958261,0.1541178674,0.3049504757,0.0221223086,0.2065178007,0.0793474615,0.1936084628,-0.0091703711,-0.1172052622,-0.2069239616,0.0976077914,-0.2002792507,-0.1501065046,0.0502693951,-0.0720527992,0.1124494746,0.0640596598,-0.0010049804,-0.0760566592,-0.0163035393,-0.3864637911,-0.0127481688,0.3588932157,-0.1032631844,0.4203503132,0.2520543039,-0.4185499847,0.0596994981,0.1802112162,-0.3688518703,-0.0338136367,-0.2121571898,0.2862130404,-0.1030440852,-0.1879812926,-0.1200915053,0.0468235761,0.4193198979,-0.2145541459,0.1575899273,-0.3543839455,-0.309322238,-0.2777701914,-0.0663041547,0.2214326262,-0.6722466946,0.0043895538,-0.1475469172,-0.0291129798,0.0528067499,0.2960192561,-0.0810045302,0.2062918693,-0.0705920011,0.0303499643,0.1395510882,-0.3906908929,-0.1104637533,0.1035439596,-0.142003268,-0.2442642599,0.070480749,-0.0403397307,0.0917311311,-0.0916265324,0.058231581,-0.1318777055,-0.1996322423,-0.0951331779,-0.1437040716,0.053811159,0.6539097428,0.0927327424,-0.0168001503,0.0756469369,-0.1002281159,-0.3448366225,-0.0425324664,-0.1009552553,0.2062191814,0.2147901803,0.0556399524,0.2983595431,0.145952329,-0.0644701123,-0.4424965978,0.1017011255,-0.040051531,0.1395646781,-0.1451429576,-0.1923436373,-0.3297650516,-0.0537857413,-0.2079764754,0.0183234606,0.1633132845,0.0683606789,-0.2214220017,-0.0204836689,0.0140847312,-0.0758842081,-0.0172138363,0.1093306914,-0.1040249839,0.2656518817,-0.095116362,0.116894044,-0.0267394595,0.2531200051,0.43260625,0.2400517166,-0.0466214903,0.3670240343,0.084860824,-0.3270230889,0.0078406269,-0.1726583391,0.0181292314,-0.0838744566,-0.0163266789,0.1145248264,0.2768397033,-0.108597219,-0.084719412,0.1411858499,-0.1772009283,0.2762840688,-0.1619747579,0.2004608065,0.2599627972,-0.1093925983,0.1251222789,-0.2127108425,0.2158810496,-0.1297983676,0.2343145311,0.007142364,-0.2617174685,0.0677190274,0.4220534861,0.0015308688,-0.0035241439,0.00977387,-0.1586679667,0.1657464951,-0.0130174728,0.5957184434,0.3528428078,0.2059558332,-0.1638755053,-0.0582680255,-0.2622591853,-0.1369839311,0.2124618143,0.0951014534,0.0298523735,0.0572964214,0.0261099748,-0.0664389655,-0.1765730977,-0.2433386892,0.2082313001,0.3464300931,-0.5103119016,-0.0029065402,-0.3452730775,-0.2318636626,-0.3185971379,-0.0987911522,-0.2351592779,-0.4590440691,0.0904380977,-0.0774912238,-0.1903926283,0.1450765133,-0.0375639684,0.0079175811,-0.0910728574,0.1520520449,-0.0975843072,-0.4702496231,-0.2045716196,0.1811123788,0.0074801398,0.0469536185,0.4813604951,-0.1431984752,-0.1245553046,-0.0785719454,-0.2953169048,0.0040265005,-0.2730655372,0.1366042942,0.1816227883,0.1038595065,0.0084614437,-0.1486846209,0.2930162847,-0.0715195164,0.0246399287,0.1069572195,0.0452331603,-0.2920734584,-0.2664699554,-0.4750190973,-0.4745302498,-0.1774818748,0.0669479668,-0.0759082735,-0.0335288644,0.5198027492,0.0684422851,0.0656808615,0.0931449682,0.0232527219,-0.3467432261,-0.0935406461,0.2490476817,-0.1790918559,-0.2102454752,0.1960575432,0.1093097106,0.1175282523,-0.2039331645,-0.3139223754,0.0728625879,-0.0576014481,0.0589448512,0.0813107938,0.1948960274,0.2138828188,0.2671016455,-0.14800331,-0.3270515203,-0.3317119777,0.138173148,-0.0719680116,0.130635187,-0.106256038,0.2940857708,-0.1520837694,0.3265088797,0.0217723008,-0.2126559913,0.1918890625,-0.0478479452,0.520732224,0.0683805421,-0.4193749726,0.0658850446,-0.1191072389,0.0269461535,0.1760430783,-0.3196393847,0.0379564203,-0.1479679644,0.3946043551,-0.2478048056,-0.0996567681,-0.1006429642,-0.1501683146,0.0405271091,0.1169005409,-0.0409752764,-0.2084079385,-0.1915848404,-0.2512800992,0.0578237921,-0.0954615101,-0.0875514075,-0.5335751772,-0.2334990948,-0.1649959534,0.2940982878,0.1574511677,0.3346641958,-0.1785171628,0.2454716265,0.0477587357,0.2458996922,0.6149811149,-0.4832392335,0.1547025442,0.2263718545,0.3931273818,-0.1780014187,-0.1272556782,-0.1692573428,0.2947264016,-0.2957773209,0.0259889998,-0.1555342674,-0.0910772309,0.2631008327,0.1554060578,-0.1223801896,-0.3149805367,-0.2309517711,-0.179393217,-0.1272854805,-0.0533969216,0.063577123,0.1866378188,-0.3784869313,0.0986479819,0.1408237815,0.0735074952,0.2715046406,0.1186350659,0.223243013,0.0302767064,0.1973610669,0.3174021542,0.258725673,0.5099819899,0.5483982563,-0.3618466854,-0.5113295317,-0.119478181,-0.1147690713,0.2007072717,0.0650011748,-0.1422152817,0.1660956591,-0.0506399386,0.1854584366,0.0961981639,0.0223078765,0.4922950566,0.1605764627,-0.4898478091,-0.5388922691,0.2713595629,0.0392415076,-0.0478404164,0.1281552613,0.0797319561,-0.3407570422,0.2059528828,-0.1559936106,0.6908727288,0.0837979019,-0.0510306507,-0.017031014,-0.2261736095,0.7132230997,-0.1187114641,0.0007714179,-0.3614292145,-0.3074664474,-0.0622299165,-0.0221187603,0.0234746374,0.4506930709,-0.3085787892,0.3788378537,-0.3285129368,0.2046867013,0.0287500564,-0.0740613043,0.2494398057,-0.1323139668,-0.0824861303,0.2010341883,-0.1576193273,-0.1343759149,-0.167434901,0.0419054776,-0.2318480462,-0.2680606246,-0.2262602746,0.0822607502,-0.2538252473,0.1415782422,-0.0856541023,-0.0167195704,0.2168118656,0.2025098354,0.4863221645,0.0789240599,-0.1482861936,0.1838955283,0.0421487652,0.137023434,0.0343485437,0.0537465625,0.4150621295,0.1160494089,-0.2699809074,0.1285299659,-0.0795674622,0.1759495735,-0.087031506,0.0838941708,-0.0983973071,-0.3653751016,0.0326474793,-0.0770390704,0.2287102938,-0.2537733018,0.1773461103,0.0519827604,-0.0466799513,0.2120789587,-0.2636912465,-0.1465975046,-0.045746807,0.189232856,0.0259635169,0.1933567077,0.2472447157,0.0259733312,-0.3513920605,-0.256511271,0.0389520451,0.0239994898,-0.6778354049,0.2459823489,0.253354311,0.0759462416,-0.0571891218,0.2232685089,-0.0158298351,0.0710847676,0.0610138401,-0.527438283,-0.2186501473,-0.0046702353,-0.184754625,0.1169300824,0.2478653193,0.4110052586,0.1337577701,0.0714501068,-0.36415416,-0.020954106,-0.1723938733,0.2980129123,0.1525105983,0.0199608039,0.2820311189,-0.1788063049,0.2182168067,0.0007441466,0.0077501768,-0.2498184443,-0.1872093379,0.112410605,0.1352095455,-0.1972158104,-0.0634119287,-0.1643811762,0.0578369237,-0.0719841644,-0.0172511209,0.3105735183,0.0032078389,0.3735311031,-0.1509218365,0.232784614,-0.0463298671,0.2182773203,0.0127173578,-0.1140016541,0.0742452592,0.1729656011,-0.1127644703,0.0194780286,-0.1750795394,0.0847949162,-0.1170051396,-0.117929846,0.1342935562,-0.5689806342,0.2959568501,0.1936379969,0.4493623376,0.278136611,-0.186825484,-0.0707663894,0.3826090693,0.3079268634,-0.3656700552,-0.1683601588,0.1630017757,0.1655610651,0.0003905857,0.0756073967,0.0607252456,-0.031059619,0.0448928699,0.00808575,0.4554388523,-0.4376547039,0.2349049002,0.5269826651,-0.06197102,0.0999842882,0.496794343,0.1731765866,0.0635327548,0.5881984234,0.0346501805,0.1773209274,0.1874703467,-0.1379599571,-0.0754207894,-0.1722174138,0.1809182018,0.0969724283,0.01885649,-0.0428314731,-0.0954475924,0.3011342883,-0.3839454353,0.0041102921,-0.0057689263,0.111196816,-0.2255032808,-0.1060157195,-0.1562632769,-0.0987094864,-0.1664461493,-0.1973159462,0.0140279196,-0.17005077,0.3556351662,-0.0742807314,-0.1959642172,-0.4565812349,-0.1867784709,0.3054198027,0.0650814474,-0.2473099232,0.0163201988,0.2487560362,-0.1455009133,0.321377635,0.0862395689,0.6039391756,0.1713752002,0.0897913277,-0.3410919011,-0.0972950011,-0.0885803476,0.1516635269,0.21556817,0.045920819,-0.0368338749,0.3922655582,0.20187819,-0.1707492769,0.308580339,0.3357421756,0.2227041125,-0.170619905,-0.0630675256,-0.1339658201,-0.2573547363,-0.326639533,0.1013189182,-0.2697950602,0.1159571186,0.4622557759,-0.041513864,0.0840903223,-0.2942293286,0.1462800652,0.1317101866,0.3118901253,0.2739445269,-0.0476241335,-0.3456507027,0.1615712643,-0.5773794055,0.1294606179,-0.1186991856,-0.1146921888,-0.0334541127,-0.1582234502,0.1593215466,0.0206150599,0.3829500973,0.1888177395,-0.0159397498,-0.1086772382,-0.1195140183,-0.1931983083,-0.0825992078,-0.1219008043,0.2010160536,-0.3226900399,-0.0618358254,-0.1457766593,0.038975969,-0.3498084545,0.0937142819,0.1062968373,-0.0201893132,0.3603175879,0.1459479779,0.7699725628,0.0488152206,0.115031004,-0.0913874805,-0.3536492884,-0.4042431116,0.3688686788,0.3255909085,0.2565357983,-0.1063768342,0.0097639496,-0.1644826978,0.0945068598,-0.0435468145,0.3211485445,-0.2544761598,0.2145043164,-0.0240109246,0.0171400364,0.0991311371,-0.0556258745,-0.0026466679,0.3410545886,-0.206756115,-0.6047098637,0.6198091507,-0.2129311413,-0.2395551354,0.1537304372,0.1545257866,0.2213794887,-0.2187477201,-0.5100531578,0.0247500017,0.3773825169,0.0190037079,-0.4423443377,-0.0037064909,-0.192757085,0.0514444485,0.1135602817,0.3062542081,0.1166288108,-0.193286106,0.1991047263,-0.2543465793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1988","title":"Readme.md is misleading about kinds of datasets?","comments":"Hi ! Yes it's possible to use image data. There are already a few of them available (MNIST, CIFAR..)","body":"Hi!\r\n\r\nAt the README.MD, you say: \"efficient data pre-processing: simple, fast and reproducible data pre-processing for the above public datasets as well as your own local datasets in CSV\/JSON\/text. \"\r\n\r\nBut here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/templates\/new_dataset_script.py#L82-L117\r\n\r\nYou mention other kinds of datasets, with images and so on. I'm confused. \r\n\r\nIs it possible to use it to store, say, imagenet locally? ","comment_length":19,"text":"Readme.md is misleading about kinds of datasets?\nHi!\r\n\r\nAt the README.MD, you say: \"efficient data pre-processing: simple, fast and reproducible data pre-processing for the above public datasets as well as your own local datasets in CSV\/JSON\/text. \"\r\n\r\nBut here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/templates\/new_dataset_script.py#L82-L117\r\n\r\nYou mention other kinds of datasets, with images and so on. I'm confused. \r\n\r\nIs it possible to use it to store, say, imagenet locally? \nHi ! Yes it's possible to use image data. There are already a few of them available (MNIST, CIFAR..)","embeddings":[-0.1048939675,-0.40778476,-0.1252106577,0.3528693914,0.2064331174,0.0773042664,0.282169342,-0.013219526,0.1014454737,-0.0921515226,-0.2971546054,-0.0767041966,-0.0591904446,0.5414652228,0.5551636815,-0.0550173968,0.1947559863,0.0077369232,-0.1606640965,0.0243537035,-0.1311549246,-0.0184780583,-0.1919618845,-0.0560768805,-0.2441978604,0.0734436214,-0.2431037128,0.2023607492,-0.3524551094,-0.3153180778,0.1265478283,0.142026186,0.2038796246,0.151556462,-0.000114474,0.0261500701,0.2794184685,-0.2648744583,-0.0799028724,-0.2754318118,-0.210036397,-0.1958542466,0.1365066469,-0.3097448051,-0.0344094075,-0.3639695942,0.3208791912,-0.2029702365,0.4163274169,0.3321006596,0.1734585315,0.1770457923,-0.0120291691,0.071558252,0.1994965523,0.7043319941,-0.235324055,0.0765626132,0.3966988623,0.2751343846,0.1317591816,0.1904014796,-0.1471002996,0.2446244359,0.6670558453,0.2847822905,-0.063408114,-0.5953469872,0.1768101603,0.3605561852,0.7247242928,-0.2514562011,-0.3772625625,-0.2556195259,-0.3492871225,-0.1461131126,-0.0531076752,0.2958315313,-0.0094538117,0.234074384,-0.5905610323,-0.2758911848,-0.1851313412,-0.0329567306,0.0618031733,0.2776287496,-0.3768876493,0.2667973042,0.1519197822,0.1478042752,0.0193916522,-0.241119206,0.0904686004,0.2136386037,0.1061239913,-0.2127500772,-0.3010907173,0.173503831,0.4054274857,0.0513445102,-0.0248880852,-0.0001729852,-0.2388933897,0.263392657,0.333024621,-0.0660579428,-0.064734444,-0.0764091685,0.2803301215,-0.0680018663,0.2875709236,-0.0922331959,-0.2132638395,-0.0572031885,-0.1317405552,-0.3184929788,0.13540034,-0.2608077824,-0.0464894362,0.0581916906,0.1385325789,-0.1853195131,-0.1996646821,0.2353790253,-0.0096433489,-0.1611839235,0.1309799552,0.2234884351,0.1141237915,-0.1701417267,0.0089839054,0.1949697286,-0.2847525775,0.1963899434,0.0610174015,-0.2041507661,0.3392297029,-0.104999952,0.2044096738,0.0342721306,0.1778145283,-0.2119627893,0.3175321519,0.4655555487,-0.0916534141,0.0578254648,0.0791012719,0.0507043339,-0.4007256329,0.1867390871,-0.4734035432,-0.2855846286,-0.1491551101,0.0366014242,-0.1271826476,0.0141973309,-0.4138632715,0.2678916454,0.1148289219,0.1325330138,0.237514466,0.0565747954,-0.3937700391,-0.2800178826,0.2161932737,0.2000338137,-0.4700542986,0.1166008115,-0.0538984425,-0.2484962493,-0.0362886488,0.2947596908,-0.193214044,0.1562079638,-0.2128867656,0.1740450412,0.1366384178,-0.4921633005,0.2029118538,0.3513148129,0.3671303689,-0.1108635813,0.0441715084,0.1537500173,0.0429131538,0.0106786927,-0.2065613121,0.2118646204,0.0240099058,-0.1647772193,-0.0510587171,-0.4474026859,-0.0594857372,0.2002526224,0.0996519998,0.1039628312,0.1326233,-0.0051940652,0.0929168984,-0.0749805421,0.3068466187,0.0156941004,0.1508465409,0.1359285116,0.1380657703,-0.0318658687,-0.4235838056,0.1685129404,0.0633087978,0.076366052,0.0761730149,-0.3972407877,-0.0179494843,-0.0966624618,-0.0226482507,-0.3566534817,0.0199158248,-0.069963634,0.1150769293,-0.1959699243,-0.5987902284,0.3285928369,-0.1511595547,0.0230142307,-0.4239824712,0.141873166,-0.1346036345,0.1803492904,-0.0286204834,-0.0403109193,-0.2251121998,-0.3223880827,0.2558919489,0.3308746219,0.0934398025,0.072260499,0.1576122344,0.565074861,0.1312624514,-0.2301090509,0.3880794346,-0.1910669655,0.2906061709,-0.0601735599,-0.4578298032,0.4475467801,-0.0550349243,0.1308474094,0.0638445094,-0.1445367187,0.2165029198,-0.0154157709,-0.2617865503,-0.151565209,-0.0750195459,0.0945293903,0.3611973524,-0.0716756284,-0.3708128929,0.0365550667,0.2601163089,-0.1820849925,-0.0158427525,0.2417083979,-0.1150560528,-0.3044469655,0.1658321023,0.0687477961,0.4986746013,0.0392027944,0.0669747517,0.287188977,0.0804619491,-0.167108357,0.133818537,0.0213156771,0.2222685367,-0.0037227371,-0.2135404199,-0.0318362936,-0.1227990538,0.0990001559,0.0228221398,0.0743144974,-0.2289332896,0.2501445115,-0.1536984742,-0.2292582691,0.0220878907,0.0896803215,-0.3034615815,0.0047843838,-0.1613828838,-0.0531568415,0.0799912587,-0.1376591921,-0.0357058197,0.4570162892,-0.1464422047,0.1405964345,-0.328478694,0.0716090351,0.0750908256,0.0750454441,0.3216892779,0.1382000893,0.400268048,-0.3740539551,-0.1360500604,-0.3646126688,-0.1546107084,0.4024239779,-0.2423143238,0.3959169388,0.342402935,0.2233851999,-0.1198301986,0.2312307209,0.1523539573,-0.2863830328,-0.1678372473,-0.3098679483,-0.0997557715,-0.2488845736,-0.3115008771,-0.0051123207,-0.3916942477,-0.1449797899,0.2766243219,0.3025512993,0.3917849064,0.2861984074,0.1617548019,0.3849445879,-0.2790232897,0.0002967174,-0.369348079,-0.660445869,0.280684948,-0.2496239096,-0.3559561968,0.2686167061,0.1082687527,0.1110761687,0.0524176806,-0.5647812486,-0.3803999126,0.073514536,-0.0298809744,0.3173148036,0.2644761801,0.1411495209,-0.3616519868,0.021610884,-0.2539398074,-0.2714328468,-0.0720840171,-0.0069525447,0.1494468302,0.384890765,0.2619087994,-0.0657447129,0.4962334633,0.1023640931,-0.0069294879,0.4028249085,0.0429874808,0.4390588999,-0.5036187768,0.1889145374,0.3229768276,-0.2280289084,-0.0893528536,0.2797310054,0.1519518793,-0.0002822188,-0.2032935321,0.0741904378,-0.2951332629,0.0860043988,0.1225500032,0.3008064628,0.0357659571,-0.0828720108,-0.2113488317,0.0658424497,0.0793485716,-0.1980214417,0.5932247639,0.3951218128,0.042381756,-0.4509610534,-0.0493150093,-0.475903511,0.2078602463,-0.2291045487,0.0366470255,-0.0622764714,0.1710507274,0.3513618112,-0.2008666396,0.3242226541,-0.447100997,-0.3030695319,0.1148628071,-0.2994038761,-0.1905008256,0.0899868459,-0.1753987968,0.193481788,-0.0422462486,0.4469906986,-0.2695123255,-0.3309996128,0.2156669348,0.0433714949,-0.2091078311,-0.1733769178,-0.023739282,-0.3795310557,-0.4606885314,-0.3016344309,0.1094654724,-0.3531163335,0.078941144,0.10455852,0.0741396919,0.2746052444,0.0787395611,0.0344861411,0.0890977755,-0.0308660269,-0.1191001013,0.2502967715,0.2833510339,0.108338885,0.6675297618,-0.2094116062,-0.1703843325,0.1666611284,-0.0685452372,0.3022337854,0.2958959639,-0.0461920984,-0.0509172641,0.2474728525,-0.0380423851,-0.5824607611,0.0347112641,0.2892960608,-0.1367737651,-0.4954515398,-0.5957931876,0.6540586948,0.0735014156,-0.0504066199,0.0816590786,0.2155452818,-0.3724491894,0.1935535222,0.5435110331,1.07117033,-0.2064470649,0.4783608317,-0.0353165679,-0.0264179688,0.3445591629,-0.1600820273,-0.2060883492,-0.2130370736,0.0525089651,-0.2116365284,-0.0434282683,0.2980111241,0.0743614882,-0.2512974739,0.0711916685,0.0467813127,0.1992858052,-0.2622813284,0.3801475763,-0.0763475299,-0.2325717658,-0.0787089542,0.1511822492,0.1660849154,0.0884912238,-0.0883435532,0.0518044271,-0.1285295039,-0.0104411077,-0.2773341537,-0.4066328108,0.0938111097,-0.2124226689,0.0439094268,-0.0421208553,0.1966655403,0.061006885,0.4875640273,-0.0119994748,-0.5061593652,-0.0909203291,-0.1689060926,-0.1601185054,0.0015376601,-0.3381931782,0.2413755655,-0.1594561785,0.0756148845,-0.0897733048,-0.0050940299,-0.2896893919,-0.1152639911,0.0730126351,0.0224787816,-0.1160960346,-0.3256324828,0.0479331352,-0.1418657154,0.0119071966,0.0715865642,0.28895998,0.0282926969,0.2428253591,-0.02214613,-0.1877589375,-0.1326957941,0.0051650698,0.1976083517,-0.0216005519,-0.0415337645,0.1155256182,-0.2569549978,-0.1879383773,0.2555366755,0.3657169044,-0.2913578153,0.0343418457,0.2049902529,-0.0385923348,-0.2323975563,0.3837562203,0.2756189704,-0.2539528608,-0.1566249281,-0.4555267096,-0.455647558,0.2915142775,0.0550087318,0.2722638845,0.0974757299,-0.1319726855,0.3784185648,0.1438316852,-0.2696527243,0.2071363032,0.1943935603,0.2276329398,0.3067966402,0.1615450084,0.1273260266,-0.1219336092,-0.1021168232,0.0005315436,-0.5738059282,-0.1093947291,-0.0829923972,0.2434397191,0.2102246881,-0.6044513583,0.2512055039,-0.1884349585,-0.4141263664,-0.1705609262,0.0567875467,0.3243681788,-0.133683756,-0.1698004156,0.3111153245,0.0358860008,0.1018795371,0.012681229,0.1406432688,0.2875841558,0.1090644747,0.0058355168,-0.1595472991,-0.0762465224,0.2651560903,0.2241549343,0.3071344495,0.0347260088,0.4503049552,-0.0871042237,0.2993300557,-0.0995816439,0.6628457308,0.4894661307,-0.2899217606,0.2158207297,0.3399309814,0.0476656444,-0.151532352,-0.0237075984,0.1133874953,0.1913732886,-0.11480093,0.325743109,0.5763192773,0.3541560173,-0.0552245229,-0.092071414,0.4977546036,-0.236181289,-0.0451121256,-0.0041507888,-0.2009322792,0.2144486308,0.3568070531,0.2959004343,0.2447049916,0.3949370682,-0.0449648052,0.3169942498,0.2117173225,0.3111312985,0.1836911142,-0.1737136692,0.3419051766,0.3521102965,-0.0029109931,0.2933187783,0.0115053738,0.2548394203,-0.0377789401,0.0649414659,-0.218103528,0.2974994183,0.0736963749,-0.2234390676,-0.106574446,-0.2485699207,0.0351705253,0.0781802833,-0.2575296462,-0.4391990602,0.4528007209,-0.1613834053,0.0814213529,-0.011728866,0.2299821526,-0.1840503067,-0.0126001993,-0.0147157451,0.4757886529,-0.038856864,0.1751498282,0.2862821519,0.4984342158,0.4124894142,0.2951298654,0.041523803,0.1162270457,-0.2170939445,-0.0240293127,-0.1073999032,0.2956203222,0.1052569374,-0.2680221796,0.3524373472,0.0467499755,-0.0150575126,0.1611464024,-0.248075977,-0.0947321951,-0.2578945458,-0.0653268024,-0.0398038253,-0.0521666855,-0.0196102802,0.0834797099,-0.4652499855,-0.1512021869,0.4083555341,-0.1822608262,0.2170265466,-0.0191248022,0.0429893993,-0.1277041286,0.5674458742,0.1758663654,0.2255419493,-0.1872519702,-0.0654124022,-0.6191353202,0.0173139125,-0.1485690475,-0.027677441,-0.034255404,0.1928011328,0.1183425859,0.3820797205,0.1264029443,-0.1649755239,0.0781743601,0.1624612063,0.0699716285,0.0281057656,-0.1390953213,0.5387628078,-0.122906968,-0.0895444974,0.3787787557,-0.0359519869,-0.117206499,-0.230039984,0.2673724294,0.125611946,0.050617177,0.451018393,0.0111474665,0.1042971611,-0.2395535856,-0.0749331042,-0.2782075703,0.0913221836,-0.3062449098,0.1464633644,0.039266672,-0.0720214769,-0.5130123496,-0.1260092407,-0.1060917303,-0.0476455204,-0.214737609,-0.1384328753,-0.2805695534,-0.0222007651,-0.3666781485,-0.1302182674,0.0799453631,0.1912415177,0.0207773317,-0.3268556297,-0.4230944812,-0.0962444395,0.5732990503,-0.0807914287,-0.0943527445,0.0504300185,-0.2250857353,0.0867056698,-0.1694974601,-0.8003807068,-0.0244751051,0.2794879377,0.114461273,-0.049496457,0.1213410348,0.0375999548,-0.147103563,-0.372990042,0.3526125252,-0.0036821575,-0.1638883948,-0.3388902545,-0.4334890246]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1986","title":"wmt datasets fail to load","comments":"caching issue, seems to work again..","body":"~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\wmt14\\43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\\wmt_utils.py in _split_generators(self, dl_manager)\r\n 758 # Extract manually downloaded files.\r\n 759 manual_files = dl_manager.extract(manual_paths_dict)\r\n--> 760 extraction_map = dict(downloaded_files, **manual_files)\r\n 761 \r\n 762 for language in self.config.language_pair:\r\n\r\nTypeError: type object argument after ** must be a mapping, not list","comment_length":6,"text":"wmt datasets fail to load\n~\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\wmt14\\43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\\wmt_utils.py in _split_generators(self, dl_manager)\r\n 758 # Extract manually downloaded files.\r\n 759 manual_files = dl_manager.extract(manual_paths_dict)\r\n--> 760 extraction_map = dict(downloaded_files, **manual_files)\r\n 761 \r\n 762 for language in self.config.language_pair:\r\n\r\nTypeError: type object argument after ** must be a mapping, not list\ncaching issue, seems to work again..","embeddings":[-0.3388853967,-0.1563400775,0.0392464399,0.5702248216,0.3619326949,0.1132854521,0.3336086273,0.3772256374,0.4356866479,0.0311077554,0.0063652312,0.1635423601,-0.3964485228,0.2130109966,0.1321742833,-0.1492633522,-0.1550919116,-0.2090711892,-0.6247566938,-0.0261804648,-0.4880731106,0.3302257657,-0.1743941754,-0.0524810851,-0.3326319158,-0.003986551,0.0660699159,0.2896155119,-0.0262576342,-0.389860332,0.1948823631,-0.1265439838,0.2346330732,0.275896132,-0.0001089429,0.0952031463,0.2280207276,-0.0845097005,-0.2599442899,-0.1657102555,-0.5657229424,-0.258523047,0.0475684702,0.0691884533,-0.0172015652,-0.0841678977,-0.056291651,-0.3773417771,0.4659874141,0.4160060585,0.2715233862,0.3688670397,0.0250798482,-0.3898190558,0.3594439924,0.0130669605,-0.1825276315,-0.0915164724,0.232977286,-0.0923134461,0.0708752722,0.3522186577,-0.0782359242,0.1595731825,0.2243726999,0.0555943325,-0.0666337162,-0.2400630116,0.1622279137,0.142931208,0.6131698489,0.0177526716,-0.3178753555,-0.0274521336,-0.3316507638,-0.29146263,0.3521073759,0.1055378541,0.0053326194,0.3410355151,-0.030496629,0.1008111686,0.2323692888,0.1528809518,-0.0199761242,0.1634058356,-0.215850547,0.1116302311,0.184881717,-0.1075047031,0.2072139233,-0.0288825911,-0.0580178984,0.0980215892,-0.2125768214,0.1143563017,0.2425225079,0.1487284154,-0.0653247237,-0.036812447,-0.0471866503,0.1937867105,-0.0684492141,0.2620108724,0.1301662624,0.3223253191,0.1221699566,-0.153032124,0.2061553895,-0.0245427284,-0.2453738749,-0.0455787629,-0.2337267548,-0.3125987649,0.1889365762,0.0432834253,0.2104675025,0.0275299381,-0.1316523552,-0.0263047516,0.0716635138,-0.3161900938,-0.0876014531,0.4200783074,0.0851868764,0.2828685641,0.1484407783,0.2556803524,-0.1024999246,-0.2287066132,-0.3322744668,0.1364493817,-0.2935177386,0.0552489832,0.3177166283,0.3159404695,0.246010825,-0.084926784,-0.0850898623,-0.15609999,0.1545800269,0.01799036,0.2353367805,0.3010546863,-0.076279141,0.3344309628,0.5337517262,0.013023627,-0.1959387064,0.0185390487,-0.3508415818,-0.2006485611,0.2434013039,0.2642399967,0.0812531188,0.1216968372,-0.1743879765,-0.0460180454,0.2699588537,0.0516423732,-0.2573116124,-0.4044613838,-0.2292932719,-0.2044887394,-0.0104886489,0.6401978731,-0.4205351174,-0.1784503311,-0.1034565195,-0.0389555953,0.2265539765,0.4219801128,-0.2955787182,0.3152286708,-0.3064777553,0.0786348954,0.2414634079,-0.1370716691,-0.1213568598,0.0113088619,-0.0624857694,0.2895672619,0.2456699759,-0.0178858917,0.1957135499,-0.2158804238,0.2070478648,0.1909878254,-0.1261783689,-0.0267336927,-0.2576395571,-0.3646636307,0.4106539786,0.0112024238,0.3808276355,0.0591872856,0.2329039723,0.260602802,0.0606748946,-0.0754262283,0.0368817598,0.0603792332,0.0335468538,-0.0355301984,0.0586558431,-0.2482159585,-0.4154435098,0.338832438,-0.1478432417,0.2518348992,0.0225054063,0.1131667271,-0.1243863627,0.1084148586,-0.4083625674,-0.2087507844,0.1520325541,0.0971758813,-0.0757928044,0.0594159104,-0.0660963506,0.0731556267,0.0591850206,0.0631009787,-0.1793776751,-0.0472442284,-0.1980560124,-0.1755319834,-0.0141887367,0.0814412534,0.1879961938,-0.2943767309,-0.2541226745,-0.106797725,0.3654383123,0.0142308287,-0.1424064785,0.1551862657,0.0786089003,-0.3979342878,0.2732883692,0.0231634155,0.120441325,-0.4718690217,-0.0119674969,0.1235987097,0.2477939427,0.2379597425,-0.0839016587,0.2052850872,0.356764406,0.0429207832,0.2606799901,-0.3454299271,0.14693515,0.0507004969,-0.0245486777,-0.0154483113,0.050430987,0.0918973386,0.7295478582,0.3369091153,0.3306079805,0.0361793749,-0.4006486535,-0.0654311925,-0.085292384,0.1598505229,0.4039934576,0.0443479791,0.0946987271,0.0090333261,0.1567132324,-0.2341935933,0.2785615027,-0.157329157,0.0293368418,0.1701325625,-0.1352719367,0.0476262458,-0.360971421,0.2541760802,-0.0863987431,0.137501657,-0.2064768374,-0.0641895905,-0.5173707604,-0.0987077504,-0.3122809529,-0.2706081569,-0.3408389986,-0.48423329,-0.2059248984,-0.0046142694,-0.0751244947,0.430990696,0.2410561144,0.0431273393,0.053121198,0.0466187373,0.0395303816,-0.3208910823,-0.2228517532,0.0227681156,0.4493676424,-0.1045164391,0.2901251614,-0.4252577722,-0.023239091,-0.0860539675,-0.1928814501,0.1055801436,-0.0290782209,0.2146313488,0.3267443776,0.2112122923,0.3802035153,-0.2122448683,0.4954159558,-0.214000687,-0.0280094892,0.139326483,0.0726897418,-0.0983089432,-0.2263500094,-0.4966059029,-0.4336785972,-0.4345676899,-0.0378917605,0.1799628586,0.1673876345,0.2239832878,0.0751128122,0.1288751215,-0.1760531813,0.1092718244,-0.3114063144,-0.1971670687,0.3201406002,-0.1813881248,-0.3238304257,0.3230157793,-0.0691390783,0.224494949,0.0961188152,-0.0936224163,-0.1407164484,-0.2153965086,0.0995296016,0.2240286916,0.346755147,0.2097037286,0.2050194293,-0.0313687697,-0.2882865071,-0.2545294166,0.1370033175,-0.0064833793,0.2539068162,0.0648909733,0.1784543842,-0.0656318739,0.6545190215,0.0086750761,0.0393158533,0.3305959404,0.2654773891,0.2299578935,-0.1722156852,-0.4679053724,-0.0539515428,-0.3365930319,-0.0113472976,0.2553058863,0.1285797954,-0.2581196427,-0.09282998,0.1328211129,-0.2144163251,-0.2973546386,0.1663918793,0.1286385357,0.1201717481,0.2331998795,0.1261652112,-0.191022858,0.049418591,0.109274298,0.3992340863,-0.0755802318,0.1322114468,-0.2438860536,-0.005502088,-0.3649451733,0.2589742541,0.0301339217,0.389285475,-0.2335260808,-0.2509506047,-0.1993843615,0.0286647715,0.5613512397,-0.3648861945,0.2238378823,0.2303847224,-0.1264564097,0.291205734,-0.0672904029,-0.1587489247,-0.0258710384,0.3056671023,0.1819134355,-0.0942946002,-0.1765313596,-0.0815170482,0.4090386331,-0.1798054129,0.0782328099,0.0294991545,-0.2381497473,-0.3609119952,-0.2278444022,0.123694703,0.0832296684,-0.2067482769,0.0651933998,-0.1512577087,-0.0409280136,0.1481562853,0.0407726727,0.2344475389,-0.0360365957,0.1018421128,-0.051134754,0.4142954051,0.2564111352,0.5204036236,-0.399586767,-0.1393707544,-0.1287464499,-0.2314675897,0.1259210706,0.2830592394,-0.2586618066,0.1469757706,0.006117336,0.1122120917,-0.2818084955,0.1326900125,0.3012812734,0.1150346845,-0.2354940325,-0.4385925531,0.1199197024,0.0276776105,0.0405982658,0.3151747584,-0.2408192158,-0.5700796247,0.1414667964,-0.1812869757,0.6553112864,-0.1610859036,0.2337189764,-0.0606625825,0.1056082994,0.3364338279,0.2162844837,0.010017436,-0.2175841033,-0.1246965453,-0.0652330741,-0.156924516,0.2384860367,0.2110469043,-0.3539441526,0.2488985062,-0.1240767241,0.3158943355,0.0413822308,0.0289901644,-0.1059274003,-0.2751176953,-0.0548970997,0.184248805,-0.1741878539,0.3417241573,-0.0929451734,-0.2915517092,0.2771413922,-0.2383934408,-0.3894301355,0.0609889664,-0.4143161774,0.1237198859,-0.0778249055,-0.0515376069,0.3886596262,0.0187870283,0.0642736778,0.1256380975,-0.1982099414,-0.0156366918,-0.0608247183,0.0068582785,-0.2149094939,0.1186459288,0.2731108069,-0.030978648,0.011590085,-0.0072157998,-0.1848579794,0.035621576,0.0876247361,0.0900985152,0.0207824931,-0.214451611,-0.4203359485,0.0980667919,-0.1910964847,-0.2341104746,0.1831472218,0.1338882148,-0.3069568872,0.5446383357,0.0084583322,-0.0681894943,-0.046435371,0.2788502276,-0.0870566666,0.0141997514,0.4689539373,0.4608744383,-0.1571596414,-0.3521206677,0.0115451301,-0.016121272,-0.3512851596,0.2735048831,0.1298124641,-0.1359515637,0.0975910425,0.4091960192,0.3144862652,-0.0064361221,-0.0331232995,-0.6006265879,-0.0701395869,0.1344459653,-0.1646348387,0.0561887771,-0.2546944916,0.2535224259,0.0964121073,0.3129808605,-0.3912235498,0.2235292941,-0.1791581362,0.0091999909,-0.0436379239,-0.147516191,0.3303283155,0.0553997755,0.0405875966,-0.0632708147,-0.4975077212,-0.2401835173,-0.1999774575,0.1221420616,-0.0468727313,-0.146143958,0.141951412,-0.1533184648,0.1123716757,-0.2411905229,0.0343132578,0.103558667,-0.2819585204,0.3256540596,0.2208948582,0.2039568126,0.0139787775,0.2516504526,-0.1553240567,0.0197263788,-0.0793997794,-0.1209991425,0.0472147614,0.0050312523,0.214392662,0.0242107473,-0.140850842,-0.1022707149,0.5038083196,-0.3771141768,-0.0654525533,0.3685275912,0.2165228128,0.3240405321,0.0231866594,0.0702738017,0.0790081173,0.1747967899,-0.3646740317,-0.1281636804,0.4066771567,-0.0049502994,0.0336250179,0.3243213892,0.3738220632,-0.1128483564,0.1286474168,0.1095463634,0.2935252786,-0.2715616822,0.1808086485,0.118851915,0.0260651335,0.2622679174,0.1813650429,0.0526101775,0.1022734642,0.2956345677,-0.1970249116,0.353841871,-0.1262918115,0.0354524441,0.0271984935,-0.6782624722,0.2208520174,0.2540223598,-0.0802531093,0.0678980425,-0.3105304241,0.2246866226,-0.336185813,-0.1228170991,-0.0463943146,0.2738315761,-0.2266457528,-0.0855176672,0.163771987,-0.1524950564,-0.1856617481,-0.1961367428,0.0128546124,-0.0794394836,0.2415431291,0.0284991357,0.0132366531,-0.272841841,-0.2647996545,-0.3008139133,-0.116881676,-0.3331807256,0.1662574857,0.1069409102,-0.0176948104,0.0075899614,0.0695131794,0.3988871574,0.5008767247,-0.2648846209,-0.0867637545,-0.196947664,-0.0482549034,-0.3694549203,0.3029950857,-0.0750766024,0.0192057714,0.2510211468,0.1331951469,-0.1847474873,-0.0919668078,-0.1261492223,-0.2398339659,-0.0938880071,0.3400411606,-0.1425893307,0.0243272409,-0.1773078889,0.2223648727,-0.3230505586,0.1655036956,0.7220769525,0.003210742,0.2474960983,-0.09102837,0.104308635,-0.0197832361,0.5697503686,0.3506541848,0.0768358707,-0.3043240905,-0.1035114005,-0.6640555859,-0.0051523084,-0.2052468657,0.1716413647,0.0405423157,0.1656543016,0.1962125003,0.2586087286,0.2245393842,0.2900209725,0.1666387022,-0.0414694883,-0.4483461678,-0.2288763821,0.0788210481,0.0647418424,0.0666648373,-0.4194889665,0.0838084817,-0.1597698331,0.0892409459,-0.4381546676,0.0088250116,0.0350856818,0.3024075925,0.1233434677,0.271546185,0.2669329345,0.0623705722,0.0152222803,-0.169302687,-0.4776983261,-0.0411378331,-0.0285590775,-0.0117203733,0.4172450006,-0.2775146663,0.0593073666,-0.4564947486,0.2788071036,0.0720058233,-0.0705013424,-0.1984367818,0.2864790559,-0.1076682881,0.0055410084,0.4469618201,0.4367139637,-0.063735351,0.0750084892,-0.2809995711,-0.2912311852,0.198244065,-0.2182110101,-0.5651194453,-0.2045198232,-0.068005845,-0.2664653063,-0.1949589998,-0.3291836083,-0.0541994423,0.2271216661,-0.040732149,-0.1602048874,0.2782325745,-0.0562099591,-0.1637247205,0.0110497652,0.0379045159,0.2311168909,-0.1238264367,-0.1699872911,-0.0333337896]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1983","title":"The size of CoNLL-2003 is not consistant with the official release.","comments":"Hi,\r\n\r\nif you inspect the raw data, you can find there are 946 occurrences of `-DOCSTART- -X- -X- O` in the train split and `14041 + 946 = 14987`, which is exactly the number of sentences the authors report. `-DOCSTART-` is a special line that acts as a boundary between two different documents and is filtered out in our implementation.\r\n\r\n@lhoestq What do you think about including these lines? ([Link](https:\/\/github.com\/flairNLP\/flair\/issues\/1097) to a similar issue in the flairNLP repo)","body":"Thanks for the dataset sharing! But when I use conll-2003, I meet some questions.\r\nThe statistics of conll-2003 in this repo is : \r\n\\#train 14041 \\#dev 3250 \\#test 3453\r\nWhile the official statistics is:\r\n\\#train 14987 \\#dev 3466 \\#test 3684\r\nWish for your reply~","comment_length":78,"text":"The size of CoNLL-2003 is not consistant with the official release.\nThanks for the dataset sharing! But when I use conll-2003, I meet some questions.\r\nThe statistics of conll-2003 in this repo is : \r\n\\#train 14041 \\#dev 3250 \\#test 3453\r\nWhile the official statistics is:\r\n\\#train 14987 \\#dev 3466 \\#test 3684\r\nWish for your reply~\nHi,\r\n\r\nif you inspect the raw data, you can find there are 946 occurrences of `-DOCSTART- -X- -X- O` in the train split and `14041 + 946 = 14987`, which is exactly the number of sentences the authors report. `-DOCSTART-` is a special line that acts as a boundary between two different documents and is filtered out in our implementation.\r\n\r\n@lhoestq What do you think about including these lines? ([Link](https:\/\/github.com\/flairNLP\/flair\/issues\/1097) to a similar issue in the flairNLP repo)","embeddings":[0.1660706699,-0.3441519439,-0.026519794,0.3718201518,-0.3747011423,-0.0020881079,0.0613716431,-0.070897989,-0.9394959211,-0.0062836837,0.1252351701,0.1568083316,0.081358403,0.0008895223,-0.0184590258,0.0292958673,0.1648515314,-0.0400795117,0.2795078754,-0.1671278775,-0.0582850873,0.4750356674,-0.5303708911,0.0231517497,-0.8211599588,0.0690073296,-0.1038991064,0.1645847559,-0.3679313362,-0.3551759124,0.2766619027,0.0723338574,0.1248623282,0.1931931823,-0.0001210675,-0.435992837,0.2224059701,-0.0981824324,-0.1973603368,0.2145414054,-0.193009004,-0.4295057356,-0.038219817,-0.1875087768,0.2144868076,0.4310618043,-0.5452502966,0.2614451051,0.376606524,0.3219401538,0.1062864661,-0.0599874817,0.1251414269,-0.0609836057,0.5427105427,0.044348523,-0.0532875881,0.2690054774,0.1820478737,0.2210755646,-0.0648549423,0.3887670636,0.0820056722,-0.2496798337,-0.0430985987,-0.0280937646,-0.2199691534,-0.2218949944,0.0012414544,0.4098291993,0.4609613121,-0.2167000175,-0.3023409545,-0.6893437505,0.2730959952,-0.2258210778,0.1093596816,-0.0073540686,-0.0039547747,-0.0349640995,-0.1026145592,-0.3313957453,-0.2891083062,-0.1153049543,-0.2888514102,0.3619834781,-0.0982110873,0.0420258865,0.1299825907,0.2171084732,0.6810007095,-0.035297446,-0.037436787,-0.059974473,0.0213083457,-0.2826972008,0.0769811496,-0.0876311585,0.4273091257,-0.3164520264,-0.3112627864,-0.4441926479,0.0422772653,-0.0517050996,0.3466984034,0.0796580166,0.1834950298,0.4299295247,0.0845104232,-0.044189319,0.1412351578,0.114253521,-0.3085027933,-0.104844898,-0.4565196335,0.0863843933,-0.4163788557,-0.4130695164,-0.1330944002,0.1724139899,-0.1113018841,-0.2724635601,0.0154795712,0.0970013514,0.2453032285,0.568240881,-0.0069987499,-0.1539082527,-0.0962361097,-0.2754597962,-0.0487606972,-0.125920862,-0.0837792233,-0.0407452844,-0.0431362875,-0.4689645767,0.1422903091,-0.0367282778,0.3424780965,0.1044097841,0.0004998242,-0.190817818,0.1123936474,-0.0760022998,-0.2280941159,-0.0431931168,0.0182659924,0.4265727103,-0.2598370016,0.1983314008,-0.0825977921,-0.3248425126,-0.263859719,0.0226160027,-0.2561658621,-0.1643792093,0.6193922758,0.0523145199,0.2874726951,-0.1028845236,0.2342781276,-0.2183289379,-0.1448468417,-0.075727649,0.2444783598,0.1550737917,-0.243281275,-0.0339366682,0.0535974391,-0.0178699493,0.0158104803,0.2335743606,-0.1454826593,0.4814392328,-0.0997478813,-0.158286795,0.3144389093,-0.5198950171,-0.1304037869,0.1959632933,0.1512755156,-0.0103271725,-0.1246015429,-0.287071228,0.5772898197,-0.0048112338,0.0324353911,0.0367222652,0.0223802384,0.0059987605,-0.3243201673,-0.1682890207,-0.4615073204,-0.311951071,-0.0392338447,-0.4152747393,-0.3570280075,-0.1325614899,0.4206111729,0.1739861518,-0.0713801682,0.1050239056,0.2062046975,0.2779878378,0.1916780323,0.3133241832,-0.1001001298,-0.0194032844,0.3702386022,0.1440892518,0.5833430886,-0.2024097294,-0.3424496949,0.0180991739,0.3618111312,-0.2660242915,0.0464325994,0.0784569159,0.2490747869,0.1045739874,-0.2230343521,0.229434073,-0.0686060861,0.0829589963,-0.022883229,-0.1679258645,0.0076258061,-0.0514643341,0.0765403956,0.3759826124,0.0501769334,0.2213463932,-0.043064557,0.1560602933,0.3733198345,0.0259110723,0.1593764722,-0.1503411084,0.1712959856,0.0713375136,0.003965931,-0.3762101233,-0.0671595261,0.1555728763,-0.1190159768,0.2018063068,0.3937220573,-0.0849646702,0.1251069754,-0.1524040103,-0.103281498,-0.0504986458,-0.043562945,-0.4440820217,-0.0775448456,0.2097178549,0.0434644893,0.1147812456,-0.1278449744,0.0794446915,0.2715919912,-0.0406669043,-0.3521394432,0.3497368991,-0.3610708117,-0.0431854911,0.1686316282,0.1574980021,0.1628735363,0.2707741559,0.4084943831,0.0136154192,-0.1261952072,-0.2134845555,0.2715884745,0.2755286396,0.2698619664,0.313046068,0.1287536621,0.2043976486,-0.0576743782,-0.0522325337,0.0049607134,-0.1004546061,-0.2025470436,-0.1476150304,0.0883112922,-0.3175346255,-0.019139152,-0.2643443644,0.1351879239,-0.2632873356,0.2045490295,-0.2654726505,-0.1398562044,-0.0968471915,-0.4157758951,0.3907619417,-0.2395300418,0.2784121335,0.0863355771,-0.2383735478,-0.4046698511,0.0646499246,-0.1899420172,-0.1168667153,0.3373421431,-0.3670263588,0.1029953063,0.1408559382,-0.3336954713,0.4816080928,-0.1656112522,-0.1645951122,0.2019170523,-0.3982416093,-0.0883902982,0.1549080759,-0.188340053,-0.0865180567,-0.010428749,-0.2035202384,-0.1064208746,0.1076328978,-0.1134156287,-0.2887963951,0.0965061337,-0.2495456636,0.0297890548,-0.1858461648,0.0634325072,-0.3321940899,-0.0946239978,0.1867458671,-0.3426503539,0.287134707,-0.1742215604,-0.3361068368,0.0613681041,-0.1169488579,-0.1123993099,0.0519327298,-0.1871530414,-0.1176806837,-0.2642972469,-0.5564438701,0.3457522392,-0.0687346905,0.1330453455,-0.2591458559,0.1160944849,0.309104681,0.025470769,0.0253795292,-0.3204879463,0.032914076,-0.0702275261,-0.0610334761,0.5930926204,-0.2280661911,0.0729788989,0.1565073133,0.1225825325,0.4253528416,0.05936886,-0.0481373966,0.0182672143,0.1644159555,-0.1071258932,0.1683371216,0.2077216208,-0.0976899341,-0.1563350707,0.4705615342,0.3646205962,-0.0054983087,-0.1388783753,0.1833245754,-0.0293371677,-0.2661377788,0.1739902049,0.2995582819,0.3880095184,-0.1546847373,0.1660907865,-0.1512525976,0.0750182271,-0.2321189195,-0.1701999754,0.3782023191,-0.222284779,0.3915688396,0.3863414228,-0.3913907409,0.2822634578,0.3291309774,-0.0832038671,0.0201571546,-0.2512384355,0.0599982329,0.1080634445,0.3921416104,-0.2249241769,-0.4162837565,0.4402129054,0.1864141077,-0.366956979,-0.0029857776,-0.4216745496,-0.6974324584,1.1738078594,0.2573472857,-0.1538091898,0.0064694877,0.3078435957,0.0786396489,0.0139561025,-0.2059752494,-0.1701974571,-0.1316949129,-0.2217651457,0.0581703819,0.0521628074,-0.0452029519,-0.2391819954,-0.1456112415,0.4471742213,-0.1733315885,-0.0507846475,0.3019070029,0.2329541743,0.035990648,0.1739749312,0.1901964098,0.2399755269,0.2120049745,0.2777101398,-0.0766570941,-0.4299620986,0.1638704687,-0.2015994787,0.5481743217,0.4593533576,0.0066290433,-0.0979801267,-0.5291199684,-0.027972864,0.304469198,0.1693527699,0.2401064932,-0.0530310348,-0.3426640034,-0.0914302468,0.1091333255,0.1910518408,-0.3062942922,-0.0053372565,-0.2550165355,-0.0913892314,0.0554537587,0.2086657584,0.9420134425,0.2601419687,0.1926536411,0.1376910955,-0.408513546,0.6010934114,0.0820223913,0.1172550917,-0.2683610618,0.1279535741,0.0676626489,-0.1992436349,-0.123735182,-0.2735461593,-0.2983320653,-0.1589807868,-0.3140909076,-0.1726135015,-0.1010144874,0.3583068252,-0.1642761081,-0.0845822468,0.4812367558,0.0399217792,0.3590171337,0.0060372469,-0.1238963604,0.0394354947,-0.3739419281,-0.0988585427,-0.5869011879,-0.247688219,-0.0175372418,-0.0215941872,-0.1326886266,-0.6107324362,0.0473341383,0.2841534615,0.2572387755,-0.0001589937,-0.1271247864,0.1364340186,0.3836013079,-0.0967625976,0.3567716777,0.1492606848,0.0368875377,-0.1886469275,-0.2704183459,-0.0527099371,-0.0406093337,0.0104834558,0.1263320148,-0.0127741275,0.2184476852,-0.5251654387,0.1823174953,0.2948208749,0.0213305354,-0.2041355968,0.1176783964,0.1039169058,-0.1102819815,0.0116428509,-0.3245984614,-0.3441010118,-0.1030216292,-0.1764565259,0.2090327591,-0.0325292088,-0.0658808947,0.2634502947,-0.1613495201,-0.0393638723,0.4047774374,-0.4816966355,-0.1757545173,0.0467466228,0.3898824751,-0.4873671532,-0.2078991979,0.718190074,0.2839851379,0.1210317835,-0.2442059368,0.0776853487,-0.2190176696,0.2853497863,-0.1448173672,0.2672012448,-0.1714926064,0.1936895251,0.1794568598,-0.255846113,-0.2345598489,-0.0517657138,-0.0645457655,0.055714801,-0.0814134628,0.0005702286,-0.1619118154,-0.0869020149,-0.0464037843,0.290756762,0.148340404,-0.1066332906,-0.0454832129,0.1994403154,0.0568390563,-0.1714213789,-0.0616395287,0.2124747634,-0.1362291723,-0.1371076703,0.1286386102,-0.0290473681,0.1667297482,0.3370328546,-0.641823709,0.2252511978,0.0984827876,0.4276833236,0.2031524628,0.3347712755,0.1589771211,0.2815193236,-0.1367619187,0.0131743802,0.2245784104,0.0837199017,-0.1395095885,0.1632847339,-0.0118575972,0.2165140659,0.0882182717,0.1616093665,-0.0195057802,0.3819689751,-0.1094944552,-0.0760948882,0.2928072214,0.109441027,-0.1039118022,0.0777936354,-0.3227869272,0.1022258997,-0.1177077144,0.1059041619,0.133848682,-0.0717743263,-0.2860386968,-0.0957183093,0.2287203074,0.4395951033,0.0452027135,0.2318582386,-0.2078228444,0.3169747293,-0.1538719982,-0.0106305173,0.0769633502,-0.3803688288,0.3780266941,-0.147817567,0.0793881491,0.439311564,0.4490016401,-0.1216979846,0.5643333197,0.0937559903,0.1662644148,0.0777896941,-0.1747880578,0.076293759,0.3394047916,0.1429419219,-0.5593453646,0.0119872093,0.0098370444,0.050772775,-0.0907640979,0.0733439475,0.3212533891,0.0812146738,-0.1045710891,-0.0404206812,-0.0034834547,0.0848082528,-0.1408283561,-0.0825388804,0.2572428584,-0.1428916305,0.2666389048,-0.2509085238,0.2829680145,-0.1334726512,0.2678283453,-0.0321186334,0.37512362,0.3795365095,0.0596000515,0.2736057639,-0.1282398552,-0.0425588973,0.186337173,-0.0076870169,0.2494965941,0.3891302645,0.1700572968,0.2802572548,0.058299493,-0.0292889606,-0.1207242683,-0.0304327831,-0.0494100749,-0.0676256716,-0.1022972167,-0.0696648732,-0.2004576772,0.1851491183,0.1658856273,-0.3086848557,0.432611078,0.2365031838,-0.0868012086,-0.0161569342,0.1564302891,0.0106880656,-0.1097473651,0.0504861288,0.2927233577,0.3041372597,0.1310093999,-0.5701604486,-0.1223326027,0.1272026449,-0.1752165705,-0.5535442233,0.1471072435,0.0704303235,0.1053016409,0.1867408007,0.1335620433,0.0324815735,-0.1700442135,0.058994785,-0.0915264115,-0.1022744402,0.3059032559,-0.07940761,0.0153195774,0.1056592017,0.0515683815,0.0710842237,-0.1111142412,0.0873573124,-0.2075468153,0.2392421663,0.2646929026,0.2493785173,0.3220409453,0.6761343479,-0.1604974717,0.0959385112,0.0055761421,-0.0548025519,-0.0238487814,0.1650029719,-0.0799583644,-0.3598213196,-0.3013978004,0.087592341,0.3000230789,0.3268484175,0.31174317,-0.1694519818,-0.3693899512,-0.3254948258,0.0781043544,-0.2393303961,0.1011588201,0.3707349002,-0.173439011,-0.2449704856,0.0340408683,-0.1015891433,0.4346277714,-0.1382065713,0.1127895638,-0.3447733521,0.1693844795,0.6491170526,-0.1134441197,-0.6402673125,-0.2678227723,0.3487847745,0.2495373338,-0.2667502463,0.1410763264,0.2826908827,-0.1879063398,-0.0119995754,0.8816466928,-0.1740211993,-0.2488283366,-0.2054694146,-0.0559658371]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1983","title":"The size of CoNLL-2003 is not consistant with the official release.","comments":"We should mention in the Conll2003 dataset card that these lines have been removed indeed.\r\n\r\nIf some users are interested in using these lines (maybe to recombine documents ?) then we can add a parameter to the conll2003 dataset to include them.\r\n\r\nBut IMO the default config should stay the current one (without the `-DOCSTART-` stuff), so that you can directly train NER models without additional preprocessing. Let me know what you think","body":"Thanks for the dataset sharing! But when I use conll-2003, I meet some questions.\r\nThe statistics of conll-2003 in this repo is : \r\n\\#train 14041 \\#dev 3250 \\#test 3453\r\nWhile the official statistics is:\r\n\\#train 14987 \\#dev 3466 \\#test 3684\r\nWish for your reply~","comment_length":73,"text":"The size of CoNLL-2003 is not consistant with the official release.\nThanks for the dataset sharing! But when I use conll-2003, I meet some questions.\r\nThe statistics of conll-2003 in this repo is : \r\n\\#train 14041 \\#dev 3250 \\#test 3453\r\nWhile the official statistics is:\r\n\\#train 14987 \\#dev 3466 \\#test 3684\r\nWish for your reply~\nWe should mention in the Conll2003 dataset card that these lines have been removed indeed.\r\n\r\nIf some users are interested in using these lines (maybe to recombine documents ?) then we can add a parameter to the conll2003 dataset to include them.\r\n\r\nBut IMO the default config should stay the current one (without the `-DOCSTART-` stuff), so that you can directly train NER models without additional preprocessing. Let me know what you think","embeddings":[-0.0790517628,0.0053060674,0.0227088984,0.1767311543,-0.1750467122,0.0091935722,0.1926335841,0.1599837542,-1.0088423491,0.0919716656,0.1161217391,0.0397600345,-0.0169360023,0.1187621579,-0.0615689158,0.3097825348,0.0787528902,0.0600117221,0.1479182392,-0.1191559359,-0.3082509637,0.358076632,-0.3310672045,0.0405783467,-0.5341429114,0.1651314795,-0.0984805524,0.1878358722,-0.434206605,-0.4354428649,0.3668316007,0.1122712269,0.4130791724,0.0164174419,-0.0001118863,-0.1192543954,0.2520217597,-0.1412217766,-0.076693818,0.1694634557,-0.050209824,-0.2738068402,0.0147658866,-0.0273164678,-0.0503348671,0.2545052767,-0.1091153324,0.2557539642,0.0737720877,0.3681807518,0.2139769644,0.0384481065,0.1702184826,-0.0514829643,0.3541508615,0.2172906995,-0.0525220372,0.3811748326,-0.19514893,0.1040109694,-0.2099635452,0.2688074112,0.1873317361,-0.2268101722,0.3166727722,0.1540002376,0.2569654584,-0.2214953005,-0.2121857852,0.3500757813,0.2903466523,-0.2142493129,-0.1941875219,-0.2296766937,0.2725404799,-0.457477659,0.060404595,0.0433200635,-0.0058561708,0.0137385046,-0.2138532996,-0.5553431511,-0.292645514,0.0585920289,-0.3667196333,0.3881666064,-0.115728572,-0.1022905111,0.1817389131,0.0599916689,0.6260008216,0.010335274,-0.1692814678,-0.1431291848,-0.1197364405,-0.2859581709,-0.0559346043,0.0654557869,0.2638859153,-0.0636547729,-0.1156925261,-0.4510706365,0.2121991366,-0.248263225,-0.0093579367,0.113195397,0.2837645411,0.239855811,0.1120828465,0.2224410921,0.2796357572,0.1129842103,-0.2029008418,-0.0597663336,-0.2434735,-0.0676073134,-0.0179156046,-0.2810794711,0.1291228384,0.1988326162,0.0639458224,-0.0796382278,-0.1012124643,0.018493589,0.0820036754,0.4512069225,0.0034321337,-0.1984367222,-0.084528923,-0.3352283835,-0.1166150644,-0.2821266055,-0.0829380751,-0.0459283814,0.0355485119,-0.2012436688,0.1888842732,-0.1283838898,0.2089136243,0.3528966904,0.1774379164,-0.1730100662,0.1446496844,-0.0073264125,-0.2975330353,-0.1950110346,-0.058565557,0.1618616581,-0.2799871266,0.0485449918,-0.0993474349,-0.3043549359,-0.4162815511,0.1212233454,-0.1019023657,0.0118265729,0.6224346757,0.1664651483,0.0826453716,-0.1743585914,0.0671644732,-0.4080204666,-0.3539324403,-0.1758713424,0.3193433881,0.2330732197,-0.2300450057,-0.1173345223,-0.0911655352,-0.1505451947,0.1310708076,0.0084172813,-0.1777452826,0.1105761901,-0.0720570385,-0.2215397954,0.2518273294,-0.3360654116,-0.3417873979,0.1112436801,0.2198238075,-0.3266861439,-0.0330128595,-0.0516684912,0.3488057852,-0.0335538462,0.0222429186,0.0991626233,0.0220309403,-0.2132713199,-0.3179375231,-0.2815461457,-0.3954845965,-0.1032078117,0.1473606974,-0.1974997967,-0.352946043,0.0315568335,0.3055689931,0.1924332082,0.0112376949,0.3021039963,0.3747840822,-0.1904214919,0.0783938915,0.2963239551,-0.6002673507,0.0333616734,-0.0880930722,0.2312198132,0.729577601,-0.0865769759,-0.0486960374,-0.1075349674,0.1938242018,-0.210212782,0.1165840477,0.0602159649,0.2370423377,0.0884938166,-0.1364407688,0.0785467029,-0.2363155037,0.1135212928,-0.2278850824,-0.1792073548,-0.0897696763,-0.0024871451,0.0092414422,0.0679097846,-0.1765299141,0.2218803465,0.0440063402,0.3089011312,0.3044915795,-0.3136169314,0.0809995681,-0.0011140064,0.007004465,0.0568197891,0.0082929349,-0.271427691,-0.3744516373,0.2555375993,-0.0846404284,0.0808002502,0.465287447,-0.069869563,0.2097191215,-0.0009587118,-0.1454469562,-0.0171781667,-0.0766737312,-0.5671272874,-0.2756134868,0.3597340882,-0.0090087485,0.3806385398,-0.3556628525,-0.0670061111,0.2865484357,0.0019143085,-0.4107649624,0.4418158531,-0.3745262027,-0.0814533606,0.1061326787,0.4837848544,0.0947721824,0.2629711628,0.2847138047,0.1773799807,-0.0606921576,-0.0977045298,0.3825392127,0.0797182918,0.6006683707,0.2228209972,0.0153301898,0.0987114608,0.0872502849,-0.163764134,0.0181887634,0.1527938843,-0.2198047489,-0.2572900653,0.0601755977,-0.067753002,-0.0747493133,-0.0177049655,-0.3243751228,-0.1396196932,0.0603800006,-0.2385694534,0.071347326,-0.0294600278,-0.5579313636,0.3389033079,-0.2254012674,0.2155260295,0.1509417444,0.08237914,-0.5002079606,0.1482927799,-0.1069184691,-0.3531935811,0.0481997505,-0.418269217,0.0668951496,0.2385117561,-0.3264962733,0.3696917892,0.1347669512,0.1904295087,0.2230056673,-0.3871462345,-0.1640752703,0.2841857672,-0.1811352521,0.0183174349,-0.2151919305,-0.0226862896,-0.3398996294,0.257486701,-0.1655430198,-0.4132044315,-0.0645678937,-0.1689619273,-0.0013992144,0.0684766918,0.0359239392,0.0072923438,-0.0268487707,-0.0757276416,-0.400631249,0.2760815918,-0.0114549706,-0.2453077883,0.255941987,0.0025324323,-0.1818640828,-0.0200049113,-0.1705810428,0.1318047941,-0.1081926003,-0.6059451103,0.2363824993,-0.0703084022,-0.0106604658,-0.1274934262,-0.0597981513,0.3178860545,0.2325284183,-0.0260879043,0.094920896,0.0383140296,0.2087802291,-0.1287229657,0.4611245692,-0.0976861566,0.0773097128,0.0459611267,0.3247405887,0.2846692502,-0.2928981781,-0.1263490021,0.0874618664,0.3358834684,0.0002202532,0.2209355682,0.0259134062,0.0353934281,0.0152609376,0.3652918637,0.2062508017,-0.1151993647,0.0174175948,0.2381312996,0.192542389,-0.1675613225,0.0327136554,0.2342021167,0.6284465194,-0.1660373062,0.0329455733,0.0276785735,-0.3331246674,-0.1734179854,-0.1224722862,0.1904447973,-0.1301648766,0.1776046306,0.5518295765,-0.5440958142,0.2639871538,-0.0514188036,0.0805388018,0.1151379943,-0.2817246616,-0.1604717821,-0.0060485834,0.4679948092,-0.0481595285,-0.4125792384,0.3504716754,0.1653051227,-0.4734293818,-0.1215546057,-0.2587657273,-0.5339019895,1.0868715048,0.5440798998,-0.0359400064,-0.2314255983,0.3389778435,-0.0286154225,-0.2221271843,-0.1624898016,-0.29480353,0.0752848238,-0.0209757276,-0.0550567769,-0.2603350282,-0.1130757183,-0.2074436098,-0.0102987727,0.2051171213,-0.1974098235,0.0964840353,0.2191097587,0.3813682199,0.1402325183,0.2431928664,0.0593849085,0.4666581154,0.4851014018,0.1532856226,-0.0573616549,-0.271948874,0.2822732329,-0.1091230884,0.5299078226,0.3107827902,-0.0572631583,-0.0213020556,-0.2038786709,0.0771388859,0.0760964751,0.2159231156,0.3242283463,-0.2165441066,-0.3204391003,-0.1204326451,0.0201634392,0.2061267793,-0.2027852237,-0.4543347061,-0.1485805064,-0.0730137154,-0.1013201252,0.4920628965,0.8532764316,0.0660228431,0.2991860509,-0.0171369966,-0.2874809802,0.4888044,-0.0867125392,0.1391884536,-0.3684087098,-0.007965561,0.07105048,-0.1088033989,-0.1466741413,0.0864087716,-0.1399547309,0.0994460881,-0.1318170875,-0.2483180463,-0.0121796895,0.3627187908,0.0624589063,0.0470756777,0.6486375332,0.0402388535,-0.0018705219,-0.1850624979,-0.0330131799,-0.0668512732,-0.1162806228,0.0411882736,-0.1368861049,-0.1115402579,0.0805262849,-0.0213483386,0.1995276809,-0.4030072093,0.0898965448,0.5373719931,0.1734157801,0.2524736524,-0.2667478323,-0.0780738965,0.2593308389,-0.1200405881,0.3983041942,0.1348486245,0.0571173877,-0.220069766,-0.2907457352,-0.0389296748,0.1188781857,0.1148881391,0.0022322631,-0.3247589171,0.1397278011,-0.5065875053,0.1417088658,0.0707658976,0.0411731824,-0.123154372,0.2083688378,-0.1243158057,-0.0224427339,0.0308007803,-0.141569078,-0.387067318,-0.0869079232,-0.1063449979,0.2709780633,0.2768917084,0.1275358945,0.1496863216,-0.0527111366,-0.0856909603,0.2901490927,-0.1613321602,-0.1506538689,0.1519982368,0.4072241783,-0.0559113622,-0.095918946,0.2289100289,0.2441454232,0.0138013652,-0.0843919441,0.2946996987,-0.4281079173,0.1821244955,-0.2010680288,0.4821609557,-0.1125019044,-0.0547959469,-0.0732149854,-0.2601736784,-0.3238525093,-0.0153868534,-0.1974707991,-0.0878957435,-0.2802780867,0.0168674756,-0.2452738434,-0.0561282076,0.0320812277,-0.0878540948,-0.0835087001,-0.1851974279,-0.0691233426,0.1099173427,-0.0350423865,-0.1135477945,0.1861979365,0.0987315252,-0.3318326175,-0.0568747967,-0.0724851191,-0.2056240737,0.2165595293,0.605278194,-0.4788680077,0.0174339861,0.2167275101,0.3572171628,0.0631686896,0.2910677195,0.191824764,0.5138154626,-0.0972709805,0.016756488,-0.0803737193,0.1394071728,-0.3710766137,-0.0375037082,-0.0178774297,0.0828925669,-0.0480662435,-0.0041854456,0.1387234479,0.114427112,-0.0810982063,-0.1132284701,0.2425559014,0.2390527576,-0.2424698621,0.0384459719,-0.1725450307,0.058764331,-0.0347023904,0.1650694609,0.165233165,-0.0666311607,-0.1837201715,0.1502304226,-0.094149746,0.4256555438,-0.014578538,0.2439222485,0.0069968603,0.3048895895,0.3469516039,-0.1299915165,0.1580550671,-0.0452580526,0.519287467,-0.0432184748,-0.1320542842,0.30897668,0.3415551186,-0.4437057674,0.4573957324,0.1989684701,0.112117745,0.0746014789,-0.0678187609,0.0970053598,0.4011941254,-0.1341477782,-0.5298088193,0.0278158896,-0.092799753,-0.1768021435,-0.1649041027,-0.0588298254,0.402444303,0.3872849345,-0.0608082525,-0.1517052352,-0.0667159036,0.0923594236,-0.2444111109,0.0574362986,0.5005825758,-0.0221061464,0.0581260137,-0.2497326136,0.0316063687,-0.1127104312,-0.004712123,-0.1360365152,0.4472354054,0.2742931247,0.0504238978,0.1626672149,-0.1984322518,0.2732845545,0.0497829616,-0.144620046,0.1420491487,0.5019465089,0.218043834,0.1841220409,0.1261109263,-0.0163302496,-0.2222193331,-0.1227293238,0.0380658284,-0.0254142806,0.136661008,0.2436605394,-0.1785573959,0.1551506966,0.2969259322,-0.4654332399,0.3182471395,0.5749578476,0.1311012357,0.0354698896,0.0578068793,0.0711148307,-0.1768959463,0.0446700417,0.2089402974,0.1623996198,-0.1082178503,-0.5213893056,-0.0811594874,0.2729403675,-0.1001609936,-0.164581582,0.1313474476,0.1835096478,0.077796191,0.0132117905,-0.1120468006,-0.0527082421,-0.43847543,0.1809064895,-0.1560251564,0.020243505,-0.1199271679,-0.005276063,0.0289472211,-0.1149220094,-0.0118543366,-0.1289931089,-0.0810680911,-0.0968911797,0.1312231421,0.384447515,0.43712008,0.2869357169,0.2150786072,0.3534070849,-0.059993051,-0.223243624,-0.0786948651,0.2101556808,-0.2013532072,-0.0585583635,-0.209385857,-0.2135195583,-0.3199549615,-0.1325805038,0.1056051701,0.2887854278,0.2079716176,-0.0856634378,-0.3484726846,-0.1450233012,0.1012752652,-0.261372149,0.2602857649,0.2764100134,-0.1039939076,-0.2661340535,-0.0656880066,-0.0074326103,0.3319243193,0.0968856141,0.2977837622,-0.4783828557,-0.0235057417,0.274410218,0.0364435241,-0.4960820377,-0.0709709898,0.4799347818,0.1348759532,-0.2606063485,0.3722867966,0.1307561398,-0.0844974816,-0.1351376027,0.5800472498,-0.3410003483,-0.3498989642,-0.5202119946,-0.157811448]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1983","title":"The size of CoNLL-2003 is not consistant with the official release.","comments":"@lhoestq Yes, I agree adding a small note should be sufficient.\r\n\r\nCurrently, NLTK's `ConllCorpusReader` ignores the `-DOCSTART-` lines so I think it's ok if we do the same. If there is an interest in the future to use these lines, then we can include them.","body":"Thanks for the dataset sharing! But when I use conll-2003, I meet some questions.\r\nThe statistics of conll-2003 in this repo is : \r\n\\#train 14041 \\#dev 3250 \\#test 3453\r\nWhile the official statistics is:\r\n\\#train 14987 \\#dev 3466 \\#test 3684\r\nWish for your reply~","comment_length":45,"text":"The size of CoNLL-2003 is not consistant with the official release.\nThanks for the dataset sharing! But when I use conll-2003, I meet some questions.\r\nThe statistics of conll-2003 in this repo is : \r\n\\#train 14041 \\#dev 3250 \\#test 3453\r\nWhile the official statistics is:\r\n\\#train 14987 \\#dev 3466 \\#test 3684\r\nWish for your reply~\n@lhoestq Yes, I agree adding a small note should be sufficient.\r\n\r\nCurrently, NLTK's `ConllCorpusReader` ignores the `-DOCSTART-` lines so I think it's ok if we do the same. If there is an interest in the future to use these lines, then we can include them.","embeddings":[0.1301080137,0.0962899402,0.0394509584,0.0533670038,-0.2169309855,0.0075207311,0.1542738676,-0.0001621469,-0.9731163979,0.0672688559,0.2056128085,0.0701422319,-0.0513292663,-0.0842716172,-0.0803269595,0.3419833183,-0.0675743446,0.327860564,0.2285470366,-0.1227636263,-0.3668461442,0.2775174081,-0.274977982,-0.0082388809,-0.393466413,0.0806119815,-0.1419335157,0.2710108459,-0.2415265888,-0.532017827,0.3974419236,0.2227784991,0.0772376209,0.0288814083,-0.000115512,-0.2126500458,0.3872515857,-0.215965271,-0.2099744678,0.4730767608,-0.1129044816,-0.7660289407,-0.044372905,-0.4447828233,0.0319344439,0.5121642351,-0.0817523897,0.1743455529,0.2542611659,0.161358431,0.1676934958,0.2393181622,0.2375080734,0.1333176941,0.3475002348,-0.02496868,-0.1696096063,0.5508455634,-0.2081135809,0.1203334332,-0.2780005038,0.3548775017,0.1467530131,-0.2336138338,0.1369725913,-0.0380336381,0.2645008266,-0.0674744695,-0.1419992447,0.2909509242,0.29439044,-0.2650747299,-0.1194988489,-0.5421310663,0.267749697,-0.4985669255,0.0995380431,-0.0365589336,-0.0151115526,0.017058922,-0.2545989752,-0.4198880196,-0.3282954991,0.1200169846,-0.1598795503,0.418084234,-0.0272401609,0.0301428102,0.1538592726,0.1586638987,0.5796195269,0.0767197609,-0.2444941103,0.0215668604,-0.0377822928,-0.4418598413,0.1902304292,-0.1309911013,0.1642335951,-0.2104474902,-0.1204381213,-0.4199705422,-0.0901070833,-0.0593098961,0.0193503629,0.0543407723,0.4128799736,0.3364706039,0.2408224642,-0.100061059,0.4114401639,0.1436223984,-0.252967,-0.1811205596,-0.1239445433,0.1675828397,-0.2642978132,-0.2406897545,0.1624811888,0.1565834135,-0.0349364914,-0.073528789,-0.0618176274,-0.0178566203,0.2747400701,0.3580447435,0.0686169714,-0.0945587531,0.0771590322,-0.2723608911,-0.0594635494,-0.3487521708,0.0059081572,-0.017153291,-0.1843848675,-0.0562592819,0.3026262522,-0.3127999902,0.2598237395,0.3169444501,0.1503162235,0.012402623,0.0943745896,0.0488491356,-0.0882604346,-0.1484766304,0.0212390274,0.1710014939,-0.3730175495,0.2206998616,0.107878916,-0.3871753216,-0.4981610477,0.0574468337,-0.3294167519,-0.0661147758,0.5692340732,0.0615522526,0.1533857584,-0.2984735668,0.2479482889,-0.1679144204,-0.1764439642,-0.1377005279,0.2795294523,0.0724272355,0.0772195309,-0.2327023596,-0.2192714065,-0.1420802176,0.2104848772,0.0464700945,-0.1911343485,0.1837269217,-0.0846440643,-0.0169071853,0.4908646345,-0.3189494014,-0.2493038177,0.4301478565,-0.0368908755,-0.219440043,-0.0988788158,0.1389304399,0.1706574112,-0.1958432049,-0.0332034044,0.1517599076,0.0310926605,0.1289447546,-0.3339414895,-0.2219436765,-0.2317954898,-0.2167307884,0.163366586,-0.2227857858,-0.2178170532,0.2956934869,0.4338988662,0.0595176332,-0.069379814,0.2499163151,0.3685540557,-0.2128676474,-0.1087667048,0.4909599721,-0.3863342702,0.0884430408,0.0807879269,0.2146434635,0.747834444,-0.265327245,-0.2305809855,-0.1599819362,0.2598367035,-0.0037722851,0.1104198247,0.1521407664,0.1595448852,0.2314088047,0.1487793028,0.0730554014,-0.1615217477,0.0371756814,0.0772234946,-0.1853340864,-0.0302614309,-0.1165647879,0.1630454957,0.3133796751,0.040984273,0.2251074761,-0.1252541244,0.369248271,0.0459581986,-0.0481309481,0.0709815845,-0.166158542,0.0461778529,0.267226845,-0.0127223209,-0.3932291269,-0.4152544737,0.0613912866,0.0116798775,0.0987932384,0.5511188507,-0.243895039,0.0746922344,-0.0841133595,-0.1902338266,-0.1499232352,-0.2079464793,-0.2950230241,-0.2136440873,0.3231470585,-0.2128845304,0.153433755,-0.3101692796,-0.0633996874,0.4416494071,-0.185323745,-0.1847084761,0.3180700541,-0.2418947369,-0.0890740305,0.248134926,0.2256144583,0.0159505494,0.2709316313,0.283038944,0.1357545555,-0.2408515513,-0.1283157915,0.1752324104,0.2388232797,0.4394446611,0.239195779,0.0922869816,0.114813894,0.0340718515,-0.0045291381,-0.0054153185,-0.0277312137,-0.2279874086,-0.3621948659,0.1602542102,-0.1504976451,-0.065082632,-0.3093796074,-0.2254503071,-0.3511163294,0.1018531621,-0.1388074458,-0.1369154751,0.1145197153,-0.5290244818,0.4189198911,-0.0392219238,0.2031296194,0.2047443986,-0.2497678101,-0.5104724765,0.0853452086,-0.0410059057,-0.306143254,0.3523700833,-0.4794403911,0.0280888788,0.2486198843,-0.4806945622,0.3634321988,-0.0830034241,-0.1581347138,0.2018323392,-0.205129683,-0.3833319247,0.1208054572,-0.2449854612,0.2246058285,-0.153999567,0.000564967,-0.0720465705,0.2123025805,-0.4482307732,-0.1638085991,0.1669665575,-0.3378913999,0.1977341026,-0.0912566856,0.0160743296,0.0817922354,-0.1622832417,0.2011844367,-0.181869328,0.1154790744,-0.1136728451,0.0167072359,0.2259347588,-0.081131734,-0.1662234217,0.0699585751,-0.2387206852,0.0106036831,-0.3078627884,-0.7261509299,0.0743746087,0.0595251843,-0.0601580255,-0.1150825992,0.0376050062,0.2131718397,0.1127520353,0.052319292,-0.0855741352,-0.1226233169,0.1621295065,0.0945110768,0.5319262743,-0.222887814,0.3096297979,-0.0022927178,0.159055993,0.2853764296,-0.0007306936,-0.2252567559,-0.0015945244,0.2930024862,0.258847326,0.3286583722,0.3251229525,0.1030992642,-0.1379554272,0.4538410008,0.0773999467,-0.1460178047,0.0553967543,0.1441570669,0.0169384107,-0.4341693819,0.0573548153,0.1168065369,0.5626435876,-0.2184776664,-0.0084214807,-0.1097343341,-0.1650135517,-0.2341182828,-0.0293201543,0.3005231619,-0.0291775092,0.2054887116,0.2458188683,-0.563102603,0.2150882781,-0.0362782069,0.0829903111,0.1746266484,-0.253567487,-0.2356996238,0.1044344604,0.0360960886,-0.0938173234,-0.2063310146,0.3459343016,0.3252570033,-0.2910985053,-0.2175762057,-0.2208033949,-0.2699076533,1.2819186449,0.4255770147,-0.3862962723,0.0430352725,0.3762326241,-0.148317799,-0.1491270363,-0.2190731764,-0.3982902467,-0.1763947606,-0.1942517608,0.1178858131,-0.0154517284,-0.0890326276,-0.1234610975,0.0912798569,0.320541054,-0.1780037731,0.0968419686,0.392963171,0.2283859253,0.5601581335,0.1672639102,0.0511213206,0.3188781142,-0.055344861,0.1946551204,0.1176549867,-0.0828341916,0.2511397898,-0.3525596857,0.4620984197,0.3488352001,0.131316632,-0.1007162705,-0.1621301174,0.0177413821,0.1748205274,0.3541529477,0.1557814926,-0.3248019516,-0.3251659274,-0.1433885545,0.2663400769,0.068508409,-0.1487399787,-0.4119541943,-0.1902715564,-0.2139978707,-0.1538045257,0.2392158061,0.8080552816,0.123983331,0.170427084,0.3559841514,-0.2942585349,0.4923358262,0.0814508274,0.0388295576,-0.3318097591,0.2694885135,0.1430426389,-0.041259408,-0.2466946244,-0.0483891107,-0.4165376723,-0.0031493849,0.0022391786,-0.2413029224,-0.0930721164,0.3237605095,0.0527842417,0.0639535487,0.2752559781,0.0389373749,0.1326100826,-0.0510024093,0.0120956246,-0.2595166266,-0.1758071333,-0.1770463735,-0.353235364,-0.3114776909,0.1412600428,0.0185092147,-0.0236557256,-0.4095527232,0.1921205968,0.2770200074,0.3589107096,0.2970498204,-0.2169280946,-0.1093113571,0.2528967559,-0.2034264952,0.3574486375,0.2469755709,0.0348186567,-0.283064574,-0.195648402,0.1517183185,-0.1001775414,-0.0376817733,0.1460778862,-0.1781592816,0.014207785,-0.3417791128,-0.0354531631,0.3411011994,-0.0877546668,-0.0100297425,0.1375862062,0.093252331,-0.0353010669,-0.1563230455,-0.2820822001,-0.2060042918,-0.0444565453,-0.3939485252,0.1372539401,0.182411626,-0.0218914114,0.0423907898,0.0228498392,0.0014152575,0.1289570183,-0.4131377935,-0.0770957544,-0.0668571517,0.5177926421,-0.2952233255,-0.2913728058,0.2481763959,0.1947964728,0.2999889851,-0.1022827029,0.2632980049,-0.4071258008,0.2513200343,-0.2720452845,0.4459058642,0.0000086172,0.4026787579,-0.0745483041,-0.1872904599,-0.2772979438,-0.2720987499,-0.2297486067,0.0107978396,-0.1879491955,0.0458862484,-0.1456429809,-0.1231525615,-0.0274072047,0.1666630805,-0.0020089617,-0.1085648313,0.0261020474,0.1678994596,0.0458098203,-0.0362757631,0.2290015817,0.001738635,-0.1395778358,-0.0205962602,0.0394503437,-0.3833727539,0.1314447224,0.6604422331,-0.5886901021,0.417170763,0.2511737049,0.2564392388,-0.0192744564,0.3532114625,-0.1148343533,0.542819798,-0.0046753651,0.1038530841,0.2478196323,0.0658432022,-0.3922309279,0.1441467851,-0.0034275728,0.1552131027,-0.0760343745,0.004132567,-0.0537703782,0.0722347125,-0.2965034246,-0.1990163922,0.2163131088,0.1753441244,-0.2990163267,0.1887123734,-0.1672668755,-0.0348735228,0.1587667465,0.1492211819,0.0316303,-0.1084925979,-0.1870098859,0.207992211,-0.115156047,0.389423728,-0.0186801199,0.2622984648,-0.0982708558,0.2752375007,0.0979673639,-0.1320033073,0.2643313408,-0.4030513763,0.5268942714,-0.1661987156,-0.1249707714,0.2696351111,0.1843376011,-0.2103803307,0.3639319241,0.3102311492,0.0992766768,0.252815783,0.0409560427,0.1101134047,0.1543637067,0.1772454381,-0.6114675999,0.1304255575,-0.0217846669,0.0371783301,0.0025651734,0.2486207634,0.1974567622,0.1135745347,-0.0705641285,-0.0309344083,-0.3614548743,0.0356952064,-0.0779794231,0.2171178162,0.3358728886,0.1498039067,0.0945758447,-0.3691613972,0.0371608511,-0.1639917195,-0.0264079962,-0.1165779755,0.423006773,0.3049890399,0.0142477183,0.1797245145,-0.0623151287,0.1979967505,0.1574591845,-0.1397766471,0.0717102289,0.3436504006,0.0809427872,0.2746039927,0.1012328267,-0.0663010925,0.040720392,-0.1917506158,-0.1818831712,0.0257629994,0.1480092555,0.1553262174,-0.2714482844,0.1351749748,0.1511494368,-0.2748745084,0.3268390894,0.3695696294,0.3335462213,0.0922205746,-0.1199246794,0.0460598357,-0.0518145114,0.1357112378,0.3351663351,0.1156705841,0.0591659695,-0.6472341418,-0.2615152299,0.2753867507,-0.1858984679,-0.6451829076,0.3044980168,0.0233297125,0.1042953655,0.0785895362,-0.1120036393,-0.1032504961,-0.4663717449,-0.1224790514,-0.1709792465,-0.1859328449,0.0332653373,-0.0808854774,0.0616828054,-0.0555320866,0.0894193128,0.0845132917,-0.0057154112,-0.0798985139,0.156445384,0.402807802,0.5088064075,0.2772645652,0.1935526431,0.503473103,-0.1609892845,-0.2187089473,0.0961429402,0.0153530836,-0.0406199209,-0.1719721854,-0.232699573,-0.0901059657,-0.2025837749,0.1903230995,-0.0134506952,0.2978093624,0.2321572751,-0.0934983864,-0.1128151119,-0.1320217103,-0.0002962068,-0.1756173223,0.0633368418,0.2933471501,-0.0990398824,-0.0328284428,-0.0693781525,-0.220418781,0.2743246257,-0.0955026299,0.1714579165,-0.2696417272,0.1478176117,0.0846032798,0.1020460352,-0.5734166503,-0.0021998198,0.4456374645,0.3907030821,-0.3774712384,0.2812387645,0.1921702772,-0.1185679361,0.0316979364,0.6229767799,-0.280760318,-0.5063766241,-0.4081284106,-0.2519994676]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1983","title":"The size of CoNLL-2003 is not consistant with the official release.","comments":"I added a mention of this in conll2003's dataset card:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc9796920da88486c3b97690969aabf03d6b4088\/datasets\/conll2003\/README.md#conll2003\r\n\r\nEdit: just saw your PR @mariosasko (noticed it too late ^^)\r\nLet me take a look at it :)","body":"Thanks for the dataset sharing! But when I use conll-2003, I meet some questions.\r\nThe statistics of conll-2003 in this repo is : \r\n\\#train 14041 \\#dev 3250 \\#test 3453\r\nWhile the official statistics is:\r\n\\#train 14987 \\#dev 3466 \\#test 3684\r\nWish for your reply~","comment_length":30,"text":"The size of CoNLL-2003 is not consistant with the official release.\nThanks for the dataset sharing! But when I use conll-2003, I meet some questions.\r\nThe statistics of conll-2003 in this repo is : \r\n\\#train 14041 \\#dev 3250 \\#test 3453\r\nWhile the official statistics is:\r\n\\#train 14987 \\#dev 3466 \\#test 3684\r\nWish for your reply~\nI added a mention of this in conll2003's dataset card:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/fc9796920da88486c3b97690969aabf03d6b4088\/datasets\/conll2003\/README.md#conll2003\r\n\r\nEdit: just saw your PR @mariosasko (noticed it too late ^^)\r\nLet me take a look at it :)","embeddings":[-0.11778,-0.1490903944,-0.1239807233,0.4063589871,-0.1421683133,-0.1860256195,0.2695673406,0.0108112395,-0.976584971,0.129761517,0.0317089558,-0.0465303175,0.0295431539,0.1740378886,-0.073604621,0.1277457923,0.062856324,-0.1194887534,-0.0711038485,-0.161968857,-0.1469121724,0.4227974713,-0.3452403247,-0.0569811948,-0.4710490704,0.2424959093,-0.1226277947,0.0701174885,-0.5567559004,-0.323337853,0.4087007642,-0.0866105109,0.1769321561,0.3554422557,-0.0001040467,-0.1748574525,0.2246979922,-0.115742363,-0.0655561239,0.2404021025,-0.2004329711,-0.1920388937,-0.1547529548,-0.0302857477,0.0281113908,0.2732361555,-0.3089978099,0.2993923724,0.182216689,0.2430881262,0.2945328057,0.0809215605,0.2179147005,-0.1760448515,0.2989802659,0.1917587072,-0.1165719777,0.2260445803,-0.1307196319,0.2275835723,-0.0872837901,0.2860423326,0.334974438,-0.172430411,0.0849063024,-0.026191758,-0.0870816037,-0.0864435136,0.0954536274,0.3211065233,0.394913137,-0.3121786416,-0.4026528597,-0.191369161,0.0310513992,-0.3462634087,0.0563204102,0.0096388916,0.0461317413,-0.0584371202,-0.4982578754,-0.1443539709,-0.20755665,-0.0425582565,-0.3624979854,0.4512245059,-0.2420071065,0.0177786704,0.1008713171,-0.0527409203,0.4142463505,0.0274848901,-0.0922444984,-0.0860951245,-0.2407745719,-0.3211268485,0.1106194928,-0.126775369,0.4134696722,-0.1761906147,-0.3213562965,-0.4195230305,0.0159686469,-0.3341854215,0.1879396141,0.2340653241,0.0104155121,0.1668908894,0.1190471947,0.0761188492,0.2723383307,0.0930758119,-0.244910121,-0.1001408249,-0.2591290176,-0.0199954938,-0.1423551887,-0.2868362069,-0.1039165035,0.0991119668,-0.0918708965,-0.1583966315,-0.0469875634,0.1526241302,-0.0079525858,0.4125003219,-0.1267403662,-0.0555189848,-0.0259355456,-0.2938326895,-0.2362737656,-0.306984812,-0.1196088865,-0.206977427,-0.1334817261,-0.1495285779,0.2941322625,-0.1133824214,0.2720679939,0.2048839331,0.0209798776,-0.131098941,0.0350976177,-0.0036285412,-0.2124439478,-0.2451636791,-0.0987363979,0.1680230498,-0.2560645342,-0.1284075975,0.057140518,-0.3249761164,-0.3823211789,0.1705315709,-0.1979279965,-0.0290262792,0.5316371322,0.0959364995,0.0505769923,0.1557330638,0.1403920799,-0.3408176601,-0.0699074641,-0.2292314768,0.3176356256,0.0836290792,-0.1907851696,-0.0787202716,-0.0876957774,-0.2094607353,0.1775773764,0.2158054709,-0.0621547401,-0.0934578031,-0.1083809659,-0.2343145609,0.1496353745,-0.3616794348,-0.4399799109,0.2209210694,0.1757685542,-0.2946114838,0.077080667,-0.0218824986,0.3208502531,0.0463332199,0.091138944,0.0918140337,0.0417531803,-0.0348821804,-0.1790225655,-0.3428456187,-0.3542283177,-0.1385772228,0.2041584253,-0.3445739448,-0.14810507,0.0730544552,0.3761892319,0.2262465805,-0.0707721785,0.2620154619,0.4964691699,-0.1420655251,0.0202060826,0.2618133724,-0.347964406,0.0389451049,0.1203958243,0.1105892137,0.7259976268,-0.1705368161,-0.2706593871,0.0465808474,0.2493647486,-0.0999445319,0.2031153738,0.0404588208,0.3650392592,-0.0040919506,-0.1104710177,0.1139745936,-0.0178981572,-0.02927568,0.1010706723,-0.1370583624,-0.001614105,-0.0653423741,0.1549784839,0.2403028458,-0.0411923267,0.1002099738,-0.0479528755,0.2280957848,0.3012150824,-0.2031024694,0.2046176642,0.042294316,0.0624803081,0.2254555225,0.0878704935,-0.4247793853,-0.2324251086,0.277583003,-0.1241131946,0.2256472111,0.4845538139,-0.2323309481,0.1195513457,-0.029378511,0.0251130182,-0.0848059431,0.0052104597,-0.3428599834,-0.2204592228,0.181563139,0.0575422943,0.2655125856,-0.3689698279,0.0382841416,0.3136286139,-0.0431068912,-0.3063026071,0.2688687742,-0.4913647175,-0.0822650492,0.2280891091,0.4648602009,0.1076152325,0.2788805068,0.4253003001,0.0682944357,0.0518131293,-0.0445777103,0.1996300668,0.0479048975,0.4186436236,0.1436415315,0.0831096172,0.0965783969,-0.0389881842,-0.0646341145,-0.0050440542,0.0042763432,-0.082438536,-0.2426883727,-0.090787217,-0.0100195725,-0.0121672619,-0.2176555991,-0.251437664,-0.1626991481,0.1214177981,-0.3009015024,0.0510753095,-0.0633108839,-0.3588641286,0.1681368649,0.0136897592,0.3135440052,0.2385151833,0.1388934851,-0.4799479544,0.2259171307,-0.1357045323,-0.3781635761,0.2379622906,-0.4166471362,0.0695397556,0.0819515511,-0.2997785807,0.3862543702,-0.1278282702,0.0064578187,0.1638875455,-0.3423283696,-0.2572751641,0.3563469648,-0.0730230883,-0.1017332077,-0.2386434525,0.0057295705,-0.303976506,0.0857309029,-0.2400321513,-0.2603770196,-0.0834300369,-0.0932432711,0.1943808496,0.0187601596,-0.027861258,0.0240296163,0.1147530302,0.15723674,-0.2134676427,0.1495198756,-0.2803869545,-0.380674988,0.1139078438,-0.0351745225,-0.2225524038,0.0481997281,0.0305072814,0.0931165442,-0.1576381326,-0.6883491874,-0.0888285488,-0.157608673,0.1290762424,-0.1252102405,0.1556968093,0.2230455875,0.0539965928,-0.1211584955,-0.1104235202,-0.1563749313,0.041126851,0.094594799,0.6660491228,-0.1945737302,0.2059250325,0.2436729968,0.4627051353,0.4733084738,-0.2580979168,-0.0867398977,-0.0067991968,0.2599477172,0.0124997627,0.3047555983,0.348764509,-0.0693453774,-0.0428561866,0.3592760265,0.2311929017,-0.0112248566,0.1118726432,0.2965572774,0.1082370654,-0.0736093596,-0.0059637553,0.4028126299,0.4447231293,-0.1062416583,0.1473757625,-0.1348290741,-0.1046175584,-0.1210756227,-0.2605172098,0.0995492414,-0.1071474925,0.251593709,0.3448478281,-0.6651976109,0.3250333965,0.051977247,0.026449278,0.0646591336,-0.1042145938,-0.1022303775,0.084546268,0.4017082751,0.0158036109,-0.1736900806,0.2601073384,-0.0285712872,-0.3924729526,-0.0987120569,-0.4133390486,-0.4186163843,1.0134056807,0.4779088199,-0.1989348829,-0.0723821148,0.2843052149,0.1422549933,-0.0698135942,-0.155986771,-0.330398649,-0.1588707417,0.0416567065,0.0585860275,-0.023408873,0.006052169,-0.1393432319,-0.1141656712,0.2211963981,-0.2400444299,0.0303374007,0.2552057207,0.4171558321,0.2711313367,0.1991782188,-0.0221704096,0.359005332,0.23648265,0.3872633576,0.079674162,-0.2558588684,0.3316531181,-0.0051721246,0.4434928894,0.3573241234,-0.0277478267,-0.1808556914,-0.2613324821,0.1665480584,0.0823952034,0.2123363167,0.3883033693,-0.0115759121,-0.2462442964,-0.1123891175,0.1828767657,0.1039711982,-0.0659534633,-0.2461930364,-0.0813808665,0.0203258209,-0.0806612074,0.2885064185,0.8127256632,-0.0819903985,0.0596108623,0.0411756635,-0.3376258314,0.2398923635,-0.0695431232,0.1148761064,-0.339015007,-0.0835401416,0.1362097561,-0.1137319058,-0.2418110371,-0.0780691579,-0.1553236097,0.0390689261,-0.1424754113,-0.0921170413,-0.0923305526,0.2410518974,-0.0133186253,0.1570474803,0.4033717215,0.2571757138,0.2467220575,0.0312094055,-0.0076873866,-0.0942278877,-0.2511580884,-0.031015778,-0.4506995976,-0.1496546715,-0.0415389165,0.0437069796,0.146395877,-0.4477825761,-0.0523128696,0.4075407386,0.1918005496,0.1762987375,-0.1778763831,0.1143680066,0.1589150876,0.1622078866,0.431904912,0.2148407102,-0.1464108378,-0.1885229498,-0.2969202399,0.137177676,0.0668575913,0.0508464873,0.0620935261,-0.183218956,0.0084642516,-0.4938527644,0.3007796407,0.0256836526,0.0652868673,-0.0152882645,0.2463904619,0.0747117326,-0.0519328229,0.0346720926,-0.0704200491,-0.436296314,-0.1606469005,-0.1092332974,0.1591437012,0.3684759736,-0.0709742904,0.1098434776,0.0327447578,-0.0309943091,0.2632474601,-0.349771589,-0.046795968,0.0411747061,0.0660474747,-0.2239822745,-0.1087350622,0.19834584,0.3862473369,-0.0009008233,-0.2026980519,0.1171668395,-0.3669370711,0.1662439108,-0.1405933946,0.4673709273,-0.2569172382,0.1729712337,-0.0228858236,-0.3501435518,-0.3971063197,-0.1263405532,-0.2659687102,-0.113525331,-0.3912788928,-0.1805749536,-0.1014028862,-0.017122196,0.1442002803,-0.0524010658,-0.1238125786,-0.2701585293,-0.0278828051,0.0554122739,-0.0799710229,-0.1237789243,0.0732189119,0.121066533,-0.3472322822,-0.0554115437,0.1460540444,-0.0395326279,0.226110056,0.4076474607,-0.5039833188,0.0725948587,0.2214102745,0.2120588869,0.3264290988,0.3990575969,0.0824617445,0.4527707994,0.1557212621,0.0685368106,-0.0174621753,0.0426176079,-0.2237725407,-0.0387073047,0.085318841,0.1677066684,0.1316862553,0.2200066745,0.2272943258,0.183295399,-0.1064093187,0.0482755005,0.1594007015,0.3050548434,-0.2743836641,0.1071229056,-0.2487889677,0.3690179288,-0.0323477611,0.1779422313,0.1711959988,0.0399644487,-0.0335226506,0.079400681,0.08340583,0.5254677534,-0.016351575,0.1120551825,-0.0185537804,0.4008262157,0.2523584366,-0.2061770856,0.1637972444,0.0349368565,0.333168596,-0.1194412559,-0.2856724858,0.2920555472,0.3565746546,-0.1682107598,0.334002763,0.1833136082,-0.0667129308,0.0179812405,-0.0859552175,0.1677901,-0.0120945722,-0.0016966613,-0.7779325843,0.0512743071,-0.0561568774,-0.1049496084,0.1208096147,-0.0815203115,0.1709187329,0.3347248435,-0.2347364277,-0.1270287037,0.0550198406,-0.0758108795,-0.1394138038,-0.0509179682,0.1089926586,0.204947263,0.0740080848,-0.2799779177,0.2123621702,0.0756098703,0.1040028334,-0.1048165038,0.4057820141,0.2850217521,-0.0817155913,0.0166609455,-0.0904394984,0.2010720223,0.1049898341,-0.2600875199,0.0940877125,0.3385106325,0.0787675679,0.2597364783,0.2011125088,-0.134329319,-0.2416916341,-0.2874895036,0.0504148416,-0.1882928312,0.1537918299,-0.1491675675,-0.0916518718,0.0543052815,0.1305515766,-0.4923566878,0.2276897877,0.3193547428,0.0556115806,0.1301309168,-0.0193793848,0.1160944849,-0.2794567943,-0.0227173045,0.2474236935,0.2013018131,0.0879413933,-0.581574738,-0.0923746973,0.3105400205,-0.1654672325,-0.3447998464,0.0532616861,0.3007673025,0.0372680761,0.037697427,0.1974076182,-0.1079811826,-0.2613868117,0.1192395836,-0.3053136766,0.0157389399,0.0029159468,0.0249562003,0.1774373949,-0.0161397811,0.0565101728,-0.1310533434,0.0159718487,-0.1169697419,0.0239997841,0.3244971335,0.2114358544,0.3351795077,0.1502851993,0.4356203377,-0.1151320413,-0.3086133003,-0.1074879766,0.0539789535,0.043238949,-0.1012954488,-0.3373307586,-0.2797778845,-0.1999591887,0.0397296138,0.2189149559,0.4644092917,0.2239390016,-0.1449688077,-0.2340929955,-0.0960400775,-0.1071276441,-0.3327950239,0.1731466651,0.2569656074,-0.1455277801,-0.2706973553,0.0175240487,-0.0046456596,0.3147620261,-0.181398809,0.1084276885,-0.3498391211,-0.154106304,0.2446233481,-0.1181630418,-0.4130525589,0.0719900876,0.3762161732,0.3241554499,-0.0967503041,0.3059367836,0.1406514198,-0.1888916343,-0.0053488361,0.7830973864,-0.3070815206,-0.4296426475,-0.1124283075,-0.0667238757]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1981","title":"wmt datasets fail to load","comments":"@stas00 Mea culpa... May I fix this tomorrow morning?","body":"on master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova ","comment_length":9,"text":"wmt datasets fail to load\non master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova \n@stas00 Mea culpa... May I fix this tomorrow morning?","embeddings":[-0.3290719688,-0.0584938265,-0.0104082478,0.5395736694,0.3104315698,0.0041043591,0.2097796947,0.0872702822,0.3129351139,0.1078865156,-0.0224996712,-0.1256612837,-0.296931684,0.1887943596,0.1114164367,0.1848195195,-0.1364242435,0.0058543072,-0.821605444,0.015868606,-0.1838497221,0.1984840035,-0.167342186,-0.0969723165,-0.5068030357,0.2375522107,-0.0300084837,0.2536428869,-0.2113710791,-0.4816319048,0.5141910315,0.0309756044,0.0518511981,0.5576848388,-0.0001141044,0.1515254825,0.2935239077,-0.01024595,-0.2466028333,-0.2974118888,-0.5779017806,-0.3827317059,0.0219598692,0.2245493531,-0.1884617954,0.0820968226,-0.2400668561,-0.3640685081,0.2820504308,0.2943232954,0.2214949429,0.4151587784,0.2789464891,-0.3091937304,0.0510975905,0.2907721698,-0.1379881054,-0.0280019324,0.0135747483,-0.1352173835,0.0232767537,0.0316457935,0.0733231753,0.3297724426,0.3457936347,-0.1587231904,0.2857407629,-0.0081027998,0.2792252004,0.0577791668,0.1964938641,-0.075779289,-0.2360399216,-0.0594310798,-0.130695492,-0.4276536703,0.2592030466,0.2658741474,0.0303441808,0.2660063207,-0.3589408398,0.0739574134,0.2212079316,0.0317872539,-0.3881673515,0.3771407902,-0.3114940822,0.0018198559,0.2749304175,-0.1902687997,0.1660659462,-0.1114594862,-0.0187486093,0.0837211162,-0.5072568655,0.2616916895,0.0255597904,0.2756205797,-0.0800576732,0.1582372934,-0.2520163655,0.014928068,0.2089373171,-0.0155652929,0.1787930727,0.1386247426,-0.109944649,-0.0626590401,0.2749227881,0.0550982095,-0.1853418797,0.1351207048,-0.1703789979,-0.4433655143,-0.089175649,-0.1894311756,0.1525510997,-0.0720274746,-0.2599219978,0.0635320917,0.135518983,-0.3142132461,0.0204186551,0.2855190933,0.0484095998,0.170236975,0.1179608032,0.00365311,-0.0042028385,-0.0357706062,-0.3564473689,-0.3906199932,-0.2991639674,0.0611203648,0.4078392386,-0.2185730487,0.2283025235,0.0004109528,-0.1310051382,-0.1273209304,-0.2269295007,-0.0077418489,0.2188936472,0.4824034274,0.071092926,0.2990789115,0.2443664521,0.1595918238,-0.0081311166,-0.0074383789,-0.0995391086,-0.2956005335,0.1295208484,0.2581728995,-0.0815234408,0.1570645422,-0.2294500917,-0.1136633456,0.1050200537,0.3215996325,-0.19287166,-0.2417548001,-0.1637251675,-0.0666386709,0.5321374536,0.5977163315,0.0095659709,-0.194243744,-0.3947505057,-0.0471861549,0.2344344705,0.332223624,-0.0298235733,0.1007020026,-0.26099509,-0.2470602542,-0.0820458457,-0.059449859,-0.1832846254,0.1371451616,-0.1313237101,0.2751179039,0.0324510708,-0.1263800412,0.0145307528,-0.2265354246,0.34289065,0.0891364664,-0.1580777168,0.0836280808,-0.079416573,-0.3164910972,0.2474278659,0.1886273175,0.1936812997,-0.1896170974,0.2157250047,0.0242878571,0.0684091523,0.0570893697,0.1292500794,0.0385188013,0.0848584697,-0.0127780912,-0.0889184177,0.0534026809,-0.4473019838,0.2964880764,0.2691924274,0.0437409878,0.1173462942,0.1208167449,-0.1622187644,0.1335769892,-0.3434743881,-0.1887686104,0.0876794457,0.1349028647,-0.0035151422,-0.1450345963,-0.0388088152,-0.023978943,-0.2259172648,0.1487141699,0.2445334345,0.0667199939,-0.1433457583,-0.1252084821,0.2802841961,0.2743451595,0.1635603011,-0.1697696,-0.1630631983,0.090857625,0.1399752051,0.395195812,-0.0871860459,0.1039211378,0.1791915148,-0.4736975133,0.1699429303,0.1361609697,-0.1402692944,-0.2992374897,0.1204287559,0.0424622707,-0.1103156656,0.2675850391,0.1175593957,0.0357345492,0.1648183763,0.0709900931,0.1453140378,-0.3879213929,0.280146718,-0.0923160762,0.3428555131,0.2670232058,0.0239535514,-0.0425498374,0.3805395365,0.1634019315,0.2022353113,0.0150986891,-0.3856002092,-0.0856616274,-0.0661873668,0.0501848124,0.3967731297,0.0735268518,0.0935864076,0.1786359698,0.1149715856,-0.302526772,0.4467021525,-0.0749787018,-0.0767071843,0.30438748,-0.0235597752,-0.0722774938,-0.2620562017,0.5669642687,0.0582100004,0.0062176031,-0.1975153834,-0.2630082965,-0.4342718422,0.2357833087,-0.3500567377,-0.5691426992,-0.2052633166,-0.2057430893,-0.2758672237,0.1181358621,-0.0988537818,0.2478714138,0.0687464401,0.2188895941,-0.0490783527,0.3465979993,-0.0977031887,-0.2259616405,-0.1064973697,0.0391054899,0.5312060118,-0.2995474339,0.327535212,-0.1957136095,0.063988708,-0.2002592385,-0.2282969952,0.0872321725,-0.0478701182,0.3465410471,0.2117685378,-0.0232825149,0.364433527,-0.1089356244,0.3259528279,-0.1572975963,-0.0153347533,0.0034697694,-0.044544246,-0.1426709145,-0.2228537053,-0.3446230888,-0.4408054054,-0.3883455396,-0.0106526911,-0.0261881314,-0.0418932587,0.091757603,-0.1342888921,-0.0532308593,-0.0178111736,-0.025891291,-0.3486146033,-0.2984477878,0.2000667304,-0.3020641506,-0.2819243073,0.2392926961,0.13543652,0.4544383287,0.040127337,-0.3937389553,0.5808199644,-0.1391113997,-0.1142625138,0.1804570556,0.3090633154,-0.0905662403,0.1654050052,-0.0495853089,-0.1657474041,-0.2646009028,-0.1806016117,-0.2320661098,0.1591044813,-0.1456702203,0.062785618,0.010418159,0.8047193289,0.0654786751,-0.2943529487,0.1827455461,0.1267124414,0.2218532413,-0.0006962097,-0.5895354152,0.2010116726,-0.0986168161,0.2215839028,0.3626628816,0.0432908051,0.1737747043,-0.1139224693,0.1939601153,-0.1887793243,-0.3513721824,-0.1281636208,0.2005107254,0.3849678338,0.0900534242,0.1219887882,0.0273914002,-0.0954923853,-0.0095352363,0.354700923,0.1222476214,0.0885325372,-0.1050639674,-0.0560128465,-0.2369755805,0.0535890795,-0.1932052821,0.4575626254,-0.0324502178,-0.1230478138,-0.0423158407,-0.0182903316,0.5026882291,0.1967388988,0.3927701116,0.1022480875,-0.4248227179,-0.239300549,-0.1275756657,0.0902222469,-0.0735764727,0.1285210401,0.1846208721,-0.2020288408,-0.5072082877,0.2025788128,0.1107909232,-0.1888125837,-0.1198265105,-0.1453297287,0.0249494147,-0.2902211249,-0.0777397379,-0.2248150408,0.1797920763,-0.4049384594,-0.023972258,-0.1545185298,-0.0788380057,0.1511588097,0.0657220259,0.0934651047,0.0443000272,0.4923811555,0.3122967482,0.3200487196,0.5650814176,0.615953207,-0.1010638103,0.0344186723,0.083505854,0.0163794942,0.0522822738,0.283159852,-0.0781762227,0.1111004427,0.0438617095,0.1107477918,-0.5960834026,0.1240014806,0.2019082308,0.0680205598,-0.3572423458,-0.0997587666,0.194607392,0.026873542,0.2006976753,0.1760098636,0.1702613533,-0.3972954154,-0.4666161835,0.0168187767,0.494730413,-0.073587589,0.222242713,-0.1423188895,-0.1103767976,0.0566480383,0.2778768539,-0.1172659621,-0.191495344,0.1790853292,0.0319309011,-0.1585556418,0.086094141,0.1152566522,-0.0740363374,0.2761102319,0.0473109744,0.5473350883,0.1399034262,0.0827611834,-0.1333879083,-0.2570032179,-0.2534717619,0.1469009668,-0.0401018672,0.3355665505,-0.2314390987,-0.1019250676,0.3119746745,-0.0665354207,-0.2452950776,0.0065798382,-0.6495981216,0.1033414304,-0.0357609317,-0.1031017154,0.3692237139,-0.2313642651,0.1811163127,0.319975853,-0.1902632564,0.1010997519,-0.3137949705,0.0383315757,-0.2302716672,0.2646715939,0.1008016616,-0.0336168595,-0.0603646114,-0.1086391211,-0.0407444052,-0.0539295301,0.0338524245,0.0507715642,-0.0552913956,-0.0493358076,-0.3194903433,0.011819873,0.0131511372,-0.1511107236,0.1543610096,0.029080661,-0.0549284481,0.3291851878,-0.0549553446,-0.1067547053,-0.0585501045,0.4145893157,-0.2623612285,0.0229859632,0.3038798869,0.4151066542,-0.0175478105,-0.281327337,-0.0685149655,0.090806812,-0.2448111773,0.2701681852,0.1164219454,-0.0899813995,0.380168438,0.5611450076,0.3422602117,-0.1498246193,0.0122458581,-0.2868360281,-0.1911179423,0.0608151965,-0.1435581148,0.1625947654,0.027714828,0.4404928684,-0.1747053266,0.3926444948,-0.3059358299,0.2749380469,-0.094757244,-0.0000577473,0.2123723328,-0.1585534215,0.1591589153,-0.0921347439,0.0412250496,-0.0340272896,-0.5342020392,-0.2137607485,-0.11785952,0.0945381075,0.0980818197,-0.2711153924,0.1690746993,-0.1080352366,0.0807959363,-0.1697109789,0.3470467329,0.0660692304,-0.1222601384,0.0229023825,0.0703012571,0.0658251643,-0.1689609438,0.1390585154,0.0714529082,-0.0677270144,-0.004992947,-0.0410775393,0.0441779569,0.0208381023,0.1150302812,0.0102870604,-0.0216585528,0.1182223186,0.5093998909,-0.3056085408,-0.2525115013,0.2638110816,0.285666734,0.1686179191,0.0513615571,0.0614277981,-0.1690228879,0.1712695211,-0.2953843772,-0.0795397311,-0.0377579108,0.1637812108,0.0857127979,0.0499729849,0.3570745885,-0.2151893675,-0.2337022722,0.3549879789,0.3251188695,-0.1116924211,0.1748533994,0.4612625837,0.0847785622,0.2086670697,0.1075345576,-0.1653207839,0.1047748998,0.3765080571,-0.2396270782,0.680370748,0.0646547824,0.006373262,0.24075526,-0.4143283367,0.1107713133,0.2690977156,-0.1568296403,0.0244206265,-0.2630802393,0.1377333105,-0.3402009606,-0.2731530666,-0.1855362207,0.1984923333,-0.0655509979,-0.125413537,-0.2556758225,-0.1166857257,-0.0780742988,0.0273622628,-0.0752226934,-0.1022672802,0.2595925927,0.173260808,-0.1564741433,-0.4046477079,-0.3694777787,0.0161443371,-0.098320432,-0.1228213236,0.256008625,0.4048367143,-0.2123374194,0.1384949088,0.4051626027,0.4715678394,0.3013881147,-0.0192856137,-0.0860338286,0.0871971771,-0.0851401165,-0.2775004804,0.4197260141,-0.2568069994,-0.157341525,0.0049490035,0.1699633896,-0.1862421781,-0.3644252717,0.1278567314,-0.3126043677,-0.2835620046,0.1925962269,-0.0886972845,-0.0576487854,-0.0464047492,0.3801314831,-0.0463565886,0.2271073312,0.6677170396,-0.0275353566,0.1280924529,-0.24235861,0.1004886553,0.085262239,0.719905436,0.2930560708,-0.0817147717,-0.286534816,-0.5099948645,-0.7815919518,0.0482606329,0.1542916,0.4362874329,0.0736527294,0.0439882986,0.1192256138,-0.0225296728,0.1537165791,0.1220184043,0.2925755382,-0.215133965,-0.3548647165,0.1796959639,-0.0950149968,-0.1906720698,-0.0053808568,-0.3220204711,0.1466882378,-0.2788923085,0.0247029401,-0.2417231798,0.423589617,0.0029747856,0.2875816226,0.2188119441,0.0880910382,0.3063708842,0.0107983965,-0.063057445,-0.4197324216,-0.3112786114,0.014537733,0.0860736221,-0.1074681729,0.3646843135,-0.0537015609,-0.2685222924,-0.4432088137,0.1050856933,0.1143686846,0.20233652,-0.1769937277,0.222210601,-0.2147685587,-0.1733141243,0.3443754017,0.4282422066,-0.0292369928,0.1878373325,-0.210676685,-0.3873319328,0.2546341419,-0.2887773514,-0.1629846394,-0.4629845619,0.1772362888,-0.0954384282,-0.1698636413,-0.4217805266,0.1528434455,0.1712173522,-0.1016812101,-0.2147470862,0.3726132214,-0.0460546836,-0.0684140325,-0.0185811818,0.1694674492,0.1803847849,-0.3343190551,-0.1139617413,-0.0598337278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1981","title":"wmt datasets fail to load","comments":"yes, of course, I reverted to the version before that and it works ;)\r\n\r\nbut since a new release was just made you will probably need to make a hotfix.\r\n\r\nand add the wmt to the tests?","body":"on master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova ","comment_length":37,"text":"wmt datasets fail to load\non master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova \nyes, of course, I reverted to the version before that and it works ;)\r\n\r\nbut since a new release was just made you will probably need to make a hotfix.\r\n\r\nand add the wmt to the tests?","embeddings":[-0.3290719688,-0.0584938265,-0.0104082478,0.5395736694,0.3104315698,0.0041043591,0.2097796947,0.0872702822,0.3129351139,0.1078865156,-0.0224996712,-0.1256612837,-0.296931684,0.1887943596,0.1114164367,0.1848195195,-0.1364242435,0.0058543072,-0.821605444,0.015868606,-0.1838497221,0.1984840035,-0.167342186,-0.0969723165,-0.5068030357,0.2375522107,-0.0300084837,0.2536428869,-0.2113710791,-0.4816319048,0.5141910315,0.0309756044,0.0518511981,0.5576848388,-0.0001141044,0.1515254825,0.2935239077,-0.01024595,-0.2466028333,-0.2974118888,-0.5779017806,-0.3827317059,0.0219598692,0.2245493531,-0.1884617954,0.0820968226,-0.2400668561,-0.3640685081,0.2820504308,0.2943232954,0.2214949429,0.4151587784,0.2789464891,-0.3091937304,0.0510975905,0.2907721698,-0.1379881054,-0.0280019324,0.0135747483,-0.1352173835,0.0232767537,0.0316457935,0.0733231753,0.3297724426,0.3457936347,-0.1587231904,0.2857407629,-0.0081027998,0.2792252004,0.0577791668,0.1964938641,-0.075779289,-0.2360399216,-0.0594310798,-0.130695492,-0.4276536703,0.2592030466,0.2658741474,0.0303441808,0.2660063207,-0.3589408398,0.0739574134,0.2212079316,0.0317872539,-0.3881673515,0.3771407902,-0.3114940822,0.0018198559,0.2749304175,-0.1902687997,0.1660659462,-0.1114594862,-0.0187486093,0.0837211162,-0.5072568655,0.2616916895,0.0255597904,0.2756205797,-0.0800576732,0.1582372934,-0.2520163655,0.014928068,0.2089373171,-0.0155652929,0.1787930727,0.1386247426,-0.109944649,-0.0626590401,0.2749227881,0.0550982095,-0.1853418797,0.1351207048,-0.1703789979,-0.4433655143,-0.089175649,-0.1894311756,0.1525510997,-0.0720274746,-0.2599219978,0.0635320917,0.135518983,-0.3142132461,0.0204186551,0.2855190933,0.0484095998,0.170236975,0.1179608032,0.00365311,-0.0042028385,-0.0357706062,-0.3564473689,-0.3906199932,-0.2991639674,0.0611203648,0.4078392386,-0.2185730487,0.2283025235,0.0004109528,-0.1310051382,-0.1273209304,-0.2269295007,-0.0077418489,0.2188936472,0.4824034274,0.071092926,0.2990789115,0.2443664521,0.1595918238,-0.0081311166,-0.0074383789,-0.0995391086,-0.2956005335,0.1295208484,0.2581728995,-0.0815234408,0.1570645422,-0.2294500917,-0.1136633456,0.1050200537,0.3215996325,-0.19287166,-0.2417548001,-0.1637251675,-0.0666386709,0.5321374536,0.5977163315,0.0095659709,-0.194243744,-0.3947505057,-0.0471861549,0.2344344705,0.332223624,-0.0298235733,0.1007020026,-0.26099509,-0.2470602542,-0.0820458457,-0.059449859,-0.1832846254,0.1371451616,-0.1313237101,0.2751179039,0.0324510708,-0.1263800412,0.0145307528,-0.2265354246,0.34289065,0.0891364664,-0.1580777168,0.0836280808,-0.079416573,-0.3164910972,0.2474278659,0.1886273175,0.1936812997,-0.1896170974,0.2157250047,0.0242878571,0.0684091523,0.0570893697,0.1292500794,0.0385188013,0.0848584697,-0.0127780912,-0.0889184177,0.0534026809,-0.4473019838,0.2964880764,0.2691924274,0.0437409878,0.1173462942,0.1208167449,-0.1622187644,0.1335769892,-0.3434743881,-0.1887686104,0.0876794457,0.1349028647,-0.0035151422,-0.1450345963,-0.0388088152,-0.023978943,-0.2259172648,0.1487141699,0.2445334345,0.0667199939,-0.1433457583,-0.1252084821,0.2802841961,0.2743451595,0.1635603011,-0.1697696,-0.1630631983,0.090857625,0.1399752051,0.395195812,-0.0871860459,0.1039211378,0.1791915148,-0.4736975133,0.1699429303,0.1361609697,-0.1402692944,-0.2992374897,0.1204287559,0.0424622707,-0.1103156656,0.2675850391,0.1175593957,0.0357345492,0.1648183763,0.0709900931,0.1453140378,-0.3879213929,0.280146718,-0.0923160762,0.3428555131,0.2670232058,0.0239535514,-0.0425498374,0.3805395365,0.1634019315,0.2022353113,0.0150986891,-0.3856002092,-0.0856616274,-0.0661873668,0.0501848124,0.3967731297,0.0735268518,0.0935864076,0.1786359698,0.1149715856,-0.302526772,0.4467021525,-0.0749787018,-0.0767071843,0.30438748,-0.0235597752,-0.0722774938,-0.2620562017,0.5669642687,0.0582100004,0.0062176031,-0.1975153834,-0.2630082965,-0.4342718422,0.2357833087,-0.3500567377,-0.5691426992,-0.2052633166,-0.2057430893,-0.2758672237,0.1181358621,-0.0988537818,0.2478714138,0.0687464401,0.2188895941,-0.0490783527,0.3465979993,-0.0977031887,-0.2259616405,-0.1064973697,0.0391054899,0.5312060118,-0.2995474339,0.327535212,-0.1957136095,0.063988708,-0.2002592385,-0.2282969952,0.0872321725,-0.0478701182,0.3465410471,0.2117685378,-0.0232825149,0.364433527,-0.1089356244,0.3259528279,-0.1572975963,-0.0153347533,0.0034697694,-0.044544246,-0.1426709145,-0.2228537053,-0.3446230888,-0.4408054054,-0.3883455396,-0.0106526911,-0.0261881314,-0.0418932587,0.091757603,-0.1342888921,-0.0532308593,-0.0178111736,-0.025891291,-0.3486146033,-0.2984477878,0.2000667304,-0.3020641506,-0.2819243073,0.2392926961,0.13543652,0.4544383287,0.040127337,-0.3937389553,0.5808199644,-0.1391113997,-0.1142625138,0.1804570556,0.3090633154,-0.0905662403,0.1654050052,-0.0495853089,-0.1657474041,-0.2646009028,-0.1806016117,-0.2320661098,0.1591044813,-0.1456702203,0.062785618,0.010418159,0.8047193289,0.0654786751,-0.2943529487,0.1827455461,0.1267124414,0.2218532413,-0.0006962097,-0.5895354152,0.2010116726,-0.0986168161,0.2215839028,0.3626628816,0.0432908051,0.1737747043,-0.1139224693,0.1939601153,-0.1887793243,-0.3513721824,-0.1281636208,0.2005107254,0.3849678338,0.0900534242,0.1219887882,0.0273914002,-0.0954923853,-0.0095352363,0.354700923,0.1222476214,0.0885325372,-0.1050639674,-0.0560128465,-0.2369755805,0.0535890795,-0.1932052821,0.4575626254,-0.0324502178,-0.1230478138,-0.0423158407,-0.0182903316,0.5026882291,0.1967388988,0.3927701116,0.1022480875,-0.4248227179,-0.239300549,-0.1275756657,0.0902222469,-0.0735764727,0.1285210401,0.1846208721,-0.2020288408,-0.5072082877,0.2025788128,0.1107909232,-0.1888125837,-0.1198265105,-0.1453297287,0.0249494147,-0.2902211249,-0.0777397379,-0.2248150408,0.1797920763,-0.4049384594,-0.023972258,-0.1545185298,-0.0788380057,0.1511588097,0.0657220259,0.0934651047,0.0443000272,0.4923811555,0.3122967482,0.3200487196,0.5650814176,0.615953207,-0.1010638103,0.0344186723,0.083505854,0.0163794942,0.0522822738,0.283159852,-0.0781762227,0.1111004427,0.0438617095,0.1107477918,-0.5960834026,0.1240014806,0.2019082308,0.0680205598,-0.3572423458,-0.0997587666,0.194607392,0.026873542,0.2006976753,0.1760098636,0.1702613533,-0.3972954154,-0.4666161835,0.0168187767,0.494730413,-0.073587589,0.222242713,-0.1423188895,-0.1103767976,0.0566480383,0.2778768539,-0.1172659621,-0.191495344,0.1790853292,0.0319309011,-0.1585556418,0.086094141,0.1152566522,-0.0740363374,0.2761102319,0.0473109744,0.5473350883,0.1399034262,0.0827611834,-0.1333879083,-0.2570032179,-0.2534717619,0.1469009668,-0.0401018672,0.3355665505,-0.2314390987,-0.1019250676,0.3119746745,-0.0665354207,-0.2452950776,0.0065798382,-0.6495981216,0.1033414304,-0.0357609317,-0.1031017154,0.3692237139,-0.2313642651,0.1811163127,0.319975853,-0.1902632564,0.1010997519,-0.3137949705,0.0383315757,-0.2302716672,0.2646715939,0.1008016616,-0.0336168595,-0.0603646114,-0.1086391211,-0.0407444052,-0.0539295301,0.0338524245,0.0507715642,-0.0552913956,-0.0493358076,-0.3194903433,0.011819873,0.0131511372,-0.1511107236,0.1543610096,0.029080661,-0.0549284481,0.3291851878,-0.0549553446,-0.1067547053,-0.0585501045,0.4145893157,-0.2623612285,0.0229859632,0.3038798869,0.4151066542,-0.0175478105,-0.281327337,-0.0685149655,0.090806812,-0.2448111773,0.2701681852,0.1164219454,-0.0899813995,0.380168438,0.5611450076,0.3422602117,-0.1498246193,0.0122458581,-0.2868360281,-0.1911179423,0.0608151965,-0.1435581148,0.1625947654,0.027714828,0.4404928684,-0.1747053266,0.3926444948,-0.3059358299,0.2749380469,-0.094757244,-0.0000577473,0.2123723328,-0.1585534215,0.1591589153,-0.0921347439,0.0412250496,-0.0340272896,-0.5342020392,-0.2137607485,-0.11785952,0.0945381075,0.0980818197,-0.2711153924,0.1690746993,-0.1080352366,0.0807959363,-0.1697109789,0.3470467329,0.0660692304,-0.1222601384,0.0229023825,0.0703012571,0.0658251643,-0.1689609438,0.1390585154,0.0714529082,-0.0677270144,-0.004992947,-0.0410775393,0.0441779569,0.0208381023,0.1150302812,0.0102870604,-0.0216585528,0.1182223186,0.5093998909,-0.3056085408,-0.2525115013,0.2638110816,0.285666734,0.1686179191,0.0513615571,0.0614277981,-0.1690228879,0.1712695211,-0.2953843772,-0.0795397311,-0.0377579108,0.1637812108,0.0857127979,0.0499729849,0.3570745885,-0.2151893675,-0.2337022722,0.3549879789,0.3251188695,-0.1116924211,0.1748533994,0.4612625837,0.0847785622,0.2086670697,0.1075345576,-0.1653207839,0.1047748998,0.3765080571,-0.2396270782,0.680370748,0.0646547824,0.006373262,0.24075526,-0.4143283367,0.1107713133,0.2690977156,-0.1568296403,0.0244206265,-0.2630802393,0.1377333105,-0.3402009606,-0.2731530666,-0.1855362207,0.1984923333,-0.0655509979,-0.125413537,-0.2556758225,-0.1166857257,-0.0780742988,0.0273622628,-0.0752226934,-0.1022672802,0.2595925927,0.173260808,-0.1564741433,-0.4046477079,-0.3694777787,0.0161443371,-0.098320432,-0.1228213236,0.256008625,0.4048367143,-0.2123374194,0.1384949088,0.4051626027,0.4715678394,0.3013881147,-0.0192856137,-0.0860338286,0.0871971771,-0.0851401165,-0.2775004804,0.4197260141,-0.2568069994,-0.157341525,0.0049490035,0.1699633896,-0.1862421781,-0.3644252717,0.1278567314,-0.3126043677,-0.2835620046,0.1925962269,-0.0886972845,-0.0576487854,-0.0464047492,0.3801314831,-0.0463565886,0.2271073312,0.6677170396,-0.0275353566,0.1280924529,-0.24235861,0.1004886553,0.085262239,0.719905436,0.2930560708,-0.0817147717,-0.286534816,-0.5099948645,-0.7815919518,0.0482606329,0.1542916,0.4362874329,0.0736527294,0.0439882986,0.1192256138,-0.0225296728,0.1537165791,0.1220184043,0.2925755382,-0.215133965,-0.3548647165,0.1796959639,-0.0950149968,-0.1906720698,-0.0053808568,-0.3220204711,0.1466882378,-0.2788923085,0.0247029401,-0.2417231798,0.423589617,0.0029747856,0.2875816226,0.2188119441,0.0880910382,0.3063708842,0.0107983965,-0.063057445,-0.4197324216,-0.3112786114,0.014537733,0.0860736221,-0.1074681729,0.3646843135,-0.0537015609,-0.2685222924,-0.4432088137,0.1050856933,0.1143686846,0.20233652,-0.1769937277,0.222210601,-0.2147685587,-0.1733141243,0.3443754017,0.4282422066,-0.0292369928,0.1878373325,-0.210676685,-0.3873319328,0.2546341419,-0.2887773514,-0.1629846394,-0.4629845619,0.1772362888,-0.0954384282,-0.1698636413,-0.4217805266,0.1528434455,0.1712173522,-0.1016812101,-0.2147470862,0.3726132214,-0.0460546836,-0.0684140325,-0.0185811818,0.1694674492,0.1803847849,-0.3343190551,-0.1139617413,-0.0598337278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1981","title":"wmt datasets fail to load","comments":"Sure, I will implement a regression test!","body":"on master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova ","comment_length":7,"text":"wmt datasets fail to load\non master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova \nSure, I will implement a regression test!","embeddings":[-0.3290719688,-0.0584938265,-0.0104082478,0.5395736694,0.3104315698,0.0041043591,0.2097796947,0.0872702822,0.3129351139,0.1078865156,-0.0224996712,-0.1256612837,-0.296931684,0.1887943596,0.1114164367,0.1848195195,-0.1364242435,0.0058543072,-0.821605444,0.015868606,-0.1838497221,0.1984840035,-0.167342186,-0.0969723165,-0.5068030357,0.2375522107,-0.0300084837,0.2536428869,-0.2113710791,-0.4816319048,0.5141910315,0.0309756044,0.0518511981,0.5576848388,-0.0001141044,0.1515254825,0.2935239077,-0.01024595,-0.2466028333,-0.2974118888,-0.5779017806,-0.3827317059,0.0219598692,0.2245493531,-0.1884617954,0.0820968226,-0.2400668561,-0.3640685081,0.2820504308,0.2943232954,0.2214949429,0.4151587784,0.2789464891,-0.3091937304,0.0510975905,0.2907721698,-0.1379881054,-0.0280019324,0.0135747483,-0.1352173835,0.0232767537,0.0316457935,0.0733231753,0.3297724426,0.3457936347,-0.1587231904,0.2857407629,-0.0081027998,0.2792252004,0.0577791668,0.1964938641,-0.075779289,-0.2360399216,-0.0594310798,-0.130695492,-0.4276536703,0.2592030466,0.2658741474,0.0303441808,0.2660063207,-0.3589408398,0.0739574134,0.2212079316,0.0317872539,-0.3881673515,0.3771407902,-0.3114940822,0.0018198559,0.2749304175,-0.1902687997,0.1660659462,-0.1114594862,-0.0187486093,0.0837211162,-0.5072568655,0.2616916895,0.0255597904,0.2756205797,-0.0800576732,0.1582372934,-0.2520163655,0.014928068,0.2089373171,-0.0155652929,0.1787930727,0.1386247426,-0.109944649,-0.0626590401,0.2749227881,0.0550982095,-0.1853418797,0.1351207048,-0.1703789979,-0.4433655143,-0.089175649,-0.1894311756,0.1525510997,-0.0720274746,-0.2599219978,0.0635320917,0.135518983,-0.3142132461,0.0204186551,0.2855190933,0.0484095998,0.170236975,0.1179608032,0.00365311,-0.0042028385,-0.0357706062,-0.3564473689,-0.3906199932,-0.2991639674,0.0611203648,0.4078392386,-0.2185730487,0.2283025235,0.0004109528,-0.1310051382,-0.1273209304,-0.2269295007,-0.0077418489,0.2188936472,0.4824034274,0.071092926,0.2990789115,0.2443664521,0.1595918238,-0.0081311166,-0.0074383789,-0.0995391086,-0.2956005335,0.1295208484,0.2581728995,-0.0815234408,0.1570645422,-0.2294500917,-0.1136633456,0.1050200537,0.3215996325,-0.19287166,-0.2417548001,-0.1637251675,-0.0666386709,0.5321374536,0.5977163315,0.0095659709,-0.194243744,-0.3947505057,-0.0471861549,0.2344344705,0.332223624,-0.0298235733,0.1007020026,-0.26099509,-0.2470602542,-0.0820458457,-0.059449859,-0.1832846254,0.1371451616,-0.1313237101,0.2751179039,0.0324510708,-0.1263800412,0.0145307528,-0.2265354246,0.34289065,0.0891364664,-0.1580777168,0.0836280808,-0.079416573,-0.3164910972,0.2474278659,0.1886273175,0.1936812997,-0.1896170974,0.2157250047,0.0242878571,0.0684091523,0.0570893697,0.1292500794,0.0385188013,0.0848584697,-0.0127780912,-0.0889184177,0.0534026809,-0.4473019838,0.2964880764,0.2691924274,0.0437409878,0.1173462942,0.1208167449,-0.1622187644,0.1335769892,-0.3434743881,-0.1887686104,0.0876794457,0.1349028647,-0.0035151422,-0.1450345963,-0.0388088152,-0.023978943,-0.2259172648,0.1487141699,0.2445334345,0.0667199939,-0.1433457583,-0.1252084821,0.2802841961,0.2743451595,0.1635603011,-0.1697696,-0.1630631983,0.090857625,0.1399752051,0.395195812,-0.0871860459,0.1039211378,0.1791915148,-0.4736975133,0.1699429303,0.1361609697,-0.1402692944,-0.2992374897,0.1204287559,0.0424622707,-0.1103156656,0.2675850391,0.1175593957,0.0357345492,0.1648183763,0.0709900931,0.1453140378,-0.3879213929,0.280146718,-0.0923160762,0.3428555131,0.2670232058,0.0239535514,-0.0425498374,0.3805395365,0.1634019315,0.2022353113,0.0150986891,-0.3856002092,-0.0856616274,-0.0661873668,0.0501848124,0.3967731297,0.0735268518,0.0935864076,0.1786359698,0.1149715856,-0.302526772,0.4467021525,-0.0749787018,-0.0767071843,0.30438748,-0.0235597752,-0.0722774938,-0.2620562017,0.5669642687,0.0582100004,0.0062176031,-0.1975153834,-0.2630082965,-0.4342718422,0.2357833087,-0.3500567377,-0.5691426992,-0.2052633166,-0.2057430893,-0.2758672237,0.1181358621,-0.0988537818,0.2478714138,0.0687464401,0.2188895941,-0.0490783527,0.3465979993,-0.0977031887,-0.2259616405,-0.1064973697,0.0391054899,0.5312060118,-0.2995474339,0.327535212,-0.1957136095,0.063988708,-0.2002592385,-0.2282969952,0.0872321725,-0.0478701182,0.3465410471,0.2117685378,-0.0232825149,0.364433527,-0.1089356244,0.3259528279,-0.1572975963,-0.0153347533,0.0034697694,-0.044544246,-0.1426709145,-0.2228537053,-0.3446230888,-0.4408054054,-0.3883455396,-0.0106526911,-0.0261881314,-0.0418932587,0.091757603,-0.1342888921,-0.0532308593,-0.0178111736,-0.025891291,-0.3486146033,-0.2984477878,0.2000667304,-0.3020641506,-0.2819243073,0.2392926961,0.13543652,0.4544383287,0.040127337,-0.3937389553,0.5808199644,-0.1391113997,-0.1142625138,0.1804570556,0.3090633154,-0.0905662403,0.1654050052,-0.0495853089,-0.1657474041,-0.2646009028,-0.1806016117,-0.2320661098,0.1591044813,-0.1456702203,0.062785618,0.010418159,0.8047193289,0.0654786751,-0.2943529487,0.1827455461,0.1267124414,0.2218532413,-0.0006962097,-0.5895354152,0.2010116726,-0.0986168161,0.2215839028,0.3626628816,0.0432908051,0.1737747043,-0.1139224693,0.1939601153,-0.1887793243,-0.3513721824,-0.1281636208,0.2005107254,0.3849678338,0.0900534242,0.1219887882,0.0273914002,-0.0954923853,-0.0095352363,0.354700923,0.1222476214,0.0885325372,-0.1050639674,-0.0560128465,-0.2369755805,0.0535890795,-0.1932052821,0.4575626254,-0.0324502178,-0.1230478138,-0.0423158407,-0.0182903316,0.5026882291,0.1967388988,0.3927701116,0.1022480875,-0.4248227179,-0.239300549,-0.1275756657,0.0902222469,-0.0735764727,0.1285210401,0.1846208721,-0.2020288408,-0.5072082877,0.2025788128,0.1107909232,-0.1888125837,-0.1198265105,-0.1453297287,0.0249494147,-0.2902211249,-0.0777397379,-0.2248150408,0.1797920763,-0.4049384594,-0.023972258,-0.1545185298,-0.0788380057,0.1511588097,0.0657220259,0.0934651047,0.0443000272,0.4923811555,0.3122967482,0.3200487196,0.5650814176,0.615953207,-0.1010638103,0.0344186723,0.083505854,0.0163794942,0.0522822738,0.283159852,-0.0781762227,0.1111004427,0.0438617095,0.1107477918,-0.5960834026,0.1240014806,0.2019082308,0.0680205598,-0.3572423458,-0.0997587666,0.194607392,0.026873542,0.2006976753,0.1760098636,0.1702613533,-0.3972954154,-0.4666161835,0.0168187767,0.494730413,-0.073587589,0.222242713,-0.1423188895,-0.1103767976,0.0566480383,0.2778768539,-0.1172659621,-0.191495344,0.1790853292,0.0319309011,-0.1585556418,0.086094141,0.1152566522,-0.0740363374,0.2761102319,0.0473109744,0.5473350883,0.1399034262,0.0827611834,-0.1333879083,-0.2570032179,-0.2534717619,0.1469009668,-0.0401018672,0.3355665505,-0.2314390987,-0.1019250676,0.3119746745,-0.0665354207,-0.2452950776,0.0065798382,-0.6495981216,0.1033414304,-0.0357609317,-0.1031017154,0.3692237139,-0.2313642651,0.1811163127,0.319975853,-0.1902632564,0.1010997519,-0.3137949705,0.0383315757,-0.2302716672,0.2646715939,0.1008016616,-0.0336168595,-0.0603646114,-0.1086391211,-0.0407444052,-0.0539295301,0.0338524245,0.0507715642,-0.0552913956,-0.0493358076,-0.3194903433,0.011819873,0.0131511372,-0.1511107236,0.1543610096,0.029080661,-0.0549284481,0.3291851878,-0.0549553446,-0.1067547053,-0.0585501045,0.4145893157,-0.2623612285,0.0229859632,0.3038798869,0.4151066542,-0.0175478105,-0.281327337,-0.0685149655,0.090806812,-0.2448111773,0.2701681852,0.1164219454,-0.0899813995,0.380168438,0.5611450076,0.3422602117,-0.1498246193,0.0122458581,-0.2868360281,-0.1911179423,0.0608151965,-0.1435581148,0.1625947654,0.027714828,0.4404928684,-0.1747053266,0.3926444948,-0.3059358299,0.2749380469,-0.094757244,-0.0000577473,0.2123723328,-0.1585534215,0.1591589153,-0.0921347439,0.0412250496,-0.0340272896,-0.5342020392,-0.2137607485,-0.11785952,0.0945381075,0.0980818197,-0.2711153924,0.1690746993,-0.1080352366,0.0807959363,-0.1697109789,0.3470467329,0.0660692304,-0.1222601384,0.0229023825,0.0703012571,0.0658251643,-0.1689609438,0.1390585154,0.0714529082,-0.0677270144,-0.004992947,-0.0410775393,0.0441779569,0.0208381023,0.1150302812,0.0102870604,-0.0216585528,0.1182223186,0.5093998909,-0.3056085408,-0.2525115013,0.2638110816,0.285666734,0.1686179191,0.0513615571,0.0614277981,-0.1690228879,0.1712695211,-0.2953843772,-0.0795397311,-0.0377579108,0.1637812108,0.0857127979,0.0499729849,0.3570745885,-0.2151893675,-0.2337022722,0.3549879789,0.3251188695,-0.1116924211,0.1748533994,0.4612625837,0.0847785622,0.2086670697,0.1075345576,-0.1653207839,0.1047748998,0.3765080571,-0.2396270782,0.680370748,0.0646547824,0.006373262,0.24075526,-0.4143283367,0.1107713133,0.2690977156,-0.1568296403,0.0244206265,-0.2630802393,0.1377333105,-0.3402009606,-0.2731530666,-0.1855362207,0.1984923333,-0.0655509979,-0.125413537,-0.2556758225,-0.1166857257,-0.0780742988,0.0273622628,-0.0752226934,-0.1022672802,0.2595925927,0.173260808,-0.1564741433,-0.4046477079,-0.3694777787,0.0161443371,-0.098320432,-0.1228213236,0.256008625,0.4048367143,-0.2123374194,0.1384949088,0.4051626027,0.4715678394,0.3013881147,-0.0192856137,-0.0860338286,0.0871971771,-0.0851401165,-0.2775004804,0.4197260141,-0.2568069994,-0.157341525,0.0049490035,0.1699633896,-0.1862421781,-0.3644252717,0.1278567314,-0.3126043677,-0.2835620046,0.1925962269,-0.0886972845,-0.0576487854,-0.0464047492,0.3801314831,-0.0463565886,0.2271073312,0.6677170396,-0.0275353566,0.1280924529,-0.24235861,0.1004886553,0.085262239,0.719905436,0.2930560708,-0.0817147717,-0.286534816,-0.5099948645,-0.7815919518,0.0482606329,0.1542916,0.4362874329,0.0736527294,0.0439882986,0.1192256138,-0.0225296728,0.1537165791,0.1220184043,0.2925755382,-0.215133965,-0.3548647165,0.1796959639,-0.0950149968,-0.1906720698,-0.0053808568,-0.3220204711,0.1466882378,-0.2788923085,0.0247029401,-0.2417231798,0.423589617,0.0029747856,0.2875816226,0.2188119441,0.0880910382,0.3063708842,0.0107983965,-0.063057445,-0.4197324216,-0.3112786114,0.014537733,0.0860736221,-0.1074681729,0.3646843135,-0.0537015609,-0.2685222924,-0.4432088137,0.1050856933,0.1143686846,0.20233652,-0.1769937277,0.222210601,-0.2147685587,-0.1733141243,0.3443754017,0.4282422066,-0.0292369928,0.1878373325,-0.210676685,-0.3873319328,0.2546341419,-0.2887773514,-0.1629846394,-0.4629845619,0.1772362888,-0.0954384282,-0.1698636413,-0.4217805266,0.1528434455,0.1712173522,-0.1016812101,-0.2147470862,0.3726132214,-0.0460546836,-0.0684140325,-0.0185811818,0.1694674492,0.1803847849,-0.3343190551,-0.1139617413,-0.0598337278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1981","title":"wmt datasets fail to load","comments":"@stas00 it is fixed. @lhoestq are you releasing the hot fix or would you prefer me to do it?","body":"on master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova ","comment_length":19,"text":"wmt datasets fail to load\non master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova \n@stas00 it is fixed. @lhoestq are you releasing the hot fix or would you prefer me to do it?","embeddings":[-0.3290719688,-0.0584938265,-0.0104082478,0.5395736694,0.3104315698,0.0041043591,0.2097796947,0.0872702822,0.3129351139,0.1078865156,-0.0224996712,-0.1256612837,-0.296931684,0.1887943596,0.1114164367,0.1848195195,-0.1364242435,0.0058543072,-0.821605444,0.015868606,-0.1838497221,0.1984840035,-0.167342186,-0.0969723165,-0.5068030357,0.2375522107,-0.0300084837,0.2536428869,-0.2113710791,-0.4816319048,0.5141910315,0.0309756044,0.0518511981,0.5576848388,-0.0001141044,0.1515254825,0.2935239077,-0.01024595,-0.2466028333,-0.2974118888,-0.5779017806,-0.3827317059,0.0219598692,0.2245493531,-0.1884617954,0.0820968226,-0.2400668561,-0.3640685081,0.2820504308,0.2943232954,0.2214949429,0.4151587784,0.2789464891,-0.3091937304,0.0510975905,0.2907721698,-0.1379881054,-0.0280019324,0.0135747483,-0.1352173835,0.0232767537,0.0316457935,0.0733231753,0.3297724426,0.3457936347,-0.1587231904,0.2857407629,-0.0081027998,0.2792252004,0.0577791668,0.1964938641,-0.075779289,-0.2360399216,-0.0594310798,-0.130695492,-0.4276536703,0.2592030466,0.2658741474,0.0303441808,0.2660063207,-0.3589408398,0.0739574134,0.2212079316,0.0317872539,-0.3881673515,0.3771407902,-0.3114940822,0.0018198559,0.2749304175,-0.1902687997,0.1660659462,-0.1114594862,-0.0187486093,0.0837211162,-0.5072568655,0.2616916895,0.0255597904,0.2756205797,-0.0800576732,0.1582372934,-0.2520163655,0.014928068,0.2089373171,-0.0155652929,0.1787930727,0.1386247426,-0.109944649,-0.0626590401,0.2749227881,0.0550982095,-0.1853418797,0.1351207048,-0.1703789979,-0.4433655143,-0.089175649,-0.1894311756,0.1525510997,-0.0720274746,-0.2599219978,0.0635320917,0.135518983,-0.3142132461,0.0204186551,0.2855190933,0.0484095998,0.170236975,0.1179608032,0.00365311,-0.0042028385,-0.0357706062,-0.3564473689,-0.3906199932,-0.2991639674,0.0611203648,0.4078392386,-0.2185730487,0.2283025235,0.0004109528,-0.1310051382,-0.1273209304,-0.2269295007,-0.0077418489,0.2188936472,0.4824034274,0.071092926,0.2990789115,0.2443664521,0.1595918238,-0.0081311166,-0.0074383789,-0.0995391086,-0.2956005335,0.1295208484,0.2581728995,-0.0815234408,0.1570645422,-0.2294500917,-0.1136633456,0.1050200537,0.3215996325,-0.19287166,-0.2417548001,-0.1637251675,-0.0666386709,0.5321374536,0.5977163315,0.0095659709,-0.194243744,-0.3947505057,-0.0471861549,0.2344344705,0.332223624,-0.0298235733,0.1007020026,-0.26099509,-0.2470602542,-0.0820458457,-0.059449859,-0.1832846254,0.1371451616,-0.1313237101,0.2751179039,0.0324510708,-0.1263800412,0.0145307528,-0.2265354246,0.34289065,0.0891364664,-0.1580777168,0.0836280808,-0.079416573,-0.3164910972,0.2474278659,0.1886273175,0.1936812997,-0.1896170974,0.2157250047,0.0242878571,0.0684091523,0.0570893697,0.1292500794,0.0385188013,0.0848584697,-0.0127780912,-0.0889184177,0.0534026809,-0.4473019838,0.2964880764,0.2691924274,0.0437409878,0.1173462942,0.1208167449,-0.1622187644,0.1335769892,-0.3434743881,-0.1887686104,0.0876794457,0.1349028647,-0.0035151422,-0.1450345963,-0.0388088152,-0.023978943,-0.2259172648,0.1487141699,0.2445334345,0.0667199939,-0.1433457583,-0.1252084821,0.2802841961,0.2743451595,0.1635603011,-0.1697696,-0.1630631983,0.090857625,0.1399752051,0.395195812,-0.0871860459,0.1039211378,0.1791915148,-0.4736975133,0.1699429303,0.1361609697,-0.1402692944,-0.2992374897,0.1204287559,0.0424622707,-0.1103156656,0.2675850391,0.1175593957,0.0357345492,0.1648183763,0.0709900931,0.1453140378,-0.3879213929,0.280146718,-0.0923160762,0.3428555131,0.2670232058,0.0239535514,-0.0425498374,0.3805395365,0.1634019315,0.2022353113,0.0150986891,-0.3856002092,-0.0856616274,-0.0661873668,0.0501848124,0.3967731297,0.0735268518,0.0935864076,0.1786359698,0.1149715856,-0.302526772,0.4467021525,-0.0749787018,-0.0767071843,0.30438748,-0.0235597752,-0.0722774938,-0.2620562017,0.5669642687,0.0582100004,0.0062176031,-0.1975153834,-0.2630082965,-0.4342718422,0.2357833087,-0.3500567377,-0.5691426992,-0.2052633166,-0.2057430893,-0.2758672237,0.1181358621,-0.0988537818,0.2478714138,0.0687464401,0.2188895941,-0.0490783527,0.3465979993,-0.0977031887,-0.2259616405,-0.1064973697,0.0391054899,0.5312060118,-0.2995474339,0.327535212,-0.1957136095,0.063988708,-0.2002592385,-0.2282969952,0.0872321725,-0.0478701182,0.3465410471,0.2117685378,-0.0232825149,0.364433527,-0.1089356244,0.3259528279,-0.1572975963,-0.0153347533,0.0034697694,-0.044544246,-0.1426709145,-0.2228537053,-0.3446230888,-0.4408054054,-0.3883455396,-0.0106526911,-0.0261881314,-0.0418932587,0.091757603,-0.1342888921,-0.0532308593,-0.0178111736,-0.025891291,-0.3486146033,-0.2984477878,0.2000667304,-0.3020641506,-0.2819243073,0.2392926961,0.13543652,0.4544383287,0.040127337,-0.3937389553,0.5808199644,-0.1391113997,-0.1142625138,0.1804570556,0.3090633154,-0.0905662403,0.1654050052,-0.0495853089,-0.1657474041,-0.2646009028,-0.1806016117,-0.2320661098,0.1591044813,-0.1456702203,0.062785618,0.010418159,0.8047193289,0.0654786751,-0.2943529487,0.1827455461,0.1267124414,0.2218532413,-0.0006962097,-0.5895354152,0.2010116726,-0.0986168161,0.2215839028,0.3626628816,0.0432908051,0.1737747043,-0.1139224693,0.1939601153,-0.1887793243,-0.3513721824,-0.1281636208,0.2005107254,0.3849678338,0.0900534242,0.1219887882,0.0273914002,-0.0954923853,-0.0095352363,0.354700923,0.1222476214,0.0885325372,-0.1050639674,-0.0560128465,-0.2369755805,0.0535890795,-0.1932052821,0.4575626254,-0.0324502178,-0.1230478138,-0.0423158407,-0.0182903316,0.5026882291,0.1967388988,0.3927701116,0.1022480875,-0.4248227179,-0.239300549,-0.1275756657,0.0902222469,-0.0735764727,0.1285210401,0.1846208721,-0.2020288408,-0.5072082877,0.2025788128,0.1107909232,-0.1888125837,-0.1198265105,-0.1453297287,0.0249494147,-0.2902211249,-0.0777397379,-0.2248150408,0.1797920763,-0.4049384594,-0.023972258,-0.1545185298,-0.0788380057,0.1511588097,0.0657220259,0.0934651047,0.0443000272,0.4923811555,0.3122967482,0.3200487196,0.5650814176,0.615953207,-0.1010638103,0.0344186723,0.083505854,0.0163794942,0.0522822738,0.283159852,-0.0781762227,0.1111004427,0.0438617095,0.1107477918,-0.5960834026,0.1240014806,0.2019082308,0.0680205598,-0.3572423458,-0.0997587666,0.194607392,0.026873542,0.2006976753,0.1760098636,0.1702613533,-0.3972954154,-0.4666161835,0.0168187767,0.494730413,-0.073587589,0.222242713,-0.1423188895,-0.1103767976,0.0566480383,0.2778768539,-0.1172659621,-0.191495344,0.1790853292,0.0319309011,-0.1585556418,0.086094141,0.1152566522,-0.0740363374,0.2761102319,0.0473109744,0.5473350883,0.1399034262,0.0827611834,-0.1333879083,-0.2570032179,-0.2534717619,0.1469009668,-0.0401018672,0.3355665505,-0.2314390987,-0.1019250676,0.3119746745,-0.0665354207,-0.2452950776,0.0065798382,-0.6495981216,0.1033414304,-0.0357609317,-0.1031017154,0.3692237139,-0.2313642651,0.1811163127,0.319975853,-0.1902632564,0.1010997519,-0.3137949705,0.0383315757,-0.2302716672,0.2646715939,0.1008016616,-0.0336168595,-0.0603646114,-0.1086391211,-0.0407444052,-0.0539295301,0.0338524245,0.0507715642,-0.0552913956,-0.0493358076,-0.3194903433,0.011819873,0.0131511372,-0.1511107236,0.1543610096,0.029080661,-0.0549284481,0.3291851878,-0.0549553446,-0.1067547053,-0.0585501045,0.4145893157,-0.2623612285,0.0229859632,0.3038798869,0.4151066542,-0.0175478105,-0.281327337,-0.0685149655,0.090806812,-0.2448111773,0.2701681852,0.1164219454,-0.0899813995,0.380168438,0.5611450076,0.3422602117,-0.1498246193,0.0122458581,-0.2868360281,-0.1911179423,0.0608151965,-0.1435581148,0.1625947654,0.027714828,0.4404928684,-0.1747053266,0.3926444948,-0.3059358299,0.2749380469,-0.094757244,-0.0000577473,0.2123723328,-0.1585534215,0.1591589153,-0.0921347439,0.0412250496,-0.0340272896,-0.5342020392,-0.2137607485,-0.11785952,0.0945381075,0.0980818197,-0.2711153924,0.1690746993,-0.1080352366,0.0807959363,-0.1697109789,0.3470467329,0.0660692304,-0.1222601384,0.0229023825,0.0703012571,0.0658251643,-0.1689609438,0.1390585154,0.0714529082,-0.0677270144,-0.004992947,-0.0410775393,0.0441779569,0.0208381023,0.1150302812,0.0102870604,-0.0216585528,0.1182223186,0.5093998909,-0.3056085408,-0.2525115013,0.2638110816,0.285666734,0.1686179191,0.0513615571,0.0614277981,-0.1690228879,0.1712695211,-0.2953843772,-0.0795397311,-0.0377579108,0.1637812108,0.0857127979,0.0499729849,0.3570745885,-0.2151893675,-0.2337022722,0.3549879789,0.3251188695,-0.1116924211,0.1748533994,0.4612625837,0.0847785622,0.2086670697,0.1075345576,-0.1653207839,0.1047748998,0.3765080571,-0.2396270782,0.680370748,0.0646547824,0.006373262,0.24075526,-0.4143283367,0.1107713133,0.2690977156,-0.1568296403,0.0244206265,-0.2630802393,0.1377333105,-0.3402009606,-0.2731530666,-0.1855362207,0.1984923333,-0.0655509979,-0.125413537,-0.2556758225,-0.1166857257,-0.0780742988,0.0273622628,-0.0752226934,-0.1022672802,0.2595925927,0.173260808,-0.1564741433,-0.4046477079,-0.3694777787,0.0161443371,-0.098320432,-0.1228213236,0.256008625,0.4048367143,-0.2123374194,0.1384949088,0.4051626027,0.4715678394,0.3013881147,-0.0192856137,-0.0860338286,0.0871971771,-0.0851401165,-0.2775004804,0.4197260141,-0.2568069994,-0.157341525,0.0049490035,0.1699633896,-0.1862421781,-0.3644252717,0.1278567314,-0.3126043677,-0.2835620046,0.1925962269,-0.0886972845,-0.0576487854,-0.0464047492,0.3801314831,-0.0463565886,0.2271073312,0.6677170396,-0.0275353566,0.1280924529,-0.24235861,0.1004886553,0.085262239,0.719905436,0.2930560708,-0.0817147717,-0.286534816,-0.5099948645,-0.7815919518,0.0482606329,0.1542916,0.4362874329,0.0736527294,0.0439882986,0.1192256138,-0.0225296728,0.1537165791,0.1220184043,0.2925755382,-0.215133965,-0.3548647165,0.1796959639,-0.0950149968,-0.1906720698,-0.0053808568,-0.3220204711,0.1466882378,-0.2788923085,0.0247029401,-0.2417231798,0.423589617,0.0029747856,0.2875816226,0.2188119441,0.0880910382,0.3063708842,0.0107983965,-0.063057445,-0.4197324216,-0.3112786114,0.014537733,0.0860736221,-0.1074681729,0.3646843135,-0.0537015609,-0.2685222924,-0.4432088137,0.1050856933,0.1143686846,0.20233652,-0.1769937277,0.222210601,-0.2147685587,-0.1733141243,0.3443754017,0.4282422066,-0.0292369928,0.1878373325,-0.210676685,-0.3873319328,0.2546341419,-0.2887773514,-0.1629846394,-0.4629845619,0.1772362888,-0.0954384282,-0.1698636413,-0.4217805266,0.1528434455,0.1712173522,-0.1016812101,-0.2147470862,0.3726132214,-0.0460546836,-0.0684140325,-0.0185811818,0.1694674492,0.1803847849,-0.3343190551,-0.1139617413,-0.0598337278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1981","title":"wmt datasets fail to load","comments":"I'll do a patch release for this issue early tomorrow.\r\n\r\nAnd yes we absolutly need tests for the wmt datasets: The missing tests for wmt are an artifact from the early development of the lib but now we have tools to generate automatically the dummy data used for tests :)","body":"on master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova ","comment_length":50,"text":"wmt datasets fail to load\non master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova \nI'll do a patch release for this issue early tomorrow.\r\n\r\nAnd yes we absolutly need tests for the wmt datasets: The missing tests for wmt are an artifact from the early development of the lib but now we have tools to generate automatically the dummy data used for tests :)","embeddings":[-0.3290719688,-0.0584938265,-0.0104082478,0.5395736694,0.3104315698,0.0041043591,0.2097796947,0.0872702822,0.3129351139,0.1078865156,-0.0224996712,-0.1256612837,-0.296931684,0.1887943596,0.1114164367,0.1848195195,-0.1364242435,0.0058543072,-0.821605444,0.015868606,-0.1838497221,0.1984840035,-0.167342186,-0.0969723165,-0.5068030357,0.2375522107,-0.0300084837,0.2536428869,-0.2113710791,-0.4816319048,0.5141910315,0.0309756044,0.0518511981,0.5576848388,-0.0001141044,0.1515254825,0.2935239077,-0.01024595,-0.2466028333,-0.2974118888,-0.5779017806,-0.3827317059,0.0219598692,0.2245493531,-0.1884617954,0.0820968226,-0.2400668561,-0.3640685081,0.2820504308,0.2943232954,0.2214949429,0.4151587784,0.2789464891,-0.3091937304,0.0510975905,0.2907721698,-0.1379881054,-0.0280019324,0.0135747483,-0.1352173835,0.0232767537,0.0316457935,0.0733231753,0.3297724426,0.3457936347,-0.1587231904,0.2857407629,-0.0081027998,0.2792252004,0.0577791668,0.1964938641,-0.075779289,-0.2360399216,-0.0594310798,-0.130695492,-0.4276536703,0.2592030466,0.2658741474,0.0303441808,0.2660063207,-0.3589408398,0.0739574134,0.2212079316,0.0317872539,-0.3881673515,0.3771407902,-0.3114940822,0.0018198559,0.2749304175,-0.1902687997,0.1660659462,-0.1114594862,-0.0187486093,0.0837211162,-0.5072568655,0.2616916895,0.0255597904,0.2756205797,-0.0800576732,0.1582372934,-0.2520163655,0.014928068,0.2089373171,-0.0155652929,0.1787930727,0.1386247426,-0.109944649,-0.0626590401,0.2749227881,0.0550982095,-0.1853418797,0.1351207048,-0.1703789979,-0.4433655143,-0.089175649,-0.1894311756,0.1525510997,-0.0720274746,-0.2599219978,0.0635320917,0.135518983,-0.3142132461,0.0204186551,0.2855190933,0.0484095998,0.170236975,0.1179608032,0.00365311,-0.0042028385,-0.0357706062,-0.3564473689,-0.3906199932,-0.2991639674,0.0611203648,0.4078392386,-0.2185730487,0.2283025235,0.0004109528,-0.1310051382,-0.1273209304,-0.2269295007,-0.0077418489,0.2188936472,0.4824034274,0.071092926,0.2990789115,0.2443664521,0.1595918238,-0.0081311166,-0.0074383789,-0.0995391086,-0.2956005335,0.1295208484,0.2581728995,-0.0815234408,0.1570645422,-0.2294500917,-0.1136633456,0.1050200537,0.3215996325,-0.19287166,-0.2417548001,-0.1637251675,-0.0666386709,0.5321374536,0.5977163315,0.0095659709,-0.194243744,-0.3947505057,-0.0471861549,0.2344344705,0.332223624,-0.0298235733,0.1007020026,-0.26099509,-0.2470602542,-0.0820458457,-0.059449859,-0.1832846254,0.1371451616,-0.1313237101,0.2751179039,0.0324510708,-0.1263800412,0.0145307528,-0.2265354246,0.34289065,0.0891364664,-0.1580777168,0.0836280808,-0.079416573,-0.3164910972,0.2474278659,0.1886273175,0.1936812997,-0.1896170974,0.2157250047,0.0242878571,0.0684091523,0.0570893697,0.1292500794,0.0385188013,0.0848584697,-0.0127780912,-0.0889184177,0.0534026809,-0.4473019838,0.2964880764,0.2691924274,0.0437409878,0.1173462942,0.1208167449,-0.1622187644,0.1335769892,-0.3434743881,-0.1887686104,0.0876794457,0.1349028647,-0.0035151422,-0.1450345963,-0.0388088152,-0.023978943,-0.2259172648,0.1487141699,0.2445334345,0.0667199939,-0.1433457583,-0.1252084821,0.2802841961,0.2743451595,0.1635603011,-0.1697696,-0.1630631983,0.090857625,0.1399752051,0.395195812,-0.0871860459,0.1039211378,0.1791915148,-0.4736975133,0.1699429303,0.1361609697,-0.1402692944,-0.2992374897,0.1204287559,0.0424622707,-0.1103156656,0.2675850391,0.1175593957,0.0357345492,0.1648183763,0.0709900931,0.1453140378,-0.3879213929,0.280146718,-0.0923160762,0.3428555131,0.2670232058,0.0239535514,-0.0425498374,0.3805395365,0.1634019315,0.2022353113,0.0150986891,-0.3856002092,-0.0856616274,-0.0661873668,0.0501848124,0.3967731297,0.0735268518,0.0935864076,0.1786359698,0.1149715856,-0.302526772,0.4467021525,-0.0749787018,-0.0767071843,0.30438748,-0.0235597752,-0.0722774938,-0.2620562017,0.5669642687,0.0582100004,0.0062176031,-0.1975153834,-0.2630082965,-0.4342718422,0.2357833087,-0.3500567377,-0.5691426992,-0.2052633166,-0.2057430893,-0.2758672237,0.1181358621,-0.0988537818,0.2478714138,0.0687464401,0.2188895941,-0.0490783527,0.3465979993,-0.0977031887,-0.2259616405,-0.1064973697,0.0391054899,0.5312060118,-0.2995474339,0.327535212,-0.1957136095,0.063988708,-0.2002592385,-0.2282969952,0.0872321725,-0.0478701182,0.3465410471,0.2117685378,-0.0232825149,0.364433527,-0.1089356244,0.3259528279,-0.1572975963,-0.0153347533,0.0034697694,-0.044544246,-0.1426709145,-0.2228537053,-0.3446230888,-0.4408054054,-0.3883455396,-0.0106526911,-0.0261881314,-0.0418932587,0.091757603,-0.1342888921,-0.0532308593,-0.0178111736,-0.025891291,-0.3486146033,-0.2984477878,0.2000667304,-0.3020641506,-0.2819243073,0.2392926961,0.13543652,0.4544383287,0.040127337,-0.3937389553,0.5808199644,-0.1391113997,-0.1142625138,0.1804570556,0.3090633154,-0.0905662403,0.1654050052,-0.0495853089,-0.1657474041,-0.2646009028,-0.1806016117,-0.2320661098,0.1591044813,-0.1456702203,0.062785618,0.010418159,0.8047193289,0.0654786751,-0.2943529487,0.1827455461,0.1267124414,0.2218532413,-0.0006962097,-0.5895354152,0.2010116726,-0.0986168161,0.2215839028,0.3626628816,0.0432908051,0.1737747043,-0.1139224693,0.1939601153,-0.1887793243,-0.3513721824,-0.1281636208,0.2005107254,0.3849678338,0.0900534242,0.1219887882,0.0273914002,-0.0954923853,-0.0095352363,0.354700923,0.1222476214,0.0885325372,-0.1050639674,-0.0560128465,-0.2369755805,0.0535890795,-0.1932052821,0.4575626254,-0.0324502178,-0.1230478138,-0.0423158407,-0.0182903316,0.5026882291,0.1967388988,0.3927701116,0.1022480875,-0.4248227179,-0.239300549,-0.1275756657,0.0902222469,-0.0735764727,0.1285210401,0.1846208721,-0.2020288408,-0.5072082877,0.2025788128,0.1107909232,-0.1888125837,-0.1198265105,-0.1453297287,0.0249494147,-0.2902211249,-0.0777397379,-0.2248150408,0.1797920763,-0.4049384594,-0.023972258,-0.1545185298,-0.0788380057,0.1511588097,0.0657220259,0.0934651047,0.0443000272,0.4923811555,0.3122967482,0.3200487196,0.5650814176,0.615953207,-0.1010638103,0.0344186723,0.083505854,0.0163794942,0.0522822738,0.283159852,-0.0781762227,0.1111004427,0.0438617095,0.1107477918,-0.5960834026,0.1240014806,0.2019082308,0.0680205598,-0.3572423458,-0.0997587666,0.194607392,0.026873542,0.2006976753,0.1760098636,0.1702613533,-0.3972954154,-0.4666161835,0.0168187767,0.494730413,-0.073587589,0.222242713,-0.1423188895,-0.1103767976,0.0566480383,0.2778768539,-0.1172659621,-0.191495344,0.1790853292,0.0319309011,-0.1585556418,0.086094141,0.1152566522,-0.0740363374,0.2761102319,0.0473109744,0.5473350883,0.1399034262,0.0827611834,-0.1333879083,-0.2570032179,-0.2534717619,0.1469009668,-0.0401018672,0.3355665505,-0.2314390987,-0.1019250676,0.3119746745,-0.0665354207,-0.2452950776,0.0065798382,-0.6495981216,0.1033414304,-0.0357609317,-0.1031017154,0.3692237139,-0.2313642651,0.1811163127,0.319975853,-0.1902632564,0.1010997519,-0.3137949705,0.0383315757,-0.2302716672,0.2646715939,0.1008016616,-0.0336168595,-0.0603646114,-0.1086391211,-0.0407444052,-0.0539295301,0.0338524245,0.0507715642,-0.0552913956,-0.0493358076,-0.3194903433,0.011819873,0.0131511372,-0.1511107236,0.1543610096,0.029080661,-0.0549284481,0.3291851878,-0.0549553446,-0.1067547053,-0.0585501045,0.4145893157,-0.2623612285,0.0229859632,0.3038798869,0.4151066542,-0.0175478105,-0.281327337,-0.0685149655,0.090806812,-0.2448111773,0.2701681852,0.1164219454,-0.0899813995,0.380168438,0.5611450076,0.3422602117,-0.1498246193,0.0122458581,-0.2868360281,-0.1911179423,0.0608151965,-0.1435581148,0.1625947654,0.027714828,0.4404928684,-0.1747053266,0.3926444948,-0.3059358299,0.2749380469,-0.094757244,-0.0000577473,0.2123723328,-0.1585534215,0.1591589153,-0.0921347439,0.0412250496,-0.0340272896,-0.5342020392,-0.2137607485,-0.11785952,0.0945381075,0.0980818197,-0.2711153924,0.1690746993,-0.1080352366,0.0807959363,-0.1697109789,0.3470467329,0.0660692304,-0.1222601384,0.0229023825,0.0703012571,0.0658251643,-0.1689609438,0.1390585154,0.0714529082,-0.0677270144,-0.004992947,-0.0410775393,0.0441779569,0.0208381023,0.1150302812,0.0102870604,-0.0216585528,0.1182223186,0.5093998909,-0.3056085408,-0.2525115013,0.2638110816,0.285666734,0.1686179191,0.0513615571,0.0614277981,-0.1690228879,0.1712695211,-0.2953843772,-0.0795397311,-0.0377579108,0.1637812108,0.0857127979,0.0499729849,0.3570745885,-0.2151893675,-0.2337022722,0.3549879789,0.3251188695,-0.1116924211,0.1748533994,0.4612625837,0.0847785622,0.2086670697,0.1075345576,-0.1653207839,0.1047748998,0.3765080571,-0.2396270782,0.680370748,0.0646547824,0.006373262,0.24075526,-0.4143283367,0.1107713133,0.2690977156,-0.1568296403,0.0244206265,-0.2630802393,0.1377333105,-0.3402009606,-0.2731530666,-0.1855362207,0.1984923333,-0.0655509979,-0.125413537,-0.2556758225,-0.1166857257,-0.0780742988,0.0273622628,-0.0752226934,-0.1022672802,0.2595925927,0.173260808,-0.1564741433,-0.4046477079,-0.3694777787,0.0161443371,-0.098320432,-0.1228213236,0.256008625,0.4048367143,-0.2123374194,0.1384949088,0.4051626027,0.4715678394,0.3013881147,-0.0192856137,-0.0860338286,0.0871971771,-0.0851401165,-0.2775004804,0.4197260141,-0.2568069994,-0.157341525,0.0049490035,0.1699633896,-0.1862421781,-0.3644252717,0.1278567314,-0.3126043677,-0.2835620046,0.1925962269,-0.0886972845,-0.0576487854,-0.0464047492,0.3801314831,-0.0463565886,0.2271073312,0.6677170396,-0.0275353566,0.1280924529,-0.24235861,0.1004886553,0.085262239,0.719905436,0.2930560708,-0.0817147717,-0.286534816,-0.5099948645,-0.7815919518,0.0482606329,0.1542916,0.4362874329,0.0736527294,0.0439882986,0.1192256138,-0.0225296728,0.1537165791,0.1220184043,0.2925755382,-0.215133965,-0.3548647165,0.1796959639,-0.0950149968,-0.1906720698,-0.0053808568,-0.3220204711,0.1466882378,-0.2788923085,0.0247029401,-0.2417231798,0.423589617,0.0029747856,0.2875816226,0.2188119441,0.0880910382,0.3063708842,0.0107983965,-0.063057445,-0.4197324216,-0.3112786114,0.014537733,0.0860736221,-0.1074681729,0.3646843135,-0.0537015609,-0.2685222924,-0.4432088137,0.1050856933,0.1143686846,0.20233652,-0.1769937277,0.222210601,-0.2147685587,-0.1733141243,0.3443754017,0.4282422066,-0.0292369928,0.1878373325,-0.210676685,-0.3873319328,0.2546341419,-0.2887773514,-0.1629846394,-0.4629845619,0.1772362888,-0.0954384282,-0.1698636413,-0.4217805266,0.1528434455,0.1712173522,-0.1016812101,-0.2147470862,0.3726132214,-0.0460546836,-0.0684140325,-0.0185811818,0.1694674492,0.1803847849,-0.3343190551,-0.1139617413,-0.0598337278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1981","title":"wmt datasets fail to load","comments":"still facing the same issue or similar:\r\nfrom datasets import load_dataset\r\nwtm14_test = load_dataset('wmt14',\"de-en\",cache_dir='.\/datasets')\r\n\r\n~.cache\\huggingface\\modules\\datasets_modules\\datasets\\wmt14\\43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\\wmt_utils.py in _split_generators(self, dl_manager)\r\n758 # Extract manually downloaded files.\r\n759 manual_files = dl_manager.extract(manual_paths_dict)\r\n--> 760 extraction_map = dict(downloaded_files, **manual_files)\r\n761\r\n762 for language in self.config.language_pair:\r\n\r\nTypeError: type object argument after ** must be a mapping, not list","body":"on master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova ","comment_length":52,"text":"wmt datasets fail to load\non master:\r\n```\r\npython -c 'from datasets import load_dataset; load_dataset(\"wmt14\", \"de-en\")'\r\nDownloading and preparing dataset wmt14\/de-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/stas\/.cache\/huggingface\/datasets\/wmt14\/de-en\/1.0.0\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 578, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 634, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt14\/43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\/wmt_utils.py\", line 760, in _split_generators\r\n extraction_map = dict(downloaded_files, **manual_files)\r\n```\r\n\r\nit worked fine recently. same problem if I try wmt16.\r\n\r\ngit bisect points to this commit from Feb 25 as the culprit https:\/\/github.com\/huggingface\/datasets\/commit\/792f1d9bb1c5361908f73e2ef7f0181b2be409fa\r\n\r\n@albertvillanova \nstill facing the same issue or similar:\r\nfrom datasets import load_dataset\r\nwtm14_test = load_dataset('wmt14',\"de-en\",cache_dir='.\/datasets')\r\n\r\n~.cache\\huggingface\\modules\\datasets_modules\\datasets\\wmt14\\43e717d978d2261502b0194999583acb874ba73b0f4aed0ada2889d1bb00f36e\\wmt_utils.py in _split_generators(self, dl_manager)\r\n758 # Extract manually downloaded files.\r\n759 manual_files = dl_manager.extract(manual_paths_dict)\r\n--> 760 extraction_map = dict(downloaded_files, **manual_files)\r\n761\r\n762 for language in self.config.language_pair:\r\n\r\nTypeError: type object argument after ** must be a mapping, not list","embeddings":[-0.3290719688,-0.0584938265,-0.0104082478,0.5395736694,0.3104315698,0.0041043591,0.2097796947,0.0872702822,0.3129351139,0.1078865156,-0.0224996712,-0.1256612837,-0.296931684,0.1887943596,0.1114164367,0.1848195195,-0.1364242435,0.0058543072,-0.821605444,0.015868606,-0.1838497221,0.1984840035,-0.167342186,-0.0969723165,-0.5068030357,0.2375522107,-0.0300084837,0.2536428869,-0.2113710791,-0.4816319048,0.5141910315,0.0309756044,0.0518511981,0.5576848388,-0.0001141044,0.1515254825,0.2935239077,-0.01024595,-0.2466028333,-0.2974118888,-0.5779017806,-0.3827317059,0.0219598692,0.2245493531,-0.1884617954,0.0820968226,-0.2400668561,-0.3640685081,0.2820504308,0.2943232954,0.2214949429,0.4151587784,0.2789464891,-0.3091937304,0.0510975905,0.2907721698,-0.1379881054,-0.0280019324,0.0135747483,-0.1352173835,0.0232767537,0.0316457935,0.0733231753,0.3297724426,0.3457936347,-0.1587231904,0.2857407629,-0.0081027998,0.2792252004,0.0577791668,0.1964938641,-0.075779289,-0.2360399216,-0.0594310798,-0.130695492,-0.4276536703,0.2592030466,0.2658741474,0.0303441808,0.2660063207,-0.3589408398,0.0739574134,0.2212079316,0.0317872539,-0.3881673515,0.3771407902,-0.3114940822,0.0018198559,0.2749304175,-0.1902687997,0.1660659462,-0.1114594862,-0.0187486093,0.0837211162,-0.5072568655,0.2616916895,0.0255597904,0.2756205797,-0.0800576732,0.1582372934,-0.2520163655,0.014928068,0.2089373171,-0.0155652929,0.1787930727,0.1386247426,-0.109944649,-0.0626590401,0.2749227881,0.0550982095,-0.1853418797,0.1351207048,-0.1703789979,-0.4433655143,-0.089175649,-0.1894311756,0.1525510997,-0.0720274746,-0.2599219978,0.0635320917,0.135518983,-0.3142132461,0.0204186551,0.2855190933,0.0484095998,0.170236975,0.1179608032,0.00365311,-0.0042028385,-0.0357706062,-0.3564473689,-0.3906199932,-0.2991639674,0.0611203648,0.4078392386,-0.2185730487,0.2283025235,0.0004109528,-0.1310051382,-0.1273209304,-0.2269295007,-0.0077418489,0.2188936472,0.4824034274,0.071092926,0.2990789115,0.2443664521,0.1595918238,-0.0081311166,-0.0074383789,-0.0995391086,-0.2956005335,0.1295208484,0.2581728995,-0.0815234408,0.1570645422,-0.2294500917,-0.1136633456,0.1050200537,0.3215996325,-0.19287166,-0.2417548001,-0.1637251675,-0.0666386709,0.5321374536,0.5977163315,0.0095659709,-0.194243744,-0.3947505057,-0.0471861549,0.2344344705,0.332223624,-0.0298235733,0.1007020026,-0.26099509,-0.2470602542,-0.0820458457,-0.059449859,-0.1832846254,0.1371451616,-0.1313237101,0.2751179039,0.0324510708,-0.1263800412,0.0145307528,-0.2265354246,0.34289065,0.0891364664,-0.1580777168,0.0836280808,-0.079416573,-0.3164910972,0.2474278659,0.1886273175,0.1936812997,-0.1896170974,0.2157250047,0.0242878571,0.0684091523,0.0570893697,0.1292500794,0.0385188013,0.0848584697,-0.0127780912,-0.0889184177,0.0534026809,-0.4473019838,0.2964880764,0.2691924274,0.0437409878,0.1173462942,0.1208167449,-0.1622187644,0.1335769892,-0.3434743881,-0.1887686104,0.0876794457,0.1349028647,-0.0035151422,-0.1450345963,-0.0388088152,-0.023978943,-0.2259172648,0.1487141699,0.2445334345,0.0667199939,-0.1433457583,-0.1252084821,0.2802841961,0.2743451595,0.1635603011,-0.1697696,-0.1630631983,0.090857625,0.1399752051,0.395195812,-0.0871860459,0.1039211378,0.1791915148,-0.4736975133,0.1699429303,0.1361609697,-0.1402692944,-0.2992374897,0.1204287559,0.0424622707,-0.1103156656,0.2675850391,0.1175593957,0.0357345492,0.1648183763,0.0709900931,0.1453140378,-0.3879213929,0.280146718,-0.0923160762,0.3428555131,0.2670232058,0.0239535514,-0.0425498374,0.3805395365,0.1634019315,0.2022353113,0.0150986891,-0.3856002092,-0.0856616274,-0.0661873668,0.0501848124,0.3967731297,0.0735268518,0.0935864076,0.1786359698,0.1149715856,-0.302526772,0.4467021525,-0.0749787018,-0.0767071843,0.30438748,-0.0235597752,-0.0722774938,-0.2620562017,0.5669642687,0.0582100004,0.0062176031,-0.1975153834,-0.2630082965,-0.4342718422,0.2357833087,-0.3500567377,-0.5691426992,-0.2052633166,-0.2057430893,-0.2758672237,0.1181358621,-0.0988537818,0.2478714138,0.0687464401,0.2188895941,-0.0490783527,0.3465979993,-0.0977031887,-0.2259616405,-0.1064973697,0.0391054899,0.5312060118,-0.2995474339,0.327535212,-0.1957136095,0.063988708,-0.2002592385,-0.2282969952,0.0872321725,-0.0478701182,0.3465410471,0.2117685378,-0.0232825149,0.364433527,-0.1089356244,0.3259528279,-0.1572975963,-0.0153347533,0.0034697694,-0.044544246,-0.1426709145,-0.2228537053,-0.3446230888,-0.4408054054,-0.3883455396,-0.0106526911,-0.0261881314,-0.0418932587,0.091757603,-0.1342888921,-0.0532308593,-0.0178111736,-0.025891291,-0.3486146033,-0.2984477878,0.2000667304,-0.3020641506,-0.2819243073,0.2392926961,0.13543652,0.4544383287,0.040127337,-0.3937389553,0.5808199644,-0.1391113997,-0.1142625138,0.1804570556,0.3090633154,-0.0905662403,0.1654050052,-0.0495853089,-0.1657474041,-0.2646009028,-0.1806016117,-0.2320661098,0.1591044813,-0.1456702203,0.062785618,0.010418159,0.8047193289,0.0654786751,-0.2943529487,0.1827455461,0.1267124414,0.2218532413,-0.0006962097,-0.5895354152,0.2010116726,-0.0986168161,0.2215839028,0.3626628816,0.0432908051,0.1737747043,-0.1139224693,0.1939601153,-0.1887793243,-0.3513721824,-0.1281636208,0.2005107254,0.3849678338,0.0900534242,0.1219887882,0.0273914002,-0.0954923853,-0.0095352363,0.354700923,0.1222476214,0.0885325372,-0.1050639674,-0.0560128465,-0.2369755805,0.0535890795,-0.1932052821,0.4575626254,-0.0324502178,-0.1230478138,-0.0423158407,-0.0182903316,0.5026882291,0.1967388988,0.3927701116,0.1022480875,-0.4248227179,-0.239300549,-0.1275756657,0.0902222469,-0.0735764727,0.1285210401,0.1846208721,-0.2020288408,-0.5072082877,0.2025788128,0.1107909232,-0.1888125837,-0.1198265105,-0.1453297287,0.0249494147,-0.2902211249,-0.0777397379,-0.2248150408,0.1797920763,-0.4049384594,-0.023972258,-0.1545185298,-0.0788380057,0.1511588097,0.0657220259,0.0934651047,0.0443000272,0.4923811555,0.3122967482,0.3200487196,0.5650814176,0.615953207,-0.1010638103,0.0344186723,0.083505854,0.0163794942,0.0522822738,0.283159852,-0.0781762227,0.1111004427,0.0438617095,0.1107477918,-0.5960834026,0.1240014806,0.2019082308,0.0680205598,-0.3572423458,-0.0997587666,0.194607392,0.026873542,0.2006976753,0.1760098636,0.1702613533,-0.3972954154,-0.4666161835,0.0168187767,0.494730413,-0.073587589,0.222242713,-0.1423188895,-0.1103767976,0.0566480383,0.2778768539,-0.1172659621,-0.191495344,0.1790853292,0.0319309011,-0.1585556418,0.086094141,0.1152566522,-0.0740363374,0.2761102319,0.0473109744,0.5473350883,0.1399034262,0.0827611834,-0.1333879083,-0.2570032179,-0.2534717619,0.1469009668,-0.0401018672,0.3355665505,-0.2314390987,-0.1019250676,0.3119746745,-0.0665354207,-0.2452950776,0.0065798382,-0.6495981216,0.1033414304,-0.0357609317,-0.1031017154,0.3692237139,-0.2313642651,0.1811163127,0.319975853,-0.1902632564,0.1010997519,-0.3137949705,0.0383315757,-0.2302716672,0.2646715939,0.1008016616,-0.0336168595,-0.0603646114,-0.1086391211,-0.0407444052,-0.0539295301,0.0338524245,0.0507715642,-0.0552913956,-0.0493358076,-0.3194903433,0.011819873,0.0131511372,-0.1511107236,0.1543610096,0.029080661,-0.0549284481,0.3291851878,-0.0549553446,-0.1067547053,-0.0585501045,0.4145893157,-0.2623612285,0.0229859632,0.3038798869,0.4151066542,-0.0175478105,-0.281327337,-0.0685149655,0.090806812,-0.2448111773,0.2701681852,0.1164219454,-0.0899813995,0.380168438,0.5611450076,0.3422602117,-0.1498246193,0.0122458581,-0.2868360281,-0.1911179423,0.0608151965,-0.1435581148,0.1625947654,0.027714828,0.4404928684,-0.1747053266,0.3926444948,-0.3059358299,0.2749380469,-0.094757244,-0.0000577473,0.2123723328,-0.1585534215,0.1591589153,-0.0921347439,0.0412250496,-0.0340272896,-0.5342020392,-0.2137607485,-0.11785952,0.0945381075,0.0980818197,-0.2711153924,0.1690746993,-0.1080352366,0.0807959363,-0.1697109789,0.3470467329,0.0660692304,-0.1222601384,0.0229023825,0.0703012571,0.0658251643,-0.1689609438,0.1390585154,0.0714529082,-0.0677270144,-0.004992947,-0.0410775393,0.0441779569,0.0208381023,0.1150302812,0.0102870604,-0.0216585528,0.1182223186,0.5093998909,-0.3056085408,-0.2525115013,0.2638110816,0.285666734,0.1686179191,0.0513615571,0.0614277981,-0.1690228879,0.1712695211,-0.2953843772,-0.0795397311,-0.0377579108,0.1637812108,0.0857127979,0.0499729849,0.3570745885,-0.2151893675,-0.2337022722,0.3549879789,0.3251188695,-0.1116924211,0.1748533994,0.4612625837,0.0847785622,0.2086670697,0.1075345576,-0.1653207839,0.1047748998,0.3765080571,-0.2396270782,0.680370748,0.0646547824,0.006373262,0.24075526,-0.4143283367,0.1107713133,0.2690977156,-0.1568296403,0.0244206265,-0.2630802393,0.1377333105,-0.3402009606,-0.2731530666,-0.1855362207,0.1984923333,-0.0655509979,-0.125413537,-0.2556758225,-0.1166857257,-0.0780742988,0.0273622628,-0.0752226934,-0.1022672802,0.2595925927,0.173260808,-0.1564741433,-0.4046477079,-0.3694777787,0.0161443371,-0.098320432,-0.1228213236,0.256008625,0.4048367143,-0.2123374194,0.1384949088,0.4051626027,0.4715678394,0.3013881147,-0.0192856137,-0.0860338286,0.0871971771,-0.0851401165,-0.2775004804,0.4197260141,-0.2568069994,-0.157341525,0.0049490035,0.1699633896,-0.1862421781,-0.3644252717,0.1278567314,-0.3126043677,-0.2835620046,0.1925962269,-0.0886972845,-0.0576487854,-0.0464047492,0.3801314831,-0.0463565886,0.2271073312,0.6677170396,-0.0275353566,0.1280924529,-0.24235861,0.1004886553,0.085262239,0.719905436,0.2930560708,-0.0817147717,-0.286534816,-0.5099948645,-0.7815919518,0.0482606329,0.1542916,0.4362874329,0.0736527294,0.0439882986,0.1192256138,-0.0225296728,0.1537165791,0.1220184043,0.2925755382,-0.215133965,-0.3548647165,0.1796959639,-0.0950149968,-0.1906720698,-0.0053808568,-0.3220204711,0.1466882378,-0.2788923085,0.0247029401,-0.2417231798,0.423589617,0.0029747856,0.2875816226,0.2188119441,0.0880910382,0.3063708842,0.0107983965,-0.063057445,-0.4197324216,-0.3112786114,0.014537733,0.0860736221,-0.1074681729,0.3646843135,-0.0537015609,-0.2685222924,-0.4432088137,0.1050856933,0.1143686846,0.20233652,-0.1769937277,0.222210601,-0.2147685587,-0.1733141243,0.3443754017,0.4282422066,-0.0292369928,0.1878373325,-0.210676685,-0.3873319328,0.2546341419,-0.2887773514,-0.1629846394,-0.4629845619,0.1772362888,-0.0954384282,-0.1698636413,-0.4217805266,0.1528434455,0.1712173522,-0.1016812101,-0.2147470862,0.3726132214,-0.0460546836,-0.0684140325,-0.0185811818,0.1694674492,0.1803847849,-0.3343190551,-0.1139617413,-0.0598337278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1977","title":"ModuleNotFoundError: No module named 'apache_beam' for wikipedia datasets ","comments":"I sometimes also get this error with other languages of the same dataset:\r\n\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n\r\n@lhoestq \r\n","body":"Hi\r\nI am trying to run run_mlm.py code [1] of huggingface with following \"wikipedia\"\/ \"20200501.aa\" dataset:\r\n\r\n`python run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.aa --do_train --do_eval --output_dir \/tmp\/test-mlm --max_seq_length 256\r\n`\r\n\r\nI am getting this error, but as per documentation, huggingface dataset provide processed version of this dataset and users can load it without requiring setup extra settings for apache-beam. could you help me please to load this dataset? \r\nDo you think I can run run_ml.py with this dataset? or anyway I could subsample and train the model? I greatly appreciate providing the processed version of all languages for this dataset, which allow the user to use them without setting up apache-beam,. thanks \r\n\r\nI really appreciate your help.\r\n@lhoestq \r\n\r\nthanks.\r\n\r\n[1] https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py\r\n\r\nerror I get: \r\n\r\n```\r\n>>> import datasets \r\n>>> datasets.load_dataset(\"wikipedia\", \"20200501.aa\")\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/dara\/temp\/cache_home_2\/datasets\/wikipedia\/20200501.aa\/1.0.0\/4021357e28509391eab2f8300d9b689e7e8f3a877ebb3d354b01577d497ebc63...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 1099, in _download_and_prepare\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n\r\n```","comment_length":55,"text":"ModuleNotFoundError: No module named 'apache_beam' for wikipedia datasets \nHi\r\nI am trying to run run_mlm.py code [1] of huggingface with following \"wikipedia\"\/ \"20200501.aa\" dataset:\r\n\r\n`python run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.aa --do_train --do_eval --output_dir \/tmp\/test-mlm --max_seq_length 256\r\n`\r\n\r\nI am getting this error, but as per documentation, huggingface dataset provide processed version of this dataset and users can load it without requiring setup extra settings for apache-beam. could you help me please to load this dataset? \r\nDo you think I can run run_ml.py with this dataset? or anyway I could subsample and train the model? I greatly appreciate providing the processed version of all languages for this dataset, which allow the user to use them without setting up apache-beam,. thanks \r\n\r\nI really appreciate your help.\r\n@lhoestq \r\n\r\nthanks.\r\n\r\n[1] https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py\r\n\r\nerror I get: \r\n\r\n```\r\n>>> import datasets \r\n>>> datasets.load_dataset(\"wikipedia\", \"20200501.aa\")\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/dara\/temp\/cache_home_2\/datasets\/wikipedia\/20200501.aa\/1.0.0\/4021357e28509391eab2f8300d9b689e7e8f3a877ebb3d354b01577d497ebc63...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 1099, in _download_and_prepare\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n\r\n```\nI sometimes also get this error with other languages of the same dataset:\r\n\r\n File \"\/dara\/libs\/anaconda3\/envs\/code\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n\r\n@lhoestq \r\n","embeddings":[-0.2131412625,-0.3382999003,0.001538666,0.3398338854,0.2197447717,0.235442251,0.2507144213,0.2680625618,0.1814851314,-0.0144523093,-0.0359319523,0.0364899971,-0.171056658,0.1019865423,0.1627780348,-0.4244243205,0.1850165129,-0.1324697435,-0.28719607,-0.0623331778,-0.2229189277,0.3467077613,-0.2787266672,0.0804548115,-0.0931125656,-0.1117043048,-0.0233219042,0.177757442,-0.054072462,-0.1476043314,0.1096976846,-0.2670561969,0.2810361981,0.3396104574,-0.0001204881,0.0703728497,0.3882136047,-0.210515365,-0.3313601017,-0.350505054,0.1729956865,-0.1294179261,0.138146013,-0.2263075113,-0.2553460896,-0.1438835114,0.1525821835,-0.2568327487,0.6282421947,-0.0624543279,0.1526159346,0.3100936413,0.3425408602,-0.0296261031,0.1350966841,0.2360670418,-0.0238424726,0.3365868032,-0.1699708253,-0.1384140104,0.1368692964,0.5407437086,-0.2211783677,-0.0125280395,0.4623476565,-0.2002913803,0.0987885296,-0.5092285275,0.116753906,0.2758406997,0.5031315088,-0.26792714,-0.1555137187,-0.0499551333,0.0338076949,-0.03536411,0.0232853089,0.22989133,-0.1179501638,0.0734570548,0.2295575142,-0.3673401475,-0.1967659593,0.2305391133,0.2532696426,0.2152464986,-0.0669510439,0.2225007564,0.1748564541,-0.1025937721,-0.4621145427,-0.0402746014,0.156126067,0.4422611892,-0.231231764,0.002823259,-0.1473004669,0.2895689607,0.2478105426,-0.1703071743,-0.4538461268,0.115305379,0.271007508,0.0321764424,0.0142410379,0.0148335658,0.1804619879,-0.1404565275,0.2022216916,0.1521080583,0.0284997057,0.0471678078,-0.0834322572,-0.0926120132,-0.6663607955,-0.0144305415,0.2571786344,-0.1104114279,-0.1701249182,0.0090121748,-0.548682034,-0.2558385134,-0.1178343967,0.4914840758,-0.0876239315,-0.0432364196,0.3066846132,0.2725155652,-0.3196043372,-0.332947433,-0.0259876121,0.3970681727,-0.4042425454,0.2580560446,0.1624427736,-0.185209617,0.3300159872,-0.0969913378,0.0612081997,-0.0126665076,0.093832843,0.0160338152,-0.2645788193,0.1376620084,0.2601518929,0.2935320139,0.3347854614,-0.1072975174,-0.1303373873,-0.0130061731,-0.2017825842,-0.0785866901,0.0849881843,0.0104481429,-0.1748300493,0.1516970396,-0.1632316411,0.2782359719,-0.0621964559,0.0589513369,-0.0485334769,0.251856178,-0.0731160566,-0.1545423865,0.4467759132,0.7152784467,-0.0924548283,-0.1832813919,-0.1248655617,0.0888795927,-0.1330818832,0.0378335416,-0.0201482009,0.2038291842,-0.1944642514,0.0074124318,0.2688691616,-0.4080123603,0.0281282626,0.0515699685,0.0758030936,0.0689504147,0.1730379313,-0.1063590124,-0.2876467109,-0.0430139713,-0.0504935347,0.2006358802,0.1387092173,-0.0972392559,-0.1221533567,-0.2608413398,0.0938873217,0.233961612,0.104605414,-0.0214352626,0.0598472171,0.5199069977,0.225365594,-0.1468617469,0.1068723127,0.3651059866,-0.103207089,0.2124211788,0.0801012442,-0.2059350163,-0.1896687299,0.0508289747,-0.2155381143,0.5097803473,-0.1102500632,-0.0496578589,-0.2062759697,-0.0623012409,-0.1766985357,-0.5256921053,0.0174633265,-0.1082244441,0.13924779,0.3834108114,-0.1024011075,0.2250570953,0.005566176,0.199066326,-0.8539184928,0.1584817022,-0.1625045389,0.0418303609,0.1188349947,0.2998978794,0.1418554336,-0.2450328022,0.1450186968,0.2261148095,0.0557804815,0.0124180801,0.0420554914,-0.1097348183,0.1478796601,-0.3845858276,0.2552294135,0.0223892294,0.0874389112,0.0594493076,0.0067111249,0.2668873668,0.154275924,0.3922149241,0.1603031754,0.1677394658,-0.0652832985,0.0154548204,-0.2106283009,-0.3001226485,0.3233882189,0.1925318241,0.331528157,-0.2222321182,-0.1197232157,-0.3948613703,0.2577231824,-0.025115706,0.227075994,-0.056015756,-0.6535443664,0.290453583,0.1373188943,-0.2333566248,0.1929884106,0.2338313311,-0.1762011796,0.2047182173,0.175831452,0.0648411438,0.2654977441,0.088253893,0.4785864651,-0.0819546282,-0.2294490039,-0.1114422157,-0.2178023607,-0.1793961972,-0.2691538036,0.1448301375,-0.3526336253,0.1320126653,-0.2323406637,-0.5582601428,-0.3051196039,-0.021738369,-0.5030117035,-0.0974111259,-0.3454745114,0.1530036777,0.1994351745,0.4281540513,0.2881691754,-0.0364836417,0.0833428353,-0.1567291617,-0.1880505979,-0.1502944678,-0.2787014246,-0.0067598391,0.3752514124,0.1393429637,0.1550534964,-0.2120730877,-0.0861392841,-0.0383539386,-0.4270922244,0.2244886011,-0.2542711496,0.2628424466,0.188561663,0.4130579233,-0.1757036746,-0.0502504744,0.3604340255,0.0225338917,-0.0979226083,-0.0710498542,0.0758107305,0.0790889561,0.0070314542,-0.287669003,-0.274710238,-0.3001061082,0.1849595755,0.113288939,-0.0709090233,0.3787707984,0.2008128613,0.1534817517,-0.0516646802,0.0577601455,-0.1449172944,0.0399068967,0.3708283007,-0.2393716276,-0.3559527993,0.1864742041,-0.1147243679,0.2194794416,-0.0391898118,-0.2681618631,-0.1775782257,0.3354676366,-0.1826422513,-0.0108241942,0.3228735626,0.2339251786,-0.1341770738,0.1562602818,-0.1174859405,-0.0410772935,-0.0801702961,-0.4929061532,0.3673178256,0.248804003,0.6314407587,-0.0803147256,0.8895260096,0.2763777077,0.2054400295,0.1781024039,-0.026158208,0.3253340125,-0.0767209008,-0.291257441,-0.1696092784,0.016621178,0.1385060251,0.3068870008,-0.0287208222,0.0690138265,-0.5301668644,-0.256786257,-0.1413567215,-0.2349941134,0.0289408118,-0.1157510951,0.4020377994,0.0677074343,0.2030601799,0.0631277487,-0.0334787332,0.3315420151,0.5468326807,0.0733549148,0.1779473871,-0.0101673659,-0.1991823167,-0.5310757756,0.402153641,-0.1498976052,0.0007159312,-0.09026061,-0.106420517,0.1710890383,0.0208133869,0.7178604603,-0.0888773277,-0.0936146677,0.1358968168,0.0015432175,-0.7305615544,0.105054751,0.0151006253,0.1899862736,0.1572462767,0.2613230944,-0.470985353,-0.1298935711,0.4364516735,0.1856455952,-0.1291808039,-0.1521936506,-0.5246492028,-0.1736597121,-0.4372420311,-0.1901437193,-0.0122950794,0.1366353631,0.1450327486,0.3733565807,0.026763536,-0.1719457209,0.2619962096,0.2708519101,0.3428224921,-0.0041506607,-0.1550676674,0.1976625025,0.3420525491,-0.0263307765,0.1929224879,-0.0081908489,-0.4109748006,-0.0481527373,0.0178303607,0.2324728817,0.1311837286,-0.1344868243,0.1006974205,0.1931822002,-0.0921949372,-0.1482954919,-0.0953785628,0.2387633771,0.2665473521,-0.3958492875,-0.416685164,0.6590346098,0.0648799166,0.0568404347,0.1441074014,0.2705354691,-0.329798907,0.3083680272,0.1000937298,1.0329909325,-0.0303210989,0.186760515,0.2142025977,0.1614112407,0.8284099698,-0.446864903,0.1807176024,-0.32085675,0.1913253516,-0.0089974497,-0.072842434,0.438472122,0.0767244548,-0.1231356338,0.3307362795,0.220532313,0.0837881789,-0.1420511007,0.4586176872,-0.0194509104,-0.2219656259,-0.4577875435,0.0059212423,-0.1063210741,0.6052085757,-0.2764023542,-0.1886194646,-0.230437845,-0.2535758317,-0.1999853551,-0.0153510477,-0.353202343,0.2437590212,-0.1092242301,-0.2311705649,0.3377098441,0.2006301433,0.2454082221,0.0233236402,-0.333147645,0.1841923594,-0.3711706996,-0.4309200644,-0.2007158101,0.0908433422,0.321277529,-0.1738765389,-0.1705289632,0.0283660591,-0.0744693056,0.0023991931,-0.1516765058,-0.1976522803,0.2312403768,0.0417475179,-0.1732554287,0.2482134402,0.1364869326,0.107802324,0.046610117,-0.1768838018,-0.1747897416,-0.0595203787,0.1811487228,-0.1508108675,-0.0135541055,0.4774492085,0.1515924633,-0.0545032322,0.5850344896,0.432033211,-0.1466626525,-0.148063302,0.115029417,0.0594775006,-0.3519896269,-0.0794999078,0.3436746299,0.2288256735,-0.1090679541,0.015570431,0.0660213009,-0.2205022871,0.1175459027,-0.475569278,-0.1650405675,0.445112735,-0.222631216,-0.1648510396,0.2192070335,0.2640362978,0.2220729738,-0.0910979137,-0.1593146324,-0.1072032452,0.1454121917,0.0105217444,0.5708838701,0.0572469123,0.1781495661,0.1490954161,0.044193089,0.0211147405,-0.2832331359,-0.0636303276,-0.0720102042,0.2172734886,-0.1876900941,-0.1258059889,0.0418041423,-0.2209444195,-0.0505203083,-0.1861306727,0.0226196516,-0.0529786982,0.0284197684,0.0480478778,0.044134561,0.2348095477,-0.436896801,0.043743033,-0.0688095614,0.1023072302,0.2306390107,-0.0978876874,0.3817752302,0.0426771007,-0.3169856071,0.009664502,0.0188739188,-0.092812784,0.2114265263,0.0173062813,-0.1756639481,-0.0178821106,0.0450209603,0.2304658592,-0.1178769395,0.0271782931,0.2184094042,0.0492903702,-0.4587966204,0.3201698065,0.4787884057,-0.2735529244,-0.087610282,0.2480540425,0.0477450676,0.3263320327,0.0658330768,0.1816511601,0.2887360156,0.1045155153,0.3736536801,0.0241869651,0.1239091307,0.1329514682,0.244455263,0.1555859447,0.2112366855,0.407526195,-0.219760865,0.2934308648,-0.2903069556,0.0693989173,-0.0584796071,-0.1634107381,0.1687483042,0.3123420775,0.0832757428,0.1723279804,-0.2653711736,0.488825351,-0.0398289859,-0.2026303262,-0.193253383,0.2618729472,-0.0716298446,-0.1267788708,-0.1107864901,-0.2217423916,-0.1446863264,-0.0342082828,0.144216314,-0.6080909371,0.1658254266,0.1780467182,-0.1325175017,-0.1697184891,-0.1060797572,0.2342267632,0.1096986756,-0.0739862248,-0.0205644127,0.022054147,0.0508319288,-0.0317686833,0.0707945973,0.2845675349,0.0694999099,-0.326143533,-0.088607505,0.0281688627,0.0773354694,0.1540694684,-0.0325725488,0.1480217278,0.2344671637,0.1302425265,0.0422345959,-0.0877723247,-0.176692009,0.115390949,0.0173509903,-0.3497005105,-0.1444838345,-0.5136170983,0.119173795,-0.3950394988,0.1520346105,-0.6309111118,0.3177124262,0.2420934588,0.1043871045,-0.2357274741,-0.2944581807,-0.0183674507,-0.2088824213,0.4825663865,0.2201556116,0.1175153553,-0.2192915827,-0.3498180509,-0.5625340343,-0.0253219567,-0.1389907598,-0.0824832767,-0.2705150545,-0.0041030897,0.2746976912,0.2762576342,-0.0568132512,-0.0429306477,-0.0837091729,0.0761819929,-0.1995844096,-0.2112319767,-0.3830574453,0.1804929376,-0.1026849598,-0.0495564826,-0.080513604,-0.029824134,-0.1119437888,-0.2530980706,-0.1765733659,0.1710343808,-0.1491254419,0.3884039223,-0.0165548455,0.3359567821,0.0793647543,-0.1874970794,-0.112381205,0.0434571169,-0.054018613,0.3798463643,0.0634100959,0.2255599797,-0.278462708,-0.3539013863,-0.3619379699,0.5960179567,0.1159183383,-0.0575564951,-0.3417952955,0.1050301641,-0.0232628975,0.176586628,-0.0566634461,0.2489111871,-0.1592449993,0.0920145363,-0.4346440434,-0.0472101308,0.5104724765,-0.8678759933,-0.2605215907,0.0031878252,0.0732161701,-0.0388213284,-0.2040875405,-0.6355007887,0.1892228425,0.3016792238,0.1750743985,-0.0031793322,0.2277702987,-0.1221880913,-0.0525795594,-0.056074027,-0.3247557282,-0.0901467651,-0.194520995,0.0034505054,-0.2350435704]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1977","title":"ModuleNotFoundError: No module named 'apache_beam' for wikipedia datasets ","comments":"Hi ! Thanks for reporting\r\nSome wikipedia configurations do require the user to have `apache_beam` in order to parse the wikimedia data.\r\n\r\nOn the other hand regarding your second issue\r\n```\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\nI've never experienced this, can you open a new issue for this specific error and provide more details please ?\r\nFor example what script did you use to get this, what language did you use, what's your environment details (os, python version, pyarrow version)..","body":"Hi\r\nI am trying to run run_mlm.py code [1] of huggingface with following \"wikipedia\"\/ \"20200501.aa\" dataset:\r\n\r\n`python run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.aa --do_train --do_eval --output_dir \/tmp\/test-mlm --max_seq_length 256\r\n`\r\n\r\nI am getting this error, but as per documentation, huggingface dataset provide processed version of this dataset and users can load it without requiring setup extra settings for apache-beam. could you help me please to load this dataset? \r\nDo you think I can run run_ml.py with this dataset? or anyway I could subsample and train the model? I greatly appreciate providing the processed version of all languages for this dataset, which allow the user to use them without setting up apache-beam,. thanks \r\n\r\nI really appreciate your help.\r\n@lhoestq \r\n\r\nthanks.\r\n\r\n[1] https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py\r\n\r\nerror I get: \r\n\r\n```\r\n>>> import datasets \r\n>>> datasets.load_dataset(\"wikipedia\", \"20200501.aa\")\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/dara\/temp\/cache_home_2\/datasets\/wikipedia\/20200501.aa\/1.0.0\/4021357e28509391eab2f8300d9b689e7e8f3a877ebb3d354b01577d497ebc63...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 1099, in _download_and_prepare\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n\r\n```","comment_length":84,"text":"ModuleNotFoundError: No module named 'apache_beam' for wikipedia datasets \nHi\r\nI am trying to run run_mlm.py code [1] of huggingface with following \"wikipedia\"\/ \"20200501.aa\" dataset:\r\n\r\n`python run_mlm.py --model_name_or_path bert-base-multilingual-cased --dataset_name wikipedia --dataset_config_name 20200501.aa --do_train --do_eval --output_dir \/tmp\/test-mlm --max_seq_length 256\r\n`\r\n\r\nI am getting this error, but as per documentation, huggingface dataset provide processed version of this dataset and users can load it without requiring setup extra settings for apache-beam. could you help me please to load this dataset? \r\nDo you think I can run run_ml.py with this dataset? or anyway I could subsample and train the model? I greatly appreciate providing the processed version of all languages for this dataset, which allow the user to use them without setting up apache-beam,. thanks \r\n\r\nI really appreciate your help.\r\n@lhoestq \r\n\r\nthanks.\r\n\r\n[1] https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py\r\n\r\nerror I get: \r\n\r\n```\r\n>>> import datasets \r\n>>> datasets.load_dataset(\"wikipedia\", \"20200501.aa\")\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/dara\/temp\/cache_home_2\/datasets\/wikipedia\/20200501.aa\/1.0.0\/4021357e28509391eab2f8300d9b689e7e8f3a877ebb3d354b01577d497ebc63...\r\nTraceback (most recent call last):\r\n File \"\", line 1, in \r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/dara\/temp\/libs\/anaconda3\/envs\/codes\/lib\/python3.7\/site-packages\/datasets-1.3.0-py3.7.egg\/datasets\/builder.py\", line 1099, in _download_and_prepare\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n\r\n```\nHi ! Thanks for reporting\r\nSome wikipedia configurations do require the user to have `apache_beam` in order to parse the wikimedia data.\r\n\r\nOn the other hand regarding your second issue\r\n```\r\nOSError: Memory mapping file failed: Cannot allocate memory\r\n```\r\nI've never experienced this, can you open a new issue for this specific error and provide more details please ?\r\nFor example what script did you use to get this, what language did you use, what's your environment details (os, python version, pyarrow version)..","embeddings":[-0.2131412625,-0.3382999003,0.001538666,0.3398338854,0.2197447717,0.235442251,0.2507144213,0.2680625618,0.1814851314,-0.0144523093,-0.0359319523,0.0364899971,-0.171056658,0.1019865423,0.1627780348,-0.4244243205,0.1850165129,-0.1324697435,-0.28719607,-0.0623331778,-0.2229189277,0.3467077613,-0.2787266672,0.0804548115,-0.0931125656,-0.1117043048,-0.0233219042,0.177757442,-0.054072462,-0.1476043314,0.1096976846,-0.2670561969,0.2810361981,0.3396104574,-0.0001204881,0.0703728497,0.3882136047,-0.210515365,-0.3313601017,-0.350505054,0.1729956865,-0.1294179261,0.138146013,-0.2263075113,-0.2553460896,-0.1438835114,0.1525821835,-0.2568327487,0.6282421947,-0.0624543279,0.1526159346,0.3100936413,0.3425408602,-0.0296261031,0.1350966841,0.2360670418,-0.0238424726,0.3365868032,-0.1699708253,-0.1384140104,0.1368692964,0.5407437086,-0.2211783677,-0.0125280395,0.4623476565,-0.2002913803,0.0987885296,-0.5092285275,0.116753906,0.2758406997,0.5031315088,-0.26792714,-0.1555137187,-0.0499551333,0.0338076949,-0.03536411,0.0232853089,0.22989133,-0.1179501638,0.0734570548,0.2295575142,-0.3673401475,-0.1967659593,0.2305391133,0.2532696426,0.2152464986,-0.0669510439,0.2225007564,0.1748564541,-0.1025937721,-0.4621145427,-0.0402746014,0.156126067,0.4422611892,-0.231231764,0.002823259,-0.1473004669,0.2895689607,0.2478105426,-0.1703071743,-0.4538461268,0.115305379,0.271007508,0.0321764424,0.0142410379,0.0148335658,0.1804619879,-0.1404565275,0.2022216916,0.1521080583,0.0284997057,0.0471678078,-0.0834322572,-0.0926120132,-0.6663607955,-0.0144305415,0.2571786344,-0.1104114279,-0.1701249182,0.0090121748,-0.548682034,-0.2558385134,-0.1178343967,0.4914840758,-0.0876239315,-0.0432364196,0.3066846132,0.2725155652,-0.3196043372,-0.332947433,-0.0259876121,0.3970681727,-0.4042425454,0.2580560446,0.1624427736,-0.185209617,0.3300159872,-0.0969913378,0.0612081997,-0.0126665076,0.093832843,0.0160338152,-0.2645788193,0.1376620084,0.2601518929,0.2935320139,0.3347854614,-0.1072975174,-0.1303373873,-0.0130061731,-0.2017825842,-0.0785866901,0.0849881843,0.0104481429,-0.1748300493,0.1516970396,-0.1632316411,0.2782359719,-0.0621964559,0.0589513369,-0.0485334769,0.251856178,-0.0731160566,-0.1545423865,0.4467759132,0.7152784467,-0.0924548283,-0.1832813919,-0.1248655617,0.0888795927,-0.1330818832,0.0378335416,-0.0201482009,0.2038291842,-0.1944642514,0.0074124318,0.2688691616,-0.4080123603,0.0281282626,0.0515699685,0.0758030936,0.0689504147,0.1730379313,-0.1063590124,-0.2876467109,-0.0430139713,-0.0504935347,0.2006358802,0.1387092173,-0.0972392559,-0.1221533567,-0.2608413398,0.0938873217,0.233961612,0.104605414,-0.0214352626,0.0598472171,0.5199069977,0.225365594,-0.1468617469,0.1068723127,0.3651059866,-0.103207089,0.2124211788,0.0801012442,-0.2059350163,-0.1896687299,0.0508289747,-0.2155381143,0.5097803473,-0.1102500632,-0.0496578589,-0.2062759697,-0.0623012409,-0.1766985357,-0.5256921053,0.0174633265,-0.1082244441,0.13924779,0.3834108114,-0.1024011075,0.2250570953,0.005566176,0.199066326,-0.8539184928,0.1584817022,-0.1625045389,0.0418303609,0.1188349947,0.2998978794,0.1418554336,-0.2450328022,0.1450186968,0.2261148095,0.0557804815,0.0124180801,0.0420554914,-0.1097348183,0.1478796601,-0.3845858276,0.2552294135,0.0223892294,0.0874389112,0.0594493076,0.0067111249,0.2668873668,0.154275924,0.3922149241,0.1603031754,0.1677394658,-0.0652832985,0.0154548204,-0.2106283009,-0.3001226485,0.3233882189,0.1925318241,0.331528157,-0.2222321182,-0.1197232157,-0.3948613703,0.2577231824,-0.025115706,0.227075994,-0.056015756,-0.6535443664,0.290453583,0.1373188943,-0.2333566248,0.1929884106,0.2338313311,-0.1762011796,0.2047182173,0.175831452,0.0648411438,0.2654977441,0.088253893,0.4785864651,-0.0819546282,-0.2294490039,-0.1114422157,-0.2178023607,-0.1793961972,-0.2691538036,0.1448301375,-0.3526336253,0.1320126653,-0.2323406637,-0.5582601428,-0.3051196039,-0.021738369,-0.5030117035,-0.0974111259,-0.3454745114,0.1530036777,0.1994351745,0.4281540513,0.2881691754,-0.0364836417,0.0833428353,-0.1567291617,-0.1880505979,-0.1502944678,-0.2787014246,-0.0067598391,0.3752514124,0.1393429637,0.1550534964,-0.2120730877,-0.0861392841,-0.0383539386,-0.4270922244,0.2244886011,-0.2542711496,0.2628424466,0.188561663,0.4130579233,-0.1757036746,-0.0502504744,0.3604340255,0.0225338917,-0.0979226083,-0.0710498542,0.0758107305,0.0790889561,0.0070314542,-0.287669003,-0.274710238,-0.3001061082,0.1849595755,0.113288939,-0.0709090233,0.3787707984,0.2008128613,0.1534817517,-0.0516646802,0.0577601455,-0.1449172944,0.0399068967,0.3708283007,-0.2393716276,-0.3559527993,0.1864742041,-0.1147243679,0.2194794416,-0.0391898118,-0.2681618631,-0.1775782257,0.3354676366,-0.1826422513,-0.0108241942,0.3228735626,0.2339251786,-0.1341770738,0.1562602818,-0.1174859405,-0.0410772935,-0.0801702961,-0.4929061532,0.3673178256,0.248804003,0.6314407587,-0.0803147256,0.8895260096,0.2763777077,0.2054400295,0.1781024039,-0.026158208,0.3253340125,-0.0767209008,-0.291257441,-0.1696092784,0.016621178,0.1385060251,0.3068870008,-0.0287208222,0.0690138265,-0.5301668644,-0.256786257,-0.1413567215,-0.2349941134,0.0289408118,-0.1157510951,0.4020377994,0.0677074343,0.2030601799,0.0631277487,-0.0334787332,0.3315420151,0.5468326807,0.0733549148,0.1779473871,-0.0101673659,-0.1991823167,-0.5310757756,0.402153641,-0.1498976052,0.0007159312,-0.09026061,-0.106420517,0.1710890383,0.0208133869,0.7178604603,-0.0888773277,-0.0936146677,0.1358968168,0.0015432175,-0.7305615544,0.105054751,0.0151006253,0.1899862736,0.1572462767,0.2613230944,-0.470985353,-0.1298935711,0.4364516735,0.1856455952,-0.1291808039,-0.1521936506,-0.5246492028,-0.1736597121,-0.4372420311,-0.1901437193,-0.0122950794,0.1366353631,0.1450327486,0.3733565807,0.026763536,-0.1719457209,0.2619962096,0.2708519101,0.3428224921,-0.0041506607,-0.1550676674,0.1976625025,0.3420525491,-0.0263307765,0.1929224879,-0.0081908489,-0.4109748006,-0.0481527373,0.0178303607,0.2324728817,0.1311837286,-0.1344868243,0.1006974205,0.1931822002,-0.0921949372,-0.1482954919,-0.0953785628,0.2387633771,0.2665473521,-0.3958492875,-0.416685164,0.6590346098,0.0648799166,0.0568404347,0.1441074014,0.2705354691,-0.329798907,0.3083680272,0.1000937298,1.0329909325,-0.0303210989,0.186760515,0.2142025977,0.1614112407,0.8284099698,-0.446864903,0.1807176024,-0.32085675,0.1913253516,-0.0089974497,-0.072842434,0.438472122,0.0767244548,-0.1231356338,0.3307362795,0.220532313,0.0837881789,-0.1420511007,0.4586176872,-0.0194509104,-0.2219656259,-0.4577875435,0.0059212423,-0.1063210741,0.6052085757,-0.2764023542,-0.1886194646,-0.230437845,-0.2535758317,-0.1999853551,-0.0153510477,-0.353202343,0.2437590212,-0.1092242301,-0.2311705649,0.3377098441,0.2006301433,0.2454082221,0.0233236402,-0.333147645,0.1841923594,-0.3711706996,-0.4309200644,-0.2007158101,0.0908433422,0.321277529,-0.1738765389,-0.1705289632,0.0283660591,-0.0744693056,0.0023991931,-0.1516765058,-0.1976522803,0.2312403768,0.0417475179,-0.1732554287,0.2482134402,0.1364869326,0.107802324,0.046610117,-0.1768838018,-0.1747897416,-0.0595203787,0.1811487228,-0.1508108675,-0.0135541055,0.4774492085,0.1515924633,-0.0545032322,0.5850344896,0.432033211,-0.1466626525,-0.148063302,0.115029417,0.0594775006,-0.3519896269,-0.0794999078,0.3436746299,0.2288256735,-0.1090679541,0.015570431,0.0660213009,-0.2205022871,0.1175459027,-0.475569278,-0.1650405675,0.445112735,-0.222631216,-0.1648510396,0.2192070335,0.2640362978,0.2220729738,-0.0910979137,-0.1593146324,-0.1072032452,0.1454121917,0.0105217444,0.5708838701,0.0572469123,0.1781495661,0.1490954161,0.044193089,0.0211147405,-0.2832331359,-0.0636303276,-0.0720102042,0.2172734886,-0.1876900941,-0.1258059889,0.0418041423,-0.2209444195,-0.0505203083,-0.1861306727,0.0226196516,-0.0529786982,0.0284197684,0.0480478778,0.044134561,0.2348095477,-0.436896801,0.043743033,-0.0688095614,0.1023072302,0.2306390107,-0.0978876874,0.3817752302,0.0426771007,-0.3169856071,0.009664502,0.0188739188,-0.092812784,0.2114265263,0.0173062813,-0.1756639481,-0.0178821106,0.0450209603,0.2304658592,-0.1178769395,0.0271782931,0.2184094042,0.0492903702,-0.4587966204,0.3201698065,0.4787884057,-0.2735529244,-0.087610282,0.2480540425,0.0477450676,0.3263320327,0.0658330768,0.1816511601,0.2887360156,0.1045155153,0.3736536801,0.0241869651,0.1239091307,0.1329514682,0.244455263,0.1555859447,0.2112366855,0.407526195,-0.219760865,0.2934308648,-0.2903069556,0.0693989173,-0.0584796071,-0.1634107381,0.1687483042,0.3123420775,0.0832757428,0.1723279804,-0.2653711736,0.488825351,-0.0398289859,-0.2026303262,-0.193253383,0.2618729472,-0.0716298446,-0.1267788708,-0.1107864901,-0.2217423916,-0.1446863264,-0.0342082828,0.144216314,-0.6080909371,0.1658254266,0.1780467182,-0.1325175017,-0.1697184891,-0.1060797572,0.2342267632,0.1096986756,-0.0739862248,-0.0205644127,0.022054147,0.0508319288,-0.0317686833,0.0707945973,0.2845675349,0.0694999099,-0.326143533,-0.088607505,0.0281688627,0.0773354694,0.1540694684,-0.0325725488,0.1480217278,0.2344671637,0.1302425265,0.0422345959,-0.0877723247,-0.176692009,0.115390949,0.0173509903,-0.3497005105,-0.1444838345,-0.5136170983,0.119173795,-0.3950394988,0.1520346105,-0.6309111118,0.3177124262,0.2420934588,0.1043871045,-0.2357274741,-0.2944581807,-0.0183674507,-0.2088824213,0.4825663865,0.2201556116,0.1175153553,-0.2192915827,-0.3498180509,-0.5625340343,-0.0253219567,-0.1389907598,-0.0824832767,-0.2705150545,-0.0041030897,0.2746976912,0.2762576342,-0.0568132512,-0.0429306477,-0.0837091729,0.0761819929,-0.1995844096,-0.2112319767,-0.3830574453,0.1804929376,-0.1026849598,-0.0495564826,-0.080513604,-0.029824134,-0.1119437888,-0.2530980706,-0.1765733659,0.1710343808,-0.1491254419,0.3884039223,-0.0165548455,0.3359567821,0.0793647543,-0.1874970794,-0.112381205,0.0434571169,-0.054018613,0.3798463643,0.0634100959,0.2255599797,-0.278462708,-0.3539013863,-0.3619379699,0.5960179567,0.1159183383,-0.0575564951,-0.3417952955,0.1050301641,-0.0232628975,0.176586628,-0.0566634461,0.2489111871,-0.1592449993,0.0920145363,-0.4346440434,-0.0472101308,0.5104724765,-0.8678759933,-0.2605215907,0.0031878252,0.0732161701,-0.0388213284,-0.2040875405,-0.6355007887,0.1892228425,0.3016792238,0.1750743985,-0.0031793322,0.2277702987,-0.1221880913,-0.0525795594,-0.056074027,-0.3247557282,-0.0901467651,-0.194520995,0.0034505054,-0.2350435704]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1973","title":"Question: what gets stored in the datasets cache and why is it so huge?","comments":"Echo'ing this observation: I have a few datasets in the neighborhood of 2GB CSVs uncompressed, and when I use something like `Dataset.save_to_disk()` it's ~18GB on disk.\r\n\r\nIf this is unexpected behavior, would be happy to help run debugging as needed.","body":"I'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!","comment_length":40,"text":"Question: what gets stored in the datasets cache and why is it so huge?\nI'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!\nEcho'ing this observation: I have a few datasets in the neighborhood of 2GB CSVs uncompressed, and when I use something like `Dataset.save_to_disk()` it's ~18GB on disk.\r\n\r\nIf this is unexpected behavior, would be happy to help run debugging as needed.","embeddings":[-0.0360606462,-0.0662569851,-0.1086274385,0.5382812619,0.1768519878,0.3046650887,-0.0687095299,0.256344825,-0.1498931944,-0.1115516722,-0.0044143545,-0.2384364158,-0.1599156111,-0.2295420617,0.1550879031,0.2438128591,0.1417650431,-0.0338777266,-0.0882983729,-0.0986522809,-0.1149732172,0.0371307395,0.0000266192,-0.1759963185,-0.5011152625,-0.1951541156,-0.1269186437,-0.0226079263,-0.1835439503,-0.1439898908,0.2084015608,-0.0900361612,0.3271475136,0.4473099709,-0.0001115048,-0.362196058,0.3227512538,-0.008409529,-0.3896471858,0.3312594295,-0.3773738742,-0.3923428059,-0.104653351,-0.1323722154,0.3790541589,-0.0305659622,-0.0775615722,-0.4155649245,-0.1001973376,0.236139521,0.2581386864,-0.2053128183,-0.3442173898,0.1373020858,0.1090577245,0.0737133399,-0.1573267132,-0.2792854011,0.2627800107,0.2356608063,-0.0979836285,0.1034083143,0.0833162814,-0.0201201383,0.2829732597,-0.0790503696,-0.3153593838,-0.2074168772,0.6619532704,0.2052862197,0.868317306,-0.3482807279,-0.0808232054,-0.2358752936,-0.1497370452,-0.230252862,0.1697393507,0.4599973857,0.0505639389,0.0040057399,-0.5909302831,-0.4143614471,-0.126802966,-0.133024171,-0.2453342229,-0.2898691297,-0.3402654529,0.0371112004,0.1638762504,-0.0343983844,0.3052972555,-0.3209097087,-0.0971701443,0.2020126134,-0.3501282036,-0.2808207273,-0.1091755852,0.4334006906,0.2432497591,0.0263104197,-0.2546927035,-0.1597532928,-0.1192888916,-0.049323611,0.2371750176,0.415301621,-0.1042509079,0.0482306182,-0.049397178,-0.348623395,-0.1449474841,-0.1352476478,0.1227397695,0.0719255283,0.4723917246,-0.4745317698,-0.1642237604,-0.3412801325,0.1787273735,0.2651431859,0.0549180731,0.0278724711,0.087542668,0.1612014174,-0.113969028,0.2380318344,-0.2866079211,-0.1712253392,0.0806446001,-0.1550655216,-0.2483567297,-0.1152959988,-0.0712154582,0.248961091,0.3250068128,-0.2040179968,0.2558794916,0.0902936906,-0.0581718944,-0.1221661717,0.1817219257,-0.3663792312,0.4594609737,0.3355002701,-0.1150589138,0.4209492803,-0.1799025387,0.1078725606,-0.2113251984,0.2094638646,-0.4173648953,-0.4349770248,-0.0622864813,0.0899464861,-0.2316663861,0.0077563892,-0.3615067303,0.0903249383,0.4243746996,0.1884943247,0.2217032462,-0.1052607223,-0.3592414558,-0.2606439292,-0.1278331131,0.1828977466,-0.5495997071,0.2316555083,-0.0171850193,0.2038089335,0.0265679266,0.3862605393,-0.0676840916,0.0514887571,-0.0738906711,-0.1387885064,-0.0273051374,0.0410684757,-0.6250160336,0.1765685976,0.3490022719,-0.0522845425,-0.0821781605,0.3963996768,0.0915794894,-0.0462785922,-0.1137974858,0.2983604968,-0.2702443004,-0.142247498,-0.3710263968,-0.1000405848,0.0486651435,-0.0800512284,0.0584715568,-0.2000578046,-0.0648278594,0.2171606123,-0.0068974877,0.1431343555,0.2213953137,0.380055666,-0.0400363021,0.2462723255,0.2545773387,0.0045563751,-0.3890997171,0.295788914,0.3068705499,-0.6919739842,0.1563736349,-0.1010126173,0.116968289,0.1049530879,-0.0330854319,-0.1894969791,-0.0202213936,0.08104489,0.0512779169,-0.1019686386,-0.0221285317,0.4722431898,0.003787111,-0.1720621735,-0.3683245778,-0.205806911,0.0847271383,0.0366322547,-0.1773812622,0.0202002451,0.0359575339,-0.1158996299,-0.1611272395,0.1403155178,0.1411769539,-0.1652727574,0.2010636926,0.5026519895,-0.0098972637,0.1133935004,0.3529167473,-0.0542698205,0.1295189708,-0.182449311,-0.2455490977,0.0227384232,-0.1056368575,0.0891257897,-0.0018707897,-0.1094690338,0.0670585856,0.0806366354,0.2521679997,0.0496883616,0.1556522697,-0.110239245,0.3004677892,0.3928951621,-0.0488347411,0.2108535618,0.4110551775,-0.2141384929,0.0022680063,0.163585484,-0.2870114148,-0.4692252874,0.0958473831,0.3619622886,0.4408290386,0.0798801929,0.3315306902,-0.1054326594,0.2634226084,-0.1867278963,0.0832477659,0.0455255657,0.0130668087,-0.1717012078,0.1634775698,-0.002815672,-0.0295009632,0.280049324,-0.0815472305,0.236097008,-0.0313673951,0.4049853981,-0.2644424736,-0.0564952083,-0.0029137924,0.3724592924,-0.2667140961,-0.1263842583,-0.1069569215,-0.0781468824,0.2320266664,-0.2643210888,-0.2530310452,0.6069620252,0.0467410535,0.1223386526,0.1749621034,-0.0454662256,-0.3047426343,0.1147944331,0.1704497039,-0.3243663311,0.2530548275,0.2005155385,0.1197219044,-0.2608419061,-0.1869264245,-0.1394843906,0.1243320853,0.2161901593,-0.3505869508,0.030197816,-0.2612145245,0.1456113309,-0.3643140197,-0.1169227809,-0.0494125262,0.1684373319,-0.0416090861,-0.0869178921,0.1603957862,-0.0047819167,-0.2482867539,0.0057886313,0.2127772719,-0.1623176187,-0.0298233572,0.1968926191,-0.1227709427,0.0286222864,-0.1135850474,-0.0954577699,-0.3736051023,-0.796487093,0.3856806755,0.1086682752,-0.2260926962,0.1569166183,0.1852319539,-0.0288626626,0.1390071064,-0.7766131163,0.2781107426,-0.2700860202,0.1859996915,0.0507349037,-0.0482592918,0.2186818868,-0.245657891,0.0444292389,-0.1558119804,-0.3363452852,-0.2957929373,0.1520014703,0.5615339875,-0.2364708334,-0.034599416,0.3850360215,0.5469372272,0.4004915059,-0.0670359656,-0.067981489,0.4184623659,0.5475918055,-0.3115632832,-0.0791823193,0.2339959443,-0.1497088969,-0.0996912047,0.3620908558,0.3107394576,-0.1633521914,0.286963433,0.0401583463,0.1326717883,0.0942769349,0.2241770476,-0.309875071,0.1005248353,0.1926442385,-0.1880115122,-0.0518749095,-0.3989800215,-0.1218361184,-0.0243604276,0.061000593,0.2646265626,-0.1970543861,-0.0600719638,-0.0649954602,0.1443314552,0.0027432337,0.0161043033,-0.0121987425,0.2155073732,0.2796203494,0.2686991096,0.4034879506,-0.26559937,-0.0943304375,-0.1338841915,-0.144049868,-0.0571437068,-0.0475823022,0.0855806172,0.2596667409,0.1097124368,0.1156005859,0.3220529258,-0.0970590934,-0.0447798856,0.4560160935,-0.3914614618,-0.294403553,0.0297587458,-0.191253826,0.2687639892,-0.0891850665,-0.0929313451,-0.2434486896,-0.0358288102,-0.0312523767,-0.1641988605,0.1342917234,0.127809003,-0.1605033726,0.1407569796,0.2580203414,0.143434599,0.0538886972,0.2126133144,0.1216577291,0.2759924829,-0.0779657513,0.1932588518,0.072306715,-0.061202459,0.1887561977,0.0845030472,-0.2157591581,-0.1994145215,-0.260589689,-0.0752066821,-0.1938834935,0.2318017781,0.0760559812,0.1843226254,-0.2688039839,-0.5139300823,0.4054386318,0.2316489816,-0.059795592,0.1844306886,-0.3162142932,-0.0129252253,-0.080029346,0.0073938412,0.6194285154,-0.2254932225,0.3763304949,-0.1509215981,-0.0443961285,0.3291864991,-0.497579366,-0.0549178943,-0.0405446254,-0.2693860233,-0.0064733955,-0.0443162695,-0.1164269522,0.0063693342,-0.0639231205,0.3358236849,-0.1385199428,0.0161025319,0.1028255522,0.0630025938,0.0982587039,-0.1343484074,0.0600645132,0.2710026801,-0.128954649,-0.1042524204,-0.0486171208,0.1003938094,0.0181983449,0.2117584795,-0.348410517,-0.1806467324,0.0193875749,0.3263842165,0.0004045458,-0.1931096315,-0.0247473046,0.1287295669,0.2211126238,0.2069506645,-0.157676518,0.24063766,-0.0560303219,0.1770827025,0.2221260667,-0.1428223252,0.1603872478,-0.0400775373,0.059972629,-0.0360330343,0.0636530295,-0.1874228865,-0.0689612255,0.2709845006,0.1704344153,-0.0534021929,-0.2605217993,0.0466165841,0.1419523954,0.066594474,0.1456883699,0.3092527986,-0.0304542445,0.5358054638,-0.243746981,-0.4441325366,-0.0491460487,0.089582704,0.4706982672,-0.2969496846,0.246140793,-0.0117872823,-0.1994111389,-0.005233733,0.3188085854,0.067256704,0.0313018747,0.1030184478,-0.2232806981,0.2156289071,-0.163562268,-0.4461269677,0.1223970279,-0.3968331814,-0.2381710112,-0.116958715,-0.4305744171,-0.0959674716,-0.2070608437,0.2224828601,0.0288554206,-0.4427157044,0.1704188585,0.4168639183,-0.2408590913,0.103058517,0.1197370961,0.215049684,0.0983555838,-0.0639102533,0.06558422,-0.1661859006,-0.1151798069,-0.0573547184,-0.1430146247,-0.2526406944,0.0401961319,0.1194192693,0.1940991431,-0.0846005753,-0.1608939916,-0.3413431346,0.0277582686,-0.2151890844,0.0458555743,0.3281026483,0.250222683,0.0525254942,-0.0190482605,-0.2545650005,0.0606618375,0.2935043871,0.0707679987,0.3296484053,0.0834345073,0.1396348923,0.1571826637,-0.2573201358,-0.421965301,0.1872330606,-0.1020592228,-0.0509157702,0.2260445803,-0.2251012027,0.2495860755,0.4410793781,-0.1233844161,0.1734838039,-0.080980055,-0.1859038323,0.0557178557,0.1488506347,0.26887694,-0.1151915044,0.0335802548,0.417607069,-0.2711031735,0.0664892644,0.1544736773,-0.0757163987,0.0487861075,0.1051445529,0.4411541224,-0.2353550345,0.1328932345,0.1203708574,0.2203483433,-0.0414518565,0.3915826082,0.4607016742,0.0668893903,0.5477506518,0.1909267306,0.2733959854,0.4535444081,-0.1988291442,-0.0916387215,-0.3442139626,-0.1206762716,0.3691903949,-0.1119835973,0.1148327067,-0.0498378277,0.2011880875,0.1710630655,-0.4117728472,-0.1741052717,-0.0463554896,-0.3184819221,-0.0562825501,0.4661230743,0.0097872335,0.1980104893,0.3031118512,-0.0922474563,-0.3561854661,0.4774303138,-0.0334118232,-0.2023450285,-0.2770073712,0.2927493453,0.1348914057,-0.17489703,-0.1015881822,0.2509457767,-0.2589648664,-0.064692758,-0.0242526233,0.0933383927,0.1487419754,0.1655650288,0.05718638,0.3473058343,0.1391281337,-0.070704706,-0.1802367121,0.2328952998,-0.1752924323,-0.1302569062,-0.0693158135,0.1011548117,0.0940848216,0.3056695163,0.0186947249,0.2601122558,-0.2928519249,-0.1882201284,0.0067084981,-0.0192584861,-0.2305589318,0.258348763,-0.3910734653,-0.1737109274,0.6758269668,-0.2580587864,0.1359237134,-0.1140784845,0.0960622579,-0.1910621226,0.4656352997,0.4458967149,-0.0579667613,-0.299136132,0.0074321884,-0.2848970294,0.1120932549,-0.2606487274,0.4734933674,-0.1006662175,0.5078715682,-0.144312188,-0.0784377232,-0.1094810665,-0.0335023962,-0.0562613457,-0.078524068,-0.2845355868,-0.0505347326,0.1140254885,0.1673985869,0.05656882,-0.1680807173,0.2948609889,-0.2146653533,-0.0311391391,-0.0888767242,0.0323958546,0.0776208788,0.3440172672,0.0808253288,0.0452024154,0.0137083866,0.1630230099,0.1854195446,-0.3548761308,0.0036779279,0.1407096982,0.063222982,0.0539239459,-0.0235884879,-0.5196059942,0.1973180771,0.143505156,0.0603236258,-0.1435730904,-0.0266829021,0.0854971334,-0.3765104115,-0.2722699046,0.347807914,0.4459002018,0.1441458613,-0.1681191325,-0.0506997891,-0.2086984217,-0.0880703628,0.2826548517,0.0581457056,-0.3323767483,-0.0007810725,0.1203754246,0.506211102,-0.1627057791,-0.335636884,0.1049752235,0.4289159775,-0.171066612,-0.2342907637,0.2778825164,-0.2623939216,0.1653283536,-0.1241283268,0.6173259616,0.0001726229,-0.0431100987,0.2192090005,-0.1861627847]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1973","title":"Question: what gets stored in the datasets cache and why is it so huge?","comments":"Thanks @ioana-blue for pointing out this problem (and thanks also @justin-yan). You are right that current implementation of the datasets caching files take too much memory. We are definitely changing this and optimizing the defaults, so that the file sizes are considerably reduced. I will come back to you as soon as this is fixed.","body":"I'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!","comment_length":55,"text":"Question: what gets stored in the datasets cache and why is it so huge?\nI'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!\nThanks @ioana-blue for pointing out this problem (and thanks also @justin-yan). You are right that current implementation of the datasets caching files take too much memory. We are definitely changing this and optimizing the defaults, so that the file sizes are considerably reduced. I will come back to you as soon as this is fixed.","embeddings":[-0.0833511651,0.0121250013,-0.1291754693,0.5186682343,0.1493166834,0.2426836044,-0.0757901371,0.3019647598,-0.0877328739,-0.0359353572,-0.0295592695,-0.2534342706,-0.0936839208,-0.2008035481,0.0723080933,0.1365639567,0.0901818648,-0.0653728023,-0.0765932277,-0.0797738656,-0.1062290147,-0.0433256477,-0.0135121122,-0.1718011051,-0.5766088963,-0.1823077798,-0.125397265,-0.0051542921,-0.2007423937,-0.1783480495,0.1808399558,-0.0291627496,0.3237573504,0.3823104501,-0.0001080761,-0.3183884621,0.3569486737,0.0258181766,-0.35998106,0.3442702293,-0.5000380874,-0.297462225,-0.1312585324,-0.1360443085,0.3404588699,-0.0457876362,-0.0374122113,-0.3942119479,-0.0077165314,0.2449931949,0.2759654224,-0.2257376611,-0.305583477,0.0927578658,0.2078104913,0.0236138646,-0.1338836402,-0.2668149173,0.2703861892,0.1654694229,-0.1730728447,0.1816094667,0.0901528075,-0.0411953926,0.2849489152,-0.0867872685,-0.2421435118,-0.203481257,0.6005621552,0.2197853029,0.9331120849,-0.3052316308,-0.100264661,-0.1678259373,-0.1438965201,-0.151291132,0.2134750932,0.4073275328,0.0708834007,-0.0055508926,-0.5161411166,-0.4197382331,-0.1441660076,-0.1212616116,-0.1735284775,-0.251204282,-0.2740025818,0.0589072704,0.161728397,-0.0618541129,0.4042038023,-0.2412802577,-0.1609716415,0.1731912047,-0.3817909956,-0.2755789757,-0.2229240835,0.4864886403,0.2402953506,0.0038443923,-0.2600577772,-0.109291926,-0.1039826646,-0.0202954579,0.2225336134,0.3948453367,-0.1059476063,0.1032147035,-0.0339701734,-0.3564936519,-0.2345282584,-0.1782754809,0.1071284637,0.0232658144,0.4347737134,-0.4517464638,-0.1963359118,-0.3539862931,0.2049482316,0.1451401711,-0.0209944863,-0.0245751273,0.1459482014,0.1945821941,-0.1634193063,0.2311792821,-0.2916731834,-0.2418981194,0.0145778842,-0.21426332,-0.2492374033,-0.062359143,-0.1481209248,0.3357031047,0.3038076162,-0.1790136844,0.280624181,0.0677447021,-0.0248508584,-0.1428686231,0.2385396063,-0.3573527038,0.4010714591,0.328265816,-0.1311090738,0.2729564607,-0.2001665831,0.1455440074,-0.190713495,0.1966866404,-0.5129279494,-0.4545753598,-0.0051716887,0.1176294163,-0.10547591,0.037894208,-0.3192513287,0.0472883806,0.4130821824,0.1691479236,0.1751920283,-0.1207905561,-0.4018245041,-0.232427299,-0.1263591647,0.1870438606,-0.389220655,0.2134615779,0.009946526,0.2114602923,0.05776041,0.3836231828,-0.1483823806,0.0776749998,-0.0870076343,-0.1209373623,0.0001559153,-0.079579778,-0.636710763,0.1218225881,0.381228596,-0.1184332147,-0.0189932287,0.4304911792,0.0283242892,-0.0962783918,-0.1710998863,0.2358875573,-0.2705536783,-0.1395442039,-0.2917768061,-0.1208130792,-0.0083782794,-0.0380461402,0.0865780488,-0.1643362194,-0.0146144964,0.330711633,-0.0058595021,0.1374899894,0.0867403299,0.3432820439,-0.0301416107,0.2613257468,0.3053285778,0.0460471138,-0.4341362715,0.27020818,0.1865427196,-0.6862898469,0.1906158924,-0.1346430629,0.1436361521,0.0420822687,-0.0126282424,-0.1946732551,0.0207596421,0.0416758806,0.0679130852,-0.2097644061,-0.0188928451,0.573993206,-0.0988553315,-0.1863345951,-0.3965398967,-0.216348961,0.0431044549,0.0391309336,-0.073134616,-0.0137702739,0.0558140986,-0.0892109498,-0.1688681096,0.1361524612,0.2321187705,-0.1308439374,0.217418775,0.5695483088,-0.0355291814,0.0977270901,0.371270448,-0.1007415727,0.0756179765,-0.1984051317,-0.2419497371,-0.0179809146,-0.1506020278,0.1063494533,0.0278175846,-0.1154013798,0.0016547996,0.0606407076,0.2964203358,0.0728689805,0.1380551755,-0.105885841,0.3245917559,0.4382680655,-0.0780995041,0.176500082,0.4605905414,-0.1509393007,-0.0410198718,0.1564587951,-0.2423934191,-0.4789295197,0.1506143659,0.331137985,0.3907627761,0.10922084,0.2850667536,-0.1033120528,0.2868413031,-0.1608241349,0.0812226087,0.0667128861,0.0219489355,-0.2884733677,0.0932127684,-0.0410421491,-0.0126916254,0.2164717466,-0.055618044,0.2983462811,-0.0109923044,0.3491840065,-0.2550541162,-0.1621956378,-0.0070126061,0.3860312402,-0.2329478413,-0.1407172233,-0.1002913192,-0.0584415197,0.2660431862,-0.1486243904,-0.2300077677,0.5778571963,0.0820519626,0.1423575133,0.088034831,-0.0799468681,-0.3060518205,0.111074701,0.24933514,-0.3045134842,0.1999036223,0.1163125038,0.0619489402,-0.2768292129,-0.1332397014,-0.1086862162,0.1835201979,0.198715046,-0.3027626276,0.0109501537,-0.3178112209,0.138032198,-0.3626973927,-0.1848442107,-0.1165110394,0.1484915912,-0.0446062349,-0.0062331846,0.2158530802,-0.0736985654,-0.162884444,-0.0196522772,0.1723953187,-0.1615498811,0.0438599549,0.3077871799,-0.1309025288,0.0314119086,-0.2782609761,-0.0471064523,-0.3229497373,-0.8187430501,0.2975212932,0.1009911448,-0.2260670811,0.1321332455,0.2460592389,0.0601855591,0.1901030242,-0.7528636456,0.1599376947,-0.2401581258,0.1551108658,-0.0018200498,-0.0307548046,0.2011798173,-0.1798158139,0.0073360302,-0.1245791614,-0.1967882961,-0.2765602469,0.1456563324,0.4915038943,-0.192502737,-0.0633852035,0.3370993137,0.5671008825,0.5190479755,0.0446704403,-0.0705253929,0.3782269955,0.4487402737,-0.3152042031,-0.0880659968,0.2686746716,-0.1435835212,-0.0421168804,0.4030549526,0.3427503109,-0.1661659479,0.2124953717,-0.0696430281,0.1701494455,0.0360478275,0.1999653131,-0.2818181217,0.1630782336,0.1915143877,-0.2135233283,-0.0805293024,-0.4090752006,-0.0097256862,-0.0490241237,0.0604406148,0.3139725029,-0.2041915059,-0.0216282755,-0.0354798734,0.0954543948,0.0373733342,0.0340191089,-0.0738481283,0.2846222818,0.2186451107,0.3076267838,0.3758858144,-0.1845826209,-0.128083095,-0.1948767155,-0.1542897224,0.0497720949,0.0045884242,0.0770994499,0.3125368357,0.1059291437,0.1206215248,0.3195889294,-0.1694954038,-0.0632763132,0.4400778711,-0.3505926132,-0.3430583477,0.0311940238,-0.1552645415,0.2621398866,-0.116670832,-0.0711956099,-0.2251215577,0.0282577183,-0.0399833918,-0.1176182553,0.0885976478,0.1425408125,-0.1457243711,0.111519821,0.2330707014,0.1177610978,0.1649422646,0.1629828662,0.1325430274,0.2992492616,-0.1056108624,0.17949,0.0435791835,-0.0517020673,0.1908410341,0.0886818394,-0.210628733,-0.1995036602,-0.21762757,-0.1043856665,-0.177401945,0.2469909489,0.0499367341,0.1593248099,-0.2842359543,-0.5892307162,0.467823863,0.2875882387,-0.0503034517,0.281650275,-0.3000889719,-0.0262130313,-0.1170723736,0.0719791353,0.6399997473,-0.2170850039,0.354742527,-0.1607198864,-0.0057511469,0.3252637982,-0.494831115,-0.0203311853,-0.031721212,-0.294303149,0.0027601528,-0.0793971494,-0.1042396352,0.0160079785,-0.0375991017,0.3806457222,-0.0770798773,0.0171461962,0.1023027077,0.2111571729,0.0683915168,-0.1219477728,0.1412032247,0.2821708024,-0.1376039684,-0.1486556381,-0.0250102468,0.1032503843,0.0334875509,0.2228507251,-0.2927027345,-0.1754133403,-0.0282164514,0.3199635744,-0.0239460468,-0.1103063598,0.0224183183,0.1508461535,0.1966027617,0.2202226669,-0.1755133867,0.2917010486,-0.0590799451,0.1441821009,0.2262175977,-0.1373253167,0.1802841276,-0.0922553912,-0.015595926,-0.0061956411,0.0756166801,-0.2075465471,-0.1815805137,0.2590361834,0.1753003299,-0.0318703428,-0.1809736192,0.0111451475,0.0968554616,0.0720027313,0.1782183796,0.3341040015,-0.0147101181,0.5450965166,-0.1917355657,-0.4085083604,-0.0390249752,0.037645679,0.495264709,-0.2649909258,0.296146065,0.0108361645,-0.2105105817,-0.0300778747,0.2401515692,0.106525436,-0.0192218032,0.0464051999,-0.1987548918,0.2384643853,-0.1002332568,-0.3699209392,0.0672204122,-0.3693825305,-0.1530644,-0.1152306199,-0.4739052057,-0.0556584485,-0.2485062778,0.2175403833,-0.0033826553,-0.4405683577,0.1212480515,0.4964890778,-0.2847179472,0.0173554905,0.0766951889,0.2561146319,0.0128195519,-0.107309401,0.0661789849,-0.1396746784,-0.0889873132,-0.0818248987,-0.1462655067,-0.2698170543,0.0640408844,0.106032446,0.1788944453,-0.0970167592,-0.1527721286,-0.3858827353,-0.0040498623,-0.2405215502,0.0508169457,0.2874113321,0.2743214667,0.0970945284,0.0281570088,-0.3459436893,0.1409607083,0.3457016647,0.0663540959,0.2746598423,0.0574159138,0.1722718477,0.15250808,-0.2675606608,-0.3174443543,0.1916094571,-0.1096866354,-0.110775277,0.1850320995,-0.2056340873,0.2195078582,0.4261027277,-0.1534730941,0.1396077871,-0.0626812577,-0.1825831532,0.1420999318,0.1769847572,0.2107807994,-0.1552992314,0.0270279218,0.3825099766,-0.2631728053,0.1017864197,0.1827222854,-0.0995817631,0.1150415838,0.0916781798,0.34887743,-0.3227354884,0.0893634781,-0.0018475763,0.2343327999,-0.1470898688,0.4342311621,0.5094912648,0.0556072854,0.5676188469,0.2129418701,0.3438628614,0.426843226,-0.1354210526,-0.0707474947,-0.2690297663,-0.2201738358,0.4318252206,-0.1616875082,0.0600060709,-0.0915101469,0.2003552467,0.2525607646,-0.3557163477,-0.1063615009,-0.0910748541,-0.284098357,-0.0544140525,0.4328331649,0.0195154808,0.16266267,0.2804258466,-0.0502699986,-0.3196659982,0.4784559906,0.0176240318,-0.2315435857,-0.2045263946,0.2657102644,0.2216133624,-0.0962501094,-0.1503149569,0.2840476036,-0.2878260911,-0.0521467216,-0.0020573363,0.1353498548,0.1079745144,0.1173431277,0.0900890976,0.2902495861,0.1260780096,-0.117747128,-0.1577920765,0.2456450462,-0.1479880512,-0.163123861,-0.0584236979,0.1051508039,0.0860580057,0.1809345782,-0.0109625906,0.204143554,-0.2903199792,-0.1878154725,0.0826536193,-0.0291631781,-0.1981458515,0.306201309,-0.3989690244,-0.1746680588,0.6400590539,-0.2770073712,0.0793247223,-0.1532568783,0.1183214784,-0.2559439838,0.5393371582,0.4638421834,0.0643789917,-0.2885505259,-0.0669268146,-0.2184087485,0.0996603891,-0.3270794451,0.4932592511,-0.02972541,0.4698668122,-0.2054979056,-0.1049922407,-0.0908684358,-0.1021397039,-0.0105253328,-0.1008916572,-0.3023269773,-0.0487516969,0.1350871772,0.18933855,0.0312538929,-0.1479459554,0.3177106977,-0.2440875769,0.0139771234,-0.1394674033,-0.006686802,0.1513634771,0.3554560244,0.0827659816,0.0488254614,-0.0478210263,0.1564616412,0.1493866891,-0.3969499767,-0.0511363074,0.0851008818,-0.0666459948,0.028011281,-0.0466875881,-0.5293648243,0.1965588331,0.0999139771,0.1286892593,-0.2723075449,-0.0669395328,0.0615590028,-0.3856152594,-0.197949335,0.3814426363,0.3347819149,0.135027051,-0.1180505753,-0.1002390012,-0.2879343629,-0.0645386875,0.2868933976,0.0080703106,-0.3462714553,0.0657007545,0.1039329097,0.4476662278,-0.1230793148,-0.3959047496,0.1256313622,0.4089435339,-0.0978845954,-0.1693587899,0.2278121263,-0.134761259,0.2264450938,-0.0989769772,0.5771958232,-0.0549586117,-0.0072089545,0.2037740797,-0.1564337015]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1973","title":"Question: what gets stored in the datasets cache and why is it so huge?","comments":"Thank you! Also I noticed that the files don't seem to be cleaned after the jobs finish. Last night I had only 3 jobs running, but the cache was still at 180GB. ","body":"I'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!","comment_length":32,"text":"Question: what gets stored in the datasets cache and why is it so huge?\nI'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!\nThank you! Also I noticed that the files don't seem to be cleaned after the jobs finish. Last night I had only 3 jobs running, but the cache was still at 180GB. ","embeddings":[-0.1613561809,0.0733159259,-0.1333291382,0.5806114078,0.0784174725,0.2742608786,-0.0560066178,0.2457107455,-0.1244978607,-0.0830651596,-0.0408940278,-0.2797769606,-0.1030301005,-0.1785742939,0.0967974588,0.2181337625,0.0879473761,-0.1207300723,-0.1814738065,-0.1622634083,-0.1023330167,-0.0179168638,0.0224896539,-0.1180053279,-0.6523393393,-0.2212353647,-0.1131067127,0.0551729165,-0.140467599,-0.0933794007,0.1439766437,-0.0460858941,0.3950193226,0.5104987025,-0.0001139643,-0.2925594747,0.3823718429,0.0545073599,-0.4045554698,0.3270026147,-0.4570840001,-0.3279615343,-0.1085926965,-0.1003948078,0.4394214749,-0.0204240996,0.0014370388,-0.3428097069,-0.01866249,0.2640453875,0.2358754277,-0.3048946857,-0.3635195196,0.1135002077,0.1681370586,0.12315844,-0.0441249125,-0.2613183856,0.245504573,0.1298613101,-0.1326352358,0.1569849402,0.1074741408,-0.0296071619,0.1850611269,-0.0502866879,-0.2164560407,-0.2998597026,0.7091450691,0.2274082601,0.8417983651,-0.2384046167,-0.0828320086,-0.27496171,-0.1479184777,-0.1580325067,0.2207850069,0.5174489021,0.1128786579,-0.0317042917,-0.5117552876,-0.3802755475,-0.06558083,-0.184029147,-0.1793046743,-0.2046904117,-0.3094809651,0.0825137123,0.1481610388,-0.0059112608,0.39160496,-0.2668062747,-0.1414568424,0.231658712,-0.3380459547,-0.2481552809,-0.172574833,0.5365595818,0.1817601472,-0.0329650678,-0.3444201648,-0.1124420539,-0.2114587128,-0.0000929759,0.2603337467,0.3116800487,-0.0861698538,0.1187398508,0.0085733477,-0.3854296803,-0.3219989538,-0.1497398019,0.0372533388,0.0221334323,0.399423331,-0.4164708555,-0.1734295487,-0.3614159822,0.1985375881,0.1956635714,-0.0042334674,-0.0722748637,0.1121164933,0.2183167189,-0.1871268451,0.2036716342,-0.2611676753,-0.2313588113,0.028411787,-0.1532808542,-0.2285506576,-0.0909060091,-0.1327641308,0.2995837033,0.2765157521,-0.1396545917,0.3309641778,0.0368876532,-0.064817749,-0.1563345194,0.2549434304,-0.358281523,0.3950999379,0.4093769789,-0.1124285534,0.3327395916,-0.2217198014,0.2115445882,-0.1827074289,0.2137183398,-0.5258376002,-0.4418341517,-0.0159431212,0.0983831584,-0.1572528034,0.0768250972,-0.345371604,-0.0189489424,0.4571041465,0.2275614291,0.2256826758,-0.1646018773,-0.3279085457,-0.1894806772,-0.2278371304,0.1464494467,-0.3910383582,0.2511361837,0.0179405548,0.1454429179,0.0711840242,0.2979474962,-0.1191538796,0.131539166,-0.0978694335,-0.1348291337,0.027983401,-0.04060819,-0.5370240211,0.1421535909,0.400799185,-0.1622817814,-0.0745797083,0.3459576368,0.0290507562,-0.067992866,-0.1514481306,0.2860537171,-0.2108335793,-0.1279174238,-0.2654138505,-0.0885379314,-0.0383824669,-0.0184705984,0.0998502299,-0.2232891172,-0.0368052199,0.1881962568,0.02879286,0.1677855104,0.0976628661,0.3305998445,-0.1397838742,0.4280255437,0.287922889,0.030646015,-0.4317771196,0.2343403548,0.1349747479,-0.6966284513,0.1623967737,-0.1269115955,0.2554288805,0.0729258284,-0.0731006488,-0.2337891012,-0.0279716104,0.1845546961,0.0301332008,-0.1857680082,-0.0711145028,0.4955204725,-0.1140736639,-0.1607788205,-0.4297209084,-0.1848450452,0.1431550533,-0.0023234328,-0.1459568292,0.0411186367,0.0786861107,-0.1117612869,-0.1637632698,0.1110548079,0.24776645,-0.1484709531,0.2682817578,0.5673520565,-0.0548567176,0.1951506138,0.3002670109,-0.0716318861,0.0636250302,-0.1734065413,-0.2402172089,-0.0508687086,-0.2077073306,0.1525839418,0.0015329618,-0.1117341816,0.0542668775,0.0190218985,0.2636958063,0.1293394417,0.1632869542,-0.0756035298,0.2168033719,0.4447779953,-0.1519056112,0.1747808456,0.4706738889,-0.1771167666,-0.0516844951,0.111319162,-0.2850257158,-0.4193945229,0.1050440893,0.4016013145,0.4452232718,0.0496107563,0.3604390621,-0.0756823644,0.277546525,-0.1837282479,0.0208872706,0.0748989582,-0.0487935692,-0.152292937,0.0795725435,-0.0078957994,-0.0303099416,0.1681210697,-0.0738716945,0.29049173,-0.0363349505,0.3572681546,-0.2580591738,-0.1316239387,-0.07329119,0.3482121825,-0.2150791883,-0.1365124881,-0.1094918251,-0.0709861517,0.2380701452,-0.1624608338,-0.2139178216,0.5364332199,-0.0142115271,0.1500573903,0.0735522732,-0.0893294141,-0.3087903261,0.0690322071,0.2572316527,-0.2098612934,0.2047890276,0.0922112092,0.0664835945,-0.2757597566,-0.1237822473,-0.1766274571,0.1951682419,0.2061213851,-0.288238287,0.000666933,-0.2626593113,0.0905258656,-0.3399087787,-0.1441929042,-0.0432294831,0.0609958395,-0.0696478263,0.0072284043,0.2078796029,-0.0054568253,-0.2065435499,-0.0260301251,0.1936479658,-0.1558229923,-0.0068773031,0.2063533664,-0.2218588293,0.016188968,-0.2453486323,-0.0151543347,-0.3820661902,-0.9283376932,0.259250015,0.036852289,-0.1697472483,0.2085366249,0.1965106726,0.023202477,0.1615057439,-0.7610763311,0.2020997107,-0.2372547388,0.1262796372,0.0077726627,0.0311026089,0.2027198821,-0.1744493991,0.0423541367,-0.1398317963,-0.2366739511,-0.3054128885,0.1870910227,0.5237814188,-0.2106374204,-0.0840959698,0.1811249852,0.536593616,0.4866471291,-0.0381651372,-0.1205075234,0.4225435853,0.4848379791,-0.2912487984,-0.0783840641,0.2911514342,-0.0866435394,-0.0904414132,0.4184948504,0.3539145887,-0.0811954737,0.2511945069,-0.0332706459,0.0404647663,0.0854173079,0.2601003647,-0.3480753005,0.1371676028,0.1285502613,-0.1664553583,-0.0566880926,-0.3869127333,-0.0410664752,-0.0087417001,0.0842422545,0.2667947114,-0.1963014156,-0.0937538072,0.0104858223,0.1013094485,-0.0088180639,-0.0358110107,-0.0618790388,0.2916465402,0.2779179215,0.3221161067,0.3650979698,-0.3677512407,-0.0626016334,-0.1375657916,-0.1767582446,0.0598027483,-0.0522746667,0.0341922343,0.2674119174,0.1784341186,0.1854330152,0.3122754991,-0.1331876516,0.0565019846,0.4340999424,-0.3935497403,-0.3147127926,0.0180363581,-0.193453297,0.2668859363,-0.096753262,-0.1134518087,-0.2543800473,-0.013445871,-0.0458248556,-0.1563433707,0.1244399101,0.1552250683,-0.1609626412,0.0811807364,0.3141678572,0.1119781807,0.1702612638,0.2174154967,0.1782796681,0.321836412,-0.1787976027,0.2074742317,0.0910573006,-0.0685985833,0.2785595655,0.0779270753,-0.1736046821,-0.2642306387,-0.236840874,-0.0679100528,-0.1778141856,0.2747413218,0.078583777,0.256066829,-0.2771654725,-0.5995448232,0.3857510388,0.2862666249,-0.109285295,0.2671396732,-0.489554435,-0.0106045473,-0.1321151853,-0.0013051083,0.646619916,-0.2087157369,0.3260278106,-0.1999161988,-0.0667585582,0.3643881083,-0.5225836039,-0.0769258142,-0.0006269134,-0.2032234967,-0.0374314263,-0.0765035376,-0.1397905946,0.0380024053,-0.0022657816,0.3384401798,-0.0816521719,0.0173256472,0.0731512383,0.1452865005,0.0633621216,-0.0883146599,0.1170081198,0.2121531665,-0.1198509634,-0.0831394866,-0.0899268091,0.1223376691,0.1065189987,0.2631841004,-0.2684571147,-0.186167106,-0.0166535676,0.2970582545,0.0237222854,-0.1567716151,-0.0804010257,0.1931591183,0.1869921982,0.1478960514,-0.1096016243,0.2880507708,-0.0981836468,0.1249374151,0.1941354871,-0.1072758213,0.2267092913,-0.0649359077,-0.0248343069,0.0245965384,0.0448238179,-0.173237294,-0.219497785,0.3592066169,0.1631358564,-0.0518307611,-0.2033403665,0.0533610173,0.0576289967,0.1359996051,0.1384758651,0.3363602757,-0.0736912265,0.6109339595,-0.2178416401,-0.4782263637,-0.0767513663,-0.0232030917,0.5032827854,-0.2337809652,0.2860405445,0.0184660424,-0.175132066,-0.0311497319,0.2559075952,0.1470770538,0.0428171419,0.0856799856,-0.1649220139,0.2176072598,-0.0642015785,-0.4341562688,0.0132348733,-0.3945584893,-0.232401967,-0.1734164655,-0.4358879924,-0.0303704739,-0.2216004431,0.2527504265,0.0294938236,-0.3938131332,0.2168478966,0.4286861718,-0.2361278832,0.1436594874,0.0652046353,0.3111394644,0.0642904416,-0.0576629788,0.0879255012,-0.1032580286,-0.1550849229,-0.0653671324,-0.1248485297,-0.2228997201,0.052591987,0.1372592747,0.1883078963,-0.0488751791,-0.1431195736,-0.3217907548,0.0851672739,-0.2121693939,0.0646228045,0.1869442612,0.209081158,0.1564133465,0.1074780673,-0.2736562788,0.0798655152,0.3291398883,0.0449234173,0.2957021892,0.0472166575,0.1226035506,0.137989983,-0.2660237849,-0.3409770727,0.0906479433,-0.0949325189,-0.0699300617,0.209503904,-0.2170521915,0.2317305803,0.432890147,-0.126177758,0.2592012584,-0.0209151711,-0.179333359,0.1794490516,0.1302594244,0.2233305871,-0.0603436939,0.0664522126,0.4383385777,-0.3042235374,0.0979854017,0.1023675427,-0.1207405478,0.0606674105,0.1509355605,0.3464744687,-0.3671177328,0.1173054129,0.0725601092,0.1537768841,-0.0700296462,0.4517393112,0.479398787,0.0183852855,0.5429626703,0.3200546205,0.4149710834,0.4656387269,-0.1855344772,-0.0896632969,-0.3549958467,-0.2353122085,0.5174940825,-0.1213669479,0.0425862558,-0.1281033009,0.1910884678,0.2133021802,-0.3449545503,-0.1182982922,-0.0223555136,-0.2834682167,-0.0617319271,0.3831990957,0.0348755755,0.2570710778,0.3279588521,-0.0391800106,-0.2646865547,0.5557457805,-0.0171339139,-0.2771163881,-0.2427119315,0.2536901534,0.1225280091,-0.1367987543,-0.1073721275,0.252615273,-0.3126508892,-0.0337105766,-0.0128571112,0.1473042965,0.1936819702,0.1900820434,-0.0141222971,0.3107691705,0.0938917324,-0.0519730635,-0.2107859999,0.3367969692,-0.1680718362,-0.1389774382,-0.0691708028,0.0962318182,0.0966415405,0.2490070611,-0.05092692,0.2233565748,-0.2308834642,-0.1931045353,0.1338278204,-0.0147658968,-0.2211968154,0.2575021386,-0.397985965,-0.1736105382,0.6114830971,-0.2598451376,0.0722599328,-0.1495042741,0.07011085,-0.2653844655,0.5645395517,0.5225402117,0.0885150731,-0.2780853808,-0.0863828436,-0.2743099928,0.0643812045,-0.3352906406,0.527192235,-0.1185422167,0.4859295487,-0.201898694,-0.1149547175,-0.0428570881,-0.122041896,-0.000512375,-0.1198476702,-0.3422717452,-0.0923207104,0.1259981394,0.1910202652,0.0899187028,-0.1604303271,0.2859534025,-0.2384636849,-0.0576173104,-0.0538929775,-0.0453089513,0.1033481583,0.3779560328,0.1462683976,0.0711490512,-0.0091992142,0.1375867873,0.2291775346,-0.3952301741,-0.0751999021,0.1247661561,0.0192075893,0.0290724616,-0.0575017333,-0.5672248006,0.2911173999,0.1137893125,0.0580924451,-0.2475099564,-0.0398728177,0.092695117,-0.4392723143,-0.1869778484,0.3382079899,0.3680446446,0.1801202893,-0.116571635,-0.1416971684,-0.2414920926,-0.0349174552,0.3328413367,0.041523248,-0.4040132165,-0.0493646562,0.0916977078,0.4154227674,-0.0620097779,-0.41349262,0.1519260705,0.4260599017,-0.0775112957,-0.1994899958,0.1650744081,-0.0829914957,0.2037351876,-0.0660081059,0.4899362624,-0.0626689941,-0.039693702,0.2204195261,-0.1378958672]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1973","title":"Question: what gets stored in the datasets cache and why is it so huge?","comments":"And to clarify, it's not memory, it's disk space. Thank you!","body":"I'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!","comment_length":11,"text":"Question: what gets stored in the datasets cache and why is it so huge?\nI'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!\nAnd to clarify, it's not memory, it's disk space. Thank you!","embeddings":[-0.0562924184,-0.0744777843,-0.1533181369,0.5605322123,0.1392017156,0.2712965906,-0.103793487,0.1924572736,-0.1215412021,-0.0535814241,-0.0198384952,-0.2450783998,-0.0932705179,-0.2444388568,0.0898417607,0.1368533671,0.1016000509,-0.0692107528,-0.1194493398,-0.1004541293,-0.1232964098,-0.0296263415,-0.0306411609,-0.1901216507,-0.5624324083,-0.177314356,-0.1243965775,0.0014009768,-0.1873768121,-0.1622399986,0.1398853511,-0.0987059176,0.3147396147,0.4090926945,-0.000110236,-0.334823966,0.3442212939,0.0813871771,-0.4406152964,0.3626205921,-0.4893383682,-0.3059493303,-0.0806634724,-0.1268419623,0.3792389631,0.0108966744,-0.0467454977,-0.4488114715,-0.1145015433,0.2249283493,0.2559868693,-0.2213608772,-0.2966934443,0.0954151228,0.1739647985,0.0541763864,-0.1401763558,-0.2864204645,0.2703980803,0.2199219167,-0.1200313941,0.1260786802,0.135957256,0.0196812674,0.2455102205,-0.0532069169,-0.274977982,-0.2100818157,0.6089799404,0.2325011492,0.9566081762,-0.3052790761,-0.0922987536,-0.1764596403,-0.1393209696,-0.1700592488,0.2371486872,0.4252000153,0.0750685111,-0.0326383188,-0.5084655881,-0.4093165696,-0.1100541428,-0.1268862337,-0.2654014826,-0.1718884706,-0.2728392184,0.0589097403,0.1790124476,-0.0518645793,0.3500959873,-0.1983054578,-0.1393890977,0.1995717883,-0.3551800549,-0.3141660988,-0.1887093633,0.4850865006,0.2018358707,-0.0138913933,-0.3047730625,-0.0940738767,-0.1251036078,-0.0070109782,0.2149579525,0.3676956594,-0.1059814021,0.1228733212,-0.0277378224,-0.4088228345,-0.2197234482,-0.1497768313,0.1345712841,0.1122433841,0.4681493044,-0.4973906577,-0.1613356024,-0.3145111203,0.1813944578,0.1700838059,-0.050271336,-0.0242434237,0.1238459945,0.1785236001,-0.1724553555,0.190039143,-0.2924372554,-0.2345407307,0.0431479849,-0.1576696932,-0.247199893,-0.1134015471,-0.1410695761,0.3161053061,0.3001384437,-0.1031678095,0.2539270222,0.104041189,-0.0586248673,-0.1305873245,0.2422409505,-0.3871128261,0.3800346255,0.3951134384,-0.1071810499,0.3413525522,-0.2308599502,0.182765767,-0.1703885198,0.2146661729,-0.4468957484,-0.4317738712,-0.0399952196,0.1027977243,-0.1249627098,0.01111749,-0.3311277032,0.0430473834,0.402783066,0.1845217645,0.2279994339,-0.1034339815,-0.394778043,-0.2560149133,-0.1391716897,0.1418177038,-0.3938845694,0.236370489,-0.0590823442,0.146881327,0.0093646534,0.3528730571,-0.1007439271,0.0371600576,-0.0525216572,-0.099825941,0.0087994291,-0.0403371714,-0.6160578728,0.0466768518,0.358481437,-0.1241397634,-0.11355073,0.4548946321,0.0170637257,0.0218658224,-0.1444569379,0.2895534933,-0.2269501388,-0.1168098375,-0.2866629064,-0.0697142631,0.0429695733,-0.038076248,0.1152125224,-0.1604979038,-0.0182470251,0.330100596,-0.0267677139,0.1765110791,0.1211749837,0.3527576029,-0.0300392117,0.3443938196,0.2838682532,-0.0062710308,-0.3910816014,0.2419979572,0.2294461876,-0.6656194925,0.1775836945,-0.1041037887,0.2164342105,0.1437695771,0.0230682194,-0.1430692822,0.0029056482,0.1156113446,0.0587029867,-0.1908686608,-0.04028669,0.571495235,-0.0174571183,-0.1097308993,-0.399605453,-0.194902122,0.0717656389,0.0020464973,-0.1177682132,0.0268071219,0.0646983609,-0.0531838574,-0.1354516298,0.132530123,0.2350403816,-0.1841284633,0.2183110118,0.5753266215,-0.0304631516,0.0793447569,0.3582178652,-0.1098541543,0.0829194412,-0.1429531872,-0.1972124428,-0.0453892387,-0.106517382,0.1040898561,0.0280806292,-0.0881634504,0.0114748012,0.0476095453,0.2900853753,0.0385831967,0.1725430936,-0.1153220683,0.3299143016,0.4539194405,-0.0662727132,0.1906682402,0.4051602781,-0.1563767642,0.034543056,0.0956183523,-0.2954479456,-0.5034760237,0.1262611747,0.2956638038,0.4432424307,0.0853862911,0.3393195271,-0.0892629623,0.2325178683,-0.1668754369,0.0442263372,0.0904783383,0.014300094,-0.2837027609,0.1120660901,-0.0194407143,-0.0117567312,0.2003641725,-0.0636239424,0.2859850526,-0.018365832,0.3545537591,-0.2499097735,-0.1208016649,0.0566588305,0.3294337988,-0.2431869656,-0.157449469,-0.1103096455,-0.103245683,0.2969869077,-0.1982882023,-0.2016782761,0.5683117509,0.0249820985,0.1459359676,0.1145493612,-0.0991263986,-0.2929771841,0.1220862418,0.1178102046,-0.2766754031,0.2385308594,0.162810266,0.0693855882,-0.2903749347,-0.1793014556,-0.1414546072,0.176029101,0.1988422424,-0.3187978268,0.0199604835,-0.2818790674,0.1486767381,-0.3509925604,-0.1390207261,-0.0820397735,0.1100302115,-0.0518451817,-0.018585179,0.2080886662,0.0231268257,-0.1956242472,-0.0005323004,0.2007437646,-0.170079723,0.0033859403,0.2077008039,-0.1590008885,0.0475089103,-0.2538766861,0.0022091139,-0.3741787076,-0.7601431012,0.2588633597,0.0699526966,-0.1604857743,0.1322582662,0.1972810328,0.0021229831,0.2399017513,-0.7863323092,0.1801281273,-0.2798417509,0.1795738786,-0.0002496035,-0.0266196113,0.2113597542,-0.2099695802,0.0019351683,-0.1480554491,-0.3139939308,-0.3401144743,0.1210073531,0.5165979266,-0.2106372565,-0.0497715995,0.2944307327,0.5086993575,0.4466643631,-0.0239654612,-0.0558316596,0.3948989213,0.4893722832,-0.3056718111,-0.0371883586,0.2432448417,-0.0672781393,-0.04586659,0.4008315504,0.3374234736,-0.1947290152,0.2188663781,-0.0247279163,0.1151547953,0.0922851413,0.211836949,-0.325827837,0.1520371437,0.1879922748,-0.1940552443,-0.0422346853,-0.4001238048,-0.0770933107,-0.0610437132,0.0010262965,0.274900943,-0.2822743952,-0.0889966786,0.0061951536,0.1054063663,0.0214612279,0.0676504597,-0.0676508769,0.2550243139,0.2702705562,0.304248035,0.3080634773,-0.2590051293,-0.0929588228,-0.2189053297,-0.2154364735,-0.0062233773,-0.0012562822,0.0505282357,0.2656677961,0.0779580399,0.0848391429,0.3116336465,-0.1354449242,-0.0031840652,0.3992989659,-0.3605487943,-0.3031994998,0.056258861,-0.169968605,0.323255688,-0.0835350007,-0.1185046807,-0.221521914,0.0527069792,-0.0353988409,-0.1982783675,0.101223506,0.1568153799,-0.1422313601,0.1170857772,0.2820791006,0.1019912809,0.1625956148,0.1719684303,0.1060465127,0.3091925383,-0.1119183898,0.2310091704,0.0773481503,0.018853223,0.1983376145,0.0860225335,-0.1403381526,-0.1737746745,-0.2807689011,-0.0849681348,-0.2064615339,0.2287168652,0.0541659929,0.1782126874,-0.3069344759,-0.5476050377,0.4587532282,0.3453297317,-0.1233141273,0.2892435789,-0.2909910083,-0.0476990491,-0.0719961599,0.0648019612,0.6531684399,-0.2257570624,0.3624620736,-0.1904183775,-0.0598420538,0.3249565959,-0.4715160728,-0.0577474572,-0.0475540236,-0.2755205929,-0.0048727612,-0.1040814072,-0.1133414358,-0.0370227173,-0.0260994397,0.280131042,-0.1219659969,0.0112949638,0.1369399875,0.1317212135,0.0766375586,-0.1126057431,0.0708226711,0.2763620913,-0.1342699826,-0.1192002296,-0.0868408382,0.1491636485,0.0299112126,0.2420289814,-0.3720339835,-0.2111899108,-0.0776275396,0.3044442534,-0.0176206455,-0.1054507494,-0.061394114,0.1349855661,0.2498844713,0.2228628993,-0.1473266929,0.2935294509,-0.0713370442,0.1180461049,0.2224439085,-0.1318589598,0.1930983365,-0.0718774423,-0.0098495753,0.0332748182,0.0694981813,-0.2168830931,-0.2165783048,0.2744106352,0.1476647407,-0.0077554998,-0.2036010772,0.0259051025,0.1392732263,0.1007394269,0.149123311,0.3413062394,-0.0586859025,0.5790477991,-0.1499800831,-0.4814817309,-0.059294574,0.0519048907,0.5293886065,-0.2362924218,0.239887625,0.0007293657,-0.1987935603,-0.0119412206,0.2358602881,0.1386660635,0.0503677316,0.0933824405,-0.1928174347,0.1818088591,-0.1206569597,-0.4481330216,0.0531995483,-0.37968117,-0.2101469636,-0.1529890895,-0.3905370831,-0.0201876648,-0.2552936673,0.2326132804,0.0299709998,-0.4740037024,0.1575727612,0.4650921524,-0.2588333189,0.0638188645,0.109214738,0.2758654952,0.046127826,-0.0868899971,0.0652781203,-0.1073557734,-0.0920866281,-0.0710886642,-0.1549720466,-0.2556705475,0.0713971704,0.1185621694,0.1875983179,-0.1030332744,-0.1938452721,-0.4041804671,0.0555707626,-0.1995748281,0.0387933478,0.3198177218,0.2814010382,0.1030797958,0.0871749893,-0.2852217853,0.1003486067,0.3705048561,0.0569959395,0.2657252848,0.0724090636,0.153349176,0.1264775544,-0.2659114003,-0.3557162285,0.1876092702,-0.1058919579,-0.172592625,0.16714257,-0.2132871151,0.2297845632,0.4358646572,-0.1427881569,0.166552037,-0.0482552424,-0.1878939867,0.0903669,0.17766276,0.282677412,-0.1383442879,0.0273642391,0.4319906235,-0.2453383803,0.1398776472,0.1514158696,-0.0406209379,0.0456165411,0.0960231945,0.3662637472,-0.3026959002,0.0946149826,-0.0084588528,0.1593742371,-0.1231359169,0.4338537157,0.4962776303,0.0600374825,0.545253098,0.2318535447,0.3972381353,0.4147687554,-0.2355530858,-0.0213723574,-0.229851827,-0.1961258203,0.4717699289,-0.1499741524,0.0256896019,-0.0789207816,0.1489372253,0.2410157472,-0.3657794297,-0.1196648926,-0.1075454056,-0.3508948088,-0.0618456192,0.4207300246,0.0200238358,0.1423901767,0.292144835,-0.0755536109,-0.3922308385,0.5520911813,-0.0204915721,-0.2191552222,-0.284040451,0.3094538748,0.2223961353,-0.224606663,-0.1140559688,0.2862431705,-0.3074787259,-0.0787994117,0.0258970056,0.1629722714,0.1248444095,0.1003853008,0.0007439702,0.3202114105,0.106704928,-0.1317388266,-0.1687612832,0.2570386231,-0.1226112917,-0.128148213,-0.0980599374,0.1124222875,0.0812974721,0.2410618216,-0.0226329658,0.2444644272,-0.2529678941,-0.1383438408,0.0570221655,-0.0144251501,-0.147761479,0.2825644016,-0.3985839486,-0.1727992296,0.6393331289,-0.3065425754,0.0360279903,-0.1399249732,0.0984030887,-0.206952408,0.5006203651,0.5337818265,0.0351054072,-0.2743776739,-0.07307069,-0.2464553267,0.1305404156,-0.2995530367,0.4579053521,-0.067268528,0.5332180262,-0.1929124892,-0.0627561137,-0.0526509546,-0.0872983113,-0.0331412777,-0.1127753705,-0.3234880269,-0.0405371934,0.1424405873,0.1619337648,0.0455936417,-0.1503854543,0.2997761071,-0.2829472423,-0.027907772,-0.1940570921,0.016735686,0.1536925584,0.3539260924,0.1022271216,0.027015062,0.0109173134,0.1718161553,0.1846287996,-0.3589402139,-0.0363112874,0.1359957755,-0.0250521954,0.0651614442,-0.0367964096,-0.5339885354,0.2101837546,0.195152998,0.0803620368,-0.2293145359,-0.0128879501,0.094461821,-0.4102738798,-0.1920652837,0.3306035995,0.3782918155,0.1532652825,-0.0947784483,-0.0753749087,-0.25687024,-0.1049028188,0.248617053,0.0219582636,-0.3535798788,-0.0171068907,0.0988304392,0.4542749524,-0.1751122773,-0.3810437322,0.1199979931,0.4483494759,-0.1446056515,-0.2124413103,0.2593078613,-0.1190321743,0.2057488114,-0.1096001789,0.6172021031,-0.0777164474,-0.0245396122,0.2117211968,-0.1280506998]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1973","title":"Question: what gets stored in the datasets cache and why is it so huge?","comments":"Hi ! As Albert said they can sometimes take more space that expected but we'll fix that soon.\r\n\r\nAlso, to give more details about caching: computations on a dataset are cached by default so that you don't have to recompute them the next time you run them.\r\n\r\nSo by default the cache files stay on your disk when you job is finished (so that if you re-execute it, it will be reloaded from the cache).\r\nFeel free to clear your cache after your job has finished, or disable caching using\r\n```python\r\nimport datasets\r\n\r\ndatasets.set_caching_enabled(False)\r\n```","body":"I'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!","comment_length":95,"text":"Question: what gets stored in the datasets cache and why is it so huge?\nI'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!\nHi ! As Albert said they can sometimes take more space that expected but we'll fix that soon.\r\n\r\nAlso, to give more details about caching: computations on a dataset are cached by default so that you don't have to recompute them the next time you run them.\r\n\r\nSo by default the cache files stay on your disk when you job is finished (so that if you re-execute it, it will be reloaded from the cache).\r\nFeel free to clear your cache after your job has finished, or disable caching using\r\n```python\r\nimport datasets\r\n\r\ndatasets.set_caching_enabled(False)\r\n```","embeddings":[-0.0611622036,-0.0334960222,-0.1297230124,0.5127940774,0.1173948124,0.2473039627,-0.0104232728,0.2427644581,-0.0756754801,-0.0384477898,-0.0576454513,-0.2856124043,-0.062841095,-0.1385336965,0.1360729337,0.0736555532,0.0571287647,-0.0924019665,-0.0630483553,-0.160524711,-0.1804127246,0.0023389754,-0.043512404,-0.1090048254,-0.6069411039,-0.1671299934,-0.1760504842,0.013318684,-0.1266205758,-0.2256072462,0.1812643111,-0.0810930282,0.3298203349,0.5128854513,-0.0001129724,-0.3193747401,0.3587917089,0.0624000914,-0.4379311204,0.2950620949,-0.3354638219,-0.4271190464,-0.0203636494,-0.1686972231,0.3214415014,-0.0767427906,-0.132932812,-0.4167048633,-0.0587260686,0.2930725813,0.2494700253,-0.1366453171,-0.3039160669,0.1178890467,0.1235289797,-0.0146650802,-0.0452343635,-0.2453737408,0.2158931941,0.1646102369,0.0025354703,0.147770077,0.0169417318,0.000306403,0.2684416175,-0.0396770462,-0.2254711986,-0.311825633,0.6033416986,0.2232895494,0.8363260031,-0.2880611718,-0.202964142,-0.2979716957,-0.1118043363,-0.2330391854,0.1009243131,0.4950154722,0.0988687202,-0.0455195792,-0.4867727458,-0.3448884785,-0.0777859464,-0.0465623252,-0.2821796536,-0.1423802972,-0.2524193227,0.1125405356,0.1055423319,-0.0284945481,0.3575477898,-0.3363456726,-0.0157869067,0.2345681787,-0.3509426117,-0.2011492401,-0.1322093159,0.5577651262,0.1247597337,0.0052215718,-0.3312217295,-0.1804830283,-0.1822399199,0.0167691,0.1956836283,0.383474052,-0.175694719,0.1844043881,0.0359155759,-0.3828605711,-0.2349632233,-0.1346769482,0.1312394291,-0.0012338117,0.4762334526,-0.4796627164,-0.098683998,-0.3029206395,0.170734942,0.2444289476,-0.0966221392,0.0273224618,0.1169028357,0.2148654312,-0.2545852959,0.2108535916,-0.2304310501,-0.1243429109,-0.0487615392,-0.1454728693,-0.2492411733,-0.0925647914,-0.067214489,0.2685973346,0.285683006,-0.2234884351,0.3373494446,0.0277760718,0.1063642949,-0.1411570907,0.2592744827,-0.3753346801,0.3710655272,0.4655985832,-0.0162447132,0.3749172986,-0.1396909803,0.1204946488,-0.270875901,0.2553671598,-0.5015741587,-0.4222280383,0.0191155411,0.104588069,-0.2029598951,-0.0228899047,-0.3181248009,-0.0490736701,0.4451746643,0.2298892289,0.2339739054,-0.2258072942,-0.3481408358,-0.2585402131,-0.1718460917,0.1132037342,-0.3837774694,0.2007804215,0.0188006144,0.2824506462,0.0434069298,0.4158514738,-0.1315063983,0.1567609906,-0.053813342,-0.1209238023,0.0304627083,-0.1566626281,-0.6271005273,0.1202648431,0.4168612063,-0.0274888612,-0.1223238111,0.4630418122,0.0704054013,-0.0158333648,-0.1096921042,0.3284002542,-0.1558927745,-0.0901832432,-0.292834729,-0.0386738218,0.1061438322,-0.0216424521,0.1188717186,-0.2305828631,-0.1309048682,0.1836871356,0.0527519099,0.1259609312,0.0694234669,0.3437756598,-0.0101463469,0.3748395443,0.3232589364,-0.0078653451,-0.3653471172,0.2573634684,0.2419636548,-0.6245020032,0.1455274075,-0.1339732111,0.1272249967,0.145066157,-0.0969081521,-0.2118736058,-0.0223334916,0.0297688544,0.1151991263,-0.2108961046,-0.0449691117,0.5452654958,-0.0635576248,-0.1103802994,-0.418202579,-0.170804143,0.0938118473,0.0492023975,-0.1359282136,0.0650078729,0.1010837629,-0.0895123556,-0.174693495,0.1081861854,0.2973416746,-0.1439442188,0.1077988073,0.5784066916,-0.0145122204,0.1164007708,0.423370868,-0.0106687611,0.0905218795,-0.1879730672,-0.1697003543,0.0567103922,-0.1616250128,0.1676454693,-0.0029339595,-0.0588439256,0.0638449341,0.1050605476,0.1870781779,0.0860394165,0.1927024722,-0.0662449822,0.3792734444,0.4196597338,-0.0819109827,0.1654995084,0.4912531972,-0.1889422238,0.0204286166,0.1143535301,-0.3488580883,-0.4260418713,0.0824612379,0.3058714271,0.4276180863,0.0918692946,0.2786064148,-0.0187046304,0.1948915273,-0.196464777,0.1110937744,0.1457482576,-0.0842692032,-0.2639130056,0.1610333174,-0.0088864975,0.0190624893,0.1868859828,-0.1532470137,0.3150743842,-0.0810955986,0.4258531332,-0.2198471129,-0.0666332394,-0.0421657376,0.1854019612,-0.2399906963,-0.2696067095,-0.164760679,-0.091042459,0.2942963541,-0.1827508509,-0.2300923616,0.4699304104,0.0185604449,0.0107689556,0.1471359432,-0.0653238669,-0.3312746882,0.0642062724,0.1377596259,-0.2550904453,0.2090715319,0.0785148069,0.0315914005,-0.2303849459,-0.1585240811,-0.1949876696,0.107504569,0.1837016493,-0.2883615196,0.0521753766,-0.2645152211,0.1547861248,-0.2806056142,-0.1142954752,-0.0577976145,0.1149192452,-0.0509900562,0.0269431025,0.2000137419,-0.0313189924,-0.2764130831,0.0005234366,0.2373341471,-0.0265397169,0.0129480707,0.2392877042,-0.1557642072,-0.0176599529,-0.1436425298,-0.0100040752,-0.3633743823,-0.7905566096,0.3151326776,0.0385604016,-0.1788747907,0.2086083293,0.1866608411,0.0924993902,0.1387996078,-0.708563447,0.0724817663,-0.3142879307,0.2180024832,0.0143601261,0.0765080154,0.1385978162,-0.1808548272,0.0806726962,-0.1142867804,-0.2508400679,-0.4035722017,0.202921316,0.3824837506,-0.2886651754,-0.0345722921,0.2057446092,0.4963887036,0.5578774214,-0.0190250184,-0.1098964736,0.3746926785,0.4645800591,-0.3987114131,-0.1185678169,0.2061817646,-0.1018320397,-0.1222377792,0.3536936641,0.3558354974,-0.0725569427,0.281808883,-0.0484957024,0.0352503061,0.0698835999,0.2234961241,-0.2844494879,0.1933543682,0.1191814765,-0.1655669659,-0.1425035745,-0.397328794,-0.0735083148,-0.0214167628,0.0456141233,0.282697767,-0.1842951328,-0.1279440373,-0.0754010305,0.1423415095,-0.0231203716,-0.0497249365,0.0514616221,0.2763727009,0.3083380759,0.3073391914,0.3131423593,-0.2744621336,-0.2138916105,-0.1058372185,-0.2411259711,-0.1186879352,-0.0318222158,-0.0057502282,0.3012592793,0.1512315869,0.1316509247,0.3211079538,-0.1440427154,0.0087752659,0.5344349742,-0.3888075054,-0.2943491638,-0.0840831399,-0.276268065,0.216451481,-0.1252090186,-0.1201285347,-0.1507781893,0.023403367,-0.0632726848,-0.240449816,0.1136998311,0.1901481897,-0.1203188375,0.156110093,0.2085818648,0.0853591189,0.1842174679,0.1479958594,0.1042301878,0.3757861853,-0.1913027465,0.1126988381,-0.0072586481,0.0352626368,0.2778845429,0.0563132688,-0.1737894714,-0.183246389,-0.2612355947,0.0143588418,-0.269115895,0.223762989,0.1496378034,0.2776620984,-0.2618519664,-0.6982131004,0.4616201222,0.276463151,-0.0716560557,0.2424008995,-0.418320626,-0.0365161747,-0.0543632433,-0.0064194291,0.6589263678,-0.2236090302,0.4306711555,-0.0395675972,0.0344164334,0.4065584242,-0.4473713934,-0.0286580455,-0.0794861987,-0.1845750362,0.0040681055,-0.1019874141,-0.177142188,-0.0575257204,-0.045195438,0.440168798,-0.1805891693,-0.0142159751,0.0162401982,0.186169982,0.0436868258,-0.1095271856,-0.0421760939,0.2492516488,-0.1284711957,-0.0425213799,-0.0616244003,0.149669975,0.0199741777,0.1891242117,-0.3140135109,-0.2012457848,-0.028866943,0.324588418,0.0098946188,-0.1806678772,-0.0580397546,0.191376403,0.2203962952,0.2901022434,-0.0961439461,0.267059654,-0.1876454651,0.1555694193,0.2435379773,-0.091469489,0.1821191907,-0.0270394664,-0.0338976756,0.0384034254,0.0826619565,-0.2128392458,-0.1591347456,0.3341325819,0.0977141708,-0.064110145,-0.2387067974,0.0309948679,0.1164923608,0.0835460946,0.1395623237,0.3734086752,-0.0597868487,0.6573856473,-0.2188907117,-0.517998755,-0.0908969641,-0.0430397391,0.5688452125,-0.2462107241,0.3715312481,0.0228577685,-0.2190306932,-0.0264745541,0.2350376695,0.0635287464,0.1049836874,0.1324856132,-0.1333607137,0.2987756729,-0.1285158545,-0.3920572996,0.0370853059,-0.3879739344,-0.2860276997,-0.1671657413,-0.4105102122,-0.0266919415,-0.0878059417,0.2242877781,0.056885805,-0.297328949,0.1887333542,0.4064845443,-0.2433436066,0.0819339976,0.0297147501,0.2258062065,0.0675609037,-0.0967761874,0.0996919274,-0.1272792965,-0.1318953782,-0.0946975052,-0.1185578182,-0.2398275584,-0.0179099459,0.1461031139,0.1713632047,-0.0941336304,-0.1098677665,-0.3558125198,0.1286489218,-0.25622949,0.0847612321,0.2919304669,0.2346148044,0.127768144,0.0235261768,-0.2593023479,0.1167764962,0.3148727417,0.0370092578,0.2348183841,0.1609542221,0.1893199533,0.2360896766,-0.1818602234,-0.2923294604,0.0981742144,-0.1392735988,-0.1110362038,0.192197606,-0.2242861837,0.2773998082,0.3868361413,-0.1051514074,0.2471598536,-0.0514642745,-0.1714699119,0.1436871439,0.1363578588,0.1666206568,-0.1629320234,0.0704922229,0.4474796653,-0.244958818,0.1808709949,0.1114569455,-0.0085003115,-0.0219189115,0.0780228004,0.3410858512,-0.3142295778,0.1521823406,0.075065285,0.1814650744,-0.1232979521,0.4441087544,0.4823721647,0.0159765333,0.5705832839,0.177680552,0.3709160089,0.3553227186,-0.2679767907,-0.0616261661,-0.3010139763,-0.2353511751,0.3864545524,-0.1725271344,0.0777339861,-0.0796904489,0.1531075537,0.1485292912,-0.3190407157,-0.2061379254,-0.0086433627,-0.284135133,-0.1284306049,0.4397257268,0.0606790036,0.2145421505,0.261476934,-0.0105376719,-0.4243508875,0.5057739615,-0.0288449023,-0.2109094113,-0.263243705,0.2805339396,0.1405210346,-0.1363978684,-0.1008689404,0.1854052842,-0.2370547354,-0.034901768,-0.1282428205,0.0883433372,0.2755081654,0.1468670517,0.0165942889,0.3816764951,0.0619968139,-0.0660636127,-0.1962122023,0.1796407849,-0.2197112441,-0.0932975858,-0.0303407665,0.1074646413,0.0647096857,0.2108067572,-0.06530229,0.2227579653,-0.3473086655,-0.0111099146,0.0606893785,-0.0644293725,-0.2514471114,0.3058980703,-0.4158521593,-0.1779727936,0.6756238341,-0.1745564044,0.0435057767,-0.2043659091,0.0827862471,-0.2363872677,0.5724013448,0.5509409308,0.1847085506,-0.2911015153,0.0006310517,-0.2649799883,0.0914617702,-0.4748069644,0.3963684738,-0.0434781574,0.4691889882,-0.2120286971,-0.0628110543,-0.0016786405,-0.0164435022,-0.0396844372,-0.1721843183,-0.3049779534,-0.0741594136,0.0393011011,0.1438498348,0.0957855359,-0.1642239243,0.2386148274,-0.3013191819,-0.0327400304,-0.1607341617,0.0522633307,0.1634505689,0.337836355,0.1687491238,0.0903688744,0.0849587545,0.2190423161,0.1733672023,-0.3463304639,-0.0603383183,0.0893499181,0.0750254318,-0.0569976866,-0.0385944284,-0.5262833238,0.1357942969,0.1050975993,0.1555252671,-0.1740638316,-0.1851957142,0.1313710213,-0.3232077658,-0.1827083379,0.3909306824,0.3479951024,0.1791638732,-0.1575751603,-0.0387184545,-0.1911489964,-0.0748231262,0.3076620102,-0.0438524224,-0.3876527846,-0.0505253524,0.1258730739,0.4357409477,-0.1410327107,-0.4110242724,0.1245218962,0.4666081369,-0.0897403434,-0.2220543772,0.2676522434,-0.1196449548,0.1683700979,-0.0717624277,0.519482851,-0.1256662905,-0.0529369265,0.2193593532,-0.1838398874]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1973","title":"Question: what gets stored in the datasets cache and why is it so huge?","comments":"Thanks for the tip, this is useful. ","body":"I'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!","comment_length":7,"text":"Question: what gets stored in the datasets cache and why is it so huge?\nI'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!\nThanks for the tip, this is useful. ","embeddings":[-0.0910820663,-0.04966598,-0.1459761411,0.533685565,0.1184812412,0.2585758269,-0.0462159067,0.235682413,-0.1354729235,-0.0779450536,-0.0469890125,-0.2340994924,-0.0860741585,-0.2003123611,0.1008797511,0.1412308961,0.0743916482,-0.1022917554,-0.0934476107,-0.1323721707,-0.1064716727,-0.0319585539,-0.014309545,-0.1588148028,-0.5555639863,-0.1575461924,-0.1048966423,-0.0151819987,-0.1581282616,-0.1658613831,0.1318885833,-0.0807019174,0.3239879608,0.4146981239,-0.0001100813,-0.335608542,0.3806997836,0.0498000234,-0.4226690233,0.3468374908,-0.511100471,-0.3109680712,-0.0702029392,-0.1205111817,0.385736227,-0.0091442605,-0.0435124598,-0.42731601,-0.0852716416,0.228198722,0.2662566602,-0.2085999995,-0.3013235927,0.0881310701,0.1605710238,0.0212307516,-0.1242138892,-0.2946600318,0.2871528864,0.2586841583,-0.1296629608,0.1348274797,0.1101032346,-0.0153515842,0.1964701563,-0.0566938594,-0.2455633134,-0.2344828695,0.6056619883,0.2451971173,0.9396653771,-0.2796750963,-0.0966032743,-0.1806423515,-0.1391808242,-0.1529354006,0.2254547775,0.4324713349,0.1049342752,-0.0143741928,-0.5043811202,-0.4141313136,-0.1130247787,-0.14016366,-0.2624779046,-0.1594463289,-0.2912850678,0.0568915196,0.1403226703,-0.0235659778,0.3676126599,-0.2304257452,-0.1443159729,0.1711045355,-0.3604377508,-0.3154287338,-0.18505916,0.4993733466,0.1758071184,0.0308775585,-0.3200520575,-0.1065963954,-0.1417331398,-0.007766644,0.2203616798,0.3254199922,-0.0804763287,0.1240057722,-0.0426207744,-0.3983849883,-0.2481525689,-0.1271366179,0.1283040792,0.0943918377,0.4448985159,-0.4489933252,-0.135785833,-0.3506948054,0.216606617,0.1541106701,-0.060498666,-0.0593818128,0.1641681641,0.1634443998,-0.1868607551,0.1819961667,-0.3235552311,-0.2428055257,0.0547444075,-0.1994986683,-0.2587067783,-0.1200448647,-0.1061788425,0.2842767537,0.2830832899,-0.1848261207,0.2813486755,0.0819389895,-0.0678442642,-0.1351094544,0.2534404695,-0.3445239067,0.3772216737,0.3833989203,-0.1007692814,0.3494978547,-0.1968421191,0.2001308352,-0.1887577027,0.2068155706,-0.4943928421,-0.4737573862,-0.0482565239,0.1123155057,-0.1450081766,-0.0195910856,-0.2921686172,0.0504633896,0.4257988334,0.1920555234,0.2395728528,-0.1090387926,-0.4032928646,-0.2532099485,-0.1626202315,0.165907681,-0.4017329812,0.2280752361,-0.0404929556,0.1383301914,0.0262749214,0.3303931355,-0.1390546113,0.0548870116,-0.0438082889,-0.1275388598,0.0379965566,-0.0281546135,-0.6311725378,0.0989643857,0.3401176929,-0.1451987028,-0.1067684367,0.4308367074,0.0076482422,-0.0601587109,-0.1733954549,0.3155531585,-0.2395171523,-0.1494519711,-0.2897186875,-0.0652492642,0.0513679609,-0.031945087,0.1063761488,-0.1619085819,-0.0223404616,0.293813169,-0.0530233346,0.179487139,0.1169540808,0.3216351867,-0.0402980298,0.3469660282,0.2915922105,0.0324185714,-0.3730482161,0.23733069,0.2102227062,-0.6842323542,0.1893998235,-0.1201594174,0.1936784685,0.1125034317,-0.0106104566,-0.1603600234,-0.0017307423,0.0956924632,0.0678614601,-0.1734548211,-0.0553622954,0.5476195812,-0.0471453182,-0.1292725056,-0.4216752648,-0.1914343983,0.0951161832,0.0200518891,-0.1175360158,0.0190828796,0.0706207156,-0.0600037239,-0.1484692246,0.1164088175,0.2167061716,-0.1793405861,0.2036768049,0.5867330432,-0.0592653826,0.105175525,0.3599257171,-0.0951660946,0.0794360712,-0.1449075192,-0.197143048,-0.0001061547,-0.1244772449,0.0941938907,0.0062657986,-0.0913332924,0.010302227,0.0534193777,0.2880220115,0.0585945211,0.1496827602,-0.1289294362,0.3330984116,0.4372338355,-0.0899266079,0.2440519333,0.4252323806,-0.1524659246,0.011455969,0.1158775538,-0.27724123,-0.4959875345,0.1091784388,0.2932763994,0.4485353231,0.0866670087,0.3302315772,-0.1061854139,0.2477095723,-0.1754382998,0.0699620098,0.0830938593,0.0037490611,-0.2721201181,0.1280982047,-0.0160132963,-0.0354691073,0.2073645443,-0.0769588128,0.2978479564,-0.0255325139,0.4027766883,-0.2888165116,-0.1361104399,0.0195092298,0.3201855421,-0.2421627939,-0.1887014359,-0.105890274,-0.1121453568,0.2898509502,-0.163115263,-0.2095083445,0.5641601682,0.046050977,0.1450116336,0.1072474346,-0.0948028043,-0.3029347658,0.1188081428,0.1693400592,-0.2972670794,0.2270589024,0.1454292685,0.0640947595,-0.3040459752,-0.2081785351,-0.1435049921,0.1677989215,0.2014836818,-0.286329627,-0.0019202861,-0.2906394899,0.1233848259,-0.3717865944,-0.1625103503,-0.0946991518,0.1093951091,-0.0659470186,0.0057900529,0.2148779184,-0.0025487291,-0.213694185,-0.0064487318,0.1955193579,-0.1660236269,0.0109814908,0.2229156047,-0.1905568391,0.0492698178,-0.2541787624,-0.0296442434,-0.3847595155,-0.7905492187,0.2807975113,0.0639935583,-0.1543558687,0.1331286132,0.1955239028,-0.0009991479,0.1785250902,-0.7959542871,0.1727334708,-0.2503736615,0.1903976947,0.0155275101,-0.0142639829,0.2251355499,-0.2180945128,0.0165938213,-0.1355332136,-0.2539791167,-0.3267677724,0.1108090952,0.5325921178,-0.2379928678,-0.0691559091,0.2687109709,0.5088127851,0.4665988684,-0.073419787,-0.0946414024,0.3972522318,0.4660296142,-0.3290318549,-0.0656503439,0.2490000576,-0.0539479814,-0.0341036916,0.4104531705,0.3295966983,-0.1937504411,0.2283280939,-0.028417794,0.1017715707,0.0871941522,0.2032328099,-0.3174817264,0.1455588937,0.1923584342,-0.1903363168,-0.0523695275,-0.4094746113,-0.0499609523,-0.0318136513,0.0248605032,0.2728543282,-0.247151196,-0.0485359207,0.0014838967,0.1380141526,0.0331369564,0.0588568561,-0.0715049878,0.2757134438,0.284350425,0.3022868335,0.3207072616,-0.330021888,-0.091828458,-0.1889237911,-0.2018698156,0.0240414888,-0.0278336275,0.0501675792,0.2897085547,0.140050292,0.0768817589,0.3207588792,-0.1241883039,0.0851448923,0.3703844249,-0.3446640074,-0.3388575315,0.0407486632,-0.1705856323,0.3129299581,-0.0881709456,-0.1251131445,-0.228940472,0.0148437433,-0.0132669332,-0.2069157362,0.119855985,0.1409682631,-0.1477362663,0.0795834363,0.3267940283,0.1304474622,0.1610313207,0.2016321421,0.0728502125,0.3022143543,-0.1035170183,0.1904076487,0.082879439,0.0087247565,0.2127504796,0.0856052935,-0.1619913876,-0.1529130787,-0.2846599817,-0.100801006,-0.1475293189,0.2429664731,0.0309699066,0.1892949343,-0.3273279965,-0.5602045655,0.461876601,0.2948145568,-0.108837083,0.2984643579,-0.3441236019,-0.0208233315,-0.0996036306,0.0241375137,0.6588848829,-0.1985240132,0.3259885609,-0.1549807787,-0.0310026258,0.350037843,-0.4661120474,-0.0975723192,-0.0590030663,-0.2707820833,-0.0117369462,-0.0877319202,-0.1108345985,-0.0386396311,-0.0476212241,0.2901198268,-0.1469714791,0.0219237618,0.1355294883,0.1516553313,0.0958696455,-0.0986580327,0.0500238463,0.2743349075,-0.1359768361,-0.1032635495,-0.069800429,0.1169296652,0.0648792386,0.23183842,-0.3391394317,-0.1721058637,-0.0407482833,0.3207064867,0.0444675311,-0.1334026307,-0.0289583709,0.105202809,0.2724507451,0.2252890319,-0.1459253579,0.2784033418,-0.0505276546,0.101111941,0.23152785,-0.16332528,0.2062135637,-0.0723882914,-0.0190225262,0.0257583428,0.0751136541,-0.2210074961,-0.2030168474,0.2917263508,0.1873632669,-0.0371936969,-0.1748412549,0.0816197991,0.1496905982,0.0911520272,0.1532594413,0.3589361608,-0.0665625408,0.5359370708,-0.1648442,-0.4788101912,-0.0649746656,0.0548428446,0.524900198,-0.275442183,0.2568592429,0.0221778844,-0.2009641081,-0.0180821661,0.2450576425,0.1426694393,-0.003763315,0.0699929893,-0.1954150647,0.2365883142,-0.1251362115,-0.4283644855,0.0205678958,-0.3713095486,-0.224219963,-0.1379992664,-0.398834765,-0.0542683937,-0.2856023014,0.2273069471,0.0265488364,-0.4286036491,0.1531362683,0.499563545,-0.2634973228,0.0811531991,0.1207542121,0.2745434046,0.0569537058,-0.0942149162,0.0668105409,-0.1022898331,-0.100806132,-0.0520418547,-0.1705560088,-0.2679602802,0.0499499701,0.1237023696,0.1758723855,-0.0937316939,-0.1905436963,-0.3502650559,0.0199971925,-0.1582224816,0.0325356349,0.3161878884,0.2877379358,0.1010708585,0.0929379985,-0.2972269654,0.0854387805,0.3532807529,0.0471327193,0.2935054302,0.0576824732,0.1476220936,0.1351096034,-0.2600308657,-0.3265872002,0.1865135282,-0.0972870737,-0.1415700316,0.2022814006,-0.2322734296,0.2480222434,0.4085486829,-0.1613511443,0.137750119,-0.0621987432,-0.2113530785,0.1210173368,0.1682346165,0.2477070242,-0.1063280255,0.0547773577,0.4899424016,-0.2557163537,0.1252621412,0.1570497453,-0.021572737,0.0826567858,0.1241510436,0.3273490667,-0.3012550771,0.115902245,0.0075740167,0.1673424244,-0.1127806306,0.4492766857,0.4861943722,0.0529379696,0.483641386,0.2748920321,0.3910236657,0.4213873744,-0.2114483863,-0.0290204752,-0.2636984587,-0.1690821648,0.4437488616,-0.1463425606,0.0340924412,-0.0575843304,0.1902150959,0.2418713272,-0.3460990489,-0.1202291474,-0.0576002374,-0.3462969363,-0.0686218888,0.3792825341,0.0341405384,0.1358830929,0.2822809219,-0.0705037192,-0.3624432087,0.5152921081,-0.0096724331,-0.1842914224,-0.3128350973,0.2834543586,0.2191067487,-0.149830997,-0.1173664853,0.304508388,-0.3286159635,-0.0580531731,0.0506835543,0.1506443769,0.1297903359,0.11233899,0.0204220768,0.3214921355,0.1179345399,-0.1188899353,-0.1585446,0.2726268768,-0.136942178,-0.1512343735,-0.0570449047,0.1080463529,0.0855169371,0.2758952081,-0.0266888198,0.2362381667,-0.2699012756,-0.1136698574,0.0881081372,-0.0085483836,-0.1838525683,0.3062767684,-0.4110687375,-0.1740508676,0.6598910093,-0.2722984254,0.0481562242,-0.1548888087,0.1019250005,-0.2297460884,0.5343679786,0.5249462128,0.0533633605,-0.2653984129,-0.0521210879,-0.2242523134,0.1481460333,-0.2804605365,0.4683699012,-0.0965060964,0.5254320502,-0.1774438322,-0.0124765299,-0.0463403314,-0.1034613624,-0.0274609905,-0.1260992289,-0.3267989755,-0.0520099178,0.1429599673,0.1676156074,0.0465233773,-0.1360290945,0.2935642004,-0.2736624479,-0.0111657437,-0.1905980408,-0.0437480137,0.1484264135,0.3059611022,0.0754083619,0.0575983487,0.0064279665,0.1611012667,0.2115343064,-0.3951300085,-0.0879589394,0.1122607663,0.0244435649,0.0718840808,-0.0286566205,-0.4976660907,0.1963293254,0.1620313674,0.0901678205,-0.2192296535,-0.0126367211,0.0693069696,-0.4221372008,-0.2116131186,0.342869103,0.3543790877,0.1606199145,-0.0919573158,-0.0486745462,-0.2471660674,-0.1070423126,0.299710691,-0.0084435567,-0.360969156,-0.0228030346,0.1055884361,0.4676610827,-0.153755784,-0.3999622762,0.1267792732,0.4711312652,-0.1352132112,-0.1754511893,0.270899415,-0.1231469139,0.2182753831,-0.1123404875,0.5687893033,-0.0791668668,-0.0100174975,0.2217682153,-0.1471127123]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1973","title":"Question: what gets stored in the datasets cache and why is it so huge?","comments":"Hi @ioana-blue, we have optimized Datasets' disk usage in the latest release v1.5.\r\n\r\nFeel free to update your Datasets version\r\n```shell\r\npip install -U datasets\r\n```\r\nand see if it better suits your needs.","body":"I'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!","comment_length":34,"text":"Question: what gets stored in the datasets cache and why is it so huge?\nI'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!\nHi @ioana-blue, we have optimized Datasets' disk usage in the latest release v1.5.\r\n\r\nFeel free to update your Datasets version\r\n```shell\r\npip install -U datasets\r\n```\r\nand see if it better suits your needs.","embeddings":[-0.1255369484,-0.0336780734,-0.148232758,0.5212608576,0.1614960432,0.2426841706,-0.0925324187,0.2584299445,-0.0802801028,-0.0226318613,-0.0793154463,-0.1993971169,-0.0890249535,-0.1937061399,0.0600060374,0.1441113949,0.1028742194,-0.0664708018,-0.0705818161,-0.0968135297,-0.1211709902,0.0194596611,0.0352946147,-0.1679355055,-0.5302938819,-0.1817645133,-0.1013382226,-0.0436650887,-0.1651499122,-0.17851533,0.177683115,-0.0290396884,0.3176939487,0.4055050015,-0.0001109538,-0.3099729419,0.3714547157,0.1150019318,-0.4696372747,0.3074622452,-0.3643327653,-0.3740325272,-0.026445087,-0.1428351551,0.3908945322,0.0068974965,-0.0497719459,-0.4210889935,-0.0684004277,0.1861254722,0.2560692728,-0.1310575604,-0.3215969205,0.1046227515,0.1198371351,0.01389091,-0.146869272,-0.2766178846,0.2572325766,0.3019111156,-0.106885761,0.2367847115,0.0618183538,0.047615312,0.2798642218,-0.0198587086,-0.1303207129,-0.1588064134,0.6128367186,0.2203997523,0.9452199936,-0.3251812756,-0.1090440676,-0.2270799428,-0.1699434817,-0.1856394112,0.2053114325,0.3774011433,0.0623725355,-0.0036444445,-0.4708918333,-0.4204160869,-0.1368875504,-0.1069780588,-0.1935634911,-0.2392660975,-0.2930272818,0.0738702193,0.2006108016,-0.0358287878,0.3408772647,-0.2122146636,-0.1762360334,0.1255204082,-0.3456109464,-0.2916094065,-0.1444860846,0.4545297325,0.166012302,0.0392155647,-0.3247304559,-0.1198164672,-0.1323130429,0.0058710747,0.2860298157,0.3430542946,-0.0669386536,0.1038033813,-0.0429320224,-0.3748679757,-0.1334179938,-0.1865023524,0.1570681483,0.0631488264,0.4441578686,-0.5209648609,-0.0817443058,-0.3450357318,0.16527614,0.1584222019,0.0572482571,-0.0358495004,0.1435464323,0.1911145449,-0.110761255,0.2182204127,-0.2996869683,-0.214574486,0.017782798,-0.1605705768,-0.2281496823,-0.131633997,-0.0949164629,0.2396894097,0.2744666934,-0.164727509,0.2371573448,0.063488923,-0.0907145888,-0.1422198415,0.234905526,-0.4032544196,0.3749283552,0.4167662859,-0.0743716359,0.3681034148,-0.1970903575,0.1527627856,-0.227425918,0.244900763,-0.4916736484,-0.4740643501,-0.1170701385,0.108580485,-0.161923483,-0.0288123973,-0.3449948728,0.0922386199,0.3634020388,0.153567642,0.1603097767,-0.0919574946,-0.4179095328,-0.228961587,-0.1571575254,0.1740286499,-0.429195255,0.1551629603,-0.1355956793,0.1866011471,0.0713289827,0.365894109,-0.1201933548,0.0764698386,-0.0518555455,-0.1532371938,0.040617872,-0.105731748,-0.6648839712,0.1029138267,0.3574771881,-0.1740622371,-0.0367924348,0.4561774433,-0.0024712256,-0.0370603167,-0.2303043306,0.290750742,-0.2054569125,-0.1454718411,-0.3278050721,-0.077662088,0.0378576815,-0.0154083706,0.0659245476,-0.2320688069,-0.037144199,0.2797748744,-0.0108113438,0.193445757,0.1148358583,0.3457117677,0.031427227,0.3540056944,0.3015316725,-0.0218943059,-0.3611468673,0.2509370744,0.2351715714,-0.7048804164,0.1734369397,-0.1321182251,0.1448073238,0.1432245374,-0.0207026955,-0.1719018817,-0.0045520132,0.0870919824,0.0700759515,-0.178489238,-0.0182370413,0.4595890641,-0.0254144967,-0.1279367208,-0.3539280295,-0.1449450701,0.0487825051,0.0091999453,-0.1537999213,0.0313860886,0.0623050369,-0.0768781304,-0.1596943587,0.1524797529,0.2265719473,-0.1928770393,0.2952666283,0.5493305326,-0.0663553551,0.1707333922,0.3852481246,-0.0702986121,0.1017337367,-0.1299525499,-0.2927914858,-0.0375097282,-0.1519739479,0.1403201371,-0.0249337684,-0.029297566,0.0416705012,0.0892701671,0.3436175287,0.0104550952,0.1606137902,-0.1470387429,0.2927565575,0.4069283009,-0.0240144711,0.2519698441,0.47829023,-0.1528939605,0.0289916713,0.1744274199,-0.2387561053,-0.5282307267,0.1187324747,0.3788320124,0.4037412703,0.1002401188,0.3153688908,-0.1507233381,0.29839921,-0.1419318765,0.0769500509,0.080037564,-0.0390971676,-0.2749990821,0.1334964931,-0.0680702552,-0.0156767294,0.1987346411,-0.0829682723,0.3320787549,-0.0365241729,0.3400944471,-0.2261348516,-0.1101819202,0.0874783173,0.3231965899,-0.2059070915,-0.2550525665,-0.1116032153,-0.0590515845,0.2781714797,-0.1644494236,-0.1198915988,0.4901140034,0.0762803853,0.101694271,0.0700767115,-0.0708680451,-0.316455096,0.1030517593,0.1625664532,-0.2699268758,0.2397953272,0.1029200628,0.0852553025,-0.3361773193,-0.151069209,-0.1490524411,0.1190942228,0.2485793829,-0.2534008324,0.0126693146,-0.3917896152,0.1810095906,-0.2936210036,-0.1614059508,-0.0395210125,0.162891984,-0.1019608229,-0.0407595113,0.1762523502,0.0103803249,-0.2491561472,-0.0023531609,0.2122996747,-0.1259872913,0.0549067594,0.2655703127,-0.1126466468,0.0555636249,-0.1928707212,-0.0497126058,-0.366995573,-0.7784826756,0.2513565123,0.1395721436,-0.1240602359,0.1695470661,0.2248826474,-0.0192549191,0.1598345339,-0.7960984707,0.0848229229,-0.2673851848,0.2629162669,0.0260320809,0.0380782112,0.2360247225,-0.1774861664,0.028453432,-0.1528524905,-0.2566244602,-0.3062720299,0.1367166787,0.4193833172,-0.231095165,-0.0271235034,0.3263312578,0.5199459791,0.4282681048,0.0144179054,-0.0571797565,0.3847506344,0.5536596775,-0.388767004,-0.0796085298,0.2006558776,-0.0727028698,-0.0731616989,0.3296693861,0.3406888843,-0.1232187748,0.2847546339,-0.0813004076,0.09379033,0.1011782214,0.198808521,-0.2678991854,0.1599436402,0.1861117184,-0.1608430445,-0.0656137988,-0.390984416,-0.0692704841,-0.0309385769,0.0783748031,0.2683601379,-0.2041399777,0.0029264241,-0.0678285882,0.1011187732,-0.0287856907,-0.0061646528,-0.0838761926,0.3064675033,0.2835105956,0.279494673,0.3214493394,-0.2633185983,-0.0617116801,-0.1646201015,-0.1650079638,-0.101812467,0.0064617535,-0.0108331544,0.2996634543,0.0837251395,0.0862341821,0.3078326881,-0.1852034628,0.0375439078,0.4310305417,-0.3830849528,-0.3002845049,0.0414095186,-0.2625495791,0.2326451987,-0.1294325143,-0.118738912,-0.2153359056,0.0714965984,0.0059646158,-0.2001736164,0.1365029216,0.1608421504,-0.1547916234,0.1932304054,0.2181787491,0.1120679528,0.1524927318,0.1146083027,0.0978649408,0.2603714466,-0.1108557507,0.2093981802,0.0343885422,-0.0138657717,0.1596274674,0.0138404556,-0.1994660795,-0.1868090034,-0.3193314373,-0.0532790348,-0.1618547738,0.2261725217,0.0621839389,0.2645357251,-0.2653847337,-0.579557538,0.4742965698,0.3399647474,-0.1296412796,0.2579099238,-0.2862511873,0.0161555074,-0.0667420179,0.0111744655,0.5658621788,-0.1947555542,0.3254500031,-0.0794114918,-0.0616300069,0.3023921251,-0.4567987621,-0.0652408078,-0.0468615368,-0.3045466542,-0.0021183726,-0.0720989034,-0.1411138028,-0.019983869,-0.0946820825,0.3443318307,-0.1653456241,-0.0544738658,0.1015557647,0.1601570547,0.0502447262,-0.125617519,0.040408846,0.2785877883,-0.1807658076,-0.1113735214,-0.0539935008,0.1307125092,0.0862976462,0.2149008512,-0.318110615,-0.2407790869,0.0139596602,0.3552249372,-0.0003474855,-0.1252682209,-0.0668669567,0.1332379729,0.2062510252,0.2343954444,-0.1449888051,0.3111366034,-0.1219340116,0.1470759958,0.2458626628,-0.1360154152,0.1184892282,-0.0817848593,-0.0365061723,-0.0434313938,0.0705959946,-0.2296567559,-0.1445905268,0.2853672504,0.1721625775,-0.0119023863,-0.2052929401,0.0354664586,0.0315034762,0.0794769973,0.1419689208,0.3844061494,-0.0609966703,0.5599350929,-0.2562152147,-0.4541349113,-0.0462835692,0.0224631652,0.5034426451,-0.2178084552,0.2553486228,-0.0312380884,-0.1767935604,-0.0469373912,0.2533653378,0.0601818264,-0.0204421878,0.1567865461,-0.1809256822,0.2217566222,-0.141422689,-0.4298364818,0.0265338123,-0.3845816851,-0.2996765375,-0.1156658009,-0.4305515587,-0.0555208474,-0.2334783822,0.1969345957,0.02845289,-0.4536219835,0.1628562808,0.4174414575,-0.2616335154,0.098854363,0.1829998791,0.2679968178,0.016152421,-0.1439229846,0.1268773675,-0.1616188288,-0.1091142073,-0.1418680996,-0.1457025111,-0.2609520555,0.0503480397,0.1333000958,0.1298932284,-0.077105023,-0.1611856371,-0.307038933,0.0476326048,-0.2632964253,0.0594224744,0.3452678919,0.2517391741,0.0951137617,0.1064113677,-0.307300061,0.127826184,0.3223632872,0.0648431703,0.2507064641,0.0393301435,0.2024416029,0.1677877158,-0.26296103,-0.3952639401,0.1813610792,-0.0667770877,-0.1152082756,0.1839221567,-0.2410745025,0.2093157619,0.4576960504,-0.0930898637,0.1318606883,-0.0612669587,-0.2261504829,0.0325866304,0.1872493029,0.25838691,-0.1771103144,0.0614770129,0.4482594132,-0.2375608981,0.0974703208,0.1723159403,-0.0793057755,0.0926527232,0.1511339098,0.3712780178,-0.3484845757,0.1774656326,0.1149255708,0.1882874221,-0.16113998,0.4405623376,0.4650237858,0.0315324664,0.5609922409,0.1833839118,0.3716109097,0.5011677742,-0.1915829331,-0.1141740903,-0.2393341213,-0.2015488148,0.4385492206,-0.1161765009,0.0592845194,-0.0714483261,0.2083077878,0.1838032454,-0.3680932522,-0.1831098795,-0.1217206791,-0.3449113369,-0.1100355238,0.4735716879,0.0752969682,0.1437609494,0.2953138351,-0.0625439435,-0.3859573603,0.4411940277,-0.014913612,-0.1918070912,-0.2876868248,0.2943648398,0.2762215734,-0.1601945311,-0.1219862849,0.2891839445,-0.2347111255,-0.0451046489,-0.0269343108,0.1330798268,0.2197633386,0.0857369751,0.0260293651,0.3150087297,0.0979131609,-0.1002512425,-0.2087868601,0.2268522978,-0.236111775,-0.0908068717,-0.0861886814,0.1070382744,0.0798826814,0.2771602571,-0.0013058876,0.2785951495,-0.2990992069,-0.1270029396,0.0565838963,-0.028875012,-0.1554894745,0.3027579486,-0.3916992545,-0.2383483499,0.6524431109,-0.2137705535,0.0422887541,-0.1606041342,0.1084501147,-0.2349987477,0.4478149116,0.468929559,0.0369879045,-0.3087177575,-0.0269653257,-0.2480551004,0.0657562837,-0.3029725552,0.4378798604,-0.0127349431,0.5008688569,-0.2398172468,-0.0382050313,-0.0074696704,-0.0681988969,-0.035166271,-0.131875217,-0.3061086833,-0.0137978168,0.1377976835,0.1543895155,0.0681297779,-0.2035728842,0.2141352594,-0.2355101705,-0.0289051719,-0.1541021019,-0.0110407071,0.1529419571,0.4067371488,0.0864679888,0.1257699132,-0.0493976995,0.152663976,0.1546432823,-0.3914330304,-0.0762771145,0.0519489832,-0.0218606591,0.0463794097,0.003566161,-0.5463690758,0.2738312185,0.1344443262,0.1515663415,-0.161591813,0.0087533742,0.1073768139,-0.3834522665,-0.204690069,0.3414939046,0.312323451,0.1437021941,-0.0722696111,-0.0394230448,-0.3186280429,-0.078977786,0.3032265007,0.0869207084,-0.4065341651,-0.003789702,0.1015667692,0.5180827379,-0.1703476906,-0.3908232152,0.147650063,0.4772286117,-0.1026958749,-0.1970788091,0.2607144415,-0.1244606301,0.2009640038,-0.0636367649,0.5886341333,-0.0591424331,-0.0374849141,0.2227338403,-0.1312392503]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1973","title":"Question: what gets stored in the datasets cache and why is it so huge?","comments":"Thank you!","body":"I'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!","comment_length":2,"text":"Question: what gets stored in the datasets cache and why is it so huge?\nI'm running several training jobs (around 10) with a relatively large dataset (3M samples). The datasets cache reached 178G and it seems really large. What is it stored in there and why is it so large? I don't think I noticed this problem before and seems to be related to the new version of the datasets library. Any insight? Thank you!\nThank you!","embeddings":[-0.093276009,-0.0431144089,-0.1500352472,0.5251605511,0.1169909835,0.2516939938,-0.0498338975,0.2222593725,-0.1382898241,-0.0774006546,-0.0397733524,-0.2352829874,-0.0839777291,-0.1979648769,0.1002384275,0.1400149614,0.0831124857,-0.097200051,-0.0941666216,-0.1367502064,-0.1051238701,-0.0279082078,-0.0203791987,-0.1615523845,-0.5561434031,-0.1613333076,-0.0992957205,-0.0143713923,-0.173880145,-0.1549727768,0.1389190555,-0.0815957114,0.3282954395,0.4119418859,-0.0001099192,-0.3336604536,0.3809490204,0.059466552,-0.4237931967,0.3471472859,-0.5078397989,-0.2993497252,-0.0722064227,-0.1197393537,0.3782676756,-0.0046722153,-0.038674973,-0.4304594994,-0.0937503725,0.2323083878,0.266400069,-0.2057155818,-0.3067412078,0.0793622881,0.1548915058,0.0190567952,-0.1256777346,-0.2989519835,0.2981007993,0.2503693998,-0.1426252127,0.1356903166,0.1248135716,-0.0012544931,0.1985639334,-0.0607421659,-0.2490263879,-0.244838357,0.6020487547,0.2472292036,0.9486544728,-0.2768333852,-0.1108004302,-0.1819037497,-0.140663147,-0.1601852924,0.240489617,0.4229886234,0.1148210317,-0.017354684,-0.5174568295,-0.4082565606,-0.1164194196,-0.1364544183,-0.2701277733,-0.1550516635,-0.2891930342,0.0476053469,0.1391430944,-0.0293958485,0.3759151697,-0.2152040452,-0.1482262611,0.1781728566,-0.3746461868,-0.3132804334,-0.1754294932,0.5183995366,0.175896883,0.0355411172,-0.3193719387,-0.1049579382,-0.1499888748,-0.0058422522,0.2283180803,0.3312994242,-0.0872190967,0.1299627274,-0.0415721983,-0.3952956498,-0.251203835,-0.1190739796,0.1382246017,0.0840223208,0.4433962703,-0.4480436742,-0.1275271326,-0.3552954197,0.2091561854,0.150101319,-0.058311902,-0.0533158295,0.1636088789,0.1649718136,-0.1942420006,0.1913871765,-0.3255336881,-0.2493953556,0.0491348356,-0.2023278773,-0.2613448203,-0.1175599396,-0.1176537871,0.2837980688,0.2783750892,-0.1868805736,0.2782343924,0.0942072272,-0.0542964749,-0.1353007853,0.2577476203,-0.3480947614,0.3744066954,0.3860787153,-0.1004103199,0.3397671282,-0.2112310976,0.1938162744,-0.1748200506,0.1912363768,-0.4954328835,-0.4762571156,-0.0555787981,0.1161494628,-0.1490110904,-0.0265013985,-0.296798259,0.048915837,0.4143257439,0.1898825467,0.2373777181,-0.1148351878,-0.3927145004,-0.2508896291,-0.1605243832,0.1593327373,-0.3977333307,0.233879447,-0.0485821813,0.1352215409,0.0289089382,0.3458126485,-0.1337621212,0.0476840883,-0.0410713255,-0.119875662,0.0387460478,-0.0323211253,-0.6220085025,0.0979867578,0.3401122987,-0.1395850629,-0.1133324802,0.4246947169,0.0139003126,-0.0529296845,-0.1726596504,0.3259518147,-0.2444830388,-0.1414567083,-0.2859779298,-0.0662527233,0.0541431867,-0.0253438298,0.1098343432,-0.1615825742,-0.0214516055,0.2977244258,-0.0521551631,0.1725430936,0.1138234586,0.320427686,-0.0534516871,0.3505656421,0.2981018126,0.0280147307,-0.3736860454,0.2298048884,0.2132073343,-0.677041471,0.1925313026,-0.1264005154,0.1930011511,0.1073375568,-0.0159968492,-0.1633145362,-0.0019701256,0.094193846,0.0692311078,-0.1790812314,-0.0584780984,0.5535147786,-0.0396658033,-0.1265941709,-0.4304448664,-0.1942278594,0.094344385,0.0151443006,-0.1126897112,0.0254178047,0.0723584965,-0.0598642677,-0.1487295181,0.127148211,0.2232443094,-0.1721601784,0.2070409805,0.5801752806,-0.0443099961,0.0983168557,0.3464978933,-0.0851986706,0.075654082,-0.1376399398,-0.2005127519,-0.0036713409,-0.1321956217,0.0896138921,0.010150075,-0.092445299,0.0108425347,0.0487444811,0.2794580162,0.0613052994,0.1610013247,-0.1444362998,0.3293243945,0.4313061237,-0.1017219052,0.2352163494,0.4231288135,-0.1432104856,0.0016113913,0.1080245152,-0.2741894424,-0.4883163869,0.1158694476,0.3100363314,0.4469021559,0.0864019096,0.333599776,-0.1080225483,0.2362550646,-0.1760383993,0.0613435246,0.078651242,-0.0024266345,-0.272269398,0.1289169192,-0.0161494352,-0.0493577383,0.2004240304,-0.0738437921,0.2921090126,-0.0273528062,0.3947324455,-0.2867271006,-0.1300338507,0.0338745713,0.3088098764,-0.2400012314,-0.192703858,-0.1028382927,-0.1125453562,0.2906282842,-0.172575444,-0.2111422122,0.5469235778,0.0419968739,0.1554718167,0.1039048508,-0.092834875,-0.2938800454,0.1249580011,0.1484294385,-0.2920488715,0.22932221,0.1391724199,0.0667569488,-0.3075315654,-0.2181171924,-0.1431798637,0.1764889956,0.2055828273,-0.2787579298,-0.0143719399,-0.2872858644,0.122208856,-0.3718659878,-0.1516460329,-0.0944587216,0.0969584137,-0.0684211552,0.0045432835,0.2092037052,-0.0072683538,-0.2206558585,-0.0041540372,0.1962669045,-0.1676088274,0.0026187426,0.2360999137,-0.18823798,0.0499798991,-0.2465439439,-0.0251868255,-0.3899688125,-0.797157824,0.2943048775,0.0502677411,-0.1699786484,0.1299383044,0.1982038468,-0.0022327369,0.1979551464,-0.7925055027,0.1653409451,-0.2428253442,0.196601972,0.0208892487,-0.017237775,0.2112317681,-0.2289224863,0.0143871456,-0.1428320408,-0.2490252256,-0.3274752796,0.109886311,0.5421879888,-0.2484795749,-0.0671518147,0.2619904876,0.5072640777,0.4594123363,-0.0740814134,-0.0782492459,0.3955213726,0.4682129622,-0.3141708672,-0.0679075569,0.2521081865,-0.0545631759,-0.0373573713,0.4049779475,0.3252908587,-0.1929706633,0.2231424451,-0.0215327814,0.0985536799,0.0911554918,0.1883597225,-0.3275538683,0.1542470306,0.1886700988,-0.1808984429,-0.0533187613,-0.4057599604,-0.0513983518,-0.0306312572,0.0280588474,0.2627258301,-0.2404523939,-0.0450364314,-0.0027814989,0.1284266859,0.0271358509,0.0745979846,-0.0780795366,0.2845902741,0.2828457057,0.2910917103,0.3174207509,-0.3396304548,-0.0846997425,-0.1930547655,-0.1963182837,0.0259428266,-0.0224295296,0.0550437123,0.2927469611,0.1410840601,0.0753163174,0.326397717,-0.1245106608,0.0797332227,0.3650948703,-0.3474428654,-0.3324900866,0.0470677242,-0.1765142083,0.3243806958,-0.0945608467,-0.1253634691,-0.2191181034,0.0165884588,-0.021505015,-0.211133644,0.128383562,0.1492261291,-0.1286559701,0.0826712549,0.3063970208,0.1286906749,0.1678551137,0.2026660293,0.0713185444,0.3028711081,-0.1005303785,0.1832405329,0.0893841535,0.0172607098,0.2023429424,0.0775390714,-0.1676412225,-0.1547068506,-0.2783543766,-0.0984492004,-0.1424423456,0.2424081564,0.041176267,0.1923865229,-0.3316679597,-0.5521623492,0.4700379074,0.3056497872,-0.1144093275,0.2945451736,-0.3459411561,-0.0206749327,-0.0994963422,0.0236073602,0.6341750026,-0.1885597557,0.3193606436,-0.1602065563,-0.0349118821,0.3534514904,-0.4713695943,-0.102197662,-0.0627890229,-0.2784768939,-0.0164858848,-0.0837881565,-0.1141886115,-0.0374989435,-0.0432551578,0.2957080901,-0.135757342,0.0166549832,0.131731078,0.1624050736,0.085610874,-0.1001110151,0.0457270779,0.2781667113,-0.1420331299,-0.0983969122,-0.0710744187,0.1140255556,0.0678911209,0.2308056802,-0.3431584537,-0.1678885072,-0.0454373993,0.3140927851,0.0495369248,-0.1369927078,-0.0279311612,0.115493536,0.2825072408,0.2087309361,-0.1547801495,0.2885030508,-0.0454521067,0.0966111273,0.2257048339,-0.1604085267,0.2028546929,-0.0742523894,-0.0151800131,0.0295604803,0.0734269246,-0.2166202962,-0.1963933408,0.2952686548,0.1863598675,-0.0282136314,-0.1638399363,0.0749168247,0.1513376981,0.0881335139,0.154373467,0.3600261211,-0.0543503985,0.537355125,-0.1618746072,-0.4731983542,-0.0721386671,0.0677511692,0.516712904,-0.2706273198,0.2418920547,0.0224652346,-0.1990827024,-0.0185677111,0.2447337657,0.1431670487,-0.0030223557,0.0718462393,-0.2016819715,0.2234190106,-0.1091552228,-0.428373307,0.0274119973,-0.3796580434,-0.2289860547,-0.1432086974,-0.3966193795,-0.0543786734,-0.2849994302,0.2259270549,0.022139851,-0.422678411,0.1597841233,0.5000872016,-0.2631738782,0.0833807737,0.1169349402,0.2854115665,0.0662835985,-0.0973123088,0.0697784722,-0.106771864,-0.1027365997,-0.0564703904,-0.1701865643,-0.2685880959,0.041633077,0.121981509,0.1881790906,-0.0928609744,-0.1890252084,-0.3496143222,0.0264189523,-0.1652932167,0.0376300178,0.3092354834,0.2898817956,0.0977420211,0.0932450294,-0.2944804132,0.0824089572,0.3522119522,0.0532983318,0.2886715829,0.0523715578,0.1505593956,0.1289164126,-0.2599317431,-0.320299238,0.1719150245,-0.1021982878,-0.1519041359,0.2000733465,-0.2343578488,0.242953375,0.4122782648,-0.1516736299,0.1414520591,-0.0604684316,-0.2107726336,0.1105462164,0.1698565781,0.2486298233,-0.1103553176,0.0520400964,0.4889526963,-0.2491535246,0.115041092,0.158088237,-0.0294891838,0.082156077,0.1283503473,0.3319914937,-0.306276381,0.1198209822,0.0120764747,0.1549917459,-0.1110656634,0.4483393133,0.4729842246,0.0416496061,0.5058625937,0.2680700421,0.3988321722,0.4374980032,-0.2157473862,-0.0218933169,-0.2589825094,-0.1739058048,0.4412644804,-0.1299243569,0.0350021943,-0.0519282185,0.1892955899,0.2301977277,-0.3319939673,-0.118498385,-0.0579402186,-0.3430294394,-0.0681128949,0.3798815012,0.031580396,0.1382098794,0.2852560282,-0.0750496686,-0.3461761773,0.5272699594,-0.0084154354,-0.1911329031,-0.3182269037,0.2836037278,0.2136116028,-0.1539754421,-0.1132363677,0.3115799725,-0.3228003681,-0.064494893,0.0528002195,0.1617187858,0.1342451423,0.1119954959,0.0052519329,0.3138703704,0.127148971,-0.1226738393,-0.1524906754,0.2763051093,-0.1418131441,-0.1579709053,-0.0520270988,0.1068974733,0.0814670622,0.2794260681,-0.0366242118,0.2377297729,-0.2635088265,-0.1167068332,0.0852560923,-0.0010798151,-0.1774182767,0.3059797883,-0.4137663841,-0.1718048453,0.6572293043,-0.2769588232,0.0466738306,-0.1554091573,0.1024168357,-0.2232591361,0.5363010764,0.5385504365,0.0594781823,-0.2584859133,-0.0594528951,-0.2399016619,0.14623712,-0.2901110053,0.4619367421,-0.0910703689,0.5299177766,-0.1890094578,-0.0065966598,-0.0372081883,-0.0913438499,-0.0227755941,-0.1192819849,-0.3284572363,-0.0476733223,0.1545762271,0.1598671526,0.0452581048,-0.1416112185,0.3050190508,-0.2754768133,-0.0113205025,-0.1833556145,-0.0434422493,0.1503605247,0.3183854818,0.07886222,0.0568814278,0.0019152631,0.156356886,0.2113881707,-0.3921956122,-0.0894638002,0.1129502952,0.0212053806,0.0789426714,-0.021748526,-0.4952470362,0.1933276355,0.1648822427,0.0909024402,-0.2228275687,-0.0061289961,0.0728331804,-0.4129471779,-0.2074353695,0.3323971331,0.3403295577,0.1628835499,-0.0909257382,-0.0492659993,-0.2456224561,-0.1233864501,0.3075986803,-0.0276221726,-0.349627614,-0.0242450349,0.1000767276,0.4839314222,-0.1577885598,-0.3946616352,0.1313188821,0.4674474895,-0.1451134533,-0.1747546196,0.2652282715,-0.1087836996,0.2228713036,-0.1194973812,0.5708802938,-0.0706261471,-0.0238189772,0.2222152203,-0.1409866363]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1972","title":"'Dataset' object has no attribute 'rename_column'","comments":"Hi ! `rename_column` has been added recently and will be available in the next release","body":"'Dataset' object has no attribute 'rename_column'","comment_length":15,"text":"'Dataset' object has no attribute 'rename_column'\n'Dataset' object has no attribute 'rename_column'\nHi ! `rename_column` has been added recently and will be available in the next release","embeddings":[-0.1480205953,0.2673698366,-0.1891879886,-0.2122284621,0.4945076108,0.2257964909,0.526695013,0.5100852847,-0.0069026817,0.2416522056,0.0692551285,0.4630435109,0.0955998227,0.2902490795,-0.1881605685,-0.1668882519,0.1073040441,0.1791319996,-0.0695754439,0.0544452891,-0.310726434,0.1311455816,-0.0807547495,0.2260072976,-0.1303207129,-0.0281751119,0.1008253545,-0.0371696092,-0.2674559951,-0.2563641071,-0.0171613228,0.2009526193,-0.045062758,0.3236352801,-0.0000926938,-0.0632152334,0.2490769774,0.0662861168,-0.3413289785,-0.006262288,-0.2956322134,-0.1352452487,0.0950806141,-0.3032380342,-0.0807811394,-0.2669169307,0.0652627945,-0.1756295264,-0.0559086651,0.3902206123,0.4301387966,-0.0322908275,0.1253099293,-0.4424621463,0.1918611526,-0.1107830033,-0.2682177722,-0.1102858856,-0.007194594,-0.2042922527,0.077012293,0.3974594176,-0.0113447849,0.0387128778,0.1085092276,0.0573984012,-0.0773152858,-0.0708466694,0.5388149619,-0.0235521905,0.5927603245,-0.3442261219,-0.1485881805,0.1486360431,0.2113902122,-0.2700507641,0.114813365,-0.1621433645,0.2576489449,0.0862666517,-0.047085572,-0.3145783842,0.0067425421,0.0226357244,-0.0003393269,0.0747529492,-0.1425303817,0.0510961711,0.0354336016,-0.1930792332,0.2353818566,0.034327928,-0.1335516274,-0.0680656955,-0.2258865088,0.0163827203,0.0826951638,-0.0045984983,-0.0762810633,0.049839545,-0.0036279326,-0.0832191035,0.3695904315,0.1721154153,-0.0216460396,0.1260512769,0.2433011979,0.2542546093,0.3802496791,0.0975107029,-0.0875926688,-0.2376025915,-0.1819169521,-0.1043573692,-0.0333304964,0.0813894793,0.2577290535,-0.0474398956,-0.3202418089,0.0599253848,0.2252569497,-0.1256399006,0.0028607675,0.3451608121,0.0479501039,0.0295492858,-0.0497932918,0.1249662936,0.0375541002,-0.1547208726,-0.3227243125,0.167449832,-0.1810479313,-0.1722711176,0.1666683555,0.1421059668,0.1395260543,-0.0943465084,-0.0464925356,0.0271551665,0.0591181554,-0.0214488674,-0.2104855031,0.2149101496,0.0345261022,0.0420472175,0.191686973,-0.2598477602,0.00537764,0.060029,-0.1884548366,-0.2045653313,-0.3990676999,0.3650212884,0.3822282851,-0.1665391922,0.0971240774,0.1239283383,0.2216399163,-0.1890992969,-0.1021855026,0.0530204698,0.1605154276,-0.1461176276,0.068583563,0.1838475168,-0.3581221402,0.0134650553,-0.2887741625,-0.0868131444,-0.0167985447,-0.2142134309,0.0198602024,0.0224956218,0.0302110668,0.0697153211,0.2355883271,-0.2709612548,-0.0858065411,-0.0233693719,-0.1932396293,-0.3842564523,0.0728418604,-0.0179815125,0.0777758732,0.0289092343,0.0001888713,0.0040612649,0.0657594055,0.0753392428,-0.0487518609,-0.1657735556,0.0360384807,0.0990325958,0.0208247676,0.2514448464,0.1152186617,-0.0176038258,-0.0835476965,-0.0945646092,0.0493910238,0.1367905736,0.2716975808,0.0668769777,-0.1081464812,-0.1380594969,-0.5242903233,-0.1051994264,0.3412159681,0.0549163818,-0.0587885901,-0.1475253254,-0.3400182724,-0.0101279775,-0.0550659709,-0.0212706942,0.3860673308,0.3193697035,-0.0134630036,-0.0553289242,-0.3478039503,0.0267904773,-0.1508700103,-0.0205303747,-0.1836133003,-0.0409139469,-0.1973558962,-0.0668869242,-0.0609186254,0.0232243706,0.1579786986,0.0522689149,-0.1797013581,0.4454961419,-0.0546321124,-0.0396266952,0.2310346514,0.0945145115,-0.2082560211,-0.0159746837,-0.057502266,-0.21474123,0.0896259397,0.1060440019,-0.2134995311,0.1804047227,-0.0159970727,0.1296051294,0.0871786252,0.0675841123,0.2908117175,-0.218098253,0.0525978878,-0.4199349582,-0.1031184867,0.0483449958,-0.0173058361,-0.144408226,-0.2945400178,0.0188788194,0.6333140731,0.1176865473,0.1106442809,-0.0410701819,0.0029520553,-0.0365373716,-0.0374096483,0.4473365843,0.3660518825,0.291713208,-0.0790926665,0.0386721529,0.0271348618,-0.1462706178,0.2341074347,-0.0970844701,0.1014942601,0.1980494708,0.2248340696,-0.2653696537,-0.3610744178,-0.2014162987,0.3461225629,0.2907543778,-0.1158341095,-0.1547301859,-0.317609489,-0.0216606222,0.1577748805,-0.0571727417,0.0258274917,-0.3772435784,0.1997126341,0.1574040651,-0.1107139736,0.3284634948,-0.1047199816,0.049320478,-0.0491220914,0.1450624466,-0.2935516536,-0.3051329851,0.0327566043,0.2360943854,-0.0038471299,-0.0952004865,0.2388827354,-0.1697970182,0.1575883925,-0.3804335892,-0.2947395742,0.0679295063,-0.1176331341,0.1239461675,0.2836106122,0.1835907698,-0.1257891357,-0.2704280019,0.3789155185,-0.0872144401,-0.0782680586,0.0786567479,0.0126770837,-0.2021655589,-0.1481722891,-0.5371339321,-0.2125718296,-0.2981596291,0.0395348296,-0.1193330809,0.0404338874,0.1268253773,-0.0288182329,0.0682808608,-0.0434517413,0.1633730382,-0.355500102,-0.2119586468,0.1214780882,-0.3827038109,-0.1991724223,0.0622373931,0.245855093,0.1395715326,-0.1114470884,-0.1340412199,-0.3259519339,-0.1066582203,0.1783331931,0.1086397171,0.0194965247,0.2268592417,0.3485840857,-0.3652878106,-0.1613431424,-0.221971035,0.116874896,0.0998080745,0.068968527,-0.1211088747,0.4541032612,-0.2014998347,0.1520923376,-0.1756862849,-0.0391680673,0.3083595634,-0.3411304057,0.2872065604,-0.2163613141,-0.3523576558,0.096776329,-0.1076208726,0.1117239371,0.1249361336,-0.2410759628,-0.0873407573,-0.1596370637,0.2966870666,-0.2368938029,-0.0857199132,-0.2308347225,0.0809220895,0.1734250933,0.1985884458,0.0865165889,-0.2180884629,-0.0792569593,0.1132837683,0.100642167,-0.2390064001,-0.2756571472,-0.4863038957,-0.2102683783,-0.2669141293,0.2848839164,-0.1865969002,0.1732610613,-0.1746491492,-0.1776549816,0.0585991628,-0.0255937539,0.6016751528,-0.2537382841,-0.1737802625,0.1780041009,0.1378784329,0.1953620166,-0.0066862507,-0.1487511396,0.227542311,-0.0614164174,0.3027277291,-0.2554592192,-0.1527983397,0.4616550207,-0.0647035316,-0.079406321,0.1208013073,-0.247358188,-0.2539885044,-0.3022773266,-0.1140802503,-0.0684003457,0.3441505134,0.0568245687,0.0054405578,-0.1038327962,-0.1635390222,0.0367310829,0.3356334269,0.2391733378,0.0446276627,0.0695558563,-0.044464279,0.2265317589,0.3430849612,0.2867485881,-0.4382374287,-0.3304234445,0.1850846559,-0.3098658919,0.4284175932,0.0306044854,-0.0725116432,0.252699703,-0.1457138658,0.2394143492,0.0675392523,0.0511419177,0.3273399472,0.0310181864,-0.1440851986,-0.4038775861,0.266962707,0.0918097347,0.007480057,0.3013288081,0.2620114982,-0.1831329316,0.0549538694,-0.1113177091,0.6837829351,-0.004706528,-0.2229882479,-0.0665146708,-0.3429149389,0.3429071307,-0.1546102017,0.2809612751,-0.3597823977,-0.3364992738,0.0832097232,0.146709159,0.3844800293,0.1262198985,-0.2122838795,0.1197039858,-0.2312849909,0.2236239165,-0.1565447897,0.1038186252,-0.0749703646,-0.2371715456,-0.0486385301,0.3325895667,-0.0908745602,-0.067587629,-0.0258903354,0.1619464606,0.058512032,-0.0740995035,-0.0594003126,0.057971973,0.1429982632,0.2783632874,-0.0219333358,-0.1894520521,-0.0546426475,-0.0100114997,0.2719328105,0.0803741291,-0.0963101238,0.297724396,0.0871086046,0.1816965789,-0.1825905591,0.1677411795,0.1787417829,-0.1118285134,-0.243840307,0.0457880013,-0.0607100502,-0.1036731526,-0.1899767816,-0.0490300283,0.3122725189,-0.244493112,-0.0786403045,0.0081410473,-0.0044303834,-0.4761295915,0.3348542452,-0.1799605489,-0.1032378301,0.072444357,0.1087355539,-0.2025539279,-0.0554285832,0.2966775298,-0.0392871685,0.1967197806,0.5504460335,0.0448470004,-0.1884283125,-0.4120585322,-0.0801275,0.3164348602,-0.3292592764,0.1787674278,0.1684602499,-0.1998119205,0.2051618695,0.2007067353,-0.2299065292,0.2235673964,-0.1491681635,-0.2777605653,-0.2230476141,0.1140504926,0.0821096003,0.156032905,-0.0997488946,0.1267519444,0.0860068277,-0.1033679768,-0.5092449188,-0.105319269,-0.2658164799,0.1509397179,0.1610097438,-0.0610453561,0.0272355918,-0.0828104615,0.4158418179,0.0533655249,-0.0202236213,-0.4396899641,-0.0818807706,0.0378424786,0.0225864798,0.1542374939,-0.2144750953,-0.1528184414,-0.3356893361,-0.0601361319,0.0256923847,0.096166119,0.0748494267,0.3337559998,-0.0444297008,-0.0105783548,0.2925128043,-0.1168747544,0.1618533134,0.0065436652,-0.1475300044,0.0515553765,-0.0166055094,0.0349573232,-0.0458759628,0.0928954333,0.0280128047,0.0560145415,0.1608013809,-0.1276749521,0.0751641989,0.1455015242,0.264865458,0.2653462589,-0.0048148641,-0.1082087904,0.2926561832,0.4193916917,-0.533172369,-0.0675504506,0.0832463875,-0.1591645777,0.1647067219,-0.1210749447,-0.0497693755,0.1405865252,-0.1378156394,0.1416704655,0.323469311,-0.2824817896,0.0907106698,0.4998244345,-0.0476826131,0.1336069703,0.3606959879,0.2890868783,-0.2043738812,0.7021962404,0.107870087,0.2113151997,0.2455170155,0.089454636,0.107606113,-0.0993981808,0.4570518136,0.0587081574,0.0520744585,0.0344241969,0.1065681875,0.0252360329,-0.123126246,0.0842268616,-0.2626786828,0.1908276677,-0.0585932136,-0.2600926161,-0.1458389461,-0.0884479135,-0.0445868373,0.0311142076,-0.0598824508,0.1458694786,0.5411097407,-0.175549522,0.1312216967,-0.430306077,0.1280246079,0.0826762542,0.1712614596,-0.0920844376,0.060225755,0.1641637087,-0.0091685392,0.1622570902,0.2394184172,0.0751857236,0.1047730446,0.1376175582,-0.1273897439,-0.209139511,-0.3048556447,-0.0004940034,0.2981129289,0.1749025136,0.1421924233,0.030074995,0.3311495185,-0.2573341727,0.0983779505,0.2216239721,0.0953385159,-0.0019299217,0.2055904567,0.1140879914,-0.2734469771,-0.0592406243,-0.0815684125,-0.2983763814,-0.1440274417,0.3373300135,-0.0916036218,-0.0182954073,-0.2200299054,0.2184695303,0.1244773939,0.2653069496,0.003936565,-0.2051738948,-0.0742229372,-0.3145595789,-0.4338268042,0.1091861725,0.087865904,-0.0965908021,0.0225743838,-0.054097984,0.1319004148,-0.1511961669,0.20744811,-0.1973023564,-0.0224528909,0.1340035498,-0.3383761942,0.0024020686,-0.0140556339,0.1266244501,0.1374220252,-0.1364718527,0.0664770678,0.2662596107,0.2039011121,-0.385979265,0.0008109563,0.1797762811,-0.1163298637,0.2950178981,-0.0843664557,0.3050220907,-0.2873917222,-0.0018566244,-0.031289652,-0.2810800076,-0.240995571,0.316033721,0.1748432219,-0.0286566988,0.1733542979,0.0222293269,-0.0884622782,0.2141535282,-0.1021218672,0.2154707462,-0.2071211785,0.1852338165,0.3606932461,-0.1293056458,0.0092875604,-0.1204326302,-0.0262249056,0.2712040544,-0.2253396511,-0.396469444,0.4225472808,-0.057189066,-0.5882441401,-0.039294783,0.1613834649,-0.0684370622,-0.1786171943,-0.4791527987,0.0589543805,0.1514200866,-0.0811301395,-0.1250235289,0.3021185994,-0.1249189004,0.1843337715,0.0491681285,0.3032671511,0.1752819866,-0.0793284625,-0.2028675228,0.0078004031]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1965","title":"Can we parallelized the add_faiss_index process over dataset shards ?","comments":"Hi !\r\nAs far as I know not all faiss indexes can be computed in parallel and then merged. \r\nFor example [here](https:\/\/github.com\/facebookresearch\/faiss\/wiki\/Special-operations-on-indexes#splitting-and-merging-indexes) is is mentioned that only IndexIVF indexes can be merged.\r\nMoreover faiss already works using multithreading to parallelize the workload over your different CPU cores. You can find more info [here](https:\/\/github.com\/facebookresearch\/faiss\/wiki\/Threads-and-asynchronous-calls#internal-threading)\r\nSo I feel like the gains we would get by implementing a parallel `add_faiss_index` would not be that important, but let me know what you think.\r\n","body":"I am thinking of making the **add_faiss_index** process faster. What if we run the add_faiss_index process on separate dataset shards and then combine them before (dataset.concatenate) saving the faiss.index file ?\r\n\r\nI feel theoretically this will reduce the accuracy of retrieval since it affects the indexing process.\r\n\r\n@lhoestq\r\n","comment_length":79,"text":"Can we parallelized the add_faiss_index process over dataset shards ?\nI am thinking of making the **add_faiss_index** process faster. What if we run the add_faiss_index process on separate dataset shards and then combine them before (dataset.concatenate) saving the faiss.index file ?\r\n\r\nI feel theoretically this will reduce the accuracy of retrieval since it affects the indexing process.\r\n\r\n@lhoestq\r\n\nHi !\r\nAs far as I know not all faiss indexes can be computed in parallel and then merged. \r\nFor example [here](https:\/\/github.com\/facebookresearch\/faiss\/wiki\/Special-operations-on-indexes#splitting-and-merging-indexes) is is mentioned that only IndexIVF indexes can be merged.\r\nMoreover faiss already works using multithreading to parallelize the workload over your different CPU cores. You can find more info [here](https:\/\/github.com\/facebookresearch\/faiss\/wiki\/Threads-and-asynchronous-calls#internal-threading)\r\nSo I feel like the gains we would get by implementing a parallel `add_faiss_index` would not be that important, but let me know what you think.\r\n","embeddings":[-0.2789407969,-0.0627319291,-0.1573459059,0.1106119156,-0.3771443665,0.2478514761,0.2097888738,0.1062705293,0.1371025592,0.1719250977,-0.1766027361,-0.1353845596,0.3414870501,0.1139966771,-0.3853704035,0.24953866,0.3513414562,-0.1033916771,0.2524386942,0.1125721633,-0.317397058,0.0103294859,0.029140437,-0.2674212158,-0.178014636,0.3508556485,-0.2582988143,0.0532396846,-0.0132865068,-0.2483889163,-0.2480458915,0.3926287889,0.067936942,0.3004924357,-0.0001269445,-0.0502257869,0.1069203466,0.0691175684,0.1533449739,0.5967572331,-0.2711214423,0.0120974537,-0.0762948021,-0.1729471833,-0.0013871079,-0.3030721545,-0.0339291841,-0.2620636523,-0.2039729506,-0.1719792485,-0.010249096,-0.1237887368,0.0848588869,0.0481446125,-0.0428055003,-0.2038855404,-0.1553710401,-0.1455569118,0.0888659507,0.223718822,0.2321635634,0.1709749401,0.0189927835,-0.1704597175,0.087762326,-0.0448629223,0.4525159895,-0.0526509918,0.0105196806,-0.1225785241,0.1086316556,0.0454834178,-0.4651236236,-0.2125894278,0.0684300289,0.0824425891,-0.2274452597,-0.1753317863,0.0390132815,0.0092715155,0.0714065358,-0.4730629027,0.049428571,0.1115148365,0.3911833167,0.0608306788,0.2611269355,0.1117506102,0.4002321959,-0.1047056168,0.0262903646,-0.1128797978,-0.0963885337,0.1314448863,-0.6142691374,0.0232872348,-0.0899243951,-0.2996670008,0.0209422838,-0.1793354005,-0.3022100329,0.2885174751,0.0202772673,0.1198161766,0.0016666742,-0.1536525786,0.2066206783,-0.250123322,-0.0125038568,-0.4881715775,-0.1144803241,0.1374482512,0.1478853822,-0.120116204,-0.6300248504,-0.1439967901,-0.3549608886,-0.1606156528,-0.0620705113,-0.1819740832,0.0238425918,-0.0995018035,0.1697277874,0.1614187956,0.3377495706,-0.0016715091,-0.1338706017,0.0112249758,-0.0672990829,-0.0540435165,-0.0022989584,-0.2795442045,0.2434519082,0.5302842259,-0.0042607989,-0.2697486579,-0.4165178835,-0.0278765894,-0.0764702484,0.1563667655,0.0336068086,-0.1561209559,0.0336155035,0.0777927712,-0.0235536657,0.0020426498,-0.0585453063,0.1687937975,-0.2429046035,0.1656707376,-0.2159646302,-0.3029153943,0.0556750037,0.0054091727,-0.0320964903,-0.1371088624,-0.1657654494,0.5699167848,0.1761478484,-0.0432697684,-0.0384708233,0.1838904023,-0.3282163739,-0.1131456792,0.1609256566,0.1564359665,-0.2961857617,-0.2146445662,-0.1436220556,-0.18578583,0.006723281,0.5707305074,-0.0211948249,0.0817070678,-0.287550211,0.5438025594,0.2934736013,-0.2815659344,0.0505191758,0.0885654539,-0.0611449778,-0.1653109938,0.0520843193,0.1922506392,0.1179279983,-0.0568645895,0.2276799828,0.2944645584,-0.0839017108,-0.1194113791,-0.2615699768,-0.3065003753,0.493680656,0.2430091202,-0.1071245745,0.0761088356,0.043990504,-0.7627868652,0.2860036194,-0.1323405355,0.1944724917,0.0383248217,0.2891156077,0.442709893,0.4458394647,-0.1140970737,0.0472650602,0.2294330001,-0.15346843,-0.001935605,0.3032384217,-0.3450814784,0.4740671217,-0.0507173128,0.0305639897,0.3955182433,-0.0676674023,-0.2117543072,0.1425748765,-0.3888503909,-0.331433624,0.4658032656,-0.3955696523,-0.1659502089,-0.0724599287,0.226380825,-0.0312096868,-0.033708334,-0.3186895251,0.0959291533,0.1591546685,-0.2405869514,-0.1059324145,0.3029584587,0.0247303136,0.0376866609,0.9103011489,0.2225018591,-0.0706852973,0.2421095669,0.0581994504,-0.25952214,-0.1585788429,-0.1893134713,0.3330977261,0.0545157604,0.088710919,0.5726435184,0.002658088,-0.1720752865,0.1380843222,0.161930114,0.0660996661,-0.0243275408,-0.0152292252,-0.1266034693,-0.1130965874,0.0886519253,-0.0277289692,0.5910162926,-0.2684080601,-0.0822856203,-0.1726530641,-0.0034399868,-0.0326986425,0.0887022391,-0.042020902,-0.1948680133,0.3121991754,0.1168773323,0.0520308949,-0.1584303528,0.380494982,-0.1136476398,-0.0397205241,0.3633241057,0.2026499957,0.0750135109,0.1684561819,-0.05808606,-0.0571013205,-0.1315051168,0.0688443705,0.0319430418,-0.011060901,-0.0850690529,0.1229411289,0.080090642,-0.1491916329,-0.2043133527,-0.396911025,-0.1968986541,0.2237753868,0.012626024,-0.2535367608,-0.0580514595,-0.0775191262,0.5329561234,-0.0946544781,-0.2684037983,-0.3124780953,-0.017047016,0.1478615552,-0.1259402335,0.2970302105,0.1887925416,-0.053231474,0.1188512146,-0.0868266001,-0.3709445,-0.0934977531,-0.1448415518,-0.0173402298,-0.1443885565,-0.2989169657,-0.2838769853,-0.1429505646,-0.1644759923,0.1537098587,-0.0450639203,-0.0964550599,-0.2686010003,0.0291599222,0.0496852957,0.1025891304,0.074838832,-0.3937217891,-0.2396276742,0.3449154496,-0.0618611723,0.0971541405,-0.3533826768,-0.2882671356,-0.1973514408,-0.2623565495,-0.2343729287,-0.3656171858,0.0202944782,0.1619838923,-0.1117467508,-0.1317776442,0.0018892486,0.0815019831,-0.2325874418,0.3353889287,0.0072120754,0.0734021515,-0.006610916,0.1284045875,-0.2605769634,0.1623670757,0.1447145343,-0.1115813255,-0.0068714102,0.0925001502,0.1554604173,0.2978833616,0.1614468992,0.1604870558,-0.1416181326,-0.0257230587,-0.1567706168,0.7340026498,0.2577253282,-0.0254654419,0.152501151,0.1177449226,0.0364684947,0.0644167885,-0.1403583139,0.2962310612,0.0880492479,-0.1183745638,0.0249202196,0.0486862212,-0.0981472284,0.0440148413,-0.1993707865,-0.1111701056,-0.1187808365,-0.0568945333,0.2207621336,0.5126085877,0.0025169072,-0.0549470671,-0.2808338702,0.0094224242,0.1832819134,0.3620166481,0.4992682934,-0.1927378774,0.1517201066,-0.2025240511,-0.465570122,0.4637286961,0.0087807486,0.2816376686,0.0165974014,-0.044790972,0.2494057417,-0.1914137453,0.8901629448,0.0034871418,-0.4772203863,0.1134795472,-0.0452145226,-0.3266536295,-0.2326374501,0.3000526428,0.2172367275,0.4511134028,0.2308520824,-0.2460670769,-0.3018375337,0.1756523997,0.0751908049,-0.2882432342,-0.5182569027,-0.1469023526,0.0661566332,-0.0823279321,0.1156659201,-0.2963603735,0.1139921099,0.0361340642,0.0092623783,-0.1929906905,0.0298596881,-0.0371094756,0.060285911,0.0510642454,0.112022981,0.4447426498,0.321939528,-0.2890278101,0.2662294507,0.0822312608,0.3286916614,0.0109019568,-0.0902361274,-0.0044672117,0.2150546759,0.071943149,0.0639618337,0.2853553891,-0.121838145,0.2022846043,-0.2137944251,0.1298509389,0.1342011094,0.2930413485,-0.1915011406,-0.1733280867,0.4962482452,0.0442023762,-0.1058535352,-0.0767576918,0.4734348357,-0.0099865813,0.7146952152,0.0455126911,0.7037382722,-0.1248852164,-0.1526597589,-0.2615527511,0.1158447191,0.0420728736,-0.64742589,0.3106974959,-0.1907190233,-0.2052680254,0.1221536845,-0.1048922762,0.1149579808,0.3697929382,0.2797479928,0.1120956093,0.2166129202,-0.2035370618,-0.2071072161,0.3400913775,0.1826967597,-0.2283104211,0.2803766727,0.0391107574,0.0797609836,-0.0306658018,0.0226754844,0.0615450889,-0.1580805331,0.3526025414,-0.1197993085,-0.0483023636,0.3265137672,0.2986385822,-0.2302096188,-0.0455086827,0.1209442392,-0.5679507256,0.227529496,-0.2574241161,-0.1487644911,0.1044511124,0.1428872943,-0.1538688093,-0.0604204983,-0.5270249248,-0.087980397,0.019898681,-0.3047129512,-0.1901982576,0.2688499689,-0.8522258997,-0.3888071477,0.0930700675,-0.0728898421,0.0655892715,-0.1263828427,0.5215659142,-0.1912914664,0.0358548164,-0.0476521328,0.2592445016,-0.4425133169,0.7070091963,-0.3910204768,-0.0085432446,-0.0221072882,0.2433186769,0.2357957363,-0.3659254909,0.0845108852,-0.0970671177,-0.1915614903,0.0398340262,-0.5570601821,0.0905365273,0.4957045913,-0.0826076344,0.028738426,-0.1235797703,0.0528359823,-0.1890909821,-0.1416686028,-0.0433611833,-0.6022043824,-0.1440006495,-0.2501831949,0.0947240666,0.2008906901,0.367053479,0.1365483999,0.1150724217,-0.3318484724,0.1516888291,-0.1088539064,-0.0053227358,0.1221120805,0.0959045589,-0.0204846244,-0.0927880555,0.1320218742,-0.081364654,-0.0900850594,0.1627068669,0.1668039262,-0.0654661506,0.2371114343,0.2261744887,0.312930733,0.1566813439,-0.1954315454,-0.2620935142,-0.1907634735,-0.2348840833,0.0444748513,0.0386032611,-0.2214903831,-0.0736925676,0.3863520324,-0.2685997784,0.1082161367,0.5082785487,0.4393986762,0.4513733685,-0.0027622064,0.5174337626,-0.2549319863,0.0965867415,0.1337032914,0.049999103,0.244960919,0.0504376478,0.0262856353,-0.0082346704,-0.2805000246,-0.0818451196,0.4896751642,0.6631741524,0.2678462863,-0.2227081656,-0.0256315451,0.0075579789,-0.0481432639,-0.0765162781,0.8391583562,-0.1111103594,0.1397631615,0.1743852496,0.3427317441,0.3477385938,-0.2052412033,0.0624891669,0.0630099773,0.3767134845,-0.1499240547,-0.6269664168,0.2316364944,0.0486394316,0.2675713599,-0.1535964161,0.1240843982,0.7165526152,-0.0639144182,0.478597492,0.0993159413,-0.0497429185,0.1411432475,-0.1296245009,0.4695599377,0.1944604516,0.3647834361,0.1691023409,0.1554943323,0.292483896,-0.0987268239,-0.5738973618,-0.1505660862,0.4674291015,-0.0120793283,-0.0454108976,0.4208975434,0.0418873727,-0.0121290572,0.0280934405,0.1323243827,-0.1338592172,0.194598794,-0.0022141966,0.1079651043,0.2100123614,-0.0201810841,0.4173367321,0.4710710943,-0.0277034529,-0.2999332845,-0.0784050971,-0.0858180374,0.1237093434,0.1651202887,-0.0054091667,0.1638244539,-0.0103433235,0.3054114282,-0.0628253818,0.137759909,0.0035015254,-0.0520890094,-0.3109496832,-0.1872943938,0.208296001,-0.0313969739,-0.0078559332,-0.5724868774,-0.1981349885,0.4705941677,-0.1560833156,-0.0974815786,0.0979627967,-0.0830199718,-0.0531942435,-0.0803229511,-0.0383732431,-0.2392767519,0.0729021654,-0.4864426851,-0.2075454146,0.134289518,-0.0296102781,-0.1673309505,0.3975626826,0.2318827659,0.5182492137,-0.1991002113,-0.1807960123,-0.2487403601,0.1420690268,-0.1421346515,0.3399437368,0.0479399897,0.5490871668,0.0789261013,-0.0603778102,0.6871289611,0.1367733777,0.1093953773,0.4636110067,-0.1878227592,0.3570115268,-0.3443448842,-0.2847066224,0.1339998692,-0.0340750664,0.4974749386,0.2011549324,-0.2478416562,0.2226835489,-0.2510336041,0.326099515,-0.4909696579,0.5764991641,0.0635076836,0.1297810674,-0.0798649713,0.0456422009,-0.1233183295,-0.2051599026,-0.3342983127,0.066215001,-0.386046946,-0.2056215405,-0.2810480893,-0.1591938436,0.0470137931,0.2368545681,-0.1396909654,-0.1990498006,-0.0603569411,0.0884797499,-0.1346513033,0.2249327749,-0.0805848315,0.1380124837,0.1908537745,0.2626543045,-0.2852629423,0.2808651328,0.2794204354,-0.1816107333,-0.0213588271,-0.4958341122,0.1355202645,0.4050750136,0.1090613678,-0.5247439742,-0.5293174982,0.1268747747,0.33361727,0.2443785518,0.1576969028,-0.0107488101,-0.3092321157,-0.2780920565,-0.3314505816,0.0037303271,0.1973460764,0.0342447534,-0.0339604467]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1965","title":"Can we parallelized the add_faiss_index process over dataset shards ?","comments":"Actually, you are right. I also had the same idea. I am trying this in the context of end-ton-end retrieval training in RAG. So far I have parallelized the embedding re-computation within the training loop by using datasets shards. \r\n\r\nThen I was thinking of can I calculate the indexes for each shard and combined them with **concatenate** before I save.","body":"I am thinking of making the **add_faiss_index** process faster. What if we run the add_faiss_index process on separate dataset shards and then combine them before (dataset.concatenate) saving the faiss.index file ?\r\n\r\nI feel theoretically this will reduce the accuracy of retrieval since it affects the indexing process.\r\n\r\n@lhoestq\r\n","comment_length":60,"text":"Can we parallelized the add_faiss_index process over dataset shards ?\nI am thinking of making the **add_faiss_index** process faster. What if we run the add_faiss_index process on separate dataset shards and then combine them before (dataset.concatenate) saving the faiss.index file ?\r\n\r\nI feel theoretically this will reduce the accuracy of retrieval since it affects the indexing process.\r\n\r\n@lhoestq\r\n\nActually, you are right. I also had the same idea. I am trying this in the context of end-ton-end retrieval training in RAG. So far I have parallelized the embedding re-computation within the training loop by using datasets shards. \r\n\r\nThen I was thinking of can I calculate the indexes for each shard and combined them with **concatenate** before I save.","embeddings":[-0.2849661112,-0.1696963161,-0.091261059,0.1680023968,-0.3235114217,0.3758886158,0.3347997963,0.0697201118,-0.0358868092,0.1546950489,-0.0876063108,0.0618726648,0.3824176192,-0.0200762153,-0.3032027185,0.1434956789,0.2809820771,-0.0584644936,0.2126163095,0.0394795276,-0.3338530064,-0.1382644624,-0.0715480521,-0.2807034552,-0.2238429785,0.1862551868,-0.3562583029,0.1115740836,-0.0008239155,-0.3895473778,-0.1900932938,0.3138128817,0.2559099495,0.0299698468,-0.0001277621,-0.0997196585,0.1779816151,0.0183813926,0.1161235645,0.6891212463,-0.2972025275,-0.026710866,-0.0404046066,-0.2445941269,0.0822643191,-0.2505580485,0.0011064857,-0.25172925,-0.0702404454,-0.2690679133,-0.0466796048,-0.0607918501,0.0575929023,0.0574581996,0.0406849682,-0.3079018891,-0.1213591099,0.1398369372,0.0839523226,0.2245052606,0.1343079507,0.0507928059,-0.0477382094,-0.1600026488,0.0726076812,0.1129783392,0.2452798784,0.1347260922,-0.0512386933,-0.1212150753,0.2552976012,-0.099278748,-0.5349696279,-0.1582330018,0.2663683295,-0.1595634073,-0.1438806504,-0.1659998596,0.104799062,0.0622216016,0.0662736371,-0.5065560341,-0.0104388818,0.0325589664,0.4056906998,0.0920149088,0.2955592871,0.1298534572,0.3458766937,0.0445572883,0.0997549519,-0.2029110193,-0.1997184157,0.1131251976,-0.6591157317,-0.0331773572,-0.2063943744,-0.2228663415,-0.1241202652,-0.0351017751,-0.1060766876,0.1871549338,0.2389025986,0.1609723717,-0.005884049,-0.0895845294,0.03438434,-0.1298580468,-0.1368224323,-0.4281221032,-0.1075878367,0.0488347597,0.0350251943,-0.1958869249,-0.5263495445,-0.0633085445,-0.4865528941,-0.0719926059,-0.102642633,-0.173439756,-0.1677077711,-0.1291918904,0.1297932118,0.0206174087,0.282286346,-0.1592240632,-0.170183748,0.0673740134,-0.0559870228,0.0045553483,-0.0279973727,-0.2161835879,0.0174257401,0.512678206,-0.0436444283,-0.2863789201,-0.3827866912,0.0505875498,-0.0481587462,0.1748678684,0.066413708,-0.2154343724,0.120924972,0.0772427544,-0.1290807575,-0.167744875,-0.0365751684,0.2454080284,-0.1966933906,0.2098246813,-0.2311400771,-0.2099299133,0.1646127403,-0.0386470743,-0.034817595,-0.1441643238,-0.0446498059,0.3352725506,0.3445595801,-0.0064366502,0.0062948894,0.1591218412,-0.28740412,-0.1396418661,0.2494934797,0.2338727713,-0.4014918208,-0.1169342399,-0.0487821698,0.0194945522,0.1672157645,0.5547193885,-0.1314323097,0.1645195782,-0.3813923597,0.5477818251,0.3399351835,-0.2826054096,-0.0182866473,0.0328421071,-0.1953346729,-0.096134454,0.263723284,0.1584376693,0.4460000992,0.0293704383,0.2704944611,0.2283853441,-0.066930443,-0.3063270748,-0.3454902768,-0.2478902787,0.2212375849,0.1413300633,-0.0717564598,0.1985466033,0.0511377342,-0.8309345841,0.3635245264,-0.0840666741,0.215919137,-0.0107421456,0.2092398107,0.3811731339,0.3770609498,-0.078118287,0.0771491155,0.1599858254,-0.0280741211,-0.0409755819,0.2488015443,-0.3538302481,0.4831629992,-0.0756302774,0.0241911225,0.4689189196,-0.1199094728,-0.2062516809,0.0621985607,-0.3314244151,-0.294170469,0.5498024225,-0.477976352,-0.0968439877,-0.1887799203,0.2802512944,0.0779903233,-0.1488750726,-0.2816527486,0.0475954078,0.0733578354,-0.2442056537,-0.0596472137,0.2386415154,-0.0809267387,0.1019926742,0.8778350949,0.1095690578,-0.1541573256,0.0711275265,0.1545335948,-0.3248102963,-0.0852531046,-0.2058341205,0.3224862218,0.2268131673,0.0249040499,0.3609706461,0.0308591612,-0.0908334851,0.0195496716,0.0234930795,0.0144608794,0.0800584033,-0.1024847478,0.0013242186,-0.0705540776,0.0450991653,-0.1120695174,0.3265722394,-0.1385033727,-0.1785280257,-0.2923904955,-0.0533653907,-0.0340082347,0.1097715273,-0.0718640909,-0.1530020833,0.3807485402,0.1627696007,0.1704768687,-0.12954624,0.2641075552,-0.194244653,0.024381429,0.2502625883,0.0645522252,0.2118131965,0.0696620122,-0.0877100155,-0.0326046273,-0.0260461178,0.0928725675,0.0639977083,-0.0158800166,-0.0416265838,0.0864127874,0.3875234723,-0.3104661107,-0.1104385033,-0.2621289492,-0.1685064137,0.1309842467,0.0351331495,-0.1759462953,-0.0140682012,-0.0472524725,0.5214400291,-0.1426884234,-0.3105710149,-0.1479194611,-0.1127294376,0.1196701005,-0.1296887845,0.2577828169,0.1593421251,0.1317600906,0.1190094948,-0.0478240699,-0.2820173502,0.0005472153,-0.170217514,0.0225036182,-0.2009871006,-0.3025679886,-0.2304276675,-0.3104281127,-0.1213891357,0.069540143,0.1559575647,-0.0151106054,-0.2252917439,0.1130053252,0.0651460513,0.1385291666,-0.1371595562,-0.3187462986,-0.2540183365,0.3161889613,-0.1976729631,-0.0086671589,-0.4010066092,-0.1311909854,-0.1447111815,-0.1944415271,-0.2484068274,-0.3096928,-0.1726858318,0.1525397152,-0.0263112243,-0.0767534673,-0.0058424822,-0.0332636461,-0.1281506121,0.6561332345,-0.0107417135,0.2173006535,0.0476656184,0.0933474228,-0.2506474555,0.2637120485,0.2129894942,-0.0281591471,0.0079064956,0.1556269079,0.2947208881,0.3695460558,0.0974454731,0.0661053285,0.0408578329,0.0288635343,-0.1932774633,0.9015139937,0.2490787357,-0.1850935221,0.1608608216,0.2594344914,0.1495266557,0.007931333,-0.1160737947,0.2399917841,0.1435940266,-0.1371480823,0.0009889953,-0.036458835,-0.0945904404,0.109618552,-0.2012953311,0.011533495,-0.1715347767,-0.0033563175,0.0544862412,0.5516700149,0.0550488904,0.0214142855,-0.202568233,0.0437315516,0.1283036619,0.3945729434,0.5861868262,-0.2756174505,-0.0562679581,-0.2580190003,-0.301384449,0.4053607583,0.0661222339,0.3546912074,0.1334126294,-0.1077653989,0.334064275,-0.1155369654,0.8974553347,-0.1313093305,-0.5274403095,0.1343751997,0.0533279628,-0.1458674073,-0.2606739104,0.3020035923,0.1860320419,0.3533156514,0.1942872107,-0.2180707306,-0.2082677037,0.3316380382,0.1343443394,-0.2656055391,-0.6529963017,-0.0941174179,0.1721439511,-0.047479853,0.2200227082,-0.3541423082,0.0357384682,0.0266461316,-0.0205887817,-0.3798432946,-0.0183668826,0.0105702942,-0.041842144,0.0875428692,0.0219420753,0.5158404708,0.3603673577,-0.0402309038,0.4263272583,0.0963509381,0.0793370754,0.0130316969,-0.1868700981,-0.1968352348,0.3067002892,0.09572386,0.0654582903,0.3111416399,-0.2788876593,0.0630096123,-0.3485806584,0.0565940104,0.1023266464,0.4389446676,-0.1225831285,-0.1930823475,0.4389567971,0.0504131131,-0.1514481455,-0.054530561,0.3455350101,-0.1894798875,0.6812269688,0.0664830357,0.8806685209,-0.1594643295,0.0483082756,-0.2123780847,0.0975702479,-0.0178218503,-0.6159808636,0.4225203097,-0.2377068102,-0.177193746,0.1049716994,-0.2092422694,0.2017694861,0.4122804403,0.1896034628,0.1242855638,-0.0764308721,-0.0992064327,-0.1765964925,0.2509697974,0.2230383754,-0.2394242287,0.3446897864,-0.0590552352,0.2382807434,-0.2323361337,0.0191419534,0.0264445804,-0.1703361422,0.3796364963,-0.1086358652,-0.0167274345,0.4481118321,0.1966137439,-0.173183158,-0.2193167657,0.2574957609,-0.3695749044,0.3011191189,-0.2077327222,-0.0317814127,-0.0070875273,-0.0048743915,-0.2725214362,-0.1024235412,-0.3807839155,-0.0331248529,0.1110463515,-0.3711657524,-0.1661395878,0.2631865442,-0.7294945717,-0.4386252761,0.1284064204,-0.0531713255,-0.1192603633,-0.1365399957,0.4132653177,-0.1144068688,0.0019169003,-0.068925418,0.2600788474,-0.4817506373,0.8244270682,-0.3634019792,0.0254209768,0.0193003472,0.288120985,0.3121020794,-0.257922262,0.2277693748,-0.1517081112,-0.2348931283,-0.0000379096,-0.5809662342,0.1541986167,0.4484098256,-0.0926689133,0.0182478391,-0.0281311758,-0.0096793724,-0.4759551585,-0.0965823978,0.0260910131,-0.571609199,-0.1279105097,-0.3554806113,0.1013093591,0.2063253373,0.4299412966,0.0584556758,0.1112203822,-0.3822533786,0.1869219244,-0.1091271564,0.0534862913,0.1014933512,0.2128516585,0.1496154368,-0.0973928347,0.1051631942,-0.0466013588,-0.0639348552,0.1806935817,0.0985817388,-0.0267344322,0.2877898216,0.2327820063,0.1426940113,0.1709017158,-0.0853703693,-0.3187400103,-0.2249472439,-0.2446722239,-0.0246261209,-0.0533241443,-0.1381592005,-0.0823489577,0.4531776607,-0.0884272009,0.037072707,0.4527213573,0.3983288109,0.3977431655,-0.0550011061,0.4254107475,-0.2170275897,0.0043881186,0.0188513771,0.1060654819,0.3288202882,0.0073814928,0.1000651568,-0.0175165646,-0.3024097085,-0.2244684845,0.5101447701,0.6108431816,0.3429232538,-0.4377302229,0.1115001515,-0.0158237685,-0.0398659073,-0.1532558203,0.7825090289,0.0298473891,0.1465705782,0.1259584576,0.2123567611,0.2996015251,-0.3592546582,-0.0047614472,0.1406220049,0.3161543608,-0.1753818393,-0.4867200255,0.3327390552,0.0838660821,0.155430004,-0.1178853139,0.1441015601,0.6216912866,-0.1454122812,0.5556603074,0.0931106284,0.0085597346,-0.0513991788,-0.3147478104,0.3127994239,0.415858835,0.347841531,0.1352378726,0.1462415457,0.3830262423,0.2558977306,-0.5684161186,-0.1476950645,0.4132967889,-0.1333668232,-0.1575539857,0.3834903538,-0.0540318079,0.1045505479,0.0788135454,0.1858241856,0.0929143503,0.37519297,-0.0736511648,0.243272543,0.2592485249,0.0974979699,0.3380092382,0.2684299052,-0.1351719946,-0.3136456013,-0.1853096932,0.0345402472,0.123591505,0.0863995329,0.0447233394,0.1819353253,0.1531486958,0.2662109733,-0.0655696541,0.1450531185,-0.0084780045,-0.0248160884,-0.2228887081,-0.1458398998,0.1872137785,-0.0870501995,0.0245117657,-0.4995052814,-0.1176992729,0.3334580958,-0.2195058763,-0.1744241863,0.2647448778,-0.3016046584,0.0676541626,-0.0331206918,-0.1562538743,-0.0631976277,0.2823480666,-0.3615163565,-0.2766749263,0.2801628709,-0.0707580671,-0.308637321,0.4806895852,0.3341833651,0.5037377477,-0.2653267384,-0.1025482416,-0.2228054106,0.197182402,-0.1594050378,0.2911324799,-0.0202686638,0.4730789065,0.1639253199,-0.1333218366,0.5242284536,0.0366239734,0.0648771748,0.5568504333,-0.251976192,0.1873479635,-0.2975609004,-0.3276863992,0.0573023558,-0.1519962698,0.4205698073,0.1887063384,-0.2505378425,0.1852189898,-0.0281145778,0.3232847154,-0.2908660173,0.529887259,-0.0202975385,0.2953534424,-0.0530069396,0.0754800439,-0.1315184236,-0.1362611949,-0.3146404028,0.0852033719,-0.2904397547,-0.0889270455,-0.4956982732,-0.1259976327,-0.026905911,0.1326652765,-0.0939806253,-0.1971268952,-0.0234793387,0.1047436595,-0.0522619709,0.2000953555,0.0460236967,0.237413168,0.1900678426,0.2453595549,-0.260297507,0.3165750504,0.2897218466,-0.1897237897,-0.0601596236,-0.3615053892,0.0874474645,0.3728887141,0.1571476609,-0.5834983587,-0.4987608194,0.2187902927,0.2269461453,0.1632259339,0.2620422244,0.3271625936,-0.2740887105,-0.2698299587,-0.5132451057,-0.0430748798,0.0780402869,-0.1016156524,-0.2167233825]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1965","title":"Can we parallelized the add_faiss_index process over dataset shards ?","comments":"@lhoestq As you mentioned faiss is already using multiprocessing. I tried to do the add_index with faiss for a dataset object inside a RAY actor and the process became very slow... if fact it takes so much time. It is because a ray actor comes with a single CPU core unless we assign it more. I also tried assigning more cores but still running add_index in the main process is very fast. ","body":"I am thinking of making the **add_faiss_index** process faster. What if we run the add_faiss_index process on separate dataset shards and then combine them before (dataset.concatenate) saving the faiss.index file ?\r\n\r\nI feel theoretically this will reduce the accuracy of retrieval since it affects the indexing process.\r\n\r\n@lhoestq\r\n","comment_length":72,"text":"Can we parallelized the add_faiss_index process over dataset shards ?\nI am thinking of making the **add_faiss_index** process faster. What if we run the add_faiss_index process on separate dataset shards and then combine them before (dataset.concatenate) saving the faiss.index file ?\r\n\r\nI feel theoretically this will reduce the accuracy of retrieval since it affects the indexing process.\r\n\r\n@lhoestq\r\n\n@lhoestq As you mentioned faiss is already using multiprocessing. I tried to do the add_index with faiss for a dataset object inside a RAY actor and the process became very slow... if fact it takes so much time. It is because a ray actor comes with a single CPU core unless we assign it more. I also tried assigning more cores but still running add_index in the main process is very fast. ","embeddings":[-0.4054992199,-0.1552561522,-0.1257984191,0.1830277443,-0.3051235974,0.2511138618,0.3391765058,0.0727008656,0.1347900778,0.1768684834,-0.1435732096,0.1960875839,0.3504758775,0.087931864,-0.2222462595,0.0771596283,0.3349466622,-0.0737509206,0.187864691,0.1442663968,-0.349216491,-0.0790117085,-0.0737032816,-0.3046618104,-0.2064707577,0.2073306143,-0.2225418389,0.0050883587,0.1136128679,-0.3396568298,-0.2113557309,0.2969415784,0.2204650193,0.2022595555,-0.0001236163,-0.0930320546,0.2447298169,0.1353631169,0.2126867473,0.7094957232,-0.2860053182,0.0842965171,0.0118541978,-0.2131094933,0.0391235761,-0.2245900482,0.0313989259,-0.307661593,-0.1945788264,-0.2164341211,0.0178864188,-0.0420150645,0.1275662631,0.0397086106,-0.0555945672,-0.2334746569,-0.1960607618,-0.1421741694,0.1881549358,0.4077614844,0.0928777382,0.1631823927,-0.0815118551,-0.1110181659,0.0442543626,0.0191483367,0.3126096725,0.1118710637,0.062026076,0.0274287108,0.1938612312,-0.1418400705,-0.5114311576,-0.181544885,0.1477668583,-0.1431189626,-0.1886232793,-0.1851349026,0.0634021312,0.0095758699,0.1191961393,-0.4742821753,0.0137948692,-0.0145699009,0.3781154752,0.1848355532,0.2552728653,0.1527383924,0.4181759655,0.0640506521,0.138956666,-0.0919338986,-0.2201531678,0.0968007743,-0.7896094918,0.0617333613,-0.1457816064,-0.1499472708,-0.0559659041,-0.1912728548,-0.3814213872,0.2947395742,0.1131915525,0.1418474019,0.0255720615,-0.1043267548,0.0632383451,-0.2763235569,-0.0681233406,-0.3750481009,-0.0584922396,0.0647536963,0.0459298603,-0.1101507246,-0.4853464365,-0.0342457332,-0.3445420265,-0.114913255,-0.0396998227,-0.2562002838,0.1214185879,-0.1795289814,0.2044482976,0.1827087104,0.322575748,0.1010054201,-0.2328573614,-0.0010052047,-0.0215399563,-0.0223837886,0.0048326245,-0.1678569168,0.1890094876,0.4930638075,-0.0246064309,-0.4299058616,-0.3655105829,0.0114615774,-0.0571343116,0.1619334072,0.021370085,-0.2168648839,-0.0300495476,0.0381839201,-0.0198899452,-0.0655273199,-0.1344530284,0.272883445,-0.2350448072,0.2151949555,-0.2092870325,-0.3914920688,0.0174666848,-0.0052477024,-0.0216319263,-0.1631601304,-0.0698066801,0.3777594566,0.1984350383,0.0365074091,-0.0853618681,0.1062015072,-0.2503991127,-0.1264425665,0.2313388139,0.1742165983,-0.4121520817,-0.0412243307,-0.3235965967,-0.1685480475,0.1864796281,0.4758313,-0.0292444862,0.1094245315,-0.3331192434,0.3905993998,0.1770891696,-0.3672056198,0.0471229404,0.1398416907,-0.1204541624,-0.1656452864,0.331510365,0.2888590991,0.3453750014,-0.0083529437,0.135558933,0.146589607,-0.0612415597,-0.1479458809,-0.2127332389,-0.2535600662,0.1667654812,0.2950843871,-0.0209490117,0.1681788713,-0.0310331155,-0.8039103746,0.3160693944,-0.1459510028,0.299919188,-0.06772241,0.2666429579,0.3727966249,0.370464921,-0.116235286,0.0769765824,0.2382863611,0.0209097359,-0.0338546485,0.2227012068,-0.3855346441,0.4876060188,0.0459406562,0.0490641929,0.4221509099,-0.0552332364,-0.1752128899,0.0095422193,-0.41629529,-0.3488960266,0.5179666281,-0.2398486584,-0.184675619,-0.069581449,0.0604914576,0.0085241962,-0.1463850737,-0.3137203157,0.0875518098,0.0763883144,-0.2518971562,-0.0870734379,0.3075044453,-0.0211861804,0.0460121855,0.8042544127,0.0774966776,-0.1413669437,0.2354428619,0.0844637901,-0.2908227742,-0.0603009388,-0.1021532938,0.34709391,0.1247011423,0.1034974009,0.4663467109,0.0484497175,-0.0102870204,0.2089307904,0.0494257584,0.1752787828,0.0404649675,-0.0039793467,-0.1105219722,-0.0594568178,0.0742544904,-0.2332384884,0.5649155378,-0.1717336476,-0.1027774215,-0.2015347779,-0.0461472236,-0.0802179947,0.1478333473,0.0841501802,-0.0952415615,0.2414818108,0.2003468275,0.030859204,-0.2746875286,0.2501146495,-0.1929262578,0.0255797803,0.2707829773,0.1545015574,0.0391260535,-0.0348359458,-0.1922845691,0.0270986166,-0.1999224126,0.0515194125,0.1211284548,-0.0166958738,-0.1093714535,0.0189686194,0.1860831827,-0.2127860337,-0.1081467196,-0.2143588066,-0.1133390293,0.1764407903,0.0975932479,-0.1917555779,-0.0609497502,-0.0961273015,0.3963818252,-0.1553369015,-0.452817291,-0.2940588295,-0.1470527947,0.1323347986,-0.0661902651,0.3142543733,0.085630998,0.0921736434,0.1264203638,-0.0404543169,-0.3440809846,0.0950988755,-0.1518468261,-0.0034035116,-0.0941034257,-0.2936231494,-0.3489646316,-0.1819126606,-0.0824747607,0.0411472283,0.0189424548,0.0167009961,-0.1476113498,-0.1336666495,-0.0711159408,0.0677145123,-0.1171513498,-0.2093511075,-0.3055463433,0.3475598097,-0.1158814728,0.1410505772,-0.4759967029,-0.1276790351,-0.0629486889,-0.3044448197,-0.1956475377,-0.2405064851,-0.1052337587,0.1640596539,0.0499347784,-0.1538969874,0.0957066119,0.0029419172,-0.0289327875,0.4696128666,0.0203414857,0.0658014864,-0.0760877952,0.1825042069,-0.1649398953,0.1721914709,0.3239159584,0.0056635202,-0.0553413182,0.1192137152,0.1172983274,0.336915493,0.0030350529,0.0329294503,0.0064069736,0.0400705561,-0.2011463642,0.7294507623,0.1223027408,-0.1957218647,0.1946378648,0.0824977532,0.0250118226,0.0861675069,-0.1782761961,0.350084126,0.1012573242,-0.0929720849,0.0026172893,-0.0438820645,-0.1264021397,0.0411880501,-0.1293791234,-0.1038943976,-0.1315627396,-0.0548463091,0.2528621554,0.5113689899,-0.0436335728,0.0077443914,-0.2366498262,-0.0065447115,0.1355667263,0.4619989097,0.5420896411,-0.3316129744,0.0905424207,-0.3064237535,-0.3880170882,0.4490930736,0.0378731377,0.4197951555,0.0841399059,0.0210977551,0.2919827104,-0.2219098657,0.9026585221,0.0270747934,-0.5150843263,0.0960619673,-0.0259989202,-0.3318488598,-0.0938113481,0.1524903327,0.2472437471,0.3681126833,0.1548551023,-0.2120522857,-0.2917934656,0.134326905,0.0831084549,-0.2784323394,-0.673858881,-0.1498617381,-0.0077256532,-0.0534089357,0.1820505559,-0.2881641686,0.1141936406,0.0948083401,-0.0059594642,-0.3301713765,-0.0232590493,0.0441849045,0.0040746555,0.0915082023,-0.0205349531,0.5346843004,0.2204899043,-0.0735487342,0.2417609245,-0.0164079517,0.222472325,0.1455619037,-0.1792441308,-0.1392751634,0.2092137039,0.0098430971,-0.0030311467,0.3491419852,-0.2979663312,0.1423682719,-0.2083738446,-0.0092653017,0.0615616925,0.4385180175,-0.0940528214,-0.284619242,0.3209881186,0.121175617,-0.1498609185,-0.1145359948,0.3715471029,-0.050134588,0.6364876628,0.1283536106,0.6783912182,-0.2733451724,-0.0964757726,-0.175793305,0.1353349537,0.0775397271,-0.5695595145,0.4079462588,-0.2618125975,-0.1729674339,0.1151987016,-0.1011656225,0.1822108328,0.267760396,0.0997643918,0.2494496703,0.0900088474,-0.2832883894,-0.1383249015,0.3152263761,0.1780462712,-0.3854479194,0.3567737043,0.0085209711,0.1185498536,-0.2574541271,0.0475545526,0.0995272547,-0.0964791551,0.4011670053,-0.1618818343,-0.062256705,0.4725167751,0.0784942508,-0.1750519425,0.0149901463,0.22652179,-0.4040831029,0.2418207973,-0.2935242653,-0.1323635578,0.0944526941,-0.0516434088,-0.1664473414,0.001274345,-0.4851131439,-0.0742492974,0.0507860705,-0.4548889101,-0.1688657254,0.275044769,-0.7210444808,-0.3166680336,0.0968268365,-0.0611467212,0.1004588455,-0.145427987,0.3827253878,-0.1340805739,-0.0466316603,-0.0264505725,0.2520075142,-0.4767291546,0.8772199154,-0.3962702751,-0.0234917086,0.0064207879,0.2088797987,0.303753823,-0.1857493818,0.0359900855,0.0318611972,-0.1825810224,-0.029721532,-0.7237133384,-0.0615831278,0.3477555811,-0.1356629729,0.0826942697,-0.2126896381,0.0430259742,-0.5923836231,-0.0986189321,-0.0355291478,-0.6348736286,-0.0941628739,-0.3619499803,0.087631464,0.1204520985,0.3645608723,0.1034472287,0.103367053,-0.347306639,0.2193742543,-0.1451957971,-0.0548931248,0.2107295841,0.1804763526,0.2133192718,-0.0586997867,0.1368571222,-0.0641082302,-0.0119606042,0.1239686534,0.1416327953,-0.0910775959,0.2457419187,0.2072181404,0.0595082603,0.0605388694,-0.1134945378,-0.2933693826,-0.2580486536,-0.227774173,0.0708859488,-0.0550412796,-0.2842026651,-0.0373875611,0.379673928,-0.2199344784,-0.0367660522,0.5711757541,0.3638447523,0.4119433761,-0.1304623336,0.5840527415,-0.1503132433,-0.0116172684,0.1101325825,0.0380049981,0.3578718305,0.1273033619,0.0257705171,0.1005515754,-0.2617841065,-0.2099003047,0.5314099789,0.5130409002,0.4117487371,-0.2642948627,0.0969002917,0.043632552,-0.0635161027,-0.1413562894,0.7605375648,-0.0652896017,0.2589676082,0.2289000452,0.3385686874,0.3246192038,-0.2439430058,0.0713546351,0.1840886474,0.3953973651,-0.2592020035,-0.5113334656,0.3347240686,0.0319291949,0.2751269639,-0.1516727954,0.1107717678,0.9230399728,-0.1979516447,0.5783293247,0.2697029114,0.0731895715,0.0477504432,-0.1127757952,0.5054671764,0.2119088769,0.2594753802,0.042558264,0.1568236798,0.4555436671,0.0862915739,-0.4791406095,-0.2030679584,0.363342762,-0.0651321039,-0.1690251082,0.3964122236,0.0012898264,0.0932264552,0.0458046272,0.2178767025,0.035757497,0.3246455789,0.049106013,0.1516363174,0.1504167765,0.0357428677,0.4240095913,0.2561796904,0.0196467973,-0.2027694136,-0.1023056135,-0.0331467167,0.1279642284,0.1681079268,0.0743427649,0.1605702341,0.0622821189,0.3273927569,-0.0757857114,0.1349636614,-0.0273566321,-0.0791054219,-0.3827724457,-0.0394778177,0.1916842312,-0.0211960338,-0.0130573437,-0.487477988,-0.2010089308,0.3620595038,-0.0939692482,-0.0635306239,0.1349419057,-0.3062334955,0.003493092,0.0126204826,-0.1436277479,-0.1309952289,0.3259009421,-0.5958315134,-0.2522018552,0.1412502527,-0.0217811204,-0.2656868398,0.4139502943,0.0934202373,0.5814159513,-0.2157100588,-0.0181516539,-0.2982735038,0.1779899597,-0.1417060494,0.0568242259,-0.0309726652,0.4734140038,-0.0127442768,-0.1003621668,0.640220046,0.0816157982,0.2973897755,0.5808256269,-0.2128428668,0.1888262779,-0.2031282634,-0.2613909543,0.1296497732,-0.1679709107,0.4393238425,0.1654769927,-0.1756121963,0.1502399743,-0.1133836731,0.4254377782,-0.4664508104,0.4777955711,0.0649799854,0.2579562664,-0.1346855015,0.1684644818,-0.132851854,-0.1170493886,-0.3317901492,-0.0999454781,-0.254770726,-0.1374467164,-0.324665904,-0.1422874033,-0.0342454948,0.3217967749,-0.1579924226,-0.0471627563,-0.1276910901,0.1993934959,-0.1976647377,0.1830365211,-0.0548234545,0.1600778401,0.207494244,0.1919039339,-0.280105412,0.2080818564,0.2239714116,-0.2235782743,-0.0168869123,-0.4630414546,0.1518833637,0.3087670207,0.1494837552,-0.436986655,-0.440608412,0.0744832158,0.2929986417,0.1704966277,0.1940241009,0.0582559034,-0.2841024995,-0.2611771822,-0.5086881518,0.0182645489,0.1092121378,-0.0496230945,-0.0873969868]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1964","title":"Datasets.py function load_dataset does not match squad dataset","comments":"Hi !\r\n\r\nTo fix 1, an you try to run this code ?\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nload_dataset(\"squad\", download_mode=\"force_redownload\")\r\n```\r\nMaybe the file your downloaded was corrupted, in this case redownloading this way should fix your issue 1.\r\n\r\nRegarding your 2nd point, you're right that loading the raw json this way doesn't give you a dataset with the column \"context\", \"question\" and \"answers\". Indeed the squad format is a very nested format so you have to preprocess the data. You can do it this way:\r\n```python\r\ndef process_squad(examples):\r\n \"\"\"\r\n Process a dataset in the squad format with columns \"title\" and \"paragraphs\"\r\n to return the dataset with columns \"context\", \"question\" and \"answers\".\r\n \"\"\"\r\n out = {\"context\": [], \"question\": [], \"answers\":[]} \r\n for paragraphs in examples[\"paragraphs\"]: \r\n for paragraph in paragraphs: \r\n for qa in paragraph[\"qas\"]: \r\n answers = [{\"answer_start\": answer[\"answer_start\"], \"text\": answer[\"text\"].strip()} for answer in qa[\"answers\"]] \r\n out[\"context\"].append(paragraph[\"context\"].strip()) \r\n out[\"question\"].append(qa[\"question\"].strip()) \r\n out[\"answers\"].append(answers) \r\n return out\r\n\r\ndatasets = load_dataset(extension, data_files=data_files, field=\"data\")\r\ncolumn_names = datasets[\"train\"].column_names\r\n\r\nif set(column_names) == {\"title\", \"paragraphs\"}:\r\n datasets = datasets.map(process_squad, batched=True, remove_columns=column_names)\r\n```\r\n\r\nHope that helps :)","body":"### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!","comment_length":170,"text":"Datasets.py function load_dataset does not match squad dataset\n### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!\nHi !\r\n\r\nTo fix 1, an you try to run this code ?\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nload_dataset(\"squad\", download_mode=\"force_redownload\")\r\n```\r\nMaybe the file your downloaded was corrupted, in this case redownloading this way should fix your issue 1.\r\n\r\nRegarding your 2nd point, you're right that loading the raw json this way doesn't give you a dataset with the column \"context\", \"question\" and \"answers\". Indeed the squad format is a very nested format so you have to preprocess the data. You can do it this way:\r\n```python\r\ndef process_squad(examples):\r\n \"\"\"\r\n Process a dataset in the squad format with columns \"title\" and \"paragraphs\"\r\n to return the dataset with columns \"context\", \"question\" and \"answers\".\r\n \"\"\"\r\n out = {\"context\": [], \"question\": [], \"answers\":[]} \r\n for paragraphs in examples[\"paragraphs\"]: \r\n for paragraph in paragraphs: \r\n for qa in paragraph[\"qas\"]: \r\n answers = [{\"answer_start\": answer[\"answer_start\"], \"text\": answer[\"text\"].strip()} for answer in qa[\"answers\"]] \r\n out[\"context\"].append(paragraph[\"context\"].strip()) \r\n out[\"question\"].append(qa[\"question\"].strip()) \r\n out[\"answers\"].append(answers) \r\n return out\r\n\r\ndatasets = load_dataset(extension, data_files=data_files, field=\"data\")\r\ncolumn_names = datasets[\"train\"].column_names\r\n\r\nif set(column_names) == {\"title\", \"paragraphs\"}:\r\n datasets = datasets.map(process_squad, batched=True, remove_columns=column_names)\r\n```\r\n\r\nHope that helps :)","embeddings":[-0.3537200987,0.0625078902,0.0333654545,0.3963988423,0.5554530621,0.0057210135,0.5457583666,0.3404688239,-0.0983101875,-0.1216232106,-0.1650660336,0.4508374333,0.1084147692,-0.0774192885,0.1859294176,0.2271618843,-0.1242239624,-0.0123368138,-0.1461081356,-0.2893434763,-0.1109781116,0.1282911897,-0.0823139772,0.2603987455,-0.1788990498,0.0166908093,0.2006343901,0.2723646462,-0.0037961667,-0.2199688405,0.3428422213,-0.2606053054,0.2361012548,0.5758861899,-0.0001148581,0.1132558137,-0.0115434257,-0.3397289217,-0.5327045321,-0.2002218962,-0.1021837071,-0.3757722974,0.1486449242,-0.2605553567,-0.1595326066,0.093737483,0.0203164779,-0.4645369053,0.3544548452,0.4883137345,0.213982597,0.2522563636,-0.1001010016,0.0775510594,0.1142548099,-0.0009760508,-0.06950894,0.1360990405,0.3347509801,-0.0582686849,0.0025541722,0.1724455059,0.0828086361,0.0568236373,0.2703095675,0.0064037591,0.199115023,-0.316155225,0.0564422533,0.3177836239,0.4627765119,-0.4630158544,-0.334705174,-0.2631344795,0.133104369,-0.2565729618,0.1320165247,0.0672778934,0.0511586368,-0.0424272642,0.0477764271,0.0376646332,0.1114287302,0.0368287303,0.0068916786,0.1999943703,0.0288672037,0.2383961529,-0.201539591,-0.1135483086,0.1148922518,-0.083149083,0.0818328112,0.392493546,-0.828517139,-0.0814113691,-0.2042951286,-0.1421720982,0.2562874556,0.0819008574,0.0386924408,-0.0269711483,0.148668915,0.1044574156,0.1687794775,0.3692035675,0.3052028716,0.2674016058,-0.0663638264,0.0668205246,-0.3717198968,0.2254103124,-0.3225816488,-0.1618057936,0.0960089266,0.0835877731,0.0062298947,-0.1927381605,-0.5517169237,0.1761802584,0.1544087082,0.0026421051,0.0974367931,0.324791342,0.0759165287,0.3802291751,0.0263226684,0.1048553437,-0.294623524,-0.3704161942,-0.2408784777,0.0330561139,-0.0880101547,0.0268361289,0.0183388889,-0.0356007777,0.3227834404,-0.1957616955,0.2114947587,-0.1790799946,0.1150045693,-0.2556711733,-0.1285898685,-0.0065955352,0.0298101194,0.0322048925,0.1742971241,-0.3730719388,-0.0482525341,0.1222925037,-0.3712684512,-0.0336521827,-0.1820591837,0.1871163845,-0.1391085088,0.0208789185,-0.2373687923,0.1698867232,0.3073554337,-0.1878401935,0.1063959599,-0.3961229026,-0.1664000899,-0.2122270316,0.3151646554,0.3882876039,-0.4074572027,-0.2117103338,0.1249395758,-0.0952139571,-0.0381632037,0.1216655001,-0.1516055316,0.1555615067,-0.20622392,-0.0070253811,0.7145847082,-0.8443276286,-0.4243383706,0.1099728644,-0.3059237599,-0.0194989499,-0.0710134581,0.066407755,0.1930091679,0.0147336535,0.0827305093,0.5066136718,-0.0362928361,0.0914031267,-0.1724580824,-0.2712134719,0.2526653707,0.1708515584,0.0318142921,-0.0380649008,0.0388377123,0.4989839792,0.332213819,0.1426163316,0.1112044156,-0.0299585946,-0.0359400138,-0.0821635574,0.114605628,-0.1481092274,-0.6798553467,0.1638657302,-0.1644516885,0.0439888574,0.2410787791,-0.1508302242,-0.2135134637,-0.0917369127,-0.4252310395,-0.1023878902,0.0743926764,0.1495984495,-0.0585200973,-0.1226450801,-0.159884721,0.3514788449,-0.2154513896,0.2284579277,-0.4849060774,-0.0965390429,-0.0637453571,-0.0362155363,0.0537013672,0.3318411708,0.087192513,-0.0709578693,-0.105753161,0.497995466,0.0180258732,0.1023961678,0.2883635163,-0.2016443759,0.0996336117,-0.0990453884,0.0020643868,0.2259092182,0.1946656704,-0.0812613964,0.0224434528,0.1418915838,0.1574391872,0.0746554807,-0.0048610461,-0.1536109,0.139052406,0.0118647469,0.0111010959,0.0971629173,0.0520384163,0.2940087616,0.5483121872,0.1111362055,-0.1868935674,0.1310922652,0.3315558434,-0.0650450215,0.2757264376,-0.0043974128,-0.171322912,-0.0783392414,0.1010062397,0.178945601,0.5017948747,-0.00362671,-0.1863486022,-0.1349211335,-0.063677758,-0.0468234979,0.0711337104,-0.1824291945,0.255372107,0.2577778697,0.1839800775,0.0547975115,-0.1531583816,-0.0497436859,0.0207577217,0.2905252874,-0.3741133809,0.0461585224,-0.1272047311,0.0979508162,-0.3777190447,0.133810699,-0.3502556384,-0.1833577752,-0.135731101,-0.1108105704,0.309304595,0.2101667076,-0.3485357165,0.1555947661,0.2402642965,-0.5443775654,0.1576044559,0.1009899378,-0.3830037117,0.0529526137,0.1205421314,-0.0696215704,-0.0969110429,-0.1528277695,0.0173257664,-0.1201020107,-0.168004334,0.1868712455,-0.0529078133,0.576833725,-0.0123318527,0.1703878939,-0.0075948159,-0.0680116788,0.166756928,-0.1843080521,-0.1167619601,-0.1657404155,-0.1121703759,0.0364880748,-0.0776862428,-0.622492671,-0.4673993886,-0.2988173664,-0.1805043668,0.0007811155,0.2649473548,0.0936486349,0.0102780862,0.3407399952,-0.0355759114,0.0361259319,-0.4035873711,-0.2818406522,0.2556055784,0.0628466681,-0.2947934866,-0.0994649678,0.0628204569,0.2072016597,-0.0956361443,-0.3958399296,-0.2739957571,0.2307893485,-0.0112170214,-0.0910095647,0.0472379848,0.1151574999,-0.1613552123,0.1061526462,-0.2067129463,-0.5061612129,0.1450342834,0.1470152438,0.6205662489,-0.1660099775,0.2980054021,-0.0526382029,0.6775237918,0.1890775263,-0.2280175686,0.2107886672,-0.2105231285,0.3032869399,0.0035384649,-0.2769274116,0.3039433956,0.0719058588,0.0459164828,0.3248111308,0.0534212366,-0.0166754816,-0.0613191389,0.3597286642,-0.1392308474,-0.1571274251,-0.0605640151,-0.0752931461,-0.2700310051,0.0741321221,0.3065279722,-0.1772015244,-0.1673002541,-0.1517760754,0.4453921914,-0.0859651938,0.2577697635,-0.6791188717,-0.0043502864,-0.1259898394,0.286806494,-0.1449967176,0.6176644564,-0.0544948392,-0.185738191,-0.0650190338,0.0791004077,0.4798837304,-0.0459979139,-0.0093406076,-0.0361170694,0.2946531475,-0.6179642081,-0.1878311783,0.177603066,0.1685312092,0.0936853364,0.3528563976,-0.1821606308,-0.0771105886,0.1613401026,0.199181214,-0.1662609279,-0.0170077868,-0.3835242391,-0.2073016763,-0.3429889977,-0.0511813834,-0.276417762,0.1328058094,0.1391488761,0.1173347458,0.2389433235,-0.1128840968,0.3738936782,0.1367881894,0.3835945725,0.1973970532,0.1353370249,0.2334738374,0.2617393732,-0.1502471417,0.4892303646,-0.4564051628,-0.3000645041,0.0343399867,-0.0737037808,0.1476985812,0.3775250018,-0.1436186284,0.0837649927,-0.3207883835,-0.0862554312,0.275416106,0.1629421562,0.2510092556,0.0771014914,-0.1133952439,-0.4098265171,0.1798558235,-0.112034753,-0.0426193438,0.2145652771,-0.4636399746,-0.2888949513,0.0964138806,-0.1039724872,0.8939831853,-0.0649739653,0.1889764965,0.3377411067,0.163262248,0.2822509706,-0.0774445832,0.1654775292,-0.4589556456,-0.0782304406,0.0744747519,-0.1392124891,-0.0647865608,0.3663628101,0.0141666662,0.5146529078,-0.2688012421,0.3116567731,0.0540146455,0.5210416317,-0.1337990165,-0.010874562,-0.3472368717,0.1949188709,-0.1809049547,0.2802450955,-0.1203895137,0.0054683955,-0.3029422462,-0.1379414052,-0.131941393,0.2274260819,-0.5376272798,0.1669765413,0.0934275091,-0.3486429155,0.0191773493,0.5643280745,-0.1002912,-0.0820765272,-0.1084556207,0.353082329,0.0207632296,0.3046307564,0.0768876597,0.0457809493,0.4243014753,-0.0168561712,-0.0948249623,0.0720392466,-0.0042471443,-0.3094080985,-0.2512997389,0.040815454,0.3062185347,-0.2756883502,-0.1972322911,-0.1188386902,0.0159458686,-0.0596131496,0.1198260859,0.0611222945,-0.0966570526,0.2068720311,0.1296120435,-0.4074319601,0.0401298888,0.3414311111,0.0581959672,0.003053996,0.6670231223,0.0741890967,0.0704765096,-0.1949026883,0.1497739553,0.301173836,-0.2336236984,-0.039200563,-0.2000790834,-0.2490169704,0.0004849006,-0.1210952997,0.2407468706,-0.0176706649,0.040279977,-0.4578694999,-0.4342662692,0.031510029,-0.0630608797,0.0092651313,0.0782028437,-0.0439870693,0.1450648308,0.0453661457,-0.2561044693,0.2107708305,-0.1966145039,0.2120202482,0.0993529782,-0.0401501283,0.249724865,-0.0228080153,0.0503343157,-0.0272150729,-0.3081597984,-0.1973445117,-0.12951231,0.1182673052,0.0220857915,-0.015638737,-0.1738801301,-0.2943677902,-0.1543195397,-0.247476548,-0.0346161649,0.0881620198,0.0274494998,0.0142343817,-0.0534722954,0.0265387595,0.0657878071,-0.0739742294,0.0375893079,0.0440010689,0.2452987581,0.1012571603,0.047981102,-0.0090396414,-0.5060158372,0.1013093889,0.1878549755,0.38351354,0.0299073067,-0.1631867141,-0.038469024,0.2729744315,0.0583594814,0.4147221148,-0.4566241503,-0.0442090221,-0.0886465684,0.143432647,-0.2171788365,0.0970659181,0.3625798225,-0.1833094358,-0.0155572481,0.2735645771,0.1266646832,0.189718172,0.0217407867,0.1068987325,0.5871457458,-0.0874021053,0.1646409184,0.2900642157,-0.2160798758,0.2205359042,0.3349782526,-0.0837548003,-0.0087310392,0.3994481862,-0.1929351389,0.3431117833,-0.0790344253,-0.1280991137,0.2440442443,-0.2488018423,0.0799286142,0.402431488,-0.1566050351,0.1699330658,0.0469692349,0.539581418,-0.4212684333,-0.1374428123,-0.0382384658,-0.1659311652,-0.1484349817,-0.0598560758,0.0648372024,-0.1156122908,-0.1001708731,-0.2388545126,-0.1484500319,-0.3574163914,-0.0503093265,-0.003967029,-0.3478863537,-0.0714054331,-0.0686984584,0.3550385833,-0.1964205354,-0.1058571562,0.3332430124,-0.1416485012,0.0370896682,0.3267581463,0.2570812404,0.195392713,0.445263952,0.0781018436,-0.1183174253,0.3155586123,0.0447255895,0.014626109,0.1478066146,-0.1330287158,-0.1975685507,0.3923157752,0.1372191608,-0.0548008196,-0.2756023109,0.3441493809,-0.066534847,-0.2472360432,0.2151201516,-0.2440628409,-0.0380831957,-0.5316839218,0.3191308975,-0.3130171299,0.0980828851,0.4218278825,0.1543100178,0.3799419999,-0.2579372823,0.0727035478,-0.0288272072,0.3897283077,0.1729896069,-0.2461687326,-0.1826694757,0.1881106496,-0.9209598899,0.4458393455,0.0589725524,-0.039084103,0.0334376767,0.3439892232,0.053749986,0.0531443954,0.2887405157,-0.0962834656,0.1624656022,-0.0415589474,-0.3789632022,-0.3341313899,-0.3106013536,0.4049323797,0.1494320929,-0.5924777985,0.289875567,0.038488958,-0.0501692854,-0.1049868017,-0.2132235318,0.1590006202,-0.1347794384,0.4179118574,-0.1402909905,0.4938715398,-0.027379442,-0.0191062782,-0.2152624726,-0.1111690998,-0.2957692444,0.2690999806,-0.1461151987,0.0767935514,0.1251032054,0.0760321543,-0.298789531,0.3234610856,0.1469838619,-0.2812111974,-0.4066532552,-0.0215115547,0.0377548337,0.0922570005,0.0601231754,0.3981387019,-0.1498904824,0.2334185094,-0.1931376159,-0.3373351097,0.5052189827,-0.3257136941,-0.1926412582,-0.2741817236,0.2520015538,0.1326462328,0.0052120443,-0.640047133,0.0537735224,0.3680356443,-0.1895976514,-0.1688360572,0.1406750828,-0.2344212979,0.4075179398,0.0046436405,0.0375304818,0.0067876321,-0.1305246651,0.1636307687,-0.2550606132]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1964","title":"Datasets.py function load_dataset does not match squad dataset","comments":"Thks for quickly answering\uff01\r\n### 1 I try the first way,but seems not work \r\n```\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 503, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 218, in main\r\n datasets = load_dataset(data_args.dataset_name, download_mode=\"force_redownload\")\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\n### 2 I try the second way,and run the examples\/question-answering\/run_qa.py,it lead to another bug orz..\r\n```\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 523, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 379, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1120, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1091, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"examples\/question-answering\/run_qa.py\", line 339, in prepare_train_features\r\n if len(answers[\"answer_start\"]) == 0:\r\nTypeError: list indices must be integers or slices, not str\r\n```\r\n## may be the function prepare_train_features in run_qa.py need to fix,I think is that the prep\r\n```python\r\nfor i, offsets in enumerate(offset_mapping):\r\n # We will label impossible answers with the index of the CLS token.\r\n input_ids = tokenized_examples[\"input_ids\"][i]\r\n cls_index = input_ids.index(tokenizer.cls_token_id)\r\n\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n answers = examples[answer_column_name][sample_index]\r\n print(examples,answers)\r\n # If no answers are given, set the cls_index as answer.\r\n if len(answers[\"answer_start\"]) == 0:\r\n tokenized_examples[\"start_positions\"].append(cls_index)\r\n tokenized_examples[\"end_positions\"].append(cls_index)\r\n else:\r\n # Start\/end character index of the answer in the text.\r\n start_char = answers[\"answer_start\"][0]\r\n end_char = start_char + len(answers[\"text\"][0])\r\n\r\n # Start token index of the current span in the text.\r\n token_start_index = 0\r\n while sequence_ids[token_start_index] != (1 if pad_on_right else 0):\r\n token_start_index += 1\r\n\r\n # End token index of the current span in the text.\r\n token_end_index = len(input_ids) - 1\r\n while sequence_ids[token_end_index] != (1 if pad_on_right else 0):\r\n token_end_index -= 1\r\n\r\n # Detect if the answer is out of the span (in which case this feature is labeled with the CLS index).\r\n if not (offsets[token_start_index][0] <= start_char and offsets[token_end_index][1] >= end_char):\r\n tokenized_examples[\"start_positions\"].append(cls_index)\r\n tokenized_examples[\"end_positions\"].append(cls_index)\r\n else:\r\n # Otherwise move the token_start_index and token_end_index to the two ends of the answer.\r\n # Note: we could go after the last offset if the answer is the last word (edge case).\r\n while token_start_index < len(offsets) and offsets[token_start_index][0] <= start_char:\r\n token_start_index += 1\r\n tokenized_examples[\"start_positions\"].append(token_start_index - 1)\r\n while offsets[token_end_index][1] >= end_char:\r\n token_end_index -= 1\r\n tokenized_examples[\"end_positions\"].append(token_end_index + 1)\r\n\r\n return tokenized_examples\r\n``` ","body":"### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!","comment_length":434,"text":"Datasets.py function load_dataset does not match squad dataset\n### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!\nThks for quickly answering\uff01\r\n### 1 I try the first way,but seems not work \r\n```\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 503, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 218, in main\r\n datasets = load_dataset(data_args.dataset_name, download_mode=\"force_redownload\")\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\n### 2 I try the second way,and run the examples\/question-answering\/run_qa.py,it lead to another bug orz..\r\n```\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 523, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 379, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1120, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1091, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"examples\/question-answering\/run_qa.py\", line 339, in prepare_train_features\r\n if len(answers[\"answer_start\"]) == 0:\r\nTypeError: list indices must be integers or slices, not str\r\n```\r\n## may be the function prepare_train_features in run_qa.py need to fix,I think is that the prep\r\n```python\r\nfor i, offsets in enumerate(offset_mapping):\r\n # We will label impossible answers with the index of the CLS token.\r\n input_ids = tokenized_examples[\"input_ids\"][i]\r\n cls_index = input_ids.index(tokenizer.cls_token_id)\r\n\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n answers = examples[answer_column_name][sample_index]\r\n print(examples,answers)\r\n # If no answers are given, set the cls_index as answer.\r\n if len(answers[\"answer_start\"]) == 0:\r\n tokenized_examples[\"start_positions\"].append(cls_index)\r\n tokenized_examples[\"end_positions\"].append(cls_index)\r\n else:\r\n # Start\/end character index of the answer in the text.\r\n start_char = answers[\"answer_start\"][0]\r\n end_char = start_char + len(answers[\"text\"][0])\r\n\r\n # Start token index of the current span in the text.\r\n token_start_index = 0\r\n while sequence_ids[token_start_index] != (1 if pad_on_right else 0):\r\n token_start_index += 1\r\n\r\n # End token index of the current span in the text.\r\n token_end_index = len(input_ids) - 1\r\n while sequence_ids[token_end_index] != (1 if pad_on_right else 0):\r\n token_end_index -= 1\r\n\r\n # Detect if the answer is out of the span (in which case this feature is labeled with the CLS index).\r\n if not (offsets[token_start_index][0] <= start_char and offsets[token_end_index][1] >= end_char):\r\n tokenized_examples[\"start_positions\"].append(cls_index)\r\n tokenized_examples[\"end_positions\"].append(cls_index)\r\n else:\r\n # Otherwise move the token_start_index and token_end_index to the two ends of the answer.\r\n # Note: we could go after the last offset if the answer is the last word (edge case).\r\n while token_start_index < len(offsets) and offsets[token_start_index][0] <= start_char:\r\n token_start_index += 1\r\n tokenized_examples[\"start_positions\"].append(token_start_index - 1)\r\n while offsets[token_end_index][1] >= end_char:\r\n token_end_index -= 1\r\n tokenized_examples[\"end_positions\"].append(token_end_index + 1)\r\n\r\n return tokenized_examples\r\n``` ","embeddings":[-0.3537200987,0.0625078902,0.0333654545,0.3963988423,0.5554530621,0.0057210135,0.5457583666,0.3404688239,-0.0983101875,-0.1216232106,-0.1650660336,0.4508374333,0.1084147692,-0.0774192885,0.1859294176,0.2271618843,-0.1242239624,-0.0123368138,-0.1461081356,-0.2893434763,-0.1109781116,0.1282911897,-0.0823139772,0.2603987455,-0.1788990498,0.0166908093,0.2006343901,0.2723646462,-0.0037961667,-0.2199688405,0.3428422213,-0.2606053054,0.2361012548,0.5758861899,-0.0001148581,0.1132558137,-0.0115434257,-0.3397289217,-0.5327045321,-0.2002218962,-0.1021837071,-0.3757722974,0.1486449242,-0.2605553567,-0.1595326066,0.093737483,0.0203164779,-0.4645369053,0.3544548452,0.4883137345,0.213982597,0.2522563636,-0.1001010016,0.0775510594,0.1142548099,-0.0009760508,-0.06950894,0.1360990405,0.3347509801,-0.0582686849,0.0025541722,0.1724455059,0.0828086361,0.0568236373,0.2703095675,0.0064037591,0.199115023,-0.316155225,0.0564422533,0.3177836239,0.4627765119,-0.4630158544,-0.334705174,-0.2631344795,0.133104369,-0.2565729618,0.1320165247,0.0672778934,0.0511586368,-0.0424272642,0.0477764271,0.0376646332,0.1114287302,0.0368287303,0.0068916786,0.1999943703,0.0288672037,0.2383961529,-0.201539591,-0.1135483086,0.1148922518,-0.083149083,0.0818328112,0.392493546,-0.828517139,-0.0814113691,-0.2042951286,-0.1421720982,0.2562874556,0.0819008574,0.0386924408,-0.0269711483,0.148668915,0.1044574156,0.1687794775,0.3692035675,0.3052028716,0.2674016058,-0.0663638264,0.0668205246,-0.3717198968,0.2254103124,-0.3225816488,-0.1618057936,0.0960089266,0.0835877731,0.0062298947,-0.1927381605,-0.5517169237,0.1761802584,0.1544087082,0.0026421051,0.0974367931,0.324791342,0.0759165287,0.3802291751,0.0263226684,0.1048553437,-0.294623524,-0.3704161942,-0.2408784777,0.0330561139,-0.0880101547,0.0268361289,0.0183388889,-0.0356007777,0.3227834404,-0.1957616955,0.2114947587,-0.1790799946,0.1150045693,-0.2556711733,-0.1285898685,-0.0065955352,0.0298101194,0.0322048925,0.1742971241,-0.3730719388,-0.0482525341,0.1222925037,-0.3712684512,-0.0336521827,-0.1820591837,0.1871163845,-0.1391085088,0.0208789185,-0.2373687923,0.1698867232,0.3073554337,-0.1878401935,0.1063959599,-0.3961229026,-0.1664000899,-0.2122270316,0.3151646554,0.3882876039,-0.4074572027,-0.2117103338,0.1249395758,-0.0952139571,-0.0381632037,0.1216655001,-0.1516055316,0.1555615067,-0.20622392,-0.0070253811,0.7145847082,-0.8443276286,-0.4243383706,0.1099728644,-0.3059237599,-0.0194989499,-0.0710134581,0.066407755,0.1930091679,0.0147336535,0.0827305093,0.5066136718,-0.0362928361,0.0914031267,-0.1724580824,-0.2712134719,0.2526653707,0.1708515584,0.0318142921,-0.0380649008,0.0388377123,0.4989839792,0.332213819,0.1426163316,0.1112044156,-0.0299585946,-0.0359400138,-0.0821635574,0.114605628,-0.1481092274,-0.6798553467,0.1638657302,-0.1644516885,0.0439888574,0.2410787791,-0.1508302242,-0.2135134637,-0.0917369127,-0.4252310395,-0.1023878902,0.0743926764,0.1495984495,-0.0585200973,-0.1226450801,-0.159884721,0.3514788449,-0.2154513896,0.2284579277,-0.4849060774,-0.0965390429,-0.0637453571,-0.0362155363,0.0537013672,0.3318411708,0.087192513,-0.0709578693,-0.105753161,0.497995466,0.0180258732,0.1023961678,0.2883635163,-0.2016443759,0.0996336117,-0.0990453884,0.0020643868,0.2259092182,0.1946656704,-0.0812613964,0.0224434528,0.1418915838,0.1574391872,0.0746554807,-0.0048610461,-0.1536109,0.139052406,0.0118647469,0.0111010959,0.0971629173,0.0520384163,0.2940087616,0.5483121872,0.1111362055,-0.1868935674,0.1310922652,0.3315558434,-0.0650450215,0.2757264376,-0.0043974128,-0.171322912,-0.0783392414,0.1010062397,0.178945601,0.5017948747,-0.00362671,-0.1863486022,-0.1349211335,-0.063677758,-0.0468234979,0.0711337104,-0.1824291945,0.255372107,0.2577778697,0.1839800775,0.0547975115,-0.1531583816,-0.0497436859,0.0207577217,0.2905252874,-0.3741133809,0.0461585224,-0.1272047311,0.0979508162,-0.3777190447,0.133810699,-0.3502556384,-0.1833577752,-0.135731101,-0.1108105704,0.309304595,0.2101667076,-0.3485357165,0.1555947661,0.2402642965,-0.5443775654,0.1576044559,0.1009899378,-0.3830037117,0.0529526137,0.1205421314,-0.0696215704,-0.0969110429,-0.1528277695,0.0173257664,-0.1201020107,-0.168004334,0.1868712455,-0.0529078133,0.576833725,-0.0123318527,0.1703878939,-0.0075948159,-0.0680116788,0.166756928,-0.1843080521,-0.1167619601,-0.1657404155,-0.1121703759,0.0364880748,-0.0776862428,-0.622492671,-0.4673993886,-0.2988173664,-0.1805043668,0.0007811155,0.2649473548,0.0936486349,0.0102780862,0.3407399952,-0.0355759114,0.0361259319,-0.4035873711,-0.2818406522,0.2556055784,0.0628466681,-0.2947934866,-0.0994649678,0.0628204569,0.2072016597,-0.0956361443,-0.3958399296,-0.2739957571,0.2307893485,-0.0112170214,-0.0910095647,0.0472379848,0.1151574999,-0.1613552123,0.1061526462,-0.2067129463,-0.5061612129,0.1450342834,0.1470152438,0.6205662489,-0.1660099775,0.2980054021,-0.0526382029,0.6775237918,0.1890775263,-0.2280175686,0.2107886672,-0.2105231285,0.3032869399,0.0035384649,-0.2769274116,0.3039433956,0.0719058588,0.0459164828,0.3248111308,0.0534212366,-0.0166754816,-0.0613191389,0.3597286642,-0.1392308474,-0.1571274251,-0.0605640151,-0.0752931461,-0.2700310051,0.0741321221,0.3065279722,-0.1772015244,-0.1673002541,-0.1517760754,0.4453921914,-0.0859651938,0.2577697635,-0.6791188717,-0.0043502864,-0.1259898394,0.286806494,-0.1449967176,0.6176644564,-0.0544948392,-0.185738191,-0.0650190338,0.0791004077,0.4798837304,-0.0459979139,-0.0093406076,-0.0361170694,0.2946531475,-0.6179642081,-0.1878311783,0.177603066,0.1685312092,0.0936853364,0.3528563976,-0.1821606308,-0.0771105886,0.1613401026,0.199181214,-0.1662609279,-0.0170077868,-0.3835242391,-0.2073016763,-0.3429889977,-0.0511813834,-0.276417762,0.1328058094,0.1391488761,0.1173347458,0.2389433235,-0.1128840968,0.3738936782,0.1367881894,0.3835945725,0.1973970532,0.1353370249,0.2334738374,0.2617393732,-0.1502471417,0.4892303646,-0.4564051628,-0.3000645041,0.0343399867,-0.0737037808,0.1476985812,0.3775250018,-0.1436186284,0.0837649927,-0.3207883835,-0.0862554312,0.275416106,0.1629421562,0.2510092556,0.0771014914,-0.1133952439,-0.4098265171,0.1798558235,-0.112034753,-0.0426193438,0.2145652771,-0.4636399746,-0.2888949513,0.0964138806,-0.1039724872,0.8939831853,-0.0649739653,0.1889764965,0.3377411067,0.163262248,0.2822509706,-0.0774445832,0.1654775292,-0.4589556456,-0.0782304406,0.0744747519,-0.1392124891,-0.0647865608,0.3663628101,0.0141666662,0.5146529078,-0.2688012421,0.3116567731,0.0540146455,0.5210416317,-0.1337990165,-0.010874562,-0.3472368717,0.1949188709,-0.1809049547,0.2802450955,-0.1203895137,0.0054683955,-0.3029422462,-0.1379414052,-0.131941393,0.2274260819,-0.5376272798,0.1669765413,0.0934275091,-0.3486429155,0.0191773493,0.5643280745,-0.1002912,-0.0820765272,-0.1084556207,0.353082329,0.0207632296,0.3046307564,0.0768876597,0.0457809493,0.4243014753,-0.0168561712,-0.0948249623,0.0720392466,-0.0042471443,-0.3094080985,-0.2512997389,0.040815454,0.3062185347,-0.2756883502,-0.1972322911,-0.1188386902,0.0159458686,-0.0596131496,0.1198260859,0.0611222945,-0.0966570526,0.2068720311,0.1296120435,-0.4074319601,0.0401298888,0.3414311111,0.0581959672,0.003053996,0.6670231223,0.0741890967,0.0704765096,-0.1949026883,0.1497739553,0.301173836,-0.2336236984,-0.039200563,-0.2000790834,-0.2490169704,0.0004849006,-0.1210952997,0.2407468706,-0.0176706649,0.040279977,-0.4578694999,-0.4342662692,0.031510029,-0.0630608797,0.0092651313,0.0782028437,-0.0439870693,0.1450648308,0.0453661457,-0.2561044693,0.2107708305,-0.1966145039,0.2120202482,0.0993529782,-0.0401501283,0.249724865,-0.0228080153,0.0503343157,-0.0272150729,-0.3081597984,-0.1973445117,-0.12951231,0.1182673052,0.0220857915,-0.015638737,-0.1738801301,-0.2943677902,-0.1543195397,-0.247476548,-0.0346161649,0.0881620198,0.0274494998,0.0142343817,-0.0534722954,0.0265387595,0.0657878071,-0.0739742294,0.0375893079,0.0440010689,0.2452987581,0.1012571603,0.047981102,-0.0090396414,-0.5060158372,0.1013093889,0.1878549755,0.38351354,0.0299073067,-0.1631867141,-0.038469024,0.2729744315,0.0583594814,0.4147221148,-0.4566241503,-0.0442090221,-0.0886465684,0.143432647,-0.2171788365,0.0970659181,0.3625798225,-0.1833094358,-0.0155572481,0.2735645771,0.1266646832,0.189718172,0.0217407867,0.1068987325,0.5871457458,-0.0874021053,0.1646409184,0.2900642157,-0.2160798758,0.2205359042,0.3349782526,-0.0837548003,-0.0087310392,0.3994481862,-0.1929351389,0.3431117833,-0.0790344253,-0.1280991137,0.2440442443,-0.2488018423,0.0799286142,0.402431488,-0.1566050351,0.1699330658,0.0469692349,0.539581418,-0.4212684333,-0.1374428123,-0.0382384658,-0.1659311652,-0.1484349817,-0.0598560758,0.0648372024,-0.1156122908,-0.1001708731,-0.2388545126,-0.1484500319,-0.3574163914,-0.0503093265,-0.003967029,-0.3478863537,-0.0714054331,-0.0686984584,0.3550385833,-0.1964205354,-0.1058571562,0.3332430124,-0.1416485012,0.0370896682,0.3267581463,0.2570812404,0.195392713,0.445263952,0.0781018436,-0.1183174253,0.3155586123,0.0447255895,0.014626109,0.1478066146,-0.1330287158,-0.1975685507,0.3923157752,0.1372191608,-0.0548008196,-0.2756023109,0.3441493809,-0.066534847,-0.2472360432,0.2151201516,-0.2440628409,-0.0380831957,-0.5316839218,0.3191308975,-0.3130171299,0.0980828851,0.4218278825,0.1543100178,0.3799419999,-0.2579372823,0.0727035478,-0.0288272072,0.3897283077,0.1729896069,-0.2461687326,-0.1826694757,0.1881106496,-0.9209598899,0.4458393455,0.0589725524,-0.039084103,0.0334376767,0.3439892232,0.053749986,0.0531443954,0.2887405157,-0.0962834656,0.1624656022,-0.0415589474,-0.3789632022,-0.3341313899,-0.3106013536,0.4049323797,0.1494320929,-0.5924777985,0.289875567,0.038488958,-0.0501692854,-0.1049868017,-0.2132235318,0.1590006202,-0.1347794384,0.4179118574,-0.1402909905,0.4938715398,-0.027379442,-0.0191062782,-0.2152624726,-0.1111690998,-0.2957692444,0.2690999806,-0.1461151987,0.0767935514,0.1251032054,0.0760321543,-0.298789531,0.3234610856,0.1469838619,-0.2812111974,-0.4066532552,-0.0215115547,0.0377548337,0.0922570005,0.0601231754,0.3981387019,-0.1498904824,0.2334185094,-0.1931376159,-0.3373351097,0.5052189827,-0.3257136941,-0.1926412582,-0.2741817236,0.2520015538,0.1326462328,0.0052120443,-0.640047133,0.0537735224,0.3680356443,-0.1895976514,-0.1688360572,0.1406750828,-0.2344212979,0.4075179398,0.0046436405,0.0375304818,0.0067876321,-0.1305246651,0.1636307687,-0.2550606132]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1964","title":"Datasets.py function load_dataset does not match squad dataset","comments":"## I have fixed it, @lhoestq \r\n### the first section change as you said and add [\"id\"]\r\n```python\r\ndef process_squad(examples):\r\n \"\"\"\r\n Process a dataset in the squad format with columns \"title\" and \"paragraphs\"\r\n to return the dataset with columns \"context\", \"question\" and \"answers\".\r\n \"\"\"\r\n # print(examples)\r\n out = {\"context\": [], \"question\": [], \"answers\":[],\"id\":[]} \r\n for paragraphs in examples[\"paragraphs\"]: \r\n for paragraph in paragraphs: \r\n for qa in paragraph[\"qas\"]: \r\n answers = [{\"answer_start\": answer[\"answer_start\"], \"text\": answer[\"text\"].strip()} for answer in qa[\"answers\"]] \r\n out[\"context\"].append(paragraph[\"context\"].strip()) \r\n out[\"question\"].append(qa[\"question\"].strip()) \r\n out[\"answers\"].append(answers) \r\n out[\"id\"].append(qa[\"id\"]) \r\n return out\r\ncolumn_names = datasets[\"train\"].column_names if training_args.do_train else datasets[\"validation\"].column_names\r\n# print(datasets[\"train\"].column_names)\r\nif set(column_names) == {\"title\", \"paragraphs\"}:\r\n datasets = datasets.map(process_squad, batched=True, remove_columns=column_names)\r\n# Preprocessing the datasets.\r\n# Preprocessing is slighlty different for training and evaluation.\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\nelse:\r\n column_names = datasets[\"validation\"].column_names\r\n# print(column_names)\r\nquestion_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\ncontext_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\nanswer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n```\r\n### the second section\r\n```python\r\ndef prepare_train_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[question_column_name if pad_on_right else context_column_name],\r\n examples[context_column_name if pad_on_right else question_column_name],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=data_args.max_seq_length,\r\n stride=data_args.doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\" if data_args.pad_to_max_length else False,\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n # The offset mappings will give us a map from token to character position in the original context. This will\r\n # help us compute the start_positions and end_positions.\r\n offset_mapping = tokenized_examples.pop(\"offset_mapping\")\r\n\r\n # Let's label those examples!\r\n tokenized_examples[\"start_positions\"] = []\r\n tokenized_examples[\"end_positions\"] = []\r\n\r\n for i, offsets in enumerate(offset_mapping):\r\n # We will label impossible answers with the index of the CLS token.\r\n input_ids = tokenized_examples[\"input_ids\"][i]\r\n cls_index = input_ids.index(tokenizer.cls_token_id)\r\n\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n answers = examples[answer_column_name][sample_index]\r\n # print(examples,answers,offset_mapping,tokenized_examples)\r\n # If no answers are given, set the cls_index as answer.\r\n if len(answers) == 0:#len(answers[\"answer_start\"]) == 0:\r\n tokenized_examples[\"start_positions\"].append(cls_index)\r\n tokenized_examples[\"end_positions\"].append(cls_index)\r\n else:\r\n # Start\/end character index of the answer in the text.\r\n start_char = answers[0][\"answer_start\"]\r\n end_char = start_char + len(answers[0][\"text\"])\r\n\r\n # Start token index of the current span in the text.\r\n token_start_index = 0\r\n while sequence_ids[token_start_index] != (1 if pad_on_right else 0):\r\n token_start_index += 1\r\n\r\n # End token index of the current span in the text.\r\n token_end_index = len(input_ids) - 1\r\n while sequence_ids[token_end_index] != (1 if pad_on_right else 0):\r\n token_end_index -= 1\r\n\r\n # Detect if the answer is out of the span (in which case this feature is labeled with the CLS index).\r\n if not (offsets[token_start_index][0] <= start_char and offsets[token_end_index][1] >= end_char):\r\n tokenized_examples[\"start_positions\"].append(cls_index)\r\n tokenized_examples[\"end_positions\"].append(cls_index)\r\n else:\r\n # Otherwise move the token_start_index and token_end_index to the two ends of the answer.\r\n # Note: we could go after the last offset if the answer is the last word (edge case).\r\n while token_start_index < len(offsets) and offsets[token_start_index][0] <= start_char:\r\n token_start_index += 1\r\n tokenized_examples[\"start_positions\"].append(token_start_index - 1)\r\n while offsets[token_end_index][1] >= end_char:\r\n token_end_index -= 1\r\n tokenized_examples[\"end_positions\"].append(token_end_index + 1)\r\n return tokenized_examples\r\n```","body":"### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!","comment_length":569,"text":"Datasets.py function load_dataset does not match squad dataset\n### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!\n## I have fixed it, @lhoestq \r\n### the first section change as you said and add [\"id\"]\r\n```python\r\ndef process_squad(examples):\r\n \"\"\"\r\n Process a dataset in the squad format with columns \"title\" and \"paragraphs\"\r\n to return the dataset with columns \"context\", \"question\" and \"answers\".\r\n \"\"\"\r\n # print(examples)\r\n out = {\"context\": [], \"question\": [], \"answers\":[],\"id\":[]} \r\n for paragraphs in examples[\"paragraphs\"]: \r\n for paragraph in paragraphs: \r\n for qa in paragraph[\"qas\"]: \r\n answers = [{\"answer_start\": answer[\"answer_start\"], \"text\": answer[\"text\"].strip()} for answer in qa[\"answers\"]] \r\n out[\"context\"].append(paragraph[\"context\"].strip()) \r\n out[\"question\"].append(qa[\"question\"].strip()) \r\n out[\"answers\"].append(answers) \r\n out[\"id\"].append(qa[\"id\"]) \r\n return out\r\ncolumn_names = datasets[\"train\"].column_names if training_args.do_train else datasets[\"validation\"].column_names\r\n# print(datasets[\"train\"].column_names)\r\nif set(column_names) == {\"title\", \"paragraphs\"}:\r\n datasets = datasets.map(process_squad, batched=True, remove_columns=column_names)\r\n# Preprocessing the datasets.\r\n# Preprocessing is slighlty different for training and evaluation.\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\nelse:\r\n column_names = datasets[\"validation\"].column_names\r\n# print(column_names)\r\nquestion_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\ncontext_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\nanswer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n```\r\n### the second section\r\n```python\r\ndef prepare_train_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[question_column_name if pad_on_right else context_column_name],\r\n examples[context_column_name if pad_on_right else question_column_name],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=data_args.max_seq_length,\r\n stride=data_args.doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\" if data_args.pad_to_max_length else False,\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n # The offset mappings will give us a map from token to character position in the original context. This will\r\n # help us compute the start_positions and end_positions.\r\n offset_mapping = tokenized_examples.pop(\"offset_mapping\")\r\n\r\n # Let's label those examples!\r\n tokenized_examples[\"start_positions\"] = []\r\n tokenized_examples[\"end_positions\"] = []\r\n\r\n for i, offsets in enumerate(offset_mapping):\r\n # We will label impossible answers with the index of the CLS token.\r\n input_ids = tokenized_examples[\"input_ids\"][i]\r\n cls_index = input_ids.index(tokenizer.cls_token_id)\r\n\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n answers = examples[answer_column_name][sample_index]\r\n # print(examples,answers,offset_mapping,tokenized_examples)\r\n # If no answers are given, set the cls_index as answer.\r\n if len(answers) == 0:#len(answers[\"answer_start\"]) == 0:\r\n tokenized_examples[\"start_positions\"].append(cls_index)\r\n tokenized_examples[\"end_positions\"].append(cls_index)\r\n else:\r\n # Start\/end character index of the answer in the text.\r\n start_char = answers[0][\"answer_start\"]\r\n end_char = start_char + len(answers[0][\"text\"])\r\n\r\n # Start token index of the current span in the text.\r\n token_start_index = 0\r\n while sequence_ids[token_start_index] != (1 if pad_on_right else 0):\r\n token_start_index += 1\r\n\r\n # End token index of the current span in the text.\r\n token_end_index = len(input_ids) - 1\r\n while sequence_ids[token_end_index] != (1 if pad_on_right else 0):\r\n token_end_index -= 1\r\n\r\n # Detect if the answer is out of the span (in which case this feature is labeled with the CLS index).\r\n if not (offsets[token_start_index][0] <= start_char and offsets[token_end_index][1] >= end_char):\r\n tokenized_examples[\"start_positions\"].append(cls_index)\r\n tokenized_examples[\"end_positions\"].append(cls_index)\r\n else:\r\n # Otherwise move the token_start_index and token_end_index to the two ends of the answer.\r\n # Note: we could go after the last offset if the answer is the last word (edge case).\r\n while token_start_index < len(offsets) and offsets[token_start_index][0] <= start_char:\r\n token_start_index += 1\r\n tokenized_examples[\"start_positions\"].append(token_start_index - 1)\r\n while offsets[token_end_index][1] >= end_char:\r\n token_end_index -= 1\r\n tokenized_examples[\"end_positions\"].append(token_end_index + 1)\r\n return tokenized_examples\r\n```","embeddings":[-0.3537200987,0.0625078902,0.0333654545,0.3963988423,0.5554530621,0.0057210135,0.5457583666,0.3404688239,-0.0983101875,-0.1216232106,-0.1650660336,0.4508374333,0.1084147692,-0.0774192885,0.1859294176,0.2271618843,-0.1242239624,-0.0123368138,-0.1461081356,-0.2893434763,-0.1109781116,0.1282911897,-0.0823139772,0.2603987455,-0.1788990498,0.0166908093,0.2006343901,0.2723646462,-0.0037961667,-0.2199688405,0.3428422213,-0.2606053054,0.2361012548,0.5758861899,-0.0001148581,0.1132558137,-0.0115434257,-0.3397289217,-0.5327045321,-0.2002218962,-0.1021837071,-0.3757722974,0.1486449242,-0.2605553567,-0.1595326066,0.093737483,0.0203164779,-0.4645369053,0.3544548452,0.4883137345,0.213982597,0.2522563636,-0.1001010016,0.0775510594,0.1142548099,-0.0009760508,-0.06950894,0.1360990405,0.3347509801,-0.0582686849,0.0025541722,0.1724455059,0.0828086361,0.0568236373,0.2703095675,0.0064037591,0.199115023,-0.316155225,0.0564422533,0.3177836239,0.4627765119,-0.4630158544,-0.334705174,-0.2631344795,0.133104369,-0.2565729618,0.1320165247,0.0672778934,0.0511586368,-0.0424272642,0.0477764271,0.0376646332,0.1114287302,0.0368287303,0.0068916786,0.1999943703,0.0288672037,0.2383961529,-0.201539591,-0.1135483086,0.1148922518,-0.083149083,0.0818328112,0.392493546,-0.828517139,-0.0814113691,-0.2042951286,-0.1421720982,0.2562874556,0.0819008574,0.0386924408,-0.0269711483,0.148668915,0.1044574156,0.1687794775,0.3692035675,0.3052028716,0.2674016058,-0.0663638264,0.0668205246,-0.3717198968,0.2254103124,-0.3225816488,-0.1618057936,0.0960089266,0.0835877731,0.0062298947,-0.1927381605,-0.5517169237,0.1761802584,0.1544087082,0.0026421051,0.0974367931,0.324791342,0.0759165287,0.3802291751,0.0263226684,0.1048553437,-0.294623524,-0.3704161942,-0.2408784777,0.0330561139,-0.0880101547,0.0268361289,0.0183388889,-0.0356007777,0.3227834404,-0.1957616955,0.2114947587,-0.1790799946,0.1150045693,-0.2556711733,-0.1285898685,-0.0065955352,0.0298101194,0.0322048925,0.1742971241,-0.3730719388,-0.0482525341,0.1222925037,-0.3712684512,-0.0336521827,-0.1820591837,0.1871163845,-0.1391085088,0.0208789185,-0.2373687923,0.1698867232,0.3073554337,-0.1878401935,0.1063959599,-0.3961229026,-0.1664000899,-0.2122270316,0.3151646554,0.3882876039,-0.4074572027,-0.2117103338,0.1249395758,-0.0952139571,-0.0381632037,0.1216655001,-0.1516055316,0.1555615067,-0.20622392,-0.0070253811,0.7145847082,-0.8443276286,-0.4243383706,0.1099728644,-0.3059237599,-0.0194989499,-0.0710134581,0.066407755,0.1930091679,0.0147336535,0.0827305093,0.5066136718,-0.0362928361,0.0914031267,-0.1724580824,-0.2712134719,0.2526653707,0.1708515584,0.0318142921,-0.0380649008,0.0388377123,0.4989839792,0.332213819,0.1426163316,0.1112044156,-0.0299585946,-0.0359400138,-0.0821635574,0.114605628,-0.1481092274,-0.6798553467,0.1638657302,-0.1644516885,0.0439888574,0.2410787791,-0.1508302242,-0.2135134637,-0.0917369127,-0.4252310395,-0.1023878902,0.0743926764,0.1495984495,-0.0585200973,-0.1226450801,-0.159884721,0.3514788449,-0.2154513896,0.2284579277,-0.4849060774,-0.0965390429,-0.0637453571,-0.0362155363,0.0537013672,0.3318411708,0.087192513,-0.0709578693,-0.105753161,0.497995466,0.0180258732,0.1023961678,0.2883635163,-0.2016443759,0.0996336117,-0.0990453884,0.0020643868,0.2259092182,0.1946656704,-0.0812613964,0.0224434528,0.1418915838,0.1574391872,0.0746554807,-0.0048610461,-0.1536109,0.139052406,0.0118647469,0.0111010959,0.0971629173,0.0520384163,0.2940087616,0.5483121872,0.1111362055,-0.1868935674,0.1310922652,0.3315558434,-0.0650450215,0.2757264376,-0.0043974128,-0.171322912,-0.0783392414,0.1010062397,0.178945601,0.5017948747,-0.00362671,-0.1863486022,-0.1349211335,-0.063677758,-0.0468234979,0.0711337104,-0.1824291945,0.255372107,0.2577778697,0.1839800775,0.0547975115,-0.1531583816,-0.0497436859,0.0207577217,0.2905252874,-0.3741133809,0.0461585224,-0.1272047311,0.0979508162,-0.3777190447,0.133810699,-0.3502556384,-0.1833577752,-0.135731101,-0.1108105704,0.309304595,0.2101667076,-0.3485357165,0.1555947661,0.2402642965,-0.5443775654,0.1576044559,0.1009899378,-0.3830037117,0.0529526137,0.1205421314,-0.0696215704,-0.0969110429,-0.1528277695,0.0173257664,-0.1201020107,-0.168004334,0.1868712455,-0.0529078133,0.576833725,-0.0123318527,0.1703878939,-0.0075948159,-0.0680116788,0.166756928,-0.1843080521,-0.1167619601,-0.1657404155,-0.1121703759,0.0364880748,-0.0776862428,-0.622492671,-0.4673993886,-0.2988173664,-0.1805043668,0.0007811155,0.2649473548,0.0936486349,0.0102780862,0.3407399952,-0.0355759114,0.0361259319,-0.4035873711,-0.2818406522,0.2556055784,0.0628466681,-0.2947934866,-0.0994649678,0.0628204569,0.2072016597,-0.0956361443,-0.3958399296,-0.2739957571,0.2307893485,-0.0112170214,-0.0910095647,0.0472379848,0.1151574999,-0.1613552123,0.1061526462,-0.2067129463,-0.5061612129,0.1450342834,0.1470152438,0.6205662489,-0.1660099775,0.2980054021,-0.0526382029,0.6775237918,0.1890775263,-0.2280175686,0.2107886672,-0.2105231285,0.3032869399,0.0035384649,-0.2769274116,0.3039433956,0.0719058588,0.0459164828,0.3248111308,0.0534212366,-0.0166754816,-0.0613191389,0.3597286642,-0.1392308474,-0.1571274251,-0.0605640151,-0.0752931461,-0.2700310051,0.0741321221,0.3065279722,-0.1772015244,-0.1673002541,-0.1517760754,0.4453921914,-0.0859651938,0.2577697635,-0.6791188717,-0.0043502864,-0.1259898394,0.286806494,-0.1449967176,0.6176644564,-0.0544948392,-0.185738191,-0.0650190338,0.0791004077,0.4798837304,-0.0459979139,-0.0093406076,-0.0361170694,0.2946531475,-0.6179642081,-0.1878311783,0.177603066,0.1685312092,0.0936853364,0.3528563976,-0.1821606308,-0.0771105886,0.1613401026,0.199181214,-0.1662609279,-0.0170077868,-0.3835242391,-0.2073016763,-0.3429889977,-0.0511813834,-0.276417762,0.1328058094,0.1391488761,0.1173347458,0.2389433235,-0.1128840968,0.3738936782,0.1367881894,0.3835945725,0.1973970532,0.1353370249,0.2334738374,0.2617393732,-0.1502471417,0.4892303646,-0.4564051628,-0.3000645041,0.0343399867,-0.0737037808,0.1476985812,0.3775250018,-0.1436186284,0.0837649927,-0.3207883835,-0.0862554312,0.275416106,0.1629421562,0.2510092556,0.0771014914,-0.1133952439,-0.4098265171,0.1798558235,-0.112034753,-0.0426193438,0.2145652771,-0.4636399746,-0.2888949513,0.0964138806,-0.1039724872,0.8939831853,-0.0649739653,0.1889764965,0.3377411067,0.163262248,0.2822509706,-0.0774445832,0.1654775292,-0.4589556456,-0.0782304406,0.0744747519,-0.1392124891,-0.0647865608,0.3663628101,0.0141666662,0.5146529078,-0.2688012421,0.3116567731,0.0540146455,0.5210416317,-0.1337990165,-0.010874562,-0.3472368717,0.1949188709,-0.1809049547,0.2802450955,-0.1203895137,0.0054683955,-0.3029422462,-0.1379414052,-0.131941393,0.2274260819,-0.5376272798,0.1669765413,0.0934275091,-0.3486429155,0.0191773493,0.5643280745,-0.1002912,-0.0820765272,-0.1084556207,0.353082329,0.0207632296,0.3046307564,0.0768876597,0.0457809493,0.4243014753,-0.0168561712,-0.0948249623,0.0720392466,-0.0042471443,-0.3094080985,-0.2512997389,0.040815454,0.3062185347,-0.2756883502,-0.1972322911,-0.1188386902,0.0159458686,-0.0596131496,0.1198260859,0.0611222945,-0.0966570526,0.2068720311,0.1296120435,-0.4074319601,0.0401298888,0.3414311111,0.0581959672,0.003053996,0.6670231223,0.0741890967,0.0704765096,-0.1949026883,0.1497739553,0.301173836,-0.2336236984,-0.039200563,-0.2000790834,-0.2490169704,0.0004849006,-0.1210952997,0.2407468706,-0.0176706649,0.040279977,-0.4578694999,-0.4342662692,0.031510029,-0.0630608797,0.0092651313,0.0782028437,-0.0439870693,0.1450648308,0.0453661457,-0.2561044693,0.2107708305,-0.1966145039,0.2120202482,0.0993529782,-0.0401501283,0.249724865,-0.0228080153,0.0503343157,-0.0272150729,-0.3081597984,-0.1973445117,-0.12951231,0.1182673052,0.0220857915,-0.015638737,-0.1738801301,-0.2943677902,-0.1543195397,-0.247476548,-0.0346161649,0.0881620198,0.0274494998,0.0142343817,-0.0534722954,0.0265387595,0.0657878071,-0.0739742294,0.0375893079,0.0440010689,0.2452987581,0.1012571603,0.047981102,-0.0090396414,-0.5060158372,0.1013093889,0.1878549755,0.38351354,0.0299073067,-0.1631867141,-0.038469024,0.2729744315,0.0583594814,0.4147221148,-0.4566241503,-0.0442090221,-0.0886465684,0.143432647,-0.2171788365,0.0970659181,0.3625798225,-0.1833094358,-0.0155572481,0.2735645771,0.1266646832,0.189718172,0.0217407867,0.1068987325,0.5871457458,-0.0874021053,0.1646409184,0.2900642157,-0.2160798758,0.2205359042,0.3349782526,-0.0837548003,-0.0087310392,0.3994481862,-0.1929351389,0.3431117833,-0.0790344253,-0.1280991137,0.2440442443,-0.2488018423,0.0799286142,0.402431488,-0.1566050351,0.1699330658,0.0469692349,0.539581418,-0.4212684333,-0.1374428123,-0.0382384658,-0.1659311652,-0.1484349817,-0.0598560758,0.0648372024,-0.1156122908,-0.1001708731,-0.2388545126,-0.1484500319,-0.3574163914,-0.0503093265,-0.003967029,-0.3478863537,-0.0714054331,-0.0686984584,0.3550385833,-0.1964205354,-0.1058571562,0.3332430124,-0.1416485012,0.0370896682,0.3267581463,0.2570812404,0.195392713,0.445263952,0.0781018436,-0.1183174253,0.3155586123,0.0447255895,0.014626109,0.1478066146,-0.1330287158,-0.1975685507,0.3923157752,0.1372191608,-0.0548008196,-0.2756023109,0.3441493809,-0.066534847,-0.2472360432,0.2151201516,-0.2440628409,-0.0380831957,-0.5316839218,0.3191308975,-0.3130171299,0.0980828851,0.4218278825,0.1543100178,0.3799419999,-0.2579372823,0.0727035478,-0.0288272072,0.3897283077,0.1729896069,-0.2461687326,-0.1826694757,0.1881106496,-0.9209598899,0.4458393455,0.0589725524,-0.039084103,0.0334376767,0.3439892232,0.053749986,0.0531443954,0.2887405157,-0.0962834656,0.1624656022,-0.0415589474,-0.3789632022,-0.3341313899,-0.3106013536,0.4049323797,0.1494320929,-0.5924777985,0.289875567,0.038488958,-0.0501692854,-0.1049868017,-0.2132235318,0.1590006202,-0.1347794384,0.4179118574,-0.1402909905,0.4938715398,-0.027379442,-0.0191062782,-0.2152624726,-0.1111690998,-0.2957692444,0.2690999806,-0.1461151987,0.0767935514,0.1251032054,0.0760321543,-0.298789531,0.3234610856,0.1469838619,-0.2812111974,-0.4066532552,-0.0215115547,0.0377548337,0.0922570005,0.0601231754,0.3981387019,-0.1498904824,0.2334185094,-0.1931376159,-0.3373351097,0.5052189827,-0.3257136941,-0.1926412582,-0.2741817236,0.2520015538,0.1326462328,0.0052120443,-0.640047133,0.0537735224,0.3680356443,-0.1895976514,-0.1688360572,0.1406750828,-0.2344212979,0.4075179398,0.0046436405,0.0375304818,0.0067876321,-0.1305246651,0.1636307687,-0.2550606132]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1964","title":"Datasets.py function load_dataset does not match squad dataset","comments":"I'm glad you managed to fix run_qa.py for your case :)\r\n\r\nRegarding the checksum error, I'm not able to reproduce on my side.\r\nThis errors says that the downloaded file doesn't match the expected file.\r\n\r\nCould you try running this and let me know if you get the same output as me ?\r\n```python\r\nfrom datasets.utils.info_utils import get_size_checksum_dict\r\nfrom datasets import cached_path\r\n\r\nget_size_checksum_dict(cached_path(\"https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json\"))\r\n# {'num_bytes': 30288272, 'checksum': '3527663986b8295af4f7fcdff1ba1ff3f72d07d61a20f487cb238a6ef92fd955'}\r\n```","body":"### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!","comment_length":69,"text":"Datasets.py function load_dataset does not match squad dataset\n### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!\nI'm glad you managed to fix run_qa.py for your case :)\r\n\r\nRegarding the checksum error, I'm not able to reproduce on my side.\r\nThis errors says that the downloaded file doesn't match the expected file.\r\n\r\nCould you try running this and let me know if you get the same output as me ?\r\n```python\r\nfrom datasets.utils.info_utils import get_size_checksum_dict\r\nfrom datasets import cached_path\r\n\r\nget_size_checksum_dict(cached_path(\"https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json\"))\r\n# {'num_bytes': 30288272, 'checksum': '3527663986b8295af4f7fcdff1ba1ff3f72d07d61a20f487cb238a6ef92fd955'}\r\n```","embeddings":[-0.3537200987,0.0625078902,0.0333654545,0.3963988423,0.5554530621,0.0057210135,0.5457583666,0.3404688239,-0.0983101875,-0.1216232106,-0.1650660336,0.4508374333,0.1084147692,-0.0774192885,0.1859294176,0.2271618843,-0.1242239624,-0.0123368138,-0.1461081356,-0.2893434763,-0.1109781116,0.1282911897,-0.0823139772,0.2603987455,-0.1788990498,0.0166908093,0.2006343901,0.2723646462,-0.0037961667,-0.2199688405,0.3428422213,-0.2606053054,0.2361012548,0.5758861899,-0.0001148581,0.1132558137,-0.0115434257,-0.3397289217,-0.5327045321,-0.2002218962,-0.1021837071,-0.3757722974,0.1486449242,-0.2605553567,-0.1595326066,0.093737483,0.0203164779,-0.4645369053,0.3544548452,0.4883137345,0.213982597,0.2522563636,-0.1001010016,0.0775510594,0.1142548099,-0.0009760508,-0.06950894,0.1360990405,0.3347509801,-0.0582686849,0.0025541722,0.1724455059,0.0828086361,0.0568236373,0.2703095675,0.0064037591,0.199115023,-0.316155225,0.0564422533,0.3177836239,0.4627765119,-0.4630158544,-0.334705174,-0.2631344795,0.133104369,-0.2565729618,0.1320165247,0.0672778934,0.0511586368,-0.0424272642,0.0477764271,0.0376646332,0.1114287302,0.0368287303,0.0068916786,0.1999943703,0.0288672037,0.2383961529,-0.201539591,-0.1135483086,0.1148922518,-0.083149083,0.0818328112,0.392493546,-0.828517139,-0.0814113691,-0.2042951286,-0.1421720982,0.2562874556,0.0819008574,0.0386924408,-0.0269711483,0.148668915,0.1044574156,0.1687794775,0.3692035675,0.3052028716,0.2674016058,-0.0663638264,0.0668205246,-0.3717198968,0.2254103124,-0.3225816488,-0.1618057936,0.0960089266,0.0835877731,0.0062298947,-0.1927381605,-0.5517169237,0.1761802584,0.1544087082,0.0026421051,0.0974367931,0.324791342,0.0759165287,0.3802291751,0.0263226684,0.1048553437,-0.294623524,-0.3704161942,-0.2408784777,0.0330561139,-0.0880101547,0.0268361289,0.0183388889,-0.0356007777,0.3227834404,-0.1957616955,0.2114947587,-0.1790799946,0.1150045693,-0.2556711733,-0.1285898685,-0.0065955352,0.0298101194,0.0322048925,0.1742971241,-0.3730719388,-0.0482525341,0.1222925037,-0.3712684512,-0.0336521827,-0.1820591837,0.1871163845,-0.1391085088,0.0208789185,-0.2373687923,0.1698867232,0.3073554337,-0.1878401935,0.1063959599,-0.3961229026,-0.1664000899,-0.2122270316,0.3151646554,0.3882876039,-0.4074572027,-0.2117103338,0.1249395758,-0.0952139571,-0.0381632037,0.1216655001,-0.1516055316,0.1555615067,-0.20622392,-0.0070253811,0.7145847082,-0.8443276286,-0.4243383706,0.1099728644,-0.3059237599,-0.0194989499,-0.0710134581,0.066407755,0.1930091679,0.0147336535,0.0827305093,0.5066136718,-0.0362928361,0.0914031267,-0.1724580824,-0.2712134719,0.2526653707,0.1708515584,0.0318142921,-0.0380649008,0.0388377123,0.4989839792,0.332213819,0.1426163316,0.1112044156,-0.0299585946,-0.0359400138,-0.0821635574,0.114605628,-0.1481092274,-0.6798553467,0.1638657302,-0.1644516885,0.0439888574,0.2410787791,-0.1508302242,-0.2135134637,-0.0917369127,-0.4252310395,-0.1023878902,0.0743926764,0.1495984495,-0.0585200973,-0.1226450801,-0.159884721,0.3514788449,-0.2154513896,0.2284579277,-0.4849060774,-0.0965390429,-0.0637453571,-0.0362155363,0.0537013672,0.3318411708,0.087192513,-0.0709578693,-0.105753161,0.497995466,0.0180258732,0.1023961678,0.2883635163,-0.2016443759,0.0996336117,-0.0990453884,0.0020643868,0.2259092182,0.1946656704,-0.0812613964,0.0224434528,0.1418915838,0.1574391872,0.0746554807,-0.0048610461,-0.1536109,0.139052406,0.0118647469,0.0111010959,0.0971629173,0.0520384163,0.2940087616,0.5483121872,0.1111362055,-0.1868935674,0.1310922652,0.3315558434,-0.0650450215,0.2757264376,-0.0043974128,-0.171322912,-0.0783392414,0.1010062397,0.178945601,0.5017948747,-0.00362671,-0.1863486022,-0.1349211335,-0.063677758,-0.0468234979,0.0711337104,-0.1824291945,0.255372107,0.2577778697,0.1839800775,0.0547975115,-0.1531583816,-0.0497436859,0.0207577217,0.2905252874,-0.3741133809,0.0461585224,-0.1272047311,0.0979508162,-0.3777190447,0.133810699,-0.3502556384,-0.1833577752,-0.135731101,-0.1108105704,0.309304595,0.2101667076,-0.3485357165,0.1555947661,0.2402642965,-0.5443775654,0.1576044559,0.1009899378,-0.3830037117,0.0529526137,0.1205421314,-0.0696215704,-0.0969110429,-0.1528277695,0.0173257664,-0.1201020107,-0.168004334,0.1868712455,-0.0529078133,0.576833725,-0.0123318527,0.1703878939,-0.0075948159,-0.0680116788,0.166756928,-0.1843080521,-0.1167619601,-0.1657404155,-0.1121703759,0.0364880748,-0.0776862428,-0.622492671,-0.4673993886,-0.2988173664,-0.1805043668,0.0007811155,0.2649473548,0.0936486349,0.0102780862,0.3407399952,-0.0355759114,0.0361259319,-0.4035873711,-0.2818406522,0.2556055784,0.0628466681,-0.2947934866,-0.0994649678,0.0628204569,0.2072016597,-0.0956361443,-0.3958399296,-0.2739957571,0.2307893485,-0.0112170214,-0.0910095647,0.0472379848,0.1151574999,-0.1613552123,0.1061526462,-0.2067129463,-0.5061612129,0.1450342834,0.1470152438,0.6205662489,-0.1660099775,0.2980054021,-0.0526382029,0.6775237918,0.1890775263,-0.2280175686,0.2107886672,-0.2105231285,0.3032869399,0.0035384649,-0.2769274116,0.3039433956,0.0719058588,0.0459164828,0.3248111308,0.0534212366,-0.0166754816,-0.0613191389,0.3597286642,-0.1392308474,-0.1571274251,-0.0605640151,-0.0752931461,-0.2700310051,0.0741321221,0.3065279722,-0.1772015244,-0.1673002541,-0.1517760754,0.4453921914,-0.0859651938,0.2577697635,-0.6791188717,-0.0043502864,-0.1259898394,0.286806494,-0.1449967176,0.6176644564,-0.0544948392,-0.185738191,-0.0650190338,0.0791004077,0.4798837304,-0.0459979139,-0.0093406076,-0.0361170694,0.2946531475,-0.6179642081,-0.1878311783,0.177603066,0.1685312092,0.0936853364,0.3528563976,-0.1821606308,-0.0771105886,0.1613401026,0.199181214,-0.1662609279,-0.0170077868,-0.3835242391,-0.2073016763,-0.3429889977,-0.0511813834,-0.276417762,0.1328058094,0.1391488761,0.1173347458,0.2389433235,-0.1128840968,0.3738936782,0.1367881894,0.3835945725,0.1973970532,0.1353370249,0.2334738374,0.2617393732,-0.1502471417,0.4892303646,-0.4564051628,-0.3000645041,0.0343399867,-0.0737037808,0.1476985812,0.3775250018,-0.1436186284,0.0837649927,-0.3207883835,-0.0862554312,0.275416106,0.1629421562,0.2510092556,0.0771014914,-0.1133952439,-0.4098265171,0.1798558235,-0.112034753,-0.0426193438,0.2145652771,-0.4636399746,-0.2888949513,0.0964138806,-0.1039724872,0.8939831853,-0.0649739653,0.1889764965,0.3377411067,0.163262248,0.2822509706,-0.0774445832,0.1654775292,-0.4589556456,-0.0782304406,0.0744747519,-0.1392124891,-0.0647865608,0.3663628101,0.0141666662,0.5146529078,-0.2688012421,0.3116567731,0.0540146455,0.5210416317,-0.1337990165,-0.010874562,-0.3472368717,0.1949188709,-0.1809049547,0.2802450955,-0.1203895137,0.0054683955,-0.3029422462,-0.1379414052,-0.131941393,0.2274260819,-0.5376272798,0.1669765413,0.0934275091,-0.3486429155,0.0191773493,0.5643280745,-0.1002912,-0.0820765272,-0.1084556207,0.353082329,0.0207632296,0.3046307564,0.0768876597,0.0457809493,0.4243014753,-0.0168561712,-0.0948249623,0.0720392466,-0.0042471443,-0.3094080985,-0.2512997389,0.040815454,0.3062185347,-0.2756883502,-0.1972322911,-0.1188386902,0.0159458686,-0.0596131496,0.1198260859,0.0611222945,-0.0966570526,0.2068720311,0.1296120435,-0.4074319601,0.0401298888,0.3414311111,0.0581959672,0.003053996,0.6670231223,0.0741890967,0.0704765096,-0.1949026883,0.1497739553,0.301173836,-0.2336236984,-0.039200563,-0.2000790834,-0.2490169704,0.0004849006,-0.1210952997,0.2407468706,-0.0176706649,0.040279977,-0.4578694999,-0.4342662692,0.031510029,-0.0630608797,0.0092651313,0.0782028437,-0.0439870693,0.1450648308,0.0453661457,-0.2561044693,0.2107708305,-0.1966145039,0.2120202482,0.0993529782,-0.0401501283,0.249724865,-0.0228080153,0.0503343157,-0.0272150729,-0.3081597984,-0.1973445117,-0.12951231,0.1182673052,0.0220857915,-0.015638737,-0.1738801301,-0.2943677902,-0.1543195397,-0.247476548,-0.0346161649,0.0881620198,0.0274494998,0.0142343817,-0.0534722954,0.0265387595,0.0657878071,-0.0739742294,0.0375893079,0.0440010689,0.2452987581,0.1012571603,0.047981102,-0.0090396414,-0.5060158372,0.1013093889,0.1878549755,0.38351354,0.0299073067,-0.1631867141,-0.038469024,0.2729744315,0.0583594814,0.4147221148,-0.4566241503,-0.0442090221,-0.0886465684,0.143432647,-0.2171788365,0.0970659181,0.3625798225,-0.1833094358,-0.0155572481,0.2735645771,0.1266646832,0.189718172,0.0217407867,0.1068987325,0.5871457458,-0.0874021053,0.1646409184,0.2900642157,-0.2160798758,0.2205359042,0.3349782526,-0.0837548003,-0.0087310392,0.3994481862,-0.1929351389,0.3431117833,-0.0790344253,-0.1280991137,0.2440442443,-0.2488018423,0.0799286142,0.402431488,-0.1566050351,0.1699330658,0.0469692349,0.539581418,-0.4212684333,-0.1374428123,-0.0382384658,-0.1659311652,-0.1484349817,-0.0598560758,0.0648372024,-0.1156122908,-0.1001708731,-0.2388545126,-0.1484500319,-0.3574163914,-0.0503093265,-0.003967029,-0.3478863537,-0.0714054331,-0.0686984584,0.3550385833,-0.1964205354,-0.1058571562,0.3332430124,-0.1416485012,0.0370896682,0.3267581463,0.2570812404,0.195392713,0.445263952,0.0781018436,-0.1183174253,0.3155586123,0.0447255895,0.014626109,0.1478066146,-0.1330287158,-0.1975685507,0.3923157752,0.1372191608,-0.0548008196,-0.2756023109,0.3441493809,-0.066534847,-0.2472360432,0.2151201516,-0.2440628409,-0.0380831957,-0.5316839218,0.3191308975,-0.3130171299,0.0980828851,0.4218278825,0.1543100178,0.3799419999,-0.2579372823,0.0727035478,-0.0288272072,0.3897283077,0.1729896069,-0.2461687326,-0.1826694757,0.1881106496,-0.9209598899,0.4458393455,0.0589725524,-0.039084103,0.0334376767,0.3439892232,0.053749986,0.0531443954,0.2887405157,-0.0962834656,0.1624656022,-0.0415589474,-0.3789632022,-0.3341313899,-0.3106013536,0.4049323797,0.1494320929,-0.5924777985,0.289875567,0.038488958,-0.0501692854,-0.1049868017,-0.2132235318,0.1590006202,-0.1347794384,0.4179118574,-0.1402909905,0.4938715398,-0.027379442,-0.0191062782,-0.2152624726,-0.1111690998,-0.2957692444,0.2690999806,-0.1461151987,0.0767935514,0.1251032054,0.0760321543,-0.298789531,0.3234610856,0.1469838619,-0.2812111974,-0.4066532552,-0.0215115547,0.0377548337,0.0922570005,0.0601231754,0.3981387019,-0.1498904824,0.2334185094,-0.1931376159,-0.3373351097,0.5052189827,-0.3257136941,-0.1926412582,-0.2741817236,0.2520015538,0.1326462328,0.0052120443,-0.640047133,0.0537735224,0.3680356443,-0.1895976514,-0.1688360572,0.1406750828,-0.2344212979,0.4075179398,0.0046436405,0.0375304818,0.0067876321,-0.1305246651,0.1636307687,-0.2550606132]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1964","title":"Datasets.py function load_dataset does not match squad dataset","comments":"I run the code,and it show below:\r\n```\r\n>>> from datasets.utils.info_utils import get_size_checksum_dict\r\n>>> from datasets import cached_path\r\n>>> get_size_checksum_dict(cached_path(\"https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json\"))\r\nDownloading: 30.3MB [04:13, 120kB\/s]\r\n{'num_bytes': 30288272, 'checksum': '3527663986b8295af4f7fcdff1ba1ff3f72d07d61a20f487cb238a6ef92fd955'}\r\n```","body":"### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!","comment_length":29,"text":"Datasets.py function load_dataset does not match squad dataset\n### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!\nI run the code,and it show below:\r\n```\r\n>>> from datasets.utils.info_utils import get_size_checksum_dict\r\n>>> from datasets import cached_path\r\n>>> get_size_checksum_dict(cached_path(\"https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json\"))\r\nDownloading: 30.3MB [04:13, 120kB\/s]\r\n{'num_bytes': 30288272, 'checksum': '3527663986b8295af4f7fcdff1ba1ff3f72d07d61a20f487cb238a6ef92fd955'}\r\n```","embeddings":[-0.3537200987,0.0625078902,0.0333654545,0.3963988423,0.5554530621,0.0057210135,0.5457583666,0.3404688239,-0.0983101875,-0.1216232106,-0.1650660336,0.4508374333,0.1084147692,-0.0774192885,0.1859294176,0.2271618843,-0.1242239624,-0.0123368138,-0.1461081356,-0.2893434763,-0.1109781116,0.1282911897,-0.0823139772,0.2603987455,-0.1788990498,0.0166908093,0.2006343901,0.2723646462,-0.0037961667,-0.2199688405,0.3428422213,-0.2606053054,0.2361012548,0.5758861899,-0.0001148581,0.1132558137,-0.0115434257,-0.3397289217,-0.5327045321,-0.2002218962,-0.1021837071,-0.3757722974,0.1486449242,-0.2605553567,-0.1595326066,0.093737483,0.0203164779,-0.4645369053,0.3544548452,0.4883137345,0.213982597,0.2522563636,-0.1001010016,0.0775510594,0.1142548099,-0.0009760508,-0.06950894,0.1360990405,0.3347509801,-0.0582686849,0.0025541722,0.1724455059,0.0828086361,0.0568236373,0.2703095675,0.0064037591,0.199115023,-0.316155225,0.0564422533,0.3177836239,0.4627765119,-0.4630158544,-0.334705174,-0.2631344795,0.133104369,-0.2565729618,0.1320165247,0.0672778934,0.0511586368,-0.0424272642,0.0477764271,0.0376646332,0.1114287302,0.0368287303,0.0068916786,0.1999943703,0.0288672037,0.2383961529,-0.201539591,-0.1135483086,0.1148922518,-0.083149083,0.0818328112,0.392493546,-0.828517139,-0.0814113691,-0.2042951286,-0.1421720982,0.2562874556,0.0819008574,0.0386924408,-0.0269711483,0.148668915,0.1044574156,0.1687794775,0.3692035675,0.3052028716,0.2674016058,-0.0663638264,0.0668205246,-0.3717198968,0.2254103124,-0.3225816488,-0.1618057936,0.0960089266,0.0835877731,0.0062298947,-0.1927381605,-0.5517169237,0.1761802584,0.1544087082,0.0026421051,0.0974367931,0.324791342,0.0759165287,0.3802291751,0.0263226684,0.1048553437,-0.294623524,-0.3704161942,-0.2408784777,0.0330561139,-0.0880101547,0.0268361289,0.0183388889,-0.0356007777,0.3227834404,-0.1957616955,0.2114947587,-0.1790799946,0.1150045693,-0.2556711733,-0.1285898685,-0.0065955352,0.0298101194,0.0322048925,0.1742971241,-0.3730719388,-0.0482525341,0.1222925037,-0.3712684512,-0.0336521827,-0.1820591837,0.1871163845,-0.1391085088,0.0208789185,-0.2373687923,0.1698867232,0.3073554337,-0.1878401935,0.1063959599,-0.3961229026,-0.1664000899,-0.2122270316,0.3151646554,0.3882876039,-0.4074572027,-0.2117103338,0.1249395758,-0.0952139571,-0.0381632037,0.1216655001,-0.1516055316,0.1555615067,-0.20622392,-0.0070253811,0.7145847082,-0.8443276286,-0.4243383706,0.1099728644,-0.3059237599,-0.0194989499,-0.0710134581,0.066407755,0.1930091679,0.0147336535,0.0827305093,0.5066136718,-0.0362928361,0.0914031267,-0.1724580824,-0.2712134719,0.2526653707,0.1708515584,0.0318142921,-0.0380649008,0.0388377123,0.4989839792,0.332213819,0.1426163316,0.1112044156,-0.0299585946,-0.0359400138,-0.0821635574,0.114605628,-0.1481092274,-0.6798553467,0.1638657302,-0.1644516885,0.0439888574,0.2410787791,-0.1508302242,-0.2135134637,-0.0917369127,-0.4252310395,-0.1023878902,0.0743926764,0.1495984495,-0.0585200973,-0.1226450801,-0.159884721,0.3514788449,-0.2154513896,0.2284579277,-0.4849060774,-0.0965390429,-0.0637453571,-0.0362155363,0.0537013672,0.3318411708,0.087192513,-0.0709578693,-0.105753161,0.497995466,0.0180258732,0.1023961678,0.2883635163,-0.2016443759,0.0996336117,-0.0990453884,0.0020643868,0.2259092182,0.1946656704,-0.0812613964,0.0224434528,0.1418915838,0.1574391872,0.0746554807,-0.0048610461,-0.1536109,0.139052406,0.0118647469,0.0111010959,0.0971629173,0.0520384163,0.2940087616,0.5483121872,0.1111362055,-0.1868935674,0.1310922652,0.3315558434,-0.0650450215,0.2757264376,-0.0043974128,-0.171322912,-0.0783392414,0.1010062397,0.178945601,0.5017948747,-0.00362671,-0.1863486022,-0.1349211335,-0.063677758,-0.0468234979,0.0711337104,-0.1824291945,0.255372107,0.2577778697,0.1839800775,0.0547975115,-0.1531583816,-0.0497436859,0.0207577217,0.2905252874,-0.3741133809,0.0461585224,-0.1272047311,0.0979508162,-0.3777190447,0.133810699,-0.3502556384,-0.1833577752,-0.135731101,-0.1108105704,0.309304595,0.2101667076,-0.3485357165,0.1555947661,0.2402642965,-0.5443775654,0.1576044559,0.1009899378,-0.3830037117,0.0529526137,0.1205421314,-0.0696215704,-0.0969110429,-0.1528277695,0.0173257664,-0.1201020107,-0.168004334,0.1868712455,-0.0529078133,0.576833725,-0.0123318527,0.1703878939,-0.0075948159,-0.0680116788,0.166756928,-0.1843080521,-0.1167619601,-0.1657404155,-0.1121703759,0.0364880748,-0.0776862428,-0.622492671,-0.4673993886,-0.2988173664,-0.1805043668,0.0007811155,0.2649473548,0.0936486349,0.0102780862,0.3407399952,-0.0355759114,0.0361259319,-0.4035873711,-0.2818406522,0.2556055784,0.0628466681,-0.2947934866,-0.0994649678,0.0628204569,0.2072016597,-0.0956361443,-0.3958399296,-0.2739957571,0.2307893485,-0.0112170214,-0.0910095647,0.0472379848,0.1151574999,-0.1613552123,0.1061526462,-0.2067129463,-0.5061612129,0.1450342834,0.1470152438,0.6205662489,-0.1660099775,0.2980054021,-0.0526382029,0.6775237918,0.1890775263,-0.2280175686,0.2107886672,-0.2105231285,0.3032869399,0.0035384649,-0.2769274116,0.3039433956,0.0719058588,0.0459164828,0.3248111308,0.0534212366,-0.0166754816,-0.0613191389,0.3597286642,-0.1392308474,-0.1571274251,-0.0605640151,-0.0752931461,-0.2700310051,0.0741321221,0.3065279722,-0.1772015244,-0.1673002541,-0.1517760754,0.4453921914,-0.0859651938,0.2577697635,-0.6791188717,-0.0043502864,-0.1259898394,0.286806494,-0.1449967176,0.6176644564,-0.0544948392,-0.185738191,-0.0650190338,0.0791004077,0.4798837304,-0.0459979139,-0.0093406076,-0.0361170694,0.2946531475,-0.6179642081,-0.1878311783,0.177603066,0.1685312092,0.0936853364,0.3528563976,-0.1821606308,-0.0771105886,0.1613401026,0.199181214,-0.1662609279,-0.0170077868,-0.3835242391,-0.2073016763,-0.3429889977,-0.0511813834,-0.276417762,0.1328058094,0.1391488761,0.1173347458,0.2389433235,-0.1128840968,0.3738936782,0.1367881894,0.3835945725,0.1973970532,0.1353370249,0.2334738374,0.2617393732,-0.1502471417,0.4892303646,-0.4564051628,-0.3000645041,0.0343399867,-0.0737037808,0.1476985812,0.3775250018,-0.1436186284,0.0837649927,-0.3207883835,-0.0862554312,0.275416106,0.1629421562,0.2510092556,0.0771014914,-0.1133952439,-0.4098265171,0.1798558235,-0.112034753,-0.0426193438,0.2145652771,-0.4636399746,-0.2888949513,0.0964138806,-0.1039724872,0.8939831853,-0.0649739653,0.1889764965,0.3377411067,0.163262248,0.2822509706,-0.0774445832,0.1654775292,-0.4589556456,-0.0782304406,0.0744747519,-0.1392124891,-0.0647865608,0.3663628101,0.0141666662,0.5146529078,-0.2688012421,0.3116567731,0.0540146455,0.5210416317,-0.1337990165,-0.010874562,-0.3472368717,0.1949188709,-0.1809049547,0.2802450955,-0.1203895137,0.0054683955,-0.3029422462,-0.1379414052,-0.131941393,0.2274260819,-0.5376272798,0.1669765413,0.0934275091,-0.3486429155,0.0191773493,0.5643280745,-0.1002912,-0.0820765272,-0.1084556207,0.353082329,0.0207632296,0.3046307564,0.0768876597,0.0457809493,0.4243014753,-0.0168561712,-0.0948249623,0.0720392466,-0.0042471443,-0.3094080985,-0.2512997389,0.040815454,0.3062185347,-0.2756883502,-0.1972322911,-0.1188386902,0.0159458686,-0.0596131496,0.1198260859,0.0611222945,-0.0966570526,0.2068720311,0.1296120435,-0.4074319601,0.0401298888,0.3414311111,0.0581959672,0.003053996,0.6670231223,0.0741890967,0.0704765096,-0.1949026883,0.1497739553,0.301173836,-0.2336236984,-0.039200563,-0.2000790834,-0.2490169704,0.0004849006,-0.1210952997,0.2407468706,-0.0176706649,0.040279977,-0.4578694999,-0.4342662692,0.031510029,-0.0630608797,0.0092651313,0.0782028437,-0.0439870693,0.1450648308,0.0453661457,-0.2561044693,0.2107708305,-0.1966145039,0.2120202482,0.0993529782,-0.0401501283,0.249724865,-0.0228080153,0.0503343157,-0.0272150729,-0.3081597984,-0.1973445117,-0.12951231,0.1182673052,0.0220857915,-0.015638737,-0.1738801301,-0.2943677902,-0.1543195397,-0.247476548,-0.0346161649,0.0881620198,0.0274494998,0.0142343817,-0.0534722954,0.0265387595,0.0657878071,-0.0739742294,0.0375893079,0.0440010689,0.2452987581,0.1012571603,0.047981102,-0.0090396414,-0.5060158372,0.1013093889,0.1878549755,0.38351354,0.0299073067,-0.1631867141,-0.038469024,0.2729744315,0.0583594814,0.4147221148,-0.4566241503,-0.0442090221,-0.0886465684,0.143432647,-0.2171788365,0.0970659181,0.3625798225,-0.1833094358,-0.0155572481,0.2735645771,0.1266646832,0.189718172,0.0217407867,0.1068987325,0.5871457458,-0.0874021053,0.1646409184,0.2900642157,-0.2160798758,0.2205359042,0.3349782526,-0.0837548003,-0.0087310392,0.3994481862,-0.1929351389,0.3431117833,-0.0790344253,-0.1280991137,0.2440442443,-0.2488018423,0.0799286142,0.402431488,-0.1566050351,0.1699330658,0.0469692349,0.539581418,-0.4212684333,-0.1374428123,-0.0382384658,-0.1659311652,-0.1484349817,-0.0598560758,0.0648372024,-0.1156122908,-0.1001708731,-0.2388545126,-0.1484500319,-0.3574163914,-0.0503093265,-0.003967029,-0.3478863537,-0.0714054331,-0.0686984584,0.3550385833,-0.1964205354,-0.1058571562,0.3332430124,-0.1416485012,0.0370896682,0.3267581463,0.2570812404,0.195392713,0.445263952,0.0781018436,-0.1183174253,0.3155586123,0.0447255895,0.014626109,0.1478066146,-0.1330287158,-0.1975685507,0.3923157752,0.1372191608,-0.0548008196,-0.2756023109,0.3441493809,-0.066534847,-0.2472360432,0.2151201516,-0.2440628409,-0.0380831957,-0.5316839218,0.3191308975,-0.3130171299,0.0980828851,0.4218278825,0.1543100178,0.3799419999,-0.2579372823,0.0727035478,-0.0288272072,0.3897283077,0.1729896069,-0.2461687326,-0.1826694757,0.1881106496,-0.9209598899,0.4458393455,0.0589725524,-0.039084103,0.0334376767,0.3439892232,0.053749986,0.0531443954,0.2887405157,-0.0962834656,0.1624656022,-0.0415589474,-0.3789632022,-0.3341313899,-0.3106013536,0.4049323797,0.1494320929,-0.5924777985,0.289875567,0.038488958,-0.0501692854,-0.1049868017,-0.2132235318,0.1590006202,-0.1347794384,0.4179118574,-0.1402909905,0.4938715398,-0.027379442,-0.0191062782,-0.2152624726,-0.1111690998,-0.2957692444,0.2690999806,-0.1461151987,0.0767935514,0.1251032054,0.0760321543,-0.298789531,0.3234610856,0.1469838619,-0.2812111974,-0.4066532552,-0.0215115547,0.0377548337,0.0922570005,0.0601231754,0.3981387019,-0.1498904824,0.2334185094,-0.1931376159,-0.3373351097,0.5052189827,-0.3257136941,-0.1926412582,-0.2741817236,0.2520015538,0.1326462328,0.0052120443,-0.640047133,0.0537735224,0.3680356443,-0.1895976514,-0.1688360572,0.1406750828,-0.2344212979,0.4075179398,0.0046436405,0.0375304818,0.0067876321,-0.1305246651,0.1636307687,-0.2550606132]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1964","title":"Datasets.py function load_dataset does not match squad dataset","comments":"Alright ! So in this case redownloading the file with `download_mode=\"force_redownload\"` should fix it. Can you try using `download_mode=\"force_redownload\"` again ?\r\n\r\nNot sure why it didn't work for you the first time though :\/","body":"### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!","comment_length":34,"text":"Datasets.py function load_dataset does not match squad dataset\n### 1 When I try to train lxmert,and follow the code in README that --dataset name:\r\n```shell \r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --dataset_name squad --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nthe bug is that:\r\n```\r\nDownloading and preparing dataset squad\/plain_text (download: 33.51 MiB, generated: 85.75 MiB, post-processed: Unknown size, total: 119.27 MiB) to \/home2\/zhenggo1\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/4c81550d83a2ac7c7ce23783bd8ff36642800e6633c1f18417fb58c3ff50cdd7...\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 217, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 746, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 573, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 633, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home2\/zhenggo1\/anaconda3\/envs\/lpot\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/rajpurkar.github.io\/SQuAD-explorer\/dataset\/train-v1.1.json']\r\n```\r\nAnd I try to find the [checksum link](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/squad\/dataset_infos.json)\r\n,is the problem plain_text do not have a checksum?\r\n\r\n### 2 When I try to train lxmert,and use local dataset:\r\n```\r\npython examples\/question-answering\/run_qa.py --model_name_or_path unc-nlp\/lxmert-base-uncased --train_file $SQUAD_DIR\/train-v1.1.json --validation_file $SQUAD_DIR\/dev-v1.1.json --do_train --do_eval --per_device_train_batch_size 12 --learning_rate 3e-5 --num_train_epochs 2 --max_seq_length 384 --doc_stride 128 --output_dir \/home2\/zhenggo1\/checkpoint\/lxmert_squad\r\n```\r\nThe bug is that \r\n```\r\n['title', 'paragraphs']\r\nTraceback (most recent call last):\r\n File \"examples\/question-answering\/run_qa.py\", line 501, in \r\n main()\r\n File \"examples\/question-answering\/run_qa.py\", line 273, in main\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\nIndexError: list index out of range\r\n```\r\nI print the answer_column_name and find that local squad dataset need the package datasets to preprocessing so that the code below can work:\r\n```\r\nif training_args.do_train:\r\n column_names = datasets[\"train\"].column_names\r\n else:\r\n column_names = datasets[\"validation\"].column_names\r\n print(datasets[\"train\"].column_names)\r\n question_column_name = \"question\" if \"question\" in column_names else column_names[0]\r\n context_column_name = \"context\" if \"context\" in column_names else column_names[1]\r\n answer_column_name = \"answers\" if \"answers\" in column_names else column_names[2]\r\n``` \r\n## Please tell me how to fix the bug,thks a lot!\nAlright ! So in this case redownloading the file with `download_mode=\"force_redownload\"` should fix it. Can you try using `download_mode=\"force_redownload\"` again ?\r\n\r\nNot sure why it didn't work for you the first time though :\/","embeddings":[-0.3537200987,0.0625078902,0.0333654545,0.3963988423,0.5554530621,0.0057210135,0.5457583666,0.3404688239,-0.0983101875,-0.1216232106,-0.1650660336,0.4508374333,0.1084147692,-0.0774192885,0.1859294176,0.2271618843,-0.1242239624,-0.0123368138,-0.1461081356,-0.2893434763,-0.1109781116,0.1282911897,-0.0823139772,0.2603987455,-0.1788990498,0.0166908093,0.2006343901,0.2723646462,-0.0037961667,-0.2199688405,0.3428422213,-0.2606053054,0.2361012548,0.5758861899,-0.0001148581,0.1132558137,-0.0115434257,-0.3397289217,-0.5327045321,-0.2002218962,-0.1021837071,-0.3757722974,0.1486449242,-0.2605553567,-0.1595326066,0.093737483,0.0203164779,-0.4645369053,0.3544548452,0.4883137345,0.213982597,0.2522563636,-0.1001010016,0.0775510594,0.1142548099,-0.0009760508,-0.06950894,0.1360990405,0.3347509801,-0.0582686849,0.0025541722,0.1724455059,0.0828086361,0.0568236373,0.2703095675,0.0064037591,0.199115023,-0.316155225,0.0564422533,0.3177836239,0.4627765119,-0.4630158544,-0.334705174,-0.2631344795,0.133104369,-0.2565729618,0.1320165247,0.0672778934,0.0511586368,-0.0424272642,0.0477764271,0.0376646332,0.1114287302,0.0368287303,0.0068916786,0.1999943703,0.0288672037,0.2383961529,-0.201539591,-0.1135483086,0.1148922518,-0.083149083,0.0818328112,0.392493546,-0.828517139,-0.0814113691,-0.2042951286,-0.1421720982,0.2562874556,0.0819008574,0.0386924408,-0.0269711483,0.148668915,0.1044574156,0.1687794775,0.3692035675,0.3052028716,0.2674016058,-0.0663638264,0.0668205246,-0.3717198968,0.2254103124,-0.3225816488,-0.1618057936,0.0960089266,0.0835877731,0.0062298947,-0.1927381605,-0.5517169237,0.1761802584,0.1544087082,0.0026421051,0.0974367931,0.324791342,0.0759165287,0.3802291751,0.0263226684,0.1048553437,-0.294623524,-0.3704161942,-0.2408784777,0.0330561139,-0.0880101547,0.0268361289,0.0183388889,-0.0356007777,0.3227834404,-0.1957616955,0.2114947587,-0.1790799946,0.1150045693,-0.2556711733,-0.1285898685,-0.0065955352,0.0298101194,0.0322048925,0.1742971241,-0.3730719388,-0.0482525341,0.1222925037,-0.3712684512,-0.0336521827,-0.1820591837,0.1871163845,-0.1391085088,0.0208789185,-0.2373687923,0.1698867232,0.3073554337,-0.1878401935,0.1063959599,-0.3961229026,-0.1664000899,-0.2122270316,0.3151646554,0.3882876039,-0.4074572027,-0.2117103338,0.1249395758,-0.0952139571,-0.0381632037,0.1216655001,-0.1516055316,0.1555615067,-0.20622392,-0.0070253811,0.7145847082,-0.8443276286,-0.4243383706,0.1099728644,-0.3059237599,-0.0194989499,-0.0710134581,0.066407755,0.1930091679,0.0147336535,0.0827305093,0.5066136718,-0.0362928361,0.0914031267,-0.1724580824,-0.2712134719,0.2526653707,0.1708515584,0.0318142921,-0.0380649008,0.0388377123,0.4989839792,0.332213819,0.1426163316,0.1112044156,-0.0299585946,-0.0359400138,-0.0821635574,0.114605628,-0.1481092274,-0.6798553467,0.1638657302,-0.1644516885,0.0439888574,0.2410787791,-0.1508302242,-0.2135134637,-0.0917369127,-0.4252310395,-0.1023878902,0.0743926764,0.1495984495,-0.0585200973,-0.1226450801,-0.159884721,0.3514788449,-0.2154513896,0.2284579277,-0.4849060774,-0.0965390429,-0.0637453571,-0.0362155363,0.0537013672,0.3318411708,0.087192513,-0.0709578693,-0.105753161,0.497995466,0.0180258732,0.1023961678,0.2883635163,-0.2016443759,0.0996336117,-0.0990453884,0.0020643868,0.2259092182,0.1946656704,-0.0812613964,0.0224434528,0.1418915838,0.1574391872,0.0746554807,-0.0048610461,-0.1536109,0.139052406,0.0118647469,0.0111010959,0.0971629173,0.0520384163,0.2940087616,0.5483121872,0.1111362055,-0.1868935674,0.1310922652,0.3315558434,-0.0650450215,0.2757264376,-0.0043974128,-0.171322912,-0.0783392414,0.1010062397,0.178945601,0.5017948747,-0.00362671,-0.1863486022,-0.1349211335,-0.063677758,-0.0468234979,0.0711337104,-0.1824291945,0.255372107,0.2577778697,0.1839800775,0.0547975115,-0.1531583816,-0.0497436859,0.0207577217,0.2905252874,-0.3741133809,0.0461585224,-0.1272047311,0.0979508162,-0.3777190447,0.133810699,-0.3502556384,-0.1833577752,-0.135731101,-0.1108105704,0.309304595,0.2101667076,-0.3485357165,0.1555947661,0.2402642965,-0.5443775654,0.1576044559,0.1009899378,-0.3830037117,0.0529526137,0.1205421314,-0.0696215704,-0.0969110429,-0.1528277695,0.0173257664,-0.1201020107,-0.168004334,0.1868712455,-0.0529078133,0.576833725,-0.0123318527,0.1703878939,-0.0075948159,-0.0680116788,0.166756928,-0.1843080521,-0.1167619601,-0.1657404155,-0.1121703759,0.0364880748,-0.0776862428,-0.622492671,-0.4673993886,-0.2988173664,-0.1805043668,0.0007811155,0.2649473548,0.0936486349,0.0102780862,0.3407399952,-0.0355759114,0.0361259319,-0.4035873711,-0.2818406522,0.2556055784,0.0628466681,-0.2947934866,-0.0994649678,0.0628204569,0.2072016597,-0.0956361443,-0.3958399296,-0.2739957571,0.2307893485,-0.0112170214,-0.0910095647,0.0472379848,0.1151574999,-0.1613552123,0.1061526462,-0.2067129463,-0.5061612129,0.1450342834,0.1470152438,0.6205662489,-0.1660099775,0.2980054021,-0.0526382029,0.6775237918,0.1890775263,-0.2280175686,0.2107886672,-0.2105231285,0.3032869399,0.0035384649,-0.2769274116,0.3039433956,0.0719058588,0.0459164828,0.3248111308,0.0534212366,-0.0166754816,-0.0613191389,0.3597286642,-0.1392308474,-0.1571274251,-0.0605640151,-0.0752931461,-0.2700310051,0.0741321221,0.3065279722,-0.1772015244,-0.1673002541,-0.1517760754,0.4453921914,-0.0859651938,0.2577697635,-0.6791188717,-0.0043502864,-0.1259898394,0.286806494,-0.1449967176,0.6176644564,-0.0544948392,-0.185738191,-0.0650190338,0.0791004077,0.4798837304,-0.0459979139,-0.0093406076,-0.0361170694,0.2946531475,-0.6179642081,-0.1878311783,0.177603066,0.1685312092,0.0936853364,0.3528563976,-0.1821606308,-0.0771105886,0.1613401026,0.199181214,-0.1662609279,-0.0170077868,-0.3835242391,-0.2073016763,-0.3429889977,-0.0511813834,-0.276417762,0.1328058094,0.1391488761,0.1173347458,0.2389433235,-0.1128840968,0.3738936782,0.1367881894,0.3835945725,0.1973970532,0.1353370249,0.2334738374,0.2617393732,-0.1502471417,0.4892303646,-0.4564051628,-0.3000645041,0.0343399867,-0.0737037808,0.1476985812,0.3775250018,-0.1436186284,0.0837649927,-0.3207883835,-0.0862554312,0.275416106,0.1629421562,0.2510092556,0.0771014914,-0.1133952439,-0.4098265171,0.1798558235,-0.112034753,-0.0426193438,0.2145652771,-0.4636399746,-0.2888949513,0.0964138806,-0.1039724872,0.8939831853,-0.0649739653,0.1889764965,0.3377411067,0.163262248,0.2822509706,-0.0774445832,0.1654775292,-0.4589556456,-0.0782304406,0.0744747519,-0.1392124891,-0.0647865608,0.3663628101,0.0141666662,0.5146529078,-0.2688012421,0.3116567731,0.0540146455,0.5210416317,-0.1337990165,-0.010874562,-0.3472368717,0.1949188709,-0.1809049547,0.2802450955,-0.1203895137,0.0054683955,-0.3029422462,-0.1379414052,-0.131941393,0.2274260819,-0.5376272798,0.1669765413,0.0934275091,-0.3486429155,0.0191773493,0.5643280745,-0.1002912,-0.0820765272,-0.1084556207,0.353082329,0.0207632296,0.3046307564,0.0768876597,0.0457809493,0.4243014753,-0.0168561712,-0.0948249623,0.0720392466,-0.0042471443,-0.3094080985,-0.2512997389,0.040815454,0.3062185347,-0.2756883502,-0.1972322911,-0.1188386902,0.0159458686,-0.0596131496,0.1198260859,0.0611222945,-0.0966570526,0.2068720311,0.1296120435,-0.4074319601,0.0401298888,0.3414311111,0.0581959672,0.003053996,0.6670231223,0.0741890967,0.0704765096,-0.1949026883,0.1497739553,0.301173836,-0.2336236984,-0.039200563,-0.2000790834,-0.2490169704,0.0004849006,-0.1210952997,0.2407468706,-0.0176706649,0.040279977,-0.4578694999,-0.4342662692,0.031510029,-0.0630608797,0.0092651313,0.0782028437,-0.0439870693,0.1450648308,0.0453661457,-0.2561044693,0.2107708305,-0.1966145039,0.2120202482,0.0993529782,-0.0401501283,0.249724865,-0.0228080153,0.0503343157,-0.0272150729,-0.3081597984,-0.1973445117,-0.12951231,0.1182673052,0.0220857915,-0.015638737,-0.1738801301,-0.2943677902,-0.1543195397,-0.247476548,-0.0346161649,0.0881620198,0.0274494998,0.0142343817,-0.0534722954,0.0265387595,0.0657878071,-0.0739742294,0.0375893079,0.0440010689,0.2452987581,0.1012571603,0.047981102,-0.0090396414,-0.5060158372,0.1013093889,0.1878549755,0.38351354,0.0299073067,-0.1631867141,-0.038469024,0.2729744315,0.0583594814,0.4147221148,-0.4566241503,-0.0442090221,-0.0886465684,0.143432647,-0.2171788365,0.0970659181,0.3625798225,-0.1833094358,-0.0155572481,0.2735645771,0.1266646832,0.189718172,0.0217407867,0.1068987325,0.5871457458,-0.0874021053,0.1646409184,0.2900642157,-0.2160798758,0.2205359042,0.3349782526,-0.0837548003,-0.0087310392,0.3994481862,-0.1929351389,0.3431117833,-0.0790344253,-0.1280991137,0.2440442443,-0.2488018423,0.0799286142,0.402431488,-0.1566050351,0.1699330658,0.0469692349,0.539581418,-0.4212684333,-0.1374428123,-0.0382384658,-0.1659311652,-0.1484349817,-0.0598560758,0.0648372024,-0.1156122908,-0.1001708731,-0.2388545126,-0.1484500319,-0.3574163914,-0.0503093265,-0.003967029,-0.3478863537,-0.0714054331,-0.0686984584,0.3550385833,-0.1964205354,-0.1058571562,0.3332430124,-0.1416485012,0.0370896682,0.3267581463,0.2570812404,0.195392713,0.445263952,0.0781018436,-0.1183174253,0.3155586123,0.0447255895,0.014626109,0.1478066146,-0.1330287158,-0.1975685507,0.3923157752,0.1372191608,-0.0548008196,-0.2756023109,0.3441493809,-0.066534847,-0.2472360432,0.2151201516,-0.2440628409,-0.0380831957,-0.5316839218,0.3191308975,-0.3130171299,0.0980828851,0.4218278825,0.1543100178,0.3799419999,-0.2579372823,0.0727035478,-0.0288272072,0.3897283077,0.1729896069,-0.2461687326,-0.1826694757,0.1881106496,-0.9209598899,0.4458393455,0.0589725524,-0.039084103,0.0334376767,0.3439892232,0.053749986,0.0531443954,0.2887405157,-0.0962834656,0.1624656022,-0.0415589474,-0.3789632022,-0.3341313899,-0.3106013536,0.4049323797,0.1494320929,-0.5924777985,0.289875567,0.038488958,-0.0501692854,-0.1049868017,-0.2132235318,0.1590006202,-0.1347794384,0.4179118574,-0.1402909905,0.4938715398,-0.027379442,-0.0191062782,-0.2152624726,-0.1111690998,-0.2957692444,0.2690999806,-0.1461151987,0.0767935514,0.1251032054,0.0760321543,-0.298789531,0.3234610856,0.1469838619,-0.2812111974,-0.4066532552,-0.0215115547,0.0377548337,0.0922570005,0.0601231754,0.3981387019,-0.1498904824,0.2334185094,-0.1931376159,-0.3373351097,0.5052189827,-0.3257136941,-0.1926412582,-0.2741817236,0.2520015538,0.1326462328,0.0052120443,-0.640047133,0.0537735224,0.3680356443,-0.1895976514,-0.1688360572,0.1406750828,-0.2344212979,0.4075179398,0.0046436405,0.0375304818,0.0067876321,-0.1305246651,0.1636307687,-0.2550606132]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1963","title":"bug in SNLI dataset ","comments":"Hi ! The labels -1 correspond to the examples without gold labels in the original snli dataset.\r\nFeel free to remove these examples if you don't need them by using\r\n```python\r\ndata = data.filter(lambda x: x[\"label\"] != -1)\r\n```","body":"Hi\r\nThere is label of -1 in train set of SNLI dataset, please find the code below:\r\n\r\n```\r\nimport numpy as np \r\nimport datasets \r\ndata = datasets.load_dataset(\"snli\")[\"train\"]\r\nlabels = []\r\nfor d in data:\r\n labels.append(d[\"label\"])\r\nprint(np.unique(labels))\r\n```\r\n\r\nand results:\r\n\r\n`[-1 0 1 2]`\r\n\r\nversion of datasets used:\r\n`datasets 1.2.1 \r\n`\r\n\r\nthanks for your help. @lhoestq ","comment_length":39,"text":"bug in SNLI dataset \nHi\r\nThere is label of -1 in train set of SNLI dataset, please find the code below:\r\n\r\n```\r\nimport numpy as np \r\nimport datasets \r\ndata = datasets.load_dataset(\"snli\")[\"train\"]\r\nlabels = []\r\nfor d in data:\r\n labels.append(d[\"label\"])\r\nprint(np.unique(labels))\r\n```\r\n\r\nand results:\r\n\r\n`[-1 0 1 2]`\r\n\r\nversion of datasets used:\r\n`datasets 1.2.1 \r\n`\r\n\r\nthanks for your help. @lhoestq \nHi ! The labels -1 correspond to the examples without gold labels in the original snli dataset.\r\nFeel free to remove these examples if you don't need them by using\r\n```python\r\ndata = data.filter(lambda x: x[\"label\"] != -1)\r\n```","embeddings":[0.1758222431,-0.2825255096,-0.1035732776,0.3496335149,0.2219778597,0.0471981168,0.3874210715,0.1183437407,0.0808190703,0.2969244123,-0.2027283907,0.6121100187,-0.1427199841,0.0958901942,0.0018212209,-0.0009762677,0.2330546677,0.3379025459,0.2097955346,-0.4966840744,-0.2359668463,-0.0235047974,-0.3817799091,0.2652831078,-0.2869722247,0.1784489006,-0.0155612463,-0.1333378255,0.1334049553,-0.4898706973,0.1604635864,-0.0986117944,0.1226140335,0.2325982898,-0.0001058666,-0.2743193209,0.1129734814,0.0678980052,-0.3696534932,-0.2605839968,-0.1101465896,-0.0220090337,-0.146892488,-0.1628938764,-0.2096339464,-0.1724426448,0.0142517593,0.0935407504,0.1124736294,0.2662241459,0.2374289483,0.0673883259,-0.1264111251,0.1347338408,0.2391382456,-0.3994213939,0.1405956298,0.1139144152,0.0496283025,0.049375765,0.4056082666,0.6512392759,-0.0620471761,-0.043695204,0.1137125939,0.090373829,0.127294898,-0.3254041672,-0.0787117556,0.3404750824,0.1091918424,-0.3263964653,-0.4188061655,0.0663319081,0.1710607111,-0.4623311758,-0.0283233002,0.2538726032,0.140697822,0.2034579664,0.0178657025,0.0447488539,-0.0742185861,0.3183240891,0.0500646271,0.659071207,-0.0445929728,0.1819166839,0.0840544328,0.0472337678,-0.1453344226,0.1597234756,0.068557404,0.3448960483,-0.5432954431,-0.0704533681,-0.0477732383,-0.0830213055,-0.1924338788,0.1565322131,0.135423407,0.0860434622,-0.1599017829,0.2858231068,-0.0184930209,0.0122394478,0.1436226815,0.3150083721,0.1404030174,-0.2455681562,0.097809203,-0.0502138287,-0.1937011033,-0.2610455155,0.4902793467,0.0590683483,0.0314709209,-0.0605653413,-0.3164850771,0.2388178259,-0.3513689339,0.0303709246,-0.0477262735,0.0027172163,-0.0162866618,-0.0127467709,-0.0026860826,0.051894065,0.0969416201,-0.2098995298,-0.2651704848,0.0638557971,-0.0640470386,-0.1516255438,-0.0483408347,-0.1381340027,0.1749141216,0.0200463757,-0.3853110075,-0.1688892096,0.1255019605,0.0578714497,0.2358458042,0.1766265929,-0.2485652566,0.4699509144,-0.1297376156,-0.1469859779,0.0430618562,-0.0063615078,-0.2545925975,-0.1627761275,-0.3255548477,0.25451985,0.0457574874,-0.03966932,0.0741639063,-0.1403101981,0.146074757,0.0239902772,0.3011030257,-0.2547019422,-0.1741402149,0.0351589397,0.1350886822,-0.0714575574,-0.1860977113,-0.1540900171,-0.1498029679,-0.220589757,0.2482090592,0.3248289227,0.0830720887,-0.0982796997,-0.1811051667,0.1455697566,0.6714295745,-0.4713222086,-0.5759915113,-0.0373700559,-0.3127920032,-0.3483288586,-0.0722914711,0.5257780552,-0.2541721463,0.1841239035,0.0559160449,-0.0284524187,0.0702485591,-0.0407907329,-0.4820291698,-0.2443044633,0.4282407165,0.2879720926,0.1184493378,0.0484846048,-0.0653521344,-0.1245332807,0.3911313117,0.000452248,0.1557037681,0.2652800381,0.5162929893,0.1337868422,0.1086844057,-0.2567136586,-0.0195428692,0.1996148825,-0.1025182009,0.3829938769,0.1140017509,-0.0327498019,-0.2263214588,-0.2294123173,-0.3714199662,-0.1633773148,0.172794342,0.1106389314,0.1249597743,-0.0002137626,-0.0690630972,0.2349174768,-0.1951018721,0.0574191883,-0.2534842491,0.1734796315,-0.0722109526,0.0002446634,-0.2396927923,0.199630022,0.1941972524,0.1150439009,-0.1352175027,0.239359051,-0.2861438096,-0.2572818995,0.1280991435,0.0865210295,0.0745254755,-0.616065979,-0.2318005562,0.5943585634,0.1214367598,0.0328817405,-0.1761128753,0.2059393972,-0.1515797824,-0.1561628282,-0.3645664155,0.2804860175,0.1399029195,-0.0466484278,-0.08440274,-0.3972603083,-0.0248351693,-0.492898494,0.3556153774,0.128847599,-0.3839584291,-0.0716296509,0.321644336,-0.0507559404,0.1607387066,-0.018025944,-0.1524935514,0.0303000286,0.2578040361,0.1180636063,0.1480195522,0.1922579706,-0.0845910609,0.1927807182,-0.152892977,-0.1079919189,-0.0311143938,0.1146148965,-0.0190106314,0.0687712729,0.2076686472,0.1067494899,-0.17784141,-0.2461322248,-0.1395559311,0.268166095,-0.232404992,-0.0760472789,-0.1674968004,-0.2094908506,-0.5057194829,-0.1521824896,-0.0507111475,-0.099207297,0.1847671568,-0.0889074653,-0.0105202412,0.3142932653,0.1960565299,0.1012288854,0.1451552063,0.1357943267,-0.1855854541,-0.3651711345,-0.2351708263,0.0717458799,-0.3185686469,0.2699109912,0.362963438,-0.1586459428,-0.0961316675,-0.0891949683,-0.2441946864,-0.0196911767,0.0663825274,0.0541803502,0.0247739851,-0.0128802359,-0.2053201646,0.2964856625,0.2392027676,0.010781995,-0.0804010257,0.165114373,-0.0892625302,-0.0484052077,-0.2911716402,-0.4643014669,-0.5519579053,0.0023524363,0.0160962772,-0.1708481461,0.1917137355,0.1217435896,0.0539649054,-0.0223809853,0.1511522979,0.1787069142,-0.2152539194,-0.1564620882,0.1699820012,0.04657932,-0.2690292001,-0.1365805119,0.0395993926,-0.0324099883,-0.1843263805,-0.2561134398,-0.1648022085,0.1473015845,0.1070359349,0.0568785146,0.0316865295,-0.0339900665,0.1185097322,-0.0793804601,-0.2076565921,0.1215779185,0.1340607852,-0.3273540437,0.4176501632,-0.0254115332,0.2975836098,-0.0089342277,0.057175722,0.6455105543,-0.1563932449,0.0258548092,-0.1597404182,0.0220553521,0.1431916803,-0.2573871613,0.1051399559,0.4043577909,-0.0375110358,0.1361239254,-0.0309336483,0.0562431701,0.0022190406,0.1778101921,-0.286626339,-0.0421882905,-0.2883038819,0.0081804562,0.0121032577,0.0024795686,-0.1297768205,-0.016993884,-0.1806281656,0.000050472,0.1284516454,-0.1970236599,0.1025353074,-0.2461900264,-0.1150915325,0.0435953885,-0.1480235755,0.1687984169,0.3643041849,0.1497539878,0.1644024849,0.2874140441,0.2821745276,0.6061435938,-0.5048749447,0.2027046978,0.2420894504,0.497779876,-0.3383361101,-0.3108196259,-0.4152241945,-0.2385454923,0.4774543643,0.1917114109,0.0423532762,-0.1758732945,0.2739935815,0.1112614498,-0.3431813419,-0.1644869596,-0.2301604748,-0.0410895124,-0.112740472,0.1340055764,-0.0636912286,0.1891910732,-0.0973167643,-0.2146484107,0.054948952,0.1077021137,0.2498282939,0.1666043401,0.1703020185,0.1368238479,0.0883757845,0.0453658402,0.124310784,0.0230896771,0.3792884052,-0.1508143544,-0.3487549424,-0.0490253419,-0.4373478293,0.0748113394,-0.0211909842,-0.0076099364,-0.0620708875,-0.1066274643,0.3135819733,0.0335746035,0.1250333488,0.077662833,0.2430589199,-0.400236249,-0.3817459643,0.564103961,0.1614262909,-0.3126894832,0.1298294514,-0.1115404293,-0.232385844,0.2495944053,0.1741063893,0.8710391521,-0.0439972244,-0.0777831078,0.1872613877,-0.5491984487,0.4160410464,-0.0962840989,-0.1567226201,-0.5003467798,0.0170137044,0.0462395214,0.02956721,-0.3262172341,0.3848616183,-0.0168425199,0.3335962892,-0.0721827596,0.1315602958,0.0626660064,-0.0263762437,0.3919155896,0.1101384312,-0.1359794587,0.1998820603,-0.0867547989,0.36509794,-0.1807789356,0.0051463242,-0.1593277007,0.0511810035,-0.1213785112,-0.0190216359,0.1898370683,0.2132984698,0.0453390814,-0.1771476865,-0.3371671438,0.360956192,0.7153919935,0.0166354738,-0.0032235868,0.0695009604,0.2405066043,0.4848549068,0.0438013673,0.2695524693,0.4295513332,-0.0175234471,-0.3549792171,0.166836068,0.1424149275,0.1523369849,-0.403462261,0.0467120819,-0.1067343727,0.1847809404,0.1023434401,-0.0712311491,0.2050365657,-0.2408463508,0.1963704675,0.0711247548,-0.1680083126,0.0077840118,-0.0060903002,-0.2926673591,-0.1471736133,0.3525077105,0.1104488596,0.1597242802,-0.0106637459,-0.1213548109,-0.0746298507,-0.262301892,-0.0025097383,0.2081130892,-0.4890862703,0.1244215295,-0.0861642733,0.0198643431,-0.1201537997,0.2211391926,0.0576067865,0.2330544889,0.1238742694,-0.293454051,-0.4125732183,-0.1974737346,0.0958210826,0.3416711092,-0.1268664151,0.1598138511,-0.1841470897,-0.1776505113,-0.3772061765,-0.0511795506,-0.2954213917,0.2258235663,0.1912419945,0.2830272019,-0.1852217168,0.0047180387,0.1705357879,-0.0288293734,-0.0857438594,-0.2370090485,-0.2462338656,0.0528249294,0.2717525363,-0.2514607906,-0.0944939032,0.1243265197,0.060784135,0.0364012644,0.2265707701,0.2477096915,0.2358562648,0.9074827433,-0.1937113851,0.0112096146,-0.0809666514,-0.0974369273,0.0123710046,0.0009785178,0.0554630831,0.069967106,0.1669416577,0.0968655124,-0.2715975642,-0.1088641956,-0.3038878441,-0.0061018267,0.010286916,-0.5755164027,0.4176025987,0.2746883333,0.2566933036,0.0816466734,-0.3946265876,-0.2472677082,0.2157200873,0.3618089557,-0.229612872,-0.0064823949,0.2608962655,0.1352830529,0.0575641952,0.0501229055,-0.0629426092,-0.1292917579,-0.0882649943,0.036922276,0.1134369299,-0.3506598175,0.0868882984,0.4115595818,0.0686953962,0.0167496037,0.1330561489,-0.0848109797,-0.1213133931,0.6410635114,-0.0264557432,0.436912328,-0.0552831329,0.1763094217,-0.2318536788,0.1728783697,-0.0638043284,0.1332531273,-0.4932233691,0.0759485289,-0.0145830801,0.2183898687,-0.0881676599,-0.2310975641,0.1531068385,0.2232792079,-0.1793226302,-0.2515395284,-0.3708983064,0.1215013489,-0.0856519341,0.2131275237,0.0081086606,-0.0478741452,-0.073274903,0.1985646486,-0.2597439885,-0.4831650257,-0.4174051583,0.484400034,0.3321519494,-0.2023145854,-0.0913058221,0.1243589148,-0.0094582485,-0.0002588507,-0.1325841695,0.1982096881,0.1807582825,0.3041765094,0.0009505772,-0.0164919775,-0.1214542761,-0.0175779369,-0.0014495916,0.2395068258,-0.0714902729,0.1017907336,0.1085182354,-0.2647216022,0.4198532701,0.0897682831,0.5018251538,-0.3040899634,0.2026570439,-0.2344893962,-0.2914105356,-0.4042404592,-0.2128189206,-0.1952725947,0.298636049,0.174780488,0.4102656841,0.2424886823,-0.2421738058,0.114418827,0.0731973723,0.174911499,0.1492746919,0.0461604111,-0.0892121792,-0.0291161183,-0.619448781,0.1014879495,0.0836119875,0.0044276305,-0.0240110196,-0.0721822307,-0.5134974718,-0.1284178495,0.1638016105,0.2362136841,0.0536452122,0.4173865318,-0.1874003559,-0.0113126263,-0.0348638408,-0.0060417703,0.2730378509,-0.1915699095,-0.0835704654,-0.2481811494,0.1709211022,-0.1321860403,0.0614212416,-0.1478126794,0.2032229155,0.463974148,-0.0967973471,0.3414817452,0.0473213866,0.2152624726,-0.1106104627,-0.0845470056,-0.3863243461,0.0817851499,-0.1389547288,0.493219018,0.1644179374,-0.1385681778,-0.2046210617,0.1893080473,0.0041485988,-0.1906443685,-0.2836728692,0.3273255229,-0.0022105484,0.1383226514,0.0360353477,0.0244637746,0.2087702304,0.2192704529,-0.2801864147,-0.541385293,0.3990264535,-0.2757382989,-0.0321880132,-0.2142979056,0.3366419375,0.3930225372,-0.2063712478,-0.6328860521,-0.0117065776,0.3883118927,-0.1035705209,-0.4395238757,0.0690477043,0.2488583326,0.2722867131,-0.0126970382,-0.1619984657,0.2330733687,-0.0640470758,0.4385647774,-0.226767078]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1959","title":"Bug in skip_rows argument of load_dataset function ?","comments":"Hi,\r\n\r\ntry `skiprows` instead. This part is not properly documented in the docs it seems.\r\n\r\n@lhoestq I'll fix this as part of a bigger PR that fixes typos in the docs.","body":"Hello everyone,\r\n\r\nI'm quite new to Git so sorry in advance if I'm breaking some ground rules of issues posting... :\/\r\nI tried to use the load_dataset function, from Huggingface datasets library, on a csv file using the skip_rows argument described on Huggingface page to skip the first row containing column names\r\n\r\n`test_dataset = load_dataset('csv', data_files=['test_wLabel.tsv'], delimiter='\\t', column_names=[\"id\", \"sentence\", \"label\"], skip_rows=1)`\r\n\r\nBut I got the following error message\r\n\r\n`__init__() got an unexpected keyword argument 'skip_rows'`\r\n\r\nHave I used the wrong argument ? Am I missing something or is this a bug ?\r\n\r\nThank you very much for your time,\r\nBest regards,\r\nArthur","comment_length":31,"text":"Bug in skip_rows argument of load_dataset function ?\nHello everyone,\r\n\r\nI'm quite new to Git so sorry in advance if I'm breaking some ground rules of issues posting... :\/\r\nI tried to use the load_dataset function, from Huggingface datasets library, on a csv file using the skip_rows argument described on Huggingface page to skip the first row containing column names\r\n\r\n`test_dataset = load_dataset('csv', data_files=['test_wLabel.tsv'], delimiter='\\t', column_names=[\"id\", \"sentence\", \"label\"], skip_rows=1)`\r\n\r\nBut I got the following error message\r\n\r\n`__init__() got an unexpected keyword argument 'skip_rows'`\r\n\r\nHave I used the wrong argument ? Am I missing something or is this a bug ?\r\n\r\nThank you very much for your time,\r\nBest regards,\r\nArthur\nHi,\r\n\r\ntry `skiprows` instead. This part is not properly documented in the docs it seems.\r\n\r\n@lhoestq I'll fix this as part of a bigger PR that fixes typos in the docs.","embeddings":[0.1184592843,-0.5157735348,-0.0002604486,0.1341336071,0.2405335009,0.3162592053,0.3996219039,0.1163900495,0.1897607744,0.276788801,0.3576669991,0.3101141155,0.0523400754,0.1391011477,0.2122827172,0.009779199,0.3222466111,-0.0971084833,-0.2278704941,-0.1425264925,-0.466517657,0.205986321,-0.4954795241,-0.0062192786,0.0046916604,0.2700179815,-0.1468370259,0.325058192,0.1460925341,-0.0820216909,0.4361990392,0.092788741,0.0622523129,0.4919847548,-0.0001279238,-0.0485202931,0.0909043476,-0.1174367517,-0.2508835495,-0.2423615158,-0.0103976633,0.1511083841,-0.1586591601,-0.3461466432,-0.0797087103,0.2208740264,-0.1931831092,-0.3198311031,0.0172095113,0.2279166728,0.1132518351,0.0995623246,-0.2322340459,0.0133915367,0.3679778278,0.1688151211,-0.2028682232,0.5136327147,0.0749797001,-0.1312933266,0.2170034051,0.3002195656,-0.2162893564,0.1819370389,0.6969504356,0.2346367985,-0.0080310414,0.0670854971,0.1547786146,0.1427004337,0.3508746028,-0.2583890557,-0.1454953402,-0.5243525505,0.1148113981,-0.4422196448,0.3829135001,-0.0757525936,-0.1903962344,0.189231962,-0.1961863041,-0.3662986755,-0.2399329841,0.1452038735,0.2731882632,-0.3786741197,-0.1079846472,0.2083946317,0.6590313911,0.016708659,-0.0740466788,-0.1450303495,0.1362966001,0.4119838774,-0.5192498565,0.0457277521,0.2632987797,0.5026040673,0.1499713361,0.1526492238,-0.0932041481,-0.4410412908,0.4102219641,0.1674042493,0.1500700563,-0.0041119023,0.1651004106,0.2374447286,0.1703535467,-0.0576963313,0.1190752462,-0.0006027379,-0.0619619414,-0.2427826673,-0.0386947244,-0.2457968295,0.3429158032,-0.2678745389,-0.2375253886,0.10176678,-0.1411713958,-0.1343102753,-0.2298791856,0.3612944186,0.3145490885,-0.1731253564,0.005755458,0.2405338734,-0.1080146208,-0.1394807249,-0.1983850896,0.1062839627,-0.0351237208,0.2528295219,0.061962828,-0.4568336308,0.0769031271,0.0545499288,0.3502005935,-0.0496462695,-0.0791885629,-0.0380574949,-0.019058466,0.2549309731,-0.1363021731,-0.144865334,0.2976911664,0.1185013577,-0.0435420014,0.5042288303,-0.2794054151,-0.222258389,0.0607963018,-0.0945083797,-0.1695889533,0.0945608169,-0.6603605151,0.2222949564,0.1386060715,-0.2015696168,0.0344223194,-0.0063756914,-0.3262874186,-0.2263970971,0.2093477994,0.5372986794,-0.5187488794,-0.3551664948,0.10339991,0.0400552712,0.0190916546,-0.011484487,-0.0296287928,0.1079793125,-0.3077189028,0.2494839579,-0.0521676131,-0.2833323479,-0.1313028932,0.0714629069,-0.0621324331,0.2180778086,0.026195541,-0.2533243001,-0.0204311609,-0.166972965,0.253333956,0.1141669825,-0.2502171993,0.0942132249,-0.1150973588,-0.1965749562,0.3166657984,0.0250527002,-0.2964205146,0.1144848019,-0.2064061761,-0.5897098184,0.3778431416,-0.0728813633,-0.1749572605,0.1832056046,0.0438149087,0.1798682958,0.2096750587,0.0379981361,-0.6307157278,0.1086895987,0.4949711263,-0.046261739,-0.0923271701,-0.2249066085,-0.2521224022,-0.0548985712,0.1884768605,0.2854731977,-0.1327737868,0.0781114548,-0.1239917502,-0.0686810985,-0.1948330998,0.1549468189,-0.3675636351,0.1537380815,-0.4071644843,0.2684524655,0.2598707378,-0.0959241614,0.0065903384,-0.0984930098,0.2952405214,-0.3557017744,-0.1320198476,0.4505610466,0.23428455,-0.0091307759,0.0868653655,0.0611072555,-0.1808294952,0.019300038,-0.1605574489,-0.1600672156,0.1238663197,-0.0958675742,-0.1965255886,0.3076041639,-0.3604245484,0.325520575,-0.2628011703,-0.0352184884,-0.0754588023,-0.034186285,-0.0842181295,-0.1979155689,0.5240824819,-0.0061132195,0.1118526235,0.0035343422,-0.5212613344,-0.2066823691,0.383084625,-0.0621211939,-0.2070294172,0.3191439509,0.1557872742,-0.0262889266,0.3411843181,-0.1251760721,0.5859801173,0.0449354835,-0.3043896854,-0.0666453093,0.1478591412,-0.0423211791,0.2223359197,0.0385463275,0.1551990956,-0.2557174265,-0.1896429509,0.0230875332,-0.0583357513,0.1618837565,0.0453286543,0.179325968,-0.6254628897,0.2487070709,-0.0588402525,-0.1682410836,0.1682819724,0.087958768,-0.3060450852,-0.2525306642,-0.0842325985,0.0649744794,-0.1292806566,0.2338948697,-0.2523553669,0.6201030016,-0.1122031361,-0.2239248306,-0.334861666,-0.0328213796,-0.0588841811,-0.0667504817,0.0557746477,0.1014679521,0.3207905293,-0.0589228198,-0.1396524906,-0.0037375474,-0.0733497813,0.1602299809,-0.3428150713,0.3066675961,0.2471903563,0.3997295797,-0.1542533934,0.151381433,0.0548964143,-0.0476550721,0.1007011086,0.0335221291,0.253308177,0.0797876641,-0.292095989,0.0357824191,-0.1299386024,-0.0000385791,0.2213915139,0.019871654,-0.0376495905,-0.0669298694,0.1179807857,0.0733961388,-0.071933873,-0.0484872945,-0.3079355359,-0.6267732382,0.2409044653,-0.026431188,-0.1951413751,-0.0236929022,0.3533588648,-0.0311958622,-0.1339343041,-0.4656538665,-0.2469318062,0.1305469722,0.1517637819,-0.1969887167,0.0388503,0.0503357835,-0.2450463772,0.1536444873,-0.4438821971,-0.4812512994,0.0825288743,0.1209163219,0.1065659299,0.4717400968,0.2879648507,-0.2733358443,0.316341877,0.1464791298,0.3638205826,0.3352499902,-0.167528674,0.2650375068,0.0734614879,-0.3172518611,0.1638061553,0.0464834161,-0.2111837715,0.178517431,0.0041291206,0.1407556236,-0.2208397239,0.0856056511,-0.0104810987,-0.389654249,0.017736787,-0.2655600011,-0.0000428804,0.066714853,0.2684609294,-0.4285849631,-0.0578813329,-0.0875502527,0.1557602286,0.631010592,0.0573401712,-0.0730302259,-0.0804754943,-0.3710915446,0.3203972876,0.0968668237,-0.033295285,0.0873855203,-0.0594167411,0.1238119081,0.1924568266,0.9702677727,0.1808316112,0.1366528422,0.1105832458,0.2175794542,-0.3055568337,-0.0134121291,0.0733852237,0.2672163844,0.3817239106,0.2656692266,-0.4033347964,-0.2226455212,0.2769913077,0.2830926776,-0.0478488617,-0.3223292828,-0.3425843716,-0.1045977697,-0.3616633117,0.1127395332,0.2983873487,0.0953745618,-0.3774100244,0.3649781644,0.2887407541,-0.1708686501,0.0014054518,0.2205379903,0.2984918654,-0.3236280084,-0.0663358793,0.4265277386,-0.252358377,0.4026009142,0.6713850498,-0.3897360861,-0.4563646913,-0.0650470257,-0.5007573366,0.5561573505,0.2915156484,-0.3470171094,0.2494370788,-0.0746390894,0.0408958346,-0.1202686727,-0.3268789053,0.4433290064,0.0008113799,-0.1251009107,-0.3575873077,0.2981074154,-0.0859074816,-0.1844010204,0.426938206,0.7055799365,-0.2070477009,0.0742951632,-0.4285740256,0.98763901,0.4479186833,0.1335487217,0.1885944456,0.2559270859,0.544903636,0.4156984985,-0.039442867,-0.1166791618,-0.3400412798,0.0358401462,-0.0436227471,0.0801737234,0.1187170371,-0.1670798063,0.2674468458,-0.0833323672,0.0977663398,-0.0778850839,-0.3814988732,0.3286150396,-0.018328201,-0.2122380733,-0.0213963911,0.1279185861,0.0291785579,0.0027434838,0.0765290409,0.0225502048,-0.4613630772,-0.3341674209,-0.0798490569,-0.1575935632,-0.35949108,0.2929357588,-0.2710047066,0.2584749162,0.188232556,0.0791061968,-0.1033836901,-0.2842763662,-0.0838426501,0.0818976536,0.1955945939,-0.3937596977,0.0037559951,0.1803208739,0.0484673157,-0.2307668924,0.0925997943,-0.1555217355,-0.4176955223,0.0138239823,0.071384266,-0.3471805453,-0.3193702102,-0.4029918015,-0.0582872145,0.3188603222,-0.272246033,-0.031917613,0.1824084371,0.0771120638,0.0252369568,-0.0104837334,-0.341566205,-0.0579080321,0.2334397584,-0.1890173852,-0.1578358561,0.2290959507,0.3940145671,-0.1259292364,0.0110728936,-0.0159361232,0.1251057982,-0.4891855419,0.5048874617,0.4031619728,0.4566207528,0.1132586673,0.1569551975,0.0025206099,-0.1316834986,0.0398233198,-0.1272894293,0.0206685569,0.1495545655,0.2238905579,0.1502295136,0.1870157272,0.0887278914,0.0358877815,-0.0848186165,-0.1349639893,0.0610037744,-0.18004632,-0.1442871839,0.7343524098,0.4233570397,0.2425292879,-0.1966070086,-0.0783286989,0.0659479275,0.0952811986,-0.0415141769,-0.3675378561,0.2147181332,-0.0072650639,-0.0344188176,0.0490244702,-0.2043952644,-0.1229929104,-0.2577928901,0.2782376111,0.0852609128,0.052520968,0.2377353609,0.3915507495,0.1032231078,-0.0592512824,0.1123413742,0.0507622585,0.2624181807,-0.1811738163,0.108792752,-0.1437572092,0.0702594891,-0.0417995378,-0.0027778295,0.0403458066,0.1710273474,-0.030533798,-0.1313512772,0.1698524952,0.0017849387,0.0675761476,0.3770513535,-0.3807772696,0.098496154,0.0769971684,-0.0105655827,-0.3598612845,-0.1977274418,0.2949619889,0.1285376847,-0.0150992339,0.2488164306,-0.2123209387,0.036275357,-0.1478246599,-0.1595070511,0.5118710399,0.0680444911,0.0086554652,0.5517320633,-0.1172772795,0.2473681569,0.0751566067,0.3716813028,0.2079107314,0.3752686381,-0.0928772911,-0.1525134593,0.0226772353,0.0332652107,-0.3393266499,-0.612711966,0.1138678789,0.4662078619,-0.4114251137,0.0233353861,0.012316037,0.3759471178,0.0782099217,-0.1980503201,-0.1014166102,0.0741265267,0.020827407,-0.1253559589,-0.2443073392,0.0554774888,-0.0471790209,-0.0114518255,-0.2020470202,-0.0929233059,0.6958201528,-0.1093798056,0.2166823298,-0.2644774616,-0.062330924,-0.2002690434,0.1111040115,0.0200602505,-0.0145325419,-0.0366858281,0.0475840159,0.3377985358,-0.0703439638,0.3302784264,-0.0045204121,0.3747518361,-0.3745860457,-0.2005807012,0.0350316912,0.0564235561,0.2205990106,-0.0823419094,0.2905298173,0.1375501454,0.0074117347,-0.0456092581,0.4056415856,0.2595394254,0.1267185956,-0.5624562502,0.2345647365,-0.4512844384,-0.2537255287,-0.465295285,0.122199975,0.0452433638,0.1385639757,0.1585767567,0.1255729347,-0.0869322643,0.0617626682,0.0072451672,0.1048962101,0.4281698167,0.2795760334,0.2275763303,-0.454030931,-0.1966570169,-0.4089965522,0.1094144285,-0.0585413314,0.1503301263,0.4332963228,-0.102670826,0.2155163586,0.0754046366,0.2592982352,0.0064719436,-0.0435530916,0.1727090925,-0.1886402965,-0.5312667489,-0.2145917565,0.0049196002,0.1421578079,-0.1503656805,0.1773815602,-0.0768594071,-0.1850160062,-0.2282260358,0.1627285779,0.1208619773,0.1198444739,0.2785858512,-0.1667264253,0.2345069349,-0.016035134,-0.063096568,-0.0092437286,-0.0470901169,-0.3542245924,-0.0529601201,-0.2087932974,-0.0009171858,-0.3818293214,-0.3481458426,0.0208477583,0.0150349392,0.1303408295,0.2017199099,-0.4272088706,0.2923900187,-0.0747928396,0.0261830073,-0.1406653523,-0.097834304,-0.184204042,0.3183946908,-0.0623561479,0.0342243314,0.8977257609,-0.2795376778,-0.1409467161,-0.3316092789,0.1682136059,0.1322269738,-0.5405359268,-0.6190083623,-0.0539388545,0.2388797551,0.3874325454,-0.2610635757,0.5420950055,-0.3744550943,0.1921135634,0.0887777433,0.4205943644,0.1564856619,-0.0732057095,0.2644813955,-0.2102509439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1958","title":"XSum dataset download link broken","comments":"Never mind, I ran it again and it worked this time. Strange.","body":"I did \r\n```\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"xsum\")\r\n```\r\n\r\nThis returns\r\n`ConnectionError: Couldn't reach http:\/\/bollin.inf.ed.ac.uk\/public\/direct\/XSUM-EMNLP18-Summary-Data-Original.tar.gz`","comment_length":12,"text":"XSum dataset download link broken\nI did \r\n```\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"xsum\")\r\n```\r\n\r\nThis returns\r\n`ConnectionError: Couldn't reach http:\/\/bollin.inf.ed.ac.uk\/public\/direct\/XSUM-EMNLP18-Summary-Data-Original.tar.gz`\nNever mind, I ran it again and it worked this time. Strange.","embeddings":[-0.2764044106,-0.2233156711,-0.103664279,0.045770403,0.418441385,-0.0618779957,-0.0929600224,0.238045916,0.5096369386,0.4056463242,-0.1850806922,0.2674150765,0.3387895823,0.2342272699,0.2290572077,0.0686125159,0.000753105,0.12320932,-0.3504887223,-0.1669686884,-0.0790779367,0.1609481126,-0.1180893183,-0.1588568538,-0.1145428419,0.399920404,-0.0615768917,0.1302329153,-0.0440710187,-0.3042329252,0.244999826,-0.0201103017,-0.0212145243,0.3748176992,-0.0001020228,0.0553762466,0.1678401679,0.0461127609,-0.2232153863,-0.0890669748,-0.1297596097,-0.0460241027,0.1417364776,-0.1960759312,-0.0611383989,0.0368307382,0.0788052157,-0.3622721136,0.2039882541,0.2256118506,0.3025926352,0.1225930601,0.2186696082,-0.2986866236,0.0914430395,-0.5091504455,-0.0541461781,0.4087373912,0.3993614316,0.2585482895,0.2649771869,0.2739112377,-0.0679040626,0.0175543707,0.0715726167,0.0039512967,0.4605130851,-0.2682837248,-0.0467209145,0.1415991336,0.7401708364,-0.1731741279,-0.1414855868,0.1855368614,0.0983031318,-0.0891142637,0.20958969,-0.0452637039,0.0262667239,0.0892333835,0.0478678495,0.0265981872,-0.1243798584,0.203158915,-0.0877642184,0.2102627605,-0.172117576,-0.0091951424,-0.0123684695,0.1390240937,0.0659035519,0.1239858866,-0.0847338289,0.0301669966,-0.2340864539,0.0589899048,-0.0093309013,-0.0113875838,0.2729692161,0.1154535711,0.1092056632,0.1804099381,-0.215729773,0.1414728761,0.1131750643,0.0783869848,0.1416383386,-0.4686625004,0.2006008625,0.0859777331,0.0460311696,-0.0147069218,-0.1881822646,-0.0424212068,-0.2981321812,0.1592081487,-0.0732510015,-0.4073847234,-0.5382404923,0.1794435531,0.2019304335,0.0341731943,-0.0757718533,0.1730283648,-0.2285261154,0.2609003484,0.0758331865,0.178989321,-0.3156870604,-0.2880953252,-0.2477270961,0.0013452768,-0.1969659179,-0.1845733821,0.2578615844,-0.0651561469,0.2595637441,-0.0926851854,0.0810177699,-0.2237947285,-0.0023535257,0.1277985424,-0.2122615427,0.40383178,0.3498871624,0.174626261,0.1072512418,-0.0705069676,-0.0612785891,0.0707310662,-0.3816494346,-0.1223044619,-0.0774999708,0.331196636,-0.1024972126,-0.2609705627,-0.1378035843,-0.1863349229,-0.1280825287,-0.130698204,-0.2952125967,0.015136919,0.0441769399,-0.1182739064,0.4215595722,0.0737007186,-0.4475826621,0.0333113298,-0.1508289725,-0.2514948845,0.0885734186,0.0999337435,-0.0202766005,0.1169048771,-0.1858592033,0.181096971,0.0160356201,-0.5269539952,-0.5331658721,0.1238722131,-0.2614545524,-0.2906809449,0.0875370502,0.1174987108,-0.0181996152,-0.0120057929,0.2138641924,0.2541196048,0.0872502476,0.0218017455,-0.2783760428,-0.1488109976,-0.0392382219,0.0949340388,-0.0420364924,-0.081703037,0.0662484914,0.2583847046,0.4059853852,-0.0068355692,-0.0438732989,0.1859772503,0.2643761039,-0.1135218292,-0.0205590427,-0.3841408491,0.0504691191,0.1124104112,0.0365784913,0.202165395,0.1719736457,-0.147190392,-0.5457947254,-0.1043143794,-0.0453878082,-0.0900173634,0.2904453874,-0.0265382826,-0.0314590521,0.1865247488,-0.1403582096,-0.007928703,-0.2141026556,0.080278039,-0.1898532361,0.4338623583,-0.2191331536,-0.0026195645,0.0289952923,0.0413343124,0.075006716,-0.2755147219,-0.1230017915,0.5091951489,-0.0973589644,-0.0153443329,0.165177688,-0.2457539737,0.2062679827,-0.4278204739,0.0595184006,0.2515350878,0.0248620138,-0.0906808972,-0.1196787953,-0.0046730242,0.005003016,0.0426867716,0.0237166137,0.2956436872,0.1815700978,-0.2109782398,0.0559542626,-0.2433481216,0.3559718728,0.2225655913,-0.2567153871,-0.3612931371,0.0574358404,0.2645579576,0.1073275283,0.0148486672,0.0731170252,-0.1362330616,-0.3165480793,0.0196062382,-0.0165489912,0.1210737899,0.2569382191,0.2263541669,0.1181029677,0.203424409,-0.0522617847,-0.2192419469,0.0794206038,0.0045057209,-0.0640518442,0.1565916389,-0.2245605141,-0.0157692526,-0.2913669944,-0.0941511393,-0.0516464077,0.2043609619,-0.2796161175,-0.2052050382,-0.327342242,-0.3389379382,0.0777488947,0.0146948975,0.1257458776,-0.5507653356,-0.0802335963,0.156023249,-0.1041695103,0.1569285691,-0.3882160783,-0.0323445015,0.0559337363,-0.104550831,-0.0437140688,-0.1256535649,-0.1199783832,0.2215936929,-0.0226989519,0.1685116142,0.3146056831,-0.4046343565,0.0504872762,-0.3088687658,-0.2650348246,0.0951296985,-0.1840364635,0.2213110179,0.0097352033,0.2241288126,0.1826987416,0.0162745118,0.3517945707,-0.1412424594,0.16457811,0.0764224902,0.0662495196,-0.2384145558,-0.186042428,-0.4612749219,-0.1751509756,-0.1923461258,-0.0455992036,0.1057623699,0.1489665955,-0.0107490001,0.1280044317,0.1324311644,0.1088374704,-0.0068928353,-0.211812377,-0.4825291038,0.2286378741,-0.2118044496,-0.7715365887,0.2630935907,0.319381088,0.2052171826,0.0850011706,-0.4112580419,-0.058949478,-0.0533760414,-0.0299679004,0.2193081379,0.2324508429,0.1466675699,-0.2037300467,-0.2504064739,-0.0734709129,0.0510923415,-0.1526043713,-0.2124257535,0.3691562712,-0.192392692,0.058414951,0.1514961571,-0.3077858984,0.0219618231,0.0900342017,0.3092048764,-0.2141511738,0.2547960281,-0.14483504,-0.2422440499,0.3726710677,-0.0816625953,0.0336632915,0.0257963818,0.1322605759,0.2185096145,-0.2938734591,0.0232551508,-0.1743822843,-0.2709324062,-0.173386842,-0.0480491333,-0.1245892569,0.2689690292,0.1434736699,-0.1070549488,-0.1090969294,0.1068169624,0.3341707587,0.0336368233,0.0359358601,-0.6460626125,0.1369527578,-0.2345936149,0.3795817494,-0.2107500881,0.1123149991,-0.1936036199,-0.0177938491,0.1300113648,-0.0932523981,0.200162068,-0.3285288513,0.1797149181,0.330660522,0.1581169516,-0.335380733,0.0676862448,-0.201903075,0.1286539435,0.539182961,0.0600065589,-0.1926287115,0.0419302098,0.2993035018,0.3723601997,-0.0958164185,-0.0701785982,-0.1488515437,-0.1742246896,-0.1317340136,-0.3690787554,-0.0231718738,0.2054982185,0.0137599558,-0.099446252,-0.1536712646,-0.1992365271,0.2447923571,-0.0157726035,0.2868978381,-0.1049039364,0.3004880548,0.0686266795,0.0376842245,0.0713640153,0.477460891,-0.0921594277,-0.1503475904,0.1439627707,-0.4003172815,0.0667612255,0.062432915,-0.1119067445,-0.2911936939,-0.0356295854,0.0385079272,0.0012290591,0.2479393631,0.209203437,0.1162730306,-0.2959610224,-0.4427827597,0.4512099028,-0.010498832,-0.0605118535,0.5520952344,0.1645716727,-0.0402668417,-0.0159875341,-0.0405098982,0.7600554228,0.1020254567,-0.0291606616,0.166036576,0.2898142338,0.4640924037,-0.3148960471,0.1504715681,-0.1436509043,-0.2371143848,-0.0317096263,-0.0875571221,0.0773668215,-0.07096719,-0.1690234542,0.2010775357,-0.0573223196,-0.0725683346,0.1508755237,0.1754481494,-0.2561741173,-0.0837746263,0.0815923959,0.291484952,0.1432785988,0.3798618019,-0.0138785113,-0.0341683105,-0.1944875419,-0.1422200203,-0.2294468433,0.4808261991,0.057832431,0.0733363703,-0.3454812765,-0.3790072203,-0.0166590679,0.2106298208,0.1370495409,0.3227264881,-0.2481309474,0.1894940734,-0.4602684975,-0.2036532611,0.3243308961,0.0934725404,0.0025174562,-0.2440018654,-0.3351403773,0.1628840715,0.1517549306,-0.1733371466,0.0747999698,-0.0490900427,0.057829164,-0.2295881808,-0.1143699214,0.0864797309,0.0790589154,-0.3293909132,0.2318040729,0.1154871657,0.0672041476,0.1333773285,0.005116683,-0.2586956024,-0.1399454027,0.4396959841,0.0930423215,-0.2489510477,0.2917980552,0.1076716632,-0.0996790603,-0.308650732,-0.168989554,0.0223063882,-0.3777389526,-0.0933288187,-0.3013107777,0.0320021473,0.0224515628,0.3412595093,0.1383705735,-0.1538671553,0.0425758325,-0.5141339898,-0.1871087998,0.083305113,0.126412183,0.1680183709,-0.05671468,0.3326092958,0.2265016586,-0.3396377861,-0.4223298132,0.1763816178,-0.0035242883,-0.181774646,0.0350776464,0.0436985791,0.4562261105,-0.1539541036,0.2955729663,0.0153698102,-0.1905777007,-0.2805426717,-0.1522511244,0.0493951254,0.1341015697,0.0024308728,0.0373213142,-0.2385322899,-0.1153899133,0.0696318522,0.0675834268,-0.0080564553,0.0220183097,0.1929484159,0.0482578278,0.2787000537,0.0287708882,0.1294963807,0.0898164958,0.5602575541,-0.0813619941,-0.038692262,-0.1323051304,-0.0662849173,-0.55384022,-0.1036054865,0.0001127057,0.1068167686,0.4119785726,-0.2794876397,0.2196392268,-0.0154830934,0.5139537454,0.4189261198,-0.3055816591,-0.0603956319,0.1614418626,0.3564123213,-0.2088758349,0.0242070947,0.0831522048,0.3083201945,-0.015725689,-0.1054596677,0.3084817827,-0.0258254036,0.3062668741,0.0789919347,0.6571202874,0.1079625413,0.1032787636,0.2854039967,0.0550014935,0.1304156035,-0.0848326609,0.0996289402,-0.0744592473,0.47083202,-0.2091995627,0.0471035056,0.1016732007,0.1772847325,-0.3722590804,-0.1400655359,0.14274773,0.1104568094,-0.027793536,-0.0893151388,0.1150517315,0.4786494076,-0.2869562209,0.1160090864,-0.3308345377,0.2082569599,0.0785619318,-0.0064413548,-0.1526996642,-0.0796728805,0.0452289209,-0.0372289084,-0.0758959576,0.0253868029,0.0582960844,0.084685646,-0.1543906331,-0.7404968739,-0.2398364991,0.0298607722,0.1219475046,-0.1450752169,0.0517201647,0.2287221104,0.2320341915,0.1003994271,0.3321994543,0.3574899435,0.2998871505,0.0171521101,0.0169069469,0.0440290421,-0.0430321693,-0.1834830791,0.0819630772,0.2755979896,0.1191816553,0.3257418573,0.3219671845,-0.3136021495,0.0502686203,0.0941158757,0.332472086,0.1180216447,0.0346288532,-0.0715984032,0.015239778,-0.2664966285,-0.2925759554,-0.5327065587,0.2229028046,-0.0283834506,0.0416000336,0.1011038125,-0.217861101,0.1713426113,-0.0281066205,0.3776103854,0.2248639464,0.3179716766,-0.2836133242,-0.3549744487,-0.7161689997,0.3535206318,-0.0692363307,-0.5048679113,-0.1602217555,0.1981882304,-0.0067467354,0.2532434762,0.1823502779,-0.0613274574,-0.077198036,0.2188005596,-0.3017931581,-0.1129925922,0.0016723787,0.1833504438,-0.0321644954,-0.2237074524,0.1441137344,0.0649875104,0.201727882,-0.1572109014,-0.0512361079,-0.0074920021,0.0391256362,0.6396716833,0.1039026156,0.62641114,-0.0832064971,-0.1765110046,-0.417095691,-0.3758992255,-0.0470225066,0.1608997434,0.3128935993,-0.0722383633,-0.0813180581,0.0357701294,-0.2359566838,0.4477524161,0.0487037711,0.0625852793,-0.3592696488,0.1898798347,0.1562644988,0.0525116287,0.1364221275,0.2010004669,-0.0943664089,0.1769389808,-0.2459845394,-0.3627747893,0.3736001849,-0.0463957004,-0.0260394104,-0.0069333916,0.2731678188,0.1373420656,-0.1534952223,-0.3705109954,0.1966354698,0.3039187491,0.0387588628,-0.4010812044,0.1908845156,-0.0291675627,-0.0518016219,-0.0390139706,0.1873755306,0.0683301091,-0.236577481,0.0421643443,-0.1397438347]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1956","title":"[distributed env] potentially unsafe parallel execution","comments":"You can pass the same `experiment_id` for all the metrics of the same group, and use another `experiment_id` for the other groups.\r\nMaybe we can add an environment variable that sets the default value for `experiment_id` ? What do you think ?","body":"```\r\nmetric = load_metric('glue', 'mrpc', num_process=num_process, process_id=rank)\r\n```\r\n\r\npresumes that there is only one set of parallel processes running - and will intermittently fail if you have multiple sets running as they will surely overwrite each other. Similar to https:\/\/github.com\/huggingface\/datasets\/issues\/1942 (but for a different reason).\r\nThat's why dist environments use some unique to a group identifier so that each group is dealt with separately. \r\n\r\ne.g. the env-way of pytorch dist syncing is done with a unique per set `MASTER_ADDRESS+MASTER_PORT`\r\n\r\nSo ideally this interface should ask for a shared secret to do the right thing.\r\n\r\nI'm not reporting an immediate need, but am only flagging that this will hit someone down the road.\r\n\r\nThis problem can be remedied by adding a new optional `shared_secret` option, which can then be used to differentiate different groups of processes. and this secret should be part of the file lock name and the experiment.\r\n\r\nThank you","comment_length":42,"text":"[distributed env] potentially unsafe parallel execution\n```\r\nmetric = load_metric('glue', 'mrpc', num_process=num_process, process_id=rank)\r\n```\r\n\r\npresumes that there is only one set of parallel processes running - and will intermittently fail if you have multiple sets running as they will surely overwrite each other. Similar to https:\/\/github.com\/huggingface\/datasets\/issues\/1942 (but for a different reason).\r\nThat's why dist environments use some unique to a group identifier so that each group is dealt with separately. \r\n\r\ne.g. the env-way of pytorch dist syncing is done with a unique per set `MASTER_ADDRESS+MASTER_PORT`\r\n\r\nSo ideally this interface should ask for a shared secret to do the right thing.\r\n\r\nI'm not reporting an immediate need, but am only flagging that this will hit someone down the road.\r\n\r\nThis problem can be remedied by adding a new optional `shared_secret` option, which can then be used to differentiate different groups of processes. and this secret should be part of the file lock name and the experiment.\r\n\r\nThank you\nYou can pass the same `experiment_id` for all the metrics of the same group, and use another `experiment_id` for the other groups.\r\nMaybe we can add an environment variable that sets the default value for `experiment_id` ? What do you think ?","embeddings":[-0.338857621,-0.4435221255,-0.0121532436,0.0259144232,-0.0003200659,-0.1188518777,0.4191748202,-0.0220881794,0.694543004,0.334757179,-0.0704577714,0.2575858235,0.0208578371,0.0921893641,-0.1329859495,-0.0292750634,-0.0619412065,-0.1000037789,-0.1330745965,-0.1348372549,-0.3230050504,0.1404807717,-0.1086490601,0.033235576,-0.1115667224,-0.0510076657,-0.2452469915,0.2668406963,-0.2192952782,-0.3962014019,-0.0470970906,0.661061883,-0.0401509851,0.3781712055,-0.0001052099,-0.0733862147,0.1403617114,-0.1786131561,0.1255243719,-0.0316817388,-0.000416236,-0.2217452377,0.1787687391,-0.584472239,0.1618598253,-0.3749356866,0.1602336913,-0.620006144,0.5229051113,-0.092047438,0.1828143597,0.4089609981,-0.3570385873,-0.1778223366,0.0755402222,-0.0741302446,0.0622762367,0.4285991788,0.4382115304,-0.1648908854,-0.3774008155,0.0479232594,-0.0095322765,0.4112260342,0.5238329172,0.1457097977,0.1289573908,-0.2262361199,-0.3590711355,0.5437494516,0.0621518306,-0.0059186695,-0.2175007612,-0.0875980034,-0.0445221327,-0.2018630058,0.0316551402,0.0730774105,-0.2755064368,0.070117034,-0.090801008,0.1095004454,-0.2334239483,-0.016613435,0.1179818287,0.4391717315,0.1565182358,0.3073607683,0.4072217047,0.1744004339,-0.5871927738,0.3229157329,-0.1573162377,-0.1374911517,-0.2163519859,-0.0715378895,-0.1345012933,-0.2474492341,0.2707520723,0.2180184275,-0.1847607195,-0.0347350873,0.5012080073,0.2264455706,0.1733484864,0.2225490212,-0.0145000555,0.2927637398,0.3371622264,0.0463934466,0.0130739938,0.1034730375,0.2154529989,-0.0866064131,0.0407120883,0.0291974191,0.0415125079,-0.01023384,-0.466539979,0.1799823195,0.1296782494,-0.0897997171,0.1274870634,0.234956637,0.2761306465,-0.1984886974,0.2240095735,0.0891115814,-0.2850963473,0.2397984564,-0.1558063179,-0.0192014258,-0.3922467828,0.0479356796,-0.0023125117,0.2354627848,0.3477457762,0.0911863297,0.3725248873,0.0711543187,0.1579835862,0.1498431712,0.1748493761,0.1164054871,-0.0279612709,-0.1533220559,0.1613998115,-0.0924462825,-0.1918702722,-0.253803134,-0.2990004122,-0.1078144684,-0.0957446992,0.2104450911,-0.4288651943,0.1794071048,0.3127835989,-0.044189129,-0.0737617016,-0.1354023516,0.4412108064,-0.1397594064,-0.0608982854,-0.1054953784,0.0603922233,0.5278873444,0.1167069003,-0.3138536215,0.1733956188,-0.2131505609,-0.0626815706,0.2512883544,-0.0731380805,-0.1633246541,-0.1726192087,0.0451325662,0.1839676499,-0.570456326,-0.4551762044,0.2267521024,-0.4629926682,-0.1833772212,0.5246551037,0.3992156386,0.4919354916,-0.1816748083,0.1842800975,0.1285000145,-0.0225711148,-0.0210043956,-0.2199544162,-0.2132083923,0.2055389136,0.1335553229,-0.0407903492,-0.0299863778,0.1842954457,0.0449085347,0.3806384802,0.0599327944,0.049430985,0.0249996204,0.2068836093,0.0262049623,0.1731613874,0.0735590011,-0.0773380697,0.1987994611,-0.2551991045,0.0786855221,0.1668829024,-0.2616333365,0.0597451739,-0.1807138175,-0.3017637134,-0.3694494665,0.1339280754,0.1007098779,-0.425352782,-0.2251143456,-0.1763160974,0.4126905203,-0.0026029355,-0.0783250853,-0.177489832,-0.1190525815,-0.286660254,0.0046237456,-0.0351308882,0.1059898138,0.1628840268,-0.1626184881,-0.183540538,0.3546155691,0.3039299548,0.109246552,0.335532099,0.1854079366,0.1129209995,-0.1525601596,0.1541462988,0.0319657102,0.0296029598,0.046851892,-0.1756990701,0.3792437017,-0.1660756469,-0.0425319895,-0.1739497036,0.2052099705,-0.0407580175,-0.1759886593,-0.3004975021,-0.1131736115,0.5896606445,0.2008987218,0.1878550947,0.0096999872,-0.0725959316,0.0389045477,0.0934056044,0.1792061925,0.0130479056,-0.1290573329,0.2640550733,0.0338388756,0.0100350939,0.3682819605,0.5299090743,0.2942911685,-0.0035119657,0.2998711467,-0.022067694,-0.2826879025,-0.0080962246,0.0303518008,-0.1510358453,0.1437180787,-0.0427940525,0.0623153821,-0.138762325,-0.2496975958,0.1067964658,0.0581836253,-0.3008233011,0.2420987338,-0.0928122401,0.2503350675,-0.0295380857,-0.1750542074,-0.1717950404,-0.3238005638,0.29145661,0.0818376243,-0.4521723092,0.1862106025,0.0592271388,0.3545638621,-0.1497493535,0.0780103132,-0.0674722493,-0.3504399955,-0.0287414007,0.0338265896,0.0693555623,0.2826054096,0.7376533151,-0.0895642415,-0.1716850102,-0.3883170784,-0.1675387621,-0.0548316911,-0.2927384079,0.5755821466,0.3770563602,0.1170972288,-0.156552121,-0.2018892765,0.1285159886,-0.4124993086,-0.1544502378,-0.1241053045,0.0733181387,-0.2878053784,-0.3132178485,-0.425173521,-0.3864906728,-0.3381042182,0.1782756448,-0.1811825633,0.1448945701,0.060122598,-0.351541698,-0.015221348,0.2928807139,-0.1243955716,-0.4747360051,-0.6132940054,0.1247877777,-0.3289964199,-0.2105765194,-0.0699182153,0.21357207,-0.1581752002,0.2809077203,-0.2412759066,-0.7833874226,0.1669145674,0.1982203573,-0.1628024727,0.0633633062,0.2734371126,-0.0901562646,-0.1465454549,-0.1769039035,0.0182819646,0.5482854247,-0.052514676,0.1894315332,-0.1148565337,0.0894955993,0.0083572557,0.5683797002,0.3369283974,0.012904115,0.4058041871,-0.1494740099,0.4014569223,0.1018189788,-0.2884704173,0.3126338422,0.2290747166,-0.2266400605,0.050981652,-0.2824375033,0.2149926424,-0.0292827431,-0.1109168157,0.1569546163,-0.3881774843,-0.1111626104,0.0122723812,0.0890299901,-0.2054055035,0.1362084001,0.0220757984,-0.0265207235,0.0318306722,0.5858515501,-0.0124930553,0.1177794784,-0.0797203928,-0.1668656021,-0.3578243554,0.4702195227,-0.0713934973,0.4873967469,-0.3471036255,0.0088333189,0.1524148881,-0.0133303432,0.5068036914,0.061241433,0.146770075,-0.0510439426,0.066610001,-0.0926588848,-0.280185461,-0.2568093538,-0.0583581738,0.2295815498,0.1282978356,-0.4007205963,-0.3233206272,0.0673335865,-0.168711707,-0.3320307434,-0.4579871297,-0.4411281049,-0.2853358388,-0.3242387176,0.2526414692,0.0713428259,-0.1816559732,-0.021465065,0.008205045,0.1470402628,0.2471085638,0.0978484899,0.0833591074,0.1141774282,-0.1949960738,0.1316143721,0.4105849564,0.1862633526,0.0891109183,0.2787957489,-0.2423746139,0.030171968,0.1517164856,-0.1628015041,0.0851587281,0.3966124058,0.0636260137,-0.0428079739,0.0545017794,0.3373203874,-0.317887485,0.0252185743,0.3418806493,0.0325722396,-0.1114490032,-0.0825199112,0.117481254,0.1140533686,-0.2343799323,0.3709228635,0.0198112987,-0.1443082839,0.304214716,0.1867202967,0.9030949473,0.0311157703,0.228094548,0.2173774838,0.0732171237,0.1633869708,-0.0176216848,0.0419676676,-0.3961948752,-0.3439328074,-0.1805066019,-0.1347697675,0.2987574041,-0.2991396189,-0.3444767892,0.3026094139,-0.0014976469,-0.2954224944,-0.4523965418,0.4312303662,-0.3331827521,-0.1938413531,0.2419086844,0.1094016358,-0.0754101127,0.2564893365,-0.0621118471,-0.2436823249,0.0105544552,-0.321307689,-0.3890616,-0.2424219698,0.0219555926,0.2903726995,-0.0039325021,-0.1580266505,-0.1069824174,-0.0413956493,0.0664166212,0.2461607009,-0.1328770518,0.2183260471,-0.2534845769,0.0385652408,0.0761466846,-0.360330075,0.3276643753,-0.1171878427,-0.5082363486,-0.1755797714,0.145401448,-0.2381673306,0.0332742743,0.1399684846,0.355722934,-0.0296970643,0.1975512207,-0.1510793865,-0.2765431702,-0.1418732405,0.1220527813,0.2428269535,-0.2791320086,0.3559331596,-0.1460072547,-0.3269095123,0.1301740855,0.1035871431,-0.0482685715,-0.057016544,0.2107543945,-0.2135885358,-0.0768894702,-0.2594355345,0.0709317699,-0.0081096906,0.0050342111,0.0122502511,0.1308852136,-0.2050805241,0.0797916725,0.139365077,0.3301217556,0.1170080453,0.0605256855,-0.471609205,-0.2771270275,0.0881797895,-0.4826113582,0.453003943,0.0873672739,0.0094039077,0.0200223997,0.2794264555,-0.3884773254,0.0165666714,-0.360101521,-0.1482967585,0.3039814234,-0.0586699247,0.0161730461,-0.0914917365,0.0403938293,0.111242041,-0.5826269984,-0.1632358581,-0.3143565655,0.1723043621,-0.1240756288,-0.0353661105,0.0486162007,-0.4985809922,-0.0422138199,-0.0605113246,0.1797477305,0.4235542715,-0.1383513808,0.0699024647,0.0984755978,0.2840332091,0.1419795156,0.2620127797,0.150504604,-0.1604099572,-0.2000433356,0.1210674942,-0.2418048084,-0.0977046937,-0.1702811867,0.1415652484,0.4924521148,0.2272942513,-0.1555066854,0.2823543549,-0.1449465901,-0.0772125348,0.4140523374,0.37487638,-0.2612327337,-0.1307837665,0.0538166277,0.2483969033,-0.2528648078,-0.0649072826,0.1949891746,0.3159147501,0.1384039968,0.3598711789,0.3957139552,-0.003083311,-0.1721450239,0.0037229052,0.3735209405,-0.1775870472,0.2339088321,-0.2443967313,-0.1510739923,0.160830915,0.3821859658,0.0139735835,0.3264163435,0.0110329054,-0.2467038482,0.3727151453,0.3607267737,0.0910366252,0.096688576,-0.0279843025,0.0726910606,-0.1473217458,-0.0689115748,0.0848786309,0.2389332056,0.6212895513,-0.5725147724,-0.1553122401,-0.2476173341,0.3540132642,-0.1840917021,-0.0106131332,0.1967908442,-0.1030417532,-0.1526864022,-0.1799846292,0.243683055,-0.0352681763,0.2975531518,0.246914193,0.1061945781,-0.2136853933,-0.1666731983,0.1843445301,0.1577887237,-0.4586446583,-0.1551937014,0.3142164946,-0.1442608535,0.4445006251,0.3467246592,0.2973738015,0.1290785819,-0.2095435262,-0.0285068452,0.0848276764,0.1700477749,-0.1163835526,0.1652919948,-0.0131251905,-0.3018900156,0.2105091959,0.1445438713,-0.1336321384,0.079003863,0.1094840392,-0.0485223196,-0.0589641482,0.5235513449,0.0981115624,0.0814317316,0.0804447606,0.1009463817,-0.6264939308,0.1366877556,0.0718844011,0.0941775143,0.2018312514,-0.3069219887,0.1171846315,-0.01182207,0.3767240345,0.1390984356,0.0868158862,-0.4619697928,0.1721453071,-0.7986824512,0.5472496152,0.053406354,-0.4315550029,-0.0160559416,0.044231303,0.0583805293,0.1264327466,0.3685828447,0.3651022315,-0.0372741558,0.3127584159,-0.511043489,0.2197709829,0.1067206413,-0.1656024158,0.212930575,-0.2172719836,0.1808338761,-0.0124271354,0.0676129684,0.0702232644,0.0709809363,0.2476882637,0.0499053039,0.2937142849,0.0575814582,0.5136566162,-0.0202782899,-0.0191470478,-0.107481122,-0.0272800196,-0.2270099521,-0.3902609348,0.1994389445,0.0170686413,-0.2086765766,-0.1834443808,-0.4254196882,0.1746605039,-0.2887333632,-0.0300679505,-0.3006604314,0.2215656638,0.0388803855,0.1303006858,0.0145916175,0.327349484,0.2044250816,0.143247515,-0.1279266477,-0.0846335813,0.5487346053,-0.2959809303,-0.186741069,-0.2512256801,0.4572097957,0.1084876508,0.1357612163,-0.4568593204,0.0080926511,0.0233042017,-0.0894661173,-0.3460148871,0.1537269354,-0.2121055424,-0.1914075166,-0.0443377867,0.0226720851,0.0970135182,-0.4199603498,0.2008335739,-0.1824046224]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1956","title":"[distributed env] potentially unsafe parallel execution","comments":"Ah, you're absolutely correct, @lhoestq - it's exactly the equivalent of the shared secret. Thank you!","body":"```\r\nmetric = load_metric('glue', 'mrpc', num_process=num_process, process_id=rank)\r\n```\r\n\r\npresumes that there is only one set of parallel processes running - and will intermittently fail if you have multiple sets running as they will surely overwrite each other. Similar to https:\/\/github.com\/huggingface\/datasets\/issues\/1942 (but for a different reason).\r\nThat's why dist environments use some unique to a group identifier so that each group is dealt with separately. \r\n\r\ne.g. the env-way of pytorch dist syncing is done with a unique per set `MASTER_ADDRESS+MASTER_PORT`\r\n\r\nSo ideally this interface should ask for a shared secret to do the right thing.\r\n\r\nI'm not reporting an immediate need, but am only flagging that this will hit someone down the road.\r\n\r\nThis problem can be remedied by adding a new optional `shared_secret` option, which can then be used to differentiate different groups of processes. and this secret should be part of the file lock name and the experiment.\r\n\r\nThank you","comment_length":16,"text":"[distributed env] potentially unsafe parallel execution\n```\r\nmetric = load_metric('glue', 'mrpc', num_process=num_process, process_id=rank)\r\n```\r\n\r\npresumes that there is only one set of parallel processes running - and will intermittently fail if you have multiple sets running as they will surely overwrite each other. Similar to https:\/\/github.com\/huggingface\/datasets\/issues\/1942 (but for a different reason).\r\nThat's why dist environments use some unique to a group identifier so that each group is dealt with separately. \r\n\r\ne.g. the env-way of pytorch dist syncing is done with a unique per set `MASTER_ADDRESS+MASTER_PORT`\r\n\r\nSo ideally this interface should ask for a shared secret to do the right thing.\r\n\r\nI'm not reporting an immediate need, but am only flagging that this will hit someone down the road.\r\n\r\nThis problem can be remedied by adding a new optional `shared_secret` option, which can then be used to differentiate different groups of processes. and this secret should be part of the file lock name and the experiment.\r\n\r\nThank you\nAh, you're absolutely correct, @lhoestq - it's exactly the equivalent of the shared secret. Thank you!","embeddings":[-0.2438624203,-0.5634172559,-0.0252403282,-0.0789237022,-0.0883313641,-0.0465874597,0.4310415983,-0.094735682,0.6884128451,0.3382450044,-0.0324464552,0.2416061014,0.0294990465,0.0651078969,-0.0662357882,0.0515502393,-0.0137192803,-0.0775892362,-0.2323585749,-0.1388348192,-0.258167088,0.1642918587,-0.1347393095,0.1182623878,-0.1310550421,-0.0675265938,-0.293854177,0.3518728316,-0.2069532126,-0.3734535575,-0.0511100814,0.5885394216,-0.1115160733,0.3470314145,-0.0001025435,-0.0290896483,0.2390752733,-0.1840211004,0.0937753469,-0.0225164928,-0.0063960506,-0.2147799581,0.2390904129,-0.6090909839,0.2548007965,-0.3178233206,0.3465066254,-0.5903226137,0.4716377556,-0.1917704493,0.2114199549,0.4353037775,-0.2383697033,-0.2103887349,0.0767698884,-0.1481731981,0.0649377778,0.3659193814,0.4074763954,-0.1650011092,-0.313985467,-0.0395437144,0.043598678,0.3771082759,0.5423665047,0.216079846,-0.0078244032,-0.3159538805,-0.3685427904,0.544993639,0.0689924657,-0.0175680891,-0.1552130133,-0.0432232507,-0.1245836392,-0.0624338761,0.0405220389,0.0801767707,-0.2795662284,0.095741272,-0.063394323,0.2075612098,-0.1429089159,-0.0452653691,0.0962147862,0.4674387872,0.1451297998,0.3589620292,0.3896103203,0.2626070678,-0.595140934,0.2887125909,-0.0776974261,-0.1011503413,-0.1418187171,-0.0618655644,-0.0439788438,-0.1890520006,0.2088943273,0.1767123789,-0.2218020707,-0.0964426249,0.4521384835,0.2948691845,0.1431004405,0.1408369839,-0.0696253553,0.2683866322,0.300517112,0.0206614006,0.0306791496,0.1363249123,0.2108891904,0.0027403666,0.0445302986,-0.0351561271,0.0238796473,0.0576753095,-0.442153126,0.1751860827,0.0797957256,-0.1001359746,0.0160644874,0.1829918027,0.3118669987,-0.2796678245,0.1751922965,-0.1004711762,-0.2744460702,0.2910319865,-0.1642244458,0.0376682244,-0.3985897303,-0.0958595946,-0.0917951539,0.0960873514,0.3988892734,0.0868708789,0.4171565175,0.1293655336,0.2241773605,0.1347905993,0.2759168446,0.178447932,-0.0017330297,-0.2029621452,0.2175432146,-0.081027478,-0.2209538668,-0.2160374969,-0.2931857705,-0.0774588361,-0.1473874003,0.2421827614,-0.3560352027,0.2010423988,0.3800360858,-0.1475385576,-0.1068175137,-0.098329626,0.5739683509,-0.1218078882,0.0830891579,-0.0840706825,0.020189194,0.4858983457,0.0717014894,-0.2658756375,0.2152179331,-0.276860863,-0.0522133298,0.2739189565,-0.0850304663,-0.2639813721,-0.2457928061,0.1607364565,0.1307996362,-0.481888175,-0.4057360888,0.2260941267,-0.4426649809,-0.1356583685,0.5243067741,0.3977626264,0.3573784232,-0.1086715981,0.182147935,0.0534409247,0.0158823803,-0.0299053807,-0.1290168166,-0.1947342157,0.0986651331,0.053240262,-0.1217258796,-0.0514579751,0.2830915451,0.0329827741,0.3958420157,0.0993602872,0.0005168345,-0.0211902335,0.097968325,0.1156114787,0.212317735,0.1543367356,-0.09075284,0.2483416796,-0.3218515217,0.1370418966,0.1254315227,-0.2333319485,0.0656790137,-0.1063255072,-0.2659734488,-0.3408946097,0.1597786844,0.1032047719,-0.4180200696,-0.275488019,-0.2062266171,0.4432168901,-0.0138518624,-0.1089850813,-0.2169679552,-0.1268142313,-0.334615767,0.0727749094,-0.1406827867,0.0365250446,0.074713625,-0.1410052478,-0.164787665,0.3544574082,0.3195937574,0.1061796099,0.3022877872,0.1927321255,0.0328771695,-0.1241667792,0.1496131718,0.0756309703,0.1013348401,0.0801173076,-0.0873717889,0.256319344,-0.1455116272,-0.1093319952,-0.1375808716,0.13144508,-0.0765736476,-0.1365867108,-0.2957275808,-0.1035109982,0.610616982,0.317109406,0.231996119,0.0240163058,-0.0837848485,-0.0262041688,0.1193377599,0.1129376218,0.0377228856,-0.1472759694,0.2499714494,-0.0389145724,-0.0289272256,0.3769492507,0.5130673647,0.2503073215,0.071699664,0.3051802218,0.0016121762,-0.2875819504,0.0358735695,0.0561415479,-0.1126169339,0.0684218258,-0.0435020328,0.0493567586,-0.1518482566,-0.3390766382,0.1330272555,-0.0304418411,-0.1809448004,0.2108364254,-0.0941149294,0.2258033901,-0.0496726111,-0.1246389225,-0.2088010162,-0.4025353193,0.2573352158,0.0917731747,-0.4607612491,0.0951692164,-0.0385735333,0.2664901316,-0.1746661365,0.0735596344,-0.1235503629,-0.3214639127,-0.0437071472,0.0795470849,-0.0880632102,0.2526984811,0.7304578424,-0.1311330795,-0.2009440213,-0.3724404573,-0.1643543392,0.0462005772,-0.3329814374,0.4429244697,0.4895553589,0.0533521473,-0.1120601445,-0.1844489574,0.0840663761,-0.3830341697,-0.2660089135,-0.1845015883,0.1068547517,-0.3374786973,-0.3387934864,-0.368558228,-0.4663522243,-0.3014016151,0.2419752926,-0.1762255877,0.1574790031,0.0984287485,-0.404407382,0.0433088169,0.2788309157,-0.1425389946,-0.499094069,-0.5869216323,0.0771412998,-0.2953599095,-0.1439429671,-0.0453416817,0.1576921195,-0.2012320012,0.1635486037,-0.2414560467,-0.7453762293,0.0937659591,0.2354438454,-0.0996864215,0.0434934199,0.3776494563,-0.0950522795,-0.2161876708,-0.2172877789,0.0611052252,0.4402588904,-0.0263326485,0.1234301403,-0.0219239611,0.0885738879,-0.0025034384,0.5244187117,0.231462732,0.0415040255,0.3437654376,-0.191254586,0.4083034396,-0.0738660693,-0.2043543309,0.2984809875,0.1865424961,-0.3711796403,0.0471345559,-0.3468583822,0.1800029874,0.0091128303,-0.0582057498,0.2116906643,-0.2857472301,-0.1605326831,0.0355200619,0.0172329526,-0.1656162441,0.0695135295,0.1110113114,-0.0095531996,0.0143053643,0.5648735762,-0.0347725302,0.1002481058,-0.1071098074,-0.2136805207,-0.3458729386,0.4169006944,-0.0235033464,0.4492670298,-0.3281523585,-0.1347951144,0.2345647663,0.0092919981,0.4790715575,0.1126458272,0.1230675951,-0.0280252751,0.1007869244,0.005192678,-0.3153510392,-0.3267291486,-0.0421085916,0.1992313415,0.081579268,-0.3377787471,-0.261932373,-0.0124277659,-0.2160033882,-0.3102512956,-0.4012083113,-0.5420239568,-0.2751145959,-0.3600599468,0.239256531,0.1309219301,-0.2203641236,-0.0489166267,0.0770777687,0.1968029439,0.303154856,0.0091499891,0.1745309979,0.0539988279,-0.1638088077,0.0084147751,0.3750188649,0.2809549272,-0.0584357977,0.2942312956,-0.1721453965,0.1452232301,0.2947836518,-0.1858273596,0.0139187835,0.2652875781,0.0881237015,-0.0277720988,0.0551487803,0.317225188,-0.2777989805,-0.0882019475,0.3498710096,0.0299350992,-0.0799063593,-0.0434259847,0.0550970808,0.0568047799,-0.2567835152,0.2959794402,0.056592267,-0.0883460641,0.2275805026,0.1434837729,0.8623201251,0.062468458,0.4144766927,0.1813774854,0.0522226728,0.2632079422,0.0515541919,0.0629204661,-0.4102436304,-0.3862266839,-0.243396461,-0.1412841827,0.2726684213,-0.3026342392,-0.4003547132,0.3024199307,0.0394350849,-0.2073438466,-0.4099218845,0.3298549354,-0.2978419662,-0.0898286253,0.2729331255,0.1163659841,0.0611584894,0.3924246132,-0.0717784092,-0.12018428,-0.0320484079,-0.3505810499,-0.3782687187,-0.2039823979,0.0113128591,0.3753530085,0.0356372334,-0.0861976668,-0.1979120076,-0.0643186942,-0.0133302854,0.236809507,-0.109384492,0.2227956206,-0.0976634696,0.0239577089,0.1493279338,-0.3480623066,0.3557846248,-0.0891308188,-0.4809229672,-0.0553521514,0.0374371745,-0.2613354027,-0.0033930517,0.2813756466,0.3861358464,0.0169835016,0.0992792323,-0.1376872063,-0.3311130702,-0.0852764472,0.1279795766,0.1529994309,-0.2878354192,0.2079714984,-0.1581050009,-0.3879677951,0.1795874536,0.1025423706,-0.0463122316,-0.031727314,0.2740699351,-0.2099541575,-0.0563407429,-0.2196567655,0.0789949596,-0.0370823145,0.0329978913,0.0156200426,0.1447085291,-0.2545282543,0.0552963838,0.1285775155,0.350636065,0.288582027,-0.0284398906,-0.4474082589,-0.3208758235,0.0160208307,-0.4509329498,0.4844259322,0.2211310714,-0.0322237425,0.1585773975,0.1567707956,-0.3978053927,0.0319292396,-0.4641882479,-0.1248669177,0.3408115506,-0.075253956,0.0603977963,-0.0641891807,0.090621464,0.1347994953,-0.5935941339,-0.1863761246,-0.3097167909,0.139476493,-0.1753057241,-0.038648203,0.0729879439,-0.4741799235,-0.0042618504,0.0098958109,0.2145979553,0.4166866243,-0.1314972788,0.0742182508,0.1948978305,0.2095610797,-0.0010700441,0.2277536243,0.1270664632,-0.1274959743,-0.1741636693,0.0659340471,-0.3609362543,-0.1272587627,-0.1918007582,0.2360032797,0.4632054269,0.2804304957,-0.1466752887,0.3265675902,-0.1062843576,-0.2187989801,0.4266735017,0.3294382095,-0.2642704546,-0.2751123309,0.0969031155,0.2293529809,-0.2391537875,-0.0873809531,0.2002388984,0.2963561416,0.1713221371,0.2899867296,0.3527077436,0.0661037043,-0.2116889805,-0.0075866268,0.3411163986,-0.1817124337,0.1577339321,-0.1876578629,-0.1835862696,0.259157598,0.3726924956,-0.1158644259,0.3025060892,-0.0012335795,-0.307197392,0.3853198588,0.4048243761,0.05814825,0.1763491482,0.014160715,0.1094284803,-0.1097700968,-0.0881034508,0.1642715633,0.175325349,0.5363361835,-0.5529485345,-0.0152972294,-0.2336349338,0.3381834626,-0.1672060043,-0.1116958112,0.2486129701,-0.0818461776,-0.1377281547,-0.1703465581,0.2336868197,0.0222307816,0.3978537917,0.2234881073,0.1507522166,-0.1550403386,-0.1522816569,0.0944682285,0.1013660431,-0.4949466288,-0.1825048625,0.3110486567,-0.1196641028,0.4652066231,0.3376262188,0.338267386,0.1054530144,-0.1465232521,0.0138811311,0.0766874254,0.1530277282,-0.1337038726,0.1422636509,-0.0048856977,-0.2454777956,0.1626679748,0.1714867502,-0.1359953284,0.0845952705,0.1394140422,-0.0188518837,-0.0488447212,0.4834093451,0.0646335781,0.0362902172,0.0416610688,0.3051947355,-0.6858347654,0.1099766418,0.0944082588,0.1007446945,0.1709935665,-0.2672555745,0.1259394437,-0.0470306464,0.4904879034,0.2188585103,0.134214595,-0.3599074185,0.2000602186,-0.6501296163,0.5871246457,0.0781986266,-0.5412569046,-0.0465625413,0.0264715683,0.0851541162,0.0336233452,0.3250840306,0.3061597943,-0.1271294802,0.3344304264,-0.3145225644,0.1504697949,0.0444562919,-0.1469235271,0.1582152843,-0.2173129469,0.1303644627,0.0323747955,0.0812971964,0.0324241593,0.0689331144,0.2904607356,-0.0080095204,0.2579007745,0.0819111988,0.4832308888,-0.0748590678,-0.0566992797,-0.0253478978,-0.0559688061,-0.1712686419,-0.3650263548,0.2243227065,-0.0869929269,-0.1901637316,-0.1799830794,-0.3332695365,0.2587376237,-0.2860462964,-0.0420758873,-0.3193355203,0.1278971732,0.0068237018,0.114600271,0.0743902549,0.3767515123,0.2553520501,0.0016886055,-0.0977996215,-0.0905388221,0.4677690566,-0.3282396495,-0.1754850894,-0.3249821663,0.3709001541,0.0376721807,0.1435147226,-0.4403663576,0.0035192121,0.0317308381,-0.0757510811,-0.3692752123,0.1333639175,-0.1457455754,-0.1425705552,-0.1155923381,0.1194954962,0.1037470773,-0.4305735528,0.1731836647,-0.1977106631]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1954","title":"add a new column ","comments":"Hi\r\nnot sure how change the lable after creation, but this is an issue not dataset request. thanks ","body":"Hi\r\nI'd need to add a new column to the dataset, I was wondering how this can be done? thanks \r\n@lhoestq ","comment_length":18,"text":"add a new column \nHi\r\nI'd need to add a new column to the dataset, I was wondering how this can be done? thanks \r\n@lhoestq \nHi\r\nnot sure how change the lable after creation, but this is an issue not dataset request. thanks ","embeddings":[-0.234251976,-0.0549111441,-0.1812909096,-0.0524510667,0.0165326744,0.0354573317,0.3398596942,-0.0366015695,0.0798499286,0.1445066929,0.0107775247,0.1978341937,0.0306363143,0.3719204962,0.1402912736,0.1070635766,-0.2483593524,0.4518121183,-0.1740532815,-0.0729186535,-0.3456433117,-0.0778268054,0.1791183352,-0.2770399153,-0.3205130994,-0.2202869207,0.0320800766,0.0979108885,-0.3490820825,-0.4073872864,-0.0071710809,0.3199305236,-0.0588898212,0.1259934604,-0.0001025004,-0.1651871055,-0.0742464662,-0.1672584862,-0.232492432,0.1484258771,-0.2225056142,-0.1204708591,-0.0649789348,-0.2121859044,-0.4337717891,0.066661641,-0.0907053128,-0.4884952009,-0.0028187665,0.3639039993,0.3648745418,0.1379202157,-0.1298270375,-0.3510278165,0.2751463056,0.0080140121,-0.1131887361,0.1160293072,0.1404919773,-0.0684450492,-0.1430787444,0.0452408716,0.5009627342,0.1005612984,0.0736441687,0.0076348106,-0.0122201797,-0.0688309371,0.2871033847,0.2532128096,0.552614212,-0.1473834217,-0.2127429694,0.0277346764,0.2715097368,-0.3950330615,-0.1023686677,0.0229894593,0.1035378948,0.0007529014,0.3265817165,-0.3593195379,-0.3072420657,0.1033377126,-0.150990352,0.3630675673,-0.0937532932,-0.0232558511,-0.1438339055,-0.0457355492,0.0405790843,0.5270217061,-0.0486872457,0.0868929923,-0.1699164808,-0.1534494311,0.12583749,0.0771977827,-0.0013115226,0.0071968413,0.0130605316,-0.1483258158,-0.1018105,-0.0491407402,-0.0985417068,-0.2586115003,0.2241183072,0.081423983,0.0712919757,-0.1250397116,-0.1409027129,-0.3387119472,0.090477258,-0.0591131896,0.0375072584,-0.0255632307,0.1908127666,-0.1325506419,-0.1077480316,-0.09479554,0.1682734936,-0.0805519447,0.0298450794,0.4504487216,0.1969485134,-0.0125306053,-0.0428345017,0.1681353748,0.0000444309,-0.3650209904,-0.2425491661,0.0406483263,0.0234598909,0.0446031727,0.0642384216,0.198407799,0.359516114,0.0659383982,-0.0261973757,0.1453186274,0.1730485559,-0.0396587737,0.2590415478,0.1630302072,-0.0337751023,0.0389916413,-0.0643853992,-0.0070228279,0.019939851,0.0474146754,0.0771668032,-0.0852071121,-0.6027121544,0.2963129282,0.3269850016,-0.298228085,0.1507370472,0.1876557469,-0.1407329291,-0.2708619237,0.0805230588,0.1842939705,-0.1445277333,-0.1888030767,-0.0908173695,-0.0513373427,-0.3442672193,-0.0079605347,-0.2085409313,-0.1071883887,0.141275838,-0.1655448973,-0.0701738894,0.1713636816,-0.0367519334,0.087621972,0.5381994843,-0.0310597699,-0.3604515493,-0.2498097122,-0.1576156616,-0.368219018,0.2996690571,0.5366459489,-0.0117452722,0.0206747949,-0.2044987977,0.0247423239,-0.2616029084,-0.0983523652,-0.1163445264,-0.0286877286,0.1365664899,0.0614786483,0.1248018965,0.1944216937,0.3542417288,0.1688765734,0.1422254145,-0.2761051953,-0.0772149563,0.1145116016,0.5605573058,0.004229879,-0.0603509583,-0.0865190476,-0.3809376657,-0.1319012493,-0.2419222742,0.0342391171,0.2410074323,-0.3780497015,-0.3833188415,-0.1641275585,-0.0116927288,-0.0470617972,0.2636746466,0.1956169307,-0.0568315908,-0.267572403,-0.0989275649,0.1625026911,-0.0958886892,-0.1135597676,-0.2682310343,0.0426566601,-0.0947994962,-0.1658973396,0.2144950479,0.1803326458,0.2492696494,0.1912123263,-0.0645207763,0.1831037402,-0.0789258778,0.2383155376,0.2969739735,0.1007824391,0.0447555631,0.168024078,0.0345779285,-0.2406416237,0.0310394187,0.2692002654,-0.2842737734,0.1433785707,0.0153192403,-0.1771539599,0.0303519275,-0.1171382815,0.207539022,-0.3155144155,0.1554968357,-0.5195201039,0.0813990608,0.0396895818,-0.0432146899,-0.0339907557,-0.3423886299,0.2929726839,0.1848214865,0.1485246271,0.2735670209,-0.024573015,-0.1060432568,-0.0345940217,-0.0283582602,0.3750405014,0.1763320118,0.3945909441,0.0966876075,-0.065639846,-0.107538268,-0.2565464377,-0.0951367393,-0.3713633716,0.0835183933,0.2370493859,0.3434161246,-0.3087705076,-0.4307762384,0.097355321,0.062979959,0.3789816201,-0.2094773501,-0.2184330374,-0.095888719,0.0126106888,0.0270041004,0.0428651012,0.1422492266,-0.2052132785,0.2989897728,0.2755517364,-0.0869216248,-0.0169358552,-0.2405190617,0.3863027394,0.1515257806,0.2785090506,-0.003132354,-0.3506252766,0.1799201816,0.2896728516,-0.2606052458,0.1840180457,0.8060863018,-0.1483347267,0.2207277864,-0.4004446566,-0.392611742,-0.0963892713,0.2622625232,0.396368444,0.1159327105,0.2579714656,-0.0414654426,-0.1163630262,0.085041903,-0.1183885485,-0.1783895195,-0.2418837398,-0.2830155492,0.3030081093,0.0470834821,-0.6182641983,0.1425399035,-0.2114256918,0.0926956311,0.0451594107,0.0243385751,-0.2476557642,-0.1087786406,-0.0807371289,-0.1766065806,0.0192312747,-0.578874588,-0.3256313205,0.2002626508,-0.4118048847,-0.2977393568,-0.1887417436,-0.1371980608,0.3828653097,-0.1759369969,-0.0962097049,-0.2526671588,-0.1949562579,0.2247782052,-0.1720819324,0.0294296462,0.5308918357,-0.1651632637,-0.2386846244,-0.2381104082,-0.1847108603,0.1396241039,0.2987613678,0.1010990292,-0.2928363979,0.309417367,-0.2106472254,0.1787432134,-0.2057984173,-0.2192953974,0.4171487689,-0.4435089231,0.440014571,0.0062042465,-0.2436644882,0.0464619808,0.3540069759,0.2039543241,0.166148439,0.0458548889,0.0117861088,0.1008761451,-0.0214096252,-0.3324842751,-0.1376765519,0.0581159592,-0.1252361834,0.2881025374,0.009422549,-0.2280369848,-0.4572380781,-0.115929395,-0.1513103694,-0.0643427297,0.0590966642,-0.3631082773,-0.4082349837,-0.0058826194,-0.3175401092,0.1657299995,0.0785015821,0.2407679558,-0.1516016871,0.0132910749,0.0675241649,-0.0793042779,0.4309128225,-0.4250861406,-0.0646549687,0.1295699328,0.172838822,-0.2631373703,0.3178607523,-0.2250889838,0.2126710415,0.0132096205,0.3453506827,-0.0616825037,0.0251382403,0.4632736444,0.2184384614,-0.3159709871,-0.2171535492,-0.1842076927,-0.2916347086,-0.2276749164,0.08149001,-0.087078318,0.1912452132,0.3089259267,-0.1160627306,-0.1407399774,-0.2207105458,0.1488073021,0.2851306498,0.0713690966,0.3882255852,0.1160017848,-0.1079921946,0.0300685074,0.1671796739,0.2251330614,-0.2639661729,0.1820215732,0.0570057705,-0.374899447,0.4564008415,0.089512907,0.1582809091,0.4401009083,-0.1284204274,0.0947444215,-0.1609490216,0.2963100076,0.3025246561,-0.1867970526,-0.0204954762,-0.2033822536,0.3991321325,-0.0663109198,-0.2036618292,0.2702282369,0.2597520649,-0.1118683591,0.1192702278,0.1386223286,0.8258507848,-0.1651404649,-0.1446479708,-0.0164528452,-0.4330645502,0.2019690275,-0.1218137965,0.1498010159,-0.3629225492,-0.1199739426,0.0581465662,0.0594906658,0.2272736132,0.0732949376,-0.040814355,-0.0277748387,-0.2128306925,-0.1205980256,-0.0634451285,0.3888405561,-0.1706169844,-0.0385121778,0.0300873332,0.2427949607,0.0967733786,-0.3202816844,-0.0654103681,0.1378675699,-0.0022840232,0.0278224349,-0.0952260271,-0.3391869664,-0.0943234488,-0.1221593767,-0.0982924029,-0.3562058508,-0.0801301822,0.1126285344,0.4637310207,0.2225897461,-0.1460873783,0.3929459453,-0.0290304609,-0.0106850145,-0.0356684178,0.1002591997,-0.0205149315,-0.2906253636,-0.3704018891,0.03855763,0.1290340424,-0.4437745214,-0.0630921349,-0.0641659051,0.136926949,-0.5037252307,-0.0970801339,0.1996070594,0.1768924594,-0.4096499383,0.2604002953,0.2085069418,-0.0506003052,0.1573241204,0.2987822592,-0.1548352689,-0.1539896578,0.1537168175,0.2887458205,0.3733950257,0.2711693048,-0.1032876596,-0.0903468505,-0.5066464543,0.1425132751,0.3526089787,-0.4496010542,0.1566504836,0.2619929612,-0.1789028794,0.1585228294,0.2552873194,-0.1255612671,0.2080289721,-0.0764535293,0.019324718,-0.1772874594,-0.1674602032,-0.1012038589,0.3335441351,-0.0086052427,0.1660338193,-0.1989145577,0.2513757646,-0.4863962531,0.0772578567,-0.2586612403,0.3740481436,0.2822453082,0.0298459325,0.0526984707,-0.201311782,0.3311571777,-0.1693150401,0.0260728654,-0.3804881275,-0.1417732239,0.0272810291,-0.0407106355,0.2253510505,-0.1291323453,-0.098043099,-0.0999181941,0.1576020569,0.095543541,-0.0132000437,0.1009495705,0.2714956105,0.084444195,-0.0818278342,0.0909768641,0.0607517213,-0.092682384,-0.0806736276,-0.2555572987,0.1211445406,-0.0813705474,-0.0933392644,0.2188299149,0.0702439174,0.1460737139,0.0843621716,-0.0059984368,0.0182248931,0.1019705236,0.3345043063,-0.0408118144,-0.1155470833,0.0686178803,0.1410688609,0.2462552339,0.411970377,-0.5275209546,0.199717164,0.1063262224,-0.207549572,0.489481777,0.2371238321,0.0950018167,0.1738279015,0.0527155362,0.067053169,-0.1209326461,-0.1525200456,-0.0445638336,0.2598795891,0.0038239346,-0.1028244421,0.4838427007,0.174403742,0.0401196294,0.2454656661,-0.0153561085,0.344330281,0.1854997277,-0.0901881084,0.1924189627,0.0522952266,0.3868966699,0.115433149,0.1420226395,0.0033466674,0.2396483868,0.3484844565,0.0982982516,-0.1710909009,-0.093546696,0.1353795826,-0.1843029112,-0.2212319225,-0.3129013479,-0.1616813838,-0.0202544425,-0.1022178903,-0.0152745787,0.0196006,0.0218567494,-0.1224043667,0.0537980609,-0.1761557013,0.2304535508,-0.0224753711,0.0916209817,0.0089870999,0.172144115,0.1660161763,0.0933442712,0.3902437985,0.511795938,0.0263725091,-0.1082327217,0.1417185813,0.3036316633,-0.0487567224,-0.3072279394,-0.0955745876,-0.1135290563,0.2628571391,0.295041889,0.0611836351,0.3272846043,-0.3016424775,0.2575121522,0.069497548,0.0337028578,-0.0859602243,0.0288314652,0.1409157962,-0.2790925801,-0.2350941151,-0.0579076447,-0.1712060422,0.0085802255,-0.0225509014,-0.3474243581,0.0604503863,-0.1142209843,0.1349505186,0.2431203872,0.3177865148,-0.0364288017,-0.0754410923,0.0554301925,-0.3573129177,-0.4765751362,0.1471113712,-0.1192049608,-0.3870978653,-0.2098788917,0.1780355126,0.1869483441,0.078063041,0.54296422,-0.0980799496,-0.0792967305,0.0778159127,-0.2451916486,-0.0572218075,0.2500936389,0.0704833865,-0.0130244233,-0.0930496454,0.1171159223,0.0759095848,0.2199334204,-0.056321729,-0.0838343203,0.0421145707,-0.0226467736,0.3171388209,0.0366458297,0.3963233829,-0.3592706919,0.0693522766,-0.1080049276,-0.1451854557,-0.2400190085,-0.0097660003,0.4331315756,0.3607579768,-0.0265036765,-0.1312754601,0.0754580349,0.244697392,-0.0946358293,0.1830011308,-0.0927543342,0.5407566428,0.2331822515,-0.1764004081,-0.0008146632,-0.0152571667,0.0184416771,0.0915627331,-0.0991520137,-0.4116050601,0.3260044158,-0.0473124608,-0.1467532963,-0.1849721223,0.2483223826,0.3317495286,-0.0286455899,-0.3332256973,0.2292403579,0.2574435472,0.0949163288,0.1930333972,0.4474113584,0.0441587828,0.1114351451,-0.0722742751,0.2434653938,0.1745956242,-0.1794503629,-0.1646870226,-0.2129769474]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1954","title":"add a new column ","comments":"Hi ! Currently you have to use `map` . You can see an example of how to do it in this comment: https:\/\/github.com\/huggingface\/datasets\/issues\/853#issuecomment-727872188\r\n\r\nIn the future we'll add support for a more native way of adding a new column ;)","body":"Hi\r\nI'd need to add a new column to the dataset, I was wondering how this can be done? thanks \r\n@lhoestq ","comment_length":40,"text":"add a new column \nHi\r\nI'd need to add a new column to the dataset, I was wondering how this can be done? thanks \r\n@lhoestq \nHi ! Currently you have to use `map` . You can see an example of how to do it in this comment: https:\/\/github.com\/huggingface\/datasets\/issues\/853#issuecomment-727872188\r\n\r\nIn the future we'll add support for a more native way of adding a new column ;)","embeddings":[-0.2675053477,-0.3267511427,-0.2293746471,-0.0306972917,0.0712796822,0.1932588071,0.2296767533,0.098611936,0.2282399833,0.1450353712,-0.155760482,0.1634853631,0.0027359948,0.51192379,0.2227695286,-0.1781963706,-0.0625528023,0.1988129318,-0.0534236208,0.0027416525,-0.3495630622,0.0484526381,0.2379430085,-0.1702793986,-0.2531002462,-0.2114895731,-0.0195244942,0.0866771787,-0.2067594081,-0.3567590415,0.0423742458,0.2098613083,-0.0749002248,0.1859726012,-0.0000973422,-0.0641719848,-0.0359068289,-0.134630084,-0.1033529937,-0.0347682685,-0.2414609045,-0.1699303687,-0.069243513,-0.0861014277,-0.418582648,-0.1042871177,-0.0217464622,-0.153125748,0.1256458312,0.1134686843,0.4063779116,0.2553823292,-0.0764760301,-0.3638369143,0.1776876897,0.1461547911,-0.2223314196,0.0608458295,0.0292487275,-0.0947456285,-0.08670111,0.2065436393,0.3743553162,0.1480084509,0.1906539947,0.2405007929,-0.1616064459,0.0865648165,0.1893065721,0.327333957,0.4445179701,-0.1060111448,-0.3236677051,-0.2550575733,0.0308138691,-0.3803779185,-0.0968920738,-0.0229365844,0.095169425,0.1734922826,0.0142113063,-0.1790034175,-0.1423597932,0.1836372018,-0.2275916338,0.2113582641,-0.2669085264,-0.0052920156,0.1706732512,-0.0678176358,-0.2897088826,0.3743713796,-0.0731748566,0.1141525805,-0.1223932579,-0.1932972223,0.1971586049,0.1913174838,0.2126955539,0.0485491827,0.0913344994,-0.054155089,-0.110760808,0.1634424031,0.0828428566,-0.1297962815,0.20901829,-0.0888698623,0.2225174159,0.0488693193,0.0825487822,-0.3772425354,0.2139924467,-0.0154408403,-0.0938181058,-0.194901064,0.2368553579,-0.1049163118,-0.1222472042,-0.1332412362,0.2561585605,-0.0918196142,0.0656536818,0.4747041166,0.0947093293,-0.158417061,-0.045620691,0.1981340647,0.080589734,-0.1449050009,-0.3359633982,-0.0149343796,0.0481394231,0.1808351576,0.034531232,-0.0946614072,0.4725374877,-0.0024814336,-0.0302333087,0.1297015548,0.0783943534,-0.0043350221,0.1934849769,0.1643794328,0.004470835,-0.0576049238,0.040354792,-0.1187347621,-0.1352625489,-0.0327263959,-0.0106876418,-0.1250885576,-0.3687202632,0.3041077256,0.104108341,-0.2859869897,-0.0987835079,0.4664135277,-0.0598139763,-0.2486292273,0.0581969544,0.3222052455,-0.3405737877,-0.0693770349,0.0408708975,-0.0870443881,-0.252587229,-0.2556756139,-0.0349228345,-0.2431697249,-0.1837906539,0.0473110788,-0.0754375681,0.079724595,-0.0735879391,0.2277758718,0.2582001686,0.1087667197,-0.2271649092,-0.1417523324,-0.2780780792,-0.09710031,0.14764902,0.3498815298,0.1436208785,0.0415607616,-0.0008696713,0.0429348759,-0.184850052,0.0126023944,0.0297190491,-0.0185100157,0.022534281,0.1027058288,-0.0757717863,0.1558429152,0.31769225,0.0230158847,0.1711735725,-0.1810259819,-0.070137389,0.2003710419,0.4800208211,0.0673372,-0.1219241694,-0.2546574771,-0.5289201736,-0.0148887308,0.0389770493,-0.0887867808,0.1192283481,-0.3768486679,-0.3831667304,-0.0615021959,-0.1084753796,-0.0569789447,0.2907108366,0.1430888921,0.067939356,-0.1913001686,-0.1594552696,0.1520783603,-0.0369494706,0.0091191288,-0.2650406361,0.1284197569,0.0173635744,-0.1359552443,0.0963114575,0.1147398278,0.1042970493,0.0309456773,0.1269075274,0.2909142971,-0.1034535021,0.3359621763,0.2832551897,0.143500492,0.1118826717,0.0508651622,0.0449773669,-0.1548962891,0.0173943769,0.0894271657,-0.4237238467,0.2668480277,0.1428694278,-0.0632410347,-0.0419426002,-0.0038181678,0.2277862132,-0.1741270572,0.0426250398,-0.4620880783,0.018292645,-0.0694311634,0.0898766369,-0.0262677725,-0.406491667,0.2828010917,0.3379993439,0.0339879729,0.0409408547,0.1030980051,-0.2593681216,0.0518031158,0.1649906337,0.2050795555,0.2228690088,0.4660227597,0.2596595883,-0.0010431162,-0.1270840913,-0.1808649898,-0.0776134878,-0.2090056092,-0.0364621021,0.1165696159,0.2806752622,-0.3271846175,-0.5827208757,0.0823420584,-0.0810477436,0.1986769587,-0.1886476874,-0.2168995738,-0.090699397,0.0099926675,-0.0009752501,-0.2157727331,-0.1718835235,-0.2708534598,0.2800835073,0.3527169228,-0.2082553655,0.0128023848,-0.0188154746,0.2867797017,0.0408826396,0.1994555146,-0.074971959,-0.2418559343,0.145731926,0.3013759255,-0.2270241231,0.0597375222,0.6364293694,-0.1969311386,0.1622603238,-0.4014016688,-0.4579341412,0.0456495844,-0.0933285877,0.2280401587,0.1930567622,0.4515183568,-0.113910526,-0.1788264662,0.2495382875,-0.0848386213,-0.3495542109,-0.2163494527,-0.1054063886,0.1298486441,-0.0546000153,-0.3900428116,0.0308279917,-0.3032822013,0.5029949546,-0.0068603801,0.0459585637,0.0246298183,0.0617438667,0.0616711304,-0.3266998827,-0.0044154814,-0.5509761572,-0.3817510009,0.3346668482,-0.4371896088,-0.2686751187,-0.0591444373,-0.0356205069,0.136780858,-0.1566551626,-0.095357053,-0.3855728507,-0.1848876774,0.2939253747,-0.0176991746,0.1278983057,0.4841794372,-0.1433973014,-0.2456701696,-0.2272544652,-0.312632829,0.1525451243,0.3414388597,0.1171690524,-0.2383913994,0.2780508697,-0.0030252272,0.294416815,-0.0667823926,-0.1000752151,0.3630546033,-0.4525365531,0.4308469892,-0.0333026275,-0.2724324465,0.1023563296,0.170372352,0.1682206392,0.3327376842,0.2434000671,0.0377826989,0.1129044071,-0.0568233468,-0.1521762311,-0.1492474228,0.0324970111,0.1975148469,0.1152206138,-0.0478092283,-0.2434740812,-0.4107986093,-0.0237178244,0.0472183973,0.1210914105,0.0176356751,-0.185810551,-0.4173747301,0.0242550895,-0.4365698695,0.2714166045,0.081738539,-0.1642325521,-0.2187956274,-0.1694921851,-0.0801295638,0.0094462857,0.5412454009,-0.194412902,-0.1243834868,-0.1509480327,-0.1719124764,-0.214743495,0.3027726114,-0.1361378431,0.0351208486,0.0809627771,0.3884940147,-0.1651658863,-0.0014951441,0.4199516773,0.2146371603,-0.2711303234,-0.2036134899,-0.2595193982,-0.266092509,-0.2679540217,0.1394275129,0.0362481847,0.0924399272,0.4106515348,-0.2905561328,0.0161661077,-0.177614823,0.1379458755,0.4061155915,0.1916994601,0.4186710715,-0.0198619179,0.2423336059,0.1439825296,0.1078943983,0.3881815076,-0.2044390887,-0.124441646,-0.0786233246,-0.2945462167,0.5106819868,0.1328920424,0.0991695449,0.3325246274,0.0147336544,0.1614180952,-0.2695331573,0.3139098585,0.1731738895,0.1405242532,0.0327329189,-0.1346599907,0.4818721414,-0.0100157922,-0.1039175391,0.2979790568,0.3700315356,-0.2019829601,0.0462892838,-0.0542661399,0.7508448362,-0.1673716754,-0.0357275046,0.0385224,-0.1774643809,0.2226021141,-0.1097051203,0.1153547689,-0.1414429992,-0.3266515732,-0.0257296059,0.1210597232,0.2492420673,0.0732878372,-0.0535686389,0.0528168827,0.057418488,0.1537076533,-0.1922420859,0.0704625547,-0.0091985343,-0.2924908698,-0.2694923878,0.3007312715,-0.0104220295,-0.0957230926,-0.0762965754,0.1155964658,0.0195107702,-0.2082198411,-0.1271173656,-0.3215965331,-0.0585806072,-0.1984926909,0.1279903799,-0.2477897406,-0.2515174448,0.1707946658,0.4153106511,0.0573318042,-0.2216247171,0.2055685967,-0.2225225568,0.0856782645,-0.2475019544,0.0399667099,0.1094514877,-0.1779582947,-0.2346386909,-0.0101332134,0.1052416712,-0.4131995142,-0.0338254981,0.0321347639,-0.2425576895,-0.4690617919,-0.1052210927,0.0798239037,0.1773300618,-0.3729863167,0.2656607926,0.1998648643,-0.1658198833,-0.0440213419,0.3708614409,-0.1272996813,-0.1514552981,0.3287456632,-0.0637816042,0.0696693435,0.2002505362,0.0091878623,-0.1790071279,-0.3545342088,-0.0733047277,0.4246886373,-0.1535011381,0.1729310602,0.2348183692,-0.2309894264,-0.0066764364,0.3626250923,0.0038933882,0.1581724882,-0.0481379852,-0.1246738657,-0.429834187,0.0749493018,0.0049051759,0.3372640312,-0.032798145,0.2446377426,0.0093610864,0.169631809,-0.5226578712,-0.0651676133,-0.2895472646,0.2796230018,0.3717902005,-0.0287870169,0.208365947,-0.1875355244,0.3329309821,-0.0352039449,-0.1669900864,-0.3831592798,-0.2214453965,0.0608606525,0.0196374021,0.1678806096,0.0087600872,0.1811007112,-0.0178544819,-0.1832400262,0.3455120623,0.1026549861,0.0202968642,0.2028232068,0.2382941395,-0.0692555681,0.1074714959,-0.1394740343,0.0720087513,-0.0467459261,-0.1000197977,-0.0731560886,-0.3047165573,-0.049512364,0.3234211802,0.1798161715,0.2336516529,-0.0063012061,0.1078683808,0.1271710694,-0.0252147242,0.3240079582,0.2929454446,-0.0965291783,0.0720095113,0.2770685554,0.111819379,0.4103547931,-0.4879571199,0.175447762,0.3631243706,-0.1352904737,0.5029091239,0.2272493541,0.0658412203,0.2021551132,0.0913497433,0.0503357947,0.0012862938,-0.211471349,0.0581787005,0.1365847588,-0.0296758618,-0.0812325776,0.47530514,0.2625532746,0.1824615747,0.1179975644,-0.2786803842,0.2116145939,-0.0583927892,0.0420126431,0.1983387917,-0.1137420982,0.3527482152,0.1639513969,0.1773966551,-0.0448389091,0.1969461292,0.412995398,-0.0848411694,-0.1146609411,-0.1688688099,0.2937822342,-0.1005148515,-0.2533902228,-0.1573494971,-0.1492339075,-0.0493522733,-0.0913517699,0.1106574163,-0.1992681175,0.3730932772,-0.1738866717,0.1257230639,-0.3164148331,0.082453385,-0.0605836362,0.0700601414,0.0404517725,0.1876727343,0.1593514085,-0.0599557385,0.3969538212,0.3891629875,-0.028474059,-0.3357331455,0.2878659666,0.4870365262,-0.2670856714,-0.3102598786,-0.1628081948,-0.0846298561,0.0545227267,0.2395138144,0.1223445684,0.3468786478,-0.3152138889,0.2867630422,-0.0915715918,0.1400623322,-0.1972039193,0.0750786215,0.0684304163,-0.3613371849,-0.1958664954,-0.1309572309,-0.2448277026,-0.158496961,-0.0164297093,-0.2886989117,0.1080749184,0.0048587476,0.1596200913,0.1991929412,0.2223717719,0.0157021359,-0.1388692707,0.0664106086,-0.2817444801,-0.4920034409,0.0352377668,-0.0334379375,-0.270737797,-0.1485718042,0.068599306,0.2237577438,0.1569606662,0.3889604807,-0.1579192877,0.0279865023,-0.2279443592,-0.2847083211,-0.0991776809,0.0880907997,0.0599071644,0.1327790618,0.1024205536,0.0482950285,0.1003052816,0.1921440661,-0.2107080072,0.1916915029,-0.2847426832,-0.2442822158,0.1931996495,-0.0159533322,0.2518629134,-0.3791727126,0.1583206505,-0.017983431,-0.1998200268,-0.3959464431,0.0735316947,0.1107605249,0.5224793553,-0.0584191941,-0.2186151892,0.0750891045,0.209973067,0.051511813,-0.0029987881,-0.1215342507,0.5358163714,0.0955800638,-0.1547905952,0.0824593902,0.2031573504,-0.040305201,0.0380424969,-0.0200888943,-0.4734790027,0.4227631986,-0.1249492317,-0.201923877,0.0458445549,0.0944090709,0.2825827897,-0.1984641701,-0.5064719915,0.1899568588,0.2214783877,0.0548424087,0.0497555733,0.349491924,-0.1279300302,0.0390071124,-0.0718112364,0.4237104058,0.2931393385,-0.1638945192,-0.0361737348,-0.2457034588]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1949","title":"Enable Fast Filtering using Arrow Dataset","comments":"Hi @gchhablani :)\r\nThanks for proposing your help !\r\n\r\nI'll be doing a refactor of some parts related to filtering in the scope of https:\/\/github.com\/huggingface\/datasets\/issues\/1877\r\nSo I would first wait for this refactor to be done before working on the filtering. In particular because I plan to make things simpler to manipulate.\r\n\r\nYour feedback on this refactor would also be appreciated since it also aims at making the core code more accessible (basically my goal is that no one's ever \"having troubles getting started\" ^^)\r\n\r\nThis will be available in a few days, I will be able to give you more details at that time if you don't mind waiting a bit !","body":"Hi @lhoestq,\r\n\r\nAs mentioned in Issue #1796, I would love to work on enabling fast filtering\/mapping. Can you please share the expectations? It would be great if you could point me to the relevant methods\/files involved. Or the docs or maybe an overview of `arrow_dataset.py`. I only ask this because I am having trouble getting started ;-;\r\n\r\nAny help would be appreciated.\r\n\r\nThanks,\r\nGunjan","comment_length":113,"text":"Enable Fast Filtering using Arrow Dataset\nHi @lhoestq,\r\n\r\nAs mentioned in Issue #1796, I would love to work on enabling fast filtering\/mapping. Can you please share the expectations? It would be great if you could point me to the relevant methods\/files involved. Or the docs or maybe an overview of `arrow_dataset.py`. I only ask this because I am having trouble getting started ;-;\r\n\r\nAny help would be appreciated.\r\n\r\nThanks,\r\nGunjan\nHi @gchhablani :)\r\nThanks for proposing your help !\r\n\r\nI'll be doing a refactor of some parts related to filtering in the scope of https:\/\/github.com\/huggingface\/datasets\/issues\/1877\r\nSo I would first wait for this refactor to be done before working on the filtering. In particular because I plan to make things simpler to manipulate.\r\n\r\nYour feedback on this refactor would also be appreciated since it also aims at making the core code more accessible (basically my goal is that no one's ever \"having troubles getting started\" ^^)\r\n\r\nThis will be available in a few days, I will be able to give you more details at that time if you don't mind waiting a bit !","embeddings":[-0.0609657727,-0.2426414639,-0.1764556915,-0.0540162101,0.0382519737,-0.2388272434,-0.1074126586,0.2039251029,0.2231049538,-0.184539929,-0.2857308686,0.4401746094,-0.1111105308,0.381758064,0.0005044458,-0.22477974,-0.1491404474,-0.1010928676,-0.1609327197,-0.0210787263,0.2228640169,-0.0318532549,-0.1598431319,0.0115776397,0.0991677195,-0.0908600837,0.4640896022,-0.0603517592,-0.386195749,-0.4257190824,0.2159529924,0.4622389376,-0.2772480249,0.3484126031,-0.0001159239,-0.1035579219,0.3350031078,0.0698489174,-0.167813465,0.1456142962,-0.2793073058,-0.4648908079,0.346085012,0.0931740031,-0.2401725352,-0.3023802638,-0.4887275398,-0.5661330223,0.4217899442,0.1795723736,0.1517729014,0.1879249364,0.1759025455,0.0575222075,0.1689590365,0.3293296695,-0.2442489564,-0.0968683958,0.7156175971,-0.1107307076,-0.0707974359,0.4140395224,-0.0608403124,-0.1602589488,0.3494179249,-0.2657895982,-0.1116122901,-0.5176677108,0.2471853048,0.3313281834,0.4620250762,-0.1396792382,-0.3891334832,-0.2872939408,-0.2797900438,-0.1300293356,0.0210040454,-0.0488146618,-0.1954260319,0.1936593801,-0.1636239588,-0.2893832922,-0.3083898127,0.022700768,0.2393592298,0.2634184957,-0.2676101327,-0.0376471356,0.2280462086,-0.3477035165,0.0239024516,-0.1890485585,0.099835366,0.3675711751,-0.2276188582,-0.0771989822,0.1686148793,0.4670672715,0.3032642603,0.1906511188,-0.1145789102,0.3745818138,0.0616545156,0.0119904717,-0.0012414742,0.0475068949,-0.0682887137,0.2453008741,0.2181309611,-0.0863909274,-0.0993872434,0.0604786724,0.0828244314,-0.2747676075,-0.146286279,-0.3215171993,-0.0542428493,-0.4708516598,-0.0017119346,-0.1667616069,-0.0813663006,-0.2673496008,0.4737507403,0.2197330296,-0.119538784,-0.057073649,-0.2483893186,-0.1011037752,-0.1768927872,-0.0029422245,0.0027063391,-0.1255301237,-0.1310469955,-0.1374052763,0.1485551447,-0.1948987246,-0.0996300355,0.0176745374,-0.033111807,0.2060005218,0.2575750649,0.0909352601,0.3814996183,0.3757706285,-0.4013728201,-0.0247621648,0.1775443703,0.0400693007,-0.2153909206,-0.003656595,-0.2171408087,-0.5327579975,0.0032878898,0.0952740386,-0.2179053277,-0.1383990645,-0.1729111969,0.6192804575,-0.2132295072,-0.000494361,0.1261023283,0.0303856488,-0.0746977776,-0.2937876284,0.3139742315,0.2567203045,-0.4902034402,-0.0608596317,-0.2886763513,-0.1347985417,0.2083828896,0.1620910317,-0.4143380523,-0.2132012546,0.1621505618,-0.0254688319,0.5568774343,-0.1211442128,-0.392944634,0.0143117821,-0.0648195371,0.3086546957,0.0995056629,0.0061432607,0.2001383901,-0.153079167,0.0048772707,0.6350794435,-0.165749684,-0.1484430283,-0.3304502964,-0.093700029,0.0300593339,0.0449815914,-0.0954458714,-0.0510530658,0.1706358194,-0.5012762547,0.1360241324,0.0712873787,0.175914079,-0.0128359497,0.470792383,0.2733278573,0.3618062735,0.217262134,-0.2060662359,0.2344364673,0.0301132947,-0.1215244457,-0.4391415715,-0.0571952984,-0.0998980775,0.1922446489,-0.1285838932,-0.0275910627,0.0785058811,-0.2916638255,0.2678441703,-0.1258757561,-0.364239037,0.273060739,-0.123048082,0.0722340941,-0.1536018401,0.0954929069,0.237029165,0.1554512978,-0.01025589,-0.0900421441,-0.0304583143,-0.1076329127,-0.1002159044,0.0061865002,0.1352829635,0.2469383776,0.4057640433,0.5706692934,0.1249560863,-0.4567500949,0.4129880667,0.0831520483,0.166469723,0.0793097764,-0.2738563716,0.7109128833,-0.4960452616,0.1653510034,0.1060892344,0.0411678143,0.0584857166,0.1099969447,-0.3593268991,0.1287246197,0.299886167,-0.3171958029,0.2916972637,-0.0638211817,-0.1837740988,0.4710711539,-0.1567672789,0.1237304658,-0.245022282,0.3046881557,0.0002448684,-0.1571689844,0.0060961582,0.3676767945,0.0476040766,0.3104174137,0.0743715018,-0.1966027021,0.1642754972,0.0204417203,0.3753753603,0.0974472016,0.3182750642,-0.0777941868,0.2552367449,-0.0056444858,-0.2419352978,-0.3230299652,-0.1648765206,0.1371055394,-0.1429522038,-0.0491912439,0.1000470817,-0.3030164838,0.0799470171,-0.4713403881,-0.0805584267,-0.0791645572,0.1630918831,-0.1481856406,-0.059070453,0.1847274005,-0.3384579718,0.2940368652,-0.0116723478,-0.2430040538,-0.3440419436,-0.1907553971,0.1860200167,0.0658502355,0.0843954086,0.2708716094,0.3436901867,0.391939044,0.0891071483,-0.110916324,-0.6208744049,0.0376508906,-0.0244235471,0.3294962943,-0.0550893992,0.1703158319,0.0902501643,0.0219805539,0.2207974046,-0.183091417,-0.0867154375,-0.2048915476,-0.2004709989,0.1766234487,0.1549229771,0.0479071029,-0.4060262442,-0.3476777077,0.2588637173,0.0216842704,0.1105113551,0.1468440294,0.1031279936,0.1100967228,-0.307074666,0.2067844123,0.0703001544,-0.0732146651,0.2750939131,-0.1457663327,-0.2073541731,-0.1615960747,-0.2363073826,-0.0982559249,0.314291507,-0.1593430787,0.0102773383,-0.4495514631,0.2834903598,0.0864341035,0.0271232463,0.3345820904,0.0846127793,0.0097950781,-0.2262037843,-0.2326354235,0.0154636921,-0.1563065797,0.1400000751,0.3474459052,0.3566597104,-0.0228054523,0.7490887642,0.2729087174,0.3102783561,-0.0216605067,0.2651241124,-0.0104615269,-0.004716868,-0.3897791207,-0.2076573074,-0.1832052171,0.0713240132,0.0647226349,0.4708165526,-0.2996092737,-0.2581763268,-0.1289068311,-0.2300778627,-0.1261652857,0.1658201218,-0.081925489,0.2298600078,0.0215348955,-0.2481342405,-0.3278767467,0.0212046038,0.0939893126,0.1197006404,0.3694306612,-0.1155922115,-0.3070031703,-0.0110324491,-0.4549897909,0.3740399182,0.0163287818,-0.0884356946,-0.0765636936,-0.1763864011,0.1765044183,0.0622372255,0.4188169539,-0.1716429889,-0.2019245028,0.0084134899,-0.2102384865,-0.3476477265,0.3733870983,-0.1467583925,0.155193001,-0.1636204571,0.1940013021,-0.2316623479,-0.4352599978,0.0153539646,0.0025666237,0.1551010013,0.1351790279,0.1571701318,-0.2165170014,-0.0974028483,0.2244462222,-0.0569239631,-0.1157417819,-0.0190441217,0.1077827662,0.0715728477,0.2140097618,-0.0328818932,0.0305704698,0.1167550534,0.1374806166,0.1845524907,0.2986932993,0.0114758415,-0.0831662491,0.4986408651,0.0181760062,-0.1970019788,-0.4065933228,0.3767666221,0.4093343616,0.6419485211,0.0083802128,0.4672681391,0.1627712101,0.1252395362,0.0951696038,-0.2514255047,0.0303575303,0.1789841354,-0.4732816517,-0.4895405769,0.5944274068,0.0797559619,-0.2069953233,0.6742953658,0.4888080359,-0.02136405,0.273150444,0.031781368,0.9381295443,0.1353502274,-0.2550575435,-0.055171784,-0.457287848,0.2276633084,0.1907256246,0.0677271634,-0.1280806065,-0.028097745,0.0904291198,-0.163553983,0.2674819231,-0.0624958314,-0.1392671764,-0.1324787438,-0.0148850568,-0.362768203,-0.1140649915,0.1413881034,0.1497001499,-0.3225058615,-0.4596540332,0.1375404149,0.029719837,0.2756161392,0.1242972165,-0.336002171,-0.1396521777,0.0116644176,-0.1610211581,-0.2538186908,-0.4586703777,0.0900092572,0.3469236493,-0.689471066,0.4541460872,-0.2189528346,-0.0276077632,0.0240766741,-0.2672510445,0.3191041946,0.3087886274,0.0923413187,0.1026990861,-0.0448272899,0.4351013601,-0.1017229557,-0.1091320813,-0.1032364368,-0.2462429851,-0.3320293725,-0.12454772,-0.1183082685,0.4498850107,0.0405549444,-0.1221791282,0.186303094,-0.2323570102,-0.0750323236,0.1450913101,0.2975463271,-0.1378024817,0.3352734745,0.2150527686,-0.0183797814,-0.0883290246,0.1541642994,0.0119425515,-0.2380981892,0.1957037449,0.4382343292,-0.3694276214,-0.0726265386,-0.042376399,0.4137723744,-0.3821433187,0.0919382423,0.0623241365,0.0000772028,0.3333826661,0.279731065,0.0447161458,-0.1187899262,-0.2067807466,-0.2249266654,-0.2354303598,0.2200051248,-0.2416066527,-0.0036625247,-0.6351540089,-0.3681564927,0.1231980398,0.2730611563,-0.3414292634,0.1466785371,0.1788661778,0.0029465612,0.106255509,0.1206168681,0.2502305508,0.0875689462,0.0558757633,-0.161180228,-0.44214046,-0.1634098589,0.0420132875,0.1547056437,-0.227390781,-0.3869334757,0.1276437491,-0.2015352547,-0.4394991994,-0.024872439,0.0112368045,0.1066522971,-0.008705968,-0.4182301164,-0.1489040554,-0.3942814767,0.1087362692,0.1185578853,0.181875363,-0.0532115623,0.2519565225,0.0725565925,-0.0979097709,-0.2889054716,0.0683789849,0.2930141389,0.0831355825,-0.1490927637,0.3740784824,-0.3148295879,-0.0208638981,0.0578756072,0.4791154861,0.1315644979,0.0495792516,0.2891439795,0.2545942366,0.1510763913,-0.2919922769,-0.0689227283,0.2296595275,0.0560886636,0.1602949947,0.0991424322,0.1550763994,0.3570220768,0.1556583047,0.1796741784,0.1728068441,-0.1825162321,0.0343784913,0.0758160502,0.2442071736,-0.2312833369,0.5170527101,0.4599158168,0.1527625173,0.2112317085,0.0077615706,0.3438524306,0.1532796621,0.3685179055,-0.0573913306,-0.3248311281,0.1044099256,0.1820527911,-0.3671522737,0.1109513491,0.0719504505,0.4479805529,0.2105858922,-0.0966868997,0.1776969433,0.2294197381,0.1842699051,-0.1984756142,0.2682887018,-0.4554393888,-0.0567954108,-0.0700842366,-0.1246259883,0.1792718917,0.5062457323,-0.033642564,0.4923925102,-0.3771255314,0.0177505743,0.1347044408,0.5880489349,0.0054553123,0.1343386769,-0.0756400898,-0.0395638943,0.0040604603,0.1912315041,0.2396367788,0.0882498845,-0.2215311527,0.1217186451,-0.1278241873,-0.3407157063,0.0758996531,0.5048363209,0.301821053,-0.3505644798,0.0060389871,0.0432858057,-0.1411370337,-0.0027779448,-0.1935842633,-0.0184468031,-0.0177154429,-0.077614069,0.0754837394,-0.1306012422,-0.4775841832,-0.1475538164,-0.1269415319,-0.2714190185,0.2402140796,-0.2049629837,-0.1174962968,-0.0226437598,0.0711854547,-0.0046641463,0.5002954006,0.2632885575,0.3848914206,-0.084428221,-0.3183431029,-0.2290221155,-0.3159655333,-0.1155213714,0.1433747411,0.2399309427,0.0960694328,0.3070605993,0.4269804955,0.1642241478,-0.3078041375,0.2154697627,0.1345229298,-0.229724139,0.2803108394,0.2107987851,0.3605071306,-0.1337548643,0.0139040789,0.0631632283,-0.3434786499,0.0417119004,-0.1177065223,0.125227198,0.273296535,0.0424076058,0.0395915993,0.1712906063,0.2689296007,0.1354904175,-0.0002494914,-0.2420800626,-0.1890338361,-0.0756736547,0.163183704,0.1734727174,0.3652724624,0.0811430886,-0.4510500431,-0.039122127,-0.0669403672,-0.2234858721,-0.0710233599,0.0192720741,-0.0203316174,0.0477410741,0.0642387569,-0.2072569728,0.2440374941,0.1739787012,0.0113017242,-0.4065947235,-0.2832552195,0.6696619391,-0.4198071957,-0.2190530449,-0.1257772744,0.065654397,0.3203105032,-0.1362008154,-0.2478348017,-0.1276458204,0.004893472,-0.2941203117,-0.0921106488,0.3316379786,0.0929634348,0.0761645585,-0.0724547282,0.1414701492,0.172723949,0.2042236626,-0.4060373902,0.0155449742]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1949","title":"Enable Fast Filtering using Arrow Dataset","comments":"Sure! I don't mind waiting. I'll check the refactor and try to understand what you're trying to do :)","body":"Hi @lhoestq,\r\n\r\nAs mentioned in Issue #1796, I would love to work on enabling fast filtering\/mapping. Can you please share the expectations? It would be great if you could point me to the relevant methods\/files involved. Or the docs or maybe an overview of `arrow_dataset.py`. I only ask this because I am having trouble getting started ;-;\r\n\r\nAny help would be appreciated.\r\n\r\nThanks,\r\nGunjan","comment_length":19,"text":"Enable Fast Filtering using Arrow Dataset\nHi @lhoestq,\r\n\r\nAs mentioned in Issue #1796, I would love to work on enabling fast filtering\/mapping. Can you please share the expectations? It would be great if you could point me to the relevant methods\/files involved. Or the docs or maybe an overview of `arrow_dataset.py`. I only ask this because I am having trouble getting started ;-;\r\n\r\nAny help would be appreciated.\r\n\r\nThanks,\r\nGunjan\nSure! I don't mind waiting. I'll check the refactor and try to understand what you're trying to do :)","embeddings":[-0.1475265026,-0.0889887959,-0.2098283023,-0.0918968394,0.0624384396,-0.2331448346,-0.0470115282,0.2564979196,0.1409168392,-0.165285185,-0.2117004097,0.5572015047,-0.1610823572,0.2340772152,-0.0721702799,-0.1115348861,-0.163025409,-0.0334027894,-0.1529833227,-0.0570407026,0.160586223,-0.0926778093,-0.1693582982,-0.0569531918,0.0641346052,-0.0823348612,0.4899257421,-0.1223512962,-0.3783577681,-0.5257975459,0.180488348,0.4125720263,-0.2424077839,0.2819820344,-0.0001122101,-0.0977285206,0.3756682873,0.0460015684,-0.1633183658,0.2018047422,-0.3833836019,-0.5163570642,0.3023416102,-0.0402294919,-0.2140232474,-0.3167768419,-0.4475963414,-0.6233633757,0.3333078921,0.2781315446,0.1977690309,0.0779646039,0.1296117902,0.0343558192,0.2299921364,0.1702783704,-0.2757145464,-0.1611308753,0.690895319,-0.1348021477,-0.1131289452,0.4136619866,-0.0534187928,-0.1600668281,0.2612384558,-0.3271079063,0.0970113128,-0.5140107274,0.2721892297,0.244067207,0.5241881013,-0.0686452687,-0.2503685355,-0.1154986694,-0.295946151,-0.0541290008,-0.0665444285,-0.059307754,-0.1523031741,0.195379585,-0.1410022974,-0.3106043935,-0.3834908009,0.0470867008,0.1975131333,0.3864691257,-0.1185574308,-0.0790665895,0.1377027333,-0.3078043759,0.2436964065,-0.1139409989,0.0077825147,0.3805395961,-0.2030074298,-0.0627390966,0.1559075713,0.4094288051,0.2585114539,0.1290134341,-0.0537548661,0.3599945307,0.1209415272,0.0324548408,-0.0365011655,0.0077360962,-0.0455789194,0.4321999252,0.150579989,-0.2071461827,-0.2090796083,0.0376978219,0.0157882031,-0.2357439697,-0.0105468379,-0.2417128086,-0.0886400864,-0.4509317875,-0.0292536505,-0.1328237951,-0.148855716,-0.2649073601,0.3727755249,0.2461594939,-0.0908480585,-0.042247802,-0.2776985765,-0.1313280314,-0.2057754397,0.0612553358,0.0159739405,-0.1020879522,-0.1534935832,-0.290271312,0.1802892983,-0.0157666206,-0.10327252,-0.0190820396,-0.0743510872,0.2151860148,0.3460691571,0.0992514789,0.411332339,0.3538830876,-0.4210656881,-0.0184229221,0.115955621,0.0654646084,-0.2028469443,0.1139784157,-0.2340939492,-0.5071300268,-0.0122117652,0.1530790627,-0.1667325348,-0.1694045961,0.0689224526,0.5219022632,-0.1484542191,-0.0539499,0.130165413,-0.0499783121,-0.0169830658,-0.3549636602,0.1976926029,0.1616173834,-0.5772004724,0.0571639501,-0.3905580044,-0.0376618318,0.2938017547,0.1009637266,-0.4094146192,-0.2548648417,0.1447502971,-0.0853995532,0.6621947289,-0.1478880197,-0.5032480955,0.0212403536,-0.1166824177,0.2362808585,0.1395219266,0.0708171725,0.2376329005,-0.1459852606,0.0106067592,0.5886858702,-0.1571557671,-0.1358809769,-0.3851483464,-0.0330214538,0.0929226056,0.0147239659,-0.0122914733,-0.0489020683,0.2204797417,-0.3853618801,0.1980082989,0.0757929757,0.1681527793,-0.046123866,0.5657210946,0.1406765878,0.3982255459,0.2781224549,-0.0813048854,0.2465830594,-0.0564617142,-0.1005464718,-0.3765888512,-0.0060677938,-0.1080690846,0.1723631918,-0.1178417206,-0.0534229502,0.14563407,-0.2586216033,0.1876114011,-0.1119330898,-0.2767269313,0.1864242703,-0.1401674151,-0.0221047364,0.0777286291,0.0473793671,0.1509610713,0.1477293372,-0.03628847,-0.2032631934,-0.0335153602,-0.0979410037,-0.1415888965,-0.0482927598,0.1992948055,0.2076708525,0.3779627681,0.5796352029,0.0671273619,-0.4598431289,0.4000549018,0.1309711635,0.1228171587,0.0600435212,-0.2230873108,0.6486538649,-0.4913371205,0.1301896721,0.0743035227,0.0571306497,0.1209030896,0.091638729,-0.3350029886,0.1061155424,0.2322738767,-0.2754030228,0.2704346478,-0.0466508307,-0.1053303406,0.4582201242,-0.1308725029,0.2124437094,-0.2018691301,0.2182039022,0.0384851173,-0.137300089,0.0497142747,0.4776943922,0.0552189313,0.3242245615,0.1072229967,-0.2351413369,0.077988036,0.0317963623,0.4003838301,0.0332389995,0.3895933926,-0.0226081554,0.2586921453,-0.0754904523,-0.1984192878,-0.3335967362,-0.0943608657,0.1433903277,-0.0976475477,-0.1084431633,0.1243613362,-0.2786762416,0.0896425322,-0.4824868739,0.0609073527,-0.1182475463,0.1991486549,-0.2700444162,-0.0258981995,0.2583579123,-0.476493448,0.2215439528,0.0698770285,-0.2258888185,-0.2692493796,-0.2329246998,0.1807035804,0.118802242,0.0096916314,0.2544655204,0.3076722622,0.4234210551,0.185579434,-0.0607866235,-0.5851409435,0.0178731307,0.0574484468,0.2471532375,-0.1005003303,0.0984463096,0.0665504485,0.0384896025,0.1948616058,-0.1776289344,0.003502131,-0.2151720375,-0.1799066067,0.1618772596,0.1814917326,-0.007470869,-0.402543813,-0.3848935664,0.0839819536,0.0953602344,0.1507592052,0.0864581913,0.1575749964,0.0498173162,-0.21996966,0.1554563046,0.1019880101,0.0315358602,0.3705923855,-0.1950341463,-0.1830601543,-0.1638777405,-0.2358260602,-0.0301395897,0.3896688521,-0.1173681915,0.1325734407,-0.3845604062,0.2650758922,0.0392279178,-0.0429855809,0.2871629,0.1398596913,-0.0358021371,-0.2890634239,-0.0778519958,0.0468321629,-0.1745110005,0.1589364707,0.2671407461,0.2515628338,-0.0135229062,0.6781641841,0.2385448664,0.3053124845,-0.0278060716,0.2382785678,-0.0804032236,0.0050881519,-0.363689661,-0.2951539457,-0.1504167318,-0.0515609607,0.0314597972,0.3922911584,-0.5023056269,-0.2534084618,-0.1076097116,-0.174099192,-0.0939800814,0.2185051292,-0.2374886423,0.2750197053,0.0177654214,-0.1958895624,-0.3312814236,0.0534339547,0.0893674791,-0.0237316992,0.3159025609,-0.1551207006,-0.2472109497,-0.0224772692,-0.3784748912,0.318292737,-0.0073684175,-0.0131222131,-0.0167227294,-0.1759759337,0.1313333213,0.0682744756,0.3394864798,-0.2564072907,-0.1299345046,0.0996434093,-0.065461345,-0.2634764016,0.3052334785,-0.1720013767,0.2245872021,-0.1720325798,0.1267334521,-0.0689731985,-0.249893412,-0.0183734205,-0.0665832534,0.1111677065,0.1037795544,0.2244229913,-0.1516264081,-0.073159039,0.2071251422,-0.0555613041,-0.1221343055,-0.0203145165,0.0570159219,0.0557578243,0.2013000697,-0.069921203,-0.0022001185,0.1236116737,0.0978464484,0.1559082568,0.1842034012,0.0010705286,-0.0571026728,0.4396472275,-0.0187204517,-0.0930866301,-0.4215840697,0.3052022755,0.417478472,0.544275701,0.0049808575,0.3545251191,0.0523051098,0.0445025787,0.2079078257,-0.27359429,0.0200398117,0.081698589,-0.4799916446,-0.5529648662,0.5566237569,0.044552654,-0.2231656462,0.6911590695,0.276450336,-0.0742662475,0.3161837459,0.109255977,0.8662745953,0.1756535321,-0.3162346482,-0.1896562725,-0.4516646862,0.1645269692,0.1649383307,0.0966427773,-0.155504033,-0.0197433587,0.1007922441,-0.157529816,0.2383474559,0.0757960752,-0.1866053045,-0.1635295451,-0.0692259669,-0.4089564979,-0.1013877392,0.156400308,0.1919609904,-0.236952126,-0.3374944031,0.1550123692,0.0376734249,0.1941005886,0.1098063141,-0.2859632671,-0.1854332238,0.0339033119,-0.1372964829,-0.2098491341,-0.4218374789,0.2017422915,0.1767585129,-0.6858209372,0.4032747149,-0.2590159774,-0.1760679185,0.0660737157,-0.2633850276,0.3766176999,0.4508697093,0.0541923046,0.1525779068,-0.0431578085,0.3689368069,-0.1130989566,-0.1333814561,-0.096992746,-0.340468049,-0.3379002213,-0.2013331503,-0.0738957673,0.5246477723,0.1063145399,-0.1258167922,0.1729200333,-0.3183770776,-0.0435608141,0.2024710476,0.2792018056,-0.1081268862,0.4356886744,0.2355086207,-0.0234056674,-0.0776552856,0.1419811547,0.1190922186,-0.1750055403,0.1872933358,0.3225567639,-0.3461342752,-0.116336152,-0.0046142139,0.3299773335,-0.2947348654,0.0382042602,0.0107051916,0.0373415947,0.3760463297,0.2310846597,0.0844665319,-0.0195484422,-0.3427930474,-0.1993971169,-0.2802716494,0.133384943,-0.2288027257,0.0179556627,-0.698708117,-0.3938132226,0.1250784695,0.2507772744,-0.3786206841,0.1363939941,0.1850523502,0.0663276464,-0.0127426628,0.1230914146,0.1247168183,0.1519988179,0.1075395867,-0.1399763525,-0.36956954,-0.1927404553,0.1210191175,0.1453264654,-0.1919303238,-0.3453676105,0.0982165858,-0.2254561186,-0.4630486667,0.0646004081,-0.1146150678,0.0931694433,-0.0822216347,-0.3298394084,-0.1761529297,-0.4999929667,0.1140099764,0.1175862923,0.1143500358,-0.0884717628,0.2574737668,0.1593891978,-0.0012377779,-0.2922892869,0.0625136644,0.2208314687,-0.0397202484,-0.0318804868,0.4271565378,-0.2684899271,0.0735095069,0.011983267,0.4289993346,0.0997653008,-0.0149694243,0.2136553675,0.237022087,0.2167619467,-0.2880711257,-0.1528851986,0.2142183185,-0.014134286,0.1783148199,0.0957960859,0.1697105467,0.2545417249,0.1227240339,0.1712327302,0.1429409683,-0.2657503784,-0.0098025678,0.2091660649,0.1680659056,-0.2541534603,0.4537973404,0.3679922223,0.1609521508,0.3294355869,0.0872919261,0.4186424017,0.2954578698,0.3099860847,-0.0988232046,-0.224089697,0.1348628849,0.1933923364,-0.3778963685,0.1893369555,0.1609071791,0.2772226632,0.2905339003,0.0100825503,0.1538794041,0.1100168973,0.1400520355,-0.1770974845,0.2699897587,-0.4617323577,-0.0446575135,-0.0875977054,-0.1264299303,0.3221154511,0.4303367734,-0.0477113053,0.4564765394,-0.2363159508,0.1498155445,0.0908822268,0.5641852617,-0.1071926728,0.0874126628,-0.1085968539,-0.0462369435,0.0002309459,0.0986825526,0.2250643969,0.2264624685,-0.1615534425,0.0607653894,-0.0552968532,-0.3745658994,-0.0088110603,0.5079402924,0.3447353244,-0.3521858454,0.0400462933,0.0511242189,-0.1720167249,0.0784049407,-0.1419505179,-0.0302248914,0.086973682,-0.1116792858,0.1773302555,-0.2569562495,-0.4046736062,-0.1024260595,-0.0818587244,-0.3288163841,0.2311626524,-0.2292255163,-0.0891651958,-0.0596248023,0.0937024578,-0.0756825581,0.5075571537,0.1621739566,0.3664547801,-0.0580714308,-0.3708943129,-0.2053261399,-0.1831036955,-0.2422365695,0.1734284014,0.279635936,0.1507984549,0.283403784,0.410600096,0.2246815562,-0.2500520647,0.1204198971,0.2884011865,-0.2290094048,0.2530551255,0.2458895147,0.385976702,-0.1710741967,-0.049793411,0.080886282,-0.327778548,0.1301947236,-0.1378729045,0.034653686,0.4280061126,0.117198281,0.1686580032,0.1115588397,0.319814831,0.1346822679,0.0524808131,-0.2619460523,-0.2163193524,-0.0217665769,0.1465340108,0.2841082215,0.3330197036,0.1076078713,-0.4194912016,0.0117204292,-0.1147373393,-0.3243845105,-0.0019164677,0.0411204882,-0.0516378358,0.0358678363,0.0699704662,-0.3306314945,0.1953562349,0.1989713907,-0.0378248431,-0.4084860682,-0.2806215286,0.5808033943,-0.3440419734,-0.2982541025,-0.0936832801,0.0427401587,0.1809702069,-0.027420247,-0.1848164052,-0.1665574312,0.1074378937,-0.3471882343,-0.0620344132,0.2725344896,0.1391054839,0.1319852471,-0.0350426063,0.0918883458,0.1254365742,0.2822115719,-0.554889679,-0.0043712636]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1948","title":"dataset loading logger level","comments":"These warnings are showed when there's a call to `.map` to say to the user that a dataset is reloaded from the cache instead of being recomputed.\r\nThey are warnings since we want to make sure the users know that it's not recomputed.","body":"on master I get this with `--dataset_name wmt16 --dataset_config ro-en`:\r\n\r\n```\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-2e01bead8cf42e26.arrow\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-ac3bebaf4f91f776.arrow\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-810c3e61259d73a9.arrow\r\n```\r\n\r\nwhy are those WARNINGs? Should be INFO, no?\r\n\r\nwarnings should only be used when a user needs to pay attention to something, this is just informative - I'd even say it should be DEBUG, but definitely not WARNING.\r\n\r\nThank you.\r\n","comment_length":43,"text":"dataset loading logger level\non master I get this with `--dataset_name wmt16 --dataset_config ro-en`:\r\n\r\n```\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-2e01bead8cf42e26.arrow\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-ac3bebaf4f91f776.arrow\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-810c3e61259d73a9.arrow\r\n```\r\n\r\nwhy are those WARNINGs? Should be INFO, no?\r\n\r\nwarnings should only be used when a user needs to pay attention to something, this is just informative - I'd even say it should be DEBUG, but definitely not WARNING.\r\n\r\nThank you.\r\n\nThese warnings are showed when there's a call to `.map` to say to the user that a dataset is reloaded from the cache instead of being recomputed.\r\nThey are warnings since we want to make sure the users know that it's not recomputed.","embeddings":[-0.151090309,-0.3636071384,-0.0170671083,0.3475191295,0.4003514051,0.4622991979,0.4706430733,0.143284902,0.2523232996,-0.0047669052,0.0243790075,-0.0168783832,-0.2481703609,-0.1975330561,-0.3672408462,0.1944217086,-0.117108494,-0.0153791243,-0.3518670797,-0.0757232457,-0.1171182394,0.0906820446,-0.0266264919,0.5081018806,-0.7216521502,-0.0836909041,0.1400509477,0.1188435033,0.0099192094,-0.7821527719,0.2707004249,-0.128009975,0.1496111751,0.0640889332,-0.000122115,0.2060291022,0.6381807923,0.0515546985,-0.4871644378,0.02924826,-0.4754450619,-0.478941083,0.2521521151,-0.1042924523,-0.0297280084,-0.3696423173,0.1922228932,-0.4077891707,0.3286340237,0.1021776944,0.1139923781,0.1136904508,-0.1395543218,0.0326823704,0.3232573569,0.2458929718,0.037366733,0.0479511358,-0.0644855797,0.070087947,-0.2977598906,0.4409292042,-0.2956631184,0.052874893,0.240166679,-0.0134788146,0.2989350855,-0.2302694768,0.1710722744,0.3026792407,0.5181794167,-0.0112023316,0.1131071821,-0.4032346308,-0.2267826349,-0.0801946446,0.1673657596,0.1750006676,0.071756959,0.1927925944,-0.377902478,0.1259964406,0.1591122448,-0.0451121181,0.2188045681,0.3621397316,-0.3730171025,0.2935863733,0.1711701751,-0.0791650489,0.1143631712,-0.2527516484,-0.3305520117,0.1376243085,-0.2099725008,0.013834741,-0.1173723415,0.392984122,-0.1036160216,0.0236903653,0.1074151769,-0.0235768147,0.1798293591,0.1378366351,0.4080687165,0.0065904092,0.4469654262,0.0536668263,-0.1491612047,-0.1865981668,0.1863411218,-0.0538994484,0.0364211015,-0.140154317,0.5328091383,0.1328554749,0.1173180267,-0.151405856,-0.0866409466,0.0324665718,-0.1064751893,-0.2918034792,0.012836,0.1817286015,-0.0469556972,-0.0455386788,-0.1222740337,-0.0642848387,0.0570406467,-0.173807919,-0.0831567347,-0.3341276944,-0.3643180132,0.1104174107,-0.0247186851,-0.207774058,0.3294540644,0.1876461804,-0.1365977526,0.0210160818,0.109282136,-0.1964774877,0.1147880629,0.627643466,-0.228299126,0.2196660042,0.3440072834,0.1959386468,-0.1997303069,0.4353930354,-0.410954088,-0.4699918926,0.0184684638,0.0884320661,-0.1389927715,0.0448096395,-0.308349669,0.1366654783,0.3585318327,0.1251950711,0.2424760759,-0.1067552045,-0.3246652782,-0.1552183926,-0.1979091167,0.5811362267,-0.1848368049,-0.3025492728,-0.3280834854,-0.4322116673,0.2125727087,0.1218341365,-0.2136898041,0.2800593078,-0.3107714653,-0.0972535238,0.3082108498,-0.1050789505,0.0822921246,0.4469822347,-0.2502985001,0.1680027246,0.2125217766,0.0836042091,-0.2156756669,-0.1224220842,0.0135816662,-0.4536339045,0.0406977534,-0.1546110809,-0.1473661363,-0.1895036399,0.2960463166,0.0144892475,-0.2710362673,-0.0140961269,0.0367450602,-0.1554628164,0.2397705764,0.1271273047,0.1040881723,0.0431242064,-0.1647823155,0.0198250562,-0.0450833403,0.3280397356,-0.6091821194,0.0961715579,0.1165381074,-0.0305937845,0.0045087794,-0.0909322351,-0.0490077138,-0.0273274016,-0.3792020679,-0.1640519053,-0.0221960694,0.0016261688,-0.0596196055,-0.0850864947,-0.0751335546,0.2245910466,-0.5344884396,0.0863711834,0.0056907381,-0.1661028564,0.0117708649,0.3507805169,-0.2235099226,-0.1054344699,0.1355479658,0.0792681947,-0.1547690928,0.2173436284,0.1086671054,0.2987656891,-0.0428354889,0.3947451711,0.0307760108,-0.1995355487,0.4566539824,-0.1920306832,-0.1919047236,-0.1184583381,0.1470098346,0.1488821208,-0.0293637179,0.2458742708,-0.3589741588,0.0014749457,0.0724731684,0.1763548106,-0.1827459037,-0.5318390727,-0.0535599738,0.1382286251,0.2236326486,0.3524547815,0.0328621119,0.0614067689,0.5036969781,-0.0225223694,-0.0350022763,0.1725946218,-0.1172601059,-0.1267620027,0.2107435018,0.0995993614,0.400642097,0.1433577538,0.0964034051,0.2135555297,0.0398832411,-0.0287604891,0.2845728695,0.0011580769,0.2919470072,0.0412642993,-0.4688796699,-0.053495463,-0.2464312762,0.0426365919,-0.1107656434,-0.1338513196,-0.548420012,-0.1352434307,-0.1794122308,-0.2473459095,-0.3247794509,-0.2180150151,-0.2057373375,-0.4174848199,0.0159907546,-0.196761623,-0.3559078276,0.1864648908,-0.1796554178,0.2185807824,0.1087459549,0.2702391744,-0.3594065309,-0.2073978037,-0.186862722,0.0056629335,-0.1753740758,-0.1987687945,0.0720702484,-0.300937891,0.2761229873,-0.2969670594,-0.1735821217,0.1628869176,-0.3161101639,-0.0158542413,0.1541595608,0.0007796765,0.0341844596,-0.1384242177,0.0944360048,0.1207671911,-0.0642790124,-0.3195793331,0.1604953706,0.1289811879,-0.155172497,-0.2121983469,-0.3349220753,-0.1513204426,-0.1797146797,-0.3641599417,0.189716205,0.3864744008,0.002457493,0.1253540367,-0.2772358358,0.2045513988,-0.351603359,-0.6407558918,0.041935984,-0.0941983685,0.0447054617,0.0233052131,0.1199380681,0.1616583914,-0.0717071369,-0.8328558803,0.0837572068,-0.1131227836,-0.2497929931,0.2192757875,-0.0311910827,-0.0892969817,0.2955489457,-0.0040108827,-0.0155311041,-0.2937012911,0.0342544802,-0.4827652276,0.2337228656,0.0650281832,0.4955634475,-0.127476573,0.667124927,0.1786766946,-0.0275564641,0.197088927,0.0470813066,0.6038406491,-0.1499833167,-0.1548581421,-0.1112248451,0.0743583962,-0.1259296089,-0.0157104116,-0.0841409713,0.0476633944,0.1504987925,0.1654736698,0.0786557719,-0.2425961345,0.0733592659,-0.4269165695,-0.0912574455,-0.1446536481,0.0739908442,0.1281628311,-0.0123730442,0.0110633047,0.4801499248,0.3755212128,0.0773743466,-0.2766473591,-0.1589408517,-0.1690036356,0.11225456,0.091774419,0.067072995,-0.3882780671,-0.3064818978,-0.072123833,0.3549329042,0.1371835917,-0.1570132971,0.29493168,0.1318044811,-0.121684216,-0.1347224712,-0.2034214884,0.0879288614,0.1541792303,0.127638042,0.171756044,-0.0838699043,-0.3960535228,0.110046424,-0.1419891864,-0.1392142177,-0.1915456206,-0.2848087847,0.2992536128,-0.2038092911,-0.0032043837,-0.0412966423,-0.1334177703,-0.4892788529,-0.1212306023,0.1779166758,0.298081547,-0.0719581023,-0.100474,-0.018829789,0.3366833627,-0.0939927772,0.3638506234,0.6472281814,0.2555452883,0.7740367651,0.0363737494,-0.1040382087,-0.013920038,-0.0484956093,0.230755955,0.1523725986,-0.0274608694,0.0000711399,0.3212532997,0.3440138102,-0.4421436191,0.1337494552,0.2313803583,0.1668226719,-0.4083360434,-0.2868032157,0.5116586089,0.2142746449,-0.1043916121,0.5442043543,0.0711134747,-0.1615944058,-0.2187599838,0.3741802275,0.8942174911,-0.1471625715,0.1002799645,-0.0160290152,-0.007123223,0.1717582494,0.1440548003,-0.2678991854,0.007318439,0.102914311,-0.0925106406,-0.0707772002,0.1881528199,0.5290831923,-0.1787948608,0.0160599034,-0.2433728278,0.3730201423,0.0617842823,-0.0121208373,-0.0502513535,-0.0936081186,-0.1660182476,0.0052699484,0.104509525,0.2041032612,-0.2394821495,0.0460051037,0.37987414,-0.2689362466,-0.1060884297,-0.2011124045,-0.2168130875,0.1850458086,-0.0572742708,-0.047016833,0.119134739,0.0937589705,-0.023439344,0.1446029097,0.011947114,-0.1187982708,0.0615762547,0.0238210689,-0.3643181622,0.0179009046,0.2325511128,-0.0072844308,-0.2646377385,0.2048374861,-0.3840926588,-0.5043620467,0.0228942223,0.0711933225,-0.1922395229,-0.2036357075,0.0533132516,0.1699159741,-0.2309156954,0.2873203754,0.0293758325,0.1502049416,0.1965529174,0.0601568222,0.2314125746,-0.6483148336,-0.1805454046,0.2899333835,0.0787507221,-0.078715153,0.4794629514,-0.2117161751,-0.0789494663,-0.1393016428,0.0622333139,0.369577527,-0.0275955796,-0.113756977,0.3495698869,0.0395739526,0.1129494533,-0.0258814506,0.2895483673,0.2128801644,-0.2625923753,-0.3203586042,-0.6266810298,0.2629424334,-0.5137580633,0.1910710037,-0.1408808082,-0.2451640964,0.3036814034,0.1079544425,-0.1916731745,-0.0999565274,0.0556900501,-0.0071934271,0.1709096879,0.1493299305,0.3320841491,0.2791546881,-0.0922301784,-0.3216710687,-0.4433014393,-0.0400057994,0.1684950739,0.2169547528,0.0094909659,-0.079517372,-0.2575777769,-0.1280809343,0.1019529626,-0.0778396353,0.3905062377,0.1049131826,-0.0095973499,0.3817896545,-0.1426332295,-0.0086786617,-0.2456882894,-0.0357544497,0.1109166145,0.0550187491,-0.1513474286,-0.1801465154,-0.0491041094,-0.0539736524,0.2511945963,0.0999220088,0.0972087979,-0.0786000565,0.1991378069,0.1278101355,0.1732063442,0.2066397369,0.1552710086,0.0559155308,0.0019713507,-0.3858051002,-0.2215704918,0.0646760911,0.0356072113,0.0378215984,-0.1296639144,0.4016621709,0.0059828754,0.0710967481,0.1350280344,0.1085011363,0.5503321886,0.1401061714,0.1703459471,-0.4319105148,0.2125662416,0.3312862813,-0.2809247673,0.347361058,0.5354616046,0.1322885305,0.2809295356,0.0803072378,0.237775743,0.5244954824,0.3298656344,0.064048335,-0.1033367738,-0.552081883,0.1015581787,0.6678978205,-0.5076916814,-0.1573924422,0.1145320535,0.0160783455,0.1352500319,-0.1314962655,-0.1689451337,-0.1418943107,-0.1557698995,-0.0667336807,0.0940255746,-0.0192205254,-0.057518512,-0.0362388119,0.1004228815,-0.0434548967,0.2444917411,-0.3040907681,-0.2544710934,-0.0272784084,-0.0317716673,0.278442055,0.1199142486,-0.2305406034,-0.0807027668,0.157921344,-0.0214184579,0.4265593588,0.5706611872,0.5717047453,0.6063525677,0.1830434352,0.1278859526,0.0131114637,0.0590676181,-0.0073184352,0.6166630983,-0.0045634615,-0.5439010859,0.2784528732,0.0439495593,0.0041392292,0.3585559428,-0.2119914144,0.3224886358,0.1105726138,0.1054093912,0.0642698258,-0.0584102347,0.1374062002,0.0827970579,-0.2589648664,0.072637029,0.9143493176,-0.134717986,-0.0051502902,-0.3627817035,0.0297406875,0.0700255632,0.752967298,0.6106796265,0.065898709,-0.3105236888,0.0288968422,-0.5796044469,0.0823533833,-0.01419002,0.329328686,0.1985842884,-0.2157303393,0.1154160425,0.3421982229,0.2262532711,-0.3068056703,-0.0698385313,-0.4682077467,-0.2701190412,-0.0169814713,0.3454255164,0.1913647801,-0.0666048154,-0.1096358672,0.3333044648,-0.0088537429,-0.0584821068,-0.1018457115,0.2546596229,0.0168317668,0.4844090641,0.363651067,0.195884347,0.3726079166,0.1436448246,0.018105302,-0.3456308842,-0.2510227263,0.0901650116,0.2232737243,0.1879864633,0.1458601803,-0.1542794406,-0.0382021554,-0.2758579552,0.0652724281,0.0706895217,0.0428656749,0.1385474205,-0.0437365249,-0.1341778636,-0.181351319,-0.0378304385,0.120915994,0.495134294,0.1649115086,-0.0142046809,-0.2301945239,0.3458403349,-0.1777028441,0.0022803659,-0.0202440117,0.1730800271,0.1003086567,-0.2233655006,-0.8071298003,0.0816920027,0.4822101593,-0.2035746723,-0.1766884774,0.0255616289,-0.0607893355,-0.0086673545,0.0344972983,0.4376494884,-0.1060258001,-0.0757307783,-0.2217496783,-0.5205211043]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1948","title":"dataset loading logger level","comments":"Thank you for explaining the intention, @lhoestq \r\n\r\n1. Could it be then made more human-friendly? Currently the hex gibberish tells me nothing of what's really going on. e.g. the following is instructive, IMHO:\r\n\r\n```\r\nWARNING: wmt16\/ro-en\/train dataset was loaded from cache instead of being recomputed\r\nWARNING: wmt16\/ro-en\/validation dataset was loaded from cache instead of being recomputed\r\nWARNING: wmt16\/ro-en\/test dataset was loaded from cache instead of being recomputed\r\n```\r\nnote that it removes the not so useful hex info and tells the user instead which split it's referring to - but probably no harm in keeping the path if it helps the debug. But the key is that now the warning is telling me what it is it's warning me about.\r\n```\r\nWarning:Loading cache path\r\n```\r\non the other hand isn't telling what it is warning about.\r\n\r\nAnd I still suggest this is INFO level, otherwise you need to turn all 'using cache' statements to WARNING to be consistent. The user is most likely well aware the cache is used for models, etc. So this feels very similar.\r\n\r\n2. Should there be a way for a user to void warranty by having a flag - `I know I'm expecting the cached version to load if it's available - please do not warn me about it=True`\r\n\r\nTo explain the need: Warnings are a problem, they constantly take attention away because they could be the harbinger of a problem. Therefore I prefer not to have any warnings in the log, and if I get any I usually try to deal with those so that my log is clean. \r\n\r\nIt's less of an issue for somebody doing long runs. It's a huge issue for someone who does a new run every few minutes and on the lookout for any potential problems which is what I have been doing a lot of integrating DeepSpeed and other things. And since there are already problems to deal with during the integration it's nice to have a clean log to start with. \r\n\r\nI hope my need is not unreasonable and I was able to explain it adequately. \r\n\r\nThank you.","body":"on master I get this with `--dataset_name wmt16 --dataset_config ro-en`:\r\n\r\n```\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-2e01bead8cf42e26.arrow\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-ac3bebaf4f91f776.arrow\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-810c3e61259d73a9.arrow\r\n```\r\n\r\nwhy are those WARNINGs? Should be INFO, no?\r\n\r\nwarnings should only be used when a user needs to pay attention to something, this is just informative - I'd even say it should be DEBUG, but definitely not WARNING.\r\n\r\nThank you.\r\n","comment_length":351,"text":"dataset loading logger level\non master I get this with `--dataset_name wmt16 --dataset_config ro-en`:\r\n\r\n```\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-2e01bead8cf42e26.arrow\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-ac3bebaf4f91f776.arrow\r\nWARNING:datasets.arrow_dataset:Loading cached processed dataset at \/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f\/cache-810c3e61259d73a9.arrow\r\n```\r\n\r\nwhy are those WARNINGs? Should be INFO, no?\r\n\r\nwarnings should only be used when a user needs to pay attention to something, this is just informative - I'd even say it should be DEBUG, but definitely not WARNING.\r\n\r\nThank you.\r\n\nThank you for explaining the intention, @lhoestq \r\n\r\n1. Could it be then made more human-friendly? Currently the hex gibberish tells me nothing of what's really going on. e.g. the following is instructive, IMHO:\r\n\r\n```\r\nWARNING: wmt16\/ro-en\/train dataset was loaded from cache instead of being recomputed\r\nWARNING: wmt16\/ro-en\/validation dataset was loaded from cache instead of being recomputed\r\nWARNING: wmt16\/ro-en\/test dataset was loaded from cache instead of being recomputed\r\n```\r\nnote that it removes the not so useful hex info and tells the user instead which split it's referring to - but probably no harm in keeping the path if it helps the debug. But the key is that now the warning is telling me what it is it's warning me about.\r\n```\r\nWarning:Loading cache path\r\n```\r\non the other hand isn't telling what it is warning about.\r\n\r\nAnd I still suggest this is INFO level, otherwise you need to turn all 'using cache' statements to WARNING to be consistent. The user is most likely well aware the cache is used for models, etc. So this feels very similar.\r\n\r\n2. Should there be a way for a user to void warranty by having a flag - `I know I'm expecting the cached version to load if it's available - please do not warn me about it=True`\r\n\r\nTo explain the need: Warnings are a problem, they constantly take attention away because they could be the harbinger of a problem. Therefore I prefer not to have any warnings in the log, and if I get any I usually try to deal with those so that my log is clean. \r\n\r\nIt's less of an issue for somebody doing long runs. It's a huge issue for someone who does a new run every few minutes and on the lookout for any potential problems which is what I have been doing a lot of integrating DeepSpeed and other things. And since there are already problems to deal with during the integration it's nice to have a clean log to start with. \r\n\r\nI hope my need is not unreasonable and I was able to explain it adequately. \r\n\r\nThank you.","embeddings":[-0.1465872675,-0.2024423927,0.0129192071,0.3102096915,0.4003248215,0.3999482989,0.4905202687,0.2311711758,0.1243601069,0.0282325111,0.0093349731,-0.147262916,-0.2483211011,-0.188297689,-0.2508576214,0.0609391592,-0.1264946759,-0.031552773,-0.385997504,-0.1280371398,0.011019526,-0.0093774199,0.0819806829,0.4615404904,-0.69535321,-0.0577312857,0.0907774121,0.269282788,0.04841673,-0.7044392824,0.2687951028,-0.0610610768,0.1792588234,0.1538811177,-0.0001220934,0.130161792,0.626830101,0.006786231,-0.6243699193,0.0161634833,-0.3979128003,-0.3823045492,0.1909533143,-0.0893724859,-0.0878836066,-0.3329226673,0.3174687624,-0.398244381,0.3191033006,0.1539319158,0.0972733125,0.130963847,-0.1569447517,0.1304767281,0.3219229877,0.2671613693,-0.066899851,0.062543571,-0.027156949,0.1869474202,-0.3544816077,0.3654177785,-0.1897873282,-0.0355945677,0.168256402,-0.025154613,0.2707450092,-0.0829854682,0.0980230868,0.457021296,0.4693352878,-0.1381201595,0.0960084051,-0.3951651752,-0.114280127,-0.3058123589,0.1722683012,0.1817942113,-0.1236345321,0.1642877609,-0.278645575,0.0962313861,0.0633188337,-0.0412001908,0.2753942609,0.5356911421,-0.2549032271,0.199491173,0.232157737,-0.1086486205,0.318099618,-0.0647453815,-0.292390883,0.0557426736,-0.3284548223,0.0961241871,-0.156981498,0.2279085368,-0.127032131,0.0550274663,0.0303926487,0.0645745769,0.1766626239,0.0465370044,0.3021320403,0.0782378837,0.3390364349,-0.0437264033,-0.2039043158,-0.0232509989,0.190676406,-0.0522891022,-0.1306730509,-0.0884445235,0.4426140189,0.2055938244,0.0061232545,-0.0329376422,-0.1966416687,0.0259849578,-0.1524892002,-0.3309821784,0.024868533,0.1910765469,-0.0790168047,-0.0314244404,-0.1334362924,-0.0265853889,-0.0438799746,-0.2736995518,-0.0795496926,-0.2813799977,-0.3722065687,0.1771147996,-0.1128016785,-0.1633096784,0.1749125272,0.1784884781,-0.0563448071,0.0769142658,-0.0438766368,-0.1474408209,0.1251858175,0.4888392091,-0.1188975126,0.2109150141,0.315344125,0.114410758,-0.2508600056,0.3219142556,-0.3088692129,-0.4429717362,-0.1388504952,0.056897305,-0.1872850657,-0.042705588,-0.2263440192,0.2049746364,0.2080996037,0.2616475523,0.3043168485,-0.0870516226,-0.2893235683,-0.1400413662,-0.1068019271,0.6045224071,-0.2300405055,-0.3448925614,-0.2233905047,-0.390848875,0.3220081627,0.1940955967,-0.1252973825,0.3429298103,-0.2669639289,-0.1275547594,0.2065176219,-0.199307546,0.1538253278,0.5722748041,-0.2666302323,0.3573934436,0.2487148046,0.0757367015,-0.3795433939,-0.1587415636,0.0931627974,-0.4061235487,0.0346498452,-0.219547376,-0.1422888488,-0.3101754487,0.3162405193,0.0956732556,-0.1243333444,-0.0407623909,-0.0270057153,-0.059658967,0.5715073943,0.0659455955,0.0881292075,0.0159739293,-0.0474747047,0.0137909576,-0.1247672737,0.3047455549,-0.4733557701,0.0033311769,0.2965785563,0.1541423351,0.0607184656,-0.0349128097,-0.1101765782,-0.041664362,-0.4651914239,-0.1592132896,-0.014823894,0.0448270179,-0.2315159291,-0.133993879,-0.1228299066,0.3283220232,-0.5877948999,0.0591800176,-0.1173660383,-0.3510161936,0.0371697359,0.3243008852,-0.2366799265,0.0116760097,0.1059909984,0.0001265008,-0.1354235411,0.2674871981,-0.0427938104,0.311644882,-0.0823953971,0.5417579412,0.038802281,-0.1560539007,0.3681598306,-0.2089685947,-0.2602341473,-0.0510693863,0.113723956,0.0264503155,0.0294059273,0.2619304955,-0.2610865831,-0.1066248864,0.0779918954,0.0519058481,-0.1736053228,-0.3629187942,0.1012513787,0.1045231819,0.2224511802,0.3565239608,-0.0605902523,0.1401377916,0.373893708,-0.0226562619,-0.0009537109,0.2139959484,-0.0660070106,-0.1223928779,0.1966656595,0.0896252021,0.1997384131,0.1502249688,0.178512007,0.2778914869,-0.0631349906,-0.057255663,0.2577065229,0.0119202621,0.4047441483,0.0357543752,-0.456192404,0.0109884143,-0.2117695659,-0.0841504335,-0.1447158456,-0.0529185422,-0.6462906003,0.003194595,-0.268383503,-0.3433909416,-0.3749732077,-0.252673924,-0.1563916355,-0.3785588741,0.0086878147,-0.3024218976,-0.3954238892,0.1360277683,-0.3061369061,0.230926007,0.1001675949,0.2513026893,-0.2852820158,-0.0661872253,-0.2680495977,-0.0152586149,-0.0835170969,-0.2519970238,0.0985292122,-0.490531534,0.1629213393,-0.1752011329,-0.2026423514,0.3088508248,-0.1975004226,0.0879951939,0.1897615939,-0.0893781185,-0.0432891361,-0.2671490312,0.0250475947,0.1185781434,0.008313857,-0.354283154,0.0778483227,0.1457306594,-0.1193108782,-0.091356039,-0.3239561319,-0.1959176064,-0.1272505373,-0.367288202,0.2959487438,0.4755595326,0.0311138406,0.0013714295,-0.1817460358,0.2356292009,-0.2804739475,-0.7693060637,-0.0325261652,-0.1196923032,0.1096624807,-0.0435089134,0.1002776027,0.1628791541,0.0031549782,-0.8037417531,0.1855158359,-0.1419694424,-0.3532150686,0.160369873,-0.0036665837,-0.1876031607,0.2181845605,-0.0146573279,0.0057902932,-0.2862210572,0.0531042032,-0.4676643908,0.4226564169,0.1411723197,0.4266456068,0.1103252023,0.4797783792,0.1754454821,-0.0096641118,0.2206352055,-0.0405065082,0.6518876553,-0.0414569713,-0.0822485909,0.0396552198,0.0212174859,-0.2239888012,-0.0989487544,-0.0745002553,0.2100459635,0.2413645089,0.0937358513,0.035237547,-0.4016301632,0.271589309,-0.4244952202,-0.1394442916,-0.1997300982,0.1225217879,0.2278078198,-0.0328869633,0.1564408988,0.4967171252,0.2978893816,0.1715159863,-0.2740747035,-0.1311542988,-0.3372915983,-0.0180857107,0.0550625697,0.1616726369,-0.3739035726,-0.3055794835,-0.110557802,0.3051782846,0.1360486448,-0.139113605,0.2344477624,0.1739260703,-0.069646351,-0.0882605165,-0.1640915722,0.0512594953,0.0295519307,0.1467779875,0.0859366208,-0.152584359,-0.4390387237,0.1397966743,-0.0961279944,-0.2378573269,-0.2504742444,-0.2054912597,0.1677544564,-0.0954988524,-0.0245507807,-0.047024101,-0.0282836985,-0.5195003152,-0.0082285758,0.1157225296,0.3064211309,0.075593181,-0.1433747709,-0.090258345,0.3614208996,-0.1069243476,0.4327397048,0.7041027546,0.2550817728,0.7299550772,0.0712156966,-0.0811747313,-0.0565037541,-0.0520723946,0.2179764658,0.3498026431,0.1193915382,-0.0570035651,0.3241608739,0.4321883321,-0.4970788956,0.3054851592,0.2432101965,0.1004246026,-0.477558583,-0.3319886029,0.4986374974,0.1059214473,-0.0401842557,0.4706761241,0.2508615851,-0.1733491868,-0.1791170985,0.4485978782,1.0204641819,-0.1413291097,0.2379662246,0.0887254328,-0.0671918094,0.2032370716,0.0498396903,-0.2287016362,-0.0339879878,0.0954981744,-0.1102621704,-0.0079029379,0.0217751805,0.4259312153,-0.0814972669,0.0552998334,-0.3799417317,0.1997650862,0.2381646335,0.0111748734,-0.1948652864,-0.1852941811,-0.1854830384,-0.0109014446,0.1929553747,0.1259739995,-0.254964292,-0.0789041594,0.3553372622,-0.1292890012,-0.2059787661,-0.1289222836,-0.2180993855,-0.0243697613,-0.1056504175,-0.0971036181,0.0242868904,-0.0689306781,0.0443328395,0.1664553881,0.0667945519,0.0058084289,0.003596168,0.0768849701,-0.3164058626,-0.0807072222,0.3426615,0.0499628149,-0.3006897271,0.2033223212,-0.2766222954,-0.5778278112,0.0519154705,-0.1594394594,-0.2073705345,-0.0885656476,0.1399164647,0.1879905462,-0.1414648592,0.221740678,0.016712511,0.1536526382,0.2042517811,0.0143995704,0.2500064671,-0.5792984962,-0.1628116518,0.0786973909,0.0129217757,-0.1310413927,0.4648008049,-0.0984340385,-0.0492203832,-0.1144694462,-0.0070027839,0.5963891149,-0.1043778285,-0.1934039444,0.2725802958,-0.048712749,0.076960437,-0.1061723977,0.3125024736,0.0999436006,-0.2827934027,-0.2267077565,-0.5409591794,0.2808478475,-0.4361152053,0.2125404477,-0.1690126508,-0.147634387,0.2259999812,0.0833062381,-0.1783976108,-0.2111179829,-0.0559465773,-0.1625299454,0.1657117158,0.2391402572,0.3696463108,0.2524444163,-0.0903135985,-0.3158454597,-0.320556879,-0.0216379575,-0.0654079393,0.2244133502,-0.0104349991,-0.117598258,-0.0966218039,-0.0900585577,0.138456434,-0.1828060001,0.3072935045,0.2030452639,0.0739672109,0.4574136436,-0.2511950731,0.1956444085,-0.3103450537,-0.0566111989,0.3130631745,0.2139307559,-0.1051529199,-0.1277214289,-0.0128357196,-0.0179275125,0.3755550981,0.1243964732,0.245598346,-0.0501767993,0.0954121426,0.2405617535,0.1526499838,0.1736780554,0.2529512048,0.1828891188,-0.0151574649,-0.4614331424,-0.2064892501,0.0549770594,0.0874793604,0.0284560695,-0.3894968033,0.3267393112,-0.0016325311,0.1633819342,0.1570379883,0.1106405482,0.4825283289,-0.0165399015,0.0794721469,-0.4653482437,0.1203662828,0.3244590759,-0.3549873233,0.2751422524,0.4656180143,-0.0184549224,0.325889349,0.2565750182,0.3138428926,0.5164299011,0.3635672629,0.1383249164,-0.1878064722,-0.4322495759,0.0178330876,0.5528194904,-0.3737696409,-0.0973557457,0.1278448254,0.0378067382,0.2741698921,-0.1087609679,-0.1625677794,-0.2601650655,-0.0821955502,-0.1250255406,0.0548071712,-0.0606838875,-0.2160615325,-0.0553320423,0.0889469087,-0.1751568466,0.1543828547,-0.2759079933,-0.1503820419,-0.082684204,-0.0239452887,0.3113580942,0.1572370976,-0.2301957309,-0.1721215099,0.3310587406,-0.1155634075,0.4097523987,0.4752766788,0.6177262068,0.5910733342,0.1977445036,0.1410229802,-0.0959316343,0.0992012769,-0.0008837936,0.466473788,0.0532714725,-0.5392205715,0.3291276991,0.0190834906,-0.0180721469,0.4304611385,-0.1307463795,0.3070867062,0.0956978723,0.1636202335,0.0522701666,0.0454727933,0.0786592811,0.1035527289,-0.2113203406,0.0828136876,0.9799102545,-0.2187812328,0.0082387188,-0.3598146439,0.0376449116,0.1337392032,0.7504997849,0.5301433206,0.0191347562,-0.2689056695,0.1098716557,-0.6785558462,0.1615492851,-0.0100104166,0.1195642725,0.2431473732,-0.1740598381,0.1362668276,0.2347297668,0.2257976532,-0.3438852131,0.0795258433,-0.3624971211,-0.2689088583,0.0313590616,0.4597531855,0.2654235363,-0.0453866571,-0.1311106682,0.2973899245,0.0306785908,-0.0711343959,-0.1218314022,0.167689994,0.1573789716,0.382961452,0.3445104063,0.1535736322,0.36864236,0.032440085,-0.123678118,-0.4942083061,-0.1948024631,-0.0650608391,0.2983183861,0.1826138943,-0.0177897662,-0.0968003944,0.0101349894,-0.3805670142,0.1060073599,0.0610493161,0.0909545496,-0.0067870063,0.1168815941,-0.1358053237,-0.0998946801,0.0673662424,0.1554569453,0.5111301541,0.1131734475,0.0115881739,-0.2730745077,0.2586715519,-0.070067443,-0.0500632115,-0.1905662566,0.1469298452,0.2262373418,-0.0769284591,-0.8204528689,0.1293432862,0.3578497469,-0.1564281434,-0.1171125993,0.1346116513,-0.0860334635,0.0065338886,0.0855472833,0.4466448426,-0.0737442449,-0.1286573559,-0.1838977039,-0.5174059272]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1945","title":"AttributeError: 'DatasetDict' object has no attribute 'concatenate_datasets'","comments":"sorry my mistake, datasets were overwritten closing now, thanks a lot","body":"Hi\r\nI am trying to concatenate a list of huggingface datastes as:\r\n\r\n` train_dataset = datasets.concatenate_datasets(train_datasets)\r\n`\r\nHere is the `train_datasets` when I print:\r\n\r\n```\r\n[Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 120361\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 2670\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 6944\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 38140\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 173711\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 1655\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 4274\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 2019\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 2109\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 11963\r\n})]\r\n```\r\n\r\nI am getting the following error:\r\n\r\n`AttributeError: 'DatasetDict' object has no attribute 'concatenate_datasets'\r\n`\r\n\r\nI was wondering if you could help me with this issue, thanks a lot ","comment_length":11,"text":"AttributeError: 'DatasetDict' object has no attribute 'concatenate_datasets'\nHi\r\nI am trying to concatenate a list of huggingface datastes as:\r\n\r\n` train_dataset = datasets.concatenate_datasets(train_datasets)\r\n`\r\nHere is the `train_datasets` when I print:\r\n\r\n```\r\n[Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 120361\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 2670\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 6944\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 38140\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 173711\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 1655\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 4274\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 2019\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 2109\r\n}), Dataset({\r\n features: ['attention_mask', 'idx', 'input_ids', 'label', 'question1', 'question2', 'token_type_ids'],\r\n num_rows: 11963\r\n})]\r\n```\r\n\r\nI am getting the following error:\r\n\r\n`AttributeError: 'DatasetDict' object has no attribute 'concatenate_datasets'\r\n`\r\n\r\nI was wondering if you could help me with this issue, thanks a lot \nsorry my mistake, datasets were overwritten closing now, thanks a lot","embeddings":[-0.1553635448,-0.3995109797,-0.1242145598,0.3105229139,0.3694525659,0.1859706491,0.1929529607,0.2077361345,0.207564652,0.052870404,-0.1874539852,0.1402713656,-0.142166391,0.4111873507,-0.053310506,-0.2501934171,0.0768456757,0.0768707618,-0.0963189527,-0.0250812136,-0.2554418743,0.3081436753,-0.2117798775,0.0072639724,-0.2831128538,0.0741345733,-0.1294038445,0.0415823087,-0.175096944,-0.1812459081,0.3635767102,0.0713296086,0.1658633053,0.3401078284,-0.0001091484,-0.0245555826,0.1182109043,0.0008767808,-0.3499364555,-0.2074406743,-0.1324796379,-0.2582700551,0.1783022285,-0.1530847847,-0.1991170347,-0.2704346478,-0.184564814,-0.3986696005,0.179839924,0.3665132821,0.242903173,0.1494325697,0.3928612471,-0.3153750002,-0.1668022871,0.0791142955,-0.0581791364,-0.121736452,-0.0883874521,-0.117890276,0.5402024984,0.188324064,-0.0782720074,-0.1424954236,0.1520389467,0.2249963582,-0.0756306499,-0.2221471667,-0.0102833202,0.1587879807,0.4116046727,-0.2373737842,-0.3039067984,0.0714036971,0.1480208039,-0.3593174517,-0.0673640296,0.0210991912,0.0812061355,0.1326166987,-0.0632015243,-0.0505946465,-0.107829906,0.2659151256,0.0878571644,0.1808361411,-0.1251336038,-0.0206059124,0.4426805675,-0.2885499299,-0.0702547133,-0.3396536112,-0.1339826733,-0.0149492351,-0.5133880377,-0.0125477938,-0.0056124888,-0.1433745772,0.2827238739,0.1443427056,-0.0765891895,-0.1423650682,0.1767922491,0.1607795805,0.0900241733,0.2656177282,0.1663806289,0.2765578926,-0.2901925147,0.1888599545,0.1523294896,-0.0769804195,-0.1295096725,-0.2320137471,-0.0534621626,-0.0789850652,0.3778482378,0.0457603186,-0.2180935293,0.0949083343,-0.1813769937,0.0341604911,0.135455668,0.1524733454,0.0245163906,0.2178997695,0.3438382149,0.1192780435,0.2107279301,0.0399646945,-0.3012008369,0.1349444836,0.1051839292,-0.1091863215,-0.0977679789,-0.2555749416,0.0882131383,-0.066877529,0.3607113957,-0.0884673819,-0.247484237,-0.1317535937,-0.0085099675,0.1012947857,-0.013582957,-0.0752314776,0.3506235182,-0.4361875057,-0.0806678608,-0.0899094418,-0.228810519,-0.3179389536,-0.4285404086,0.1898999065,0.1536555886,-0.0625742152,-0.3370428085,0.2471288443,0.2433793247,0.3274158239,-0.0993457064,-0.0210212115,-0.0687862188,-0.1318064928,0.385245055,0.2192941904,-0.1389419287,-0.0089721214,0.2220339477,0.0433916785,0.217297405,0.3328602016,-0.0363839455,0.1510230005,-0.2066963613,0.1375653744,0.2532048821,-0.2636224627,-0.0808965564,0.0082756709,-0.0236042589,0.0630581602,0.1982234716,-0.0189242419,-0.0136356354,0.1244800687,0.523394227,0.2455238253,0.0863857791,-0.1915378422,0.0075868629,-0.1053271741,0.4928790033,0.0146981049,-0.4134127796,0.3073050678,-0.2627158463,-0.1642296314,0.2333242744,-0.2625823915,-0.0151456669,0.2997291386,0.1987410933,0.146633774,-0.0954837278,-0.0311120544,-0.4100517929,-0.0599722117,0.3106036782,0.1345285624,-0.2050267905,-0.2143428922,-0.1739224494,0.0790161565,-0.1929142475,-0.1563899219,0.203286469,0.0325134471,0.066198431,-0.0886109695,-0.2620039582,0.2638612986,-0.0221967958,0.0062814886,-0.5264207721,0.3115148544,-0.2673886716,0.0320747234,0.0575026385,0.2273767143,0.1453247815,-0.0078952685,0.0362668335,0.3336415887,-0.2272820026,-0.3630015552,0.2058277875,0.0180656519,-0.0709077269,-0.066741623,-0.306699872,-0.135014832,0.0028041427,-0.0644584894,0.0594662763,0.2359501421,0.1903809905,0.3338133395,-0.1356976032,0.3407630324,0.2145909667,0.0128109781,-0.0933014378,-0.0983740017,0.3283233345,-0.2312618643,0.0752196237,0.0177170616,-0.5187623501,0.0891328827,0.3289399147,0.1075401306,-0.0636239126,0.0161542799,-0.0888593048,0.234727338,-0.1328078657,-0.1498552859,0.3219728768,0.3021977246,-0.2806299627,0.0603181124,0.0039801938,0.0561994016,0.162978828,0.1012855545,0.1885715872,0.1025961787,0.059269879,-0.0850191265,-0.2646084428,-0.4832986593,-0.1892193407,0.1590166539,-0.222497195,0.0781256035,-0.1928979903,-0.0226629544,0.0364633314,-0.5499449372,-0.2939560711,-0.2333604246,-0.2800451815,0.1243622303,0.0515944697,0.1902521998,0.0055239019,0.1351149529,0.318002373,-0.167574361,-0.24426274,-0.2018204331,0.0830773935,0.0800069943,-0.0461948626,-0.1443715543,0.0286513213,-0.2217649519,0.1630960703,-0.1631821096,-0.0481904708,-0.0521990173,-0.4023903012,-0.048076462,0.2004349381,0.3328292668,-0.1743370891,-0.4240826964,0.4525981843,0.4647889137,-0.2419528216,0.2221052945,0.2307612598,-0.387991488,-0.0300217979,-0.3273943067,-0.285516113,-0.334443599,0.2169170231,0.0400660932,0.1278458089,0.3682623804,0.3376397192,0.0231013522,-0.0353032723,0.2109106332,-0.1813537925,-0.3246292472,0.3397473693,-0.1891771555,-0.4015987813,-0.0776924118,-0.0167039465,0.1271897852,-0.110907279,-0.2983245254,-0.2105138004,-0.0630159006,0.060216248,-0.1997409016,0.14690274,0.3387866616,0.2343356013,-0.1529866606,0.0529616289,-0.2567719221,0.2314902395,0.0057827295,0.3582767248,-0.0563242659,0.2210365683,-0.1576875299,0.1767101884,0.4676401615,-0.2753359377,0.1967468262,-0.1436666548,0.2698884904,-0.1478315592,-0.7538875937,-0.1605125517,-0.0706772953,0.0761237368,-0.0718755871,-0.1166711971,0.5005452037,-0.2064363956,0.3281053305,0.0096712299,-0.104681246,-0.1598448455,-0.1041951627,0.2121590972,-0.105103761,-0.0993189663,-0.1714398265,0.1372460872,0.106034629,0.4097722769,0.0052118707,-0.0366506204,-0.5925052762,-0.3142340779,-0.5939047337,0.2892137468,0.1304304451,0.3410022855,0.0011680761,0.0878459662,-0.2699686587,0.0885331109,0.8688221574,0.1718490273,-0.5497944355,0.2882564366,-0.0580344386,-0.1228223443,-0.0319609605,-0.1035866588,0.1320337355,-0.0887176618,0.4401442111,-0.4043752253,-0.1828104854,0.1653003693,0.1574501395,-0.2034188658,-0.0092924731,-0.2596355081,-0.3819734156,-0.2956288159,-0.2908945382,-0.017464241,0.5890978575,-0.0275027324,0.0422428101,-0.1878645122,-0.2542864382,0.29640764,-0.1896687895,0.3984369338,0.3125735521,0.0654364675,0.000567082,0.4081735015,0.360522747,0.4486038685,0.0093665309,-0.4961603284,-0.2903381884,-0.0734360293,0.5020400286,-0.204292044,-0.2565837204,0.1835908592,-0.0093592871,0.3295362592,-0.0565719157,-0.0017727143,0.5006633401,0.3596221805,-0.4966524839,-0.4199634194,0.2004321814,0.2232732028,0.2295217961,0.1138129681,0.1530111134,-0.134685263,-0.0098130992,-0.1960530132,0.7867659926,0.089282535,0.2373338193,0.1965730488,0.2462375164,0.4825965464,0.1866048276,0.2257986665,-0.1298598349,-0.3495658338,-0.1591959894,-0.165658921,0.1963473558,-0.1071754918,-0.3925723732,0.12003766,-0.2373864204,0.1112573147,-0.1901235878,-0.0093469825,0.3010618985,-0.2964861393,-0.4294170737,0.1363446712,-0.1184609085,0.1290178299,-0.1316586882,-0.145661518,-0.1879282445,-0.1534754783,-0.1705775857,0.1421431452,-0.1703364253,0.3630063832,0.3771669865,-0.4069009721,0.0442419387,0.2135673016,0.2267694026,0.0968414992,-0.0715590939,-0.2619841099,-0.123308599,-0.1463087499,-0.2092655003,-0.2096257955,0.2813212574,0.0582912751,-0.2563529909,0.2149608433,0.2208893299,-0.2024576515,-0.1327985972,0.0166961104,-0.0112161729,-0.5740742087,-0.206064716,0.292096138,-0.0343218893,-0.1786274612,0.132495001,0.2042160481,-0.4272687435,-0.0424611159,0.2339830995,-0.1603784263,-0.0142945778,0.4138047993,0.1688677669,0.0697013736,0.5084963441,0.1992431879,-0.019848682,-0.1690449864,-0.133111462,0.2591382563,-0.0500452593,0.180186674,0.0638165027,-0.0072349804,-0.1813568771,0.4084774256,0.0394733399,0.1148696989,0.1163274273,-0.2929786146,-0.4712376595,0.2106634378,-0.0466424339,0.1756320149,0.0594972782,0.4334875047,-0.1212393045,0.0462078378,-0.3490440249,-0.3054604232,-0.3621222377,0.1543505937,0.2158899307,-0.2102532685,0.2698872089,-0.0167464018,0.2230248004,0.0010576413,-0.1082389578,-0.3297482133,-0.052272208,0.0436589308,0.2298210114,-0.1080715954,-0.1586970389,-0.0456734188,-0.2111779153,-0.2189432532,-0.2444191277,0.0918581784,-0.0624302812,0.1491883844,-0.014986123,0.0401080847,0.1214091554,-0.1108681038,0.0919976532,0.0342511423,-0.0294003636,-0.0356422327,-0.2587109208,-0.1238615066,-0.3060120344,0.1467031986,0.1193286777,-0.0083570797,0.2514860928,-0.38034603,-0.164575994,0.0358979516,0.5426744223,0.3503537178,-0.1354963034,0.3164409399,0.1010203809,0.2150979042,-0.385143429,-0.1468336135,0.3506208658,-0.0273296498,0.0454375222,0.1195212826,-0.2036484033,0.0200891066,-0.1230171397,0.0064362627,0.0939654112,0.007387294,0.1834734529,-0.0001134111,0.1506797373,0.0699913204,0.2821568847,0.5563673973,0.1764346212,0.3241117895,-0.0405711904,0.2870617211,0.2195057869,-0.066354394,0.0787392259,-0.0410854183,0.1003847793,-0.0541945994,0.138329044,-0.0911908299,-0.1494788677,0.2972196639,-0.1750122011,-0.1908718646,0.0208670199,0.3293990791,-0.2064218521,-0.0830659121,0.1967383921,0.1228054464,0.0778612122,-0.3429261446,0.3401705027,-0.3024844229,0.4166293442,-0.1387972832,0.1691201627,-0.4439661801,0.0022571944,0.5147246122,0.171378538,-0.1830505431,0.0110842343,0.41735515,0.005650857,0.1081515551,0.4019402266,0.6873853803,0.1790600568,0.0756566525,0.0999369249,-0.2188079357,-0.1867724359,0.0161064006,-0.0421460867,-0.0199195016,-0.2726674378,0.2549723387,0.2175862789,-0.073750779,0.2158401608,-0.0633467957,0.1589209586,-0.350828737,0.3025413752,-0.2164506316,-0.289021343,-0.381339103,0.122420907,-0.4448777139,-0.1334553659,0.4066427052,0.1785845459,0.2382153422,-0.0646538958,0.1651756763,0.0887548253,0.4596246183,0.36904791,0.3348042965,-0.2569288015,-0.0771055222,-0.0783622712,-0.0332979783,0.08732301,-0.2059233636,-0.0607178994,0.2371718138,0.043378748,0.2012426853,0.0333545655,-0.019454876,-0.0518961996,-0.0965346247,-0.4304777682,-0.2350200266,0.020700274,0.1444832087,0.276890099,0.0052560274,0.2164730877,0.0195458196,0.1211233661,-0.3047235906,0.3678897023,-0.0900340006,0.1900854856,-0.1598725319,0.2939372063,0.3117020428,-0.1207104772,-0.0348363295,0.055853162,-0.322586,-0.1462715417,0.4029141068,0.0916959792,0.1139220893,-0.2778934538,-0.2524081469,-0.1537589133,0.2722933292,0.3012140989,-0.3593222797,-0.3023630381,0.2865928113,0.0207218677,-0.0854139626,-0.0221012421,0.3620015681,-0.0449029952,0.3693901002,0.0208166,-0.4574992061,0.6101882458,-0.0914575979,-0.358097285,0.0126109999,-0.0638374388,0.313822031,-0.1132792756,-0.3750820458,-0.0363007523,0.2389461845,-0.121288538,-0.2083401084,0.4147646427,-0.1596445739,0.2183290869,0.0187695529,0.2537869513,0.0524349287,-0.1328873336,0.1422332525,-0.1400771141]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"Hi !\r\n\r\nThe cache at `~\/.cache\/huggingface\/metrics` stores the users data for metrics computations (hence the arrow files).\r\n\r\nHowever python modules (i.e. dataset scripts, metric scripts) are stored in `~\/.cache\/huggingface\/modules\/datasets_modules`.\r\n\r\nIn particular the metrics are cached in `~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/`\r\n\r\nFeel free to take a look at your cache and let me know if you find any issue that would help explaining why you had an issue with `rouge` with no connection. I'm doing some tests on my side to try to reproduce the issue you have\r\n","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":84,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nHi !\r\n\r\nThe cache at `~\/.cache\/huggingface\/metrics` stores the users data for metrics computations (hence the arrow files).\r\n\r\nHowever python modules (i.e. dataset scripts, metric scripts) are stored in `~\/.cache\/huggingface\/modules\/datasets_modules`.\r\n\r\nIn particular the metrics are cached in `~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/`\r\n\r\nFeel free to take a look at your cache and let me know if you find any issue that would help explaining why you had an issue with `rouge` with no connection. I'm doing some tests on my side to try to reproduce the issue you have\r\n","embeddings":[-0.0358740725,-0.0009466419,0.0741713271,0.1184607968,0.0837402493,-0.1097532511,0.1720347852,0.2317183316,0.263871789,0.1391707808,0.0678788871,0.1733187735,-0.2854747474,-0.0121780438,0.1369281113,0.0551351197,-0.0628577545,-0.0301498175,-0.325932622,-0.1252780259,-0.2200645804,0.5307022333,-0.0628679842,0.0645004809,-0.4562329352,0.1752285659,-0.1284715831,0.3620730042,-0.0833548233,-0.584066689,0.3293269575,0.1146462336,0.0539533086,0.4142387807,-0.000119533,-0.1037964821,0.2400082946,-0.1888287961,-0.2738633454,-0.2778948545,-0.075361982,-0.2522022128,0.5798787475,-0.0966041684,-0.1344479173,0.0916116387,0.0100134797,-0.4268659651,0.4767811894,0.121173963,0.1166027337,0.4053646624,0.0331894234,-0.3333912194,0.2213087976,-0.104605563,0.0415684804,0.4594830275,0.0687569603,0.0936744362,-0.1464713663,0.2114187777,0.3470683396,-0.184419781,0.309864521,0.0112119317,0.00131537,-0.1081184596,-0.0284116585,0.3066345155,0.2365594357,-0.1305871606,-0.2608057261,0.1283788681,-0.0879927501,-0.4039272666,0.1538643092,-0.0939178318,0.0763653293,0.0891124755,0.0873675048,-0.1980054379,-0.0160592515,0.1474253088,-0.1284569502,-0.1398061663,-0.1876245439,0.0414875671,-0.0495051667,0.0219498575,-0.5056313872,0.4271095395,-0.183152318,0.1320415884,-0.4020826221,0.1370484233,0.2792122364,0.1743569076,0.0485688336,0.5940818787,0.0544064604,0.0976389945,0.3240977824,0.1152189597,-0.1439858377,0.4637179077,0.1471063793,-0.0318703651,0.229068175,0.2400322706,-0.0997861177,-0.2906394005,0.2207398117,-0.2854764462,0.1802102029,0.2259304225,0.2352595031,-0.3656279445,0.0109472349,0.2513976395,-0.1731025726,-0.0880765393,0.0778518245,0.3774317801,0.0209159143,-0.0662802085,0.3336496055,0.3538492024,-0.2321899384,-0.0483000167,-0.4542716742,-0.1579865664,-0.1675678939,0.3084725738,0.1213161945,-0.0429608561,0.5225880146,-0.137724787,0.3675760031,-0.0634157658,0.2979551554,0.1495726556,-0.0578732304,0.2460677922,0.0045490917,0.1730860323,0.3100202382,-0.0152883707,-0.1046919227,-0.313659668,-0.214396596,-0.4901871383,0.086323522,0.0288708042,-0.3260264695,0.1841583699,0.3849876821,0.0545696914,-0.2149495631,-0.2581464648,0.1204283237,0.0444046967,-0.0732085928,-0.1320746392,0.5114512444,0.7114232183,-0.2573743463,-0.4081455171,0.2153412402,0.0144806914,-0.1083219424,0.1975735277,-0.0630953982,0.1563919634,-0.3403097093,-0.1043980196,0.4113843739,-0.5171518922,-0.4667468071,-0.0168590825,-0.3854107857,-0.1010334939,0.1008030176,-0.0357385799,0.0745937452,0.2175187916,0.1196304932,-0.1724984199,0.2414471209,-0.3219715357,-0.2602002919,-0.2055730373,0.1193033457,-0.0603425615,0.2362509519,0.1523997933,-0.0052174143,-0.073207669,0.1255847514,0.1020373926,0.1981779188,0.378146112,0.2921807468,0.1836390048,0.1978509873,0.1566252261,-0.1678197384,0.2583167851,-0.4689180255,0.1527780443,-0.0503249019,-0.0147543289,-0.2192664593,-0.2279687673,-0.0850848034,-0.4767327011,0.0177706815,0.0121323569,0.2931835949,0.3102684021,-0.1623553485,0.2714135349,0.0924602225,0.2920165062,-0.6656681895,0.0322583541,-0.1113724336,-0.33567366,-0.0986841321,0.1647350043,0.0611554831,-0.008172608,-0.1790619493,0.3883354664,-0.0169802532,0.1749452502,0.3794831336,0.0891622975,-0.0574563257,-0.4079510868,-0.0328812003,0.1797966808,0.0542467199,-0.0016229018,-0.2007852644,0.323792547,-0.0825717226,0.1331495494,-0.2450239509,0.3666039705,-0.0399597995,-0.0688230768,-0.5726243258,-0.0933298841,0.3485019505,-0.1906701922,0.4904285371,-0.1709937453,-0.0320257135,0.0218921844,0.0991401896,0.1401264817,0.2501839697,-0.0037819657,-0.2171487063,-0.0424428843,0.0269454364,-0.1459605545,0.4352152348,0.136262551,0.1439883709,0.1243445575,0.2096870393,-0.1694747359,0.0424716659,-0.0767036676,-0.1214765087,0.0951082706,0.1849317551,0.168072179,-0.2335238755,-0.1956410855,-0.3971342742,0.0322263464,-0.3484332263,0.0946538001,-0.2037186772,0.3464001119,0.3030623794,-0.0559965037,-0.3411486447,-0.2638901174,0.1385770738,-0.1877672374,0.0301673356,0.2229873985,-0.2555914521,0.4870513678,0.1352511793,-0.1863757521,-0.3071427047,-0.2004612088,-0.1395705789,-0.150298059,-0.2870164514,0.0871473402,0.0249529704,-0.2135687172,0.1292314827,-0.0343739353,-0.3483106792,0.0483233929,0.051832296,0.5172300935,0.3936597109,0.0021313401,-0.2833370268,0.2200516164,0.5371545553,-0.3732428253,-0.309099704,0.0990039334,0.0583365485,0.2021277249,-0.2552041411,-0.3111086488,-0.0462787375,-0.3856249452,0.6103236079,0.1012190729,-0.1091394797,-0.0645339042,0.012322762,0.1742912531,-0.2101198584,0.3627804518,-0.3324130177,-0.6493901014,0.3374326825,-0.1905549616,-0.3684265316,0.0284811556,0.0337616466,0.3815257549,-0.0200926159,-0.5352692008,-0.3893999755,0.0445665605,0.2492183745,-0.1060669869,0.0623193979,0.580855906,-0.0551422276,-0.0207334682,-0.0587935299,-0.1301974803,0.434307009,-0.1079049259,0.1379557103,-0.1012509242,0.4365177751,0.0528162457,0.7605785131,0.482316494,0.2873231769,0.2661584616,0.035839349,0.4865826666,-0.2143168747,-0.1602762789,0.1947723031,0.1949837506,-0.0684700832,0.1258611083,0.1112143323,0.2847439945,-0.2191254646,-0.0268837307,-0.3354534507,-0.330055356,-0.0206896923,-0.2873105705,0.2177421749,-0.0690016672,0.0413902663,-0.2042773217,0.0539595596,0.3671606481,0.3646484911,0.2758910656,0.2595493495,-0.2775645554,0.0905445442,-0.5864399672,0.3170912564,-0.5184065104,0.1670907885,-0.2591429651,-0.2926996052,0.0993681401,-0.016317483,0.5642932057,0.0010579751,-0.1702984124,0.0686848983,-0.2607215643,-0.5911595821,-0.0195031427,-0.0470783673,-0.0120294839,0.2525356114,0.2193756402,-0.3621320426,-0.1756046861,-0.1045859903,-0.1455084682,-0.216161266,-0.1501789987,-0.3854457438,0.0320448317,-0.1962873638,0.103323482,-0.2960127592,0.0171897896,0.0963180214,0.0973912552,-0.2283386141,-0.0517571457,-0.0071311258,0.0747468099,0.3057206571,-0.3038360775,0.1706122607,0.4674375653,0.0272756927,-0.0002296802,0.5265596509,0.0551016405,-0.2797267437,0.2987195253,-0.0631925315,0.2123971283,0.5377953053,-0.2559638619,0.1114687845,-0.0735079646,0.3654127121,-0.1908652931,-0.1126985103,0.3272386491,-0.1773870438,0.3741321564,0.1637040675,0.0672192499,0.1586410403,-0.1380216032,0.1574546248,0.0404474027,-0.1430930793,0.3999297023,-0.1042415947,0.9887375832,0.2283452004,0.2489135563,0.131403774,0.0513559543,0.4545087814,0.0168125276,0.022108702,-0.240444392,-0.2034012824,-0.0113504147,-0.1185370088,0.0253494717,-0.2338592112,-0.2589440942,0.4140233099,-0.3903059661,-0.0517876595,-0.2063387036,-0.1757092178,-0.2596443594,-0.1164498106,0.0121736946,0.0868628994,0.0721112788,0.6109398603,-0.2037316114,-0.1589169353,-0.4563352168,-0.3336650133,-0.0674571469,0.0130486479,-0.3301522136,0.088163361,0.2974457145,-0.1824513376,-0.1842072308,0.3558039069,0.3632246554,0.1726472527,-0.2810915709,0.0607846603,0.0395961814,0.0463073999,0.1197913513,0.0193876028,0.2255722433,0.0099451868,-0.1676051319,0.1641512513,-0.2100360841,0.0512366854,0.0139879445,-0.0534159243,-0.1955824494,-0.1137569025,-0.0202226248,0.1057808995,0.0967819095,-0.1549049318,0.0860570222,0.3396830261,-0.1765406281,-0.0151249385,0.1682721674,-0.2041364908,-0.054596331,0.579885602,-0.3019783199,-0.1088530794,0.3987403214,0.0334341787,-0.0022550276,-0.1496373266,0.1192860007,0.4620956182,-0.7651275992,0.0641346499,0.0818871781,0.0467187762,0.0998905376,0.1696162522,0.3279047906,-0.1738369018,0.0599951856,-0.2481648326,-0.4355520308,0.3564048111,-0.1347399056,0.1367656142,-0.3409821987,-0.0871532038,-0.0929877833,0.2171245515,-0.2840718329,0.143404901,-0.2832876742,-0.118183583,-0.0723359063,-0.1272996366,0.397706151,-0.0696901605,-0.044878561,0.0667278543,-0.4174093604,-0.1143618003,-0.1739067137,0.1288141161,-0.1347729266,0.133372128,-0.0616762154,0.2145868838,0.0732686222,-0.169459939,0.1197837666,0.2172562182,0.0872647241,0.0729128122,-0.0300873164,0.0965830013,-0.1226501316,0.2455373704,0.1651061922,-0.1488167644,-0.0761009157,0.0207682718,-0.3127219379,-0.0915166587,0.1688801795,0.105879724,-0.1790584028,0.1402069628,0.136084199,-0.0907972157,-0.3047740459,-0.2188349068,0.400414288,0.1239721105,-0.1787181944,0.0696369112,-0.0565709136,0.1327507049,-0.085242942,0.3196436763,0.0953551531,0.3395170271,0.1357048154,0.2389128059,0.1446427405,-0.0525933132,0.0786919221,0.1679394394,-0.0646350384,-0.1023811176,0.3450419605,-0.1604188383,0.0050425618,0.2077207118,0.5121745467,0.1764217913,0.0708955973,0.0435595997,-0.0611322895,0.2262661755,-0.1423236281,0.2956945598,-0.1136340275,-0.2759338319,-0.142478466,0.0311708935,-0.0090724835,-0.2327331752,-0.4316939116,0.7024177313,-0.2806922197,-0.3212979734,-0.1667735279,0.140896365,-0.132092461,-0.2029651999,-0.1144987121,-0.1392981708,-0.1908464432,-0.1848258227,0.2290685773,-0.0132926665,0.0634706765,0.1439854503,0.0835380256,-0.2804713249,-0.4855870306,0.1773281544,0.1425659657,-0.041545663,0.0629001856,0.1436041445,-0.1707988381,0.1758076847,0.6104492545,0.2904420793,0.0435973965,-0.1393885911,0.2393365204,0.2593470514,-0.0496915355,-0.1322289705,0.1564975232,0.2695860565,-0.4744098485,0.0466503762,0.085416764,-0.1447212845,0.3480509222,-0.3163058758,0.3591749668,-0.1538483649,0.3500984907,-0.2364439517,0.1475689262,-0.1308512539,-0.1009435579,-0.7179360986,0.1895143837,-0.1232710183,0.0928874984,0.0719829649,-0.3484199345,0.0471337922,-0.1072719693,0.2144947201,0.430064559,0.1629473269,-0.2852113843,-0.3108713031,-0.3476507962,0.2463074178,0.0730978101,0.2230472267,-0.1770630032,0.4440940619,-0.1031215563,0.0922273472,0.2410821468,0.1612479538,-0.0969933718,0.0524557419,-0.4813613594,0.1660904884,-0.1789522767,-0.056241408,0.0814190358,0.2535749674,-0.1228970662,-0.1291749179,0.0204990096,0.1910954565,-0.2418768257,-0.2137036473,0.1613465697,0.5954045653,0.1722678989,0.1285201162,0.0685032606,-0.1416748613,-0.3088629544,-0.242384851,0.0062135551,0.0871116668,-0.171327576,0.3439906538,-0.2581760287,-0.422747761,-0.3795552552,0.3892647624,0.1324902773,-0.1534838825,-0.1188772619,0.3110212386,-0.071071893,0.0651843771,0.1813978553,0.2737930119,-0.097420834,0.3329169452,-0.3157681227,-0.1971639097,0.6771646738,-0.3671158254,0.2805627584,-0.2694847882,0.5114606619,0.3368268907,-0.3409715295,-0.3263522387,0.1121220365,0.0807697028,0.0419810526,-0.1084783673,0.1534227282,-0.345520556,0.0323721021,-0.121545352,0.3072331846,0.1546478868,0.1746728569,0.0677532926,-0.1550117731]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"Thank you for clarifying that the metrics files are to be found elsewhere, @lhoestq \r\n\r\n> The cache at ~\/.cache\/huggingface\/metrics stores the users data for metrics computations (hence the arrow files).\r\n\r\ncould it be renamed to reflect that? otherwise it misleadingly suggests that it's the metrics. Perhaps `~\/.cache\/huggingface\/metrics-user-data`?\r\n\r\nAnd there are so many `.lock` files w\/o corresponding files under `~\/.cache\/huggingface\/metrics\/`. Why are they there? \r\n\r\nfor example after I wipe out the dir completely and do one training I end up with:\r\n```\r\n~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock\r\n```\r\nwhat is that lock file locking when nothing is running?","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":93,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nThank you for clarifying that the metrics files are to be found elsewhere, @lhoestq \r\n\r\n> The cache at ~\/.cache\/huggingface\/metrics stores the users data for metrics computations (hence the arrow files).\r\n\r\ncould it be renamed to reflect that? otherwise it misleadingly suggests that it's the metrics. Perhaps `~\/.cache\/huggingface\/metrics-user-data`?\r\n\r\nAnd there are so many `.lock` files w\/o corresponding files under `~\/.cache\/huggingface\/metrics\/`. Why are they there? \r\n\r\nfor example after I wipe out the dir completely and do one training I end up with:\r\n```\r\n~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock\r\n```\r\nwhat is that lock file locking when nothing is running?","embeddings":[0.2474486679,0.0509269387,0.0548085459,0.2140230685,0.1152904257,0.1876896173,0.2276507616,0.3381584883,0.2461999655,0.1004048139,0.1633262634,0.120030582,-0.3513062,-0.1723562777,0.0993687585,-0.0273564663,0.108128421,-0.0378123932,-0.2317160219,-0.164974153,-0.2525990903,0.5168857574,-0.0808430165,0.0938802287,-0.564119637,0.2315038443,-0.1601383835,0.2439415455,-0.0434489399,-0.5871254206,0.3680899441,0.1198711395,-0.0636295155,0.5067222118,-0.0001260311,-0.1185866445,0.2515877485,-0.1634837091,-0.2404909581,-0.1307276189,-0.0129830427,-0.2789452374,0.4915622771,-0.2036509514,0.1352380216,-0.0842312649,0.1214475334,-0.4259888232,0.3512769043,0.0431348793,0.0366241485,0.2477326244,-0.1989923418,-0.288210541,0.1527168602,-0.0924047977,0.0252721328,0.3664670289,-0.0836111903,-0.0094820866,-0.2560975552,0.3051272333,0.201936543,-0.1374050081,0.3521669805,0.0150463143,-0.0426552668,-0.2301341146,0.0013805062,0.2264509797,0.2374706417,-0.1862585992,-0.2228959501,-0.0533670969,-0.0899376571,-0.2590976655,0.2228358388,-0.0521903224,0.0696589276,0.1424809545,0.0115555236,-0.1976010054,0.0148122869,-0.0037729971,-0.1053485796,-0.0057458789,-0.1575381905,-0.0152483471,-0.0514961481,0.1989749968,-0.3450107872,0.3006658554,-0.0482684039,0.2071722299,-0.2239120901,0.0468739755,0.1978999823,0.1199527979,0.0460014939,0.4275059104,0.0228465553,0.0263509676,0.4165912867,0.0909777731,-0.1048306525,0.4695401788,0.3464441597,-0.0673513412,0.0851758644,0.1602268964,-0.1593470424,-0.2561867535,0.1269487292,-0.2231046706,0.2775918245,0.0905745476,0.0148455482,-0.409976095,0.0961333141,0.3147747219,-0.2490134984,-0.23452878,0.0893958881,0.3153007329,0.2518271208,-0.0041998359,0.1808401793,0.209176138,-0.2426620275,-0.0500778817,-0.3978200257,-0.1054066643,-0.2283210009,0.3502309918,0.0507635511,-0.1394725144,0.7004565001,-0.260506779,0.3893989325,-0.1397276372,0.2826374471,0.2533380091,0.0007684092,0.4032491148,-0.0881246626,0.2115938067,0.3516604602,0.0514885075,-0.1892664135,-0.2525210977,-0.2373767644,-0.5181609392,0.0860111713,-0.0284101684,-0.5039556623,0.2390986979,0.4408912063,0.1475693733,-0.0292732287,-0.1374207586,0.2848838866,0.0987269208,0.0174057744,-0.1247298494,0.3923794925,0.7586100101,-0.2284236103,-0.4756804407,0.1044251546,-0.0117908381,-0.0543393828,0.147732392,-0.0291076768,0.2321294397,-0.28400141,0.1421148628,0.3668580651,-0.3609597385,-0.3620804846,-0.0106194112,-0.3919947743,-0.1379084438,0.1821041554,0.0160973482,0.0728418529,0.143812865,0.1258191019,-0.3850254416,0.1877111942,-0.4215300977,-0.3467738926,-0.1896634847,-0.0304805059,-0.1046682894,0.2787402868,0.037039917,0.0111542493,0.1009099483,0.103630051,0.174416706,0.1748632342,0.2626431882,0.2613887191,0.1620635241,0.2222009897,0.1930808276,-0.311049521,0.2377901077,-0.3628157973,-0.0176313967,0.1102191657,-0.073212944,-0.1398175955,-0.2352026701,-0.1244361848,-0.527023077,-0.0462953858,-0.0360660926,0.1183855608,0.1227987111,-0.2726368308,0.1473123133,-0.0587887019,0.3274554312,-0.5532151461,0.0320394821,0.0278662313,-0.1970100999,-0.2257992178,0.1093442887,-0.0025995036,0.0232681595,-0.0975060761,0.380703032,-0.0499365255,0.1832061112,0.5586856604,0.3541171253,-0.1420735121,-0.3481737375,0.0158954021,0.0016878804,-0.1003440693,-0.0012896619,-0.2387945056,0.3765332997,0.134847939,0.1348783821,-0.3614094555,0.2761767507,-0.2093584388,-0.1233218312,-0.5444701314,-0.1353482604,0.4068960547,-0.2476063371,0.3068441749,-0.0123205194,0.0024220273,0.0304550156,0.1419316679,0.0483441949,0.0429760255,0.0081629893,-0.2377903163,-0.053385146,-0.0278111082,0.003766177,0.4933545589,0.1662803441,0.2230534703,0.014451649,0.1677996218,-0.2444672883,0.0901625752,0.1739695817,-0.2210897356,0.1015084758,0.1378706992,0.2083474994,-0.2946519852,-0.050339099,-0.2348016798,0.0318766832,-0.3016398251,0.1512337625,-0.1539792418,0.3830914795,0.0986963734,-0.0210696273,-0.2364776731,-0.333298862,0.2775138319,-0.3183354437,-0.0270291232,0.2005967796,-0.3842666745,0.6387848854,0.018266527,0.016717406,-0.2889690101,-0.0855253935,0.0128287217,-0.2142503858,-0.1412100494,0.0268610641,0.0817524493,-0.2668958008,0.0615631044,-0.1273287684,-0.4955730736,0.0512227751,-0.0161663014,0.6132103205,0.2828388214,0.0391496904,-0.3447293341,0.0620119385,0.2875090539,-0.2682441473,-0.3798858821,0.0450980328,0.0512684956,0.2367150038,-0.3826585412,-0.333468914,0.1506506801,-0.3048092723,0.5795782208,-0.0784145147,-0.0105534084,-0.1236330047,-0.134855777,0.2229509801,-0.34937343,0.2714141905,-0.3231592476,-0.7829341292,0.4090441465,-0.0861361027,-0.0492391251,0.0389753282,0.0890199468,0.3141541183,-0.1025536954,-0.7185192108,-0.4109581411,0.0587699935,0.1837124825,-0.042348247,0.0645665973,0.4620451033,-0.0869352445,-0.0113906544,-0.1259307265,-0.1665929556,0.4881491363,-0.1238272935,0.186942786,-0.1846831143,0.3211714923,0.1527525187,1.0194034576,0.5859872103,0.3531093001,0.1362784952,0.2488948405,0.4330224097,-0.1152088344,-0.0847556815,0.2251204997,0.2476467788,-0.1349827945,0.3065981269,0.2231494188,0.1870115548,-0.0333176404,0.0093365358,-0.3177512586,-0.2941355705,-0.0049865786,-0.1403052658,0.2279344946,-0.076974161,0.1744736433,-0.1848286986,0.0769245327,0.3933352232,0.5813077688,0.3059713542,0.261300385,-0.1821739674,-0.1309935451,-0.5380686522,0.1992972493,-0.4458011687,0.0942804292,-0.2295228839,-0.3128388226,0.2112236172,0.1381264627,0.5694957972,0.1833454072,-0.2201473713,0.0576127656,-0.1568565667,-0.2315971553,-0.1577763259,-0.0046724193,0.0672969297,0.2795846164,0.3941031396,-0.3905998766,-0.3081789017,-0.1242815703,-0.1054901257,-0.3238491416,-0.2797122002,-0.2802339494,0.02357479,-0.0901425183,0.0684608221,-0.2081900835,-0.0828768462,0.017027054,0.111315839,-0.0379926972,-0.0876011997,0.1582567692,0.074530527,0.2948908806,-0.233771503,0.1438097805,0.5626406074,-0.055148486,0.0503553301,0.4550115764,-0.0244702697,-0.0061214357,0.3603450358,-0.15289931,0.3670763671,0.6862469912,-0.1970186979,0.1456226557,0.06505806,0.3557294309,-0.0740640014,-0.0981433764,0.3586219549,-0.1224850044,0.3750058711,0.1454334855,0.247368753,0.243081972,-0.1910924315,0.306197077,0.1086222678,-0.1157655567,0.3354496062,0.0150604593,1.0408422947,0.244695887,0.2499573082,0.0714620873,0.1312486082,0.3983026147,0.064664945,0.0467557274,-0.1198779717,-0.1944516152,-0.0132674538,-0.1940737516,-0.0282721445,-0.2923018336,-0.1856505871,0.4099218547,-0.4929338396,-0.155075267,-0.1195691004,-0.3021264076,-0.2614345551,-0.0750590488,0.1547853053,-0.0412603728,0.1292809695,0.5553107858,-0.2514209747,-0.223262459,-0.3335103691,-0.2589264512,-0.148975879,-0.2111898661,-0.1891939342,-0.023680266,0.3486911356,-0.1595155746,-0.1513006836,0.2677080631,0.397124052,0.0755348653,-0.3020040393,0.0638313144,0.1680329293,0.0085620219,0.0653498694,0.0340459757,0.3502122462,0.0003313319,0.0511186793,0.0249727294,-0.3353211284,0.0492204316,-0.0794571042,-0.083564274,-0.0526298247,-0.0755612701,0.0267641935,0.204214558,-0.0968355313,-0.1874864846,0.03314659,0.3781017363,-0.1960804462,-0.0299340989,0.1211581752,-0.3434658349,-0.0560230091,0.434222877,-0.1925750226,-0.1195441559,0.5066856742,-0.0645434037,-0.1223668829,-0.0741554722,0.1144800708,0.4430142939,-0.7368456125,0.0767200142,0.0242282525,0.027922526,0.076529488,-0.007154345,0.5015127659,-0.1989104003,-0.2411555052,-0.2157037556,-0.3882722855,0.3828385174,-0.304048866,0.1324845403,-0.6194877028,0.0594516695,-0.0539754406,0.1296790838,-0.2124767751,0.2109172344,-0.2437586039,-0.1170596927,-0.0746393576,0.0346618779,0.2913802564,-0.0565141439,-0.1333194971,0.0286500026,-0.3847769797,-0.0517192818,-0.1915149391,0.1899384111,-0.1081529036,0.2442276031,0.0066732313,0.3041701615,0.0198468082,-0.0583771504,0.0101103438,0.2571412325,0.1279606819,0.1851983666,0.1432306468,0.2480728626,-0.0729077011,0.2049413323,0.1763910353,-0.1000491232,-0.1000215933,-0.1039726287,-0.428586185,-0.0848947763,0.1566678882,0.1853019148,-0.3170948923,0.099970974,0.0618540309,0.0173467807,-0.235360831,0.0403751731,0.211148113,0.224532783,-0.2050518841,0.0644706413,-0.0084532341,0.0521701872,0.0229844693,0.2303782552,-0.0552917682,0.4240315855,-0.026842339,0.2237032205,0.1670225263,-0.1279317737,0.1427684724,0.1989018768,-0.0372638181,-0.0433573648,0.3380188644,-0.0348775722,0.1034381688,0.3432768583,0.5569766164,0.1540989727,0.0934894681,-0.0954653397,0.0071917418,0.2507784069,-0.0900275186,0.2582259178,-0.2793651819,-0.1373146027,-0.221054703,0.1737846285,0.0515683591,-0.1402488947,-0.3882140517,0.8121156096,-0.1609298587,-0.2608853281,-0.0700840876,0.1249645352,-0.1619194895,-0.1566223651,-0.1197467893,-0.1842712611,-0.1253316104,-0.1607531756,0.2283628881,-0.0062963702,0.1427665353,0.0652464554,-0.0398544818,-0.1074960232,-0.4368215203,0.2725354433,0.2284072787,-0.2035957575,0.0297514796,0.1274764836,-0.2528050542,0.2495367825,0.5566775203,0.2099296004,-0.0059626494,-0.1163639277,-0.0035180978,0.288309902,-0.0202122722,-0.0466433503,0.240210861,0.1830725074,-0.6153120995,-0.0621646047,0.0086093564,-0.107665062,0.3888005316,-0.2084245831,0.3229672313,-0.1921549588,0.3530454934,-0.2691180408,0.0926729515,0.0585445426,-0.0918286666,-0.7468347549,0.213738054,-0.135936141,0.0162219051,0.0330396555,-0.2474638373,0.0152216572,-0.0955320373,0.1805478781,0.553049624,0.0532680713,-0.3335349858,-0.3067726791,-0.4826698899,0.2709026039,0.1446658522,0.4335974455,-0.0228388924,0.3380472958,-0.0630271062,-0.0161664151,0.2461424321,-0.0855208933,-0.161103189,0.0040011504,-0.5652293563,0.1719986647,0.0195857324,-0.0025388629,0.1001019552,0.292696923,0.0207593013,-0.233472392,-0.0914474502,0.2584428191,-0.2112590671,-0.0553373806,0.3531959951,0.4249581099,0.0721086636,0.0359956324,0.0304970238,-0.1994164884,-0.2580087185,-0.1420315653,-0.0411230624,-0.1456901729,-0.2933090031,0.3412899673,-0.4688861966,-0.4392541051,-0.2739019394,0.2539839745,0.0947203115,-0.1432622969,-0.3065039814,0.3340900838,-0.212044552,0.0159971807,0.3520705104,0.3827580512,0.0340382159,0.2756767869,-0.3647694886,-0.1141464263,0.7994841933,-0.5541844368,0.2426104844,-0.1927599162,0.3955248594,0.5046318769,-0.3508092761,-0.4224117994,-0.096780844,0.1382532716,0.0550413579,-0.2188415974,0.1441938281,-0.4663903415,-0.0743133128,-0.0288978759,0.2651972175,0.1610632688,0.0547087938,0.0375731848,-0.1937296838]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"The lock files come from an issue with filelock (see comment in the code [here](https:\/\/github.com\/benediktschmitt\/py-filelock\/blob\/master\/filelock.py#L394-L398)). Basically on unix there're always .lock files left behind. I haven't dove into this issue","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":30,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nThe lock files come from an issue with filelock (see comment in the code [here](https:\/\/github.com\/benediktschmitt\/py-filelock\/blob\/master\/filelock.py#L394-L398)). Basically on unix there're always .lock files left behind. I haven't dove into this issue","embeddings":[0.0978651121,0.0017499296,0.1070125327,0.1418968439,-0.0015061367,0.0359290205,0.2448608726,0.2713269293,0.2091092318,0.1213858202,0.1688691676,0.1361201853,-0.3981262147,-0.0413279571,0.0657574683,0.077555269,0.0212283693,-0.0395922288,-0.3166533113,-0.1807789952,-0.2565910518,0.5491791368,-0.1192011982,0.0093765138,-0.5109118223,0.2307155579,-0.2306171954,0.277212292,-0.0274775904,-0.581561625,0.389434129,0.121536456,0.0478550903,0.4788641036,-0.0001274045,-0.1663458943,0.1572808176,-0.1869959533,-0.1530368477,-0.1362828463,0.0305324215,-0.296793133,0.5233107805,-0.2339410633,0.079851374,-0.0288459454,0.102354154,-0.2997952998,0.4383219481,0.0624674,0.0467604399,0.3075246513,-0.0342115685,-0.2570320666,0.0702645779,-0.0924271196,0.0015265134,0.3444680572,-0.2050896734,0.0921004117,-0.2054682672,0.2364491075,0.2473953664,-0.1546228379,0.2657466829,0.0201909356,-0.0640991554,-0.1387817711,-0.0949208513,0.3408105969,0.1928483993,-0.1521597654,-0.23427926,0.0592422895,-0.1122282296,-0.3646720052,0.2042401731,-0.0701380074,0.1202769727,0.1426341534,0.0680519566,-0.1823931187,-0.0471961349,0.0707622021,-0.1521102935,0.0407557935,-0.1404026747,0.0206547938,-0.0408887789,0.1827445179,-0.4113306701,0.3544477224,0.0115481168,0.1267995685,-0.3444469571,0.1168110892,0.2434752136,0.028262224,0.0463358462,0.4056189358,-0.058036983,0.0199387167,0.3959121406,0.1504532993,-0.0784413069,0.4660251439,0.2361447364,0.0168702435,0.1764601469,0.1849360615,-0.2115632892,-0.2638486624,0.1365805715,-0.236214444,0.4281723499,0.1122872084,0.110580571,-0.4319956005,0.1628943384,0.2799922526,-0.1437161267,-0.1873987168,0.0960946009,0.3751057386,0.1744754314,0.0033347695,0.2302564681,0.3915508389,-0.2153754234,0.0076531689,-0.3904958367,-0.2105622739,-0.0540188849,0.3026128113,0.135630101,-0.1125969067,0.6726797223,-0.2044869661,0.4261627495,-0.0901329294,0.3285324275,0.2119804323,0.0645420924,0.3416905999,-0.0264643971,0.2123863548,0.2896437049,0.1480878145,-0.2393878549,-0.2758550048,-0.1812229455,-0.5380879641,0.0798784643,-0.0424434766,-0.4659706652,0.1792515814,0.4528223276,0.1464153528,-0.1015214771,-0.3035165668,0.2009056956,-0.0098103043,-0.008477659,-0.1502698809,0.5102133155,0.6958425641,-0.3465329707,-0.3804927468,0.0367823429,0.0076954146,-0.0054186331,0.2611059546,0.0363228954,0.1693779826,-0.3158248365,0.1305027753,0.3877377808,-0.3373550475,-0.4568701684,-0.0679737255,-0.3683251441,-0.1110464409,0.1929325461,0.0133991111,0.1307076961,0.1547204852,0.0300893057,-0.219019711,0.2606108487,-0.3739937544,-0.3569114506,-0.2306001633,0.0067292596,-0.0920921713,0.3358702362,0.0176148601,0.0747954398,0.0229984261,0.0826772675,0.0752030835,0.162635982,0.278436482,0.3372353315,0.1000661328,0.2661374509,0.1485668123,-0.2092305124,0.2349851131,-0.4146922231,0.0708448216,0.1174756885,0.0061449613,-0.1154760197,-0.2034732699,-0.0905117095,-0.4177539051,-0.0547221527,-0.049519483,0.2109538019,0.2493541539,-0.1767959446,0.0389834866,0.086997278,0.2594942451,-0.4751551449,0.0103413193,-0.0947172791,-0.335790813,-0.0653162375,0.1485408694,0.0108669186,0.0354320034,-0.1438509971,0.3710696399,-0.1724669188,0.2730033696,0.450623095,0.2415375561,-0.1196843088,-0.3655080199,0.0325872302,0.1424403638,-0.1088574454,0.0363418609,-0.2451748252,0.3829999268,0.21202676,0.0713483319,-0.3359493613,0.3146027327,-0.0740355998,-0.0987766758,-0.4815725982,-0.1280817837,0.3917382658,-0.2381811887,0.2410218567,-0.1055556089,0.0105510326,0.0025617529,0.2497153431,0.07995525,0.233252883,0.0780286416,-0.2227705419,-0.0949146673,0.0184847414,-0.1300748289,0.4449999034,0.190678373,0.227855593,0.0538468696,0.1052548885,-0.188812986,0.1054110155,-0.0117074698,-0.16059497,0.1285531968,0.2171354592,0.2108034641,-0.2892757952,-0.1394120455,-0.4038005769,-0.0391597971,-0.3721092343,0.0799339265,-0.2586631775,0.4104159176,0.183658272,0.173447758,-0.2760920227,-0.3279844224,0.2431277186,-0.2901171744,-0.0558715053,0.2372846007,-0.3406104147,0.5479469299,0.1450073868,-0.2062622458,-0.295977056,-0.1340124011,-0.0942150056,-0.2200794071,-0.2425167114,0.0989216641,0.1037221998,-0.1725022644,0.0372332409,-0.0137783829,-0.4823435843,0.0696525127,0.0511408858,0.6196216345,0.3521490395,0.0815300345,-0.3057146072,0.083496958,0.4136028588,-0.4095363319,-0.2688114941,0.1824415028,0.0434782021,0.2113901079,-0.2620833814,-0.3617442846,0.0265498981,-0.2976148427,0.6853885651,0.096109271,-0.0340798348,-0.1620347947,-0.0633253679,0.1737497002,-0.323364228,0.3805911243,-0.2930787206,-0.6595548987,0.5047425032,-0.1025177017,-0.2188171595,-0.0036964265,0.0586478226,0.3344171047,-0.089447394,-0.6739205122,-0.4296785295,0.0067402273,0.2661030293,-0.0254283547,0.1197582409,0.5252697468,-0.0839203671,0.0137430709,-0.0674872324,-0.2023647428,0.5829992294,-0.0621240363,0.1375797838,-0.082530953,0.3729041815,0.12414217,0.911796689,0.4719629884,0.3536292315,0.2457000464,0.0920005068,0.3813191652,-0.1479361802,-0.1129252985,0.3361164033,0.2327181101,-0.1589669287,0.2437696308,0.1992550939,0.1921728849,-0.0608860664,-0.0730623081,-0.3863719702,-0.3076612055,-0.0325332433,-0.0471550412,0.2321032286,-0.0069956719,0.0424787961,-0.2280805558,0.0112633677,0.4220144153,0.4312678277,0.2322890311,0.2378754318,-0.1441177726,0.0589368418,-0.5730276108,0.2734699547,-0.5298247337,0.0832474679,-0.2206547707,-0.3305592239,0.1510061175,0.0413568094,0.5898941159,0.1778952181,-0.1399519295,-0.0725890324,-0.2002184987,-0.4257732332,-0.1523622572,-0.1288876981,-0.117143251,0.3681684434,0.3341633081,-0.3990162611,-0.1482420862,-0.1328407526,-0.0242101662,-0.2809101045,-0.3642933071,-0.2995695174,0.0558415838,-0.2109502554,0.079452239,-0.2678171992,-0.0450141095,0.0515149496,0.0376873612,-0.1677218527,-0.1312997192,0.0200150106,0.0744418949,0.3540909886,-0.3674106896,0.1572969705,0.4705135226,-0.0993365124,0.0339576118,0.3758528233,0.0386760496,0.007548043,0.323161751,-0.1632548422,0.2435810864,0.5914896727,-0.3155204058,0.0485774055,0.0017071113,0.4769366384,-0.1545539647,-0.1676850319,0.3743362129,-0.1851758659,0.3994726539,0.217222631,0.1465723962,0.2307453603,-0.1425453275,0.213699311,0.0921448246,-0.1486644,0.4187479615,-0.0191742424,1.0156917572,0.2073533237,0.2827003896,0.132918939,0.1772165596,0.459148854,0.1079502553,0.0556265265,-0.209124878,-0.2929944396,0.0010578317,-0.1498106271,-0.0807728469,-0.2313542366,-0.2762808502,0.4570978582,-0.3854297996,-0.1603196114,-0.1962326169,-0.2233338952,-0.3173439503,-0.105138436,0.0847922638,-0.0389764048,0.1487234831,0.5471583605,-0.222628504,-0.1768111289,-0.3493130207,-0.3877286017,-0.1265837997,-0.033755485,-0.2098589391,0.0311239045,0.3574776947,-0.1190107688,-0.2099216133,0.2792097926,0.3486991227,0.1478072107,-0.3286235332,-0.0536580309,0.1387420744,0.0693156719,0.1134880632,-0.0162831098,0.2863367796,-0.0592635758,-0.0206086561,0.0139444247,-0.2617380917,0.0625092834,0.0623129234,-0.1763906926,-0.2089602947,-0.04637748,-0.0278090704,0.0901816934,-0.0493521243,-0.2061769813,0.0301746074,0.350042522,-0.238143146,0.0757837519,0.1948084086,-0.2536814511,-0.1046510115,0.5353069901,-0.2839657366,-0.0404862165,0.5257986188,0.0521452501,-0.0739264041,-0.0879664049,0.1231449842,0.3842419684,-0.8295775652,0.0623208731,-0.0147827556,0.0206483658,0.0930647627,0.0981906578,0.4433169067,-0.1498861462,-0.1371233463,-0.2625291049,-0.2995174825,0.2950277925,-0.3477450907,0.1463387907,-0.4390567243,0.0171232522,-0.1422357559,0.1940521598,-0.2237335593,0.1365989149,-0.2280653566,-0.1690292358,-0.1042840555,-0.0442445725,0.3885759413,-0.0331422053,-0.084089838,0.0814567134,-0.3113950491,-0.0289357062,-0.2361383885,0.1762454212,-0.1520950347,0.1542677134,-0.0213325545,0.2418326437,-0.0256441422,-0.1351262033,0.0199913736,0.174633652,0.0819578022,0.18294999,0.1014149264,0.1096100286,-0.1310430169,0.2293001264,0.082787618,-0.2001680285,-0.0907208025,0.0441642441,-0.4078601599,-0.0517290346,0.1881525964,0.1526675522,-0.2726956308,0.1853771508,0.0912855491,-0.0026348277,-0.3078107238,0.0038793411,0.4702510834,0.1285609007,-0.2048867941,0.0074314252,-0.0171015058,0.052266404,-0.1066366136,0.256721288,0.0800266117,0.4084855318,-0.0540620275,0.2842751443,0.1380118728,-0.0800259262,0.0952327177,0.2558289766,-0.0737559572,-0.1462478936,0.3656685948,-0.0703354552,0.0768483505,0.3500859737,0.47741431,0.1101920158,0.1348928809,-0.0343452282,-0.0622371025,0.1586259753,-0.0870699137,0.2827890813,-0.1443686336,-0.200066179,-0.1374969184,0.0946215987,0.0407014005,-0.1743674725,-0.4197376668,0.7717667818,-0.2196433544,-0.2426420599,-0.1869886816,0.2187612504,-0.1312939525,-0.102885671,-0.1330425888,-0.1627594084,-0.1881944239,-0.1369980574,0.2413048595,0.0748141631,0.0135711208,0.1831946522,0.0779207796,-0.2401618809,-0.4097238779,0.1471295357,0.2348361909,-0.1723956168,0.0443609171,0.2696973383,-0.2988671064,0.18265903,0.604993701,0.3143845201,0.0188652258,-0.1037897468,0.1164325178,0.2726605833,0.0104301004,-0.1231349632,0.1442331225,0.2494451255,-0.4571605027,0.0450880826,0.0002661294,-0.1097639054,0.4263656139,-0.2372503728,0.3066159785,-0.1923083812,0.4076460898,-0.1710709184,0.1075957641,0.0466245301,-0.139591217,-0.7387015224,0.2041074038,-0.1014896855,0.0151310833,0.068896085,-0.2466745526,0.0107893581,-0.148622185,0.1965500265,0.5074281693,0.0630350709,-0.2921724021,-0.1924867034,-0.4941076636,0.312924087,0.127784282,0.2551854849,-0.0394156538,0.3570337296,-0.0903412253,0.0310946181,0.2803325057,-0.0243291166,-0.110245496,0.0334424935,-0.4978340566,0.1731984019,-0.0518372804,-0.0960942507,0.064528577,0.2707467377,-0.044634264,-0.2914806306,-0.0436215252,0.2637952268,-0.2875222564,-0.1453561634,0.1927300245,0.4304181933,0.1727121323,0.1035242975,-0.0241545495,-0.1447631568,-0.175772652,-0.1006726399,-0.0385693237,-0.1139795259,-0.2793585956,0.4239613712,-0.399751544,-0.4653254747,-0.3727272153,0.3133134842,0.0486235693,-0.1338500679,-0.2125529796,0.3537247479,-0.1586045921,0.0163044762,0.2923788726,0.2744099796,-0.0121014258,0.3496609628,-0.2905235589,-0.1873505563,0.722386539,-0.5237223506,0.3236785829,-0.2664921582,0.4198946655,0.3544078171,-0.3350405097,-0.3600738645,-0.0202656686,0.127798453,0.0886094347,-0.155157119,0.0890782326,-0.3894797564,-0.0616638288,0.0164001454,0.239748925,0.168612659,-0.0078574549,-0.0522728525,-0.1597507149]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"are you sure you need an external lock file? if it's a single purpose locking in the same scope you can lock the caller `__file__` instead, e.g. here is how one can `flock` the script file itself to ensure atomic printing:\r\n\r\n```\r\nimport fcntl\r\ndef printflock(*msgs):\r\n \"\"\" print in multiprocess env so that the outputs from different processes don't get interleaved \"\"\"\r\n with open(__file__, \"r\") as fh:\r\n fcntl.flock(fh, fcntl.LOCK_EX)\r\n try:\r\n print(*msgs)\r\n finally:\r\n fcntl.flock(fh, fcntl.LOCK_UN)\r\n```\r\n","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":75,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nare you sure you need an external lock file? if it's a single purpose locking in the same scope you can lock the caller `__file__` instead, e.g. here is how one can `flock` the script file itself to ensure atomic printing:\r\n\r\n```\r\nimport fcntl\r\ndef printflock(*msgs):\r\n \"\"\" print in multiprocess env so that the outputs from different processes don't get interleaved \"\"\"\r\n with open(__file__, \"r\") as fh:\r\n fcntl.flock(fh, fcntl.LOCK_EX)\r\n try:\r\n print(*msgs)\r\n finally:\r\n fcntl.flock(fh, fcntl.LOCK_UN)\r\n```\r\n","embeddings":[-0.0958084464,-0.0942348838,0.1255570352,0.0927403942,-0.0868547782,-0.0296325516,0.2317219824,0.2375220805,0.2348071039,0.1782363057,0.0034624427,0.14741005,-0.2378918976,-0.0178872645,-0.0204138942,0.1772465557,0.007517321,-0.0802016482,-0.3421198428,-0.1448375136,-0.2741203606,0.53156811,-0.1308241487,-0.0719346926,-0.4290018976,0.2123154998,-0.2263918221,0.1211149171,-0.0081585767,-0.3395671546,0.1049853936,0.3507529199,0.1030155346,0.4625885189,-0.0001223227,-0.2479463816,0.1706754416,-0.1718239933,-0.1011753306,-0.0645022318,0.1405719966,-0.3108265102,0.4765364528,-0.2978203893,0.1028483659,0.0841091052,0.1408373266,-0.3395661116,0.43585217,-0.0378726833,0.0893844143,0.2628902197,0.1200164706,-0.1775288135,-0.0331903175,-0.1780336201,0.0121771665,0.2705992162,-0.1119368449,0.0844490603,-0.2585883141,0.2509045303,0.2237693965,-0.2024630159,0.1423094124,0.0610956363,-0.1400081366,-0.2547037601,-0.1634257287,0.2697402239,0.1364804059,-0.3204964697,-0.3008389771,0.0492605083,-0.117218554,-0.3431062102,0.0262462962,-0.0340269618,-0.0005137632,0.0618584864,0.1232936308,-0.1993052363,-0.0104406094,0.0003190125,-0.013987557,0.0958361998,-0.1606515646,0.0895802081,-0.1322297305,0.2888585627,-0.3445695341,0.3877014816,0.0099186832,0.1042271033,-0.2629313469,0.1574040651,0.2871651947,-0.1045001149,0.1025784984,0.3618665636,-0.0343261063,0.0118681435,0.5354156494,0.1580042988,-0.0143520003,0.3216001093,0.2761745751,0.0193977691,0.2049499601,0.4004882574,-0.2367353141,-0.2947693467,0.2996625304,-0.209830761,0.3451402485,0.1116673872,0.1468068808,-0.474060744,0.3149008751,0.3492853642,-0.0679861754,-0.0676057488,-0.0274083279,0.3686693311,0.2322140932,0.0972456336,0.359259367,0.2219015509,-0.2242739499,-0.0752867386,-0.3204058111,-0.1706400663,-0.0248745047,0.4072385132,-0.0777525827,-0.0594260283,0.5295740962,-0.0586507991,0.5135621428,-0.0885761976,0.0644367635,0.1758888364,0.0725545138,0.2391681373,0.0254030153,0.2357241213,0.3454734087,0.2439838797,-0.2318499535,-0.3414825499,0.0639522895,-0.4619214237,0.1058227047,-0.0182386171,-0.3980556726,0.195152536,0.6058903337,0.1725607663,-0.1805835366,-0.2601388395,0.3114195168,0.0852632076,-0.0760378465,-0.0874109194,0.4565057456,0.7431334853,-0.3583593667,-0.3069776893,0.1075284332,-0.1093930081,0.0635075793,0.1491250396,0.0973032042,0.096509248,-0.2766647339,-0.117821686,0.3571416438,-0.4726914167,-0.472574532,0.0852729604,-0.4719013572,-0.0274127889,0.2066820413,0.007604328,0.2238250226,0.2170172036,0.1575410813,-0.1981640011,0.2113953978,-0.2767271996,-0.3929208815,-0.1785394549,-0.1324971318,-0.0649424791,0.2047929466,0.0115920985,0.0840649679,0.0313538127,0.2286920995,0.0897837356,0.188824147,0.2832199037,0.3775959313,0.1742916405,0.2464331836,0.1135068908,-0.2764574587,0.2135853469,-0.4569099247,0.1178642586,0.0813397989,-0.0489264615,-0.021800328,-0.1058617309,-0.0418673828,-0.4147400856,0.0085285567,-0.0434777662,0.3030488491,0.249597609,-0.1796232611,0.2385921478,-0.0130311493,0.291190058,-0.4203335643,-0.1497296542,-0.0477628708,-0.3882257342,0.0575981699,0.1718961,0.0015680999,0.0748586655,-0.1377814859,0.4197938442,-0.057103619,0.2876296043,0.2521417141,0.1678893566,-0.1971282214,-0.2420713753,-0.0125151826,0.0657090023,-0.0356098711,0.0254984871,-0.1123302802,0.490195483,0.0970405266,0.0857904926,-0.3372980058,0.2576314807,-0.1290608197,-0.06232173,-0.5252739787,-0.1358704865,0.3787258565,-0.2251196802,0.2366429567,-0.131506294,0.0301113445,0.07980223,0.1785757542,0.2459708154,0.1715176404,-0.0005889279,-0.1394901127,-0.0125992363,0.0347564183,-0.0848471075,0.2582921386,0.2357061058,0.1050706059,0.0881278664,0.1087510213,-0.1732836664,0.1077506989,-0.0638256297,-0.1016960293,0.1692502648,0.1857384294,0.1188376173,-0.2927369773,-0.177698791,-0.3255646229,-0.0320762694,-0.2366850227,-0.0204165373,0.0345232971,0.3458484411,0.009169586,0.0865000635,-0.2900829911,-0.404525429,0.217493996,-0.1798904985,-0.0688815787,0.1823814362,-0.4417120814,0.4122616053,0.1074886024,-0.2418729216,-0.3294233084,-0.1597177833,-0.078658022,-0.2074675411,-0.2837306261,0.1251960248,0.222012341,-0.0486109182,0.0800969303,-0.1392987221,-0.3460685909,0.0014418731,-0.0510540158,0.6285656095,0.2911643088,-0.0305407196,-0.1663736403,0.0547243059,0.449611634,-0.5593898296,-0.1918304265,0.1030515358,-0.0518236309,0.1852306873,-0.255188942,-0.2687283754,0.0448057204,-0.4148482084,0.6636121869,0.0234176144,-0.0101746758,-0.2235938907,-0.1076810881,0.1781018823,-0.2553531229,0.4349594414,-0.1497935802,-0.5847468972,0.4244233072,-0.1281866878,-0.2230434418,-0.1408714801,0.0822282284,0.17599231,-0.1449057758,-0.5932175517,-0.5006266832,-0.1014444679,0.2448946387,-0.0263824984,0.1439140439,0.4210920036,-0.1415075511,-0.0179313254,-0.0129031828,-0.1239335686,0.532720685,-0.1963026226,0.0532477461,-0.0795462579,0.4616315663,0.0842699707,0.8287326097,0.5620367527,0.3317268491,0.1832953542,0.0219699498,0.4228357673,-0.136825338,-0.1904710233,0.3991360962,0.3047939539,-0.2714653313,0.2112302333,0.0435512699,0.2230718881,-0.1163756475,0.0106749507,-0.4465951622,-0.4564572871,-0.0719129443,0.0920207202,0.2174009681,-0.0941524059,0.0339871347,-0.0975612998,0.0202384014,0.5146849751,0.3764460683,0.1630998403,0.1277968138,-0.1811331511,-0.0057472438,-0.6595851183,0.3577421904,-0.4646302462,0.1572023332,-0.1583257616,-0.3413119316,0.1660341918,-0.0564945415,0.6355431676,0.1735782176,-0.1956252307,-0.0046772887,-0.2547474802,-0.4369533062,-0.1027991101,-0.2421707809,-0.1507626027,0.3620065153,0.0918471664,-0.4735630155,-0.1324457675,-0.1301804781,-0.1474509537,-0.3104038835,-0.3958629966,-0.2934837043,0.0456174985,-0.3154401481,0.0738444403,-0.3079227209,0.0215210319,0.0181093942,0.0927206278,0.0271873772,-0.107778497,0.0773513913,0.1022581384,0.2887762785,-0.213611424,0.0983378142,0.437238425,-0.191536963,-0.0192965735,0.3132380545,0.1018751115,-0.0999717265,0.1759297103,-0.0794544742,0.2799620032,0.6032480001,-0.1729381979,0.0549094006,-0.0586394183,0.4818816185,-0.1178966537,-0.2855547071,0.2373222411,-0.1547115296,0.39270401,0.2192185521,0.048576083,0.3099289536,-0.1057202816,0.3555969596,-0.020597877,-0.14115192,0.4236711562,0.1191990823,0.950653851,0.0976581275,0.1860413998,0.2422059029,-0.0376854353,0.4492138028,0.0296556354,0.0142837288,-0.3872033954,-0.2756673992,0.0917524844,-0.1032314971,-0.0816485211,-0.3444148302,-0.2914729714,0.315526545,-0.3911909759,-0.3037447333,-0.3794370294,-0.1960083544,-0.4446938634,-0.2635651231,-0.0477042831,-0.0090243351,0.1433115751,0.5026950836,-0.3078819215,-0.1107050553,-0.2886677682,-0.3712597489,-0.2273342907,-0.0222581364,-0.2027351707,-0.0640979186,0.2401205152,0.0918198079,-0.3199104965,0.3298313618,0.381528765,0.1273179948,-0.2358826399,-0.1780478209,0.2861911356,0.038246911,0.1124107838,0.0014286628,0.2514294982,0.0359148122,-0.0307594091,0.1709094197,-0.1664930731,0.0872220099,0.1005959958,-0.1597257107,-0.1423913389,-0.195840463,-0.0657531768,0.1554861218,-0.0663395226,-0.2104568779,0.0482374355,0.2457901239,-0.314684391,0.1481910199,0.0379270576,-0.2771352232,-0.0750765651,0.5247073174,-0.2640314996,-0.0248469766,0.3970454931,0.2828394771,-0.086016193,-0.1441112459,0.1293323636,0.2986586392,-0.7690415978,0.2018031776,0.1477666795,-0.0192428753,0.0434805043,0.2532590926,0.3413795829,-0.1568590701,-0.0955142155,-0.2206643522,-0.3461404443,0.3158053756,-0.2180731595,0.2099333704,-0.2942487597,0.1017868593,-0.1563838273,0.2841731608,-0.2742088735,0.0098687606,-0.238579616,-0.1825028807,-0.0759445131,-0.0353177227,0.3743842542,0.1295120269,-0.0323035307,0.2220456749,-0.293264389,-0.0894105136,-0.207783401,0.1688523591,-0.1008318067,0.2274570465,-0.0931678638,0.1616178453,-0.1409089118,-0.1505561471,0.121502921,0.3491714299,0.0348307118,0.1579634994,0.2235538512,0.0524991415,-0.1588917822,0.1782961041,0.0967275128,-0.1647049487,-0.0673635155,0.0238149855,-0.4683925509,-0.075499557,0.1976671815,0.1554300785,-0.1301897764,0.1547016501,0.0438508838,0.0733110905,-0.3498418033,0.0681984127,0.5667790771,0.2293316871,-0.1947151423,-0.0845655575,-0.0761872455,0.0748456419,-0.0797744468,0.4089386463,0.0196242929,0.2220222354,0.0884395912,0.3918869197,0.0069045769,-0.1549846232,0.088721931,0.1567456126,-0.1267592311,-0.1499550343,0.2738578916,-0.0977543667,0.1879412383,0.3699198961,0.5017709732,0.1857874393,0.1736930311,-0.1599157602,-0.0085363053,-0.0041767419,-0.1832286865,0.3578150272,-0.0631645918,-0.1351080686,-0.0619017519,0.0247357953,0.0999882817,-0.1619314998,-0.4995589554,0.7212763429,-0.1596976817,-0.0727414787,-0.1729461849,0.4202396274,-0.1257977635,-0.120153293,-0.0807232261,-0.1388192326,-0.0810431316,-0.1081053019,0.3586574197,0.1451632082,0.063832514,0.2073590308,0.1183850467,-0.27884233,-0.4896351695,0.2174432278,0.2856251895,-0.2475684434,0.012894379,0.2346803546,-0.2322447449,0.1432935297,0.6515596509,0.2464326769,0.0117482534,-0.0120477537,0.2025607526,0.2649378479,0.0501916856,-0.1223700643,0.1640230119,0.2685831189,-0.4571163654,0.2159979194,0.0437234603,-0.1576764286,0.2495294064,-0.233338654,0.2681138515,-0.0029878228,0.5281680822,-0.0881277919,0.289267838,-0.090634279,-0.1573887467,-0.6749100089,0.1098422408,-0.1533483565,-0.1440505981,0.0397293754,-0.3130022287,0.0267414805,-0.0766397268,0.0271469355,0.5202213526,0.1198398396,-0.2506419718,-0.2004321963,-0.4098311067,0.3709433079,0.1497470587,0.1613408923,-0.072733663,0.4148488343,-0.0166738834,0.0786470994,0.2746058404,-0.0772282779,-0.2901545465,0.236041829,-0.4297748506,0.2515313029,-0.2711088657,-0.0724577755,0.0736432523,0.304646045,-0.0841522217,-0.2813910842,-0.006545485,0.2192519456,-0.2696215212,-0.1493303329,0.0387555622,0.3670339882,0.1418906152,0.0027355996,-0.1245593876,-0.1818845719,-0.1920184046,-0.0750168115,-0.1639817804,-0.1211038902,-0.2894135118,0.4088885486,-0.3357611895,-0.5149744153,-0.3786815107,0.3133416176,0.1088762507,-0.1906796247,-0.1274161488,0.3311877847,-0.1567309946,0.0528101362,0.2602564394,0.4714582264,0.0496161021,0.4661207795,-0.2932118773,-0.19221811,0.6927307844,-0.464907974,0.2622734606,-0.195201695,0.4434483945,0.1342340559,-0.3789175749,-0.3714771867,0.0214508418,0.0577610917,0.1420640796,-0.0396045744,0.1356401592,-0.3098528087,-0.1109305099,-0.0667539835,0.1820854694,0.1642801464,0.1262947172,-0.1123879999,-0.0309414659]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"OK, this issue is not about caching but some internal conflict\/race condition it seems, I have just run into it on my normal env:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 356, in _finalize\r\n self.data = Dataset(**reader.read_files([{\"filename\": f} for f in file_paths]))\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/home\/stas\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 655, in \r\n main()\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 619, in main\r\n test_results = trainer.predict(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-master\/src\/transformers\/trainer_seq2seq.py\", line 121, in predict\r\n return super().predict(test_dataset, ignore_keys=ignore_keys, metric_key_prefix=metric_key_prefix)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-master\/src\/transformers\/trainer.py\", line 1706, in predict\r\n output = self.prediction_loop(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-master\/src\/transformers\/trainer.py\", line 1813, in prediction_loop\r\n metrics = self.compute_metrics(EvalPrediction(predictions=preds, label_ids=label_ids))\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 556, in compute_metrics\r\n result = metric.compute(predictions=decoded_preds, references=decoded_labels)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 388, in compute\r\n self._finalize()\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 358, in _finalize\r\n raise ValueError(\r\nValueError: Error in finalize: another metric instance is already using the local cache file. Please specify an experiment_id to avoid colision between distributed metric instances.\r\n```\r\n\r\nI'm just running `run_seq2seq.py` under DeepSpeed:\r\n\r\n```\r\nexport BS=16; rm -r output_dir; PYTHONPATH=src USE_TF=0 CUDA_VISIBLE_DEVICES=0,1 deepspeed --num_gpus=2 examples\/seq2seq\/run_seq2seq.py --model_name_or_path t5-small --output_dir output_dir --adam_eps 1e-06 --do_eval --do_train --do_predict --evaluation_strategy=steps --label_smoothing 0.1 --learning_rate 3e-5 --logging_first_step --logging_steps 1000 --max_source_length 128 --max_target_length 128 --num_train_epochs 1 --overwrite_output_dir --per_device_eval_batch_size $BS --per_device_train_batch_size $BS --predict_with_generate --eval_steps 25000 --sortish_sampler --task translation_en_to_ro --val_max_target_length 128 --warmup_steps 500 --max_train_samples 100 --max_val_samples 100 --max_test_samples 100 --dataset_name wmt16 --dataset_config ro-en --source_prefix \"translate English to Romanian: \" --deepspeed examples\/tests\/deepspeed\/ds_config.json\r\n```\r\n\r\nIt finished the evaluation OK and crashed on the prediction part of the Trainer. But the eval \/ predict parts no longer run under Deepspeed, it's just plain ddp.\r\n\r\nIs this some kind of race condition? It happens intermittently - there is nothing else running at the same time.\r\n\r\nBut if 2 independent instances of the same script were to run at the same time it's clear to see that this problem would happen. Perhaps it'd help to create a unique hash which is shared between all processes in the group and use that as the default experiment id?\r\n","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":409,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nOK, this issue is not about caching but some internal conflict\/race condition it seems, I have just run into it on my normal env:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 356, in _finalize\r\n self.data = Dataset(**reader.read_files([{\"filename\": f} for f in file_paths]))\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/arrow_reader.py\", line 236, in read_files\r\n pa_table = self._read_files(files, in_memory=in_memory)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/arrow_reader.py\", line 171, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/arrow_reader.py\", line 302, in _get_dataset_from_filename\r\n pa_table = ArrowReader.read_table(filename, in_memory=in_memory)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/arrow_reader.py\", line 322, in read_table\r\n stream = stream_from(filename)\r\n File \"pyarrow\/io.pxi\", line 782, in pyarrow.lib.memory_map\r\n File \"pyarrow\/io.pxi\", line 743, in pyarrow.lib.MemoryMappedFile._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 97, in pyarrow.lib.check_status\r\nFileNotFoundError: [Errno 2] Failed to open local file '\/home\/stas\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow'. Detail: [errno 2] No such file or directory\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 655, in \r\n main()\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 619, in main\r\n test_results = trainer.predict(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-master\/src\/transformers\/trainer_seq2seq.py\", line 121, in predict\r\n return super().predict(test_dataset, ignore_keys=ignore_keys, metric_key_prefix=metric_key_prefix)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-master\/src\/transformers\/trainer.py\", line 1706, in predict\r\n output = self.prediction_loop(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-master\/src\/transformers\/trainer.py\", line 1813, in prediction_loop\r\n metrics = self.compute_metrics(EvalPrediction(predictions=preds, label_ids=label_ids))\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 556, in compute_metrics\r\n result = metric.compute(predictions=decoded_preds, references=decoded_labels)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 388, in compute\r\n self._finalize()\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 358, in _finalize\r\n raise ValueError(\r\nValueError: Error in finalize: another metric instance is already using the local cache file. Please specify an experiment_id to avoid colision between distributed metric instances.\r\n```\r\n\r\nI'm just running `run_seq2seq.py` under DeepSpeed:\r\n\r\n```\r\nexport BS=16; rm -r output_dir; PYTHONPATH=src USE_TF=0 CUDA_VISIBLE_DEVICES=0,1 deepspeed --num_gpus=2 examples\/seq2seq\/run_seq2seq.py --model_name_or_path t5-small --output_dir output_dir --adam_eps 1e-06 --do_eval --do_train --do_predict --evaluation_strategy=steps --label_smoothing 0.1 --learning_rate 3e-5 --logging_first_step --logging_steps 1000 --max_source_length 128 --max_target_length 128 --num_train_epochs 1 --overwrite_output_dir --per_device_eval_batch_size $BS --per_device_train_batch_size $BS --predict_with_generate --eval_steps 25000 --sortish_sampler --task translation_en_to_ro --val_max_target_length 128 --warmup_steps 500 --max_train_samples 100 --max_val_samples 100 --max_test_samples 100 --dataset_name wmt16 --dataset_config ro-en --source_prefix \"translate English to Romanian: \" --deepspeed examples\/tests\/deepspeed\/ds_config.json\r\n```\r\n\r\nIt finished the evaluation OK and crashed on the prediction part of the Trainer. But the eval \/ predict parts no longer run under Deepspeed, it's just plain ddp.\r\n\r\nIs this some kind of race condition? It happens intermittently - there is nothing else running at the same time.\r\n\r\nBut if 2 independent instances of the same script were to run at the same time it's clear to see that this problem would happen. Perhaps it'd help to create a unique hash which is shared between all processes in the group and use that as the default experiment id?\r\n","embeddings":[-0.0576908737,-0.0025045027,0.1063458398,0.3295958638,-0.0730611011,0.0291781742,0.1474635452,0.2418277711,0.3056671023,0.159903422,0.2047879845,0.0719605088,-0.3148393333,-0.1378425062,-0.0102979355,0.2024058104,0.0270134211,0.0037926601,-0.20881854,-0.1798234433,-0.3237953186,0.5144024491,-0.0927538127,-0.030579187,-0.3576315939,0.3367961645,-0.1782272458,0.3024886847,0.0206062924,-0.6091864109,0.4519519508,0.1594463885,0.1245748922,0.4872964323,-0.0001303917,-0.1621840298,0.2800582647,-0.1859594285,-0.1424783468,-0.2011274397,-0.0963112786,-0.2803775966,0.4510991871,-0.2343733907,-0.023727281,-0.0773367509,0.1758160144,-0.5915822983,0.3542677164,0.113642998,-0.0048828861,0.3083095551,-0.1144847423,-0.2425568104,0.1217077598,-0.0425462052,0.0690073967,0.2845221162,-0.0860336721,-0.1388474703,-0.2688547969,0.2733645141,0.1988227069,-0.0540879108,0.435208559,0.0169216935,0.0313846096,-0.0497766659,-0.1019930467,0.3329291642,0.1740120351,-0.193828553,-0.208735168,0.0363131464,-0.1447734088,-0.4275946319,0.2937574983,-0.0938262194,0.0557499453,0.1103091612,0.0745091364,-0.0691117272,0.0193976741,0.1733286232,-0.1224157661,0.1237840131,-0.0248199441,0.0248816963,-0.0011232562,0.2460562736,-0.1643101573,0.1366795152,-0.0891157091,0.2079615146,-0.3737040758,0.1329750717,0.1770112813,-0.0221957676,-0.0012672452,0.2791175544,-0.0404296406,0.0242971219,0.3617475927,0.1396155655,-0.011977314,0.4315187633,0.2799905539,0.1480378658,0.2010217458,0.2454798222,-0.0930142477,-0.1489395499,0.1177478805,-0.2112675607,0.3843245804,0.1562823653,-0.0006581515,-0.4661150575,0.068159543,0.3652066886,-0.2267410159,-0.1440903693,0.2043337226,0.3926241994,0.196039781,-0.0364983268,0.1631375998,0.282530725,-0.2508915663,-0.0268494915,-0.442496717,-0.2588551641,-0.1373984665,0.2259525806,0.1951779574,-0.032690838,0.5749140382,-0.2132206708,0.3801062107,-0.0748561546,0.2046850771,0.1003378332,0.0419853739,0.2588788569,-0.0775679201,0.2467142791,0.4501030147,0.0853139982,-0.2225542367,-0.1783111691,-0.1147635654,-0.5404675603,0.1880502105,-0.0822353363,-0.493271172,0.2703615427,0.4431524873,0.1586527675,0.0580164418,-0.3409621119,0.2449738532,-0.0816504136,-0.2438091189,-0.0823501498,0.4338897169,0.7405845523,-0.3018999994,-0.3919824064,0.0709851831,0.1103327125,-0.0848064274,0.2470318675,-0.1676895916,-0.05899271,-0.3711156845,0.0297606532,0.3290179968,-0.4905431569,-0.647685349,-0.1112071723,-0.363750726,-0.0950264633,0.2161112279,0.1401781291,0.0239364151,-0.0563677289,0.1230901927,-0.1741127372,0.2157103568,-0.3362263739,-0.254781276,-0.3030847609,0.092170231,0.011652329,0.3164173663,-0.0692317486,-0.0339647867,-0.1159876212,0.1047608778,0.1246332675,0.1933831722,0.3314185739,0.2837236226,-0.0269378442,0.3410241306,0.2112776935,-0.3488750458,0.3641500175,-0.5019794703,0.0062016589,0.1006616876,0.0107668405,-0.079102844,-0.2361928672,-0.1411092877,-0.4627401531,-0.098077327,-0.1100679561,0.0589427575,0.1902162433,-0.1977372766,0.1868091077,-0.0470207445,0.2514115572,-0.4648362696,0.0201665591,-0.081868954,-0.2810691297,-0.0736851767,0.1205331162,0.0165863577,0.0521006137,-0.2455527186,0.3516991138,-0.0886233225,0.2942867577,0.2905912101,0.2052253783,-0.0483786166,-0.4059664607,-0.0110340612,0.196951732,-0.1073714644,0.02527209,-0.3456165791,0.4294168055,0.0680319965,0.1740290523,-0.3788430691,0.1567651033,-0.1176290885,-0.1660424769,-0.428858906,-0.1451912671,0.543369174,-0.1508431882,0.3000460565,-0.0514612868,0.0512116924,-0.0064014466,0.2151771635,0.2164016664,0.1246335283,0.0957661048,-0.2813659906,-0.1090495512,0.0874325931,0.0135735022,0.5847152472,0.1977362931,0.2524625659,0.1471095085,0.0353333242,-0.1976366639,0.1717872024,-0.0318980329,-0.0335673727,0.2177739143,0.2405479699,0.159772262,-0.2981685698,-0.1236059293,-0.2751501501,0.066438131,-0.3651408255,0.1122677848,-0.2560161054,0.3395058513,0.1566374749,0.0014964637,-0.2696926296,-0.2396304458,0.1786824465,-0.1410636008,0.0234252252,0.2518822551,-0.4109794497,0.4310765862,0.0332370251,-0.2168571353,-0.2495376766,-0.3102408051,-0.190364629,-0.2834860086,-0.1112968624,0.0557817593,0.0091719152,-0.0543887764,-0.0524966531,0.0212187432,-0.3649503291,-0.0103318281,0.1506583542,0.6111653447,0.2543150187,0.0139052402,-0.3683998287,0.0578797571,0.3914343119,-0.2630782425,-0.3664728403,0.0817095339,0.0431456044,0.2250199169,-0.259280026,-0.3096568286,0.0280254688,-0.3285584748,0.4363337457,-0.0327876955,0.0186580662,-0.0697610676,-0.0786047876,0.0785873532,-0.243889153,0.1752252281,-0.372603327,-0.5516355634,0.395483464,-0.011283529,-0.1737852246,-0.1263961792,0.0514004007,0.3583467603,0.0741554871,-0.7520511746,-0.4106815159,0.0507171415,0.2256350815,-0.1302632689,-0.069943659,0.6061971188,-0.0339118429,0.0616026074,-0.0909521878,-0.1484256685,0.6115368605,-0.1508138478,0.1900129318,-0.096024029,0.4599855542,0.2075083852,0.9944367409,0.5102089643,0.1887706965,0.2738809586,0.2238533497,0.4091358185,-0.0956359506,-0.1983337104,0.2562207878,0.2452370375,-0.0865911767,0.2047558278,0.0359231941,0.1220131293,-0.0613914505,-0.0637567267,-0.2538520098,-0.3101133704,-0.0132152634,-0.1616043895,0.3300979733,0.0614185669,0.0212093852,-0.2757084072,-0.0623924881,0.3666175306,0.4859733582,0.1757149547,0.2632332146,-0.2251618356,0.004415981,-0.4434452057,0.2522543669,-0.3854750395,0.2858226895,-0.2330488414,-0.1856447756,0.0981952772,-0.0833995491,0.6718565822,0.1430876404,-0.0848102942,-0.0781539381,-0.0773273781,-0.3918814361,-0.1582129896,0.0495565832,0.0107552912,0.2382263541,0.434158802,-0.3735689819,-0.2426632345,-0.0838242546,-0.1122098863,-0.2599928379,-0.4525399208,-0.2815349996,0.1446144134,-0.1928810477,0.1833460629,-0.2113405019,-0.0374569669,0.0612428635,-0.0035395171,-0.2129247934,-0.1019909829,0.0182372965,-0.0392021723,0.3360673785,-0.4024116099,0.2283705026,0.5376234651,-0.1064085811,0.037613593,0.3673107624,0.1663352698,-0.1827934384,0.268301338,-0.0340071879,0.2161348164,0.6718818545,-0.1646300852,0.1388864219,0.1991302371,0.2080640197,-0.1390986592,-0.0884974673,0.3013365865,-0.1133786142,0.2461322993,0.154859826,0.236022234,0.2951615453,-0.1596689224,0.4172230959,-0.0116957445,-0.1624189615,0.2946782708,0.0393434316,1.0834728479,0.2922216356,0.3179513216,0.1359237134,-0.0182772838,0.3813758194,0.0950419307,0.0987548232,-0.2544327378,-0.3131214082,-0.0655984804,-0.2473014593,-0.0368871763,-0.2755259275,-0.1136833429,0.4412731528,-0.2594443262,-0.0031390667,-0.1601794362,-0.2166808397,-0.2920574546,-0.1391657293,0.1713820845,-0.0610483922,0.0194266941,0.5495744944,-0.1914759129,-0.1742052734,-0.4280658662,-0.2743673027,-0.1341761947,-0.0437697954,-0.3516485095,0.0898426846,0.2500086725,-0.3409355879,-0.105503127,0.3168901503,0.4565448463,0.1514362842,-0.3181406558,0.1140336618,0.2022183239,0.1065969393,0.09550675,-0.1106804386,0.1808836162,-0.0044128918,-0.0609434731,0.0249476992,-0.2868688107,0.0445359461,0.008089846,-0.1284440011,-0.0748335868,0.0710110664,-0.1148347259,0.0865437835,0.0567273609,-0.2192517221,-0.0001807713,0.3327897787,-0.1402470469,-0.054850027,0.0633069947,-0.3442890048,-0.1063787565,0.7176709771,-0.0840655416,-0.1729815304,0.6388201714,-0.01084443,-0.0114887943,-0.0330064222,0.0686272159,0.4259324074,-0.8050131798,-0.0211885013,-0.0666623563,-0.0668366477,0.1689558923,0.0528477691,0.4529182315,-0.1656748056,-0.0827274844,-0.2919704914,-0.4071454704,0.205439344,-0.4431143105,0.2485615611,-0.3803116381,-0.0345143825,-0.3471587002,0.1759299189,-0.1712403744,0.2049064934,-0.1429181695,-0.1194588616,-0.1073107272,-0.0242529828,0.2813020647,-0.107729584,-0.159074679,0.0560277663,-0.4286457598,-0.0027018965,-0.2546121776,0.21965231,-0.080688253,0.1286475211,-0.0928509161,0.0388123877,0.0574886315,-0.1133800596,-0.0858350173,0.168055892,0.0827878043,0.031674087,0.0651550442,0.2205137461,-0.1086527631,0.3121115267,0.0699838847,-0.2269551754,-0.0963369831,0.0597958341,-0.3726491332,-0.0195322763,0.0980868861,0.0736281872,-0.2188918442,0.1540913135,0.112855576,0.0042940029,-0.3112641275,0.0228297971,0.339735955,0.060158994,-0.1754278839,-0.045322679,0.0236441065,0.0185236353,-0.0767244101,0.2651647329,0.1048920602,0.407500416,0.0353325568,0.2018357664,0.2429069728,-0.0319002382,0.138836354,0.2965883315,0.0931647271,-0.2163159251,0.3850467205,-0.1120211482,0.1703716964,0.3353476226,0.5294114351,0.1533012986,0.3033120036,0.0558894128,-0.0029089297,0.1036750302,0.0792399347,0.2830422521,-0.275130868,-0.2732942104,-0.2105268091,0.1802259833,0.0099641662,-0.1593455672,-0.2688842118,0.8103463054,-0.2867060006,-0.1412347406,-0.1363691092,0.1648873836,-0.1185229272,-0.1748317033,-0.1090282723,-0.1780403852,-0.0861646011,-0.0919659063,0.1850930154,0.149837181,-0.0308984891,0.2067728937,-0.0318866968,-0.2017472684,-0.3404049873,0.3621706367,0.162451759,-0.1593182236,0.0908347368,0.2921066582,-0.3177046478,0.1487344503,0.5308632255,0.2546340525,0.0195457283,-0.2139033228,0.0672453269,0.4024716616,0.0209689457,-0.1070486382,0.1095999405,0.2688015401,-0.4691239893,-0.0098234145,-0.0169435646,-0.0519172065,0.3932632804,-0.1417685151,0.2539665997,-0.162134409,0.3582155108,-0.1994121671,0.1481004655,0.1612881571,-0.0726243779,-0.6035972238,0.2233772278,-0.0262330547,0.0818551406,0.1133687794,-0.1424368024,-0.0047132666,-0.0867540389,0.1382466406,0.5117942691,-0.0139972493,-0.4154541194,-0.2745571136,-0.6609945893,0.388194412,0.0333934464,0.3113510907,0.000975939,0.2808103859,-0.0223052874,-0.0369713083,0.2481424809,0.0789000988,-0.0846418366,0.1082658321,-0.4920291305,0.098280333,0.0012777366,-0.1708384603,0.034142606,0.2196375281,0.0084805703,-0.1435922831,-0.0773816109,0.2027598768,-0.1723472625,-0.0357827395,0.2105973363,0.4661997259,0.0523876101,0.0963559821,0.0934226811,-0.1339044124,-0.2228318155,-0.0566935353,-0.0889864489,-0.0846365318,-0.3063417673,0.4797859788,-0.5571389794,-0.5037087798,-0.3182576895,0.2528630495,0.0200796332,-0.26037395,-0.2814155519,0.3865290582,-0.1927475184,0.1288113892,0.2811796069,0.3434962034,0.1242486462,0.3990010023,-0.4151052833,-0.1510776579,0.7318447232,-0.4763924479,0.3372202516,-0.1236952543,0.362672478,0.4212175012,-0.310390085,-0.2815282047,-0.0851890445,0.1817666441,-0.0512084514,-0.1511390507,0.0933481902,-0.2861341834,0.0352644809,-0.0865418315,0.1908096522,0.2411007285,-0.1176284477,0.0448560528,-0.0937296227]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"When you're using metrics in a distributed setup, there are two cases:\r\n1. you're doing two completely different experiments (two evaluations) and the 2 metrics jobs have nothing to do with each other\r\n2. you're doing one experiment (one evaluation) but use multiple processes to feed the data to the metric.\r\n\r\nIn case 1. you just need to provide two different `experiment_id` so that the metrics don't collide.\r\nIn case 2. they must have the same experiment_id (or use the default one), but in this case you also need to provide the `num_processes` and `process_id`\r\n\r\nIf understand correctly you're in situation 2.\r\n\r\nIf so, you make sure that you instantiate the metrics with both the right `num_processes` and `process_id` parameters ?\r\n\r\nIf they're not set, then the cache files of the two metrics collide it can cause issues. For example if one metric finishes before the other, then the cache file is deleted and the other metric gets a FileNotFoundError\r\nThere's more information in the [documentation](https:\/\/huggingface.co\/docs\/datasets\/loading_metrics.html#distributed-setups) if you want\r\n\r\nHope that helps !","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":173,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nWhen you're using metrics in a distributed setup, there are two cases:\r\n1. you're doing two completely different experiments (two evaluations) and the 2 metrics jobs have nothing to do with each other\r\n2. you're doing one experiment (one evaluation) but use multiple processes to feed the data to the metric.\r\n\r\nIn case 1. you just need to provide two different `experiment_id` so that the metrics don't collide.\r\nIn case 2. they must have the same experiment_id (or use the default one), but in this case you also need to provide the `num_processes` and `process_id`\r\n\r\nIf understand correctly you're in situation 2.\r\n\r\nIf so, you make sure that you instantiate the metrics with both the right `num_processes` and `process_id` parameters ?\r\n\r\nIf they're not set, then the cache files of the two metrics collide it can cause issues. For example if one metric finishes before the other, then the cache file is deleted and the other metric gets a FileNotFoundError\r\nThere's more information in the [documentation](https:\/\/huggingface.co\/docs\/datasets\/loading_metrics.html#distributed-setups) if you want\r\n\r\nHope that helps !","embeddings":[-0.0598983429,0.0068152775,0.084484525,0.1580613703,0.0199416559,-0.0320760161,0.2774336934,0.2207652032,0.3088784218,0.1602413505,0.0177518763,0.1031419486,-0.2257237434,-0.0206120964,-0.0592386834,-0.0310602225,-0.0435301624,-0.0667761266,-0.3691173196,-0.1325116158,-0.3067215085,0.501986146,-0.0801183656,0.0445248485,-0.4807972014,0.1278136522,-0.1756111234,0.2225741297,-0.0789252147,-0.558731854,0.3777546883,0.1831186414,0.0421271063,0.6000635624,-0.0001176384,-0.1433923841,0.1927302778,-0.124524422,-0.083281219,-0.1930171698,0.0741741657,-0.3329316378,0.4511277974,-0.2442206591,0.0344935618,-0.051198341,0.0247724187,-0.3721569777,0.4274136722,-0.0064346162,0.1307448,0.3117668629,-0.068226777,-0.3215267956,-0.0999848619,-0.1143518686,0.0776162222,0.4171811044,-0.1312239319,0.0894623995,-0.2742326856,0.1245770603,0.212353453,0.0292280708,0.2388039976,0.024607148,-0.0142720928,-0.1256713569,-0.0685603768,0.2887807488,0.1676787883,-0.1594269574,-0.2014690191,0.0085315043,-0.2289424837,-0.4411575794,0.1714061648,-0.124580346,0.0746701732,0.103214927,-0.0762275234,-0.154982388,-0.0232046694,0.1175038964,-0.1456239074,0.0324369818,-0.1697130799,0.1212585792,-0.027921401,0.1497502178,-0.3855251968,0.3516908884,-0.0194257125,0.0714746565,-0.3793978989,0.136480391,0.2671203017,0.0217041448,0.1065052971,0.4761018157,-0.0047422666,-0.0132049061,0.3891662359,0.1737164408,-0.0644987375,0.5233271122,0.2027726769,0.0692863762,0.2185038477,0.2179413289,-0.1748858243,-0.3413075507,0.2836264968,-0.3860961497,0.3315577805,0.1842790097,0.0930345803,-0.4868882298,0.0585246794,0.1844706088,-0.0363664255,-0.1577919573,0.0875825584,0.3153523207,0.1501623392,0.0135483872,0.2400891483,0.3981119096,-0.2120680511,0.0780819207,-0.3880885541,-0.1634971648,-0.1441083103,0.4293542206,0.0517114885,-0.1105698496,0.6856907606,-0.1622716486,0.4642171562,-0.061395552,0.1907998472,0.1755948961,0.0188385695,0.2674921453,0.0719169155,0.175892666,0.2669182122,0.10486193,-0.2120070457,-0.3701926768,-0.188737154,-0.5549445748,0.1263828725,0.0321527384,-0.4145857096,0.2830307484,0.3431289494,0.2120162845,-0.0794245973,-0.2337050438,0.1338794082,0.0511640869,-0.0858735442,-0.1279808432,0.4915291369,0.7377231717,-0.2220460027,-0.3446711898,0.0168370027,-0.0487126447,-0.0864723176,0.3050107956,-0.0926363841,0.2284962535,-0.2772685885,0.1222094968,0.3515832722,-0.4234384298,-0.4113765061,0.0080032414,-0.4272915125,-0.1334397495,0.1083738208,-0.0730173066,0.2581312656,0.1308469474,0.0548113696,-0.1267488599,0.1568037421,-0.283675164,-0.3865198493,-0.1858029962,-0.0045480025,-0.1014779657,0.300691992,0.0563218743,0.0986357033,0.0875678733,0.0597034022,0.1062118262,0.26060763,0.2477435917,0.2262085676,0.0003543365,0.289907515,0.1102003902,-0.1867316067,0.2562479675,-0.3439131379,0.1210736707,0.1683987677,-0.0155254724,-0.200071454,-0.2549752593,-0.1508955508,-0.4215324521,0.0472918451,-0.0709691867,0.3166153729,0.1520330608,-0.1438020319,0.0916495696,0.1583401114,0.1830408126,-0.4161327779,0.0770388246,-0.0725939348,-0.3368439674,-0.0147532811,0.0729850233,0.0808983073,0.0420653485,-0.1817387789,0.4776448011,-0.2574470937,0.2789967656,0.383679986,0.1822809577,-0.0755051225,-0.3026510477,0.0588997863,0.0483248942,-0.0166241489,0.001170759,-0.1728834212,0.3591372073,0.0592466183,0.095292896,-0.3097137511,0.3031127155,-0.046105627,-0.1465944946,-0.4951242507,-0.1190972924,0.3968292475,-0.3081524968,0.2047468275,-0.1520290971,0.0571071133,0.0319924727,0.1416758448,0.1343099326,0.126549229,-0.0061651319,-0.161784932,-0.0100206947,-0.0072231968,-0.1396538019,0.5526068211,0.1933413148,0.2127412409,0.1210232303,0.0394823477,-0.203192085,0.0009544448,0.0007292124,-0.1463176459,0.1554473341,0.2229654342,0.133715868,-0.2336518914,-0.0772265121,-0.3163039982,0.0065659764,-0.3755759597,0.0749985501,-0.1550010145,0.4108365178,0.0599399842,0.0543511435,-0.2262229621,-0.2606708407,0.2092506886,-0.1175970435,-0.0695596486,0.3137815297,-0.183540374,0.5583069324,0.0173088554,-0.0202201214,-0.2246251702,-0.2112009376,-0.0854687914,-0.1483573616,-0.1897707731,0.2323055267,0.13009803,-0.166483596,0.0585788451,-0.0695105493,-0.3264140785,0.0377147757,0.0303655788,0.6761788726,0.2347936928,0.0001523599,-0.251847893,0.028367741,0.4868867397,-0.4140199721,-0.2675405741,0.0591720492,0.0195234343,0.0564181395,-0.2959034443,-0.3545251191,0.070414193,-0.3689333797,0.564774394,0.0072399732,-0.0687930733,-0.2838418186,-0.0883293897,0.1034515724,-0.2374043167,0.2703514099,-0.3315576911,-0.7032841444,0.3768460155,-0.3030439019,-0.2822323442,-0.0451202691,0.0785690397,0.4289895594,-0.0625871941,-0.5393745899,-0.4895910025,0.0535351411,0.1895357817,-0.0105361529,0.1493699104,0.5207619667,-0.1010604426,-0.0189073421,-0.0783482045,-0.2318599075,0.6520729065,-0.0872166455,0.1458659172,-0.1006416231,0.416723609,0.0971859023,0.8765502572,0.5474234819,0.2963705659,0.2949184179,0.121151261,0.3381679952,-0.0262518153,-0.1907776147,0.4884671271,0.2340406775,-0.144677192,0.2166243494,0.1231645942,0.1313864291,-0.1734934598,-0.0926730707,-0.3810849488,-0.327729255,-0.1055307388,-0.117643334,0.2585914135,-0.0223617591,-0.0434567444,-0.2274356931,0.0631563365,0.4233258963,0.3457271159,0.200604111,0.2726470828,-0.1781516522,0.019640632,-0.5355622768,0.3346014917,-0.4541780949,0.0489891209,-0.2220091522,-0.2432731688,0.1035628468,0.0479661897,0.5427538753,0.0469464026,-0.1899243295,-0.0557613522,-0.2372996062,-0.4443702996,-0.1240651608,-0.047235813,-0.0626907721,0.2470576018,0.3423611224,-0.5057767034,-0.1438983232,-0.1255820841,-0.1004821509,-0.3064891994,-0.3933877945,-0.3682079017,-0.0740963742,-0.2561747432,0.1216928512,-0.2380999327,0.0357332714,-0.0201439355,0.0438737124,-0.1827234328,-0.0074398722,-0.0261025839,0.190134272,0.3149793446,-0.3515060842,0.2579291165,0.5438994169,-0.0235520434,0.0557703972,0.4056897759,0.0491004102,-0.0081697134,0.2634596825,-0.0709712133,0.287120074,0.5680272579,-0.2031901181,0.0198505148,-0.0444087982,0.4116784334,-0.1413420439,-0.1564923376,0.2634869218,-0.1303112656,0.3417813182,0.2045196295,0.0988926739,0.2688887417,-0.1753828228,0.2153662443,-0.1017574221,-0.1341546923,0.3826420307,0.0164262485,0.9436200857,0.2054094672,0.0592285171,0.1934207827,0.1771509051,0.3641620576,0.0876191854,0.0328884423,-0.2055558115,-0.2237051725,0.0278852507,-0.1202304885,-0.0501211882,-0.2236301899,-0.25450477,0.4385682046,-0.267675519,-0.2376937717,-0.3451876342,-0.1525627822,-0.2159319371,-0.2082403898,0.0347567834,0.0756383762,0.0542347655,0.534883976,-0.2138372362,-0.1893496215,-0.3338544071,-0.3082471788,-0.0714996681,-0.0873035938,-0.106776692,0.006939027,0.3344351351,-0.0596463457,-0.2383830696,0.2095327228,0.4598333836,0.0625675842,-0.2789725959,-0.0697868094,0.0656185299,0.0188155267,0.0810708553,-0.0724339262,0.240883559,-0.1024614871,-0.1515270621,0.0943483263,-0.1460154504,0.0819114149,0.0776296332,-0.1399116218,-0.1492311656,-0.089636147,0.0508769527,0.1414059699,0.0980152041,-0.2040065825,0.1001997516,0.2666704655,-0.192235589,0.1376900822,0.1894948483,-0.2131112218,-0.1155994087,0.5188007355,-0.2788771987,-0.0439733788,0.3037275076,-0.0226094611,-0.0551314391,-0.0628999472,0.0001369011,0.464610517,-0.8037570119,0.0456759967,0.0220151749,0.0338212363,0.0662963614,0.2331276238,0.43972826,-0.1197047159,-0.059818916,-0.2463515848,-0.3375791311,0.3076078296,-0.3285470605,0.1088334471,-0.354359746,0.1391602308,-0.1507173628,0.352865845,-0.3035137057,0.0882375389,-0.2990910411,-0.0992693752,-0.0445940308,-0.0650521293,0.2502366602,0.0224681776,-0.0775944665,0.0911414176,-0.3674952388,-0.1012228057,-0.2215259969,0.140322566,-0.0527469516,0.0371609926,-0.0516503677,0.1735046804,-0.0011846134,-0.1697573662,0.0934676453,0.2170441002,-0.0037538293,0.0508844182,0.1405664831,0.0473802388,-0.113699913,0.2166002095,0.1527908295,-0.2643914819,-0.1714894176,0.0890158266,-0.3603459299,-0.0956997573,0.2447437197,0.0772195831,-0.1432616264,0.20820041,0.047675617,0.0165245347,-0.2506276965,0.009715328,0.4669169188,0.1286633313,-0.139187634,0.0958867297,0.0040334836,0.1172938198,-0.0484067351,0.3132454753,0.0363499783,0.3366869688,0.0685615763,0.2757846117,0.2371047586,-0.1068969369,0.1015352383,0.2575312257,0.0356713459,-0.1651239395,0.228510052,-0.1312101781,0.0824568719,0.3728130162,0.6446869373,0.1242340356,0.2143917084,-0.1020230129,-0.0754533708,0.2085901797,-0.1400790513,0.3282520473,-0.1582538784,-0.1204887629,-0.1251689494,-0.0651015937,0.0011574796,-0.2720160186,-0.3687119186,0.6771073937,-0.3241759837,-0.3469700515,-0.1790501773,0.3376560807,-0.1456151605,-0.1069582701,-0.2152796984,-0.1884310842,-0.1496008635,-0.1087996215,0.2406614572,0.0445790328,0.2287884057,0.0562940873,0.0164168719,-0.2694598436,-0.4721544385,0.2357537597,0.1594562232,-0.1011680663,0.0534244999,0.2343874276,-0.3595524728,0.1360555291,0.7672570348,0.2582760155,0.075595662,-0.0845426545,0.0233774893,0.2670136988,0.0479377843,-0.1087179333,0.1379797012,0.1853430867,-0.4984890819,0.0855320916,0.106067799,-0.172519356,0.3995271921,-0.3463371992,0.330455035,-0.193364948,0.4167951941,-0.1988388151,0.2309960425,0.029863432,-0.2229295969,-0.7316349745,0.3039991558,-0.1173734739,-0.0747346058,0.116549477,-0.3806368709,0.0477004051,-0.0271401554,0.189659372,0.3733307123,0.0494939461,-0.3342296183,-0.2568193972,-0.5867595673,0.3230748475,0.1014959216,0.3235936165,-0.089816168,0.3570632637,-0.0974761844,0.1376839131,0.1949172467,0.1135395244,-0.1029613167,0.0712881982,-0.4963091314,0.1593194902,-0.1107420921,-0.1466667652,0.1599315703,0.2845371366,-0.0483200215,-0.2016771287,0.0270367414,0.267298758,-0.1220200211,-0.1107842699,0.1535866559,0.3878780901,0.2212015837,0.1139356419,0.0036389637,0.0192444883,-0.2425035685,-0.1357331574,-0.1199793294,-0.1062790528,-0.2875246406,0.3848083317,-0.3711648881,-0.3915485144,-0.3861628771,0.2148721218,0.0746119469,-0.1054601744,-0.112957567,0.301743567,-0.0959620699,-0.0575611442,0.2854055166,0.314861834,0.0014858003,0.4434122741,-0.2427063286,-0.1596809328,0.7152579427,-0.4597457945,0.2641885281,-0.2539711297,0.3937128782,0.3925572932,-0.4172923863,-0.3282020092,0.0121974358,0.0605196729,0.0799088031,-0.1610705405,0.0920374617,-0.390629679,-0.1147467345,-0.0314645208,0.1399295628,0.2236700803,-0.0557530932,0.0957488567,-0.0571781881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"Thank you for explaining that in a great way, @lhoestq \r\n\r\nSo the bottom line is that the `transformers` examples are broken since they don't do any of that. At least `run_seq2seq.py` just does `metric = load_metric(metric_name)`\r\n\r\nWhat test would you recommend to reliably reproduce this bug in `examples\/seq2seq\/run_seq2seq.py`?","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":48,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nThank you for explaining that in a great way, @lhoestq \r\n\r\nSo the bottom line is that the `transformers` examples are broken since they don't do any of that. At least `run_seq2seq.py` just does `metric = load_metric(metric_name)`\r\n\r\nWhat test would you recommend to reliably reproduce this bug in `examples\/seq2seq\/run_seq2seq.py`?","embeddings":[0.0750928149,-0.1511457562,0.1568090171,0.1809438467,-0.0645122975,-0.0121280113,0.3743320405,0.2074542642,0.2033585608,0.0637144893,0.232380569,0.1092878655,-0.3682080805,-0.2463029474,0.1191040576,-0.1598070711,0.0618006103,0.0787240714,-0.3983020484,-0.1459854245,-0.2288619876,0.5977004766,-0.2141059488,0.0435038507,-0.5726758242,0.278558135,-0.2246072292,0.1813751459,0.1039820015,-0.510097146,0.4914008677,0.1075301394,-0.0393089764,0.6184602976,-0.0001310607,-0.1159780473,0.2905943692,-0.2518160641,-0.145431608,-0.2269473076,-0.1584118903,-0.2461212128,0.4491089284,-0.2053947747,0.0618131571,0.0542072617,0.1260381341,-0.1677697599,0.4823168218,0.0367282964,0.0278775282,0.4897834957,-0.1625120193,-0.3433550298,0.0686691403,0.145645678,0.0055331509,0.2566850483,-0.0682889149,0.011801051,-0.2210907787,0.2391419262,0.2916741371,-0.0231562499,0.2404492795,0.0032988689,-0.0292709004,-0.0683813989,-0.1375892758,0.3957450092,0.2470199168,-0.1506510675,-0.3314959705,-0.0105179334,-0.3707772791,-0.1938928515,0.3266067505,-0.1777870804,0.0734513476,0.1111949384,-0.0882413089,-0.1193780005,-0.0133474991,0.0675627515,-0.2427281737,0.0775505379,-0.088917084,0.0202900227,-0.0826743543,0.0575080924,-0.2245983928,0.2535657585,-0.0392755792,0.1483736187,-0.3013089895,-0.007587533,0.1568772644,-0.0926359668,0.0377053246,0.3995665312,-0.0197044369,0.1337978989,0.3218588531,0.145534873,-0.0883822143,0.5639249086,0.2473172098,0.1956648976,0.1423614025,0.2767653465,-0.0769011751,-0.2230588645,0.1054571271,-0.1692461222,0.3477960229,0.0648900047,0.0919375047,-0.4264973402,0.1201483607,0.2431017458,-0.0429426841,-0.1472582519,0.1478135586,0.4604143202,0.1981285214,0.0761523247,0.1781300604,0.3603515625,-0.1025502235,-0.0604590587,-0.3435696661,-0.1752094775,-0.2113113105,0.2515896857,0.2330026925,0.1246434376,0.6372714043,-0.2465672493,0.361093998,-0.1116089821,0.306776613,0.285213083,0.1577807218,0.4128853679,-0.1351288855,0.2474038452,0.3334746063,0.0782879218,-0.2614819705,-0.1910322607,-0.108800821,-0.4993456602,0.209586665,-0.0593761206,-0.6320038438,0.2099979818,0.3444989324,0.1252350211,-0.088011682,-0.29063043,0.1899120957,0.0113037731,0.0194324162,0.0363644734,0.5021131635,0.8084053397,-0.1071282178,-0.5354682207,-0.0404285192,0.0006152861,-0.0503945276,0.1723441631,-0.0216381773,0.1558131725,-0.2820476592,0.0270247981,0.3047235012,-0.3449210227,-0.3754405975,-0.0424837582,-0.2123759389,-0.0130924257,0.1726590544,-0.099288635,0.159601137,0.0654157996,0.0390030555,-0.1574855894,0.178449437,-0.1477150023,-0.2838848233,-0.3624432683,0.0209673345,-0.0074382541,0.3962644041,0.1487722546,-0.0420768484,0.2471259087,-0.0140904114,0.1546952873,0.1208314672,0.2546764612,0.1531657875,-0.2394436002,0.3301900029,-0.0040534297,-0.3316984475,0.3137979209,-0.4128647149,-0.0016717316,0.171761021,0.0299317986,-0.3266585171,-0.1333886981,-0.0559235513,-0.4057338536,-0.0640614405,-0.0681678653,0.1189898998,0.0956791639,-0.1376004517,0.0342143402,0.293838948,0.2461948246,-0.4998989403,0.0841354206,0.0144144064,-0.2128111571,-0.2256056517,0.01389293,-0.0110580968,-0.1588858217,-0.2033104748,0.3690865934,-0.0667352453,0.3717402816,0.2778237462,0.3218532503,0.0117869796,-0.4765178561,-0.0597838424,0.2749159336,-0.053749986,-0.0168479159,-0.4565024376,0.3429724574,0.1099388376,0.0574722849,-0.2757979333,0.3847721517,-0.036700666,-0.2347462922,-0.5471624732,-0.1155283675,0.4180024564,-0.2595297992,0.2554793954,-0.106464237,0.1143487394,-0.2071104646,0.1366577446,-0.0392934233,0.2872008383,0.1524625272,-0.3554177284,-0.0619288348,0.033507321,-0.080422543,0.3691939116,0.0703881159,0.2867376208,-0.0314374417,0.0858655795,-0.1533981711,0.168610692,0.1308805943,-0.118894428,0.1236365065,0.2123601288,0.1744197607,-0.283901304,-0.166707918,-0.2365742922,0.064098157,-0.3988458514,0.1490962207,-0.011422853,0.5295215249,0.0435043611,-0.0604522265,-0.2244169116,-0.2906612754,0.1571871042,-0.2209062129,-0.0536589622,0.3262691796,-0.222744599,0.6643281579,0.0525766462,-0.1865573078,-0.4059668183,-0.0892629623,-0.1126825809,-0.1945126504,-0.2627268136,0.0734461546,-0.1226369664,-0.1925716251,0.0805480257,0.1065049395,-0.483492583,0.1850721389,0.2666202784,0.6751850843,0.3146781921,-0.0235575959,-0.137926966,-0.0140281953,0.3056758046,-0.4094474614,-0.2437088788,0.308364898,-0.0299442541,0.1579077989,-0.2975551784,-0.253415823,0.1630838513,-0.2594065666,0.5369822979,0.0357925594,-0.039002385,0.0486771017,-0.0482484996,0.1972010434,-0.2278416455,0.2416341901,-0.0970920995,-0.4955857396,0.4939000905,-0.0899777114,-0.232549876,-0.233001709,-0.259604156,0.2908662558,-0.0499920622,-0.4904942513,-0.2846312523,-0.1554356068,0.1077007949,0.0514273718,0.0468786247,0.5173733234,-0.0740085691,0.0814283565,-0.1966250241,-0.1054342091,0.4829234183,-0.0403454676,0.2072467208,-0.1482871175,0.252377063,0.1457008719,0.9683125019,0.4021627605,0.2108834833,0.1378668994,0.1212735698,0.2691400051,-0.03276572,-0.1612855941,0.1290480494,0.1736795902,-0.2389682233,0.1625221074,0.1619470865,0.0524048693,-0.0170534886,0.0480599105,-0.3372254372,-0.3736346066,0.0258931294,-0.2867444158,0.2674738765,0.0084657408,0.1298524588,-0.2105980366,0.0496194735,0.3940189183,0.3825774193,0.2130386978,0.1041270345,-0.1515610665,0.1164900735,-0.4340966344,0.1125115305,-0.4435677528,0.295886457,-0.1493600309,-0.3472633064,-0.0203620419,-0.0741200298,0.7864434123,0.1294671595,-0.0121841673,-0.0056254594,-0.1805309951,-0.527990818,-0.1744875908,-0.2424480915,-0.0621795431,0.328620553,0.4173190892,-0.3465723097,-0.091442056,-0.0846183673,-0.1157487258,-0.2575946152,-0.1866010576,-0.3307994008,0.0978285372,-0.1791491657,0.1057741493,-0.1725385636,-0.136350885,0.0036878448,0.1023926735,-0.1907083541,-0.0807406157,0.0790346861,0.3291590214,0.3670138717,-0.4262426794,0.2056540549,0.5175892711,-0.3233252764,-0.1374791414,0.3022696376,0.147537306,-0.1070956662,0.4146627784,-0.0502339303,0.1603396386,0.5822029114,-0.2052153945,-0.0157068875,0.0857775584,0.3517509103,-0.1212828681,-0.0598318242,0.2842303813,-0.1013543606,0.2932528853,0.278011024,0.4115163684,0.2014752775,-0.2698430419,0.2848837078,0.0203258526,-0.1558415294,0.5100288391,0.0391432792,1.0379647017,0.1548095345,0.2015151829,0.1303993762,0.1648876369,0.4351435006,0.120360516,0.0437967591,-0.2623807788,-0.3548333943,-0.0181121994,-0.2812707126,-0.0030667644,-0.1445852667,-0.4378490746,0.3533287942,-0.342638135,0.0178091321,-0.0403170884,-0.2422981411,-0.2960453928,-0.1201758832,0.1541963369,-0.04566608,0.1980380267,0.6490924358,-0.2589025199,-0.156810075,-0.3486701846,-0.4128616154,-0.2985284925,0.0445986316,-0.0711567774,-0.0084000789,0.5609113574,-0.0884365216,-0.0502094328,0.264380306,0.4029280245,0.1458511651,-0.4677396417,0.1415465474,-0.059056595,0.163937524,0.176322937,0.0302410778,0.1974542141,0.0159928817,-0.0811609328,0.031744957,-0.2748938203,0.1612436324,0.089596495,-0.1340851039,-0.2504485548,-0.0825361162,0.0357696339,0.1338226646,-0.0820522085,-0.2147385031,0.026969878,0.2467697859,-0.2292547077,0.0371180885,0.1516253352,-0.2628640532,-0.0579269677,0.4928119779,-0.362527132,-0.0921284854,0.3880810738,-0.0145975165,0.0954929218,-0.0928230584,0.1040695608,0.4269271493,-0.9760483503,0.1369957328,-0.3507162631,-0.1869861931,0.0663210675,0.2257801592,0.6225939393,-0.1898290217,-0.13349168,-0.1952226013,-0.3736699224,0.3917945027,-0.1390203834,0.1900529712,-0.5208064318,-0.0400962941,-0.0909901187,0.2423184514,-0.1821935922,0.2231492102,-0.2955841124,-0.0602543876,-0.1686761826,-0.095660165,0.3259902894,-0.1957744211,-0.144848913,-0.0011052121,-0.3109070659,-0.0424052142,-0.3904665709,0.1731740981,-0.0693069547,0.0203189012,-0.1117977649,0.4331068993,0.042607192,-0.2253696173,0.042176351,0.0808046386,0.025740318,0.2968802154,0.1680519879,0.0712303221,-0.2817444503,0.088296704,0.1818272173,-0.1884128153,0.0025738929,0.1377200633,-0.3446248472,-0.0645113215,0.0776904002,0.2567935586,-0.1901199818,0.0443105027,0.1996380538,-0.0354052112,-0.1188756227,-0.055423066,0.3305149674,0.1367082745,-0.1593909413,0.0950238928,-0.0886044055,0.0568297356,-0.0802782476,0.1048179194,0.0006983211,0.321878016,0.085381344,0.3245587349,0.1636347771,-0.1622797251,0.3574334681,0.1928835809,0.193733871,-0.1373789161,0.3820769489,0.0233727954,-0.0269822571,0.4478304982,0.4515419304,-0.0703796744,0.2798916698,-0.1402748376,0.0753537193,0.2244552225,-0.2527284324,0.3090746999,-0.0772216767,-0.1782063246,-0.1137577221,0.0177020915,0.0819642916,0.0149574988,-0.2574413419,0.59964782,-0.36095047,-0.1359236091,-0.305848062,0.0015193509,0.0109866038,-0.2694576681,-0.0101670576,-0.1457814127,-0.2194039375,-0.2321443111,0.2223371267,0.0849191993,0.0927917063,0.086917147,0.0281846244,-0.1800113767,-0.2946361899,0.3112851977,0.2292453796,-0.250225246,0.1192082465,0.2937873602,-0.3188090324,0.0803370327,0.6347109675,0.2948594689,-0.0884070918,-0.0761146843,0.126560092,0.1849356443,-0.0109211495,-0.0974577963,0.2389882058,0.1733166873,-0.3825125992,-0.047026962,0.0349703059,-0.1180279031,0.3365169168,-0.2178979367,0.1758395135,-0.2750437558,0.5571632385,-0.2242746204,0.0447612442,0.1200994849,-0.1157301292,-0.6041264534,0.2573820651,-0.1305813044,-0.0253201108,0.1239655614,-0.0469639935,-0.0171992686,0.0293214563,0.2097137719,0.4323589504,0.232756421,-0.270385623,-0.2339159548,-0.5324943066,0.2602303028,0.1255425662,0.3936369419,0.2322482765,0.2296855748,-0.0150977178,-0.0309845936,0.1089071557,-0.0064848065,-0.1906322688,0.0576795824,-0.4498601556,0.0906724706,-0.0608780645,-0.0421842933,0.0336841457,0.0912768468,-0.0197495371,-0.0872878581,-0.0868772343,0.1656349301,-0.2633648515,-0.0535269976,0.1447160095,0.3367942572,0.161567986,0.0700882077,0.0182316899,-0.1633973271,-0.2470243126,-0.1918805093,0.0415443741,-0.2905655503,-0.3666417897,0.4213339388,-0.292668432,-0.3824874759,-0.3238886297,0.1042756885,0.1808907688,-0.104052946,-0.3937341869,0.316043824,-0.2724983692,0.0433944948,0.2825978994,0.3950480223,-0.1257580519,0.3722699583,-0.2413482815,-0.0346612744,0.6833050847,-0.5341154337,0.1998460591,-0.3363792896,0.3415124416,0.2691192627,-0.339043349,-0.3327305615,-0.1527008563,0.152698949,-0.0082969228,-0.1418135613,0.0664726421,-0.1462513357,-0.0786665455,-0.080192849,0.4183965921,0.168691799,-0.0924171731,-0.027184451,-0.1010505855]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"To give more context, we are just using the metrics for the `comput_metric` function and nothing else. Is there something else we can use that just applies the function to the full arrays of predictions and labels? Because that's all we need, all the gathering has already been done because the datasets Metric multiprocessing relies on file storage and thus does not work in a multi-node distributed setup (whereas the Trainer does).\r\n\r\nOtherwise, we'll have to switch to something else to compute the metrics :-(","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":85,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nTo give more context, we are just using the metrics for the `comput_metric` function and nothing else. Is there something else we can use that just applies the function to the full arrays of predictions and labels? Because that's all we need, all the gathering has already been done because the datasets Metric multiprocessing relies on file storage and thus does not work in a multi-node distributed setup (whereas the Trainer does).\r\n\r\nOtherwise, we'll have to switch to something else to compute the metrics :-(","embeddings":[0.0360684097,0.0379045084,0.0943586305,0.1253585219,0.176023975,0.0507231429,0.1819939315,0.2530153394,0.2025461346,0.251921773,0.0820863023,0.1800011545,-0.3751119971,0.0665903017,0.1416402161,-0.0772908852,0.0167336687,0.0033793277,-0.3133099973,-0.1381399632,-0.2079617828,0.4481794238,-0.1228016689,0.0743648484,-0.5695955753,0.1498474181,-0.1701162308,0.0805468708,-0.1255107522,-0.5812905431,0.3919239938,0.1458094865,0.0017238059,0.4191245139,-0.0001262029,-0.1343241781,0.1936454922,-0.2333316952,-0.1047410443,-0.2225060165,0.1268722415,-0.3101156354,0.5224650502,-0.146082744,-0.0495458841,-0.0838488117,0.0543664955,-0.4214364886,0.4781624377,0.1183948815,0.0403270535,0.3758564293,0.034930747,-0.2492558956,-0.0176795013,-0.0774545372,0.0102454899,0.3299077451,-0.0517302752,-0.0297793802,-0.2794109285,0.2436843216,0.2794153988,-0.0088509116,0.3128017187,-0.0297786128,-0.0145224724,-0.1737946719,-0.1312471628,0.2608571649,0.1949307024,-0.2089758515,-0.3556877673,-0.0132943876,-0.0411426052,-0.5314995646,0.1682944447,-0.1075419858,-0.0336157791,0.0474928319,-0.1316332966,-0.1672554612,-0.0719782338,0.0874651,-0.1760182083,0.0398989767,-0.1421070397,0.0187715013,0.0959144533,0.2716259956,-0.2097330242,0.4107071459,0.1014254019,0.1463738382,-0.3606158197,0.0825564787,0.1721960157,-0.0282386262,0.0502630398,0.3950385451,0.0560898148,0.1017731205,0.375590086,0.0602638312,-0.045978725,0.4311771989,0.2344252318,0.1357567757,0.1160663068,0.2224400491,-0.1367132515,-0.2810336053,0.1833153069,-0.1449779719,0.3832835853,0.2245602608,-0.1369666755,-0.3769397736,0.063728869,0.2464854419,-0.0436970703,-0.2041589022,0.1965016872,0.2750817239,0.1341587901,0.1387406886,0.1868602186,0.3029897213,-0.17224738,-0.1656983942,-0.3989240825,-0.0746448487,-0.2094953656,0.3997785449,0.1894150972,-0.0411997586,0.6616476774,-0.3590566218,0.5086171031,-0.2140405327,0.1602294296,0.145807758,0.097427167,0.2898179591,-0.0214464571,0.0991971567,0.3203160167,0.0583865009,-0.1933315545,-0.3334262073,-0.2335591316,-0.5762785673,0.1160737276,-0.0552396998,-0.5121427774,0.2240834087,0.1648786515,0.3089993,-0.1530394852,-0.1348861158,0.1365174651,-0.0361995623,-0.1778376698,-0.0181775913,0.6248615384,0.5545912981,-0.3065955639,-0.4664386511,0.109280616,-0.0107354699,-0.1018487811,0.1589731723,-0.0603356697,0.3344116509,-0.1820726693,0.0870406851,0.4544463754,-0.408008486,-0.4083961844,-0.2533523142,-0.2293063104,-0.1433107108,0.0884948522,-0.0339344069,0.1266498864,0.2276443839,0.0750585496,-0.2100956887,0.1428106725,-0.3406258523,-0.2292552292,-0.256660223,-0.0860029906,0.0165199991,0.4695231915,-0.0004947771,-0.0502117053,0.066992186,0.0552925058,0.0273389686,0.2316122353,0.2274737656,0.1849201769,-0.0041497685,0.1867880225,-0.0504863523,-0.2215746939,0.3026397526,-0.293017894,0.0360219739,0.1714074165,-0.0174638852,-0.1159898937,-0.0948145688,-0.0791351125,-0.3731910884,-0.0357883871,-0.0878942832,0.232910797,0.1419439316,-0.2297814637,0.1230704486,-0.0462224931,0.2593328953,-0.3241704702,0.1597561389,-0.0609496757,-0.3210313618,-0.0040195277,0.2121324092,-0.1006768197,0.042405732,-0.1941397637,0.3698192537,-0.1435660422,0.1763956845,0.3263867497,0.1948139369,-0.0946418121,-0.4680911601,-0.1868380457,0.0226582251,-0.0780186951,-0.0425728224,-0.191508159,0.4211587608,0.1166608259,0.1690180749,-0.2548603117,0.3526454568,-0.1358468682,-0.0465467684,-0.513710916,-0.1119693369,0.3400008082,-0.4222242832,0.3011046052,-0.183504805,-0.025069749,-0.0542991757,0.2423658371,0.1083917469,0.1255672574,0.1039023697,-0.2200978696,-0.070863083,-0.0215853434,-0.1932946146,0.4366632402,0.1965451688,0.2774853706,0.0860513449,0.0942830667,-0.1230531856,0.0770650208,0.1215845793,-0.0986283645,0.0737261772,0.1942550987,0.148744002,-0.3226958513,-0.0074757859,-0.3011007309,0.0195339154,-0.3337210119,0.0596701726,-0.2368711233,0.3630669713,0.1683836579,-0.0021910882,-0.2333547324,-0.2706065476,0.2153623998,-0.1962648332,-0.0221892428,0.3150249422,-0.2984651625,0.6049127579,0.061546158,-0.022624407,-0.2279758751,-0.1495145708,-0.0368123651,-0.1780066639,-0.1940803528,0.096849516,0.0374841876,-0.1499424875,0.1426575482,-0.1025383621,-0.3601033986,0.0933407098,0.0195365828,0.6119935513,0.3463463485,-0.0454063676,-0.3527431786,0.0991575643,0.4500066638,-0.407998234,-0.2767021954,0.1133303493,0.0314443186,0.1227029637,-0.2230903208,-0.2890613377,0.01126655,-0.2815251946,0.5991318226,0.0410658494,-0.0020701666,-0.2288421988,-0.0237935055,0.1773279607,-0.3290063441,0.2469102144,-0.2935864925,-0.7149354815,0.4781392515,-0.0908914655,-0.2723424137,-0.1058706567,0.067531243,0.434768647,-0.0653754249,-0.6561143994,-0.4437874556,0.0903977379,0.0918385908,-0.0425770208,0.1582767665,0.5028115511,-0.0640258715,0.0077045537,-0.1060191914,-0.2926110923,0.5380766988,-0.1749769151,0.1555449963,-0.2270885706,0.2927685976,0.2596622407,0.9461427331,0.5229724646,0.1494284719,0.2371035963,0.1094772294,0.4445635676,-0.0699854121,-0.1495136172,0.3707888126,0.1129481867,-0.0162508562,0.2179201394,0.0888906717,0.1427087188,-0.1463135183,0.0634615421,-0.2214714736,-0.3302626014,-0.023503473,-0.139239043,0.1961394697,0.0105115296,0.1161011457,-0.2785869241,-0.0160785206,0.3919546008,0.2068582475,0.1829012036,0.2792763114,-0.2079496533,-0.1412262172,-0.5621311665,0.2452997565,-0.3804301023,0.0453098193,-0.1028631479,-0.2098947018,0.1184139177,0.0432683155,0.4675579667,0.0269831009,-0.2499115616,0.0441787876,-0.1463387609,-0.3195356131,-0.1433723271,-0.0926066563,-0.0362703763,0.223690331,0.3636891544,-0.3087708056,-0.2344554365,-0.0264697019,-0.0768967047,-0.2622457147,-0.2304192781,-0.3326176107,0.018139923,-0.1006646156,0.2380574644,-0.2081853747,0.0290682819,0.0581601635,0.0281244665,-0.1481357813,-0.0308295451,0.0667479262,0.1422157288,0.2505323887,-0.2525008023,0.1854340732,0.5488838553,-0.0489839241,0.0482748635,0.3769323528,-0.0075408332,-0.0765798539,0.2972931266,-0.1136460528,0.4294684827,0.5456281304,-0.2386378944,-0.0288395081,0.0150823826,0.3863064945,-0.1887574941,0.0419147946,0.3080283105,-0.0701061636,0.3514488637,0.2013497055,0.2649248838,0.3189358711,-0.1359730959,0.3833771646,0.1452753097,-0.1684723645,0.244407326,0.1406889707,0.9382529855,0.0755253211,0.189523831,0.1168168411,0.1613009721,0.5568839908,0.0864995942,0.0911395848,-0.0336713158,-0.3376883268,-0.0298734251,-0.1386371553,-0.0565793402,-0.2417147756,-0.186282292,0.4872969091,-0.2020582259,-0.1370249987,-0.183180958,-0.1983545125,-0.034543775,-0.2534297705,0.1010975465,-0.0277635474,-0.0001073782,0.4575577974,-0.197728917,-0.1139506102,-0.2750441134,-0.3137764037,-0.2320908308,-0.0686383769,-0.3299613893,-0.0637362972,0.3890515864,-0.0742628425,-0.0691311136,0.2970839739,0.4629917443,0.0839800239,-0.3456386328,0.0233303215,0.0002667142,0.0664639398,0.0784111395,0.0073582432,0.3216218948,-0.0662749112,0.0102050863,0.064731054,-0.2347325385,0.0867333859,0.1223054677,-0.1585008353,-0.02577601,-0.0164247584,0.1262877434,0.0988366976,0.0305557903,-0.3260999322,0.041287113,0.2831495702,-0.2210921347,0.1518936902,0.1960550994,-0.2191640884,-0.0440745093,0.5630209446,-0.2119454592,-0.1510917991,0.4201502204,-0.0155509384,-0.0242564753,-0.0876443908,-0.0401977189,0.541163981,-0.6515188813,0.0948456451,-0.0788772255,-0.0676449686,0.0241858419,0.2146998197,0.4476061463,-0.1635663211,-0.0357842594,-0.166727975,-0.4597932696,0.3591097593,-0.2916518748,0.143611148,-0.4845038354,0.2159706205,-0.1050217897,0.3118296266,-0.2253036052,0.1461495608,-0.3180457056,-0.1291919053,-0.0933290869,0.0110474052,0.2815721631,-0.0890991092,-0.1196709797,0.0586082451,-0.3794817626,-0.0509908348,-0.238400504,0.1548076868,-0.0630956069,0.1762827486,0.0774987191,0.2939431667,-0.1219545379,-0.2028156072,-0.0756871849,0.1594339162,0.0939315781,0.1515529901,0.0464457683,0.1661389619,-0.0703114718,0.0964262262,0.2369101644,-0.1221284121,-0.1685438752,0.0712456554,-0.2272638232,-0.1083660275,0.1468651444,0.0859441385,-0.2312791795,0.1614066213,0.1170423105,0.0157431457,-0.3260185122,-0.0447355211,0.4571228027,0.1518584788,-0.1727563143,0.1071678475,-0.0483281314,0.0419554561,-0.1785603315,0.2204388082,0.0308612287,0.3870626986,0.0750013813,0.2414397597,0.2413665056,-0.0906783566,0.1284103841,0.1479792893,-0.1151852757,-0.1851312667,0.3271128535,-0.0866592899,0.1272324324,0.297246635,0.580282867,0.1529762894,0.18360883,0.0263477731,-0.1465575248,0.2086515427,-0.1388533562,0.2844023407,-0.2059945315,-0.0926633999,-0.241040498,0.0577584542,0.0592740439,-0.1637322903,-0.5557774901,0.8412664533,-0.2907381654,-0.3601021469,-0.1830139905,0.192828849,-0.0918845013,-0.1444141567,-0.1394797862,-0.1471370906,-0.1995597482,-0.152138263,0.2422818094,-0.0227823351,0.2497125715,0.0456516445,0.0298936348,-0.2345262021,-0.5137044191,0.2872682214,0.2645595074,-0.1797736734,0.1278647333,0.1644396335,-0.3647380471,0.1035095528,0.7005337477,0.2391315401,-0.057740543,-0.1001672372,0.0456188358,0.2528724968,-0.0410716906,-0.0698783696,0.1960646361,0.3556524813,-0.593210578,0.1043868959,-0.0007090137,-0.0965858996,0.402623862,-0.2877337039,0.3789049089,-0.2412760854,0.4647481143,-0.1182563081,0.0442552604,0.0336359665,-0.1767843664,-0.6990652084,0.3790702522,-0.0600344241,0.0051059178,0.0743904933,-0.257763654,0.0131136058,-0.0343388282,0.2901058197,0.4205664694,0.0609129779,-0.2994609475,-0.2646945417,-0.5396564007,0.1927631795,0.1414522082,0.34311077,-0.0039898288,0.3530112803,0.0309116282,0.0716796219,0.2382591516,-0.0196619853,-0.0600884296,-0.0279571507,-0.637863338,0.1246493161,-0.1350907683,0.0022307422,0.1786823124,0.3266239762,0.0166260209,-0.2267824411,-0.0830882192,0.3012852073,-0.0886904001,-0.1383722723,0.3098233044,0.4241041839,0.0609222949,0.1810896993,0.0859954581,-0.0158585329,-0.1867468506,-0.1026432812,-0.0907712579,-0.0847365633,-0.2794478536,0.4343644381,-0.4776818752,-0.3170902133,-0.3601943851,0.070885621,0.1506469697,-0.1097966358,-0.3143194616,0.3568704426,-0.1146318465,0.0508094653,0.3609861732,0.3676590621,-0.0981448069,0.3505352139,-0.3764446676,-0.1778484732,0.77211833,-0.4845305681,0.2451476455,-0.3426685035,0.3337124288,0.6285014153,-0.2959862649,-0.5779026151,-0.0101888534,0.1667891294,0.0562025867,-0.3047447205,0.1543692052,-0.4789434075,0.0177680869,-0.0675173476,0.3459015191,0.1138315424,-0.0296077505,-0.0420089923,-0.183147043]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"OK, it definitely leads to a race condition in how it's used right now. Here is how you can reproduce it - by injecting a random sleep time different for each process before the locks are acquired. \r\n```\r\n--- a\/src\/datasets\/metric.py\r\n+++ b\/src\/datasets\/metric.py\r\n@@ -348,6 +348,16 @@ class Metric(MetricInfoMixin):\r\n\r\n elif self.process_id == 0:\r\n # Let's acquire a lock on each node files to be sure they are finished writing\r\n+\r\n+ import time\r\n+ import random\r\n+ import os\r\n+ pid = os.getpid()\r\n+ random.seed(pid)\r\n+ secs = random.randint(1, 15)\r\n+ time.sleep(secs)\r\n+ print(f\"sleeping {secs}\")\r\n+\r\n file_paths, filelocks = self._get_all_cache_files()\r\n\r\n # Read the predictions and references\r\n@@ -385,7 +395,10 @@ class Metric(MetricInfoMixin):\r\n\r\n if predictions is not None:\r\n self.add_batch(predictions=predictions, references=references)\r\n+ print(\"FINALIZE START\")\r\n+\r\n self._finalize()\r\n+ print(\"FINALIZE END\")\r\n\r\n self.cache_file_name = None\r\n self.filelock = None\r\n```\r\n\r\nthen run with 2 procs: `python -m torch.distributed.launch --nproc_per_node=2`\r\n```\r\nexport BS=16; rm -r output_dir; PYTHONPATH=src USE_TF=0 CUDA_VISIBLE_DEVICES=0,1 python -m torch.distributed.launch --nproc_per_node=2 examples\/seq2seq\/run_seq2seq.py --model_name_or_path t5-small --output_dir output_dir --adam_eps 1e-06 --do_eval --do_train --do_predict --evaluation_strategy=steps --label_smoothing 0.1 --learning_rate 3e-5 --logging_first_step --logging_steps 1000 --max_source_length 128 --max_target_length 128 --num_train_epochs 1 --overwrite_output_dir --per_device_eval_batch_size $BS --per_device_train_batch_size $BS --predict_with_generate --eval_steps 25000 --sortish_sampler --task translation_en_to_ro --val_max_target_length 128 --warmup_steps 500 --max_train_samples 10 --max_val_samples 10 --max_test_samples 10 --dataset_name wmt16 --dataset_config ro-en --source_prefix \"translate English to Romanian: \"\r\n```\r\n\r\n```\r\n***** Running Evaluation *****\r\n Num examples = 10\r\n Batch size = 16\r\n 0%| | 0\/1 [00:00\r\n main()\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 601, in main\r\n metrics = trainer.evaluate(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer_seq2seq.py\", line 74, in evaluate\r\n return super().evaluate(eval_dataset, ignore_keys=ignore_keys, metric_key_prefix=metric_key_prefix)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer.py\", line 1703, in evaluate\r\n output = self.prediction_loop(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer.py\", line 1876, in prediction_loop\r\n metrics = self.compute_metrics(EvalPrediction(predictions=preds, label_ids=label_ids))\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 556, in compute_metrics\r\n result = metric.compute(predictions=decoded_preds, references=decoded_labels)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 402, in compute\r\n self._finalize()\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 370, in _finalize\r\n raise ValueError(\r\nValueError: Error in finalize: another metric instance is already using the local cache file. Please specify an experiment_id to avoid colision between distributed metric instances.\r\n```","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":452,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nOK, it definitely leads to a race condition in how it's used right now. Here is how you can reproduce it - by injecting a random sleep time different for each process before the locks are acquired. \r\n```\r\n--- a\/src\/datasets\/metric.py\r\n+++ b\/src\/datasets\/metric.py\r\n@@ -348,6 +348,16 @@ class Metric(MetricInfoMixin):\r\n\r\n elif self.process_id == 0:\r\n # Let's acquire a lock on each node files to be sure they are finished writing\r\n+\r\n+ import time\r\n+ import random\r\n+ import os\r\n+ pid = os.getpid()\r\n+ random.seed(pid)\r\n+ secs = random.randint(1, 15)\r\n+ time.sleep(secs)\r\n+ print(f\"sleeping {secs}\")\r\n+\r\n file_paths, filelocks = self._get_all_cache_files()\r\n\r\n # Read the predictions and references\r\n@@ -385,7 +395,10 @@ class Metric(MetricInfoMixin):\r\n\r\n if predictions is not None:\r\n self.add_batch(predictions=predictions, references=references)\r\n+ print(\"FINALIZE START\")\r\n+\r\n self._finalize()\r\n+ print(\"FINALIZE END\")\r\n\r\n self.cache_file_name = None\r\n self.filelock = None\r\n```\r\n\r\nthen run with 2 procs: `python -m torch.distributed.launch --nproc_per_node=2`\r\n```\r\nexport BS=16; rm -r output_dir; PYTHONPATH=src USE_TF=0 CUDA_VISIBLE_DEVICES=0,1 python -m torch.distributed.launch --nproc_per_node=2 examples\/seq2seq\/run_seq2seq.py --model_name_or_path t5-small --output_dir output_dir --adam_eps 1e-06 --do_eval --do_train --do_predict --evaluation_strategy=steps --label_smoothing 0.1 --learning_rate 3e-5 --logging_first_step --logging_steps 1000 --max_source_length 128 --max_target_length 128 --num_train_epochs 1 --overwrite_output_dir --per_device_eval_batch_size $BS --per_device_train_batch_size $BS --predict_with_generate --eval_steps 25000 --sortish_sampler --task translation_en_to_ro --val_max_target_length 128 --warmup_steps 500 --max_train_samples 10 --max_val_samples 10 --max_test_samples 10 --dataset_name wmt16 --dataset_config ro-en --source_prefix \"translate English to Romanian: \"\r\n```\r\n\r\n```\r\n***** Running Evaluation *****\r\n Num examples = 10\r\n Batch size = 16\r\n 0%| | 0\/1 [00:00\r\n main()\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 601, in main\r\n metrics = trainer.evaluate(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer_seq2seq.py\", line 74, in evaluate\r\n return super().evaluate(eval_dataset, ignore_keys=ignore_keys, metric_key_prefix=metric_key_prefix)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer.py\", line 1703, in evaluate\r\n output = self.prediction_loop(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer.py\", line 1876, in prediction_loop\r\n metrics = self.compute_metrics(EvalPrediction(predictions=preds, label_ids=label_ids))\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 556, in compute_metrics\r\n result = metric.compute(predictions=decoded_preds, references=decoded_labels)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 402, in compute\r\n self._finalize()\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 370, in _finalize\r\n raise ValueError(\r\nValueError: Error in finalize: another metric instance is already using the local cache file. Please specify an experiment_id to avoid colision between distributed metric instances.\r\n```","embeddings":[-0.0844155699,-0.0657195449,0.0652180836,0.2722371519,-0.0870862305,-0.032664571,0.1643525809,0.2697967291,0.3311934769,0.0832053274,0.098261185,0.1941852719,-0.3580646217,-0.0889754966,-0.0156052159,0.0498044565,0.0008351138,-0.1240264401,-0.4149836302,-0.1004132852,-0.3499764502,0.5602373481,-0.1571679413,0.0035177716,-0.4620630443,0.1669693291,-0.1928113848,0.2385937572,-0.1055250093,-0.5698463917,0.2600385845,0.3488971293,-0.0173991658,0.5673115849,-0.000121866,-0.156548962,0.2898033559,-0.1259326637,-0.0928893313,-0.1496436298,0.0516416617,-0.3085162044,0.4341697991,-0.1971871704,-0.0820829719,-0.0098867286,0.075831078,-0.495130688,0.4286834002,0.0644462407,0.0891190469,0.4689293504,-0.0374805965,-0.1865046918,-0.0093006808,-0.1247958094,0.0236409735,0.3594599664,-0.0119030736,0.0050282986,-0.2551394999,0.2341878414,0.2350952029,-0.0686758906,0.3630651534,0.0002415854,-0.0232570134,-0.1683153063,-0.1944670528,0.3529827297,0.1351855695,-0.1981303096,-0.2590878606,0.0288155973,-0.1893539131,-0.4624345005,0.1429929435,-0.1314374804,0.0301056616,0.088087298,0.0458046198,0.0006789111,0.006420983,0.0480430722,-0.0750627741,0.0268831421,-0.1102879792,0.0966649428,-0.0770767033,0.267449975,-0.3081120551,0.21201244,-0.0161630362,0.1204218641,-0.43568331,0.0894821882,0.2162855715,0.0534023084,0.0726227984,0.4313857853,-0.0319301262,0.1022296101,0.397911638,0.1139199957,-0.0671368837,0.4132384956,0.2391605079,-0.0104756225,0.1874959916,0.2739362419,-0.0854610801,-0.2305916995,0.2011027634,-0.2459424585,0.3607820272,0.0959146693,0.1355988532,-0.3899984956,0.0805313736,0.3011382222,-0.0750805512,-0.1568284333,0.1580274254,0.3855433464,0.085239619,-0.0463728569,0.3243296742,0.3055721819,-0.2333544642,0.0019363248,-0.4296098948,-0.2353733927,-0.119487904,0.3071151078,0.0591079742,-0.0130565641,0.6048250794,-0.1873202324,0.3196305037,-0.0907878578,0.3079951704,0.1744770408,0.0746124089,0.2866277099,-0.0819513202,0.1851965189,0.3427527249,0.1244142652,-0.2457848489,-0.2860185802,-0.0978619531,-0.4152011573,0.1885588765,0.000531118,-0.5068399906,0.239495337,0.4041298926,0.2187768668,-0.0533623956,-0.1476347297,0.2692391574,-0.0410024524,-0.158067435,-0.017097719,0.437728554,0.8433099985,-0.2321900427,-0.4020535648,0.1269695163,-0.0551433563,0.0210924614,0.2261457443,-0.0151569322,0.131389752,-0.2769354284,0.1363208145,0.3145765662,-0.3304304481,-0.5092270374,-0.0667772964,-0.437472254,-0.0340688936,0.1925942451,0.109986119,0.0847023278,0.1234846264,0.1166357398,-0.2643081546,0.2656232119,-0.3458692431,-0.3824667335,-0.2897080481,-0.0370660089,0.0175562743,0.3021219373,0.0015976339,-0.0653040186,0.0374565944,0.1378031969,0.1916084439,0.1767644733,0.2403720915,0.2350229025,0.0061784941,0.2478129268,0.1133008152,-0.2364605367,0.3414187133,-0.3823010921,0.0915210992,0.1842734665,0.0344474316,-0.1589950174,-0.1162987277,-0.1790624857,-0.4125052691,0.0269138478,-0.1299059093,0.1790639162,0.1748148352,-0.1957106292,0.1845989525,-0.0311742108,0.2445206791,-0.4866950214,0.0809158459,-0.0441642366,-0.3607316613,-0.0655888617,0.0390097611,0.0376205966,0.0411496051,-0.138989225,0.3688038886,-0.1038145944,0.2514378428,0.2193906009,0.2252401263,-0.0583164766,-0.3309867978,-0.1056120545,0.1239272431,0.0039939252,-0.045171421,-0.1759320796,0.3884069026,0.090015173,0.14617531,-0.3499466777,0.2182744592,-0.0972633138,-0.1663038135,-0.555668056,-0.0380490161,0.4728535712,-0.2647351027,0.2770218849,-0.1631315947,0.0137206158,0.0223108213,0.233317107,0.154551968,0.1547574848,0.0433734059,-0.233309418,-0.0644800887,0.0418287627,-0.1615593135,0.4600605667,0.1955893338,0.2431957573,0.1327683926,0.087455593,-0.2459524572,0.0757683069,0.0206983369,-0.1691020578,0.159304902,0.1514809728,0.1435250938,-0.3288494051,-0.1380015016,-0.2743545771,0.0329185165,-0.3341398835,0.1020790935,-0.1349844486,0.4001249671,0.154652074,0.0645313635,-0.1272897869,-0.2716541886,0.1869915426,-0.1014484242,0.0305638071,0.2809856236,-0.2944428027,0.5300469995,0.110001944,-0.181198895,-0.3561827242,-0.2111852467,-0.0076867859,-0.1622644514,-0.1794365942,0.19308348,0.0939458907,-0.1835871041,-0.0286585875,-0.0898886323,-0.354308933,0.0646219328,0.0579995736,0.680811584,0.2862638831,0.0012738783,-0.2250635922,0.0685060695,0.4061901569,-0.4061918259,-0.3184202313,0.0798092559,-0.0271717571,0.1367471367,-0.3171879351,-0.2778091431,0.1093105599,-0.3724223077,0.6279050708,0.0351591259,-0.0284729768,-0.1151442304,-0.006565087,0.2452505231,-0.2081536949,0.2181218863,-0.3708445728,-0.6790494919,0.3761582971,-0.0701826364,-0.2920279801,-0.1314426363,0.0402270742,0.3392747641,-0.0791691542,-0.5926974416,-0.522321403,-0.0324288048,0.3241592348,0.0222350433,-0.0101590138,0.5464061499,-0.0936023146,-0.0244744457,-0.0935700014,-0.1157884747,0.6498821378,-0.1600061059,0.125362426,-0.041815564,0.5212675333,0.1402976364,0.9083513021,0.5984097719,0.1078086942,0.2427328527,0.0859564245,0.3766086996,-0.1323872656,-0.1446452439,0.3193695247,0.1811810881,-0.1200467646,0.1948687732,0.0696764365,0.1736306995,-0.0670456439,0.0006568957,-0.2872377336,-0.3173183501,-0.0860211104,-0.0220395755,0.1946453005,0.0260201897,0.0464021116,-0.1714588404,0.0210230071,0.4557762146,0.4667409062,0.2445893735,0.151806578,-0.2346044928,0.150272131,-0.6847786307,0.2808535695,-0.4843690991,0.09266188,-0.2083668411,-0.2226665765,0.1251804084,-0.1038854867,0.5619880557,0.1111256108,-0.0980999917,0.0389469489,-0.2003228962,-0.4362590313,-0.0758717284,-0.1520861238,-0.0548392199,0.3091390729,0.4101701975,-0.437271148,-0.2195557803,-0.1674850434,-0.1153562516,-0.263875246,-0.3710044324,-0.3840120137,-0.0693475381,-0.1728612483,0.2685464621,-0.1494049579,-0.0337936729,-0.0002119843,-0.0176423471,-0.177866444,-0.1178585738,0.0093377149,0.1683959514,0.3239504397,-0.3101280332,0.2613334954,0.5088202357,-0.2863684893,0.0070695933,0.3720149994,0.1125952974,0.0114657218,0.3537889421,-0.0389570892,0.1451955289,0.6060222983,-0.2288959026,0.1373125911,0.0775088817,0.3824473619,-0.1733544618,-0.0328599103,0.3560639918,-0.0646378472,0.36572963,0.1943790019,0.213457644,0.2551449537,-0.2164087892,0.3798713684,0.0024198729,-0.1254675984,0.289678067,0.0662679747,0.9412876368,0.1287704408,0.1316235662,0.3074283302,0.0300386325,0.3580708206,0.1575554758,0.0769383833,-0.2155002952,-0.2878910005,0.0281255189,-0.1426008344,0.0565012842,-0.2786584496,-0.1992247254,0.3468231261,-0.3472928107,-0.1368992031,-0.2096108645,-0.1000206545,-0.2952027917,-0.1441333294,0.1108286306,0.0379240811,0.0561159588,0.5053175092,-0.196272254,-0.1456767321,-0.3237811923,-0.2361002415,-0.0989164487,-0.0192418303,-0.2128232867,0.0031836552,0.2255653143,-0.0905264392,-0.1694483608,0.2623282373,0.4520953894,0.1580691785,-0.2806242406,0.0806226283,0.130550161,0.0593532436,0.1180900484,-0.0961390361,0.239487648,-0.1169218719,-0.0971455425,0.0897525549,-0.2411247194,0.0596833192,0.0625083894,-0.1247530133,-0.06132783,-0.0006805723,0.0187084042,0.1744980812,-0.0133294342,-0.2417227179,0.0776691362,0.3355978727,-0.1957993507,0.0846767426,0.1114365906,-0.2621316016,-0.1653986871,0.5599821806,-0.1930417269,-0.043804802,0.4942462146,0.0620817132,-0.0214790087,-0.0895953104,0.066549547,0.4614343345,-0.7272011638,0.1340647936,-0.0469323993,-0.0091641629,0.1100817844,0.2127385885,0.3792605698,-0.1826424152,-0.0886157155,-0.1702752709,-0.349769026,0.2320081294,-0.3465448022,0.1859749705,-0.4120010436,0.086351119,-0.2553619444,0.2320757806,-0.2822775245,0.1403428763,-0.3109701872,-0.1574933976,-0.0858474523,-0.1043038294,0.3523840904,-0.0494926497,-0.0698431283,0.1268845499,-0.4691964984,-0.1015590876,-0.2797078192,0.1388323605,-0.130494535,0.0779665112,-0.0051861191,0.2057712078,-0.0304877907,-0.0940724909,0.0187169164,0.1957746148,0.0266360268,0.0420336649,0.2091130763,0.1399056762,-0.1949334145,0.1630113274,0.0852947906,-0.2907069623,-0.115603976,-0.0107713751,-0.2982536852,-0.072202608,0.1672507674,0.1254228204,-0.1080607101,0.1442044377,0.1264520139,-0.0076734037,-0.300779134,-0.0219408814,0.4440127611,0.1493700147,-0.1383358836,-0.1004374251,-0.1468465328,0.0966836885,-0.0956381112,0.2848831713,-0.0055988426,0.3573443592,0.1428943127,0.2579458356,0.1733795553,0.0006109166,0.0838580877,0.287468493,0.0338028185,-0.1745362431,0.2908103168,-0.2125217617,0.112110272,0.3444501162,0.6617323756,0.0436841846,0.1996881068,-0.1315075755,-0.0667857826,0.2362422049,-0.1090037003,0.3268863857,-0.2201714516,-0.1009963304,-0.1156372502,0.0700995922,-0.1066155359,-0.133678779,-0.3784928024,0.6685897112,-0.3151126504,-0.2031569481,-0.1847664267,0.2426720858,-0.135775581,-0.2470302284,-0.1185152754,-0.1595748663,-0.0961114466,-0.1728561968,0.2104006261,0.0722266808,0.1171074212,0.052264642,0.0329216421,-0.2532563508,-0.4568680227,0.2618017495,0.2318311185,-0.2234422565,0.0250683147,0.2808709741,-0.3499053419,0.1182914078,0.6035231352,0.2323000133,-0.0582313836,-0.0834590867,0.1073073819,0.3267668784,0.0169561841,-0.0994201601,0.1862422079,0.3138247132,-0.5439269543,0.0629945844,0.0679738596,-0.1795365065,0.3942139149,-0.2318457216,0.3350016475,-0.253605485,0.4101294875,-0.2309689522,0.1492196023,0.044267673,-0.1309387088,-0.6219240427,0.2565176487,-0.0546287671,-0.0505176038,0.0629609227,-0.2929469347,0.0455986708,-0.1245393902,0.2503241599,0.3252754509,0.0252176952,-0.3522538543,-0.1959876716,-0.597715497,0.2946241498,0.145552054,0.2486119717,-0.0141235683,0.3637274206,-0.0974159986,0.0827243179,0.1792985201,-0.0695216507,-0.1520158648,0.0536880456,-0.5792644024,0.1903704703,-0.1522354037,-0.1525033265,0.1285752654,0.1937535256,0.003823478,-0.2565794885,0.0009941093,0.2496807128,-0.2043685317,0.0098005123,0.2409560531,0.4613412321,0.1374110281,0.0890942886,0.0119939493,-0.0801084191,-0.2506814599,-0.1601435989,-0.1050342768,-0.1749790162,-0.3328834176,0.4853777587,-0.3128299117,-0.5159752369,-0.2895016968,0.2952622175,0.0879574716,-0.3334015012,-0.236761421,0.3818435073,-0.1371771544,0.0356156118,0.2386960387,0.3839943707,0.0910454392,0.3809011281,-0.3290142417,-0.1753951311,0.7739758492,-0.5265752673,0.2481269538,-0.2379906625,0.3535308242,0.4005734622,-0.3686007559,-0.2991082072,-0.0865050331,0.1388595402,0.0136334766,-0.1585733891,0.1545866728,-0.3073345721,-0.0271864738,-0.048386246,0.1458697468,0.1749285161,-0.049109105,-0.0030156351,-0.1477310956]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"I tried to adjust `run_seq2seq.py` and trainer to use the suggested dist env:\r\n```\r\n import torch.distributed as dist\r\n metric = load_metric(metric_name, num_process=dist.get_world_size(), process_id=dist.get_rank())\r\n```\r\nand in `trainer.py` added to call just for rank 0:\r\n```\r\n if self.is_world_process_zero() and self.compute_metrics is not None and preds is not None and label_ids is not None:\r\n metrics = self.compute_metrics(EvalPrediction(predictions=preds, label_ids=label_ids))\r\n```\r\nand then the process hangs in a deadlock. \r\n\r\nHere is the tb:\r\n```\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/filelock.py\", line 275 in acquire\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 306 in _check_all_processes_locks\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 501 in _init_writer\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 440 in add_batch\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 397 in compute\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 558 in compute_metrics\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer.py\", line 1876 in prediction_loop\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer.py\", line 1703 in evaluate\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer_seq2seq.py\", line 74 in evaluate\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 603 in main\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 651 in \r\n```\r\n\r\nBut this sounds right, since in the above diff I set up a distributed metric and only called one process - so it's blocking on waiting for other processes to do the same.\r\n\r\nSo one working solution is to leave:\r\n\r\n```\r\n metric = load_metric(metric_name)\r\n```\r\nalone, and only call `compute_metrics` from rank 0\r\n```\r\n if self.is_world_process_zero() and self.compute_metrics is not None and preds is not None and label_ids is not None:\r\n metrics = self.compute_metrics(EvalPrediction(predictions=preds, label_ids=label_ids))\r\n```\r\n\r\nso we now no longer use the distributed env as far as `datasets` is concerned, it's just a single process.\r\n\r\nAre there any repercussions\/side-effects to this proposed change in Trainer? If it always gathers all inputs on rank 0 then this is how it should have been done in first place - i.e. only run for rank 0. It appears that currently it was re-calculating the metrics on all processes on the same data just to throw the results away other than for rank 0. Unless I missed something.\r\n","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":302,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nI tried to adjust `run_seq2seq.py` and trainer to use the suggested dist env:\r\n```\r\n import torch.distributed as dist\r\n metric = load_metric(metric_name, num_process=dist.get_world_size(), process_id=dist.get_rank())\r\n```\r\nand in `trainer.py` added to call just for rank 0:\r\n```\r\n if self.is_world_process_zero() and self.compute_metrics is not None and preds is not None and label_ids is not None:\r\n metrics = self.compute_metrics(EvalPrediction(predictions=preds, label_ids=label_ids))\r\n```\r\nand then the process hangs in a deadlock. \r\n\r\nHere is the tb:\r\n```\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/filelock.py\", line 275 in acquire\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 306 in _check_all_processes_locks\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 501 in _init_writer\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 440 in add_batch\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/metric.py\", line 397 in compute\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 558 in compute_metrics\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer.py\", line 1876 in prediction_loop\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer.py\", line 1703 in evaluate\r\n File \"\/mnt\/nvme1\/code\/huggingface\/transformers-mp-pp\/src\/transformers\/trainer_seq2seq.py\", line 74 in evaluate\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 603 in main\r\n File \"examples\/seq2seq\/run_seq2seq.py\", line 651 in \r\n```\r\n\r\nBut this sounds right, since in the above diff I set up a distributed metric and only called one process - so it's blocking on waiting for other processes to do the same.\r\n\r\nSo one working solution is to leave:\r\n\r\n```\r\n metric = load_metric(metric_name)\r\n```\r\nalone, and only call `compute_metrics` from rank 0\r\n```\r\n if self.is_world_process_zero() and self.compute_metrics is not None and preds is not None and label_ids is not None:\r\n metrics = self.compute_metrics(EvalPrediction(predictions=preds, label_ids=label_ids))\r\n```\r\n\r\nso we now no longer use the distributed env as far as `datasets` is concerned, it's just a single process.\r\n\r\nAre there any repercussions\/side-effects to this proposed change in Trainer? If it always gathers all inputs on rank 0 then this is how it should have been done in first place - i.e. only run for rank 0. It appears that currently it was re-calculating the metrics on all processes on the same data just to throw the results away other than for rank 0. Unless I missed something.\r\n","embeddings":[0.0881400257,-0.1092867628,0.1158301681,0.2856828272,0.0845733956,0.0806555972,0.2349973619,0.2377158999,0.2347908914,0.2356441766,0.0788630545,0.166712448,-0.373960644,-0.2242649049,0.0668825433,0.0503915213,-0.0005046903,-0.1439883858,-0.221030131,-0.1951075941,-0.2040551752,0.5319712758,-0.088830635,-0.0502556562,-0.5913400054,0.2881834209,-0.1651010513,0.1823531091,-0.0684032291,-0.617151022,0.3700884283,0.1490116119,0.0829517618,0.5546138883,-0.0001318881,-0.1994537562,0.2629815638,-0.2211295068,-0.1534065604,-0.2687932551,0.1527822465,-0.2575644553,0.571503222,-0.1630466133,-0.1416184306,-0.0879457891,0.0900792032,-0.4483617544,0.4259662926,0.1269228458,-0.0067988681,0.3528015912,-0.0033912866,-0.2833948433,0.097712405,-0.1208251268,0.0339141265,0.3804101944,-0.0241650287,0.0620026514,-0.327988416,0.2068395466,0.2323785722,-0.0643238276,0.3330114484,-0.0780323222,-0.1014687717,-0.1370963901,-0.0787223428,0.3096213937,0.2080832869,-0.1693377495,-0.2303849608,0.0603585951,0.0139125641,-0.343267709,0.1953567713,-0.0388792939,-0.0130721033,0.04169387,0.0024041771,-0.1951526105,0.0043453164,0.0744460374,-0.1773838401,0.1295827478,-0.0661413819,0.0334934853,0.0162886679,0.2591890097,-0.1607431322,0.3018882573,0.0713185444,0.1371362656,-0.3619210124,0.170331955,0.2666286528,0.057864666,-0.0056573586,0.319526583,-0.0361856818,0.0114500206,0.4184844494,0.0381960087,-0.1389734447,0.362706542,0.2786945701,0.1029436886,0.1162616462,0.1878528744,-0.0855451375,-0.2297569364,0.0198366959,-0.1975078732,0.3149439096,0.2442042977,-0.0334214009,-0.4205731153,0.1195659563,0.295242995,-0.1290903389,-0.0933247954,0.1693427861,0.3671609759,0.0882651433,0.0787011161,0.3353294134,0.2471677959,-0.199189499,-0.0173158702,-0.4135643542,-0.0986426994,-0.1242088079,0.2724477649,0.2196628004,-0.1628999412,0.6351852417,-0.1787698269,0.4696894288,-0.0300205424,0.1313095689,0.1857780814,0.1316873729,0.2409881353,-0.0306473374,0.2101943344,0.3361890912,0.3006996512,-0.1925242394,-0.1458505243,-0.1428042948,-0.5472254157,0.1147262827,-0.096416086,-0.4572981298,0.2726358473,0.3576215804,0.1117123365,-0.1076127216,-0.1904318482,0.1932600141,-0.0918901116,-0.0997676253,-0.0660918951,0.5843692422,0.7293358445,-0.2147573531,-0.4012365341,0.098839514,0.0325715914,-0.1022926718,0.1529448926,-0.0788165107,0.1933204234,-0.3543674052,-0.0127081899,0.3378298879,-0.4275298417,-0.5466860533,-0.1503923386,-0.3531804681,-0.1286658049,0.2151797116,0.0550851822,0.129709959,0.1625282913,0.1186125427,-0.1818707138,0.1395101547,-0.3199318945,-0.2878010869,-0.1782021075,0.054649163,0.0048474343,0.328759104,-0.0217356887,-0.0645741001,0.1194041222,0.1203885376,0.1731573492,0.1427398622,0.3467543721,0.1570518166,0.0381508358,0.2471780181,0.1363053173,-0.3056511879,0.305378437,-0.3536006808,0.0743503794,0.0327291712,0.0219307262,-0.1138970107,-0.2049700618,-0.1658814549,-0.4732397497,-0.1058141738,-0.0443967432,0.1530443281,0.101691477,-0.1681249887,0.1105814353,-0.0655051917,0.2719031572,-0.4879302084,0.0612640902,-0.0638747662,-0.3398908675,-0.0635269657,0.16161789,0.0212047491,0.0049574156,-0.1978687644,0.3824717104,-0.1466091573,0.2167610377,0.2523966432,0.1297360212,-0.0266656876,-0.5362629294,-0.024970945,0.1701551974,-0.0606264137,0.0106334873,-0.2244264483,0.3717922568,0.1169154719,0.1528419852,-0.3292899132,0.2574849725,-0.1934431642,-0.0311923902,-0.4751604199,-0.0524111949,0.461251229,-0.1249243543,0.3616275787,-0.0769227818,0.0478943437,-0.0536687672,0.1575220674,0.1602757126,0.0998461619,0.0825526267,-0.1949329972,-0.0521236435,0.0845899135,-0.2958513796,0.418176353,0.1859461218,0.2945014834,0.0671907365,0.138365075,-0.1661348194,0.0406900011,-0.0327739678,-0.1795000285,0.2156122923,0.2768219709,0.191608727,-0.2365643233,-0.0302204862,-0.2584660649,0.0539821386,-0.339664638,0.0545165725,-0.2114023566,0.5269723535,0.1609437466,-0.001784038,-0.3060412407,-0.2587700784,0.1985844225,-0.1509752721,0.1073668972,0.2856147289,-0.1824020594,0.366370976,0.1217021942,-0.1664306521,-0.2195236683,-0.2145324796,-0.0526186749,-0.2615052164,-0.2276444286,-0.0176186599,0.1051164493,-0.181295082,-0.0146132056,-0.0838483274,-0.3236866295,0.0789873451,0.0525829904,0.5157334208,0.3757797182,0.0781337321,-0.3018793762,0.1728724241,0.4036900997,-0.4335684776,-0.2911327779,0.0549028963,0.0282635931,0.2603513598,-0.2404197305,-0.3637054861,0.0195961576,-0.3122047782,0.4976966977,0.0147305662,-0.0670966133,-0.2401181161,0.0159149691,0.2207764983,-0.1401596814,0.3164168,-0.2179742008,-0.7374212146,0.4974017739,-0.0844749361,-0.2427448332,-0.0719342381,0.0304230992,0.3310313225,0.0592322648,-0.674808383,-0.3775847554,0.0442798994,0.2911629379,-0.1235283613,0.138741374,0.5976380706,-0.0984463319,0.0906325802,-0.084837988,-0.164619714,0.514677465,-0.2370685041,0.2622246742,-0.0529930331,0.3953697681,0.205136925,1.0660384893,0.4365729094,0.1375798136,0.2404038608,0.0790980682,0.3862739801,-0.1463313997,-0.1985452324,0.2925935388,0.2944891751,-0.0125393216,0.1614468694,0.1040411443,0.2343330532,-0.1193147451,-0.0580517873,-0.2382648885,-0.2966466844,0.0124176964,-0.246753037,0.2885679901,-0.0768272951,0.1307217479,-0.2336004227,-0.0309700537,0.4748420715,0.3607445061,0.3400712013,0.2659281194,0.0138886757,0.0185916759,-0.5698681474,0.2131749541,-0.3507131338,0.0432460122,-0.1114050597,-0.2091226876,0.199576512,-0.0166746136,0.6270531416,0.2594262064,-0.1323656589,-0.0344221927,-0.3146463633,-0.4235260487,-0.0951624364,0.0202626009,-0.0807242095,0.261980474,0.244354099,-0.4367316365,-0.2846646607,-0.0777887031,-0.0575864054,-0.2513468266,-0.331643343,-0.2910419405,0.072215341,-0.2103280276,0.1533391327,-0.1221069545,0.0151933907,-0.0006407483,0.0610141866,-0.1972708702,-0.0585675277,0.0051170075,0.0079776077,0.3492991328,-0.3426237404,0.1846986711,0.5938441753,0.0462311283,0.0694633201,0.3731180429,0.0241030529,-0.0076116379,0.3295188248,-0.1096442044,0.1984226108,0.6668339968,-0.2204794735,0.0092969118,-0.1022757515,0.2657603025,-0.0840354264,-0.024089478,0.275171876,-0.1857896,0.3366446495,0.1338977963,0.1865536273,0.3631666601,-0.0754490793,0.3928913474,0.0175169986,-0.1314404905,0.1697278917,0.0900112391,1.0596708059,0.1733086109,0.288056463,0.0718040615,0.1134336144,0.2868271768,0.0770513862,0.0670693964,-0.2336565703,-0.2930971682,-0.0237972289,-0.2060994804,-0.0268658586,-0.2503251135,-0.0878033638,0.4228726625,-0.3943589628,-0.0798699409,-0.2098240703,-0.1599922627,-0.215791136,-0.1572387666,0.1809965968,-0.0825239196,0.0263455622,0.5747039318,-0.2157615572,-0.1310919821,-0.3793376684,-0.3467714787,-0.1881611496,-0.0683698505,-0.3320905268,0.0041430867,0.332113117,-0.3493465483,-0.1050181761,0.405166477,0.4119630158,0.1500435024,-0.3158900142,0.0133931451,0.1522022337,0.0163300801,0.0662481487,0.0037695591,0.2145133317,-0.0231949016,-0.0367353521,0.1018836126,-0.2719015181,0.0721195191,0.0464071892,-0.1299458146,0.1035778597,0.0079809558,-0.0395117998,0.0416478962,0.0088516558,-0.2246104926,0.0092117814,0.3180173934,-0.3020229042,-0.1260164231,0.1327933818,-0.2384745479,-0.0902607068,0.7388558388,-0.1895920187,-0.0833806396,0.508664906,0.1132063568,-0.0950124711,-0.0279414672,0.0534411073,0.5871481895,-0.7964068651,0.0375650674,-0.0280949753,-0.0733696371,0.0531302951,0.157377854,0.4190034568,-0.1953767538,-0.0511974134,-0.1429456025,-0.5170990229,0.1747768372,-0.4015873969,0.2136110812,-0.4579894245,0.1037899777,-0.1205526292,0.1368554384,-0.1713806838,0.2145242989,-0.1843110472,-0.0760746449,-0.0796444267,-0.0408797599,0.2494169623,-0.0830168873,-0.132385239,0.1427082419,-0.2992048264,-0.0110410983,-0.259768635,0.2003601789,-0.1354681253,0.1764083058,-0.028694151,0.1227868423,-0.0617896058,-0.1407510042,0.0586236343,0.238239795,0.1416047066,0.1680943072,0.183171764,0.205873549,-0.1489621252,0.1856940389,0.1167300865,-0.1866595,-0.1223106533,-0.0412707701,-0.3114250898,-0.0305346865,0.0741111413,0.1024200544,-0.1661493331,0.2045538127,0.0971289128,0.0333639421,-0.2555373311,0.0163982231,0.3647414148,0.0902378857,-0.2043006718,0.055267211,-0.0069349422,0.021738844,-0.1006231233,0.2891527712,0.0024499164,0.3737566471,-0.0071059661,0.2339390218,0.1555727422,-0.0368139558,0.0814675018,0.2663898766,-0.0957861692,-0.164601177,0.3494300246,-0.0394395106,0.1696405858,0.291565299,0.5672032237,0.1148113981,0.1399044096,-0.0718218163,-0.0939018354,0.3427062035,-0.1072866991,0.2498675138,-0.1057759374,-0.2143326253,-0.2015596032,-0.010069808,0.1364575475,-0.2415562272,-0.4650448859,0.7891871333,-0.2494346499,-0.2857100368,-0.1221793145,0.2552675307,-0.1764252037,-0.1262228936,-0.0582016706,-0.1612551361,-0.1629287004,-0.0820104331,0.217740342,-0.0160791352,0.1115121022,0.2160970122,-0.0037840668,-0.3101449907,-0.4137935936,0.2377390116,0.2219977081,-0.1553967595,0.0707510263,0.1623302251,-0.2955522835,0.1464754194,0.5265617967,0.1610320657,0.0471725352,-0.1688614488,0.0138867684,0.3292209804,-0.0358106829,-0.0913413092,0.1298604906,0.2126795799,-0.4804819226,-0.0339946151,-0.0184382349,-0.104944706,0.3532821238,-0.2245303243,0.2486274093,-0.1422259361,0.5203970075,-0.1859227866,0.0720942095,0.0460008495,-0.1357634366,-0.6475103498,0.2183962017,-0.1087092757,0.0747861862,0.0200169627,-0.2638200819,-0.0101508098,-0.1398228854,0.1454683095,0.4622712135,0.0498209037,-0.4014340639,-0.2744064629,-0.5686835647,0.3184739053,0.2554924786,0.2237138748,-0.039021302,0.276802361,-0.0909414589,0.0218144972,0.1413944513,-0.0164726935,0.1134108528,0.1315147132,-0.5518431664,0.1067018509,-0.1301143765,-0.080223456,0.0674612969,0.3303523958,-0.1294987053,-0.1872288883,-0.076754503,0.1827464253,-0.288012445,-0.0719876811,0.3570813835,0.3660238981,-0.0069942293,0.1416062862,0.0799733698,-0.129911229,-0.2005256861,-0.0048077665,-0.0236915667,-0.0690515935,-0.2420410216,0.4297691882,-0.442310214,-0.5072241426,-0.4338132143,0.2354567796,0.0585268699,-0.1644812375,-0.3899991214,0.3686194122,-0.2135487348,0.1273375452,0.2562362254,0.3571557701,0.0298299789,0.3313847482,-0.4535017908,-0.1204369664,0.7448203564,-0.5261362791,0.3190010786,-0.261610806,0.3419785798,0.45168221,-0.3614073098,-0.5016885996,0.007907927,0.1330846846,-0.0542376786,-0.1616004705,0.0992287844,-0.3082868755,0.1013120189,-0.0707926676,0.2413412184,0.1822642386,0.0541245043,-0.0666989386,-0.1139496788]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"But no, since \r\n`\r\n metric = load_metric(metric_name)\r\n`\r\nis called for each process, the race condition is still there. So still getting:\r\n\r\n```\r\nValueError: Error in finalize: another metric instance is already using the local cache file. Please specify an experiment_id to avoid colision between distributed metric instances.\r\n```\r\n\r\ni.e. the only way to fix this is to `load_metric` only for rank 0, but this requires huge changes in the code and all end users' code.\r\n","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":76,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nBut no, since \r\n`\r\n metric = load_metric(metric_name)\r\n`\r\nis called for each process, the race condition is still there. So still getting:\r\n\r\n```\r\nValueError: Error in finalize: another metric instance is already using the local cache file. Please specify an experiment_id to avoid colision between distributed metric instances.\r\n```\r\n\r\ni.e. the only way to fix this is to `load_metric` only for rank 0, but this requires huge changes in the code and all end users' code.\r\n","embeddings":[-0.0743820369,0.0189331546,0.0986257046,0.2385591269,0.0015655096,0.0210894197,0.1744567901,0.2997217178,0.3662584722,0.1215545535,0.0272279736,0.1309887916,-0.3144579828,-0.0137835871,-0.0090082521,-0.0354207791,-0.0174290687,-0.0118403193,-0.2157492489,-0.1407078803,-0.3549723625,0.4909443557,-0.094041802,-0.0196759515,-0.4025662541,0.1862027049,-0.1208144575,0.1642288417,-0.0712104589,-0.6548614502,0.3760336339,0.2119780928,0.0550227165,0.5668932199,-0.0001228389,-0.1647618562,0.2224575281,-0.1654272527,-0.0895339325,-0.2089778632,0.0419953763,-0.3018239141,0.4746413529,-0.1672160625,-0.0354671106,-0.066387251,0.0834047124,-0.4613867998,0.4682587683,-0.0254345387,0.0765321851,0.3800902069,-0.1043446735,-0.2316598892,-0.0495066196,-0.1613818258,0.0241747182,0.4295062721,-0.0905898958,0.0036928728,-0.3325567842,0.2633509636,0.1676269174,0.0209394712,0.2453759462,-0.0559759326,0.064929001,-0.1188937649,-0.1050406396,0.3209834695,0.1520304829,-0.1883409619,-0.2298358232,0.0978398696,-0.1813545078,-0.5551991463,0.1535269469,-0.0972424373,0.0039800592,0.0392485633,-0.0380287506,-0.0952162445,-0.0568385944,0.056629993,-0.1502937675,0.0717534721,-0.1363277435,0.0569492243,-0.0620997027,0.2655829191,-0.3586181104,0.3487266302,-0.0165708736,0.1322408617,-0.4776332974,0.1030918807,0.233364895,0.0192754772,0.063808322,0.3710057735,0.0047831195,0.1104969233,0.4419555068,0.0989984944,-0.0648535118,0.4883598387,0.2653134167,-0.097262986,0.172475934,0.1843720973,-0.1093547717,-0.3044114709,0.1303544492,-0.2305703312,0.3590710759,0.1428924501,0.0501234122,-0.4300651252,0.0268946644,0.3125571609,-0.0175362881,-0.1046422049,0.1151105165,0.3204920292,0.0802922249,0.0376737975,0.2430389971,0.3135303259,-0.2338243127,0.0523923822,-0.4096278548,-0.209206,-0.0599776506,0.3604540527,0.0806869939,-0.0867773667,0.6622604728,-0.2452789098,0.4363716841,-0.0098047946,0.2046338618,0.1801846921,-0.0522345975,0.335319519,0.0205427725,0.2025182992,0.4229381979,0.1180685014,-0.2378182411,-0.222247839,-0.1965861768,-0.4890339077,0.1327867061,-0.0115742479,-0.4982536137,0.2183241993,0.2810570896,0.236883074,-0.0750576779,-0.1865112633,0.1740934998,0.0632326603,-0.1740942299,-0.0775207281,0.5013920665,0.8000811934,-0.3106375635,-0.4090100527,0.0621440969,0.0198931396,0.059744738,0.1834789366,-0.0264085419,0.1122234389,-0.2398672551,0.043061886,0.4500413835,-0.3818278015,-0.44047153,-0.0102439271,-0.4845240712,-0.0927990898,0.1529062241,0.0130148763,0.1167048067,0.1743211001,0.0596258231,-0.2304722071,0.1583907753,-0.2232081443,-0.4099032879,-0.3159253597,0.0827811882,0.0175182074,0.351108253,0.0312786065,-0.0216585081,0.0502975211,0.0938851088,0.2171548158,0.1900270581,0.2495134324,0.2989781201,0.0074488814,0.2794372141,0.0626640096,-0.1863881797,0.3330546618,-0.3847153783,0.1210732087,0.1558227986,0.0101862922,-0.2554594278,-0.233665213,-0.2074742615,-0.3817099631,0.0072999345,-0.1816835999,0.2119551748,0.2074861526,-0.2354964465,0.1523828954,0.0227620788,0.2163387686,-0.4231482446,0.0148601681,-0.0297127701,-0.331325978,-0.0202226397,0.0606493354,0.0434454679,0.0521167554,-0.2072704732,0.4505540133,-0.2041242868,0.271848917,0.3489960134,0.1949971169,-0.073145166,-0.3825898468,-0.0086809983,0.1035968363,-0.0373466983,0.003466916,-0.2115324885,0.3656711578,0.0740349218,0.1468586773,-0.351149112,0.2701358497,-0.0933459401,-0.1494020373,-0.5120645761,-0.0564063489,0.3746648431,-0.2972075343,0.3142363429,-0.1348352283,0.0955447927,-0.0313753709,0.2216869891,0.2303462178,0.1324336529,0.0526792556,-0.2068916112,-0.0816363618,0.0431526639,-0.2377364486,0.377917558,0.1831710339,0.2262164205,0.080045715,0.1042996049,-0.1794141382,0.0618315712,0.0174040738,-0.0939754769,0.1558281928,0.2129647136,0.1486931294,-0.3308373392,-0.0466651432,-0.2257911265,0.0777691454,-0.3252052367,0.0463929921,-0.1676025093,0.3622933328,0.1042036414,0.0760788769,-0.320507735,-0.200661242,0.2137132585,-0.1206795275,0.0558301024,0.3729785383,-0.3125834167,0.5460637808,0.1275182664,-0.1835206598,-0.2804273367,-0.2134806514,-0.0769835263,-0.1774238348,-0.1685278118,0.1996259987,0.1353262067,-0.1900614798,0.0364903957,0.0679248273,-0.2923831046,0.0072688437,0.0544490032,0.6009891033,0.2198510915,0.1127616242,-0.3618911207,0.0661276281,0.4660863578,-0.4653955102,-0.2988180816,0.1108197719,-0.005610731,0.2263551503,-0.2566636801,-0.3030112386,0.1503475606,-0.3555740416,0.6306397915,0.0361626931,0.0040374417,-0.1828644872,-0.0212518219,0.1606975347,-0.145521313,0.1835410893,-0.3220936954,-0.7370195389,0.3779354692,-0.0818951651,-0.2262859195,-0.0287562869,0.1227793247,0.3893444538,-0.0262093954,-0.5944405198,-0.4810172617,0.0726495683,0.2060054392,0.0201442447,0.0549132936,0.5922463536,-0.0786325336,-0.0039966004,-0.0683099702,-0.114582628,0.5944651961,-0.1417892873,0.1444726586,-0.0960227996,0.4891734719,0.1768221408,0.9024285078,0.5595239401,0.3208563328,0.285921365,0.1434484273,0.3294373453,-0.0935019553,-0.1837992519,0.3400885463,0.2125222236,-0.1073298231,0.2406834662,0.1164697483,0.1155291945,-0.1355736107,-0.030692596,-0.283249706,-0.2555078566,-0.1000093296,-0.1111311689,0.3113298118,-0.013860696,-0.0408909954,-0.2783927619,0.0155795161,0.4612799287,0.426333487,0.2830251753,0.2465917319,-0.2043099552,0.1160475016,-0.60267061,0.2750054598,-0.422406733,0.0926009417,-0.156595245,-0.229576692,0.088854529,-0.0187624302,0.5770187378,0.1072899178,-0.1236119047,0.0174960624,-0.142274037,-0.4782928824,-0.0497222953,-0.0212901179,-0.0193778574,0.2601453066,0.3804856837,-0.4534812272,-0.2462220043,-0.0953102559,-0.0955966935,-0.3262084723,-0.3852430582,-0.4633516669,-0.0032970144,-0.1292663515,0.2367421091,-0.2321139723,0.0642477423,-0.0269815624,0.0604758412,-0.1549195349,-0.0765350088,0.0366816521,0.1297768056,0.3405137062,-0.2411499023,0.200292021,0.524379611,-0.2138910443,0.0506991893,0.381957382,0.0767322332,-0.0481712408,0.3114059865,-0.0057785995,0.2684099376,0.5744407773,-0.2220171839,0.066104427,-0.0402308553,0.3440594971,-0.1733614057,-0.0811486244,0.2343105078,-0.1451415867,0.3937273324,0.1732892245,0.2727482319,0.2667104006,-0.1511027962,0.3070549965,-0.0249732826,-0.0635656938,0.3386529088,0.0756162927,0.9486906528,0.1284909546,0.0594738014,0.1395249665,0.1365728229,0.3250000179,0.0548069216,0.114858456,-0.2098522186,-0.2724112272,0.0157611705,-0.1055022627,0.0776122957,-0.2158324569,-0.1212308705,0.4221641123,-0.3239582181,-0.171465382,-0.2416449338,-0.1891413331,-0.2325046211,-0.2224490792,0.0643667579,0.0192141719,0.0016657843,0.5368846655,-0.2295260429,-0.1801932454,-0.2658546865,-0.2863274217,-0.0787065923,-0.1080938429,-0.2214947045,-0.0420682617,0.1932160407,-0.1637636125,-0.2138139158,0.2345352769,0.4904159307,0.1107980236,-0.3359957039,-0.0386441052,0.1185732037,0.0472981967,0.1092753857,-0.0165890399,0.1812068373,-0.0596783422,-0.0700883269,0.0230897628,-0.2034308761,0.029456608,0.1439750046,-0.1831580698,-0.0728730336,0.0270324051,0.0658264235,0.126166746,0.149388209,-0.2243226916,0.0851812437,0.2755255401,-0.2201800346,0.0531320386,0.1045536622,-0.2454176992,-0.1072542071,0.6161940098,-0.1266837269,-0.0828660056,0.3956391811,-0.0293205325,-0.028158173,-0.0929279327,0.0045109889,0.4716783166,-0.7318607569,0.0216784775,-0.00432728,-0.018871462,0.0596460253,0.2084546238,0.431427896,-0.1141748205,-0.019610649,-0.1784307808,-0.4242085516,0.2923364639,-0.2917732298,0.1231569424,-0.4502267838,0.1341909468,-0.1835298836,0.293942064,-0.2639543116,0.1568065435,-0.2775064707,-0.1364926547,-0.1669444144,-0.0555276386,0.3321179748,-0.0281812567,-0.0911530107,0.1156947613,-0.3989805877,-0.0762549713,-0.2312700599,0.1596203595,-0.1257577538,0.0551049933,0.0076610632,0.1867506802,-0.0734079033,-0.1189985871,0.0783567056,0.251432538,0.0211228114,0.076216273,0.2156767994,0.1833855361,-0.1267372668,0.1507460475,0.1054920256,-0.1982359886,-0.1485909522,-0.0037620352,-0.3089409769,-0.0378239602,0.20452784,0.0469588488,-0.1688239574,0.1653288156,0.0948095396,-0.0197555274,-0.337571919,0.0840761289,0.4339145422,0.1084734946,-0.1139044762,0.017161401,-0.0484705679,0.094181098,-0.1545313299,0.2720601559,0.03887292,0.3606870174,0.0561297983,0.2945477664,0.2639497817,-0.0546168573,0.1945337206,0.2769582868,0.0772711709,-0.1692988276,0.3492948115,-0.2080000043,0.1232765317,0.2905894518,0.6146329641,0.1501113027,0.1940245479,-0.1213558912,-0.0478095114,0.1758172512,-0.1244245619,0.3346286714,-0.3251515627,-0.1684168577,-0.2493241131,0.03564246,-0.0446152911,-0.1554951072,-0.4124151468,0.6928099394,-0.3596712649,-0.3196227551,-0.2396950126,0.2137310207,-0.16219154,-0.1622505337,-0.1247979626,-0.1749601811,-0.1221320033,-0.129580155,0.255628556,-0.0027531607,0.1339078993,0.0584408008,-0.012934654,-0.1976453066,-0.5636003613,0.3360029757,0.2369899899,-0.1671219915,0.0380651355,0.2539327443,-0.3203011453,0.0502868034,0.6612645388,0.1400390416,0.020517474,-0.1053515002,-0.0068349526,0.3291796744,0.0089210654,-0.0917339996,0.1179396436,0.212397933,-0.4631313682,0.0149258077,0.0473552644,-0.1401910037,0.4126127958,-0.2860366702,0.3127577007,-0.252772361,0.4388633966,-0.2093796134,0.1386310458,0.0330520533,-0.2090283483,-0.6382529736,0.2721561491,-0.1211095229,0.005566726,0.0730724558,-0.3161962032,0.0315936059,-0.0653400049,0.1874742955,0.3362928033,-0.0472614728,-0.3999146521,-0.3157871664,-0.6288834214,0.3125652969,0.1188525259,0.3920563161,0.0013999417,0.3996372223,-0.1393110454,0.1077261418,0.08064688,0.0115249399,-0.1084975451,0.0430235416,-0.5370870233,0.2043398321,-0.120561108,-0.0920396149,0.0809262618,0.2269995064,0.019760184,-0.1926590204,-0.003235905,0.2365329117,-0.1842131317,-0.032818757,0.225468114,0.3922176361,0.0917389691,-0.002602176,0.0817171112,-0.0382619128,-0.2293901145,-0.122765556,-0.1318761855,-0.0989051834,-0.3089475036,0.4359286726,-0.3517574966,-0.4784049392,-0.3654709458,0.2429835796,0.0612316169,-0.2057872713,-0.2005639821,0.2870566845,-0.0907272995,0.018358646,0.2171033025,0.3107776344,0.0868568867,0.3682156801,-0.3748184741,-0.1778047085,0.7422968745,-0.5126376748,0.1951844841,-0.2612099349,0.4384709895,0.476701051,-0.3749290705,-0.3257740438,-0.0008096767,0.1407495141,0.0945585519,-0.1339272559,0.1337970346,-0.4526482821,-0.0406459905,-0.0503086448,0.1235828176,0.143606171,0.0090163285,0.0242891163,-0.1226512715]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"OK, here is a workaround that works. The onus here is absolutely on the user:\r\n\r\n```\r\ndiff --git a\/examples\/seq2seq\/run_seq2seq.py b\/examples\/seq2seq\/run_seq2seq.py\r\nindex 2a060dac5..c82fd83ea 100755\r\n--- a\/examples\/seq2seq\/run_seq2seq.py\r\n+++ b\/examples\/seq2seq\/run_seq2seq.py\r\n@@ -520,7 +520,11 @@ def main():\r\n\r\n # Metric\r\n metric_name = \"rouge\" if data_args.task.startswith(\"summarization\") else \"sacrebleu\"\r\n- metric = load_metric(metric_name)\r\n+ import torch.distributed as dist\r\n+ if dist.is_initialized():\r\n+ metric = load_metric(metric_name, num_process=dist.get_world_size(), process_id=dist.get_rank())\r\n+ else:\r\n+ metric = load_metric(metric_name)\r\n\r\n def postprocess_text(preds, labels):\r\n preds = [pred.strip() for pred in preds]\r\n@@ -548,12 +552,17 @@ def main():\r\n # Some simple post-processing\r\n decoded_preds, decoded_labels = postprocess_text(decoded_preds, decoded_labels)\r\n\r\n+ kwargs = dict(predictions=decoded_preds, references=decoded_labels)\r\n+ if metric_name == \"rouge\":\r\n+ kwargs.update(use_stemmer=True)\r\n+ result = metric.compute(**kwargs) # must call for all processes\r\n+ if result is None: # only process with rank-0 will return metrics, others None\r\n+ return {}\r\n+\r\n if metric_name == \"rouge\":\r\n- result = metric.compute(predictions=decoded_preds, references=decoded_labels, use_stemmer=True)\r\n # Extract a few results from ROUGE\r\n result = {key: value.mid.fmeasure * 100 for key, value in result.items()}\r\n else:\r\n- result = metric.compute(predictions=decoded_preds, references=decoded_labels)\r\n result = {\"bleu\": result[\"score\"]}\r\n\r\n prediction_lens = [np.count_nonzero(pred != tokenizer.pad_token_id) for pred in preds]\r\n```\r\n\r\nThis is not user-friendly to say the least. And it's still wasteful as we don't need other processes to do anything.\r\n\r\nBut it solves the current race condition.\r\n\r\nClearly this calls for a design discussion as it's the responsibility of the Trainer to handle this and not user's. Perhaps in the `transformers` land?","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":233,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nOK, here is a workaround that works. The onus here is absolutely on the user:\r\n\r\n```\r\ndiff --git a\/examples\/seq2seq\/run_seq2seq.py b\/examples\/seq2seq\/run_seq2seq.py\r\nindex 2a060dac5..c82fd83ea 100755\r\n--- a\/examples\/seq2seq\/run_seq2seq.py\r\n+++ b\/examples\/seq2seq\/run_seq2seq.py\r\n@@ -520,7 +520,11 @@ def main():\r\n\r\n # Metric\r\n metric_name = \"rouge\" if data_args.task.startswith(\"summarization\") else \"sacrebleu\"\r\n- metric = load_metric(metric_name)\r\n+ import torch.distributed as dist\r\n+ if dist.is_initialized():\r\n+ metric = load_metric(metric_name, num_process=dist.get_world_size(), process_id=dist.get_rank())\r\n+ else:\r\n+ metric = load_metric(metric_name)\r\n\r\n def postprocess_text(preds, labels):\r\n preds = [pred.strip() for pred in preds]\r\n@@ -548,12 +552,17 @@ def main():\r\n # Some simple post-processing\r\n decoded_preds, decoded_labels = postprocess_text(decoded_preds, decoded_labels)\r\n\r\n+ kwargs = dict(predictions=decoded_preds, references=decoded_labels)\r\n+ if metric_name == \"rouge\":\r\n+ kwargs.update(use_stemmer=True)\r\n+ result = metric.compute(**kwargs) # must call for all processes\r\n+ if result is None: # only process with rank-0 will return metrics, others None\r\n+ return {}\r\n+\r\n if metric_name == \"rouge\":\r\n- result = metric.compute(predictions=decoded_preds, references=decoded_labels, use_stemmer=True)\r\n # Extract a few results from ROUGE\r\n result = {key: value.mid.fmeasure * 100 for key, value in result.items()}\r\n else:\r\n- result = metric.compute(predictions=decoded_preds, references=decoded_labels)\r\n result = {\"bleu\": result[\"score\"]}\r\n\r\n prediction_lens = [np.count_nonzero(pred != tokenizer.pad_token_id) for pred in preds]\r\n```\r\n\r\nThis is not user-friendly to say the least. And it's still wasteful as we don't need other processes to do anything.\r\n\r\nBut it solves the current race condition.\r\n\r\nClearly this calls for a design discussion as it's the responsibility of the Trainer to handle this and not user's. Perhaps in the `transformers` land?","embeddings":[-0.0640703589,-0.0787641034,0.0965643525,0.2534787655,-0.0056881383,-0.0212866962,0.3123223782,0.2638015747,0.2110244334,0.2581807971,0.0451734066,0.3624677956,-0.3526070714,-0.1786394417,0.0343574695,-0.034716554,-0.0441434048,-0.1339790523,-0.3583231866,-0.1239942759,-0.190798983,0.4987677038,-0.0344169177,-0.0297840927,-0.3931267858,0.224177435,-0.164009124,0.1788709909,0.0643501431,-0.5618361831,0.3226022422,0.1360853016,0.0452935137,0.5764806867,-0.0001236324,-0.1678349376,0.1872864068,-0.2450377643,-0.1874425858,-0.271746397,0.2271195054,-0.3114353418,0.4968193769,-0.2364772558,-0.1577860862,0.0378682949,-0.0054120263,-0.3434016407,0.4236991405,0.0571904667,0.085931316,0.3330073953,-0.032325916,-0.1901235878,0.1212329939,-0.0653862208,-0.0368981287,0.391304642,-0.0438003577,0.0700681061,-0.2391371131,0.2996984422,0.1755198389,-0.0402260795,0.314748466,-0.0494450517,-0.2011662573,-0.0775264129,-0.105209291,0.308021009,0.2359141856,-0.223406136,-0.2375836819,0.0531364717,-0.1129199192,-0.3445173502,0.1445093304,-0.0051630242,-0.0187254827,0.0235215779,0.0591396205,-0.1390450746,-0.0627024695,0.0998460129,-0.0452240445,0.0139156533,-0.1335917711,0.1054865494,-0.0017543795,0.1352349818,-0.3275977373,0.306905359,0.0391775332,0.1289857626,-0.3677838743,0.1243339702,0.2337081581,0.0210840479,0.0918095335,0.4560725987,0.0586389378,0.0463430211,0.4054613411,0.1031668037,-0.1804311275,0.4910764992,0.2882263064,0.1206449419,0.1309707463,0.2508909106,-0.0607221946,-0.2949403524,0.1342707276,-0.2586985826,0.2855698466,0.2531366348,0.0631746277,-0.4151259363,0.0464722887,0.1691827774,-0.0770844817,-0.1774351299,0.1314891577,0.4097194076,0.1488403976,0.0084704608,0.3483400047,0.3792473972,-0.1620044112,0.005470606,-0.3775833845,-0.1619015038,-0.0832369179,0.3270745277,0.1530539989,-0.1119453609,0.6152184606,-0.2334931195,0.4344537258,-0.0676151067,0.2086048126,0.2173037529,0.0877950042,0.1757381558,-0.0002522572,0.1726161093,0.3499711752,0.3042697906,-0.2953206599,-0.1659086645,-0.1075638756,-0.5697390437,0.1044747531,-0.0256729499,-0.3686076999,0.252402693,0.3646577597,0.1765206307,-0.0878772363,-0.1766688824,0.2183123082,0.0424640067,-0.1899272203,-0.1322890371,0.5401469469,0.795589745,-0.223470822,-0.3928968012,0.0861313418,-0.0302689224,-0.0688859746,0.1711060703,-0.1304824352,0.1734703034,-0.3124321699,0.0132106543,0.4526499212,-0.4541595578,-0.5230979919,-0.0551201962,-0.3473708034,-0.1022441834,0.1810635328,0.0192643367,0.1447815597,0.1427844316,0.1297154278,-0.1455388516,0.2052158266,-0.2678942382,-0.3874981999,-0.2713224292,0.0696757361,-0.0526352227,0.2598974407,0.019360641,0.0312664509,-0.0005676097,0.1223189384,0.112574324,0.1530454457,0.3094593287,0.1673349887,0.1094968691,0.2271713763,0.0797586888,-0.2535483241,0.280608058,-0.2641236186,0.096486561,0.0547574796,-0.0591282099,-0.2017363757,-0.191734314,-0.1189767569,-0.4788504839,-0.0093121016,0.0025924593,0.1778894514,0.098006621,-0.0833217427,0.0272526108,0.0022785261,0.2093143463,-0.4971334338,0.0486030243,-0.1358307302,-0.3562149107,-0.0538285002,0.162473321,0.082434766,0.0275184531,-0.2172619998,0.4142546058,-0.2254364938,0.2963485718,0.2505712509,0.0866480023,-0.1558147371,-0.4436963201,0.0706837699,0.1534367651,-0.0243405979,0.0595748425,-0.2230667919,0.4005185962,0.057605084,0.088164255,-0.3915153742,0.320607543,-0.0622925423,-0.1256732643,-0.5161961913,-0.065331772,0.4644509256,-0.2338357419,0.306891948,-0.1485267282,-0.1147532538,0.0304505322,0.0552103035,0.1526011825,0.2161269635,0.0492043793,-0.2106391788,0.0539288856,0.0803134516,-0.2700403631,0.4026062787,0.2301347256,0.2446660399,0.0396017358,0.2069373727,-0.1649561971,0.0762510151,-0.0633525997,-0.1690059602,0.1789561957,0.2994227707,0.16940552,-0.2687003314,-0.0116007859,-0.3858012557,-0.0177335758,-0.316115737,0.1085886434,-0.1061817184,0.3600185215,0.0628945455,0.027881762,-0.2497003078,-0.2115068734,0.2172032595,-0.128764689,-0.0082145529,0.3403942883,-0.1438351423,0.4247652292,0.1005562171,-0.2894504964,-0.2572137713,-0.1998008192,-0.0760467201,-0.1793352067,-0.2656765878,0.0456999801,0.218057245,-0.1610600501,0.0553274378,-0.1083599702,-0.2996193469,0.0607741922,0.0864420161,0.5465068817,0.294896543,0.0005504977,-0.1870801449,0.0301565975,0.4708809257,-0.463945508,-0.2469661385,0.0697957352,0.0687668622,0.3005344272,-0.3028265834,-0.4108238518,0.0369482189,-0.4303351343,0.6223106384,0.0505601317,0.0192916226,-0.2766919434,0.0470691025,0.1442534924,-0.1902227104,0.3671061695,-0.3098250329,-0.6642648578,0.4911518693,-0.153203994,-0.2969996631,-0.0493560918,0.0273516737,0.4338546991,0.0113272937,-0.6472849846,-0.427665025,0.0734205619,0.3311350048,-0.0820795298,0.1234300286,0.5292959213,-0.1036516801,0.0432882756,-0.0859871879,-0.1932816356,0.4761613905,-0.2479908019,0.2904284596,-0.0502362959,0.4710008502,0.2042387277,1.0362510681,0.4318812191,0.1980848312,0.2408500165,0.0621297508,0.3678739667,-0.1329009831,-0.2530627847,0.3594072163,0.215783909,-0.065166086,0.1750031561,0.1401123554,0.1193047762,-0.1175367534,-0.1354067624,-0.3230404556,-0.3765500784,0.0584392063,-0.3119445443,0.2293654233,-0.102417089,0.1097870767,-0.2834073007,-0.0170254987,0.4524648488,0.3811736405,0.2687610984,0.2686800659,-0.1061641872,0.1557906419,-0.6143527627,0.3195683658,-0.4150725901,0.0204222631,-0.196459353,-0.2585929036,0.1327742785,0.0286134817,0.5701928735,0.1704184413,-0.0772307515,-0.0223533362,-0.2900950313,-0.572894454,-0.1306662261,0.0335995331,0.0039162594,0.3430426419,0.2983974814,-0.4971288145,-0.2174235731,-0.1325904429,-0.0143522611,-0.2948710322,-0.2986010313,-0.3033482134,0.0376956053,-0.3638680875,0.1257952899,-0.1481666416,0.00197746,0.0124056628,0.0938546509,-0.1730405688,-0.0333019346,-0.0419658236,0.0819390863,0.445299834,-0.3680725098,0.2138540745,0.4355837703,-0.0814052597,0.0465961471,0.4527378082,0.0202477966,-0.0322044529,0.249416396,-0.1702366769,0.1760369986,0.5683686137,-0.2360541075,-0.0065520392,-0.1683749557,0.3352352083,-0.1813479513,-0.094721213,0.2148928046,-0.1399389654,0.4028420746,0.0797779486,0.1491568536,0.2951569259,-0.1458461881,0.308953315,-0.0896981433,-0.1584067643,0.3474822044,0.0283556525,1.0431015491,0.1858016402,0.1939702779,0.0941930264,0.0594871193,0.3286902606,0.0322481021,0.0423596241,-0.2455371171,-0.1855905354,0.0552304313,-0.1338486224,-0.0892541856,-0.2128922194,-0.2421695441,0.4602898061,-0.3985310793,-0.1187767833,-0.2751251161,-0.216184333,-0.3653043509,-0.1540292054,0.1623811275,0.0106246006,0.0389807895,0.5765026808,-0.206303075,-0.1709365398,-0.3550463617,-0.4208046496,-0.1701827496,-0.0027211853,-0.2345432937,0.0126432553,0.2596086264,-0.3003920317,-0.127521053,0.3978976905,0.3753563166,0.1119137406,-0.2954158783,-0.0671827942,0.0727764145,0.0780795887,-0.0095543237,-0.0105061559,0.202223599,-0.0534298383,-0.0567144565,0.0370832346,-0.1553322524,0.0644343272,0.0799800009,-0.1045582816,-0.0424971394,-0.1324025244,0.0671937838,0.1465180516,0.0625550076,-0.2441199422,0.0631633103,0.3170138001,-0.3118456304,-0.0164044984,0.1300032288,-0.1914991736,-0.0526274256,0.6006056666,-0.2180666476,-0.0574906059,0.4439824224,0.0863898695,-0.090012081,-0.0896645561,-0.0095595466,0.4701999128,-0.8980280161,0.0632024109,-0.0824749097,-0.0375495255,0.0195322521,0.2173477113,0.3759930432,-0.0564713068,-0.0513753816,-0.1332810074,-0.4421541989,0.2491102368,-0.2822010517,0.2810609937,-0.3498198688,0.0752844065,-0.1540609896,0.2654627562,-0.264413178,0.1684677899,-0.1766667515,-0.039287474,-0.0815755948,0.0030675286,0.3155998588,-0.0181364007,-0.0913404152,0.1194010451,-0.3476508558,-0.0852037594,-0.2056672871,0.1578653902,-0.1684205979,0.1116118431,-0.1268759519,0.177016601,0.0223468654,-0.1894455701,0.1027841568,0.270388484,0.0493438318,0.2141861618,0.1147242486,0.238899827,-0.0491266921,0.2069228292,0.0891241804,-0.0931501389,-0.127140224,0.0269986,-0.3199724853,-0.0794886649,0.176806286,0.0992085263,-0.0745762736,0.226169765,0.0143868802,0.0950899497,-0.3218647838,-0.0831813514,0.3944787681,0.0252385996,-0.2240364999,0.0014872645,-0.0390673876,0.0907803923,-0.0961546525,0.2872498631,-0.0495013297,0.2823807299,0.0991164222,0.2835998833,0.2086952925,-0.0566607267,0.0748728812,0.1841032654,-0.0485476106,-0.1143180504,0.387547493,-0.066637598,0.0531389341,0.2888018787,0.5470498204,0.1314865798,0.2106275111,-0.1264861524,-0.0628759637,0.3586797118,-0.0406755172,0.2327472717,-0.068258591,-0.1420321018,-0.1085531041,-0.1151635572,0.1268028915,-0.2630985379,-0.3883004189,0.6905946732,-0.1910868734,-0.2771374583,-0.0456669778,0.2752059698,-0.1335234344,-0.1646227539,-0.1777421236,-0.1472785026,-0.1358105242,-0.1351949424,0.2524801493,-0.0429370552,0.1753880084,0.1447418481,0.0329223424,-0.2331388146,-0.3482346833,0.2942021787,0.1979301423,-0.1560441107,-0.0297636501,0.2149160504,-0.2448703498,0.1404075772,0.5754084587,0.2674567103,-0.0427276529,-0.0580169484,0.0885282233,0.2895345688,0.0024859894,-0.1617273539,0.0645731688,0.1674539149,-0.4141650796,0.0965312645,0.0538550839,-0.1895053238,0.4136652052,-0.2827782929,0.2279515117,-0.1521187723,0.4434229136,-0.2251646966,0.1118193418,-0.0354707614,-0.1718774736,-0.620079875,0.2504784167,-0.1109717712,0.0327468663,0.09214136,-0.2865742147,0.0349231884,-0.0817037746,0.1180626377,0.3765927553,0.0629151911,-0.3114416897,-0.2652311325,-0.5255100131,0.3029744029,0.2050180435,0.2848012447,-0.0820878968,0.3575700521,-0.0763025284,0.1108622104,0.2054360658,-0.0856378153,0.0611880273,0.1663652658,-0.5453774929,0.0388831981,-0.1388038397,-0.0357953906,0.1995832473,0.2482260913,-0.2193721086,-0.1935058534,0.0216775034,0.1656119823,-0.3292373419,-0.1445855796,0.2926346064,0.4417029023,0.0879337192,0.0814847648,0.0020326278,-0.0957131162,-0.1901303977,-0.1178382859,-0.0816707462,-0.11695721,-0.2638622522,0.3734414876,-0.3695460558,-0.4859572053,-0.4547900259,0.2592763007,0.0549923964,-0.1221695542,-0.2049329281,0.334738642,-0.1531445086,0.0585971884,0.1520989388,0.3737296462,0.0079819169,0.4060660303,-0.4049213529,-0.1603150815,0.7884654403,-0.4327733517,0.2788589299,-0.2969713807,0.356649369,0.4510624409,-0.3289364874,-0.4853739738,0.0553708076,0.1302498579,0.0357781611,-0.003226201,0.0996122882,-0.3638705015,0.0264715627,-0.0803087801,0.1829400361,0.1846149862,0.0688116252,-0.0152423708,-0.1749408394]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"I don't see how this could be the responsibility of `Trainer`, who hasn't the faintest idea of what a `datasets.Metric` is. The trainer takes a function `compute_metrics` that goes from predictions + labels to metric results, there is nothing there. That computation is done on all processes \r\n\r\nThe fact a `datasets.Metric` object cannot be used as a simple compute function in a multi-process environment is, in my opinion, a bug in `datasets`. Especially since, as I mentioned before, the multiprocessing part of `datasets.Metric` has a deep flaw since it can't work in a multinode environment. So you actually need to do the job of gather predictions and labels yourself.\r\n\r\nThe changes you are proposing Stas are making the code less readable and also concatenate all the predictions and labels `number_of_processes` times I believe, which is not going to make the metric computation any faster.\r\n\r\n","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":144,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nI don't see how this could be the responsibility of `Trainer`, who hasn't the faintest idea of what a `datasets.Metric` is. The trainer takes a function `compute_metrics` that goes from predictions + labels to metric results, there is nothing there. That computation is done on all processes \r\n\r\nThe fact a `datasets.Metric` object cannot be used as a simple compute function in a multi-process environment is, in my opinion, a bug in `datasets`. Especially since, as I mentioned before, the multiprocessing part of `datasets.Metric` has a deep flaw since it can't work in a multinode environment. So you actually need to do the job of gather predictions and labels yourself.\r\n\r\nThe changes you are proposing Stas are making the code less readable and also concatenate all the predictions and labels `number_of_processes` times I believe, which is not going to make the metric computation any faster.\r\n\r\n","embeddings":[-0.0986766294,0.0874648094,0.0754722878,0.1241439432,0.1474277377,-0.0701845735,0.3525932133,0.2356343269,0.1611245424,0.2098139673,0.000421444,0.214056015,-0.3455352783,0.1935886294,0.0740699172,-0.0832463354,-0.0075628199,-0.1163947061,-0.2869692147,0.0001233007,-0.261204958,0.4958016276,-0.1959237009,0.1165345833,-0.4939318895,0.0160643514,-0.1405585408,0.0436572582,-0.1571676433,-0.6827939153,0.1779743582,0.2850542068,0.0590012819,0.6136043668,-0.0001197262,-0.1275459379,0.1432348788,-0.0766494647,-0.1142276749,-0.101868622,0.0062036146,-0.3750232458,0.5041076541,-0.2027844936,-0.0246382952,-0.1871017665,-0.000907287,-0.3866353333,0.4851273,0.1960166842,0.1148971319,0.3724759221,-0.0633555204,-0.2893129587,-0.0304380469,-0.179746449,-0.0004408027,0.3063096106,-0.1180348918,0.1041778475,-0.2638342679,0.2987607419,0.1747556776,-0.1579250693,0.3985736668,0.0108563146,0.0237378236,-0.0750850588,-0.026344303,0.228096053,0.3640272319,-0.2081997395,-0.4027554691,0.0213231351,-0.1136104017,-0.4449650645,0.0867719799,-0.1544950157,-0.1124913022,0.017352581,-0.0512008294,-0.0918699801,-0.1205045357,-0.1065767035,-0.1400079429,0.0415050797,-0.1515364647,0.1079932898,0.0986205488,0.2940391898,-0.2250009924,0.4625895619,0.107808806,0.0874904841,-0.5861869454,0.051559452,0.1689260602,-0.137893185,0.026183527,0.3476356268,0.1393156648,0.1647456735,0.4749913812,0.0469469056,-0.0859803259,0.495952338,0.1158409417,-0.0413894244,0.2166435719,0.2328991294,-0.1324020177,-0.3073429167,0.1509372443,-0.2894469798,0.4075081646,0.1650965512,-0.0240274984,-0.3971653283,0.0111141028,0.2321667373,-0.1615175307,-0.3076352179,0.1635745168,0.3015667498,-0.0188990217,0.1546543688,0.2649487555,0.2677026987,-0.2153810561,0.0336757265,-0.3627006114,-0.1388621628,-0.2221658081,0.5311321616,-0.033613015,-0.1031351015,0.4908144474,-0.2240339071,0.3613184988,-0.114314951,0.1868821234,0.1512250751,0.0729259402,0.1448817104,-0.0949600264,0.0191628709,0.1993149817,-0.0007020317,-0.2345672101,-0.3136923611,-0.1904584765,-0.5587620735,0.0460788868,0.0113418596,-0.4121508598,0.168701455,0.1926000267,0.4390551746,-0.0838328302,-0.0827364922,0.1543191671,-0.0658370554,-0.2503941059,-0.0533928685,0.538708806,0.6572615504,-0.2708118856,-0.3530547917,0.1573342681,-0.0764822289,0.0305980127,0.1866123974,-0.0675412193,0.3315973282,-0.1351325661,0.0338992961,0.318702817,-0.3188119233,-0.2820299566,-0.1126638576,-0.2808205187,-0.0668343529,0.1820889562,-0.0332553238,0.2036238611,0.1777180731,0.0793947801,-0.2812536359,0.199252516,-0.2796541154,-0.3488397598,-0.2571011186,-0.0160810836,0.0350257978,0.4142493606,0.1447869688,-0.0155540835,0.2191104591,0.0989591479,0.0124070393,0.1865232289,0.1369633824,0.2173769474,0.0301675629,0.2122123092,-0.0668710098,-0.259488225,0.1811954081,-0.2249533683,0.2644392848,0.181269899,-0.1109844595,-0.1313766092,-0.0319106393,-0.0450462811,-0.3800312877,0.038146954,0.0193699189,0.2461988032,0.063606061,-0.1863696277,0.2601462603,-0.0229211096,0.1970857531,-0.3077025115,0.0054115532,-0.0493264832,-0.2693814635,-0.0472342819,0.2843676805,0.0135070439,0.0620762482,-0.1563108563,0.4471603036,-0.0971780792,0.1208168939,0.2460498959,0.2522501945,-0.1001493707,-0.2472563386,0.016244486,-0.01773897,0.0309684929,-0.0384248942,-0.1867370754,0.4048735499,-0.0063228183,0.1802697033,-0.1444068253,0.2286265343,-0.1040408611,-0.0146845682,-0.6104436517,-0.0441228263,0.1817787588,-0.358833164,0.2854744196,-0.0732167214,-0.1879296005,0.0873126239,0.2204131186,0.0784531012,-0.0388826355,0.1840817332,-0.1221867502,-0.1293780357,-0.0313919559,-0.0752816051,0.4372566938,0.198172912,0.265868485,0.136862576,0.1789865494,-0.2482047379,0.1564263254,0.1467076093,-0.19835338,0.0213169325,0.1952255666,0.1843651831,-0.2286498398,-0.1335852593,-0.2828978896,0.0912203491,-0.3479496241,0.1114817262,-0.1617618948,0.3325738907,-0.0256056022,0.0918035433,-0.1845480949,-0.2695588171,0.2477204055,-0.1749944687,-0.0806614682,0.3019294441,-0.1997397393,0.5645967126,0.0458001383,-0.0543940477,-0.1539081335,-0.1289447099,0.024809733,-0.1263793409,-0.1107395217,0.2380958647,0.144310236,-0.1409785599,0.0804135874,-0.1126384661,-0.3777036667,0.0318094678,-0.0719139948,0.6136100292,0.3686548471,-0.0671826899,-0.3057754934,0.0558075421,0.4334175289,-0.4616743624,-0.1821401119,-0.0158735793,0.0210346337,0.1119773015,-0.2284412086,-0.4090143442,-0.0333864987,-0.3924374878,0.6301872134,-0.0302866921,0.0650706887,-0.1423585862,0.0048807203,0.3127186894,-0.2132500559,0.2855618596,-0.3027739823,-0.6883821487,0.3615089357,-0.1526074857,-0.2784527242,-0.0922783613,0.1246677935,0.4610554278,-0.1010721028,-0.639030993,-0.48743999,-0.0583331324,0.1837807447,-0.0854850635,0.1019595116,0.5934659839,0.0402632505,-0.1249963343,-0.0401615165,-0.3275840878,0.6002625227,-0.2267073095,-0.0799337775,-0.0480876155,0.3807918429,0.1421984136,0.9527528882,0.5312842727,0.0835819617,0.2418905199,0.0609008297,0.4276578426,-0.1242132187,-0.2087624371,0.3594114184,0.0553975031,-0.0696334168,0.2196852118,0.0753751472,0.2284238786,-0.0930808038,0.0747022778,-0.1595893651,-0.3905131817,0.0516245253,-0.070996128,0.2339848727,-0.0320989117,0.0994047821,-0.2225835025,0.0105921747,0.3239010572,0.3276197314,0.2672769129,0.1773552895,-0.4027299881,-0.0141273271,-0.6917689443,0.2999631763,-0.3306577504,0.0550329648,-0.1726821065,-0.3012733757,0.1569828242,-0.0032647848,0.5102081895,0.1670262367,-0.325607419,0.0979515314,-0.2732122242,-0.3518054783,-0.0322374329,-0.1057447866,-0.0924548507,0.1303734183,0.4627729356,-0.3822813034,-0.4021385908,-0.0575436503,-0.0055291038,-0.2024351507,-0.2122659385,-0.2753818333,-0.1906089038,-0.1053456962,0.2593511939,-0.1286329925,0.1493752599,-0.0751169771,0.010561944,-0.1109575108,0.0294923633,0.0498249307,0.2006752193,0.3328441381,-0.2227474898,0.3211791217,0.5052848458,0.0824888274,0.1085553393,0.3950394392,-0.0712283105,-0.0762939155,0.2625710666,-0.0787664428,0.3977535367,0.5973274112,-0.3462336957,0.050473962,0.0018158519,0.4982075095,-0.1901556253,0.0729652122,0.2902306318,0.0637293011,0.3196473122,0.0237539317,0.2319914997,0.3206911683,-0.240091756,0.1433272511,0.0838350579,-0.2308394015,0.2619361579,0.1453868896,0.9682911634,0.0424495712,0.1907757521,0.224296689,0.0791933015,0.4160702825,0.0028481516,0.0480352975,-0.2100685239,-0.3447068036,0.0477675758,-0.0816003382,0.0021893855,-0.2576920092,-0.3582127094,0.3531533778,-0.1213833988,-0.0985353664,-0.2188064158,-0.0329664275,-0.0591558106,-0.3148884177,0.0249197632,0.0341158584,0.0702982247,0.262604475,-0.1506989896,-0.1671604663,-0.2209928036,-0.2636078298,-0.1547154784,-0.2230544835,-0.2821092904,-0.1693060696,0.2403291166,-0.0131516559,0.0276419204,0.2988243997,0.4105807245,0.0932678282,-0.3142998517,-0.1085325703,-0.1605774164,0.2130302936,0.1821775436,-0.0810614154,0.279545635,-0.131883353,-0.0807683691,-0.0057212049,-0.2108217031,0.0140894521,0.0495157987,-0.1452664584,-0.0465457439,-0.1562491655,0.1377622634,0.0298856311,-0.0652444884,-0.3307300508,0.0995554328,0.2345898002,-0.2563966513,0.2924687862,0.1632495224,-0.2699543536,-0.0011781426,0.4867061973,-0.0743930489,-0.0469132476,0.5342655778,0.0417936556,-0.0784231052,-0.1546751261,-0.0377027094,0.3748117089,-0.6188102365,0.2553607523,-0.0921285748,-0.1034337804,0.0324718542,0.2574296296,0.2556154132,-0.1610131711,-0.0511721857,-0.195092231,-0.3997427225,0.3095521033,-0.2061042041,0.2349155843,-0.4314939678,0.2607114017,-0.1639536172,0.2664735317,-0.3096978664,0.0712418854,-0.3173698783,0.0048189606,-0.0426048897,-0.0775404871,0.3540878892,0.0196155384,-0.0791099146,0.1467722803,-0.3419515491,-0.1227808148,-0.2055000514,0.1432722062,-0.1030794457,0.0037045388,0.1053148434,0.2841434479,-0.2889694571,-0.2163692117,0.1446666121,0.2201308757,-0.080318138,0.164986521,0.0227368008,0.1731150895,0.0217941459,0.1275438964,0.3193958402,-0.1167129055,-0.2552144527,0.1674158573,-0.3102363944,-0.1582752466,0.2583626509,0.2446475625,-0.053016562,0.2476225942,-0.0005029342,0.039084956,-0.4149771035,0.015203163,0.5152156353,0.2113122046,0.0148268444,-0.0132391714,0.0306367837,0.062968865,-0.2263818979,0.2248078138,-0.0241948441,0.1465190649,0.1827189028,0.3415971398,0.20041807,-0.0087897144,0.1131294742,0.1668815911,-0.0762682781,-0.0972013101,0.392945528,-0.0642776415,0.0135834282,0.3271149993,0.6098663807,0.0715303048,0.1015054062,0.1042543128,-0.1152170375,0.3192860186,-0.0454615764,0.3231517076,-0.1962460428,-0.020482948,-0.1880552471,0.1027822644,0.0078947982,-0.2648672163,-0.4510085583,0.7601667643,-0.2982218266,-0.4281980097,-0.1595361382,0.2162976265,-0.2203803509,-0.2502735853,-0.1180726439,-0.2273586988,-0.187136963,-0.1366971582,0.1595191956,0.0055200043,0.317237556,-0.0935515687,0.1166381985,-0.2562332153,-0.6462206244,0.1233225465,0.2115249783,-0.1426584572,0.0580332018,0.1994390786,-0.2637606561,0.1976294667,0.7228573561,0.2519356906,0.0261539277,-0.0786250755,0.0529665835,0.0839556232,0.0876960158,-0.0001555772,0.2521165311,0.2659590244,-0.6128060222,0.1696699262,-0.0114007443,-0.1481031775,0.4118888378,-0.2929542959,0.3268871605,-0.3496899605,0.495154202,-0.0461821854,0.0140403928,0.0405657478,-0.1425222307,-0.5717719793,0.2246554941,-0.0200581551,-0.0718957782,0.0829666406,-0.3604297936,0.0544083863,-0.0736921802,0.4507356286,0.2547645569,0.1308719814,-0.3133822083,-0.0882295296,-0.6293569803,0.0641285405,0.2504545748,0.312289238,-0.0201935321,0.2240371555,0.0060548577,0.1538954228,0.2459771186,-0.0620279834,-0.0188194253,-0.0547912419,-0.6547977924,0.2398648113,-0.1182163283,-0.0279670842,0.183256343,0.173082903,0.0813002586,-0.2520494461,-0.0084853498,0.244053632,-0.0625068918,-0.0889036655,0.2176330984,0.4322580695,0.1216438115,0.0933319703,0.0375456363,-0.0467441455,-0.278555721,-0.1413880885,-0.1935823113,-0.0083428165,-0.1165120751,0.4362838864,-0.3418883681,-0.2784533799,-0.389836967,0.207449019,0.0935021266,-0.0974027887,-0.1647613347,0.4062645733,-0.0399049819,-0.0485423878,0.2773154974,0.3931531012,-0.0078771627,0.3691285253,-0.4815252423,-0.2000590414,0.7874458432,-0.5091014504,0.162209183,-0.2533195615,0.3058442175,0.453786999,-0.3390708268,-0.6162970662,0.0236155055,0.1255334914,0.046710588,-0.1886403114,0.2845034897,-0.4269403815,-0.1252786517,-0.0973918959,0.2794103026,0.0648686066,0.0003610959,-0.0046301405,-0.2103040516]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"Right, to clarify, I meant it'd be good to have it sorted on the library side and not requiring the user to figure it out. This is too complex and error-prone and if not coded correctly the bug will be intermittent which is even worse.\r\n\r\nOh I guess I wasn't clear in my message - in no way I'm proposing that we use this workaround code - I was just showing what I had to do to make it work.\r\n\r\nWe are on the same page.\r\n\r\n> The changes you are proposing Stas are making the code less readable and also concatenate all the predictions and labels number_of_processes times I believe, which is not going to make the metric computation any faster.\r\n\r\nAnd yes, this is another problem that my workaround introduces. Thank you for pointing it out, @sgugger \r\n","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":139,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nRight, to clarify, I meant it'd be good to have it sorted on the library side and not requiring the user to figure it out. This is too complex and error-prone and if not coded correctly the bug will be intermittent which is even worse.\r\n\r\nOh I guess I wasn't clear in my message - in no way I'm proposing that we use this workaround code - I was just showing what I had to do to make it work.\r\n\r\nWe are on the same page.\r\n\r\n> The changes you are proposing Stas are making the code less readable and also concatenate all the predictions and labels number_of_processes times I believe, which is not going to make the metric computation any faster.\r\n\r\nAnd yes, this is another problem that my workaround introduces. Thank you for pointing it out, @sgugger \r\n","embeddings":[0.0017203506,0.0483337678,0.0741655752,0.0518631414,0.0032177684,-0.0890675113,0.2733488679,0.3028243482,0.2559508085,0.157837823,0.0953509733,0.1828419864,-0.3078093231,0.1192955226,0.0508190766,-0.0088097211,-0.0289928019,-0.1110823229,-0.2872489989,-0.0546385124,-0.2833441198,0.574783206,-0.1937366277,0.0275069624,-0.4175572991,0.180085063,-0.0967399552,0.0959717706,-0.0746753663,-0.6333745122,0.3431346118,0.1919608712,0.0437812433,0.4339032471,-0.00011918,-0.153466478,0.2158040702,-0.1456783116,-0.0593994819,-0.1108042374,0.0721555278,-0.3260098696,0.472986728,-0.146902591,-0.013560731,-0.0872117355,0.0065102926,-0.4241975546,0.3669990599,0.0343662463,0.1324311495,0.2945926785,-0.0216743201,-0.292167455,0.071168609,-0.148108542,0.0353575274,0.3607213795,-0.1339002997,0.0896664858,-0.2697998285,0.266195178,0.2188452482,-0.1134704798,0.3259405792,0.0134763289,-0.0287550781,-0.0886272714,-0.1024975181,0.2506771386,0.2891428173,-0.1953328699,-0.2615005672,0.0619682595,-0.2091754377,-0.379871279,0.1182642654,-0.1241420805,-0.0386012234,0.0183159318,0.0663574338,-0.1046716347,-0.0890833363,-0.0081161847,-0.0986792445,0.0158464666,-0.1301837862,0.0467522182,0.018719878,0.1890254319,-0.3581730425,0.4645082653,0.0015239388,0.1163826883,-0.4526612163,0.0794150531,0.2340442836,0.0034457755,0.1275369823,0.4340135157,0.0486906394,0.0966068804,0.4923918843,0.0736068636,-0.1517188698,0.5450241566,0.1368220001,-0.0942575857,0.2153577954,0.2788958251,-0.0592239425,-0.3349413276,0.2765874863,-0.2964963317,0.3546170294,0.1382145137,0.0487446971,-0.5286619663,0.0650700331,0.1770535409,-0.1106246486,-0.2184835076,0.1539464444,0.3556601703,0.0642174557,0.0728146583,0.3231688142,0.2419796586,-0.2418756336,0.0730488822,-0.3921151161,-0.0962854847,-0.156184569,0.3900668323,0.042701792,-0.067355752,0.5413695574,-0.2072089463,0.3952532709,-0.0542634018,0.2074732333,0.2733632624,0.0856185108,0.2450954765,-0.097115308,0.1106680036,0.2532865107,0.0348394178,-0.2827269435,-0.3384743333,-0.163116917,-0.5935837626,0.0358846262,0.0273312069,-0.4786864817,0.1723166704,0.3216293454,0.3057265282,-0.1627590805,-0.1374487877,0.127164498,0.0738493204,-0.1078673005,-0.0605453402,0.5470593572,0.6318542957,-0.3070908487,-0.3632516861,0.0832097679,-0.0528882816,0.0849456191,0.1514370143,0.0182663072,0.1231782883,-0.2710851729,0.066081062,0.3751960099,-0.2643640339,-0.4211470485,-0.095526062,-0.3790399432,-0.1714232266,0.1483892649,0.0196404122,0.2099822611,0.1557330191,0.0462164879,-0.2413901687,0.2183355242,-0.2098482847,-0.4395233095,-0.348697722,0.018681718,-0.0433584191,0.2190101296,0.0419359468,0.0363063253,0.1706880331,0.1256624311,0.0660397112,0.1798494458,0.1231299415,0.3237704933,-0.0355295874,0.2978153229,0.0444417521,-0.2541936636,0.1762868762,-0.3174492717,0.213624835,0.1256953478,-0.1005328298,-0.2092880607,-0.1439248621,-0.0070462106,-0.4147659838,0.0263381079,-0.0470577665,0.270124197,0.1364537627,-0.1191124693,0.1284606904,0.107264854,0.2114131749,-0.4049317241,0.0130384667,-0.0703825802,-0.31916219,0.0298931058,0.1800133735,0.032949958,0.1019104421,-0.1698641926,0.3908302188,-0.1540256292,0.2272182852,0.3048533797,0.2851654887,-0.1629730016,-0.3259815872,0.0220399052,0.1105295494,-0.0162725355,0.0648799911,-0.202249378,0.4049359858,0.0456023328,0.1391701996,-0.2539230287,0.2988235056,-0.0855141878,-0.0431870483,-0.596321106,-0.0990096331,0.3305991292,-0.3289653659,0.2105221152,-0.1463254839,-0.0314475074,0.0823649466,0.2473679781,0.123350367,0.0772458389,0.0978807807,-0.1947132051,-0.085629411,0.0829138905,-0.1318768561,0.304479152,0.2182020247,0.1607309878,0.1080423743,0.125892967,-0.1936116517,0.1297408491,0.0885624886,-0.1993184537,-0.0519842841,0.2548471093,0.153946653,-0.2797272503,-0.0694240481,-0.4146265388,0.0284549408,-0.3391500115,0.0933108181,-0.1681764126,0.2760876417,0.0427147858,0.0185734294,-0.2692074776,-0.2986362875,0.2447029054,-0.2117600143,-0.1059977636,0.3040297925,-0.3279455006,0.6078708768,0.0554264411,-0.1463516653,-0.2999764979,-0.0674372017,-0.0466689654,-0.1146391928,-0.2106843442,0.2572805583,0.2162186056,-0.113994725,0.1512454599,-0.0240464527,-0.4987678826,0.0363867991,0.0080175931,0.6357993484,0.3888975978,0.0157717094,-0.2617827058,0.0316111185,0.3991930783,-0.4438582659,-0.2224477828,0.0916183367,-0.014788514,0.2465290576,-0.3239530027,-0.4238109887,0.1243079305,-0.4012077749,0.7106443644,-0.0169693604,0.0168244522,-0.2196318805,-0.0416192003,0.215982601,-0.2111341655,0.3168085217,-0.3373783827,-0.6845728159,0.4224869013,-0.1047696993,-0.2446500212,-0.0255974308,0.0576751344,0.3486339748,-0.1622273475,-0.5894123316,-0.4763513803,-0.0748244822,0.199990958,-0.0194934551,0.061285425,0.5578100085,-0.0331993625,-0.051837597,-0.0575106777,-0.3019850254,0.6153208613,-0.205465734,0.0410426259,-0.0076326211,0.4433331192,0.1814221442,0.8166956306,0.5588751435,0.2393333018,0.2656949461,0.1434507966,0.3056056499,-0.1414476931,-0.0946223736,0.3185797036,0.1920531392,-0.0673523694,0.1947146505,0.1703399718,0.1952806562,-0.1114070117,0.0084031364,-0.2526726723,-0.3594488204,-0.0285918526,-0.0122556416,0.1676210165,-0.0411501899,0.0335739739,-0.2244511396,-0.0431748815,0.4540455639,0.3833142817,0.2945558727,0.2433973253,-0.1909148395,0.1245266348,-0.7305420637,0.3448733687,-0.4099700749,-0.001410301,-0.1546721607,-0.329013288,0.1191457659,-0.0597264878,0.5468218327,0.2264779955,-0.1941045076,0.0241146199,-0.2036367059,-0.4357721806,-0.0704817995,-0.0595813356,-0.1430882215,0.2688424587,0.3386008143,-0.4449524581,-0.2397596985,-0.1034353971,-0.031264618,-0.2191040069,-0.3313040137,-0.2937081158,-0.0996049494,-0.2220920622,0.2061285228,-0.1314238608,0.0842838436,-0.0277879275,0.0280270278,-0.0842378661,-0.1578317434,0.0256605968,0.1693505496,0.3800826073,-0.2912943363,0.3020869493,0.5443145037,-0.0482206643,0.0338874795,0.3722675741,0.0880227759,-0.0148380762,0.2595535517,-0.0796129853,0.3189831972,0.5834521055,-0.3298600614,0.1268918514,-0.0379151031,0.4513748288,-0.2243210226,-0.1023803204,0.2226600945,-0.0440628305,0.3950959444,0.1574694663,0.2453731596,0.2686672509,-0.2394672334,0.139265269,0.0856610611,-0.1921716481,0.3677600324,0.07692305,0.9748388529,0.1425094903,0.1349177659,0.1659119278,0.0843514875,0.4501752555,0.1098287776,0.0794715807,-0.2773195505,-0.2535167933,0.0752464458,-0.1201173365,-0.1260693222,-0.3618940711,-0.3251042366,0.4081982076,-0.2714810073,-0.1056302339,-0.229064554,-0.1442239583,-0.219987154,-0.2697547972,0.0447138846,0.0453383885,0.0907435343,0.4384637773,-0.1836979538,-0.1840985566,-0.2652368248,-0.3652840853,-0.1289610863,-0.0673230439,-0.1766261905,-0.1138669997,0.2795245051,0.0044970801,-0.069608666,0.194792822,0.3574464917,0.1417988837,-0.3306594491,-0.0981989875,0.0538104847,0.1065853909,0.1469284743,0.0098678051,0.254425168,-0.1431809366,-0.0182388499,-0.0244800486,-0.1513181329,0.043899104,0.0058422452,-0.2144088447,-0.0606862269,-0.10186968,0.0982544422,0.1071010754,0.0112332618,-0.2338057458,0.110572055,0.3235704005,-0.186555028,0.1633548141,0.2400679886,-0.2253583968,-0.0512893461,0.464343667,-0.1365624219,-0.0627510324,0.4527926147,0.0421356708,-0.1334981322,-0.0853750855,-0.0759426579,0.3869663477,-0.8485310674,0.1576581895,-0.0338791125,-0.0515598841,0.0432767048,0.2604058087,0.3408826888,-0.1588791609,-0.0822825208,-0.2311331928,-0.4042378664,0.381724745,-0.3052514195,0.2306506038,-0.5099554062,0.190286696,-0.2812499702,0.1481530219,-0.319031924,0.0882178172,-0.2205171883,-0.1557302028,-0.1763935387,-0.0018246479,0.3478405476,0.0095737604,-0.0726349354,0.2380474359,-0.3705165684,-0.0926233903,-0.2068836838,0.1353428364,-0.1777234226,0.0089452211,0.0565367453,0.2277377248,-0.1598329544,-0.1410602331,0.093887426,0.1851567477,-0.0096971486,0.0880966038,0.1260210872,0.0756725073,-0.0135944607,0.1977690756,0.2378551811,-0.160340026,-0.2263823301,0.1338219345,-0.4602914155,-0.073757641,0.3483999968,0.2329486459,-0.1067643985,0.2461019754,-0.0235695969,0.0950317532,-0.448233515,0.045112282,0.4445843399,0.0489915088,-0.094012402,-0.0108754616,-0.0413266085,0.0979490355,-0.1402585059,0.3291450441,-0.0622720271,0.3363081217,0.1206548586,0.2768872678,0.2116221189,0.0139382081,0.1684416085,0.2232990265,-0.0547307283,-0.0914750025,0.3676172793,-0.1182740331,0.0249968674,0.32894665,0.5605508089,0.0763844773,0.1496204883,-0.1066402048,-0.0672027692,0.14843373,-0.0661066025,0.3817591965,-0.2465208471,-0.0910573229,-0.1572404653,0.1240019202,0.0061645443,-0.1901739389,-0.331569612,0.6917667389,-0.1093515232,-0.2469641119,-0.1880482137,0.2186513245,-0.1606260687,-0.1849393845,-0.0838897079,-0.2105691731,-0.2191709727,-0.1509547383,0.1695631444,0.018552186,0.1450956017,-0.0328529738,0.1160524189,-0.1858572662,-0.5609817505,0.141820997,0.2941439152,-0.1176028848,0.0181687735,0.2244018316,-0.2845525444,0.1979002804,0.6598454118,0.2762863636,-0.0298537668,-0.088421315,0.0559420623,0.1904031932,0.0715142637,-0.074406743,0.1938427836,0.2686851323,-0.6026164293,0.1062565893,0.0361006297,-0.1714393198,0.4032694101,-0.3423991203,0.3624660373,-0.2444796413,0.4835073352,-0.1571031213,0.1179201603,0.0028645699,-0.1929027885,-0.5489095449,0.1697090417,-0.1121426299,-0.0506036729,0.0383754335,-0.3354283273,0.0636244267,-0.106464766,0.2458087057,0.264192313,0.1838691831,-0.228505522,-0.2114779651,-0.5184434056,0.1700246334,0.2240597159,0.4091748893,0.0069524827,0.2396068871,-0.0212804452,0.1017889902,0.2442372888,0.0565914884,-0.0967075825,-0.0916238651,-0.5518030524,0.1963032484,-0.0378328972,-0.1260572076,0.1373006999,0.1882489771,-0.0334711373,-0.2720347643,0.0144160949,0.1643147767,-0.094495222,-0.0992466286,0.1843332499,0.4404118359,0.2046839595,0.0925298706,-0.01925347,-0.0986650586,-0.2265908569,-0.1590010971,-0.1725432724,-0.1005912274,-0.2516169846,0.4177378714,-0.284011066,-0.4044748545,-0.3830950856,0.2652684152,0.0418745726,-0.0697606951,-0.2140797824,0.4274216592,-0.1089170948,-0.0424474329,0.1821325272,0.2948850095,0.0220186096,0.3942229748,-0.3259976208,-0.2536972463,0.796841681,-0.4785505831,0.2696469426,-0.194009304,0.3755391836,0.4057736695,-0.3926139176,-0.4662135243,-0.0183522049,0.0767981857,0.1142917573,-0.1075978652,0.2290247232,-0.4001036286,-0.0505050309,-0.0820426643,0.3052809834,0.168076247,0.0183528364,-0.0207116622,-0.1692501158]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"> The fact a datasets.Metric object cannot be used as a simple compute function in a multi-process environment is, in my opinion, a bug in datasets\r\n\r\nYes totally, this use case is supposed to be supported by `datasets`. And in this case there shouldn't be any collision between the metrics. I'm looking into it :)\r\nMy guess is that at one point the metric isn't using the right file name. It's supposed to use one with a unique uuid in order to avoid the collisions.","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":85,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\n> The fact a datasets.Metric object cannot be used as a simple compute function in a multi-process environment is, in my opinion, a bug in datasets\r\n\r\nYes totally, this use case is supposed to be supported by `datasets`. And in this case there shouldn't be any collision between the metrics. I'm looking into it :)\r\nMy guess is that at one point the metric isn't using the right file name. It's supposed to use one with a unique uuid in order to avoid the collisions.","embeddings":[-0.1335230619,0.0371194929,0.0925977081,0.216525197,0.1155614927,0.0259055011,0.2892284095,0.2465747297,0.2430669516,0.1485799849,0.0223329589,0.1771802157,-0.2399053425,0.1134241223,-0.0412850939,-0.009081373,-0.0321525261,-0.0648063049,-0.2784070075,-0.0532979928,-0.3712781966,0.4944949448,-0.213693738,0.0575829558,-0.4752726257,0.0689635426,-0.198795557,0.1876269132,-0.120708622,-0.48742342,0.3815600574,0.1813322008,0.0492469296,0.648494184,-0.0001204536,-0.1653455496,0.2496981621,-0.0945594534,-0.1752892435,-0.2627056539,-0.0832363367,-0.3161787391,0.4207959473,-0.2127367109,0.1441112161,-0.0956146568,0.032539878,-0.4734930098,0.3829439878,0.0713166073,0.1092243865,0.4045968354,-0.0290133543,-0.3105271459,-0.0231044702,-0.0660713911,0.0089851022,0.3649280965,-0.1367458403,0.0707322732,-0.2014990449,0.1797533333,0.1654919833,0.0047445553,0.2488140315,0.0069307359,-0.0829743892,-0.1220625713,0.0122205429,0.3364715576,0.2785382569,-0.254947722,-0.3165333271,0.0869153515,-0.21462892,-0.4481872022,0.1836500317,-0.1002554223,0.0678673834,0.0524204895,-0.0563672669,-0.1756006628,-0.0516794063,0.0565954186,-0.1596506387,-0.1350136995,-0.2554564476,0.1798077077,0.0026671889,0.1714439839,-0.3266620338,0.2590613663,0.0486182012,0.1027816683,-0.5279275775,0.2013274133,0.173094213,-0.0289445519,0.1072831377,0.4302815199,-0.0310643651,0.1589917988,0.4029908478,0.1944817603,-0.0551076867,0.4064210057,0.1782255322,0.0014902785,0.2801542282,0.1749690473,-0.2232084125,-0.3422784209,0.1502284855,-0.3470233083,0.3246978819,0.0795400366,0.1799817979,-0.3473767638,-0.0161977392,0.1810603887,-0.1118371859,-0.1612889022,0.0996878743,0.2822615504,0.186663121,-0.0516057201,0.2157453597,0.4055103362,-0.1794903576,0.0225586556,-0.420535773,-0.1213239208,-0.1182068884,0.4025931954,0.0458073281,-0.1762454361,0.5597066879,-0.176439777,0.3955861926,-0.149099499,0.3421090245,0.13213934,0.0209955033,0.3017513156,0.0507131368,0.2459193766,0.2705455124,-0.0035898776,-0.1020763442,-0.3254132569,-0.1897502989,-0.5088412762,0.0911111012,-0.0049069198,-0.451523304,0.2914340496,0.2894193232,0.2298967391,-0.072845839,-0.2428895235,0.0437052138,0.0391706675,-0.2443353534,-0.1741629839,0.5107157826,0.671084702,-0.3306004107,-0.3469324112,-0.077282019,-0.0398799777,-0.0461158194,0.2499621809,-0.0406612679,0.2462102622,-0.2826662362,0.1334010065,0.2764788866,-0.3176396489,-0.4375066161,-0.0121191032,-0.36036852,-0.1362518817,0.0509688668,-0.0041754171,0.1422252357,0.1615346223,0.0578495823,-0.1849369705,0.1697222888,-0.3091491461,-0.3021737039,-0.2162054777,0.0122868987,-0.0623264462,0.3494416773,0.0934797451,0.0866662934,0.053355027,0.0463506356,0.0644136518,0.2311803699,0.2167264968,0.2976898849,-0.0360914208,0.30195117,-0.0816925466,-0.2597101331,0.3623234034,-0.163040027,0.0962197557,0.102365531,0.026928924,-0.175643295,-0.0892352238,-0.1704858094,-0.4365580082,0.0410097279,-0.0181949623,0.2263961136,0.1631458849,-0.1785077155,0.2397717983,0.1039646044,0.1861555874,-0.4449467361,0.0735413954,-0.0304206479,-0.2937315404,-0.0068508256,0.1123702824,0.0569214076,0.0660690591,-0.197527796,0.4633436799,-0.0460408553,0.261434406,0.4305955172,0.1458352953,-0.1093317121,-0.3408472538,-0.0262473654,-0.00136146,0.0535816252,-0.0572734661,-0.2175344676,0.3621328175,0.0230479632,0.1479338855,-0.2466674149,0.3745265603,-0.0810135901,-0.0561903939,-0.5400524139,-0.0346522927,0.3048755229,-0.293292582,0.3406456709,-0.1490434855,-0.0125545645,-0.0973657221,0.2400292605,0.1414795667,0.1339054704,0.0312433224,-0.3177606463,-0.0283261798,0.0000272399,-0.0193968154,0.5301395059,0.1484722495,0.2678799033,0.1348848641,0.0870616436,-0.162880674,0.0027407247,0.0954414681,0.005738534,0.0730264336,0.2060080916,0.1222143024,-0.2972860932,-0.0545394905,-0.2019452006,0.0336649604,-0.3863704801,0.1017447934,-0.247481674,0.4286676049,0.1342421174,0.109720774,-0.220001623,-0.2977715731,0.1170777902,-0.1271331459,-0.0012805379,0.3124436736,-0.1825164855,0.5101054311,0.1200421602,-0.1082284674,-0.2282790542,-0.1558557898,-0.0329200365,-0.161180079,-0.1028815433,0.2642015815,0.1315704435,-0.1125555485,0.1509696394,-0.2350515574,-0.4278851748,0.0632872656,0.0025523594,0.7433798313,0.4111458063,-0.0036345341,-0.2799821198,0.1605406702,0.4311717749,-0.4666101635,-0.2351148427,0.1392445266,-0.0111166295,0.0810798034,-0.2477896959,-0.3338740468,0.067698814,-0.3375909925,0.6766374111,-0.0679348186,-0.0718750283,-0.2298683375,-0.0328819603,0.2590990067,-0.2438746393,0.2232509404,-0.3715209961,-0.6017815471,0.3536577523,-0.1585828215,-0.3658097386,-0.0293399096,0.08941558,0.4120741487,0.0269325301,-0.5248664021,-0.4424886703,-0.0557454824,0.2325976044,0.0255852006,0.0923559219,0.4470262825,-0.1033083424,0.0136129595,-0.2129089534,-0.3032337427,0.5885711908,-0.1253105402,0.0313164629,-0.0246884041,0.4544475973,0.1785299927,0.8193620443,0.5739848018,0.2213240117,0.2435995489,0.0272405241,0.41929546,0.010075666,-0.2407452911,0.3913172185,0.0835299939,-0.1487617642,0.1663925946,0.1160771102,0.1952006966,-0.1400610358,0.0190413874,-0.3352613151,-0.2709818184,-0.1398512572,-0.0637856275,0.2285211682,0.1551153362,0.0282760579,-0.25396052,0.0925767869,0.3488354981,0.3272807002,0.1989052445,0.1866468489,-0.2482557446,0.0456898995,-0.5986585021,0.3302707672,-0.4330099225,0.1360402405,-0.1070090756,-0.1760908812,0.1209396347,-0.0360520966,0.6881152987,-0.0618485995,-0.226888001,-0.0087779369,-0.2107588053,-0.5308983326,0.0313063599,-0.1429647654,0.0433675423,0.2762994766,0.4072860777,-0.4166150391,-0.1702871323,-0.1740849912,0.0234401841,-0.2182136923,-0.3281387389,-0.3009985983,-0.0614093989,-0.2629109025,0.0574811548,-0.2308470309,0.0075310161,-0.0831387565,0.0304853208,-0.2600259185,0.0171634089,0.0119883055,0.1035649776,0.3259920776,-0.372078985,0.2957855165,0.5614343286,-0.0697107241,0.1372144371,0.4020881951,0.0194239561,-0.1499101371,0.3707313538,-0.0855315775,0.2840798497,0.453030467,-0.3448780477,-0.0066800844,-0.0185194239,0.4108576775,-0.1840827316,-0.0980240107,0.3913733065,0.0029589501,0.2447964996,0.1449107081,0.2343632281,0.3837971687,-0.1564331204,0.2894923091,0.0861308053,-0.1000663266,0.2511699498,0.0149509571,0.8027763963,0.0547763444,0.015825307,0.1936580688,0.0883169994,0.5205696225,0.1113066673,-0.0199033711,-0.245632723,-0.3328893185,0.0005449316,-0.1819361597,0.0331191532,-0.2277821153,-0.2538833022,0.3600265384,-0.1687324643,-0.130089879,-0.1787787825,-0.0495316871,-0.2385253608,-0.1991713345,0.135170877,0.0546528436,0.1097514927,0.4981806278,-0.1706255227,-0.1447525024,-0.2722235024,-0.3178369105,-0.2482748479,-0.0900001153,-0.1977530569,-0.0111124814,0.2499724329,-0.0639043972,-0.0252631325,0.267781347,0.5074549913,0.0472657792,-0.3525550663,0.0034764183,0.0130431205,0.1437121928,0.1161226854,-0.1211206764,0.2373104692,-0.068992734,-0.0926968083,-0.0657146424,-0.1648800075,0.1208203584,0.1018045247,-0.1147975475,-0.2134728581,-0.0577048324,0.034478914,0.0207513701,0.0414717831,-0.3736036718,0.0758247972,0.2722466886,-0.1465555727,0.1722982079,0.1113971621,-0.2608970702,-0.0141501585,0.5147318244,-0.2373294979,-0.0303424858,0.4104905725,0.1308338791,-0.028320238,-0.1000231802,0.0894256383,0.4136506319,-0.7102579474,0.1728532016,-0.0983264744,0.1189663857,0.0743293092,0.1602973938,0.2923335731,-0.2233011276,-0.0529579669,-0.3134144545,-0.3251955211,0.2963859737,-0.3277485073,0.1088097692,-0.3996322155,0.1098277271,-0.1114411652,0.2352122068,-0.2688808143,0.2411135435,-0.2456686497,-0.1734372973,0.0200265851,0.0298224259,0.3511940539,-0.0595263205,-0.0846800432,0.2033030242,-0.3578532636,-0.0760751814,-0.2357937247,0.1430660188,-0.0966833755,0.0890799239,0.0271421988,0.173092261,-0.1648473442,-0.1966081858,0.0914934427,0.2636885941,-0.0502028577,0.0855073482,0.1101140231,0.1596431881,-0.0471670479,0.1796099395,0.1540251821,-0.187758252,-0.1319302619,0.1650542021,-0.2592048347,-0.0534538589,0.2167648226,0.0128264725,-0.1290295571,0.1538245827,0.06963557,-0.0815031603,-0.30344522,-0.040924456,0.5869249701,0.2290361822,-0.148273617,0.0973780081,-0.0870540068,0.0747678056,-0.1308400333,0.2859936655,0.0804749876,0.2450648248,0.1103253588,0.194886595,0.147214666,0.0447329655,-0.0332639441,0.1988371611,-0.0040262341,-0.2192634791,0.3405556977,-0.1342507452,0.0446110889,0.395119369,0.5971511602,0.173056215,0.1452834159,0.1526366472,-0.0955010504,0.2259518653,-0.1051749513,0.3171729147,-0.0828347802,-0.1924597323,-0.091211088,0.0132808834,-0.0581707954,-0.2416953146,-0.4781826735,0.750474453,-0.3104272485,-0.2048026174,-0.2497939765,0.2282947004,-0.1533305496,-0.1966213137,-0.1678117961,-0.1948921382,-0.100598678,-0.162246272,0.1900172234,0.0440237001,0.2180059254,-0.0209905356,0.1075953767,-0.3651963174,-0.5040792227,0.2400596142,0.0598785542,-0.1469106823,-0.0378542617,0.2424749434,-0.3101002872,0.1404374838,0.6258374453,0.3339041471,0.0355106369,-0.1565678716,0.0957118198,0.2010085732,0.0026190197,-0.1123624668,0.1926767975,0.1783116162,-0.4704280794,0.2172089666,0.0601484701,-0.1306458712,0.2637719512,-0.2489306778,0.3911230564,-0.3739433885,0.4142304361,-0.3403817713,0.1366842538,0.0856260434,-0.1665976495,-0.5777480602,0.2350998819,0.0391676165,-0.0788595974,0.1047069356,-0.3096340001,0.0495043173,-0.1320972145,0.2266010344,0.3706276119,0.1260538399,-0.3011863232,-0.2492923141,-0.7142771482,0.3186253309,0.1262675673,0.3032756746,-0.0899856687,0.3800609708,-0.0494272411,0.0609031469,0.2452533543,0.0120628886,-0.0164225101,0.0251141228,-0.5366042852,0.0548594072,-0.2019567341,-0.1890062839,0.1295386702,0.2296334207,0.0115954028,-0.2051220536,-0.005605774,0.0943918154,-0.0876883417,-0.1065062732,0.104528442,0.4748960137,0.2246723622,0.1173732728,-0.044018209,-0.0084056044,-0.2767057717,-0.1092071757,-0.0843692347,-0.0301332884,-0.3517562747,0.4323344529,-0.3502947688,-0.3416685164,-0.337552458,0.1549044997,0.1130455956,-0.1122550741,-0.259562254,0.406701088,-0.1336854845,0.0444320701,0.2433538586,0.3771452606,-0.0471835248,0.4603345394,-0.2675015628,-0.177642554,0.7530431747,-0.4552138448,0.1326742619,-0.2201703042,0.3438367248,0.3547530174,-0.4648526609,-0.3702152967,0.0004914533,0.1647120267,0.0972570032,-0.2037421316,0.1568378508,-0.3975250125,-0.0358708911,-0.0410871729,0.2923150361,0.1856916249,-0.0826075152,0.1278854758,-0.0314362608]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"I just opened #1966 to fix this :)\r\n@stas00 if have a chance feel free to try it !","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":19,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nI just opened #1966 to fix this :)\r\n@stas00 if have a chance feel free to try it !","embeddings":[0.0502678864,-0.0307857376,0.058525674,0.1668777913,0.0344583355,0.0303145554,0.1830068529,0.2939564586,0.2358811647,0.1017371118,0.1023575366,0.1887697875,-0.3214577138,0.0011921655,0.0528264791,-0.0287232734,-0.050264582,-0.0096711554,-0.3583105803,-0.125402987,-0.2325507402,0.5447103381,-0.1571903378,-0.0045523304,-0.5084875226,0.2233060896,-0.1877898276,0.1402115971,-0.096882537,-0.5538378358,0.4047330916,0.1505976319,0.0910343751,0.4610123634,-0.0001223977,-0.1811482906,0.1949375272,-0.1929721832,-0.1423965693,-0.1068891063,-0.0061707539,-0.2863018513,0.5193983316,-0.1492537558,-0.0622213967,-0.0420732498,0.0282986779,-0.2860839665,0.5111199617,0.0969570428,0.1117404699,0.3820098341,0.0355157331,-0.3040623665,0.0467137098,-0.0588730462,-0.0233318191,0.3352954686,-0.1401589215,0.0310185049,-0.1860093772,0.2797896564,0.2617025375,-0.0712250322,0.2279260457,0.0110841403,-0.1231674775,-0.1699431539,0.0016832431,0.2876955271,0.2455520481,-0.1002671272,-0.2289027572,0.1350904703,-0.1488545388,-0.2994295061,0.1515563577,-0.1007180512,0.0996650085,0.1133290753,0.0295816734,-0.1865183264,-0.1075744331,0.0627107322,-0.1474566013,-0.0327554233,-0.1966677457,0.0187708233,-0.0397600308,0.0962033048,-0.3239345551,0.4986526668,-0.0832981765,0.1876079142,-0.3713005483,0.0709057674,0.2739272714,0.0947960839,0.0921184048,0.3463573158,-0.0080726137,0.0798768252,0.4184749126,0.081495665,-0.0967774466,0.4281893671,0.2606991231,-0.0058043739,0.160658434,0.2437297851,-0.2130867392,-0.2938705683,0.1996711642,-0.2233803868,0.3632273972,0.0774252191,0.0587498061,-0.4959126115,0.111743018,0.248695761,-0.1679585725,-0.2072851658,0.1042115614,0.3869486749,0.1424576193,-0.0502152331,0.2787387967,0.3803991377,-0.1684308052,-0.0360849202,-0.4457574785,-0.1796839535,-0.0967056304,0.3202261627,0.1431293339,-0.0712345764,0.666742444,-0.1645593494,0.3357947469,-0.1272892505,0.2507025301,0.2788630724,0.0503678694,0.2963223755,0.017531028,0.1571660787,0.3135041893,0.1565272659,-0.1912126988,-0.3347271681,-0.1577663422,-0.514759481,0.0790068284,0.0095946388,-0.3871156573,0.1222413555,0.4855627418,0.2077401876,-0.1191087812,-0.1844083816,0.1640950143,0.0471462198,-0.0238042064,-0.0801770017,0.5420817733,0.615870893,-0.3337564468,-0.3810562491,0.0857886076,-0.0635162294,-0.0396002382,0.2201705575,-0.0056287884,0.1396712959,-0.2328994572,0.1088684872,0.3703120351,-0.2679551244,-0.5212201476,-0.0552315712,-0.3564217985,-0.1850541383,0.0904623941,0.0134211509,0.0928364769,0.225457117,0.0927810594,-0.2245485932,0.2309311479,-0.3117347658,-0.3569087386,-0.2340022773,0.0308409315,-0.062772423,0.3183653355,0.0701388493,0.0489349738,0.0875613615,0.1221265048,0.0490468703,0.1379181296,0.228475377,0.405058831,0.0881538838,0.2707894444,0.1175322682,-0.1957643479,0.212717846,-0.3725407422,0.1472752243,0.1539965719,-0.0014504673,-0.2130684406,-0.1569034904,-0.0808195323,-0.497097522,0.0302161854,-0.1321008503,0.2546531856,0.1813732684,-0.160805583,0.0999855921,0.070633769,0.2504447103,-0.448924154,0.0808878168,-0.1008740142,-0.3557091355,0.0105345771,0.1248174831,0.0659178793,0.0568635762,-0.2100691795,0.360947758,-0.1960226744,0.2700865269,0.3957691789,0.1259018481,-0.1700840443,-0.4662604928,0.0135637289,0.096373342,-0.1102107391,0.0343088657,-0.2294193059,0.3718851805,0.1129395738,0.0852602422,-0.3511004746,0.3622631133,-0.0410507955,-0.1646935195,-0.4694804549,-0.1569866389,0.3380329013,-0.2630411983,0.1648102999,-0.1352573931,-0.0066636335,0.0267155673,0.1505910605,0.1934215277,0.192119509,0.0640547499,-0.2622757554,-0.1365644187,0.0338335112,-0.1024969444,0.3670924306,0.1950183511,0.2114621401,0.0737834275,0.1446101367,-0.1691469848,0.0960002169,-0.0084996074,-0.1350520849,0.089232862,0.2438503951,0.1896131933,-0.3362028003,-0.0146998204,-0.3934499025,0.073306106,-0.3259215355,-0.0099369492,-0.1636276543,0.3344520628,0.1839309931,0.0512290783,-0.2567327023,-0.2762008011,0.2306934446,-0.2379369885,-0.0242273547,0.2272982746,-0.3410358429,0.5317052007,0.1348362118,-0.06541612,-0.285535723,-0.1330365539,-0.0507565513,-0.1544408351,-0.2762285769,0.2058389485,0.1834726036,-0.1706162989,0.0736595616,-0.0385266058,-0.4664661884,0.0471297354,-0.0269774441,0.6298754215,0.3298671246,0.0796626508,-0.267336458,0.0408608392,0.4414919317,-0.431804806,-0.2710913122,0.0875409544,0.0048042042,0.1961859167,-0.2463932484,-0.3488059342,0.0998698547,-0.3313794732,0.6622585654,0.0467878245,-0.0492481515,-0.2069789171,-0.0708792135,0.2037699521,-0.2940333486,0.3168635368,-0.355140835,-0.684522748,0.45145455,-0.1522807628,-0.2791167796,0.0029129069,0.0694520399,0.411183387,-0.104393661,-0.5750929117,-0.4180468023,-0.0496871099,0.2071723938,-0.0043233573,0.0652834028,0.4947388172,-0.1225787476,-0.0312670246,-0.0818334594,-0.2440459132,0.5067594647,-0.0428347141,0.1758790463,-0.124929063,0.4491825998,0.1300961524,0.8017882109,0.5838682652,0.2834761441,0.2731365561,0.0615963787,0.3118466437,-0.1129289567,-0.2101451457,0.3428444266,0.2255459428,-0.0879663229,0.2291420251,0.1611828506,0.1728362739,-0.1125848368,-0.0583819151,-0.3652356863,-0.3408969343,-0.0870222375,-0.0281336755,0.2828761041,-0.0352769941,0.0498061888,-0.2052808553,0.0224271156,0.4554180205,0.3214437962,0.1747520864,0.2012364417,-0.1860411316,0.0822226405,-0.6240657568,0.288346827,-0.5249202847,0.1255550236,-0.2078340203,-0.3026982248,0.1475297213,0.0284114759,0.5838211775,0.096759811,-0.1193180904,-0.0351820514,-0.2145917118,-0.4352736771,-0.0572161525,-0.1044352353,-0.0560941137,0.3624502718,0.3101021647,-0.3499494195,-0.1848714799,0.0001565909,-0.1191484705,-0.2800283432,-0.3077387512,-0.2760510445,0.0416327752,-0.1664548367,0.1114428416,-0.2791818678,-0.0038866687,0.0408117324,0.0349706337,-0.1482830048,-0.1122353151,0.0149488393,0.1882759631,0.3704337478,-0.2815084159,0.2173379213,0.5262382627,-0.1605110615,0.0300573353,0.3995460272,0.0934050977,0.0072200228,0.216224432,-0.1260620654,0.3122652769,0.4790814221,-0.3064969778,0.0584022664,-0.0449388698,0.4254012406,-0.1694805175,-0.1294699907,0.3659024835,-0.1286430359,0.4067678452,0.2168536335,0.2520103157,0.1982664466,-0.1372206807,0.1922990531,0.1228666753,-0.1298318356,0.4037837684,0.0725145191,0.906662941,0.228497237,0.1569655091,0.1142856628,0.1287258267,0.4183024764,0.1660075635,0.0168894846,-0.2391742468,-0.2163564265,0.021836875,-0.129456073,-0.0922039747,-0.334130913,-0.3491729796,0.3339420557,-0.4162360728,-0.171432063,-0.230045408,-0.2589849532,-0.2749912441,-0.1125248149,0.08329238,0.03937231,0.1273489445,0.5895770192,-0.1984666586,-0.186569348,-0.3087351918,-0.3782607615,-0.1994603276,-0.0075616613,-0.1780516058,-0.0446494669,0.3563719392,-0.1140281335,-0.1081576496,0.274751991,0.4248383343,0.141020909,-0.4095390141,-0.0447904915,0.0970984548,-0.0064653698,0.088215746,0.0607834682,0.2227434069,-0.0427292734,-0.0432672016,0.0264807232,-0.1708238721,0.1513321251,-0.0115302447,-0.1838580072,-0.2010068893,-0.1706019491,-0.0206781775,0.1199315488,0.0131870182,-0.2438447624,0.0945434347,0.3195438385,-0.2366118878,0.1028821617,0.2528238297,-0.2230640948,-0.0671368614,0.5789999962,-0.3150370419,-0.066442892,0.4722227454,0.089262791,-0.0668759122,-0.1505775005,0.0398676097,0.4476163089,-0.7930822372,0.0682794154,-0.0052738097,0.0292086471,0.1369697005,0.1824982762,0.3927782178,-0.2080865949,-0.0718280971,-0.1729663163,-0.4252092838,0.2752403319,-0.3139511943,0.184644863,-0.5226011872,0.0904927924,-0.113483727,0.3245965242,-0.2781523764,0.18216452,-0.2456912249,-0.1479725838,-0.1570417732,-0.0566651039,0.3577843308,-0.0033507482,-0.0518474653,0.1187390015,-0.3425809145,-0.0974133089,-0.2557145357,0.1442090124,-0.1051663831,0.1344410032,-0.0343740806,0.2559404671,-0.078572005,-0.0954964012,0.0442787483,0.1784220636,0.0884616598,0.2135743201,0.1516379267,0.0300446376,-0.0655759498,0.1676625758,0.1536464989,-0.2112171352,-0.0985519513,0.0039766603,-0.3737252057,-0.0803641304,0.256691277,0.1499453038,-0.2190535069,0.1602817327,0.0316479728,0.0055185561,-0.3093929887,-0.0150977857,0.4742203951,0.1363734454,-0.1317465454,0.0929520279,-0.0725804344,0.1205275357,-0.1349424571,0.3159368038,0.0219198074,0.3718644977,-0.0062719989,0.2041681558,0.0908575654,-0.0251444578,0.0929933935,0.2155624926,-0.0399737433,-0.1805283427,0.3490175605,-0.0985132381,0.0237438548,0.2782061696,0.5448624492,0.1011172682,0.1475640833,-0.0531077944,-0.0803421885,0.171030134,-0.1602045894,0.3541837037,-0.1122851893,-0.1044135019,-0.1984600276,0.0329712853,0.0545213856,-0.1761419773,-0.4291506708,0.8031209111,-0.1808707863,-0.2441591918,-0.1937084198,0.2028096467,-0.1682163626,-0.1527668983,-0.1624449641,-0.2118790597,-0.1720935702,-0.1443358809,0.2267824262,0.036352776,0.0690791681,0.1380734146,0.0689271018,-0.3447806835,-0.4604189396,0.2070623934,0.1764552146,-0.1624523252,-0.0082209725,0.1841730475,-0.3044584692,0.1713963002,0.6305410862,0.2583577335,0.0060694506,-0.1108448878,0.1110413894,0.2097603679,-0.04998146,-0.1143164784,0.1608441919,0.2634715736,-0.500848949,0.0362585373,0.0499518961,-0.143764779,0.3807952404,-0.3335390985,0.3961320519,-0.1687199771,0.4106985629,-0.2003553361,0.1380613148,-0.0227754749,-0.2151282728,-0.665073514,0.2170359343,-0.1144522503,-0.0244503319,0.0889878944,-0.2698170841,0.0400799625,-0.1257957071,0.2337582558,0.4144931734,0.1051801518,-0.2935039103,-0.3102200329,-0.5083488822,0.2386898994,0.2074893713,0.3287257254,-0.0308868103,0.3978417218,-0.099537842,0.037392959,0.2820653617,-0.0539015979,-0.0763697252,-0.036537677,-0.5322999358,0.2037280798,-0.0945926234,-0.0557686202,0.1298418045,0.3552437425,-0.0351372696,-0.2536826432,-0.0055233804,0.1682240218,-0.203499943,-0.1541195661,0.1702323258,0.4273458123,0.1514543891,0.0878897384,-0.0181997344,-0.0674417317,-0.2374912351,-0.1905182749,-0.0524034202,-0.15061979,-0.2223515958,0.4425070584,-0.3113907278,-0.4462624788,-0.3600740433,0.2234850228,0.0713064149,-0.1215926632,-0.2035275251,0.3574284911,-0.0377724916,-0.0298089609,0.278734684,0.3011468351,-0.0210172497,0.3851296008,-0.3222494423,-0.2338648587,0.7657587528,-0.4776708186,0.2892950475,-0.2233699113,0.3952086866,0.3707591891,-0.4478728175,-0.4201605916,-0.0033490148,0.1377255172,0.0659002662,-0.094123438,0.1670298278,-0.2764947116,0.0034387955,-0.0241549481,0.3220012486,0.2488743216,0.0636228547,-0.0355923921,-0.1095625833]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1942","title":"[experiment] missing default_experiment-1-0.arrow","comments":"Thank you, @lhoestq - I will experiment and report back. \r\n\r\nedit: It works! Thank you","body":"the original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.","comment_length":15,"text":"[experiment] missing default_experiment-1-0.arrow\nthe original report was pretty bad and incomplete - my apologies!\r\n\r\nPlease see the complete version here: https:\/\/github.com\/huggingface\/datasets\/issues\/1942#issuecomment-786336481\r\n\r\n------------\r\n\r\nAs mentioned here https:\/\/github.com\/huggingface\/datasets\/issues\/1939 metrics don't get cached, looking at my local `~\/.cache\/huggingface\/metrics` - there are many `*.arrow.lock` files but zero metrics files.\r\n\r\nw\/o the network I get:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow\r\n```\r\nthere is just `~\/.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow.lock`\r\n\r\nI did run the same `run_seq2seq.py` script on the instance with network and it worked just fine, but only the lock file was left behind.\r\n\r\nthis is with master.\r\n\r\nThank you.\nThank you, @lhoestq - I will experiment and report back. \r\n\r\nedit: It works! Thank you","embeddings":[0.0530284643,-0.0160480198,0.0690020919,0.1493227035,0.0107738739,0.0104319835,0.2262237519,0.2791083455,0.2140974551,0.1024512351,0.0898374617,0.1662594974,-0.3189935982,0.027089553,0.0486126803,-0.0156555306,-0.0611110628,-0.0139249777,-0.3136204779,-0.1463938802,-0.2088242024,0.5517245531,-0.1529769897,-0.0209766086,-0.5255408883,0.2609625161,-0.1898949295,0.1473518014,-0.0834888369,-0.5555571914,0.4081381559,0.1047194153,0.0988777801,0.4777060747,-0.0001239213,-0.1717230678,0.1813993305,-0.2075423598,-0.1246178001,-0.1537092924,-0.027442202,-0.3326763213,0.4967187047,-0.1787758172,-0.0639811158,-0.007995612,0.0282780845,-0.2284790426,0.5079200864,0.1058161482,0.0929994211,0.3394871652,0.0588178299,-0.2831996977,0.0330822058,-0.0905507356,0.0051503321,0.2902565598,-0.1839210093,0.0723215044,-0.2034839541,0.2765860558,0.2661857605,-0.0525257662,0.2360172421,0.0097891884,-0.113349095,-0.1363295168,0.0047200564,0.2672788203,0.3008900285,-0.143958956,-0.2297505736,0.1419091076,-0.150279209,-0.3227819204,0.1594252139,-0.0864905193,0.1198247075,0.1210821122,0.0573465414,-0.1782152653,-0.1153752357,0.0943819657,-0.1627408564,-0.0137343546,-0.2190897018,0.0114504518,-0.0803233311,0.1196820214,-0.3777919114,0.4586606026,-0.0573252365,0.1761979759,-0.4027523398,0.0534301102,0.2504758835,0.0726154521,0.0583632886,0.4224841595,-0.0098689059,0.0775411651,0.4253779054,0.1009115651,-0.1109181121,0.4971424937,0.2863775492,-0.0123322923,0.1780929565,0.2536886036,-0.2057204694,-0.3051637411,0.201737076,-0.2274441272,0.4072896242,0.0938275009,0.1033251137,-0.4950875342,0.1318174601,0.2394347638,-0.1752534062,-0.244053781,0.1077289283,0.3957390189,0.1389407516,-0.0703240037,0.2585566044,0.3971364498,-0.142260924,-0.0342284814,-0.4243409038,-0.1769605875,-0.0900696442,0.3117819726,0.172831133,-0.0550283678,0.6602275372,-0.19816719,0.3620591462,-0.1110952348,0.2596358657,0.2776535749,0.0715484172,0.2864853144,0.0207959488,0.1853353977,0.2984788418,0.129904002,-0.1763179898,-0.334148854,-0.2054360211,-0.5386745334,0.1031149179,-0.0126619693,-0.358697325,0.1368581653,0.4928449392,0.1927814931,-0.0777483284,-0.2290932089,0.1878876388,0.0900730863,-0.0053315037,-0.0721998215,0.5400422812,0.608046174,-0.3268814981,-0.4060615003,0.0563770197,-0.0615029335,-0.0281167161,0.2436828613,0.0355425626,0.1406904459,-0.2321488559,0.1402564049,0.3658519685,-0.2745782137,-0.539262116,-0.0522775501,-0.3938910663,-0.1868904531,0.1386560351,-0.0106840078,0.1208342239,0.2324938774,0.0717758611,-0.2263977826,0.2879318893,-0.3329198956,-0.346011281,-0.2300477773,0.0348986499,-0.0751586556,0.3323730826,0.0698657557,0.0588531718,0.1029050797,0.1170106754,0.0507231802,0.1663779467,0.2380164713,0.3865634203,0.0873437151,0.2853321135,0.1242259964,-0.1947783083,0.2273243517,-0.4057549238,0.1488321871,0.1629823893,0.0135508152,-0.1952039748,-0.174373582,-0.1079485938,-0.4924083948,0.0176211279,-0.1006393507,0.2854163945,0.1654936373,-0.146558091,0.0612561963,0.1174865067,0.2530274987,-0.4141819477,0.0692490116,-0.1078382507,-0.3465871513,-0.0023672986,0.1437533647,0.0744288266,0.0578936748,-0.1972722858,0.3352254033,-0.2186820805,0.2658824623,0.371319592,0.1242166907,-0.1607363522,-0.4700486958,0.0195444301,0.1124824509,-0.0977362618,0.039985951,-0.1854148954,0.3564597666,0.1572318226,0.0682926476,-0.3777579665,0.3484071791,-0.0678526536,-0.1668175459,-0.4660649002,-0.1620980501,0.3576309085,-0.2789849639,0.2035492957,-0.1403975934,0.0119919665,0.0131228939,0.1550483406,0.1689918637,0.2122148275,0.044043228,-0.2486723512,-0.1136638671,0.0683899224,-0.1658792347,0.3285855353,0.1947811395,0.2141816914,0.0608224533,0.1628137082,-0.1768020242,0.1081261337,0.0002517631,-0.1498255581,0.1192423627,0.2224060446,0.1849992126,-0.3228563964,-0.0047681141,-0.4039285481,0.0286790356,-0.3488660157,-0.0138969785,-0.1858678609,0.3286344409,0.2171792537,0.0518233553,-0.2774496377,-0.2361330241,0.2195572704,-0.2605090141,0.0174155738,0.2638673782,-0.2950529158,0.5226531029,0.1304601282,-0.0973458588,-0.3099012375,-0.1005875915,-0.0493998751,-0.1723790616,-0.2647532821,0.1496191025,0.128203392,-0.1620803922,0.0916673094,-0.0384878628,-0.4578766227,0.0451449975,-0.0160957742,0.6010417342,0.3081532419,0.0610171743,-0.2483258545,0.077884756,0.4227271378,-0.431867063,-0.2639607489,0.1343554854,0.0005381812,0.2076465338,-0.2232670337,-0.3197921216,0.0541237444,-0.3064563572,0.6670658588,0.0968331397,-0.0405136719,-0.2120511383,-0.0913318321,0.2088548839,-0.2869220376,0.3692118227,-0.3518076837,-0.676003933,0.4575407207,-0.1490148604,-0.2631347179,-0.0068330201,0.0805734545,0.4331932962,-0.153840974,-0.5491352677,-0.3989624083,-0.0375186615,0.2518237531,-0.0203163698,0.0920146406,0.5247725844,-0.126942277,-0.0120769264,-0.0607511215,-0.2334377915,0.5128337741,-0.0418005548,0.1538528204,-0.1134990677,0.4474056661,0.1449232846,0.8587962985,0.536613822,0.3179355562,0.3232057691,0.0690223873,0.2967537045,-0.1217105836,-0.1966748983,0.3493185043,0.2244125754,-0.088823393,0.2535344362,0.1484118253,0.173803106,-0.1174712479,-0.1031360999,-0.387493372,-0.3084761798,-0.1228357479,0.0059073921,0.264814198,-0.0123821935,0.0448047854,-0.2134616673,0.0204893351,0.4786902368,0.3412490785,0.1560891867,0.211485818,-0.1639944911,0.0492308214,-0.6495614052,0.3202866614,-0.5207639337,0.0661031604,-0.2111539692,-0.3088837564,0.1460712403,0.0243327171,0.5834698081,0.1222550347,-0.0972264558,-0.0430510156,-0.2499232143,-0.4018514752,-0.0719796792,-0.1046482027,-0.0572994575,0.3666612208,0.3168306351,-0.3949627578,-0.1739231348,-0.039723929,-0.1177941933,-0.2673517466,-0.3102994561,-0.2667182088,0.1071896628,-0.240885213,0.1153007969,-0.3058530986,0.0238019507,0.067705825,0.0594696216,-0.1743971407,-0.1006905809,0.0112086283,0.1511523128,0.3522874713,-0.2827352583,0.1949660629,0.4679219723,-0.1379637271,0.0058981967,0.3911335766,0.0858728364,0.0104828868,0.2230029255,-0.1398701072,0.3000159562,0.4908545911,-0.3067614436,0.0700802878,-0.0474110395,0.4236694574,-0.1835255772,-0.1934931725,0.3262196779,-0.112550959,0.3926888704,0.237464726,0.2116914243,0.1799035519,-0.1150676459,0.1524745971,0.0959748104,-0.1356598288,0.3794244826,0.0544518083,0.9139834642,0.2294770628,0.2096172869,0.1152469963,0.1707020551,0.4021791518,0.1505739242,0.0198923554,-0.2222251743,-0.2606551647,-0.007048673,-0.1182865649,-0.0995652974,-0.2877553999,-0.3400573432,0.3567262292,-0.3996930718,-0.1402613074,-0.236695677,-0.268908143,-0.2612554431,-0.1223160997,0.0343365036,0.022329187,0.1307484061,0.6034716368,-0.1952624321,-0.1886188388,-0.3215142787,-0.4012112617,-0.1478893459,-0.0035599989,-0.2024101913,-0.024034461,0.3291046023,-0.1503533721,-0.1365438849,0.2533171177,0.3774065375,0.1440952122,-0.3918331861,-0.0749658048,0.1197247505,0.0067370743,0.0556371696,0.03669554,0.2433629632,-0.0621769279,-0.0256709307,0.0667109564,-0.1895771325,0.1487134844,0.0489069149,-0.1859183311,-0.1910699308,-0.1324882358,-0.0560968928,0.1310995072,-0.006444579,-0.2230911106,0.0766917169,0.2993971109,-0.2721615434,0.0597745031,0.2282117307,-0.2293671519,-0.077878654,0.6224727035,-0.3129539192,-0.0068699694,0.5105515718,0.0825104639,-0.0532475039,-0.1368771344,0.0461645909,0.4430187047,-0.834520638,0.0495914668,0.0264324229,0.0182014909,0.0851424783,0.1873038262,0.4106439054,-0.1507997513,-0.0645109862,-0.1905778497,-0.4066484869,0.2583598793,-0.300560683,0.1543877274,-0.4546395242,0.0955723748,-0.0862620622,0.2974578738,-0.2611766458,0.1482568532,-0.2478500903,-0.1595347077,-0.1593501717,-0.0607593767,0.3503407538,0.0200631302,-0.065885976,0.1011899263,-0.329754144,-0.0854228809,-0.2368927747,0.1594503522,-0.0745474175,0.160889551,-0.0299938247,0.2547236979,-0.1119280234,-0.0923389345,0.0564472489,0.1654249877,0.085316807,0.2073580027,0.139518559,-0.0110622477,-0.1186528951,0.1496536285,0.1445297748,-0.2485118806,-0.1124760062,0.0430292226,-0.383915931,-0.0573817492,0.2169474065,0.1526251435,-0.2329845279,0.1506479383,0.0332970954,-0.0083723823,-0.3137623966,-0.0016784056,0.4485861957,0.0790398642,-0.1359830201,0.0520742126,-0.0866719261,0.0965154395,-0.1428188533,0.3265905976,-0.0089052841,0.3387452662,0.0024490729,0.2145982683,0.0864074752,-0.0495481528,0.1020356342,0.2454911619,-0.0877587274,-0.1683283597,0.3294771314,-0.102321513,0.037644431,0.3276984692,0.5661908388,0.1154537499,0.1226814091,-0.0989383385,-0.0815664083,0.1801287532,-0.1583233476,0.3534008563,-0.1367229819,-0.1280311942,-0.1877104789,0.0457687005,0.0673491135,-0.2072540522,-0.4382883608,0.7649759054,-0.1675099432,-0.2265948355,-0.2085262388,0.2537218034,-0.1612797827,-0.1442864388,-0.1842089295,-0.1971043795,-0.1438867897,-0.1522115916,0.233172968,0.0709913373,0.0378103405,0.1406013072,0.0546418875,-0.3246084154,-0.465305686,0.2115962058,0.1670997888,-0.1508572847,0.0238391701,0.1826548278,-0.2962293327,0.1739457548,0.6317142844,0.2703976929,-0.0220080763,-0.0838794932,0.1107020304,0.2334987968,-0.0621768981,-0.1147419512,0.1793136448,0.2737958431,-0.5111684203,0.0394843929,0.0295919646,-0.1384340376,0.3690729141,-0.2814406753,0.3652097583,-0.2134120166,0.4403071702,-0.1833730191,0.1313392818,-0.0299817026,-0.1995068192,-0.6798967719,0.2709859908,-0.1317396164,-0.0320486762,0.0729617476,-0.2919365764,0.027432112,-0.1381022781,0.2248945534,0.3910744488,0.0905789137,-0.2797752917,-0.2613033652,-0.5100674629,0.2513461709,0.2254179716,0.3319685161,-0.0507252924,0.3823292851,-0.1034601033,0.0434141532,0.309048593,-0.043778725,-0.0741627961,-0.0186925642,-0.508797884,0.2340993881,-0.0603275448,-0.0463620871,0.1229965165,0.3462249935,-0.0519236512,-0.2408712804,-0.0194102135,0.1882900298,-0.22010912,-0.1649547964,0.1550042927,0.4570683241,0.1186340526,0.0820321217,-0.0636416525,-0.0485141873,-0.1681472808,-0.1305892915,-0.063582018,-0.1433063149,-0.2328721136,0.4607042968,-0.3149006069,-0.4100903273,-0.358969897,0.2712211311,0.0819939747,-0.1255527139,-0.1570492536,0.364779681,-0.0779932365,-0.019671347,0.2716139853,0.2628602386,-0.0135426242,0.3932921588,-0.3367212415,-0.2338114679,0.727686584,-0.5018507242,0.2962461114,-0.2486705035,0.4007821977,0.3242767751,-0.44358325,-0.3566764295,0.0007660027,0.1400144994,0.0953736305,-0.1084916294,0.1438907087,-0.3240505755,0.0245998632,-0.055635836,0.314332068,0.2554893494,0.0668172911,-0.0600044765,-0.1039852798]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1941","title":"Loading of FAISS index fails for index_name = 'exact'","comments":"Thanks for reporting ! I'm taking a look","body":"Hi,\r\n\r\nIt looks like loading of FAISS index now fails when using index_name = 'exact'.\r\n\r\nFor example, from the RAG [model card](https:\/\/huggingface.co\/facebook\/rag-token-nq?fbclid=IwAR3bTfhls5U_t9DqsX2Vzb7NhtRHxJxfQ-uwFT7VuCPMZUM2AdAlKF_qkI8#usage).\r\n\r\nRunning `transformers==4.3.2` and datasets installed from source on latest `master` branch.\r\n\r\n```bash\r\n(venv) sergey_mkrtchyan datasets (master) $ python\r\nPython 3.8.6 (v3.8.6:db455296be, Sep 23 2020, 13:31:39)\r\n[Clang 6.0 (clang-600.0.57)] on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n>>> tokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\n>>> retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=True)\r\nUsing custom data configuration dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\nUsing custom data configuration dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\n 0%| | 0\/10 [00:00\", line 1, in \r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 425, in from_pretrained\r\n return cls(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 387, in __init__\r\n self.init_retrieval()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 458, in init_retrieval\r\n self.index.init_index()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 284, in init_index\r\n self.dataset = load_dataset(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 734, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/utils\/py_utils.py\", line 195, in map_nested\r\n return function(data_struct)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 769, in _build_single_dataset\r\n post_processed = self._post_process(ds, resources_paths)\r\n File \"\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb\/wiki_dpr.py\", line 205, in _post_process\r\n dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/arrow_dataset.py\", line 2516, in add_faiss_index\r\n super().add_faiss_index(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 416, in add_faiss_index\r\n faiss_index.add_vectors(self, column=column, train_size=train_size, faiss_verbose=faiss_verbose)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 281, in add_vectors\r\n self.faiss_index.add(vecs)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/__init__.py\", line 104, in replacement_add\r\n self.add_c(n, swig_ptr(x))\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/swigfaiss.py\", line 3263, in add\r\n return _swigfaiss.IndexHNSW_add(self, n, x)\r\nRuntimeError: Error in virtual void faiss::IndexHNSW::add(faiss::Index::idx_t, const float *) at \/Users\/runner\/work\/faiss-wheels\/faiss-wheels\/faiss\/faiss\/IndexHNSW.cpp:356: Error: 'is_trained' failed\r\n>>>\r\n```\r\n\r\nThe issue seems to be related to the scalar quantization in faiss added in this commit: 8c5220307c33f00e01c3bf7b8. Reverting it fixes the issue.\r\n\r\n\r\n","comment_length":8,"text":"Loading of FAISS index fails for index_name = 'exact'\nHi,\r\n\r\nIt looks like loading of FAISS index now fails when using index_name = 'exact'.\r\n\r\nFor example, from the RAG [model card](https:\/\/huggingface.co\/facebook\/rag-token-nq?fbclid=IwAR3bTfhls5U_t9DqsX2Vzb7NhtRHxJxfQ-uwFT7VuCPMZUM2AdAlKF_qkI8#usage).\r\n\r\nRunning `transformers==4.3.2` and datasets installed from source on latest `master` branch.\r\n\r\n```bash\r\n(venv) sergey_mkrtchyan datasets (master) $ python\r\nPython 3.8.6 (v3.8.6:db455296be, Sep 23 2020, 13:31:39)\r\n[Clang 6.0 (clang-600.0.57)] on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n>>> tokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\n>>> retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=True)\r\nUsing custom data configuration dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\nUsing custom data configuration dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\n 0%| | 0\/10 [00:00\", line 1, in \r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 425, in from_pretrained\r\n return cls(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 387, in __init__\r\n self.init_retrieval()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 458, in init_retrieval\r\n self.index.init_index()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 284, in init_index\r\n self.dataset = load_dataset(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 734, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/utils\/py_utils.py\", line 195, in map_nested\r\n return function(data_struct)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 769, in _build_single_dataset\r\n post_processed = self._post_process(ds, resources_paths)\r\n File \"\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb\/wiki_dpr.py\", line 205, in _post_process\r\n dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/arrow_dataset.py\", line 2516, in add_faiss_index\r\n super().add_faiss_index(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 416, in add_faiss_index\r\n faiss_index.add_vectors(self, column=column, train_size=train_size, faiss_verbose=faiss_verbose)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 281, in add_vectors\r\n self.faiss_index.add(vecs)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/__init__.py\", line 104, in replacement_add\r\n self.add_c(n, swig_ptr(x))\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/swigfaiss.py\", line 3263, in add\r\n return _swigfaiss.IndexHNSW_add(self, n, x)\r\nRuntimeError: Error in virtual void faiss::IndexHNSW::add(faiss::Index::idx_t, const float *) at \/Users\/runner\/work\/faiss-wheels\/faiss-wheels\/faiss\/faiss\/IndexHNSW.cpp:356: Error: 'is_trained' failed\r\n>>>\r\n```\r\n\r\nThe issue seems to be related to the scalar quantization in faiss added in this commit: 8c5220307c33f00e01c3bf7b8. Reverting it fixes the issue.\r\n\r\n\r\n\nThanks for reporting ! I'm taking a look","embeddings":[-0.0889257565,-0.1103842705,0.0109983413,0.0540016554,0.4101668894,-0.0569174886,0.2968126535,0.2508972585,0.2696110904,0.1997553408,-0.2217194587,0.139844045,0.136316672,-0.1376480311,-0.1762496382,0.2843903303,0.2129584849,0.123024039,-0.0870195404,-0.1029342785,-0.2702940702,0.2029289007,-0.2149687111,0.2133810818,-0.3408923149,0.4708951414,-0.1025647521,0.1511468738,-0.1406457871,-0.5798454881,0.323237896,-0.0542232022,0.2287583798,0.279283613,-0.0001177264,0.1951242834,0.5912719369,-0.0638967305,-0.3142662942,0.0912119225,0.1551293135,0.2524677217,0.188416183,-0.1187293231,-0.2891574204,-0.2747856379,0.169850722,-0.0614712201,0.1880251914,0.209688589,0.1545539945,0.1129425019,0.3915553093,-0.2391127646,0.3550436497,-0.4309919775,-0.2338058501,0.4001970589,0.0305760968,-0.0565604307,-0.0222198851,0.0928998664,0.1818179339,-0.0270642173,-0.0330323465,-0.1413348466,0.3804668784,0.0840317309,0.2681783438,0.1212422028,0.2935319543,-0.0311049707,-0.4443510175,0.0353066139,0.1401630938,0.0986513719,0.2189644426,-0.4058190286,0.0952564403,0.2862180769,0.0806031749,0.0194833223,0.0560727865,0.2368140668,-0.0019370971,0.390316844,-0.1297442168,0.0979245082,0.0764907449,-0.2108365446,-0.2144456953,0.0950188935,-0.2728638649,-0.0225361474,-0.6444880366,0.1160787269,0.0512801819,-0.2159349918,0.0352995507,0.0424009375,-0.2709305882,0.2906615436,0.1008517146,0.0380018242,-0.0305454731,0.0233890675,-0.0013722524,0.0068335258,0.1125111952,0.0044880873,-0.1529048979,-0.0514266565,-0.2755156159,-0.0540647916,-0.6907650828,-0.3722564876,-0.0145713603,-0.2811401784,-0.4073481262,-0.0867943242,-0.2471618354,0.1289564669,0.2523566484,0.7027882934,0.0485681109,-0.0792193636,0.0272349175,0.1643967777,-0.135249719,0.1639370322,-0.2256950885,-0.3065262437,-0.2085856944,-0.0223727431,0.3425043821,-0.2723258734,0.2192728817,0.0867487192,-0.0316480026,0.0625711307,-0.2385466397,-0.0671575963,0.1852024198,0.319519192,-0.2793678939,0.1558532715,0.2347674817,0.0704580098,0.0292036925,-0.0564666688,-0.1806442291,-0.3224920928,0.11471311,0.172033444,-0.0485996492,-0.1230743751,0.3416301906,0.0278548207,-0.1601530164,0.1821374446,-0.0173215792,-0.2101818025,-0.2364784926,0.0118759815,0.4402866662,0.3116513193,-0.1029743105,-0.2342440337,-0.0569787994,0.0164476484,-0.0192490257,0.1451807171,-0.1825048625,0.1064566746,-0.3530344367,0.0480398051,0.2476074845,-0.2162941098,-0.0942614898,0.0434443876,0.0055873683,0.048692733,0.0026495929,-0.2383570224,-0.0799198523,0.1082673445,0.1997901797,0.1565131098,0.2076743692,-0.1675224453,-0.2941005528,-0.454654783,0.0983941481,0.3206001818,0.0272102952,0.0472556986,0.0218186025,-0.1154953465,0.3780491352,-0.159472838,-0.0510468259,0.1109238639,0.3147558868,0.2252455205,0.5496624708,-0.1197111905,-0.0283096433,0.4137910604,-0.1743458658,0.049115818,0.0225034487,0.0811982453,-0.3368809223,0.062507309,-0.0826765969,-0.0540839732,0.050944034,0.1031437144,0.1682233065,-0.2200391442,-0.2104469389,0.2421289533,-0.3746232688,0.2127416283,-0.2988705635,0.1358976513,-0.1310333312,-0.3043181598,-0.0259588826,0.1280353069,0.3351846933,-0.2089738995,-0.1438956708,0.3537579477,0.1275344789,-0.5363496542,0.3885994852,-0.0711898431,-0.2205817252,-0.2585739493,-0.0819249153,0.4026681781,0.0992996469,0.0287675895,-0.1217490137,0.4608269036,-0.0165670477,0.2404149622,-0.0125429183,-0.13944453,0.3590926826,-0.1018888429,0.0247064345,-0.2612526119,0.1761949062,0.1161095127,0.1975640804,-0.2353008091,0.0291063692,-0.1079016179,-0.1206863746,0.1251834333,-0.0890994966,0.1797404587,-0.1933805048,-0.018068783,-0.1605866998,-0.4791119993,0.2939761579,0.2113705724,-0.2033737153,-0.2445664257,-0.0433400683,-0.2511078119,0.4474205375,0.0702858269,-0.1768762469,0.1726860702,0.175490424,-0.2408617139,-0.4516462386,-0.1435257047,0.1325661689,-0.0127125662,-0.3138774931,0.168381229,-0.0572857261,0.0419997945,-0.0414767638,-0.2650158107,-0.3554300368,-0.1907569021,0.1494047791,0.0260543432,-0.0674646646,0.4098317027,-0.1587007791,-0.0234234408,0.1372434646,-0.2612144947,-0.3281145096,-0.1754921675,-0.3357377648,0.0243652146,-0.1582147181,-0.1349006593,-0.0027408276,0.0413393416,-0.1111522913,-0.4048219919,-0.2892029285,0.2777305543,-0.0530993342,0.1911062896,0.0515837409,-0.1222086698,-0.0527465828,-0.0224399511,0.3487251997,-0.2121926695,0.0221102089,0.107708171,-0.0921973512,-0.0403692089,-0.1590484232,-0.0769054443,0.0546558984,-0.2552226484,0.2541982234,-0.098238118,0.0092389965,0.1670096666,0.095018357,0.1037694886,-0.2616454363,0.0031433697,-0.3037983179,-0.0857035294,0.3209959269,0.0239907205,-0.272444427,-0.0610823706,0.0665887445,-0.2728331387,0.2826738656,-0.0803059042,-0.2875355184,-0.0009769005,0.2117819935,0.1499058455,0.1871914864,-0.0716412887,-0.2176822573,-0.1821729988,-0.0779760256,-0.0572831295,0.3550330698,-0.1895646453,0.2287379354,0.0150354514,-0.0614278838,-0.3030913472,0.6028997898,0.0793468505,-0.2462953031,0.4174587429,-0.0564926416,0.2219815701,-0.1502774358,-0.170541361,0.2374226153,0.0112562394,0.2298963666,0.211857453,-0.0392461084,0.1808157414,0.0655932501,0.2212321162,-0.063307263,-0.2813162804,0.1477481425,0.2113944888,0.1398189068,0.0216098186,0.280366689,-0.131276533,-0.044824522,0.4365172088,0.4124360383,-0.0938058496,-0.0584578067,-0.0734546855,-0.0941996947,-0.2126452625,0.3466752768,-0.0420315489,0.6930013299,-0.0107859932,0.0099755824,0.31167835,-0.0148644494,1.0188843012,0.0143519966,0.1796170324,0.4301277101,0.0961003155,-0.5665025115,-0.1661121845,-0.1035225838,-0.1321617067,0.6401345134,0.5198693871,-0.314419508,-0.3671022356,0.1703525782,-0.014736224,-0.129188776,-0.4554663599,-0.4305740893,0.0800214857,-0.132115379,0.047169894,-0.063252978,0.2533337772,0.0612495877,0.3186868727,-0.5520010591,-0.1595086455,-0.0162109584,0.130431667,0.2088962495,-0.0035232641,0.4091517329,0.4179935455,0.2740361989,-0.0023827108,0.5052620769,-0.1862425059,-0.3868599832,0.3231752217,0.1895134002,-0.1025222763,0.4031794071,-0.1244615242,0.0519955344,-0.1901735067,0.1065804884,-0.1335679591,0.1749097258,0.0733134225,0.0210778285,-0.0387684181,-0.3867224455,0.6415421367,-0.1238808557,-0.0281991381,0.137066707,0.5084984303,-0.2649716437,0.8774344325,-0.1680248976,0.7281212211,-0.1413463056,-0.0527823642,0.2257831097,-0.1320694238,0.2003129721,-0.4170898795,0.1278454959,-0.5279301405,0.0115585346,0.1474643201,-0.0227250401,0.0728066638,0.2921898067,0.013980601,0.2591739893,0.0157226771,0.2496107966,0.0998791307,0.4061026573,0.1601365954,-0.2743454874,-0.2626038194,0.1274069995,0.2499642223,0.3748722374,0.0817507356,-0.1256569326,-0.158352226,0.0780409127,-0.1681931168,0.4189148843,-0.2086645961,0.2526067495,0.4359012246,-0.0881342143,0.0852468386,-0.385986805,0.6924887896,-0.01351876,-0.2834363878,0.4745703936,0.0331090279,-0.1972595453,-0.18509157,-0.1159422547,0.0572610758,0.1282583177,-0.1469925642,-0.110553883,-0.0610210449,-0.1273223609,-0.1444333047,-0.0581149869,0.0174226481,-0.338471055,0.0029890165,-0.1440852731,0.2218132168,-0.2125500143,0.1255456358,-0.0039676805,-0.1156826541,0.0639207587,-0.0838062391,-0.3261520267,-0.0504815504,0.4979910553,-0.3048427999,0.0334972031,0.4874071181,-0.0755845681,-0.224524796,-0.0711387247,-0.5798049569,0.4502623975,-0.4499554634,-0.060462784,0.2341181487,-0.3475360274,0.2754639685,-0.0626302361,0.1548950225,-0.3116759062,-0.1879795939,-0.1831904054,-0.0942262486,0.0886205062,0.0754399821,0.2557691634,-0.0081494488,0.1618411839,-0.2867745757,0.0140153905,-0.2770027816,0.1075770482,-0.0721936673,0.2139977068,-0.2397189438,-0.4832122624,0.2989865243,-0.0265961476,0.1560488045,0.0845014304,-0.1904972941,-0.1558907926,-0.2289702147,0.1160232946,0.0690034181,-0.1123571843,-0.4639086425,-0.0516459867,0.0010403504,-0.1568174958,0.223797664,0.2394790351,-0.0870598406,0.0214177314,0.1709785461,-0.1172266081,-0.0764697045,0.2197501659,-0.0321822092,0.3317801952,0.3021886945,0.21044074,-0.6520290971,0.0660031214,-0.1195573732,0.1622678638,0.0261330493,-0.0447577126,0.1173154116,-0.268381238,-0.4649256766,0.0705434754,0.1218460724,0.5309670568,-0.1870386451,-0.4047635496,0.166649431,0.1591293663,-0.2034615278,-0.1616839767,0.0730252191,-0.1118111014,0.1455474794,0.0349668674,0.4118672609,-0.3259596229,-0.1759105027,0.1035068184,0.2216585875,0.3152202666,-0.3731226325,-0.0197300781,0.206229344,0.0610960834,0.5360981226,0.0945846662,0.2477027476,0.6870844364,-0.0722828582,0.174462229,0.055084765,0.2124176174,-0.0244814269,-0.2739833891,0.2910193205,0.3236261308,-0.0894857869,0.0141977835,-0.0831028298,0.3698725402,0.070572041,-0.1630812287,-0.122359708,-0.0045941318,-0.0648388863,-0.4671503603,0.0585067309,-0.0793082491,-0.2404517084,0.1137233078,-0.2591261268,-0.0938860551,-0.0230938941,0.1891686618,-0.0880200714,-0.3320332766,-0.5754647255,0.2790208161,0.3300842047,-0.2184932828,0.1983409673,0.2257538587,-0.2282550484,0.3201926649,0.0007114494,0.4408040941,0.0708033517,-0.0725468919,-0.0531330444,0.0030455142,0.0104163187,-0.0953289717,0.0662751049,0.0054900516,-0.2781654298,0.2291246653,0.0866308436,-0.211282894,-0.6455811858,-0.0891911536,0.4423399866,-0.0844531581,0.1159545183,0.1088769734,0.2360854894,-0.1458783299,0.019432079,-0.4001848698,0.325729847,0.3039851189,-0.0512608439,0.0937304124,0.2689379752,0.0631137565,-0.0552850328,0.2625035644,0.2940624952,0.0623607934,-0.4068162143,-0.173008129,-0.2244133651,0.2328393459,-0.1030965075,0.3240920901,0.4435863793,0.3735142052,-0.0393700041,-0.1675020456,0.1411899477,-0.0199434068,0.2401421964,0.3795910776,-0.2141230106,0.3025226891,-0.2922196984,0.0490273647,0.0471590459,-0.3806379139,0.0363535918,-0.0276333764,-0.0515837409,-0.0018895994,-0.2796105146,0.2360591441,-0.3397530019,0.3566699326,-0.0204692706,0.0925085321,-0.0411743298,-0.2193613499,-0.3757417202,-0.3629160523,-0.144807443,0.0370066576,-0.2165197581,-0.0487192459,-0.0600931086,-0.2090692222,-0.2312405556,0.3326083124,0.0224145316,-0.1425608993,-0.1646382064,0.1618474573,-0.1754695326,0.2252351642,-0.1536867768,0.4404304922,0.1206704229,0.3056820035,-0.2830803692,-0.1984687001,0.7958436608,-0.303399086,-0.2339416444,-0.2163999677,0.3711208701,0.1836148053,0.2305790782,-0.6851292849,0.2322601229,0.2757996023,0.1451227516,-0.2586943507,-0.1075070202,0.292255342,0.0834405571,-0.3089339733,-0.0486338623,0.4374202192,0.001816342,-0.009879089,-0.071689263]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1941","title":"Loading of FAISS index fails for index_name = 'exact'","comments":"Index training was missing, I fixed it here: https:\/\/github.com\/huggingface\/datasets\/commit\/f5986c46323583989f6ed1dabaf267854424a521\r\n\r\nCan you try again please ?","body":"Hi,\r\n\r\nIt looks like loading of FAISS index now fails when using index_name = 'exact'.\r\n\r\nFor example, from the RAG [model card](https:\/\/huggingface.co\/facebook\/rag-token-nq?fbclid=IwAR3bTfhls5U_t9DqsX2Vzb7NhtRHxJxfQ-uwFT7VuCPMZUM2AdAlKF_qkI8#usage).\r\n\r\nRunning `transformers==4.3.2` and datasets installed from source on latest `master` branch.\r\n\r\n```bash\r\n(venv) sergey_mkrtchyan datasets (master) $ python\r\nPython 3.8.6 (v3.8.6:db455296be, Sep 23 2020, 13:31:39)\r\n[Clang 6.0 (clang-600.0.57)] on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n>>> tokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\n>>> retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=True)\r\nUsing custom data configuration dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\nUsing custom data configuration dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\n 0%| | 0\/10 [00:00\", line 1, in \r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 425, in from_pretrained\r\n return cls(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 387, in __init__\r\n self.init_retrieval()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 458, in init_retrieval\r\n self.index.init_index()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 284, in init_index\r\n self.dataset = load_dataset(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 734, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/utils\/py_utils.py\", line 195, in map_nested\r\n return function(data_struct)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 769, in _build_single_dataset\r\n post_processed = self._post_process(ds, resources_paths)\r\n File \"\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb\/wiki_dpr.py\", line 205, in _post_process\r\n dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/arrow_dataset.py\", line 2516, in add_faiss_index\r\n super().add_faiss_index(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 416, in add_faiss_index\r\n faiss_index.add_vectors(self, column=column, train_size=train_size, faiss_verbose=faiss_verbose)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 281, in add_vectors\r\n self.faiss_index.add(vecs)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/__init__.py\", line 104, in replacement_add\r\n self.add_c(n, swig_ptr(x))\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/swigfaiss.py\", line 3263, in add\r\n return _swigfaiss.IndexHNSW_add(self, n, x)\r\nRuntimeError: Error in virtual void faiss::IndexHNSW::add(faiss::Index::idx_t, const float *) at \/Users\/runner\/work\/faiss-wheels\/faiss-wheels\/faiss\/faiss\/IndexHNSW.cpp:356: Error: 'is_trained' failed\r\n>>>\r\n```\r\n\r\nThe issue seems to be related to the scalar quantization in faiss added in this commit: 8c5220307c33f00e01c3bf7b8. Reverting it fixes the issue.\r\n\r\n\r\n","comment_length":15,"text":"Loading of FAISS index fails for index_name = 'exact'\nHi,\r\n\r\nIt looks like loading of FAISS index now fails when using index_name = 'exact'.\r\n\r\nFor example, from the RAG [model card](https:\/\/huggingface.co\/facebook\/rag-token-nq?fbclid=IwAR3bTfhls5U_t9DqsX2Vzb7NhtRHxJxfQ-uwFT7VuCPMZUM2AdAlKF_qkI8#usage).\r\n\r\nRunning `transformers==4.3.2` and datasets installed from source on latest `master` branch.\r\n\r\n```bash\r\n(venv) sergey_mkrtchyan datasets (master) $ python\r\nPython 3.8.6 (v3.8.6:db455296be, Sep 23 2020, 13:31:39)\r\n[Clang 6.0 (clang-600.0.57)] on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n>>> tokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\n>>> retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=True)\r\nUsing custom data configuration dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\nUsing custom data configuration dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\n 0%| | 0\/10 [00:00\", line 1, in \r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 425, in from_pretrained\r\n return cls(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 387, in __init__\r\n self.init_retrieval()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 458, in init_retrieval\r\n self.index.init_index()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 284, in init_index\r\n self.dataset = load_dataset(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 734, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/utils\/py_utils.py\", line 195, in map_nested\r\n return function(data_struct)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 769, in _build_single_dataset\r\n post_processed = self._post_process(ds, resources_paths)\r\n File \"\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb\/wiki_dpr.py\", line 205, in _post_process\r\n dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/arrow_dataset.py\", line 2516, in add_faiss_index\r\n super().add_faiss_index(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 416, in add_faiss_index\r\n faiss_index.add_vectors(self, column=column, train_size=train_size, faiss_verbose=faiss_verbose)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 281, in add_vectors\r\n self.faiss_index.add(vecs)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/__init__.py\", line 104, in replacement_add\r\n self.add_c(n, swig_ptr(x))\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/swigfaiss.py\", line 3263, in add\r\n return _swigfaiss.IndexHNSW_add(self, n, x)\r\nRuntimeError: Error in virtual void faiss::IndexHNSW::add(faiss::Index::idx_t, const float *) at \/Users\/runner\/work\/faiss-wheels\/faiss-wheels\/faiss\/faiss\/IndexHNSW.cpp:356: Error: 'is_trained' failed\r\n>>>\r\n```\r\n\r\nThe issue seems to be related to the scalar quantization in faiss added in this commit: 8c5220307c33f00e01c3bf7b8. Reverting it fixes the issue.\r\n\r\n\r\n\nIndex training was missing, I fixed it here: https:\/\/github.com\/huggingface\/datasets\/commit\/f5986c46323583989f6ed1dabaf267854424a521\r\n\r\nCan you try again please ?","embeddings":[-0.0889257565,-0.1103842705,0.0109983413,0.0540016554,0.4101668894,-0.0569174886,0.2968126535,0.2508972585,0.2696110904,0.1997553408,-0.2217194587,0.139844045,0.136316672,-0.1376480311,-0.1762496382,0.2843903303,0.2129584849,0.123024039,-0.0870195404,-0.1029342785,-0.2702940702,0.2029289007,-0.2149687111,0.2133810818,-0.3408923149,0.4708951414,-0.1025647521,0.1511468738,-0.1406457871,-0.5798454881,0.323237896,-0.0542232022,0.2287583798,0.279283613,-0.0001177264,0.1951242834,0.5912719369,-0.0638967305,-0.3142662942,0.0912119225,0.1551293135,0.2524677217,0.188416183,-0.1187293231,-0.2891574204,-0.2747856379,0.169850722,-0.0614712201,0.1880251914,0.209688589,0.1545539945,0.1129425019,0.3915553093,-0.2391127646,0.3550436497,-0.4309919775,-0.2338058501,0.4001970589,0.0305760968,-0.0565604307,-0.0222198851,0.0928998664,0.1818179339,-0.0270642173,-0.0330323465,-0.1413348466,0.3804668784,0.0840317309,0.2681783438,0.1212422028,0.2935319543,-0.0311049707,-0.4443510175,0.0353066139,0.1401630938,0.0986513719,0.2189644426,-0.4058190286,0.0952564403,0.2862180769,0.0806031749,0.0194833223,0.0560727865,0.2368140668,-0.0019370971,0.390316844,-0.1297442168,0.0979245082,0.0764907449,-0.2108365446,-0.2144456953,0.0950188935,-0.2728638649,-0.0225361474,-0.6444880366,0.1160787269,0.0512801819,-0.2159349918,0.0352995507,0.0424009375,-0.2709305882,0.2906615436,0.1008517146,0.0380018242,-0.0305454731,0.0233890675,-0.0013722524,0.0068335258,0.1125111952,0.0044880873,-0.1529048979,-0.0514266565,-0.2755156159,-0.0540647916,-0.6907650828,-0.3722564876,-0.0145713603,-0.2811401784,-0.4073481262,-0.0867943242,-0.2471618354,0.1289564669,0.2523566484,0.7027882934,0.0485681109,-0.0792193636,0.0272349175,0.1643967777,-0.135249719,0.1639370322,-0.2256950885,-0.3065262437,-0.2085856944,-0.0223727431,0.3425043821,-0.2723258734,0.2192728817,0.0867487192,-0.0316480026,0.0625711307,-0.2385466397,-0.0671575963,0.1852024198,0.319519192,-0.2793678939,0.1558532715,0.2347674817,0.0704580098,0.0292036925,-0.0564666688,-0.1806442291,-0.3224920928,0.11471311,0.172033444,-0.0485996492,-0.1230743751,0.3416301906,0.0278548207,-0.1601530164,0.1821374446,-0.0173215792,-0.2101818025,-0.2364784926,0.0118759815,0.4402866662,0.3116513193,-0.1029743105,-0.2342440337,-0.0569787994,0.0164476484,-0.0192490257,0.1451807171,-0.1825048625,0.1064566746,-0.3530344367,0.0480398051,0.2476074845,-0.2162941098,-0.0942614898,0.0434443876,0.0055873683,0.048692733,0.0026495929,-0.2383570224,-0.0799198523,0.1082673445,0.1997901797,0.1565131098,0.2076743692,-0.1675224453,-0.2941005528,-0.454654783,0.0983941481,0.3206001818,0.0272102952,0.0472556986,0.0218186025,-0.1154953465,0.3780491352,-0.159472838,-0.0510468259,0.1109238639,0.3147558868,0.2252455205,0.5496624708,-0.1197111905,-0.0283096433,0.4137910604,-0.1743458658,0.049115818,0.0225034487,0.0811982453,-0.3368809223,0.062507309,-0.0826765969,-0.0540839732,0.050944034,0.1031437144,0.1682233065,-0.2200391442,-0.2104469389,0.2421289533,-0.3746232688,0.2127416283,-0.2988705635,0.1358976513,-0.1310333312,-0.3043181598,-0.0259588826,0.1280353069,0.3351846933,-0.2089738995,-0.1438956708,0.3537579477,0.1275344789,-0.5363496542,0.3885994852,-0.0711898431,-0.2205817252,-0.2585739493,-0.0819249153,0.4026681781,0.0992996469,0.0287675895,-0.1217490137,0.4608269036,-0.0165670477,0.2404149622,-0.0125429183,-0.13944453,0.3590926826,-0.1018888429,0.0247064345,-0.2612526119,0.1761949062,0.1161095127,0.1975640804,-0.2353008091,0.0291063692,-0.1079016179,-0.1206863746,0.1251834333,-0.0890994966,0.1797404587,-0.1933805048,-0.018068783,-0.1605866998,-0.4791119993,0.2939761579,0.2113705724,-0.2033737153,-0.2445664257,-0.0433400683,-0.2511078119,0.4474205375,0.0702858269,-0.1768762469,0.1726860702,0.175490424,-0.2408617139,-0.4516462386,-0.1435257047,0.1325661689,-0.0127125662,-0.3138774931,0.168381229,-0.0572857261,0.0419997945,-0.0414767638,-0.2650158107,-0.3554300368,-0.1907569021,0.1494047791,0.0260543432,-0.0674646646,0.4098317027,-0.1587007791,-0.0234234408,0.1372434646,-0.2612144947,-0.3281145096,-0.1754921675,-0.3357377648,0.0243652146,-0.1582147181,-0.1349006593,-0.0027408276,0.0413393416,-0.1111522913,-0.4048219919,-0.2892029285,0.2777305543,-0.0530993342,0.1911062896,0.0515837409,-0.1222086698,-0.0527465828,-0.0224399511,0.3487251997,-0.2121926695,0.0221102089,0.107708171,-0.0921973512,-0.0403692089,-0.1590484232,-0.0769054443,0.0546558984,-0.2552226484,0.2541982234,-0.098238118,0.0092389965,0.1670096666,0.095018357,0.1037694886,-0.2616454363,0.0031433697,-0.3037983179,-0.0857035294,0.3209959269,0.0239907205,-0.272444427,-0.0610823706,0.0665887445,-0.2728331387,0.2826738656,-0.0803059042,-0.2875355184,-0.0009769005,0.2117819935,0.1499058455,0.1871914864,-0.0716412887,-0.2176822573,-0.1821729988,-0.0779760256,-0.0572831295,0.3550330698,-0.1895646453,0.2287379354,0.0150354514,-0.0614278838,-0.3030913472,0.6028997898,0.0793468505,-0.2462953031,0.4174587429,-0.0564926416,0.2219815701,-0.1502774358,-0.170541361,0.2374226153,0.0112562394,0.2298963666,0.211857453,-0.0392461084,0.1808157414,0.0655932501,0.2212321162,-0.063307263,-0.2813162804,0.1477481425,0.2113944888,0.1398189068,0.0216098186,0.280366689,-0.131276533,-0.044824522,0.4365172088,0.4124360383,-0.0938058496,-0.0584578067,-0.0734546855,-0.0941996947,-0.2126452625,0.3466752768,-0.0420315489,0.6930013299,-0.0107859932,0.0099755824,0.31167835,-0.0148644494,1.0188843012,0.0143519966,0.1796170324,0.4301277101,0.0961003155,-0.5665025115,-0.1661121845,-0.1035225838,-0.1321617067,0.6401345134,0.5198693871,-0.314419508,-0.3671022356,0.1703525782,-0.014736224,-0.129188776,-0.4554663599,-0.4305740893,0.0800214857,-0.132115379,0.047169894,-0.063252978,0.2533337772,0.0612495877,0.3186868727,-0.5520010591,-0.1595086455,-0.0162109584,0.130431667,0.2088962495,-0.0035232641,0.4091517329,0.4179935455,0.2740361989,-0.0023827108,0.5052620769,-0.1862425059,-0.3868599832,0.3231752217,0.1895134002,-0.1025222763,0.4031794071,-0.1244615242,0.0519955344,-0.1901735067,0.1065804884,-0.1335679591,0.1749097258,0.0733134225,0.0210778285,-0.0387684181,-0.3867224455,0.6415421367,-0.1238808557,-0.0281991381,0.137066707,0.5084984303,-0.2649716437,0.8774344325,-0.1680248976,0.7281212211,-0.1413463056,-0.0527823642,0.2257831097,-0.1320694238,0.2003129721,-0.4170898795,0.1278454959,-0.5279301405,0.0115585346,0.1474643201,-0.0227250401,0.0728066638,0.2921898067,0.013980601,0.2591739893,0.0157226771,0.2496107966,0.0998791307,0.4061026573,0.1601365954,-0.2743454874,-0.2626038194,0.1274069995,0.2499642223,0.3748722374,0.0817507356,-0.1256569326,-0.158352226,0.0780409127,-0.1681931168,0.4189148843,-0.2086645961,0.2526067495,0.4359012246,-0.0881342143,0.0852468386,-0.385986805,0.6924887896,-0.01351876,-0.2834363878,0.4745703936,0.0331090279,-0.1972595453,-0.18509157,-0.1159422547,0.0572610758,0.1282583177,-0.1469925642,-0.110553883,-0.0610210449,-0.1273223609,-0.1444333047,-0.0581149869,0.0174226481,-0.338471055,0.0029890165,-0.1440852731,0.2218132168,-0.2125500143,0.1255456358,-0.0039676805,-0.1156826541,0.0639207587,-0.0838062391,-0.3261520267,-0.0504815504,0.4979910553,-0.3048427999,0.0334972031,0.4874071181,-0.0755845681,-0.224524796,-0.0711387247,-0.5798049569,0.4502623975,-0.4499554634,-0.060462784,0.2341181487,-0.3475360274,0.2754639685,-0.0626302361,0.1548950225,-0.3116759062,-0.1879795939,-0.1831904054,-0.0942262486,0.0886205062,0.0754399821,0.2557691634,-0.0081494488,0.1618411839,-0.2867745757,0.0140153905,-0.2770027816,0.1075770482,-0.0721936673,0.2139977068,-0.2397189438,-0.4832122624,0.2989865243,-0.0265961476,0.1560488045,0.0845014304,-0.1904972941,-0.1558907926,-0.2289702147,0.1160232946,0.0690034181,-0.1123571843,-0.4639086425,-0.0516459867,0.0010403504,-0.1568174958,0.223797664,0.2394790351,-0.0870598406,0.0214177314,0.1709785461,-0.1172266081,-0.0764697045,0.2197501659,-0.0321822092,0.3317801952,0.3021886945,0.21044074,-0.6520290971,0.0660031214,-0.1195573732,0.1622678638,0.0261330493,-0.0447577126,0.1173154116,-0.268381238,-0.4649256766,0.0705434754,0.1218460724,0.5309670568,-0.1870386451,-0.4047635496,0.166649431,0.1591293663,-0.2034615278,-0.1616839767,0.0730252191,-0.1118111014,0.1455474794,0.0349668674,0.4118672609,-0.3259596229,-0.1759105027,0.1035068184,0.2216585875,0.3152202666,-0.3731226325,-0.0197300781,0.206229344,0.0610960834,0.5360981226,0.0945846662,0.2477027476,0.6870844364,-0.0722828582,0.174462229,0.055084765,0.2124176174,-0.0244814269,-0.2739833891,0.2910193205,0.3236261308,-0.0894857869,0.0141977835,-0.0831028298,0.3698725402,0.070572041,-0.1630812287,-0.122359708,-0.0045941318,-0.0648388863,-0.4671503603,0.0585067309,-0.0793082491,-0.2404517084,0.1137233078,-0.2591261268,-0.0938860551,-0.0230938941,0.1891686618,-0.0880200714,-0.3320332766,-0.5754647255,0.2790208161,0.3300842047,-0.2184932828,0.1983409673,0.2257538587,-0.2282550484,0.3201926649,0.0007114494,0.4408040941,0.0708033517,-0.0725468919,-0.0531330444,0.0030455142,0.0104163187,-0.0953289717,0.0662751049,0.0054900516,-0.2781654298,0.2291246653,0.0866308436,-0.211282894,-0.6455811858,-0.0891911536,0.4423399866,-0.0844531581,0.1159545183,0.1088769734,0.2360854894,-0.1458783299,0.019432079,-0.4001848698,0.325729847,0.3039851189,-0.0512608439,0.0937304124,0.2689379752,0.0631137565,-0.0552850328,0.2625035644,0.2940624952,0.0623607934,-0.4068162143,-0.173008129,-0.2244133651,0.2328393459,-0.1030965075,0.3240920901,0.4435863793,0.3735142052,-0.0393700041,-0.1675020456,0.1411899477,-0.0199434068,0.2401421964,0.3795910776,-0.2141230106,0.3025226891,-0.2922196984,0.0490273647,0.0471590459,-0.3806379139,0.0363535918,-0.0276333764,-0.0515837409,-0.0018895994,-0.2796105146,0.2360591441,-0.3397530019,0.3566699326,-0.0204692706,0.0925085321,-0.0411743298,-0.2193613499,-0.3757417202,-0.3629160523,-0.144807443,0.0370066576,-0.2165197581,-0.0487192459,-0.0600931086,-0.2090692222,-0.2312405556,0.3326083124,0.0224145316,-0.1425608993,-0.1646382064,0.1618474573,-0.1754695326,0.2252351642,-0.1536867768,0.4404304922,0.1206704229,0.3056820035,-0.2830803692,-0.1984687001,0.7958436608,-0.303399086,-0.2339416444,-0.2163999677,0.3711208701,0.1836148053,0.2305790782,-0.6851292849,0.2322601229,0.2757996023,0.1451227516,-0.2586943507,-0.1075070202,0.292255342,0.0834405571,-0.3089339733,-0.0486338623,0.4374202192,0.001816342,-0.009879089,-0.071689263]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1941","title":"Loading of FAISS index fails for index_name = 'exact'","comments":"Works great \ud83d\udc4d I just put a minor comment on the commit, I think you meant to pass the `train_size` from the one obtained from the config.\r\n\r\nThanks for a quick response!","body":"Hi,\r\n\r\nIt looks like loading of FAISS index now fails when using index_name = 'exact'.\r\n\r\nFor example, from the RAG [model card](https:\/\/huggingface.co\/facebook\/rag-token-nq?fbclid=IwAR3bTfhls5U_t9DqsX2Vzb7NhtRHxJxfQ-uwFT7VuCPMZUM2AdAlKF_qkI8#usage).\r\n\r\nRunning `transformers==4.3.2` and datasets installed from source on latest `master` branch.\r\n\r\n```bash\r\n(venv) sergey_mkrtchyan datasets (master) $ python\r\nPython 3.8.6 (v3.8.6:db455296be, Sep 23 2020, 13:31:39)\r\n[Clang 6.0 (clang-600.0.57)] on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n>>> tokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\n>>> retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=True)\r\nUsing custom data configuration dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\nUsing custom data configuration dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\n 0%| | 0\/10 [00:00\", line 1, in \r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 425, in from_pretrained\r\n return cls(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 387, in __init__\r\n self.init_retrieval()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 458, in init_retrieval\r\n self.index.init_index()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 284, in init_index\r\n self.dataset = load_dataset(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 734, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/utils\/py_utils.py\", line 195, in map_nested\r\n return function(data_struct)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 769, in _build_single_dataset\r\n post_processed = self._post_process(ds, resources_paths)\r\n File \"\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb\/wiki_dpr.py\", line 205, in _post_process\r\n dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/arrow_dataset.py\", line 2516, in add_faiss_index\r\n super().add_faiss_index(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 416, in add_faiss_index\r\n faiss_index.add_vectors(self, column=column, train_size=train_size, faiss_verbose=faiss_verbose)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 281, in add_vectors\r\n self.faiss_index.add(vecs)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/__init__.py\", line 104, in replacement_add\r\n self.add_c(n, swig_ptr(x))\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/swigfaiss.py\", line 3263, in add\r\n return _swigfaiss.IndexHNSW_add(self, n, x)\r\nRuntimeError: Error in virtual void faiss::IndexHNSW::add(faiss::Index::idx_t, const float *) at \/Users\/runner\/work\/faiss-wheels\/faiss-wheels\/faiss\/faiss\/IndexHNSW.cpp:356: Error: 'is_trained' failed\r\n>>>\r\n```\r\n\r\nThe issue seems to be related to the scalar quantization in faiss added in this commit: 8c5220307c33f00e01c3bf7b8. Reverting it fixes the issue.\r\n\r\n\r\n","comment_length":32,"text":"Loading of FAISS index fails for index_name = 'exact'\nHi,\r\n\r\nIt looks like loading of FAISS index now fails when using index_name = 'exact'.\r\n\r\nFor example, from the RAG [model card](https:\/\/huggingface.co\/facebook\/rag-token-nq?fbclid=IwAR3bTfhls5U_t9DqsX2Vzb7NhtRHxJxfQ-uwFT7VuCPMZUM2AdAlKF_qkI8#usage).\r\n\r\nRunning `transformers==4.3.2` and datasets installed from source on latest `master` branch.\r\n\r\n```bash\r\n(venv) sergey_mkrtchyan datasets (master) $ python\r\nPython 3.8.6 (v3.8.6:db455296be, Sep 23 2020, 13:31:39)\r\n[Clang 6.0 (clang-600.0.57)] on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n>>> tokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\n>>> retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=True)\r\nUsing custom data configuration dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.no_index-dummy=True,with_index=False\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\nUsing custom data configuration dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\r\nReusing dataset wiki_dpr (\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/datasets\/wiki_dpr\/dummy.psgs_w100.nq.exact-50b6cda57ff32ab4\/0.0.0\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb)\r\n 0%| | 0\/10 [00:00\", line 1, in \r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 425, in from_pretrained\r\n return cls(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 387, in __init__\r\n self.init_retrieval()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 458, in init_retrieval\r\n self.index.init_index()\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/transformers\/models\/rag\/retrieval_rag.py\", line 284, in init_index\r\n self.dataset = load_dataset(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/load.py\", line 750, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 734, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/utils\/py_utils.py\", line 195, in map_nested\r\n return function(data_struct)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/builder.py\", line 769, in _build_single_dataset\r\n post_processed = self._post_process(ds, resources_paths)\r\n File \"\/Users\/sergey_mkrtchyan\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/8a97e0f4fa5bc46e179474db6a61b09d5d2419d2911835bd3f91d110c936d8bb\/wiki_dpr.py\", line 205, in _post_process\r\n dataset.add_faiss_index(\"embeddings\", custom_index=index)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/arrow_dataset.py\", line 2516, in add_faiss_index\r\n super().add_faiss_index(\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 416, in add_faiss_index\r\n faiss_index.add_vectors(self, column=column, train_size=train_size, faiss_verbose=faiss_verbose)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/huggingface\/datasets\/src\/datasets\/search.py\", line 281, in add_vectors\r\n self.faiss_index.add(vecs)\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/__init__.py\", line 104, in replacement_add\r\n self.add_c(n, swig_ptr(x))\r\n File \"\/Users\/sergey_mkrtchyan\/workspace\/cformers\/venv\/lib\/python3.8\/site-packages\/faiss\/swigfaiss.py\", line 3263, in add\r\n return _swigfaiss.IndexHNSW_add(self, n, x)\r\nRuntimeError: Error in virtual void faiss::IndexHNSW::add(faiss::Index::idx_t, const float *) at \/Users\/runner\/work\/faiss-wheels\/faiss-wheels\/faiss\/faiss\/IndexHNSW.cpp:356: Error: 'is_trained' failed\r\n>>>\r\n```\r\n\r\nThe issue seems to be related to the scalar quantization in faiss added in this commit: 8c5220307c33f00e01c3bf7b8. Reverting it fixes the issue.\r\n\r\n\r\n\nWorks great \ud83d\udc4d I just put a minor comment on the commit, I think you meant to pass the `train_size` from the one obtained from the config.\r\n\r\nThanks for a quick response!","embeddings":[-0.0889257565,-0.1103842705,0.0109983413,0.0540016554,0.4101668894,-0.0569174886,0.2968126535,0.2508972585,0.2696110904,0.1997553408,-0.2217194587,0.139844045,0.136316672,-0.1376480311,-0.1762496382,0.2843903303,0.2129584849,0.123024039,-0.0870195404,-0.1029342785,-0.2702940702,0.2029289007,-0.2149687111,0.2133810818,-0.3408923149,0.4708951414,-0.1025647521,0.1511468738,-0.1406457871,-0.5798454881,0.323237896,-0.0542232022,0.2287583798,0.279283613,-0.0001177264,0.1951242834,0.5912719369,-0.0638967305,-0.3142662942,0.0912119225,0.1551293135,0.2524677217,0.188416183,-0.1187293231,-0.2891574204,-0.2747856379,0.169850722,-0.0614712201,0.1880251914,0.209688589,0.1545539945,0.1129425019,0.3915553093,-0.2391127646,0.3550436497,-0.4309919775,-0.2338058501,0.4001970589,0.0305760968,-0.0565604307,-0.0222198851,0.0928998664,0.1818179339,-0.0270642173,-0.0330323465,-0.1413348466,0.3804668784,0.0840317309,0.2681783438,0.1212422028,0.2935319543,-0.0311049707,-0.4443510175,0.0353066139,0.1401630938,0.0986513719,0.2189644426,-0.4058190286,0.0952564403,0.2862180769,0.0806031749,0.0194833223,0.0560727865,0.2368140668,-0.0019370971,0.390316844,-0.1297442168,0.0979245082,0.0764907449,-0.2108365446,-0.2144456953,0.0950188935,-0.2728638649,-0.0225361474,-0.6444880366,0.1160787269,0.0512801819,-0.2159349918,0.0352995507,0.0424009375,-0.2709305882,0.2906615436,0.1008517146,0.0380018242,-0.0305454731,0.0233890675,-0.0013722524,0.0068335258,0.1125111952,0.0044880873,-0.1529048979,-0.0514266565,-0.2755156159,-0.0540647916,-0.6907650828,-0.3722564876,-0.0145713603,-0.2811401784,-0.4073481262,-0.0867943242,-0.2471618354,0.1289564669,0.2523566484,0.7027882934,0.0485681109,-0.0792193636,0.0272349175,0.1643967777,-0.135249719,0.1639370322,-0.2256950885,-0.3065262437,-0.2085856944,-0.0223727431,0.3425043821,-0.2723258734,0.2192728817,0.0867487192,-0.0316480026,0.0625711307,-0.2385466397,-0.0671575963,0.1852024198,0.319519192,-0.2793678939,0.1558532715,0.2347674817,0.0704580098,0.0292036925,-0.0564666688,-0.1806442291,-0.3224920928,0.11471311,0.172033444,-0.0485996492,-0.1230743751,0.3416301906,0.0278548207,-0.1601530164,0.1821374446,-0.0173215792,-0.2101818025,-0.2364784926,0.0118759815,0.4402866662,0.3116513193,-0.1029743105,-0.2342440337,-0.0569787994,0.0164476484,-0.0192490257,0.1451807171,-0.1825048625,0.1064566746,-0.3530344367,0.0480398051,0.2476074845,-0.2162941098,-0.0942614898,0.0434443876,0.0055873683,0.048692733,0.0026495929,-0.2383570224,-0.0799198523,0.1082673445,0.1997901797,0.1565131098,0.2076743692,-0.1675224453,-0.2941005528,-0.454654783,0.0983941481,0.3206001818,0.0272102952,0.0472556986,0.0218186025,-0.1154953465,0.3780491352,-0.159472838,-0.0510468259,0.1109238639,0.3147558868,0.2252455205,0.5496624708,-0.1197111905,-0.0283096433,0.4137910604,-0.1743458658,0.049115818,0.0225034487,0.0811982453,-0.3368809223,0.062507309,-0.0826765969,-0.0540839732,0.050944034,0.1031437144,0.1682233065,-0.2200391442,-0.2104469389,0.2421289533,-0.3746232688,0.2127416283,-0.2988705635,0.1358976513,-0.1310333312,-0.3043181598,-0.0259588826,0.1280353069,0.3351846933,-0.2089738995,-0.1438956708,0.3537579477,0.1275344789,-0.5363496542,0.3885994852,-0.0711898431,-0.2205817252,-0.2585739493,-0.0819249153,0.4026681781,0.0992996469,0.0287675895,-0.1217490137,0.4608269036,-0.0165670477,0.2404149622,-0.0125429183,-0.13944453,0.3590926826,-0.1018888429,0.0247064345,-0.2612526119,0.1761949062,0.1161095127,0.1975640804,-0.2353008091,0.0291063692,-0.1079016179,-0.1206863746,0.1251834333,-0.0890994966,0.1797404587,-0.1933805048,-0.018068783,-0.1605866998,-0.4791119993,0.2939761579,0.2113705724,-0.2033737153,-0.2445664257,-0.0433400683,-0.2511078119,0.4474205375,0.0702858269,-0.1768762469,0.1726860702,0.175490424,-0.2408617139,-0.4516462386,-0.1435257047,0.1325661689,-0.0127125662,-0.3138774931,0.168381229,-0.0572857261,0.0419997945,-0.0414767638,-0.2650158107,-0.3554300368,-0.1907569021,0.1494047791,0.0260543432,-0.0674646646,0.4098317027,-0.1587007791,-0.0234234408,0.1372434646,-0.2612144947,-0.3281145096,-0.1754921675,-0.3357377648,0.0243652146,-0.1582147181,-0.1349006593,-0.0027408276,0.0413393416,-0.1111522913,-0.4048219919,-0.2892029285,0.2777305543,-0.0530993342,0.1911062896,0.0515837409,-0.1222086698,-0.0527465828,-0.0224399511,0.3487251997,-0.2121926695,0.0221102089,0.107708171,-0.0921973512,-0.0403692089,-0.1590484232,-0.0769054443,0.0546558984,-0.2552226484,0.2541982234,-0.098238118,0.0092389965,0.1670096666,0.095018357,0.1037694886,-0.2616454363,0.0031433697,-0.3037983179,-0.0857035294,0.3209959269,0.0239907205,-0.272444427,-0.0610823706,0.0665887445,-0.2728331387,0.2826738656,-0.0803059042,-0.2875355184,-0.0009769005,0.2117819935,0.1499058455,0.1871914864,-0.0716412887,-0.2176822573,-0.1821729988,-0.0779760256,-0.0572831295,0.3550330698,-0.1895646453,0.2287379354,0.0150354514,-0.0614278838,-0.3030913472,0.6028997898,0.0793468505,-0.2462953031,0.4174587429,-0.0564926416,0.2219815701,-0.1502774358,-0.170541361,0.2374226153,0.0112562394,0.2298963666,0.211857453,-0.0392461084,0.1808157414,0.0655932501,0.2212321162,-0.063307263,-0.2813162804,0.1477481425,0.2113944888,0.1398189068,0.0216098186,0.280366689,-0.131276533,-0.044824522,0.4365172088,0.4124360383,-0.0938058496,-0.0584578067,-0.0734546855,-0.0941996947,-0.2126452625,0.3466752768,-0.0420315489,0.6930013299,-0.0107859932,0.0099755824,0.31167835,-0.0148644494,1.0188843012,0.0143519966,0.1796170324,0.4301277101,0.0961003155,-0.5665025115,-0.1661121845,-0.1035225838,-0.1321617067,0.6401345134,0.5198693871,-0.314419508,-0.3671022356,0.1703525782,-0.014736224,-0.129188776,-0.4554663599,-0.4305740893,0.0800214857,-0.132115379,0.047169894,-0.063252978,0.2533337772,0.0612495877,0.3186868727,-0.5520010591,-0.1595086455,-0.0162109584,0.130431667,0.2088962495,-0.0035232641,0.4091517329,0.4179935455,0.2740361989,-0.0023827108,0.5052620769,-0.1862425059,-0.3868599832,0.3231752217,0.1895134002,-0.1025222763,0.4031794071,-0.1244615242,0.0519955344,-0.1901735067,0.1065804884,-0.1335679591,0.1749097258,0.0733134225,0.0210778285,-0.0387684181,-0.3867224455,0.6415421367,-0.1238808557,-0.0281991381,0.137066707,0.5084984303,-0.2649716437,0.8774344325,-0.1680248976,0.7281212211,-0.1413463056,-0.0527823642,0.2257831097,-0.1320694238,0.2003129721,-0.4170898795,0.1278454959,-0.5279301405,0.0115585346,0.1474643201,-0.0227250401,0.0728066638,0.2921898067,0.013980601,0.2591739893,0.0157226771,0.2496107966,0.0998791307,0.4061026573,0.1601365954,-0.2743454874,-0.2626038194,0.1274069995,0.2499642223,0.3748722374,0.0817507356,-0.1256569326,-0.158352226,0.0780409127,-0.1681931168,0.4189148843,-0.2086645961,0.2526067495,0.4359012246,-0.0881342143,0.0852468386,-0.385986805,0.6924887896,-0.01351876,-0.2834363878,0.4745703936,0.0331090279,-0.1972595453,-0.18509157,-0.1159422547,0.0572610758,0.1282583177,-0.1469925642,-0.110553883,-0.0610210449,-0.1273223609,-0.1444333047,-0.0581149869,0.0174226481,-0.338471055,0.0029890165,-0.1440852731,0.2218132168,-0.2125500143,0.1255456358,-0.0039676805,-0.1156826541,0.0639207587,-0.0838062391,-0.3261520267,-0.0504815504,0.4979910553,-0.3048427999,0.0334972031,0.4874071181,-0.0755845681,-0.224524796,-0.0711387247,-0.5798049569,0.4502623975,-0.4499554634,-0.060462784,0.2341181487,-0.3475360274,0.2754639685,-0.0626302361,0.1548950225,-0.3116759062,-0.1879795939,-0.1831904054,-0.0942262486,0.0886205062,0.0754399821,0.2557691634,-0.0081494488,0.1618411839,-0.2867745757,0.0140153905,-0.2770027816,0.1075770482,-0.0721936673,0.2139977068,-0.2397189438,-0.4832122624,0.2989865243,-0.0265961476,0.1560488045,0.0845014304,-0.1904972941,-0.1558907926,-0.2289702147,0.1160232946,0.0690034181,-0.1123571843,-0.4639086425,-0.0516459867,0.0010403504,-0.1568174958,0.223797664,0.2394790351,-0.0870598406,0.0214177314,0.1709785461,-0.1172266081,-0.0764697045,0.2197501659,-0.0321822092,0.3317801952,0.3021886945,0.21044074,-0.6520290971,0.0660031214,-0.1195573732,0.1622678638,0.0261330493,-0.0447577126,0.1173154116,-0.268381238,-0.4649256766,0.0705434754,0.1218460724,0.5309670568,-0.1870386451,-0.4047635496,0.166649431,0.1591293663,-0.2034615278,-0.1616839767,0.0730252191,-0.1118111014,0.1455474794,0.0349668674,0.4118672609,-0.3259596229,-0.1759105027,0.1035068184,0.2216585875,0.3152202666,-0.3731226325,-0.0197300781,0.206229344,0.0610960834,0.5360981226,0.0945846662,0.2477027476,0.6870844364,-0.0722828582,0.174462229,0.055084765,0.2124176174,-0.0244814269,-0.2739833891,0.2910193205,0.3236261308,-0.0894857869,0.0141977835,-0.0831028298,0.3698725402,0.070572041,-0.1630812287,-0.122359708,-0.0045941318,-0.0648388863,-0.4671503603,0.0585067309,-0.0793082491,-0.2404517084,0.1137233078,-0.2591261268,-0.0938860551,-0.0230938941,0.1891686618,-0.0880200714,-0.3320332766,-0.5754647255,0.2790208161,0.3300842047,-0.2184932828,0.1983409673,0.2257538587,-0.2282550484,0.3201926649,0.0007114494,0.4408040941,0.0708033517,-0.0725468919,-0.0531330444,0.0030455142,0.0104163187,-0.0953289717,0.0662751049,0.0054900516,-0.2781654298,0.2291246653,0.0866308436,-0.211282894,-0.6455811858,-0.0891911536,0.4423399866,-0.0844531581,0.1159545183,0.1088769734,0.2360854894,-0.1458783299,0.019432079,-0.4001848698,0.325729847,0.3039851189,-0.0512608439,0.0937304124,0.2689379752,0.0631137565,-0.0552850328,0.2625035644,0.2940624952,0.0623607934,-0.4068162143,-0.173008129,-0.2244133651,0.2328393459,-0.1030965075,0.3240920901,0.4435863793,0.3735142052,-0.0393700041,-0.1675020456,0.1411899477,-0.0199434068,0.2401421964,0.3795910776,-0.2141230106,0.3025226891,-0.2922196984,0.0490273647,0.0471590459,-0.3806379139,0.0363535918,-0.0276333764,-0.0515837409,-0.0018895994,-0.2796105146,0.2360591441,-0.3397530019,0.3566699326,-0.0204692706,0.0925085321,-0.0411743298,-0.2193613499,-0.3757417202,-0.3629160523,-0.144807443,0.0370066576,-0.2165197581,-0.0487192459,-0.0600931086,-0.2090692222,-0.2312405556,0.3326083124,0.0224145316,-0.1425608993,-0.1646382064,0.1618474573,-0.1754695326,0.2252351642,-0.1536867768,0.4404304922,0.1206704229,0.3056820035,-0.2830803692,-0.1984687001,0.7958436608,-0.303399086,-0.2339416444,-0.2163999677,0.3711208701,0.1836148053,0.2305790782,-0.6851292849,0.2322601229,0.2757996023,0.1451227516,-0.2586943507,-0.1075070202,0.292255342,0.0834405571,-0.3089339733,-0.0486338623,0.4374202192,0.001816342,-0.009879089,-0.071689263]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1940","title":"Side effect when filtering data due to `does_function_return_dict` call in `Dataset.map()`","comments":"Thanks for the report !\r\n\r\nCurrently we don't have a way to let the user easily disable this behavior.\r\nHowever I agree that we should support stateful processing functions, ideally by removing `does_function_return_dict`.\r\n\r\nWe needed this function in order to know whether the `map` functions needs to write data or not. if `does_function_return_dict` returns False then we don't write anything.\r\n\r\nInstead of checking the output of the processing function outside of the for loop that iterates through the dataset to process it, we can check the output of the first processed example and at that point decide if we need to write data or not.\r\n\r\nTherefore it's definitely possible to fix this unwanted behavior, any contribution going into this direction is welcome :)","body":"Hi there!\r\n\r\nIn my codebase I have a function to filter rows in a dataset, selecting only a certain number of examples per class. The function passes a extra argument to maintain a counter of the number of dataset rows\/examples already selected per each class, which are the ones I want to keep in the end:\r\n\r\n```python\r\n def fill_train_examples_per_class(example, per_class_limit: int, counter: collections.Counter):\r\n label = int(example['label'])\r\n current_counter = counter.get(label, 0)\r\n if current_counter < per_class_limit:\r\n counter[label] = current_counter + 1\r\n return True\r\n return False\r\n```\r\n\r\nAt some point I invoke it through the `Dataset.filter()` method in the `arrow_dataset.py` module like this:\r\n\r\n```python\r\n...\r\nkwargs = {\"per_class_limit\": train_examples_per_class_limit, \"counter\": Counter()}\r\ndatasets['train'] = datasets['train'].filter(fill_train_examples_per_class, num_proc=1, fn_kwargs=kwargs)\r\n...\r\n```\r\n\r\nThe problem is that, passing a stateful container (the counter,) provokes a side effect in the new filtered dataset obtained. This is due to the fact that at some point in `filter()`, the `map()`'s function `does_function_return_dict` is invoked in line [1290](https:\/\/github.com\/huggingface\/datasets\/blob\/96578adface7e4bc1f3e8bafbac920d72ca1ca60\/src\/datasets\/arrow_dataset.py#L1290). \r\n\r\nWhen this occurs, the state of the counter is initially modified by the effects of the function call on the 1 or 2 rows selected in lines 1288 and 1289 of the same file (which are marked as `test_inputs` & `test_indices` respectively in lines 1288 and 1289. This happens out of the control of the user (which for example can't reset the state of the counter before continuing the execution,) provoking in the end an undesired side effect in the results obtained. \r\n\r\nIn my case, the resulting dataset -despite of the counter results are ok- lacks an instance of the classes 0 and 1 (which happen to be the classes of the first two examples of my dataset.) The rest of the classes I have in my dataset, contain the right number of examples as they were not affected by the effects of `does_function_return_dict` call.\r\n\r\nI've debugged my code extensively and made a workaround myself hardcoding the necessary stuff (basically putting `update_data=True` in line 1290,) and then I obtain the results I expected without the side effect.\r\n\r\nIs there a way to avoid that call to `does_function_return_dict` in map()'s line 1290 ? (e.g. extracting the required information that `does_function_return_dict` returns without making the testing calls to the user function on dataset rows 0 & 1) \r\n\r\nThanks in advance,\r\n\r\nFrancisco Perez-Sorrosal\r\n\r\n","comment_length":123,"text":"Side effect when filtering data due to `does_function_return_dict` call in `Dataset.map()`\nHi there!\r\n\r\nIn my codebase I have a function to filter rows in a dataset, selecting only a certain number of examples per class. The function passes a extra argument to maintain a counter of the number of dataset rows\/examples already selected per each class, which are the ones I want to keep in the end:\r\n\r\n```python\r\n def fill_train_examples_per_class(example, per_class_limit: int, counter: collections.Counter):\r\n label = int(example['label'])\r\n current_counter = counter.get(label, 0)\r\n if current_counter < per_class_limit:\r\n counter[label] = current_counter + 1\r\n return True\r\n return False\r\n```\r\n\r\nAt some point I invoke it through the `Dataset.filter()` method in the `arrow_dataset.py` module like this:\r\n\r\n```python\r\n...\r\nkwargs = {\"per_class_limit\": train_examples_per_class_limit, \"counter\": Counter()}\r\ndatasets['train'] = datasets['train'].filter(fill_train_examples_per_class, num_proc=1, fn_kwargs=kwargs)\r\n...\r\n```\r\n\r\nThe problem is that, passing a stateful container (the counter,) provokes a side effect in the new filtered dataset obtained. This is due to the fact that at some point in `filter()`, the `map()`'s function `does_function_return_dict` is invoked in line [1290](https:\/\/github.com\/huggingface\/datasets\/blob\/96578adface7e4bc1f3e8bafbac920d72ca1ca60\/src\/datasets\/arrow_dataset.py#L1290). \r\n\r\nWhen this occurs, the state of the counter is initially modified by the effects of the function call on the 1 or 2 rows selected in lines 1288 and 1289 of the same file (which are marked as `test_inputs` & `test_indices` respectively in lines 1288 and 1289. This happens out of the control of the user (which for example can't reset the state of the counter before continuing the execution,) provoking in the end an undesired side effect in the results obtained. \r\n\r\nIn my case, the resulting dataset -despite of the counter results are ok- lacks an instance of the classes 0 and 1 (which happen to be the classes of the first two examples of my dataset.) The rest of the classes I have in my dataset, contain the right number of examples as they were not affected by the effects of `does_function_return_dict` call.\r\n\r\nI've debugged my code extensively and made a workaround myself hardcoding the necessary stuff (basically putting `update_data=True` in line 1290,) and then I obtain the results I expected without the side effect.\r\n\r\nIs there a way to avoid that call to `does_function_return_dict` in map()'s line 1290 ? (e.g. extracting the required information that `does_function_return_dict` returns without making the testing calls to the user function on dataset rows 0 & 1) \r\n\r\nThanks in advance,\r\n\r\nFrancisco Perez-Sorrosal\r\n\r\n\nThanks for the report !\r\n\r\nCurrently we don't have a way to let the user easily disable this behavior.\r\nHowever I agree that we should support stateful processing functions, ideally by removing `does_function_return_dict`.\r\n\r\nWe needed this function in order to know whether the `map` functions needs to write data or not. if `does_function_return_dict` returns False then we don't write anything.\r\n\r\nInstead of checking the output of the processing function outside of the for loop that iterates through the dataset to process it, we can check the output of the first processed example and at that point decide if we need to write data or not.\r\n\r\nTherefore it's definitely possible to fix this unwanted behavior, any contribution going into this direction is welcome :)","embeddings":[-0.4214114249,0.0289576147,-0.1575038433,0.0184577908,-0.084425889,-0.3201851845,0.2051187158,0.1946579814,0.2660327554,0.1206628382,0.2073837817,0.5199099183,-0.0788978338,0.0967270359,0.0301455557,0.2262496203,0.1668142974,-0.032501027,-0.1896086782,-0.1237709373,-0.2070304006,-0.0553835072,-0.2351946831,0.0294677727,0.1121999249,-0.2066931427,0.2833378315,0.0864987075,0.1280570924,-0.233986333,0.379776001,0.2778170407,-0.3832846582,0.4839723408,-0.0001190907,0.0343095772,0.1299140453,-0.0802711844,-0.2356485426,-0.4382512271,-0.2481366247,0.0078954305,0.1348444819,-0.1925271153,0.2354865968,-0.1303691268,-0.4192903936,-0.3500099182,0.5882077217,0.2834120989,0.1079882756,0.105306901,-0.343233794,0.1123894006,0.4360818267,0.3156699836,0.1410355121,0.3769839108,0.2226104736,-0.4172928631,-0.089796707,0.6554317474,-0.419348985,0.1199867725,-0.1365879774,0.1172127947,0.1939310879,0.0126645528,0.3233865798,0.1725642383,0.1055694222,-0.3248958588,-0.1459566951,-0.420864135,-0.1007794887,0.0005761604,0.0750974044,-0.2886882424,-0.2021785378,-0.1206648573,-0.1208331883,-0.012387082,-0.0200270172,0.1681484133,-0.2169067562,0.1855434477,0.0811168924,0.2631781995,0.0817807019,-0.0514533445,0.2090521157,-0.3612852395,0.2060837448,0.2147341669,0.1196141541,0.0732956082,0.3768185079,-0.4789855778,0.1241235957,-0.13665995,0.1146184802,0.1747404039,-0.2579136491,0.2301839143,0.4253011346,0.085000217,-0.1979258806,0.6640278101,0.1729907244,-0.1068491042,-0.0037359183,0.0037666485,0.2818769813,0.0269299001,0.2999613285,-0.1395051032,0.1103360206,-0.1578751206,-0.2358426154,0.4175386727,-0.3590551019,0.2780917585,0.0907847881,0.0966719389,0.3004828691,0.4036149085,0.1682981998,-0.1376380175,0.003388444,0.5379956365,-0.1066150218,-0.3244780898,-0.2710621357,-0.0770937875,-0.011990536,0.0786589086,0.0169853605,0.1104542688,-0.0109124994,-0.0541719161,0.0757818967,-0.2363334149,0.7913022637,0.4118511677,-0.3480505049,0.1737432331,0.2374269217,-0.2243672758,-0.1260404438,0.2483870536,-0.0546916015,-0.1973662376,0.1852192581,0.1147135645,0.0059861485,0.3745835125,-0.0160887148,0.1991495788,0.4185599089,-0.471998781,0.3922097385,-0.5384453535,-0.3335294724,-0.2172567099,0.1258204728,0.525978744,-0.819714427,-0.1214056835,-0.2428824008,-0.1178346276,0.0761058852,0.1740315706,-0.2996686697,-0.0011380088,-0.2255411148,0.1341030598,0.3071460724,-0.4831515849,-0.705484271,0.0982058793,-0.2767548561,0.499204278,-0.1460386366,0.0213056188,0.3568475246,-0.008165176,0.3267332315,0.1796566099,-0.0669032261,0.1762675196,0.0203271583,0.1293668896,0.3243874311,-0.0954314321,-0.0437130406,-0.077088058,-0.074971199,-0.4875876904,0.281059891,0.0821083933,0.1924644411,-0.3172644973,0.4052571058,0.1898687333,0.0079568848,-0.2883609235,-0.2915593386,0.0308367405,0.0936341137,-0.24900195,-0.2050819397,-0.1570568383,-0.1389681697,0.3914354146,-0.0558801778,0.1572161317,0.1144860759,-0.0418155044,0.00553917,0.0628047958,0.0852909014,0.080448173,-0.4057528675,-0.0257785544,0.0938627049,0.2385653853,0.3390497267,-0.2808124125,-0.2351316363,0.1347765326,0.0226554684,-0.0478546508,-0.2866383493,0.219219178,0.5091303587,-0.0345370322,-0.2714932263,0.0800034404,-0.1027276367,0.4487369657,-0.0084447358,0.3058345914,0.1903862655,-0.2705367208,-0.0997638702,0.4954223633,0.0765313879,0.4916887283,-0.1645125449,0.136452347,0.2814358473,0.2763494253,-0.3366055191,-0.5444409847,-0.2632301152,-0.0805498734,0.1870295554,-0.1061945185,-0.155970186,-0.0380207151,0.0843561813,-0.0368897617,0.1383838505,0.103854686,-0.0823239088,-0.000345651,0.4746937752,0.4552512467,0.5759060979,0.0555185191,-0.1408450902,0.0276647117,-0.0354824476,-0.0613747053,0.1117442772,0.2467335612,0.1628268063,0.3262067139,0.3910287619,-0.1081223041,-0.2600623667,-0.1121191531,-0.1188485101,-0.1379218549,-0.2318265587,0.3826011419,-0.1168807521,-0.2817385197,-0.2118333727,-0.1495019346,0.051916115,-0.1294428259,-0.0069106244,0.4019418955,-0.3745889664,0.0674525797,-0.0789706782,0.3998449147,0.1533491313,-0.4363389611,0.1452000886,-0.3576052785,-0.1495454907,-0.0052298307,0.0020299521,0.0059776395,0.4475282729,0.282324791,-0.210486576,0.0630706325,-0.1092976108,0.1017176583,-0.007605616,-0.0080981012,0.3318776488,0.0187597405,-0.2166742086,0.0701481327,0.0484458171,0.0242532827,-0.1498593092,-0.1106217429,0.0191021357,0.2351116985,-0.2889119983,0.1924333423,-0.1723965108,-0.1195878163,0.1122066975,-0.1309978515,0.2291354388,0.1243820488,0.2222427875,0.0680092126,-0.0073051597,-0.2129320949,-0.4279069602,-0.0134776644,-0.1785435081,-0.144994691,0.0711149648,-0.2463677824,-0.2123691887,0.14317213,0.597771585,-0.5670164824,-0.198012501,0.1690313816,0.280459553,-0.1854302883,0.0043560239,0.3448859155,0.1838807613,-0.0802197531,-0.3217480481,-0.0644872263,0.1970244944,0.3125908077,0.0231229048,0.0797691122,0.1033611,0.0153850578,0.6843212247,0.2735551894,-0.1441142112,-0.0708047748,0.2465561479,0.2138903141,-0.3259014189,-0.0774517357,-0.34776932,-0.233711198,-0.1915657371,-0.1238080114,0.0483663194,-0.2100283802,0.1475331187,-0.1117216647,-0.2634606957,-0.457383126,0.2509599328,-0.2607240379,0.3891144395,0.0045573306,-0.0390345082,-0.3957884908,0.0963751078,-0.0857025608,-0.2023214102,0.3929489851,-0.2158955038,-0.2131988704,-0.0444583893,0.0876910016,0.2771394849,0.3724202514,0.2909666598,-0.006964514,-0.1946190298,-0.0075320178,0.1920891851,0.921217382,-0.1532570869,0.1145694852,0.3606545031,-0.163496539,-0.4281956255,-0.0288701728,-0.3811837137,0.3040297031,0.1288411915,0.136218667,-0.1125409603,0.0810653344,-0.1752343178,0.1800236255,-0.2305070162,-0.0291542169,-0.0458450951,0.0430700332,-0.2418702394,0.4140961468,-0.042652607,-0.3411511481,-0.0270895604,-0.3010748923,0.1418939084,-0.2788738906,-0.1683320403,0.0251771566,0.3258301616,0.0281419717,-0.033845149,0.4214897752,-0.044580318,0.3595966697,0.2097106278,-0.1842389107,-0.099346891,-0.3385727406,0.2808583081,0.0884962678,0.6351366043,-0.0662128776,0.292578578,0.0539201871,0.0340458862,0.0468430892,0.0528456457,0.1087765023,0.3228063583,0.0372582451,-0.5029551387,0.2584246993,0.1158447117,-0.4988925457,0.5093582869,0.070716247,-0.4502280354,0.500374794,0.0967449993,0.8437870145,0.3119137287,0.1658471674,0.2260128111,-0.1717870831,0.0884205475,0.2525969744,-0.0557101145,0.1190145686,0.152756229,-0.0358197615,-0.2901842594,-0.0203851238,-0.1074931622,0.0465300791,0.0796530992,-0.5716618896,0.3465054035,-0.1252949089,-0.2016382217,0.2813823521,-0.3633782268,-0.0759189278,0.1183448583,0.1558741033,0.0072763762,-0.117448777,0.1039035842,0.0316315517,-0.1071537808,-0.1751724333,-0.2920619845,-0.4194130003,0.4033285081,0.0758899152,-0.4115694761,-0.0596608073,-0.0202501994,-0.4361622334,0.2700872421,0.0537092462,-0.0371725373,0.4989475608,0.5238935351,0.1007662788,-0.269499898,0.4783497453,0.1840208918,-0.2807706296,0.0061873645,-0.0833480135,0.0214813873,-0.1123805195,0.0961127952,0.375865072,-0.2928257287,0.0643375739,0.0271966532,-0.0029291124,-0.1024997309,0.0415786356,-0.0710347444,-0.0433687679,0.2548505962,0.0600099824,-0.035591241,-0.0283804033,0.2195603997,0.1335608512,0.2223019749,0.37287274,-0.1903460473,-0.1306002885,-0.1904115379,-0.2018484175,0.1169957742,0.0854392275,0.4244581461,0.0115424357,-0.2601805031,-0.3701823056,0.1487354338,-0.0050651985,0.2873671055,-0.5665975213,-0.3418369591,-0.2227964103,0.1422403306,0.121300377,0.1461369097,-0.2059533745,0.2034112662,-0.3286547363,0.1982479841,-0.2763057649,-0.088812381,0.0760137886,0.3025518358,0.072930567,0.2834870219,-0.2598807812,-0.1926088929,0.022234913,0.4459893405,-0.245225504,-0.1933341473,-0.0131521011,0.1412336379,-0.0306064282,-0.3593326807,0.0489733629,-0.3224500418,0.1845193654,-0.2329692096,0.4509948194,0.1140588447,-0.0886011794,-0.2661243379,-0.0722008049,0.152037099,0.3301406503,0.4199129045,-0.282502681,-0.266082108,-0.0245819353,0.1808160841,0.065408662,-0.0692419931,-0.06781739,0.188092038,0.1825224161,0.1244096756,0.1734904349,-0.0926332772,-0.2610250115,0.271304518,0.3550498486,0.1851463169,-0.3361179233,-0.1769832224,0.2839801311,0.1789635122,-0.2347437888,-0.2329941094,-0.0508185513,-0.1223188341,0.1850056052,0.3649440408,-0.0237663891,-0.2316484004,-0.0074903606,0.1311026812,-0.0525209345,-0.0459918827,0.1973968595,0.3307221234,-0.0605216287,0.0740697607,0.203822583,0.3356961906,0.1484804451,0.4495689869,0.390204668,0.4044875801,0.2652183175,0.1011540592,0.0109557817,-0.3769709468,0.1383010745,0.2418542057,-0.6026106477,0.2768874168,0.2579662204,-0.1434610486,0.0388517864,-0.1774974465,-0.1597563773,0.0448752604,-0.2500063777,-0.1211033538,-0.0018091797,-0.1462201029,0.2350991815,-0.1145027205,-0.0338259041,-0.125802502,0.460375309,-0.0914185718,-0.0093528582,-0.1481794268,-0.3311117291,-0.2172561586,0.2395831794,-0.1705412716,-0.2877802551,0.1675358415,0.2746525407,-0.1172754616,-0.1353558153,0.2621349096,0.2212348878,-0.2120891362,-0.0261261743,-0.0590664037,-0.0073863207,-0.1609937549,0.1352122426,0.035655722,-0.1427993178,0.0698155537,0.1292563379,-0.1309076697,0.018817015,0.3850506842,0.2576455176,-0.3966679573,0.2288762778,-0.0653828382,0.0065334723,-0.3422942162,-0.0566753261,0.071760647,-0.0921381935,0.3632926941,-0.1265348494,0.3559886813,0.0683030859,0.0675616339,-0.1695962995,0.2754689157,-0.0937736034,0.1342546046,-0.1768610328,-0.1477606446,-0.3707118332,-0.0324157476,-0.2415802032,-0.0101434235,0.0788472444,-0.0222217161,0.1696500927,0.3857278824,-0.0734427571,-0.0960481539,-0.2717599869,-0.1786078662,-0.0953544751,-0.240993008,-0.5097748637,-0.2365776598,0.3167963028,-0.3936298192,0.0446066,-0.2630873919,0.1287750006,0.1057690755,0.2196846306,-0.1327527314,0.2146078795,0.3450618982,0.4004009068,0.0616496801,0.0566399209,0.153494969,0.0584256239,-0.0649762079,-0.311940968,0.1824387163,-0.222593233,0.3479798436,-0.340234071,-0.2343781739,0.1301048696,-0.2715990245,0.2384704202,-0.1920745373,-0.2710632682,0.0018826236,-0.2772067785,0.0974600613,-0.2875549495,0.1892949343,-0.001522604,0.1094881371,0.022503484,-0.2070251703,0.6020870805,-0.3814713061,-0.3790595829,-0.2499885559,0.2925056219,-0.1566465795,-0.3191484809,-0.1879810542,-0.168021068,0.3404830992,-0.2098423541,-0.2546118498,-0.1149111316,-0.0779231191,-0.0458497256,-0.2578428686,0.0601824373,-0.0492761098,-0.1487095356,0.3046515286,-0.2471387386]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1940","title":"Side effect when filtering data due to `does_function_return_dict` call in `Dataset.map()`","comments":"Thanks @mariosasko for the PR!","body":"Hi there!\r\n\r\nIn my codebase I have a function to filter rows in a dataset, selecting only a certain number of examples per class. The function passes a extra argument to maintain a counter of the number of dataset rows\/examples already selected per each class, which are the ones I want to keep in the end:\r\n\r\n```python\r\n def fill_train_examples_per_class(example, per_class_limit: int, counter: collections.Counter):\r\n label = int(example['label'])\r\n current_counter = counter.get(label, 0)\r\n if current_counter < per_class_limit:\r\n counter[label] = current_counter + 1\r\n return True\r\n return False\r\n```\r\n\r\nAt some point I invoke it through the `Dataset.filter()` method in the `arrow_dataset.py` module like this:\r\n\r\n```python\r\n...\r\nkwargs = {\"per_class_limit\": train_examples_per_class_limit, \"counter\": Counter()}\r\ndatasets['train'] = datasets['train'].filter(fill_train_examples_per_class, num_proc=1, fn_kwargs=kwargs)\r\n...\r\n```\r\n\r\nThe problem is that, passing a stateful container (the counter,) provokes a side effect in the new filtered dataset obtained. This is due to the fact that at some point in `filter()`, the `map()`'s function `does_function_return_dict` is invoked in line [1290](https:\/\/github.com\/huggingface\/datasets\/blob\/96578adface7e4bc1f3e8bafbac920d72ca1ca60\/src\/datasets\/arrow_dataset.py#L1290). \r\n\r\nWhen this occurs, the state of the counter is initially modified by the effects of the function call on the 1 or 2 rows selected in lines 1288 and 1289 of the same file (which are marked as `test_inputs` & `test_indices` respectively in lines 1288 and 1289. This happens out of the control of the user (which for example can't reset the state of the counter before continuing the execution,) provoking in the end an undesired side effect in the results obtained. \r\n\r\nIn my case, the resulting dataset -despite of the counter results are ok- lacks an instance of the classes 0 and 1 (which happen to be the classes of the first two examples of my dataset.) The rest of the classes I have in my dataset, contain the right number of examples as they were not affected by the effects of `does_function_return_dict` call.\r\n\r\nI've debugged my code extensively and made a workaround myself hardcoding the necessary stuff (basically putting `update_data=True` in line 1290,) and then I obtain the results I expected without the side effect.\r\n\r\nIs there a way to avoid that call to `does_function_return_dict` in map()'s line 1290 ? (e.g. extracting the required information that `does_function_return_dict` returns without making the testing calls to the user function on dataset rows 0 & 1) \r\n\r\nThanks in advance,\r\n\r\nFrancisco Perez-Sorrosal\r\n\r\n","comment_length":5,"text":"Side effect when filtering data due to `does_function_return_dict` call in `Dataset.map()`\nHi there!\r\n\r\nIn my codebase I have a function to filter rows in a dataset, selecting only a certain number of examples per class. The function passes a extra argument to maintain a counter of the number of dataset rows\/examples already selected per each class, which are the ones I want to keep in the end:\r\n\r\n```python\r\n def fill_train_examples_per_class(example, per_class_limit: int, counter: collections.Counter):\r\n label = int(example['label'])\r\n current_counter = counter.get(label, 0)\r\n if current_counter < per_class_limit:\r\n counter[label] = current_counter + 1\r\n return True\r\n return False\r\n```\r\n\r\nAt some point I invoke it through the `Dataset.filter()` method in the `arrow_dataset.py` module like this:\r\n\r\n```python\r\n...\r\nkwargs = {\"per_class_limit\": train_examples_per_class_limit, \"counter\": Counter()}\r\ndatasets['train'] = datasets['train'].filter(fill_train_examples_per_class, num_proc=1, fn_kwargs=kwargs)\r\n...\r\n```\r\n\r\nThe problem is that, passing a stateful container (the counter,) provokes a side effect in the new filtered dataset obtained. This is due to the fact that at some point in `filter()`, the `map()`'s function `does_function_return_dict` is invoked in line [1290](https:\/\/github.com\/huggingface\/datasets\/blob\/96578adface7e4bc1f3e8bafbac920d72ca1ca60\/src\/datasets\/arrow_dataset.py#L1290). \r\n\r\nWhen this occurs, the state of the counter is initially modified by the effects of the function call on the 1 or 2 rows selected in lines 1288 and 1289 of the same file (which are marked as `test_inputs` & `test_indices` respectively in lines 1288 and 1289. This happens out of the control of the user (which for example can't reset the state of the counter before continuing the execution,) provoking in the end an undesired side effect in the results obtained. \r\n\r\nIn my case, the resulting dataset -despite of the counter results are ok- lacks an instance of the classes 0 and 1 (which happen to be the classes of the first two examples of my dataset.) The rest of the classes I have in my dataset, contain the right number of examples as they were not affected by the effects of `does_function_return_dict` call.\r\n\r\nI've debugged my code extensively and made a workaround myself hardcoding the necessary stuff (basically putting `update_data=True` in line 1290,) and then I obtain the results I expected without the side effect.\r\n\r\nIs there a way to avoid that call to `does_function_return_dict` in map()'s line 1290 ? (e.g. extracting the required information that `does_function_return_dict` returns without making the testing calls to the user function on dataset rows 0 & 1) \r\n\r\nThanks in advance,\r\n\r\nFrancisco Perez-Sorrosal\r\n\r\n\nThanks @mariosasko for the PR!","embeddings":[-0.4214114249,0.0289576147,-0.1575038433,0.0184577908,-0.084425889,-0.3201851845,0.2051187158,0.1946579814,0.2660327554,0.1206628382,0.2073837817,0.5199099183,-0.0788978338,0.0967270359,0.0301455557,0.2262496203,0.1668142974,-0.032501027,-0.1896086782,-0.1237709373,-0.2070304006,-0.0553835072,-0.2351946831,0.0294677727,0.1121999249,-0.2066931427,0.2833378315,0.0864987075,0.1280570924,-0.233986333,0.379776001,0.2778170407,-0.3832846582,0.4839723408,-0.0001190907,0.0343095772,0.1299140453,-0.0802711844,-0.2356485426,-0.4382512271,-0.2481366247,0.0078954305,0.1348444819,-0.1925271153,0.2354865968,-0.1303691268,-0.4192903936,-0.3500099182,0.5882077217,0.2834120989,0.1079882756,0.105306901,-0.343233794,0.1123894006,0.4360818267,0.3156699836,0.1410355121,0.3769839108,0.2226104736,-0.4172928631,-0.089796707,0.6554317474,-0.419348985,0.1199867725,-0.1365879774,0.1172127947,0.1939310879,0.0126645528,0.3233865798,0.1725642383,0.1055694222,-0.3248958588,-0.1459566951,-0.420864135,-0.1007794887,0.0005761604,0.0750974044,-0.2886882424,-0.2021785378,-0.1206648573,-0.1208331883,-0.012387082,-0.0200270172,0.1681484133,-0.2169067562,0.1855434477,0.0811168924,0.2631781995,0.0817807019,-0.0514533445,0.2090521157,-0.3612852395,0.2060837448,0.2147341669,0.1196141541,0.0732956082,0.3768185079,-0.4789855778,0.1241235957,-0.13665995,0.1146184802,0.1747404039,-0.2579136491,0.2301839143,0.4253011346,0.085000217,-0.1979258806,0.6640278101,0.1729907244,-0.1068491042,-0.0037359183,0.0037666485,0.2818769813,0.0269299001,0.2999613285,-0.1395051032,0.1103360206,-0.1578751206,-0.2358426154,0.4175386727,-0.3590551019,0.2780917585,0.0907847881,0.0966719389,0.3004828691,0.4036149085,0.1682981998,-0.1376380175,0.003388444,0.5379956365,-0.1066150218,-0.3244780898,-0.2710621357,-0.0770937875,-0.011990536,0.0786589086,0.0169853605,0.1104542688,-0.0109124994,-0.0541719161,0.0757818967,-0.2363334149,0.7913022637,0.4118511677,-0.3480505049,0.1737432331,0.2374269217,-0.2243672758,-0.1260404438,0.2483870536,-0.0546916015,-0.1973662376,0.1852192581,0.1147135645,0.0059861485,0.3745835125,-0.0160887148,0.1991495788,0.4185599089,-0.471998781,0.3922097385,-0.5384453535,-0.3335294724,-0.2172567099,0.1258204728,0.525978744,-0.819714427,-0.1214056835,-0.2428824008,-0.1178346276,0.0761058852,0.1740315706,-0.2996686697,-0.0011380088,-0.2255411148,0.1341030598,0.3071460724,-0.4831515849,-0.705484271,0.0982058793,-0.2767548561,0.499204278,-0.1460386366,0.0213056188,0.3568475246,-0.008165176,0.3267332315,0.1796566099,-0.0669032261,0.1762675196,0.0203271583,0.1293668896,0.3243874311,-0.0954314321,-0.0437130406,-0.077088058,-0.074971199,-0.4875876904,0.281059891,0.0821083933,0.1924644411,-0.3172644973,0.4052571058,0.1898687333,0.0079568848,-0.2883609235,-0.2915593386,0.0308367405,0.0936341137,-0.24900195,-0.2050819397,-0.1570568383,-0.1389681697,0.3914354146,-0.0558801778,0.1572161317,0.1144860759,-0.0418155044,0.00553917,0.0628047958,0.0852909014,0.080448173,-0.4057528675,-0.0257785544,0.0938627049,0.2385653853,0.3390497267,-0.2808124125,-0.2351316363,0.1347765326,0.0226554684,-0.0478546508,-0.2866383493,0.219219178,0.5091303587,-0.0345370322,-0.2714932263,0.0800034404,-0.1027276367,0.4487369657,-0.0084447358,0.3058345914,0.1903862655,-0.2705367208,-0.0997638702,0.4954223633,0.0765313879,0.4916887283,-0.1645125449,0.136452347,0.2814358473,0.2763494253,-0.3366055191,-0.5444409847,-0.2632301152,-0.0805498734,0.1870295554,-0.1061945185,-0.155970186,-0.0380207151,0.0843561813,-0.0368897617,0.1383838505,0.103854686,-0.0823239088,-0.000345651,0.4746937752,0.4552512467,0.5759060979,0.0555185191,-0.1408450902,0.0276647117,-0.0354824476,-0.0613747053,0.1117442772,0.2467335612,0.1628268063,0.3262067139,0.3910287619,-0.1081223041,-0.2600623667,-0.1121191531,-0.1188485101,-0.1379218549,-0.2318265587,0.3826011419,-0.1168807521,-0.2817385197,-0.2118333727,-0.1495019346,0.051916115,-0.1294428259,-0.0069106244,0.4019418955,-0.3745889664,0.0674525797,-0.0789706782,0.3998449147,0.1533491313,-0.4363389611,0.1452000886,-0.3576052785,-0.1495454907,-0.0052298307,0.0020299521,0.0059776395,0.4475282729,0.282324791,-0.210486576,0.0630706325,-0.1092976108,0.1017176583,-0.007605616,-0.0080981012,0.3318776488,0.0187597405,-0.2166742086,0.0701481327,0.0484458171,0.0242532827,-0.1498593092,-0.1106217429,0.0191021357,0.2351116985,-0.2889119983,0.1924333423,-0.1723965108,-0.1195878163,0.1122066975,-0.1309978515,0.2291354388,0.1243820488,0.2222427875,0.0680092126,-0.0073051597,-0.2129320949,-0.4279069602,-0.0134776644,-0.1785435081,-0.144994691,0.0711149648,-0.2463677824,-0.2123691887,0.14317213,0.597771585,-0.5670164824,-0.198012501,0.1690313816,0.280459553,-0.1854302883,0.0043560239,0.3448859155,0.1838807613,-0.0802197531,-0.3217480481,-0.0644872263,0.1970244944,0.3125908077,0.0231229048,0.0797691122,0.1033611,0.0153850578,0.6843212247,0.2735551894,-0.1441142112,-0.0708047748,0.2465561479,0.2138903141,-0.3259014189,-0.0774517357,-0.34776932,-0.233711198,-0.1915657371,-0.1238080114,0.0483663194,-0.2100283802,0.1475331187,-0.1117216647,-0.2634606957,-0.457383126,0.2509599328,-0.2607240379,0.3891144395,0.0045573306,-0.0390345082,-0.3957884908,0.0963751078,-0.0857025608,-0.2023214102,0.3929489851,-0.2158955038,-0.2131988704,-0.0444583893,0.0876910016,0.2771394849,0.3724202514,0.2909666598,-0.006964514,-0.1946190298,-0.0075320178,0.1920891851,0.921217382,-0.1532570869,0.1145694852,0.3606545031,-0.163496539,-0.4281956255,-0.0288701728,-0.3811837137,0.3040297031,0.1288411915,0.136218667,-0.1125409603,0.0810653344,-0.1752343178,0.1800236255,-0.2305070162,-0.0291542169,-0.0458450951,0.0430700332,-0.2418702394,0.4140961468,-0.042652607,-0.3411511481,-0.0270895604,-0.3010748923,0.1418939084,-0.2788738906,-0.1683320403,0.0251771566,0.3258301616,0.0281419717,-0.033845149,0.4214897752,-0.044580318,0.3595966697,0.2097106278,-0.1842389107,-0.099346891,-0.3385727406,0.2808583081,0.0884962678,0.6351366043,-0.0662128776,0.292578578,0.0539201871,0.0340458862,0.0468430892,0.0528456457,0.1087765023,0.3228063583,0.0372582451,-0.5029551387,0.2584246993,0.1158447117,-0.4988925457,0.5093582869,0.070716247,-0.4502280354,0.500374794,0.0967449993,0.8437870145,0.3119137287,0.1658471674,0.2260128111,-0.1717870831,0.0884205475,0.2525969744,-0.0557101145,0.1190145686,0.152756229,-0.0358197615,-0.2901842594,-0.0203851238,-0.1074931622,0.0465300791,0.0796530992,-0.5716618896,0.3465054035,-0.1252949089,-0.2016382217,0.2813823521,-0.3633782268,-0.0759189278,0.1183448583,0.1558741033,0.0072763762,-0.117448777,0.1039035842,0.0316315517,-0.1071537808,-0.1751724333,-0.2920619845,-0.4194130003,0.4033285081,0.0758899152,-0.4115694761,-0.0596608073,-0.0202501994,-0.4361622334,0.2700872421,0.0537092462,-0.0371725373,0.4989475608,0.5238935351,0.1007662788,-0.269499898,0.4783497453,0.1840208918,-0.2807706296,0.0061873645,-0.0833480135,0.0214813873,-0.1123805195,0.0961127952,0.375865072,-0.2928257287,0.0643375739,0.0271966532,-0.0029291124,-0.1024997309,0.0415786356,-0.0710347444,-0.0433687679,0.2548505962,0.0600099824,-0.035591241,-0.0283804033,0.2195603997,0.1335608512,0.2223019749,0.37287274,-0.1903460473,-0.1306002885,-0.1904115379,-0.2018484175,0.1169957742,0.0854392275,0.4244581461,0.0115424357,-0.2601805031,-0.3701823056,0.1487354338,-0.0050651985,0.2873671055,-0.5665975213,-0.3418369591,-0.2227964103,0.1422403306,0.121300377,0.1461369097,-0.2059533745,0.2034112662,-0.3286547363,0.1982479841,-0.2763057649,-0.088812381,0.0760137886,0.3025518358,0.072930567,0.2834870219,-0.2598807812,-0.1926088929,0.022234913,0.4459893405,-0.245225504,-0.1933341473,-0.0131521011,0.1412336379,-0.0306064282,-0.3593326807,0.0489733629,-0.3224500418,0.1845193654,-0.2329692096,0.4509948194,0.1140588447,-0.0886011794,-0.2661243379,-0.0722008049,0.152037099,0.3301406503,0.4199129045,-0.282502681,-0.266082108,-0.0245819353,0.1808160841,0.065408662,-0.0692419931,-0.06781739,0.188092038,0.1825224161,0.1244096756,0.1734904349,-0.0926332772,-0.2610250115,0.271304518,0.3550498486,0.1851463169,-0.3361179233,-0.1769832224,0.2839801311,0.1789635122,-0.2347437888,-0.2329941094,-0.0508185513,-0.1223188341,0.1850056052,0.3649440408,-0.0237663891,-0.2316484004,-0.0074903606,0.1311026812,-0.0525209345,-0.0459918827,0.1973968595,0.3307221234,-0.0605216287,0.0740697607,0.203822583,0.3356961906,0.1484804451,0.4495689869,0.390204668,0.4044875801,0.2652183175,0.1011540592,0.0109557817,-0.3769709468,0.1383010745,0.2418542057,-0.6026106477,0.2768874168,0.2579662204,-0.1434610486,0.0388517864,-0.1774974465,-0.1597563773,0.0448752604,-0.2500063777,-0.1211033538,-0.0018091797,-0.1462201029,0.2350991815,-0.1145027205,-0.0338259041,-0.125802502,0.460375309,-0.0914185718,-0.0093528582,-0.1481794268,-0.3311117291,-0.2172561586,0.2395831794,-0.1705412716,-0.2877802551,0.1675358415,0.2746525407,-0.1172754616,-0.1353558153,0.2621349096,0.2212348878,-0.2120891362,-0.0261261743,-0.0590664037,-0.0073863207,-0.1609937549,0.1352122426,0.035655722,-0.1427993178,0.0698155537,0.1292563379,-0.1309076697,0.018817015,0.3850506842,0.2576455176,-0.3966679573,0.2288762778,-0.0653828382,0.0065334723,-0.3422942162,-0.0566753261,0.071760647,-0.0921381935,0.3632926941,-0.1265348494,0.3559886813,0.0683030859,0.0675616339,-0.1695962995,0.2754689157,-0.0937736034,0.1342546046,-0.1768610328,-0.1477606446,-0.3707118332,-0.0324157476,-0.2415802032,-0.0101434235,0.0788472444,-0.0222217161,0.1696500927,0.3857278824,-0.0734427571,-0.0960481539,-0.2717599869,-0.1786078662,-0.0953544751,-0.240993008,-0.5097748637,-0.2365776598,0.3167963028,-0.3936298192,0.0446066,-0.2630873919,0.1287750006,0.1057690755,0.2196846306,-0.1327527314,0.2146078795,0.3450618982,0.4004009068,0.0616496801,0.0566399209,0.153494969,0.0584256239,-0.0649762079,-0.311940968,0.1824387163,-0.222593233,0.3479798436,-0.340234071,-0.2343781739,0.1301048696,-0.2715990245,0.2384704202,-0.1920745373,-0.2710632682,0.0018826236,-0.2772067785,0.0974600613,-0.2875549495,0.1892949343,-0.001522604,0.1094881371,0.022503484,-0.2070251703,0.6020870805,-0.3814713061,-0.3790595829,-0.2499885559,0.2925056219,-0.1566465795,-0.3191484809,-0.1879810542,-0.168021068,0.3404830992,-0.2098423541,-0.2546118498,-0.1149111316,-0.0779231191,-0.0458497256,-0.2578428686,0.0601824373,-0.0492761098,-0.1487095356,0.3046515286,-0.2471387386]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1939","title":"[firewalled env] OFFLINE mode","comments":"Thanks for reporting and for all the details and suggestions.\r\n\r\nI'm totally in favor of having a HF_DATASETS_OFFLINE env variable to disable manually all the connection checks, remove retries etc.\r\n\r\nMoreover you may know that the use case that you are mentioning is already supported from `datasets` 1.3.0, i.e. you already can:\r\n- first load datasets and metrics from an instance with internet connection\r\n- then be able to reload datasets and metrics from another instance without connection (as long as the filesystem is shared)\r\n\r\nThis is already implemented, but currently it only works if the requests return a `ConnectionError` (or any error actually). Not sure why it would hang instead of returning an error.\r\n\r\nMaybe this is just a issue with the timeout value being not set or too high ?\r\nIs there a way I can have access to one of the instances on which there's this issue (we can discuss this offline) ?\r\n","body":"This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.","comment_length":156,"text":"[firewalled env] OFFLINE mode\nThis issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.\nThanks for reporting and for all the details and suggestions.\r\n\r\nI'm totally in favor of having a HF_DATASETS_OFFLINE env variable to disable manually all the connection checks, remove retries etc.\r\n\r\nMoreover you may know that the use case that you are mentioning is already supported from `datasets` 1.3.0, i.e. you already can:\r\n- first load datasets and metrics from an instance with internet connection\r\n- then be able to reload datasets and metrics from another instance without connection (as long as the filesystem is shared)\r\n\r\nThis is already implemented, but currently it only works if the requests return a `ConnectionError` (or any error actually). Not sure why it would hang instead of returning an error.\r\n\r\nMaybe this is just a issue with the timeout value being not set or too high ?\r\nIs there a way I can have access to one of the instances on which there's this issue (we can discuss this offline) ?\r\n","embeddings":[-0.4748800695,0.1685334891,-0.0625536442,0.1050562486,0.0443113707,-0.2284921706,0.4647973776,0.0716796964,0.2545911074,0.0485309139,0.0184395313,0.0636148751,0.0845065862,0.1993038058,0.0180914346,0.1079676971,0.0199452993,-0.1096389145,-0.2528404295,0.0637163967,-0.2635913491,-0.0536728576,-0.2544517219,-0.2575809062,0.078367807,-0.5080999136,-0.2168778479,-0.0105694095,0.0096480669,-0.2275560498,0.3089389205,0.2612626255,0.2518863976,0.1530115753,-0.0001084152,-0.0537566915,-0.1393844485,-0.0361674428,-0.272887677,0.0223880373,-0.3179339468,-0.3802416921,0.1650776267,-0.4613099992,-0.171355173,-0.174725309,0.1464560628,-0.6594020128,0.5154665709,-0.0763156116,0.0945165008,0.0003517794,-0.3278796375,-0.0994149074,-0.3226190805,-0.1980506182,-0.1814403385,0.1710348725,0.0883782059,-0.2174991071,-0.0455865599,0.3952884078,-0.3122881055,0.1565717608,0.3260605037,0.0182255022,-0.338113457,-0.1188222617,-0.1281385869,0.3376224041,0.1812187731,-0.2216326147,-0.4262379706,-0.4555667341,-0.0790792406,-0.370331347,0.3025760651,0.0889306292,-0.1169174388,0.2809565067,-0.3825724125,-0.2138668299,-0.1247217432,-0.0012391666,-0.0308274534,0.3738206923,0.0736688673,0.1736761779,-0.0692528114,0.1060303599,0.0134036681,-0.3340767622,0.2569811642,0.0270416345,-0.5665541291,0.0265186541,0.1224480569,-0.0067842589,0.0168159846,0.4544952512,-0.1779251099,-0.0429254211,0.4983155131,0.195036903,0.0398950391,0.2975893617,0.1519620121,-0.1790611893,0.3776472807,0.1602634937,-0.0568716712,-0.1345084161,-0.1136473343,-0.0706924796,0.1811163574,0.0640624613,0.1197685003,-0.1377031356,-0.1166463196,0.046965342,-0.0323215127,0.2401375026,0.3313703537,0.3093738258,-0.0555984639,-0.2358305752,0.2333134562,0.1667176485,-0.1382888556,-0.06039121,0.072590515,-0.0984998196,-0.1313179582,0.082668893,0.3286071122,0.2563716769,0.2747864425,-0.1755116731,0.1528570056,0.0238279738,0.5519109368,-0.2458668649,0.1138633564,0.3021980822,0.1545313895,-0.2306014448,0.0980398506,0.2813174427,-0.0375363044,0.0506099239,0.0799426436,-0.3858196735,0.4044837654,0.1439911574,-0.1975332648,-0.0476032272,0.1782198101,0.0742536411,-0.0721478388,-0.4519341588,0.1831779182,0.0667779595,-0.1193479002,-0.0543197319,0.4079632163,0.4936104715,-0.4202961922,0.0228122212,-0.221729517,0.019100355,-0.1197147667,0.1713059396,-0.2794617414,0.1356073618,-0.1801660359,-0.2079320103,0.2443795949,-0.3084554672,-0.3225480914,0.4538064003,-0.1916149855,0.2314610779,0.3152425885,0.1994496435,0.2296581119,-0.0338119008,0.1222665012,0.3941685557,-0.2495333552,-0.0643453002,-0.1496422589,-0.1908909678,0.2760188878,0.1059188098,0.2374101728,-0.1076447815,-0.086140126,-0.2940383255,-0.0350805186,-0.083207719,0.2003037333,-0.1944311261,0.0856611282,0.0736428425,0.0493785702,0.1411918253,-0.4604979455,0.3037669957,-0.2521933019,0.0747477561,-0.3391274512,-0.3241582215,0.2313053608,-0.0713331848,-0.1031431407,-0.3441320658,0.0740442052,0.0512403548,-0.1017183214,-0.0374327078,-0.0173988789,0.0242225807,-0.0870222002,0.1135079041,0.1581827104,-0.0672886223,0.0744352862,0.0992200971,-0.2595793903,-0.1118372455,-0.0186913777,-0.1419194341,-0.0033473987,0.507107079,0.0670183226,0.434106648,0.1801059842,0.43449682,0.100204356,0.3320184648,0.2241401672,0.0292811133,0.0667587593,-0.0281432383,-0.5865287185,0.8847305179,-0.0772907659,0.4949632585,-0.0253681988,-0.0896625817,0.2581321895,-0.070058696,-0.5748965144,-0.0896323398,-0.0292923599,0.2472509444,0.1246782467,-0.0695684403,-0.3645498455,0.1309467256,0.0615998544,0.2113251835,0.4444034696,-0.2755915225,0.2542671561,-0.0863385573,0.4632284939,0.2097619325,0.2965559363,0.2528910935,0.0414115787,0.1776760668,-0.0744988695,-0.2386737466,0.3642019629,0.1546070129,0.3154804409,0.1600142121,0.0825260282,-0.1450218856,-0.1656669974,-0.2701164484,0.1441674232,-0.2285314947,-0.052432321,0.3188200295,-0.3138054311,-0.5460381508,-0.0330595821,0.2567465603,-0.1731722355,-0.1265536398,0.232597664,0.1215340123,-0.3687728345,0.3979503214,-0.2202257961,0.4702404141,-0.0649726614,-0.4042862654,0.2372114658,-0.3736272752,0.0975892395,0.0727022961,0.2432580888,-0.090477258,0.6828110218,0.2652270794,-0.0664931238,-0.172499001,-0.2048208117,0.1804469377,0.0453945622,0.5188974738,0.1410698295,0.097885415,-0.1078779474,0.0392978527,0.266864717,-0.131492421,0.2002531737,-0.1014679745,0.0052114641,0.1161197722,-0.2464787811,-0.5637351274,-0.2847306728,-0.2684567571,0.1167265549,-0.2222549319,0.2011254579,-0.0231130943,0.1001847163,-0.050306119,0.1435732543,0.0995229259,-0.0322472528,-0.568808198,0.1139889956,-0.1637168378,0.0997769907,0.0580960028,-0.0110152299,0.3906186521,0.5869692564,-0.4705568254,-0.2902962863,0.3895433247,0.286469698,0.0516619869,-0.3373523057,0.6150392294,0.0365746021,-0.0880156383,-0.0884277374,0.3548750579,0.4860231876,-0.1864490658,-0.0539083518,0.1791349649,0.7724372149,0.202256009,0.2877366543,-0.1666050255,0.4881629646,0.2260150015,0.1102786735,0.304939419,-0.1625722647,0.0179588497,0.2034069151,-0.0815214068,-0.5336046219,0.0716209486,0.0470552854,-0.0527305119,-0.3611359894,-0.2582453787,-0.1032524407,-0.3916658759,0.2961251736,-0.1127902269,0.4995014668,0.117896229,0.372337997,-0.2182418406,0.1661288887,-0.1418527663,0.1501836479,0.5369586945,-0.0128860902,-0.1799632609,0.0348873436,-0.22274369,0.2784568965,0.0099335965,0.3417262137,-0.036391452,0.0174978469,0.0015469281,-0.0763285756,0.4463128746,-0.1146823615,0.0632685199,0.0081591588,-0.2919878364,-0.2112849802,-0.1305243969,-0.0428579152,-0.0211985148,0.0649969876,0.1952529997,-0.2032026201,-0.0390476733,-0.3187787533,0.1351860166,-0.3197744787,-0.2367112786,0.0234858245,0.1267216206,-0.6222805977,0.420473069,-0.1571975201,-0.1905483156,-0.2480874062,-0.0747949705,-0.1783072352,0.4812121689,-0.145950079,0.0565128624,0.1470396221,-0.5123636127,0.1858813465,0.1123727337,0.1500660181,0.3838641942,0.0461331084,0.1363827586,0.284540683,0.0637651309,0.066484876,0.2051711679,0.4261293709,-0.1089276075,-0.0030097992,0.1616743952,0.3030042052,-0.3588068783,-0.3979021609,0.1224461347,0.1558630913,0.0335068256,-0.4153938591,0.0671153069,0.04605335,-0.2596212626,0.0445624702,0.1521447897,-0.3728854656,0.2617143989,0.2630714774,0.9972520471,0.2139621526,0.2179925889,0.4170042574,0.2532133162,0.4726442695,-0.6332621574,0.2983065844,0.1272928864,-0.1670437455,-0.1499634385,-0.02558963,0.2321767211,0.099870272,-0.2685555816,0.5433549285,-0.0701307207,0.1014339775,-0.2433125675,-0.0296450164,-0.2376535684,-0.4036455154,-0.2663605511,0.1260658652,0.2334099859,-0.0104813557,-0.3751176894,-0.068979308,0.1925515532,0.1631508023,0.14545542,0.0721764639,0.0428949408,0.2766603231,-0.419208467,0.0559381731,-0.2763850689,0.1090795174,0.0180981327,0.2108995914,-0.2886296809,0.1259953529,-0.0419275537,0.1140722334,-0.1539172381,-0.2668178678,0.372919023,-0.1662880629,-0.2487163991,-0.2205847353,-0.3582707942,-0.0771528035,0.2514949441,-0.3500300646,0.0939490199,-0.0691653416,0.0619472638,0.1937758923,-0.1183083802,-0.215838626,0.0828801244,0.1098431647,-0.0477760732,0.2642475665,0.0833750889,-0.0296078119,-0.1844729483,0.5224729776,-0.0124881873,0.2983217239,0.329159528,0.17099756,-0.2422726303,-0.3916774988,-0.1179054156,-0.0944124162,0.1713488996,0.0689709634,0.0006624407,-0.0254697464,0.0580743439,-0.2008399367,0.1766274869,-0.0651128218,-0.1013034061,-0.4263277054,-0.0957949758,0.3383998871,0.2018853426,0.0942364931,0.1214297861,0.1285647601,0.0334427096,0.011660737,-0.3435808122,-0.1933899075,-0.2666880488,0.2216063738,-0.0113147693,-0.0477256738,0.1774714738,-0.0150593249,0.0563594587,-0.0292195305,-0.2538238168,-0.1619534642,-0.1136552319,0.2258027345,-0.1483045071,-0.3157637417,0.2160577774,-0.423892051,-0.3664358854,-0.3153072298,0.0918554738,-0.0170711968,-0.1677893996,-0.4335837364,-0.1071414053,0.3017937243,-0.2243884355,0.0949885547,-0.0977727994,-0.2725069225,-0.3345758319,0.0538960434,-0.0734193325,-0.1096151099,-0.0395689383,0.1027320102,0.2106037289,0.205578953,0.4639981687,0.0892977566,-0.1204226613,0.1604589671,0.6960189939,0.1433717161,0.1893894821,0.1112304628,0.0307057649,0.2147427052,-0.4058090746,0.1708363444,-0.2121228129,-0.1652800739,0.2011716664,0.2182677984,0.4211681187,0.1228163764,0.2115367651,-0.0768140182,0.3960933387,-0.3639710844,0.2452284545,-0.0232135504,-0.0485098064,0.1692207754,-0.0433972329,0.1081223264,0.3069615662,0.0682089925,-0.269303441,0.1668248624,0.4557292759,0.2022744119,-0.0908759907,-0.4782039225,-0.1009483114,0.0956822485,-0.0603445247,0.0244362708,-0.464867115,-0.0542507693,-0.3395825922,-0.1913392246,-0.2145864964,0.6027536392,-0.3315451741,-0.1462776661,-0.4204238951,-0.4396710098,0.0110949399,0.0038176409,0.1870268434,-0.138487637,-0.1736916751,0.0796872526,0.2774659395,-0.2712741792,0.0969131291,-0.0292456578,0.1431989819,-0.323631078,0.1587519348,0.5755221844,0.0011660997,0.0627885014,0.2858873308,0.2266620696,0.0022768998,-0.1239224076,0.1276148558,0.1615405679,0.0533404723,-0.3756064475,-0.0559492894,-0.016708402,0.2881965935,0.3220778406,0.1017410532,-0.0451429375,0.503982842,-0.0574037246,-0.3350550532,-0.217862308,-0.0598231368,0.4229668081,0.1130177602,-0.0961265415,-0.0181783196,-0.1123341396,-0.0917096362,0.2774384022,-0.1616678834,0.0388007686,0.0190163963,0.0622579455,-0.1917547733,0.5251255035,-0.0828251243,0.1001403928,-0.3585446477,-0.0513545163,-0.6031898856,-0.2333036214,-0.0481282994,0.050825011,-0.1225795448,0.1338446736,0.2245391905,0.1548327804,-0.2501180172,0.024877632,0.1090888679,0.0784571469,-0.3683554232,0.0319491997,0.0217686146,-0.0349099971,0.1462561339,-0.2422401905,0.2108056098,-0.1415773928,0.0672279671,0.307381779,0.0241460316,0.0134564163,0.1558464766,0.7136694193,0.0963875428,0.3541609049,0.0330818854,-0.0392261669,-0.1355392486,0.0061807684,-0.1747146398,0.0417080708,0.042096138,0.2525860369,-0.1677740067,-0.1356784999,-0.1732335538,0.0797532499,-0.0598759577,-0.006825991,-0.1402485371,-0.1277015358,-0.1939341277,0.1754536778,-0.0835498869,0.2089730799,0.1280008405,0.227099821,-0.2550806999,-0.1245743409,0.4326334,-0.0996425748,0.1882971972,-0.3496024013,0.2039181441,0.229376331,0.170839116,-0.0759380162,-0.0538152941,0.1604664326,-0.0303299837,0.107823357,-0.0209005252,0.0539114326,-0.1801201999,-0.1716643721,0.2066053599,-0.3249644637,-0.3378339112,-0.2800188065,-0.2434223294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1939","title":"[firewalled env] OFFLINE mode","comments":"I'm on master, so using all the available bells and whistles already.\r\n\r\nIf you look at the common issues - it for example tries to look up files if they appear in `_PACKAGED_DATASETS_MODULES` which it shouldn't do.\r\n\r\n--------------\r\n\r\nYes, there is a nuance to it. As I mentioned it's firewalled - that is it has a network but making any calls outside - it just hangs in:\r\n\r\n```\r\nsin_addr=inet_addr(\"xx.xx.xx.xx\")}, [28->16]) = 0\r\nclose(5) = 0\r\nsocket(AF_INET, SOCK_STREAM|SOCK_CLOEXEC, IPPROTO_TCP) = 5\r\nconnect(5, {sa_family=AF_INET, sin_port=htons(3128), sin_addr=inet_addr(\"yy.yy.yy.yy\")}, 16^C) = ? ERESTARTSYS (To be restarted if SA_RESTART is set)\r\n```\r\nuntil it times out.\r\n\r\nThat's why we need to be able to tell the software that there is no network to rely on even if there is one (good for testing too).\r\n\r\nSo what I'm thinking is that this is a simple matter of pre-ambling any network call wrappers with:\r\n\r\n```\r\nif HF_DATASETS_OFFLINE:\r\n assert \"Attempting to make a network call under Offline mode\"\r\n```\r\n\r\nand then fixing up if there is anything else to fix to make it work.\r\n\r\n--------------\r\n\r\nOtherwise I think the only other problem I encountered is that we need to find a way to pre-cache metrics, for some reason it's not caching it and wanting to fetch it from online.\r\n\r\nWhich is extra strange since it already has those files in the `datasets` repo itself that is on the filesystem.\r\n\r\nThe workaround I had to do is to copy `rouge\/rouge.py` (with the parent folder) from the datasets repo to the current dir - and then it proceeded.","body":"This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.","comment_length":257,"text":"[firewalled env] OFFLINE mode\nThis issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.\nI'm on master, so using all the available bells and whistles already.\r\n\r\nIf you look at the common issues - it for example tries to look up files if they appear in `_PACKAGED_DATASETS_MODULES` which it shouldn't do.\r\n\r\n--------------\r\n\r\nYes, there is a nuance to it. As I mentioned it's firewalled - that is it has a network but making any calls outside - it just hangs in:\r\n\r\n```\r\nsin_addr=inet_addr(\"xx.xx.xx.xx\")}, [28->16]) = 0\r\nclose(5) = 0\r\nsocket(AF_INET, SOCK_STREAM|SOCK_CLOEXEC, IPPROTO_TCP) = 5\r\nconnect(5, {sa_family=AF_INET, sin_port=htons(3128), sin_addr=inet_addr(\"yy.yy.yy.yy\")}, 16^C) = ? ERESTARTSYS (To be restarted if SA_RESTART is set)\r\n```\r\nuntil it times out.\r\n\r\nThat's why we need to be able to tell the software that there is no network to rely on even if there is one (good for testing too).\r\n\r\nSo what I'm thinking is that this is a simple matter of pre-ambling any network call wrappers with:\r\n\r\n```\r\nif HF_DATASETS_OFFLINE:\r\n assert \"Attempting to make a network call under Offline mode\"\r\n```\r\n\r\nand then fixing up if there is anything else to fix to make it work.\r\n\r\n--------------\r\n\r\nOtherwise I think the only other problem I encountered is that we need to find a way to pre-cache metrics, for some reason it's not caching it and wanting to fetch it from online.\r\n\r\nWhich is extra strange since it already has those files in the `datasets` repo itself that is on the filesystem.\r\n\r\nThe workaround I had to do is to copy `rouge\/rouge.py` (with the parent folder) from the datasets repo to the current dir - and then it proceeded.","embeddings":[-0.4748800695,0.1685334891,-0.0625536442,0.1050562486,0.0443113707,-0.2284921706,0.4647973776,0.0716796964,0.2545911074,0.0485309139,0.0184395313,0.0636148751,0.0845065862,0.1993038058,0.0180914346,0.1079676971,0.0199452993,-0.1096389145,-0.2528404295,0.0637163967,-0.2635913491,-0.0536728576,-0.2544517219,-0.2575809062,0.078367807,-0.5080999136,-0.2168778479,-0.0105694095,0.0096480669,-0.2275560498,0.3089389205,0.2612626255,0.2518863976,0.1530115753,-0.0001084152,-0.0537566915,-0.1393844485,-0.0361674428,-0.272887677,0.0223880373,-0.3179339468,-0.3802416921,0.1650776267,-0.4613099992,-0.171355173,-0.174725309,0.1464560628,-0.6594020128,0.5154665709,-0.0763156116,0.0945165008,0.0003517794,-0.3278796375,-0.0994149074,-0.3226190805,-0.1980506182,-0.1814403385,0.1710348725,0.0883782059,-0.2174991071,-0.0455865599,0.3952884078,-0.3122881055,0.1565717608,0.3260605037,0.0182255022,-0.338113457,-0.1188222617,-0.1281385869,0.3376224041,0.1812187731,-0.2216326147,-0.4262379706,-0.4555667341,-0.0790792406,-0.370331347,0.3025760651,0.0889306292,-0.1169174388,0.2809565067,-0.3825724125,-0.2138668299,-0.1247217432,-0.0012391666,-0.0308274534,0.3738206923,0.0736688673,0.1736761779,-0.0692528114,0.1060303599,0.0134036681,-0.3340767622,0.2569811642,0.0270416345,-0.5665541291,0.0265186541,0.1224480569,-0.0067842589,0.0168159846,0.4544952512,-0.1779251099,-0.0429254211,0.4983155131,0.195036903,0.0398950391,0.2975893617,0.1519620121,-0.1790611893,0.3776472807,0.1602634937,-0.0568716712,-0.1345084161,-0.1136473343,-0.0706924796,0.1811163574,0.0640624613,0.1197685003,-0.1377031356,-0.1166463196,0.046965342,-0.0323215127,0.2401375026,0.3313703537,0.3093738258,-0.0555984639,-0.2358305752,0.2333134562,0.1667176485,-0.1382888556,-0.06039121,0.072590515,-0.0984998196,-0.1313179582,0.082668893,0.3286071122,0.2563716769,0.2747864425,-0.1755116731,0.1528570056,0.0238279738,0.5519109368,-0.2458668649,0.1138633564,0.3021980822,0.1545313895,-0.2306014448,0.0980398506,0.2813174427,-0.0375363044,0.0506099239,0.0799426436,-0.3858196735,0.4044837654,0.1439911574,-0.1975332648,-0.0476032272,0.1782198101,0.0742536411,-0.0721478388,-0.4519341588,0.1831779182,0.0667779595,-0.1193479002,-0.0543197319,0.4079632163,0.4936104715,-0.4202961922,0.0228122212,-0.221729517,0.019100355,-0.1197147667,0.1713059396,-0.2794617414,0.1356073618,-0.1801660359,-0.2079320103,0.2443795949,-0.3084554672,-0.3225480914,0.4538064003,-0.1916149855,0.2314610779,0.3152425885,0.1994496435,0.2296581119,-0.0338119008,0.1222665012,0.3941685557,-0.2495333552,-0.0643453002,-0.1496422589,-0.1908909678,0.2760188878,0.1059188098,0.2374101728,-0.1076447815,-0.086140126,-0.2940383255,-0.0350805186,-0.083207719,0.2003037333,-0.1944311261,0.0856611282,0.0736428425,0.0493785702,0.1411918253,-0.4604979455,0.3037669957,-0.2521933019,0.0747477561,-0.3391274512,-0.3241582215,0.2313053608,-0.0713331848,-0.1031431407,-0.3441320658,0.0740442052,0.0512403548,-0.1017183214,-0.0374327078,-0.0173988789,0.0242225807,-0.0870222002,0.1135079041,0.1581827104,-0.0672886223,0.0744352862,0.0992200971,-0.2595793903,-0.1118372455,-0.0186913777,-0.1419194341,-0.0033473987,0.507107079,0.0670183226,0.434106648,0.1801059842,0.43449682,0.100204356,0.3320184648,0.2241401672,0.0292811133,0.0667587593,-0.0281432383,-0.5865287185,0.8847305179,-0.0772907659,0.4949632585,-0.0253681988,-0.0896625817,0.2581321895,-0.070058696,-0.5748965144,-0.0896323398,-0.0292923599,0.2472509444,0.1246782467,-0.0695684403,-0.3645498455,0.1309467256,0.0615998544,0.2113251835,0.4444034696,-0.2755915225,0.2542671561,-0.0863385573,0.4632284939,0.2097619325,0.2965559363,0.2528910935,0.0414115787,0.1776760668,-0.0744988695,-0.2386737466,0.3642019629,0.1546070129,0.3154804409,0.1600142121,0.0825260282,-0.1450218856,-0.1656669974,-0.2701164484,0.1441674232,-0.2285314947,-0.052432321,0.3188200295,-0.3138054311,-0.5460381508,-0.0330595821,0.2567465603,-0.1731722355,-0.1265536398,0.232597664,0.1215340123,-0.3687728345,0.3979503214,-0.2202257961,0.4702404141,-0.0649726614,-0.4042862654,0.2372114658,-0.3736272752,0.0975892395,0.0727022961,0.2432580888,-0.090477258,0.6828110218,0.2652270794,-0.0664931238,-0.172499001,-0.2048208117,0.1804469377,0.0453945622,0.5188974738,0.1410698295,0.097885415,-0.1078779474,0.0392978527,0.266864717,-0.131492421,0.2002531737,-0.1014679745,0.0052114641,0.1161197722,-0.2464787811,-0.5637351274,-0.2847306728,-0.2684567571,0.1167265549,-0.2222549319,0.2011254579,-0.0231130943,0.1001847163,-0.050306119,0.1435732543,0.0995229259,-0.0322472528,-0.568808198,0.1139889956,-0.1637168378,0.0997769907,0.0580960028,-0.0110152299,0.3906186521,0.5869692564,-0.4705568254,-0.2902962863,0.3895433247,0.286469698,0.0516619869,-0.3373523057,0.6150392294,0.0365746021,-0.0880156383,-0.0884277374,0.3548750579,0.4860231876,-0.1864490658,-0.0539083518,0.1791349649,0.7724372149,0.202256009,0.2877366543,-0.1666050255,0.4881629646,0.2260150015,0.1102786735,0.304939419,-0.1625722647,0.0179588497,0.2034069151,-0.0815214068,-0.5336046219,0.0716209486,0.0470552854,-0.0527305119,-0.3611359894,-0.2582453787,-0.1032524407,-0.3916658759,0.2961251736,-0.1127902269,0.4995014668,0.117896229,0.372337997,-0.2182418406,0.1661288887,-0.1418527663,0.1501836479,0.5369586945,-0.0128860902,-0.1799632609,0.0348873436,-0.22274369,0.2784568965,0.0099335965,0.3417262137,-0.036391452,0.0174978469,0.0015469281,-0.0763285756,0.4463128746,-0.1146823615,0.0632685199,0.0081591588,-0.2919878364,-0.2112849802,-0.1305243969,-0.0428579152,-0.0211985148,0.0649969876,0.1952529997,-0.2032026201,-0.0390476733,-0.3187787533,0.1351860166,-0.3197744787,-0.2367112786,0.0234858245,0.1267216206,-0.6222805977,0.420473069,-0.1571975201,-0.1905483156,-0.2480874062,-0.0747949705,-0.1783072352,0.4812121689,-0.145950079,0.0565128624,0.1470396221,-0.5123636127,0.1858813465,0.1123727337,0.1500660181,0.3838641942,0.0461331084,0.1363827586,0.284540683,0.0637651309,0.066484876,0.2051711679,0.4261293709,-0.1089276075,-0.0030097992,0.1616743952,0.3030042052,-0.3588068783,-0.3979021609,0.1224461347,0.1558630913,0.0335068256,-0.4153938591,0.0671153069,0.04605335,-0.2596212626,0.0445624702,0.1521447897,-0.3728854656,0.2617143989,0.2630714774,0.9972520471,0.2139621526,0.2179925889,0.4170042574,0.2532133162,0.4726442695,-0.6332621574,0.2983065844,0.1272928864,-0.1670437455,-0.1499634385,-0.02558963,0.2321767211,0.099870272,-0.2685555816,0.5433549285,-0.0701307207,0.1014339775,-0.2433125675,-0.0296450164,-0.2376535684,-0.4036455154,-0.2663605511,0.1260658652,0.2334099859,-0.0104813557,-0.3751176894,-0.068979308,0.1925515532,0.1631508023,0.14545542,0.0721764639,0.0428949408,0.2766603231,-0.419208467,0.0559381731,-0.2763850689,0.1090795174,0.0180981327,0.2108995914,-0.2886296809,0.1259953529,-0.0419275537,0.1140722334,-0.1539172381,-0.2668178678,0.372919023,-0.1662880629,-0.2487163991,-0.2205847353,-0.3582707942,-0.0771528035,0.2514949441,-0.3500300646,0.0939490199,-0.0691653416,0.0619472638,0.1937758923,-0.1183083802,-0.215838626,0.0828801244,0.1098431647,-0.0477760732,0.2642475665,0.0833750889,-0.0296078119,-0.1844729483,0.5224729776,-0.0124881873,0.2983217239,0.329159528,0.17099756,-0.2422726303,-0.3916774988,-0.1179054156,-0.0944124162,0.1713488996,0.0689709634,0.0006624407,-0.0254697464,0.0580743439,-0.2008399367,0.1766274869,-0.0651128218,-0.1013034061,-0.4263277054,-0.0957949758,0.3383998871,0.2018853426,0.0942364931,0.1214297861,0.1285647601,0.0334427096,0.011660737,-0.3435808122,-0.1933899075,-0.2666880488,0.2216063738,-0.0113147693,-0.0477256738,0.1774714738,-0.0150593249,0.0563594587,-0.0292195305,-0.2538238168,-0.1619534642,-0.1136552319,0.2258027345,-0.1483045071,-0.3157637417,0.2160577774,-0.423892051,-0.3664358854,-0.3153072298,0.0918554738,-0.0170711968,-0.1677893996,-0.4335837364,-0.1071414053,0.3017937243,-0.2243884355,0.0949885547,-0.0977727994,-0.2725069225,-0.3345758319,0.0538960434,-0.0734193325,-0.1096151099,-0.0395689383,0.1027320102,0.2106037289,0.205578953,0.4639981687,0.0892977566,-0.1204226613,0.1604589671,0.6960189939,0.1433717161,0.1893894821,0.1112304628,0.0307057649,0.2147427052,-0.4058090746,0.1708363444,-0.2121228129,-0.1652800739,0.2011716664,0.2182677984,0.4211681187,0.1228163764,0.2115367651,-0.0768140182,0.3960933387,-0.3639710844,0.2452284545,-0.0232135504,-0.0485098064,0.1692207754,-0.0433972329,0.1081223264,0.3069615662,0.0682089925,-0.269303441,0.1668248624,0.4557292759,0.2022744119,-0.0908759907,-0.4782039225,-0.1009483114,0.0956822485,-0.0603445247,0.0244362708,-0.464867115,-0.0542507693,-0.3395825922,-0.1913392246,-0.2145864964,0.6027536392,-0.3315451741,-0.1462776661,-0.4204238951,-0.4396710098,0.0110949399,0.0038176409,0.1870268434,-0.138487637,-0.1736916751,0.0796872526,0.2774659395,-0.2712741792,0.0969131291,-0.0292456578,0.1431989819,-0.323631078,0.1587519348,0.5755221844,0.0011660997,0.0627885014,0.2858873308,0.2266620696,0.0022768998,-0.1239224076,0.1276148558,0.1615405679,0.0533404723,-0.3756064475,-0.0559492894,-0.016708402,0.2881965935,0.3220778406,0.1017410532,-0.0451429375,0.503982842,-0.0574037246,-0.3350550532,-0.217862308,-0.0598231368,0.4229668081,0.1130177602,-0.0961265415,-0.0181783196,-0.1123341396,-0.0917096362,0.2774384022,-0.1616678834,0.0388007686,0.0190163963,0.0622579455,-0.1917547733,0.5251255035,-0.0828251243,0.1001403928,-0.3585446477,-0.0513545163,-0.6031898856,-0.2333036214,-0.0481282994,0.050825011,-0.1225795448,0.1338446736,0.2245391905,0.1548327804,-0.2501180172,0.024877632,0.1090888679,0.0784571469,-0.3683554232,0.0319491997,0.0217686146,-0.0349099971,0.1462561339,-0.2422401905,0.2108056098,-0.1415773928,0.0672279671,0.307381779,0.0241460316,0.0134564163,0.1558464766,0.7136694193,0.0963875428,0.3541609049,0.0330818854,-0.0392261669,-0.1355392486,0.0061807684,-0.1747146398,0.0417080708,0.042096138,0.2525860369,-0.1677740067,-0.1356784999,-0.1732335538,0.0797532499,-0.0598759577,-0.006825991,-0.1402485371,-0.1277015358,-0.1939341277,0.1754536778,-0.0835498869,0.2089730799,0.1280008405,0.227099821,-0.2550806999,-0.1245743409,0.4326334,-0.0996425748,0.1882971972,-0.3496024013,0.2039181441,0.229376331,0.170839116,-0.0759380162,-0.0538152941,0.1604664326,-0.0303299837,0.107823357,-0.0209005252,0.0539114326,-0.1801201999,-0.1716643721,0.2066053599,-0.3249644637,-0.3378339112,-0.2800188065,-0.2434223294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1939","title":"[firewalled env] OFFLINE mode","comments":"Ok understand better the hanging issue.\r\nI guess catching connection errors is not enough, we should also avoid all the hangings.\r\nCurrently the offline mode tests are only done by simulating an instant connection fail that returns an error, let's have another connection mock that hangs instead.\r\n\r\nI'll also take a look at why you had to do this for `rouge`.\r\n","body":"This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.","comment_length":61,"text":"[firewalled env] OFFLINE mode\nThis issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.\nOk understand better the hanging issue.\r\nI guess catching connection errors is not enough, we should also avoid all the hangings.\r\nCurrently the offline mode tests are only done by simulating an instant connection fail that returns an error, let's have another connection mock that hangs instead.\r\n\r\nI'll also take a look at why you had to do this for `rouge`.\r\n","embeddings":[-0.4748800695,0.1685334891,-0.0625536442,0.1050562486,0.0443113707,-0.2284921706,0.4647973776,0.0716796964,0.2545911074,0.0485309139,0.0184395313,0.0636148751,0.0845065862,0.1993038058,0.0180914346,0.1079676971,0.0199452993,-0.1096389145,-0.2528404295,0.0637163967,-0.2635913491,-0.0536728576,-0.2544517219,-0.2575809062,0.078367807,-0.5080999136,-0.2168778479,-0.0105694095,0.0096480669,-0.2275560498,0.3089389205,0.2612626255,0.2518863976,0.1530115753,-0.0001084152,-0.0537566915,-0.1393844485,-0.0361674428,-0.272887677,0.0223880373,-0.3179339468,-0.3802416921,0.1650776267,-0.4613099992,-0.171355173,-0.174725309,0.1464560628,-0.6594020128,0.5154665709,-0.0763156116,0.0945165008,0.0003517794,-0.3278796375,-0.0994149074,-0.3226190805,-0.1980506182,-0.1814403385,0.1710348725,0.0883782059,-0.2174991071,-0.0455865599,0.3952884078,-0.3122881055,0.1565717608,0.3260605037,0.0182255022,-0.338113457,-0.1188222617,-0.1281385869,0.3376224041,0.1812187731,-0.2216326147,-0.4262379706,-0.4555667341,-0.0790792406,-0.370331347,0.3025760651,0.0889306292,-0.1169174388,0.2809565067,-0.3825724125,-0.2138668299,-0.1247217432,-0.0012391666,-0.0308274534,0.3738206923,0.0736688673,0.1736761779,-0.0692528114,0.1060303599,0.0134036681,-0.3340767622,0.2569811642,0.0270416345,-0.5665541291,0.0265186541,0.1224480569,-0.0067842589,0.0168159846,0.4544952512,-0.1779251099,-0.0429254211,0.4983155131,0.195036903,0.0398950391,0.2975893617,0.1519620121,-0.1790611893,0.3776472807,0.1602634937,-0.0568716712,-0.1345084161,-0.1136473343,-0.0706924796,0.1811163574,0.0640624613,0.1197685003,-0.1377031356,-0.1166463196,0.046965342,-0.0323215127,0.2401375026,0.3313703537,0.3093738258,-0.0555984639,-0.2358305752,0.2333134562,0.1667176485,-0.1382888556,-0.06039121,0.072590515,-0.0984998196,-0.1313179582,0.082668893,0.3286071122,0.2563716769,0.2747864425,-0.1755116731,0.1528570056,0.0238279738,0.5519109368,-0.2458668649,0.1138633564,0.3021980822,0.1545313895,-0.2306014448,0.0980398506,0.2813174427,-0.0375363044,0.0506099239,0.0799426436,-0.3858196735,0.4044837654,0.1439911574,-0.1975332648,-0.0476032272,0.1782198101,0.0742536411,-0.0721478388,-0.4519341588,0.1831779182,0.0667779595,-0.1193479002,-0.0543197319,0.4079632163,0.4936104715,-0.4202961922,0.0228122212,-0.221729517,0.019100355,-0.1197147667,0.1713059396,-0.2794617414,0.1356073618,-0.1801660359,-0.2079320103,0.2443795949,-0.3084554672,-0.3225480914,0.4538064003,-0.1916149855,0.2314610779,0.3152425885,0.1994496435,0.2296581119,-0.0338119008,0.1222665012,0.3941685557,-0.2495333552,-0.0643453002,-0.1496422589,-0.1908909678,0.2760188878,0.1059188098,0.2374101728,-0.1076447815,-0.086140126,-0.2940383255,-0.0350805186,-0.083207719,0.2003037333,-0.1944311261,0.0856611282,0.0736428425,0.0493785702,0.1411918253,-0.4604979455,0.3037669957,-0.2521933019,0.0747477561,-0.3391274512,-0.3241582215,0.2313053608,-0.0713331848,-0.1031431407,-0.3441320658,0.0740442052,0.0512403548,-0.1017183214,-0.0374327078,-0.0173988789,0.0242225807,-0.0870222002,0.1135079041,0.1581827104,-0.0672886223,0.0744352862,0.0992200971,-0.2595793903,-0.1118372455,-0.0186913777,-0.1419194341,-0.0033473987,0.507107079,0.0670183226,0.434106648,0.1801059842,0.43449682,0.100204356,0.3320184648,0.2241401672,0.0292811133,0.0667587593,-0.0281432383,-0.5865287185,0.8847305179,-0.0772907659,0.4949632585,-0.0253681988,-0.0896625817,0.2581321895,-0.070058696,-0.5748965144,-0.0896323398,-0.0292923599,0.2472509444,0.1246782467,-0.0695684403,-0.3645498455,0.1309467256,0.0615998544,0.2113251835,0.4444034696,-0.2755915225,0.2542671561,-0.0863385573,0.4632284939,0.2097619325,0.2965559363,0.2528910935,0.0414115787,0.1776760668,-0.0744988695,-0.2386737466,0.3642019629,0.1546070129,0.3154804409,0.1600142121,0.0825260282,-0.1450218856,-0.1656669974,-0.2701164484,0.1441674232,-0.2285314947,-0.052432321,0.3188200295,-0.3138054311,-0.5460381508,-0.0330595821,0.2567465603,-0.1731722355,-0.1265536398,0.232597664,0.1215340123,-0.3687728345,0.3979503214,-0.2202257961,0.4702404141,-0.0649726614,-0.4042862654,0.2372114658,-0.3736272752,0.0975892395,0.0727022961,0.2432580888,-0.090477258,0.6828110218,0.2652270794,-0.0664931238,-0.172499001,-0.2048208117,0.1804469377,0.0453945622,0.5188974738,0.1410698295,0.097885415,-0.1078779474,0.0392978527,0.266864717,-0.131492421,0.2002531737,-0.1014679745,0.0052114641,0.1161197722,-0.2464787811,-0.5637351274,-0.2847306728,-0.2684567571,0.1167265549,-0.2222549319,0.2011254579,-0.0231130943,0.1001847163,-0.050306119,0.1435732543,0.0995229259,-0.0322472528,-0.568808198,0.1139889956,-0.1637168378,0.0997769907,0.0580960028,-0.0110152299,0.3906186521,0.5869692564,-0.4705568254,-0.2902962863,0.3895433247,0.286469698,0.0516619869,-0.3373523057,0.6150392294,0.0365746021,-0.0880156383,-0.0884277374,0.3548750579,0.4860231876,-0.1864490658,-0.0539083518,0.1791349649,0.7724372149,0.202256009,0.2877366543,-0.1666050255,0.4881629646,0.2260150015,0.1102786735,0.304939419,-0.1625722647,0.0179588497,0.2034069151,-0.0815214068,-0.5336046219,0.0716209486,0.0470552854,-0.0527305119,-0.3611359894,-0.2582453787,-0.1032524407,-0.3916658759,0.2961251736,-0.1127902269,0.4995014668,0.117896229,0.372337997,-0.2182418406,0.1661288887,-0.1418527663,0.1501836479,0.5369586945,-0.0128860902,-0.1799632609,0.0348873436,-0.22274369,0.2784568965,0.0099335965,0.3417262137,-0.036391452,0.0174978469,0.0015469281,-0.0763285756,0.4463128746,-0.1146823615,0.0632685199,0.0081591588,-0.2919878364,-0.2112849802,-0.1305243969,-0.0428579152,-0.0211985148,0.0649969876,0.1952529997,-0.2032026201,-0.0390476733,-0.3187787533,0.1351860166,-0.3197744787,-0.2367112786,0.0234858245,0.1267216206,-0.6222805977,0.420473069,-0.1571975201,-0.1905483156,-0.2480874062,-0.0747949705,-0.1783072352,0.4812121689,-0.145950079,0.0565128624,0.1470396221,-0.5123636127,0.1858813465,0.1123727337,0.1500660181,0.3838641942,0.0461331084,0.1363827586,0.284540683,0.0637651309,0.066484876,0.2051711679,0.4261293709,-0.1089276075,-0.0030097992,0.1616743952,0.3030042052,-0.3588068783,-0.3979021609,0.1224461347,0.1558630913,0.0335068256,-0.4153938591,0.0671153069,0.04605335,-0.2596212626,0.0445624702,0.1521447897,-0.3728854656,0.2617143989,0.2630714774,0.9972520471,0.2139621526,0.2179925889,0.4170042574,0.2532133162,0.4726442695,-0.6332621574,0.2983065844,0.1272928864,-0.1670437455,-0.1499634385,-0.02558963,0.2321767211,0.099870272,-0.2685555816,0.5433549285,-0.0701307207,0.1014339775,-0.2433125675,-0.0296450164,-0.2376535684,-0.4036455154,-0.2663605511,0.1260658652,0.2334099859,-0.0104813557,-0.3751176894,-0.068979308,0.1925515532,0.1631508023,0.14545542,0.0721764639,0.0428949408,0.2766603231,-0.419208467,0.0559381731,-0.2763850689,0.1090795174,0.0180981327,0.2108995914,-0.2886296809,0.1259953529,-0.0419275537,0.1140722334,-0.1539172381,-0.2668178678,0.372919023,-0.1662880629,-0.2487163991,-0.2205847353,-0.3582707942,-0.0771528035,0.2514949441,-0.3500300646,0.0939490199,-0.0691653416,0.0619472638,0.1937758923,-0.1183083802,-0.215838626,0.0828801244,0.1098431647,-0.0477760732,0.2642475665,0.0833750889,-0.0296078119,-0.1844729483,0.5224729776,-0.0124881873,0.2983217239,0.329159528,0.17099756,-0.2422726303,-0.3916774988,-0.1179054156,-0.0944124162,0.1713488996,0.0689709634,0.0006624407,-0.0254697464,0.0580743439,-0.2008399367,0.1766274869,-0.0651128218,-0.1013034061,-0.4263277054,-0.0957949758,0.3383998871,0.2018853426,0.0942364931,0.1214297861,0.1285647601,0.0334427096,0.011660737,-0.3435808122,-0.1933899075,-0.2666880488,0.2216063738,-0.0113147693,-0.0477256738,0.1774714738,-0.0150593249,0.0563594587,-0.0292195305,-0.2538238168,-0.1619534642,-0.1136552319,0.2258027345,-0.1483045071,-0.3157637417,0.2160577774,-0.423892051,-0.3664358854,-0.3153072298,0.0918554738,-0.0170711968,-0.1677893996,-0.4335837364,-0.1071414053,0.3017937243,-0.2243884355,0.0949885547,-0.0977727994,-0.2725069225,-0.3345758319,0.0538960434,-0.0734193325,-0.1096151099,-0.0395689383,0.1027320102,0.2106037289,0.205578953,0.4639981687,0.0892977566,-0.1204226613,0.1604589671,0.6960189939,0.1433717161,0.1893894821,0.1112304628,0.0307057649,0.2147427052,-0.4058090746,0.1708363444,-0.2121228129,-0.1652800739,0.2011716664,0.2182677984,0.4211681187,0.1228163764,0.2115367651,-0.0768140182,0.3960933387,-0.3639710844,0.2452284545,-0.0232135504,-0.0485098064,0.1692207754,-0.0433972329,0.1081223264,0.3069615662,0.0682089925,-0.269303441,0.1668248624,0.4557292759,0.2022744119,-0.0908759907,-0.4782039225,-0.1009483114,0.0956822485,-0.0603445247,0.0244362708,-0.464867115,-0.0542507693,-0.3395825922,-0.1913392246,-0.2145864964,0.6027536392,-0.3315451741,-0.1462776661,-0.4204238951,-0.4396710098,0.0110949399,0.0038176409,0.1870268434,-0.138487637,-0.1736916751,0.0796872526,0.2774659395,-0.2712741792,0.0969131291,-0.0292456578,0.1431989819,-0.323631078,0.1587519348,0.5755221844,0.0011660997,0.0627885014,0.2858873308,0.2266620696,0.0022768998,-0.1239224076,0.1276148558,0.1615405679,0.0533404723,-0.3756064475,-0.0559492894,-0.016708402,0.2881965935,0.3220778406,0.1017410532,-0.0451429375,0.503982842,-0.0574037246,-0.3350550532,-0.217862308,-0.0598231368,0.4229668081,0.1130177602,-0.0961265415,-0.0181783196,-0.1123341396,-0.0917096362,0.2774384022,-0.1616678834,0.0388007686,0.0190163963,0.0622579455,-0.1917547733,0.5251255035,-0.0828251243,0.1001403928,-0.3585446477,-0.0513545163,-0.6031898856,-0.2333036214,-0.0481282994,0.050825011,-0.1225795448,0.1338446736,0.2245391905,0.1548327804,-0.2501180172,0.024877632,0.1090888679,0.0784571469,-0.3683554232,0.0319491997,0.0217686146,-0.0349099971,0.1462561339,-0.2422401905,0.2108056098,-0.1415773928,0.0672279671,0.307381779,0.0241460316,0.0134564163,0.1558464766,0.7136694193,0.0963875428,0.3541609049,0.0330818854,-0.0392261669,-0.1355392486,0.0061807684,-0.1747146398,0.0417080708,0.042096138,0.2525860369,-0.1677740067,-0.1356784999,-0.1732335538,0.0797532499,-0.0598759577,-0.006825991,-0.1402485371,-0.1277015358,-0.1939341277,0.1754536778,-0.0835498869,0.2089730799,0.1280008405,0.227099821,-0.2550806999,-0.1245743409,0.4326334,-0.0996425748,0.1882971972,-0.3496024013,0.2039181441,0.229376331,0.170839116,-0.0759380162,-0.0538152941,0.1604664326,-0.0303299837,0.107823357,-0.0209005252,0.0539114326,-0.1801201999,-0.1716643721,0.2066053599,-0.3249644637,-0.3378339112,-0.2800188065,-0.2434223294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1939","title":"[firewalled env] OFFLINE mode","comments":"FWIW, I think instant failure on the behalf of a network call is the simplest solution to correctly represent the environment and having the caller to sort it out is the next thing to do, since here it is the case of having no functional network, it's just that the software doesn't know this is the case, because there is some network. So we just need to help it to bail out instantly rather than hang waiting for it to time out. And afterwards everything else you said.","body":"This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.","comment_length":88,"text":"[firewalled env] OFFLINE mode\nThis issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.\nFWIW, I think instant failure on the behalf of a network call is the simplest solution to correctly represent the environment and having the caller to sort it out is the next thing to do, since here it is the case of having no functional network, it's just that the software doesn't know this is the case, because there is some network. So we just need to help it to bail out instantly rather than hang waiting for it to time out. And afterwards everything else you said.","embeddings":[-0.4748800695,0.1685334891,-0.0625536442,0.1050562486,0.0443113707,-0.2284921706,0.4647973776,0.0716796964,0.2545911074,0.0485309139,0.0184395313,0.0636148751,0.0845065862,0.1993038058,0.0180914346,0.1079676971,0.0199452993,-0.1096389145,-0.2528404295,0.0637163967,-0.2635913491,-0.0536728576,-0.2544517219,-0.2575809062,0.078367807,-0.5080999136,-0.2168778479,-0.0105694095,0.0096480669,-0.2275560498,0.3089389205,0.2612626255,0.2518863976,0.1530115753,-0.0001084152,-0.0537566915,-0.1393844485,-0.0361674428,-0.272887677,0.0223880373,-0.3179339468,-0.3802416921,0.1650776267,-0.4613099992,-0.171355173,-0.174725309,0.1464560628,-0.6594020128,0.5154665709,-0.0763156116,0.0945165008,0.0003517794,-0.3278796375,-0.0994149074,-0.3226190805,-0.1980506182,-0.1814403385,0.1710348725,0.0883782059,-0.2174991071,-0.0455865599,0.3952884078,-0.3122881055,0.1565717608,0.3260605037,0.0182255022,-0.338113457,-0.1188222617,-0.1281385869,0.3376224041,0.1812187731,-0.2216326147,-0.4262379706,-0.4555667341,-0.0790792406,-0.370331347,0.3025760651,0.0889306292,-0.1169174388,0.2809565067,-0.3825724125,-0.2138668299,-0.1247217432,-0.0012391666,-0.0308274534,0.3738206923,0.0736688673,0.1736761779,-0.0692528114,0.1060303599,0.0134036681,-0.3340767622,0.2569811642,0.0270416345,-0.5665541291,0.0265186541,0.1224480569,-0.0067842589,0.0168159846,0.4544952512,-0.1779251099,-0.0429254211,0.4983155131,0.195036903,0.0398950391,0.2975893617,0.1519620121,-0.1790611893,0.3776472807,0.1602634937,-0.0568716712,-0.1345084161,-0.1136473343,-0.0706924796,0.1811163574,0.0640624613,0.1197685003,-0.1377031356,-0.1166463196,0.046965342,-0.0323215127,0.2401375026,0.3313703537,0.3093738258,-0.0555984639,-0.2358305752,0.2333134562,0.1667176485,-0.1382888556,-0.06039121,0.072590515,-0.0984998196,-0.1313179582,0.082668893,0.3286071122,0.2563716769,0.2747864425,-0.1755116731,0.1528570056,0.0238279738,0.5519109368,-0.2458668649,0.1138633564,0.3021980822,0.1545313895,-0.2306014448,0.0980398506,0.2813174427,-0.0375363044,0.0506099239,0.0799426436,-0.3858196735,0.4044837654,0.1439911574,-0.1975332648,-0.0476032272,0.1782198101,0.0742536411,-0.0721478388,-0.4519341588,0.1831779182,0.0667779595,-0.1193479002,-0.0543197319,0.4079632163,0.4936104715,-0.4202961922,0.0228122212,-0.221729517,0.019100355,-0.1197147667,0.1713059396,-0.2794617414,0.1356073618,-0.1801660359,-0.2079320103,0.2443795949,-0.3084554672,-0.3225480914,0.4538064003,-0.1916149855,0.2314610779,0.3152425885,0.1994496435,0.2296581119,-0.0338119008,0.1222665012,0.3941685557,-0.2495333552,-0.0643453002,-0.1496422589,-0.1908909678,0.2760188878,0.1059188098,0.2374101728,-0.1076447815,-0.086140126,-0.2940383255,-0.0350805186,-0.083207719,0.2003037333,-0.1944311261,0.0856611282,0.0736428425,0.0493785702,0.1411918253,-0.4604979455,0.3037669957,-0.2521933019,0.0747477561,-0.3391274512,-0.3241582215,0.2313053608,-0.0713331848,-0.1031431407,-0.3441320658,0.0740442052,0.0512403548,-0.1017183214,-0.0374327078,-0.0173988789,0.0242225807,-0.0870222002,0.1135079041,0.1581827104,-0.0672886223,0.0744352862,0.0992200971,-0.2595793903,-0.1118372455,-0.0186913777,-0.1419194341,-0.0033473987,0.507107079,0.0670183226,0.434106648,0.1801059842,0.43449682,0.100204356,0.3320184648,0.2241401672,0.0292811133,0.0667587593,-0.0281432383,-0.5865287185,0.8847305179,-0.0772907659,0.4949632585,-0.0253681988,-0.0896625817,0.2581321895,-0.070058696,-0.5748965144,-0.0896323398,-0.0292923599,0.2472509444,0.1246782467,-0.0695684403,-0.3645498455,0.1309467256,0.0615998544,0.2113251835,0.4444034696,-0.2755915225,0.2542671561,-0.0863385573,0.4632284939,0.2097619325,0.2965559363,0.2528910935,0.0414115787,0.1776760668,-0.0744988695,-0.2386737466,0.3642019629,0.1546070129,0.3154804409,0.1600142121,0.0825260282,-0.1450218856,-0.1656669974,-0.2701164484,0.1441674232,-0.2285314947,-0.052432321,0.3188200295,-0.3138054311,-0.5460381508,-0.0330595821,0.2567465603,-0.1731722355,-0.1265536398,0.232597664,0.1215340123,-0.3687728345,0.3979503214,-0.2202257961,0.4702404141,-0.0649726614,-0.4042862654,0.2372114658,-0.3736272752,0.0975892395,0.0727022961,0.2432580888,-0.090477258,0.6828110218,0.2652270794,-0.0664931238,-0.172499001,-0.2048208117,0.1804469377,0.0453945622,0.5188974738,0.1410698295,0.097885415,-0.1078779474,0.0392978527,0.266864717,-0.131492421,0.2002531737,-0.1014679745,0.0052114641,0.1161197722,-0.2464787811,-0.5637351274,-0.2847306728,-0.2684567571,0.1167265549,-0.2222549319,0.2011254579,-0.0231130943,0.1001847163,-0.050306119,0.1435732543,0.0995229259,-0.0322472528,-0.568808198,0.1139889956,-0.1637168378,0.0997769907,0.0580960028,-0.0110152299,0.3906186521,0.5869692564,-0.4705568254,-0.2902962863,0.3895433247,0.286469698,0.0516619869,-0.3373523057,0.6150392294,0.0365746021,-0.0880156383,-0.0884277374,0.3548750579,0.4860231876,-0.1864490658,-0.0539083518,0.1791349649,0.7724372149,0.202256009,0.2877366543,-0.1666050255,0.4881629646,0.2260150015,0.1102786735,0.304939419,-0.1625722647,0.0179588497,0.2034069151,-0.0815214068,-0.5336046219,0.0716209486,0.0470552854,-0.0527305119,-0.3611359894,-0.2582453787,-0.1032524407,-0.3916658759,0.2961251736,-0.1127902269,0.4995014668,0.117896229,0.372337997,-0.2182418406,0.1661288887,-0.1418527663,0.1501836479,0.5369586945,-0.0128860902,-0.1799632609,0.0348873436,-0.22274369,0.2784568965,0.0099335965,0.3417262137,-0.036391452,0.0174978469,0.0015469281,-0.0763285756,0.4463128746,-0.1146823615,0.0632685199,0.0081591588,-0.2919878364,-0.2112849802,-0.1305243969,-0.0428579152,-0.0211985148,0.0649969876,0.1952529997,-0.2032026201,-0.0390476733,-0.3187787533,0.1351860166,-0.3197744787,-0.2367112786,0.0234858245,0.1267216206,-0.6222805977,0.420473069,-0.1571975201,-0.1905483156,-0.2480874062,-0.0747949705,-0.1783072352,0.4812121689,-0.145950079,0.0565128624,0.1470396221,-0.5123636127,0.1858813465,0.1123727337,0.1500660181,0.3838641942,0.0461331084,0.1363827586,0.284540683,0.0637651309,0.066484876,0.2051711679,0.4261293709,-0.1089276075,-0.0030097992,0.1616743952,0.3030042052,-0.3588068783,-0.3979021609,0.1224461347,0.1558630913,0.0335068256,-0.4153938591,0.0671153069,0.04605335,-0.2596212626,0.0445624702,0.1521447897,-0.3728854656,0.2617143989,0.2630714774,0.9972520471,0.2139621526,0.2179925889,0.4170042574,0.2532133162,0.4726442695,-0.6332621574,0.2983065844,0.1272928864,-0.1670437455,-0.1499634385,-0.02558963,0.2321767211,0.099870272,-0.2685555816,0.5433549285,-0.0701307207,0.1014339775,-0.2433125675,-0.0296450164,-0.2376535684,-0.4036455154,-0.2663605511,0.1260658652,0.2334099859,-0.0104813557,-0.3751176894,-0.068979308,0.1925515532,0.1631508023,0.14545542,0.0721764639,0.0428949408,0.2766603231,-0.419208467,0.0559381731,-0.2763850689,0.1090795174,0.0180981327,0.2108995914,-0.2886296809,0.1259953529,-0.0419275537,0.1140722334,-0.1539172381,-0.2668178678,0.372919023,-0.1662880629,-0.2487163991,-0.2205847353,-0.3582707942,-0.0771528035,0.2514949441,-0.3500300646,0.0939490199,-0.0691653416,0.0619472638,0.1937758923,-0.1183083802,-0.215838626,0.0828801244,0.1098431647,-0.0477760732,0.2642475665,0.0833750889,-0.0296078119,-0.1844729483,0.5224729776,-0.0124881873,0.2983217239,0.329159528,0.17099756,-0.2422726303,-0.3916774988,-0.1179054156,-0.0944124162,0.1713488996,0.0689709634,0.0006624407,-0.0254697464,0.0580743439,-0.2008399367,0.1766274869,-0.0651128218,-0.1013034061,-0.4263277054,-0.0957949758,0.3383998871,0.2018853426,0.0942364931,0.1214297861,0.1285647601,0.0334427096,0.011660737,-0.3435808122,-0.1933899075,-0.2666880488,0.2216063738,-0.0113147693,-0.0477256738,0.1774714738,-0.0150593249,0.0563594587,-0.0292195305,-0.2538238168,-0.1619534642,-0.1136552319,0.2258027345,-0.1483045071,-0.3157637417,0.2160577774,-0.423892051,-0.3664358854,-0.3153072298,0.0918554738,-0.0170711968,-0.1677893996,-0.4335837364,-0.1071414053,0.3017937243,-0.2243884355,0.0949885547,-0.0977727994,-0.2725069225,-0.3345758319,0.0538960434,-0.0734193325,-0.1096151099,-0.0395689383,0.1027320102,0.2106037289,0.205578953,0.4639981687,0.0892977566,-0.1204226613,0.1604589671,0.6960189939,0.1433717161,0.1893894821,0.1112304628,0.0307057649,0.2147427052,-0.4058090746,0.1708363444,-0.2121228129,-0.1652800739,0.2011716664,0.2182677984,0.4211681187,0.1228163764,0.2115367651,-0.0768140182,0.3960933387,-0.3639710844,0.2452284545,-0.0232135504,-0.0485098064,0.1692207754,-0.0433972329,0.1081223264,0.3069615662,0.0682089925,-0.269303441,0.1668248624,0.4557292759,0.2022744119,-0.0908759907,-0.4782039225,-0.1009483114,0.0956822485,-0.0603445247,0.0244362708,-0.464867115,-0.0542507693,-0.3395825922,-0.1913392246,-0.2145864964,0.6027536392,-0.3315451741,-0.1462776661,-0.4204238951,-0.4396710098,0.0110949399,0.0038176409,0.1870268434,-0.138487637,-0.1736916751,0.0796872526,0.2774659395,-0.2712741792,0.0969131291,-0.0292456578,0.1431989819,-0.323631078,0.1587519348,0.5755221844,0.0011660997,0.0627885014,0.2858873308,0.2266620696,0.0022768998,-0.1239224076,0.1276148558,0.1615405679,0.0533404723,-0.3756064475,-0.0559492894,-0.016708402,0.2881965935,0.3220778406,0.1017410532,-0.0451429375,0.503982842,-0.0574037246,-0.3350550532,-0.217862308,-0.0598231368,0.4229668081,0.1130177602,-0.0961265415,-0.0181783196,-0.1123341396,-0.0917096362,0.2774384022,-0.1616678834,0.0388007686,0.0190163963,0.0622579455,-0.1917547733,0.5251255035,-0.0828251243,0.1001403928,-0.3585446477,-0.0513545163,-0.6031898856,-0.2333036214,-0.0481282994,0.050825011,-0.1225795448,0.1338446736,0.2245391905,0.1548327804,-0.2501180172,0.024877632,0.1090888679,0.0784571469,-0.3683554232,0.0319491997,0.0217686146,-0.0349099971,0.1462561339,-0.2422401905,0.2108056098,-0.1415773928,0.0672279671,0.307381779,0.0241460316,0.0134564163,0.1558464766,0.7136694193,0.0963875428,0.3541609049,0.0330818854,-0.0392261669,-0.1355392486,0.0061807684,-0.1747146398,0.0417080708,0.042096138,0.2525860369,-0.1677740067,-0.1356784999,-0.1732335538,0.0797532499,-0.0598759577,-0.006825991,-0.1402485371,-0.1277015358,-0.1939341277,0.1754536778,-0.0835498869,0.2089730799,0.1280008405,0.227099821,-0.2550806999,-0.1245743409,0.4326334,-0.0996425748,0.1882971972,-0.3496024013,0.2039181441,0.229376331,0.170839116,-0.0759380162,-0.0538152941,0.1604664326,-0.0303299837,0.107823357,-0.0209005252,0.0539114326,-0.1801201999,-0.1716643721,0.2066053599,-0.3249644637,-0.3378339112,-0.2800188065,-0.2434223294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1939","title":"[firewalled env] OFFLINE mode","comments":"Update on this: \r\n\r\nI managed to create a mock environment for tests that makes the connections hang until timeout.\r\nI managed to reproduce the issue you're having in this environment.\r\n\r\nI'll update the offline test cases to also test the robustness to connection hangings, and make sure we set proper timeouts where it's needed in the code. This should cover the _automatic_ section you mentioned.","body":"This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.","comment_length":65,"text":"[firewalled env] OFFLINE mode\nThis issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.\nUpdate on this: \r\n\r\nI managed to create a mock environment for tests that makes the connections hang until timeout.\r\nI managed to reproduce the issue you're having in this environment.\r\n\r\nI'll update the offline test cases to also test the robustness to connection hangings, and make sure we set proper timeouts where it's needed in the code. This should cover the _automatic_ section you mentioned.","embeddings":[-0.4748800695,0.1685334891,-0.0625536442,0.1050562486,0.0443113707,-0.2284921706,0.4647973776,0.0716796964,0.2545911074,0.0485309139,0.0184395313,0.0636148751,0.0845065862,0.1993038058,0.0180914346,0.1079676971,0.0199452993,-0.1096389145,-0.2528404295,0.0637163967,-0.2635913491,-0.0536728576,-0.2544517219,-0.2575809062,0.078367807,-0.5080999136,-0.2168778479,-0.0105694095,0.0096480669,-0.2275560498,0.3089389205,0.2612626255,0.2518863976,0.1530115753,-0.0001084152,-0.0537566915,-0.1393844485,-0.0361674428,-0.272887677,0.0223880373,-0.3179339468,-0.3802416921,0.1650776267,-0.4613099992,-0.171355173,-0.174725309,0.1464560628,-0.6594020128,0.5154665709,-0.0763156116,0.0945165008,0.0003517794,-0.3278796375,-0.0994149074,-0.3226190805,-0.1980506182,-0.1814403385,0.1710348725,0.0883782059,-0.2174991071,-0.0455865599,0.3952884078,-0.3122881055,0.1565717608,0.3260605037,0.0182255022,-0.338113457,-0.1188222617,-0.1281385869,0.3376224041,0.1812187731,-0.2216326147,-0.4262379706,-0.4555667341,-0.0790792406,-0.370331347,0.3025760651,0.0889306292,-0.1169174388,0.2809565067,-0.3825724125,-0.2138668299,-0.1247217432,-0.0012391666,-0.0308274534,0.3738206923,0.0736688673,0.1736761779,-0.0692528114,0.1060303599,0.0134036681,-0.3340767622,0.2569811642,0.0270416345,-0.5665541291,0.0265186541,0.1224480569,-0.0067842589,0.0168159846,0.4544952512,-0.1779251099,-0.0429254211,0.4983155131,0.195036903,0.0398950391,0.2975893617,0.1519620121,-0.1790611893,0.3776472807,0.1602634937,-0.0568716712,-0.1345084161,-0.1136473343,-0.0706924796,0.1811163574,0.0640624613,0.1197685003,-0.1377031356,-0.1166463196,0.046965342,-0.0323215127,0.2401375026,0.3313703537,0.3093738258,-0.0555984639,-0.2358305752,0.2333134562,0.1667176485,-0.1382888556,-0.06039121,0.072590515,-0.0984998196,-0.1313179582,0.082668893,0.3286071122,0.2563716769,0.2747864425,-0.1755116731,0.1528570056,0.0238279738,0.5519109368,-0.2458668649,0.1138633564,0.3021980822,0.1545313895,-0.2306014448,0.0980398506,0.2813174427,-0.0375363044,0.0506099239,0.0799426436,-0.3858196735,0.4044837654,0.1439911574,-0.1975332648,-0.0476032272,0.1782198101,0.0742536411,-0.0721478388,-0.4519341588,0.1831779182,0.0667779595,-0.1193479002,-0.0543197319,0.4079632163,0.4936104715,-0.4202961922,0.0228122212,-0.221729517,0.019100355,-0.1197147667,0.1713059396,-0.2794617414,0.1356073618,-0.1801660359,-0.2079320103,0.2443795949,-0.3084554672,-0.3225480914,0.4538064003,-0.1916149855,0.2314610779,0.3152425885,0.1994496435,0.2296581119,-0.0338119008,0.1222665012,0.3941685557,-0.2495333552,-0.0643453002,-0.1496422589,-0.1908909678,0.2760188878,0.1059188098,0.2374101728,-0.1076447815,-0.086140126,-0.2940383255,-0.0350805186,-0.083207719,0.2003037333,-0.1944311261,0.0856611282,0.0736428425,0.0493785702,0.1411918253,-0.4604979455,0.3037669957,-0.2521933019,0.0747477561,-0.3391274512,-0.3241582215,0.2313053608,-0.0713331848,-0.1031431407,-0.3441320658,0.0740442052,0.0512403548,-0.1017183214,-0.0374327078,-0.0173988789,0.0242225807,-0.0870222002,0.1135079041,0.1581827104,-0.0672886223,0.0744352862,0.0992200971,-0.2595793903,-0.1118372455,-0.0186913777,-0.1419194341,-0.0033473987,0.507107079,0.0670183226,0.434106648,0.1801059842,0.43449682,0.100204356,0.3320184648,0.2241401672,0.0292811133,0.0667587593,-0.0281432383,-0.5865287185,0.8847305179,-0.0772907659,0.4949632585,-0.0253681988,-0.0896625817,0.2581321895,-0.070058696,-0.5748965144,-0.0896323398,-0.0292923599,0.2472509444,0.1246782467,-0.0695684403,-0.3645498455,0.1309467256,0.0615998544,0.2113251835,0.4444034696,-0.2755915225,0.2542671561,-0.0863385573,0.4632284939,0.2097619325,0.2965559363,0.2528910935,0.0414115787,0.1776760668,-0.0744988695,-0.2386737466,0.3642019629,0.1546070129,0.3154804409,0.1600142121,0.0825260282,-0.1450218856,-0.1656669974,-0.2701164484,0.1441674232,-0.2285314947,-0.052432321,0.3188200295,-0.3138054311,-0.5460381508,-0.0330595821,0.2567465603,-0.1731722355,-0.1265536398,0.232597664,0.1215340123,-0.3687728345,0.3979503214,-0.2202257961,0.4702404141,-0.0649726614,-0.4042862654,0.2372114658,-0.3736272752,0.0975892395,0.0727022961,0.2432580888,-0.090477258,0.6828110218,0.2652270794,-0.0664931238,-0.172499001,-0.2048208117,0.1804469377,0.0453945622,0.5188974738,0.1410698295,0.097885415,-0.1078779474,0.0392978527,0.266864717,-0.131492421,0.2002531737,-0.1014679745,0.0052114641,0.1161197722,-0.2464787811,-0.5637351274,-0.2847306728,-0.2684567571,0.1167265549,-0.2222549319,0.2011254579,-0.0231130943,0.1001847163,-0.050306119,0.1435732543,0.0995229259,-0.0322472528,-0.568808198,0.1139889956,-0.1637168378,0.0997769907,0.0580960028,-0.0110152299,0.3906186521,0.5869692564,-0.4705568254,-0.2902962863,0.3895433247,0.286469698,0.0516619869,-0.3373523057,0.6150392294,0.0365746021,-0.0880156383,-0.0884277374,0.3548750579,0.4860231876,-0.1864490658,-0.0539083518,0.1791349649,0.7724372149,0.202256009,0.2877366543,-0.1666050255,0.4881629646,0.2260150015,0.1102786735,0.304939419,-0.1625722647,0.0179588497,0.2034069151,-0.0815214068,-0.5336046219,0.0716209486,0.0470552854,-0.0527305119,-0.3611359894,-0.2582453787,-0.1032524407,-0.3916658759,0.2961251736,-0.1127902269,0.4995014668,0.117896229,0.372337997,-0.2182418406,0.1661288887,-0.1418527663,0.1501836479,0.5369586945,-0.0128860902,-0.1799632609,0.0348873436,-0.22274369,0.2784568965,0.0099335965,0.3417262137,-0.036391452,0.0174978469,0.0015469281,-0.0763285756,0.4463128746,-0.1146823615,0.0632685199,0.0081591588,-0.2919878364,-0.2112849802,-0.1305243969,-0.0428579152,-0.0211985148,0.0649969876,0.1952529997,-0.2032026201,-0.0390476733,-0.3187787533,0.1351860166,-0.3197744787,-0.2367112786,0.0234858245,0.1267216206,-0.6222805977,0.420473069,-0.1571975201,-0.1905483156,-0.2480874062,-0.0747949705,-0.1783072352,0.4812121689,-0.145950079,0.0565128624,0.1470396221,-0.5123636127,0.1858813465,0.1123727337,0.1500660181,0.3838641942,0.0461331084,0.1363827586,0.284540683,0.0637651309,0.066484876,0.2051711679,0.4261293709,-0.1089276075,-0.0030097992,0.1616743952,0.3030042052,-0.3588068783,-0.3979021609,0.1224461347,0.1558630913,0.0335068256,-0.4153938591,0.0671153069,0.04605335,-0.2596212626,0.0445624702,0.1521447897,-0.3728854656,0.2617143989,0.2630714774,0.9972520471,0.2139621526,0.2179925889,0.4170042574,0.2532133162,0.4726442695,-0.6332621574,0.2983065844,0.1272928864,-0.1670437455,-0.1499634385,-0.02558963,0.2321767211,0.099870272,-0.2685555816,0.5433549285,-0.0701307207,0.1014339775,-0.2433125675,-0.0296450164,-0.2376535684,-0.4036455154,-0.2663605511,0.1260658652,0.2334099859,-0.0104813557,-0.3751176894,-0.068979308,0.1925515532,0.1631508023,0.14545542,0.0721764639,0.0428949408,0.2766603231,-0.419208467,0.0559381731,-0.2763850689,0.1090795174,0.0180981327,0.2108995914,-0.2886296809,0.1259953529,-0.0419275537,0.1140722334,-0.1539172381,-0.2668178678,0.372919023,-0.1662880629,-0.2487163991,-0.2205847353,-0.3582707942,-0.0771528035,0.2514949441,-0.3500300646,0.0939490199,-0.0691653416,0.0619472638,0.1937758923,-0.1183083802,-0.215838626,0.0828801244,0.1098431647,-0.0477760732,0.2642475665,0.0833750889,-0.0296078119,-0.1844729483,0.5224729776,-0.0124881873,0.2983217239,0.329159528,0.17099756,-0.2422726303,-0.3916774988,-0.1179054156,-0.0944124162,0.1713488996,0.0689709634,0.0006624407,-0.0254697464,0.0580743439,-0.2008399367,0.1766274869,-0.0651128218,-0.1013034061,-0.4263277054,-0.0957949758,0.3383998871,0.2018853426,0.0942364931,0.1214297861,0.1285647601,0.0334427096,0.011660737,-0.3435808122,-0.1933899075,-0.2666880488,0.2216063738,-0.0113147693,-0.0477256738,0.1774714738,-0.0150593249,0.0563594587,-0.0292195305,-0.2538238168,-0.1619534642,-0.1136552319,0.2258027345,-0.1483045071,-0.3157637417,0.2160577774,-0.423892051,-0.3664358854,-0.3153072298,0.0918554738,-0.0170711968,-0.1677893996,-0.4335837364,-0.1071414053,0.3017937243,-0.2243884355,0.0949885547,-0.0977727994,-0.2725069225,-0.3345758319,0.0538960434,-0.0734193325,-0.1096151099,-0.0395689383,0.1027320102,0.2106037289,0.205578953,0.4639981687,0.0892977566,-0.1204226613,0.1604589671,0.6960189939,0.1433717161,0.1893894821,0.1112304628,0.0307057649,0.2147427052,-0.4058090746,0.1708363444,-0.2121228129,-0.1652800739,0.2011716664,0.2182677984,0.4211681187,0.1228163764,0.2115367651,-0.0768140182,0.3960933387,-0.3639710844,0.2452284545,-0.0232135504,-0.0485098064,0.1692207754,-0.0433972329,0.1081223264,0.3069615662,0.0682089925,-0.269303441,0.1668248624,0.4557292759,0.2022744119,-0.0908759907,-0.4782039225,-0.1009483114,0.0956822485,-0.0603445247,0.0244362708,-0.464867115,-0.0542507693,-0.3395825922,-0.1913392246,-0.2145864964,0.6027536392,-0.3315451741,-0.1462776661,-0.4204238951,-0.4396710098,0.0110949399,0.0038176409,0.1870268434,-0.138487637,-0.1736916751,0.0796872526,0.2774659395,-0.2712741792,0.0969131291,-0.0292456578,0.1431989819,-0.323631078,0.1587519348,0.5755221844,0.0011660997,0.0627885014,0.2858873308,0.2266620696,0.0022768998,-0.1239224076,0.1276148558,0.1615405679,0.0533404723,-0.3756064475,-0.0559492894,-0.016708402,0.2881965935,0.3220778406,0.1017410532,-0.0451429375,0.503982842,-0.0574037246,-0.3350550532,-0.217862308,-0.0598231368,0.4229668081,0.1130177602,-0.0961265415,-0.0181783196,-0.1123341396,-0.0917096362,0.2774384022,-0.1616678834,0.0388007686,0.0190163963,0.0622579455,-0.1917547733,0.5251255035,-0.0828251243,0.1001403928,-0.3585446477,-0.0513545163,-0.6031898856,-0.2333036214,-0.0481282994,0.050825011,-0.1225795448,0.1338446736,0.2245391905,0.1548327804,-0.2501180172,0.024877632,0.1090888679,0.0784571469,-0.3683554232,0.0319491997,0.0217686146,-0.0349099971,0.1462561339,-0.2422401905,0.2108056098,-0.1415773928,0.0672279671,0.307381779,0.0241460316,0.0134564163,0.1558464766,0.7136694193,0.0963875428,0.3541609049,0.0330818854,-0.0392261669,-0.1355392486,0.0061807684,-0.1747146398,0.0417080708,0.042096138,0.2525860369,-0.1677740067,-0.1356784999,-0.1732335538,0.0797532499,-0.0598759577,-0.006825991,-0.1402485371,-0.1277015358,-0.1939341277,0.1754536778,-0.0835498869,0.2089730799,0.1280008405,0.227099821,-0.2550806999,-0.1245743409,0.4326334,-0.0996425748,0.1882971972,-0.3496024013,0.2039181441,0.229376331,0.170839116,-0.0759380162,-0.0538152941,0.1604664326,-0.0303299837,0.107823357,-0.0209005252,0.0539114326,-0.1801201999,-0.1716643721,0.2066053599,-0.3249644637,-0.3378339112,-0.2800188065,-0.2434223294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1939","title":"[firewalled env] OFFLINE mode","comments":"Fabulous! I'm glad you were able to reproduce the issues, @lhoestq!","body":"This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.","comment_length":11,"text":"[firewalled env] OFFLINE mode\nThis issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.\nFabulous! I'm glad you were able to reproduce the issues, @lhoestq!","embeddings":[-0.4748800695,0.1685334891,-0.0625536442,0.1050562486,0.0443113707,-0.2284921706,0.4647973776,0.0716796964,0.2545911074,0.0485309139,0.0184395313,0.0636148751,0.0845065862,0.1993038058,0.0180914346,0.1079676971,0.0199452993,-0.1096389145,-0.2528404295,0.0637163967,-0.2635913491,-0.0536728576,-0.2544517219,-0.2575809062,0.078367807,-0.5080999136,-0.2168778479,-0.0105694095,0.0096480669,-0.2275560498,0.3089389205,0.2612626255,0.2518863976,0.1530115753,-0.0001084152,-0.0537566915,-0.1393844485,-0.0361674428,-0.272887677,0.0223880373,-0.3179339468,-0.3802416921,0.1650776267,-0.4613099992,-0.171355173,-0.174725309,0.1464560628,-0.6594020128,0.5154665709,-0.0763156116,0.0945165008,0.0003517794,-0.3278796375,-0.0994149074,-0.3226190805,-0.1980506182,-0.1814403385,0.1710348725,0.0883782059,-0.2174991071,-0.0455865599,0.3952884078,-0.3122881055,0.1565717608,0.3260605037,0.0182255022,-0.338113457,-0.1188222617,-0.1281385869,0.3376224041,0.1812187731,-0.2216326147,-0.4262379706,-0.4555667341,-0.0790792406,-0.370331347,0.3025760651,0.0889306292,-0.1169174388,0.2809565067,-0.3825724125,-0.2138668299,-0.1247217432,-0.0012391666,-0.0308274534,0.3738206923,0.0736688673,0.1736761779,-0.0692528114,0.1060303599,0.0134036681,-0.3340767622,0.2569811642,0.0270416345,-0.5665541291,0.0265186541,0.1224480569,-0.0067842589,0.0168159846,0.4544952512,-0.1779251099,-0.0429254211,0.4983155131,0.195036903,0.0398950391,0.2975893617,0.1519620121,-0.1790611893,0.3776472807,0.1602634937,-0.0568716712,-0.1345084161,-0.1136473343,-0.0706924796,0.1811163574,0.0640624613,0.1197685003,-0.1377031356,-0.1166463196,0.046965342,-0.0323215127,0.2401375026,0.3313703537,0.3093738258,-0.0555984639,-0.2358305752,0.2333134562,0.1667176485,-0.1382888556,-0.06039121,0.072590515,-0.0984998196,-0.1313179582,0.082668893,0.3286071122,0.2563716769,0.2747864425,-0.1755116731,0.1528570056,0.0238279738,0.5519109368,-0.2458668649,0.1138633564,0.3021980822,0.1545313895,-0.2306014448,0.0980398506,0.2813174427,-0.0375363044,0.0506099239,0.0799426436,-0.3858196735,0.4044837654,0.1439911574,-0.1975332648,-0.0476032272,0.1782198101,0.0742536411,-0.0721478388,-0.4519341588,0.1831779182,0.0667779595,-0.1193479002,-0.0543197319,0.4079632163,0.4936104715,-0.4202961922,0.0228122212,-0.221729517,0.019100355,-0.1197147667,0.1713059396,-0.2794617414,0.1356073618,-0.1801660359,-0.2079320103,0.2443795949,-0.3084554672,-0.3225480914,0.4538064003,-0.1916149855,0.2314610779,0.3152425885,0.1994496435,0.2296581119,-0.0338119008,0.1222665012,0.3941685557,-0.2495333552,-0.0643453002,-0.1496422589,-0.1908909678,0.2760188878,0.1059188098,0.2374101728,-0.1076447815,-0.086140126,-0.2940383255,-0.0350805186,-0.083207719,0.2003037333,-0.1944311261,0.0856611282,0.0736428425,0.0493785702,0.1411918253,-0.4604979455,0.3037669957,-0.2521933019,0.0747477561,-0.3391274512,-0.3241582215,0.2313053608,-0.0713331848,-0.1031431407,-0.3441320658,0.0740442052,0.0512403548,-0.1017183214,-0.0374327078,-0.0173988789,0.0242225807,-0.0870222002,0.1135079041,0.1581827104,-0.0672886223,0.0744352862,0.0992200971,-0.2595793903,-0.1118372455,-0.0186913777,-0.1419194341,-0.0033473987,0.507107079,0.0670183226,0.434106648,0.1801059842,0.43449682,0.100204356,0.3320184648,0.2241401672,0.0292811133,0.0667587593,-0.0281432383,-0.5865287185,0.8847305179,-0.0772907659,0.4949632585,-0.0253681988,-0.0896625817,0.2581321895,-0.070058696,-0.5748965144,-0.0896323398,-0.0292923599,0.2472509444,0.1246782467,-0.0695684403,-0.3645498455,0.1309467256,0.0615998544,0.2113251835,0.4444034696,-0.2755915225,0.2542671561,-0.0863385573,0.4632284939,0.2097619325,0.2965559363,0.2528910935,0.0414115787,0.1776760668,-0.0744988695,-0.2386737466,0.3642019629,0.1546070129,0.3154804409,0.1600142121,0.0825260282,-0.1450218856,-0.1656669974,-0.2701164484,0.1441674232,-0.2285314947,-0.052432321,0.3188200295,-0.3138054311,-0.5460381508,-0.0330595821,0.2567465603,-0.1731722355,-0.1265536398,0.232597664,0.1215340123,-0.3687728345,0.3979503214,-0.2202257961,0.4702404141,-0.0649726614,-0.4042862654,0.2372114658,-0.3736272752,0.0975892395,0.0727022961,0.2432580888,-0.090477258,0.6828110218,0.2652270794,-0.0664931238,-0.172499001,-0.2048208117,0.1804469377,0.0453945622,0.5188974738,0.1410698295,0.097885415,-0.1078779474,0.0392978527,0.266864717,-0.131492421,0.2002531737,-0.1014679745,0.0052114641,0.1161197722,-0.2464787811,-0.5637351274,-0.2847306728,-0.2684567571,0.1167265549,-0.2222549319,0.2011254579,-0.0231130943,0.1001847163,-0.050306119,0.1435732543,0.0995229259,-0.0322472528,-0.568808198,0.1139889956,-0.1637168378,0.0997769907,0.0580960028,-0.0110152299,0.3906186521,0.5869692564,-0.4705568254,-0.2902962863,0.3895433247,0.286469698,0.0516619869,-0.3373523057,0.6150392294,0.0365746021,-0.0880156383,-0.0884277374,0.3548750579,0.4860231876,-0.1864490658,-0.0539083518,0.1791349649,0.7724372149,0.202256009,0.2877366543,-0.1666050255,0.4881629646,0.2260150015,0.1102786735,0.304939419,-0.1625722647,0.0179588497,0.2034069151,-0.0815214068,-0.5336046219,0.0716209486,0.0470552854,-0.0527305119,-0.3611359894,-0.2582453787,-0.1032524407,-0.3916658759,0.2961251736,-0.1127902269,0.4995014668,0.117896229,0.372337997,-0.2182418406,0.1661288887,-0.1418527663,0.1501836479,0.5369586945,-0.0128860902,-0.1799632609,0.0348873436,-0.22274369,0.2784568965,0.0099335965,0.3417262137,-0.036391452,0.0174978469,0.0015469281,-0.0763285756,0.4463128746,-0.1146823615,0.0632685199,0.0081591588,-0.2919878364,-0.2112849802,-0.1305243969,-0.0428579152,-0.0211985148,0.0649969876,0.1952529997,-0.2032026201,-0.0390476733,-0.3187787533,0.1351860166,-0.3197744787,-0.2367112786,0.0234858245,0.1267216206,-0.6222805977,0.420473069,-0.1571975201,-0.1905483156,-0.2480874062,-0.0747949705,-0.1783072352,0.4812121689,-0.145950079,0.0565128624,0.1470396221,-0.5123636127,0.1858813465,0.1123727337,0.1500660181,0.3838641942,0.0461331084,0.1363827586,0.284540683,0.0637651309,0.066484876,0.2051711679,0.4261293709,-0.1089276075,-0.0030097992,0.1616743952,0.3030042052,-0.3588068783,-0.3979021609,0.1224461347,0.1558630913,0.0335068256,-0.4153938591,0.0671153069,0.04605335,-0.2596212626,0.0445624702,0.1521447897,-0.3728854656,0.2617143989,0.2630714774,0.9972520471,0.2139621526,0.2179925889,0.4170042574,0.2532133162,0.4726442695,-0.6332621574,0.2983065844,0.1272928864,-0.1670437455,-0.1499634385,-0.02558963,0.2321767211,0.099870272,-0.2685555816,0.5433549285,-0.0701307207,0.1014339775,-0.2433125675,-0.0296450164,-0.2376535684,-0.4036455154,-0.2663605511,0.1260658652,0.2334099859,-0.0104813557,-0.3751176894,-0.068979308,0.1925515532,0.1631508023,0.14545542,0.0721764639,0.0428949408,0.2766603231,-0.419208467,0.0559381731,-0.2763850689,0.1090795174,0.0180981327,0.2108995914,-0.2886296809,0.1259953529,-0.0419275537,0.1140722334,-0.1539172381,-0.2668178678,0.372919023,-0.1662880629,-0.2487163991,-0.2205847353,-0.3582707942,-0.0771528035,0.2514949441,-0.3500300646,0.0939490199,-0.0691653416,0.0619472638,0.1937758923,-0.1183083802,-0.215838626,0.0828801244,0.1098431647,-0.0477760732,0.2642475665,0.0833750889,-0.0296078119,-0.1844729483,0.5224729776,-0.0124881873,0.2983217239,0.329159528,0.17099756,-0.2422726303,-0.3916774988,-0.1179054156,-0.0944124162,0.1713488996,0.0689709634,0.0006624407,-0.0254697464,0.0580743439,-0.2008399367,0.1766274869,-0.0651128218,-0.1013034061,-0.4263277054,-0.0957949758,0.3383998871,0.2018853426,0.0942364931,0.1214297861,0.1285647601,0.0334427096,0.011660737,-0.3435808122,-0.1933899075,-0.2666880488,0.2216063738,-0.0113147693,-0.0477256738,0.1774714738,-0.0150593249,0.0563594587,-0.0292195305,-0.2538238168,-0.1619534642,-0.1136552319,0.2258027345,-0.1483045071,-0.3157637417,0.2160577774,-0.423892051,-0.3664358854,-0.3153072298,0.0918554738,-0.0170711968,-0.1677893996,-0.4335837364,-0.1071414053,0.3017937243,-0.2243884355,0.0949885547,-0.0977727994,-0.2725069225,-0.3345758319,0.0538960434,-0.0734193325,-0.1096151099,-0.0395689383,0.1027320102,0.2106037289,0.205578953,0.4639981687,0.0892977566,-0.1204226613,0.1604589671,0.6960189939,0.1433717161,0.1893894821,0.1112304628,0.0307057649,0.2147427052,-0.4058090746,0.1708363444,-0.2121228129,-0.1652800739,0.2011716664,0.2182677984,0.4211681187,0.1228163764,0.2115367651,-0.0768140182,0.3960933387,-0.3639710844,0.2452284545,-0.0232135504,-0.0485098064,0.1692207754,-0.0433972329,0.1081223264,0.3069615662,0.0682089925,-0.269303441,0.1668248624,0.4557292759,0.2022744119,-0.0908759907,-0.4782039225,-0.1009483114,0.0956822485,-0.0603445247,0.0244362708,-0.464867115,-0.0542507693,-0.3395825922,-0.1913392246,-0.2145864964,0.6027536392,-0.3315451741,-0.1462776661,-0.4204238951,-0.4396710098,0.0110949399,0.0038176409,0.1870268434,-0.138487637,-0.1736916751,0.0796872526,0.2774659395,-0.2712741792,0.0969131291,-0.0292456578,0.1431989819,-0.323631078,0.1587519348,0.5755221844,0.0011660997,0.0627885014,0.2858873308,0.2266620696,0.0022768998,-0.1239224076,0.1276148558,0.1615405679,0.0533404723,-0.3756064475,-0.0559492894,-0.016708402,0.2881965935,0.3220778406,0.1017410532,-0.0451429375,0.503982842,-0.0574037246,-0.3350550532,-0.217862308,-0.0598231368,0.4229668081,0.1130177602,-0.0961265415,-0.0181783196,-0.1123341396,-0.0917096362,0.2774384022,-0.1616678834,0.0388007686,0.0190163963,0.0622579455,-0.1917547733,0.5251255035,-0.0828251243,0.1001403928,-0.3585446477,-0.0513545163,-0.6031898856,-0.2333036214,-0.0481282994,0.050825011,-0.1225795448,0.1338446736,0.2245391905,0.1548327804,-0.2501180172,0.024877632,0.1090888679,0.0784571469,-0.3683554232,0.0319491997,0.0217686146,-0.0349099971,0.1462561339,-0.2422401905,0.2108056098,-0.1415773928,0.0672279671,0.307381779,0.0241460316,0.0134564163,0.1558464766,0.7136694193,0.0963875428,0.3541609049,0.0330818854,-0.0392261669,-0.1355392486,0.0061807684,-0.1747146398,0.0417080708,0.042096138,0.2525860369,-0.1677740067,-0.1356784999,-0.1732335538,0.0797532499,-0.0598759577,-0.006825991,-0.1402485371,-0.1277015358,-0.1939341277,0.1754536778,-0.0835498869,0.2089730799,0.1280008405,0.227099821,-0.2550806999,-0.1245743409,0.4326334,-0.0996425748,0.1882971972,-0.3496024013,0.2039181441,0.229376331,0.170839116,-0.0759380162,-0.0538152941,0.1604664326,-0.0303299837,0.107823357,-0.0209005252,0.0539114326,-0.1801201999,-0.1716643721,0.2066053599,-0.3249644637,-0.3378339112,-0.2800188065,-0.2434223294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1939","title":"[firewalled env] OFFLINE mode","comments":"I lost access to the firewalled setup, but I emulated it with:\r\n\r\n```\r\nsudo ufw enable\r\nsudo ufw default deny outgoing\r\n```\r\n(thanks @mfuntowicz)\r\n\r\nI was able to test `HF_DATASETS_OFFLINE=1` and it worked great - i.e. didn't try to reach out with it and used the cached files instead.\r\n\r\nThank you!","body":"This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.","comment_length":51,"text":"[firewalled env] OFFLINE mode\nThis issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.\r\n\r\nI propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 possible ways to going about it.\r\n\r\n## 1. Manual\r\n\r\nmanually prepare data and metrics files, that is transfer to the firewalled instance the dataset and the metrics and run:\r\n\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --train_file xyz.csv --validation_file xyz.csv ...\r\n```\r\n\r\n`datasets` must not make any network calls and if there is a logic to do that and something is missing it should assert that this or that action requires network and therefore it can't proceed.\r\n\r\n## 2. Automatic\r\n\r\nIn some clouds one can prepare a datastorage ahead of time with a normal networked environment but which doesn't have gpus and then one switches to the gpu instance which is firewalled, but it can access all the cached data. This is the ideal situation, since in this scenario we don't have to do anything manually, but simply run the same application twice:\r\n\r\n1. on the non-firewalled instance:\r\n```\r\nrun_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nwhich should download and cached everything.\r\n\r\n2. and then immediately after on the firewalled instance, which shares the same filesystem\r\n```\r\nDATASETS_OFFLINE=1 run_seq2seq.py --dataset_name wmt16 --dataset_config ro-en ...\r\n```\r\n\r\nand the metrics and datasets should be cached by the invocation number 1 and any network calls be skipped and if the logic is missing data it should assert and not try to fetch any data from online.\r\n\r\n## Common Issues\r\n\r\n1. for example currently `datasets` tries to look up online datasets if the files contain json or csv, despite the paths already provided\r\n\r\n```\r\n if dataset and path in _PACKAGED_DATASETS_MODULES:\r\n```\r\n\r\n2. it has an issue with metrics. e.g. I had to manually copy `rouge\/rouge.py` from the `datasets` repo to the current dir - or it was hanging.\r\n\r\nI had to comment out `head_hf_s3(...)` calls to make things work. So all those `try: head_hf_s3(...)` shouldn't be tried with `DATASETS_OFFLINE=1`\r\n\r\nHere is the corresponding issue for `transformers`: https:\/\/github.com\/huggingface\/transformers\/issues\/10379\r\n\r\nThanks.\nI lost access to the firewalled setup, but I emulated it with:\r\n\r\n```\r\nsudo ufw enable\r\nsudo ufw default deny outgoing\r\n```\r\n(thanks @mfuntowicz)\r\n\r\nI was able to test `HF_DATASETS_OFFLINE=1` and it worked great - i.e. didn't try to reach out with it and used the cached files instead.\r\n\r\nThank you!","embeddings":[-0.4748800695,0.1685334891,-0.0625536442,0.1050562486,0.0443113707,-0.2284921706,0.4647973776,0.0716796964,0.2545911074,0.0485309139,0.0184395313,0.0636148751,0.0845065862,0.1993038058,0.0180914346,0.1079676971,0.0199452993,-0.1096389145,-0.2528404295,0.0637163967,-0.2635913491,-0.0536728576,-0.2544517219,-0.2575809062,0.078367807,-0.5080999136,-0.2168778479,-0.0105694095,0.0096480669,-0.2275560498,0.3089389205,0.2612626255,0.2518863976,0.1530115753,-0.0001084152,-0.0537566915,-0.1393844485,-0.0361674428,-0.272887677,0.0223880373,-0.3179339468,-0.3802416921,0.1650776267,-0.4613099992,-0.171355173,-0.174725309,0.1464560628,-0.6594020128,0.5154665709,-0.0763156116,0.0945165008,0.0003517794,-0.3278796375,-0.0994149074,-0.3226190805,-0.1980506182,-0.1814403385,0.1710348725,0.0883782059,-0.2174991071,-0.0455865599,0.3952884078,-0.3122881055,0.1565717608,0.3260605037,0.0182255022,-0.338113457,-0.1188222617,-0.1281385869,0.3376224041,0.1812187731,-0.2216326147,-0.4262379706,-0.4555667341,-0.0790792406,-0.370331347,0.3025760651,0.0889306292,-0.1169174388,0.2809565067,-0.3825724125,-0.2138668299,-0.1247217432,-0.0012391666,-0.0308274534,0.3738206923,0.0736688673,0.1736761779,-0.0692528114,0.1060303599,0.0134036681,-0.3340767622,0.2569811642,0.0270416345,-0.5665541291,0.0265186541,0.1224480569,-0.0067842589,0.0168159846,0.4544952512,-0.1779251099,-0.0429254211,0.4983155131,0.195036903,0.0398950391,0.2975893617,0.1519620121,-0.1790611893,0.3776472807,0.1602634937,-0.0568716712,-0.1345084161,-0.1136473343,-0.0706924796,0.1811163574,0.0640624613,0.1197685003,-0.1377031356,-0.1166463196,0.046965342,-0.0323215127,0.2401375026,0.3313703537,0.3093738258,-0.0555984639,-0.2358305752,0.2333134562,0.1667176485,-0.1382888556,-0.06039121,0.072590515,-0.0984998196,-0.1313179582,0.082668893,0.3286071122,0.2563716769,0.2747864425,-0.1755116731,0.1528570056,0.0238279738,0.5519109368,-0.2458668649,0.1138633564,0.3021980822,0.1545313895,-0.2306014448,0.0980398506,0.2813174427,-0.0375363044,0.0506099239,0.0799426436,-0.3858196735,0.4044837654,0.1439911574,-0.1975332648,-0.0476032272,0.1782198101,0.0742536411,-0.0721478388,-0.4519341588,0.1831779182,0.0667779595,-0.1193479002,-0.0543197319,0.4079632163,0.4936104715,-0.4202961922,0.0228122212,-0.221729517,0.019100355,-0.1197147667,0.1713059396,-0.2794617414,0.1356073618,-0.1801660359,-0.2079320103,0.2443795949,-0.3084554672,-0.3225480914,0.4538064003,-0.1916149855,0.2314610779,0.3152425885,0.1994496435,0.2296581119,-0.0338119008,0.1222665012,0.3941685557,-0.2495333552,-0.0643453002,-0.1496422589,-0.1908909678,0.2760188878,0.1059188098,0.2374101728,-0.1076447815,-0.086140126,-0.2940383255,-0.0350805186,-0.083207719,0.2003037333,-0.1944311261,0.0856611282,0.0736428425,0.0493785702,0.1411918253,-0.4604979455,0.3037669957,-0.2521933019,0.0747477561,-0.3391274512,-0.3241582215,0.2313053608,-0.0713331848,-0.1031431407,-0.3441320658,0.0740442052,0.0512403548,-0.1017183214,-0.0374327078,-0.0173988789,0.0242225807,-0.0870222002,0.1135079041,0.1581827104,-0.0672886223,0.0744352862,0.0992200971,-0.2595793903,-0.1118372455,-0.0186913777,-0.1419194341,-0.0033473987,0.507107079,0.0670183226,0.434106648,0.1801059842,0.43449682,0.100204356,0.3320184648,0.2241401672,0.0292811133,0.0667587593,-0.0281432383,-0.5865287185,0.8847305179,-0.0772907659,0.4949632585,-0.0253681988,-0.0896625817,0.2581321895,-0.070058696,-0.5748965144,-0.0896323398,-0.0292923599,0.2472509444,0.1246782467,-0.0695684403,-0.3645498455,0.1309467256,0.0615998544,0.2113251835,0.4444034696,-0.2755915225,0.2542671561,-0.0863385573,0.4632284939,0.2097619325,0.2965559363,0.2528910935,0.0414115787,0.1776760668,-0.0744988695,-0.2386737466,0.3642019629,0.1546070129,0.3154804409,0.1600142121,0.0825260282,-0.1450218856,-0.1656669974,-0.2701164484,0.1441674232,-0.2285314947,-0.052432321,0.3188200295,-0.3138054311,-0.5460381508,-0.0330595821,0.2567465603,-0.1731722355,-0.1265536398,0.232597664,0.1215340123,-0.3687728345,0.3979503214,-0.2202257961,0.4702404141,-0.0649726614,-0.4042862654,0.2372114658,-0.3736272752,0.0975892395,0.0727022961,0.2432580888,-0.090477258,0.6828110218,0.2652270794,-0.0664931238,-0.172499001,-0.2048208117,0.1804469377,0.0453945622,0.5188974738,0.1410698295,0.097885415,-0.1078779474,0.0392978527,0.266864717,-0.131492421,0.2002531737,-0.1014679745,0.0052114641,0.1161197722,-0.2464787811,-0.5637351274,-0.2847306728,-0.2684567571,0.1167265549,-0.2222549319,0.2011254579,-0.0231130943,0.1001847163,-0.050306119,0.1435732543,0.0995229259,-0.0322472528,-0.568808198,0.1139889956,-0.1637168378,0.0997769907,0.0580960028,-0.0110152299,0.3906186521,0.5869692564,-0.4705568254,-0.2902962863,0.3895433247,0.286469698,0.0516619869,-0.3373523057,0.6150392294,0.0365746021,-0.0880156383,-0.0884277374,0.3548750579,0.4860231876,-0.1864490658,-0.0539083518,0.1791349649,0.7724372149,0.202256009,0.2877366543,-0.1666050255,0.4881629646,0.2260150015,0.1102786735,0.304939419,-0.1625722647,0.0179588497,0.2034069151,-0.0815214068,-0.5336046219,0.0716209486,0.0470552854,-0.0527305119,-0.3611359894,-0.2582453787,-0.1032524407,-0.3916658759,0.2961251736,-0.1127902269,0.4995014668,0.117896229,0.372337997,-0.2182418406,0.1661288887,-0.1418527663,0.1501836479,0.5369586945,-0.0128860902,-0.1799632609,0.0348873436,-0.22274369,0.2784568965,0.0099335965,0.3417262137,-0.036391452,0.0174978469,0.0015469281,-0.0763285756,0.4463128746,-0.1146823615,0.0632685199,0.0081591588,-0.2919878364,-0.2112849802,-0.1305243969,-0.0428579152,-0.0211985148,0.0649969876,0.1952529997,-0.2032026201,-0.0390476733,-0.3187787533,0.1351860166,-0.3197744787,-0.2367112786,0.0234858245,0.1267216206,-0.6222805977,0.420473069,-0.1571975201,-0.1905483156,-0.2480874062,-0.0747949705,-0.1783072352,0.4812121689,-0.145950079,0.0565128624,0.1470396221,-0.5123636127,0.1858813465,0.1123727337,0.1500660181,0.3838641942,0.0461331084,0.1363827586,0.284540683,0.0637651309,0.066484876,0.2051711679,0.4261293709,-0.1089276075,-0.0030097992,0.1616743952,0.3030042052,-0.3588068783,-0.3979021609,0.1224461347,0.1558630913,0.0335068256,-0.4153938591,0.0671153069,0.04605335,-0.2596212626,0.0445624702,0.1521447897,-0.3728854656,0.2617143989,0.2630714774,0.9972520471,0.2139621526,0.2179925889,0.4170042574,0.2532133162,0.4726442695,-0.6332621574,0.2983065844,0.1272928864,-0.1670437455,-0.1499634385,-0.02558963,0.2321767211,0.099870272,-0.2685555816,0.5433549285,-0.0701307207,0.1014339775,-0.2433125675,-0.0296450164,-0.2376535684,-0.4036455154,-0.2663605511,0.1260658652,0.2334099859,-0.0104813557,-0.3751176894,-0.068979308,0.1925515532,0.1631508023,0.14545542,0.0721764639,0.0428949408,0.2766603231,-0.419208467,0.0559381731,-0.2763850689,0.1090795174,0.0180981327,0.2108995914,-0.2886296809,0.1259953529,-0.0419275537,0.1140722334,-0.1539172381,-0.2668178678,0.372919023,-0.1662880629,-0.2487163991,-0.2205847353,-0.3582707942,-0.0771528035,0.2514949441,-0.3500300646,0.0939490199,-0.0691653416,0.0619472638,0.1937758923,-0.1183083802,-0.215838626,0.0828801244,0.1098431647,-0.0477760732,0.2642475665,0.0833750889,-0.0296078119,-0.1844729483,0.5224729776,-0.0124881873,0.2983217239,0.329159528,0.17099756,-0.2422726303,-0.3916774988,-0.1179054156,-0.0944124162,0.1713488996,0.0689709634,0.0006624407,-0.0254697464,0.0580743439,-0.2008399367,0.1766274869,-0.0651128218,-0.1013034061,-0.4263277054,-0.0957949758,0.3383998871,0.2018853426,0.0942364931,0.1214297861,0.1285647601,0.0334427096,0.011660737,-0.3435808122,-0.1933899075,-0.2666880488,0.2216063738,-0.0113147693,-0.0477256738,0.1774714738,-0.0150593249,0.0563594587,-0.0292195305,-0.2538238168,-0.1619534642,-0.1136552319,0.2258027345,-0.1483045071,-0.3157637417,0.2160577774,-0.423892051,-0.3664358854,-0.3153072298,0.0918554738,-0.0170711968,-0.1677893996,-0.4335837364,-0.1071414053,0.3017937243,-0.2243884355,0.0949885547,-0.0977727994,-0.2725069225,-0.3345758319,0.0538960434,-0.0734193325,-0.1096151099,-0.0395689383,0.1027320102,0.2106037289,0.205578953,0.4639981687,0.0892977566,-0.1204226613,0.1604589671,0.6960189939,0.1433717161,0.1893894821,0.1112304628,0.0307057649,0.2147427052,-0.4058090746,0.1708363444,-0.2121228129,-0.1652800739,0.2011716664,0.2182677984,0.4211681187,0.1228163764,0.2115367651,-0.0768140182,0.3960933387,-0.3639710844,0.2452284545,-0.0232135504,-0.0485098064,0.1692207754,-0.0433972329,0.1081223264,0.3069615662,0.0682089925,-0.269303441,0.1668248624,0.4557292759,0.2022744119,-0.0908759907,-0.4782039225,-0.1009483114,0.0956822485,-0.0603445247,0.0244362708,-0.464867115,-0.0542507693,-0.3395825922,-0.1913392246,-0.2145864964,0.6027536392,-0.3315451741,-0.1462776661,-0.4204238951,-0.4396710098,0.0110949399,0.0038176409,0.1870268434,-0.138487637,-0.1736916751,0.0796872526,0.2774659395,-0.2712741792,0.0969131291,-0.0292456578,0.1431989819,-0.323631078,0.1587519348,0.5755221844,0.0011660997,0.0627885014,0.2858873308,0.2266620696,0.0022768998,-0.1239224076,0.1276148558,0.1615405679,0.0533404723,-0.3756064475,-0.0559492894,-0.016708402,0.2881965935,0.3220778406,0.1017410532,-0.0451429375,0.503982842,-0.0574037246,-0.3350550532,-0.217862308,-0.0598231368,0.4229668081,0.1130177602,-0.0961265415,-0.0181783196,-0.1123341396,-0.0917096362,0.2774384022,-0.1616678834,0.0388007686,0.0190163963,0.0622579455,-0.1917547733,0.5251255035,-0.0828251243,0.1001403928,-0.3585446477,-0.0513545163,-0.6031898856,-0.2333036214,-0.0481282994,0.050825011,-0.1225795448,0.1338446736,0.2245391905,0.1548327804,-0.2501180172,0.024877632,0.1090888679,0.0784571469,-0.3683554232,0.0319491997,0.0217686146,-0.0349099971,0.1462561339,-0.2422401905,0.2108056098,-0.1415773928,0.0672279671,0.307381779,0.0241460316,0.0134564163,0.1558464766,0.7136694193,0.0963875428,0.3541609049,0.0330818854,-0.0392261669,-0.1355392486,0.0061807684,-0.1747146398,0.0417080708,0.042096138,0.2525860369,-0.1677740067,-0.1356784999,-0.1732335538,0.0797532499,-0.0598759577,-0.006825991,-0.1402485371,-0.1277015358,-0.1939341277,0.1754536778,-0.0835498869,0.2089730799,0.1280008405,0.227099821,-0.2550806999,-0.1245743409,0.4326334,-0.0996425748,0.1882971972,-0.3496024013,0.2039181441,0.229376331,0.170839116,-0.0759380162,-0.0538152941,0.1604664326,-0.0303299837,0.107823357,-0.0209005252,0.0539114326,-0.1801201999,-0.1716643721,0.2066053599,-0.3249644637,-0.3378339112,-0.2800188065,-0.2434223294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1937","title":"CommonGen dataset page shows an error OSError: [Errno 28] No space left on device","comments":"Facing the same issue for [Squad](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=squad) and [TriviaQA](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=trivia_qa) datasets as well.","body":"The page of the CommonGen data https:\/\/huggingface.co\/datasets\/viewer\/?dataset=common_gen shows \r\n![image](https:\/\/user-images.githubusercontent.com\/10104354\/108959311-1865e600-7629-11eb-868c-cf4cb27034ea.png)\r\n","comment_length":11,"text":"CommonGen dataset page shows an error OSError: [Errno 28] No space left on device\nThe page of the CommonGen data https:\/\/huggingface.co\/datasets\/viewer\/?dataset=common_gen shows \r\n![image](https:\/\/user-images.githubusercontent.com\/10104354\/108959311-1865e600-7629-11eb-868c-cf4cb27034ea.png)\r\n\nFacing the same issue for [Squad](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=squad) and [TriviaQA](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=trivia_qa) datasets as well.","embeddings":[-0.0371066444,-0.7277044058,-0.0643555894,0.3471886814,0.3455493748,0.2450598776,0.139766857,0.1919169575,0.0533625148,0.4051840901,-0.1042112336,-0.1051004604,0.1274256557,-0.0246365536,0.0325752497,-0.0515596867,-0.072950013,0.2982448637,-0.0637304187,0.0033932624,-0.0307162162,0.2042712867,-0.18158333,-0.1119373366,-0.175521493,-0.1310201585,-0.3065968752,0.0901676118,0.0566331707,-0.4230850339,0.293720603,-0.1226030812,0.0345175005,0.3597543836,-0.0001193126,-0.0482880883,0.1198864281,-0.0549285114,-0.1158830151,0.0613365695,0.044759918,-0.1279976815,0.0495730229,0.169330731,-0.003603807,-0.1158716306,0.1184552833,-0.3641517162,0.2921280265,0.1602797508,0.1618177295,0.1443664432,0.367898643,-0.0370014608,0.2100429684,0.1013436764,-0.1765626818,0.3184151351,0.4184165895,0.148171708,0.1798471659,0.3926923275,0.3360721767,-0.1393052042,0.2504782677,0.0370891802,0.3939348161,-0.525146544,0.3715389073,0.1761273891,0.2438695133,-0.2753061354,-0.0624825843,-0.3469039798,0.1114153937,0.1203087419,0.2995107472,0.370174855,-0.2872421741,0.0067794756,-0.2296364903,-0.2950042486,-0.0958231091,0.1173552349,-0.0825959519,0.0128180599,-0.0849807784,0.0768427476,0.245437935,-0.115148291,-0.1379693896,-0.0187409595,0.097971946,0.114309825,-0.2908746898,-0.169340089,-0.2272607386,-0.1641017795,0.2925755978,-0.4004248381,-0.0890739635,-0.4012104571,-0.0025876288,0.0538854748,0.3853557706,0.0996643454,0.0271330718,-0.0880583674,-0.0941145346,0.2242338359,0.2225245684,-0.0876265168,-0.1700957417,-0.1901226044,-0.0621909983,-0.4690943658,0.146588102,-0.1762777865,-0.4803974628,0.3524149656,0.234287262,0.2944587469,0.1850014031,0.5516119003,0.2357548028,0.15434286,-0.0067399899,0.1066055074,-0.0662721023,-0.2350685149,-0.1104255617,0.1501301527,-0.1873968542,0.0531770587,0.1386820972,-0.2763881981,0.1142680496,0.0828535184,0.2100125253,-0.0877215639,0.0473091602,-0.3139556944,-0.0065680505,0.248085171,0.2312964499,0.0970281065,0.209556818,-0.0855550468,-0.2394898385,0.2608163357,-0.2484851778,-0.1314193755,-0.3544407487,0.1019268557,-0.0186035112,0.0870704949,0.1587355137,-0.1992030293,0.5694010854,-0.3311134279,0.1568794549,-0.0647392571,-0.2909047008,0.0825877786,-0.0044867112,-0.0026521748,-0.1231050938,-0.1305025965,0.0479223616,0.1483278126,0.0210197195,0.2523830235,0.2228738368,-0.1959376037,-0.3356046379,0.2198466212,0.1766916066,-0.323045224,-0.3951907158,0.0682959184,-0.1094960496,-0.1526805013,0.173490718,-0.1320239156,-0.0147165703,0.1678975523,0.279179126,0.1184750795,-0.0434480906,0.169941321,-0.3568478525,-0.3826210797,-0.0614122078,0.2171603739,0.0287755299,-0.365433991,0.0422571488,0.0351260006,0.2866052985,-0.2608788908,0.0437148027,0.4189171493,0.6117413044,-0.1796193719,-0.0464769341,0.0245839655,-0.6670953631,0.0378405601,-0.1079499274,-0.0223946385,0.1572635323,-0.2308731377,-0.1573639363,-0.0100382715,-0.3417517543,-0.2438987195,0.0361120328,0.0682755113,0.2450990826,-0.1359256655,-0.0491704009,-0.1612724066,-0.1872478575,0.281511277,-0.4758988917,0.7208092213,-0.3970691562,0.0945934355,0.0610737838,0.0526206121,-0.0417416096,-0.0763759539,-0.0681196973,0.3936880231,0.174215734,0.2329404354,0.2937013805,0.0233935528,0.3391291201,-0.4047809541,0.0756145343,0.1275743395,-0.1813566983,0.2718836665,-0.5229043961,-0.0259506796,-0.0620607734,0.2886085808,0.0279309489,0.1983329058,0.2086573988,0.1064177826,0.4015279114,-0.3711340129,0.1167173907,-0.1467512101,0.3884682357,0.2455463707,-0.0208014026,0.2122652382,0.3905693591,0.3606899083,0.1135032624,0.2973338664,-0.172375083,-0.3511391878,0.4975039363,0.1439281255,0.1765219867,0.2251676023,-0.1036506519,0.1496193409,0.3822692633,-0.0547608398,0.1919005513,-0.0316728763,0.320048511,0.1579496861,0.1068802774,0.1632661968,-0.3695676327,-0.2622111738,0.1825890839,0.28148067,-0.1545584798,-0.2450593263,0.0649170205,-0.0525409356,0.1261697859,0.4955257177,-0.5756959915,-0.2728329003,0.3711250126,0.118715547,0.1207982451,0.1140314341,-0.2104639113,0.2808659077,0.0590332225,0.2155060768,-0.1803479642,0.1253486425,-0.2068506777,0.0136887515,0.2389111668,-0.2162069976,0.3188436627,-0.2209903449,0.0970996618,-0.2774519622,-0.4910782576,0.2934611738,-0.0228376146,-0.0940466821,0.1393589526,0.5265669227,-0.3623134196,0.060286954,0.1641134769,-0.2459428161,-0.2475943714,-0.0911050588,-0.051447928,-0.0270084348,-0.3853808045,-0.1374876648,-0.3055696785,-0.3278987408,0.1232673675,-0.3155274689,0.1697466522,0.3462600708,0.0898694694,-0.0501502194,-0.1388591975,0.27947855,-0.306773454,-0.2540832162,-0.2072079033,-0.1608607918,-0.0706543624,0.183448568,0.3263358474,0.386320889,0.0463383757,-0.8021862507,-0.3492587805,-0.3515963256,0.1841829419,-0.1472003013,-0.1147406399,0.3189388812,0.0757250413,-0.0935723484,-0.0673306584,-0.0475024953,-0.3464223146,-0.0614788085,0.2954241931,-0.2584116161,0.3204969168,0.3509953618,0.5191604495,0.3317056894,-0.1902632415,0.2993983924,-0.2020637244,0.6267129183,-0.3736895323,0.0398561992,0.3594235182,0.3485282063,-0.0310899075,0.5342381001,0.27736485,0.1674593389,0.0970407203,0.032599818,-0.1190154329,-0.1070726737,-0.0480284281,0.2431617528,0.2638537586,-0.0135911815,-0.0675044209,0.0906744227,-0.4112641811,-0.2730492353,0.2119695991,0.0500023589,0.0499006696,-0.274992764,-0.1753108799,-0.3815764487,0.3775395453,0.0121627441,0.2768786848,-0.5317758322,-0.00840287,0.2087222487,0.068630293,0.3978365958,-0.0990408808,-0.1231185943,-0.3036735058,-0.0225430634,-0.4008125067,0.06158638,0.1457875818,0.1954033524,0.565143168,0.5005791783,-0.2739661634,-0.3184965849,-0.0208426137,0.0984144285,-0.0301294681,-0.1245303601,-0.1285288036,-0.0452893488,-0.5294141769,-0.0589871332,-0.1294941604,0.1528802067,-0.1532124579,-0.1278877705,0.3017373383,-0.1895927191,0.0397365205,0.1125238985,0.4392372072,0.1025977209,0.4039860666,0.1776151806,0.272878617,0.4640432596,0.4589373171,-0.2846638262,-0.1919920146,0.4358006716,-0.0338138193,0.1028953791,0.3194747269,0.0328035839,-0.21644032,-0.0177085195,0.5839995742,-0.0922106579,0.3101250231,0.1745637804,-0.0130418325,-0.2021275908,-0.0410112664,-0.0012959088,0.2255742699,0.0419716574,0.2043476105,0.3536215127,-0.2969720066,0.1162735894,-0.1133963689,0.9039855599,-0.201163426,0.2220166773,0.1584686339,0.123953335,0.3643602431,-0.0652857274,0.0143586332,-0.3139036596,-0.3527930677,0.1156968027,-0.0069093453,0.0795857608,-0.0469703153,-0.1966790408,0.1597642154,0.0366235897,0.1755373925,-0.0211653989,0.1101985648,-0.1688172966,-0.2013396472,-0.3453652859,0.0735255778,-0.1863065511,0.2632056475,-0.1176259071,0.1856416464,-0.1524582207,-0.148345679,-0.139524132,0.07418859,0.1982402951,0.2659704089,0.126055643,-0.2828586698,0.084938094,0.3843862712,0.195562765,0.249803558,-0.4358243048,0.4141407907,-0.0292463638,-0.1070406586,0.0340835229,0.1718976796,-0.150050506,-0.14825508,-0.0499440394,-0.0334390402,-0.1028088108,-0.0840351284,-0.4245301485,-0.1161950231,-0.0475748554,-0.1756401807,-0.0361596197,-0.5126248598,-0.1878898442,-0.1316136867,0.0906926617,0.0318608955,-0.0783230066,-0.2349028587,0.3620882034,-0.2038057745,-0.0448439755,0.1393103153,0.2349600345,-0.2556627393,0.5390514135,-0.0460060462,-0.0740157142,-0.281552732,0.200145632,0.1889965087,-0.6781333089,-0.0435289405,0.0106717832,-0.0268014669,-0.027638156,0.0429416671,0.2614259124,-0.0696532205,0.2603668571,-0.3286043704,-0.2995316386,0.1748076826,-0.3574552238,0.4751408994,0.1332615018,-0.1025138795,0.0857901871,-0.1413593292,-0.2407751828,0.132039547,-0.2083179653,0.116452761,-0.2125300169,-0.1951358467,0.0263259448,-0.3980256617,-0.0052177911,0.1243356466,-0.0006490008,-0.147717759,0.1771693677,0.1289777905,0.1492973715,0.0298653524,-0.1375554651,0.1317008436,-0.1633868665,-0.392182827,0.1876769811,0.090001151,-0.1158590764,-0.3012962639,0.0743189156,0.071962133,-0.0746045038,0.0592930987,0.1316411495,0.1472416222,0.2144149244,0.3704710305,-0.3433310986,0.0871322528,-0.0165975019,0.2220983952,0.4402986169,-0.2097571641,0.0691564605,-0.2712257802,0.0483521223,0.1087801084,0.3244306743,0.0952875316,-0.5820509791,0.2211446464,-0.0395045206,0.1410136372,-0.1285522878,0.0701118857,0.2822138071,0.1231793389,-0.1515432596,0.1842641085,0.3540320992,-0.2795535028,0.2841908336,0.1916446984,0.27205953,0.0047683464,0.3676289916,0.1329089254,-0.0387973376,-0.0164120216,0.1046910286,-0.0017849412,0.0821173117,0.2164423615,-0.5631693006,0.2151815295,0.0654265732,-0.1795391738,0.290854454,-0.5052908063,-0.1648344398,0.2075485587,-0.0208177846,0.0276390314,0.1378442496,0.4407661557,-0.009386518,-0.4788568318,-0.3848580122,0.1586794108,-0.284851253,0.2277045399,0.1372904629,-0.1071314737,-0.2579507828,0.0878830776,0.0667374358,-0.5059937239,0.2082117498,0.0070994943,-0.4344204068,-0.2809408605,-0.0341315195,0.6781238317,-0.1428991556,-0.005550494,0.4498954117,0.0574923977,-0.1085301638,0.0507366769,0.2701074183,0.5949308872,0.002575695,0.1320284307,0.1032331735,0.1823693365,0.0651025698,-0.2419361323,-0.0862727165,0.1838927269,0.2155934125,0.251401186,0.1048664227,-0.0838915259,-0.0803294778,0.0020152477,0.3839266896,-0.1301589161,-0.3313196301,-0.5470066667,-0.1683015078,-0.0352371968,-0.1721921712,-0.5586609244,-0.051042743,0.1829718947,0.023752939,0.0760921389,-0.1897554994,0.0208240692,-0.041124437,0.0363576226,0.3585242629,-0.0946613997,0.0097645298,-0.589410305,-0.6001143456,0.2375193387,-0.18171601,-0.1425997913,0.1211403832,0.1706345081,-0.1372907162,-0.2440579385,0.2767210603,-0.0947912261,0.013074046,-0.062381655,-0.3302640617,-0.0968077183,0.0304697715,-0.0460107774,0.0725487173,-0.1769266278,0.1896629184,-0.0901327878,-0.1527957767,-0.0880704448,0.2800997496,-0.4749968946,-0.1960314363,0.2378982902,0.0480385609,0.6161639094,0.0735350996,-0.2006046325,-0.0231962185,-0.428435564,0.0221516229,0.1909897923,0.0454263724,0.3342825174,-0.2008292228,-0.3604723215,-0.0379148871,0.1565743536,0.1454167515,-0.1614546776,0.0555581041,0.1071973965,-0.118876718,0.2528007925,0.3854166865,0.0818922222,0.0732249022,-0.1175346375,-0.2002107799,-0.3098943532,0.448127985,0.0379524231,-0.1425083727,-0.2008562982,0.2241461873,0.1914752275,-0.3918489814,-0.4388789237,-0.0631639436,0.2255984396,0.1159111857,-0.1574043185,-0.1273723841,-0.2146362811,-0.1072386503,0.052181311,0.7352699637,0.0764378309,-0.4514698088,0.3330565691,-0.1091022789]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1937","title":"CommonGen dataset page shows an error OSError: [Errno 28] No space left on device","comments":"We just fixed the issue, thanks for reporting !","body":"The page of the CommonGen data https:\/\/huggingface.co\/datasets\/viewer\/?dataset=common_gen shows \r\n![image](https:\/\/user-images.githubusercontent.com\/10104354\/108959311-1865e600-7629-11eb-868c-cf4cb27034ea.png)\r\n","comment_length":9,"text":"CommonGen dataset page shows an error OSError: [Errno 28] No space left on device\nThe page of the CommonGen data https:\/\/huggingface.co\/datasets\/viewer\/?dataset=common_gen shows \r\n![image](https:\/\/user-images.githubusercontent.com\/10104354\/108959311-1865e600-7629-11eb-868c-cf4cb27034ea.png)\r\n\nWe just fixed the issue, thanks for reporting !","embeddings":[-0.2168094665,-0.8186897635,-0.1532910168,0.2456236631,0.3866601288,0.1788641959,0.1706050932,0.2139260173,0.073083736,0.439462781,-0.066067189,-0.0802875683,0.0973653942,0.1756025553,0.0391897745,-0.1075841263,-0.0624366663,0.346965611,-0.1179798171,0.0243874863,-0.0497391894,0.2251305133,-0.1368663013,-0.0512149781,-0.1058038697,-0.1257779747,-0.1990371943,0.1514787972,0.0068648411,-0.330686748,0.2843216956,-0.1884208173,-0.0609633215,0.4002468884,-0.0001118768,0.0302232839,0.1039313525,-0.0066811689,-0.0275118612,0.0394476764,0.0769780949,-0.1956631392,-0.0480154529,0.1287626028,-0.1173984036,-0.2421863228,0.0567002222,-0.2010760456,0.2938131392,0.0938491225,0.2730531693,-0.0214489251,0.4035378397,0.072398372,0.1780162752,0.0686719418,-0.1340936571,0.2636840045,0.3082714677,0.1371113062,0.2478962839,0.4572532475,0.2707586884,-0.1396931708,0.2640158534,0.0744826868,0.5146688223,-0.4848741591,0.3274132907,0.0791771933,0.3050457239,-0.2980371416,0.0325063691,-0.2302158922,0.0021827777,-0.0511022806,0.3623480499,0.2909736335,-0.2620467842,0.0783496127,-0.2530494034,-0.4153077602,-0.0999582335,0.040215835,-0.1354917139,0.0200567432,-0.0783354267,0.0153764794,0.2521363497,-0.0809207335,-0.1139356196,0.1157206967,-0.0148962298,0.1424235106,-0.2676033676,-0.2082410604,-0.2093379349,-0.082198225,0.2657619715,-0.3279723823,-0.1683755666,-0.3186999261,-0.0037043015,-0.0575544648,0.2942853868,0.0819398835,-0.1889507473,-0.1144847721,-0.0584825538,0.2389303148,0.3092429042,-0.0688241646,-0.1049433202,-0.2501226962,-0.0378783904,-0.4504525959,0.2047510892,-0.2397476733,-0.3501015902,0.319768846,0.3335659504,0.2233361751,0.1373767108,0.5658308268,0.1513384879,0.1929547936,0.1296551228,-0.0239322949,-0.0385579802,-0.1278088838,-0.145155713,0.1616987139,-0.2116710991,0.0019018889,-0.0357650183,-0.1972795427,0.120044753,0.1282221079,0.1675605327,0.038731195,0.0484819077,-0.1823068261,-0.1014708504,0.3256656528,0.1822958887,0.0596916825,0.1577034444,-0.0681932494,-0.1980290711,0.1949933171,-0.3234319091,-0.1802427322,-0.3160531819,0.2118639499,0.0624634475,0.1389618218,0.1652616113,-0.1659286618,0.5152139068,-0.3311287463,0.1388339847,0.020451583,-0.255673945,0.0243679564,0.0723622441,-0.1889380962,-0.1001608223,-0.1706874818,0.0573193058,0.0097698951,0.1232379973,0.29534477,0.2312224656,-0.1675324738,-0.351223439,0.2567287982,0.1999067068,-0.1285005212,-0.5261874199,0.1373963505,-0.2090214342,-0.2687667608,0.1637566537,-0.1139360443,-0.0714995638,0.1232949272,0.3362640142,0.0992944315,-0.0224724486,0.2612015605,-0.4623736739,-0.4037704766,-0.0513680689,0.1726771891,0.0844555795,-0.3269016445,0.082069777,-0.0063521201,0.3844094276,-0.2927235067,0.0381006747,0.458779335,0.6576005816,-0.1885106415,0.0194963366,-0.0017562057,-0.5453286171,-0.003799327,-0.0872013792,-0.0993222967,0.2304789871,-0.2361997366,-0.1658479869,-0.0461479425,-0.2936156094,-0.1987378746,0.1366499215,-0.0089549143,0.09781757,-0.1804184467,-0.0967651829,-0.1255838871,-0.1510161161,0.2704274356,-0.4222493768,0.7886171341,-0.4120775163,0.0446342863,0.1277244538,-0.0371336602,-0.0322193988,-0.1226743534,-0.0794267356,0.255905509,0.1182090715,0.2138890177,0.1718530655,0.0180904306,0.2771151364,-0.3981200159,0.0602721982,0.0597248375,-0.1296058297,0.2829613686,-0.3869802058,-0.0598495416,-0.041120261,0.1861249208,-0.0420927331,0.1310271919,0.2394969463,0.0098825824,0.4432578683,-0.26363343,0.2209330797,-0.1817786694,0.2593132555,0.1438275576,-0.1194179282,0.2541792691,0.3288524747,0.3404072523,0.0491895787,0.287561357,-0.1915466338,-0.3186422884,0.4535158277,0.1556653678,0.1878195554,0.2691259682,-0.0731743127,0.2652206719,0.3331106305,-0.0349787176,0.2424940914,-0.121585913,0.3249031901,0.2223101109,0.0059734876,0.0886971056,-0.4568737447,-0.2299518287,0.2014754117,0.3793136477,-0.1668829024,-0.2501884997,0.0446698554,0.009437575,0.2339250147,0.3419100046,-0.4598270953,-0.1964195371,0.330622524,0.1868198812,0.1359382719,0.1307810396,-0.1216628328,0.2421572059,0.1062381938,0.3103389442,-0.2865147889,0.112432532,-0.1709619462,0.1367435008,0.1949550509,-0.1796920002,0.402602911,-0.2577437162,-0.048684001,-0.2564508915,-0.4592579901,0.2617638707,-0.1141461805,-0.0526007898,0.211114049,0.5055345893,-0.2242104709,0.1444621235,0.0678856298,-0.2310438603,-0.331895858,-0.1331077069,-0.0816811398,0.0240117982,-0.4078484774,-0.1643335819,-0.2516196072,-0.2854740322,0.1430790871,-0.2530637681,0.1767539233,0.1875883043,0.1555704027,-0.076367639,-0.0732975006,0.1861123294,-0.412101239,-0.2778962553,-0.2274161577,-0.0942137167,-0.1793845892,0.1581914723,0.255726546,0.4004504979,0.0170912836,-0.8478398919,-0.3248330951,-0.299135685,0.1206479892,-0.1977964044,-0.0308268853,0.3564553857,0.1528513283,-0.1662874073,-0.024855366,0.0170527827,-0.3536676764,-0.1205809936,0.291339606,-0.1877669692,0.3857827783,0.2338583469,0.4491418302,0.243935734,-0.2276366353,0.2380734086,-0.2830122709,0.5588680506,-0.3239809871,0.0037709209,0.3467084467,0.3735772669,-0.0172323547,0.413033098,0.1587423533,0.0830663815,0.0739988536,-0.1531870365,-0.0825018138,-0.0720004067,0.0009111607,0.3129196465,0.2121030539,0.0193181764,-0.1659513265,0.0762920603,-0.3645926416,-0.0965407118,0.2978362739,-0.0708377883,-0.0049900734,-0.1782595068,-0.1960586905,-0.4301601946,0.3066400886,0.046772752,0.0566468686,-0.4846808612,0.0047120983,0.188363418,-0.0662763193,0.4130117297,-0.1791920215,-0.0939894393,-0.1971548498,0.015315868,-0.2188730836,0.1236937866,0.0881945565,0.2223389894,0.4836180508,0.4964018166,-0.2581639886,-0.3417085111,-0.0213226005,0.166251123,-0.1854681373,-0.1168784946,-0.169256866,-0.0340829827,-0.5279605985,-0.000780469,-0.1434282511,0.133117944,-0.1576189399,-0.1336166561,0.187634483,-0.2138563097,0.0221807435,0.1780868322,0.4660642147,0.0363125168,0.3925412595,0.1543625146,0.3062667549,0.3658607304,0.4862293005,-0.1503633261,-0.0955573767,0.3650470078,-0.0271750987,0.0663823262,0.1938154101,-0.0296924394,-0.2304403782,0.0661847517,0.6470083594,-0.077885434,0.3327542245,0.2215340137,0.1444753557,-0.3330666125,-0.1154892817,0.020912664,0.1776442826,-0.0332655795,0.2684893906,0.4100492895,-0.3196966946,-0.0107884984,-0.0679792315,0.8436338305,-0.034831211,0.1712500453,0.3351177275,0.1629513353,0.2048090398,-0.1287787706,0.0413953215,-0.3531040251,-0.3450030386,0.0493533872,0.0249393992,0.0562025495,-0.1280156225,-0.1890036762,0.138894558,0.1727903336,-0.0059903548,-0.0386299677,0.0435160026,0.0107331201,-0.1098451242,-0.2672280669,0.1370902359,-0.1771621555,0.2574095726,-0.1872772276,0.1398834586,-0.1132068112,-0.2107560784,-0.1768612564,0.0562444143,0.1625495702,0.1861705184,0.0570835099,-0.2973457575,0.063923724,0.2093394101,0.2081397176,0.2727320492,-0.4215897322,0.3225466609,-0.0717777982,-0.1218273416,-0.0585621744,0.0907502845,-0.1382073015,-0.2263236493,-0.1680245548,0.012827835,-0.0857556909,-0.0277710874,-0.2167381048,-0.153604582,-0.0171378553,-0.2819541693,0.0345001258,-0.3385381997,-0.0635388941,-0.1276970059,0.1661129892,0.064143911,-0.1901186109,-0.1921638548,0.4364754856,-0.1817346066,-0.0750915185,0.106847018,0.3403317332,-0.2032791674,0.4763286114,-0.0827659145,-0.1230335832,-0.3811211288,0.1494606584,0.3268787563,-0.6355054379,-0.0542984717,0.0313308649,0.0020845477,0.0202050395,0.1349167675,0.3093496561,-0.0739419982,0.1983305514,-0.2839328647,-0.3859999478,0.1469189525,-0.4365637898,0.3969067037,0.1614441127,-0.0916481987,0.1137206033,-0.2212848514,-0.3606903553,0.1830333173,-0.1696034968,0.0537523217,-0.1036186963,-0.202817291,0.0302618463,-0.3651135564,0.1210050285,0.0859073997,-0.0484625325,-0.2719352245,0.13642326,0.0843022838,0.0684452131,-0.0323546603,-0.0933424607,0.2066884488,-0.1922902763,-0.1824374646,0.2643355727,-0.0284645297,-0.0154687222,-0.2976878881,0.1698185354,-0.0403713733,-0.1420922875,0.0549063943,0.0622635484,0.1633230746,0.136835292,0.3143016696,-0.3336694837,-0.0030382888,0.0163324066,0.1789251566,0.4551702738,-0.1466173232,0.1860788763,-0.2518864274,0.046732638,0.0950345472,0.357600987,0.050843358,-0.5674385428,0.1733662337,0.0795140564,0.2429518104,-0.2657651603,-0.0053452887,0.2520882487,0.0589039959,-0.0523970686,0.1822541952,0.3525527418,-0.2223055363,0.2296389639,0.2284063846,0.2739516795,-0.001555449,0.2047388405,0.0983073041,-0.1196881607,0.0783313885,0.1528793275,0.0978297666,0.0869987532,0.2946607172,-0.5620661974,0.1896393895,0.0522390902,-0.0473052785,0.1052413955,-0.5263912678,-0.0780543387,0.0872668549,-0.1740516126,-0.0565504245,0.2065705061,0.3204136491,0.097242415,-0.4487623274,-0.2887787521,0.3061585128,-0.4001098573,0.2217199653,0.1786717772,-0.048632361,-0.2135446221,0.0593282692,0.0635451078,-0.3613875508,0.2754186094,-0.0133579187,-0.3275069296,-0.2001144737,-0.0430393256,0.6938011646,-0.114504233,-0.0358828418,0.4734570086,0.0726231858,-0.1695219129,-0.0303363781,0.3387397826,0.5704157352,-0.0651936308,0.1733290404,0.0522644855,0.2001823634,0.050538782,-0.2509518862,-0.121166721,0.1557346433,0.2567350566,0.1808181405,0.2048124522,-0.1677437425,0.0899707302,-0.0856772736,0.310136348,-0.1037508473,-0.3237806857,-0.4459539056,-0.1738522053,-0.1494963318,-0.2065624148,-0.5698300004,0.0131414169,0.2049328834,0.0381456763,0.1246638,-0.1438594162,0.0887777805,0.0146235581,0.0524716564,0.2618640363,-0.0810922906,0.0464999899,-0.5845843554,-0.593850255,0.2643868625,-0.1603444964,-0.2391425222,0.0671630278,0.04182069,-0.0766819566,-0.1839275509,0.3914036751,0.0103833023,-0.0426143855,-0.1254459769,-0.3123367131,-0.0809249133,0.0939121991,0.0401515886,0.0324271321,-0.2033738792,0.1921676993,-0.047948394,-0.049196057,-0.1247798651,0.3439123333,-0.5425662994,-0.144488886,0.2515626252,0.0472703427,0.4137514234,-0.0639180392,-0.1729413271,0.0602880381,-0.3672413826,-0.0515214168,0.2315686941,0.1666563302,0.2415644825,-0.209898293,-0.3652320206,-0.1113158837,0.0603880025,0.209894821,-0.133084178,-0.0028529833,0.2031837404,-0.0071489755,0.2655232251,0.3117284477,0.0755308494,0.1269062757,-0.1885558069,-0.2119236588,-0.2781872451,0.4581196308,0.0567617491,-0.1256155074,-0.1653666794,0.2096101791,0.2247540653,-0.3838548362,-0.3656026125,0.0049669375,0.2320416272,0.0370318852,-0.1204366833,-0.0081967805,-0.181891948,-0.0773857981,0.045634184,0.7375470996,0.1435072571,-0.4418088198,0.2364855558,-0.1404327601]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1934","title":"Add Stanford Sentiment Treebank (SST)","comments":"Dataset added in release [1.5.0](https:\/\/github.com\/huggingface\/datasets\/releases\/tag\/1.5.0), I think I can close this.","body":"I am going to add SST:\r\n\r\n- **Name:** The Stanford Sentiment Treebank\r\n- **Description:** The first corpus with fully labeled parse trees that allows for a complete analysis of the compositional effects of sentiment in language\r\n- **Paper:** [Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank](https:\/\/nlp.stanford.edu\/~socherr\/EMNLP2013_RNTN.pdf)\r\n- **Data:** https:\/\/nlp.stanford.edu\/sentiment\/index.html\r\n- **Motivation:** Already requested in #353, SST is a popular dataset for Sentiment Classification\r\n\r\nWhat's the difference with the [_SST-2_](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=glue&config=sst2) dataset included in GLUE? Essentially, SST-2 is a version of SST where:\r\n- the labels were mapped from real numbers in [0.0, 1.0] to a binary label: {0, 1}\r\n- the labels of the *sub-sentences* were included only in the training set\r\n- the labels in the test set are obfuscated\r\n\r\nSo there is a lot more information in the original SST. The tricky bit is, the data is scattered into many text files and, for one in particular, I couldn't find the original encoding ([*but I'm not the only one*](https:\/\/groups.google.com\/g\/word2vec-toolkit\/c\/QIUjLw6RqFk\/m\/_iEeyt428wkJ) \ud83c\udfb5). The only solution I found was to manually replace all the \u00e8, \u00eb, \u00e7 and so on into an `utf-8` copy of the text file. I uploaded the result in my Dropbox and I am using that as the main repo for the dataset.\r\n\r\nAlso, the _sub-sentences_ are built at run-time from the information encoded in several text files, so generating the examples is a bit more cumbersome than usual. Luckily, the dataset is not enormous.\r\n\r\nI plan to divide the dataset in 2 configs: one with just whole sentences with their labels, the other with sentences _and their sub-sentences_ with their labels. Each config will be split in train, validation and test. Hopefully this makes sense, we may discuss it in the PR I'm going to submit.\r\n\r\n\r\n\r\n","comment_length":11,"text":"Add Stanford Sentiment Treebank (SST)\nI am going to add SST:\r\n\r\n- **Name:** The Stanford Sentiment Treebank\r\n- **Description:** The first corpus with fully labeled parse trees that allows for a complete analysis of the compositional effects of sentiment in language\r\n- **Paper:** [Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank](https:\/\/nlp.stanford.edu\/~socherr\/EMNLP2013_RNTN.pdf)\r\n- **Data:** https:\/\/nlp.stanford.edu\/sentiment\/index.html\r\n- **Motivation:** Already requested in #353, SST is a popular dataset for Sentiment Classification\r\n\r\nWhat's the difference with the [_SST-2_](https:\/\/huggingface.co\/datasets\/viewer\/?dataset=glue&config=sst2) dataset included in GLUE? Essentially, SST-2 is a version of SST where:\r\n- the labels were mapped from real numbers in [0.0, 1.0] to a binary label: {0, 1}\r\n- the labels of the *sub-sentences* were included only in the training set\r\n- the labels in the test set are obfuscated\r\n\r\nSo there is a lot more information in the original SST. The tricky bit is, the data is scattered into many text files and, for one in particular, I couldn't find the original encoding ([*but I'm not the only one*](https:\/\/groups.google.com\/g\/word2vec-toolkit\/c\/QIUjLw6RqFk\/m\/_iEeyt428wkJ) \ud83c\udfb5). The only solution I found was to manually replace all the \u00e8, \u00eb, \u00e7 and so on into an `utf-8` copy of the text file. I uploaded the result in my Dropbox and I am using that as the main repo for the dataset.\r\n\r\nAlso, the _sub-sentences_ are built at run-time from the information encoded in several text files, so generating the examples is a bit more cumbersome than usual. Luckily, the dataset is not enormous.\r\n\r\nI plan to divide the dataset in 2 configs: one with just whole sentences with their labels, the other with sentences _and their sub-sentences_ with their labels. Each config will be split in train, validation and test. Hopefully this makes sense, we may discuss it in the PR I'm going to submit.\r\n\r\n\r\n\r\n\nDataset added in release [1.5.0](https:\/\/github.com\/huggingface\/datasets\/releases\/tag\/1.5.0), I think I can close this.","embeddings":[0.0112798819,0.1192583814,-0.0384877957,0.0895807073,-0.1235435531,0.0041984264,-0.0643986538,0.331428051,-0.0082973838,0.1232367158,-0.1510062814,-0.0503281951,-0.1747421473,0.0092151193,-0.0243232343,-0.1746755689,0.1758180559,0.0721649304,0.20083718,-0.2114398181,0.0086534936,0.3530318439,0.3495486379,-0.1575782448,-0.4798354208,0.0219505038,-0.3393312395,0.0572050959,-0.2454832941,-0.2264818996,-0.0220576432,0.1692650616,0.2136270702,0.3234196603,-0.0001097931,-0.0467267558,-0.0878914297,-0.0702384114,-0.0494960807,-0.0560395122,0.1228134558,-0.3275004029,-0.1960346401,-0.2914278507,-0.3313576281,0.0354921632,0.1182006821,-0.4680806994,0.3969146907,0.0490526482,0.1314833611,-0.2269693613,-0.2417730838,-0.0211666413,0.4285118878,0.0752754137,-0.2057125866,0.0384733118,0.2158942223,0.0852221325,-0.1031058878,0.2458280176,-0.0556601584,-0.1349495351,0.1626085639,0.0046964949,0.3424948156,-0.2709479928,0.2253102064,0.2150636613,0.4312856793,0.031700775,-0.2100484371,-0.4132013023,-0.195575282,-0.0735763758,-0.0824564621,0.0736861527,-0.0564692579,0.2975814044,0.0754755512,0.0905292034,-0.0331035927,0.0018298132,0.1095241159,0.2613213956,0.0166818108,-0.0209462252,0.0091383979,0.0073259799,-0.1416196972,-0.0180235095,-0.1218041629,0.0635258928,-0.263047576,-0.0163766947,0.2049950361,-0.1212208122,-0.1336365938,0.0699821487,0.1295101345,0.0264451019,0.3932116926,-0.1462831497,0.2172733098,0.279576838,0.0693887323,-0.0662046149,0.1564803571,-0.2990593314,-0.1084357575,0.0463691093,-0.3227201998,0.1991577595,-0.0595175661,-0.0971804261,-0.122473754,-0.183228299,-0.301894635,0.1513046771,-0.0439412072,-0.05944141,-0.1932768226,0.0872102305,0.2687995434,0.2426601946,-0.0856457874,-0.0523308292,-0.2324338853,-0.2377403229,-0.1760891974,0.2392558455,-0.4137883484,-0.0128490357,0.165639326,-0.0951641798,0.1478443593,-0.2595514953,-0.1446399242,0.0700544789,-0.4958370328,-0.2163425535,0.3694247603,-0.2475334406,-0.0921374559,0.1072521061,0.0750113949,-0.1486422718,-0.3017863631,-0.2305287868,-0.0504846498,-0.1388576031,-0.1169377416,0.1084909886,-0.5719519854,-0.1299607605,0.0320318304,0.6869637966,0.0697574094,0.113776356,0.1400628835,-0.0591961369,-0.2731010616,-0.0700209364,-0.1299939603,0.2110172212,-0.5694032907,-0.2650302052,-0.0912726969,0.1479579061,0.2791951001,0.4231003821,0.0170329511,0.2481164485,0.0002448008,0.5123127699,0.1964296699,0.0748057291,-0.2291440219,0.2969909608,-0.1413710415,0.3057640791,0.3425171971,0.1910930723,0.4229764044,-0.2710974216,0.0192479137,0.0855986923,0.1080737188,0.1160047725,-0.180905506,-0.2458872944,-0.1870650798,0.1197864935,-0.0076932143,-0.2369240373,-0.0289533101,0.1350192875,0.0477567576,-0.2560800314,0.2696814239,0.1024737656,0.1703469008,0.333902657,0.0468061082,-0.2138257623,0.2014058083,0.200720191,0.0972477049,0.4005851448,-0.0165911242,-0.277408421,-0.0217102934,-0.1775601655,0.0605730265,-0.2961050868,0.1532041281,0.440708071,0.3853669763,-0.0320356525,-0.3150791824,0.0746740103,0.0248824582,-0.0361564197,-0.0002978328,0.3181526065,0.2043041289,-0.4153083563,-0.0474801399,0.4842116833,-0.2368757725,-0.1496642828,-0.0064541376,0.4069054127,-0.3605393767,0.1970271468,-0.342190057,0.2783021033,-0.1142998412,-0.4313264191,0.0953032747,0.0758283064,-0.1641089767,-0.021644868,-0.0344171077,0.1987181604,0.2786545455,0.034716662,0.043798279,-0.2967430353,-0.220746994,-0.1168123037,-0.2512762249,-0.1402556747,0.2758912146,-0.4384073019,0.2842908204,0.0756899565,-0.1789025366,0.3452252448,0.4391682744,-0.0816558003,-0.0815190822,0.0605939962,-0.2587482035,-0.162834987,0.020614963,0.3056252599,0.1336439252,0.3033626974,0.2671150565,-0.0739379078,0.122191906,-0.1099769473,0.32742998,0.1718069613,0.0075331102,-0.064836733,0.2899579704,0.2163148075,-0.3010205626,0.1674779952,-0.1094603911,-0.0603115521,-0.0907523185,0.1104788855,-0.0647240803,-0.8813092709,-0.2019512057,0.2461913377,0.3136714995,-0.1259864718,0.3303887844,-0.3191558123,-0.2755281329,-0.031646695,0.1521807611,0.5010756254,-0.1796292514,-0.2361843437,0.288892597,-0.2545360923,-0.2063586861,0.128442049,0.0864742994,0.1716228575,0.1225602329,-0.2219211012,0.0132856285,-0.2261556685,-0.3961024284,0.3497328162,-0.4464884996,-0.1309728175,0.1899065524,-0.1678454131,-0.18155092,-0.4652253389,-0.0424661189,-0.281858623,-0.0581852831,0.1937846243,-0.0346710272,-0.3212600648,-0.232145533,-0.3512441814,-0.093295157,-0.2339584976,0.4032623172,-0.0072659599,0.0121446094,-0.0134367803,-0.1036311314,-0.008050777,0.046579808,0.247530058,-0.0933031589,-0.6117536426,0.4081140161,-0.0870466828,-0.1862791032,-0.1620958894,-0.0290495772,0.0252390001,-0.1972189099,-0.4274520278,0.3164747953,-0.4015045166,0.2918103635,0.216587469,-0.0193030313,0.3641271591,0.1073143482,-0.2019039989,-0.1293396503,-0.0527275279,0.449215591,0.0809502378,0.31931144,-0.0374340937,0.3219330311,0.3447920382,0.2939225733,0.0979869813,-0.1076745763,0.1996205747,0.0751843378,0.1228551567,-0.0145744309,-0.0843418613,0.4467892349,-0.130627349,-0.167804122,0.7083497047,0.1606945544,-0.3120301366,0.0236514416,0.0749032497,0.0017536622,-0.1970581412,0.1318641752,-0.0194742009,0.1194981411,-0.0780112892,-0.2459765971,-0.2520020902,-0.1354495138,0.1807561815,-0.2153787017,0.1523097456,0.1551080644,-0.4818228185,0.3262999356,-0.4751217663,0.0387698077,0.1615716219,-0.3188147247,-0.2837565541,-0.2723180354,-0.0321609527,-0.036125578,0.0991083384,-0.0412064455,0.065479815,0.154103756,0.0679922327,-0.280470252,-0.1939625293,-0.3909927905,-0.1564810425,0.7242668271,0.2994261682,-0.3492622375,-0.1079903767,0.2787587345,-0.0144496756,-0.1087837294,0.0934435502,-0.0039565773,-0.0916921347,0.1541236192,-0.1951795667,0.3688026965,0.2536669374,0.1023527756,0.4647565186,0.1815450937,-0.2139131129,0.4730787575,0.2187859267,0.098831825,-0.0350847058,-0.0027584631,0.2618695199,0.2586822808,-0.1691660285,0.7590197921,-0.0965479091,-0.4454907477,0.3889067173,-0.0709320083,0.1964934021,0.4117993116,-0.0210539661,-0.0448528454,-0.2916581631,-0.0613364466,-0.2347286344,0.1775606424,0.0021507286,0.1530530453,-0.1977614164,-0.2779597342,0.3355336189,0.2348188758,-0.240416795,-0.6540489793,0.4699546993,-0.4808736742,0.6152433157,-0.0068604653,0.9738184214,0.2406240553,0.0756209195,0.1862126142,0.1945261508,0.8159403205,-0.3254711926,-0.1459634304,0.2518002987,-0.0564008802,0.0199969877,0.0962995663,-0.0471560061,0.3474225104,-0.2402160615,0.2837764919,0.3312977552,0.1096833199,0.1515101641,0.2711781859,0.1631530076,-0.192283541,0.0210199021,0.14644894,-0.0008678964,-0.0281520244,0.1261155307,-0.2328004092,-0.2937428355,0.1205662712,-0.3868802488,-0.0269755404,-0.2062440366,-0.0598147996,-0.0452341177,0.0565337129,0.3580532074,0.1823706478,0.5915020704,-0.0451863222,-0.0393077508,0.0438510813,0.2364031971,0.0503293127,0.2368572056,-0.1921856701,0.4849185944,-0.0509319231,-0.0648451298,0.1485805064,-0.0491418205,-0.1113169044,-0.4014542401,-0.4292569757,0.2828301191,-0.1761767268,0.0897038057,0.3294538558,-0.1359443218,-0.2061397433,0.107684806,-0.2161507457,-0.03886199,-0.0858887136,-0.1092247069,-0.0657979921,-0.1213394031,0.16281946,0.0046632756,-0.1190033406,0.2968683243,0.1112079099,-0.2199805081,-0.1470145285,-0.1252937019,0.2480339855,-0.4628822505,0.0330650993,-0.3658758402,-0.5363413095,0.0706134588,0.3297194839,0.5352128148,0.1748828292,-0.2412402481,0.0755360425,-0.0525752679,0.3484301567,0.3149802387,0.3542430699,-0.0420862548,0.2739790678,-0.2029297203,0.1251733452,-0.3771889806,-0.001891009,0.035700839,0.1249698997,0.2111822814,-0.0788631067,-0.1059978604,-0.0484320894,0.0438988209,0.1600702256,-0.0209657531,-0.1719918102,-0.1439905018,0.1226456612,-0.0382191688,-0.0527708381,0.3488315046,0.124433212,0.1541611403,-0.1197044998,0.3907666206,0.0019008276,-0.0947886929,0.0676119104,0.1955424547,-0.0288350489,0.1638769805,-0.0466976315,-0.1064906493,0.2432933301,0.174799934,0.2537864447,-0.1198059618,-0.1528309584,0.0559826605,0.2236577123,0.2706699967,0.1359575391,0.0209847596,-0.1497798264,-0.0141215557,0.1014575809,0.1175143272,-0.1352743506,-0.1448087841,-0.2437280416,0.0982068256,0.2335881442,-0.1265488118,0.0511780493,-0.4640073478,-0.0643715039,0.1395639181,0.3388160467,0.430893898,-0.036968179,-0.1727529615,-0.0484457314,0.7477487326,0.3190096319,0.1648743302,0.2736594677,-0.001732765,-0.0894720182,0.055721797,-0.1706320494,0.0423385426,-0.079270944,-0.0321223103,0.3462046981,0.6103425026,0.3031258583,-0.0756670758,0.4391905367,0.4845722914,0.4814592302,-0.1635466218,0.3022128046,-0.032769978,0.3423241377,-0.0382039621,-0.0721719041,-0.0343911499,0.5861446261,0.0283275452,-0.4237352014,-0.1394070387,-0.2527531385,-0.2173991501,0.0064956532,-0.3790783584,-0.3607700169,-0.138627246,-0.0629319176,-0.0175553001,-0.0623013154,-0.2217354029,-0.4406215847,0.491353929,-0.1431449503,0.5266364813,0.2590594292,-0.212746501,0.2561628222,0.53714782,0.2045851499,-0.0188634843,0.1289850175,0.006131452,0.0377056338,-0.1607207656,0.0372364298,0.064440608,-0.0379803069,-0.1927178204,0.322994113,0.0941634774,-0.1480181068,-0.1050899327,0.1399338543,0.1732362062,-0.1509922892,0.0871012583,-0.0413707346,-0.0266352445,0.1267888993,-0.0250871144,-0.3877917826,-0.30365026,0.141978085,-0.1652466208,0.1369761378,0.0289212465,0.0890304893,-0.1401530951,0.360235244,0.281947881,-0.1693965644,0.0916568488,-0.2331808954,-0.1792958081,-0.1964948177,-0.0736974403,-0.5881227851,-0.1721567065,0.1165200472,0.3494550884,0.0603447035,-0.379016608,0.1645309627,-0.4040742218,-0.0388545431,-0.3441653848,0.195051685,0.4417001903,0.3965959251,0.0322996974,0.0143397013,-0.1241382062,0.304014802,0.0465943106,0.0237881988,0.0360691249,-0.1295627803,-0.4040894806,-0.0360269211,0.0887745693,0.0363586769,-0.0754033327,-0.068658106,-0.2349932641,-0.2282786071,-0.1975767016,0.2237847149,-0.0482903868,0.6510016918,-0.2202948183,0.2103970647,0.2667027414,-0.449837178,0.0504957214,0.0548466519,-0.5308574438,0.0429505296,-0.2391296774,0.1314696968,-0.0098720985,0.3480008245,-0.0590456203,-0.1156236827,-0.0889889225,-0.5249195099,0.3044824898,-0.4829391539,0.0031633689,0.1065554023,0.2050375193,0.5696105361,0.18554914,-0.4168031514,-0.1150379479,0.2296257764,0.1689773202,-0.4160438776,0.4046005309,0.0235162396,-0.1792050749,-0.0051348805,-0.1631866395,0.1353644431,-0.2089085132,-0.1329484731,-0.4596967697]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1924","title":"Anonymous Dataset Addition (i.e Anonymous PR?)","comments":"Hi !\r\nI guess you can add a dataset without the fields that must be kept anonymous, and then update those when the anonymity period is over.\r\nYou can also make the PR from an anonymous org.\r\nPinging @yjernite just to make sure it's ok","body":"Hello,\r\nThanks a lot for your librairy.\r\nWe plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ? \r\nCheers \r\n@eusip","comment_length":45,"text":"Anonymous Dataset Addition (i.e Anonymous PR?)\nHello,\r\nThanks a lot for your librairy.\r\nWe plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ? \r\nCheers \r\n@eusip\nHi !\r\nI guess you can add a dataset without the fields that must be kept anonymous, and then update those when the anonymity period is over.\r\nYou can also make the PR from an anonymous org.\r\nPinging @yjernite just to make sure it's ok","embeddings":[-0.2978561223,0.5536879897,-0.0238176081,-0.1903211623,-0.0758951455,-0.1079585776,0.5019997954,0.0080968365,0.0642958507,0.0534695312,0.0712104365,0.1038329452,0.0041556037,0.1634862423,-0.038204141,0.0795786381,0.0454185903,-0.0966099054,0.0962188765,-0.2013753802,-0.4753124416,-0.2437654734,0.0791259259,-0.5910321474,0.1504401416,-0.2484027892,-0.2008707821,-0.2064752281,-0.0254556611,-0.4402409494,-0.2508310676,0.2532382011,0.0878261477,0.1687669903,-0.0001136982,0.0365013182,0.2275120318,0.0229101274,-0.2287050039,-0.1465033889,-0.4085339904,-0.1241108775,-0.0218316149,-0.1218808517,0.0511337407,-0.3378623426,0.1586951762,-0.4428370297,-0.2920340896,0.0187059697,0.187534377,0.2154204845,0.1018245146,-0.1916258484,0.1721318513,0.3090620339,-0.0694505945,-0.0765855312,-0.2525062859,0.2380914986,0.1200364232,0.1911775321,0.2108571678,-0.3427754939,0.2185998857,0.255967319,0.135845229,-0.0866364315,0.0983041152,-0.0139201935,0.5642251968,0.1111585796,-0.2762329876,0.0066860383,0.1424970478,-0.2253250182,0.1219130531,0.360276401,-0.1878941357,0.0065119327,-0.3873705268,-0.4332360923,-0.1487916559,0.2721284628,0.3960341513,-0.0827069655,0.2204151154,0.219899103,-0.1080729589,-0.0076145832,-0.0657849163,-0.131126374,-0.1911907196,0.0271343086,0.2725528479,-0.0197917856,0.0595823154,0.0854259059,-0.0407377742,0.1603176296,0.0219724867,-0.2648992836,-0.266192764,-0.101888001,0.1051796004,-0.2567987442,-0.0329098739,0.2582083642,0.1456508487,-0.1770040691,0.5126866102,-0.1028508022,0.2065980434,0.2185723186,-0.1635549068,-0.4257535934,0.0924938023,-0.2446938902,-0.0193248596,0.1227926016,-0.1386152506,0.1385225058,-0.3871082067,0.1158605218,0.1389353573,-0.3016420007,0.2873123884,-0.058875829,0.0300295111,-0.107908763,0.0076900152,-0.146346271,-0.1829976141,0.0220316928,0.1334694922,0.1433653831,0.0655953661,0.0237891339,0.327932179,0.2830860019,0.3779454827,0.0594570599,0.1582080573,0.0805631131,0.0714396834,-0.2646244764,-0.0486896634,-0.5071710944,0.0606984794,0.2869585752,0.4154135287,0.3485382497,-0.2451441735,0.1902703643,-0.1741662472,-0.2699770629,0.2741080821,0.2919701338,-0.0598984323,-0.1097665876,0.3346720636,0.2572618723,0.0371020697,-0.0198328253,-0.0654725432,0.3630783856,-0.1588656306,-0.2072477937,0.1318625063,-0.1387540698,-0.2335201204,-0.0035389329,-0.2095681131,-0.1997006834,0.0260429531,-0.2894255519,-0.0541625917,-0.2797907591,-0.3725432754,-0.3235584497,-0.255299896,-0.2175352722,0.4006320238,0.5792213678,0.2217822671,-0.1478639841,-0.2846814692,-0.0699517876,-0.1696240604,-0.3042348027,-0.4500306249,-0.1257377714,0.132505551,0.0718635246,-0.2021477968,0.3229564726,0.3146029711,-0.1585004628,0.2660982311,-0.2799395323,0.1645666063,0.1382402778,0.5464157462,-0.0631281361,0.1361641288,0.2277239114,-0.1811691076,-0.2020771503,-0.328779161,0.4175469577,0.1670813411,-0.5736303329,0.2955128849,-0.2195345163,0.0891285613,-0.1777299941,0.0052080504,0.2768193781,-0.284775883,-0.2003248781,-0.07579308,0.1646350026,0.023851119,0.1015504226,-0.4024479687,-0.2272196859,0.127779007,0.0510368347,-0.2499553263,-0.0462137721,-0.1039530337,-0.1574058086,0.1313714683,0.1555237472,-0.1213012636,-0.0245550033,0.7049100995,0.4160323739,-0.0812954009,-0.2752478421,0.1455888301,0.133646518,-0.1570249349,0.2209073603,-0.6624383926,0.5039723516,-0.0585588031,-0.0115834353,-0.1644259244,0.2266272008,0.4233987331,-0.2268246859,-0.1009383202,-0.2740346789,0.2443836033,0.5010877252,0.2135253847,-0.1423729211,-0.4439190924,0.2676128149,-0.4807438254,-0.0149002178,-0.0389987268,-0.0193484724,-0.1518962383,0.2312796414,-0.0101371743,0.1605871618,0.1751214862,0.3742703497,-0.2108206898,0.0669731125,0.2131949812,-0.1076783016,0.276350379,-0.0201313589,-0.3504347205,-0.1547375917,0.4559481442,0.2296287417,0.0018173329,0.0319887139,0.2925581336,0.171434626,-0.0211071018,-0.0420381501,0.2286534905,-0.1715062857,0.1336688846,0.2258008569,-0.2456106395,-0.007508724,0.3010831773,0.0826040581,-0.1294431984,0.2552319169,-0.2911655903,0.812263608,-0.3899851143,0.1093774661,0.0818701759,-0.242402941,0.0765795931,0.1667841077,0.158855617,-0.3950011134,0.5738779902,0.1488127112,0.0059595197,-0.4898711443,-0.4616575539,-0.033103928,0.185569495,0.0171722677,-0.1285817325,-0.2001145035,-0.2259712666,-0.0200328324,-0.0314349271,-0.2088885158,-0.2280631363,-0.132307902,-0.1854385287,-0.0510286056,-0.1308766752,-0.206087634,-0.0663331598,0.2258476913,0.4678015113,0.4428364933,-0.0063269543,0.2578885853,-0.1487094909,-0.2320683002,-0.1050001904,-0.320115149,-0.2184946388,-0.4413897097,0.28193295,-0.3696287274,-0.2477011979,0.2761833072,-0.1013437733,-0.0418128707,0.3652010858,-0.529271841,-0.0816141069,0.1441634595,0.3721830547,0.1676898897,-0.1993681788,0.5032022595,0.1920815557,-0.2018285692,-0.1152354255,-0.1001846567,0.21195364,-0.072530508,0.5440939665,-0.3213922381,-0.050977923,-0.070249334,0.6683695912,0.4097921848,-0.0773719773,-0.031078646,-0.1300684661,0.5164962411,0.0133107482,-0.0727206469,-0.0798474401,0.1519530118,-0.0881566033,0.006335109,-0.0149987414,-0.4001375139,-0.0334775746,-0.094628945,-0.1591327786,-0.2538811862,0.0483876728,-0.0147431949,0.6296268106,0.1038355678,-0.1671636403,-0.568908751,-0.61127913,-0.016611727,0.2107746303,0.097506687,-0.371168226,-0.1986097544,0.4038897753,-0.3924840689,0.4433428943,-0.2406228483,0.0674040467,-0.0974951684,0.0603990443,0.0572580546,0.2134022266,0.4243105352,-0.4223254323,-0.3225738108,0.1250233352,0.3005471528,0.135809347,0.2328307629,-0.2882535756,0.3327237964,0.1789940447,0.0821344629,-0.1581816971,0.269654423,0.5318704247,-0.1897789687,-0.3482929468,-0.1841911077,-0.27729249,0.0357519388,-0.0944797695,-0.0110327639,-0.3482284248,-0.0295591094,-0.0094439369,0.2832340896,-0.1524950117,0.4053567946,0.2272452116,-0.3571989834,-0.0322472565,0.5289489031,0.2385687679,0.2302054614,0.1427942365,0.4895375669,0.4533395767,-0.1047912389,-0.453261137,-0.0820490643,-0.4671744108,0.2918991745,0.4869161248,0.3948281109,0.4081510901,0.1663912535,0.0477416031,-0.0323490277,0.0627355129,-0.0819478929,-0.1082182154,-0.145276919,-0.4273102582,0.1304296851,-0.057151407,-0.1948762983,0.330611974,-0.0339324884,-0.1191854849,0.5634446144,0.1458991617,1.1023107767,0.1004173532,0.2189497054,-0.2815106511,-0.1477197856,0.4086607099,-0.0295336954,0.1976179034,0.0389001109,-0.0740398169,0.0169318095,-0.046031896,0.106662102,-0.0323695764,-0.0779184923,-0.1003015861,-0.1667453796,0.0400974266,-0.1215349808,0.406752944,0.1403793246,-0.1238909811,0.2923855782,0.1070315093,0.1914249212,-0.301886797,-0.0045081056,-0.280326277,-0.1367900223,-0.1383551359,-0.0493660346,-0.0047558742,0.2829570174,-0.2809303999,0.0539188199,-0.3358972669,0.314494282,0.2217281908,0.4640049338,0.0496843234,-0.2218732834,0.0630485117,0.0547405519,0.0449578911,0.2886951268,-0.0992085114,-0.1164653525,-0.0068295887,-0.2520050108,0.4243529439,0.0123268533,-0.469453454,-0.3177482486,-0.2418293506,0.3339381218,-0.3270104527,0.5733966231,0.2793072164,-0.3338207603,-0.1607431024,0.133107394,0.2079848945,0.0421002731,0.1667079031,0.2086300701,-0.0422763713,-0.1200368404,0.2904540002,0.2368242294,-0.2219737619,0.3192644715,-0.0267113615,-0.174716264,-0.1491135061,0.1113802269,0.0908229426,-0.5180429816,-0.1389474273,0.215021953,0.5172666311,-0.1865574121,0.0901883692,-0.0313067436,0.2822612226,-0.3256951571,-0.1173382625,-0.3173075616,0.1764895767,-0.0433794521,0.0434846319,0.1321836412,0.1180731431,-0.0843110532,0.1214717925,-0.3499597013,-0.1642439812,0.0788089409,0.1941703409,-0.0780221,0.1549396217,0.2916842997,-0.3930835426,0.0790205151,-0.3343438506,0.1350063235,-0.1498425901,-0.0629085675,0.1342587322,0.0922139958,-0.3548137844,-0.2271988392,-0.2885832489,0.0102207204,0.1557580829,0.0311393477,0.0854372531,-0.1410771757,-0.5126219988,0.0328138433,0.163709715,-0.0851256996,-0.2230620086,-0.0045075379,0.2441371828,-0.092936106,0.1326643974,-0.2574223876,0.0306833219,0.3204216957,0.1555722952,0.3057708442,0.1402736157,0.1473254561,-0.1621403396,0.2386338115,0.0103780553,0.1027916893,-0.0022573615,-0.3916282058,0.1462599784,0.2996270955,0.2058556229,-0.396943301,-0.0135459099,-0.0898562968,0.3488265276,0.1341765076,0.3458150327,0.2564822435,-0.1308520585,-0.0300770439,0.0337997749,-0.098115027,-0.3168902993,0.0707860216,0.0121033555,0.07762447,0.074657768,0.3548707068,0.5277468562,-0.1667529345,-0.0421070084,0.1757559776,0.1529548019,0.0668763295,-0.0905277431,-0.0371429212,0.0062319939,0.2874759138,0.0308411345,0.3539128602,0.4143075943,0.2471048981,0.0791501924,0.1839893907,0.3479206562,-0.0506792627,0.2434340119,0.2524256408,0.1177736446,-0.4141272306,-0.0449197367,0.037203867,0.0642395541,0.1411059201,-0.2051169574,0.0313310139,0.1824617535,-0.1503529102,-0.2200534195,0.174153924,0.1781613082,0.3247277737,-0.2266158164,0.0150916744,0.1836255938,-0.0367107466,0.3735908568,0.7002019286,-0.0724854246,-0.0966845378,0.0389509574,0.4729221165,0.0671328604,-0.0448374003,0.0145800607,0.1043502092,0.369075954,-0.0846209228,-0.1591265649,0.0622224361,-0.0456527695,0.3700082004,0.1289583743,-0.0270934403,-0.2512516081,0.2030456662,0.1691949964,-0.1808620095,-0.1311328709,-0.1924048811,-0.0200504363,-0.0679725707,0.5362581611,-0.0872044414,-0.1096591651,-0.079727158,0.0757281855,-0.0774374381,0.0412250757,0.0676566362,0.1005141884,-0.2797425985,-0.2162727863,-0.0529522821,0.4314086437,-0.2296760678,-0.1711851954,-0.2231504172,0.029984206,0.4677249789,0.0866039693,0.2552157342,-0.0955875888,-0.3125907481,0.4435620308,-0.1287368685,-0.0321025029,-0.2192429453,0.3003244102,0.0772808567,0.1784941852,0.2000684887,0.2640064061,0.0090312911,-0.1725067049,-0.069413431,-0.0369098447,0.3054794967,0.2970537245,-0.1720961332,0.2769836783,-0.1313273162,-0.263774842,0.1774761379,-0.0812476128,-0.4348013401,0.1676120162,0.2094636261,-0.2985935509,-0.346267581,-0.1960131079,0.2539476752,0.0089696646,-0.3364588618,-0.0469110124,-0.2473500073,0.0109994113,-0.0972685739,-0.0149678579,-0.0300883707,-0.0220288914,0.0159425009,-0.0463811457,-0.2196148932,-0.0853785276,0.2916968167,-0.5352232456,0.2132209539,0.0055197626,0.3361277282,-0.3151449263,-0.3718156219,-0.4345390499,0.0649122447,0.2189659625,-0.0057221577,0.0519212596,-0.2206096947,-0.0004879076,0.0172971282,-0.1764282882,0.4233128428,-0.1038066074,-0.0114996405,-0.1624679118,-0.1279369295]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1924","title":"Anonymous Dataset Addition (i.e Anonymous PR?)","comments":"Hello,\r\nI would prefer to do the reverse: adding a link to an anonymous paper without the people names\/institution in the PR. Would it be conceivable ?\r\nCheers\r\n","body":"Hello,\r\nThanks a lot for your librairy.\r\nWe plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ? \r\nCheers \r\n@eusip","comment_length":28,"text":"Anonymous Dataset Addition (i.e Anonymous PR?)\nHello,\r\nThanks a lot for your librairy.\r\nWe plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ? \r\nCheers \r\n@eusip\nHello,\r\nI would prefer to do the reverse: adding a link to an anonymous paper without the people names\/institution in the PR. Would it be conceivable ?\r\nCheers\r\n","embeddings":[-0.2011668831,0.4423335195,0.0201971103,-0.0765717253,-0.0947154462,-0.0944121331,0.4387371838,-0.0027232436,0.0362781845,0.0934396684,-0.0507460982,0.0669827163,0.0037659982,0.0874343961,0.1370116174,0.1177665591,0.1308901608,-0.0895970315,0.2099281102,-0.2495173812,-0.5651833415,-0.1786230206,0.0916349292,-0.5738092065,0.23158741,-0.207276538,-0.2125302702,-0.1660725027,0.0858120248,-0.3669403493,-0.2833172381,0.1583455801,-0.0273778979,0.2683500051,-0.0001262668,0.1422165483,0.2734455168,0.0246185511,-0.2487211525,-0.1675764769,-0.5025137067,-0.0314126499,-0.0990062207,-0.1659391075,0.0814038664,-0.3552331924,0.1146846637,-0.3773088753,-0.3385237753,0.0273261201,0.0901736543,0.1165459156,0.0729364231,-0.0961636305,0.1395261139,0.4059197903,-0.018656116,0.0075163785,-0.2865670323,0.2626431882,0.2199782282,0.0724457428,0.1970395595,-0.3708330393,0.2289111912,0.1912171692,0.0853175521,-0.0382951051,0.0687063485,0.1095335111,0.6216089725,0.0948709175,-0.2760731578,0.0327650793,0.147463575,-0.0852832496,0.0675747171,0.3558513522,-0.1859707385,0.0162687916,-0.4244497716,-0.506978333,-0.1157958657,0.2613269687,0.3833504021,-0.1147986129,0.3029622138,0.2377171516,-0.1090036109,0.0225542542,-0.1252333224,-0.1423774809,-0.2535927296,0.1280827224,0.3269246221,-0.0503585227,0.0798263326,0.0344047956,-0.0289014559,0.0962052792,0.0563223921,-0.2872754633,-0.2855663896,-0.0859597772,0.0643912703,-0.3160105646,-0.0880991966,0.2874583602,0.0498653986,-0.158363983,0.5381066799,-0.0784931779,0.1278113425,0.2147779316,-0.230950281,-0.4649635255,0.0294684209,-0.2643438578,-0.0542346984,0.0519972518,-0.2531796098,0.1072175652,-0.4038323164,0.1170241907,0.0738791749,-0.3461569548,0.3234764934,-0.0996323004,-0.0508287996,0.0475881547,0.0394910723,-0.0566453002,-0.2179220915,-0.0798602328,0.127335012,0.2161314934,0.0571418367,0.044197984,0.2492774278,0.2355629802,0.2972955406,-0.0089328056,0.2083230168,0.0413716249,0.1397211403,-0.2625546455,-0.0491480008,-0.6527947187,0.0888416544,0.3641912341,0.2272872776,0.2947358787,-0.2152953744,0.0771596283,-0.1591785252,-0.2361368537,0.1870181262,0.1990443319,-0.1162700206,-0.1539996713,0.2643208504,0.3141643107,-0.0525802039,-0.0466035381,0.0177290533,0.4519748688,-0.2121914476,-0.2798964679,0.2454179972,-0.1773500592,-0.1309855729,0.1275194585,-0.1967158169,-0.1245191246,-0.0778549537,-0.3614664078,0.0002923247,-0.2300691456,-0.3864842951,-0.221759975,-0.3067471087,-0.148767069,0.3727555573,0.6480480433,0.1632297337,-0.0901934057,-0.1348420382,-0.0588171333,-0.0446889363,-0.2744155526,-0.4228567481,-0.1009271741,0.1049097553,-0.0332545787,-0.3364382386,0.4607168734,0.3744395673,-0.1109757423,0.3731579781,-0.269818902,0.1481581628,0.2017128915,0.5984317064,-0.0929670185,0.1491685212,0.3131767809,-0.0584281608,-0.231814146,-0.3104812801,0.3493329585,0.1211912334,-0.5198779702,0.2872042954,-0.130935654,0.1329517066,-0.0196609981,-0.1120867133,0.1630742401,-0.3360396922,-0.1026534066,-0.1313270479,0.2828582823,0.0319539905,0.1305522025,-0.3676549494,-0.1331425756,0.223424688,0.1517508924,-0.228390485,0.0358835943,-0.0845753103,-0.1171071529,0.1633034945,0.1684955657,-0.1520997286,-0.0657869726,0.710524261,0.4017373919,-0.0259381682,-0.3726828694,0.1247756258,0.1605218947,-0.0716081113,0.0807434842,-0.6590612531,0.4140938818,-0.0702203438,0.0705798864,-0.108390823,0.1661629975,0.2895028293,-0.2371481508,0.0395896994,-0.1682233214,0.3143056035,0.5559506416,0.3530573845,-0.1535458863,-0.4435145855,0.1051558256,-0.6972011328,-0.0495921783,-0.026042847,-0.0027955971,-0.2676233053,0.2278954983,-0.0403534733,0.2359807938,0.2688292563,0.2979460061,-0.1101177037,0.1627380103,0.2760838568,-0.0626733378,0.2161822915,-0.1061522886,-0.3277377188,-0.1313817054,0.4049280584,0.2382092774,0.0116231069,0.0085196709,0.3042285144,0.1753628254,-0.1000528708,-0.0469267294,0.2736749351,-0.1836227924,0.1325827837,0.2519230843,-0.205432117,-0.0217876714,0.2373282909,0.1182722375,-0.0469168164,0.1121253148,-0.3568478525,0.778847754,-0.4367331564,-0.0640919358,0.0964713767,-0.1648805439,0.1710062176,0.1025388539,0.0354846083,-0.1941321641,0.4717495739,0.1212736145,-0.009404582,-0.4361022711,-0.3261881769,-0.021046428,0.2483045608,-0.0025580148,-0.0462470613,-0.151955381,-0.1663593352,-0.0428271741,-0.0476744473,-0.0424825996,-0.2063824087,-0.1596759707,-0.2316061109,-0.1420589238,-0.0228530951,-0.0611937232,-0.1075326949,0.3066417575,0.4745171964,0.3733968139,0.0308912918,0.1602347046,-0.1567804515,-0.2718474567,-0.0966227427,-0.3991087675,-0.0479865856,-0.4468705952,0.2575535476,-0.311026752,-0.2784337699,0.3430234492,-0.1163650304,-0.0179087929,0.4083646536,-0.5690831542,-0.0278029591,0.127594009,0.3076961339,0.1703378856,-0.1437247097,0.3931418657,0.2275318652,-0.0822774693,-0.0902018547,-0.1061115414,0.2046216428,-0.1738822907,0.6187602282,-0.1944050044,-0.0548236817,-0.0545141436,0.6473221183,0.3969494998,-0.1100327671,-0.0457413495,-0.2120825797,0.555854857,0.0750215128,-0.0713288561,-0.1281473488,0.2026109695,-0.1250326633,-0.0744449794,0.0493145175,-0.4832840264,-0.1288197935,-0.1193007231,-0.1127936691,-0.1850577891,-0.0140776513,-0.074696593,0.5667656064,0.0115627302,-0.2575372756,-0.5660694242,-0.6033068895,0.0378774181,0.3398478329,0.1747936904,-0.3895838857,-0.3328221738,0.4341337383,-0.4290961623,0.3453397155,-0.2231281698,0.0909887031,-0.1166270375,0.0546613149,0.1256177574,0.1978280246,0.4831276238,-0.4202390015,-0.2678390741,0.0679990575,0.3786558807,0.0961247757,0.2741270363,-0.2173538506,0.3001220822,0.2447852045,-0.02580221,-0.1339326799,0.2969381511,0.5221366286,-0.2260958552,-0.3237432837,-0.2412983179,-0.2852934301,0.0302050281,0.0038252654,-0.0078397561,-0.3472956419,-0.041092854,0.09232077,0.1792479455,-0.1938165575,0.3915732801,0.3434727788,-0.3964323103,-0.0315436162,0.5583869219,0.2310682833,0.2091256678,0.3514365256,0.4544342458,0.504378736,-0.129390955,-0.5598922968,-0.1983911693,-0.4881317019,0.4070118964,0.4231113791,0.3770060837,0.3981897831,0.195366472,-0.0679779723,0.0634853691,0.1406250894,-0.1637904197,-0.047614038,-0.3536589444,-0.3614882827,0.1960397512,-0.0669640452,-0.2215824872,0.3081970513,0.0110019427,-0.1212780699,0.4656127095,0.0476928093,1.238451004,0.1484943181,0.2753532231,-0.2489492297,-0.2222335488,0.3537950814,0.0325664133,0.1756271869,-0.0171394963,-0.0671021417,-0.0676107481,-0.1809380502,0.007624459,-0.0023860307,0.0968144611,-0.1408824921,-0.244505614,0.0033184292,-0.0653605089,0.3175387084,0.2996050715,-0.0296012368,0.4043979645,-0.0052492958,0.2315820754,-0.2421016842,-0.0196749624,-0.2431137115,-0.213837117,-0.1342335194,-0.0893028304,0.068193011,0.1340523958,-0.3112969398,0.1171625257,-0.447260201,0.3500976861,0.3386390805,0.4484483004,0.1044705808,-0.1836323142,0.0932413265,0.0463461839,0.1008624136,0.3841872215,-0.0712479129,-0.2078350037,0.089288041,-0.3228288889,0.4911769629,-0.0261771437,-0.5840344429,-0.2656639218,-0.2430948466,0.2465312928,-0.3386537135,0.410841912,0.3529491425,-0.2228950262,-0.0658854395,0.0265276991,0.2364750654,0.0466577411,0.1349187642,0.1815422475,-0.0505538285,-0.0214173384,0.4060080647,0.2141217291,-0.3812195361,0.3060287535,-0.1139824465,-0.1865628362,-0.0384655073,0.0719917119,0.1743496954,-0.4641196728,-0.2478681207,0.189018786,0.6258869767,-0.2097504437,0.0078869518,-0.0960952044,0.3676452637,-0.3700941205,-0.1063323542,-0.2826554179,0.2064353228,-0.1157424748,0.0402161852,0.0490543805,0.0395378508,-0.0438173003,0.0087191509,-0.2136474848,-0.1027079672,0.10558144,0.1609470546,0.0182482116,0.3611710668,0.3024438322,-0.4145246446,-0.0036734235,-0.4310128987,0.1613376588,-0.0555240996,-0.0313927196,0.1859136969,0.1197883785,-0.3572769761,-0.2396789789,-0.3576963842,0.0524080992,0.1926746964,0.0640170649,0.1225618273,-0.1955527663,-0.5717161894,0.1367693841,0.1894464046,-0.2239125967,-0.1017910838,0.1240301132,0.3057495356,-0.1502554864,0.0802023709,-0.252109617,0.0395031832,0.3099381328,0.1254525036,0.3709874153,0.2314793319,0.1872221529,-0.1610586792,0.2410395741,-0.061075341,0.0096612433,-0.0496630706,-0.4228898585,0.159531638,0.3035745919,0.1041655913,-0.3014527857,-0.0901671499,-0.0021961667,0.4176018536,0.1107093468,0.290423274,0.2070523351,-0.1417419314,-0.0988257527,-0.0486247875,-0.0972351879,-0.3975598812,-0.0232782364,0.092094779,0.0603303313,0.1328025311,0.2289080024,0.5250607133,-0.2251092046,0.0746501535,0.1632907242,0.1939112395,0.2640430331,-0.0256815776,-0.0109260948,0.1031349078,0.2265006453,0.0186885633,0.3572964072,0.3870846331,0.2988618016,-0.0848570839,0.2265387923,0.4148775041,0.019761743,0.2069386095,0.2649187744,0.0552266873,-0.4714702666,-0.0823117048,0.0401577726,0.0133798821,0.1034477651,-0.2998307943,-0.0817496628,0.162090376,-0.0102200946,-0.1705797315,0.1244451329,0.2051188946,0.2919789553,-0.2277558893,-0.0460027717,0.069398433,0.1477602869,0.4056123793,0.754799068,-0.0657600313,-0.0475356728,0.2017750889,0.4158101678,0.0178022776,-0.0005626645,0.0580632649,0.0738010257,0.3980828822,-0.1837501973,-0.1494775265,-0.0211104099,0.0649081543,0.4278395176,0.1346671134,0.0423235372,-0.2120675445,0.2639534473,-0.0169524867,-0.168698594,-0.1762657613,-0.1584697217,0.095615387,0.0273567662,0.5971824527,-0.1300956607,-0.1346501261,0.0912899002,-0.0000076541,-0.056462463,-0.0904306397,0.1892032474,0.140395537,-0.2802189589,-0.2236589044,-0.0279042386,0.3802796006,-0.1861335486,-0.2230957597,-0.332660377,-0.0618388355,0.3705800176,0.0530229658,0.0935091004,0.0273806732,-0.3798119724,0.4845603704,-0.1119275317,-0.1011023,-0.2941746712,0.3170862496,-0.0849195421,0.1796787977,0.1528979838,0.3086315989,-0.0950192064,-0.1213019639,-0.0762388036,-0.1598168612,0.3058774173,0.2785101533,-0.1957733333,0.3706883788,-0.1379019022,-0.2090939134,0.2905963659,-0.0943000093,-0.3691655397,0.2377327383,0.1939804405,-0.3000845611,-0.2975242138,-0.0845024064,0.3556576669,-0.0637442619,-0.2942965627,-0.0366496257,-0.3430096507,-0.0321009904,-0.1096607372,-0.0095350258,0.0639310628,-0.0277892556,0.0190485716,-0.0419584289,-0.1561746597,0.0116826938,0.2327777445,-0.5458312631,0.3114074171,-0.0218565483,0.4344706237,-0.4064275622,-0.3519240022,-0.3572448194,-0.0659033656,0.2870924473,-0.044731576,0.0349989645,-0.2243577242,-0.0411719643,0.0790597871,-0.1507356018,0.3140515089,-0.0783637613,-0.0839317441,-0.1645740867,-0.1246090457]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1924","title":"Anonymous Dataset Addition (i.e Anonymous PR?)","comments":"Sure, I think it's ok on our side","body":"Hello,\r\nThanks a lot for your librairy.\r\nWe plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ? \r\nCheers \r\n@eusip","comment_length":8,"text":"Anonymous Dataset Addition (i.e Anonymous PR?)\nHello,\r\nThanks a lot for your librairy.\r\nWe plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ? \r\nCheers \r\n@eusip\nSure, I think it's ok on our side","embeddings":[-0.247134909,0.3878291249,-0.0250061769,-0.0712201297,-0.1435951293,-0.144092679,0.4446018338,0.0007834382,0.0178448521,0.0311326161,-0.0967483297,0.0530138761,0.0171098281,0.1391016543,0.0187266227,0.1147655025,0.0662259683,-0.1187853739,0.1887228191,-0.253136307,-0.5374703407,-0.2428633422,0.0495522469,-0.5661013126,0.1603137404,-0.2404482663,-0.196773082,-0.1826818138,0.0294838827,-0.4255110025,-0.2858841717,0.2867761254,0.068298988,0.2416846156,-0.0001196361,0.1164926365,0.233651638,-0.0246826448,-0.2259674668,-0.1233626902,-0.4074278772,-0.0922171101,-0.0592250489,-0.1801940948,0.010512827,-0.3904939294,0.1158190146,-0.4062104225,-0.3286239803,0.0273723565,0.1746528298,0.190651983,0.0339517146,-0.0990606174,0.1273151785,0.338211745,0.029196376,-0.0319917798,-0.2435337007,0.2782224715,0.1297217906,0.0919820517,0.2623674572,-0.365986377,0.1682933867,0.2013637722,0.1694687903,-0.0137226982,0.0875010192,0.0093896184,0.6295799613,0.0846999809,-0.2952815592,0.0398822017,0.1128078476,-0.142457217,0.0903782323,0.3977229595,-0.2013644129,0.0601423681,-0.441927731,-0.5339925289,-0.1371515393,0.2333774716,0.431704253,-0.0828318372,0.2713038027,0.2172265649,-0.169267118,0.0472210944,-0.0588965751,-0.15275608,-0.2825243473,0.003796248,0.280657351,-0.0067813243,0.104036741,0.0656816363,-0.0659599006,0.1678720266,0.0995775312,-0.2339675575,-0.2842834294,-0.1019187942,0.1794334501,-0.2871128619,-0.0772777498,0.2998698354,0.1184046566,-0.150330916,0.4833081067,-0.0480022021,0.1742841303,0.205271244,-0.1832642108,-0.4568749964,0.074585177,-0.2460968643,-0.0260456465,0.0774206072,-0.2003284544,0.0934754387,-0.3263406456,0.1827337742,0.0560162216,-0.3611070812,0.3267019689,-0.1611869484,-0.0312204976,-0.0254487284,0.0031735429,-0.0831116736,-0.2269147933,-0.0916290283,0.0670697242,0.1837407202,0.0601073466,0.044126451,0.3239905536,0.2723436058,0.2890750468,0.0445964001,0.2387387455,0.0594689958,0.1178275943,-0.2840260863,-0.1034773961,-0.573319912,0.0649437681,0.3512658775,0.327886939,0.2998686135,-0.2512911856,0.1549491435,-0.1931495368,-0.2523591518,0.2112556547,0.2580756545,-0.0669518188,-0.1334109753,0.2852809727,0.242818594,-0.0345345028,-0.097720556,-0.045740962,0.3926412463,-0.1981772035,-0.2074654251,0.1860650331,-0.2115226388,-0.1140638068,0.0807060823,-0.2520679235,-0.1588671803,-0.0562465414,-0.3016118705,-0.0020913237,-0.1984258145,-0.4574971497,-0.2800042629,-0.250442028,-0.2061260492,0.4397351742,0.6227098107,0.1916397065,-0.1582782269,-0.2280854434,-0.0420787521,-0.0692400411,-0.2434900403,-0.4580893219,-0.0662558749,0.1408385336,0.0295751616,-0.2851476073,0.3538032174,0.4317177832,-0.1706714779,0.3868232071,-0.3253704011,0.1264896393,0.1295206398,0.6475499272,-0.1005043462,0.1545937657,0.367025286,-0.1393336505,-0.2277813852,-0.360243082,0.3657832146,0.1392658353,-0.5584985018,0.3219586015,-0.1826287061,0.0459738784,-0.1592522413,-0.0148743046,0.275764972,-0.3737313449,-0.1763480157,-0.087414071,0.2921238542,0.0117968814,0.025107367,-0.316500783,-0.1514278203,0.1306669414,0.0948107541,-0.2174645215,-0.0581727363,-0.0894684494,-0.13367486,0.1499872804,0.1305726469,-0.1509994119,0.015549548,0.6848812103,0.3416575789,-0.0386318453,-0.3269557953,0.1385549456,0.0975708514,-0.0833190382,0.1653500646,-0.6286693215,0.4655435383,-0.0478317291,-0.0755471587,-0.1237164587,0.1607293487,0.339654088,-0.1958124638,-0.0243449025,-0.1682256907,0.3509458303,0.5593212843,0.2586764991,-0.1044109538,-0.4522874057,0.2192164063,-0.5436038971,-0.0453082845,-0.0200441685,-0.0394814089,-0.1596883982,0.2102915347,0.0297145788,0.1423166543,0.223809287,0.3619076312,-0.1660782844,0.1804982424,0.2240943313,-0.0667850971,0.2089779228,-0.0576311089,-0.3477261662,-0.1378743798,0.442809999,0.1364219189,-0.0443148874,0.0657370538,0.3422340155,0.1607662141,-0.0806729198,-0.0138315773,0.312450856,-0.0932484344,0.0751168579,0.2070452571,-0.2428253293,0.0068052639,0.2679630816,0.0638475567,-0.1182503924,0.2193146646,-0.3655565083,0.7769959569,-0.4439932704,0.0011095647,0.1311145127,-0.2073430419,0.0937609747,0.1581333429,0.1004305854,-0.2936827242,0.530751586,0.1009925827,0.0495381169,-0.5021622181,-0.3919862211,-0.0373085849,0.2690249383,0.0120603899,-0.0635631308,-0.2577194273,-0.1695854068,-0.0398253761,-0.1018259525,-0.1053671688,-0.2735604346,-0.1864105612,-0.2025663406,-0.0899034739,-0.0585305654,-0.2621882856,-0.12454772,0.2302824408,0.510273695,0.3811747432,0.0098160971,0.1737980545,-0.1751248091,-0.2425146699,-0.1245711818,-0.3452533484,-0.1522209495,-0.4224831462,0.2079289556,-0.3231139481,-0.3195494413,0.2264228165,-0.1044688821,0.0257215109,0.3909400702,-0.5271616578,-0.0473069437,0.1302896291,0.3801364899,0.1524700224,-0.1292843223,0.430719167,0.2023510486,-0.148413077,-0.122822471,-0.0963207781,0.1783896685,-0.1429001689,0.5568922758,-0.2366476953,-0.0503837243,-0.0583829843,0.6825946569,0.4599709809,-0.190600425,0.0321341939,-0.1758282781,0.5764404535,0.112809211,-0.1212912649,-0.0643527359,0.1485604197,-0.1356455088,-0.0010611439,-0.026607113,-0.5009939671,-0.069979094,-0.1082001925,-0.1828837842,-0.2122738361,0.0376274772,0.0198484566,0.5788437128,0.0731545091,-0.2022375911,-0.5691615939,-0.6170281172,0.0505143367,0.291359812,0.1880910099,-0.3982411623,-0.2764868736,0.4477979541,-0.3822238743,0.4073144197,-0.1923403293,0.1430618018,-0.0960611403,0.0411262661,0.1556975543,0.1777589768,0.5320597291,-0.4465730786,-0.3082463741,0.0906887874,0.267370522,0.1521223038,0.2357754111,-0.2937586904,0.3772712052,0.218202427,-0.0040045185,-0.1257352829,0.2918021977,0.5262829065,-0.2259114832,-0.3145754039,-0.2562882006,-0.3114239573,-0.0007688018,-0.013015396,0.0065737013,-0.3320064247,-0.0510374606,0.0261645298,0.2586666644,-0.2188333571,0.3757217526,0.2784825563,-0.3407128751,-0.019493049,0.5227956176,0.2416499406,0.199187696,0.2875306308,0.4390183985,0.5387373567,-0.1048912928,-0.5265433192,-0.1606948078,-0.4511874616,0.3810036778,0.4330092371,0.3982286751,0.3644188643,0.1823946685,-0.0172405951,0.0252286587,0.136470601,-0.0966661051,-0.0645044148,-0.2686911523,-0.388228476,0.1494316161,-0.0490486808,-0.2483307719,0.292866081,-0.0372418128,-0.1279868484,0.4344697893,0.0914479345,1.1573667526,0.1562568694,0.2694735229,-0.2529120445,-0.2665753961,0.3270263076,-0.0570668355,0.1811705828,-0.0526945554,-0.0759033561,0.0096496874,-0.1688012779,0.0069588008,-0.0839265287,0.0072756629,-0.1184525415,-0.1725475043,-0.0100366613,-0.1159510911,0.365668416,0.2679741085,-0.0664500892,0.3326010108,0.0403640606,0.1878806502,-0.283804208,-0.0325797796,-0.2545093596,-0.1460153908,-0.1383777112,-0.0575865433,0.0286042932,0.1981846988,-0.1883324385,0.0771149173,-0.4420200288,0.3684043586,0.2938214242,0.4399767816,0.0537237711,-0.162436828,0.0685815439,0.046625372,0.082693696,0.2684399784,-0.07972496,-0.1075058728,0.020922916,-0.3490535021,0.469040513,0.0855296329,-0.5520849824,-0.3311719,-0.2070747018,0.2979760766,-0.3116955161,0.5076208115,0.2860383689,-0.3047467172,-0.0944499448,0.0825801194,0.279276818,0.0021736862,0.1837478727,0.257306844,-0.0855762288,-0.0908067599,0.4112284184,0.287938863,-0.3583945632,0.3205325603,-0.0538666844,-0.2063628882,-0.1072256491,0.1216684133,0.1523890048,-0.5358930826,-0.2215095311,0.1651886553,0.5708414316,-0.2428628504,0.081803903,-0.0265789237,0.3054182529,-0.3707421422,-0.0715501159,-0.2899203002,0.1770952046,-0.161389187,0.0713035688,0.0842435062,0.0611530468,-0.0458604991,0.0702017322,-0.2968958318,-0.1036129966,0.0447721854,0.1579910666,-0.0137746632,0.2560126483,0.2832299769,-0.35874933,0.0637806803,-0.3667784035,0.0649034828,-0.1174022555,-0.0109725092,0.1505605131,0.1228832081,-0.3171796799,-0.2434571981,-0.3172389567,0.0134462388,0.2519538403,0.0603128225,0.0894595012,-0.1599818021,-0.509976089,0.06334126,0.0684121996,-0.1157004163,-0.1592136621,0.0693928078,0.2128814012,-0.1084030718,0.1062788516,-0.2780923843,0.0269001499,0.3566057384,0.0973525643,0.3519695103,0.2698073387,0.1139391661,-0.1357005239,0.2257449776,-0.0432752036,0.0847396106,-0.0649368316,-0.4326902032,0.0993225425,0.333717227,0.1619793624,-0.4017883539,-0.0926814526,-0.0308312047,0.4182392657,0.1350175291,0.3176068664,0.2352963388,-0.0629673898,-0.0599842146,0.0038229758,-0.0988468677,-0.3940961957,0.0139084021,0.0470186956,0.0110542616,0.1054485068,0.3172875345,0.5286137462,-0.2010546625,0.1163037047,0.125619188,0.2374517769,0.1373951882,-0.0632128268,0.0248543303,0.0816247091,0.2765976489,-0.0231217369,0.341678828,0.432620883,0.294072181,0.0562957115,0.2233486027,0.3975689411,-0.0734668225,0.2213743329,0.1830568612,0.068246074,-0.3777292073,-0.1015120819,0.0128558753,0.0412384272,0.0933644846,-0.1653019786,0.0142817944,0.1538720876,-0.0857991129,-0.1657959521,0.1501193494,0.1704204977,0.2899797559,-0.27199471,-0.0108243311,0.1304588467,0.0977789909,0.3695704937,0.7195887566,-0.0525287613,-0.068455711,0.1315503865,0.4088155031,0.0895131156,0.0067292126,0.0176601037,0.0793750733,0.3860146701,-0.1518455446,-0.1179470271,0.0401255265,-0.0038713852,0.4501042068,0.1512707472,-0.0524081104,-0.2787723541,0.249772653,0.0238135345,-0.1770060062,-0.1516076475,-0.2316629887,0.0219865851,-0.0506060198,0.5310348868,-0.0871829391,-0.1123189181,-0.0082122907,0.0478252545,-0.0721098334,0.0696004555,0.0867989883,0.1493126303,-0.2463832498,-0.1516132802,-0.0874263868,0.5023704171,-0.0902161524,-0.2187040895,-0.2808861434,-0.0138585167,0.4165892601,0.1398543119,0.2429224551,0.0499764867,-0.3009789586,0.4355744421,-0.2064579874,-0.0875840038,-0.1942467541,0.2704803646,0.0097617051,0.1988959908,0.1922329068,0.2320564836,-0.0094159488,-0.1932210922,-0.0893390477,-0.0447755903,0.3605258167,0.3338338733,-0.180049181,0.3943703175,-0.1311672926,-0.204353258,0.226256907,-0.1279355586,-0.383343935,0.2104909867,0.1950385571,-0.2814119458,-0.3043899834,-0.1843663007,0.2522423267,-0.009772134,-0.2593125105,0.0274275038,-0.2086183578,-0.0295853075,-0.0977472514,-0.0516777113,0.0586770214,-0.0450552814,0.0177550875,-0.0511657894,-0.1895511299,-0.0678883493,0.3175985813,-0.523860693,0.2509985864,-0.0274012201,0.3858340383,-0.3570177257,-0.3705475628,-0.3947063684,0.0158088598,0.2980502844,-0.0566373803,0.0420756303,-0.1655331403,0.0158264283,0.0540001914,-0.1318888664,0.3929895163,-0.1209091991,-0.08490742,-0.1144813076,-0.1260319203]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1924","title":"Anonymous Dataset Addition (i.e Anonymous PR?)","comments":"Yup, sounds good!","body":"Hello,\r\nThanks a lot for your librairy.\r\nWe plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ? \r\nCheers \r\n@eusip","comment_length":3,"text":"Anonymous Dataset Addition (i.e Anonymous PR?)\nHello,\r\nThanks a lot for your librairy.\r\nWe plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ? \r\nCheers \r\n@eusip\nYup, sounds good!","embeddings":[-0.2626509368,0.4015284181,-0.0504989438,-0.1253037751,-0.1580169946,-0.1648795456,0.4638371766,0.0262678266,0.0307678767,0.0466941595,-0.1124388948,0.073402144,-0.0135629633,0.1564850211,0.0422838107,0.1102689356,0.063233234,-0.0892885029,0.168613106,-0.2312205434,-0.5188638568,-0.2539670169,0.0544439889,-0.5515534282,0.159351632,-0.2366060466,-0.1894864887,-0.1624655128,0.0295834932,-0.4324190021,-0.3030962646,0.2972929776,0.0548268966,0.2280926257,-0.0001130332,0.1126938686,0.1965181231,-0.0243414044,-0.244378984,-0.1077605188,-0.3854706883,-0.0911417454,-0.0346736275,-0.1985535324,-0.0100034624,-0.3882982433,0.1392218769,-0.3787285089,-0.3069884181,0.006885624,0.2241524607,0.2046720088,0.0026336934,-0.1582001299,0.1236249283,0.2979676723,0.0167796873,-0.014691446,-0.2402915657,0.2690574825,0.1002743989,0.1449889392,0.263328284,-0.3236764073,0.1512383521,0.179003045,0.1770873964,0.0021325848,0.0565300211,0.0149978893,0.5649096966,0.0745220035,-0.2905502915,0.0005503514,0.0985778794,-0.1359340847,0.0514273904,0.3862766623,-0.2116685808,0.0652322471,-0.4457839429,-0.4874722958,-0.1370580643,0.2025703937,0.3688219488,-0.0840114355,0.2220504582,0.153909713,-0.1241599396,0.0496863388,-0.0596240126,-0.0995288268,-0.2601954639,-0.0111142993,0.2734091282,-0.0072588651,0.1735072881,0.0657663494,-0.08281596,0.1688817143,0.1186835393,-0.2116629183,-0.2837513387,-0.0677249581,0.1750847101,-0.2858922184,-0.1035743281,0.2591291666,0.1410415918,-0.1462766528,0.4627949893,-0.0350351632,0.1618887931,0.2208849341,-0.2041787356,-0.4643006325,0.0882523209,-0.2290827036,-0.0291971415,0.0810779855,-0.1612815857,0.107927449,-0.3247381449,0.2256261259,0.0408777706,-0.340998441,0.2974004447,-0.1780753136,-0.0597767793,-0.0133431964,-0.0263836067,-0.0787431747,-0.2322201729,-0.0844304264,0.072285153,0.178096056,0.0806065947,0.0427377149,0.321311444,0.2581573129,0.3044337034,0.068255432,0.2249989808,0.0738726556,0.0804393589,-0.2920084,-0.1006323025,-0.5406374335,0.045285888,0.2788287699,0.340888381,0.2855674922,-0.2167698145,0.2154028416,-0.2026088238,-0.2664228678,0.2376119345,0.3178715706,-0.1444329172,-0.1577579081,0.2228542268,0.2445729971,0.0254207719,-0.0913185775,-0.0203366634,0.3662511408,-0.1719528139,-0.2039457709,0.1768949628,-0.2215764821,-0.154390499,0.052149497,-0.26150617,-0.1755541712,-0.0487520359,-0.2875377238,-0.0050121471,-0.1941390187,-0.4179427028,-0.2913920879,-0.2553350031,-0.1943253577,0.4336124063,0.5731931925,0.2078680694,-0.2156210989,-0.2248894572,-0.0701590255,-0.0825287998,-0.2014776319,-0.4620446563,-0.0632349625,0.0713890567,0.0844863728,-0.2832010984,0.3337907195,0.4241825342,-0.1730530113,0.3987813592,-0.3498981893,0.0919773504,0.0800955147,0.6248120666,-0.0811459571,0.1042848676,0.3133194745,-0.111247763,-0.2428982407,-0.3812782764,0.3342036307,0.1665903628,-0.5542097092,0.2989407182,-0.1906853914,0.0506214425,-0.1982035041,0.0610587336,0.2946097851,-0.371029824,-0.1414901912,-0.0872526467,0.2500436604,0.0126340454,-0.0001580327,-0.2916484773,-0.1538148522,0.125850603,0.0850344822,-0.1699796766,-0.0559231155,-0.0998217016,-0.1384304464,0.1939631402,0.1460846066,-0.148952961,0.0366800912,0.6731448174,0.3345088661,-0.0503468178,-0.2976899147,0.1260708719,0.1104531437,-0.0841835737,0.2169567496,-0.6657384038,0.4782570601,0.0220829155,-0.1029118001,-0.0810836554,0.1690955907,0.3406691551,-0.215533942,-0.0326123983,-0.1927470863,0.3212936223,0.5436180234,0.2417767346,-0.0888056681,-0.4803644419,0.2673237622,-0.5183224678,-0.0096025476,0.0070798676,-0.0585717969,-0.1732994318,0.2002928853,0.0441349223,0.1664008349,0.1980833858,0.3949071467,-0.1411576867,0.1498659402,0.171323508,-0.090189971,0.2036939263,-0.0628557429,-0.3252154291,-0.1450202465,0.4354495704,0.1035651416,-0.0554929152,0.045721896,0.3011760116,0.0967799723,-0.0532708094,-0.0050148787,0.2992325723,-0.1026296616,0.0948285162,0.2024654597,-0.2052326202,-0.017673403,0.3023852408,0.0992249399,-0.1923673302,0.2217485011,-0.3797129095,0.7704243064,-0.473354429,0.0549358651,0.1412692815,-0.2151580304,0.0992140099,0.2206952423,0.0686642826,-0.3022278249,0.5763546824,0.1223662049,0.0820018947,-0.5111482739,-0.3933106065,-0.0405384377,0.2430749387,0.0115547171,-0.0458564125,-0.2571169734,-0.1637569219,-0.0620063581,-0.0999426097,-0.1152975783,-0.2693549097,-0.2195746899,-0.1852113158,-0.064544633,-0.0638465658,-0.3020082712,-0.100309968,0.1677125096,0.5203088522,0.3900313675,0.0122500369,0.1704714149,-0.1208104268,-0.1833198518,-0.0962313488,-0.3090976179,-0.1812371761,-0.4057359099,0.2095426023,-0.3691759408,-0.3100214899,0.1896382421,-0.1187938228,-0.0116595672,0.3775458038,-0.5130351782,-0.0811709017,0.1364114583,0.3640540838,0.185716778,-0.1566789299,0.4573351145,0.2218340635,-0.2325732112,-0.1578566879,-0.0662488565,0.1560688764,-0.1522312462,0.5086657405,-0.2719679773,-0.1186367571,-0.0529238172,0.6839931011,0.4052784741,-0.1630107313,0.0290854294,-0.1682232916,0.5643977523,0.0976184532,-0.0884746686,-0.0632352829,0.1119738519,-0.1005267948,0.0426419489,0.0075463811,-0.538151741,-0.0633718818,-0.1101437062,-0.1887186617,-0.2265864909,0.0465617888,0.0473604202,0.571449399,0.0818355978,-0.1914182901,-0.5849758387,-0.5647132397,0.040078748,0.2324153185,0.1933318079,-0.4350435436,-0.2546269596,0.4843989015,-0.359169364,0.3859283328,-0.1873663962,0.106593892,-0.0992764682,-0.0081645949,0.1339827478,0.1548752785,0.4749927819,-0.4148231745,-0.2910833657,0.1095021144,0.2510960102,0.1284506768,0.2394743264,-0.301122427,0.3459842503,0.194961369,-0.0030739827,-0.1031772122,0.2667869627,0.5545663834,-0.2308452278,-0.3155401945,-0.2256566435,-0.2788955867,-0.0199697502,-0.0043713138,0.0305044558,-0.3120690584,-0.0529095381,0.0046109268,0.2704331577,-0.2043247968,0.3772785366,0.2700770795,-0.2986817956,-0.0470155329,0.4869456589,0.2244849652,0.1947565973,0.2719412148,0.4356605113,0.5252122283,-0.0861665085,-0.5235154033,-0.1108068675,-0.4205373228,0.336116612,0.4640169144,0.3988571167,0.3554478884,0.1476940513,0.01616662,0.0082760332,0.1150138304,-0.0959034413,-0.1138095856,-0.221081242,-0.3862004876,0.1429820061,-0.0364853032,-0.2778068483,0.2821336389,-0.0285670236,-0.1170628145,0.4474495351,0.0942902863,1.1027923822,0.1788707078,0.256841749,-0.2424026132,-0.2895278633,0.2940419912,-0.1005280465,0.19741638,-0.050279405,-0.0916116312,0.0353956111,-0.1291717589,0.0198140852,-0.083117187,0.0100812269,-0.1390423477,-0.1446795166,-0.015185914,-0.0841536447,0.3481968939,0.2337446809,-0.0812218711,0.2933125496,0.1113772392,0.2062716335,-0.333451122,-0.0155706517,-0.2543244064,-0.1217950061,-0.0805265456,-0.0468146913,0.0439436026,0.2161108255,-0.1884003729,0.0408884175,-0.4244702458,0.3078849018,0.2919597924,0.3972119391,0.040149156,-0.1710205823,0.0776836649,0.0508346297,0.0822169408,0.2467938364,-0.0814787298,-0.0433701165,-0.0142996935,-0.321733892,0.48044765,0.0625113249,-0.5177263021,-0.3387183547,-0.1637647152,0.3225481808,-0.3368664384,0.5551453233,0.2757951319,-0.3263565004,-0.1401033849,0.1335932016,0.2494939119,0.0065757651,0.1703778356,0.2498508394,-0.0706915855,-0.1061946303,0.414172858,0.2776949406,-0.3212720454,0.2818195522,-0.074882634,-0.2260642946,-0.1720010638,0.1243650317,0.1351262629,-0.5310297608,-0.1638185978,0.1570721269,0.5056236386,-0.1942727715,0.1081710309,0.0180105083,0.2657480538,-0.3905247152,-0.0357779562,-0.2612576783,0.1759454757,-0.1687119305,0.0991804749,0.0595523417,0.0671048686,-0.03329622,0.0476428047,-0.3746834397,-0.0988369584,0.008970933,0.1755702347,-0.0501974784,0.1933920234,0.2740625143,-0.3227054775,0.1358535737,-0.3485456109,0.0290474482,-0.1773178726,-0.0319323279,0.1218788773,0.0732395202,-0.3082714379,-0.2044844925,-0.2319301218,0.0127424095,0.2600525022,0.0867933109,0.0728392676,-0.1544577479,-0.5035931468,0.055860281,0.0665636882,-0.0512345955,-0.1558870226,0.0581975058,0.2196475863,-0.1181579009,0.0589682385,-0.3189415038,-0.0010190782,0.365998894,0.1214030907,0.351243794,0.2435104847,0.1409765184,-0.0701366588,0.2257468849,-0.0370733365,0.115001902,-0.0720778629,-0.4241215885,0.1222152337,0.3066512346,0.220068872,-0.4064371884,-0.1045260057,-0.0292294342,0.3752980232,0.1751669645,0.323317945,0.2628151774,-0.0710748285,-0.0347139984,0.0161557794,-0.0774525702,-0.3805306554,0.0285727438,0.0316989161,-0.0014235082,0.0685844049,0.3267396092,0.4974457622,-0.1626035869,0.1247773394,0.0851913542,0.2355245948,0.1178890765,-0.0561846159,0.0429606922,0.0936092511,0.3211867213,-0.0411869958,0.3021328449,0.4416776001,0.2359760851,0.0710136816,0.1681247056,0.4049632847,-0.1139232591,0.2302694768,0.2110021114,0.0569325835,-0.3669137657,-0.1209518611,0.0197701417,0.0266999844,0.104148373,-0.1632339358,0.0477570668,0.1515251398,-0.0940832123,-0.1893306971,0.1463987529,0.1243217587,0.3203229606,-0.265688926,0.0091423783,0.166310817,0.0842251256,0.3762978017,0.7163056135,-0.0498112813,-0.1155114174,0.1107793972,0.3845354319,0.0776188001,-0.0198165085,-0.0111767696,0.0534567237,0.3691917658,-0.1123560518,-0.0864349529,0.1064715087,-0.0654347986,0.4662733674,0.1303942949,-0.0798005164,-0.2929105461,0.2131121308,0.0453450009,-0.188920483,-0.1528919935,-0.2329912633,-0.0048643732,-0.0975584537,0.4986766875,-0.1148471981,-0.1141225174,-0.0252965409,0.086361587,-0.0390995257,0.1097822264,0.0254601929,0.1391851604,-0.2015144229,-0.1651790738,-0.0814534575,0.4965391457,-0.0696383566,-0.2107804865,-0.2908932865,0.019165583,0.3946292698,0.1838238388,0.2034607083,0.0424308442,-0.2754522562,0.4290665388,-0.2242332697,-0.0492458493,-0.1554783583,0.2718611658,0.0342986621,0.1706387103,0.1736121178,0.1905906349,0.069239907,-0.1530960053,-0.0951768607,-0.0238672011,0.354970783,0.3473921716,-0.1838816553,0.3876902759,-0.1265530437,-0.2361266762,0.2022468448,-0.10822846,-0.370244801,0.1631869823,0.2210122198,-0.2686350346,-0.2921260595,-0.2127028108,0.2675874233,0.024054857,-0.2199881673,0.0837324634,-0.1928919852,-0.0391905122,-0.0634881407,-0.0352406688,0.0660374835,-0.0245186482,-0.0204746444,-0.0700689107,-0.1810761839,-0.1173243821,0.3328267336,-0.5002976656,0.2323708683,-0.0207689013,0.3811979294,-0.3691374063,-0.3220351934,-0.3721812963,0.052908875,0.2962576747,-0.0563212074,0.0427122451,-0.1542441249,0.0007027978,0.0263939928,-0.1371687949,0.3887121975,-0.1335439831,-0.0700125843,-0.1263689101,-0.149820745]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1922","title":"How to update the \"wino_bias\" dataset","comments":"Hi @JieyuZhao !\r\n\r\nYou can edit the dataset card of wino_bias to update the URL via a Pull Request. This would be really appreciated :)\r\n\r\nThe dataset card is the README.md file you can find at https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wino_bias\r\nAlso the homepage url is also mentioned in the wino_bias.py so feel free to update it there as well.\r\n\r\nYou can create a Pull Request directly from the github interface by editing the files you want and submit a PR, or from a local clone of the repository.\r\n\r\nThanks for noticing !","body":"Hi all,\r\n\r\nThanks for the efforts to collect all the datasets! But I think there is a problem with the wino_bias dataset. The current link is not correct. How can I update that?\r\n\r\nThanks!","comment_length":89,"text":"How to update the \"wino_bias\" dataset\nHi all,\r\n\r\nThanks for the efforts to collect all the datasets! But I think there is a problem with the wino_bias dataset. The current link is not correct. How can I update that?\r\n\r\nThanks!\nHi @JieyuZhao !\r\n\r\nYou can edit the dataset card of wino_bias to update the URL via a Pull Request. This would be really appreciated :)\r\n\r\nThe dataset card is the README.md file you can find at https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wino_bias\r\nAlso the homepage url is also mentioned in the wino_bias.py so feel free to update it there as well.\r\n\r\nYou can create a Pull Request directly from the github interface by editing the files you want and submit a PR, or from a local clone of the repository.\r\n\r\nThanks for noticing !","embeddings":[-0.3595491946,0.1663745791,-0.0910964832,0.1053340286,0.0010983016,0.1791955829,-0.0393261202,0.0637117624,0.0622652695,-0.1287824512,-0.2621755898,-0.0001704918,0.2578838766,0.1593094617,0.2776818871,-0.3877265155,0.1113262102,0.1100791767,0.1167294681,-0.2110313922,-0.309186995,-0.0822069719,-0.2789872587,0.0970853269,-0.0021148934,-0.0778445676,-0.5115855932,0.2775803506,-0.270675987,-0.4728704989,0.3751768172,0.4130046964,-0.0921292379,0.3960892856,-0.0001064684,-0.2544566095,0.4334570467,0.117040202,-0.151261121,-0.0557776429,-0.1133984029,-0.185517177,-0.3816221058,-0.1367891133,0.0211595744,-0.1777675748,0.1367512643,0.1966270208,0.1025712192,0.4926286042,0.2142109126,0.2089529485,0.3577846587,-0.1554494798,0.3749152422,0.2227902114,0.2564052641,0.3468582332,0.0206908993,-0.1691053063,0.1622935236,0.2506525218,-0.1467977911,-0.1894206107,-0.0663957,-0.0445823632,0.0470025316,-0.1570632309,0.1275359094,0.1428323239,0.154863894,-0.0296167023,-0.3587404191,-0.0778496116,0.1910637617,-0.7335625291,0.165397957,0.0577254631,0.1552003771,-0.0117981331,0.1111088023,-0.518104434,-0.1214332953,0.3081212044,0.1125403568,0.1010735631,-0.0518354401,0.1025832146,0.1344674975,0.021949796,0.3273358345,0.0410104468,-0.2244393677,-0.0568372905,-0.2383382022,-0.1983612031,-0.1284375042,0.4627984762,0.1581492424,-0.0870212168,-0.2016758025,-0.0480877534,-0.2221380323,-0.160610646,0.3178705871,-0.0264144875,-0.139250353,-0.1350999027,0.1666113138,0.1985975355,-0.0380894132,0.086371921,-0.0215295386,-0.349781245,-0.1778565645,-0.0127986157,0.2721118331,-0.0768700242,-0.3750704527,-0.0947862789,0.0259466898,-0.1254574507,0.0305153821,0.0824875981,-0.1555774212,0.2190558761,-0.0637154728,0.1419501454,-0.2471201718,-0.4623448253,-0.2801394761,0.0670003891,-0.2803134024,-0.0033133586,0.1079281121,-0.1923789382,0.4917120934,-0.0540024936,-0.1442541778,-0.0563039929,-0.2679831088,0.0572684705,-0.0964815021,0.0256993696,-0.0513382815,0.0214462988,0.0096340887,0.2765556872,-0.0625869259,0.1380599141,-0.382538259,-0.2012334913,-0.0192093402,0.0240553021,-0.0979103968,-0.3637704551,-0.3967584372,0.2190155387,-0.1084562987,-0.0248063877,0.0903782323,-0.083959572,-0.325486511,-0.1047406942,0.3892816901,0.4097493589,0.0005988537,0.2222694606,-0.0285612568,0.0435792096,0.0198257845,0.1556759924,-0.1559701562,0.0572348237,-0.0255526211,-0.2854391038,0.2380369902,-0.4642197192,-0.4162271023,-0.2647117674,-0.2668766677,-0.3149490654,-0.2621735334,0.1892431974,-0.2789892256,0.0009048752,-0.1267001629,0.4232994914,-0.1128286645,-0.0350897089,-0.2599846125,-0.13261199,-0.1460773647,0.1583602279,0.0443662256,0.1021287441,-0.0254891552,-0.0300826151,0.4650369287,0.1059122086,-0.0444547124,0.0974038318,0.4143883884,0.1672912985,0.0513468683,-0.0249313638,-0.1223677546,-0.1004391462,0.0400929749,-0.0212420598,0.325835228,-0.097712025,-0.4292385578,0.0725061521,-0.0063626524,-0.1980825663,0.1160103604,0.032533098,0.1374223381,0.1323659271,-0.1745865196,-0.051063668,-0.1280091703,0.226971373,0.201635465,0.3748253882,-0.144493565,-0.0283053089,0.4443071485,0.2948619425,0.1618736684,-0.0843273774,-0.1209082156,0.5419946909,-0.1632207036,0.1217008978,0.482005626,0.25588274,0.2863232791,-0.5517087579,0.1105966493,0.1785252541,-0.1446702033,0.3147725761,-0.1308719516,0.1476735622,-0.2084826827,0.0985418633,-0.3138096035,0.4175610244,0.2844434977,-0.0370110124,0.1749881357,0.2364445925,-0.0953710377,-0.0353623927,0.0069961031,-0.2239562571,-0.3241182566,0.3490578532,0.10750328,-0.0751613975,-0.1951504648,0.0845110044,-0.4713924229,-0.18563582,-0.1845969707,0.12506558,0.2852442861,0.0596687049,0.0369316712,0.1783207804,0.0880993828,-0.1388009042,0.2750367522,0.253480643,-0.1154989824,-0.1864898056,0.1920571923,0.1174200997,-0.270031482,0.250079602,0.10697864,0.3870291114,-0.1194834039,-0.0423878431,0.10714975,-0.2513252795,-0.1118524969,-0.2089509666,-0.3875476122,-0.250233978,0.126520291,0.2735853493,-0.0377792493,0.323630482,-0.097894758,0.4073313475,0.2222720981,0.2817659676,-0.0579492003,-0.1924304664,-0.0709382668,-0.0261849798,0.1622069478,0.0077293878,0.1705101728,-0.3521750569,0.3545436263,-0.501827538,-0.2015229911,-0.0719844997,0.0605509207,0.4237366915,0.1837680042,0.0672684982,-0.0619829632,-0.0454567485,0.2018213719,-0.5429760814,0.1377966851,-0.1011143848,0.0560781769,0.1222836599,0.2462984025,-0.2002692819,0.0757783279,-0.1129235327,-0.0110854274,0.0062888195,0.2626053989,0.083349444,0.0803580731,-0.1830531061,-0.1386593133,-0.1123519912,-0.2118853331,-0.5318339467,0.294545114,-0.3595643342,-0.4036683738,0.0698623732,-0.0449920334,0.129984349,-0.2164371461,-0.4553367794,-0.181292206,-0.1102129892,-0.0260291696,-0.0176295452,0.2905609906,0.2023494542,-0.2502673268,-0.2868277729,-0.0613818429,-0.2773211002,-0.0220623054,0.1599816978,0.4155290425,0.2630163729,-0.1384329945,-0.0856813565,0.7499744296,0.1584288925,-0.3041239679,0.2625769973,-0.1049516127,0.3394186497,0.2625924051,-0.4821255803,0.0951007381,0.0758165643,0.3618260026,0.4159470201,0.2147670984,-0.0936452374,-0.1254832894,-0.4458161294,-0.3002961576,-0.349491179,-0.0799214244,-0.2334310412,0.2280255407,0.3162684143,0.0644520968,-0.3139523268,-0.3706806302,0.4793214202,0.2478992939,0.0487987995,-0.0248303805,-0.0488768108,-0.1790971756,-0.2587465048,0.100715965,-0.3959196508,0.3303524852,0.155930981,0.2400741577,0.1629255563,-0.0136303222,0.5150212049,-0.3189906478,0.21450679,0.0065915803,0.4062411189,0.0342263617,0.0536104925,-0.1371552646,0.4261339903,0.3031489551,0.0954746976,-0.0536196753,-0.3796596825,0.5107845664,0.0088636791,-0.3696417511,-0.1735326499,0.0432640128,-0.173238188,0.1945694238,0.1986497194,0.0033818609,0.2432301939,-0.1732665151,0.3855418861,-0.1760899574,0.1895332932,0.1059629321,-0.1900244057,0.1902855337,0.7589063644,-0.0688142106,0.21692954,0.359329164,0.0102544511,0.2289912999,-0.0221396517,-0.0018987349,-0.3797623217,-0.02913264,0.3100044429,0.3471744657,0.1730179489,-0.1439306289,-0.0786145031,-0.0761658102,-0.2665877938,0.2564159036,0.4641891718,0.0551711433,0.0940661356,-0.451341629,0.325763613,-0.096344091,-0.2576679587,0.3563808799,0.6770893931,-0.0954731405,0.1394957751,0.4004173875,0.9049196839,0.0827682614,0.1737271696,0.0031513402,-0.339312315,0.5974237919,0.2014868408,0.2052599341,-0.2889653444,-0.1158253551,0.0493202657,-0.0234536417,-0.1391366124,-0.1694779396,-0.0496767536,0.1778852195,0.1009409353,-0.076254718,0.2097299099,0.0811125413,0.329287827,0.0470666699,0.0342058241,0.131862253,-0.0655901134,0.3627304435,-0.0238695461,-0.0672173798,0.1665795594,-0.0531840064,-0.4540297687,-0.1060017198,-0.3741123378,-0.1598393023,-0.2068896443,-0.2787751555,0.0899747908,0.2630541623,0.6369230151,0.1170048416,-0.2022819519,0.2968300283,0.2092126161,-0.3399344087,-0.0385754965,-0.0325204991,0.0729542747,-0.046971865,-0.4824802279,0.0205870681,0.0569699518,-0.3097609878,-0.2386480123,-0.3444824219,-0.2155516595,-0.0426008515,-0.0406221263,0.1557121426,0.2891097069,-0.2364450395,0.1308779269,0.2382089198,-0.021580603,0.0094799474,0.1763736904,-0.1295785308,-0.0420931242,0.3489992321,-0.2122966349,-0.2513976991,0.4640514851,0.0533144586,0.0571857169,-0.2307000011,-0.0966273546,0.0322856233,-0.1737257093,-0.2553484738,-0.2653808296,0.2728780508,-0.0655220151,0.0965374857,0.1904210448,0.1249186918,0.1873113215,-0.3786254525,-0.1960243136,-0.027977515,0.0701155066,0.0200254675,-0.1266750395,-0.3186695576,0.3089510798,0.0238870122,-0.3352238834,0.2886683941,0.0549338758,-0.09019541,0.0819560289,0.1310851425,0.0562633723,-0.166114375,0.0168970153,-0.6879580617,-0.1011054292,-0.1165050939,-0.2289210111,0.1809024513,0.0273049455,-0.1927419007,-0.1123912558,-0.3219532669,-0.0460023619,0.1075630561,-0.2701082528,-0.000701706,-0.0720349252,0.1723218262,-0.2173937261,0.339117676,0.0378724895,0.160591051,0.054761041,0.3583118916,0.1378597021,0.2793379426,0.1034464464,-0.3125390112,0.1207464859,0.1390585452,-0.2034007162,0.1799464524,0.0968893021,-0.1967434883,0.0019371372,0.5520454645,0.3395652771,0.5244819522,0.0260485131,0.1232304722,0.3496303558,0.2413620204,-0.3485979736,0.2696170211,0.3537509143,0.1770886481,-0.0470032245,-0.0910046026,0.4353304207,-0.003416609,-0.2699351311,0.2203965038,0.429174006,-0.1460224688,0.0769175366,0.0206868611,0.0055982172,0.0343426615,0.22240448,0.5141208172,0.0484518446,0.2780649662,0.0842245296,0.1572732329,0.0436353125,-0.0590576492,-0.1049678251,0.1966388822,0.0795727149,0.1122419909,0.0878396034,-0.0910459459,-0.1542141587,0.7406972647,-0.1039512232,-0.2170755267,-0.2388750464,0.124853991,0.1798283458,-0.3467885554,-0.6718619466,-0.1101686358,-0.0743154213,-0.1182857379,0.1658358574,-0.1971797496,0.385758847,-0.1451541185,-0.1421555579,-0.4889583588,0.0479588732,0.2954272032,0.2731299102,0.0242915656,-0.0154147204,0.2184958905,0.0382713675,0.2569575012,0.6370867491,0.0527471751,-0.031829685,-0.3823573589,0.0200688336,-0.0201665033,-0.1323195696,0.1026285067,-0.0901871473,-0.0984991044,0.0167688411,0.1813509315,0.1151835397,-0.0449462458,0.2071427852,0.1966823786,0.2826370299,-0.0761736035,0.1413924396,-0.0195684955,-0.1620721966,-0.1833899021,-0.2084055692,-0.050103493,0.0226610247,0.2130999863,-0.2613892257,-0.0142903812,-0.2763113678,0.0691142529,0.0182071459,0.4819695354,0.4473837912,0.2496091723,-0.2478417605,-0.26063627,-0.6751897335,0.0702117383,0.0466932617,-0.2329179198,0.0193472318,-0.0835056603,-0.1857014447,0.123835288,0.256457746,-0.1176079288,0.0093840091,0.3501382172,-0.3854677677,-0.036820814,-0.1467262506,0.4041577876,0.1732931137,0.1981582195,0.1174555346,-0.0660074204,-0.0030955651,-0.0030507168,0.1023303345,-0.103005603,0.1763143092,0.1233723387,-0.000933202,0.0599504896,0.0501687527,-0.0178513229,-0.6541119814,-0.2283634096,-0.4662816226,0.0908214897,-0.1706441641,0.0184335075,-0.2503196001,-0.5449128151,-0.017903503,0.1748427898,0.0182504971,-0.4272343516,-0.334302932,0.2882147431,0.145425871,-0.0655941665,0.238757506,-0.0146009382,-0.0937185436,0.188016504,-0.2161310464,-0.2531245351,0.451177597,-0.7091348171,-0.003316476,-0.2580904365,0.4379968345,0.4351271987,0.1337576061,-0.1770675778,0.1712412536,0.1312957406,0.2485462129,-0.3428422511,0.1815029383,-0.2439756989,-0.0710948929,0.1230073422,0.6898620129,0.1534520686,-0.4521771371,-0.1873071194,0.0191911627]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1919","title":"Failure to save with save_to_disk","comments":"Hi thanks for reporting and for proposing a fix :)\r\n\r\nI just merged a fix, feel free to try it from the master branch !","body":"When I try to save a dataset locally using the `save_to_disk` method I get the error:\r\n\r\n```bash\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/content\/squad\/train\/squad-train.arrow'\r\n```\r\n\r\nTo replicate:\r\n\r\n1. Install `datasets` from master\r\n2. Run this code:\r\n\r\n ```python\r\n from datasets import load_dataset\r\n squad = load_dataset(\"squad\") # or any other dataset\r\n squad.save_to_disk(\"squad\") # error here\r\n ```\r\n\r\nThe problem is that the method is not creating a directory with the name `dataset_path` for saving the dataset in (i.e. it's not creating the *train* and *validation* directories in this case). After creating the directory the problem resolves.\r\nI'll open a PR soon doing that and linking this issue.\r\n","comment_length":25,"text":"Failure to save with save_to_disk\nWhen I try to save a dataset locally using the `save_to_disk` method I get the error:\r\n\r\n```bash\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/content\/squad\/train\/squad-train.arrow'\r\n```\r\n\r\nTo replicate:\r\n\r\n1. Install `datasets` from master\r\n2. Run this code:\r\n\r\n ```python\r\n from datasets import load_dataset\r\n squad = load_dataset(\"squad\") # or any other dataset\r\n squad.save_to_disk(\"squad\") # error here\r\n ```\r\n\r\nThe problem is that the method is not creating a directory with the name `dataset_path` for saving the dataset in (i.e. it's not creating the *train* and *validation* directories in this case). After creating the directory the problem resolves.\r\nI'll open a PR soon doing that and linking this issue.\r\n\nHi thanks for reporting and for proposing a fix :)\r\n\r\nI just merged a fix, feel free to try it from the master branch !","embeddings":[-0.0436869934,0.142949298,0.0570114776,0.1895621121,0.5194172263,0.2885248661,0.192212835,0.2388131469,-0.1313640922,0.1859047413,0.141578272,0.3680593073,-0.3196108043,-0.2508209646,0.0698751137,0.0474057831,0.2884759605,0.0203925744,-0.0373371616,-0.0828118548,-0.1238706633,0.264870733,0.0996688977,0.0421990082,-0.5081453919,-0.0061072712,-0.1618735492,0.31453076,0.001071627,-0.2673496306,0.1534147561,-0.3283669651,0.2090518028,0.5840054154,-0.0001187165,0.0369130969,-0.0382387005,-0.1583705544,-0.4442630112,-0.4077847898,-0.221984148,-0.1730088443,0.1148869023,-0.4092869163,0.0701678395,-0.0506983846,0.2495624125,-0.1575463563,0.3871827722,0.2822141647,0.1549114436,-0.0382053778,0.1740321517,-0.2288421541,-0.1663120538,0.6161901355,-0.3467484713,0.2955402732,-0.0347836353,-0.1686111987,0.3699814379,0.0192853771,-0.0258871969,-0.1304126829,0.28187567,0.1215031296,0.0884760767,-0.330727458,-0.013628141,0.0122261383,0.614715457,-0.6311994791,-0.5074194074,0.224433586,0.0965238065,0.019315023,0.1897397637,0.2495189756,-0.2263125926,0.1998587698,-0.1867233664,-0.3792485595,-0.3171930611,0.2024367154,0.2337694913,-0.1623820364,-0.0914282575,-0.0537081771,0.1051189676,-0.0304780751,-0.0507375374,-0.2843352258,0.0843666345,0.1800297052,-0.1861409843,-0.1386528015,-0.1079159379,-0.0201213434,-0.137375772,0.120416373,0.1969716698,-0.2985674143,0.0002792501,0.3721286058,0.1506853849,0.0598145351,0.2832764387,0.3988432884,0.3854604661,0.0902727023,0.0267784055,-0.0459410772,-0.041089762,-0.1973222941,0.5454720259,-0.0097808056,0.3190205693,-0.1826929003,-0.2289980054,0.4254073203,-0.1825689673,0.1510066986,0.0385164618,0.2409756631,0.111803785,-0.0846430436,-0.0101934876,0.4569025636,0.0640194342,0.3838762045,-0.1783630103,0.1913471967,-0.0088196388,0.2253187895,0.2253583223,-0.1830714494,0.1848752797,0.0567193106,-0.0188269503,-0.0164779108,-0.0026833476,-0.2802461088,0.0282900464,0.3156400323,0.072548911,0.1089627668,0.1198396832,-0.4458800554,-0.1219691038,0.2643061876,-0.0774846897,-0.4749075174,-0.2428726852,0.0788345858,0.0020117979,-0.0123701124,-0.2620709538,-0.1679551005,0.205144152,-0.115492247,-0.064013809,0.0571926869,-0.0576198176,-0.416885674,0.4127009511,0.3620338142,-0.506675899,0.2732725441,0.0345090143,-0.123007603,0.159774214,0.5587685108,-0.0151765971,0.3559604883,-0.2971852124,0.311873883,0.5128965974,-0.40454036,-0.2994139194,0.1477932185,-0.1420666873,-0.3448434174,0.0593093149,0.1312277913,-0.1309546828,-0.1517180353,0.0372945443,0.0817090347,-0.1437611431,0.0410480499,-0.4923837483,0.1228282973,-0.0063296775,0.0801651329,-0.1582182348,-0.0098179309,0.2346519828,0.0741199479,0.3167810738,-0.2465216964,0.1298096031,0.4385485947,0.7910145521,-0.1569506973,-0.1734226346,-0.0515779667,-0.3099271953,0.1068462878,0.2395672798,-0.0529066622,-0.1526204497,-0.0953178927,-0.1524760127,0.0408612192,-0.4206057489,0.0266656391,0.0151800141,-0.0559189469,0.1668254435,0.1911595017,-0.3386030197,0.1531254202,-0.3734537661,0.0766417757,0.054072801,0.4180463254,-0.3401333988,-0.3832112253,-0.2234160751,0.0831986815,0.0262242835,-0.0200842619,-0.1568193138,0.4635842144,0.0581358857,0.1698818505,0.1769722253,0.059483245,0.1188471317,-0.1710497886,-0.0211702865,0.3229103386,0.3793407083,0.0707868263,-0.6439212561,-0.0796288475,-0.0559942573,0.0869714543,0.1420308352,0.0986385494,0.1159187704,0.0188671127,0.008736846,-0.0919285342,0.1190077364,0.0041110129,0.1784569323,-0.2022560239,-0.3054654002,-0.0537458211,0.0045893178,0.0100243511,0.4323045611,0.1164100617,0.1207392812,-0.0221861303,-0.0730500519,0.4694502652,0.539755702,-0.0449205339,-0.0467782505,0.0906326398,-0.2236265689,-0.132350415,0.1095442101,-0.1445541531,0.2041365057,0.578312695,0.0744415969,0.1398697793,-0.237594232,-0.0113389026,-0.0262815934,0.2607398629,-0.293199718,-0.1521639973,-0.2273589224,0.1628912985,-0.2301882356,0.0590158775,-0.0771367252,-0.2169804573,0.0288824532,0.3544380963,-0.0217200462,-0.0699991062,0.007992289,0.0586082898,-0.1687670797,-0.359392643,-0.0122328019,-0.0431214832,-0.0647554994,-0.0405164324,0.2432489395,-0.093178913,0.407433331,-0.1362606883,0.0064516044,-0.6529046893,0.0417982414,-0.0146363229,0.006026695,0.3284969628,0.0468836725,0.3231469393,-0.2639580965,-0.2300036401,0.3766773045,0.0304342322,-0.1758868098,0.1453618705,0.1346272379,-0.3108990192,-0.1181407645,-0.3630990684,-0.3233220577,-0.2766486704,0.1131914258,-0.2195637077,0.1284826845,-0.0175772347,0.2915680707,0.2227851897,0.1160493568,-0.0855979994,-0.0912680626,-0.3929232657,0.6367709041,-0.310070008,-0.4007944167,0.1726480722,0.1183610633,0.029365208,-0.0310871731,-0.3465647399,-0.3220780492,0.0218669232,0.1416007876,-0.1745146364,0.2809893787,0.3479137719,-0.0128053017,0.0259512234,0.019551171,-0.3645197451,0.1986777633,0.5634405613,0.5095309615,-0.2741743624,0.4664591253,0.0063155796,0.5923214555,0.1623983383,0.2559560239,0.4696496427,0.0460422821,0.3521324098,-0.2076819688,-0.3169225156,0.1021856144,-0.1217054576,-0.5222668052,-0.1713111699,0.1042052284,0.0405895412,-0.0762661472,-0.0276084449,-0.1950878799,-0.1804351658,-0.0646873266,0.149326399,0.2803890109,-0.1482632458,0.1099720076,0.2512289286,-0.0083691981,0.0718713999,0.6023398638,0.3001723886,0.0348981805,-0.5934087634,-0.332179755,-0.3693142235,0.0797094628,0.0514101014,0.380574882,-0.2038463652,-0.0741521046,-0.0068888916,0.0717543066,0.4177282155,-0.2680716515,-0.2005081773,0.0655285493,0.047638718,-0.4446979761,-0.166714862,0.330550611,0.3490884602,-0.1856458783,0.4512579739,-0.0317449421,-0.114383094,-0.093333818,0.2862459719,-0.1979819834,-0.2021401972,-0.1551747918,-0.2627649903,-0.3537598252,-0.1133114547,-0.0027021794,0.0357784778,0.3868080676,0.0661626011,0.0273665357,0.0027788326,0.1347556114,-0.1732167751,0.4466359615,-0.2300791591,0.1627964973,-0.0883875862,0.0243435316,0.3703680933,0.3753552437,-0.2585771084,-0.3253380358,0.0951187983,-0.2750011981,0.4798430204,0.1903957874,-0.1140700579,-0.4817686081,-0.0603415146,-0.0389532596,-0.2283413112,-0.2157603055,0.2810062766,-0.1916750669,-0.1766624749,-0.0019361373,0.1306398511,0.2120267004,0.221573323,0.106867753,0.0158487167,-0.066400066,0.331007272,0.1926985085,0.8347555399,-0.021092739,0.3929268122,0.104682669,-0.278688103,0.1136672869,-0.1644261926,0.1742984504,-0.5152051449,0.0956633314,-0.0492588505,-0.0825595632,0.2677944601,0.256013155,-0.2164886296,0.3953602314,-0.3980212212,0.1224348173,-0.1871150732,0.0445941649,-0.2941901982,-0.2515767217,-0.2279292345,0.1340621114,0.0702633932,0.1735769808,-0.0390295796,-0.0485874526,-0.3068465889,-0.0457439534,-0.2921003401,0.2171729952,-0.283955425,0.4295626581,-0.1510763913,-0.5352062583,-0.0205071568,0.7050192952,0.4556487501,0.0605424047,-0.2067081183,-0.0671710521,-0.1587240249,0.1630186141,-0.0978155285,-0.1728441417,0.3130180538,-0.076165244,-0.0021598481,-0.0566107221,-0.1997144222,-0.2288071215,-0.1366551667,-0.0496367775,0.1111476794,-0.1250342578,-0.5298013687,0.012467077,-0.125045836,-0.1918313652,0.0934787244,0.1643523723,-0.1178807691,0.3749890924,-0.3534035981,0.0724096075,0.1048418358,0.4314040244,-0.3103451133,0.0340765864,0.4717000425,-0.1298153102,0.0600452535,-0.3047808707,0.4256797433,0.0011194173,-0.1358040124,0.1446162015,-0.2254025936,-0.0124130007,-0.0816781521,-0.0176228974,0.3689910769,-0.2554948032,-0.085753873,-0.2979804575,-0.5189905167,-0.0992813036,0.4103376269,0.1895849556,-0.0605897531,-0.1244573891,0.1943563223,-0.2750883102,-0.2191739827,-0.0253715385,-0.3023048341,0.0205048379,0.0584355332,0.0555098951,0.3383267522,-0.3429073095,0.0912236869,0.0168084782,-0.1370325238,-0.1676114798,-0.038636487,0.1491854489,0.1423889697,0.1044512391,0.0018484246,-0.3341522515,-0.073188372,-0.0511822216,-0.2740119994,0.4773292542,0.0286548939,0.0739975572,-0.008724709,0.0509583615,0.1287954599,-0.196923539,0.052765552,0.2355601043,0.0527856797,-0.1015891135,0.1415501088,-0.2127240747,-0.324586004,-0.0788838938,-0.1340298206,0.1693494469,-0.0163829345,-0.377715826,-0.1813931465,0.0101543311,0.4741109014,0.3265454471,-0.3118917644,0.2228902578,0.38209337,0.1247721985,0.023510782,-0.2684881389,0.2837251723,-0.1158549041,-0.2796919048,0.0808056444,-0.0752426907,0.0503998101,-0.4147300124,0.0888026431,0.3177274168,-0.3490390778,0.1107111424,0.1824828535,-0.0235072542,0.3092257082,-0.1106093004,-0.1716782153,0.1599222273,0.2186932713,-0.1734767556,0.2385257185,0.3012813032,0.1493503898,0.295352906,-0.0135152265,-0.006918814,-0.4171992838,-0.0982951671,-0.0134436488,-0.0001410936,0.3464176059,-0.4165669084,-0.2585897744,-0.1719698608,0.0722397417,-0.2669572234,0.0741031989,-0.0763530359,-0.1190554723,0.2702915072,-0.0594432428,-0.1385216862,-0.1295191795,-0.052325055,0.1955135763,0.1951879859,-0.0229412466,0.1453984231,0.2591456175,0.027171772,-0.1961701661,-0.0074240253,0.0663925037,-0.3078882396,-0.399821192,0.3634181321,0.4397539198,0.5535860062,-0.0218441095,0.2691555023,-0.2226386815,0.0166128762,0.2642598152,0.0386485867,-0.1029789001,-0.0857439786,0.4220949113,0.0959093273,0.0597936027,0.2503314018,-0.0150832199,0.1046721861,-0.0081345923,0.2377560288,-0.0240383763,-0.2090127319,-0.0002202796,0.085436143,-0.219811067,-0.0536932535,0.5168391466,0.0682664141,0.1285763681,-0.1078775227,0.0383699536,-0.0365350321,0.3760023713,0.2762748003,-0.2577250898,-0.5194141865,0.0330716744,-0.6200771332,-0.1018305421,0.0556321628,0.2281337082,-0.2808684111,0.263874054,-0.1216255352,0.2111622244,0.190923959,0.1290503889,-0.1083809063,0.5609630942,-0.1707858592,-0.0435360037,0.0092591085,0.068142578,0.1391112506,-0.3969303071,0.174302116,0.178389892,-0.0282298196,0.0432105698,-0.0229311101,-0.2206307352,0.3385377824,0.4254108071,-0.0666353256,0.1587461829,0.0038032338,-0.3034677804,0.1541690081,0.0924081281,-0.0381421447,0.1844716668,-0.0478865169,0.5128154159,-0.3749782741,-0.0932090804,-0.4210633934,-0.0058483714,-0.0919632912,-0.0159097873,-0.0792490914,0.0060059689,0.1355394125,0.1587672234,0.2504431009,-0.0072605573,-0.1995280981,0.3668052554,-0.3987038136,-0.2346246392,0.5763828754,0.0544463992,-0.0320315324,-0.0823913291,0.2976126671,0.257694006,0.1391412467,-0.3864381015,-0.2610426247,0.2535917759,-0.0958445072,-0.2886846066,0.0853897706,-0.1614722908,0.1608714759,0.1543487906,0.3828740716,-0.0747896135,-0.0117755774,0.1511539519,-0.3348569274]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1919","title":"Failure to save with save_to_disk","comments":"Closing since this has been fixed by #1923","body":"When I try to save a dataset locally using the `save_to_disk` method I get the error:\r\n\r\n```bash\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/content\/squad\/train\/squad-train.arrow'\r\n```\r\n\r\nTo replicate:\r\n\r\n1. Install `datasets` from master\r\n2. Run this code:\r\n\r\n ```python\r\n from datasets import load_dataset\r\n squad = load_dataset(\"squad\") # or any other dataset\r\n squad.save_to_disk(\"squad\") # error here\r\n ```\r\n\r\nThe problem is that the method is not creating a directory with the name `dataset_path` for saving the dataset in (i.e. it's not creating the *train* and *validation* directories in this case). After creating the directory the problem resolves.\r\nI'll open a PR soon doing that and linking this issue.\r\n","comment_length":8,"text":"Failure to save with save_to_disk\nWhen I try to save a dataset locally using the `save_to_disk` method I get the error:\r\n\r\n```bash\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/content\/squad\/train\/squad-train.arrow'\r\n```\r\n\r\nTo replicate:\r\n\r\n1. Install `datasets` from master\r\n2. Run this code:\r\n\r\n ```python\r\n from datasets import load_dataset\r\n squad = load_dataset(\"squad\") # or any other dataset\r\n squad.save_to_disk(\"squad\") # error here\r\n ```\r\n\r\nThe problem is that the method is not creating a directory with the name `dataset_path` for saving the dataset in (i.e. it's not creating the *train* and *validation* directories in this case). After creating the directory the problem resolves.\r\nI'll open a PR soon doing that and linking this issue.\r\n\nClosing since this has been fixed by #1923","embeddings":[-0.0454722121,0.1459745169,0.0602632575,0.1778056622,0.4819115102,0.3136120439,0.2128188908,0.2358687222,-0.1218468174,0.2012174428,0.1596109867,0.3447875977,-0.3220589459,-0.281961292,0.0766139999,0.0534634739,0.3039935529,0.0099745002,-0.0236092657,-0.0860661641,-0.1675192118,0.2575318515,0.1083029062,0.0558859818,-0.5031904578,-0.0176541694,-0.1729001403,0.3241853118,0.0385291204,-0.2517320812,0.1330118328,-0.3309397399,0.2090246975,0.5483575463,-0.0001191118,0.0250290092,-0.0279018674,-0.1432851404,-0.4836084247,-0.3924627304,-0.2169181108,-0.1650694907,0.1284265071,-0.4142520726,0.1227133051,-0.060999278,0.2796038389,-0.1797349006,0.3705977499,0.258995384,0.152259469,-0.0254510622,0.1678916216,-0.2386840433,-0.1582733542,0.6525206566,-0.3766581714,0.2813692093,-0.0456080809,-0.1411252022,0.3722431362,-0.0034799818,-0.0376885645,-0.1021190882,0.2788050771,0.1218921617,0.0669316053,-0.2903158069,0.0162196457,0.0356163941,0.6141216755,-0.654518187,-0.5000389814,0.2072982341,0.0760903507,0.0121843265,0.1850981265,0.2224955112,-0.2341009974,0.1947090179,-0.1931081265,-0.3909176886,-0.3147275448,0.2212132514,0.2260747701,-0.1785453856,-0.0906339809,-0.0318120345,0.1004273742,-0.0397195034,-0.0262208749,-0.3224375248,0.1119300202,0.1638354212,-0.1328068078,-0.1061607823,-0.0928485021,0.0145391878,-0.1522725672,0.1282441914,0.1964695007,-0.3097831607,-0.0192321315,0.3657628,0.1193145141,0.0651824623,0.2449765652,0.3959979415,0.3762278557,0.0825780779,0.0548074767,-0.0390957259,-0.0616130643,-0.1546515524,0.5732768178,-0.0563444607,0.3043907583,-0.1715777218,-0.2581066787,0.438588202,-0.1874002665,0.1412317455,0.0006820663,0.2139499635,0.1371718794,-0.1285811812,-0.020804096,0.4611298144,0.0747337565,0.3890019059,-0.1984332502,0.1936920285,0.0045446763,0.2487003207,0.2262179255,-0.1594564766,0.1781109571,0.0130283292,-0.0296577159,0.0139853721,-0.0122977924,-0.2827103734,0.0605338514,0.3154584765,0.0742643103,0.1167129204,0.1199509054,-0.4700453877,-0.1547076702,0.2668646276,-0.0556033403,-0.4356417358,-0.2185701132,0.0825063363,-0.0076145059,-0.0300826579,-0.2690930665,-0.1774781942,0.2064899653,-0.1384309381,-0.0734006241,0.0497052185,-0.0470183864,-0.4198392034,0.4122245014,0.4154115319,-0.5102511644,0.2813529968,-0.0128277782,-0.1025190651,0.1699428707,0.5522985458,0.0119768297,0.3633149564,-0.3168611825,0.3343156576,0.5220881701,-0.3667138219,-0.2781837285,0.167043522,-0.1294516623,-0.3222136199,0.0602138937,0.1279399246,-0.1371601224,-0.1501688361,0.0377227068,0.0711101964,-0.1441951841,0.0132864928,-0.4627874792,0.1287294626,-0.0117363008,0.0723878369,-0.1950644404,0.006530561,0.2397696674,0.0769972727,0.2938893735,-0.2080800235,0.1304306537,0.4335074723,0.7482835054,-0.1726855785,-0.1939040869,-0.0367122702,-0.3209858835,0.1026858687,0.2750215232,-0.0918409079,-0.1368254572,-0.0967847258,-0.1538455933,0.0538301021,-0.4147994518,0.055497963,0.0071209199,-0.0227852799,0.2031404525,0.2034980804,-0.316490382,0.1373868585,-0.3487656116,0.0584803373,0.0340002999,0.3982741535,-0.3314959407,-0.3733037114,-0.2407271862,0.1033391356,0.0285996571,-0.0112988008,-0.1358210295,0.4597263634,0.0517299064,0.1977541745,0.1541383415,0.1074563712,0.0998554528,-0.172226578,-0.0044258269,0.3394437432,0.3830580711,0.0587014295,-0.6437901258,-0.1190300956,-0.0280297324,0.1038063616,0.1650286317,0.0944975168,0.0950790346,0.0423139147,0.0053073927,-0.1112206504,0.1403886527,0.0009486425,0.1965453029,-0.2025052309,-0.3031795323,-0.0487529971,-0.0164065473,-0.0113807796,0.481254667,0.1079405844,0.1076412946,-0.025880998,-0.0577867515,0.4716669619,0.519558847,-0.0427597687,-0.0185085461,0.0905289948,-0.2661867142,-0.1626723856,0.0772649646,-0.1229642928,0.1961182356,0.5765924454,0.0605233796,0.1594773233,-0.2288222611,-0.0445816517,-0.0314901695,0.2541310787,-0.2917925417,-0.1287752986,-0.2262894064,0.1596861184,-0.2583131492,0.0842352733,-0.0576556884,-0.2411369681,0.0124629624,0.3800755441,-0.0372499898,-0.0754060745,-0.0304455142,0.0820811316,-0.1811150461,-0.3938151002,-0.0295608751,-0.004344529,-0.1032468006,-0.0370467156,0.2306065708,-0.1096808016,0.4133143723,-0.154722929,0.0300074201,-0.6621431112,0.1065167412,-0.0086608054,0.0029624375,0.3009829223,0.0403677747,0.3215003014,-0.281658411,-0.2423710823,0.360943377,0.0562095083,-0.1532462239,0.1556725055,0.1469996125,-0.3220201433,-0.1032376066,-0.3450143039,-0.3054874241,-0.2564385235,0.1405747831,-0.2092833221,0.1078964248,-0.0140333651,0.2886074185,0.2410936803,0.1358706206,-0.0757923499,-0.0916232169,-0.3560595512,0.6322084665,-0.2912359834,-0.3576440811,0.1447479278,0.1018239558,-0.0007020758,-0.0050398163,-0.3537961543,-0.2946646512,0.0089860922,0.1397423148,-0.1608416736,0.3058066964,0.372367084,0.0187164638,0.0260370541,0.0251429826,-0.3686543703,0.2113608569,0.5452209115,0.483109355,-0.2605043352,0.4375837445,-0.0234601367,0.5586878061,0.179729104,0.2529701889,0.4923608303,0.0536940657,0.3575519323,-0.2235596776,-0.2752074599,0.0948073119,-0.1364291459,-0.5605276823,-0.1917362213,0.1022811234,0.036217127,-0.0580450706,0.0074059195,-0.1905190647,-0.1723173112,-0.0977829024,0.1481289268,0.3047075272,-0.1463681757,0.1228773743,0.2856322527,-0.0108019533,0.0635066926,0.5805811882,0.3296270072,0.0018231072,-0.5693743229,-0.3468470871,-0.3900059164,0.0748784542,0.0564814024,0.383621484,-0.2300029546,-0.1219168603,-0.0256777778,0.0932151601,0.4042369723,-0.2689138055,-0.2154144794,0.0761848167,0.0641702041,-0.5036776066,-0.1508371234,0.3282284439,0.3099474907,-0.216328606,0.4668899179,-0.0463003255,-0.0684974939,-0.0899212807,0.3347346783,-0.2023601681,-0.2061738372,-0.1429482102,-0.3008933067,-0.3324765265,-0.1551550478,-0.0242723897,0.0308593772,0.4000303745,0.0892190188,0.0071853646,0.0083433613,0.1571943909,-0.188042745,0.429417491,-0.2334521562,0.1643823087,-0.0756522566,0.038639266,0.365355432,0.3657547832,-0.291513592,-0.3161011636,0.099616468,-0.2370423526,0.4481694996,0.1968546212,-0.1149711013,-0.4994105399,-0.0285205953,-0.0736558884,-0.280276984,-0.196261093,0.2861952186,-0.2132455558,-0.1674471647,-0.00820396,0.1524488479,0.2312515676,0.2152250111,0.0931615829,0.0203166828,-0.0664671212,0.3325028718,0.1634415537,0.8270385861,-0.0122713698,0.4100121856,0.1234663352,-0.2621618807,0.0699314252,-0.1349287778,0.1605491489,-0.4996156096,0.1168651432,-0.0490786992,-0.0737631619,0.2748275995,0.2490328699,-0.1966339797,0.3810817897,-0.4197617471,0.1351208538,-0.1608535796,0.0291859396,-0.3250791728,-0.2663833201,-0.1911587417,0.1373316348,0.1014308855,0.1636351049,-0.0258752629,-0.0109849079,-0.3057770729,-0.0496426523,-0.2940428555,0.2207651287,-0.2662471831,0.4255137742,-0.1460680217,-0.5219295025,-0.0259394459,0.6845982671,0.4853636026,0.0770735368,-0.2034773976,-0.0907206461,-0.1595537066,0.1479286551,-0.0876029432,-0.1755505651,0.2993941307,-0.0449024066,0.0259103291,-0.0606523715,-0.1972535849,-0.2024756372,-0.1542836279,-0.0700370148,0.1309524626,-0.1507246196,-0.5251845121,0.0113415895,-0.1365833282,-0.1978164762,0.0901776925,0.1499551982,-0.0860435665,0.3564593196,-0.3762525618,0.0772266015,0.0980765894,0.426628679,-0.3306312561,0.0461883806,0.4672518671,-0.1662420332,0.0400086492,-0.2887984514,0.4274754226,0.003404845,-0.1334697902,0.1579136848,-0.21842435,-0.0021141681,-0.1344150752,0.00645144,0.3820420802,-0.252325505,-0.0869283304,-0.2662516534,-0.4764588773,-0.0822011977,0.4412848055,0.1886109412,-0.0422996283,-0.163359642,0.1782086343,-0.2954446077,-0.2131765485,-0.0518867224,-0.2743162811,-0.0092698019,0.0617510341,0.0702770427,0.3478434086,-0.327033937,0.1014054492,0.011155271,-0.1235881671,-0.1696092188,-0.0574440658,0.1375873536,0.1425720751,0.0808647722,-0.0035085247,-0.3329453766,-0.0564024933,-0.0763497204,-0.2665266395,0.4911968112,-0.0008873722,0.0562777482,0.0039254911,0.0844468847,0.1089706868,-0.1851009727,0.0340944864,0.2446621954,0.0362888649,-0.1092753932,0.1474728137,-0.2059413493,-0.3537523448,-0.0568958968,-0.1093250066,0.1253881603,-0.0323316753,-0.3643604517,-0.1579608619,0.0091602402,0.4833977818,0.3309371173,-0.3173161745,0.1788804531,0.410805434,0.1141000018,0.0495412461,-0.2803115249,0.2659290135,-0.1064414084,-0.2595844567,0.0978504792,-0.0837041661,0.0405851193,-0.4375405014,0.0393004678,0.3193021715,-0.3401571512,0.0904817358,0.2073451728,-0.010860703,0.332641542,-0.1172659248,-0.1465524137,0.1753389686,0.1949026585,-0.1684360802,0.2371272296,0.2984831929,0.1527365595,0.2807841003,-0.0457310639,0.0317036211,-0.4608289599,-0.0952862948,0.0173489098,-0.0196610019,0.291934222,-0.4380250871,-0.2204205841,-0.1941017807,0.0428760089,-0.2558835149,0.0371472687,-0.0298123397,-0.1344526708,0.277705133,-0.0647342429,-0.1376666874,-0.1596805155,-0.045706939,0.160433054,0.2268493474,0.0031126551,0.1786895841,0.2251681089,0.0214584824,-0.1803481728,-0.0078377668,0.0464366451,-0.3065976202,-0.4252684414,0.3595761359,0.4357853234,0.5265779495,-0.0077403677,0.2576757967,-0.2866071165,0.0290604811,0.2520585656,0.0556304418,-0.1251506656,-0.0884988084,0.4381090403,0.0957569405,0.0656031668,0.25080055,0.0002621464,0.1339402497,-0.020066075,0.2686880529,-0.0679888278,-0.1949611455,0.0318692029,0.1051264182,-0.2346801311,-0.0530953221,0.5320307612,0.0625414774,0.1352676451,-0.0860144868,0.0355862863,-0.0484785214,0.3853313923,0.2863916755,-0.2630216777,-0.516884625,0.0418937355,-0.5929528475,-0.0861770511,0.0385478772,0.1810338646,-0.2844314873,0.2571778297,-0.1351568252,0.2157465667,0.1435604393,0.1577586681,-0.1072184369,0.5891895294,-0.1236224324,-0.0578571409,0.0183941089,0.0427981019,0.120187372,-0.4123603106,0.174312219,0.1906407177,-0.0326112621,0.0463872254,-0.0362940095,-0.221627146,0.3495963514,0.4338771105,-0.0611273535,0.1477762759,-0.0121057928,-0.2985063195,0.1323631257,0.1081031114,-0.00095705,0.1673992127,-0.0795164332,0.4619467854,-0.413168788,-0.0685465336,-0.4056153595,-0.007219228,-0.0839685798,-0.01200727,-0.0666933954,-0.0062802862,0.0973209888,0.1398613453,0.2986194193,-0.0238795821,-0.20003438,0.3677126169,-0.362391293,-0.2318203896,0.5680642724,0.0796045884,-0.0192601588,-0.0893948227,0.2922913134,0.2538231909,0.1429369003,-0.3502804041,-0.2651204765,0.2798573077,-0.1064845622,-0.2676223814,0.082207337,-0.1894234717,0.1364842355,0.1611816883,0.4127109945,-0.1003586352,-0.0148975244,0.1569976211,-0.3237011433]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1917","title":"UnicodeDecodeError: windows 10 machine","comments":"upgraded to php 3.9.2 and it works!","body":"Windows 10\r\nPhp 3.6.8\r\n\r\nwhen running\r\n\r\n```\r\nimport datasets\r\n\r\noscar_am = datasets.load_dataset(\"oscar\", \"unshuffled_deduplicated_am\")\r\nprint(oscar_am[\"train\"][0])\r\n```\r\nI get the following error\r\n\r\n```\r\nfile \"C:\\PYTHON\\3.6.8\\lib\\encodings\\cp1252.py\", line 23, in decode\r\n return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 58: character maps to \r\n```","comment_length":7,"text":"UnicodeDecodeError: windows 10 machine\nWindows 10\r\nPhp 3.6.8\r\n\r\nwhen running\r\n\r\n```\r\nimport datasets\r\n\r\noscar_am = datasets.load_dataset(\"oscar\", \"unshuffled_deduplicated_am\")\r\nprint(oscar_am[\"train\"][0])\r\n```\r\nI get the following error\r\n\r\n```\r\nfile \"C:\\PYTHON\\3.6.8\\lib\\encodings\\cp1252.py\", line 23, in decode\r\n return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 58: character maps to \r\n```\nupgraded to php 3.9.2 and it works!","embeddings":[-0.3856231272,0.0980772153,-0.1237038895,0.4771643281,0.4591267109,0.0767546445,0.1410268396,0.3478401899,-0.2554140389,0.1840734631,0.2676636279,0.2224644125,-0.1075476632,-0.4143674076,0.0257732887,-0.0647161603,0.2888023853,0.239443928,-0.0674791411,-0.0858589038,-0.4802003801,0.3050513268,-0.2515510023,0.2307684869,-0.1449909806,-0.1191773862,-0.0656924546,0.0846950114,-0.3173426986,-0.5594245791,0.0756158531,-0.093593277,0.3571541309,0.2990395725,-0.0001112805,-0.0985494778,0.2768119276,-0.0292477198,-0.0734337196,-0.0614212528,-0.2926857769,0.0341933668,0.0027999107,-0.0326391608,0.2925029397,-0.1035518944,-0.0790895522,-0.4253410995,0.1875777096,0.4606081247,0.2305106968,-0.1116035581,-0.0997159332,-0.1409738958,-0.0615660436,0.056264665,-0.0426060744,0.0248322021,0.1054643393,-0.2690535784,0.2722634673,0.1682191789,-0.3016638458,-0.018376261,-0.2031937987,0.0443209112,-0.0563550554,-0.4133683741,0.4125463367,0.0658865646,0.363027066,-0.338411659,-0.2151494473,0.0745142177,0.2320626378,-0.2901382446,0.2454329431,0.0826834217,-0.0681428909,0.2009836137,0.079318963,0.1262541115,-0.080040738,0.0647683367,-0.0928989425,0.2330993563,-0.1017987728,-0.0147603862,-0.0401433595,-0.1501978189,0.2740641832,-0.0391106904,-0.060916964,0.2259772122,0.0604588278,0.1643537134,-0.2485349029,-0.2018023431,0.0677089542,0.219633311,0.1775224358,0.2290952951,0.2551826239,0.0018677551,0.115802139,0.0509103537,0.0239196792,0.1763727516,0.244814977,0.1926914603,-0.0680054128,-0.09628167,-0.450437516,-0.1927208006,0.1736155748,0.2972631156,0.4827865362,-0.0855369419,-0.1553178132,0.1482302397,-0.0228499454,-0.1229048595,-0.0761768669,0.3465724289,-0.0659588724,0.3822956979,-0.1871381998,0.1520625651,-0.2398327291,-0.1730767339,-0.1211345717,-0.0303407926,-0.069613561,-0.0326519907,0.129533127,-0.1674832851,0.01421561,0.1100767851,0.2575817704,-0.0308221299,-0.0772308558,-0.1261075586,0.0395605974,0.0480407849,0.0775521398,0.4086249471,0.2005735934,-0.2471066713,-0.2138561457,0.2215946317,-0.305988431,-0.0828007981,-0.22773242,0.2212927788,0.0138790635,-0.113516517,-0.0011969419,0.1250077039,0.2258231491,-0.1907804757,-0.0585096963,-0.245715484,-0.0582372397,-0.2870610952,0.2367257476,0.3141480088,-0.566398859,0.1946065277,0.1419872493,-0.2376205325,0.2334515303,0.0437130406,-0.1698141247,0.1103600338,-0.1392560601,0.1720301211,0.2161302269,-0.1829609871,-0.4020714164,0.0413830914,0.3971996009,-0.0138773555,-0.0980606452,-0.2717710733,0.2356907725,0.1993918568,-0.0377843976,0.2965170741,-0.0538776591,0.1220554709,-0.1776152402,-0.2450730354,0.3240061104,0.1710414886,0.1876946092,0.0310245734,0.1897466481,0.2943571508,0.215053454,-0.1174103841,0.2779630125,0.1032426506,0.4007765055,-0.0073642344,0.1019489914,-0.3867780864,0.2517250478,-0.0252806153,-0.0076294886,0.2368594855,-0.6547681093,-0.2647831142,-0.3121625483,-0.025830904,-0.4846376777,-0.0283334851,0.2063694447,0.0481296256,0.0439985655,0.4396088421,-0.4373794198,0.2771134973,0.1048197001,-0.0162666105,-0.0870932117,0.2333818823,-0.1232160553,-0.2223884612,-0.0360008366,0.3260163963,0.0275085066,0.0844180807,-0.1239229664,0.151439026,-0.0062860139,-0.1012277976,0.0877758339,-0.2745110989,0.2716356218,-0.3854341209,0.0108550973,0.3584052622,0.2544040084,0.012458425,0.5229176879,0.3607985377,0.1309864819,0.0047399239,0.1596141011,-0.0360543653,0.1644055843,0.2437644601,0.1556479186,-0.2882475257,0.1080994159,0.2060247958,0.4651649594,-0.1006724983,-0.0061814347,-0.0981715322,0.6944160461,-0.1729173362,0.0731927454,0.0689731315,-0.2127287239,0.2158112675,-0.1108153313,-0.2230937481,0.282985568,0.0023209217,0.0603173412,0.2798644006,0.0221882686,0.0645358041,0.1168215051,-0.0624730512,0.0952360928,0.0431768484,-0.279270947,0.1629126817,-0.2930317223,-0.2341297567,-0.0624012202,0.5793997645,-0.1489324272,0.2303509116,-0.4628950655,-0.2045208961,-0.0787363201,-0.3974444568,-0.4104942083,-0.1899440438,-0.0040211538,-0.1877396256,0.1462003291,0.2206022739,-0.5228921771,0.0443228409,0.1362039596,-0.1953602731,-0.126956597,-0.1723050326,-0.3590372503,0.0936886668,0.3631295562,-0.127447322,0.019843448,-0.2478466183,0.254952848,0.2545464635,-0.1223841459,0.0468113832,-0.1097017154,0.0464895442,-0.0705101043,-0.0044845832,-0.1012573391,-0.2933575511,0.3167978823,-0.2502653599,-0.1345066279,0.0111015877,0.2080468982,-0.2112153769,0.0153184813,-0.2787295282,-0.4464733303,-0.1309375763,-0.2526050806,-0.0893685967,0.0770280585,-0.0592101552,0.0670390576,0.2440885305,-0.0313663967,-0.0570359603,-0.1043252647,-0.0945007652,0.480217129,-0.216340512,-0.4246936142,0.1244038418,0.076908879,0.1552265137,-0.1643548012,-0.2264216542,-0.0248855632,-0.2398474067,0.0204472244,-0.156555444,0.0771142468,0.257055223,0.0026030068,-0.0252068192,0.0266012345,0.0749124438,-0.2962510884,-0.200671941,0.1363382488,-0.0919154733,0.1938224882,-0.0410265699,0.1907352209,0.3083446026,0.2765741944,0.540653944,-0.3078549206,0.2696657181,-0.0855971426,-0.2215681523,-0.143584758,-0.0091617992,-0.0120479474,0.2584413886,-0.1071535572,-0.2774141133,-0.2172447294,0.0290748924,-0.4561845064,-0.2686645687,0.023831008,-0.0893477499,-0.0169112552,-0.0891364589,0.1383647025,0.014703813,-0.2177511603,0.068209894,0.2951210141,-0.3689623475,0.2421076298,-0.2130000889,-0.1225254759,-0.2664454281,0.247591272,0.1506994665,0.4262364209,-0.1236199215,-0.0755208731,0.0265095346,-0.1941903979,0.2333406955,-0.1603285968,-0.2908462584,0.1687159836,0.1706056297,0.1095482931,-0.0844957978,-0.3665852249,-0.0097888531,-0.2287279814,-0.0926206484,0.0360179693,-0.0472488813,-0.0161391888,0.1193227097,-0.1108767167,0.1764355153,-0.5984653234,-0.7269862294,-0.2578582168,0.0455533154,0.1361999959,0.55233109,-0.004642081,0.0259218551,-0.2695075572,-0.1148437038,0.3324126303,0.3213649988,0.1127392352,-0.0899587721,0.0383546837,-0.3145249188,-0.135554269,0.4009406567,0.5842265487,0.1452472657,-0.7271106839,-0.0935817808,-0.3014317453,0.4599072337,0.0289451201,-0.2441748232,0.0562446602,0.1340471953,0.0106605822,-0.2896457911,-0.0162553675,0.3499532342,0.0206712168,-0.6139425635,-0.3685237467,0.1280777603,-0.0576741286,0.1933114976,0.1910597533,0.1926906407,-0.3214296401,0.2847873271,-0.0704885051,0.7287112474,0.068506591,0.232834816,-0.0546927601,-0.2685354352,0.0284548122,0.2458198667,0.0890586376,-0.4026048779,-0.2122039944,-0.1592353135,0.2188973129,0.5947277546,0.0032554753,-0.2596394122,0.0598291047,0.0298777856,0.1070341319,0.037523631,-0.0048475196,-0.2880981565,-0.0778032467,-0.4771361053,0.1656149775,-0.1385527849,-0.024462644,-0.1220873594,-0.0433546044,-0.104458645,-0.3514779508,-0.1510646045,0.0526792407,-0.3079713285,0.2075068504,-0.3629766405,-0.4374901056,0.1110944301,0.1455349773,0.3107249737,-0.0231625997,0.0923141688,0.1514184177,-0.095418185,-0.0872194171,-0.0031580473,0.110830605,0.0896452889,-0.1044409871,-0.3822081983,0.2586684823,0.1068968177,-0.0460569449,0.0553289726,0.2123970091,-0.1310105324,-0.2901581824,-0.1119994521,-0.2300676554,0.1816224605,-0.0707860738,0.1582632214,-0.0451896191,-0.0678532273,0.2614602149,-0.0076245042,-0.3708268106,-0.1672649086,0.1855658293,-0.0071158288,-0.4556158781,0.4556062222,-0.0839898065,-0.3163056076,-0.1903872341,0.3991205394,0.1614010334,-0.22005485,-0.2605897188,0.01003478,0.1220600381,0.1406049728,0.1072001383,-0.0155569091,-0.1925764084,0.071719259,-0.4087239206,-0.2751372159,0.1600391716,0.2458664328,0.094780609,0.1726830453,-0.1984698325,0.1485823244,-0.3255025148,-0.2900404036,0.2444757521,-0.0866242275,-0.0245897602,-0.4669628441,-0.0229876284,0.1285440773,0.0860244855,0.2022146434,0.2526505589,-0.012699767,-0.2199431658,-0.0820389241,0.1241432503,0.1190057173,-0.2253132761,0.0365872495,0.0153022725,0.0347682685,-0.1840181947,0.1882107705,0.2569800913,-0.0764494166,0.2404691577,0.1349008977,0.1008123234,-0.1115205437,0.0436901934,-0.0604537874,0.5660281777,0.2114326358,0.1548069566,0.157154575,0.1376875788,0.2274455726,0.0209664553,0.016540911,-0.1112936661,0.3320977986,-0.5293729305,0.2882075608,0.2282841653,0.1719616503,0.5190086961,-0.3012439013,-0.0315728672,0.4205056429,0.1547213793,-0.2704833448,0.0710881278,0.1919422299,-0.121701628,-0.2566027045,0.3519894481,0.195398733,-0.027906958,0.0902830884,-0.1604171693,0.702839911,0.0118458057,0.2560504675,0.2967597544,0.0514383018,0.1482069343,0.0243935566,0.178605631,0.007100754,1.0043616295,-0.5104754567,0.3897270858,-0.0343943872,0.1863347739,-0.1949543208,-0.2114784569,0.0562511496,-0.1234765202,-0.1014328152,0.2496176064,0.1374088824,0.2604541779,-0.3318978846,-0.0127904704,-0.1395208091,-0.1367085427,-0.2057547122,0.1248675287,0.1255959272,0.0101653207,-0.1870059371,-0.0602751821,0.1468210518,-0.2216507047,0.1812556386,0.1561868042,-0.1281467825,-0.0791221783,0.0163754039,0.0753801242,0.1020826623,-0.1398314387,0.1466614455,0.4155284166,0.1663506627,0.1675386429,0.1882580966,0.513690114,0.4934977889,-0.330716759,0.113944456,-0.1418102384,0.0140924416,0.0935492814,0.2457866222,0.2724190354,0.8603190184,0.034591306,0.1887731403,-0.0464425981,-0.2097913772,0.1136624441,0.1802144647,0.0939098373,0.1978257895,-0.1627750099,0.1165151671,-0.1892609596,-0.0461454466,-0.4863938987,0.3011235297,0.3250532448,0.1874185205,0.0117905652,-0.0385037363,0.0622423291,0.1979704648,0.2654522061,0.3120368719,0.1771979928,-0.0010989318,-0.234133929,-0.6304964423,0.287078321,-0.0147751393,0.0629694387,-0.0836824998,-0.1353456676,-0.0116071068,0.1800624728,0.2472201139,0.0226373971,0.1034061909,0.2277410328,-0.2585066855,-0.118205145,0.0679697841,0.0853260159,-0.1791678816,-0.1498943716,0.1264433563,-0.1880733073,0.0745991096,-0.1221612021,-0.232432723,-0.3099809289,-0.2677236795,0.4300990105,0.3035269678,0.6953615546,-0.1201181635,0.1233739853,-0.2261851579,-0.2087058723,0.0519554354,0.6460496187,0.1535214931,-0.0110486913,-0.0754143894,-0.3035059273,-0.3414487243,0.4483339787,0.0505028181,0.1355827004,-0.2378664166,-0.0854559839,0.3214973509,0.0148258843,-0.0272913147,0.1159836203,-0.0649366304,0.2085013688,-0.1612676233,-0.434489727,0.205968082,-0.3151429296,-0.5329882503,-0.0730269253,0.0878177211,0.0448532626,-0.0629050583,-0.8263720274,0.1969081461,0.214216888,-0.1804354936,-0.1294161081,0.2301895171,-0.2065946907,0.001194089,-0.0582196787,0.6145447493,-0.1501440704,-0.0995580256,0.2943917513,-0.2253649086]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1915","title":"Unable to download `wiki_dpr`","comments":"Thanks for reporting ! This is a bug. For now feel free to set `ignore_verifications=False` in `load_dataset`.\r\nI'm working on a fix","body":"I am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings\/no index. In order to do so, I ran:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", embeddings_name=\"multiset\", index_name=\"no_index\")` \r\n\r\nHowever, I got the following error:\r\n`datasets.utils.info_utils.UnexpectedDownloadedFile: {'embeddings_index'}`\r\n\r\nI tried adding in flags `with_embeddings=False` and `with_index=False`:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False, embeddings_name=\"multiset\", index_name=\"no_index\")`\r\n\r\nBut I got the following error:\r\n`raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums)))\r\ndatasets.utils.info_utils.ExpectedMoreDownloadedFiles: {\u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_5\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_15\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_30\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_36\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_18\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_41\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_13\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_48\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_10\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_23\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_14\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_34\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_43\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_40\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_47\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_3\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_24\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_7\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_33\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_46\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_42\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_27\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_29\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_26\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_22\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_4\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_20\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_39\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_6\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_16\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_8\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_35\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_49\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_17\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_25\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_0\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_38\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_12\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_44\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_1\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_32\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_19\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_31\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_37\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_9\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_11\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_21\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_28\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_45\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_2\u2019}`\r\n\r\nIs there anything else I need to set to download the dataset?\r\n\r\n**UPDATE**: just running `curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False)` gives me the same error.\r\n","comment_length":22,"text":"Unable to download `wiki_dpr`\nI am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings\/no index. In order to do so, I ran:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", embeddings_name=\"multiset\", index_name=\"no_index\")` \r\n\r\nHowever, I got the following error:\r\n`datasets.utils.info_utils.UnexpectedDownloadedFile: {'embeddings_index'}`\r\n\r\nI tried adding in flags `with_embeddings=False` and `with_index=False`:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False, embeddings_name=\"multiset\", index_name=\"no_index\")`\r\n\r\nBut I got the following error:\r\n`raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums)))\r\ndatasets.utils.info_utils.ExpectedMoreDownloadedFiles: {\u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_5\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_15\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_30\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_36\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_18\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_41\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_13\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_48\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_10\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_23\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_14\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_34\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_43\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_40\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_47\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_3\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_24\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_7\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_33\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_46\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_42\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_27\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_29\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_26\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_22\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_4\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_20\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_39\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_6\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_16\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_8\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_35\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_49\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_17\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_25\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_0\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_38\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_12\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_44\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_1\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_32\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_19\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_31\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_37\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_9\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_11\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_21\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_28\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_45\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_2\u2019}`\r\n\r\nIs there anything else I need to set to download the dataset?\r\n\r\n**UPDATE**: just running `curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False)` gives me the same error.\r\n\nThanks for reporting ! This is a bug. For now feel free to set `ignore_verifications=False` in `load_dataset`.\r\nI'm working on a fix","embeddings":[-0.3591079712,-0.3630983829,-0.0504424088,0.3148263693,0.3717034757,0.4270488024,0.3198949397,0.0625339374,0.2385216206,0.1804990321,0.1647016555,-0.0206091162,0.0954112709,-0.0210907198,-0.1673092544,-0.0334004611,0.1001311168,0.1070557833,0.3710588217,-0.0622109734,-0.329595685,0.2548207641,-0.3576647937,-0.231895268,-0.1536287069,-0.143520534,-0.0590852164,-0.0766759217,-0.4447588623,-0.4724355638,0.5257073641,0.2139309347,0.456645906,0.2094175071,-0.000123334,-0.0081851818,0.2570953667,-0.276065439,-0.2267439216,-0.1994098872,-0.5099616051,-0.2134321183,0.1228160188,-0.2729622424,0.2952129245,-0.2128572315,-0.0749510676,-0.0928959846,0.074530825,0.1510064006,0.1449033767,-0.0625398681,0.1508243978,-0.2355563939,0.0945451707,-0.1769310385,-0.1318905056,0.1710799336,-0.0272569992,0.1341178864,0.2691080272,0.0560868569,-0.2518814504,0.2277733535,0.1170533374,0.0164909512,0.2614754736,-0.304164052,0.4016195536,0.5526133776,0.923730731,0.1302185357,-0.2386934608,0.0382870622,0.211455673,-0.3757750988,0.1635375172,0.2258236408,-0.2530870736,0.200611487,-0.2965897322,-0.0683538169,-0.0212516002,0.3237015009,-0.0989618227,0.3529938459,0.2022208571,0.1463739276,0.1856656522,-0.2036290169,0.0743476376,-0.1115095615,-0.2810247838,0.0874456316,-0.2282613516,0.0073461859,-0.1403028518,0.0571529195,0.2315741628,0.3335160315,0.0137926796,0.1705392301,0.4314410686,0.1014831588,0.2653803229,0.2508428395,-0.1195526347,0.2461987734,0.3145948052,0.547288239,0.2549657226,-0.0376866497,-0.0095993467,-0.0076847337,-0.0618706718,-0.1101422459,-0.0867578164,-0.3650256991,-0.299870044,0.1207373217,-0.2721976042,-0.1857746094,-0.2455056757,0.3989276588,0.0512363166,0.4941689968,0.1345018893,0.1692253351,-0.0389331952,0.045549307,-0.096598126,-0.0991690308,-0.1295962483,0.0827541575,0.3832916021,-0.3321835995,0.3954805434,0.0359180942,0.0670721903,0.0898475721,-0.0672724321,-0.0200424772,0.2051387131,0.3828620315,0.0285898894,0.2937918305,0.107425645,-0.2261042297,-0.0246159267,0.2774973512,-0.1818651706,-0.4335758388,-0.0587910637,0.124901168,-0.2760751247,-0.0717422366,-0.3263556063,0.1071712226,0.0178034566,-0.1447193474,0.158366397,0.0138511891,-0.2004513741,-0.3192541003,0.3162532151,0.5299692154,-0.53788656,-0.0565230511,-0.0876784995,-0.2763260901,0.1355169863,0.0153356371,-0.2887880504,-0.0896552205,-0.3782499731,0.1199998185,0.1308039427,-0.4048190117,-0.5335974097,0.4260773361,0.0165747441,0.1425583661,0.3088404834,0.0687623024,0.2286393195,0.0907653645,-0.109613128,0.360917598,0.1335101128,-0.2713665366,-0.0036884013,-0.3318436444,0.3804715872,0.0646600276,-0.0475530401,-0.0033561576,0.118311286,0.1810515225,0.4451654851,0.0042786784,0.031223679,0.2706511021,0.1180464923,0.0901454762,0.1339274943,0.1229389459,-0.51377666,0.2290322632,0.1504512876,0.0457544737,-0.3320191801,-0.094936803,-0.3201955557,-0.0873651728,-0.3744758666,-0.2137416899,-0.0082012471,0.1704927832,-0.0325004943,-0.1602595001,-0.0984629914,0.0415371247,-0.05279148,0.0277336389,-0.1782583445,0.1469732225,-0.1374594718,0.014372942,0.1123541743,-0.0683239251,0.318995446,-0.0505682491,0.0679063499,0.3047980666,0.2043274045,-0.0166690685,-0.0066899275,-0.005671245,0.258731097,-0.5035472512,-0.0205891393,0.3803394437,0.1576838642,-0.0605140664,-0.2632324398,0.224481225,-0.0290870704,0.3685749769,-0.2061018944,0.0416202657,0.3276460767,-0.1760827154,0.2287088186,-0.3677020967,0.2900660038,0.5791848898,-0.2493056208,-0.1534954756,-0.0482501648,0.0391812027,0.1322937459,-0.2087436318,-0.0871027559,0.0236714669,0.0552035943,-0.0481585823,0.1448453516,0.0600990131,0.3681735992,0.247254774,0.0936580449,-0.1286984682,0.1636072844,-0.3218720853,0.2134936303,-0.0119457636,-0.1579790264,0.5060125589,-0.0821752176,-0.0530164614,-0.2347956598,-0.0330483504,0.2623149455,0.2831532955,-0.3046387136,-0.1779405475,-0.6171994209,0.0053771483,0.0949762836,-0.1705055088,-0.2037487328,-0.2974213064,0.1051949412,-0.074800849,0.124833256,0.1406850666,-0.2971715331,-0.1109372154,-0.1286422014,-0.202541694,-0.0407080688,-0.0536280163,-0.331849426,-0.0249450263,0.2382584959,0.3014809191,0.2516848743,0.058719039,-0.3458539844,-0.3976098895,-0.0931589454,0.0518293828,-0.1394761503,0.0614196137,0.336607933,0.481333077,-0.0528129339,-0.3433493674,0.0972987488,-0.0614315197,-0.0334287025,-0.0621375032,-0.0878678113,-0.0453506224,-0.1144122407,-0.6167302728,-0.1330221742,-0.3135217726,-0.0966890827,0.0850574449,-0.0683745295,-0.0727720186,0.1985067874,-0.1263793856,0.2616435885,-0.0850995928,-0.1912420541,-0.1607696265,0.5849156976,-0.1287938356,-0.3757933974,0.5222776532,0.0111422595,-0.0262258314,0.3314102292,-0.5823156238,0.1069680825,0.1025796235,0.2082475722,0.0025481533,0.2310561985,0.305845499,0.0286073294,-0.0586488731,-0.0875662118,-0.093996942,-0.0564806722,-0.2831987143,0.509218514,-0.0105568729,0.2368230969,-0.1851426214,0.6094356179,0.3000097275,0.1178231388,0.3601537943,-0.0355622433,0.1675811857,0.1629151553,-0.1403379142,0.0368618518,-0.1783125699,0.3249070346,0.2848487496,-0.161845535,-0.3321040869,-0.3920755088,0.1011412516,-0.0891899094,-0.0727448985,-0.0491276309,0.3427481949,0.1343055069,-0.1356298327,0.302932173,0.0821175426,-0.3334839344,0.2686665356,0.2534911036,-0.1429877877,0.1484562904,0.1589110345,-0.0379800834,-0.3448107541,0.3444143534,0.1638859808,0.3087622523,0.0045746486,0.1655189246,0.1895901561,-0.0242543779,0.5605655909,-0.4350599051,0.3597978055,0.1688041836,0.0517789498,-0.2673882544,-0.1922028959,-0.2611429691,-0.276125133,-0.1781174242,0.2215978503,-0.6027700901,0.0161780752,0.1669195592,0.1612285525,-0.1238549352,-0.359395355,-0.3704822659,-0.4479207993,-0.4778935909,-0.1964293718,0.0075105694,0.3619404733,-0.2136188,-0.0925567895,-0.0091365511,0.1171005815,-0.0015813193,-0.0716375336,0.243196398,0.4718736708,-0.0696359649,0.2932132781,0.1787507236,0.5283367634,0.6097086668,-0.0627239421,-0.1071096957,0.048447229,-0.2018221468,0.0192043819,0.2239023149,-0.0920598656,0.1804151833,0.1474184692,0.0573410019,-0.0472669639,-0.2881349623,0.3782993555,-0.2557798028,-0.3657977283,-0.4157612324,0.4662528336,0.0413501114,0.1092932597,0.1936930418,0.2587582171,-0.1668642461,0.0226928052,0.35775581,1.0175694227,0.1065527275,-0.0188474376,0.2060739398,0.0146241486,0.2137940377,-0.2915165722,0.2040702999,-0.0980691463,-0.4103773236,-0.2600039244,0.0024992549,0.1532847285,0.3295856416,-0.4880572259,0.3298540115,-0.2777177691,0.1029102132,-0.0117820604,0.2915962636,-0.2209283859,-0.1896228939,-0.2440835685,0.0250631962,0.106212154,0.1561797112,-0.1630843133,-0.0633788332,0.0575042218,-0.3713615537,-0.4828009009,0.0583305024,-0.2181004435,0.4266487062,-0.3948787451,-0.2592290044,0.2925424874,0.4851695001,0.1716979891,-0.0103366533,-0.3101628423,0.2248512655,-0.12282224,-0.4960898459,-0.2240558714,0.0299589187,0.0838920698,-0.0808789134,-0.5226157904,0.1020061448,-0.0371094123,-0.0976731181,0.06950441,-0.0406803861,-0.1487551332,-0.2177066952,-0.0691503957,-0.0016990738,0.1558312923,-0.0840194449,0.0362307951,-0.001989631,-0.1585756689,0.1211213395,0.1971923411,-0.1881342381,-0.1666731983,0.6800674796,0.0213053767,-0.2687130272,0.6212598085,0.2700417042,-0.3788090646,-0.2264257371,-0.0876422077,0.1271988004,-0.2000413984,0.0818250701,0.3339788318,0.1755504757,0.1250366569,0.4918984771,0.211436078,0.0397883654,-0.0355801135,-0.5814552307,-0.0940757543,0.0889578387,-0.0807976201,0.188050881,-0.4698204398,-0.3154388368,0.2788203657,0.0934570134,-0.1850284189,0.0440964662,-0.4358551204,-0.0807070807,0.1609925628,0.0234947689,0.1360045671,0.1081752554,-0.0183176082,0.051795017,-0.0362442806,-0.0480694063,-0.0843037143,0.2136483788,0.0365364328,-0.1229959428,0.0750056878,-0.4660222232,-0.3347616196,0.01751546,0.2713816166,0.3464633524,0.011237816,0.0264794286,0.3368587792,0.0637375265,-0.5087180138,0.0850754827,0.1238567755,0.3832836747,0.3522448242,-0.0603408739,0.0815015659,0.0552273132,-0.0290255304,0.0364017412,0.1451057494,-0.0185582973,0.4161353409,-0.2846483886,-0.0397414342,0.3598666489,0.080249615,0.4701498449,-0.1421664953,0.1650274247,0.1924473643,0.0728161857,-0.4741274416,-0.2790483236,0.3225473166,0.0880576894,-0.1475120634,0.0822515488,0.1976872981,-0.5292577744,-0.1985706389,0.0116609186,0.2377842665,-0.2428717166,0.1522594392,0.5658448935,-0.0603007786,0.026681494,-0.1705094576,0.1049014479,0.3029583395,-0.022607306,-0.3302282989,0.3238980174,-0.0517212488,0.2576176524,-0.2167513371,-0.5816491246,0.1976688504,0.2486557811,0.1091314703,-0.0623876937,0.1428005546,0.2813912332,0.1161995754,0.1848139763,-0.5019873381,0.1146367341,-0.0847099647,-0.0868122652,-0.1550206989,-0.1631114036,-0.3408384025,0.3697252572,0.2506076992,-0.3877323568,0.1391340941,0.2233997583,-0.3519901037,-0.3351708353,-0.0046411222,0.0475750901,0.1984971613,-0.2501712441,-0.0406993702,0.4264978766,-0.0355310589,0.5507053137,0.1158855483,0.2798188329,0.1993611753,-0.0190243442,0.0986388847,0.1593259424,0.0265295263,0.0264789164,0.3988074362,0.1845425963,0.1278179884,0.1725085676,-0.0023972353,-0.1505349725,-0.3022055626,0.1991573721,0.1582883149,0.1563813984,-0.0906068981,0.0002232369,0.1193928644,-0.1225259081,-0.0520095155,-0.2637984157,0.2559801638,0.2613166571,0.096773833,0.0997288898,0.1218362823,-0.0077642389,-0.2113354355,0.4893983305,0.3439127505,-0.0329992585,-0.3170140982,-0.165727362,-0.8635811806,0.1909140944,-0.0018567108,-0.1987917572,0.2767450213,0.1448274702,0.0629281104,0.1501017809,0.0517331958,0.0493551344,0.3558023274,0.5278563499,-0.0388547406,-0.3366440237,0.0577609614,0.0212414321,-0.2305419147,-0.2928699553,0.0717013255,-0.0456007048,-0.1349543482,-0.1366087794,-0.0527431183,-0.1673202962,0.2791327536,0.2438106388,0.0376392491,0.6168964505,-0.0421943404,0.0340050422,-0.2160889953,0.0339432918,-0.1781509966,0.3692152798,0.1981420517,0.040288534,-0.2612348199,0.1893491596,-0.2551579475,0.1336954534,0.1048166454,0.3364696503,-0.4344368279,-0.0075040176,-0.0616459772,0.0245793927,0.1752414554,-0.1018626615,-0.1526125222,-0.0320739858,-0.1232203245,-0.3546847999,0.6694049835,-0.445109129,-0.1035618335,-0.0715580881,0.1026608571,-0.2589513361,0.1466145515,-0.782879591,0.2994582653,0.2722882926,-0.0191888474,-0.5205340385,0.0897959992,-0.196808368,0.0039333892,-0.1593840271,0.1601531953,0.1927386224,-0.2259242535,-0.1817126572,-0.21335195]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1915","title":"Unable to download `wiki_dpr`","comments":"I just merged a fix :)\r\n\r\nWe'll do a patch release soon. In the meantime feel free to try it from the master branch\r\nThanks again for reporting !","body":"I am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings\/no index. In order to do so, I ran:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", embeddings_name=\"multiset\", index_name=\"no_index\")` \r\n\r\nHowever, I got the following error:\r\n`datasets.utils.info_utils.UnexpectedDownloadedFile: {'embeddings_index'}`\r\n\r\nI tried adding in flags `with_embeddings=False` and `with_index=False`:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False, embeddings_name=\"multiset\", index_name=\"no_index\")`\r\n\r\nBut I got the following error:\r\n`raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums)))\r\ndatasets.utils.info_utils.ExpectedMoreDownloadedFiles: {\u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_5\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_15\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_30\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_36\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_18\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_41\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_13\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_48\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_10\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_23\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_14\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_34\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_43\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_40\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_47\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_3\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_24\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_7\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_33\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_46\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_42\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_27\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_29\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_26\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_22\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_4\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_20\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_39\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_6\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_16\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_8\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_35\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_49\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_17\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_25\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_0\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_38\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_12\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_44\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_1\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_32\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_19\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_31\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_37\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_9\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_11\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_21\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_28\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_45\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_2\u2019}`\r\n\r\nIs there anything else I need to set to download the dataset?\r\n\r\n**UPDATE**: just running `curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False)` gives me the same error.\r\n","comment_length":29,"text":"Unable to download `wiki_dpr`\nI am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings\/no index. In order to do so, I ran:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", embeddings_name=\"multiset\", index_name=\"no_index\")` \r\n\r\nHowever, I got the following error:\r\n`datasets.utils.info_utils.UnexpectedDownloadedFile: {'embeddings_index'}`\r\n\r\nI tried adding in flags `with_embeddings=False` and `with_index=False`:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False, embeddings_name=\"multiset\", index_name=\"no_index\")`\r\n\r\nBut I got the following error:\r\n`raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums)))\r\ndatasets.utils.info_utils.ExpectedMoreDownloadedFiles: {\u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_5\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_15\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_30\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_36\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_18\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_41\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_13\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_48\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_10\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_23\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_14\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_34\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_43\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_40\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_47\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_3\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_24\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_7\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_33\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_46\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_42\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_27\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_29\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_26\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_22\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_4\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_20\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_39\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_6\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_16\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_8\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_35\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_49\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_17\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_25\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_0\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_38\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_12\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_44\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_1\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_32\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_19\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_31\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_37\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_9\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_11\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_21\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_28\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_45\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_2\u2019}`\r\n\r\nIs there anything else I need to set to download the dataset?\r\n\r\n**UPDATE**: just running `curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False)` gives me the same error.\r\n\nI just merged a fix :)\r\n\r\nWe'll do a patch release soon. In the meantime feel free to try it from the master branch\r\nThanks again for reporting !","embeddings":[-0.3591079712,-0.3630983829,-0.0504424088,0.3148263693,0.3717034757,0.4270488024,0.3198949397,0.0625339374,0.2385216206,0.1804990321,0.1647016555,-0.0206091162,0.0954112709,-0.0210907198,-0.1673092544,-0.0334004611,0.1001311168,0.1070557833,0.3710588217,-0.0622109734,-0.329595685,0.2548207641,-0.3576647937,-0.231895268,-0.1536287069,-0.143520534,-0.0590852164,-0.0766759217,-0.4447588623,-0.4724355638,0.5257073641,0.2139309347,0.456645906,0.2094175071,-0.000123334,-0.0081851818,0.2570953667,-0.276065439,-0.2267439216,-0.1994098872,-0.5099616051,-0.2134321183,0.1228160188,-0.2729622424,0.2952129245,-0.2128572315,-0.0749510676,-0.0928959846,0.074530825,0.1510064006,0.1449033767,-0.0625398681,0.1508243978,-0.2355563939,0.0945451707,-0.1769310385,-0.1318905056,0.1710799336,-0.0272569992,0.1341178864,0.2691080272,0.0560868569,-0.2518814504,0.2277733535,0.1170533374,0.0164909512,0.2614754736,-0.304164052,0.4016195536,0.5526133776,0.923730731,0.1302185357,-0.2386934608,0.0382870622,0.211455673,-0.3757750988,0.1635375172,0.2258236408,-0.2530870736,0.200611487,-0.2965897322,-0.0683538169,-0.0212516002,0.3237015009,-0.0989618227,0.3529938459,0.2022208571,0.1463739276,0.1856656522,-0.2036290169,0.0743476376,-0.1115095615,-0.2810247838,0.0874456316,-0.2282613516,0.0073461859,-0.1403028518,0.0571529195,0.2315741628,0.3335160315,0.0137926796,0.1705392301,0.4314410686,0.1014831588,0.2653803229,0.2508428395,-0.1195526347,0.2461987734,0.3145948052,0.547288239,0.2549657226,-0.0376866497,-0.0095993467,-0.0076847337,-0.0618706718,-0.1101422459,-0.0867578164,-0.3650256991,-0.299870044,0.1207373217,-0.2721976042,-0.1857746094,-0.2455056757,0.3989276588,0.0512363166,0.4941689968,0.1345018893,0.1692253351,-0.0389331952,0.045549307,-0.096598126,-0.0991690308,-0.1295962483,0.0827541575,0.3832916021,-0.3321835995,0.3954805434,0.0359180942,0.0670721903,0.0898475721,-0.0672724321,-0.0200424772,0.2051387131,0.3828620315,0.0285898894,0.2937918305,0.107425645,-0.2261042297,-0.0246159267,0.2774973512,-0.1818651706,-0.4335758388,-0.0587910637,0.124901168,-0.2760751247,-0.0717422366,-0.3263556063,0.1071712226,0.0178034566,-0.1447193474,0.158366397,0.0138511891,-0.2004513741,-0.3192541003,0.3162532151,0.5299692154,-0.53788656,-0.0565230511,-0.0876784995,-0.2763260901,0.1355169863,0.0153356371,-0.2887880504,-0.0896552205,-0.3782499731,0.1199998185,0.1308039427,-0.4048190117,-0.5335974097,0.4260773361,0.0165747441,0.1425583661,0.3088404834,0.0687623024,0.2286393195,0.0907653645,-0.109613128,0.360917598,0.1335101128,-0.2713665366,-0.0036884013,-0.3318436444,0.3804715872,0.0646600276,-0.0475530401,-0.0033561576,0.118311286,0.1810515225,0.4451654851,0.0042786784,0.031223679,0.2706511021,0.1180464923,0.0901454762,0.1339274943,0.1229389459,-0.51377666,0.2290322632,0.1504512876,0.0457544737,-0.3320191801,-0.094936803,-0.3201955557,-0.0873651728,-0.3744758666,-0.2137416899,-0.0082012471,0.1704927832,-0.0325004943,-0.1602595001,-0.0984629914,0.0415371247,-0.05279148,0.0277336389,-0.1782583445,0.1469732225,-0.1374594718,0.014372942,0.1123541743,-0.0683239251,0.318995446,-0.0505682491,0.0679063499,0.3047980666,0.2043274045,-0.0166690685,-0.0066899275,-0.005671245,0.258731097,-0.5035472512,-0.0205891393,0.3803394437,0.1576838642,-0.0605140664,-0.2632324398,0.224481225,-0.0290870704,0.3685749769,-0.2061018944,0.0416202657,0.3276460767,-0.1760827154,0.2287088186,-0.3677020967,0.2900660038,0.5791848898,-0.2493056208,-0.1534954756,-0.0482501648,0.0391812027,0.1322937459,-0.2087436318,-0.0871027559,0.0236714669,0.0552035943,-0.0481585823,0.1448453516,0.0600990131,0.3681735992,0.247254774,0.0936580449,-0.1286984682,0.1636072844,-0.3218720853,0.2134936303,-0.0119457636,-0.1579790264,0.5060125589,-0.0821752176,-0.0530164614,-0.2347956598,-0.0330483504,0.2623149455,0.2831532955,-0.3046387136,-0.1779405475,-0.6171994209,0.0053771483,0.0949762836,-0.1705055088,-0.2037487328,-0.2974213064,0.1051949412,-0.074800849,0.124833256,0.1406850666,-0.2971715331,-0.1109372154,-0.1286422014,-0.202541694,-0.0407080688,-0.0536280163,-0.331849426,-0.0249450263,0.2382584959,0.3014809191,0.2516848743,0.058719039,-0.3458539844,-0.3976098895,-0.0931589454,0.0518293828,-0.1394761503,0.0614196137,0.336607933,0.481333077,-0.0528129339,-0.3433493674,0.0972987488,-0.0614315197,-0.0334287025,-0.0621375032,-0.0878678113,-0.0453506224,-0.1144122407,-0.6167302728,-0.1330221742,-0.3135217726,-0.0966890827,0.0850574449,-0.0683745295,-0.0727720186,0.1985067874,-0.1263793856,0.2616435885,-0.0850995928,-0.1912420541,-0.1607696265,0.5849156976,-0.1287938356,-0.3757933974,0.5222776532,0.0111422595,-0.0262258314,0.3314102292,-0.5823156238,0.1069680825,0.1025796235,0.2082475722,0.0025481533,0.2310561985,0.305845499,0.0286073294,-0.0586488731,-0.0875662118,-0.093996942,-0.0564806722,-0.2831987143,0.509218514,-0.0105568729,0.2368230969,-0.1851426214,0.6094356179,0.3000097275,0.1178231388,0.3601537943,-0.0355622433,0.1675811857,0.1629151553,-0.1403379142,0.0368618518,-0.1783125699,0.3249070346,0.2848487496,-0.161845535,-0.3321040869,-0.3920755088,0.1011412516,-0.0891899094,-0.0727448985,-0.0491276309,0.3427481949,0.1343055069,-0.1356298327,0.302932173,0.0821175426,-0.3334839344,0.2686665356,0.2534911036,-0.1429877877,0.1484562904,0.1589110345,-0.0379800834,-0.3448107541,0.3444143534,0.1638859808,0.3087622523,0.0045746486,0.1655189246,0.1895901561,-0.0242543779,0.5605655909,-0.4350599051,0.3597978055,0.1688041836,0.0517789498,-0.2673882544,-0.1922028959,-0.2611429691,-0.276125133,-0.1781174242,0.2215978503,-0.6027700901,0.0161780752,0.1669195592,0.1612285525,-0.1238549352,-0.359395355,-0.3704822659,-0.4479207993,-0.4778935909,-0.1964293718,0.0075105694,0.3619404733,-0.2136188,-0.0925567895,-0.0091365511,0.1171005815,-0.0015813193,-0.0716375336,0.243196398,0.4718736708,-0.0696359649,0.2932132781,0.1787507236,0.5283367634,0.6097086668,-0.0627239421,-0.1071096957,0.048447229,-0.2018221468,0.0192043819,0.2239023149,-0.0920598656,0.1804151833,0.1474184692,0.0573410019,-0.0472669639,-0.2881349623,0.3782993555,-0.2557798028,-0.3657977283,-0.4157612324,0.4662528336,0.0413501114,0.1092932597,0.1936930418,0.2587582171,-0.1668642461,0.0226928052,0.35775581,1.0175694227,0.1065527275,-0.0188474376,0.2060739398,0.0146241486,0.2137940377,-0.2915165722,0.2040702999,-0.0980691463,-0.4103773236,-0.2600039244,0.0024992549,0.1532847285,0.3295856416,-0.4880572259,0.3298540115,-0.2777177691,0.1029102132,-0.0117820604,0.2915962636,-0.2209283859,-0.1896228939,-0.2440835685,0.0250631962,0.106212154,0.1561797112,-0.1630843133,-0.0633788332,0.0575042218,-0.3713615537,-0.4828009009,0.0583305024,-0.2181004435,0.4266487062,-0.3948787451,-0.2592290044,0.2925424874,0.4851695001,0.1716979891,-0.0103366533,-0.3101628423,0.2248512655,-0.12282224,-0.4960898459,-0.2240558714,0.0299589187,0.0838920698,-0.0808789134,-0.5226157904,0.1020061448,-0.0371094123,-0.0976731181,0.06950441,-0.0406803861,-0.1487551332,-0.2177066952,-0.0691503957,-0.0016990738,0.1558312923,-0.0840194449,0.0362307951,-0.001989631,-0.1585756689,0.1211213395,0.1971923411,-0.1881342381,-0.1666731983,0.6800674796,0.0213053767,-0.2687130272,0.6212598085,0.2700417042,-0.3788090646,-0.2264257371,-0.0876422077,0.1271988004,-0.2000413984,0.0818250701,0.3339788318,0.1755504757,0.1250366569,0.4918984771,0.211436078,0.0397883654,-0.0355801135,-0.5814552307,-0.0940757543,0.0889578387,-0.0807976201,0.188050881,-0.4698204398,-0.3154388368,0.2788203657,0.0934570134,-0.1850284189,0.0440964662,-0.4358551204,-0.0807070807,0.1609925628,0.0234947689,0.1360045671,0.1081752554,-0.0183176082,0.051795017,-0.0362442806,-0.0480694063,-0.0843037143,0.2136483788,0.0365364328,-0.1229959428,0.0750056878,-0.4660222232,-0.3347616196,0.01751546,0.2713816166,0.3464633524,0.011237816,0.0264794286,0.3368587792,0.0637375265,-0.5087180138,0.0850754827,0.1238567755,0.3832836747,0.3522448242,-0.0603408739,0.0815015659,0.0552273132,-0.0290255304,0.0364017412,0.1451057494,-0.0185582973,0.4161353409,-0.2846483886,-0.0397414342,0.3598666489,0.080249615,0.4701498449,-0.1421664953,0.1650274247,0.1924473643,0.0728161857,-0.4741274416,-0.2790483236,0.3225473166,0.0880576894,-0.1475120634,0.0822515488,0.1976872981,-0.5292577744,-0.1985706389,0.0116609186,0.2377842665,-0.2428717166,0.1522594392,0.5658448935,-0.0603007786,0.026681494,-0.1705094576,0.1049014479,0.3029583395,-0.022607306,-0.3302282989,0.3238980174,-0.0517212488,0.2576176524,-0.2167513371,-0.5816491246,0.1976688504,0.2486557811,0.1091314703,-0.0623876937,0.1428005546,0.2813912332,0.1161995754,0.1848139763,-0.5019873381,0.1146367341,-0.0847099647,-0.0868122652,-0.1550206989,-0.1631114036,-0.3408384025,0.3697252572,0.2506076992,-0.3877323568,0.1391340941,0.2233997583,-0.3519901037,-0.3351708353,-0.0046411222,0.0475750901,0.1984971613,-0.2501712441,-0.0406993702,0.4264978766,-0.0355310589,0.5507053137,0.1158855483,0.2798188329,0.1993611753,-0.0190243442,0.0986388847,0.1593259424,0.0265295263,0.0264789164,0.3988074362,0.1845425963,0.1278179884,0.1725085676,-0.0023972353,-0.1505349725,-0.3022055626,0.1991573721,0.1582883149,0.1563813984,-0.0906068981,0.0002232369,0.1193928644,-0.1225259081,-0.0520095155,-0.2637984157,0.2559801638,0.2613166571,0.096773833,0.0997288898,0.1218362823,-0.0077642389,-0.2113354355,0.4893983305,0.3439127505,-0.0329992585,-0.3170140982,-0.165727362,-0.8635811806,0.1909140944,-0.0018567108,-0.1987917572,0.2767450213,0.1448274702,0.0629281104,0.1501017809,0.0517331958,0.0493551344,0.3558023274,0.5278563499,-0.0388547406,-0.3366440237,0.0577609614,0.0212414321,-0.2305419147,-0.2928699553,0.0717013255,-0.0456007048,-0.1349543482,-0.1366087794,-0.0527431183,-0.1673202962,0.2791327536,0.2438106388,0.0376392491,0.6168964505,-0.0421943404,0.0340050422,-0.2160889953,0.0339432918,-0.1781509966,0.3692152798,0.1981420517,0.040288534,-0.2612348199,0.1893491596,-0.2551579475,0.1336954534,0.1048166454,0.3364696503,-0.4344368279,-0.0075040176,-0.0616459772,0.0245793927,0.1752414554,-0.1018626615,-0.1526125222,-0.0320739858,-0.1232203245,-0.3546847999,0.6694049835,-0.445109129,-0.1035618335,-0.0715580881,0.1026608571,-0.2589513361,0.1466145515,-0.782879591,0.2994582653,0.2722882926,-0.0191888474,-0.5205340385,0.0897959992,-0.196808368,0.0039333892,-0.1593840271,0.1601531953,0.1927386224,-0.2259242535,-0.1817126572,-0.21335195]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1915","title":"Unable to download `wiki_dpr`","comments":"Closing since this has been fixed by #1925","body":"I am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings\/no index. In order to do so, I ran:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", embeddings_name=\"multiset\", index_name=\"no_index\")` \r\n\r\nHowever, I got the following error:\r\n`datasets.utils.info_utils.UnexpectedDownloadedFile: {'embeddings_index'}`\r\n\r\nI tried adding in flags `with_embeddings=False` and `with_index=False`:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False, embeddings_name=\"multiset\", index_name=\"no_index\")`\r\n\r\nBut I got the following error:\r\n`raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums)))\r\ndatasets.utils.info_utils.ExpectedMoreDownloadedFiles: {\u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_5\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_15\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_30\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_36\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_18\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_41\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_13\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_48\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_10\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_23\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_14\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_34\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_43\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_40\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_47\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_3\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_24\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_7\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_33\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_46\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_42\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_27\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_29\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_26\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_22\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_4\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_20\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_39\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_6\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_16\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_8\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_35\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_49\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_17\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_25\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_0\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_38\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_12\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_44\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_1\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_32\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_19\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_31\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_37\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_9\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_11\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_21\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_28\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_45\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_2\u2019}`\r\n\r\nIs there anything else I need to set to download the dataset?\r\n\r\n**UPDATE**: just running `curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False)` gives me the same error.\r\n","comment_length":8,"text":"Unable to download `wiki_dpr`\nI am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings\/no index. In order to do so, I ran:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", embeddings_name=\"multiset\", index_name=\"no_index\")` \r\n\r\nHowever, I got the following error:\r\n`datasets.utils.info_utils.UnexpectedDownloadedFile: {'embeddings_index'}`\r\n\r\nI tried adding in flags `with_embeddings=False` and `with_index=False`:\r\n\r\n`curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False, embeddings_name=\"multiset\", index_name=\"no_index\")`\r\n\r\nBut I got the following error:\r\n`raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums)))\r\ndatasets.utils.info_utils.ExpectedMoreDownloadedFiles: {\u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_5\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_15\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_30\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_36\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_18\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_41\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_13\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_48\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_10\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_23\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_14\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_34\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_43\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_40\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_47\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_3\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_24\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_7\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_33\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_46\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_42\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_27\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_29\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_26\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_22\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_4\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_20\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_39\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_6\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_16\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_8\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_35\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_49\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_17\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_25\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_0\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_38\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_12\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_44\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_1\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_32\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_19\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_31\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_37\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_9\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_11\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_21\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_28\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_45\u2019, \u2018https:\/\/dl.fbaipublicfiles.com\/rag\/rag_multiset_embeddings\/wiki_passages_2\u2019}`\r\n\r\nIs there anything else I need to set to download the dataset?\r\n\r\n**UPDATE**: just running `curr_dataset = load_dataset(\"wiki_dpr\", with_embeddings=False, with_index=False)` gives me the same error.\r\n\nClosing since this has been fixed by #1925","embeddings":[-0.3591079712,-0.3630983829,-0.0504424088,0.3148263693,0.3717034757,0.4270488024,0.3198949397,0.0625339374,0.2385216206,0.1804990321,0.1647016555,-0.0206091162,0.0954112709,-0.0210907198,-0.1673092544,-0.0334004611,0.1001311168,0.1070557833,0.3710588217,-0.0622109734,-0.329595685,0.2548207641,-0.3576647937,-0.231895268,-0.1536287069,-0.143520534,-0.0590852164,-0.0766759217,-0.4447588623,-0.4724355638,0.5257073641,0.2139309347,0.456645906,0.2094175071,-0.000123334,-0.0081851818,0.2570953667,-0.276065439,-0.2267439216,-0.1994098872,-0.5099616051,-0.2134321183,0.1228160188,-0.2729622424,0.2952129245,-0.2128572315,-0.0749510676,-0.0928959846,0.074530825,0.1510064006,0.1449033767,-0.0625398681,0.1508243978,-0.2355563939,0.0945451707,-0.1769310385,-0.1318905056,0.1710799336,-0.0272569992,0.1341178864,0.2691080272,0.0560868569,-0.2518814504,0.2277733535,0.1170533374,0.0164909512,0.2614754736,-0.304164052,0.4016195536,0.5526133776,0.923730731,0.1302185357,-0.2386934608,0.0382870622,0.211455673,-0.3757750988,0.1635375172,0.2258236408,-0.2530870736,0.200611487,-0.2965897322,-0.0683538169,-0.0212516002,0.3237015009,-0.0989618227,0.3529938459,0.2022208571,0.1463739276,0.1856656522,-0.2036290169,0.0743476376,-0.1115095615,-0.2810247838,0.0874456316,-0.2282613516,0.0073461859,-0.1403028518,0.0571529195,0.2315741628,0.3335160315,0.0137926796,0.1705392301,0.4314410686,0.1014831588,0.2653803229,0.2508428395,-0.1195526347,0.2461987734,0.3145948052,0.547288239,0.2549657226,-0.0376866497,-0.0095993467,-0.0076847337,-0.0618706718,-0.1101422459,-0.0867578164,-0.3650256991,-0.299870044,0.1207373217,-0.2721976042,-0.1857746094,-0.2455056757,0.3989276588,0.0512363166,0.4941689968,0.1345018893,0.1692253351,-0.0389331952,0.045549307,-0.096598126,-0.0991690308,-0.1295962483,0.0827541575,0.3832916021,-0.3321835995,0.3954805434,0.0359180942,0.0670721903,0.0898475721,-0.0672724321,-0.0200424772,0.2051387131,0.3828620315,0.0285898894,0.2937918305,0.107425645,-0.2261042297,-0.0246159267,0.2774973512,-0.1818651706,-0.4335758388,-0.0587910637,0.124901168,-0.2760751247,-0.0717422366,-0.3263556063,0.1071712226,0.0178034566,-0.1447193474,0.158366397,0.0138511891,-0.2004513741,-0.3192541003,0.3162532151,0.5299692154,-0.53788656,-0.0565230511,-0.0876784995,-0.2763260901,0.1355169863,0.0153356371,-0.2887880504,-0.0896552205,-0.3782499731,0.1199998185,0.1308039427,-0.4048190117,-0.5335974097,0.4260773361,0.0165747441,0.1425583661,0.3088404834,0.0687623024,0.2286393195,0.0907653645,-0.109613128,0.360917598,0.1335101128,-0.2713665366,-0.0036884013,-0.3318436444,0.3804715872,0.0646600276,-0.0475530401,-0.0033561576,0.118311286,0.1810515225,0.4451654851,0.0042786784,0.031223679,0.2706511021,0.1180464923,0.0901454762,0.1339274943,0.1229389459,-0.51377666,0.2290322632,0.1504512876,0.0457544737,-0.3320191801,-0.094936803,-0.3201955557,-0.0873651728,-0.3744758666,-0.2137416899,-0.0082012471,0.1704927832,-0.0325004943,-0.1602595001,-0.0984629914,0.0415371247,-0.05279148,0.0277336389,-0.1782583445,0.1469732225,-0.1374594718,0.014372942,0.1123541743,-0.0683239251,0.318995446,-0.0505682491,0.0679063499,0.3047980666,0.2043274045,-0.0166690685,-0.0066899275,-0.005671245,0.258731097,-0.5035472512,-0.0205891393,0.3803394437,0.1576838642,-0.0605140664,-0.2632324398,0.224481225,-0.0290870704,0.3685749769,-0.2061018944,0.0416202657,0.3276460767,-0.1760827154,0.2287088186,-0.3677020967,0.2900660038,0.5791848898,-0.2493056208,-0.1534954756,-0.0482501648,0.0391812027,0.1322937459,-0.2087436318,-0.0871027559,0.0236714669,0.0552035943,-0.0481585823,0.1448453516,0.0600990131,0.3681735992,0.247254774,0.0936580449,-0.1286984682,0.1636072844,-0.3218720853,0.2134936303,-0.0119457636,-0.1579790264,0.5060125589,-0.0821752176,-0.0530164614,-0.2347956598,-0.0330483504,0.2623149455,0.2831532955,-0.3046387136,-0.1779405475,-0.6171994209,0.0053771483,0.0949762836,-0.1705055088,-0.2037487328,-0.2974213064,0.1051949412,-0.074800849,0.124833256,0.1406850666,-0.2971715331,-0.1109372154,-0.1286422014,-0.202541694,-0.0407080688,-0.0536280163,-0.331849426,-0.0249450263,0.2382584959,0.3014809191,0.2516848743,0.058719039,-0.3458539844,-0.3976098895,-0.0931589454,0.0518293828,-0.1394761503,0.0614196137,0.336607933,0.481333077,-0.0528129339,-0.3433493674,0.0972987488,-0.0614315197,-0.0334287025,-0.0621375032,-0.0878678113,-0.0453506224,-0.1144122407,-0.6167302728,-0.1330221742,-0.3135217726,-0.0966890827,0.0850574449,-0.0683745295,-0.0727720186,0.1985067874,-0.1263793856,0.2616435885,-0.0850995928,-0.1912420541,-0.1607696265,0.5849156976,-0.1287938356,-0.3757933974,0.5222776532,0.0111422595,-0.0262258314,0.3314102292,-0.5823156238,0.1069680825,0.1025796235,0.2082475722,0.0025481533,0.2310561985,0.305845499,0.0286073294,-0.0586488731,-0.0875662118,-0.093996942,-0.0564806722,-0.2831987143,0.509218514,-0.0105568729,0.2368230969,-0.1851426214,0.6094356179,0.3000097275,0.1178231388,0.3601537943,-0.0355622433,0.1675811857,0.1629151553,-0.1403379142,0.0368618518,-0.1783125699,0.3249070346,0.2848487496,-0.161845535,-0.3321040869,-0.3920755088,0.1011412516,-0.0891899094,-0.0727448985,-0.0491276309,0.3427481949,0.1343055069,-0.1356298327,0.302932173,0.0821175426,-0.3334839344,0.2686665356,0.2534911036,-0.1429877877,0.1484562904,0.1589110345,-0.0379800834,-0.3448107541,0.3444143534,0.1638859808,0.3087622523,0.0045746486,0.1655189246,0.1895901561,-0.0242543779,0.5605655909,-0.4350599051,0.3597978055,0.1688041836,0.0517789498,-0.2673882544,-0.1922028959,-0.2611429691,-0.276125133,-0.1781174242,0.2215978503,-0.6027700901,0.0161780752,0.1669195592,0.1612285525,-0.1238549352,-0.359395355,-0.3704822659,-0.4479207993,-0.4778935909,-0.1964293718,0.0075105694,0.3619404733,-0.2136188,-0.0925567895,-0.0091365511,0.1171005815,-0.0015813193,-0.0716375336,0.243196398,0.4718736708,-0.0696359649,0.2932132781,0.1787507236,0.5283367634,0.6097086668,-0.0627239421,-0.1071096957,0.048447229,-0.2018221468,0.0192043819,0.2239023149,-0.0920598656,0.1804151833,0.1474184692,0.0573410019,-0.0472669639,-0.2881349623,0.3782993555,-0.2557798028,-0.3657977283,-0.4157612324,0.4662528336,0.0413501114,0.1092932597,0.1936930418,0.2587582171,-0.1668642461,0.0226928052,0.35775581,1.0175694227,0.1065527275,-0.0188474376,0.2060739398,0.0146241486,0.2137940377,-0.2915165722,0.2040702999,-0.0980691463,-0.4103773236,-0.2600039244,0.0024992549,0.1532847285,0.3295856416,-0.4880572259,0.3298540115,-0.2777177691,0.1029102132,-0.0117820604,0.2915962636,-0.2209283859,-0.1896228939,-0.2440835685,0.0250631962,0.106212154,0.1561797112,-0.1630843133,-0.0633788332,0.0575042218,-0.3713615537,-0.4828009009,0.0583305024,-0.2181004435,0.4266487062,-0.3948787451,-0.2592290044,0.2925424874,0.4851695001,0.1716979891,-0.0103366533,-0.3101628423,0.2248512655,-0.12282224,-0.4960898459,-0.2240558714,0.0299589187,0.0838920698,-0.0808789134,-0.5226157904,0.1020061448,-0.0371094123,-0.0976731181,0.06950441,-0.0406803861,-0.1487551332,-0.2177066952,-0.0691503957,-0.0016990738,0.1558312923,-0.0840194449,0.0362307951,-0.001989631,-0.1585756689,0.1211213395,0.1971923411,-0.1881342381,-0.1666731983,0.6800674796,0.0213053767,-0.2687130272,0.6212598085,0.2700417042,-0.3788090646,-0.2264257371,-0.0876422077,0.1271988004,-0.2000413984,0.0818250701,0.3339788318,0.1755504757,0.1250366569,0.4918984771,0.211436078,0.0397883654,-0.0355801135,-0.5814552307,-0.0940757543,0.0889578387,-0.0807976201,0.188050881,-0.4698204398,-0.3154388368,0.2788203657,0.0934570134,-0.1850284189,0.0440964662,-0.4358551204,-0.0807070807,0.1609925628,0.0234947689,0.1360045671,0.1081752554,-0.0183176082,0.051795017,-0.0362442806,-0.0480694063,-0.0843037143,0.2136483788,0.0365364328,-0.1229959428,0.0750056878,-0.4660222232,-0.3347616196,0.01751546,0.2713816166,0.3464633524,0.011237816,0.0264794286,0.3368587792,0.0637375265,-0.5087180138,0.0850754827,0.1238567755,0.3832836747,0.3522448242,-0.0603408739,0.0815015659,0.0552273132,-0.0290255304,0.0364017412,0.1451057494,-0.0185582973,0.4161353409,-0.2846483886,-0.0397414342,0.3598666489,0.080249615,0.4701498449,-0.1421664953,0.1650274247,0.1924473643,0.0728161857,-0.4741274416,-0.2790483236,0.3225473166,0.0880576894,-0.1475120634,0.0822515488,0.1976872981,-0.5292577744,-0.1985706389,0.0116609186,0.2377842665,-0.2428717166,0.1522594392,0.5658448935,-0.0603007786,0.026681494,-0.1705094576,0.1049014479,0.3029583395,-0.022607306,-0.3302282989,0.3238980174,-0.0517212488,0.2576176524,-0.2167513371,-0.5816491246,0.1976688504,0.2486557811,0.1091314703,-0.0623876937,0.1428005546,0.2813912332,0.1161995754,0.1848139763,-0.5019873381,0.1146367341,-0.0847099647,-0.0868122652,-0.1550206989,-0.1631114036,-0.3408384025,0.3697252572,0.2506076992,-0.3877323568,0.1391340941,0.2233997583,-0.3519901037,-0.3351708353,-0.0046411222,0.0475750901,0.1984971613,-0.2501712441,-0.0406993702,0.4264978766,-0.0355310589,0.5507053137,0.1158855483,0.2798188329,0.1993611753,-0.0190243442,0.0986388847,0.1593259424,0.0265295263,0.0264789164,0.3988074362,0.1845425963,0.1278179884,0.1725085676,-0.0023972353,-0.1505349725,-0.3022055626,0.1991573721,0.1582883149,0.1563813984,-0.0906068981,0.0002232369,0.1193928644,-0.1225259081,-0.0520095155,-0.2637984157,0.2559801638,0.2613166571,0.096773833,0.0997288898,0.1218362823,-0.0077642389,-0.2113354355,0.4893983305,0.3439127505,-0.0329992585,-0.3170140982,-0.165727362,-0.8635811806,0.1909140944,-0.0018567108,-0.1987917572,0.2767450213,0.1448274702,0.0629281104,0.1501017809,0.0517331958,0.0493551344,0.3558023274,0.5278563499,-0.0388547406,-0.3366440237,0.0577609614,0.0212414321,-0.2305419147,-0.2928699553,0.0717013255,-0.0456007048,-0.1349543482,-0.1366087794,-0.0527431183,-0.1673202962,0.2791327536,0.2438106388,0.0376392491,0.6168964505,-0.0421943404,0.0340050422,-0.2160889953,0.0339432918,-0.1781509966,0.3692152798,0.1981420517,0.040288534,-0.2612348199,0.1893491596,-0.2551579475,0.1336954534,0.1048166454,0.3364696503,-0.4344368279,-0.0075040176,-0.0616459772,0.0245793927,0.1752414554,-0.1018626615,-0.1526125222,-0.0320739858,-0.1232203245,-0.3546847999,0.6694049835,-0.445109129,-0.1035618335,-0.0715580881,0.1026608571,-0.2589513361,0.1466145515,-0.782879591,0.2994582653,0.2722882926,-0.0191888474,-0.5205340385,0.0897959992,-0.196808368,0.0039333892,-0.1593840271,0.1601531953,0.1927386224,-0.2259242535,-0.1817126572,-0.21335195]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1911","title":"Saving processed dataset running infinitely","comments":"@thomwolf @lhoestq can you guys please take a look and recommend some solution.","body":"I have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.","comment_length":13,"text":"Saving processed dataset running infinitely\nI have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.\n@thomwolf @lhoestq can you guys please take a look and recommend some solution.","embeddings":[-0.3462689817,0.1112205088,-0.1123075634,0.3321032226,0.2287198901,-0.0315706395,0.1957826763,0.2457539886,-0.207043916,-0.077350907,0.1137549207,0.1284730583,-0.1262958944,0.3329091072,-0.0082139438,0.1650309861,0.2835009098,0.1011736095,0.3245293796,0.1469843835,-0.0754511729,0.021536367,-0.1527943164,0.0416437909,-0.3447156847,-0.3817797303,0.1388849467,-0.2361930311,-0.315928489,-0.2921616137,-0.0329730362,0.3505786955,0.0886600763,0.5876352787,-0.0001302904,-0.3339567482,0.2800740302,-0.0642243922,-0.3001139164,0.1140513718,0.1857902557,-0.0684731528,-0.0688635111,-0.3427966237,0.2439277023,-0.0056540133,0.0650246367,-0.1403125077,0.3274538219,-0.0871174857,0.0368967652,0.04061893,-0.111470148,0.4226695597,0.0790355057,0.5443825126,-0.004128207,0.0382268466,0.3585308194,-0.118232213,-0.0285114907,0.2880208492,-0.4349061549,-0.1572247744,0.2131676823,-0.1154781282,-0.1539889574,-0.480148226,0.3200002909,0.1433316469,0.3409271538,-0.5267712474,-0.2359001338,-0.2272756696,-0.2031021863,-0.5277699828,0.3498126566,0.3446081579,-0.1571394205,0.1236172989,-0.4454064071,-0.1049970761,-0.2088370472,0.031187024,-0.1964515299,-0.0352307409,-0.2409812212,0.1953012347,0.0391045101,0.0138807856,0.2675268352,-0.419044435,0.1405345351,0.0688068643,-0.3199298084,-0.0528214984,-0.3904061615,0.3017805517,0.2126227617,0.0899465531,-0.29725945,0.1585399657,0.253484726,-0.0113677094,0.1558794975,-0.0100917043,0.1206870526,0.0639086142,0.2652877271,-0.3977831602,-0.2180280685,0.0884655491,0.0269364044,0.1522716731,0.366314292,-0.4956862032,0.0104759671,-0.1983377784,-0.06957075,0.326287359,-0.1176813319,-0.0555313304,0.0048630876,0.3380567729,-0.2640655339,-0.1682436019,-0.1559184343,-0.1021788269,-0.1503475457,-0.0472473651,0.059852086,-0.0424917303,-0.0017760688,0.1866793185,0.1193042174,-0.1290482283,-0.0688992143,0.0114264553,-0.2875577807,-0.05662705,0.0743240491,-0.6766244769,0.3467288911,0.3672560751,0.1442293823,0.4238489568,-0.2752934694,-0.0534883477,-0.056578692,0.3877505362,-0.000708464,-0.0911062956,0.0405864865,-0.0181568563,-0.1473418921,0.0696737096,-0.4181420207,0.0755584836,0.4679705501,-0.1213260293,-0.0761059448,-0.2114838511,-0.4208420217,-0.163152352,0.0943418741,0.2364164144,-0.5544015169,0.2326695472,-0.1332175136,0.0852983817,0.3265242875,0.421977967,0.0622518212,0.34236601,-0.2341058105,0.2924084961,0.3488395512,0.2054667026,-0.5819138288,0.145796597,0.0080100261,-0.1210385785,-0.134838298,0.2811975479,0.6140083671,-0.2130707949,0.1732332706,0.1245537847,0.0293324683,0.0238549262,-0.2897364497,-0.2323378026,0.0487736426,0.1644203365,-0.1300416887,-0.1006065235,-0.0017898318,-0.1818434596,0.290332526,0.097328864,0.3074536026,0.374578625,0.324895829,0.3565075696,0.0265387353,-0.3290907741,0.0621473305,-0.1142954901,0.1963784397,-0.2449512035,-0.3188972771,-0.3323393166,-0.0582028665,0.349424839,-0.1261119395,0.1193740442,-0.0852314904,0.1333748698,0.0051406454,0.278973341,-0.0863778889,0.3613173962,-0.2680849731,0.0128620276,-0.0708487704,0.096514523,0.1093686745,-0.2580110729,0.0112937456,0.0019685749,0.1490408033,-0.0817830637,0.025204381,0.2016450167,0.0562403388,0.0234272201,-0.1535659134,0.1709238887,0.0828458816,-0.1179192588,-0.0065380451,-0.173402667,0.312982738,-0.0407051966,-0.1425972432,0.1646902561,-0.0448422655,0.2705032825,0.0430182554,-0.2275513113,0.1965032667,0.037880078,0.0281038731,0.0705855116,0.2386953384,0.1462380737,0.250344485,0.1763491631,-0.1731234491,0.3850982785,0.6649882197,-0.2655313313,0.0198610462,0.2812173665,0.0132434238,-0.1775903255,-0.0834731609,0.0360750854,0.2095482349,0.1622571796,-0.0115479864,-0.111901097,0.1140808389,-0.1199284047,0.3320640326,0.1452700347,0.2954078913,0.2153036445,0.2438083142,-0.1247175261,-0.2861071825,-0.0786817819,0.30715096,0.3383691907,-0.1680182964,0.2085847408,-0.3333526552,-0.3527363837,0.0108462283,0.0311214793,-0.2404712737,-0.0932406411,-0.2269631773,0.3507361412,0.1099768952,0.1162968874,0.3041036129,0.1544539183,0.1149298623,-0.3052272797,-0.2681237757,-0.0610204376,-0.2256550789,-0.0912838653,0.4155008793,0.1512745619,0.602722764,0.0009671272,-0.2049450427,-0.5224997997,-0.3132186532,0.1833232343,0.1084619984,0.4322623014,-0.1099165455,0.3570844233,-0.2580294013,0.0854083523,-0.0577036738,-0.1052925512,-0.1852759421,-0.0851418972,0.0961970538,-0.1023807302,0.0426919833,0.0798613355,-0.4667095542,0.0363719016,0.3104667366,-0.4122330546,0.2908778787,-0.4167514145,-0.1292856336,-0.0196780171,-0.0533650257,-0.1564752012,-0.1622196287,-0.4193166792,0.3422170579,0.0910416692,-0.3131945431,0.1246146336,0.0880812854,-0.1545636952,0.3032865524,-0.5087324977,0.1958821863,-0.2251634002,0.1396950632,0.1194979548,0.0071949135,0.4061894119,-0.1032689363,0.1607913077,0.0510781333,-0.5065678358,-0.0250950418,-0.1966403723,0.1960191429,0.2598112524,0.7609847188,0.0487995781,0.5090444088,0.0426963568,0.0866486579,0.1990516633,-0.0951332375,0.0993893445,-0.1843753308,-0.3475402296,0.1350349635,-0.0583773665,-0.3017497063,0.1435715854,-0.1674912423,-0.0908360705,0.1540742517,-0.3456570208,-0.2929798067,-0.2391265482,-0.0114753796,0.4214158356,0.3753612936,0.3577443659,-0.0624562055,-0.0897326842,-0.3099236786,-0.2550873756,0.3171872199,0.2111603171,-0.0205908772,-0.3067134321,-0.080435954,-0.498071909,0.2298799008,0.0280217957,0.2766464949,-0.0287789311,0.098367922,0.3522484004,-0.3751344681,1.0997821093,0.0412317961,0.1122361049,0.0704323724,-0.2550372779,-0.2936953902,0.1199757829,0.1335675269,0.0856399015,0.3609296381,0.246976614,-0.282468915,-0.2438847721,0.112008296,0.3873696029,-0.2585331798,-0.2977850139,-0.1255546659,-0.3985819817,-0.32180655,0.2505300939,0.0092894668,0.1406272948,0.0759959221,-0.19362095,-0.2554508448,-0.1333056688,0.2487089932,-0.0520473383,0.1911732703,-0.2868556976,0.2934429646,-0.2749109864,0.1640561372,0.4442838728,-0.0910997614,0.0515956134,0.1343945414,0.3350647688,-0.1668715924,0.3163836598,0.3138032258,-0.1137191579,0.1268326491,-0.0096630203,0.3928796351,-0.3282928169,0.1034102291,-0.0016200795,0.3349971175,-0.6982312202,-0.1868077517,0.0866688266,0.2449615598,-0.1140252352,0.2108346075,-0.1462631673,-0.1523392648,0.3102091551,0.1795013398,1.2192177773,-0.3855558634,0.1935942322,0.08631479,-0.0402686074,0.0553885661,-0.5265199542,0.0338719822,-0.2266726047,-0.1286094487,-0.1037948132,-0.1317565441,0.0447866768,0.0826315954,-0.115995571,0.1744969338,0.0033348757,0.0921769813,0.0312409215,0.4691443741,-0.2218597531,-0.2493133843,-0.3254625797,-0.011740909,-0.0068113124,-0.0958136395,-0.0755479559,0.0079894224,0.1820365936,0.0063726483,-0.1322175711,-0.2159815431,-0.3851326406,-0.075620167,-0.3696689308,-0.0954224318,0.1888213903,-0.1044624299,0.1096026152,0.3891012669,0.0706137046,0.1522417068,-0.0959189683,0.1193147749,-0.0088725211,-0.390283227,0.0913992226,-0.1593626589,-0.0146056637,-0.0812668055,-0.0683836415,-0.25179407,-0.0574068986,-0.0056943377,0.5581324697,-0.2117469758,-0.1766024083,-0.2562556863,-0.0714484006,-0.1805589944,-0.0352133363,0.1604262441,0.0753412172,0.7354882956,-0.4020230174,-0.2554617226,-0.2805970013,0.3271191418,0.1778017581,-0.0108325928,0.3210655749,0.1576960832,-0.3095927238,-0.2292636484,0.2449497581,-0.1753445715,-0.1126234457,0.4065895081,-0.0329935625,0.0664948151,-0.0425621867,0.0190474242,-0.0726533905,-0.3024794161,-0.2139541656,-0.5561480522,-0.3190757036,-0.1081016809,0.2901566327,-0.0655131638,0.4148560464,0.2095452696,0.009966176,0.048701521,-0.1501694918,0.1182927787,-0.0673669577,0.2234449983,-0.0307379123,-0.0767033547,0.1211765409,-0.1331681162,-0.1081463844,0.4487351477,-0.0360665433,-0.0934509635,-0.068462193,0.2429277897,-0.1422200352,-0.0974869654,-0.1006681249,-0.4213645458,-0.1489633173,-0.1279077083,0.0121273408,0.1561644822,0.0416913256,-0.6981227994,0.2029936016,-0.0765425265,-0.17414397,0.0471718311,-0.0550015941,0.0291479677,-0.0267979894,0.2616246939,0.1442092806,-0.0581739284,-0.3075599372,-0.0345452987,0.2265169173,-0.2154630721,0.3061604798,-0.3790401518,-0.2054189742,0.2820777595,0.2120906264,0.5549833179,-0.0561194643,-0.3385288715,0.3481304348,0.0247800052,-0.1099910289,-0.041744113,0.1706786007,-0.1481862962,-0.2019422054,0.1447660476,0.0072219293,0.2070357203,-0.1836110204,0.1962678581,0.6882261634,-0.1799512208,-0.1950706691,0.2641071379,0.1905099154,0.2998276651,0.35625121,0.3778895736,0.0092532979,0.6284761429,0.1030139402,0.3522012532,0.3930136859,0.4763846695,-0.078013584,-0.3016082346,-0.0448353663,0.4150340557,-0.2448970079,0.1772336215,-0.063111499,0.0515901968,0.2122164071,-0.4803639054,-0.1790392846,0.6716665626,-0.1294308603,0.0703581423,-0.1476971507,-0.1093361005,0.3141917288,0.204258278,-0.0465743616,-0.081053175,-0.0472272821,0.2473636717,0.0290614571,-0.2316239029,-0.04608557,-0.0631528497,0.0438597798,-0.1713939309,0.0932157934,-0.1694827676,-0.0234029312,-0.3052830994,0.1204627976,0.4610807598,0.3933627605,0.110608831,0.1083964631,-0.1290009469,-0.0088128522,-0.161459133,0.3447709978,0.07641761,0.3605462313,0.1700503826,-0.0337123685,0.0277602542,-0.3234816492,0.218619734,0.2693846524,-0.0903745815,0.0407240242,0.1000638232,-0.0956286266,-0.2320312858,0.0207259301,-0.4128190875,0.3123778105,0.7299881577,-0.3081651628,0.2114313692,-0.0481600575,0.0234605428,-0.338817358,0.4558164179,0.1190049797,-0.0160978362,-0.2184236646,-0.2037891746,-0.5707247853,-0.1950216442,-0.0664853081,0.0400631279,0.0191994254,0.0641751289,0.1446711719,0.0083159851,-0.0882055461,-0.0540873222,-0.3324717879,0.3031989932,-0.1834214926,0.2354883403,-0.1537622362,0.2677846849,-0.1510539651,-0.4354768097,0.6668779254,-0.2291222364,-0.0730198324,-0.0860635042,0.1902728826,-0.0948882103,-0.0059107277,0.1987279803,0.2186683267,0.0210202243,0.0016057179,0.3654521704,0.2141619623,0.4159851968,-0.0123841474,0.2336122692,0.2500815094,0.2826732695,-0.166619271,0.1861836761,-0.0829034373,0.1299488991,-0.0635156333,0.0359909944,0.1797095984,-0.0437195115,0.1408159286,0.4216271937,0.084576048,0.3161844313,0.1325236261,0.2234113365,-0.6091109514,-0.132842198,0.4472812414,-0.5442948341,-0.4699992836,-0.1501910239,0.1825283468,0.497475028,-0.1947529316,-0.3893520534,-0.0757900029,0.3746799529,-0.0396187641,-0.3052856922,0.3742066026,-0.2455496341,0.1844296306,-0.096541658,0.318808496,0.0014437741,-0.3656584322,-0.0391951576,-0.1730715632]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1911","title":"Saving processed dataset running infinitely","comments":"am suspicious of this thing? what's the purpose of this? pickling and unplickling\r\n`self = pickle.loads(pickle.dumps(self))`\r\n\r\n```\r\n def save_to_disk(self, dataset_path: str, fs=None):\r\n \"\"\"\r\n Saves a dataset to a dataset directory, or in a filesystem using either :class:`datasets.filesystem.S3FileSystem` or any implementation of ``fsspec.spec.AbstractFileSystem``.\r\n\r\n Args:\r\n dataset_path (``str``): path (e.g. ``dataset\/train``) or remote uri (e.g. ``s3:\/\/my-bucket\/dataset\/train``) of the dataset directory where the dataset will be saved to\r\n fs (Optional[:class:`datasets.filesystem.S3FileSystem`,``fsspec.spec.AbstractFileSystem``], `optional`, defaults ``None``): instance of :class:`datasets.filesystem.S3FileSystem` or ``fsspec.spec.AbstractFileSystem`` used to download the files from remote filesystem.\r\n \"\"\"\r\n assert (\r\n not self.list_indexes()\r\n ), \"please remove all the indexes using `dataset.drop_index` before saving a dataset\"\r\n self = pickle.loads(pickle.dumps(self))\r\n ```","body":"I have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.","comment_length":103,"text":"Saving processed dataset running infinitely\nI have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.\nam suspicious of this thing? what's the purpose of this? pickling and unplickling\r\n`self = pickle.loads(pickle.dumps(self))`\r\n\r\n```\r\n def save_to_disk(self, dataset_path: str, fs=None):\r\n \"\"\"\r\n Saves a dataset to a dataset directory, or in a filesystem using either :class:`datasets.filesystem.S3FileSystem` or any implementation of ``fsspec.spec.AbstractFileSystem``.\r\n\r\n Args:\r\n dataset_path (``str``): path (e.g. ``dataset\/train``) or remote uri (e.g. ``s3:\/\/my-bucket\/dataset\/train``) of the dataset directory where the dataset will be saved to\r\n fs (Optional[:class:`datasets.filesystem.S3FileSystem`,``fsspec.spec.AbstractFileSystem``], `optional`, defaults ``None``): instance of :class:`datasets.filesystem.S3FileSystem` or ``fsspec.spec.AbstractFileSystem`` used to download the files from remote filesystem.\r\n \"\"\"\r\n assert (\r\n not self.list_indexes()\r\n ), \"please remove all the indexes using `dataset.drop_index` before saving a dataset\"\r\n self = pickle.loads(pickle.dumps(self))\r\n ```","embeddings":[-0.2176190168,0.1967649609,-0.1150818914,0.263695389,0.1485698819,-0.0731196851,0.1913734376,0.2333744466,-0.1932332069,-0.0838291422,0.0687725544,0.3565655351,-0.1087135077,0.2635340989,0.0180146117,0.2948513329,0.2866332531,0.1167820469,0.2218978852,0.0443398654,-0.1144503951,-0.039229136,-0.1053009704,0.0729433745,-0.2676576376,-0.2402478009,0.0895973071,-0.0787024647,-0.3543228209,-0.2827144265,-0.0402493291,0.2020813376,-0.020740699,0.5996127725,-0.0001320559,-0.3027279377,0.2690651119,-0.0898026377,-0.500980556,0.0918379053,0.1252096891,-0.0274518523,-0.0236187503,-0.4111998379,0.3176500797,0.0550618842,0.1007526144,-0.2602964044,0.2392185777,0.0291388854,0.0483903848,0.1627709866,-0.1344563365,0.4667539597,0.3288221657,0.5879324675,-0.0920103565,0.068196021,0.2598213553,-0.1367048174,-0.0111972671,0.213149175,-0.4668712914,-0.1813862175,0.3238168061,-0.0626365021,-0.1642286032,-0.485502094,0.1900697201,0.1010173261,0.3124753237,-0.5746484995,-0.3016631305,-0.2907087207,-0.3085360825,-0.4616621733,0.303160429,0.2828127742,-0.1148828343,0.0892646834,-0.3573212624,-0.0885956511,-0.0714923143,0.013924689,-0.1595258266,-0.0218514595,-0.2625794709,0.2816871703,0.1571230441,0.1349899769,0.2138314247,-0.5275251269,0.2432615906,0.1830554456,-0.2409775853,-0.0892678052,-0.128101632,0.2305015922,0.068454504,0.2216311097,-0.2834169865,0.0942212194,0.1737063825,0.0587529503,0.3496392965,-0.1559289396,0.1240393147,0.1688316613,0.1533984393,-0.3493698835,-0.1055040732,0.0989832655,0.1423719376,0.1292568445,0.2541599572,-0.4392640591,0.0998979211,-0.1352922022,-0.1503771544,0.293505311,-0.1735006124,-0.0440574475,-0.0193763841,0.2404387593,-0.189639315,-0.1738820374,-0.213664934,-0.1226866096,-0.282875061,0.1256302446,0.0212364867,-0.0136175109,-0.0764163733,0.189699471,-0.0462357253,-0.0983829647,-0.0973695442,-0.0382032655,-0.2951149642,0.0666370541,0.1504892856,-0.6224698424,0.2766810954,0.5634760857,0.0076077431,0.3507173359,-0.1286654323,-0.1917122602,-0.1637352556,0.3982414305,-0.0726503059,-0.1950869113,-0.0195657127,-0.0449417122,-0.2271591276,0.0960474312,-0.4880655706,0.0129513796,0.4187315106,-0.3198018074,0.113004379,-0.2247399688,-0.4087262154,-0.2898054123,0.0422482193,0.3129444122,-0.6525022388,0.1624652743,-0.2219654173,0.0037870104,0.3684933782,0.2907514572,-0.0184265543,0.3519964516,-0.2924845219,0.4722724259,0.4642398655,0.1513800025,-0.5824525952,0.2184977829,-0.0032660235,0.0871213526,-0.1679913849,0.2806184292,0.3480526209,-0.1936855912,0.0448203869,0.1299171746,0.0902841464,-0.0739290118,-0.2451011091,-0.1558860838,0.1333705038,0.1133300513,-0.1586099267,-0.0949898362,0.1699920446,-0.2811574638,0.3692251742,0.1862333864,0.2380425632,0.2262436748,0.4387608171,0.376616925,-0.0299835764,-0.2525387704,0.0410481878,-0.0314515792,0.2027600706,-0.2562841177,-0.3878393769,-0.3444662988,-0.0672249421,0.3747113049,-0.1951195449,0.0566943064,-0.1595136523,0.1894023567,0.0622430369,0.2978353202,-0.1763309389,0.1621255726,-0.428573966,-0.0729774162,-0.2045764029,0.1255356222,0.0948560983,-0.0802917704,-0.2075641155,-0.0644673705,0.111855194,-0.0987707898,0.0527718514,0.1788642108,0.3130970597,0.0153764123,-0.0880960226,0.2765067816,0.0521606021,-0.0698793381,-0.0566345677,-0.1350646168,0.4371303916,0.0080777425,-0.1692182124,0.1955073178,-0.0692287609,0.269998163,0.0169115625,-0.2877047956,0.1833025068,0.1248816699,-0.0894945711,0.1032610908,0.1317162216,0.2835094631,0.3211947978,0.2160081118,-0.1547553986,0.4590097368,0.6726176143,-0.3963784575,0.0546120107,0.3500018418,0.1655372381,-0.3123154938,-0.1193112507,0.1344985813,0.4746424258,0.0471061245,0.1163870096,-0.0722296908,0.0566727556,-0.1296166629,0.3597228825,0.1670878828,0.2787831724,0.266392678,0.2064008117,-0.0926338956,-0.3180205226,-0.2046259493,0.2067177743,0.3337043822,-0.2496601641,0.4450703561,-0.2661706805,-0.1963251382,-0.0248068664,-0.0397417173,-0.1011844352,-0.1752339751,-0.2606779933,0.2353005111,-0.0188816637,-0.001637742,0.2302643359,0.29997316,0.1067367792,-0.3236914575,-0.2955575287,-0.0433314107,-0.2829555273,-0.0933734104,0.3848826885,0.1502899081,0.4872848094,-0.0357298031,-0.2362879962,-0.6306734085,-0.3318595886,0.1534067988,-0.002999156,0.5537433028,0.0621285886,0.2313983291,-0.2958818972,0.1094467267,-0.0570730716,-0.0468581729,-0.2712249756,-0.1202104017,0.0726316348,-0.1153623015,0.0249984972,0.0948592946,-0.4184019864,0.0370903127,0.3668830693,-0.3680104911,0.3206292391,-0.2364815921,-0.1106603146,0.1673944741,-0.0467545763,-0.1276377589,-0.2547926009,-0.3890378773,0.2348028421,0.1060612202,-0.2417595685,-0.0012645205,0.0302166902,-0.2923387587,0.2514457405,-0.6670500636,0.1034713611,-0.2575093806,0.1251004934,0.0880426168,0.0954703465,0.3562363684,-0.1353866607,0.1612423211,-0.0044977837,-0.4499634802,-0.090728581,-0.0605865605,0.1561434269,0.2193526477,0.6643550992,0.0217189845,0.6045765877,0.0560661219,0.1380069256,0.2675991356,-0.0902759358,0.3034040928,-0.2566714883,-0.3080062866,0.1277370155,-0.1734944284,-0.5134416223,0.048415564,-0.2671698332,-0.0654825717,0.2267850935,-0.1906284243,-0.2014591992,-0.1861759573,0.0661621466,0.39812693,0.1973192692,0.260833174,-0.0373023897,-0.007418673,-0.1218075678,-0.27530393,0.3208586872,0.3104401231,-0.0027598219,-0.3376794457,-0.1947736889,-0.4489061832,0.167063728,0.0814194977,0.1557722986,-0.0812253803,-0.0599585287,0.3341170847,-0.3577575088,1.154493928,-0.0068498831,0.1128057614,0.2271645218,-0.1863550693,-0.4312819541,0.0836007223,0.0544862933,0.124398604,0.2215548158,0.2606627941,-0.2197369337,-0.1267499179,0.0469347611,0.3274624944,-0.1696225256,-0.1616916656,-0.2280384749,-0.5100156069,-0.2329775691,0.2316285074,0.0015790251,0.0847233012,0.0655161291,-0.0458717495,-0.2126406133,-0.0030375819,0.1310043186,-0.091773048,0.1487095207,-0.1269322634,0.1700147092,-0.3102150857,0.0544443615,0.3360923231,0.1814393103,0.0254547279,0.2124770582,0.4325071871,-0.1095378995,0.3074274957,0.4214728475,-0.0701554939,0.1852405667,-0.0510538332,0.3726031482,-0.3217012584,0.0800599232,0.1405324191,0.2599298656,-0.48649472,-0.355561316,0.2350625694,0.1999847144,-0.0872698277,0.262347579,-0.0226397365,-0.0963536948,0.2893201113,0.1828494519,1.3086409569,-0.1450115144,0.3485222459,0.354860574,-0.0636642426,0.0992142111,-0.4697084427,-0.0752163678,-0.1502972096,-0.0368536375,-0.1703922898,-0.1730159521,0.1589586437,0.0185335279,-0.0591552034,0.1629075259,-0.1580059975,-0.0016330195,0.0588591695,0.3480600417,-0.2200704068,-0.0782796293,-0.2369823605,-0.0147665106,0.107382603,-0.0454220921,-0.0758806244,0.0863569528,0.0776351169,-0.0572424568,-0.1449046731,-0.1484409571,-0.2265643626,0.1288760751,-0.2935394347,-0.1118460819,0.1564644724,-0.1301866919,-0.1242422089,0.2798473537,0.1314188987,0.1870657057,0.0642460585,0.222928986,0.0615288578,-0.493162334,0.2082460225,-0.0806246623,-0.0249039736,-0.0510763638,-0.1839712709,-0.3373163342,-0.0620780028,0.1841899604,0.5335732102,-0.2247230709,-0.34626019,-0.1785236746,-0.1496079862,-0.1413445473,-0.0690938756,0.1150557622,0.1077008024,0.6988791227,-0.5259568691,-0.3785372972,-0.2415050715,0.2715919018,0.0896575525,0.1711594462,0.3277649581,-0.0070922775,-0.3488295972,-0.1859962195,0.31301561,-0.1552764177,-0.0183439832,0.4162682295,-0.0939846486,0.1282294393,-0.1468785405,0.027375279,-0.0787057653,-0.1212723777,-0.3803351223,-0.5441784859,-0.1124778241,-0.132888183,0.3512420356,0.0402090289,0.3490844667,0.0829245448,0.0400354117,-0.0158003271,-0.1107846573,0.1925852299,-0.1334849447,0.233353585,0.0765910596,-0.0061084446,0.1610463113,-0.1598842591,-0.1560815275,0.3592751026,-0.162875399,-0.0988477618,-0.1282745898,0.2784002423,-0.0087861111,-0.1810813844,-0.0937768593,-0.3695252836,-0.1319839209,-0.1144260615,-0.0155978324,0.2512640953,0.0957467109,-0.5581236482,0.1915045232,-0.190706268,-0.1094852164,0.0942842215,-0.0859411657,-0.0040376186,0.0022507741,0.2778499722,0.1359617263,-0.0499164052,-0.2847098112,0.0633430406,0.2024172395,-0.1726548225,0.2768602669,-0.2688003778,-0.1116399914,0.0199811514,0.3108309209,0.6949669719,-0.1796551794,-0.4286446273,0.3348477185,-0.0296763033,-0.0848098248,-0.1932088882,0.0858063549,-0.0390490033,-0.1912904084,0.0999151617,-0.0937543213,0.2890966535,-0.3029480577,0.1010221764,0.5465027094,-0.069145754,-0.2038593143,0.3043861985,0.0957866982,0.3228174448,0.3069991767,0.2502877116,0.0553601161,0.6206972003,0.0220014248,0.3816455007,0.3313220739,0.3392212391,-0.1184374541,-0.3111564219,0.1606081277,0.4570246041,-0.4098156989,0.1371446848,0.0804123059,0.1331509799,0.0033823426,-0.3533461988,-0.0817739293,0.6286841035,-0.0968444869,-0.1043323502,-0.09129522,-0.1429573447,0.3363591731,0.1976307333,-0.0739195198,-0.1162398607,0.0362838991,0.21063447,0.0515612774,-0.1486438513,-0.0090335151,-0.1144569367,0.0577787235,-0.1716121882,-0.0355437621,-0.1341075599,0.100166969,-0.3189406097,0.0707945004,0.518648982,0.4143279195,0.096057266,0.088321276,-0.092451714,0.001873311,-0.1517194062,0.3243767917,0.0075166873,0.2404128611,0.2899964154,-0.0352593027,0.0400818884,-0.2816096246,0.2757959962,0.3581038117,-0.2363889515,0.1167822033,0.0016280126,0.0580269881,-0.172130093,0.2312397659,-0.505982399,0.2320760638,0.7975662947,-0.2769778967,0.2970381379,-0.0013093017,-0.010084454,-0.312954247,0.5281896591,0.0810494795,0.0797507167,-0.2492765337,-0.0456883572,-0.53944695,-0.1482831687,-0.0268201735,-0.0935609713,0.0284004174,0.1371170282,0.1103594452,0.0513768233,-0.0754792914,-0.0340376347,-0.4058742821,0.2675481141,-0.0699576214,0.1924789995,-0.2088578343,0.2203048021,-0.189762339,-0.5497687459,0.5792176723,-0.1915824711,-0.103992112,-0.0177278705,0.0784666464,0.0395194888,0.0994429514,0.3121793866,0.137277022,0.1301537305,-0.0450148918,0.2122042179,0.1963948309,0.3255818784,-0.0211298931,0.1904253513,0.0014327141,0.1711434871,-0.2003148645,0.3008941114,-0.1542052627,0.2673222125,0.0009623835,0.091515474,0.2174800336,-0.099898003,-0.0176703222,0.4172852933,0.0027415194,0.3088708222,0.1381883025,0.2813307643,-0.4911982119,-0.1034869626,0.5883191228,-0.5534768105,-0.3947317302,-0.1415627748,0.1200660691,0.4194647372,-0.1703362465,-0.4737051129,-0.2043703943,0.4016354978,-0.0639476255,-0.3335826993,0.3400070965,-0.3495184481,0.2096223831,-0.1204767972,0.5618261695,-0.1222997084,-0.3039699197,0.0734524876,-0.3393236995]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1911","title":"Saving processed dataset running infinitely","comments":"It's been 24 hours and sadly it's still running. With not a single byte written","body":"I have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.","comment_length":15,"text":"Saving processed dataset running infinitely\nI have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.\nIt's been 24 hours and sadly it's still running. With not a single byte written","embeddings":[-0.3297096789,0.1424022913,-0.12852633,0.2731113136,0.2379973084,-0.0472650416,0.2240568399,0.2563262284,-0.2344348729,-0.0836633518,0.186755985,0.0835028663,-0.1240532249,0.2571455538,0.0449179523,0.2680698037,0.2727478743,0.0804258883,0.3616465032,0.0918493494,-0.0587723665,0.0208334997,-0.0952349454,0.0293303151,-0.371545881,-0.358822763,0.1103805453,-0.1849191636,-0.3356473744,-0.26647982,-0.0018282232,0.3159895241,0.0290562492,0.5698403716,-0.0001298194,-0.2882886529,0.2986748815,-0.0548315197,-0.3562124968,0.1398755312,0.1537554711,-0.1009532884,-0.0912809446,-0.3484634161,0.2817374766,-0.0311013851,0.1400818527,-0.1649760455,0.3057920039,-0.0531787798,0.0425951928,0.0707928911,-0.1032269746,0.416564256,0.094712995,0.6043565869,-0.0011229983,0.057670176,0.3096903563,-0.1523675025,-0.060891699,0.2575120628,-0.4589864016,-0.1112056002,0.2213466614,-0.0761444494,-0.1363668144,-0.4309905171,0.3297851384,0.1296487898,0.3007091582,-0.5301759243,-0.2460941523,-0.2463755459,-0.2089576572,-0.5371947289,0.3537603617,0.3532669246,-0.1822617054,0.0968947485,-0.4256733656,-0.0868634209,-0.1570457816,0.0346310548,-0.2576577365,-0.0051727705,-0.2485288978,0.1814807653,0.013632427,0.0612363443,0.272939533,-0.4238608181,0.1223608702,0.0913740546,-0.2675905824,-0.0211146511,-0.2876037061,0.290705502,0.1783635765,0.0219791345,-0.2739258707,0.1134378091,0.2208641917,0.0050065308,0.1454209089,-0.0544353798,0.1411344111,0.0553645119,0.263581425,-0.3815296292,-0.1889055967,0.1006428078,0.0567780919,0.1594394445,0.3537156284,-0.4944783747,0.0918615311,-0.1643034369,-0.1034905389,0.3457174599,-0.1381813139,-0.0744404346,-0.0421606749,0.3073130846,-0.2305865884,-0.2064673454,-0.1699006855,-0.1252318174,-0.1804243028,-0.0315907188,0.0211319551,-0.0604348257,-0.0070722732,0.1579724103,0.1477620453,-0.073522754,-0.1069960222,0.0452449471,-0.2723116875,-0.0412787683,0.0789143518,-0.6873937249,0.4294087291,0.4291079044,0.0932732746,0.4574785829,-0.2269351929,-0.117323339,-0.044369176,0.4025470614,0.0589408353,-0.0726512522,0.040625371,-0.0160244424,-0.1812247187,0.0905254483,-0.3773212731,0.0633519813,0.5016575456,-0.1563331932,-0.033689104,-0.2192283422,-0.4052931666,-0.1665128022,0.1044852287,0.2676697969,-0.5897531509,0.2135879844,-0.1442498565,0.120100677,0.3167187572,0.403429091,0.0466590822,0.3543677628,-0.1974437833,0.3049887419,0.3142447174,0.1962144226,-0.5912645459,0.166687876,-0.0037220439,-0.0669125617,-0.1114930063,0.2831867933,0.6190411448,-0.2196117043,0.1917886287,0.0445450433,0.0231877621,-0.0156708192,-0.2888511419,-0.2308747768,0.0528307557,0.1410274804,-0.1071242467,-0.1180416867,0.0095251221,-0.1803518534,0.3274611533,0.1060855687,0.3396356404,0.3355779648,0.3286719918,0.3271795809,0.0031540019,-0.2983071208,0.0427794717,-0.1007520705,0.2340429723,-0.2410679758,-0.3500641882,-0.3467324674,-0.0035734181,0.344068557,-0.1176244617,0.0676883757,-0.0781916156,0.1602630764,-0.0492836386,0.3055345714,-0.1590861976,0.3407440186,-0.2511481345,-0.0206758752,-0.0753229335,0.0542030185,0.1494503766,-0.2818360627,-0.0169728696,0.0167249683,0.1326036751,-0.0632392466,0.0596703254,0.1890763789,-0.0020006117,0.0485614613,-0.1562067717,0.1631949246,0.0629353523,-0.1159440205,0.0053809695,-0.1784935445,0.31553334,-0.0240085311,-0.2224375308,0.1309906989,-0.0092730466,0.2512265444,0.0380896404,-0.2569566369,0.1794318259,0.0563939624,-0.0100783287,0.0535734445,0.2626291215,0.1169320792,0.2410032004,0.1757248789,-0.1307740808,0.396406889,0.7074303627,-0.2991807461,-0.0026568021,0.283349961,0.0449701287,-0.1802396625,-0.0742575973,0.0757772252,0.2231163681,0.1491415054,0.0229176488,-0.1573479623,0.0582925789,-0.1341576576,0.3119965792,0.1317681819,0.2735721469,0.1959267855,0.2171970308,-0.0959397182,-0.3046144843,-0.0858253911,0.2932009995,0.3251830339,-0.1888864487,0.2451591343,-0.3274155259,-0.3436899185,-0.0042954665,0.021107085,-0.2568872869,-0.1545768827,-0.2492871583,0.2779390216,0.1241913438,0.0420722067,0.2712286711,0.1545711458,0.0926618427,-0.2983962297,-0.2918512821,-0.0768423229,-0.211337924,-0.0767434835,0.4055637717,0.1525024921,0.5768886805,-0.0172312893,-0.2575844228,-0.502645731,-0.3270561993,0.1784654558,0.0736636072,0.4721550643,-0.0863901228,0.3955770135,-0.2926429808,0.0788385421,-0.0228141565,-0.1053583175,-0.1758050919,-0.1002444997,0.1233835444,-0.0750689134,-0.0014184749,0.14672032,-0.4228385091,0.0398340449,0.324372232,-0.406974107,0.2966657877,-0.390837729,-0.1489543617,-0.0052552945,-0.0364439487,-0.185508728,-0.1702359468,-0.4257723093,0.3393367827,0.1048429087,-0.279175669,0.1473000944,0.1208996251,-0.1657225937,0.3274570405,-0.5214036107,0.2171139717,-0.2352367789,0.1130636707,0.1526228935,0.0127324238,0.3990937471,-0.0850784704,0.1320053935,0.0092366599,-0.504245162,-0.0338894799,-0.18472673,0.189937517,0.1969634444,0.7578648925,0.0769206285,0.4938227236,0.0513332821,0.094557777,0.1742371619,-0.0950363874,0.1287570596,-0.1528608352,-0.3467243314,0.176655069,-0.0547736213,-0.3505257666,0.1174543723,-0.1528026909,-0.0835829303,0.1961493045,-0.2936436534,-0.3279715478,-0.259077251,0.0050923955,0.4180191159,0.3436193764,0.3429773152,-0.0366867408,-0.0733157471,-0.2665355802,-0.268435359,0.3439901769,0.2368249595,0.0013042204,-0.2327722907,-0.0547189973,-0.5444891453,0.2029835731,-0.0203806348,0.284240067,-0.0995345637,0.0514071174,0.3361436725,-0.4322921634,1.1148775816,0.0361098275,0.1349782646,0.1353654563,-0.195383966,-0.3135349751,0.1273670197,0.1354184598,0.1465166807,0.3530646563,0.1946006864,-0.2500933111,-0.2110372931,0.1420263499,0.3878874481,-0.2620953321,-0.2509109974,-0.1871654838,-0.4516673088,-0.2775382102,0.2018153518,0.0180741623,0.1237089932,0.0825124979,-0.168654263,-0.2898334265,-0.1649524719,0.2930144966,-0.080268234,0.1632921249,-0.2711595595,0.2283226848,-0.2229060829,0.1750877947,0.4321938157,-0.0832448974,0.0926654488,0.1671524644,0.387388438,-0.2093241513,0.2922061682,0.3781285286,-0.1068221033,0.1079042554,0.0740349665,0.3698235154,-0.3393636942,0.1017060801,0.0346850716,0.3320007026,-0.7182207704,-0.1857410669,0.1058411375,0.2541768253,-0.1101085097,0.2081141174,-0.1446826458,-0.1441427022,0.2766614258,0.2099832147,1.1998814344,-0.4041949511,0.2164201289,0.1340779662,0.0028647366,0.0602796264,-0.546420753,0.0563627966,-0.160033226,-0.1135050282,-0.1236360297,-0.1387907714,0.1093526632,0.0377539434,-0.0898051187,0.1329973936,0.0050503588,0.1404448301,0.0760040805,0.3954976201,-0.2565298378,-0.2116338164,-0.3240529299,-0.0107448995,0.0297870673,-0.078690581,-0.052178584,-0.0050329161,0.1249791831,-0.0203903019,-0.1340772212,-0.1996145844,-0.3611172736,-0.1090103984,-0.4145851731,-0.1076501757,0.1188901365,-0.1210897639,0.1450740397,0.3561914563,0.0926869959,0.1824738979,-0.0805006102,0.144442901,-0.0083581209,-0.4545828104,0.1005401909,-0.1361110657,-0.0278675109,-0.0501097627,-0.0978148133,-0.2377777845,-0.0219422076,0.0334922522,0.5325298905,-0.2047560364,-0.1433975846,-0.2515298426,-0.0314531885,-0.1572749168,-0.0334988907,0.1566637903,0.0951339006,0.720107913,-0.4890384376,-0.2588404417,-0.3007694483,0.3307547867,0.1926663667,0.0006721088,0.3300257623,0.1241748556,-0.2870666385,-0.228913933,0.2318741828,-0.1645646393,-0.1324971914,0.4048203826,-0.0564655215,0.0760584921,-0.0409008563,0.0363219827,-0.0987913162,-0.2716424763,-0.278151691,-0.5511589646,-0.3076512814,-0.0989568904,0.3239641786,-0.0677680001,0.4056753218,0.2368669808,0.0762931108,0.0314345509,-0.1551580131,0.0998657793,-0.0410087816,0.2210818231,-0.0131538874,-0.0714888647,0.1764442772,-0.1952680051,-0.0989018306,0.4231313765,-0.0336455293,-0.0981620252,-0.0841276497,0.249213025,-0.1457284093,-0.0681148916,-0.0758762583,-0.3890574574,-0.0932583138,-0.1382397264,0.0363238901,0.1418420076,0.0492627881,-0.7289839983,0.200726226,-0.0130151827,-0.155237779,-0.0300696827,-0.0566782579,0.0184770897,0.0029518572,0.2712486386,0.133833155,-0.016576549,-0.265917629,-0.0227518678,0.135442093,-0.2055039257,0.322016865,-0.4032432735,-0.2155200839,0.2433153391,0.2207262814,0.5596863031,-0.0931691378,-0.3822940886,0.2886338234,0.0205564462,-0.0721250847,-0.049068816,0.1694927365,-0.1168900505,-0.1901004016,0.1687067449,-0.0183814298,0.1590145677,-0.1936968118,0.1996811181,0.6564341187,-0.2235852629,-0.171803385,0.282579571,0.1846308857,0.320726186,0.3309204578,0.3395738602,0.0209501255,0.6970062852,0.1130838096,0.3697090149,0.4053201675,0.473338902,-0.1606832296,-0.3386050761,0.0238087624,0.4170040488,-0.2762649357,0.2156032324,-0.0329777263,0.0740699917,0.1509241462,-0.4475080669,-0.1848196387,0.6678310633,-0.1342783272,0.028733436,-0.1467298865,-0.135316819,0.3280400634,0.1869159788,-0.0531411953,-0.052870702,-0.0363359675,0.2322191298,0.0210284907,-0.2007909566,-0.0540457517,-0.13285622,0.0500720292,-0.157833755,0.1202641353,-0.1436671913,-0.041792322,-0.3405930698,0.1075992659,0.4964114428,0.3997760415,0.1354496479,0.1628125906,-0.1277668327,-0.0247839522,-0.2019581497,0.3321648836,0.0834558308,0.3931239843,0.1676693857,-0.0094369072,0.0331868082,-0.3098169267,0.2056339681,0.2801144421,-0.0650479123,0.0366547331,0.0418867543,-0.0802719146,-0.2159325778,0.0285560563,-0.4107843041,0.3032264411,0.6998465061,-0.2657906413,0.2379603386,-0.0351712108,0.0229508542,-0.3568628132,0.4547575712,0.0956432894,-0.0049805678,-0.2210353464,-0.1757514626,-0.5987207294,-0.2077641487,-0.1416371912,-0.0123060653,0.0006594149,0.1151490137,0.173068732,-0.0255805645,-0.087635003,-0.1008707434,-0.3014923632,0.2486161739,-0.1369232088,0.2424452752,-0.1698524058,0.2453712225,-0.1836574376,-0.4434060156,0.68035537,-0.263918817,-0.0542280413,-0.0838979259,0.134545356,-0.080400303,-0.0438418277,0.2369212657,0.2200989723,0.0042251102,-0.0271653514,0.2827035785,0.1414732337,0.4761386216,0.016711833,0.237207979,0.2433839142,0.323130101,-0.191240862,0.2551395893,-0.0845643133,0.1636112481,-0.0614707619,0.06267149,0.2137396783,-0.0361168869,0.1173311174,0.3821262419,0.0866436586,0.324508369,0.1322168261,0.2526768148,-0.6004786491,-0.1375020444,0.4529815912,-0.5878813267,-0.4281166196,-0.1689145863,0.1872006357,0.5068283677,-0.1822630763,-0.418751955,-0.0939290076,0.3750267625,-0.0451282561,-0.3550537229,0.3378444016,-0.2188799232,0.1696395725,-0.0916004628,0.3078427911,-0.0136940563,-0.4050228596,-0.016858995,-0.1534182131]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1911","title":"Saving processed dataset running infinitely","comments":"Tried finding the root cause but was unsuccessful.\r\nI am using lazy tokenization with `dataset.set_transform()`, it works like a charm with almost same performance as pre-compute.","body":"I have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.","comment_length":26,"text":"Saving processed dataset running infinitely\nI have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.\nTried finding the root cause but was unsuccessful.\r\nI am using lazy tokenization with `dataset.set_transform()`, it works like a charm with almost same performance as pre-compute.","embeddings":[-0.3561964035,0.1683889478,-0.0956779867,0.2705768049,0.2809066176,-0.055566512,0.2274563462,0.2229010761,-0.3108579814,-0.1395974904,0.1082137525,0.175986141,-0.1438532025,0.2344981879,0.0023731389,0.2562197745,0.3514589071,0.1079189405,0.370406419,0.1087035164,-0.0698908716,0.0259365421,-0.1417382807,0.0397690423,-0.3694447577,-0.3420152068,0.1376417428,-0.2580528855,-0.3062804937,-0.2669488788,-0.0369216688,0.3850825429,0.0711701214,0.5333968997,-0.0001309843,-0.3164223433,0.2313465625,-0.0396906398,-0.2950721979,0.147365585,0.1879237145,-0.0768317059,-0.0761477426,-0.3226693869,0.228089273,-0.0560969152,0.0666025877,-0.1789620221,0.3069119453,-0.099914059,0.024462603,0.0963327587,-0.2024442405,0.3944402933,0.0781609491,0.5052488446,-0.0299684722,-0.0143226553,0.2417495102,-0.1317525357,-0.0599417277,0.2605648637,-0.4225959182,-0.1757376939,0.2483236045,-0.0603317432,-0.1893056929,-0.4330962598,0.3536810875,0.1250943244,0.2908191085,-0.5002196431,-0.2704515457,-0.2737931609,-0.2251416445,-0.496601522,0.3358698487,0.2875787914,-0.0780523568,0.0703711137,-0.450178355,-0.0643272996,-0.1219843104,-0.0093971854,-0.2441366762,0.0015158034,-0.2214839607,0.1797708869,-0.0281713419,0.0159979779,0.2974232733,-0.4438245296,0.1197837591,0.0765056983,-0.2945407033,-0.0409833752,-0.3404235244,0.2206659764,0.1488776505,0.1337817609,-0.247090444,0.1759064794,0.2328191847,-0.0440793559,0.1266121417,0.0395753346,0.0933191627,0.137498796,0.2706351876,-0.4037394226,-0.1725402325,0.0876222551,0.0085124671,0.1798761189,0.4154669046,-0.5015193224,-0.0385920517,-0.1568151265,-0.1673543453,0.2999088466,-0.1609703898,-0.0266306419,-0.0335096046,0.3120589554,-0.255743742,-0.1600605845,-0.1566638499,-0.0816945583,-0.1733277142,-0.00999643,0.0484994277,-0.0495843478,-0.0274326168,0.2023169845,0.0920676515,-0.0578010231,-0.0601685122,0.0159271527,-0.2328036875,-0.0403911062,0.1218574494,-0.6531044841,0.3523284495,0.3579154611,0.0929572433,0.4303992093,-0.2356288135,-0.1576525122,-0.0987181365,0.3413205743,0.0289170183,-0.1152943,0.0900052935,-0.0103598442,-0.22735551,0.045630239,-0.3860265315,0.1007698476,0.4957956374,-0.1020928547,-0.0312961563,-0.2901702523,-0.4178996384,-0.1631322056,0.1399657577,0.2322128564,-0.5348827243,0.1970506907,-0.1168977544,0.1143586263,0.3382962346,0.4622505605,0.0029967006,0.391759485,-0.253549099,0.3449238539,0.3348969221,0.217957437,-0.5527255535,0.1113338172,0.0105000501,-0.0713317022,-0.1493799686,0.2181277275,0.6505599022,-0.2163791656,0.1967260391,0.144339487,0.0347693004,0.0030682553,-0.3002889454,-0.2102754861,0.0704109669,0.1161040068,-0.0734507367,-0.1079778746,-0.0262140706,-0.0873260722,0.2607068717,0.087332502,0.3065142632,0.3071137667,0.2978533506,0.3108500838,0.0385424942,-0.2755227685,0.0926343054,-0.0664934963,0.2263993472,-0.2249599546,-0.3080353141,-0.3852899075,-0.0595657602,0.331705451,-0.1536318213,0.0673741996,-0.079326421,0.1915822327,-0.0316917934,0.2812978625,-0.0916687921,0.3534356058,-0.2316103876,0.0081601106,-0.0361627936,0.0677646548,0.1614734083,-0.2569529712,0.0035996104,0.0137703959,0.1429991126,-0.0656035691,0.0499034375,0.170520857,0.0855399221,0.0511174388,-0.1799753755,0.2203267515,0.0789977387,-0.1262023151,0.0091753472,-0.129791826,0.2911994159,-0.0562337823,-0.137433663,0.209573105,-0.0264990907,0.2317998707,0.0407148078,-0.2505154312,0.2457159907,0.0342696197,-0.0115949623,0.0257075839,0.2258248925,0.186943233,0.2858631909,0.1908427179,-0.1367448419,0.4230031073,0.638071835,-0.2912369967,-0.0326803103,0.2630530298,0.0070483703,-0.2236629277,-0.0833157003,0.0391303375,0.2409757823,0.1381784528,0.0359574258,-0.0949811414,0.0806814805,-0.1128686145,0.3657848835,0.1423152834,0.3048703969,0.2377583236,0.2720299661,-0.0556193516,-0.3268786073,-0.1154103503,0.312830776,0.3043539524,-0.1911308318,0.249332279,-0.3187403083,-0.2655820251,-0.0087982062,0.0150732314,-0.2098757923,-0.1644063592,-0.2142040133,0.2902170122,0.0964989066,0.1142695695,0.2372078151,0.1645357758,0.0896114781,-0.3337947428,-0.2886410058,-0.0336719826,-0.2719524801,-0.1009411514,0.3449321985,0.1193480939,0.5602348447,-0.0377671458,-0.2010662854,-0.5184054375,-0.3823202848,0.1836897135,0.095488131,0.4375211,-0.0720780939,0.2666137516,-0.2688071728,0.0775148794,-0.0349436626,-0.1127791032,-0.1897428185,-0.0965391099,0.0818462297,-0.1162786931,0.0815770254,0.1189096645,-0.414113611,0.0816564634,0.2215959877,-0.4189993441,0.2398495376,-0.4049991965,-0.1667506844,-0.0364277437,-0.0301092751,-0.1684183627,-0.1815478951,-0.3694897294,0.404592365,0.1032464877,-0.2782887518,0.0854258388,0.0241050757,-0.1648419201,0.3543712199,-0.5036894679,0.2920398116,-0.2637021542,0.1855168194,0.1028890386,-0.0470127612,0.4037770331,-0.054596547,0.1448313147,0.034324158,-0.5332224965,-0.0501898043,-0.1899865866,0.1441877633,0.2329204381,0.7339385152,0.0965076461,0.5828547478,0.0113101359,0.0956182927,0.188200593,-0.0941795409,0.1587119699,-0.1883860677,-0.3327953815,0.0815476552,-0.0654175654,-0.3301579356,0.1679265648,-0.177642256,-0.1029427499,0.2416671664,-0.3118292987,-0.2376957983,-0.2351974845,0.0346930735,0.3311674893,0.3532544672,0.3883739412,0.0201953016,-0.0997952446,-0.292807132,-0.2255048752,0.2581325471,0.2102113217,-0.0397794731,-0.2918073237,-0.0236838218,-0.5580818057,0.2260975689,-0.0011218035,0.3068577945,0.0162966158,0.0465436466,0.3327513337,-0.3871513009,1.0841895342,0.0571326427,0.1120703071,0.1409702301,-0.3501552045,-0.3101097345,0.0681102723,0.0868621841,0.0650254413,0.4142943919,0.3152958751,-0.2865530252,-0.2286938429,0.0804544836,0.3835465014,-0.2443111837,-0.2377669811,-0.1378658563,-0.4442843795,-0.3062330484,0.2386859953,0.0170177203,0.1890956014,0.0787003785,-0.1602080017,-0.3057706654,-0.1577475667,0.3004525602,-0.0766151175,0.2009885758,-0.2700323164,0.3185217977,-0.2576541007,0.1249518767,0.3866300285,-0.1018247157,0.1028979048,0.0692249462,0.3182426691,-0.1466758549,0.3119281232,0.3617830575,-0.1181293055,0.1253963858,-0.0126623949,0.3950273693,-0.3305844367,0.1631429791,0.022236567,0.3152641356,-0.6721392274,-0.2041846812,0.0585968941,0.2648921609,-0.1384116262,0.1423313767,-0.1764369458,-0.2093856335,0.3884707391,0.2383251935,1.1850349903,-0.3527068794,0.2435462773,0.0962542072,-0.0012013066,0.1224101707,-0.5848091841,0.0443328135,-0.1987950206,-0.201774016,-0.1024803743,-0.1412135363,0.0470496193,0.0428310707,-0.0479663983,0.1785007566,-0.0131100183,0.1646735221,0.0457173958,0.4429638684,-0.1965352595,-0.2629642487,-0.2347229868,-0.029272683,0.0182908867,-0.0972593427,-0.053350471,0.0304032508,0.1121018752,0.0048080841,-0.1400815696,-0.2070683688,-0.2857974768,-0.0379699171,-0.3523786664,-0.1300926954,0.1904640496,-0.1281269193,0.156473577,0.3663034737,0.0848677084,0.174513638,-0.0607506931,0.1701619327,0.0446171463,-0.457350105,0.0846157968,-0.1292459369,0.0339114629,-0.1513441801,-0.0740293786,-0.2199123949,-0.1011068523,-0.0188080929,0.5681501031,-0.2446748018,-0.146063447,-0.2114471346,-0.006342967,-0.191219911,-0.0381627679,0.1370900571,0.080205217,0.7246432304,-0.4206167459,-0.2836773992,-0.2777165473,0.3438790739,0.1798158139,-0.0019146091,0.3709084094,0.0575133823,-0.3377861381,-0.1966435909,0.2740467191,-0.1098374054,-0.0943763107,0.3875785768,-0.0744715407,0.0293177292,-0.0829980895,0.0094571551,-0.0782549456,-0.2915322781,-0.2505192757,-0.5259460211,-0.282207638,-0.1161347032,0.3359451592,-0.0582361668,0.3865965605,0.2560594976,0.0006350586,0.0651101097,-0.1440390944,0.0952232182,-0.0528994836,0.2378822267,-0.1445232928,-0.0852377787,0.0821091682,-0.1076259688,-0.1013878584,0.4470528364,-0.0315738097,-0.0804555491,-0.1055288017,0.2624735832,-0.1276788414,-0.1190457344,-0.0676208586,-0.3544049859,-0.1504694819,-0.2015370876,0.0297063719,0.1752888709,0.0456521437,-0.7417253256,0.2148528397,-0.0661022887,-0.1919349432,-0.0027079061,-0.0215966497,0.0346857682,-0.0384617038,0.3330565691,0.1243117601,-0.0545377024,-0.2384004593,-0.0041587283,0.2349292636,-0.2059056014,0.3036861122,-0.3696604669,-0.2139859945,0.2145573199,0.2094155699,0.587788105,-0.0653170869,-0.4183367193,0.3807541132,0.005293095,-0.0946513861,-0.0843190029,0.1214046702,-0.1532374471,-0.204800725,0.1553560644,0.0319791362,0.1579072475,-0.1818763465,0.1740534753,0.7348417044,-0.205214113,-0.184009105,0.2037713528,0.1761213094,0.3332895339,0.3962142766,0.3413486183,0.023367241,0.6390288472,0.1276102066,0.366517216,0.4090359509,0.5006256104,-0.1466990113,-0.3013933003,-0.0524770692,0.4722627997,-0.2299551815,0.1820887178,-0.017243227,-0.0034151345,0.2032730579,-0.4290625155,-0.1983248889,0.7001779079,-0.1493081897,0.0222713966,-0.1310429275,-0.1185642183,0.2914765477,0.2307433933,-0.0646342188,-0.0569516011,-0.0748470202,0.2232402861,0.0209224876,-0.1831173897,-0.0129006645,-0.1347297877,0.0248769745,-0.1974101663,0.0989496261,-0.1873244792,-0.0104270512,-0.2813012004,0.1564614773,0.4901657999,0.4058198333,0.1036321223,0.1096377149,-0.1395994723,-0.0003490832,-0.1513182372,0.3752331138,0.0393170528,0.3430398703,0.13738662,-0.0394751206,0.016496107,-0.3794522285,0.2590906322,0.2656539679,-0.0957700387,0.1269501299,0.0949981213,-0.0701044351,-0.1572813541,0.067624405,-0.3923461437,0.3463858962,0.7331264019,-0.300365448,0.2092431039,-0.0336298682,0.0196808688,-0.3313454688,0.403632462,0.0891081095,-0.0353798158,-0.2302408367,-0.196440652,-0.5508621931,-0.1725220084,-0.0550139137,0.0583997406,0.0377011113,0.100885734,0.1468235403,0.0002133864,-0.1338802427,-0.0363393761,-0.3546831906,0.2725558281,-0.1193493083,0.2490170449,-0.2044496834,0.2637105584,-0.1720294058,-0.454308033,0.6403278112,-0.2540274858,-0.0691938475,-0.0592381097,0.1509843171,-0.0337429941,-0.0266966354,0.2175126076,0.2582066655,0.0115284547,-0.0121624516,0.3897282183,0.2276836634,0.4205097258,-0.0297248736,0.2628470659,0.185128361,0.2558895946,-0.1835886538,0.2455414087,-0.0848851427,0.1309694201,-0.0105230222,0.049428951,0.1391809583,0.0279731899,0.057698179,0.4323113859,0.0486310795,0.3516111076,0.1439342201,0.2393044084,-0.6242003441,-0.1817731708,0.4588896632,-0.6276181936,-0.5190659761,-0.1725154221,0.1477303952,0.5300548077,-0.1651887745,-0.3037884235,-0.0740636364,0.397131294,-0.056713853,-0.3319044411,0.3381716013,-0.2208518535,0.176168412,-0.1379645616,0.3688358963,-0.0206712224,-0.3688697219,-0.0293013081,-0.1754115969]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1911","title":"Saving processed dataset running infinitely","comments":"Hi ! This very probably comes from the hack you used.\r\n\r\nThe pickling line was added an a sanity check because save_to_disk uses the same assumptions as pickling for a dataset object. The main assumption is that memory mapped pyarrow tables must be reloadable from the disk. In your case it's not possible since you altered the pyarrow table.\r\nI would suggest you to rebuild a valid Dataset object from your new pyarrow table. To do so you must first save your new table to a file, and then make a new Dataset object from that arrow file.\r\n\r\nYou can save the raw arrow table (without all the `datasets.Datasets` metadata) by calling `map` with `cache_file_name=\"path\/to\/outut.arrow\"` and `function=None`. Having `function=None` makes the `map` write your dataset on disk with no data transformation.\r\n\r\nOnce you have your new arrow file, load it with `datasets.Dataset.from_file` to have a brand new Dataset object :)\r\n\r\nIn the future we'll have a better support for the fast filtering method from pyarrow so you don't have to do this very unpractical workaround. Since it breaks somes assumptions regarding the core behavior of Dataset objects, this is very discouraged.","body":"I have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.","comment_length":191,"text":"Saving processed dataset running infinitely\nI have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.\nHi ! This very probably comes from the hack you used.\r\n\r\nThe pickling line was added an a sanity check because save_to_disk uses the same assumptions as pickling for a dataset object. The main assumption is that memory mapped pyarrow tables must be reloadable from the disk. In your case it's not possible since you altered the pyarrow table.\r\nI would suggest you to rebuild a valid Dataset object from your new pyarrow table. To do so you must first save your new table to a file, and then make a new Dataset object from that arrow file.\r\n\r\nYou can save the raw arrow table (without all the `datasets.Datasets` metadata) by calling `map` with `cache_file_name=\"path\/to\/outut.arrow\"` and `function=None`. Having `function=None` makes the `map` write your dataset on disk with no data transformation.\r\n\r\nOnce you have your new arrow file, load it with `datasets.Dataset.from_file` to have a brand new Dataset object :)\r\n\r\nIn the future we'll have a better support for the fast filtering method from pyarrow so you don't have to do this very unpractical workaround. Since it breaks somes assumptions regarding the core behavior of Dataset objects, this is very discouraged.","embeddings":[-0.3835667074,0.2207914442,-0.0753484517,0.2524212301,0.1676334143,-0.0859722495,0.0995910391,0.2425324768,-0.1649823487,-0.0584910735,0.0330239199,0.4020829201,-0.0848712102,0.2319572419,0.0172859058,0.2511723638,0.190382123,0.0988868326,0.1988591403,0.1922226101,-0.0973070562,-0.0212861095,-0.1073065996,0.0238847807,-0.3752510846,-0.3421777189,0.1020547897,-0.176468417,-0.2869735658,-0.3915145099,0.025844872,0.2738236785,0.0088607669,0.4969805777,-0.0001254971,-0.2490330487,0.2621245682,-0.0777269378,-0.3850048184,0.1543710083,0.1895141304,-0.0919579417,-0.0632527322,-0.3167696893,0.2532064319,-0.0840291232,-0.0716398135,-0.148009181,0.3763958514,-0.0836270675,0.093504861,0.1139939278,0.000027077,0.3896321356,0.1677815616,0.5001136661,-0.0288456958,0.0879574567,0.3362618983,-0.0670923293,-0.1371211708,0.27834481,-0.4867144525,-0.1411877722,0.2645096481,-0.09666688,-0.1675312221,-0.3948975801,0.3739179671,0.2074767649,0.261888206,-0.5467400551,-0.2214086056,-0.3172277808,-0.170903787,-0.5744575858,0.3465484381,0.3022803962,-0.1129226759,0.0919190571,-0.3701660931,-0.175999552,-0.1269787252,0.025989458,-0.1026383266,-0.043541003,-0.2155543417,0.253354758,0.1353450119,-0.0528708026,0.2719113231,-0.3670327961,0.0432741791,0.1267949194,-0.348074615,-0.0000376811,-0.4093147516,0.2114434987,0.1972282976,0.1091903672,-0.17486009,0.1767202616,0.2158230692,-0.0245413166,0.1190609559,-0.0190227367,0.0418545082,0.1248172522,0.1343432516,-0.341491878,-0.1594240218,0.0596138239,0.0896649361,0.0389786735,0.342050761,-0.4655125737,0.0720364526,-0.1747670472,-0.0930221528,0.273494482,-0.151959613,-0.0404583961,-0.0540088713,0.304574728,-0.2286708951,-0.1745485961,-0.1305063218,-0.0846120268,-0.1934614629,-0.0743912011,0.0262205694,-0.0696877688,-0.1318622977,0.2878908515,0.1656465828,-0.096797511,-0.0424892455,-0.0020873728,-0.2887094617,-0.0223020725,0.024013916,-0.5841047764,0.408166796,0.4192660153,0.0451572165,0.3685524166,-0.2013343126,-0.1447459161,-0.095655553,0.4664761424,-0.0685854778,-0.1822187304,0.0261735991,0.0107502295,-0.1690286845,0.0607698634,-0.557675302,0.0929521918,0.4866020977,-0.0743257329,-0.0206744783,-0.2441198379,-0.3163408339,-0.2649134398,0.1202724874,0.255748421,-0.6958663464,0.2083716094,-0.2188268602,0.1379354745,0.3864471018,0.4180851877,-0.0557118058,0.2867590487,-0.209656924,0.4172917306,0.233522743,0.2801867723,-0.6572457552,0.0674076602,-0.0122792264,-0.0770693049,-0.2024056166,0.2173897028,0.5708351731,-0.1448692977,0.0347846188,0.1399073899,0.0250293668,0.0227942038,-0.3262345493,-0.2562339008,0.0538614988,0.114188686,-0.1550353765,-0.2059192061,0.0014636974,-0.266328305,0.3237319589,0.0752424151,0.3228730261,0.3021536767,0.2688813806,0.2784599364,0.0033835014,-0.3233175278,0.0001042504,0.0160897318,0.1061026752,-0.2561635077,-0.4138259888,-0.3335952759,-0.0771510154,0.3796461225,-0.1549665183,0.1034822166,-0.0588509478,0.0857046396,0.0420032218,0.1270328313,-0.0741478056,0.2337340117,-0.2826394737,-0.0001098073,-0.115852721,0.0829696134,0.0813676342,-0.3049838245,0.0064164116,-0.0618377775,0.1211540177,-0.0209339373,0.0009330697,0.2589646578,0.1091075316,0.0138087384,-0.1826982647,0.1420538872,0.0493600667,-0.2159519643,-0.0053009465,-0.2205975205,0.3107066154,-0.0643308535,-0.1287136227,0.2295851111,-0.0448109694,0.2971573472,-0.0409098938,-0.2173621655,0.1645674556,-0.0057499204,0.0359961577,0.0714560226,0.0766403005,0.1678945571,0.2786027491,0.2042236626,-0.1401461512,0.4174443185,0.6351094246,-0.2277691811,-0.0020448521,0.2752580047,-0.0467976145,-0.2161701173,-0.0883960351,0.0545103662,0.2103078514,0.1702259332,-0.0626562163,-0.125707835,0.0470386706,-0.1102022976,0.3756606877,0.1766787022,0.3735621274,0.1716313064,0.2496705353,-0.1166419983,-0.3490358591,-0.0377165973,0.2515732646,0.3713640869,-0.2111776173,0.2102584392,-0.3709145188,-0.2967136502,-0.024760358,-0.0927341506,-0.1991704106,-0.1360981762,-0.1633742005,0.2816187143,0.0414899699,0.213025853,0.2939043939,0.2184886485,0.0265323035,-0.2598508894,-0.3517384529,-0.1167461351,-0.3186908364,-0.0677342489,0.4512522519,0.1509422958,0.562354207,0.1044084132,-0.1226967275,-0.5189983845,-0.2806610167,0.1452239603,0.0051963651,0.407317102,-0.0914313048,0.3319068551,-0.3093451262,0.0313942134,-0.0224245191,-0.1245275512,-0.2509892881,-0.0310691986,0.0188751332,-0.0936461911,0.1211017221,-0.0205935799,-0.3942092955,0.0286190175,0.2887539864,-0.3527413011,0.3154911101,-0.2839599848,-0.0240322258,0.027449118,-0.0858804733,-0.193931967,-0.2213399559,-0.3033426106,0.3021124303,0.1005964056,-0.271892637,0.1194721982,0.0966610909,-0.022728093,0.2998028696,-0.4964134097,0.3107545674,-0.3027847111,0.1872274876,0.0311478153,-0.0686372668,0.4508425295,-0.0096569639,0.0962054878,0.0648788735,-0.4937604368,-0.069287926,-0.191551432,0.1666082889,0.1861975342,0.6179507971,0.0621536411,0.6712687612,0.0298438482,0.0587151684,0.1744910181,-0.0372415073,0.1861652434,-0.2480670959,-0.3244780004,0.0117498096,-0.1864834428,-0.2262557,0.0575556904,-0.1923340112,-0.1853723824,0.2674930692,-0.1907465309,-0.3303467035,-0.2938373387,0.0812425539,0.3420722783,0.3041879237,0.3475703001,-0.0553813986,-0.1528906524,-0.3027538657,-0.1691553295,0.1518634558,0.2205616683,-0.0215393789,-0.3734182715,-0.1219277754,-0.5277500153,0.2789167166,-0.0333701856,0.2458916008,0.0629912466,0.0062378356,0.3658824861,-0.4122897685,0.9922737479,0.0536895692,0.0677828714,0.1914999336,-0.1122041941,-0.3611142635,0.0798027068,0.1436760873,0.1254973114,0.2814244628,0.1921367496,-0.3499973416,-0.167574957,0.1012263522,0.4437808394,-0.2382509261,-0.1989262551,-0.1163623482,-0.4190782607,-0.3543668687,0.2736730576,0.0291247685,0.173857674,0.1009286568,-0.0984695181,-0.2437912077,-0.1644492298,0.1603194922,-0.0630564913,0.2521674335,-0.1471993327,0.3327230513,-0.2219741493,0.1688555628,0.4433543682,0.0932507515,0.0593430214,0.1484785974,0.26075086,-0.0668406636,0.2663196027,0.436547935,-0.0338201411,0.244259119,0.0333699435,0.3509745002,-0.2852980793,0.1120601371,-0.0206122119,0.241535008,-0.6335099339,-0.2218270302,0.1848451793,0.2303270549,-0.1434342116,0.3211612105,-0.0970837399,-0.1378961056,0.3642775118,0.3419322073,1.3057318926,-0.4013215899,0.1721456051,0.1258508116,-0.1251710802,0.0361407697,-0.3602975905,0.0009560375,-0.2630103528,-0.0570890345,-0.0198608693,-0.1382505298,0.0301313289,0.1163865328,-0.0877325609,0.0985184684,-0.0142498454,0.0629530326,0.0384843536,0.4255832136,-0.1980213225,-0.310516119,-0.1420005262,0.0185979661,-0.0941060036,-0.1785994172,-0.0325995758,0.0235860683,0.1079502776,-0.0005845182,-0.2096889764,-0.1743991375,-0.3854602575,-0.0478232615,-0.2607677579,-0.0625413656,0.124436222,-0.1577663571,-0.0354483984,0.3297653794,0.0586733297,0.2710851133,-0.0003176222,0.1032804847,0.0483013541,-0.4770299494,0.1158820242,-0.122364141,-0.0531713516,-0.0253540911,-0.1195457503,-0.3043126464,-0.157233119,0.0488527268,0.6313869357,-0.1920303553,-0.1679515094,-0.1920223534,-0.1115256846,-0.2200874388,-0.0146434773,0.1289254129,0.1512727737,0.7495426536,-0.3222368956,-0.2972276509,-0.2453289777,0.3345036805,0.2095808089,0.0449316204,0.3827110231,0.1830306351,-0.3001340032,-0.225105077,0.2039886117,-0.0447254591,-0.2874950767,0.354431659,-0.0558972731,0.1160110459,0.0527332574,0.0653968677,0.0482638851,-0.196893096,-0.2552427649,-0.4429489672,-0.3174460828,-0.0582753681,0.3190534413,-0.0096054533,0.3441122174,0.2328621894,-0.0540831611,0.1911137998,-0.1861876398,0.1707054377,-0.0204903577,0.2614901066,-0.1396443099,-0.0499599464,0.1205209047,-0.1675790697,-0.0861779377,0.4423304498,-0.0704083219,-0.1326019466,-0.0543186069,0.2066002339,-0.1244501173,-0.1713791043,-0.1712515354,-0.4104076922,-0.0668746158,-0.1763727069,0.0321454294,0.1435269415,-0.0098113352,-0.7529999614,0.2587635517,-0.1436858773,-0.107342191,0.0183955263,-0.1062027514,0.0172518175,0.0108023603,0.3042663038,0.1916284859,-0.106121175,-0.2100142837,-0.0201886389,0.1744356155,-0.1991694421,0.2817911506,-0.3842066824,-0.2681801021,0.1518539488,0.2423105836,0.6086739302,0.0183302537,-0.3965167403,0.4168857634,0.0804784819,-0.1517599523,-0.0787130371,0.1096322909,-0.1024193913,-0.1905668676,0.128428176,0.0097169783,0.2112805545,-0.2009210587,0.2207413167,0.6288778782,-0.2152168006,-0.2538819015,0.3068976104,0.1504750848,0.2075532824,0.4781116247,0.3338552713,0.1287807524,0.5598870516,0.0369232148,0.4385352135,0.4149489403,0.4798612595,-0.0740118697,-0.3737938404,-0.0722601861,0.4444386065,-0.1859976351,0.1629435569,0.028421307,0.1739250422,0.2753212154,-0.5637277365,-0.1189806461,0.6395537853,-0.1324681193,0.0351376012,-0.2842962146,-0.1248129606,0.2628364861,0.2157151252,-0.087598376,-0.1145733446,0.079730548,0.1870063543,-0.0750070587,-0.2530838251,-0.0606540069,-0.1116712317,0.050792031,-0.1934712678,0.0569865145,-0.1657603979,-0.0400586203,-0.2671598792,0.121903114,0.4641966522,0.3988071978,0.0146517614,0.1256377697,-0.07091178,0.0146945026,-0.1193478331,0.3011447489,0.1262501329,0.229254365,0.2591567338,0.0042548114,-0.0350130461,-0.3886093795,0.1917257756,0.3899106085,-0.0059602829,-0.0864131823,0.1085750759,-0.1012055725,-0.1906961054,0.0765016675,-0.3310644925,0.3115876317,0.7496406436,-0.3366106153,0.1730404943,-0.0520019419,0.041475866,-0.2468413264,0.5603487492,0.1954560131,-0.0417723805,-0.308290273,-0.196368441,-0.5810984969,-0.176803425,-0.1830585897,0.0588070974,0.0534620322,0.1389306337,0.1888881177,0.1059289351,-0.048495166,-0.0238035358,-0.3080116212,0.3683887422,-0.1822069436,0.2372138947,-0.0119722551,0.2303601205,-0.099985905,-0.4714621007,0.6159074306,-0.090680033,-0.051641617,-0.1053693891,0.074947089,0.039868813,0.1452372074,0.3274850249,0.2118887603,0.0348478667,0.0449153446,0.3216598928,0.0835499093,0.3057261109,-0.0545348339,0.2176975906,0.1881708801,0.302115202,-0.0890842006,0.2686496973,-0.1741846651,0.1284353435,-0.0778664351,0.1572755575,0.1311726272,-0.036222674,0.1362513453,0.4410628378,-0.0110285981,0.3235621452,0.1653807461,0.3450427353,-0.6625384092,-0.2042625099,0.5496140718,-0.4917227924,-0.440238148,-0.0959463716,0.2014236748,0.4864785075,-0.1716353297,-0.3719892204,-0.1293404102,0.3550194502,-0.0672078058,-0.3207606673,0.2895030081,-0.2783170044,0.2445469052,-0.1638460606,0.4990861416,-0.0096895546,-0.3690645397,-0.1437183022,-0.2491849661]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1911","title":"Saving processed dataset running infinitely","comments":"Thanks, @lhoestq for your response. Will try your solution and let you know.","body":"I have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.","comment_length":13,"text":"Saving processed dataset running infinitely\nI have a text dataset of size 220M.\r\n\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.\r\n\r\nfilter() function was way to slow, so I used a hack to use pyarrow filter table function, which is damm fast. Mentioned [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1796)\r\n\r\n```dataset._data = dataset._data.filter(...)```\r\nIt took 1 hr for the filter.\r\n\r\nThen i use `save_to_disk()` on processed dataset and it is running forever.\r\n\r\nI have been waiting since 8 hrs, it has not written a single byte. \r\n\r\nInfact it has actually read from disk more than 100GB, screenshot below shows the stats using `iotop`. \r\nSecond process is the one.\r\n\"Screenshot\r\n\r\n\r\nI am not able to figure out, whether this is some issue with dataset library or that it is due to my hack for filter() function.\nThanks, @lhoestq for your response. Will try your solution and let you know.","embeddings":[-0.3597593307,0.1054213569,-0.1111195758,0.3128801286,0.2211879939,-0.0300027505,0.2119499147,0.2385794073,-0.2019536942,-0.0772697777,0.1144533157,0.1193091869,-0.1160765961,0.3145315647,0.005771507,0.1776215136,0.2804865837,0.0960509032,0.3201403916,0.1582006812,-0.092254594,0.0329874828,-0.1535490602,0.03622647,-0.3603553474,-0.3808619082,0.1245507076,-0.2353426516,-0.3119152486,-0.2971180677,-0.0218299273,0.3504550755,0.089894779,0.5613131523,-0.0001294822,-0.3301271498,0.2810339332,-0.0546688549,-0.3094198108,0.1067515239,0.1822481304,-0.0783540532,-0.0676387846,-0.3384917974,0.2419880927,-0.0127866929,0.0628531575,-0.1438934654,0.3332122862,-0.0883401632,0.0393424146,0.0491744392,-0.1112133339,0.4080750346,0.0684425831,0.5446000099,0.0019138718,0.0293755382,0.3507192731,-0.1100794449,-0.0251225475,0.2801214755,-0.4348613024,-0.1405546367,0.2109708786,-0.1102473661,-0.1463636607,-0.4818505347,0.3428038955,0.1627472639,0.3462658823,-0.5211233497,-0.2293142229,-0.2183847725,-0.2037366182,-0.5381008387,0.3435384929,0.3489048183,-0.1451383084,0.1140599549,-0.4484113157,-0.0914645642,-0.194166705,0.0346008353,-0.2238059342,-0.0346903913,-0.2512477934,0.2032380998,0.0275383741,-0.0053895093,0.2505341768,-0.4423814118,0.1243684068,0.070415929,-0.3239110708,-0.0574067459,-0.3998301327,0.2915915847,0.2212855369,0.0737189725,-0.2985982895,0.1395414174,0.249293223,-0.0028316458,0.1449559182,-0.0084970351,0.1300870478,0.0686268136,0.2542479038,-0.4209403098,-0.220846653,0.0884132311,0.0246843658,0.1548254937,0.3649069369,-0.5077899694,0.0206093788,-0.1909214705,-0.0847468749,0.3307364881,-0.1145631447,-0.0589406975,-0.0124642365,0.320022732,-0.2498162687,-0.1739005297,-0.1675587893,-0.0970191509,-0.1356807947,-0.0414903872,0.0490906313,-0.0419146195,-0.0012089999,0.195576936,0.1264496148,-0.1172161847,-0.0645689815,0.0061924527,-0.2858996093,-0.0477497466,0.0858709589,-0.68455863,0.3599770963,0.3753085434,0.1470837742,0.4268293679,-0.2785609663,-0.0581173748,-0.0602076612,0.3959425986,0.0171609074,-0.0867892057,0.0561278798,-0.0124058416,-0.1486766934,0.0676337183,-0.4135324657,0.0693867132,0.4720985591,-0.1148443744,-0.0672661886,-0.2293657064,-0.4211976826,-0.1542957425,0.1005317941,0.2162790298,-0.5453290343,0.2333512008,-0.1347437501,0.0944514275,0.306932807,0.4288157523,0.0481356494,0.327460587,-0.229488492,0.3114904463,0.330576539,0.2128310502,-0.5719065666,0.148039192,0.0142431827,-0.1257253438,-0.1503320336,0.2757492065,0.6334018111,-0.2191340178,0.1805928648,0.1372405738,0.028149588,0.0076837358,-0.280597508,-0.2188080996,0.0488904156,0.1629419923,-0.1410835236,-0.1100685298,-0.0127399517,-0.170173794,0.2864417732,0.0931410491,0.3084415197,0.3696156144,0.3115650713,0.3505732417,0.0433638953,-0.3346327841,0.0684195533,-0.0916573405,0.2141959816,-0.2587138414,-0.3049888313,-0.330729872,-0.039176587,0.3567990661,-0.1211112663,0.1027016714,-0.0793895796,0.1322471052,-0.008127436,0.2799836993,-0.0922295228,0.3627993166,-0.2525633276,0.0172147192,-0.0625320598,0.0745420381,0.1065604389,-0.253690511,0.0202278811,-0.0062352316,0.1377633512,-0.080599159,0.0456250794,0.1982908845,0.0634310469,0.0245925244,-0.1546587348,0.1622640789,0.0796072781,-0.1302889884,0.0049024215,-0.1859749258,0.3041968644,-0.0543787964,-0.1387647092,0.1747485846,-0.0329565257,0.2667893171,0.0390749834,-0.2264470905,0.2026651949,0.0332531519,0.0180662628,0.0651547313,0.2343499362,0.1501993984,0.2606929839,0.1694227755,-0.1619215757,0.3740854263,0.6605485082,-0.292019397,0.0052237664,0.2767118812,-0.0019462616,-0.177304551,-0.0860488489,0.0368173197,0.2132088691,0.1610411555,-0.0079728374,-0.1166154221,0.1072064042,-0.1209591627,0.3384363949,0.1358503103,0.2971608937,0.2092590034,0.234596312,-0.1111435369,-0.2889982164,-0.0722805932,0.3130457997,0.3249441683,-0.1488227397,0.2251377702,-0.3158389926,-0.3490330577,0.0129083702,0.0285957344,-0.2482076138,-0.0868941918,-0.2378250808,0.3402898312,0.1219408661,0.1281036586,0.2950276434,0.1547338665,0.1116388068,-0.2981856167,-0.2660737932,-0.060786169,-0.2194406688,-0.0838351995,0.3953531682,0.1440834999,0.6054932475,0.0042056064,-0.1970841438,-0.5157345533,-0.3023740351,0.1900635213,0.1018257067,0.4048639834,-0.1112729385,0.3760762811,-0.2508971095,0.098465085,-0.0621488802,-0.1062086001,-0.1947844326,-0.0795624629,0.0965875685,-0.1159193963,0.0464825854,0.0994974822,-0.4645868838,0.0600151457,0.2981568575,-0.4114679396,0.2864318788,-0.4205574691,-0.1372179389,-0.0144094881,-0.0481236167,-0.1793636531,-0.1731070727,-0.4029321373,0.3528703749,0.0952376127,-0.2950040996,0.1345790476,0.0873484761,-0.1466000676,0.3153682649,-0.4979821742,0.215862602,-0.2244187742,0.1464547664,0.1230418384,-0.0042797658,0.4046213031,-0.1045506448,0.1495567262,0.0486008488,-0.5149697065,-0.0361069553,-0.1809903532,0.1851040274,0.2615905106,0.7603002787,0.0545843355,0.5108706951,0.0484450869,0.0874438882,0.1832389235,-0.0851930827,0.1021743268,-0.1946123987,-0.3352130651,0.1347627491,-0.068697311,-0.3061743677,0.1440570205,-0.1621428877,-0.1045364514,0.1677456051,-0.3383278251,-0.286082834,-0.2327581644,-0.0136865191,0.4081383944,0.3922974467,0.3660130799,-0.0641401559,-0.0828572735,-0.3070748448,-0.2686931789,0.3141343594,0.2079207599,-0.0101666041,-0.3038491905,-0.0787532404,-0.5130470991,0.2295326889,0.0053335372,0.2757646441,-0.0156171601,0.0962912515,0.3449009657,-0.3922867179,1.0903298855,0.0437595546,0.120795086,0.0787350163,-0.2569639981,-0.3013694584,0.1211455315,0.135469377,0.0844297931,0.3634844124,0.2498572767,-0.2752649486,-0.2396759689,0.1015236452,0.3932771683,-0.2509185076,-0.2989666462,-0.1258153915,-0.4069059491,-0.3230697215,0.2508509457,0.014534791,0.135403797,0.0657873303,-0.1860847026,-0.2580133379,-0.1397654563,0.2475002706,-0.0622195788,0.180452764,-0.2845987082,0.2873571813,-0.259349376,0.1520914137,0.4403588772,-0.1011734083,0.0490416028,0.141113013,0.3251040578,-0.171111688,0.2993180156,0.3193327188,-0.1190699041,0.1140351817,0.0046322406,0.3792167008,-0.3236221969,0.1105604395,0.0163424481,0.3397674859,-0.7052352428,-0.1935235858,0.0769362822,0.2443071455,-0.1151420176,0.2037349492,-0.1601251513,-0.1533433199,0.3071887791,0.1880859733,1.204892993,-0.3915423155,0.207319051,0.0724232495,-0.0384581909,0.050756257,-0.5071493387,0.0403578021,-0.2188887596,-0.1193417907,-0.1032298729,-0.1380858719,0.0555654131,0.0975881219,-0.1126570329,0.1554732472,0.0109391669,0.0925586373,0.0330126584,0.4777514935,-0.2122525275,-0.2529829443,-0.2980323136,-0.0029155842,-0.0069768382,-0.0919113308,-0.0653399229,0.0215941146,0.187270686,-0.0034946403,-0.1415566057,-0.2086235136,-0.3718006611,-0.0738620684,-0.3743020892,-0.0864778608,0.1757733226,-0.1031241566,0.1173218861,0.3839810491,0.076214619,0.1630580574,-0.0789019093,0.1163585559,-0.0148399323,-0.4047977626,0.0863532424,-0.1716443002,0.0012195447,-0.0781266317,-0.0746804029,-0.2364904135,-0.0564091392,-0.0041622897,0.5495881438,-0.2085425705,-0.1735353172,-0.2467993796,-0.0685876906,-0.1664274931,-0.0309171956,0.1622531265,0.0776151195,0.7389782667,-0.4247893989,-0.2539009452,-0.2794258595,0.3157607913,0.1950534731,-0.0094160615,0.3288805783,0.1557084024,-0.3067361712,-0.2323424667,0.2511332929,-0.1702623665,-0.1036404967,0.4087438881,-0.0224623121,0.0752266943,-0.0372504853,0.0237790495,-0.0706722736,-0.3030922711,-0.2074283063,-0.5520346165,-0.3196726441,-0.1276892573,0.2888016701,-0.0559750088,0.4196579158,0.2146143466,0.0144812791,0.0538745038,-0.1545840055,0.1229638457,-0.0635664612,0.2287177444,-0.0456675664,-0.0908407569,0.1228771582,-0.1230902299,-0.098207742,0.438775301,-0.0507366098,-0.0962698162,-0.0770148337,0.2390509844,-0.1372156441,-0.1001189351,-0.0949343219,-0.4149073958,-0.1360651851,-0.1321190894,0.018033376,0.1577410698,0.0367625244,-0.7231053114,0.2029173225,-0.0618304163,-0.1737445295,0.0285851657,-0.048642192,0.018210955,-0.0148764635,0.2758390903,0.149655968,-0.0528057404,-0.3137412667,-0.0312157478,0.228632316,-0.2301286459,0.3059849441,-0.3897563517,-0.2145376205,0.2620078623,0.2192458212,0.55558002,-0.0617609024,-0.3459027112,0.346146673,0.0323687419,-0.0986690745,-0.0478812568,0.1733779311,-0.1442540288,-0.2148090601,0.1338495165,0.0082891183,0.2133862227,-0.192592591,0.1912616491,0.6946160793,-0.200569585,-0.2033275664,0.2523305416,0.1941632926,0.3093204796,0.3646641374,0.3876368999,0.0254350901,0.634450376,0.1103037372,0.3517306447,0.4038519859,0.471650213,-0.0723033324,-0.3100889325,-0.0349691994,0.4051694572,-0.2589799166,0.17881082,-0.0670034215,0.0376467742,0.1973958015,-0.48692891,-0.1985945851,0.6722016335,-0.1384541243,0.0627903119,-0.1338501573,-0.1102734208,0.3184683621,0.2062447965,-0.0527345687,-0.0703593269,-0.0477138348,0.2357457131,0.0247941595,-0.2170370519,-0.0395752713,-0.0857587457,0.0106725954,-0.1700833738,0.097055167,-0.1765278876,-0.0313595012,-0.304114908,0.1100031361,0.4628582895,0.3941033483,0.1119000837,0.1116276532,-0.1353266239,-0.0094126919,-0.1543616652,0.3341702819,0.0809994787,0.3470247686,0.1831064373,-0.0294152796,0.027413521,-0.3326728642,0.2106412649,0.2673999071,-0.0808303952,0.0422161818,0.0948247537,-0.087428309,-0.2326333523,0.0361121632,-0.420027405,0.3090634048,0.7239662409,-0.3023775518,0.1989325583,-0.0445111282,0.026096208,-0.3554971516,0.4656997621,0.1101826131,-0.0180798899,-0.2189454138,-0.2099904865,-0.5598361492,-0.1950412393,-0.0926848799,0.0350554325,0.0204279609,0.0726561695,0.1589581519,0.0209808256,-0.0990496427,-0.0448680371,-0.3444212377,0.3077263236,-0.1664430201,0.2439672798,-0.1442698687,0.2751395106,-0.158106491,-0.4187309146,0.6580687165,-0.2280644476,-0.0681007951,-0.0908299461,0.1846681386,-0.0794261172,-0.0132404109,0.2019705176,0.2183785439,0.0144839259,0.004565062,0.3586002588,0.2293722183,0.4281020463,-0.0048968587,0.2238651812,0.2486083061,0.2848536968,-0.1669168621,0.1928020716,-0.070679687,0.1269425154,-0.0724253729,0.0266124774,0.1859895438,-0.0445758514,0.1274199933,0.4266431928,0.0901257545,0.3239371479,0.1339163184,0.2252250016,-0.610145092,-0.1288598329,0.4468807578,-0.5451056957,-0.4777399302,-0.1633597016,0.1802479178,0.4998636544,-0.1790253222,-0.3670514524,-0.0672344342,0.3778812885,-0.0381875671,-0.3214938939,0.3635118306,-0.2444079667,0.1782049239,-0.1048882604,0.3113435209,-0.0023716881,-0.3814167082,-0.0597062483,-0.1669084579]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1907","title":"DBPedia14 Dataset Checksum bug?","comments":"Hi ! :)\r\n\r\nThis looks like the same issue as https:\/\/github.com\/huggingface\/datasets\/issues\/1856 \r\nBasically google drive has quota issues that makes it inconvenient for downloading files.\r\n\r\nIf the quota of a file is exceeded, you have to wait 24h for the quota to reset (which is painful).\r\n\r\nThe error says that the checksum of the downloaded file doesn't match because google drive returns a text file with the \"Quota Exceeded\" error instead of the actual data file.","body":"Hi there!!!\r\n\r\nI've been using successfully the DBPedia dataset (https:\/\/huggingface.co\/datasets\/dbpedia_14) with my codebase in the last couple of weeks, but in the last couple of days now I get this error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/conditional_classification\/basic_pipeline.py\", line 178, in \r\n main()\r\n File \".\/conditional_classification\/basic_pipeline.py\", line 128, in main\r\n corpus.load_data(limit_train_examples_per_class=args.data_args.train_examples_per_class,\r\n File \"\/home\/fp\/dev\/conditional_classification\/conditional_classification\/datasets_base.py\", line 83, in load_data\r\n datasets = load_dataset(self.name, split=dataset_split)\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 609, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 526, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 586, in _download_and_prepare\r\n verify_checksums(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/uc?export=download&id=0Bz8a_Dbh9QhbQ2Vic1kxMmZZQ1k']\r\n```\r\n\r\nI've seen this has happened before in other datasets as reported in #537.\r\n\r\nI've tried clearing my cache and call again `load_dataset` but still is not working. My same codebase is successfully downloading and using other datasets (e.g. AGNews) without any problem, so I guess something has happened specifically to the DBPedia dataset in the last few days. \r\n\r\nCan you please check if there's a problem with the checksums? \r\n\r\nOr this is related to any other stuff? I've seen that the path in the cache for the dataset is `\/home\/fp\/.cache\/huggingface\/datasets\/d_bpedia14\/dbpedia_14\/2.0.0\/a70413e39e7a716afd0e90c9e53cb053691f56f9ef5fe317bd07f2c368e8e897...` and includes `d_bpedia14` instead maybe of `dbpedia_14`. Was this maybe a bug introduced recently?\r\n\r\nThanks!","comment_length":75,"text":"DBPedia14 Dataset Checksum bug?\nHi there!!!\r\n\r\nI've been using successfully the DBPedia dataset (https:\/\/huggingface.co\/datasets\/dbpedia_14) with my codebase in the last couple of weeks, but in the last couple of days now I get this error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/conditional_classification\/basic_pipeline.py\", line 178, in \r\n main()\r\n File \".\/conditional_classification\/basic_pipeline.py\", line 128, in main\r\n corpus.load_data(limit_train_examples_per_class=args.data_args.train_examples_per_class,\r\n File \"\/home\/fp\/dev\/conditional_classification\/conditional_classification\/datasets_base.py\", line 83, in load_data\r\n datasets = load_dataset(self.name, split=dataset_split)\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 609, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 526, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 586, in _download_and_prepare\r\n verify_checksums(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/uc?export=download&id=0Bz8a_Dbh9QhbQ2Vic1kxMmZZQ1k']\r\n```\r\n\r\nI've seen this has happened before in other datasets as reported in #537.\r\n\r\nI've tried clearing my cache and call again `load_dataset` but still is not working. My same codebase is successfully downloading and using other datasets (e.g. AGNews) without any problem, so I guess something has happened specifically to the DBPedia dataset in the last few days. \r\n\r\nCan you please check if there's a problem with the checksums? \r\n\r\nOr this is related to any other stuff? I've seen that the path in the cache for the dataset is `\/home\/fp\/.cache\/huggingface\/datasets\/d_bpedia14\/dbpedia_14\/2.0.0\/a70413e39e7a716afd0e90c9e53cb053691f56f9ef5fe317bd07f2c368e8e897...` and includes `d_bpedia14` instead maybe of `dbpedia_14`. Was this maybe a bug introduced recently?\r\n\r\nThanks!\nHi ! :)\r\n\r\nThis looks like the same issue as https:\/\/github.com\/huggingface\/datasets\/issues\/1856 \r\nBasically google drive has quota issues that makes it inconvenient for downloading files.\r\n\r\nIf the quota of a file is exceeded, you have to wait 24h for the quota to reset (which is painful).\r\n\r\nThe error says that the checksum of the downloaded file doesn't match because google drive returns a text file with the \"Quota Exceeded\" error instead of the actual data file.","embeddings":[-0.2483563125,0.35080567,-0.1227644235,0.2624529004,0.1046657935,-0.0048445282,0.3205699623,0.5194298029,-0.087741904,-0.0413587727,0.0776256993,-0.1982264817,0.0031295526,0.3474114835,-0.2444945425,-0.0099745383,0.098670274,-0.1543822438,-0.0066188341,0.0506980233,-0.176279366,0.0808001161,-0.277286768,-0.3061199486,-0.0021763062,0.2340680808,0.014913098,-0.1259736717,-0.001067427,-0.2701264322,0.5489282608,0.1795443147,0.0238150284,0.4143281579,-0.0001126272,-0.0021212655,0.3832170963,-0.0588617101,-0.328229636,0.105318971,-0.3156419694,-0.1499086767,-0.0790167302,-0.0825071856,-0.1530301273,-0.5050341487,-0.2010857463,-0.1114249155,-0.104029946,0.261859864,0.2175973356,0.2089748532,0.0917270258,-0.0789880306,-0.1643640399,0.0381903574,-0.0346372761,0.0000728814,0.1079570428,-0.1750721484,0.0917512923,0.1509220302,-0.1619757265,0.0288006309,0.1156822592,-0.2075259537,0.0970106125,-0.2640064955,0.2317044139,0.240515843,0.2926249504,-0.128769815,-0.3720597327,-0.2977039814,-0.0750023425,-0.245641768,0.2302350849,0.1716050953,-0.067303285,-0.0363245681,-0.3061080277,0.252392441,0.1045450494,0.0709883124,-0.1234941557,0.1498817354,0.0577276759,-0.0749311298,0.1063356474,-0.0413521416,0.6009082198,-0.1663401127,-0.2048625946,0.0807095915,-0.4355798364,-0.1634497792,0.0560006686,0.6840990782,0.1832090616,0.4219745696,-0.1995084137,0.1657820642,-0.1209183633,-0.0922681913,0.1504204273,-0.0343498252,-0.0519279428,0.4296023548,0.2829835415,0.3001469374,0.1284109205,0.1946967095,0.1358951181,-0.2661388516,0.2309339345,0.0963650122,0.0433475263,-0.2907250226,-0.2783105969,0.276026845,-0.1270309091,-0.1380530298,0.2978390455,0.2877974212,-0.4330213964,0.1959685087,0.0082532791,0.0210149046,0.0442288406,-0.2511606216,-0.1368148774,-0.1471276432,-0.0469176024,0.137075603,0.1180858687,0.031462837,0.2896967828,-0.0035000346,0.1563295871,-0.1104827821,0.2000196576,-0.3637543023,0.2052349597,0.5173292756,-0.0956809148,0.2081980556,0.089403756,-0.3298082948,-0.0150805023,-0.1024182662,-0.2392274141,-0.1406963617,-0.3736905754,0.2286854237,-0.3321997821,0.0105260788,-0.5877588987,-0.2102768719,0.3507718146,-0.2136421949,0.0752197951,-0.3401685059,-0.2169393599,-0.1177302599,0.3215620518,0.3882475495,-0.2872908115,0.1926524937,-0.5058418512,-0.0116746584,0.1311635673,0.0566376485,-0.3111607134,0.3517674804,-0.3844766021,-0.2458629757,0.0601757877,-0.5608974099,-0.3146588802,0.1328949034,0.1920131296,0.3282431662,-0.1377325356,-0.2058589309,0.0681966692,-0.2593610585,0.1219714582,0.0948211327,0.0682110935,0.1584233344,-0.2493803948,-0.0130494172,0.2194643766,0.1278644204,0.1831167638,0.00329765,0.0568596125,0.1512579173,0.2725539207,-0.1036928073,-0.0148238773,0.1306653172,0.2006486058,0.0827575177,0.0509694889,-0.0756189004,-0.3890631199,0.1855844557,-0.2650212049,0.0332782269,-0.1553544253,-0.0455868579,-0.19924815,-0.1255375147,-0.4297518432,-0.3456073105,0.2111902684,0.2249656916,-0.1547851264,0.1536783427,0.046917703,0.1888696998,-0.3862114549,-0.0471783876,-0.0983198136,0.3950824738,-0.0024916609,0.0562320389,0.0013942948,0.0921594203,0.1276094764,-0.0379030854,-0.0082374858,0.4010099769,0.3606117666,0.038932316,0.190829441,0.3519996405,0.256611973,-0.2802025974,-0.2643060088,0.3049584329,0.1222953796,-0.0713119507,0.1305744052,0.3840272129,-0.2733516693,0.0554502793,0.126211822,0.0377751514,0.4955723882,-0.1079363376,-0.2369183451,-0.2576626241,0.176587835,0.1049608141,0.0675177202,0.1393899769,-0.0053886087,0.0798845366,0.2055527419,-0.080689393,-0.1177334562,-0.042633567,0.121971339,-0.035609182,0.0879070237,0.7510873079,0.4312459826,-0.0813290998,0.0834221691,0.1749548316,-0.0722506121,-0.0217371378,0.0707214326,0.078366138,0.1058773994,0.3351549804,0.058534801,0.0462715179,-0.1494187713,-0.2088983655,0.173096478,0.3105882406,-0.3180760741,0.0239008889,-0.3514176905,-0.2420360297,-0.3531622291,-0.1688136905,-0.2286220491,-0.438159287,-0.0157749671,0.3220815957,0.0368187949,0.2074562609,-0.4492830634,0.0925884247,0.1287255585,-0.2775695026,0.0639834106,0.1320360452,-0.0243535209,0.0451560877,0.57372576,0.2313758731,0.2702275813,-0.3618479073,-0.1297322214,-0.523191154,-0.3683932126,0.0594919212,0.0910136774,0.3205645382,0.1584272683,0.1651458293,0.1059408858,-0.3236399591,0.0281092543,0.317777127,-0.0200341064,-0.0149845202,-0.0636154339,0.0657196939,0.0275884718,0.0795964599,-0.2562216222,-0.0792468712,-0.2330019623,0.1119934544,0.0559557192,0.0892030075,-0.0778987408,0.1029612124,-0.0143481353,0.0297972932,-0.1901176423,-0.376108855,0.1693154871,-0.09028887,-0.441747874,-0.0958381593,-0.0300432798,0.371925056,-0.0784174949,-0.3715616465,0.006258986,-0.1335083842,0.1186246425,0.2444375604,-0.3846087158,0.2823690474,0.0073235747,0.016855523,-0.1849660724,0.0201206245,-0.0567025393,-0.2161255777,0.5088825226,-0.1245597601,-0.1940550208,0.0573147088,0.3122811317,0.4735923111,-0.1680437326,0.2759514749,0.0246083308,0.4678413868,-0.1390938908,-0.4603390992,-0.1930557489,-0.095010139,-0.0335226022,-0.0393403471,-0.247411266,-0.1724460572,-0.1374422163,0.0313692875,-0.3403830826,-0.2995963395,-0.1047632396,-0.3917798996,0.2362755239,0.0033101952,0.0916021243,0.1571807712,-0.3026350141,0.1535651088,0.3441468775,-0.1502449811,-0.1311242729,-0.4588471949,-0.0238263458,0.2610839307,0.168688789,0.0576811805,0.5540510416,-0.0254554302,0.0201969072,-0.227447778,-0.033008229,0.3528869152,-0.5139679313,0.3113613129,-0.0135799488,0.0258567762,-0.0678247586,-0.2351488024,0.0483183749,-0.1841357201,0.1932884157,0.5447885394,-0.0482577458,-0.1264222264,0.2902929783,0.3781235218,-0.1861049831,-0.0063399635,-0.2953607142,-0.1861956418,-0.0172106177,0.097582601,0.1861983538,0.2329001129,-0.3318367302,-0.1802390516,-0.0489787683,0.0665417835,0.0224901978,0.0602437556,0.2055643797,0.0275994856,0.3083541393,0.1284879595,-0.138871789,0.7748520374,0.2862086892,-0.2472219318,-0.2631367445,0.0567611381,-0.1023992449,-0.0323166214,0.0993448496,-0.1670053154,-0.0142985918,0.3959570527,-0.092776753,-0.1880429387,0.0810126886,0.2037273943,-0.0244847592,-0.4803299606,-0.5200728178,-0.1251573712,0.1447148174,0.011095834,0.4486398101,0.0692506507,-0.2791030407,0.3141688406,0.0963624492,0.4851434231,0.1838276982,-0.0944215655,-0.0706914365,-0.0477726683,-0.0933511555,-0.1017453745,-0.0060383556,-0.3348731399,-0.6285871863,-0.1254462004,-0.1433281898,0.2475311011,-0.2586276829,0.113837935,0.3659522533,-0.0879793391,0.0187855698,0.0249823723,0.0654744133,-0.0623459704,-0.0027059449,-0.3052218258,0.1355485916,0.037938185,0.1731657088,-0.1275607347,0.0060586161,-0.1111586466,-0.1593545675,-0.5275295377,0.0832000524,0.0327207483,0.0402562805,0.2117589414,0.0777599216,-0.2082785666,0.5434353948,0.2335897386,0.3035466969,-0.1847197562,0.1648023874,0.0613129213,0.2169542164,0.0758589804,-0.2602525651,0.5399428606,-0.1387327313,-0.2067426741,-0.0027815464,-0.1383007765,-0.3202257752,0.1114301831,0.0230444074,-0.1734686941,-0.1759435982,-0.0219599716,-0.1599989086,0.1275666356,-0.2879592776,0.1467720866,0.0805835724,-0.0266166534,0.1897024065,0.1364036947,-0.3712579608,-0.0933347121,0.6231891513,-0.0370385312,-0.1861687601,0.3555686772,-0.0803420693,-0.1102819517,-0.2094681263,0.1944050938,0.2133210897,-0.3284565806,-0.1627129614,-0.4114899039,0.2058758438,0.0871242508,0.4304659367,0.2160063684,0.0801099166,0.1100860611,-0.7900116444,-0.03053152,0.098693423,-0.1488034278,0.1077324301,-0.1543703228,0.1304602921,0.0958669558,-0.2670515478,-0.2456792891,0.0759127736,-0.4292776287,0.0968834236,-0.3269150257,0.014633107,0.2100938112,-0.09852539,0.1508654654,0.015134925,-0.0596978031,-0.1588754803,-0.2603534162,0.1143746451,0.205262512,-0.1401935965,0.1857998222,-0.1429645121,-0.0799993202,-0.1082240865,-0.0873486027,0.0758004487,0.0700319111,0.1423600465,0.3404824436,0.0620095208,-0.0689224601,-0.2004760355,-0.0650647581,-0.0370064452,0.2050034702,0.1779527068,-0.0698505938,-0.025923213,0.0382795744,0.0719280839,-0.2412770391,0.1140172109,0.3307912052,-0.2690100074,0.0512698814,0.4562601149,0.5259580612,0.4900167584,-0.4191825986,-0.1510872245,0.3108334541,0.1734564006,-0.3344335556,-0.1327138394,0.3407932818,0.1940482706,-0.0793226883,0.1123545542,0.3362826407,-0.3540165126,0.3828054965,0.1053666994,0.1520605385,-0.5242892504,0.5086352229,0.4963040054,-0.3074856997,0.1020008773,0.2030338049,-0.189843446,-0.1762794107,0.5503347516,-0.0935347229,0.3178268671,-0.1157834008,0.007693103,-0.0532061085,-0.5167142749,0.4916111529,0.1496077925,-0.0728207901,-0.0603758581,0.0231312476,0.0525243022,-0.105610095,0.0569683053,-0.3598283529,-0.0102919452,-0.0256157313,0.0309297163,0.1982888281,-0.3819050193,-0.1348468065,0.1083562896,0.1899412423,0.0246690884,0.1944099069,0.0915195569,-0.5324066281,-0.4129911363,0.1037879735,-0.0287562143,0.0480688289,-0.1690666676,0.1855584383,0.2086615562,-0.0036593641,-0.0355615094,0.1527326405,0.242548719,0.4049273729,0.2420446873,-0.1356989145,-0.0280001089,0.0791085958,0.0403711349,0.479392767,0.2239826918,0.1611828506,0.3548166156,0.157780692,-0.1032981202,0.1443248987,0.03648993,0.2156119198,-0.2185446173,0.5506973863,0.0730593503,0.1857665777,-0.2539486587,0.1162541807,-0.0487986058,-0.0968164355,0.3136620224,0.2028799057,0.1977779418,-0.0651572123,0.0681772083,0.1485725641,0.4755242765,0.2595829368,-0.1090379804,-0.1856627613,-0.3761809766,-0.7282562256,0.1537902504,-0.055522956,0.5085514784,0.0304756537,0.1653366685,-0.1357877702,0.1360773742,0.3877424896,-0.4676198065,0.2673336267,0.070268482,-0.1864013374,-0.0820699036,-0.2114674598,0.074621737,-0.0708472729,-0.2945510447,0.2579188645,-0.0986893773,0.1454250664,0.177061975,-0.5136868954,0.0208940338,0.0947820619,0.2509298325,0.069316864,0.7262417674,0.0100471638,-0.103388235,-0.4000116587,-0.0479570143,0.0073248157,0.5848952532,0.0964269564,0.0247241743,0.1275389344,-0.0883762613,-0.3506602347,0.4802132249,0.0932057947,-0.3719081879,-0.4586521089,-0.0097270124,0.1669662893,-0.0739346445,-0.0917095467,-0.1543525308,-0.1256267875,0.453749299,-0.0129294349,-0.2346231639,0.398748368,-0.2917445898,0.0939348713,-0.1376038343,0.2120365351,-0.1579277515,-0.2208517492,-0.6547482014,0.0706725195,0.497736752,0.0051654596,-0.0674565881,0.0658459663,-0.1770766824,0.3712732792,-0.0562769324,0.4282302856,-0.0078680124,-0.2420659661,0.2888298631,0.022523554]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1907","title":"DBPedia14 Dataset Checksum bug?","comments":"Thanks @lhoestq! Yes, it seems back to normal after a couple of days.","body":"Hi there!!!\r\n\r\nI've been using successfully the DBPedia dataset (https:\/\/huggingface.co\/datasets\/dbpedia_14) with my codebase in the last couple of weeks, but in the last couple of days now I get this error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/conditional_classification\/basic_pipeline.py\", line 178, in \r\n main()\r\n File \".\/conditional_classification\/basic_pipeline.py\", line 128, in main\r\n corpus.load_data(limit_train_examples_per_class=args.data_args.train_examples_per_class,\r\n File \"\/home\/fp\/dev\/conditional_classification\/conditional_classification\/datasets_base.py\", line 83, in load_data\r\n datasets = load_dataset(self.name, split=dataset_split)\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 609, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 526, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 586, in _download_and_prepare\r\n verify_checksums(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/uc?export=download&id=0Bz8a_Dbh9QhbQ2Vic1kxMmZZQ1k']\r\n```\r\n\r\nI've seen this has happened before in other datasets as reported in #537.\r\n\r\nI've tried clearing my cache and call again `load_dataset` but still is not working. My same codebase is successfully downloading and using other datasets (e.g. AGNews) without any problem, so I guess something has happened specifically to the DBPedia dataset in the last few days. \r\n\r\nCan you please check if there's a problem with the checksums? \r\n\r\nOr this is related to any other stuff? I've seen that the path in the cache for the dataset is `\/home\/fp\/.cache\/huggingface\/datasets\/d_bpedia14\/dbpedia_14\/2.0.0\/a70413e39e7a716afd0e90c9e53cb053691f56f9ef5fe317bd07f2c368e8e897...` and includes `d_bpedia14` instead maybe of `dbpedia_14`. Was this maybe a bug introduced recently?\r\n\r\nThanks!","comment_length":13,"text":"DBPedia14 Dataset Checksum bug?\nHi there!!!\r\n\r\nI've been using successfully the DBPedia dataset (https:\/\/huggingface.co\/datasets\/dbpedia_14) with my codebase in the last couple of weeks, but in the last couple of days now I get this error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/conditional_classification\/basic_pipeline.py\", line 178, in \r\n main()\r\n File \".\/conditional_classification\/basic_pipeline.py\", line 128, in main\r\n corpus.load_data(limit_train_examples_per_class=args.data_args.train_examples_per_class,\r\n File \"\/home\/fp\/dev\/conditional_classification\/conditional_classification\/datasets_base.py\", line 83, in load_data\r\n datasets = load_dataset(self.name, split=dataset_split)\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 609, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 526, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 586, in _download_and_prepare\r\n verify_checksums(\r\n File \"\/home\/fp\/anaconda3\/envs\/conditional\/lib\/python3.8\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/uc?export=download&id=0Bz8a_Dbh9QhbQ2Vic1kxMmZZQ1k']\r\n```\r\n\r\nI've seen this has happened before in other datasets as reported in #537.\r\n\r\nI've tried clearing my cache and call again `load_dataset` but still is not working. My same codebase is successfully downloading and using other datasets (e.g. AGNews) without any problem, so I guess something has happened specifically to the DBPedia dataset in the last few days. \r\n\r\nCan you please check if there's a problem with the checksums? \r\n\r\nOr this is related to any other stuff? I've seen that the path in the cache for the dataset is `\/home\/fp\/.cache\/huggingface\/datasets\/d_bpedia14\/dbpedia_14\/2.0.0\/a70413e39e7a716afd0e90c9e53cb053691f56f9ef5fe317bd07f2c368e8e897...` and includes `d_bpedia14` instead maybe of `dbpedia_14`. Was this maybe a bug introduced recently?\r\n\r\nThanks!\nThanks @lhoestq! Yes, it seems back to normal after a couple of days.","embeddings":[-0.2483563125,0.35080567,-0.1227644235,0.2624529004,0.1046657935,-0.0048445282,0.3205699623,0.5194298029,-0.087741904,-0.0413587727,0.0776256993,-0.1982264817,0.0031295526,0.3474114835,-0.2444945425,-0.0099745383,0.098670274,-0.1543822438,-0.0066188341,0.0506980233,-0.176279366,0.0808001161,-0.277286768,-0.3061199486,-0.0021763062,0.2340680808,0.014913098,-0.1259736717,-0.001067427,-0.2701264322,0.5489282608,0.1795443147,0.0238150284,0.4143281579,-0.0001126272,-0.0021212655,0.3832170963,-0.0588617101,-0.328229636,0.105318971,-0.3156419694,-0.1499086767,-0.0790167302,-0.0825071856,-0.1530301273,-0.5050341487,-0.2010857463,-0.1114249155,-0.104029946,0.261859864,0.2175973356,0.2089748532,0.0917270258,-0.0789880306,-0.1643640399,0.0381903574,-0.0346372761,0.0000728814,0.1079570428,-0.1750721484,0.0917512923,0.1509220302,-0.1619757265,0.0288006309,0.1156822592,-0.2075259537,0.0970106125,-0.2640064955,0.2317044139,0.240515843,0.2926249504,-0.128769815,-0.3720597327,-0.2977039814,-0.0750023425,-0.245641768,0.2302350849,0.1716050953,-0.067303285,-0.0363245681,-0.3061080277,0.252392441,0.1045450494,0.0709883124,-0.1234941557,0.1498817354,0.0577276759,-0.0749311298,0.1063356474,-0.0413521416,0.6009082198,-0.1663401127,-0.2048625946,0.0807095915,-0.4355798364,-0.1634497792,0.0560006686,0.6840990782,0.1832090616,0.4219745696,-0.1995084137,0.1657820642,-0.1209183633,-0.0922681913,0.1504204273,-0.0343498252,-0.0519279428,0.4296023548,0.2829835415,0.3001469374,0.1284109205,0.1946967095,0.1358951181,-0.2661388516,0.2309339345,0.0963650122,0.0433475263,-0.2907250226,-0.2783105969,0.276026845,-0.1270309091,-0.1380530298,0.2978390455,0.2877974212,-0.4330213964,0.1959685087,0.0082532791,0.0210149046,0.0442288406,-0.2511606216,-0.1368148774,-0.1471276432,-0.0469176024,0.137075603,0.1180858687,0.031462837,0.2896967828,-0.0035000346,0.1563295871,-0.1104827821,0.2000196576,-0.3637543023,0.2052349597,0.5173292756,-0.0956809148,0.2081980556,0.089403756,-0.3298082948,-0.0150805023,-0.1024182662,-0.2392274141,-0.1406963617,-0.3736905754,0.2286854237,-0.3321997821,0.0105260788,-0.5877588987,-0.2102768719,0.3507718146,-0.2136421949,0.0752197951,-0.3401685059,-0.2169393599,-0.1177302599,0.3215620518,0.3882475495,-0.2872908115,0.1926524937,-0.5058418512,-0.0116746584,0.1311635673,0.0566376485,-0.3111607134,0.3517674804,-0.3844766021,-0.2458629757,0.0601757877,-0.5608974099,-0.3146588802,0.1328949034,0.1920131296,0.3282431662,-0.1377325356,-0.2058589309,0.0681966692,-0.2593610585,0.1219714582,0.0948211327,0.0682110935,0.1584233344,-0.2493803948,-0.0130494172,0.2194643766,0.1278644204,0.1831167638,0.00329765,0.0568596125,0.1512579173,0.2725539207,-0.1036928073,-0.0148238773,0.1306653172,0.2006486058,0.0827575177,0.0509694889,-0.0756189004,-0.3890631199,0.1855844557,-0.2650212049,0.0332782269,-0.1553544253,-0.0455868579,-0.19924815,-0.1255375147,-0.4297518432,-0.3456073105,0.2111902684,0.2249656916,-0.1547851264,0.1536783427,0.046917703,0.1888696998,-0.3862114549,-0.0471783876,-0.0983198136,0.3950824738,-0.0024916609,0.0562320389,0.0013942948,0.0921594203,0.1276094764,-0.0379030854,-0.0082374858,0.4010099769,0.3606117666,0.038932316,0.190829441,0.3519996405,0.256611973,-0.2802025974,-0.2643060088,0.3049584329,0.1222953796,-0.0713119507,0.1305744052,0.3840272129,-0.2733516693,0.0554502793,0.126211822,0.0377751514,0.4955723882,-0.1079363376,-0.2369183451,-0.2576626241,0.176587835,0.1049608141,0.0675177202,0.1393899769,-0.0053886087,0.0798845366,0.2055527419,-0.080689393,-0.1177334562,-0.042633567,0.121971339,-0.035609182,0.0879070237,0.7510873079,0.4312459826,-0.0813290998,0.0834221691,0.1749548316,-0.0722506121,-0.0217371378,0.0707214326,0.078366138,0.1058773994,0.3351549804,0.058534801,0.0462715179,-0.1494187713,-0.2088983655,0.173096478,0.3105882406,-0.3180760741,0.0239008889,-0.3514176905,-0.2420360297,-0.3531622291,-0.1688136905,-0.2286220491,-0.438159287,-0.0157749671,0.3220815957,0.0368187949,0.2074562609,-0.4492830634,0.0925884247,0.1287255585,-0.2775695026,0.0639834106,0.1320360452,-0.0243535209,0.0451560877,0.57372576,0.2313758731,0.2702275813,-0.3618479073,-0.1297322214,-0.523191154,-0.3683932126,0.0594919212,0.0910136774,0.3205645382,0.1584272683,0.1651458293,0.1059408858,-0.3236399591,0.0281092543,0.317777127,-0.0200341064,-0.0149845202,-0.0636154339,0.0657196939,0.0275884718,0.0795964599,-0.2562216222,-0.0792468712,-0.2330019623,0.1119934544,0.0559557192,0.0892030075,-0.0778987408,0.1029612124,-0.0143481353,0.0297972932,-0.1901176423,-0.376108855,0.1693154871,-0.09028887,-0.441747874,-0.0958381593,-0.0300432798,0.371925056,-0.0784174949,-0.3715616465,0.006258986,-0.1335083842,0.1186246425,0.2444375604,-0.3846087158,0.2823690474,0.0073235747,0.016855523,-0.1849660724,0.0201206245,-0.0567025393,-0.2161255777,0.5088825226,-0.1245597601,-0.1940550208,0.0573147088,0.3122811317,0.4735923111,-0.1680437326,0.2759514749,0.0246083308,0.4678413868,-0.1390938908,-0.4603390992,-0.1930557489,-0.095010139,-0.0335226022,-0.0393403471,-0.247411266,-0.1724460572,-0.1374422163,0.0313692875,-0.3403830826,-0.2995963395,-0.1047632396,-0.3917798996,0.2362755239,0.0033101952,0.0916021243,0.1571807712,-0.3026350141,0.1535651088,0.3441468775,-0.1502449811,-0.1311242729,-0.4588471949,-0.0238263458,0.2610839307,0.168688789,0.0576811805,0.5540510416,-0.0254554302,0.0201969072,-0.227447778,-0.033008229,0.3528869152,-0.5139679313,0.3113613129,-0.0135799488,0.0258567762,-0.0678247586,-0.2351488024,0.0483183749,-0.1841357201,0.1932884157,0.5447885394,-0.0482577458,-0.1264222264,0.2902929783,0.3781235218,-0.1861049831,-0.0063399635,-0.2953607142,-0.1861956418,-0.0172106177,0.097582601,0.1861983538,0.2329001129,-0.3318367302,-0.1802390516,-0.0489787683,0.0665417835,0.0224901978,0.0602437556,0.2055643797,0.0275994856,0.3083541393,0.1284879595,-0.138871789,0.7748520374,0.2862086892,-0.2472219318,-0.2631367445,0.0567611381,-0.1023992449,-0.0323166214,0.0993448496,-0.1670053154,-0.0142985918,0.3959570527,-0.092776753,-0.1880429387,0.0810126886,0.2037273943,-0.0244847592,-0.4803299606,-0.5200728178,-0.1251573712,0.1447148174,0.011095834,0.4486398101,0.0692506507,-0.2791030407,0.3141688406,0.0963624492,0.4851434231,0.1838276982,-0.0944215655,-0.0706914365,-0.0477726683,-0.0933511555,-0.1017453745,-0.0060383556,-0.3348731399,-0.6285871863,-0.1254462004,-0.1433281898,0.2475311011,-0.2586276829,0.113837935,0.3659522533,-0.0879793391,0.0187855698,0.0249823723,0.0654744133,-0.0623459704,-0.0027059449,-0.3052218258,0.1355485916,0.037938185,0.1731657088,-0.1275607347,0.0060586161,-0.1111586466,-0.1593545675,-0.5275295377,0.0832000524,0.0327207483,0.0402562805,0.2117589414,0.0777599216,-0.2082785666,0.5434353948,0.2335897386,0.3035466969,-0.1847197562,0.1648023874,0.0613129213,0.2169542164,0.0758589804,-0.2602525651,0.5399428606,-0.1387327313,-0.2067426741,-0.0027815464,-0.1383007765,-0.3202257752,0.1114301831,0.0230444074,-0.1734686941,-0.1759435982,-0.0219599716,-0.1599989086,0.1275666356,-0.2879592776,0.1467720866,0.0805835724,-0.0266166534,0.1897024065,0.1364036947,-0.3712579608,-0.0933347121,0.6231891513,-0.0370385312,-0.1861687601,0.3555686772,-0.0803420693,-0.1102819517,-0.2094681263,0.1944050938,0.2133210897,-0.3284565806,-0.1627129614,-0.4114899039,0.2058758438,0.0871242508,0.4304659367,0.2160063684,0.0801099166,0.1100860611,-0.7900116444,-0.03053152,0.098693423,-0.1488034278,0.1077324301,-0.1543703228,0.1304602921,0.0958669558,-0.2670515478,-0.2456792891,0.0759127736,-0.4292776287,0.0968834236,-0.3269150257,0.014633107,0.2100938112,-0.09852539,0.1508654654,0.015134925,-0.0596978031,-0.1588754803,-0.2603534162,0.1143746451,0.205262512,-0.1401935965,0.1857998222,-0.1429645121,-0.0799993202,-0.1082240865,-0.0873486027,0.0758004487,0.0700319111,0.1423600465,0.3404824436,0.0620095208,-0.0689224601,-0.2004760355,-0.0650647581,-0.0370064452,0.2050034702,0.1779527068,-0.0698505938,-0.025923213,0.0382795744,0.0719280839,-0.2412770391,0.1140172109,0.3307912052,-0.2690100074,0.0512698814,0.4562601149,0.5259580612,0.4900167584,-0.4191825986,-0.1510872245,0.3108334541,0.1734564006,-0.3344335556,-0.1327138394,0.3407932818,0.1940482706,-0.0793226883,0.1123545542,0.3362826407,-0.3540165126,0.3828054965,0.1053666994,0.1520605385,-0.5242892504,0.5086352229,0.4963040054,-0.3074856997,0.1020008773,0.2030338049,-0.189843446,-0.1762794107,0.5503347516,-0.0935347229,0.3178268671,-0.1157834008,0.007693103,-0.0532061085,-0.5167142749,0.4916111529,0.1496077925,-0.0728207901,-0.0603758581,0.0231312476,0.0525243022,-0.105610095,0.0569683053,-0.3598283529,-0.0102919452,-0.0256157313,0.0309297163,0.1982888281,-0.3819050193,-0.1348468065,0.1083562896,0.1899412423,0.0246690884,0.1944099069,0.0915195569,-0.5324066281,-0.4129911363,0.1037879735,-0.0287562143,0.0480688289,-0.1690666676,0.1855584383,0.2086615562,-0.0036593641,-0.0355615094,0.1527326405,0.242548719,0.4049273729,0.2420446873,-0.1356989145,-0.0280001089,0.0791085958,0.0403711349,0.479392767,0.2239826918,0.1611828506,0.3548166156,0.157780692,-0.1032981202,0.1443248987,0.03648993,0.2156119198,-0.2185446173,0.5506973863,0.0730593503,0.1857665777,-0.2539486587,0.1162541807,-0.0487986058,-0.0968164355,0.3136620224,0.2028799057,0.1977779418,-0.0651572123,0.0681772083,0.1485725641,0.4755242765,0.2595829368,-0.1090379804,-0.1856627613,-0.3761809766,-0.7282562256,0.1537902504,-0.055522956,0.5085514784,0.0304756537,0.1653366685,-0.1357877702,0.1360773742,0.3877424896,-0.4676198065,0.2673336267,0.070268482,-0.1864013374,-0.0820699036,-0.2114674598,0.074621737,-0.0708472729,-0.2945510447,0.2579188645,-0.0986893773,0.1454250664,0.177061975,-0.5136868954,0.0208940338,0.0947820619,0.2509298325,0.069316864,0.7262417674,0.0100471638,-0.103388235,-0.4000116587,-0.0479570143,0.0073248157,0.5848952532,0.0964269564,0.0247241743,0.1275389344,-0.0883762613,-0.3506602347,0.4802132249,0.0932057947,-0.3719081879,-0.4586521089,-0.0097270124,0.1669662893,-0.0739346445,-0.0917095467,-0.1543525308,-0.1256267875,0.453749299,-0.0129294349,-0.2346231639,0.398748368,-0.2917445898,0.0939348713,-0.1376038343,0.2120365351,-0.1579277515,-0.2208517492,-0.6547482014,0.0706725195,0.497736752,0.0051654596,-0.0674565881,0.0658459663,-0.1770766824,0.3712732792,-0.0562769324,0.4282302856,-0.0078680124,-0.2420659661,0.2888298631,0.022523554]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1906","title":"Feature Request: Support for Pandas `Categorical`","comments":"We already have a ClassLabel type that does this kind of mapping between the label ids (integers) and actual label values (strings).\r\n\r\nI wonder if actually we should use the DictionaryType from Arrow and the Categorical type from pandas for the `datasets` ClassLabel feature type.\r\nCurrently ClassLabel corresponds to `pa.int64()` in pyarrow and `dtype('int64')` in pandas (so the label names are lost during conversions).\r\n\r\nWhat do you think ?","body":"```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.Series([\"a\", \"b\", \"c\", \"a\"], dtype=\"category\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws NotImplementedError\r\n# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_table\r\n```\r\n\r\nI'm curious if https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L796 could be built out in a way similar to `Sequence`?\r\n\r\ne.g. a `Map` class (or whatever name the maintainers might prefer) that can accept:\r\n\r\n```\r\nindex_type = generate_from_arrow_type(pa_type.index_type)\r\nvalue_type = generate_from_arrow_type(pa_type.value_type)\r\n```\r\n\r\nand then additional code points to modify:\r\n\r\n- FeatureType: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L694\r\n- A branch to handle Map in get_nested_type: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L719\r\n- I don't quite understand what `encode_nested_example` does but perhaps a branch there? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L755\r\n- Similarly, I don't quite understand why `Sequence` is used this way in `generate_from_dict`, but perhaps a branch here? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L775\r\n\r\nI couldn't find other usages of `Sequence` outside of defining specific datasets, so I'm not sure if that's a comprehensive set of touchpoints.","comment_length":69,"text":"Feature Request: Support for Pandas `Categorical`\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.Series([\"a\", \"b\", \"c\", \"a\"], dtype=\"category\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws NotImplementedError\r\n# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_table\r\n```\r\n\r\nI'm curious if https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L796 could be built out in a way similar to `Sequence`?\r\n\r\ne.g. a `Map` class (or whatever name the maintainers might prefer) that can accept:\r\n\r\n```\r\nindex_type = generate_from_arrow_type(pa_type.index_type)\r\nvalue_type = generate_from_arrow_type(pa_type.value_type)\r\n```\r\n\r\nand then additional code points to modify:\r\n\r\n- FeatureType: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L694\r\n- A branch to handle Map in get_nested_type: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L719\r\n- I don't quite understand what `encode_nested_example` does but perhaps a branch there? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L755\r\n- Similarly, I don't quite understand why `Sequence` is used this way in `generate_from_dict`, but perhaps a branch here? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L775\r\n\r\nI couldn't find other usages of `Sequence` outside of defining specific datasets, so I'm not sure if that's a comprehensive set of touchpoints.\nWe already have a ClassLabel type that does this kind of mapping between the label ids (integers) and actual label values (strings).\r\n\r\nI wonder if actually we should use the DictionaryType from Arrow and the Categorical type from pandas for the `datasets` ClassLabel feature type.\r\nCurrently ClassLabel corresponds to `pa.int64()` in pyarrow and `dtype('int64')` in pandas (so the label names are lost during conversions).\r\n\r\nWhat do you think ?","embeddings":[0.0230096243,-0.0102735767,-0.1426058412,0.2076647431,0.2370132208,0.1672788262,0.1267411113,0.2487185448,-0.0738426745,-0.2155069113,0.1673993021,0.2824072838,-0.2035112977,0.4252623618,0.1327462345,-0.2236522436,0.1246828139,0.0284454841,0.077850081,0.3117170632,-0.3771744668,-0.0838787854,-0.0995519832,0.2389727235,-0.1983468533,-0.1293813884,-0.45648247,-0.3532522023,-0.2593736649,-0.8477243185,0.3983210921,0.3807965219,0.1931584775,0.1735817343,-0.000115819,-0.3079189658,0.2058762759,-0.0425236411,-0.0503565595,-0.4431983232,-0.2918069661,-0.2464248687,0.3244947791,-0.2901013196,-0.2068761438,-0.292574048,0.0067172511,-0.2727668285,0.2317002714,0.0666810945,0.1667516083,-0.0497691333,-0.0274424106,-0.1615715176,0.1318348497,0.4021803737,-0.237100631,0.284689635,0.5595126152,-0.322232008,0.2314834446,0.0884723142,0.0117629515,-0.3249572814,0.6005215049,0.0490307808,0.2645990551,-0.5582161546,-0.3079175055,0.1526240557,0.3408336937,-0.3584251702,-0.5394771099,-0.4483566582,-0.0119229034,-0.2346872836,-0.0061525847,-0.1371355653,-0.1556818336,0.1616322845,0.112262167,-0.0731579587,-0.2399319559,0.1873058379,-0.0844331756,0.1476809233,-0.0443651713,-0.1675475389,0.2131358683,-0.099540472,0.003308367,-0.1552573591,0.0259761289,0.6048634052,-0.2247818261,-0.1363115907,0.0729947016,0.2108622342,0.0177890956,-0.0083071152,0.1002103984,0.01343753,-0.231400162,-0.0256741904,0.3791758418,0.292435199,0.1800874174,-0.086328879,-0.1288450658,0.3581373394,-0.0041660462,0.0536369346,-0.1282648146,-0.1226120666,-0.1723236442,0.1386386752,-0.0707128495,0.2454463691,-0.0304598641,-0.0372074991,-0.0177116375,-0.0263489336,0.0152251087,0.2628429532,-0.065954715,0.0928890333,-0.0641002133,0.0835099369,0.2764406502,0.120650351,-0.0680939406,0.0020986979,-0.2102169693,0.1605477333,-0.0775292516,-0.185875833,-0.1492251903,-0.0344319195,0.1192694381,0.1085190102,-0.2276078761,-0.0936202332,0.1229106933,0.0074040387,-0.4257945716,0.1622291803,0.2187957615,-0.0850061551,-0.3504912853,0.1319193393,0.1542150974,-0.2472275048,-0.2392049581,0.1113661602,-0.2031580359,0.1457660347,-0.3309437335,0.4760418832,-0.008096843,0.0497261994,0.2327813506,0.0547782816,-0.263841778,-0.3284746408,-0.1710059941,0.452237606,-0.3215060234,-0.2309907079,0.0531622022,-0.0306914076,0.0361553393,0.0772850141,-0.1583134085,0.1192366183,-0.0716489926,0.1776524633,0.5333985686,-0.1753790528,-0.1190234646,0.3577374816,0.0055855042,0.1976584643,0.2812623382,0.2622909248,0.1901309639,-0.2151019275,0.0931514278,0.128313452,-0.1040712446,-0.2355068028,0.149151817,-0.3304395676,0.1761069745,0.1930086315,-0.0684674233,0.2835514843,0.0743933991,-0.3360271156,0.3749298751,-0.2894718051,0.1078301296,0.063677147,0.306604445,0.4409077764,0.0205318797,-0.479870975,-0.1257184148,0.3306059539,-0.0952279717,0.1742290556,-0.3567067087,-0.4952701032,-0.0397885852,0.2739695907,-0.2659189999,0.1416194439,-0.0091220392,-0.2109060735,0.2343253344,0.2042154521,-0.3539479673,0.5252873302,-0.0650035888,0.2001345307,-0.2408441156,0.1756791472,-0.0458536372,0.0768375322,-0.1055376828,0.3452531993,-0.0846971571,-0.119718574,0.2849961519,0.0615107976,0.2428294271,-0.0946210027,-0.3645075262,0.229264617,0.526425004,0.04913757,-0.0908039883,-0.1106586307,0.0843873024,-0.0667051449,0.016678663,0.582357347,-0.4817159772,0.2271652371,0.1415827572,0.1412526667,0.0998748615,0.2575952411,-0.2499683946,-0.1974934191,0.1438122392,0.2578533888,-0.2217444628,0.0667169839,-0.1658438444,0.191597715,0.3040448725,-0.1511575133,0.1042441651,0.103822872,0.1065929979,-0.1703601927,0.2263924032,0.1434833258,0.2757690549,0.0754674077,0.0635722578,0.1523475945,-0.1569283754,-0.0685348958,0.364499718,0.0272788499,-0.0856467262,-0.2118576765,0.0292044468,0.0771163031,0.0599482432,-0.3495509624,-0.0224032812,-0.0369299576,-0.1203360185,0.1702971458,-0.1160883382,-0.5219683051,0.1412714422,-0.5717269778,0.007809924,-0.1792367697,-0.0621801689,0.0817525163,0.0744452327,0.1140420958,-0.3413842618,-0.0354355201,0.1047067568,-0.2842971385,-0.2089479566,-0.4510963261,0.2566649616,0.0999105349,-0.0136359641,-0.1809960306,0.2080218941,0.0345660634,-0.0310896784,-0.2123503685,-0.2790402174,0.0493886098,-0.1699927449,0.0264709871,0.2983824909,-0.253008157,-0.1749667674,-0.2625124156,0.1982944012,0.1198830456,-0.1365111768,-0.2098926306,-0.1893094927,0.0895449668,-0.1031980589,0.1020331457,-0.4095129073,-0.378429234,0.676147759,0.2192848772,0.1426874399,0.1680996716,-0.0058762957,0.0730166435,-0.0896329507,-0.1047137678,-0.1668846905,0.2988793254,0.2073049098,-0.1608503014,-0.3180474043,-0.0798370764,-0.364660114,0.1057950407,0.2121289074,-0.1410655975,-0.3041145802,0.0048825415,0.5083081722,0.1135894135,0.0284402966,0.2844410539,0.2777577043,0.0575979911,0.1198542416,0.0421048477,0.0113995001,0.177919805,-0.2589248419,0.2157589197,0.4879332483,0.1209347025,0.2910847962,0.0937247127,-0.013683538,0.4954909682,-0.1015301943,0.2374844849,-0.1354480833,-0.2925463319,-0.3849201202,-0.3812013865,0.0873693228,0.0922504589,-0.1177489534,-0.1566373408,-0.2011428326,-0.0049440479,0.1318271309,-0.0870393738,-0.0268941969,-0.355404675,0.4113953114,-0.0421568081,-0.1222862974,-0.4082291126,-0.0289858393,-0.0885416195,0.124010466,0.3442616761,-0.1650990099,-0.3631921411,-0.1629422754,-0.4993342161,0.2302030325,0.6471207738,-0.1424650699,0.0527764745,-0.0200615469,-0.0053712907,0.1159427091,0.560677588,-0.0718634129,-0.3375798166,0.2686222196,-0.4713844359,-0.3781880438,0.3169171214,0.0455295332,0.044110775,-0.0156897027,0.102161333,-0.0899460167,-0.2031950802,-0.0265187807,-0.0172946416,0.0012240851,-0.0552293807,0.0234429929,-0.0369381197,-0.1320169866,-0.1758876294,0.4480686486,0.1625965387,0.0418878235,-0.1095055193,-0.2154184133,0.0515596643,0.2643419206,-0.0085911881,-0.0637493804,-0.0592568219,0.1627687365,0.0402163938,0.0491839387,0.1073396951,0.2975523472,-0.1458135247,-0.5790703893,-0.0394760035,0.0585537031,0.4294224083,0.3097966313,-0.0215763934,0.3014404774,-0.0531078354,-0.032504946,-0.3707037866,-0.2006960511,0.3816450536,-0.0846341252,-0.344147265,-0.7334685922,0.3509335518,0.1426517963,-0.3018722236,0.1473779529,0.3631760478,-0.5190951228,0.2938227355,0.040890459,0.8400938511,-0.0927528068,0.3819609284,0.3813296854,0.3095453978,0.7505726814,0.2775804698,0.0987701789,-0.3746117353,-0.2893296182,-0.3023834825,-0.1062904596,0.2692957819,0.3666596711,-0.3419593573,0.3305617869,-0.1324571818,0.2882319987,0.109804593,-0.1822684258,-0.1189226881,-0.2034186572,-0.7257787585,0.0480664074,0.4525593817,-0.0365069993,-0.2351940423,-0.2062583268,-0.1167573556,0.0931266919,-0.2607623637,-0.2176077515,-0.4321043789,0.1444190443,-0.1006666347,0.1561190635,0.2696223259,0.470572263,0.3740429282,-0.3842300177,0.0270237196,-0.1776852012,-0.1497761011,-0.1718022078,-0.0811984837,-0.0599159747,0.475372076,0.0336377285,0.0971907079,0.3608364761,0.0548866503,-0.2822104692,-0.3045469522,0.2037369162,-0.0772633478,-0.2618447244,-0.2907627225,-0.2406383306,0.0460178331,-0.1168738231,0.0699091554,0.2150891572,-0.0369708873,0.164935872,-0.3538252711,0.1511003822,-0.2132048607,0.1253356487,0.2026967406,-0.033026319,0.1077268049,0.1574068367,-0.2605880499,-0.0517194346,-0.0053987964,0.2587834895,-0.4805978239,0.1313628852,0.2005820572,-0.265095681,-0.1276811957,0.5652495623,0.2517309189,0.2128712684,-0.0211022813,-0.1289938837,-0.2973949015,0.3678028584,0.1430948228,0.2439884841,-0.1040255874,0.1322651803,-0.1852490604,-0.1136159152,-0.2206497937,-0.0021180212,-0.0052057137,-0.0675509349,0.4329418242,0.0207916331,0.1636742502,0.2392761111,0.1212816462,0.2636927664,-0.1925227195,-0.132017985,-0.0871845409,0.1703848541,-0.0201830156,-0.0114978524,0.0209324118,-0.0894585252,-0.2565029263,-0.2649773955,0.094342649,0.2512863576,0.1210567057,-0.0448327065,0.399472177,-0.0346534289,-0.1810763329,0.3869203925,0.0380889513,0.4918481112,0.0317566842,0.0474308319,-0.3425094187,0.0545476004,0.0307748616,0.3291054964,-0.0392379351,-0.4100055993,0.6409253478,0.0848434493,-0.1820202768,0.1968024522,0.5251167417,0.2113816291,0.1651661396,0.0732804313,0.1216478497,0.0837516412,-0.333778888,-0.3142723441,0.1178408861,-0.1556552798,0.1714912206,0.5202995539,0.3050468564,0.2484450638,0.1247217357,-0.142963782,0.2160950899,0.198144421,0.0745482072,0.3195811212,0.3900642097,-0.0091482764,0.356150955,0.3804430068,0.5041666627,0.183292985,-0.0932919085,0.0783436,0.0882698596,-0.0172891822,0.6047267318,0.0295796096,0.0089012086,0.4108660221,0.0399434976,-0.0082164165,-0.0557898246,0.0577757806,-0.1989376992,0.0129611809,-0.0629524291,0.2387240976,0.2161091566,-0.1524887234,0.4791820645,-0.1766897142,-0.1289842725,-0.3476141691,-0.1384478658,-0.2974571288,0.4343164563,-0.1308518052,0.565055728,-0.0634816661,0.1978338361,-0.0791861415,0.3597240448,-0.0898838118,-0.0796705261,-0.0958931521,-0.0673906654,0.0075925766,0.2095282674,0.4526987076,0.180991888,-0.1542158425,0.1083929986,-0.4261741936,-0.2367450297,-0.0348737501,0.2622194886,0.1265501231,0.0943511501,0.2722560465,0.0530812331,-0.0310173091,0.0222226586,0.0101432232,0.2748621404,-0.3390365839,0.7011240125,-0.2181453556,-0.2350433767,0.1938218474,0.3255448341,-0.3201076984,0.0481975228,0.0591655709,-0.169170022,0.1492698789,0.1201524213,0.0416862145,-0.0675468519,0.1855736673,0.1969929487,0.4075565636,0.0107357595,-0.2408178449,-0.4748313427,-0.1409271061,0.254796952,-0.2541853189,0.2475463003,-0.0389153212,-0.0236245953,0.091067113,-0.0815633759,-0.3627574146,-0.1309396625,0.214824155,-0.0924031287,-0.2767859101,0.184160322,0.1239567176,-0.1801171154,-0.2226980478,0.0785784051,-0.2537121177,-0.0861013308,-0.1989878565,0.1434233934,-0.1959645003,0.0455567054,0.2694401741,0.222721979,0.180764541,0.2077305764,-0.1693365723,-0.0346198343,-0.3444247544,0.0613573007,0.5801233053,-0.0249673575,0.3234552741,-0.3697853386,-0.2753036916,-0.0232932251,0.2424982786,0.2755203545,-0.193352133,-0.5966461897,0.0874607041,-0.4430664182,-0.0298179649,0.169570297,0.3679272532,-0.1307651103,0.3309237063,0.1300217658,-0.2077973485,0.3422713876,-0.4958495498,-0.2886917889,0.0350957923,-0.2006175965,0.3782958388,-0.1855077744,-0.1591355652,-0.0128236292,0.4850818515,0.1197656393,-0.1062523574,0.1743545085,-0.2196145952,-0.108369641,-0.3893577456,0.0817014873,-0.2865290642,-0.0213867761,0.0420439988,-0.4636683464]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1906","title":"Feature Request: Support for Pandas `Categorical`","comments":"Now that I've heard you explain ClassLabel, that makes a lot of sense! While DictionaryType for Arrow (I think) can have arbitrarily typed keys, so it won't cover all potential cases, pandas' Category is *probably* the most common use for that pyarrow type, and ClassLabel should match that perfectly?\r\n\r\nOther thoughts:\r\n\r\n- changing the resulting patype on ClassLabel might be backward-incompatible? I'm not totally sure if users of the `datasets` library tend to directly access the `patype` attribute (I don't think we really do, but we haven't been using it for very long yet).\r\n- would ClassLabel's dtype change to `dict[int64, string]`? It seems like in practice a ClassLabel (when not explicitly specified) would be constructed from the DictionaryType branch of `generate_from_arrow_type`, so it's not totally clear to me that anyone ever actually accesses\/uses that dtype?\r\n- I don't quite know how `.int2str` and `.str2int` are used in practice - would those be kept? Perhaps the implementation might actually be substantially smaller if we can just delegate to pyarrow's dict methods?\r\n\r\nAnother idea that just occurred to me: add a branch in here to generate a ClassLabel if the dict key is int64 and the values are string: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L932 , and then don't touch anything else.\r\n\r\nIn practice, I don't think this would be backward-incompatible in a way anyone would care about since the current behavior just throws an exception, and this way, we could support *reading* a pandas Categorical into a `Dataset` as a ClassLabel. I *think* from there, while it would require some custom glue it wouldn't be too hard to convert the ClassLabel into a pandas Category if we want to go back - I think this would improve on the current behavior without risking changing the behavior of ClassLabel in a backward-incompat way.\r\n\r\nThoughts? I'm not sure if this is overly cautious. Whichever approach you think is better, I'd be happy to take it on!\r\n","body":"```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.Series([\"a\", \"b\", \"c\", \"a\"], dtype=\"category\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws NotImplementedError\r\n# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_table\r\n```\r\n\r\nI'm curious if https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L796 could be built out in a way similar to `Sequence`?\r\n\r\ne.g. a `Map` class (or whatever name the maintainers might prefer) that can accept:\r\n\r\n```\r\nindex_type = generate_from_arrow_type(pa_type.index_type)\r\nvalue_type = generate_from_arrow_type(pa_type.value_type)\r\n```\r\n\r\nand then additional code points to modify:\r\n\r\n- FeatureType: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L694\r\n- A branch to handle Map in get_nested_type: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L719\r\n- I don't quite understand what `encode_nested_example` does but perhaps a branch there? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L755\r\n- Similarly, I don't quite understand why `Sequence` is used this way in `generate_from_dict`, but perhaps a branch here? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L775\r\n\r\nI couldn't find other usages of `Sequence` outside of defining specific datasets, so I'm not sure if that's a comprehensive set of touchpoints.","comment_length":319,"text":"Feature Request: Support for Pandas `Categorical`\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.Series([\"a\", \"b\", \"c\", \"a\"], dtype=\"category\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws NotImplementedError\r\n# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_table\r\n```\r\n\r\nI'm curious if https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L796 could be built out in a way similar to `Sequence`?\r\n\r\ne.g. a `Map` class (or whatever name the maintainers might prefer) that can accept:\r\n\r\n```\r\nindex_type = generate_from_arrow_type(pa_type.index_type)\r\nvalue_type = generate_from_arrow_type(pa_type.value_type)\r\n```\r\n\r\nand then additional code points to modify:\r\n\r\n- FeatureType: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L694\r\n- A branch to handle Map in get_nested_type: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L719\r\n- I don't quite understand what `encode_nested_example` does but perhaps a branch there? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L755\r\n- Similarly, I don't quite understand why `Sequence` is used this way in `generate_from_dict`, but perhaps a branch here? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L775\r\n\r\nI couldn't find other usages of `Sequence` outside of defining specific datasets, so I'm not sure if that's a comprehensive set of touchpoints.\nNow that I've heard you explain ClassLabel, that makes a lot of sense! While DictionaryType for Arrow (I think) can have arbitrarily typed keys, so it won't cover all potential cases, pandas' Category is *probably* the most common use for that pyarrow type, and ClassLabel should match that perfectly?\r\n\r\nOther thoughts:\r\n\r\n- changing the resulting patype on ClassLabel might be backward-incompatible? I'm not totally sure if users of the `datasets` library tend to directly access the `patype` attribute (I don't think we really do, but we haven't been using it for very long yet).\r\n- would ClassLabel's dtype change to `dict[int64, string]`? It seems like in practice a ClassLabel (when not explicitly specified) would be constructed from the DictionaryType branch of `generate_from_arrow_type`, so it's not totally clear to me that anyone ever actually accesses\/uses that dtype?\r\n- I don't quite know how `.int2str` and `.str2int` are used in practice - would those be kept? Perhaps the implementation might actually be substantially smaller if we can just delegate to pyarrow's dict methods?\r\n\r\nAnother idea that just occurred to me: add a branch in here to generate a ClassLabel if the dict key is int64 and the values are string: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L932 , and then don't touch anything else.\r\n\r\nIn practice, I don't think this would be backward-incompatible in a way anyone would care about since the current behavior just throws an exception, and this way, we could support *reading* a pandas Categorical into a `Dataset` as a ClassLabel. I *think* from there, while it would require some custom glue it wouldn't be too hard to convert the ClassLabel into a pandas Category if we want to go back - I think this would improve on the current behavior without risking changing the behavior of ClassLabel in a backward-incompat way.\r\n\r\nThoughts? I'm not sure if this is overly cautious. Whichever approach you think is better, I'd be happy to take it on!\r\n","embeddings":[0.0230096243,-0.0102735767,-0.1426058412,0.2076647431,0.2370132208,0.1672788262,0.1267411113,0.2487185448,-0.0738426745,-0.2155069113,0.1673993021,0.2824072838,-0.2035112977,0.4252623618,0.1327462345,-0.2236522436,0.1246828139,0.0284454841,0.077850081,0.3117170632,-0.3771744668,-0.0838787854,-0.0995519832,0.2389727235,-0.1983468533,-0.1293813884,-0.45648247,-0.3532522023,-0.2593736649,-0.8477243185,0.3983210921,0.3807965219,0.1931584775,0.1735817343,-0.000115819,-0.3079189658,0.2058762759,-0.0425236411,-0.0503565595,-0.4431983232,-0.2918069661,-0.2464248687,0.3244947791,-0.2901013196,-0.2068761438,-0.292574048,0.0067172511,-0.2727668285,0.2317002714,0.0666810945,0.1667516083,-0.0497691333,-0.0274424106,-0.1615715176,0.1318348497,0.4021803737,-0.237100631,0.284689635,0.5595126152,-0.322232008,0.2314834446,0.0884723142,0.0117629515,-0.3249572814,0.6005215049,0.0490307808,0.2645990551,-0.5582161546,-0.3079175055,0.1526240557,0.3408336937,-0.3584251702,-0.5394771099,-0.4483566582,-0.0119229034,-0.2346872836,-0.0061525847,-0.1371355653,-0.1556818336,0.1616322845,0.112262167,-0.0731579587,-0.2399319559,0.1873058379,-0.0844331756,0.1476809233,-0.0443651713,-0.1675475389,0.2131358683,-0.099540472,0.003308367,-0.1552573591,0.0259761289,0.6048634052,-0.2247818261,-0.1363115907,0.0729947016,0.2108622342,0.0177890956,-0.0083071152,0.1002103984,0.01343753,-0.231400162,-0.0256741904,0.3791758418,0.292435199,0.1800874174,-0.086328879,-0.1288450658,0.3581373394,-0.0041660462,0.0536369346,-0.1282648146,-0.1226120666,-0.1723236442,0.1386386752,-0.0707128495,0.2454463691,-0.0304598641,-0.0372074991,-0.0177116375,-0.0263489336,0.0152251087,0.2628429532,-0.065954715,0.0928890333,-0.0641002133,0.0835099369,0.2764406502,0.120650351,-0.0680939406,0.0020986979,-0.2102169693,0.1605477333,-0.0775292516,-0.185875833,-0.1492251903,-0.0344319195,0.1192694381,0.1085190102,-0.2276078761,-0.0936202332,0.1229106933,0.0074040387,-0.4257945716,0.1622291803,0.2187957615,-0.0850061551,-0.3504912853,0.1319193393,0.1542150974,-0.2472275048,-0.2392049581,0.1113661602,-0.2031580359,0.1457660347,-0.3309437335,0.4760418832,-0.008096843,0.0497261994,0.2327813506,0.0547782816,-0.263841778,-0.3284746408,-0.1710059941,0.452237606,-0.3215060234,-0.2309907079,0.0531622022,-0.0306914076,0.0361553393,0.0772850141,-0.1583134085,0.1192366183,-0.0716489926,0.1776524633,0.5333985686,-0.1753790528,-0.1190234646,0.3577374816,0.0055855042,0.1976584643,0.2812623382,0.2622909248,0.1901309639,-0.2151019275,0.0931514278,0.128313452,-0.1040712446,-0.2355068028,0.149151817,-0.3304395676,0.1761069745,0.1930086315,-0.0684674233,0.2835514843,0.0743933991,-0.3360271156,0.3749298751,-0.2894718051,0.1078301296,0.063677147,0.306604445,0.4409077764,0.0205318797,-0.479870975,-0.1257184148,0.3306059539,-0.0952279717,0.1742290556,-0.3567067087,-0.4952701032,-0.0397885852,0.2739695907,-0.2659189999,0.1416194439,-0.0091220392,-0.2109060735,0.2343253344,0.2042154521,-0.3539479673,0.5252873302,-0.0650035888,0.2001345307,-0.2408441156,0.1756791472,-0.0458536372,0.0768375322,-0.1055376828,0.3452531993,-0.0846971571,-0.119718574,0.2849961519,0.0615107976,0.2428294271,-0.0946210027,-0.3645075262,0.229264617,0.526425004,0.04913757,-0.0908039883,-0.1106586307,0.0843873024,-0.0667051449,0.016678663,0.582357347,-0.4817159772,0.2271652371,0.1415827572,0.1412526667,0.0998748615,0.2575952411,-0.2499683946,-0.1974934191,0.1438122392,0.2578533888,-0.2217444628,0.0667169839,-0.1658438444,0.191597715,0.3040448725,-0.1511575133,0.1042441651,0.103822872,0.1065929979,-0.1703601927,0.2263924032,0.1434833258,0.2757690549,0.0754674077,0.0635722578,0.1523475945,-0.1569283754,-0.0685348958,0.364499718,0.0272788499,-0.0856467262,-0.2118576765,0.0292044468,0.0771163031,0.0599482432,-0.3495509624,-0.0224032812,-0.0369299576,-0.1203360185,0.1702971458,-0.1160883382,-0.5219683051,0.1412714422,-0.5717269778,0.007809924,-0.1792367697,-0.0621801689,0.0817525163,0.0744452327,0.1140420958,-0.3413842618,-0.0354355201,0.1047067568,-0.2842971385,-0.2089479566,-0.4510963261,0.2566649616,0.0999105349,-0.0136359641,-0.1809960306,0.2080218941,0.0345660634,-0.0310896784,-0.2123503685,-0.2790402174,0.0493886098,-0.1699927449,0.0264709871,0.2983824909,-0.253008157,-0.1749667674,-0.2625124156,0.1982944012,0.1198830456,-0.1365111768,-0.2098926306,-0.1893094927,0.0895449668,-0.1031980589,0.1020331457,-0.4095129073,-0.378429234,0.676147759,0.2192848772,0.1426874399,0.1680996716,-0.0058762957,0.0730166435,-0.0896329507,-0.1047137678,-0.1668846905,0.2988793254,0.2073049098,-0.1608503014,-0.3180474043,-0.0798370764,-0.364660114,0.1057950407,0.2121289074,-0.1410655975,-0.3041145802,0.0048825415,0.5083081722,0.1135894135,0.0284402966,0.2844410539,0.2777577043,0.0575979911,0.1198542416,0.0421048477,0.0113995001,0.177919805,-0.2589248419,0.2157589197,0.4879332483,0.1209347025,0.2910847962,0.0937247127,-0.013683538,0.4954909682,-0.1015301943,0.2374844849,-0.1354480833,-0.2925463319,-0.3849201202,-0.3812013865,0.0873693228,0.0922504589,-0.1177489534,-0.1566373408,-0.2011428326,-0.0049440479,0.1318271309,-0.0870393738,-0.0268941969,-0.355404675,0.4113953114,-0.0421568081,-0.1222862974,-0.4082291126,-0.0289858393,-0.0885416195,0.124010466,0.3442616761,-0.1650990099,-0.3631921411,-0.1629422754,-0.4993342161,0.2302030325,0.6471207738,-0.1424650699,0.0527764745,-0.0200615469,-0.0053712907,0.1159427091,0.560677588,-0.0718634129,-0.3375798166,0.2686222196,-0.4713844359,-0.3781880438,0.3169171214,0.0455295332,0.044110775,-0.0156897027,0.102161333,-0.0899460167,-0.2031950802,-0.0265187807,-0.0172946416,0.0012240851,-0.0552293807,0.0234429929,-0.0369381197,-0.1320169866,-0.1758876294,0.4480686486,0.1625965387,0.0418878235,-0.1095055193,-0.2154184133,0.0515596643,0.2643419206,-0.0085911881,-0.0637493804,-0.0592568219,0.1627687365,0.0402163938,0.0491839387,0.1073396951,0.2975523472,-0.1458135247,-0.5790703893,-0.0394760035,0.0585537031,0.4294224083,0.3097966313,-0.0215763934,0.3014404774,-0.0531078354,-0.032504946,-0.3707037866,-0.2006960511,0.3816450536,-0.0846341252,-0.344147265,-0.7334685922,0.3509335518,0.1426517963,-0.3018722236,0.1473779529,0.3631760478,-0.5190951228,0.2938227355,0.040890459,0.8400938511,-0.0927528068,0.3819609284,0.3813296854,0.3095453978,0.7505726814,0.2775804698,0.0987701789,-0.3746117353,-0.2893296182,-0.3023834825,-0.1062904596,0.2692957819,0.3666596711,-0.3419593573,0.3305617869,-0.1324571818,0.2882319987,0.109804593,-0.1822684258,-0.1189226881,-0.2034186572,-0.7257787585,0.0480664074,0.4525593817,-0.0365069993,-0.2351940423,-0.2062583268,-0.1167573556,0.0931266919,-0.2607623637,-0.2176077515,-0.4321043789,0.1444190443,-0.1006666347,0.1561190635,0.2696223259,0.470572263,0.3740429282,-0.3842300177,0.0270237196,-0.1776852012,-0.1497761011,-0.1718022078,-0.0811984837,-0.0599159747,0.475372076,0.0336377285,0.0971907079,0.3608364761,0.0548866503,-0.2822104692,-0.3045469522,0.2037369162,-0.0772633478,-0.2618447244,-0.2907627225,-0.2406383306,0.0460178331,-0.1168738231,0.0699091554,0.2150891572,-0.0369708873,0.164935872,-0.3538252711,0.1511003822,-0.2132048607,0.1253356487,0.2026967406,-0.033026319,0.1077268049,0.1574068367,-0.2605880499,-0.0517194346,-0.0053987964,0.2587834895,-0.4805978239,0.1313628852,0.2005820572,-0.265095681,-0.1276811957,0.5652495623,0.2517309189,0.2128712684,-0.0211022813,-0.1289938837,-0.2973949015,0.3678028584,0.1430948228,0.2439884841,-0.1040255874,0.1322651803,-0.1852490604,-0.1136159152,-0.2206497937,-0.0021180212,-0.0052057137,-0.0675509349,0.4329418242,0.0207916331,0.1636742502,0.2392761111,0.1212816462,0.2636927664,-0.1925227195,-0.132017985,-0.0871845409,0.1703848541,-0.0201830156,-0.0114978524,0.0209324118,-0.0894585252,-0.2565029263,-0.2649773955,0.094342649,0.2512863576,0.1210567057,-0.0448327065,0.399472177,-0.0346534289,-0.1810763329,0.3869203925,0.0380889513,0.4918481112,0.0317566842,0.0474308319,-0.3425094187,0.0545476004,0.0307748616,0.3291054964,-0.0392379351,-0.4100055993,0.6409253478,0.0848434493,-0.1820202768,0.1968024522,0.5251167417,0.2113816291,0.1651661396,0.0732804313,0.1216478497,0.0837516412,-0.333778888,-0.3142723441,0.1178408861,-0.1556552798,0.1714912206,0.5202995539,0.3050468564,0.2484450638,0.1247217357,-0.142963782,0.2160950899,0.198144421,0.0745482072,0.3195811212,0.3900642097,-0.0091482764,0.356150955,0.3804430068,0.5041666627,0.183292985,-0.0932919085,0.0783436,0.0882698596,-0.0172891822,0.6047267318,0.0295796096,0.0089012086,0.4108660221,0.0399434976,-0.0082164165,-0.0557898246,0.0577757806,-0.1989376992,0.0129611809,-0.0629524291,0.2387240976,0.2161091566,-0.1524887234,0.4791820645,-0.1766897142,-0.1289842725,-0.3476141691,-0.1384478658,-0.2974571288,0.4343164563,-0.1308518052,0.565055728,-0.0634816661,0.1978338361,-0.0791861415,0.3597240448,-0.0898838118,-0.0796705261,-0.0958931521,-0.0673906654,0.0075925766,0.2095282674,0.4526987076,0.180991888,-0.1542158425,0.1083929986,-0.4261741936,-0.2367450297,-0.0348737501,0.2622194886,0.1265501231,0.0943511501,0.2722560465,0.0530812331,-0.0310173091,0.0222226586,0.0101432232,0.2748621404,-0.3390365839,0.7011240125,-0.2181453556,-0.2350433767,0.1938218474,0.3255448341,-0.3201076984,0.0481975228,0.0591655709,-0.169170022,0.1492698789,0.1201524213,0.0416862145,-0.0675468519,0.1855736673,0.1969929487,0.4075565636,0.0107357595,-0.2408178449,-0.4748313427,-0.1409271061,0.254796952,-0.2541853189,0.2475463003,-0.0389153212,-0.0236245953,0.091067113,-0.0815633759,-0.3627574146,-0.1309396625,0.214824155,-0.0924031287,-0.2767859101,0.184160322,0.1239567176,-0.1801171154,-0.2226980478,0.0785784051,-0.2537121177,-0.0861013308,-0.1989878565,0.1434233934,-0.1959645003,0.0455567054,0.2694401741,0.222721979,0.180764541,0.2077305764,-0.1693365723,-0.0346198343,-0.3444247544,0.0613573007,0.5801233053,-0.0249673575,0.3234552741,-0.3697853386,-0.2753036916,-0.0232932251,0.2424982786,0.2755203545,-0.193352133,-0.5966461897,0.0874607041,-0.4430664182,-0.0298179649,0.169570297,0.3679272532,-0.1307651103,0.3309237063,0.1300217658,-0.2077973485,0.3422713876,-0.4958495498,-0.2886917889,0.0350957923,-0.2006175965,0.3782958388,-0.1855077744,-0.1591355652,-0.0128236292,0.4850818515,0.1197656393,-0.1062523574,0.1743545085,-0.2196145952,-0.108369641,-0.3893577456,0.0817014873,-0.2865290642,-0.0213867761,0.0420439988,-0.4636683464]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1906","title":"Feature Request: Support for Pandas `Categorical`","comments":"I think we can first keep the int64 precision but with an arrow Dictionary for ClassLabel, and focus on the connection with arrow and pandas.\r\n\r\nIn this scope, I really like the idea of checking for the dictionary type:\r\n\r\n> Another idea that just occurred to me: add a branch in here to generate a ClassLabel if the dict key is int64 and the values are string: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L932 , and then don't touch anything else.\r\n\r\nThis looks like a great start.\r\n\r\nThen as you said we'd have to add the conversion from classlabel to the correct arrow dictionary type. Arrow is already able to convert from arrow Dictionary to pandas Categorical so it should be enough.\r\n\r\nI can see two things that we must take case of to make this change backward compatible:\r\n- first we must still be able to load an arrow file with arrow int64 dtype and `datasets` ClassLabel type without crashing. This can be fixed by casting the arrow int64 array to an arrow Dictionary array on-the-fly when loading the table in the ArrowReader.\r\n- then we still have to return integers when accessing examples from a ClassLabel column. Currently it would return the strings values since it's based on the pandas behavior for converting from pandas to python\/numpy. To do so we just have to adapt the python\/numpy extractors in formatting.py (it takes care of converting an arrow table to a dictionary of python objects by doing arrow table -> pandas dataframe -> python dictionary)\r\n\r\nAny help on this matter is very much welcome :)","body":"```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.Series([\"a\", \"b\", \"c\", \"a\"], dtype=\"category\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws NotImplementedError\r\n# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_table\r\n```\r\n\r\nI'm curious if https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L796 could be built out in a way similar to `Sequence`?\r\n\r\ne.g. a `Map` class (or whatever name the maintainers might prefer) that can accept:\r\n\r\n```\r\nindex_type = generate_from_arrow_type(pa_type.index_type)\r\nvalue_type = generate_from_arrow_type(pa_type.value_type)\r\n```\r\n\r\nand then additional code points to modify:\r\n\r\n- FeatureType: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L694\r\n- A branch to handle Map in get_nested_type: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L719\r\n- I don't quite understand what `encode_nested_example` does but perhaps a branch there? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L755\r\n- Similarly, I don't quite understand why `Sequence` is used this way in `generate_from_dict`, but perhaps a branch here? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L775\r\n\r\nI couldn't find other usages of `Sequence` outside of defining specific datasets, so I'm not sure if that's a comprehensive set of touchpoints.","comment_length":260,"text":"Feature Request: Support for Pandas `Categorical`\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.Series([\"a\", \"b\", \"c\", \"a\"], dtype=\"category\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws NotImplementedError\r\n# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_table\r\n```\r\n\r\nI'm curious if https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L796 could be built out in a way similar to `Sequence`?\r\n\r\ne.g. a `Map` class (or whatever name the maintainers might prefer) that can accept:\r\n\r\n```\r\nindex_type = generate_from_arrow_type(pa_type.index_type)\r\nvalue_type = generate_from_arrow_type(pa_type.value_type)\r\n```\r\n\r\nand then additional code points to modify:\r\n\r\n- FeatureType: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L694\r\n- A branch to handle Map in get_nested_type: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L719\r\n- I don't quite understand what `encode_nested_example` does but perhaps a branch there? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L755\r\n- Similarly, I don't quite understand why `Sequence` is used this way in `generate_from_dict`, but perhaps a branch here? https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L775\r\n\r\nI couldn't find other usages of `Sequence` outside of defining specific datasets, so I'm not sure if that's a comprehensive set of touchpoints.\nI think we can first keep the int64 precision but with an arrow Dictionary for ClassLabel, and focus on the connection with arrow and pandas.\r\n\r\nIn this scope, I really like the idea of checking for the dictionary type:\r\n\r\n> Another idea that just occurred to me: add a branch in here to generate a ClassLabel if the dict key is int64 and the values are string: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L932 , and then don't touch anything else.\r\n\r\nThis looks like a great start.\r\n\r\nThen as you said we'd have to add the conversion from classlabel to the correct arrow dictionary type. Arrow is already able to convert from arrow Dictionary to pandas Categorical so it should be enough.\r\n\r\nI can see two things that we must take case of to make this change backward compatible:\r\n- first we must still be able to load an arrow file with arrow int64 dtype and `datasets` ClassLabel type without crashing. This can be fixed by casting the arrow int64 array to an arrow Dictionary array on-the-fly when loading the table in the ArrowReader.\r\n- then we still have to return integers when accessing examples from a ClassLabel column. Currently it would return the strings values since it's based on the pandas behavior for converting from pandas to python\/numpy. To do so we just have to adapt the python\/numpy extractors in formatting.py (it takes care of converting an arrow table to a dictionary of python objects by doing arrow table -> pandas dataframe -> python dictionary)\r\n\r\nAny help on this matter is very much welcome :)","embeddings":[0.0230096243,-0.0102735767,-0.1426058412,0.2076647431,0.2370132208,0.1672788262,0.1267411113,0.2487185448,-0.0738426745,-0.2155069113,0.1673993021,0.2824072838,-0.2035112977,0.4252623618,0.1327462345,-0.2236522436,0.1246828139,0.0284454841,0.077850081,0.3117170632,-0.3771744668,-0.0838787854,-0.0995519832,0.2389727235,-0.1983468533,-0.1293813884,-0.45648247,-0.3532522023,-0.2593736649,-0.8477243185,0.3983210921,0.3807965219,0.1931584775,0.1735817343,-0.000115819,-0.3079189658,0.2058762759,-0.0425236411,-0.0503565595,-0.4431983232,-0.2918069661,-0.2464248687,0.3244947791,-0.2901013196,-0.2068761438,-0.292574048,0.0067172511,-0.2727668285,0.2317002714,0.0666810945,0.1667516083,-0.0497691333,-0.0274424106,-0.1615715176,0.1318348497,0.4021803737,-0.237100631,0.284689635,0.5595126152,-0.322232008,0.2314834446,0.0884723142,0.0117629515,-0.3249572814,0.6005215049,0.0490307808,0.2645990551,-0.5582161546,-0.3079175055,0.1526240557,0.3408336937,-0.3584251702,-0.5394771099,-0.4483566582,-0.0119229034,-0.2346872836,-0.0061525847,-0.1371355653,-0.1556818336,0.1616322845,0.112262167,-0.0731579587,-0.2399319559,0.1873058379,-0.0844331756,0.1476809233,-0.0443651713,-0.1675475389,0.2131358683,-0.099540472,0.003308367,-0.1552573591,0.0259761289,0.6048634052,-0.2247818261,-0.1363115907,0.0729947016,0.2108622342,0.0177890956,-0.0083071152,0.1002103984,0.01343753,-0.231400162,-0.0256741904,0.3791758418,0.292435199,0.1800874174,-0.086328879,-0.1288450658,0.3581373394,-0.0041660462,0.0536369346,-0.1282648146,-0.1226120666,-0.1723236442,0.1386386752,-0.0707128495,0.2454463691,-0.0304598641,-0.0372074991,-0.0177116375,-0.0263489336,0.0152251087,0.2628429532,-0.065954715,0.0928890333,-0.0641002133,0.0835099369,0.2764406502,0.120650351,-0.0680939406,0.0020986979,-0.2102169693,0.1605477333,-0.0775292516,-0.185875833,-0.1492251903,-0.0344319195,0.1192694381,0.1085190102,-0.2276078761,-0.0936202332,0.1229106933,0.0074040387,-0.4257945716,0.1622291803,0.2187957615,-0.0850061551,-0.3504912853,0.1319193393,0.1542150974,-0.2472275048,-0.2392049581,0.1113661602,-0.2031580359,0.1457660347,-0.3309437335,0.4760418832,-0.008096843,0.0497261994,0.2327813506,0.0547782816,-0.263841778,-0.3284746408,-0.1710059941,0.452237606,-0.3215060234,-0.2309907079,0.0531622022,-0.0306914076,0.0361553393,0.0772850141,-0.1583134085,0.1192366183,-0.0716489926,0.1776524633,0.5333985686,-0.1753790528,-0.1190234646,0.3577374816,0.0055855042,0.1976584643,0.2812623382,0.2622909248,0.1901309639,-0.2151019275,0.0931514278,0.128313452,-0.1040712446,-0.2355068028,0.149151817,-0.3304395676,0.1761069745,0.1930086315,-0.0684674233,0.2835514843,0.0743933991,-0.3360271156,0.3749298751,-0.2894718051,0.1078301296,0.063677147,0.306604445,0.4409077764,0.0205318797,-0.479870975,-0.1257184148,0.3306059539,-0.0952279717,0.1742290556,-0.3567067087,-0.4952701032,-0.0397885852,0.2739695907,-0.2659189999,0.1416194439,-0.0091220392,-0.2109060735,0.2343253344,0.2042154521,-0.3539479673,0.5252873302,-0.0650035888,0.2001345307,-0.2408441156,0.1756791472,-0.0458536372,0.0768375322,-0.1055376828,0.3452531993,-0.0846971571,-0.119718574,0.2849961519,0.0615107976,0.2428294271,-0.0946210027,-0.3645075262,0.229264617,0.526425004,0.04913757,-0.0908039883,-0.1106586307,0.0843873024,-0.0667051449,0.016678663,0.582357347,-0.4817159772,0.2271652371,0.1415827572,0.1412526667,0.0998748615,0.2575952411,-0.2499683946,-0.1974934191,0.1438122392,0.2578533888,-0.2217444628,0.0667169839,-0.1658438444,0.191597715,0.3040448725,-0.1511575133,0.1042441651,0.103822872,0.1065929979,-0.1703601927,0.2263924032,0.1434833258,0.2757690549,0.0754674077,0.0635722578,0.1523475945,-0.1569283754,-0.0685348958,0.364499718,0.0272788499,-0.0856467262,-0.2118576765,0.0292044468,0.0771163031,0.0599482432,-0.3495509624,-0.0224032812,-0.0369299576,-0.1203360185,0.1702971458,-0.1160883382,-0.5219683051,0.1412714422,-0.5717269778,0.007809924,-0.1792367697,-0.0621801689,0.0817525163,0.0744452327,0.1140420958,-0.3413842618,-0.0354355201,0.1047067568,-0.2842971385,-0.2089479566,-0.4510963261,0.2566649616,0.0999105349,-0.0136359641,-0.1809960306,0.2080218941,0.0345660634,-0.0310896784,-0.2123503685,-0.2790402174,0.0493886098,-0.1699927449,0.0264709871,0.2983824909,-0.253008157,-0.1749667674,-0.2625124156,0.1982944012,0.1198830456,-0.1365111768,-0.2098926306,-0.1893094927,0.0895449668,-0.1031980589,0.1020331457,-0.4095129073,-0.378429234,0.676147759,0.2192848772,0.1426874399,0.1680996716,-0.0058762957,0.0730166435,-0.0896329507,-0.1047137678,-0.1668846905,0.2988793254,0.2073049098,-0.1608503014,-0.3180474043,-0.0798370764,-0.364660114,0.1057950407,0.2121289074,-0.1410655975,-0.3041145802,0.0048825415,0.5083081722,0.1135894135,0.0284402966,0.2844410539,0.2777577043,0.0575979911,0.1198542416,0.0421048477,0.0113995001,0.177919805,-0.2589248419,0.2157589197,0.4879332483,0.1209347025,0.2910847962,0.0937247127,-0.013683538,0.4954909682,-0.1015301943,0.2374844849,-0.1354480833,-0.2925463319,-0.3849201202,-0.3812013865,0.0873693228,0.0922504589,-0.1177489534,-0.1566373408,-0.2011428326,-0.0049440479,0.1318271309,-0.0870393738,-0.0268941969,-0.355404675,0.4113953114,-0.0421568081,-0.1222862974,-0.4082291126,-0.0289858393,-0.0885416195,0.124010466,0.3442616761,-0.1650990099,-0.3631921411,-0.1629422754,-0.4993342161,0.2302030325,0.6471207738,-0.1424650699,0.0527764745,-0.0200615469,-0.0053712907,0.1159427091,0.560677588,-0.0718634129,-0.3375798166,0.2686222196,-0.4713844359,-0.3781880438,0.3169171214,0.0455295332,0.044110775,-0.0156897027,0.102161333,-0.0899460167,-0.2031950802,-0.0265187807,-0.0172946416,0.0012240851,-0.0552293807,0.0234429929,-0.0369381197,-0.1320169866,-0.1758876294,0.4480686486,0.1625965387,0.0418878235,-0.1095055193,-0.2154184133,0.0515596643,0.2643419206,-0.0085911881,-0.0637493804,-0.0592568219,0.1627687365,0.0402163938,0.0491839387,0.1073396951,0.2975523472,-0.1458135247,-0.5790703893,-0.0394760035,0.0585537031,0.4294224083,0.3097966313,-0.0215763934,0.3014404774,-0.0531078354,-0.032504946,-0.3707037866,-0.2006960511,0.3816450536,-0.0846341252,-0.344147265,-0.7334685922,0.3509335518,0.1426517963,-0.3018722236,0.1473779529,0.3631760478,-0.5190951228,0.2938227355,0.040890459,0.8400938511,-0.0927528068,0.3819609284,0.3813296854,0.3095453978,0.7505726814,0.2775804698,0.0987701789,-0.3746117353,-0.2893296182,-0.3023834825,-0.1062904596,0.2692957819,0.3666596711,-0.3419593573,0.3305617869,-0.1324571818,0.2882319987,0.109804593,-0.1822684258,-0.1189226881,-0.2034186572,-0.7257787585,0.0480664074,0.4525593817,-0.0365069993,-0.2351940423,-0.2062583268,-0.1167573556,0.0931266919,-0.2607623637,-0.2176077515,-0.4321043789,0.1444190443,-0.1006666347,0.1561190635,0.2696223259,0.470572263,0.3740429282,-0.3842300177,0.0270237196,-0.1776852012,-0.1497761011,-0.1718022078,-0.0811984837,-0.0599159747,0.475372076,0.0336377285,0.0971907079,0.3608364761,0.0548866503,-0.2822104692,-0.3045469522,0.2037369162,-0.0772633478,-0.2618447244,-0.2907627225,-0.2406383306,0.0460178331,-0.1168738231,0.0699091554,0.2150891572,-0.0369708873,0.164935872,-0.3538252711,0.1511003822,-0.2132048607,0.1253356487,0.2026967406,-0.033026319,0.1077268049,0.1574068367,-0.2605880499,-0.0517194346,-0.0053987964,0.2587834895,-0.4805978239,0.1313628852,0.2005820572,-0.265095681,-0.1276811957,0.5652495623,0.2517309189,0.2128712684,-0.0211022813,-0.1289938837,-0.2973949015,0.3678028584,0.1430948228,0.2439884841,-0.1040255874,0.1322651803,-0.1852490604,-0.1136159152,-0.2206497937,-0.0021180212,-0.0052057137,-0.0675509349,0.4329418242,0.0207916331,0.1636742502,0.2392761111,0.1212816462,0.2636927664,-0.1925227195,-0.132017985,-0.0871845409,0.1703848541,-0.0201830156,-0.0114978524,0.0209324118,-0.0894585252,-0.2565029263,-0.2649773955,0.094342649,0.2512863576,0.1210567057,-0.0448327065,0.399472177,-0.0346534289,-0.1810763329,0.3869203925,0.0380889513,0.4918481112,0.0317566842,0.0474308319,-0.3425094187,0.0545476004,0.0307748616,0.3291054964,-0.0392379351,-0.4100055993,0.6409253478,0.0848434493,-0.1820202768,0.1968024522,0.5251167417,0.2113816291,0.1651661396,0.0732804313,0.1216478497,0.0837516412,-0.333778888,-0.3142723441,0.1178408861,-0.1556552798,0.1714912206,0.5202995539,0.3050468564,0.2484450638,0.1247217357,-0.142963782,0.2160950899,0.198144421,0.0745482072,0.3195811212,0.3900642097,-0.0091482764,0.356150955,0.3804430068,0.5041666627,0.183292985,-0.0932919085,0.0783436,0.0882698596,-0.0172891822,0.6047267318,0.0295796096,0.0089012086,0.4108660221,0.0399434976,-0.0082164165,-0.0557898246,0.0577757806,-0.1989376992,0.0129611809,-0.0629524291,0.2387240976,0.2161091566,-0.1524887234,0.4791820645,-0.1766897142,-0.1289842725,-0.3476141691,-0.1384478658,-0.2974571288,0.4343164563,-0.1308518052,0.565055728,-0.0634816661,0.1978338361,-0.0791861415,0.3597240448,-0.0898838118,-0.0796705261,-0.0958931521,-0.0673906654,0.0075925766,0.2095282674,0.4526987076,0.180991888,-0.1542158425,0.1083929986,-0.4261741936,-0.2367450297,-0.0348737501,0.2622194886,0.1265501231,0.0943511501,0.2722560465,0.0530812331,-0.0310173091,0.0222226586,0.0101432232,0.2748621404,-0.3390365839,0.7011240125,-0.2181453556,-0.2350433767,0.1938218474,0.3255448341,-0.3201076984,0.0481975228,0.0591655709,-0.169170022,0.1492698789,0.1201524213,0.0416862145,-0.0675468519,0.1855736673,0.1969929487,0.4075565636,0.0107357595,-0.2408178449,-0.4748313427,-0.1409271061,0.254796952,-0.2541853189,0.2475463003,-0.0389153212,-0.0236245953,0.091067113,-0.0815633759,-0.3627574146,-0.1309396625,0.214824155,-0.0924031287,-0.2767859101,0.184160322,0.1239567176,-0.1801171154,-0.2226980478,0.0785784051,-0.2537121177,-0.0861013308,-0.1989878565,0.1434233934,-0.1959645003,0.0455567054,0.2694401741,0.222721979,0.180764541,0.2077305764,-0.1693365723,-0.0346198343,-0.3444247544,0.0613573007,0.5801233053,-0.0249673575,0.3234552741,-0.3697853386,-0.2753036916,-0.0232932251,0.2424982786,0.2755203545,-0.193352133,-0.5966461897,0.0874607041,-0.4430664182,-0.0298179649,0.169570297,0.3679272532,-0.1307651103,0.3309237063,0.1300217658,-0.2077973485,0.3422713876,-0.4958495498,-0.2886917889,0.0350957923,-0.2006175965,0.3782958388,-0.1855077744,-0.1591355652,-0.0128236292,0.4850818515,0.1197656393,-0.1062523574,0.1743545085,-0.2196145952,-0.108369641,-0.3893577456,0.0817014873,-0.2865290642,-0.0213867761,0.0420439988,-0.4636683464]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1898","title":"ALT dataset has repeating instances in all splits","comments":"Thanks for reporting. This looks like a very bad issue. I'm looking into it","body":"The [ALT](https:\/\/huggingface.co\/datasets\/alt) dataset has all the same instances within each split :\/\r\nSeemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.\r\n\r\nWould be great if this could be fixed :)\r\n\r\nAdded a snapshot of the contents from `explore-datset` feature, for quick reference.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/33179372\/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)\r\n","comment_length":14,"text":"ALT dataset has repeating instances in all splits\nThe [ALT](https:\/\/huggingface.co\/datasets\/alt) dataset has all the same instances within each split :\/\r\nSeemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.\r\n\r\nWould be great if this could be fixed :)\r\n\r\nAdded a snapshot of the contents from `explore-datset` feature, for quick reference.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/33179372\/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)\r\n\nThanks for reporting. This looks like a very bad issue. I'm looking into it","embeddings":[-0.1631651521,-0.7289551497,-0.0662916452,0.3419412971,0.2512514293,-0.1623471081,0.3814066052,0.1100149453,0.0985512957,0.2535694242,-0.0940162018,0.1431563348,0.0478812791,0.0091341753,-0.097737886,-0.0388371795,0.0357439481,0.2375667989,-0.0193260536,-0.3644630015,-0.0506417342,-0.0380549245,-0.3200644255,-0.1316971481,-0.3216892779,-0.2752453089,-0.105615437,-0.117725797,0.1748771816,-0.2252466083,0.1172012016,0.1168540642,-0.0370866433,0.5170089006,-0.0001172285,0.1029115692,0.4137586653,0.034278553,-0.2644333541,-0.0530044548,-0.4652634561,0.2490053028,0.2411475927,-0.1660094112,0.0642974302,-0.3429347277,-0.2051727772,-0.4327793121,0.3821235895,-0.147645399,0.1793243289,0.384370327,-0.4000789225,-0.2148509622,0.3235217631,0.7567286491,-0.1054928452,-0.3033079803,0.4386866391,0.344730258,-0.0145450383,0.4138398767,0.1362956017,0.3397969306,0.0350655727,0.0755516887,0.0352784358,-0.1396577209,-0.0736130029,0.3006031811,0.4313908815,-0.1073700935,-0.0926095918,-0.1887903064,0.0325331204,0.018392358,0.1699793786,0.1389453411,0.0308139399,0.070566006,0.1658791453,0.3442645967,0.0758970529,-0.1634944677,-0.0434422456,-0.1004548892,-0.1338233501,-0.0279220622,0.0239180233,-0.1545775533,0.4095433056,-0.1576184481,-0.1496071219,-0.2447432131,-0.2803131938,-0.1528839469,-0.2339380085,-0.0777174234,0.1628651768,0.4836684763,0.3364042342,0.0937424302,-0.3747744858,0.0314476788,0.3411440849,0.0411593691,0.2565410435,0.0477833264,0.1245988831,-0.0621359237,-0.2254869342,0.1344414651,-0.1133179739,0.0573506653,-0.1375525296,-0.0618819855,0.2900682092,-0.1169427708,-0.4158878028,-0.058493942,-0.0758559853,-0.1796581149,0.0264941547,0.0976755172,0.3720098734,0.1319071651,-0.2740635872,0.0575530455,-0.0321177617,-0.4481333792,-0.1692385823,-0.1977692246,-0.3170105517,0.0038307086,-0.0233333744,-0.0946159959,0.1927935481,0.4945669174,-0.0005785761,-0.2038546056,0.2380753011,-0.0100878477,0.2227056772,-0.1046459004,-0.0605234876,0.3490628302,0.0974074081,0.0620094165,0.0025338868,0.0714012608,-0.0918679312,0.0758141205,0.0842893794,0.1909156889,-0.1866382509,0.0441555865,0.0393140353,0.233616814,-0.0880465657,0.1142031103,-0.0215788279,0.0872974172,0.135596171,-0.0231523216,0.2440759093,0.4709062874,-0.2902562916,-0.2034665644,-0.5472716093,0.0741846934,0.2167263329,0.5320841074,-0.0206142217,-0.0627961606,-0.3931255937,0.2154884338,0.0309925508,0.2233020365,-0.1881516129,0.0521497205,-0.0194941927,-0.1596402824,0.1417791396,0.0343840197,0.4422266185,-0.0358003415,-0.009782739,0.1807652265,0.2210448682,-0.0719608143,-0.3207636774,-0.0423183106,0.0373281278,-0.0625519454,-0.2631533742,-0.3144912124,0.1078336239,0.0127710644,0.1139290184,0.0262004398,-0.1203728542,-0.0420103967,0.3050195277,0.2051718086,0.3295312226,-0.0502216741,-0.1873775572,-0.0950101018,-0.1545414478,-0.1509712785,0.1986654252,-0.374371618,-0.176349327,0.0437802933,-0.3803083003,-0.0643635988,0.1378361881,0.2440615147,-0.1773788035,0.071812883,0.0432734005,-0.3208737671,0.1759228855,0.0108709941,-0.1343391836,0.4218702316,-0.0616695099,0.0388678797,0.0509284958,0.2172107846,0.2195648253,-0.2327410579,0.0059772898,0.2691893578,0.4275459945,0.26080814,0.1565421075,-0.3347748816,0.4673957527,-0.3614204228,0.1626018733,0.1923951954,0.1030133963,-0.1680822521,-0.1621756107,0.1256401688,0.117727071,0.2895122766,-0.1589815617,0.2823730707,0.2099374384,-0.2082327604,0.2379570007,-0.5056691766,0.1422518641,-0.0053919326,0.0904456377,0.3325628936,-0.1817397922,0.384571135,-0.08412873,0.1144240052,0.213992849,-0.0995822847,-0.4711393714,-0.0479861796,0.0586359203,0.1333133429,0.0915257856,0.2596873641,-0.0478288531,-0.1813358366,0.307540834,-0.1500649005,0.2265015692,0.2347643077,-0.1031617597,0.1229772344,0.148899734,-0.0760293156,-0.5468763113,-0.1584145278,0.3575976789,0.1424584389,-0.1135968119,-0.2120998204,-0.7171415687,-0.0719872266,-0.2348685265,0.1343805492,-0.1488770396,-0.1345389336,0.2325718701,-0.0292579271,-0.0162923131,0.3588981628,0.4100904167,0.5319163799,-0.003900853,0.5492991805,-0.3083837032,-0.1055174619,-0.2065490633,0.071795024,0.4206090569,0.3094550073,0.4547947049,-0.0109557398,-0.0253266301,-0.3409042656,-0.3052348495,0.1471370608,0.0660010129,0.0009968966,0.0953037366,-0.0947012082,-0.031965632,-0.1133880094,0.0222256407,-0.2687592506,-0.3944505751,0.1378272474,-0.1266771108,-0.1416569948,-0.0645620674,-0.1156658754,0.0222649463,-0.1918635815,0.0751186237,-0.01034893,0.0278531127,0.0238510836,-0.199648276,-0.0884224325,-0.2103114277,-0.1789892912,-0.5075007081,0.0303963274,0.0963890329,-0.1631279588,0.0054593021,-0.1585733742,-0.2581519783,-0.1319605112,0.0863713101,-0.4287010729,0.1689084619,-0.5534602404,-0.1583281904,0.264403373,0.1046649739,0.1845236421,-0.0040387949,-0.1697713137,-0.016444033,-0.1946198642,0.1435853541,0.0437066928,0.2003095448,-0.0094003901,-0.1000794694,0.0668483675,0.3522066772,0.5448688865,-0.157846719,0.0531630293,-0.1383739263,0.0794875696,-0.3109108508,-0.2010263354,0.0866555125,0.010001393,-0.2054494768,0.3264258504,0.0029436122,0.0533879101,0.2523938715,0.0954437703,0.0192166008,-0.3053220809,-0.1290235668,-0.2215144634,0.1312076151,0.0036891608,-0.038667284,0.0046512024,-0.0867157653,0.1101562679,0.0918815061,-0.1360677928,-0.182177946,-0.6610919833,0.120233804,-0.154666394,0.2951977849,-0.2116066068,0.2428035885,0.0070957397,-0.0783520415,0.280113548,0.3349652588,0.8090581298,-0.2927730381,0.0482469462,-0.1670667976,0.0511316471,-0.3500305414,-0.1544348896,-0.2348483652,0.09369535,-0.0204944294,0.5390534997,0.0412141606,0.0582307242,0.3225772977,0.0844903812,-0.1670930684,-0.1874313653,0.0937394351,-0.1936037391,-0.0270929914,-0.1144269556,0.1424950957,-0.1233154833,0.1432500035,-0.1767794639,-0.3039346337,-0.0156434029,0.3241500556,0.1054370552,0.5784534812,0.0671518669,0.2436881363,0.4591932893,0.0520401038,0.2709104121,0.6183913946,-0.15324682,0.0039065084,0.1388940513,-0.5353448987,0.1842255443,0.3231850564,0.2063046545,0.3103055358,-0.2497599721,0.0269570872,-0.3228527308,-0.0533182211,0.0169490296,0.1604757309,-0.039546825,-0.3794274032,0.1899178475,0.0555486344,-0.0060271183,0.515032053,0.0639688522,0.1535671055,0.5109503865,-0.0331524462,0.6710172892,0.2344679683,-0.1419735849,0.0938532203,0.0702048168,0.2008218616,0.1896062493,0.0191098675,-0.4727521241,-0.4597455263,0.0550567769,-0.2335486561,0.0776508301,0.009348236,-0.4669885933,-0.3283096254,-0.0263060667,0.2590354979,-0.0220026895,0.1212726831,-0.2179686725,0.0243733022,0.5267267823,0.0787560195,-0.0973056927,0.2348811328,0.1601954699,-0.0441317633,0.2221727669,-0.0728273541,-0.4262633324,-0.0731842443,-0.246781975,0.4182291925,-0.0864363313,-0.4795523882,0.3422911465,-0.2530901432,0.4477908015,0.0289879553,0.0867598951,0.3190314174,0.4859369397,0.1612641364,-0.0752566829,-0.1901926547,0.0352378339,0.3000510931,-0.2625970244,-0.3228490651,0.4309956431,0.2422151715,-0.4912870824,-0.0825457945,0.0161078442,-0.115683794,0.0231804792,0.1835472137,-0.2027212083,-0.3466083705,0.0961791426,-0.2492551804,0.0062454846,0.2501300871,-0.1438393891,-0.019246744,0.0305124372,0.0313749313,0.4452058077,0.1609508544,0.2590193748,0.204231143,-0.1818805486,-0.4756581783,0.7170437574,0.1750987023,-0.651491046,0.4187364876,-0.5128411055,-0.0500036441,0.0072623328,0.0484087728,0.0050200718,0.145466134,-0.3754453659,-0.0699497908,0.0602704734,0.1139782444,-0.3955006897,0.3749454916,-0.1546097398,0.1686704755,-0.0028410743,0.1781879961,-0.2679579258,0.3152910471,0.1381236166,0.2182859182,-0.6414310336,-0.3956805766,0.0825460032,-0.2133760601,0.0399491079,0.1047915667,-0.0208281502,-0.2153434753,-0.0912228972,0.1373777837,-0.2316706181,0.0588125698,-0.0791252702,-0.0007565921,0.0493079424,-0.2401863039,0.1165437177,0.1567592919,0.0096034184,0.0326312557,0.2198516726,-0.3900797665,-0.0929195881,0.1003969908,0.1002061069,-0.0578012541,0.2769751847,0.0283478256,0.0412748605,-0.1130092517,-0.2849675715,0.1624340564,-0.2829884589,0.1025912166,0.0242857859,-0.146208927,-0.199680239,0.2354584634,0.3317643404,0.0857051387,-0.115798004,0.0578809641,0.0400463529,0.1937897205,-0.4071137607,-0.2291003615,0.3625342846,0.3321420252,0.1062600687,0.216132462,0.2679526508,-0.1967030168,-0.315030843,0.0680386424,0.3958882987,-0.3142405152,0.7396067977,0.4856785238,0.0144475726,-0.2123617977,0.323135078,0.1159296036,0.0158464909,0.2964069843,0.1458117068,0.2606302202,0.3932551742,-0.0904434174,0.4815217257,-0.2485916913,0.0441295579,0.0699524507,0.0383013524,0.2308662236,0.3229465783,0.2821926773,-0.2512858212,-0.2832986712,-0.2358257324,-0.3964208364,-0.156490162,0.1373060942,-0.1924154609,-0.1204698682,-0.3637733757,-0.156526342,-0.023391135,-0.0308167469,0.3355796337,0.0241651088,-0.1057668775,-0.4859954119,-0.1690469533,0.274187386,0.2819873095,-0.3544804752,0.0869446322,0.2014077902,-0.0435846783,0.2817973197,0.5439087152,0.2516171336,0.1218982637,-0.1708224416,-0.0479084738,-0.0584518872,-0.1852670461,-0.0509226918,-0.1371868253,-0.1373125762,-0.2188018113,0.2876554728,0.0765415207,0.0068388665,-0.0936339125,0.4312238693,0.0096748201,-0.2501610219,0.1148549169,-0.1878355145,-0.0408800915,-0.0570583269,0.1177902222,-0.5034052134,-0.3102568388,0.454906255,-0.4434607923,0.0052168383,-0.1944149286,0.0678190291,-0.3350276351,0.2533739805,0.2653913498,-0.4544321299,-0.0848062485,-0.3181562126,-0.3731744587,0.0911652371,-0.2540838122,-0.0742387995,0.4235868454,0.2457554042,-0.0157016832,0.0326420888,0.4061009586,0.147262454,-0.4325443506,0.1995514035,-0.5272194147,0.0336528644,-0.0036390629,0.0868428946,-0.0465877056,-0.0638393164,0.4602612257,0.1479310244,0.002493931,0.119927384,-0.0704661161,0.0786214471,-0.070617348,0.0431030467,0.1967476755,0.0051398203,0.0158445444,0.2411822677,-0.2310293019,-0.4597846568,-0.2495854944,0.2447735071,0.5059544444,0.165153563,-0.2029283494,-0.3836670518,-0.0870936066,-0.1329806298,0.1877358705,0.4171221554,-0.2784481347,0.2841784358,-0.1943355948,0.1377507895,0.1491039544,0.2222228348,-0.0418184139,-0.2150107473,0.1695346832,-0.6131685972,0.3822545409,-0.0084805693,-0.2326036841,-0.2491041124,0.2279046327,0.3132239282,0.0803362355,-0.3357806504,-0.1982931346,0.4098865688,-0.2011238933,0.0731699541,0.0707581639,0.119523935,-0.1865908206,-0.1204974726,0.1009377763,0.4566224813,-0.1776718646,0.3259519041,-0.1516489238]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1898","title":"ALT dataset has repeating instances in all splits","comments":"I just merged a fix, we'll do a patch release soon. Thanks again for reporting, and sorry for the inconvenience.\r\nIn the meantime you can load `ALT` using `datasets` from the master branch","body":"The [ALT](https:\/\/huggingface.co\/datasets\/alt) dataset has all the same instances within each split :\/\r\nSeemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.\r\n\r\nWould be great if this could be fixed :)\r\n\r\nAdded a snapshot of the contents from `explore-datset` feature, for quick reference.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/33179372\/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)\r\n","comment_length":33,"text":"ALT dataset has repeating instances in all splits\nThe [ALT](https:\/\/huggingface.co\/datasets\/alt) dataset has all the same instances within each split :\/\r\nSeemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.\r\n\r\nWould be great if this could be fixed :)\r\n\r\nAdded a snapshot of the contents from `explore-datset` feature, for quick reference.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/33179372\/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)\r\n\nI just merged a fix, we'll do a patch release soon. Thanks again for reporting, and sorry for the inconvenience.\r\nIn the meantime you can load `ALT` using `datasets` from the master branch","embeddings":[-0.2530450225,-0.7312532663,-0.0555956438,0.4320774376,0.3174535334,-0.1736770868,0.3799054027,0.1424697042,0.2794243693,0.331989646,-0.1614301652,0.0980525613,0.0777056143,-0.0084827486,-0.0531923398,-0.0700326785,-0.0420137346,0.2138660252,-0.001591517,-0.2287296951,-0.136103645,0.1931893826,-0.3067333102,0.0349796563,-0.1967672557,-0.2158623934,-0.0085895853,0.0501792505,0.0790567994,-0.278033793,0.0863051564,0.1110689417,0.0143353166,0.4517409205,-0.0001077245,0.1526554823,0.4967345595,0.0332457721,-0.2704990506,-0.2112117112,-0.3161489069,0.2112973183,0.2924031019,-0.1687503904,-0.0563588105,-0.2315426916,-0.1243325248,-0.4026524723,0.3349613249,-0.0742322877,0.2620726824,0.4766892195,-0.3479264975,-0.4059582055,0.438123703,0.5822082758,-0.1347050071,-0.1824561656,0.3820548654,0.4519361854,0.0182801615,0.3951468766,0.1519043744,0.2581334412,0.1004771739,0.1850468218,0.1533161849,-0.1135968566,0.0046806978,0.267268002,0.3437399864,-0.0407965966,-0.0840745196,-0.208226651,0.0375719853,0.0121965185,0.242644608,0.0158100352,0.020386735,0.1143068373,0.2770204544,0.2895072699,0.0548395067,-0.0553695373,0.0077201026,-0.0340844393,-0.1891286373,0.0874743387,0.1575710177,-0.1861014962,0.3643367887,-0.0746586546,-0.1419699788,-0.202799961,-0.3039722443,-0.0242452417,-0.1317781806,-0.1549829096,0.1851761192,0.3924604952,0.3807055652,0.2239830196,-0.3758049607,0.0164204333,0.2636727095,0.0733937696,0.303794086,-0.0638524815,0.1854224503,0.078543663,-0.1949545741,0.0184880849,-0.2299306393,-0.0099541424,-0.2303521335,-0.0250952318,0.248890236,-0.119489044,-0.4183261395,-0.116393365,0.1127997339,-0.1395734698,0.0736995637,0.3457525373,0.3099732399,0.1333232224,-0.0907328874,0.0623501018,-0.1016265303,-0.4191182554,-0.2110074311,-0.2572085261,-0.3241529167,-0.0920962244,-0.0299138203,-0.114004679,0.268192023,0.2965404391,0.0488082208,-0.1254488081,0.2154405564,0.0645246506,0.1204699874,-0.0578569062,0.1036735997,0.3334433138,0.153557539,-0.0137160979,-0.0288760252,0.0631649271,-0.111020796,-0.0017046058,0.1181939021,0.2778229713,-0.0948350877,0.1378462166,0.0151464585,0.2947776318,-0.1778623164,0.0043671816,-0.1145090386,0.0638617799,0.0688562542,-0.0965424031,0.254160434,0.4378035069,-0.228737399,-0.3377067447,-0.5115407705,-0.0721741095,0.1281719357,0.5561010242,-0.1574354917,-0.1233198792,-0.399990648,0.2835876048,0.1919445693,0.2184081227,-0.2081331313,0.1348673403,-0.1094422191,-0.0671303421,0.0755431131,0.1349571198,0.3912534416,0.008007139,0.0901445225,0.2794891,0.287633121,-0.0885440782,-0.2506302297,-0.0650307685,0.0467608906,-0.041387476,-0.2071636319,-0.32281515,0.1217947751,0.0372686833,0.1147334799,-0.0210297983,-0.0985308513,0.0281672217,0.2101825625,0.2042817771,0.3248484135,-0.0377073213,-0.2207812667,0.0477528796,-0.1736790687,-0.147826463,0.2951383889,-0.418310672,-0.2752805054,0.1193790808,-0.3506122231,-0.0463987626,0.1903411895,0.2829102576,-0.1916191727,0.0162029378,-0.0466967709,-0.2341548949,0.1645257622,0.0080304416,-0.1809080094,0.2991646528,-0.1965728998,0.0187053531,0.0256757382,0.1080784872,0.1730462611,-0.2686104476,-0.0383830518,0.4057914615,0.3260442615,0.1635503173,0.0293142907,-0.4810571373,0.4122965932,-0.2348932475,0.1248606667,0.2347392589,0.1278838664,-0.2357152849,-0.1550873518,0.1593185216,0.1920385361,0.2863152921,-0.0333648995,0.208559826,0.3136180639,-0.1635531336,0.2406594902,-0.5274125338,0.1388350278,-0.0694969669,0.0954828709,0.1965302825,-0.2810149491,0.3699359298,0.1038458869,0.10755077,0.2117594332,0.0147755118,-0.5166527629,-0.0949606746,-0.0444386676,0.0235481486,0.1444617659,0.3443719447,0.0418831855,-0.1660429984,0.3011958897,-0.1309051663,0.3186788857,0.1758503616,-0.1268957257,0.092217356,0.1232438982,-0.1161591187,-0.5793968439,-0.1676859856,0.3131718636,0.063004531,-0.2004716843,-0.22012043,-0.5879811049,-0.013966009,-0.1368472576,0.010807205,-0.1271443069,-0.1715045869,0.2827455997,0.0558801629,0.0002116532,0.366748035,0.3949826956,0.4391366541,-0.0003229588,0.442353636,-0.3167379797,-0.1895712167,-0.3355871141,0.1189157069,0.4297174513,0.3283501565,0.4452002645,-0.0780442134,-0.021039784,-0.2582357824,-0.3862937689,0.1428810507,0.0395937078,-0.0219456907,0.1286536753,-0.1326865405,0.0025743707,-0.1702553481,0.1888428479,-0.4004678428,-0.4218910933,0.1067008898,-0.1009612754,-0.2328895628,-0.1279150695,-0.296223253,0.0629430562,-0.3463539779,0.1321719587,0.0558892526,0.0171966497,0.0587159172,-0.1139376312,-0.0360047854,-0.0893377736,-0.1253001839,-0.5492932796,0.0673271716,0.0928752869,-0.2612448931,-0.133163482,-0.0414444283,-0.2199798375,-0.1416226327,0.032470651,-0.4009286761,0.187947154,-0.4680107534,-0.0788509622,0.1960986406,0.0292475149,0.2848639786,-0.0054312022,-0.2143379301,-0.0154975187,-0.296048075,0.1560645252,0.0158260241,0.1873741597,-0.0449537896,0.0046860641,0.0418864228,0.3934039474,0.318161577,-0.1329952329,0.0471848287,-0.1528198421,0.1869278997,-0.4131629765,-0.1799942404,0.0803353935,0.0075400821,-0.1402645111,0.3866060078,0.0220710523,-0.0186825376,0.1082531586,0.0504925177,0.0193109419,-0.3562465012,-0.1019512117,-0.100329496,-0.0155535294,-0.0355559438,0.0361449234,-0.1161257848,-0.0161619112,0.1786687374,0.115796417,-0.2331679612,-0.1852466911,-0.629769206,0.1757918,-0.2929056287,0.3453254998,-0.2199638784,0.1585962176,0.0163655337,-0.2085305452,0.3482231796,0.1503788531,0.7905905247,-0.2985009849,0.0180559754,-0.0152582871,0.0137595441,-0.5073357821,-0.1102741361,-0.2534873784,0.0568084046,0.0848111138,0.59610641,-0.0769744813,-0.0178145785,0.2569271028,0.0763158575,-0.2049290985,-0.1959679425,-0.0262698587,-0.1808037907,-0.2594660819,-0.2019487619,0.066461578,-0.0534828864,0.1716034859,-0.1570577472,-0.3182406127,-0.0022011907,0.1731117815,0.308816582,0.522115767,0.0686081648,0.1972514391,0.4771078825,0.1118460894,0.1605866253,0.7080816031,-0.1246948838,-0.0885445625,0.179395631,-0.4542679787,0.0541717149,0.310457021,0.2080195099,0.2227960676,-0.194497183,0.0548258349,-0.1967514604,0.1551366597,0.1092845798,0.0995221287,0.0047365646,-0.337159127,0.154225111,-0.1057449058,-0.0794187561,0.462381959,0.0169350319,0.0483624339,0.5934584737,-0.1379651278,0.7147182226,0.0982610881,-0.0499804765,0.139275983,0.0757338479,0.1597912312,0.21658732,0.0462790392,-0.5335631967,-0.4241665304,0.1115132123,-0.1202279478,0.0085769901,-0.0140779596,-0.4247624278,-0.2648171186,0.0503543131,0.3281959295,-0.0736606121,0.1685643196,-0.1620366126,-0.0568720065,0.4204368591,0.17028898,-0.0780799985,0.218998149,0.1155496091,-0.0110787423,0.1699695885,-0.1600702852,-0.3022381067,0.0105894813,-0.3480478525,0.4214714468,-0.1170589924,-0.5092543364,0.2436768264,-0.2589383125,0.3577935994,0.0583618954,0.0356465839,0.3490626216,0.318775326,0.0343276747,-0.1135340929,-0.17302531,-0.0191466566,0.2346480489,-0.285566479,-0.2195286304,0.3488925099,0.1695678085,-0.3726007044,0.0135493744,0.0414603949,-0.1194157973,-0.2098167539,0.1362935454,-0.1320102215,-0.3592993021,0.1538861543,-0.3178063333,-0.0152981849,0.0411127619,0.0042615877,-0.0640338361,0.0624863878,0.0726291612,0.3321679235,0.071181193,0.2105039656,0.2428441644,-0.1890864223,-0.4889020622,0.5547899008,0.2250992656,-0.6193469763,0.370090574,-0.3495398164,-0.1074079201,0.0922145024,0.1170544177,0.1182190329,0.088555418,-0.332921207,-0.1235658005,-0.0467914306,0.139230445,-0.3650645316,0.3558060527,-0.0435734205,0.1386948675,-0.0411707871,0.1824305505,-0.3591291904,0.294811815,0.1066087335,0.1958191991,-0.5335167646,-0.354647547,0.1680217534,-0.1661370695,0.1139765829,0.0865002871,-0.0731942579,-0.270996213,-0.1109666377,0.0990340337,-0.2938289344,0.1494212151,-0.0456151329,-0.0791164935,0.1114271954,-0.3093267083,0.2880663276,0.1603247821,-0.0819973797,0.0103599206,0.1283068508,-0.2510236502,-0.1187815145,0.2032217979,0.0138759511,0.040772073,0.2367196977,0.0036928044,-0.0783742443,-0.1281886697,-0.1712080985,0.1973726749,-0.1173798442,0.1716750562,0.0816531405,-0.1096390709,-0.1526998878,0.1485708654,0.3307651281,0.0729482323,-0.0844853818,0.0714306086,0.0353539884,0.2475076765,-0.4756090045,-0.2290028036,0.3186691403,0.1778696626,0.1701244712,0.2099043876,0.1794622838,-0.1407875866,-0.2270637602,0.1378560662,0.4304172993,-0.1530753523,0.4993042648,0.4427487552,0.0093532586,-0.2456501275,0.3271171749,0.0901493058,0.139837712,0.3966960609,0.000159427,0.2757720649,0.3659015,-0.0076311748,0.3976854086,-0.3081564903,-0.0114769423,0.0781335011,0.053365875,0.136476621,0.3035897017,0.2825964391,-0.3126272857,-0.2198749483,-0.2563137412,-0.3658160865,-0.1523784995,0.0193885136,-0.3127085567,-0.0757380947,-0.4050969481,-0.1092406437,-0.0566347577,0.0324506126,0.3366054296,0.0102391262,-0.0523800366,-0.4539518654,-0.1849164069,0.1961039156,0.1046566814,-0.3900808096,0.1740124077,0.2441935092,-0.0297510438,0.3311133385,0.4414886236,0.2335396707,0.1364921182,-0.2641112804,-0.0189036112,-0.0341617428,-0.2060284764,-0.155176267,-0.1845023334,-0.0049265292,-0.0138345379,0.3606172502,0.1489454359,-0.1105716527,-0.0897165686,0.3804370463,0.0569639876,-0.191024974,0.0530697927,-0.264929682,-0.0524504073,-0.1745592952,0.077349931,-0.5230470896,-0.3014148772,0.3586508632,-0.406314373,0.0645207986,-0.1936033219,0.1006172001,-0.2720007598,0.2932375968,0.1757775247,-0.4867835045,-0.0614389591,-0.2895151079,-0.4792466462,0.2440948337,-0.1913367063,-0.1096368581,0.3604549766,0.2841494679,0.0968423262,-0.0035263628,0.4418776035,0.2153237015,-0.2735732794,0.1427687407,-0.5009114146,-0.009266343,0.0609094985,0.0025331264,-0.0517024025,-0.1849664748,0.3137893379,0.0846940055,0.0423715487,0.0779922977,-0.0587282255,0.1510836929,-0.1502036601,0.1557246298,0.1671287566,0.0088142296,-0.1092448607,0.261847198,-0.1659757942,-0.5043036938,-0.1913361698,0.1892676502,0.5310876966,0.2051565498,-0.1579915434,-0.3544663191,-0.1397819221,-0.0417206548,0.2464240044,0.3472797871,-0.310659796,0.3282321692,-0.1817416549,0.0989636406,0.1563435942,0.2996597588,-0.0352362394,-0.2530741096,0.1073725745,-0.5553367734,0.4459258914,0.0170266721,-0.3018951416,-0.1467086375,0.2151526064,0.3208595812,0.0979312807,-0.3039985895,-0.1042323783,0.3273189664,-0.1401027888,-0.06310004,0.1081161797,0.0052376566,-0.1605123132,-0.0747159272,0.1002990007,0.4923512042,-0.2181967646,0.2114478648,-0.2496324331]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1898","title":"ALT dataset has repeating instances in all splits","comments":"Thanks!!! works perfectly in the bleading edge master version","body":"The [ALT](https:\/\/huggingface.co\/datasets\/alt) dataset has all the same instances within each split :\/\r\nSeemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.\r\n\r\nWould be great if this could be fixed :)\r\n\r\nAdded a snapshot of the contents from `explore-datset` feature, for quick reference.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/33179372\/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)\r\n","comment_length":9,"text":"ALT dataset has repeating instances in all splits\nThe [ALT](https:\/\/huggingface.co\/datasets\/alt) dataset has all the same instances within each split :\/\r\nSeemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.\r\n\r\nWould be great if this could be fixed :)\r\n\r\nAdded a snapshot of the contents from `explore-datset` feature, for quick reference.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/33179372\/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)\r\n\nThanks!!! works perfectly in the bleading edge master version","embeddings":[-0.1432338953,-0.7411319613,-0.0646167919,0.3409085274,0.1455446631,-0.2264077514,0.3666744828,0.0980923846,0.0733458921,0.3188048601,-0.1350822896,0.1315943897,-0.0026312051,0.1397401541,-0.1287334263,-0.0102504585,0.0539634712,0.2943821847,0.1015710756,-0.3700869679,-0.0299461782,0.0056629297,-0.2776165903,-0.1522094309,-0.410451442,-0.1317205578,-0.1005455032,-0.0302187074,0.1420179009,-0.1902011931,0.0918512642,0.1052669063,0.0342212208,0.4780634642,-0.0001190875,0.0802431926,0.366317451,-0.0022656913,-0.3532449901,-0.1287859082,-0.3791440129,0.2279408723,0.2505927086,-0.1519715339,-0.0237446055,-0.263797611,-0.1902197897,-0.3847571909,0.3048505187,-0.1322153956,0.1603767574,0.4273771346,-0.3755769432,-0.2642551064,0.4237337112,0.6270629764,-0.141246587,-0.2901128232,0.4227785766,0.4091949165,-0.0623730198,0.45213449,0.1751375794,0.3899441957,0.1907288581,0.1209822595,0.1085247397,-0.0706855953,-0.0077864509,0.3318018615,0.30782938,-0.0692772493,-0.1130189747,-0.2167091519,-0.0367142223,-0.0053766388,0.2613369823,0.1937548816,-0.012104,0.0723597184,0.0865674093,0.3616466522,0.1277176738,-0.1421602517,-0.0444375537,0.0089976871,-0.1349588782,-0.0251098052,0.1040217802,-0.1090827733,0.4190514386,-0.1374999881,-0.1357961744,-0.3034139574,-0.3474066257,-0.0815404654,-0.1618430018,-0.0440053903,0.136575833,0.6368730664,0.3755488396,0.1411592364,-0.3979195058,0.0704346523,0.2938709855,0.0634369552,0.3264969885,0.0657068938,0.17612423,0.0111243855,-0.1428340524,0.1315774173,-0.1427951455,0.0417957045,-0.1464542896,-0.1196214706,0.2442979962,-0.147298038,-0.5001684427,0.0030808195,-0.0291624088,-0.180323422,-0.0274770614,0.1366508156,0.3374715745,-0.0340258293,-0.1678562462,0.10998974,-0.0256194267,-0.4465983212,-0.1868380606,-0.2938410938,-0.3700720668,-0.0788435116,-0.0006995763,-0.1391824335,0.3710677326,0.4498294592,0.045439478,-0.088487111,0.1333485544,0.1015287191,0.234340772,-0.1114969924,-0.0746894628,0.2642549574,0.0526960976,-0.051366087,0.0141583411,0.0612602606,-0.0019852275,0.0460162312,0.0292076785,0.1909938306,-0.2174370736,0.061039988,0.0363732092,0.2263648957,-0.1124545708,0.0346033014,0.026346378,0.0665886253,0.1646341681,-0.0697525144,0.3459428251,0.3631860018,-0.259534508,-0.2581822574,-0.4387571514,0.0042496375,0.2390059829,0.5514923334,-0.0146233533,-0.2183582634,-0.3362366855,0.1625882238,0.016921673,0.3442063928,-0.2639377117,0.0345235541,-0.0550046191,-0.1096635908,0.1319912374,0.0702938512,0.4521463811,0.0467483215,0.0476264395,0.1865565926,0.1972427815,-0.0659856573,-0.3596084714,-0.0473439395,0.1003187001,-0.0363798104,-0.2923872769,-0.394471705,0.0990745947,0.1154211462,0.1360983551,-0.008111042,-0.1151730791,-0.1889947653,0.226318419,0.1357957125,0.2831089795,-0.0682637393,-0.2537719011,-0.0073366943,-0.1389950663,-0.0699323937,0.1837476939,-0.3936514258,-0.1771345586,0.0453827828,-0.4503329396,-0.25984779,0.132233724,0.309240222,-0.1696923822,-0.0273769647,0.0332893059,-0.4133384526,0.1222707555,0.0324669369,-0.0451002382,0.3809635043,-0.1358428597,-0.0088419765,0.0353596695,0.2302884161,0.1116384864,-0.3148207664,-0.044265423,0.2403662056,0.3326075971,0.1774942875,0.1210042313,-0.2514394522,0.5117627382,-0.3314775229,0.1781380624,0.2846139371,0.0725368485,-0.2682233155,-0.312007159,0.1659219116,0.1567597836,0.3052766025,-0.1241572425,0.168108657,0.1871665418,-0.224597767,0.2686387599,-0.5152732134,0.2406061441,-0.0724184811,0.0495937876,0.2547550499,-0.2705640793,0.4221197963,0.0255390741,0.1047367454,0.219971627,-0.1135721505,-0.4306195974,-0.0885522515,0.1558296382,0.0773419142,0.0856610611,0.2558582127,0.0455654897,-0.0490546301,0.2864246666,-0.2281389683,0.2358247042,0.1421984732,-0.2308425158,0.107758157,0.1702852249,-0.1101902351,-0.5562086701,-0.0963869244,0.3570446372,0.1080325767,-0.0583282299,-0.2081913948,-0.7342938185,-0.0807880089,-0.1901641488,0.1073013842,-0.1430390924,-0.1708264053,0.2784512341,0.0807668418,-0.076060228,0.3710619211,0.31013152,0.4586260617,-0.0210415125,0.4584572315,-0.2633452415,-0.1465984583,-0.2611463368,0.074670054,0.425305903,0.3340824544,0.4629066288,-0.1232818738,-0.0184792113,-0.4074267745,-0.3877984285,0.166063264,0.0857153311,-0.0123913437,0.1521285027,-0.1569224,-0.0698596388,-0.213842243,0.1588779837,-0.2804556191,-0.4500385523,0.0999519303,-0.0196362343,-0.2104210407,-0.1022665352,-0.1495605856,0.072674714,-0.224817872,0.0078073279,-0.0280408431,0.0970501155,-0.0792023987,-0.1314918101,-0.0223323926,-0.2143274248,-0.1712163389,-0.5031940937,0.0106487256,0.0699591935,-0.2100305259,-0.0355332568,-0.1238685697,-0.2258290499,-0.0865992308,0.1188603565,-0.409240365,0.1831256002,-0.5295407772,-0.0226387493,0.3959909081,0.1157881692,0.1802433878,0.0326820463,-0.146506384,-0.0099652465,-0.222399205,0.1394902021,-0.0434854329,0.3012242019,0.027503252,-0.0385654643,0.0969520137,0.4696353078,0.5537309647,-0.1567809284,0.1234932393,-0.0730413571,0.1262533963,-0.3891055584,-0.215358153,0.2263391465,0.005026157,-0.1447591335,0.3670523763,-0.0008044109,-0.0533599034,0.2010237575,0.0002366802,0.0753648952,-0.3477862477,-0.1027078331,-0.1860586107,0.0486810356,-0.0357455164,0.0121148722,0.0607815273,0.0059154835,0.0886048004,0.1137845144,-0.1301876307,-0.1528149545,-0.547406137,0.1143865958,-0.1796865314,0.3247638047,-0.2216786742,0.2033331245,0.0272792075,-0.141910091,0.2780212164,0.2891900539,0.745103538,-0.3100730777,0.0858501047,-0.1492773145,-0.022874793,-0.3589737415,-0.1262821108,-0.2036880404,0.0719319955,0.1100452244,0.4793020189,-0.1261904389,0.0159279648,0.4004838169,0.0757742599,-0.1206655353,-0.0810493976,0.1170996875,-0.1580604017,-0.1536934078,-0.1200234592,0.0929979309,-0.0356179699,0.1418088675,-0.1358755082,-0.3257707059,0.0175010096,0.3060266972,0.1948249489,0.4759511948,0.0290990621,0.2479220033,0.5123000145,0.1617462933,0.164609611,0.6935783029,-0.182485953,-0.1117542684,0.1219561696,-0.4966381788,0.1296454817,0.3607435524,0.2685115635,0.2468133122,-0.1590206027,0.0047961036,-0.2270912528,0.0316256322,-0.0360769369,0.2277975827,0.0228127185,-0.3826573193,0.2058195323,-0.0258823242,-0.018087808,0.4762520492,0.0950901732,0.0914964825,0.5054714084,-0.0756395161,0.7240462303,0.3841404915,-0.0917470008,0.1096851006,0.0896741524,0.1416185647,0.2551501393,0.0143851899,-0.4136134386,-0.5025152564,0.0455847047,-0.1818876863,-0.0183293093,-0.0199754164,-0.3928911686,-0.249718681,-0.0488870516,0.2686213553,0.0486051179,0.1471062303,-0.2318058312,0.0552381128,0.5516294837,0.0726266429,-0.0726470202,0.2775883377,0.1465629935,-0.0721459985,0.1444833428,-0.1146306098,-0.4348649979,0.0019705931,-0.3141396344,0.3636945784,-0.218091011,-0.6139375567,0.3599208295,-0.3392064869,0.4624397457,-0.0276405215,0.0742625818,0.2692219317,0.3994348943,0.1321975738,-0.0938873515,-0.2020085156,0.033493042,0.2230104357,-0.3255194128,-0.2477235943,0.4298632443,0.133787781,-0.4374690056,-0.0552570634,-0.0795704946,-0.0378453285,-0.0562808029,0.1894203722,-0.2205643207,-0.3917036653,0.0888563842,-0.2704361379,0.0167239606,0.0789595246,-0.0368744321,0.0069031501,-0.0105611822,0.0693240985,0.3529342413,0.1684878767,0.2861364484,0.204136759,-0.1923140585,-0.4003774226,0.5768387318,0.2435779124,-0.6102576852,0.4501030445,-0.5535581708,-0.1550338864,0.0502788723,0.2308396846,0.1312877536,0.2005809993,-0.3268123567,-0.0768531859,0.0508580618,0.1326956302,-0.3997341394,0.4009024799,-0.2076987028,0.2257539183,0.0641600937,0.15808855,-0.2686696351,0.2481979281,0.1832541823,0.2988264561,-0.5795439482,-0.4352315664,0.1162197664,-0.2023260444,-0.0033534898,0.1071764901,-0.0518312603,-0.2113156915,-0.1382948458,0.1563345641,-0.1482896358,0.0446572751,-0.0862434953,0.0554771498,0.0520194322,-0.2768657804,0.2777357697,0.0238890648,-0.0001388249,0.0488247052,0.1055317223,-0.3861584663,-0.0345458128,0.0048468029,0.1662408262,-0.017636627,0.1819462329,-0.0187927354,0.0150501253,-0.131857872,-0.1298942566,0.1863972545,-0.3092861474,0.1392334104,0.0335520729,-0.1612041146,-0.1995785087,0.1763872355,0.3167331517,0.0603254214,-0.017347835,0.0395080037,0.0687590986,0.1906350255,-0.4308918118,-0.1999690682,0.4343127012,0.3829276264,0.1917889714,0.1994040012,0.2319154888,-0.1848633587,-0.2726876736,0.0487705767,0.3629497588,-0.4285967052,0.6921288967,0.5360705256,-0.0367042981,-0.1325149387,0.4052261412,0.1044078767,0.0335037373,0.2952120602,0.1058389768,0.3496291041,0.4809883833,-0.0674118251,0.5377588868,-0.1930002421,-0.1409263909,0.0499363244,0.1003704518,0.3057473302,0.2853627503,0.1967334449,-0.2972610891,-0.2367017269,-0.2788190544,-0.2865012288,-0.1237248704,0.061144501,-0.264521718,-0.1614059806,-0.3372541964,-0.1373053044,-0.0693401247,-0.042624522,0.417216748,-0.0188650116,-0.1319416761,-0.4661605358,-0.1289836615,0.2112163603,0.2459658533,-0.3809912205,0.0867763385,0.3292019367,-0.0905091092,0.2369182557,0.5934254527,0.2728251815,0.2029117942,-0.2005637288,-0.060925521,-0.0956432074,-0.1770843714,-0.1805595458,-0.096007973,-0.1068999171,-0.2837380767,0.2961031497,0.0732611343,-0.0483254306,-0.0728036687,0.396130532,0.0214655846,-0.2630032301,0.099733822,-0.2442317605,-0.0287479497,-0.0358146466,0.069875516,-0.4617167413,-0.200421229,0.3167520463,-0.3267872632,-0.0013620561,-0.1962488294,0.0554501936,-0.3099507987,0.2671869695,0.202949971,-0.3237589002,-0.1261396855,-0.3141415119,-0.4640206099,0.0807216614,-0.1672983468,-0.0955172405,0.4626304507,0.3064720929,0.0208574478,0.0450285822,0.402905494,0.1984961778,-0.3592259884,0.2187169045,-0.4093839824,0.0220743679,0.0038096695,0.0131476792,-0.046780251,-0.1415130049,0.4371832907,0.1419619769,0.0024409944,0.0963439718,0.023183519,0.1140695959,-0.1196811125,0.0738984719,0.1995916665,0.0847052783,0.0286647864,0.2310502231,-0.3371454179,-0.3893413544,-0.3118960857,0.2671481371,0.5295559168,0.3146478236,-0.1881343573,-0.3344070017,-0.1446192414,0.1001897007,0.2120617777,0.3289917707,-0.2096469402,0.2989166677,-0.2176069617,0.1033018604,0.1485135555,0.2747234702,-0.0902638137,-0.3013493419,0.1231502518,-0.6432666183,0.3550566435,0.0008719725,-0.1998066455,-0.226019159,0.1696320325,0.4530137479,0.0435891412,-0.4025931358,-0.1951213777,0.3077701926,-0.2401515841,0.0207517315,0.0597915649,0.1902772486,-0.2086837143,-0.1027081311,0.177965343,0.4185393751,-0.3055689633,0.2569785118,-0.2193935215]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1898","title":"ALT dataset has repeating instances in all splits","comments":"Closed by #1899","body":"The [ALT](https:\/\/huggingface.co\/datasets\/alt) dataset has all the same instances within each split :\/\r\nSeemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.\r\n\r\nWould be great if this could be fixed :)\r\n\r\nAdded a snapshot of the contents from `explore-datset` feature, for quick reference.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/33179372\/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)\r\n","comment_length":3,"text":"ALT dataset has repeating instances in all splits\nThe [ALT](https:\/\/huggingface.co\/datasets\/alt) dataset has all the same instances within each split :\/\r\nSeemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.\r\n\r\nWould be great if this could be fixed :)\r\n\r\nAdded a snapshot of the contents from `explore-datset` feature, for quick reference.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/33179372\/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)\r\n\nClosed by #1899","embeddings":[-0.202688545,-0.7161698341,-0.101162903,0.3325876594,0.2076653242,-0.1354822069,0.3829605281,0.1558819115,0.0895967409,0.2992260754,-0.1059225202,0.0893386006,0.0451681092,-0.030216489,-0.0956859738,-0.0314799994,0.0361991264,0.2305938452,-0.0053989575,-0.3822104931,-0.0660940409,-0.0205500014,-0.2966687679,-0.146784544,-0.3256379962,-0.2459123433,-0.1061831489,-0.0799544901,0.1280632466,-0.2083444446,0.1071116403,0.1368983537,-0.062758103,0.5199160576,-0.0001150815,0.1047902182,0.4041944146,0.0098844562,-0.2859147489,-0.0507381633,-0.4377415776,0.2116918415,0.1915025711,-0.1825297326,0.0611280538,-0.3465020061,-0.133185178,-0.4057281911,0.320894748,-0.1030381843,0.2110275328,0.3777181804,-0.3906783164,-0.2974714041,0.3689129055,0.7189519405,-0.1204951704,-0.2425626218,0.4264728725,0.3656361103,-0.048348587,0.4043780267,0.0990466103,0.2929365635,0.0306628421,0.0903539583,0.0482151881,-0.1325431168,-0.0158582367,0.3224765062,0.3690006137,-0.0873706192,-0.0939566121,-0.190826565,-0.0460108966,0.0403057896,0.1617869288,0.1478199214,0.0020003095,0.1097778007,0.1569439769,0.3768596947,0.0879084542,-0.1166817397,-0.0680803582,-0.0620149635,-0.1265340745,-0.0112406248,0.0484648421,-0.1395207494,0.424567908,-0.1327881962,-0.0803248063,-0.2956769764,-0.2373849303,-0.1147551313,-0.1771525741,-0.1264804453,0.1922277063,0.4989924431,0.3251841068,0.1127131358,-0.3637488186,0.0292520653,0.2612804472,0.062511839,0.2181607932,0.0550864041,0.1055111811,-0.034748245,-0.2595981956,0.1553868502,-0.1289288402,0.0457668044,-0.124827221,-0.0964298621,0.238551572,-0.1079389825,-0.4295912087,-0.0389150605,-0.0400209539,-0.1516146809,0.0104882075,0.1623571217,0.3267984986,0.0987722948,-0.2626851797,0.0903422087,-0.0180046279,-0.4428600669,-0.1965819448,-0.168050155,-0.3000223041,-0.0341912769,-0.01092395,-0.0853872523,0.2645308077,0.4280027151,-0.0165610034,-0.1535366774,0.2240602672,0.0062306738,0.2648015916,-0.097827889,-0.0265017506,0.3196493387,0.1067201793,0.0271268282,-0.0227408409,0.0360054187,-0.0466852225,0.0941285118,0.0979376286,0.2262067795,-0.1923806667,0.0167058073,0.0738931522,0.2818030715,-0.1342010051,0.078075178,0.0133331176,0.0949452966,0.1456508934,-0.0916038528,0.2577318847,0.4203606844,-0.2971400023,-0.1908222139,-0.5387561917,0.0684899613,0.2240524739,0.5419839025,0.0172649343,-0.0422137901,-0.3643981814,0.2574512064,0.0517334677,0.1914350092,-0.2369787395,0.0658722445,-0.0015073415,-0.1377456039,0.1526763886,0.0401475951,0.4620674551,-0.036152374,0.0430387184,0.1690574288,0.1855122149,-0.1026228294,-0.3262394667,-0.0700607896,0.0513733886,-0.0602581315,-0.1946315914,-0.3675562441,0.1409749091,0.0284575708,0.1171198338,0.0310898647,-0.1312018186,-0.0588846803,0.2827779055,0.1376745701,0.2420160174,-0.0946074277,-0.1827645898,-0.0555392914,-0.1610679924,-0.1038974226,0.2103390247,-0.3387687206,-0.2169366926,0.0349161141,-0.3631244898,-0.1392222047,0.189728573,0.2997423708,-0.1688533276,0.0477658808,0.0632281303,-0.3010536134,0.1986758262,0.0160260126,-0.0952098072,0.4047405124,-0.1050057411,0.0382626094,0.0519875549,0.2038112581,0.1873730719,-0.2616489828,0.0128793651,0.2922493815,0.3954139054,0.2591448128,0.1350082606,-0.3222782612,0.405521214,-0.3386874199,0.1324847192,0.1891297549,0.0677183047,-0.1655183285,-0.2611801326,0.0935561955,0.1645116508,0.2681492567,-0.0991472304,0.2509047091,0.1982537657,-0.2006964684,0.2294321507,-0.5128518939,0.1915762275,-0.0508476011,0.126376465,0.3277891874,-0.2168403715,0.3685017228,0.0258782264,0.0730272308,0.2752084136,-0.1076655686,-0.4524433315,-0.1018606946,0.0940312073,0.13225995,0.0913672596,0.3081658781,0.0151483463,-0.140831247,0.2460036427,-0.1578074992,0.2351581007,0.2202262282,-0.0943025574,0.0898437575,0.1858490855,-0.0458089784,-0.5447576642,-0.1777642667,0.2925497591,0.1203366295,-0.034336254,-0.216149509,-0.7330652475,-0.0340131149,-0.1970010698,0.1012780741,-0.110374108,-0.1762114316,0.2736002207,-0.0224829931,-0.0597601533,0.3339950442,0.350458771,0.4971061945,0.0074464064,0.4783541858,-0.3245426118,-0.0905284658,-0.2133940458,0.1004221663,0.4238580167,0.3719182909,0.4951200187,-0.0396295302,0.0080789691,-0.3355074823,-0.3686025441,0.1557384282,0.0461049378,-0.001350619,0.1467974484,-0.0734512135,-0.0287462417,-0.1199162155,0.0686413571,-0.2481957078,-0.3877100646,0.1072674915,-0.105117701,-0.1983692646,-0.06223749,-0.1494958848,0.055049669,-0.2839042246,0.061549332,0.0073672766,0.0383277014,0.0624648929,-0.1240568087,-0.0724500641,-0.1773536354,-0.1093214974,-0.5126647949,0.0374531932,0.1123473793,-0.1995929629,-0.0348566696,-0.197803542,-0.269117415,-0.1056673899,0.101622723,-0.4269977808,0.1791707575,-0.5763946176,-0.1295300275,0.2940059006,0.0639302284,0.2215138227,0.0082819555,-0.2006286085,-0.0628703982,-0.2191715986,0.1117746457,0.0292152632,0.1789409965,-0.0316228867,-0.0902842507,0.1129957438,0.3712133765,0.5309602618,-0.1422596872,0.0574783459,-0.1158839911,0.1340875179,-0.3771362901,-0.20935224,0.1344402134,0.0601319894,-0.1882447302,0.3518439829,0.0227295067,0.0415974781,0.2382805347,0.0957568884,-0.0104335193,-0.3102916181,-0.1118428409,-0.1433180869,0.0821125135,0.0093648965,0.0204155874,-0.0131324548,-0.0244538616,0.1058972478,0.0938302577,-0.1866975576,-0.1888062656,-0.6277314425,0.1478975564,-0.1956106573,0.2822773457,-0.230009988,0.2345071137,-0.0183770433,-0.1582256556,0.2294249684,0.3058253527,0.7938646078,-0.3571519554,0.0334887169,-0.1460759789,0.0840379447,-0.3862646222,-0.1231765747,-0.2352197468,0.071044974,-0.0093955481,0.5702295303,-0.0048345639,0.0423759297,0.3307508528,0.1047166139,-0.1241951212,-0.158497408,0.1069349274,-0.2348849326,-0.0645142794,-0.1632273048,0.140400663,-0.1262714863,0.1816483289,-0.1530652195,-0.2959682345,-0.0248315819,0.3182707429,0.1663443595,0.4955883324,0.0406691544,0.2449928522,0.4133763611,0.0887402892,0.271784991,0.6212050915,-0.1827764958,0.0003751641,0.1738278419,-0.5075185299,0.1064237282,0.3277516663,0.2514624,0.2533440292,-0.2316041291,0.0664522499,-0.2817179561,0.0183800682,0.0327746905,0.142437309,-0.0422356911,-0.3483804166,0.1831749529,0.0480573997,-0.0359343253,0.47536093,0.0587792657,0.1175362989,0.5338934064,-0.0408375449,0.7115446925,0.2526963651,-0.1143575162,0.1240061894,0.062445309,0.19392398,0.1893161833,0.0028493861,-0.4139647484,-0.4884984195,0.0611251891,-0.2067425102,0.0375745632,-0.0274843257,-0.4680762887,-0.3129974306,0.0349494033,0.2858318388,0.0163763128,0.1628247052,-0.2004487067,0.011907069,0.4998903573,0.1131540239,-0.0720998421,0.2397081256,0.1421412677,-0.0480197668,0.1149773672,-0.0588597395,-0.4039325714,-0.0975260064,-0.2621707618,0.4430781305,-0.094447732,-0.4981281459,0.2750097215,-0.2807173133,0.4163926542,0.0163350031,0.0444665775,0.3077631593,0.457554251,0.1445823312,-0.040623486,-0.2506301701,0.0559913814,0.2520436049,-0.2147709876,-0.2938264608,0.398568362,0.242247358,-0.4782817066,-0.0818516761,0.0184907429,-0.11549142,-0.0367559642,0.1469837278,-0.2320756018,-0.3868403733,0.1132361814,-0.2739658952,0.0853125304,0.1801199168,-0.0837924778,-0.0534275286,0.0062190997,0.0036755602,0.3757433593,0.1464045048,0.2301342487,0.1947802901,-0.1685251445,-0.4992214441,0.7249481082,0.187923342,-0.6645482779,0.4152348042,-0.5926182866,-0.0975003541,0.018559359,0.0750939175,0.0913069472,0.1305346936,-0.3314328194,-0.0626808256,0.0193739533,0.1418280602,-0.3883522153,0.4006213844,-0.184187904,0.1764028072,-0.0068229726,0.1565921307,-0.3094122708,0.2813387215,0.1256444901,0.1859951168,-0.6349822283,-0.3963675499,0.1017129794,-0.203508392,0.0734289512,0.1088269725,-0.0567016415,-0.2499236315,-0.1186283305,0.1249881908,-0.2432569563,0.0681026354,-0.0357738324,0.0512123257,0.0588426553,-0.2939887643,0.1132899597,0.1498596519,-0.0010446082,0.0190525055,0.1493361294,-0.3508696258,-0.088975437,0.0730254203,0.11210078,-0.0147928279,0.2364004701,0.0122344736,0.0618710704,-0.1323663741,-0.2046422213,0.2039265484,-0.2525182068,0.1091999561,0.0315728076,-0.096859917,-0.158345744,0.2479764223,0.3089214563,0.0742744729,-0.1063593924,0.0874212086,0.0083319545,0.2337899357,-0.4056389928,-0.1903107464,0.3689658642,0.2954415381,0.2008636743,0.2110202909,0.2073889822,-0.1393021345,-0.290696919,0.0740201473,0.4134844542,-0.2731982768,0.721617043,0.503546834,0.0336251557,-0.1767079383,0.3458384275,0.0674599409,0.0623523369,0.3285690248,0.1471339464,0.3004637957,0.3881609738,-0.1212980747,0.5118225217,-0.1894971579,0.0512277074,0.0973377675,0.0219249055,0.2297852486,0.2850725949,0.2617585659,-0.2613041997,-0.2151097059,-0.2266894132,-0.4083160162,-0.1197575554,0.1052452847,-0.2379235029,-0.1495076716,-0.3927215934,-0.1995561123,-0.045484256,-0.1014358401,0.34226197,-0.019004995,-0.1157050952,-0.4549862444,-0.1677140743,0.2384355217,0.1792158931,-0.3914793134,0.073939085,0.2507441938,-0.0947633833,0.2820994854,0.5637273192,0.2956740856,0.1354662627,-0.2192551494,-0.0522691049,-0.0687265545,-0.2237273306,-0.1305772066,-0.1607714444,-0.0984271467,-0.2071688324,0.3236301243,0.1271722913,-0.0511732362,-0.0578405112,0.4083117247,-0.0469181538,-0.257332027,0.0439029373,-0.2360372692,-0.0350836031,-0.0192108843,0.1272952706,-0.5087063909,-0.3370307982,0.3923962116,-0.4343366027,0.0494831726,-0.1700672954,0.0786047354,-0.3279531002,0.255158633,0.2591193318,-0.4408549368,-0.0875779763,-0.3193072677,-0.3819572628,0.1186597198,-0.2583672702,-0.1136335209,0.3748677075,0.2687956393,0.0388370231,0.0121087348,0.4229905605,0.1843013465,-0.3646041751,0.2250579149,-0.4783328772,0.0080318702,0.0546013415,0.0696869418,-0.0709013119,-0.1301875263,0.4494130313,0.1452189982,0.0451134332,0.1023077369,-0.0495584197,0.1228892952,-0.1102641001,0.0613370314,0.24945952,0.0550905354,-0.04664566,0.1468426734,-0.2963097692,-0.4016417861,-0.2142000943,0.2366359979,0.5126778483,0.2268670201,-0.2126492113,-0.3995494843,-0.0733442083,-0.0713725388,0.2082241774,0.4261969924,-0.2668526769,0.3109864295,-0.1845207661,0.1164305508,0.1804144233,0.2022798061,-0.0714788735,-0.2593640685,0.1447174549,-0.6502587795,0.3783638477,0.0070334938,-0.2603648901,-0.2389973253,0.1434078515,0.3303184509,0.1517160833,-0.3037117124,-0.1993384063,0.3814299107,-0.1813216507,0.0264559723,0.0717017949,0.1154285446,-0.1950414032,-0.0596238412,0.1469551474,0.4326779544,-0.2566993535,0.3158218563,-0.1873208731]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1895","title":"Bug Report: timestamp[ns] not recognized","comments":"Thanks for reporting !\r\n\r\nYou're right, `string_to_arrow` should be able to take `\"timestamp[ns]\"` as input and return the right pyarrow timestamp type.\r\nFeel free to suggest a fix for `string_to_arrow` and open a PR if you want to contribute ! This would be very appreciated :)\r\n\r\nTo give you more context:\r\n\r\nAs you may know we define the features types of a dataset using the `Features` object in combination with feature types like `Value`. For example\r\n```python\r\nfeatures = Features({\r\n \"age\": Value(\"int32\")\r\n})\r\n```\r\nHowever under the hood we are actually using pyarrow to store the data, and so we have a mapping between the feature types of `datasets` and the types of pyarrow.\r\n\r\nFor example, the `Value` feature types are created from a pyarrow type with `Value(str(pa_type))`.\r\nHowever it looks like the conversion back to a pyarrow type doesn't work with `\"timestamp[ns]\"`.\r\nThis is the `string_to_arrow` function you highlighted that does this conversion, so we should fix that.\r\n\r\n","body":"Repro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```","comment_length":159,"text":"Bug Report: timestamp[ns] not recognized\nRepro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```\nThanks for reporting !\r\n\r\nYou're right, `string_to_arrow` should be able to take `\"timestamp[ns]\"` as input and return the right pyarrow timestamp type.\r\nFeel free to suggest a fix for `string_to_arrow` and open a PR if you want to contribute ! This would be very appreciated :)\r\n\r\nTo give you more context:\r\n\r\nAs you may know we define the features types of a dataset using the `Features` object in combination with feature types like `Value`. For example\r\n```python\r\nfeatures = Features({\r\n \"age\": Value(\"int32\")\r\n})\r\n```\r\nHowever under the hood we are actually using pyarrow to store the data, and so we have a mapping between the feature types of `datasets` and the types of pyarrow.\r\n\r\nFor example, the `Value` feature types are created from a pyarrow type with `Value(str(pa_type))`.\r\nHowever it looks like the conversion back to a pyarrow type doesn't work with `\"timestamp[ns]\"`.\r\nThis is the `string_to_arrow` function you highlighted that does this conversion, so we should fix that.\r\n\r\n","embeddings":[-0.1527938545,0.3287227452,0.004078465,0.0871678293,0.094773151,-0.0773305222,0.4513304532,0.3623484373,-0.491699934,-0.2842443287,0.2675407529,0.5988483429,-0.1803753823,0.0181875285,0.0292804502,-0.1218024567,0.1887533069,0.2048997879,0.0169893932,0.2929658294,-0.411757797,0.0337461047,-0.0582276806,0.4490803778,-0.2121623605,-0.098631613,-0.0603366084,0.0371372662,-0.1723987162,-0.6164935827,-0.0294905044,-0.2923269272,0.2255375087,0.458345443,-0.0001204796,-0.0532897338,0.3580789566,-0.003380205,-0.1952470243,0.2531554103,0.2438883036,-0.2023748159,0.2693036795,-0.2410607934,0.0408186913,-0.2710432708,-0.0724558979,-0.0755451173,0.0478876121,0.5628821254,0.1174847186,0.3736271858,0.4515483677,0.2590070069,0.8312047124,0.269094348,-0.2306951582,0.056339018,0.477792263,0.0477021076,0.2724705338,0.0612961426,0.01842664,-0.2222912312,0.3558754027,0.1927733719,0.2080835104,-0.2612546384,-0.2831192613,0.0733712092,0.5608472228,-0.4021904171,-0.3237973154,-0.2525907159,-0.3008702695,-0.3398887813,0.1143042594,-0.0021664116,0.0236200877,0.1635642946,0.2240136117,0.1570541412,-0.2538725734,0.3059591949,0.0345591567,0.1951960921,0.036505729,0.2330444902,-0.1538335681,-0.163309142,0.4218271077,-0.0682979152,0.2391823679,0.2117653489,0.0548948757,-0.0026040752,0.0709574968,0.0192168001,-0.0629317686,0.2086346149,-0.3227862716,-0.055237975,-0.0514091104,0.0474792868,0.1782118976,0.1103378609,-0.0465708449,0.1961290538,0.1691955924,0.0495364591,0.2670128345,-0.075864315,0.0448122546,-0.3755866885,0.2224445045,0.1253210306,0.5929380655,-0.1710079014,-0.1433101594,0.0585863478,-0.3615436554,-0.1077894643,0.1066295952,0.1574509144,-0.1743690819,0.404202342,0.2209697515,0.106683895,-0.0398546904,-0.1873346418,-0.1595071107,0.0632436201,-0.312923789,-0.0837123692,-0.0525706373,0.08560922,-0.1687335223,0.1228703037,-0.2593176663,-0.0080446023,-0.2724782825,-0.0224871561,-0.0080480557,0.259834379,-0.3691702485,0.1590273976,0.0179573316,-0.2260663509,-0.3272784352,0.2760610878,-0.3891295493,-0.3840898871,-0.2853721082,0.0803797394,-0.2141950727,-0.0022991954,-0.1201459393,0.0653199926,0.3260180354,-0.2303648144,0.1859744936,0.0410010777,-0.0157311149,-0.4460178018,-0.0160409063,-0.0757956058,-0.5357329845,-0.0214967299,-0.0532195196,-0.1864697337,0.2656601369,0.1523074657,0.0214728136,0.1077250987,-0.0525948666,0.1655796766,0.1830980778,-0.0440341681,-0.1000685766,-0.0628414825,0.1198739707,-0.1564222574,-0.0947118253,0.1092776731,-0.0736897215,-0.3927891552,0.0568929277,0.1541986614,-0.1479741037,-0.1493551284,-0.2447480261,-0.2944935262,0.3030258119,0.1998855472,0.2393239588,-0.0407982394,0.0045123682,-0.5796675086,0.1131130308,-0.0502080023,0.1263220906,-0.0548149981,0.5200362206,0.2141897976,-0.070763059,-0.3474243283,-0.4297995865,0.1461998075,-0.0691066831,0.1610938907,-0.4859769046,-0.3641879857,-0.2271838784,0.2800627053,0.1993660182,0.1383026838,-0.0436110385,-0.0297358986,0.1434597224,0.3332371116,-0.1515100002,0.0788408443,-0.1270583719,0.2203107327,-0.2441525012,0.2964921892,-0.2026634216,-0.2622415125,-0.12086045,0.3751481175,0.1991807073,0.0832859799,0.0865875706,0.2168290019,0.15634799,0.0290441755,-0.4961464107,0.2458914518,0.1520669758,-0.2512366176,-0.2621866167,0.3554765582,0.2291227132,-0.0469216704,0.0833044574,0.4930122793,-0.2353567034,0.031903673,-0.1030333191,0.1757520139,-0.0606435575,0.0314014256,-0.213278532,0.0498013049,-0.2454682589,0.1282763481,0.1705575883,0.0274255667,-0.4170612395,0.2885534763,0.4863225818,-0.1517165005,0.1151833758,0.1635690182,-0.2545307577,-0.1554028541,0.0685561299,0.0153659713,0.1739568114,0.1825347543,-0.0270138849,-0.0015853908,-0.3035330772,0.0931105688,0.2514898479,0.131937921,0.1806957722,-0.0920817405,0.1278416663,0.1615659148,0.0097596887,-0.1166407019,-0.0649031997,0.3374016285,-0.6451544166,0.1693527997,-0.2670386434,-0.2664468884,-0.0653513521,-0.5234569907,-0.2097799778,-0.4976215065,0.2060505748,0.0130747203,-0.129035458,0.1445627809,-0.2396755666,0.3295800686,0.0367025547,-0.0488287732,-0.1746848822,-0.3294133246,-0.3490490615,0.0426957756,0.1227638125,0.1496353447,0.238105759,-0.2039757818,0.2300887406,-0.1850446165,-0.5263569951,0.0271414816,-0.2452932596,0.5031446815,0.109993659,0.128013432,0.1552547216,-0.1925951242,0.1817753166,-0.2458168268,-0.3332672715,0.037178956,-0.1570665687,0.1477027982,-0.1943172812,-0.0874928012,-0.5084860325,-0.3264511824,0.2179849893,-0.1214900985,0.2835149765,0.4560049772,-0.1136700958,0.0235686321,-0.0690424815,0.0206422918,-0.2456250638,-0.0137555143,0.4213277698,-0.1864165813,-0.3269939423,-0.0897910222,-0.1323457062,-0.4214044809,0.043569304,-0.1114263237,-0.5367970467,-0.0529440045,0.4217863679,0.1056533083,0.0251024012,0.3719080985,0.178548485,0.0630282909,-0.1003422514,-0.2586354613,0.0566188991,0.151097104,0.1743383557,0.2761527002,0.2530393004,-0.0747738034,0.965936482,0.1138239056,-0.4457321763,0.389862448,-0.3384912312,0.1157409027,-0.0298454762,-0.0037478837,-0.2835678458,-0.2376618832,-0.0190818906,0.1120849401,-0.0420456454,-0.2181739658,-0.0301214568,0.1389134079,-0.6106889844,-0.3204309344,0.158517614,0.1101700962,0.2406476289,-0.1108918041,-0.0829683021,-0.3057219684,-0.2346674353,0.057837069,0.1387049109,0.0329600722,0.0643219873,-0.42959553,-0.0718574673,-0.4033311605,0.2331348956,0.0878611803,0.2045992017,0.0860947818,-0.4413841367,-0.0051744166,-0.0607057177,0.3204250336,-0.1267918199,0.0915321559,0.24192895,0.1802393943,-0.4087596238,-0.1198688671,-0.0064532589,-0.0864054784,-0.1325336546,0.2273883969,-0.6342525482,-0.11655882,0.1460369527,-0.0613224469,0.0395364761,0.1096673384,0.0564646944,-0.1275828183,-0.0743717179,-0.066792354,0.3582876325,0.3050360382,0.0262295958,0.123598665,-0.2668554187,0.1806503534,0.1099917293,0.1013183594,0.2407742441,-0.4582191706,0.2066280991,-0.1807590276,0.0120178424,0.4070287943,0.7095594406,0.0807230696,-0.3991869688,-0.0225861799,-0.0491637215,0.2766858339,0.4198618829,-0.1451694071,-0.0110789705,0.0248727221,0.0342696868,-0.3001809418,-0.1395234317,0.2934096754,-0.0880783871,-0.318710655,-0.04025729,0.3679546118,0.2904569507,-0.1857378334,0.2262479961,0.3401787877,-0.1935942322,0.2715738118,0.1173752397,1.1871386766,0.0625183284,0.290237844,0.691859901,-0.2262700498,0.3874614537,0.1182625145,-0.1172614396,-0.4156978428,0.0445860773,-0.0881238431,-0.0749851763,0.127363652,0.1820988357,-0.2661907077,0.1274767965,-0.109205395,0.5751085877,0.1092147008,0.0284855124,0.3513977528,0.0581372567,-0.3950576186,0.0595080033,-0.1463330239,-0.0669677258,-0.1203710884,-0.360003829,-0.4582643211,-0.2313321531,-0.3740951121,-0.1418720335,-0.1527080536,0.1279595792,0.3998519182,-0.2702067494,-0.1875254065,-0.042872645,0.1339929402,-0.4898433983,0.0362872593,0.0534932017,0.1385068148,0.0911614075,-0.0083295349,-0.0276889428,0.407791853,-0.1116941422,0.0431237817,0.1187888682,-0.1682581604,-0.40023458,0.097821489,0.1089321747,-0.3288662732,-0.1873814017,-0.3726100028,0.0175438132,0.121521458,-0.2492669821,0.0417999178,-0.0047181076,0.1482410282,0.300370872,-0.2605713904,-0.1535680741,-0.1635639369,-0.008784458,-0.1962376386,0.2959886491,0.0721460953,0.4499379396,-0.2437382489,0.0095925815,0.2266792506,0.27746436,-1.1125524044,0.1213632822,-0.2176548094,-0.0115715554,0.173529163,0.4090932608,0.2185940146,0.0233569052,0.0249557849,-0.3650189042,-0.2235943973,0.4313726127,0.0291006267,0.1161478236,-0.1057684794,0.1679746062,-0.0146390153,-0.1159441397,-0.2461662591,-0.0055374275,-0.0731100515,0.358199656,-0.1237032562,-0.1004959345,0.0893472359,-0.1796742082,0.0018194026,0.0249231178,-0.1933645159,-0.09214136,-0.1605201811,0.2176162004,0.0471408367,-0.2346194386,0.0659838915,0.006506897,-0.1924472004,0.0714405477,-0.0881938338,0.3197444379,-0.1182507947,0.3038511574,0.5535709262,-0.0406600796,-0.0375190638,0.0780653805,-0.0039796624,0.5921232104,0.0860234648,0.3497187197,-0.1798067242,0.027392026,-0.2085113376,0.298938632,-0.0330960192,-0.23612912,0.276789695,-0.2176705748,-0.1418225914,0.0069901729,0.3088608384,0.3089553714,-0.1613729894,-0.057214357,0.1093274951,0.0551286452,-0.3600951135,0.0201887842,0.1924487352,-0.009106555,0.0326310508,0.3184308112,0.2696886361,0.4179322124,0.0814510062,0.048755493,0.1866779923,0.092055656,0.0785365701,0.5822566152,-0.1735043377,0.1097609103,0.4740933478,0.2046062499,0.2165544033,-0.0329340957,0.1398327798,-0.062883094,0.533425808,0.1327573508,0.3355410099,-0.1961832792,-0.270468235,0.0710249618,0.0618186891,-0.0300514717,0.4114815891,0.4589707851,0.1298111081,-0.0805714056,-0.0094673419,0.3307631612,0.1050335169,-0.1596028656,0.1815194935,-0.1845346242,-0.3742822707,-0.2979975641,-0.3258631825,-0.2028909475,0.5982089043,-0.028054215,0.0776655301,-0.2835755944,0.2046006769,0.0635140166,0.4366495311,0.0131074172,0.0410349444,-0.0268681832,-0.2891232669,0.2568720281,0.3931136131,0.6102513671,0.2314992845,-0.3602035046,0.061978966,-0.1825956851,-0.0692616254,0.1624374539,0.1441892534,0.2920224071,-0.0268787406,0.3253534138,0.0861779749,0.0038223118,0.0794863179,-0.2405523807,0.2377735376,-0.0654830933,0.6614518762,-0.206013456,-0.1796910614,0.0107787643,0.1719624698,0.0381427594,-0.0641417354,0.3810345531,0.2532756031,0.1315001994,-0.1354820728,0.0227107406,-0.0019411991,0.2765326202,0.3934857547,0.5602554679,-0.0075215693,-0.051890213,-0.5826063156,0.0651911944,-0.2603628039,0.0356161334,0.5112119913,0.0183603317,0.2061680108,0.2826227844,0.3524319232,-0.0367427729,-0.2190715522,0.2220223397,-0.2706006765,0.1071839482,0.0888694301,0.0353242457,-0.0358551219,-0.2874413431,0.2049768716,-0.1357737482,-0.0541669838,-0.2018861622,-0.2639063597,0.0873279572,0.0622774735,0.6423357129,0.2027242333,0.241353631,0.0302191395,-0.1367004216,-0.178445518,-0.287396729,-0.3819350004,0.3547775149,-0.3083920777,0.1692904383,-0.2298476547,-0.0136597706,-0.2985231876,0.1240811124,-0.0712545291,-0.0334206484,-0.2776940465,0.248026818,-0.2848936021,0.0948361009,0.1316543967,0.0741187185,-0.0224254541,0.3093962669,-0.1988300234,-0.6252144575,0.4662373364,-0.0962081105,-0.1410774142,0.1713165641,-0.0737684295,0.2998205423,-0.0596598014,-0.4555017054,-0.3096269667,0.3819800019,-0.0605066083,-0.2291325778,-0.1350292116,-0.1367224753,-0.120448336,-0.1010311022,0.3519510627,-0.2100378126,0.1753135771,-0.0738834515,-0.41962865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1895","title":"Bug Report: timestamp[ns] not recognized","comments":"Thanks for the clarification @lhoestq !\r\n\r\nThis may be a little bit of a stupid question, but I wanted to clarify one more thing before I took a stab at this:\r\n\r\nWhen the features get inferred, I believe they already have a pyarrow schema (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_dataset.py#L234).\r\n\r\nWe then convert it to a string (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L778) only to convert it back into the arrow type (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L143, and https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L35). Is there a reason for this round-trip?\r\n\r\nI'll open a PR later to add `timestamp` support to `string_to_arrow`, but I'd be curious to understand since it feels like there may be some opportunities to simplify!","body":"Repro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```","comment_length":100,"text":"Bug Report: timestamp[ns] not recognized\nRepro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```\nThanks for the clarification @lhoestq !\r\n\r\nThis may be a little bit of a stupid question, but I wanted to clarify one more thing before I took a stab at this:\r\n\r\nWhen the features get inferred, I believe they already have a pyarrow schema (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_dataset.py#L234).\r\n\r\nWe then convert it to a string (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L778) only to convert it back into the arrow type (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L143, and https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L35). Is there a reason for this round-trip?\r\n\r\nI'll open a PR later to add `timestamp` support to `string_to_arrow`, but I'd be curious to understand since it feels like there may be some opportunities to simplify!","embeddings":[-0.0739534423,0.2633909583,0.0234291479,0.1172331348,0.0697290748,-0.102800779,0.3482510448,0.2648758888,-0.53681463,-0.2700381577,0.2399619818,0.4913219213,-0.1497713774,-0.0816278756,0.0596875735,-0.1228537112,0.1500014812,0.1810293347,0.0497963615,0.2760607302,-0.4647652805,0.0268085431,-0.1178036183,0.4554687142,-0.1349889189,-0.0517363809,-0.1678003371,0.1312940568,-0.222699672,-0.6291897893,-0.006809386,-0.2124488503,0.1671675146,0.5190418959,-0.0001214839,-0.0243573692,0.3058465421,-0.081429705,-0.0674782395,0.2833605409,0.3476228714,-0.0902878717,0.2400758564,-0.2225877345,0.0005373039,-0.2519564331,-0.0341052264,0.0672973394,0.144416064,0.490727663,0.1079129949,0.3850309551,0.4924409389,0.2931482792,0.8187010288,0.2305018455,-0.2327064723,0.2020156682,0.4911263287,-0.0119446041,0.3040035367,0.1110465974,-0.0202678852,-0.3036396503,0.4174173176,0.2343477756,0.2308800668,-0.2514504492,-0.3207892478,0.012445624,0.4684594572,-0.4220660925,-0.273563236,-0.2206660956,-0.2268998325,-0.3210521936,0.0590606742,0.0244471561,-0.09441825,0.1667205542,0.1804581732,0.1114420369,-0.2595479786,0.1966615617,0.0458390526,0.2566615045,0.0918182582,0.2258583903,-0.1358292997,-0.0887856483,0.3104703724,0.0069876313,0.197939083,0.260856241,0.1146385446,-0.0238555409,0.1147630811,0.1258624345,-0.054354351,0.218019411,-0.2752971053,-0.1187834144,0.0749938563,0.0247224476,0.1478242427,0.1771616638,-0.0778387636,0.1958505064,0.1620578617,0.1409114301,0.2483191192,-0.0972781628,0.0117753046,-0.3109066486,0.0915321112,0.0369888917,0.4998414516,-0.1672780067,-0.0820687935,0.0809872672,-0.3534942567,-0.0999075323,0.1619462818,0.2031747252,-0.1754030287,0.2231131196,0.2975910306,0.064635396,-0.0271315984,-0.1359474659,-0.1771470457,0.0740092099,-0.3025508523,0.0419389606,-0.1343196034,0.1324589252,-0.1791499108,0.1216739267,-0.2732574046,0.0485023148,-0.294860363,-0.0083999876,-0.0298526324,0.289180398,-0.3410614431,0.0869282633,0.0180822741,-0.2282739878,-0.2970764339,0.2654173672,-0.261515975,-0.4053736925,-0.3469693363,0.0767558888,-0.2768322527,0.0730395988,-0.0575086288,0.042672962,0.318020165,-0.2170475274,0.1977300346,0.1241649389,0.0056941085,-0.4275840223,0.0414000712,0.0141745731,-0.468375653,-0.007270257,0.0773429349,-0.2030068785,0.2537609339,0.1049124599,-0.001687185,0.0774715096,-0.0125497961,0.2026353031,-0.0127511136,-0.0462616347,-0.0595328547,-0.0833548754,0.0376259685,-0.1311553568,-0.0338489637,0.0006076856,-0.1404957622,-0.3639112413,0.029856557,-0.0254544672,-0.0889196321,-0.2139908075,-0.3913457692,-0.3040256202,0.1618695408,0.2460020185,0.2437745333,-0.0502096824,-0.0592442937,-0.5101994872,0.2055175751,-0.0247060806,0.0671349466,0.0306335464,0.4889732301,0.263794899,-0.0420458168,-0.2111530751,-0.3733316958,0.1900897622,-0.1035777479,0.2703511715,-0.5419317484,-0.4424498081,-0.2253599316,0.3193966448,0.1861850768,0.0407054089,-0.0255179685,-0.0455303937,0.2115215063,0.3166394532,-0.0906739682,0.0358716398,-0.1910474747,0.218768537,-0.321942389,0.388936758,-0.2348883301,-0.264787972,-0.1389497519,0.3878708482,0.2070742697,0.0228760764,0.1657654643,0.2556868494,0.1249759197,0.062481232,-0.4905864,0.2075892389,0.1242336109,-0.1908746958,-0.3281776011,0.2835147083,0.1718753725,-0.0430423878,0.0834822729,0.4734211266,-0.2386840582,-0.0334388949,-0.0330949835,0.1673658341,-0.0946146101,-0.0246200506,-0.1773063391,0.0732662678,-0.2078598142,0.1559969336,0.1349333823,0.0551781915,-0.3709949255,0.2192381918,0.4112526476,-0.1022621021,-0.0292528383,0.0717290714,-0.230445683,-0.0914861858,0.1116986871,-0.0674773976,0.1516866386,0.1919134855,-0.0590074658,0.1059635431,-0.2441176176,0.0980612412,0.266556412,0.1425211281,0.2044153363,-0.1282246262,0.0713001937,0.1824331731,-0.0108701093,-0.1943412274,-0.0333507173,0.3838225007,-0.6351848245,0.1070168242,-0.1585384458,-0.3386041224,-0.0780429393,-0.5197437406,-0.2630250454,-0.4474776089,0.2224151045,-0.0186964963,-0.1537584811,0.0871340036,-0.2732034028,0.2935717404,-0.0245447215,-0.0188590027,-0.2927134037,-0.3856007159,-0.2980498374,0.0385946408,0.1031715795,0.1414165944,0.1863087565,-0.2234013975,0.1416748613,-0.0752855316,-0.5809757113,0.0517930016,-0.2343017459,0.4214703143,0.0943552181,0.1354868263,-0.0083564147,-0.2265824825,0.2086318433,-0.1833892167,-0.2840283811,0.0269487742,-0.1109141409,0.0774557143,-0.256203115,0.0519671366,-0.4750356674,-0.3110540509,0.3278072774,-0.0538620502,0.244488582,0.4900130033,-0.2033354938,0.0783633962,-0.1862731129,-0.0581388883,-0.1955919415,-0.0609742589,0.4027327597,-0.1493617743,-0.3285061717,-0.1424711496,-0.0957720131,-0.3075481653,0.0162439365,-0.2464179248,-0.5779743195,0.0321428441,0.3910689056,0.0669242963,0.0162401516,0.3755374551,0.0941712707,0.0749978945,-0.0748711303,-0.1395760924,0.0567207858,0.1027210876,0.1628661454,0.1843212396,0.2286666632,-0.1184219047,1.0252785683,0.1548281908,-0.3985675573,0.2973780036,-0.4068078101,0.1511539221,0.100807488,0.0101968441,-0.2597646415,-0.1983235627,-0.0663356632,0.1671133488,-0.0206998233,-0.0396879613,-0.1277154088,0.1601393372,-0.5489759445,-0.3292185962,0.1012688503,0.1339627206,0.2427392751,-0.0709929243,-0.0134777175,-0.2422405481,-0.2760460973,-0.007135849,0.1489183009,0.1206922829,0.0424133204,-0.347494632,-0.0746206045,-0.540081799,0.2000551969,0.0904131681,0.1046069786,0.129386887,-0.440654099,0.072600767,-0.093158029,0.3412593603,-0.0809386075,0.0096042547,0.2593657374,0.2304246426,-0.4890086651,-0.1646184474,0.0079608588,-0.1361946464,-0.0026931926,0.2836537659,-0.6594840884,-0.0884748101,0.1195068508,-0.0813739225,0.0034233788,0.1037502065,-0.0359482951,-0.2243492752,-0.0199493971,-0.0011437978,0.3446558714,0.2038823366,0.0385996141,0.2203025669,-0.1765737385,0.209939912,0.0870150477,0.2514794469,0.2378824353,-0.4663807452,0.2183127105,-0.1073825285,-0.0237033833,0.4341363013,0.8395712972,0.0945457146,-0.3643737137,-0.0120032886,-0.1064651161,0.2241790891,0.4501791,-0.1059037969,0.0466725938,0.0237456467,0.1539503634,-0.2946677804,-0.2546074688,0.3219574392,-0.2613391876,-0.2040185928,0.099651061,0.4065245092,0.2346213311,-0.2085297406,0.1910944432,0.3937905729,-0.1838937253,0.3359527886,0.1511063278,1.2702790499,0.0139766308,0.2444759011,0.8142849207,-0.268304199,0.4792139828,0.0589140318,-0.1071836203,-0.4414278567,-0.0196415745,-0.0512077771,-0.0609592348,0.0706434175,0.2082356066,-0.2985694408,-0.004078594,-0.0550605878,0.614382863,0.1036397964,0.0107557662,0.4065075517,0.1416151971,-0.4835610688,0.0149958385,-0.0665139779,-0.0387884192,-0.2164726108,-0.3448263705,-0.443339318,-0.2966761589,-0.3549309969,-0.0789487362,-0.2381044328,0.0777387917,0.3928997815,-0.1791936308,-0.2161377221,-0.0108062746,0.1059905067,-0.4689138234,-0.0151649946,0.0966897383,0.2415173501,0.0263712984,-0.0251894165,0.0395323448,0.4224720001,-0.1926269084,0.0911812037,0.0659673363,-0.2271892726,-0.4013898969,-0.0803041831,0.1792922616,-0.2816666365,-0.1221654862,-0.2321108431,0.0369481929,0.1103408858,-0.2227006108,0.0303714667,0.000072179,0.15900442,0.3048906922,-0.327396512,-0.1706457436,-0.1958446652,0.0490120389,-0.152226299,0.181783855,0.1040563658,0.4240525067,-0.2806765437,0.0272601824,0.1039105952,0.3103799224,-1.208091855,0.0783213228,-0.1443828791,0.1256974638,0.1736299992,0.3578544855,0.3502129018,-0.0629290268,0.0975492522,-0.3580776453,-0.1638460904,0.5015025139,0.1614501774,0.0878614411,-0.0682930127,0.312572062,-0.0154462429,-0.0080323415,-0.2533616126,0.0375949368,-0.0609772466,0.3026740849,-0.0879840255,-0.0993148983,0.081822224,-0.2171735317,0.021618031,0.110018298,-0.1998908371,-0.0590150915,-0.1486640871,0.2173300087,-0.0103567876,-0.2744107544,0.1058250964,0.0243900605,-0.1482447535,0.1286582202,-0.0399119183,0.3466383815,-0.0478722341,0.1967081726,0.6293019056,0.0310005918,-0.2173807174,0.0282821674,0.0387978815,0.6772246957,0.1184674352,0.3900330067,-0.2114505917,0.0099746278,-0.1494379491,0.2731032968,-0.1109382138,-0.2645560503,0.3118435442,-0.1926226914,-0.1457064599,-0.1146255881,0.3807465434,0.3744283617,-0.1657448262,-0.1251213551,0.1148403361,0.05030277,-0.3930952847,0.0686075911,0.1412355006,0.0108563267,0.0223656576,0.3505520821,0.2716420293,0.3389990628,0.0833527446,0.0825151205,0.2047491968,0.2121264935,0.0339282677,0.688072741,-0.2037144303,0.0673743039,0.4253971875,0.2280193716,0.2544561028,-0.1298799813,0.1423652619,-0.0018907,0.2977822125,0.1041293219,0.2363485843,-0.1910442263,-0.2957153022,0.1016258299,0.0756592304,-0.0666564405,0.268263787,0.5727297664,0.1583220363,-0.0059439978,-0.0279541276,0.3884640336,0.1275228709,-0.0853436589,0.1714215279,-0.2963217795,-0.4126920104,-0.2512806952,-0.2750636339,-0.1715150326,0.5290336609,-0.0194255598,0.1003441364,-0.1836778373,0.1278781444,0.0605847165,0.4979167879,-0.066281803,0.0501498394,0.0980923101,-0.2756039202,0.2209875435,0.3142209053,0.6572133303,0.1727277488,-0.2627300322,0.0539606661,-0.1522942334,-0.0122877462,0.1789769232,0.1686475426,0.381444782,-0.0905959606,0.3732955754,0.0567316525,-0.0017730402,0.0363793038,-0.2228270769,0.245572567,-0.0188541953,0.6408087611,-0.2512758672,-0.284841001,0.0452862084,0.0861149579,0.0581634417,0.0230365731,0.087326318,0.2503396869,0.1181759164,-0.1517624706,0.0167905781,0.015795432,0.3172742128,0.4708019495,0.5868931413,-0.0127285318,-0.2436187267,-0.6182447076,0.1379017681,-0.2962483764,-0.040631108,0.5741755366,-0.003480965,0.1780946851,0.2483145744,0.2168854624,-0.1460634321,-0.4045743346,0.2246893942,-0.3030552864,0.1460393518,-0.0170184635,-0.0292106848,-0.0141216731,-0.1510019898,0.2226449847,-0.1600375921,-0.0567626767,-0.214396894,-0.2009408921,0.1222724542,0.0438151285,0.6215819716,0.1074527055,0.2562095523,0.031369701,-0.1709304154,-0.1546253562,-0.2967070043,-0.3674125969,0.3668497801,-0.2256978452,0.1504351497,-0.146248281,-0.2550419867,-0.3340625465,0.1019861549,-0.0748043433,-0.0062332931,-0.1942546964,0.1319129765,-0.2435483634,0.0780956298,0.1977069229,0.0747331381,-0.0230465159,0.3005979955,-0.2489381582,-0.6309579015,0.6126278043,-0.0882751122,-0.0696323961,0.1810095012,-0.0121719399,0.1211140826,-0.1286634803,-0.4617311358,-0.2831819952,0.3873360157,-0.0093719875,-0.2162742615,-0.1066667289,-0.1331150234,-0.193874076,-0.1164909899,0.2906976938,-0.241782099,0.1810143441,-0.0667143613,-0.3776301742]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1895","title":"Bug Report: timestamp[ns] not recognized","comments":"The objective in terms of design is to make it easy to create Features in a pythonic way. So for example we use a string to define a Value type.\r\nThat's why when inferring the Features from an arrow schema we have to find the right string definitions for Value types. I guess we could also have a constructor `Value.from_arrow_type` to avoid recreating the arrow type, but this could create silent errors if the pyarrow type doesn't have a valid mapping with the string definition. The \"round-trip\" is used to enforce that the ground truth is the string definition, not the pyarrow type, and also as a sanity check.\r\n\r\nLet me know if that makes sense ","body":"Repro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```","comment_length":116,"text":"Bug Report: timestamp[ns] not recognized\nRepro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```\nThe objective in terms of design is to make it easy to create Features in a pythonic way. So for example we use a string to define a Value type.\r\nThat's why when inferring the Features from an arrow schema we have to find the right string definitions for Value types. I guess we could also have a constructor `Value.from_arrow_type` to avoid recreating the arrow type, but this could create silent errors if the pyarrow type doesn't have a valid mapping with the string definition. The \"round-trip\" is used to enforce that the ground truth is the string definition, not the pyarrow type, and also as a sanity check.\r\n\r\nLet me know if that makes sense ","embeddings":[-0.1146750972,0.1922078729,0.0449614786,0.1444818825,-0.0285734795,-0.1225972176,0.4444709122,0.2822777629,-0.5322682858,-0.3254840076,0.3281251788,0.565227747,-0.2068839967,0.0883828998,0.0604489595,-0.1087237075,0.1322407722,0.1888604015,0.0694065616,0.302650243,-0.4672093093,-0.0324615985,-0.0768617019,0.5418869257,-0.2132998854,-0.1348686069,-0.0447074324,0.121338971,-0.1925995052,-0.62851125,-0.0460072421,-0.2513629794,0.1139964163,0.4543082118,-0.0001248188,-0.0610761233,0.3190858066,-0.0606154129,-0.1071549878,0.3047243953,0.2991985977,-0.1720588505,0.226745978,-0.2034840733,0.064833872,-0.2897109091,-0.0865030661,0.0162862763,0.1138690934,0.5138632059,0.0579914749,0.3322004676,0.4559001625,0.31643489,0.7315247655,0.2873693705,-0.3010259569,0.1178452149,0.4375963211,-0.0148453191,0.2742497623,0.060122788,-0.0338262543,-0.2455758899,0.4660461247,0.1590055823,0.2337595224,-0.2711871862,-0.2636854649,0.0157378986,0.6011711955,-0.3657228947,-0.3060893714,-0.2891941071,-0.2802906036,-0.3191669285,0.1780137569,0.0389166363,-0.0652729124,0.1685661823,0.2455635369,0.0536339693,-0.292694509,0.210945636,-0.0086206812,0.1789677739,0.1051902846,0.2517607212,-0.1219961643,-0.1045323908,0.4222159088,-0.0869216621,0.2680025697,0.2776350677,0.1601040512,-0.0247629751,0.2341123074,0.0900533125,-0.0380102582,0.251591742,-0.2591480017,-0.1095598042,-0.0133118834,0.0278955474,0.1763338,0.0915726647,-0.0460614227,0.254768312,0.2294872254,0.0362674035,0.2801785469,-0.0112602459,0.0494234189,-0.3504867256,0.1208941862,0.0874278545,0.5513427854,-0.1017854214,-0.0562895648,0.0878310949,-0.4067278802,-0.190067485,0.1278896779,0.1757331789,-0.1061478257,0.3503423631,0.1927413344,0.0354352854,-0.0859617963,-0.142204091,-0.1247508526,0.0912913904,-0.3831536472,-0.0046767071,-0.1261662692,0.1497956812,-0.1982554048,0.1039191037,-0.3030178249,0.0053041922,-0.2102535367,-0.0027507837,-0.0778624341,0.2609841824,-0.4186582863,0.041618865,-0.0050415224,-0.1475328505,-0.2964429557,0.3294450343,-0.3006254733,-0.4777580202,-0.3351682723,0.0246195924,-0.2629366815,0.0942768678,-0.0529211611,0.0758389011,0.3659572005,-0.2005764544,0.207698673,0.1488821954,0.0198466983,-0.4542139173,-0.0389558561,0.0858147144,-0.4874397218,0.0180766508,0.0117791034,-0.1783303916,0.2273170352,0.14925237,-0.0544851683,0.2853111923,-0.069012247,0.030484328,0.173755154,-0.0537420884,-0.0090095745,-0.0453506149,0.0166329201,-0.1620360464,-0.054685168,0.0518301055,-0.1548094004,-0.3739098012,0.1024479941,-0.0272694062,-0.1559444666,-0.1869696081,-0.3173422515,-0.2053689212,0.2707279027,0.1471916586,0.297213763,-0.074116312,-0.0182959605,-0.5894552469,0.0808063671,-0.0670727417,0.1290226728,-0.1090822592,0.526750505,0.2543950081,-0.019559348,-0.3033373356,-0.4691953659,0.14857319,-0.0458119474,0.2215669453,-0.4678382277,-0.3387564719,-0.2100580633,0.3160874546,0.1810052097,0.0692281276,-0.1275725216,-0.0425604582,0.1440313607,0.3177388906,-0.2035161257,0.0742744058,-0.2066335827,0.2454376221,-0.229748562,0.295658499,-0.2230956107,-0.2008726597,-0.1884634793,0.4087274075,0.1667343825,0.0398220718,0.1125768423,0.1886122823,0.2222637534,0.0503239669,-0.4616852403,0.2739307582,0.1825553626,-0.2483138293,-0.2325449437,0.2102227062,0.2238045782,-0.0391978994,0.0242674369,0.5012859106,-0.2386719286,-0.0381102934,-0.0554042123,0.1768053025,-0.0629528612,-0.0053473348,-0.2213292867,0.0950103328,-0.1456543207,0.1314743459,0.1823014766,0.0988718942,-0.419465065,0.3065633476,0.4412782192,-0.184552893,0.0657002479,0.1669250727,-0.1256091148,-0.1821093857,0.0243963469,0.0633508861,0.1932552904,0.1093817353,-0.0737958029,0.0870247856,-0.3350427151,0.1240950078,0.3025604784,0.1057851538,0.1506128758,-0.0427631475,0.0878644735,0.1648917645,0.0059595495,-0.2358991951,0.0087930551,0.361700058,-0.6444141269,0.083459191,-0.1029474288,-0.2071101815,-0.1139989272,-0.4375006258,-0.1725714803,-0.4349935353,0.2647053003,-0.0309934802,-0.2029861808,0.1580693126,-0.2468392998,0.4031077325,-0.020223612,-0.0244638566,-0.189421922,-0.3819478154,-0.27837044,-0.0091069937,0.1031072736,0.161631763,0.2546700835,-0.1953088045,0.1725180596,-0.2559407353,-0.5900082588,0.1090604439,-0.1908489615,0.4780797958,0.2302556932,0.0542719923,0.1150306016,-0.1919831336,0.1805397868,-0.2277277708,-0.3124533296,0.0068508368,-0.1336709708,0.1433003396,-0.1865099818,-0.0971053839,-0.4087033272,-0.3184026182,0.3017260134,-0.1244097203,0.2925082445,0.3748553395,-0.1885424703,0.0390923992,-0.0794436336,-0.0317176022,-0.2190653533,-0.0592352599,0.3438011706,-0.2040378898,-0.2717078626,-0.0715547055,-0.2190256566,-0.3240414858,0.1017378196,-0.1886050105,-0.5865526795,0.0506291986,0.4496210515,0.0195630845,0.0574493892,0.3803966939,0.2366747558,0.1188609749,-0.0545691028,-0.2511970103,0.0852106139,0.1458350569,0.1823966056,0.2963669598,0.2636641562,-0.1237480789,1.0641577244,0.1239254102,-0.3048214912,0.3154619634,-0.3813141882,0.1278883219,0.0245105661,0.0315708704,-0.2884289622,-0.1302001923,-0.1269648224,0.0954817235,-0.0099108079,-0.1974669844,-0.0531251691,0.1143811718,-0.5511309505,-0.3061090112,0.1621596813,0.0865288004,0.3027016222,-0.1337644458,-0.0555713177,-0.366207242,-0.1868463755,-0.0149501134,0.1488548219,0.1591360867,-0.0144721614,-0.4274669588,-0.1349991113,-0.4412764013,0.1400615871,0.1244695932,0.1612965465,0.0797282904,-0.3914797902,0.0157234296,-0.0967811942,0.3329073191,-0.2039231956,0.0405733623,0.1984584779,0.2034590542,-0.3924633563,-0.1231595576,0.0506468974,-0.0776617303,-0.1415311545,0.2285326719,-0.666691184,-0.0832034722,0.1930227429,-0.0938978866,-0.00491283,0.1773246527,0.0341729969,-0.2593523562,-0.0734561607,-0.0109661128,0.3833399117,0.1678201109,0.1851971895,0.2016892731,-0.1962920427,0.244792819,0.08637207,0.2237502933,0.1684031934,-0.513153255,0.2636957765,-0.1476707906,-0.0958304927,0.4380196333,0.7881335616,0.0549326912,-0.367303431,-0.0192520116,-0.1670998186,0.3739964366,0.4537484348,-0.1495277137,-0.0694990158,-0.0603554361,0.1180324778,-0.3618154824,-0.2566982508,0.2628566921,-0.1266414672,-0.2913715839,-0.0681965649,0.3376440108,0.284760654,-0.2532699704,0.3074707389,0.351467073,-0.1588598043,0.294798255,0.1517979205,1.2444746494,0.0614700317,0.247240752,0.7527216673,-0.1919377595,0.426687777,0.0509966537,-0.154717207,-0.397549957,0.060054671,-0.1092934757,-0.1424699277,0.174272716,0.2239438444,-0.219053477,0.0858552083,-0.0274110436,0.5473250151,0.1140354574,0.0093630552,0.3956258595,0.0952996612,-0.397062391,-0.0251928065,-0.1131135672,-0.1045871899,-0.1930862814,-0.2934211791,-0.3712642491,-0.3107536733,-0.4376641214,-0.1997319013,-0.1791955233,0.1093505174,0.3794304729,-0.1861806661,-0.1206920967,-0.0383758992,0.0713254437,-0.495221287,0.0719388276,0.0182908773,0.1696642935,0.0685536712,-0.0681995153,0.0092355907,0.4175358415,-0.1730699241,0.082803458,0.1338689178,-0.2195994705,-0.4088677168,0.0619026087,0.1559414566,-0.2963092029,-0.1515341401,-0.2595840096,0.0117759313,0.0861451775,-0.2296878844,-0.014954309,0.012776251,0.1038637161,0.3013719022,-0.3263362944,-0.1554414779,-0.1420365572,-0.1136321425,-0.1120792329,0.2243082672,0.0475020409,0.3986507058,-0.2858334482,-0.0046174121,0.2311015278,0.1710645556,-1.1378570795,0.1551970094,-0.1599299461,-0.0787613988,0.1355428994,0.3849827051,0.3726135194,0.0192530341,0.069763951,-0.3278522491,-0.1900340766,0.440749079,0.1129593179,0.1254664809,-0.164476186,0.3093409538,0.0484085791,0.0023709333,-0.2117620558,-0.0141785638,-0.0780037493,0.3418265879,-0.0410260893,-0.0964813754,-0.009172217,-0.2040373683,-0.0508271381,0.0441773497,-0.2312684059,-0.0100584691,-0.1532641947,0.2547923326,-0.0071606049,-0.2541337311,0.1426767558,0.0120169092,-0.2261398882,0.1298964918,-0.0766764209,0.3577623367,-0.1088051349,0.2292369604,0.5094114542,-0.0565434434,-0.156470418,0.1130509526,0.0622364245,0.6591524482,0.1150816754,0.3603155613,-0.1920635551,-0.0039916541,-0.2051598132,0.2868708968,-0.1598638892,-0.1706921011,0.2907277346,-0.1167994067,-0.1360352933,-0.0547789335,0.311476171,0.372276634,-0.13614133,-0.0720524192,0.1595931947,-0.0159564484,-0.404815495,-0.01873146,0.2147328109,-0.0864780247,0.069122754,0.4023573995,0.2419651598,0.3705163598,-0.0420309342,0.0478400402,0.2017497718,0.0450045243,0.1762165278,0.6065785885,-0.2892410457,0.1085496098,0.5195780396,0.2501336336,0.2066774219,-0.0102568958,0.2667907476,-0.0877649337,0.4458202124,0.1883162558,0.3032173216,-0.1875769347,-0.1511214375,0.1049322635,0.002915899,-0.0958702639,0.3808432221,0.3216409087,0.0992823243,-0.0064416076,-0.0491470769,0.3589018583,0.1302800179,-0.1068479419,0.1717421263,-0.3093046248,-0.3062058389,-0.2060706019,-0.2572782934,-0.1837344766,0.6696666479,-0.0245140251,0.1644701809,-0.1349965632,0.2490410209,0.0209417492,0.468937397,-0.0211453382,0.050972335,0.0349876955,-0.2753680646,0.1461730003,0.3481431305,0.553965807,0.2513148487,-0.3344511986,0.0559434369,-0.2266022116,-0.0308700986,0.1379650533,0.2152829617,0.3123206794,-0.0863276199,0.3279177547,0.0120886788,0.0131511902,0.1103933379,-0.174635306,0.1903260052,-0.0719167218,0.6846841574,-0.2405485809,-0.2191602886,0.0798335373,0.2028457522,0.1270466447,-0.065557085,0.3195228577,0.2477747053,0.1120904237,-0.1479393989,-0.0037353004,0.0849955082,0.3358626962,0.4570589364,0.5510883927,0.0002882683,-0.0565858111,-0.6022316217,0.1106288135,-0.2849782109,-0.0564054362,0.5855398178,0.0052759014,0.1396760345,0.2635880411,0.3131507337,0.0294622425,-0.2463753521,0.2116759419,-0.2150538117,0.0883643329,0.121181272,0.0210590698,-0.0386589877,-0.2935075462,0.1734870523,-0.0886847228,-0.120708406,-0.1904480755,-0.2710628808,0.0782652497,0.0362037383,0.624199152,0.2537045777,0.2888682187,0.0458205752,-0.2056276053,-0.1399153918,-0.3122873306,-0.363579005,0.394923836,-0.2031986266,0.1283416301,-0.2055520713,-0.0619658493,-0.3853028715,0.0676858276,-0.0603742935,0.0399143361,-0.2818533778,0.2246341407,-0.2620017827,0.0083154542,0.1375729293,0.1264619827,-0.0457650758,0.2847646177,-0.1562806666,-0.5460674167,0.5290355086,-0.065191336,-0.1494631171,0.2081088722,-0.0053828587,0.2078639269,-0.1047968119,-0.5080913901,-0.351857692,0.3884011805,-0.0272086225,-0.2498048097,-0.1115684956,-0.187611714,-0.2963377833,-0.0898066759,0.347784847,-0.3409019411,0.1478934437,-0.0169024151,-0.4204331934]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1895","title":"Bug Report: timestamp[ns] not recognized","comments":"OK I think I understand now:\r\n\r\nFeatures are datasets' internal representation of a schema type, distinct from pyarrow's schema.\r\nValue() corresponds to pyarrow's \"primitive\" types (e.g. `int` or `string`, but not things like `list` or `dict`).\r\n`get_nested_type()` (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L698) and `generate_from_arrow_type()` (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L778) *should* be inverses of each other, and similarly, for the primitive values, `string_to_arrow()` and `Value.__call__` (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L146) should be inverses of each other?\r\n\r\nThanks for taking the time to answer - I just wanted to make sure I understood before opening a PR so I'm not disrupting anything about how the codebase is expected to work!","body":"Repro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```","comment_length":96,"text":"Bug Report: timestamp[ns] not recognized\nRepro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```\nOK I think I understand now:\r\n\r\nFeatures are datasets' internal representation of a schema type, distinct from pyarrow's schema.\r\nValue() corresponds to pyarrow's \"primitive\" types (e.g. `int` or `string`, but not things like `list` or `dict`).\r\n`get_nested_type()` (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L698) and `generate_from_arrow_type()` (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L778) *should* be inverses of each other, and similarly, for the primitive values, `string_to_arrow()` and `Value.__call__` (https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L146) should be inverses of each other?\r\n\r\nThanks for taking the time to answer - I just wanted to make sure I understood before opening a PR so I'm not disrupting anything about how the codebase is expected to work!","embeddings":[-0.1425731182,0.3246956468,-0.0001188292,0.1199892834,0.0674353838,-0.0765276998,0.3947218657,0.2770192325,-0.4754169285,-0.2832741439,0.2034531385,0.5407942533,-0.1618396789,0.0006553546,0.0657191873,-0.1031982005,0.16099599,0.2012021244,0.0746155903,0.2967378497,-0.5015542507,0.0138936304,-0.0977718383,0.4974121749,-0.11246714,-0.1238003075,-0.1548793018,0.1586416662,-0.2299353778,-0.6235607266,0.0846156627,-0.2241215706,0.2440818846,0.5928259492,-0.0001198721,0.0047001434,0.3186536133,-0.0295721721,-0.0745582134,0.2892119884,0.3251001835,-0.1471627504,0.1617639065,-0.2490819097,0.0260687321,-0.3570151627,-0.0386652276,0.0843394399,0.0631872639,0.5538393259,0.1039331779,0.380754441,0.5164234638,0.2587676942,0.791349709,0.2428612709,-0.2411892265,0.163986057,0.4420024157,-0.0212106947,0.3544448018,0.0041861106,-0.0741578713,-0.3006661832,0.4204317927,0.1880339831,0.1718084663,-0.2704502344,-0.3142211437,0.0297426507,0.5606445074,-0.4373676181,-0.2953945994,-0.177010268,-0.2630363107,-0.3189742267,0.1110462248,0.0432382002,-0.0600381382,0.1891306937,0.1954855174,0.1987176389,-0.2890897095,0.1917297244,0.0323539339,0.1801727265,0.0781724304,0.2091457844,-0.1302066743,-0.1141496673,0.3688918948,-0.0431594849,0.1968961507,0.2251400501,0.0955176577,-0.0600453764,0.1935303956,0.0480217934,-0.0821159407,0.1949988306,-0.3199300766,-0.1341323704,0.0023784265,0.0736258551,0.1694238186,0.0672128573,-0.049991414,0.2378932834,0.2057169825,0.1243259311,0.2681448758,-0.088194117,-0.0147982286,-0.2870153487,0.135250017,0.067397058,0.4890877008,-0.1417488456,-0.1287476569,0.0631199107,-0.3517041206,-0.1332259476,0.1493448168,0.2231163234,-0.1738203317,0.3179601133,0.2768835425,0.0402333811,-0.0383706167,-0.1485457122,-0.1716760993,0.1480696052,-0.2860532999,-0.009941984,-0.1201558486,0.1745955199,-0.2415946722,0.2068478614,-0.3203718662,0.0915210694,-0.2103444487,-0.0939712003,-0.0475952365,0.3335162699,-0.3074407876,0.0906602815,-0.0102368863,-0.2213965356,-0.3277479708,0.2611623704,-0.2281775177,-0.4078338146,-0.3511123359,0.0891685635,-0.1868232042,0.099014625,-0.1131187081,0.0563398115,0.3458622396,-0.262634933,0.1782426089,0.0927590951,-0.0486754738,-0.4944669604,0.0261735301,-0.0464595631,-0.5584202409,0.0517700948,0.06821955,-0.240746215,0.2964725196,0.1136727706,0.0055103558,0.1590037346,-0.0400644392,0.0783476382,0.0067199171,-0.0000014846,-0.0172602944,-0.0826756433,0.1535087079,-0.1110581085,-0.0318773873,-0.0040984913,-0.1063617468,-0.413466841,0.0506184064,0.0435642786,-0.1433010101,-0.2232991904,-0.3076542318,-0.2335506976,0.219019413,0.2242191434,0.2780328989,-0.0184108522,-0.0233402047,-0.4954235256,0.1061686203,-0.0543606095,0.0125319781,0.0036810304,0.4620353281,0.244200632,-0.052324675,-0.3410769999,-0.3519198596,0.2165426314,-0.1543737352,0.266556114,-0.4333074391,-0.4322145879,-0.2469987124,0.2843517959,0.2359921336,0.0394693278,-0.0127509618,-0.0075019482,0.1634652317,0.313750416,-0.1070332974,0.0478356034,-0.2006307989,0.1544554085,-0.2843419909,0.3436804116,-0.2499451786,-0.2929549217,-0.165871799,0.370357126,0.1957437694,0.0318404697,0.1319417506,0.2155221552,0.1618633717,0.0285041574,-0.4760746062,0.2356525213,0.1555476487,-0.1008736789,-0.3313786089,0.246456936,0.1859114766,-0.0580180734,0.0594786443,0.4856963158,-0.1908375472,-0.0566003099,-0.0343563482,0.1692690998,-0.0513303131,-0.0013236309,-0.101103507,0.1046526656,-0.2391401827,0.230587095,0.0001366253,0.056585785,-0.3509805799,0.2909885943,0.4456031024,-0.1674019396,0.0147179514,0.083494164,-0.1784943491,-0.1631649584,0.0915302783,-0.0366803855,0.1846927851,0.1808855832,-0.0249650907,0.0427777357,-0.3133356273,0.1275976747,0.2737653852,0.1087989062,0.1261407733,-0.1336376816,0.0752126053,0.2009641379,0.00074131,-0.213247776,-0.0554308631,0.4063256681,-0.6351138353,0.1275184751,-0.1732435524,-0.2968247235,0.0163698308,-0.4004116356,-0.1705930978,-0.4447685778,0.1812200397,0.0509731323,-0.1545256078,0.0874485895,-0.2516877651,0.307404995,0.0014126778,0.0575638525,-0.2588644624,-0.3532381058,-0.308522135,0.0650167465,0.1218183935,0.1679161787,0.2874653339,-0.2560241222,0.1563140452,-0.1229113117,-0.5098377466,0.0441641547,-0.2387231588,0.4635874629,0.1750925034,0.140240863,0.0216932073,-0.2076889724,0.200883761,-0.1902018338,-0.2915955484,0.0148491627,-0.1831306368,0.1171564236,-0.2185306996,0.0000787072,-0.5082910657,-0.3376797736,0.2521597445,-0.0579733998,0.3331135511,0.4523911178,-0.2341374159,0.0739863068,-0.1119891182,-0.0105879586,-0.2284050286,-0.0708668455,0.3865037262,-0.1658471972,-0.3249434233,-0.1007075086,-0.1360381544,-0.3380487859,0.0097035924,-0.1703654379,-0.6069763303,-0.0062818001,0.4324361384,0.0875600278,0.0493195206,0.3641599417,0.0953375474,0.0434650518,-0.0619656146,-0.201996401,-0.0165941697,0.1031663939,0.1257831603,0.2682313919,0.244355008,-0.1529213339,0.9732429385,0.1925334632,-0.4252699912,0.2978717983,-0.3933077753,0.1986407191,0.1053328216,-0.0577903911,-0.2490445673,-0.1892093718,-0.0495350361,0.1236293092,-0.0490287468,-0.1206420735,-0.083937116,0.1164637804,-0.5242831111,-0.3016832769,0.1639481336,0.1452866793,0.2797560692,-0.1104723066,-0.0359601937,-0.3287208974,-0.2361146361,-0.015007359,0.1915815622,0.1116151661,0.0231647585,-0.4009042084,-0.0878890827,-0.497033298,0.1842302829,0.0980044156,0.14782314,0.1063738465,-0.389472127,0.0219026357,-0.0667378902,0.4054003358,-0.2339537144,-0.0019958252,0.3085433543,0.1621932089,-0.4127195179,-0.1350307465,-0.0340216272,-0.0879730359,-0.0286866985,0.2912923992,-0.6167088151,-0.1203338504,0.1549595594,-0.0830820948,0.0244572684,0.0715189055,0.0052372208,-0.2415812463,-0.0116968928,0.0000388363,0.3617350757,0.232165724,0.0151417665,0.1140036508,-0.265626967,0.2865315378,0.1159244925,0.2600768507,0.2419897169,-0.4722299576,0.1895926744,-0.1549279988,-0.0320113935,0.4399710298,0.8017591238,0.0013362864,-0.3774531484,0.0692614019,-0.0825815722,0.1664783359,0.3741441965,-0.1755113602,-0.0466297269,-0.0018876952,0.0905059054,-0.3189404607,-0.2506583929,0.3491077125,-0.2284453064,-0.2749233246,-0.0753925294,0.3444890082,0.3377458453,-0.2223604172,0.1533702463,0.2905141413,-0.1511124969,0.2589245439,0.1570722759,1.2064077854,-0.0153215826,0.232927233,0.8293604255,-0.2247523069,0.4708929658,-0.0154075501,-0.1143924668,-0.4183699489,-0.0701894239,-0.0909998864,-0.0843942538,0.1406822801,0.2364738286,-0.3014637232,0.0941990018,-0.0105568133,0.6234744787,0.1450230777,0.0283696093,0.3794142008,0.1386442631,-0.4682632685,0.0396065004,-0.048986569,-0.1091404632,-0.1863295883,-0.2961185575,-0.3392006159,-0.2683630288,-0.4254045784,-0.1056793332,-0.1911338717,0.1342063397,0.3609844446,-0.1450232118,-0.150358364,-0.0021088151,0.1601801515,-0.4690327048,0.0325323977,0.0656602159,0.2302646488,0.0883297771,-0.07449729,0.0340193026,0.457459867,-0.1669483185,0.0878560543,0.0973758698,-0.2251967043,-0.3736436069,-0.0168211274,0.1957816929,-0.3043113053,-0.182414636,-0.2569025457,-0.0583664402,0.1101156324,-0.2604668438,0.0407186151,0.0249340869,0.1772096008,0.3647125363,-0.4047940969,-0.1612392813,-0.1925474405,-0.0092803566,-0.0929734185,0.2434468567,0.0639283285,0.4381604493,-0.3122686744,-0.007938371,0.1511159539,0.2893581986,-1.1213930845,0.1736577898,-0.1421826929,0.0422297604,0.2248421907,0.3096892536,0.3495355844,-0.0235738531,0.0698694438,-0.3813153505,-0.118830502,0.4501488507,0.1593347937,0.0845624134,0.0077525582,0.2527994514,-0.0186622888,-0.0796027482,-0.2650850713,0.0204533022,-0.1176598147,0.274137795,-0.1268693358,-0.0997330174,0.0859550387,-0.2266238183,0.0168990251,0.0621908456,-0.1944761276,-0.0668078661,-0.1436803341,0.2118020207,-0.0286762156,-0.2419742495,0.1111505181,0.0409476608,-0.20070301,0.126041919,-0.0283884052,0.3093132675,-0.073716253,0.2115762085,0.5683375001,0.022435315,-0.2005319744,0.0826722682,-0.0269080941,0.6481403708,0.0997174233,0.4128547609,-0.180968523,-0.01548866,-0.2070545405,0.3070147336,-0.0667616874,-0.2506654859,0.3473496437,-0.1496489495,-0.1441102326,-0.0238031764,0.3532019556,0.3636676967,-0.1636703014,-0.0987512171,0.1394599676,0.0525334179,-0.3912105262,0.0406718291,0.1080649719,-0.0568064973,0.0505802967,0.3678475022,0.2521849573,0.3781638741,0.0158221591,0.0582321025,0.189845413,0.1869028509,0.100984484,0.6963248849,-0.1746333688,0.0840028003,0.4598173201,0.2462665439,0.207175836,-0.080582343,0.2349165678,-0.0046843868,0.3384930789,0.0963289365,0.3023405671,-0.2732301056,-0.1873209774,0.1139715761,0.0383944884,-0.0926932991,0.335911423,0.4884853065,0.0057594366,-0.0479990095,-0.0669928938,0.3490701616,0.1283691972,-0.0297104064,0.1874418855,-0.2616684437,-0.3701867461,-0.2805118561,-0.2940536141,-0.1471425146,0.4771131277,-0.0832236633,0.1124220118,-0.2286419123,0.2214773595,0.0046589035,0.4474379718,-0.0118356943,0.0205226131,0.087958701,-0.2916251719,0.160386309,0.3764319122,0.6810601354,0.2150320709,-0.299410671,0.0295351949,-0.1387153417,-0.031643305,0.1745839119,0.1355935037,0.2617887259,-0.0924993604,0.2974662781,0.0751300454,0.0038921614,0.023082111,-0.2199760079,0.2572871447,-0.0387223959,0.7626436949,-0.254390955,-0.2159519792,0.1172535643,0.1648068875,0.0974282175,-0.0208561737,0.1874935776,0.2575365901,0.1307535022,-0.160488084,0.024116993,0.0179905426,0.295624733,0.4160703719,0.5409519672,0.0188523456,-0.0909032896,-0.5557599068,0.1335933208,-0.2954984009,-0.1014697179,0.585302949,-0.0006335062,0.1555573344,0.2369497269,0.2687373459,-0.1316557378,-0.3562443554,0.2115816176,-0.2996397913,0.1014943197,-0.0165896155,-0.0062038987,-0.0589291565,-0.2343893945,0.2052758336,-0.101925768,-0.0483356453,-0.2327211946,-0.2531912029,0.1642938107,0.0329702683,0.5891050696,0.1821759492,0.2321147621,0.0006473033,-0.1848768294,-0.1043766588,-0.2903867066,-0.3705573082,0.4406276941,-0.2583458722,0.1234623119,-0.1702501923,-0.1884611547,-0.3143103123,0.1531979293,-0.1151576564,-0.0582043789,-0.2314805984,0.1538600922,-0.2495630234,0.0250833202,0.1734547168,0.0568795837,-0.0339188129,0.2845543325,-0.1821431071,-0.6252738237,0.5489295721,-0.0608155057,-0.129671067,0.1641037911,-0.0447652638,0.0542829223,-0.0908272266,-0.3985231817,-0.2870026529,0.3825747669,-0.0081696268,-0.1730787754,-0.0893271714,-0.1535514891,-0.1996258199,-0.1315865815,0.2530616224,-0.2761401832,0.1293635368,0.0007363827,-0.3553196192]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1895","title":"Bug Report: timestamp[ns] not recognized","comments":"Yes you're totally right :)","body":"Repro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```","comment_length":5,"text":"Bug Report: timestamp[ns] not recognized\nRepro:\r\n\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\nimport pyarrow\r\n\r\ndf = pd.DataFrame(pd.date_range(\"2018-01-01\", periods=3, freq=\"H\"))\r\npyarrow.Table.from_pandas(df)\r\nDataset.from_pandas(df)\r\n# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.\r\n```\r\n\r\nThe factory function seems to be just \"timestamp\": https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.timestamp.html#pyarrow.timestamp\r\n\r\nIt seems like https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L36-L43 could have a little bit of additional structure for handling these cases? I'd be happy to take a shot at opening a PR if I could receive some guidance on whether parsing something like `timestamp[ns]` and resolving it to timestamp('ns') is the goal of this method.\r\n\r\nAlternatively, if I'm using this incorrectly (e.g. is the expectation that we always provide a schema when timestamps are involved?), that would be very helpful to know as well!\r\n\r\n```\r\n$ pip list # only the relevant libraries\/versions\r\ndatasets 1.2.1\r\npandas 1.0.3\r\npyarrow 3.0.0\r\n```\nYes you're totally right :)","embeddings":[-0.1712163389,0.3689662814,0.003292226,0.1200065538,0.0389690474,0.0063596917,0.4177592993,0.3191994429,-0.5414348841,-0.2883336842,0.1708868295,0.4181439281,-0.1971515268,-0.0701071173,0.0968245789,-0.0721213967,0.1526492089,0.1948472261,0.0042301482,0.278873384,-0.4898158312,-0.0086926986,-0.160146296,0.4114893377,-0.0979179591,-0.0846081078,-0.1701647937,0.1511279792,-0.2236874551,-0.5807994008,0.0905103236,-0.1734608412,0.3139378428,0.5964645743,-0.0001206944,-0.0602834113,0.2654674649,-0.0582142249,-0.0869484022,0.2695898414,0.2911416292,-0.166187793,0.1631792337,-0.2210086733,-0.0031523786,-0.2590198815,-0.0266215485,0.0606098808,0.1177497655,0.5725889802,0.1116792709,0.3596913517,0.4053273797,0.269172132,0.8239446282,0.175139755,-0.1652662456,0.1319540143,0.5636393428,-0.0837067887,0.2992212772,-0.0009874894,-0.0765660033,-0.2952079773,0.3304923475,0.191089958,0.2567849755,-0.2762508094,-0.2971894741,0.0136371376,0.5969539285,-0.4742133915,-0.2829780579,-0.2331835032,-0.2118597776,-0.3467919827,0.1277925372,0.0166729484,0.0058834851,0.1879542619,0.0947409347,0.1735738516,-0.2729296088,0.2689314187,0.0163247027,0.2588616312,0.0926153585,0.1863214225,-0.1061083302,-0.1057699099,0.4148502946,-0.0338957831,0.2159719467,0.2340483367,0.0060429326,-0.0650783479,0.1588430703,0.2339073271,-0.0101868119,0.1859593987,-0.4148418307,-0.1419945508,0.0197583586,0.0233705398,0.1594313979,0.0949772,-0.0978199467,0.2346450388,0.2858369946,0.0727374405,0.3084191084,-0.0286313165,-0.0049750968,-0.3471416831,0.0285806507,0.092614159,0.4899860322,-0.1412359029,-0.1604263037,0.0552529283,-0.3619656563,-0.2051770985,0.1963819265,0.1992468536,-0.2182422727,0.2137087882,0.2011802644,0.0269601587,-0.0524330214,-0.2357276827,-0.1947674006,0.1632246971,-0.325937748,-0.0044013904,-0.0697133839,0.0834411681,-0.2602510154,0.1745148003,-0.2776673138,0.1033933312,-0.2527114451,-0.1092472672,-0.0871887431,0.3319549561,-0.2476770133,0.1015351042,-0.0234976001,-0.2988457382,-0.3137090504,0.28019315,-0.270804584,-0.3724544048,-0.3620957732,0.0856426582,-0.2378625274,0.0581535771,-0.0505387783,0.0747648552,0.3474942446,-0.3229500353,0.1912234277,0.052648548,-0.0998871624,-0.4426660836,0.0632723123,-0.0278244,-0.5200600028,0.0289077796,-0.044458393,-0.2612947822,0.2198861688,0.1145030484,0.0504762866,0.2109780312,-0.0775353163,0.040351674,0.0455742739,-0.1334454119,-0.1160097197,-0.1439096034,0.1102397218,-0.1481423676,-0.1186655983,0.0803177878,-0.0341916084,-0.4841594696,0.1131182685,0.1099988893,-0.1668995917,-0.1930762082,-0.2836003602,-0.3124386966,0.3115412593,0.3100444078,0.315464884,-0.0124182263,-0.0120155253,-0.5159145594,0.1398466825,-0.049790211,0.062073689,0.0085804565,0.5028082132,0.2655887902,-0.0710094869,-0.308501035,-0.4346484542,0.1792693585,-0.0132150492,0.2713472247,-0.3855877221,-0.391913861,-0.1591580659,0.1858913898,0.1716033667,0.0549747013,-0.0241008345,-0.0139735322,0.2975762486,0.3910107911,-0.1097623557,0.1277098954,-0.1737927645,0.204668805,-0.3012341261,0.3584569693,-0.2320429236,-0.2474815547,-0.0753977597,0.396563828,0.2196329981,0.0440811813,0.1611410677,0.1804697663,0.1023626328,0.0255111177,-0.508138895,0.3058562875,0.1570288688,-0.1067987159,-0.3130812049,0.1815741956,0.1639288664,-0.0477707461,0.0890898481,0.4016560614,-0.2578517795,-0.1514287442,-0.0879751071,0.1153595001,-0.1101846918,-0.0064856005,-0.1047864482,0.0477867499,-0.1570212692,0.1719448119,0.2637302577,0.0017159851,-0.343736589,0.2375441641,0.5016986132,-0.1995997429,0.0288545564,0.1195334792,-0.1818961799,-0.1606799513,0.0819393173,0.0861116722,0.1992159784,0.1731415242,-0.039982114,0.1240331158,-0.2740451097,0.05201938,0.2621642649,0.1248745546,0.1029666811,-0.1009511277,0.1064923182,0.2438187301,-0.0406373627,-0.1789779067,-0.0354533531,0.4443127811,-0.5867643356,0.0881659314,-0.2698512077,-0.3931251168,-0.0134528279,-0.401250869,-0.2389084399,-0.4459261298,0.2156419158,0.0290031899,-0.0856268778,0.0737136528,-0.1158504486,0.3515297472,-0.0225711372,0.0576807261,-0.2410780042,-0.3672991097,-0.2936166525,0.0782089233,0.1348718107,0.0445528999,0.2480589449,-0.3825285435,0.178156212,-0.1365739107,-0.5566642284,0.0415281802,-0.2646552622,0.508526504,0.1333207488,0.1465233862,0.0676574856,-0.1519384831,0.1448756456,-0.245396331,-0.2541875839,-0.0476774722,-0.152546376,0.1671404392,-0.2427244633,0.0063439771,-0.6110576391,-0.287122488,0.1713874489,-0.1518876553,0.2658904195,0.4004629254,-0.2815297544,0.0652818829,-0.199742049,0.0100220917,-0.2344829738,-0.1802892834,0.4237622917,-0.2036313862,-0.3659539521,-0.0618682094,-0.0637695044,-0.3685590625,-0.0605194382,-0.2288625091,-0.5742839575,0.0363865234,0.353282094,0.0923584476,0.1224893108,0.3748053312,0.0537482612,0.1030293554,-0.0606655031,-0.2759569883,-0.065116629,0.1605800539,0.1628348082,0.1907928139,0.2865312994,-0.1095836684,0.9504398108,0.2172856778,-0.4353504777,0.3302107751,-0.3422561884,0.2199934423,0.0635874346,-0.0498655289,-0.1792584807,-0.1103058383,-0.053880807,0.2223873734,-0.0119532766,-0.0808624402,-0.1130946055,0.1014318988,-0.5949831009,-0.3255494237,0.0544935241,0.2390346229,0.2171629071,-0.053264536,-0.0564144179,-0.2351148278,-0.3530791402,-0.0699241459,0.2208587676,0.0650967658,0.0636057705,-0.346767664,-0.1617218256,-0.5193020105,0.2152635455,0.080042772,0.1107955202,0.0319780558,-0.3380623162,0.0189990308,-0.045512706,0.4486453533,-0.2430839241,0.0613118894,0.2571922839,0.1821389943,-0.3860387206,-0.1263476312,0.0127791986,-0.121714063,-0.0074608205,0.1662485152,-0.5437633395,-0.1029359922,0.1957963407,-0.0487228036,0.0665401295,0.0139637794,0.0027535222,-0.2400275469,-0.0425048023,-0.0302318819,0.2813857794,0.2459445894,-0.0211642031,0.2098035514,-0.0890597552,0.2605317235,0.1655850559,0.2443134189,0.1777708828,-0.4257642627,0.2575263977,-0.0786437616,-0.1060435623,0.5185486078,0.7284455299,-0.074080795,-0.3300741613,0.061011821,-0.0490744486,0.2246278971,0.3502838612,-0.1157915145,-0.0573390275,-0.0321082808,0.0734164789,-0.3459991217,-0.2488318384,0.3256963789,-0.1799391061,-0.4076094031,-0.0537618138,0.3039690554,0.4024892449,-0.1858420223,0.2023984045,0.270509094,-0.1833848059,0.1874643266,0.0116080241,1.1638237238,-0.0721763596,0.2634496093,0.7411692739,-0.1460213065,0.5090304017,-0.0178908221,-0.0566175133,-0.4304604232,-0.0472796895,-0.085336864,-0.0363121256,0.165028736,0.2280774415,-0.2681397498,0.1692124307,0.0063895038,0.6298398376,0.0904826522,0.0446504168,0.4225158393,0.1603684574,-0.5168052316,0.0765145272,-0.0958628282,-0.042623803,-0.2608085573,-0.2561250329,-0.325519681,-0.2521010935,-0.4270889759,-0.1366885453,-0.0476043783,0.0824127719,0.3654866815,-0.1868573427,-0.1265697181,0.0466823168,0.2154381126,-0.4125201106,-0.003884776,0.0395429879,0.0597999841,0.139662236,-0.0602992326,0.0497021265,0.4080988467,-0.1151726022,0.0471481308,0.1151576042,-0.1259016842,-0.4068232477,-0.0193249583,0.188159585,-0.3480127454,-0.1408096403,-0.2347355783,-0.0273125172,0.2241288722,-0.2770579755,0.051412344,0.0578946471,0.2018315792,0.391906321,-0.4326790869,-0.1345729083,-0.2106649578,0.0847493261,-0.2191836387,0.1898237765,0.0029415926,0.5468834043,-0.2559200525,-0.0016308064,0.2118882835,0.1858240813,-1.0505319834,0.1141708791,-0.1551207751,0.1786454618,0.2602198422,0.3121308982,0.2324507087,-0.0938932002,0.1838261038,-0.4210488796,-0.1673953086,0.4584985673,0.1066416353,0.0531858131,0.1013858542,0.2549583316,0.035846781,-0.1112970188,-0.2669356763,0.0474660099,-0.1613498181,0.2921188772,-0.1178986281,-0.108323738,0.0657419786,-0.2237425297,0.0066882581,0.0913093984,-0.1875941455,-0.0869052336,-0.1469251513,0.2193279415,0.0000361276,-0.2647080421,0.1042259187,-0.0091218064,-0.2208395898,0.1153884381,0.0064554331,0.3029393256,-0.0191416945,0.2541543245,0.5963889956,-0.0297234822,-0.1379506886,0.0317118801,0.0059258933,0.6255276799,0.1596845537,0.436950922,-0.1903410703,0.0354017727,-0.2326796949,0.2355890125,0.0223688986,-0.3386207819,0.3317260146,-0.2019913793,-0.0964061767,0.0703661367,0.3752290308,0.2908653915,-0.2350554764,-0.0729565099,0.1494571865,0.043149434,-0.3526556194,0.0343780108,0.2417684644,0.0164456461,0.0312238242,0.2903963923,0.1973565668,0.3996395767,0.0839410275,0.0605247244,0.3632286787,0.1542024761,0.1140419021,0.6070529819,-0.2462800145,0.1013473868,0.4045569599,0.2248598933,0.1430251151,0.0210061669,0.1837040633,-0.0541263856,0.2680233717,0.0953389853,0.2811519504,-0.3394353688,-0.1548633873,0.1710010767,0.0417942181,-0.1644572318,0.2826384306,0.3999171257,0.0810352564,0.0241980273,-0.0365038477,0.4071493447,0.1262088269,-0.0478410423,0.2126591355,-0.2242040485,-0.2674947083,-0.2039017975,-0.3527131975,-0.2125602961,0.4762694538,-0.0051593059,0.0883947313,-0.203635335,0.1818143278,0.0646126941,0.3783961535,0.0281303618,0.0640391484,0.0927318931,-0.2887279391,0.2264472991,0.3718537986,0.6563258171,0.1718581021,-0.27845186,0.0698856488,-0.2321823239,-0.026480725,0.1885379106,0.2383593768,0.2889409959,-0.0322028659,0.3244104087,0.0748000443,0.0254868753,0.0562879816,-0.1683303714,0.2525105774,-0.0944247693,0.6780033708,-0.2196514904,-0.1749730706,0.0307048932,0.1894093901,0.0201420765,0.0690670162,0.264220655,0.2671748996,0.1165718287,-0.2152452171,0.0189078301,0.0339967459,0.2600192428,0.4770720899,0.5898712873,0.0057974844,-0.2198035568,-0.649102211,0.0751180202,-0.3181274235,0.0281934366,0.468167603,-0.0720829815,0.1225585416,0.2659867108,0.3522367775,-0.0005349132,-0.3501638174,0.1730748117,-0.2570110559,0.1387224942,-0.0875528529,-0.0257680677,-0.0693929791,-0.2017223835,0.2339670062,-0.2053847462,-0.075685963,-0.3458893001,-0.1399026215,0.0634431541,0.0765710548,0.6215052009,0.2365681827,0.286724478,-0.0008243248,-0.131377086,-0.1618960202,-0.2432719916,-0.3222280443,0.3669305146,-0.3413616717,0.0808060989,-0.2449175566,-0.2048253864,-0.2688115835,0.128012538,-0.1085958257,0.0493109673,-0.2868924141,0.1390610784,-0.1962713301,0.0688624382,0.1990922093,0.0340349451,0.0041455701,0.3002775908,-0.1847196966,-0.5789723396,0.4849747419,-0.1388847828,-0.0964651033,0.0805752128,-0.0394600779,0.0112224286,-0.1528961062,-0.4339449108,-0.2382514477,0.4028992355,-0.024852166,-0.2168011069,0.0175870936,-0.0998443216,-0.1691541225,-0.0784406736,0.2024047524,-0.1806793511,0.0786857754,0.0489263609,-0.3800324798]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1894","title":"benchmarking against MMapIndexedDataset","comments":"Hi sam !\r\nIndeed we can expect the performances to be very close since both MMapIndexedDataset and the `datasets` implem use memory mapping. With memory mapping what determines the I\/O performance is the speed of your hard drive\/SSD.\r\n\r\nIn terms of performance we're pretty close to the optimal speed for reading text, even though I found recently that we could still slightly improve speed for big datasets (see [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1803)).\r\n\r\nIn terms of number of examples and example sizes, the only limit is the available disk space you have.\r\n\r\nI haven't used `psrecord` yet but it seems to be a very interesting tool for benchmarking. Currently for benchmarks we only have github actions to avoid regressions in terms of speed. But it would be cool to have benchmarks with comparisons with other dataset tools ! This would be useful to many people","body":"I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB of tokens).\r\n\r\nQuestions:\r\n1) Is this (basically identical) performance expected? \r\n2) Is there a scenario where this library will outperform `MMapIndexedDataset`? (maybe more examples\/larger examples?)\r\n3) Should I be using different benchmarking tools than `psrecord`\/how do you guys do benchmarks?\r\n\r\nThanks in advance! Sam","comment_length":141,"text":"benchmarking against MMapIndexedDataset\nI am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB of tokens).\r\n\r\nQuestions:\r\n1) Is this (basically identical) performance expected? \r\n2) Is there a scenario where this library will outperform `MMapIndexedDataset`? (maybe more examples\/larger examples?)\r\n3) Should I be using different benchmarking tools than `psrecord`\/how do you guys do benchmarks?\r\n\r\nThanks in advance! Sam\nHi sam !\r\nIndeed we can expect the performances to be very close since both MMapIndexedDataset and the `datasets` implem use memory mapping. With memory mapping what determines the I\/O performance is the speed of your hard drive\/SSD.\r\n\r\nIn terms of performance we're pretty close to the optimal speed for reading text, even though I found recently that we could still slightly improve speed for big datasets (see [here](https:\/\/github.com\/huggingface\/datasets\/issues\/1803)).\r\n\r\nIn terms of number of examples and example sizes, the only limit is the available disk space you have.\r\n\r\nI haven't used `psrecord` yet but it seems to be a very interesting tool for benchmarking. Currently for benchmarks we only have github actions to avoid regressions in terms of speed. But it would be cool to have benchmarks with comparisons with other dataset tools ! This would be useful to many people","embeddings":[-0.3810420632,-0.0305888336,-0.1813578159,0.2504057288,-0.2935993969,-0.0467076786,0.0398227051,0.1909921169,-0.1505501568,-0.090393059,-0.1634379625,0.3356112242,0.0311309136,-0.4226370454,0.2201004773,-0.1310540438,0.365021497,0.0921234712,-0.2799274027,0.0275711939,-0.0007041279,0.0569512434,-0.2673867643,0.0297653079,-0.2419723123,0.0386086777,-0.0066018244,0.084529914,-0.0699152052,-0.2794293761,0.2004752308,0.2412211001,0.0896887183,0.59246099,-0.0001197653,-0.1850560457,-0.0232878737,0.1665841192,0.0442655943,0.3798983991,-0.334876895,-0.2034229338,0.16586411,-0.1334248632,0.1651628464,-0.2025379241,-0.0441732407,-0.7676182985,0.0728932098,0.1517505944,0.0765773132,0.4518498778,-0.4108696878,0.0904258117,0.0827235654,-0.0421303883,-0.1984526217,0.2391930372,0.5925350785,-0.0382881835,-0.0415330492,-0.0342429578,-0.2421227098,0.0677972808,0.156667605,-0.05456613,-0.2205545157,-0.0158528481,0.1271639764,0.5387344956,0.241151154,-0.427827239,-0.2604738176,-0.3222860992,-0.2503652871,0.0612294264,0.1345780641,-0.2026060969,0.0095424447,0.0385263078,-0.5764119625,0.1521897465,0.1874468476,-0.0834676996,0.0837983787,0.2970573604,0.2520069182,0.0690655559,0.478025347,-0.2811423242,0.2511444986,-0.3926388919,-0.1077352241,0.0773697346,-0.5279675722,-0.2925722003,-0.1000031233,-0.0710350126,0.1268396229,0.353518784,-0.0782521516,0.4015808105,0.4423339963,0.0821368694,0.1381019652,0.3148056865,0.1134951413,0.1276748031,0.0943863019,-0.1098216847,-0.3615859151,0.1438644975,0.2758626938,0.0285086334,-0.1372569203,-0.2606762648,-0.4354254901,-0.5328081846,-0.2650946975,0.1021634638,0.1557455361,-0.3873815835,-0.0231652297,0.1626877934,-0.080462873,0.1054965928,-0.4629507363,0.1284536272,-0.2905658185,-0.2079000771,-0.0178972352,0.1329868883,-0.0984145477,0.4099316597,0.1127967015,-0.0160043538,0.2330067009,-0.0401802622,0.1472288966,0.0203946382,0.608997941,-0.3896597028,-0.0092049912,-0.0509615205,-0.0262422822,0.1041255146,0.0037979241,0.2796248794,-0.3651382327,-0.0016902328,-0.0569008403,-0.1275285929,0.3806732893,0.0101187434,-0.4040935636,-0.3670137823,-0.6880221963,0.4135314822,-0.1943508089,-0.2400924265,0.1414815485,-0.0057902974,-0.2251413614,-0.1442043781,0.0685545504,-0.065207392,-0.2823380828,0.0577856749,-0.2562697232,0.0964200497,-0.1205662936,0.4209596813,-0.341702193,0.2553822994,0.0339739807,0.1158019304,0.2943266332,-0.522521317,-0.4324796796,0.0042804005,0.2473973632,0.2319125831,0.0130043169,0.1114886478,0.3896990418,-0.2509157658,-0.0663521886,-0.018187508,0.1446327567,0.2138108015,-0.1329520941,-0.377420783,0.1667663455,0.1763423234,-0.2033471018,-0.3884898126,-0.0828033462,-0.2697426677,-0.0737745091,-0.1108160838,-0.2474601418,0.0625223294,-0.0462348983,-0.2136875838,-0.0365388133,-0.3376493454,-0.0658715144,0.2914369404,-0.2398122251,0.0201900098,0.324151516,0.0338926353,0.0636358187,-0.1058214158,0.0409154706,-0.2654241025,-0.0032363592,-0.172893554,0.308970958,0.1328320056,-0.307633549,-0.179654941,0.1738008112,0.0090148868,0.1496795118,-0.0337287411,0.3308536112,-0.1629496962,-0.102573894,0.1462418288,-0.2228517383,0.0144666284,-0.192617923,0.0757529214,0.2122276425,0.3306454122,0.5596898794,0.4596720934,-0.0421850681,-0.2652954161,0.2174792886,0.0666743368,0.1624614745,-0.195156008,-0.1541749239,0.6028689146,0.3290905356,0.4167497754,0.089675568,-0.4029943645,-0.2495696545,0.1632981449,0.19101201,0.2234076262,0.1918010563,0.5014183521,0.3815578818,0.1069945395,0.2281945199,0.1957737952,0.4031579792,-0.0130919581,-0.0780945495,0.0731680095,-0.4537008405,-0.2498407662,0.358250916,-0.0080586197,0.6080089211,0.0452938229,0.3489488959,-0.0202739481,0.0384119004,0.2139678895,0.2788416445,0.1956564784,-0.3231135905,-0.2122598439,0.0817539841,-0.0996227786,0.0957184881,-0.3293740153,-0.1609760374,0.1362846047,-0.0265140701,0.1889923513,-0.1423367262,0.2636851966,0.1514907926,-0.0919877291,-0.2330502868,-0.2190002948,0.0291683003,0.0568972118,0.1278783679,0.0836860836,-0.0180447083,0.2556385696,-0.503836453,-0.3434745669,0.1368796229,-0.1046149507,-0.4280404747,0.0491807386,0.4018220305,0.2501460314,0.3435859978,-0.0171820819,0.1546420306,-0.0059221131,-0.0546263978,0.0654993132,-0.0305257924,0.2342978418,-0.2400772125,-0.2851139307,-0.3815671504,-0.2068900764,-0.0194856543,-0.2105273008,-0.0706485286,-0.3994475007,0.0426964909,-0.0061722267,0.3142234981,-0.0383286141,-0.1715778708,0.0247891769,0.1571281999,-0.15209499,0.0684840754,-0.6228299737,-0.0046539069,0.3014576733,-0.3341757953,0.2447912991,-0.015212358,-0.6034364104,0.1193514168,-0.0281742029,-0.1584166586,-0.0312690176,0.0633481517,0.2792357504,0.1650233865,-0.2042656988,0.0022451428,-0.144772917,0.2476503849,0.205782786,0.123665005,0.346288085,-0.0707701594,0.0550051406,0.0003557114,-0.1202043593,0.0172309559,-0.2281658351,0.0850014165,-0.0106464885,0.2428559661,-0.1317143738,0.9007663131,0.3081695437,-0.2566815615,-0.0362229943,-0.1160542145,-0.0649496466,-0.202250883,-0.0032782564,-0.1313838214,0.0836992562,0.0835961252,0.4366495311,0.0672881007,-0.1921498775,-0.2168181837,0.2444776744,-0.1167559251,-0.1476192176,0.1422252059,0.0472918861,0.0865886733,0.0697297826,0.2408092022,-0.4479861259,-0.3257124424,0.1657262444,-0.1058409885,-0.2687660754,-0.208068341,-0.4598976076,-0.2253926396,-0.4417966902,0.3698550165,0.2283757031,0.3858377039,0.0419383682,-0.0493478328,0.1058860719,-0.0716939941,-0.1567471623,-0.1842283607,-0.2512066066,0.1228974909,-0.1194097996,-0.2804475129,0.0593971163,-0.1287492365,0.0730540678,0.1754046828,0.0934483707,-0.1135771275,-0.1453138739,-0.1706183702,0.1141082421,0.0440540984,-0.1558131427,-0.1495450884,0.026927378,0.2580502927,0.1313979775,-0.334325105,0.0245797709,-0.0810637772,0.0945228264,-0.2703286111,-0.0469448939,0.4475466609,0.5082722902,0.2300990224,-0.4144336283,0.4280772805,0.1398906559,-0.2195221782,0.2852533758,0.2222192734,0.1147792041,0.1803373843,0.0512736291,0.0061850473,0.1618946046,0.054343231,0.061525993,0.1596080363,-0.1863074154,-0.0937585533,-0.2137928754,0.0187278613,0.2513189912,0.374475807,-0.3786364794,-0.3073340654,0.6385368705,0.5235541463,-0.0965763852,0.0525805578,0.0129983127,-0.1762463748,0.0544871502,0.2187593281,0.893015027,-0.1143616512,0.1361534745,-0.0392666832,-0.0860790312,0.3093455434,-0.3286449611,-0.1282592416,-0.132983759,-0.3400751054,0.0043992144,-0.1569239497,0.3266392946,0.2135235667,-0.000568721,0.3174332678,0.3206454515,-0.1586578637,-0.045405034,0.3780540526,0.1787365079,-0.1435991228,-0.1427614838,0.0819604769,0.062893942,-0.1736571342,0.0926641598,0.1084843874,-0.4816479385,0.3293642998,-0.4015117884,-0.2510002851,0.0079932474,0.2396191359,-0.079739809,-0.1534740776,0.1838406026,-0.0968568325,0.1813536882,-0.1740348041,-0.164255783,0.1746462584,0.0699734017,0.3781450987,0.4506776929,-0.2396115214,-0.0631688684,-0.1362427622,-0.137851581,0.2459464222,0.0079497136,-0.4742002189,-0.2062947899,0.0371296853,0.1810423434,-0.092326723,-0.1455951333,0.1721229255,-0.0562836342,0.1244474724,0.0392775089,0.1677746773,0.0337306559,0.8375976682,-0.2117079496,-0.3055508435,-0.1786480546,0.5786249638,0.0612991452,-0.3443079591,0.189752087,-0.0601633266,-0.2026634216,-0.1124159247,0.1679845303,0.1435300708,-0.120711118,-0.0137460334,-0.4844608307,-0.1354613155,0.300834775,0.1886303425,0.3548614085,-0.2195270061,-0.063579075,0.1740076244,-0.2841738164,0.2056913674,0.1625823081,0.2614249885,-0.1952732652,-0.2161904573,-0.200156644,-0.0141646517,-0.1311893761,0.1216102019,0.0355168693,0.0125301182,-0.0895135105,-0.0025198432,0.2382045537,0.1458501816,-0.1338468194,0.1199318394,-0.2701048553,-0.0916061327,-0.0081749698,0.1948789805,-0.1186747998,-0.4352926314,-0.0065529281,0.0658313632,0.0443832353,-0.2980178595,-0.1076629609,-0.3384864032,-0.1209913269,-0.212324366,0.4334196448,-0.0772559494,-0.1824731529,0.3083030283,-0.0912151411,0.1343858391,0.1567104757,0.3019170761,0.1051746905,-0.2173040509,0.2668798864,0.2612120509,0.2187346518,-0.1041395366,0.1611050069,-0.1335749477,0.012031002,0.3733333349,0.1930890828,0.7797372937,-0.0087356353,-0.4619467854,0.2343355864,0.0584151037,0.2420035005,-0.2574577332,0.3182439804,0.1881200522,0.1099917218,0.1034220755,0.3604651392,0.118177928,0.229156673,-0.1195742488,0.4428493381,0.0472076274,0.1710546464,-0.4130825996,0.1040693969,-0.0341942757,0.1235078052,0.216870591,0.2585467398,0.1693549454,-0.2604327202,0.6944693923,0.4254680276,0.2541790903,-0.0289790668,-0.2021910995,0.3091746867,0.3924271166,0.1401467472,0.2696558535,0.2136562914,0.364964664,-0.2042591423,-0.09722846,-0.1767688543,-0.0155599024,0.1200892478,-0.4284583926,0.3444975019,-0.1480668634,0.0189351719,0.1770252436,-0.24061203,-0.1653125584,0.1838980764,-0.1320498586,0.0844628438,-0.1065875739,-0.1226844117,0.0567378029,0.1678584814,0.0700049996,0.1653280705,0.0613432154,-0.0631594509,0.249922201,0.3911224306,0.0704721585,-0.14906165,-0.579934597,-0.044205334,-0.2540729642,0.2481929511,0.1321778148,0.2059017569,-0.3404233456,-0.1203946099,0.2132739425,0.0382157527,0.0154140564,-0.5190162659,-0.217396304,-0.0050257812,-0.3723578155,-0.3217624724,0.0190562215,0.1719231308,0.1021031365,0.1325762123,-0.0789241344,-0.3048042059,0.1418816298,-0.2075071186,-0.0879505426,0.1146052852,0.0292846449,-0.1355268657,0.6446287036,0.2345737219,0.1028292924,-0.1635216475,-0.4154815078,-0.3512727618,0.0232473705,-0.1005347148,-0.1174554452,0.1458581239,0.2632142901,-0.1220303401,0.3209252656,0.3944303989,0.520883739,-0.1863090992,0.11976704,-0.4308366477,-0.006904162,-0.3667601645,-0.2619175613,-0.0267945305,-0.0647680461,0.0552319959,-0.1110080257,-0.1463379562,-0.1283796728,-0.0659102872,0.0990431458,0.0858683139,0.5319930911,0.2502122521,0.5302524567,0.3108712435,-0.07150466,-0.0706610382,0.1376149505,-0.0680530444,0.1617410332,-0.066642262,0.047036089,-0.1838991046,0.3131071329,0.1243215278,0.6157354712,0.1630045474,-0.3072164059,-0.1744776666,-0.2664985955,0.2768178582,0.0297354087,-0.0782983899,0.3815819919,0.117402032,0.4984681904,-0.13551566,0.161210075,0.1391355395,-0.4457457662,0.056701269,-0.1604028493,0.0219391361,0.1593650728,-0.0639591664,-0.507874012,-0.324285537,0.3920134604,-0.0146357985,0.1442724913,0.2979818881,0.0034138227,0.1368669122,-0.1357822269,0.0685589835,-0.341779381,-0.0102835065,-0.0427371487,-0.0472321697]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1894","title":"benchmarking against MMapIndexedDataset","comments":"Also I would be interested to know what data types `MMapIndexedDataset` supports. Is there some documentation somewhere ?","body":"I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB of tokens).\r\n\r\nQuestions:\r\n1) Is this (basically identical) performance expected? \r\n2) Is there a scenario where this library will outperform `MMapIndexedDataset`? (maybe more examples\/larger examples?)\r\n3) Should I be using different benchmarking tools than `psrecord`\/how do you guys do benchmarks?\r\n\r\nThanks in advance! Sam","comment_length":18,"text":"benchmarking against MMapIndexedDataset\nI am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB of tokens).\r\n\r\nQuestions:\r\n1) Is this (basically identical) performance expected? \r\n2) Is there a scenario where this library will outperform `MMapIndexedDataset`? (maybe more examples\/larger examples?)\r\n3) Should I be using different benchmarking tools than `psrecord`\/how do you guys do benchmarks?\r\n\r\nThanks in advance! Sam\nAlso I would be interested to know what data types `MMapIndexedDataset` supports. Is there some documentation somewhere ?","embeddings":[-0.2623206377,-0.0587809421,-0.1896093339,0.2777880728,-0.227726087,-0.0737620816,0.0111633465,0.1357609332,-0.2933735251,-0.1568026692,-0.1532005072,0.3774213493,0.0027406048,-0.3708997667,0.2234075814,-0.1318493038,0.3597462773,0.1027279869,-0.2230797857,0.0388148762,0.004757049,0.0907562524,-0.3878263533,-0.0007551571,-0.308994323,0.1608976424,0.0237744935,0.1252493113,-0.0787538067,-0.327116996,0.2054045796,0.2334433347,0.0119257383,0.6243491173,-0.0001231747,-0.2159653604,-0.0060990448,0.1145242155,0.1292808205,0.4733529687,-0.355368048,-0.1873503625,0.1760196388,-0.0643589869,0.1672016978,-0.1405496299,-0.1546870023,-0.7546443939,0.0663225204,0.166835174,0.0619084761,0.3860285878,-0.4022429585,0.0922332853,0.0853737816,-0.0240259543,-0.2072671652,0.189033702,0.5677084923,-0.0037737319,-0.0012585642,-0.0574584752,-0.1934132576,0.0176083986,0.1063681766,-0.0935725346,-0.1897952706,-0.0806949511,0.1602991521,0.5714413524,0.3033574522,-0.3853615224,-0.1944885999,-0.3469295502,-0.2907654643,0.1559844762,0.1440193355,-0.304181397,-0.0016855652,0.0526980609,-0.5018880367,0.1885734648,0.1384049654,-0.166680187,0.0034871246,0.4123747349,0.2059561908,0.0534100756,0.463853389,-0.2791369557,0.2155865878,-0.2781588435,-0.0903226882,0.1010188982,-0.5098439455,-0.2299691886,-0.0520547517,-0.0863875672,0.135603115,0.3474746048,-0.2032877505,0.5046434402,0.429818213,0.0884598643,0.1067866907,0.2343621403,0.1511718035,0.0173080042,0.0804942399,-0.1772749275,-0.372879535,0.096013248,0.1723684818,0.0604582168,-0.210566476,-0.2381728142,-0.3924916089,-0.4892415702,-0.2800405324,0.0489859879,0.3160374463,-0.3449766636,0.045167394,0.2040766776,0.0596403033,0.1809966713,-0.4886460304,0.055304911,-0.2804618478,-0.2519147098,0.0053385948,0.0612695068,-0.0932258442,0.3304063082,0.117045477,-0.0538370311,0.2595546842,-0.0688250065,0.2348473221,0.0566518083,0.51203233,-0.3035805225,-0.0164029244,0.0229024831,-0.1128176227,0.1444650739,0.0295833126,0.3873952627,-0.3637986481,0.0013563817,-0.1206983998,-0.1722417772,0.3445302844,-0.0133164339,-0.3269232213,-0.3736244738,-0.575011909,0.3550662398,-0.1666037738,-0.2867623866,0.2065920681,-0.0011886382,-0.1739472747,-0.2382938266,0.0253138449,-0.1241655499,-0.235348165,0.0866830572,-0.3662692606,-0.0079210345,-0.1065398455,0.357750237,-0.3426887989,0.2114575654,0.0139793782,0.1618296057,0.3885022402,-0.5090394616,-0.4399956763,-0.0519217774,0.2221059501,0.2691320479,0.0002896641,0.1311459839,0.2970203459,-0.2753646374,-0.0625803918,0.0255434271,0.1016711816,0.2055302113,-0.1206864789,-0.2674798071,0.2851322293,0.1665535122,-0.2390076518,-0.3818077445,-0.1458150148,-0.259470731,-0.0230305456,-0.0676745251,-0.2381089032,0.0109963994,0.0580346175,-0.1880514324,-0.0189225599,-0.2777186036,-0.0237315167,0.2902144194,-0.256510675,-0.0430307873,0.3490846157,-0.0300245993,0.0151800364,-0.1060819998,0.0208048839,-0.2491456419,-0.0057928399,-0.1321569979,0.2593689263,0.123262465,-0.3601054251,-0.2201318592,0.269644469,-0.0733439028,0.2368864119,-0.1186155677,0.3427739739,-0.1157050282,-0.0664257482,0.2260990441,-0.1223699376,-0.0454593264,-0.1864117533,0.1043651998,0.3241366446,0.2530105114,0.5111225247,0.4740772545,-0.1185918376,-0.1677387506,0.1293274164,0.1104747951,0.1646763533,-0.2225419283,-0.0931732133,0.6572032571,0.3265316486,0.3121195138,0.0560516044,-0.3682840168,-0.1889240742,0.2153066844,0.1976000369,0.1536179483,0.1227065697,0.4277952909,0.3569436371,0.0901561007,0.2018396854,0.3611363173,0.436733067,0.0122092217,-0.0282971486,0.1319124252,-0.4084214866,-0.2810285985,0.3035345972,-0.0571420006,0.6194142103,-0.0001335044,0.3581352234,0.0060083275,0.1161910594,0.1531037837,0.3110364974,0.1907286346,-0.3497518897,-0.1504902393,0.1073554382,-0.0877887234,0.1404762864,-0.2416676432,-0.1674517244,0.1516312212,-0.0311502051,0.2706885934,-0.2101578265,0.2350755781,0.1818676889,-0.1810306013,-0.240647614,-0.2521364987,0.0457765795,-0.0088928426,0.170753926,0.0854838118,-0.0188961383,0.3148322403,-0.3668530285,-0.3901950717,0.1375721544,-0.0929022059,-0.4670560062,0.0459489413,0.4081196785,0.1970107406,0.3088477254,0.0678155348,0.2102432698,-0.0006256719,-0.1477908194,0.0600795187,-0.112185061,0.1130770296,-0.1586785614,-0.3379225731,-0.3725426495,-0.1332513541,-0.0509335175,-0.2197098434,-0.0060843122,-0.3206317127,0.0545925386,-0.0176921636,0.1768091023,0.0380966999,-0.1530308127,0.0944231898,0.1511808634,-0.1855606139,0.1327320933,-0.62751472,-0.0694225803,0.2902387679,-0.4144516289,0.2188192904,-0.1059561744,-0.6133030057,0.1753400266,-0.0784829706,-0.2060005367,0.001224825,0.1281179935,0.2923304439,0.1140855104,-0.1864194721,-0.007726239,-0.0248297248,0.2191719711,0.173627153,0.0818920508,0.3506791294,-0.1272312999,0.0651778728,-0.0971101522,-0.1086509973,-0.021585824,-0.2666110098,0.1335418075,-0.0696092322,0.2099888325,-0.1552820355,0.7983125448,0.2943360209,-0.1711910665,-0.0251084603,-0.1154659018,-0.0341901518,-0.2214478999,0.0346540734,-0.0982867926,0.1160926074,0.1295556277,0.5040053725,0.0420982689,-0.2055519223,-0.2370728552,0.2103705853,-0.1381924897,-0.125543192,0.1617178172,0.0916195735,0.1419389993,0.0423240922,0.2302965075,-0.4181652963,-0.2926326096,0.1091580465,-0.0408109166,-0.2304724753,-0.2813099921,-0.3877126575,-0.2973513305,-0.4328524768,0.3820945323,0.184148103,0.4901478291,0.0667019561,-0.0495630875,0.1984018236,-0.033615768,-0.1258167922,-0.1876072437,-0.2699111402,0.192062363,-0.1328732967,-0.3253301978,0.0336240493,-0.0957036465,0.1680741459,0.1610640287,0.0147240032,-0.1274618804,-0.1373986304,-0.2789736092,0.0331875086,0.0262859482,-0.1226210892,-0.2699660659,-0.0181864891,0.3601047397,0.0197657198,-0.3271099329,0.0494053848,-0.0690035596,0.1861640811,-0.2409917861,-0.0343741477,0.3886441886,0.5452684164,0.2817429304,-0.4422087073,0.3140861392,0.1073643267,-0.2230251133,0.2349049598,0.1926438212,0.1615495235,0.2590863109,0.117810972,-0.0231582858,0.1559976935,0.1223499924,0.0014794386,0.1502127349,-0.2321808636,-0.1356612146,-0.1750068814,0.002425944,0.2685618401,0.24770841,-0.4353989065,-0.3170287907,0.6888301373,0.4300852716,-0.0938147083,0.0343573205,0.0793807432,-0.2249603271,0.0344740525,0.2207663357,0.9849880934,0.0195464771,0.1792137027,0.0678455979,-0.0657152161,0.2816603184,-0.3302939534,-0.1168099716,-0.1556586623,-0.2901778221,-0.0779815391,-0.1615458727,0.2968316078,0.1133745462,-0.0163850486,0.2415135056,0.17051965,-0.1562662572,-0.0078398203,0.3491708934,0.2862446606,-0.0430462025,-0.208005324,0.0738809556,0.1003976911,-0.1788126528,0.093554154,0.0736101717,-0.4495070875,0.2725325227,-0.4372515678,-0.3201458156,0.1095449254,0.2533123195,0.015576981,-0.1597276628,0.1760913879,-0.2382441163,0.243627876,-0.194328025,-0.2094154954,0.1973567754,0.0914246291,0.385846585,0.3977943361,-0.245992437,-0.1021103412,-0.143546626,-0.1216748357,0.2774524689,-0.0152428215,-0.528481245,-0.2779016793,0.1535318941,0.2374778092,-0.0695360079,-0.20802553,0.1414768845,-0.0295072906,0.0897424296,0.0274556633,0.1424980164,-0.0217177868,0.8859041929,-0.1997342706,-0.2523300648,-0.1323381066,0.5498276353,0.09271954,-0.3913935721,0.1513200253,-0.126446113,-0.2121489495,-0.0967690423,0.1890246272,0.1080343798,-0.0423379093,-0.0924679041,-0.4025084078,-0.1561019272,0.322642684,0.2322563082,0.2806727886,-0.1581177413,-0.1462302655,0.16242899,-0.248647809,0.1582395136,0.048861403,0.2107313871,-0.2525596619,-0.1154350713,-0.1653578728,-0.0204238351,-0.1105907187,0.1441789865,0.1088577285,0.044520583,-0.1210544705,0.0115019092,0.2281009257,0.1904027909,-0.164916873,0.252232343,-0.3012043834,-0.0569662936,-0.0036375897,0.2221192718,-0.0319837742,-0.4015704095,-0.0436995961,0.1358124167,-0.0425883643,-0.1925137043,-0.1190372407,-0.3740125299,-0.0797029138,-0.2223741412,0.4970264435,-0.0573125966,-0.098382324,0.2753205895,-0.0541344434,0.1241820306,0.2635473907,0.2435032874,0.037581604,-0.224066928,0.3111835122,0.3066109419,0.246731624,-0.1525162756,0.1661751866,-0.1351745576,0.0642865077,0.326738596,0.1774222255,0.819607079,-0.0564901233,-0.477450639,0.2959795594,0.017035516,0.1751724184,-0.312990278,0.3569063842,0.2060686648,0.0312646814,0.0897602513,0.413036257,0.1084635928,0.2423648089,-0.0868822709,0.5108745694,0.1430851966,0.1881145835,-0.4495840073,-0.0083315056,-0.0644420162,0.1368062496,0.2627196312,0.2217084616,0.248002395,-0.2684145272,0.6961628199,0.4385241866,0.1038485318,-0.1020707935,-0.1935295761,0.2840628922,0.4061124921,0.1730378568,0.2769913077,0.194364503,0.3954158723,-0.2286614776,-0.091996938,-0.1883071214,0.0622647405,0.1141953543,-0.4302681983,0.4157553613,-0.1572193354,0.0111829853,0.0721845478,-0.2298713773,-0.15696989,0.2616047263,-0.0762803182,0.1084914058,-0.0931721479,-0.1474349052,0.0843063593,0.181541428,0.1187275648,0.1750241518,0.1565627754,-0.0186472479,0.3655472398,0.2379738241,0.0692332759,-0.060986314,-0.6656449437,-0.0637142137,-0.332441926,0.2670287788,0.1411351711,0.266456604,-0.2655002177,-0.0878869221,0.1292444468,0.0156821553,0.0900103003,-0.4702937603,-0.1247409359,0.0434832275,-0.3978414536,-0.2820212245,-0.0617983043,0.2348828763,0.0428595655,0.1404763311,-0.0633875206,-0.3730198741,0.0760592744,-0.104985632,-0.0716553181,0.1470345706,-0.0024818862,-0.0942311436,0.6994123459,0.0653570369,0.061892204,-0.0799616575,-0.5321529508,-0.4118786752,0.0667626858,-0.0329950936,-0.1383026838,0.194188714,0.2579255104,-0.1062392518,0.2881212831,0.4280996919,0.5550525188,-0.2165674567,0.0425789803,-0.4298307896,0.0392039716,-0.384167254,-0.2963057458,-0.0905965492,-0.0491233654,0.1268496811,-0.129475981,-0.1205532029,-0.1357968152,-0.1788153797,0.0664918572,0.1710092574,0.5717977285,0.3197728693,0.4699353278,0.3191304803,-0.1181835905,-0.1592604667,0.1190493256,-0.0503680259,0.156670019,-0.0535300635,0.0268940553,-0.17647475,0.3074465692,0.0814195797,0.6996270418,0.1489475071,-0.3488850594,-0.1889693588,-0.2172174156,0.2014439553,0.0730847418,-0.1378266364,0.3830969632,0.1412885338,0.5235023499,-0.1728391796,0.1528406292,0.3073652983,-0.4860737324,-0.0532757565,-0.1688021421,0.0878736004,0.1353642344,-0.0636115298,-0.5485002398,-0.3553925455,0.3390137851,0.0616851337,0.1476610452,0.2592873275,-0.0902005956,0.0979384482,-0.1953221262,0.0274974145,-0.3222491443,0.0253422298,-0.0194342323,-0.0621364415]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1894","title":"benchmarking against MMapIndexedDataset","comments":"no docs haha, it's written to support integer numpy arrays.\r\n\r\nYou can build one in fairseq with, roughly:\r\n```bash\r\n\r\nwget https:\/\/s3.amazonaws.com\/research.metamind.io\/wikitext\/wikitext-103-raw-v1.zip\r\nunzip wikitext-103-raw-v1.zip\r\nexport dd=$HOME\/fairseq-py\/wikitext-103-raw\r\n\r\nexport mm_dir=$HOME\/mmap_wikitext2\r\nmkdir -p gpt2_bpe\r\nwget -O gpt2_bpe\/encoder.json https:\/\/dl.fbaipublicfiles.com\/fairseq\/gpt2_bpe\/encoder.json\r\nwget -O gpt2_bpe\/vocab.bpe https:\/\/dl.fbaipublicfiles.com\/fairseq\/gpt2_bpe\/vocab.bpe\r\nwget -O gpt2_bpe\/dict.txt https:\/\/dl.fbaipublicfiles.com\/fairseq\/gpt2_bpe\/dict.txt\r\nfor SPLIT in train valid; do \\\r\n python -m examples.roberta.multiprocessing_bpe_encoder \\\r\n --encoder-json gpt2_bpe\/encoder.json \\\r\n --vocab-bpe gpt2_bpe\/vocab.bpe \\\r\n --inputs \/scratch\/stories_small\/${SPLIT}.txt \\\r\n --outputs \/scratch\/stories_small\/${SPLIT}.bpe \\\r\n --keep-empty \\\r\n --workers 60; \\\r\ndone\r\n\r\nmkdir -p $mm_dir\r\nfairseq-preprocess \\\r\n --only-source \\\r\n --srcdict gpt2_bpe\/dict.txt \\\r\n --trainpref $dd\/wiki.train.bpe \\\r\n --validpref $dd\/wiki.valid.bpe \\\r\n --destdir $mm_dir \\\r\n --workers 60 \\\r\n --dataset-impl mmap\r\n```\r\n\r\nI'm noticing in my benchmarking that it's much smaller on disk than arrow (200mb vs 900mb), and that both incur significant cost by increasing the number of data loader workers. \r\nThis somewhat old [post](https:\/\/ray-project.github.io\/2017\/10\/15\/fast-python-serialization-with-ray-and-arrow.html) suggests there are some gains to be had from using `pyarrow.serialize(array).tobuffer()`. I haven't yet figured out how much of this stuff `pa.Table` does under the hood.\r\n\r\nThe `MMapIndexedDataset` bottlenecks we are working on improving (by using arrow) are:\r\n1) `MMapIndexedDataset`'s index, which stores offsets, basically gets read in its entirety by each dataloading process.\r\n2) we have separate, identical, `MMapIndexedDatasets` on each dataloading worker, so there's redundancy there; we wonder if there is a way that arrow can somehow dedupe these in shared memory.\r\n\r\nIt will take me a few hours to get `MMapIndexedDataset` benchmarks out of `fairseq`\/onto a branch in this repo, but I'm happy to invest the time if you're interested in collaborating on some performance hacking.","body":"I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB of tokens).\r\n\r\nQuestions:\r\n1) Is this (basically identical) performance expected? \r\n2) Is there a scenario where this library will outperform `MMapIndexedDataset`? (maybe more examples\/larger examples?)\r\n3) Should I be using different benchmarking tools than `psrecord`\/how do you guys do benchmarks?\r\n\r\nThanks in advance! Sam","comment_length":249,"text":"benchmarking against MMapIndexedDataset\nI am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB of tokens).\r\n\r\nQuestions:\r\n1) Is this (basically identical) performance expected? \r\n2) Is there a scenario where this library will outperform `MMapIndexedDataset`? (maybe more examples\/larger examples?)\r\n3) Should I be using different benchmarking tools than `psrecord`\/how do you guys do benchmarks?\r\n\r\nThanks in advance! Sam\nno docs haha, it's written to support integer numpy arrays.\r\n\r\nYou can build one in fairseq with, roughly:\r\n```bash\r\n\r\nwget https:\/\/s3.amazonaws.com\/research.metamind.io\/wikitext\/wikitext-103-raw-v1.zip\r\nunzip wikitext-103-raw-v1.zip\r\nexport dd=$HOME\/fairseq-py\/wikitext-103-raw\r\n\r\nexport mm_dir=$HOME\/mmap_wikitext2\r\nmkdir -p gpt2_bpe\r\nwget -O gpt2_bpe\/encoder.json https:\/\/dl.fbaipublicfiles.com\/fairseq\/gpt2_bpe\/encoder.json\r\nwget -O gpt2_bpe\/vocab.bpe https:\/\/dl.fbaipublicfiles.com\/fairseq\/gpt2_bpe\/vocab.bpe\r\nwget -O gpt2_bpe\/dict.txt https:\/\/dl.fbaipublicfiles.com\/fairseq\/gpt2_bpe\/dict.txt\r\nfor SPLIT in train valid; do \\\r\n python -m examples.roberta.multiprocessing_bpe_encoder \\\r\n --encoder-json gpt2_bpe\/encoder.json \\\r\n --vocab-bpe gpt2_bpe\/vocab.bpe \\\r\n --inputs \/scratch\/stories_small\/${SPLIT}.txt \\\r\n --outputs \/scratch\/stories_small\/${SPLIT}.bpe \\\r\n --keep-empty \\\r\n --workers 60; \\\r\ndone\r\n\r\nmkdir -p $mm_dir\r\nfairseq-preprocess \\\r\n --only-source \\\r\n --srcdict gpt2_bpe\/dict.txt \\\r\n --trainpref $dd\/wiki.train.bpe \\\r\n --validpref $dd\/wiki.valid.bpe \\\r\n --destdir $mm_dir \\\r\n --workers 60 \\\r\n --dataset-impl mmap\r\n```\r\n\r\nI'm noticing in my benchmarking that it's much smaller on disk than arrow (200mb vs 900mb), and that both incur significant cost by increasing the number of data loader workers. \r\nThis somewhat old [post](https:\/\/ray-project.github.io\/2017\/10\/15\/fast-python-serialization-with-ray-and-arrow.html) suggests there are some gains to be had from using `pyarrow.serialize(array).tobuffer()`. I haven't yet figured out how much of this stuff `pa.Table` does under the hood.\r\n\r\nThe `MMapIndexedDataset` bottlenecks we are working on improving (by using arrow) are:\r\n1) `MMapIndexedDataset`'s index, which stores offsets, basically gets read in its entirety by each dataloading process.\r\n2) we have separate, identical, `MMapIndexedDatasets` on each dataloading worker, so there's redundancy there; we wonder if there is a way that arrow can somehow dedupe these in shared memory.\r\n\r\nIt will take me a few hours to get `MMapIndexedDataset` benchmarks out of `fairseq`\/onto a branch in this repo, but I'm happy to invest the time if you're interested in collaborating on some performance hacking.","embeddings":[-0.2432021052,0.023352541,-0.1691768616,0.1820509285,-0.1022186801,-0.0226155818,0.1294504404,0.2604651451,-0.2088269889,-0.0699171424,-0.155699119,0.5033259988,-0.0702501908,-0.4188563228,0.2817451656,-0.0300162956,0.3354423642,0.1515544802,-0.1571365446,0.0486279726,-0.1405944675,0.1506153196,-0.3253050447,0.0204201993,-0.3509258628,0.122545585,-0.0095216176,0.2243740261,-0.0806345567,-0.3892317712,0.2625464201,0.1312130243,-0.0679286122,0.5093360543,-0.0001197014,-0.1623321921,-0.0542118922,0.1078669131,0.0268320143,0.540053606,-0.1131937727,-0.3049707711,0.0386117063,-0.1792382598,0.0806340948,-0.1571841538,-0.1240352243,-0.6968595386,0.0376429297,0.112498045,0.0433407277,0.3202779293,-0.287357986,0.1792372763,0.0780043751,-0.0878257155,-0.2255390435,0.2030549645,0.6038826704,-0.0669367611,-0.0622575395,0.0722511038,-0.1325347275,0.0774309486,0.0730090886,0.0020603244,-0.1156363711,-0.0859563053,0.0608329996,0.4867009521,0.1345411241,-0.3546428084,-0.2642897069,-0.4171714485,-0.3238388896,0.0939705223,0.2001515329,-0.2242229581,-0.045118209,0.1037718952,-0.5512924194,0.2541922927,0.1437454224,-0.0920725763,-0.0109138638,0.4918263257,0.3039819896,0.0423975363,0.4399318397,-0.2640122771,0.1324423403,-0.2057048827,0.0011939199,0.1028615758,-0.4656966925,-0.2803414762,-0.0441006422,-0.250459969,0.1140584946,0.2019072175,-0.1421616822,0.515226841,0.4101502001,0.0859821811,0.075697735,0.2415622324,0.0921152011,0.0314107649,0.0018228889,-0.09552598,-0.2015071213,0.1789547205,0.1996313035,0.0112764891,-0.2609249651,-0.1755254269,-0.3801255226,-0.3332433403,-0.3777097166,0.0147047155,0.3565771878,-0.3525733352,0.0801032111,0.2087020427,0.0533464961,0.2673592567,-0.4570310116,0.0394971967,-0.3195516467,-0.2048750669,-0.0434496328,0.1353322864,-0.0892045274,0.2397130728,0.1953056157,0.1026565209,0.2620453835,-0.0591993704,0.1804069132,0.1563943475,0.5779836774,-0.2525075972,-0.0220910367,-0.0524834134,-0.1771284789,0.1634490192,0.0495934784,0.4309071898,-0.3159751594,0.0107475482,-0.1292083561,-0.1222500205,0.374181211,-0.0066924528,-0.2167480588,-0.301915437,-0.3807257414,0.3237059414,-0.0806941465,-0.2796670496,0.2008896619,-0.0380026847,-0.2599069178,-0.2586502433,0.1129405051,-0.104804866,-0.1829633862,0.0902451426,-0.2681652009,0.1017137095,-0.0574376918,0.4228224456,-0.3233219385,0.2451714128,-0.0476449318,0.2953229547,0.3699166477,-0.4035232663,-0.4890653193,-0.0134465992,0.18437998,0.2378026843,-0.0289858598,0.2881448269,0.2071392089,-0.3476954997,0.0604116172,0.086655207,0.1298025548,0.3404920101,-0.191495046,-0.4372926652,0.4101485014,0.1844226867,-0.2632358372,-0.5926275849,-0.2096512914,-0.220100373,0.0141147422,-0.1993325502,-0.2652760148,-0.0197391007,0.0547744371,-0.3296548724,-0.0481884144,-0.2119642049,-0.0164226443,0.4075005352,-0.3809975088,-0.0671109483,0.410477072,-0.0325264037,0.0160775185,-0.126000762,-0.0352367759,-0.3404727876,0.0188441556,-0.124628529,0.2839566469,0.1574340165,-0.372543931,-0.2051105797,0.2095497102,-0.0340516791,0.0869028717,-0.1760457903,0.2704811096,-0.1397456527,-0.1292769313,0.2784688175,-0.1737473607,-0.0782441348,-0.1303406656,0.1673095971,0.262606591,0.0895880014,0.4840814173,0.36742571,-0.0200807415,-0.1213861853,0.1851388961,0.1798506826,0.1355849355,-0.2332924604,-0.0173051562,0.6673050523,0.158408612,0.3197748661,0.0334665813,-0.3281886876,-0.2300444841,0.2341326475,0.2983640432,0.1582586169,0.1329226196,0.3186577857,0.3469450474,0.0745346472,0.2600978017,0.43803671,0.5110450983,0.0919650421,0.0119462172,0.0950751305,-0.3309466243,-0.3392916918,0.2181062102,-0.1530645341,0.6769615412,0.0056273658,0.3541246653,-0.0272403862,0.0668689162,0.1295981407,0.3824511468,0.2104080319,-0.4022667706,-0.1451212466,0.0917438716,-0.1520504057,0.1312493086,-0.1228138506,-0.138861984,0.1244563162,-0.0374373756,0.227590546,-0.0976811647,0.3413497806,0.1647017747,-0.3163818717,-0.1967754215,-0.2952255309,0.1427739859,-0.0410444699,0.1379371881,0.1017521098,0.0208522659,0.2121637464,-0.2811616361,-0.2880559564,0.0261986479,-0.0989259854,-0.4190021753,0.041986037,0.4446631372,0.1643384844,0.3288487196,0.0893413201,0.0642520711,0.0240686946,-0.1532045007,0.0006648721,-0.1560738534,0.1611467451,-0.0476363078,-0.3093944788,-0.4757262766,-0.2025446296,0.0364159048,-0.2672000527,-0.0642173439,-0.2295372635,0.2396284491,-0.0640902743,0.1067321599,-0.107920602,-0.1804136485,0.064130567,0.0804081485,-0.140421167,0.1453793645,-0.5900020003,0.0552156344,0.3555607498,-0.3177753389,0.1301364899,-0.1445241868,-0.4670259058,0.2165990025,-0.0551492088,-0.1931108981,-0.049044691,0.0232459009,0.2404927164,0.1181667224,-0.1977451593,-0.0716168731,-0.0266945828,0.1901101768,0.1161490977,0.1605773717,0.3275685012,-0.1320878863,0.0220072344,-0.1042530164,0.0397158265,0.0453550145,-0.2836326659,0.2459726781,-0.0552967489,0.1432182044,-0.1069723666,0.8057386279,0.2422968745,-0.2431987375,-0.0703333765,-0.0903486609,-0.0702882782,-0.1221860126,-0.0307255723,0.0003347387,0.1754679233,0.1227691248,0.5753460526,0.0227228627,-0.2618599236,-0.2323166281,0.2291735113,-0.0984151661,-0.1660814881,0.2451482415,0.2188758552,0.0562770963,0.0476948284,0.3273079693,-0.4191041887,-0.3323987424,0.04468951,-0.1760578454,-0.2805054784,-0.2391751558,-0.2002647817,-0.3777946234,-0.4251481295,0.343288064,0.2635824978,0.5489294529,0.1243810207,-0.0968295708,0.2191800773,-0.1522560418,0.065535292,-0.2366161048,-0.2895755768,0.1795128137,-0.0333263502,-0.3219075203,0.0134373605,-0.1971853077,-0.0295849815,0.3278869987,-0.015387659,-0.1670003384,-0.0544806197,-0.209921971,0.0499691665,0.0047979853,0.0105062369,-0.3588391244,-0.1914762706,0.3003986776,0.1187164187,-0.3400736153,0.077657558,0.0090765283,0.1560637653,-0.2711514533,0.0005420088,0.3867950439,0.5396571159,0.2772964239,-0.3942848444,0.2889390886,0.0025328186,-0.1356240213,0.2091222256,0.1200833097,0.1594482362,0.364718914,0.0862561762,-0.1609552801,0.1497536898,0.1717626899,0.0109470775,0.0738902912,-0.2870318294,-0.0327258706,-0.0553467013,0.056418635,0.23353751,0.2024370879,-0.3133951426,-0.3247179687,0.5959884524,0.380327642,-0.0817591622,0.0591212474,-0.1594303399,-0.2124493569,0.0031840277,0.332075268,0.9616857767,-0.0361314341,0.2336978912,0.2174376696,-0.0309352707,0.3985089958,-0.3488456309,-0.0284826308,-0.2169908583,-0.2417110801,-0.07720723,-0.1510023922,0.2551012933,0.1418097913,-0.1065920964,0.29047966,0.3111560941,0.0454586819,-0.0738774389,0.2952062786,0.1969210804,-0.0388143919,-0.1757237762,0.0622950532,-0.0421181172,-0.2161159068,0.1218047962,-0.0506019928,-0.5397326946,0.1987795681,-0.3056119084,-0.2535447478,-0.0059525925,0.4464239776,-0.0064624371,-0.1360346079,0.1693731993,-0.1459734738,0.367300272,-0.2042052448,-0.1845328957,0.0985409692,0.1169233546,0.3119471371,0.4654184282,-0.2429742366,0.0305172633,-0.1376772225,-0.1448931098,0.2899270952,0.024312688,-0.4574985802,-0.3292911649,0.1733497083,0.3085516691,-0.1377101392,-0.1473702788,0.143372938,-0.1439525038,0.0311075393,0.0259093512,0.0657194704,-0.1187400818,0.8549541235,-0.2891100347,-0.2632473409,-0.0914647356,0.4993326962,0.0986840278,-0.4388745427,0.2042728215,-0.0517666899,-0.2537934184,-0.1439123452,0.1367648691,0.0153956767,-0.1277050078,-0.0665010512,-0.1762321442,-0.1240036115,0.3249712288,0.1699852496,0.2428448796,-0.0837258175,-0.1774937809,0.2468436956,-0.328785032,0.0891061276,0.2111873776,0.3494646251,-0.2351031303,-0.1996961385,-0.2085604519,-0.1003244147,-0.1379260421,0.1160773113,0.0410959236,0.0242041647,-0.2370050848,-0.0795884281,0.1163436621,0.2017842233,-0.1304215044,0.2277389616,-0.3855400085,-0.0815220922,0.068775326,0.1902280301,-0.1212153286,-0.4181318283,-0.0599127114,0.1629431993,-0.0284064189,-0.2897573411,-0.0809319541,-0.4819076061,-0.0330374204,-0.1748729944,0.5092713237,-0.0284544993,-0.0630905777,0.3277779818,-0.2139818072,0.1407270283,0.1937383413,0.2686721385,0.0791731924,-0.2487015426,0.3357217312,0.2605231106,0.1490178257,-0.036126025,0.1514648497,-0.1979037374,-0.0279117972,0.2792246044,0.1092881188,0.8128835559,-0.1213120744,-0.595482409,0.2628180385,0.0550585426,0.070903413,-0.2267865688,0.4238925278,0.1857876927,0.0937285721,0.2167644352,0.4243469536,-0.0296487194,0.0790285692,-0.0533586442,0.5252417922,0.05597727,0.0300830808,-0.164886713,0.0284829699,-0.1760407537,-0.0447766334,0.1498751789,0.2445606142,0.2723895311,-0.3558055758,0.6884691715,0.3570601344,0.1591429412,-0.2962267399,-0.007938562,0.2951882482,0.5422820449,0.1094152853,0.2572138906,0.2081545293,0.3968319893,-0.2889023721,-0.0473377518,-0.1906132102,0.1498117745,0.1400303096,-0.3140506446,0.2555698156,-0.1108250245,-0.030933911,0.0641805977,-0.2243410498,-0.1642580777,0.2001765221,0.0384884514,0.0602515116,-0.0766350254,-0.0091736298,0.0860621557,0.251047343,0.0271171797,0.1162607446,0.3386173844,0.0252006166,0.3081500232,0.2748593688,0.1436080486,-0.0221874826,-0.6469067931,-0.0227144174,-0.2772479057,0.2943775356,0.0195346363,0.1883046627,-0.3704979122,-0.0827376917,0.2145932168,0.003111982,0.0610564761,-0.5787854791,-0.0727083907,-0.0034162027,-0.2699400187,-0.1305489987,-0.0787749365,0.1848434359,0.0313293822,0.1325552166,-0.1934871823,-0.3762678206,-0.0294351354,-0.0934248567,0.0093004145,0.236623913,0.0202467721,-0.1909415871,0.7200247049,0.0936730728,-0.0714851469,-0.1348982006,-0.4660586119,-0.4094492495,-0.0068373745,-0.1019363329,-0.2489138246,0.1159808934,0.2929784358,-0.1527727395,0.3135237694,0.3988290131,0.6963889003,-0.2271549404,0.1801040918,-0.536499083,0.1016406342,-0.3078744113,-0.2435192615,-0.0656288192,-0.1456205249,0.121539019,-0.1681292653,-0.1349673122,-0.0808500499,-0.0165278874,0.0096097169,0.1161242053,0.4607246518,0.3132439852,0.3438243866,0.2764532566,-0.1613562256,-0.1428869367,0.1805913895,-0.0459165238,0.1733429283,0.0033166863,0.1465429068,-0.1652986556,0.1458811611,0.0261309687,0.7369502187,0.1684111059,-0.2778198123,-0.1064319015,-0.2087538242,0.1732776016,0.118063435,-0.1529623419,0.3568540812,0.171987325,0.5304063559,-0.190996781,-0.0618096255,0.3194054663,-0.4830409586,-0.1014674902,-0.2315475345,0.0702895001,0.0974435136,0.0605839379,-0.5089956522,-0.3161846697,0.3136837482,0.1816224307,0.0581687689,0.2628867626,-0.1175676957,0.0733460113,-0.2202983797,-0.077541545,-0.3205032945,0.0767266527,-0.0646989122,-0.0776448324]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1893","title":"wmt19 is broken","comments":"This was also mentioned in https:\/\/github.com\/huggingface\/datasets\/issues\/488 \r\n\r\nThe bucket where is data was stored seems to be unavailable now. Maybe we can change the URL to the ones in https:\/\/conferences.unite.un.org\/uncorpus\/en\/downloadoverview ?","body":"1. Check which lang pairs we have: `--dataset_name wmt19`:\r\n\r\nPlease pick one among the available configs: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']\r\n\r\n \r\n2. OK, let's pick `ru-en`:\r\n\r\n`--dataset_name wmt19 --dataset_config \"ru-en\"`\r\n\r\nno cookies:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/run_seq2seq.py\", line 661, in \r\n main()\r\n File \".\/run_seq2seq.py\", line 317, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 628, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt19\/436092de5f3faaf0fc28bc84875475b384e90a5470fa6afaee11039ceddc5052\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 276, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 191, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 233, in map_nested\r\n mapped = [\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 234, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 190, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 190, in \r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 172, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 211, in _download\r\n return cached_path(url_or_filename, download_config=download_config)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/file_utils.py\", line 274, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/file_utils.py\", line 584, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-ru.tar.gz\r\n```","comment_length":30,"text":"wmt19 is broken\n1. Check which lang pairs we have: `--dataset_name wmt19`:\r\n\r\nPlease pick one among the available configs: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']\r\n\r\n \r\n2. OK, let's pick `ru-en`:\r\n\r\n`--dataset_name wmt19 --dataset_config \"ru-en\"`\r\n\r\nno cookies:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/run_seq2seq.py\", line 661, in \r\n main()\r\n File \".\/run_seq2seq.py\", line 317, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 628, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt19\/436092de5f3faaf0fc28bc84875475b384e90a5470fa6afaee11039ceddc5052\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 276, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 191, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 233, in map_nested\r\n mapped = [\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 234, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 190, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 190, in \r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 172, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 211, in _download\r\n return cached_path(url_or_filename, download_config=download_config)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/file_utils.py\", line 274, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/file_utils.py\", line 584, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-ru.tar.gz\r\n```\nThis was also mentioned in https:\/\/github.com\/huggingface\/datasets\/issues\/488 \r\n\r\nThe bucket where is data was stored seems to be unavailable now. Maybe we can change the URL to the ones in https:\/\/conferences.unite.un.org\/uncorpus\/en\/downloadoverview ?","embeddings":[-0.3531678915,-0.4596051574,-0.041260805,0.3225629628,0.1783245504,-0.0362047665,0.1937312186,0.2335278839,0.0660974607,0.0606783777,0.0147966258,0.1090325788,-0.1255616546,0.5707479715,0.1197273284,-0.0096024554,0.006727763,0.0419421494,-0.4998805523,0.0689936355,-0.1902456582,0.3333269358,-0.240805164,0.0314803496,-0.1947323233,0.0247913152,-0.1083271578,0.1399192214,-0.2531927824,-0.2184594125,0.1407962143,0.03599016,-0.0891286358,0.4259700477,-0.0001029476,-0.0148938447,0.2780917883,0.0244727302,-0.2462794483,-0.2544905841,-0.3194821775,-0.2060382217,0.0056866487,-0.00197129,-0.0526040867,-0.2070801854,-0.1086709499,-0.1944910437,0.5796175003,0.2073678076,0.3498454392,0.3110418916,0.0414667688,-0.3367854059,0.1314578205,0.1913625002,-0.075214833,-0.0526342764,0.1527308673,0.0621753149,0.0765529945,0.345037818,0.0654329211,-0.1494609565,0.2028799951,-0.1226101592,0.0630022287,-0.1020159349,0.2366117835,0.2687706947,0.3145791888,-0.1319089383,-0.157293722,0.1063634828,-0.2688269913,-0.6988610625,0.1989784688,0.1747250706,-0.0829618424,0.1859146655,-0.0768822357,-0.1307185888,-0.0139767723,-0.114756003,-0.2573106289,0.6457993984,-0.1298800111,0.0660456195,0.0168189295,-0.2135042101,-0.1319770068,0.3256309628,-0.0784671456,-0.2144065797,-0.1820536107,0.182136178,0.0829403177,0.1146459728,0.1386961937,-0.0364187658,-0.1732154489,0.2740972042,0.1333288997,0.0846897662,-0.037811555,0.3722313344,0.1994921565,-0.1968185753,0.2560356557,0.2929574251,-0.0020001449,0.0089129545,0.0502564311,-0.5564665794,-0.0951240286,0.0572429746,0.093136996,-0.0478079356,-0.1790327877,0.0198320504,0.0519164056,-0.1261248142,0.0150336111,0.4379251003,0.0404039361,-0.0652516857,0.0184146632,0.1667163819,-0.1312855929,-0.222632736,-0.2503516674,-0.1263069808,-0.2839661837,0.1710803658,0.161487937,-0.1287928522,0.3098284304,0.059067037,-0.2042987943,-0.1414279342,-0.1652156562,0.0173629746,0.1748155355,0.2738702893,-0.0161503609,0.2083771825,0.2654596269,-0.2715440392,-0.0453215539,-0.0837185457,-0.126306355,-0.1878836304,-0.0786041543,0.332367748,0.0418706909,0.1754484922,-0.1218371466,0.20036982,-0.0080903638,-0.0647755265,-0.0314939655,-0.1032682657,-0.4300821424,-0.1006188989,0.4257931709,0.4848943353,-0.5752804875,-0.2235403508,-0.0691839606,-0.1647978127,0.0550692901,0.2180525661,-0.0357789248,-0.0072339587,-0.3086130023,0.1856991351,-0.2164037377,-0.1183210537,-0.0606821328,-0.011784601,-0.1158541739,-0.0222994089,0.0701588169,-0.2123704106,-0.0756179765,0.020681439,0.0376714058,0.0554947928,0.0734893084,-0.1147501096,-0.2230088115,-0.4322020411,0.531208992,-0.034019731,0.1467173398,-0.3156560659,0.169720158,-0.1503376067,0.3136665523,-0.0254015904,0.0116742151,0.0524743386,0.2096423656,0.0273616482,0.0132029941,-0.0901462957,-0.5520675778,0.2051399499,-0.0901376158,0.3927583992,-0.212714076,0.0717063025,-0.2444405109,0.0683690831,-0.3136653006,-0.3684312701,0.2931483984,0.1580297351,-0.1629101783,-0.1037695035,-0.0295058955,-0.019819025,0.0508935116,0.1874933988,-0.4635505974,0.0206466895,-0.2454169989,-0.1044289246,0.156505987,0.1352606118,0.3007023931,-0.0551719964,-0.1239377335,0.3911254406,0.0701720193,0.5023868084,-0.1645805538,-0.1782219857,0.249526009,-0.6714123487,-0.0299860816,-0.0705696866,-0.0209423117,-0.1671116948,-0.0189833883,0.1479646415,-0.1536420584,0.2499715835,0.2225042433,0.1215657592,0.2497499287,-0.0424224213,-0.2755948007,-0.3043380976,0.1753047556,-0.1138344705,-0.0894696191,0.0540231355,0.0358070731,0.0644626617,0.3603795767,0.1643043011,0.1321129352,-0.043757055,-0.2861381173,-0.0584824234,0.0737614557,0.1758922786,0.2659896314,0.135631606,-0.0020136386,0.2280427963,0.1120297983,-0.267542392,0.346276015,-0.0811365917,0.0762071684,0.2666009665,-0.0454606488,-0.1460240334,-0.3090715408,0.2241084129,-0.1355576515,0.0238552969,-0.2537421882,-0.074804537,-0.5700973272,-0.1503544897,-0.3576549292,-0.2965597212,-0.3699661195,-0.2056931853,-0.0746994689,-0.2369470596,-0.2057539523,0.3813068867,0.2306605577,-0.0097506074,-0.0720566809,0.2905026376,-0.1970252395,-0.0966326296,-0.0782292858,0.1418487877,0.2689707577,0.4118449688,0.3864142001,-0.3926121294,-0.0617833138,-0.0894832015,-0.4258926511,0.1954809129,-0.1968224645,0.5000601411,0.2598654628,0.0435456745,0.003262074,-0.3089444339,0.5313377976,-0.3635278642,0.0224792399,-0.0262083299,0.0161467921,-0.1683609337,-0.27094841,-0.4824743867,-0.3764455318,-0.3136797547,0.1400423348,0.0832739696,0.1126022264,0.5596601963,-0.0394125246,0.1088424399,-0.2227118462,0.2569581866,-0.2365367562,-0.1614936739,0.137923345,-0.3226335943,-0.3150461018,0.1665690243,0.1618140042,0.5677848458,-0.0527581684,-0.3567388058,0.1910223067,-0.0590969175,0.0631303266,0.1176173165,0.1084913611,0.2975240946,0.2475846559,-0.1623527855,-0.1300195754,-0.443903178,0.1164973229,-0.2083269507,0.1733241528,-0.1027667895,0.2727052867,-0.0579380058,0.5740627646,0.0900356397,0.1567035466,0.380767554,0.0893212259,0.344758004,0.0577331036,-0.4066468179,0.1074610129,-0.1301034838,0.0447860137,0.4078933299,-0.0203962512,0.2605615854,-0.3096646369,-0.0563562438,-0.3117498159,-0.3440133333,-0.1111934036,0.3289459646,0.2659176588,-0.0373780318,0.0181176476,0.0653881878,0.0588199198,0.1717750132,0.2928853035,-0.0539857633,0.0369721651,-0.2738019526,-0.276831001,-0.3234131038,0.1019824222,-0.0129103325,0.314891398,-0.1771434397,-0.1706010699,-0.115519315,-0.0268453434,0.4513019919,0.0661667511,0.1779454648,0.0517488196,-0.0555670075,-0.1506130099,-0.0096938051,-0.158163324,-0.2395304441,0.0718367323,0.5307104588,-0.301789999,-0.1330113709,-0.0683953837,0.3462325931,-0.2519260943,-0.1010977849,-0.051916413,-0.3258300126,-0.2018239349,-0.1577467024,0.0122022359,0.0089348461,-0.2125344872,0.0324793644,0.0873223841,-0.0878785029,0.0805717483,0.1848243326,0.3329838812,-0.0717549771,0.1573340297,0.1696560085,0.3938362896,0.3397711217,0.4407480955,-0.1332304031,0.0009336532,-0.0843146294,-0.0430407338,0.2639267743,0.3094120324,-0.215477109,0.2849526405,0.1207702085,0.4189075828,-0.3216806948,-0.0813540146,0.4164524376,0.1362739354,-0.2010646164,-0.1323176026,0.0803623572,0.0103407558,0.031541504,-0.1423971802,0.1290774345,-0.3507485092,0.0609088875,0.078151755,0.827927947,0.1066750288,0.0505959168,0.2272731066,0.0661897212,0.32757622,0.1382712573,-0.1211392954,-0.2004662007,-0.3317951858,0.0248154569,-0.099222362,0.1703279912,-0.0614861883,-0.5772724152,0.3057090044,0.0568097457,0.2166914493,-0.0032007033,0.1019637585,-0.1573070437,-0.2348063141,-0.1303106695,0.2480469346,-0.1064357087,0.0918127447,-0.1769693643,-0.0333318971,0.2297402024,-0.1552753597,-0.2749805748,0.0140248826,-0.4671751857,-0.0161021519,-0.0833164677,0.0032322081,0.4410224259,0.0363410376,0.0427438356,0.4214236736,-0.1386109143,0.0637238547,-0.0429337136,0.1349081844,-0.0841732323,-0.0047630817,0.3453175128,-0.1905305684,-0.1987347156,0.0011532097,-0.0088481866,0.2760788798,-0.128066361,-0.1807636619,0.0528138243,-0.0886569396,0.0750101507,-0.0534617491,0.1075574532,-0.3861637115,0.2380411774,-0.0373641886,-0.2498450875,0.2838935256,0.0724158138,-0.1974202245,-0.2185720056,0.1967156976,-0.0967931375,0.0722089186,0.4667076766,0.6660541892,-0.1134741083,-0.3804566264,-0.0415860973,0.0729243234,-0.4744254649,0.0584273413,0.2235012203,0.0947011188,0.0354972444,0.5245188475,0.2763046026,-0.1558062732,0.0191001408,-0.2525489926,-0.2125600725,0.2726362944,0.1221406013,0.1264694631,0.0170031767,0.0855208859,-0.0519933999,0.4380584657,-0.4479949772,0.0722157061,-0.0592745207,0.1075921878,-0.0114648519,-0.1288196743,0.1122962758,-0.0021178836,0.1891260445,0.1201670468,-0.2339137346,-0.3181345463,-0.2554555237,0.0638008565,0.0222388711,-0.3724784553,0.1332535595,0.0038004601,0.1911079884,-0.0588761866,0.1395943612,0.2633418739,0.0707135946,0.2352755517,0.3863000572,0.1705931127,-0.0358113386,0.2202048004,0.229640305,0.1827299893,-0.120643653,0.0040958808,-0.1112098023,-0.1543441564,0.2066453993,0.0368227065,0.0823291615,0.1238961667,0.2084758133,-0.2427893728,-0.1854782701,0.0495932698,0.6940181851,0.3251621723,-0.0039949967,0.1033432409,0.2818107009,0.3300274312,-0.4054022729,0.2187286466,0.0474143773,-0.0549256802,0.1595040262,0.0900707692,0.1510621011,-0.2287932187,0.0520993397,0.1556001604,-0.0261918847,-0.1324225515,0.2067637146,0.3938092887,-0.1332139224,0.1219130605,0.2359653413,0.1771038175,0.1712319404,0.1749379039,-0.1229380146,0.2797958851,-0.1340799034,0.2047886699,0.2102097422,-0.1498601586,-0.0360555723,0.2317603976,-0.172191456,-0.2055539191,-0.1639109403,0.7347185016,-0.1869656444,-0.2501268685,-0.1461642683,0.4038397968,-0.1279132813,-0.2281083316,-0.0500654727,-0.1608845145,-0.402636081,-0.0049357046,-0.063475877,-0.1126711592,-0.1450135708,0.2226958275,-0.0667210594,-0.2337391973,-0.3430245817,-0.0320747867,-0.1722007245,-0.2374601215,0.2469195276,0.4927988052,-0.039252162,-0.0443843864,0.2488892376,0.3715668321,0.2062989175,-0.0362031683,0.1271968037,0.0061100456,-0.0001468372,-0.238723591,0.2048417926,0.0573939048,-0.0297745559,0.3240711093,0.276162982,-0.3107548952,-0.1467660069,-0.0071531287,0.0752686858,-0.1834906042,0.0840032771,-0.2293253839,-0.0188462306,-0.2724256217,-0.0399923958,-0.4678646922,0.29306072,0.4881744087,-0.0980252028,0.2136202306,-0.3128412664,0.1499489248,-0.009569481,0.5410996675,0.344787091,0.0321860872,-0.3470342457,-0.3034185767,-0.7088698745,0.1239585206,0.1814087182,0.3573852777,-0.2041451782,-0.1512265801,0.1522311866,0.0626366511,0.0859171003,0.2302500457,-0.0085127698,-0.1255458295,-0.2412825376,0.0564879924,-0.0381073989,-0.0573259145,0.0210198257,-0.054989107,0.2110695094,-0.2144033313,0.1125235185,-0.3159071505,0.0951716825,-0.0801588818,0.0805482343,0.1205660701,0.2367544323,0.1540481597,-0.2392389625,-0.057379093,-0.1824198961,-0.2834384739,-0.1731160134,0.338611722,-0.0379037187,0.2053391039,0.008053802,0.0012945826,-0.3797041774,0.1562884301,0.0118329693,-0.1514143795,-0.2223190367,0.2561254501,-0.0427351184,-0.2844480276,0.3420321047,0.2845244408,-0.0065560443,0.2018401772,-0.2968755364,-0.3999528885,0.40293625,-0.1829158515,-0.1783638,-0.2296224684,0.0490622744,0.1954258531,-0.1067993939,-0.1924477071,0.0228688605,0.1301557571,0.0905380249,-0.0717682615,0.215465188,-0.1068774462,-0.1212995872,-0.0754636526,0.248825416,0.533864975,-0.2447043657,-0.0007777357,-0.0159923304]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1893","title":"wmt19 is broken","comments":"Closing since this has been fixed by #1912","body":"1. Check which lang pairs we have: `--dataset_name wmt19`:\r\n\r\nPlease pick one among the available configs: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']\r\n\r\n \r\n2. OK, let's pick `ru-en`:\r\n\r\n`--dataset_name wmt19 --dataset_config \"ru-en\"`\r\n\r\nno cookies:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/run_seq2seq.py\", line 661, in \r\n main()\r\n File \".\/run_seq2seq.py\", line 317, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 628, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt19\/436092de5f3faaf0fc28bc84875475b384e90a5470fa6afaee11039ceddc5052\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 276, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 191, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 233, in map_nested\r\n mapped = [\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 234, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 190, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 190, in \r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 172, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 211, in _download\r\n return cached_path(url_or_filename, download_config=download_config)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/file_utils.py\", line 274, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/file_utils.py\", line 584, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-ru.tar.gz\r\n```","comment_length":8,"text":"wmt19 is broken\n1. Check which lang pairs we have: `--dataset_name wmt19`:\r\n\r\nPlease pick one among the available configs: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']\r\n\r\n \r\n2. OK, let's pick `ru-en`:\r\n\r\n`--dataset_name wmt19 --dataset_config \"ru-en\"`\r\n\r\nno cookies:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/run_seq2seq.py\", line 661, in \r\n main()\r\n File \".\/run_seq2seq.py\", line 317, in main\r\n datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/load.py\", line 740, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 572, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/builder.py\", line 628, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/stas\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt19\/436092de5f3faaf0fc28bc84875475b384e90a5470fa6afaee11039ceddc5052\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 276, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 191, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 233, in map_nested\r\n mapped = [\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 234, in \r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 190, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 190, in \r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/py_utils.py\", line 172, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/download_manager.py\", line 211, in _download\r\n return cached_path(url_or_filename, download_config=download_config)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/file_utils.py\", line 274, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/mnt\/nvme1\/code\/huggingface\/datasets-master\/src\/datasets\/utils\/file_utils.py\", line 584, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-ru.tar.gz\r\n```\nClosing since this has been fixed by #1912","embeddings":[-0.3531678915,-0.4596051574,-0.041260805,0.3225629628,0.1783245504,-0.0362047665,0.1937312186,0.2335278839,0.0660974607,0.0606783777,0.0147966258,0.1090325788,-0.1255616546,0.5707479715,0.1197273284,-0.0096024554,0.006727763,0.0419421494,-0.4998805523,0.0689936355,-0.1902456582,0.3333269358,-0.240805164,0.0314803496,-0.1947323233,0.0247913152,-0.1083271578,0.1399192214,-0.2531927824,-0.2184594125,0.1407962143,0.03599016,-0.0891286358,0.4259700477,-0.0001029476,-0.0148938447,0.2780917883,0.0244727302,-0.2462794483,-0.2544905841,-0.3194821775,-0.2060382217,0.0056866487,-0.00197129,-0.0526040867,-0.2070801854,-0.1086709499,-0.1944910437,0.5796175003,0.2073678076,0.3498454392,0.3110418916,0.0414667688,-0.3367854059,0.1314578205,0.1913625002,-0.075214833,-0.0526342764,0.1527308673,0.0621753149,0.0765529945,0.345037818,0.0654329211,-0.1494609565,0.2028799951,-0.1226101592,0.0630022287,-0.1020159349,0.2366117835,0.2687706947,0.3145791888,-0.1319089383,-0.157293722,0.1063634828,-0.2688269913,-0.6988610625,0.1989784688,0.1747250706,-0.0829618424,0.1859146655,-0.0768822357,-0.1307185888,-0.0139767723,-0.114756003,-0.2573106289,0.6457993984,-0.1298800111,0.0660456195,0.0168189295,-0.2135042101,-0.1319770068,0.3256309628,-0.0784671456,-0.2144065797,-0.1820536107,0.182136178,0.0829403177,0.1146459728,0.1386961937,-0.0364187658,-0.1732154489,0.2740972042,0.1333288997,0.0846897662,-0.037811555,0.3722313344,0.1994921565,-0.1968185753,0.2560356557,0.2929574251,-0.0020001449,0.0089129545,0.0502564311,-0.5564665794,-0.0951240286,0.0572429746,0.093136996,-0.0478079356,-0.1790327877,0.0198320504,0.0519164056,-0.1261248142,0.0150336111,0.4379251003,0.0404039361,-0.0652516857,0.0184146632,0.1667163819,-0.1312855929,-0.222632736,-0.2503516674,-0.1263069808,-0.2839661837,0.1710803658,0.161487937,-0.1287928522,0.3098284304,0.059067037,-0.2042987943,-0.1414279342,-0.1652156562,0.0173629746,0.1748155355,0.2738702893,-0.0161503609,0.2083771825,0.2654596269,-0.2715440392,-0.0453215539,-0.0837185457,-0.126306355,-0.1878836304,-0.0786041543,0.332367748,0.0418706909,0.1754484922,-0.1218371466,0.20036982,-0.0080903638,-0.0647755265,-0.0314939655,-0.1032682657,-0.4300821424,-0.1006188989,0.4257931709,0.4848943353,-0.5752804875,-0.2235403508,-0.0691839606,-0.1647978127,0.0550692901,0.2180525661,-0.0357789248,-0.0072339587,-0.3086130023,0.1856991351,-0.2164037377,-0.1183210537,-0.0606821328,-0.011784601,-0.1158541739,-0.0222994089,0.0701588169,-0.2123704106,-0.0756179765,0.020681439,0.0376714058,0.0554947928,0.0734893084,-0.1147501096,-0.2230088115,-0.4322020411,0.531208992,-0.034019731,0.1467173398,-0.3156560659,0.169720158,-0.1503376067,0.3136665523,-0.0254015904,0.0116742151,0.0524743386,0.2096423656,0.0273616482,0.0132029941,-0.0901462957,-0.5520675778,0.2051399499,-0.0901376158,0.3927583992,-0.212714076,0.0717063025,-0.2444405109,0.0683690831,-0.3136653006,-0.3684312701,0.2931483984,0.1580297351,-0.1629101783,-0.1037695035,-0.0295058955,-0.019819025,0.0508935116,0.1874933988,-0.4635505974,0.0206466895,-0.2454169989,-0.1044289246,0.156505987,0.1352606118,0.3007023931,-0.0551719964,-0.1239377335,0.3911254406,0.0701720193,0.5023868084,-0.1645805538,-0.1782219857,0.249526009,-0.6714123487,-0.0299860816,-0.0705696866,-0.0209423117,-0.1671116948,-0.0189833883,0.1479646415,-0.1536420584,0.2499715835,0.2225042433,0.1215657592,0.2497499287,-0.0424224213,-0.2755948007,-0.3043380976,0.1753047556,-0.1138344705,-0.0894696191,0.0540231355,0.0358070731,0.0644626617,0.3603795767,0.1643043011,0.1321129352,-0.043757055,-0.2861381173,-0.0584824234,0.0737614557,0.1758922786,0.2659896314,0.135631606,-0.0020136386,0.2280427963,0.1120297983,-0.267542392,0.346276015,-0.0811365917,0.0762071684,0.2666009665,-0.0454606488,-0.1460240334,-0.3090715408,0.2241084129,-0.1355576515,0.0238552969,-0.2537421882,-0.074804537,-0.5700973272,-0.1503544897,-0.3576549292,-0.2965597212,-0.3699661195,-0.2056931853,-0.0746994689,-0.2369470596,-0.2057539523,0.3813068867,0.2306605577,-0.0097506074,-0.0720566809,0.2905026376,-0.1970252395,-0.0966326296,-0.0782292858,0.1418487877,0.2689707577,0.4118449688,0.3864142001,-0.3926121294,-0.0617833138,-0.0894832015,-0.4258926511,0.1954809129,-0.1968224645,0.5000601411,0.2598654628,0.0435456745,0.003262074,-0.3089444339,0.5313377976,-0.3635278642,0.0224792399,-0.0262083299,0.0161467921,-0.1683609337,-0.27094841,-0.4824743867,-0.3764455318,-0.3136797547,0.1400423348,0.0832739696,0.1126022264,0.5596601963,-0.0394125246,0.1088424399,-0.2227118462,0.2569581866,-0.2365367562,-0.1614936739,0.137923345,-0.3226335943,-0.3150461018,0.1665690243,0.1618140042,0.5677848458,-0.0527581684,-0.3567388058,0.1910223067,-0.0590969175,0.0631303266,0.1176173165,0.1084913611,0.2975240946,0.2475846559,-0.1623527855,-0.1300195754,-0.443903178,0.1164973229,-0.2083269507,0.1733241528,-0.1027667895,0.2727052867,-0.0579380058,0.5740627646,0.0900356397,0.1567035466,0.380767554,0.0893212259,0.344758004,0.0577331036,-0.4066468179,0.1074610129,-0.1301034838,0.0447860137,0.4078933299,-0.0203962512,0.2605615854,-0.3096646369,-0.0563562438,-0.3117498159,-0.3440133333,-0.1111934036,0.3289459646,0.2659176588,-0.0373780318,0.0181176476,0.0653881878,0.0588199198,0.1717750132,0.2928853035,-0.0539857633,0.0369721651,-0.2738019526,-0.276831001,-0.3234131038,0.1019824222,-0.0129103325,0.314891398,-0.1771434397,-0.1706010699,-0.115519315,-0.0268453434,0.4513019919,0.0661667511,0.1779454648,0.0517488196,-0.0555670075,-0.1506130099,-0.0096938051,-0.158163324,-0.2395304441,0.0718367323,0.5307104588,-0.301789999,-0.1330113709,-0.0683953837,0.3462325931,-0.2519260943,-0.1010977849,-0.051916413,-0.3258300126,-0.2018239349,-0.1577467024,0.0122022359,0.0089348461,-0.2125344872,0.0324793644,0.0873223841,-0.0878785029,0.0805717483,0.1848243326,0.3329838812,-0.0717549771,0.1573340297,0.1696560085,0.3938362896,0.3397711217,0.4407480955,-0.1332304031,0.0009336532,-0.0843146294,-0.0430407338,0.2639267743,0.3094120324,-0.215477109,0.2849526405,0.1207702085,0.4189075828,-0.3216806948,-0.0813540146,0.4164524376,0.1362739354,-0.2010646164,-0.1323176026,0.0803623572,0.0103407558,0.031541504,-0.1423971802,0.1290774345,-0.3507485092,0.0609088875,0.078151755,0.827927947,0.1066750288,0.0505959168,0.2272731066,0.0661897212,0.32757622,0.1382712573,-0.1211392954,-0.2004662007,-0.3317951858,0.0248154569,-0.099222362,0.1703279912,-0.0614861883,-0.5772724152,0.3057090044,0.0568097457,0.2166914493,-0.0032007033,0.1019637585,-0.1573070437,-0.2348063141,-0.1303106695,0.2480469346,-0.1064357087,0.0918127447,-0.1769693643,-0.0333318971,0.2297402024,-0.1552753597,-0.2749805748,0.0140248826,-0.4671751857,-0.0161021519,-0.0833164677,0.0032322081,0.4410224259,0.0363410376,0.0427438356,0.4214236736,-0.1386109143,0.0637238547,-0.0429337136,0.1349081844,-0.0841732323,-0.0047630817,0.3453175128,-0.1905305684,-0.1987347156,0.0011532097,-0.0088481866,0.2760788798,-0.128066361,-0.1807636619,0.0528138243,-0.0886569396,0.0750101507,-0.0534617491,0.1075574532,-0.3861637115,0.2380411774,-0.0373641886,-0.2498450875,0.2838935256,0.0724158138,-0.1974202245,-0.2185720056,0.1967156976,-0.0967931375,0.0722089186,0.4667076766,0.6660541892,-0.1134741083,-0.3804566264,-0.0415860973,0.0729243234,-0.4744254649,0.0584273413,0.2235012203,0.0947011188,0.0354972444,0.5245188475,0.2763046026,-0.1558062732,0.0191001408,-0.2525489926,-0.2125600725,0.2726362944,0.1221406013,0.1264694631,0.0170031767,0.0855208859,-0.0519933999,0.4380584657,-0.4479949772,0.0722157061,-0.0592745207,0.1075921878,-0.0114648519,-0.1288196743,0.1122962758,-0.0021178836,0.1891260445,0.1201670468,-0.2339137346,-0.3181345463,-0.2554555237,0.0638008565,0.0222388711,-0.3724784553,0.1332535595,0.0038004601,0.1911079884,-0.0588761866,0.1395943612,0.2633418739,0.0707135946,0.2352755517,0.3863000572,0.1705931127,-0.0358113386,0.2202048004,0.229640305,0.1827299893,-0.120643653,0.0040958808,-0.1112098023,-0.1543441564,0.2066453993,0.0368227065,0.0823291615,0.1238961667,0.2084758133,-0.2427893728,-0.1854782701,0.0495932698,0.6940181851,0.3251621723,-0.0039949967,0.1033432409,0.2818107009,0.3300274312,-0.4054022729,0.2187286466,0.0474143773,-0.0549256802,0.1595040262,0.0900707692,0.1510621011,-0.2287932187,0.0520993397,0.1556001604,-0.0261918847,-0.1324225515,0.2067637146,0.3938092887,-0.1332139224,0.1219130605,0.2359653413,0.1771038175,0.1712319404,0.1749379039,-0.1229380146,0.2797958851,-0.1340799034,0.2047886699,0.2102097422,-0.1498601586,-0.0360555723,0.2317603976,-0.172191456,-0.2055539191,-0.1639109403,0.7347185016,-0.1869656444,-0.2501268685,-0.1461642683,0.4038397968,-0.1279132813,-0.2281083316,-0.0500654727,-0.1608845145,-0.402636081,-0.0049357046,-0.063475877,-0.1126711592,-0.1450135708,0.2226958275,-0.0667210594,-0.2337391973,-0.3430245817,-0.0320747867,-0.1722007245,-0.2374601215,0.2469195276,0.4927988052,-0.039252162,-0.0443843864,0.2488892376,0.3715668321,0.2062989175,-0.0362031683,0.1271968037,0.0061100456,-0.0001468372,-0.238723591,0.2048417926,0.0573939048,-0.0297745559,0.3240711093,0.276162982,-0.3107548952,-0.1467660069,-0.0071531287,0.0752686858,-0.1834906042,0.0840032771,-0.2293253839,-0.0188462306,-0.2724256217,-0.0399923958,-0.4678646922,0.29306072,0.4881744087,-0.0980252028,0.2136202306,-0.3128412664,0.1499489248,-0.009569481,0.5410996675,0.344787091,0.0321860872,-0.3470342457,-0.3034185767,-0.7088698745,0.1239585206,0.1814087182,0.3573852777,-0.2041451782,-0.1512265801,0.1522311866,0.0626366511,0.0859171003,0.2302500457,-0.0085127698,-0.1255458295,-0.2412825376,0.0564879924,-0.0381073989,-0.0573259145,0.0210198257,-0.054989107,0.2110695094,-0.2144033313,0.1125235185,-0.3159071505,0.0951716825,-0.0801588818,0.0805482343,0.1205660701,0.2367544323,0.1540481597,-0.2392389625,-0.057379093,-0.1824198961,-0.2834384739,-0.1731160134,0.338611722,-0.0379037187,0.2053391039,0.008053802,0.0012945826,-0.3797041774,0.1562884301,0.0118329693,-0.1514143795,-0.2223190367,0.2561254501,-0.0427351184,-0.2844480276,0.3420321047,0.2845244408,-0.0065560443,0.2018401772,-0.2968755364,-0.3999528885,0.40293625,-0.1829158515,-0.1783638,-0.2296224684,0.0490622744,0.1954258531,-0.1067993939,-0.1924477071,0.0228688605,0.1301557571,0.0905380249,-0.0717682615,0.215465188,-0.1068774462,-0.1212995872,-0.0754636526,0.248825416,0.533864975,-0.2447043657,-0.0007777357,-0.0159923304]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1892","title":"request to mirror wmt datasets, as they are really slow to download","comments":"Yes that would be awesome. Not only the download speeds are awful, but also some files are missing.\r\nWe list all the URLs in the datasets\/wmt19\/wmt_utils.py so we can make a script to download them all and host on S3.\r\nAlso I think most of the materials are under the CC BY-NC-SA 3.0 license (must double check) so it should be possible to redistribute the data with no issues.\r\n\r\ncc @patrickvonplaten who knows more about the wmt scripts","body":"Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!","comment_length":78,"text":"request to mirror wmt datasets, as they are really slow to download\nWould it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!\nYes that would be awesome. Not only the download speeds are awful, but also some files are missing.\r\nWe list all the URLs in the datasets\/wmt19\/wmt_utils.py so we can make a script to download them all and host on S3.\r\nAlso I think most of the materials are under the CC BY-NC-SA 3.0 license (must double check) so it should be possible to redistribute the data with no issues.\r\n\r\ncc @patrickvonplaten who knows more about the wmt scripts","embeddings":[-0.194238469,-0.3241634369,0.0704936907,0.2124052793,0.0100606102,0.0652933568,0.1663089246,0.4544945955,0.056202095,-0.1142761186,-0.2518411875,-0.0738516971,-0.0602755286,0.310782969,-0.0818374977,-0.0826895609,-0.0564791039,0.1170890555,-0.6240917444,-0.0960814133,0.0491623133,-0.1243582368,-0.0437549204,-0.0403349809,-0.0081339916,0.2965478599,-0.220490545,0.0053677587,-0.2378391623,0.0302223265,0.0825243667,0.5002441406,0.2514034212,0.3773466349,-0.0001185067,-0.1264594942,0.1999659538,0.1051914617,-0.2532367408,0.0950692743,-0.4915260077,-0.3680053353,-0.235564217,-0.016106084,-0.1161408946,-0.1211741045,-0.1031256467,-0.0419021584,0.2931009531,-0.1038033813,0.0933474675,0.4699967206,-0.4896414876,-0.0356685296,0.5301848054,0.3358390927,-0.2962783575,0.3221007884,0.5346037745,-0.138974309,0.3522563875,0.423499465,-0.1568779647,0.3637658954,0.6082298756,-0.188751027,0.0055280928,-0.0134862866,-0.248511374,0.652790606,0.720821023,-0.1008875892,-0.5376054645,-0.0442308113,-0.0979527459,0.1767011881,-0.3057165146,0.1743835956,0.0661836788,0.2955296636,-0.4052235484,-0.3033839762,-0.057906244,0.0288335346,-0.0891483799,0.4804267585,0.0012672525,-0.0247035343,-0.1784487516,-0.1355646998,0.447550416,-0.4463375807,0.2125969678,0.0525240749,0.0689759851,-0.3961450458,-0.4048391283,-0.1257929951,0.1217897981,0.0788985491,-0.0996380746,0.0322684087,0.060873691,-0.0059303278,0.053396605,-0.0918024257,-0.2332044095,-0.2772554755,0.2760227323,0.1725937873,0.0821032822,-0.0665007085,-0.1882131249,-0.1074698642,-0.4842549264,-0.0299393963,0.0422553979,-0.3843735158,-0.1701832265,-0.2568981647,0.1291088313,-0.1630939841,-0.2392384708,0.0911244228,-0.2050314546,0.2176314592,0.3381380737,0.1368322372,-0.1975701302,-0.1293524206,-0.0453770421,-0.1888690144,0.1018535569,0.4096263945,0.1961645782,0.1012696177,0.0809571892,0.01252288,0.2291604728,0.3779648542,0.2429531813,0.2257295847,0.0011985829,0.2455646694,-0.2264270484,0.5534642935,-0.0571046621,0.494664669,-0.2405672669,0.0189298727,-0.395958662,-0.1411177516,0.1163505614,0.1434739679,-0.2485123426,-0.2411218286,-0.3235337436,-0.1215084046,-0.358140111,0.1436905265,-0.0027514824,-0.160331279,-0.1871885955,0.0441947542,0.1369813532,0.5623070002,-0.3566907942,-0.0392406173,0.1318229735,0.0986153781,0.424931556,0.4538913071,-0.1545871943,0.0098788086,-0.2259876579,-0.0569346696,-0.0896128193,-0.2199738175,0.0895927176,0.4570536613,-0.2056543082,-0.0597877093,0.2275102884,0.1248131469,0.1572809815,-0.3333368301,-0.0803302228,0.4827212393,-0.0005445417,-0.0532289445,-0.240019381,-0.5139433146,0.321305871,0.3085314929,0.1036258489,0.135584861,-0.1118637845,0.0921425149,0.4986327887,-0.0057322043,0.030869225,0.0384275615,0.2461281717,0.4436832666,0.021693971,-0.0548787564,0.1014436707,0.0964070857,0.1379815936,0.0048364475,0.3148323596,-0.2072585076,0.0021639727,-0.1151764542,0.0240758881,0.2152531892,0.0360904336,-0.2255754471,-0.1383960992,-0.3562449813,-0.0284167528,0.0065095904,-0.0618620403,0.035450656,0.3565728962,0.3909591436,0.0855044574,0.3153090477,0.4274924099,-0.1052074805,0.2611545622,0.0331897177,-0.0113981282,-0.2118983418,-0.0324237496,0.380189538,0.48837322,0.4331637621,0.2952303588,-0.540738225,0.5896484852,0.1838252544,0.1421567798,-0.2557565272,-0.181696713,0.2997314334,-0.1090607792,0.1404949427,0.0827980563,-0.077605173,-0.0244501363,-0.1666247845,-0.169337675,0.0878170207,0.2612858415,-0.061919976,-0.2795540392,0.0631749704,0.0156494044,0.2879464626,0.1335858405,-0.1323908418,0.0646268725,-0.0883881748,-0.3523101211,0.1846619248,0.1971083879,0.1618145406,0.1227264032,0.2624022365,0.081486091,0.1348206401,0.3998582363,-0.2348161787,0.0997165963,-0.0268699564,0.0908876956,0.1765918881,-0.157308042,-0.0949692652,-0.2644998431,-0.1340385973,-0.0339329764,-0.1693633795,-0.205760777,-0.2036472559,-0.3950927556,-0.1671643704,-0.5224645138,-0.0712078735,-0.1530241221,-0.2038117051,0.0044244444,0.0359447189,-0.0924441144,-0.0787629038,-0.15335913,0.1934635043,-0.3043530881,0.1676543653,-0.2028636634,0.037273936,0.0826232433,-0.0348152369,0.1339907497,0.1073842794,0.3420969546,-0.1645022482,0.024506405,-0.6451830268,-0.1008385792,-0.089119181,0.2413382679,-0.00529119,-0.0128091332,0.4223403037,0.3099903464,0.1228703186,0.1148653924,-0.1887265444,-0.057935629,0.026614951,-0.1791533381,-0.1181851849,-0.0117393313,-0.1499120891,-0.4867734015,-0.1689320356,-0.0910814926,-0.2901754081,-0.0047472958,-0.0703091174,-0.0402645543,-0.1997772604,-0.2292350531,0.0202789493,-0.1326332986,-0.4535301626,0.4614348114,0.0416385792,-0.237315923,0.1984031945,0.1946207881,-0.2962462008,0.0509123318,-0.4404743314,-0.0259731617,-0.0599781126,-0.1563954204,0.1378271133,0.3616072536,0.0748511702,-0.0016155969,-0.0182547197,-0.1750897914,-0.170382753,-0.3151818514,0.0791268945,0.3384957612,0.0421886668,0.1538708359,0.1167562827,0.9036080837,0.21061261,0.086070329,0.1693006009,0.3139078617,0.0126044797,0.330495894,-0.1501098573,0.0602103658,-0.1646620482,0.1526878923,0.3023821712,0.1654993296,-0.0322428271,-0.1236022413,-0.041088894,-0.0721399486,-0.1531147659,-0.1666941643,0.0026164528,0.1032639667,0.2143984586,-0.1322298646,-0.043378409,-0.2789007723,0.567160964,0.5380114317,0.0727010816,0.1175924838,0.0948472843,-0.2196182311,-0.3919327855,-0.0763272196,-0.0541766733,0.2786936164,0.0569840968,0.1415188909,-0.0046377224,-0.0162834469,0.5470106006,-0.356440872,0.2949028909,-0.1442961395,-0.3694461584,-0.1728060097,-0.3832759261,-0.1095415652,-0.5588153601,-0.0772586688,0.3528344929,0.0084962016,-0.1516514868,0.0020079701,0.0590502732,-0.3691159785,-0.3575654626,0.2556676567,-0.27191329,-0.0996837318,-0.3776426911,-0.006233878,-0.2807677984,-0.2527923286,-0.0268178172,0.0922123492,0.4408909976,0.3354249895,-0.0946624279,-0.0925570875,0.3012338579,0.3603005409,0.0374417752,0.2391992062,0.360665977,0.5046021342,0.271708101,0.0542822145,-0.4194884896,-0.1345648766,0.2902258933,0.4311195314,0.057327088,0.2111226916,0.172272265,0.3007184267,-0.5591008663,-0.0802775249,0.094421044,-0.063487418,0.0172772761,-0.3364315331,0.2843790948,0.1423128992,0.0836704075,-0.2427265942,0.0492841229,0.1078369692,-0.1778744906,0.2658794522,1.1352498531,-0.3869089186,0.3767457008,-0.2859295309,-0.4389281273,0.1930590421,-0.2761818469,-0.327101171,-0.2529634535,-0.3486732543,-0.1005761102,-0.1680055559,-0.077090241,0.2319845259,-0.0434464738,0.3802526891,0.2873779833,0.145206809,0.0846103802,0.2755835354,-0.1763863862,-0.1767053902,0.0174672566,0.085803695,-0.1933781058,0.3361291289,-0.147295922,-0.2969940603,0.0805649534,0.1121013239,-0.2119934857,0.1694366783,-0.0968711823,-0.12840949,-0.3832776845,0.0316645391,0.5623496175,-0.6436485648,-0.0283580981,0.4671011567,-0.4409530461,-0.0093005011,-0.1883154064,0.0606653057,-0.2078014612,0.0433315039,0.1382687837,-0.1444744915,-0.1458599269,-0.2769358456,0.2813341022,0.0168469865,-0.0551267229,-0.4118452966,0.1743118167,0.1672117859,-0.0446522236,0.1962048262,0.0632110164,-0.0272574965,0.0819345042,0.2307800651,-0.4344766438,0.3629769683,0.3060402572,-0.0408442058,-0.0279145129,0.6274658442,0.0486871637,0.1014703885,0.3580069244,0.5374692082,-0.2415651977,-0.1683886498,0.0307497308,-0.0980749428,-0.1062976643,0.1364814639,-0.0502545238,0.0660019964,-0.0690907538,0.1344996989,0.3593295217,-0.4689964354,-0.2436367422,-0.3942449093,0.0242588129,0.0535472706,0.0327044316,0.1865341663,-0.2131736577,0.0048583858,0.0788602829,-0.1300333142,-0.2711833715,0.2341395319,0.1094055101,0.1262315214,-0.0035661156,-0.0821687505,-0.1568306983,-0.0009771427,-0.0895652547,-0.1557528973,-0.1373562813,-0.1435674727,-0.0431494527,0.2149405032,0.2033580989,-0.2076737881,0.2152032554,-0.4240143299,-0.2052709311,0.2071904689,0.0006216474,0.2079426944,0.0590290539,0.307104826,0.4449148178,-0.4507373571,-0.3996251822,0.5348269343,0.3253591657,0.3255485594,-0.0671706796,-0.2354677767,0.2650063038,-0.0123565504,-0.0109522771,-0.0831557512,0.0932816863,0.2349167168,0.6755410433,-0.2223104686,0.2469659746,0.3122739494,0.5689752102,0.165848434,0.0841733068,-0.0623439774,0.3127582669,0.1107252613,-0.2655971646,-0.2527962625,0.0075161299,0.1948583424,0.0039293575,0.0310061611,0.1690722853,0.0471986532,-0.3080037832,-0.1037285849,0.1783949286,0.2707657516,-0.2716914713,0.1359703243,0.3350222409,0.0492308959,-0.0688226745,0.1073077023,-0.0878664181,-0.0548633374,0.2023762017,0.3228906989,0.1374689192,0.1845917106,0.000463529,-0.272870183,-0.0769895986,0.1307688206,0.3602410555,0.1649551243,-0.1015258953,-0.2240437716,0.3066639304,-0.4101093411,-0.1396852583,0.4005136788,0.1548840106,-0.0920049772,-0.0353103578,-0.0261293668,-0.0149947284,0.1356625557,-0.0866739899,-0.1096817181,0.2018984258,0.2682113647,0.3090913594,-0.1002072692,0.0417803265,0.0281454977,0.2311125994,-0.185875684,-0.3661698997,-0.1566649675,0.1162600815,0.2669984698,0.3494156301,0.0623406582,0.0763204396,-0.1407997906,0.0924163982,-0.0587394312,-0.0367527343,0.1045638099,0.1865001619,-0.1557814032,-0.2483767718,0.1958192736,0.0466492139,-0.1271168888,-0.1421160847,-0.3193194568,-0.4644872546,-0.2384031415,0.1430113167,0.2091125101,-0.0635755658,-0.1490928829,0.1420563012,-0.2940702736,0.0162316021,0.6448610425,-0.2574046552,0.1037262082,0.0241501909,0.0357919335,-0.3664202988,0.4402461052,0.5229842663,0.363378346,-0.2721739411,-0.0472266935,-0.2062498778,-0.1155872047,-0.3925127685,-0.0154038509,0.104728438,0.0767885596,0.2676976621,0.0844927058,0.1572554708,0.4931562841,-0.1959601641,0.1494657248,-0.1139352322,-0.0947789103,-0.189532131,0.2923419178,-0.1569333524,-0.3024151325,0.2051188201,0.0730021745,-0.1011532322,-0.0674514696,0.1848720312,0.2188967764,0.22321558,0.0867180973,-0.3278851211,0.2595078945,0.0074989772,0.0889079943,0.026173735,-0.0786310285,-0.0229801033,-0.0662236735,-0.2500491738,-0.1942213178,0.0299948752,0.0709095597,-0.1323094666,0.0677479357,-0.1909638494,0.0218404606,-0.2647408247,0.2825659215,-0.5296422839,0.1088234857,0.089814879,-0.0256011523,-0.1168296263,-0.2822602689,-0.2311119437,-0.184853375,0.5103880763,-0.1827995479,-0.1760086864,-0.3596641421,0.1117762327,-0.3630663455,0.3120629489,-0.1571564227,0.1808515489,0.1460754722,-0.0923871845,0.2985776663,0.5052083135,0.0931392759,-0.0751510412,-0.0936294049,0.2878225744,-0.0517684817,0.0559036545,-0.5135704279,0.0270245131]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1892","title":"request to mirror wmt datasets, as they are really slow to download","comments":"Yeah, the scripts are pretty ugly! A big refactor would make sense here...and I also remember that the datasets were veeery slow to download","body":"Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!","comment_length":24,"text":"request to mirror wmt datasets, as they are really slow to download\nWould it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!\nYeah, the scripts are pretty ugly! A big refactor would make sense here...and I also remember that the datasets were veeery slow to download","embeddings":[-0.380508244,-0.1753539741,0.0075874529,0.1037696823,-0.0104693696,0.0599777699,0.1670318693,0.5077069998,0.2716593742,-0.0307254717,-0.2506194115,-0.221631512,0.033997003,0.24326846,-0.1237284541,-0.2390306145,-0.1749914736,0.1651735455,-0.7791355252,-0.1799131185,0.106156148,-0.1956078261,-0.1282520443,0.0072468398,-0.0354572758,0.1832383424,-0.128082782,0.0092026293,-0.281499505,0.1942979246,0.0807311609,0.4943597913,0.1801086366,0.2579829693,-0.0001139865,-0.1233249083,0.3037301004,0.1550533772,-0.1294935197,0.2597908378,-0.3924541473,-0.3952484727,-0.1610328406,-0.0658745393,-0.0436388627,-0.1273960024,0.0423421934,-0.1791125238,0.2340647131,-0.1549867839,0.1716459394,0.3462652862,-0.4791024029,-0.0595961213,0.3936366737,0.286072433,-0.3739835024,0.3643722832,0.5096567273,-0.1086651608,0.0914191604,0.3541181982,0.0158592202,0.2641918957,0.4950572848,-0.2407833487,0.0755990446,0.0502780229,-0.2511317134,0.5559908152,0.7872530222,-0.0566438399,-0.3280628622,0.1233147085,-0.1975928098,0.2542445362,-0.2248382717,0.056848038,0.0246378183,0.3577070534,-0.4675727785,-0.305555284,0.0396127999,-0.07427378,-0.1059214771,0.2843854427,0.0716390535,-0.012037307,-0.0921839774,-0.1142882854,0.2289706022,-0.4278957248,0.084317714,0.0660024062,0.0435412303,-0.407456249,-0.3956961632,0.0255869534,0.2399980873,0.1858924627,0.0770375058,0.1213208511,-0.0796723068,-0.06023518,0.1222130656,0.2509377301,-0.2132793516,-0.3227464557,0.348944962,0.0415046737,-0.0507071316,0.0060407449,-0.0399916917,-0.1021285579,-0.3949147761,0.123759985,0.0564692654,-0.3286577463,-0.2215178907,-0.2023866326,0.0951999277,-0.2182261646,-0.1477472037,0.1160875112,-0.2152334452,0.324593991,0.3515886962,-0.0812580511,-0.2032058984,-0.106238246,-0.0876734629,-0.0906698555,-0.1020311862,0.3286113441,0.2291557342,0.1653606147,-0.0103698168,-0.0122763645,0.2687145174,0.3333777785,0.1240342855,0.1441444457,-0.0162177868,0.1667734385,-0.3101998568,0.4530290365,-0.0016281067,0.5766244531,-0.2522424459,0.1281220168,-0.4634134471,-0.1512120515,0.093342565,0.2227725685,-0.0691977069,-0.1474102587,-0.3374592364,-0.1135531589,-0.3930585086,0.083896108,-0.0484141186,-0.0942333713,-0.1931762695,0.1001078635,-0.0327635333,0.507394135,-0.4661897719,0.1300623417,0.0478365198,-0.0847914591,0.5431494117,0.5093151927,-0.2762378752,0.1490685791,-0.3273758292,-0.1876093745,0.0762817785,-0.2453272492,0.0465539619,0.5941820145,-0.2964412868,-0.0330521204,0.2292263508,0.1116807237,0.2743132412,-0.2730723917,-0.1289038211,0.3666533828,-0.0583093874,-0.0247745924,-0.2994121909,-0.5489884019,0.222985968,0.2651913762,0.1280792356,0.3034206331,-0.0386638604,0.0988603085,0.6538231969,-0.008927796,-0.0234397613,-0.0003190556,0.0472183228,0.3054360747,0.1835785806,-0.0747923404,0.0838956013,0.0323624983,0.0159604382,-0.0498866178,0.5514832735,-0.2607074678,-0.1371174753,-0.1333980262,0.1238410249,0.2173400074,0.0750877187,-0.1869909018,-0.333196789,-0.4164792895,-0.0838466212,0.3183391094,-0.1707099974,-0.023818193,0.3621007204,0.3282922208,0.0042818831,0.4077354968,0.2721312046,-0.2106322348,0.2551041245,-0.0920156762,-0.0419658013,-0.1011896655,-0.08891581,0.3310208917,0.5127508044,0.2911525071,0.3215765953,-0.4997344613,0.5782851577,0.2840396166,0.0891665742,-0.3063915074,-0.1912351251,0.2503745258,-0.1309273541,0.074878104,0.1060547233,-0.1318771243,-0.0196299646,-0.2404821962,-0.1276592314,0.0532319993,0.4599595666,0.123571977,-0.1596833915,0.1282088459,-0.042849768,0.2617393732,0.0907394588,-0.1311024576,-0.0340095125,-0.0717116669,-0.1973294616,0.2812795937,0.0401278697,0.010960673,0.153315112,0.3378612101,0.0244841855,0.2259532362,0.3778088391,-0.2540162504,0.0651439354,-0.091539897,0.0669928715,0.1442798227,-0.054981485,-0.104156442,-0.4047582448,-0.181286037,0.0463717021,-0.0976759791,-0.1688509881,-0.3446543515,-0.4082675874,-0.0992145315,-0.5792316198,0.0557994395,-0.0046382234,-0.2675746083,0.1047198325,0.0755682513,-0.1571830809,0.1201933175,-0.1703465283,0.2722745538,-0.2807755172,0.3172829449,-0.288212359,0.0609800443,0.0805845857,0.0599450469,0.1933198422,0.059323702,0.3335085809,-0.1804134101,0.0672978908,-0.5876153111,-0.1100950018,-0.0758057684,0.3115236163,-0.0588632636,-0.0066559305,0.4242179692,0.2609001398,-0.0444026515,-0.074484773,-0.1252694577,-0.1272205561,-0.1251815408,-0.1961204261,-0.1452279985,-0.0458365791,-0.2629895508,-0.4468740821,-0.1661346406,-0.0215503182,-0.2584233284,0.098836638,0.0352506004,-0.1489526033,-0.3318729699,-0.1897393167,-0.0466475934,-0.1492340118,-0.3901198804,0.4578678906,0.0261714738,-0.2646767795,0.1498559713,0.2164316326,-0.5094377995,0.0981018767,-0.2976005971,-0.0140467947,-0.1650477201,-0.2264684439,0.1445492655,0.2360325009,0.0886883587,-0.0517388694,-0.0951496214,-0.0671269596,-0.096291624,-0.0946700722,-0.015709905,0.3045810163,0.0664640591,0.1675754488,-0.0576209649,0.9596548676,0.0216907207,0.1754843593,0.0679390207,0.271738112,0.0398697183,0.2014282495,-0.2668823004,0.0835354403,-0.1249187812,0.0521827079,0.3039142489,0.1249792501,-0.1708259284,-0.0700295493,-0.0757935122,0.0712349638,-0.1591444314,-0.1414767206,-0.1161920205,-0.0078626554,0.3021861315,-0.0106127821,-0.1702643186,-0.2661761343,0.4968166947,0.5332590938,-0.0456053838,-0.0154966637,0.092160821,-0.0783625394,-0.3821448088,-0.0482672192,0.0627895519,0.2886771858,0.014483585,0.1875742078,0.0463960841,-0.0074963467,0.545327425,-0.1237715483,0.3349527419,-0.1271114349,-0.2761808634,-0.0125233037,-0.3755603135,-0.0906999111,-0.4050471187,0.0913536623,0.1981764138,0.0271171555,-0.1265956908,0.3056598604,0.0780078694,-0.3180401325,-0.4010608196,0.2760358155,-0.1486581415,-0.0660198629,-0.2761563361,-0.0371491835,-0.3086219728,-0.2919811904,0.0455662906,0.0917203054,0.3058342338,0.3838367462,-0.2576649487,-0.0059410324,0.2473198771,0.3797289133,-0.0763757601,0.4432216287,0.1999130249,0.4128928483,0.2050317228,0.1331308037,-0.2609005272,-0.1755689234,0.1446676701,0.5312457085,0.0300274808,0.0413486287,0.1008016318,0.2563358545,-0.4412170351,-0.0150062535,0.0300202742,-0.1436484158,-0.1330210865,-0.3638939559,0.4546463192,0.2197041512,0.0021270763,0.1150176749,0.0837419704,0.123145178,-0.2407303452,0.1318782717,1.2271447182,-0.326398015,0.1647959799,-0.3563350737,-0.486818701,0.0334769003,-0.4651010334,-0.2087593079,-0.3127733469,-0.4907945096,-0.131943807,-0.1764509976,0.0160791669,0.2103881836,-0.0678397566,0.3187938333,0.2265433818,0.1773725301,0.0438166857,0.4219678938,-0.2083986849,-0.1405159384,0.2325104326,0.1023765802,-0.1130534261,0.2913128436,-0.1807505786,-0.2523027062,0.0849611163,0.0883075148,-0.3463482857,0.1626554281,-0.1756799668,-0.0730063915,-0.4221749008,0.0382067822,0.4540902078,-0.5974770188,-0.0403135829,0.4229358733,-0.4424490333,0.1386766732,-0.2240240425,0.1539204866,-0.1801401079,0.1034055129,0.1297213733,-0.1762741953,-0.2432310283,-0.3747296035,0.1003049389,-0.125598371,-0.2650673389,-0.3074172139,0.3010378778,0.2093650252,-0.0423919559,0.296833843,0.0057017822,-0.004228049,0.1096119583,0.1353621334,-0.3067114055,0.4592967927,0.4877766967,-0.1900091171,-0.0723130554,0.6181307435,0.2643424273,-0.0587382354,0.3968245387,0.3254404664,-0.324373126,-0.226950407,-0.0082668765,0.1197329015,-0.1654708982,0.048173558,-0.2693102956,0.0307194237,0.0650308579,0.2906700671,0.20030123,-0.3191117346,-0.3266791999,-0.2999357581,-0.1410443038,0.036391817,-0.1130056307,0.1646086872,-0.0948690847,0.147203356,0.0882496387,-0.0622582883,-0.3280103207,0.1313577741,0.1928316653,0.0267670024,0.1010590196,-0.0162746832,-0.1377652884,0.0935381353,-0.0342332609,-0.0437326394,-0.2401137501,-0.1801445633,0.0280419122,0.1960351467,0.0092212874,-0.1458552182,0.1458788216,-0.4362347424,-0.2423060089,0.2092887312,-0.0175634138,0.1895725727,0.0336885713,0.0602163896,0.4817601442,-0.4044144154,-0.2955757082,0.4526641369,0.3606831431,0.4359558523,-0.1289246678,-0.2172566503,0.1653545648,-0.0583169237,-0.0379864462,-0.0083540967,0.2533789277,0.3885247409,0.5414743423,-0.0975567251,0.2074246258,0.3328344226,0.5465680957,0.1835826635,0.1167751923,-0.2574371994,0.196217224,0.1262336075,-0.2941477597,-0.178119421,0.2401687056,0.1639474034,-0.0225158501,-0.0333560854,0.1968991607,-0.0072687282,-0.0731644481,-0.0811792165,0.2850218117,0.1132991612,-0.3163086474,0.1843852103,0.2956701219,0.0016439751,0.0151046179,0.2206532359,0.004664958,0.2000059336,0.1815547943,0.3975974619,0.289358288,0.2947398126,-0.0303024016,-0.2451177537,0.0555301011,0.2859324515,0.2938249111,0.1808366627,0.1104941368,-0.3314799666,0.3333823085,-0.3425584733,-0.2218653411,0.3521631658,0.1152308285,0.0076440042,0.0382804237,-0.1178940162,-0.0955639556,0.0974492729,-0.1117019579,0.1432792246,0.2680998147,0.1403879523,0.1448079944,-0.0640906245,0.0713649839,-0.0701779053,0.2951077521,-0.2350680083,-0.3420348465,-0.2025130838,0.1914536655,0.4404243529,0.5600758195,-0.1249941289,0.2510111034,0.0793602243,-0.0655985996,-0.1637025923,-0.0240777284,0.1137699559,0.206111446,-0.1341822743,-0.2268022597,0.2327245474,0.0812334344,-0.139989078,-0.0647991523,-0.2271707207,-0.5099568963,-0.297840029,0.1872094274,0.2960654795,-0.0391263291,-0.1910185963,-0.0217959024,-0.1943466216,-0.1120372713,0.3948348761,-0.2685810626,0.0481979698,0.0001683157,0.0641377568,-0.317589879,0.3527601063,0.5571141839,0.2938530445,-0.2924822569,-0.1421654373,-0.3117378354,-0.0294770394,-0.3457061052,0.0882662982,0.0843097866,0.0187884066,0.3116803169,0.0353750996,0.3401125669,0.3215561211,-0.0861686841,0.0670943037,-0.1379242688,-0.033606343,-0.097779125,0.2399317324,-0.1622874141,-0.420037806,0.3297258019,0.1149041504,-0.0178159475,-0.0010443287,0.0221915506,0.2019759417,0.0660205036,0.0297746174,-0.2607840598,0.3256411254,-0.0511151515,0.1036875322,-0.0411339737,-0.1182778105,-0.088549532,-0.1841375232,-0.0987280086,-0.2638864219,0.0237838142,-0.0384649336,-0.0475370772,0.1023280174,-0.2719499469,0.1443005055,-0.3689766228,0.0997441709,-0.4695650935,0.0733556598,0.0477818474,0.0111497296,-0.0262464862,-0.1484719664,-0.1620293558,-0.1092313677,0.4935832918,0.1556566656,-0.3027803004,-0.360182941,0.1839686781,-0.3446590006,0.4294776022,-0.1602382511,0.0225390531,0.1177980378,-0.1433000118,0.3758054674,0.4795022309,0.1060111374,-0.1197685301,-0.1107084453,0.1266928017,0.0223802719,0.0692919716,-0.5538512468,0.0501204692]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1892","title":"request to mirror wmt datasets, as they are really slow to download","comments":"I'm downloading them.\r\nI'm starting with the ones hosted on http:\/\/data.statmt.org which are the slowest ones","body":"Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!","comment_length":16,"text":"request to mirror wmt datasets, as they are really slow to download\nWould it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!\nI'm downloading them.\r\nI'm starting with the ones hosted on http:\/\/data.statmt.org which are the slowest ones","embeddings":[-0.4048575759,-0.3140096664,0.0369965173,0.1695623994,-0.0139842182,0.1483641565,0.0976714566,0.4406719506,0.201249823,-0.0260525476,-0.3467169404,-0.2557108998,0.0894272178,0.1072438285,-0.1605696976,-0.2548781633,-0.0824236795,0.1284085363,-0.6802170873,-0.1375779659,0.0916178077,-0.2926720381,-0.1066290364,0.1005208194,-0.1244999692,0.1187914237,-0.1336611956,-0.0403247587,-0.2713430822,0.2231260687,0.0854254514,0.3694525957,0.1696945727,0.2424310148,-0.0001144641,-0.1100508869,0.1767383218,0.2066615075,-0.2290118337,0.0147697991,-0.4497104585,-0.4104295671,-0.2166112363,0.010225093,-0.048070807,-0.2220502496,-0.0168272033,-0.1880827248,0.203047052,-0.2183157802,0.1400485188,0.29976511,-0.4613548517,-0.1171535552,0.3193843365,0.2256761342,-0.2811764777,0.3676026165,0.3717106283,0.0632952526,0.1310904473,0.284540832,-0.0112572461,0.5359501243,0.3685594499,-0.166326046,0.1632967442,0.1413364708,-0.2258806676,0.6267239451,0.8606649637,-0.0362282246,-0.3107337058,0.1644583493,-0.104365848,0.3088447154,-0.1453811675,0.2164489031,0.0218988359,0.3345198631,-0.3599195778,-0.2622787058,0.0219189562,-0.0385216847,-0.1772525609,0.2454934269,0.0386833213,-0.0167623181,-0.1030218899,-0.1719418317,0.2029070258,-0.5088528395,0.1482295394,0.0559971072,0.047647588,-0.3436841965,-0.602165997,-0.0119132502,0.2365089208,0.1315677762,-0.0645153001,0.0113891168,-0.1155203953,-0.0040341085,0.1955769211,0.021670023,-0.3318411708,-0.3245331943,0.2537920475,-0.1227808073,-0.1207693815,-0.0051750974,-0.0560189411,-0.0316803418,-0.3978105783,-0.1201777682,-0.0525082275,-0.4321282804,-0.2400128841,-0.2409909964,0.0043355208,-0.2385877371,-0.1764444262,0.0435199924,-0.2159591615,0.1803094298,0.2785016894,-0.055914633,-0.1992187947,-0.1912965626,-0.1133068204,-0.2688353062,0.0762467459,0.399210304,0.1202946007,0.0832927898,-0.0349252783,0.0040843477,0.255731225,0.3186903596,0.2280763239,0.0805885345,-0.0507082418,0.163294211,-0.1923894435,0.5492449403,0.006165639,0.5649519563,-0.2161950916,0.0496356189,-0.4902123213,-0.0830659866,0.166442126,0.1738640219,0.014853837,-0.1995098591,-0.3133862317,-0.0792067796,-0.5097865462,0.1552453041,-0.1281242371,-0.0444266796,-0.1475669146,0.1225653365,0.0442835763,0.4127905965,-0.2928535938,0.015791852,0.0534084216,-0.0462496355,0.4894139171,0.505053699,-0.2413398921,0.125004977,-0.2376469225,-0.1175844669,0.0515243858,-0.1229912266,-0.0325720869,0.5273026824,-0.3150410652,-0.2354463786,0.1770163327,0.2023500055,0.2160562724,-0.2299128622,-0.0736211091,0.3995553255,-0.0265602395,-0.0433853231,-0.2944585383,-0.4869575202,0.2586955726,0.2374853492,0.0530503057,0.195983842,-0.0057679331,0.1948939264,0.4590452313,0.0916238353,-0.0071796691,-0.04296951,0.0914264917,0.2943187654,0.0716167986,-0.0905445144,0.1710486561,-0.0536939576,0.1256409287,-0.0315473638,0.590339601,-0.1503253132,0.0074809697,-0.0631246865,0.1456245482,0.3890587986,0.0475445203,-0.1497329921,-0.25374493,-0.2909080684,0.048538968,0.3243517876,-0.2179533392,-0.0125326514,0.5686083436,0.4224882126,0.0976043567,0.4273555577,0.3591609895,-0.1608591229,0.212768212,-0.0247060694,-0.1188272759,-0.1153676957,-0.0895709693,0.3114093542,0.5767362714,0.185341537,0.3630093932,-0.5933862329,0.5415121317,0.1586470902,0.035349384,-0.2379420698,-0.1708571762,0.2270596772,-0.0726936087,0.0169404428,0.0890948474,-0.0511591621,0.0289126188,-0.1342879236,-0.0268503148,0.0258786548,0.4235852361,0.0438396372,-0.1089968458,0.1734325141,-0.0720521435,0.1587706506,0.0192032922,-0.087114118,0.0364314839,-0.0771665052,-0.399472177,0.2457520962,0.1586457938,0.0353569873,0.1561485231,0.309667021,0.1416003257,0.1711187363,0.2627328634,-0.1946193278,0.0222405493,-0.0450107977,-0.0230077766,0.1137780026,-0.1154776141,-0.1428893805,-0.2855250835,-0.0421554036,0.0820326731,-0.0846410245,-0.0614049211,-0.3466530144,-0.5324671268,-0.1312478483,-0.617423296,-0.0382481627,-0.1032452136,-0.2420078218,0.0315125249,0.0466377065,0.0030832421,0.055459667,-0.1959473044,0.263687849,-0.2539597154,0.2536245584,-0.2014015466,0.117071569,0.0113308877,0.0463702902,0.1688543409,0.2607702315,0.3691004515,-0.1120979339,0.1326038986,-0.54004848,0.0408738106,-0.0130174793,0.3866237402,-0.0338755883,-0.2162379622,0.4594663978,0.1832925081,0.0527516827,0.0272366572,-0.1683708727,-0.0866200924,-0.0670042783,-0.1649515331,-0.0319061428,0.0660424307,-0.2619406879,-0.5064035058,-0.1169011891,-0.0841074064,-0.257537365,-0.0112296492,-0.0393469259,-0.1664268821,-0.1902399808,-0.2238335162,-0.0161364991,-0.2360783815,-0.4750446677,0.4797661006,0.0775588006,-0.2248677462,0.1456604749,0.22342287,-0.4729546607,0.1071243435,-0.3786622286,0.0849280134,-0.1833435297,-0.1696145236,0.1855350733,0.279622376,-0.056123849,-0.1013823748,-0.0752929002,-0.0664281473,-0.1520847082,-0.1004991606,0.0640997887,0.390765965,0.0871325582,0.0543973558,-0.0144240502,0.9082939625,0.2233340591,0.0690710992,0.0383333601,0.2785713375,-0.0878232867,0.2137203664,-0.2204850763,0.0769744962,-0.0739088431,0.1334906071,0.4275669754,0.2644349337,-0.0535896309,-0.0824767724,-0.0538632683,0.0244774316,-0.1276631802,-0.3043903708,0.056410037,0.1423767656,0.3426665664,-0.1209506691,-0.1897440255,-0.3379052877,0.4688180387,0.6513611078,-0.0609649047,0.0805081576,0.080832839,-0.0314890742,-0.2758725882,-0.0102398684,0.1072125807,0.3037173152,-0.0362743661,0.2422356606,0.0506251976,0.062750347,0.2902603149,-0.0846781582,0.3149998784,-0.2368753403,-0.3210120201,0.1225262806,-0.2114388347,-0.0314983018,-0.4593895674,0.1168687567,0.2564661801,0.0187606066,-0.0877767801,0.1811024547,0.0854194015,-0.3570493162,-0.3601974547,0.3576740026,-0.1333040297,0.1113687679,-0.2920585871,-0.1578163505,-0.1651885509,-0.1639720798,0.0473700687,0.0987864807,0.2592989802,0.4022815228,-0.2731802464,0.0022825825,0.2368913293,0.4151025414,0.1408010572,0.5216087103,0.2833666205,0.3856368661,0.2477420419,0.135935396,-0.3158938587,-0.2188782394,0.1133713871,0.3797287643,0.0028565454,0.2040544599,0.1859217733,0.2156556845,-0.4926126599,-0.0066154976,-0.0361992531,-0.0276058111,-0.111558415,-0.2969650924,0.438880235,0.2685517967,0.0987454355,0.1024798676,-0.0013477979,0.2189721018,-0.4021963477,0.1367766708,1.0715600252,-0.3632158339,0.1096803322,-0.3807941377,-0.4657898843,0.1052370518,-0.4329254329,-0.2255306095,-0.2089772969,-0.4920128286,-0.1338612884,-0.1546847224,0.1261092424,0.1819249988,-0.0065345392,0.1903235465,0.3940100074,0.2490646541,0.0605362728,0.3680457473,-0.1053008959,-0.1515583098,0.1496782601,0.0964277461,-0.1311601549,0.2217648476,-0.0750767142,-0.2779450119,0.0856968686,0.1454675794,-0.2411844879,0.0366284736,-0.1807596534,-0.0174850114,-0.3960857391,0.0377732776,0.4888110757,-0.686460793,0.0188975111,0.3415190578,-0.4918785989,0.0173330288,-0.2555533946,0.0967732295,-0.2075228095,0.1782932281,0.1384382993,-0.1854407191,-0.1331966519,-0.2244444788,0.181885317,-0.0898350105,-0.1923565269,-0.4273499846,0.2352772951,0.1829027086,0.0969031081,0.2842636406,0.0437414385,-0.0054004444,0.0888818055,0.1895171851,-0.3294634521,0.420753032,0.5548223257,-0.1979381293,-0.1259873062,0.6357402205,0.204756856,0.1227298155,0.3180736303,0.2730902135,-0.309709698,-0.181944713,0.0837123767,-0.1224996522,-0.0889851153,0.0476451814,-0.2326230258,0.2021251172,0.0015359161,0.3044347167,0.2734978497,-0.4514809549,-0.350833416,-0.332285583,-0.097191304,0.0723167881,-0.2590216994,0.1174583957,-0.1480484456,0.2740131021,0.0842933506,-0.0004575704,-0.2636137605,0.16949597,0.2028912008,-0.0421103239,0.1046393514,-0.031433735,-0.1653338969,0.2372364998,-0.0471886545,-0.055654835,-0.1970707774,-0.1565998495,0.0797374323,0.2054679245,0.0912319794,-0.2244706601,0.0881510898,-0.4742835462,-0.2115296572,0.2286708653,-0.1130283847,0.0611490123,0.0590620711,0.0175807867,0.3800590634,-0.333848834,-0.4026545584,0.5654033422,0.3676562607,0.316298008,-0.1362122297,-0.3233819306,0.1768166721,0.0032183614,-0.2345848978,-0.0586541034,0.189022541,0.1380799562,0.5165481567,-0.151564762,0.2381733209,0.4325608909,0.5199490786,0.0538495071,0.186631754,-0.1702868342,0.253813833,0.0755517259,-0.2593781352,-0.1948936731,0.2842742205,0.2228807956,-0.0430594534,-0.0183991939,0.2436733991,0.0698681101,-0.2601143718,-0.1258624047,0.1752544343,0.1359932721,-0.2857545018,0.2617629766,0.3761045635,0.1796094626,0.0130057605,0.2156082541,-0.091237396,0.107367076,0.2476565838,0.3638363779,0.2184050381,0.1920839548,0.0094175758,-0.2894505262,0.0534280092,0.3052604198,0.2470102459,0.0937669575,0.0228027385,-0.1456734389,0.1766122878,-0.406226337,-0.2408550233,0.2607614696,0.0568463989,-0.0434246697,0.1096576303,-0.0884288996,-0.0807548612,0.1009493992,-0.0840277895,0.0368060134,0.2215911746,0.1124881953,0.2072285861,-0.1745325625,0.0337692164,0.1124484837,0.1183237061,-0.1860309243,-0.3728147149,-0.3661412895,0.1247574762,0.4770665467,0.4669599831,-0.2627882063,0.0623641983,-0.0314197317,-0.1631105691,-0.1262384653,-0.0515965335,0.1546846777,0.2232962996,-0.197398454,-0.395658344,0.3107919991,0.0714399368,-0.0971719772,-0.0413184911,-0.2396615893,-0.4381754398,-0.4239861369,0.2285808474,0.2297338545,0.0753645897,-0.1485553682,-0.1019665748,-0.0453389995,-0.1195874438,0.3612485528,-0.3295561075,0.0384022221,0.0436413549,0.0518504344,-0.3699964583,0.2957012057,0.4970721304,0.308634907,-0.2605168223,-0.2158732414,-0.2554790676,-0.0673455,-0.3787695467,0.0402849801,0.0965735614,-0.0228543486,0.3310455978,0.1006317884,0.2956652045,0.4210744202,-0.0227038153,0.1257092357,-0.1323169768,-0.0828522518,-0.1275241226,0.2189684212,-0.2469501495,-0.3628223538,0.3034452498,0.1365980953,-0.0504155196,0.0687016398,0.0925953016,0.2002922893,0.0921232104,0.0337069929,-0.2981725633,0.2842431664,0.03695333,0.1631704718,0.0750778988,-0.0555993915,-0.0350967124,-0.265278399,-0.1652207375,-0.2003612369,-0.0574180819,0.0681429431,-0.009916041,0.0072871125,-0.1741359979,0.1290452629,-0.3380168974,0.1175001934,-0.5509541631,-0.0493194312,0.1299543232,-0.0076273032,-0.0067697563,-0.2790979445,-0.05415795,0.0086204484,0.3676296771,-0.0132627115,-0.1212072372,-0.408597827,0.0892248526,-0.1533352435,0.3862652779,-0.075510487,-0.0240538139,0.2824198306,-0.1754864752,0.4065974653,0.4313983321,0.1577373445,-0.0796525925,-0.1086573079,0.1649679393,-0.0654714778,0.0759764612,-0.4208132327,0.1529712975]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1892","title":"request to mirror wmt datasets, as they are really slow to download","comments":"@lhoestq better to use our new git-based system than just raw S3, no? (that way we have built-in CDN etc.)","body":"Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!","comment_length":20,"text":"request to mirror wmt datasets, as they are really slow to download\nWould it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!\n@lhoestq better to use our new git-based system than just raw S3, no? (that way we have built-in CDN etc.)","embeddings":[-0.2889099419,-0.330678761,0.0120119527,0.1093232706,0.0525675081,-0.0814258605,0.161522463,0.4271324575,0.0938708335,-0.0190335866,-0.3289594054,-0.0868706703,0.0092277732,0.3145236373,-0.2300491333,-0.13434048,0.0110245394,0.1879598051,-0.5668653846,-0.1316286623,0.1182784885,-0.1686478406,-0.0159086976,-0.0560586043,0.033592239,0.2922874689,-0.1292638481,-0.0466057733,-0.4010473788,0.2060649842,0.0686515346,0.4430897832,0.2044611275,0.423340559,-0.0001147051,-0.1628362983,0.2551142871,0.1594582349,-0.2941845357,0.1020636782,-0.5625475645,-0.2940482795,-0.1721081883,0.0206452962,-0.2282908559,0.0565156303,-0.0642556846,-0.1407652646,0.2692017853,-0.2676649988,0.1218051687,0.3570899963,-0.569530189,-0.0760531351,0.4883638024,0.3561578095,-0.3749561608,0.3744753301,0.4599542916,0.0103055267,0.2937456071,0.3776448667,0.0498915091,0.4903302789,0.5588130951,-0.1442912519,0.1084049791,0.1010665521,-0.2936745882,0.6387222409,0.8014050722,-0.0934529752,-0.4955377579,0.0299208537,-0.1658012271,0.1371797621,-0.1666875035,0.0896818563,-0.0822557807,0.2906906009,-0.5258622766,-0.4199448228,-0.0245539527,0.0971614346,-0.1622619927,0.3614453375,-0.0499788225,-0.0575058535,-0.02159914,-0.0868925452,0.2871056199,-0.5339570045,0.0277758427,0.0092751235,0.0158333238,-0.5079792142,-0.5281723738,-0.1124781296,0.1427506804,0.0715579167,-0.2179968804,-0.0615719035,0.0635953322,0.0075886487,0.1665517688,0.0226126723,-0.3363720179,-0.1470309347,0.2428098321,-0.0826789066,-0.0626366735,-0.1126049981,-0.1358261853,-0.058529038,-0.4983781874,-0.059105143,-0.0641234145,-0.3968077898,-0.0560140759,-0.3704043329,0.0905914307,-0.2271958143,-0.2425270975,-0.0228242557,-0.1544200182,0.1946366131,0.230892241,0.0609242953,-0.1811841577,-0.0949192941,-0.0341663994,-0.3538129628,0.0227362309,0.3785385191,0.1024874747,0.0440883413,0.0467234477,0.0125861587,0.2723509967,0.3997442126,0.1198748946,0.1355108321,-0.0892270878,0.1303317547,-0.3001799285,0.5513956547,-0.082259573,0.5401500463,-0.2474909574,0.0239216164,-0.3828434348,-0.1484455466,-0.0855847448,0.1434902996,-0.1461866945,-0.2256681025,-0.4768670499,-0.1261459887,-0.3702212572,0.0962527469,-0.0732416958,-0.0139557952,-0.1799243689,0.1466061622,0.009652365,0.2471744865,-0.2330057025,-0.0535838269,0.0196600463,-0.0174877122,0.5455886722,0.604721725,-0.2435891926,0.0085157417,-0.1989867091,-0.1130551696,0.0107872877,-0.2240668386,-0.0656671226,0.5325779915,-0.2225827873,-0.0993728042,0.1051367372,0.1408649385,0.1468102038,-0.2809913456,-0.2059138715,0.5454112291,-0.0261678938,0.0190587342,-0.2581395805,-0.5230137706,0.1062318683,0.2587064505,0.0255726036,0.268631041,0.0157456771,0.16113621,0.4051992297,-0.0458651483,-0.0070546712,-0.0064185131,0.3837785423,0.3038677573,-0.0046905219,-0.0034467319,0.1089395434,0.0643636212,0.0773331672,0.0847619474,0.4987878501,-0.1055862233,0.0438254699,-0.1553094983,0.1709591746,0.3255119324,0.0487534925,-0.1046043038,-0.1730397791,-0.3651247025,-0.0260843728,0.0542201363,-0.0070962016,-0.0475674719,0.5461928248,0.3452306688,0.0251750313,0.3276378214,0.2912020385,-0.2134888321,0.1773380786,0.0235403571,-0.0474236421,-0.2182141691,-0.1485566348,0.4687215984,0.5035502315,0.4017015696,0.3517154753,-0.437682271,0.4014915526,0.0752760693,0.0569213666,-0.20259206,-0.2804099619,0.1560456455,-0.0736740455,0.0032426224,-0.0218051374,-0.1877586246,0.0269898977,-0.173040092,-0.0699176639,0.1036519706,0.3284134865,0.0244786292,-0.2184794843,0.2030522674,0.0279873461,0.1575505286,0.210099861,-0.130979538,0.0774587244,-0.0152056394,-0.2841055393,0.1486790776,0.1963616461,0.1138747856,0.1733242422,0.3031970561,0.1898444742,0.1547533423,0.4193894863,-0.2129912972,0.0615872256,-0.0796338916,0.2978202701,0.2055939734,-0.0357468911,-0.1473859251,-0.1455067992,0.0025367574,0.0777263343,-0.1255670935,-0.0882344469,-0.242634654,-0.3631213307,-0.1498048156,-0.4077998996,-0.0824510381,-0.0949652493,-0.1468745023,0.0854016468,0.0987483859,-0.0776861385,-0.0030245779,-0.0181440376,0.3205004632,-0.2589112818,0.0728623345,-0.2940064073,0.034422975,0.074730821,0.0384311713,0.1803443283,-0.1175260246,0.3160448968,-0.1246628985,0.0540366024,-0.5954364538,-0.0566791892,0.023898758,0.3337820172,0.1331888288,-0.1336387694,0.4270215034,0.1770924181,0.0450295471,0.037895523,-0.1569055021,-0.2003164887,-0.1181856394,-0.1908142865,-0.0698712468,-0.0480867401,-0.1044219434,-0.4379579127,-0.1385015845,-0.0337491333,-0.3620450795,-0.0011296557,0.0098884907,-0.0695373937,-0.2653432786,-0.2764092386,-0.0574303046,-0.216629833,-0.5555568337,0.3413765728,0.080528833,-0.1662790179,0.0229264461,0.3018043935,-0.3650659621,0.0230336301,-0.4307643771,-0.0529677644,-0.0700539649,-0.0424139202,0.2023964077,0.2894767225,0.0451646335,-0.1432104856,-0.0733305216,-0.1906097531,-0.0912860483,-0.21619533,0.176628232,0.3960534036,0.0840989128,0.2189153433,0.1232558563,1.0079665184,0.2507517934,0.0240488034,0.0811173096,0.3469982743,-0.0633275062,0.1654553115,-0.1328163147,0.0492813066,-0.1263699681,0.1854526401,0.4246993959,0.1636006236,-0.3049119711,-0.1610038728,-0.0788758993,-0.0187024996,-0.0902349576,-0.2591300011,0.0133083416,0.1059597284,0.2703246772,-0.1721437275,-0.0569164157,-0.2881118953,0.5134935379,0.5014878511,-0.0035765425,0.0948700756,0.1205557883,-0.1066680774,-0.4163663387,-0.1466144323,0.0493514314,0.2472967803,0.0047215088,0.2376878709,0.0504564717,-0.0322898962,0.505035162,-0.243818298,0.2876742184,-0.2740414739,-0.4379419684,0.0138898389,-0.2338367254,-0.1396898776,-0.4957695305,0.1366998404,0.3703024387,-0.0324072018,-0.1467790753,0.0312684029,-0.1002137214,-0.2395892441,-0.4307837188,0.2779767513,-0.1126506105,-0.0150617752,-0.2945275009,-0.0675639957,-0.1654196829,-0.1310248822,0.0797211528,0.0226147138,0.4109747708,0.2506613433,-0.1289608777,-0.1178849414,0.2919684052,0.3045597076,0.0532699712,0.3760148883,0.2645440996,0.5690954924,0.3403799534,0.0349323861,-0.2294002622,-0.1585590839,0.1091764122,0.3894104958,0.1780187637,0.2084598243,0.1500378996,0.2509880662,-0.5666327477,-0.0764778033,-0.002303611,-0.0262864102,0.0252304766,-0.1212276816,0.3590238094,0.1255481541,0.0238728374,-0.1480847597,0.1376604289,0.1760539711,-0.3433087766,0.2689771354,1.107686758,-0.3656045496,0.2305892408,-0.3309267461,-0.6006500721,0.1223600134,-0.3119621873,-0.2270287424,-0.2131426185,-0.3130872548,-0.0980494991,-0.1495820284,-0.0412907712,0.1944207251,-0.0749912485,0.2023566961,0.4440805316,0.1437163204,0.0984948501,0.4232386649,-0.0224348139,-0.1782529503,0.1104894131,0.1282366961,-0.2176444381,0.3711437881,-0.112985447,-0.2857261598,0.032713268,0.2146435976,-0.2116666287,0.1427698582,0.0150580313,-0.1655507684,-0.300799489,-0.0127583966,0.4979439378,-0.6503034234,0.0528458282,0.3720949292,-0.4815836549,0.0828290954,-0.1513359696,0.1640633345,-0.2222639173,0.0016877876,0.0279363785,-0.230705604,-0.1817238182,-0.2478939742,0.2220163941,-0.2231402546,-0.0563500635,-0.3715164065,0.1247260943,0.2312968969,0.0375285298,0.2623560727,0.0632217303,0.0235481597,0.0910902917,0.2097056061,-0.3416237533,0.3736175895,0.4086952507,-0.1830561757,-0.1204476431,0.6932265759,0.0599627383,0.1400994807,0.3065079451,0.3477091193,-0.2966594398,-0.1432987601,0.0931545049,-0.0949296132,-0.2008296996,0.0257395934,-0.1383600235,-0.0713489205,0.0436998829,0.2620227337,0.3745442927,-0.4309707582,-0.2084338963,-0.2575330436,0.0982459858,-0.0194727611,-0.2694956064,0.2521566153,-0.0434695072,-0.0051662955,0.043202389,-0.0195719618,-0.2829038203,0.1779470891,0.0899410248,0.1331410557,-0.0084639285,-0.1326048821,-0.1195361018,0.0352881737,-0.0872822702,-0.1757516265,-0.1880126297,-0.1848124415,0.0541844033,0.2005055547,0.1780557334,-0.2121697217,0.219863534,-0.3553383648,-0.2397073805,0.2403867245,-0.0003095021,0.0835642815,0.051109124,0.1653579473,0.395906508,-0.4697081745,-0.3244507015,0.5378158092,0.3684096932,0.296150893,-0.2040382028,-0.2357987463,0.3141792417,-0.0028276336,-0.0207214337,-0.0554047674,0.2595159113,0.2902944088,0.5963610411,-0.1955122501,0.2901704907,0.4197302163,0.5349561572,-0.1353764534,0.0373879783,-0.1808124781,0.3128833473,0.1392176598,-0.1906391829,-0.3360894322,0.0202596504,0.2062437087,0.0407676548,0.0474689081,0.2516182959,0.1388719678,-0.1557547599,-0.1798747778,0.1865212172,0.2619705498,-0.2746182084,0.1914222986,0.2551439404,0.0179772396,-0.0672473311,0.1334267855,-0.1012008861,0.0294645149,0.2103702724,0.1706470847,0.1291402131,0.3148681223,0.0914536119,-0.1671327651,-0.0086862911,0.1696699411,0.3133716583,0.1615932733,0.0091584139,-0.420160085,0.2893745899,-0.4594337642,-0.1767069697,0.3615793884,0.1225368679,-0.1365913302,-0.1155404598,0.005283501,-0.0001123362,0.2047948688,-0.2032438815,0.0247592255,0.130185619,0.2216451615,0.2809840441,-0.0325115956,0.0774461702,0.1081179902,0.2180994898,-0.172107175,-0.1806696802,-0.3341346085,0.0832063407,0.3756327927,0.5546928048,-0.1806427389,0.137149483,-0.003132215,-0.0899247825,-0.0146248629,-0.0397587866,0.0984542444,0.1521309614,-0.0985548049,-0.2382955849,0.2103476971,0.0660591498,-0.1263799369,-0.0434932709,-0.3883406222,-0.4532511234,-0.2978823483,0.1055048034,0.2279828191,0.0872058496,-0.0487627871,0.1077936888,-0.1841795892,0.0548812561,0.3799105883,-0.3713881373,0.0885182247,0.0712898597,0.0576950721,-0.3803709745,0.4389807582,0.4249639213,0.3350913823,-0.2078576535,-0.1208073124,-0.0419812389,-0.0675685704,-0.3867940009,0.0545179807,0.0605669953,0.1189957261,0.2939837873,0.2040805519,0.1446166635,0.391386807,-0.1518393308,0.219924584,-0.1296265125,-0.0603255071,-0.0497748442,0.2328106016,-0.1578113288,-0.3550445139,0.3176903129,0.0918005854,-0.0439624712,0.0779464766,0.1598030925,0.2010507733,0.1578125358,0.0118871517,-0.2411862761,0.215612635,0.010236932,0.0855984315,0.1212218776,-0.0877584666,-0.1366536766,-0.2215114236,-0.2333175838,-0.1392724514,-0.0417704768,0.1220543981,-0.0421441048,0.0073033194,-0.1979624182,0.1851385534,-0.2294759452,0.0798231438,-0.5772967339,0.0448308811,0.1001190841,-0.0192213375,-0.0352138542,-0.2984090447,-0.2601257563,0.0002441441,0.4037048817,-0.1191800907,-0.1109490842,-0.3758733273,0.0425349995,-0.1837888956,0.2944350243,-0.1605956703,0.0931734592,0.1711373329,-0.0880122185,0.3685884178,0.5453279018,0.2308086455,-0.0512096882,-0.1287275553,0.3854849041,-0.0311306212,0.0547293276,-0.5184085965,0.0736178234]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1892","title":"request to mirror wmt datasets, as they are really slow to download","comments":"Closing since the urls were changed to mirror urls in #1912 ","body":"Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!","comment_length":11,"text":"request to mirror wmt datasets, as they are really slow to download\nWould it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.\r\n\r\nThank you!\nClosing since the urls were changed to mirror urls in #1912 ","embeddings":[-0.3201408684,-0.2878118753,0.0268324371,0.1421597302,0.0106379557,0.1133044809,0.1318190247,0.4422803819,0.1877802908,-0.0621385351,-0.3263046145,-0.2276874036,0.0757907331,0.0749308392,-0.2242674977,-0.3100861013,-0.137197271,0.1327812374,-0.6658096313,-0.1379882693,0.0306709744,-0.2067516893,-0.1030230001,0.0847297832,-0.0930906907,0.1120902821,-0.1556050628,-0.0157693587,-0.3748447895,0.3074751496,-0.0089850882,0.4012153149,0.1368270218,0.2935676277,-0.0001129378,-0.099204056,0.2300017029,0.1748823225,-0.2311171144,0.0956003442,-0.535531342,-0.3375250995,-0.1658560485,-0.0182706341,-0.0250715073,-0.1312534511,-0.0136193912,-0.0258734692,0.2186397463,-0.1779873073,0.1700704396,0.2847053409,-0.3552664518,-0.1211486906,0.3252402246,0.1695856452,-0.3358907104,0.3851730525,0.3924822211,0.0853654891,0.1287096888,0.2957683504,0.0774516165,0.3790658712,0.3420926034,-0.2025957704,0.0862705633,0.1134861484,-0.2057214528,0.5278211832,0.8242468834,-0.0183595363,-0.32455948,0.1885299236,-0.1809529215,0.3742679656,-0.1438590735,0.184198007,0.0141325882,0.3497406244,-0.3529739678,-0.3738212883,0.0138915796,-0.0097283311,-0.2112909704,0.323364228,0.0623479709,-0.0203316361,-0.0639661998,-0.1474560201,0.2259700298,-0.4838762581,0.1644107252,0.036563728,0.0548369139,-0.387614131,-0.5696231723,0.0255236793,0.2244374901,0.0502159931,-0.0021938244,0.0001673,-0.0795378312,0.0530818477,0.1049923822,0.0707287118,-0.3700878918,-0.2895984948,0.2618850172,-0.0328756124,-0.1887653768,-0.0251486022,-0.0782101303,-0.051132571,-0.5012040734,-0.1053433493,-0.0491803959,-0.4042754471,-0.086934559,-0.2621425092,0.0034311477,-0.1942895353,-0.155181706,0.0428291485,-0.1646634191,0.2633424997,0.3198452592,-0.0643479228,-0.1009789333,-0.1244635135,-0.1096655503,-0.230734393,0.0151077239,0.3793432415,0.1607760042,0.1835088283,-0.0264970083,0.0714589208,0.1871489584,0.2979238033,0.1626498103,0.0999860764,-0.09548489,0.1537763923,-0.2235677391,0.4898343384,-0.0068899621,0.539978683,-0.1829464287,0.1306385398,-0.5362301469,-0.0872573033,0.1179237589,0.2108526528,0.0035420777,-0.2029482424,-0.3437066376,-0.1009114683,-0.4951937497,0.1625241637,-0.0783379152,0.0547622405,-0.1744444221,0.0745491087,-0.0290956683,0.2967530787,-0.2971054018,0.0096940128,0.1010530218,-0.0786645785,0.4400476813,0.56788975,-0.2806084454,0.0694608912,-0.2635285258,-0.133734867,0.1111119241,-0.1084509119,-0.0592944138,0.54352355,-0.3293906152,-0.2201854587,0.1481132209,0.096671626,0.2547278106,-0.2159998566,-0.0261982381,0.4056242108,-0.0409076735,-0.0200780481,-0.3630282581,-0.5139619112,0.2218669653,0.2335493118,0.0658873469,0.2670324445,-0.0446992293,0.219145909,0.5100453496,-0.0130067077,-0.0341541544,-0.0031469041,0.1182924211,0.3207374215,0.1727683544,-0.0021015825,0.1946770102,0.012115228,0.0071187555,0.0108610447,0.5763463378,-0.1128931865,-0.0906414688,-0.0952935144,0.1483992934,0.3308257163,0.0764268637,-0.2565792799,-0.3150259852,-0.273101747,0.0386660732,0.3470095396,-0.2914427817,-0.0492936857,0.4785698354,0.4035878181,0.0491395406,0.5107378364,0.2956504822,-0.1559608579,0.2321447134,0.0272000879,-0.1007668823,-0.1767116189,-0.1278298348,0.3549263179,0.648266077,0.224910304,0.3852214515,-0.6528519988,0.4892866313,0.2645688653,0.118695356,-0.2394483536,-0.200606361,0.1796885282,-0.0564641841,-0.0081601199,0.0878449082,-0.1209948435,0.0065613119,-0.2456300855,-0.064915359,0.1080361381,0.3763082922,0.0884733647,-0.1005980521,0.2181912959,-0.0276720747,0.092574805,-0.0241211411,-0.078388907,0.0999161974,-0.028587101,-0.2825834453,0.2173990309,0.0763569102,0.0671221465,0.1517035216,0.3023700118,0.124123916,0.2473087907,0.2882807851,-0.2529309094,0.0441164933,-0.059447296,0.079712458,0.1977356523,-0.0318736359,-0.1359433681,-0.3088808358,-0.058803115,0.1121968776,-0.0954856426,-0.0609200709,-0.4091857672,-0.4363724291,-0.1340856552,-0.6206345558,0.0629134178,-0.1419766545,-0.2275445014,0.0822052211,0.0729139671,-0.0365143567,0.0819649398,-0.2517956793,0.2632343471,-0.2626538277,0.2251274586,-0.2201605886,0.0935336724,0.109805733,0.0789356381,0.0904804915,0.151131317,0.3001580536,-0.0575323217,0.1111627072,-0.5687984824,-0.0717429146,-0.0752982348,0.3864685297,-0.0887643471,-0.1853168458,0.4462538064,0.1665978432,-0.0541751534,0.0585122071,-0.1235360727,-0.1062498167,-0.1829089671,-0.1778020263,0.030258134,0.0675507337,-0.2415568233,-0.4418850541,-0.1213040426,-0.0686212555,-0.4142852724,0.0459743664,0.0509504862,-0.2035459578,-0.2873601317,-0.1412247568,-0.10078419,-0.2341542393,-0.4118723869,0.4387646019,0.0744708031,-0.2113506198,0.1126198322,0.1420237124,-0.4650194943,0.1762064546,-0.3002626002,-0.0296161883,-0.1267944872,-0.2078947723,0.1997316629,0.1984221786,-0.0270043761,-0.1070001423,-0.0953338891,-0.0951264426,-0.1457083672,-0.0711874068,0.1057383195,0.4388487041,0.1428515166,0.1185459644,0.0119685819,0.9772946835,0.1499729156,0.234776929,0.0945313498,0.3088816106,-0.1360781491,0.2179625481,-0.2374171168,0.0928851962,-0.0539269596,0.105664514,0.4357517362,0.2621480525,-0.2151492834,-0.1010228842,-0.1224421412,-0.0629500225,-0.1598923057,-0.2332875878,0.0668892637,0.2118618488,0.2878441215,-0.1680623293,-0.0912533477,-0.2855665684,0.5250576138,0.6199007034,-0.1092921793,0.0015480587,0.1533150673,-0.0251475722,-0.3112509549,-0.0673388094,0.1414321959,0.3474780917,-0.0192969218,0.2752978802,0.0336303897,0.0507650413,0.5013995171,-0.1344690025,0.4031098783,-0.2277424634,-0.3345046937,0.065394856,-0.1744205803,-0.0449002758,-0.4933713973,0.0577102751,0.2268834412,0.0055875955,-0.0893752053,0.1865164787,0.0136763798,-0.3401004076,-0.3759701848,0.3130290806,-0.1662904173,0.0258938521,-0.3066671193,-0.1111586019,-0.1634335667,-0.2077139169,0.0666865706,0.0570926107,0.2721537352,0.3255658746,-0.275205791,-0.0429448187,0.2955211699,0.316510886,0.1076434478,0.5364947319,0.182333082,0.3519440293,0.2243199348,0.0642141998,-0.3080330193,-0.1656359583,0.0892765895,0.4703482389,0.0500975139,0.1892940253,0.2066586763,0.1883696914,-0.3920236528,-0.015182958,-0.0581860319,-0.0903045982,-0.2000687569,-0.2003120631,0.4898576438,0.2675738633,0.0391195714,0.0615449883,0.0723369569,0.1098116413,-0.3412899077,0.1174326316,1.1421962976,-0.3415795565,0.0819600746,-0.400545001,-0.5496628284,-0.0561825149,-0.3183085322,-0.141850546,-0.2618855834,-0.4626096487,-0.1672808081,-0.1333797872,0.0433199257,0.181987375,-0.0118877217,0.2120315582,0.3940558732,0.2004892677,0.2084562033,0.4233233929,-0.0878353268,-0.1043526307,0.1177266911,0.1153621674,-0.106932655,0.255074501,-0.1436698288,-0.3016398549,0.0917617381,0.1066976339,-0.2414003909,0.1056214944,-0.2431866676,0.0329690613,-0.432736665,-0.0491292253,0.4668371379,-0.6663906574,-0.0133018671,0.3800630271,-0.4930589497,0.1136756763,-0.3295829892,0.0709359199,-0.2764283419,0.1464346349,0.0483733453,-0.1872126609,-0.2041689605,-0.2599578798,0.1510371119,-0.2203843892,-0.244539395,-0.3250757754,0.229109183,0.1948619187,0.0021743337,0.2550162375,-0.0051341387,-0.0131336218,0.1012198925,0.1107691079,-0.3170960844,0.3409248888,0.522890687,-0.205870226,-0.0643102154,0.6444104314,0.175502345,0.0224682949,0.295042783,0.2583117783,-0.2794430852,-0.1670136452,0.1228691339,-0.1009619683,-0.1046884209,0.037028905,-0.1843492538,0.1788042784,-0.0213558655,0.3223059475,0.2598019838,-0.3408122063,-0.2941895127,-0.3789357245,-0.0501657277,0.0879002437,-0.2777827382,0.0233422071,-0.0946675614,0.1944852173,0.1626742333,-0.1082793847,-0.2977977693,0.2999429405,0.1435577124,0.0241341665,0.0278409924,-0.099064067,-0.1392769814,0.2116160244,-0.0321011692,-0.0941747054,-0.1682745516,-0.1945363581,0.0316668935,0.1984601021,0.1413717419,-0.121359922,0.0238189269,-0.4569860995,-0.2013712078,0.3306091428,-0.1136308461,0.1297505647,0.0622294247,-0.0035243039,0.4377677143,-0.3607697785,-0.4622147679,0.523589015,0.3787530959,0.3950980604,-0.0800738782,-0.3016563654,0.1708412915,0.0190627351,-0.1168514788,-0.0827491805,0.2424776256,0.1647912264,0.5014294386,-0.1129535139,0.215316847,0.4649429619,0.5074689984,-0.0114162955,0.1282438189,-0.1914574653,0.2106886506,0.1079068035,-0.2145968974,-0.1927174181,0.3673309684,0.1509806365,-0.0447233282,-0.0178194549,0.2347281277,0.0313559957,-0.1679919064,-0.1251224577,0.2257544696,0.1281460226,-0.2875154018,0.2649760842,0.339176476,0.0926984623,0.0185875278,0.2276931107,-0.0069013173,0.1171706319,0.2411837578,0.3843455613,0.2051965892,0.2365907729,0.126942277,-0.1961148977,-0.0848794281,0.2315552682,0.2305351496,0.0941048935,-0.0247597899,-0.2736377716,0.2797476649,-0.3926746249,-0.1108977348,0.3093758821,0.1065089107,-0.0564616323,0.0857296735,-0.1172195897,-0.133807227,0.1319513619,-0.0982970595,0.0415782221,0.2698492408,0.0972465947,0.2037105709,-0.1728853285,0.0542571321,0.2205591202,0.1487330794,-0.189927727,-0.3218618631,-0.3390257955,0.1345348209,0.4706587791,0.4898872077,-0.2722158134,0.0841081664,-0.0052668559,-0.103488788,-0.1526565552,-0.0481266342,0.1111126542,0.1752470434,-0.0632076114,-0.4407536685,0.3156870008,0.0949254856,-0.1165117323,-0.0429488085,-0.2313705385,-0.4570764303,-0.3354213834,0.18377994,0.2698967159,0.078813158,-0.1063330472,-0.03414765,-0.0973730981,-0.0100471759,0.3311527967,-0.3071192205,0.0497625358,0.1144782826,0.0548596047,-0.3324650526,0.3118374646,0.5370134115,0.3597958088,-0.2904850245,-0.2868002057,-0.2471642494,-0.069685325,-0.3411109746,0.130442515,0.112750493,-0.0380306356,0.2658918798,0.0896183178,0.272785455,0.4220817983,-0.0403515473,0.0986664519,-0.153328985,-0.0246562622,0.009806335,0.2113315463,-0.2478384674,-0.3801898062,0.3214846551,0.1399858594,-0.0223506335,0.0654015243,0.0853591561,0.0865804181,0.0317286514,0.0030252484,-0.2462696731,0.2878150344,-0.0009056222,0.212360248,0.102294594,-0.1151326671,-0.0570633933,-0.2549096346,-0.0963427052,-0.2262342423,-0.0047011459,0.042723354,0.0452063717,0.0499341637,-0.1918734014,0.0775703415,-0.3268816173,0.1058405712,-0.4386679232,0.005843115,0.0483884476,-0.001305114,-0.0709745735,-0.2507548928,-0.1037413254,-0.097372666,0.4216864705,0.0733325928,-0.22131069,-0.360807091,0.0826870352,-0.2779904902,0.4248167574,-0.1094128713,-0.045282878,0.2479967326,-0.1688022166,0.3487218618,0.4388926029,0.2332284153,-0.0915351138,-0.1108137742,0.0748712942,-0.0440808497,0.0345026739,-0.4516977072,0.1096459031]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1877","title":"Allow concatenation of both in-memory and on-disk datasets","comments":"I started working on this. My idea is to first add the pyarrow Table wrappers InMemoryTable and MemoryMappedTable that both implement what's necessary regarding copy\/pickle. Then have another wrapper that takes the concatenation of InMemoryTable\/MemoryMappedTable objects.\r\n\r\nWhat's important here is that concatenating two tables into one doesn't double the memory used (`total_allocated_bytes()` stays the same).","body":"This is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova ","comment_length":55,"text":"Allow concatenation of both in-memory and on-disk datasets\nThis is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova \nI started working on this. My idea is to first add the pyarrow Table wrappers InMemoryTable and MemoryMappedTable that both implement what's necessary regarding copy\/pickle. Then have another wrapper that takes the concatenation of InMemoryTable\/MemoryMappedTable objects.\r\n\r\nWhat's important here is that concatenating two tables into one doesn't double the memory used (`total_allocated_bytes()` stays the same).","embeddings":[-0.363280803,0.1712489426,-0.0166254789,0.1944024116,0.0991945714,0.0310435519,-0.2240344882,0.2879283726,-0.2779931128,0.169799298,-0.0141437314,0.6244731545,-0.0076511414,0.3612303436,0.3155474961,0.0208593663,0.1660713404,0.3219620287,-0.444509238,0.2120204419,-0.210962221,-0.265068531,-0.0310649611,-0.3578844965,-0.0414918177,0.1385575235,-0.3737605214,0.0541282594,-0.4198373854,-0.4219732285,0.0363923497,0.2514683306,0.0210895445,-0.0215105508,-0.0001128412,-0.0243956912,-0.0898242071,-0.159600094,-0.4942531884,-0.0600703321,-0.1329773813,-0.3831800222,0.052267462,-0.0217930917,0.2780458033,-0.2891717255,0.1262459904,0.0966007411,-0.0420539565,-0.0283585805,0.1490674913,0.2059526592,0.3069906831,-0.1393374056,0.4071874022,0.4137899578,-0.2051131576,0.076786302,0.5211777687,-0.2946157157,0.1044791862,-0.2324473411,-0.1127735004,-0.0687844232,0.4034471214,0.2039737552,-0.2281581461,-0.0668898076,-0.0968059525,0.2737793028,0.4659924507,-0.5767402053,-0.3419583142,-0.3257187605,0.2530631423,-0.3041679859,0.0833871812,0.2824950814,-0.1709522754,0.1324291825,0.009371982,-0.3181267083,-0.248142153,0.2170400918,0.3161291182,0.1951753795,0.3829263449,0.0876913071,0.1879309416,-0.0128811775,0.1668949425,-0.1282310337,-0.235522747,0.1022403538,-0.2337997556,-0.0629916787,0.1070392355,-0.031367939,0.6003126502,0.0461593308,0.3129302263,0.0161987264,-0.2231709808,0.2855436504,0.1177802533,0.1156653985,-0.340587467,0.0699794367,-0.0161275622,-0.1747214794,0.224742353,0.0428512581,-0.1124511585,0.0882590339,0.1821427494,-0.2519498169,-0.0506821685,0.1571142972,0.0636103377,-0.2553876042,-0.072543703,-0.1010019034,-0.009234787,0.1719469279,0.1038401723,-0.0815271437,0.0429695882,0.3130611181,0.2300454676,-0.1362041682,0.0179162342,0.0584186204,-0.2758913338,0.4048903883,0.4093623757,0.1198432297,-0.18954283,0.256219238,0.0313365832,0.3089534044,0.0265606288,-0.1521234959,0.3627638221,0.173697412,0.0343497619,-0.5147961378,-0.1038978845,-0.04992944,-0.3467097878,0.5826544762,-0.0074535361,-0.3105188608,-0.2788467705,0.1124335229,-0.1399570107,-0.1520400345,-0.2362221926,0.5187286735,0.2234690338,-0.319429189,0.0532796942,0.1524678022,-0.0710656792,-0.487328738,0.1770173907,0.2461451739,-0.6097146273,0.0000970873,0.1346835494,-0.0355317593,-0.0081753982,0.1414324045,-0.3650355339,0.0534462892,-0.329562366,0.3390133679,0.1483229548,0.051443547,-0.3597134054,-0.1292899996,-0.1312126666,-0.0572623573,0.4833108783,0.4032011628,0.2118994445,0.1453692019,0.2845694423,0.353921026,-0.2226081491,-0.101739347,-0.0969366729,-0.5097315311,0.3827188909,0.0436721146,-0.1091200486,0.055742424,0.1622281224,-0.5239160061,0.262868464,-0.2847255468,0.222206369,-0.1562278867,0.4910975993,-0.0552522168,-0.1267486811,-0.1807769537,-0.3993273675,0.1532471925,-0.2859075963,0.1690541357,-0.2946665883,-0.3228557706,0.2363322079,0.1322880238,0.0230403095,0.1099831238,0.1279896945,0.0328161791,0.0467838012,-0.3443705142,-0.1753308177,0.0709775239,0.197288543,0.050254114,-0.0657289326,0.4222063422,-0.0130110113,-0.2735289633,0.0111992788,0.13805978,-0.0577251241,0.1894703954,0.1517652869,0.4081635177,-0.1975198835,0.1372388005,0.0834626779,0.3099541366,0.1968977302,0.275786072,0.0593566224,-0.547847569,0.1720447242,-0.1497484446,-0.1026823148,0.2931601107,0.1232740954,0.1167148277,0.205556035,-0.1688681245,0.1156383827,-0.0519309305,0.0044764811,-0.1561864316,-0.0496018566,0.2407089472,0.0706171766,0.316755861,-0.3992090225,0.2313849628,0.2559007108,-0.0257454403,0.2315573096,0.1496711373,-0.1381302327,-0.2459967881,0.0611862764,0.0652551278,0.6371228695,0.401491642,-0.0435564183,0.0230936408,-0.0723105967,-0.1035803407,0.1566609144,0.0239395797,0.1919994205,0.2960342169,0.5149371028,0.019223135,-0.1312813908,0.0068405038,0.2437663376,-0.0876248777,-0.1007825807,-0.233731553,-0.3976410031,-0.0855071619,-0.2734545767,-0.2590794265,-0.1498129517,-0.0000415692,0.2618874907,0.3025763035,-0.1651454419,0.131574139,0.1604930311,0.5026214123,-0.4176717997,-0.4203869402,-0.0996631086,-0.1015155166,0.0813330114,0.0966110379,0.3859767318,-0.0796462074,0.4720601141,0.4112360179,-0.0381253064,-0.3536835611,-0.3720272481,0.1493061036,-0.0701523423,0.1334905326,0.1294969618,-0.0610316023,0.1223479807,-0.4504505098,0.0899821892,0.2273865491,-0.3001632094,0.2916746438,-0.030743029,-0.1589653939,-0.3208263516,-0.1860148907,-0.2604093254,-0.4231286347,0.5149369836,-0.0762138367,0.1177768111,0.021703206,0.2126738131,-0.2492764741,0.0497663803,0.1925605685,-0.0910727009,-0.041425854,0.239865467,-0.1205269843,-0.0846003518,-0.158255741,-0.1886270791,0.1327323616,0.4877551198,-0.2162237167,-0.165286243,-0.0899309665,0.4463778138,-0.078002125,0.0622832552,0.628362,0.4974433184,-0.1046437249,-0.068352446,-0.1453668773,0.1930408031,0.2069644779,-0.0469186977,0.0595548339,-0.1752711385,0.0368184932,0.37630862,-0.1202298328,0.1142815202,0.5036056638,-0.0323612876,0.3150956631,-0.1933971494,0.0417080857,-0.195370242,-0.365473479,-0.0698202625,-0.0185965709,-0.1062516794,-0.2606143057,-0.0054411748,0.1965700984,-0.2024159431,-0.246523872,0.3875079453,-0.0537447408,0.2299595028,0.0281378552,-0.4425182939,-0.3240288496,0.054487817,-0.0609602816,0.0671938509,0.1160012335,-0.197233066,-0.3786022365,-0.491250813,-0.3914147317,0.0640260875,0.2047518492,0.1848560274,0.1703853607,-0.416223079,-0.0712061599,-0.0527942851,0.7359611392,0.0948453769,-0.45966205,0.0931817517,0.085306637,0.1662989855,0.0113305543,0.1150954813,-0.0030770381,-0.22919406,0.1979467869,-0.4122438729,0.1610734612,0.1100379527,0.4241368771,0.0007143976,-0.2770532072,-0.0968427137,-0.0321943387,-0.4565469921,-0.3765699565,0.0682289079,0.127480343,0.1043492854,-0.0086606042,-0.22388421,-0.0173047837,0.0462397002,-0.0654447526,0.3290600181,-0.072959803,0.3960427344,-0.3747377694,-0.0979481936,0.8761479855,0.5169813037,-0.2389469892,-0.1651127934,-0.0853904337,-0.1969732046,0.2817299664,0.0562429242,0.1388367414,0.5503252745,-0.0712960139,-0.0063715596,-0.4769187868,-0.1635629088,-0.185014531,-0.2612448633,-0.2375259101,-0.2697760463,0.4483609796,0.2505084574,-0.1940823048,0.3499965668,0.08893314,-0.4363665879,0.3113612235,0.0886842087,1.1590025425,0.0148961684,0.380012691,0.105602324,-0.3035753071,-0.0273733679,-0.1147421226,0.3279806376,-0.2095256448,-0.1091612875,-0.0539950468,-0.2491599321,-0.0816703066,-0.03036399,-0.5228952169,-0.1054128706,-0.0092082778,0.0421193577,0.0221540648,0.1669178754,-0.2668741047,-0.3497603536,-0.0076731304,0.0527729392,0.0839632601,-0.4466324449,-0.1137774214,-0.1879317313,-0.194729507,0.0875311196,-0.2319943905,-0.1305643767,-0.3072328866,0.1639656276,-0.0106166685,-0.1232779846,-0.1258396059,-0.0341868289,-0.1937228292,0.0677935854,-0.1358166039,-0.0937975124,0.0887528583,0.020943474,-0.1650480181,-0.3840017319,0.4343760908,0.0698953569,-0.176159367,0.0155025488,-0.0875432268,-0.3180531263,-0.1743966192,-0.0645118505,0.2385873795,-0.1678339988,-0.1451399773,0.33309412,0.0488711596,-0.2631314397,0.1011422202,0.1563953608,-0.0900350735,0.1993999481,-0.0967756659,-0.0139273675,-0.2291684002,0.0532123223,0.3081642985,0.1532835066,0.1787961423,0.0110157356,0.0155308442,-0.1112135872,0.0246139783,0.1981816292,-0.2538961172,-0.0508713722,0.1319766343,-0.0862403736,-0.047914464,0.1837478131,0.3293776214,0.3658771813,-0.153763473,-0.0920518637,-0.4576088488,0.4486490488,0.1664474607,0.375132829,0.1310441792,0.2656696141,-0.0705385283,0.3580209017,-0.3201295733,-0.1697633415,-0.0794121996,0.1343297362,-0.0008918769,0.1290601045,-0.1456327289,-0.2106573582,0.0542132854,0.1350273937,-0.0068687187,-0.1512795538,0.056904085,0.1238347664,0.2363279164,0.2013338506,-0.3025515974,-0.2956584096,0.0827311575,-0.2491502166,0.1192171648,-0.0134786367,-0.3082534075,-0.0282565299,0.5552067757,0.0543291159,-0.1269169152,0.3359418809,0.0900058597,0.1041655317,-0.0440357067,0.2061877102,-0.400832653,-0.0117807966,-0.0299184509,0.3622285426,0.2243481725,-0.066108115,0.1021795571,0.2828692794,-0.3487553,-0.1331785917,0.178449586,0.2610363364,0.0997982919,0.2106981725,0.286446929,0.1986654699,-0.3570131958,-0.253536284,-0.2535090744,-0.2603913546,0.2006376386,0.2700420618,0.1721268892,0.1027308255,-0.289110899,0.1050470471,0.0649804547,0.0560054183,-0.007339261,0.1163393855,-0.0184892137,-0.0767278597,0.0518959276,0.3210088015,0.2086158246,-0.0023519325,-0.1749981344,0.4798824787,0.1716327667,-0.2350007296,0.2683984935,0.0498675629,0.124815695,0.3608212173,0.3734305799,-0.0093324203,0.2327892631,0.457444042,0.1498824805,-0.367636472,0.2930667698,0.1533664912,0.0267952941,0.0329519324,-0.0564519167,-0.2492745668,-0.2568692267,-0.1258594692,0.0693878084,-0.2811411023,0.4982883632,-0.0051182825,0.1935657412,-0.2537730932,0.3826335073,0.6115159392,-0.0449689887,-0.4045726657,-0.1106404662,0.3154496551,-0.3348842263,0.0536020622,0.1520833522,0.4882132709,0.2802979052,0.0297362767,0.1499826759,-0.137053594,0.0802362487,-0.1953178942,-0.1656139344,0.2606082857,0.1206902936,0.119293876,0.1675411761,-0.0253524873,0.1888162494,0.1565557122,-0.0386964306,-0.4066993892,-0.1166395396,0.4306771159,-0.6115733385,0.3541147411,-0.0224138871,-0.2569141984,0.0828355327,0.3463213742,-0.0350436345,0.0142797362,0.0515129082,0.0750957727,0.3211847544,0.281421423,0.298643887,0.1896282285,-0.3491724133,-0.2060583234,-0.3519130349,-0.0636398494,-0.0258577485,-0.3163741827,-0.0514050089,0.1494465619,0.2918800712,0.0563042201,0.035611216,-0.0549109615,-0.0545144528,0.2851235569,-0.2741577327,-0.1244374141,0.2524220347,-0.0996354893,0.0223415792,-0.2017434835,0.0379134938,0.1373586953,0.0389975905,-0.3926256895,0.0246429294,0.2244713157,0.2718222141,0.5434365273,0.0922377557,0.3047643006,-0.2285113931,0.0017848081,0.0915162712,-0.1626474708,-0.3749506474,-0.032452032,-0.0192977842,-0.0970246643,-0.3878608644,-0.0701935962,-0.1041528881,-0.040986605,-0.1244583726,0.0511250086,-0.2187108845,0.0896853954,0.1657111794,-0.0279463157,-0.0913172066,0.1592532843,0.1434430033,0.0932312459,-0.091666609,-0.2565828562,0.0419634357,0.2645942569,-0.0430804454,-0.0967150852,-0.0352426097,-0.1707375795,0.1734446585,-0.8052763343,-0.2288548052,0.2185244262,0.0149887893,-0.0133392261,0.4682545066,0.0607344173,0.007089593,-0.3281799257,0.3623627126,-0.1387654692,-0.3489482403,-0.1699950397,-0.5305551887]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1877","title":"Allow concatenation of both in-memory and on-disk datasets","comments":"Hi @lhoestq @albertvillanova,\r\n\r\nI checked the linked issues and PR, this seems like a great idea. Would you mind elaborating on the in-memory and memory-mapped datasets? \r\nBased on my understanding, it is something like this, please correct me if I am wrong:\r\n1. For in-memory datasets, we don't have any dataset files so the entire dataset is pickled to the cache during loading, and then whenever required it is unpickled .\r\n2. For on-disk\/memory-mapped datasets, we have the data files provided, so they can be re-loaded from the paths, and only the file-paths are stored while pickling.\r\n\r\nIf this is correct, will the feature also handle pickling\/unpickling of a concatenated dataset? Will this be cached?\r\n\r\nThis also leads me to ask whether datasets are chunked during pickling? \r\n\r\nThanks,\r\nGunjan","body":"This is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova ","comment_length":129,"text":"Allow concatenation of both in-memory and on-disk datasets\nThis is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova \nHi @lhoestq @albertvillanova,\r\n\r\nI checked the linked issues and PR, this seems like a great idea. Would you mind elaborating on the in-memory and memory-mapped datasets? \r\nBased on my understanding, it is something like this, please correct me if I am wrong:\r\n1. For in-memory datasets, we don't have any dataset files so the entire dataset is pickled to the cache during loading, and then whenever required it is unpickled .\r\n2. For on-disk\/memory-mapped datasets, we have the data files provided, so they can be re-loaded from the paths, and only the file-paths are stored while pickling.\r\n\r\nIf this is correct, will the feature also handle pickling\/unpickling of a concatenated dataset? Will this be cached?\r\n\r\nThis also leads me to ask whether datasets are chunked during pickling? \r\n\r\nThanks,\r\nGunjan","embeddings":[-0.3151938915,0.0239811186,-0.0091471439,0.3324131966,-0.0929977,-0.0160011314,-0.1415656209,0.0513885133,-0.1234135851,-0.0343258008,0.012674707,0.508906126,0.015509028,0.4306793809,0.1224206537,0.0571519397,0.1958594322,0.1693573892,-0.4769207835,0.1094922423,-0.3694703281,-0.2283259928,-0.2241987735,-0.4162859917,-0.2465887219,0.0351609178,-0.458506614,0.1797388792,-0.5705538988,-0.2730188668,0.0891333893,0.3404117227,0.0331604257,0.0554767735,-0.0001229169,-0.1215621307,-0.1313509792,-0.1438875347,-0.4560647607,-0.1900413632,-0.4508099258,-0.4511183202,0.0536833927,0.0278961174,0.4002720118,-0.1607562602,0.1130949184,-0.1156750619,-0.0251444653,-0.04614041,0.0328180306,0.136922732,0.2183957547,-0.06498041,0.1313886642,0.4564218819,-0.2105983645,-0.0154868495,0.4654994309,-0.1170650795,0.1297414303,-0.2641114891,-0.0361640975,-0.2052618563,0.4992964268,0.0941745117,-0.419972986,-0.2418172657,-0.1013240814,0.2512696087,0.6391003728,-0.3239811063,-0.4200568199,-0.3742827773,0.1631272137,-0.3263413906,0.1764194816,0.2312669903,-0.0021583738,0.102301985,-0.0673531815,-0.4811716974,-0.2130007297,0.2608930171,0.2544135451,0.2590686679,0.2200955153,0.1209522262,0.0647674575,0.0149428332,0.2294029444,-0.3686328828,-0.1657604724,0.2453332692,-0.3499107659,-0.0451852307,0.1129574478,-0.0423486419,0.5332869291,0.3057990372,0.3247354627,0.0516036898,-0.5262040496,0.2660189271,0.2431846261,0.1561827958,-0.0963965878,0.1706820726,-0.0694975629,-0.1371653527,0.237611115,0.0166346114,-0.1367421299,-0.0226723664,0.0251740646,-0.1259246022,-0.0738968551,0.1945527792,0.1652775854,-0.1814727932,-0.0189234577,-0.2162984908,0.0778371245,-0.0743919238,0.0621999353,0.0568561107,-0.0136168804,0.351400882,0.3560393751,-0.1627322435,0.0549262799,-0.081479162,-0.2937661111,0.3898349404,0.3329309821,0.0588356704,-0.1926192939,0.2399522662,0.0409079604,0.2022750527,-0.1028457433,-0.2446871698,0.4061219394,0.0637177378,0.0030735596,-0.5130050778,-0.0187864676,-0.0329550914,-0.3379784524,0.5166099668,-0.1849387288,-0.3788852096,-0.3063006699,-0.0302096177,-0.1421592981,-0.0467224903,-0.217796579,0.5460851789,0.3333918154,-0.215039894,0.1148192137,0.1583981216,-0.1153829321,-0.484054774,0.1431287825,0.4034638405,-0.4882220924,-0.0285890736,0.0979509056,0.0205815285,0.062108051,0.131243214,-0.3371044993,-0.0256573148,-0.3812057078,0.2984582484,0.330996722,0.059259899,-0.1576823592,-0.007598164,0.0237680133,0.1462474763,0.4264763594,0.3102689385,0.0304750167,0.0982170478,0.3803618252,0.3373641968,-0.172657162,-0.383688271,-0.0291046184,-0.4970319271,0.443430692,-0.138687551,-0.1300171912,0.2080399692,0.2209337801,-0.6818778515,0.1491347998,-0.1081970036,0.2218994796,-0.1341847777,0.3370450437,0.0576104783,-0.0608253144,-0.1623553038,-0.6230329871,0.1108442172,-0.3560378253,0.0800078213,-0.1722413898,-0.2401232719,0.2953734696,0.0905296355,-0.0372997113,0.1557226926,-0.036793761,0.0580456965,0.0346238762,-0.4091594517,-0.4521028399,0.2373283058,0.0837838277,0.0530395843,-0.2692818642,0.334718734,0.1345411092,-0.0918150395,-0.1463078856,0.217555359,0.0298361201,0.1256483644,0.0788572729,0.4706090391,-0.157760486,0.0567779727,0.1230105907,0.3960999846,0.1758621186,0.2842260301,0.069804579,-0.7420755029,0.200221926,-0.2985211611,-0.0355480425,0.3061635196,-0.1743585914,0.2201001048,0.1943643987,-0.1398901492,0.0687406063,-0.0497543104,-0.2310173362,-0.2290154696,-0.0816831142,0.215775758,0.0627810135,0.3361774087,-0.525601685,0.1454780698,0.1831400841,0.0464512818,0.133184284,0.1387510598,-0.0849193856,-0.2781789601,-0.0172249656,0.0637630895,0.7007017732,0.2160703391,0.0262388047,-0.0180966146,0.0740410909,-0.125470683,0.2335386723,-0.0067183003,0.1687889248,0.4369444549,0.3996045291,0.11076653,-0.0637710169,0.0234743711,0.2618205845,-0.2111879587,-0.1233824268,-0.2116711885,-0.436606288,-0.0754052773,-0.2309647202,-0.3501520753,-0.1998724639,-0.0431916304,0.0764245018,0.2593502998,0.0600147992,0.1693040878,0.1015223116,0.7371382713,-0.4011029899,-0.3640144169,-0.0570496209,0.0777585357,0.1655644327,-0.0318345577,0.3715845942,-0.0787614882,0.3381142914,0.2861672938,-0.0335315466,-0.2892336547,-0.2818521559,0.1511435062,0.0167326778,0.0454625934,0.173060447,-0.0999790356,0.0313671529,-0.4025486708,0.0488804393,0.4723199904,-0.2581701875,0.1338032633,0.0513441041,-0.1683002412,-0.2078544796,-0.1897789985,-0.2680081129,-0.3550593853,0.4333010614,0.0087036677,0.084102869,0.1416249424,0.191974014,-0.3095085323,0.223487705,0.1859374791,-0.2188668698,-0.2029943466,0.3382450938,-0.1542316526,-0.0722792447,-0.1777091026,-0.1409837008,0.1397572607,0.4569010437,-0.1309050024,-0.0475999899,0.0050769402,0.4843650758,-0.093959555,0.1180795208,0.7127976418,0.4727385044,0.0355076343,-0.1927271485,-0.1481276155,0.2134426981,0.3343856633,0.0561791174,0.3356004655,-0.2804927528,-0.0089960704,0.34156546,0.1196296737,0.0275281612,0.4800341129,0.1667901576,0.3584747016,-0.1939470172,-0.0502044261,-0.2021125555,-0.4524939656,0.0759777725,-0.1718984842,0.0070285625,-0.117406249,-0.1100477874,0.3091024458,-0.1377835125,-0.1023493484,0.3258583248,-0.3790817857,0.4001640379,0.0627165958,-0.5139105916,-0.3369116187,0.1134469658,-0.1384169608,0.199614197,0.3677077591,-0.2280968726,-0.423145473,-0.3757846355,-0.2592116296,0.0711827129,0.2041254044,0.2274395972,0.1916835457,-0.2750765979,-0.0513800196,0.0553244129,0.7124958634,-0.0236196723,-0.5619778037,0.0524925478,0.0781707838,0.3489019573,0.0581909269,0.2769466937,0.1562776715,-0.2433077395,0.3843099773,-0.3124511242,0.1250409335,0.0606208369,0.4003783166,0.0890041366,-0.2711163759,0.049205903,0.1524653137,-0.3478347957,-0.5306872725,-0.0281419046,0.0754002705,0.1031538546,-0.012837667,-0.3287605643,0.0601395257,0.0363639481,-0.1553476453,0.1983362883,-0.0498030633,0.3714060187,-0.2416535318,-0.0307153668,0.7655333281,0.6723148823,-0.2359549701,-0.2399857342,-0.1107772589,-0.1825509071,0.2655691803,0.0883513391,0.1202687547,0.4290006757,0.0296990499,-0.0531857461,-0.4937260747,-0.134796828,-0.0538786501,-0.0830792636,-0.3994481564,-0.4151044786,0.4616016746,0.1421400309,-0.1688081622,0.5554350615,0.0548573099,-0.4117569923,0.132878527,0.0513511114,1.3276767731,0.2869278789,0.238622725,-0.038914565,-0.2739723921,-0.0141870175,-0.0437281281,0.2706170082,-0.0841446221,-0.1475721449,-0.2070041448,-0.253297627,0.0391065367,0.0096353497,-0.3656955957,-0.1243709102,-0.0491564348,-0.0999156907,0.0220851768,0.0850069374,-0.2020624727,-0.354175359,0.2298353761,-0.0861267745,0.130862385,-0.2966295779,-0.0865637064,-0.140936017,-0.1197372973,0.0594427176,-0.3053862453,-0.1838281453,-0.2568381429,0.1923812479,0.1816962808,-0.1746618897,0.0791647583,0.1379819065,-0.1640588641,-0.074780494,0.0173848271,-0.1131511033,0.2929673493,-0.1268592179,-0.166672647,-0.2928451896,0.5314502716,0.1283067763,-0.0943869576,0.0102132643,-0.1836506426,-0.3001394272,-0.2620705366,-0.0297706537,0.241097331,-0.2740913033,-0.3464109004,0.5255739093,0.111817956,-0.1333803684,-0.0171922706,0.282140851,0.0041771922,0.2388521731,-0.1153029129,0.0447538048,-0.1724219471,0.1028251648,0.3626397252,-0.0572487116,0.1238453239,-0.012466833,-0.0947057158,-0.0052469773,0.1010325477,0.088878803,-0.1637595296,-0.0873330906,0.1111692041,-0.0554686226,-0.1481943429,0.2667476833,0.2715303302,0.536375761,-0.095954746,-0.2259961218,-0.3361230195,0.3077058494,0.0108236177,0.3015132248,-0.0119777666,0.056874685,-0.0652878433,0.4264638126,-0.1813466698,-0.1138811484,0.0087323375,0.0344174355,0.1005261168,0.182461679,-0.2513325512,-0.1998897493,-0.0870505646,0.0121300872,-0.0571954846,-0.0201098751,0.0253750868,0.2010982633,0.3936626613,0.0385296419,-0.3091892898,-0.4045878351,-0.0642538667,-0.301109463,0.0180380922,0.006746714,-0.2550587654,-0.0453521721,0.3187661171,0.1587715298,-0.2169712931,0.4139187634,0.2465452254,0.1790568084,-0.0495940745,0.1369213909,-0.3104901314,-0.0004518197,-0.2512000799,0.3576909304,0.2786886692,-0.0583894737,0.2566403449,0.1928193569,-0.1270216852,-0.2254950106,0.0935561433,0.3229034245,0.1165196598,0.178593263,0.2229982615,0.0470432751,-0.3327697515,-0.2509397566,-0.2289906144,-0.1038742661,0.1372491419,0.2295454592,0.2040006518,0.0853357613,-0.2376068383,0.0376168676,0.0781697333,-0.008050588,0.210460797,0.1018413827,0.0391521044,0.0759052932,0.1027447507,0.443402648,0.1278669387,-0.0603465848,-0.0536372252,0.4059923887,0.3354986012,-0.156401515,0.2840719521,-0.1464847475,0.1905841976,0.4117327034,0.2667860091,0.1027265042,0.1858244091,0.3560812771,-0.090308696,-0.3604811132,0.377574712,0.191888541,0.1005822048,0.0141638611,-0.1952614188,-0.3438718021,-0.3334765136,-0.1246994734,0.0684071332,-0.1662838906,0.5352700949,-0.1510177255,0.1251711845,-0.18672508,0.4833075404,0.4300200343,0.0550655089,-0.307012558,-0.0052052978,0.2363697141,-0.2021860629,0.1345390826,0.1701176167,0.4026604891,0.3949691653,0.1552561074,0.1541089714,-0.1276641935,0.0246710051,-0.034211643,-0.0366836041,0.1181239933,-0.0916448459,0.1320817173,0.0459388457,0.0718483329,0.4185603261,0.1211922914,0.0658814237,-0.480781287,-0.0521919876,0.3336154222,-0.5089948177,0.4396710694,0.0326912142,-0.105412513,0.0539264977,0.4718613327,-0.027870968,0.0252903048,0.0926606432,0.0147380196,0.3638404906,0.2257442176,0.4549018145,0.1053721234,-0.2259773314,-0.2966226041,-0.261364609,-0.1315375268,0.3128865957,-0.3798814416,0.000133438,0.0425660647,0.3286978304,0.0753523037,-0.0420673974,-0.0879049078,0.0292956606,0.2934948504,-0.2248123437,-0.1151871756,0.2837558985,0.1098931581,-0.1114573032,-0.2010266483,0.0119722877,0.3292579353,-0.1495505422,-0.2576143444,-0.0294732731,0.1874913722,0.2973900735,0.3829784095,0.2143354565,0.3231746554,-0.2412548065,0.0565773472,0.1228623316,-0.1036887541,-0.2429079711,0.208593145,0.0174816828,-0.1900770962,-0.5419860482,0.044584427,-0.0664346516,-0.055745434,-0.1266641617,-0.0023196405,-0.2712285221,0.0344975777,0.0491988435,-0.1256289035,0.1763835251,0.3343782723,0.1156432554,0.0935378075,0.0536944941,-0.098595202,-0.0166143365,0.2888440788,-0.0382963642,-0.0881180018,-0.1418971717,-0.0524469353,0.218778491,-0.8657224178,-0.3556312323,0.307570219,-0.1007833406,0.1090641767,0.3817704916,0.0892796069,-0.1302482486,-0.3879982233,0.3571617007,-0.1040115878,-0.3392214775,-0.0915777385,-0.4220617115]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1877","title":"Allow concatenation of both in-memory and on-disk datasets","comments":"Hi ! Yes you're totally right about your two points :)\r\n\r\nAnd in the case of a concatenated dataset, then we should reload each sub-table depending on whether it's in-memory or memory mapped. That means the dataset will be made of several blocks in order to keep track of what's from memory and what's memory mapped. This allows to pickle\/unpickle concatenated datasets","body":"This is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova ","comment_length":62,"text":"Allow concatenation of both in-memory and on-disk datasets\nThis is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova \nHi ! Yes you're totally right about your two points :)\r\n\r\nAnd in the case of a concatenated dataset, then we should reload each sub-table depending on whether it's in-memory or memory mapped. That means the dataset will be made of several blocks in order to keep track of what's from memory and what's memory mapped. This allows to pickle\/unpickle concatenated datasets","embeddings":[-0.3308897913,0.0809664726,-0.0875295699,0.1288803965,0.103787519,0.021383103,-0.0953433961,0.2642484903,-0.0994139612,0.1963127553,-0.0066629774,0.5743992329,-0.0273892563,0.474514991,0.1390598565,0.0199648086,0.1339086592,0.2706728578,-0.4599793255,0.0995579213,-0.2808738351,-0.2470986694,-0.058446072,-0.3346604109,-0.0523938648,-0.0421133265,-0.3391370475,0.1539800167,-0.3864444494,-0.4221208096,0.0044404808,0.2818987072,0.1375907063,-0.0432056189,-0.0001011826,-0.0590702482,-0.0670984015,-0.1675768644,-0.5165483356,-0.096780099,-0.3099037707,-0.3545522094,0.0518829934,-0.134839341,0.250743866,-0.3207660615,0.1727227271,-0.0237123817,0.0291028172,-0.0417876169,0.2448408008,0.2011366487,0.1695047468,-0.1530706137,0.1990844011,0.3594318628,-0.1890633553,0.0671692938,0.414209038,-0.1301229298,0.0097996546,-0.1467058212,-0.0309299231,-0.0901516378,0.3890366554,0.1517840773,-0.1840360612,-0.2165958434,-0.1419532597,0.2824298739,0.5254635215,-0.3413601518,-0.4619012773,-0.2848484218,0.1841118783,-0.3126706779,0.1059342474,0.2865487933,-0.0368905552,0.2204462588,0.0238914266,-0.2196092904,-0.3139277399,0.1313325167,0.241442129,0.2709457576,0.2284069657,0.0479884073,0.0606652237,0.0512919053,0.1425949037,-0.1905992031,-0.1348852515,0.123174265,-0.1949246675,-0.1008890271,0.0449645668,-0.0745258182,0.5886337757,0.2289017588,0.336588949,0.1677314788,-0.3017885983,0.3579454124,0.2023162395,0.0898768529,-0.2345991284,0.0051620994,0.128000021,-0.2076895535,0.117369324,-0.0112337815,-0.1240905672,0.1520697176,0.0693647787,-0.1491604,-0.0841786116,0.1468493044,0.0327503383,-0.226170212,-0.0130927078,-0.1123067364,0.0061755949,0.1253177673,0.0704290867,-0.0281159375,-0.0210920367,0.2245664597,0.2642063498,-0.0813479722,0.0046337512,-0.085058026,-0.2081477344,0.3224073648,0.3119046092,0.0075857211,-0.1021549106,0.1868179888,0.0192195494,0.2130554467,0.1430291384,-0.151851207,0.431381762,0.1036882624,0.0568898804,-0.4961796999,-0.1049027368,-0.0678688213,-0.3789528608,0.4363248944,-0.1076943055,-0.2815758884,-0.3184344471,0.2236296386,-0.0530512221,-0.1553883702,-0.1535331607,0.4602693319,0.098652184,-0.2554016709,0.052607961,0.1726684421,-0.0268212166,-0.4418768585,0.1139305457,0.1807295829,-0.458358556,0.0623890944,0.155356586,-0.1281030625,-0.0389838144,0.1168562397,-0.3931361139,0.0349425897,-0.3964233398,0.341822803,0.2655103207,0.029062802,-0.1269691586,-0.047787372,-0.0275525041,0.0150119141,0.4528053403,0.3481975198,0.2118473798,0.0749009177,0.2160172462,0.3431636989,-0.1340663135,-0.1320328265,-0.0794656426,-0.4177032113,0.3380465806,-0.0189005584,-0.230333671,0.1269377917,0.2450880706,-0.522721529,0.2441468686,-0.3239215612,0.1687000245,-0.1750642061,0.4148915112,0.0176649075,-0.0822701827,-0.1235155165,-0.5488771796,0.1035035551,-0.3118017912,0.0981165469,-0.1812515408,-0.3896619081,0.2531530857,0.0668020546,0.0256447699,0.1171815246,0.24580957,0.0483007319,-0.0197433867,-0.394582063,-0.3163654208,0.0991427004,0.0952568576,-0.0008600721,-0.1155858859,0.2568218708,-0.0641727448,-0.1052489504,0.059379831,0.1430232674,-0.0712554529,0.077618435,0.1616188586,0.433228761,-0.0909077898,0.0953142643,0.1977407485,0.3503227234,0.1969361007,0.3186226785,0.0536822081,-0.43033126,0.1388086975,-0.1494997889,-0.2212162018,0.409717232,0.0697912499,0.1340528876,0.2553195655,-0.1307528317,0.2639632821,-0.0489563569,-0.1614898443,-0.2157241553,-0.1472558677,0.1894982606,0.1447420418,0.3336493671,-0.4741411805,0.2637764215,0.2495567203,-0.0116501376,0.1791906655,0.1175614297,-0.1045674831,-0.2616479397,0.0555108674,0.1884200573,0.5312662125,0.4590286613,0.0591647848,-0.0612866357,0.0004426084,-0.1253446639,0.2002017796,0.0703893527,0.1017497405,0.3079035878,0.3970175385,0.037156824,-0.1925429106,-0.0835520476,0.1319690347,-0.1614086628,-0.0441866405,-0.2051670104,-0.3770576119,-0.0965348855,-0.1838059425,-0.201302588,-0.1883923858,-0.1703035384,0.2789214849,0.2475674897,-0.2998245955,0.1375082731,0.1295550913,0.5335016847,-0.3754701018,-0.1796849072,-0.1084835976,-0.0895538703,0.1045133695,0.1845104247,0.3080826104,-0.0592202544,0.5809780955,0.3431640267,0.0040002665,-0.3314961493,-0.3079931438,0.0987155065,-0.0493037254,0.0598140098,0.112049371,-0.0000598997,0.1649868339,-0.4930720627,0.0386009328,0.2502211034,-0.2884955406,0.1441666037,0.0897160619,-0.171761483,-0.3596302569,-0.3185691833,-0.3232247233,-0.5316622853,0.5087660551,-0.0124017047,0.218919605,0.1244732589,0.083114773,-0.2596382499,0.0898743942,0.1883158237,-0.2038047463,-0.1830642968,0.2191313356,-0.2883405983,-0.142227903,-0.1156650633,-0.1629634798,0.1477711946,0.3719266355,-0.1589659899,-0.1688541621,-0.0243642684,0.3907033801,-0.0688614473,0.0487955287,0.5856058002,0.5143858194,-0.2438393682,-0.1338955313,0.0273836628,0.1994644552,0.2149878591,0.0123744514,0.0917401612,-0.2447971255,-0.0339085497,0.2672213912,-0.0449006222,-0.0083456682,0.4802699685,0.0296960138,0.3378029764,-0.1988239288,0.0846577808,-0.1542255133,-0.4396434426,-0.0395383984,0.0810745731,-0.0181267932,-0.2163593769,-0.0275655817,0.2161120623,-0.1812887341,-0.2459593415,0.3671936393,-0.1107377037,0.2527348995,-0.0106522618,-0.4284535348,-0.2671545744,0.1762675345,-0.0477775149,0.1248852536,0.2901671529,-0.2416695356,-0.4778904915,-0.3409610391,-0.3137471378,0.1387543082,0.2407148927,0.1306317151,0.2050717622,-0.4547716975,-0.0807619989,-0.008548134,0.6017181277,-0.0057809106,-0.4994778931,0.0726114437,0.0704672188,0.2290144861,0.091851823,0.0698179081,0.0356098451,-0.1553216279,0.2879039347,-0.340005219,0.0505487844,0.0386342481,0.3517752588,-0.0046850042,-0.289387852,-0.0051389653,0.0565502569,-0.4865798354,-0.4271866977,-0.0341630727,0.012153646,0.0056691994,-0.1888551414,-0.2518093884,-0.0151931457,0.0212318711,0.0047650649,0.2259651721,-0.0365257896,0.3290243447,-0.2873803675,0.1535702497,0.7977833152,0.5349408388,-0.3412919044,-0.0700025335,-0.0637580827,-0.246235922,0.2890383303,-0.0018982087,0.1382666081,0.4825799167,-0.0300123151,-0.0126197673,-0.5641716123,-0.1735904962,-0.1112964675,-0.1793776751,-0.3051040769,-0.3777842224,0.4603811204,0.120141767,-0.2521436512,0.3317560256,-0.0150275948,-0.4221543968,0.3290462792,0.0887539163,1.1225887537,0.1779592335,0.3433840573,0.0340351164,-0.2753273845,0.0926774517,-0.1129648611,0.2278164774,-0.186744228,-0.1238155439,-0.1344875395,-0.1599748433,-0.0595982075,0.029115947,-0.5210747719,-0.1095253825,0.0125245107,-0.0728692487,-0.1100044623,0.2566878796,-0.2936607301,-0.3604159057,-0.0380547941,0.1260079592,0.1341444701,-0.3623820543,-0.0884259641,-0.2095151842,-0.1346682459,0.087835446,-0.1085635796,-0.1660427004,-0.221591264,0.1632595062,-0.0204982515,-0.0961077958,-0.1961719394,-0.0393272638,-0.1402605176,0.0502508432,-0.1339435428,-0.1060590521,0.0702747032,0.0893014595,-0.1681358367,-0.2926282883,0.5270134807,0.0715628341,-0.1033334658,-0.0052105705,-0.117736496,-0.311301291,-0.1454510093,-0.0112222414,0.1119057238,-0.2358527035,-0.0929558203,0.3725411892,-0.1225112453,-0.2287290841,0.1796438396,0.1737488955,-0.0622076578,0.2706395686,0.0244432986,0.0373503827,-0.2022334784,0.1054246649,0.2561534941,0.0977666602,0.1157249734,-0.0616130307,-0.0979964584,-0.2150928378,0.1079407111,0.1257083267,-0.0852989405,-0.0342697799,0.1577449143,-0.1305825263,-0.0722839758,0.2747027576,0.2969365716,0.3914657235,-0.1860088855,-0.1561176926,-0.4850620031,0.3692664504,0.0839186236,0.4723112285,0.0426459797,0.0909435749,-0.0972829238,0.3439353406,-0.4360330999,-0.2447360307,-0.1786685884,0.1886477172,0.0621032938,0.1072232202,-0.1068103164,-0.163298592,0.1243667156,0.0660277009,-0.0521617085,-0.2705054581,0.0005366561,0.0886474475,0.2666463852,0.1510404646,-0.2268205583,-0.238851577,-0.009021447,-0.2859176993,0.0984269604,-0.0472494923,-0.2804762125,0.047379557,0.4312934875,0.1661451161,-0.0911777765,0.2814756036,0.1302234232,0.1757025272,-0.0284208152,0.077483058,-0.526586175,-0.0447006188,0.0518370606,0.3191321492,0.2118453383,0.006655802,0.2070166022,0.3998593092,-0.2470119447,-0.15974693,0.2610473335,0.247299701,0.0784810111,0.1721436977,0.2526983321,0.320116967,-0.3753237128,-0.2055064589,-0.2421291023,-0.3017142415,0.2398989797,0.2074735463,0.2079684734,0.0633592755,-0.2070703804,-0.0001582508,0.1082522124,-0.0299702249,0.0974069834,0.0044747237,-0.0765742213,-0.0357471071,0.031070279,0.2621293068,0.207734704,0.0396182537,-0.1450176835,0.4661249816,0.3491328061,-0.1293599308,0.2394710034,0.0307021216,0.2176368535,0.3183438182,0.3328393102,0.0937069356,0.2749662101,0.2630879283,-0.0392217971,-0.3378157914,0.3260612488,0.1355720311,0.0584687814,-0.0337924436,-0.2106035501,-0.2294219881,-0.2200936079,-0.2160734534,0.0881617069,-0.2070176303,0.3964641392,0.0546578094,0.2351972312,-0.1943439692,0.3281579316,0.3741763532,0.1182761639,-0.3072229624,-0.0797842965,0.3700535893,-0.1761287749,0.1467443258,0.198633045,0.3817968071,0.310108006,0.0776183009,0.1682849377,-0.153634578,0.0742752701,-0.2141852677,-0.1156191751,0.1510536373,-0.0136951562,0.2354594022,0.2452613711,-0.0899587125,0.3030724823,0.1031544209,-0.0936660841,-0.3618723154,-0.0470553972,0.5336930752,-0.5586107969,0.2597325742,-0.0543600172,-0.257538259,0.0155447843,0.3132825494,-0.0733169913,0.100257203,0.0567995086,0.1455907375,0.289175123,0.4204463065,0.3056252003,0.1715645641,-0.2166493237,-0.1849778593,-0.2979390025,-0.0502006412,0.1012847722,-0.2598644197,-0.0731844679,0.0629110783,0.2843534946,0.0627029836,-0.045171883,-0.0960493162,0.029861005,0.2053635567,-0.3371421099,-0.1072490737,0.248449564,-0.0365940668,0.0122252163,-0.2142079324,0.0440489538,0.2075444311,0.107770279,-0.2285483479,0.0429880507,0.1554294676,0.2202722877,0.5512210727,0.1182145551,0.2511795163,-0.2976635396,0.0068757026,0.0886663571,-0.1955954731,-0.3251899779,0.0286133531,0.0679929033,-0.1258863807,-0.2877640724,-0.1205191165,-0.0775560588,-0.0381673872,-0.1025922224,0.0462992452,-0.227287516,0.1190911606,0.1525843143,-0.1543568671,-0.0258812569,0.2455315888,0.1390735954,-0.0061064102,-0.0663329586,-0.2600612342,0.0703028589,0.266022712,-0.0264014751,-0.0838250369,0.0239696577,-0.0870698392,0.2527748644,-0.763025403,-0.2411406785,0.2996515036,-0.1082538515,-0.0248060133,0.4120289683,0.0739523843,-0.0951400027,-0.3477320671,0.3045030534,-0.0662197843,-0.3062162101,-0.1457571536,-0.4974688888]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1877","title":"Allow concatenation of both in-memory and on-disk datasets","comments":"Hi @lhoestq\r\n\r\nThanks, that sounds nice. Can you explain where the issue of the double memory may arise? Also, why is the existing `concatenate_datasets` not sufficient for this purpose?","body":"This is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova ","comment_length":29,"text":"Allow concatenation of both in-memory and on-disk datasets\nThis is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova \nHi @lhoestq\r\n\r\nThanks, that sounds nice. Can you explain where the issue of the double memory may arise? Also, why is the existing `concatenate_datasets` not sufficient for this purpose?","embeddings":[-0.3921350837,0.1095163152,-0.0102479458,0.3699243367,0.0255508721,0.1648331285,-0.2444256991,0.199427858,-0.1821034551,0.1366394758,0.0495681241,0.4666678607,0.0200924296,0.3617293537,0.1092143208,-0.0306011587,0.160487473,0.2564076483,-0.401722312,0.1629248708,-0.3001021743,-0.1728203893,-0.1916837394,-0.4789800048,-0.1616440713,0.1648575813,-0.4099677205,0.1326929331,-0.505335331,-0.3240083456,-0.0337188505,0.2305967659,0.1441221833,0.1116464809,-0.0001183449,-0.139306128,-0.0883800983,-0.1393118948,-0.4981412292,-0.13762559,-0.3153511882,-0.4476878643,0.0293858964,0.0132387895,0.3109204471,-0.307611227,0.0527275875,-0.1188952774,-0.0023293418,-0.0532878786,0.1089298204,0.060827665,0.2846040428,-0.1830016375,0.1651664227,0.4658094347,-0.2200627178,-0.0311097912,0.4206276834,-0.1952694058,0.2418729365,-0.2360848486,-0.1178748235,-0.1633916348,0.3987491727,0.1727331281,-0.3037188649,-0.1314041764,-0.0412971489,0.2176211774,0.5865792036,-0.4311450422,-0.3342105448,-0.2628730536,0.2460080832,-0.2954332232,0.1332033873,0.2357889414,-0.1375348568,0.1732293665,0.0094016939,-0.4855881631,-0.1796577573,0.2952824235,0.27376315,0.2759509087,0.3529571593,0.1340430677,0.214188993,0.0620447025,0.2141938359,-0.2129987031,-0.2546743751,0.1283379346,-0.3424053192,-0.0140579948,0.0599339753,-0.1500598043,0.5483281016,0.1297712177,0.2871190608,-0.0448517241,-0.4067798257,0.3076332211,0.0919863284,0.2553535104,-0.2252778858,0.1636935472,-0.0637872741,-0.2130728662,0.3011719882,-0.0163453389,-0.1694171429,-0.0730684251,0.2098757029,-0.2031953931,-0.0720810965,0.1567242891,0.0766474977,-0.2511428297,-0.1469202191,-0.1524340361,0.093894586,-0.0379308611,0.139798969,0.133142367,0.1158497259,0.4211546779,0.302969873,-0.1989288926,0.0078565152,0.0276407059,-0.2613214254,0.3779640794,0.353638798,0.1148882285,-0.1973733306,0.3130430281,-0.0070334114,0.1946331263,-0.0352153406,-0.1502510458,0.312420547,0.1093138233,0.0563113615,-0.4541546404,-0.0355091467,-0.0233203173,-0.2534182668,0.6008579731,-0.1377720535,-0.3231899738,-0.503111124,0.0469376184,-0.0669114292,-0.023708418,-0.2471259087,0.5314822793,0.3372127414,-0.2383345664,0.061134547,0.1914597005,-0.1751529723,-0.4846293628,0.221705094,0.3638979793,-0.4608438015,0.0812958777,0.0106251603,0.028483903,0.0968002081,0.1960922331,-0.3160741627,-0.0223341174,-0.4253902137,0.2173309475,0.2514075935,-0.0291145034,-0.0864312351,-0.0521700643,-0.158220917,0.0194218103,0.5092132092,0.4111901522,0.053894639,0.1372525841,0.4314586222,0.2517196834,-0.1865962893,-0.2832247615,-0.0806560069,-0.5113966465,0.4175637066,-0.1091271415,-0.1119985878,0.1417613477,0.1748517305,-0.5082270503,0.2905801535,-0.1933499873,0.2741513252,-0.144450143,0.3921687007,-0.0364353508,-0.1589792222,-0.1589272618,-0.5136506557,0.161558643,-0.1798149347,0.1487389952,-0.1148836538,-0.2364256084,0.2844363153,0.0667648837,0.0482148789,0.1300471872,0.0389023907,0.0784216598,0.0149993943,-0.352771759,-0.2117245048,0.2260375321,0.1711194813,-0.0062981127,-0.1863906682,0.3505262733,0.0912903994,-0.0637104735,-0.0793347806,0.201258868,0.0118027693,0.236667648,0.0625833422,0.4659431875,-0.1970269829,0.0314246342,0.1769622713,0.3687469065,0.142022863,0.2543242276,0.0953699052,-0.5583907366,0.16449821,-0.2625587583,-0.0583055876,0.2322974205,0.0301949233,0.1346611083,0.054290507,-0.1179548576,0.0882850662,-0.0218894891,-0.0929088742,-0.2552785277,0.0091869179,0.2155604511,-0.0128443493,0.3394584358,-0.5138894916,0.1045169681,0.1417919844,0.0801378936,0.1494958401,0.1846862286,-0.0975309014,-0.2499236465,0.0110410387,0.1050735936,0.7278345823,0.3327358961,-0.0270749852,0.024019016,-0.0232421663,-0.1357819289,0.246051237,-0.0523551889,0.2259603143,0.3271225095,0.5175660849,0.0978712067,-0.0231261924,0.0357372314,0.280864507,-0.0925224349,-0.1162450835,-0.2990508974,-0.4070843458,-0.0361857377,-0.2963101268,-0.2642375529,-0.1862796843,-0.031216355,0.1388061345,0.3211254179,-0.0032379187,0.074707143,0.1011650413,0.5987221599,-0.3522721231,-0.4385094345,-0.0326087326,0.0550653003,0.0571304299,0.0250616223,0.4078399241,-0.1361310035,0.4487454295,0.4095326066,-0.1099669784,-0.363945663,-0.3495539129,0.0442864634,0.0677511171,0.0573991463,0.0872165337,-0.1416820586,0.1894736737,-0.4799762666,0.0716041028,0.4400590956,-0.2252619267,0.262676388,0.0469375364,-0.1753203124,-0.3075061142,-0.1543862969,-0.1919998527,-0.4514460266,0.4154670537,-0.1236037612,0.0698882118,0.0775341094,0.195018664,-0.3288708627,0.1795968115,0.1767162085,-0.2035758644,-0.1530013978,0.3337669969,-0.1142297164,0.0010321037,-0.1347980052,-0.1396705955,0.1908705533,0.4543942511,-0.1955454499,-0.1592120081,-0.0649244711,0.3299049735,-0.2098087221,0.1074899286,0.6054486036,0.4714114666,-0.0591280423,-0.0853744671,-0.2051151842,0.254046917,0.2081807405,0.1479589194,0.1304111481,-0.1937831938,0.0110789984,0.2495637983,0.0844728053,0.1532485932,0.4600158334,0.0520453788,0.3349918425,-0.2104030102,-0.1420198083,-0.2575081289,-0.3768157661,-0.0843170211,-0.1151893064,0.000957619,-0.21539177,-0.1009144932,0.3079533279,-0.178628251,-0.212446928,0.3111824691,-0.2671615779,0.3104088902,0.0059292465,-0.4796599746,-0.3252386153,0.0819748417,-0.1227235794,0.1870945096,0.1271075904,-0.1675873846,-0.3169616461,-0.412825197,-0.326481998,0.0829285234,0.1736974269,0.2987652719,0.151991874,-0.2542053759,-0.0798253939,0.0172684528,0.8591117263,0.0638372302,-0.5751929283,0.0738103986,0.0964723304,0.2353251278,0.0099884002,0.2178262174,0.1040728539,-0.2128963917,0.249107182,-0.3420879841,0.1586687565,0.1269883662,0.4117609859,0.0143966395,-0.3251320422,-0.0357558951,0.1081974879,-0.3782021105,-0.5106711984,0.1149502322,0.2176660597,0.0696090385,-0.0149386246,-0.2205721885,-0.0238578264,0.137072444,-0.1835566461,0.3332089484,0.0676083788,0.4472166002,-0.2835330963,-0.1221532375,0.8011996746,0.6519256234,-0.2067798227,-0.2765523195,-0.0635087043,-0.1825507879,0.3542743623,0.0533775613,0.0887246951,0.4914827943,-0.0025876504,0.0044300817,-0.4980612099,-0.1264448315,-0.0918327719,-0.1753398329,-0.4002375305,-0.36676386,0.45652017,0.2283899337,-0.1706124246,0.5208793879,0.0523545146,-0.4474896491,0.195593372,0.0214555264,1.2458560467,0.1081591174,0.3011713326,0.0293310154,-0.3577950895,0.0946925282,-0.0979986638,0.3106344342,-0.1265079826,-0.0006995346,-0.0910703093,-0.2204681337,-0.0275888126,-0.0188457631,-0.5232799053,-0.0967271328,-0.0838886052,-0.1347823143,0.0285350177,0.2104777992,-0.2344261855,-0.3641813695,0.2192934901,-0.0080284402,0.0928410441,-0.4051949084,-0.133577913,-0.1685975194,-0.131226629,0.0406673588,-0.4389455318,-0.1573922187,-0.2707947791,0.289229691,-0.0143481418,-0.1908162683,0.0505688563,0.0598167293,-0.0464057326,0.0665043145,-0.075132668,-0.1513553411,0.1128890142,-0.0322073102,-0.2315134406,-0.4815507829,0.4041180313,0.131048888,-0.2263631374,-0.0356833003,-0.0952384621,-0.3578651845,-0.2171877772,-0.0105715962,0.2572179735,-0.148679167,-0.3082504272,0.4176059961,0.1546215117,-0.2415665686,0.0456593968,0.3068842888,-0.1014374569,0.098888129,-0.1634719521,-0.0184570979,-0.1666097641,0.1040518954,0.3685367405,0.0014765022,0.1319414377,0.0362788551,0.04892499,-0.0144577464,0.0155183645,0.1414501518,-0.1418099254,-0.0327542387,0.1646201909,-0.1074483395,-0.1131366342,0.1744044274,0.3290681243,0.4423860013,0.0320083424,-0.2129263133,-0.4008006155,0.3655231297,0.0545670986,0.3841343224,0.0187724978,0.2157429606,-0.1020368785,0.4042104185,-0.2356408089,-0.1241043881,0.0039592679,0.0871532261,0.0489385277,0.189533785,-0.1527881026,-0.1534685194,-0.020730868,0.0457588769,-0.0048320992,-0.0742734894,0.0333284698,0.1710641086,0.2575355768,0.1527629644,-0.219035849,-0.4636068046,0.0118623031,-0.2358222008,0.0221935958,0.1075063497,-0.2809910476,-0.003731346,0.3735970855,0.0267493334,-0.1658876985,0.3852543831,0.1266210526,0.1624551117,-0.0302153286,0.2613482773,-0.3178707659,0.0073678801,-0.244767189,0.3067189157,0.1848434508,-0.0801896602,0.0112203863,0.2754110098,-0.2506646514,-0.1051611826,0.1765163243,0.2280173004,0.0846747085,0.284799993,0.2059252709,0.1205364689,-0.3507790565,-0.2350205183,-0.3103940189,-0.2013255358,0.1200513989,0.2938491404,0.1429552883,0.0684170648,-0.2992199957,0.0443540961,0.0560022034,0.0071748164,0.1238171309,-0.014543633,-0.0549441166,-0.0123574762,-0.0364688076,0.2724403739,0.1355479807,0.0482492037,-0.0523110963,0.4605272114,0.3122769296,-0.193119511,0.1669672281,-0.0077454899,0.2087924629,0.3916075826,0.2493345886,-0.0104128271,0.2095508724,0.392691493,-0.0135431383,-0.2855477333,0.2348793596,0.136574626,0.0337786861,0.0279449169,0.0244911294,-0.316383332,-0.2760235667,-0.1336278021,0.1761220396,-0.2296824902,0.5138269067,-0.0857443139,0.0736973807,-0.209164843,0.3786339462,0.6538555026,-0.0343051031,-0.4184261262,-0.0995008945,0.3698887825,-0.3635704517,0.1004637182,0.278180331,0.531565845,0.3679253757,0.217149958,0.0727916807,-0.1131784767,0.0552785695,-0.1523777246,-0.1277520508,0.2600960732,-0.0165481493,0.0902413502,0.1100357249,0.021406116,0.2997889817,0.1268080771,0.0509665534,-0.5596282482,-0.0497354418,0.354636848,-0.5050176978,0.3200531304,0.0705103502,-0.2276426405,-0.003229338,0.3698559999,-0.0468129441,-0.0114682037,0.077433452,0.0423710234,0.3519233763,0.1842328906,0.3905685246,0.1013742611,-0.3784219325,-0.3028891087,-0.3700412512,-0.0803645775,0.0273215827,-0.2456855178,-0.080717735,0.1697350442,0.2687168717,0.0672250763,0.0479575507,-0.0700803697,-0.0411723703,0.2574210465,-0.2703314722,-0.1060350165,0.3794722259,-0.0613822117,0.0117305527,-0.2595916986,0.0768876299,0.1721324921,-0.0374334827,-0.3302075565,0.1232138053,0.2427935898,0.3225736618,0.5413458943,0.2077284157,0.2530894578,-0.2301815599,-0.0335408263,0.0080999658,-0.135127306,-0.3178330362,0.1035705209,-0.0106344819,-0.05775024,-0.6203406453,-0.0652145371,-0.0586421192,-0.0344107971,-0.1914424002,0.0050816415,-0.3712550402,0.1478951573,0.0911393985,0.0203422494,0.0294464212,0.2493439764,0.1452340484,0.1393389404,-0.0469747297,-0.2207223624,-0.0059856917,0.3501097858,-0.0742328241,-0.1108440161,-0.0415952951,-0.1768174171,0.2330031097,-0.8800831437,-0.3228906989,0.2580394447,0.0318023302,0.0130913146,0.5286701918,-0.0296338666,-0.1114498749,-0.2562058866,0.3106476963,-0.1890333742,-0.371129632,-0.0093600918,-0.4192746878]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1877","title":"Allow concatenation of both in-memory and on-disk datasets","comments":"Hi @lhoestq,\r\n\r\nWill the `add_item` feature also help with lazy writing (or no caching) during `map`\/`filter`?","body":"This is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova ","comment_length":16,"text":"Allow concatenation of both in-memory and on-disk datasets\nThis is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova \nHi @lhoestq,\r\n\r\nWill the `add_item` feature also help with lazy writing (or no caching) during `map`\/`filter`?","embeddings":[-0.4765581787,0.0608365983,-0.0707229972,0.0521826521,0.0756110996,0.0150534501,-0.2226113826,0.343570739,0.0733319372,0.0877744108,0.0297729317,0.613764286,-0.0133163817,0.4729108214,0.1156108305,0.0217176229,0.1311561614,0.2873371243,-0.4156746864,0.158914268,-0.242046088,-0.2265802175,-0.0592831299,-0.4308310449,-0.0040398268,0.0221989062,-0.2348327935,0.0460517295,-0.4015104473,-0.3713645041,0.0051823263,0.3903060555,0.0715252161,0.004581973,-0.0001091117,-0.0961322635,-0.0213993974,-0.1660830975,-0.4541787803,-0.1732280552,-0.3239759505,-0.3942295909,0.0978121907,-0.05641184,0.2198666036,-0.1939130872,0.1246484667,-0.0921968073,0.0510853529,-0.0005946931,0.1770361215,0.1682467163,0.1792136729,-0.0644020513,0.2513578534,0.4772345722,-0.2069153339,-0.016978953,0.5437409878,-0.2866382599,0.0655305386,-0.0498586558,-0.1694198102,-0.1345374435,0.5407468677,0.1566751301,-0.0944058001,-0.3209829032,-0.0573862791,0.2175630182,0.4746558964,-0.4267606437,-0.4392644465,-0.2462813109,0.1049217731,-0.1887442172,0.0964329839,0.2361290753,-0.1394093335,0.2127306908,-0.0762295052,-0.3423056901,-0.1845762432,0.2520519197,0.3293735683,0.3210098445,0.3376605511,0.0152412867,0.1598050743,0.0632552952,0.2139872462,-0.3032201827,-0.2217297703,0.1727042794,-0.2810246944,-0.0871396512,0.0919322819,0.0622821264,0.4882809222,0.0839881524,0.3856853843,0.114587307,-0.2556549609,0.3995407224,0.1065206602,0.049226664,-0.1860399395,0.1075042263,0.0387987047,-0.2714679539,0.113678135,-0.0038706104,0.0327558108,0.0663645566,0.1317946315,-0.2102599889,-0.1473457217,0.1261440814,0.1737753451,-0.2988483906,-0.1008138955,-0.145976156,0.044921305,0.0515245646,0.1264670789,0.0187107194,0.0435861163,0.2391199768,0.3044169247,-0.0500586219,0.0550269522,-0.1532796025,-0.1885490119,0.3325956762,0.3076521456,-0.0385489762,-0.1532024145,0.2341603935,-0.0660620555,0.2304021418,0.1649900824,-0.168360576,0.4443362653,0.2302446365,0.0097289821,-0.4723738432,-0.0464850999,-0.0752672181,-0.399423033,0.4876351058,-0.1724241227,-0.3116793931,-0.2956554592,0.1612209082,-0.0140455505,-0.0826084539,-0.1136311889,0.549939692,0.1494811028,-0.4128620923,0.0774163008,0.1391511112,-0.2023865432,-0.4482164085,0.163980633,0.2338346839,-0.5422617197,-0.0117075182,-0.0156527236,-0.011032261,0.0200100616,0.1657969952,-0.3622570932,0.0428361334,-0.3063444793,0.260096401,0.3495652974,-0.0053011328,-0.2179085463,0.0879691169,-0.1191324964,0.0412280634,0.4386925697,0.4519631565,0.2561971843,0.0754784569,0.4115336239,0.3871443272,-0.2556595802,-0.1236731485,-0.0714348406,-0.395652622,0.3593601584,-0.0549358614,-0.2022868544,0.1958160698,0.2372691184,-0.6348543167,0.1601854116,-0.2363996208,0.2605082691,-0.1372983158,0.5229123831,-0.0187939592,-0.1351219267,-0.1498906463,-0.476847142,0.0531755909,-0.2135912329,0.0618376061,-0.2344727963,-0.3581313491,0.2459336817,0.1108099297,0.0520032384,0.1000967547,0.1329471767,-0.0339875035,0.0112846186,-0.3725003302,-0.2047203183,0.1839380264,0.1429634541,-0.0163461883,-0.0828609243,0.3200978637,-0.0363274552,-0.1325560659,-0.043002598,0.1616273969,-0.0701386184,0.1692525446,0.0912618712,0.3741360307,-0.2031278759,0.1750642359,0.1850980967,0.462964505,0.2235046774,0.2856938541,0.0593158044,-0.4190938175,0.198394984,-0.2558090687,-0.2348174304,0.3836270273,0.0461267196,0.1875648052,0.1859799474,-0.0598320961,0.2270471454,-0.024106767,-0.1483356804,-0.2353376895,-0.0343403146,0.1370093822,0.0364806391,0.3631222248,-0.318318218,0.237388432,0.2140039951,0.023440402,0.1837716103,0.2534226179,-0.0868161023,-0.3094965518,0.0938750431,0.1604813486,0.599206686,0.4242342412,-0.0878359452,-0.0488775447,0.0394226462,-0.0914463773,0.2606401145,0.0424238332,0.1299591511,0.3358275294,0.5084390044,0.1565901637,-0.1096974462,-0.1188924462,0.1821382493,-0.1932054609,-0.0549293421,-0.1851835698,-0.2611820102,-0.0383874923,-0.1816255152,-0.1657211334,-0.2075845599,-0.0977215841,0.2218884379,0.3331531882,-0.1846591234,0.1372748315,0.1484766454,0.6496680975,-0.2763092816,-0.4283191264,-0.1805220842,-0.1743451208,0.1853334308,0.0614532158,0.3481795192,-0.0459122695,0.4973226488,0.4200582504,-0.0873132199,-0.3659249246,-0.4064377844,0.0658799931,0.0051850188,0.1079395935,0.0827780366,-0.0497721769,0.1289793551,-0.5469331741,0.0883527696,0.2471522987,-0.3253917396,0.1705880612,0.0624139532,-0.1377744526,-0.3601431847,-0.1490687877,-0.2621841133,-0.5986356139,0.4414026141,-0.1484771073,0.1800319105,0.0249648765,0.1966509968,-0.4081294537,0.0214182399,0.2013935149,-0.1242209822,-0.1794524491,0.2523640394,-0.15564394,-0.0787100419,-0.1698462367,-0.2005185932,0.0577689856,0.5631287098,-0.1712462604,-0.207875967,-0.0578831322,0.4450137913,-0.0747942254,0.0141645949,0.5879991651,0.5294589996,-0.1370869577,-0.1355694234,-0.1235170662,0.1481075883,0.2665321231,-0.0376133546,0.1700484306,-0.069788307,-0.109271057,0.3424431682,-0.0055072121,0.0933607072,0.4374438226,0.152698189,0.3345212638,-0.2667293549,-0.0711629689,-0.2311915904,-0.4024605453,-0.051034078,0.0386339352,0.0045275395,-0.3420058489,-0.0554988012,0.225002557,-0.1947951764,-0.2088430822,0.3794307113,-0.1374755651,0.3802072406,-0.0168420766,-0.4671304524,-0.3867400587,0.0967373997,-0.0300535224,0.1141522154,0.2807101607,-0.2234766036,-0.4105630815,-0.3334210813,-0.4103838205,0.0940469354,0.2343839407,0.1793721765,0.0772437379,-0.3287862837,-0.1084486842,0.0249274578,0.7993981242,-0.1190702841,-0.4681967497,0.0222506523,-0.0270182006,0.2350582778,0.1083079949,0.1514734626,0.1005419791,-0.0470658056,0.3932628632,-0.2227915078,0.1073616669,-0.0357364081,0.2548595369,0.0443054922,-0.2726075947,0.0777119324,0.0755954683,-0.4649639428,-0.4787923694,0.07301981,0.1387930959,0.1561548263,-0.1237868816,-0.1623479575,-0.003798814,0.0645300373,-0.0983877406,0.2371671945,0.0146076428,0.3409413099,-0.2411071062,-0.0266855266,0.6412107348,0.4557238221,-0.3451505005,-0.1704420745,-0.1250288785,-0.2065743059,0.325262934,0.0364197977,0.1780779958,0.5699098706,0.0409947895,0.012361601,-0.5910294056,-0.1712291837,-0.1487493515,-0.0998117849,-0.3636200726,-0.3279721439,0.466222167,0.2714466155,-0.238114208,0.4522881806,-0.1000937074,-0.4977033436,0.4367558062,0.0994019136,1.1413235664,0.1355963796,0.3166122735,-0.1556839049,-0.2941762209,0.0338794887,-0.156363979,0.3443221152,-0.0829754323,-0.1609973758,-0.1484382749,-0.1936960816,-0.0601992831,-0.0071586007,-0.5996420979,-0.0995764881,0.0189619903,0.0549626872,0.0502853431,0.1950850934,-0.2405226529,-0.4536107183,-0.0444339849,0.0430857725,0.1018247306,-0.3034266233,-0.0788565427,-0.2548260689,-0.0428844281,0.0972741917,-0.2213405669,-0.2444440871,-0.2668029666,0.2151144296,0.006361491,-0.1317189783,-0.0095420089,-0.0237924214,-0.1792587042,0.0970389843,-0.1027620956,-0.1221287176,0.20269081,0.0986105949,-0.1388660222,-0.3898612857,0.4284768403,0.1266145557,-0.038052585,-0.0174508952,-0.1661405116,-0.4061594009,-0.1345542818,-0.0337523334,0.2091445327,-0.118443273,-0.262583077,0.30845052,-0.0782311708,-0.2305266708,0.1201270446,0.1115125492,-0.097948201,0.2626572549,-0.0174360126,-0.0027792526,-0.1698483229,0.0950914025,0.2190250158,0.1056209654,0.1143998727,-0.046669066,-0.0904561505,-0.1804322302,-0.0006052702,0.0840085372,-0.0884604082,0.0542770997,0.1276235878,-0.2729560435,-0.0776810646,0.2229896039,0.3230027854,0.4300011098,-0.1942081898,-0.2773551941,-0.428699702,0.35606125,0.0843128413,0.3598292768,0.0427450761,0.1024750918,-0.1521518826,0.4299740791,-0.3599181473,-0.2498050481,-0.062408559,0.1829363257,0.0389307253,0.126891017,-0.1238600463,-0.1420397609,0.0421582423,0.1408634037,-0.072979413,-0.1976670921,0.0948442519,0.1418450475,0.1906209886,0.100148201,-0.2256940454,-0.3016279042,-0.0575749017,-0.3511753082,0.103514269,-0.0092987167,-0.3125889897,-0.1604615301,0.375433892,0.0410337262,-0.1293419003,0.3695023656,0.0883188024,0.0355710089,-0.0848018304,0.1509502232,-0.4421796799,-0.0200401004,-0.0005445252,0.3523772061,0.1863220781,-0.0164087024,0.2933133245,0.399145782,-0.2534354925,0.0088061811,0.2165987343,0.1170322746,0.0714503527,0.2712439895,0.2000564039,0.2315378338,-0.3683590591,-0.3234624267,-0.1228905171,-0.2538639903,0.2898574769,0.1271502227,0.1676930338,0.0901787505,-0.1946223527,0.1269264668,0.0460952707,-0.0471631922,0.1355329901,0.078291513,0.1367095411,-0.0010414658,0.0903619602,0.2777533531,0.1658720374,0.0078186048,-0.0568729602,0.4941709936,0.3845891953,-0.0948342234,0.2010027766,-0.0306095295,0.2637595236,0.2859853804,0.2809197605,0.1700624973,0.2751068771,0.1701160818,-0.0839604735,-0.3715267777,0.2919419706,0.1412462741,0.0187650528,-0.028316021,-0.1185658947,-0.245038256,-0.1537515372,-0.142740801,0.1537356824,-0.1843530089,0.5169635415,0.0133686503,0.3105067313,-0.1622338593,0.3076369464,0.3823401332,0.1365792453,-0.3617655337,-0.0999965295,0.1555233151,-0.2377637476,0.1416983455,0.1683663577,0.4571424425,0.4111610651,0.0801919326,0.1701584011,-0.1596920788,-0.0086721489,-0.1871578544,-0.0559565276,0.2579289675,-0.0538845286,0.143966943,0.1886513978,-0.0323325396,0.2839425504,0.0560536124,-0.0276472215,-0.3535073102,-0.0604583882,0.5578275919,-0.5780310631,0.3011538982,-0.0372194797,-0.1526598632,-0.0206112005,0.4134000242,-0.1001086831,0.0743933097,0.0755791143,0.0957457051,0.2858551741,0.3127804697,0.3436603248,0.1175781786,-0.3295322061,-0.1820982397,-0.2553746402,-0.1938426942,0.0749252588,-0.1611101478,0.0014268872,0.0788863152,0.3104854822,0.1100076884,-0.0542918779,-0.240780741,-0.0073100319,0.1459174305,-0.2963370681,-0.0806993544,0.2281942815,-0.066347681,0.0275355168,-0.2774797082,0.0943286866,0.1722065359,0.0837301761,-0.3141050041,0.0166981015,0.1884007007,0.1818265468,0.4190168083,0.1478976607,0.2240951359,-0.2689065933,0.072252132,0.160931617,-0.203279838,-0.3194253147,0.0030532982,0.0187073927,-0.0467037633,-0.5168744922,-0.1386774033,-0.1059471592,0.0141491182,-0.169720754,-0.0443772338,-0.2834773362,0.0758285746,0.0713371038,-0.074320443,-0.0535411537,0.2164406925,0.162305668,0.0218602549,-0.0475969762,-0.213684544,0.0684086308,0.1741075814,-0.0640859902,-0.0976745412,-0.0648060367,-0.1147760376,0.3184703588,-0.7344168425,-0.3051047623,0.2667624652,-0.1206688359,0.0448516533,0.4788630307,0.0444674529,0.0331965424,-0.3538219035,0.2330826968,-0.1090683192,-0.3129502535,-0.1438241452,-0.5232966542]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1877","title":"Allow concatenation of both in-memory and on-disk datasets","comments":"> Can you explain where the issue of the double memory may arise?\r\n\r\nWe have to keep each block (in-memory vs memory mapped) separated in order to be able to reload them with pickle.\r\nOn the other hand we also need to have the full table from mixed in-memory and memory mapped data in order to iterate or extract data conveniently. That means that each block is accessible twice: once in the full table, and once in the separated blocks. But since pyarrow tables concatenation doesn't double the memory, then building the full table doesn't cost memory which is what we want :)\r\n\r\n> Also, why is the existing concatenate_datasets not sufficient for this purpose?\r\n\r\nThe existing `concatenate_datasets` doesn't support having both in-memory and memory mapped data together (there's no fancy block separation logic). It works for datasets fully in-memory or fully memory mapped but not a mix of the two.\r\n\r\n> Will the add_item feature also help with lazy writing (or no caching) during map\/filter?\r\n\r\nIt will enable the implementation of the fast, masked filter from this discussion: https:\/\/github.com\/huggingface\/datasets\/issues\/1949\r\nHowever I don't think this will affect map.","body":"This is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova ","comment_length":188,"text":"Allow concatenation of both in-memory and on-disk datasets\nThis is a prerequisite for the addition of the `add_item` feature (see #1870).\r\nCurrently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).\r\nThis assumption is used for pickling for example:\r\n- in-memory dataset can just be pickled\/unpickled in-memory\r\n- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling\r\n\r\nMaybe let's have a design that allows a Dataset to have a Table that can be rebuilt from heterogenous sources like in-memory tables or on-disk tables ? This could also be further extended in the future\r\n\r\nOne idea would be to define a list of sources and each source implements a way to reload its corresponding pyarrow Table.\r\nThen the dataset would be the concatenation of all these tables.\r\n\r\nDepending on the source type, the serialization using pickle would be different. In-memory data would be copied while on-disk data would simply be replaced by the path to these data.\r\n\r\nIf you have some ideas you would like to share about the design\/API feel free to do so :)\r\n\r\ncc @albertvillanova \n> Can you explain where the issue of the double memory may arise?\r\n\r\nWe have to keep each block (in-memory vs memory mapped) separated in order to be able to reload them with pickle.\r\nOn the other hand we also need to have the full table from mixed in-memory and memory mapped data in order to iterate or extract data conveniently. That means that each block is accessible twice: once in the full table, and once in the separated blocks. But since pyarrow tables concatenation doesn't double the memory, then building the full table doesn't cost memory which is what we want :)\r\n\r\n> Also, why is the existing concatenate_datasets not sufficient for this purpose?\r\n\r\nThe existing `concatenate_datasets` doesn't support having both in-memory and memory mapped data together (there's no fancy block separation logic). It works for datasets fully in-memory or fully memory mapped but not a mix of the two.\r\n\r\n> Will the add_item feature also help with lazy writing (or no caching) during map\/filter?\r\n\r\nIt will enable the implementation of the fast, masked filter from this discussion: https:\/\/github.com\/huggingface\/datasets\/issues\/1949\r\nHowever I don't think this will affect map.","embeddings":[-0.4205480516,0.1177791432,-0.0135113075,0.2829381526,0.0721307024,0.0939797312,-0.1494367421,0.2320156842,-0.1620398611,0.1522656828,0.0206330512,0.510204792,-0.0248010047,0.3577524722,0.153740406,-0.0333436169,0.1533055007,0.2407618612,-0.432516396,0.2023283541,-0.2718376219,-0.2322957516,-0.1211226955,-0.4407121241,-0.1328163743,0.1058929712,-0.3701851666,0.085054554,-0.4612610042,-0.417299509,0.0272538252,0.2392499447,0.1475351006,0.0204422213,-0.0001143608,-0.0304890182,-0.0908041373,-0.1432089806,-0.4931333959,-0.144099012,-0.302980274,-0.4612712562,-0.001470246,-0.0359222107,0.2839404941,-0.3497850001,0.0704570338,0.0037824237,-0.0004687344,-0.0690577403,0.1282015294,0.0832635909,0.2490193993,-0.1535838544,0.2830585241,0.4180071056,-0.2433202416,0.0545799695,0.4837106466,-0.2297801673,0.1604688168,-0.185549736,-0.1114508659,-0.1264527738,0.3167591691,0.2640161514,-0.2564167678,-0.1504638195,-0.062841706,0.2610198557,0.4842351377,-0.460285604,-0.3712317944,-0.3262190223,0.1695735902,-0.2965211868,0.1417034566,0.2897109389,-0.1246990263,0.181557402,-0.0053669689,-0.3174619675,-0.1614489257,0.2725508809,0.295029074,0.268396467,0.3663906455,0.1452648193,0.2541914582,0.0573609583,0.2632886469,-0.1695788652,-0.2025769353,0.1410523355,-0.2625716031,-0.0054378361,0.0481507108,-0.1600821912,0.5918854475,0.0995221511,0.2872602642,0.022782702,-0.2712804079,0.3238212466,0.1234028339,0.1842763871,-0.2756664753,0.1439854503,-0.058196716,-0.1266408861,0.2939161956,0.0151917348,-0.1816811413,-0.0323906243,0.1392959952,-0.1682103276,-0.1076435372,0.202600196,0.0467827283,-0.2848707438,-0.1391761601,-0.1701781899,0.0513958298,0.0725707412,0.1154181659,0.0110624637,0.0228989609,0.3052885532,0.3008713424,-0.1711031049,0.0266159531,0.0167589281,-0.3064671159,0.3657286167,0.3710740209,0.0442401133,-0.1348829567,0.271768719,-0.0841675028,0.3043233156,-0.0053500663,-0.1916876435,0.3680019379,0.1365817338,0.0408077426,-0.4994348586,-0.0858418569,-0.0360954702,-0.2986930907,0.6121868491,-0.0858688578,-0.3561389148,-0.4303496182,0.0681181476,-0.0598043427,-0.0137253208,-0.263925761,0.5533694029,0.3309740424,-0.2825146019,0.040916305,0.1475729644,-0.1118465886,-0.5068644881,0.2198455483,0.2586522698,-0.5528047085,0.0585637093,0.12899445,-0.0345776305,0.1111791581,0.2120650113,-0.397354871,0.0458505228,-0.3841030002,0.2939403951,0.1699150652,0.0413726941,-0.2576788664,-0.1045229286,-0.0269715153,0.0448917523,0.4616477787,0.4100912511,0.0950642824,0.16308707,0.3194943368,0.3140036464,-0.2028822452,-0.1930931658,-0.091433093,-0.5326014161,0.5056709647,-0.0021548409,-0.1473309547,0.0495729074,0.2052704245,-0.5217081308,0.3012337089,-0.2306774706,0.2859789133,-0.1803401411,0.3845082223,-0.088817656,-0.1506255567,-0.1423925906,-0.5311309695,0.2147483081,-0.291357547,0.1576973051,-0.2815657854,-0.2635592818,0.2890248895,0.0733938068,-0.000241285,0.0676864684,0.0810715258,0.0158447046,0.0758395642,-0.3457562029,-0.2225687504,0.1293768734,0.1590712517,-0.0178132951,-0.151096493,0.3943977654,0.0196057074,-0.1607129574,-0.0763068572,0.195953995,-0.0492699035,0.1961109638,0.1199967042,0.427000314,-0.1109359264,0.0082428129,0.0712700784,0.272744596,0.1453538686,0.293161422,0.0296722818,-0.5536985397,0.1499249488,-0.1731536537,-0.0905447677,0.2890941799,0.1202573702,0.1383823007,0.1417947263,-0.1665379405,0.1052098721,-0.0745431334,0.0005404817,-0.2177435458,-0.0516589284,0.246276319,0.1018992364,0.3133161068,-0.4512616396,0.2090350091,0.2587777078,0.0302924812,0.1629493982,0.1270131767,-0.1180138215,-0.270550102,0.0862549841,0.148076877,0.6196273565,0.3672582805,-0.0011291382,-0.0006010781,-0.0432569049,-0.0995766819,0.2032078654,-0.049649857,0.2204270959,0.3111623228,0.4858122766,0.0462777354,-0.0847223774,0.0187398307,0.2424550354,-0.0236544032,-0.0614503361,-0.2882328629,-0.4230681658,-0.0768891722,-0.2263135463,-0.207323119,-0.1776858419,-0.0387442037,0.2044095099,0.3084001839,-0.1329572052,0.1332468241,0.1385305077,0.4870517552,-0.3491360545,-0.3730885386,-0.107596986,-0.0817747265,0.0371586047,0.0833978504,0.4131780565,-0.1326893866,0.4439466,0.4218161404,-0.1286580265,-0.3170163631,-0.3468725979,0.1252694428,-0.039119333,0.0798845813,0.0871986449,-0.0926862359,0.1265868396,-0.4788066447,0.1119153574,0.3859839737,-0.2844462097,0.2566552758,0.0273776781,-0.1551009566,-0.3382604718,-0.2619175613,-0.3027262986,-0.4580030143,0.4380853772,-0.1251059324,0.1152242571,0.0953672007,0.2327349186,-0.2977924347,0.0700946674,0.1275105923,-0.1762500554,-0.1304361075,0.3235555589,-0.135790199,-0.0962289423,-0.1681300104,-0.1917956173,0.1267280579,0.4574133754,-0.2145803869,-0.0556686707,-0.1127313972,0.4080162644,-0.1579324305,0.0697523355,0.6263289452,0.5570223927,-0.0811090991,-0.0401119292,-0.1109493598,0.2213413715,0.2556411624,0.1116929799,0.118866533,-0.1989126951,0.0869155452,0.3723876476,-0.0004664562,0.0835760385,0.4383431077,-0.0265469048,0.297645092,-0.2404049039,-0.0903695971,-0.2160644233,-0.4245319366,-0.031698361,-0.05708462,-0.0455007181,-0.2517730892,-0.0882366225,0.2317118049,-0.1693067253,-0.2507647574,0.346331507,-0.1190845147,0.2554132342,0.0230748188,-0.4499953389,-0.2813352942,0.0694618672,-0.0932985693,0.0908188298,0.1557243615,-0.152404353,-0.3255567253,-0.3939990997,-0.3772607744,0.1050971448,0.3018627167,0.1467111409,0.1842963398,-0.3467965722,-0.0701855049,-0.0196163412,0.8354034424,0.0623193532,-0.5227494836,0.0703704879,0.131728068,0.2117169946,0.0114878817,0.1598885953,0.0968080387,-0.2083498389,0.1697030067,-0.4544526041,0.1512103081,0.0658319965,0.4569223523,0.0042451979,-0.3004601598,-0.0685631335,0.0551145189,-0.4431818426,-0.4144229591,0.0910973102,0.1811735183,0.0586873032,-0.0761776865,-0.1391134709,-0.007849399,0.0387607515,-0.0608312152,0.3464483917,0.0200289693,0.4484291375,-0.2765004039,-0.0650269762,0.8608862162,0.6849475503,-0.2380896211,-0.1312052161,-0.0917247385,-0.1726324409,0.3094537258,0.0511384942,0.0947381929,0.4905812144,0.0457090475,0.0370524079,-0.4452128708,-0.1305461973,-0.0909067243,-0.2205581963,-0.3609577417,-0.3468234241,0.3938198984,0.1752981544,-0.1744603068,0.4534808397,0.0231146794,-0.4244272411,0.2573759854,0.0153103629,1.1907246113,0.1039270908,0.3567136228,0.1535402685,-0.227843076,0.0615459792,-0.0887165144,0.2979052067,-0.1494070143,-0.0145541243,-0.0680171475,-0.2100125253,-0.0067493892,0.0345537402,-0.4873152375,-0.1321270764,-0.0438016877,-0.0675915182,-0.0297099184,0.1946683377,-0.2373500615,-0.3399893641,0.1423704177,0.020313818,0.0962796062,-0.4308600426,-0.0875311717,-0.1844056845,-0.1948050112,0.0438084416,-0.297635287,-0.1206348836,-0.3254969418,0.3182812035,-0.028027121,-0.1441754103,-0.0765076876,0.0056239869,-0.164966926,-0.0034902371,-0.1046124175,-0.1205439195,0.1330459118,-0.0177097246,-0.1660441905,-0.4333419502,0.4511227012,0.0919252336,-0.2152074575,-0.0488369837,-0.1193017736,-0.3479339182,-0.1846932918,0.011517032,0.1461817324,-0.2090687007,-0.2435186505,0.4039715528,0.1063711122,-0.2542291582,0.0761181638,0.2311345339,-0.053938169,0.1534841061,-0.1724506766,-0.0014171106,-0.2240436971,0.0823869631,0.3592762947,0.1005133316,0.1333336085,0.0378350504,-0.0127528142,-0.0389260836,0.0005595041,0.183571443,-0.1900925338,-0.0854351595,0.1291294694,-0.0664899722,-0.1210852191,0.1926624328,0.3050109446,0.3547001183,-0.0487308986,-0.1518024206,-0.4447176754,0.4528210759,0.1614197344,0.3631177545,0.0886069164,0.3154402077,-0.0791257173,0.3983821571,-0.2913195193,-0.121158205,-0.0197574869,0.1517448127,-0.0519757569,0.1548065543,-0.0444980599,-0.1612505168,0.0085243015,0.0477791689,-0.0568659566,-0.1307118237,-0.0081736306,0.1341316253,0.3104193211,0.1546635181,-0.258486867,-0.4188140333,0.0273020286,-0.2525815666,0.0860088095,0.0694740936,-0.2653083801,0.0038917188,0.4690990746,0.046466466,-0.1550290436,0.3445216417,0.0580561459,0.132303834,0.0329741873,0.223534897,-0.365991354,-0.009011209,-0.1355912387,0.2857983708,0.2216772139,-0.0888825655,0.1399485767,0.2560234964,-0.2851628065,-0.2245733589,0.1921325773,0.2152834684,0.0877314135,0.2562374473,0.2321071029,0.1659255028,-0.3981326818,-0.1963470131,-0.2837144732,-0.252707541,0.2140795439,0.2647175789,0.1927173585,0.132044062,-0.2951056957,0.0354305804,0.1063152403,-0.0030813348,-0.0055838688,0.0424556173,-0.0371844694,-0.0206310935,-0.0071206228,0.2337068915,0.1869412959,0.0185629949,-0.1811137646,0.4842152297,0.2601227164,-0.1952259243,0.1806902736,-0.0248801783,0.2080717832,0.4031640589,0.2998482287,-0.0317456983,0.1795702428,0.4649269581,0.0617630184,-0.3081249595,0.3061023355,0.1629644334,0.0695119947,0.0276513472,-0.0616922677,-0.2765995562,-0.2785306871,-0.1142990142,0.1149340346,-0.30453825,0.5424286723,-0.0381481312,0.1117517948,-0.219407931,0.3140153289,0.5822793841,0.0182736497,-0.3843547404,-0.0587344952,0.3683798611,-0.3806055486,0.0478154048,0.234635964,0.5222939253,0.3223424256,0.1351946145,0.0562999584,-0.0919187367,0.0474996455,-0.207973972,-0.0770851821,0.2428851426,0.0241848975,0.144309029,0.1389949024,-0.0102108112,0.2100463212,0.1940659583,0.0517657474,-0.4397470355,-0.1522545218,0.3874554038,-0.5414434671,0.4035972655,0.0093684737,-0.2633351386,0.0363720767,0.3410353661,-0.0467891507,0.0430983342,0.0376731567,0.0665062666,0.3350123167,0.2623302042,0.2968886793,0.1776204556,-0.373018384,-0.2074378878,-0.3266669512,-0.1075489148,-0.0852779597,-0.2132118642,-0.0943394899,0.1443772614,0.2860146165,0.1380279064,0.0169367623,-0.0303850006,-0.0389752686,0.255659014,-0.2881510258,-0.1359728277,0.3313158751,0.0026478597,0.0365236178,-0.275409162,0.1030916497,0.1382570714,-0.0082948999,-0.3732219636,0.0620639957,0.311485827,0.3021305501,0.6179407239,0.1816809475,0.2282115519,-0.2298468053,-0.0382055864,0.0162861254,-0.208839193,-0.2918352187,0.0776328444,0.0476223715,-0.0637338907,-0.4909763038,-0.134386465,-0.0859147012,0.0522648506,-0.1825255007,0.1082524434,-0.2580526769,0.0688755587,0.1717408895,0.0054738205,-0.0727895498,0.1789417118,0.1221984923,0.0560632199,-0.11989858,-0.2423341274,0.0153738428,0.2471015006,-0.0438111052,-0.0469027683,-0.1139092147,-0.0990875736,0.2313878536,-0.8545036912,-0.2810121477,0.2284170687,0.0029589513,-0.004220366,0.444512248,0.0436203182,-0.0501921475,-0.2877586782,0.3475422859,-0.1791473776,-0.4259859622,-0.1129206493,-0.4514321089]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1876","title":" load_dataset(\"multi_woz_v22\") NonMatchingChecksumError","comments":"Thanks for reporting !\r\nThis is due to the changes made in the data files in the multiwoz repo: https:\/\/github.com\/budzianowski\/multiwoz\/pull\/59\r\nI'm opening a PR to update the checksums of the data files.","body":"Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n\r\n`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`\r\n\r\n\r\nThis will give the following error:\r\n\r\n```\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_003.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_004.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_005.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_006.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_007.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_008.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_009.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_010.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_012.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_013.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_014.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_015.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_016.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_017.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_002.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_002.json']\r\n```\r\n","comment_length":32,"text":" load_dataset(\"multi_woz_v22\") NonMatchingChecksumError\nHi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n\r\n`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`\r\n\r\n\r\nThis will give the following error:\r\n\r\n```\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_003.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_004.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_005.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_006.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_007.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_008.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_009.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_010.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_012.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_013.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_014.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_015.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_016.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_017.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_002.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_002.json']\r\n```\r\n\nThanks for reporting !\r\nThis is due to the changes made in the data files in the multiwoz repo: https:\/\/github.com\/budzianowski\/multiwoz\/pull\/59\r\nI'm opening a PR to update the checksums of the data files.","embeddings":[-0.2244368494,0.1764702946,-0.0287194345,0.1495722383,0.1895723343,0.0018984437,0.3650578558,0.4820295274,0.2514897287,0.1666063964,-0.0970349461,0.1781593114,-0.0822444186,0.0992015228,-0.1845492125,0.2732406855,0.0765193701,-0.0256412663,-0.1682780683,-0.0625795946,-0.2430096716,0.3292944729,-0.1593268216,-0.0338407457,-0.3486862779,0.130385235,0.173908785,0.336238414,-0.0838693827,-0.1523166895,0.4145028293,0.2369774878,0.0439615399,0.3686157465,-0.0001172875,0.163338542,0.2778438032,-0.1433111131,-0.1339644343,-0.1817747802,-0.5618035793,-0.3552265167,-0.1035682037,-0.1154623255,0.0706329644,0.3033826053,0.0572008304,-0.0867421255,0.2431910932,0.1712264121,0.1535670757,0.3319171369,0.2157697976,0.096905008,0.0455294512,-0.0774404854,-0.1376802474,0.0922583491,0.2744395733,-0.1569582969,-0.2395652235,0.2651384175,-0.1804820746,0.2693834305,-0.08042676,-0.151098907,0.1479792297,-0.1058241054,0.2105871737,0.5727722049,0.2205697149,-0.0554821864,-0.1762661785,-0.031950675,-0.2201738805,0.0833650157,0.3634192646,0.1834743917,-0.2078651488,0.0045875246,-0.4487929642,0.1443090439,-0.0263798088,0.2792589366,0.2121898234,0.3145956695,0.2003566772,0.0726620853,0.2279323786,-0.2064414769,0.1992104501,-0.3010312915,-0.1609574109,0.1293215007,-0.6801476479,-0.0664964691,0.0034521031,0.3319556117,0.4404266775,0.2289481759,0.4523443282,0.3736301064,-0.1520804018,0.1170620769,0.3968931139,0.1744454056,0.0657053292,-0.0545444787,0.1692391336,0.2095874697,-0.0953853428,0.23737818,-0.0305045638,-0.3612305224,0.349316746,0.1929925084,0.2069615573,-0.3434550762,-0.1813292205,0.0824042559,-0.2454786599,-0.1945316195,0.2718029022,0.2862129211,-0.0944880024,0.5524677038,-0.1653458029,0.2156011462,-0.1538057327,-0.184701696,-0.2573401332,-0.1516558081,0.0730739161,0.1531955004,0.2943665981,0.0656549037,0.3595133722,-0.0321129337,0.3835370243,-0.2031213194,0.333525002,0.0902094916,-0.069919236,0.2554130256,0.0134760654,0.2010277361,0.2036009133,-0.0260661189,-0.0622884482,0.2220196426,-0.332572788,-0.6167247295,0.1093568653,0.1607629508,-0.3595766425,-0.0804914013,0.0163278151,-0.2801302671,0.3050760031,-0.4042750597,-0.1545268744,-0.3094389141,-0.0430206098,-0.1179552451,0.0098633282,0.0942928717,-0.0934405178,0.0649290308,-0.0455776192,-0.1832666099,0.3468402326,0.2667627037,-0.2780414522,-0.1645423472,-0.1394770145,0.0187303592,0.2207069397,-0.0911147445,-0.3638109565,0.2024427652,-0.071666114,0.6232899427,0.1856193095,0.1097929999,-0.347364068,-0.1108146757,0.2366856933,0.0823844671,0.1350486577,0.1764032245,-0.2450098842,-0.3122074902,0.3476401865,0.1528952271,0.0161080807,-0.2277660966,0.0021591219,0.0893484354,0.5888454914,-0.229704082,-0.0971514285,0.173809126,0.0959679857,-0.2113368064,0.0982792377,0.0511620082,-0.4866259396,0.3225809634,-0.1565800011,-0.0987997502,0.3440404832,-0.0598260574,-0.0504282936,-0.1761434972,-0.3390328586,0.0107523073,0.0809487924,0.2748901248,0.083496578,-0.2535201013,-0.0927836373,0.3571139276,-0.3864992857,-0.0857931376,-0.4520960152,0.1969179064,-0.0879131109,0.0851735622,0.0764899477,0.1571474075,0.1050598249,-0.1094213575,-0.1870695502,0.5129078031,0.321377933,0.238428399,-0.0831710547,0.3510250747,-0.1519880891,-0.0498669744,-0.1007987484,0.5829588175,-0.043825414,-0.2062136233,-0.1941188127,0.4879204333,-0.1729985923,0.4535035193,-0.0467061065,0.1363264918,0.3366037309,-0.1314040124,-0.267632544,-0.094632037,0.4545117021,-0.1980919838,-0.166961506,0.2530014515,-0.2133246213,0.0163294896,0.3733881712,-0.1440781504,0.1259907186,-0.0151448352,0.2458733171,-0.2552571595,-0.0198079348,0.492271781,0.3632516265,0.1011830643,-0.017946247,-0.0273172073,-0.2294692695,-0.1744433939,0.1194066629,0.0973706469,0.3296886981,0.5370066166,0.0413935557,-0.3653722107,-0.4267033637,-0.0148492791,0.0290301628,0.1656199098,-0.441488415,-0.2537639439,-0.4441935718,0.2288448811,-0.3637366891,-0.3436980546,-0.4129472375,-0.4648984671,0.1210602671,0.2281651497,-0.019953087,0.3193200827,-0.3111640215,0.0683446303,-0.13192828,0.0605431758,0.0400538072,-0.2777514756,-0.0817120597,-0.0046348502,0.395745039,0.0018831765,-0.0578058697,-0.3077501059,-0.0692631528,-0.1458399743,-0.1667571664,-0.0308500472,-0.1108122692,-0.0727645904,0.258897841,0.3025155365,0.1404629946,-0.3886933029,0.2458182424,-0.0367405601,-0.2834911942,0.1877612323,-0.1721364707,-0.0189697389,0.1081124917,-0.3866266012,-0.0947575793,-0.3375903666,-0.0101160659,0.0664086565,0.2717181146,0.2532503307,0.0835985392,-0.067855671,0.135645777,0.2740627229,-0.3200554848,-0.8128382564,0.4206303358,-0.0225661304,-0.2601571679,-0.0509394854,0.0214568581,0.2018916756,0.0964379087,-0.4472939074,-0.2804222107,-0.0651813745,-0.3137300909,0.1682505161,-0.3037391305,0.3359626532,0.118827641,-0.0578323677,-0.244994387,-0.294922471,0.3670241833,0.0858874843,0.5151624084,-0.2128699422,0.2413334697,0.0072190906,0.59493047,0.2273386568,0.2582996488,0.0617827326,0.1863201261,0.1531594396,-0.0817795843,-0.3245143592,0.1173866391,-0.2651423812,0.004429182,0.1789305359,0.0948021784,-0.273284018,-0.1016846523,0.1297237724,-0.1512695104,-0.2311442494,0.0542182997,-0.0111719985,0.0999288857,0.2145664692,0.0766038224,-0.3172064722,-0.3501291573,0.1731933951,0.2011538148,-0.1653898358,-0.1424287409,-0.5950775743,0.0641577691,0.024697274,0.4123165309,0.2499673218,0.4031144381,0.0229396354,-0.3206753731,0.0325473547,-0.2579710484,0.5845400095,-0.3834357262,0.3248415589,0.0977153778,-0.0471826047,-0.2630657256,-0.1649410576,-0.0380526185,-0.0256651193,0.3266568482,0.3074330389,-0.2269275337,-0.2858472764,0.1157697365,0.337846756,-0.2878535688,-0.1682776213,-0.3097966015,-0.1108990908,-0.3571147323,-0.2201311439,-0.3995670676,0.3472657502,-0.1648672968,-0.0812498257,0.1711757928,-0.1453643888,0.2767299116,0.03994295,0.3021826744,0.3644610047,0.225110203,0.0248756167,0.338565141,0.2234014571,0.4796167314,-0.1350188404,-0.0993258879,-0.0563185886,-0.2143767029,0.0410142802,0.2235033959,0.0236763805,-0.0814053416,0.2965885103,-0.1108627766,-0.0433097072,0.0397988297,0.2845616937,0.0299271457,-0.3399115801,-0.301807642,0.0595081486,0.0007752888,-0.178718999,0.2506121993,-0.1772620231,-0.0745188445,0.0401590206,0.3154736757,0.8823344707,0.0781968459,0.0889303312,0.0817513838,-0.1757192612,-0.1717552245,-0.4199728668,0.1754765809,-0.2713772655,-0.2802045345,-0.0789559931,-0.2028966546,0.1161567271,0.3723813295,-0.0140839126,0.3294216692,0.0507585853,0.1718657315,0.015210106,0.0928249508,-0.1202962399,-0.1144701317,0.0429769382,0.1203733087,-0.1465322077,0.1447477788,-0.1230822057,-0.0494485982,0.2085236609,-0.2759677768,-0.1316172779,0.1984799504,-0.1262031645,0.4007564783,0.2463366538,-0.2069448829,-0.290071398,0.4346532226,0.0933009014,-0.0674183592,-0.2564155459,0.1446241438,0.3509513736,0.0936215445,-0.0919712707,-0.0624782331,0.3511782587,-0.1872355193,-0.2813732922,0.0104376664,-0.1517420858,-0.6179360747,-0.0681770965,-0.1264027506,0.0984783098,-0.154596284,-0.1535268128,0.1786635965,0.0546398424,-0.3420567513,0.1221394762,0.1330333799,-0.3626447618,0.3028178513,0.1293569803,-0.2770988643,-0.0944160596,0.554800272,0.0386738665,-0.001004266,0.7191022038,0.0710801035,-0.1204014197,-0.2845618129,0.3099564314,0.2341080755,-0.2122221738,-0.0682358071,-0.1181255877,-0.0549927317,0.040189594,0.5675612092,0.2766489387,0.3637252152,-0.0064766975,-0.6050378084,-0.122251153,0.0505504832,0.2656698227,0.3840138316,-0.3860800564,-0.0716586486,-0.120571062,0.2204845697,-0.2977004051,0.0136333136,-0.3014138937,-0.0014124378,-0.3319445252,-0.0595821477,0.1033880115,0.0466479473,0.1038635299,-0.2260658592,-0.2882824838,-0.1636383384,-0.2158522606,0.1587166041,0.0363478102,-0.0632579476,0.0698035806,-0.0939673111,-0.1066646129,-0.1600662321,-0.0529938377,0.2080393583,-0.0953483358,0.125788644,0.2332840264,-0.1949185878,-0.1487827599,-0.261736393,-0.2963095307,-0.0960339904,0.0908312425,-0.1021253839,0.0535941757,-0.169658795,0.4369630814,0.1701050401,0.024545135,-0.0090559432,0.2811068594,-0.2891628742,-0.0970124528,0.2284561545,0.3137277067,0.3578693867,-0.2104902118,0.1913179606,0.000825197,0.2238859683,-0.2928492129,-0.1559717953,0.3667841554,-0.112051852,0.0143712647,0.0317219533,-0.0716917664,-0.3091430366,0.5745686293,0.1000950933,-0.0016480027,-0.3915390968,0.1874298453,0.3277118504,-0.3724373281,-0.0696389079,-0.004597913,-0.2296410054,0.2119155675,-0.2145881206,-0.1123061702,0.4519319832,0.0501495712,0.250569135,-0.1284569353,-0.4928975999,0.0944712386,0.1954616159,-0.0912432149,0.0177286398,0.0348484144,0.4064955413,-0.2336776704,0.182787776,-0.702688992,-0.0180880725,0.0231124628,-0.2515980303,-0.3489803374,-0.3787186146,-0.1243415698,0.1621755213,-0.0076932451,0.1223282889,0.3172167838,0.0278804507,-0.0302277654,-0.5480276346,-0.1554399431,-0.072140418,-0.1325324923,-0.2546501458,0.2110592127,0.3817438781,-0.1168670058,0.3411186635,0.232888937,0.3793445826,0.3968301713,0.0158854928,-0.0075071477,-0.0547184944,0.0275844373,-0.151116848,0.7324585915,0.1668840051,-0.1373345107,0.4767388999,0.0521655828,-0.1582669616,0.0700558797,-0.0142776398,-0.0927828625,-0.0598390847,0.7789545655,0.0388093628,0.1453841925,-0.0719826072,0.2336157411,-0.3082112372,-0.3224420547,0.0849673524,0.0910015479,0.2019011974,-0.0258247387,0.0734821185,-0.1888687611,0.6922151446,-0.0983080342,-0.2206100374,-0.2996817231,-0.2596409619,-0.8195856214,0.2501328588,-0.1412824243,0.3424734771,-0.0689983666,0.1562742442,0.1727872491,0.2854847014,0.2575885057,-0.2129989117,-0.1863705069,-0.0521587506,-0.3973252773,0.0739042535,0.4234284163,0.0349841602,-0.0069678947,-0.2355769873,0.1889984906,0.2631671429,0.0424308777,0.1342669725,0.0111747002,-0.0327080488,-0.0246862136,-0.0812196285,0.1433468908,0.5623226762,-0.0031299428,-0.011459128,-0.1989398003,-0.2911090851,-0.0620972998,0.1848046333,-0.050203722,0.5143117309,-0.1861067712,-0.0308319367,-0.2220528424,0.2004486471,-0.0561484024,-0.2992598414,-0.1287595034,0.0325104445,-0.0127721233,0.084331587,0.1155453026,0.0505930148,0.0815804675,0.2093015313,-0.0561158136,-0.0594237521,0.7326385975,-0.1133245677,-0.1869971454,-0.3649466038,0.1383291632,-0.3404731452,-0.0648737326,-0.4481017292,0.0900497064,0.3190638721,0.1263615787,-0.2260794342,0.1400440037,-0.2574895024,0.0272709914,-0.0578922816,0.0775431693,0.2296063006,-0.2987921536,-0.1100094467,-0.1358176172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1876","title":" load_dataset(\"multi_woz_v22\") NonMatchingChecksumError","comments":"I just merged the fix. It will be available in the new release of `datasets` later today.\r\nYou'll be able to get the new version with\r\n```\r\npip install --upgrade datasets\r\n```","body":"Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n\r\n`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`\r\n\r\n\r\nThis will give the following error:\r\n\r\n```\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_003.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_004.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_005.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_006.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_007.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_008.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_009.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_010.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_012.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_013.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_014.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_015.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_016.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_017.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_002.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_002.json']\r\n```\r\n","comment_length":32,"text":" load_dataset(\"multi_woz_v22\") NonMatchingChecksumError\nHi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n\r\n`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`\r\n\r\n\r\nThis will give the following error:\r\n\r\n```\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_003.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_004.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_005.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_006.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_007.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_008.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_009.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_010.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_012.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_013.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_014.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_015.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_016.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_017.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_002.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_002.json']\r\n```\r\n\nI just merged the fix. It will be available in the new release of `datasets` later today.\r\nYou'll be able to get the new version with\r\n```\r\npip install --upgrade datasets\r\n```","embeddings":[-0.2244368494,0.1764702946,-0.0287194345,0.1495722383,0.1895723343,0.0018984437,0.3650578558,0.4820295274,0.2514897287,0.1666063964,-0.0970349461,0.1781593114,-0.0822444186,0.0992015228,-0.1845492125,0.2732406855,0.0765193701,-0.0256412663,-0.1682780683,-0.0625795946,-0.2430096716,0.3292944729,-0.1593268216,-0.0338407457,-0.3486862779,0.130385235,0.173908785,0.336238414,-0.0838693827,-0.1523166895,0.4145028293,0.2369774878,0.0439615399,0.3686157465,-0.0001172875,0.163338542,0.2778438032,-0.1433111131,-0.1339644343,-0.1817747802,-0.5618035793,-0.3552265167,-0.1035682037,-0.1154623255,0.0706329644,0.3033826053,0.0572008304,-0.0867421255,0.2431910932,0.1712264121,0.1535670757,0.3319171369,0.2157697976,0.096905008,0.0455294512,-0.0774404854,-0.1376802474,0.0922583491,0.2744395733,-0.1569582969,-0.2395652235,0.2651384175,-0.1804820746,0.2693834305,-0.08042676,-0.151098907,0.1479792297,-0.1058241054,0.2105871737,0.5727722049,0.2205697149,-0.0554821864,-0.1762661785,-0.031950675,-0.2201738805,0.0833650157,0.3634192646,0.1834743917,-0.2078651488,0.0045875246,-0.4487929642,0.1443090439,-0.0263798088,0.2792589366,0.2121898234,0.3145956695,0.2003566772,0.0726620853,0.2279323786,-0.2064414769,0.1992104501,-0.3010312915,-0.1609574109,0.1293215007,-0.6801476479,-0.0664964691,0.0034521031,0.3319556117,0.4404266775,0.2289481759,0.4523443282,0.3736301064,-0.1520804018,0.1170620769,0.3968931139,0.1744454056,0.0657053292,-0.0545444787,0.1692391336,0.2095874697,-0.0953853428,0.23737818,-0.0305045638,-0.3612305224,0.349316746,0.1929925084,0.2069615573,-0.3434550762,-0.1813292205,0.0824042559,-0.2454786599,-0.1945316195,0.2718029022,0.2862129211,-0.0944880024,0.5524677038,-0.1653458029,0.2156011462,-0.1538057327,-0.184701696,-0.2573401332,-0.1516558081,0.0730739161,0.1531955004,0.2943665981,0.0656549037,0.3595133722,-0.0321129337,0.3835370243,-0.2031213194,0.333525002,0.0902094916,-0.069919236,0.2554130256,0.0134760654,0.2010277361,0.2036009133,-0.0260661189,-0.0622884482,0.2220196426,-0.332572788,-0.6167247295,0.1093568653,0.1607629508,-0.3595766425,-0.0804914013,0.0163278151,-0.2801302671,0.3050760031,-0.4042750597,-0.1545268744,-0.3094389141,-0.0430206098,-0.1179552451,0.0098633282,0.0942928717,-0.0934405178,0.0649290308,-0.0455776192,-0.1832666099,0.3468402326,0.2667627037,-0.2780414522,-0.1645423472,-0.1394770145,0.0187303592,0.2207069397,-0.0911147445,-0.3638109565,0.2024427652,-0.071666114,0.6232899427,0.1856193095,0.1097929999,-0.347364068,-0.1108146757,0.2366856933,0.0823844671,0.1350486577,0.1764032245,-0.2450098842,-0.3122074902,0.3476401865,0.1528952271,0.0161080807,-0.2277660966,0.0021591219,0.0893484354,0.5888454914,-0.229704082,-0.0971514285,0.173809126,0.0959679857,-0.2113368064,0.0982792377,0.0511620082,-0.4866259396,0.3225809634,-0.1565800011,-0.0987997502,0.3440404832,-0.0598260574,-0.0504282936,-0.1761434972,-0.3390328586,0.0107523073,0.0809487924,0.2748901248,0.083496578,-0.2535201013,-0.0927836373,0.3571139276,-0.3864992857,-0.0857931376,-0.4520960152,0.1969179064,-0.0879131109,0.0851735622,0.0764899477,0.1571474075,0.1050598249,-0.1094213575,-0.1870695502,0.5129078031,0.321377933,0.238428399,-0.0831710547,0.3510250747,-0.1519880891,-0.0498669744,-0.1007987484,0.5829588175,-0.043825414,-0.2062136233,-0.1941188127,0.4879204333,-0.1729985923,0.4535035193,-0.0467061065,0.1363264918,0.3366037309,-0.1314040124,-0.267632544,-0.094632037,0.4545117021,-0.1980919838,-0.166961506,0.2530014515,-0.2133246213,0.0163294896,0.3733881712,-0.1440781504,0.1259907186,-0.0151448352,0.2458733171,-0.2552571595,-0.0198079348,0.492271781,0.3632516265,0.1011830643,-0.017946247,-0.0273172073,-0.2294692695,-0.1744433939,0.1194066629,0.0973706469,0.3296886981,0.5370066166,0.0413935557,-0.3653722107,-0.4267033637,-0.0148492791,0.0290301628,0.1656199098,-0.441488415,-0.2537639439,-0.4441935718,0.2288448811,-0.3637366891,-0.3436980546,-0.4129472375,-0.4648984671,0.1210602671,0.2281651497,-0.019953087,0.3193200827,-0.3111640215,0.0683446303,-0.13192828,0.0605431758,0.0400538072,-0.2777514756,-0.0817120597,-0.0046348502,0.395745039,0.0018831765,-0.0578058697,-0.3077501059,-0.0692631528,-0.1458399743,-0.1667571664,-0.0308500472,-0.1108122692,-0.0727645904,0.258897841,0.3025155365,0.1404629946,-0.3886933029,0.2458182424,-0.0367405601,-0.2834911942,0.1877612323,-0.1721364707,-0.0189697389,0.1081124917,-0.3866266012,-0.0947575793,-0.3375903666,-0.0101160659,0.0664086565,0.2717181146,0.2532503307,0.0835985392,-0.067855671,0.135645777,0.2740627229,-0.3200554848,-0.8128382564,0.4206303358,-0.0225661304,-0.2601571679,-0.0509394854,0.0214568581,0.2018916756,0.0964379087,-0.4472939074,-0.2804222107,-0.0651813745,-0.3137300909,0.1682505161,-0.3037391305,0.3359626532,0.118827641,-0.0578323677,-0.244994387,-0.294922471,0.3670241833,0.0858874843,0.5151624084,-0.2128699422,0.2413334697,0.0072190906,0.59493047,0.2273386568,0.2582996488,0.0617827326,0.1863201261,0.1531594396,-0.0817795843,-0.3245143592,0.1173866391,-0.2651423812,0.004429182,0.1789305359,0.0948021784,-0.273284018,-0.1016846523,0.1297237724,-0.1512695104,-0.2311442494,0.0542182997,-0.0111719985,0.0999288857,0.2145664692,0.0766038224,-0.3172064722,-0.3501291573,0.1731933951,0.2011538148,-0.1653898358,-0.1424287409,-0.5950775743,0.0641577691,0.024697274,0.4123165309,0.2499673218,0.4031144381,0.0229396354,-0.3206753731,0.0325473547,-0.2579710484,0.5845400095,-0.3834357262,0.3248415589,0.0977153778,-0.0471826047,-0.2630657256,-0.1649410576,-0.0380526185,-0.0256651193,0.3266568482,0.3074330389,-0.2269275337,-0.2858472764,0.1157697365,0.337846756,-0.2878535688,-0.1682776213,-0.3097966015,-0.1108990908,-0.3571147323,-0.2201311439,-0.3995670676,0.3472657502,-0.1648672968,-0.0812498257,0.1711757928,-0.1453643888,0.2767299116,0.03994295,0.3021826744,0.3644610047,0.225110203,0.0248756167,0.338565141,0.2234014571,0.4796167314,-0.1350188404,-0.0993258879,-0.0563185886,-0.2143767029,0.0410142802,0.2235033959,0.0236763805,-0.0814053416,0.2965885103,-0.1108627766,-0.0433097072,0.0397988297,0.2845616937,0.0299271457,-0.3399115801,-0.301807642,0.0595081486,0.0007752888,-0.178718999,0.2506121993,-0.1772620231,-0.0745188445,0.0401590206,0.3154736757,0.8823344707,0.0781968459,0.0889303312,0.0817513838,-0.1757192612,-0.1717552245,-0.4199728668,0.1754765809,-0.2713772655,-0.2802045345,-0.0789559931,-0.2028966546,0.1161567271,0.3723813295,-0.0140839126,0.3294216692,0.0507585853,0.1718657315,0.015210106,0.0928249508,-0.1202962399,-0.1144701317,0.0429769382,0.1203733087,-0.1465322077,0.1447477788,-0.1230822057,-0.0494485982,0.2085236609,-0.2759677768,-0.1316172779,0.1984799504,-0.1262031645,0.4007564783,0.2463366538,-0.2069448829,-0.290071398,0.4346532226,0.0933009014,-0.0674183592,-0.2564155459,0.1446241438,0.3509513736,0.0936215445,-0.0919712707,-0.0624782331,0.3511782587,-0.1872355193,-0.2813732922,0.0104376664,-0.1517420858,-0.6179360747,-0.0681770965,-0.1264027506,0.0984783098,-0.154596284,-0.1535268128,0.1786635965,0.0546398424,-0.3420567513,0.1221394762,0.1330333799,-0.3626447618,0.3028178513,0.1293569803,-0.2770988643,-0.0944160596,0.554800272,0.0386738665,-0.001004266,0.7191022038,0.0710801035,-0.1204014197,-0.2845618129,0.3099564314,0.2341080755,-0.2122221738,-0.0682358071,-0.1181255877,-0.0549927317,0.040189594,0.5675612092,0.2766489387,0.3637252152,-0.0064766975,-0.6050378084,-0.122251153,0.0505504832,0.2656698227,0.3840138316,-0.3860800564,-0.0716586486,-0.120571062,0.2204845697,-0.2977004051,0.0136333136,-0.3014138937,-0.0014124378,-0.3319445252,-0.0595821477,0.1033880115,0.0466479473,0.1038635299,-0.2260658592,-0.2882824838,-0.1636383384,-0.2158522606,0.1587166041,0.0363478102,-0.0632579476,0.0698035806,-0.0939673111,-0.1066646129,-0.1600662321,-0.0529938377,0.2080393583,-0.0953483358,0.125788644,0.2332840264,-0.1949185878,-0.1487827599,-0.261736393,-0.2963095307,-0.0960339904,0.0908312425,-0.1021253839,0.0535941757,-0.169658795,0.4369630814,0.1701050401,0.024545135,-0.0090559432,0.2811068594,-0.2891628742,-0.0970124528,0.2284561545,0.3137277067,0.3578693867,-0.2104902118,0.1913179606,0.000825197,0.2238859683,-0.2928492129,-0.1559717953,0.3667841554,-0.112051852,0.0143712647,0.0317219533,-0.0716917664,-0.3091430366,0.5745686293,0.1000950933,-0.0016480027,-0.3915390968,0.1874298453,0.3277118504,-0.3724373281,-0.0696389079,-0.004597913,-0.2296410054,0.2119155675,-0.2145881206,-0.1123061702,0.4519319832,0.0501495712,0.250569135,-0.1284569353,-0.4928975999,0.0944712386,0.1954616159,-0.0912432149,0.0177286398,0.0348484144,0.4064955413,-0.2336776704,0.182787776,-0.702688992,-0.0180880725,0.0231124628,-0.2515980303,-0.3489803374,-0.3787186146,-0.1243415698,0.1621755213,-0.0076932451,0.1223282889,0.3172167838,0.0278804507,-0.0302277654,-0.5480276346,-0.1554399431,-0.072140418,-0.1325324923,-0.2546501458,0.2110592127,0.3817438781,-0.1168670058,0.3411186635,0.232888937,0.3793445826,0.3968301713,0.0158854928,-0.0075071477,-0.0547184944,0.0275844373,-0.151116848,0.7324585915,0.1668840051,-0.1373345107,0.4767388999,0.0521655828,-0.1582669616,0.0700558797,-0.0142776398,-0.0927828625,-0.0598390847,0.7789545655,0.0388093628,0.1453841925,-0.0719826072,0.2336157411,-0.3082112372,-0.3224420547,0.0849673524,0.0910015479,0.2019011974,-0.0258247387,0.0734821185,-0.1888687611,0.6922151446,-0.0983080342,-0.2206100374,-0.2996817231,-0.2596409619,-0.8195856214,0.2501328588,-0.1412824243,0.3424734771,-0.0689983666,0.1562742442,0.1727872491,0.2854847014,0.2575885057,-0.2129989117,-0.1863705069,-0.0521587506,-0.3973252773,0.0739042535,0.4234284163,0.0349841602,-0.0069678947,-0.2355769873,0.1889984906,0.2631671429,0.0424308777,0.1342669725,0.0111747002,-0.0327080488,-0.0246862136,-0.0812196285,0.1433468908,0.5623226762,-0.0031299428,-0.011459128,-0.1989398003,-0.2911090851,-0.0620972998,0.1848046333,-0.050203722,0.5143117309,-0.1861067712,-0.0308319367,-0.2220528424,0.2004486471,-0.0561484024,-0.2992598414,-0.1287595034,0.0325104445,-0.0127721233,0.084331587,0.1155453026,0.0505930148,0.0815804675,0.2093015313,-0.0561158136,-0.0594237521,0.7326385975,-0.1133245677,-0.1869971454,-0.3649466038,0.1383291632,-0.3404731452,-0.0648737326,-0.4481017292,0.0900497064,0.3190638721,0.1263615787,-0.2260794342,0.1400440037,-0.2574895024,0.0272709914,-0.0578922816,0.0775431693,0.2296063006,-0.2987921536,-0.1100094467,-0.1358176172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1876","title":" load_dataset(\"multi_woz_v22\") NonMatchingChecksumError","comments":"Hi, I still meet the error when loading the datasets after upgradeing datasets.\r\n\r\nraise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json']","body":"Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n\r\n`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`\r\n\r\n\r\nThis will give the following error:\r\n\r\n```\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_003.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_004.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_005.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_006.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_007.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_008.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_009.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_010.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_012.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_013.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_014.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_015.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_016.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_017.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_002.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_002.json']\r\n```\r\n","comment_length":27,"text":" load_dataset(\"multi_woz_v22\") NonMatchingChecksumError\nHi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n\r\n`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`\r\n\r\n\r\nThis will give the following error:\r\n\r\n```\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_003.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_004.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_005.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_006.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_007.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_008.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_009.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_010.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_012.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_013.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_014.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_015.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_016.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_017.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_002.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_002.json']\r\n```\r\n\nHi, I still meet the error when loading the datasets after upgradeing datasets.\r\n\r\nraise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json']","embeddings":[-0.2244368494,0.1764702946,-0.0287194345,0.1495722383,0.1895723343,0.0018984437,0.3650578558,0.4820295274,0.2514897287,0.1666063964,-0.0970349461,0.1781593114,-0.0822444186,0.0992015228,-0.1845492125,0.2732406855,0.0765193701,-0.0256412663,-0.1682780683,-0.0625795946,-0.2430096716,0.3292944729,-0.1593268216,-0.0338407457,-0.3486862779,0.130385235,0.173908785,0.336238414,-0.0838693827,-0.1523166895,0.4145028293,0.2369774878,0.0439615399,0.3686157465,-0.0001172875,0.163338542,0.2778438032,-0.1433111131,-0.1339644343,-0.1817747802,-0.5618035793,-0.3552265167,-0.1035682037,-0.1154623255,0.0706329644,0.3033826053,0.0572008304,-0.0867421255,0.2431910932,0.1712264121,0.1535670757,0.3319171369,0.2157697976,0.096905008,0.0455294512,-0.0774404854,-0.1376802474,0.0922583491,0.2744395733,-0.1569582969,-0.2395652235,0.2651384175,-0.1804820746,0.2693834305,-0.08042676,-0.151098907,0.1479792297,-0.1058241054,0.2105871737,0.5727722049,0.2205697149,-0.0554821864,-0.1762661785,-0.031950675,-0.2201738805,0.0833650157,0.3634192646,0.1834743917,-0.2078651488,0.0045875246,-0.4487929642,0.1443090439,-0.0263798088,0.2792589366,0.2121898234,0.3145956695,0.2003566772,0.0726620853,0.2279323786,-0.2064414769,0.1992104501,-0.3010312915,-0.1609574109,0.1293215007,-0.6801476479,-0.0664964691,0.0034521031,0.3319556117,0.4404266775,0.2289481759,0.4523443282,0.3736301064,-0.1520804018,0.1170620769,0.3968931139,0.1744454056,0.0657053292,-0.0545444787,0.1692391336,0.2095874697,-0.0953853428,0.23737818,-0.0305045638,-0.3612305224,0.349316746,0.1929925084,0.2069615573,-0.3434550762,-0.1813292205,0.0824042559,-0.2454786599,-0.1945316195,0.2718029022,0.2862129211,-0.0944880024,0.5524677038,-0.1653458029,0.2156011462,-0.1538057327,-0.184701696,-0.2573401332,-0.1516558081,0.0730739161,0.1531955004,0.2943665981,0.0656549037,0.3595133722,-0.0321129337,0.3835370243,-0.2031213194,0.333525002,0.0902094916,-0.069919236,0.2554130256,0.0134760654,0.2010277361,0.2036009133,-0.0260661189,-0.0622884482,0.2220196426,-0.332572788,-0.6167247295,0.1093568653,0.1607629508,-0.3595766425,-0.0804914013,0.0163278151,-0.2801302671,0.3050760031,-0.4042750597,-0.1545268744,-0.3094389141,-0.0430206098,-0.1179552451,0.0098633282,0.0942928717,-0.0934405178,0.0649290308,-0.0455776192,-0.1832666099,0.3468402326,0.2667627037,-0.2780414522,-0.1645423472,-0.1394770145,0.0187303592,0.2207069397,-0.0911147445,-0.3638109565,0.2024427652,-0.071666114,0.6232899427,0.1856193095,0.1097929999,-0.347364068,-0.1108146757,0.2366856933,0.0823844671,0.1350486577,0.1764032245,-0.2450098842,-0.3122074902,0.3476401865,0.1528952271,0.0161080807,-0.2277660966,0.0021591219,0.0893484354,0.5888454914,-0.229704082,-0.0971514285,0.173809126,0.0959679857,-0.2113368064,0.0982792377,0.0511620082,-0.4866259396,0.3225809634,-0.1565800011,-0.0987997502,0.3440404832,-0.0598260574,-0.0504282936,-0.1761434972,-0.3390328586,0.0107523073,0.0809487924,0.2748901248,0.083496578,-0.2535201013,-0.0927836373,0.3571139276,-0.3864992857,-0.0857931376,-0.4520960152,0.1969179064,-0.0879131109,0.0851735622,0.0764899477,0.1571474075,0.1050598249,-0.1094213575,-0.1870695502,0.5129078031,0.321377933,0.238428399,-0.0831710547,0.3510250747,-0.1519880891,-0.0498669744,-0.1007987484,0.5829588175,-0.043825414,-0.2062136233,-0.1941188127,0.4879204333,-0.1729985923,0.4535035193,-0.0467061065,0.1363264918,0.3366037309,-0.1314040124,-0.267632544,-0.094632037,0.4545117021,-0.1980919838,-0.166961506,0.2530014515,-0.2133246213,0.0163294896,0.3733881712,-0.1440781504,0.1259907186,-0.0151448352,0.2458733171,-0.2552571595,-0.0198079348,0.492271781,0.3632516265,0.1011830643,-0.017946247,-0.0273172073,-0.2294692695,-0.1744433939,0.1194066629,0.0973706469,0.3296886981,0.5370066166,0.0413935557,-0.3653722107,-0.4267033637,-0.0148492791,0.0290301628,0.1656199098,-0.441488415,-0.2537639439,-0.4441935718,0.2288448811,-0.3637366891,-0.3436980546,-0.4129472375,-0.4648984671,0.1210602671,0.2281651497,-0.019953087,0.3193200827,-0.3111640215,0.0683446303,-0.13192828,0.0605431758,0.0400538072,-0.2777514756,-0.0817120597,-0.0046348502,0.395745039,0.0018831765,-0.0578058697,-0.3077501059,-0.0692631528,-0.1458399743,-0.1667571664,-0.0308500472,-0.1108122692,-0.0727645904,0.258897841,0.3025155365,0.1404629946,-0.3886933029,0.2458182424,-0.0367405601,-0.2834911942,0.1877612323,-0.1721364707,-0.0189697389,0.1081124917,-0.3866266012,-0.0947575793,-0.3375903666,-0.0101160659,0.0664086565,0.2717181146,0.2532503307,0.0835985392,-0.067855671,0.135645777,0.2740627229,-0.3200554848,-0.8128382564,0.4206303358,-0.0225661304,-0.2601571679,-0.0509394854,0.0214568581,0.2018916756,0.0964379087,-0.4472939074,-0.2804222107,-0.0651813745,-0.3137300909,0.1682505161,-0.3037391305,0.3359626532,0.118827641,-0.0578323677,-0.244994387,-0.294922471,0.3670241833,0.0858874843,0.5151624084,-0.2128699422,0.2413334697,0.0072190906,0.59493047,0.2273386568,0.2582996488,0.0617827326,0.1863201261,0.1531594396,-0.0817795843,-0.3245143592,0.1173866391,-0.2651423812,0.004429182,0.1789305359,0.0948021784,-0.273284018,-0.1016846523,0.1297237724,-0.1512695104,-0.2311442494,0.0542182997,-0.0111719985,0.0999288857,0.2145664692,0.0766038224,-0.3172064722,-0.3501291573,0.1731933951,0.2011538148,-0.1653898358,-0.1424287409,-0.5950775743,0.0641577691,0.024697274,0.4123165309,0.2499673218,0.4031144381,0.0229396354,-0.3206753731,0.0325473547,-0.2579710484,0.5845400095,-0.3834357262,0.3248415589,0.0977153778,-0.0471826047,-0.2630657256,-0.1649410576,-0.0380526185,-0.0256651193,0.3266568482,0.3074330389,-0.2269275337,-0.2858472764,0.1157697365,0.337846756,-0.2878535688,-0.1682776213,-0.3097966015,-0.1108990908,-0.3571147323,-0.2201311439,-0.3995670676,0.3472657502,-0.1648672968,-0.0812498257,0.1711757928,-0.1453643888,0.2767299116,0.03994295,0.3021826744,0.3644610047,0.225110203,0.0248756167,0.338565141,0.2234014571,0.4796167314,-0.1350188404,-0.0993258879,-0.0563185886,-0.2143767029,0.0410142802,0.2235033959,0.0236763805,-0.0814053416,0.2965885103,-0.1108627766,-0.0433097072,0.0397988297,0.2845616937,0.0299271457,-0.3399115801,-0.301807642,0.0595081486,0.0007752888,-0.178718999,0.2506121993,-0.1772620231,-0.0745188445,0.0401590206,0.3154736757,0.8823344707,0.0781968459,0.0889303312,0.0817513838,-0.1757192612,-0.1717552245,-0.4199728668,0.1754765809,-0.2713772655,-0.2802045345,-0.0789559931,-0.2028966546,0.1161567271,0.3723813295,-0.0140839126,0.3294216692,0.0507585853,0.1718657315,0.015210106,0.0928249508,-0.1202962399,-0.1144701317,0.0429769382,0.1203733087,-0.1465322077,0.1447477788,-0.1230822057,-0.0494485982,0.2085236609,-0.2759677768,-0.1316172779,0.1984799504,-0.1262031645,0.4007564783,0.2463366538,-0.2069448829,-0.290071398,0.4346532226,0.0933009014,-0.0674183592,-0.2564155459,0.1446241438,0.3509513736,0.0936215445,-0.0919712707,-0.0624782331,0.3511782587,-0.1872355193,-0.2813732922,0.0104376664,-0.1517420858,-0.6179360747,-0.0681770965,-0.1264027506,0.0984783098,-0.154596284,-0.1535268128,0.1786635965,0.0546398424,-0.3420567513,0.1221394762,0.1330333799,-0.3626447618,0.3028178513,0.1293569803,-0.2770988643,-0.0944160596,0.554800272,0.0386738665,-0.001004266,0.7191022038,0.0710801035,-0.1204014197,-0.2845618129,0.3099564314,0.2341080755,-0.2122221738,-0.0682358071,-0.1181255877,-0.0549927317,0.040189594,0.5675612092,0.2766489387,0.3637252152,-0.0064766975,-0.6050378084,-0.122251153,0.0505504832,0.2656698227,0.3840138316,-0.3860800564,-0.0716586486,-0.120571062,0.2204845697,-0.2977004051,0.0136333136,-0.3014138937,-0.0014124378,-0.3319445252,-0.0595821477,0.1033880115,0.0466479473,0.1038635299,-0.2260658592,-0.2882824838,-0.1636383384,-0.2158522606,0.1587166041,0.0363478102,-0.0632579476,0.0698035806,-0.0939673111,-0.1066646129,-0.1600662321,-0.0529938377,0.2080393583,-0.0953483358,0.125788644,0.2332840264,-0.1949185878,-0.1487827599,-0.261736393,-0.2963095307,-0.0960339904,0.0908312425,-0.1021253839,0.0535941757,-0.169658795,0.4369630814,0.1701050401,0.024545135,-0.0090559432,0.2811068594,-0.2891628742,-0.0970124528,0.2284561545,0.3137277067,0.3578693867,-0.2104902118,0.1913179606,0.000825197,0.2238859683,-0.2928492129,-0.1559717953,0.3667841554,-0.112051852,0.0143712647,0.0317219533,-0.0716917664,-0.3091430366,0.5745686293,0.1000950933,-0.0016480027,-0.3915390968,0.1874298453,0.3277118504,-0.3724373281,-0.0696389079,-0.004597913,-0.2296410054,0.2119155675,-0.2145881206,-0.1123061702,0.4519319832,0.0501495712,0.250569135,-0.1284569353,-0.4928975999,0.0944712386,0.1954616159,-0.0912432149,0.0177286398,0.0348484144,0.4064955413,-0.2336776704,0.182787776,-0.702688992,-0.0180880725,0.0231124628,-0.2515980303,-0.3489803374,-0.3787186146,-0.1243415698,0.1621755213,-0.0076932451,0.1223282889,0.3172167838,0.0278804507,-0.0302277654,-0.5480276346,-0.1554399431,-0.072140418,-0.1325324923,-0.2546501458,0.2110592127,0.3817438781,-0.1168670058,0.3411186635,0.232888937,0.3793445826,0.3968301713,0.0158854928,-0.0075071477,-0.0547184944,0.0275844373,-0.151116848,0.7324585915,0.1668840051,-0.1373345107,0.4767388999,0.0521655828,-0.1582669616,0.0700558797,-0.0142776398,-0.0927828625,-0.0598390847,0.7789545655,0.0388093628,0.1453841925,-0.0719826072,0.2336157411,-0.3082112372,-0.3224420547,0.0849673524,0.0910015479,0.2019011974,-0.0258247387,0.0734821185,-0.1888687611,0.6922151446,-0.0983080342,-0.2206100374,-0.2996817231,-0.2596409619,-0.8195856214,0.2501328588,-0.1412824243,0.3424734771,-0.0689983666,0.1562742442,0.1727872491,0.2854847014,0.2575885057,-0.2129989117,-0.1863705069,-0.0521587506,-0.3973252773,0.0739042535,0.4234284163,0.0349841602,-0.0069678947,-0.2355769873,0.1889984906,0.2631671429,0.0424308777,0.1342669725,0.0111747002,-0.0327080488,-0.0246862136,-0.0812196285,0.1433468908,0.5623226762,-0.0031299428,-0.011459128,-0.1989398003,-0.2911090851,-0.0620972998,0.1848046333,-0.050203722,0.5143117309,-0.1861067712,-0.0308319367,-0.2220528424,0.2004486471,-0.0561484024,-0.2992598414,-0.1287595034,0.0325104445,-0.0127721233,0.084331587,0.1155453026,0.0505930148,0.0815804675,0.2093015313,-0.0561158136,-0.0594237521,0.7326385975,-0.1133245677,-0.1869971454,-0.3649466038,0.1383291632,-0.3404731452,-0.0648737326,-0.4481017292,0.0900497064,0.3190638721,0.1263615787,-0.2260794342,0.1400440037,-0.2574895024,0.0272709914,-0.0578922816,0.0775431693,0.2296063006,-0.2987921536,-0.1100094467,-0.1358176172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1876","title":" load_dataset(\"multi_woz_v22\") NonMatchingChecksumError","comments":"This must be related to https:\/\/github.com\/budzianowski\/multiwoz\/pull\/72\r\nThose files have changed, let me update the checksums for this dataset.\r\n\r\nFor now you can use `ignore_verifications=True` in `load_dataset` to skip the checksum verification.","body":"Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n\r\n`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`\r\n\r\n\r\nThis will give the following error:\r\n\r\n```\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_003.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_004.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_005.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_006.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_007.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_008.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_009.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_010.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_012.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_013.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_014.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_015.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_016.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_017.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_002.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_002.json']\r\n```\r\n","comment_length":31,"text":" load_dataset(\"multi_woz_v22\") NonMatchingChecksumError\nHi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n\r\n`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`\r\n\r\n\r\nThis will give the following error:\r\n\r\n```\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dialog_acts.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_003.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_004.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_005.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_006.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_007.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_008.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_009.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_010.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_012.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_013.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_014.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_015.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_016.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/train\/dialogues_017.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/dev\/dialogues_002.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_001.json', 'https:\/\/github.com\/budzianowski\/multiwoz\/raw\/master\/data\/MultiWOZ_2.2\/test\/dialogues_002.json']\r\n```\r\n\nThis must be related to https:\/\/github.com\/budzianowski\/multiwoz\/pull\/72\r\nThose files have changed, let me update the checksums for this dataset.\r\n\r\nFor now you can use `ignore_verifications=True` in `load_dataset` to skip the checksum verification.","embeddings":[-0.2244368494,0.1764702946,-0.0287194345,0.1495722383,0.1895723343,0.0018984437,0.3650578558,0.4820295274,0.2514897287,0.1666063964,-0.0970349461,0.1781593114,-0.0822444186,0.0992015228,-0.1845492125,0.2732406855,0.0765193701,-0.0256412663,-0.1682780683,-0.0625795946,-0.2430096716,0.3292944729,-0.1593268216,-0.0338407457,-0.3486862779,0.130385235,0.173908785,0.336238414,-0.0838693827,-0.1523166895,0.4145028293,0.2369774878,0.0439615399,0.3686157465,-0.0001172875,0.163338542,0.2778438032,-0.1433111131,-0.1339644343,-0.1817747802,-0.5618035793,-0.3552265167,-0.1035682037,-0.1154623255,0.0706329644,0.3033826053,0.0572008304,-0.0867421255,0.2431910932,0.1712264121,0.1535670757,0.3319171369,0.2157697976,0.096905008,0.0455294512,-0.0774404854,-0.1376802474,0.0922583491,0.2744395733,-0.1569582969,-0.2395652235,0.2651384175,-0.1804820746,0.2693834305,-0.08042676,-0.151098907,0.1479792297,-0.1058241054,0.2105871737,0.5727722049,0.2205697149,-0.0554821864,-0.1762661785,-0.031950675,-0.2201738805,0.0833650157,0.3634192646,0.1834743917,-0.2078651488,0.0045875246,-0.4487929642,0.1443090439,-0.0263798088,0.2792589366,0.2121898234,0.3145956695,0.2003566772,0.0726620853,0.2279323786,-0.2064414769,0.1992104501,-0.3010312915,-0.1609574109,0.1293215007,-0.6801476479,-0.0664964691,0.0034521031,0.3319556117,0.4404266775,0.2289481759,0.4523443282,0.3736301064,-0.1520804018,0.1170620769,0.3968931139,0.1744454056,0.0657053292,-0.0545444787,0.1692391336,0.2095874697,-0.0953853428,0.23737818,-0.0305045638,-0.3612305224,0.349316746,0.1929925084,0.2069615573,-0.3434550762,-0.1813292205,0.0824042559,-0.2454786599,-0.1945316195,0.2718029022,0.2862129211,-0.0944880024,0.5524677038,-0.1653458029,0.2156011462,-0.1538057327,-0.184701696,-0.2573401332,-0.1516558081,0.0730739161,0.1531955004,0.2943665981,0.0656549037,0.3595133722,-0.0321129337,0.3835370243,-0.2031213194,0.333525002,0.0902094916,-0.069919236,0.2554130256,0.0134760654,0.2010277361,0.2036009133,-0.0260661189,-0.0622884482,0.2220196426,-0.332572788,-0.6167247295,0.1093568653,0.1607629508,-0.3595766425,-0.0804914013,0.0163278151,-0.2801302671,0.3050760031,-0.4042750597,-0.1545268744,-0.3094389141,-0.0430206098,-0.1179552451,0.0098633282,0.0942928717,-0.0934405178,0.0649290308,-0.0455776192,-0.1832666099,0.3468402326,0.2667627037,-0.2780414522,-0.1645423472,-0.1394770145,0.0187303592,0.2207069397,-0.0911147445,-0.3638109565,0.2024427652,-0.071666114,0.6232899427,0.1856193095,0.1097929999,-0.347364068,-0.1108146757,0.2366856933,0.0823844671,0.1350486577,0.1764032245,-0.2450098842,-0.3122074902,0.3476401865,0.1528952271,0.0161080807,-0.2277660966,0.0021591219,0.0893484354,0.5888454914,-0.229704082,-0.0971514285,0.173809126,0.0959679857,-0.2113368064,0.0982792377,0.0511620082,-0.4866259396,0.3225809634,-0.1565800011,-0.0987997502,0.3440404832,-0.0598260574,-0.0504282936,-0.1761434972,-0.3390328586,0.0107523073,0.0809487924,0.2748901248,0.083496578,-0.2535201013,-0.0927836373,0.3571139276,-0.3864992857,-0.0857931376,-0.4520960152,0.1969179064,-0.0879131109,0.0851735622,0.0764899477,0.1571474075,0.1050598249,-0.1094213575,-0.1870695502,0.5129078031,0.321377933,0.238428399,-0.0831710547,0.3510250747,-0.1519880891,-0.0498669744,-0.1007987484,0.5829588175,-0.043825414,-0.2062136233,-0.1941188127,0.4879204333,-0.1729985923,0.4535035193,-0.0467061065,0.1363264918,0.3366037309,-0.1314040124,-0.267632544,-0.094632037,0.4545117021,-0.1980919838,-0.166961506,0.2530014515,-0.2133246213,0.0163294896,0.3733881712,-0.1440781504,0.1259907186,-0.0151448352,0.2458733171,-0.2552571595,-0.0198079348,0.492271781,0.3632516265,0.1011830643,-0.017946247,-0.0273172073,-0.2294692695,-0.1744433939,0.1194066629,0.0973706469,0.3296886981,0.5370066166,0.0413935557,-0.3653722107,-0.4267033637,-0.0148492791,0.0290301628,0.1656199098,-0.441488415,-0.2537639439,-0.4441935718,0.2288448811,-0.3637366891,-0.3436980546,-0.4129472375,-0.4648984671,0.1210602671,0.2281651497,-0.019953087,0.3193200827,-0.3111640215,0.0683446303,-0.13192828,0.0605431758,0.0400538072,-0.2777514756,-0.0817120597,-0.0046348502,0.395745039,0.0018831765,-0.0578058697,-0.3077501059,-0.0692631528,-0.1458399743,-0.1667571664,-0.0308500472,-0.1108122692,-0.0727645904,0.258897841,0.3025155365,0.1404629946,-0.3886933029,0.2458182424,-0.0367405601,-0.2834911942,0.1877612323,-0.1721364707,-0.0189697389,0.1081124917,-0.3866266012,-0.0947575793,-0.3375903666,-0.0101160659,0.0664086565,0.2717181146,0.2532503307,0.0835985392,-0.067855671,0.135645777,0.2740627229,-0.3200554848,-0.8128382564,0.4206303358,-0.0225661304,-0.2601571679,-0.0509394854,0.0214568581,0.2018916756,0.0964379087,-0.4472939074,-0.2804222107,-0.0651813745,-0.3137300909,0.1682505161,-0.3037391305,0.3359626532,0.118827641,-0.0578323677,-0.244994387,-0.294922471,0.3670241833,0.0858874843,0.5151624084,-0.2128699422,0.2413334697,0.0072190906,0.59493047,0.2273386568,0.2582996488,0.0617827326,0.1863201261,0.1531594396,-0.0817795843,-0.3245143592,0.1173866391,-0.2651423812,0.004429182,0.1789305359,0.0948021784,-0.273284018,-0.1016846523,0.1297237724,-0.1512695104,-0.2311442494,0.0542182997,-0.0111719985,0.0999288857,0.2145664692,0.0766038224,-0.3172064722,-0.3501291573,0.1731933951,0.2011538148,-0.1653898358,-0.1424287409,-0.5950775743,0.0641577691,0.024697274,0.4123165309,0.2499673218,0.4031144381,0.0229396354,-0.3206753731,0.0325473547,-0.2579710484,0.5845400095,-0.3834357262,0.3248415589,0.0977153778,-0.0471826047,-0.2630657256,-0.1649410576,-0.0380526185,-0.0256651193,0.3266568482,0.3074330389,-0.2269275337,-0.2858472764,0.1157697365,0.337846756,-0.2878535688,-0.1682776213,-0.3097966015,-0.1108990908,-0.3571147323,-0.2201311439,-0.3995670676,0.3472657502,-0.1648672968,-0.0812498257,0.1711757928,-0.1453643888,0.2767299116,0.03994295,0.3021826744,0.3644610047,0.225110203,0.0248756167,0.338565141,0.2234014571,0.4796167314,-0.1350188404,-0.0993258879,-0.0563185886,-0.2143767029,0.0410142802,0.2235033959,0.0236763805,-0.0814053416,0.2965885103,-0.1108627766,-0.0433097072,0.0397988297,0.2845616937,0.0299271457,-0.3399115801,-0.301807642,0.0595081486,0.0007752888,-0.178718999,0.2506121993,-0.1772620231,-0.0745188445,0.0401590206,0.3154736757,0.8823344707,0.0781968459,0.0889303312,0.0817513838,-0.1757192612,-0.1717552245,-0.4199728668,0.1754765809,-0.2713772655,-0.2802045345,-0.0789559931,-0.2028966546,0.1161567271,0.3723813295,-0.0140839126,0.3294216692,0.0507585853,0.1718657315,0.015210106,0.0928249508,-0.1202962399,-0.1144701317,0.0429769382,0.1203733087,-0.1465322077,0.1447477788,-0.1230822057,-0.0494485982,0.2085236609,-0.2759677768,-0.1316172779,0.1984799504,-0.1262031645,0.4007564783,0.2463366538,-0.2069448829,-0.290071398,0.4346532226,0.0933009014,-0.0674183592,-0.2564155459,0.1446241438,0.3509513736,0.0936215445,-0.0919712707,-0.0624782331,0.3511782587,-0.1872355193,-0.2813732922,0.0104376664,-0.1517420858,-0.6179360747,-0.0681770965,-0.1264027506,0.0984783098,-0.154596284,-0.1535268128,0.1786635965,0.0546398424,-0.3420567513,0.1221394762,0.1330333799,-0.3626447618,0.3028178513,0.1293569803,-0.2770988643,-0.0944160596,0.554800272,0.0386738665,-0.001004266,0.7191022038,0.0710801035,-0.1204014197,-0.2845618129,0.3099564314,0.2341080755,-0.2122221738,-0.0682358071,-0.1181255877,-0.0549927317,0.040189594,0.5675612092,0.2766489387,0.3637252152,-0.0064766975,-0.6050378084,-0.122251153,0.0505504832,0.2656698227,0.3840138316,-0.3860800564,-0.0716586486,-0.120571062,0.2204845697,-0.2977004051,0.0136333136,-0.3014138937,-0.0014124378,-0.3319445252,-0.0595821477,0.1033880115,0.0466479473,0.1038635299,-0.2260658592,-0.2882824838,-0.1636383384,-0.2158522606,0.1587166041,0.0363478102,-0.0632579476,0.0698035806,-0.0939673111,-0.1066646129,-0.1600662321,-0.0529938377,0.2080393583,-0.0953483358,0.125788644,0.2332840264,-0.1949185878,-0.1487827599,-0.261736393,-0.2963095307,-0.0960339904,0.0908312425,-0.1021253839,0.0535941757,-0.169658795,0.4369630814,0.1701050401,0.024545135,-0.0090559432,0.2811068594,-0.2891628742,-0.0970124528,0.2284561545,0.3137277067,0.3578693867,-0.2104902118,0.1913179606,0.000825197,0.2238859683,-0.2928492129,-0.1559717953,0.3667841554,-0.112051852,0.0143712647,0.0317219533,-0.0716917664,-0.3091430366,0.5745686293,0.1000950933,-0.0016480027,-0.3915390968,0.1874298453,0.3277118504,-0.3724373281,-0.0696389079,-0.004597913,-0.2296410054,0.2119155675,-0.2145881206,-0.1123061702,0.4519319832,0.0501495712,0.250569135,-0.1284569353,-0.4928975999,0.0944712386,0.1954616159,-0.0912432149,0.0177286398,0.0348484144,0.4064955413,-0.2336776704,0.182787776,-0.702688992,-0.0180880725,0.0231124628,-0.2515980303,-0.3489803374,-0.3787186146,-0.1243415698,0.1621755213,-0.0076932451,0.1223282889,0.3172167838,0.0278804507,-0.0302277654,-0.5480276346,-0.1554399431,-0.072140418,-0.1325324923,-0.2546501458,0.2110592127,0.3817438781,-0.1168670058,0.3411186635,0.232888937,0.3793445826,0.3968301713,0.0158854928,-0.0075071477,-0.0547184944,0.0275844373,-0.151116848,0.7324585915,0.1668840051,-0.1373345107,0.4767388999,0.0521655828,-0.1582669616,0.0700558797,-0.0142776398,-0.0927828625,-0.0598390847,0.7789545655,0.0388093628,0.1453841925,-0.0719826072,0.2336157411,-0.3082112372,-0.3224420547,0.0849673524,0.0910015479,0.2019011974,-0.0258247387,0.0734821185,-0.1888687611,0.6922151446,-0.0983080342,-0.2206100374,-0.2996817231,-0.2596409619,-0.8195856214,0.2501328588,-0.1412824243,0.3424734771,-0.0689983666,0.1562742442,0.1727872491,0.2854847014,0.2575885057,-0.2129989117,-0.1863705069,-0.0521587506,-0.3973252773,0.0739042535,0.4234284163,0.0349841602,-0.0069678947,-0.2355769873,0.1889984906,0.2631671429,0.0424308777,0.1342669725,0.0111747002,-0.0327080488,-0.0246862136,-0.0812196285,0.1433468908,0.5623226762,-0.0031299428,-0.011459128,-0.1989398003,-0.2911090851,-0.0620972998,0.1848046333,-0.050203722,0.5143117309,-0.1861067712,-0.0308319367,-0.2220528424,0.2004486471,-0.0561484024,-0.2992598414,-0.1287595034,0.0325104445,-0.0127721233,0.084331587,0.1155453026,0.0505930148,0.0815804675,0.2093015313,-0.0561158136,-0.0594237521,0.7326385975,-0.1133245677,-0.1869971454,-0.3649466038,0.1383291632,-0.3404731452,-0.0648737326,-0.4481017292,0.0900497064,0.3190638721,0.1263615787,-0.2260794342,0.1400440037,-0.2574895024,0.0272709914,-0.0578922816,0.0775431693,0.2296063006,-0.2987921536,-0.1100094467,-0.1358176172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1872","title":"Adding a new column to the dataset after set_format was called","comments":"Hi ! Indeed if you add a column to a formatted dataset, then the new dataset gets a new formatting in which:\r\n```\r\nnew formatted columns = (all columns - previously unformatted columns)\r\n```\r\nTherefore the new column is going to be formatted using the `torch` formatting.\r\n\r\nIf you want your new column to be unformatted you can re-run this line:\r\n```python\r\ndata.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n```","body":"Hi, \r\n\r\nthanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side. \r\n\r\nI load some lists of strings and integers, then call `data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)`. This converts the integer columns into tensors, but keeps the lists of strings as they are. I then call `map` to add a new column to my dataset, which is a **list of strings**. Once I iterate through my dataset, I get an error that the new column can't be converted into a tensor (which is probably caused by `set_format`). \r\n\r\nBelow some pseudo code:\r\n```python\r\n def augment_func(sample: Dict) -> Dict:\r\n # do something\r\n return {\r\n \"some_integer_column1\" : augmented_data[\"some_integer_column1\"], # <-- tensor\r\n \"some_integer_column2\" : augmented_data[\"some_integer_column2\"], # <-- tensor\r\n \"NEW_COLUMN\": targets, # <-- list of strings\r\n }\r\n\r\n\r\n data = datasets.load_dataset(__file__, data_dir=\"...\", split=\"train\")\r\n data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n\r\n augmented_dataset = data.map(augment_func, batched=False)\r\n \r\n for sample in augmented_dataset:\r\n print(sample) # fails\r\n\r\n```\r\n\r\nand the exception:\r\n```python\r\nTraceback (most recent call last):\r\n File \"dataset.py\", line 487, in \r\n main()\r\n File \"dataset.py\", line 471, in main\r\n for sample in augmented_dataset:\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 697, in __iter__\r\n yield self._getitem(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1069, in _getitem\r\n outputs = self._convert_outputs(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nThanks!\r\n","comment_length":67,"text":"Adding a new column to the dataset after set_format was called\nHi, \r\n\r\nthanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side. \r\n\r\nI load some lists of strings and integers, then call `data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)`. This converts the integer columns into tensors, but keeps the lists of strings as they are. I then call `map` to add a new column to my dataset, which is a **list of strings**. Once I iterate through my dataset, I get an error that the new column can't be converted into a tensor (which is probably caused by `set_format`). \r\n\r\nBelow some pseudo code:\r\n```python\r\n def augment_func(sample: Dict) -> Dict:\r\n # do something\r\n return {\r\n \"some_integer_column1\" : augmented_data[\"some_integer_column1\"], # <-- tensor\r\n \"some_integer_column2\" : augmented_data[\"some_integer_column2\"], # <-- tensor\r\n \"NEW_COLUMN\": targets, # <-- list of strings\r\n }\r\n\r\n\r\n data = datasets.load_dataset(__file__, data_dir=\"...\", split=\"train\")\r\n data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n\r\n augmented_dataset = data.map(augment_func, batched=False)\r\n \r\n for sample in augmented_dataset:\r\n print(sample) # fails\r\n\r\n```\r\n\r\nand the exception:\r\n```python\r\nTraceback (most recent call last):\r\n File \"dataset.py\", line 487, in \r\n main()\r\n File \"dataset.py\", line 471, in main\r\n for sample in augmented_dataset:\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 697, in __iter__\r\n yield self._getitem(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1069, in _getitem\r\n outputs = self._convert_outputs(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nThanks!\r\n\nHi ! Indeed if you add a column to a formatted dataset, then the new dataset gets a new formatting in which:\r\n```\r\nnew formatted columns = (all columns - previously unformatted columns)\r\n```\r\nTherefore the new column is going to be formatted using the `torch` formatting.\r\n\r\nIf you want your new column to be unformatted you can re-run this line:\r\n```python\r\ndata.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n```","embeddings":[-0.163837567,-0.1748755127,-0.0401732326,-0.0492338166,0.4729719162,0.2625342906,0.6921063662,0.4177582264,0.16875045,-0.2805859745,0.1421070397,0.3377909064,-0.234039858,-0.0259004012,0.1978132278,-0.1026414856,0.1268425137,0.0638930127,0.2461459339,0.0586879924,-0.3777551651,0.1032360643,-0.1540207863,-0.2058055401,-0.1765669137,-0.3008010983,-0.0024790808,-0.1966911554,-0.1602552682,-0.2089734375,0.1871719062,-0.239169538,-0.0005563378,0.6491594315,-0.0001178686,0.1240975857,0.1696617603,0.031614352,-0.0399997793,-0.1928997934,-0.2241786718,-0.3875743151,-0.1664160639,-0.3292916119,0.1402370483,-0.090086028,-0.4265715182,-0.7763667107,0.2612175345,0.5016534925,0.1658125967,0.1779137254,-0.0243937224,-0.0470924303,0.2402329147,0.2883871496,-0.2155929655,-0.0182730332,-0.0107632391,0.1259519309,0.5092816353,0.1936923265,-0.2842826843,0.1020674407,0.106398195,0.2241102755,-0.2291458696,-0.0425992012,-0.0292203575,0.1993377954,0.6996715069,-0.254180491,-0.3642950058,-0.2796714902,0.0794099048,-0.4402051568,0.119070828,-0.0155838113,-0.0778750554,-0.1121089906,0.0913000554,-0.1839411706,-0.0351652093,0.3412312865,-0.1856348664,0.0189567804,-0.0303626508,0.1920932084,-0.1479987055,-0.0397582315,-0.1474709958,0.1262124032,0.0943790153,0.2419148088,-0.4515655339,-0.1876330823,0.033643987,-0.5349569917,-0.2338584661,-0.1157705784,0.1772542596,0.0031550087,-0.187983036,0.1016870588,0.1837739646,-0.0243844315,-0.2377842367,0.202795282,-0.058971107,-0.2029019147,0.0786278844,-0.1089506671,0.2219047099,-0.2733579278,0.1720267087,0.0052718245,0.415762037,0.1181959957,-0.0770012587,0.2277686745,-0.2703928947,-0.2729817927,0.0384290256,0.0916894153,-0.027375767,0.1280095875,0.140044421,0.2697374821,-0.0306830835,0.1996581256,-0.0518040881,0.0073224418,-0.1836284101,0.0456944257,0.1025165766,0.1123385131,-0.1570651531,0.2218846977,-0.13670744,-0.0665040687,0.0907948315,-0.2595322132,0.3443582356,0.1951227039,-0.1425132006,0.3772317767,0.1710849255,-0.0373196006,-0.1732723415,0.2277605534,-0.2527373135,-0.3213145733,-0.4081810117,0.0944307446,-0.0362261981,-0.1098731384,0.0837554261,0.1692531705,0.5626441836,-0.1914578974,0.1816345602,-0.2311049104,-0.3539713025,-0.2420078516,-0.0327599496,0.1081718057,-0.6691618562,0.0941578373,0.0791116133,0.3833465278,0.0772274733,0.1891558468,-0.0684040338,0.3359542787,-0.2292346805,0.0447732247,0.6186287999,0.0500695035,-0.1436017752,0.065374814,-0.1200989783,0.1486835629,-0.0731995106,0.1894952059,0.0845489725,-0.0850279555,0.3703552485,0.0103158476,-0.1337237805,-0.1136995554,-0.0130396793,0.1656283587,0.5390791893,-0.2064686865,0.0425463095,0.1790866256,-0.1250628978,0.2282993644,-0.017285509,-0.0337756947,-0.0073143095,0.1034121662,0.2751893401,0.0653935894,0.0355755165,-0.2091534138,-0.3796861172,-0.0080606909,0.4330004454,0.021443544,-0.00806686,-0.1484539062,-0.3776355088,0.0075415815,-0.0111386655,0.2967942953,0.0175653826,0.166564554,-0.1642278135,-0.1772773117,-0.338902384,0.0949272737,-0.1961665452,-0.0513706133,-0.2434315234,0.1282278299,0.1284624189,-0.2729171216,-0.1948995143,0.3867044747,0.1357645839,-0.0230477694,-0.1044372097,0.2309470475,0.0756757632,-0.1251296699,-0.4924203455,0.2001788765,0.1332603395,0.1238413602,-0.1664903164,-0.1851825267,-0.0113888839,-0.2167836577,-0.0754593462,0.050736133,0.0049540126,0.2880735099,-0.3090748787,0.1987280101,0.1029938012,0.0727731362,-0.2182666212,-0.3290410638,-0.0560583398,0.0218479354,0.0440452658,-0.2663611472,-0.3668257594,0.0163654499,0.6044391394,0.0032727038,0.1298659891,0.1421494484,0.1310873926,0.0832142532,0.0108531453,0.0890294835,0.4790549278,0.0814551562,-0.0663492233,0.023984842,-0.2592504323,-0.098174952,0.0403772667,0.2914122641,0.2057984024,0.0817342773,0.2835640609,0.0091479477,0.0822448358,0.0117116934,0.2326989472,0.1617716104,-0.5018972754,-0.0692816153,-0.127254352,0.2658206522,-0.2941197157,-0.6862946153,0.0144574651,-0.3176656961,-0.2221376598,0.0013733351,0.0248833746,0.0532939658,-0.0381292999,0.1013478041,0.2265940011,-0.4410072267,-0.0932101756,-0.2546205223,-0.0083848387,0.0140890237,0.3546907604,-0.3220357895,0.4234782755,0.1450334787,-0.1792608052,-0.1153635308,-0.3000035286,-0.0179306921,-0.3884204924,-0.0696965829,0.257756561,0.1516362876,-0.1477588415,-0.1888682097,0.3712129593,0.1980449259,0.2000721395,0.1980653405,0.1245261729,0.0569173545,-0.0044003814,-0.5688668489,-0.1075387448,-0.0420352966,0.0825474933,-0.0878118873,0.0984746888,0.0212098062,0.2226237953,0.1303482354,0.2398899049,-0.0894926563,-0.3793279827,-0.0635955632,0.3975252807,-0.3226929605,-0.2073272765,0.1397510022,-0.0274451841,0.181092605,0.2993002832,-0.3082209527,-0.1508528888,-0.1311350614,0.173268497,-0.2809301317,-0.0994715616,0.5064876676,0.0120290006,-0.0551418066,-0.1404468119,-0.2117807865,0.1956448406,0.2700185776,0.0633992553,0.2628600001,0.547018826,0.088467814,0.3008495867,0.0786023214,-0.4974924028,0.2550722659,-0.30017066,0.1966451108,-0.2971628308,-0.3937258124,-0.0502180122,-0.234202832,-0.0278095137,-0.1490178406,-0.1602451056,-0.0322328135,0.014737457,0.5779079795,-0.1564670354,-0.1584125012,0.3528257012,-0.4837839007,0.3728365898,0.0443864465,0.1514329016,-0.5375984907,-0.1796579808,0.1274439245,-0.0810083896,0.2100242227,-0.1992975175,-0.456897676,-0.1856812686,-0.4243353009,0.2998924851,0.3504037857,0.7206824422,-0.1428639144,-0.0488887243,-0.1048103273,0.1505411416,0.8693706393,-0.5131114125,-0.512216568,0.4151348174,0.2547292411,-0.6216635108,-0.0057002967,-0.1735690981,0.0902562961,0.0768063143,0.6951243877,-0.0104244826,0.1401380748,0.1530417055,0.3333971202,0.0032469353,-0.1188208982,-0.1659090966,-0.1762067527,-0.1543740183,0.0938118175,0.1468466073,0.3185719252,0.3098247945,-0.2787633836,-0.08227963,-0.1979757547,0.2580941916,-0.1493848711,0.2094887495,0.4666827023,0.1809091419,0.0732228234,-0.0207394809,-0.0134861525,0.2105140388,-0.2058895677,-0.0426405296,0.151575014,-0.3012671173,0.3134796023,0.1650590003,-0.0190843772,0.1892253757,-0.2329683304,-0.0420833416,-0.3310903907,0.1621735394,0.5876699686,0.210772112,-0.3579185307,-0.7363798022,0.2013388574,0.0528041907,0.0794652179,0.5957023501,-0.0100870384,-0.1674926728,0.4298977554,0.1278735399,0.8010354042,-0.2321156859,0.0219281632,0.2473872602,-0.1963775903,0.4466971457,-0.2197206169,0.2085921913,-0.1105169877,-0.0548965931,-0.1673021764,-0.2323620021,0.3513605893,0.0045412122,-0.1806060672,-0.1170175001,-0.2128399611,0.6637679935,-0.0228089467,0.105200097,0.148751229,-0.4528320432,-0.3152978718,0.1043530777,-0.1289207488,-0.2021534294,-0.0943826064,0.1820892394,-0.1783920676,-0.208489418,-0.3607927859,-0.0663562045,-0.0690187514,-0.0468821786,0.0452600047,-0.6082686186,0.4602432847,0.3592826128,0.5291901827,0.0879840255,-0.1030571461,-0.0565395281,0.0255032834,0.1595400572,0.2783373892,-0.4173798561,0.3079517186,0.1974584162,-0.1584801227,0.3220860362,0.0993196443,-0.2854143679,-0.1614233255,0.0934315622,0.5090245008,-0.6628907323,-0.3925191462,0.0586135685,0.2449562699,-0.2578346431,0.0919436365,-0.0922592208,0.0138549255,0.376259625,-0.3261744976,-0.0639200583,0.0640054941,0.2678512037,0.0133870263,-0.073792614,0.5867010355,0.2021547109,-0.1515664756,-0.1159019545,0.0242914502,0.1093551666,-0.2628646493,0.0673720762,-0.0161277931,0.1062919125,-0.1352067292,0.0582707115,-0.1605557799,0.2047499418,-0.0174369756,-0.2758021653,-0.1192292348,-0.0141024506,0.275123775,0.2663635314,0.1359473318,0.2398539782,-0.1381954402,0.0437084623,-0.235125497,-0.1114267334,0.1152756885,0.2109810859,0.0536024421,0.056384366,0.153965503,-0.4075026512,0.0870691091,0.4069389105,0.0289605074,-0.2118223459,-0.1445440799,0.1017587483,0.117402859,-0.2523890734,0.1890082806,-0.2974910736,-0.1747690439,-0.2785748243,0.019304812,0.3081176281,-0.2520097494,-0.0051214676,-0.1206487492,0.1459065378,-0.0088552516,-0.0315668806,-0.1970043629,0.0310068186,0.1981580853,0.2192884535,0.2022118717,0.0440381356,-0.3146225512,0.217801109,0.1843610257,0.0563184284,0.2494207919,-0.4320304096,0.0137219429,0.000993419,0.3522247374,-0.0245949402,-0.2163707316,-0.0196516439,0.2225755006,0.1150809228,-0.2363813668,0.0546980016,0.3840655386,0.1135786548,0.1640511006,0.2766043842,0.3523633182,0.1498499066,-0.2969774902,-0.0283383615,0.1019247696,-0.3147266805,0.3177816272,0.636660099,-0.0096956622,0.0432766713,0.2121617943,0.0385080129,-0.0184341837,0.2452431619,0.1546956152,0.1323665231,0.553768754,-0.0233231504,-0.1268625706,-0.3789405227,0.3849281669,0.0655262396,0.1893929541,0.356595099,0.2170319408,0.4540134072,-0.1332535744,-0.0467361659,-0.0272156466,0.0026210472,-0.140909031,0.0477861241,-0.1536695212,0.0255794786,0.0613307096,-0.2604688704,-0.0698648915,-0.2468580753,0.2972401083,-0.0647751167,0.0898903459,0.0116090039,-0.0867135897,-0.1450734586,-0.1023249179,-0.2088188231,0.1428881139,-0.0000564576,-0.0023569872,0.1685059816,0.0797035322,0.3124370277,0.3186620772,0.0319544822,0.3550901115,0.1073564887,-0.01895307,0.2020024359,0.1107800528,-0.1152407229,0.0459671766,0.2320191562,0.1205592304,-0.1050471067,0.1272472143,0.0860555097,0.3713683784,-0.1742448211,0.1659411341,-0.0186983813,-0.2119493484,-0.281200707,-0.0256917533,-0.0536646545,-0.2551568151,0.5833733678,-0.0171415377,0.0827371329,0.1473907977,0.0789523497,0.016913807,0.1726586819,0.1242832616,-0.0562416874,-0.2194082886,-0.1105198041,-0.4495842755,0.0777268559,-0.2444209903,-0.1423133016,0.1331401169,0.1083743572,0.3426305652,-0.0715117231,0.3546713591,-0.4199867249,-0.3278799355,0.2354024798,-0.0737499967,-0.3474462926,-0.1158577353,-0.0218217112,0.0036887003,-0.194622457,0.1039047241,0.3052025735,-0.0131549789,-0.28851372,-0.0521491319,-0.1672354192,-0.0539736487,0.5608680248,0.0929078907,0.4217212498,-0.1400011033,0.0661325604,-0.0567793697,0.0003704348,-0.2170345485,0.4287863374,-0.1763373911,0.4928040504,-0.2431970835,0.087903291,-0.3595514297,0.233363688,-0.0096810572,0.0360782631,-0.2453804612,0.4265148938,-0.1822773665,-0.051156316,0.160516575,0.3011433184,0.0987175107,0.6137595177,-0.1845930517,-0.2232504487,0.4984788001,-0.4976068437,-0.424870342,0.0814748928,-0.0286826529,0.1198043823,0.0639730245,-0.56647259,-0.295845747,0.4186934233,-0.0994040743,-0.0072000106,0.088709414,-0.0215018094,0.1368441135,-0.1779536754,0.1924846768,-0.169606626,-0.1580830216,-0.0421707109,-0.2822973728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1872","title":"Adding a new column to the dataset after set_format was called","comments":"Hi, thanks that solved my problem. Maybe mention that in the documentation. ","body":"Hi, \r\n\r\nthanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side. \r\n\r\nI load some lists of strings and integers, then call `data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)`. This converts the integer columns into tensors, but keeps the lists of strings as they are. I then call `map` to add a new column to my dataset, which is a **list of strings**. Once I iterate through my dataset, I get an error that the new column can't be converted into a tensor (which is probably caused by `set_format`). \r\n\r\nBelow some pseudo code:\r\n```python\r\n def augment_func(sample: Dict) -> Dict:\r\n # do something\r\n return {\r\n \"some_integer_column1\" : augmented_data[\"some_integer_column1\"], # <-- tensor\r\n \"some_integer_column2\" : augmented_data[\"some_integer_column2\"], # <-- tensor\r\n \"NEW_COLUMN\": targets, # <-- list of strings\r\n }\r\n\r\n\r\n data = datasets.load_dataset(__file__, data_dir=\"...\", split=\"train\")\r\n data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n\r\n augmented_dataset = data.map(augment_func, batched=False)\r\n \r\n for sample in augmented_dataset:\r\n print(sample) # fails\r\n\r\n```\r\n\r\nand the exception:\r\n```python\r\nTraceback (most recent call last):\r\n File \"dataset.py\", line 487, in \r\n main()\r\n File \"dataset.py\", line 471, in main\r\n for sample in augmented_dataset:\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 697, in __iter__\r\n yield self._getitem(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1069, in _getitem\r\n outputs = self._convert_outputs(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nThanks!\r\n","comment_length":12,"text":"Adding a new column to the dataset after set_format was called\nHi, \r\n\r\nthanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side. \r\n\r\nI load some lists of strings and integers, then call `data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)`. This converts the integer columns into tensors, but keeps the lists of strings as they are. I then call `map` to add a new column to my dataset, which is a **list of strings**. Once I iterate through my dataset, I get an error that the new column can't be converted into a tensor (which is probably caused by `set_format`). \r\n\r\nBelow some pseudo code:\r\n```python\r\n def augment_func(sample: Dict) -> Dict:\r\n # do something\r\n return {\r\n \"some_integer_column1\" : augmented_data[\"some_integer_column1\"], # <-- tensor\r\n \"some_integer_column2\" : augmented_data[\"some_integer_column2\"], # <-- tensor\r\n \"NEW_COLUMN\": targets, # <-- list of strings\r\n }\r\n\r\n\r\n data = datasets.load_dataset(__file__, data_dir=\"...\", split=\"train\")\r\n data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n\r\n augmented_dataset = data.map(augment_func, batched=False)\r\n \r\n for sample in augmented_dataset:\r\n print(sample) # fails\r\n\r\n```\r\n\r\nand the exception:\r\n```python\r\nTraceback (most recent call last):\r\n File \"dataset.py\", line 487, in \r\n main()\r\n File \"dataset.py\", line 471, in main\r\n for sample in augmented_dataset:\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 697, in __iter__\r\n yield self._getitem(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1069, in _getitem\r\n outputs = self._convert_outputs(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nThanks!\r\n\nHi, thanks that solved my problem. Maybe mention that in the documentation. ","embeddings":[-0.163837567,-0.1748755127,-0.0401732326,-0.0492338166,0.4729719162,0.2625342906,0.6921063662,0.4177582264,0.16875045,-0.2805859745,0.1421070397,0.3377909064,-0.234039858,-0.0259004012,0.1978132278,-0.1026414856,0.1268425137,0.0638930127,0.2461459339,0.0586879924,-0.3777551651,0.1032360643,-0.1540207863,-0.2058055401,-0.1765669137,-0.3008010983,-0.0024790808,-0.1966911554,-0.1602552682,-0.2089734375,0.1871719062,-0.239169538,-0.0005563378,0.6491594315,-0.0001178686,0.1240975857,0.1696617603,0.031614352,-0.0399997793,-0.1928997934,-0.2241786718,-0.3875743151,-0.1664160639,-0.3292916119,0.1402370483,-0.090086028,-0.4265715182,-0.7763667107,0.2612175345,0.5016534925,0.1658125967,0.1779137254,-0.0243937224,-0.0470924303,0.2402329147,0.2883871496,-0.2155929655,-0.0182730332,-0.0107632391,0.1259519309,0.5092816353,0.1936923265,-0.2842826843,0.1020674407,0.106398195,0.2241102755,-0.2291458696,-0.0425992012,-0.0292203575,0.1993377954,0.6996715069,-0.254180491,-0.3642950058,-0.2796714902,0.0794099048,-0.4402051568,0.119070828,-0.0155838113,-0.0778750554,-0.1121089906,0.0913000554,-0.1839411706,-0.0351652093,0.3412312865,-0.1856348664,0.0189567804,-0.0303626508,0.1920932084,-0.1479987055,-0.0397582315,-0.1474709958,0.1262124032,0.0943790153,0.2419148088,-0.4515655339,-0.1876330823,0.033643987,-0.5349569917,-0.2338584661,-0.1157705784,0.1772542596,0.0031550087,-0.187983036,0.1016870588,0.1837739646,-0.0243844315,-0.2377842367,0.202795282,-0.058971107,-0.2029019147,0.0786278844,-0.1089506671,0.2219047099,-0.2733579278,0.1720267087,0.0052718245,0.415762037,0.1181959957,-0.0770012587,0.2277686745,-0.2703928947,-0.2729817927,0.0384290256,0.0916894153,-0.027375767,0.1280095875,0.140044421,0.2697374821,-0.0306830835,0.1996581256,-0.0518040881,0.0073224418,-0.1836284101,0.0456944257,0.1025165766,0.1123385131,-0.1570651531,0.2218846977,-0.13670744,-0.0665040687,0.0907948315,-0.2595322132,0.3443582356,0.1951227039,-0.1425132006,0.3772317767,0.1710849255,-0.0373196006,-0.1732723415,0.2277605534,-0.2527373135,-0.3213145733,-0.4081810117,0.0944307446,-0.0362261981,-0.1098731384,0.0837554261,0.1692531705,0.5626441836,-0.1914578974,0.1816345602,-0.2311049104,-0.3539713025,-0.2420078516,-0.0327599496,0.1081718057,-0.6691618562,0.0941578373,0.0791116133,0.3833465278,0.0772274733,0.1891558468,-0.0684040338,0.3359542787,-0.2292346805,0.0447732247,0.6186287999,0.0500695035,-0.1436017752,0.065374814,-0.1200989783,0.1486835629,-0.0731995106,0.1894952059,0.0845489725,-0.0850279555,0.3703552485,0.0103158476,-0.1337237805,-0.1136995554,-0.0130396793,0.1656283587,0.5390791893,-0.2064686865,0.0425463095,0.1790866256,-0.1250628978,0.2282993644,-0.017285509,-0.0337756947,-0.0073143095,0.1034121662,0.2751893401,0.0653935894,0.0355755165,-0.2091534138,-0.3796861172,-0.0080606909,0.4330004454,0.021443544,-0.00806686,-0.1484539062,-0.3776355088,0.0075415815,-0.0111386655,0.2967942953,0.0175653826,0.166564554,-0.1642278135,-0.1772773117,-0.338902384,0.0949272737,-0.1961665452,-0.0513706133,-0.2434315234,0.1282278299,0.1284624189,-0.2729171216,-0.1948995143,0.3867044747,0.1357645839,-0.0230477694,-0.1044372097,0.2309470475,0.0756757632,-0.1251296699,-0.4924203455,0.2001788765,0.1332603395,0.1238413602,-0.1664903164,-0.1851825267,-0.0113888839,-0.2167836577,-0.0754593462,0.050736133,0.0049540126,0.2880735099,-0.3090748787,0.1987280101,0.1029938012,0.0727731362,-0.2182666212,-0.3290410638,-0.0560583398,0.0218479354,0.0440452658,-0.2663611472,-0.3668257594,0.0163654499,0.6044391394,0.0032727038,0.1298659891,0.1421494484,0.1310873926,0.0832142532,0.0108531453,0.0890294835,0.4790549278,0.0814551562,-0.0663492233,0.023984842,-0.2592504323,-0.098174952,0.0403772667,0.2914122641,0.2057984024,0.0817342773,0.2835640609,0.0091479477,0.0822448358,0.0117116934,0.2326989472,0.1617716104,-0.5018972754,-0.0692816153,-0.127254352,0.2658206522,-0.2941197157,-0.6862946153,0.0144574651,-0.3176656961,-0.2221376598,0.0013733351,0.0248833746,0.0532939658,-0.0381292999,0.1013478041,0.2265940011,-0.4410072267,-0.0932101756,-0.2546205223,-0.0083848387,0.0140890237,0.3546907604,-0.3220357895,0.4234782755,0.1450334787,-0.1792608052,-0.1153635308,-0.3000035286,-0.0179306921,-0.3884204924,-0.0696965829,0.257756561,0.1516362876,-0.1477588415,-0.1888682097,0.3712129593,0.1980449259,0.2000721395,0.1980653405,0.1245261729,0.0569173545,-0.0044003814,-0.5688668489,-0.1075387448,-0.0420352966,0.0825474933,-0.0878118873,0.0984746888,0.0212098062,0.2226237953,0.1303482354,0.2398899049,-0.0894926563,-0.3793279827,-0.0635955632,0.3975252807,-0.3226929605,-0.2073272765,0.1397510022,-0.0274451841,0.181092605,0.2993002832,-0.3082209527,-0.1508528888,-0.1311350614,0.173268497,-0.2809301317,-0.0994715616,0.5064876676,0.0120290006,-0.0551418066,-0.1404468119,-0.2117807865,0.1956448406,0.2700185776,0.0633992553,0.2628600001,0.547018826,0.088467814,0.3008495867,0.0786023214,-0.4974924028,0.2550722659,-0.30017066,0.1966451108,-0.2971628308,-0.3937258124,-0.0502180122,-0.234202832,-0.0278095137,-0.1490178406,-0.1602451056,-0.0322328135,0.014737457,0.5779079795,-0.1564670354,-0.1584125012,0.3528257012,-0.4837839007,0.3728365898,0.0443864465,0.1514329016,-0.5375984907,-0.1796579808,0.1274439245,-0.0810083896,0.2100242227,-0.1992975175,-0.456897676,-0.1856812686,-0.4243353009,0.2998924851,0.3504037857,0.7206824422,-0.1428639144,-0.0488887243,-0.1048103273,0.1505411416,0.8693706393,-0.5131114125,-0.512216568,0.4151348174,0.2547292411,-0.6216635108,-0.0057002967,-0.1735690981,0.0902562961,0.0768063143,0.6951243877,-0.0104244826,0.1401380748,0.1530417055,0.3333971202,0.0032469353,-0.1188208982,-0.1659090966,-0.1762067527,-0.1543740183,0.0938118175,0.1468466073,0.3185719252,0.3098247945,-0.2787633836,-0.08227963,-0.1979757547,0.2580941916,-0.1493848711,0.2094887495,0.4666827023,0.1809091419,0.0732228234,-0.0207394809,-0.0134861525,0.2105140388,-0.2058895677,-0.0426405296,0.151575014,-0.3012671173,0.3134796023,0.1650590003,-0.0190843772,0.1892253757,-0.2329683304,-0.0420833416,-0.3310903907,0.1621735394,0.5876699686,0.210772112,-0.3579185307,-0.7363798022,0.2013388574,0.0528041907,0.0794652179,0.5957023501,-0.0100870384,-0.1674926728,0.4298977554,0.1278735399,0.8010354042,-0.2321156859,0.0219281632,0.2473872602,-0.1963775903,0.4466971457,-0.2197206169,0.2085921913,-0.1105169877,-0.0548965931,-0.1673021764,-0.2323620021,0.3513605893,0.0045412122,-0.1806060672,-0.1170175001,-0.2128399611,0.6637679935,-0.0228089467,0.105200097,0.148751229,-0.4528320432,-0.3152978718,0.1043530777,-0.1289207488,-0.2021534294,-0.0943826064,0.1820892394,-0.1783920676,-0.208489418,-0.3607927859,-0.0663562045,-0.0690187514,-0.0468821786,0.0452600047,-0.6082686186,0.4602432847,0.3592826128,0.5291901827,0.0879840255,-0.1030571461,-0.0565395281,0.0255032834,0.1595400572,0.2783373892,-0.4173798561,0.3079517186,0.1974584162,-0.1584801227,0.3220860362,0.0993196443,-0.2854143679,-0.1614233255,0.0934315622,0.5090245008,-0.6628907323,-0.3925191462,0.0586135685,0.2449562699,-0.2578346431,0.0919436365,-0.0922592208,0.0138549255,0.376259625,-0.3261744976,-0.0639200583,0.0640054941,0.2678512037,0.0133870263,-0.073792614,0.5867010355,0.2021547109,-0.1515664756,-0.1159019545,0.0242914502,0.1093551666,-0.2628646493,0.0673720762,-0.0161277931,0.1062919125,-0.1352067292,0.0582707115,-0.1605557799,0.2047499418,-0.0174369756,-0.2758021653,-0.1192292348,-0.0141024506,0.275123775,0.2663635314,0.1359473318,0.2398539782,-0.1381954402,0.0437084623,-0.235125497,-0.1114267334,0.1152756885,0.2109810859,0.0536024421,0.056384366,0.153965503,-0.4075026512,0.0870691091,0.4069389105,0.0289605074,-0.2118223459,-0.1445440799,0.1017587483,0.117402859,-0.2523890734,0.1890082806,-0.2974910736,-0.1747690439,-0.2785748243,0.019304812,0.3081176281,-0.2520097494,-0.0051214676,-0.1206487492,0.1459065378,-0.0088552516,-0.0315668806,-0.1970043629,0.0310068186,0.1981580853,0.2192884535,0.2022118717,0.0440381356,-0.3146225512,0.217801109,0.1843610257,0.0563184284,0.2494207919,-0.4320304096,0.0137219429,0.000993419,0.3522247374,-0.0245949402,-0.2163707316,-0.0196516439,0.2225755006,0.1150809228,-0.2363813668,0.0546980016,0.3840655386,0.1135786548,0.1640511006,0.2766043842,0.3523633182,0.1498499066,-0.2969774902,-0.0283383615,0.1019247696,-0.3147266805,0.3177816272,0.636660099,-0.0096956622,0.0432766713,0.2121617943,0.0385080129,-0.0184341837,0.2452431619,0.1546956152,0.1323665231,0.553768754,-0.0233231504,-0.1268625706,-0.3789405227,0.3849281669,0.0655262396,0.1893929541,0.356595099,0.2170319408,0.4540134072,-0.1332535744,-0.0467361659,-0.0272156466,0.0026210472,-0.140909031,0.0477861241,-0.1536695212,0.0255794786,0.0613307096,-0.2604688704,-0.0698648915,-0.2468580753,0.2972401083,-0.0647751167,0.0898903459,0.0116090039,-0.0867135897,-0.1450734586,-0.1023249179,-0.2088188231,0.1428881139,-0.0000564576,-0.0023569872,0.1685059816,0.0797035322,0.3124370277,0.3186620772,0.0319544822,0.3550901115,0.1073564887,-0.01895307,0.2020024359,0.1107800528,-0.1152407229,0.0459671766,0.2320191562,0.1205592304,-0.1050471067,0.1272472143,0.0860555097,0.3713683784,-0.1742448211,0.1659411341,-0.0186983813,-0.2119493484,-0.281200707,-0.0256917533,-0.0536646545,-0.2551568151,0.5833733678,-0.0171415377,0.0827371329,0.1473907977,0.0789523497,0.016913807,0.1726586819,0.1242832616,-0.0562416874,-0.2194082886,-0.1105198041,-0.4495842755,0.0777268559,-0.2444209903,-0.1423133016,0.1331401169,0.1083743572,0.3426305652,-0.0715117231,0.3546713591,-0.4199867249,-0.3278799355,0.2354024798,-0.0737499967,-0.3474462926,-0.1158577353,-0.0218217112,0.0036887003,-0.194622457,0.1039047241,0.3052025735,-0.0131549789,-0.28851372,-0.0521491319,-0.1672354192,-0.0539736487,0.5608680248,0.0929078907,0.4217212498,-0.1400011033,0.0661325604,-0.0567793697,0.0003704348,-0.2170345485,0.4287863374,-0.1763373911,0.4928040504,-0.2431970835,0.087903291,-0.3595514297,0.233363688,-0.0096810572,0.0360782631,-0.2453804612,0.4265148938,-0.1822773665,-0.051156316,0.160516575,0.3011433184,0.0987175107,0.6137595177,-0.1845930517,-0.2232504487,0.4984788001,-0.4976068437,-0.424870342,0.0814748928,-0.0286826529,0.1198043823,0.0639730245,-0.56647259,-0.295845747,0.4186934233,-0.0994040743,-0.0072000106,0.088709414,-0.0215018094,0.1368441135,-0.1779536754,0.1924846768,-0.169606626,-0.1580830216,-0.0421707109,-0.2822973728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1872","title":"Adding a new column to the dataset after set_format was called","comments":"Ok cool :) \r\nAlso I just did a PR to mention this behavior in the documentation","body":"Hi, \r\n\r\nthanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side. \r\n\r\nI load some lists of strings and integers, then call `data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)`. This converts the integer columns into tensors, but keeps the lists of strings as they are. I then call `map` to add a new column to my dataset, which is a **list of strings**. Once I iterate through my dataset, I get an error that the new column can't be converted into a tensor (which is probably caused by `set_format`). \r\n\r\nBelow some pseudo code:\r\n```python\r\n def augment_func(sample: Dict) -> Dict:\r\n # do something\r\n return {\r\n \"some_integer_column1\" : augmented_data[\"some_integer_column1\"], # <-- tensor\r\n \"some_integer_column2\" : augmented_data[\"some_integer_column2\"], # <-- tensor\r\n \"NEW_COLUMN\": targets, # <-- list of strings\r\n }\r\n\r\n\r\n data = datasets.load_dataset(__file__, data_dir=\"...\", split=\"train\")\r\n data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n\r\n augmented_dataset = data.map(augment_func, batched=False)\r\n \r\n for sample in augmented_dataset:\r\n print(sample) # fails\r\n\r\n```\r\n\r\nand the exception:\r\n```python\r\nTraceback (most recent call last):\r\n File \"dataset.py\", line 487, in \r\n main()\r\n File \"dataset.py\", line 471, in main\r\n for sample in augmented_dataset:\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 697, in __iter__\r\n yield self._getitem(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1069, in _getitem\r\n outputs = self._convert_outputs(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nThanks!\r\n","comment_length":16,"text":"Adding a new column to the dataset after set_format was called\nHi, \r\n\r\nthanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side. \r\n\r\nI load some lists of strings and integers, then call `data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)`. This converts the integer columns into tensors, but keeps the lists of strings as they are. I then call `map` to add a new column to my dataset, which is a **list of strings**. Once I iterate through my dataset, I get an error that the new column can't be converted into a tensor (which is probably caused by `set_format`). \r\n\r\nBelow some pseudo code:\r\n```python\r\n def augment_func(sample: Dict) -> Dict:\r\n # do something\r\n return {\r\n \"some_integer_column1\" : augmented_data[\"some_integer_column1\"], # <-- tensor\r\n \"some_integer_column2\" : augmented_data[\"some_integer_column2\"], # <-- tensor\r\n \"NEW_COLUMN\": targets, # <-- list of strings\r\n }\r\n\r\n\r\n data = datasets.load_dataset(__file__, data_dir=\"...\", split=\"train\")\r\n data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n\r\n augmented_dataset = data.map(augment_func, batched=False)\r\n \r\n for sample in augmented_dataset:\r\n print(sample) # fails\r\n\r\n```\r\n\r\nand the exception:\r\n```python\r\nTraceback (most recent call last):\r\n File \"dataset.py\", line 487, in \r\n main()\r\n File \"dataset.py\", line 471, in main\r\n for sample in augmented_dataset:\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 697, in __iter__\r\n yield self._getitem(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1069, in _getitem\r\n outputs = self._convert_outputs(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nThanks!\r\n\nOk cool :) \r\nAlso I just did a PR to mention this behavior in the documentation","embeddings":[-0.163837567,-0.1748755127,-0.0401732326,-0.0492338166,0.4729719162,0.2625342906,0.6921063662,0.4177582264,0.16875045,-0.2805859745,0.1421070397,0.3377909064,-0.234039858,-0.0259004012,0.1978132278,-0.1026414856,0.1268425137,0.0638930127,0.2461459339,0.0586879924,-0.3777551651,0.1032360643,-0.1540207863,-0.2058055401,-0.1765669137,-0.3008010983,-0.0024790808,-0.1966911554,-0.1602552682,-0.2089734375,0.1871719062,-0.239169538,-0.0005563378,0.6491594315,-0.0001178686,0.1240975857,0.1696617603,0.031614352,-0.0399997793,-0.1928997934,-0.2241786718,-0.3875743151,-0.1664160639,-0.3292916119,0.1402370483,-0.090086028,-0.4265715182,-0.7763667107,0.2612175345,0.5016534925,0.1658125967,0.1779137254,-0.0243937224,-0.0470924303,0.2402329147,0.2883871496,-0.2155929655,-0.0182730332,-0.0107632391,0.1259519309,0.5092816353,0.1936923265,-0.2842826843,0.1020674407,0.106398195,0.2241102755,-0.2291458696,-0.0425992012,-0.0292203575,0.1993377954,0.6996715069,-0.254180491,-0.3642950058,-0.2796714902,0.0794099048,-0.4402051568,0.119070828,-0.0155838113,-0.0778750554,-0.1121089906,0.0913000554,-0.1839411706,-0.0351652093,0.3412312865,-0.1856348664,0.0189567804,-0.0303626508,0.1920932084,-0.1479987055,-0.0397582315,-0.1474709958,0.1262124032,0.0943790153,0.2419148088,-0.4515655339,-0.1876330823,0.033643987,-0.5349569917,-0.2338584661,-0.1157705784,0.1772542596,0.0031550087,-0.187983036,0.1016870588,0.1837739646,-0.0243844315,-0.2377842367,0.202795282,-0.058971107,-0.2029019147,0.0786278844,-0.1089506671,0.2219047099,-0.2733579278,0.1720267087,0.0052718245,0.415762037,0.1181959957,-0.0770012587,0.2277686745,-0.2703928947,-0.2729817927,0.0384290256,0.0916894153,-0.027375767,0.1280095875,0.140044421,0.2697374821,-0.0306830835,0.1996581256,-0.0518040881,0.0073224418,-0.1836284101,0.0456944257,0.1025165766,0.1123385131,-0.1570651531,0.2218846977,-0.13670744,-0.0665040687,0.0907948315,-0.2595322132,0.3443582356,0.1951227039,-0.1425132006,0.3772317767,0.1710849255,-0.0373196006,-0.1732723415,0.2277605534,-0.2527373135,-0.3213145733,-0.4081810117,0.0944307446,-0.0362261981,-0.1098731384,0.0837554261,0.1692531705,0.5626441836,-0.1914578974,0.1816345602,-0.2311049104,-0.3539713025,-0.2420078516,-0.0327599496,0.1081718057,-0.6691618562,0.0941578373,0.0791116133,0.3833465278,0.0772274733,0.1891558468,-0.0684040338,0.3359542787,-0.2292346805,0.0447732247,0.6186287999,0.0500695035,-0.1436017752,0.065374814,-0.1200989783,0.1486835629,-0.0731995106,0.1894952059,0.0845489725,-0.0850279555,0.3703552485,0.0103158476,-0.1337237805,-0.1136995554,-0.0130396793,0.1656283587,0.5390791893,-0.2064686865,0.0425463095,0.1790866256,-0.1250628978,0.2282993644,-0.017285509,-0.0337756947,-0.0073143095,0.1034121662,0.2751893401,0.0653935894,0.0355755165,-0.2091534138,-0.3796861172,-0.0080606909,0.4330004454,0.021443544,-0.00806686,-0.1484539062,-0.3776355088,0.0075415815,-0.0111386655,0.2967942953,0.0175653826,0.166564554,-0.1642278135,-0.1772773117,-0.338902384,0.0949272737,-0.1961665452,-0.0513706133,-0.2434315234,0.1282278299,0.1284624189,-0.2729171216,-0.1948995143,0.3867044747,0.1357645839,-0.0230477694,-0.1044372097,0.2309470475,0.0756757632,-0.1251296699,-0.4924203455,0.2001788765,0.1332603395,0.1238413602,-0.1664903164,-0.1851825267,-0.0113888839,-0.2167836577,-0.0754593462,0.050736133,0.0049540126,0.2880735099,-0.3090748787,0.1987280101,0.1029938012,0.0727731362,-0.2182666212,-0.3290410638,-0.0560583398,0.0218479354,0.0440452658,-0.2663611472,-0.3668257594,0.0163654499,0.6044391394,0.0032727038,0.1298659891,0.1421494484,0.1310873926,0.0832142532,0.0108531453,0.0890294835,0.4790549278,0.0814551562,-0.0663492233,0.023984842,-0.2592504323,-0.098174952,0.0403772667,0.2914122641,0.2057984024,0.0817342773,0.2835640609,0.0091479477,0.0822448358,0.0117116934,0.2326989472,0.1617716104,-0.5018972754,-0.0692816153,-0.127254352,0.2658206522,-0.2941197157,-0.6862946153,0.0144574651,-0.3176656961,-0.2221376598,0.0013733351,0.0248833746,0.0532939658,-0.0381292999,0.1013478041,0.2265940011,-0.4410072267,-0.0932101756,-0.2546205223,-0.0083848387,0.0140890237,0.3546907604,-0.3220357895,0.4234782755,0.1450334787,-0.1792608052,-0.1153635308,-0.3000035286,-0.0179306921,-0.3884204924,-0.0696965829,0.257756561,0.1516362876,-0.1477588415,-0.1888682097,0.3712129593,0.1980449259,0.2000721395,0.1980653405,0.1245261729,0.0569173545,-0.0044003814,-0.5688668489,-0.1075387448,-0.0420352966,0.0825474933,-0.0878118873,0.0984746888,0.0212098062,0.2226237953,0.1303482354,0.2398899049,-0.0894926563,-0.3793279827,-0.0635955632,0.3975252807,-0.3226929605,-0.2073272765,0.1397510022,-0.0274451841,0.181092605,0.2993002832,-0.3082209527,-0.1508528888,-0.1311350614,0.173268497,-0.2809301317,-0.0994715616,0.5064876676,0.0120290006,-0.0551418066,-0.1404468119,-0.2117807865,0.1956448406,0.2700185776,0.0633992553,0.2628600001,0.547018826,0.088467814,0.3008495867,0.0786023214,-0.4974924028,0.2550722659,-0.30017066,0.1966451108,-0.2971628308,-0.3937258124,-0.0502180122,-0.234202832,-0.0278095137,-0.1490178406,-0.1602451056,-0.0322328135,0.014737457,0.5779079795,-0.1564670354,-0.1584125012,0.3528257012,-0.4837839007,0.3728365898,0.0443864465,0.1514329016,-0.5375984907,-0.1796579808,0.1274439245,-0.0810083896,0.2100242227,-0.1992975175,-0.456897676,-0.1856812686,-0.4243353009,0.2998924851,0.3504037857,0.7206824422,-0.1428639144,-0.0488887243,-0.1048103273,0.1505411416,0.8693706393,-0.5131114125,-0.512216568,0.4151348174,0.2547292411,-0.6216635108,-0.0057002967,-0.1735690981,0.0902562961,0.0768063143,0.6951243877,-0.0104244826,0.1401380748,0.1530417055,0.3333971202,0.0032469353,-0.1188208982,-0.1659090966,-0.1762067527,-0.1543740183,0.0938118175,0.1468466073,0.3185719252,0.3098247945,-0.2787633836,-0.08227963,-0.1979757547,0.2580941916,-0.1493848711,0.2094887495,0.4666827023,0.1809091419,0.0732228234,-0.0207394809,-0.0134861525,0.2105140388,-0.2058895677,-0.0426405296,0.151575014,-0.3012671173,0.3134796023,0.1650590003,-0.0190843772,0.1892253757,-0.2329683304,-0.0420833416,-0.3310903907,0.1621735394,0.5876699686,0.210772112,-0.3579185307,-0.7363798022,0.2013388574,0.0528041907,0.0794652179,0.5957023501,-0.0100870384,-0.1674926728,0.4298977554,0.1278735399,0.8010354042,-0.2321156859,0.0219281632,0.2473872602,-0.1963775903,0.4466971457,-0.2197206169,0.2085921913,-0.1105169877,-0.0548965931,-0.1673021764,-0.2323620021,0.3513605893,0.0045412122,-0.1806060672,-0.1170175001,-0.2128399611,0.6637679935,-0.0228089467,0.105200097,0.148751229,-0.4528320432,-0.3152978718,0.1043530777,-0.1289207488,-0.2021534294,-0.0943826064,0.1820892394,-0.1783920676,-0.208489418,-0.3607927859,-0.0663562045,-0.0690187514,-0.0468821786,0.0452600047,-0.6082686186,0.4602432847,0.3592826128,0.5291901827,0.0879840255,-0.1030571461,-0.0565395281,0.0255032834,0.1595400572,0.2783373892,-0.4173798561,0.3079517186,0.1974584162,-0.1584801227,0.3220860362,0.0993196443,-0.2854143679,-0.1614233255,0.0934315622,0.5090245008,-0.6628907323,-0.3925191462,0.0586135685,0.2449562699,-0.2578346431,0.0919436365,-0.0922592208,0.0138549255,0.376259625,-0.3261744976,-0.0639200583,0.0640054941,0.2678512037,0.0133870263,-0.073792614,0.5867010355,0.2021547109,-0.1515664756,-0.1159019545,0.0242914502,0.1093551666,-0.2628646493,0.0673720762,-0.0161277931,0.1062919125,-0.1352067292,0.0582707115,-0.1605557799,0.2047499418,-0.0174369756,-0.2758021653,-0.1192292348,-0.0141024506,0.275123775,0.2663635314,0.1359473318,0.2398539782,-0.1381954402,0.0437084623,-0.235125497,-0.1114267334,0.1152756885,0.2109810859,0.0536024421,0.056384366,0.153965503,-0.4075026512,0.0870691091,0.4069389105,0.0289605074,-0.2118223459,-0.1445440799,0.1017587483,0.117402859,-0.2523890734,0.1890082806,-0.2974910736,-0.1747690439,-0.2785748243,0.019304812,0.3081176281,-0.2520097494,-0.0051214676,-0.1206487492,0.1459065378,-0.0088552516,-0.0315668806,-0.1970043629,0.0310068186,0.1981580853,0.2192884535,0.2022118717,0.0440381356,-0.3146225512,0.217801109,0.1843610257,0.0563184284,0.2494207919,-0.4320304096,0.0137219429,0.000993419,0.3522247374,-0.0245949402,-0.2163707316,-0.0196516439,0.2225755006,0.1150809228,-0.2363813668,0.0546980016,0.3840655386,0.1135786548,0.1640511006,0.2766043842,0.3523633182,0.1498499066,-0.2969774902,-0.0283383615,0.1019247696,-0.3147266805,0.3177816272,0.636660099,-0.0096956622,0.0432766713,0.2121617943,0.0385080129,-0.0184341837,0.2452431619,0.1546956152,0.1323665231,0.553768754,-0.0233231504,-0.1268625706,-0.3789405227,0.3849281669,0.0655262396,0.1893929541,0.356595099,0.2170319408,0.4540134072,-0.1332535744,-0.0467361659,-0.0272156466,0.0026210472,-0.140909031,0.0477861241,-0.1536695212,0.0255794786,0.0613307096,-0.2604688704,-0.0698648915,-0.2468580753,0.2972401083,-0.0647751167,0.0898903459,0.0116090039,-0.0867135897,-0.1450734586,-0.1023249179,-0.2088188231,0.1428881139,-0.0000564576,-0.0023569872,0.1685059816,0.0797035322,0.3124370277,0.3186620772,0.0319544822,0.3550901115,0.1073564887,-0.01895307,0.2020024359,0.1107800528,-0.1152407229,0.0459671766,0.2320191562,0.1205592304,-0.1050471067,0.1272472143,0.0860555097,0.3713683784,-0.1742448211,0.1659411341,-0.0186983813,-0.2119493484,-0.281200707,-0.0256917533,-0.0536646545,-0.2551568151,0.5833733678,-0.0171415377,0.0827371329,0.1473907977,0.0789523497,0.016913807,0.1726586819,0.1242832616,-0.0562416874,-0.2194082886,-0.1105198041,-0.4495842755,0.0777268559,-0.2444209903,-0.1423133016,0.1331401169,0.1083743572,0.3426305652,-0.0715117231,0.3546713591,-0.4199867249,-0.3278799355,0.2354024798,-0.0737499967,-0.3474462926,-0.1158577353,-0.0218217112,0.0036887003,-0.194622457,0.1039047241,0.3052025735,-0.0131549789,-0.28851372,-0.0521491319,-0.1672354192,-0.0539736487,0.5608680248,0.0929078907,0.4217212498,-0.1400011033,0.0661325604,-0.0567793697,0.0003704348,-0.2170345485,0.4287863374,-0.1763373911,0.4928040504,-0.2431970835,0.087903291,-0.3595514297,0.233363688,-0.0096810572,0.0360782631,-0.2453804612,0.4265148938,-0.1822773665,-0.051156316,0.160516575,0.3011433184,0.0987175107,0.6137595177,-0.1845930517,-0.2232504487,0.4984788001,-0.4976068437,-0.424870342,0.0814748928,-0.0286826529,0.1198043823,0.0639730245,-0.56647259,-0.295845747,0.4186934233,-0.0994040743,-0.0072000106,0.088709414,-0.0215018094,0.1368441135,-0.1779536754,0.1924846768,-0.169606626,-0.1580830216,-0.0421707109,-0.2822973728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1872","title":"Adding a new column to the dataset after set_format was called","comments":"Closed by #1888","body":"Hi, \r\n\r\nthanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side. \r\n\r\nI load some lists of strings and integers, then call `data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)`. This converts the integer columns into tensors, but keeps the lists of strings as they are. I then call `map` to add a new column to my dataset, which is a **list of strings**. Once I iterate through my dataset, I get an error that the new column can't be converted into a tensor (which is probably caused by `set_format`). \r\n\r\nBelow some pseudo code:\r\n```python\r\n def augment_func(sample: Dict) -> Dict:\r\n # do something\r\n return {\r\n \"some_integer_column1\" : augmented_data[\"some_integer_column1\"], # <-- tensor\r\n \"some_integer_column2\" : augmented_data[\"some_integer_column2\"], # <-- tensor\r\n \"NEW_COLUMN\": targets, # <-- list of strings\r\n }\r\n\r\n\r\n data = datasets.load_dataset(__file__, data_dir=\"...\", split=\"train\")\r\n data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n\r\n augmented_dataset = data.map(augment_func, batched=False)\r\n \r\n for sample in augmented_dataset:\r\n print(sample) # fails\r\n\r\n```\r\n\r\nand the exception:\r\n```python\r\nTraceback (most recent call last):\r\n File \"dataset.py\", line 487, in \r\n main()\r\n File \"dataset.py\", line 471, in main\r\n for sample in augmented_dataset:\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 697, in __iter__\r\n yield self._getitem(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1069, in _getitem\r\n outputs = self._convert_outputs(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nThanks!\r\n","comment_length":3,"text":"Adding a new column to the dataset after set_format was called\nHi, \r\n\r\nthanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side. \r\n\r\nI load some lists of strings and integers, then call `data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)`. This converts the integer columns into tensors, but keeps the lists of strings as they are. I then call `map` to add a new column to my dataset, which is a **list of strings**. Once I iterate through my dataset, I get an error that the new column can't be converted into a tensor (which is probably caused by `set_format`). \r\n\r\nBelow some pseudo code:\r\n```python\r\n def augment_func(sample: Dict) -> Dict:\r\n # do something\r\n return {\r\n \"some_integer_column1\" : augmented_data[\"some_integer_column1\"], # <-- tensor\r\n \"some_integer_column2\" : augmented_data[\"some_integer_column2\"], # <-- tensor\r\n \"NEW_COLUMN\": targets, # <-- list of strings\r\n }\r\n\r\n\r\n data = datasets.load_dataset(__file__, data_dir=\"...\", split=\"train\")\r\n data.set_format(\"torch\", columns=[\"some_integer_column1\", \"some_integer_column2\"], output_all_columns=True)\r\n\r\n augmented_dataset = data.map(augment_func, batched=False)\r\n \r\n for sample in augmented_dataset:\r\n print(sample) # fails\r\n\r\n```\r\n\r\nand the exception:\r\n```python\r\nTraceback (most recent call last):\r\n File \"dataset.py\", line 487, in \r\n main()\r\n File \"dataset.py\", line 471, in main\r\n for sample in augmented_dataset:\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 697, in __iter__\r\n yield self._getitem(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 1069, in _getitem\r\n outputs = self._convert_outputs(\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in command\r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 850, in \r\n return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n File \"lib\/python3.8\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"lib\/python3.8\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nThanks!\r\n\nClosed by #1888","embeddings":[-0.163837567,-0.1748755127,-0.0401732326,-0.0492338166,0.4729719162,0.2625342906,0.6921063662,0.4177582264,0.16875045,-0.2805859745,0.1421070397,0.3377909064,-0.234039858,-0.0259004012,0.1978132278,-0.1026414856,0.1268425137,0.0638930127,0.2461459339,0.0586879924,-0.3777551651,0.1032360643,-0.1540207863,-0.2058055401,-0.1765669137,-0.3008010983,-0.0024790808,-0.1966911554,-0.1602552682,-0.2089734375,0.1871719062,-0.239169538,-0.0005563378,0.6491594315,-0.0001178686,0.1240975857,0.1696617603,0.031614352,-0.0399997793,-0.1928997934,-0.2241786718,-0.3875743151,-0.1664160639,-0.3292916119,0.1402370483,-0.090086028,-0.4265715182,-0.7763667107,0.2612175345,0.5016534925,0.1658125967,0.1779137254,-0.0243937224,-0.0470924303,0.2402329147,0.2883871496,-0.2155929655,-0.0182730332,-0.0107632391,0.1259519309,0.5092816353,0.1936923265,-0.2842826843,0.1020674407,0.106398195,0.2241102755,-0.2291458696,-0.0425992012,-0.0292203575,0.1993377954,0.6996715069,-0.254180491,-0.3642950058,-0.2796714902,0.0794099048,-0.4402051568,0.119070828,-0.0155838113,-0.0778750554,-0.1121089906,0.0913000554,-0.1839411706,-0.0351652093,0.3412312865,-0.1856348664,0.0189567804,-0.0303626508,0.1920932084,-0.1479987055,-0.0397582315,-0.1474709958,0.1262124032,0.0943790153,0.2419148088,-0.4515655339,-0.1876330823,0.033643987,-0.5349569917,-0.2338584661,-0.1157705784,0.1772542596,0.0031550087,-0.187983036,0.1016870588,0.1837739646,-0.0243844315,-0.2377842367,0.202795282,-0.058971107,-0.2029019147,0.0786278844,-0.1089506671,0.2219047099,-0.2733579278,0.1720267087,0.0052718245,0.415762037,0.1181959957,-0.0770012587,0.2277686745,-0.2703928947,-0.2729817927,0.0384290256,0.0916894153,-0.027375767,0.1280095875,0.140044421,0.2697374821,-0.0306830835,0.1996581256,-0.0518040881,0.0073224418,-0.1836284101,0.0456944257,0.1025165766,0.1123385131,-0.1570651531,0.2218846977,-0.13670744,-0.0665040687,0.0907948315,-0.2595322132,0.3443582356,0.1951227039,-0.1425132006,0.3772317767,0.1710849255,-0.0373196006,-0.1732723415,0.2277605534,-0.2527373135,-0.3213145733,-0.4081810117,0.0944307446,-0.0362261981,-0.1098731384,0.0837554261,0.1692531705,0.5626441836,-0.1914578974,0.1816345602,-0.2311049104,-0.3539713025,-0.2420078516,-0.0327599496,0.1081718057,-0.6691618562,0.0941578373,0.0791116133,0.3833465278,0.0772274733,0.1891558468,-0.0684040338,0.3359542787,-0.2292346805,0.0447732247,0.6186287999,0.0500695035,-0.1436017752,0.065374814,-0.1200989783,0.1486835629,-0.0731995106,0.1894952059,0.0845489725,-0.0850279555,0.3703552485,0.0103158476,-0.1337237805,-0.1136995554,-0.0130396793,0.1656283587,0.5390791893,-0.2064686865,0.0425463095,0.1790866256,-0.1250628978,0.2282993644,-0.017285509,-0.0337756947,-0.0073143095,0.1034121662,0.2751893401,0.0653935894,0.0355755165,-0.2091534138,-0.3796861172,-0.0080606909,0.4330004454,0.021443544,-0.00806686,-0.1484539062,-0.3776355088,0.0075415815,-0.0111386655,0.2967942953,0.0175653826,0.166564554,-0.1642278135,-0.1772773117,-0.338902384,0.0949272737,-0.1961665452,-0.0513706133,-0.2434315234,0.1282278299,0.1284624189,-0.2729171216,-0.1948995143,0.3867044747,0.1357645839,-0.0230477694,-0.1044372097,0.2309470475,0.0756757632,-0.1251296699,-0.4924203455,0.2001788765,0.1332603395,0.1238413602,-0.1664903164,-0.1851825267,-0.0113888839,-0.2167836577,-0.0754593462,0.050736133,0.0049540126,0.2880735099,-0.3090748787,0.1987280101,0.1029938012,0.0727731362,-0.2182666212,-0.3290410638,-0.0560583398,0.0218479354,0.0440452658,-0.2663611472,-0.3668257594,0.0163654499,0.6044391394,0.0032727038,0.1298659891,0.1421494484,0.1310873926,0.0832142532,0.0108531453,0.0890294835,0.4790549278,0.0814551562,-0.0663492233,0.023984842,-0.2592504323,-0.098174952,0.0403772667,0.2914122641,0.2057984024,0.0817342773,0.2835640609,0.0091479477,0.0822448358,0.0117116934,0.2326989472,0.1617716104,-0.5018972754,-0.0692816153,-0.127254352,0.2658206522,-0.2941197157,-0.6862946153,0.0144574651,-0.3176656961,-0.2221376598,0.0013733351,0.0248833746,0.0532939658,-0.0381292999,0.1013478041,0.2265940011,-0.4410072267,-0.0932101756,-0.2546205223,-0.0083848387,0.0140890237,0.3546907604,-0.3220357895,0.4234782755,0.1450334787,-0.1792608052,-0.1153635308,-0.3000035286,-0.0179306921,-0.3884204924,-0.0696965829,0.257756561,0.1516362876,-0.1477588415,-0.1888682097,0.3712129593,0.1980449259,0.2000721395,0.1980653405,0.1245261729,0.0569173545,-0.0044003814,-0.5688668489,-0.1075387448,-0.0420352966,0.0825474933,-0.0878118873,0.0984746888,0.0212098062,0.2226237953,0.1303482354,0.2398899049,-0.0894926563,-0.3793279827,-0.0635955632,0.3975252807,-0.3226929605,-0.2073272765,0.1397510022,-0.0274451841,0.181092605,0.2993002832,-0.3082209527,-0.1508528888,-0.1311350614,0.173268497,-0.2809301317,-0.0994715616,0.5064876676,0.0120290006,-0.0551418066,-0.1404468119,-0.2117807865,0.1956448406,0.2700185776,0.0633992553,0.2628600001,0.547018826,0.088467814,0.3008495867,0.0786023214,-0.4974924028,0.2550722659,-0.30017066,0.1966451108,-0.2971628308,-0.3937258124,-0.0502180122,-0.234202832,-0.0278095137,-0.1490178406,-0.1602451056,-0.0322328135,0.014737457,0.5779079795,-0.1564670354,-0.1584125012,0.3528257012,-0.4837839007,0.3728365898,0.0443864465,0.1514329016,-0.5375984907,-0.1796579808,0.1274439245,-0.0810083896,0.2100242227,-0.1992975175,-0.456897676,-0.1856812686,-0.4243353009,0.2998924851,0.3504037857,0.7206824422,-0.1428639144,-0.0488887243,-0.1048103273,0.1505411416,0.8693706393,-0.5131114125,-0.512216568,0.4151348174,0.2547292411,-0.6216635108,-0.0057002967,-0.1735690981,0.0902562961,0.0768063143,0.6951243877,-0.0104244826,0.1401380748,0.1530417055,0.3333971202,0.0032469353,-0.1188208982,-0.1659090966,-0.1762067527,-0.1543740183,0.0938118175,0.1468466073,0.3185719252,0.3098247945,-0.2787633836,-0.08227963,-0.1979757547,0.2580941916,-0.1493848711,0.2094887495,0.4666827023,0.1809091419,0.0732228234,-0.0207394809,-0.0134861525,0.2105140388,-0.2058895677,-0.0426405296,0.151575014,-0.3012671173,0.3134796023,0.1650590003,-0.0190843772,0.1892253757,-0.2329683304,-0.0420833416,-0.3310903907,0.1621735394,0.5876699686,0.210772112,-0.3579185307,-0.7363798022,0.2013388574,0.0528041907,0.0794652179,0.5957023501,-0.0100870384,-0.1674926728,0.4298977554,0.1278735399,0.8010354042,-0.2321156859,0.0219281632,0.2473872602,-0.1963775903,0.4466971457,-0.2197206169,0.2085921913,-0.1105169877,-0.0548965931,-0.1673021764,-0.2323620021,0.3513605893,0.0045412122,-0.1806060672,-0.1170175001,-0.2128399611,0.6637679935,-0.0228089467,0.105200097,0.148751229,-0.4528320432,-0.3152978718,0.1043530777,-0.1289207488,-0.2021534294,-0.0943826064,0.1820892394,-0.1783920676,-0.208489418,-0.3607927859,-0.0663562045,-0.0690187514,-0.0468821786,0.0452600047,-0.6082686186,0.4602432847,0.3592826128,0.5291901827,0.0879840255,-0.1030571461,-0.0565395281,0.0255032834,0.1595400572,0.2783373892,-0.4173798561,0.3079517186,0.1974584162,-0.1584801227,0.3220860362,0.0993196443,-0.2854143679,-0.1614233255,0.0934315622,0.5090245008,-0.6628907323,-0.3925191462,0.0586135685,0.2449562699,-0.2578346431,0.0919436365,-0.0922592208,0.0138549255,0.376259625,-0.3261744976,-0.0639200583,0.0640054941,0.2678512037,0.0133870263,-0.073792614,0.5867010355,0.2021547109,-0.1515664756,-0.1159019545,0.0242914502,0.1093551666,-0.2628646493,0.0673720762,-0.0161277931,0.1062919125,-0.1352067292,0.0582707115,-0.1605557799,0.2047499418,-0.0174369756,-0.2758021653,-0.1192292348,-0.0141024506,0.275123775,0.2663635314,0.1359473318,0.2398539782,-0.1381954402,0.0437084623,-0.235125497,-0.1114267334,0.1152756885,0.2109810859,0.0536024421,0.056384366,0.153965503,-0.4075026512,0.0870691091,0.4069389105,0.0289605074,-0.2118223459,-0.1445440799,0.1017587483,0.117402859,-0.2523890734,0.1890082806,-0.2974910736,-0.1747690439,-0.2785748243,0.019304812,0.3081176281,-0.2520097494,-0.0051214676,-0.1206487492,0.1459065378,-0.0088552516,-0.0315668806,-0.1970043629,0.0310068186,0.1981580853,0.2192884535,0.2022118717,0.0440381356,-0.3146225512,0.217801109,0.1843610257,0.0563184284,0.2494207919,-0.4320304096,0.0137219429,0.000993419,0.3522247374,-0.0245949402,-0.2163707316,-0.0196516439,0.2225755006,0.1150809228,-0.2363813668,0.0546980016,0.3840655386,0.1135786548,0.1640511006,0.2766043842,0.3523633182,0.1498499066,-0.2969774902,-0.0283383615,0.1019247696,-0.3147266805,0.3177816272,0.636660099,-0.0096956622,0.0432766713,0.2121617943,0.0385080129,-0.0184341837,0.2452431619,0.1546956152,0.1323665231,0.553768754,-0.0233231504,-0.1268625706,-0.3789405227,0.3849281669,0.0655262396,0.1893929541,0.356595099,0.2170319408,0.4540134072,-0.1332535744,-0.0467361659,-0.0272156466,0.0026210472,-0.140909031,0.0477861241,-0.1536695212,0.0255794786,0.0613307096,-0.2604688704,-0.0698648915,-0.2468580753,0.2972401083,-0.0647751167,0.0898903459,0.0116090039,-0.0867135897,-0.1450734586,-0.1023249179,-0.2088188231,0.1428881139,-0.0000564576,-0.0023569872,0.1685059816,0.0797035322,0.3124370277,0.3186620772,0.0319544822,0.3550901115,0.1073564887,-0.01895307,0.2020024359,0.1107800528,-0.1152407229,0.0459671766,0.2320191562,0.1205592304,-0.1050471067,0.1272472143,0.0860555097,0.3713683784,-0.1742448211,0.1659411341,-0.0186983813,-0.2119493484,-0.281200707,-0.0256917533,-0.0536646545,-0.2551568151,0.5833733678,-0.0171415377,0.0827371329,0.1473907977,0.0789523497,0.016913807,0.1726586819,0.1242832616,-0.0562416874,-0.2194082886,-0.1105198041,-0.4495842755,0.0777268559,-0.2444209903,-0.1423133016,0.1331401169,0.1083743572,0.3426305652,-0.0715117231,0.3546713591,-0.4199867249,-0.3278799355,0.2354024798,-0.0737499967,-0.3474462926,-0.1158577353,-0.0218217112,0.0036887003,-0.194622457,0.1039047241,0.3052025735,-0.0131549789,-0.28851372,-0.0521491319,-0.1672354192,-0.0539736487,0.5608680248,0.0929078907,0.4217212498,-0.1400011033,0.0661325604,-0.0567793697,0.0003704348,-0.2170345485,0.4287863374,-0.1763373911,0.4928040504,-0.2431970835,0.087903291,-0.3595514297,0.233363688,-0.0096810572,0.0360782631,-0.2453804612,0.4265148938,-0.1822773665,-0.051156316,0.160516575,0.3011433184,0.0987175107,0.6137595177,-0.1845930517,-0.2232504487,0.4984788001,-0.4976068437,-0.424870342,0.0814748928,-0.0286826529,0.1198043823,0.0639730245,-0.56647259,-0.295845747,0.4186934233,-0.0994040743,-0.0072000106,0.088709414,-0.0215018094,0.1368441135,-0.1779536754,0.1924846768,-0.169606626,-0.1580830216,-0.0421707109,-0.2822973728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1867","title":"ERROR WHEN USING SET_TRANSFORM() ","comments":"Hi @alejandrocros it looks like an incompatibility with the current Trainer @sgugger \r\nIndeed currently the Trainer of `transformers` doesn't support a dataset with a transform\r\n\r\nIt looks like it comes from this line: https:\/\/github.com\/huggingface\/transformers\/blob\/f51188cbe74195c14c5b3e2e8f10c2f435f9751a\/src\/transformers\/trainer.py#L442\r\n\r\nThis line sets the format to not return certain unused columns. But this has two issues:\r\n1. it forgets to also set the format_kwargs (this causes the error you got):\r\n```python\r\ndataset.set_format(type=dataset.format[\"type\"], columns=columns, format_kwargs=dataset.format[\"format_kwargs\"])\r\n```\r\n2. the Trainer wants to keep only the fields that are used as input for a model. However for a dataset with a transform, the output fields are often different from the columns fields. For example from a column \"text\" in the dataset, the strings can be transformed on-the-fly into \"input_ids\". If you want your dataset to only output certain fields and not other you must change your transform function.\r\n","body":"Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n","comment_length":139,"text":"ERROR WHEN USING SET_TRANSFORM() \nHi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n\nHi @alejandrocros it looks like an incompatibility with the current Trainer @sgugger \r\nIndeed currently the Trainer of `transformers` doesn't support a dataset with a transform\r\n\r\nIt looks like it comes from this line: https:\/\/github.com\/huggingface\/transformers\/blob\/f51188cbe74195c14c5b3e2e8f10c2f435f9751a\/src\/transformers\/trainer.py#L442\r\n\r\nThis line sets the format to not return certain unused columns. But this has two issues:\r\n1. it forgets to also set the format_kwargs (this causes the error you got):\r\n```python\r\ndataset.set_format(type=dataset.format[\"type\"], columns=columns, format_kwargs=dataset.format[\"format_kwargs\"])\r\n```\r\n2. the Trainer wants to keep only the fields that are used as input for a model. However for a dataset with a transform, the output fields are often different from the columns fields. For example from a column \"text\" in the dataset, the strings can be transformed on-the-fly into \"input_ids\". If you want your dataset to only output certain fields and not other you must change your transform function.\r\n","embeddings":[-0.1826189756,-0.2732517719,0.1393030435,0.1847398877,0.7292689681,0.1197050363,0.6009586453,0.2287764698,-0.3227846622,0.0544771031,0.2355112433,0.0333268568,-0.0883406326,-0.0234047007,0.1892216057,-0.5358201861,-0.0138219129,0.1288610101,-0.4634343982,-0.1831161827,-0.27669698,0.3134228885,-0.3415367901,0.1193406507,-0.3186003268,-0.0096082669,-0.047695376,0.016471222,0.0844151899,-0.1718301773,0.2334779501,-0.2800752223,0.1231957451,0.3974503875,-0.0001255372,0.1195578203,0.0606352575,-0.2890718281,-0.2367867827,-0.3003558218,-0.1437315047,-0.1552786827,-0.0154345548,-0.2408748716,-0.4479370713,-0.1890559047,0.0066991188,-0.2687551975,0.5653699636,0.7264817953,0.1215873957,0.2663597167,0.0996964872,-0.2382416725,-0.1261355132,0.1903086156,-0.1291115582,-0.0224978216,0.0317756385,0.2165266722,0.1336240321,0.1615502834,0.2006310076,-0.1752620935,0.4064543247,-0.0619789995,-0.1365289241,0.0435062647,0.0006193888,0.0104680918,0.2298866063,-0.3454308212,-0.420854032,-0.2735254467,0.1654442102,-0.2141008228,0.2700577974,0.1503871977,-0.066397734,0.22794348,-0.2600681484,-0.1098380312,-0.0588664711,0.0313304402,-0.1728429347,0.1623857617,-0.0065073655,0.1303315759,0.0931790695,0.1813862473,0.2739946544,0.1153549775,0.1876621246,0.3649872243,-0.265356034,-0.05478606,-0.1434788406,0.0108277686,-0.2889722586,-0.2492735982,-0.1723645478,-0.2085095048,0.0230882131,-0.0149517776,0.1071462706,0.6613349319,-0.2454049438,0.4977247119,0.0151178716,-0.1120734289,0.2516205609,-0.0635674,-0.0569484457,-0.3780822754,-0.0096246395,0.4175055921,0.0937968418,-0.006041795,-0.2874678373,-0.1932771057,-0.2089736909,-0.0578821562,0.1327706873,0.217985943,0.0588070415,0.2713345587,0.5032314658,0.3485563695,-0.1059520543,-0.2050584853,-0.1715610176,0.135013476,-0.3857185841,-0.1006987318,0.1464199424,-0.1412298381,0.3286191225,0.0532120019,0.0755415633,0.3815326989,0.1520244032,-0.191182673,0.2819835246,0.1436703354,-0.0143271564,0.2014978528,0.4693849087,0.10144265,-0.1861768216,0.2664014995,-0.2484502494,-0.4411483705,0.0609503128,0.0914024562,-0.0530007631,-0.0377959199,-0.1594837457,0.1583796293,0.4205289483,-0.0675125122,0.0763092414,-0.4352326989,0.0418038815,0.1099921837,0.090025045,0.0778380185,-0.2328578383,-0.3426035345,0.3258585334,0.1530645639,-0.143343091,0.2205605805,-0.4285855889,0.2562271655,-0.0625739917,-0.1584211737,0.2098882198,-0.3840966821,-0.0710864812,-0.0994279832,-0.1054709256,-0.219163999,0.0696848631,-0.1317368597,-0.0530189537,-0.1544256955,-0.2207259983,0.5086032748,-0.0608968958,0.2324942946,0.0340902545,-0.0222444125,0.1885959953,0.1376521885,0.0248874463,0.1576400846,-0.320507586,0.5120714307,0.0987583771,-0.2441777289,0.2444199026,0.4102393985,-0.244097501,-0.1919647157,0.0468020588,0.2764771283,-0.0697960332,-0.148902759,0.1784002781,0.1963512599,-0.0604719073,0.0484519824,-0.2369389385,-0.0641728118,-0.2133747637,-0.1585211456,-0.0149899647,0.1101620644,0.0207734834,-0.0464728251,-0.2484817207,0.2723915279,-0.095426932,0.1349620223,-0.5259901881,0.2280209661,-0.2145735472,-0.0914447233,-0.0678843781,0.2231209576,0.2404938191,-0.2305194438,-0.3247486055,0.3036762774,-0.2122558355,0.0371924043,-0.3943189681,0.2457401305,0.3090952635,-0.5342994928,0.0032722938,0.2634411156,0.1639743596,-0.127455622,-0.383022517,0.1136109233,0.0757724941,0.3314516544,0.1057743058,0.3340388238,0.0536119379,-0.0884053633,-0.2020374686,-0.0776441023,-0.0154921887,-0.0838234276,0.1649821252,0.0597501025,-0.2288423777,-0.1983881891,0.4673170149,0.148841694,0.0216302704,0.1293901354,-0.1397807598,0.0306708906,0.1618632525,-0.0950354412,0.562422514,0.0324369743,-0.2305446267,0.1875147969,0.0697727501,-0.1226138026,0.4002860487,-0.0731957108,0.0252318047,-0.106548816,0.0615078025,0.0952067301,-0.1649064571,-0.289491564,0.25625211,0.2521972656,-0.4736778736,0.2136499882,0.0259740539,0.1696037799,-0.5800776482,-0.3827403784,0.0345899239,-0.2872623503,-0.1098969206,-0.0208649188,0.1248787344,0.4439495802,0.1025984958,0.072272338,0.2448716611,-0.158022657,-0.0618830435,0.0459858961,-0.1911002547,0.0143575687,-0.1871763021,-0.2879667282,-0.0504698642,-0.3237556815,-0.0552909076,-0.1035404056,-0.1793443263,0.1855938435,-0.2955768704,0.0113008786,0.3231825829,0.1871942282,-0.1992335618,-0.3188306391,0.4316528141,0.0009160679,0.0959324092,-0.0127255181,-0.0703296512,-0.2425712794,-0.0997440368,-0.4264292717,-0.0446449108,-0.1689539254,0.0303072091,-0.2177828997,-0.1313049346,0.4203010499,0.2371348441,0.0843030065,0.2997512817,0.1740505844,-0.0198321659,-0.2111925334,0.4776291847,-0.2067952901,-0.3576395214,-0.1271426529,-0.0129219079,0.2218758017,0.2340473831,-0.1764791906,-0.178256005,-0.2040158212,-0.1126637682,-0.3014701605,0.435983181,0.4326145947,0.0603171624,0.1686085016,-0.050865747,-0.5156285167,0.3171764314,0.1424028128,0.4169045985,-0.0929954201,0.4798959196,0.0866109133,0.7751232386,0.0111009004,-0.3219003677,0.260661304,-0.1630743444,0.2260283828,-0.1403172314,-0.4625391364,0.2768841088,0.2438232154,0.0316104107,-0.0728213787,-0.0910808519,-0.0698518232,-0.1445435584,0.2413931191,-0.1052515581,-0.152235657,0.141644001,-0.1808562279,0.5583443642,0.0370699987,0.3895744383,-0.2105321437,0.0093455995,0.0729864463,-0.1095194891,0.0688644648,0.1238157228,-0.1383382678,-0.5897532701,-0.1902554631,0.1065888256,-0.0117324274,0.3856185079,-0.1194277927,-0.0214349553,0.1513940543,0.1890711784,0.7310069203,0.0729371309,-0.3278804123,-0.0309844855,-0.2211033851,-0.6239109635,-0.0709218159,0.0394208506,-0.1653980464,-0.1467238814,0.1922552288,-0.1503565162,-0.1719058305,0.3554315269,0.0347531512,-0.2583989501,0.0776256397,-0.0642109886,-0.523164928,-0.3752950728,0.0817698017,0.2867938876,0.1763403565,0.2706166506,0.3705551624,-0.025202468,-0.0146308187,0.18252334,0.0229771715,0.2532839775,0.1200980097,-0.0711414665,0.2466866821,0.2663377821,0.0747498721,0.3784388602,0.0218263883,-0.1386081427,-0.0059653665,-0.0273004491,0.3917846084,0.3158438504,-0.1038671583,-0.2018494606,0.000328526,0.0675849691,-0.0518511422,0.2731793523,0.4600665271,0.0707793012,-0.1523297429,-0.5686461329,0.1935224235,0.1283531487,0.1165830716,0.0412615761,0.1834687442,-0.4985372424,0.3000656366,0.2441835999,0.9627013803,0.0045670066,0.4122143984,0.4066116214,-0.1718395501,-0.0448778979,-0.113391012,0.1925232708,-0.2689211369,-0.3740708232,-0.1267440766,-0.1576801091,0.1537520885,0.1447482109,-0.4471739829,0.2574018836,-0.1016518995,0.1324464083,-0.0517069362,-0.0063035074,-0.1191574186,-0.3786025941,-0.4207054675,0.0605765507,0.0085383756,0.2239290327,-0.021146439,-0.0518368334,-0.0725239962,-0.3276558518,-0.1989789903,0.1438138634,-0.3425531387,0.1162754223,0.0215491783,-0.3508206308,0.5728157759,0.476177901,0.5377926826,0.0480021685,-0.0448533259,0.2743342817,-0.2238800377,0.2336572111,0.0026795755,0.1144546121,0.4311787486,-0.0262231417,0.0106902737,0.2838855386,-0.1292874068,-0.0491685048,-0.1847702265,-0.2700214386,0.0675761774,-0.372892797,-0.2156337798,0.1479793638,-0.0138652613,-0.1213464215,0.0629487187,-0.1453512013,-0.2010597736,0.1748773158,-0.1926721185,-0.3175874949,-0.0471323766,0.5520834327,-0.0154436249,-0.0259577204,0.9083886147,0.144269377,0.0290533174,-0.0506250225,0.0027593984,0.5083619952,-0.7060852647,0.1902077049,0.0442791618,-0.1163032502,-0.2586606443,0.2225267291,0.0910592005,-0.1223974228,0.106648542,0.0268161763,-0.3815151155,0.0435700901,0.1826019138,0.1099413484,-0.006349633,0.4157488346,0.3691781163,-0.160919562,-0.2136099041,0.0854345262,0.0152923148,0.2436425686,0.1348816752,-0.1812804937,0.1046393886,-0.2158272713,0.0344824716,-0.0117325801,0.0288340431,-0.1036318764,-0.303047955,0.0960220471,0.1902741194,-0.1844199449,-0.0305881146,-0.3210200965,-0.2273693979,-0.2802014351,0.148508355,0.2875715792,-0.0231884699,0.4670304656,-0.0447168574,-0.1073348075,-0.1710858345,-0.3624248505,-0.055877734,-0.2588469684,0.2030044794,0.1647436172,0.0901968554,-0.0337181799,-0.5971805453,0.2267414331,0.2485940754,0.2303075343,0.0678243041,-0.2015693933,0.1524341106,0.1655961126,0.2326402515,-0.0341038108,-0.2902600765,0.0206330046,0.1774314642,0.0617807992,-0.2854016721,0.3496537805,-0.0410753638,0.0037135417,0.0104907202,0.2545004785,0.0580310784,-0.0115493983,0.0494867526,0.07942947,0.2100776881,-0.060544882,0.3345696926,0.7754502892,-0.2804121077,0.2012109309,0.2750228941,0.1458674669,0.2543017566,0.3205238581,-0.1166608557,0.344928503,-0.0188068897,-0.0199496411,-0.0652343482,-0.1228847727,-0.3104800284,-0.1817116886,0.0402105004,0.0242764875,0.0616917051,0.5748892426,-0.2628111541,-0.3483905494,-0.2398824245,0.27559039,-0.1278471202,-0.0154439704,-0.0172689799,0.0287494026,-0.3195879459,-0.1499734968,-0.2125629634,-0.1377667636,0.6361117959,0.0198255535,-0.0023895337,-0.1892204583,-0.0454767235,0.2628235817,0.0819954276,-0.1269645691,0.2411371469,0.1011123285,-0.1883557737,0.0976076052,0.4976045489,0.2201207876,0.2194078565,-0.0463458151,-0.0102585051,-0.0782288536,-0.07122951,0.1536559314,0.2663584054,-0.1075949073,0.3812067509,0.2045420259,0.0881867185,-0.0691197589,0.1516282856,-0.2430780083,0.311342746,-0.2424247712,0.5272917151,-0.0208063293,-0.1454467177,-0.1479764134,0.0642576069,-0.1720504463,0.1480787694,0.3423065841,0.007757517,0.1103841811,-0.1328207552,0.0115130777,0.2430801243,0.2766731679,0.5207222104,-0.2535755932,-0.3213942349,-0.2156719267,-0.3150944412,-0.0384822972,0.1294478774,0.0174163729,-0.0245442092,-0.1764499247,0.3041365147,-0.0779176876,0.0809708983,-0.1065159515,-0.1017209142,0.211242795,-0.0689955726,-0.3980531991,-0.1088816151,0.3790934682,-0.0227049384,-0.0859454423,-0.1128532961,0.4846835732,-0.0293582678,-0.369301945,0.0441125371,-0.1199799553,-0.1286889464,0.2373430282,0.0764206648,0.2086417675,0.0588344671,0.2778173089,0.0176058225,-0.2506176531,-0.3266896904,0.0686159506,-0.0857321024,0.4292593896,-0.2401910871,-0.1764764786,-0.4258554578,0.0071116677,0.1781212837,0.0464417078,-0.4590160847,0.4230368733,-0.367138505,0.0823734626,0.2416835129,0.3583958447,0.0034974401,0.1476848125,-0.3863488436,-0.4244388938,0.6188313961,-0.1814698875,-0.4834167063,0.1206571609,-0.1362357289,-0.0846379399,-0.1597628593,-0.5339651108,0.0789652839,0.1070881411,-0.2078465074,-0.3984488547,0.2419902831,0.1348467618,0.0488820933,-0.0512382947,0.4404880404,-0.0221921708,-0.202132076,-0.1810315847,-0.0351451151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1867","title":"ERROR WHEN USING SET_TRANSFORM() ","comments":"FYI that option can be removed with `remove_unused_columns = False` in your `TrainingArguments`, so there is a workaround @alexvaca0 while the fix in `Trainer` is underway.\r\n\r\n@lhoestq I think I will just use the line you suggested and if someone is using the columns that are removed in their transform they will need to change `remove_unused_columns` to `False`. We might switch the default of that argument in the next version if that proves too bug-proof.","body":"Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n","comment_length":75,"text":"ERROR WHEN USING SET_TRANSFORM() \nHi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n\nFYI that option can be removed with `remove_unused_columns = False` in your `TrainingArguments`, so there is a workaround @alexvaca0 while the fix in `Trainer` is underway.\r\n\r\n@lhoestq I think I will just use the line you suggested and if someone is using the columns that are removed in their transform they will need to change `remove_unused_columns` to `False`. We might switch the default of that argument in the next version if that proves too bug-proof.","embeddings":[-0.1826189756,-0.2732517719,0.1393030435,0.1847398877,0.7292689681,0.1197050363,0.6009586453,0.2287764698,-0.3227846622,0.0544771031,0.2355112433,0.0333268568,-0.0883406326,-0.0234047007,0.1892216057,-0.5358201861,-0.0138219129,0.1288610101,-0.4634343982,-0.1831161827,-0.27669698,0.3134228885,-0.3415367901,0.1193406507,-0.3186003268,-0.0096082669,-0.047695376,0.016471222,0.0844151899,-0.1718301773,0.2334779501,-0.2800752223,0.1231957451,0.3974503875,-0.0001255372,0.1195578203,0.0606352575,-0.2890718281,-0.2367867827,-0.3003558218,-0.1437315047,-0.1552786827,-0.0154345548,-0.2408748716,-0.4479370713,-0.1890559047,0.0066991188,-0.2687551975,0.5653699636,0.7264817953,0.1215873957,0.2663597167,0.0996964872,-0.2382416725,-0.1261355132,0.1903086156,-0.1291115582,-0.0224978216,0.0317756385,0.2165266722,0.1336240321,0.1615502834,0.2006310076,-0.1752620935,0.4064543247,-0.0619789995,-0.1365289241,0.0435062647,0.0006193888,0.0104680918,0.2298866063,-0.3454308212,-0.420854032,-0.2735254467,0.1654442102,-0.2141008228,0.2700577974,0.1503871977,-0.066397734,0.22794348,-0.2600681484,-0.1098380312,-0.0588664711,0.0313304402,-0.1728429347,0.1623857617,-0.0065073655,0.1303315759,0.0931790695,0.1813862473,0.2739946544,0.1153549775,0.1876621246,0.3649872243,-0.265356034,-0.05478606,-0.1434788406,0.0108277686,-0.2889722586,-0.2492735982,-0.1723645478,-0.2085095048,0.0230882131,-0.0149517776,0.1071462706,0.6613349319,-0.2454049438,0.4977247119,0.0151178716,-0.1120734289,0.2516205609,-0.0635674,-0.0569484457,-0.3780822754,-0.0096246395,0.4175055921,0.0937968418,-0.006041795,-0.2874678373,-0.1932771057,-0.2089736909,-0.0578821562,0.1327706873,0.217985943,0.0588070415,0.2713345587,0.5032314658,0.3485563695,-0.1059520543,-0.2050584853,-0.1715610176,0.135013476,-0.3857185841,-0.1006987318,0.1464199424,-0.1412298381,0.3286191225,0.0532120019,0.0755415633,0.3815326989,0.1520244032,-0.191182673,0.2819835246,0.1436703354,-0.0143271564,0.2014978528,0.4693849087,0.10144265,-0.1861768216,0.2664014995,-0.2484502494,-0.4411483705,0.0609503128,0.0914024562,-0.0530007631,-0.0377959199,-0.1594837457,0.1583796293,0.4205289483,-0.0675125122,0.0763092414,-0.4352326989,0.0418038815,0.1099921837,0.090025045,0.0778380185,-0.2328578383,-0.3426035345,0.3258585334,0.1530645639,-0.143343091,0.2205605805,-0.4285855889,0.2562271655,-0.0625739917,-0.1584211737,0.2098882198,-0.3840966821,-0.0710864812,-0.0994279832,-0.1054709256,-0.219163999,0.0696848631,-0.1317368597,-0.0530189537,-0.1544256955,-0.2207259983,0.5086032748,-0.0608968958,0.2324942946,0.0340902545,-0.0222444125,0.1885959953,0.1376521885,0.0248874463,0.1576400846,-0.320507586,0.5120714307,0.0987583771,-0.2441777289,0.2444199026,0.4102393985,-0.244097501,-0.1919647157,0.0468020588,0.2764771283,-0.0697960332,-0.148902759,0.1784002781,0.1963512599,-0.0604719073,0.0484519824,-0.2369389385,-0.0641728118,-0.2133747637,-0.1585211456,-0.0149899647,0.1101620644,0.0207734834,-0.0464728251,-0.2484817207,0.2723915279,-0.095426932,0.1349620223,-0.5259901881,0.2280209661,-0.2145735472,-0.0914447233,-0.0678843781,0.2231209576,0.2404938191,-0.2305194438,-0.3247486055,0.3036762774,-0.2122558355,0.0371924043,-0.3943189681,0.2457401305,0.3090952635,-0.5342994928,0.0032722938,0.2634411156,0.1639743596,-0.127455622,-0.383022517,0.1136109233,0.0757724941,0.3314516544,0.1057743058,0.3340388238,0.0536119379,-0.0884053633,-0.2020374686,-0.0776441023,-0.0154921887,-0.0838234276,0.1649821252,0.0597501025,-0.2288423777,-0.1983881891,0.4673170149,0.148841694,0.0216302704,0.1293901354,-0.1397807598,0.0306708906,0.1618632525,-0.0950354412,0.562422514,0.0324369743,-0.2305446267,0.1875147969,0.0697727501,-0.1226138026,0.4002860487,-0.0731957108,0.0252318047,-0.106548816,0.0615078025,0.0952067301,-0.1649064571,-0.289491564,0.25625211,0.2521972656,-0.4736778736,0.2136499882,0.0259740539,0.1696037799,-0.5800776482,-0.3827403784,0.0345899239,-0.2872623503,-0.1098969206,-0.0208649188,0.1248787344,0.4439495802,0.1025984958,0.072272338,0.2448716611,-0.158022657,-0.0618830435,0.0459858961,-0.1911002547,0.0143575687,-0.1871763021,-0.2879667282,-0.0504698642,-0.3237556815,-0.0552909076,-0.1035404056,-0.1793443263,0.1855938435,-0.2955768704,0.0113008786,0.3231825829,0.1871942282,-0.1992335618,-0.3188306391,0.4316528141,0.0009160679,0.0959324092,-0.0127255181,-0.0703296512,-0.2425712794,-0.0997440368,-0.4264292717,-0.0446449108,-0.1689539254,0.0303072091,-0.2177828997,-0.1313049346,0.4203010499,0.2371348441,0.0843030065,0.2997512817,0.1740505844,-0.0198321659,-0.2111925334,0.4776291847,-0.2067952901,-0.3576395214,-0.1271426529,-0.0129219079,0.2218758017,0.2340473831,-0.1764791906,-0.178256005,-0.2040158212,-0.1126637682,-0.3014701605,0.435983181,0.4326145947,0.0603171624,0.1686085016,-0.050865747,-0.5156285167,0.3171764314,0.1424028128,0.4169045985,-0.0929954201,0.4798959196,0.0866109133,0.7751232386,0.0111009004,-0.3219003677,0.260661304,-0.1630743444,0.2260283828,-0.1403172314,-0.4625391364,0.2768841088,0.2438232154,0.0316104107,-0.0728213787,-0.0910808519,-0.0698518232,-0.1445435584,0.2413931191,-0.1052515581,-0.152235657,0.141644001,-0.1808562279,0.5583443642,0.0370699987,0.3895744383,-0.2105321437,0.0093455995,0.0729864463,-0.1095194891,0.0688644648,0.1238157228,-0.1383382678,-0.5897532701,-0.1902554631,0.1065888256,-0.0117324274,0.3856185079,-0.1194277927,-0.0214349553,0.1513940543,0.1890711784,0.7310069203,0.0729371309,-0.3278804123,-0.0309844855,-0.2211033851,-0.6239109635,-0.0709218159,0.0394208506,-0.1653980464,-0.1467238814,0.1922552288,-0.1503565162,-0.1719058305,0.3554315269,0.0347531512,-0.2583989501,0.0776256397,-0.0642109886,-0.523164928,-0.3752950728,0.0817698017,0.2867938876,0.1763403565,0.2706166506,0.3705551624,-0.025202468,-0.0146308187,0.18252334,0.0229771715,0.2532839775,0.1200980097,-0.0711414665,0.2466866821,0.2663377821,0.0747498721,0.3784388602,0.0218263883,-0.1386081427,-0.0059653665,-0.0273004491,0.3917846084,0.3158438504,-0.1038671583,-0.2018494606,0.000328526,0.0675849691,-0.0518511422,0.2731793523,0.4600665271,0.0707793012,-0.1523297429,-0.5686461329,0.1935224235,0.1283531487,0.1165830716,0.0412615761,0.1834687442,-0.4985372424,0.3000656366,0.2441835999,0.9627013803,0.0045670066,0.4122143984,0.4066116214,-0.1718395501,-0.0448778979,-0.113391012,0.1925232708,-0.2689211369,-0.3740708232,-0.1267440766,-0.1576801091,0.1537520885,0.1447482109,-0.4471739829,0.2574018836,-0.1016518995,0.1324464083,-0.0517069362,-0.0063035074,-0.1191574186,-0.3786025941,-0.4207054675,0.0605765507,0.0085383756,0.2239290327,-0.021146439,-0.0518368334,-0.0725239962,-0.3276558518,-0.1989789903,0.1438138634,-0.3425531387,0.1162754223,0.0215491783,-0.3508206308,0.5728157759,0.476177901,0.5377926826,0.0480021685,-0.0448533259,0.2743342817,-0.2238800377,0.2336572111,0.0026795755,0.1144546121,0.4311787486,-0.0262231417,0.0106902737,0.2838855386,-0.1292874068,-0.0491685048,-0.1847702265,-0.2700214386,0.0675761774,-0.372892797,-0.2156337798,0.1479793638,-0.0138652613,-0.1213464215,0.0629487187,-0.1453512013,-0.2010597736,0.1748773158,-0.1926721185,-0.3175874949,-0.0471323766,0.5520834327,-0.0154436249,-0.0259577204,0.9083886147,0.144269377,0.0290533174,-0.0506250225,0.0027593984,0.5083619952,-0.7060852647,0.1902077049,0.0442791618,-0.1163032502,-0.2586606443,0.2225267291,0.0910592005,-0.1223974228,0.106648542,0.0268161763,-0.3815151155,0.0435700901,0.1826019138,0.1099413484,-0.006349633,0.4157488346,0.3691781163,-0.160919562,-0.2136099041,0.0854345262,0.0152923148,0.2436425686,0.1348816752,-0.1812804937,0.1046393886,-0.2158272713,0.0344824716,-0.0117325801,0.0288340431,-0.1036318764,-0.303047955,0.0960220471,0.1902741194,-0.1844199449,-0.0305881146,-0.3210200965,-0.2273693979,-0.2802014351,0.148508355,0.2875715792,-0.0231884699,0.4670304656,-0.0447168574,-0.1073348075,-0.1710858345,-0.3624248505,-0.055877734,-0.2588469684,0.2030044794,0.1647436172,0.0901968554,-0.0337181799,-0.5971805453,0.2267414331,0.2485940754,0.2303075343,0.0678243041,-0.2015693933,0.1524341106,0.1655961126,0.2326402515,-0.0341038108,-0.2902600765,0.0206330046,0.1774314642,0.0617807992,-0.2854016721,0.3496537805,-0.0410753638,0.0037135417,0.0104907202,0.2545004785,0.0580310784,-0.0115493983,0.0494867526,0.07942947,0.2100776881,-0.060544882,0.3345696926,0.7754502892,-0.2804121077,0.2012109309,0.2750228941,0.1458674669,0.2543017566,0.3205238581,-0.1166608557,0.344928503,-0.0188068897,-0.0199496411,-0.0652343482,-0.1228847727,-0.3104800284,-0.1817116886,0.0402105004,0.0242764875,0.0616917051,0.5748892426,-0.2628111541,-0.3483905494,-0.2398824245,0.27559039,-0.1278471202,-0.0154439704,-0.0172689799,0.0287494026,-0.3195879459,-0.1499734968,-0.2125629634,-0.1377667636,0.6361117959,0.0198255535,-0.0023895337,-0.1892204583,-0.0454767235,0.2628235817,0.0819954276,-0.1269645691,0.2411371469,0.1011123285,-0.1883557737,0.0976076052,0.4976045489,0.2201207876,0.2194078565,-0.0463458151,-0.0102585051,-0.0782288536,-0.07122951,0.1536559314,0.2663584054,-0.1075949073,0.3812067509,0.2045420259,0.0881867185,-0.0691197589,0.1516282856,-0.2430780083,0.311342746,-0.2424247712,0.5272917151,-0.0208063293,-0.1454467177,-0.1479764134,0.0642576069,-0.1720504463,0.1480787694,0.3423065841,0.007757517,0.1103841811,-0.1328207552,0.0115130777,0.2430801243,0.2766731679,0.5207222104,-0.2535755932,-0.3213942349,-0.2156719267,-0.3150944412,-0.0384822972,0.1294478774,0.0174163729,-0.0245442092,-0.1764499247,0.3041365147,-0.0779176876,0.0809708983,-0.1065159515,-0.1017209142,0.211242795,-0.0689955726,-0.3980531991,-0.1088816151,0.3790934682,-0.0227049384,-0.0859454423,-0.1128532961,0.4846835732,-0.0293582678,-0.369301945,0.0441125371,-0.1199799553,-0.1286889464,0.2373430282,0.0764206648,0.2086417675,0.0588344671,0.2778173089,0.0176058225,-0.2506176531,-0.3266896904,0.0686159506,-0.0857321024,0.4292593896,-0.2401910871,-0.1764764786,-0.4258554578,0.0071116677,0.1781212837,0.0464417078,-0.4590160847,0.4230368733,-0.367138505,0.0823734626,0.2416835129,0.3583958447,0.0034974401,0.1476848125,-0.3863488436,-0.4244388938,0.6188313961,-0.1814698875,-0.4834167063,0.1206571609,-0.1362357289,-0.0846379399,-0.1597628593,-0.5339651108,0.0789652839,0.1070881411,-0.2078465074,-0.3984488547,0.2419902831,0.1348467618,0.0488820933,-0.0512382947,0.4404880404,-0.0221921708,-0.202132076,-0.1810315847,-0.0351451151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1867","title":"ERROR WHEN USING SET_TRANSFORM() ","comments":"I've tried your solutions @sgugger @lhoestq and the good news is that it throws no error. However, TPU training is taking forever, in 1 hour it has only trained 1 batch of 8192 elements, which doesn't make much sense... Is it possible that \"on the fly\" tokenization of batches is slowing down TPU training to that extent?","body":"Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n","comment_length":57,"text":"ERROR WHEN USING SET_TRANSFORM() \nHi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n\nI've tried your solutions @sgugger @lhoestq and the good news is that it throws no error. However, TPU training is taking forever, in 1 hour it has only trained 1 batch of 8192 elements, which doesn't make much sense... Is it possible that \"on the fly\" tokenization of batches is slowing down TPU training to that extent?","embeddings":[-0.1826189756,-0.2732517719,0.1393030435,0.1847398877,0.7292689681,0.1197050363,0.6009586453,0.2287764698,-0.3227846622,0.0544771031,0.2355112433,0.0333268568,-0.0883406326,-0.0234047007,0.1892216057,-0.5358201861,-0.0138219129,0.1288610101,-0.4634343982,-0.1831161827,-0.27669698,0.3134228885,-0.3415367901,0.1193406507,-0.3186003268,-0.0096082669,-0.047695376,0.016471222,0.0844151899,-0.1718301773,0.2334779501,-0.2800752223,0.1231957451,0.3974503875,-0.0001255372,0.1195578203,0.0606352575,-0.2890718281,-0.2367867827,-0.3003558218,-0.1437315047,-0.1552786827,-0.0154345548,-0.2408748716,-0.4479370713,-0.1890559047,0.0066991188,-0.2687551975,0.5653699636,0.7264817953,0.1215873957,0.2663597167,0.0996964872,-0.2382416725,-0.1261355132,0.1903086156,-0.1291115582,-0.0224978216,0.0317756385,0.2165266722,0.1336240321,0.1615502834,0.2006310076,-0.1752620935,0.4064543247,-0.0619789995,-0.1365289241,0.0435062647,0.0006193888,0.0104680918,0.2298866063,-0.3454308212,-0.420854032,-0.2735254467,0.1654442102,-0.2141008228,0.2700577974,0.1503871977,-0.066397734,0.22794348,-0.2600681484,-0.1098380312,-0.0588664711,0.0313304402,-0.1728429347,0.1623857617,-0.0065073655,0.1303315759,0.0931790695,0.1813862473,0.2739946544,0.1153549775,0.1876621246,0.3649872243,-0.265356034,-0.05478606,-0.1434788406,0.0108277686,-0.2889722586,-0.2492735982,-0.1723645478,-0.2085095048,0.0230882131,-0.0149517776,0.1071462706,0.6613349319,-0.2454049438,0.4977247119,0.0151178716,-0.1120734289,0.2516205609,-0.0635674,-0.0569484457,-0.3780822754,-0.0096246395,0.4175055921,0.0937968418,-0.006041795,-0.2874678373,-0.1932771057,-0.2089736909,-0.0578821562,0.1327706873,0.217985943,0.0588070415,0.2713345587,0.5032314658,0.3485563695,-0.1059520543,-0.2050584853,-0.1715610176,0.135013476,-0.3857185841,-0.1006987318,0.1464199424,-0.1412298381,0.3286191225,0.0532120019,0.0755415633,0.3815326989,0.1520244032,-0.191182673,0.2819835246,0.1436703354,-0.0143271564,0.2014978528,0.4693849087,0.10144265,-0.1861768216,0.2664014995,-0.2484502494,-0.4411483705,0.0609503128,0.0914024562,-0.0530007631,-0.0377959199,-0.1594837457,0.1583796293,0.4205289483,-0.0675125122,0.0763092414,-0.4352326989,0.0418038815,0.1099921837,0.090025045,0.0778380185,-0.2328578383,-0.3426035345,0.3258585334,0.1530645639,-0.143343091,0.2205605805,-0.4285855889,0.2562271655,-0.0625739917,-0.1584211737,0.2098882198,-0.3840966821,-0.0710864812,-0.0994279832,-0.1054709256,-0.219163999,0.0696848631,-0.1317368597,-0.0530189537,-0.1544256955,-0.2207259983,0.5086032748,-0.0608968958,0.2324942946,0.0340902545,-0.0222444125,0.1885959953,0.1376521885,0.0248874463,0.1576400846,-0.320507586,0.5120714307,0.0987583771,-0.2441777289,0.2444199026,0.4102393985,-0.244097501,-0.1919647157,0.0468020588,0.2764771283,-0.0697960332,-0.148902759,0.1784002781,0.1963512599,-0.0604719073,0.0484519824,-0.2369389385,-0.0641728118,-0.2133747637,-0.1585211456,-0.0149899647,0.1101620644,0.0207734834,-0.0464728251,-0.2484817207,0.2723915279,-0.095426932,0.1349620223,-0.5259901881,0.2280209661,-0.2145735472,-0.0914447233,-0.0678843781,0.2231209576,0.2404938191,-0.2305194438,-0.3247486055,0.3036762774,-0.2122558355,0.0371924043,-0.3943189681,0.2457401305,0.3090952635,-0.5342994928,0.0032722938,0.2634411156,0.1639743596,-0.127455622,-0.383022517,0.1136109233,0.0757724941,0.3314516544,0.1057743058,0.3340388238,0.0536119379,-0.0884053633,-0.2020374686,-0.0776441023,-0.0154921887,-0.0838234276,0.1649821252,0.0597501025,-0.2288423777,-0.1983881891,0.4673170149,0.148841694,0.0216302704,0.1293901354,-0.1397807598,0.0306708906,0.1618632525,-0.0950354412,0.562422514,0.0324369743,-0.2305446267,0.1875147969,0.0697727501,-0.1226138026,0.4002860487,-0.0731957108,0.0252318047,-0.106548816,0.0615078025,0.0952067301,-0.1649064571,-0.289491564,0.25625211,0.2521972656,-0.4736778736,0.2136499882,0.0259740539,0.1696037799,-0.5800776482,-0.3827403784,0.0345899239,-0.2872623503,-0.1098969206,-0.0208649188,0.1248787344,0.4439495802,0.1025984958,0.072272338,0.2448716611,-0.158022657,-0.0618830435,0.0459858961,-0.1911002547,0.0143575687,-0.1871763021,-0.2879667282,-0.0504698642,-0.3237556815,-0.0552909076,-0.1035404056,-0.1793443263,0.1855938435,-0.2955768704,0.0113008786,0.3231825829,0.1871942282,-0.1992335618,-0.3188306391,0.4316528141,0.0009160679,0.0959324092,-0.0127255181,-0.0703296512,-0.2425712794,-0.0997440368,-0.4264292717,-0.0446449108,-0.1689539254,0.0303072091,-0.2177828997,-0.1313049346,0.4203010499,0.2371348441,0.0843030065,0.2997512817,0.1740505844,-0.0198321659,-0.2111925334,0.4776291847,-0.2067952901,-0.3576395214,-0.1271426529,-0.0129219079,0.2218758017,0.2340473831,-0.1764791906,-0.178256005,-0.2040158212,-0.1126637682,-0.3014701605,0.435983181,0.4326145947,0.0603171624,0.1686085016,-0.050865747,-0.5156285167,0.3171764314,0.1424028128,0.4169045985,-0.0929954201,0.4798959196,0.0866109133,0.7751232386,0.0111009004,-0.3219003677,0.260661304,-0.1630743444,0.2260283828,-0.1403172314,-0.4625391364,0.2768841088,0.2438232154,0.0316104107,-0.0728213787,-0.0910808519,-0.0698518232,-0.1445435584,0.2413931191,-0.1052515581,-0.152235657,0.141644001,-0.1808562279,0.5583443642,0.0370699987,0.3895744383,-0.2105321437,0.0093455995,0.0729864463,-0.1095194891,0.0688644648,0.1238157228,-0.1383382678,-0.5897532701,-0.1902554631,0.1065888256,-0.0117324274,0.3856185079,-0.1194277927,-0.0214349553,0.1513940543,0.1890711784,0.7310069203,0.0729371309,-0.3278804123,-0.0309844855,-0.2211033851,-0.6239109635,-0.0709218159,0.0394208506,-0.1653980464,-0.1467238814,0.1922552288,-0.1503565162,-0.1719058305,0.3554315269,0.0347531512,-0.2583989501,0.0776256397,-0.0642109886,-0.523164928,-0.3752950728,0.0817698017,0.2867938876,0.1763403565,0.2706166506,0.3705551624,-0.025202468,-0.0146308187,0.18252334,0.0229771715,0.2532839775,0.1200980097,-0.0711414665,0.2466866821,0.2663377821,0.0747498721,0.3784388602,0.0218263883,-0.1386081427,-0.0059653665,-0.0273004491,0.3917846084,0.3158438504,-0.1038671583,-0.2018494606,0.000328526,0.0675849691,-0.0518511422,0.2731793523,0.4600665271,0.0707793012,-0.1523297429,-0.5686461329,0.1935224235,0.1283531487,0.1165830716,0.0412615761,0.1834687442,-0.4985372424,0.3000656366,0.2441835999,0.9627013803,0.0045670066,0.4122143984,0.4066116214,-0.1718395501,-0.0448778979,-0.113391012,0.1925232708,-0.2689211369,-0.3740708232,-0.1267440766,-0.1576801091,0.1537520885,0.1447482109,-0.4471739829,0.2574018836,-0.1016518995,0.1324464083,-0.0517069362,-0.0063035074,-0.1191574186,-0.3786025941,-0.4207054675,0.0605765507,0.0085383756,0.2239290327,-0.021146439,-0.0518368334,-0.0725239962,-0.3276558518,-0.1989789903,0.1438138634,-0.3425531387,0.1162754223,0.0215491783,-0.3508206308,0.5728157759,0.476177901,0.5377926826,0.0480021685,-0.0448533259,0.2743342817,-0.2238800377,0.2336572111,0.0026795755,0.1144546121,0.4311787486,-0.0262231417,0.0106902737,0.2838855386,-0.1292874068,-0.0491685048,-0.1847702265,-0.2700214386,0.0675761774,-0.372892797,-0.2156337798,0.1479793638,-0.0138652613,-0.1213464215,0.0629487187,-0.1453512013,-0.2010597736,0.1748773158,-0.1926721185,-0.3175874949,-0.0471323766,0.5520834327,-0.0154436249,-0.0259577204,0.9083886147,0.144269377,0.0290533174,-0.0506250225,0.0027593984,0.5083619952,-0.7060852647,0.1902077049,0.0442791618,-0.1163032502,-0.2586606443,0.2225267291,0.0910592005,-0.1223974228,0.106648542,0.0268161763,-0.3815151155,0.0435700901,0.1826019138,0.1099413484,-0.006349633,0.4157488346,0.3691781163,-0.160919562,-0.2136099041,0.0854345262,0.0152923148,0.2436425686,0.1348816752,-0.1812804937,0.1046393886,-0.2158272713,0.0344824716,-0.0117325801,0.0288340431,-0.1036318764,-0.303047955,0.0960220471,0.1902741194,-0.1844199449,-0.0305881146,-0.3210200965,-0.2273693979,-0.2802014351,0.148508355,0.2875715792,-0.0231884699,0.4670304656,-0.0447168574,-0.1073348075,-0.1710858345,-0.3624248505,-0.055877734,-0.2588469684,0.2030044794,0.1647436172,0.0901968554,-0.0337181799,-0.5971805453,0.2267414331,0.2485940754,0.2303075343,0.0678243041,-0.2015693933,0.1524341106,0.1655961126,0.2326402515,-0.0341038108,-0.2902600765,0.0206330046,0.1774314642,0.0617807992,-0.2854016721,0.3496537805,-0.0410753638,0.0037135417,0.0104907202,0.2545004785,0.0580310784,-0.0115493983,0.0494867526,0.07942947,0.2100776881,-0.060544882,0.3345696926,0.7754502892,-0.2804121077,0.2012109309,0.2750228941,0.1458674669,0.2543017566,0.3205238581,-0.1166608557,0.344928503,-0.0188068897,-0.0199496411,-0.0652343482,-0.1228847727,-0.3104800284,-0.1817116886,0.0402105004,0.0242764875,0.0616917051,0.5748892426,-0.2628111541,-0.3483905494,-0.2398824245,0.27559039,-0.1278471202,-0.0154439704,-0.0172689799,0.0287494026,-0.3195879459,-0.1499734968,-0.2125629634,-0.1377667636,0.6361117959,0.0198255535,-0.0023895337,-0.1892204583,-0.0454767235,0.2628235817,0.0819954276,-0.1269645691,0.2411371469,0.1011123285,-0.1883557737,0.0976076052,0.4976045489,0.2201207876,0.2194078565,-0.0463458151,-0.0102585051,-0.0782288536,-0.07122951,0.1536559314,0.2663584054,-0.1075949073,0.3812067509,0.2045420259,0.0881867185,-0.0691197589,0.1516282856,-0.2430780083,0.311342746,-0.2424247712,0.5272917151,-0.0208063293,-0.1454467177,-0.1479764134,0.0642576069,-0.1720504463,0.1480787694,0.3423065841,0.007757517,0.1103841811,-0.1328207552,0.0115130777,0.2430801243,0.2766731679,0.5207222104,-0.2535755932,-0.3213942349,-0.2156719267,-0.3150944412,-0.0384822972,0.1294478774,0.0174163729,-0.0245442092,-0.1764499247,0.3041365147,-0.0779176876,0.0809708983,-0.1065159515,-0.1017209142,0.211242795,-0.0689955726,-0.3980531991,-0.1088816151,0.3790934682,-0.0227049384,-0.0859454423,-0.1128532961,0.4846835732,-0.0293582678,-0.369301945,0.0441125371,-0.1199799553,-0.1286889464,0.2373430282,0.0764206648,0.2086417675,0.0588344671,0.2778173089,0.0176058225,-0.2506176531,-0.3266896904,0.0686159506,-0.0857321024,0.4292593896,-0.2401910871,-0.1764764786,-0.4258554578,0.0071116677,0.1781212837,0.0464417078,-0.4590160847,0.4230368733,-0.367138505,0.0823734626,0.2416835129,0.3583958447,0.0034974401,0.1476848125,-0.3863488436,-0.4244388938,0.6188313961,-0.1814698875,-0.4834167063,0.1206571609,-0.1362357289,-0.0846379399,-0.1597628593,-0.5339651108,0.0789652839,0.1070881411,-0.2078465074,-0.3984488547,0.2419902831,0.1348467618,0.0488820933,-0.0512382947,0.4404880404,-0.0221921708,-0.202132076,-0.1810315847,-0.0351451151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1867","title":"ERROR WHEN USING SET_TRANSFORM() ","comments":"I'm pretty sure this is because of padding but @sgugger might know better","body":"Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n","comment_length":13,"text":"ERROR WHEN USING SET_TRANSFORM() \nHi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n\nI'm pretty sure this is because of padding but @sgugger might know better","embeddings":[-0.1826189756,-0.2732517719,0.1393030435,0.1847398877,0.7292689681,0.1197050363,0.6009586453,0.2287764698,-0.3227846622,0.0544771031,0.2355112433,0.0333268568,-0.0883406326,-0.0234047007,0.1892216057,-0.5358201861,-0.0138219129,0.1288610101,-0.4634343982,-0.1831161827,-0.27669698,0.3134228885,-0.3415367901,0.1193406507,-0.3186003268,-0.0096082669,-0.047695376,0.016471222,0.0844151899,-0.1718301773,0.2334779501,-0.2800752223,0.1231957451,0.3974503875,-0.0001255372,0.1195578203,0.0606352575,-0.2890718281,-0.2367867827,-0.3003558218,-0.1437315047,-0.1552786827,-0.0154345548,-0.2408748716,-0.4479370713,-0.1890559047,0.0066991188,-0.2687551975,0.5653699636,0.7264817953,0.1215873957,0.2663597167,0.0996964872,-0.2382416725,-0.1261355132,0.1903086156,-0.1291115582,-0.0224978216,0.0317756385,0.2165266722,0.1336240321,0.1615502834,0.2006310076,-0.1752620935,0.4064543247,-0.0619789995,-0.1365289241,0.0435062647,0.0006193888,0.0104680918,0.2298866063,-0.3454308212,-0.420854032,-0.2735254467,0.1654442102,-0.2141008228,0.2700577974,0.1503871977,-0.066397734,0.22794348,-0.2600681484,-0.1098380312,-0.0588664711,0.0313304402,-0.1728429347,0.1623857617,-0.0065073655,0.1303315759,0.0931790695,0.1813862473,0.2739946544,0.1153549775,0.1876621246,0.3649872243,-0.265356034,-0.05478606,-0.1434788406,0.0108277686,-0.2889722586,-0.2492735982,-0.1723645478,-0.2085095048,0.0230882131,-0.0149517776,0.1071462706,0.6613349319,-0.2454049438,0.4977247119,0.0151178716,-0.1120734289,0.2516205609,-0.0635674,-0.0569484457,-0.3780822754,-0.0096246395,0.4175055921,0.0937968418,-0.006041795,-0.2874678373,-0.1932771057,-0.2089736909,-0.0578821562,0.1327706873,0.217985943,0.0588070415,0.2713345587,0.5032314658,0.3485563695,-0.1059520543,-0.2050584853,-0.1715610176,0.135013476,-0.3857185841,-0.1006987318,0.1464199424,-0.1412298381,0.3286191225,0.0532120019,0.0755415633,0.3815326989,0.1520244032,-0.191182673,0.2819835246,0.1436703354,-0.0143271564,0.2014978528,0.4693849087,0.10144265,-0.1861768216,0.2664014995,-0.2484502494,-0.4411483705,0.0609503128,0.0914024562,-0.0530007631,-0.0377959199,-0.1594837457,0.1583796293,0.4205289483,-0.0675125122,0.0763092414,-0.4352326989,0.0418038815,0.1099921837,0.090025045,0.0778380185,-0.2328578383,-0.3426035345,0.3258585334,0.1530645639,-0.143343091,0.2205605805,-0.4285855889,0.2562271655,-0.0625739917,-0.1584211737,0.2098882198,-0.3840966821,-0.0710864812,-0.0994279832,-0.1054709256,-0.219163999,0.0696848631,-0.1317368597,-0.0530189537,-0.1544256955,-0.2207259983,0.5086032748,-0.0608968958,0.2324942946,0.0340902545,-0.0222444125,0.1885959953,0.1376521885,0.0248874463,0.1576400846,-0.320507586,0.5120714307,0.0987583771,-0.2441777289,0.2444199026,0.4102393985,-0.244097501,-0.1919647157,0.0468020588,0.2764771283,-0.0697960332,-0.148902759,0.1784002781,0.1963512599,-0.0604719073,0.0484519824,-0.2369389385,-0.0641728118,-0.2133747637,-0.1585211456,-0.0149899647,0.1101620644,0.0207734834,-0.0464728251,-0.2484817207,0.2723915279,-0.095426932,0.1349620223,-0.5259901881,0.2280209661,-0.2145735472,-0.0914447233,-0.0678843781,0.2231209576,0.2404938191,-0.2305194438,-0.3247486055,0.3036762774,-0.2122558355,0.0371924043,-0.3943189681,0.2457401305,0.3090952635,-0.5342994928,0.0032722938,0.2634411156,0.1639743596,-0.127455622,-0.383022517,0.1136109233,0.0757724941,0.3314516544,0.1057743058,0.3340388238,0.0536119379,-0.0884053633,-0.2020374686,-0.0776441023,-0.0154921887,-0.0838234276,0.1649821252,0.0597501025,-0.2288423777,-0.1983881891,0.4673170149,0.148841694,0.0216302704,0.1293901354,-0.1397807598,0.0306708906,0.1618632525,-0.0950354412,0.562422514,0.0324369743,-0.2305446267,0.1875147969,0.0697727501,-0.1226138026,0.4002860487,-0.0731957108,0.0252318047,-0.106548816,0.0615078025,0.0952067301,-0.1649064571,-0.289491564,0.25625211,0.2521972656,-0.4736778736,0.2136499882,0.0259740539,0.1696037799,-0.5800776482,-0.3827403784,0.0345899239,-0.2872623503,-0.1098969206,-0.0208649188,0.1248787344,0.4439495802,0.1025984958,0.072272338,0.2448716611,-0.158022657,-0.0618830435,0.0459858961,-0.1911002547,0.0143575687,-0.1871763021,-0.2879667282,-0.0504698642,-0.3237556815,-0.0552909076,-0.1035404056,-0.1793443263,0.1855938435,-0.2955768704,0.0113008786,0.3231825829,0.1871942282,-0.1992335618,-0.3188306391,0.4316528141,0.0009160679,0.0959324092,-0.0127255181,-0.0703296512,-0.2425712794,-0.0997440368,-0.4264292717,-0.0446449108,-0.1689539254,0.0303072091,-0.2177828997,-0.1313049346,0.4203010499,0.2371348441,0.0843030065,0.2997512817,0.1740505844,-0.0198321659,-0.2111925334,0.4776291847,-0.2067952901,-0.3576395214,-0.1271426529,-0.0129219079,0.2218758017,0.2340473831,-0.1764791906,-0.178256005,-0.2040158212,-0.1126637682,-0.3014701605,0.435983181,0.4326145947,0.0603171624,0.1686085016,-0.050865747,-0.5156285167,0.3171764314,0.1424028128,0.4169045985,-0.0929954201,0.4798959196,0.0866109133,0.7751232386,0.0111009004,-0.3219003677,0.260661304,-0.1630743444,0.2260283828,-0.1403172314,-0.4625391364,0.2768841088,0.2438232154,0.0316104107,-0.0728213787,-0.0910808519,-0.0698518232,-0.1445435584,0.2413931191,-0.1052515581,-0.152235657,0.141644001,-0.1808562279,0.5583443642,0.0370699987,0.3895744383,-0.2105321437,0.0093455995,0.0729864463,-0.1095194891,0.0688644648,0.1238157228,-0.1383382678,-0.5897532701,-0.1902554631,0.1065888256,-0.0117324274,0.3856185079,-0.1194277927,-0.0214349553,0.1513940543,0.1890711784,0.7310069203,0.0729371309,-0.3278804123,-0.0309844855,-0.2211033851,-0.6239109635,-0.0709218159,0.0394208506,-0.1653980464,-0.1467238814,0.1922552288,-0.1503565162,-0.1719058305,0.3554315269,0.0347531512,-0.2583989501,0.0776256397,-0.0642109886,-0.523164928,-0.3752950728,0.0817698017,0.2867938876,0.1763403565,0.2706166506,0.3705551624,-0.025202468,-0.0146308187,0.18252334,0.0229771715,0.2532839775,0.1200980097,-0.0711414665,0.2466866821,0.2663377821,0.0747498721,0.3784388602,0.0218263883,-0.1386081427,-0.0059653665,-0.0273004491,0.3917846084,0.3158438504,-0.1038671583,-0.2018494606,0.000328526,0.0675849691,-0.0518511422,0.2731793523,0.4600665271,0.0707793012,-0.1523297429,-0.5686461329,0.1935224235,0.1283531487,0.1165830716,0.0412615761,0.1834687442,-0.4985372424,0.3000656366,0.2441835999,0.9627013803,0.0045670066,0.4122143984,0.4066116214,-0.1718395501,-0.0448778979,-0.113391012,0.1925232708,-0.2689211369,-0.3740708232,-0.1267440766,-0.1576801091,0.1537520885,0.1447482109,-0.4471739829,0.2574018836,-0.1016518995,0.1324464083,-0.0517069362,-0.0063035074,-0.1191574186,-0.3786025941,-0.4207054675,0.0605765507,0.0085383756,0.2239290327,-0.021146439,-0.0518368334,-0.0725239962,-0.3276558518,-0.1989789903,0.1438138634,-0.3425531387,0.1162754223,0.0215491783,-0.3508206308,0.5728157759,0.476177901,0.5377926826,0.0480021685,-0.0448533259,0.2743342817,-0.2238800377,0.2336572111,0.0026795755,0.1144546121,0.4311787486,-0.0262231417,0.0106902737,0.2838855386,-0.1292874068,-0.0491685048,-0.1847702265,-0.2700214386,0.0675761774,-0.372892797,-0.2156337798,0.1479793638,-0.0138652613,-0.1213464215,0.0629487187,-0.1453512013,-0.2010597736,0.1748773158,-0.1926721185,-0.3175874949,-0.0471323766,0.5520834327,-0.0154436249,-0.0259577204,0.9083886147,0.144269377,0.0290533174,-0.0506250225,0.0027593984,0.5083619952,-0.7060852647,0.1902077049,0.0442791618,-0.1163032502,-0.2586606443,0.2225267291,0.0910592005,-0.1223974228,0.106648542,0.0268161763,-0.3815151155,0.0435700901,0.1826019138,0.1099413484,-0.006349633,0.4157488346,0.3691781163,-0.160919562,-0.2136099041,0.0854345262,0.0152923148,0.2436425686,0.1348816752,-0.1812804937,0.1046393886,-0.2158272713,0.0344824716,-0.0117325801,0.0288340431,-0.1036318764,-0.303047955,0.0960220471,0.1902741194,-0.1844199449,-0.0305881146,-0.3210200965,-0.2273693979,-0.2802014351,0.148508355,0.2875715792,-0.0231884699,0.4670304656,-0.0447168574,-0.1073348075,-0.1710858345,-0.3624248505,-0.055877734,-0.2588469684,0.2030044794,0.1647436172,0.0901968554,-0.0337181799,-0.5971805453,0.2267414331,0.2485940754,0.2303075343,0.0678243041,-0.2015693933,0.1524341106,0.1655961126,0.2326402515,-0.0341038108,-0.2902600765,0.0206330046,0.1774314642,0.0617807992,-0.2854016721,0.3496537805,-0.0410753638,0.0037135417,0.0104907202,0.2545004785,0.0580310784,-0.0115493983,0.0494867526,0.07942947,0.2100776881,-0.060544882,0.3345696926,0.7754502892,-0.2804121077,0.2012109309,0.2750228941,0.1458674669,0.2543017566,0.3205238581,-0.1166608557,0.344928503,-0.0188068897,-0.0199496411,-0.0652343482,-0.1228847727,-0.3104800284,-0.1817116886,0.0402105004,0.0242764875,0.0616917051,0.5748892426,-0.2628111541,-0.3483905494,-0.2398824245,0.27559039,-0.1278471202,-0.0154439704,-0.0172689799,0.0287494026,-0.3195879459,-0.1499734968,-0.2125629634,-0.1377667636,0.6361117959,0.0198255535,-0.0023895337,-0.1892204583,-0.0454767235,0.2628235817,0.0819954276,-0.1269645691,0.2411371469,0.1011123285,-0.1883557737,0.0976076052,0.4976045489,0.2201207876,0.2194078565,-0.0463458151,-0.0102585051,-0.0782288536,-0.07122951,0.1536559314,0.2663584054,-0.1075949073,0.3812067509,0.2045420259,0.0881867185,-0.0691197589,0.1516282856,-0.2430780083,0.311342746,-0.2424247712,0.5272917151,-0.0208063293,-0.1454467177,-0.1479764134,0.0642576069,-0.1720504463,0.1480787694,0.3423065841,0.007757517,0.1103841811,-0.1328207552,0.0115130777,0.2430801243,0.2766731679,0.5207222104,-0.2535755932,-0.3213942349,-0.2156719267,-0.3150944412,-0.0384822972,0.1294478774,0.0174163729,-0.0245442092,-0.1764499247,0.3041365147,-0.0779176876,0.0809708983,-0.1065159515,-0.1017209142,0.211242795,-0.0689955726,-0.3980531991,-0.1088816151,0.3790934682,-0.0227049384,-0.0859454423,-0.1128532961,0.4846835732,-0.0293582678,-0.369301945,0.0441125371,-0.1199799553,-0.1286889464,0.2373430282,0.0764206648,0.2086417675,0.0588344671,0.2778173089,0.0176058225,-0.2506176531,-0.3266896904,0.0686159506,-0.0857321024,0.4292593896,-0.2401910871,-0.1764764786,-0.4258554578,0.0071116677,0.1781212837,0.0464417078,-0.4590160847,0.4230368733,-0.367138505,0.0823734626,0.2416835129,0.3583958447,0.0034974401,0.1476848125,-0.3863488436,-0.4244388938,0.6188313961,-0.1814698875,-0.4834167063,0.1206571609,-0.1362357289,-0.0846379399,-0.1597628593,-0.5339651108,0.0789652839,0.1070881411,-0.2078465074,-0.3984488547,0.2419902831,0.1348467618,0.0488820933,-0.0512382947,0.4404880404,-0.0221921708,-0.202132076,-0.1810315847,-0.0351451151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1867","title":"ERROR WHEN USING SET_TRANSFORM() ","comments":"I don't know what the value of `padding` is in your lines of code pasted above so I can't say for sure. The first batch will be very slow on TPU since it compiles everything, so that's normal (1 hour is long but 8192 elements is also large). Then if your batches are not of the same lengths, it will recompile everything at each step instead of using the same graph, which will be very slow, so you should double check you are using padding to make everything the exact same shape. ","body":"Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n","comment_length":92,"text":"ERROR WHEN USING SET_TRANSFORM() \nHi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n\nI don't know what the value of `padding` is in your lines of code pasted above so I can't say for sure. The first batch will be very slow on TPU since it compiles everything, so that's normal (1 hour is long but 8192 elements is also large). Then if your batches are not of the same lengths, it will recompile everything at each step instead of using the same graph, which will be very slow, so you should double check you are using padding to make everything the exact same shape. ","embeddings":[-0.1826189756,-0.2732517719,0.1393030435,0.1847398877,0.7292689681,0.1197050363,0.6009586453,0.2287764698,-0.3227846622,0.0544771031,0.2355112433,0.0333268568,-0.0883406326,-0.0234047007,0.1892216057,-0.5358201861,-0.0138219129,0.1288610101,-0.4634343982,-0.1831161827,-0.27669698,0.3134228885,-0.3415367901,0.1193406507,-0.3186003268,-0.0096082669,-0.047695376,0.016471222,0.0844151899,-0.1718301773,0.2334779501,-0.2800752223,0.1231957451,0.3974503875,-0.0001255372,0.1195578203,0.0606352575,-0.2890718281,-0.2367867827,-0.3003558218,-0.1437315047,-0.1552786827,-0.0154345548,-0.2408748716,-0.4479370713,-0.1890559047,0.0066991188,-0.2687551975,0.5653699636,0.7264817953,0.1215873957,0.2663597167,0.0996964872,-0.2382416725,-0.1261355132,0.1903086156,-0.1291115582,-0.0224978216,0.0317756385,0.2165266722,0.1336240321,0.1615502834,0.2006310076,-0.1752620935,0.4064543247,-0.0619789995,-0.1365289241,0.0435062647,0.0006193888,0.0104680918,0.2298866063,-0.3454308212,-0.420854032,-0.2735254467,0.1654442102,-0.2141008228,0.2700577974,0.1503871977,-0.066397734,0.22794348,-0.2600681484,-0.1098380312,-0.0588664711,0.0313304402,-0.1728429347,0.1623857617,-0.0065073655,0.1303315759,0.0931790695,0.1813862473,0.2739946544,0.1153549775,0.1876621246,0.3649872243,-0.265356034,-0.05478606,-0.1434788406,0.0108277686,-0.2889722586,-0.2492735982,-0.1723645478,-0.2085095048,0.0230882131,-0.0149517776,0.1071462706,0.6613349319,-0.2454049438,0.4977247119,0.0151178716,-0.1120734289,0.2516205609,-0.0635674,-0.0569484457,-0.3780822754,-0.0096246395,0.4175055921,0.0937968418,-0.006041795,-0.2874678373,-0.1932771057,-0.2089736909,-0.0578821562,0.1327706873,0.217985943,0.0588070415,0.2713345587,0.5032314658,0.3485563695,-0.1059520543,-0.2050584853,-0.1715610176,0.135013476,-0.3857185841,-0.1006987318,0.1464199424,-0.1412298381,0.3286191225,0.0532120019,0.0755415633,0.3815326989,0.1520244032,-0.191182673,0.2819835246,0.1436703354,-0.0143271564,0.2014978528,0.4693849087,0.10144265,-0.1861768216,0.2664014995,-0.2484502494,-0.4411483705,0.0609503128,0.0914024562,-0.0530007631,-0.0377959199,-0.1594837457,0.1583796293,0.4205289483,-0.0675125122,0.0763092414,-0.4352326989,0.0418038815,0.1099921837,0.090025045,0.0778380185,-0.2328578383,-0.3426035345,0.3258585334,0.1530645639,-0.143343091,0.2205605805,-0.4285855889,0.2562271655,-0.0625739917,-0.1584211737,0.2098882198,-0.3840966821,-0.0710864812,-0.0994279832,-0.1054709256,-0.219163999,0.0696848631,-0.1317368597,-0.0530189537,-0.1544256955,-0.2207259983,0.5086032748,-0.0608968958,0.2324942946,0.0340902545,-0.0222444125,0.1885959953,0.1376521885,0.0248874463,0.1576400846,-0.320507586,0.5120714307,0.0987583771,-0.2441777289,0.2444199026,0.4102393985,-0.244097501,-0.1919647157,0.0468020588,0.2764771283,-0.0697960332,-0.148902759,0.1784002781,0.1963512599,-0.0604719073,0.0484519824,-0.2369389385,-0.0641728118,-0.2133747637,-0.1585211456,-0.0149899647,0.1101620644,0.0207734834,-0.0464728251,-0.2484817207,0.2723915279,-0.095426932,0.1349620223,-0.5259901881,0.2280209661,-0.2145735472,-0.0914447233,-0.0678843781,0.2231209576,0.2404938191,-0.2305194438,-0.3247486055,0.3036762774,-0.2122558355,0.0371924043,-0.3943189681,0.2457401305,0.3090952635,-0.5342994928,0.0032722938,0.2634411156,0.1639743596,-0.127455622,-0.383022517,0.1136109233,0.0757724941,0.3314516544,0.1057743058,0.3340388238,0.0536119379,-0.0884053633,-0.2020374686,-0.0776441023,-0.0154921887,-0.0838234276,0.1649821252,0.0597501025,-0.2288423777,-0.1983881891,0.4673170149,0.148841694,0.0216302704,0.1293901354,-0.1397807598,0.0306708906,0.1618632525,-0.0950354412,0.562422514,0.0324369743,-0.2305446267,0.1875147969,0.0697727501,-0.1226138026,0.4002860487,-0.0731957108,0.0252318047,-0.106548816,0.0615078025,0.0952067301,-0.1649064571,-0.289491564,0.25625211,0.2521972656,-0.4736778736,0.2136499882,0.0259740539,0.1696037799,-0.5800776482,-0.3827403784,0.0345899239,-0.2872623503,-0.1098969206,-0.0208649188,0.1248787344,0.4439495802,0.1025984958,0.072272338,0.2448716611,-0.158022657,-0.0618830435,0.0459858961,-0.1911002547,0.0143575687,-0.1871763021,-0.2879667282,-0.0504698642,-0.3237556815,-0.0552909076,-0.1035404056,-0.1793443263,0.1855938435,-0.2955768704,0.0113008786,0.3231825829,0.1871942282,-0.1992335618,-0.3188306391,0.4316528141,0.0009160679,0.0959324092,-0.0127255181,-0.0703296512,-0.2425712794,-0.0997440368,-0.4264292717,-0.0446449108,-0.1689539254,0.0303072091,-0.2177828997,-0.1313049346,0.4203010499,0.2371348441,0.0843030065,0.2997512817,0.1740505844,-0.0198321659,-0.2111925334,0.4776291847,-0.2067952901,-0.3576395214,-0.1271426529,-0.0129219079,0.2218758017,0.2340473831,-0.1764791906,-0.178256005,-0.2040158212,-0.1126637682,-0.3014701605,0.435983181,0.4326145947,0.0603171624,0.1686085016,-0.050865747,-0.5156285167,0.3171764314,0.1424028128,0.4169045985,-0.0929954201,0.4798959196,0.0866109133,0.7751232386,0.0111009004,-0.3219003677,0.260661304,-0.1630743444,0.2260283828,-0.1403172314,-0.4625391364,0.2768841088,0.2438232154,0.0316104107,-0.0728213787,-0.0910808519,-0.0698518232,-0.1445435584,0.2413931191,-0.1052515581,-0.152235657,0.141644001,-0.1808562279,0.5583443642,0.0370699987,0.3895744383,-0.2105321437,0.0093455995,0.0729864463,-0.1095194891,0.0688644648,0.1238157228,-0.1383382678,-0.5897532701,-0.1902554631,0.1065888256,-0.0117324274,0.3856185079,-0.1194277927,-0.0214349553,0.1513940543,0.1890711784,0.7310069203,0.0729371309,-0.3278804123,-0.0309844855,-0.2211033851,-0.6239109635,-0.0709218159,0.0394208506,-0.1653980464,-0.1467238814,0.1922552288,-0.1503565162,-0.1719058305,0.3554315269,0.0347531512,-0.2583989501,0.0776256397,-0.0642109886,-0.523164928,-0.3752950728,0.0817698017,0.2867938876,0.1763403565,0.2706166506,0.3705551624,-0.025202468,-0.0146308187,0.18252334,0.0229771715,0.2532839775,0.1200980097,-0.0711414665,0.2466866821,0.2663377821,0.0747498721,0.3784388602,0.0218263883,-0.1386081427,-0.0059653665,-0.0273004491,0.3917846084,0.3158438504,-0.1038671583,-0.2018494606,0.000328526,0.0675849691,-0.0518511422,0.2731793523,0.4600665271,0.0707793012,-0.1523297429,-0.5686461329,0.1935224235,0.1283531487,0.1165830716,0.0412615761,0.1834687442,-0.4985372424,0.3000656366,0.2441835999,0.9627013803,0.0045670066,0.4122143984,0.4066116214,-0.1718395501,-0.0448778979,-0.113391012,0.1925232708,-0.2689211369,-0.3740708232,-0.1267440766,-0.1576801091,0.1537520885,0.1447482109,-0.4471739829,0.2574018836,-0.1016518995,0.1324464083,-0.0517069362,-0.0063035074,-0.1191574186,-0.3786025941,-0.4207054675,0.0605765507,0.0085383756,0.2239290327,-0.021146439,-0.0518368334,-0.0725239962,-0.3276558518,-0.1989789903,0.1438138634,-0.3425531387,0.1162754223,0.0215491783,-0.3508206308,0.5728157759,0.476177901,0.5377926826,0.0480021685,-0.0448533259,0.2743342817,-0.2238800377,0.2336572111,0.0026795755,0.1144546121,0.4311787486,-0.0262231417,0.0106902737,0.2838855386,-0.1292874068,-0.0491685048,-0.1847702265,-0.2700214386,0.0675761774,-0.372892797,-0.2156337798,0.1479793638,-0.0138652613,-0.1213464215,0.0629487187,-0.1453512013,-0.2010597736,0.1748773158,-0.1926721185,-0.3175874949,-0.0471323766,0.5520834327,-0.0154436249,-0.0259577204,0.9083886147,0.144269377,0.0290533174,-0.0506250225,0.0027593984,0.5083619952,-0.7060852647,0.1902077049,0.0442791618,-0.1163032502,-0.2586606443,0.2225267291,0.0910592005,-0.1223974228,0.106648542,0.0268161763,-0.3815151155,0.0435700901,0.1826019138,0.1099413484,-0.006349633,0.4157488346,0.3691781163,-0.160919562,-0.2136099041,0.0854345262,0.0152923148,0.2436425686,0.1348816752,-0.1812804937,0.1046393886,-0.2158272713,0.0344824716,-0.0117325801,0.0288340431,-0.1036318764,-0.303047955,0.0960220471,0.1902741194,-0.1844199449,-0.0305881146,-0.3210200965,-0.2273693979,-0.2802014351,0.148508355,0.2875715792,-0.0231884699,0.4670304656,-0.0447168574,-0.1073348075,-0.1710858345,-0.3624248505,-0.055877734,-0.2588469684,0.2030044794,0.1647436172,0.0901968554,-0.0337181799,-0.5971805453,0.2267414331,0.2485940754,0.2303075343,0.0678243041,-0.2015693933,0.1524341106,0.1655961126,0.2326402515,-0.0341038108,-0.2902600765,0.0206330046,0.1774314642,0.0617807992,-0.2854016721,0.3496537805,-0.0410753638,0.0037135417,0.0104907202,0.2545004785,0.0580310784,-0.0115493983,0.0494867526,0.07942947,0.2100776881,-0.060544882,0.3345696926,0.7754502892,-0.2804121077,0.2012109309,0.2750228941,0.1458674669,0.2543017566,0.3205238581,-0.1166608557,0.344928503,-0.0188068897,-0.0199496411,-0.0652343482,-0.1228847727,-0.3104800284,-0.1817116886,0.0402105004,0.0242764875,0.0616917051,0.5748892426,-0.2628111541,-0.3483905494,-0.2398824245,0.27559039,-0.1278471202,-0.0154439704,-0.0172689799,0.0287494026,-0.3195879459,-0.1499734968,-0.2125629634,-0.1377667636,0.6361117959,0.0198255535,-0.0023895337,-0.1892204583,-0.0454767235,0.2628235817,0.0819954276,-0.1269645691,0.2411371469,0.1011123285,-0.1883557737,0.0976076052,0.4976045489,0.2201207876,0.2194078565,-0.0463458151,-0.0102585051,-0.0782288536,-0.07122951,0.1536559314,0.2663584054,-0.1075949073,0.3812067509,0.2045420259,0.0881867185,-0.0691197589,0.1516282856,-0.2430780083,0.311342746,-0.2424247712,0.5272917151,-0.0208063293,-0.1454467177,-0.1479764134,0.0642576069,-0.1720504463,0.1480787694,0.3423065841,0.007757517,0.1103841811,-0.1328207552,0.0115130777,0.2430801243,0.2766731679,0.5207222104,-0.2535755932,-0.3213942349,-0.2156719267,-0.3150944412,-0.0384822972,0.1294478774,0.0174163729,-0.0245442092,-0.1764499247,0.3041365147,-0.0779176876,0.0809708983,-0.1065159515,-0.1017209142,0.211242795,-0.0689955726,-0.3980531991,-0.1088816151,0.3790934682,-0.0227049384,-0.0859454423,-0.1128532961,0.4846835732,-0.0293582678,-0.369301945,0.0441125371,-0.1199799553,-0.1286889464,0.2373430282,0.0764206648,0.2086417675,0.0588344671,0.2778173089,0.0176058225,-0.2506176531,-0.3266896904,0.0686159506,-0.0857321024,0.4292593896,-0.2401910871,-0.1764764786,-0.4258554578,0.0071116677,0.1781212837,0.0464417078,-0.4590160847,0.4230368733,-0.367138505,0.0823734626,0.2416835129,0.3583958447,0.0034974401,0.1476848125,-0.3863488436,-0.4244388938,0.6188313961,-0.1814698875,-0.4834167063,0.1206571609,-0.1362357289,-0.0846379399,-0.1597628593,-0.5339651108,0.0789652839,0.1070881411,-0.2078465074,-0.3984488547,0.2419902831,0.1348467618,0.0488820933,-0.0512382947,0.4404880404,-0.0221921708,-0.202132076,-0.1810315847,-0.0351451151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1867","title":"ERROR WHEN USING SET_TRANSFORM() ","comments":"I have tried now on a GPU and it goes smooth! Amazing feature .set_transform() instead of .map()! Now I can pre-train my model without the hard disk limitation. Thanks for your work all HuggingFace team!! :clap: ","body":"Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n","comment_length":36,"text":"ERROR WHEN USING SET_TRANSFORM() \nHi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n\nI have tried now on a GPU and it goes smooth! Amazing feature .set_transform() instead of .map()! Now I can pre-train my model without the hard disk limitation. Thanks for your work all HuggingFace team!! :clap: ","embeddings":[-0.1826189756,-0.2732517719,0.1393030435,0.1847398877,0.7292689681,0.1197050363,0.6009586453,0.2287764698,-0.3227846622,0.0544771031,0.2355112433,0.0333268568,-0.0883406326,-0.0234047007,0.1892216057,-0.5358201861,-0.0138219129,0.1288610101,-0.4634343982,-0.1831161827,-0.27669698,0.3134228885,-0.3415367901,0.1193406507,-0.3186003268,-0.0096082669,-0.047695376,0.016471222,0.0844151899,-0.1718301773,0.2334779501,-0.2800752223,0.1231957451,0.3974503875,-0.0001255372,0.1195578203,0.0606352575,-0.2890718281,-0.2367867827,-0.3003558218,-0.1437315047,-0.1552786827,-0.0154345548,-0.2408748716,-0.4479370713,-0.1890559047,0.0066991188,-0.2687551975,0.5653699636,0.7264817953,0.1215873957,0.2663597167,0.0996964872,-0.2382416725,-0.1261355132,0.1903086156,-0.1291115582,-0.0224978216,0.0317756385,0.2165266722,0.1336240321,0.1615502834,0.2006310076,-0.1752620935,0.4064543247,-0.0619789995,-0.1365289241,0.0435062647,0.0006193888,0.0104680918,0.2298866063,-0.3454308212,-0.420854032,-0.2735254467,0.1654442102,-0.2141008228,0.2700577974,0.1503871977,-0.066397734,0.22794348,-0.2600681484,-0.1098380312,-0.0588664711,0.0313304402,-0.1728429347,0.1623857617,-0.0065073655,0.1303315759,0.0931790695,0.1813862473,0.2739946544,0.1153549775,0.1876621246,0.3649872243,-0.265356034,-0.05478606,-0.1434788406,0.0108277686,-0.2889722586,-0.2492735982,-0.1723645478,-0.2085095048,0.0230882131,-0.0149517776,0.1071462706,0.6613349319,-0.2454049438,0.4977247119,0.0151178716,-0.1120734289,0.2516205609,-0.0635674,-0.0569484457,-0.3780822754,-0.0096246395,0.4175055921,0.0937968418,-0.006041795,-0.2874678373,-0.1932771057,-0.2089736909,-0.0578821562,0.1327706873,0.217985943,0.0588070415,0.2713345587,0.5032314658,0.3485563695,-0.1059520543,-0.2050584853,-0.1715610176,0.135013476,-0.3857185841,-0.1006987318,0.1464199424,-0.1412298381,0.3286191225,0.0532120019,0.0755415633,0.3815326989,0.1520244032,-0.191182673,0.2819835246,0.1436703354,-0.0143271564,0.2014978528,0.4693849087,0.10144265,-0.1861768216,0.2664014995,-0.2484502494,-0.4411483705,0.0609503128,0.0914024562,-0.0530007631,-0.0377959199,-0.1594837457,0.1583796293,0.4205289483,-0.0675125122,0.0763092414,-0.4352326989,0.0418038815,0.1099921837,0.090025045,0.0778380185,-0.2328578383,-0.3426035345,0.3258585334,0.1530645639,-0.143343091,0.2205605805,-0.4285855889,0.2562271655,-0.0625739917,-0.1584211737,0.2098882198,-0.3840966821,-0.0710864812,-0.0994279832,-0.1054709256,-0.219163999,0.0696848631,-0.1317368597,-0.0530189537,-0.1544256955,-0.2207259983,0.5086032748,-0.0608968958,0.2324942946,0.0340902545,-0.0222444125,0.1885959953,0.1376521885,0.0248874463,0.1576400846,-0.320507586,0.5120714307,0.0987583771,-0.2441777289,0.2444199026,0.4102393985,-0.244097501,-0.1919647157,0.0468020588,0.2764771283,-0.0697960332,-0.148902759,0.1784002781,0.1963512599,-0.0604719073,0.0484519824,-0.2369389385,-0.0641728118,-0.2133747637,-0.1585211456,-0.0149899647,0.1101620644,0.0207734834,-0.0464728251,-0.2484817207,0.2723915279,-0.095426932,0.1349620223,-0.5259901881,0.2280209661,-0.2145735472,-0.0914447233,-0.0678843781,0.2231209576,0.2404938191,-0.2305194438,-0.3247486055,0.3036762774,-0.2122558355,0.0371924043,-0.3943189681,0.2457401305,0.3090952635,-0.5342994928,0.0032722938,0.2634411156,0.1639743596,-0.127455622,-0.383022517,0.1136109233,0.0757724941,0.3314516544,0.1057743058,0.3340388238,0.0536119379,-0.0884053633,-0.2020374686,-0.0776441023,-0.0154921887,-0.0838234276,0.1649821252,0.0597501025,-0.2288423777,-0.1983881891,0.4673170149,0.148841694,0.0216302704,0.1293901354,-0.1397807598,0.0306708906,0.1618632525,-0.0950354412,0.562422514,0.0324369743,-0.2305446267,0.1875147969,0.0697727501,-0.1226138026,0.4002860487,-0.0731957108,0.0252318047,-0.106548816,0.0615078025,0.0952067301,-0.1649064571,-0.289491564,0.25625211,0.2521972656,-0.4736778736,0.2136499882,0.0259740539,0.1696037799,-0.5800776482,-0.3827403784,0.0345899239,-0.2872623503,-0.1098969206,-0.0208649188,0.1248787344,0.4439495802,0.1025984958,0.072272338,0.2448716611,-0.158022657,-0.0618830435,0.0459858961,-0.1911002547,0.0143575687,-0.1871763021,-0.2879667282,-0.0504698642,-0.3237556815,-0.0552909076,-0.1035404056,-0.1793443263,0.1855938435,-0.2955768704,0.0113008786,0.3231825829,0.1871942282,-0.1992335618,-0.3188306391,0.4316528141,0.0009160679,0.0959324092,-0.0127255181,-0.0703296512,-0.2425712794,-0.0997440368,-0.4264292717,-0.0446449108,-0.1689539254,0.0303072091,-0.2177828997,-0.1313049346,0.4203010499,0.2371348441,0.0843030065,0.2997512817,0.1740505844,-0.0198321659,-0.2111925334,0.4776291847,-0.2067952901,-0.3576395214,-0.1271426529,-0.0129219079,0.2218758017,0.2340473831,-0.1764791906,-0.178256005,-0.2040158212,-0.1126637682,-0.3014701605,0.435983181,0.4326145947,0.0603171624,0.1686085016,-0.050865747,-0.5156285167,0.3171764314,0.1424028128,0.4169045985,-0.0929954201,0.4798959196,0.0866109133,0.7751232386,0.0111009004,-0.3219003677,0.260661304,-0.1630743444,0.2260283828,-0.1403172314,-0.4625391364,0.2768841088,0.2438232154,0.0316104107,-0.0728213787,-0.0910808519,-0.0698518232,-0.1445435584,0.2413931191,-0.1052515581,-0.152235657,0.141644001,-0.1808562279,0.5583443642,0.0370699987,0.3895744383,-0.2105321437,0.0093455995,0.0729864463,-0.1095194891,0.0688644648,0.1238157228,-0.1383382678,-0.5897532701,-0.1902554631,0.1065888256,-0.0117324274,0.3856185079,-0.1194277927,-0.0214349553,0.1513940543,0.1890711784,0.7310069203,0.0729371309,-0.3278804123,-0.0309844855,-0.2211033851,-0.6239109635,-0.0709218159,0.0394208506,-0.1653980464,-0.1467238814,0.1922552288,-0.1503565162,-0.1719058305,0.3554315269,0.0347531512,-0.2583989501,0.0776256397,-0.0642109886,-0.523164928,-0.3752950728,0.0817698017,0.2867938876,0.1763403565,0.2706166506,0.3705551624,-0.025202468,-0.0146308187,0.18252334,0.0229771715,0.2532839775,0.1200980097,-0.0711414665,0.2466866821,0.2663377821,0.0747498721,0.3784388602,0.0218263883,-0.1386081427,-0.0059653665,-0.0273004491,0.3917846084,0.3158438504,-0.1038671583,-0.2018494606,0.000328526,0.0675849691,-0.0518511422,0.2731793523,0.4600665271,0.0707793012,-0.1523297429,-0.5686461329,0.1935224235,0.1283531487,0.1165830716,0.0412615761,0.1834687442,-0.4985372424,0.3000656366,0.2441835999,0.9627013803,0.0045670066,0.4122143984,0.4066116214,-0.1718395501,-0.0448778979,-0.113391012,0.1925232708,-0.2689211369,-0.3740708232,-0.1267440766,-0.1576801091,0.1537520885,0.1447482109,-0.4471739829,0.2574018836,-0.1016518995,0.1324464083,-0.0517069362,-0.0063035074,-0.1191574186,-0.3786025941,-0.4207054675,0.0605765507,0.0085383756,0.2239290327,-0.021146439,-0.0518368334,-0.0725239962,-0.3276558518,-0.1989789903,0.1438138634,-0.3425531387,0.1162754223,0.0215491783,-0.3508206308,0.5728157759,0.476177901,0.5377926826,0.0480021685,-0.0448533259,0.2743342817,-0.2238800377,0.2336572111,0.0026795755,0.1144546121,0.4311787486,-0.0262231417,0.0106902737,0.2838855386,-0.1292874068,-0.0491685048,-0.1847702265,-0.2700214386,0.0675761774,-0.372892797,-0.2156337798,0.1479793638,-0.0138652613,-0.1213464215,0.0629487187,-0.1453512013,-0.2010597736,0.1748773158,-0.1926721185,-0.3175874949,-0.0471323766,0.5520834327,-0.0154436249,-0.0259577204,0.9083886147,0.144269377,0.0290533174,-0.0506250225,0.0027593984,0.5083619952,-0.7060852647,0.1902077049,0.0442791618,-0.1163032502,-0.2586606443,0.2225267291,0.0910592005,-0.1223974228,0.106648542,0.0268161763,-0.3815151155,0.0435700901,0.1826019138,0.1099413484,-0.006349633,0.4157488346,0.3691781163,-0.160919562,-0.2136099041,0.0854345262,0.0152923148,0.2436425686,0.1348816752,-0.1812804937,0.1046393886,-0.2158272713,0.0344824716,-0.0117325801,0.0288340431,-0.1036318764,-0.303047955,0.0960220471,0.1902741194,-0.1844199449,-0.0305881146,-0.3210200965,-0.2273693979,-0.2802014351,0.148508355,0.2875715792,-0.0231884699,0.4670304656,-0.0447168574,-0.1073348075,-0.1710858345,-0.3624248505,-0.055877734,-0.2588469684,0.2030044794,0.1647436172,0.0901968554,-0.0337181799,-0.5971805453,0.2267414331,0.2485940754,0.2303075343,0.0678243041,-0.2015693933,0.1524341106,0.1655961126,0.2326402515,-0.0341038108,-0.2902600765,0.0206330046,0.1774314642,0.0617807992,-0.2854016721,0.3496537805,-0.0410753638,0.0037135417,0.0104907202,0.2545004785,0.0580310784,-0.0115493983,0.0494867526,0.07942947,0.2100776881,-0.060544882,0.3345696926,0.7754502892,-0.2804121077,0.2012109309,0.2750228941,0.1458674669,0.2543017566,0.3205238581,-0.1166608557,0.344928503,-0.0188068897,-0.0199496411,-0.0652343482,-0.1228847727,-0.3104800284,-0.1817116886,0.0402105004,0.0242764875,0.0616917051,0.5748892426,-0.2628111541,-0.3483905494,-0.2398824245,0.27559039,-0.1278471202,-0.0154439704,-0.0172689799,0.0287494026,-0.3195879459,-0.1499734968,-0.2125629634,-0.1377667636,0.6361117959,0.0198255535,-0.0023895337,-0.1892204583,-0.0454767235,0.2628235817,0.0819954276,-0.1269645691,0.2411371469,0.1011123285,-0.1883557737,0.0976076052,0.4976045489,0.2201207876,0.2194078565,-0.0463458151,-0.0102585051,-0.0782288536,-0.07122951,0.1536559314,0.2663584054,-0.1075949073,0.3812067509,0.2045420259,0.0881867185,-0.0691197589,0.1516282856,-0.2430780083,0.311342746,-0.2424247712,0.5272917151,-0.0208063293,-0.1454467177,-0.1479764134,0.0642576069,-0.1720504463,0.1480787694,0.3423065841,0.007757517,0.1103841811,-0.1328207552,0.0115130777,0.2430801243,0.2766731679,0.5207222104,-0.2535755932,-0.3213942349,-0.2156719267,-0.3150944412,-0.0384822972,0.1294478774,0.0174163729,-0.0245442092,-0.1764499247,0.3041365147,-0.0779176876,0.0809708983,-0.1065159515,-0.1017209142,0.211242795,-0.0689955726,-0.3980531991,-0.1088816151,0.3790934682,-0.0227049384,-0.0859454423,-0.1128532961,0.4846835732,-0.0293582678,-0.369301945,0.0441125371,-0.1199799553,-0.1286889464,0.2373430282,0.0764206648,0.2086417675,0.0588344671,0.2778173089,0.0176058225,-0.2506176531,-0.3266896904,0.0686159506,-0.0857321024,0.4292593896,-0.2401910871,-0.1764764786,-0.4258554578,0.0071116677,0.1781212837,0.0464417078,-0.4590160847,0.4230368733,-0.367138505,0.0823734626,0.2416835129,0.3583958447,0.0034974401,0.1476848125,-0.3863488436,-0.4244388938,0.6188313961,-0.1814698875,-0.4834167063,0.1206571609,-0.1362357289,-0.0846379399,-0.1597628593,-0.5339651108,0.0789652839,0.1070881411,-0.2078465074,-0.3984488547,0.2419902831,0.1348467618,0.0488820933,-0.0512382947,0.4404880404,-0.0221921708,-0.202132076,-0.1810315847,-0.0351451151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1867","title":"ERROR WHEN USING SET_TRANSFORM() ","comments":"In the end, to make it work I turned to A-100 gpus instead of TPUS, among other changes. Set_transform doesn't work as expected and slows down training very much even in GPUs, and applying map destroys the disk, as it multiplies by 100 the size of the data passed to it (due to inefficient implementation converting strings to int64 floats I guess). For that reason, I chose to use datasets to load the data as text, and then edit the Collator from Transformers to tokenize every batch it receives before processing it. That way, I'm being able to train fast, without memory breaks, without the disk being unnecessarily filled, while making use of GPUs almost all the time I'm paying for them (the map function over the whole dataset took ~15hrs, in which you're not training at all). I hope this info helps others that are looking for training a language model from scratch cheaply, I'm going to close the issue as the optimal solution I found after many experiments to the problem posted in it is explained above. ","body":"Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n","comment_length":179,"text":"ERROR WHEN USING SET_TRANSFORM() \nHi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n\nIn the end, to make it work I turned to A-100 gpus instead of TPUS, among other changes. Set_transform doesn't work as expected and slows down training very much even in GPUs, and applying map destroys the disk, as it multiplies by 100 the size of the data passed to it (due to inefficient implementation converting strings to int64 floats I guess). For that reason, I chose to use datasets to load the data as text, and then edit the Collator from Transformers to tokenize every batch it receives before processing it. That way, I'm being able to train fast, without memory breaks, without the disk being unnecessarily filled, while making use of GPUs almost all the time I'm paying for them (the map function over the whole dataset took ~15hrs, in which you're not training at all). I hope this info helps others that are looking for training a language model from scratch cheaply, I'm going to close the issue as the optimal solution I found after many experiments to the problem posted in it is explained above. ","embeddings":[-0.1826189756,-0.2732517719,0.1393030435,0.1847398877,0.7292689681,0.1197050363,0.6009586453,0.2287764698,-0.3227846622,0.0544771031,0.2355112433,0.0333268568,-0.0883406326,-0.0234047007,0.1892216057,-0.5358201861,-0.0138219129,0.1288610101,-0.4634343982,-0.1831161827,-0.27669698,0.3134228885,-0.3415367901,0.1193406507,-0.3186003268,-0.0096082669,-0.047695376,0.016471222,0.0844151899,-0.1718301773,0.2334779501,-0.2800752223,0.1231957451,0.3974503875,-0.0001255372,0.1195578203,0.0606352575,-0.2890718281,-0.2367867827,-0.3003558218,-0.1437315047,-0.1552786827,-0.0154345548,-0.2408748716,-0.4479370713,-0.1890559047,0.0066991188,-0.2687551975,0.5653699636,0.7264817953,0.1215873957,0.2663597167,0.0996964872,-0.2382416725,-0.1261355132,0.1903086156,-0.1291115582,-0.0224978216,0.0317756385,0.2165266722,0.1336240321,0.1615502834,0.2006310076,-0.1752620935,0.4064543247,-0.0619789995,-0.1365289241,0.0435062647,0.0006193888,0.0104680918,0.2298866063,-0.3454308212,-0.420854032,-0.2735254467,0.1654442102,-0.2141008228,0.2700577974,0.1503871977,-0.066397734,0.22794348,-0.2600681484,-0.1098380312,-0.0588664711,0.0313304402,-0.1728429347,0.1623857617,-0.0065073655,0.1303315759,0.0931790695,0.1813862473,0.2739946544,0.1153549775,0.1876621246,0.3649872243,-0.265356034,-0.05478606,-0.1434788406,0.0108277686,-0.2889722586,-0.2492735982,-0.1723645478,-0.2085095048,0.0230882131,-0.0149517776,0.1071462706,0.6613349319,-0.2454049438,0.4977247119,0.0151178716,-0.1120734289,0.2516205609,-0.0635674,-0.0569484457,-0.3780822754,-0.0096246395,0.4175055921,0.0937968418,-0.006041795,-0.2874678373,-0.1932771057,-0.2089736909,-0.0578821562,0.1327706873,0.217985943,0.0588070415,0.2713345587,0.5032314658,0.3485563695,-0.1059520543,-0.2050584853,-0.1715610176,0.135013476,-0.3857185841,-0.1006987318,0.1464199424,-0.1412298381,0.3286191225,0.0532120019,0.0755415633,0.3815326989,0.1520244032,-0.191182673,0.2819835246,0.1436703354,-0.0143271564,0.2014978528,0.4693849087,0.10144265,-0.1861768216,0.2664014995,-0.2484502494,-0.4411483705,0.0609503128,0.0914024562,-0.0530007631,-0.0377959199,-0.1594837457,0.1583796293,0.4205289483,-0.0675125122,0.0763092414,-0.4352326989,0.0418038815,0.1099921837,0.090025045,0.0778380185,-0.2328578383,-0.3426035345,0.3258585334,0.1530645639,-0.143343091,0.2205605805,-0.4285855889,0.2562271655,-0.0625739917,-0.1584211737,0.2098882198,-0.3840966821,-0.0710864812,-0.0994279832,-0.1054709256,-0.219163999,0.0696848631,-0.1317368597,-0.0530189537,-0.1544256955,-0.2207259983,0.5086032748,-0.0608968958,0.2324942946,0.0340902545,-0.0222444125,0.1885959953,0.1376521885,0.0248874463,0.1576400846,-0.320507586,0.5120714307,0.0987583771,-0.2441777289,0.2444199026,0.4102393985,-0.244097501,-0.1919647157,0.0468020588,0.2764771283,-0.0697960332,-0.148902759,0.1784002781,0.1963512599,-0.0604719073,0.0484519824,-0.2369389385,-0.0641728118,-0.2133747637,-0.1585211456,-0.0149899647,0.1101620644,0.0207734834,-0.0464728251,-0.2484817207,0.2723915279,-0.095426932,0.1349620223,-0.5259901881,0.2280209661,-0.2145735472,-0.0914447233,-0.0678843781,0.2231209576,0.2404938191,-0.2305194438,-0.3247486055,0.3036762774,-0.2122558355,0.0371924043,-0.3943189681,0.2457401305,0.3090952635,-0.5342994928,0.0032722938,0.2634411156,0.1639743596,-0.127455622,-0.383022517,0.1136109233,0.0757724941,0.3314516544,0.1057743058,0.3340388238,0.0536119379,-0.0884053633,-0.2020374686,-0.0776441023,-0.0154921887,-0.0838234276,0.1649821252,0.0597501025,-0.2288423777,-0.1983881891,0.4673170149,0.148841694,0.0216302704,0.1293901354,-0.1397807598,0.0306708906,0.1618632525,-0.0950354412,0.562422514,0.0324369743,-0.2305446267,0.1875147969,0.0697727501,-0.1226138026,0.4002860487,-0.0731957108,0.0252318047,-0.106548816,0.0615078025,0.0952067301,-0.1649064571,-0.289491564,0.25625211,0.2521972656,-0.4736778736,0.2136499882,0.0259740539,0.1696037799,-0.5800776482,-0.3827403784,0.0345899239,-0.2872623503,-0.1098969206,-0.0208649188,0.1248787344,0.4439495802,0.1025984958,0.072272338,0.2448716611,-0.158022657,-0.0618830435,0.0459858961,-0.1911002547,0.0143575687,-0.1871763021,-0.2879667282,-0.0504698642,-0.3237556815,-0.0552909076,-0.1035404056,-0.1793443263,0.1855938435,-0.2955768704,0.0113008786,0.3231825829,0.1871942282,-0.1992335618,-0.3188306391,0.4316528141,0.0009160679,0.0959324092,-0.0127255181,-0.0703296512,-0.2425712794,-0.0997440368,-0.4264292717,-0.0446449108,-0.1689539254,0.0303072091,-0.2177828997,-0.1313049346,0.4203010499,0.2371348441,0.0843030065,0.2997512817,0.1740505844,-0.0198321659,-0.2111925334,0.4776291847,-0.2067952901,-0.3576395214,-0.1271426529,-0.0129219079,0.2218758017,0.2340473831,-0.1764791906,-0.178256005,-0.2040158212,-0.1126637682,-0.3014701605,0.435983181,0.4326145947,0.0603171624,0.1686085016,-0.050865747,-0.5156285167,0.3171764314,0.1424028128,0.4169045985,-0.0929954201,0.4798959196,0.0866109133,0.7751232386,0.0111009004,-0.3219003677,0.260661304,-0.1630743444,0.2260283828,-0.1403172314,-0.4625391364,0.2768841088,0.2438232154,0.0316104107,-0.0728213787,-0.0910808519,-0.0698518232,-0.1445435584,0.2413931191,-0.1052515581,-0.152235657,0.141644001,-0.1808562279,0.5583443642,0.0370699987,0.3895744383,-0.2105321437,0.0093455995,0.0729864463,-0.1095194891,0.0688644648,0.1238157228,-0.1383382678,-0.5897532701,-0.1902554631,0.1065888256,-0.0117324274,0.3856185079,-0.1194277927,-0.0214349553,0.1513940543,0.1890711784,0.7310069203,0.0729371309,-0.3278804123,-0.0309844855,-0.2211033851,-0.6239109635,-0.0709218159,0.0394208506,-0.1653980464,-0.1467238814,0.1922552288,-0.1503565162,-0.1719058305,0.3554315269,0.0347531512,-0.2583989501,0.0776256397,-0.0642109886,-0.523164928,-0.3752950728,0.0817698017,0.2867938876,0.1763403565,0.2706166506,0.3705551624,-0.025202468,-0.0146308187,0.18252334,0.0229771715,0.2532839775,0.1200980097,-0.0711414665,0.2466866821,0.2663377821,0.0747498721,0.3784388602,0.0218263883,-0.1386081427,-0.0059653665,-0.0273004491,0.3917846084,0.3158438504,-0.1038671583,-0.2018494606,0.000328526,0.0675849691,-0.0518511422,0.2731793523,0.4600665271,0.0707793012,-0.1523297429,-0.5686461329,0.1935224235,0.1283531487,0.1165830716,0.0412615761,0.1834687442,-0.4985372424,0.3000656366,0.2441835999,0.9627013803,0.0045670066,0.4122143984,0.4066116214,-0.1718395501,-0.0448778979,-0.113391012,0.1925232708,-0.2689211369,-0.3740708232,-0.1267440766,-0.1576801091,0.1537520885,0.1447482109,-0.4471739829,0.2574018836,-0.1016518995,0.1324464083,-0.0517069362,-0.0063035074,-0.1191574186,-0.3786025941,-0.4207054675,0.0605765507,0.0085383756,0.2239290327,-0.021146439,-0.0518368334,-0.0725239962,-0.3276558518,-0.1989789903,0.1438138634,-0.3425531387,0.1162754223,0.0215491783,-0.3508206308,0.5728157759,0.476177901,0.5377926826,0.0480021685,-0.0448533259,0.2743342817,-0.2238800377,0.2336572111,0.0026795755,0.1144546121,0.4311787486,-0.0262231417,0.0106902737,0.2838855386,-0.1292874068,-0.0491685048,-0.1847702265,-0.2700214386,0.0675761774,-0.372892797,-0.2156337798,0.1479793638,-0.0138652613,-0.1213464215,0.0629487187,-0.1453512013,-0.2010597736,0.1748773158,-0.1926721185,-0.3175874949,-0.0471323766,0.5520834327,-0.0154436249,-0.0259577204,0.9083886147,0.144269377,0.0290533174,-0.0506250225,0.0027593984,0.5083619952,-0.7060852647,0.1902077049,0.0442791618,-0.1163032502,-0.2586606443,0.2225267291,0.0910592005,-0.1223974228,0.106648542,0.0268161763,-0.3815151155,0.0435700901,0.1826019138,0.1099413484,-0.006349633,0.4157488346,0.3691781163,-0.160919562,-0.2136099041,0.0854345262,0.0152923148,0.2436425686,0.1348816752,-0.1812804937,0.1046393886,-0.2158272713,0.0344824716,-0.0117325801,0.0288340431,-0.1036318764,-0.303047955,0.0960220471,0.1902741194,-0.1844199449,-0.0305881146,-0.3210200965,-0.2273693979,-0.2802014351,0.148508355,0.2875715792,-0.0231884699,0.4670304656,-0.0447168574,-0.1073348075,-0.1710858345,-0.3624248505,-0.055877734,-0.2588469684,0.2030044794,0.1647436172,0.0901968554,-0.0337181799,-0.5971805453,0.2267414331,0.2485940754,0.2303075343,0.0678243041,-0.2015693933,0.1524341106,0.1655961126,0.2326402515,-0.0341038108,-0.2902600765,0.0206330046,0.1774314642,0.0617807992,-0.2854016721,0.3496537805,-0.0410753638,0.0037135417,0.0104907202,0.2545004785,0.0580310784,-0.0115493983,0.0494867526,0.07942947,0.2100776881,-0.060544882,0.3345696926,0.7754502892,-0.2804121077,0.2012109309,0.2750228941,0.1458674669,0.2543017566,0.3205238581,-0.1166608557,0.344928503,-0.0188068897,-0.0199496411,-0.0652343482,-0.1228847727,-0.3104800284,-0.1817116886,0.0402105004,0.0242764875,0.0616917051,0.5748892426,-0.2628111541,-0.3483905494,-0.2398824245,0.27559039,-0.1278471202,-0.0154439704,-0.0172689799,0.0287494026,-0.3195879459,-0.1499734968,-0.2125629634,-0.1377667636,0.6361117959,0.0198255535,-0.0023895337,-0.1892204583,-0.0454767235,0.2628235817,0.0819954276,-0.1269645691,0.2411371469,0.1011123285,-0.1883557737,0.0976076052,0.4976045489,0.2201207876,0.2194078565,-0.0463458151,-0.0102585051,-0.0782288536,-0.07122951,0.1536559314,0.2663584054,-0.1075949073,0.3812067509,0.2045420259,0.0881867185,-0.0691197589,0.1516282856,-0.2430780083,0.311342746,-0.2424247712,0.5272917151,-0.0208063293,-0.1454467177,-0.1479764134,0.0642576069,-0.1720504463,0.1480787694,0.3423065841,0.007757517,0.1103841811,-0.1328207552,0.0115130777,0.2430801243,0.2766731679,0.5207222104,-0.2535755932,-0.3213942349,-0.2156719267,-0.3150944412,-0.0384822972,0.1294478774,0.0174163729,-0.0245442092,-0.1764499247,0.3041365147,-0.0779176876,0.0809708983,-0.1065159515,-0.1017209142,0.211242795,-0.0689955726,-0.3980531991,-0.1088816151,0.3790934682,-0.0227049384,-0.0859454423,-0.1128532961,0.4846835732,-0.0293582678,-0.369301945,0.0441125371,-0.1199799553,-0.1286889464,0.2373430282,0.0764206648,0.2086417675,0.0588344671,0.2778173089,0.0176058225,-0.2506176531,-0.3266896904,0.0686159506,-0.0857321024,0.4292593896,-0.2401910871,-0.1764764786,-0.4258554578,0.0071116677,0.1781212837,0.0464417078,-0.4590160847,0.4230368733,-0.367138505,0.0823734626,0.2416835129,0.3583958447,0.0034974401,0.1476848125,-0.3863488436,-0.4244388938,0.6188313961,-0.1814698875,-0.4834167063,0.1206571609,-0.1362357289,-0.0846379399,-0.1597628593,-0.5339651108,0.0789652839,0.1070881411,-0.2078465074,-0.3984488547,0.2419902831,0.1348467618,0.0488820933,-0.0512382947,0.4404880404,-0.0221921708,-0.202132076,-0.1810315847,-0.0351451151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1867","title":"ERROR WHEN USING SET_TRANSFORM() ","comments":"Great comment @alexvaca0 . I think that we could re-open the issue as a reformulation of why it takes so much space to save the arrow. Saving a 1% of oscar corpus takes more thank 600 GB (it breaks when it pass 600GB because it is the free memory that I have at this moment) when the full dataset is 1,3 TB. I have a 1TB M.2 NVMe disk that I can not train on because the saved .arrow files goes crazily big. If you can share your Collator I will be grateful. ","body":"Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n","comment_length":93,"text":"ERROR WHEN USING SET_TRANSFORM() \nHi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1825#issuecomment-774202797\r\n\r\nHowever, when I try to use Trainer from transformers with such dataset, it throws an error:\r\n\r\n```\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n[INFO|trainer.py:357] 2021-02-12 10:18:09,893 >> The following columns in the training set don't have a corresponding argument in `AlbertForMaskedLM.forward` and have been ignored: text.\r\nException in device=TPU:0: __init__() missing 1 required positional argument: 'transform'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 330, in _mp_start_fn\r\n _start_fn(index, pf_cfg, fn, args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 324, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 368, in _mp_fn\r\n main()\r\n File \"\/home\/alejandro_vaca\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 332, in main\r\n data_collator=data_collator,\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 286, in __init__\r\n self._remove_unused_columns(self.train_dataset, description=\"training\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.7\/lib\/python3.6\/site-packages\/transformers\/trainer.py\", line 359, in _remove_unused_columns\r\n dataset.set_format(type=dataset.format[\"type\"], columns=columns)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/fingerprint.py\", line 312, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/arrow_dataset.py\", line 818, in set_format\r\n _ = get_formatter(type, **format_kwargs)\r\n File \"\/home\/alejandro_vaca\/datasets\/src\/datasets\/formatting\/__init__.py\", line 112, in get_formatter\r\n return _FORMAT_TYPES[format_type](**format_kwargs)\r\nTypeError: __init__() missing 1 required positional argument: 'transform'\r\n```\r\n\r\nThe code I'm using:\r\n\r\n```{python}\r\n\r\n def tokenize_function(examples):\r\n # Remove empty lines\r\n examples[\"text\"] = [line for line in examples[\"text\"] if len(line) > 0 and not line.isspace()]\r\n return tokenizer(examples[\"text\"], padding=padding, truncation=True, max_length=data_args.max_seq_length)\r\n\r\n datasets.set_transform(tokenize_function)\r\n\r\n data_collator = DataCollatorForWholeWordMask(tokenizer=tokenizer, mlm_probability=data_args.mlm_probability)\r\n\r\n # Initialize our Trainer\r\n trainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n train_dataset=datasets[\"train\"] if training_args.do_train else None,\r\n eval_dataset=datasets[\"val\"] if training_args.do_eval else None,\r\n tokenizer=tokenizer,\r\n data_collator=data_collator,\r\n )\r\n```\r\n\r\nI've installed from source, master branch.\r\n\nGreat comment @alexvaca0 . I think that we could re-open the issue as a reformulation of why it takes so much space to save the arrow. Saving a 1% of oscar corpus takes more thank 600 GB (it breaks when it pass 600GB because it is the free memory that I have at this moment) when the full dataset is 1,3 TB. I have a 1TB M.2 NVMe disk that I can not train on because the saved .arrow files goes crazily big. If you can share your Collator I will be grateful. ","embeddings":[-0.1826189756,-0.2732517719,0.1393030435,0.1847398877,0.7292689681,0.1197050363,0.6009586453,0.2287764698,-0.3227846622,0.0544771031,0.2355112433,0.0333268568,-0.0883406326,-0.0234047007,0.1892216057,-0.5358201861,-0.0138219129,0.1288610101,-0.4634343982,-0.1831161827,-0.27669698,0.3134228885,-0.3415367901,0.1193406507,-0.3186003268,-0.0096082669,-0.047695376,0.016471222,0.0844151899,-0.1718301773,0.2334779501,-0.2800752223,0.1231957451,0.3974503875,-0.0001255372,0.1195578203,0.0606352575,-0.2890718281,-0.2367867827,-0.3003558218,-0.1437315047,-0.1552786827,-0.0154345548,-0.2408748716,-0.4479370713,-0.1890559047,0.0066991188,-0.2687551975,0.5653699636,0.7264817953,0.1215873957,0.2663597167,0.0996964872,-0.2382416725,-0.1261355132,0.1903086156,-0.1291115582,-0.0224978216,0.0317756385,0.2165266722,0.1336240321,0.1615502834,0.2006310076,-0.1752620935,0.4064543247,-0.0619789995,-0.1365289241,0.0435062647,0.0006193888,0.0104680918,0.2298866063,-0.3454308212,-0.420854032,-0.2735254467,0.1654442102,-0.2141008228,0.2700577974,0.1503871977,-0.066397734,0.22794348,-0.2600681484,-0.1098380312,-0.0588664711,0.0313304402,-0.1728429347,0.1623857617,-0.0065073655,0.1303315759,0.0931790695,0.1813862473,0.2739946544,0.1153549775,0.1876621246,0.3649872243,-0.265356034,-0.05478606,-0.1434788406,0.0108277686,-0.2889722586,-0.2492735982,-0.1723645478,-0.2085095048,0.0230882131,-0.0149517776,0.1071462706,0.6613349319,-0.2454049438,0.4977247119,0.0151178716,-0.1120734289,0.2516205609,-0.0635674,-0.0569484457,-0.3780822754,-0.0096246395,0.4175055921,0.0937968418,-0.006041795,-0.2874678373,-0.1932771057,-0.2089736909,-0.0578821562,0.1327706873,0.217985943,0.0588070415,0.2713345587,0.5032314658,0.3485563695,-0.1059520543,-0.2050584853,-0.1715610176,0.135013476,-0.3857185841,-0.1006987318,0.1464199424,-0.1412298381,0.3286191225,0.0532120019,0.0755415633,0.3815326989,0.1520244032,-0.191182673,0.2819835246,0.1436703354,-0.0143271564,0.2014978528,0.4693849087,0.10144265,-0.1861768216,0.2664014995,-0.2484502494,-0.4411483705,0.0609503128,0.0914024562,-0.0530007631,-0.0377959199,-0.1594837457,0.1583796293,0.4205289483,-0.0675125122,0.0763092414,-0.4352326989,0.0418038815,0.1099921837,0.090025045,0.0778380185,-0.2328578383,-0.3426035345,0.3258585334,0.1530645639,-0.143343091,0.2205605805,-0.4285855889,0.2562271655,-0.0625739917,-0.1584211737,0.2098882198,-0.3840966821,-0.0710864812,-0.0994279832,-0.1054709256,-0.219163999,0.0696848631,-0.1317368597,-0.0530189537,-0.1544256955,-0.2207259983,0.5086032748,-0.0608968958,0.2324942946,0.0340902545,-0.0222444125,0.1885959953,0.1376521885,0.0248874463,0.1576400846,-0.320507586,0.5120714307,0.0987583771,-0.2441777289,0.2444199026,0.4102393985,-0.244097501,-0.1919647157,0.0468020588,0.2764771283,-0.0697960332,-0.148902759,0.1784002781,0.1963512599,-0.0604719073,0.0484519824,-0.2369389385,-0.0641728118,-0.2133747637,-0.1585211456,-0.0149899647,0.1101620644,0.0207734834,-0.0464728251,-0.2484817207,0.2723915279,-0.095426932,0.1349620223,-0.5259901881,0.2280209661,-0.2145735472,-0.0914447233,-0.0678843781,0.2231209576,0.2404938191,-0.2305194438,-0.3247486055,0.3036762774,-0.2122558355,0.0371924043,-0.3943189681,0.2457401305,0.3090952635,-0.5342994928,0.0032722938,0.2634411156,0.1639743596,-0.127455622,-0.383022517,0.1136109233,0.0757724941,0.3314516544,0.1057743058,0.3340388238,0.0536119379,-0.0884053633,-0.2020374686,-0.0776441023,-0.0154921887,-0.0838234276,0.1649821252,0.0597501025,-0.2288423777,-0.1983881891,0.4673170149,0.148841694,0.0216302704,0.1293901354,-0.1397807598,0.0306708906,0.1618632525,-0.0950354412,0.562422514,0.0324369743,-0.2305446267,0.1875147969,0.0697727501,-0.1226138026,0.4002860487,-0.0731957108,0.0252318047,-0.106548816,0.0615078025,0.0952067301,-0.1649064571,-0.289491564,0.25625211,0.2521972656,-0.4736778736,0.2136499882,0.0259740539,0.1696037799,-0.5800776482,-0.3827403784,0.0345899239,-0.2872623503,-0.1098969206,-0.0208649188,0.1248787344,0.4439495802,0.1025984958,0.072272338,0.2448716611,-0.158022657,-0.0618830435,0.0459858961,-0.1911002547,0.0143575687,-0.1871763021,-0.2879667282,-0.0504698642,-0.3237556815,-0.0552909076,-0.1035404056,-0.1793443263,0.1855938435,-0.2955768704,0.0113008786,0.3231825829,0.1871942282,-0.1992335618,-0.3188306391,0.4316528141,0.0009160679,0.0959324092,-0.0127255181,-0.0703296512,-0.2425712794,-0.0997440368,-0.4264292717,-0.0446449108,-0.1689539254,0.0303072091,-0.2177828997,-0.1313049346,0.4203010499,0.2371348441,0.0843030065,0.2997512817,0.1740505844,-0.0198321659,-0.2111925334,0.4776291847,-0.2067952901,-0.3576395214,-0.1271426529,-0.0129219079,0.2218758017,0.2340473831,-0.1764791906,-0.178256005,-0.2040158212,-0.1126637682,-0.3014701605,0.435983181,0.4326145947,0.0603171624,0.1686085016,-0.050865747,-0.5156285167,0.3171764314,0.1424028128,0.4169045985,-0.0929954201,0.4798959196,0.0866109133,0.7751232386,0.0111009004,-0.3219003677,0.260661304,-0.1630743444,0.2260283828,-0.1403172314,-0.4625391364,0.2768841088,0.2438232154,0.0316104107,-0.0728213787,-0.0910808519,-0.0698518232,-0.1445435584,0.2413931191,-0.1052515581,-0.152235657,0.141644001,-0.1808562279,0.5583443642,0.0370699987,0.3895744383,-0.2105321437,0.0093455995,0.0729864463,-0.1095194891,0.0688644648,0.1238157228,-0.1383382678,-0.5897532701,-0.1902554631,0.1065888256,-0.0117324274,0.3856185079,-0.1194277927,-0.0214349553,0.1513940543,0.1890711784,0.7310069203,0.0729371309,-0.3278804123,-0.0309844855,-0.2211033851,-0.6239109635,-0.0709218159,0.0394208506,-0.1653980464,-0.1467238814,0.1922552288,-0.1503565162,-0.1719058305,0.3554315269,0.0347531512,-0.2583989501,0.0776256397,-0.0642109886,-0.523164928,-0.3752950728,0.0817698017,0.2867938876,0.1763403565,0.2706166506,0.3705551624,-0.025202468,-0.0146308187,0.18252334,0.0229771715,0.2532839775,0.1200980097,-0.0711414665,0.2466866821,0.2663377821,0.0747498721,0.3784388602,0.0218263883,-0.1386081427,-0.0059653665,-0.0273004491,0.3917846084,0.3158438504,-0.1038671583,-0.2018494606,0.000328526,0.0675849691,-0.0518511422,0.2731793523,0.4600665271,0.0707793012,-0.1523297429,-0.5686461329,0.1935224235,0.1283531487,0.1165830716,0.0412615761,0.1834687442,-0.4985372424,0.3000656366,0.2441835999,0.9627013803,0.0045670066,0.4122143984,0.4066116214,-0.1718395501,-0.0448778979,-0.113391012,0.1925232708,-0.2689211369,-0.3740708232,-0.1267440766,-0.1576801091,0.1537520885,0.1447482109,-0.4471739829,0.2574018836,-0.1016518995,0.1324464083,-0.0517069362,-0.0063035074,-0.1191574186,-0.3786025941,-0.4207054675,0.0605765507,0.0085383756,0.2239290327,-0.021146439,-0.0518368334,-0.0725239962,-0.3276558518,-0.1989789903,0.1438138634,-0.3425531387,0.1162754223,0.0215491783,-0.3508206308,0.5728157759,0.476177901,0.5377926826,0.0480021685,-0.0448533259,0.2743342817,-0.2238800377,0.2336572111,0.0026795755,0.1144546121,0.4311787486,-0.0262231417,0.0106902737,0.2838855386,-0.1292874068,-0.0491685048,-0.1847702265,-0.2700214386,0.0675761774,-0.372892797,-0.2156337798,0.1479793638,-0.0138652613,-0.1213464215,0.0629487187,-0.1453512013,-0.2010597736,0.1748773158,-0.1926721185,-0.3175874949,-0.0471323766,0.5520834327,-0.0154436249,-0.0259577204,0.9083886147,0.144269377,0.0290533174,-0.0506250225,0.0027593984,0.5083619952,-0.7060852647,0.1902077049,0.0442791618,-0.1163032502,-0.2586606443,0.2225267291,0.0910592005,-0.1223974228,0.106648542,0.0268161763,-0.3815151155,0.0435700901,0.1826019138,0.1099413484,-0.006349633,0.4157488346,0.3691781163,-0.160919562,-0.2136099041,0.0854345262,0.0152923148,0.2436425686,0.1348816752,-0.1812804937,0.1046393886,-0.2158272713,0.0344824716,-0.0117325801,0.0288340431,-0.1036318764,-0.303047955,0.0960220471,0.1902741194,-0.1844199449,-0.0305881146,-0.3210200965,-0.2273693979,-0.2802014351,0.148508355,0.2875715792,-0.0231884699,0.4670304656,-0.0447168574,-0.1073348075,-0.1710858345,-0.3624248505,-0.055877734,-0.2588469684,0.2030044794,0.1647436172,0.0901968554,-0.0337181799,-0.5971805453,0.2267414331,0.2485940754,0.2303075343,0.0678243041,-0.2015693933,0.1524341106,0.1655961126,0.2326402515,-0.0341038108,-0.2902600765,0.0206330046,0.1774314642,0.0617807992,-0.2854016721,0.3496537805,-0.0410753638,0.0037135417,0.0104907202,0.2545004785,0.0580310784,-0.0115493983,0.0494867526,0.07942947,0.2100776881,-0.060544882,0.3345696926,0.7754502892,-0.2804121077,0.2012109309,0.2750228941,0.1458674669,0.2543017566,0.3205238581,-0.1166608557,0.344928503,-0.0188068897,-0.0199496411,-0.0652343482,-0.1228847727,-0.3104800284,-0.1817116886,0.0402105004,0.0242764875,0.0616917051,0.5748892426,-0.2628111541,-0.3483905494,-0.2398824245,0.27559039,-0.1278471202,-0.0154439704,-0.0172689799,0.0287494026,-0.3195879459,-0.1499734968,-0.2125629634,-0.1377667636,0.6361117959,0.0198255535,-0.0023895337,-0.1892204583,-0.0454767235,0.2628235817,0.0819954276,-0.1269645691,0.2411371469,0.1011123285,-0.1883557737,0.0976076052,0.4976045489,0.2201207876,0.2194078565,-0.0463458151,-0.0102585051,-0.0782288536,-0.07122951,0.1536559314,0.2663584054,-0.1075949073,0.3812067509,0.2045420259,0.0881867185,-0.0691197589,0.1516282856,-0.2430780083,0.311342746,-0.2424247712,0.5272917151,-0.0208063293,-0.1454467177,-0.1479764134,0.0642576069,-0.1720504463,0.1480787694,0.3423065841,0.007757517,0.1103841811,-0.1328207552,0.0115130777,0.2430801243,0.2766731679,0.5207222104,-0.2535755932,-0.3213942349,-0.2156719267,-0.3150944412,-0.0384822972,0.1294478774,0.0174163729,-0.0245442092,-0.1764499247,0.3041365147,-0.0779176876,0.0809708983,-0.1065159515,-0.1017209142,0.211242795,-0.0689955726,-0.3980531991,-0.1088816151,0.3790934682,-0.0227049384,-0.0859454423,-0.1128532961,0.4846835732,-0.0293582678,-0.369301945,0.0441125371,-0.1199799553,-0.1286889464,0.2373430282,0.0764206648,0.2086417675,0.0588344671,0.2778173089,0.0176058225,-0.2506176531,-0.3266896904,0.0686159506,-0.0857321024,0.4292593896,-0.2401910871,-0.1764764786,-0.4258554578,0.0071116677,0.1781212837,0.0464417078,-0.4590160847,0.4230368733,-0.367138505,0.0823734626,0.2416835129,0.3583958447,0.0034974401,0.1476848125,-0.3863488436,-0.4244388938,0.6188313961,-0.1814698875,-0.4834167063,0.1206571609,-0.1362357289,-0.0846379399,-0.1597628593,-0.5339651108,0.0789652839,0.1070881411,-0.2078465074,-0.3984488547,0.2419902831,0.1348467618,0.0488820933,-0.0512382947,0.4404880404,-0.0221921708,-0.202132076,-0.1810315847,-0.0351451151]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1864","title":"Add Winogender Schemas","comments":"Nevermind, this one is already available on the hub under the name `'wino_bias'`: https:\/\/huggingface.co\/datasets\/wino_bias","body":"## Adding a Dataset\r\n- **Name:** Winogender Schemas\r\n- **Description:** Winogender Schemas (inspired by Winograd Schemas) are minimal pairs of sentences that differ only by the gender of one pronoun in the sentence, designed to test for the presence of gender bias in automated coreference resolution systems.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1804.09301\r\n- **Data:** https:\/\/github.com\/rudinger\/winogender-schemas (see data directory)\r\n- **Motivation:** Testing gender bias in automated coreference resolution systems, improve coreference resolution in general.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":14,"text":"Add Winogender Schemas\n## Adding a Dataset\r\n- **Name:** Winogender Schemas\r\n- **Description:** Winogender Schemas (inspired by Winograd Schemas) are minimal pairs of sentences that differ only by the gender of one pronoun in the sentence, designed to test for the presence of gender bias in automated coreference resolution systems.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1804.09301\r\n- **Data:** https:\/\/github.com\/rudinger\/winogender-schemas (see data directory)\r\n- **Motivation:** Testing gender bias in automated coreference resolution systems, improve coreference resolution in general.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nNevermind, this one is already available on the hub under the name `'wino_bias'`: https:\/\/huggingface.co\/datasets\/wino_bias","embeddings":[-0.3111324012,-0.1223415881,-0.1506437361,-0.1477670074,0.0376245156,-0.2091957033,0.2435115278,0.0809530094,0.1477920413,0.29605636,0.0450613946,0.0423981398,-0.1369290799,0.0410834886,0.0158626288,-0.1046132445,-0.1254138798,0.3697438836,0.1942304224,-0.0442929864,0.1058407128,0.0166977737,0.0456707701,0.1576350778,-0.1911803037,-0.2512027323,-0.0436748378,0.1509488076,-0.3451169431,-0.1797563732,0.1306218803,0.4377282262,-0.375254184,0.1767809689,-0.0001014193,-0.19611983,0.0849243477,0.000093535,-0.0829663351,0.1911304146,0.0905902311,-0.0026722699,-0.1904382557,-0.1853649318,-0.2259798497,0.1977828294,0.0962058902,0.0201797727,0.2737496197,0.2469010949,0.2721018791,0.434471637,-0.0995626152,-0.2098513991,0.0767075568,0.2507292032,-0.046547357,0.1322425455,0.1439907998,-0.0031622085,-0.4140337706,0.2618872821,-0.006104318,0.0610825866,-0.2525632381,-0.0522227325,0.2747611701,-0.2779034972,0.1135258377,0.2270492613,-0.0493161865,-0.206888482,-0.2641969919,-0.337533474,0.100484997,-0.4982972443,0.1866339296,0.0934517607,-0.2442584038,-0.0246095527,-0.2139632255,-0.2424431294,-0.055694215,0.0854664743,-0.0561057255,0.3870002031,0.005024083,-0.1379666477,0.0117822979,-0.0382353738,0.0093054157,-0.1260811836,-0.2116778195,-0.1458807737,-0.2278147042,-0.3566716611,-0.0167989973,0.2560401559,0.2135466635,-0.1246084645,-0.085010916,0.1706413627,-0.1032040864,0.008101848,0.0521219894,0.240571484,0.367584765,-0.005828701,0.0902784243,-0.0641364381,-0.008207906,0.2162537128,0.1372815967,-0.2796420753,-0.1100750938,0.1492447257,-0.1685900539,0.1466801614,-0.4264695942,-0.0442305654,0.0416251719,0.013395939,0.2262862474,0.3273386657,0.226290077,0.3769505024,0.1604203135,0.2802196145,-0.1494142413,-0.5447602868,-0.2183998376,0.1096804067,-0.2626082301,0.1605751812,0.3425990641,0.0844990164,0.21081011,-0.0413803756,0.0927007869,-0.1342148483,-0.2443131804,-0.0059679556,0.2002988458,-0.256362766,0.021456778,-0.0909658372,0.0267593265,-0.2415896952,-0.1331943572,-0.2676970363,0.1987241805,0.0296477526,-0.179957062,0.2358525991,-0.1113438904,-0.0523003675,0.2154138833,0.605085969,-0.0183624737,-0.011309945,0.1092552766,0.0821021497,-0.4612615108,0.0686738342,0.1961282194,0.3109408319,-0.0556349605,-0.0488877706,0.237180531,0.0171817299,0.0449483991,-0.0959973931,0.1836510599,0.257614702,-0.0450536273,0.1543824524,0.0180860553,-0.4368788898,-0.3538578153,-0.3431141675,-0.4318683147,0.0876902044,0.0839271173,-0.0406520329,0.2525442541,0.0356668718,0.1052397713,0.162624836,0.0160058104,0.1528294086,-0.1808015555,-0.2751531899,-0.3593993783,0.0784354135,0.0296072979,-0.207965821,-0.0022577415,0.0385518223,0.3904957175,-0.3577960134,0.1356218159,-0.2303302586,0.062954925,-0.0924325883,-0.1055393293,-0.4670884907,-0.0573837906,-0.1094674096,-0.0078103808,0.096732229,0.3186022043,-0.2921331525,-0.2123345286,-0.0663043857,-0.0252415296,0.142738387,0.3464359641,0.2789817154,0.1728418618,-0.0479547307,0.1281430572,0.0844777077,-0.4225111306,-0.0063451445,0.1094993204,0.06258367,0.0278679002,-0.2169515342,0.1091244668,0.3236815035,-0.2097385675,0.1422026902,0.0463907458,0.2946658432,-0.0563195311,0.4445663989,0.1092790067,0.1640347689,0.0428922549,-0.2279420942,0.0084107099,0.0546104722,0.0119560575,0.1291141212,-0.1322940737,0.2825600505,0.4328561723,-0.2643295825,-0.1759186089,0.132048443,0.3613295853,-0.1423321515,-0.0836176574,-0.4381397665,0.2614216506,0.0685280114,-0.3585900664,-0.1624337733,-0.2987911701,0.1554536968,0.5109546781,0.0168273151,0.5802515745,0.0332452431,-0.2878004313,-0.1192048118,-0.2460386902,-0.1612318754,0.0462158248,0.314997673,-0.0732244998,-0.0995751992,0.0573353879,-0.1053054184,0.2988854945,-0.0100046098,-0.1139703766,0.234041661,0.203125298,-0.103056252,-0.5978714824,-0.056449227,-0.2409467995,-0.2740018368,-0.2643963099,-0.2594861984,0.1638227105,0.1534384638,0.1212752238,0.0346593596,0.1855064034,-0.0138310799,0.5186085105,0.2829941809,-0.1945309788,0.1479712725,0.0317986012,0.7553442717,0.0993117541,0.2731579244,-0.0918737128,-0.24257797,0.0150182322,0.1890481561,0.1120913401,-0.0880817622,0.4231035411,-0.3451015353,0.3203162551,-0.2348245233,-0.526863873,0.0826109871,-0.0873650312,0.2867395878,0.2726042569,-0.0926693454,-0.0624623783,-0.3587361276,0.0415161625,-0.2615879774,-0.268761456,0.0534969456,-0.0348222964,-0.1775320321,-0.3194874525,-0.1461620778,-0.1656021476,-0.1957797855,0.2701155841,0.011756123,0.167135343,0.2066203803,-0.1129849404,-0.0382451303,-0.4261061847,0.3350615203,-0.2828384936,-0.2180576473,-0.0061382339,-0.2324671745,-0.2410275489,-0.4036731124,-0.4692922831,0.097107254,-0.1955397278,-0.2056851089,-0.1532286257,-0.1019081548,0.0910596177,0.2588322461,-0.0602737255,0.3999316096,0.0386020988,-0.2263841927,-0.2445722222,-0.0168806687,0.0707095787,0.1074107736,-0.0266543329,-0.1318338215,-0.2643313706,-0.0061836634,0.1259480715,0.014280661,-0.2783602774,0.2787393034,-0.0875231251,0.1093911827,0.0635337755,-0.2796401978,0.2552881241,-0.0886585042,0.24581635,0.7867630124,-0.1203080863,-0.3230628967,0.0035955862,-0.227496177,-0.3736884892,-0.3063054979,0.0817212239,-0.210654676,0.0689733252,-0.0791146755,-0.0376953632,-0.2599276304,-0.3228301406,0.1977745295,-0.1191091463,-0.0947799683,-0.3039632142,-0.2032380253,-0.0167098586,-0.2105005234,0.2073528618,0.089390479,0.0175199676,-0.0937194228,-0.0086438088,-0.0814527571,-0.0983941108,0.2527912557,-0.3646980524,0.2847574949,-0.1952623874,0.1934806108,-0.1234177351,0.2706457675,-0.4318377972,-0.0072515444,0.1618883461,0.2619520128,-0.4123615026,0.0655792803,0.1505727321,-0.0467719585,-0.203610763,-0.1377185732,-0.063316226,0.2652275562,-0.0332424678,0.4958551824,-0.0222369023,0.2181075662,-0.1366543174,-0.0859582648,-0.127258122,0.1477032304,0.400913775,0.5024082065,0.088701725,0.0749925971,-0.1889169216,0.142340526,0.2925342023,-0.152052328,-0.0061651897,0.1624269933,-0.0305358153,-0.0528322272,-0.1393069476,0.3142230213,0.2774372697,0.1487404853,-0.0160676222,-0.3090571761,0.0873590037,-0.6388782263,0.0586519502,0.3420395553,0.2586528957,0.1310645193,0.0950577557,0.114945516,0.0768198743,-0.2748823166,0.2212093621,0.6285663843,-0.3871157765,0.2640437484,0.1633511633,0.9728955626,0.2253869623,0.0993081182,0.1366998106,-0.0715045854,0.3681707978,-0.0924856961,0.225133121,-0.0157958288,-0.0300706662,0.099915646,0.1469020694,0.0601706915,0.1502644569,-0.6459121108,-0.2055728436,0.0260414761,0.0366094075,0.1789438725,0.182380721,-0.2642895579,-0.3666077256,-0.2170652151,0.2816362679,0.1240946874,-0.1898097992,0.0129313758,0.1507308185,-0.0894152224,-0.2285475731,-0.2225745171,0.0155598046,-0.0858448148,-0.2532702684,-0.133164987,0.2070593834,0.2232525498,0.2565119565,0.3936180472,0.340566963,-0.1693143994,0.1566661,0.0584357567,0.1262791008,0.227088213,-0.0424542874,0.0344721004,0.0418022238,-0.2898930311,0.0213531666,0.1185859367,-0.0082485452,-0.0888468921,-0.0914695039,0.0658241808,-0.1928819716,0.0023745131,0.0842379108,-0.1540639251,-0.2085360438,0.216037482,0.0087347394,-0.1505924165,0.1868311465,0.2121028453,-0.1489987671,0.0339090787,0.010153627,-0.0705272928,0.1477279216,0.2812233865,-0.1863527447,-0.2292063832,-0.5011111498,-0.3078337908,0.0528512262,-0.4250307679,-0.0668775737,0.1293718219,-0.5863796473,0.0766911805,0.2471939027,0.0884727389,0.2929263413,0.0240598973,-0.3884311914,-0.0572598688,0.0856191441,-0.1908696294,0.3104537725,0.1536928713,0.0625215694,-0.1571274698,0.2901481986,-0.484228611,-0.376555711,-0.2000883371,0.1573019922,-0.0807795301,-0.1328651458,0.0455373749,-0.0821603164,0.2339514941,0.1069047377,0.0545998402,-0.3364736438,-0.0613877736,0.0750116855,-0.1574610323,0.0665254146,0.4336796999,0.2859508395,0.1672497839,-0.2381448746,0.1880469769,0.0848681703,0.0679322705,0.0958095118,0.0700329766,0.3056826591,-0.0290725008,-0.1535682976,-0.044199504,0.0284220017,-0.2368972301,0.2739984095,-0.1222198978,-0.1661885828,0.3534010351,0.4656022489,0.4622932971,0.2464289218,-0.0608616807,0.2721383572,-0.1222703457,0.4582117796,0.2745559812,-0.0056035304,-0.0227365959,-0.1182778105,-0.0865723118,0.4763235152,-0.3801175058,0.1748147607,0.3582953513,-0.1875040382,0.4075791836,0.0668451786,0.1683205068,-0.2054982781,-0.0076048844,-0.0774826705,0.0983514413,-0.2572859228,0.0166916326,0.2644664049,0.1574621499,-0.0675367787,0.2874825895,0.117920801,0.2388056964,0.2470190078,0.007441205,0.3313239515,0.0093773585,0.3157010078,0.2658502758,0.4049664736,0.0810603201,-0.0379042774,-0.090882346,0.0599305369,-0.2749577463,0.3034971952,-0.0761803091,-0.0319895297,-0.4566192031,0.0974395573,-0.0728987381,-0.2265222818,-0.7448361516,-0.0819691941,-0.1600871235,-0.2373749316,0.2702182233,0.111989744,0.2704706788,-0.0109361103,0.0143488096,-0.1902953237,0.0640633404,0.0316879451,0.3597327173,-0.1756412834,0.3102312982,0.3302374482,0.1855844259,0.1869924217,0.720195055,0.1895421594,-0.0127331829,-0.2168774456,0.0275140367,-0.3248380423,-0.1628329456,-0.3571715653,0.2908222079,0.1495018899,-0.0266180132,0.2081859708,0.3554236293,-0.2950564921,-0.1567055583,0.2387396544,-0.0077259927,0.0401796997,-0.0901437476,0.2680261731,0.138619557,-0.1359923631,-0.228256315,-0.2850211561,-0.0715629607,0.1465923935,-0.1691199392,0.1783895046,-0.299957484,0.142497927,0.195516035,0.2737124562,0.1307840347,0.0149193648,0.0501843989,-0.1485399455,-0.2599642873,0.2726962566,0.132280305,-0.5165960789,-0.3158485591,0.1586721241,0.1149498224,-0.0740365461,-0.0622373223,0.148278445,-0.0541036278,-0.0756316781,-0.2514532506,0.3446326852,0.0791681558,0.072971411,0.3192301989,-0.248422429,0.0511232242,0.1157744899,0.1474984437,0.1070168465,0.1811039746,0.0355329365,0.079815127,0.0107935518,0.025454741,0.4776384234,-0.3644377291,-0.0484219566,-0.2148435563,-0.0314216129,-0.5005764961,0.0295196325,0.2169032097,0.3884496391,-0.1867879033,-0.003668105,0.3597296476,0.1774297953,0.2077928931,-0.1559929252,-0.1569238007,0.1684156805,0.0488859601,-0.1303520948,-0.082460016,0.2540156245,-0.0530028641,0.156689316,0.0877179876,-0.5936799049,0.139491275,-0.2510969937,-0.1447713077,-0.150167048,0.3362686932,0.2080435753,0.0004796902,-0.2991192937,0.0576789044,0.0948864967,0.1995245516,0.1000102833,0.224744767,0.1050251126,-0.1967421919,-0.1659430712,0.3079936504,0.1332525015,-0.0317125842,-0.0647813305,-0.3845325708]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1863","title":"Add WikiCREM","comments":"Hi @NielsRogge I would like to work on this dataset.\r\n\r\nThanks!","body":"## Adding a Dataset\r\n- **Name:** WikiCREM\r\n- **Description:** A large unsupervised corpus for coreference resolution.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1905.06290\r\n- **Github repo:**: https:\/\/github.com\/vid-koci\/bert-commonsense\r\n- **Data:** https:\/\/ora.ox.ac.uk\/objects\/uuid:c83e94bb-7584-41a1-aef9-85b0e764d9e3\r\n- **Motivation:** Coreference resolution, common sense reasoning\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":11,"text":"Add WikiCREM\n## Adding a Dataset\r\n- **Name:** WikiCREM\r\n- **Description:** A large unsupervised corpus for coreference resolution.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1905.06290\r\n- **Github repo:**: https:\/\/github.com\/vid-koci\/bert-commonsense\r\n- **Data:** https:\/\/ora.ox.ac.uk\/objects\/uuid:c83e94bb-7584-41a1-aef9-85b0e764d9e3\r\n- **Motivation:** Coreference resolution, common sense reasoning\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHi @NielsRogge I would like to work on this dataset.\r\n\r\nThanks!","embeddings":[-0.4407314956,-0.1599087268,-0.0999490991,-0.1801345199,0.0894528776,0.0274574049,-0.0447961353,0.3095080853,0.1454717219,0.3904922903,-0.0194433313,0.1824821532,-0.1553196758,0.1059726179,0.1349793524,-0.1939857155,0.0841979682,0.0411375687,0.4077956676,-0.1898164302,-0.3976635337,-0.1622177809,-0.2521415651,-0.4947379827,-0.3915042281,-0.3610392809,-0.3319664299,-0.0410158671,-0.2966036201,-0.3139800429,0.1917515993,0.2971969545,0.0988208055,0.3994063735,-0.0001035392,-0.1101281866,0.2812848389,-0.115616776,0.0468345173,0.2783694267,-0.1436743438,-0.2489386648,-0.3597214818,-0.350843966,-0.4317418337,0.1768944114,-0.0643603802,-0.3029634356,0.0359678976,0.111438781,0.2586870193,0.0236728918,0.1575330049,-0.2494108081,-0.1121732146,-0.0249303821,-0.3254310787,0.0860088766,0.2586782873,0.0961899534,-0.0250706133,0.1995141804,-0.1443886459,-0.0334497467,-0.0511933789,-0.1522969753,-0.2051431984,-0.1782246679,0.2767952681,0.384937644,0.4957941473,-0.1799378991,-0.2506143153,0.090871796,0.1013325378,-0.2681334615,0.1048922092,0.2591715753,-0.0554458387,0.2391004711,-0.2031579912,-0.0466334149,-0.2778356373,0.142296195,-0.0696110651,0.4528657198,-0.2122524381,0.0312634744,0.1426391602,0.047508385,-0.4933908582,-0.0748015344,-0.0398239233,0.0749354511,-0.3202104867,-0.2851356864,0.166029647,0.2398615927,0.4968706667,-0.2060883045,0.0850152299,0.0348710343,-0.0548174977,0.0324459933,0.122579135,-0.0495610572,-0.1779506952,0.0905847326,0.0080133118,0.0378491841,-0.0384466834,0.1776128858,-0.1901453584,0.0308457036,-0.2640621364,-0.2709587514,-0.1413232833,-0.3037443161,0.0315508768,-0.0192265343,-0.0418915227,-0.0411654003,-0.2742781937,0.2961383462,-0.0608267784,0.284971565,-0.0461669974,0.2777099609,-0.0781152025,0.0601172373,-0.2268766463,0.3212426305,-0.106877476,-0.0041831881,0.2936427295,-0.0717859194,0.0816602409,-0.020064678,0.0841760188,-0.0583187267,0.1517410576,-0.0151427314,0.0596552528,0.0953526124,-0.1270074099,0.0002532357,-0.2384724766,-0.0615398921,-0.1981245279,0.187645942,-0.0469342843,-0.1659969091,-0.4477865696,0.2395411283,-0.3009365797,-0.2284845859,0.0895396769,0.3879596293,0.0074264188,0.1182814091,0.0981447548,0.2752722502,-0.2103003263,-0.2621613443,0.1325328797,0.0803160742,-0.4427700639,0.103853032,-0.1713512093,0.1063628122,0.164522171,0.114521496,-0.0217083003,0.2421056777,0.0940716341,0.4064642489,-0.0826255903,-0.4448413253,-0.4682834446,-0.1435405761,-0.200960815,-0.1186698675,0.0907715335,0.1628204733,0.4664723873,-0.2005954981,0.0845846012,0.2308134735,-0.1183307469,-0.0635471418,-0.3397328258,-0.5084086657,-0.1497400701,0.2207913399,0.101884298,-0.2608160973,0.2788939178,-0.0059982198,0.1873816103,-0.2610117793,0.1383561641,0.3026478887,0.251835674,-0.4050545394,0.0275308248,-0.3881405592,-0.1039086878,0.0697704032,0.1581071466,0.4538139701,0.2078117281,-0.3468476832,-0.2269730866,-0.1566688418,-0.1420059055,-0.1831564754,0.2310025543,0.204402715,0.1687270403,0.0040898034,-0.2303096205,-0.4396666288,-0.2459777892,-0.2190028429,0.0836377963,0.1529903412,-0.0733813196,-0.0125384713,0.1017293334,0.2282531857,0.2066943944,0.014688205,0.0370269641,0.0126641113,-0.1705151647,0.027428614,0.311507225,0.4896600246,0.1118253693,-0.3998038769,0.2281902581,-0.1884535104,0.1793444306,-0.2681579292,-0.1078749374,0.1652228236,0.3070494235,-0.0022616128,-0.2196760625,0.0205411781,0.3762958944,-0.2551033795,0.3242551088,-0.2036981285,0.1324060261,0.3472542465,-0.2570860982,-0.1404097974,-0.0364367701,0.3359801471,0.1423147619,-0.3314623833,0.1709473729,-0.0372203551,-0.2017327249,-0.097325772,0.0680720955,0.1559993774,0.1876150817,0.3487676084,0.1981163919,-0.030741781,0.1273055673,-0.0128506394,0.1467029154,0.0961681455,-0.0254890453,0.0900782496,0.0876811445,0.0879914686,-0.2031331956,0.0040416471,0.1200463027,0.2018092275,0.113038294,-0.3147417307,0.108361125,0.3484838605,0.0957791507,-0.1272100508,-0.0673594028,-0.1370140165,0.1531420052,0.0245978963,-0.0508033372,0.2282087356,-0.1897694618,0.4302666187,-0.1523922384,-0.0952960625,-0.1558071524,-0.4620195329,0.0156335402,0.3002491295,0.0297451876,-0.0028314451,0.6374166608,-0.0410414673,0.3687342107,-0.1997547746,-0.5954600573,0.1451208442,-0.1365985572,0.2586898804,0.2262195647,0.2869263589,-0.1349432021,-0.1770446599,0.0334907882,-0.1180679649,-0.1092781425,-0.2940087914,-0.291947186,-0.1096892357,-0.1604087502,-0.413531512,-0.1793057919,-0.3016083837,0.0295026563,0.3547047675,0.0041641691,0.0260604508,-0.0059117884,0.3117597699,0.0028497493,0.243976593,-0.2872231305,-0.3574642539,0.2055399865,-0.2297451347,-0.3318353295,0.1258043349,-0.3196533322,0.3030266464,0.1264776886,-0.3386478722,0.0257469695,-0.0916181281,0.2352367043,0.2250819057,0.295389533,0.3932166696,-0.0013648377,-0.2089255452,-0.20550327,-0.2649210095,0.0296576787,-0.0081939399,0.1791798174,-0.0224129241,-0.1537304521,0.1341263354,0.6297284365,0.0509412698,-0.2344979942,0.2515153587,0.0594050996,0.2307500541,-0.1362513304,-0.1067593247,0.8309819102,0.1338877529,0.0169956423,0.373693049,0.2689469755,-0.3893512189,-0.0342108347,-0.3465065062,-0.3625879586,-0.1566365957,-0.0593790114,0.5361384153,0.1901840568,-0.1085010469,-0.0522052012,-0.2126131505,-0.3971962333,0.1869889945,0.2103909701,-0.0727277398,-0.228800416,-0.2326825112,0.4399664998,-0.5430250168,0.3954854012,0.2108593434,-0.1601977199,-0.1823779196,-0.0294173881,0.0930648893,-0.1963631511,0.2791250646,-0.5521399379,0.0841889605,0.1073124111,0.122786127,0.1487591416,0.3898605704,-0.3024753034,-0.1475185156,0.2381363958,-0.0527989827,-0.2848704755,0.1586049497,0.2330538034,0.2176512331,-0.0325030126,-0.2616946101,0.1382077038,-0.3769758344,-0.4929214418,0.0327617712,0.0412818566,0.3038755953,0.0644138157,-0.0827691928,-0.1958381981,0.1285353601,0.2046431154,-0.0661050975,0.0275186244,0.3624041378,0.1502272338,-0.0203852113,0.128879562,0.1617427617,0.2432396561,-0.0386202335,0.2289523333,0.0022880405,-0.1217049882,0.3606544137,0.1046968997,0.193627283,0.0230075978,0.1177191809,-0.0183693599,-0.4465163946,0.304390341,0.0803707764,0.1441161931,-0.6762186885,-0.0994041115,0.1839721501,0.3502432108,-0.135570541,0.2697901726,0.5367873907,-0.0305862203,-0.0402063169,0.323870182,1.117946744,-0.1863863617,-0.0461597294,0.1155973971,0.0718968436,0.3849251866,-0.1668240577,0.0044086338,-0.1430930197,-0.0136157554,0.0236961506,0.1303164363,0.2243366092,0.1143514216,-0.2979783714,-0.0066021546,0.0546591319,0.1335836649,0.0361081287,0.1466515958,0.2833259404,-0.2794210911,-0.2620079815,0.2672333717,-0.0348894261,-0.0114170797,-0.0161424633,0.0456298329,0.1566741168,-0.0962052792,-0.4005265832,0.0379346721,0.1746740341,-0.3006767035,-0.2017898411,0.042991098,0.2862248421,-0.001345832,0.3645139337,0.3377839029,-0.4034851193,0.2747937739,-0.2414198965,-0.3803391457,0.5242644548,0.0068128319,0.1586697847,-0.192829445,-0.3527013063,0.1757999957,-0.0215557814,-0.1650134623,-0.0688075796,-0.0411253348,0.3997370899,-0.3249343634,-0.0182551946,0.1573459059,-0.1726592928,-0.1892230511,0.2227850705,0.0897957608,-0.0692487881,0.1884468645,0.0499881953,-0.1288991719,-0.2412356436,0.3357365727,0.476959914,-0.4968990088,-0.058302667,0.0477662273,-0.3090817034,-0.4039881527,-0.2840791643,0.0331288204,-0.1855129898,-0.104642041,0.4165259004,-0.263271302,0.2468353361,0.679317832,0.2035892308,-0.1252930015,-0.0673448294,-0.3587679565,-0.5001820922,-0.0658699796,-0.2144863456,0.3404162228,0.1076648086,-0.1237260997,-0.0380552262,0.1771518141,-0.4541389346,0.0151897138,-0.3422978222,0.1529209465,0.1921239942,-0.0084482683,0.1876777411,0.1454505473,0.2594677806,0.0459584482,-0.1883171052,-0.3129834533,-0.1191283613,0.0963699371,0.1103080735,-0.0785999149,0.3491221964,-0.0980250314,-0.3335924745,-0.0809830129,0.2677793205,0.0364364237,-0.1072737128,0.0236559883,0.0188162513,-0.0313230343,0.0340790302,0.0334584527,0.1263934523,0.2504929602,-0.1372995228,-0.0571785122,0.0625062063,-0.1346618235,0.2377215773,0.1422775239,0.3981235623,0.2491879463,0.2804582417,0.0574074984,-0.1547867507,0.4639063478,0.239550069,-0.0659613758,-0.1560662836,0.0442869253,0.2180904001,0.3783308268,-0.3977388442,-0.0598729216,0.3731211424,-0.0497250259,0.2492633462,0.153979376,0.1551462263,-0.4958420098,-0.033301495,-0.0745335743,0.0419028513,-0.3184362352,0.1260908395,0.1325431466,0.0154190538,-0.1565835476,0.0350785367,0.0220130421,0.2204629481,0.1853109747,-0.1484041512,0.4166986644,0.043711815,0.0866500884,0.3432574868,-0.1041007712,0.7040758729,0.2969291508,0.1390071958,0.1923387498,0.1173237562,0.1657307148,0.2424838245,-0.0318779834,-0.2587847412,0.2718677223,-0.2066556066,0.2189414203,-0.3708701432,-0.2094176114,0.1691771448,-0.1592833102,-0.0910420939,-0.0029548181,0.4754193723,-0.060181953,-0.0044951336,-0.0712870136,-0.0651420131,0.0763048604,0.351466924,-0.1475037932,0.1202212274,0.2450440526,-0.0226860009,0.154788807,0.5030417442,0.3711786866,0.0935289115,-0.0217613354,0.0747647956,0.0071182982,-0.0784740746,-0.061129231,0.3980685472,0.0579172075,-0.0644916743,0.3585013151,0.1501916647,-0.2003632933,-0.0160034783,0.0299207028,0.0716478303,-0.0429702625,-0.2618921697,0.0528502911,-0.0264433492,-0.3314476013,0.1403939724,-0.2425013781,-0.0158767216,0.0521834344,-0.2351879627,0.0171528589,-0.2297913879,0.1510123909,-0.0875680521,0.2629375458,0.3375997841,-0.1868363321,-0.1274234504,-0.1306875199,-0.3394628167,0.2497768253,-0.1547258645,-0.4437461197,-0.1284370571,0.016586937,0.1658088714,0.1649753898,0.0608813837,0.1865144521,0.1372607648,0.1402982175,-0.3315558434,0.2100038677,0.1103314385,-0.1191146374,0.1022672728,-0.3408853412,0.0199099705,0.0253163557,0.1270096004,-0.1134325042,0.0822825506,-0.1478722095,-0.0650045425,0.2503086925,0.1132491454,0.433578074,-0.3894268274,-0.1632464677,-0.1437621117,0.178375572,-0.1482659429,0.0391603261,0.1048066542,0.1566553861,-0.0326245651,0.3559236526,0.2468889356,-0.2990797162,0.0350387804,0.2552298307,-0.1551950574,0.0801931545,-0.0969370008,0.055627238,0.197859928,0.3308553696,0.0969521403,0.0200582165,-0.1714352518,-0.4403534234,0.4269277751,-0.2368391752,-0.2428275347,-0.0685105026,0.1247554645,-0.1316450536,0.035386391,-0.5663987994,0.1959188879,0.27421996,0.3386182189,0.0023759326,0.2220983654,-0.2189546674,-0.159287259,-0.0780503005,0.1866428703,-0.0035417974,-0.1112445965,-0.1060951874,-0.2588943243]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1863","title":"Add WikiCREM","comments":"Hi @udapy, are you working on this?","body":"## Adding a Dataset\r\n- **Name:** WikiCREM\r\n- **Description:** A large unsupervised corpus for coreference resolution.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1905.06290\r\n- **Github repo:**: https:\/\/github.com\/vid-koci\/bert-commonsense\r\n- **Data:** https:\/\/ora.ox.ac.uk\/objects\/uuid:c83e94bb-7584-41a1-aef9-85b0e764d9e3\r\n- **Motivation:** Coreference resolution, common sense reasoning\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":7,"text":"Add WikiCREM\n## Adding a Dataset\r\n- **Name:** WikiCREM\r\n- **Description:** A large unsupervised corpus for coreference resolution.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1905.06290\r\n- **Github repo:**: https:\/\/github.com\/vid-koci\/bert-commonsense\r\n- **Data:** https:\/\/ora.ox.ac.uk\/objects\/uuid:c83e94bb-7584-41a1-aef9-85b0e764d9e3\r\n- **Motivation:** Coreference resolution, common sense reasoning\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHi @udapy, are you working on this?","embeddings":[-0.4077638686,-0.205807969,-0.0988060534,-0.2392525524,0.0346783176,-0.0113330791,0.0751565695,0.3809154034,0.1212940365,0.4436633289,-0.038902279,0.2468754649,-0.1091112494,0.1012573093,0.1549977809,-0.1249022111,0.0805579349,0.093813628,0.29372558,-0.1393387765,-0.3762649894,-0.121868521,-0.1612662673,-0.4609536827,-0.2754552364,-0.234033525,-0.3111154139,-0.0590635501,-0.2107639015,-0.3432930112,0.0519858822,0.2547081411,0.0182998292,0.3130804002,-0.0000994487,-0.1110078022,0.2080167681,-0.1099197641,0.0529746488,0.3043293357,-0.0758741722,-0.1439179331,-0.3006351888,-0.317329824,-0.5135996342,0.1668092608,-0.0316615179,-0.1975912899,0.0594412088,0.0467274897,0.3197562695,0.0472453348,0.0879576132,-0.279430449,-0.110361062,-0.0402012579,-0.2068031281,0.0997380093,0.2196082473,0.0766661614,-0.1418928504,0.3368791938,0.0026644573,0.0289097894,-0.064967081,-0.0536094382,-0.1552438885,-0.0848126784,0.1953252256,0.3484485447,0.3779142499,-0.2195069641,-0.276255697,0.0688441768,0.00718938,-0.3482172489,-0.0425836146,0.2347936183,-0.0256828852,0.2060023099,-0.1693071127,-0.0646835491,-0.2618379295,0.1254189909,-0.07713411,0.4296009243,-0.222291559,-0.0796642825,0.1207824945,0.0310776066,-0.4710711837,-0.0302187968,0.0390850566,-0.0277639609,-0.2232083529,-0.3255957067,0.1717935801,0.1405705959,0.5156207681,-0.2325478941,0.1638853252,0.0709455237,-0.0233926605,0.093535766,0.1340420693,-0.0283063632,-0.2073621899,0.0782231018,0.0925559029,0.1027844697,-0.0138675114,0.1995332241,-0.0817440376,0.1103855968,-0.2250864059,-0.1607063115,-0.2622678876,-0.2896385193,-0.0091428142,-0.0117970901,-0.0127246929,0.1042085886,-0.270532608,0.3341981769,-0.04000948,0.2690737247,0.0065498147,0.2451579124,-0.1030224115,-0.0009120557,-0.2158444375,0.190609768,-0.1396459192,0.0730037689,0.3568342328,-0.0195532274,0.1512200236,-0.0943876132,0.0908565521,-0.0157427788,0.0994291157,0.0290356465,0.161209479,0.0236667916,-0.0682906136,-0.1472090036,-0.2026321292,-0.0638897866,-0.199055776,0.0931191519,0.0078827543,-0.123224996,-0.4064459503,0.3104184568,-0.1398646086,-0.1979910433,0.2308237106,0.4591111541,-0.0231020786,-0.0234947987,0.0857784823,0.286418736,-0.1976640373,-0.206639275,0.083448194,0.1108500734,-0.325931251,-0.0194297358,-0.0862330869,0.1331335157,0.0026222444,0.1451297253,-0.0721894056,0.1019437835,0.0075700423,0.4074426293,-0.1625498235,-0.4612466991,-0.4476382136,-0.1357137859,-0.2744984627,-0.0723336488,0.1172101349,0.2298997194,0.4492276311,-0.1861773282,0.1586811543,0.2008193284,-0.1466075182,0.0189532228,-0.3592533767,-0.4677701294,-0.2614105046,0.2164385617,0.1483157873,-0.2325045615,0.3176487684,0.0535159893,0.2632814944,-0.2881196141,0.1640310287,0.2253045589,0.2054742128,-0.3424519002,0.0651866272,-0.4240676761,-0.0923776403,0.0077714836,0.0834631398,0.409442395,0.2136125863,-0.4916167855,-0.1583684981,-0.1952052265,-0.0880761221,-0.1037088782,0.3057529628,0.2273789346,0.1011631787,0.0407457091,-0.16509673,-0.4185503423,-0.1780884713,-0.1318175644,0.1383481473,0.1066794544,-0.092082046,-0.0139201479,0.149353981,0.1734752208,0.0752497688,0.013473561,0.0997923985,0.0910715833,-0.0985889509,0.0782054216,0.2984572053,0.4659199119,0.1262162179,-0.3325410485,0.2027386874,-0.0858426392,0.1148561388,-0.1473759562,-0.2065532953,0.23092179,0.3846161664,-0.0599152632,-0.0642100275,0.0105864955,0.3201976717,-0.2718378007,0.24895823,-0.3035152256,0.0704170987,0.3102131188,-0.2855428159,-0.1459620744,-0.0879385248,0.3402736187,0.1380183548,-0.2412950397,0.2524059713,0.0696373954,-0.2065127939,-0.1811681986,0.1035562903,0.1179132089,0.1744139344,0.4337342083,0.2602734864,-0.084724158,0.1115362793,-0.0399302281,0.1513885111,-0.0366595909,-0.0773890093,0.2068917453,0.0965763777,-0.0739368722,-0.3251163363,0.0236723293,0.0102797756,0.0368791372,0.1109577492,-0.2369595617,0.1050554588,0.2115240246,0.0959396064,-0.1127219424,-0.0024941869,-0.0897304714,0.2962302566,0.087920025,-0.1560901701,0.1623190641,-0.2009939551,0.4561264217,-0.1237073317,0.0213737562,-0.0545677468,-0.416775018,0.0362091288,0.3390698731,-0.1098536998,-0.102574639,0.6233806014,-0.0126132257,0.3913969696,-0.2147289068,-0.6188639402,0.2255098671,-0.1025129631,0.1161777675,0.245871976,0.1143373176,-0.12842682,-0.101572752,0.0280533861,-0.1576159745,-0.1187616363,-0.2788777947,-0.2247140855,-0.068672128,-0.21844998,-0.4074412286,-0.2077162564,-0.3283400238,0.0930915624,0.3551411927,0.0207858365,0.0489591435,-0.0019187803,0.1847237349,-0.062835291,0.2636861801,-0.3206385374,-0.2905136049,0.1431726813,-0.3450678885,-0.3356272876,-0.0579644553,-0.3895283341,0.3305683732,0.1331154704,-0.2935569286,-0.080513671,0.0119440397,0.1864883155,0.1481633037,0.2084597647,0.3978950679,0.0560756363,-0.3066022396,-0.2230830938,-0.1476482004,0.0843561664,-0.0467290059,0.13054052,-0.0932736844,-0.2166330814,0.1979939789,0.5052394867,0.0432392992,-0.2054990381,0.3556256592,0.0426843092,0.278221786,-0.1076478809,0.0366373286,0.6706678867,0.1533032656,0.0359614156,0.5118537545,0.302559793,-0.4134701788,-0.0758531392,-0.3085207045,-0.3737484515,-0.2135562748,-0.0611034445,0.4238473475,0.2065554559,-0.0730456784,-0.0393766239,-0.2124172896,-0.3112090826,0.2269527614,0.1646388322,-0.0567145571,-0.2540764511,-0.1519553661,0.3400621414,-0.4402955174,0.328442812,0.262609601,-0.2102814615,-0.1872894317,-0.1548252106,-0.0355117284,-0.1821598709,0.2202617824,-0.4549578428,0.0897883251,0.0042892345,0.0008833446,0.1504564136,0.3925900757,-0.2747558057,-0.2607506514,0.1493119895,0.0166875198,-0.3690779805,0.2162151635,0.2319925278,0.1734744906,-0.0795122534,-0.2585974932,0.1107180789,-0.2556708753,-0.4173486531,0.0879677013,-0.0447423831,0.2250300795,0.0547549054,-0.1566604376,-0.13496916,-0.0034829471,0.1024612859,0.0649412125,0.1177830994,0.2879518569,0.1287029088,-0.0358289629,0.1928113997,0.1937588304,0.2322993577,-0.0382260121,0.1442051232,-0.0980684161,-0.1473768204,0.2768398523,0.1012690216,0.1958398372,0.067730844,0.1215423793,-0.0191838127,-0.3842844069,0.2794705033,0.0764273703,0.1024325192,-0.4863859117,-0.0049012979,0.1954122782,0.3242048621,-0.151915729,0.2949093282,0.4272574186,-0.1672812551,0.0186211858,0.2361995876,1.0328341722,-0.136559695,0.0239676461,0.1584387273,-0.0134253604,0.3355565369,-0.1921784729,0.1085271537,-0.1307767034,-0.0727516562,0.0396622829,0.1887578815,0.1185131669,0.1440252215,-0.3859009743,0.0262186024,0.074739702,0.1375958771,-0.0594806448,0.1093910113,0.1475106776,-0.3034875691,-0.2808917463,0.3073491156,-0.0447129197,-0.0401199646,-0.0630042031,-0.0150932735,0.1283406615,-0.1440418065,-0.2346902788,0.0673750937,0.235418573,-0.2452817708,-0.222697705,0.0907002464,0.0471853763,-0.0936517045,0.3010088503,0.3013942838,-0.3733263016,0.2287617773,-0.176249072,-0.2098056823,0.4446069598,0.0664169267,0.1496900022,-0.1647577286,-0.2754538953,0.1735024452,0.0593606047,-0.2308817655,-0.0201572701,-0.0853141993,0.384827733,-0.2472136766,0.1141272038,0.1730683446,-0.2638600767,-0.2070954293,0.2689556479,0.0423929878,-0.0373793803,0.1814082712,0.1138628423,-0.0595052764,-0.2836385667,0.400172174,0.4384227395,-0.4283724129,-0.1911134422,0.048675362,-0.3080506623,-0.4142374694,-0.4425738454,-0.0108989291,-0.1340936869,0.0102144899,0.3136352301,-0.3374761045,0.2704055905,0.5956162214,0.2535864115,-0.0851942152,-0.1448263377,-0.2032132745,-0.4807205498,0.0387034155,-0.1993643045,0.446364224,0.1789088845,-0.0864979103,-0.0922113284,0.0843789876,-0.5146923065,-0.0953183621,-0.2592647076,0.1713119149,0.0007960685,-0.0379747599,0.1682861745,0.1591706723,0.3078216016,0.0861056149,-0.1038957685,-0.3587715626,-0.1043481529,0.0733084008,0.0723568201,0.0051030261,0.2836004496,-0.0185867175,-0.2648432255,-0.1091645658,0.3356852233,0.009554904,-0.0899251401,0.0236340594,0.0242940001,-0.0380575284,0.2061522305,0.0323656239,0.0850776583,0.2819442451,-0.1190773249,-0.1368392706,-0.0151214181,-0.1470354199,0.1840705127,0.2090016305,0.3853526711,0.1259191334,0.3337939382,0.0598944798,-0.1644434482,0.3879371285,0.3616994023,-0.0754584,-0.1593722403,-0.0093056038,0.16023238,0.4749807119,-0.438421011,-0.0894710496,0.2101851553,-0.0808399171,0.3618840277,0.2185579687,0.1248944849,-0.4961799681,0.0131323077,0.0106525449,0.0175341181,-0.2215996087,0.0440348648,0.121637851,0.0785127953,-0.0876083001,-0.0229873341,-0.0117578171,0.3087129891,0.1461340785,-0.1316762418,0.3724793196,0.0442643575,0.0834809765,0.3977062702,0.0294246059,0.6038467884,0.1800377369,0.1410353333,0.2301468253,0.1694865674,0.1485455781,0.1573044956,-0.0339697227,-0.3479909003,0.1955111474,-0.2034083158,0.2029089481,-0.313136518,-0.1647763103,0.1733789891,-0.1211948767,-0.0017371319,0.0515026413,0.4097982645,-0.0725826696,0.0642750859,-0.1830426902,-0.0359882489,0.0580609813,0.4127602577,-0.1487865299,0.1106857583,0.3017064631,-0.0715014488,0.1934209019,0.496450156,0.2459880859,0.0333878025,-0.0006685274,0.0056506791,-0.066860579,-0.0254244562,-0.1698480248,0.2500841916,0.1790136099,-0.0137956236,0.3562876284,0.2165046483,-0.2637440264,0.0388241634,-0.0029923888,-0.0073129092,0.0487936921,-0.2256844193,0.1695144475,-0.0032938826,-0.3303879499,0.065791972,-0.3926138282,-0.1072265953,0.0268990304,-0.2166593671,0.0430137403,-0.2569118738,0.1591167897,0.0130631421,0.2076710314,0.2188807279,-0.1980095506,-0.1045340598,-0.156967774,-0.2332954556,0.3279353678,-0.1391162276,-0.4429382086,-0.2375700474,0.0951624215,0.1669791788,0.1585065275,-0.035439834,0.153932929,0.1392980069,0.1424828321,-0.3694817424,0.2487863451,0.0605093949,-0.0811578855,0.1459499598,-0.3438447714,0.0799975321,-0.0236566383,0.2079273462,-0.1281655282,0.0932697728,-0.1996394545,-0.0357468314,0.3014785647,0.0167488139,0.3266522586,-0.4614873528,-0.2716183662,-0.0955085978,0.1414510459,-0.1381998062,0.0168076046,0.1141944081,0.1353174448,0.010042157,0.2494295239,0.2706552446,-0.215249151,0.1234369501,0.2064025849,-0.093148604,0.1147077903,-0.0878318623,-0.027637748,0.0998094007,0.3559868038,0.0705399364,-0.0102553088,-0.0803446844,-0.5434737206,0.3916168511,-0.2346883416,-0.1309483051,-0.0451919697,0.1598547995,-0.1458108574,-0.0802700892,-0.4972158074,0.2129828632,0.2630094886,0.3575159907,0.0254281368,0.2474963069,-0.2499663979,-0.2067599595,-0.080073975,0.2000027001,0.0412563533,-0.1130964458,-0.1929794252,-0.28468436]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1859","title":"Error \"in void don't know how to serialize this type of index\" when saving index to disk when device=0 (GPU)","comments":"Hi @corticalstack ! Thanks for reporting. Indeed in the recent versions of Faiss we must use `getDevice` to check if the index in on GPU.\r\n\r\nI'm opening a PR","body":"Error serializing faiss index. Error as follows:\r\n\r\n`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at \/home\/conda\/feedstock_root\/build_artifacts\/faiss-split_1612472484670\/work\/faiss\/impl\/index_write.cpp:453: don't know how to serialize this type of index`\r\n\r\n\r\nNote:\r\n\r\n`torch.cuda.is_available()` reports:\r\n\r\n```\r\nCuda is available\r\ncuda:0\r\n\r\n```\r\n\r\nAdding index, device=0 for GPU.\r\n\r\n`dataset.add_faiss_index(column='embeddings', index_name='idx_embeddings', device=0)`\r\n\r\nHowever, during a quick debug, self.faiss_index has no attr \"device\" when checked in` search.py, method save`, so fails to transform gpu index to cpu index. If I add index without device, index is saved OK.\r\n\r\n\r\n```\r\ndef save(self, file: str):\r\n \"\"\"Serialize the FaissIndex on disk\"\"\"\r\n import faiss # noqa: F811\r\n\r\n if (\r\n hasattr(self.faiss_index, \"device\")\r\n and self.faiss_index.device is not None\r\n and self.faiss_index.device > -1\r\n ):\r\n index = faiss.index_gpu_to_cpu(self.faiss_index)\r\n else:\r\n index = self.faiss_index\r\n faiss.write_index(index, file)\r\n```\r\n","comment_length":29,"text":"Error \"in void don't know how to serialize this type of index\" when saving index to disk when device=0 (GPU)\nError serializing faiss index. Error as follows:\r\n\r\n`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at \/home\/conda\/feedstock_root\/build_artifacts\/faiss-split_1612472484670\/work\/faiss\/impl\/index_write.cpp:453: don't know how to serialize this type of index`\r\n\r\n\r\nNote:\r\n\r\n`torch.cuda.is_available()` reports:\r\n\r\n```\r\nCuda is available\r\ncuda:0\r\n\r\n```\r\n\r\nAdding index, device=0 for GPU.\r\n\r\n`dataset.add_faiss_index(column='embeddings', index_name='idx_embeddings', device=0)`\r\n\r\nHowever, during a quick debug, self.faiss_index has no attr \"device\" when checked in` search.py, method save`, so fails to transform gpu index to cpu index. If I add index without device, index is saved OK.\r\n\r\n\r\n```\r\ndef save(self, file: str):\r\n \"\"\"Serialize the FaissIndex on disk\"\"\"\r\n import faiss # noqa: F811\r\n\r\n if (\r\n hasattr(self.faiss_index, \"device\")\r\n and self.faiss_index.device is not None\r\n and self.faiss_index.device > -1\r\n ):\r\n index = faiss.index_gpu_to_cpu(self.faiss_index)\r\n else:\r\n index = self.faiss_index\r\n faiss.write_index(index, file)\r\n```\r\n\nHi @corticalstack ! Thanks for reporting. Indeed in the recent versions of Faiss we must use `getDevice` to check if the index in on GPU.\r\n\r\nI'm opening a PR","embeddings":[-0.0712384582,-0.3793422878,-0.0126395551,0.1242901012,0.3250301182,0.1440671384,0.3014210463,0.5358146429,0.4775857925,0.3618176281,0.0682528615,-0.0320642032,0.1739509106,-0.0848504752,-0.1260111183,0.1587210149,0.4708423615,0.2794695199,0.4284847975,-0.0465701222,-0.4897720814,0.0811465085,0.0083706705,0.0710052326,0.0153226191,-0.0360257663,-0.0760059804,-0.0318541415,-0.1387515366,-0.4813572764,-0.1281306595,-0.2600694299,0.4309383333,0.1854503453,-0.0001175451,0.0909124911,0.2596594691,-0.1979711354,-0.0848168135,-0.066420801,0.3481684029,0.0831488296,0.1509593874,-0.3121950924,-0.2275109142,-0.3267764449,-0.0085452357,-0.1260353476,-0.179897964,0.2844127417,0.165733844,-0.1552814245,0.4602343142,-0.0657822639,-0.0353392623,0.0855674967,-0.2641057074,-0.1356429309,0.0376558676,0.5125663877,0.2819295824,0.3212586641,0.0330720544,-0.280249238,-0.1809240431,-0.000955855,0.6283458471,-0.3459066153,0.1626062989,-0.155659467,0.0611039288,-0.1890729964,-0.3456079066,0.2895677388,0.2570062876,-0.4825872779,-0.0632231086,0.0443118513,-0.2356054336,0.1721524298,0.6439068317,-0.2015724033,0.0719451681,-0.1651588976,0.2031634748,-0.0524165481,-0.0817488804,0.0932567418,0.0007190898,-0.1320931762,-0.1938658804,0.0544876792,0.159401387,0.1064695045,-0.3842536211,-0.1285613775,0.0794626027,-0.4333073199,-0.1163278893,0.1132654771,-0.4117616713,0.0138393734,0.1884579062,0.0297547486,-0.2788296342,-0.0231560301,-0.0122408941,-0.202277571,-0.0987560973,0.0644984469,0.0437913686,-0.3351603448,-0.1625306606,0.0169744045,-0.3138410151,0.1340081394,0.2640893757,-0.3600541353,-0.5404109359,0.1515235156,-0.2051830441,0.2541145682,0.1186715141,0.3394500017,0.3243599832,0.1193249375,0.2600001693,0.2940564752,-0.2567103505,0.0723281801,-0.1546052098,-0.1204363033,0.3157219589,0.1273838133,-0.0999876782,-0.4395748675,-0.055944223,0.0610690489,-0.2190777659,0.0148686077,0.0390028544,-0.3672314882,0.137136668,0.4929429293,0.0687847063,0.1333899051,0.3695411086,-0.1329853237,-0.1448982358,0.3369978368,0.0195288118,-0.2087257951,-0.423784852,0.1341565996,0.2424482405,-0.0107827382,0.3436016738,0.0485057347,0.0278701372,-0.0147276456,0.0952939317,0.133586064,0.0630662665,-0.2642409503,0.4410449862,0.0570823178,-0.0643290058,-0.0238724928,0.1438179016,-0.1553903669,0.243175894,0.4800496995,0.4059357047,0.2361027747,-0.3073669672,0.2976962626,0.4374848008,-0.2929921448,-0.0848619044,0.0937048346,-0.1128476784,-0.4517088532,0.3368116319,0.0265279058,0.1492991447,-0.0534780622,0.221041128,-0.1418749839,-0.0379879512,-0.1587764174,-0.3545044661,-0.3207323253,0.3053153157,-0.0515019819,-0.0426942334,0.1113008261,-0.1175483987,-0.5868125558,0.1098725423,-0.0735581592,-0.0875956491,-0.0569470823,0.9322971702,0.0278815273,0.2512435615,-0.0496051162,0.2162654847,-0.0534441769,0.0089386702,-0.1493970305,-0.399633497,-0.0829541832,-0.0047222939,-0.0199991483,0.3082149625,0.1228873208,0.0533143319,-0.1235305741,-0.1038722843,0.2014550269,-0.2821288407,-0.1687683761,-0.1973066479,0.0413984358,0.0628545806,0.1378224194,-0.337028265,-0.4472599626,-0.1559965909,0.2468762398,0.3647454381,-0.239956677,-0.0033568651,0.3587985933,0.1815855205,-0.0744804665,0.2085341811,0.0854679346,-0.0874268785,-0.0320986323,-0.0945054591,0.4734265208,0.4213306904,-0.03208239,0.1105897427,0.3010748327,0.2076686323,0.0950360745,-0.1479107738,0.1323513389,0.4306560457,0.0921608582,0.0992562771,-0.131262362,-0.0976164341,0.3995724618,0.1208027825,-0.2026572078,-0.6101469994,0.1275203973,-0.0361715369,-0.0758984387,0.2764007151,-0.0513894372,-0.1054836139,0.0300882608,0.0499136001,-0.8405379057,0.1827556491,0.043939665,-0.2287994474,-0.1709542572,-0.0537621602,-0.1136793569,0.3696154654,-0.1543500423,-0.1558543742,0.1671735495,-0.1639797837,-0.0550496988,-0.0069373245,0.0262612067,-0.1074864268,0.0510736667,-0.3698427379,0.0085143363,0.0789709836,0.2077314854,-0.2933821976,-0.1897430867,-0.0669160187,0.0192803759,0.1494904459,0.2111279964,-0.026634505,0.2719708085,-0.2457418591,0.2465221733,0.2640500367,0.0451867878,-0.1872443706,-0.068279393,-0.050330665,-0.0573591441,0.0380743295,0.0472577699,0.1168290004,-0.1141843423,0.082290642,-0.2946701944,0.0565170422,0.0235375203,-0.2850291729,-0.1330879778,-0.1121563911,0.2326910049,0.0897386521,0.115924485,0.179619357,-0.0158458371,-0.2538018227,0.1694006622,-0.2895036042,0.0810748264,-0.2157639414,-0.093047902,-0.0153563842,-0.2302937508,0.0575717315,-0.0329140499,0.1174361855,-0.1404099315,0.2576200366,0.2002503276,0.3634287417,-0.2162118554,-0.1813200414,-0.1422646195,0.1777137816,0.0998021141,-0.4149801731,0.3305240273,0.1879826933,-0.0797430426,0.0312867276,0.0676988661,0.1965959668,0.1595302075,0.1310822964,-0.3357588351,0.1694398671,0.4078317285,0.0772916228,-0.1620270014,-0.0135013033,-0.0939559937,0.0710933357,0.0425379202,0.2906353772,-0.1924156696,0.3714401722,-0.1209215075,0.527189374,-0.0674376711,-0.4186014235,0.2030136883,-0.0951289833,0.1291924268,-0.082832776,-0.1675729603,0.4965711534,0.3667693138,-0.3319823146,0.0488788523,-0.4631192386,0.0197123215,0.1105112284,0.2581325769,-0.0941582397,-0.0665519163,0.0787322521,-0.0464896411,0.2993482053,-0.5147420168,0.3564274311,-0.0554458275,0.2113454342,0.232053861,0.4537919164,0.2656055689,-0.1263950318,0.2204959244,-0.2902864218,-0.3792003393,0.4615423381,-0.0284014046,0.3594944477,-0.0985932574,-0.1411703527,0.2867295742,0.0377819091,0.5474432707,0.2116204649,-0.3290473521,0.4518081546,0.1415424496,-0.3843215704,-0.1611052006,-0.1790841967,0.2378777117,0.0109718777,0.4116705358,-0.1174162328,-0.3329501748,-0.3428003192,-0.0415191166,-0.2220499814,-0.2910752892,-0.421022743,-0.3375881016,-0.1586618423,0.4674514234,0.0171478465,0.3032742441,-0.1070557386,0.1508081853,-0.4915829897,-0.107655853,-0.0780055299,-0.1760778576,0.2934147716,-0.1969268471,-0.091050528,0.0964758545,0.2985581458,0.4689682126,-0.1375478804,0.3453898132,0.1153774858,0.2162029296,0.048441235,0.0482280105,0.15357548,-0.0105878087,0.1670198441,-0.1051018238,0.1631274074,-0.5299460888,-0.1543177664,0.2289403379,-0.042661678,-0.2358109206,0.1510487497,0.3084451258,-0.1525323093,0.0727142245,0.031839285,0.2313715518,-0.1411706209,0.7170218229,0.4281056523,0.9727648497,-0.045755621,-0.1917944551,0.1572518945,-0.4654707611,0.2642842233,-0.1838365644,0.4179236889,-0.5206820369,-0.0384491347,-0.0848185271,0.0788293406,0.2849553823,-0.1983235478,-0.4596638381,0.078482531,-0.4503545761,-0.1716126204,-0.1771043837,0.0814976394,0.0242341328,-0.2860445678,-0.1874105334,0.064548105,0.0486763902,0.2211966962,0.1126721948,-0.0673128366,-0.1354816705,0.0555884354,-0.2241005301,0.0320326649,-0.1085607931,0.3049998879,-0.1301370561,-0.2326379716,-0.1463970393,-0.3277444839,0.4981686771,-0.3548031151,-0.2450336069,-0.1663228422,0.0345945209,-0.1157079116,-0.1562319696,-0.1925418079,0.2612090409,0.0618066937,-0.1635155827,0.0063223382,0.0223619975,0.0268577598,0.0025241682,-0.1410477757,-0.3572549224,-0.3139790297,-0.4244965315,-0.0923329666,0.2318003923,0.0967235714,0.0967092216,0.3062935472,-0.271713078,0.1072380468,-0.2875131965,-0.1340830028,0.0060452507,0.3354347646,0.0758272558,0.0202163253,0.376619786,-0.3060869575,-0.1705391407,-0.1801343858,-0.2744092941,0.1611406356,-0.0199051201,0.0499291234,-0.0482469052,-0.2829883397,-0.048085317,0.1993077993,-0.0310297646,-0.1004890576,-0.6292185187,-0.0166368559,-0.4327389598,0.0409639664,-0.2383730412,0.2573255599,-0.1825343072,0.3144520819,-0.0718005002,-0.305644542,-0.2587477267,0.0695555732,-0.4101376832,0.2657018006,-0.4145555794,-0.1361314058,0.0721296594,-0.0040343911,0.1892806143,0.0023424039,-0.0447903983,-0.1383941919,-0.1356264651,0.1331773698,0.0876084939,0.2126123309,-0.152917251,-0.0640684143,-0.1222837567,-0.0091961231,-0.0924090967,0.2961057425,-0.0596884415,0.1375635564,0.1278684288,0.2047434747,-0.1311385483,-0.0529431179,0.2364308983,0.3273602724,-0.1198692322,0.1237516031,-0.3980755806,0.1246480867,-0.3058548272,0.1360518932,0.366917789,0.1037764177,0.0543196797,0.0582993329,-0.1199313253,0.1112457961,-0.1013382599,0.2528004944,-0.169813782,-0.0065537877,0.1894349009,0.2006902993,-0.002072399,0.1720246375,-0.0062767612,-0.2850208282,-0.1459521651,0.1465512216,0.3709390163,0.0773445964,-0.1627675593,-0.1122219786,0.1676335931,0.1602598876,-0.1290872991,-0.2250767499,0.2475020587,0.2212368697,0.168211773,0.0048405617,0.420807451,0.4026451111,0.0048014489,0.1288187504,0.5096390843,0.1529030204,-0.0295121782,-0.0068310034,0.2534535527,-0.1015483066,-0.0050085406,-0.1515843421,-0.085988231,0.0324034095,-0.4543075562,-0.3302355111,0.0706347227,0.2346277386,-0.4260687828,-0.0737660453,0.1905769706,0.2589171529,-0.0605845451,-0.3084826171,0.210009113,0.1119435951,-0.0233672597,0.1212687045,-0.1889848113,0.323212713,-0.4511882365,0.5020250082,0.3923589885,-0.0609437078,0.02530526,-0.0984660462,-0.1560994089,0.0754402503,0.3508640826,0.5285477638,0.4314030707,-0.3667829633,0.2937222719,0.2441829145,-0.0594789796,0.114927493,-0.2624688148,-0.0268694293,0.2984538972,0.1137851998,0.0836742148,0.0034765706,-0.3272370398,-0.0920965001,0.3319786787,0.0844725296,0.1815083176,-0.3820916116,-0.0899768323,-0.2460793704,0.0526116751,-0.2649683654,0.0140136508,0.0373253599,-0.2223136425,0.0476381071,0.2294640243,0.039745722,-0.1147418916,0.0392754823,0.088831991,0.0258671977,-0.3434681892,-0.0146147376,0.0055436888,0.2804381251,0.2627859712,0.0692646205,0.2042123526,0.5844687223,-0.0433292091,-0.2042547464,0.0936052054,-0.0386313386,0.3985429108,0.4201744497,0.0109178899,0.1280185729,-0.508053422,-0.2697492242,-0.0381361246,0.0823268294,0.2540043294,0.106594637,0.1364918947,0.015240198,-0.0982608423,-0.0470896028,0.0751611367,0.6740577817,0.0045946524,-0.0781618059,-0.0839427635,-0.2627613842,0.1842107028,0.014303267,-0.3596515059,0.0784462616,-0.0378655344,-0.036199823,-0.0588833243,-0.2341378927,-0.2605669796,0.7164087892,-0.0687383562,-0.3527014256,-0.1285432577,0.2969983518,-0.3471461833,-0.2056799084,0.3894695342,0.0512245446,0.0723171756,0.4429090023,-0.2615904808,-0.3091915846,0.3434889913,0.0125176087,0.0193110723,-0.0837199017,0.3828177452,0.2854709625,0.5122018456,-0.8527299762,-0.0467742532,0.1794889569,0.0681958497,-0.0979242474,-0.2164348662,0.0820890665,-0.0615525134,-0.102163747,-0.0230054446,0.0768950358,0.4291138649,-0.0582010373,-0.1954105794]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1859","title":"Error \"in void don't know how to serialize this type of index\" when saving index to disk when device=0 (GPU)","comments":"I fixed this issue. It should work fine now.\r\nFeel free to try it out by installing `datasets` from source.\r\nOtherwise you can wait for the next release of `datasets` (in a few days)","body":"Error serializing faiss index. Error as follows:\r\n\r\n`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at \/home\/conda\/feedstock_root\/build_artifacts\/faiss-split_1612472484670\/work\/faiss\/impl\/index_write.cpp:453: don't know how to serialize this type of index`\r\n\r\n\r\nNote:\r\n\r\n`torch.cuda.is_available()` reports:\r\n\r\n```\r\nCuda is available\r\ncuda:0\r\n\r\n```\r\n\r\nAdding index, device=0 for GPU.\r\n\r\n`dataset.add_faiss_index(column='embeddings', index_name='idx_embeddings', device=0)`\r\n\r\nHowever, during a quick debug, self.faiss_index has no attr \"device\" when checked in` search.py, method save`, so fails to transform gpu index to cpu index. If I add index without device, index is saved OK.\r\n\r\n\r\n```\r\ndef save(self, file: str):\r\n \"\"\"Serialize the FaissIndex on disk\"\"\"\r\n import faiss # noqa: F811\r\n\r\n if (\r\n hasattr(self.faiss_index, \"device\")\r\n and self.faiss_index.device is not None\r\n and self.faiss_index.device > -1\r\n ):\r\n index = faiss.index_gpu_to_cpu(self.faiss_index)\r\n else:\r\n index = self.faiss_index\r\n faiss.write_index(index, file)\r\n```\r\n","comment_length":34,"text":"Error \"in void don't know how to serialize this type of index\" when saving index to disk when device=0 (GPU)\nError serializing faiss index. Error as follows:\r\n\r\n`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at \/home\/conda\/feedstock_root\/build_artifacts\/faiss-split_1612472484670\/work\/faiss\/impl\/index_write.cpp:453: don't know how to serialize this type of index`\r\n\r\n\r\nNote:\r\n\r\n`torch.cuda.is_available()` reports:\r\n\r\n```\r\nCuda is available\r\ncuda:0\r\n\r\n```\r\n\r\nAdding index, device=0 for GPU.\r\n\r\n`dataset.add_faiss_index(column='embeddings', index_name='idx_embeddings', device=0)`\r\n\r\nHowever, during a quick debug, self.faiss_index has no attr \"device\" when checked in` search.py, method save`, so fails to transform gpu index to cpu index. If I add index without device, index is saved OK.\r\n\r\n\r\n```\r\ndef save(self, file: str):\r\n \"\"\"Serialize the FaissIndex on disk\"\"\"\r\n import faiss # noqa: F811\r\n\r\n if (\r\n hasattr(self.faiss_index, \"device\")\r\n and self.faiss_index.device is not None\r\n and self.faiss_index.device > -1\r\n ):\r\n index = faiss.index_gpu_to_cpu(self.faiss_index)\r\n else:\r\n index = self.faiss_index\r\n faiss.write_index(index, file)\r\n```\r\n\nI fixed this issue. It should work fine now.\r\nFeel free to try it out by installing `datasets` from source.\r\nOtherwise you can wait for the next release of `datasets` (in a few days)","embeddings":[-0.2384788841,-0.2405906171,-0.0386406891,0.1195868924,0.3427581191,0.152492255,0.3444892168,0.5221388936,0.4608196318,0.2513044775,-0.079141818,0.1412855983,0.1216126382,-0.0570589788,-0.1395932287,0.1362419277,0.4380633831,0.1934212744,0.3012018204,-0.0583262667,-0.400860697,0.0912531316,0.0359445699,0.0129263112,-0.18736431,-0.0472335294,-0.0613629334,-0.0423774272,-0.1474945396,-0.4240972698,-0.0577419847,-0.154525429,0.4029407203,0.3433766663,-0.0001137439,0.0171603374,0.2634907365,-0.198621437,-0.1538046747,-0.1396831423,0.2965035141,0.0760403275,0.1519305706,-0.3508684635,-0.2409550399,-0.2891043127,-0.0485756733,-0.2439465076,-0.1022812054,0.2978606224,0.2040632516,-0.0875000507,0.4052405655,-0.1325298846,0.0500694141,0.110026218,-0.2817083001,-0.1140927747,0.0078178281,0.4504535794,0.2506672144,0.298507601,-0.0069536157,-0.2596974075,-0.0916546956,-0.0183353834,0.3743934631,-0.3835607767,0.2296165824,-0.1212559193,0.1396841705,-0.3347601891,-0.361761421,0.2229541838,0.1872790754,-0.4804652035,-0.0755150393,0.0607192926,-0.2271499634,0.1341762394,0.5808408856,-0.1446946859,0.0683102831,-0.1654362082,0.1633944362,0.0585158244,-0.1198725328,0.1065055951,0.0583300814,-0.1843249947,-0.0953072831,0.064227432,0.1779869199,0.1287582815,-0.5028725266,-0.0867344141,0.0247371625,-0.4174596369,-0.0460305735,0.072453253,-0.3467902243,0.1453314573,0.1766988635,0.0592344068,-0.2870437801,0.0337667428,-0.0274505001,-0.1457376629,0.0066589694,0.0383950844,0.0344570316,-0.330335319,-0.1660042107,-0.0315668434,-0.3789484203,0.0761541799,0.1963506192,-0.3188083172,-0.5184018612,0.1290775836,-0.0136288051,0.2290712446,0.141697824,0.4266785085,0.3175201416,0.0353451818,0.2645550966,0.2988888323,-0.2037705034,0.0669840127,-0.1860590875,-0.1827611625,0.2399476618,0.1644988358,0.0138555197,-0.4622967541,0.064864412,0.0778760836,-0.1108913347,0.1539482325,0.0939454958,-0.2833533287,0.094420217,0.4206935763,0.0270249993,0.2014466673,0.3102342188,-0.0711252019,-0.1758990884,0.3338373601,-0.00733579,-0.2411532253,-0.5131024122,0.1491003633,0.213733837,-0.0529846251,0.2680781186,0.0859570205,0.1033834144,-0.0308677033,0.0688061267,0.0914394855,0.0358478613,-0.3160327077,0.495166719,0.0128610181,-0.1686315089,-0.0478117578,0.1192981452,-0.1202775836,0.1759350151,0.5109654665,0.3312971294,0.1718973964,-0.2703888118,0.3007789254,0.4135612845,-0.3300421238,-0.1987008601,0.064755708,-0.0192649029,-0.3254494071,0.1825144887,0.0729461759,0.2850210071,-0.0070310356,0.2279463261,-0.0312298518,0.0020127299,-0.1611647159,-0.3785850108,-0.3233622015,0.4342168868,0.0403834768,-0.0712841079,0.0773933083,-0.1179749966,-0.6022039056,0.1653825641,-0.1568247229,-0.0420191102,-0.0337346904,0.9180984497,0.0664086193,0.248209998,-0.0480834469,0.0745991617,0.021119928,0.1085083857,-0.1187084839,-0.4229264855,-0.1303247958,0.0156875271,-0.0758009329,0.2126974463,0.0030957297,0.0767877474,-0.1611407697,-0.1121302396,0.1742811054,-0.3099654913,-0.0134581411,-0.1755194664,0.0850166082,-0.0379409082,0.1815442592,-0.3314419687,-0.4518476129,-0.1660560519,0.2707016468,0.3499483168,-0.2712790072,-0.0366942734,0.3937572837,0.0932658836,-0.0825445727,0.1525270343,0.0257431716,0.0043484732,-0.0181564875,0.0768706128,0.4259207845,0.3218758106,-0.0369861759,0.0271981694,0.3717030287,0.2110736221,0.0548550002,-0.1873320192,0.1108266041,0.4963282049,0.1157274246,0.0822515935,-0.1317607909,-0.0508645326,0.2979769707,0.1687548608,-0.2600298524,-0.6055720448,0.2507573962,0.0219843611,-0.007665365,0.2330774665,-0.0285571907,-0.1289513558,0.0609241389,0.1440113038,-0.7042905092,0.1790257096,0.0768935233,-0.170310691,-0.0954483673,-0.1197600514,-0.1216633692,0.3036922812,-0.2471762896,-0.1520445347,0.1871303916,-0.1039698794,-0.0682562292,-0.0682849586,-0.0093313809,-0.1152828783,0.0440150984,-0.3456337154,0.0185122639,0.0293475483,0.2506072521,-0.2507233918,-0.267932415,-0.0913451836,-0.0071544861,0.1249451339,0.2417454869,-0.007397505,0.3755143285,-0.1992133558,0.2434767932,0.282722801,0.0155903231,-0.1905139834,-0.1179165095,0.0238807704,-0.0235805847,0.043543987,0.0319900215,0.1436030865,-0.0923393294,0.1365600228,-0.320015043,-0.0451029986,-0.0297245719,-0.3586837649,-0.036515478,-0.0903935507,0.1915640086,0.0737377703,0.010656327,0.1889198869,-0.0424713157,-0.1706620604,0.1441256702,-0.235605821,0.0102608604,-0.2453027666,-0.1029474437,-0.0885342658,-0.3024118543,0.0760996938,0.0420689322,0.1523383707,-0.2104491144,0.2786490917,0.1540673226,0.3944187462,-0.2255447656,-0.2214528769,-0.2535198629,0.3299390674,-0.1207605526,-0.4436139166,0.3380658627,0.2211083919,0.0755026266,0.0768547654,-0.0015967459,0.32285586,0.1641940624,0.2143816799,-0.2751164138,0.2191844583,0.3376360834,0.0541074239,-0.1252539158,-0.0836514011,-0.0990648493,0.0410788544,-0.00585037,0.3779468238,-0.2007446736,0.4935574532,-0.0686324537,0.5453831553,0.0227312129,-0.4466867447,0.2302495539,-0.0553989448,0.1944939196,-0.1707660854,-0.2438559234,0.5023388863,0.2270676196,-0.1800374985,-0.054065872,-0.4180530906,0.026112197,0.0925810486,0.1768018156,-0.0753671154,-0.1083805412,0.1398132145,-0.0579094961,0.2643485367,-0.4492532015,0.3536348641,-0.1488878727,0.1971689165,0.2340363115,0.346843034,0.2115702182,-0.1303756237,0.1627738923,-0.1963834614,-0.5711723566,0.4759761989,-0.1632489562,0.3026893139,-0.0298790298,-0.1258692294,0.2209758908,0.0263586473,0.6625293493,0.177660957,-0.3577192724,0.4814569652,0.1116923094,-0.5040145516,-0.1274021417,-0.1381823868,0.3122319579,0.0321445353,0.412440598,-0.1865827441,-0.341437608,-0.3256055415,-0.0043906933,-0.245097667,-0.2345226407,-0.3622905314,-0.2913392782,-0.238020733,0.3276261389,-0.0114638526,0.3348127306,-0.1287940294,0.0263398364,-0.494604528,-0.1166376993,-0.0544775389,-0.0450536944,0.3728664517,-0.1497163326,0.0129617117,0.0797666982,0.3228396773,0.4381982684,-0.012070247,0.2011637539,0.1447052062,0.1515206546,-0.0496981405,0.0190384202,0.1335830241,-0.0207422022,0.1335235983,-0.2026960999,0.174954921,-0.450063318,-0.1119607165,0.2253980339,0.0185722522,-0.3241938353,0.0502264947,0.3169851899,-0.1579330862,0.0697199851,0.1620562226,0.192193225,-0.1824704409,0.710005343,0.3957362771,0.9556046724,-0.0562951341,-0.0954110026,0.1049596742,-0.4326781332,0.2646969557,-0.1638128608,0.3374716938,-0.5618676543,-0.0877074152,-0.0307710972,0.0842926279,0.2589505911,-0.1019880995,-0.4649597108,0.0948774293,-0.4682738483,-0.1472354531,-0.2312412113,0.0857866332,0.0226013865,-0.3416077197,-0.1620435715,0.1439854056,0.105465658,0.1562862843,0.0815876424,-0.0745507926,-0.1078782529,0.0366040617,-0.1978430897,-0.0010732736,-0.1509224921,0.3344519138,-0.1510011554,-0.3497722149,-0.0803262964,-0.2889745235,0.5595018864,-0.2791683376,-0.2173137516,-0.1567486078,-0.0037706664,-0.1009950787,-0.1617235988,-0.207930088,0.25107041,0.0504242592,-0.2247953117,0.0285593607,0.0856410787,-0.0402889401,0.0778372288,-0.1377175003,-0.3344858289,-0.302164346,-0.4617123902,-0.0396448188,0.2221156955,-0.0165144037,0.1410473138,0.3233231306,-0.2969527841,0.2007445842,-0.3336692452,-0.0935504064,-0.0683642328,0.4073933661,0.0004040862,0.0350844748,0.3433612883,-0.1498123109,-0.2154189497,-0.1667424589,-0.2526399493,0.223308146,-0.0360389091,0.0633851811,-0.1675495654,-0.2924132645,-0.0721021146,0.1874263138,-0.0347481072,-0.0919861421,-0.5462034345,-0.0152212577,-0.4797928333,0.0193543099,-0.1716100574,0.3138584793,-0.2071558833,0.214841336,-0.0435705557,-0.2430595607,-0.3138628304,0.1492476314,-0.3146577775,0.2671282887,-0.3454406559,-0.1002850756,0.1255985945,0.015047715,0.1757668257,0.0711520016,-0.0695873797,-0.1896984428,-0.1923274249,0.1164015979,0.1130902842,0.1149253845,-0.156016916,-0.1207174212,-0.2156414688,-0.0848517567,0.0179863442,0.329067409,-0.0386980958,0.1870074272,0.140423581,0.1338136792,0.0028008157,-0.0495496839,0.3241001666,0.3158858418,-0.0934157819,0.0927629918,-0.382597357,0.0782682374,-0.2551828623,0.0975089148,0.3156036139,0.0965930745,0.0957464799,-0.0200307276,-0.1255624145,0.2182789892,0.0375530273,0.3016256094,-0.1857342422,0.015393124,0.2364439815,0.2276838273,-0.0599945635,0.0629108548,0.0485123992,-0.2579556704,-0.0974107161,0.144182682,0.3739403188,0.062202584,-0.0634855106,-0.0815519914,0.2122589499,0.0998330414,0.0115573574,-0.2535571158,0.1916528791,0.1791359186,0.2260335088,0.0954765975,0.4428193867,0.5543105006,-0.0909990594,0.2256232947,0.5728285313,0.0962594077,0.0338152945,-0.0322630033,0.2113486379,-0.1478863955,-0.0385961458,-0.1483103782,-0.0765078068,0.090602234,-0.463067472,-0.4397488236,0.0545198508,0.2481703907,-0.4364412725,-0.0726338848,0.1797018796,0.1748315096,-0.0361979082,-0.2937135994,0.1201873198,0.0101771457,0.0292139091,0.0758814886,-0.1556364447,0.1881146431,-0.2491922081,0.4484763741,0.3650005162,0.0280819815,0.0773645267,-0.0085192919,-0.1375135332,0.0097256694,0.2962383926,0.5330884457,0.4502035677,-0.3733889759,0.3845908642,0.218331039,-0.055716645,0.135407567,-0.2168995291,-0.0477172472,0.1838824153,0.1874000281,0.1257199049,-0.0600480959,-0.351975441,-0.0242308453,0.4029407799,0.0626426488,0.0988755822,-0.3894557059,-0.0816499442,-0.2280411124,-0.0090593714,-0.3260864019,0.0363086052,0.1406638771,-0.2598185837,0.0531366579,0.1167775393,0.0698552057,-0.1794427782,0.1243246272,-0.0248709787,0.0554880723,-0.3125730157,-0.1278846264,-0.1540424824,0.2752576172,0.2603755295,0.1083726138,0.126684621,0.5640339255,-0.036120519,-0.2216990888,0.2161679566,-0.1263417006,0.4234747887,0.354295671,-0.0601988435,0.1872889549,-0.5837618113,-0.2151583284,0.0077524204,0.0600415654,0.264893055,0.0928942487,0.1579354852,-0.0195696447,-0.0677329004,-0.0965492502,0.0310988761,0.7332779169,-0.0058208136,-0.0975356698,-0.0607817397,-0.2373835295,0.1060011685,-0.0161630884,-0.2784540057,0.0840936899,0.0688180402,0.0703045204,-0.0650185943,-0.2544137239,-0.3064646423,0.6995818615,-0.0112484815,-0.3947990239,-0.1250035614,0.3383899033,-0.302824676,-0.194757849,0.3656844795,0.1907244325,0.0350593701,0.4770624638,-0.2215889394,-0.3079065681,0.4277323186,0.0150559358,-0.1008341908,-0.0568256155,0.2920419276,0.348646313,0.3909554482,-0.8917971253,-0.0410390683,0.1061531156,0.0174173359,-0.1242297515,-0.0979671106,0.0490947962,-0.0471265502,-0.1576712877,0.0149087012,0.1378743947,0.3350586593,-0.0517826118,-0.1984497458]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1859","title":"Error \"in void don't know how to serialize this type of index\" when saving index to disk when device=0 (GPU)","comments":"Thanks for such a quick fix and merge to master, pip installed git master, tested all OK","body":"Error serializing faiss index. Error as follows:\r\n\r\n`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at \/home\/conda\/feedstock_root\/build_artifacts\/faiss-split_1612472484670\/work\/faiss\/impl\/index_write.cpp:453: don't know how to serialize this type of index`\r\n\r\n\r\nNote:\r\n\r\n`torch.cuda.is_available()` reports:\r\n\r\n```\r\nCuda is available\r\ncuda:0\r\n\r\n```\r\n\r\nAdding index, device=0 for GPU.\r\n\r\n`dataset.add_faiss_index(column='embeddings', index_name='idx_embeddings', device=0)`\r\n\r\nHowever, during a quick debug, self.faiss_index has no attr \"device\" when checked in` search.py, method save`, so fails to transform gpu index to cpu index. If I add index without device, index is saved OK.\r\n\r\n\r\n```\r\ndef save(self, file: str):\r\n \"\"\"Serialize the FaissIndex on disk\"\"\"\r\n import faiss # noqa: F811\r\n\r\n if (\r\n hasattr(self.faiss_index, \"device\")\r\n and self.faiss_index.device is not None\r\n and self.faiss_index.device > -1\r\n ):\r\n index = faiss.index_gpu_to_cpu(self.faiss_index)\r\n else:\r\n index = self.faiss_index\r\n faiss.write_index(index, file)\r\n```\r\n","comment_length":17,"text":"Error \"in void don't know how to serialize this type of index\" when saving index to disk when device=0 (GPU)\nError serializing faiss index. Error as follows:\r\n\r\n`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at \/home\/conda\/feedstock_root\/build_artifacts\/faiss-split_1612472484670\/work\/faiss\/impl\/index_write.cpp:453: don't know how to serialize this type of index`\r\n\r\n\r\nNote:\r\n\r\n`torch.cuda.is_available()` reports:\r\n\r\n```\r\nCuda is available\r\ncuda:0\r\n\r\n```\r\n\r\nAdding index, device=0 for GPU.\r\n\r\n`dataset.add_faiss_index(column='embeddings', index_name='idx_embeddings', device=0)`\r\n\r\nHowever, during a quick debug, self.faiss_index has no attr \"device\" when checked in` search.py, method save`, so fails to transform gpu index to cpu index. If I add index without device, index is saved OK.\r\n\r\n\r\n```\r\ndef save(self, file: str):\r\n \"\"\"Serialize the FaissIndex on disk\"\"\"\r\n import faiss # noqa: F811\r\n\r\n if (\r\n hasattr(self.faiss_index, \"device\")\r\n and self.faiss_index.device is not None\r\n and self.faiss_index.device > -1\r\n ):\r\n index = faiss.index_gpu_to_cpu(self.faiss_index)\r\n else:\r\n index = self.faiss_index\r\n faiss.write_index(index, file)\r\n```\r\n\nThanks for such a quick fix and merge to master, pip installed git master, tested all OK","embeddings":[-0.1146762744,-0.3390221596,-0.0211910177,0.1105611771,0.3647998273,0.1579398662,0.2933622599,0.5287505984,0.504611671,0.3331744075,0.0193728264,0.0713495985,0.1117654741,-0.0630019233,-0.1396043301,0.1634267718,0.4838933647,0.2276034504,0.33570081,-0.0381595157,-0.401009798,0.1175061092,0.0855751932,0.0203524493,-0.0746072978,0.027452521,-0.0272676796,-0.0571344234,-0.1761827022,-0.430722177,-0.0974790081,-0.2497538477,0.3530046046,0.2987298071,-0.0001167819,0.0765003487,0.2537879348,-0.2185033411,-0.1212787703,-0.1918153167,0.3125521541,0.0238258801,0.2112206519,-0.3889563382,-0.2219991833,-0.2291416824,-0.0148390858,-0.0824066401,-0.0860950053,0.2563896179,0.1728011221,-0.1268796772,0.426910609,-0.0993105397,0.048668839,0.1324171722,-0.2699905038,-0.081328027,0.0310128517,0.4924758375,0.3509413004,0.3124850094,-0.0470326245,-0.3270618916,-0.1348573714,-0.0353984609,0.6270079613,-0.4113226831,0.1714873165,-0.1755583286,0.0252775438,-0.2577891648,-0.3364470005,0.2810265422,0.1917248368,-0.4463648498,-0.1149210483,0.0594530776,-0.2262235135,0.1179195642,0.6050459146,-0.2198892385,0.0826855674,-0.1929790527,0.2028477788,0.0529569313,-0.1074484065,0.0764327869,0.0183136519,-0.1368501484,-0.2525518239,0.0414270312,0.1735977978,0.1504006833,-0.4110235572,-0.1585874707,0.0316705741,-0.4751079082,-0.1077526957,0.1082183123,-0.3547216654,0.0536165573,0.2421877533,0.0392326675,-0.302978158,0.0532989688,-0.0838188231,-0.1678367406,-0.0528734811,0.0116879614,0.0138864294,-0.3690967262,-0.1901766658,0.0614837259,-0.3394028842,0.1539418995,0.2605766356,-0.3318852782,-0.5334686637,0.1121128052,-0.0934316814,0.224335596,0.1077442244,0.3150539994,0.3123532832,0.0913958251,0.2175917774,0.3172948658,-0.2520931661,0.1195910946,-0.1601220071,-0.1714918613,0.2880524397,0.0833575502,-0.0394231007,-0.4399599731,0.0343991257,0.1377116144,-0.1891623735,0.0371744297,0.0833467767,-0.3589977026,0.0460222997,0.4789451957,0.0850206465,0.1684921682,0.3972036839,-0.1274073869,-0.1334162652,0.3661162853,-0.0772539601,-0.192571193,-0.5117437243,0.1412725896,0.2128684968,-0.0281150304,0.1784328818,0.031009173,0.1252801865,-0.0399937928,0.0840364173,0.1249946132,0.1171110272,-0.2871429026,0.4271494448,-0.0141213136,-0.035585437,-0.0640763864,0.1413884312,-0.1511024386,0.2421639115,0.5180215836,0.4163944721,0.2536924183,-0.2635124922,0.2852970064,0.4269224405,-0.3314129412,-0.1441108733,0.0950621888,-0.0796561539,-0.3813628256,0.2808741033,0.0293669421,0.2075395882,-0.1095142886,0.1899430454,-0.1288816929,-0.0122308601,-0.1432453096,-0.3757482767,-0.3430812955,0.3114051223,-0.0660417452,-0.0519273803,0.1272798926,-0.1354821026,-0.5923423767,0.1196643636,-0.1409867257,-0.0515021943,-0.0383368507,0.9742364883,0.0753188357,0.2244439572,-0.0890067667,0.1671759337,-0.082271859,0.0197584089,-0.1304292828,-0.4603835642,-0.0925274566,0.0079037333,-0.1039909199,0.2397685349,0.0991542861,0.0758389011,-0.1709728241,-0.1432718188,0.1668157279,-0.296772778,-0.1551455408,-0.1298324615,0.1190659776,0.069975771,0.1621151268,-0.3236829937,-0.4899028242,-0.2084936202,0.2800624073,0.363855809,-0.2614922225,-0.0434338637,0.3414683044,0.0304946825,-0.0277129523,0.1776825786,0.097763218,-0.044090502,-0.0421458036,-0.0168216266,0.5092594028,0.4029051363,-0.0575097315,0.0993845984,0.2719661593,0.1814074367,0.0969095379,-0.1989655048,0.0933970511,0.4091677666,0.0790104792,0.092893891,-0.1172558963,0.0038422062,0.3509766757,0.1438655257,-0.2183193564,-0.5096514821,0.1285246015,-0.0430186875,-0.0327102318,0.2649704218,0.0011569771,-0.1233770475,0.0679899082,0.0005137897,-0.7569183707,0.1782088876,0.035366565,-0.2462668121,-0.1474381536,-0.0495026298,-0.1168398783,0.3600614965,-0.1883478165,-0.1926207244,0.2656208873,-0.0701297,-0.0270441975,0.018068919,0.0745554492,-0.1314165741,0.0339823961,-0.3221395612,0.0470151119,0.1045816466,0.2126232237,-0.2320215255,-0.2859646082,-0.0888023004,-0.0023957214,0.1029872373,0.2418350726,-0.0353085399,0.2771553099,-0.1721949875,0.1879617125,0.2446773946,0.0301045831,-0.2019225508,-0.1163958833,0.0249054618,-0.0574817173,0.056234289,-0.0197934061,0.1009035558,-0.1188319772,0.0695459992,-0.2965686321,-0.0026126639,-0.0111070527,-0.3667127788,-0.0308046229,-0.1012416705,0.2145470828,0.0998611748,0.0570703968,0.2294715494,0.0136418613,-0.2189856768,0.1472661495,-0.2851096988,0.0995737687,-0.2638249099,-0.031423822,-0.0933679268,-0.277186662,0.046300251,-0.0066175321,0.1611276418,-0.1750767231,0.2500221431,0.1096313223,0.3826157749,-0.1994668096,-0.1604940444,-0.2403205186,0.2455213368,0.0141066639,-0.4077498913,0.3803664148,0.2327859402,0.0825452954,-0.0168326776,0.0531429425,0.2793097496,0.1768025309,0.1367385536,-0.3660625219,0.2124079615,0.3130120337,0.0957227051,-0.1149212345,-0.0534125566,-0.026912706,0.0467293113,0.0268329103,0.3343476951,-0.1983141303,0.4334018528,-0.1179565266,0.6137663722,-0.0294546019,-0.3339177072,0.2581992447,-0.0253974665,0.1039876044,-0.1415958405,-0.2537579536,0.4257482886,0.305231154,-0.2542210817,-0.0381653383,-0.4093991816,0.0523647368,0.1372279525,0.2153998613,-0.0554612353,-0.0866810754,0.1114593297,-0.0794077292,0.2802918851,-0.5015254617,0.2859975994,-0.133761391,0.1897230446,0.2340593785,0.4296455979,0.2158713043,-0.1023618504,0.1824279577,-0.3492436409,-0.5001865625,0.4314860702,-0.115353018,0.2361050993,-0.0570210628,-0.0995658711,0.2641097903,0.0280012637,0.5748898387,0.2505704761,-0.3606672883,0.4638264477,0.1107289419,-0.4079251587,-0.1878312975,-0.111082159,0.3629105091,0.0571906865,0.4573310614,-0.2006698847,-0.3740849197,-0.3333072364,-0.0210817493,-0.2284048647,-0.2749685645,-0.4151601791,-0.3075054288,-0.1949049383,0.3872945309,0.0377008058,0.3573037386,-0.1309336573,0.1222716719,-0.4832380414,-0.0675470009,-0.0815928727,-0.2251808196,0.3386273086,-0.2394334376,-0.0769014806,-0.0219500475,0.2859575152,0.4177362323,-0.0491309538,0.3201033175,0.1861423403,0.1888227612,0.000461801,0.0391882434,0.1098860726,0.0540728271,0.1841254681,-0.1428143978,0.2230829895,-0.4668781757,-0.1193054095,0.2227195203,-0.0280994549,-0.263017416,0.1655523628,0.3173095882,-0.1281889826,0.1277455389,0.1194433868,0.1563385874,-0.1931025833,0.6817430854,0.3997468948,0.9798754454,0.0252916664,-0.0914613828,0.0858017281,-0.416723758,0.2120485455,-0.1581191421,0.3874966204,-0.5338541865,-0.0146614769,-0.0704162866,0.0637785792,0.2272361964,-0.2236260027,-0.4985609055,0.120772481,-0.4409584105,-0.1280460507,-0.1978638917,0.0189629961,-0.0129705705,-0.2784303725,-0.1050788537,0.0792925581,0.0551649854,0.2143960595,0.0763614848,-0.1182075143,-0.1765968353,0.0644143522,-0.23040241,-0.0265106522,-0.1730618924,0.3446368873,-0.0993860736,-0.3877467215,-0.1328043789,-0.3671295643,0.4722646475,-0.2943333983,-0.2061363012,-0.18809174,0.0174351297,-0.0513620302,-0.1856402755,-0.1617178023,0.2138828337,0.0408917665,-0.1677394658,0.016889615,0.0026912563,-0.0189810582,0.0521006621,-0.1252658367,-0.3506316543,-0.3085312247,-0.4875867665,-0.0376002267,0.2412182093,0.0331749693,0.1144584641,0.3455030322,-0.3167863488,0.1765204817,-0.3581039011,-0.1160884425,-0.0079591209,0.364125073,0.0735053122,0.0168952122,0.4057067633,-0.2747248113,-0.1524932534,-0.1529943645,-0.2313351184,0.131937772,-0.0770363063,0.1316164881,-0.1380922645,-0.3396422565,-0.0646581948,0.2435810119,0.0010776407,-0.0346703716,-0.6063767076,0.0421925634,-0.4536468983,0.0480275601,-0.1868368685,0.2728901803,-0.2377519161,0.2444854528,-0.0362342522,-0.3157740235,-0.2671248019,0.1573026627,-0.3896790445,0.2999816835,-0.3915306628,-0.1758677959,0.0934170112,0.0215161685,0.1660155207,0.0189267043,-0.0389734767,-0.163183555,-0.1572702527,0.1302211285,0.0838363543,0.16329059,-0.1456517279,0.0065036872,-0.1009747237,-0.0033866602,-0.0315148421,0.2484490722,-0.0204781517,0.1504319608,0.095167622,0.2308656573,-0.0633921623,-0.0546432808,0.278667897,0.2993788421,-0.1352014989,0.0992387831,-0.3457974792,0.1191849932,-0.3002997339,0.0524715036,0.3520539999,0.1087465286,0.098834835,-0.0381993391,-0.1049235016,0.2200045288,-0.0865344852,0.219607532,-0.1954570562,0.0227086097,0.1914832741,0.2191763967,0.0411762781,0.1304051131,0.0471723936,-0.2500390708,-0.1444497705,0.1082358286,0.3742764592,0.0224191807,-0.1011441723,-0.0451785959,0.2204563767,0.1478135884,-0.0020097869,-0.209974885,0.243030414,0.2296785712,0.1241613552,0.0390943773,0.3866074979,0.4460996985,0.0162555482,0.1503826678,0.5617634058,0.0576331913,0.0258298647,0.0887884796,0.1186500415,-0.1494784504,-0.0015852492,-0.1095031723,-0.0852054656,0.017559303,-0.4877850711,-0.4000405371,0.1134632975,0.2814671993,-0.4551028907,-0.0638953447,0.1260296702,0.2360826582,-0.005589617,-0.3023852706,0.1427742541,0.1391485482,-0.0389774665,0.1191958785,-0.1901588887,0.3054991663,-0.3720206022,0.5323474407,0.4078015089,-0.0345836803,0.0908470824,0.0159076266,-0.1674892157,-0.0154257119,0.4096110761,0.5584622025,0.4596506655,-0.4080401361,0.3030127585,0.2462727129,-0.0842656642,0.1901925951,-0.2424072474,-0.0100639202,0.2692169547,0.0473013856,0.0936596245,-0.0415211543,-0.2905817926,-0.0935946777,0.291346401,0.0783217475,0.1853542924,-0.3981410563,-0.157191053,-0.1800523549,0.0423380435,-0.353824079,0.04928286,0.0853732154,-0.1960124671,0.0353722908,0.1670022905,0.0526831262,-0.1624989957,0.0931417495,0.0648179054,0.0099918079,-0.3806734085,-0.0841453448,0.0213937461,0.2346542925,0.2559273243,0.1386160403,0.1663260013,0.553581655,-0.0884348601,-0.2022334337,0.1693255007,-0.1345931143,0.4113422334,0.3900162876,-0.0215614438,0.170901075,-0.4755473137,-0.2536104023,0.0076707364,0.1050755605,0.2578587532,0.0936070383,0.1616384089,0.007990025,-0.0863537192,-0.0575325489,0.1221468896,0.6473215818,0.0037714019,-0.1314731389,-0.0434658378,-0.3159878254,0.2457980514,0.0247578491,-0.3558964133,0.0329874419,-0.029260492,0.0657521412,-0.0683964565,-0.2140351236,-0.3025994301,0.6512049437,-0.0724927559,-0.3786836565,-0.1211362928,0.2815712988,-0.2921560705,-0.2217398733,0.3673098385,0.1116155535,0.0860565007,0.519209981,-0.3003044128,-0.2952014506,0.3583335876,0.0133008463,-0.0779225901,-0.1047354862,0.3780346215,0.3131186664,0.4897960424,-0.8266086578,-0.050542444,0.1462625712,0.059830226,-0.106938459,-0.0695538968,0.1198457405,-0.0517983176,-0.0920977667,-0.009607425,0.119287014,0.3867179155,-0.0559398048,-0.2090591341]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1857","title":"Unable to upload \"community provided\" dataset - 400 Client Error","comments":"Hi ! We're in the process of switching the community datasets to git repos, exactly like what we're doing for models.\r\nYou can find an example here:\r\nhttps:\/\/huggingface.co\/datasets\/lhoestq\/custom_squad\/tree\/main\r\n\r\nWe'll update the CLI in the coming days and do a new release :)\r\n\r\nAlso cc @julien-c maybe we can make improve the error message ?","body":"Hi,\r\ni'm trying to a upload a dataset as described [here](https:\/\/huggingface.co\/docs\/datasets\/v1.2.0\/share_dataset.html#sharing-a-community-provided-dataset). This is what happens:\r\n\r\n``` \r\n$ datasets-cli login\r\n$ datasets-cli upload_dataset my_dataset\r\nAbout to upload file \/path\/to\/my_dataset\/dataset_infos.json to S3 under filename my_dataset\/dataset_infos.json and namespace username\r\nAbout to upload file \/path\/to\/my_dataset\/my_dataset.py to S3 under filename my_dataset\/my_dataset.py and namespace username\r\nProceed? [Y\/n] Y\r\nUploading... This might take a while if files are large\r\n400 Client Error: Bad Request for url: https:\/\/huggingface.co\/api\/datasets\/presign\r\nhuggingface.co migrated to a new model hosting system.\r\nYou need to upgrade to transformers v3.5+ to upload new models.\r\nMore info at https:\/\/discuss.hugginface.co or https:\/\/twitter.com\/julien_c. Thank you! \r\n```\r\nI'm using the latest releases of datasets and transformers.","comment_length":54,"text":"Unable to upload \"community provided\" dataset - 400 Client Error\nHi,\r\ni'm trying to a upload a dataset as described [here](https:\/\/huggingface.co\/docs\/datasets\/v1.2.0\/share_dataset.html#sharing-a-community-provided-dataset). This is what happens:\r\n\r\n``` \r\n$ datasets-cli login\r\n$ datasets-cli upload_dataset my_dataset\r\nAbout to upload file \/path\/to\/my_dataset\/dataset_infos.json to S3 under filename my_dataset\/dataset_infos.json and namespace username\r\nAbout to upload file \/path\/to\/my_dataset\/my_dataset.py to S3 under filename my_dataset\/my_dataset.py and namespace username\r\nProceed? [Y\/n] Y\r\nUploading... This might take a while if files are large\r\n400 Client Error: Bad Request for url: https:\/\/huggingface.co\/api\/datasets\/presign\r\nhuggingface.co migrated to a new model hosting system.\r\nYou need to upgrade to transformers v3.5+ to upload new models.\r\nMore info at https:\/\/discuss.hugginface.co or https:\/\/twitter.com\/julien_c. Thank you! \r\n```\r\nI'm using the latest releases of datasets and transformers.\nHi ! We're in the process of switching the community datasets to git repos, exactly like what we're doing for models.\r\nYou can find an example here:\r\nhttps:\/\/huggingface.co\/datasets\/lhoestq\/custom_squad\/tree\/main\r\n\r\nWe'll update the CLI in the coming days and do a new release :)\r\n\r\nAlso cc @julien-c maybe we can make improve the error message ?","embeddings":[-0.091571033,-0.0363663472,0.0344169848,0.0672418028,0.364969939,-0.0492849946,0.1734051406,0.0044843452,-0.2230454385,-0.1468119323,-0.1488375515,-0.0205998104,0.0050854953,0.3581641018,-0.057110548,-0.1069303602,-0.0269479491,0.0025079621,-0.4462412596,0.0053096106,-0.1092077494,0.1308574378,0.1841463447,0.0065866653,-0.3465132713,-0.075896278,-0.0069127912,0.2234145552,-0.2710898221,-0.1376961768,0.4473814964,0.1389370859,0.2372483611,0.5959687233,-0.0001169556,0.0284550507,0.354206264,0.0238105897,-0.6256381273,-0.3368651569,-0.0726407021,-0.194068566,0.0081896242,0.1364579648,-0.1094309613,0.0796100572,0.1506982744,-0.0484017506,0.598502934,0.0964820758,0.1131024733,0.2700642943,-0.0149941603,-0.2242386788,-0.2202052027,0.3592706919,-0.1043352261,0.2172357738,0.2370122522,0.0222127046,0.1814651936,0.0335429125,0.1154125184,-0.116887629,0.3684593439,-0.1654307693,-0.4199377298,-0.2600347996,0.1162151769,0.3317759335,0.5857560635,0.0770779029,-0.5686392784,-0.2074169666,-0.0260345507,-0.0701275766,0.1815169454,0.0815205425,-0.0584621355,0.1197283491,-0.4374508262,-0.1708433479,-0.3639187515,-0.0612374917,0.1857609749,0.2177210599,-0.1954705566,0.2356920242,-0.1176391691,-0.5343660712,-0.122730352,-0.0399159156,0.1515682936,0.2014334649,-0.324742347,-0.1058338061,0.0912819654,0.2499752045,0.3493946195,0.4778476357,-0.1200058162,-0.3294019997,-0.1893023849,-0.0660497323,0.2054422647,-0.1204180494,-0.0417398587,0.0452685431,0.0842115879,0.4836541414,0.3961295187,-0.0219664145,-0.2512885332,-0.0405339673,-0.2929287255,-0.1570432335,0.4355102181,-0.0129264444,-0.0357220136,0.0570332743,0.2894047797,0.0918491185,0.0366360471,0.2619793415,0.0178698115,-0.0107141742,-0.0824559852,0.0342847966,0.0659229606,-0.363963455,-0.039450489,-0.1896135211,0.0428131297,0.0224771313,0.151420787,-0.2134020776,0.093679443,-0.0295447558,0.4580491185,0.1253284216,-0.1442784071,0.1845533401,-0.1110058203,0.384172976,0.1887934208,0.4600846469,0.0868834704,-0.1932172775,-0.0594399013,-0.1539819241,-0.2070763707,-0.4029405415,-0.4708251655,0.027756175,0.207651481,-0.1514526904,-0.3534536064,-0.1519583464,0.0665723681,-0.0373009555,-0.0678822473,0.1048246026,-0.0090033179,-0.0408445485,0.2429062277,0.629003942,-0.1543011516,-0.136814177,-0.0070554758,0.2116970271,-0.1128645837,0.4283787012,-0.1202709675,-0.3136930466,-0.1316830069,-0.0307029951,0.1514888108,-0.4831223786,-0.5233066678,0.3477802873,-0.1778277159,-0.1530867815,0.0853089914,0.1088344157,-0.1908691674,-0.0401852429,-0.0610545985,0.346468389,0.0755402148,0.1005179435,-0.1324571073,-0.333335191,0.0300015826,0.0861646608,-0.2761984169,-0.0550341159,0.0439736955,0.005162186,0.1278568059,-0.1416758746,0.2658134103,0.3532611728,0.6264160872,-0.0552132018,0.139233157,0.1916906089,-0.465521425,0.155237928,-0.0301758368,0.1750134081,-0.3612290919,-0.0907869637,-0.4938573837,0.1524475664,-0.2330334187,0.1688339859,0.0031360392,-0.1407907456,0.0996627361,-0.2074013501,-0.0429635309,0.4761528075,-0.1568289548,0.3111292422,-0.4384240508,0.2664834261,-0.3046866059,-0.0398059785,0.3985339999,-0.1630729884,0.3846707046,0.0297985319,-0.1488204747,0.3252856135,-0.4254599214,0.2014298141,0.1874761432,0.3789373934,0.3259228766,-0.191376701,-0.1845211387,0.0234061237,-0.0551307797,0.3489126265,-0.3237706721,0.006853892,-0.1706615984,0.2133798748,0.004942392,0.3508688211,0.3105170131,0.1342519224,-0.1620022357,0.1888069063,0.1251220405,-0.0281079262,0.4954760969,-0.1988340467,0.1358230412,-0.0043301769,0.0218349788,0.0227265861,-0.0362495445,0.1160141081,-0.1007650346,-0.0427422561,0.2822441459,0.3077149093,0.0753518417,-0.0590367354,-0.2393871993,-0.0900595412,0.1027620137,-0.0979990289,-0.0190006904,-0.1057624519,0.1955442876,-0.0377549864,0.2462148517,0.2434728742,-0.2217589319,-0.1577916741,-0.0150159663,0.1437564641,-0.3067656159,0.0647429004,-0.0432588868,-0.2571393549,0.1904011667,0.0039525875,-0.5359175205,-0.1126174256,-0.2029633969,0.0531677715,0.0151663981,0.1205603629,0.2327250093,0.3003382683,0.3323481083,-0.0865685493,-0.3027799726,0.2590976655,0.0402396768,0.0501028784,0.1002198085,-0.4291399121,0.1166410148,-0.0433383435,0.352088809,-0.3359856606,-0.3338156939,0.1310446858,-0.1267152727,0.49113217,0.1695733815,0.4597931206,-0.1194784939,-0.1157093123,0.241499275,-0.1196787581,-0.0532944575,0.0056243422,-0.168696925,-0.1025479734,-0.0004041885,0.1713876575,-0.0893216282,-0.2928875089,0.3449076414,0.1513797641,-0.0110914605,0.3846877813,0.1992631555,0.0767356679,-0.0966850817,0.1581422091,-0.0142404111,-0.5848600864,0.381619513,-0.2026971728,-0.3429449499,-0.0246028695,0.1481397301,0.2809788287,-0.2376184613,-0.1241214797,-0.4074200094,-0.1366187483,0.2894022167,0.0676823482,0.0077136569,0.3739700317,-0.0090630921,0.0658066645,-0.0294284821,-0.4492641687,-0.1344391704,0.3306126893,0.3022273779,-0.0583615676,0.1823432446,0.0515648089,0.8003917933,0.520750463,-0.1554859281,0.4999686778,-0.0043584011,0.3545721769,0.0832152739,-0.153094843,-0.0170857105,0.2352315933,0.039930027,0.0317251831,0.2795096338,0.2463772148,-0.1541664004,-0.16770643,-0.1231180727,-0.0498752855,-0.3460411727,0.0586698167,0.1963111162,0.1161699742,-0.2394078672,-0.1919602007,-0.2947542965,0.0026092133,0.36731264,0.320844084,0.1394205093,-0.343528688,0.0137963761,-0.4768286347,0.2295169383,-0.0217435546,0.0656604841,-0.2315968573,0.1908889264,-0.0122329062,-0.0165007319,0.4541062117,-0.4889232814,-0.0582799651,-0.431142956,0.2210528105,-0.3148051202,0.196871236,0.2207861692,-0.0102750668,0.1239133924,0.49034518,-0.3415438831,-0.0220447127,-0.0865177065,-0.1912461519,-0.1529007703,-0.159262538,-0.2356317937,-0.4699708223,-0.4909074008,-0.2192814052,0.2149493843,0.1797940582,0.0512502566,0.1261921674,0.3211120069,0.1570871323,0.1690648347,-0.0818528607,0.1993985623,0.3902917504,0.0657231063,0.0789734349,0.2132924795,0.345892489,0.7234184146,0.0248379074,-0.4888086021,0.0911087021,0.2046039999,0.2360623926,0.2299363166,0.1936471611,0.0403192192,0.4569897652,0.2494083941,-0.2306302488,0.1356469691,0.3005810678,-0.2050448209,-0.3913976848,-0.0677413493,0.4168677628,-0.1035319567,0.1218625531,-0.0426351279,0.4895393848,0.0788981169,0.0654914901,0.1265354306,1.035789609,-0.2924604118,0.2455256581,-0.1198339611,-0.5208818316,0.7754577994,-0.0201138239,-0.1492321193,-0.4048445523,-0.1508547068,-0.0616820231,-0.0208097734,0.036529012,0.0276635084,-0.1775850207,0.0893840864,-0.3146385849,0.1403439641,-0.0592532121,0.5753946304,-0.3577100933,-0.2095098048,-0.2162095755,0.1285811663,-0.1896327585,0.199802354,-0.0824748278,-0.2414723337,-0.2121481895,-0.3160802424,-0.2313428968,0.1291580051,0.008255423,-0.0934870988,0.4783602953,-0.2157315761,0.3904020488,0.2486741394,0.2368652225,0.0964201242,-0.4975145757,0.2585036457,-0.3437275589,-0.1752979755,-0.1725774705,0.0312004276,0.1753028035,-0.0954129174,-0.0261518117,-0.1201224476,0.0313037112,-0.0615019239,-0.2928437889,-0.1749279946,-0.1124114618,-0.5553346872,0.1550160646,-0.2224547863,0.0837304518,-0.0646316782,0.0215722695,0.0818035901,0.0439640433,-0.398160845,0.5290651321,0.0469098091,-0.007492593,0.3999333382,-0.2207385302,-0.1064387634,0.5063541532,0.0868058428,-0.0409621298,0.0550989509,0.0000201892,0.1553798765,-0.7436199784,-0.010905968,0.2651931942,0.0752898306,-0.1520250589,0.1781099886,0.2044998705,-0.0692892671,0.2705127895,0.0362765118,-0.2207256556,-0.0655042008,-0.3735987246,-0.0167082716,0.1623820812,-0.0873606429,0.43687886,0.0182802211,-0.217092365,-0.1081428751,-0.3802972436,0.0241258815,0.4103033543,-0.2786548436,0.0579219498,-0.0799691156,-0.0164358001,0.0136636039,-0.3531111479,-0.0814256072,-0.28954041,0.2039218843,0.3594950736,-0.3609602749,-0.1264454126,-0.1718744487,-0.1880278885,-0.0993752256,-0.1974612921,-0.0555646755,0.0901344717,-0.2200163603,0.1662517041,0.0927822962,-0.3215282857,0.1665591896,0.2040168494,0.1928849667,0.1309165657,0.134337008,0.0815955698,-0.0592803806,-0.1065384671,0.2511464059,0.4674801528,-0.019813681,0.3256031275,-0.3629270494,-0.073623091,-0.0257286876,0.4536920786,-0.0972912833,-0.3470792472,0.0412581638,0.0792223662,0.1049641967,-0.0659509823,-0.0650133565,0.0684915483,0.1194636449,-0.0792730525,0.0845511332,0.2587446868,0.0058865603,0.2516885996,0.1026485041,0.5311475992,-0.0180532765,0.0580717623,0.0093473354,-0.168247655,-0.0520484298,0.473230511,0.0507225469,-0.0654959455,-0.1305395216,-0.2922395468,0.0339563452,-0.3159373701,-0.0270484835,0.1887206435,-0.2837962806,0.2764680684,-0.1694781929,0.163800627,0.1581928432,0.0279098935,0.5838848948,-0.2007367313,0.02593619,-0.3901529312,0.1971658468,-0.0810164958,-0.2365208864,-0.1915607005,-0.1182718948,-0.2009131759,0.0433617197,-0.0816271454,-0.3174012303,0.1202808395,-0.0248026066,0.0807824358,-0.4145545959,-0.0669585988,0.4527414441,0.0931211188,-0.112456128,0.3525934219,0.366630286,-0.1180029511,0.0978855342,0.2067520469,0.6063757539,0.0039431006,0.2722609639,0.3868841231,0.255473882,0.1195414513,0.1990561485,0.0681631118,0.0598044582,0.2382235378,0.0997470319,0.0371566042,-0.1150524765,-0.1154233441,-0.4080911577,0.3273947835,-0.3633913994,0.120013006,-0.4371233881,-0.1757614464,-0.2189849466,0.1765093654,-0.2445448041,0.3110892773,0.5170588493,-0.0518473461,0.0027796603,-0.5339076519,0.0247855857,-0.0718070716,0.4974350035,0.6792612076,0.023111511,-0.0135587221,-0.2177400887,-0.3219079077,0.1379478127,0.0576829202,0.2602822185,-0.0974857733,-0.0088565201,-0.0073967506,-0.1254077852,0.1286444366,-0.2437612861,-0.2440421134,-0.0267158337,-0.0501533747,-0.0899466574,-0.0394924656,0.0243941713,0.0450284369,0.0611171909,0.2203259766,0.235164538,-0.0289455932,-0.2344932854,-0.1655996591,-0.1057446375,-0.2576320767,0.1671301723,0.3643932343,0.2754126191,-0.0435290076,-0.3752656579,-0.2265497893,-0.1034767777,-0.2364065945,0.0197883267,-0.0029046743,0.3410148621,0.0548066534,-0.1026082113,-0.3389887512,-0.1286716163,-0.0888457224,-0.1113579273,-0.1724304706,-0.1531215459,-0.2435727417,0.0413385928,0.1082698107,-0.1378850341,0.1694569737,0.0257008485,-0.2529639304,-0.3108346164,0.5455137491,-0.3568969965,-0.1854249239,0.0470241793,0.1835143268,-0.1315599531,-0.1658132523,-0.3201593459,0.1859065443,0.2529086471,-0.0348474868,-0.0790374056,0.2278553098,-0.1121339872,-0.0029700617,-0.0513183624,0.3933723271,0.1405708641,-0.221369043,0.2974922061,0.0010211039]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1856","title":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError","comments":"Hi ! This issue may be related to #996 \r\nThis comes probably from the Quota Exceeded error from Google Drive.\r\nCan you try again tomorrow and see if you still have the error ?\r\n\r\nOn my side I didn't get any error today with `load_dataset(\"amazon_polarity\")`","body":"Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```","comment_length":45,"text":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError\nHi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```\nHi ! This issue may be related to #996 \r\nThis comes probably from the Quota Exceeded error from Google Drive.\r\nCan you try again tomorrow and see if you still have the error ?\r\n\r\nOn my side I didn't get any error today with `load_dataset(\"amazon_polarity\")`","embeddings":[-0.1446352303,0.1181077734,-0.121307157,0.2410930544,0.128062129,-0.0035925508,0.344353348,0.0653002113,0.2865491509,0.23939161,0.0281222612,-0.0647892952,0.0097159557,0.1506800503,-0.1385413408,0.2059793174,0.0107034752,-0.0445432663,-0.1888069808,0.1773511469,-0.368370235,0.3735696971,-0.0902100056,-0.2902166247,-0.0568014644,0.0482145026,0.2492216974,0.1227689907,-0.0999508873,-0.3141925335,0.2892129421,0.0641637966,0.1367656142,0.3216083646,-0.0001227321,-0.0760453939,0.4037206769,0.0215304326,-0.3489623964,-0.3050134182,-0.3421134949,-0.2206279337,0.0789371431,-0.1413576603,0.1984624118,0.226269424,-0.0134389466,-0.0203781053,0.2310121506,0.1465004981,0.164618969,0.3472891152,0.4247147143,0.2376425117,0.336298883,0.0346841216,0.0294039361,0.4239823818,0.1685522795,-0.1180216447,-0.151582852,0.1364068836,-0.0350834653,0.4699849188,0.0731047764,-0.0709814727,0.1299203187,-0.3009367287,0.2216953486,0.2402327508,0.5259546638,-0.1580471992,-0.3384304941,0.0157191157,-0.2012751997,-0.3522212207,0.3682412207,0.3873554468,-0.0813419223,-0.2479474545,-0.4934211373,0.127337411,-0.1564018875,0.2325276136,-0.1409677267,0.1816923171,-0.0141904699,0.1430938542,0.2745828927,-0.256382674,0.4774478078,-0.1528468728,-0.0680743158,0.1682090759,-0.5600778461,-0.1012885347,-0.0274246838,0.4298371375,0.2686576247,0.5032892823,0.1269349754,0.2249221951,0.0311237425,0.2137777954,0.3584070206,0.2740326822,-0.2249949425,0.3298181891,0.1838583648,0.3255170286,0.0375998206,0.0343653485,-0.129116416,-0.3544429243,0.670812428,0.0256587323,0.3402949572,-0.4990726411,-0.4904987812,0.3258250058,-0.2727227211,-0.1630820781,-0.0561086535,0.1552018225,-0.1133430749,0.0365046225,0.0853408575,-0.0136228995,-0.1336150765,0.1910001934,-0.1091789976,0.1557029933,-0.1040167212,0.2002133131,0.3021281362,-0.1143079102,0.3013885915,-0.1079606935,0.3744319677,-0.2520242631,0.5605089664,-0.2755386233,-0.1011085734,0.5101605654,-0.0266280323,-0.0405234657,-0.0029191871,0.2849735022,-0.1475282609,0.2997645438,-0.2269744575,-0.2836966217,0.0125993993,0.1088242903,-0.512270093,-0.1845045388,-0.2347881645,-0.3862113953,0.2250039428,-0.3119521439,-0.077526167,-0.3307490051,0.030206101,-0.1781544983,0.1285468638,0.4043677747,-0.3409209549,0.0198289491,-0.3198651373,-0.0909231827,0.3040816784,0.3598715365,-0.1228100136,0.0390357226,-0.257070601,-0.3024713099,-0.1342588812,-0.1597056836,-0.4444053173,0.0475131422,-0.1507032216,0.2654606402,-0.0001137547,-0.000171975,0.0030774677,0.0456521623,0.2740553021,0.1631098688,-0.0405791737,0.1599233598,-0.162685439,-0.3306627274,0.2296402901,0.230322063,0.2603027821,-0.0007047705,0.2450987697,-0.1309292316,0.0556210726,-0.0284383371,-0.0195899066,0.3621577919,0.3402877748,-0.094012484,-0.0205230787,-0.0889068022,-0.5740942359,0.1598230749,-0.309572041,0.1014345512,-0.1388898939,0.1446614712,-0.1564300656,-0.2206208706,-0.0964881182,0.0483622141,0.0302265398,0.2930377126,0.1999130845,-0.1771686673,0.0637950823,-0.0643372759,-0.2649933398,0.0469801612,-0.382153064,0.4421381056,-0.0644131154,-0.0300862696,0.1348583996,-0.0121622793,0.1962991208,-0.1785702854,-0.0394668728,0.2556773126,0.099367097,0.3484787643,-0.0179110821,0.6057217717,0.0546131022,-0.0496350229,-0.053094238,0.4482397735,0.0615543276,-0.1253736615,-0.2357844561,0.541423142,-0.190831691,0.2926019132,-0.0797188357,0.0411804132,0.2328560054,-0.0834397301,-0.1304754317,-0.145867005,0.3699472845,0.1853739768,-0.1306202114,-0.0103301555,-0.028405631,-0.0796814859,0.2859376073,-0.1639742255,0.0210786089,0.0837235078,-0.0233296901,-0.1471817791,0.1689704061,0.723451972,0.4756741226,0.1242609769,-0.1934708059,-0.2028280646,-0.0221904963,-0.0442498028,0.1061170921,0.1723461747,0.0280860495,0.4393385947,0.2144734263,-0.1984520257,-0.3193229735,-0.2452728152,0.2342373133,0.3231551349,-0.4198349118,-0.0584383309,0.0039443872,0.1874771267,-0.2392983437,0.0287349336,-0.2095084935,-0.3734353483,-0.264025718,0.5433931351,-0.0794598758,0.0803993419,-0.2748629153,0.1581510007,0.1703451723,-0.4426254332,0.003161696,-0.0818632841,-0.0020267947,-0.0054456177,0.3528393209,-0.1443244964,0.3782065809,-0.1511015892,0.0522559211,-0.599588275,-0.310205847,0.0240140911,-0.0560188256,0.2117903382,0.1166588143,0.2397703528,-0.054990977,-0.0856558233,-0.014656513,0.1234324649,-0.1965638101,0.1444491893,-0.1703918427,0.0828320831,0.0760672092,-0.1404228806,-0.0595666803,-0.3863202929,0.1085538566,0.413972944,0.2004535645,-0.0529636778,0.0042032301,0.066558741,0.0606756583,0.1433402598,-0.2855795324,-0.5523830652,0.4413995743,-0.0899475738,-0.2269182652,0.057241369,0.009802782,0.2371595502,0.4438067675,-0.462461412,-0.2290543914,-0.3841306865,-0.1036187857,-0.0784264207,0.111617744,0.2703568339,0.1271469742,-0.0600731634,0.0835442543,-0.4889653027,-0.0108485837,0.2261229157,0.3510299027,-0.0546096601,0.2301254123,-0.3080968261,0.5490029454,-0.0987386033,0.0003188615,0.1891236454,0.2644511759,0.3137302697,0.0006335465,-0.1188187823,-0.011117544,-0.1578209549,0.0573498979,0.0866157115,-0.0329533704,-0.2033592612,-0.0622321852,-0.0875488818,-0.3918566406,-0.1945627928,0.0748152584,-0.139214918,0.4881857634,0.1346015781,-0.0214913916,-0.2688154578,-0.5046188235,-0.188937813,-0.0647781193,0.0630994365,-0.1009809002,-0.2334393412,-0.0225733425,-0.1421685815,0.3643379509,-0.03538201,0.4224090576,0.0193761811,-0.0043754415,0.0461301133,-0.246693179,0.1373498887,-0.1803703457,0.1608525068,0.0215204004,-0.2174907029,-0.1500654519,-0.1809921265,-0.3337965906,0.2890722156,-0.064665772,0.3597169816,-0.4139510989,0.1336410493,0.1360066235,0.4626384377,-0.2675513327,-0.2452470064,-0.3585158587,-0.2845741212,-0.5091453791,0.0698035061,-0.0466539226,0.2883152962,-0.1866549104,0.0993350893,0.1112002134,-0.1078603193,0.1417367607,-0.1796577126,0.4755511582,-0.0155174471,0.1578666866,0.2025622278,-0.3639868796,0.4272719324,0.6419882774,-0.0033067765,-0.078378655,0.2667135298,-0.0911349282,-0.3595597148,0.2407397777,0.1143725663,-0.0302240625,0.4233919382,0.172584489,-0.0554398559,-0.260854274,0.1437499523,0.1891074479,-0.2571281195,-0.3033946753,0.0415695459,-0.1116676107,-0.0516260825,0.4951725304,0.2201733589,-0.1880419105,0.1546063423,0.4347303212,1.0237908363,-0.0492086485,0.2442313135,-0.0516957939,-0.1728122383,0.3330746591,-0.1814734042,0.2314369082,-0.493619591,-0.0283009224,-0.002531776,-0.3672087193,0.0202415846,-0.0002853708,0.0402520336,0.370502919,-0.0152052566,0.0481119342,0.1073374003,0.1459222585,-0.3727532625,0.0209448803,0.1946272254,0.1011239514,-0.2460949421,0.2926782072,-0.2635617256,0.129431054,0.2280171514,-0.2702026367,-0.4946398437,-0.0237862915,-0.1600282192,0.4113853872,0.2236571759,-0.0913551375,-0.1175961569,-0.022400897,0.1198652908,0.138779074,-0.2773254216,0.0700237155,0.2710630298,-0.1021819934,-0.0784414336,-0.0544876866,0.2245148569,-0.096388936,-0.2602992654,-0.0037649816,0.0314115472,-0.5006491542,-0.3535880446,0.2000948489,-0.2833966315,-0.3202573359,-0.2009006143,-0.255900383,-0.3804832101,-0.1931124032,0.1067073941,0.239674747,-0.1619654149,0.3134112358,0.1729707271,-0.2827112973,-0.2536541224,0.53038764,0.0574306771,-0.1936882734,0.5586514473,0.0570110828,-0.1526134461,-0.2115243524,0.0699741468,0.0815589726,-0.4405606687,0.2268458158,-0.2302013338,0.199733302,-0.1039063632,0.3816934228,0.3563076556,0.3587630391,0.1471941918,-0.6384472847,-0.1401275843,0.0637325495,0.0598139614,0.3195614517,-0.2418227494,-0.3160450161,0.0211862344,-0.2031254917,-0.2665422559,0.1864969283,-0.4010072947,0.0106705343,0.2122327238,-0.2468951195,0.2790992856,0.0829901621,0.0980512798,0.1335290223,-0.1468684524,-0.1415511072,-0.1775091439,0.1733580381,0.069386147,-0.1976430863,0.0750410855,-0.0982388631,0.0330030844,0.0126408963,0.0078632459,0.1390440017,-0.1186559573,-0.0819458142,0.1011997014,-0.1600312293,-0.080513522,-0.08792492,-0.2624417245,-0.0095140329,0.2319003195,0.2567796111,0.3299060762,0.0053126621,-0.0946292654,0.0166896433,0.1174840182,-0.0694535449,0.1595929861,-0.2571022213,-0.2200364172,0.0047605396,0.3969479799,0.4484581649,-0.3414472044,-0.0623532645,0.2812820375,0.1999392211,-0.1703756452,-0.0777103528,0.024561489,-0.124617666,0.0377965607,0.1070602983,-0.0396772176,-0.2999564707,0.3996058702,0.2185706496,0.2358448654,-0.6298623085,0.0965889469,0.0054112412,-0.1926877201,-0.1233177036,0.3981250226,-0.0484687425,0.0148385875,-0.1310550421,-0.0493785702,0.1746674925,-0.2538429201,0.1621407121,-0.230324477,-0.5396325588,0.437032491,0.2160810381,0.0331490897,0.1858433038,0.1410859823,0.3841734529,-0.0615310185,0.3127914667,-0.5757319927,-0.0219482854,-0.1935696304,-0.018974971,-0.080769226,-0.3122321367,0.0624359921,0.412076205,0.0994741097,-0.1274212301,0.327339977,0.1988075972,-0.2762842178,-0.4806446731,0.1116910651,0.0681001842,-0.0663456023,-0.3008173108,0.1126565486,0.4403692484,-0.0299149416,-0.0137705104,0.2593500018,0.5040763021,0.3554439247,-0.0053285216,0.0508302227,-0.0166127775,0.0016250174,0.0226878598,0.5821024776,0.1404578388,0.085850805,0.2327135503,0.1287442148,-0.2031431943,0.2377802581,-0.0915486664,-0.3499816656,-0.1010764316,0.3797106147,0.0623786859,-0.0163545031,-0.2361407429,0.0732196122,-0.4497527182,-0.3263024986,-0.0062239263,0.1135476232,-0.0069786916,0.0771601871,0.0390185006,-0.1217343137,0.4378733933,0.3708485663,0.2917272449,-0.4959544241,-0.0260206815,-0.6134394407,0.2469732612,-0.2785331607,0.2613338828,0.0232956409,0.2809860706,-0.2379380316,0.1994634271,0.189442113,-0.0120889768,-0.1458255947,-0.1587488651,-0.2953234911,-0.0323910303,0.4535721838,-0.1372143179,-0.0707074329,-0.3728673756,0.0866375118,-0.0651711598,0.0292798486,-0.2280119956,-0.0862385333,-0.0900863856,0.364605397,0.2749117017,0.1641651839,0.7399237156,-0.0142191062,-0.1903964281,-0.3088877797,-0.2489171028,-0.0874358863,0.0422236919,0.1307715476,0.5580548048,-0.0710164085,-0.4577326179,-0.3680815995,0.3573647439,-0.2444417477,-0.4376500547,0.0216133725,-0.0160228219,0.0019513447,0.023479702,-0.1015519723,0.1670088023,0.2402370125,0.1430894285,-0.257378906,-0.2453723401,0.4923164248,-0.2891906798,-0.2053520232,0.0971357673,0.2429452091,0.0788672194,-0.1983155906,-0.470184803,0.456746757,0.2685395479,0.0661137253,-0.1953128427,0.2482628822,-0.2461411208,0.2198327035,0.0062482734,0.1118361354,0.2052615136,-0.2532483041,-0.2365482301,-0.0854921937]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1856","title":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError","comments":"+1 encountering this issue as well","body":"Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```","comment_length":6,"text":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError\nHi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```\n+1 encountering this issue as well","embeddings":[-0.2250885963,0.076731883,-0.064804554,0.1695934683,0.2001733035,0.040594127,0.2751486003,0.187905848,0.2695752084,0.1107619852,0.0015163886,0.1326853484,-0.0287981313,-0.1502168477,-0.2530457973,0.248814255,0.0478339978,0.0672525391,-0.2112784833,0.072815083,-0.318139255,0.2823181748,-0.0256928746,-0.2754576206,-0.0798858255,0.2020083368,0.2110439837,0.0612465665,-0.0773533657,-0.3580642045,0.4978875816,0.1145544648,0.0764976367,0.2203702629,-0.000119347,0.0796893984,0.3349234462,0.0695688799,-0.3520457447,-0.3454497457,-0.4815134704,-0.2950345874,0.0918402895,-0.0753720328,0.0777980536,0.2695018351,0.1183858663,-0.1615502387,0.0342997722,0.117386952,0.2079418302,0.4360889196,0.1372526288,0.1996222734,0.2846588492,0.2759159505,-0.0792347491,0.220105499,0.1539587379,-0.0898141041,-0.2394572496,0.1415972859,-0.0873871967,0.4106123447,0.1029985547,-0.0067237811,0.3030147552,-0.1253305525,0.2309733182,0.3783274293,0.3772535026,-0.1524417698,-0.3207696974,0.0113689331,-0.1718390137,-0.2661002278,0.3038751483,0.3049181998,-0.1360975057,-0.2294645905,-0.3821815848,0.1992192119,-0.0510050543,0.2017719746,0.0429364853,0.2815959454,0.0940981284,0.1019361168,0.1405905634,-0.2489580512,0.4996942282,-0.3159530461,0.0001530419,0.211418584,-0.5378153324,0.0448053628,-0.0471408702,0.4704831541,0.2136866599,0.5300042033,0.1440232843,0.2561103702,0.04553416,0.1990092397,0.176613301,0.1001254544,0.0138424234,0.3264424801,0.1768940389,0.2471064627,0.0556653999,0.1190195829,0.0605339445,-0.2356237024,0.755854845,0.0236200113,0.3285298944,-0.488045454,-0.4174036086,0.2627676427,-0.1400986463,-0.1077218503,0.1001200527,0.1696061492,-0.1635462195,0.2428719252,0.0417018756,-0.0506725647,-0.2140608281,0.1194284856,-0.2032587081,-0.0475262068,-0.0552201681,0.2171226293,0.2866111994,-0.0192884933,0.3025390804,0.0484429337,0.2369210422,-0.1257388145,0.4250385165,-0.1052431539,-0.0686630756,0.3942902982,-0.1767568886,0.0645479187,0.1035768166,0.1936035305,-0.1814714521,0.1804849058,-0.1145071834,-0.1773196757,0.0078976555,0.1471541673,-0.5448837876,-0.1859565079,0.1192204803,-0.4711104333,0.2571556866,-0.4039901793,-0.0672560036,-0.4754796028,-0.0605797507,-0.1578425914,0.0999695435,0.3058144748,-0.1072055101,-0.069692038,-0.2516229451,-0.1098586097,0.4513156712,0.25693506,-0.0770454928,-0.0644333735,-0.2110654265,-0.3225081265,-0.0559925288,-0.1632230133,-0.3385608494,0.08949662,-0.144110322,0.5519915223,0.0089698965,0.3081128597,-0.179934144,0.046058286,0.2681049109,0.1586003006,-0.0253751818,0.1480659246,-0.2331869155,-0.223135829,0.3979045153,0.2493879944,0.108389914,-0.027500499,0.0777470842,-0.095918715,0.217902109,-0.0545549206,-0.066508092,0.0995218232,0.2915270627,-0.0368007757,-0.0139718773,-0.1122395471,-0.5903803706,0.2317750603,-0.1558704078,0.0793760866,-0.1641157269,0.0996027514,-0.1016563475,-0.2222622186,-0.2332544625,0.1004611179,0.0312377997,0.4524867237,0.1561381668,-0.0822483897,0.0693807453,0.1923408508,-0.3761925995,0.0692076087,-0.5538233519,0.3280290663,0.0837696642,0.0933739766,0.0699319094,0.0909372941,0.052533593,-0.0128623052,-0.0901793018,0.3472244143,0.155425325,0.1804797947,-0.0208076034,0.5027162433,-0.0905460119,-0.1482663751,-0.1669492573,0.4497154355,0.1175168902,-0.1578345001,-0.1788064688,0.6174514294,-0.164665848,0.4060372114,-0.0943282098,0.1108536273,0.278391093,-0.0894509628,-0.3801782429,-0.2291381061,0.4588930905,-0.1066001728,-0.1646127254,0.1251154393,-0.2034100294,-0.1545231044,0.1039141119,-0.1623341739,0.0529166237,0.1160744056,0.0970427394,-0.1087477431,0.1216676757,0.6843627095,0.5313081145,0.0968352631,-0.1494176835,-0.0406537317,-0.350255549,-0.0370823108,0.0794153661,0.1601330042,0.0667908788,0.4127055407,0.1499239653,-0.166498363,-0.4900137186,-0.2799476087,0.0477287211,0.2749968469,-0.4560048878,0.0792689845,-0.3234074414,0.1973025501,-0.3749614656,-0.1009683013,-0.2287079394,-0.4537575543,-0.1703415513,0.3959552348,-0.056187328,0.2191454768,-0.5849730968,0.1086793169,0.015038982,-0.4684254527,0.0154071273,-0.1580300927,0.0297400765,-0.0277547445,0.361084938,-0.0261436421,0.4567725956,-0.3252692819,0.1089977697,-0.568166256,-0.2964089215,-0.0768270716,-0.0686087757,-0.0217505656,0.2306248397,0.1409736127,0.0728965998,-0.254646033,0.1855376661,0.0360522158,-0.2252131999,0.2616174817,-0.2118110061,-0.0549847372,0.118847087,-0.1612678021,-0.0456255265,-0.237349093,0.0052702227,0.3044779897,0.1336456686,0.0298982747,-0.0337766819,-0.0053854794,-0.129609555,0.3159192502,-0.2435557097,-0.6007629633,0.4780753553,-0.0015765774,-0.1143220961,-0.0135347443,0.0261702016,0.2810082734,0.4022573233,-0.491664201,-0.053290505,-0.3677075207,-0.0600164868,0.0618661456,0.0176251642,0.2620430887,0.1838421822,-0.0202766787,-0.0052218516,-0.4612856507,0.0493663214,-0.096008122,0.5152311325,-0.1660987586,0.1556836218,-0.2366997451,0.4685258269,0.0155187957,-0.1988294423,-0.0550353378,0.146520257,0.3458606601,-0.1022410989,-0.2498234063,-0.1518725902,-0.1086972207,0.0488136038,-0.0070799724,0.0141991451,-0.2092012465,-0.0348537266,0.1581667662,-0.1648043096,-0.2482783794,0.0270718429,-0.2430202067,0.4352553785,0.1412229538,0.0992230922,-0.2979245484,-0.5274120569,-0.1109163463,0.1231679246,-0.0623908266,-0.1347864866,-0.3814020753,-0.0325925164,-0.0348637067,0.3347353637,0.1213547587,0.5968704224,0.0056365742,-0.2821841538,0.0711991042,-0.1052578688,0.0336885005,-0.1763169765,0.1929178983,0.0828305706,-0.2185994238,-0.3280678391,-0.1699917614,-0.1326868385,0.217424795,-0.0495403819,0.3800929785,-0.2958737314,-0.0002563173,0.1241276413,0.4704477489,-0.2354771346,-0.3043919206,-0.3524944484,-0.0863404721,-0.2161052674,-0.0572923012,-0.2322838604,0.1554466635,-0.1232960597,0.1649760306,-0.0292985998,0.0208521206,0.3132496178,0.0045281705,0.3159149885,-0.0456329957,0.3559559584,0.1698106974,-0.1779014319,0.1723959446,0.646582365,-0.1042154804,-0.3704963028,0.102010712,-0.0875274166,-0.2644770443,0.2236969769,0.1206545159,-0.0575275198,0.4007931054,-0.0765669271,-0.1705048978,-0.1761162132,0.1911170632,0.1068798229,-0.3308697939,-0.2099461406,-0.0075995252,-0.0560103059,-0.1028774753,0.4602232277,0.1001626849,-0.1091961265,0.2810291946,0.4155534506,0.8621155024,-0.0272741877,0.2889660001,0.0371906161,-0.2216851562,0.2454454303,-0.0498783737,0.1756289154,-0.4704478681,-0.0611897074,-0.0498888344,-0.2730338573,0.1443859041,-0.0564442277,0.167547971,0.3379246593,0.0851329416,0.1058019325,0.0894836411,0.0293584783,-0.190225035,-0.1105333939,0.1396246254,0.1610452086,-0.2275767326,0.2526022792,-0.1454082578,0.2003412694,0.1538546532,-0.2497354001,-0.3570319414,0.0904810876,0.11693234,0.444709897,0.4004594088,0.016140433,-0.1963677555,0.0581610724,0.1552620232,0.0741927177,-0.2351881266,0.1508303434,0.4351428151,0.1137916073,-0.0126792351,-0.1484872848,0.2704454362,-0.0006335658,-0.1707801968,-0.0108917383,0.0215519518,-0.5391872525,-0.2096562535,0.0230181143,-0.1192047596,-0.2234150022,-0.125690043,-0.1307750791,-0.0216629449,-0.2801505029,0.1491042227,0.2940140665,-0.1519973427,0.3228521049,0.164577201,-0.3934350908,-0.1657423228,0.5581394434,-0.0524916947,-0.2320254296,0.4815107286,-0.0082210675,-0.2350170463,-0.2183789909,0.2079294771,0.0064746039,-0.5276332498,0.0149990376,-0.2202656418,0.1958942562,-0.0444698147,0.317931205,0.2691926062,0.1195092723,0.0982406959,-0.5594802499,-0.080996722,-0.0230040364,0.1357378215,0.4403987229,-0.2651758492,-0.3168162405,-0.2081654519,-0.1880725324,-0.2810768187,0.0172315035,-0.3004452288,0.0469481908,-0.0496212617,-0.0938209295,0.2330237925,0.0236996971,0.1921802461,-0.010384772,-0.159156099,-0.1588350981,-0.2360826433,0.1447756886,0.0456201248,-0.1164229438,0.0490257964,-0.0882111192,0.0906146169,-0.0829597488,-0.0640107021,0.3415862024,-0.2005649209,0.0741784126,0.0021135835,-0.0681785494,0.0746967718,-0.1422455162,-0.367241174,-0.0821838379,0.2453424484,-0.0085414881,0.0199948698,-0.0081134047,0.1631763428,0.1610419452,0.0430779159,-0.1377586573,0.3099133968,-0.2693898976,-0.1078333482,0.0800843835,0.2453332245,0.4371735752,-0.3119061887,-0.0938737169,0.2375618517,0.2239438146,-0.1928546876,-0.0912000686,0.1285325736,0.0811859965,0.0870433301,0.1048167273,-0.1443973184,-0.3393930495,0.5064846873,0.1266261041,0.1903393716,-0.7166602612,0.136276111,0.2253235281,-0.1923126727,-0.0878706872,0.2379967868,-0.0698391795,-0.1021710187,-0.1045716703,-0.0845684484,0.2801551521,-0.3009580672,0.2002409548,-0.3098140955,-0.40661183,0.3633381128,0.0308528394,0.0029970577,0.1790235639,0.2425090373,0.4073970318,-0.1077391207,0.2966654897,-0.5963937044,-0.0931222737,-0.1258669347,-0.1755936742,-0.307890743,-0.2528649271,0.0189747866,0.2218114287,0.0074996501,0.057498727,0.2740175724,0.1635297835,-0.17374219,-0.5598787069,0.0716221854,-0.0453827269,-0.0432609431,-0.189927578,0.0333668515,0.3887264132,-0.0708334744,0.1376362741,0.0637725294,0.4085103273,0.3661541641,0.0076681599,0.0491684154,-0.2757106721,0.0087122722,0.0253661834,0.7732428908,0.0328074098,-0.0283914898,0.3182451725,0.1481941938,-0.1918527335,0.3897079229,-0.1919863224,-0.1013873369,0.0346332379,0.73007828,0.0769656971,0.2296725661,-0.2099964619,0.2093302011,-0.3319430947,-0.4132502079,0.10272412,0.152440384,0.026416257,0.0462336168,0.0515701771,-0.1446708888,0.4821435213,0.2566469014,0.0236625094,-0.4520733953,-0.0249651279,-0.6759163141,0.2390690595,-0.3022977412,0.1391062587,-0.0113593917,0.2993178368,0.032044407,0.2368614972,0.1235118434,-0.2095144242,-0.162431851,-0.1206244752,-0.1801688373,-0.0190075263,0.3118014038,-0.2157643884,-0.0647248179,-0.282999903,0.0891877636,0.1019544229,0.0812597945,-0.0267289095,-0.1298771799,-0.1214852929,0.2167259008,0.1267948747,0.1198971719,0.7084317207,0.0639197826,-0.1620446742,-0.3878141046,-0.2808282971,-0.0057211169,0.2121156603,-0.0361770205,0.6009325385,-0.0044473619,-0.32307899,-0.3126295805,0.3023803234,-0.1683024466,-0.5676315427,-0.1111982167,0.1247734651,-0.1132828444,0.2396723628,-0.219635725,0.1194171757,0.218316406,0.2779963911,-0.1286899894,-0.2827096581,0.6120929718,-0.2388294935,-0.1360079795,0.024559876,0.1968936324,-0.0818692371,-0.1385934055,-0.5419692993,0.2295688987,0.2558017075,0.006991508,-0.1098815873,0.1264554709,-0.328461051,0.3246332705,-0.0561399944,0.2013751566,0.1050240323,-0.1627770662,-0.2421953827,-0.1300474554]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1856","title":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError","comments":"@lhoestq Hi! I encounter the same error when loading `yelp_review_full`.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset_yp = load_dataset(\"yelp_review_full\")\r\n```\r\n\r\nWhen you say the \"Quota Exceeded from Google drive\". Is this a quota from the dataset owner? or the quota from our (the runner) Google Drive?","body":"Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```","comment_length":45,"text":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError\nHi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```\n@lhoestq Hi! I encounter the same error when loading `yelp_review_full`.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset_yp = load_dataset(\"yelp_review_full\")\r\n```\r\n\r\nWhen you say the \"Quota Exceeded from Google drive\". Is this a quota from the dataset owner? or the quota from our (the runner) Google Drive?","embeddings":[-0.0611734577,0.1423272789,-0.1086319983,0.221383661,0.1263362318,0.1217507571,0.2452765852,0.0103803724,0.3931340277,0.1096663773,-0.1395590156,-0.0201449152,0.0363817997,0.0252378229,-0.1712938547,0.364323169,0.0185451135,-0.0113467071,-0.0788380429,-0.0102700591,-0.30445382,0.2593735754,-0.0600827113,-0.0840556398,-0.0598397739,0.1535959691,0.3166562915,0.0442227907,-0.0833672881,-0.3201696277,0.4210704565,0.1378976405,-0.0156709719,0.3289815187,-0.0001259096,-0.0500239134,0.4167665839,0.0320632271,-0.3987612724,-0.3099810779,-0.4581512809,-0.3251411617,0.1509957314,-0.100701496,0.1499893963,0.1633685529,0.0053270529,-0.2619084418,0.0953765139,0.1001130268,0.117274262,0.3501996994,0.105993107,0.2015099078,0.3452163935,0.1226299033,-0.0187581033,0.2913204134,0.1881213188,-0.1821518689,-0.2705043852,0.0405861363,-0.0015808828,0.524055779,0.2074133009,-0.065812543,0.2831406891,-0.2702280879,0.2464372814,0.4334842265,0.5518841147,-0.1611353755,-0.3142518401,-0.1084907204,-0.1443194747,-0.2082730383,0.2683827281,0.4797148108,-0.1679522544,-0.1982161701,-0.5444876552,0.1608868092,-0.0506577082,0.1889034659,-0.132513836,0.246868059,0.034653727,0.1829144806,0.1495759487,-0.17243433,0.6086639166,-0.3743198216,0.0349230245,0.2907087505,-0.5965977907,-0.00776981,0.0340772346,0.5633322001,0.2247852087,0.5783308148,0.1639446467,0.2058051378,0.0257192608,0.2746677697,0.506039083,0.2200191021,-0.1086446121,0.3596937954,0.1931715757,0.2153612673,0.1479297131,0.1244066581,-0.0030034045,-0.2710251212,0.618827939,-0.0734223798,0.2509362698,-0.5128179193,-0.5945706367,0.3519128561,-0.2508152127,-0.1227589026,0.0529643372,0.2121903747,-0.1759034097,0.1633972973,-0.0488658249,-0.0933008939,-0.1332674772,0.0890614912,-0.2012096941,0.1213541999,-0.0276430454,0.2276746333,0.2279391736,-0.0355741829,0.3027086854,-0.1383718103,0.3369037807,-0.203156963,0.4234705269,-0.1530131102,-0.1576988995,0.4656793773,0.0188060924,0.1361016035,0.077129744,0.1752080619,-0.2254126668,0.317854166,-0.1811051071,-0.191312477,0.1732186079,0.0843120068,-0.6098444462,-0.1870330423,-0.2142927945,-0.4069159031,0.3283523619,-0.3127764165,0.0074014249,-0.3569791019,0.0626097918,-0.14311409,-0.0513427816,0.4348807931,-0.2782552838,-0.1656878293,-0.3917514384,-0.1966852993,0.3818414509,0.4646115899,-0.1238603368,0.0903289914,-0.2280271351,-0.16861929,0.0388513617,-0.1546451896,-0.4479780197,0.0987027809,-0.1993888617,0.4591367841,0.0024011473,0.1677615941,-0.097883068,-0.0189726278,0.311175108,0.1539816707,-0.0105006089,0.0756677464,-0.1189428195,-0.1285549104,0.3238051236,0.263197124,0.1939359158,0.0001208673,0.1493890733,-0.0769864544,0.0875307024,-0.0165592395,-0.0044122222,0.2684107125,0.3694301248,0.12750341,-0.0572713837,-0.036968559,-0.4602802694,0.1719496548,-0.419696331,0.0224219933,-0.0614067167,0.0927605703,-0.1885295808,-0.1335383952,-0.2406832576,0.0501293167,-0.0554182567,0.3447307944,0.1498196125,-0.1016497537,0.075372301,0.0877255276,-0.4383293092,0.030736668,-0.4001354277,0.2532045841,0.0287610237,0.1389985085,0.0798331574,0.0776893198,0.1592817754,-0.0172329172,-0.020363912,0.2922261655,0.1412290931,0.4036790729,-0.0789693147,0.4830584824,-0.0287256408,-0.1002454236,-0.0511055626,0.4246054292,0.1246158928,-0.2297820151,-0.2653988898,0.6342540979,-0.1246944517,0.3354456127,-0.0368466116,0.1040458903,0.2729403675,0.0248604082,-0.2758755684,-0.1860741675,0.4675994813,0.1056900993,0.0031508841,0.1219839305,0.0486664921,-0.1620047539,0.3148164451,-0.3001989126,0.0459326021,0.0504410788,-0.0494226478,-0.2442610413,0.1841947436,0.6842710376,0.5695011616,0.0790135413,-0.0502621643,-0.1501575261,-0.170257628,-0.1112228259,0.028954763,0.1859032512,-0.0549864322,0.4992842674,0.1912119985,-0.2243994772,-0.4243579805,-0.1603443772,0.1699371189,0.3191200793,-0.4955602586,-0.0845628306,-0.1221753657,0.2251867056,-0.284299165,-0.0633714572,-0.1669766754,-0.5736444592,-0.1112596244,0.4169124067,-0.0569943935,0.181970194,-0.4180186987,0.1734418124,-0.0535651036,-0.4603134096,0.0476913266,-0.2303589582,-0.0511448421,-0.0123480652,0.3434800506,-0.0785725713,0.3597759306,-0.2547629476,0.1363714337,-0.6831150055,-0.2296449393,0.0053691883,-0.1194750294,0.1411395818,0.041313339,0.1324565709,-0.0009103297,-0.0025763779,0.0279338472,0.0969524235,-0.2848416269,0.1665333658,-0.1978720427,0.0331816413,0.1929992884,0.0042981002,-0.1131908,-0.3015261889,0.1296286583,0.4311295152,0.0800789297,-0.0818815231,-0.1446488351,0.1002122015,0.1218437776,0.307833463,-0.3872552812,-0.4542714953,0.5385531783,-0.0599849783,-0.1861253232,-0.0017487928,0.0950822532,0.2571623027,0.4127817452,-0.4884949028,-0.1079789326,-0.3561542928,-0.1380737573,0.0689078793,-0.0369922258,0.2624653876,0.0366223045,-0.0105526224,0.0430209152,-0.4865606129,0.0323370732,0.0484947935,0.3351130784,-0.0962446406,0.1513513476,-0.2685446739,0.5760503411,0.071352534,0.0197706576,0.093815133,0.271107465,0.3968070745,0.0199433714,-0.0723116323,-0.0836176351,-0.0744836256,0.037522953,0.120005168,0.0165874641,-0.2322214693,0.0097657777,-0.0028807507,-0.2798798382,-0.1282061785,-0.0318137184,-0.1907246709,0.3495222628,0.1516786069,0.0363003947,-0.2485708296,-0.5477306843,-0.201163426,0.1152138188,-0.0155694038,-0.1009352058,-0.3231308162,-0.0268980265,-0.0342555381,0.3341203332,-0.0238246247,0.416454196,-0.0473058745,-0.2034258395,0.2210251987,-0.1795638204,0.1422536373,-0.0516005866,0.1998466551,-0.0585345216,-0.1162847206,-0.2715664804,-0.206076771,-0.2619056106,0.2453435808,-0.0776946247,0.2267118245,-0.402666837,0.0427866541,0.0687754899,0.3809310198,-0.2973003685,-0.2171104997,-0.3158603013,-0.2742903233,-0.2834303081,-0.0560931265,-0.1054622531,0.2077483833,-0.1203184873,0.1751268506,-0.0006153196,0.0093421303,0.3233749568,-0.0668394342,0.3630559146,-0.1080761403,0.1413898319,0.0862660334,-0.3430889547,0.3026038408,0.7589162588,-0.0703877434,-0.229164362,0.2544202507,-0.0593636967,-0.3124517202,0.2692412436,0.1170491427,-0.0899834409,0.3634094,-0.0209082663,-0.0537425429,-0.2774036825,0.146838367,0.1181071773,-0.3462929726,-0.4122238457,0.1307946891,0.0961119384,-0.1092520878,0.4117068648,0.2758307159,-0.1401700377,0.0539927669,0.4594937563,0.9954117537,-0.1312534213,0.2847009599,-0.0603051521,-0.087182045,0.3122884333,-0.248934865,0.10440211,-0.3933152258,0.1471042782,-0.126330629,-0.332987994,0.0744592398,0.0401632003,0.2220612317,0.4867366254,0.0150607275,0.1255379766,0.1986963004,0.093246147,-0.3384930491,-0.0146011747,0.1880654693,0.0834587142,-0.2254415005,0.3900942206,-0.132897675,0.2234778553,0.0194447823,-0.1187901571,-0.5311700106,-0.036327228,-0.0562916286,0.4472949505,0.2374154478,-0.2032884955,-0.1788794547,0.1081016883,0.0135929557,0.0053537376,-0.2681933939,0.154819414,0.3955296874,0.0751511082,-0.0286445357,-0.0019437727,0.2729851007,-0.0566019118,-0.1518229842,0.0878179818,-0.010166808,-0.5436390638,-0.3430452645,0.2885993421,-0.233512342,-0.3450496197,-0.2680541575,-0.2246433645,-0.1828403026,-0.2051273137,0.0665159598,0.2343987823,-0.1421090961,0.2173697501,0.1332423687,-0.3970557153,-0.2534718513,0.5514842868,0.1613256782,-0.1161005124,0.535987854,-0.1115483865,-0.2558026612,-0.1674052179,0.0322989151,0.0732364953,-0.5194632411,0.0728615373,-0.2384830564,0.1712496281,-0.1284858137,0.3552261889,0.3510367572,0.3508452177,-0.0265103038,-0.561318934,-0.0033330654,-0.0142845325,0.0619172715,0.322763592,-0.2390571684,-0.2200428993,-0.0399372503,-0.1807187945,-0.2042579651,0.1239503175,-0.4284902811,-0.0301426742,0.1998591125,-0.0875157118,0.16646716,0.0576776601,0.0722068101,0.0649721026,-0.2155700177,-0.1194583178,-0.138119176,0.2042919844,0.1090286896,-0.1423733681,-0.0239586867,-0.033593677,0.0521026589,-0.0478164926,0.0341451913,0.3134577274,-0.1016386896,-0.0161755849,0.155046612,-0.1265354455,-0.1068242267,-0.0469964482,-0.2977286875,-0.0476188064,0.1859017462,0.0903444961,0.3465736508,0.0605114251,-0.0392325446,0.1167162806,0.180992946,-0.1556670368,0.1471493542,-0.2417294979,-0.0341071449,-0.0012865985,0.3312526345,0.578892231,-0.3147875369,-0.180444181,0.2346495837,0.1194177344,-0.1353293508,-0.0913541391,0.0758088678,-0.0935797617,0.0894161984,0.1361348331,-0.0180283841,-0.2474400103,0.4463033676,0.1169827729,0.3723030388,-0.6105338931,0.171087265,0.1364411563,-0.1175372154,0.0375966243,0.3581705987,-0.0394604243,-0.0296013821,-0.1714909375,-0.1217372417,0.1995048225,-0.3597776294,0.1593436301,-0.2145534456,-0.6058475375,0.4413034618,0.3200616539,-0.072470963,0.2271359861,0.1714421958,0.3912146091,-0.139546603,0.4165734351,-0.5694961548,-0.0349033773,-0.1443037093,-0.1760994494,-0.1324055046,-0.3570284247,0.0547092333,0.3048049212,0.0694902465,-0.1116636097,0.1567234844,0.0529072545,-0.2609964013,-0.4940488636,0.1442116052,-0.098430343,-0.0799138099,-0.2667756677,0.0581328347,0.3363921344,-0.0411696993,0.117006585,0.187348485,0.4015208781,0.3223712444,0.0609501526,-0.0089962576,-0.1105556861,-0.0168924499,0.0459278077,0.7341306806,0.011859118,-0.1686287224,0.2282496691,0.1050968841,-0.162401557,0.3463482261,-0.0672928467,-0.1969519854,-0.2495878637,0.5849602222,-0.0744033456,0.0405418426,-0.2656804025,0.1646089256,-0.3968412876,-0.4330810308,-0.0256162006,0.001106879,0.0062215556,-0.0262205657,0.0161198583,-0.1713746041,0.4875509143,0.4068420231,0.0580718853,-0.4211580157,0.0195165295,-0.6684360504,0.3273186386,-0.3808624446,0.1056365222,-0.0429514199,0.3075509369,-0.1990820765,0.3060510159,0.1197364479,0.0222013518,-0.3295070231,-0.2277282774,-0.1944898814,-0.1070672274,0.3338200748,-0.2107223868,-0.1293594837,-0.3248287737,0.1042675227,0.0008915893,0.0082118874,-0.1230815798,-0.0285041872,-0.0221031383,0.3299268484,0.1967958063,0.1362971067,0.7609401941,0.0671186298,-0.1331506968,-0.2573164403,-0.2398208231,-0.1036633775,0.1768930554,-0.0367276445,0.4463853538,-0.0211850703,-0.4058427513,-0.1812153161,0.4622705579,-0.1463851184,-0.4938641191,0.0005719427,-0.0510119349,-0.1060828865,0.0742335618,-0.1120769903,0.093068786,0.2723671496,0.2488854825,-0.1138962582,-0.2186215073,0.5264893174,-0.3776214123,-0.0317405425,0.1192818359,0.1963829845,0.0453493297,-0.2616777122,-0.4460185468,0.3479745388,0.3618851602,0.061042957,-0.1220168471,0.2319945097,-0.4572667778,0.1608504653,-0.0770439282,0.0566178262,0.1084925681,-0.2246620804,-0.1531855762,-0.1641315818]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1856","title":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError","comments":"+1 Also encountering this issue","body":"Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```","comment_length":5,"text":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError\nHi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```\n+1 Also encountering this issue","embeddings":[-0.2178745568,0.0554040819,-0.064928472,0.1871751398,0.1897960603,0.0605423711,0.2724092901,0.1821097434,0.2669914365,0.0945816562,-0.0005953679,0.1338512003,-0.03037332,-0.1514435261,-0.2481457591,0.2435114533,0.0482975841,0.0691834316,-0.2024333924,0.0750953704,-0.3096033931,0.2842549682,-0.0377480499,-0.2623851299,-0.082422182,0.1999594867,0.215073064,0.0540608019,-0.067720212,-0.3531710207,0.4997682869,0.1102389991,0.0904275104,0.2218971252,-0.0001206106,0.0773112401,0.3365374506,0.0592141598,-0.3413579762,-0.3580171168,-0.5039172173,-0.296883285,0.0930138454,-0.0769762248,0.0825937837,0.2733370364,0.1172112897,-0.1631205231,0.0393582657,0.1118925065,0.2008530349,0.4259656668,0.12885423,0.2188681662,0.2776581645,0.2788429856,-0.0738042146,0.2238055468,0.1559256464,-0.0835598484,-0.2277692854,0.1441743821,-0.0969619006,0.4131450653,0.1029466093,-0.0081603834,0.3177075386,-0.134849757,0.2336220592,0.3813249171,0.395409286,-0.1605687588,-0.3164846003,0.017511962,-0.1672884226,-0.266122818,0.3026234806,0.3161084652,-0.1344576925,-0.2379841059,-0.3881217241,0.1841726154,-0.0627293512,0.1972432286,0.0598149896,0.2679462731,0.090262495,0.1066287085,0.1268402934,-0.2452400178,0.5148397088,-0.3396483958,0.0037300398,0.2236133963,-0.5432439446,0.0434418507,-0.0740641654,0.4915336668,0.2200702429,0.5344147086,0.140569523,0.2440538555,0.045836281,0.2060581744,0.1794531643,0.1077670977,0.0173543505,0.3295859694,0.1763882488,0.2370858341,0.0405865535,0.1123761013,0.05432529,-0.227460742,0.7772920132,0.0244350247,0.3349606991,-0.5027849078,-0.4132243097,0.2557024062,-0.1501734704,-0.1224995255,0.1041047424,0.1704564393,-0.1738851815,0.2396553606,0.0381027907,-0.0528810434,-0.2120942771,0.1041609347,-0.2031277567,-0.0519851781,-0.053574767,0.217077747,0.2755151689,-0.0297782421,0.3001874089,0.0471727401,0.238130331,-0.1332133114,0.4429187179,-0.1067754403,-0.0749877542,0.3930728137,-0.1757900715,0.0655454099,0.1107219756,0.1983457357,-0.1881795526,0.2063942254,-0.1355583817,-0.1886826903,0.013167399,0.1361986846,-0.5564047694,-0.1839848459,0.1152159423,-0.4772114158,0.2781608999,-0.4164579213,-0.0590034872,-0.4747435451,-0.0600926951,-0.1583661735,0.0895674378,0.3023730516,-0.1104579195,-0.0815355331,-0.2466866225,-0.1230524704,0.4650461078,0.2589406073,-0.0639864877,-0.0601934716,-0.2174759358,-0.3271944225,-0.0404383056,-0.1540954411,-0.3420242965,0.0912783742,-0.1428965777,0.5588669777,0.0145261269,0.321306318,-0.1898764372,0.0501968935,0.2622364163,0.1476598531,-0.0228625368,0.1470791101,-0.2252560556,-0.2295219451,0.4173057079,0.2482072562,0.1088691279,-0.018791832,0.0693831667,-0.1045441031,0.2170234323,-0.0526572727,-0.070621267,0.0989063382,0.2976103127,-0.0336249769,-0.0019603814,-0.1133741587,-0.5976336598,0.222367242,-0.1609450877,0.0676306114,-0.1494893134,0.1032281891,-0.104489103,-0.2297403812,-0.2406028658,0.1175107956,0.0198070351,0.4462444186,0.1596745998,-0.0857806429,0.0703165308,0.205697462,-0.3823071122,0.0671319142,-0.5609131455,0.337711066,0.08678177,0.1049913839,0.0610157885,0.1016894281,0.0541471578,-0.0113872644,-0.0983933434,0.336096406,0.1517018825,0.1691040993,-0.0288114101,0.5085158348,-0.0977093205,-0.1557233483,-0.1655844003,0.4433358312,0.1189268827,-0.1608149111,-0.1664123684,0.6206329465,-0.1738571972,0.4118745625,-0.1046169326,0.1159796566,0.2689355314,-0.0908705294,-0.38115412,-0.2212152034,0.4814460278,-0.1125113368,-0.1719045192,0.127786383,-0.1985134631,-0.1686115861,0.0878599957,-0.1587539166,0.0595486574,0.1169676334,0.0886384621,-0.1073943228,0.1214296594,0.6769244671,0.5398643613,0.0935227796,-0.1603294015,-0.0387552753,-0.3330653906,-0.0220770314,0.0724555477,0.1599492431,0.0686489195,0.418274194,0.1476784647,-0.1687997282,-0.4865092039,-0.2735460401,0.0481290296,0.2818510234,-0.4619216621,0.0803784952,-0.3105114698,0.1911713034,-0.3737375736,-0.1031896472,-0.2391005009,-0.4456830919,-0.1752128154,0.3929042518,-0.0344686173,0.2194630653,-0.5884147286,0.1130117923,0.0028451928,-0.4747439623,-0.0002839612,-0.1679385155,0.0361029841,-0.0350508802,0.3594091833,-0.0211303812,0.4505541027,-0.3260692358,0.1110915244,-0.5618536472,-0.2934463024,-0.0808973089,-0.0591672026,-0.0283722747,0.2249014825,0.1403338015,0.0621079132,-0.2500926852,0.1799130738,0.0377177782,-0.2338918895,0.2585438192,-0.211644277,-0.0459613763,0.1336752772,-0.1478007138,-0.058554396,-0.2226848751,0.013533798,0.3099861741,0.1398198009,0.0229718294,-0.0583661459,-0.0164197348,-0.1327498704,0.3191753626,-0.248862192,-0.6189054847,0.4743742645,0.0019400835,-0.0994028002,-0.0092319222,0.0340995379,0.284686625,0.3977892101,-0.4953569472,-0.0536244698,-0.3588047028,-0.0627469197,0.0522361062,0.0204044804,0.263068229,0.1859682053,-0.0038162894,-0.0021125514,-0.4584467113,0.0525366217,-0.0858374834,0.5262308121,-0.1672768593,0.1593538523,-0.2505889833,0.4776856601,0.0279360972,-0.1937171668,-0.0523909256,0.1461376995,0.3373754025,-0.0928896815,-0.2515045404,-0.1518340707,-0.0996649265,0.0488732643,-0.016009653,0.0143158622,-0.205829218,-0.0328440145,0.1405609548,-0.1758420765,-0.2387253344,0.01270073,-0.2501254678,0.440931648,0.1430865973,0.0733075514,-0.308953613,-0.5363458395,-0.1071169823,0.1336047351,-0.0502079353,-0.1342522055,-0.3801734746,-0.0427635983,-0.0235148389,0.3379953206,0.1229480952,0.5948275328,0.0067244829,-0.269942224,0.0806010142,-0.0939595252,0.0454640687,-0.1798044145,0.1958059818,0.0754683167,-0.2214194834,-0.3173527718,-0.1706830263,-0.1200192794,0.2347998321,-0.0489016399,0.37717852,-0.2862829864,0.0061668437,0.1301852465,0.4764517844,-0.2486736923,-0.3306413293,-0.3622151911,-0.0722150803,-0.2113661021,-0.0639038682,-0.2452948689,0.1547139883,-0.1111119315,0.1726577282,-0.0394472592,0.0242073927,0.3233990669,0.0033016147,0.3177635074,-0.0530276559,0.3567244411,0.1725349575,-0.171056971,0.1675015986,0.6376372576,-0.1025596038,-0.3744504452,0.1006586179,-0.0843937695,-0.243328318,0.1961911023,0.1209743395,-0.0490209982,0.4042466283,-0.084277235,-0.1705224961,-0.1925237775,0.1835253537,0.1257869303,-0.3364682496,-0.2201020718,0.0002932706,-0.0461723618,-0.1013564393,0.4643617868,0.0918563455,-0.1034092531,0.2647438347,0.4115453064,0.879853487,-0.0362383164,0.2789280415,0.0217466243,-0.2213857025,0.2513014674,-0.0535693914,0.172350958,-0.4822317958,-0.066031605,-0.0580274016,-0.2925001085,0.1545434445,-0.0535324886,0.1656517833,0.3480384648,0.0800513998,0.1019719616,0.0937176347,0.0190364756,-0.1762449741,-0.0979330987,0.1452869326,0.1444221437,-0.2360604107,0.2668321431,-0.148008734,0.2115222961,0.1564919949,-0.241954565,-0.3526753485,0.0864393786,0.1136166155,0.4504359365,0.406236738,0.0073404475,-0.1775670648,0.0793955401,0.1626346856,0.0686415806,-0.2348012924,0.1452098936,0.4441324174,0.1097422317,-0.0177524537,-0.138197124,0.260206461,-0.002250154,-0.1736400723,0.0018198377,0.025153514,-0.5385318398,-0.2084507942,0.0208293647,-0.1153780594,-0.2228443027,-0.1367407888,-0.1245195568,-0.007335511,-0.2672893405,0.1394557208,0.3018868268,-0.1626272351,0.3270903528,0.1709275991,-0.3994530141,-0.160016194,0.56162256,-0.0381297097,-0.2386693209,0.4898665547,-0.0231455136,-0.2375335246,-0.2122086138,0.2074767798,0.0022804749,-0.5358330011,0.0175639577,-0.2180843204,0.2091597915,-0.0571239479,0.3211675584,0.2754170001,0.1331746131,0.1021709815,-0.559304893,-0.0904083177,-0.0300303996,0.1342077255,0.4324396253,-0.2666622698,-0.3314816356,-0.2024728954,-0.1854554713,-0.2689685225,0.0123255067,-0.2999761403,0.0428119935,-0.0476857275,-0.0822202787,0.2238813043,0.0249198712,0.1818886548,-0.0190197509,-0.1569295973,-0.150951907,-0.232030645,0.1529759169,0.049497243,-0.1152828112,0.0469012037,-0.0839086547,0.0881139338,-0.0728897527,-0.0743782595,0.3436106145,-0.1954153329,0.0835900903,0.0015198318,-0.0838368088,0.0518295765,-0.1243935451,-0.363204211,-0.0880024657,0.2480521202,-0.0139972204,0.0337432474,-0.0075060953,0.1590684354,0.1519680917,0.0534285828,-0.1480500549,0.3107746243,-0.275675267,-0.0973936096,0.083781302,0.232463941,0.4287669957,-0.3155188859,-0.0872615352,0.237482205,0.2143419385,-0.1977393478,-0.0876059681,0.1407948881,0.0790852606,0.0720903575,0.1051666215,-0.1471204758,-0.3289375305,0.5094257593,0.1189016476,0.1898252964,-0.7150640488,0.1457044184,0.2271502018,-0.1831375957,-0.084131211,0.2408305705,-0.0503774919,-0.1048385575,-0.122482352,-0.0798509344,0.2789274156,-0.3003844917,0.2017603815,-0.3060791194,-0.4116814733,0.3636362255,0.0416848287,0.0089655584,0.1754401177,0.2418015301,0.3942921162,-0.0933181718,0.2988741696,-0.5901844501,-0.097027652,-0.1262466162,-0.1750353128,-0.300909996,-0.2519982457,0.0262667201,0.2150810212,0.0018890786,0.0568493754,0.2738512754,0.1662657559,-0.1606713533,-0.5553162098,0.0727898404,-0.0152588608,-0.050921943,-0.1951958239,0.0224617459,0.375982672,-0.0708821267,0.1427759975,0.0711200535,0.4011163414,0.3711749613,0.0136834569,0.0529907942,-0.269854039,0.008999113,0.0487838499,0.7850109339,0.0406080931,-0.0257806461,0.3020814955,0.1359072626,-0.1802984327,0.3963080049,-0.1902419031,-0.1048244163,0.034128733,0.7346903682,0.0708432868,0.2199714482,-0.2086880505,0.2120710015,-0.324649781,-0.4037729502,0.1002421379,0.1546900868,0.0163029525,0.0604104288,0.0446966588,-0.1609225273,0.4730005562,0.2568392754,0.0345162712,-0.4656062126,-0.0315698311,-0.6796703935,0.2415541261,-0.2944031954,0.1424510926,0.0005370466,0.2860098183,0.0236352812,0.233856827,0.1276287138,-0.1993961185,-0.1672694087,-0.1205180213,-0.1788385659,-0.0248440485,0.3228877187,-0.2120677233,-0.065877296,-0.2774015069,0.0846820846,0.1126926467,0.0656573847,-0.0291861333,-0.1179038286,-0.1330115497,0.2247759402,0.1174020767,0.1151411012,0.7283633351,0.0701670945,-0.141096577,-0.3645971715,-0.2860898077,-0.0098950863,0.2241288126,-0.0376554318,0.583379209,-0.0091647776,-0.3205572963,-0.3167699873,0.294560194,-0.1722947657,-0.5752869248,-0.1107217744,0.1148878708,-0.1128908321,0.2570157647,-0.214720279,0.1015291438,0.2277433723,0.2801733911,-0.1267982721,-0.2605792284,0.6177116632,-0.2484771907,-0.1319354922,0.0186465438,0.201417163,-0.0842356458,-0.1458451152,-0.546684742,0.2290164977,0.2817784846,0.0068823146,-0.0929112583,0.1316911429,-0.3325563371,0.3384543061,-0.059525311,0.20012936,0.0999475271,-0.1574689746,-0.2472424656,-0.1409941912]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1856","title":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError","comments":"> When you say the \"Quota Exceeded from Google drive\". Is this a quota from the dataset owner? or the quota from our (the runner) Google Drive?\r\n\r\nEach file on Google Drive can be downloaded only a certain amount of times per day because of a quota. The quota is reset every day. So if too many people download the dataset the same day, then the quota is likely to exceed.\r\nThat's a really bad limitations of Google Drive and we should definitely find another host for these dataset than Google Drive.\r\nFor now I would suggest to wait and try again later..\r\n\r\nSo far the issue happened with CNN DailyMail, Amazon Polarity and Yelp Reviews. \r\nAre you experiencing the issue with other datasets ? @calebchiam @dtch1997 ","body":"Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```","comment_length":127,"text":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError\nHi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```\n> When you say the \"Quota Exceeded from Google drive\". Is this a quota from the dataset owner? or the quota from our (the runner) Google Drive?\r\n\r\nEach file on Google Drive can be downloaded only a certain amount of times per day because of a quota. The quota is reset every day. So if too many people download the dataset the same day, then the quota is likely to exceed.\r\nThat's a really bad limitations of Google Drive and we should definitely find another host for these dataset than Google Drive.\r\nFor now I would suggest to wait and try again later..\r\n\r\nSo far the issue happened with CNN DailyMail, Amazon Polarity and Yelp Reviews. \r\nAre you experiencing the issue with other datasets ? @calebchiam @dtch1997 ","embeddings":[-0.1349428147,0.0474774651,-0.1110621244,0.2588101327,0.0624821633,-0.0367827639,0.2863380015,0.0119939046,0.3435023725,0.2581289709,0.0767338723,-0.0375789516,0.0030820963,0.0379523821,-0.2337089777,0.1695692539,-0.0217135455,-0.0736348629,-0.1117966995,0.0406751484,-0.2747719884,0.2887174785,0.0070180912,-0.2034480572,-0.0396618061,0.008733741,0.3278360665,0.1175962165,-0.1759135723,-0.3116758466,0.330021739,0.0826124772,0.0683188289,0.3464248478,-0.0001204181,-0.0683180317,0.3942154646,0.0965760574,-0.41895926,-0.2462104112,-0.3514712751,-0.3533970714,0.0768758953,-0.1483185291,0.1762081534,0.3158309758,0.131492883,-0.2369056493,0.17296049,0.1381318271,0.1324190795,0.4214162529,0.1527190655,0.1810043305,0.3624846339,0.0484379791,-0.0363256857,0.2699772418,0.2151565552,-0.062193431,-0.2929163873,0.1053554639,-0.0394045301,0.5324478149,0.1441996247,-0.0262994152,0.1012791097,-0.3418424726,0.1716118306,0.3872762322,0.5805895329,-0.1249104738,-0.2338673174,-0.093341291,-0.2541105747,-0.3007954955,0.3701971471,0.4087914228,-0.1825312823,-0.2420866638,-0.4668247998,0.2356185466,-0.1145986989,0.1640466601,-0.1773631275,0.2125671208,0.0296117086,0.2180982977,0.2534275055,-0.0722891688,0.5507889986,-0.3385269046,0.0278644338,0.2165496349,-0.5382955074,-0.059662506,0.0265298504,0.475604713,0.1846510619,0.5643036366,0.1895441562,0.1655359119,0.0494824313,0.2911800444,0.4096359313,0.1855698973,-0.2252947837,0.2780072093,0.2126202881,0.1978701949,0.1698426306,0.098749958,-0.0183397476,-0.3449646533,0.6129296422,-0.0725433379,0.1491273493,-0.4413726032,-0.568980217,0.3616555035,-0.2959178388,-0.1476629972,0.0714619532,0.1275595427,-0.1886182874,-0.0284606107,0.0131302457,-0.1142851785,-0.1554235369,0.0914227888,-0.1961929947,0.1193031967,-0.0973106027,0.2034059614,0.208208099,-0.0458631888,0.2794568539,-0.0408518538,0.2950496078,-0.2197562903,0.5377064347,-0.2721767128,-0.1807202101,0.554646492,-0.0523272678,0.0112610497,0.1048941389,0.1803134084,-0.1602941304,0.2673656642,-0.1479073167,-0.1279229075,0.1871571541,0.1310720742,-0.5833469629,-0.2072044313,-0.1647089273,-0.3109378219,0.2320268154,-0.2921413183,0.0238497984,-0.3568731248,0.0165804457,-0.212259993,0.0246520508,0.4080348909,-0.2196417153,-0.0292678587,-0.3386438191,-0.1784811467,0.4241977334,0.3356624246,-0.1659255922,0.1827775091,-0.2545887232,-0.1609527916,0.0107212383,-0.1088565663,-0.3976554871,0.0501730777,-0.2954813838,0.405782342,0.1434173137,0.1928779036,0.0143453786,0.0030359826,0.2835543156,0.1582210809,-0.0147035662,0.0361357853,-0.1703040898,-0.1570073813,0.1708415151,0.2926814556,0.2981024683,0.0041951658,0.272792846,-0.1237247661,0.0197496768,-0.0045746616,-0.0226048306,0.2537461519,0.1342649609,-0.0235799309,-0.0728575885,-0.0011671089,-0.5085259676,0.2030844092,-0.3583157957,0.0499369241,-0.0378025062,0.1098610908,-0.1467158049,-0.2290976197,-0.142865479,0.093814142,0.0016223545,0.3955658078,0.1550281346,-0.1600453705,0.1011650041,0.1278325021,-0.4371520281,0.0473451577,-0.489506036,0.2992715538,0.0183333922,0.0552750826,0.1059115008,-0.0228187405,0.1357090175,-0.1147482097,0.0340183452,0.3085660934,0.1717772037,0.3997481763,0.0409694351,0.6185170412,-0.1008119956,-0.0422909521,-0.0050376407,0.5124958158,0.0809207335,-0.2270922661,-0.3972212672,0.6124414802,-0.2382662594,0.2321486175,-0.1331163496,0.0842777565,0.2649222612,-0.0140435025,-0.2797190249,0.0376058295,0.4447349608,0.214877367,-0.1085133255,0.0594823286,-0.1257260293,-0.0875002667,0.3686031699,-0.2639417052,-0.0099635394,0.1531330049,-0.0149100414,-0.1758394688,0.0952373371,0.8160720468,0.5828450322,0.1383719742,0.0651865974,-0.1881671548,-0.147227332,-0.1379794627,0.0860718787,0.1756961793,-0.0989140198,0.4286357164,0.1037989333,-0.2119631618,-0.4885817766,-0.2679806352,0.1950267851,0.4274774492,-0.431117177,0.0080336919,-0.0060762088,0.2277985066,-0.3302200139,0.0214811247,-0.2290285528,-0.4285721481,-0.1362854689,0.5601523519,-0.0212817974,0.0625669435,-0.320874393,0.2279883921,0.0183125902,-0.4872389436,-0.0421905965,-0.0939055383,-0.1066752151,0.0311306342,0.4095044732,-0.1010930389,0.4162184298,-0.3030728996,0.1113957241,-0.568146348,-0.2399628311,-0.0442503393,-0.1286785901,0.1164732501,0.0898447335,0.2863659263,-0.0945336595,0.0060075377,-0.038362477,0.165768519,-0.2480380833,0.1248965636,-0.1673418134,0.0428206138,0.1129246429,-0.1804883927,-0.1032415256,-0.3451417089,0.1504686624,0.4686197042,0.1605377942,-0.050380908,0.0320634879,0.1036319509,0.0961376578,0.2149112672,-0.3637515903,-0.6516868472,0.4263007343,0.0288697742,-0.1900146604,0.0495832078,0.0850529298,0.2140693069,0.4236083627,-0.5703600645,-0.1277110726,-0.3932390213,-0.1056847423,-0.0140199326,0.1253590137,0.3836190999,0.1056822687,-0.0706488788,0.053199891,-0.4380121827,0.0243644733,0.0625939518,0.330132097,-0.0929753855,0.2045581341,-0.1958513111,0.6013499498,0.0388450734,-0.0149959438,0.0843245462,0.2517703772,0.4348466694,-0.000299089,-0.0227025822,0.0428795256,-0.064292945,-0.0343608037,0.1285892874,-0.0404769033,-0.2122768909,0.0525671989,-0.038389951,-0.1745993197,-0.1617653966,0.1090541705,-0.0566072762,0.3451625705,0.0706645846,0.0425659083,-0.2898287475,-0.4241083562,-0.2228246778,0.1104777306,0.0040395521,-0.0751193911,-0.3085340858,0.0598783754,-0.1164573506,0.3225312233,-0.0794917718,0.3666817546,-0.1161500216,-0.0807974264,0.2194276303,-0.1463888288,0.1137590334,-0.1069362089,0.1028719097,-0.0610684641,-0.046993047,-0.1138119623,-0.2126427293,-0.3106776476,0.3325226307,-0.0191534683,0.2525075674,-0.4487139583,0.0448544137,0.0557432845,0.3680000603,-0.2528471947,-0.2449684888,-0.3577002585,-0.3070174456,-0.3425726593,0.1855710298,0.0011230507,0.2138730884,-0.1341525316,0.0814422071,0.0966469944,0.0366723314,0.2058677673,-0.0807240754,0.3047992587,-0.0512738526,0.0288868919,0.2029771358,-0.3895156085,0.2778032422,0.6610252857,-0.0437913127,-0.0609285608,0.2704164684,-0.0926650912,-0.3631850779,0.2875838876,0.0679405257,-0.0884697214,0.4182348847,0.1258313954,-0.0699599832,-0.2813867331,0.2164290845,0.1614836007,-0.2466440052,-0.3831723928,0.1160350665,0.0151441898,-0.1279821396,0.5681807995,0.1725110263,-0.1188843474,0.0885016695,0.4173097014,1.1071872711,-0.1167678982,0.2464447618,-0.003125652,-0.1765982956,0.3316381872,-0.2680203915,0.1921709627,-0.39882195,0.0398376994,-0.1273453385,-0.3764095902,0.1355912983,-0.0464972891,0.0921546817,0.4239909649,-0.0516012274,0.0821549296,0.1611681283,0.0710147172,-0.4011039138,0.0109411683,0.2179982215,0.1014790684,-0.212803036,0.3707577288,-0.2021793872,0.1133664846,0.1107489392,-0.222391367,-0.4860476255,-0.0960520357,-0.0327839181,0.2836486399,0.2205233723,-0.1124428436,-0.2335146517,0.0826688856,0.1213995889,-0.0617823265,-0.1754774898,0.1419451237,0.2989319563,0.0259446967,0.0317302346,-0.0853983462,0.2706813514,-0.1188420504,-0.1628625691,0.0528548434,-0.0360289253,-0.5696084499,-0.2322643846,0.2447643131,-0.1292956471,-0.2774000764,-0.1481331438,-0.1568342894,-0.3212555647,-0.2062811106,0.1114614755,0.3105113804,-0.1522542089,0.388468951,0.0536009967,-0.3761644065,-0.2494042814,0.4735057056,0.0903814062,-0.0472882092,0.4643446207,-0.0493585654,-0.2557545602,-0.2168323547,0.0545881912,0.0519261956,-0.4644057453,0.1567270458,-0.2019971013,0.2047417462,-0.1024396345,0.330983907,0.338080138,0.3494670689,-0.0752472803,-0.6224424839,-0.0548293814,-0.1099106818,0.0315387882,0.3769244552,-0.2181615233,-0.34974581,0.0394920968,-0.2079380602,-0.2735701203,0.0982374027,-0.4265048504,-0.0077453563,0.1820399612,-0.1430696398,0.2588160932,0.0930425823,0.0909409076,0.0455214232,-0.3256878555,-0.1388732493,-0.1334293485,0.192550838,0.0386026986,-0.1288579702,0.0252980478,-0.092192173,0.0449752435,-0.0188251324,-0.0558863655,0.2361635566,-0.1481453925,-0.1354154348,0.1923213601,-0.0158994999,-0.1197670177,-0.1880185604,-0.3041902483,-0.1043626964,0.1823477745,0.1488694549,0.1872992069,-0.0496610366,-0.0694655329,0.1183414161,0.2726778686,-0.1055048183,0.2189337313,-0.1947792172,-0.1653708816,-0.0994819701,0.3385939598,0.5190999508,-0.2795707285,-0.2123018354,0.2391746491,0.1850921661,-0.081366919,-0.1164119467,0.0985699967,0.0218693279,0.0697929412,0.1930089593,-0.0143177016,-0.3089471161,0.3804351389,0.1134406254,0.3322176635,-0.7211831808,0.0382433943,0.094849363,-0.2230162472,-0.048974257,0.4857473969,-0.0627155975,-0.0028130207,-0.1344220191,-0.099513486,0.1665256172,-0.2666638196,0.1025170609,-0.3384125233,-0.5586660504,0.4284240305,0.3406714797,0.0492289774,0.1477314234,0.2172038704,0.2795032859,-0.1171722561,0.3882199526,-0.5804319382,0.0083502624,-0.1292090714,-0.1199181229,-0.215075314,-0.323345989,0.0437495485,0.3426990807,0.0991773084,-0.1191890165,0.2262880206,0.0652913377,-0.2811642587,-0.3712739646,0.0387421511,-0.0329999775,-0.0986481905,-0.2912286222,0.0770832822,0.3421229422,-0.029564172,0.0117454296,0.2095718682,0.4396032691,0.2685383558,0.0999382511,0.0597683527,-0.0781120434,0.0021205533,-0.0750579908,0.6987162232,-0.0252974667,-0.1015169099,0.2111515999,0.1382064521,-0.2075856179,0.3503798246,-0.1719439924,-0.3512225747,-0.1474416405,0.5249900818,0.0583561324,0.0321251266,-0.2822971642,0.1320524365,-0.4383369684,-0.3146593273,-0.00948196,0.058175806,0.0133782085,0.0018400972,0.0307002682,-0.1343363523,0.4461662769,0.3831959665,0.1674164236,-0.4753990173,0.0328942128,-0.6384221911,0.2768553197,-0.3937015235,0.100226067,-0.0358738229,0.3426169455,-0.1548323929,0.2197476476,0.2084041685,-0.0953624099,-0.1761244386,-0.0730258822,-0.1464689225,-0.0286832433,0.3783764541,-0.1787625253,-0.1578896046,-0.308483392,0.2094821334,-0.1234287992,0.0764032081,-0.133964166,-0.0156035442,-0.0620611086,0.2542665899,0.2515361309,0.1244906783,0.6347792745,0.0061344113,-0.0981874242,-0.3610522449,-0.0923288465,-0.2160479575,0.0614580028,0.0253805984,0.6215923429,-0.1001171619,-0.3572173119,-0.2692157328,0.5112687349,-0.1436240375,-0.4518866539,0.0984652936,-0.0592589155,-0.1322195232,0.1841618568,-0.1086142659,0.1654422134,0.3128494918,0.2070147097,-0.2399944514,-0.2264770567,0.5082708597,-0.4022962749,-0.1337482929,0.1396023929,0.1908556968,0.1651875824,-0.1824489534,-0.439394176,0.3990353942,0.2949404418,0.0157096796,-0.1631641537,0.169516921,-0.4015009105,0.1822026074,0.0383053347,0.0392191634,0.1061062068,-0.2178327888,-0.1542906016,-0.1948410273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1856","title":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError","comments":"@lhoestq Gotcha, that is quite problematic...for what it's worth, I've had no issues with the other datasets I tried, such as `yelp_reviews_full` and `amazon_reviews_multi`.","body":"Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```","comment_length":24,"text":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError\nHi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in ()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```\n@lhoestq Gotcha, that is quite problematic...for what it's worth, I've had no issues with the other datasets I tried, such as `yelp_reviews_full` and `amazon_reviews_multi`.","embeddings":[-0.194786638,0.0554474331,-0.0836697891,0.1578223705,0.1719484329,0.0804431736,0.2375732958,0.099151589,0.2416200787,-0.0136137661,-0.0839820951,0.0607005432,-0.0039978959,-0.1410388201,-0.3183495402,0.3028692007,0.0808663368,0.0319708996,-0.1793727428,0.036818359,-0.3233240247,0.2446756661,-0.0489678606,-0.2615421116,-0.1571846455,0.1928241253,0.1952106357,0.0186479744,-0.1012964174,-0.3044863641,0.5374702811,0.199249506,-0.0120766601,0.318572551,-0.0001195672,0.0853057429,0.3367518187,0.0248466954,-0.3156446218,-0.3974169791,-0.3859312832,-0.2944520414,0.0969609991,-0.0750944391,0.1068804562,0.1773181409,0.1115883663,-0.1239965186,0.1156422347,0.1848811209,0.2053506076,0.3852194548,0.1166317239,0.1972644776,0.3125816286,0.3064412773,-0.1069201529,0.2679606676,0.1251927316,-0.1011527255,-0.1667243242,0.0929330066,-0.1170698479,0.3990475833,0.1447162926,-0.0405137092,0.3200861216,-0.1548224092,0.2880885601,0.4979780316,0.38207376,-0.1294228435,-0.3618169129,0.0308115426,-0.2295023799,-0.2095724046,0.2629090548,0.3812124133,-0.0578768328,-0.2178773135,-0.4667551517,0.2266641259,-0.0254905,0.2627213001,0.0525064915,0.3400252759,0.1289118528,0.1233245432,0.1598106176,-0.2095304579,0.4995460212,-0.244139865,-0.0085651474,0.1588152498,-0.5495276451,0.0090561146,-0.0314955376,0.4228042066,0.2175273746,0.526178062,0.1308601648,0.2861514688,0.0186714996,0.2018932998,0.2349801213,0.1130001545,-0.0022747032,0.3624728918,0.1853784174,0.2451706976,0.0670079961,0.1450264454,0.0917718858,-0.1510998011,0.6740022302,-0.1330978721,0.2948665917,-0.4654795825,-0.4914773405,0.3247948289,-0.1839044094,-0.0913163647,0.0314727761,0.2384021729,-0.145765394,0.2540999353,0.0444367416,0.0352183171,-0.1544977129,0.2191961408,-0.1957147866,-0.0111426571,-0.0717064664,0.2342016697,0.2611872256,0.0155249527,0.3998512626,-0.0320103317,0.2115763128,-0.1109667793,0.3717316389,-0.1075839102,-0.0641368777,0.3863778412,-0.0871023536,0.0800858811,0.1076514125,0.120585382,-0.1914372593,0.1178879291,-0.0960544124,-0.1347409189,-0.0384671241,0.1501204371,-0.564566493,-0.1816153079,0.0755826533,-0.3771919906,0.3030200303,-0.3675339818,-0.0443699211,-0.4381237924,-0.0404126644,-0.1279566884,0.0071101333,0.3058854938,-0.113079384,-0.0507401749,-0.3328308761,-0.1052065939,0.4337944388,0.2970632315,-0.0183742121,-0.0539242849,-0.1458172351,-0.2737306952,-0.0362811908,-0.0873362944,-0.2959800065,0.1222153977,-0.1102306172,0.48588413,-0.0137295835,0.3003837466,-0.2076560706,0.0164326988,0.2473590672,0.0874254778,0.0156229073,0.127114743,-0.2721618414,-0.2168179154,0.4197224975,0.3042122722,0.1447952539,-0.0549958386,0.0315452851,-0.0376000069,0.1919880211,0.0012869322,-0.0891600624,0.1462565213,0.3222911358,0.0391501933,0.028895136,-0.0679283366,-0.4993548393,0.2035862803,-0.2444885522,0.0501396395,-0.0717329159,0.0586968586,-0.144768253,-0.2020163834,-0.2211794704,0.0414226055,0.0379387103,0.3838430047,0.1825157702,-0.1179737598,0.0241606478,0.1218319535,-0.3390520215,0.0490049012,-0.4581525922,0.2769997418,0.041090291,0.1103897989,0.111227043,0.0895713717,0.0747309923,-0.0100377146,-0.0309221353,0.3922051489,0.0827847123,0.20969522,-0.0304660294,0.4936658442,-0.1033023968,-0.1565171629,-0.1688938886,0.5129125714,0.1365161389,-0.1889180839,-0.2576062679,0.5926942229,-0.1301007122,0.3979545534,-0.0644440204,0.1393132955,0.2574175596,-0.076796256,-0.3569566309,-0.2310631573,0.5517593622,-0.1549686044,-0.1069904193,0.0832182616,-0.157275185,-0.1169730276,0.0496916696,-0.1312956363,0.1258497685,0.0575028695,0.1097996756,-0.1443409771,0.1292147636,0.5851309896,0.5111247897,0.1082654819,-0.1515368074,-0.0196426623,-0.3857214451,-0.0863465518,0.0441454612,0.1460323781,0.040510416,0.3825343847,0.1978774816,-0.1548850238,-0.4980463684,-0.1963295192,0.0285993442,0.2656152248,-0.4287972748,0.0746725351,-0.3176436424,0.2282875776,-0.3036859334,-0.083618477,-0.1786009371,-0.570351541,-0.0427637249,0.3845753372,-0.0711474419,0.2192089409,-0.5256514549,0.1033714265,-0.043353498,-0.4788769484,0.0370989516,-0.247161299,-0.0079333717,-0.0047993744,0.3693635762,0.0695809573,0.440527916,-0.2635098994,0.0232031848,-0.5357100964,-0.260789901,-0.0754989013,-0.096307613,0.0138617987,0.145665437,0.1418268979,0.0332270749,-0.2047023028,0.136900261,0.0476083942,-0.1899101436,0.2861983478,-0.2202901244,-0.1222561076,0.1410218179,-0.1501275003,-0.0176324341,-0.22995013,0.0179057028,0.3132324517,0.112975575,-0.0485048927,-0.0028420032,0.0647495463,-0.1155693829,0.3681407571,-0.2924708128,-0.5026839375,0.5586988926,0.001054663,-0.1893103272,0.0034450891,0.0447999425,0.2240406722,0.3211057484,-0.5155474544,-0.0771052018,-0.3473516107,-0.1309498996,0.1703898162,-0.0433537029,0.3131451011,0.1213181391,-0.0129036149,-0.0031683033,-0.4504831433,0.0553054027,-0.0679176524,0.5599042177,-0.2180964351,0.0785483643,-0.2356140614,0.4960045218,0.124429442,-0.1527095437,-0.0077510071,0.1388252676,0.3288232982,-0.1095913649,-0.2257803679,-0.1408218443,-0.1820833236,0.0983357355,0.0209302865,0.038820602,-0.3366082907,-0.1184429303,0.2828342617,-0.2413039804,-0.2129884958,-0.0104347663,-0.2640120685,0.3414264321,0.1733110398,0.1471133232,-0.1718312651,-0.5536513925,-0.1006071717,0.0657591745,-0.1263186783,-0.1281958371,-0.3953514993,-0.0828723535,0.0268551186,0.3618725836,0.0303685889,0.5633289814,-0.0331771187,-0.2938784361,0.0765326247,-0.1728953272,0.0719009489,-0.2082015276,0.1885582656,0.0929391161,-0.1472606063,-0.3871896267,-0.2017133236,-0.1949769557,0.2555857897,-0.0511925071,0.3263749778,-0.365424484,0.0443696119,0.098878853,0.3995558918,-0.211952731,-0.27420035,-0.3901991248,-0.1497893035,-0.170501247,-0.0632799044,-0.2262337357,0.1313824952,-0.1788987517,0.1958269328,0.0281610303,0.0461335257,0.3407544196,0.0278885551,0.2946108282,-0.0845742896,0.3027666211,0.1313607097,-0.1791003346,0.1014210954,0.7728282809,-0.0677529201,-0.3344687819,0.1764199436,-0.0949003324,-0.2939121127,0.3029403687,0.2060928196,-0.0789722875,0.3323389292,-0.0642491654,-0.0421120375,-0.1474913955,0.1818183362,0.0336117297,-0.3301462829,-0.1793639064,0.0923853666,-0.0224105883,-0.0460614711,0.4420637786,0.1555276066,-0.0700891316,0.3109847605,0.4530811012,0.9685325027,0.0399154127,0.2407519519,0.0082418174,-0.2067376822,0.169952333,-0.137561962,0.1208817661,-0.4418897927,0.0415292159,-0.066954568,-0.2064253241,0.0555014275,-0.024467418,0.1306659132,0.3444656134,0.0549289212,0.1411395371,0.0872213989,0.0500618108,-0.2777734101,-0.0530601256,0.1067032889,0.1831928641,-0.139330253,0.2137444764,-0.112701267,0.2057655603,0.0610125363,-0.1952586621,-0.4594775736,0.0500412285,0.1386649013,0.4351476431,0.4156053066,-0.0492290072,-0.2195991278,0.0527029224,0.1416323036,0.0440181941,-0.3519321084,0.1094768792,0.4287138581,0.0711221173,0.0184635539,-0.1683943123,0.300139606,-0.0241657216,-0.1445745528,0.0099768527,0.0385397412,-0.4615333378,-0.2100619674,0.0240343679,-0.2081833184,-0.2153235078,-0.1697953492,-0.1131547913,-0.0830768272,-0.2635005414,0.1405860484,0.2836170495,-0.0946101397,0.3060813248,0.2413632423,-0.357301861,-0.1883834153,0.5763757229,-0.0343189202,-0.1531593353,0.3977429569,0.0097034117,-0.2667919099,-0.2258932889,0.2647129595,-0.0087203169,-0.5552688837,0.0053584343,-0.2442552,0.0947042108,-0.0603761338,0.3846195042,0.2534619868,0.1525187045,0.065535374,-0.5687195063,-0.0102743711,0.0688736513,0.1708159298,0.4163554609,-0.3672188222,-0.3088107705,-0.0868051276,-0.1136470586,-0.2778407931,0.0114196055,-0.3757010996,0.0113293966,-0.1058376431,-0.0557271764,0.186875686,0.0056309486,0.1596178561,-0.0025282833,-0.2202316076,-0.1630436629,-0.2161616534,0.1567677557,0.1282968223,-0.1293980926,0.1067018285,-0.0669196546,0.051574178,-0.0949597508,-0.1071287096,0.3279501796,-0.1974289566,0.068808943,0.0284851287,-0.0793260559,0.0542711206,-0.2081377506,-0.3427086771,-0.1541710347,0.2902843058,-0.0608393215,0.1340801865,-0.0150655042,0.1764710248,0.1667009592,0.0597369187,-0.1213481128,0.3109100163,-0.2965931892,-0.0690154433,0.0695432052,0.3078793883,0.5859784484,-0.3248097897,-0.07010369,0.2360535562,0.1994351149,-0.2696749866,-0.1061143428,0.0882230699,0.0964634344,0.1012218073,0.0777798668,-0.1369308531,-0.2971315086,0.4733383358,0.1976398975,0.194473356,-0.7444229722,0.1211017668,0.1118926033,-0.2094274759,-0.0126876822,0.221334815,-0.1534295678,-0.0909123123,-0.2060396969,-0.0669832826,0.3816144466,-0.363848567,0.209220916,-0.2261677533,-0.3045947254,0.3876189888,-0.0022028389,-0.0506288968,0.1873215586,0.2213814557,0.4170788825,-0.1789329648,0.3228991628,-0.6740233898,-0.1435574293,-0.0505537391,-0.1997704953,-0.3083727062,-0.2873631716,-0.0626987293,0.170352906,-0.0142220473,-0.0185894109,0.3029105067,0.0746686161,-0.2003745437,-0.6321527362,0.1177443117,-0.1633833349,0.0233811345,-0.1873856187,-0.005406308,0.4042905867,-0.070431821,0.0978627726,0.1443993896,0.3789830208,0.3532222807,-0.0114153624,0.0215936359,-0.2928701639,-0.0741365105,0.0963459536,0.8004985452,0.1237498075,-0.1162975207,0.3406516612,0.1436722577,-0.2166982293,0.5199922323,-0.1479876786,-0.0157215651,-0.0350822397,0.7006475329,0.0540913604,0.1681730896,-0.2048673034,0.1254870594,-0.3451822996,-0.4681116343,0.0220951475,0.1330071539,0.0437428169,-0.0183665846,0.0575387254,-0.0470975302,0.5547432303,0.279066503,0.0183346756,-0.4084367454,0.0108955828,-0.6790779233,0.2003026754,-0.3235703111,0.1987060606,-0.0086516589,0.3224900961,0.0390089788,0.4053060412,0.0974019468,-0.1242955327,-0.1956339777,-0.0393807553,-0.2221725732,-0.0104110138,0.2749225795,-0.1489663273,-0.077845186,-0.2355686426,0.0726659149,0.0336348265,0.0702484399,-0.0422693007,-0.2567169666,-0.0293313134,0.2860662937,0.1323132068,0.0755300149,0.6710899472,0.0472505987,-0.1162920445,-0.4180580378,-0.2712435424,0.0032964556,0.1223937422,-0.0506834313,0.5432870984,-0.0329473168,-0.3503838181,-0.180683881,0.2976987064,-0.2001122832,-0.5811730623,-0.130095467,0.0512183458,-0.0686125532,0.2098849118,-0.2360986918,0.0555030815,0.170375064,0.1922392547,-0.0983936265,-0.2720018029,0.6404886842,-0.1993482411,-0.058099445,0.0215242784,0.1502723098,-0.0441082269,-0.1477539539,-0.5046070218,0.1836385876,0.2008164674,0.0420109741,-0.1002981886,0.1392054856,-0.3728256226,0.2734102011,-0.0902340859,0.2114885449,0.0914051682,-0.1839486063,-0.2237300724,-0.155693382]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1856","title":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError","comments":"Same issue today with \"big_patent\", though the symptoms are slightly different.\r\n\r\nWhen running\r\n\r\n```py\r\nfrom datasets import load_dataset\r\nload_dataset(\"big_patent\", split=\"validation\")\r\n```\r\n\r\nI get the following\r\n`FileNotFoundError: Local file \\huggingface\\datasets\\downloads\\6159313604f4f2c01e7d1cac52139343b6c07f73f6de348d09be6213478455c5\\bigPatentData\\train.tar.gz doesn't exist`\r\n\r\nI had to look into `6159313604f4f2c01e7d1cac52139343b6c07f73f6de348d09be6213478455c5` (which is a file instead of a folder) and got the following:\r\n\r\n`Google Drive - Quota exceeded<\/title><meta http-equiv=\"content-type\" content=\"text\/html; charset=utf-8\"\/><link href=/static/doclist/client/css/4033072956-untrustedcontent.css rel=\"stylesheet\" nonce=\"JV0t61Smks2TEKdFCGAUFA\"><link rel=\"icon\" href=\"\/\/ssl.gstatic.com\/images\/branding\/product\/1x\/drive_2020q4_32dp.png\"\/><style nonce=\"JV0t61Smks2TEKdFCGAUFA\">#gbar,#guser{font-size:13px;padding-top:0px !important;}#gbar{height:22px}#guser{padding-bottom:7px !important;text-align:right}.gbh,.gbd{border-top:1px solid #c9d7f1;font-size:1px}.gbh{height:0;position:absolute;top:24px;width:100%}@media all{.gb1{height:22px;margin-right:.5em;vertical-align:top}#gbar{float:left}}a.gb1,a.gb4{text-decoration:underline !important}a.gb1,a.gb4{color:#00c !important}.gbi .gb4{color:#dd8e27 !important}.gbf .gb4{color:#900 !important}\r\n<\/style><script nonce=\"iNUHigT+ENVQ3UZrLkFtRw\"><\/script><\/head><body><div id=gbar><nobr><a target=_blank class=gb1 href=\"https:\/\/www.google.fr\/webhp?tab=ow\">Search<\/a> <a target=_blank class=gb1 href=\"http:\/\/www.google.fr\/imghp?hl=en&tab=oi\">Images<\/a> <a target=_blank class=gb1 href=\"https:\/\/maps.google.fr\/maps?hl=en&tab=ol\">Maps<\/a> <a target=_blank class=gb1 href=\"https:\/\/play.google.com\/?hl=en&tab=o8\">Play<\/a> <a target=_blank class=gb1 href=\"https:\/\/www.youtube.com\/?gl=FR&tab=o1\">YouTube<\/a> <a target=_blank class=gb1 href=\"https:\/\/news.google.com\/?tab=on\">News<\/a> <a target=_blank class=gb1 href=\"https:\/\/mail.google.com\/mail\/?tab=om\">Gmail<\/a> <b class=gb1>Drive<\/b> <a target=_blank class=gb1 style=\"text-decoration:none\" href=\"https:\/\/www.google.fr\/intl\/en\/about\/products?tab=oh\"><u>More<\/u> »<\/a><\/nobr><\/div><div id=guser width=100%><nobr><span id=gbn class=gbi><\/span><span id=gbf class=gbf><\/span><span id=gbe><\/span><a target=\"_self\" href=\"\/settings?hl=en_US\" class=gb4>Settings<\/a> | <a target=_blank href=\"\/\/support.google.com\/drive\/?p=web_home&hl=en_US\" class=gb4>Help<\/a> | <a target=_top id=gb_70 href=\"https:\/\/accounts.google.com\/ServiceLogin?hl=en&passive=true&continue=https:\/\/drive.google.com\/uc%3Fexport%3Ddownload%26id%3D1J3mucMFTWrgAYa3LuBZoLRR3CzzYD3fa&service=writely&ec=GAZAMQ\" class=gb4>Sign in<\/a><\/nobr><\/div><div class=gbh style=left:0><\/div><div class=gbh style=right:0><\/div><div class=\"uc-main\"><div id=\"uc-text\"><p class=\"uc-error-caption\">Sorry, you can't view or download this file at this time.<\/p><p class=\"uc-error-subcaption\">Too many users have viewed or downloaded this file recently. Please try accessing the file again later. If the file you are trying to access is particularly large or is shared with many people, it may take up to 24 hours to be able to view or download the file. If you still can't access a file after 24 hours, contact your domain administrator.<\/p><\/div><\/div><div class=\"uc-footer\"><hr class=\"uc-footer-divider\">© 2021 Google - <a class=\"goog-link\" href=\"\/\/support.google.com\/drive\/?p=web_home\">Help<\/a> - <a class=\"goog-link\" href=\"\/\/support.google.com\/drive\/bin\/answer.py?hl=en_US&answer=2450387\">Privacy & Terms<\/a><\/div><\/body><\/html>`","body":"Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-3-8559a03fe0f8> in <module>()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```","comment_length":230,"text":"load_dataset(\"amazon_polarity\") NonMatchingChecksumError\nHi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.\r\n\r\nTo reproduce:\r\n```\r\nload_dataset(\"amazon_polarity\")\r\n```\r\nThis will give the following error:\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-3-8559a03fe0f8> in <module>()\r\n----> 1 dataset = load_dataset(\"amazon_polarity\")\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/drive.google.com\/u\/0\/uc?id=0Bz8a_Dbh9QhbaW12WVVZS2drcnM&export=download']\r\n```\nSame issue today with \"big_patent\", though the symptoms are slightly different.\r\n\r\nWhen running\r\n\r\n```py\r\nfrom datasets import load_dataset\r\nload_dataset(\"big_patent\", split=\"validation\")\r\n```\r\n\r\nI get the following\r\n`FileNotFoundError: Local file \\huggingface\\datasets\\downloads\\6159313604f4f2c01e7d1cac52139343b6c07f73f6de348d09be6213478455c5\\bigPatentData\\train.tar.gz doesn't exist`\r\n\r\nI had to look into `6159313604f4f2c01e7d1cac52139343b6c07f73f6de348d09be6213478455c5` (which is a file instead of a folder) and got the following:\r\n\r\n`<!DOCTYPE html><html><head><title>Google Drive - Quota exceeded<\/title><meta http-equiv=\"content-type\" content=\"text\/html; charset=utf-8\"\/><link href=/static/doclist/client/css/4033072956-untrustedcontent.css rel=\"stylesheet\" nonce=\"JV0t61Smks2TEKdFCGAUFA\"><link rel=\"icon\" href=\"\/\/ssl.gstatic.com\/images\/branding\/product\/1x\/drive_2020q4_32dp.png\"\/><style nonce=\"JV0t61Smks2TEKdFCGAUFA\">#gbar,#guser{font-size:13px;padding-top:0px !important;}#gbar{height:22px}#guser{padding-bottom:7px !important;text-align:right}.gbh,.gbd{border-top:1px solid #c9d7f1;font-size:1px}.gbh{height:0;position:absolute;top:24px;width:100%}@media all{.gb1{height:22px;margin-right:.5em;vertical-align:top}#gbar{float:left}}a.gb1,a.gb4{text-decoration:underline !important}a.gb1,a.gb4{color:#00c !important}.gbi .gb4{color:#dd8e27 !important}.gbf .gb4{color:#900 !important}\r\n<\/style><script nonce=\"iNUHigT+ENVQ3UZrLkFtRw\"><\/script><\/head><body><div id=gbar><nobr><a target=_blank class=gb1 href=\"https:\/\/www.google.fr\/webhp?tab=ow\">Search<\/a> <a target=_blank class=gb1 href=\"http:\/\/www.google.fr\/imghp?hl=en&tab=oi\">Images<\/a> <a target=_blank class=gb1 href=\"https:\/\/maps.google.fr\/maps?hl=en&tab=ol\">Maps<\/a> <a target=_blank class=gb1 href=\"https:\/\/play.google.com\/?hl=en&tab=o8\">Play<\/a> <a target=_blank class=gb1 href=\"https:\/\/www.youtube.com\/?gl=FR&tab=o1\">YouTube<\/a> <a target=_blank class=gb1 href=\"https:\/\/news.google.com\/?tab=on\">News<\/a> <a target=_blank class=gb1 href=\"https:\/\/mail.google.com\/mail\/?tab=om\">Gmail<\/a> <b class=gb1>Drive<\/b> <a target=_blank class=gb1 style=\"text-decoration:none\" href=\"https:\/\/www.google.fr\/intl\/en\/about\/products?tab=oh\"><u>More<\/u> »<\/a><\/nobr><\/div><div id=guser width=100%><nobr><span id=gbn class=gbi><\/span><span id=gbf class=gbf><\/span><span id=gbe><\/span><a target=\"_self\" href=\"\/settings?hl=en_US\" class=gb4>Settings<\/a> | <a target=_blank href=\"\/\/support.google.com\/drive\/?p=web_home&hl=en_US\" class=gb4>Help<\/a> | <a target=_top id=gb_70 href=\"https:\/\/accounts.google.com\/ServiceLogin?hl=en&passive=true&continue=https:\/\/drive.google.com\/uc%3Fexport%3Ddownload%26id%3D1J3mucMFTWrgAYa3LuBZoLRR3CzzYD3fa&service=writely&ec=GAZAMQ\" class=gb4>Sign in<\/a><\/nobr><\/div><div class=gbh style=left:0><\/div><div class=gbh style=right:0><\/div><div class=\"uc-main\"><div id=\"uc-text\"><p class=\"uc-error-caption\">Sorry, you can't view or download this file at this time.<\/p><p class=\"uc-error-subcaption\">Too many users have viewed or downloaded this file recently. Please try accessing the file again later. If the file you are trying to access is particularly large or is shared with many people, it may take up to 24 hours to be able to view or download the file. If you still can't access a file after 24 hours, contact your domain administrator.<\/p><\/div><\/div><div class=\"uc-footer\"><hr class=\"uc-footer-divider\">© 2021 Google - <a class=\"goog-link\" href=\"\/\/support.google.com\/drive\/?p=web_home\">Help<\/a> - <a class=\"goog-link\" href=\"\/\/support.google.com\/drive\/bin\/answer.py?hl=en_US&answer=2450387\">Privacy & Terms<\/a><\/div><\/body><\/html>`","embeddings":[-0.1977470964,0.1699575484,-0.0510927998,0.2398218662,0.1494399905,-0.0500105023,0.2648400664,0.2264509648,0.3186415434,0.0602953546,-0.0621266589,0.0072547733,-0.0048490805,-0.1367188245,-0.2075065523,0.3518310189,0.1385649145,-0.0169535168,-0.2259685993,0.0192921013,-0.2470269948,0.2507219613,-0.0814839229,-0.1623295397,-0.0814451873,0.2453400195,0.185918808,0.1196510941,-0.0924422145,-0.4332829714,0.3541623354,0.1207953468,0.0554162674,0.3615809679,-0.000124711,0.1088824868,0.3887200952,0.0579437539,-0.2862367332,-0.1727553755,-0.4470102191,-0.3646294475,-0.0332945846,0.0008812062,0.1849218458,0.0702594444,0.0694613382,-0.063148044,0.0055411682,0.2387625873,0.1421351284,0.4269422889,0.2395886481,0.1177865267,0.2658692002,0.2129864693,-0.0901237801,0.2691429257,0.2103940696,-0.2142736167,-0.2508116066,0.1191412657,-0.1859516799,0.414689064,0.103681609,0.0116963284,0.2535794973,-0.1878900826,0.2364885658,0.5302149653,0.3473127186,-0.0848420411,-0.4574839473,-0.1686614156,-0.1927171648,-0.3431209922,0.3405172825,0.2637237608,-0.0864364654,-0.1382098645,-0.3554598391,0.1351724267,-0.0891586468,0.2401632071,-0.0999407172,0.1585044861,0.1089744642,0.0511102416,0.0896147639,-0.283146441,0.4739673436,-0.4253416359,-0.0152056171,0.1840294003,-0.5253770947,0.0538717397,0.0191307981,0.5155975819,0.3433676362,0.4453732371,0.1638211012,0.1542468518,-0.113208957,0.1749989688,0.1005771756,0.164857313,-0.1246671528,0.2837972641,0.334449023,0.2328185886,0.0663328096,0.1549246907,-0.0565032437,-0.2741813958,0.5267841816,-0.1108523384,0.313787967,-0.5724887252,-0.4306190014,0.3364908397,-0.1928468943,-0.0282177832,0.0675675794,0.2280802429,-0.2387456298,0.3099794388,-0.0643480495,0.0004804438,-0.1664965302,0.0502480827,-0.2033563405,-0.0187495016,-0.0810727403,0.2882333994,0.3058470488,-0.0299485289,0.3222577274,-0.0148619581,0.2578918934,-0.0937811062,0.2736911178,-0.1426098198,0.0097336601,0.4259118736,-0.2763498724,0.0317265801,0.0826817825,0.0599154346,-0.14994964,0.1888548881,-0.1219960228,-0.1037561074,0.0341112874,0.0855986401,-0.7176007032,-0.1909341216,0.0236068051,-0.4370221794,0.2319756001,-0.4660434127,-0.0455491692,-0.579919517,0.0250526201,-0.1446424425,0.1525873095,0.3106122613,-0.1575115323,-0.1112049595,-0.1945011467,-0.0421186164,0.3214511275,0.3376633227,-0.0974220037,0.0249704383,-0.2604079545,-0.1538295597,0.0182919856,-0.2152289599,-0.4789546132,0.0974535346,-0.1639339775,0.5817306042,0.1618592888,0.164980486,-0.2130648941,0.0506925024,0.2002895921,0.0676911175,-0.0125752799,0.163019225,-0.2348095179,-0.1929581016,0.3607847989,0.2327778339,0.2313600928,-0.024613047,0.0942839235,-0.0469427705,0.2663872242,-0.0295545198,-0.1264053881,0.1404914707,0.2977230251,-0.0565797649,-0.0152171683,-0.0363365822,-0.451076448,0.28875646,-0.1545195878,0.0236894451,-0.1039606258,0.1078216061,-0.129286468,-0.2308788747,-0.2409704179,-0.0934643894,-0.0329290666,0.4628794491,0.1729547232,0.0034063782,0.028061077,0.2453964949,-0.2640742064,0.1012686566,-0.58715415,0.2537441254,0.109455049,0.0982457623,0.0388326831,0.0806021243,0.0688670129,-0.0402693488,-0.0483747236,0.3387963176,0.2107178122,0.1517692059,-0.0045256522,0.5949782729,-0.0514891185,-0.0401122868,-0.0681998283,0.4075601697,0.1674464047,-0.1222714782,-0.2599912584,0.6478692889,-0.1487545967,0.4013353586,-0.0244220961,-0.0253933091,0.2595122159,-0.1413267404,-0.4156427681,-0.0606277175,0.4904745519,-0.0587507896,-0.1178299859,0.0945851952,-0.2039511204,-0.1534876078,0.2384014726,-0.1865176409,-0.0035637394,0.0259025954,-0.1144873649,-0.0729258582,0.1115812436,0.6173869371,0.6931012869,0.0906759426,-0.1532764137,-0.0719935223,-0.2084608376,-0.1319147497,0.0589391328,0.1659445763,-0.0344939902,0.4840489924,0.2609243989,-0.174152717,-0.454282701,-0.2900517881,0.0969517305,0.1641105264,-0.4628485143,0.0809603855,-0.3064672649,0.1440186054,-0.447399646,0.0089243902,-0.3047971427,-0.4628069103,-0.1676448882,0.5472712517,-0.0157070011,0.2009985298,-0.6645430326,0.1647136211,-0.0594837628,-0.4675349593,0.0693726167,-0.1088770702,-0.0143261012,-0.0495197065,0.3760409653,-0.0842460617,0.3908686936,-0.191077143,-0.0181511585,-0.5628102422,-0.4042610824,-0.0315472782,-0.0961567909,-0.0866458192,0.1487677097,0.0661067739,0.2045079023,-0.1082295701,0.1219889745,0.0034571928,-0.2515302896,0.1921658814,-0.13921386,-0.0867666751,0.1303486377,-0.1573179215,-0.0555603839,-0.2416144013,0.1264010817,0.2230749875,0.0426238962,0.0844557211,-0.0340708643,0.1342974901,-0.184631452,0.2868455052,-0.2110422105,-0.4735352993,0.4939221442,0.0137702432,-0.208632797,-0.1078974828,0.0114796618,0.203142941,0.3830661178,-0.512294054,-0.0500408672,-0.3821271956,-0.0122855278,0.1495184898,-0.0381680317,0.2434116453,0.0776330978,0.000671882,-0.0634186715,-0.3752563596,0.0119352052,-0.0515981019,0.5554400682,-0.1658086777,0.1307981014,-0.2195640504,0.520567596,0.0204533804,-0.1773886681,-0.0150971878,0.1209319681,0.4579964578,-0.0820356086,-0.2375092506,-0.0136436708,-0.0928557515,0.0310114902,0.0167738423,-0.0459583513,-0.0546080209,-0.0233056676,0.175757274,-0.1513470411,-0.1952749044,0.0736327022,-0.2062538415,0.266628772,0.149187848,0.0829503462,-0.2736793458,-0.4601077437,-0.088315852,0.0990170762,0.0131992623,-0.1141730025,-0.3601973057,0.0028565181,-0.0865938887,0.3694247007,0.0677432865,0.537417531,0.0034999822,-0.4062408209,-0.0274816472,-0.1980948597,0.1366263181,0.0413336381,0.1518649757,0.0760885179,-0.2618225515,-0.2618720233,-0.3011713624,-0.1455934197,0.19725281,0.0102241104,0.4263956845,-0.311190486,0.0090129096,0.0770427734,0.5044949651,-0.1736919582,-0.1661171466,-0.3287209868,-0.1103904769,-0.2056304365,-0.0670005828,-0.1488405615,0.1898534149,-0.276004374,0.2880747616,-0.1159316003,-0.0420524962,0.3881800175,0.0811637193,0.2963171303,-0.0851844773,0.3596169949,0.186571151,-0.0997020006,0.2935777307,0.7036969662,-0.0759230629,-0.4403658807,0.137074858,-0.0812451616,-0.2529205382,0.3446392417,0.0421708897,-0.0207267478,0.4141941667,-0.0693849549,-0.1543785036,0.0517622791,0.2886786163,0.1154750884,-0.3386641741,-0.2388708442,0.1438050568,0.0697232559,-0.1526653469,0.4629602432,0.0509198271,-0.1079159006,0.1011270806,0.1203639582,0.8393642306,0.1140443757,0.3778189421,0.0267085116,-0.0391329005,0.2921770513,-0.0136149749,0.200063169,-0.4918188155,-0.231799826,-0.066881448,-0.3209642768,0.1960917115,-0.1047178134,0.2709470987,0.4352650344,0.219897151,0.2370737195,0.1346770376,0.0994503722,-0.1265523136,0.0080762887,0.0543688424,0.0991846025,-0.1956755519,0.4031546116,-0.0468659289,0.0923537761,-0.0931581333,-0.2360897809,-0.4514454603,0.1142913848,0.0259427801,0.4191182256,0.3971019387,-0.0595751666,-0.2479570806,0.1414895207,0.1691755205,0.0425854027,-0.2766382396,0.1784698069,0.4671690464,0.0909846425,0.0286954213,-0.217767626,0.3202312589,0.0191051345,-0.1145873368,-0.0044737221,-0.0086997626,-0.4215155244,-0.2650597394,-0.0038227499,-0.017931439,-0.2658550143,-0.1391578019,-0.1247063354,-0.0123555632,-0.2472330034,0.0871947631,0.3706959486,-0.1243250668,0.230215475,0.1520808786,-0.3544494808,-0.2047877759,0.5908538699,-0.1227731332,-0.2588196397,0.5218557119,-0.1136019304,-0.1334782541,-0.1409383416,0.2275197059,0.0963099152,-0.4837852418,0.0500919037,-0.2060242742,0.0261208992,0.0078299483,0.2911912799,0.224415347,-0.0170446094,0.0997881144,-0.5264368057,-0.079675585,0.095192641,0.0575174987,0.4152068496,-0.2721105516,-0.2381439954,-0.3541296422,-0.275493741,-0.227135554,0.0920687243,-0.2097597271,-0.0396223813,-0.018890826,-0.0343643464,0.2426287085,-0.0293570571,0.1448670775,0.0459998511,-0.1230620444,-0.121265389,-0.2768180966,0.2021892816,0.1157276183,-0.09818241,0.0693174452,-0.094187282,0.1100542322,-0.112745598,0.1177112684,0.2905290723,-0.1701913476,-0.0598585382,0.0158686377,-0.0449269935,0.1417570412,-0.0080648074,-0.3633030951,-0.0902774855,0.3748225868,0.0704593882,0.0243402254,0.0274133142,0.0599086136,0.1317327768,-0.031007763,-0.1302216053,0.4147023261,-0.3255168498,-0.1937369704,0.0743361413,0.2790163159,0.6166602373,-0.3277213871,-0.1052146778,0.2799766064,0.1788962334,-0.2058532685,-0.1430945098,0.1972789317,-0.0152627882,0.1052272022,0.15600501,-0.0999647751,-0.2674521208,0.3703506589,0.1576055288,0.2714168131,-0.4941527843,0.1963050812,0.2657337189,-0.1613769382,-0.0107379332,0.2556597292,-0.0921969339,-0.0656126291,0.0056155967,-0.0532908477,0.3297795355,-0.4341211319,0.0854645222,-0.1895896941,-0.4409664273,0.2892761827,0.0798092261,-0.0337830223,0.24952133,0.1416720301,0.3144389093,-0.216625765,0.3021011949,-0.5998132229,-0.0176931396,-0.0304031726,-0.1813414991,-0.1573509127,-0.3583907783,0.092484206,0.2413350791,-0.0094321864,0.0255111028,0.3337188363,0.1591835916,-0.251145035,-0.5507925153,-0.083507821,-0.0567499623,0.0359930247,-0.1781905293,0.0337067731,0.4150089025,-0.1086321622,0.1495471895,0.2184256613,0.5513748527,0.3730784655,0.0504292473,-0.1465508938,-0.2240879536,-0.0177874751,0.0449721962,0.7161060572,-0.014010489,-0.0573871024,0.3180462122,0.0889363438,-0.1314575523,0.3808566332,-0.2900702357,-0.1717511117,-0.086082153,0.621068418,0.0097138453,0.1610928327,-0.243621543,0.1875264049,-0.3157789707,-0.398067534,0.1399549991,0.1968461871,0.0464000814,0.0836229697,0.0227409936,-0.1517202854,0.5273543,0.3188101053,0.0647680759,-0.4174135923,0.0885208771,-0.7221711278,0.3024897873,-0.3389039934,0.1521541774,-0.0938103944,0.2113615274,0.0834675655,0.2010390908,0.0466543287,-0.2246107459,-0.1208973676,-0.1212607324,-0.2425459027,-0.0915076956,0.323215723,-0.216842249,-0.0514252856,-0.3011354804,0.1313190013,-0.0132247182,0.0739568397,0.012301378,-0.0289639607,-0.1769033074,0.2476486266,0.1143122017,0.1731823832,0.7652651668,0.0657588169,-0.2970636487,-0.4681603611,-0.2858740389,-0.0980945379,0.2642611861,-0.1359898746,0.6043872833,-0.0054838397,-0.3212574124,-0.2743748426,0.3681740165,-0.159987241,-0.5791293383,0.0425796807,0.0364802517,-0.1174707785,0.235993281,-0.1278226525,0.1046727449,0.1012215689,0.2595416009,-0.16834566,-0.31324175,0.643245697,-0.3836359382,-0.0290614609,-0.0133755496,0.2157868892,-0.0461704023,-0.0788291618,-0.5784170628,0.2646857202,0.275985539,0.0526711568,-0.2628392279,0.1017507762,-0.2769649029,0.2739332616,-0.0319270119,0.2957463562,0.022110695,-0.3565128744,-0.1994646043,-0.030063143]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1854","title":"Feature Request: Dataset.add_item","comments":"Hi @sshleifer.\r\n\r\nI am not sure of understanding the need of the `add_item` approach...\r\n\r\nBy just reading your \"Desired API\" section, I would say you could (nearly) get it with a 1-column Dataset:\r\n```python\r\ndata = {\"input_ids\": [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])]}\r\nds = Dataset.from_dict(data)\r\nassert (ds[\"input_ids\"][0] == np.array([4,4,2])).all()\r\n```","body":"I'm trying to integrate `huggingface\/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L318, as opposed to loading all the text into arrow, and then `dataset.map(binarizer)`.\r\nIs this possible at the moment? Is there an example? I'm happy to use raw `pa.Table` but not sure whether it will support uneven length entries.\r\n\r\n### Desired API\r\n\r\n```python\r\nimport numpy as np\r\ntokenized: List[np.NDArray[np.int64]] = [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])\r\n\r\ndef build_dataset_from_tokenized(tokenized: List[np.NDArray[int]]) -> Dataset:\r\n \"\"\"FIXME\"\"\"\r\n dataset = EmptyDataset()\r\n for t in tokenized: dataset.append(t)\r\n return dataset\r\nds = build_dataset_from_tokenized(tokenized)\r\nassert (ds[0] == np.array([4,4,2])).all()\r\n```\r\n\r\n### What I tried\r\ngrep, google for \"add one entry at a time\", \"datasets.append\"\r\n\r\n### Current Code\r\nThis code achieves the same result but doesn't fit into the `add_item` abstraction.\r\n\r\n```python\r\n dataset = load_dataset('text', data_files={'train': 'train.txt'})\r\n tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base', max_length=4096)\r\n def tokenize_function(examples):\r\n ids = tokenizer(examples['text'], return_attention_mask=False)['input_ids']\r\n return {'input_ids': [x[1:] for x in ids]}\r\n ds = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=['text'], load_from_cache_file=not overwrite_cache)\r\n\tprint(ds['train'][0]) => np array\r\n```\r\n\r\nThanks in advance!","comment_length":48,"text":"Feature Request: Dataset.add_item\nI'm trying to integrate `huggingface\/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L318, as opposed to loading all the text into arrow, and then `dataset.map(binarizer)`.\r\nIs this possible at the moment? Is there an example? I'm happy to use raw `pa.Table` but not sure whether it will support uneven length entries.\r\n\r\n### Desired API\r\n\r\n```python\r\nimport numpy as np\r\ntokenized: List[np.NDArray[np.int64]] = [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])\r\n\r\ndef build_dataset_from_tokenized(tokenized: List[np.NDArray[int]]) -> Dataset:\r\n \"\"\"FIXME\"\"\"\r\n dataset = EmptyDataset()\r\n for t in tokenized: dataset.append(t)\r\n return dataset\r\nds = build_dataset_from_tokenized(tokenized)\r\nassert (ds[0] == np.array([4,4,2])).all()\r\n```\r\n\r\n### What I tried\r\ngrep, google for \"add one entry at a time\", \"datasets.append\"\r\n\r\n### Current Code\r\nThis code achieves the same result but doesn't fit into the `add_item` abstraction.\r\n\r\n```python\r\n dataset = load_dataset('text', data_files={'train': 'train.txt'})\r\n tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base', max_length=4096)\r\n def tokenize_function(examples):\r\n ids = tokenizer(examples['text'], return_attention_mask=False)['input_ids']\r\n return {'input_ids': [x[1:] for x in ids]}\r\n ds = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=['text'], load_from_cache_file=not overwrite_cache)\r\n\tprint(ds['train'][0]) => np array\r\n```\r\n\r\nThanks in advance!\nHi @sshleifer.\r\n\r\nI am not sure of understanding the need of the `add_item` approach...\r\n\r\nBy just reading your \"Desired API\" section, I would say you could (nearly) get it with a 1-column Dataset:\r\n```python\r\ndata = {\"input_ids\": [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])]}\r\nds = Dataset.from_dict(data)\r\nassert (ds[\"input_ids\"][0] == np.array([4,4,2])).all()\r\n```","embeddings":[-0.2646298409,0.1603154093,-0.0458905734,0.1386155337,-0.0252806228,0.1201700866,0.1569381356,0.1071401313,0.058670368,0.0440863669,0.1095488444,0.5823287368,-0.1573915929,0.138007164,0.4451274276,-0.1356466115,0.1021329612,0.2048072368,-0.1752964556,0.0704091564,-0.2645285428,-0.0905520767,-0.1519214809,-0.1790196747,-0.0579325259,-0.061322201,-0.4128628969,-0.3023573458,-0.2655974329,-0.7929652929,-0.0123855649,-0.0009123574,-0.0839308575,0.2624169588,-0.0001283345,0.0117334425,0.1670986712,-0.0539960302,-0.5640701652,-0.0456967317,-0.1553967893,-0.2387477756,0.1454959214,-0.3138717115,0.2260569036,-0.7180809379,0.3229591548,-0.1545696706,0.089084737,0.125013724,-0.0003758298,0.494810909,0.1991044879,-0.3639475703,0.2621561289,-0.0453420877,-0.1395283341,-0.1721628904,0.3763104081,-0.2068813294,0.0824214742,-0.1125964671,0.2658927441,-0.0915565714,0.4066393673,0.1262135953,-0.0127451709,-0.1426652074,-0.1565271169,0.2903876007,0.2264159918,-0.419044286,-0.6362015009,-0.4910083711,0.3017780781,-0.0399683453,-0.1348128319,0.060155388,-0.0711733475,-0.0760381594,-0.069809854,-0.3770206571,-0.111912109,0.2381709218,0.0717944801,0.4428234696,0.1425431371,0.1643607467,0.1441977918,-0.1271944046,-0.2345662117,0.0151866274,-0.1232940331,0.3333287835,-0.1031110138,-0.1166460291,0.1092310101,-0.0470127016,0.4755124152,-0.061299704,0.173726812,0.1157122627,-0.3675787151,0.0866340697,0.171476528,-0.1008807644,0.2137441039,-0.2119158655,-0.3128740191,0.0772754475,0.4092112184,0.0620396063,-0.1824602634,0.1073736846,-0.2618193328,-0.3582352996,-0.0681386217,0.4037789404,-0.1207616106,-0.0273630079,-0.0106998915,0.0841955394,-0.0472879857,0.173677966,0.1717312634,-0.3679118156,-0.2393834144,0.2014809549,0.1263076216,-0.0102789896,0.0816346034,0.217958957,-0.1080401167,0.1765950322,0.1008231193,-0.001051515,-0.014450158,-0.0649894327,0.0691936165,0.0783881024,-0.0114106443,-0.1804757416,0.2521993816,0.0106602665,0.0532554165,-0.0456522033,-0.1014854684,-0.4404917955,-0.4781494141,0.2198809236,-0.1078879684,-0.1812904775,-0.2244119793,-0.0641979277,0.0042391834,-0.1642895639,-0.1952681541,0.5241122246,0.3831294477,-0.2358407676,0.1302087307,-0.1283583939,-0.5635914207,-0.1791516542,0.1944267601,0.3399145603,-0.4468756914,-0.3274410367,-0.1016851589,-0.0158187244,-0.4132902026,0.1295030862,-0.0944460928,0.2800901532,-0.4338248074,0.4084133506,0.3091626167,-0.1658173501,-0.2207370549,0.1308985353,-0.1782537401,0.0376557037,0.0842344984,0.3958645165,0.2953098118,0.060019996,0.4510013461,0.2585290372,-0.0795216337,-0.0899351388,0.0736893639,-0.4146664441,0.3007451296,-0.0111934626,-0.1263205856,-0.0905314982,0.0279579051,-0.5407937765,0.1644920558,-0.1066171676,0.3368983269,-0.0383560583,0.4019224942,-0.0076305624,-0.1058239862,-0.5203376412,-0.3627278507,0.1384117454,-0.0773245841,0.1222194284,-0.1124037206,-0.4358099401,-0.1482965052,0.4267143309,-0.1347390115,-0.0214686058,-0.0691511258,0.0572068244,-0.0093465028,-0.1200825647,-0.4359641373,0.1428783089,0.2283416539,0.3076541424,-0.4393087029,-0.049837932,0.2686482966,-0.1849064827,-0.1217021868,0.1733721644,-0.141075328,-0.1443530619,0.208371371,0.2823854685,-0.0867410675,-0.27010414,0.1203915998,0.2604833245,0.2900448143,0.3060085773,-0.1001439542,-0.2057063431,0.3140789568,-0.2532829642,-0.2291707546,0.5499863029,0.0419301353,0.4261950552,0.1370570213,0.0392848514,0.0201952625,0.2511620224,-0.2537923455,-0.2578029633,-0.0499058366,0.1488565952,0.2123851925,-0.1196054667,-0.1045179814,0.1626659185,0.1422627568,0.1596186012,0.1770738661,0.1443223208,-0.2144739032,0.0623016506,0.0679695085,0.0149866864,0.4575430453,0.0544013269,0.1145165935,0.0659159645,0.111045748,0.0568909645,0.0589827783,0.031461373,-0.1170743406,0.0384728573,0.5389946699,0.0615468398,-0.0218392145,-0.1388209462,0.1136970893,-0.0432560146,-0.1218187734,0.0397563912,0.0448243991,0.0722149983,-0.2530289888,-0.3716776073,-0.0849188417,-0.0843767971,0.2689409852,0.0314313322,0.0156413484,0.301081568,0.137987569,0.7158766985,0.1047442034,-0.4731730223,-0.0859230235,0.0065621044,0.2182613462,-0.101071611,-0.1031731069,0.3030089736,0.402112186,0.0869909003,-0.0316570625,-0.3475198448,-0.5039143562,0.1095275655,-0.051787585,0.4337313473,0.3948709071,-0.335601896,-0.1123425141,-0.2517050803,0.1836480349,0.0373475961,-0.1394076496,-0.1860643178,-0.0976857767,-0.2373919785,-0.0496559888,0.1030326337,-0.206016019,-0.1766934991,0.4733741581,0.1536125243,0.2519126236,-0.2228932679,0.0982955918,0.0336947665,-0.1396936774,-0.1208668947,0.0718317702,-0.0042209281,0.4730607271,-0.3619567752,-0.1762785763,-0.1606230736,-0.5109411478,0.1114847213,0.377494216,-0.2879175246,-0.3993910253,-0.2613127232,0.3807624578,-0.0936847404,0.2705568075,0.4955573678,0.0041033556,0.1533956528,0.0573997162,-0.287943095,0.374073267,0.1856687218,0.0599000379,0.0809817091,0.1704524159,0.0867763981,0.663498342,0.1845435947,-0.1084073484,0.1433631629,-0.5918194056,0.3627535999,-0.2168656737,-0.2220130712,-0.1796467155,-0.1106616184,0.0941357389,0.0607794598,-0.0263928659,0.1260304004,0.0712459236,0.3468834758,0.0541964993,-0.1985840499,0.1793527603,0.0210639574,0.3373360038,-0.0193252079,-0.0224008486,-0.3871428668,0.017365003,-0.0409920514,0.2764789164,0.2389587611,-0.1770690531,-0.4978469908,-0.140987426,-0.307357192,0.0694869161,0.3090179861,0.4473903477,0.2480776906,-0.1597179025,0.0432345122,-0.2020529509,0.7236328125,0.325334698,-0.4594327211,0.0084282393,0.0938177109,-0.2563486099,0.173294425,0.0139204469,0.5422018766,0.1814727485,0.2524054945,-0.4542917013,0.2550082803,0.2051475048,0.179388687,-0.181665659,-0.1778385043,-0.1138078868,-0.1289127022,-0.2315636277,0.1871423721,0.1932213008,0.2209282368,0.4691876769,-0.0989606008,-0.4084217846,0.0668774918,0.1020533815,-0.0333174728,0.0044961032,0.1497972906,0.5072354674,-0.1110389158,-0.2078583688,0.4400340915,0.4042310417,-0.1330380738,-0.4836511612,-0.1611758918,-0.2042241693,0.4172766507,0.2799086571,0.1667552888,0.2738693357,0.048915837,0.167960912,-0.2383995652,0.2513348758,0.0925280228,-0.1277251542,-0.358201474,-0.3206029534,0.4401080608,-0.0505429246,-0.0510863662,0.0931559503,0.1860111803,-0.3132160604,0.3559414148,-0.0383071564,0.8532413244,0.0140183745,0.4227761626,0.2781120539,-0.0398610048,0.3147439659,-0.0383674316,0.211047858,-0.4547417462,-0.213548556,-0.0393988378,-0.1463226527,0.160040319,0.2031617761,-0.1510424614,0.2056574374,0.2487069666,0.0024850443,-0.1176300347,0.2441386729,-0.2815453708,-0.4994777739,-0.3187270463,-0.078942515,0.1857927144,-0.0679338351,-0.1185278744,-0.0023444057,-0.0017808152,-0.2800907791,-0.3747290373,-0.3991499245,-0.3093171418,-0.1589401513,0.6127160788,-0.264888227,0.2600618899,-0.0347001292,-0.0342629179,0.1709604114,0.1167391017,0.0811085179,-0.0193566158,-0.1360162348,-0.0407432057,-0.390902698,0.1759610772,0.2466171384,0.0166306775,0.2744945884,-0.1171826869,-0.2790036798,-0.2106477618,0.2183814049,0.2384328395,-0.2930262983,-0.2510787547,0.1993103027,0.1179536805,-0.2569961548,-0.1151394546,0.1747874022,-0.2499647886,0.1552372724,0.0525324047,-0.0369467661,-0.0058731325,0.2975792289,-0.1896295846,-0.2781789303,0.3539299667,0.142874673,-0.0727733895,-0.0792888477,-0.0266168173,0.4377448261,-0.2136408538,-0.0719576329,0.0547067747,-0.0483799987,-0.2827552557,0.0705747306,0.0742127374,0.1839034557,-0.105127506,-0.2911681533,-0.4044686854,0.2763894796,0.3002138138,0.3721829951,0.1848497391,0.1702437699,-0.252456069,0.4358135462,-0.1169064119,0.1479174644,0.1459783763,0.150271371,0.1854646951,0.1160024181,-0.0826214179,0.1280495822,-0.0782819465,0.1550616622,-0.1440820694,0.0151926642,-0.1747512966,0.2127382159,0.2331584394,-0.0112262378,-0.1478844434,-0.1146837622,-0.176208809,-0.3947841823,0.3163344562,-0.0946518704,-0.2700203061,-0.1996149123,0.6224370599,0.2577295005,-0.0989504829,0.5540859103,0.2836504579,0.4706646502,0.1808962822,0.2874778509,-0.352455318,0.0764467418,0.3385862112,0.4616112411,0.3684318066,-0.0464078262,0.0408415087,-0.206032604,-0.0081247808,0.1402797848,0.2286501378,0.5841594338,-0.2424769998,0.0834469348,0.338927418,0.0132566243,-0.1886738092,0.0095983101,0.2829734981,-0.1192660555,0.2885172665,0.4438749254,-0.1419064254,0.1382748783,-0.156646654,0.0861744583,0.0100405058,0.0146586439,-0.1547353566,-0.0930767283,-0.0351524539,-0.1431533396,0.1480107158,0.3850335181,0.3257227838,0.5992687345,-0.2377139628,0.4989932179,0.5984037519,0.006602589,0.1962082088,0.0195396338,0.0539919026,0.5260339975,0.6944862008,0.1154351905,0.3511893451,0.4129686952,-0.0707862154,-0.0713928416,0.0020068907,0.1799687743,0.1909280419,-0.3396001756,0.1210675761,-0.1997504383,-0.2255669534,-0.0936933458,-0.0438627005,-0.3908543885,0.0556977428,-0.0328667648,0.3594781458,-0.4826160967,0.0741701871,0.2190117836,0.0676174909,-0.3067223728,0.1683755219,0.0193257648,0.1612189412,0.0985937864,0.0885514244,0.1539871991,0.2236357629,-0.0412270315,0.3775160313,-0.5520080328,0.0923720151,-0.0642093644,-0.0626926571,-0.0128712729,-0.050201308,0.2981148064,-0.0703727081,0.0275407471,0.1684700698,0.2442592531,-0.1106707528,-0.58131814,-0.0054786503,0.0077514676,-0.276887387,-0.0820569247,-0.014521867,-0.0441375114,-0.0687596053,0.2984167933,-0.3923063278,0.0352506787,0.1657793969,-0.014820463,-0.1587476432,0.163619712,0.5065175295,0.2900012434,-0.0710300431,-0.2468840182,-0.5102574825,-0.0098479483,-0.1119040251,-0.1957143694,-0.0556481071,0.1179482937,0.0269699115,0.2001409233,0.161121726,-0.009064951,-0.0340892337,0.3953381181,-0.178308934,-0.0462915748,-0.0857851058,0.1022696123,0.1350229979,-0.2631911635,-0.1652940661,-0.0707383081,-0.1621331573,-0.1025682688,-0.0092492159,-0.0537614636,-0.124688983,0.5808807611,0.2099566162,0.1434181631,-0.1043718234,-0.0140569117,-0.2172804922,-0.1125323474,-0.3362767696,0.60516572,0.0282509923,0.0629272088,-0.1458367407,0.0355633162,0.0187913217,0.0046900376,0.0892241299,-0.1565242857,-0.1970908344,0.3591445684,-0.396266222,0.1175956801,-0.1363485456,0.3045372665,0.2100229561,0.2169963717,0.1847030967,-0.0915804729,0.2319037169,-0.454172641,-0.323569566,-0.0779609904,0.0218928661,0.0010353213,-0.3787177801,-0.5221460462,-0.287507683,0.3396661878,-0.0111592943,-0.0437442325,0.0462231226,-0.1847174615,-0.1237373427,-0.4769803584,0.4785841107,-0.0198921319,-0.0834573656,-0.0986276343,-0.5801851153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1854","title":"Feature Request: Dataset.add_item","comments":"Hi @sshleifer :) \r\n\r\nWe don't have methods like `Dataset.add_batch` or `Dataset.add_entry\/add_item` yet.\r\nBut that's something we'll add pretty soon. Would an API that looks roughly like this help ? Do you have suggestions ?\r\n```python\r\nimport numpy as np\r\nfrom datasets import Dataset\r\n\r\ntokenized = [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])\r\n\r\n# API suggestion (not available yet)\r\nd = Dataset()\r\nfor input_ids in tokenized:\r\n d.add_item({\"input_ids\": input_ids})\r\n\r\nprint(d[0][\"input_ids\"])\r\n# [4, 4, 2]\r\n```\r\n\r\nCurrently you can define a dataset with what @albertvillanova suggest, or via a generator using dataset builders. It's also possible to [concatenate datasets](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?highlight=concatenate#datasets.concatenate_datasets).","body":"I'm trying to integrate `huggingface\/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L318, as opposed to loading all the text into arrow, and then `dataset.map(binarizer)`.\r\nIs this possible at the moment? Is there an example? I'm happy to use raw `pa.Table` but not sure whether it will support uneven length entries.\r\n\r\n### Desired API\r\n\r\n```python\r\nimport numpy as np\r\ntokenized: List[np.NDArray[np.int64]] = [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])\r\n\r\ndef build_dataset_from_tokenized(tokenized: List[np.NDArray[int]]) -> Dataset:\r\n \"\"\"FIXME\"\"\"\r\n dataset = EmptyDataset()\r\n for t in tokenized: dataset.append(t)\r\n return dataset\r\nds = build_dataset_from_tokenized(tokenized)\r\nassert (ds[0] == np.array([4,4,2])).all()\r\n```\r\n\r\n### What I tried\r\ngrep, google for \"add one entry at a time\", \"datasets.append\"\r\n\r\n### Current Code\r\nThis code achieves the same result but doesn't fit into the `add_item` abstraction.\r\n\r\n```python\r\n dataset = load_dataset('text', data_files={'train': 'train.txt'})\r\n tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base', max_length=4096)\r\n def tokenize_function(examples):\r\n ids = tokenizer(examples['text'], return_attention_mask=False)['input_ids']\r\n return {'input_ids': [x[1:] for x in ids]}\r\n ds = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=['text'], load_from_cache_file=not overwrite_cache)\r\n\tprint(ds['train'][0]) => np array\r\n```\r\n\r\nThanks in advance!","comment_length":92,"text":"Feature Request: Dataset.add_item\nI'm trying to integrate `huggingface\/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L318, as opposed to loading all the text into arrow, and then `dataset.map(binarizer)`.\r\nIs this possible at the moment? Is there an example? I'm happy to use raw `pa.Table` but not sure whether it will support uneven length entries.\r\n\r\n### Desired API\r\n\r\n```python\r\nimport numpy as np\r\ntokenized: List[np.NDArray[np.int64]] = [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])\r\n\r\ndef build_dataset_from_tokenized(tokenized: List[np.NDArray[int]]) -> Dataset:\r\n \"\"\"FIXME\"\"\"\r\n dataset = EmptyDataset()\r\n for t in tokenized: dataset.append(t)\r\n return dataset\r\nds = build_dataset_from_tokenized(tokenized)\r\nassert (ds[0] == np.array([4,4,2])).all()\r\n```\r\n\r\n### What I tried\r\ngrep, google for \"add one entry at a time\", \"datasets.append\"\r\n\r\n### Current Code\r\nThis code achieves the same result but doesn't fit into the `add_item` abstraction.\r\n\r\n```python\r\n dataset = load_dataset('text', data_files={'train': 'train.txt'})\r\n tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base', max_length=4096)\r\n def tokenize_function(examples):\r\n ids = tokenizer(examples['text'], return_attention_mask=False)['input_ids']\r\n return {'input_ids': [x[1:] for x in ids]}\r\n ds = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=['text'], load_from_cache_file=not overwrite_cache)\r\n\tprint(ds['train'][0]) => np array\r\n```\r\n\r\nThanks in advance!\nHi @sshleifer :) \r\n\r\nWe don't have methods like `Dataset.add_batch` or `Dataset.add_entry\/add_item` yet.\r\nBut that's something we'll add pretty soon. Would an API that looks roughly like this help ? Do you have suggestions ?\r\n```python\r\nimport numpy as np\r\nfrom datasets import Dataset\r\n\r\ntokenized = [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])\r\n\r\n# API suggestion (not available yet)\r\nd = Dataset()\r\nfor input_ids in tokenized:\r\n d.add_item({\"input_ids\": input_ids})\r\n\r\nprint(d[0][\"input_ids\"])\r\n# [4, 4, 2]\r\n```\r\n\r\nCurrently you can define a dataset with what @albertvillanova suggest, or via a generator using dataset builders. It's also possible to [concatenate datasets](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?highlight=concatenate#datasets.concatenate_datasets).","embeddings":[-0.2646298409,0.1603154093,-0.0458905734,0.1386155337,-0.0252806228,0.1201700866,0.1569381356,0.1071401313,0.058670368,0.0440863669,0.1095488444,0.5823287368,-0.1573915929,0.138007164,0.4451274276,-0.1356466115,0.1021329612,0.2048072368,-0.1752964556,0.0704091564,-0.2645285428,-0.0905520767,-0.1519214809,-0.1790196747,-0.0579325259,-0.061322201,-0.4128628969,-0.3023573458,-0.2655974329,-0.7929652929,-0.0123855649,-0.0009123574,-0.0839308575,0.2624169588,-0.0001283345,0.0117334425,0.1670986712,-0.0539960302,-0.5640701652,-0.0456967317,-0.1553967893,-0.2387477756,0.1454959214,-0.3138717115,0.2260569036,-0.7180809379,0.3229591548,-0.1545696706,0.089084737,0.125013724,-0.0003758298,0.494810909,0.1991044879,-0.3639475703,0.2621561289,-0.0453420877,-0.1395283341,-0.1721628904,0.3763104081,-0.2068813294,0.0824214742,-0.1125964671,0.2658927441,-0.0915565714,0.4066393673,0.1262135953,-0.0127451709,-0.1426652074,-0.1565271169,0.2903876007,0.2264159918,-0.419044286,-0.6362015009,-0.4910083711,0.3017780781,-0.0399683453,-0.1348128319,0.060155388,-0.0711733475,-0.0760381594,-0.069809854,-0.3770206571,-0.111912109,0.2381709218,0.0717944801,0.4428234696,0.1425431371,0.1643607467,0.1441977918,-0.1271944046,-0.2345662117,0.0151866274,-0.1232940331,0.3333287835,-0.1031110138,-0.1166460291,0.1092310101,-0.0470127016,0.4755124152,-0.061299704,0.173726812,0.1157122627,-0.3675787151,0.0866340697,0.171476528,-0.1008807644,0.2137441039,-0.2119158655,-0.3128740191,0.0772754475,0.4092112184,0.0620396063,-0.1824602634,0.1073736846,-0.2618193328,-0.3582352996,-0.0681386217,0.4037789404,-0.1207616106,-0.0273630079,-0.0106998915,0.0841955394,-0.0472879857,0.173677966,0.1717312634,-0.3679118156,-0.2393834144,0.2014809549,0.1263076216,-0.0102789896,0.0816346034,0.217958957,-0.1080401167,0.1765950322,0.1008231193,-0.001051515,-0.014450158,-0.0649894327,0.0691936165,0.0783881024,-0.0114106443,-0.1804757416,0.2521993816,0.0106602665,0.0532554165,-0.0456522033,-0.1014854684,-0.4404917955,-0.4781494141,0.2198809236,-0.1078879684,-0.1812904775,-0.2244119793,-0.0641979277,0.0042391834,-0.1642895639,-0.1952681541,0.5241122246,0.3831294477,-0.2358407676,0.1302087307,-0.1283583939,-0.5635914207,-0.1791516542,0.1944267601,0.3399145603,-0.4468756914,-0.3274410367,-0.1016851589,-0.0158187244,-0.4132902026,0.1295030862,-0.0944460928,0.2800901532,-0.4338248074,0.4084133506,0.3091626167,-0.1658173501,-0.2207370549,0.1308985353,-0.1782537401,0.0376557037,0.0842344984,0.3958645165,0.2953098118,0.060019996,0.4510013461,0.2585290372,-0.0795216337,-0.0899351388,0.0736893639,-0.4146664441,0.3007451296,-0.0111934626,-0.1263205856,-0.0905314982,0.0279579051,-0.5407937765,0.1644920558,-0.1066171676,0.3368983269,-0.0383560583,0.4019224942,-0.0076305624,-0.1058239862,-0.5203376412,-0.3627278507,0.1384117454,-0.0773245841,0.1222194284,-0.1124037206,-0.4358099401,-0.1482965052,0.4267143309,-0.1347390115,-0.0214686058,-0.0691511258,0.0572068244,-0.0093465028,-0.1200825647,-0.4359641373,0.1428783089,0.2283416539,0.3076541424,-0.4393087029,-0.049837932,0.2686482966,-0.1849064827,-0.1217021868,0.1733721644,-0.141075328,-0.1443530619,0.208371371,0.2823854685,-0.0867410675,-0.27010414,0.1203915998,0.2604833245,0.2900448143,0.3060085773,-0.1001439542,-0.2057063431,0.3140789568,-0.2532829642,-0.2291707546,0.5499863029,0.0419301353,0.4261950552,0.1370570213,0.0392848514,0.0201952625,0.2511620224,-0.2537923455,-0.2578029633,-0.0499058366,0.1488565952,0.2123851925,-0.1196054667,-0.1045179814,0.1626659185,0.1422627568,0.1596186012,0.1770738661,0.1443223208,-0.2144739032,0.0623016506,0.0679695085,0.0149866864,0.4575430453,0.0544013269,0.1145165935,0.0659159645,0.111045748,0.0568909645,0.0589827783,0.031461373,-0.1170743406,0.0384728573,0.5389946699,0.0615468398,-0.0218392145,-0.1388209462,0.1136970893,-0.0432560146,-0.1218187734,0.0397563912,0.0448243991,0.0722149983,-0.2530289888,-0.3716776073,-0.0849188417,-0.0843767971,0.2689409852,0.0314313322,0.0156413484,0.301081568,0.137987569,0.7158766985,0.1047442034,-0.4731730223,-0.0859230235,0.0065621044,0.2182613462,-0.101071611,-0.1031731069,0.3030089736,0.402112186,0.0869909003,-0.0316570625,-0.3475198448,-0.5039143562,0.1095275655,-0.051787585,0.4337313473,0.3948709071,-0.335601896,-0.1123425141,-0.2517050803,0.1836480349,0.0373475961,-0.1394076496,-0.1860643178,-0.0976857767,-0.2373919785,-0.0496559888,0.1030326337,-0.206016019,-0.1766934991,0.4733741581,0.1536125243,0.2519126236,-0.2228932679,0.0982955918,0.0336947665,-0.1396936774,-0.1208668947,0.0718317702,-0.0042209281,0.4730607271,-0.3619567752,-0.1762785763,-0.1606230736,-0.5109411478,0.1114847213,0.377494216,-0.2879175246,-0.3993910253,-0.2613127232,0.3807624578,-0.0936847404,0.2705568075,0.4955573678,0.0041033556,0.1533956528,0.0573997162,-0.287943095,0.374073267,0.1856687218,0.0599000379,0.0809817091,0.1704524159,0.0867763981,0.663498342,0.1845435947,-0.1084073484,0.1433631629,-0.5918194056,0.3627535999,-0.2168656737,-0.2220130712,-0.1796467155,-0.1106616184,0.0941357389,0.0607794598,-0.0263928659,0.1260304004,0.0712459236,0.3468834758,0.0541964993,-0.1985840499,0.1793527603,0.0210639574,0.3373360038,-0.0193252079,-0.0224008486,-0.3871428668,0.017365003,-0.0409920514,0.2764789164,0.2389587611,-0.1770690531,-0.4978469908,-0.140987426,-0.307357192,0.0694869161,0.3090179861,0.4473903477,0.2480776906,-0.1597179025,0.0432345122,-0.2020529509,0.7236328125,0.325334698,-0.4594327211,0.0084282393,0.0938177109,-0.2563486099,0.173294425,0.0139204469,0.5422018766,0.1814727485,0.2524054945,-0.4542917013,0.2550082803,0.2051475048,0.179388687,-0.181665659,-0.1778385043,-0.1138078868,-0.1289127022,-0.2315636277,0.1871423721,0.1932213008,0.2209282368,0.4691876769,-0.0989606008,-0.4084217846,0.0668774918,0.1020533815,-0.0333174728,0.0044961032,0.1497972906,0.5072354674,-0.1110389158,-0.2078583688,0.4400340915,0.4042310417,-0.1330380738,-0.4836511612,-0.1611758918,-0.2042241693,0.4172766507,0.2799086571,0.1667552888,0.2738693357,0.048915837,0.167960912,-0.2383995652,0.2513348758,0.0925280228,-0.1277251542,-0.358201474,-0.3206029534,0.4401080608,-0.0505429246,-0.0510863662,0.0931559503,0.1860111803,-0.3132160604,0.3559414148,-0.0383071564,0.8532413244,0.0140183745,0.4227761626,0.2781120539,-0.0398610048,0.3147439659,-0.0383674316,0.211047858,-0.4547417462,-0.213548556,-0.0393988378,-0.1463226527,0.160040319,0.2031617761,-0.1510424614,0.2056574374,0.2487069666,0.0024850443,-0.1176300347,0.2441386729,-0.2815453708,-0.4994777739,-0.3187270463,-0.078942515,0.1857927144,-0.0679338351,-0.1185278744,-0.0023444057,-0.0017808152,-0.2800907791,-0.3747290373,-0.3991499245,-0.3093171418,-0.1589401513,0.6127160788,-0.264888227,0.2600618899,-0.0347001292,-0.0342629179,0.1709604114,0.1167391017,0.0811085179,-0.0193566158,-0.1360162348,-0.0407432057,-0.390902698,0.1759610772,0.2466171384,0.0166306775,0.2744945884,-0.1171826869,-0.2790036798,-0.2106477618,0.2183814049,0.2384328395,-0.2930262983,-0.2510787547,0.1993103027,0.1179536805,-0.2569961548,-0.1151394546,0.1747874022,-0.2499647886,0.1552372724,0.0525324047,-0.0369467661,-0.0058731325,0.2975792289,-0.1896295846,-0.2781789303,0.3539299667,0.142874673,-0.0727733895,-0.0792888477,-0.0266168173,0.4377448261,-0.2136408538,-0.0719576329,0.0547067747,-0.0483799987,-0.2827552557,0.0705747306,0.0742127374,0.1839034557,-0.105127506,-0.2911681533,-0.4044686854,0.2763894796,0.3002138138,0.3721829951,0.1848497391,0.1702437699,-0.252456069,0.4358135462,-0.1169064119,0.1479174644,0.1459783763,0.150271371,0.1854646951,0.1160024181,-0.0826214179,0.1280495822,-0.0782819465,0.1550616622,-0.1440820694,0.0151926642,-0.1747512966,0.2127382159,0.2331584394,-0.0112262378,-0.1478844434,-0.1146837622,-0.176208809,-0.3947841823,0.3163344562,-0.0946518704,-0.2700203061,-0.1996149123,0.6224370599,0.2577295005,-0.0989504829,0.5540859103,0.2836504579,0.4706646502,0.1808962822,0.2874778509,-0.352455318,0.0764467418,0.3385862112,0.4616112411,0.3684318066,-0.0464078262,0.0408415087,-0.206032604,-0.0081247808,0.1402797848,0.2286501378,0.5841594338,-0.2424769998,0.0834469348,0.338927418,0.0132566243,-0.1886738092,0.0095983101,0.2829734981,-0.1192660555,0.2885172665,0.4438749254,-0.1419064254,0.1382748783,-0.156646654,0.0861744583,0.0100405058,0.0146586439,-0.1547353566,-0.0930767283,-0.0351524539,-0.1431533396,0.1480107158,0.3850335181,0.3257227838,0.5992687345,-0.2377139628,0.4989932179,0.5984037519,0.006602589,0.1962082088,0.0195396338,0.0539919026,0.5260339975,0.6944862008,0.1154351905,0.3511893451,0.4129686952,-0.0707862154,-0.0713928416,0.0020068907,0.1799687743,0.1909280419,-0.3396001756,0.1210675761,-0.1997504383,-0.2255669534,-0.0936933458,-0.0438627005,-0.3908543885,0.0556977428,-0.0328667648,0.3594781458,-0.4826160967,0.0741701871,0.2190117836,0.0676174909,-0.3067223728,0.1683755219,0.0193257648,0.1612189412,0.0985937864,0.0885514244,0.1539871991,0.2236357629,-0.0412270315,0.3775160313,-0.5520080328,0.0923720151,-0.0642093644,-0.0626926571,-0.0128712729,-0.050201308,0.2981148064,-0.0703727081,0.0275407471,0.1684700698,0.2442592531,-0.1106707528,-0.58131814,-0.0054786503,0.0077514676,-0.276887387,-0.0820569247,-0.014521867,-0.0441375114,-0.0687596053,0.2984167933,-0.3923063278,0.0352506787,0.1657793969,-0.014820463,-0.1587476432,0.163619712,0.5065175295,0.2900012434,-0.0710300431,-0.2468840182,-0.5102574825,-0.0098479483,-0.1119040251,-0.1957143694,-0.0556481071,0.1179482937,0.0269699115,0.2001409233,0.161121726,-0.009064951,-0.0340892337,0.3953381181,-0.178308934,-0.0462915748,-0.0857851058,0.1022696123,0.1350229979,-0.2631911635,-0.1652940661,-0.0707383081,-0.1621331573,-0.1025682688,-0.0092492159,-0.0537614636,-0.124688983,0.5808807611,0.2099566162,0.1434181631,-0.1043718234,-0.0140569117,-0.2172804922,-0.1125323474,-0.3362767696,0.60516572,0.0282509923,0.0629272088,-0.1458367407,0.0355633162,0.0187913217,0.0046900376,0.0892241299,-0.1565242857,-0.1970908344,0.3591445684,-0.396266222,0.1175956801,-0.1363485456,0.3045372665,0.2100229561,0.2169963717,0.1847030967,-0.0915804729,0.2319037169,-0.454172641,-0.323569566,-0.0779609904,0.0218928661,0.0010353213,-0.3787177801,-0.5221460462,-0.287507683,0.3396661878,-0.0111592943,-0.0437442325,0.0462231226,-0.1847174615,-0.1237373427,-0.4769803584,0.4785841107,-0.0198921319,-0.0834573656,-0.0986276343,-0.5801851153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1854","title":"Feature Request: Dataset.add_item","comments":"Your API looks perfect @lhoestq, thanks!","body":"I'm trying to integrate `huggingface\/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L318, as opposed to loading all the text into arrow, and then `dataset.map(binarizer)`.\r\nIs this possible at the moment? Is there an example? I'm happy to use raw `pa.Table` but not sure whether it will support uneven length entries.\r\n\r\n### Desired API\r\n\r\n```python\r\nimport numpy as np\r\ntokenized: List[np.NDArray[np.int64]] = [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])\r\n\r\ndef build_dataset_from_tokenized(tokenized: List[np.NDArray[int]]) -> Dataset:\r\n \"\"\"FIXME\"\"\"\r\n dataset = EmptyDataset()\r\n for t in tokenized: dataset.append(t)\r\n return dataset\r\nds = build_dataset_from_tokenized(tokenized)\r\nassert (ds[0] == np.array([4,4,2])).all()\r\n```\r\n\r\n### What I tried\r\ngrep, google for \"add one entry at a time\", \"datasets.append\"\r\n\r\n### Current Code\r\nThis code achieves the same result but doesn't fit into the `add_item` abstraction.\r\n\r\n```python\r\n dataset = load_dataset('text', data_files={'train': 'train.txt'})\r\n tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base', max_length=4096)\r\n def tokenize_function(examples):\r\n ids = tokenizer(examples['text'], return_attention_mask=False)['input_ids']\r\n return {'input_ids': [x[1:] for x in ids]}\r\n ds = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=['text'], load_from_cache_file=not overwrite_cache)\r\n\tprint(ds['train'][0]) => np array\r\n```\r\n\r\nThanks in advance!","comment_length":6,"text":"Feature Request: Dataset.add_item\nI'm trying to integrate `huggingface\/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https:\/\/github.com\/pytorch\/fairseq\/blob\/master\/fairseq\/data\/indexed_dataset.py#L318, as opposed to loading all the text into arrow, and then `dataset.map(binarizer)`.\r\nIs this possible at the moment? Is there an example? I'm happy to use raw `pa.Table` but not sure whether it will support uneven length entries.\r\n\r\n### Desired API\r\n\r\n```python\r\nimport numpy as np\r\ntokenized: List[np.NDArray[np.int64]] = [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])\r\n\r\ndef build_dataset_from_tokenized(tokenized: List[np.NDArray[int]]) -> Dataset:\r\n \"\"\"FIXME\"\"\"\r\n dataset = EmptyDataset()\r\n for t in tokenized: dataset.append(t)\r\n return dataset\r\nds = build_dataset_from_tokenized(tokenized)\r\nassert (ds[0] == np.array([4,4,2])).all()\r\n```\r\n\r\n### What I tried\r\ngrep, google for \"add one entry at a time\", \"datasets.append\"\r\n\r\n### Current Code\r\nThis code achieves the same result but doesn't fit into the `add_item` abstraction.\r\n\r\n```python\r\n dataset = load_dataset('text', data_files={'train': 'train.txt'})\r\n tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base', max_length=4096)\r\n def tokenize_function(examples):\r\n ids = tokenizer(examples['text'], return_attention_mask=False)['input_ids']\r\n return {'input_ids': [x[1:] for x in ids]}\r\n ds = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=['text'], load_from_cache_file=not overwrite_cache)\r\n\tprint(ds['train'][0]) => np array\r\n```\r\n\r\nThanks in advance!\nYour API looks perfect @lhoestq, thanks!","embeddings":[-0.2646298409,0.1603154093,-0.0458905734,0.1386155337,-0.0252806228,0.1201700866,0.1569381356,0.1071401313,0.058670368,0.0440863669,0.1095488444,0.5823287368,-0.1573915929,0.138007164,0.4451274276,-0.1356466115,0.1021329612,0.2048072368,-0.1752964556,0.0704091564,-0.2645285428,-0.0905520767,-0.1519214809,-0.1790196747,-0.0579325259,-0.061322201,-0.4128628969,-0.3023573458,-0.2655974329,-0.7929652929,-0.0123855649,-0.0009123574,-0.0839308575,0.2624169588,-0.0001283345,0.0117334425,0.1670986712,-0.0539960302,-0.5640701652,-0.0456967317,-0.1553967893,-0.2387477756,0.1454959214,-0.3138717115,0.2260569036,-0.7180809379,0.3229591548,-0.1545696706,0.089084737,0.125013724,-0.0003758298,0.494810909,0.1991044879,-0.3639475703,0.2621561289,-0.0453420877,-0.1395283341,-0.1721628904,0.3763104081,-0.2068813294,0.0824214742,-0.1125964671,0.2658927441,-0.0915565714,0.4066393673,0.1262135953,-0.0127451709,-0.1426652074,-0.1565271169,0.2903876007,0.2264159918,-0.419044286,-0.6362015009,-0.4910083711,0.3017780781,-0.0399683453,-0.1348128319,0.060155388,-0.0711733475,-0.0760381594,-0.069809854,-0.3770206571,-0.111912109,0.2381709218,0.0717944801,0.4428234696,0.1425431371,0.1643607467,0.1441977918,-0.1271944046,-0.2345662117,0.0151866274,-0.1232940331,0.3333287835,-0.1031110138,-0.1166460291,0.1092310101,-0.0470127016,0.4755124152,-0.061299704,0.173726812,0.1157122627,-0.3675787151,0.0866340697,0.171476528,-0.1008807644,0.2137441039,-0.2119158655,-0.3128740191,0.0772754475,0.4092112184,0.0620396063,-0.1824602634,0.1073736846,-0.2618193328,-0.3582352996,-0.0681386217,0.4037789404,-0.1207616106,-0.0273630079,-0.0106998915,0.0841955394,-0.0472879857,0.173677966,0.1717312634,-0.3679118156,-0.2393834144,0.2014809549,0.1263076216,-0.0102789896,0.0816346034,0.217958957,-0.1080401167,0.1765950322,0.1008231193,-0.001051515,-0.014450158,-0.0649894327,0.0691936165,0.0783881024,-0.0114106443,-0.1804757416,0.2521993816,0.0106602665,0.0532554165,-0.0456522033,-0.1014854684,-0.4404917955,-0.4781494141,0.2198809236,-0.1078879684,-0.1812904775,-0.2244119793,-0.0641979277,0.0042391834,-0.1642895639,-0.1952681541,0.5241122246,0.3831294477,-0.2358407676,0.1302087307,-0.1283583939,-0.5635914207,-0.1791516542,0.1944267601,0.3399145603,-0.4468756914,-0.3274410367,-0.1016851589,-0.0158187244,-0.4132902026,0.1295030862,-0.0944460928,0.2800901532,-0.4338248074,0.4084133506,0.3091626167,-0.1658173501,-0.2207370549,0.1308985353,-0.1782537401,0.0376557037,0.0842344984,0.3958645165,0.2953098118,0.060019996,0.4510013461,0.2585290372,-0.0795216337,-0.0899351388,0.0736893639,-0.4146664441,0.3007451296,-0.0111934626,-0.1263205856,-0.0905314982,0.0279579051,-0.5407937765,0.1644920558,-0.1066171676,0.3368983269,-0.0383560583,0.4019224942,-0.0076305624,-0.1058239862,-0.5203376412,-0.3627278507,0.1384117454,-0.0773245841,0.1222194284,-0.1124037206,-0.4358099401,-0.1482965052,0.4267143309,-0.1347390115,-0.0214686058,-0.0691511258,0.0572068244,-0.0093465028,-0.1200825647,-0.4359641373,0.1428783089,0.2283416539,0.3076541424,-0.4393087029,-0.049837932,0.2686482966,-0.1849064827,-0.1217021868,0.1733721644,-0.141075328,-0.1443530619,0.208371371,0.2823854685,-0.0867410675,-0.27010414,0.1203915998,0.2604833245,0.2900448143,0.3060085773,-0.1001439542,-0.2057063431,0.3140789568,-0.2532829642,-0.2291707546,0.5499863029,0.0419301353,0.4261950552,0.1370570213,0.0392848514,0.0201952625,0.2511620224,-0.2537923455,-0.2578029633,-0.0499058366,0.1488565952,0.2123851925,-0.1196054667,-0.1045179814,0.1626659185,0.1422627568,0.1596186012,0.1770738661,0.1443223208,-0.2144739032,0.0623016506,0.0679695085,0.0149866864,0.4575430453,0.0544013269,0.1145165935,0.0659159645,0.111045748,0.0568909645,0.0589827783,0.031461373,-0.1170743406,0.0384728573,0.5389946699,0.0615468398,-0.0218392145,-0.1388209462,0.1136970893,-0.0432560146,-0.1218187734,0.0397563912,0.0448243991,0.0722149983,-0.2530289888,-0.3716776073,-0.0849188417,-0.0843767971,0.2689409852,0.0314313322,0.0156413484,0.301081568,0.137987569,0.7158766985,0.1047442034,-0.4731730223,-0.0859230235,0.0065621044,0.2182613462,-0.101071611,-0.1031731069,0.3030089736,0.402112186,0.0869909003,-0.0316570625,-0.3475198448,-0.5039143562,0.1095275655,-0.051787585,0.4337313473,0.3948709071,-0.335601896,-0.1123425141,-0.2517050803,0.1836480349,0.0373475961,-0.1394076496,-0.1860643178,-0.0976857767,-0.2373919785,-0.0496559888,0.1030326337,-0.206016019,-0.1766934991,0.4733741581,0.1536125243,0.2519126236,-0.2228932679,0.0982955918,0.0336947665,-0.1396936774,-0.1208668947,0.0718317702,-0.0042209281,0.4730607271,-0.3619567752,-0.1762785763,-0.1606230736,-0.5109411478,0.1114847213,0.377494216,-0.2879175246,-0.3993910253,-0.2613127232,0.3807624578,-0.0936847404,0.2705568075,0.4955573678,0.0041033556,0.1533956528,0.0573997162,-0.287943095,0.374073267,0.1856687218,0.0599000379,0.0809817091,0.1704524159,0.0867763981,0.663498342,0.1845435947,-0.1084073484,0.1433631629,-0.5918194056,0.3627535999,-0.2168656737,-0.2220130712,-0.1796467155,-0.1106616184,0.0941357389,0.0607794598,-0.0263928659,0.1260304004,0.0712459236,0.3468834758,0.0541964993,-0.1985840499,0.1793527603,0.0210639574,0.3373360038,-0.0193252079,-0.0224008486,-0.3871428668,0.017365003,-0.0409920514,0.2764789164,0.2389587611,-0.1770690531,-0.4978469908,-0.140987426,-0.307357192,0.0694869161,0.3090179861,0.4473903477,0.2480776906,-0.1597179025,0.0432345122,-0.2020529509,0.7236328125,0.325334698,-0.4594327211,0.0084282393,0.0938177109,-0.2563486099,0.173294425,0.0139204469,0.5422018766,0.1814727485,0.2524054945,-0.4542917013,0.2550082803,0.2051475048,0.179388687,-0.181665659,-0.1778385043,-0.1138078868,-0.1289127022,-0.2315636277,0.1871423721,0.1932213008,0.2209282368,0.4691876769,-0.0989606008,-0.4084217846,0.0668774918,0.1020533815,-0.0333174728,0.0044961032,0.1497972906,0.5072354674,-0.1110389158,-0.2078583688,0.4400340915,0.4042310417,-0.1330380738,-0.4836511612,-0.1611758918,-0.2042241693,0.4172766507,0.2799086571,0.1667552888,0.2738693357,0.048915837,0.167960912,-0.2383995652,0.2513348758,0.0925280228,-0.1277251542,-0.358201474,-0.3206029534,0.4401080608,-0.0505429246,-0.0510863662,0.0931559503,0.1860111803,-0.3132160604,0.3559414148,-0.0383071564,0.8532413244,0.0140183745,0.4227761626,0.2781120539,-0.0398610048,0.3147439659,-0.0383674316,0.211047858,-0.4547417462,-0.213548556,-0.0393988378,-0.1463226527,0.160040319,0.2031617761,-0.1510424614,0.2056574374,0.2487069666,0.0024850443,-0.1176300347,0.2441386729,-0.2815453708,-0.4994777739,-0.3187270463,-0.078942515,0.1857927144,-0.0679338351,-0.1185278744,-0.0023444057,-0.0017808152,-0.2800907791,-0.3747290373,-0.3991499245,-0.3093171418,-0.1589401513,0.6127160788,-0.264888227,0.2600618899,-0.0347001292,-0.0342629179,0.1709604114,0.1167391017,0.0811085179,-0.0193566158,-0.1360162348,-0.0407432057,-0.390902698,0.1759610772,0.2466171384,0.0166306775,0.2744945884,-0.1171826869,-0.2790036798,-0.2106477618,0.2183814049,0.2384328395,-0.2930262983,-0.2510787547,0.1993103027,0.1179536805,-0.2569961548,-0.1151394546,0.1747874022,-0.2499647886,0.1552372724,0.0525324047,-0.0369467661,-0.0058731325,0.2975792289,-0.1896295846,-0.2781789303,0.3539299667,0.142874673,-0.0727733895,-0.0792888477,-0.0266168173,0.4377448261,-0.2136408538,-0.0719576329,0.0547067747,-0.0483799987,-0.2827552557,0.0705747306,0.0742127374,0.1839034557,-0.105127506,-0.2911681533,-0.4044686854,0.2763894796,0.3002138138,0.3721829951,0.1848497391,0.1702437699,-0.252456069,0.4358135462,-0.1169064119,0.1479174644,0.1459783763,0.150271371,0.1854646951,0.1160024181,-0.0826214179,0.1280495822,-0.0782819465,0.1550616622,-0.1440820694,0.0151926642,-0.1747512966,0.2127382159,0.2331584394,-0.0112262378,-0.1478844434,-0.1146837622,-0.176208809,-0.3947841823,0.3163344562,-0.0946518704,-0.2700203061,-0.1996149123,0.6224370599,0.2577295005,-0.0989504829,0.5540859103,0.2836504579,0.4706646502,0.1808962822,0.2874778509,-0.352455318,0.0764467418,0.3385862112,0.4616112411,0.3684318066,-0.0464078262,0.0408415087,-0.206032604,-0.0081247808,0.1402797848,0.2286501378,0.5841594338,-0.2424769998,0.0834469348,0.338927418,0.0132566243,-0.1886738092,0.0095983101,0.2829734981,-0.1192660555,0.2885172665,0.4438749254,-0.1419064254,0.1382748783,-0.156646654,0.0861744583,0.0100405058,0.0146586439,-0.1547353566,-0.0930767283,-0.0351524539,-0.1431533396,0.1480107158,0.3850335181,0.3257227838,0.5992687345,-0.2377139628,0.4989932179,0.5984037519,0.006602589,0.1962082088,0.0195396338,0.0539919026,0.5260339975,0.6944862008,0.1154351905,0.3511893451,0.4129686952,-0.0707862154,-0.0713928416,0.0020068907,0.1799687743,0.1909280419,-0.3396001756,0.1210675761,-0.1997504383,-0.2255669534,-0.0936933458,-0.0438627005,-0.3908543885,0.0556977428,-0.0328667648,0.3594781458,-0.4826160967,0.0741701871,0.2190117836,0.0676174909,-0.3067223728,0.1683755219,0.0193257648,0.1612189412,0.0985937864,0.0885514244,0.1539871991,0.2236357629,-0.0412270315,0.3775160313,-0.5520080328,0.0923720151,-0.0642093644,-0.0626926571,-0.0128712729,-0.050201308,0.2981148064,-0.0703727081,0.0275407471,0.1684700698,0.2442592531,-0.1106707528,-0.58131814,-0.0054786503,0.0077514676,-0.276887387,-0.0820569247,-0.014521867,-0.0441375114,-0.0687596053,0.2984167933,-0.3923063278,0.0352506787,0.1657793969,-0.014820463,-0.1587476432,0.163619712,0.5065175295,0.2900012434,-0.0710300431,-0.2468840182,-0.5102574825,-0.0098479483,-0.1119040251,-0.1957143694,-0.0556481071,0.1179482937,0.0269699115,0.2001409233,0.161121726,-0.009064951,-0.0340892337,0.3953381181,-0.178308934,-0.0462915748,-0.0857851058,0.1022696123,0.1350229979,-0.2631911635,-0.1652940661,-0.0707383081,-0.1621331573,-0.1025682688,-0.0092492159,-0.0537614636,-0.124688983,0.5808807611,0.2099566162,0.1434181631,-0.1043718234,-0.0140569117,-0.2172804922,-0.1125323474,-0.3362767696,0.60516572,0.0282509923,0.0629272088,-0.1458367407,0.0355633162,0.0187913217,0.0046900376,0.0892241299,-0.1565242857,-0.1970908344,0.3591445684,-0.396266222,0.1175956801,-0.1363485456,0.3045372665,0.2100229561,0.2169963717,0.1847030967,-0.0915804729,0.2319037169,-0.454172641,-0.323569566,-0.0779609904,0.0218928661,0.0010353213,-0.3787177801,-0.5221460462,-0.287507683,0.3396661878,-0.0111592943,-0.0437442325,0.0462231226,-0.1847174615,-0.1237373427,-0.4769803584,0.4785841107,-0.0198921319,-0.0834573656,-0.0986276343,-0.5801851153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1849","title":"Add TIMIT","comments":"@patrickvonplaten Could you please help me with how the output text has to be represented in the data? TIMIT has Words, Phonemes and texts. Also has lot on info on the speaker and the dialect. Could you please help me? An example of how to arrange it would be super helpful!\r\n\r\n","body":"## Adding a Dataset\r\n- **Name:** *TIMIT*\r\n- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*\r\n\r\n- **Paper:** *Homepage*: http:\/\/groups.inf.ed.ac.uk\/ami\/corpus\/ \/ *Wikipedia*: https:\/\/en.wikipedia.org\/wiki\/TIMIT\r\n- **Data:** *https:\/\/deepai.org\/dataset\/timit*\r\n- **Motivation:** Important speech dataset\r\n\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":51,"text":"Add TIMIT\n## Adding a Dataset\r\n- **Name:** *TIMIT*\r\n- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*\r\n\r\n- **Paper:** *Homepage*: http:\/\/groups.inf.ed.ac.uk\/ami\/corpus\/ \/ *Wikipedia*: https:\/\/en.wikipedia.org\/wiki\/TIMIT\r\n- **Data:** *https:\/\/deepai.org\/dataset\/timit*\r\n- **Motivation:** Important speech dataset\r\n\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n@patrickvonplaten Could you please help me with how the output text has to be represented in the data? TIMIT has Words, Phonemes and texts. Also has lot on info on the speaker and the dialect. Could you please help me? An example of how to arrange it would be super helpful!\r\n\r\n","embeddings":[0.0733412355,-0.3110238612,-0.1762830466,0.1887156516,0.0858767852,-0.1072419658,0.1176897883,0.0632870495,-0.4261901975,-0.1358919293,-0.2996531427,0.1368214786,-0.0700436831,0.3138222992,0.2315153927,-0.3812035322,-0.0924161598,0.1239683703,-0.2485858947,-0.1681149155,-0.0206748378,0.2014280111,0.0602086633,-0.017016774,-0.0788321272,-0.1857829541,-0.1064575762,-0.19129695,-0.2009379417,-0.4111648798,-0.1293073744,0.0393785276,0.0996622592,0.521894455,-0.0001090629,-0.1116432026,-0.0847964659,-0.2752177417,-0.048206795,-0.1977025419,0.0026013448,0.0172410663,-0.1876087934,0.0332623161,-0.1161857545,0.0043508285,0.1767341495,-0.2566524446,0.152622968,0.2560269833,0.1796585619,-0.2397240102,0.0037524428,0.120319806,0.1837881505,0.3532900214,-0.1659121662,-0.1861263961,0.442364037,0.3691049516,0.154524371,0.2054741234,-0.2338324636,-0.1602454036,0.1172219515,0.1588024646,-0.1280131787,-0.4249416292,-0.2001318932,0.1614860892,1.0803807974,-0.2730861008,-0.2102272958,-0.1682152897,-0.0067150947,-0.0388446786,0.0545272641,0.0947851166,-0.1075404435,0.2822042704,-0.1284112781,-0.0873562098,-0.4036233127,0.2998380661,0.1751087606,-0.0903957784,-0.1651895642,-0.0817210227,-0.2652519941,-0.0224783812,-0.1265884787,-0.0691512898,0.1829085946,0.1443703324,0.1264306903,-0.2070368081,-0.0722235516,-0.2396785468,0.1297903657,0.0799331218,0.1578906029,0.1282496452,-0.3613839447,-0.035532821,0.3212117255,-0.1483234763,0.1604641378,-0.3020412922,0.0304557309,-0.3613665104,0.1621885449,-0.2847991586,-0.2428912669,0.2361746877,0.1377900392,-0.0467887186,0.1481570601,-0.2279863954,0.0026647793,0.0527142324,-0.0180348344,0.0589291118,-0.0303587299,0.0837378055,-0.1445214301,0.1464492083,0.1182098612,0.2381437719,-0.0150136463,-0.1115037352,0.0840686038,-0.0457720794,-0.2555977702,0.1817157567,0.0520271473,-0.1000453383,0.1685810685,0.1960580498,-0.011818029,-0.0784360245,-0.1366565526,-0.1784195602,0.2777072787,-0.1461721361,0.0527808405,0.1390263438,-0.1903186888,-0.0470301062,-0.3633146882,0.1211081743,0.1679523736,-0.0422931872,-0.0453132279,0.1867241114,-0.0124824429,-0.156264171,-0.1769187301,0.711476624,0.2749207914,-0.1027399525,-0.0075083468,0.1939911097,-0.5860921741,-0.105024077,0.1776429713,0.0585591421,-0.6696823835,0.0212520659,-0.0378686897,0.1611167639,0.2975309491,0.2125638127,0.3395764232,0.586378336,0.0210056063,0.2667877078,0.171755597,-0.2706251442,-0.1431101263,0.1369380355,-0.1598664969,-0.1170849726,-0.0473485291,0.1449374706,0.4792860448,-0.101877138,0.0294850636,0.1940071881,-0.2014424205,-0.0235529151,-0.0290610548,-0.0371894427,-0.1204591691,0.0778586268,-0.1989880204,-0.106796056,-0.1251796186,0.217751056,0.2246510237,0.0652947649,0.3470469713,0.069211565,0.3847998381,0.1909002364,0.129735589,-0.4917135239,0.4033924043,-0.1964217871,0.1526651829,0.4782302082,-0.1473187506,-0.167142421,-0.3596304655,-0.2206088901,-0.0516645722,0.2606676221,0.2205816954,0.1991176605,-0.009456072,0.0857204795,0.0355820432,0.4501103759,-0.0075565879,-0.2419346571,-0.3201273382,0.2504458129,0.0951381922,-0.3446833789,-0.014947948,0.4989426136,-0.0831925943,0.1493805498,0.1380867064,0.2466288507,0.0038797308,0.0135975964,-0.2022036761,-0.1101445854,0.2628404498,-0.4228825867,0.3313645124,0.06269072,0.0637856349,-0.2258459479,0.0393554047,-0.0205577482,0.1559752524,0.0706196427,-0.1424725205,-0.1016159207,-0.1329055876,0.0496608876,-0.0569060482,-0.1106606722,0.0648706481,0.1193826869,-0.0562707148,-0.2600478232,-0.262647301,0.2423307896,0.2686450779,0.006195752,0.092106007,0.0466340072,-0.2050528377,0.2062917203,-0.0478995517,-0.4742149115,-0.0054341042,0.2985425293,0.193499878,-0.0386918783,0.1178182811,-0.0000569245,0.0734971911,0.144157365,-0.1406797022,0.2691238225,-0.168408528,0.0144770807,-0.2006818354,-0.0646961257,-0.2875880897,0.0392232202,-0.1764378697,-0.1992070526,0.0765008181,-0.1841205955,-0.2453136891,-0.3167244196,0.099512428,0.0610859059,0.2686392367,0.1480069607,0.0491264388,-0.0101225134,0.2124205083,0.1507262439,0.0232352819,-0.0001346469,-0.0215184875,-0.0320803262,-0.1561594307,0.1696841866,0.0760947093,0.3061435819,0.4528852105,-0.3062444031,0.0146126365,0.0802450404,-0.0178312529,0.2304171473,-0.3757922351,0.2834007144,0.0289359558,-0.0000253201,-0.3605325818,-0.4115389287,0.0161460415,-0.0315226577,-0.291928798,0.0743878484,-0.0956783667,-0.1340010762,0.1157216802,-0.4445347786,-0.2971717417,-0.3025036752,0.3352980316,0.0821444616,0.2249576449,0.1699913144,-0.1208080798,0.1753780097,0.1636454761,0.4341614246,-0.3202284276,-0.3624622822,0.3046229482,-0.2548819482,-0.297619313,-0.0871013105,0.0060070879,0.0260227639,0.0635265186,-0.2088310272,0.1419960558,-0.0100438697,0.0497111231,0.0847021714,0.1780128628,0.1583238989,0.0765187666,-0.0798770636,-0.2320663631,0.0328554511,0.1476105601,0.1669758856,0.1966291964,0.2978936434,0.0147548905,0.0068026502,0.3493381441,0.4952243865,0.0325859189,0.062058378,0.2822734118,0.0033729149,-0.0174730867,-0.3049688637,0.1944867671,0.0497228764,0.1094197631,0.6189506054,0.2590442002,-0.0316263661,-0.1296888888,-0.1573352218,-0.3830251396,-0.1344005615,0.2337817103,-0.2662259042,0.1417679936,-0.3733877838,-0.1500866711,-0.0214477573,-0.1725009829,0.0644913614,-0.0231989417,0.0695002154,0.0170307979,-0.4599190056,-0.1179009005,-0.4404782653,-0.1500291526,0.3699202538,0.066468522,-0.2545385361,0.0589631312,0.0166574195,0.4282952547,0.2871984243,-0.3595501781,-0.081750907,-0.0595481396,0.320838213,-0.2337534428,0.3633396924,-0.3790821135,0.0611296073,0.2409206927,-0.0161084142,-0.1946840286,0.1699624956,0.0684409589,0.3552654386,-0.1093612984,-0.3979184926,0.1107330546,-0.3221363127,-0.0377206802,0.0476120524,0.297573477,0.0108385431,0.2765029967,-0.3337193429,0.3182638288,0.2296828032,0.5717841387,-0.0792506784,0.1295107603,0.171101585,0.1002243012,0.4372742176,0.0810578018,0.0613370016,0.3355130851,-0.0084579596,-0.0838704258,0.0423019044,-0.6474964023,0.6383487582,0.1778749824,0.5054211617,0.0578814037,-0.098855339,-0.17916511,-0.3123470247,0.0829518661,0.2765512466,0.2287977636,-0.5406431556,-0.4827101231,0.3688135147,-0.050815016,-0.016092727,0.3932924271,0.0866099522,-0.5273903608,0.4000514746,0.13467893,1.2645546198,-0.0729820505,0.0805896819,0.0166744348,0.2139329016,0.1104693413,-0.3328994811,-0.1002239734,-0.1184967905,0.0039252243,-0.1840423346,0.1489253789,-0.1124997661,0.4466062784,-0.6558457613,-0.2342239022,0.1134621203,0.1801486164,0.0316668637,0.2309565097,-0.0875614807,-0.5443079472,-0.1068469211,0.1333287507,-0.0943770409,-0.1051522791,0.0779341981,-0.1407402307,-0.0990864858,0.1872478873,-0.5839387178,-0.1935067922,-0.2705624104,-0.253999263,-0.2642699778,-0.4549129903,0.4614437222,0.3034361601,0.471196264,0.2648222148,-0.0988334343,-0.2700282037,0.2565156221,-0.2440985292,0.1132158414,-0.0564184003,0.2361492813,-0.2565663159,-0.4020327628,0.1199065223,0.2340469509,0.0558138527,-0.1997070312,-0.2428251505,0.2773273885,-0.6060074568,0.0332642384,0.4039945304,0.0392997749,0.0561624058,0.159371078,0.3219446838,-0.2703963816,0.3919323087,-0.0728839412,-0.1171324477,-0.0917012095,-0.0140200332,0.3209598362,-0.095053874,0.2667022347,0.3177777529,-0.2066060007,-0.1901085377,0.1254203469,-0.0025458976,-0.0699186549,-0.1668092161,-0.0008290529,0.0364960693,-0.2243149281,0.2582896948,0.2697301209,-0.1312970966,-0.2389479578,-0.162515223,-0.4066056907,0.1496085823,0.2464058548,0.2030090094,0.0568583682,0.1091360152,0.105846867,0.1297781467,-0.3844206929,-0.356223017,0.1625213325,0.3911123574,0.2898442745,-0.0993201435,0.052622173,-0.1396072954,0.1151624769,0.1162933856,-0.1275236607,-0.211836502,-0.1285014451,0.1377587616,0.1193389222,-0.2392579913,-0.0457863174,-0.1553720534,-0.205681473,-0.1629978567,-0.0004477899,0.4357185662,0.03952742,0.3720488548,0.206815511,0.1207290888,-0.1114281118,-0.1343018115,0.0886631384,0.0658335164,0.1747948825,-0.0144564221,0.0639540181,-0.1519779116,0.0520474464,0.3738812506,0.3549928665,-0.1232147589,0.1929378361,-0.1140261963,0.2656604648,0.3561338484,0.2540929615,0.1947013587,0.1525964588,-0.1367121339,0.1833355725,0.2633679807,0.0052772625,0.0747997686,-0.2822592854,-0.1367001981,0.0513610803,0.276914686,0.1008640975,-0.0852714032,0.2514654696,-0.073127225,0.1488074362,-0.0841932818,-0.0641893148,0.3689873219,0.1300589144,-0.0457837991,0.0508011207,0.2202302068,0.0176396407,0.0883082002,0.0042287083,0.0974259079,-0.1153716445,0.100322701,0.174871996,0.0657179132,-0.0259541273,0.2681545913,-0.1054071113,0.1885498017,0.205889672,0.6595245004,-0.0601865761,-0.1049587354,0.1661816537,-0.0065031312,-0.0552187227,0.0787562951,-0.6642002463,-0.2857768834,-0.0530471392,-0.4137738049,0.0484932736,-0.3519783914,0.0782280415,0.1137441546,0.0702374354,-0.4151626825,-0.0527815334,-0.2030508518,0.3201818168,-0.2529435456,0.23392497,0.2223710716,-0.0441343449,0.1609780341,0.3641623855,0.1534331143,-0.0533096939,-0.375795424,0.4693291783,-0.2915986478,-0.2523393631,0.0952130035,0.3439056873,0.0758944973,-0.0035801101,0.1196693778,0.1978463829,-0.0468992814,-0.1937182397,0.4577820301,-0.3645666242,0.1280644387,0.1383056343,0.1133919433,-0.3378146887,-0.1734227389,-0.3441180289,-0.4632585049,0.0062943385,0.7435656786,-0.2706799805,-0.0127983317,-0.047272969,0.1016563773,0.034843497,0.05373431,0.5934016705,0.2932862341,0.1423203796,-0.1733773947,-0.1467894614,0.108046487,0.1389687508,-0.3366042972,-0.0652379915,-0.0520298108,0.368201524,0.3634085655,-0.0564650483,-0.3060190678,-0.7328798771,0.239866659,-0.1863745451,0.0345147066,-0.0666829944,0.6858640909,-0.1334248036,0.0540283695,-0.2784759402,0.3473326564,0.076316759,-0.4032463431,0.1232481524,-0.0656646267,0.2084628195,-0.1524846256,-0.0060836873,0.4159908593,-0.3321259916,0.1745471507,0.1244741008,0.0556591824,-0.328928113,0.2753641009,0.165295884,0.3569642305,-0.2938200831,0.1285813451,0.0174262542,0.0891979635,0.125061363,0.0432401299,-0.2158611268,0.0552389175,-0.2938522995,-0.1764225513,0.0560261831,0.3817706704,-0.0473241583,-0.0257543102,-0.3988314271,-0.4108399451,0.005938177,0.0089297676,-0.4403369427,0.1504620165,-0.0303868577,-0.0181618165,0.2231646329,-0.7285835147,-0.0874039307,0.4551881254,0.020122109,0.0890050009,0.2369014472,0.0475302897,-0.157026872,-0.0915603861,0.4929834604,0.1033172384,-0.1523524225,-0.19843705,-0.5478033423]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1849","title":"Add TIMIT","comments":"Hey @vrindaprabhu - sure I'll help you :-) Could you open a first PR for TIMIT where you copy-paste more or less the `librispeech_asr` script: https:\/\/github.com\/huggingface\/datasets\/blob\/28be129db862ec89a87ac9349c64df6b6118aff4\/datasets\/librispeech_asr\/librispeech_asr.py#L93 (obviously replacing all the naming and links correctly...) and then you can list all possible outputs in the features dict: https:\/\/github.com\/huggingface\/datasets\/blob\/28be129db862ec89a87ac9349c64df6b6118aff4\/datasets\/librispeech_asr\/librispeech_asr.py#L104 (words, phonemes should probably be of kind `datasets.Sequence(datasets.Value(\"string\"))` and texts I think should be of type `\"text\": datasets.Value(\"string\")`.\r\n\r\nWhen you've opened a first PR, I think it'll be much easier for us to take a look together :-) ","body":"## Adding a Dataset\r\n- **Name:** *TIMIT*\r\n- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*\r\n\r\n- **Paper:** *Homepage*: http:\/\/groups.inf.ed.ac.uk\/ami\/corpus\/ \/ *Wikipedia*: https:\/\/en.wikipedia.org\/wiki\/TIMIT\r\n- **Data:** *https:\/\/deepai.org\/dataset\/timit*\r\n- **Motivation:** Important speech dataset\r\n\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":85,"text":"Add TIMIT\n## Adding a Dataset\r\n- **Name:** *TIMIT*\r\n- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*\r\n\r\n- **Paper:** *Homepage*: http:\/\/groups.inf.ed.ac.uk\/ami\/corpus\/ \/ *Wikipedia*: https:\/\/en.wikipedia.org\/wiki\/TIMIT\r\n- **Data:** *https:\/\/deepai.org\/dataset\/timit*\r\n- **Motivation:** Important speech dataset\r\n\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHey @vrindaprabhu - sure I'll help you :-) Could you open a first PR for TIMIT where you copy-paste more or less the `librispeech_asr` script: https:\/\/github.com\/huggingface\/datasets\/blob\/28be129db862ec89a87ac9349c64df6b6118aff4\/datasets\/librispeech_asr\/librispeech_asr.py#L93 (obviously replacing all the naming and links correctly...) and then you can list all possible outputs in the features dict: https:\/\/github.com\/huggingface\/datasets\/blob\/28be129db862ec89a87ac9349c64df6b6118aff4\/datasets\/librispeech_asr\/librispeech_asr.py#L104 (words, phonemes should probably be of kind `datasets.Sequence(datasets.Value(\"string\"))` and texts I think should be of type `\"text\": datasets.Value(\"string\")`.\r\n\r\nWhen you've opened a first PR, I think it'll be much easier for us to take a look together :-) ","embeddings":[-0.2446853518,-0.4731037915,-0.174794212,0.0511885472,0.1257324815,-0.0904694572,0.0721059144,0.2155632228,-0.2834814489,0.1218893826,-0.3020024598,0.1912916899,-0.3662675917,0.1645146161,0.1355866939,-0.0494615212,-0.1529108435,0.1166824773,-0.065175876,-0.1085638627,-0.0938299447,0.2661955655,-0.0076833549,-0.0831073225,0.0740496665,0.0012327221,0.0271210168,-0.2614464164,-0.2260967046,-0.4311261475,-0.2807672322,0.109782964,0.0420130379,0.2606547177,-0.0001012167,-0.0995493382,0.1022698656,-0.2563877106,-0.2364086807,-0.2771131694,0.0944201052,0.2249715328,-0.1020732373,0.0330961421,-0.3512912989,-0.0870842263,0.1736273617,-0.299061954,0.09423583,0.4483915865,0.2870048583,0.1087668538,0.0172399729,-0.0470734425,0.3513868749,0.0892287791,-0.1679859906,-0.1017498076,0.2633745968,0.171956107,-0.1154143438,0.2804053426,-0.073449783,-0.06494008,0.1699990928,0.0632101223,-0.1534417421,-0.4414430857,-0.1258067042,0.191982016,0.5023941398,-0.3778818548,-0.2496569008,-0.2343409508,0.0014096359,-0.238330707,0.0316988714,0.0852811933,-0.4321370423,0.244062975,0.0006114896,-0.1649410427,-0.1142207012,0.210340783,0.3562372625,-0.066118449,-0.203621164,-0.2457416803,0.0624934174,-0.112938568,-0.1985174119,0.0873564929,0.1447731853,0.1188490838,-0.0219811723,-0.1750058383,0.2242131233,-0.1459815949,0.2494655252,0.1811535507,0.2915926576,0.2477516234,-0.27515921,-0.008245795,0.0662302971,0.0162236616,0.2081102282,-0.3157474101,0.1748314649,0.0749867037,0.1083327532,-0.1047930419,-0.1158122271,0.3235115111,0.0545835011,-0.0948979482,0.1270539463,-0.1975926757,-0.1494171321,-0.0654081032,-0.0892636701,-0.0529341735,0.0261622295,0.3333387077,-0.0332881138,0.1628894508,0.187749818,0.1965328008,-0.1016970128,-0.1391807348,-0.0169424955,0.071937494,-0.0720686391,0.1417998672,0.2965699136,-0.1536196321,0.2145942152,0.0117725255,0.0531396791,0.0227478798,-0.0899613798,-0.0020375766,0.2373252958,-0.1218462363,0.0231860615,-0.0782553181,-0.2007280737,-0.0473286845,-0.2776868045,-0.1355700493,0.3456792831,-0.0079596573,-0.0829637572,0.3332653642,-0.0892810896,-0.1117862388,0.075203307,0.6522989869,0.1052544862,-0.1494070292,-0.0124878837,0.1218355894,-0.5117614865,-0.082250692,0.1544604897,0.1537688524,-0.178316623,-0.1735667288,0.1369705647,0.0502703488,0.0950465947,0.1664884537,0.1406133175,0.1440724134,-0.1438378245,0.3682314754,0.1132042259,-0.2234618962,-0.2475328743,-0.0302935541,-0.1698493361,-0.0815486088,0.2441520095,0.1658109128,0.3559544384,-0.0554690696,0.1300072819,0.1874440461,0.0038795331,-0.0674785972,-0.2416634411,-0.2077523917,-0.1962818354,0.2166894972,-0.0796581581,0.0269908104,0.010064492,0.2276171893,0.2825474143,-0.1527392417,0.3140565753,0.0815103352,0.3197703958,0.0782899559,0.177743569,-0.5899131894,0.2397315502,-0.2106185704,-0.0954130664,0.5726776123,-0.1458019912,-0.3300564289,-0.1602546126,-0.2373498529,0.0323942304,0.1230166852,0.3529720008,0.4523493946,0.0681819543,0.1242334172,-0.0714032575,0.2873777747,0.0565586686,-0.1308630705,-0.2722588181,0.0438408367,-0.0734725296,-0.2239449918,0.2974459827,0.3255224228,-0.1548393369,0.0911623687,0.1202835068,0.2255719453,-0.0497888327,0.0233637076,-0.1530899853,0.0925231427,0.1362103522,-0.4039678574,-0.0200870391,0.1621841192,0.0843086615,-0.0932024792,-0.005054215,0.1134380251,0.4156939387,0.0289969034,0.2748575211,0.099829495,0.0727254003,-0.0646615997,-0.2692393064,-0.0880861878,0.2396902442,-0.2024722546,0.0311079789,-0.1333856881,-0.4374273419,0.092172198,0.2956879735,-0.051810421,0.1979669631,0.1582334936,-0.2827548385,0.0848320946,-0.0343025438,-0.3078743517,-0.0230967458,0.3893951774,0.092301026,-0.0883150101,-0.0653998256,0.0106037864,0.0936963633,0.1997484565,-0.2894505262,0.2492717057,0.1273010671,-0.135072276,-0.5136486292,-0.1554001868,-0.2910377085,0.0527033992,-0.1608417481,-0.1769292504,0.0188021511,-0.2057586014,0.0059271045,-0.1677866578,0.0387004279,0.0547487624,0.3567626774,0.236237064,-0.1175415143,0.0963414088,-0.0778380185,0.1724716723,0.0335903168,-0.08172521,-0.1140225902,-0.0078570973,-0.0065495418,0.2402787209,0.1958111674,0.2084497064,0.4720792472,-0.0974733531,0.04280065,-0.0827950686,-0.1179530174,0.2676103711,-0.2731459439,0.1502684802,0.128431946,-0.2301835418,-0.1692804694,-0.5684183836,0.028790947,-0.1986536533,-0.381359607,0.1055131257,-0.0323807746,-0.0563612059,-0.1398995966,-0.5737683773,-0.1626630127,-0.5080012679,0.4297767282,0.3065857589,0.1278098673,0.2718521655,-0.1408876032,0.1621195078,0.0003225153,0.4808383882,-0.2825733721,-0.1073479354,0.2177516073,-0.3385890722,-0.2448087782,-0.3268170953,-0.1765284985,0.1740167886,0.0678549856,-0.2126875669,-0.0678556114,-0.0893665478,0.2051257044,0.0219392348,-0.0329811685,0.2971903682,0.2930394113,-0.2229768485,-0.1700256765,0.0129995337,0.2324623913,-0.0746984333,0.0605608374,0.2563892007,-0.045488134,-0.0539106168,0.3795776665,0.1761077791,-0.0381145962,0.356025815,-0.0658747479,0.1278821975,-0.0495062359,-0.2565574944,0.1938413978,0.0296881013,0.0429239981,0.5046966076,0.0830382332,-0.121720992,-0.1212025881,0.0329715796,-0.2757847607,-0.2715727389,0.0381218754,-0.1754383743,0.2203903943,-0.2052238733,0.0756849274,-0.0262293871,-0.1750739068,0.2910814583,-0.0916306749,0.0453099795,-0.1528000236,-0.3904526532,0.0952796564,-0.4769840539,0.1211924851,0.3720566332,0.1703338176,-0.1272700727,-0.3410014808,-0.1047666892,0.1463623345,0.0802302212,-0.3026004136,-0.1044278145,0.199041158,0.2179022282,-0.0960018411,0.3850281835,-0.4127274156,-0.1140462086,0.1458284408,0.1484453082,-0.3244846761,0.1359201372,0.1403809339,0.2401788235,0.0372512117,-0.0919630304,0.0676757842,-0.2784214914,-0.2120233029,0.0796412006,0.1409565359,0.0751523077,0.0609364621,0.0171629675,0.1617179513,0.015728863,0.6219562888,0.1262027323,0.2305989712,0.0345698148,0.3014844358,0.1798699349,0.067793712,-0.0350250751,0.3268337846,0.0359696001,-0.3196800351,-0.0307558328,-0.3156150579,0.3362610936,0.0105597591,0.1693540663,0.275796622,0.020734366,0.0480330773,-0.2243027836,0.2289400995,0.2475953251,0.1215527728,-0.2556460798,-0.3392496705,0.2284768522,-0.1506510824,-0.1627786458,0.1830177605,-0.0899857581,-0.3301520348,0.5516926646,-0.0942658186,1.0282052755,-0.0590444952,0.0768178627,0.1052366346,0.2488816381,0.1212231293,-0.381313622,0.1350329518,-0.0621702448,-0.1011077464,-0.0695617869,0.1310069114,-0.2805915475,0.2087641209,-0.6060715318,-0.126758039,0.2254345566,0.1357457936,0.046754241,0.4178176522,-0.210723117,-0.5592319369,-0.2329226732,0.2671326399,0.0596365444,0.1550032943,0.3905996084,-0.1615734547,0.0116558131,-0.1391179264,-0.4064032435,-0.0549853146,-0.3689361811,-0.2309956402,-0.1944556534,-0.2307686061,0.2400191426,0.2623232901,0.3207446635,0.3633094728,-0.2477014959,-0.191238448,0.2563922107,-0.0942576453,0.0220479444,0.0694200397,0.2842489779,-0.1292541772,-0.1618177891,0.257302314,0.2658286393,-0.0859095529,-0.1815224141,-0.2416186333,0.1066544726,-0.4168110192,0.1245778352,0.1638920605,-0.3157116473,-0.1993839741,0.244388774,0.1866870373,-0.371160984,0.1189567223,0.2077975869,-0.0500675403,-0.1599428505,0.222669512,0.2622888088,-0.169254154,0.3335668743,0.1494542062,-0.2238979489,-0.3878361583,-0.3032690585,-0.1698224545,-0.1835508496,0.0154229682,0.15174523,-0.211646229,-0.0663785189,0.2987136543,0.1045893654,0.064161174,-0.2635017931,0.0145297498,-0.5142236948,0.2843659222,0.163855359,0.3496056497,-0.0192259792,0.1884596646,-0.0024221181,-0.0450584218,-0.5011376143,-0.2810227573,-0.2862842679,0.2542143762,0.215879485,-0.2318916768,0.2132052332,0.0431461632,0.3190140426,0.1415691227,-0.0317833424,-0.2971872389,-0.1771234572,0.0787459165,-0.0148755638,0.0569592491,-0.0445689224,0.0414820053,-0.1327000558,-0.0937612578,-0.0420436226,0.3803985715,0.011144015,0.2373114824,0.1808613241,-0.0908333659,0.2514047623,-0.0949690714,-0.0125642074,0.1085036844,0.1166459993,0.0050611338,-0.0760420635,-0.1288555712,0.2272413224,0.324349761,0.3210984468,-0.2039889842,0.1382939667,0.0497858003,-0.0581773669,0.4174616933,0.3369735479,0.1587646306,-0.044992812,-0.0270832181,0.2871141434,0.3827195168,-0.2486952394,0.0387311913,-0.3137306273,-0.2704774141,0.370323956,0.3311015964,0.0026799527,-0.1017881781,0.3540240526,-0.0562111847,-0.0927354991,-0.173517853,-0.1017996892,0.2736458182,0.2307824194,-0.1702155322,0.0418854803,0.0315530449,0.1582623273,0.2494809777,-0.2579782009,0.187767297,-0.1976953894,0.1082634553,0.3382607698,0.1366483271,-0.0676263124,0.1859726906,0.0656733587,0.3725672662,0.0044227797,0.7724063396,0.0109552098,0.0456986129,0.0433558226,0.220591262,-0.1191807836,-0.1363265365,-0.2501583993,-0.2826465964,-0.1302423775,-0.2654426396,0.1874406487,-0.0811684504,0.1288659871,0.0495907664,0.0923413485,-0.2246034741,-0.1778470129,-0.1649048775,0.4222962558,-0.2740442455,0.2454617321,0.125713557,0.0839334801,0.2143175155,0.3574656248,0.2837608457,-0.0000713888,-0.3314853013,0.359774977,-0.3336437941,-0.1156687513,-0.206890896,0.3654175699,0.3015878499,0.1280875206,0.3135041595,0.2218193561,-0.2273353338,-0.0564452559,0.0746689886,-0.2550722361,-0.0650054216,0.1765339971,0.0582255274,-0.1135064438,-0.4047627449,-0.2626597285,-0.5181845427,0.1170379072,0.3340686262,-0.0781930089,0.0896390676,-0.1309967339,0.1565668881,0.01636618,0.1215924025,0.4205240309,0.2626699507,0.0787535682,-0.2565956414,-0.3723429739,0.0445839651,0.185798645,-0.1927731633,-0.1286540926,0.0697534755,0.4065481424,0.3313052356,-0.1891488582,-0.1789573431,-0.5281731486,0.2662967145,-0.2141922116,0.1076403558,-0.0734585896,0.2644220293,0.1340915859,-0.1748874336,-0.1715740114,0.1485577971,0.2026426792,-0.3142870367,0.1665398479,-0.1248771176,0.0549171567,0.0170081612,-0.0173098538,0.3602311611,-0.513923943,-0.109518379,-0.0154577326,-0.1967178881,-0.3190577328,0.0358889922,0.1754819751,0.4160326719,-0.0598506778,-0.0791895539,-0.0858869925,0.1916798651,0.2438005805,0.0317477547,-0.1619962007,0.1378461272,-0.2386995703,-0.0164068267,-0.0654043183,0.3601736426,-0.0269399527,-0.0483553261,-0.3428578079,-0.6533893347,0.3332822919,-0.1210155562,-0.2877959311,0.0354304835,0.2070018798,0.0498094931,-0.0692166165,-0.7117774487,-0.0491668433,0.369114548,0.1050426587,-0.006304123,0.2704651952,-0.0664257482,-0.0887218788,-0.1651454121,0.4005799592,0.2827533782,-0.1001332179,-0.0985568687,-0.4512105286]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1849","title":"Add TIMIT","comments":"I am sorry! I created the PR [#1903](https:\/\/github.com\/huggingface\/datasets\/pull\/1903#). Requesting your comments! CircleCI tests are failing, will address them along with your comments!","body":"## Adding a Dataset\r\n- **Name:** *TIMIT*\r\n- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*\r\n\r\n- **Paper:** *Homepage*: http:\/\/groups.inf.ed.ac.uk\/ami\/corpus\/ \/ *Wikipedia*: https:\/\/en.wikipedia.org\/wiki\/TIMIT\r\n- **Data:** *https:\/\/deepai.org\/dataset\/timit*\r\n- **Motivation:** Important speech dataset\r\n\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":22,"text":"Add TIMIT\n## Adding a Dataset\r\n- **Name:** *TIMIT*\r\n- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*\r\n\r\n- **Paper:** *Homepage*: http:\/\/groups.inf.ed.ac.uk\/ami\/corpus\/ \/ *Wikipedia*: https:\/\/en.wikipedia.org\/wiki\/TIMIT\r\n- **Data:** *https:\/\/deepai.org\/dataset\/timit*\r\n- **Motivation:** Important speech dataset\r\n\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nI am sorry! I created the PR [#1903](https:\/\/github.com\/huggingface\/datasets\/pull\/1903#). Requesting your comments! CircleCI tests are failing, will address them along with your comments!","embeddings":[-0.2418492585,-0.3422915041,-0.1415262073,0.071031943,-0.0340815671,-0.0950556621,0.1323646009,0.1322909743,-0.3182575405,0.2184933424,-0.3416487277,0.185334608,-0.2274647206,0.2172827274,0.1103418171,-0.274307251,-0.0984166339,0.0761334673,0.0409776382,-0.0154816629,0.0028428342,0.1751541942,-0.1123562455,-0.2232402563,-0.030483596,-0.1709174365,-0.0104662273,-0.1952399015,-0.2521417737,-0.4360906482,-0.2043971121,0.1632164568,-0.0707419217,0.478176862,-0.000098801,-0.0432505235,0.3008354902,-0.2335778475,-0.2098337859,-0.0855702981,0.1470597982,0.0775862336,-0.2077464014,0.0137366457,-0.3632184267,0.081985265,0.0991953835,-0.1189360246,0.1455515176,0.3298049867,0.302805841,0.2105756998,0.1843785495,-0.14840509,0.291999191,0.1259447336,-0.2449664474,-0.0080540609,0.3158089221,0.2650029063,-0.0971473306,0.3574872017,-0.0624654032,-0.0115335323,0.088795051,-0.0562791713,-0.1725324392,-0.287977159,-0.0953781977,0.2019649148,0.6051986814,-0.3522334695,-0.3222532272,-0.1541470587,0.1114054844,-0.2598361075,0.0426243395,0.0666830689,-0.3443680406,0.2692187726,-0.1005103216,-0.1166076884,-0.1049024016,0.1386529356,0.1667456627,-0.0524291359,-0.1955411732,-0.2295540124,0.049006924,-0.1806072146,-0.3074808121,0.1258800626,0.0600780919,0.0210451595,0.0164981037,-0.2695724368,0.0644150525,-0.0103564216,0.1720127612,0.272112608,0.2065294534,0.2726242542,-0.1945527643,0.0149006238,0.0328458287,-0.0967329964,-0.013960856,-0.270267874,0.2339609116,-0.029196674,0.0716520175,-0.0332951508,0.0932587683,0.2528772056,-0.030330563,-0.1892641783,-0.0086388923,-0.1349222511,-0.1671446264,-0.0076821512,-0.1076601967,-0.031766437,0.0806055963,0.3081170917,-0.0160189029,0.1887349188,0.0280749369,0.2908256948,-0.1090196073,-0.1577567458,-0.0441772155,0.0036540849,0.0693574324,0.2889721096,0.3754378259,0.0159705095,0.3058305085,-0.010897629,0.0807985663,0.0981955677,-0.0059638731,-0.0427165478,0.2109492421,0.0302290097,0.1373911202,-0.0289445873,-0.1801170111,-0.0946732759,-0.2171858251,-0.1117169261,0.1574805975,-0.006221666,-0.1678809971,0.310387224,-0.036877688,-0.2231855094,0.062778011,0.5657606125,0.1443596929,-0.2489610463,0.0116325552,0.2420761734,-0.5268597603,0.0170172211,0.1285327524,0.2361277491,-0.1103185117,-0.1806617528,0.028469773,0.1733750254,0.2071208805,0.3799465299,0.0689258352,0.1884655207,-0.0609649643,0.2828649879,0.0272428263,-0.2947927415,-0.3006776571,-0.0205902513,-0.2514826655,-0.0225665607,0.1542674303,0.1046828181,0.1874707043,-0.0937609375,0.2120353878,-0.0455918014,0.0191751495,-0.0072499989,-0.2553142309,-0.1871530861,-0.1492057443,0.1607584357,-0.0763448551,-0.093991518,0.0210798513,0.1162800565,0.2667903602,-0.0516692996,0.2920298278,0.2032511681,0.3211770654,-0.0704505816,0.1548584849,-0.4971804023,0.3612621129,-0.1764655858,-0.0748251304,0.5619053841,-0.1617320925,-0.3172180355,-0.2374285758,-0.1571286321,0.0772495642,0.1063983142,0.3205331862,0.2693744302,0.0062883026,0.247822836,-0.0563230887,0.3235151172,-0.0477462113,-0.0809992924,-0.1075146124,-0.0702726319,-0.1081745699,-0.198239997,0.2134743631,0.3048537076,-0.2491975129,0.100148946,0.1128685176,0.2063148022,0.0032749476,0.1444820464,-0.0971877947,-0.0155652482,0.3050767183,-0.5114291906,-0.0984993353,0.1117218584,0.1653170139,-0.1006155685,-0.0515682511,0.0679313317,0.4120126367,-0.2124266922,0.0480910204,0.010254438,0.1641913056,-0.000184682,-0.1825164557,-0.1345059276,0.2180036604,0.0489832386,0.0848791227,-0.1027614549,-0.2898696363,0.0887442753,0.1933972687,-0.0700599179,0.2708614469,0.1330530494,-0.2582097054,-0.0150856543,-0.0243688151,-0.2392414808,0.069826968,0.3799447417,0.1702325642,-0.099079825,0.063788116,-0.1242570803,0.1739594191,0.1781301647,-0.3330646455,0.2277119756,0.1620798707,0.0017155075,-0.5846610069,-0.0186314862,-0.1945096254,0.0477673672,-0.220249474,-0.2804267704,0.0822401643,-0.1402014494,0.0397124551,-0.0347185098,-0.0160561763,0.0751506463,0.3684802353,0.2897178829,-0.1119676158,0.0800262764,-0.0768050104,0.3573745489,0.0550491661,-0.0802040324,-0.1860490292,-0.1291750222,-0.0031819162,0.2529430389,0.1803863198,0.0536898151,0.4089682996,-0.1718646884,0.2351417989,-0.0699668303,-0.2305305153,0.2565394938,-0.3240991235,0.1748987734,0.0680630133,-0.0792354718,-0.3135059476,-0.5873499513,-0.0457584821,-0.2532635927,-0.3723116219,0.0103995232,-0.2013486475,-0.1547277421,-0.2350603044,-0.4306767285,-0.1491475105,-0.458687216,0.4364933968,0.2045632303,-0.0967970192,0.3382937014,-0.2340962738,0.1816582531,-0.0209201481,0.5309724212,-0.3852647841,-0.3563811183,0.2105848044,-0.3785663247,-0.3260809481,-0.277775377,-0.0822068602,0.1823603362,0.0284910779,-0.2786485851,-0.045653794,-0.1603798419,0.2161771357,0.0745411366,0.0627595335,0.190913707,0.2142303288,-0.2686978579,-0.1128789186,-0.0405480973,0.1360976398,0.0716317147,0.1438774168,0.1299643368,-0.135065034,0.0226934608,0.5563359261,0.2371796072,0.011820606,0.4369502366,-0.0162489023,0.3077051938,-0.1899001747,-0.1996940672,0.2642294765,0.0733269081,0.0486607589,0.4626654685,0.1490514427,-0.3474370837,-0.1562966108,-0.139870882,-0.2432698309,-0.3702931404,-0.0504724942,-0.0547599271,0.2536336482,-0.1029003486,0.0904585421,-0.1287587434,-0.2342312038,0.2971387804,-0.074398756,-0.0081655439,-0.2143931389,-0.4703220725,-0.0235816091,-0.5126332045,0.0893503502,0.4468321502,0.2200338691,-0.2683324218,-0.3191055357,-0.0592535213,0.166914314,0.156460464,-0.3530772328,-0.1464752108,0.1531501263,0.2743315101,-0.2008212656,0.4838940799,-0.338524431,-0.0998316333,0.0766772777,0.1265443265,-0.253058821,0.1895697713,0.1489555985,0.2315547317,0.0622262247,-0.061058253,0.0871215165,-0.4036605656,-0.130987078,0.2098195255,0.2338943779,0.1643754244,0.2413668782,0.048471123,0.1633082628,0.0364835672,0.4714270234,0.1970905513,0.1634080112,0.2929776311,0.2803602517,0.1741773933,-0.0531518497,-0.0616635047,0.2690221071,0.0211436227,-0.2379797399,-0.1562975794,-0.3631966412,0.2750830948,0.112673521,0.2954686284,0.2113288492,0.0046410388,0.0090167057,-0.2982659638,0.20844993,0.1584712714,0.0363975316,-0.3152125478,-0.1567947268,0.0635420457,0.0378518514,-0.0335090756,0.274230957,0.0955369398,-0.2251514792,0.5279750824,-0.0991850495,1.0494732857,-0.0231630635,0.1142951176,0.2058928907,0.1653481871,0.0297465194,-0.2530485094,0.0583162233,0.102855593,-0.0705305934,-0.0058252811,0.1721448302,-0.2054262459,0.0366010219,-0.4967792928,-0.2816056311,0.253069669,0.1201017573,0.053974703,0.348063767,-0.3170424998,-0.483120352,-0.1706659198,0.3155546188,0.1884326488,0.1158039942,0.257029742,-0.1254667938,0.0130307972,-0.1530569792,-0.500433445,-0.1846955568,-0.347548902,-0.106135048,-0.1295690984,-0.3379963338,0.2349828482,0.2575419247,0.4467601478,0.3000093699,-0.4185540378,0.0933607891,0.1563783437,-0.0842720047,0.081512101,0.0669746324,0.2712974846,-0.1806640625,-0.1890214086,0.2053755373,0.1132972687,-0.1487609744,-0.1699919552,-0.1286781579,0.0244224258,-0.2915286124,0.1959647834,0.1739813089,-0.214886561,-0.2101839036,0.2649528086,0.1141244024,-0.1910846829,0.1638026237,0.0705646127,-0.1502559483,-0.1323538125,0.3076867461,0.1299338639,-0.0764308274,0.2930130661,0.053018637,-0.2316914797,-0.4250808656,-0.0415260307,-0.2792168558,-0.2017115504,0.0445269383,0.2725240588,-0.1493817568,0.0377689041,0.3083542585,0.0711307898,-0.1782240272,-0.1862171143,0.1185571179,-0.5278725624,0.2509503663,0.0587301329,0.2626579106,0.1905263066,0.226774618,0.0489658341,-0.1448141336,-0.5076662302,-0.2179989517,-0.2365243882,0.1522621661,0.1029300168,-0.330398351,0.2165377587,-0.0312763266,0.2940456569,0.1068748236,-0.0670449436,-0.3291207552,-0.1718227267,0.0825052783,-0.060948398,-0.0372560807,0.0675364733,0.0461288877,-0.1899998933,-0.0798482895,0.0884204283,0.3399567306,0.0525496118,0.204546079,0.0548166037,0.0351565592,0.0885067359,-0.0793366283,0.0393370353,0.1243656501,0.0693720356,-0.0636519641,-0.0231764261,-0.1475737095,0.2409811616,0.358311832,0.3597871363,-0.1791795939,0.1444511116,0.0110834874,-0.0464875139,0.4243443906,0.3072153926,0.1391829103,-0.0690601096,0.0093764933,0.3417519033,0.3756938577,-0.2730284035,-0.0133551015,-0.2209351212,-0.2101661861,0.3712542951,0.2329466939,0.0274006948,-0.1402460337,0.2200193405,-0.0339208618,0.0619054511,-0.2646507025,-0.1478272378,0.1841284484,0.2676275373,-0.1314225048,-0.0275135599,0.0526320562,0.172698617,0.1547852308,-0.0378277227,0.3055897653,-0.1212310791,0.2179777622,0.2608344555,0.101679869,-0.0220486894,0.0591823794,-0.0334418043,0.3391269445,0.1304327399,0.7346745729,-0.0284307469,-0.1605461687,-0.1413049251,0.1200404763,-0.0749060363,-0.1871409416,-0.3938402534,-0.1689349562,-0.0201309677,-0.2207297981,0.1600558311,0.0920049772,0.0608782209,0.0411275327,-0.0018079131,-0.4389417768,-0.2392476052,-0.1338041723,0.513122797,-0.262247026,0.1553128064,0.1738094389,-0.1143363491,0.1716643572,0.4402119219,0.3661359847,-0.2168471813,-0.2235011309,0.2903327942,-0.3593687117,-0.1105686873,-0.2341605574,0.4193156064,0.3314057291,0.0462879427,0.2952575386,0.2443084121,-0.2006036639,-0.1042799875,0.1925419867,-0.1796628535,-0.1794962734,0.1184096187,-0.1055968553,-0.0757155865,-0.372528106,-0.2455049306,-0.3838872015,0.2319805473,0.3348481655,-0.207920745,0.0855823532,-0.1815817803,0.1607810557,0.1219336465,0.2953212559,0.3858650923,0.0911169946,0.1655750275,-0.2164632529,-0.4014790952,0.0440534428,0.1756741852,0.1183046624,-0.1942743361,-0.0400505178,0.3752782345,0.2182672322,-0.1899239272,-0.2212154716,-0.5325388312,0.146890521,-0.2821151316,0.2319547236,-0.1006964445,0.2195285112,0.0863625854,-0.1229924709,-0.0546258278,0.1186649874,0.2042069882,-0.3010460436,0.0513894372,-0.1029091626,0.0987621471,0.1282826662,-0.0332266204,0.2983717322,-0.4984864295,-0.0636937618,-0.1608519852,-0.1427331716,-0.2007310987,0.0725727081,0.1483576149,0.3701210618,-0.1252955496,-0.0370343067,0.1789205074,0.2303701937,0.2046281844,0.1254641414,-0.1938413978,0.1363549531,-0.2121981978,-0.0349906757,-0.1203691214,0.3638459146,-0.0856244937,-0.0155381784,-0.1982891858,-0.6546860933,0.3485665917,-0.0960072875,-0.2456286401,0.0618923828,0.1563708186,-0.088796474,-0.0537872761,-0.6362185478,0.0021297548,0.2910182178,0.0620850436,-0.1628780067,0.3317240775,0.0392941907,-0.238017723,-0.1869782805,0.586060524,0.1730203778,-0.0988855287,-0.044994548,-0.279304415]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1844","title":"Update Open Subtitles corpus with original sentence IDs","comments":"Hi ! You're right this can can useful.\r\nThis should be easy to add, so feel free to give it a try if you want to contribute :)\r\nI think we just need to add it to the _generate_examples method of the OpenSubtitles dataset builder [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/open_subtitles\/open_subtitles.py#L103)","body":"Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!","comment_length":46,"text":"Update Open Subtitles corpus with original sentence IDs\nHi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!\nHi ! You're right this can can useful.\r\nThis should be easy to add, so feel free to give it a try if you want to contribute :)\r\nI think we just need to add it to the _generate_examples method of the OpenSubtitles dataset builder [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/open_subtitles\/open_subtitles.py#L103)","embeddings":[0.2011127472,0.1604523808,-0.048341617,-0.2089405507,-0.1301655918,0.2428105474,0.2711240649,0.2571239769,-0.2969917953,-0.0386743955,-0.3145644665,0.2453640103,0.215855971,-0.0763090253,0.0854991674,-0.3325570524,0.117993623,0.0025806152,0.5615056753,-0.2076058537,-0.3436259627,0.1963971108,0.0123375561,-0.1273038834,-0.1921984106,0.3085808158,-0.4058002532,0.0476240404,0.0575446412,-0.5313264728,0.0231936891,0.53368783,0.2939296663,0.0675309375,-0.0001140425,-0.1719845384,0.2067976594,-0.2393595129,-0.2624618709,-0.0709322691,0.087852411,-0.3210977018,-0.233664602,-0.3475590646,0.024387246,0.0029209892,0.0848526284,-0.3639572561,-0.1932421178,0.1132182404,0.1002750471,-0.0030477983,0.0137683395,-0.1201956421,0.576749146,-0.0308856275,0.1840917468,0.310872823,0.5937849283,0.2088736743,0.1791096926,0.3953019381,-0.426784724,-0.4061588049,-0.2225939482,-0.059054561,-0.0444161557,-0.2765921354,-0.0753083751,0.1490181535,0.4300260246,-0.1163203642,-0.7273470163,-0.4193949997,0.0371321552,-0.2022924572,-0.2925474346,0.4444063604,0.3025398254,0.3318837881,0.1208859906,-0.130856052,0.087955229,-0.0798967779,0.3085599244,0.1870670319,0.2493821979,-0.1455203444,0.1123643667,-0.1316785812,-0.2491819113,0.1970633268,-0.4026860297,0.0355916992,-0.0091518778,-0.1193122715,-0.1048696861,-0.4867628813,-0.0857825354,-0.0360756852,0.0594599918,0.0893293247,-0.2421433032,-0.1071767956,0.1065069735,0.298519969,0.4347006679,0.2094463259,-0.0822046176,0.043052882,-0.1378408521,-0.0407979973,-0.2322012633,0.2316170037,-0.318601191,0.2469429672,0.139196381,0.0557126105,-0.0039227339,0.1390759498,-0.4542080164,-0.0586221181,-0.1460408717,0.0145326732,0.0881806687,0.2371002883,0.2280703038,0.1176126897,-0.0117670847,-0.3454347253,-0.0980366468,-0.0226032846,0.0525417402,0.0726901889,-0.1043014079,-0.3754671514,0.3222724497,-0.2243033648,-0.1549776644,-0.1410826445,-0.2971782684,0.1499846727,0.0397361554,-0.1925652623,0.041253943,0.0150707783,-0.1407560706,-0.2130510658,0.0375022851,0.2365282476,-0.1183135211,-0.3088016808,-0.382745564,0.1078483015,-0.3966288269,-0.3141866028,0.3781790137,0.372253418,0.192082867,-0.2534308434,0.2706927359,0.1460497379,-0.263715148,0.1551269293,-0.1551273167,0.6188797355,-0.0313073024,-0.1125205979,0.4339675903,0.161625877,0.1579474211,0.1029895842,0.4129118025,0.0553356409,-0.010922513,0.0593733303,-0.4170093834,-0.6448662877,-0.013277702,0.0296999384,0.0956353769,0.0773247108,0.3346988559,-0.0057369992,0.6432266235,-0.3904713392,-0.1378060728,-0.0248264931,0.0137326214,0.1115522534,-0.1739796996,-0.1675906926,-0.2826334834,0.1758721322,-0.0657504275,0.1716849804,0.3092030883,0.0702335387,0.4772826433,-0.1497455686,0.1135960072,0.0308297016,-0.085541591,0.4020220339,0.1651075482,0.1393790692,0.1653700471,0.1226354837,0.10773094,0.3007267714,0.1516304612,-0.297226578,-0.0077399779,-0.0184939653,0.1842532307,-0.2402290702,0.0268614292,0.0802457035,-0.1575202346,0.2302193791,-0.3226493597,0.1063893214,-0.1910547167,0.0320156887,-0.2304773033,-0.1093338057,0.1902425736,0.0240315404,-0.0246546753,0.347586602,-0.1366307586,0.1259577721,-0.136586979,0.483628124,-0.070655711,0.4251635969,0.1638557762,-0.1920099556,0.349667877,-0.2763001919,0.1353270859,-0.0511862487,-0.1747100949,-0.0120673822,-0.0625551268,0.0382844992,-0.0326817855,-0.0537497289,-0.0300624482,0.0676899701,-0.1735156178,0.0570390671,-0.5073088408,-0.3154097497,-0.0423308052,0.065723151,-0.4276809096,0.0606065989,-0.1491335034,0.191170752,0.1527017057,-0.3350786567,-0.016220564,0.3151860535,-0.030840382,0.0488446653,0.0017458451,0.1214143336,0.0330554061,0.2184406966,0.1731770933,0.1966905296,0.5540255904,-0.3212427497,0.2545955181,0.3876331151,-0.2462096065,-0.0123349885,0.2942351103,0.2663767636,-0.2216776013,0.5088400841,0.0155414334,0.0300173871,-0.0983900577,0.0024214564,-0.2300807536,-0.4314886928,0.0451223589,-0.4368382394,-0.0910854712,-0.028428724,0.2096208334,-0.2737998366,-0.3143653274,0.3792945743,0.2540464997,0.2135637403,-0.1455390751,0.0765380114,0.0547238924,-0.563849926,0.2563215494,0.0055905236,-0.1694342494,-0.1402970254,0.1002022177,-0.3154429495,0.094026193,-0.352427721,-0.2280275971,-0.2467195839,-0.0456149764,-0.0267667286,0.4491200447,-0.297927469,-0.1281398982,-0.2595208585,-0.0963046625,-0.0421307608,-0.1636430472,-0.2893582284,-0.0660986826,0.0033795612,-0.2084629685,-0.1378657818,-0.2410983741,-0.4264033139,0.4629023969,-0.0151848234,0.1453615576,0.3218548894,-0.1029266864,-0.1351843178,-0.0863540471,0.1237719581,-0.0471829474,-0.5483480096,0.1013475955,-0.2635506988,-0.0949600115,-0.220122844,-0.1992489845,-0.1949390173,-0.3742797673,-0.1243757531,0.1396112144,0.1885885447,-0.1656323075,0.513258934,0.343097806,0.0108562298,0.0152145019,-0.0075512682,-0.1736867726,-0.2486013919,0.1173659265,0.376344651,0.097822383,0.0116750374,0.1348611265,0.1020317748,0.0141495904,0.0617834479,0.0384618863,0.3445845246,0.0913360715,0.2637014985,0.0993967056,-0.4850506485,0.1564117968,-0.1270903945,0.0058016605,0.4693220258,0.2017864287,-0.2209984809,-0.0731401443,-0.4381106198,-0.1139941439,-0.1537272334,-0.2047859281,-0.3517327309,0.2849874794,0.2482715547,-0.1306210756,-0.2383817434,-0.0996629149,0.0969209522,0.0416922271,0.5833782554,0.0728279874,-0.0838385597,0.153433159,0.2956742048,0.1817954183,0.4823200703,-0.131395027,0.3425044715,-0.1902772188,0.0716255605,0.2895000875,0.3205429316,-0.1360015124,-0.3276411891,0.0279125795,0.0550176613,-0.0351121388,-0.1570400149,0.2297179252,-0.3935571015,0.5220837593,-0.1318019331,-0.0809177458,0.219526723,0.4611623585,-0.1875399351,-0.1322612613,-0.1874809712,-0.0238992684,-0.1624841541,-0.6518249512,0.0132779367,0.0223756693,-0.207948029,0.0520438142,0.0171089377,0.0400806442,0.2363800555,0.0720119402,-0.1433660239,0.0447876677,0.5176129937,0.2186272293,0.1045110449,-0.1403055191,0.168489933,0.4528078735,-0.4579145312,-0.0011402102,-0.2072575092,0.019594552,0.3598827124,0.3509316742,0.4492091537,0.0455076583,0.1835978776,-0.2512091398,-0.0417621396,-0.0971119776,-0.1415278614,-0.000177679,-0.1103364751,-0.6272033453,0.0798150599,-0.3005562425,-0.080884479,-0.6132518649,0.1487607509,-0.0729990155,0.381727308,0.3588342965,0.9718952179,0.1164781377,-0.0031878653,-0.1129505858,0.023306448,0.5999672413,0.1288739294,-0.1429368705,-0.1546358317,-0.0774729326,0.1147394702,0.1504567415,-0.2465163022,-0.2554111481,-0.3286195695,-0.0118870214,0.4563322961,-0.0810527429,0.2831675708,0.1663943082,0.1266185641,-0.4115233123,0.2408711016,0.0531746596,-0.0161105003,0.122285381,0.0293136891,-0.34165892,0.2991670668,-0.0420870967,-0.2901454866,-0.0062627494,-0.0229117908,-0.1737760603,0.0083445413,-0.0163085926,-0.031418018,0.4820683002,0.4039408267,0.2079443783,0.3259915411,0.1558142006,0.6453317404,-0.2428015321,0.2031202018,0.1655896455,0.3280934393,-0.1876985133,-0.3354986906,-0.108337,0.2084201872,-0.097455211,-0.1542138606,-0.1353330761,-0.0644014254,0.1207849458,-0.0536721312,0.1561109424,-0.1290608943,-0.2635783255,0.0345703922,0.0693497881,-0.2068424523,-0.1044678912,-0.2120332122,-0.2099656165,-0.1732967198,0.254755795,0.0511985272,-0.3391388357,0.2468519658,0.4144501984,0.0182321947,0.1121769696,0.0647427291,-0.4947630167,-0.3805944026,0.0818598494,-0.1477153748,-0.105388701,-0.0382381529,0.003004472,0.065443635,0.2761074603,0.0773888454,0.0610191971,-0.5125808716,0.14177683,-0.2169882804,0.1952539086,-0.0705380887,-0.4104269147,-0.0972804204,0.0849565864,-0.2472590655,0.1167720109,0.0008764057,0.0148648042,0.5638489723,-0.0699846074,-0.2095900327,-0.3524312973,0.030410694,0.1792789102,0.1376092732,-0.073066324,-0.3910711706,0.1469332725,-0.0028447884,0.2071704417,-0.0267399382,-0.0675701723,0.1064996347,-0.2079851031,-0.1104261652,0.3379045129,-0.0205730498,0.4321059585,0.1507407278,0.1159458235,0.0684215873,0.0946973786,0.1331356317,0.2541627586,-0.0755660161,0.108178243,-0.2199726403,-0.1154977083,0.3980039358,0.1126652211,0.3249570131,0.1341479272,0.3710054159,0.16622594,0.2087059468,0.3805686235,-0.0865018889,-0.0261102039,-0.1307563782,-0.0456683822,0.0539928824,0.1476750672,-0.2826770842,0.2530395389,-0.1312070489,-0.0767034069,-0.0560725555,0.390016824,0.2105636597,-0.110101752,0.0204217192,0.2587749362,0.1919855326,0.224138841,0.2455048263,0.3673419058,0.117802605,-0.1513081789,-0.2345248014,-0.1215379834,0.236301586,-0.0868081152,-0.1539433748,0.096325703,0.636919558,0.2987172604,0.3417818546,0.2459930629,0.2233007699,0.163158834,0.1514851451,0.1554124504,0.1427953392,0.1542404741,0.3192991316,-0.0636926964,0.1114107966,0.8260490298,0.1878916919,-0.3855328858,-0.3495058715,0.0746236369,-0.0554977506,-0.2768781483,-0.014108141,0.03825403,-0.4017593563,0.0768729746,0.0168505218,-0.1180940643,-0.490229845,-0.4592857957,0.5727486014,-0.2730342448,-0.0835941434,0.3894760907,-0.0597271957,0.4841767848,0.6211237907,0.3014133573,0.0503501669,0.0807081982,0.4370158315,-0.2049175203,0.0170685314,0.2713007331,0.2428085506,0.1779883951,0.133982271,0.0421238057,0.033400476,-0.0892528147,0.087678954,-0.0986577272,-0.0115445126,-0.0875255093,0.48056826,0.1768518835,-0.2267641276,-0.1539185345,0.2814254463,-0.3128556311,0.1296314746,0.0703220591,-0.0597440563,-0.1107532829,0.0999845341,0.034561608,0.1225449368,0.2274809182,0.5293461084,0.0486861654,0.0473001041,0.0671935901,-0.3705396354,-0.2324555963,0.3131976128,0.1587543637,0.1072214171,-0.2613006234,0.5000709295,0.3267735541,-0.3188733757,-0.1576477587,-0.4547405839,0.1657609046,-0.1472991854,0.0039096088,0.1736499816,0.1379600018,0.0377808362,0.3426951766,-0.264274925,-0.4845814109,-0.0519278571,0.1741729081,0.0393904746,0.031681709,-0.0891750976,0.0342593715,-0.1033950672,-0.0534462146,-0.1380849928,-0.1036244556,-0.1395942271,-0.5137281418,-0.1348689198,0.0003541179,-0.1134848297,0.3835103214,-0.026215842,-0.1779441237,0.1534843594,-0.2015990913,0.0693004504,-0.0074516204,-0.5337436795,0.0549646765,-0.0353371985,-0.157474637,0.2266076356,0.1493812203,0.1326276064,0.2581684291,-0.4373934269,-0.4970725477,0.5725143552,-0.239777863,0.1296602637,-0.2525005341,0.2774398327,0.1957231611,0.0759675875,-0.5542806387,0.0596116148,0.3096804023,0.0061817649,-0.3936907649,0.1828479767,0.1197393462,-0.0011750598,-0.1891246289,-0.3156067133,0.2881157994,-0.1943174154,-0.2394292951,-0.020411348]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1844","title":"Update Open Subtitles corpus with original sentence IDs","comments":"Hey @lhoestq , absolutely yes! Just one question before I start implementing. The ids found in the zip file have this format: \r\n(the following is line `22497315` of the `ids` file of the `de-en` dump)\r\n\r\n\r\n`de\/2017\/7006210\/7063319.xml.gz en\/2017\/7006210\/7050201.xml.gz 335 339 340` (every space is actually a tab, aside from the space between `339` and `340`)\r\n\r\n\r\nWhere filenames encode the information like this: `lang\/year\/imdb_id\/opensubtitles_id.xml.gz` whereas the numbers correspond to the sentence ids which are linked together (i.e. sentence `335` of the German subtitle corresponds to lines `339` and `340` of the English file)\r\n\r\nThat being said, do you think I should stick to the raw sentence id (and replace the current sequential id) or should I include more detailed metadata (or both things maybe)?\r\n\r\nGoing with raw ID is surely simpler, but including `year`, `imdbId` and `subtitleId` should save space as they're just integers; besides, any operation (like filtering or grouping) will be much easier if users don't have to manually parse the ids every time.\r\nAs for the language-specific sentenceIds, what could be the best option? A list of integers or a comma-separated string?\r\n\r\n**Note:** I did not find any official information about this encoding, but it appears to check out:\r\nhttps:\/\/www.imdb.com\/title\/tt7006210\/, https:\/\/www.opensubtitles.org\/en\/subtitles\/7063319 and https:\/\/www.opensubtitles.org\/en\/subtitles\/7050201 all link to the same episode, so I guess (I hope!) it's correct.\r\n\r\n","body":"Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!","comment_length":217,"text":"Update Open Subtitles corpus with original sentence IDs\nHi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!\nHey @lhoestq , absolutely yes! Just one question before I start implementing. The ids found in the zip file have this format: \r\n(the following is line `22497315` of the `ids` file of the `de-en` dump)\r\n\r\n\r\n`de\/2017\/7006210\/7063319.xml.gz en\/2017\/7006210\/7050201.xml.gz 335 339 340` (every space is actually a tab, aside from the space between `339` and `340`)\r\n\r\n\r\nWhere filenames encode the information like this: `lang\/year\/imdb_id\/opensubtitles_id.xml.gz` whereas the numbers correspond to the sentence ids which are linked together (i.e. sentence `335` of the German subtitle corresponds to lines `339` and `340` of the English file)\r\n\r\nThat being said, do you think I should stick to the raw sentence id (and replace the current sequential id) or should I include more detailed metadata (or both things maybe)?\r\n\r\nGoing with raw ID is surely simpler, but including `year`, `imdbId` and `subtitleId` should save space as they're just integers; besides, any operation (like filtering or grouping) will be much easier if users don't have to manually parse the ids every time.\r\nAs for the language-specific sentenceIds, what could be the best option? A list of integers or a comma-separated string?\r\n\r\n**Note:** I did not find any official information about this encoding, but it appears to check out:\r\nhttps:\/\/www.imdb.com\/title\/tt7006210\/, https:\/\/www.opensubtitles.org\/en\/subtitles\/7063319 and https:\/\/www.opensubtitles.org\/en\/subtitles\/7050201 all link to the same episode, so I guess (I hope!) it's correct.\r\n\r\n","embeddings":[0.2530398667,0.2482580245,-0.0133868707,-0.1263538003,-0.2649042308,0.1654839814,0.2578907013,0.2178697288,-0.3385523558,-0.1133102104,-0.3191735744,0.321539402,0.1591855288,-0.148518905,0.0773393437,-0.2948979139,0.1316313446,-0.1250012219,0.4648747444,-0.230000779,-0.3619815707,0.3236779869,-0.0370355844,-0.0936009884,-0.0615350418,0.3529788256,-0.4526572227,0.263805449,-0.0446696505,-0.4046934247,-0.0694766417,0.5757681727,0.3239983916,0.1294881254,-0.00011813,-0.2874706686,0.1034887284,-0.3369553983,-0.1963880211,0.0130040152,-0.0088700457,-0.4548185766,-0.3890919685,-0.4192702472,0.1541495472,-0.0261845551,0.1112106964,-0.3906293809,0.1209641024,0.0818594918,0.0413355418,-0.2076831311,0.1595655531,-0.0756243318,0.5919961333,0.0116517739,0.2608609796,0.3210159838,0.6891136169,0.1703936756,0.0037902112,0.3019774258,-0.2869798839,-0.3634003401,-0.1178448573,-0.052475512,0.1650882661,-0.3370415568,-0.0982493311,0.0535925254,0.3832520545,-0.1271126866,-0.6749776602,-0.4048579931,-0.1287252605,-0.2527377903,-0.2395202219,0.2474868596,0.2107637078,0.3021943271,0.2062903345,-0.083272174,0.0339344703,-0.0094237775,0.1293780506,0.5042916536,0.3008296788,-0.136718899,0.1875466853,-0.1222259253,-0.1583456546,0.1058051065,-0.3015627563,0.0585046262,0.0942883044,-0.2353825122,-0.1615969092,-0.6308349371,0.0744334906,-0.0400858186,0.0709191784,0.1509939283,-0.1224713326,-0.1361684352,0.1698249876,0.4294911325,0.4613443017,0.1024640873,-0.1521368176,0.0124330735,0.0659148395,0.0766409338,-0.2118076831,0.1515748054,-0.3587512374,0.4623745978,-0.0251668692,0.0540593192,0.0812720433,0.1748130471,-0.3938657343,-0.1647633761,-0.1746284962,-0.1631082594,0.089661397,0.3201333284,0.1945272982,0.0853317603,-0.0248157345,-0.3484392464,-0.0304052774,-0.0231606606,-0.0796527043,-0.0301865581,-0.0908204019,-0.2682775259,0.2477035969,-0.2299614251,-0.1800294667,-0.1032726467,-0.2297187299,0.1282507926,0.0665439293,-0.235731259,-0.0581265166,0.005192549,-0.1624842137,-0.2596414983,-0.0665081665,0.2597441673,-0.1997034699,-0.2225962579,-0.376270622,0.0550482906,-0.3740643859,-0.2558541,0.4713926017,0.410369277,0.1603391767,-0.3092692494,0.2056641579,0.2361136079,-0.2818244994,0.1293491572,-0.2269668281,0.678419888,-0.2738695145,0.0262140688,0.3277464211,0.2242723852,0.2916445434,0.0869118124,0.348831445,-0.0180480406,0.1029496267,0.1082908139,-0.3351129889,-0.5341982841,0.0047101206,0.1675186008,0.0081042051,0.1024184525,0.329133153,-0.0977428705,0.535001874,-0.4903379679,-0.005577025,-0.0045308685,-0.0041214139,0.1031665057,-0.2395101935,-0.2100030184,-0.1895283908,0.1265672892,-0.0182916187,0.0177376457,0.2712003291,0.0873961747,0.4465692341,-0.1188609228,0.1369723529,0.0870912522,-0.1791390479,0.4175780714,0.181468606,0.3354213238,-0.0550518185,0.0875709504,-0.0712575391,0.2913719416,0.0699675158,-0.2244154811,0.0047409148,-0.0830375627,0.1683905721,-0.2202780694,0.0211334862,0.2027320117,-0.0942851231,0.0535805114,-0.3407467008,0.0740640387,0.0300336331,0.0245651677,-0.3404153883,-0.0929094553,0.0669452921,0.0342075676,-0.0637577176,0.2528347373,-0.0690941364,-0.041800756,-0.0741921291,0.4374390543,0.081607312,0.1990164816,0.0798434019,-0.1302948743,0.4036843777,-0.1952096969,0.3103134632,0.0083686532,-0.1633295864,-0.1201769859,-0.1196015403,0.0521892793,-0.1344232559,-0.1201385334,-0.1318626255,-0.0429544412,-0.1800505221,0.0462143347,-0.5524294376,-0.30593431,-0.1259777546,0.0782290027,-0.3559444249,0.2043480575,-0.0337421484,0.0700286329,0.1798924059,-0.3022330105,-0.183556661,0.2993932664,0.1039041579,-0.0285239536,0.0837333649,0.1877774298,0.1098017022,0.2179263532,0.2721732259,0.1856481731,0.5387432575,-0.3212424517,0.1539241523,0.3775050044,-0.2700375021,0.0931314901,0.166388616,0.337448746,-0.0746063367,0.4613173008,0.1328694373,-0.0213731285,-0.0950568318,0.1000728086,-0.1170583591,-0.4870334268,0.0527894236,-0.4062760472,-0.0933156684,-0.0032705336,0.2948794663,-0.4022113681,-0.4208998978,0.3982215524,0.4475120306,0.2819494009,-0.1660983562,0.0499904528,-0.055149138,-0.535300374,0.1435553581,-0.0170987472,0.026980307,-0.0771987438,-0.0382113047,-0.3761484921,0.0811533481,-0.275706023,-0.3715783656,-0.2372110486,-0.102462925,0.0563879237,0.325153023,-0.1816831231,-0.1408331394,-0.3250474632,-0.0982267335,0.0337392986,-0.1249266788,-0.2722986937,0.0482243709,-0.0597654581,-0.2288581133,-0.1042749733,-0.3540494442,-0.3807478249,0.3775185943,0.0417255238,0.2504079342,0.2632560134,-0.1098387688,-0.2220441252,0.0035443078,0.0477158874,-0.0502441227,-0.5488939881,0.1463500708,-0.2529658377,-0.0806219429,-0.239562124,-0.2055959404,-0.0889166594,-0.2992272377,-0.0037737396,0.1729840636,0.2357706577,-0.1624863744,0.3506118357,0.353094548,0.1605845392,0.012619881,0.0374525525,-0.1594702005,-0.1314894259,0.233860001,0.4142292142,0.1699800342,0.1112870499,0.0068856925,0.2042997479,-0.0031313354,0.0956231579,0.0647799298,0.2745959461,0.2683442235,0.1946626157,0.095864661,-0.3804878294,-0.0181422625,-0.0491972901,-0.0404909365,0.5621243715,0.1404630691,-0.3641775548,-0.2375995219,-0.2099023312,-0.08275529,-0.1877187639,-0.1030845121,-0.1791369468,0.1677153707,0.1812946647,-0.1290346235,-0.3685337901,-0.0173307322,0.0836431682,0.0225300025,0.5093563795,0.1372778267,-0.0270878933,0.2378229499,0.1579574645,0.2556654215,0.4453361034,-0.1813922375,0.3057239056,-0.2214791179,0.0665064752,0.1207905412,0.1276637763,-0.1393958479,-0.2735548913,0.1430666,0.1358129978,0.0053561782,-0.1844864935,0.0793708414,-0.5605741739,0.6572230458,-0.2063097954,-0.3128501773,0.1138931438,0.3463481963,-0.2300786972,-0.0502514914,-0.1849716455,0.127347976,-0.2202014029,-0.5741534233,-0.0089016594,0.1334369481,-0.2674989998,0.189199239,0.090397872,0.1831371635,0.2104641348,0.1201453656,-0.1454614848,-0.1112167537,0.5019413829,0.0752010569,0.1396638006,-0.1322804242,0.1509726942,0.3216265142,-0.4710198045,-0.0569669716,-0.1252305955,-0.0659663305,0.2320081443,0.4689921439,0.378462851,-0.2232717425,0.1639681756,-0.2812081277,-0.0888963491,-0.1528408527,-0.0513982177,0.0142657263,-0.0724856779,-0.5687928796,0.0256694406,-0.1905834377,-0.2089550495,-0.4614226818,-0.0136133591,-0.1713098884,0.5519624352,0.4034868181,1.0733832121,0.1448941529,-0.018529864,-0.1481234431,0.0640101731,0.4912926853,-0.0183081366,-0.1322436929,-0.0908878967,0.0339586399,0.0206843261,0.1917446852,-0.1598878056,0.0304867718,-0.332767427,0.0120823607,0.4895657599,-0.3025802672,0.1646775901,0.2287435234,0.0793667734,-0.3593931198,0.2810902297,0.0307748429,-0.07432127,0.1336574405,0.0235899054,-0.4508494735,0.0630009249,-0.1176923364,-0.177118212,-0.1022900343,-0.1640401632,-0.1470663548,-0.157105431,0.0631596297,-0.0823280141,0.4446716011,0.3506683111,0.1507035494,0.3611936867,0.030915169,0.7040262818,-0.1592651755,0.257650435,0.0856702775,0.372604996,-0.1405430883,-0.1735804379,-0.1560653448,0.1223667338,-0.0519307069,-0.0958983079,-0.1476422101,0.0959709808,0.0769844502,0.0065423027,0.3046070635,-0.3487659097,-0.0853958651,0.0194904692,-0.0126291411,-0.2387197465,0.0513565354,-0.3527488112,-0.1641180664,-0.1140380651,0.231455937,-0.047793556,-0.3762437105,0.0849552453,0.3765699565,-0.0680618733,0.1589422226,0.0315556973,-0.5533143878,-0.4605738819,-0.0533404127,-0.202347666,-0.2377945483,-0.0632308125,0.0316795595,0.2875443995,0.2954855263,0.2739046812,0.138817504,-0.481105715,0.2581267357,-0.1733326465,0.209906593,-0.0442113206,-0.5055652261,-0.1244230047,0.0745819882,-0.2233343273,0.2171487361,0.1428261846,-0.0337036103,0.5198439956,-0.1664749831,-0.3323398829,-0.2878307998,-0.0012587014,0.2772496641,0.1212463006,-0.0247074757,-0.2537344396,0.1573596746,0.0411160924,-0.0572441518,0.0772309229,-0.0840781406,0.060328383,-0.1545600295,-0.0704467148,0.2445056289,0.0040303688,0.3842096627,0.3261342645,0.2246763408,0.0569054075,0.2061653882,0.2081815153,0.3820981681,-0.1588955671,0.1095411628,-0.2445545197,-0.1486317962,0.4011065066,0.2084989846,0.3131086528,0.0799797997,0.3748272955,0.155956313,0.1776863933,0.4541484118,-0.1597455144,0.1450910121,-0.1101409197,-0.1271136105,0.1100172848,0.1496826559,-0.1734428853,0.2525629997,-0.0896005258,0.1228641123,-0.0768061429,0.3847633898,0.2813182473,-0.1984005421,0.0618180148,0.2656445503,0.251360625,0.348665297,0.2069728225,0.4152640402,0.0125699937,-0.0708378479,-0.3461160064,-0.0414786488,0.2416089922,-0.2297036052,-0.1193147302,-0.0093490491,0.6945890784,0.3947252035,0.1883924305,0.0523942821,0.0906520486,0.3460868001,0.286516577,0.1587851793,0.2790278196,0.1509080976,0.3806928992,-0.0695968792,0.0784194916,0.8199089766,0.1561669558,-0.2231380343,-0.4024161994,0.0478143692,-0.1816150695,-0.2417518348,-0.159665823,-0.0478023738,-0.3144462705,0.077819556,0.1436912566,0.0147256246,-0.4442596138,-0.6542370319,0.5127457976,-0.2960859537,0.1453993618,0.452932775,-0.1335956007,0.5429689884,0.5608369708,0.2101543546,0.0149205737,0.1236911044,0.3338797688,-0.2333038896,0.0264719371,0.1982179731,0.3579053581,0.1713339984,0.0617686771,0.0360179991,-0.0044262484,-0.071046181,-0.1227645501,-0.0512354299,-0.0765376911,0.151315406,0.4317429662,0.2671118975,-0.1425494999,0.0683981106,0.0813967437,-0.3937858343,0.0318681374,-0.093443349,-0.0117609296,0.0356452912,0.1442168504,0.0090418914,0.1034895778,0.1283718795,0.4288342893,-0.0354287103,-0.0362746119,-0.0281196553,-0.3973760903,-0.242124036,0.358913511,0.2464030832,-0.0135068437,-0.2603784204,0.4524381459,0.4628766477,-0.3483905196,-0.0927273333,-0.3676121235,0.3355752528,-0.3471516669,0.2463448644,0.4343974292,0.1993956119,0.0708334446,0.3511961699,-0.2094086856,-0.3332326412,-0.0804387107,0.2059751153,-0.0785257593,-0.0017291512,0.0544990636,-0.0438442975,-0.077815637,0.0744740218,-0.1033187732,-0.0786031634,-0.0953914151,-0.2745011151,-0.1010476276,0.0002010494,-0.0492281802,0.3249865472,-0.0884421989,-0.1082547083,0.0911930725,-0.2899893522,-0.0729307309,0.0258130878,-0.6710774302,0.0293551069,-0.0577792898,-0.1040101871,0.2080308348,0.2772983313,0.2454519123,0.2480411828,-0.4210081697,-0.468588531,0.577604413,-0.0889924616,0.0967506915,-0.2642005086,0.1231489405,0.2944205999,0.1640370339,-0.3554765582,0.0609459467,0.1970304251,-0.0203459617,-0.2716576755,0.0776118264,-0.0175608676,-0.1964315176,-0.2448689938,-0.2925896645,0.2346774638,-0.1828187108,-0.312143445,-0.0581053272]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1844","title":"Update Open Subtitles corpus with original sentence IDs","comments":"I like the idea of having `year`, `imdbId` and `subtitleId` as columns for filtering for example.\r\nAnd for the `sentenceIds` a list of integers is fine.","body":"Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!","comment_length":26,"text":"Update Open Subtitles corpus with original sentence IDs\nHi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!\nI like the idea of having `year`, `imdbId` and `subtitleId` as columns for filtering for example.\r\nAnd for the `sentenceIds` a list of integers is fine.","embeddings":[0.2025861591,0.1520482749,-0.0463982224,-0.2151798755,-0.2097586691,0.221863687,0.2927139401,0.3270958364,-0.2807451785,-0.0863660574,-0.3423935175,0.1748225391,0.1811573803,-0.1334889829,0.0285317414,-0.3397961855,0.1369057894,-0.0015652234,0.6896283627,-0.2206569463,-0.3582674265,0.2426954508,-0.0653146356,-0.0562278926,-0.1881902516,0.2943160236,-0.441863507,0.1019460931,-0.0228532888,-0.4655096531,-0.1128995717,0.5621749759,0.3112178743,0.0426083691,-0.0001207517,-0.2834954262,0.1895882338,-0.2500925064,-0.2641490102,0.0642997324,0.0472476892,-0.2889767289,-0.1915716976,-0.3414881825,0.0659717992,-0.0099867871,0.0176816415,-0.372854054,-0.1840544194,0.1396956146,0.0585583486,-0.1017709374,0.0272392314,-0.0978069603,0.5829442143,-0.0597430728,0.1945432276,0.2558287978,0.7497327924,0.1804830432,0.2259276658,0.3277651966,-0.5435162783,-0.4296785593,-0.1445403695,-0.0794596225,0.0035767611,-0.2353162467,0.0001521291,0.1074103117,0.4455190897,-0.0436719432,-0.6220408082,-0.3864548802,-0.0650598407,-0.2182623148,-0.2932415307,0.3142329752,0.2823574245,0.3710572124,0.1216522455,-0.1338570863,0.1120792702,-0.0174103528,0.2732359469,0.2782105207,0.2881120741,-0.138828367,0.1407511085,-0.1199690849,-0.2138815522,0.1529730707,-0.3576210141,0.0235266052,0.0143027995,-0.2208352685,-0.1321441531,-0.556173265,-0.1693895906,-0.0990579873,0.0849318728,0.1106009781,-0.0765191391,-0.102568455,0.1607458293,0.2703847885,0.4041804075,0.1389502883,-0.1216187254,-0.0647513568,-0.1501909643,-0.0333112814,-0.2252694368,0.3126197755,-0.2768231928,0.257481277,0.059220884,0.028795721,0.0050731483,0.1269140691,-0.45851776,-0.0567226782,-0.1381124407,-0.0049004755,0.117999807,0.2369013429,0.208266601,0.1077228338,0.0242550969,-0.4484051466,-0.0871648341,0.024688499,0.0525525399,-0.0359734483,-0.114019379,-0.3677076995,0.2324878424,-0.2234974205,-0.2207295895,-0.2021284103,-0.2647818029,0.166938886,0.1120090112,-0.1280201077,0.0895144045,0.0608766302,-0.1455751806,-0.1422111094,0.0009645282,0.2889186144,-0.1694345772,-0.3882222772,-0.419380784,0.0715942681,-0.3491366506,-0.3035748303,0.418615371,0.3618099093,0.2122471482,-0.2627322972,0.2700008452,0.1401197612,-0.303457886,0.0973143578,-0.2333872616,0.5517137051,-0.1110342965,0.0092204958,0.2950673401,0.2318435162,0.1940508634,0.1153714731,0.461801976,0.0065679187,0.1139685959,0.0176448878,-0.3239230812,-0.5405447483,0.0531943701,0.0289634187,0.0920103043,0.036595203,0.3335027099,0.024992967,0.6625003219,-0.4460938573,-0.0947645307,-0.0549961925,-0.0807454512,0.0978237167,-0.1732608676,-0.1785653532,-0.3534193039,0.1526336074,0.0105059454,0.1988874376,0.3615283966,0.0638976917,0.4545430243,-0.1058073267,0.0929382294,-0.0068921787,-0.0940113887,0.4595704377,0.1689643413,0.1794946045,0.1452911496,0.0608795434,0.1969762295,0.2431055456,0.2571724951,-0.2650688887,-0.0098635172,-0.0165852923,0.3412771523,-0.1586156934,-0.0077307052,0.1010709926,-0.133377701,0.1511722505,-0.4317301512,0.1067726538,-0.2014320791,-0.0007564338,-0.1344542652,-0.1698817462,0.1681282967,-0.0471914858,-0.0203962885,0.2770138979,-0.1128405854,0.1062748656,-0.0983023867,0.4456545115,-0.0735293105,0.3537281156,0.0585220531,-0.0232346877,0.2886241078,-0.2466838956,0.1673655659,-0.1667594314,-0.1978946477,-0.0774780214,0.053305015,0.0161478352,-0.0939232185,-0.0808402225,-0.1326182932,0.0365873277,-0.2571313381,0.0656199157,-0.4624585211,-0.2456923574,-0.0678145438,0.0629478246,-0.5987548232,0.0871661901,-0.1913071573,0.1686562151,0.0748032406,-0.2727589905,0.0040816404,0.2686508894,0.0531393364,-0.0229542404,-0.0408203602,0.0461186431,0.007834537,0.1784646958,0.1472988427,0.1579871327,0.528418839,-0.2943545878,0.2633446753,0.4139411151,-0.3146460652,-0.0506787598,0.2791536748,0.3343706429,-0.1581017226,0.528850913,0.007489115,0.058632493,-0.0998241454,0.0074828048,-0.1781918555,-0.4455892444,0.2280278951,-0.4807472527,0.0258938931,0.0110601997,0.2136840075,-0.2000710219,-0.3091619313,0.296297878,0.157410875,0.3554836512,-0.1057544574,0.1136808097,-0.0338432416,-0.4956925809,0.2197566777,-0.0266554169,-0.1890275329,-0.0995025858,0.0717255846,-0.2529875338,0.1459833533,-0.3089468777,-0.2556652725,-0.3172927499,-0.0103713507,0.0000219001,0.399818182,-0.1479943991,-0.2118763477,-0.2598623335,-0.1157898605,-0.034712933,-0.134236142,-0.2672845721,0.0388929397,-0.075001806,-0.159788698,-0.0255379602,-0.169921726,-0.418992579,0.2884587348,-0.0339388698,0.1766667217,0.2320150286,-0.068786867,-0.2220520526,-0.0449988693,0.1599311829,-0.0339272842,-0.5086304545,0.0821540207,-0.1738778055,-0.0609973446,-0.2035119683,-0.184882164,-0.3007088304,-0.285441488,-0.0657982305,0.2014685273,0.2741651237,-0.2688775361,0.5612764955,0.2155175209,0.0484091453,-0.0393469669,0.0312082842,-0.1745467186,-0.3107941151,0.0905998051,0.4777968228,0.1161536574,0.0455280729,0.1407916099,0.0875350535,-0.0364047624,0.0177185312,0.0103449132,0.3364094794,0.1912406534,0.1911735833,0.0350188054,-0.5634853244,0.0837580785,-0.0302502178,-0.0681339875,0.5521844625,0.1922180653,-0.2857973278,-0.1458038837,-0.2934899628,-0.114809446,-0.0804420486,-0.1262031198,-0.2518871725,0.3526644409,0.2575189471,-0.1440758556,-0.2897815704,-0.1491623819,0.0508484133,0.0984703079,0.5742333531,0.1250554621,-0.0238495655,0.1902288347,0.3251264095,0.2206346095,0.4722902477,-0.1029443219,0.290004462,-0.2374357134,0.178675279,0.3376524448,0.2217260003,-0.2346447259,-0.3685336411,0.0910918787,0.0209178608,0.0453170761,-0.1380934119,0.2159674317,-0.5401023626,0.6022507548,-0.1985394359,-0.0314700566,0.2185075432,0.4779143035,-0.2694036365,-0.1137057394,-0.203218326,0.0512437895,-0.1224550009,-0.589353323,-0.0019228118,0.0421613753,-0.2644689679,0.151761651,-0.0246821865,0.071737811,0.208633855,0.0964044407,-0.1431051046,-0.0627297834,0.5997660756,0.1291745901,0.1122450829,-0.1474158615,0.1794982851,0.2560945749,-0.4961569011,0.0931706578,-0.1482700109,-0.0122069363,0.3469993174,0.3949159682,0.4357192516,0.0169924963,0.1802895665,-0.3192499876,0.0696510449,-0.148743242,-0.1462892592,-0.0506361276,-0.0944407508,-0.6300880313,0.0256787688,-0.2706144452,-0.1397592723,-0.5500949621,0.1543126702,-0.0456875078,0.4206373394,0.410929203,1.0612920523,0.0454476997,-0.1464203149,-0.2469635159,0.0419370346,0.5492524505,0.1620164961,-0.0913032591,-0.160762921,0.0321069695,0.0989378393,0.1161470264,-0.2794188559,-0.254139483,-0.3178933561,-0.087288335,0.4687195718,-0.1088776812,0.2544683516,0.1779150069,0.2230877429,-0.3008314371,0.3246660531,0.0144247068,-0.0300621893,0.1073245481,0.0366573706,-0.3281346262,0.2605344653,-0.0539077222,-0.3064029813,-0.1965384334,0.0592799112,-0.1814666092,-0.0050466866,-0.0354714356,-0.035126511,0.432123214,0.4008662701,0.1190458909,0.3698362708,0.0241539218,0.7271578312,-0.2729674578,0.1730531901,0.1580008715,0.3656669259,-0.1582157612,-0.2347856909,-0.0431228355,0.2181022018,-0.243164748,-0.181212455,-0.084831208,0.1039575338,0.0822017491,-0.0676287264,0.0914621949,-0.1539804786,-0.2264296263,-0.002695462,0.0611706004,-0.0550510474,-0.089151822,-0.1379127651,-0.1850065887,-0.1715220511,0.2025743127,0.0605082177,-0.4358052313,0.1626009047,0.424685657,-0.0069942744,0.1647747159,-0.0002613117,-0.6238164306,-0.284828186,0.1141531989,-0.2611319125,-0.2134479433,0.0011068508,0.0495729223,0.0999459773,0.3431379199,0.1333882809,0.1001592278,-0.4750415385,0.1921489984,-0.2563412189,0.1356143802,-0.154953301,-0.4797373712,-0.1053514853,0.0146619435,-0.1941798925,0.0978009999,0.0607111193,0.0454815738,0.5873957872,-0.0277258884,-0.3805016279,-0.3255670965,-0.0307498612,0.1600893438,0.172245115,-0.0233361498,-0.2756186426,0.1958900094,0.0542424992,0.1653216779,-0.0756326839,-0.1026798263,0.0672932044,-0.1668602675,-0.2070133835,0.3138858676,0.014466471,0.2859558165,0.1922881305,0.0958615541,0.0852893516,0.1718426943,0.1404540241,0.202404052,-0.1502433121,0.115749903,-0.1683477312,-0.140915215,0.5077627897,0.1434131861,0.305277586,0.0637398437,0.3995049894,0.2278867215,0.2209694386,0.4789087772,-0.2055095285,0.025129376,-0.1554472893,-0.0812236667,-0.0376960374,0.1249285117,-0.1966225803,0.2986119688,-0.1273751259,0.1797408164,-0.1088070869,0.3950021863,0.226870656,-0.216406256,0.0876006261,0.2835348845,0.24829714,0.2138542682,0.1780238897,0.3404660225,0.1227643639,-0.1538021713,-0.2184629589,-0.1127519757,0.1651268601,-0.1620827764,-0.0731364712,-0.0563388541,0.7049536705,0.3048973382,0.3184443712,0.22545214,0.2580075264,0.2470722198,0.1487858593,0.1365235597,0.2202591002,0.0321096517,0.4621669352,-0.0490972362,0.0792935789,0.7958602309,0.093100965,-0.2446578443,-0.4160209596,0.0545521118,-0.0208306126,-0.3062519431,-0.0351627693,0.1146582663,-0.384249866,0.0652934387,0.0482195653,-0.0895356685,-0.4640112519,-0.4744280279,0.5227257013,-0.2781196237,-0.112893872,0.3038090467,0.0084643373,0.6416428685,0.5664156079,0.2642949224,0.1010474712,0.0864192694,0.4559990764,-0.255684942,-0.0514409617,0.2270275652,0.321972549,0.2384667695,0.1485010833,-0.0434489511,-0.0042673019,-0.0499787629,0.0832953528,-0.1263563335,0.0581449196,-0.0181444623,0.4207744598,0.2323185503,-0.2503961623,-0.0348533131,0.2278524488,-0.2158733904,0.1268191338,0.0458422638,-0.065767087,-0.125066638,0.1317737103,-0.0000037913,0.1272660196,0.0859877393,0.5072776079,-0.0002076738,0.0875705555,0.0228957068,-0.2934554815,-0.2539841235,0.3089723587,0.2383249998,0.1209153235,-0.2224777639,0.5440028906,0.4117054641,-0.2848435938,-0.1010249332,-0.5488210917,0.2706492543,-0.1794653982,0.0458364636,0.2599027157,0.1698554307,-0.0033224439,0.3813747168,-0.1600140631,-0.5111722946,-0.0679681972,0.1554610133,-0.0892324448,0.0950694457,0.1021826714,-0.0599105172,-0.0974347666,0.0022093246,-0.140312627,0.0257146042,-0.1727075279,-0.4073498249,-0.0814875066,-0.0174175147,-0.0932307914,0.3397627771,-0.026092045,-0.1358444244,0.217507273,-0.1546816975,0.0144024836,-0.0258144252,-0.5638818741,0.0654300302,-0.0396908931,-0.066131629,0.2783844769,0.0630614161,0.1904576868,0.2627518177,-0.4643498659,-0.5056293011,0.583791554,-0.2052414119,0.0938970745,-0.2969118357,0.3089847267,0.2418888509,0.1643211544,-0.4174595773,-0.0529584438,0.2758620977,-0.0863204822,-0.3164333105,0.1886339337,0.1310429573,0.0213288739,-0.1573973894,-0.3898937404,0.3080559373,-0.127613768,-0.3563329577,-0.0417051762]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1844","title":"Update Open Subtitles corpus with original sentence IDs","comments":"Thanks for improving it @Valahaar :) ","body":"Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!","comment_length":6,"text":"Update Open Subtitles corpus with original sentence IDs\nHi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!\nThanks for improving it @Valahaar :) ","embeddings":[0.2385091782,0.0411326587,-0.0701048672,-0.2531498075,-0.1922883242,0.1757838279,0.2946802378,0.2600993216,-0.3184936345,0.0109033845,-0.2936185598,0.1806429029,0.2559108734,-0.1260382831,0.0902219564,-0.3178736269,0.134591803,0.0340377502,0.6356292367,-0.2170878202,-0.3480929136,0.205214873,-0.0305393767,-0.0341973454,-0.2056984752,0.3563005924,-0.3835218549,0.0717274472,0.0410290584,-0.4665675163,0.0006836414,0.517952919,0.2569956481,-0.0020481618,-0.0001154967,-0.2161250114,0.2384640872,-0.2298586965,-0.1795046329,0.0829081982,0.0227941032,-0.3031349182,-0.2399098426,-0.3245181143,0.04683882,0.0534785502,0.09820728,-0.3507095575,-0.2469207048,0.0844334438,0.0883111209,-0.1019420624,-0.0219713673,-0.1227945313,0.545017302,-0.0729537681,0.1833640188,0.2879111469,0.6866904497,0.1826027036,0.208931163,0.398609817,-0.4904149771,-0.4037691355,-0.2028311193,-0.102940172,0.0491471328,-0.2461628914,-0.0367174521,0.144434765,0.4072777033,-0.0422054604,-0.5839229226,-0.3817543983,-0.0185396038,-0.2281208485,-0.3028188944,0.3121531606,0.3281913698,0.3908637166,0.1141157523,-0.1336661875,0.1503916681,-0.0255924668,0.3166157305,0.2049567103,0.2722668946,-0.1611500978,0.1246933341,-0.0867971852,-0.2579764724,0.2171046734,-0.4275130033,0.0021242178,0.0116551723,-0.1304299831,-0.0544841252,-0.547694087,-0.172547847,-0.0878853872,0.1059511825,0.1102436781,-0.1111580282,-0.049533315,0.0856462643,0.3350864053,0.4306411743,0.1613087356,-0.0795726627,-0.001668884,-0.1586594433,-0.0396795459,-0.2141612172,0.2852985859,-0.2947285771,0.2993528545,0.1290148795,0.063563779,0.0010623827,0.1182539463,-0.5114836693,-0.0631519035,-0.1670681536,0.006845912,0.1610931307,0.2974570692,0.2182536423,0.0681420788,-0.0191320796,-0.3927626312,-0.116158314,-0.009771985,0.0321407318,0.0145615842,-0.152750656,-0.3573973179,0.2818400562,-0.2572140396,-0.1895296425,-0.1785754859,-0.3307063878,0.1432187557,0.0860432461,-0.1927486956,0.1265302747,0.0915521681,-0.1549510956,-0.1316148192,0.0243580472,0.2906859219,-0.2040315568,-0.3742019832,-0.4007829428,0.1133049428,-0.3372978866,-0.2970984578,0.364118129,0.3589531481,0.2066531926,-0.288490057,0.3259074986,0.1620547771,-0.3102030754,0.1204795837,-0.2303573042,0.5183686614,-0.0470857471,-0.0333184525,0.3082391322,0.1387682706,0.210665077,0.1213375479,0.3869475424,0.0830719993,-0.0152505282,0.0645430386,-0.3925609589,-0.5839337707,-0.0170859545,0.0165246855,0.0215645023,0.076401405,0.2943691611,-0.0328595005,0.6133800149,-0.3988158107,-0.1424721926,-0.0260069761,0.0006882207,0.1541462392,-0.1843598336,-0.1476617604,-0.3535014987,0.0906849876,-0.0172825493,0.1518434882,0.3906272948,0.117488414,0.4809117913,-0.1198158264,0.0897193328,-0.0145510864,-0.0570486672,0.3883296847,0.1592976302,0.2126238048,0.1515559703,0.1101642326,0.102803342,0.2254133672,0.2810317576,-0.2433047444,-0.0069149793,-0.0112444246,0.2662564814,-0.272587508,0.0250928737,0.137260139,-0.2240221798,0.2387795746,-0.3977879584,0.122705996,-0.2104646415,-0.0429910906,-0.1231779605,-0.1685695797,0.1682492793,-0.0235272292,-0.0539365299,0.2959879041,-0.132343784,0.1012573764,-0.1059155613,0.4738667011,-0.1175363958,0.4543252885,0.0547806993,-0.1752753109,0.3107583821,-0.3206009269,0.1173540056,-0.1554889232,-0.2073256671,-0.0567621365,0.0247859415,-0.0192818101,-0.0030324985,-0.1311942786,-0.11042092,-0.0058437036,-0.2598792017,0.0146666765,-0.4244394898,-0.2736353278,-0.0219771378,0.1110182554,-0.52647686,0.1162608191,-0.1188218966,0.1550189257,0.191035077,-0.3198629618,0.020661192,0.3066881597,0.0394551493,-0.0016552174,-0.0067490684,0.0262860991,-0.0116849579,0.1909207851,0.1671220064,0.1565981209,0.5541487932,-0.3357284665,0.2629226744,0.3759615421,-0.224731788,-0.0276619885,0.3084382415,0.2802300453,-0.1632442921,0.5623325706,-0.0016572871,0.0352598429,-0.1053145304,-0.0300331973,-0.2091951817,-0.4872114956,0.0922109783,-0.4566916525,-0.0245172381,-0.0437641442,0.1952558458,-0.2903054655,-0.3330367208,0.3418602049,0.1836884916,0.2433516234,-0.1538806111,0.1000775546,0.0307262708,-0.4876215756,0.1971478015,0.0130258678,-0.1993417889,-0.1304058731,0.0415975749,-0.2987534404,0.1401598006,-0.2720151246,-0.228289783,-0.2484737784,-0.0185657144,-0.0638382882,0.3909180462,-0.2481505126,-0.1649878472,-0.3098790944,-0.1184802949,-0.0433695465,-0.1666132212,-0.2663245797,-0.0077072121,-0.0339578912,-0.2025760412,-0.0728955045,-0.1505842209,-0.4363654554,0.401165545,-0.110296607,0.1847876608,0.3019618988,-0.1198024601,-0.163358584,-0.0788805038,0.1149823964,-0.004862288,-0.5530983806,0.0660185218,-0.1972526163,-0.0160613507,-0.225268513,-0.1602047086,-0.2381657809,-0.3858613968,-0.0842840746,0.1761093885,0.2225780189,-0.2235019356,0.5305313468,0.2365478873,0.0346593112,-0.0551666543,-0.0092939092,-0.1420290619,-0.2715884447,0.1005478129,0.4705878198,0.155006662,0.0335771032,0.0786971226,0.1607081443,-0.0566118173,-0.0258319546,0.0703584105,0.3339280188,0.2102528065,0.2435269356,0.0312900953,-0.5540301204,0.1678706259,-0.0785398111,-0.0561442748,0.5279413462,0.1510350257,-0.3121855557,-0.1183246002,-0.3969126642,-0.0905434564,-0.1144219562,-0.1627002954,-0.2712501585,0.3043800592,0.2671272159,-0.1156417131,-0.1667663008,-0.1145156622,0.1509253532,0.1028749198,0.5525217056,0.1169351414,-0.0229198989,0.1743762344,0.3549973369,0.1774701923,0.4905994236,-0.1180895194,0.2756935358,-0.2498512268,0.1610048562,0.3390714526,0.2054677308,-0.2031674385,-0.3194183111,0.0449123681,0.0659054667,0.1212133393,-0.1741269827,0.2643869221,-0.481757164,0.6143407822,-0.1892843246,0.0063243299,0.2148978114,0.4515080452,-0.3117020726,-0.10284812,-0.1715048552,-0.0366139859,-0.0997329056,-0.622386992,0.0276330765,0.0231673736,-0.2361362576,0.1182998791,-0.0029472823,0.0214632079,0.2271289825,0.0391359963,-0.1377294958,-0.0270420834,0.5901305079,0.1298469156,0.1285037249,-0.141662091,0.140473336,0.2922388017,-0.4841948748,0.0749524832,-0.1673447639,-0.0134416576,0.38721627,0.3830912411,0.4540326595,-0.0046867132,0.1761793643,-0.2662887573,0.011255037,-0.1040527597,-0.1559546292,-0.0669044405,-0.0552754998,-0.5254816413,-0.0060316385,-0.3075252473,-0.096185863,-0.5990251899,0.1691590399,-0.0759027451,0.3313898444,0.3994527757,1.0144375563,0.1059865505,-0.1389833987,-0.2029488385,0.0218475759,0.4987595677,0.1450105757,-0.0703390539,-0.1524982005,-0.0216667801,0.1015972197,0.1463370919,-0.2779397368,-0.3051986098,-0.3045749068,-0.061979603,0.5323594213,-0.0876522362,0.2683463097,0.1637995988,0.2194395363,-0.326821506,0.31367293,0.0412719548,-0.0486634038,0.1685410291,0.001594628,-0.3356537223,0.2557192445,-0.0353132822,-0.2714336514,-0.0622713268,-0.0166777279,-0.1593730748,-0.0387663245,-0.011349353,-0.0766671002,0.4405507147,0.3620148003,0.1600417644,0.347171694,0.1123821735,0.7345730066,-0.2357628793,0.1464499533,0.1076552197,0.3475754559,-0.2040508837,-0.265378803,-0.0519819707,0.1877532601,-0.15441145,-0.2039305866,-0.0492577143,-0.0239459835,0.1649335921,-0.062732935,0.1454785019,-0.1203253865,-0.232327193,0.0325593576,0.0676883608,-0.1272209883,-0.1631771773,-0.1857771724,-0.2226776332,-0.1597347558,0.2693522573,0.007820704,-0.4306696653,0.1606290638,0.3982305527,-0.0330099165,0.1485873461,0.0564298593,-0.5457495451,-0.3335159123,0.0815541819,-0.2583537698,-0.2005814761,0.0095950877,0.0494471528,0.0529197045,0.3100442886,0.1168635637,0.1144182384,-0.4965238869,0.1680394262,-0.3061989546,0.1724036038,-0.0731726512,-0.4266828001,-0.0367332622,0.080889158,-0.2357034683,0.0780762807,0.0327157453,0.0132563747,0.5680030584,-0.1083020866,-0.2993906438,-0.3272877038,0.0053616981,0.2006707788,0.1547565013,-0.0653199926,-0.3154698908,0.161599651,0.0401425995,0.2216972113,-0.0244088285,-0.0758310631,0.1802580059,-0.1611094177,-0.186288327,0.2946015298,0.0162854977,0.3817545474,0.2027260065,0.0998959318,0.0715590641,0.1142633781,0.1217121556,0.2033158392,-0.1574086547,0.056017004,-0.2083933651,-0.1193263158,0.5329641104,0.0922766924,0.3439492583,0.1412659734,0.3125681579,0.218083784,0.2278085947,0.4295506179,-0.1999932677,-0.0434290096,-0.185567975,-0.12297488,0.0001834125,0.1502810717,-0.1811010987,0.2851026356,-0.1227156296,0.0448092446,-0.1428352892,0.3482668996,0.2083467096,-0.1628255844,0.0723519623,0.3153122962,0.1523907185,0.1584870815,0.1884978116,0.3367717862,0.0784239098,-0.1673433632,-0.2395655811,-0.1621944904,0.2461918741,-0.1101718396,-0.0916701108,-0.0000119158,0.6350054145,0.3492153585,0.3281184435,0.2653258741,0.1950273812,0.1572322994,0.1438699514,0.1412124336,0.2157698274,0.1015776172,0.4566719532,-0.0288942512,0.1157562882,0.8362417221,0.0553356372,-0.3165635467,-0.3832954764,0.1118422896,-0.0537902191,-0.2928276956,-0.0228917077,0.1544607282,-0.394066751,0.1146070808,-0.0389907733,-0.0344292484,-0.5191469789,-0.4843764901,0.5946622491,-0.2842087448,-0.102975592,0.3167232275,-0.0248732325,0.4706858993,0.6085184216,0.257353574,0.0735916942,0.0999008641,0.4314030111,-0.1741608977,-0.0197153892,0.2433860153,0.2741528451,0.2439575642,0.1683317274,0.0020905088,0.0368645601,-0.0982513353,0.1135346889,-0.0941349491,0.0427703261,-0.0313002355,0.4309895337,0.2159252167,-0.1932898462,-0.1059212759,0.3057878613,-0.2408320606,0.2011101693,-0.0122856582,-0.0531909689,-0.1547969729,0.091315791,0.0275279302,0.175770849,0.1582134217,0.4809264839,-0.0523560755,0.0742564574,0.0401333384,-0.3249293268,-0.1865442246,0.3077182472,0.2574026287,0.1236952394,-0.2324753553,0.4804792702,0.3620324135,-0.3110800385,-0.1031671762,-0.45188272,0.2455206215,-0.162539646,0.0319047533,0.2275693119,0.0711590126,0.025435375,0.379670918,-0.1981659532,-0.4616101384,-0.0238851104,0.201094687,-0.0671643764,0.0799499676,0.0543915369,-0.0027986502,-0.1112092212,-0.118661128,-0.1714072227,-0.0392578952,-0.1571739614,-0.4354531169,-0.0651259497,0.0148991914,-0.1253341138,0.3766658604,0.0109541779,-0.1860478967,0.1677650064,-0.1757780164,0.0707363859,0.0151564004,-0.5658383965,0.0207876246,0.0049436274,-0.1267190576,0.2653449178,0.0612233244,0.167854473,0.2084303647,-0.4922416508,-0.5205033422,0.5479870439,-0.185571909,0.1460298449,-0.2529903948,0.3454279602,0.2135539651,0.131020546,-0.48575598,-0.0342681184,0.2746103406,-0.0272570848,-0.3609381616,0.21026057,0.1381355077,0.0070361705,-0.1669247597,-0.334885031,0.3777521849,-0.1830803156,-0.2830196321,-0.0283158813]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1844","title":"Update Open Subtitles corpus with original sentence IDs","comments":"Something like this? (adapted from [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/open_subtitles\/open_subtitles.py#L114))\r\n\r\n```python\r\nresult = (\r\n sentence_counter,\r\n {\r\n \"id\": str(sentence_counter),\r\n \"meta\": {\r\n \"year\": year,\r\n \"imdbId\": imdb_id,\r\n \"subtitleId\": {l1: l1_sub_id, l2: l2_sub_id},\r\n \"sentenceIds\": {l1: [... source_sids ...], l2: [... target_sids ...]},\r\n # or maybe src\/tgt? I'd go with the first one for consistency with 'translation'\r\n \"subtitleId\": {\"src\": l1_sub_id, \"tgt\": l2_sub_id},\r\n \"sentenceIds\": {\"src\": [... source_sids ...], \"tgt\": [... target_sids ...]},\r\n },\r\n \"translation\": {l1: x, l2: y},\r\n },\r\n )\r\n```\r\nOr at top level, avoiding nesting into 'meta'?","body":"Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!","comment_length":79,"text":"Update Open Subtitles corpus with original sentence IDs\nHi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!\nSomething like this? (adapted from [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/open_subtitles\/open_subtitles.py#L114))\r\n\r\n```python\r\nresult = (\r\n sentence_counter,\r\n {\r\n \"id\": str(sentence_counter),\r\n \"meta\": {\r\n \"year\": year,\r\n \"imdbId\": imdb_id,\r\n \"subtitleId\": {l1: l1_sub_id, l2: l2_sub_id},\r\n \"sentenceIds\": {l1: [... source_sids ...], l2: [... target_sids ...]},\r\n # or maybe src\/tgt? I'd go with the first one for consistency with 'translation'\r\n \"subtitleId\": {\"src\": l1_sub_id, \"tgt\": l2_sub_id},\r\n \"sentenceIds\": {\"src\": [... source_sids ...], \"tgt\": [... target_sids ...]},\r\n },\r\n \"translation\": {l1: x, l2: y},\r\n },\r\n )\r\n```\r\nOr at top level, avoiding nesting into 'meta'?","embeddings":[0.2209786177,0.0347748846,-0.0174859799,-0.1517959833,-0.2049596608,0.3104095161,0.3499246836,0.2389916182,-0.4182053208,-0.1190219447,-0.399220258,0.2717107236,0.2253196388,-0.1207069606,0.1057459116,-0.2710324526,0.1577949226,-0.0647950619,0.5990246534,-0.1972804368,-0.3916414976,0.2315596342,0.0232855491,-0.1052898169,-0.191948697,0.3326811492,-0.5179935098,0.1893836856,0.0558610819,-0.4463605881,-0.0519319028,0.5385537744,0.2594017088,0.0686634034,-0.000120175,-0.1887046397,0.1532710344,-0.2464017272,-0.247608915,-0.0682022721,0.1630222648,-0.2673374414,-0.2945556939,-0.3069764972,0.0979103446,-0.0597781427,0.0081208786,-0.4214204252,-0.1565286368,0.0395493954,0.0464882366,-0.041784782,0.0353239141,-0.1026243046,0.5562952161,-0.0020258531,0.2494301498,0.321056664,0.6099584103,0.2494790703,0.2338652015,0.3001668155,-0.4769385457,-0.4708833694,-0.1287600398,-0.0272079315,0.0282106437,-0.1922350079,-0.0698113441,0.041339118,0.3280802369,-0.1330782175,-0.7111257911,-0.4584240913,-0.0461190492,-0.2136909068,-0.4144753218,0.3474140763,0.3577758074,0.3390949667,0.0551942997,-0.1647026688,0.1730967015,0.0233025476,0.2741304934,0.2407243252,0.234087944,-0.0719094723,0.1593380421,-0.1263435483,-0.3018881083,0.096115306,-0.3530394435,-0.0231116414,0.0552212633,-0.180324018,-0.0399188921,-0.5612187982,-0.0688753128,-0.0634699985,0.1200460643,0.036731869,-0.1203038245,-0.0787198544,0.1772689968,0.3284093738,0.3404316306,0.0905413851,-0.1424031109,0.026658006,-0.1131897494,-0.0110521941,-0.1618149579,0.2903782725,-0.3519211113,0.2706139684,0.1015037969,0.1630832106,-0.0103811556,0.0979341194,-0.4913239181,-0.0882689059,-0.1628619134,0.0120426901,0.1813017279,0.2778910995,0.2103706896,0.2165264189,0.019492548,-0.3489281535,-0.1100478694,0.0282664336,0.0772974715,0.0654936135,-0.1279768646,-0.3086250126,0.3220805228,-0.2250380516,-0.0650683418,-0.2030833811,-0.3065249026,0.1305768937,0.073270835,-0.1596947312,0.091713272,0.0978728831,-0.1367485672,-0.1229436547,-0.0138521027,0.2216383666,-0.1644517183,-0.3561444581,-0.2158158869,0.0465602167,-0.3473252058,-0.3419598341,0.2420505136,0.4023239613,0.2266892791,-0.2590724826,0.2422270477,0.1845205873,-0.3498832881,0.0592204221,-0.131371513,0.6238114834,-0.0077665779,-0.0832960904,0.3877950013,0.2720652521,0.0859149992,0.1677537709,0.4032240808,0.0536360554,0.0470075458,0.0432600379,-0.4101873636,-0.6244882941,0.0383599438,0.0482292734,-0.0345022529,0.1305969507,0.3557191789,0.0006950055,0.522342205,-0.4122362733,-0.0007474608,-0.0331968367,-0.0157671981,0.0964461267,-0.1680337191,-0.2356581688,-0.2997710705,0.1134860888,-0.0649364069,0.1316911429,0.2948752642,0.0013579959,0.4646084011,-0.1661217362,0.0793835521,-0.0345390551,-0.0843929499,0.3948904276,0.1212474778,0.1921689659,-0.0429225489,0.1107505858,0.165411666,0.2414632291,0.1915427893,-0.2982216775,0.0493588373,-0.0343862846,0.2178050429,-0.2670888901,-0.0152094038,0.1529762447,-0.1249362901,0.1704556495,-0.3960731924,-0.0161923468,-0.0776912048,0.0775526389,-0.1972882748,-0.1050522923,0.1627872884,0.0376423076,-0.0499208607,0.3225724995,-0.0850336775,0.0583695322,0.0013717192,0.4663871527,-0.0264587458,0.4859485626,0.0477542058,-0.1175577193,0.3558054566,-0.2490113676,0.1458485723,-0.0890548155,-0.166750282,-0.1384648681,-0.0018220597,0.0463398993,-0.0706598535,-0.0386064425,-0.037169911,-0.0073784562,-0.3098450601,0.0987724587,-0.4832093418,-0.3038065434,-0.0789630115,0.0930195823,-0.4492996931,0.1416916251,-0.1375737786,0.23616229,0.1031445265,-0.3040553331,-0.0771716908,0.2178610712,-0.0448907763,-0.0042250142,-0.0443806984,0.0655253753,0.0941631272,0.2101776302,0.1771959811,0.2408743948,0.6037320495,-0.2918301821,0.2677742541,0.421508193,-0.2798588872,0.0090122642,0.3630819321,0.3062956035,-0.1610841006,0.500187993,0.0151149537,-0.029345803,-0.155715242,0.0255987309,-0.1557438672,-0.3463326991,0.022558514,-0.509444356,-0.0530931763,-0.0738402233,0.1577112228,-0.2463462055,-0.3293325305,0.2478004098,0.2494488508,0.242607519,-0.1832521111,0.0030818116,0.0344854742,-0.528622508,0.3113759756,-0.0645060688,-0.1091002375,-0.1835632175,-0.0054513593,-0.3063260317,0.0851275697,-0.3004768491,-0.2967128158,-0.2904579937,-0.1231262237,-0.030331267,0.4129615426,-0.2559935153,-0.1472994089,-0.2908369005,0.0264723469,0.0178984199,-0.1615315825,-0.3175318241,0.0234167539,-0.1008801013,-0.2108381987,-0.0929469317,-0.2280549705,-0.3786453009,0.4878309965,-0.0483484156,0.1076342687,0.269292444,-0.0781845748,-0.2370766252,-0.1841226816,0.147393629,-0.0502740331,-0.5810415149,0.0722607672,-0.1957204938,-0.0286291912,-0.2978210747,-0.1449042261,-0.2162144184,-0.3066577315,-0.0989226922,0.1704230309,0.2413598299,-0.234220475,0.5072835684,0.3699317276,-0.0104725845,-0.0442712493,0.0359912775,-0.1981700063,-0.2408281565,0.1244890839,0.4949030578,0.0304683782,0.0134623917,0.1655344963,0.1943969876,0.1037394106,0.0954170525,-0.0017627431,0.3459439874,0.1287500709,0.233825773,0.0904993191,-0.6101649404,0.0591402911,-0.1277847141,-0.0400544517,0.4969538748,0.176601246,-0.2033743262,-0.1965546757,-0.3467213213,-0.0426354781,-0.1534173042,-0.1873701662,-0.1745566428,0.2824782729,0.2552609146,-0.1768402457,-0.2442917228,-0.0625845417,0.0728166774,0.1015824899,0.5623089075,0.1568129808,-0.0802055225,0.1029935926,0.2030635178,0.2130466998,0.5398970246,-0.1010341942,0.3852148652,-0.2962692082,0.0736889169,0.2467345744,0.2528200746,-0.1431400031,-0.3250364959,0.0906572193,-0.0771454647,-0.0479840077,-0.0847766399,0.238801226,-0.4861287177,0.6551349759,-0.0446692482,-0.138396278,0.2207531333,0.4505432844,-0.2518523037,-0.0614429377,-0.2097943723,-0.0086911665,-0.1219846681,-0.6096362472,-0.0574548095,0.0375247635,-0.2138030827,0.1593482345,0.0025439379,0.0881799757,0.1599778235,0.1477336287,-0.1215681061,-0.082762219,0.5291163921,0.1702375114,0.1728867143,-0.0909093171,0.3126510978,0.4004724026,-0.4411028326,0.0555752367,-0.2909812331,0.0360929556,0.44816944,0.3489748836,0.3569923341,-0.003615879,0.1456637233,-0.2892709672,-0.0880408362,-0.1421385854,-0.1054025814,-0.0229249708,-0.0424398296,-0.5799977779,0.0873589143,-0.2452296764,-0.0597561486,-0.6179398894,0.165803194,-0.1191089526,0.3421133161,0.3598659337,1.0053174496,0.1238172576,0.0880203769,-0.1164554805,0.1319574714,0.5445368886,0.1574399322,-0.1165827066,-0.1165342927,0.0590397604,0.1269251257,0.0888627768,-0.2667463124,-0.2439644188,-0.2833565474,-0.0237207767,0.4424541593,-0.0461775772,0.229575932,0.130550921,0.1871202588,-0.2922727764,0.286282748,-0.0227555782,0.0141836097,0.0568493158,0.0094985208,-0.312980473,0.1403810829,-0.0934900492,-0.2328750938,-0.1617426127,-0.0839832574,-0.1442180723,-0.1318127662,-0.0091923708,-0.0283598825,0.3801967204,0.3602820933,0.1380673498,0.3586059809,-0.0239597838,0.6386119723,-0.2633193135,0.1427511871,0.0863418058,0.2632280588,-0.0997211337,-0.2440456748,-0.1177793518,0.2379366755,-0.1531135887,-0.2227316648,-0.0734129697,-0.0193213709,0.1067313701,-0.1066040173,0.2211957276,-0.1674104929,-0.2496603131,-0.0274972897,0.0662308335,-0.1441971362,-0.1204332933,-0.2571228743,-0.2271402776,-0.152100876,0.2999041378,0.0110114319,-0.3845518827,0.1977353692,0.403840512,-0.0152919861,0.2056400925,0.0737703368,-0.5600082874,-0.3261530697,0.154315576,-0.205531314,-0.1241179258,-0.0972466543,0.0944913849,0.1175405234,0.3015242815,0.1594979316,0.0825542957,-0.4828440845,0.2753205895,-0.2664932609,0.1754506528,-0.030631585,-0.3450266421,-0.1609974205,0.1279798448,-0.1914197206,0.1795903742,0.129014045,-0.0228785016,0.57009238,-0.1385350227,-0.2587908804,-0.3195570409,-0.0174117386,0.1505300105,0.1474249363,-0.0189135391,-0.3108358681,0.1858166307,0.1050821245,0.2238490134,0.0075439732,-0.028108051,0.0899483711,-0.2263014615,-0.0896043479,0.3430848122,0.0869025663,0.2806044817,0.183249563,0.108940497,-0.011505127,0.1862086058,0.2204469293,0.1432758123,-0.1162975058,0.0819215328,-0.2621044815,-0.1171312332,0.5180973411,0.1331297606,0.3920755684,0.1230287179,0.3813702464,0.1814099699,0.1523991972,0.3915373385,-0.0962995142,0.0503208041,-0.0888623372,-0.0469534285,-0.0595205612,0.1010369137,-0.180071339,0.2891724408,0.0040014377,0.1104367599,-0.0288084,0.4164351225,0.226710692,-0.1072962135,-0.0603063405,0.2415639609,0.247913003,0.3363710046,0.243416816,0.2799069583,0.1374022067,-0.1132819131,-0.1982467175,0.0317735374,0.2913256586,-0.1701089591,-0.1376565844,0.0323030576,0.6526049376,0.1721726805,0.4123957753,0.2316215783,0.2014798969,0.2239348292,0.3061021864,0.1105275005,0.10288807,0.1833375841,0.444932282,-0.0468248688,0.1626707762,0.894777298,0.1134372428,-0.3787639439,-0.3137107193,0.066119507,-0.1021624058,-0.296834439,-0.0913482606,-0.096784398,-0.332601428,0.0546542965,0.0810169503,-0.1187318042,-0.5291758776,-0.4242374897,0.5176994205,-0.3008616567,-0.0995451584,0.4325691164,-0.0177163929,0.5202099681,0.5960307121,0.3152929544,0.0676189736,0.0706434473,0.4289952517,-0.2621090114,0.0018366459,0.2203690857,0.257060051,0.0967845321,0.0593927614,0.072272636,-0.0040696464,-0.0510766618,-0.0314323567,-0.0542352274,0.0711520463,-0.1385304481,0.4740909636,0.0246090647,-0.1848056465,0.0072492049,0.3113256693,-0.3124317825,0.1374709755,0.0962599441,-0.0826243013,-0.1098148152,0.11039076,-0.0001245058,0.1049865335,0.1413684785,0.4995455444,0.0022363982,0.0146101285,0.0369580239,-0.362864852,-0.2546319067,0.3638893664,0.2339007258,0.0720459297,-0.1506690234,0.54851228,0.4051665962,-0.228887409,0.0162167978,-0.526078403,0.2718011141,-0.1391752809,0.0799043551,0.1306475252,0.0314917825,0.0767936632,0.4003415704,-0.3214027286,-0.4901407063,-0.0951819643,0.1373241544,0.0201029107,-0.0601183549,-0.0938891992,-0.0482377484,-0.0940653309,-0.0181730725,-0.0797819868,-0.1192055494,-0.2002194226,-0.408672601,-0.1055347919,0.0288716983,-0.1562330276,0.3772963881,-0.1519808918,-0.118994832,0.1803653091,-0.195037365,0.1351660788,-0.1404505521,-0.5362731814,0.1425039321,-0.0960062668,-0.143671751,0.2505479455,0.2519667745,0.153834343,0.3209253848,-0.3875094056,-0.5163571239,0.6283265352,-0.2368851006,0.1326721013,-0.2854484916,0.2763117552,0.2571352124,0.1662466824,-0.5226237774,0.0320394933,0.2893506289,-0.0206170566,-0.3545163274,0.1928046048,0.0381172635,-0.1157735065,-0.1874916404,-0.3579196632,0.3021303117,-0.1233161837,-0.248265624,-0.0722471401]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1844","title":"Update Open Subtitles corpus with original sentence IDs","comments":"Merged in #1865, closing. Thanks :)","body":"Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!","comment_length":6,"text":"Update Open Subtitles corpus with original sentence IDs\nHi! It would be great if you could add the original sentence ids to [Open Subtitles](https:\/\/huggingface.co\/datasets\/open_subtitles).\r\n\r\nI can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat allowing for document-level machine translation (and other document-level stuff which could be cool to have); second, it's possible to have parallel sentences in multiple languages, as they share the same ids across bitexts.\r\n\r\nI think I should tag @abhishekkrthakur as he's the one who added it in the first place.\r\n\r\nThanks!\nMerged in #1865, closing. Thanks :)","embeddings":[0.2266675681,0.0961129367,-0.0610729493,-0.2533901036,-0.2080410421,0.1960118115,0.3298336864,0.288374126,-0.3152397871,-0.0051977979,-0.3032506704,0.1815488935,0.2368659973,-0.1456588954,0.0872383341,-0.3059326112,0.1514590681,0.0355428979,0.6395528316,-0.2217523605,-0.3472386003,0.2100997418,-0.0203145333,-0.040407151,-0.2054813355,0.3697547019,-0.4182570279,0.0875305757,0.038500987,-0.4456968904,-0.0368550755,0.5314381123,0.2634710968,-0.0058271796,-0.0001172084,-0.2237396389,0.2270976305,-0.2258344144,-0.2104466856,0.0876461118,0.0077032521,-0.2904906273,-0.258907795,-0.3330146968,0.0707928017,0.0541527048,0.0968478099,-0.3834755719,-0.2592775524,0.0822214559,0.0785769597,-0.0898821801,-0.0129200919,-0.1351801008,0.5450792909,-0.0601167046,0.2104171216,0.2701212764,0.6688584685,0.1796451956,0.2080600113,0.3877301812,-0.5008256435,-0.4227332771,-0.1936825961,-0.1063669771,0.0503375642,-0.229466036,-0.0437658392,0.1198043004,0.3974434733,-0.0692140609,-0.5729649663,-0.3557766974,-0.0282967333,-0.2307359725,-0.315723002,0.3064050972,0.3191663623,0.37535128,0.1376736611,-0.1544909477,0.13392663,-0.0220055431,0.3165286779,0.2053925097,0.28766644,-0.1493558586,0.1163444519,-0.0708056092,-0.2727918327,0.1916266978,-0.4007903039,-0.0199033264,0.0442712046,-0.1438712925,-0.0676596016,-0.5494454503,-0.1737187207,-0.0629545599,0.0869295746,0.1075489819,-0.1202870011,-0.0525174551,0.1011947915,0.3311165571,0.4152396023,0.1521139443,-0.1021027416,-0.0077195275,-0.1704122424,-0.0258095618,-0.2092092931,0.3025957942,-0.2825450301,0.2819241881,0.099867709,0.0674197227,0.0126796104,0.1064004079,-0.511336863,-0.0558979474,-0.1845504493,0.0004668288,0.1734020114,0.2913931906,0.2122107595,0.0734632909,-0.0050897622,-0.3970583081,-0.1062291414,-0.0061745029,0.0423392393,0.0014912017,-0.1395910829,-0.3595682979,0.2654149532,-0.24435848,-0.1832754314,-0.1689907461,-0.330345124,0.1506271511,0.0812588111,-0.2013196945,0.1199627593,0.0810546875,-0.1693928689,-0.135247618,0.0125557026,0.293541193,-0.1779156774,-0.375110954,-0.3833838701,0.1009605452,-0.3302427232,-0.2905547321,0.3775184155,0.3553792536,0.2064130753,-0.3001562655,0.3189355433,0.1502826363,-0.3006418049,0.1182799488,-0.242122218,0.5414482355,-0.0610909238,-0.038451422,0.3030451834,0.1473896205,0.1951971352,0.1444495022,0.4129763544,0.0275257174,0.0003948498,0.0660741031,-0.3778739274,-0.6018633246,-0.0082744444,-0.0036013857,0.0158954505,0.0532177277,0.3116686046,-0.0182921123,0.6112906337,-0.4156974852,-0.1293900311,-0.0533254892,-0.0238340087,0.1292007565,-0.1857534647,-0.167202577,-0.3289549649,0.0977385566,-0.0069717052,0.1578283608,0.401753664,0.1180433407,0.4730700254,-0.1156558096,0.108657904,-0.034971416,-0.0870690644,0.408950597,0.1758527309,0.2164292932,0.1580551863,0.0928553045,0.1233429238,0.2382349223,0.2992716134,-0.236653775,-0.0097862836,-0.0140773896,0.2960935235,-0.2421437055,0.0114945164,0.1366036236,-0.1965332478,0.2464746982,-0.4164170325,0.0919578373,-0.2004403174,-0.0331888199,-0.1106828451,-0.178469032,0.1721657068,-0.0126676774,-0.0701936036,0.3035718203,-0.1406551898,0.1030653045,-0.0914080814,0.4498193562,-0.1308388114,0.455176115,0.0491808616,-0.1426068097,0.3174208105,-0.3138044178,0.1425863802,-0.1595686674,-0.2105895579,-0.0736767799,0.0217557605,-0.0251588952,-0.0146513702,-0.1288327426,-0.1102885082,0.0028420975,-0.2820556462,0.0191051085,-0.4357894361,-0.2651578784,-0.0442738272,0.1097877845,-0.5248072147,0.1062591597,-0.0993931815,0.1774947494,0.1871041507,-0.2995183468,0.0107487952,0.2919065058,0.0286162477,-0.000517086,-0.0168404747,0.0285134036,0.0098488927,0.1857821941,0.1852181852,0.1710330844,0.5501621366,-0.334587872,0.2706845999,0.3707135022,-0.2330439985,-0.025050642,0.3103238046,0.2981428206,-0.1622712165,0.5508930087,-0.0093288673,0.0266478043,-0.0916920528,-0.0197964124,-0.1875453591,-0.4859907925,0.0851847008,-0.4665163159,-0.0228461958,-0.0198715962,0.2053230554,-0.2883334458,-0.3226696849,0.3237888515,0.1727171689,0.27791816,-0.1601648629,0.0836877674,0.0100620836,-0.4759066403,0.195547387,-0.0002310737,-0.2054438591,-0.137571618,0.0373517461,-0.2757684886,0.1489578784,-0.2719773054,-0.2380905151,-0.2841235697,-0.0282456335,-0.0897317752,0.3820701241,-0.2703148425,-0.1482572258,-0.3051162958,-0.1191422567,-0.0393779874,-0.1621567607,-0.2906461656,0.0300271343,-0.0097808214,-0.2012432963,-0.0699167475,-0.1537702084,-0.4188416898,0.4084514976,-0.0828578845,0.185544163,0.2949794531,-0.1227726191,-0.1895586848,-0.0554985106,0.1086567417,-0.0081752809,-0.542530179,0.0489866808,-0.2038503736,-0.0014222628,-0.2283306569,-0.1692972481,-0.2523999214,-0.3662465215,-0.0921796933,0.1908521056,0.2508549392,-0.219335109,0.5613185763,0.2374192029,0.0461464636,-0.0557127334,-0.0022120953,-0.1528980434,-0.2760706842,0.1041829288,0.4779337943,0.1433964223,0.0295027271,0.078185074,0.1570641845,-0.0499333218,-0.0210997351,0.0817258134,0.333078891,0.2103044987,0.2469362617,0.0270170067,-0.5501020551,0.156699121,-0.0884419307,-0.0458549671,0.5217340589,0.1794874668,-0.3193450868,-0.1200279444,-0.3871988952,-0.0730691925,-0.110883832,-0.1722304374,-0.2672497928,0.3219150603,0.2869455814,-0.1410707831,-0.1854616106,-0.110750407,0.1393930316,0.1020903811,0.5539689064,0.1212701797,-0.0175278913,0.1812206954,0.3587890565,0.181149438,0.4964658022,-0.1299861073,0.2829046547,-0.2740935981,0.1460861266,0.3384253383,0.2069538236,-0.1894409657,-0.3122293949,0.0403739214,0.0813099593,0.1108110398,-0.1659404784,0.2647279203,-0.4990411699,0.6232892275,-0.1732192039,0.0001296238,0.225525856,0.4704305828,-0.2995387912,-0.0888335258,-0.1592455655,-0.0118684461,-0.101259239,-0.6195725799,0.0295654777,-0.0029178811,-0.2289932817,0.1076930091,0.0051439297,0.0342102051,0.2037623823,0.0573490523,-0.1691962928,-0.054775089,0.6106162071,0.1040534303,0.1281930357,-0.1264776736,0.1612739414,0.2948854864,-0.4877707362,0.0504942201,-0.1687455177,0.0053600008,0.3654616177,0.3904620707,0.4715961814,-0.00190893,0.1809901148,-0.2773429453,0.0232704952,-0.0993089899,-0.1660382897,-0.0766707286,-0.0725416392,-0.5313318372,-0.0175539311,-0.3092061579,-0.0851463675,-0.6162431836,0.1560307592,-0.0761258453,0.3257183135,0.3903220594,1.0226155519,0.0994442776,-0.1257812083,-0.2100691795,0.0197743345,0.5192056894,0.1537477523,-0.0576546565,-0.1630829573,-0.0085630659,0.1035297737,0.1421118379,-0.2800079584,-0.2913126647,-0.2955609262,-0.0719867721,0.5186886787,-0.0733005628,0.2747086287,0.171430558,0.1932589114,-0.3108049631,0.3413510919,0.0285841245,-0.0284242667,0.169074133,0.0163438767,-0.3344206512,0.24747473,-0.0440864488,-0.280742079,-0.0849325061,0.0015238065,-0.136650905,-0.0617499128,-0.0111715831,-0.0575985797,0.427780509,0.3577673733,0.145366922,0.3714849353,0.0881448612,0.753452301,-0.2594947517,0.1695380658,0.1095667779,0.3496147394,-0.1961409003,-0.260003984,-0.0523266643,0.1819554418,-0.1913557947,-0.2484266013,-0.0396011397,0.0032073043,0.1616085321,-0.0301974732,0.1640594602,-0.1351219565,-0.2264246643,0.0181043334,0.0418283045,-0.1059266925,-0.138936758,-0.1917280108,-0.2258468121,-0.1563274115,0.2722494304,0.0079279905,-0.4374083877,0.1792270094,0.3943561018,-0.0201227665,0.1592507511,0.0406106673,-0.5578819513,-0.3435297012,0.104804948,-0.2584578395,-0.1778627783,0.0130444495,0.0347298235,0.0469400287,0.3125509024,0.1187940314,0.1106939912,-0.4963864386,0.1644187421,-0.3170605302,0.1706574261,-0.0695773959,-0.4217741489,-0.0590875149,0.0743329227,-0.2172401249,0.116861403,0.0617718548,0.0199468639,0.5759844184,-0.1009899005,-0.3107156456,-0.3178302646,-0.0096070413,0.1827659756,0.1530735791,-0.0516670682,-0.2952768803,0.1716642678,0.0219101589,0.2288182676,-0.0610149875,-0.0787936375,0.1633898169,-0.1415265948,-0.1805848628,0.3002555966,-0.0022119624,0.3615913391,0.1895528138,0.1226974353,0.0916780233,0.1293099523,0.1409707069,0.2155978233,-0.1645562351,0.0642111227,-0.2125830203,-0.1224615648,0.532212913,0.1072477251,0.362480551,0.1254189759,0.3303294778,0.2054566741,0.2309922427,0.4232214391,-0.2021824569,-0.0157645028,-0.1777436435,-0.1146001518,-0.0075071543,0.1395908296,-0.1642387509,0.2929917872,-0.1143338084,0.0716730654,-0.1471979916,0.3654154241,0.2147885859,-0.1752768159,0.0816737637,0.3114589751,0.1478606164,0.183692798,0.2009969354,0.3445523977,0.0876538604,-0.1505468935,-0.2444978505,-0.1490015984,0.2239199728,-0.119268477,-0.0689075738,0.0027119527,0.6447910666,0.3033047915,0.3264443874,0.2585507929,0.174790591,0.1593655795,0.1561409086,0.1522556841,0.2167493701,0.1003121883,0.4630613327,-0.0177034624,0.1171132103,0.8268604279,0.0690511093,-0.3161115646,-0.3639288843,0.0951005369,-0.0369135812,-0.2817739248,-0.0380720347,0.1552259475,-0.4117537141,0.1102225482,-0.0289416183,-0.0515502132,-0.5228483081,-0.4769150913,0.5894099474,-0.2835116088,-0.1462560296,0.3292131424,-0.0020673554,0.4966101944,0.5852130651,0.2478203923,0.0709975511,0.0958108902,0.4321252406,-0.1910568178,-0.0157371126,0.2315895557,0.2678388655,0.2616108656,0.136600107,-0.0109059056,0.026673086,-0.0787223354,0.098418802,-0.085030131,0.0467135385,-0.0543701947,0.4417090714,0.212637037,-0.1919662058,-0.1084343642,0.3064401746,-0.226550132,0.1862792671,0.0009973609,-0.0558624975,-0.1616747975,0.087808758,0.0192022081,0.1698750854,0.1479251534,0.4875727296,-0.0344739072,0.0698390678,0.0283550359,-0.2923331857,-0.1981783211,0.3184305131,0.2689498067,0.1113547608,-0.2175956666,0.4823172987,0.3598522246,-0.2960387468,-0.1094464287,-0.4749259353,0.2489556074,-0.1503498256,0.0490289889,0.2394469082,0.0674564242,0.0308323596,0.365542382,-0.1984442025,-0.4743885696,-0.0372800082,0.1954960525,-0.0552057065,0.0754246414,0.0747180134,-0.0165859442,-0.1154499874,-0.1058367565,-0.1486495137,-0.047807388,-0.1633820981,-0.4088921845,-0.0779222101,0.0232410934,-0.1414065063,0.3535666168,-0.0125414934,-0.1746035069,0.1853125691,-0.1614974141,0.0823777914,0.0047251405,-0.5624653101,0.0200219788,-0.0190261491,-0.1037326604,0.2624367476,0.0823912323,0.1709932387,0.2272083014,-0.491686821,-0.5078980327,0.5372983813,-0.1994408071,0.1119079441,-0.2724790573,0.3373183608,0.2231881768,0.1182097495,-0.4816975594,-0.0626672581,0.2790354192,-0.0346917473,-0.367934972,0.2098010927,0.1251184195,0.0039397152,-0.1557531506,-0.3659047484,0.3538681567,-0.1658850908,-0.2937555015,-0.0462321118]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Hi @patrickvonplaten I would like to work on this dataset. \r\n\r\nThanks! ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":11,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHi @patrickvonplaten I would like to work on this dataset. \r\n\r\nThanks! ","embeddings":[-0.3252358735,0.2317551672,-0.1150876209,0.031390693,-0.0583076514,-0.0101283202,0.0463694409,0.1522034556,-0.2828116417,0.1391946375,-0.2658597827,-0.1802043915,-0.1007143632,0.1219184101,0.1699670404,-0.1249362007,0.0982453898,0.0200028159,-0.3483632803,-0.2926430702,-0.113005124,0.025028659,-0.11403732,-0.1677553654,-0.249308601,-0.084705539,-0.3333165646,-0.1959509999,-0.3096325397,-0.1792253405,0.0768076628,0.1868458092,0.1147160083,0.4253309965,-0.0001053406,-0.1908281446,0.2119104266,-0.2452110201,-0.3276067376,-0.1459660083,-0.2834720314,-0.2253167331,-0.6151729226,-0.0266561601,-0.278821677,-0.1962692738,0.0227497853,-0.0080628237,0.1928633004,0.63523525,0.221903041,0.0711576194,0.2239126861,-0.1675584316,0.2597446442,0.1130032837,-0.1754041463,0.0599878058,0.2985134721,0.1111912653,0.079766199,0.3830013871,0.0262215715,-0.2354392856,-0.0226389356,-0.3774754405,-0.1484856606,-0.3440245688,0.1389355361,0.4007177353,0.6692109108,-0.1782402247,-0.1420404315,-0.1054142267,0.2476442009,-0.1427988559,-0.0327653214,0.0090220161,-0.136634782,0.2949760556,0.1480698287,-0.2520001829,-0.2070117891,0.3224077523,-0.0013260955,0.348370254,-0.0858643502,-0.0818349719,0.0784106702,-0.3064756691,-0.2852117419,-0.0746058598,0.0094414391,0.1182609797,-0.2716406882,-0.3252157569,0.2960608304,-0.1249320433,0.1420637965,-0.3202508688,-0.2072568387,-0.0842453837,-0.1804461926,0.0329260342,0.0487757064,0.3608903885,0.0349349268,-0.0145834638,0.0593756326,-0.3025403619,-0.0267368257,0.1213758811,-0.0318470895,0.1318352968,-0.3857717812,0.2605239153,-0.0040037767,-0.1809806228,-0.280677557,0.1154547557,-0.0511341095,0.0814030021,-0.0199513901,0.3654233515,0.1019674689,0.3160405159,-0.0987796262,0.3779761493,-0.0256130788,-0.1428061277,-0.0937397107,0.1761652082,-0.2416591197,0.101327531,0.2614851296,0.4330510795,0.0786162913,0.0801676437,0.2170094103,0.0601927117,-0.1789112091,-0.098340556,0.3212031424,-0.0077991127,0.0095617417,0.0705680326,-0.1993501335,-0.1756005287,-0.2367872745,0.05665683,0.1484333426,-0.0369592421,-0.2406645119,0.2562040985,-0.1418921351,-0.2206034809,-0.266467005,0.7312400341,-0.1449717432,-0.0662569106,0.0842465609,0.1134038866,-0.4138914943,0.0436527058,-0.0655789301,0.3331822157,-0.3128215373,-0.0064183837,0.0777387321,0.0443874598,0.0745785683,0.1872344464,-0.0256681051,0.3265066743,0.0101590836,0.3267500997,0.0831607953,-0.3654386699,-0.3595517278,0.1627316922,-0.1524015367,-0.0427294746,0.2522424459,-0.1276650876,0.6115841866,-0.2735266984,-0.0894004852,0.4217371345,-0.0540224016,0.0955884233,-0.1409783512,-0.3009580672,0.2889202237,0.150836423,0.0073043704,0.0217620321,-0.0184933711,0.3160351813,0.2818931639,-0.3174212873,0.2583468854,0.3189358413,0.0762697011,-0.1889031082,0.046614375,-0.1800592244,-0.1400978267,-0.0832197517,0.1910083443,0.5284779668,0.0903524756,-0.1888991594,-0.1674448103,-0.2145587057,-0.1710858643,-0.1570858508,0.2655880451,0.3211666644,0.1073440686,0.0226313174,-0.0696652755,0.2938750982,-0.3539303839,-0.2640421987,-0.0591576621,0.2875662148,0.0617442392,-0.3004541695,0.2471022755,0.4402299523,0.2418808937,-0.0384007804,-0.0348386839,0.0567716621,-0.224748686,0.2154184878,0.0984899923,0.2100404352,0.2333219647,-0.5387427807,0.1867393404,0.3271800876,0.1702099591,-0.2570531964,-0.1208795905,0.2038626671,0.28657493,0.1439254284,0.0515282489,-0.0601842925,0.4473277926,-0.044491969,-0.1174186915,-0.0474835858,0.1665917337,0.2720105648,-0.1291771084,0.0292000379,-0.2810339034,0.2077922374,0.5577809811,-0.150267899,0.2838072479,0.2025559843,-0.0321190432,-0.1471167207,0.0506783836,-0.2231017649,0.2560169399,0.2950713634,0.2904379368,-0.0130779706,0.1838820279,-0.0538282357,0.176537469,0.0168807022,-0.2383036315,0.1503623873,0.1650748104,-0.1328569055,-0.3251347244,0.0302239507,0.0179522801,0.2464697361,-0.1195606366,-0.3102584779,0.0517587513,-0.2014872134,-0.214497,0.0272333324,-0.17559053,-0.0037450138,0.3587637246,0.040934965,-0.1795310974,0.2778262794,0.2756908834,0.2567353249,-0.3953422308,-0.046821408,-0.0484784544,-0.4324026704,-0.177452296,0.2087559253,0.1871005148,0.251632303,0.1937296391,-0.1379779875,0.2160965204,-0.235275358,-0.350802362,0.1563293934,-0.259726882,0.1296606362,-0.2056757957,0.0402600542,-0.0239420515,-0.480125159,-0.2061776817,0.021698609,-0.0749441758,-0.0207713451,-0.4632716775,-0.1125666052,-0.00426492,-0.4758831263,-0.4442295134,-0.4368864596,0.3516378403,-0.2771875262,-0.0050511458,0.0255708471,-0.0045862757,0.1052760631,-0.0367466547,0.207755059,-0.2988173068,-0.1852051169,0.0888729542,-0.255315125,-0.3119958937,-0.246262446,-0.0280780159,0.1244899333,-0.0891682804,-0.2579401433,0.2285364568,-0.1514382362,0.3486223221,0.1522127986,0.0232235789,0.2420250028,-0.1500999779,-0.1329706162,-0.2150990218,-0.1945963055,0.1010482237,0.3908594251,0.2422583848,0.0133043099,0.0703058094,0.0218688436,0.586392045,0.1737699509,-0.0315079428,0.2380056381,0.0480913259,0.3574700654,-0.2079113722,-0.4680106938,0.5076212883,0.1132701263,-0.0091350581,0.4611871541,0.0069482261,-0.4669715762,-0.0857212469,-0.2822850347,-0.2654619813,-0.2488132119,0.3471039534,0.137162298,0.1765658557,0.0019097088,-0.1701016873,-0.3031850159,-0.1907344311,0.2088171244,0.1002636924,0.0206379592,-0.0605214387,-0.2106287479,-0.0514795929,-0.4207980931,-0.0692154765,0.1510304064,0.3278211057,-0.5225634575,0.0928999484,-0.057438951,0.1273660511,0.3809527755,-0.3608942628,-0.024752697,0.2979223728,0.5242812037,-0.0329076983,0.2314738333,-0.2796304822,-0.2927841544,0.6428184509,0.3276265264,-0.3227493465,-0.0102839107,0.160190165,0.3519277871,0.0165210981,-0.1711612791,0.1097740903,-0.3165455461,-0.3566650748,0.2902503312,0.5057368279,0.0370456725,0.2248194814,0.0931768641,0.1550894231,-0.0542238168,0.4874705374,-0.0598288067,-0.0406981856,0.5257422328,0.1390163898,-0.0350365266,-0.1157075167,-0.1331234276,0.2494657785,-0.0636177212,-0.3425088525,0.0783956721,-0.2893178463,0.4163156152,0.3416365683,0.1780306399,0.0484135784,-0.1591749936,-0.0215215348,-0.2354768217,0.2219952792,0.2301195562,0.2147311419,-0.3236573637,-0.2807464302,0.3742752671,-0.0077389278,-0.152735889,0.0791780651,0.1498393118,-0.2292791456,0.2664255798,-0.16473113,1.2589037418,0.267832756,-0.0241986625,0.1904282421,-0.2968116403,0.0404320993,-0.4022964835,-0.1608626097,0.2481238842,0.1416180581,0.0160677861,0.2071271539,-0.0262818877,0.2639268339,-0.4657739103,0.2198962867,0.0029370331,0.0912618116,0.1650131345,0.4521242082,-0.1491527557,-0.3923906088,-0.3389139771,0.2582337856,0.1002044231,0.1350788772,0.1591887176,-0.2785897553,0.1670818329,-0.0254665911,-0.6498436332,0.1594258249,-0.2534203529,-0.1268800646,-0.4293371737,-0.1505884528,0.3421536684,0.1754689962,0.4531047642,0.0308250673,-0.2210153192,0.2507046163,0.0494986251,0.0747125074,0.1660192907,-0.0821827501,0.377379626,-0.310601294,-0.3498829305,0.1722042561,0.1191675439,0.0005175723,-0.3766294718,-0.2563710809,0.134979859,0.0733864754,0.2136080712,0.6085217595,-0.0157199763,-0.3480754197,0.200313583,-0.0566948764,-0.4105423391,0.2483098805,-0.2254187614,-0.2836976647,-0.1779119223,0.2085810304,0.2565142214,-0.0002839529,0.377898246,0.2816507816,-0.1897423714,-0.3974358141,0.0000780557,0.1965750456,-0.5117422342,-0.0396984704,0.10953933,-0.165944308,-0.2645043731,0.2890134752,0.0558197238,-0.0210036244,-0.0427723788,-0.084660694,-0.2768094838,-0.0911268592,0.1100521609,0.0800732896,-0.12952061,0.2981410325,0.0604834817,-0.0288414694,-0.4356740117,0.0645204186,-0.1415490508,0.2692892253,0.2509800494,-0.1958243847,0.1537632346,0.0305427704,0.1638659239,-0.0574734174,-0.135766983,-0.2449367344,-0.1832194924,0.0876250118,0.0959978476,0.2219024301,-0.0018211385,-0.1235598847,-0.1879384816,0.066659309,0.311465323,0.2169691771,-0.0461621545,0.2016122192,-0.0375757888,-0.0027609346,0.1760403812,-0.0453467518,0.3013708889,0.2822418511,0.0670144781,0.1858175099,0.0038757476,-0.1778298616,0.0211065393,0.2847496867,0.1202277169,0.0588339716,-0.0162875149,0.1201375648,0.1068463996,0.3955292404,0.0188418645,-0.0272217281,-0.0326079875,-0.0640291125,0.3410852551,0.276966095,-0.1924555302,0.2138824761,-0.1667295545,-0.2658732831,0.2044311911,0.0703783408,0.0338494033,-0.1026023105,0.3154390156,-0.1256915629,-0.1558418125,0.0203059148,-0.0117223104,0.2038015127,0.087327309,-0.09296152,0.0971725136,0.0599302985,0.0605161712,0.0839569345,-0.1295386106,0.5030611157,0.1053845957,0.1237734556,0.2765104175,-0.0921485499,0.1274425685,0.1462565213,0.0614778027,0.2124518901,0.0668025985,0.2968791723,-0.0060236435,-0.2971627414,-0.0919908583,0.2316410393,-0.0208960027,0.0068411347,-0.0736488551,-0.1743391752,-0.2662583888,-0.0898543447,-0.0118021332,-0.2356344014,-0.0354976915,0.1640803367,-0.1289979815,-0.1175634861,-0.1293547899,-0.2455948591,0.1412427127,-0.2579570711,0.3941335976,0.1573672742,-0.0622291826,0.2545884848,0.4983292222,0.4850618839,-0.0353095382,-0.1974170953,0.2359006256,-0.0595958605,-0.1955352277,0.0000796194,0.4755224288,0.0525792204,0.162395075,0.2062884569,0.2292785645,-0.2021404207,-0.2196927816,0.1794706136,-0.1324526221,-0.129275918,0.1450049877,-0.0947089046,0.1165913939,-0.5573278666,-0.1924097538,-0.4859951138,0.0183403771,0.3966507614,-0.3189879954,0.1124765053,-0.151795581,0.1256681085,-0.2423669696,0.3077968657,0.2714524865,0.0520967394,-0.1639000177,-0.158607319,-0.3952848315,-0.0936686695,0.1422474831,0.0792995021,-0.3046785593,0.0781578124,0.4660860896,0.3079106212,-0.1014494523,-0.0727731511,-0.3677815199,0.1206787229,-0.4073684812,0.2045165598,0.0590264574,0.5437811613,0.1031462103,-0.1334346682,-0.2211433947,0.2984727323,0.0959752873,-0.1223821566,-0.0994634852,-0.1797263473,-0.2241125703,0.1319277734,-0.0021699367,0.5371836424,-0.2939719856,-0.0072043398,-0.1125286669,-0.1230541021,-0.1556865722,0.0096677504,-0.0201004427,0.2755509913,-0.1324941367,0.3488264382,0.0223284718,-0.1428486258,-0.0195266269,0.0580572523,-0.2546382248,0.1428652704,-0.1483009607,0.0295622423,-0.1703567207,0.2236115783,0.0488572642,-0.010205688,-0.2460256815,-0.2235571146,0.4589654803,-0.1154457852,-0.4038242698,-0.1147400439,0.375997901,0.0761377886,0.0504223369,-0.6893838644,0.2311622351,0.2644991577,0.1367573738,0.0704144016,0.191973865,0.0991934836,0.0309507251,-0.0851603895,-0.0759379715,0.1104019582,-0.2112125605,0.1174541712,-0.3179817796]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"That's awesome! Actually, I just noticed that this dataset might become a bit too big!\r\n\r\nMuST-C is the main dataset used for IWSLT19 and should probably be added as a standalone dataset. Would you be interested also in adding `datasets\/MuST-C` instead?\r\n\r\nDescription: \r\n_MuST-C is a multilingual speech translation corpus whose size and quality facilitates the training of end-to-end systems for speech translation from English into several languages. For each target language, MuST-C comprises several hundred hours of audio recordings from English TED Talks, which are automatically aligned at the sentence level with their manual transcriptions and translations._\r\n\r\nPaper: https:\/\/www.aclweb.org\/anthology\/N19-1202.pdf\r\n\r\nDataset: https:\/\/ict.fbk.eu\/must-c\/ (One needs to fill out a short from to download the data, but it's very easy).\r\n\r\nIt would be awesome if you're interested in adding this datates. I'm very happy to guide you through the PR! I think the easiest way to start would probably be to read [this README on how to add a dataset](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md) and open a PR. Think you can copy & paste some code from:\r\n\r\n- Librispeech_asr: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/librispeech_asr\/librispeech_asr.py\r\n- Flores Translation: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/flores\/flores.py\r\n\r\nThink all the rest can be handled on the PR :-) ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":188,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nThat's awesome! Actually, I just noticed that this dataset might become a bit too big!\r\n\r\nMuST-C is the main dataset used for IWSLT19 and should probably be added as a standalone dataset. Would you be interested also in adding `datasets\/MuST-C` instead?\r\n\r\nDescription: \r\n_MuST-C is a multilingual speech translation corpus whose size and quality facilitates the training of end-to-end systems for speech translation from English into several languages. For each target language, MuST-C comprises several hundred hours of audio recordings from English TED Talks, which are automatically aligned at the sentence level with their manual transcriptions and translations._\r\n\r\nPaper: https:\/\/www.aclweb.org\/anthology\/N19-1202.pdf\r\n\r\nDataset: https:\/\/ict.fbk.eu\/must-c\/ (One needs to fill out a short from to download the data, but it's very easy).\r\n\r\nIt would be awesome if you're interested in adding this datates. I'm very happy to guide you through the PR! I think the easiest way to start would probably be to read [this README on how to add a dataset](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md) and open a PR. Think you can copy & paste some code from:\r\n\r\n- Librispeech_asr: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/librispeech_asr\/librispeech_asr.py\r\n- Flores Translation: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/flores\/flores.py\r\n\r\nThink all the rest can be handled on the PR :-) ","embeddings":[-0.2875134349,0.1752746999,-0.0523378514,-0.0160808023,-0.1027401313,0.0413462967,0.0322461016,0.1573132724,-0.3544582129,0.2396767288,-0.2381985486,-0.11117699,-0.1982450634,0.104688406,0.1299680322,-0.1173933297,0.0626323596,0.1621661633,-0.2575624287,-0.2426289916,-0.0848404765,-0.0210144799,-0.167618081,-0.2154306471,-0.2636874616,-0.2343586236,-0.3444107175,-0.0222435165,-0.2228296101,-0.1175916269,0.1175155267,0.248413533,0.2399591357,0.4444422722,-0.0001051568,-0.192142427,0.1900986284,-0.323317498,-0.3289248049,-0.0075077889,-0.2392946482,-0.1538301855,-0.5299216509,-0.0503164232,-0.0830948502,-0.3202906847,0.0407133475,0.1645170152,0.2044690251,0.5778822899,0.2140561491,0.1666963398,0.3482290804,-0.1188550666,0.2372097373,0.1014449,-0.1206199601,0.1359511912,0.3214258254,0.0502922796,0.129828766,0.3519296646,0.0751625523,-0.2078210413,-0.0013175285,-0.2970611453,-0.1703730822,-0.4697276354,0.0546199232,0.3466715813,0.5668892264,-0.2300912887,-0.1379353851,-0.1885658205,0.2385323197,-0.2753940225,-0.14257963,0.1017543897,-0.2468654662,0.2201553732,0.1338834018,-0.1983614564,-0.2617568374,0.4235546887,-0.0393225178,0.2688905299,-0.0210252348,-0.0664190501,0.1583765447,-0.2404517382,-0.2738946676,-0.022596024,0.1015821248,-0.0083620613,-0.2182196379,-0.3157845736,0.2978577018,-0.1357144415,0.1732509285,-0.3756062686,-0.1892125905,-0.1522246748,-0.168646872,0.0308303311,0.1053945199,0.3480731547,-0.1274257898,0.1075954363,-0.0360942595,-0.1672763228,0.0026849431,0.212713331,-0.13084203,0.1524341702,-0.3370411992,0.258107692,-0.1394635141,-0.1006495059,-0.2270867676,0.1519316584,-0.1549450308,0.1976893991,-0.0497684442,0.1678732038,0.1502224654,0.4123967588,-0.0852502957,0.3266165555,0.0435949676,-0.1859340668,-0.0500112697,0.1548000574,-0.2542715967,0.2566569149,0.2378215045,0.3129874766,0.1783684045,0.005379519,0.3276020586,0.0617621541,-0.0991261825,-0.0511569791,0.2883894145,0.1007132158,0.1363552213,0.0000261789,-0.2535656989,-0.1737268865,-0.2774477899,0.0093008215,0.1357427686,-0.1794618666,-0.1582837254,0.2379549891,-0.1018656865,-0.1741172224,-0.3040414751,0.7007139921,-0.0606367476,-0.1210627779,0.0528959334,-0.053221371,-0.3657281697,0.034251757,-0.0872518718,0.3126232922,-0.2872094214,0.0393940248,0.0574680567,0.0726874322,0.1355343014,0.1585348696,-0.1287816465,0.3593234122,-0.0331008919,0.3460199833,-0.0488750264,-0.4397496879,-0.281599164,0.1535723209,-0.2127689719,0.0960734412,0.1783140153,-0.1073975861,0.541461885,-0.3718987107,0.0349775143,0.3706716597,-0.1727273762,0.0340493694,-0.1894364953,-0.284686178,0.0788009986,0.1839009672,-0.0323208496,-0.0800227076,-0.043562822,0.3727610707,0.2667350471,-0.3871524036,0.2202354372,0.2563530207,-0.044742398,-0.0748777464,-0.0283267424,-0.1960770339,-0.1836107224,-0.0590261631,0.1912146062,0.6219031215,0.1550298035,-0.2531658709,-0.1419557184,-0.2369056642,-0.1443944573,-0.0620132796,0.2281483561,0.3150978386,0.1107614115,-0.1125979275,-0.0861270577,0.2071199417,-0.5647342205,-0.2647086382,-0.1538940817,0.1814260632,0.0084680552,-0.2195859849,0.308077991,0.4227020144,0.0146599868,0.0535548367,0.0285549555,0.1354120821,-0.1619793475,0.2579609454,0.078119278,0.1648444086,0.2727852464,-0.5677766204,0.1332100034,0.4646582901,0.0776263028,-0.178242594,-0.0820300952,0.1244486943,0.2989769876,0.1829657406,0.1640458554,-0.1839011312,0.1819776893,-0.0004319315,-0.2123808265,-0.0435389206,0.1498964578,0.1638146639,-0.2349280119,0.0365414806,-0.354216665,0.2235673368,0.4518024623,-0.2263447344,0.264849931,0.3474594951,-0.0424363911,-0.1756834537,-0.0004732814,-0.1034013107,0.3319273591,0.3627647161,0.2422538549,0.007105187,0.0717024654,-0.0730774924,0.1388951689,0.1237823069,-0.1127777919,0.2940680385,0.1811237335,-0.0422680788,-0.3254457414,-0.0892077833,-0.0382538624,0.1844967455,-0.1044627056,-0.2978939712,0.0133783473,-0.2083917111,-0.2377456725,0.0129815377,0.0028975399,0.0303690974,0.3540036976,0.0479798242,-0.2579814494,0.2546873689,0.2618288696,0.2677576244,-0.3453737199,-0.0791167691,0.0545896851,-0.4009774923,-0.3158776462,0.2158412486,0.3016784489,0.2556003034,0.1860269308,-0.1914077252,0.1003520489,-0.1630121171,-0.396995306,0.1623779982,-0.3061778545,0.061260324,-0.1400402337,0.1018904224,0.0210825074,-0.58485502,-0.2191078663,0.1448171884,-0.1159877256,-0.1049957797,-0.5161013603,-0.1699137837,-0.0485322475,-0.6011019349,-0.4975522459,-0.4463849366,0.2768109739,-0.1984015256,0.0573747233,0.0979484841,0.1252353191,0.0677642152,0.0089160195,0.1787888706,-0.2338744849,-0.0812643766,0.1727400571,-0.2196901143,-0.29603526,-0.2503241599,-0.1269465983,0.0345077701,-0.0453706533,-0.3076547682,0.1568709761,-0.1172697768,0.1862730086,0.0695896372,0.1646303535,0.3058905602,-0.0366118848,-0.1327184588,-0.1065293923,-0.1093277112,0.0363105051,0.3674372137,0.2547686398,0.0032196797,0.0030396425,-0.0236488804,0.5740106106,0.2956697643,-0.121155113,0.1002704576,0.0473234951,0.2798265517,-0.1951166391,-0.3320175409,0.4544562399,0.054715924,-0.0594976656,0.5838135481,0.0157409869,-0.4595153332,-0.0494788624,-0.2481126636,-0.2401512861,-0.2929203808,0.3575483561,0.1684376448,0.2732201219,-0.073899731,-0.2694478631,-0.2592241764,-0.2080506831,0.1178256869,0.1426766962,0.0319772996,0.035044495,-0.2550529838,-0.0148428502,-0.240407303,-0.0514478162,0.3507369161,0.266936481,-0.3556983173,0.0085828118,-0.1341502368,0.1201198921,0.2352528423,-0.3290524483,-0.1241108924,0.1691449434,0.6236915588,0.058091417,0.2449863404,-0.2782992125,-0.4077101648,0.5713484883,0.32709378,-0.3180730939,0.0295583885,0.1805692762,0.4551047683,0.0196315851,-0.1199941561,0.0669296756,-0.418856591,-0.3954565823,0.2638201416,0.4192442894,-0.03022095,0.144436419,0.0513747446,0.273914963,-0.0993725061,0.5432238579,0.0444293357,0.0042275051,0.5168458223,0.0652425364,-0.0213872418,-0.1969531775,0.1559512019,0.2522524297,-0.0759334862,-0.2559378147,0.0856123567,-0.1964671761,0.3721098602,0.386441499,0.1251036078,0.01525106,-0.0959912911,0.0340302214,-0.2213311642,0.3231512904,0.1718505621,0.0787984282,-0.2946381569,-0.1797348708,0.4043148458,0.0214406159,-0.0749257207,0.1412546039,0.1843288988,-0.2057354152,0.3513554931,-0.13850905,1.2044336796,0.2380200773,0.0568892471,0.2659080029,-0.1096374765,0.0883284211,-0.4995722175,-0.0849643648,0.3907920718,0.259765625,-0.0190223362,0.2098603249,-0.1178991795,0.2537156045,-0.4399802983,0.2426095605,0.1150812134,-0.0074011986,0.2306423485,0.5852320194,-0.1713493019,-0.5080519319,-0.2403679192,0.2198496461,0.0945750773,0.0671871379,0.1623477489,-0.2743901312,0.1177940071,-0.0123203462,-0.6150582433,0.0930603147,-0.2432058603,-0.2029714882,-0.4296462536,-0.1472198069,0.3799270988,0.2104634792,0.2985901535,0.0168839339,-0.2263958156,0.1732626259,0.0865626112,-0.0223947186,0.0827471241,-0.172946915,0.4044258296,-0.2710454762,-0.2950927019,0.2002263963,0.0704178512,0.0133646913,-0.4002252221,-0.260035038,0.2456319481,0.1467288882,0.2403481752,0.6309653521,-0.0359696895,-0.2770215273,0.1873919815,-0.1814451665,-0.3884576857,0.2241791636,-0.276250869,-0.1189770699,-0.2200241536,0.0615964308,0.2748681009,-0.0635086,0.3337149322,0.322540313,-0.0669055879,-0.3963093162,0.0304450709,-0.0104544992,-0.4409249723,-0.0201865453,0.0905111358,-0.1279532909,-0.2117040753,0.4416200519,0.2073620558,-0.0716666877,-0.014258041,-0.046450559,-0.3322637379,-0.0164139308,0.0912043974,0.1912334263,-0.0992760509,0.3792113066,-0.032559175,0.1184500456,-0.4336080253,0.0761987269,-0.1626812965,0.2437880784,0.2632617652,-0.142660886,0.2488957494,0.0414087884,0.1591390073,0.0102063464,-0.0619051903,-0.2223884016,-0.1657424122,0.0845461488,0.1184759662,0.1900925338,0.0874157548,-0.0435399264,-0.1722626835,0.1013928205,0.2535380125,0.1184065044,-0.0353023447,0.1590712368,-0.0247711968,-0.0573767647,0.1267610639,-0.0222570729,0.3171494305,0.3422956467,0.0590778254,0.1731783003,0.0098036109,-0.2161372304,-0.0024720067,0.3561342657,0.2103466392,0.0230447762,0.0413769148,0.157468006,0.0288916696,0.235168308,0.0491055958,0.054847829,-0.0470543243,-0.0214708913,0.2766654789,0.2646977603,-0.1533526033,0.2425720245,-0.2537624836,-0.3012732863,0.1291771382,0.1017609239,0.103192918,-0.131192416,0.2717269063,-0.1053742915,-0.2317266017,0.049028866,-0.0357234813,0.2377203405,0.1284561455,-0.0859471336,-0.056269519,0.031223435,0.2207509875,-0.0386748388,-0.165040791,0.5045075417,0.0553666204,0.0940549001,0.2753019631,-0.1637172997,0.0964045525,0.0568495914,0.219476521,0.1539333165,0.0517207198,0.4231321812,0.0482201129,-0.3117497265,-0.1460020244,0.2333996743,0.1063835546,0.0108453315,-0.2279701829,-0.2169210017,-0.2533043027,-0.0314634517,0.0169393104,-0.3470692933,-0.0472034924,0.158401072,-0.1348187476,-0.1955413967,-0.1827453673,-0.2227243036,0.1587722003,-0.1729852408,0.3621605337,0.2358754128,-0.1519256681,0.1546730399,0.4195310175,0.5034295917,0.0164715555,-0.2659992874,0.2552727163,0.0722158477,-0.1134966612,0.0042057452,0.3613189161,0.0438305251,0.0530941263,0.238330856,0.2213024348,-0.2607592642,-0.2357934117,0.0998978689,-0.1937328279,-0.1134302467,0.092946291,-0.0539523959,0.190490365,-0.4325957298,-0.0372578725,-0.6131446958,0.1492977291,0.5136852264,-0.23913908,0.0791094601,-0.1282246411,0.1162267923,-0.3067201078,0.3376326859,0.3489734828,0.1395462155,-0.2194070667,-0.2014395297,-0.320910573,-0.0147350254,-0.0631304309,-0.1013231575,-0.3647857904,0.0981366411,0.4524946809,0.3658346236,-0.2682141364,-0.2221841514,-0.3807547092,0.182752341,-0.314142704,0.2570056021,-0.0253152251,0.4205223322,0.0902483016,-0.241368413,-0.2607533634,0.1864530146,0.0894530937,-0.1324123293,-0.0881005153,-0.1244737357,-0.0515834801,0.1009027138,-0.0401927195,0.5253703594,-0.2989623249,-0.1234956533,-0.1656338423,0.0010451504,-0.1861561984,-0.0595180951,-0.0427821651,0.3104833663,-0.222074002,0.391545862,-0.0350499339,-0.2125217468,0.1208043247,0.0612907112,-0.2969560027,0.1123981103,-0.2163973749,0.0407533795,0.0370623656,0.3508244455,0.0020811332,-0.0512159318,-0.2930359542,-0.220029667,0.512966454,-0.1426170468,-0.2901852429,-0.0267886296,0.3327946067,0.1220952123,0.1083657891,-0.688135922,0.2756843269,0.2302373797,0.1484369785,-0.0474420413,0.1926651448,0.058399789,-0.0780559629,-0.0438356958,-0.0487915799,0.0565764904,-0.219215855,0.0728060976,-0.3567443788]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Hi @patrickvonplaten \r\nI have tried downloading this dataset, but the connection seems to reset all the time. I have tried it via the browser, wget, and using gdown . But it gives me an error message. _\"The server is busy or down, pls try again\"_ (rephrasing the message here)\r\n\r\nI have completed adding 4 datasets in the previous data sprint (including the IWSLT dataset #1676 ) ...so just checking if you are able to download it at your end. Otherwise will write to the dataset authors to update the links. \r\n\r\n\r\n\r\n\r\n","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":90,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHi @patrickvonplaten \r\nI have tried downloading this dataset, but the connection seems to reset all the time. I have tried it via the browser, wget, and using gdown . But it gives me an error message. _\"The server is busy or down, pls try again\"_ (rephrasing the message here)\r\n\r\nI have completed adding 4 datasets in the previous data sprint (including the IWSLT dataset #1676 ) ...so just checking if you are able to download it at your end. Otherwise will write to the dataset authors to update the links. \r\n\r\n\r\n\r\n\r\n","embeddings":[-0.4057501853,0.1195164844,-0.0039104824,0.1509025097,0.0220463183,-0.0853217021,-0.1335308552,-0.0062558954,-0.2400020957,0.1257295012,-0.2750507295,-0.3189477026,0.1023326293,0.1398473829,0.1449174583,-0.0939304903,0.120543018,0.025905041,-0.3361466527,-0.2657552063,-0.0817216411,0.0491207801,-0.2169910222,-0.288932234,-0.2452686727,-0.0799722299,-0.3513832688,-0.1246387288,-0.4012856185,-0.1996175051,0.1150277704,0.2857916951,0.0613894835,0.3938317895,-0.0001092176,-0.1187215894,0.2797668576,-0.2949183583,-0.4690204859,-0.2130618989,-0.4673829377,-0.2018324584,-0.3875800669,0.0961183831,-0.0633576289,-0.1115117967,0.2050208896,0.1171565801,0.3408107162,0.5179665685,0.1969754994,0.0698204041,0.4441018701,-0.2930302918,0.0693575963,-0.0417867489,-0.0196989011,0.2912706137,0.253423363,0.02117474,0.1961011142,0.0892119184,-0.0523033552,-0.1398488432,-0.0679250136,-0.2415895164,-0.0877483785,-0.4039773047,0.4737977386,0.3026017249,0.885436058,-0.1235516518,-0.0572129674,0.0557519533,0.3473540843,-0.1545779705,0.0594960861,-0.0361596756,-0.2303109765,0.2189463526,0.0370013006,-0.3095414639,-0.254293859,0.4501921535,-0.0263029877,0.32997334,-0.0426365212,0.0961557552,0.1398076266,-0.2091662586,-0.1120555326,-0.0645513311,-0.1094444394,0.132955417,-0.3483422399,-0.0717624128,0.1117440313,-0.0141074546,0.0450829305,-0.2468639165,-0.2309931815,0.0057506594,-0.0893442258,-0.0871781036,0.1881148964,0.3799712956,-0.0295595024,-0.115954712,0.1522289217,-0.0173908081,0.0141991302,0.0099379616,-0.1114291325,0.0547410883,-0.3408961594,0.1652857363,0.0806239992,-0.2036334872,-0.2136302739,0.0881192088,-0.2003660947,-0.046621304,-0.1008245349,0.4376762509,-0.0591083951,0.3093274236,-0.0603768192,0.1930448562,-0.1479225904,-0.3695376515,-0.087126255,-0.0167545434,-0.2718517482,0.0807316676,0.2328395993,0.2596525252,0.0590137839,0.0937125012,0.0492124557,-0.0665861368,-0.2398619503,-0.0384640992,0.2115903497,0.1285411716,0.0698484033,0.2205139101,-0.0920844674,-0.2250446826,-0.0426823683,0.100471355,0.0837401226,-0.0745024607,-0.1287995279,0.1920661777,-0.1797726303,-0.3141009808,-0.3180534244,0.5080114007,-0.2414616644,-0.0652867705,0.0323659144,0.1024618074,-0.4688642323,-0.0210200418,0.0407257043,0.5596446395,-0.4436152279,-0.0307838526,0.0880186334,-0.0309036765,0.1453659236,0.1575247496,-0.154580906,0.3193620145,-0.0214490909,0.301163584,-0.028509561,-0.223050043,-0.6605327129,0.0781061873,-0.0966477618,-0.0666845366,0.0618451983,-0.0090193376,0.6679314971,0.0456297025,-0.1564008594,0.4466237724,0.0494121611,-0.0715120807,-0.1420307904,-0.3319181204,0.289217025,-0.0197947584,0.1687025875,-0.0537235737,0.1330806613,0.3083949387,0.3169301152,-0.0140785202,0.2964548469,0.3474662602,0.0816651136,-0.0188949984,0.0484906547,0.0245989747,-0.3104139864,0.1486809105,0.2196630239,0.4836666882,-0.0978778079,0.0032812564,-0.2344828248,-0.1308935732,-0.3603484333,-0.1023183763,0.1876636147,0.3728756607,0.1479760408,0.0966864675,-0.0775567219,0.3941006958,-0.2766844034,-0.1055152342,-0.0879714563,0.424913317,-0.0061780321,-0.2714403868,0.1861996055,0.2040861696,0.3342067599,-0.0953066051,-0.0261943303,0.2505721748,-0.2482030541,0.3651418388,0.1121957004,0.0992307588,0.2652243078,-0.5659905076,0.2199379355,0.4432961941,0.2206995487,-0.3255328834,0.0056900908,-0.1550021023,0.2652459145,0.1565917134,0.0905761272,0.029094642,0.455296427,-0.0814787969,-0.1654933542,0.0345614478,0.2561272681,0.225124985,-0.1797230989,-0.1153329685,-0.2786634862,0.2571583986,0.3868488967,-0.3051303029,0.2623223364,0.157049492,-0.0361990221,-0.2613961101,-0.0196962506,-0.1009156406,0.2762663066,0.1733915657,0.3040486276,0.0963478163,0.0565766357,-0.0622084476,0.2041077316,0.0213307906,-0.1722148806,0.2142886817,0.1372830868,-0.2012933046,-0.399173677,0.0878325775,0.1849177033,0.3150724471,-0.0980789214,-0.2744605839,-0.2113194466,-0.2693474293,-0.3065424263,-0.0919738412,-0.3515456617,-0.080779314,0.1875156909,0.0719188303,0.1317643225,0.2187929153,0.2243707925,0.2338600457,-0.3013218939,-0.0422889367,0.0280266888,-0.3736102283,-0.4183723629,0.0749583468,0.3446518183,0.2606092095,0.2495827228,-0.1936534345,0.0172339696,-0.3335565627,-0.205454722,0.269944489,-0.1151761264,0.049855832,-0.2566846907,0.3304830194,-0.0554927289,-0.253680855,-0.1510607302,-0.0729965419,0.0780905187,-0.0291395858,-0.3341872096,-0.1840234846,0.0849074498,-0.5110989809,-0.4248243272,-0.2990658879,0.3247789443,-0.3723615706,-0.0713724643,0.1569874883,0.0470274054,0.1290336102,-0.079925105,0.0045363926,-0.4204454422,-0.0920868441,0.162025407,-0.0662740991,-0.4101423323,0.1232520416,0.1440681964,0.1636272073,-0.0456789844,-0.3998613656,0.2495915741,-0.1116474122,0.288593322,0.2177091539,0.1170541868,0.3275977373,-0.1510184258,-0.0764590949,-0.1397576332,-0.1971049458,0.0362469554,0.3786502779,0.4517078102,0.1034541726,0.0949351266,0.0015626032,0.5036545992,0.2424418181,-0.0031210566,0.1995590478,-0.0323325731,0.345185101,-0.1202696934,-0.4784105122,0.3422201872,-0.1209315136,-0.1080135703,0.3984332383,-0.023738049,-0.4457185268,-0.2867484093,-0.2088177502,-0.1730729789,-0.3948392272,0.2035857439,0.324883759,0.1942492127,0.0894131735,-0.0677352473,-0.069924809,-0.2438024729,0.253031671,0.1920572966,-0.1503223032,0.0457890555,-0.2262023836,-0.0615707226,-0.5227594376,-0.0031612138,0.0883577913,0.6162902713,-0.5153847933,0.278806448,0.1637890041,0.1118777543,0.4498525858,-0.4674552977,0.0256744437,0.2579941452,0.5658134222,-0.0244936328,0.085250169,-0.1826390028,-0.3291039467,0.7147038579,0.1901452392,-0.2403814048,0.1496539414,0.0743119642,0.4221892357,-0.0912758112,-0.1062732488,0.0818332657,-0.4652076662,-0.3623044789,0.0167630184,0.219795391,0.023223605,0.3030980229,0.16810368,0.2868337631,-0.0858305022,0.2700436413,-0.0696946904,-0.0653882921,0.6191542745,0.1840454191,0.1778921187,-0.0954327807,-0.1222427338,0.3881083131,-0.1137280539,-0.3406713903,0.0815933868,-0.2186896354,0.2896744907,0.4477070272,0.1923992038,0.1342247874,0.0914104581,0.0541124977,-0.2861167192,0.24857907,0.2936599851,0.1774266213,-0.3353132606,-0.4066302478,0.4526827335,-0.1107048169,-0.079278931,0.1087721735,0.1136894673,-0.1474588662,0.1033294499,-0.2280929834,1.3418978453,0.2603317201,-0.1036196202,0.1596807539,-0.3964629471,0.0489220135,-0.4883713126,0.0577523746,0.241898343,0.1264494359,-0.0792926773,0.2386053652,-0.0768337622,0.0488621704,-0.4408449829,0.4517140985,0.0355285965,0.0824258551,0.2013553232,0.5475209355,-0.1408499926,-0.2970794737,-0.1880217195,0.2220740318,0.1191573963,0.4247291088,0.0701708123,-0.1490742415,0.1583411992,-0.0846646354,-0.8224692941,0.1091837883,-0.339235574,-0.0396402068,-0.4534524083,-0.0467163138,0.1759988666,0.0454419069,0.3028603196,-0.0096316142,-0.3284364343,0.3235789239,-0.1158071607,-0.0665796399,0.1293949634,-0.105898194,0.2552307248,-0.2061971873,-0.4552624226,0.1773906648,0.1539212912,0.0073598409,-0.3866434693,-0.2827108502,0.0224936362,0.0962963477,0.2535409331,0.5415590405,0.0558383986,-0.1877815574,0.1313236058,-0.1040622965,-0.2573297024,0.2665681243,-0.2013322413,-0.2123295218,-0.2359618843,0.4419167936,0.0364745483,-0.0723954961,0.4562741518,0.5152431726,-0.1842770129,-0.2891137898,-0.1190693155,-0.0773093849,-0.5514754057,-0.0808721185,-0.0493548177,0.0920355842,-0.2553323209,0.3498761058,0.0698791668,0.0173560698,-0.0959339067,-0.1991743743,-0.1558380425,-0.0977614298,0.2887160778,0.2710490823,-0.2001205534,0.1303313524,0.2551153004,-0.1390504539,-0.3407117724,0.1193634048,-0.1413017064,0.1880838722,0.2851235569,-0.2964829206,0.3054423928,0.005374705,0.0820787698,-0.1342370361,-0.0878636762,-0.1190016791,-0.0458477922,0.1082145497,0.0690947995,-0.0024013007,-0.1677383035,-0.3603922427,-0.1549487859,0.2408121377,0.1800763309,0.2694815993,-0.1145241931,0.0228337571,-0.0059335604,0.0147907836,0.1496975124,0.0004005376,0.2844318748,0.5005568862,0.1575224251,0.2174274176,0.0594820492,-0.0265239328,-0.2028452456,0.338210851,0.1522680521,-0.1635859907,0.1632893234,-0.0093589695,0.0980006531,0.1910671592,0.1548510194,0.2413558513,-0.0416816361,-0.1847534925,0.4256339967,0.1649837643,-0.1744551659,0.0280381646,-0.1418947577,-0.0508343652,0.0231786687,0.0487715192,-0.0049963365,-0.3202770948,0.1585065573,-0.0347119756,0.0083874119,0.0832318664,0.0358868055,0.2588763535,0.0304734539,-0.173949495,0.0840402842,0.0398259684,-0.0776164383,-0.0406231843,-0.165727511,0.3330057859,-0.0622349419,0.2478661686,0.1206947267,-0.2034358978,0.0679400042,0.1265934855,0.0678182468,0.0898024514,0.0554197393,0.4078163505,0.0387848094,-0.1463383734,-0.2957153916,0.3513891697,0.144403249,-0.0664983243,-0.0986273587,-0.3691887259,-0.3593110144,0.1443934739,0.0553787313,-0.2998837829,0.0408998169,0.2840964794,-0.175061807,-0.1625925004,-0.3025871813,-0.2162395567,0.0407693833,-0.3084471822,0.303530246,0.1948907971,-0.0240650252,0.1623738259,0.5255322456,0.4240308404,0.1362578869,-0.2181157172,0.4244427085,0.105322808,-0.0706630275,0.0954016224,0.4309615493,0.1197750494,0.0317173079,0.2487389147,0.1331909299,-0.13525258,-0.0908410996,0.0301531926,0.0016615189,-0.1724865884,0.0589377098,0.0017152714,0.3237992227,-0.5256592035,-0.170433417,-0.6494346261,0.1055361703,0.3118796945,-0.2809125185,0.2777458131,-0.2200428843,0.0765103549,-0.3534982204,0.3516307473,0.3177817762,0.1568410695,-0.2055823803,-0.2269177139,-0.4249224961,-0.1397572905,-0.135337621,0.0934217498,-0.280035913,0.0973803625,0.2640997171,0.2254630029,0.0399104916,-0.0119217057,-0.4218786657,0.3082427979,-0.1878544539,0.1051817536,0.1375590414,0.541703403,-0.0504850708,-0.0852683261,-0.0351543948,0.1570748985,0.0131978383,-0.0613506772,-0.2157157212,-0.1166856289,-0.2601788044,0.111974746,0.0490540229,0.5883555412,-0.1400180757,-0.0703615993,-0.437220335,-0.1042068228,-0.0301355086,-0.0435722359,-0.0192247722,0.1632123888,-0.0883646011,0.3698633611,-0.2326092422,-0.244353503,-0.0902954563,0.3523517549,-0.3444108367,-0.0661149248,-0.1640645266,0.1683388203,-0.0599141829,0.1166470721,-0.0266477652,-0.0575413257,-0.2933105826,-0.0979802758,0.4137923717,-0.1922775507,-0.377227962,-0.2222601324,0.3068431318,-0.0149476919,0.1455416977,-0.6260253787,0.2453350574,0.2247815728,0.0561953112,-0.0877757743,0.148687318,0.14266482,0.1332047582,-0.0693176463,0.0554347858,0.2344230264,-0.2237687856,0.2015021294,-0.2066237926]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Let me check tomorrow! Thanks for leaving this message!","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":9,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nLet me check tomorrow! Thanks for leaving this message!","embeddings":[-0.2856391668,0.1633064896,-0.1317873448,-0.0331960879,-0.0231015589,-0.0723085552,0.0579205155,0.1924501657,-0.3027140796,0.2374878526,-0.2450919896,-0.1144219637,-0.0066451491,0.0723022744,0.1607536227,-0.1524203271,0.1237532571,0.0681254938,-0.3650565743,-0.289465189,-0.1095258296,0.076492995,-0.1141237319,-0.1312866509,-0.1967135221,-0.0543398671,-0.237476185,-0.2019926161,-0.2966832221,-0.1384563148,0.0025638388,0.1212082952,0.0777821243,0.3630088866,-0.0001010175,-0.15700607,0.2614805996,-0.2524596155,-0.3121373951,-0.1187733859,-0.1919900924,-0.1224848852,-0.5495233536,0.0337910317,-0.2725066841,-0.194755584,0.0470327362,0.0932687074,0.2024245113,0.623403132,0.2656264603,0.0583438575,0.2412303239,-0.2010711282,0.2539067268,0.1071956605,-0.1555710137,0.0547960587,0.2479884177,0.085801743,0.0809115171,0.3983794749,0.0501922183,-0.1777345836,-0.1243541092,-0.3750897646,-0.0985557139,-0.3697888553,0.1481964737,0.3550561965,0.5841485262,-0.1603987366,-0.0684378594,-0.0553301647,0.2480223179,-0.1668183208,-0.0089089442,0.0008190327,-0.1373753846,0.2875059843,0.1870175153,-0.205347985,-0.1546850652,0.3117708266,0.053517662,0.3175325692,-0.1047516987,-0.1329772919,0.0905575231,-0.3088515401,-0.304925859,-0.0359768309,-0.0036290905,0.0576293319,-0.2075772882,-0.2677211761,0.3042556047,-0.1310103685,0.1368422061,-0.3248894513,-0.1777551621,-0.0685017779,-0.1404032111,0.0899588913,0.0271016229,0.4020976424,0.0484152958,0.0157331098,0.070434466,-0.205728367,-0.0153466407,0.0958886519,-0.0043859128,0.1313463002,-0.3620999157,0.2895817459,-0.0279997569,-0.1595762223,-0.2776524425,0.0823110715,-0.0565447919,0.1462440491,-0.0302649532,0.3559885919,0.0810487047,0.3340353966,-0.0715533122,0.3587631285,-0.0714595839,-0.1454914063,-0.0989443511,0.1628498137,-0.2574824393,0.1261955202,0.2180879712,0.3924459517,0.1906028986,0.054120034,0.2297237962,0.0238077044,-0.1899750829,-0.0473817885,0.2634067833,0.0513496399,0.0863399059,0.051596541,-0.1367209405,-0.1994287968,-0.2418517023,0.0270088576,0.0978683233,-0.029916618,-0.2256229818,0.304164052,-0.0368399248,-0.1998565644,-0.2476514876,0.6678962708,-0.1488871574,-0.1391501725,0.0983619541,0.1016187817,-0.3634717166,0.0947761908,-0.0748513788,0.2563478649,-0.2272710949,-0.0522273667,0.1240413189,0.0269636102,0.0900420099,0.2007398158,-0.0517515317,0.2592617869,-0.0842806771,0.3430352211,0.0934314653,-0.4258820117,-0.3105736375,0.1082232445,-0.179462865,-0.0486734286,0.2208053023,-0.180781737,0.6067197919,-0.2387393117,-0.0317285918,0.4102734625,-0.0055863401,0.1247012764,-0.2023807466,-0.2967993617,0.201573059,0.1317943633,-0.0150840152,0.0428443737,0.0149204675,0.3503391743,0.2898407876,-0.2343604863,0.2273850143,0.2961333692,0.0751731992,-0.1842745245,0.0778201297,-0.1171106622,-0.063670367,-0.1493693143,0.1258476973,0.4999672174,0.1078639328,-0.2317792326,-0.1555262357,-0.2418736666,-0.1555646509,-0.1630941182,0.3212541938,0.2954792082,0.0654955804,0.09666536,-0.03858326,0.234196201,-0.3385366201,-0.259180367,-0.0219033305,0.2231100798,-0.0272439383,-0.3182163835,0.2678953707,0.3651417494,0.20955652,0.0176712144,-0.05303495,0.0944855288,-0.1808695197,0.2394014597,0.0422936529,0.1487890333,0.2364812344,-0.5072137117,0.1276944131,0.4144308269,0.1588810086,-0.2189746499,-0.0769400075,0.1560238749,0.3886389732,0.1007166356,0.1512297094,-0.0517807603,0.4313502014,-0.0204293672,-0.162581712,-0.1107490212,0.1528676599,0.2242011875,-0.1748860925,0.0309011545,-0.2916494608,0.1760224253,0.6214765906,-0.1516982764,0.3430842757,0.2252093703,-0.0507970788,-0.1892836392,0.0614079274,-0.2848861516,0.2171165049,0.3107837141,0.2919835746,-0.0575880893,0.1586072594,-0.1135099679,0.1626087725,-0.0241679493,-0.2352177799,0.1664084941,0.1652378142,-0.257830739,-0.3967554569,0.0552810729,-0.0125616984,0.1833819747,-0.1181164309,-0.3113757074,0.0347328298,-0.2094092667,-0.18636401,0.0505847,-0.1639024168,-0.0073671751,0.435361594,0.1077572927,-0.1897539347,0.28068766,0.3017904758,0.2073782086,-0.3687796891,-0.0260419063,-0.0761817992,-0.3541452289,-0.2397752106,0.2395503074,0.0986054391,0.2335148156,0.1517514735,-0.162142247,0.1753920913,-0.2377748638,-0.2847909033,0.2241194993,-0.2617709637,0.0688706115,-0.2369811088,0.0149305928,-0.038236782,-0.4739055634,-0.1936577857,-0.0112000555,-0.0990166888,0.0168403219,-0.4449462593,-0.1231711954,-0.0708493143,-0.4875477552,-0.4077456295,-0.4610809684,0.3148328066,-0.2484512329,-0.0280042812,0.0462734811,0.0251182448,0.07572148,-0.0619051494,0.2172350883,-0.315846324,-0.2114031315,0.0817744657,-0.3130907714,-0.3068004251,-0.2626135647,0.0249177869,0.1077224836,-0.1796005517,-0.2366686463,0.0866269544,-0.1305063218,0.314994067,0.1194118336,0.0131442165,0.2735008001,-0.1450885087,-0.1978711784,-0.2050811797,-0.1455636322,0.1097359061,0.3634960949,0.2226115465,-0.0504287295,0.0659335405,-0.0038968902,0.5311444402,0.1689791977,0.0191353075,0.2651659846,0.0142000932,0.4160391688,-0.2975606024,-0.3449871838,0.4802240133,0.1115666479,-0.0203042645,0.4960016012,-0.0108879097,-0.4459588528,-0.0922799483,-0.2863521874,-0.3142743707,-0.2518981397,0.2920818925,0.1426917315,0.1237162501,0.0098489299,-0.1353355795,-0.2109840065,-0.1483186334,0.2661404312,0.1215988025,0.0133146355,-0.0050121611,-0.1799531132,-0.063539356,-0.3240480721,-0.1201501116,0.1310222149,0.3191864789,-0.5553612113,0.0408992544,-0.0494888946,0.1610035449,0.3229970336,-0.3436067998,-0.0031694141,0.2685288787,0.5231115818,0.0371712521,0.1860289425,-0.3070530891,-0.3497695327,0.666929841,0.3248840272,-0.3043201864,0.0164226741,0.1559573412,0.2815506458,0.0312793516,-0.1065847054,0.0015840951,-0.3043818474,-0.3849747479,0.3064019978,0.5321170092,0.0293493234,0.1606896818,0.0375488363,0.1806548983,-0.1243413761,0.4313430786,-0.0036598595,0.0276656356,0.4640321136,0.0527379103,-0.0422030762,-0.0807140023,-0.146170482,0.1681770831,-0.0056602526,-0.3521020412,0.0504001677,-0.3033138812,0.332850486,0.2621144354,0.1565977484,0.0543603934,-0.1546443701,0.0343320295,-0.158806175,0.2173501402,0.2817341089,0.1761366874,-0.2756204605,-0.2105078995,0.3226342499,-0.0261467062,-0.1374529451,0.0610131659,0.1151962131,-0.2163311839,0.2875320613,-0.1945926696,1.1859164238,0.2576404512,-0.0122283334,0.2131811827,-0.2907941043,0.0302415881,-0.4413577914,-0.1030029953,0.2916038632,0.1287510842,0.0400441289,0.2460236847,-0.04145547,0.216496557,-0.4922026992,0.2203012705,-0.0447867662,0.0825890973,0.1206859201,0.3867156804,-0.1717712879,-0.3591281474,-0.3095202744,0.2961108983,0.0985039994,0.1858029217,0.1857928038,-0.2785589397,0.2281391174,-0.062553905,-0.5616278052,0.2172744125,-0.2123617679,-0.1030458137,-0.4252543449,-0.1067392603,0.1792345792,0.1815079153,0.4212223589,0.0508493446,-0.2191141546,0.2738002837,0.0829174444,0.0783318952,0.0921333358,-0.025375342,0.3653206527,-0.3371700048,-0.3400430381,0.1704835147,0.1031239927,-0.0358519293,-0.3938149214,-0.2602199316,0.0893107131,0.1515876353,0.1987166405,0.5595204234,-0.0749498382,-0.3089916408,0.2381711602,-0.1106470153,-0.3929108083,0.1939388961,-0.1711106002,-0.2862193584,-0.2005277574,0.2460622191,0.2204613537,-0.013671333,0.3441177905,0.2517158389,-0.1926268935,-0.429713428,-0.083121337,0.188823387,-0.456486702,0.003515743,0.1013042107,-0.1986517608,-0.2390930206,0.303129524,-0.0090193544,0.0036707686,-0.0912311003,-0.0046560438,-0.2763293982,-0.074278228,0.0948653147,0.0753725022,-0.0837513879,0.2624467611,0.1158161089,-0.0718824044,-0.469943434,0.0157047138,-0.1945856214,0.2283819318,0.2006520778,-0.1785490215,0.1585137248,0.05015333,0.2080312818,-0.0632245988,-0.0900179595,-0.3062080443,-0.1723488718,0.0700092092,0.1295055002,0.2841500938,-0.0159116611,-0.0934715718,-0.1872984916,0.0690241084,0.3510090113,0.1956052631,-0.0182056408,0.2001442611,-0.1006025299,0.0294591319,0.264775306,-0.1116677821,0.239126876,0.2468034774,0.0346708335,0.0719895437,0.0389123224,-0.1786587983,-0.0186444372,0.2482374609,0.1561352313,0.07637088,-0.054995548,0.1263247728,0.1379526258,0.3233583272,0.0204060934,-0.1039864495,-0.0437199548,-0.131130591,0.3751435876,0.3287830651,-0.2098583132,0.1869631261,-0.2338539809,-0.2602289319,0.1941657662,0.0399124064,-0.0198595393,-0.0732566714,0.3554494977,-0.0680537,-0.2076733708,0.0028981774,0.0041181655,0.1912288517,0.0758495256,-0.0863619819,0.0401926227,-0.0219771899,0.084445715,0.1009654179,-0.153045401,0.4942649007,0.1331812441,0.1907344311,0.3137116134,0.0068272138,0.0501459278,0.0956340358,0.056406498,0.2425267696,0.0603156537,0.282412827,-0.0470266305,-0.2770374417,-0.1898006052,0.193657279,-0.031524647,-0.0206685755,-0.07459566,-0.1446172595,-0.2284139544,-0.0707023665,0.0633106157,-0.0837831199,-0.1030144393,0.1643240154,-0.1646775752,-0.1001448482,-0.1559727639,-0.2531581819,0.1999870837,-0.1967109293,0.4050533473,0.1586656868,-0.0554167964,0.1960554719,0.4884906709,0.4734392762,-0.0156335551,-0.1732738763,0.1633333117,-0.0383942015,-0.1688638031,-0.0063811629,0.4247526824,0.1486067623,0.148705557,0.1472287327,0.2865831852,-0.264098525,-0.1723692715,0.1527969539,-0.133746922,-0.0423160307,0.141483441,-0.0866879821,0.163078025,-0.5935208201,-0.1809074283,-0.5620452762,0.0822886154,0.3517209589,-0.2423039675,0.0841366872,-0.219199121,0.1429883242,-0.2031248659,0.2784118354,0.2171230167,0.030547468,-0.0666891262,-0.1650202721,-0.3301177919,-0.0261383131,0.1094590724,0.1156638414,-0.3325906694,0.0448058359,0.4244823158,0.2833173573,-0.1372639686,-0.090821743,-0.3754463494,0.1242533177,-0.3613914549,0.2495110631,0.0309363361,0.5286003351,0.1421683729,-0.1462256461,-0.2103559822,0.2376049012,0.146867767,-0.0984312668,-0.0428780019,-0.1939031631,-0.2500329316,0.1672465354,-0.0041708318,0.4114120007,-0.3175730705,0.0001230404,-0.0533122271,-0.1428469867,-0.1225024462,-0.0365682095,-0.0494482443,0.2665835023,-0.1010472029,0.29857409,0.0005038838,-0.0778287277,0.0354612879,0.0617628843,-0.2456126064,0.1167517528,-0.0909445137,-0.004925746,-0.1934348047,0.2066486329,0.0124319186,-0.0786826462,-0.245902136,-0.2891391516,0.4203591645,-0.1203993708,-0.3700068593,-0.0946994573,0.3964122534,0.1107814386,0.004567455,-0.6658568978,0.2259707153,0.2033445239,0.1335596144,0.0544594899,0.1924689114,0.0627558455,0.0260886718,-0.0841973945,-0.0511246361,0.18668136,-0.1644176245,0.1314736903,-0.307946533]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"cc @patil-suraj for notification ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":4,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\ncc @patil-suraj for notification ","embeddings":[-0.3044423163,0.2005738616,-0.1330436319,-0.0003878294,-0.0525170453,-0.0534180515,0.1113866791,0.1956667602,-0.3056173921,0.2338666618,-0.2486449927,-0.1109912172,-0.029684227,0.1287885755,0.1691138893,-0.1248226613,0.1237180606,0.045715943,-0.3355683088,-0.2708309293,-0.0948083252,0.0594704971,-0.1014416367,-0.1650190502,-0.2262085527,-0.042748455,-0.233525604,-0.2356780469,-0.2837334573,-0.1318155974,0.002971892,0.1435183138,0.0817181543,0.3875293136,-0.0001023634,-0.1485596895,0.2234328687,-0.2469525039,-0.3330296576,-0.1265739202,-0.1968506426,-0.1241198182,-0.5734244585,0.0394988544,-0.3046047091,-0.2156670988,0.0154631715,0.107975632,0.1945064962,0.5951433778,0.2486138642,0.0739221647,0.2416187674,-0.1712726057,0.2441120148,0.1085196957,-0.1731166095,0.0537688583,0.2704944313,0.1225475818,0.0628264248,0.4256951511,0.0466931276,-0.1806395948,-0.1212852597,-0.3692339659,-0.132763803,-0.3705955446,0.1435629278,0.3780076504,0.5728987455,-0.1526521295,-0.1122579649,-0.1113121212,0.2288615406,-0.1933954209,-0.0181991123,0.0144502232,-0.1359731555,0.3159310222,0.1721234322,-0.2056819499,-0.1615149379,0.3042834401,0.0522614866,0.3572616279,-0.1003805846,-0.1400381923,0.0637626946,-0.3051520884,-0.2907769382,-0.0602123737,0.0037537126,0.0476756804,-0.2077556849,-0.3227247596,0.2989279628,-0.1559243649,0.1487972438,-0.3133231699,-0.166247949,-0.0407950357,-0.139950335,0.0805172622,0.0263446569,0.3897664845,0.0401229002,0.0105324788,0.1207119003,-0.2041190863,0.0100175031,0.0991179571,-0.0059713293,0.2014252692,-0.3864099681,0.2811419368,-0.025304148,-0.1762153506,-0.3136895597,0.0654419959,-0.0529319793,0.1331987977,-0.0190466587,0.352120012,0.0993069261,0.3359657228,-0.1023964807,0.3737513721,-0.0675907284,-0.1350414157,-0.0770532638,0.1443920285,-0.2242296338,0.1299819946,0.2397750914,0.4106637239,0.1597126424,0.0463073924,0.2339979112,0.0206318125,-0.1659454852,-0.0345378816,0.2790391147,0.0285071936,0.0761030316,0.0131812906,-0.1533017308,-0.1914300919,-0.2482024729,0.0181688238,0.1029277891,-0.0212277845,-0.222529009,0.2942441404,-0.0475415364,-0.2048506439,-0.242674157,0.726786375,-0.148591131,-0.1237121373,0.1116543636,0.1356570423,-0.3872209489,0.1007670462,-0.0883883312,0.2928372025,-0.2065102905,-0.0913946033,0.145828262,0.0662153959,0.0606707744,0.1870315969,-0.061668206,0.2766954303,-0.0792547166,0.3336174786,0.1170910001,-0.420055151,-0.3128738105,0.1212161258,-0.1867426038,-0.0419852138,0.2426124662,-0.1385384798,0.6358507872,-0.2620573342,-0.0174459759,0.4321750402,-0.0000427684,0.1275777072,-0.1707070768,-0.3119196892,0.2161106467,0.1360423863,0.0224318095,0.0366595984,0.0288834441,0.3339502513,0.2831148803,-0.2669495642,0.2563773394,0.2690968513,0.0746246204,-0.2002023906,0.0726201162,-0.1436789483,-0.1032423824,-0.1408191025,0.1170156151,0.4878193736,0.1062037125,-0.2620977163,-0.1277538985,-0.2413671166,-0.184410587,-0.1713061035,0.302372545,0.2980331481,0.0953687802,0.0923935547,-0.0455246605,0.2277601212,-0.3388955891,-0.2209117413,-0.0106088417,0.2241858095,0.0268124044,-0.3185555935,0.2668296099,0.3995208442,0.194795385,0.0035125178,-0.0586536229,0.0842573792,-0.2078598589,0.2559249401,0.0644696802,0.1797704399,0.251334399,-0.4822784364,0.1411233842,0.4031176865,0.1478757113,-0.2123778015,-0.1096171439,0.2103890032,0.3967447877,0.1043858007,0.1181512475,-0.0353756621,0.3972280025,-0.005920622,-0.154178381,-0.1162717864,0.1341600716,0.2231917083,-0.1617115587,0.0317028277,-0.2968840897,0.2062350363,0.6123940945,-0.143398568,0.3488172591,0.2364427447,-0.018163899,-0.1851643473,0.0693393722,-0.3037065566,0.2189243287,0.3111042976,0.2975141406,-0.0584453233,0.1886100173,-0.1038814187,0.1681158692,-0.0189934857,-0.2695644498,0.1849504709,0.1999980658,-0.2501428425,-0.4017956853,0.0703816265,-0.0171471369,0.1629048735,-0.1190703809,-0.2956649661,0.0497704595,-0.2351997048,-0.205129832,0.0556664653,-0.1772665679,0.0150277251,0.4463767111,0.1089808792,-0.1893759817,0.2723469138,0.2861568332,0.2307478935,-0.3879211545,-0.0363193937,-0.0908346176,-0.3630057871,-0.2091256827,0.2257041186,0.13981691,0.2105215639,0.174261108,-0.155129239,0.1450534314,-0.2708490193,-0.2968089879,0.2253436893,-0.2719904184,0.0850240737,-0.2630524337,-0.0110128289,-0.0542645864,-0.4443923533,-0.2023138404,-0.0074507287,-0.0839788318,0.0018728158,-0.445420295,-0.0970213413,-0.0667411983,-0.5018101931,-0.4160226285,-0.4638018906,0.3455711007,-0.2590662539,-0.0248060301,0.0393966623,0.0055039814,0.0551367588,-0.0416124836,0.2086606771,-0.3130107522,-0.2151189893,0.0821749419,-0.3270351589,-0.3017486036,-0.272588104,0.0322644636,0.0902311429,-0.1493213177,-0.2301482558,0.0903851762,-0.1170146391,0.3167545199,0.1487284303,0.016567938,0.2585158646,-0.1627722234,-0.1852938086,-0.1932590455,-0.1599075794,0.1207259223,0.3617868721,0.2004140466,-0.0078947581,0.059601184,0.0067772325,0.5447929502,0.1623085737,0.0055376985,0.2778539658,0.0292126164,0.3892539144,-0.2870731652,-0.3311444521,0.5082709193,0.1187840477,-0.0115657272,0.4872771204,0.0032890902,-0.4555382133,-0.1129935011,-0.2913231254,-0.3167012036,-0.2508221567,0.301674366,0.1353458017,0.1617551446,0.0199061055,-0.1371427625,-0.2044941932,-0.173110649,0.2833042443,0.1066310704,0.0206023455,0.0015264825,-0.1755260825,-0.0524238572,-0.3445534408,-0.0967851728,0.1367710829,0.2988578379,-0.5548205376,0.028928712,-0.0318761542,0.16450423,0.3013784587,-0.345995903,-0.0010776472,0.2483111322,0.5042901635,0.0461895168,0.1976511925,-0.2953968644,-0.3544014394,0.6332861185,0.3635220826,-0.3464398086,-0.0023029561,0.1917526126,0.290518105,0.0159229301,-0.083221525,0.0395281315,-0.3232010603,-0.3579851687,0.309653312,0.5458112955,0.0396067537,0.1722356826,0.0353012979,0.1711579114,-0.1071084663,0.4388350248,-0.007533628,0.0114639141,0.4615053236,0.0679740533,-0.0357719474,-0.0814349279,-0.1232450083,0.1591210216,-0.0238020606,-0.3453144431,0.0554072075,-0.3178050816,0.3425120413,0.3072019517,0.155159682,0.0735658631,-0.141948849,0.0250609722,-0.1843127161,0.1974527687,0.2651609778,0.2007759213,-0.2954106629,-0.2272714823,0.3384841979,-0.0086356839,-0.122498922,0.0628705025,0.0901539624,-0.2389660925,0.2833459675,-0.1780062914,1.1970597506,0.2848080099,0.0052091461,0.2150357217,-0.2852778733,0.0285197347,-0.4387690127,-0.110888496,0.284735322,0.1736479849,0.0225536358,0.2396896929,-0.0637169182,0.2376484722,-0.4977327287,0.2453109622,-0.0143597443,0.1149270982,0.1137811244,0.3941930532,-0.2094036043,-0.3661506176,-0.3277177513,0.2839579582,0.0802824721,0.1721853316,0.1708164215,-0.2946581542,0.2076925784,-0.0793023556,-0.5617780685,0.1728745699,-0.2197450846,-0.091651015,-0.4098140299,-0.1817807555,0.1810394973,0.165499419,0.4319888055,0.0144776488,-0.2094681561,0.2442868948,0.0365063809,0.1226534396,0.1161848605,-0.0540164039,0.3687973022,-0.3152298629,-0.3052566648,0.1656764448,0.1367480159,-0.0275291093,-0.3849696815,-0.2742160261,0.0588196069,0.1544520557,0.2348367125,0.5532253981,-0.0526330881,-0.3304071724,0.2300472856,-0.1039025262,-0.4071670771,0.1848594993,-0.1659210324,-0.2762008607,-0.2104218155,0.2162855119,0.210073784,-0.0262832027,0.3086798191,0.260439992,-0.1873161942,-0.4145176411,-0.0659022555,0.1725262254,-0.4493046403,0.0154112671,0.0977573991,-0.1927992404,-0.2655993998,0.3142947853,-0.0336172692,0.0025151768,-0.099489823,-0.0001698517,-0.2419644445,-0.0517479703,0.0965691954,0.1065371037,-0.1244809851,0.2620987594,0.0994166955,-0.0929879919,-0.4585853815,0.0054940647,-0.1596587002,0.2302361727,0.216141358,-0.1652606726,0.1630549431,0.0747220367,0.1789746433,-0.0556346886,-0.0890237615,-0.2893693447,-0.186034739,0.0722815469,0.0970814452,0.2933942974,-0.0470174439,-0.0917554796,-0.2016554624,0.0495290272,0.3919723034,0.1906604618,-0.0094396481,0.2000129372,-0.1052573249,0.0254171565,0.2793049812,-0.0788303986,0.2539890409,0.2819824815,0.0471976362,0.1062498018,0.0414419137,-0.1656614989,-0.0114657544,0.2564173043,0.1738785207,0.0352983996,-0.0558349602,0.1152243912,0.1401368082,0.3271294236,0.0269172601,-0.0739429221,-0.0477852486,-0.1064207479,0.3743627369,0.3199340999,-0.1853891462,0.2042570561,-0.2143217325,-0.2863636017,0.2010538727,0.0716371983,0.035302911,-0.074866809,0.2932724953,-0.0732731894,-0.21820876,0.0292512588,-0.0004553592,0.1756939739,0.1118798703,-0.0561864823,0.0455899276,-0.0096632354,0.0798381045,0.0429032855,-0.1369231641,0.4975193143,0.1231197268,0.1573180854,0.3080084026,0.0077935839,0.057044398,0.071175985,0.0455180481,0.2285431623,0.0612611845,0.3036381304,-0.0489611998,-0.2997124791,-0.1959433109,0.214756459,-0.0371699221,-0.0058779889,-0.0643873587,-0.1533048153,-0.2376190573,-0.0575956404,0.0607489794,-0.1160656512,-0.093388319,0.177514106,-0.1515936702,-0.073306933,-0.1521521807,-0.2561348081,0.2185831368,-0.198653385,0.3672944605,0.1370830387,-0.0562731735,0.1891048998,0.5279836655,0.4707911313,-0.0429469086,-0.1713173687,0.1718326211,-0.0779615268,-0.1639911383,-0.0363774449,0.4503370821,0.1144084856,0.1280117184,0.1620825976,0.2742313743,-0.2522874475,-0.1971605718,0.1443254799,-0.1412536502,-0.0884326473,0.1466563642,-0.0885531604,0.1649661213,-0.5708591342,-0.2356663495,-0.5621415377,0.0672064126,0.3753933609,-0.2472829223,0.0844005272,-0.205257982,0.1351705492,-0.193760246,0.2715344131,0.2009926736,0.0308800843,-0.0708071217,-0.1581800878,-0.3424617648,-0.0458441079,0.1302273571,0.1486837566,-0.3451304138,0.0598753504,0.3965277672,0.3219383955,-0.1484483182,-0.1166994944,-0.3817603588,0.1402785629,-0.3669436872,0.2267673463,0.0450228453,0.5420415998,0.1394570321,-0.1441449523,-0.2185443789,0.2684558034,0.1265518963,-0.1097196788,-0.0415939949,-0.1988706887,-0.26586923,0.1836326271,0.0000899846,0.4266269505,-0.314886272,-0.0204918832,-0.0558456182,-0.1517912894,-0.1406574994,-0.0449293479,-0.0843415856,0.272362709,-0.099417001,0.2801978588,0.0142039228,-0.0901821852,0.0257031824,0.0478083566,-0.2372066975,0.1041954011,-0.1227978095,-0.00890736,-0.1741229892,0.2388543636,0.0170616116,-0.0823108703,-0.2353151739,-0.29352355,0.4107241929,-0.1283787042,-0.3755359948,-0.0935110897,0.3944763243,0.0869803205,-0.0084339324,-0.6728065014,0.2393997908,0.2105534822,0.1451282799,0.0761187673,0.1943640411,0.0791142359,0.013360532,-0.0921971649,-0.0864250883,0.1489816904,-0.169238463,0.1338575631,-0.3066866994]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"@skyprince999, I think I'm getting the same error you're getting :-\/\r\n\r\n```\r\nSorry, you can't view or download this file at this time.\r\n\r\nToo many users have viewed or downloaded this file recently. Please try accessing the file again later. If the file you are trying to access is particularly large or is shared with many people, it may take up to 24 hours to be able to view or download the file. If you still can't access a file after 24 hours, contact your domain administrator.\r\n```\r\n\r\nIt would be great if you could write the authors to see whether they can fix it.\r\nAlso cc @lhoestq - do you think we could mirror the dataset? ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":117,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n@skyprince999, I think I'm getting the same error you're getting :-\/\r\n\r\n```\r\nSorry, you can't view or download this file at this time.\r\n\r\nToo many users have viewed or downloaded this file recently. Please try accessing the file again later. If the file you are trying to access is particularly large or is shared with many people, it may take up to 24 hours to be able to view or download the file. If you still can't access a file after 24 hours, contact your domain administrator.\r\n```\r\n\r\nIt would be great if you could write the authors to see whether they can fix it.\r\nAlso cc @lhoestq - do you think we could mirror the dataset? ","embeddings":[-0.3741523027,0.0519908965,-0.0482892953,0.0795517936,-0.0135703487,0.0174991563,0.0815657079,0.1886139065,-0.2064038217,0.3855748475,-0.3391426802,-0.5373642445,-0.0532581173,-0.0104364725,0.1305620521,-0.210103333,0.0005456007,-0.06197026,-0.2461218834,-0.106369406,-0.1054471284,0.0369486585,-0.1291830093,-0.0238434896,-0.2057791203,-0.1044686362,-0.3355354667,-0.0673639774,-0.3276143968,-0.2688932419,0.0196583364,0.2734157741,0.1154996157,0.4012147784,-0.0001045743,0.0001495671,0.4067375362,-0.0481949747,-0.3458155692,-0.1742513627,-0.1645576358,-0.28988868,-0.4356117845,0.0596620329,-0.2253366113,-0.1871693432,0.2630777955,0.0461974666,0.3312250674,0.5269789696,0.2198282033,-0.0575602539,0.3618278801,-0.3024641275,0.3061889708,-0.1263407618,-0.0886302441,0.3788725138,0.3622792363,0.1281085312,0.1928506345,0.2564085722,-0.0494399443,-0.1281773299,0.0093107624,-0.323219806,-0.0619877614,-0.3781906366,0.112661846,0.2508860528,0.8849615455,0.1165097132,-0.2376434207,-0.0453701615,0.1858688891,-0.0820282251,0.1716369689,0.081996493,-0.0974980742,0.2420727611,0.0106665101,-0.2699157596,-0.2761213481,0.2845781446,0.0610823222,0.1404553801,-0.0721468627,0.1089524031,0.2183655649,-0.2003873289,-0.2603643835,0.1827072352,-0.095295921,0.2284316868,-0.1151503325,-0.0316391625,0.2358843088,-0.1671473235,-0.0537558533,-0.171032235,-0.1305034906,-0.1916558594,-0.2057008147,0.098752439,-0.0011438088,0.4225839078,-0.0258220676,-0.0939659774,0.2431910187,-0.1216177717,-0.0408768356,0.06720265,-0.1611659676,-0.1563165188,-0.4741425514,0.2587382197,0.0858987495,-0.157445401,-0.2232072204,0.0081407251,-0.0643859804,-0.0268376898,0.0266043972,0.3763580322,-0.0467444658,0.2172406763,0.0722464398,0.2495343387,-0.072535865,-0.3502615988,-0.0644993782,0.0966746882,-0.5013253689,0.1774252504,0.3367813826,0.1270944774,0.022837922,-0.0218097791,0.1430231482,-0.152000159,-0.1542091519,-0.1316495091,0.0718260407,0.1473982036,0.1632300764,0.2887646854,-0.0635252893,-0.2075773031,-0.0414888896,0.0201892052,-0.1272518039,-0.3853573501,-0.135878548,0.2640460432,-0.1239970624,-0.05386604,-0.3378983438,0.4284421802,-0.2500939369,-0.2704096437,-0.1836604774,0.219341293,-0.4041494131,-0.0989936143,-0.0616128147,0.5527360439,-0.2626290321,0.10055352,-0.0642991289,-0.0722765699,0.0727322921,0.338503927,-0.3098070025,0.2924597561,-0.1130012199,0.2416450977,0.0820234194,-0.3433909416,-0.6367425323,0.1574754417,-0.3702493012,-0.1231483817,0.3622512817,-0.0813827366,0.7645028234,-0.3859058022,-0.0415783674,0.2580481768,0.2841435075,0.0082385167,-0.2456496954,-0.2950386107,0.1285574883,0.0993288159,0.0345422961,0.1200488731,0.258818686,0.3030824065,0.3145304024,-0.1403214037,0.1904460788,0.3736027181,0.0648749471,0.1424502134,-0.0410016105,-0.033505071,-0.0772662833,0.0428794287,0.1194464341,0.2878559828,0.1323676705,-0.1839862913,-0.3479401171,-0.2124460489,-0.3899978399,-0.0671338961,0.2137281746,0.3639135957,0.1334801763,0.0546990894,0.1492206752,0.3630463183,-0.343339026,-0.168185398,-0.1203388721,0.3763389587,-0.1943972558,-0.097244516,0.2011359781,0.1818703264,0.3316040337,-0.0948617533,-0.0286076069,0.188295424,-0.2663230598,0.2544405162,0.2089383155,0.1866558194,0.2824369371,-0.7307382226,0.3003783226,0.3212945461,0.2425289899,-0.1243080348,-0.1257095039,-0.030530056,0.2592164874,0.0685495734,0.1028057635,-0.0509606376,0.4734241664,0.0043289405,-0.074156709,-0.0254054796,0.2493099123,0.2675668299,-0.330817461,-0.1631774902,-0.1564379632,0.0893236771,0.395845145,-0.1409756392,0.0304459818,0.2175443769,0.0073720249,-0.1856971532,-0.0386355445,-0.1548451334,0.1878665388,0.2940442562,0.2710776925,0.0418884829,0.1936337054,-0.1156006604,0.2262291759,0.0844288394,-0.2344631106,0.1612132937,0.1995627433,0.0113987885,-0.1993491501,-0.1330867112,0.2303616107,0.2885450125,-0.1960631013,-0.3173702359,0.0356169343,-0.2672513723,-0.1209780276,-0.0798082352,-0.2716400027,-0.1120494157,0.1761445701,0.2219205648,-0.1300326437,0.0525014736,0.3018539548,0.2663785219,-0.2395298779,0.1697392613,-0.0898995847,-0.2039330006,-0.2418347597,0.1517958343,0.3139507174,-0.023502782,0.2608110905,-0.1703244746,0.2581349015,-0.4193348885,-0.2484351546,0.0725554228,-0.156543687,0.1284316629,-0.3858169317,0.2875353098,0.0410725325,-0.4558401704,-0.0480211675,-0.0496227071,-0.0067900335,0.0223000906,-0.2981385291,-0.1353980899,0.0043047937,-0.518535018,-0.5353162885,-0.4876825809,0.3409767151,-0.2174546719,0.0469317175,0.1393304169,0.1168725491,0.159300521,-0.0993602499,0.1808668226,-0.2902183235,-0.2010566741,0.1514411122,-0.3761223257,-0.3692210913,0.0950604603,0.2483451217,-0.0426920801,-0.1107676625,-0.4823299944,0.168835178,-0.341481626,0.1768948585,0.0542202778,0.026152825,0.1799420118,-0.1657060534,-0.1313550919,-0.0023244345,-0.2000384182,-0.0619425774,0.3084882796,0.2828818262,-0.1465216875,0.1768688709,0.1262538135,0.7308599353,0.3372609019,0.2193733156,0.225480929,0.0592510514,0.3217624724,-0.257629931,-0.4247913659,0.4319262207,-0.195953235,0.0024231356,0.4522396922,0.0829087645,-0.325620681,-0.2853494287,-0.4064799249,-0.3902540803,-0.4201671183,0.0551182888,0.327070117,0.0952719301,0.1288486719,-0.1303878725,-0.1808485538,-0.2825467587,0.2367132902,0.3933317661,-0.0613271892,0.0482302047,-0.0554152839,-0.1063750535,-0.2948363721,0.1687136292,0.1141622588,0.3505269885,-0.5199877024,0.2094108909,0.2939937413,0.1402665079,0.594584167,-0.3036943078,0.2389882952,0.085753642,0.4931543469,0.0129783694,0.0822298154,-0.195335567,-0.2238650024,0.6836931705,0.2362921089,-0.224705413,0.2338266671,0.1791153997,0.2137952894,-0.0391467102,-0.109755531,0.0356496684,-0.3531957269,-0.4895443916,0.1220329255,0.3398804069,0.1282110363,0.0662600696,0.0115317712,0.2779923379,-0.1156476587,0.3644886613,0.0395303033,0.0247256905,0.6241257787,0.1497076899,0.1796764135,-0.0925610438,-0.1748545319,0.5049250722,-0.0045216023,-0.272161901,0.0529044047,-0.1828312278,0.152613014,0.4952747524,0.0034144386,-0.0024309375,0.151977852,0.0838479176,-0.2381065786,0.3139930069,0.2047501653,0.0051215701,-0.4276718199,-0.3296124935,0.481682837,0.0816810578,-0.1838089377,0.2121181041,0.1973602772,-0.0482705012,0.1212409586,-0.3054203689,1.2670203447,-0.0345081314,-0.0262739975,0.143627882,-0.2191204429,0.0564315952,-0.5175658464,0.0886532217,0.211225301,-0.0125613315,0.0200398061,0.0909227803,0.0053873062,0.1498712897,-0.3562783301,0.0733415857,0.0318918377,0.0326334424,-0.0236462541,0.5749592781,-0.1809690595,-0.1283803582,-0.1721372157,0.203258574,0.023101788,0.3418016434,0.0594903454,-0.2116236091,0.038442906,-0.1193577349,-0.7108005881,0.1404067576,-0.2019814551,-0.0142486282,-0.5004389286,-0.1042032763,0.4435341954,0.1194512397,0.1769238859,0.2440028191,-0.3941906095,0.2416638881,-0.1466080397,-0.1694501042,-0.0521142893,-0.0279145222,0.3019589484,-0.2829476595,-0.3729143143,0.2836357653,-0.0017941408,0.0282253213,-0.4600768983,-0.051117491,0.1826188564,0.0197839793,0.2207045704,0.3797985613,-0.1771155,-0.2517739534,0.1449237168,-0.0685282499,-0.3642730117,-0.0433896333,-0.1673242748,-0.244377926,0.027283879,0.3833366036,0.1490146965,-0.1059536487,0.542447567,0.3601782918,-0.2853142917,-0.2442198545,-0.1411839724,-0.1164885908,-0.7241874337,-0.0106912563,0.0764481127,0.0580210313,-0.1022744253,0.2253740132,0.2350110114,-0.0569151305,-0.0101075806,-0.3057074845,-0.252535224,-0.0836961195,-0.0783474147,0.0156389773,-0.0557722822,0.2327729166,0.2338805199,-0.051649522,-0.3866991401,0.0963929445,-0.1417026222,0.157254532,0.3518303335,-0.262668848,0.2583806813,0.0998058021,0.1214568987,0.0368537977,-0.1825639307,-0.1610731184,-0.0565744415,0.1114050597,0.0681372583,-0.0288578086,-0.0404418968,-0.3640963435,-0.0717707425,0.2755315006,0.2100090235,0.1929802746,-0.0068820994,-0.0666660443,-0.0738979951,-0.0178614371,0.2682754397,0.2190890312,0.0859695077,0.405148685,0.1010470912,0.2030219287,0.0525421016,-0.0890607238,-0.2841539979,0.1198029369,0.1499392539,0.0518202074,0.1336272657,-0.0791227669,0.2745363116,0.2660778761,0.1616653353,0.0666780174,-0.0928050876,-0.0901125818,0.4874474406,0.2030070871,-0.2415479422,0.0762771294,-0.1121362001,-0.184881717,0.0124606611,-0.1028384641,0.1134571508,-0.2010246366,0.1946469098,-0.0380837135,-0.0281703267,-0.019415861,0.0293576252,0.3393165767,-0.0418850295,-0.2344798446,0.116925247,0.1433206201,-0.0137236407,0.2548773885,-0.180804953,0.2845473588,-0.0537581369,0.2220527977,0.030296959,-0.3563716114,0.1120446101,0.0706906691,-0.066773966,0.2324639112,-0.0055032913,0.2380449474,-0.0018269551,0.0617609844,-0.2219633311,0.3045001626,0.1409576237,0.0731002763,0.0690347478,-0.2575113773,-0.2445528954,0.0742554069,-0.0776849166,-0.2890951335,-0.1114569679,0.2190827578,-0.2013315111,-0.1336750537,-0.1104746461,-0.1631097645,0.1939913929,-0.3072598875,0.3899513483,0.0809237435,-0.0012229361,0.3901760578,0.3995128274,0.4556732476,0.0792815089,-0.1206799373,0.2690308988,-0.0715339109,-0.1090680957,0.0018282729,0.5251642466,0.3019497693,0.1344434321,0.3727327585,0.157420516,-0.194133386,-0.1215395853,0.0120799402,-0.0314948447,-0.195530951,0.1012361571,-0.2375436872,-0.0593921058,-0.4787736237,-0.1821279526,-0.7445862293,0.0120567055,0.4457138777,-0.2951313257,0.1217688918,-0.2555833161,0.0848087594,-0.1629760265,0.2783347368,0.4391409159,0.2204273939,-0.0607465766,-0.0793667734,-0.5453093648,0.0916669294,-0.1029383242,0.2013344467,-0.2300073802,0.1344800442,0.1855961233,0.1058829874,0.0845326483,-0.0155988196,-0.2895371616,0.1473764032,-0.1530165523,0.0604421608,0.115501456,0.4367619753,-0.0457753241,-0.161592871,-0.0311631933,0.0709580854,0.0324437395,-0.057638716,-0.1000953391,-0.273329407,-0.2293982804,0.2840521038,0.1085324213,0.5500635505,-0.1439040303,-0.0561835095,-0.4176398218,-0.1313671023,0.008235557,0.0296615064,0.1782099307,0.2597198784,-0.1403371245,0.2682725489,-0.0455152132,-0.2081946731,-0.1168610379,0.176353693,-0.3187257349,-0.0397253186,-0.1087751761,0.1464364678,0.0685597733,0.0935342461,0.1244539097,-0.2427161038,-0.3076225221,-0.1533102244,0.5325167775,-0.2489630431,-0.4053457379,0.0748478994,0.4408685267,-0.0181050412,0.1790773571,-0.5474327207,0.3000046313,0.317977339,0.1242563799,0.0236966498,0.1211317554,-0.0514779352,-0.0300046876,-0.0176148061,-0.0739435256,0.1150659099,-0.1863280535,0.2279142588,-0.2254253775]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Also there are huge those datasets. Think downloading MuST-C v1.2 amounts to ~ 1000GB... because there are 14 possible configs each around 60-70GB. I think users mostly will only use one of the 14 configs so that they would only need, in theory, will have to download ~60GB which is ok. But I think this functionality doesn't exist yet in `datasets` no? cc @lhoestq ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":64,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nAlso there are huge those datasets. Think downloading MuST-C v1.2 amounts to ~ 1000GB... because there are 14 possible configs each around 60-70GB. I think users mostly will only use one of the 14 configs so that they would only need, in theory, will have to download ~60GB which is ok. But I think this functionality doesn't exist yet in `datasets` no? cc @lhoestq ","embeddings":[-0.3945277929,0.2756198347,-0.0875779614,0.0062152874,-0.0652416423,0.0286354832,-0.1824718416,0.2685896754,-0.2595131099,0.322671026,-0.2936765254,-0.2957715094,-0.1945616156,0.1870187521,0.2089396715,0.091052793,0.0300666578,0.2597108185,-0.4889114499,-0.1047810093,-0.0244903415,0.1415159255,-0.1815806776,-0.166879341,-0.2045273036,-0.0136402538,-0.3177773952,-0.2916285396,-0.31662184,-0.3389554918,0.2591308355,0.4024871886,0.2682633996,0.2643739581,-0.0001165047,-0.1814153641,0.1977115273,-0.2872515619,-0.4984134138,0.0147202974,-0.2707028091,-0.3995997906,-0.4279793203,0.0496154055,0.0131067624,-0.258695215,0.2993300557,0.1313707232,0.1540212631,0.6043412685,0.1189385355,0.1227991804,0.1979771852,-0.2026145607,0.3769967854,0.0362648554,-0.0884151608,0.2486227006,0.211314708,0.0940595195,0.3303393722,0.3243901134,0.2032753825,-0.1683722734,0.0405362621,-0.3225006163,-0.2506564856,-0.5101647973,-0.0138992425,0.4631100595,0.6668586135,-0.3101228476,-0.1322305351,-0.1701138318,0.2439388782,-0.1759988964,0.0123595623,0.0486717932,-0.2894198596,0.2103223056,0.0058159232,-0.342540741,-0.2323828191,0.3928020298,-0.1653021127,0.0985151157,-0.1363165975,-0.0438101143,0.2220848799,-0.2304327637,-0.2993947566,-0.0486107469,-0.0103297075,0.0844988674,-0.2337271869,-0.4959618747,0.1698435396,-0.0315787084,0.1768458635,-0.4561896622,-0.1986819953,-0.2189819217,-0.2775390148,0.0029941748,0.1145826578,0.3686311543,0.0729001462,0.0945083648,0.1670213342,-0.3023116887,0.1047606245,0.0855693594,-0.1426624656,0.1373975426,-0.4395268559,0.1589257419,-0.1422384381,-0.0756253973,-0.3051211834,0.1805654317,-0.0362609737,0.1540668607,-0.0433287024,0.1592666954,0.0638662949,0.3321661353,0.0086356821,0.2712830901,0.0737075731,-0.2359882146,-0.0156380441,0.0743166059,-0.1880131662,0.2972606719,0.3724639714,0.1695252955,0.2306044251,-0.1234051734,0.2562737465,0.2554161847,-0.1575928777,-0.0771213993,0.0842519179,0.2003100663,-0.0076149185,0.0768662989,-0.3242165744,0.0385288857,-0.3530516326,0.1165569499,0.1461156756,-0.168192476,-0.0042422153,0.1208604276,-0.1454240978,-0.184690237,-0.3141504526,0.6670888066,-0.1658895314,-0.1884400547,-0.0326198004,-0.0620025992,-0.335694164,0.0928697586,-0.2054013461,0.3628073633,-0.4599133134,-0.0312914215,0.0516736023,-0.0194718856,-0.016583059,0.2870403528,-0.1658418179,0.2281011641,-0.1118805334,0.1431228518,0.1121348441,-0.3574815989,-0.4023281932,0.2297324091,-0.0250995513,-0.0528222844,0.3132527173,-0.0154700074,0.6000576615,-0.4032661021,-0.1561005861,0.3050542474,0.0663902983,-0.0649881065,-0.2170980722,-0.5506214499,0.0757225007,0.0820106342,0.2562983334,0.0670404881,0.132770136,0.4450987577,0.4237921536,-0.2013008446,0.1830306649,0.1810790449,0.100528188,-0.076288715,-0.0904819742,-0.3603584468,-0.2245747,0.0752416402,0.1699463874,0.2961947024,0.2997466326,-0.3767054975,0.056624867,-0.2456870228,0.0667728335,-0.0151307788,0.0987125784,0.217659384,0.2734742463,-0.1541106403,-0.1304287612,0.3823821843,-0.4641740322,-0.2312881798,0.0643941164,0.1478845179,0.0651064143,-0.1028501242,0.1676033586,0.0703044012,0.117324762,0.05399625,0.0847353414,0.1348595023,-0.156897679,0.1102482751,-0.0723443776,0.0968137234,0.3385638595,-0.4492703378,0.2706457078,0.36655581,0.1374399364,-0.098910518,-0.2362752557,0.0285662115,0.2820813656,0.1219366714,0.2826952636,-0.1613173783,0.2331681401,0.1416084021,-0.1827303469,-0.3336019516,0.1005608886,0.3341189623,-0.2534903586,0.132493645,-0.2808016837,-0.0448419452,0.5708273053,-0.0869840682,0.3257897794,0.401941359,-0.0270678829,-0.2196687609,0.1617392153,-0.1772956252,0.390566349,0.4227868021,0.4028218985,0.0894212127,0.0876602679,-0.142158553,0.1728164405,-0.048912514,-0.2792555392,0.1039379165,0.0364374854,-0.0634938553,-0.3835653067,-0.1189086139,0.0626145825,0.174564451,-0.1854741871,-0.4444788694,-0.1693077087,-0.1456028521,-0.0709977373,0.0894658938,-0.00649055,-0.0500765033,0.2922297716,0.1339638084,-0.1373102218,0.1132405102,0.0293735042,0.4036889374,-0.3597381711,-0.0985198542,-0.0108860973,-0.1740555018,-0.3488500714,0.1027490795,0.1526207179,0.0351412371,0.0346483253,-0.2288288027,0.1461686343,-0.1769774258,-0.4210931361,0.2096335441,-0.1606576443,0.031497471,-0.1883480847,0.1851782054,0.0184183549,-0.2307129204,-0.3356032073,0.0843583792,-0.1430410296,-0.0794477686,-0.471226871,-0.0872177333,-0.0490198322,-0.3670856059,-0.4792002439,-0.4064483047,0.4187769592,-0.1318330318,0.1176661626,0.1957024634,0.1825722158,0.0393263698,-0.244108364,0.1794232726,-0.1795466244,-0.2061038315,0.3044100404,-0.2531206906,-0.1693060696,-0.1619277447,0.1287616938,0.0721429884,0.1599912196,-0.4716989994,0.2716136575,-0.3734459579,0.3990853131,0.0809294805,0.1973962337,0.3172207177,-0.2313880026,-0.0570689291,-0.036183238,-0.1424661726,0.037234012,0.4006441832,0.1502839774,-0.1070050672,0.0874566585,0.2645626664,0.6685683131,0.2341575176,0.0370725654,0.1252324134,0.1093166023,0.6374338269,-0.2302099019,-0.18133609,0.4156431556,0.2194651067,0.0160645042,0.4394763112,0.0398073755,-0.3308630884,-0.1140271425,-0.2987889051,-0.0439490639,-0.263680011,0.2918449342,-0.0301596634,0.1868848354,0.0441701375,-0.1829177588,-0.1949278563,-0.1114958152,0.0028252872,0.1889383942,0.0259738434,0.153516382,-0.1168704554,-0.1004865095,-0.2401667982,-0.0825168118,0.0902085379,0.3649488389,-0.3448270559,0.0559172444,-0.1013513952,0.1461903453,0.3217300475,-0.2075681537,-0.1045048609,0.0868100822,0.6164704561,-0.2222564816,0.099129118,-0.1585009098,-0.4502093792,0.5935949683,0.1860441417,-0.2508350015,-0.0868009329,0.2934371233,0.3685335517,0.0084277932,-0.1660437286,0.0170777105,-0.2766517699,-0.4073575735,0.3156574368,0.3849785924,-0.1538506597,0.2192331553,-0.0300270431,0.0518158153,-0.2847202718,0.3886928558,-0.0245376732,-0.1007304192,0.5650273561,0.1747563928,-0.0981945023,-0.4175640047,0.00926999,0.3048287034,-0.0742583945,-0.3260020018,0.0351705104,-0.0817231089,0.3339954317,0.4596893489,-0.1333194673,-0.0458886921,-0.1416230798,0.0858749822,-0.1006511152,0.4539083838,0.1133180857,-0.0340534784,-0.2912560105,-0.3026304543,0.4794850349,0.0659982935,-0.1266180128,0.202790916,0.0990391448,-0.3326151967,0.0656473562,-0.1355231553,1.2328360081,0.1060664356,0.1131637916,0.021870207,-0.2788861394,0.2806355357,-0.6616119742,-0.01412261,0.4549615681,0.1489109844,-0.0561077818,0.2230758667,0.1939227432,0.3626054227,-0.3357535005,0.3182626963,0.1121503338,0.1734455973,0.2129337043,0.5456153154,-0.5706213713,-0.3597718179,-0.1181831211,0.1514992416,0.1357637048,0.1687722206,0.0115254419,-0.1516532004,0.3683602214,-0.0621875711,-0.638694942,0.1000993475,-0.3431992829,-0.1112855598,-0.5482454896,0.0690919012,0.3296595216,0.279799372,0.2434121966,0.369620949,-0.3438278437,0.2961897552,-0.0447206907,0.029232081,0.1249240488,-0.179690972,0.1681115031,-0.3176144063,-0.1660713255,0.1971087307,0.1828442663,-0.0068474649,-0.3814734519,-0.1982239932,0.1744095534,0.2231499106,0.1668538898,0.633222878,-0.0268852413,-0.1847649515,0.0941915885,-0.0539355539,-0.2978715897,0.2821992338,-0.4007060528,-0.2325912416,-0.3334186077,0.2964728177,0.2167977542,-0.0397381373,0.2869243026,0.0810226798,-0.2079164237,-0.3785052598,0.0075603887,-0.2319057286,-0.4202309549,0.0416329168,0.0250498764,-0.1195923463,-0.3540137708,0.4248914123,0.2778798342,-0.1271970123,0.1037697643,-0.0079147415,-0.3289754987,-0.1241019145,0.1137878522,0.2326824516,-0.0208346285,0.2385651916,0.037386924,0.0770753175,-0.3000097275,0.0434967615,0.1029208899,0.1660790145,0.1094023213,0.1252080947,0.2947096527,0.0223572291,0.0217253678,0.1234148815,-0.0627958477,-0.0628796816,-0.1510152072,0.1382300407,0.1991150528,0.2511877418,-0.0107895434,-0.1329907179,-0.1390428096,0.0840051174,0.374818027,0.3341172338,-0.1051124707,0.2455533296,-0.0621534064,-0.1095393822,0.1172342002,0.0981831551,0.3158512712,0.3510113955,0.2124059349,0.3487104177,0.0388750173,-0.1696397066,-0.0879378468,0.4134814441,-0.02031257,-0.0497466065,-0.0338473432,0.1199689135,0.0941870436,0.2626902163,-0.0582996756,0.2162435353,0.0737694055,-0.0963565484,0.4845879078,0.1590505689,-0.1836368442,0.1744567603,-0.3760052621,-0.1898871511,0.0168813597,0.1298238337,0.0204854757,-0.196072638,0.439422071,-0.0681829005,-0.2496244609,0.1491870731,0.1073622629,0.2614358664,0.1854943782,-0.1375294477,0.1784852147,0.0797981098,0.234360069,0.2207279205,-0.3789799809,0.2936010957,0.1280773282,0.1260855943,0.193224147,-0.3421127796,0.1357569993,0.2370891124,0.0199661311,0.116063714,0.0408359244,0.0434151739,0.002709623,-0.0969308242,-0.3338383436,0.0546516553,0.0608656183,0.0003472995,-0.1674048454,-0.2518033981,-0.0856498554,0.178696081,0.0157079715,-0.5163857937,-0.034134116,0.2421743572,-0.1619973779,-0.3491897881,0.0915440395,-0.2302365303,0.1470195502,-0.3118005395,0.4185451269,0.0030815573,-0.1227405742,0.2632627189,0.3552525938,0.3930789828,0.0928429812,0.0519843288,0.1928476244,-0.0420673788,-0.0652753413,-0.1556446105,0.345625788,0.0366828367,0.4378799498,0.2163713872,0.1574176401,-0.157379508,-0.1653935611,0.0571069121,-0.1900061369,-0.362746805,0.2912068069,0.0606487431,0.1902568638,-0.2029947788,0.0283360295,-0.5914896727,-0.0446676202,0.4822023511,-0.3417671919,0.2164019197,-0.2335925251,0.0536475405,-0.2062036097,0.2428196669,0.3525647521,-0.0174205899,-0.2116129696,-0.0569505692,-0.2593861222,-0.28693223,-0.1189725846,0.0825195983,-0.3457997143,0.2447002977,0.2578915954,0.3945356309,-0.2088105232,0.2064208835,-0.2922370732,0.0576983206,-0.2314187735,0.2513568401,0.0288168378,0.6024891734,0.0707646832,-0.3821687698,-0.1606886685,0.0845788792,-0.0325326137,-0.2092237025,-0.0033529487,0.0119187525,-0.1906455457,0.3746901155,-0.0606209114,0.5547391772,-0.1299277097,-0.0191289708,-0.2614654303,0.128442362,-0.1860150695,-0.1669572294,-0.0351503715,0.2119604051,-0.3294379711,0.3177522421,-0.0082024653,-0.0343666896,0.0126503278,0.0495773628,0.025830362,-0.0817406178,-0.207535401,0.1316098422,0.0134994406,0.2541102171,-0.0143941855,-0.1189340428,-0.2897935808,-0.2972885668,0.3701490164,-0.1548674405,-0.2318077236,-0.149801895,0.3886859417,0.2876119316,-0.0012168131,-0.5777561069,0.2842466831,0.3728408217,0.1047302485,0.0706949458,0.2855329216,-0.0053192275,-0.0531589612,-0.1251319051,0.1858729869,0.078461796,-0.383341074,-0.0848382711,-0.4800366759]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"> Also cc @lhoestq - do you think we could mirror the dataset?\r\n\r\nYes we can mirror it if the authors are fine with it. You can create a dataset repo on huggingface.co (possibly under the relevant org) and add the mirrored data files.\r\n\r\n> I think users mostly will only use one of the 14 configs so that they would only need, in theory, will have to download ~60GB which is ok. But I think this functionality doesn't exist yet in datasets no? cc @lhoestq\r\n\r\nIf there are different download links for each configuration we can make the dataset builder download only the files related to the requested configuration.","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":110,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n> Also cc @lhoestq - do you think we could mirror the dataset?\r\n\r\nYes we can mirror it if the authors are fine with it. You can create a dataset repo on huggingface.co (possibly under the relevant org) and add the mirrored data files.\r\n\r\n> I think users mostly will only use one of the 14 configs so that they would only need, in theory, will have to download ~60GB which is ok. But I think this functionality doesn't exist yet in datasets no? cc @lhoestq\r\n\r\nIf there are different download links for each configuration we can make the dataset builder download only the files related to the requested configuration.","embeddings":[-0.3701684475,0.0756932944,-0.0500222258,0.1428010464,-0.1143720597,-0.0225355364,0.0245045293,0.2528821826,-0.2171293497,0.2163129747,-0.3634768128,-0.2713863552,-0.0952886492,0.2548368871,0.1178233773,-0.2256662697,0.0715246871,0.0628933534,-0.4934399128,-0.1806575656,-0.0167998634,0.0800376758,-0.077132009,-0.160868004,-0.2286911756,0.0514315665,-0.3497377634,-0.2192535102,-0.3488781154,-0.2112865299,0.0303797573,0.2346605062,0.2090707272,0.353375107,-0.0001047189,-0.0938933566,0.2101813108,-0.2275529355,-0.4023526311,-0.1474176943,-0.1694556326,-0.216504693,-0.4638263285,0.0432070866,-0.2301793396,-0.193292141,0.1055440456,0.0908902064,0.3507657051,0.3985465169,0.2088294625,0.1681186706,0.1384950131,-0.1770462096,0.1294429749,-0.027113013,-0.0644722506,0.1953025311,0.2781899273,0.2423403263,0.0250680596,0.4822728038,0.0510125756,-0.1121770144,0.0645824596,-0.3023509383,-0.1228587478,-0.4970511198,0.0585182793,0.4549770355,0.7421863079,-0.2327513993,-0.2601963878,-0.3296455145,0.2417101711,-0.0608051941,-0.0412625857,0.1268950254,-0.2630253732,0.3054477572,-0.0407549068,-0.3075760901,-0.1850809753,0.3004446924,-0.0246442109,0.2349756211,-0.1525672972,-0.0494456552,0.0938344225,-0.3000545502,-0.3330549598,-0.1387972087,-0.0195224248,0.1839544177,-0.1306201965,-0.3575485945,0.1271692067,-0.0699242428,0.2695179582,-0.2468396872,-0.1844830364,-0.1602424681,-0.108653605,-0.0174748898,0.1974957436,0.4110043943,0.050860066,-0.0448150188,0.0490414798,-0.2103945464,0.083898291,0.0806303397,-0.0542864203,0.2022230923,-0.4636357725,0.0314629823,-0.0657023564,-0.1599666476,-0.1069284752,0.0018301914,0.0456215851,0.0221909508,0.0139318323,0.2629641593,-0.0254197288,0.2773423493,-0.0889125466,0.3318893909,0.0076956139,-0.2727166414,-0.0556416027,0.0387540944,-0.2634615898,0.2528849244,0.2420615256,0.24993819,0.2341479063,-0.0182297472,0.2714515328,0.0734630302,-0.1332444698,-0.0212271083,0.1294781268,0.0178689677,0.0261993352,0.1165113747,-0.2889254987,-0.1574309021,-0.3001947701,-0.0070913052,0.0725396276,-0.1672895104,-0.1503563076,0.2312588841,-0.2765887678,-0.2198792547,-0.3784280121,0.7613043785,-0.157165423,-0.0671317205,-0.0258608311,0.1604849994,-0.4984615147,0.0312190149,-0.0184656791,0.572507143,-0.2136995643,-0.2364332378,0.1852799058,-0.0407377779,-0.0329288356,0.3206631839,-0.1548604965,0.2370501459,-0.064322941,0.3969767392,0.0862114877,-0.3365670145,-0.3337926269,0.200361982,-0.163074404,0.0126820942,0.2968087792,-0.1083985344,0.6236365438,-0.3085472584,-0.0656214729,0.4407989085,0.0436888039,0.0763146728,-0.2191424817,-0.441157192,0.1737977117,0.2197506577,0.0945778117,0.0077126264,-0.0541920774,0.330504328,0.2851980925,-0.277430892,0.3457450271,0.1845484078,-0.0375552997,-0.0314101912,-0.0289490111,-0.249455139,-0.2396082282,0.0701973289,0.1038628593,0.4097924531,0.2049586326,-0.2575562298,-0.143522203,-0.2272382975,-0.2329358906,-0.1436645091,0.2024192512,0.1730025411,0.2226167321,-0.0140872439,-0.1622647941,0.3276186883,-0.3089619577,-0.1129497215,-0.1202503964,0.2134663165,0.0581184216,-0.1602449566,0.2480307221,0.326151222,0.1055110171,0.0229459014,0.081945397,0.1655614227,-0.2360561639,0.2517371774,0.1058901474,0.232981652,0.3496467471,-0.4438665807,0.3718279004,0.3226096332,0.1047069579,-0.1649798155,-0.2093898207,0.111028187,0.3269998133,0.1179435924,0.1451898664,-0.0675860122,0.2856110334,0.0367305614,-0.2883065939,-0.1736773551,0.111066319,0.1211387739,0.0097599858,0.1580423117,-0.3687121868,0.1698787063,0.4562642574,-0.1424197406,0.1689661145,0.3246617019,-0.2069304883,-0.1832818389,0.0871323794,-0.1043376699,0.2261552662,0.3713853061,0.3932768404,0.1242209226,0.2300581932,-0.0834785402,0.1232806593,-0.1006213874,-0.2778021693,0.1717960685,-0.006828893,-0.0828555673,-0.3244183958,0.0117567005,-0.0339590013,0.1024274155,-0.1919405013,-0.4238792062,-0.0389805734,-0.4275493026,-0.3429191113,0.0687207654,-0.2192068547,0.0812050998,0.3772016764,0.0171472691,-0.2011986673,0.19204171,0.2446411103,0.2857397199,-0.3791689873,-0.0679683313,-0.0292636752,-0.2193164825,-0.1448599696,0.159105584,0.2893210351,0.1332319826,0.2628844678,-0.2597988844,0.0545173064,-0.2878354788,-0.3296543062,0.259139508,-0.1816742271,0.0342454389,-0.2146514505,0.1287184358,-0.0098619349,-0.3360490799,-0.1346302032,0.1360162795,-0.0662714913,-0.1435022801,-0.3694386482,-0.1905807704,0.0158170145,-0.4421801865,-0.5470240712,-0.4363861084,0.4418897927,-0.2328890115,-0.0117504233,0.0389080197,0.1485845745,0.0861272812,-0.261469692,0.199953258,-0.2780721784,-0.2521601319,0.1967475414,-0.2384424657,-0.3239270747,-0.2167912126,0.1482231617,0.0329869427,-0.0588274375,-0.3675884604,0.0656409115,-0.2171866298,0.3849049211,0.1657076925,0.3021327853,0.2084798664,-0.0548925586,-0.0997457579,-0.0484945402,-0.1371175051,0.0944610164,0.2940378487,0.120740898,-0.0373637341,-0.013512589,0.0706404895,0.7331622243,0.4433308244,0.0359318964,0.1378315985,0.0668264329,0.4425567687,-0.2520739734,-0.2575766146,0.4864844978,0.1110296994,0.0480158217,0.5023177862,0.1446164995,-0.3531343639,-0.1621280313,-0.382789433,-0.1612782031,-0.3055439293,0.3122988045,0.1487029046,0.1491455585,0.0018653232,-0.2088144273,-0.1799126267,-0.0995178893,0.1411103457,0.1865564436,-0.0672528148,0.1155247986,-0.2071308494,-0.0609681085,-0.4196372926,-0.1190639734,0.1540755183,0.1639531404,-0.3308215737,0.0645854846,-0.0311972443,0.1847550422,0.3390976489,-0.1980142593,-0.0449628644,-0.0339771286,0.3296681345,-0.0582750626,0.2281969339,-0.1363882422,-0.3484121561,0.5243244767,0.3274693489,-0.4339472651,-0.0851388648,0.1836047769,0.4520294666,-0.045954261,-0.2049321383,0.168930456,-0.3000575304,-0.4342628717,0.2032826692,0.4088584483,-0.111446999,0.1911326051,0.0783083588,0.2157871127,-0.0504506901,0.3658590913,0.1175390333,-0.0030350937,0.4130165577,0.2029377818,0.0896573216,-0.0690987632,-0.0898935869,0.3772088885,-0.0236030445,-0.304323256,-0.1228891537,-0.2131722867,0.2043744177,0.3947955668,0.093825236,-0.0437687598,-0.0077697416,0.1236337423,-0.2165962756,0.1542898864,0.1270649433,0.2224366516,-0.316065371,-0.3919180036,0.5125018954,0.0450153649,-0.0917449743,0.0586715378,0.2783919573,-0.2230504751,0.1936855316,-0.2150403112,1.2153538465,0.1881245971,0.1148182079,0.2017800659,-0.2469555587,0.202308625,-0.3942833543,-0.1897048652,0.272105217,0.2457346916,0.0009764633,0.2117398381,-0.081332162,0.2784785032,-0.4962832928,0.2711896002,0.1113902852,0.0455697402,0.0883855447,0.5809120536,-0.4232159555,-0.4742906094,-0.3706196845,0.2387131155,0.0902496427,0.1875306666,0.1156489104,-0.2339893579,0.1649766713,-0.0428300388,-0.6325404644,0.1084037274,-0.2850007415,-0.2105822563,-0.3871232569,-0.0091045601,0.3368227184,0.2709598839,0.2840493917,0.0197892487,-0.3268595934,0.2202637047,-0.0845438316,0.0496505238,0.0946395844,-0.0991226435,0.4290618002,-0.3332925141,-0.2859706581,0.0680474713,0.1607716829,0.1462593526,-0.4129219949,-0.2794246078,0.0710077137,-0.0287159905,0.134100914,0.5576699972,0.0208147522,-0.1288644373,0.1656281352,-0.0068200561,-0.4491666555,0.2423336804,-0.1395210773,-0.2203085423,-0.1989901513,0.2189774662,0.2148139626,-0.0585966669,0.3552798331,0.2256357074,-0.1796319038,-0.3044047356,-0.0684861839,0.0781839788,-0.5221439004,0.0359207056,0.0717630535,-0.1466501951,-0.3378514349,0.2661674619,0.2490734607,-0.1029828116,0.0247283597,0.0002229747,-0.2718348503,0.0264808387,0.1696318537,0.1759607643,-0.1232426092,0.2195599228,0.1706393659,-0.0780082792,-0.4171719849,0.0327237323,-0.0236655567,0.1987486184,0.3878056407,-0.0310051329,0.1562348902,0.0350583605,0.1031294912,-0.0903012678,-0.2246175706,-0.204034701,-0.2099707723,0.095698446,0.1128562912,0.1228678003,0.0327092372,-0.1171437055,-0.1364217103,0.0591907427,0.3521561027,0.3154357076,-0.0766702965,0.2034246325,0.0385507531,-0.0821070969,0.0648195371,0.1983171999,0.3428341746,0.3074676991,0.1040443033,0.2209674567,0.0071879402,-0.1154131442,0.0740618482,0.3046558797,0.1857195497,0.0020022306,0.0726080835,0.1045009196,0.2016433179,0.2957168221,0.1713349968,0.1734622121,-0.0500459969,-0.0748887956,0.4209257066,0.253965795,-0.1479903907,0.172418654,-0.0125950035,-0.247897014,0.1724132746,0.1482937485,0.1070315018,-0.0411763377,0.2748100162,-0.1327527165,-0.115359962,0.1135308817,0.0660153553,0.2415055484,0.1296439171,-0.0722672716,0.0194690879,0.1619671881,0.1729011238,0.025602689,-0.2666923404,0.3512324691,-0.0096801007,0.2321716696,0.2002370656,-0.1344335526,0.1186261252,0.1756958663,0.1315770596,0.2086507082,-0.0427125022,0.3043555021,0.0406306349,-0.2704007924,-0.1132573932,0.3184616268,0.090886727,-0.0968172252,-0.1420929581,-0.2933769524,-0.3084364831,0.1166450232,-0.069963634,-0.4248723388,-0.0010498769,0.1826612502,-0.0767158791,-0.197878167,-0.0527083352,-0.1791321635,0.1575374007,-0.2776763737,0.3628771603,0.1634909809,-0.1249058768,0.3019227386,0.559583962,0.4070015252,-0.0130083645,-0.0583064742,0.3310625553,0.0331274197,-0.0445070602,-0.1409436613,0.2963184118,0.0173040666,0.0616049618,0.2934182882,0.1977273673,-0.2278093249,-0.2308119982,-0.028046919,-0.2064989358,-0.2112592459,0.0871775746,-0.07415957,0.1321627349,-0.4074139595,-0.1699429303,-0.6608303189,0.1593287289,0.4730758369,-0.2615089417,0.1429408789,-0.1474593282,0.1028150544,-0.1227620617,0.359965235,0.3447949588,0.0295283515,-0.1041114703,-0.1327614188,-0.4292390347,-0.2139898241,0.0377320088,0.0893614665,-0.2855095267,0.150934875,0.2879434228,0.3707047999,-0.1318272948,0.0075488249,-0.4060830474,0.0943480954,-0.2334735245,0.1123217344,-0.0499989204,0.628600657,0.1807159632,-0.13378492,-0.178953886,0.1858935803,0.029987378,-0.0430675559,0.1070524678,-0.202740252,-0.4029411972,0.0968330279,0.0081002032,0.4688538611,-0.225196436,-0.0255574733,-0.1893644333,-0.034189254,-0.3663251698,-0.0267943814,-0.0259112623,0.2615010738,-0.2757114768,0.2454531342,-0.0735746026,-0.1564108878,0.0753840506,-0.0169627387,-0.2175990045,0.0414990447,-0.1102007106,0.0860516056,-0.02242193,0.3933387399,-0.1096891463,-0.0977682322,-0.2531040311,-0.2490006983,0.4857107401,-0.146196574,-0.2277169377,-0.0556891374,0.1905224919,0.1770900488,-0.0430363081,-0.7272621989,0.3224762082,0.2632123232,0.1082136929,0.0752158836,0.2462015748,0.0293012168,-0.0844959766,-0.1843973696,0.0642651394,0.1196352392,-0.2461562306,0.1379872411,-0.3459208608]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"I have written to the dataset authors, highlighting this issue. Waiting for their response. \r\n\r\nUpdate on 25th Feb: \r\nThe authors have replied back, they are updating the download link and will revert back shortly! \r\n\r\n```\r\nfirst of all thanks a lot for being interested in MuST-C and for building the data-loader.\r\n\r\nBefore answering your request, I'd like to clarify that the creation, maintenance, and expansion of MuST-c are not supported by any funded project, so this means that we need to find economic support for all these activities. This also includes permanently moving all the data to AWS or GCP. We are working at this with the goal of facilitating the use of MuST-C, but this is not something that can happen today. We hope to have some news ASAP and you will be among the first to be informed.\r\n\r\nI hope you understand our situation.\r\n```\r\n\r\n","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":147,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nI have written to the dataset authors, highlighting this issue. Waiting for their response. \r\n\r\nUpdate on 25th Feb: \r\nThe authors have replied back, they are updating the download link and will revert back shortly! \r\n\r\n```\r\nfirst of all thanks a lot for being interested in MuST-C and for building the data-loader.\r\n\r\nBefore answering your request, I'd like to clarify that the creation, maintenance, and expansion of MuST-c are not supported by any funded project, so this means that we need to find economic support for all these activities. This also includes permanently moving all the data to AWS or GCP. We are working at this with the goal of facilitating the use of MuST-C, but this is not something that can happen today. We hope to have some news ASAP and you will be among the first to be informed.\r\n\r\nI hope you understand our situation.\r\n```\r\n\r\n","embeddings":[-0.1994687915,0.2436533868,-0.0974727422,-0.0784687102,-0.0106336242,-0.1451659054,0.0524760596,0.0750111043,-0.1210070625,0.3005029559,-0.1688125432,-0.1214635521,-0.0929058343,0.1014922336,0.049240537,0.0520469919,0.0159369204,0.1417456716,-0.4235424995,-0.2137577981,0.0676920414,0.1652442366,-0.1971707344,-0.1185603291,-0.2417682558,-0.0051750476,-0.1987970322,-0.1627392024,-0.2773719728,-0.1755132079,0.3096407056,0.1861113161,-0.0493674017,0.4138047099,-0.0001041711,-0.0886685774,0.268043071,-0.2699055076,-0.5759402514,-0.1869329661,-0.269900769,-0.301694572,-0.3962218165,0.0005752879,-0.1565587819,-0.2501755059,0.1901202202,0.084521383,0.1982961744,0.5663022399,0.2318571657,0.1294274628,0.1595150083,-0.2813528776,0.2131444067,0.1161144152,-0.1201322004,0.3304124773,0.3230338991,-0.0537693277,0.0512337647,0.335177809,0.1430394351,-0.1815527827,-0.0345126614,-0.4015649855,-0.1171030998,-0.2742467523,0.0207959507,0.202059269,0.6257651448,-0.2948607802,-0.1283823252,-0.0489043817,0.2616833746,-0.0491095744,0.073444508,-0.0219563041,-0.1568531096,0.2644263804,0.2554604709,-0.1469140202,-0.3002218008,0.3491802216,-0.1034195945,0.2759414613,-0.1623297334,-0.0958297029,0.0670067444,-0.1347406358,-0.19822523,0.0434117056,0.0794893727,0.0268600471,-0.2630914748,-0.2374757975,0.19051148,-0.129990682,-0.0011346381,-0.2518422604,-0.033065442,-0.1378911436,-0.2423718125,0.1690661758,0.0275314618,0.3915992975,0.0478057414,0.025896851,0.0454382002,-0.1120334715,0.016895419,0.0804202929,-0.0751743764,0.1211415753,-0.4399229586,0.2164539099,-0.11259491,-0.0830751881,-0.3828395903,0.2213929892,0.0121347317,0.2058012635,-0.0827772468,0.2824524045,0.112904951,0.3342103958,0.0786044076,0.3183695972,-0.0523516759,-0.2399520874,-0.0853364095,0.1575847566,-0.3082746267,0.1993889362,0.314044863,0.1970307082,0.2421513349,0.0385026112,0.3347852528,0.1033802405,-0.2505058944,0.0173218437,0.162882641,0.1811018288,0.0541346706,0.1432269216,-0.126215741,-0.107607834,-0.1103474945,-0.007238606,0.1177454144,-0.1387742013,-0.1597446948,0.2793841064,-0.0914549604,-0.1464096904,-0.2564215958,0.4864542186,-0.1666961461,-0.2322057635,-0.0334202461,-0.0739752576,-0.2427843958,0.0647967979,-0.0570576787,0.3744244277,-0.4999311864,-0.0448739864,-0.0608987287,0.0390914753,0.215345636,0.2850775421,-0.1969279945,0.2209675163,-0.1583905965,0.1526072174,-0.0754710436,-0.3909397125,-0.4706937075,0.2447117716,-0.1340881139,0.0226223562,0.2044219971,-0.1413873583,0.4970313311,-0.3135374486,-0.0905554816,0.2267127484,0.1310690939,0.02090111,-0.3485865891,-0.263844192,0.2214559317,0.0960479453,-0.0052674101,0.0703194365,0.148497045,0.2947321534,0.2475513816,-0.1966798306,0.1935806721,0.2508864701,0.0429260284,-0.1177087203,0.0268251635,-0.0362832285,-0.2171421945,0.0160815232,0.0533794314,0.5594960451,0.0495687462,-0.1825216413,-0.2243374437,-0.0805802122,-0.0910207182,-0.0941779092,0.2362570465,0.2504659295,-0.0004784898,-0.0016528046,0.0261907969,0.3673695922,-0.5647367239,-0.3607930243,-0.0539700687,0.3357362747,-0.0739974231,-0.1965979487,0.1826985776,0.1574588567,0.2013040483,0.0631332994,-0.0389977545,0.0641008988,-0.1801032275,0.3866425455,-0.0082447873,0.0721091703,0.2638548911,-0.611143589,0.2364506274,0.3763166368,0.2113754749,-0.2172627896,-0.1177735776,0.1943618804,0.2179944962,0.1046343371,0.1812203974,-0.1003988683,0.4180644453,0.0332072824,-0.1543388665,-0.218464002,0.1855140924,0.251880765,-0.2344928682,0.0502601229,-0.2499331385,0.1158877611,0.5303115845,-0.0432329327,0.2990735471,0.2960605025,-0.0204640534,-0.2250282317,0.143171221,-0.1725830883,0.3313243985,0.3123635352,0.1839494556,0.0182741731,0.1111947298,-0.231856972,0.2586854994,0.0642221645,-0.1848547012,0.266246587,0.1669276655,-0.0963255689,-0.4054495394,-0.0904437304,0.1006516665,0.1656437814,-0.2015103102,-0.2808836401,-0.0934414566,-0.0040671653,-0.034043774,0.0120926248,-0.1151717305,-0.1763385832,0.369687438,0.1451015621,-0.2763325274,0.2118778378,0.1701673269,0.2720695436,-0.3504620492,-0.1271258742,-0.2708378732,-0.3859514296,-0.3327582479,0.1698759198,0.1569398791,0.2146903276,0.1392999589,-0.1036741734,0.1219480932,-0.2698876262,-0.3286662102,0.1964510828,-0.2829557955,0.192641288,-0.1522850692,0.1086609364,-0.0662890077,-0.3872899115,-0.1975033581,-0.039969027,-0.1109140515,0.0299118198,-0.4460125864,-0.0419856459,-0.1716645807,-0.4508725703,-0.5141522884,-0.487308234,0.1961442381,-0.2305824757,0.0891241953,0.2859636545,0.1802773476,-0.0689475536,0.0484408662,0.1605442464,-0.3384575546,-0.3026748002,0.2091612816,-0.2544575334,-0.2924198508,-0.1641889513,0.0963803083,0.2745873034,-0.1203303561,-0.3925468028,0.2880357206,-0.164059937,0.3043846488,0.047377564,0.0017230014,0.3503230214,-0.1856275499,-0.1792620718,-0.2170024067,-0.1752566546,0.0425246954,0.3976612389,0.2070634514,-0.0541818067,0.0527150743,0.1086416692,0.5772771239,0.1353759617,-0.0217989199,0.2615281641,0.0909397826,0.4808558524,-0.2286502868,-0.3318850994,0.309971422,0.1430054009,-0.0537471808,0.4422291219,-0.0603722446,-0.3396673799,-0.0721123889,-0.2483985722,-0.2171343714,-0.2148475796,0.289090544,0.0168242287,0.2655291259,0.0736431628,-0.1033728793,-0.2026956081,-0.1360356063,0.1813147664,0.1113092676,0.0707899928,-0.01741785,-0.0759070516,0.0236570127,-0.2593643367,-0.0969192237,0.1112842783,0.3616998196,-0.4803868234,0.030534368,0.0340186544,0.0966955125,0.3674912751,-0.2934884131,0.0435046926,0.3052765131,0.6362546086,-0.1275075376,0.1157651991,-0.2674595416,-0.3658494949,0.6204192042,0.1698512882,-0.2620614469,-0.0169065576,0.0119900014,0.2751437128,-0.015639903,-0.2267960757,0.0338333249,-0.3036283553,-0.3658412695,0.2329225093,0.4387885332,0.0009313525,0.0862246975,0.0877385959,0.1206988841,-0.1233061701,0.4253797829,-0.0216817856,0.0568758994,0.46729213,0.1263452172,-0.0997461453,-0.2278054357,-0.0194397699,0.4284514487,0.0437315442,-0.2520927489,0.1931254119,-0.1505157202,0.3266113102,0.3590150774,0.0778014511,-0.0811978355,-0.10516458,0.0309057869,-0.1180353165,0.2863197625,0.1681916267,-0.0060648839,-0.2975889444,-0.1450776011,0.2908435762,-0.1543271393,-0.1423891932,0.0550687313,0.1246234179,-0.1998981535,0.2963703573,-0.1650328934,1.2194262743,0.2236503363,0.0149555309,0.1721476018,-0.2562721968,0.0460315049,-0.5465513468,-0.1011935174,0.285153836,0.0596445724,0.0313027538,0.2506734133,0.0618708804,0.3207922876,-0.3977722526,0.1666983813,0.0624936111,0.0478258245,0.2526816428,0.4872466326,-0.2954547405,-0.4569347799,-0.4111875296,0.2819785476,0.1554124355,0.1700683981,0.1263743937,-0.2567602992,0.3119806349,-0.0656572431,-0.4868173897,0.2926011384,-0.2412975281,-0.0754534975,-0.414952755,-0.0019799753,0.5006420612,0.1465868205,0.1877941638,0.1889207959,-0.2083657384,0.3673810959,0.0740620717,0.0009789366,0.1497964561,-0.0317180716,0.3177677691,-0.2601958215,-0.296261102,0.2222318798,0.0458432026,-0.128356725,-0.3481412828,-0.1422816217,0.134333089,0.1277084798,0.1980601847,0.4806861281,-0.1119991988,-0.3046712279,0.1981636584,-0.1300949603,-0.3827234209,0.2528731227,-0.2519673407,-0.2997049689,-0.2446004897,0.381395638,0.1768239588,0.0369800143,0.4371746182,0.2896382809,-0.2361137718,-0.4695902765,-0.1110738963,-0.0607451089,-0.6077985168,0.0801746994,0.1317833215,-0.1747783422,-0.1215533987,0.4586311579,0.1244470552,-0.0960196108,-0.0495023802,-0.1959708482,-0.3087390661,-0.1802862436,0.0763669387,0.1393088549,0.08406014,0.2793385088,-0.0072218957,-0.0307104159,-0.4479880631,0.0897060782,-0.15544568,0.1697259098,0.0266741458,-0.1896656752,0.4018848538,0.0606179237,0.1980850399,0.105988048,-0.0371597074,-0.2394753397,-0.147430703,0.0904164091,0.190102011,0.3105753958,-0.1229280531,-0.1192799658,-0.1477823406,0.1248560622,0.3903586268,0.2951332033,-0.1126846895,0.2286197245,-0.1275033951,0.1538881809,0.207176432,-0.1238397434,0.2368000597,0.3739461899,0.0692305118,0.1296351552,-0.0193296615,-0.1698899865,-0.0997857004,0.3650918603,-0.0363295749,0.1720930934,0.137066558,0.036598891,0.1279011965,0.1474026591,0.016156638,-0.019126758,-0.0250654649,-0.2337155938,0.4090495408,0.2981198728,-0.3287235796,0.1198657379,-0.4833682477,-0.1793806255,0.123743996,-0.0178300384,0.1064965799,-0.1830837131,0.4820239544,-0.0061358423,-0.1658016294,-0.0534811914,-0.0371487103,0.1106409356,0.0659909695,-0.0492978655,0.1264679879,-0.0264140628,0.0328797773,0.1168873012,-0.2144123465,0.4423989356,0.1650188863,0.2033480704,0.1549123377,-0.2464078516,0.063534908,0.1194192097,0.0035676623,0.2820298076,0.0636543855,0.165080294,-0.07727395,-0.1518947929,-0.2547220588,0.0581931621,-0.0191448554,0.0379880145,-0.2861798704,-0.199913159,-0.1709945649,0.048519671,0.0050861086,-0.2091922313,0.0001082246,0.1914760321,-0.2331335247,-0.2815388739,-0.1237732023,-0.1230673045,0.1728279144,-0.2622758746,0.4665756524,0.1278894246,0.0108112367,0.1999354661,0.3350885212,0.4866558313,0.0178168267,-0.0781432688,0.0981005132,-0.0484185033,-0.1608693451,-0.0345884264,0.4915324748,0.0595955104,0.2903469801,0.2333680838,0.2629184425,-0.2601223588,-0.2237098515,0.2366567552,-0.0909164324,-0.1318209618,0.2377182841,0.0385786183,0.1602681577,-0.4732820392,-0.0502791479,-0.6061915159,-0.0987738967,0.3786172569,-0.2255300879,0.2581450939,-0.2684156895,0.1497335434,-0.2498272955,0.33779037,0.3482749164,-0.014710783,-0.2086762488,-0.1243183091,-0.2474950105,-0.0501402356,0.0780658126,0.0436137803,-0.4200605452,0.1117460504,0.4651495516,0.1475233734,-0.1727710515,0.1116666049,-0.294639796,0.073542282,-0.3131275773,0.3269434869,0.1025161371,0.4720641077,0.0635336265,-0.2327403128,-0.3049162626,0.2680441737,0.0959119201,-0.1254855692,-0.1032614857,-0.1006206647,-0.0835478678,0.3423159122,-0.0357189998,0.4685291648,-0.2829302847,-0.0095803337,-0.2334048748,-0.1146837771,-0.046146255,-0.035692215,-0.0001259329,0.2495294213,-0.0652912855,0.2796507478,-0.0938510001,-0.067636095,0.0154786864,0.063247107,-0.2129353285,0.1720929146,-0.2163684219,0.0971574262,-0.2414959967,0.2002528459,0.103934288,-0.153853938,-0.3313707113,-0.2756563127,0.3822062016,-0.1417807639,-0.3627890348,0.0031981878,0.3467352688,0.1424879432,0.0543960258,-0.5818463564,0.204076916,0.2936166525,0.0626483932,0.0974664837,0.1681451797,0.0184589382,-0.0131430561,-0.1100514084,0.0396537818,0.1952479929,-0.2644456029,0.0578300692,-0.334122628]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Awesome, actually @lhoestq let's just ask the authors if we should host the dataset no? They could just use our links then as well for their website - what do you think? Is it fine to use our AWS dataset storage also as external links? ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":45,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nAwesome, actually @lhoestq let's just ask the authors if we should host the dataset no? They could just use our links then as well for their website - what do you think? Is it fine to use our AWS dataset storage also as external links? ","embeddings":[-0.1768893898,0.2350078523,-0.0860683545,0.2041037232,-0.2330973893,-0.0434642397,0.3742262423,-0.044316683,0.0654879585,0.1011237577,-0.4423070848,-0.2948478162,-0.053013131,0.19022277,0.1885551363,0.0627990216,0.1907830089,-0.223871097,-0.1670011878,-0.1965385377,-0.1766091138,-0.0732584894,0.0706343129,-0.0558961108,-0.109869808,-0.0346703865,-0.2770484686,-0.2518584132,-0.1346012652,-0.1093846038,-0.0887257904,0.0714744255,0.2122815698,0.3748491406,-0.0001111458,-0.0383174494,0.1957092285,-0.0565414764,-0.3509976268,-0.1693969071,-0.573291719,0.0660453215,-0.5719038844,-0.1017439514,-0.1267914474,-0.2210606933,-0.0247829501,-0.0262521189,-0.0217784103,0.4211543202,0.1724234968,0.0767065063,-0.0757349357,0.0712172985,0.5016147494,0.1903977543,-0.1208541468,0.0486076213,0.1575179398,0.2447096258,0.0380539745,0.4838590622,-0.0023881842,-0.0943237469,0.1407907754,-0.0652164891,-0.2729624808,-0.2105958909,-0.0584889725,0.1272958815,0.9238364697,-0.0804873928,-0.3618767858,-0.0867211595,0.1270469576,0.0256298669,-0.0225610528,0.3293753266,0.0880346298,0.4814621508,0.0730028749,-0.5656747818,-0.2542563379,0.4220225811,-0.007557611,0.4286477268,-0.1196794659,0.0645305142,-0.2250846177,-0.007290449,0.1006183922,-0.3901226521,0.0825692788,-0.1444491297,0.021903744,-0.2602435052,0.139044717,-0.0016058547,-0.0352103487,-0.1127782613,-0.0280629732,-0.2498501241,-0.1836351156,0.1759288907,0.1128900945,0.1946852654,-0.0557106249,0.1234207079,0.3331668973,-0.1184011772,0.1024449542,0.0535904728,0.0506216362,0.3487089872,-0.619453311,-0.1546131521,0.053544417,-0.173026979,-0.253205806,0.2715059221,0.0761090219,0.0811212212,0.0938500315,0.2281717807,0.1154907718,0.0970761776,-0.0982920006,0.4322949946,0.2728835642,-0.2129614502,0.0688251257,0.0687561929,-0.1255346984,0.2854409218,-0.0134511394,0.4373062253,0.0432112888,0.0723351166,0.4536833167,-0.0496404171,0.0844456404,-0.0790762901,0.1410625428,0.1844744682,0.1113333702,0.2003936768,-0.2696185112,-0.0829067901,-0.0982417464,0.139248088,0.0410661809,-0.4283362925,-0.2054236978,0.1530169249,-0.2411671877,-0.1772434562,-0.3937396705,0.4677197635,-0.1089806408,-0.1580183506,0.0668157861,0.3648848534,-0.5243881345,0.0320259295,-0.2461238205,0.3035157919,-0.2656866908,0.162467584,0.1506852508,0.3610349596,-0.0026280412,0.363070786,-0.3029224873,0.3589448333,-0.046887029,0.2667205334,0.196294561,-0.2316158563,-0.3109975457,0.3833294809,-0.1581404358,-0.0803027302,0.1202453822,-0.0930226222,0.3154138327,-0.2549368143,0.0380148143,0.3598023653,0.3089077175,0.1584502757,-0.2977127731,-0.3035258055,0.1937769055,0.2918933034,-0.0053833895,0.0401416011,0.2056317329,0.1810541749,0.0813514516,-0.4483517706,0.3314917088,0.1831410676,0.064724952,-0.0849556625,-0.0895155966,0.0210448671,-0.23346062,-0.0143252993,0.1769770086,0.5444411635,0.0163460188,-0.3388365209,-0.1066398025,-0.0903188512,-0.1211052835,-0.2375422865,0.1293758154,0.1012214646,-0.022610737,0.2113003582,-0.1514019072,0.1217543632,-0.3733397126,-0.1286690533,-0.0627999902,0.3903954327,-0.0269161724,0.160422802,0.0997128189,0.2631116211,0.0181514863,-0.1116654575,0.1229363605,0.1779538095,-0.1991259307,0.4674479365,0.2875537276,0.2718807757,0.453772366,-0.4568029046,0.4307044148,0.1117926314,0.0557060316,-0.2708335519,-0.0953097269,0.1411738545,0.0712656751,0.279833734,-0.0812823474,-0.0139914155,0.1227945909,0.1529698819,-0.3194538653,-0.067554526,0.119297497,0.3907163739,-0.0500784591,0.0840572044,-0.2486940622,0.2649683356,0.2224398404,-0.1401569545,0.1298113167,0.205111891,-0.1618110389,-0.3060455024,0.3169536889,0.0108898198,0.3427012563,0.2259627432,0.3396810293,-0.0380176194,0.2025637329,-0.1947642416,0.3998696506,0.1212736443,-0.2317920923,0.0261172559,0.0598338284,-0.2210960686,-0.1962629408,-0.1436424851,0.0685110912,-0.0220153462,-0.2928755581,-0.1260191947,0.353106916,-0.22138457,-0.1454388648,-0.0889578536,-0.2236862481,-0.0737028271,0.2939588428,0.0261520352,-0.3882161677,0.0527427606,0.1434535086,0.426738739,-0.4990487695,-0.224945873,-0.1499339789,-0.3044839799,-0.0568174906,0.1155419946,0.4216364026,0.0927081779,0.1239765435,0.0199194252,0.1607158482,-0.5759221911,-0.0878570452,0.1158383787,-0.2258940488,0.1047840118,-0.1758355796,0.00570346,-0.1109208688,-0.3367815614,-0.1069707423,0.2574764192,0.3168449402,-0.2773308754,-0.1600438654,-0.1456625164,0.0277714208,-0.327439934,-0.4971727729,-0.2563106418,0.2121649981,-0.1605940312,-0.036346145,-0.1856092215,0.1209088713,0.02765828,0.2640106678,-0.1219032183,-0.357771188,-0.4457055926,0.2771977484,-0.2792561352,-0.318675667,-0.2165585756,0.2202288508,0.3044364452,0.0072865817,-0.505076468,0.1221696436,0.1136534363,0.2336242348,0.1658106744,0.2572152019,-0.1163074598,-0.0456307344,-0.0972232297,-0.0335045159,-0.0529938713,-0.0145321861,0.3511269689,0.1141988337,0.1847343743,-0.065275386,0.0465935245,0.8114858866,0.2052149028,0.0942456201,0.2072391808,0.1222642362,0.2584118843,-0.1096866801,-0.232677117,0.4507557154,-0.0110750692,-0.453101486,0.4092404246,0.1936361343,-0.4243686497,-0.1811833233,-0.1927424669,-0.2558113039,-0.2215340137,0.2201714069,-0.0004187744,0.2325729728,0.2550107837,-0.340923965,-0.150100708,-0.2251624614,0.1738248616,0.1080894321,0.1500467211,-0.0339228138,-0.1573180556,0.0072915847,-0.4244354069,0.1223909333,0.1036445051,-0.0535738207,-0.1505927593,0.0437564664,0.2378766388,0.0467573442,0.2373417318,-0.6042851806,-0.2336084247,-0.1005127281,0.5441204906,-0.0872382894,0.2938613296,-0.07094533,-0.3217417896,0.4324373305,0.0796397179,-0.3160794675,0.1735689491,-0.037849322,0.1786252707,0.0154050589,-0.2483059913,0.1622094065,-0.262696296,-0.3507334888,0.0245759469,0.4427526593,-0.2547818124,0.1167401671,0.1065266132,0.1845450997,0.1730900258,0.3746153414,0.0278273225,-0.2092146724,0.268452704,-0.1993256211,-0.0012183947,-0.0303351805,0.1117545739,0.5347274542,0.1244179681,-0.0920630693,-0.065951176,-0.3150817454,0.2478329241,0.47891289,0.1541990489,-0.0192751437,-0.1379915476,0.0182344653,-0.2363636047,0.1149400324,0.0447646715,0.1696814001,-0.1691154242,-0.3124704361,0.4716693759,-0.1161426157,-0.2748441398,0.1353507936,0.461056143,-0.3331757784,0.3812474608,-0.0777214468,1.2574037313,0.2091267705,0.3034158051,0.1667199731,-0.2147902548,0.3622977436,-0.8011132479,-0.3374129534,0.246702075,0.4070867896,-0.1861028671,0.0539330207,-0.116869539,0.258873105,-0.2602069974,0.1085949689,0.3884543777,0.2679695785,0.047313042,0.684899807,-0.1614899039,-0.3734571934,-0.3993861079,0.2234205157,-0.0549959056,0.1194544286,-0.1127426475,-0.0717189834,0.0267594587,-0.0208314434,-0.4345801473,0.0507643446,-0.1732655466,0.1041235849,-0.7388635278,-0.1987111419,0.4851576388,0.2617512047,0.3026669621,-0.1319672018,-0.3229519427,0.1003111899,-0.1360894144,0.2056321651,0.0788020641,-0.2793516815,0.249021247,-0.1946907341,-0.1699622273,0.3036076128,0.182676807,-0.1257749498,-0.1869677752,-0.2314064801,-0.0634152144,-0.1014573723,0.0645371974,0.7086158991,-0.1242265701,-0.1423468441,0.0898191482,-0.0355188809,-0.1823471338,0.0630241781,-0.1724221557,-0.2583627105,-0.2958759069,0.262740165,0.1436107606,-0.165779531,0.1606542468,0.2066292316,-0.2030226141,-0.233234331,0.2971604764,-0.1968921572,-0.4040932357,-0.0978973135,0.1039421409,0.0089094872,-0.6479458213,0.0122775314,0.3014065921,0.0565671809,0.0124742724,-0.0032776387,-0.1057602167,0.1153566837,0.1963127255,0.1957065463,-0.0041364748,0.0221762843,-0.0040189154,-0.2936589122,-0.325768888,0.1216531321,0.1322056949,0.01514275,0.230473116,-0.055925265,0.3730338812,-0.1303085685,0.0037819934,-0.2780872583,-0.132163167,-0.1266551018,-0.090246819,0.1339966655,0.3759601116,-0.1065244079,-0.3196533322,-0.3233650029,-0.1000311002,0.1722379029,0.2496543229,0.2512970865,-0.2086667418,0.0241669789,-0.0898846611,-0.0730440617,-0.1285452992,0.0532142855,0.2011703849,0.4249559045,-0.0661376566,0.2151745111,0.2274166644,-0.1533635706,-0.0283840541,0.3125064075,0.0812128633,0.3080126047,0.2752333879,0.0044603664,0.1644084454,0.1426968277,0.1786899269,0.2465071529,-0.1580422223,-0.248329103,0.2400391698,0.1879075021,-0.014929547,0.1248806044,-0.2457345277,-0.1061554253,0.1178541407,0.1235412657,0.4039233327,0.1374643445,-0.0056473012,-0.0377063118,-0.0336981826,-0.0729511529,-0.0440827198,0.0375486389,-0.0481002443,0.2122240365,-0.056210041,0.1301835775,0.0564599521,0.1402943581,-0.2529905438,0.4280245006,0.395853579,0.0609932691,0.2322162092,0.1827021688,0.0395139232,-0.1088227928,0.0987528116,0.1106765121,0.1932629198,0.1029565111,0.0499597751,-0.0883621797,-0.1628955305,0.2212507129,0.1559380293,-0.0546393543,-0.3222887516,-0.2264621407,-0.2032473981,0.0393442102,-0.2180349082,-0.3061272204,-0.0036360766,0.1827471703,0.0185775831,-0.1639657319,0.2126777917,-0.138937965,0.1095732599,-0.1786133349,0.2214951366,0.1773696542,0.0518766344,0.1455042064,0.5054803491,0.4030255973,-0.0954717398,-0.1177581698,0.2130497843,-0.0850209743,-0.0415262878,-0.050150875,0.3698762655,-0.0947636217,-0.0204168465,0.4580847919,0.1208322495,-0.1228445992,0.0358988531,-0.0657833591,-0.0155150993,-0.1848288625,0.2088818401,-0.0647201985,-0.088861756,-0.197219491,-0.2323090583,-0.4567830563,0.0873133913,0.6285592914,-0.4415358603,0.1877529472,-0.1688579619,0.0637906492,-0.141700536,0.4888125062,0.2291922122,0.1223844364,-0.3486630023,-0.0772763118,-0.3244709671,-0.0844903886,0.0308248997,0.1885183901,-0.5234082937,0.1661741585,0.395747155,0.4661047757,-0.3010932207,0.0578131527,-0.7391779423,0.1325211227,-0.2632068694,0.2868499756,-0.2400152534,0.7252261043,-0.0947534814,0.0952450559,-0.0840793476,0.2825780213,-0.0699172094,-0.2538467944,0.0970698744,-0.1875065863,-0.2575334907,0.3402135968,-0.0026477629,0.2787730694,-0.1899044216,0.0229580682,-0.2257594168,-0.2839412987,-0.0533421002,0.1128788218,-0.1534742713,0.2266210169,-0.21433644,0.0166559145,-0.0317105912,-0.2789266407,-0.0395812839,0.2307523489,-0.1291417331,-0.0628192052,-0.2441754937,0.1129018143,-0.091831021,0.3301850259,-0.0257874541,-0.0822930038,-0.2801320255,-0.2600485981,0.3812657893,-0.015017991,-0.2919608951,0.3405650556,0.3671382368,0.2214136273,-0.0475076698,-0.7940045595,0.1044555902,0.3324848115,0.1010682732,-0.0450565256,0.0702169836,0.1605493724,-0.0245241448,-0.2514024675,0.0185313206,-0.1145182177,-0.2685368359,-0.1067239791,-0.3115546703]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Yes definitely. Shall we suggest them to create a dataset repository under their org on huggingface.co ? @julien-c \r\nThe dataset is around 1TB","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":23,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nYes definitely. Shall we suggest them to create a dataset repository under their org on huggingface.co ? @julien-c \r\nThe dataset is around 1TB","embeddings":[-0.2819742262,0.1316514462,-0.114289701,0.1891249716,-0.0761730298,0.0238953866,-0.015186389,0.1323354542,-0.292786479,0.1568823755,-0.2882780433,-0.227353245,-0.1356900185,0.177145943,0.1977107376,-0.1647572368,0.1265198737,-0.0231165942,-0.3271674216,-0.2998180985,-0.0985968709,0.076150097,-0.0886971802,-0.1920019239,-0.2697936594,-0.0368123688,-0.311208725,-0.2177725881,-0.3672375083,-0.2110482156,0.0392684266,0.0594678856,0.1656423807,0.5584837794,-0.0001052459,-0.1206114292,0.1992835402,-0.2385989279,-0.2963815629,-0.1344755143,-0.2026971877,-0.2349897027,-0.6540787816,0.0331206843,-0.2515168786,-0.1497751623,0.0242077503,-0.0051272535,0.1849745363,0.5032078028,0.2287637293,0.1169008538,0.2910440862,-0.1472678483,0.2438445836,0.0753873661,-0.1609886885,0.155164212,0.2891928554,0.2260188609,0.1128324866,0.508826077,0.049854219,-0.2827511132,-0.0035755008,-0.3058139384,-0.1964981854,-0.4243428707,0.1494263262,0.2748365104,0.547150135,-0.2278184444,-0.2290821224,-0.1697379351,0.2611936331,-0.1568030864,-0.0499189682,0.1122894138,-0.168975696,0.3771129847,0.101904802,-0.3955154419,-0.211272195,0.2854039371,0.0232806113,0.2817807198,-0.157686606,-0.0990013704,0.1004225016,-0.2646625936,-0.3853080273,-0.0833880454,-0.007706542,0.0923128948,-0.1722644866,-0.3142679334,0.2135915905,-0.0067905239,0.2582921088,-0.3227830529,-0.2262986153,-0.151602596,-0.193173781,-0.0800545737,0.0986224115,0.3844141662,-0.0672855973,-0.0160431433,0.1091509312,-0.2292179614,0.0716155097,0.1443646401,-0.0746077672,0.1786736399,-0.4144827127,0.0734653249,0.0291879494,-0.2194577008,-0.1896154135,0.120626308,-0.0307190884,0.0599309839,-0.0112485364,0.3262389004,0.0525543429,0.2642126083,-0.1074369103,0.3535129428,-0.0494072251,-0.2231225371,-0.090708904,0.1249392033,-0.1816266775,0.1585262865,0.2224279791,0.2521287501,0.0808241218,-0.0307100322,0.2543049455,0.1120356917,-0.2170091569,-0.0438588075,0.2266108245,0.0343673863,-0.0065230625,0.0574704446,-0.2188868374,-0.1199041978,-0.2017616034,0.0561885647,0.0737984404,-0.1303486526,-0.2450791001,0.2372362018,-0.1433439106,-0.2346129566,-0.3658261001,0.6624663472,-0.0989387482,-0.0618829094,0.0915619135,0.2291860431,-0.4113318026,0.0413746201,-0.0740232915,0.3984498382,-0.1373505741,-0.1309414059,0.1131423563,0.0383187346,0.0138813425,0.2907513976,-0.0383745469,0.1933073103,-0.0324502029,0.3480066657,0.0341861285,-0.404607296,-0.4445278049,0.1492796689,-0.2617088556,-0.0849773213,0.230523169,-0.0753984526,0.5631130934,-0.2706470191,-0.0062919785,0.2914619446,-0.0247240309,0.0928732306,-0.2349769324,-0.4155744612,0.2283686846,0.169958204,0.0837148726,-0.0453879014,-0.0390071794,0.3225706816,0.1939514577,-0.264046967,0.2739546001,0.263848424,0.1122649089,-0.1178820357,-0.0105329249,-0.226553008,-0.1622185558,-0.0458204113,0.1651752293,0.5588303208,0.0393424667,-0.248207882,-0.1093644872,-0.1796623915,-0.1966807395,-0.1557609141,0.237955749,0.2530170381,0.1944989264,0.1405171752,-0.1418555081,0.2613063455,-0.3185066581,-0.1850554645,-0.1366381794,0.2391347736,-0.0062543959,-0.2149600089,0.2408408672,0.4956155121,0.0868682861,-0.0789923891,0.0718966648,0.055457864,-0.3503284752,0.173853904,0.1244394928,0.2339427918,0.3274999261,-0.5012203455,0.2246926427,0.2709726095,0.1081655324,-0.1598505229,-0.1166042313,0.1506719291,0.2914888561,0.0830967277,0.0569923185,-0.0883582607,0.3350683451,0.0963170305,-0.1294327676,-0.082415767,0.1184362248,0.1973696053,0.0082634343,0.1231061146,-0.3071097434,0.1754607409,0.5225203037,-0.18223387,0.3080971241,0.3316549659,-0.0987378433,-0.135523051,0.0856194496,-0.1874738187,0.2695055306,0.3026119769,0.2179222256,-0.0037574889,0.3238419592,-0.0748646334,0.1274349242,0.0160299223,-0.2442697436,0.1969398707,0.1947351992,-0.1007377133,-0.3468711078,0.0454010665,-0.0385369882,0.2338499874,-0.0725072771,-0.3474930525,0.1309443563,-0.2900649011,-0.263168782,0.0869787112,-0.2649092078,0.0776728168,0.3261661232,-0.0220975652,-0.079115063,0.2016181946,0.244778499,0.3428672254,-0.4011153877,-0.0804300681,-0.0175808799,-0.3277688622,-0.1439275891,0.1810654402,0.2856897116,0.1181080714,0.2969580293,-0.1818919182,0.147623986,-0.2822354138,-0.3307644427,0.2242473364,-0.2791380882,0.0409083478,-0.1901490241,0.0344603434,-0.0575047024,-0.3769896924,-0.1464236826,0.0630080551,-0.1137537658,-0.1005490869,-0.4217520356,-0.1302698255,0.0265244674,-0.4483323991,-0.5188705921,-0.4268541932,0.4857845902,-0.2541282773,-0.0696323812,0.0043212459,0.022377247,0.1944224089,-0.0926828682,0.1866732687,-0.2760030031,-0.2769815922,0.1621173024,-0.2088313252,-0.3052783608,-0.2177145183,0.0425678976,0.0888220519,-0.1137360334,-0.4125672281,0.1012409553,-0.1500974,0.3791904747,0.1974512488,0.1598887146,0.2622482181,-0.1320288628,-0.1125875562,-0.1074818224,-0.2057563215,0.0209976658,0.3534580767,0.2022938281,-0.0050759776,0.1006621644,0.0742811114,0.6177754998,0.3007921278,-0.014930604,0.2079491913,0.1091752201,0.4607527256,-0.2662662864,-0.4412497282,0.5007209182,0.1288642138,0.0066467421,0.4606542587,0.1416971534,-0.3381603956,-0.1234240681,-0.4403937161,-0.1857142001,-0.2916959524,0.3045727611,0.1676311642,0.2078150809,-0.0655809641,-0.2340807617,-0.2560780346,-0.1412127018,0.2151404768,0.1932208389,-0.0028482061,0.0096075209,-0.1996282637,-0.0454896428,-0.5053685904,-0.0774148554,0.1700080931,0.2866506577,-0.4697832763,0.1019584313,0.0151189985,0.1103733405,0.3322482109,-0.2976499796,-0.0814968422,0.1211557612,0.3847764134,0.0415494442,0.3428761959,-0.1924239248,-0.1800599694,0.5656817555,0.3200426102,-0.4232358336,-0.0473005995,0.2894099951,0.3699550033,-0.0223509092,-0.1071488857,0.115070805,-0.3763307631,-0.367184341,0.2561834753,0.4782500267,0.0149077782,0.298766017,0.1430377066,0.1744925231,-0.0785180554,0.4322351813,-0.0849007741,-0.0709436163,0.4809972942,0.1570385098,0.0498442054,-0.0119312573,-0.0471392572,0.3187128007,-0.0961809829,-0.3610298336,0.0101587055,-0.2590028048,0.3601127565,0.3817853034,0.1650250703,0.0477344319,-0.0667490959,0.0995852053,-0.2544426322,0.1495524496,0.2385512441,0.2091663778,-0.3974754512,-0.231734544,0.3972523212,0.0181857217,-0.0362248607,0.0221506413,0.3118011951,-0.2192683071,0.2298456132,-0.1993248314,1.2992119789,0.1632505059,0.0903647393,0.219796747,-0.2111111879,0.1376828253,-0.3781129122,-0.176121816,0.2771657705,0.2005887926,0.0001146443,0.2628122866,-0.0071454891,0.1968727708,-0.4309047461,0.2430021167,0.0824818239,0.1120306998,0.0918789357,0.5037056804,-0.2123636305,-0.436863482,-0.4565170705,0.265935868,0.1328690052,0.2141060978,0.1090772077,-0.2810109556,0.1247555837,-0.0385291427,-0.6611127257,0.0185883325,-0.2485932708,-0.0987371728,-0.3499998152,-0.1787732095,0.3545901775,0.2069097757,0.3911926746,0.0065712435,-0.255477488,0.1793650538,-0.0539304949,-0.0166115034,0.2345947772,-0.1129865646,0.3398457468,-0.3037159145,-0.3038263917,0.1185796186,0.178913787,-0.0587759092,-0.4278465509,-0.197593376,0.0824829489,0.0208462588,0.1324299723,0.5716294646,-0.0725074634,-0.1865527481,0.1920007765,0.0037562731,-0.3681871295,0.1715861559,-0.1264358461,-0.2424495965,-0.2466192693,0.2281707972,0.2111569941,-0.0708634257,0.3206344545,0.2639295459,-0.2118222415,-0.3511908352,-0.0277414788,0.1581448913,-0.6010571718,-0.0237756092,0.115309462,-0.1478649527,-0.2918296158,0.2249107659,0.1622529477,-0.1301775426,-0.0217072684,0.0005141714,-0.3193922937,0.0408909433,0.1909769326,0.1280002743,-0.0940309688,0.2617275715,0.0842409432,-0.1070800275,-0.426466167,0.1015553847,-0.0928600952,0.1503619105,0.3520087004,-0.1229469031,0.1642499715,-0.0084653795,0.1471755952,-0.0499450862,-0.1253854036,-0.2292287052,-0.1027220488,0.0912542045,0.046807006,0.153768301,0.0253817588,-0.1104473695,-0.2151512653,0.1217315197,0.324562639,0.2861365974,0.0006523764,0.0900776461,-0.0167495478,-0.0522099249,0.1707626283,0.0709109753,0.365476191,0.3459137976,0.1199684516,0.16755642,0.0461910404,-0.1234662384,-0.0178634487,0.3209616542,0.2223625779,0.0574946292,0.0683491677,0.0039023943,0.1513945013,0.4949732125,0.0559026077,-0.0193854161,-0.1128311455,-0.0122186411,0.4042258859,0.2655508518,-0.2175688446,0.1577069312,-0.1165616512,-0.2681825161,0.1453456283,0.1360539943,0.025958484,0.0493242629,0.2134719491,-0.0500252694,-0.1125881597,0.1113784611,0.0332529061,0.1568785757,0.1774015278,-0.040456634,0.0792257711,0.1081677675,0.1449200362,0.0022453994,-0.2128182352,0.4793628156,0.0601195879,0.1790766865,0.2776876986,-0.05780169,0.0460843518,0.2010461986,0.1733366847,0.159144327,-0.0010345228,0.4170062244,0.0103538921,-0.3126984239,-0.1372697055,0.2817866802,-0.0407529287,-0.0422516242,0.0398400947,-0.2205162048,-0.2530546486,0.0342658535,-0.0721731782,-0.3266580105,-0.0358871035,0.1790011525,-0.1568446606,-0.2047487646,-0.1355964541,-0.174443692,0.1936496496,-0.1843628436,0.4101902544,0.1932412088,-0.0974404886,0.1663823277,0.5504941344,0.5049684048,-0.0284648575,-0.113241002,0.2732325196,-0.0776151791,-0.0731150359,-0.0507643782,0.404709816,0.0104424218,0.1142541617,0.2026081532,0.2117382437,-0.1936115772,-0.1927449554,0.1268787086,-0.1338594258,-0.2348514795,0.1727145612,-0.1869641393,0.090551354,-0.5943123102,-0.1308286488,-0.5543190837,0.1238442063,0.4093554616,-0.3034743071,0.106618464,-0.1998762339,0.1063452214,-0.1901714802,0.3664016128,0.3114938438,0.0490184948,-0.1517065614,-0.1676571667,-0.4584262371,-0.1184764653,0.066152446,0.1342472285,-0.369192183,0.107272014,0.3487349451,0.3570177853,-0.1373610049,-0.1399362981,-0.4183528125,0.0804428756,-0.3068990111,0.1913910359,0.0125664799,0.5848876238,0.170880124,-0.1409070939,-0.1698655188,0.2786668837,0.0636013746,-0.1539649367,0.0959112048,-0.3254545629,-0.3092284799,0.1015457958,0.0018900647,0.3967323005,-0.2579964399,-0.1228033379,-0.1681188196,-0.1251890659,-0.1553907692,0.0849936977,-0.0137850828,0.2810283601,-0.2272489965,0.219228968,-0.0145044792,-0.1148026884,0.0029678775,0.037186753,-0.2482935786,0.1114626601,-0.1198016852,0.0042513395,-0.1065740436,0.308783561,-0.0339314155,-0.0553761311,-0.247954756,-0.2358028591,0.4588891566,-0.11825075,-0.2737709284,-0.0208788123,0.3357271552,0.1670122147,-0.1263698488,-0.7813284397,0.2783124447,0.2228353471,0.2146983445,-0.0092783347,0.2461497635,-0.0099820085,-0.1109784767,-0.1276862621,0.007531554,-0.0001103626,-0.215310961,0.2524287403,-0.2783285379]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Sounds good! \r\n\r\nOrder of magnitude is storage costs ~$20 per TB per month (not including bandwidth). \r\n\r\nHappy to provide this to the community as I feel this is an important dataset. Let us know what the authors want to do!\r\n\r\n","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":40,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nSounds good! \r\n\r\nOrder of magnitude is storage costs ~$20 per TB per month (not including bandwidth). \r\n\r\nHappy to provide this to the community as I feel this is an important dataset. Let us know what the authors want to do!\r\n\r\n","embeddings":[-0.2674563229,0.3468384743,-0.1150629446,0.0743555352,-0.1094922796,0.0280111264,0.0628126711,0.2007724792,-0.3066087365,0.2189829797,-0.2691016197,-0.206533283,-0.1899873912,0.1393337846,0.1780733615,-0.1505705714,0.1918693185,0.0074416245,-0.2874937057,-0.2704440951,-0.0860427767,-0.0386056602,-0.1379612833,-0.2685125172,-0.229930982,-0.0946117565,-0.3455687165,-0.2869116366,-0.3232319057,-0.1962370276,-0.0556502976,0.1350371689,0.1240529045,0.4940181375,-0.0001045567,-0.2592286766,0.1250719279,-0.2645716369,-0.3532478213,-0.0380953699,-0.1996164322,-0.1757065058,-0.6599071622,0.0304601323,-0.0956022218,-0.2508485913,0.0672225356,-0.0263301507,0.1266006827,0.5544310212,0.2140147686,-0.0130658662,0.1716801971,-0.1371425092,0.2871406376,0.0758833066,-0.2084584683,0.0772271976,0.3049242795,0.1136055216,0.0316927955,0.4821645319,-0.0075356467,-0.2882493138,-0.0318442546,-0.3790358901,-0.1772185713,-0.4207233191,0.1050826088,0.3116881847,0.6529287696,-0.2026355416,-0.1981211603,-0.0961150676,0.266823262,-0.159811765,-0.049897138,0.1737100035,-0.1241329759,0.3310909867,0.1392672956,-0.2483858317,-0.2472933233,0.3075198233,0.054465156,0.30402565,-0.086230889,-0.0989881605,0.0883807614,-0.274225086,-0.2766878605,-0.0626630485,0.0072822026,0.1095095798,-0.1418611854,-0.3701093495,0.2169772238,-0.1320178509,0.0793473497,-0.3674499094,-0.1592274159,-0.1363395154,-0.0830795243,0.0713027418,-0.0183546804,0.3584363759,-0.080280818,-0.07807675,0.1004720926,-0.2737582326,0.0465741605,0.1279942989,-0.0759718493,0.3356284499,-0.3972549438,0.0821795613,-0.0861617476,-0.1665338278,-0.2463089526,0.1953467131,-0.0466941856,0.0915033594,-0.089351885,0.3172927499,0.0441723429,0.3904410899,-0.0830957815,0.3368884027,0.0517918617,-0.192899242,-0.0603897274,0.2241698503,-0.2017291188,0.2546380758,0.2324511409,0.4521965683,0.049443651,0.0183007848,0.2040760517,0.0686422884,-0.0888890028,-0.0491717458,0.2716386318,0.0259583276,-0.0311120562,-0.0047629378,-0.2496473342,-0.1610568315,-0.203546837,-0.0107688364,0.1420149356,-0.1101741716,-0.2979933619,0.2533713281,-0.0716987997,-0.2677494287,-0.2793786526,0.7518096566,-0.1782979518,-0.1115564406,0.0697045997,0.1738110781,-0.46041888,0.0459398553,-0.1368118078,0.3723906577,-0.2363218665,-0.0399011075,0.0980855674,0.20547387,0.0443731844,0.2060778439,-0.0810586736,0.2622190416,0.0549096651,0.3178981543,0.0442697592,-0.4065377712,-0.4463209212,0.1585043073,-0.1760579348,-0.1846080869,0.2989750803,-0.0001517988,0.6347365975,-0.3393875659,-0.0137222102,0.3644084036,-0.0631121844,0.0718456954,-0.2582840025,-0.4786252379,0.1988205761,0.1503828168,0.0412006006,-0.0575377829,0.06441921,0.369481653,0.2183164656,-0.327760756,0.240490675,0.2785543501,0.0622734018,-0.1536896229,-0.0451561287,-0.1693333983,-0.158728525,-0.083061032,0.1781626195,0.4612787664,0.1114623472,-0.2092922181,-0.0072706006,-0.2170674354,-0.1183109656,-0.1257411689,0.2582389414,0.2570565343,0.1128951386,0.118896462,-0.1245715842,0.1912956089,-0.2764888108,-0.247979179,0.0048344885,0.234055236,0.0556535311,-0.2928238213,0.2693490982,0.4156476557,0.1235693842,0.0075204344,-0.035468515,0.1215612739,-0.2071395963,0.2657057047,0.1701867282,0.3067458868,0.2318335623,-0.4714109302,0.3065223694,0.3407137692,0.1727475524,-0.2098979354,-0.218340531,0.1196065396,0.3287129104,0.0522700585,0.0505121872,-0.091818139,0.2911438644,0.0929449946,-0.1188122034,-0.1293872148,0.1732148677,0.2709854841,-0.0985399559,0.1335847527,-0.2680142522,0.2048004866,0.5615863204,-0.1304108948,0.3069731891,0.1990514398,-0.0527817309,-0.2780837715,0.1204248443,-0.1962444186,0.2916373312,0.2810159922,0.2256419808,0.016477989,0.2985712886,-0.1521057636,0.1900838912,-0.0507359542,-0.2516885698,0.1603920162,0.203517586,-0.1073486209,-0.3217076957,0.0468225628,0.0054197176,0.1432108879,0.0397218801,-0.264711231,0.1132286265,-0.2363813072,-0.1619854718,0.0370658934,-0.159675926,-0.0032517808,0.3915063143,0.0485030971,-0.2069420218,0.1535012573,0.2165166736,0.3693400621,-0.394174397,-0.0828145817,0.0096045109,-0.3672121167,-0.1786138117,0.1865738779,0.1984503269,0.1628403366,0.2797743082,-0.1282354444,0.1851775497,-0.2126670927,-0.3121667504,0.1550998539,-0.2662827075,0.0524388626,-0.2734869123,0.018325977,-0.0566889942,-0.3820500076,-0.265337199,0.0315834396,0.0263812412,0.0323769487,-0.3927341104,-0.1335774809,0.0208352692,-0.3891116679,-0.5565213561,-0.4471624494,0.3865203559,-0.2517313063,-0.0092638088,-0.0017902148,0.0099090226,0.1746537685,-0.0799356401,0.2039350122,-0.2852186859,-0.2310419977,0.1599523574,-0.2169172615,-0.3349026442,-0.1540761292,0.039112553,0.0620787479,-0.1150843576,-0.4363949597,0.1783678383,-0.098703444,0.3334511817,0.1652608812,0.010133327,0.2976228893,-0.1061075404,-0.1565746218,-0.1825390607,-0.1687907726,-0.0481080674,0.4235719144,0.1204992831,-0.0134769389,-0.0380831659,0.1223323792,0.5573822856,0.2054240108,0.0274663251,0.2063635737,0.1498259306,0.4568346143,-0.2330971062,-0.3974888325,0.5278069973,0.136878714,-0.0743880719,0.5241379738,0.0732344761,-0.5568887591,-0.1071173176,-0.3339170218,-0.0818362609,-0.2224315256,0.3396263421,0.0992458314,0.1561368853,0.0206615459,-0.1405636519,-0.2554541528,-0.2717418671,0.1895900816,0.0959591568,-0.0478783026,0.000699715,-0.1602261662,-0.0620183311,-0.3980846405,-0.1072697267,0.1569577307,0.2464651018,-0.530356586,0.0870729834,0.004621054,0.1763382107,0.1798848361,-0.3207392395,-0.08123523,0.234244749,0.5497144461,-0.0188403744,0.1853069663,-0.2950846553,-0.3423018456,0.5709077716,0.3086677492,-0.1830256432,0.0339107402,0.1677019447,0.4645702541,-0.0490050726,-0.0457137674,0.1397290379,-0.3151711226,-0.3120527864,0.3089373112,0.4163502157,-0.0146757131,0.2681359053,0.1516004354,0.182018131,-0.0082620457,0.4942662418,-0.077151157,-0.1044253632,0.4753471017,0.0918057337,-0.0378000773,-0.1117261276,-0.0847388878,0.1528491229,-0.061277315,-0.2044252604,0.0375261866,-0.359580338,0.3538605273,0.4210065901,0.1246670187,0.0118692396,-0.1096945107,0.1038219631,-0.3049685061,0.1741950512,0.2645324171,0.1056654826,-0.3500519991,-0.2346191853,0.3728401959,0.0606273748,-0.1213382483,0.0969010517,0.1682978868,-0.2165053189,0.3552684784,-0.1398266554,1.3028799295,0.2360524833,0.152630344,0.191718787,-0.2324958742,0.0736016035,-0.3725529611,-0.1548819691,0.3520877063,0.1767763346,-0.0249604471,0.2205830216,-0.0468529537,0.2085399181,-0.5711346865,0.2264648676,-0.0092791012,0.1269484609,0.1225560009,0.4747870862,-0.2686780393,-0.4689059258,-0.2684135437,0.29126665,0.1350039691,0.0400314741,0.150964275,-0.2544361949,0.1512274444,0.0626962036,-0.7009258866,0.1211337894,-0.1179800481,-0.027463397,-0.5845332146,-0.0470114201,0.2115438581,0.1814661324,0.3393006921,0.0038952534,-0.2309700698,0.1532793641,-0.0033100243,0.0046116225,0.2276838422,-0.1418342292,0.4154863656,-0.3145345449,-0.2073202431,0.2223330587,0.1595188677,0.0323574618,-0.459344089,-0.1939759552,0.1733255684,0.0496667176,0.2545650601,0.5486976504,-0.1140187234,-0.2271647751,0.1972653121,-0.0021888965,-0.3667445779,0.279343456,-0.2162353247,-0.2433760315,-0.2037613094,0.2019093186,0.2641421258,-0.0414036363,0.2854686975,0.2260183543,-0.2272043228,-0.4182305038,0.062583223,-0.0068318723,-0.4837263525,0.0208749231,0.0344056152,-0.2015659064,-0.3309279382,0.2069539726,0.0655929074,-0.1283622235,-0.0847934633,-0.0409497917,-0.2633304596,0.0133850304,0.1992594898,0.1133462489,-0.1020509377,0.1707706004,0.0079120295,-0.0708649606,-0.430911839,0.1298056096,-0.044153221,0.1981446743,0.2045233548,-0.1416718066,0.1859093904,0.0071455413,0.1405664682,-0.0905823782,-0.1317285597,-0.2335890681,-0.1661424488,0.0932709426,0.1177336127,0.1945551336,0.0184493139,-0.1887735277,-0.2421473563,0.0682541057,0.2440643609,0.2484223396,-0.0795718804,0.1264837831,-0.0289050732,0.0478287712,0.2515592277,0.057637535,0.3243322372,0.4395715892,-0.056537237,0.1609668881,0.0865144953,-0.2113108039,-0.0256000459,0.3494980931,0.057361424,0.0000105554,0.1125313863,0.0271209627,0.224764809,0.3390503824,-0.0655513406,0.0401953347,-0.0590094104,-0.071136646,0.3415955305,0.279717207,-0.1062905565,0.0915033296,-0.1858474463,-0.1766531616,0.1557305902,0.1183879673,0.1122665778,-0.1109628454,0.2080169767,-0.0815588012,-0.1713954657,0.0737800747,0.0376246721,0.1666731536,0.178557381,-0.1019874364,-0.0332603529,0.1056908667,0.137499094,0.001478944,-0.1647669226,0.4625000954,0.1857182533,0.1034674197,0.1984197795,-0.0124749197,0.1337440908,0.1642797291,0.0860118866,0.2504077852,0.0465904735,0.2660349607,-0.0584483594,-0.307838887,-0.1115982607,0.2143587321,0.0498426482,0.0185079072,-0.0203730147,-0.1985639483,-0.2563594282,0.0705057532,-0.0141697749,-0.2123469114,-0.0601292327,0.2392932177,-0.0922075734,-0.1058363542,-0.0526459813,-0.1634941101,0.2898860574,-0.2593402565,0.3642909825,0.1571874917,-0.0409320891,0.2152029276,0.4409629405,0.4797421694,-0.081410639,-0.2035112381,0.317204684,-0.0881628394,-0.1225336865,-0.013026908,0.4353638887,0.0154376375,0.1728734374,0.1735206395,0.2133731097,-0.1836387217,-0.1990482211,0.1639766395,-0.0699501634,-0.1323809475,0.1319433153,-0.0558499582,0.1001103371,-0.5145985484,-0.1135330275,-0.5816864967,0.1198292524,0.4090068042,-0.319031775,0.1662038416,-0.1266600192,0.1159002334,-0.1748052835,0.3187251985,0.2675618827,0.0338461101,-0.0751664937,-0.1109833494,-0.2891764641,-0.1746115535,0.131836772,0.2151110172,-0.3938134015,0.1322588325,0.3189955354,0.3898982108,-0.1670452207,-0.0444372781,-0.5264083147,0.1136265695,-0.3879907727,0.2637178004,0.1613398939,0.6632789373,0.0919784456,-0.1655247062,-0.1893275678,0.2305866033,0.0948499143,-0.1951650381,-0.0920037478,-0.2894489467,-0.2229043096,0.1501597017,-0.0516282059,0.3096761107,-0.2479177117,-0.1190460473,-0.182057187,-0.0348702073,-0.1049269512,-0.0867210701,0.0125222653,0.268578589,-0.249343887,0.3319284916,0.0971076041,-0.0566968061,-0.0484256074,0.0773956552,-0.2554419041,0.0929590836,-0.167076394,0.1511126906,-0.186954096,0.2303742915,-0.0432429053,-0.086873062,-0.254858911,-0.1970850676,0.4478434324,-0.107051447,-0.3662914038,-0.0358731933,0.3192018569,0.1614575982,0.0437976532,-0.6474663019,0.2208114862,0.2300790399,0.1481014043,0.0356305242,0.2190492302,0.0091545563,-0.085805133,-0.1326689571,-0.0387298502,0.0105596781,-0.1824440211,0.1066593081,-0.3101616204]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Great! @skyprince999, do you think you could ping the authors here or link to this thread? I think it could be a cool idea to host the dataset on our side then","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":32,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nGreat! @skyprince999, do you think you could ping the authors here or link to this thread? I think it could be a cool idea to host the dataset on our side then","embeddings":[-0.2638154328,0.255651027,-0.1020891964,0.067426078,-0.1163012162,0.0049871961,0.2088760734,0.1324665993,-0.2380465269,0.1958873272,-0.2925175428,-0.1947898269,-0.1235653386,0.1762092859,0.2179549187,-0.0965976119,0.1338677108,-0.0480607823,-0.2836484611,-0.3060131967,-0.0538708344,-0.0600476265,-0.1137716994,-0.1840386093,-0.2231121659,-0.031440556,-0.3102774918,-0.2115971297,-0.2723539472,-0.2186130881,0.0249478724,0.143304795,0.1645143926,0.4807238579,-0.000103108,-0.1160424873,0.1976706982,-0.1972872615,-0.2979801595,-0.113288179,-0.2479342371,-0.1520431191,-0.6573248506,0.0125980163,-0.3099523187,-0.2509469688,0.0265708677,0.0222529732,0.1163543761,0.5167062879,0.2408426702,0.055473119,0.1673075557,-0.134918943,0.2340431362,0.1127671525,-0.2015993148,0.1049611196,0.3035980761,0.1237042695,0.0485425889,0.4394952953,-0.0261507332,-0.2713295817,-0.0585949644,-0.3038038909,-0.2033460438,-0.3597015738,0.0331385471,0.347229898,0.666474998,-0.1610847712,-0.1860350221,-0.1502617002,0.2188349813,-0.1186533347,-0.0508356765,0.0833063647,-0.1209420189,0.3679973483,0.161706239,-0.3044969141,-0.2177587897,0.2728983462,0.0427058563,0.4257723689,-0.0621481575,-0.0788865462,-0.0586894676,-0.2083937079,-0.213214919,-0.0897469819,0.0146057568,0.0752056241,-0.1660140008,-0.3493968248,0.3036048412,-0.1503922194,0.1601200104,-0.3179901242,-0.1644588858,-0.1152175739,-0.1738077402,0.028868787,0.0701214522,0.3264145851,-0.023237342,-0.0438519493,0.1637263447,-0.1892008632,0.0626120418,0.1544866562,-0.0323628299,0.2342712879,-0.4513018131,0.2286697626,0.0270993803,-0.1980846822,-0.238577798,0.1215214133,-0.0300242435,0.0816451088,0.0075868592,0.2975362539,0.0489691496,0.3509753346,-0.1010926291,0.32719329,-0.0031223649,-0.2137394696,-0.0711701661,0.1336993277,-0.2383320481,0.17205818,0.2494342178,0.4215917289,0.0535096489,0.0821038634,0.3226335347,0.0747349858,-0.1555283219,0.0260720532,0.2722114623,-0.0337342247,0.0475034229,0.0337505341,-0.2039002925,-0.1663172394,-0.2246453762,0.0592587739,0.1313941479,-0.1161021218,-0.2872762382,0.2791669369,-0.1298475713,-0.2352979928,-0.2771536708,0.7251247168,-0.1624794304,-0.0774156302,0.1132643595,0.2274977416,-0.4088668823,0.0521916486,-0.0945313126,0.3656329215,-0.2517017126,-0.0169540811,0.1709535122,0.1352590024,0.0225041509,0.1482346654,-0.0828428566,0.3158314228,-0.0581727885,0.349792093,0.0392581969,-0.3666366339,-0.3150377274,0.1317054778,-0.2357329726,-0.0350298472,0.2565309107,-0.1328445971,0.6240297556,-0.3125279844,-0.0122896107,0.4339906275,0.0065104268,0.1316176504,-0.1982861757,-0.3214910328,0.23603715,0.1815402508,-0.005643649,-0.0182655342,0.0441294461,0.2750940025,0.2775231898,-0.3302541077,0.2896015048,0.2214820534,0.0354990028,-0.1381884366,0.0164784417,-0.1379673481,-0.1545863748,-0.0750247315,0.1970406026,0.5770363212,0.1224235892,-0.2467807382,-0.1696626693,-0.1954882592,-0.1923277974,-0.1729772389,0.2730987966,0.2822175324,0.0562135093,0.1398088634,-0.1364933401,0.2650596499,-0.3373715878,-0.21615161,-0.0441902131,0.2488891929,0.0546995029,-0.2353997529,0.1652757376,0.4716205001,0.1654156893,-0.031769935,0.0046441965,0.0743105859,-0.273206532,0.2347431779,0.1541242898,0.2190746218,0.3090022802,-0.5264615417,0.2203454375,0.2899042964,0.0991034657,-0.2010914087,-0.0736402646,0.1721486598,0.2992254794,0.1554922909,0.0093087321,-0.011654702,0.3077354133,0.0305271167,-0.1802012324,-0.0600098111,0.132446155,0.2860775292,-0.1437016129,0.0011757519,-0.2945253849,0.2530354261,0.4543097913,-0.1788542271,0.2692893147,0.1650078148,-0.0455254577,-0.1397265494,0.1442493945,-0.2406744957,0.2386839539,0.3034240007,0.3386998773,0.023740489,0.2088297755,-0.0828210264,0.1922583878,-0.0472447239,-0.3074480891,0.1286107302,0.1818548292,-0.1868811846,-0.3317626119,-0.0036381716,-0.0241872352,0.1092714295,-0.1317532957,-0.2908289135,0.1634647846,-0.2110120803,-0.2189671844,-0.0618507229,-0.2294008285,0.0262314454,0.4298076928,0.0383161567,-0.2114010006,0.2007107139,0.2322894037,0.3449686766,-0.4511022568,-0.0329894498,-0.0573889762,-0.3996440768,-0.1822762191,0.2001166493,0.214673683,0.1815421879,0.2603405416,-0.1186480001,0.1492659301,-0.2710829377,-0.3136157691,0.1868878007,-0.2719041109,0.120898962,-0.1901511848,0.0092401784,-0.0203032251,-0.4290495813,-0.2016885132,0.088761963,-0.030420633,-0.08557491,-0.4141586423,-0.0888712928,-0.0415390395,-0.4858826995,-0.4796178043,-0.4333893359,0.2988663316,-0.271073699,-0.0247242376,-0.0173538141,0.0402075648,0.0702034682,-0.0206962842,0.1687847525,-0.3169855177,-0.2643295825,0.111153923,-0.3181254864,-0.2912344337,-0.2173937112,0.0507119857,0.1230551153,-0.1351741254,-0.3304983079,0.1405244023,-0.0506784394,0.359983623,0.1792303622,0.0934568122,0.2319622487,-0.0519551486,-0.1444593072,-0.1684286743,-0.1450237185,0.0257617887,0.377843231,0.1647441387,0.0651211143,-0.0047894511,0.05896255,0.6301655173,0.1760899723,0.047284212,0.2631708384,0.0810749903,0.3796306849,-0.2192723751,-0.4116697311,0.5353423357,0.1132875308,-0.0897989497,0.4373247623,0.0851418003,-0.4881726503,-0.1443247795,-0.3178869188,-0.2233303636,-0.255317837,0.2832995355,0.0877095461,0.1678839624,0.0171090104,-0.1764429808,-0.2634767294,-0.1980425715,0.2678225338,0.1296084076,0.0484409966,-0.0313366354,-0.1926924735,-0.0355720147,-0.4021430016,-0.0952867568,0.1271766871,0.2549352646,-0.4831461012,0.0363273993,0.0602484234,0.1474111974,0.3275741637,-0.3593376577,-0.0768371373,0.1758432984,0.5461909175,0.0617824495,0.2076323479,-0.1776831895,-0.3074544072,0.5304546356,0.3000192642,-0.3463909328,0.0579849817,0.1593796462,0.3236677945,-0.0097098798,-0.1332707554,0.1316142231,-0.3605438173,-0.3184635937,0.2939844429,0.5474277735,-0.0116180386,0.2071876526,0.1299876124,0.2318964452,0.0447096974,0.4415347278,0.0112903994,-0.1038311198,0.4746813178,0.0690966696,0.0515172556,-0.0045310771,-0.170958221,0.2669605613,-0.0338778459,-0.3202100098,0.01585426,-0.2754023075,0.4106704295,0.4015175402,0.2056024075,-0.0078711733,-0.1076532826,-0.006329474,-0.2898706794,0.1949425787,0.1890248209,0.1542851925,-0.3492989242,-0.3161615133,0.3706451654,-0.0070748185,-0.1426126212,0.1222494319,0.2197231799,-0.2660684884,0.3476285934,-0.1886654496,1.2954566479,0.270172298,0.1189190224,0.2170176208,-0.2496719062,0.0793893412,-0.4509885013,-0.1310372353,0.2621338367,0.2415788919,-0.0260866415,0.1937321573,-0.1302744746,0.2497690618,-0.4515494704,0.2120072991,0.0375548303,0.1302318424,0.1148661003,0.4553402066,-0.1565761417,-0.4293521047,-0.3479397893,0.2875838578,0.0790944397,0.1252751201,0.106623292,-0.2575728297,0.1208453104,-0.0314747877,-0.6170525551,0.1191837192,-0.18575266,-0.1361580044,-0.4633797705,-0.1735295802,0.3035571277,0.1740549803,0.4051637053,0.0084427362,-0.24363029,0.2013786733,0.0321877934,0.0925615504,0.1775135696,-0.1293593645,0.3848473132,-0.2787671983,-0.3417513669,0.240965426,0.1634778678,-0.0591214634,-0.4578181207,-0.2084450871,0.1368906945,0.0406026021,0.261434108,0.5941087008,0.0673772171,-0.2787318528,0.2033431083,-0.0538002849,-0.4027836025,0.1477177292,-0.1425073594,-0.2563803196,-0.1985164434,0.1851537377,0.2306480855,-0.0645662025,0.2740060985,0.3334341347,-0.1919670105,-0.3687168956,-0.0361462235,0.1379782259,-0.5049196482,-0.0674840882,0.1038753092,-0.1910448819,-0.3406735957,0.2093654871,0.0821783915,-0.0386002846,-0.0855515003,-0.032741785,-0.2290893644,-0.0254448541,0.129596144,0.1540955454,-0.0890791118,0.2274134606,0.0232216343,-0.1621649861,-0.4529195726,0.0261783171,-0.0793691576,0.16656138,0.2855002582,-0.160101667,0.1523312032,0.0281261764,0.1663667709,-0.1331900358,-0.0903979242,-0.2454134375,-0.1575043052,0.0915341452,0.125660494,0.1855226308,-0.0839325041,-0.1678895205,-0.2096366882,0.0896461755,0.3444300294,0.2475690097,-0.0392636172,0.1491406262,-0.0606044866,0.0110180983,0.1845854819,0.041136153,0.3104196787,0.3498603404,0.0484572388,0.145867303,0.0241687801,-0.1723857075,0.0252199378,0.3243961632,0.1631802619,0.1204419658,0.0427912772,0.0964939073,0.161920622,0.3878925443,0.0490579605,-0.0044768713,-0.0939858705,-0.0907652006,0.3871547878,0.3003925681,-0.1589111537,0.1513037384,-0.1665877402,-0.2660663426,0.1886512339,0.1115134731,0.1221079081,-0.0199779756,0.1961945742,-0.0571290106,-0.1632780582,0.066949673,-0.0034572675,0.1657529026,0.0572290756,-0.0530496836,0.0404480547,0.0304169804,0.0658825859,0.0716320649,-0.1389865428,0.4980093241,0.1549905539,0.0617301799,0.2575602829,-0.0103115579,0.1385445148,0.0412833244,0.150995031,0.1869845837,0.0523678847,0.3182862401,-0.0822582245,-0.289735347,-0.1189627796,0.2435652167,0.0293787178,0.0245356746,-0.0946952775,-0.2003645748,-0.2621451616,-0.063634485,-0.0260881335,-0.2240579873,-0.0124691846,0.1375160068,-0.1763918251,-0.0828602463,-0.1658290178,-0.2099787146,0.1894332916,-0.1993057728,0.339278996,0.1630556583,-0.0761632025,0.218065843,0.5704175234,0.46189031,-0.0398195833,-0.2486362457,0.2552427053,-0.0810579062,-0.1269970536,-0.0386592932,0.4618124664,-0.0094139231,0.1187777445,0.2067557126,0.2340582907,-0.2017280012,-0.1915644854,0.1524474472,-0.1433837414,-0.1096106917,0.1693624109,-0.0820122734,0.0917655602,-0.5450370908,-0.2100186199,-0.575600028,0.0418751203,0.4760661721,-0.2757143378,0.1057838872,-0.2261268497,0.1240856349,-0.1559839845,0.3006063402,0.2919948995,0.0539027192,-0.1142707989,-0.0973610058,-0.4213633835,-0.0655744523,0.1224700212,0.1246526241,-0.3954902589,0.0311454032,0.4504888058,0.3158788085,-0.1747768819,-0.1054390371,-0.4539692104,0.1466941684,-0.3867595494,0.2035492659,0.0492763892,0.6177328825,0.0677587911,-0.1090951562,-0.1967786551,0.3445982039,0.0974616408,-0.1278110743,-0.08028844,-0.1693009883,-0.2848241329,0.183917135,-0.0156083014,0.4005243182,-0.2893491685,-0.1023574471,-0.2012463361,-0.1380590051,-0.1710938513,-0.0229493435,-0.1017470956,0.2447318435,-0.1519684941,0.2543819547,0.0081156148,-0.1654618382,-0.0068544871,0.1092787609,-0.2397540361,0.0936172158,-0.1706962585,0.0245552622,-0.1309187412,0.269503355,-0.0044629271,-0.0716412291,-0.2301649451,-0.2515613735,0.4586344659,-0.1334103793,-0.385856688,-0.0362836644,0.3588258326,0.0566278361,0.0253875535,-0.696367681,0.238462016,0.2274999171,0.1636926979,0.0862355605,0.1958488375,0.0844282657,-0.0169707462,-0.1064812317,-0.0184868705,0.0402728729,-0.1692264229,0.1003186852,-0.296977371]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Done. They replied back, and they want to have a call over a meet\/ skype. Is that possible ? \r\nBtw @patrickvonplaten you are looped in that email (_pls check you gmail account_) ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":32,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nDone. They replied back, and they want to have a call over a meet\/ skype. Is that possible ? \r\nBtw @patrickvonplaten you are looped in that email (_pls check you gmail account_) ","embeddings":[-0.2442644387,0.0763138682,-0.1269343793,0.129224807,-0.0602062903,-0.1081656888,0.0954102203,0.0292590968,-0.1983942389,0.2287621796,-0.2163419574,-0.3068770468,-0.0361023583,0.2048379481,0.1446814537,-0.1205037832,-0.0763826296,0.0352106839,-0.2530809343,-0.3930398524,-0.0704426318,0.0100739123,-0.0025773447,-0.1456602216,-0.1974948049,-0.1006906107,-0.4669629335,0.0041094711,-0.2524196506,0.1215428188,0.0929485857,0.1669034511,-0.0617976375,0.2232315838,-0.0001050485,-0.1927651763,0.3581828475,-0.1252590716,-0.2843259871,-0.1371054351,-0.17150563,-0.1608715206,-0.5996175408,-0.0578071997,-0.3517877162,-0.0551304743,-0.0297531597,0.017230235,0.3423307836,0.5917037129,0.200846523,0.0822755545,0.4183850586,-0.1769323349,0.1135307401,0.1289752722,-0.0161012188,0.1953627914,0.0792288557,0.0627070963,0.4114525914,0.1769375801,-0.0387606211,-0.3774368763,-0.1423314065,-0.2057385594,-0.263856113,-0.4930153787,0.1322350502,0.2659471631,0.6440945864,0.013218699,-0.1394030899,-0.0239853859,0.2538210452,-0.2253908515,-0.0104750125,-0.1446998268,-0.2681514323,0.202164039,0.0474273935,-0.3013005555,-0.1538135111,0.4731569588,0.0228785742,0.1806114763,-0.0149981482,0.05533012,-0.0503499992,-0.2177546769,-0.3165231943,0.0157902073,-0.0045937295,0.1769250631,-0.2343796045,-0.4343778789,0.3938580453,-0.0072425478,0.0716760159,-0.1693592072,-0.1928205341,-0.0508283228,-0.1116915941,-0.0749313906,0.248502627,0.3352703452,0.1783816069,-0.0656065866,-0.0119934967,-0.264751941,0.1880943328,0.0778417289,-0.0760272965,0.1250343025,-0.4210846126,0.4101141691,0.0816383734,-0.1434594244,-0.258728981,0.1653543711,-0.188483879,0.1703430563,-0.0882416964,0.3126603365,-0.0846306235,0.1850612164,0.0221623778,0.1708114743,-0.0871563479,-0.1974159032,-0.1289276034,0.2370904386,-0.2407013774,0.1692819893,0.0548068359,0.4120897353,-0.0461805575,0.1059573516,0.42140311,-0.0272014048,-0.1701066196,0.0301685929,0.152664423,-0.0049024359,0.1104628295,0.1051588431,-0.1240601242,0.042750977,-0.1227745339,0.0407650135,0.0630776808,0.0138620408,-0.17445077,0.2680434585,-0.0474986993,-0.1905680746,-0.3329369128,0.4064121246,-0.2053744793,0.0053358376,0.2562300563,0.1369540095,-0.3320629895,0.1441873461,-0.0461158231,0.1919253021,-0.2280549407,0.0026738737,-0.0838654712,0.1308823526,0.2019740343,0.1595882922,-0.1489666849,0.2493917197,-0.1373122782,0.3556198776,0.0983255506,-0.4499589205,-0.3781096637,0.0318884142,-0.3757972121,-0.0230688266,0.0745265931,-0.2259061933,0.5838212967,-0.1840153039,0.0712356716,0.3171503544,-0.1127626896,-0.050895907,-0.0999286994,-0.3260999024,0.1334498078,0.0884797722,-0.0340640694,0.0560394414,0.0298629198,0.3126594722,0.1648179889,-0.253580004,0.0642090067,0.3471505642,0.1375506669,-0.0733180046,-0.0467644259,-0.0150378663,-0.1890818626,-0.0601273552,0.130899623,0.601962924,0.1781746894,-0.3046553731,-0.1507047564,-0.1705738008,-0.0035080879,-0.1373824924,0.2381709814,0.1877821833,0.1591787189,0.0274825972,0.0561035722,0.2445227206,-0.3581733406,-0.3407771885,-0.0449752212,0.2606641054,-0.0048399922,-0.2591896951,0.1548840702,0.3563862145,0.3486472368,0.037699692,0.057180956,-0.066099681,-0.2662833929,0.1654438376,-0.0018788103,0.0922554731,0.1486015171,-0.4431711137,0.1591223031,0.2065896243,0.2204488367,-0.3144567013,-0.0670315698,0.4534632266,0.4891100228,0.0362771563,0.1073851809,0.0254023448,0.3179667294,-0.0415212177,-0.1860003471,-0.0038363975,0.1625526845,0.2636128366,0.0410091877,-0.0934468359,-0.4189381003,0.2829908729,0.7185627818,-0.1966499239,0.2558173835,0.2333911955,0.0314025432,-0.2271121591,0.0466247201,-0.4440743327,0.2549783885,0.2788485885,0.2681042254,0.1341449916,0.1781595647,0.0483044647,0.1395656168,-0.0660702959,-0.2460415959,0.1130757332,0.0899870768,-0.0876070932,-0.2514942586,0.0853269622,-0.0450357646,0.1253264546,-0.2006136924,-0.3033833504,0.1786473393,-0.1797257364,-0.163015157,-0.1822063774,-0.3148111403,-0.0592153035,0.1858022064,0.0432194248,-0.0497578308,0.0558334813,0.3885269165,0.4585472047,-0.1845391393,0.069118239,-0.1597392112,-0.2830003798,-0.2647771835,0.1832286119,0.1687724292,0.2608466446,0.3231380582,-0.1066260114,0.0818492845,-0.3442077637,-0.3076981604,0.2393375039,-0.2616442442,0.1837855726,-0.0070955832,0.1523656845,-0.0198834781,-0.2366356105,-0.1570148766,0.2015193999,-0.264965713,-0.0971160904,-0.387085706,0.0635427162,-0.0555428863,-0.4665721357,-0.5568025708,-0.2878600359,0.4521857798,-0.2267304361,-0.0383325331,0.0272004046,0.0805587992,0.1978017986,-0.0809101015,0.0113973608,-0.4675084651,-0.2478359491,0.0229905192,-0.1271911263,-0.3666255176,-0.2452191263,0.0304946117,0.1717573255,-0.3337719142,-0.3176520765,-0.0713899508,-0.2055303901,0.2488624901,0.1033208296,-0.1296607852,0.3361968994,-0.3622721434,-0.2165345997,-0.1875326484,-0.0991905406,0.1613567024,0.3437241018,0.1981696635,0.1143776253,0.2781979144,-0.074097544,0.6019591093,0.1301018745,0.0468022898,0.1444503367,-0.0279523209,0.208512187,-0.2176887244,-0.5086151361,0.4292479753,0.1167939827,-0.2174568027,0.2087634206,-0.1543009728,-0.377273947,-0.1395045668,-0.4591851532,-0.0745690465,-0.2913773358,0.319124043,0.0843889117,-0.0995169282,0.0068290392,-0.1941947788,-0.2185028791,-0.1206646413,0.1302910149,0.126764223,0.0362565517,0.0884379074,-0.320933789,-0.1458193511,-0.4038719237,0.0179745946,0.1851765066,0.4507627189,-0.57643646,0.2884740233,0.085970901,0.0088289585,0.3867119849,-0.2152963728,0.0027770663,0.2585440278,0.4470023811,0.2073511332,0.2832023799,-0.196040839,-0.2514173985,0.4858677387,0.3070957363,-0.2200409174,0.0301257242,0.2143297046,0.1848351061,-0.1372194588,0.0297858417,0.0585819483,-0.3915015161,-0.3112423718,0.3392796218,0.5661070943,-0.036862772,-0.0179797299,0.0295436345,0.2167558223,0.0054741213,0.4134863615,-0.2228557467,-0.0817753524,0.3747658134,-0.0040299636,0.1148486808,-0.0174321923,-0.0260494575,0.3623434305,0.1521311104,-0.267906636,0.0653052405,-0.3942693174,0.474129647,0.3924329579,0.0945782959,-0.0311346315,-0.0038033118,0.1406337023,-0.2975140214,0.2862843573,0.3217409253,0.2720928192,-0.3405787349,-0.3129267097,0.3925609887,0.0251367837,0.0352513045,0.2253882885,0.2407797575,-0.1957075149,-0.0217290241,-0.0940744653,1.3331975937,0.4469963312,0.1533782035,0.1418835223,-0.2485107332,0.1071061343,-0.2331623733,-0.2041023225,0.2581066787,0.2154225558,-0.0954051912,0.2230089158,-0.0015439409,0.3154425621,-0.3548340797,0.2400844246,-0.2109332234,-0.0999525934,0.1728613973,0.4671241045,-0.0979210958,-0.2079847008,-0.2588572502,0.258687526,0.0988280326,0.2239103913,0.1647272706,-0.2662480772,-0.0426422916,-0.3018900156,-0.6961417794,0.1005690396,-0.3184010983,-0.3417184651,-0.5454695225,-0.1299562901,0.2826796174,0.0361798219,0.1627371162,0.0652981102,-0.2807626128,0.2030222118,0.0098909345,0.0995789617,0.1526010484,-0.1482622623,0.4521445334,-0.2773120701,-0.3156797886,0.2550986707,0.0058666025,-0.0623251349,-0.5047739744,-0.0513575487,0.0118712513,0.0162314959,0.1043771282,0.4319214523,0.0432415158,-0.2051658779,0.1985978931,-0.1214138195,-0.2531970143,0.0750246942,-0.0975746363,-0.2755206227,-0.1646804959,0.4504112899,0.1498967856,-0.2265397459,0.480900526,0.3105866015,-0.1643419266,-0.3908768296,-0.0836863145,0.2494299114,-0.5234825611,0.0770237297,-0.028225461,0.0369154215,-0.339037627,0.2942001224,-0.0409768373,0.0386597291,-0.099250637,-0.2383690923,-0.194749549,-0.0597779304,0.0821996629,0.1365941614,-0.0713142008,0.3921561837,0.1377859414,-0.2278036177,-0.4338139296,-0.0222944319,-0.2262010723,0.2120445818,0.4315142632,-0.1325905919,0.255800873,-0.0497281775,0.1494175941,-0.105143033,0.0277972445,-0.2623108923,-0.0429909751,0.0823613629,0.1180299446,0.1856115013,0.1323131025,-0.1052460149,-0.1197077036,0.2483598739,0.2296343595,0.1773598343,0.0611490048,0.0486340858,-0.0989586934,0.150998503,0.112355046,-0.1394876689,0.3923622966,0.4167675972,-0.1348117739,0.1931633055,0.0181413442,-0.0772503838,0.0899370909,0.1283002794,0.2243924141,0.0542402901,-0.099573262,0.1447252035,0.1009740308,0.2663749754,0.077474907,-0.0208057147,-0.1561101675,0.0022620682,0.2824702859,0.2942680418,-0.0485587716,0.1614083201,-0.1919209063,-0.3825166225,0.0310917068,0.0224927254,-0.043314185,0.1234690696,0.1932398528,-0.2046183199,-0.1561030447,0.0526252389,0.0568562783,0.1983299553,-0.1778887957,-0.0611464642,0.2848471105,0.0491128787,-0.0254123788,0.1481495351,-0.1519914269,0.2648295164,0.1013054922,0.2109750211,0.2045420259,-0.1911028624,0.1317051798,0.1000525653,0.1552893221,0.3295289278,-0.0076283808,0.39781636,-0.0441617705,-0.2146419585,-0.036216259,0.3101043999,-0.0988664404,0.0822315067,-0.0502833761,-0.2062361687,-0.4170131385,0.0113114044,-0.0209571384,-0.1331252158,-0.0127668492,0.0111771934,-0.2615368664,0.0350765884,-0.3658246994,-0.269230336,-0.003394299,-0.391061604,0.4116290808,0.2446995527,-0.1039444208,0.1712745875,0.4762534797,0.4604896307,0.013183658,-0.1329052895,0.2254962027,0.1744073927,-0.1647973955,0.0581590421,0.3664777577,0.0562983453,0.0434994437,0.0573599078,0.2743001282,-0.2606218159,0.010690649,0.0669599622,0.1382452548,-0.2532538772,0.1898958385,-0.1079583913,0.2467454374,-0.7149050236,-0.0774592832,-0.5943197012,-0.0391994677,0.4606373906,-0.0898026079,0.1058060527,-0.3524137437,0.1392451525,-0.3943318129,0.3255026937,0.2473961413,0.2696197927,-0.0399879403,0.0124636833,-0.174514696,-0.0506958067,0.1536047608,0.1097062752,-0.331490159,-0.0374084078,0.3865396678,0.1827248484,0.0390680656,-0.1203209534,-0.4215041101,0.1672347337,-0.3105468452,0.2724182308,0.0003340663,0.5711430907,0.1239023954,-0.1267716736,-0.0959695429,0.323210001,0.1145708561,-0.2118297368,-0.2216749489,-0.0772077516,-0.0915918797,0.0529427864,0.0604762174,0.804970324,-0.2886036336,-0.083520852,0.0689305663,0.0784924999,-0.1489465535,-0.0074676699,-0.0729132593,0.2028146833,0.036535915,0.3054816425,-0.0990558192,-0.2375438809,-0.0839894861,0.0064912601,-0.2925960422,0.0100081312,-0.1415822506,-0.0667386129,-0.1746595204,0.241853416,-0.012543234,-0.0749787688,-0.2496631742,-0.1244628504,0.4570589066,-0.0845226496,-0.2663291991,-0.0924235508,0.3227207661,-0.0391572528,-0.0024180375,-0.5286850333,0.1586453468,0.2303323299,0.0859344825,0.1709735245,0.2076752484,0.2388163507,0.1207788587,-0.111650221,-0.0106024239,0.2412818819,0.0256334282,0.1639561802,-0.3027096689]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Hello! Any news on this?","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":5,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHello! Any news on this?","embeddings":[-0.2956989408,0.2305583954,-0.0986739472,-0.0109735616,-0.0570029281,-0.0460772589,0.104497768,0.1535619348,-0.3449279666,0.1502928734,-0.2438719273,-0.1663403064,-0.1198064461,0.1203589365,0.2385118306,-0.0689408854,0.1566284597,-0.065976657,-0.3301940262,-0.2982098162,-0.1564448923,0.0650769547,-0.1644750237,-0.1840538532,-0.2196027786,0.0072394554,-0.3235074878,-0.2503673136,-0.2616667449,-0.2138739824,0.0626911521,0.1596456617,0.1075023338,0.440939188,-0.0001062752,-0.2200396359,0.2543959022,-0.2249466479,-0.3224855661,-0.1284592301,-0.2907099724,-0.2596996427,-0.5821906924,0.0489324257,-0.2430917025,-0.2514850497,0.0824828222,0.0615176782,0.0933476388,0.5752994418,0.2333545536,0.0619434267,0.2784287333,-0.1507459432,0.2929346859,0.1742874086,-0.1982347816,0.0906938985,0.2785720527,0.0988118947,0.1574473977,0.3573486209,0.0834388509,-0.1727536917,-0.0460724272,-0.3057814538,-0.131794706,-0.3586950302,0.0645952821,0.3904899061,0.7198499441,-0.1466852725,-0.207567811,-0.1006862447,0.2785320282,-0.1944948733,-0.0467448719,0.0299332794,-0.145578593,0.2507144511,0.1885866672,-0.3156897128,-0.2057638615,0.2763458788,0.0082482249,0.4687860608,-0.080335997,-0.088071458,0.0806108564,-0.2893324196,-0.2616129816,-0.0363472626,-0.0558825471,0.1316368878,-0.2387029827,-0.3420553803,0.292309165,-0.0792281553,0.0953241661,-0.4076322019,-0.192552194,-0.0770188868,-0.2106669992,0.0228790119,0.14441441,0.4009996653,-0.0162566975,-0.0011110976,0.1337204576,-0.3209312558,0.0504213236,0.1097804457,0.0576902628,0.182541579,-0.4096331,0.2281478792,0.0094401147,-0.2250670642,-0.2252515852,0.1095725447,-0.0224691834,0.0564328879,-0.1005212069,0.2819555104,0.1522694826,0.344045341,-0.0669627562,0.3172640204,-0.0162122119,-0.1469669491,-0.0576514713,0.0921782181,-0.2720851302,0.1674881577,0.2114157379,0.3986349404,0.0210573114,0.0446054675,0.2830730975,0.1081777588,-0.1930010766,0.0007148999,0.3274924159,0.0163218789,-0.0234250836,0.0887124836,-0.2283210456,-0.1782959998,-0.229201287,0.0901916772,0.1319716126,-0.0483669266,-0.2584284842,0.2465515584,-0.1388818175,-0.2486284375,-0.2886458933,0.7779040337,-0.1482392848,-0.1235402152,0.108815074,0.0523566864,-0.4833891392,0.0437839404,-0.1257819235,0.2632654905,-0.3276234269,-0.0689105317,0.0598533191,0.0737278983,0.018573219,0.1770528704,-0.1099017411,0.2971953154,-0.101222977,0.3227799237,0.0118560325,-0.408110857,-0.3176472187,0.1511346847,-0.1372952163,0.01460288,0.2022836357,-0.1139922664,0.6271263957,-0.3166701198,-0.0954003483,0.3339485824,-0.03671287,0.0843460038,-0.1912005693,-0.3413763046,0.3294296861,0.1087894738,0.0455801822,0.0313256271,0.0520923771,0.3437634706,0.3203746974,-0.2463217229,0.2328122705,0.2598080039,0.1313596368,-0.1795992255,0.1374517381,-0.1214859858,-0.123389177,-0.0811716989,0.1887345612,0.5802809,0.1473055631,-0.1994154006,-0.027114993,-0.1975615174,-0.1587149948,-0.2043090463,0.2288692892,0.294759661,0.0592361949,0.0843114331,-0.1517544091,0.2889250517,-0.3232259154,-0.2672577202,-0.0012290083,0.3019957542,0.1164445132,-0.3025855422,0.1288427114,0.3991461992,0.162652269,-0.0790680647,-0.0197336879,0.0674959123,-0.2179573774,0.2270544171,0.0966427922,0.175154224,0.3279239237,-0.5631665587,0.1716937572,0.3560693562,0.1607898623,-0.2412670851,-0.0988941342,0.1943202913,0.3042269945,0.1304118931,0.0504515469,-0.0897070244,0.4006076455,0.0177065954,-0.1792045534,-0.0522362813,0.1295823008,0.2734588385,-0.2030187845,-0.0230700243,-0.2812561691,0.1726363003,0.526979506,-0.1985149682,0.2605901062,0.232957527,-0.010322568,-0.160489291,0.0724448115,-0.2421514392,0.354285568,0.2986808419,0.2612219751,-0.0102310991,0.1743431389,-0.062076252,0.2079936117,0.0363048837,-0.2853108346,0.1875076294,0.1964538246,-0.1746705025,-0.3631158173,0.1085606441,0.0595936589,0.1615969986,-0.1192358434,-0.3293590546,0.1033086851,-0.203886345,-0.232380569,0.0092933215,-0.2198770642,-0.0490813404,0.3962935507,0.074004285,-0.1427656263,0.2445901483,0.2405261844,0.3260232508,-0.4569060504,0.0095518725,-0.0455612689,-0.3960086703,-0.1918857247,0.200962618,0.186078459,0.1336111724,0.1455569118,-0.1217517257,0.1685850471,-0.3306859732,-0.3644226193,0.185467869,-0.2945003808,0.1276147813,-0.2409420758,-0.0730881914,-0.0067862249,-0.4413414896,-0.2512781322,0.0031568224,-0.0837650597,-0.0703210607,-0.4539618194,-0.0698327646,-0.0991457999,-0.4489509463,-0.5105497837,-0.3732939959,0.3780188262,-0.2527760863,-0.0540707968,0.0772762373,-0.0542430021,0.0750687867,-0.0701686516,0.0441267155,-0.2930374444,-0.1399548352,0.091787979,-0.30402565,-0.3492863476,-0.211294502,-0.0069296719,0.0955719873,-0.1531015486,-0.3011255562,0.2184252441,-0.1033780202,0.3771996498,0.1344978213,0.0117596267,0.2771832943,-0.1506641656,-0.1341818422,-0.1623888016,-0.1498472393,0.0546154529,0.4331317246,0.2963982224,0.0242197234,0.0308436304,0.0169453621,0.5340762734,0.1473045647,-0.0298484564,0.2131877989,-0.0027455806,0.4338149726,-0.2372909188,-0.4277289808,0.5333976746,0.0685156509,-0.0300053786,0.4468079209,0.0522650145,-0.4814766645,-0.1166478917,-0.2037035972,-0.1750022024,-0.2422665507,0.275976032,0.1386599243,0.1943489611,0.0849423483,-0.1574476957,-0.2719851136,-0.2585580945,0.2030788064,0.1161344126,0.0671662241,-0.1047269329,-0.2149492502,-0.056236133,-0.4126322567,-0.0953792408,0.189037472,0.3350880146,-0.5270231366,0.0376123786,-0.039085459,0.1653494835,0.3278123438,-0.3649829328,-0.0899017081,0.3301155269,0.5309472084,-0.0051639602,0.1453696191,-0.2519061267,-0.2881290019,0.6618431807,0.3467318118,-0.2426891476,0.0359117836,0.202759549,0.3115834594,-0.0161200166,-0.1665756255,0.0314951502,-0.273609221,-0.2712801099,0.2991302609,0.4530153275,-0.0183764398,0.1506319493,0.1012261435,0.1939867586,-0.1219295189,0.4447730184,-0.072006233,-0.0241095275,0.5696542263,0.1101612896,0.0081401076,-0.169799149,-0.1448353529,0.193726629,-0.0704288632,-0.3649483025,0.1190790981,-0.3278214335,0.4001267254,0.3690628409,0.1602093875,0.0594765358,-0.1464326382,-0.0268235169,-0.2452083379,0.1565323472,0.2808238566,0.2119756341,-0.3384618461,-0.2336264104,0.3480722308,-0.0154090645,-0.134098351,0.1589498222,0.1396758705,-0.2616274059,0.2646701336,-0.1992669106,1.2451448441,0.2985204756,0.0640053451,0.1842274815,-0.319832027,0.0730815157,-0.4284185767,-0.0851328149,0.3048595786,0.1828883886,0.0238328092,0.2045857608,-0.0075935936,0.2116768509,-0.4403014779,0.2281213701,0.035420768,0.1782235056,0.136243552,0.4779459238,-0.1678078473,-0.3523461223,-0.261246115,0.2676633,0.0938324109,0.1632834971,0.1410367638,-0.2629902065,0.2041913271,-0.0231007971,-0.6556848288,0.1664325446,-0.1092993245,-0.0816405863,-0.3931180537,-0.1343807876,0.3120235801,0.1669974774,0.4421290159,-0.0112133212,-0.249113217,0.2887278795,0.0325063169,0.1793992519,0.1986146569,-0.0909873843,0.3797642887,-0.2942943275,-0.3459847867,0.2326270342,0.1478459388,-0.0825069547,-0.4550441802,-0.2049534619,0.1309652925,0.0727090687,0.2789282203,0.6814756989,0.0272291191,-0.3165540993,0.193432644,-0.0566017926,-0.4473202229,0.286654979,-0.2161513716,-0.3347171545,-0.2324706912,0.2618374527,0.2472141832,-0.0398840234,0.3973144889,0.2909885943,-0.165901646,-0.3595155776,0.0426891781,0.1478690654,-0.5135236382,-0.0011374662,0.1095852852,-0.103932783,-0.2477984279,0.3231518567,-0.0132726766,-0.0039870897,-0.1518813372,-0.0183082931,-0.2409847826,-0.063380599,0.0556718707,0.1400678903,-0.1087462604,0.2809093893,0.1101925969,-0.106043756,-0.4191497564,0.05068928,-0.0202926639,0.2180067301,0.2367084175,-0.1596224457,0.1649279743,0.0008124203,0.1657047272,-0.0358917415,-0.0668511614,-0.2228568792,-0.1817135066,0.0816267133,0.1460428387,0.1806349903,-0.0677597001,-0.1590324342,-0.1951942146,0.1128323525,0.3026595414,0.2311593592,-0.0712193772,0.211478129,-0.0098083206,-0.0302665159,0.3123317659,0.0452382639,0.3317789733,0.359121114,0.1310112923,0.2056396306,0.0642623007,-0.1018678173,-0.0428925343,0.2830486894,0.0604752824,0.0179369915,0.0060994499,0.0562063381,0.1425454021,0.325058192,0.0478976965,-0.0557328053,-0.0421590917,-0.1509943455,0.3588652611,0.2491127104,-0.2302398235,0.1407287419,-0.2041653693,-0.1961517036,0.1321987957,0.0339570828,0.1091388911,-0.0873634666,0.3852107525,-0.0818129554,-0.1917756647,0.0124920178,0.0389830507,0.1891668886,0.0846242905,-0.0782449394,0.0876205638,-0.0412016809,0.0346378982,0.1107839942,-0.1182224602,0.4903016686,0.1049344912,0.1439719349,0.3262884021,-0.0574943088,0.0748922825,0.0718528554,0.0069628488,0.2374482602,0.1160138324,0.2819061279,-0.0663613603,-0.2984737158,-0.1743558198,0.21875301,-0.0201167278,0.0160964634,0.0101283537,-0.1534364671,-0.2117804885,-0.0231155865,-0.0141435955,-0.171453476,-0.0651118606,0.1820529997,-0.2075541914,-0.0761687309,-0.0661388636,-0.2953993976,0.2066977024,-0.1987203956,0.3713853061,0.1141204238,-0.0521661863,0.2474440187,0.5845698118,0.4274490476,-0.0388191044,-0.231845662,0.2258069813,-0.0940103009,-0.1389348209,0.0671281964,0.5133994222,0.0712930188,0.1592387706,0.1614628434,0.2340639085,-0.1606775969,-0.156359002,0.1747955382,-0.1414051503,-0.2109396756,0.1788773835,-0.0752569139,0.1203282699,-0.5953325033,-0.1529376358,-0.5126851201,0.0312176701,0.4356252253,-0.2938156426,0.0662272647,-0.1695996076,0.1099321395,-0.2114325464,0.271492064,0.2459295541,0.0472728759,-0.1341008693,-0.1617973894,-0.3878321648,-0.0966424868,0.1650045365,0.1568145454,-0.3925331533,0.1310368031,0.4047833085,0.2824552953,-0.1050011218,-0.1160803363,-0.4150939882,0.090273872,-0.3977094591,0.215828225,0.0147119882,0.5993248224,0.033278726,-0.0638782978,-0.1906842589,0.273409456,0.0655221567,-0.1187033653,-0.1593221724,-0.1848053038,-0.1720726788,0.2090589404,0.0053843041,0.491111815,-0.2274378538,-0.0099183721,-0.1649256349,-0.2205870301,-0.0861662626,0.0341953896,-0.1114344746,0.2120389193,-0.1512003988,0.3231144547,0.006040337,-0.1851031184,-0.0566714332,0.0650394633,-0.265848726,0.1228578612,-0.1693635732,0.0009912789,-0.15273498,0.2100715041,0.0277081709,-0.0556952059,-0.1969781816,-0.1856122762,0.441521585,-0.1462082565,-0.3689059913,-0.1455581635,0.3443982005,0.0736668333,-0.0502180532,-0.7440784574,0.1335210204,0.273227483,0.1709235311,0.1391694397,0.1807017624,0.113115564,0.0053456444,-0.1382143945,-0.0798649788,0.1073936969,-0.1603657752,0.1336980909,-0.2696097195]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"@gegallego there were some concerns regarding dataset usage & attribution by a for-profit company, so couldn't take it forward. Also the download links were unstable. \r\nBut I guess if you want to test the fairseq benchmarks, you can connect with them directly for downloading the dataset. ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":46,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n@gegallego there were some concerns regarding dataset usage & attribution by a for-profit company, so couldn't take it forward. Also the download links were unstable. \r\nBut I guess if you want to test the fairseq benchmarks, you can connect with them directly for downloading the dataset. ","embeddings":[-0.3551790416,0.2733169794,-0.0848916322,0.0439710058,-0.2106078863,-0.0997216403,0.0143609112,0.1664302051,-0.3334971368,0.1858729571,-0.2543601096,-0.0773932114,0.0546253063,-0.1191610023,0.1883631647,-0.2674564421,0.1272602677,0.0159447584,-0.4990527034,-0.2887564301,-0.0022719416,0.082528919,-0.3534404039,-0.0658291206,-0.2267103493,-0.0015485308,-0.3274188936,-0.1423062831,-0.2602281868,-0.2481784523,-0.059322793,0.1865078807,0.072331503,0.503970325,-0.0001057738,-0.1889287084,0.3732982874,-0.2488802522,-0.3008495271,0.0820415393,-0.2667974234,0.0794181228,-0.5366252065,-0.0851109251,-0.227863729,-0.2843920588,0.0211572014,-0.054533191,0.1495311707,0.6374393106,0.2225167006,0.2601231039,0.0848627537,-0.1991487443,0.1856092811,-0.1639837325,-0.0690774843,0.1905391067,0.2750355601,-0.0210324321,-0.0819738954,0.2998503447,0.1042219549,-0.0445661023,-0.2373772264,-0.3053140938,-0.1707397848,-0.3454199433,0.1246996969,0.537383914,0.5419611931,-0.1456474215,-0.3013556898,-0.0213013608,0.2159660459,-0.0139286295,-0.0258551277,0.0006631022,-0.137916103,0.224688977,-0.0034156383,-0.0748566762,-0.2043162882,0.1544621736,-0.0655704364,0.4908934236,-0.021509219,-0.1548563987,0.1846320629,-0.2481155694,-0.2701608837,-0.0257912558,-0.2289420962,0.0444236659,-0.1947868019,-0.3074363172,0.2198108435,-0.1459584981,0.1053978354,-0.5148681998,-0.2520375848,0.0767925829,-0.0298419483,0.0380183831,0.0287979525,0.4303641021,0.0822535083,0.0139246676,0.0143891051,-0.1286827326,-0.0248530935,0.2848423719,0.0085046822,0.1234905124,-0.6589146852,0.2396982759,-0.1956278682,-0.342302382,-0.3911923468,0.1041132733,0.1085189506,0.0121728731,-0.1071960703,0.3425152898,-0.0321455114,0.118641898,-0.2888519168,0.3779186904,-0.2201810777,-0.3948869407,-0.0638384148,0.2210025936,-0.2729827166,0.0772538185,0.3387737572,0.4771769941,0.1867390275,-0.0889019147,0.2831927538,0.0326979794,-0.0412758626,-0.0643995702,0.2550360858,-0.0618495941,0.0918401331,0.0108091421,-0.1270463318,-0.2597824037,-0.2591841817,-0.0066581913,0.1293067187,-0.0338126682,-0.1639392376,0.2301292568,-0.1351456195,-0.2654020786,-0.2642100155,0.6950773001,-0.2816907763,-0.1759266257,0.1762406677,0.0513033383,-0.4213475883,0.2301777005,-0.0083591752,0.3726105988,-0.3870705366,-0.0569722503,0.1288463771,0.1356600821,-0.020216912,0.1602196693,-0.1545455009,0.284265399,-0.0705448464,0.3429217935,0.1554594636,-0.4663978517,-0.4266594648,0.1362527758,-0.0537789501,-0.0149403289,0.1475269943,-0.2671592236,0.7280392051,-0.2025388926,-0.0357537717,0.3976615667,0.1310447901,0.1978919804,-0.2383537441,-0.2894970477,0.3342309892,0.1956883371,-0.0242709238,-0.0691482723,-0.0218083207,0.0838326588,0.157940492,-0.2965834737,0.1482635736,0.1980035454,-0.0169710424,-0.3243277967,0.1154374704,-0.1188992932,0.014397583,0.0877882913,-0.1610722542,0.5759599209,0.3111337423,-0.1469516456,-0.2139986455,-0.2347021848,-0.0771126822,-0.3021524549,0.2228336036,0.1802975386,0.1322675645,0.2183843106,-0.062743023,0.182993874,-0.1797519475,-0.30597651,0.0632365942,0.1546317786,0.0335872434,-0.2636297643,0.2943400443,0.2889519036,-0.0088170739,-0.0170700569,-0.1283515096,0.1494235247,-0.1889591068,0.3021014035,0.3540577888,0.0878100991,0.2557058334,-0.529687345,0.1047673374,0.4869792163,0.1743325889,-0.0821095556,-0.2045048773,0.4014042318,0.3025280237,0.0774848387,0.2597222626,-0.171565637,0.2747779787,0.0110660503,-0.1812476814,0.0398397781,0.373065412,0.2577475309,-0.1858857274,-0.0040043979,-0.2110439241,0.2028787434,0.6933574677,-0.1499216259,0.446418643,0.2341549397,-0.147367999,-0.2394641191,0.1144062355,-0.3307747245,0.3838,0.2252084613,0.3287408054,0.0317139067,0.1154776663,-0.0175304655,0.2093215138,-0.0489870422,-0.4404384792,0.0404124707,0.2045613825,-0.2765510678,-0.2801454961,-0.0269128866,0.0115179829,0.1130231842,-0.1419904232,-0.1876505762,0.0164453164,-0.0435457528,-0.1173078939,-0.0280658212,-0.2799026072,-0.0129600978,0.4996726811,0.0351664834,-0.0776687562,0.294532001,0.2281565964,0.409661293,-0.207163617,-0.0615271926,-0.1380939931,-0.4016678333,-0.3557342291,0.2032827884,0.1243805811,0.4340555668,0.0677910149,-0.1708564758,0.2648128867,-0.1478478611,-0.1642362028,0.2659855485,-0.2536759377,0.3071198165,-0.355535984,-0.0992638096,-0.2620576322,-0.2119558454,-0.193489328,-0.1172834635,-0.0193766765,0.0092816073,-0.3579931259,-0.2484480739,0.0018924987,-0.4207038283,-0.357475698,-0.3441028595,0.2667474449,-0.3142565489,-0.0457539633,-0.2331577986,0.0538049452,0.2283514589,-0.1181003526,0.1282993555,-0.308658272,-0.3723421097,0.2361965328,-0.3731283247,-0.434738785,-0.1901666969,0.1128448844,0.2009081692,-0.1962303817,-0.2799493372,0.0898401886,-0.0757873431,0.2916795909,0.1934456378,0.0735822096,0.2507287264,-0.2208983451,-0.1460872144,-0.0767676681,-0.1419749856,0.1074151099,0.4329066575,0.1741035283,-0.1641847938,-0.0780729726,-0.039777901,0.7484592795,0.181869626,-0.0361330807,0.1550901681,-0.1157272384,0.2861981094,-0.1794919372,-0.2003703564,0.455927074,0.0702138096,0.0796504915,0.5873463154,-0.0596508831,-0.4037899673,-0.1314626783,-0.0634213239,-0.3109101355,-0.2188238502,0.2448400408,0.1556647718,0.120245263,0.0925120339,0.0067120041,-0.149914816,-0.2149372399,0.3076753318,0.063006714,0.0040882481,-0.0539754182,-0.1839215457,-0.0986998901,-0.2330609709,0.0450905673,0.1255947948,0.4172136486,-0.3388797045,0.1059628874,0.01413895,0.0712674558,0.3088402152,-0.2470687777,0.0061388179,0.267468363,0.4904707074,-0.0374626145,0.1244219244,-0.3506262004,-0.2744368613,0.5050278306,0.0443203673,-0.2060312033,0.1372445673,0.184447512,0.1801815182,-0.0555897504,-0.0532872081,-0.1777570099,-0.3515924513,-0.1624947339,0.4372068942,0.2988068759,-0.0007254976,-0.0671832338,0.245292291,0.1170035973,-0.081197992,0.3922603428,0.1679964662,-0.0293732081,0.3057389259,0.0886811391,0.0284566171,-0.2849123776,-0.1788627207,0.176321134,0.2273960412,-0.1764005125,0.2287473232,-0.2436403781,0.2854171693,0.2611733675,0.1136235371,-0.0513847433,-0.094604522,-0.0017239543,-0.1671428829,0.2643266022,0.2562303245,0.273662746,-0.2923051715,-0.0853915736,0.4494597614,-0.0086117396,-0.0877857134,0.0416331179,0.1138876975,-0.2260627598,0.2638059855,-0.1847439259,1.1430470943,0.1728739738,0.0363627262,0.2057212293,-0.1620518565,0.145541802,-0.5122431517,-0.0285773166,0.2491752952,-0.0642716959,-0.0600987561,0.1466674507,0.1465670913,0.1577019542,-0.6139907837,0.4243773818,0.15433608,-0.1389874667,0.2433599979,0.537838757,-0.0495267548,-0.3192803264,-0.3945882618,0.2510227859,0.0403646044,0.2110508829,0.2064958513,-0.1221386865,-0.0591707937,-0.0198474377,-0.5190370679,0.2081212103,-0.0438884757,-0.1314903349,-0.1924881041,-0.035058409,0.293199867,0.083400555,0.2890949845,0.0464947261,-0.271497637,0.3969090581,-0.0166866891,0.0681994185,0.3073756397,-0.1166910157,0.2369596809,-0.2375442833,-0.393628329,0.4314379394,0.1235586703,-0.0235867612,-0.4086910188,-0.3094980121,0.0325172469,0.114695482,0.2270229608,0.4223170578,-0.0373032615,-0.21045506,0.1770060062,-0.1110258624,-0.4937492013,0.3757799864,-0.0994275212,-0.3258656561,-0.1752554327,0.3120812774,0.0449592434,-0.0160839614,0.2687599063,0.3269920647,-0.0578739829,-0.358114481,-0.1832266897,0.1534422189,-0.5673998594,-0.0913339704,-0.1173205078,-0.1728059202,-0.1916518509,0.3114621937,0.0914899036,-0.0470951796,-0.0389835685,0.0308084525,-0.1744134277,-0.0822182372,0.1760894209,0.1084947661,-0.1208394021,0.1085435078,0.0495063514,-0.063469395,-0.4071820378,0.0170549359,-0.1513054818,0.1756139547,0.1496413797,-0.1506814361,0.1005388647,0.1411656141,0.1392618269,0.0018434766,-0.1265974641,-0.241028443,-0.1531740427,0.079324834,0.0233224779,0.0961774886,-0.1620374024,-0.0405350029,-0.1994916052,0.0279436242,0.1965480894,-0.0815808773,-0.1364135146,0.1721741259,0.0433891416,-0.1003298461,0.1493702978,0.0396924652,0.2879952788,0.3841156065,0.0771039054,0.3171363175,0.0695785433,-0.1613859534,-0.0354915597,0.3563213348,0.0811975002,0.1574619263,0.0134866666,-0.0027524666,0.2572341859,0.1577939242,0.1114850789,0.0853400975,-0.1115772426,-0.3110021949,0.3851221502,0.2449854612,-0.1600892842,0.0531223528,-0.1723873317,-0.1483217478,0.2773464918,0.0131062195,-0.005473299,-0.0127045466,0.4179860055,-0.0069043068,-0.1311264187,0.1008449048,-0.061286442,-0.1240780577,-0.0041741007,-0.0530043617,0.0302392151,0.0252931844,0.0865380615,-0.0241770931,-0.0087013505,0.5746207833,0.2038387805,0.3148146868,0.1140264943,-0.0255135521,0.1445300579,0.1809156984,0.2090104669,0.3537558019,0.0860843882,0.1479681283,-0.2432384342,-0.3147220612,-0.385666728,0.0855127797,0.205939427,-0.1256522983,0.1257867962,-0.2386727333,-0.2533558309,0.0628036782,-0.070408605,-0.013311943,-0.0797583759,0.0935593992,-0.0858781636,-0.1089212969,-0.2355791777,-0.2524141073,0.1721628755,-0.1859836876,0.4769241512,0.106545344,-0.0075563407,0.1642943472,0.6191945672,0.2939888537,-0.0460428447,-0.4676012993,0.1814693809,-0.1488854885,-0.0172345489,0.1076630652,0.4670607746,0.128986448,-0.0136792632,0.2001292408,0.2269101441,-0.1932051182,-0.1744666398,0.0669059604,-0.1085856408,-0.2420452386,-0.0239252318,-0.0304644573,0.3559219241,-0.4321718514,-0.2356850803,-0.5966399312,-0.032560885,0.2817835212,-0.29851228,0.0642384067,-0.1308932453,0.1120688692,-0.212472707,0.4738662541,0.2863326669,0.4016743004,-0.1726588607,-0.1763755977,-0.3388219178,0.088556923,0.0695935637,0.1894707978,-0.2420598119,0.2347683907,0.2082526833,0.4279851317,0.009279334,0.038901709,-0.4901141226,0.1881483495,-0.3655131757,0.2048383802,-0.1874970049,0.4944732189,0.0081430888,-0.1845751107,-0.1346410364,0.1167535931,0.0916789994,0.033401154,-0.1726953238,0.037582837,-0.2308252156,0.2265029401,0.009993284,0.454685539,-0.1057840511,-0.1265780032,-0.1931913793,-0.0731503218,0.0316827893,-0.1231088564,-0.0400572456,0.1766234934,-0.0161212683,0.2943894267,0.0081639169,0.0185416564,0.1709729731,0.0798277259,-0.2254880965,-0.0110623911,-0.0963843092,0.0271436349,-0.1095998883,0.2335269749,0.0589974113,0.127686426,-0.1161570847,0.039712362,0.4405524135,-0.3146034777,-0.3092559576,-0.1535411924,0.3522535264,0.0573392324,-0.1025385633,-0.5595287681,0.1816343814,0.3045867383,0.1808069646,0.1467268467,0.1256610006,0.0741314292,0.0858886912,-0.1564196497,0.0440293439,0.0584053583,-0.0969724804,0.1654687375,-0.2397267073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"Yes, that dataset is not easy to download... I had to copy it to my Google Drive and use `rsync` to be able to download it.\r\nHowever, we could add the dataset with a manual download, right?","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":37,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nYes, that dataset is not easy to download... I had to copy it to my Google Drive and use `rsync` to be able to download it.\r\nHowever, we could add the dataset with a manual download, right?","embeddings":[-0.3261700571,0.258577466,-0.0921744332,-0.0333687477,-0.0554307587,0.0201777853,0.0042871945,0.1221487597,-0.3325349391,0.2870291173,-0.2463117093,-0.1663521677,-0.0883385614,0.0430673994,0.1519542783,-0.046092838,0.100588128,0.0084114205,-0.281600982,-0.3170711398,-0.0999767557,0.0702602789,-0.1038936526,-0.1798479259,-0.1023112983,-0.1136202365,-0.2879876792,-0.1292617768,-0.3628219962,-0.1610624194,0.0828778818,0.2118594646,0.105418317,0.3385773599,-0.0001049832,-0.1688212454,0.2188206315,-0.2101087868,-0.4650355875,-0.1227784529,-0.2633396387,-0.1900159568,-0.5593047142,-0.0589392446,-0.1318376213,-0.1538991183,0.190631181,0.1063847914,0.1001442298,0.5843331814,0.2289637476,0.0718084648,0.1717671901,-0.2216983438,0.3009495735,0.1132302135,-0.1445614547,0.2121609002,0.2020394355,0.1368398666,0.187990889,0.2498551309,0.0623006821,-0.1541409194,-0.0312925652,-0.261718303,-0.1831347942,-0.4605362415,0.1067013592,0.4221014082,0.6699703336,-0.1604393721,-0.2226737291,-0.1112763435,0.3161411583,-0.1318680793,-0.041566968,0.0762252808,-0.1802818477,0.2422288358,0.0878616571,-0.3472367525,-0.1816349775,0.2786091268,0.0030743231,0.3151161075,-0.098924309,-0.0851302147,0.0590563603,-0.1815015078,-0.3160818815,-0.0776870623,0.0200668816,0.1641677916,-0.048442252,-0.2551894188,0.260302633,-0.1156946421,0.1274390221,-0.3338730931,-0.190826565,-0.112936534,-0.1809577346,0.0327018015,-0.0087539991,0.3533169627,-0.0389137715,0.0067527103,0.1603179574,-0.2577740848,0.1317540556,0.0949309617,-0.1060991213,0.1061939597,-0.4424651563,0.2819646299,0.0175171215,-0.1521286815,-0.3608736992,0.0009045305,-0.0983712077,0.0952640176,-0.1840298623,0.2711251378,0.0594117008,0.2206312567,-0.029945977,0.3511986434,-0.0816177949,-0.1676343679,-0.0635019541,0.1821158528,-0.1835427135,0.1410641223,0.2064755857,0.4040096402,0.1354864538,-0.0382832438,0.2191948742,0.1065601483,-0.2609094679,-0.0230158325,0.1315094233,0.0475583784,0.1305355281,0.2011219114,-0.2909487486,-0.2401168495,-0.3164714575,0.0372383147,0.1165111288,-0.0324811749,-0.1799694747,0.2699035406,-0.1149367914,-0.2823370993,-0.3300674558,0.6548683047,-0.1671614498,-0.1585751772,0.0269490536,0.1318160892,-0.407247752,0.0298855491,-0.0919469222,0.3723148108,-0.3409881592,-0.0958041251,0.040777918,0.0129324039,0.1060925797,0.1649013758,-0.0454858318,0.3797605336,-0.1272123456,0.3340143859,0.17381154,-0.2973581553,-0.4280662537,0.1087315232,-0.1839875877,0.0562566221,0.2124160379,-0.0888002664,0.6740357876,-0.2896543741,-0.0530236699,0.5153852701,-0.0072862851,0.0726594105,-0.186473161,-0.387589246,0.2624821663,0.116539456,0.1313602924,0.108828038,0.1810255498,0.3644085824,0.3500255942,-0.270799011,0.1428896338,0.3335073292,0.1223838404,-0.1555511355,0.0128864944,-0.037968386,-0.2598908544,0.0124909235,0.0837915912,0.5217408538,0.0192428827,-0.2437839061,-0.1016326249,-0.2844207883,-0.1674671471,-0.0977725983,0.2394647449,0.3028955758,0.1553970575,0.0371332243,0.0348586328,0.2901745439,-0.3526088297,-0.2636925578,-0.0762509704,0.1294235885,-0.0167735424,-0.1908477396,0.1762052029,0.2271558791,0.1990326196,-0.0735970438,-0.017036872,0.1629077792,-0.232838124,0.3128562272,0.046399992,0.1513501555,0.3106359839,-0.5115110874,0.1572942436,0.416193217,0.1467899382,-0.2736437023,-0.2458366305,0.1721504927,0.3308235109,0.1557471156,0.2433720678,-0.0184064247,0.4958963096,0.1071761325,-0.1974094659,-0.0540437028,0.1498032212,0.3694210351,-0.0956173837,-0.0542924516,-0.3067092001,0.2138965428,0.6212291121,-0.1969293207,0.2406687289,0.252047807,-0.078820996,-0.1463530511,0.0290928129,-0.1906289458,0.2503053248,0.352750659,0.3592058718,0.1053783,0.153202638,-0.1018713862,0.1213005856,0.0863634795,-0.2057517022,0.1337576658,0.1239680275,-0.0941411182,-0.4168773293,0.0197334085,0.0448354296,0.1779794991,-0.1667906344,-0.3204413056,0.1141039059,-0.1600840837,-0.176881507,0.0601245798,-0.1996935457,-0.1120592356,0.3303552568,0.0985862464,-0.1896595657,0.2049132138,0.2834319472,0.2777814567,-0.4166877866,-0.152523607,-0.0159285106,-0.3203995228,-0.2732067704,0.1860576868,0.2324534059,0.2177059203,0.2699485719,-0.2892925739,0.1873935014,-0.3356782496,-0.3198469579,0.2317831963,-0.2485967875,0.0674072877,-0.2411665469,0.0807213262,0.0027066029,-0.3551658988,-0.299734503,-0.0125589371,-0.1087993756,-0.0983541757,-0.4978320003,-0.2145543098,-0.0878387541,-0.5328082442,-0.5634722114,-0.3825628757,0.3526053429,-0.1424562335,-0.0276196953,0.0433124378,0.004082893,0.1098218262,0.029335482,0.2357150018,-0.2832208872,-0.0561409518,0.2433838248,-0.2432825565,-0.4001317918,-0.1083380654,0.0537583083,0.0501201563,-0.0360996947,-0.3312400877,0.2080515027,-0.1621937305,0.3929625154,0.1319549531,0.1293594539,0.3479273319,-0.1876190156,-0.1273127049,-0.1407837123,-0.1103456616,0.0769040734,0.4300549328,0.2432563901,-0.0991315767,0.1157812253,0.0324397944,0.6805483699,0.1629751176,0.0503169149,0.1335610896,-0.0671359673,0.4736748636,-0.2270119637,-0.2605061829,0.457993865,0.04270925,-0.1145168319,0.4364804029,0.0087467739,-0.4343087077,-0.1088702381,-0.1994495094,-0.2343461215,-0.2816282809,0.2488924414,0.1271236986,0.0976998061,0.2011646628,-0.0411907025,-0.1799499393,-0.1331418604,0.2895573974,0.1178946346,0.0575452782,0.0288985651,-0.1530006677,-0.0646922663,-0.5273178816,-0.053155683,0.2005153447,0.2830536067,-0.4398043454,0.0145412516,-0.0356193148,0.0901983082,0.3260068595,-0.2868438363,-0.0256013498,0.2564719617,0.6237996221,-0.0327490456,0.0954099968,-0.3140833974,-0.3194309175,0.6827363968,0.3148851693,-0.3409000337,0.0824546814,0.170137763,0.4130093455,-0.0726604387,-0.0695415065,0.0070249909,-0.5352041721,-0.3132354915,0.2133857757,0.4288982749,-0.0136697534,0.1045398191,0.1903572679,0.1677885652,0.0067857727,0.3933689892,-0.0670755506,-0.0267136749,0.5132446885,0.0416399427,0.001666063,-0.0651830658,-0.1873146743,0.4569407701,-0.015506627,-0.3477780223,0.0907885209,-0.3210055232,0.2761228681,0.4049687684,0.1350339204,-0.0484028719,-0.0832800195,0.0696668625,-0.2109884024,0.2506788969,0.2178114951,0.0417517908,-0.3729911745,-0.3129323721,0.4149708152,0.0086601013,-0.1362976134,0.1076715887,0.2048522085,-0.2556139827,0.2249554545,-0.2107982039,1.2595272064,0.2037075609,0.0730994418,0.2657436132,-0.3186558783,0.1554660797,-0.5350136161,-0.0945437402,0.3271141946,0.1984900236,-0.0224854406,0.2317279279,-0.0671636164,0.209056586,-0.4089983106,0.3646213114,0.0560593866,0.1115287617,0.2294220477,0.5157217383,-0.1826923639,-0.3492077589,-0.0984334499,0.2707855105,0.1393754035,0.222013697,0.1090829074,-0.2663983405,0.2003507167,-0.1368193626,-0.6051577926,0.2714419365,-0.2138937712,-0.0572075471,-0.506934166,-0.1218893752,0.2273203135,0.1364028454,0.3495376706,0.0322386324,-0.249032259,0.2839080989,0.0242669825,0.1254819334,0.2574923933,-0.0583834611,0.3492230177,-0.2727707326,-0.1953815967,0.1996145248,-0.0340935551,-0.0502968021,-0.4092877805,-0.2993158698,0.0274301041,0.2117248774,0.1471366733,0.6069585085,-0.1461934298,-0.2696576118,0.1862942129,-0.1487136185,-0.3434027433,0.2365069538,-0.2207871228,-0.3309103549,-0.2700929642,0.3704411685,0.1846064478,-0.0184137821,0.3758647144,0.2097630203,-0.2423903197,-0.3236143589,-0.083728902,0.0862243697,-0.5519273281,-0.080966264,0.0311017353,-0.0119347544,-0.3298691511,0.3230453134,0.0454690345,0.0178418849,-0.1266109198,-0.0471952632,-0.1949789077,-0.0875415429,0.1722123921,0.0628001839,-0.0167923141,0.1922252327,0.0846422613,-0.1631113291,-0.4355678856,-0.0034868217,-0.0785138234,0.1691964269,0.2199091315,-0.0283615068,0.235035032,-0.013018148,0.1801449358,-0.0255129114,-0.0644685328,-0.2282114029,-0.1437916756,0.0878662169,0.1438540369,0.2226475179,0.0158053115,-0.1309378445,-0.2310610563,0.1258000582,0.301975131,0.2664954066,-0.0947366655,0.0670531914,-0.1655734777,0.0737226754,0.1628746539,0.0094173402,0.2201153785,0.4262754023,0.0311132129,0.2439985871,0.0128678288,-0.0977617651,-0.0377023779,0.2497140467,0.1582270116,0.0288494416,0.0082284408,0.0374670215,0.1866350174,0.225908339,0.0351679102,0.0453833863,-0.108349666,-0.1105623841,0.4241489768,0.2845558226,-0.2480105013,0.1591357887,-0.3266995847,-0.2526808977,0.1391421556,0.0563153438,0.0221230369,-0.1091681644,0.3056883514,-0.1232889667,-0.1443874836,0.0170916487,0.0236962382,0.243568927,0.0832450166,-0.0350537561,0.0318379253,0.0566625707,0.0050731804,0.1370354146,-0.2571680844,0.3858582675,0.0796105564,0.1519748718,0.1933548003,-0.1582797766,0.0465597101,0.0729775727,0.0578014404,0.2665905952,0.1312709302,0.1293390691,-0.1678402722,-0.1768575609,-0.2061100751,0.1253010631,0.0634603724,0.0756326988,-0.0524189733,-0.2231829762,-0.2190256864,0.111264877,-0.012349871,-0.2395032793,-0.1248284057,0.1331843585,-0.1199338287,-0.044873558,-0.1700447947,-0.3792723417,0.1237501353,-0.2863716781,0.3407936692,0.1289919019,-0.0486205854,0.1464823931,0.4481001794,0.3994743824,0.1007555276,-0.1546638757,0.2085211426,0.0182339046,-0.1115250811,0.0180642791,0.413802743,0.1149520501,0.0738476068,0.1876872927,0.2296492606,-0.224199295,-0.0325604305,0.120686695,-0.0927176327,-0.094052285,0.2184560299,-0.1428853422,0.1763338447,-0.5103023052,-0.1138624549,-0.5821381807,0.0136061683,0.3610704243,-0.2312156707,0.1926645935,-0.1597992182,0.1138643026,-0.2110678554,0.3706650734,0.3844020367,0.1041263863,-0.0818357393,-0.0960431173,-0.3968750536,-0.1387018263,0.0472263843,0.0885294601,-0.319357127,0.202883333,0.3951250315,0.2992738485,-0.0766491517,0.0462789349,-0.3797616065,0.1133727506,-0.2828350067,0.1392635107,0.0671831965,0.5116134882,0.0102853021,-0.2501457334,-0.1518009752,0.1326688677,0.1015393585,-0.1446295232,-0.0052602072,-0.1481231898,-0.1814876646,0.2580218017,-0.0673453361,0.4292948544,-0.2882859111,-0.1358592212,-0.1584326923,-0.0069513633,-0.1121569127,-0.0225912444,-0.1224219352,0.1981574595,-0.0691536888,0.3408109546,-0.044421237,-0.1471341252,-0.0413158685,0.0786045566,-0.3951094747,0.0934315845,-0.2726104856,0.1314876229,-0.1290509254,0.1855121106,-0.0007183875,-0.1341258734,-0.2878398001,-0.166486457,0.4465268552,-0.1318198591,-0.2964193225,-0.0705263466,0.3546962142,0.1290847212,0.0069307936,-0.5353893042,0.2554196715,0.2577602863,0.0432399884,0.0403939486,0.1596994251,0.0577773936,-0.0230262186,-0.0984330997,-0.0063538644,0.0906568617,-0.2219184339,0.0700239688,-0.4112515748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1843","title":"MustC Speech Translation","comments":"yes that is possible. I couldn't unfortunately complete this PR, If you would like to add it, please feel free to do it. ","body":"## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":23,"text":"MustC Speech Translation\n## Adding a Dataset\r\n- **Name:** *IWSLT19*\r\n- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*\r\n- **Hompage:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation*\r\n- **Data:** *https:\/\/sites.google.com\/view\/iwslt-evaluation-2019\/speech-translation* - all data under \"Allowed Training Data\" and \"Development and Evalutaion Data for TED\/How2\"\r\n- **Motivation:** Important speech dataset\r\n\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nyes that is possible. I couldn't unfortunately complete this PR, If you would like to add it, please feel free to do it. ","embeddings":[-0.3144163489,0.2407132685,-0.0985176414,0.0024260038,-0.0414731726,-0.0363331959,0.086102657,0.1563123465,-0.3312877715,0.1687449664,-0.3144812286,-0.0872284397,-0.0800842121,0.1177140847,0.2429228276,-0.2341304123,0.1259043515,0.0504562743,-0.342484355,-0.2494594455,-0.0213390272,0.0672673211,-0.0788427293,-0.1344392598,-0.1448366344,0.0321697854,-0.279551059,-0.2341867089,-0.3020347953,-0.1368792057,-0.0195759367,0.1097846255,0.1033439115,0.3815741837,-0.0001027024,-0.1275224835,0.2527956963,-0.2727607191,-0.338896513,-0.127855137,-0.1810855865,-0.0817273632,-0.5909472704,0.0052008377,-0.3022561371,-0.2398953736,-0.0072254254,0.0997401029,0.2261944115,0.5919898748,0.2505137622,0.0701196417,0.2464676052,-0.1765385419,0.2097941041,0.1342858821,-0.1943410188,0.0748918205,0.2492660284,0.1058408991,0.134198606,0.3547525406,0.0958012044,-0.1933060735,-0.0836795196,-0.2805369198,-0.086278595,-0.34722808,0.0963443443,0.3862764537,0.5706558228,-0.1490597725,-0.200554058,-0.1098023653,0.3191327751,-0.2181977183,-0.1030058935,0.0713576525,-0.2441433668,0.2761051059,0.1657111645,-0.301453203,-0.1601063162,0.2586252987,0.081465587,0.3450281024,-0.0852068812,-0.1223038882,0.0375883579,-0.2455186844,-0.2540193796,0.0060669743,-0.0461182483,0.0812641382,-0.1872427464,-0.2806805372,0.2276121974,-0.1033081561,0.1497879773,-0.3863224089,-0.1564146429,-0.1269344091,-0.1921891868,0.0440052636,0.1150007024,0.3788427114,-0.0264629908,0.0377745926,0.0474920534,-0.2518242598,0.0497714616,0.1109476462,0.0034480935,0.1197193861,-0.4559198022,0.2956073582,-0.0350663327,-0.1681340039,-0.225004524,0.0103607131,-0.0961563289,0.1125093251,-0.0918740854,0.2835104167,0.1321020275,0.2797350287,-0.039771907,0.3609430492,-0.0224744957,-0.1811030209,-0.0242144447,0.2069827616,-0.2713601291,0.1940347999,0.1961509734,0.4122071564,0.1024266854,0.0247090049,0.2995328009,0.0639061779,-0.2043536454,0.0302976985,0.2913410366,-0.0242092609,0.0846227184,0.0053418945,-0.2197915465,-0.239873752,-0.2191122621,0.0841147974,0.1576313525,0.0178692583,-0.256384939,0.2600449324,-0.0684518516,-0.2423559874,-0.2785376012,0.7449660301,-0.1015027314,-0.1297561824,0.0934245214,0.116459392,-0.4105376899,0.1627673656,-0.0697759315,0.2469461411,-0.1938316971,-0.1282405257,0.1705827564,0.034694761,-0.0142726507,0.1706645787,-0.0773132965,0.2162147462,-0.122164242,0.3464404941,0.0698225126,-0.3657917976,-0.2103001624,0.0861999318,-0.1694951057,-0.0346869007,0.2378326952,-0.1547812819,0.6538490653,-0.2722487748,0.0153084816,0.3843728006,-0.0367776267,0.0969174281,-0.1929135025,-0.3223569989,0.3236677349,0.1244852245,0.0395553038,0.0809627399,-0.047145348,0.3634016514,0.3307479918,-0.3230010569,0.2720314264,0.2625043392,0.0517397225,-0.194038555,0.0653447807,-0.1876687407,-0.0207854155,-0.1414247751,0.1056060493,0.6179709435,0.2154016197,-0.2642193735,-0.1249610037,-0.3051586449,-0.1202379763,-0.2097997665,0.2888029516,0.269425571,0.0702564791,0.0411871038,-0.0988671184,0.2438776493,-0.2970069945,-0.2975799739,-0.0182531644,0.2478787452,0.0196483918,-0.2429094017,0.222289905,0.4088077843,0.1572903693,0.0352284759,-0.0504622944,0.0414014347,-0.2443641573,0.2295623273,0.0951736942,0.1721728295,0.2641621828,-0.4821306169,0.1307530403,0.3708100021,0.0634712726,-0.1844882965,-0.0978765115,0.2177422196,0.4106996953,0.1522424817,0.1553031206,-0.0280253701,0.4273427725,-0.0319923125,-0.2193302065,-0.0738402382,0.0980245024,0.2268742025,-0.1545484811,-0.0578117445,-0.3017581701,0.1953023374,0.5034919381,-0.1534005105,0.2943387926,0.2214704305,-0.0552389473,-0.1172668636,0.0382296667,-0.2572184503,0.2364936769,0.3023200035,0.3069224954,0.0311732721,0.1960681826,-0.0299286172,0.1378580779,-0.0343159549,-0.2566353083,0.1840289682,0.2056990862,-0.2389389127,-0.3877452612,0.0946670026,0.0309749376,0.1088092327,-0.1727689803,-0.3300414383,0.0964136347,-0.2699221969,-0.2532964051,0.0454060212,-0.1398635656,0.0259383526,0.4245880544,0.0558309294,-0.1760515571,0.283551842,0.340562731,0.2605621219,-0.4604918957,0.0429377183,-0.1058837324,-0.3384135664,-0.15018031,0.2364138812,0.1870855689,0.2369820625,0.1945245713,-0.1454205811,0.0693440735,-0.2870320678,-0.2716109455,0.2488449812,-0.2595183551,0.0837701261,-0.2471225113,-0.0334548242,0.0305460766,-0.434556067,-0.2007727325,0.0139663061,-0.1261313856,-0.1129611209,-0.4465985298,-0.1231418923,-0.1197625771,-0.4519715607,-0.5385044813,-0.4178848267,0.3924363852,-0.196870178,-0.0586535782,-0.0799917206,-0.0288544111,0.0384505503,-0.0433630012,0.1207184419,-0.3419520557,-0.1426251978,0.0722502545,-0.3547128439,-0.3084486723,-0.2175134122,0.025479449,0.1171075404,-0.190325439,-0.1839789003,0.0840391964,-0.0518385321,0.343990624,0.0942208096,0.0520406999,0.2340598702,-0.1334278435,-0.1558547318,-0.1684498638,-0.147410199,0.0870958269,0.4020906389,0.2011591047,-0.0141405277,-0.0152146639,-0.0106174378,0.5695930123,0.2129810303,0.0032868332,0.1964784861,-0.0691174045,0.4202748239,-0.3030357659,-0.2824953496,0.4688383937,0.0441655181,-0.0698982924,0.4359495044,0.0320686102,-0.4763175547,-0.0940274224,-0.2504986823,-0.2837279737,-0.2808144987,0.300473541,0.1049331576,0.1418058574,0.0175237451,-0.1853658557,-0.2366909534,-0.1386936605,0.3092558086,0.095227845,0.1058480442,-0.010956482,-0.238320455,-0.0322098918,-0.3911242783,-0.1759310812,0.1685509831,0.2604579031,-0.4749266803,0.1247686669,-0.0968390778,0.1260519177,0.3391442001,-0.3525826633,-0.031048445,0.3051852286,0.4792062342,0.0872476175,0.1870060563,-0.2485233843,-0.3464483321,0.628909409,0.3033208847,-0.3392402828,0.0734625831,0.1714071035,0.3540358245,-0.056714341,-0.1258448809,0.1036254466,-0.3130624592,-0.3513386548,0.2941850126,0.5863972902,-0.0044181701,0.1327321529,0.0856683776,0.1573927104,-0.0330057591,0.4120715857,-0.0565564781,0.0049822889,0.5083364844,0.0741575137,-0.0585307926,-0.0996992439,-0.136097312,0.2360038012,-0.0294902679,-0.3625955582,0.0410941243,-0.3179109991,0.341019839,0.3419692814,0.1969083399,0.0554200783,-0.1151663885,-0.0049267118,-0.1786254346,0.1787164211,0.2119587362,0.2197776586,-0.2665579915,-0.2907713652,0.3327901661,0.0238480493,-0.1356429309,0.0800258443,0.1426535547,-0.2639533281,0.302380532,-0.2055172324,1.2013126612,0.2142941207,0.0708872676,0.1853222698,-0.2761240005,0.0556959771,-0.4282705188,-0.1648578346,0.2509607971,0.1383715272,0.0014110631,0.2332966477,-0.0927015394,0.1877402216,-0.4681071937,0.2074917406,-0.0331504233,0.0549174137,0.1473742723,0.4590066373,-0.1414240599,-0.4073213637,-0.3770550787,0.2764765322,0.1617033631,0.1001480743,0.1867157221,-0.3114002943,0.1649118364,-0.0229208004,-0.6706446409,0.1773018688,-0.1828348935,-0.1868105084,-0.4717199802,-0.1523205191,0.235088557,0.2190215588,0.406921953,-0.0199954137,-0.2431349158,0.2917390466,0.023937162,0.1518756449,0.1312960237,-0.0602537841,0.3363066912,-0.3078973591,-0.2724235356,0.2259281576,0.0713616163,-0.0757800564,-0.3984653354,-0.2754771113,0.0963084698,0.1583934426,0.2444709688,0.5786366463,-0.0220875759,-0.3135793209,0.2143933177,-0.1159983054,-0.4537069499,0.1347676367,-0.1388079226,-0.242162928,-0.1848852187,0.2223583609,0.2763391137,0.0056740148,0.3082564473,0.250069201,-0.1052585989,-0.3605339825,-0.1390833557,0.18911542,-0.4564025998,-0.0212939195,0.1425506622,-0.1546455324,-0.2860095799,0.2714068592,0.0331897177,0.0567180738,-0.1672994643,-0.0124395518,-0.2301465869,-0.0923010185,0.1138135642,0.0898480788,-0.0456600338,0.3371570706,0.0558027588,-0.0954171345,-0.4461348355,0.0046957848,-0.1053118855,0.2245199531,0.2490898669,-0.1189706773,0.0917533338,0.0116871027,0.1757113636,-0.0036669101,-0.0317338444,-0.2695858479,-0.1937435269,0.0764046237,0.1505878717,0.2530787885,-0.0092549436,-0.0582828447,-0.1696678251,0.0840196908,0.334613651,0.2620305717,-0.0697772205,0.1393532753,-0.0488858148,0.0339894108,0.2634316683,-0.0036815463,0.3496932983,0.2920739651,0.041692365,0.154003635,0.0411523767,-0.0830360204,-0.0032431912,0.230028227,0.1934446543,0.1171964481,-0.0611629561,0.1571452022,0.1814698726,0.2435405254,0.0742650107,0.0299630743,-0.0770980865,-0.0151863424,0.3908577561,0.3026229441,-0.2279476076,0.1720368117,-0.202579543,-0.329187721,0.1894780844,0.0683883354,0.0471173115,-0.0691949055,0.3555964231,-0.122952953,-0.275238812,0.0246599372,0.0792317241,0.1658139825,0.0947474167,-0.0498004444,0.0218524896,-0.0467494614,0.0743333548,0.0160367433,-0.1718967855,0.4392328262,0.129554823,0.1724376082,0.2956699431,0.081551522,0.0978177413,0.0274230558,0.1782457978,0.2709984183,0.1174812466,0.2897264063,-0.0029127821,-0.3623558581,-0.1611575186,0.2655168474,0.0546869487,-0.023885319,-0.0480707735,-0.1641686559,-0.2151495367,-0.0387396403,-0.0059952447,-0.1757559627,-0.1309611797,0.1293317825,-0.1634711474,-0.0372238494,-0.1534476876,-0.2295427322,0.1345555335,-0.2233674675,0.3498359621,0.1768841743,-0.0774556175,0.2122640312,0.5102564096,0.4451505244,-0.0017901406,-0.1606406868,0.1599898636,-0.0294546131,-0.0854602829,0.0235630628,0.3419022858,0.1673723757,0.0670711547,0.1742616594,0.2570086718,-0.2493707985,-0.1158064753,0.1314186901,-0.0973993912,-0.0916336402,0.1299996674,-0.0769674256,0.1917337328,-0.6140168905,-0.2166836709,-0.5189560652,0.100844875,0.3167093694,-0.2451290339,0.0856406689,-0.0976706967,0.1314334869,-0.1694405675,0.2627632022,0.2098893374,0.0779373869,-0.0809132978,-0.1441600919,-0.3445735276,-0.0600717887,0.116244629,0.0893861353,-0.3636893034,0.0120038241,0.4930253625,0.3169737458,-0.1438654065,-0.1319644302,-0.3537128866,0.1792702377,-0.4054432213,0.2229891717,0.067499049,0.6136319041,0.1962837428,-0.1255868971,-0.2397262752,0.2737295926,0.1132913008,-0.0416438878,-0.0516675226,-0.2366295904,-0.2126127928,0.1456016749,-0.0530781634,0.4831128418,-0.3111390471,-0.0611145161,-0.0497695915,-0.1624682695,-0.1949472278,0.0023771145,-0.0452821143,0.2194710523,-0.1070498377,0.2709193826,-0.040550258,-0.1780353636,-0.0465668328,0.0609510019,-0.292506367,0.1395405233,-0.1016311646,0.0202333834,-0.1936718673,0.200961709,-0.0456454195,-0.0938913226,-0.2283538729,-0.186920777,0.4489099085,-0.1410411298,-0.3384423554,-0.0648675635,0.3041795194,0.0984933525,-0.067930527,-0.6517695189,0.2181300223,0.2331311256,0.1671431065,0.1549941599,0.1583144218,0.0599935949,0.0549564585,-0.1076406389,-0.0592298768,0.1667318642,-0.1569559574,0.0428398289,-0.3223870397]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1840","title":"Add common voice","comments":"I have started working on adding this dataset.","body":"## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":8,"text":"Add common voice\n## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nI have started working on adding this dataset.","embeddings":[-0.1986446232,-0.2702407241,-0.0677535012,-0.1871911287,0.1138091013,-0.1876987666,0.4035984576,0.3044218123,-0.2726596296,0.32479918,-0.3224400878,0.1522413641,-0.0523534827,-0.030831065,0.0226880852,-0.1701612622,-0.1684961915,0.1530555338,0.0794557109,-0.1830599755,0.2358211875,0.0914489031,0.0094152447,-0.0615204796,-0.1931130737,-0.0603697971,0.1552869529,-0.1437284052,0.0328677632,-0.0320457146,-0.1293242872,0.2172251642,-0.2654833198,0.5041013956,-0.0000981533,0.022254372,0.1077704653,-0.1566758305,-0.0638837293,-0.0269621592,-0.1120303795,0.1518278122,-0.3239215314,0.1463613659,-0.3190734982,-0.0583319589,0.0007174523,-0.1766736209,0.3651863337,0.3695534468,0.3072449863,0.1790877581,-0.1339848489,0.0721334368,0.1872765124,0.1243255436,-0.3021790683,0.0029829252,0.2430054396,-0.1207270771,-0.1430469155,0.4142365754,0.0865239054,-0.0811778754,-0.0860068426,-0.0295925252,-0.1644873023,-0.3714516163,0.0669073611,0.5412355661,0.4696133733,-0.3007895052,-0.1978662312,-0.2661865056,0.0936901495,-0.1324830502,0.0678513497,-0.0928892195,-0.2862649262,0.2211689502,0.1102420986,-0.0409234129,-0.078577742,0.0838393569,0.0890986323,0.3197731376,-0.0604909435,-0.3318625093,0.0152510842,-0.1988622248,-0.2641705275,0.1627449095,0.3235060573,0.0671297088,-0.3810548484,-0.1805519611,0.1859658211,-0.5511543751,0.2714396417,-0.2165005654,0.2525296807,0.2455227524,-0.291651547,0.1338457763,-0.0461610183,0.0690868199,0.0477513783,-0.2260652184,0.3325885832,-0.0611184165,-0.0046978258,0.1133923754,-0.0015039851,0.0151012903,-0.165983513,-0.0122104688,0.1386322379,-0.0024195651,-0.3064165413,0.0506103113,-0.1107279435,0.0035154615,0.0907097012,0.2134263963,-0.0354995318,0.3311030269,-0.0420792438,0.1612447649,-0.0549619272,-0.3290218115,-0.1683806926,-0.1273577213,-0.078060165,0.2068199217,0.2919307351,-0.0178703647,0.1809429973,0.0783001706,0.2682179809,0.0682879314,-0.018546151,0.1772682816,-0.0538627356,-0.0121577196,0.0324794874,0.0397860967,-0.057535924,0.0445247702,-0.1097576693,-0.1701645255,0.125000149,-0.1337899715,-0.3408316672,0.3576396406,0.1999698132,-0.0961622968,0.2229352146,0.4949534535,0.1731401086,-0.2081322521,0.133040458,0.1614362746,-0.3832354248,-0.0063836686,0.1800701618,-0.0042474489,-0.1995174885,-0.1080283225,-0.0389732942,0.0375897475,-0.0288509037,0.3656671345,-0.0154756922,0.0062895697,-0.0614280291,0.4665953815,0.356174469,-0.2734332681,-0.3905410469,0.1536207795,-0.2753944397,0.0921524987,0.2471610457,-0.0947505012,0.2111154646,0.00800661,0.4093611836,0.3448544443,-0.1737828255,0.1259950697,-0.2442999631,-0.1821595877,-0.0254446678,0.1774439365,-0.1739993095,-0.0782415718,-0.0902987272,-0.0389682055,0.2559513748,-0.327291429,0.1544577628,0.1271404475,0.4471979439,0.0064489376,0.0714422241,-0.2218536586,-0.0737177581,-0.0928288996,0.0634143427,0.2532556057,0.0101811318,-0.5207300186,-0.1401411593,-0.2908411026,-0.0363825113,0.1550834626,0.3599822819,0.2726039886,-0.0636665523,0.0162695125,-0.1563930213,0.1808231473,-0.0696744621,-0.0514257923,0.0654834732,0.1441056728,-0.2052323818,-0.138814643,0.2684591115,0.2881702781,-0.2564217448,-0.0723955408,0.0023945845,0.2414828986,-0.335498482,0.3797133565,-0.0582440756,0.3710648119,0.1536706835,-0.2068423927,0.0197438654,0.1568196118,-0.0105708875,-0.0976667777,0.2037792653,0.1170980707,0.5079960823,-0.2260580957,0.1145115495,-0.1471326053,0.3420547545,-0.0279571563,-0.1536441743,-0.0651111454,0.0454031378,-0.0782862902,0.0857338831,-0.0309938006,-0.6059052348,0.2662473619,0.6899664998,-0.1454766244,0.4327820837,0.1966296583,-0.2462561131,-0.1177844629,0.0513627231,-0.2119968683,0.0690456927,0.2764208913,0.221496582,-0.0616012178,-0.0787716582,-0.1925234646,0.0888303146,-0.0464046672,-0.0894131735,0.4977206588,0.2536010146,-0.0026640641,-0.6556220651,-0.2100590914,-0.0875944346,-0.0352637544,-0.2188842893,-0.1436338127,-0.1543637216,-0.2314421982,0.1013410836,0.0808103159,-0.107625857,-0.0312710293,0.4576940835,0.3972775638,-0.0600466728,0.3516437709,0.0657707453,0.3705534935,0.1458188891,0.0060868915,-0.0277243555,-0.0786625296,-0.1288496852,0.2931647301,0.1201775223,-0.1914924234,0.2134720981,0.0174790435,0.1495841146,-0.0541008189,-0.427385509,0.2271090895,-0.3181462884,-0.222731173,-0.0933929831,-0.0137206698,0.0670683905,-0.4379041493,-0.0320000313,-0.2286294103,-0.2779521346,0.0100521296,-0.1456549317,-0.0653946474,-0.3305634856,-0.4415305257,-0.0637457594,-0.5374345779,0.1363499761,-0.3388372958,0.0143812345,0.2015039027,-0.185856238,0.0973069668,-0.227938056,0.4542818367,-0.3658855557,-0.225672394,0.2258993387,-0.2975204587,-0.352766484,-0.2365264893,-0.1538805515,0.3741814792,-0.0311604962,-0.1700322777,-0.2505283952,-0.1093294248,0.2806351483,-0.0224856883,-0.0021423257,0.1296777129,0.2084167451,-0.2104384154,-0.1809736192,-0.0841122046,0.1006345302,0.2589063942,-0.0100909835,0.0132192066,0.1765241176,0.1539761126,0.789229095,-0.0059013432,-0.2341099828,0.2911610603,-0.2184878439,0.1637604535,0.0053008217,-0.0976832062,0.4323175251,0.1965673864,0.2562316656,0.5622221828,-0.0096276999,-0.179913193,0.065154843,0.0420621634,-0.2475855052,-0.3041238189,0.1064371988,-0.0491365641,0.1149724796,-0.0001915427,-0.038714353,-0.2012141496,-0.0219379533,0.3649641573,0.051542718,-0.0984420031,-0.2523604631,0.0539505333,0.1097136363,-0.3693659902,0.0538857467,0.2484938204,0.182403788,-0.339750886,-0.1791252494,-0.1785593033,0.1500855833,0.3159578145,-0.1979112029,-0.1752618253,0.0551898293,0.1567425132,-0.0036085737,0.3067418337,-0.193928346,-0.2028653622,0.1090112478,0.1153787822,-0.3073253334,-0.0440447517,-0.1657929868,0.185974285,0.0039204177,0.0494512618,0.0478565283,-0.2067334503,-0.2256194055,0.0514887869,0.1338160336,0.0417503268,-0.0129852584,-0.1772919595,0.3160912097,-0.2737542093,0.380204916,0.3245734274,-0.1092958376,0.2548288107,0.2012137324,0.056662444,0.1410678476,-0.0507707894,0.0920661986,0.1536083519,-0.239683643,-0.1142126396,-0.1550845951,0.2741810679,0.0659557134,0.1782815009,0.113354072,-0.084908016,0.1158984974,-0.0813614354,0.316019237,0.3757628798,0.2995266318,-0.113349922,-0.2024073005,0.1090324596,-0.0434034206,-0.1490422338,0.3395322263,-0.0510978214,-0.3260756433,0.0618981123,-0.3408779204,0.9444889426,0.1075059995,0.2077683359,0.416148603,0.2436181158,0.1349065304,-0.4964186847,0.1885138005,0.1153620407,-0.3188168705,0.0612446256,0.1844537705,-0.243292287,0.0726417676,-0.3232880235,-0.0846667215,0.043294549,0.0671275631,0.1072074547,0.2903809249,-0.0484329611,-0.441196233,-0.2806527913,0.242704913,0.0223207213,0.0864957795,0.2192836851,-0.2423390746,-0.1092183292,-0.0618397072,-0.2218150198,0.2551082373,-0.2084823251,-0.0709982291,-0.0052872458,0.0394224524,0.2425567955,0.078312695,0.3367683589,0.061300572,-0.2109241188,0.1508344859,0.237321049,0.1526782215,-0.1557234973,0.1298008561,0.3023453057,-0.2681306005,-0.1894078255,0.113468647,0.2168308645,-0.1409569979,-0.3187899888,-0.3196696043,0.0544774421,-0.1211361438,0.2912330925,0.2848578393,-0.1772722751,-0.1538570225,0.275531441,0.1042161509,-0.3494126797,0.0549425036,0.2183156163,0.014872524,-0.1303351372,0.1448873281,0.154302448,0.0253791902,0.234602496,0.0425663441,-0.1490213722,-0.3849725127,-0.3094070256,-0.0152441841,-0.1408680975,-0.2013738602,-0.0094971862,-0.4245219827,-0.0877138078,0.2024510056,-0.0327361673,0.1659881324,-0.1063209847,0.0142313587,-0.3837935328,0.0543099865,-0.0533021875,0.2428237796,-0.1427143961,0.2455167621,-0.0188259333,0.0577208549,-0.5296253562,-0.2513979077,0.0071153119,0.1486109942,0.2413728684,-0.3450950682,0.294817239,0.0338660516,0.3787585497,0.0650289059,-0.0579790808,-0.4010250568,-0.1811364293,0.064716652,0.203830868,0.0057912897,0.2677342296,0.1573988944,-0.2622781396,-0.0850755721,0.3288598657,0.0510977097,0.0240494758,0.1057930738,-0.0530084707,-0.005369016,0.2640350163,-0.2087622583,0.0935154632,0.0835146084,0.0521282256,-0.0076958709,-0.0747618377,-0.2504503727,-0.0650152564,0.1770306081,0.2479649484,-0.0052749487,0.101579465,0.061307352,-0.195863992,0.2379144728,0.3398456573,-0.1045615822,-0.172892198,-0.0666774362,0.234355405,0.4177899361,-0.2740674019,0.0067332289,0.0986415818,-0.1748791337,0.4062717259,0.107194066,-0.09053725,0.0767719746,0.101875253,-0.0862646922,-0.206687212,-0.1596663147,0.0178166572,0.0658676103,-0.0343157388,-0.0042668502,0.0843939632,-0.08950717,0.3162606657,0.2064748555,-0.1792734563,0.264706403,0.1010584161,-0.0617927238,0.4885054827,0.0303434357,0.0707060248,0.0684073493,-0.112390697,0.219506681,0.1003487855,0.312916249,-0.116333209,-0.157130897,-0.0236071553,0.3051162064,-0.077370353,-0.1479397416,-0.1200211719,0.0135528855,-0.0997512341,-0.2238639742,0.1862945706,0.1243922785,0.0814346969,-0.0085232574,-0.0211427454,-0.0196849704,-0.0057882192,0.0427259691,0.4382206202,0.0148924626,0.2688478231,-0.0355735458,-0.3004412949,0.1228511035,0.3547056913,0.1774179488,-0.1004418433,-0.19936046,0.3946743608,-0.1203175932,-0.0469407365,-0.1672067344,0.1334955245,0.3184208572,0.0579741895,0.1276078671,0.2662326992,-0.3003811538,-0.1237038225,0.0458786152,-0.1335057318,-0.1529314369,0.1256733388,-0.104824163,-0.0016187653,-0.3503442407,-0.1668829769,-0.5072212219,-0.1520282626,0.4344621301,-0.0321353264,0.1094205379,-0.3360773325,0.1648523659,-0.0356584601,0.2219433486,0.2604973018,0.045208022,0.3654291034,0.0075230128,-0.3636440635,0.3637878597,0.2044038922,-0.0065124524,-0.3167701662,0.0376092866,0.3980906904,0.1035405621,0.1184511632,-0.359618187,-0.2525045276,-0.2596485317,-0.2781997919,0.0994920582,-0.2290146798,0.1559567004,0.2564797401,-0.3086348772,0.0824337527,-0.1679216623,0.287545383,-0.1370633692,0.1252716929,-0.1213577688,-0.2146490067,0.1636648476,-0.0279508699,0.2212213725,-0.3761177063,-0.1884293556,0.0206133202,-0.1844237596,-0.2672770917,-0.0323386416,0.081849061,0.3988796473,0.0255798083,-0.0539650768,-0.1137632355,0.1991142035,0.1898878962,-0.1433342099,-0.0208879374,0.2544485033,-0.0044438676,-0.0148546519,0.0835781246,0.2099679708,-0.1756470352,-0.0621530563,-0.0790978968,-0.5515398979,0.3363562524,-0.268052727,-0.2764106393,-0.2194882482,0.3106554449,0.0377374515,-0.0343369469,-0.4072751105,0.1057876199,0.2119914889,-0.0092453882,-0.1334059238,0.1833200902,0.1225710288,-0.2605471909,-0.1302891672,0.2813132107,0.0968771279,-0.0874428079,0.1092066094,-0.3661887646]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1840","title":"Add common voice","comments":"Hey @BirgerMoell - awesome that you started working on Common Voice. Common Voice is a bit special since, there is no direct download link to download the data. In these cases we usually consider two options:\r\n\r\n1) Find a hacky solution to extract the download link somehow from the XLM tree of the website \r\n2) If this doesn't work we force the user to download the data himself and add a `\"data_dir\"` as an input parameter. E.g. you can take a look at how it is done for [this](https:\/\/github.com\/huggingface\/datasets\/blob\/66f2a7eece98d2778bd22bb5034cb7c2376032d4\/datasets\/arxiv_dataset\/arxiv_dataset.py#L66) \r\n\r\nAlso the documentation here: https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html?highlight=data_dir#downloading-data-files-and-organizing-splits (especially the \"note\") might be helpful.","body":"## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":100,"text":"Add common voice\n## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHey @BirgerMoell - awesome that you started working on Common Voice. Common Voice is a bit special since, there is no direct download link to download the data. In these cases we usually consider two options:\r\n\r\n1) Find a hacky solution to extract the download link somehow from the XLM tree of the website \r\n2) If this doesn't work we force the user to download the data himself and add a `\"data_dir\"` as an input parameter. E.g. you can take a look at how it is done for [this](https:\/\/github.com\/huggingface\/datasets\/blob\/66f2a7eece98d2778bd22bb5034cb7c2376032d4\/datasets\/arxiv_dataset\/arxiv_dataset.py#L66) \r\n\r\nAlso the documentation here: https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html?highlight=data_dir#downloading-data-files-and-organizing-splits (especially the \"note\") might be helpful.","embeddings":[-0.1914141327,-0.2330227643,-0.0579186045,-0.1640419066,0.1464050263,-0.077341415,0.2238348275,0.2807310224,-0.2090623528,0.3071741164,-0.4487543404,0.0414794385,-0.0504973307,-0.061810866,0.1337499022,-0.1799681336,-0.2989138067,0.0928419083,0.0731951296,-0.089301683,0.1916294396,0.0703539327,0.006779355,0.010302417,-0.0648635328,-0.0836325213,0.0658920258,-0.0585368574,0.0435861982,-0.1326310337,-0.1019375101,0.2207966596,-0.1319609284,0.4909032583,-0.0000958251,-0.0298786294,0.1239590496,-0.171576485,-0.1875044107,-0.1483193487,-0.2629153728,0.1999739707,-0.3399197459,0.1386548281,-0.2324215323,-0.0841777921,0.0835847184,-0.173622787,0.367010653,0.4226413667,0.321483463,0.1754924357,-0.068599917,0.0331244133,0.1989008486,0.1055693999,-0.186442107,0.1450786144,0.1981421858,-0.0047167116,-0.0269427747,0.3730793893,0.0401432291,-0.110422723,-0.1406654716,0.0541818626,-0.0711617246,-0.4051722288,0.0466371551,0.4960680008,0.5813153386,-0.2943419814,-0.1319930851,-0.2549909055,0.0939375609,-0.0744588748,-0.0119608771,0.1040292308,-0.3731024861,0.2526135147,0.1400448531,-0.1729530841,-0.0926651284,0.1806473434,0.0982044414,0.2873306274,-0.0902398825,-0.3477494717,0.035027843,-0.129786849,-0.2782455981,0.1741712093,0.3745585084,0.0719093382,-0.2025437951,-0.2045180351,0.1722052991,-0.4943517745,0.3373634219,-0.2338496894,0.3088485599,0.1937015057,-0.2733047903,0.1480517685,-0.0929492861,0.1415398717,0.0336594768,-0.2610833943,0.47900635,0.0796311498,0.1017048433,0.0869898796,-0.0355172679,0.0955113247,-0.1683771014,0.008145717,0.1157443076,0.1286963373,-0.2746973634,-0.0248591751,-0.0774223804,0.0341903716,0.1007621586,0.3143828213,-0.1674354374,0.2423110604,0.012937298,0.1006355435,-0.0871869996,-0.4383678138,-0.1132012904,-0.0901504755,-0.0422652811,0.2203017473,0.2512874901,0.0081313821,0.1794687361,-0.0383709297,0.3635953367,-0.0449105389,-0.0580287762,0.1659001857,-0.1160877869,-0.0063175331,0.1776334792,0.1012066081,-0.0947201923,-0.0506712049,-0.1521701962,-0.0814013034,0.0066563096,-0.2400241345,-0.234915942,0.3818749487,0.2067500353,-0.091620408,0.1750999093,0.4059943259,0.0205771737,-0.1133105829,0.0576032996,0.2704983652,-0.3937545717,0.0019520298,0.2353151143,0.1386979967,-0.1030028164,-0.2127184272,-0.0887004435,0.0259471405,-0.0899717435,0.2812158465,-0.0547237471,0.1473571807,-0.0783701465,0.5168426037,0.3876639605,-0.3474629521,-0.3929833472,0.1568109095,-0.3701220155,-0.0025552269,0.1129033267,-0.0325602703,0.2321348339,0.0054643154,0.3086158633,0.2736943364,-0.0658044145,0.2518997788,-0.2528394759,-0.2755353451,-0.1130043194,0.2201479971,-0.2039204538,-0.0475509055,-0.021995414,0.0114409169,0.3305029869,-0.2774728835,0.2274192274,0.1372638792,0.3657888472,0.0801173374,0.0349297076,-0.1751105189,-0.0472088046,-0.0706851706,-0.0469404459,0.2627592683,-0.1196191162,-0.5422856212,-0.3234484494,-0.2906109691,-0.0741249472,0.0678827167,0.3533306718,0.2043393552,0.1240557283,0.090327315,-0.1320174336,0.1620310396,-0.0152115356,0.0198962148,0.0743757933,0.0913552046,-0.2356636822,0.0062168064,0.2639211118,0.1740455031,-0.1094267294,-0.0470192805,0.0121951774,0.4601745903,-0.3491534591,0.4136956036,-0.0707971677,0.2818537951,0.2362557054,-0.2704651058,0.0932089537,0.237088114,0.030820692,-0.0098860031,0.194771111,0.020402411,0.4368931055,-0.1829052567,0.2546635568,-0.0160999708,0.3013763726,-0.0786268562,-0.1048425436,-0.0826984495,-0.027450094,-0.0776250064,0.1535747349,-0.1632675081,-0.5607640147,0.2125445753,0.5365672708,-0.1590388864,0.3975815475,0.2310604602,-0.3788747787,-0.0340253115,0.0417590626,-0.1818191707,0.0579486452,0.3344874978,0.2248823792,0.0508948825,-0.0263416916,-0.2014659196,0.1271747053,-0.0536994971,-0.2016319036,0.3989920616,0.1725071222,-0.0104918908,-0.5637577176,-0.1607815325,-0.1286462247,-0.098063007,-0.2647343278,-0.1571461707,-0.1429148316,-0.4420362115,-0.1284156144,0.0389535166,-0.130406186,-0.1147743762,0.4792996347,0.3162687719,-0.0680870786,0.2069611549,0.0379830599,0.3899508417,0.0894132927,-0.0279653538,0.0022464218,-0.0561508201,-0.168734163,0.2636246085,0.0540253855,-0.1238390505,0.2293983698,-0.0452980548,0.1493048817,-0.0334438048,-0.3091451824,0.3005314171,-0.1655344516,-0.3047591448,-0.120697327,0.0727315396,0.1589967012,-0.3374803662,-0.0105486307,-0.1580181122,-0.2190054506,-0.0341279544,0.022927044,-0.0131090013,-0.3320544958,-0.4506056309,-0.2056814879,-0.5207069516,0.1198896468,-0.2591120899,0.1161835939,0.1122034118,-0.2233251482,0.1942960471,-0.1867551208,0.4712706208,-0.3216316998,-0.186349526,0.3625208139,-0.368829459,-0.3724723458,-0.1673838049,-0.0606763586,0.3532105982,-0.0775346234,-0.2152739614,-0.1116482764,-0.0815466046,0.1689911634,-0.0473101176,0.0326159559,0.1993018538,0.1591008902,-0.2522543669,-0.1729519963,0.0519597717,0.0556859188,0.2103472948,0.0103447344,0.0118938144,0.0778008848,0.2278718203,0.7560041547,0.0810778067,-0.0760716572,0.1922338158,-0.2419106513,0.0715439692,0.0979513451,-0.0580223612,0.3411958516,0.135403201,0.1990298182,0.5835257173,0.1346820444,-0.1079585329,-0.0935439467,0.0194064919,-0.3015474677,-0.4096614122,0.0197023153,-0.0651295781,0.1474786699,-0.0145646036,-0.0210399088,-0.0140760057,-0.0204695016,0.4538123906,0.1323343217,-0.186484918,-0.1799759269,0.2131440789,0.1475814581,-0.3176474273,0.1330755204,0.2769827247,0.0023005593,-0.2478591055,-0.2487751245,-0.1171029806,0.131393984,0.2856900096,-0.139919728,-0.1206069961,-0.1598569304,0.2348412573,-0.0035144207,0.3497113585,-0.1370619386,-0.3063135445,0.126397714,-0.0062396987,-0.32530725,-0.0157699827,-0.1843082458,0.0775360316,-0.0926557258,-0.0158221927,-0.004391633,-0.2653481662,-0.2305484563,0.0171839781,0.086743921,0.0055582412,-0.0511264764,-0.1443826407,0.4201830029,-0.2546163499,0.3220595121,0.2204236984,-0.0859737545,0.248171851,0.07314834,0.1424105912,0.1460678428,0.0163868256,0.148281306,0.2338846475,-0.2542985976,-0.2496322542,-0.1874730736,0.1962057352,0.1179956272,0.1918427646,0.0696486756,0.0070575876,0.1785699427,-0.1386413276,0.2552043796,0.286757797,0.1282340139,-0.0800250992,-0.2555853128,0.1926199943,-0.1053228229,-0.1015549973,0.2771307528,0.0240957886,-0.3037805557,-0.0706029609,-0.2882122099,1.0026525259,0.1183435619,0.1908113062,0.3410577476,0.0726653934,0.1709722728,-0.4545092285,0.2086755037,0.1632049829,-0.1537919343,-0.054654371,0.1420119405,-0.2863587141,0.126838997,-0.3761081994,-0.061679069,-0.0214050878,0.0627399012,0.1325074285,0.4239200354,-0.1510881186,-0.4788647294,-0.3154270053,0.2642374933,-0.086728625,0.2362656444,0.2626686692,-0.2529901564,-0.1149283797,-0.0637134686,-0.1541435868,0.2917878032,-0.292461127,-0.0493436381,-0.2072124928,0.0998129025,0.2420376837,0.1741273999,0.2320064604,0.1393321604,-0.306749016,0.1588183045,0.1939471066,0.1440843642,-0.0995002389,0.0591383241,0.3088829219,-0.2716981471,-0.1698353738,0.2394302785,0.257065922,-0.0698952824,-0.2673352659,-0.3158339262,-0.0035010958,-0.0966454521,0.3100909591,0.3235730529,-0.1549088806,-0.1027553827,0.2644545734,0.0209740475,-0.2815410197,-0.1108626425,0.2335283756,0.0327757411,-0.150126487,0.2513998747,0.0610241517,-0.0098823234,0.2471794784,0.0241858196,-0.1909644455,-0.3668530285,-0.2549306154,-0.2306808382,-0.1314147562,-0.1306543052,-0.1345628053,-0.2067371905,-0.2458793074,0.1564690471,-0.1442661136,0.1952605695,-0.0851778239,0.0032714936,-0.3673198819,0.0618113838,0.1946365684,0.1336250454,-0.1124667376,0.3103489876,0.0774402171,0.0023243192,-0.52476722,-0.2034437656,0.0491655432,-0.0354599245,0.3234440982,-0.1927637756,0.4031952918,0.012477641,0.325809896,0.0486925542,-0.077448003,-0.37932235,-0.2583362162,0.0795325413,0.2377000451,-0.0187286399,0.0788434148,0.0648451522,-0.1420834661,-0.0164444372,0.3288491666,0.0724922493,0.0079336148,0.0902178138,-0.0863252059,0.0552414879,0.0965061858,-0.1942552924,-0.0503233746,0.2289917916,0.1543940753,0.0242106616,-0.1239215732,-0.2402877212,-0.1021340787,0.1974298656,0.3567436337,-0.0804348364,0.232996285,0.0097967414,-0.1402047724,0.129304409,0.3861314952,-0.0911689177,-0.2412335724,-0.1275300533,0.253695935,0.4083325267,-0.2138858885,0.1225509718,-0.060386613,-0.3291770816,0.3680534363,0.1795025319,-0.130282104,0.2072899342,0.0437912196,-0.0316705666,-0.121708408,-0.117145732,0.0613425039,0.1134322882,-0.0481743999,0.0056476695,0.0096609518,-0.0278044771,0.3023861349,0.0677325428,-0.3583764434,0.2018923014,0.0399664752,-0.0862172395,0.4024357796,0.0703883618,-0.0735239759,0.0153384386,-0.1396112442,0.2165043503,0.1023059487,0.4589159191,-0.1830164641,-0.0195774566,0.0281379782,0.3724358678,0.0333435163,-0.0442899391,-0.2614049315,-0.047341872,-0.2303518802,-0.1636780947,0.2782407105,-0.0706574023,0.0411744602,0.0363365375,-0.0496931933,-0.1757365465,-0.0177373011,0.0439616963,0.4665777087,-0.0818017274,0.1076580361,0.0078297136,-0.2480560392,0.1456405967,0.2919992208,0.0938511416,-0.0734765232,-0.2089615464,0.3558445275,-0.0780873448,-0.0767829046,-0.110415861,0.0555465706,0.2791467607,0.1199438721,0.211884737,0.2674845457,-0.3482512832,-0.1066893265,0.0184003543,-0.0197681542,-0.0893204808,0.1595226973,-0.1001286656,0.0307629276,-0.3655953109,-0.2546173334,-0.5871138573,-0.0298457667,0.3238992393,-0.1353872418,0.139384523,-0.3521345258,0.1749688834,0.0543313064,0.2641286552,0.2829347253,0.0492221378,0.3644104004,-0.0616458245,-0.4385338426,0.259562552,0.2388585806,-0.0972764641,-0.3690240085,0.0233482197,0.2873706818,0.1467652917,-0.0101037445,-0.2942483723,-0.2715741992,-0.215579927,-0.1658015102,0.1621247232,-0.3059232831,0.2832655013,0.2129464149,-0.3117226064,0.0671119615,-0.0784238875,0.2411318272,-0.0460233204,0.179601863,-0.1786737144,-0.3356564343,0.2401758879,-0.1234487891,0.2430974692,-0.395334065,-0.2057670802,0.1390499771,-0.0050312448,-0.2132121027,-0.0577723235,0.1126819849,0.3333126307,0.0756597519,-0.0800971091,-0.2306825668,0.167631343,0.1834653765,-0.1587756574,-0.0100846458,0.0971832797,-0.0229349174,-0.0381478034,0.086287722,0.1610184461,-0.2550325394,-0.0684329495,-0.133165136,-0.5474577546,0.2972139418,-0.1807413399,-0.1177975088,-0.1195981279,0.34742257,0.0213252027,0.0512089171,-0.3489221931,0.1687342376,0.2328534871,0.0347478427,-0.1949571669,0.1563552469,0.0093388567,-0.2200108021,-0.128311336,0.4098034501,0.2252566814,-0.0901622176,0.0976260453,-0.3563468158]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1840","title":"Add common voice","comments":"Let me know if you have any other questions","body":"## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":9,"text":"Add common voice\n## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nLet me know if you have any other questions","embeddings":[-0.2382236421,-0.2678191066,-0.0747092664,-0.2108193189,0.0883072391,-0.1280787438,0.3936897814,0.287578851,-0.2716715634,0.3026846647,-0.2857793868,0.1673590094,-0.0559147894,-0.0289889202,0.0087706577,-0.1595616937,-0.1767227948,0.1919882447,0.1013432816,-0.1971339136,0.1974736303,0.0249868967,0.0537406206,-0.0777602717,-0.1867675185,-0.1038363054,0.1538931429,-0.1469563693,0.0110157738,-0.0391430221,-0.1404046714,0.2559618056,-0.2115445286,0.471898973,-0.0000982532,-0.0029122778,0.0735937133,-0.1754899919,-0.0366154313,-0.0341927931,-0.1581026763,0.1725229025,-0.3245674074,0.1247059479,-0.2975684404,-0.0820175484,-0.0193571709,-0.1947485209,0.3683937192,0.3935413957,0.31129691,0.1311984956,-0.1663995385,0.0485569909,0.1760065109,0.1176249683,-0.2727398276,0.0104297474,0.2128439695,-0.106319204,-0.0831306055,0.4527285397,0.0536216497,-0.0689998493,-0.0570783503,-0.0184975248,-0.1617293954,-0.3956336081,0.0562633723,0.5271155834,0.5149642229,-0.3329765499,-0.155970484,-0.2387735844,0.0781372935,-0.1948537529,0.054799784,-0.0787407532,-0.254101485,0.2337895036,0.1096742377,-0.0785657912,-0.0756629184,0.1284011453,0.1119491011,0.3024908602,-0.0370476022,-0.3258404732,0.0185145754,-0.1877236217,-0.2702428997,0.2021368593,0.2867081165,0.0631090775,-0.3444045782,-0.2104883194,0.1489462852,-0.5666844845,0.2675295472,-0.1788342446,0.2426070571,0.2107140869,-0.2808881998,0.1893716305,-0.0654862002,0.0519322455,0.0790921748,-0.1993547231,0.3228347898,-0.0619957373,0.0042437874,0.0943063349,-0.0105269942,0.079160057,-0.1152660325,-0.006314171,0.1389518976,0.0346722975,-0.2961285412,0.0670859069,-0.113810353,0.0000616228,0.0959327817,0.2303646952,-0.0082467031,0.3303398788,-0.0118401749,0.181236282,-0.0288520642,-0.3438987136,-0.1362223625,-0.0910706818,-0.0743168816,0.2282446325,0.2842825055,-0.0043803365,0.2239324749,0.0726715028,0.2280563861,0.0044928854,0.0068305181,0.143604964,-0.062411014,-0.0489692353,0.0371386111,0.0324104503,-0.0779040456,0.0270219445,-0.1234966069,-0.1423603296,0.1102478653,-0.095477201,-0.3472796679,0.3663103878,0.2521957159,-0.1400751472,0.2302874625,0.5386869311,0.140455544,-0.2377386689,0.1479980201,0.1864100397,-0.3681662679,0.002961545,0.1636038423,0.0203448292,-0.2359216511,-0.1182373464,-0.08390183,0.0562380217,-0.0619437024,0.3340632617,0.0074057011,0.0562886074,-0.039279107,0.464640677,0.40171808,-0.2794468403,-0.3844754994,0.1350608766,-0.2910127044,0.0707285926,0.2410091162,-0.0320668817,0.1923723817,0.0128371529,0.4100187123,0.3514208496,-0.2066759914,0.097323142,-0.2118415087,-0.1888464689,-0.0136363981,0.17784217,-0.1367955953,-0.0226215255,-0.0954380929,-0.017519325,0.2285344154,-0.32302472,0.1642069072,0.1213367507,0.4600415826,0.0000769859,0.0191716701,-0.27780056,-0.0727261603,-0.1138765216,0.0632389709,0.2095928639,0.0392244644,-0.5249679685,-0.1761268824,-0.2864405215,-0.0210308433,0.1469575316,0.3566933274,0.2836820185,-0.0654959977,-0.0321484916,-0.1826448143,0.1670834869,-0.0678853169,-0.0241763461,0.1074257046,0.1151986495,-0.2139552385,-0.1653514653,0.2419538349,0.2584077716,-0.2042247653,-0.0478267781,0.0226375107,0.2648958564,-0.3567682207,0.3918891251,-0.0352044404,0.3495094776,0.1195249707,-0.2125895917,0.0203500129,0.173713088,-0.0269179028,-0.0532440506,0.197110489,0.1048903316,0.5478461385,-0.2419927269,0.0998563394,-0.1302845925,0.2862187922,-0.0658916384,-0.1163056046,-0.0795817152,0.0484531485,-0.0511933528,0.0946758017,-0.0596934818,-0.6177271605,0.2386604548,0.6700246334,-0.1406178772,0.434856236,0.1967803836,-0.2523568273,-0.082129851,0.0514996387,-0.2096174061,0.064298518,0.2990114093,0.1732086837,-0.0128203901,-0.0564529225,-0.159454003,0.1024216637,-0.0732132867,-0.0947678536,0.4619303048,0.2561556697,-0.0150685478,-0.6138242483,-0.2034734488,-0.1044143811,-0.0184672102,-0.202953428,-0.1147368327,-0.1389848739,-0.2419732809,0.0849612504,0.0748541951,-0.1227845848,-0.0028570616,0.4810583889,0.3840859532,-0.0510128886,0.3442371786,0.054133039,0.4054787159,0.1513215303,0.0313036628,-0.0272550657,-0.0587842278,-0.0930643082,0.285811007,0.0859403238,-0.1829951853,0.2515233755,0.0164428353,0.1496094465,-0.0518244281,-0.3714227974,0.2229581326,-0.3197425306,-0.2146982551,-0.1301200986,-0.0014517028,0.0478600003,-0.4632029533,-0.0172709282,-0.2196966857,-0.2276558727,-0.0211057346,-0.1298241466,-0.031329751,-0.278537482,-0.4668720961,-0.0894095823,-0.5374430418,0.119352214,-0.3069636822,0.0557770915,0.1825977266,-0.1932152063,0.0621634386,-0.2316227406,0.4421444833,-0.3724253178,-0.2166428566,0.2298760861,-0.3108026385,-0.3016857803,-0.2083777785,-0.1690123081,0.3761355281,0.0088705523,-0.1607681811,-0.2161200047,-0.1200521365,0.2622130811,-0.0526589006,-0.0013504913,0.1476102918,0.1870803088,-0.2250598669,-0.1597432345,-0.0868693739,0.0912497714,0.2776323259,-0.021173466,0.0191120282,0.1974036396,0.0896310061,0.7798036337,0.0030741501,-0.2544307411,0.2998552918,-0.2455364764,0.1621954739,0.0262348298,-0.1391484737,0.403365463,0.2146314532,0.2594225109,0.5721776485,-0.0412044898,-0.1684058756,0.0686147511,0.0598848015,-0.2934357226,-0.2760421634,0.0948270336,-0.1109370515,0.1431526244,-0.0320742726,-0.0423095748,-0.2203027457,-0.0152586419,0.3456614912,0.0650441721,-0.1241401508,-0.2581171095,0.0395885929,0.0968482196,-0.291011095,0.0813542157,0.2313378453,0.1801365167,-0.3637230992,-0.1618646085,-0.1837718487,0.1683288366,0.3122244179,-0.2164858133,-0.1954807937,0.0179755203,0.2271992564,0.038455639,0.3115446568,-0.2119858116,-0.1947025955,0.1013610736,0.0754442737,-0.2594752014,-0.0058653723,-0.129339844,0.206886813,-0.0282495078,0.0458421782,0.0198283456,-0.174311474,-0.2403411865,0.0663837269,0.1033089384,0.0197822303,-0.0205058847,-0.1656936407,0.3303748071,-0.2615110576,0.4084160924,0.3044135571,-0.087068975,0.2918069661,0.1434080601,0.0504095443,0.1230741218,-0.0436116755,0.0761445388,0.2005804181,-0.2438696325,-0.1680830717,-0.1751441509,0.3378288746,0.0548024029,0.1754476726,0.1617284268,-0.0724396408,0.0985545963,-0.0901052505,0.3090377152,0.3734277189,0.2678983808,-0.1603151262,-0.2177915126,0.1339592487,-0.0141784819,-0.13505885,0.3071519732,-0.0654562265,-0.3352682292,0.0497856289,-0.3560748994,0.9404855371,0.0967246592,0.184204042,0.3640790582,0.266089052,0.172692582,-0.5206198096,0.2037952989,0.1337493509,-0.3132113218,0.0223865192,0.1593757868,-0.2072203308,0.0894248411,-0.3803370297,-0.056971319,-0.0039334162,0.0684275776,0.1262995899,0.3074982762,-0.0537265055,-0.4448840022,-0.2847499847,0.2471996099,-0.01625528,0.0791521072,0.1782318205,-0.2045378238,-0.1261852384,-0.0709156394,-0.2027577162,0.2340609729,-0.1972101778,-0.1038346663,-0.0600356422,0.0570050217,0.2419944555,0.0930318311,0.3585981131,0.1113185883,-0.2175731659,0.1122375056,0.2688136995,0.1100721583,-0.1851909161,0.1041144207,0.2805192471,-0.2624199092,-0.1844248474,0.1536230594,0.2289160043,-0.1666930169,-0.3544749022,-0.3298838139,0.0712255463,-0.1176898777,0.2789879441,0.2944736779,-0.1688928008,-0.1672341526,0.2717521191,0.1298841238,-0.3767681122,0.0381564349,0.2295949161,0.0220148545,-0.1180714816,0.0997902006,0.1715379357,-0.0009436772,0.2659237385,0.0518171936,-0.1247306466,-0.4087865949,-0.3009690344,-0.020266749,-0.0983244926,-0.1901747733,0.0175762158,-0.4362936318,-0.1209758818,0.2004640698,-0.0620975755,0.1863871068,-0.0868361741,0.0099098617,-0.4397216737,0.0630102977,-0.0052013574,0.2504269779,-0.1540269405,0.2527239323,-0.0217550248,0.0974726006,-0.5342223644,-0.2396019995,-0.0305801462,0.1716658473,0.2419621646,-0.311165154,0.2496322691,0.0304601006,0.3737986088,0.0133679118,-0.0504463017,-0.4013290703,-0.2029195726,0.0709627792,0.2086209059,0.0176564232,0.2480520457,0.1149219498,-0.2293572277,-0.0970981419,0.2992959023,0.0245133154,0.0131763238,0.1347276419,-0.0273320917,0.0038161778,0.2183496058,-0.2522547543,0.0493124351,0.0594569854,0.0200800579,-0.0057399343,-0.041519884,-0.261695087,-0.0813249648,0.1762252897,0.3107417524,0.0008242982,0.0798378959,0.0552892499,-0.1575899124,0.2364946753,0.3327212632,-0.1030114144,-0.1769136637,-0.0182224754,0.2542023063,0.4395594001,-0.2807128131,0.067829676,0.1304859668,-0.2437882423,0.4216021299,0.1325263381,-0.1088773683,0.1037549973,0.089577809,-0.0940179899,-0.2333921343,-0.1905630827,0.0552088767,0.0901983976,-0.0208283421,-0.0029317534,0.1126643419,-0.0473081507,0.308252573,0.1653174013,-0.182118848,0.2899439931,0.1121810898,-0.0831332803,0.5081995726,0.05959142,0.0535347015,0.0297685266,-0.0856696665,0.1934975386,0.0945783108,0.3338249922,-0.1504305303,-0.1856483519,-0.0123799732,0.2919334173,-0.0765179917,-0.1177664399,-0.1619239748,-0.0056995465,-0.1206094101,-0.2692968845,0.2496197969,0.0874820724,0.0559513532,0.0337265693,-0.0286684539,-0.034168683,0.0086518032,0.0561746918,0.4016864896,-0.0231404752,0.2302810103,-0.0686261058,-0.2958490551,0.1599273533,0.3511986732,0.144282043,-0.07825277,-0.2009251118,0.4466331005,-0.1266705841,-0.0784544647,-0.1457654238,0.1082864925,0.3470347822,0.0934155881,0.0924698487,0.2763318121,-0.3151222169,-0.1480315775,0.0549496934,-0.1316834688,-0.1560070962,0.1256276518,-0.0636690184,-0.0326499753,-0.353878051,-0.1790257692,-0.5263889432,-0.1165021881,0.4589375257,-0.044170659,0.0961078107,-0.3061572015,0.1664530337,-0.0302783232,0.1684652269,0.2527271211,0.0619892776,0.3395503759,0.0096888626,-0.347291261,0.3597622812,0.2079606056,-0.0399530306,-0.2994641066,-0.0228798948,0.3948984742,0.1054984331,0.089666456,-0.3412857652,-0.266066879,-0.2469094694,-0.2780716419,0.0940278322,-0.2375137657,0.1955080479,0.2650447786,-0.2850655615,0.065227136,-0.1292154044,0.2803738713,-0.1332630217,0.1227201596,-0.0979363322,-0.2600450218,0.2062563747,-0.0581347197,0.2401164621,-0.3883155584,-0.1680299938,0.0521804839,-0.157791391,-0.3000992239,-0.0511393994,0.1162719131,0.4455352128,-0.0121318623,-0.0002323747,-0.166478008,0.1810343713,0.1803265959,-0.1446742117,-0.0047501735,0.2498239577,-0.0265495237,-0.016276015,0.0590801723,0.1962776333,-0.187193349,-0.0371850394,-0.0746863112,-0.5044651031,0.3564579487,-0.2429114133,-0.2908271849,-0.2250937968,0.3436104953,0.034076158,-0.0162363742,-0.3575397432,0.1171083078,0.2304656059,-0.0074445377,-0.1364469081,0.2068141699,0.0806964412,-0.253585875,-0.0957234651,0.3005552888,0.0982016101,-0.0755610168,0.0516241826,-0.3502775133]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1840","title":"Add common voice","comments":"I added a Work in Progress pull request (hope that is ok). I've made a card for the dataset and filled out the common_voice.py file with information about the datset (not completely).\r\n\r\nI didn't manage to get the tagging tool working locally on my machine but will look into that later.\r\n\r\nLeft to do.\r\n\r\n- Tag the dataset\r\n- Add missing information and update common_voice.py\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/1886","body":"## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":66,"text":"Add common voice\n## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nI added a Work in Progress pull request (hope that is ok). I've made a card for the dataset and filled out the common_voice.py file with information about the datset (not completely).\r\n\r\nI didn't manage to get the tagging tool working locally on my machine but will look into that later.\r\n\r\nLeft to do.\r\n\r\n- Tag the dataset\r\n- Add missing information and update common_voice.py\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/1886","embeddings":[-0.1344472319,-0.2228563577,0.0001928245,-0.1578856409,0.1937064826,0.0049486011,0.2681259215,0.3020458221,-0.2495004237,0.1795224696,-0.3955019414,0.2300344706,-0.0425767265,-0.0264860932,0.0067098695,-0.100829035,-0.1025349498,0.1595255136,0.104188852,-0.2480209321,0.1949149221,0.066846244,0.113306269,-0.079002507,-0.3241567016,-0.1644456834,0.0231045038,-0.1435564458,-0.0003262647,-0.1549275666,-0.0453543141,0.271556288,-0.2102986276,0.6402723193,-0.0001004912,-0.0444369279,0.034191668,-0.2457211912,-0.0831898525,-0.1936043054,-0.1155194193,0.1076655388,-0.2947714627,0.1147531345,-0.2959162891,-0.2513290942,-0.0181471184,-0.2323185205,0.3075140715,0.3086764514,0.2740311027,0.0791771412,0.0037059225,0.1020501032,0.1946553439,0.150016889,-0.2205498368,0.1714189798,0.3242516816,-0.1366104037,-0.0471704826,0.466268301,0.0955390558,-0.2063768059,0.072122775,0.0336857252,-0.1408482492,-0.4390898943,0.074878417,0.4949438274,0.5199844241,-0.4264089763,-0.3617741764,-0.3173406422,0.1082020923,-0.2514179945,0.101029627,-0.117975235,-0.2557395697,0.2239129543,0.1083872914,-0.1283470839,-0.0489875153,0.0811170936,0.1294689775,0.4218024313,0.031774193,-0.2574902177,-0.0835851654,-0.1344355196,-0.4091210961,0.2767671645,0.2451603264,0.2131911367,-0.2955951393,-0.1637175083,0.2193738818,-0.5752841234,0.2983843386,-0.1418928951,0.2043370605,0.1789840609,-0.3365133703,0.118501775,-0.0730809122,0.1383013278,0.1229962558,-0.1810256541,0.2307826877,-0.0775236487,-0.0508183204,0.0256609339,-0.0354845338,0.1108120829,-0.0737333149,0.0099485992,0.1576343775,0.0028997003,-0.2117083967,0.1503167301,-0.1971748471,-0.0134349549,0.0216207597,0.2945629358,-0.1580664068,0.3248334229,-0.0321693234,0.0962950736,0.0461165011,-0.1675426513,-0.1163627058,-0.1051338688,-0.0598575883,0.1783371121,0.2899332941,-0.2306269258,0.2474514097,0.0906155556,0.2151056379,0.0239891596,-0.1130202934,0.1488958597,0.0352118611,0.0130945509,-0.0530398674,0.1488322169,-0.0161975995,-0.0461042896,-0.1122690737,-0.1488526613,0.1025853455,-0.099144429,-0.4022328258,0.3375139534,0.1065637469,-0.0816783831,0.3401318192,0.3973253667,0.1810902953,-0.2607650459,0.0863159448,0.2044627815,-0.4028556049,-0.1337232888,0.298826009,0.0428128727,-0.250143826,-0.1295451522,-0.003745341,0.0662233084,-0.0445564017,0.220628798,0.0588412844,0.0657696277,-0.0801879242,0.4226998389,0.2906529307,-0.2918366492,-0.3610621095,0.1677829325,-0.2368706614,0.1416334808,0.1983561218,-0.1059633046,0.2048781216,-0.0048461971,0.5065509677,0.2191433907,-0.2216200233,0.0596206151,-0.2887995541,-0.1531593353,0.0818340927,0.0917185843,-0.1394603699,-0.0696659982,-0.1532015353,-0.110891059,0.2634885609,-0.2050185502,0.179952383,0.1328167766,0.4978140891,0.1184108183,0.1031948924,-0.203864947,-0.1884199977,0.0100561492,0.0054535698,0.3142545819,-0.0859167278,-0.523368597,-0.2934357226,-0.3056114614,-0.1241282672,0.0104292948,0.3078225255,0.2625297308,0.0256660394,0.0203962699,-0.2492913753,0.121369414,0.0112154894,0.079527095,0.0230166875,0.0897838324,-0.2401885241,-0.1604061872,0.228862375,0.3455804884,-0.1591655463,-0.0602729991,-0.0589507483,0.3045244515,-0.3686720431,0.3352653086,-0.0782301873,0.3569921255,0.184901759,-0.1701756865,0.0201045815,0.1942465454,0.074120447,-0.1760630906,0.1142837107,0.1050393581,0.337246269,-0.1475910842,0.0875649154,-0.160045132,0.3091047108,0.0102095837,-0.0980091617,-0.0631413087,-0.0762401596,-0.1810005456,0.2161413133,-0.0155315371,-0.6039136648,0.2193379253,0.6066966057,-0.0502643771,0.2664574981,0.189058423,-0.305285275,0.0091059478,0.0964586809,-0.2371928096,0.077658385,0.2682398558,0.1686557978,0.0385631993,-0.0325515419,-0.1328755319,0.0287766457,-0.061764881,-0.0482396446,0.5199928284,0.2396480143,0.1290621608,-0.4704148173,-0.2076030225,-0.103740707,0.0820092782,-0.2697705328,-0.1634817868,-0.0728347301,-0.3710719645,-0.0508547053,-0.0506706573,-0.3408683538,-0.1255007684,0.3960850239,0.3298542202,0.0087162899,0.4096944034,0.0511541069,0.1571533084,0.0600988045,-0.1292621493,-0.0713003501,-0.0899616331,-0.1671464294,0.2367101908,0.1432028264,-0.1959049702,0.233914569,-0.0050524822,0.1264352053,0.035935659,-0.4879772663,0.1857514232,-0.4026751518,-0.2273771614,0.0057523223,0.0000285043,0.0893843547,-0.4461905658,0.0706716999,-0.2756741345,-0.2109452337,-0.0913038775,-0.1745148599,0.0008462273,-0.300188303,-0.5085436106,-0.1244488209,-0.5271833539,0.1792955995,-0.2819218934,0.1161316559,0.2779878676,-0.1688766479,0.1067907289,-0.2514390945,0.4454709291,-0.3080193102,-0.225850746,0.2266342491,-0.1953063458,-0.3069180846,0.0314145386,-0.1508218646,0.5303356647,-0.0159355123,-0.0882406831,-0.1784488112,-0.0283808243,0.2868050933,-0.031425938,0.0683951601,0.2362618744,0.2672896385,-0.2033075839,-0.1630727947,-0.1307861358,0.0234084204,0.1973066032,0.0784574971,-0.0149753941,0.1801187843,0.1350790262,0.8288182616,0.0600885041,-0.2452115417,0.2792580128,-0.2965868711,0.2424110472,0.012059669,-0.2315934896,0.348926127,0.1363293678,0.1873923689,0.4802671969,0.0852959678,-0.0140556982,-0.0935033262,0.0953903049,-0.3411554098,-0.3090341985,0.212346226,-0.1360484213,0.1384043992,0.0214224216,0.0657255873,-0.1959523857,-0.1094430462,0.3681486249,0.1309851259,-0.0633036345,-0.1535940617,0.1443222761,-0.0149173308,-0.3111484647,0.0432718843,0.2908510566,0.1650206745,-0.3033664227,-0.1587877721,-0.204971984,0.2093769014,0.2548339963,-0.20755364,-0.1739871204,0.1600529701,0.2465028614,0.0190077089,0.2726394236,-0.0969249085,-0.1166559979,0.1592395604,0.1634501815,-0.350053668,-0.0576791689,-0.2100788951,0.2365627885,-0.0404832624,0.0786974877,-0.0489070453,-0.2119986862,-0.2692859471,-0.0303384233,0.0934455544,-0.0216288939,-0.0165639631,-0.0174633525,0.3813253641,-0.2355314195,0.4151549935,0.3084321916,0.0095391199,0.2811662257,0.2500530779,0.1180639043,0.2223774493,0.0232110452,0.3118491769,0.1129244789,-0.2161222249,-0.1879103631,-0.1973955184,0.4451722205,0.1190070435,0.2169222832,0.0825161412,-0.0761467889,0.0291802902,-0.1812075675,0.3324918449,0.307197094,0.2214367539,-0.1813539714,-0.2902571559,0.1426408738,-0.006095794,-0.1787079722,0.3708269,-0.0113310311,-0.3125833869,0.1201831102,-0.3805450201,1.0539363623,0.1060054004,0.1705633253,0.3440305889,0.1708270013,0.2144341618,-0.4325200617,0.1412156671,0.0622890107,-0.2115291208,0.0575007051,0.1662165672,-0.1885233223,0.0331700556,-0.3631230891,0.0602660663,0.0159086473,0.0602484345,0.1308225989,0.3993856609,-0.0664177313,-0.4873315394,-0.2496803403,0.1965699345,-0.127259627,0.2601580024,0.1786094904,-0.2498539239,-0.2813642025,-0.0841066092,-0.2218987197,0.2095036209,-0.2499833554,-0.2165773213,-0.0586412847,0.0542161539,0.2628611624,0.0235474501,0.263543427,0.1367248595,-0.2656131685,0.1696123332,0.2035770863,0.1425755024,-0.0395372696,0.1031611264,0.4867222011,-0.2618868947,-0.1868196577,0.1463110894,0.0967751071,-0.008075132,-0.3674502969,-0.3758533001,0.0626390651,-0.2191749662,0.3816480339,0.2362188995,-0.074613601,-0.1278438568,0.2453705817,0.1814095676,-0.3947041333,0.012283179,0.1293305904,0.0621429272,-0.1587810516,0.1683374345,0.1127039269,-0.0496092141,0.3067156076,0.1275480837,-0.0575342737,-0.3017158806,-0.3166985512,0.0235545933,-0.2233227789,-0.3351829648,0.0105897384,-0.3136936128,-0.1052810475,0.2668185234,0.0611059107,0.1855639964,0.0070173531,-0.0553820245,-0.5129721165,0.1031635627,0.0867874324,0.1530074179,-0.2143234462,0.3322760463,0.1086516976,0.1653893888,-0.5192533731,-0.2462426573,0.0332908817,0.1924853623,0.1842329353,-0.2702376544,0.3270450532,-0.1153580472,0.3327232897,0.0361935496,-0.0305365566,-0.3145813346,-0.2067761719,0.0717497692,0.2016597837,-0.0041058012,0.1488688588,0.1970037967,-0.2191012949,-0.1433822364,0.2331741899,0.0607710853,-0.0100695239,0.1859636456,0.0054921401,0.0829414576,0.2363807708,-0.3351299763,0.0773636028,0.1577954143,0.1253006607,-0.0204265919,-0.2257257402,-0.2216388136,0.0209905282,0.1940127462,0.1762096435,-0.0812989771,0.2119425088,-0.0126772569,-0.1316134483,0.2430470437,0.4936256409,0.0372073464,-0.1053428203,-0.0177303739,0.3020399213,0.3756219447,-0.3024636805,0.155786559,0.1194752678,-0.181477353,0.3449592292,0.1164901778,-0.0015551514,0.1204197332,-0.0463930219,-0.0269505363,-0.0555053465,-0.2370213121,0.0296387468,0.0303875878,-0.0536548607,0.0260658655,0.1019398049,-0.0088625476,0.3165412545,0.2017363161,-0.219779402,0.3501244783,0.1147186682,-0.1532222927,0.4321584105,-0.0566321686,0.0155755393,0.0331125818,-0.0650251135,0.2184816599,0.0824351832,0.5198343396,-0.1819770038,-0.2765978277,0.1574695706,0.3043002784,-0.0608052164,-0.0519479476,-0.138087973,-0.0286540631,-0.1211852431,-0.2810142338,0.2142596096,-0.0345568247,0.1360968947,-0.0227321014,-0.1195059419,-0.0947603136,-0.2054433823,0.0951797813,0.5703049898,-0.0488632023,0.2297981679,0.0943031013,-0.3598150015,0.1032425612,0.3485669494,0.2855004668,-0.1141801849,-0.1702528596,0.4518784285,-0.1236323416,-0.0518497154,-0.0853940099,0.1682268232,0.2482925653,0.037216235,0.0963401794,0.2396829128,-0.2887351215,-0.1879407167,0.0163461845,0.0142062725,-0.0523939244,0.1453014314,-0.1879596412,-0.1568939537,-0.3550959229,-0.1093550846,-0.5188398361,-0.126915887,0.4577679634,0.0951128006,0.1440895945,-0.3524836004,0.1503072977,0.0776492208,0.1650584787,0.3841638565,0.1056901664,0.2553605437,0.0124903983,-0.4982586801,0.2732191086,0.1742070019,-0.0787291229,-0.3084369898,-0.0202901848,0.3806377947,0.0694624186,0.0583566539,-0.2635870874,-0.1606087685,-0.3106971681,-0.2541221678,-0.0120296171,-0.3438553512,0.1505676508,0.2203052342,-0.1960609853,0.0125344023,-0.2732777297,0.247805655,-0.1533526778,0.0855223387,-0.057532534,-0.1711981595,0.2229479253,-0.0461711362,0.2796880603,-0.3314433694,-0.2411097884,-0.0077004484,-0.1423580199,-0.3083875179,0.1225637347,0.0703886375,0.5162594914,0.0552989282,-0.1387808472,-0.2611919641,0.1268989891,0.1418101639,-0.2411366403,-0.1124154553,0.3134895265,-0.0218003131,0.0789325312,0.1139827445,0.2645880878,-0.1810564548,-0.0235568844,-0.2825867236,-0.590260148,0.4051390886,-0.2649807036,-0.271805197,-0.2290439457,0.261384517,0.0491724722,0.0327177122,-0.5370386839,0.1134290099,0.2650887072,0.1048496291,-0.2321518362,0.0350141004,-0.0113587352,-0.271363318,-0.0993941501,0.2948484719,0.0960317701,-0.053183239,0.1073330566,-0.270847708]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1840","title":"Add common voice","comments":"Awesome! I left a longer comment on the PR :-)","body":"## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":10,"text":"Add common voice\n## Adding a Dataset\r\n- **Name:** *common voice*\r\n- **Description:** *Mozilla Common Voice Dataset*\r\n- **Paper:** Homepage: https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Data:** https:\/\/voice.mozilla.org\/en\/datasets\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/common_voice\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nAwesome! I left a longer comment on the PR :-)","embeddings":[-0.2226333469,-0.2290469259,-0.0770423934,-0.2084888965,0.1246274784,-0.1418179721,0.3705703914,0.350219667,-0.2901816368,0.3204793036,-0.2739417553,0.2019272,-0.0805767179,-0.0278763846,0.0207771901,-0.1629784852,-0.1768213063,0.174622491,0.1106891036,-0.1632323414,0.1972066313,0.0134652937,0.0632737055,-0.1052028537,-0.1438063234,-0.0436145365,0.1582066119,-0.1783482432,0.0089564482,-0.069856666,-0.1606133878,0.2376480997,-0.2060225606,0.4640005827,-0.0000988434,0.004416984,0.0658350885,-0.1882659197,-0.0454987735,-0.013920879,-0.1597845405,0.151212737,-0.3608586192,0.1308468133,-0.2983477116,-0.0912652016,0.0164545998,-0.1856850237,0.2969444394,0.4057072699,0.3025569916,0.1039947644,-0.1787906289,0.0818576962,0.1749000251,0.1285139322,-0.3160152137,-0.000010363,0.2565429509,-0.106873177,-0.1099789441,0.4412932992,0.0078622876,-0.1008394286,-0.0472868271,-0.0358729288,-0.1157791615,-0.3856057823,0.0671870187,0.5178171396,0.5020552278,-0.2926698923,-0.1130460054,-0.264192462,0.0872042775,-0.1635008603,0.0302145239,-0.0883074403,-0.2655741274,0.2566002011,0.0902115107,-0.068027243,-0.0679979175,0.1440623105,0.1597747207,0.341936022,0.0052369237,-0.3179692924,0.0067295125,-0.1473907679,-0.2583811879,0.2168437093,0.3037406504,0.0189357158,-0.348970294,-0.2076293826,0.1660055518,-0.5722863674,0.276142925,-0.1919868737,0.2168848217,0.2126398832,-0.2905551493,0.1466013193,-0.0623870417,0.0529233851,0.0729636997,-0.2165575475,0.3092233241,-0.0310564581,0.0156372637,0.1127173156,-0.0247531626,0.0954819992,-0.124078624,0.0002176053,0.2047645301,0.0565138608,-0.3048030734,0.0766942203,-0.1564541161,-0.0435683653,0.0696879551,0.2325110137,-0.0239768233,0.3679578304,-0.0509224236,0.1728673726,-0.0061579118,-0.3805988729,-0.1468554437,-0.1078763902,-0.0448294915,0.2110242099,0.2529570162,-0.0015960005,0.1817160547,0.0808100998,0.2361490726,-0.0041841636,-0.0150774391,0.1928868741,-0.0573205724,-0.0364454091,0.0104708225,0.009710853,-0.0427258238,-0.0011054402,-0.1330307573,-0.0993293449,0.0993312299,-0.1305955648,-0.3824472129,0.3664086759,0.2424589247,-0.1251050085,0.2269904464,0.5649292469,0.1684244573,-0.2136739343,0.2020111978,0.1930303574,-0.3718775809,-0.0111632999,0.1607573181,-0.0021509808,-0.2383367866,-0.0950737223,-0.1140718833,0.0681246072,-0.0395988412,0.3261310458,0.011179206,0.085625954,-0.0476415008,0.4340124726,0.3991634846,-0.2393786162,-0.4209574759,0.1179229692,-0.3225770593,0.0846640617,0.267131269,-0.0406597406,0.1898494065,-0.0236795451,0.4274797142,0.3321292698,-0.2370413542,0.1173251867,-0.2362869531,-0.2249902189,0.0239153281,0.1835041344,-0.1407333612,-0.0276748575,-0.0764470398,0.0013421305,0.276294589,-0.3254215419,0.1503465474,0.0719806701,0.4849914312,-0.0184271373,0.0358415544,-0.2714368403,-0.0976754278,-0.1448551863,0.1067218482,0.2289406061,-0.0025775563,-0.558411181,-0.1675481945,-0.2780793309,-0.0086607216,0.1289990097,0.3566849828,0.2712466419,-0.0845517442,0.0004593974,-0.1828338057,0.155088231,-0.086521484,-0.0378631353,0.1119125709,0.1076905802,-0.1539505124,-0.1326427013,0.2093821317,0.2893129289,-0.2107879519,-0.0533389933,0.0061197118,0.2358878553,-0.3635912538,0.3728990555,-0.0470726788,0.3741320372,0.1225299239,-0.2147866637,0.0043944325,0.1728391349,-0.0368863903,-0.0594242029,0.2074999809,0.0906101167,0.5196115375,-0.2014039904,0.0397373848,-0.1135318577,0.2646074891,-0.0428138636,-0.1535006613,-0.0600573532,0.0431999378,-0.0176881775,0.0771803334,-0.0004583724,-0.589910686,0.3049131632,0.7078103423,-0.1561803818,0.4018055499,0.2181809247,-0.2165738195,-0.1059894338,0.0729520842,-0.2144417167,0.0248244386,0.2925292253,0.1712110788,0.0040898961,-0.0239080563,-0.1653024554,0.0980009362,-0.0381008387,-0.1010659039,0.4702448249,0.2672325373,0.0225708522,-0.6138058901,-0.200622499,-0.0875616297,0.0216336697,-0.195210278,-0.1151304543,-0.1198053807,-0.2573671639,0.0944436565,0.0403944887,-0.1243240759,-0.0285063423,0.4808182716,0.4108341932,-0.0470931679,0.3321724236,0.0451832339,0.4387847185,0.1466405839,-0.0198493637,-0.0508785658,-0.0731909499,-0.06722074,0.2858459949,0.1244147718,-0.2063043118,0.2270344049,0.0175905153,0.1067440957,-0.066623196,-0.381960541,0.2131465971,-0.3545047045,-0.2386723608,-0.1137009412,-0.0471638367,0.0444415547,-0.4872060418,-0.0633698553,-0.1858779043,-0.2350442857,-0.0015418058,-0.1288415343,0.0197281949,-0.2857066393,-0.470440805,-0.0576378219,-0.5588482022,0.1377782226,-0.3459689915,0.035731189,0.1512639821,-0.2053277045,0.0704065189,-0.2438812703,0.4260413349,-0.3689607382,-0.2075457871,0.2135149688,-0.2544625998,-0.2875722349,-0.2418199927,-0.1712273806,0.3463823795,0.0078559099,-0.1721844673,-0.2488575876,-0.1305852681,0.2871814668,-0.0622497685,-0.0244068895,0.1384524703,0.1968735307,-0.215423286,-0.1178963184,-0.0601867773,0.0592678711,0.2517463565,-0.0027354243,0.0246086959,0.1922071576,0.1622520685,0.7889600396,0.0478856862,-0.2245557159,0.3034654856,-0.2248675525,0.1652219594,0.0320787244,-0.1543114185,0.4570647776,0.2345911264,0.2348009497,0.5374256372,-0.0614027828,-0.1532455832,0.1001426503,0.0719575062,-0.2589485347,-0.2761175931,0.1266602278,-0.1475321501,0.0913903937,-0.0123669049,-0.013102456,-0.1734261215,-0.0364248976,0.3347083032,0.0705050454,-0.1203340366,-0.2401481271,0.0662894771,0.131806314,-0.3243114948,0.0627390742,0.2253239751,0.191621244,-0.3653546274,-0.194295615,-0.1520607919,0.1896805167,0.3545732498,-0.2239930034,-0.209108755,-0.002410671,0.236685425,0.0746734813,0.3254806697,-0.1706316918,-0.1973135918,0.0986377075,0.0978557691,-0.2499479204,-0.0132211642,-0.1238846034,0.2010729909,-0.0048263744,0.0543432236,0.0432421714,-0.2022645622,-0.2367568165,0.0852359384,0.1395018995,0.0121882744,0.0097872531,-0.1767690778,0.329166472,-0.2248865217,0.3643877208,0.3009396493,-0.1210318059,0.3092730641,0.1365009248,0.0348760113,0.1279607862,-0.0798887238,0.0744017214,0.2202071697,-0.277125597,-0.1621266156,-0.1837916374,0.3701195419,0.0695960894,0.1799470782,0.1509771347,-0.080082342,0.1290866435,-0.104487583,0.3043946326,0.3581317365,0.2461159527,-0.1946095675,-0.2472268641,0.1353132129,-0.0262838453,-0.1271769702,0.3406808078,-0.0764262453,-0.3231713772,0.0509641841,-0.3456787467,0.9530736208,0.0809388235,0.2009085268,0.3685696423,0.2844898999,0.229531914,-0.5568885207,0.2117883414,0.1355103403,-0.2975157797,0.0208602194,0.1683393419,-0.243351236,0.0600250997,-0.4034921527,-0.0590952411,0.0113501092,0.0622364543,0.1747679263,0.2800404131,-0.0689108893,-0.4439918101,-0.2653532624,0.2283468395,-0.056021817,0.0796161667,0.1914303601,-0.2188002914,-0.1268146634,-0.0726896599,-0.2544878721,0.2312150747,-0.1856659502,-0.0994521603,-0.0524435416,0.031074563,0.2245290428,0.1105073541,0.357139796,0.0933724791,-0.2114784569,0.1152805462,0.2820362151,0.1313298941,-0.1364856064,0.0506700464,0.3115553558,-0.2507077157,-0.1562648714,0.1370263845,0.2445797622,-0.1896929294,-0.3634392917,-0.312880069,0.0908688754,-0.10684257,0.2612343431,0.3012030721,-0.1926568449,-0.1759650856,0.2679816484,0.1239947677,-0.3646299839,0.014758531,0.2350839674,-0.0134400502,-0.108598873,0.08089865,0.1659953743,0.0169873089,0.2095856816,0.0471363254,-0.1150551289,-0.3969002962,-0.2705089748,-0.0287044067,-0.1143131629,-0.227528885,-0.0297744721,-0.4169602692,-0.1326606274,0.1846920103,-0.081126757,0.1956051737,-0.0959915072,0.0191765651,-0.4222129583,0.0553550683,-0.0393245891,0.2401393205,-0.1331290305,0.247322008,-0.0651919097,0.0714587867,-0.5276107788,-0.2711085081,-0.0053973957,0.1403066963,0.2462263703,-0.3047358394,0.2539516687,0.0067723645,0.3645387888,0.0602430664,-0.0357118547,-0.3963227272,-0.2067503631,0.0764583275,0.219605118,0.0265016407,0.2560463846,0.1283284277,-0.2505615056,-0.0918861777,0.2991907299,0.02919719,0.0219014138,0.1447609216,-0.0272273272,0.0121621517,0.2431429774,-0.2734985352,0.0348646529,0.0666250959,0.0459707752,0.0006363466,-0.0649084076,-0.2327651232,-0.1043543965,0.1757947356,0.3103207946,0.0224830676,0.0701475814,0.1031914428,-0.1927523315,0.2162709832,0.3122358024,-0.0801046044,-0.1914384812,-0.0386073329,0.2668325603,0.4309481978,-0.2849965692,0.0450741462,0.1174256206,-0.1989818513,0.3952830434,0.1460401863,-0.1043536514,0.1323099434,0.0830771774,-0.1015666798,-0.224127844,-0.2035752833,0.0873854458,0.1003790721,-0.0118843978,-0.0045905216,0.0987700075,-0.0912168324,0.2996628582,0.1701090932,-0.1792448312,0.3097628057,0.1133677587,-0.0833967254,0.5009977221,0.0980709866,0.0878793523,0.0255937334,-0.1347993463,0.1872495115,0.1346872002,0.3172205091,-0.1295847744,-0.1606818289,0.0064513264,0.2784840167,-0.0457328372,-0.0992045626,-0.1433821321,0.0146804145,-0.1432586908,-0.2470686734,0.240956068,0.0752104595,0.0938374996,0.0476316884,-0.0426590405,-0.0568066239,0.0074049472,0.0399901234,0.4452252388,-0.0444696434,0.2046228796,-0.0994325504,-0.3252308667,0.1341963559,0.3314499259,0.1390901059,-0.0948276967,-0.1754549891,0.4211797118,-0.1568010151,-0.050404828,-0.1859866381,0.1173883602,0.3302864432,0.0993864387,0.085347943,0.2720995545,-0.2975463569,-0.1121356785,0.0785678774,-0.1475992948,-0.1590441763,0.1690138578,-0.0465112627,-0.032360021,-0.3552810848,-0.1566990614,-0.5204899311,-0.1145130843,0.4464803636,-0.0252518244,0.1295678765,-0.3468650579,0.1629621387,-0.0002357468,0.140884608,0.2860236466,0.0500750728,0.3663334548,0.0480839536,-0.3783964515,0.3710702062,0.222507447,0.0036283303,-0.2618438303,-0.0158092007,0.412337929,0.1652400494,0.0817114636,-0.3322127759,-0.2714355588,-0.2398886085,-0.2692937255,0.0971725434,-0.2216737866,0.1932360381,0.2515974343,-0.3080844879,0.1013156027,-0.1183656603,0.2678826451,-0.1777870357,0.1230058149,-0.0996507481,-0.236341089,0.1954914331,-0.035887327,0.1934611648,-0.4147116542,-0.190362066,0.0120070409,-0.1568370014,-0.3140705526,-0.0473385565,0.0857278183,0.4211135209,-0.0298719704,-0.0388883464,-0.195713535,0.1977369487,0.1551088393,-0.1336372048,-0.0309883058,0.2695300877,-0.0205102358,-0.0136764245,0.0162107255,0.1738127023,-0.1793684661,-0.0531246476,-0.0707961544,-0.5259862542,0.3386063874,-0.2059840113,-0.29532516,-0.2113706768,0.3218530715,0.0515075922,-0.0251810309,-0.3395134509,0.0767257884,0.2367912084,-0.0035685906,-0.1496741772,0.2271329463,0.1026417166,-0.2621096075,-0.1315720081,0.3072858751,0.0826981813,-0.097842738,0.0942423791,-0.3949256241]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1838","title":"Add tedlium","comments":"Hi @patrickvonplaten \r\nI can have a look to this dataset later since I am trying to add the OpenSLR dataset https:\/\/github.com\/huggingface\/datasets\/pull\/2173\r\nHopefully I have enough space since the compressed file is 21GB. The release 3 is even bigger: 54GB :-0","body":"## Adding a Dataset\r\n- **Name:** *tedlium*\r\n- **Description:** *The TED-LIUM 1-3 corpus is English-language TED talks, with transcriptions, sampled at 16kHz. It contains about 118 hours of speech.*\r\n- **Paper:** Homepage: http:\/\/www.openslr.org\/7\/, https:\/\/lium.univ-lemans.fr\/en\/ted-lium2\/ &, https:\/\/www.openslr.org\/51\/\r\n- **Data:** http:\/\/www.openslr.org\/7\/\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/tedlium\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":40,"text":"Add tedlium\n## Adding a Dataset\r\n- **Name:** *tedlium*\r\n- **Description:** *The TED-LIUM 1-3 corpus is English-language TED talks, with transcriptions, sampled at 16kHz. It contains about 118 hours of speech.*\r\n- **Paper:** Homepage: http:\/\/www.openslr.org\/7\/, https:\/\/lium.univ-lemans.fr\/en\/ted-lium2\/ &, https:\/\/www.openslr.org\/51\/\r\n- **Data:** http:\/\/www.openslr.org\/7\/\r\n- **Motivation:** Important speech dataset\r\n- **TFDatasets Implementation**: https:\/\/www.tensorflow.org\/datasets\/catalog\/tedlium\r\nIf interested in tackling this issue, feel free to tag @patrickvonplaten\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHi @patrickvonplaten \r\nI can have a look to this dataset later since I am trying to add the OpenSLR dataset https:\/\/github.com\/huggingface\/datasets\/pull\/2173\r\nHopefully I have enough space since the compressed file is 21GB. The release 3 is even bigger: 54GB :-0","embeddings":[-0.3431777358,0.0705063418,-0.0683075488,0.1333120912,0.0683671609,-0.020419864,0.0353584178,0.3795070648,-0.429474771,0.2744410634,-0.2579486966,0.3504010439,-0.2084373534,-0.0917881429,0.226974383,-0.3618197143,-0.1412836015,0.1973717064,-0.1658695489,-0.1800617725,0.047073029,0.1284918189,-0.0539657511,-0.4520364404,-0.2660470009,-0.0807280019,-0.1142628044,0.0259981919,-0.180357486,-0.0326378085,-0.0109999729,0.1045881361,0.0020434626,0.5920302868,-0.0001060044,0.0762010664,0.2563407719,-0.1970420927,-0.1261660904,-0.0920040682,0.2239137292,-0.1876969635,-0.4069092572,0.0164639726,-0.2588905394,-0.3076374531,0.0834539384,-0.0051604914,0.2092911601,0.3986495435,0.2353914082,0.5247448087,0.3459710479,-0.1376192272,0.2615657747,0.0599569045,-0.187962994,-0.0069509097,0.1280572861,0.1660992652,0.0979942828,0.3711438775,-0.0610381812,-0.1315061897,-0.0925385579,0.0455805473,-0.3500027955,-0.2385309786,0.0007924585,0.5430955291,0.6054988503,-0.0139034195,-0.1867126077,-0.1310666203,0.0636987314,-0.432633251,0.0834901333,0.1428719461,-0.238803491,0.1338607967,0.149444133,-0.1192983016,-0.3330072761,0.4302216172,0.137561962,0.2157858163,-0.145215407,-0.0925252214,0.1593595445,-0.1750651598,-0.2267788202,0.1349192709,0.0857594162,0.0308885165,-0.2997762561,-0.1899608672,0.0482355021,-0.3286361396,0.3431006968,-0.1068293005,0.0370586552,0.0968602896,-0.197213456,-0.0967442691,-0.1431961507,0.1244440675,-0.1961450428,0.056293238,0.0263551921,-0.0768797323,-0.0339645818,0.0444282517,-0.1801440418,0.005402517,-0.198944211,-0.1756489724,0.0056034042,0.0284628253,0.0730405822,0.0255739838,-0.1170153171,-0.0077279923,-0.0521265678,0.2615172863,-0.073761709,0.1935297549,0.0640940294,0.1793647856,-0.0668099597,-0.2273440808,-0.2411783934,0.202768594,0.0086696483,0.1834232062,0.2094158679,-0.1675250381,0.4441908598,-0.0717104748,0.3050296307,0.1925251186,0.2671249509,-0.0930966362,0.1217700467,0.3277336955,0.086761713,0.1370448917,-0.018653417,-0.1039518714,-0.0937542543,0.1236060932,-0.0426069945,-0.1835628003,-0.3241637945,0.2468047142,0.0472360067,-0.1724043936,-0.2665527165,0.4495433271,0.0254369136,0.011575317,-0.0133621879,0.091232717,-0.4628695548,0.0088451225,0.1388831139,0.066054292,-0.1856429279,0.0794804096,-0.0098764095,0.2739853859,0.3011516333,0.3396631479,-0.1163510829,0.1494064182,-0.0522679649,0.3731703162,0.3246429563,-0.5450291038,-0.4533584118,0.1047696397,-0.3069199622,0.0519352108,0.2519710958,0.1499703825,0.3737607598,-0.083045736,0.4227801859,0.2034450471,0.0110581201,-0.1320343763,-0.3299342394,-0.2108197808,0.0721389651,0.233444199,-0.0465111211,-0.1430468857,-0.1827776283,0.5555557013,0.2120675594,-0.1718606353,0.11728549,0.328137815,0.059160769,0.0588179342,-0.0172087252,-0.2287879735,-0.0749335736,-0.1233118027,0.2061934024,0.5123475194,0.3007350564,-0.2830674052,-0.1193273365,-0.2881749272,-0.0058933925,-0.0049295621,0.2434927523,0.1954974234,0.1445725858,0.0592778288,-0.175257951,0.2067081332,-0.3404854536,0.0865433142,-0.377746582,0.1474134922,-0.0758847222,-0.2832172215,0.3221106231,0.0795254186,-0.0662517026,-0.221248731,0.1154898256,0.1340413541,-0.2328927964,0.2099721432,-0.0679047331,0.2508223951,0.2430242151,-0.2084759027,-0.0438278243,0.3286625147,-0.0376211964,-0.1305457205,-0.045052629,-0.0507194921,0.2623155415,-0.244640246,-0.0181983672,-0.0320662931,0.3046453297,-0.0850438997,-0.1413047761,-0.1966558993,0.2957748473,0.2194461524,-0.1282253414,-0.0062848977,-0.4803938866,0.2296601981,0.341532737,-0.1680434942,0.1300327629,0.2549462914,-0.288251847,-0.0319423974,-0.0485630706,-0.1839367747,0.2958205938,0.3423232436,0.1862992048,-0.0109877614,0.0324562825,-0.2777961493,0.159609884,-0.0229557622,0.0170587786,0.3979536891,0.2667993009,0.102761142,-0.4513935447,-0.0100574531,0.0496091805,0.2514970005,-0.1109845489,-0.231736064,-0.2771235704,-0.2104730457,-0.1024403796,-0.1342397779,0.0055293334,-0.0401664153,0.257499665,0.3070298433,-0.1106538773,0.1858116537,0.1710909307,0.0387198776,0.1562957615,-0.1112624779,0.0210467242,0.0019961558,-0.1498077363,0.1764420122,0.1308522373,0.1032979414,0.3310488462,-0.0567965023,0.0055520684,-0.1606166363,-0.5656263232,0.279306978,-0.3847056627,0.0875758529,-0.0339027233,0.2460289299,0.1505428702,-0.4598914385,0.0292263329,0.0852551535,-0.25328812,-0.0471895598,-0.1832279861,-0.103262715,-0.1788061857,-0.4804840982,-0.1146158949,-0.5522738695,0.1962411553,0.0653061271,0.0634834245,0.4290138483,-0.0963242352,0.1765257567,0.0646982491,0.1573942751,-0.3818121254,-0.2910321951,0.5326831937,-0.3773007989,-0.4238978326,-0.093412362,-0.1186183542,0.1072217897,-0.0012675964,-0.3916936815,-0.3313255608,-0.1392493248,-0.0491273478,-0.0511928275,0.0205049627,0.2427377552,-0.1608285457,-0.2102477401,-0.186390534,0.0076031582,0.1623863727,0.1766283214,0.0679415166,-0.1042773202,0.1803442389,-0.0273172539,0.2320993692,0.3803949058,-0.1308205575,0.2389179468,-0.1293284744,0.2053575963,-0.2161738425,-0.150118798,0.3080955744,0.2239318788,0.1004341543,0.442348212,0.0839701667,-0.0068221255,0.0487060696,-0.291867137,-0.0460131913,-0.2875136435,0.2097434849,0.3304284215,0.0942363366,-0.0393197648,-0.1384403408,-0.2583664954,-0.3234783709,0.1980765313,0.2372493148,-0.1127188951,-0.0732653886,-0.4805634618,0.3063938916,-0.6605846882,0.0469133481,0.1685167253,0.239718318,-0.2350105196,-0.0548119545,-0.090299949,0.0348619446,0.5300660729,-0.239096567,-0.1194755882,0.1587282866,0.1821472794,-0.1653635502,0.3114029467,-0.2298677266,-0.2269604802,0.2978044748,0.4735633135,-0.3214170337,0.0065116598,0.3584797978,0.3582867682,0.0707357079,-0.0315974392,-0.1629653871,-0.5779341459,-0.1653149426,-0.1091419905,0.223202318,0.2961626947,0.2642088234,-0.2336868942,0.2393158078,-0.1497079581,0.383500576,-0.1288926154,0.1693979651,0.4644765258,0.1934645772,-0.0260809101,-0.0012997846,0.1690479964,0.1843126416,0.1255097389,-0.3187140822,0.0965065584,-0.4157080352,0.0775485933,0.1177679822,-0.00813697,0.2131837159,-0.3196205795,0.3337206542,-0.237582773,0.4872477353,0.3090242445,0.0380470306,-0.3609592319,-0.226214394,0.3978452086,0.0200646203,0.1630872488,0.616381526,0.1752377003,-0.251721561,0.2007861137,-0.1126387492,1.0684710741,0.0045059831,0.149381727,0.498411566,0.1519502252,0.4557748735,-0.4085035026,0.091128923,0.3247399628,-0.2284524739,0.0470404513,0.1397041082,-0.1672049016,-0.085850656,-0.4916424453,-0.0697975308,-0.0202877764,-0.0910472795,0.0303605255,0.5705029368,0.0527481399,-0.454518795,-0.4049417973,0.2312329262,0.2329306602,0.1113296449,0.1414178908,-0.1698577255,-0.2318777889,-0.0795402229,-0.3921783268,-0.0085454723,-0.1524394751,-0.1035653353,-0.0707162023,-0.3387396634,0.1403520107,0.2051342279,0.4466013014,0.0653495565,-0.2453508526,0.1592052132,-0.3543911278,-0.2051929981,0.0209933911,-0.1798770279,0.3663350642,-0.3109208941,0.0339893028,0.2256439626,0.1066408604,-0.2248929292,-0.153149277,-0.2863200903,-0.1165719256,-0.1228956208,0.0417501107,0.4327595532,-0.2238211781,-0.1410496384,0.2016900033,-0.0093418574,-0.3379741609,-0.0702147782,0.1814672798,-0.166337654,-0.0131688397,0.2401930988,0.33086133,-0.0417894609,0.4671463966,0.2751145363,-0.1293351799,-0.2495833039,-0.225166738,-0.0619663931,-0.2933136821,0.0238815881,0.1427774876,0.024140723,0.0520797707,0.5436133146,0.0070556342,-0.1038081646,-0.1791743785,0.0176180825,-0.4704028368,0.0602729805,-0.1382642388,0.1104068458,-0.0739547908,0.4950857162,-0.0599422418,0.2071554661,-0.4044064879,0.0280526709,-0.1069911942,0.1881234199,0.3210683465,-0.1531080455,0.3761625886,-0.0098047378,0.1706484109,0.0929283202,-0.1838894933,-0.2533866465,-0.2344512343,0.0626202971,0.1027791649,-0.0843745098,0.0830771923,-0.1179246902,-0.288489908,-0.0309626833,0.3303532004,0.0796075836,-0.063919872,-0.00828424,0.0102663357,-0.0215128846,0.1327534765,-0.0385549217,0.1275246739,0.2622133195,0.0845993459,-0.0269373935,0.0792344213,-0.1881303936,-0.1115222797,0.2177870423,0.1796777248,-0.2291119695,0.2372367978,-0.1142953113,-0.092100665,0.1545267403,0.1713132411,-0.1187018678,-0.1974865198,0.0427705534,0.2852045596,0.2801287472,-0.3212323785,-0.02420366,0.1220372766,-0.0426601619,0.2153819799,0.0679904893,0.024606796,0.0851312354,0.0855639502,0.0616787411,0.0512073487,0.0290992297,0.1072217971,0.0947461277,0.0521772392,0.0136732096,-0.0881628245,-0.0236838292,0.14385885,-0.0494967699,-0.0420292914,0.4759680033,0.0604075454,0.0208893754,0.1528878063,-0.0818339139,-0.1796126366,0.2707375288,0.148234278,0.3688292503,-0.0102098389,0.5753251314,0.0780445412,-0.0832348093,-0.0323194824,0.3308324516,-0.0300841611,-0.105160363,-0.2205192149,-0.0847394615,-0.1759761274,-0.0009997488,0.1767510027,-0.0995973349,-0.3618557453,0.2024531364,-0.1753326356,-0.0102820052,-0.1182260588,0.0362013094,0.0217511356,-0.1454071403,0.3191003501,0.1787221134,-0.1570981741,0.0954256952,0.3117292821,0.3932819963,-0.034617655,-0.0325845145,0.3784401119,-0.0071038334,-0.0426684469,-0.1095701084,0.3652693033,0.3000549972,-0.1450001448,0.2046091259,0.1762261242,-0.3138993979,-0.0101081571,0.0198193658,-0.0988438204,-0.2141602933,-0.113359265,-0.0454506464,0.0276538432,-0.4261870384,-0.2730330229,-0.4859870076,0.2697599232,0.4382817447,-0.1422918141,0.0545795672,-0.08690162,0.1318080425,-0.3137738705,0.3943662047,0.3168160915,0.1392355859,0.0983059928,-0.4539985657,-0.444016844,0.3314886987,-0.1936247796,-0.3071641624,-0.1699742824,0.2537257075,0.4797628522,0.2120730579,-0.0576485731,-0.655850172,-0.1115338728,0.0202349536,-0.323879391,0.2111012936,-0.1695866287,0.1898113638,0.104422532,-0.122984454,-0.0006070281,0.028497614,0.1451640278,-0.4149310291,0.2897208929,-0.2087987065,0.0127343647,0.1909811497,-0.0722945333,0.2729309499,-0.3162320852,-0.1306893528,-0.1338270307,0.130764544,-0.2490925044,-0.201773122,0.159150824,0.3778850138,-0.1294214427,0.2802935243,-0.1098819152,0.3118391633,0.0115891406,-0.0270205345,-0.3153156042,0.2775222361,-0.1067406163,-0.1065659001,0.1228875294,0.4937791228,-0.0332551748,-0.0316835605,-0.2956219912,-0.5551957488,0.4337227643,-0.5333991051,-0.2663990557,0.1070915759,0.0360705033,0.1670577079,-0.0112620983,-0.7686284781,0.2083081901,0.2306216061,0.2212302685,-0.1843512505,0.3145614266,0.0685674623,-0.0572856292,-0.0314870439,0.1010667235,0.1317491233,-0.2307504863,-0.1234086975,-0.3093243539]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1836","title":"test.json has been removed from the limit dataset repo (breaks dataset)","comments":"Thanks for the heads up ! I'm opening a PR to fix that","body":"https:\/\/github.com\/huggingface\/datasets\/blob\/16042b233dbff2a7585110134e969204c69322c3\/datasets\/limit\/limit.py#L51\r\n\r\nThe URL is not valid anymore since test.json has been removed in master for some reason. Directly referencing the last commit works:\r\n\r\n`https:\/\/raw.githubusercontent.com\/ilmgut\/limit_dataset\/0707d3989cd8848f0f11527c77dcf168fefd2b23\/data`","comment_length":13,"text":"test.json has been removed from the limit dataset repo (breaks dataset)\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/16042b233dbff2a7585110134e969204c69322c3\/datasets\/limit\/limit.py#L51\r\n\r\nThe URL is not valid anymore since test.json has been removed in master for some reason. Directly referencing the last commit works:\r\n\r\n`https:\/\/raw.githubusercontent.com\/ilmgut\/limit_dataset\/0707d3989cd8848f0f11527c77dcf168fefd2b23\/data`\nThanks for the heads up ! I'm opening a PR to fix that","embeddings":[-0.0109288655,0.3094912767,-0.0923572779,-0.1577003598,0.0084777093,-0.0095927436,0.2062960416,0.4119395614,-0.1712233126,0.0565039739,0.1634871364,-0.1083242297,0.0013712744,0.2522862256,-0.0849231184,0.0299900435,-0.1627081931,-0.1398808807,0.0038428288,0.116309531,-0.1638636738,0.1403089613,-0.0626552477,0.2258493006,0.0069293617,0.1504515111,-0.0568243973,0.2602863014,-0.2913572192,-0.4906516075,0.3221632838,0.0938178003,-0.0771094114,0.4222998023,-0.0001061119,0.1205834076,0.5260327458,-0.0094663352,-0.3933196068,-0.5382500887,-0.3952264786,-0.0173074845,0.0842590183,-0.0278989356,0.1080243662,0.1597742736,0.069804728,0.0054836217,0.0923294276,0.5850142241,0.2820364237,0.1575161666,0.3259352446,-0.157308355,0.3321537077,0.0726541355,0.0406034514,0.461661309,0.162511602,-0.0407777913,0.2663975954,0.1521705836,0.1789362878,-0.0582365468,-0.0141852694,-0.1352257431,0.1967254132,-0.0314138308,0.2071001977,0.2375185043,0.3874586821,-0.2965471447,-0.2600788474,0.0534035712,-0.1950726211,-0.0113911629,0.2430444658,0.2144961357,0.101994969,0.1534056216,-0.3985633254,-0.3137033284,-0.1380166411,0.1481027901,-0.2509353459,0.2196244001,-0.0248466879,-0.0336641185,0.1893254817,-0.11387714,-0.3017702103,-0.094014734,-0.2380697578,0.0007901069,0.0516709574,-0.117126517,-0.007320174,-0.2611226141,0.5107996464,0.2271280885,0.0569382757,0.0075028385,-0.2453984618,0.0311023202,0.5669777989,0.1985512078,0.0545602888,0.0936508924,0.4472285509,0.4807001948,-0.0278277192,0.0552023835,0.1884819716,-0.2206877172,-0.0785264745,0.0270800702,0.4291928709,-0.3006854951,-0.1369591802,0.0573391952,-0.377530843,-0.1478599012,-0.0400414653,0.2185223401,-0.0605291426,-0.0449661985,0.0088302726,0.2919821739,0.0001437468,-0.0411993265,-0.2616354823,0.1080471054,-0.0455313846,-0.2621949613,0.1331668794,-0.1538153887,0.5269491076,-0.1204788834,-0.002113529,-0.2623295784,-0.0427385755,0.105382584,-0.0251729991,0.3528271317,0.050725542,0.1310840994,0.1005339175,-0.4289901257,-0.135202378,0.0400038213,-0.2719546258,-0.5531669259,0.0009175905,0.2811228633,-0.2128613889,-0.0944407359,-0.3993630707,-0.1957691014,-0.085196577,0.1919860542,0.0110858381,-0.2623322308,0.2691375017,-0.1167709976,0.1521477103,0.268990308,-0.1116796881,-0.2156557739,-0.1785464436,-0.1178478748,-0.0249044336,0.2013214529,-0.3774856329,0.0942756459,-0.3569557965,0.0745093077,0.0936129466,-0.4332113564,-0.3620328605,0.3549993336,-0.2183541358,-0.134966433,-0.445122242,-0.3316715658,0.1716037691,-0.1994174123,0.2475769222,0.1351167113,0.2178747356,-0.0541514196,-0.1856604367,-0.3220585287,-0.0999715403,0.0240180586,0.0678933635,0.1879159212,0.3300004005,0.1608101726,0.2576721907,0.1269006878,0.121371448,0.4919353426,0.4691879153,-0.1812723875,-0.1384885758,-0.0480864905,-0.2049775422,0.0478253625,0.0444182493,0.1771543622,-0.1456756145,-0.1148471683,-0.4425048828,0.1475944817,-0.2608279884,-0.3512989283,0.1706815809,-0.0422722921,-0.0223519448,0.2020108253,-0.106998235,-0.0363016687,-0.3412429988,0.145683676,-0.3127666712,0.3656833768,-0.2764018476,-0.016810799,0.4088414907,0.1403151453,0.1743165553,-0.4291144311,-0.156452179,0.5749092102,0.0149997873,0.1195717752,0.2236797363,0.0599149689,0.4830204546,-0.2898099422,-0.1367132813,0.3694257736,-0.1739999503,0.1256079078,-0.3294793367,0.3566485345,-0.0365233757,0.0932560265,0.2833323777,0.0010704747,0.4858244956,-0.1946219802,-0.0800640956,-0.2613082528,0.0886043236,0.3499429822,-0.1226847246,-0.2618888021,-0.0123857474,0.0515027158,0.2603545189,-0.2127923369,0.0852127671,0.2709658742,0.0145884017,-0.2358141094,-0.0455459505,0.5093911886,0.257470727,0.0554484092,0.1827999949,0.0744915083,0.0678588673,-0.2791389823,0.27130723,0.0313188955,-0.1123050079,0.0818642229,0.3418090045,-0.3010960817,-0.2481990308,-0.0930112526,-0.0280954931,0.3554309905,-0.0338754244,-0.0117664123,-0.5418950915,-0.1253025979,0.064822346,-0.4101050794,-0.539716661,-0.347805053,0.1078453362,0.1616818607,-0.132179603,0.1826475263,-0.0258414689,-0.0624317154,0.0643332154,-0.1983563751,0.0704485103,-0.1431531161,-0.0911361948,0.109959811,0.001978867,-0.0543283336,-0.0849298686,-0.1141414866,-0.2554239929,-0.5308839083,-0.3663043976,0.1227355972,-0.0110711586,0.2815186381,0.3721878827,0.1439297348,0.3151711226,0.1866136342,0.2043242157,-0.324347645,-0.0622855462,0.0899262354,0.1096202433,0.0755640343,-0.1927650124,-0.220176667,-0.0916307196,-0.2294224501,0.1631869078,0.2451374382,0.0261211172,0.1684898436,0.1712942421,0.0902349055,0.1712110639,-0.3185037374,-0.2599452436,-0.3976639509,0.086782217,-0.1013064012,-0.4773408473,0.2257731557,0.1427584141,0.2746831477,-0.0347239412,-0.5702361465,-0.3527767658,0.0644860491,0.1823917776,0.1284661144,0.0633369833,0.1092003062,-0.1337751448,-0.1903948188,-0.1130131111,-0.3238709867,0.2828414738,0.2126527876,0.2966175973,0.0007706246,0.5261009336,0.1415368319,0.449341476,0.1688037813,-0.0615642928,0.2201280445,-0.1427747458,0.288998574,-0.0491861701,-0.0878536999,0.1688676924,-0.013872548,0.0243352186,0.1461009383,0.2546213567,0.087127775,-0.0880391449,-0.1174212694,-0.2183206379,-0.2936145663,-0.2273040861,0.1381081194,0.1822808236,0.3110196292,-0.1270979047,-0.0488390252,-0.011076075,0.1149760857,0.4667929113,0.0607835911,-0.0415372849,-0.589912653,-0.0311663039,-0.2651078105,0.2900630832,0.0204684474,0.4159531295,-0.0500770323,0.0928483829,0.0314807408,-0.177598089,0.5821406841,-0.2425104231,0.3473289311,0.1354245842,-0.191878587,-0.0157043785,0.0268517248,0.0974649787,0.3108721972,0.5515390038,0.3271017671,-0.308629781,-0.3085367978,0.2392790169,-0.1225758195,-0.028272355,0.200987637,-0.2288399041,-0.484852016,-0.3253540695,-0.2336674929,0.0102700284,-0.0393101349,-0.1687666923,-0.0610628687,0.1820582747,-0.3098095953,-0.264962554,-0.0987993255,0.1658854932,-0.0072225523,-0.0860045031,-0.0429527797,-0.049972441,0.4645078182,0.4625291824,0.1494679153,-0.0397058092,0.0595114194,0.0151343476,-0.0645753965,0.281057179,0.0896966904,0.198625043,0.0682463571,0.1683229804,0.1001155972,0.065027386,0.1164663807,-0.2811867893,-0.2859715521,-0.2276241332,0.165522173,-0.1367802471,0.0915333331,0.2283719033,0.0467083268,-0.405441314,0.014412553,0.0086572887,0.6994220614,0.0704677925,-0.0272702333,-0.0421293415,-0.2424473464,0.2048669457,-0.033239089,0.0917162523,-0.2030578107,0.1646093726,-0.0740504116,-0.1837385893,0.030137226,-0.1313894242,-0.1530456543,0.1965422779,-0.0590464883,0.4375255704,0.1689644307,0.3466405272,-0.2862257063,0.3009475172,-0.3785423934,0.2891438901,-0.1375840455,0.2199976295,-0.2258019447,-0.2631915212,-0.0212027505,-0.2012030035,-0.2781691253,0.231305927,-0.1082002297,0.1171493456,0.0271635205,-0.3534875512,-0.0126809571,-0.1995676905,0.32215029,0.1351676732,-0.2993149459,0.3058156073,-0.3631394506,-0.0892225504,0.0837844685,0.2514723539,-0.0081039844,-0.2934261858,-0.3303700686,0.1814527065,0.0651182309,-0.6371983886,0.020105144,0.0400764868,-0.2511359453,-0.1447103769,0.1178139895,-0.0149181746,-0.1354431212,-0.4786892831,0.1484293938,0.1727302969,-0.0247397162,-0.3258293271,-0.0368396081,-0.195195809,-0.2569152713,0.3153433502,-0.0581993498,0.3233926892,0.3248978853,0.1317464858,0.2382626534,-0.2811474204,0.1129432842,0.1196505725,-0.2088673711,0.292794615,-0.5139477253,-0.023795845,-0.0481466912,0.4000206888,0.0857402757,0.1701927036,0.0942142233,-0.5861810446,-0.1788284481,0.08888264,-0.2860866487,0.0958690494,-0.2721700668,0.2587830722,0.0280465409,-0.1592685133,-0.3539270759,0.3791739643,-0.2649120688,0.0868104473,-0.0027639046,-0.1797840148,0.3934028745,-0.2188183963,0.1120001674,-0.1318623871,-0.3064839542,-0.2311203778,-0.0850260332,0.0675817505,0.0053646578,-0.2011274695,-0.32207039,-0.2382813394,-0.1297645569,0.2546701431,0.0817712918,-0.0744861364,0.1125228852,0.0771312192,0.1622006297,-0.2705673873,-0.001593704,-0.0886433646,0.1006777138,0.3166829348,0.118542105,0.1584345698,-0.1172731146,-0.076097928,-0.3107618392,0.1530896276,-0.2145283818,0.5459501743,0.2931928039,-0.31489411,-0.1441551298,0.0749055743,0.5291779041,0.2539230585,-0.2609059215,0.1553837359,0.0854765698,0.2261523157,-0.0199703481,-0.1492171586,0.2679268122,0.1190499291,0.0158060119,0.0304123443,0.1372587383,0.1890201122,0.183738634,0.4024943709,0.5694633126,-0.0247551147,0.104565829,0.3553735614,-0.157377094,-0.0156124895,0.1619163454,0.2314712852,0.3316363096,0.0761202797,0.2722610831,0.3893729448,0.3265689015,0.1007167026,0.1590490937,-0.101163663,-0.0888148844,0.1637147665,-0.043521218,0.0312800631,0.3399223387,0.2798509896,-0.2327384204,-0.1636291146,-0.2075873017,0.1342456341,0.1537656039,-0.203067556,-0.3722724319,-0.2799301147,-0.0614021309,0.133614704,-0.1129373685,-0.207594201,-0.0265551917,-0.0458057672,-0.2575523555,-0.4632114768,0.0654739514,0.1013161987,0.0982346386,-0.0231095906,0.060248442,0.1976893097,-0.1393025815,0.2267959267,0.1657529771,0.3838031292,-0.1215742007,0.0192054026,0.0680160969,0.1153089553,-0.1822448522,0.0561938882,0.3214838803,0.2291706949,-0.1246004328,0.161004737,0.2458321899,-0.1561207026,0.2781552076,0.1853844076,-0.0805150568,-0.3243801892,0.5787655115,-0.0680500865,-0.1999376714,-0.132138595,-0.088592656,-0.2192539126,0.0633262768,0.1608342826,-0.1583397686,0.1846741736,-0.0972876251,0.1495916098,-0.0559928156,0.2134156078,0.4517925382,0.0750283226,0.0014947841,-0.3916741908,-0.4470764995,0.1156279147,-0.2165634483,-0.1180246994,0.2384563684,0.0655967891,-0.070093289,-0.0125042591,0.1517870426,-0.0245846398,-0.1677528024,-0.1339065731,-0.1759207547,-0.0516440757,-0.0537563339,0.20813106,-0.050736919,-0.1571603417,0.2786634862,-0.3101501167,0.1425068676,0.0161406733,0.1383268833,0.0456937253,-0.2053636909,0.2786599398,0.2197556794,-0.2630351782,-0.0586787798,-0.2688447535,-0.3487125933,-0.279376179,-0.1257323027,0.05129309,-0.1881134361,0.3420862556,0.1545470953,-0.0708343908,-0.3823770881,0.4479513168,0.0576333515,-0.0838210955,-0.1397755742,0.2395510823,-0.0153982127,-0.0265323445,0.1257272065,0.2953270078,-0.0996501893,0.1627531052,-0.2005624324,-0.0376287252,0.4760784507,-0.31350559,-0.1579256803,0.1618402451,0.3197045028,-0.1036342755,-0.3652911186,-0.2992825508,0.1931446344,0.3190315962,0.1394722164,-0.6502301097,0.1478154063,-0.0819335356,0.0349609628,0.2211327851,0.2730174959,0.347697705,-0.4209606349,0.2150211185,0.0707149282]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1831","title":"Some question about raw dataset download info in the project .","comments":"Hi ! The `dl_manager` is a `DownloadManager` object and is responsible for downloading the raw data files.\r\nIt is used by dataset builders in their `_split_generators` method to download the raw data files that are necessary to build the datasets splits.\r\n\r\nThe `Conll2003` class is a dataset builder, and so you can download all the raw data files by calling `_split_generators` with a download manager:\r\n```python\r\nfrom datasets import DownloadManager\r\nfrom datasets.load import import_main_class\r\n\r\nconll2003_builder = import_main_class(...)\r\n\r\ndl_manager = DownloadManager()\r\nsplis_generators = conll2003_builder._split_generators(dl_manager)\r\n```\r\n\r\nThen you can see what files have been downloaded with\r\n```python\r\ndl_manager.get_recorded_sizes_checksums()\r\n```\r\nIt returns a dictionary with the format {url: {num_bytes: int, checksum: str}}\r\n\r\nThen you can get the actual location of the downloaded files with\r\n```python\r\nfrom datasets import cached_path\r\n\r\nlocal_path_to_downloaded_file = cached_path(url)\r\n```\r\n\r\n------------------\r\n\r\nNote that you can also get the urls from the Dataset object:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nconll2003 = load_dataset(\"conll2003\")\r\nprint(conll2003[\"train\"].download_checksums)\r\n```\r\nIt returns the same dictionary with the format {url: {num_bytes: int, checksum: str}}","body":"Hi , i review the code in \r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py\r\nin the _split_generators function is the truly logic of download raw datasets with dl_manager\r\nand use Conll2003 cls by use import_main_class in load_dataset function\r\nMy question is that , with this logic it seems that i can not have the raw dataset download location\r\nin variable in downloaded_files in _split_generators.\r\nIf someone also want use huggingface datasets as raw dataset downloader,\r\nhow can he retrieve the raw dataset download path from attributes in \r\ndatasets.dataset_dict.DatasetDict ?","comment_length":166,"text":"Some question about raw dataset download info in the project .\nHi , i review the code in \r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py\r\nin the _split_generators function is the truly logic of download raw datasets with dl_manager\r\nand use Conll2003 cls by use import_main_class in load_dataset function\r\nMy question is that , with this logic it seems that i can not have the raw dataset download location\r\nin variable in downloaded_files in _split_generators.\r\nIf someone also want use huggingface datasets as raw dataset downloader,\r\nhow can he retrieve the raw dataset download path from attributes in \r\ndatasets.dataset_dict.DatasetDict ?\nHi ! The `dl_manager` is a `DownloadManager` object and is responsible for downloading the raw data files.\r\nIt is used by dataset builders in their `_split_generators` method to download the raw data files that are necessary to build the datasets splits.\r\n\r\nThe `Conll2003` class is a dataset builder, and so you can download all the raw data files by calling `_split_generators` with a download manager:\r\n```python\r\nfrom datasets import DownloadManager\r\nfrom datasets.load import import_main_class\r\n\r\nconll2003_builder = import_main_class(...)\r\n\r\ndl_manager = DownloadManager()\r\nsplis_generators = conll2003_builder._split_generators(dl_manager)\r\n```\r\n\r\nThen you can see what files have been downloaded with\r\n```python\r\ndl_manager.get_recorded_sizes_checksums()\r\n```\r\nIt returns a dictionary with the format {url: {num_bytes: int, checksum: str}}\r\n\r\nThen you can get the actual location of the downloaded files with\r\n```python\r\nfrom datasets import cached_path\r\n\r\nlocal_path_to_downloaded_file = cached_path(url)\r\n```\r\n\r\n------------------\r\n\r\nNote that you can also get the urls from the Dataset object:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nconll2003 = load_dataset(\"conll2003\")\r\nprint(conll2003[\"train\"].download_checksums)\r\n```\r\nIt returns the same dictionary with the format {url: {num_bytes: int, checksum: str}}","embeddings":[-0.1392953694,-0.0969949961,-0.0575437061,0.5315549374,0.1806542575,-0.0776938498,0.1291483194,-0.0684930757,0.1307476461,0.1263669282,-0.4010899365,0.1903991401,-0.0717179701,0.5283398628,0.1642400473,-0.0650983974,-0.1106831804,0.1475783736,-0.0933762714,-0.073043488,-0.3944445848,0.2827210128,-0.3126272559,0.1286266744,-0.1304444373,0.005232105,-0.0375590362,0.2881941497,-0.4633617699,-0.2766463757,0.3559412956,0.2283115536,0.4103102088,0.2549616396,-0.0001227949,0.0453443974,0.2947669923,-0.3216656446,-0.5304590464,-0.3504639268,-0.6374027729,-0.0066081779,-0.0876589119,-0.3026112616,0.0727007687,-0.0444140546,-0.070448868,-0.2377196699,0.1179704964,0.1020431444,0.0789081827,0.2801706195,-0.0258048195,0.0432848558,0.284897089,0.280359298,0.0525379293,0.1077032536,-0.0747682676,0.4011949599,0.2505822778,0.3100157976,0.1377087682,0.1574471444,0.5022882223,0.3205779493,-0.2726316154,-0.3046389818,0.1046134606,0.1935164928,0.4040815234,-0.3673968315,-0.5436560512,-0.3483605981,-0.1683370918,-0.5926069021,0.0322045647,0.1467377096,-0.364554137,0.3069461882,-0.5524965525,-0.4242562354,-0.1069234833,-0.0295140743,-0.1963638216,0.3718147278,-0.3595739901,0.0653247014,0.0780688748,0.4294739366,0.1998236477,-0.2933289707,-0.0176404715,0.2759217024,-0.172680676,-0.0890779421,-0.0210981928,0.3046434224,0.191677928,-0.0031788051,-0.1185906902,-0.19603239,0.064478837,0.0911550969,0.3602831364,0.1348402798,0.1220054775,0.271299839,0.2210540175,0.1582034677,0.1628750563,-0.2994869947,-0.0399511047,-0.0499701649,-0.3093303144,0.1616626084,0.0709266588,-0.122664623,-0.1993661374,-0.2141618729,-0.1070422083,-0.2923630178,0.3221350908,0.4048178792,-0.112061955,0.0731321871,-0.0197029356,0.0484817922,0.1325499117,-0.2103043348,-0.1068648547,-0.1425915211,-0.1424525976,0.1309209466,-0.2759810388,-0.0752524063,0.3806269765,0.0523516014,0.0155299008,0.0070278249,0.020547647,-0.1392773241,0.2036897093,0.1511138529,-0.0131356213,0.3002581596,-0.1582263708,-0.3350555897,-0.1950610578,-0.0940052196,-0.3649282455,-0.4216024578,0.0158243943,0.0255181789,-0.2253227979,0.1036860868,-0.3064567745,-0.0258364938,0.1035278887,0.3552072942,0.0736041069,-0.0413041785,-0.4622364342,-0.2665116787,0.4400364161,0.481897682,-0.2481839955,-0.1428897679,0.0488598049,-0.3707982302,0.1104975864,0.0247115418,-0.3937753737,0.5463496447,-0.411098212,-0.1288508028,0.4111630619,-0.4205340147,-0.274371475,0.5594000816,0.1189885736,-0.0040633497,0.3508140147,-0.0029530041,0.2521328032,-0.0863127857,-0.1183066815,0.5005720854,0.1785057485,-0.1154285073,-0.0150466897,-0.1998094022,-0.150423497,-0.1076831594,-0.1324987113,0.2510327101,0.0363789052,-0.0930608064,0.5309388041,-0.1145965233,0.1999672502,0.3154318035,0.0265686102,0.2461875826,-0.0191205963,0.0292345788,-0.6519911289,0.1660543978,-0.3580187559,-0.2422895133,-0.3038727343,-0.3468744755,-0.3359973133,0.152637884,-0.2951542735,-0.3711603582,-0.0584716909,0.1649305075,0.4927286208,-0.0335013196,-0.3086248338,0.5376631021,-0.0771123543,0.1103712842,-0.1524742693,0.0215066615,0.1099479273,0.1840830594,-0.0664960295,0.1019263789,0.1083981693,-0.1885489374,0.0598816797,0.5309789181,0.2396335304,0.181833297,0.154087469,0.0292210523,0.2496578097,0.1244924366,-0.0728938803,-0.2197057754,0.2130739838,0.2021097839,0.0957582444,0.4194434881,0.0039161164,0.2770685554,0.1396989077,-0.078147687,0.1378965229,-0.0618099682,0.1554043293,-0.1600990444,-0.1753717214,0.2047282308,0.3687287271,0.1324179322,-0.2014482766,0.1954443306,0.3634999692,-0.0819677263,-0.2479622066,0.0165177435,-0.2928124666,-0.0403007418,0.0202466343,0.78809762,0.4842380881,0.009911241,0.1930085868,0.0817181617,0.2897280455,-0.1300808638,0.4044666588,0.1461005658,-0.0709935278,-0.033552099,-0.4765409827,-0.1736866087,0.1612807214,-0.0182557534,0.1560421586,0.1374084949,-0.3770774007,0.1228537112,-0.2453245223,-0.3593226373,-0.2825257182,-0.182490766,0.0796568468,-0.1788537651,-0.3023648858,-0.0960525125,-0.086574018,-0.1961852759,-0.0284911226,0.0078612622,-0.2690333128,0.0346265137,-0.1725916415,-0.1846226007,-0.1305118501,-0.0738441944,0.1662775278,0.0342106074,0.0837573633,-0.4706437588,0.0689177811,-0.16951482,0.17596367,0.324734807,0.0570176393,0.1776961833,0.3906327784,0.1428868771,-0.1417596787,0.0228357054,-0.0706352517,0.1046682373,-0.1442241222,-0.3927357793,0.1336805969,0.0608388707,-0.1003025547,-0.1797951609,-0.701849699,-0.0942709669,0.1484006345,-0.1141361371,0.2285626233,-0.063475661,-0.1833595335,0.1142905504,-0.273940444,0.0497957654,-0.234608233,-0.7871106863,0.2894405723,-0.3476832509,-0.266212672,-0.0440538488,-0.0977244303,0.1984011531,0.01549144,-0.2537752688,-0.0317412391,-0.2202959359,0.171145469,0.1069121584,0.3687538207,0.0127614979,-0.1026366353,0.1300041527,-0.1975700706,-0.0617833212,0.1827081293,-0.3824943006,0.3085665405,0.3824217319,-0.0335212052,0.1748531759,0.7821301818,0.2444395274,-0.1355072707,0.1809847653,-0.0758559927,0.4101742506,0.2839820981,-0.2186225802,-0.0489328243,-0.0339052193,-0.107848458,0.144003287,0.060469918,0.1896632314,-0.3165882528,-0.1123692617,-0.1500528753,-0.1517049372,-0.12551561,0.5231513381,0.1893754452,0.0615993142,0.1974451989,-0.006582329,0.0912793279,0.0618570521,0.4306045473,0.3460768461,0.1975971907,-0.1960721314,-0.0857728869,-0.2044927627,0.1164887175,0.2858866155,-0.1161119714,0.1796921343,0.1200830638,0.1790281981,0.1177058145,0.6446791887,0.0226204805,-0.0401950367,-0.1275093406,-0.3101218343,-0.076078929,0.1280727834,-0.1500325352,0.0737530887,0.4213221669,0.1637877673,-0.4036840498,-0.3825345039,0.0745411292,0.1352339089,-0.2881725729,-0.1495440751,-0.1488921791,-0.2091282606,-0.1788700819,-0.2308760583,0.1296275109,0.1278059334,0.0346970707,0.0025444718,0.0505958647,0.1193702146,0.0030728362,0.0747295618,0.3675093055,-0.0561991297,-0.1057411358,0.3754268885,0.3393510878,0.1690374762,0.4490528107,-0.3288882375,-0.1851526648,-0.0648366213,0.0975713953,0.6481601596,0.1496813148,-0.489810735,0.0493488275,0.0348544568,-0.1490059942,-0.3897596896,0.2432527244,0.2638326883,0.0071812272,-0.0993559957,-0.7190077901,0.4855560958,-0.00733996,-0.2285361886,0.1606065929,0.0833797231,-0.3551582694,-0.1382281482,0.0349195302,0.819853127,0.1342671961,0.0822530314,0.1007773355,-0.4830279052,0.5889974236,-0.0602166019,-0.1731592864,-0.2225249112,0.0222349856,-0.1520213932,-0.2860232592,0.0705218613,-0.0760802552,-0.1098827124,0.3645530641,-0.0983016118,0.0498235635,0.0435956493,0.5104305148,-0.0831178054,-0.1409682781,0.0753675178,-0.0840002447,0.1918082982,0.4800057709,-0.114002876,-0.4310254157,-0.2550567687,0.0795700476,-0.0654510036,-0.1017837003,-0.4333568513,0.0302237105,0.0133735901,-0.3217694759,-0.0371241532,0.6608507037,0.2790494859,0.0164364427,-0.2199330926,0.0636703596,-0.0920160264,0.127336815,0.2503714263,0.1239112616,-0.0576398447,-0.0534837507,-0.1629353762,0.1557441354,-0.0058320803,-0.1810483187,0.1748495251,-0.1016854644,0.2125195116,-0.250741303,-0.0561154783,-0.0108205983,0.062607266,-0.0681609586,-0.0006634277,0.2437827587,-0.3698791862,0.0706730783,-0.2798644006,-0.4227114916,-0.1907727569,0.360604465,-0.0719441846,0.0788180009,0.2684695125,-0.1025508866,-0.2249783874,-0.0309097227,0.288657099,-0.1091570482,-0.3159406483,0.2659707367,0.3049023151,0.1788382381,-0.4056986868,0.2560827732,0.0011078222,-0.281742692,-0.2222938389,-0.4993056357,-0.2070183158,0.1097566336,-0.2958486974,0.1551875621,-0.012738023,0.1706501991,0.1040893197,0.0283844396,-0.195524767,0.0157606807,-0.0842282623,-0.1064331159,0.337510258,0.0971307009,0.2482390553,-0.003077623,-0.0589603297,-0.0390096977,-0.0900466442,-0.0314556621,-0.2268635482,0.1887499243,-0.0839849859,-0.2017305344,0.0876312479,-0.0617088974,-0.0786994994,0.0590821728,0.2508067191,0.2661225498,0.129343614,0.3978659511,-0.0577159785,-0.0034345011,-0.5778408647,0.335506469,0.0692228153,0.4035030901,-0.0779304579,0.4459483922,-0.3849085271,-0.1046703458,0.1254520416,-0.1211467907,0.0126707098,-0.0886189491,0.4348567724,-0.0435952134,0.1081254482,0.1174717844,0.2370528728,0.2382538468,-0.089698799,0.0296730418,0.1940105557,-0.0273523461,-0.0085727423,0.0239256807,-0.1370310932,-0.0406361744,0.024102509,0.3481292427,0.3333757222,0.0385599434,-0.0778069869,-0.2710223794,0.401126653,0.4406712353,0.046021387,0.5709141493,0.1283136755,0.3627936542,0.211736992,0.6090320349,0.156406045,0.3393489718,-0.2610030472,-0.0089800246,-0.08741007,0.2450071722,0.3107307553,-0.4485318959,0.1722418219,0.2022880465,-0.2262049615,0.0219259895,-0.299412787,0.5222713947,-0.1551600546,0.1328142285,-0.1060105786,0.4710695148,0.0457520895,-0.2884922922,-0.1762336195,-0.1881162524,-0.1317508668,0.1523735821,-0.0461350009,-0.4370724261,0.1362341642,-0.1464451998,0.0427229404,0.0183533672,0.0530934222,0.1040964797,0.021595154,-0.2122959942,0.1915363818,-0.0416562259,0.2909667492,0.3288476467,0.2829504609,0.0673963726,0.3388293087,0.1462308615,-0.2136947066,0.2214282155,-0.0688987672,-0.0874905437,0.0249388833,-0.2010525465,-0.0747039169,0.4496802986,-0.0103028314,-0.0228589419,-0.3053920269,0.1411831975,0.3131197989,-0.0141031761,0.5945333242,-0.0259308666,0.0670638904,0.0925704688,0.0610271916,-0.4617903233,0.0695710853,0.2748192549,-0.2647775412,0.2141080201,-0.0019733135,0.0121392729,-0.0241732616,0.371170193,0.2092071623,0.3382583559,-0.4282611609,-0.2127803415,-0.4922032058,-0.2341073006,0.3432761133,-0.2409981936,-0.0614523292,0.3055997193,-0.222242415,0.2440619171,-0.0424089208,0.2761822045,-0.0096673546,0.0954615474,-0.0504731387,-0.1580987871,-0.3776186407,0.1484346688,0.1120984778,-0.2221130729,0.3392997086,-0.2188185304,-0.1105876788,-0.0895138979,0.2348069549,0.1685887575,-0.2278386354,0.1703927815,-0.017974643,0.14916192,-0.060574539,0.0319153406,0.251495868,0.2419125438,-0.2968153656,-0.0082007367,-0.1475500911,-0.1339500695,-0.3201524317,0.0209584199,-0.4668420553,0.3759160638,0.3198999166,0.0002513771,-0.0486250296,-0.0242109671,-0.4064563811,0.0382361375,0.5310792923,0.1629974544,-0.2717091739,0.0698959157,-0.2187839895,-0.0578167364,0.5587483644,-0.2671597004,0.1210986376,0.0403500199,0.2279321551,0.3333331048,-0.0913674682,-0.5478304029,0.2801440656,0.5396885872,-0.0836962685,0.03364674,0.2122835219,0.0887805521,-0.0016809605,-0.1219748631,0.4751988947,0.1540817618,-0.3386337757,0.0972762629,-0.0298983585]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1831","title":"Some question about raw dataset download info in the project .","comments":"I am afraid that there is not a very straightforward way to get that location.\r\n\r\nAnother option, from _split_generators would be to use:\r\n- `dl_manager._download_config.cache_dir` to get the directory where all the raw downloaded files are:\r\n ```python\r\n download_dir = dl_manager._download_config.cache_dir\r\n ```\r\n- the function `datasets.utils.file_utils.hash_url_to_filename` to get the filenames of the raw downloaded files:\r\n ```python\r\n filenames = [hash_url_to_filename(url) for url in urls_to_download.values()]\r\n ```\r\nTherefore the complete path to the raw downloaded files would be the join of both:\r\n```python\r\ndownloaded_paths = [os.path.join(download_dir, filename) for filename in filenames]\r\n```\r\n\r\nMaybe it would be interesting to make these paths accessible more easily. I could work on this. What do you think, @lhoestq ?","body":"Hi , i review the code in \r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py\r\nin the _split_generators function is the truly logic of download raw datasets with dl_manager\r\nand use Conll2003 cls by use import_main_class in load_dataset function\r\nMy question is that , with this logic it seems that i can not have the raw dataset download location\r\nin variable in downloaded_files in _split_generators.\r\nIf someone also want use huggingface datasets as raw dataset downloader,\r\nhow can he retrieve the raw dataset download path from attributes in \r\ndatasets.dataset_dict.DatasetDict ?","comment_length":111,"text":"Some question about raw dataset download info in the project .\nHi , i review the code in \r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py\r\nin the _split_generators function is the truly logic of download raw datasets with dl_manager\r\nand use Conll2003 cls by use import_main_class in load_dataset function\r\nMy question is that , with this logic it seems that i can not have the raw dataset download location\r\nin variable in downloaded_files in _split_generators.\r\nIf someone also want use huggingface datasets as raw dataset downloader,\r\nhow can he retrieve the raw dataset download path from attributes in \r\ndatasets.dataset_dict.DatasetDict ?\nI am afraid that there is not a very straightforward way to get that location.\r\n\r\nAnother option, from _split_generators would be to use:\r\n- `dl_manager._download_config.cache_dir` to get the directory where all the raw downloaded files are:\r\n ```python\r\n download_dir = dl_manager._download_config.cache_dir\r\n ```\r\n- the function `datasets.utils.file_utils.hash_url_to_filename` to get the filenames of the raw downloaded files:\r\n ```python\r\n filenames = [hash_url_to_filename(url) for url in urls_to_download.values()]\r\n ```\r\nTherefore the complete path to the raw downloaded files would be the join of both:\r\n```python\r\ndownloaded_paths = [os.path.join(download_dir, filename) for filename in filenames]\r\n```\r\n\r\nMaybe it would be interesting to make these paths accessible more easily. I could work on this. What do you think, @lhoestq ?","embeddings":[-0.1346984506,-0.0291421283,-0.0565953217,0.5307526588,0.118836984,-0.1144693121,0.1020086557,-0.0082435263,0.0415084288,0.0833945125,-0.4197905958,0.207283318,-0.1245639473,0.5449789166,0.1708198935,-0.0960756987,-0.0993765071,0.1221581697,-0.1015595347,-0.0675221384,-0.3862855434,0.2964140475,-0.2933220863,0.1700219661,-0.1204365417,-0.0437129624,-0.0389077775,0.3107964396,-0.4215294421,-0.2855050862,0.2851445973,0.2104239166,0.38972646,0.2601925135,-0.0001252634,0.0339560695,0.2576283216,-0.3376943469,-0.5291388035,-0.3045315742,-0.6149904728,-0.0184274614,-0.0735696927,-0.2560712695,0.1428723633,-0.0621147715,-0.0816688538,-0.1980143338,0.0248030238,0.0257190745,0.0569270216,0.3073020279,-0.0705493763,0.0733200833,0.2743446529,0.2889507413,0.0069416389,0.1065994874,-0.0094362842,0.3097071648,0.1909423769,0.3141844571,0.1917632371,0.1161840111,0.4544143081,0.3511873484,-0.3605449498,-0.3333880901,0.0019012884,0.2414935082,0.3311742246,-0.3499562144,-0.6002609134,-0.4451579154,-0.1447591633,-0.5862122178,0.0866652057,0.1451350749,-0.4294036329,0.2783696651,-0.5737096667,-0.470205754,-0.107822679,-0.0630755946,-0.1245901063,0.3833278418,-0.3021124601,0.0366093814,0.1222617403,0.366537869,0.1831099838,-0.2040058076,0.0064977114,0.2628435791,-0.1273599416,-0.0968480706,-0.0692084581,0.2800370157,0.1793465018,0.0004258784,-0.0932023153,-0.2625410855,-0.0224481337,0.0367786027,0.3256850243,0.0887944326,0.0588160753,0.2508298755,0.2419353873,0.1852357239,0.2261740714,-0.2565537691,-0.0756491125,-0.0928039774,-0.3982129097,0.1689798385,0.023065459,-0.0881559849,-0.2564201951,-0.2584711909,-0.117320776,-0.3153946102,0.3284130096,0.3631953597,-0.1064461619,0.0919000506,-0.0218641907,0.0300095454,0.1342038512,-0.2513070703,-0.0602543615,-0.1366773695,-0.1498392075,0.1586862206,-0.2355437279,-0.0740527585,0.3717945516,0.0417128019,0.0303050894,0.0919628292,-0.0051716962,-0.1027109027,0.1987164915,0.1128025353,0.0319495089,0.2906243801,-0.2086739689,-0.2854421735,-0.2164021283,-0.1238752306,-0.2959266603,-0.4238964915,0.010931341,-0.026596522,-0.1496953666,0.1380343139,-0.2426622361,0.0063026431,0.0641364083,0.3371734619,0.0211319569,0.006673615,-0.4359383285,-0.2875217795,0.4370478988,0.4985270202,-0.2790235579,-0.1094861701,0.091829516,-0.3653411865,0.0410041586,0.0893710405,-0.3769408464,0.526019156,-0.4255958796,-0.1558313668,0.4934948087,-0.4139986038,-0.2834487259,0.548224926,0.0819413885,-0.0119388681,0.3946042657,-0.0027310946,0.2809688449,-0.0996628776,-0.06136683,0.5419443846,0.1237987727,-0.1109535769,-0.0387092419,-0.2857447565,-0.1188867688,-0.0995088667,-0.1401036531,0.2085418105,0.0630515516,-0.1487575024,0.5694390535,-0.0648537576,0.2024183124,0.3641732931,0.016141776,0.2217395306,-0.039749153,0.032757625,-0.6237203479,0.1450575739,-0.291588068,-0.2566764951,-0.3464303017,-0.3692541122,-0.2835113108,0.1359041929,-0.2205851376,-0.3565536439,-0.0813381076,0.1932128072,0.4799261391,-0.0377768092,-0.3269172013,0.5276400447,-0.0264157411,0.1004448459,-0.1584222466,0.0395960882,0.1531455368,0.1568753421,-0.0285879895,0.1156432182,0.1160011515,-0.2031235099,0.0879380926,0.4732969105,0.1667213589,0.2244895995,0.2488667816,0.0866522789,0.2675855458,0.1815995276,-0.050037697,-0.2840552032,0.1166388243,0.2693530917,0.0217542127,0.405077666,-0.0402296744,0.2749062181,0.2071458101,-0.0586074106,0.1375477016,-0.0461245924,0.1909854561,-0.1717652828,-0.1450582147,0.1957189739,0.3370574415,0.1461298615,-0.2831913829,0.2967513204,0.4335455,-0.0422255136,-0.2803872824,0.0072161951,-0.2794591486,-0.0143303955,0.0462974794,0.7960909009,0.4438539743,0.0230253264,0.191422984,0.0300603881,0.2835537791,-0.0816387832,0.3543806672,0.1328087896,-0.080934912,-0.019625986,-0.4542382061,-0.1663334519,0.193083778,-0.0439952202,0.1845147312,0.1720055491,-0.3527051508,0.0789799988,-0.243078649,-0.3965581954,-0.247439757,-0.1886555552,0.0342917405,-0.1788604558,-0.2145890892,-0.046575807,-0.1373553574,-0.1525721103,-0.0752802789,-0.0230884254,-0.284653604,0.0519280583,-0.1761706173,-0.1079143286,-0.0871832892,-0.0735015348,0.2006447762,0.008561397,0.0978720486,-0.4697065651,0.0869958177,-0.2451253086,0.1868368238,0.3407445252,0.0611883104,0.162994504,0.3850090206,0.1543998122,-0.1284118295,-0.0355447195,-0.1043887436,0.1108019426,-0.1283786148,-0.4418491423,0.126455456,0.0732035413,-0.0949317366,-0.207849592,-0.7182449102,-0.1001876965,0.2222401351,-0.1430765241,0.2521550357,-0.1681589484,-0.223976627,0.0424416177,-0.3285970688,0.0512887016,-0.1883378178,-0.7643768787,0.2837579846,-0.361350745,-0.241732046,-0.0683541372,-0.1362402737,0.1521255076,0.0578375719,-0.1992673874,-0.0171373226,-0.1693499684,0.2199527472,0.1156684533,0.3086595237,0.0356929079,-0.0309625939,0.1584858745,-0.150248006,-0.0465518758,0.2151976377,-0.342245698,0.3589676917,0.3947508335,-0.0425758064,0.2158914804,0.7951549292,0.2289339453,-0.1130506098,0.1516202539,-0.1019391418,0.390619874,0.3349591196,-0.2302670628,-0.0480876714,0.0123575199,-0.0399531014,0.132403627,0.0998543352,0.2279895246,-0.2753505409,-0.1235034093,-0.1770438701,-0.1852116287,-0.0710343942,0.5313087106,0.1815864295,0.0737839788,0.2022848129,-0.028091982,0.0728429779,0.0605884045,0.4210453033,0.4186688364,0.192148298,-0.1343727708,-0.0396623984,-0.1815310419,0.0533238873,0.3352300823,-0.1567851752,0.1472989768,0.0737626031,0.1796702296,0.1191362441,0.711619556,0.0238892883,-0.0698443204,-0.143161878,-0.2801820338,-0.0707878396,0.1149270087,-0.0957360789,0.0242426191,0.4746965766,0.2359755039,-0.3926457763,-0.3926123083,0.128238976,0.1697732806,-0.299464792,-0.1154548675,-0.1036774218,-0.1777816564,-0.1907158047,-0.2225705832,0.1076339558,0.0630090386,-0.0058769481,-0.0068223462,0.1004826352,0.0917314813,0.014331568,0.1632556021,0.4221811295,-0.0397359543,-0.0188680869,0.4088662863,0.3936689198,0.1898284703,0.4766064286,-0.4284280837,-0.1742584556,-0.0423083007,0.0491969809,0.6706671119,0.2639828324,-0.4991540611,0.0411934145,-0.0293762162,-0.1205298454,-0.3901128173,0.222776413,0.2994392812,-0.0055162297,-0.162582323,-0.6853826046,0.4479741156,-0.0421150289,-0.2754621208,0.1875998676,0.0526285581,-0.3134810328,-0.1579358876,-0.0429699421,0.8905719519,0.1091684848,0.1326388568,0.0964360908,-0.4127989411,0.6062426567,-0.0939917639,-0.1276177764,-0.2241677344,-0.0084786043,-0.1541097015,-0.2611880302,-0.018129576,-0.0639140531,-0.0427860245,0.3958505094,-0.0676030368,0.031684,0.0292265825,0.4977276325,-0.0441744067,-0.1707649827,0.0448242426,-0.1175881252,0.2282164395,0.440125227,-0.0789952651,-0.4989715517,-0.3009524047,0.1147170737,-0.0545501001,-0.0710054711,-0.4479240477,0.0286985449,0.0353882238,-0.3340227902,-0.0853986889,0.6262838244,0.2829657793,-0.0037522847,-0.1767268181,0.0725065023,-0.057849247,0.0756959319,0.2738863528,0.158611685,-0.0788758174,-0.0385617502,-0.1738841534,0.0879259109,-0.0382084511,-0.1904807091,0.1107859835,-0.1293090433,0.2101064771,-0.2470291853,-0.0184248388,-0.0158430338,0.0816949382,-0.0674472898,-0.0187769514,0.2836293578,-0.3048092723,0.0630547181,-0.282097131,-0.3546585143,-0.1788024455,0.3114256263,-0.1139961705,0.0560835674,0.2382426709,-0.1230002344,-0.2812078893,0.0032602465,0.2775327563,-0.1380116791,-0.3678483963,0.2783263922,0.3043947816,0.1569563597,-0.3993863463,0.255890429,0.0512776598,-0.3115510046,-0.2324522883,-0.4871122539,-0.2240505666,0.1930624098,-0.3125158846,0.1029729918,-0.0695718378,0.1195984185,0.0709021166,-0.0229240153,-0.1774468869,0.0112336632,-0.0516063534,-0.1422383636,0.3580642939,0.1030118465,0.3002460599,0.0284352303,-0.0640930533,-0.0792622119,-0.0325408094,0.002625529,-0.2550612688,0.19151555,-0.1290727854,-0.2119191736,0.0896774754,-0.1073907614,-0.0728524923,0.0386839025,0.2345136106,0.2677178383,0.1994189918,0.3762886226,-0.0558115728,-0.0718111843,-0.5359951258,0.4292730093,0.0665143728,0.4807568789,-0.1121403351,0.4839844406,-0.4124172926,-0.0912070423,0.1744728088,-0.1353558898,-0.0491278805,-0.0516893566,0.4121029675,0.0165324546,0.1032280028,0.1227181405,0.2486456633,0.2484258711,-0.0589140803,0.0733958408,0.2588761151,-0.0567365028,-0.0575201958,0.0019685891,-0.1295770258,-0.0149247572,0.0410098843,0.3463781178,0.3314770162,-0.0162728913,0.007801787,-0.2843848169,0.3758378029,0.44763726,0.0098603312,0.6277903914,0.151385203,0.3864418566,0.197964713,0.6132414937,0.1873982549,0.3135612011,-0.2671074867,-0.0985007063,-0.1000677124,0.2215511352,0.3323751688,-0.4487762153,0.1322830319,0.1527832896,-0.1746725887,0.0269176923,-0.2832137346,0.581245482,-0.1307188123,0.1158180758,-0.1224219725,0.4958434403,0.0782141089,-0.2474391013,-0.096474722,-0.2381599844,-0.1213553846,0.2078820914,-0.0611493587,-0.4505814016,0.1027271822,-0.079687953,0.1519846767,0.0232105795,0.072372444,0.1393904537,0.0882883444,-0.1974715889,0.1637422442,-0.0188082755,0.2937120795,0.3334792852,0.2842668593,0.0494500734,0.303371191,0.1744872481,-0.2400011718,0.245766893,-0.0123861646,-0.0853254199,-0.0283780992,-0.137529403,-0.079041779,0.4382867813,-0.0368901342,-0.0050124479,-0.2990782261,0.1053156331,0.2747851014,0.0507461838,0.5575402379,-0.0524819754,0.0360306837,0.1257701665,-0.0050424947,-0.4562192261,0.0532479361,0.2740341425,-0.2587765753,0.2118051499,0.021720292,-0.0001455999,0.0238951631,0.2618886828,0.1903114766,0.3403914273,-0.3425281942,-0.2439440042,-0.506909132,-0.2015317231,0.3620833755,-0.2535526752,-0.1272854656,0.2939567864,-0.2235213518,0.2132787555,0.0510136224,0.2868585289,-0.0318668634,0.1422691941,-0.1117262915,-0.1762109399,-0.3298960924,0.1201421544,0.1010260507,-0.1956543624,0.3169558346,-0.1772095263,-0.1324809641,-0.1015202105,0.2413177192,0.1422840506,-0.1664532721,0.1181592792,-0.0227623694,0.09941715,-0.0737084374,-0.0611808337,0.1847687662,0.1632815599,-0.2971652448,0.0032469204,-0.1449440271,-0.1646412015,-0.325243026,0.0519827716,-0.5005987287,0.3418357074,0.2640046477,0.0453134477,-0.0486570187,-0.0413432419,-0.3196692467,0.0502284095,0.5476945043,0.1139640361,-0.2953311205,0.0001216609,-0.1793000251,-0.0046625258,0.5082623959,-0.2689806223,0.1456076801,0.0743311271,0.1820535958,0.3494511545,-0.045969829,-0.5170993805,0.2281156033,0.5126157999,-0.07096605,0.1337634176,0.1736740917,0.0446925461,-0.0202741548,-0.1174549758,0.4670486152,0.179374367,-0.3825502992,0.1229937747,-0.0028742559]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1831","title":"Some question about raw dataset download info in the project .","comments":"Sure it would be nice to have an easier access to these paths !\r\nThe dataset builder could have a method to return those, what do you think ?\r\nFeel free to work on this @albertvillanova , it would be a nice addition :) \r\n\r\nYour suggestion does work as well @albertvillanova if you complete it by specifying `etag=` to `hash_url_to_filename`.\r\n\r\nThe ETag is obtained by a HEAD request and is used to know if the file on the remote host has changed. Therefore if a file is updated on the remote host, then the hash returned by `hash_url_to_filename` is different.","body":"Hi , i review the code in \r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py\r\nin the _split_generators function is the truly logic of download raw datasets with dl_manager\r\nand use Conll2003 cls by use import_main_class in load_dataset function\r\nMy question is that , with this logic it seems that i can not have the raw dataset download location\r\nin variable in downloaded_files in _split_generators.\r\nIf someone also want use huggingface datasets as raw dataset downloader,\r\nhow can he retrieve the raw dataset download path from attributes in \r\ndatasets.dataset_dict.DatasetDict ?","comment_length":100,"text":"Some question about raw dataset download info in the project .\nHi , i review the code in \r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py\r\nin the _split_generators function is the truly logic of download raw datasets with dl_manager\r\nand use Conll2003 cls by use import_main_class in load_dataset function\r\nMy question is that , with this logic it seems that i can not have the raw dataset download location\r\nin variable in downloaded_files in _split_generators.\r\nIf someone also want use huggingface datasets as raw dataset downloader,\r\nhow can he retrieve the raw dataset download path from attributes in \r\ndatasets.dataset_dict.DatasetDict ?\nSure it would be nice to have an easier access to these paths !\r\nThe dataset builder could have a method to return those, what do you think ?\r\nFeel free to work on this @albertvillanova , it would be a nice addition :) \r\n\r\nYour suggestion does work as well @albertvillanova if you complete it by specifying `etag=` to `hash_url_to_filename`.\r\n\r\nThe ETag is obtained by a HEAD request and is used to know if the file on the remote host has changed. Therefore if a file is updated on the remote host, then the hash returned by `hash_url_to_filename` is different.","embeddings":[-0.1199871525,-0.1007275581,-0.0441223904,0.5240066648,0.0438982211,-0.14139992,0.1183401495,0.0170494411,0.0113154026,0.1344257444,-0.3461932838,0.1511439532,-0.1354893446,0.5920690298,0.1894122958,-0.0379635394,-0.0841844305,0.1217005774,-0.1077714339,-0.0879005864,-0.3565678596,0.2822732031,-0.2687031925,0.165711835,-0.1946666539,-0.0248464216,-0.0181015562,0.3191277683,-0.4727252126,-0.2737624943,0.3005141318,0.2128322572,0.4005346894,0.1742072105,-0.0001249202,0.0375019796,0.2728613615,-0.3402907848,-0.5269052982,-0.2696060538,-0.6916824579,-0.0082495483,-0.0579732992,-0.2070541084,0.1547249407,0.0112637961,-0.053100694,-0.1325006634,0.0858320147,0.1186293438,0.0518584922,0.3697153628,-0.0970060304,0.0601943322,0.3183321357,0.2908330262,0.007793223,0.1266438365,-0.0761737525,0.3405940533,0.1056883559,0.302442044,0.2225824594,0.1028109491,0.4406277239,0.3373490274,-0.3781242967,-0.2769117951,0.0373095162,0.2423309535,0.3537122309,-0.2812837064,-0.58412534,-0.4083456099,-0.1191254482,-0.4685427547,0.1080773845,0.121445097,-0.4281236529,0.3086461127,-0.5675059557,-0.4780251682,-0.1658395231,-0.0852148011,-0.137549445,0.3164402246,-0.277533561,0.0663057417,0.0792562366,0.3843625188,0.2701301575,-0.2281227857,-0.021869285,0.2174347788,-0.1073611826,-0.1102793217,-0.0512238331,0.2954008281,0.1776958704,0.0195236113,-0.1246257648,-0.2577974796,-0.0622397587,0.0344964676,0.3408131301,0.0334263556,0.130103305,0.2412094921,0.2612249851,0.1766114682,0.1660885662,-0.2614012361,-0.115149036,-0.0674973726,-0.3462032378,0.1671791971,0.0185234584,-0.1103438139,-0.1785894781,-0.2391816229,-0.0279649924,-0.3496949673,0.314170599,0.3573299646,-0.1220672652,0.0427074023,-0.1295641065,-0.0033858558,0.1085045636,-0.3434873223,-0.0575917736,-0.2041842341,-0.1811356246,0.1146542653,-0.2467762232,-0.1190599427,0.3445102274,0.0368776731,0.0044905422,0.0511671118,-0.0212293789,-0.0554253049,0.2633837759,0.1153134927,0.0225037448,0.2498401701,-0.172318995,-0.2595724463,-0.194195807,-0.1401342303,-0.3265097737,-0.4692923427,-0.0550042801,-0.0365222543,-0.1397644877,0.0882554948,-0.2045928538,-0.0359483957,-0.0043262164,0.2903865874,0.0153338183,0.0368744805,-0.437574327,-0.261407584,0.4150717854,0.5498954654,-0.2640757263,-0.124882713,0.1626147479,-0.4113045931,0.0055077816,0.0851147249,-0.376221627,0.4481256902,-0.3954285383,-0.1491193324,0.4284542203,-0.3486127555,-0.2835324407,0.5134182572,0.0685734376,-0.0092668291,0.4227092564,0.0049932599,0.3071328998,-0.0831540376,-0.0749478117,0.5160323977,0.1115894988,-0.112585254,-0.0146319792,-0.2521874011,-0.1562729627,-0.0869760215,-0.0983398035,0.2540663779,0.1340302378,-0.1878966391,0.5806202292,-0.0906960666,0.2048040032,0.2849582434,0.0666126609,0.1446066797,-0.0434155166,0.0592394546,-0.6398258209,0.2061683983,-0.2713476121,-0.2402375489,-0.3789283931,-0.3321031332,-0.2896721363,0.1395770609,-0.2251721472,-0.3142985404,-0.0874109492,0.1475146711,0.4725340903,0.0423079804,-0.3148836493,0.5093697309,-0.0809046626,0.1222967356,-0.1639381498,0.0552342087,0.1615948528,0.1928499341,-0.0075578629,0.0963905826,0.117884092,-0.1511209607,0.0378694758,0.4245799184,0.1222041473,0.2514210641,0.3347149789,0.0576162823,0.2651583254,0.1458984911,-0.0056864223,-0.2740897834,0.0745327771,0.2104532719,0.0384140052,0.4083472192,-0.0198515505,0.2785645425,0.1696805805,-0.0084310053,0.103900522,-0.0308023803,0.1452540606,-0.1656621993,-0.1733821779,0.1735544205,0.3252809644,0.0908264816,-0.2493944466,0.2692500353,0.4680553973,-0.0680283308,-0.3477786183,0.1148413718,-0.2691214681,-0.055716712,0.124627687,0.8814340234,0.401753217,0.0123439897,0.2361507863,0.0718542114,0.2702481449,-0.0683172345,0.364923209,0.0819866583,-0.0913176388,-0.0149565851,-0.4808924198,-0.191566214,0.1524692923,-0.0226064268,0.1827385724,0.1612289995,-0.3683389723,0.033082407,-0.2861868441,-0.3713711798,-0.2221476436,-0.1441488564,-0.0842059329,-0.1802142113,-0.2356018871,-0.0917684063,-0.1567493081,-0.2011985332,-0.1180167273,0.0290215742,-0.222662434,0.0688054413,-0.1916803718,-0.155370906,-0.0985752493,-0.0907019079,0.1838143915,-0.0397568643,0.0875819921,-0.4963960648,0.1469944268,-0.3051645756,0.2047581673,0.3494753838,0.0336925536,0.176849544,0.4060946107,0.2000522166,-0.0665533319,0.0030451966,-0.1094028354,0.0911620408,-0.1093048006,-0.389409095,0.1404078156,0.0943241268,-0.0475681536,-0.2293685675,-0.6929691434,-0.0416314527,0.2408768535,-0.1981221735,0.2762157321,-0.1055136174,-0.1378174424,0.0386818759,-0.4083726108,-0.0131057324,-0.186048612,-0.8774716854,0.2753805518,-0.3544985354,-0.2303793877,-0.0539300665,-0.0597018078,0.178771615,0.0648877248,-0.2197252661,-0.0151435025,-0.1151275188,0.2784604132,0.1276265383,0.2927650511,0.0518944338,-0.0566531457,0.164882943,-0.1588708907,-0.0507449247,0.2370913327,-0.3477135003,0.3656221628,0.4061171114,-0.0131742787,0.2261146009,0.7650014758,0.185265854,-0.0783657506,0.1470417082,-0.1283288151,0.3387116194,0.3169541061,-0.1909824908,-0.0454266481,-0.0715172514,-0.078953214,0.175182417,0.123224929,0.2392908037,-0.2906849384,-0.2120719403,-0.165500164,-0.1859315336,-0.1148143038,0.5175088048,0.1606009454,0.1139920428,0.2111278921,-0.1149088964,0.0466600917,0.0916074514,0.3727550805,0.4440428019,0.1854564101,-0.1331199557,0.0172860213,-0.2204596996,0.0872576609,0.3334241211,-0.13853091,0.1563544571,0.0564962104,0.1949282736,0.1494938731,0.72339499,-0.0107273087,-0.0319373906,-0.2197545469,-0.2569633126,-0.0265562385,0.1234021857,-0.0482277758,0.03397673,0.4549541473,0.2488847673,-0.3597502112,-0.3623929918,0.1546041518,0.1162098721,-0.2746767104,-0.1146768928,-0.0833122283,-0.0899407342,-0.1801081449,-0.2004739642,0.0539680272,0.0568371303,-0.0310874358,-0.0269079953,0.1527353227,0.1428893805,-0.0335348137,0.1749982089,0.348239392,-0.0658729374,-0.0234589651,0.4482232034,0.4415008426,0.1750423908,0.4515222311,-0.3798852563,-0.1904298216,-0.0910222828,0.0152674746,0.6671972275,0.270257026,-0.4598641992,0.0253374763,0.0921915248,-0.0970593989,-0.4243964553,0.2011386305,0.2834628224,-0.0252960138,-0.1949806064,-0.6230821013,0.4908493161,-0.112314932,-0.2402643263,0.1811758578,0.1288716942,-0.3148922324,-0.1643142849,-0.0058821444,0.8531172276,0.1205510199,0.1551862508,0.1111792549,-0.4877255857,0.5681891441,-0.1102220938,-0.1549280435,-0.2622860968,-0.0139862485,-0.2005352378,-0.2348829806,-0.0075674965,-0.0535391048,0.0036479831,0.3770986795,0.0088334233,0.0962404013,0.0190934725,0.5144584775,-0.0574416891,-0.1568570435,0.1019321308,-0.112036787,0.2022064477,0.4311731756,-0.0611427203,-0.5024582744,-0.251190424,0.082346186,-0.0367888138,-0.0575543009,-0.4359017313,-0.0646287128,0.101085864,-0.3668353558,-0.1016825512,0.5698289275,0.2479190379,0.0213813484,-0.210635528,0.1094166115,-0.0936087668,0.1404230446,0.2428720891,0.1114888638,-0.0798583627,-0.0651342049,-0.2068641931,0.0714728013,-0.0393271334,-0.2618876994,0.1217391118,-0.1065542996,0.1592170596,-0.2659187913,-0.0238227379,-0.0651179254,0.0663717538,-0.0832174197,-0.0198666267,0.2281327695,-0.2856937945,0.0371865183,-0.2354148626,-0.3530788124,-0.1704949141,0.3665004075,-0.1845432222,0.0369889811,0.2687327266,-0.0928183943,-0.2668078542,0.0150241144,0.3082936704,-0.107783854,-0.3853513002,0.2481783032,0.2917563021,0.1974307597,-0.4033753574,0.2362273782,0.0175397359,-0.2571884096,-0.181803003,-0.4481992424,-0.2201611102,0.1389454007,-0.3590069115,0.1187703311,-0.0564319268,0.1030537635,0.0875074193,-0.0609076433,-0.1775811166,0.0292511899,-0.0316627994,-0.1708803177,0.3003230393,0.1231866106,0.3215070367,0.0304651167,-0.0757867098,-0.0375029035,-0.0238966234,0.0123707745,-0.267465502,0.2046534419,-0.1177913249,-0.1721881926,0.098898679,-0.066928573,-0.1082697362,0.111572355,0.2172593027,0.2519302964,0.1715552956,0.4327419102,-0.0578060076,-0.0781407878,-0.5655527115,0.4370854795,0.0744239762,0.5529670119,-0.1422433257,0.4968703091,-0.4778572321,-0.0598212294,0.1709110439,-0.0679049417,-0.0351617634,-0.0924729034,0.4812836647,0.0271645617,0.082426779,0.1139085069,0.2881810665,0.2134973109,-0.0559949167,-0.0220770482,0.2925488651,-0.0658350214,-0.0243379474,-0.0060554915,-0.1152768955,-0.0515512452,-0.0074527571,0.2672930062,0.3170594871,-0.057473924,0.0366612226,-0.2835755348,0.3031590283,0.4111518264,0.0160916038,0.6047902703,0.1351972222,0.3596574366,0.2057717443,0.5650690794,0.1462512165,0.3469216228,-0.2732594311,-0.1061102003,-0.0679322779,0.2807175517,0.3305304348,-0.4702317119,0.1446443498,0.113478899,-0.2296659201,0.0006697136,-0.2954060435,0.6380744576,-0.0482070819,0.1427692324,-0.1898059398,0.5034026504,0.0923021957,-0.2670196295,-0.1357155889,-0.260368228,-0.1646879762,0.2709100246,-0.0680674911,-0.4204149246,0.0985172316,-0.0504771136,0.1767946929,0.0280137658,0.0287282486,0.0661977902,0.0950494856,-0.2267214656,0.1004459634,0.0161683206,0.2924774885,0.4172537625,0.2327012867,0.0323164687,0.258326143,0.2004224211,-0.2206254601,0.2401108742,-0.0181815457,-0.0658183172,0.0275006052,-0.1293638647,0.0138201918,0.4370181561,-0.0464095436,0.0211230349,-0.2937388122,0.0402429625,0.2716841102,-0.0134611325,0.4885117412,0.0111889457,0.0218151864,0.0729123726,0.0148397889,-0.4192560017,0.0781184211,0.3170060217,-0.2312473506,0.1861359626,-0.0094240857,-0.004024114,-0.0653238371,0.3525848389,0.2203418314,0.3803894818,-0.3470634222,-0.2478012145,-0.4514292479,-0.214824602,0.3679114878,-0.2235813737,-0.0821258426,0.2524747849,-0.2352206856,0.1795069128,0.0145755336,0.2517454326,-0.0515870154,0.1812433749,-0.0746955499,-0.1761794686,-0.3598528802,0.128580004,0.0889082253,-0.1742634177,0.3584089875,-0.1178860664,-0.1673536152,-0.0971778184,0.2376361638,0.174100399,-0.2355187237,0.1722303331,0.0413751304,0.1344534606,-0.064554058,-0.0557097197,0.121246703,0.1673983783,-0.3241881728,-0.0388959274,-0.1537386775,-0.1308430433,-0.1989883631,0.0394288525,-0.5205610394,0.3817767203,0.2414392233,0.0903250203,-0.0592815503,-0.0564866588,-0.2958540022,0.0116280066,0.5659484267,0.1310495138,-0.2498774081,0.0230484605,-0.1982488185,0.0172985103,0.5022480488,-0.3008303046,0.1593635678,0.0607874617,0.2231656462,0.3416747451,-0.1027992964,-0.5669338107,0.2001698762,0.5056933761,-0.1171722859,0.1527397186,0.1659211069,0.142445609,0.0287007242,-0.1023585945,0.508535862,0.1950047612,-0.4277584553,0.1314696074,-0.0252335034]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1831","title":"Some question about raw dataset download info in the project .","comments":"Once #1846 will be merged, the paths to the raw downloaded files will be accessible as:\r\n```python\r\nbuilder_instance.dl_manager.downloaded_paths\r\n``` ","body":"Hi , i review the code in \r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py\r\nin the _split_generators function is the truly logic of download raw datasets with dl_manager\r\nand use Conll2003 cls by use import_main_class in load_dataset function\r\nMy question is that , with this logic it seems that i can not have the raw dataset download location\r\nin variable in downloaded_files in _split_generators.\r\nIf someone also want use huggingface datasets as raw dataset downloader,\r\nhow can he retrieve the raw dataset download path from attributes in \r\ndatasets.dataset_dict.DatasetDict ?","comment_length":19,"text":"Some question about raw dataset download info in the project .\nHi , i review the code in \r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/conll2003\/conll2003.py\r\nin the _split_generators function is the truly logic of download raw datasets with dl_manager\r\nand use Conll2003 cls by use import_main_class in load_dataset function\r\nMy question is that , with this logic it seems that i can not have the raw dataset download location\r\nin variable in downloaded_files in _split_generators.\r\nIf someone also want use huggingface datasets as raw dataset downloader,\r\nhow can he retrieve the raw dataset download path from attributes in \r\ndatasets.dataset_dict.DatasetDict ?\nOnce #1846 will be merged, the paths to the raw downloaded files will be accessible as:\r\n```python\r\nbuilder_instance.dl_manager.downloaded_paths\r\n``` ","embeddings":[-0.2227526158,-0.0727905929,-0.0795345679,0.5368555784,0.1130919904,-0.0717310086,0.1194330305,-0.0408380963,0.0301584732,0.1258056462,-0.3729727268,0.1903695017,-0.1131497398,0.6030976176,0.1462550908,-0.0494616218,-0.0528516881,0.1931553036,-0.0936021209,-0.0438142344,-0.4198140502,0.3098753393,-0.3674631417,0.1116438881,-0.183257401,0.0009527694,0.0041938564,0.2849134803,-0.4855917692,-0.2597879171,0.2608330548,0.2013038546,0.4007953405,0.2447413802,-0.0001246907,0.0452720039,0.2902215719,-0.3360069692,-0.4938015044,-0.2695348263,-0.6499729753,0.0080848392,-0.0744823068,-0.1832295656,0.0880008712,-0.0940373987,-0.0886582211,-0.1186727509,0.1216425449,0.0911427811,0.0624107942,0.3518903852,-0.021356253,0.0244734734,0.225017339,0.2522042096,0.058083307,0.061889872,-0.0252658818,0.3873560429,0.1903803051,0.2815833092,0.2330066413,0.1151762083,0.4400799274,0.2746642232,-0.2849640548,-0.3441517651,0.0646265,0.1897077411,0.3417406678,-0.3277815878,-0.480050087,-0.35111323,-0.1297976822,-0.580676496,0.0638201162,0.1226147786,-0.4566832185,0.2788124681,-0.516233325,-0.5137311816,-0.1360852122,-0.094853878,-0.2339905202,0.4693616033,-0.3634061813,0.0339230672,0.078788884,0.4059506357,0.252389133,-0.1982115507,-0.0329441838,0.2363315076,-0.1470575482,-0.1451294124,-0.1241967902,0.2437325716,0.1872886717,-0.0102504175,-0.1592288017,-0.2508050203,0.0632321388,0.0414232761,0.3790727258,0.0483655259,0.0730750933,0.2580432892,0.2510024011,0.1209764183,0.1441133022,-0.2335914224,-0.0929673463,-0.0628621578,-0.3814372718,0.1422059685,0.0251014978,-0.0544403046,-0.2750721872,-0.2144883871,-0.0897702277,-0.3215322793,0.3057874441,0.3743990958,-0.1019364148,0.022056818,-0.0423464738,0.0158646069,0.1265101731,-0.3291881084,-0.0845960081,-0.2330509126,-0.1551715136,0.0351234376,-0.2879267931,-0.1038957462,0.4116896093,0.1091846451,0.0140773728,0.0599597916,0.0134301223,-0.0717188939,0.1696660072,0.1859183162,0.0490511879,0.2148469239,-0.1917363852,-0.25927791,-0.1738078892,-0.0748295859,-0.322701335,-0.4281926751,-0.1209012046,0.0081474725,-0.1694132537,0.1197640374,-0.2129068524,0.0280262325,0.1235994175,0.2871412635,0.0152981626,-0.0111901397,-0.4282560647,-0.2514013052,0.3835564256,0.4348743558,-0.2215136141,-0.1230010241,0.0940126255,-0.431432277,0.0629917607,0.0478170104,-0.3582910299,0.5085852146,-0.4254652858,-0.182280615,0.5232151747,-0.3267568052,-0.2476575077,0.5593610406,0.1116131172,-0.0978967324,0.4094040692,-0.0305032916,0.2824288607,-0.0730031952,-0.0714553371,0.5319201946,0.1348209977,-0.1183862314,0.0307501685,-0.242930606,-0.1591515541,-0.1199134365,-0.1136741862,0.2524617016,0.0738943741,-0.1367600113,0.526057601,-0.0497717112,0.2035022527,0.3350777328,-0.0111465948,0.1369570643,0.0173293781,-0.015499102,-0.6657516956,0.0838522613,-0.2974469662,-0.2576330602,-0.2690088749,-0.3559127152,-0.3377182782,0.1751461327,-0.2877247632,-0.332760632,-0.0792820156,0.0922700241,0.4710853696,-0.0258780457,-0.3952095509,0.5251181722,-0.0418808423,0.1615414321,-0.088803716,-0.0046257917,0.1427607536,0.139867276,0.0007156456,0.1126954183,0.0803142264,-0.1782867014,0.0629307032,0.4662190676,0.1516308635,0.1959070712,0.1718022972,0.0093580056,0.2583298683,0.1817045361,-0.0510871932,-0.2297551781,0.0971029177,0.2127887309,0.0565138943,0.3524635434,0.0223180223,0.2595730722,0.165094763,-0.0690742135,0.1472309232,-0.0168987513,0.1959601194,-0.2239766866,-0.1486384124,0.206876412,0.3971292377,0.1536176205,-0.2714310586,0.1967890412,0.4208644331,-0.0679546371,-0.2724224627,0.0342276357,-0.318757683,-0.0333493762,0.0566567108,0.7819696665,0.4830187559,0.0190001447,0.194326207,0.0896532536,0.2920681536,-0.1252113432,0.4004474878,0.1224451214,-0.032063745,0.0760392845,-0.4785551727,-0.2069208622,0.1402295679,0.0287683439,0.2454356849,0.1555006951,-0.323954016,0.0535907038,-0.2396930307,-0.3992612362,-0.281919986,-0.1486748755,-0.015991414,-0.1645134985,-0.243656382,-0.0574240834,-0.0981531069,-0.1697708666,-0.048838269,0.0056508719,-0.2327387035,0.0742060319,-0.1707400084,-0.0941132233,-0.1039258018,-0.0643206909,0.1774672717,-0.0226612631,0.0893995613,-0.4694087207,0.1217619628,-0.1666531116,0.1832148433,0.3540813625,0.0813799724,0.1517674625,0.4431499243,0.1289682388,-0.0204903074,0.0068593398,-0.1250994354,0.124535948,-0.1264799982,-0.4716933668,0.1030035987,0.098529093,-0.0831040815,-0.2064294964,-0.7116601467,-0.0992264524,0.1218955442,-0.1045544967,0.2185328156,-0.1333873868,-0.1711411476,0.0174398627,-0.2863847017,-0.0070472839,-0.2202371508,-0.740229249,0.2095378637,-0.3735067248,-0.2058056295,-0.0553150065,-0.1200563684,0.2111456841,-0.0086781941,-0.1650973409,0.0514048338,-0.1619434953,0.1768596172,0.0831115246,0.376240015,-0.0347805582,-0.0497439653,0.1611572057,-0.1935867667,-0.0018390776,0.1826922446,-0.391580224,0.3662184477,0.3890748024,-0.0243424345,0.1229492724,0.75611341,0.2864711881,-0.1502739191,0.1753432453,-0.1364036798,0.4252972603,0.2925858796,-0.1773474365,-0.0488879457,0.0727672279,-0.0693987161,0.1823141426,0.0921332836,0.2284848541,-0.2628238797,-0.1667081118,-0.1121023893,-0.130259797,-0.1420613229,0.549648881,0.1925711334,0.0929878503,0.1902042329,-0.0035492803,0.0613080077,0.0210150331,0.4095702171,0.3880938888,0.1618987173,-0.0672065616,-0.0749615952,-0.2247973233,0.0742219463,0.2897367179,-0.1773893833,0.1776889861,0.0986660942,0.2031379342,0.1185432002,0.7021445036,-0.0180740729,-0.0924170837,-0.1328479797,-0.2192426026,-0.011309104,0.1038035527,-0.1278323531,0.1078302562,0.5110150576,0.2077199072,-0.372115314,-0.4381561875,0.2151828855,0.1094943285,-0.2812062502,-0.1837836951,-0.1544685364,-0.1078191772,-0.2226545662,-0.2091927528,0.1499244571,0.0881495252,-0.0057039484,-0.0566783622,0.082651332,0.0841821656,-0.0086777816,0.1462733448,0.3609340191,-0.0120865358,-0.0733288303,0.3574350178,0.3906067312,0.1943040043,0.463727653,-0.3728158772,-0.1768314093,-0.0265096314,0.0531405099,0.7556014657,0.1763658226,-0.4882671833,0.0601237044,-0.0116341747,-0.0755682886,-0.37209782,0.2422152907,0.2992217839,0.0321992971,-0.1801103354,-0.6387315989,0.4800695181,-0.0520989969,-0.2018402964,0.1666305065,0.0069741779,-0.3069559634,-0.1485978067,-0.0105797043,0.7889214754,0.1614478081,0.1121999398,0.0857862756,-0.4110881388,0.5621554255,-0.1268182397,-0.1707640439,-0.2967331409,0.0537740253,-0.1420812309,-0.2615424395,0.0487701334,-0.0964144468,-0.0835701451,0.3538786471,-0.1031554118,0.0783122554,0.0351161771,0.5478224158,-0.0928407758,-0.1032031551,0.0988958925,-0.1113560647,0.2575158477,0.4692693055,-0.1261555105,-0.4839955568,-0.1835500449,0.1057132408,-0.0513833649,-0.0979699194,-0.4327759147,-0.0170005951,0.0926157236,-0.3537166417,0.012873237,0.6208472848,0.2836615145,0.0231477823,-0.1648344398,0.0673446879,-0.0495202877,0.0769015625,0.2480778098,0.2085042,-0.0699047148,-0.078473039,-0.158148393,0.1368762404,0.0107301353,-0.2063127458,0.0896549672,-0.1035653055,0.181635499,-0.3317039907,-0.0347604007,-0.0417184718,0.0668420717,-0.0687279403,-0.0109379767,0.2227708995,-0.3475803435,0.0621425807,-0.2775212526,-0.3875669837,-0.1929181218,0.2877932191,-0.05374052,0.0600106232,0.2207129151,-0.0982339531,-0.2024487108,-0.0398914441,0.3363055885,-0.1070859954,-0.2793478668,0.3372612,0.3225170672,0.0940122008,-0.3637564182,0.3148418963,-0.0405122787,-0.2393659949,-0.2452914566,-0.4822891057,-0.2695065737,0.1232939363,-0.3319842815,0.1249401867,-0.0620176196,0.195829004,0.0651968718,-0.0243790969,-0.1795801222,0.0090976721,-0.0362625867,-0.1489288956,0.3070869446,0.098624073,0.2730362713,0.094840005,-0.0722218379,-0.0209141076,-0.072385855,-0.0107932696,-0.22540088,0.1856942773,-0.1288864464,-0.1764189303,0.1162186563,-0.0721574947,-0.1746608615,0.1263108402,0.2937785685,0.1577110142,0.173881039,0.3827465475,-0.0763128996,-0.0502565429,-0.54657197,0.3560209274,0.0966606885,0.4793519676,-0.1747328639,0.4857376814,-0.4541743994,-0.044782225,0.1853281856,-0.1238309294,0.0049047,-0.1167225465,0.4232525229,-0.0213853046,0.1024012715,0.1084321141,0.294776082,0.2523184717,-0.1078114584,0.0440975726,0.1966435462,-0.0459347293,-0.0586760454,-0.0020265393,-0.1170229837,-0.0076994877,0.0213974323,0.3223966658,0.3394872248,0.0151304994,-0.0119264973,-0.2416768372,0.3374834955,0.4738022685,0.0307134725,0.6178550124,0.1547664106,0.3950329125,0.2319709659,0.6090508699,0.1564681381,0.3764047325,-0.251188904,-0.0926406682,-0.0720532537,0.2094828486,0.3799166679,-0.4324010611,0.1841048598,0.1872385293,-0.2192243934,0.0289910063,-0.247770831,0.5808362961,-0.1430685967,0.1118899882,-0.1777613759,0.4941423535,-0.0063958564,-0.2583303154,-0.1401859522,-0.2041163445,-0.0913490504,0.2170603424,-0.064143382,-0.4420026839,0.0254701599,-0.1010132656,0.1205315292,0.0508955792,0.0709128752,0.0971238166,0.0195931401,-0.2196977437,0.208930552,-0.0922005177,0.3163809776,0.3709083796,0.2975154519,0.0974186957,0.312690258,0.159586221,-0.2844440043,0.2018164843,-0.0475477688,-0.090561375,0.0061739916,-0.1210865155,0.0058415672,0.41826123,-0.0461422913,-0.0200621467,-0.364814043,0.1689004004,0.2512781322,0.0000591171,0.5455234051,-0.0535514913,0.0083286846,0.0869262889,-0.0172512438,-0.4771107137,0.0602712519,0.302937746,-0.2579323053,0.2056422979,0.0049277418,-0.0021013271,-0.0642879158,0.3228241503,0.1336645037,0.3569436669,-0.3690541089,-0.2826679051,-0.4720335305,-0.1578612775,0.3902359009,-0.2169174403,-0.0885373652,0.2607622743,-0.2557821274,0.2271687239,0.0340682268,0.3226421475,0.0107520148,0.1497735083,-0.0656410307,-0.161481753,-0.3975390494,0.1098094881,0.114342995,-0.1939413399,0.3625616729,-0.2283436358,-0.1399366111,-0.1116103008,0.2486997545,0.2083444297,-0.2049413472,0.1192519963,-0.0392198563,0.1562286764,-0.0664401725,0.0166159812,0.2328225523,0.2091890424,-0.3784237206,-0.0006283663,-0.1004873961,-0.1940930188,-0.304305464,0.046580337,-0.4616069198,0.3110299706,0.2901493013,0.0687093437,-0.0569677874,-0.0396671295,-0.3614220023,0.0534360781,0.5975099206,0.1792013794,-0.2954566181,-0.0016953355,-0.20202896,0.0181056391,0.4721162021,-0.2760750651,0.0974705294,-0.0395485237,0.1857443452,0.3565402925,-0.1293281913,-0.5400487781,0.2439939529,0.5118327141,-0.1200243831,0.1001455709,0.2415116429,0.1531035155,-0.0162189882,-0.0992011577,0.5105974078,0.201197952,-0.4127675891,0.1326322407,-0.0148094008]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"Hi @wumpusman \r\n`datasets` has a caching mechanism that allows to cache the results of `.map` so that when you want to re-run it later it doesn't recompute it again.\r\nSo when you do `.map`, what actually happens is:\r\n1. compute the hash used to identify your `map` for the cache\r\n2. apply your function on every batch\r\n\r\nThis can explain the time difference between your different experiments.\r\n\r\nThe hash computation time depends of how complex your function is. For a tokenizer, the hash computation scans the lists of the words in the tokenizer to identify this tokenizer. Usually it takes 2-3 seconds.\r\n\r\nAlso note that you can disable caching though using\r\n```python\r\nimport datasets\r\n\r\ndatasets.set_caching_enabled(False)\r\n```","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":116,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\nHi @wumpusman \r\n`datasets` has a caching mechanism that allows to cache the results of `.map` so that when you want to re-run it later it doesn't recompute it again.\r\nSo when you do `.map`, what actually happens is:\r\n1. compute the hash used to identify your `map` for the cache\r\n2. apply your function on every batch\r\n\r\nThis can explain the time difference between your different experiments.\r\n\r\nThe hash computation time depends of how complex your function is. For a tokenizer, the hash computation scans the lists of the words in the tokenizer to identify this tokenizer. Usually it takes 2-3 seconds.\r\n\r\nAlso note that you can disable caching though using\r\n```python\r\nimport datasets\r\n\r\ndatasets.set_caching_enabled(False)\r\n```","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"Hi @lhoestq ,\r\n\r\nThanks for the reply. It's entirely possible that is the issue. Since it's a side project I won't be looking at it till later this week, but, I'll verify it by disabling caching and hopefully I'll see the same runtime. \r\n\r\nAppreciate the reference,\r\n\r\nMichael","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":47,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\nHi @lhoestq ,\r\n\r\nThanks for the reply. It's entirely possible that is the issue. Since it's a side project I won't be looking at it till later this week, but, I'll verify it by disabling caching and hopefully I'll see the same runtime. \r\n\r\nAppreciate the reference,\r\n\r\nMichael","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"I believe this is an actual issue, tokenizing a ~4GB txt file went from an hour and a half to ~10 minutes when I switched from my pre-trained tokenizer(on the same dataset) to the default gpt2 tokenizer.\r\nBoth were loaded using:\r\n```\r\nAutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)\r\n```\r\nI trained the tokenizer using ByteLevelBPETokenizer from the Tokenizers library and save it to a tokenizer.json file.\r\n\r\nI have tested the caching ideas above, changing the number of process, the TOKENIZERS_PARALLELISM env variable, keep_in_memory=True and batching with different sizes.\r\n\r\nApologies I can't really upload much code, but wanted to back up the finding and hopefully a fix\/the problem can be found.\r\nI will comment back if I find a fix as well.","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":117,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\nI believe this is an actual issue, tokenizing a ~4GB txt file went from an hour and a half to ~10 minutes when I switched from my pre-trained tokenizer(on the same dataset) to the default gpt2 tokenizer.\r\nBoth were loaded using:\r\n```\r\nAutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)\r\n```\r\nI trained the tokenizer using ByteLevelBPETokenizer from the Tokenizers library and save it to a tokenizer.json file.\r\n\r\nI have tested the caching ideas above, changing the number of process, the TOKENIZERS_PARALLELISM env variable, keep_in_memory=True and batching with different sizes.\r\n\r\nApologies I can't really upload much code, but wanted to back up the finding and hopefully a fix\/the problem can be found.\r\nI will comment back if I find a fix as well.","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"Hi @johncookds do you think this can come from one tokenizer being faster than the other one ? Can you try to compare their speed without using `datasets` just to make sure ?","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":33,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\nHi @johncookds do you think this can come from one tokenizer being faster than the other one ? Can you try to compare their speed without using `datasets` just to make sure ?","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"Hi yes, I'm closing the loop here with some timings below. The issue seems to be at least somewhat\/mainly with the tokenizer's themselves. Moreover legacy saves of the trainer tokenizer perform faster but differently than the new tokenizer.json saves(note nothing about the training process\/adding of special tokens changed between the top two trained tokenizer tests, only the way it was saved). This is only a 3x slowdown vs like a 10x but I think the slowdown is most likely due to this.\r\n\r\n```\r\ntrained tokenizer - tokenizer.json save (same results for AutoTokenizer legacy_format=False):\r\nTokenizer time(seconds): 0.32767510414123535\r\nTokenized avg. length: 323.01\r\n\r\ntrained tokenizer - AutoTokenizer legacy_format=True:\r\nTokenizer time(seconds): 0.09258866310119629\r\nTokenized avg. length: 301.01\r\n\r\nGPT2 Tokenizer from huggingface\r\nTokenizer time(seconds): 0.1010282039642334\r\nTokenized avg. length: 461.21\r\n```","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":124,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\nHi yes, I'm closing the loop here with some timings below. The issue seems to be at least somewhat\/mainly with the tokenizer's themselves. Moreover legacy saves of the trainer tokenizer perform faster but differently than the new tokenizer.json saves(note nothing about the training process\/adding of special tokens changed between the top two trained tokenizer tests, only the way it was saved). This is only a 3x slowdown vs like a 10x but I think the slowdown is most likely due to this.\r\n\r\n```\r\ntrained tokenizer - tokenizer.json save (same results for AutoTokenizer legacy_format=False):\r\nTokenizer time(seconds): 0.32767510414123535\r\nTokenized avg. length: 323.01\r\n\r\ntrained tokenizer - AutoTokenizer legacy_format=True:\r\nTokenizer time(seconds): 0.09258866310119629\r\nTokenized avg. length: 301.01\r\n\r\nGPT2 Tokenizer from huggingface\r\nTokenizer time(seconds): 0.1010282039642334\r\nTokenized avg. length: 461.21\r\n```","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"@lhoestq ,\r\n\r\nHi, which version of datasets has datasets.set_caching_enabled(False)? I get \r\nmodule 'datasets' has no attribute 'set_caching_enabled'. To hopefully get around this, I reran my code on a new set of data, and did so only once.\r\n\r\n@johncookds , thanks for chiming in, it looks this might be an issue of Tokenizer.\r\n\r\n**Tokenizer**: The runtime of GPT2TokenizerFast.from_pretrained(\"gpt2\") on 1000 chars is: **143 ms**\r\n**SlowTokenizer**: The runtime of a locally saved and loaded Tokenizer using the same vocab on 1000 chars is: **4.43 s**\r\n\r\nThat being said, I compared performance on the map function:\r\n\r\nRunning Tokenizer versus using it in the map function for 1000 chars goes from **141 ms** to **356 ms** \r\nRunning SlowTokenizer versus using it in the map function for 1000 chars with a single element goes from **4.43 s** to **9.76 s**\r\n\r\nI'm trying to figure out why the overhead of map would increase the time by double (figured it would be a fixed increase in time)? Though maybe this is expected behavior.\r\n\r\n@lhoestq, do you by chance know how I can redirect this issue to Tokenizer?\r\n\r\nRegards,\r\n\r\nMichael","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":182,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\n@lhoestq ,\r\n\r\nHi, which version of datasets has datasets.set_caching_enabled(False)? I get \r\nmodule 'datasets' has no attribute 'set_caching_enabled'. To hopefully get around this, I reran my code on a new set of data, and did so only once.\r\n\r\n@johncookds , thanks for chiming in, it looks this might be an issue of Tokenizer.\r\n\r\n**Tokenizer**: The runtime of GPT2TokenizerFast.from_pretrained(\"gpt2\") on 1000 chars is: **143 ms**\r\n**SlowTokenizer**: The runtime of a locally saved and loaded Tokenizer using the same vocab on 1000 chars is: **4.43 s**\r\n\r\nThat being said, I compared performance on the map function:\r\n\r\nRunning Tokenizer versus using it in the map function for 1000 chars goes from **141 ms** to **356 ms** \r\nRunning SlowTokenizer versus using it in the map function for 1000 chars with a single element goes from **4.43 s** to **9.76 s**\r\n\r\nI'm trying to figure out why the overhead of map would increase the time by double (figured it would be a fixed increase in time)? Though maybe this is expected behavior.\r\n\r\n@lhoestq, do you by chance know how I can redirect this issue to Tokenizer?\r\n\r\nRegards,\r\n\r\nMichael","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"Thanks for the experiments @johncookds and @wumpusman ! \r\n\r\n> Hi, which version of datasets has datasets.set_caching_enabled(False)?\r\n\r\nCurrently you have to install `datasets` from source to have this feature, but this will be available in the next release in a few days.\r\n\r\n> I'm trying to figure out why the overhead of map would increase the time by double (figured it would be a fixed increase in time)? Though maybe this is expected behavior.\r\n\r\nCould you also try with double the number of characters ? This should let us have an idea of the fixed cost (hashing) and the dynamic cost (actual tokenization, grows with the size of the input)\r\n\r\n> @lhoestq, do you by chance know how I can redirect this issue to Tokenizer?\r\n\r\nFeel free to post an issue on the `transformers` repo. Also I'm sure there should be related issues so you can also look for someone with the same concerns on the `transformers` repo.","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":157,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\nThanks for the experiments @johncookds and @wumpusman ! \r\n\r\n> Hi, which version of datasets has datasets.set_caching_enabled(False)?\r\n\r\nCurrently you have to install `datasets` from source to have this feature, but this will be available in the next release in a few days.\r\n\r\n> I'm trying to figure out why the overhead of map would increase the time by double (figured it would be a fixed increase in time)? Though maybe this is expected behavior.\r\n\r\nCould you also try with double the number of characters ? This should let us have an idea of the fixed cost (hashing) and the dynamic cost (actual tokenization, grows with the size of the input)\r\n\r\n> @lhoestq, do you by chance know how I can redirect this issue to Tokenizer?\r\n\r\nFeel free to post an issue on the `transformers` repo. Also I'm sure there should be related issues so you can also look for someone with the same concerns on the `transformers` repo.","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"@lhoestq,\r\n\r\nI just checked that previous run time was actually 3000 chars. I increased it to 6k chars, again, roughly double.\r\n\r\nSlowTokenizer **7.4 s** to **15.7 s**\r\nTokenizer: **276 ms** to **616 ms**\r\n\r\nI'll post this issue on Tokenizer, seems it hasn't quite been raised (albeit I noticed a similar issue that might relate).\r\n\r\nRegards,\r\n\r\nMichael","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":56,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\n@lhoestq,\r\n\r\nI just checked that previous run time was actually 3000 chars. I increased it to 6k chars, again, roughly double.\r\n\r\nSlowTokenizer **7.4 s** to **15.7 s**\r\nTokenizer: **276 ms** to **616 ms**\r\n\r\nI'll post this issue on Tokenizer, seems it hasn't quite been raised (albeit I noticed a similar issue that might relate).\r\n\r\nRegards,\r\n\r\nMichael","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1830","title":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?","comments":"Hi, \r\nI'm following up here as I found my exact issue. It was with saving and re-loading the tokenizer. When I trained then processed the data without saving and reloading it, it was 10x-100x faster than when I saved and re-loaded it.\r\nBoth resulted in the exact same tokenized datasets as well. \r\nThere is additionally a bug where the older legacy tokenizer save does not preserve a learned tokenizing behavior if trained from scratch.\r\nUnderstand its not exactly Datasets related but hope it can help someone if they have the same issue.\r\nThanks!","body":"This could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n","comment_length":93,"text":"using map on loaded Tokenizer 10x - 100x slower than default Tokenizer?\nThis could total relate to me misunderstanding particular call functions, but I added words to a GPT2Tokenizer, and saved it to disk (note I'm only showing snippets but I can share more) and the map function ran much slower: \r\n\r\n````\r\ndef save_tokenizer(original_tokenizer,text,path=\"simpledata\/tokenizer\"):\r\n words_unique = set(text.split(\" \"))\r\n for i in words_unique:\r\n original_tokenizer.add_tokens(i)\r\n original_tokenizer.save_pretrained(path)\r\n\r\ntokenizer2 = GPT2Tokenizer.from_pretrained(os.path.join(experiment_path,experiment_name,\"tokenizer_squad\"))\r\n\r\ntrain_set_baby=Dataset.from_dict({\"text\":[train_set[\"text\"][0][0:50]]})\r\n````\r\n\r\nI then applied the dataset map function on a fairly small set of text:\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\n\r\n\r\nThe run time for train_set_baby.map was 6 seconds, and the batch itself was 2.6 seconds\r\n\r\n**100% 1\/1 [00:02<00:00, 2.60s\/ba] CPU times: user 5.96 s, sys: 36 ms, total: 5.99 s Wall time: 5.99 s**\r\n\r\nIn comparison using (even after adding additional tokens): \r\n`\r\ntokenizer = GPT2TokenizerFast.from_pretrained(\"gpt2\")`\r\n\r\n```\r\n%%time\r\ntrain_set_baby = train_set_baby.map(lambda d:tokenizer2(d[\"text\"]),batched=True)\r\n\r\n```\r\nThe time is \r\n**100% 1\/1 [00:00<00:00, 34.09ba\/s] CPU times: user 68.1 ms, sys: 16 \u00b5s, total: 68.1 ms Wall time: 62.9 ms**\r\n\r\nIt seems this might relate to the tokenizer save or load function, however, the issue appears to come up when I apply the loaded tokenizer to the map function. \r\n\r\nI should also add that playing around with the amount of words I add to the tokenizer before I save it to disk and load it into memory appears to impact the time it takes to run the map function. \r\n\r\n\r\n\nHi, \r\nI'm following up here as I found my exact issue. It was with saving and re-loading the tokenizer. When I trained then processed the data without saving and reloading it, it was 10x-100x faster than when I saved and re-loaded it.\r\nBoth resulted in the exact same tokenized datasets as well. \r\nThere is additionally a bug where the older legacy tokenizer save does not preserve a learned tokenizing behavior if trained from scratch.\r\nUnderstand its not exactly Datasets related but hope it can help someone if they have the same issue.\r\nThanks!","embeddings":[-0.4040468037,-0.0130073428,-0.1024706438,0.1034291983,0.1453053504,-0.1225539744,0.2550541461,0.189088881,0.1306775063,-0.0065023075,-0.0542231686,0.4988860488,-0.1724810004,-0.2727595866,-0.1320426911,0.0650426894,0.2197590619,0.0224171188,0.2416453809,-0.03428654,-0.1218727604,0.1023019925,-0.1767257303,0.2413996309,-0.3152630031,-0.0779930353,0.0812313482,-0.2096335888,-0.0027567474,-0.5689615011,-0.2084299624,0.1581248492,0.0209429283,0.5331252813,-0.0001148737,0.0403426066,-0.0291637182,0.2731475234,0.1347175092,0.0443611667,0.3529542685,-0.1942853481,-0.1685831547,-0.2257665098,-0.0541709252,-0.2808631957,-0.0221225843,-0.2241773009,0.0683081374,0.03004192,0.1608587801,0.1151950732,-0.3964318633,-0.0595159791,-0.0024588848,-0.1042250693,-0.0991546437,0.0140264928,0.071329318,-0.255187422,-0.2116566598,0.1949048191,-0.2130429149,-0.1478140801,0.2687232494,0.2163694054,0.0346116014,-0.2425245792,0.1680867523,0.2316348851,0.0670110211,-0.0272130258,-0.017125072,-0.3350791633,-0.370691061,-0.0284823496,0.146122545,-0.4334974587,-0.1206515282,-0.0896027014,-0.4140653312,0.2037731558,0.2642425895,-0.1223938689,-0.2039308697,0.435243994,0.0954279751,0.1497347653,0.1041521281,-0.0997518599,-0.2807916701,-0.2643776536,-0.1595768332,0.3733043671,-0.3415869772,-0.1739550233,0.2692955434,-0.3091349304,-0.14078857,0.1142799705,-0.1424314529,0.194665283,0.2083414793,-0.0471201427,0.0663070083,0.2460043877,0.0384478755,0.173846364,0.0296903811,-0.4802925289,-0.3160505593,-0.0903609022,-0.188874945,0.0202136673,0.3458716273,-0.2424587309,-0.2370802015,-0.072583504,-0.0205964632,-0.171307236,-0.3799443841,-0.0983371884,0.0658612549,0.185454756,0.0587473363,0.0610382333,-0.3647983074,-0.2021667808,-0.0928319693,-0.0767263323,-0.128627494,-0.1720007658,-0.1044918671,0.2542574704,-0.1882735491,0.0913059115,0.2601749003,0.0789951906,-0.1153759062,0.2103088349,0.2196046263,-0.3643229902,0.1047284082,-0.1130543575,0.074800998,0.436026901,-0.0055075004,0.0017841377,-0.2595285773,0.2082757056,-0.2649086416,-0.2264890373,0.2413854897,0.1191601008,-0.0831423551,-0.122161001,-0.4942295253,0.4385454953,0.2539224625,-0.1163537055,-0.0749134794,-0.1809133887,-0.5298196077,-0.2167420387,0.2679248154,-0.0187716689,-0.0722125247,-0.1937800497,0.3418038189,0.1921720505,0.3103512228,0.6430558562,-0.386482656,0.7953374386,-0.1250498444,0.7483612299,0.2579936981,-0.2477056235,-0.377818048,0.116638504,-0.1784688383,0.2722213864,-0.2818516493,-0.1200631186,0.7026370764,-0.2252786309,0.3419845998,0.1061444208,0.2469955236,0.3040959239,-0.2483731955,-0.0491901487,0.5504582524,-0.0802227855,-0.2242528498,-0.150981456,-0.1534067094,0.1111330166,-0.0315542296,-0.1203731075,-0.1554094404,0.0617320798,0.0151116839,0.0270558968,-0.0637358427,0.2192265391,0.1471125484,0.2171940356,0.0741173998,0.0683925077,0.4290306866,-0.006116644,0.1055165529,-0.0645505115,-0.1504328996,-0.1419277936,0.1726531982,0.1282207072,0.0678987801,0.0891928524,0.2311721146,-0.1069797724,0.0154438466,0.1270231307,0.4437769949,-0.3041421175,0.1406869292,-0.160289079,-0.1061222926,-0.1936032474,0.1408739388,0.1087328643,-0.1834660321,0.1624329686,0.1247908548,0.3094665408,-0.1730695963,0.2074124515,0.1145593002,0.2654830515,0.0448613353,0.1117050797,0.0164180808,-0.1618009508,-0.101143539,0.2487526983,0.7665551305,0.293423444,-0.0981309041,-0.1918504685,-0.0074495161,0.2052231431,0.1254599541,0.0902863443,-0.0172139611,0.3012900651,0.4247050881,0.1712665409,0.350858897,0.4743336141,0.5267609954,-0.1261985302,-0.2109235376,0.1169244349,-0.3413845897,-0.4390027523,-0.0324556194,-0.4132475555,0.1679466665,0.1043307185,0.223832801,-0.168983534,-0.0138754928,0.0025527822,0.2727616429,0.2319285125,-0.0062529156,-0.0347080603,0.1677317023,0.2354774922,-0.3278361559,-0.1683433354,-0.017744882,0.2466843575,-0.0202035345,-0.0880928412,0.0660286173,0.1790383607,0.4836559892,-0.0730335638,0.0438799784,-0.2336488962,-0.086826764,0.0112860957,0.0730059966,0.1429340839,0.2016051859,0.0416048281,0.0553793311,-0.1324757189,-0.0326357819,-0.2036225647,-0.3690441251,0.0673273802,0.0478155538,-0.1053127348,0.3012902737,0.1895495206,-0.0706468001,-0.0284742992,-0.4572891295,0.0877392069,-0.1803108454,0.0098784408,-0.1076653674,-0.1534436494,-0.1983680278,0.0400028937,0.1168603525,-0.2781751454,-0.2616514862,-0.3224228621,-0.1180729493,-0.1744088978,0.1405443698,-0.0200541224,0.0511423238,-0.0312802456,0.1222996935,0.1008714736,0.129345417,-0.2491520643,-0.1672082841,0.1939766407,-0.0520039871,-0.0121717555,-0.3335229158,-0.2497054636,0.3890587389,-0.0699702948,-0.0141936084,-0.2141370177,0.0942139179,0.1669799536,0.1398722082,-0.2460222244,0.2176393867,-0.3044712245,0.2614798546,0.0384436436,-0.2167224288,0.394102484,0.2282124758,-0.1222438663,0.1697704941,-0.2478057742,0.1691340357,-0.0590032451,-0.235063985,0.0723133609,0.484038353,0.2467962503,0.6211673617,0.0489818566,-0.0133375674,-0.0001047923,-0.3113279939,-0.1315821707,-0.431189537,-0.0755139217,0.2071558088,-0.0991364494,0.025744576,0.4865767062,-0.085055843,-0.3330836892,0.2221700698,-0.1734428555,0.0490759537,-0.1999390274,0.1273285151,-0.0947731584,0.059305381,0.0013975933,0.4046107233,-0.5339136124,-0.1491030902,0.2374673486,-0.2787155807,0.2525823414,-0.2517894804,-0.945554018,0.351292491,-0.8524875045,0.0949762315,0.0165472962,0.2276023775,0.0475074612,-0.0270619392,0.1163009703,-0.0324518308,0.1658131778,-0.0538085625,-0.0845551342,0.2370497435,-0.5604740381,-0.1172572821,-0.1456517726,-0.0787851587,0.1936267316,0.5949997306,0.7412645817,-0.094430469,-0.2862768471,-0.3504448831,-0.0479068533,0.069677338,0.0392845869,-0.0138419038,-0.1023690999,-0.1808014512,0.0108176265,0.1023748368,0.1822670698,0.1334195733,-0.1541096568,-0.0875527263,0.1505203098,0.3036968112,0.2778019309,0.2423069924,-0.0387686603,0.3437796235,0.3173721135,0.1055742353,-0.1110975519,-0.15754655,0.2181521505,-0.369530946,0.0900226384,0.0998720303,0.0671070442,0.1695057899,-0.0323256105,0.2138267905,-0.2696321309,0.3467109501,-0.1918300986,0.3990117013,0.5268699527,0.1344355047,-0.5008938909,-0.3026986122,0.0710381046,0.5278068781,-0.2671489716,0.436906755,-0.4802762568,-0.4048759937,0.4786460102,0.2622500062,0.5893496871,-0.0001973089,0.2131251246,0.2467374355,0.2820867002,0.1871795803,-0.4422437251,-0.128349632,-0.2351831198,0.2189823538,0.0432998389,-0.0735478327,0.0681041032,0.2392938137,-0.1190579012,-0.0707464367,0.1343742311,0.8829224706,-0.0130431885,0.3027336597,0.4876137972,-0.399556309,0.3429582417,0.1393883824,0.2241421044,0.0627057776,0.0940346494,0.097355485,-0.0682936758,-0.1550722569,-0.2499284148,-0.2025585026,-0.5686978102,0.2977935672,-0.2047139555,0.0308005661,0.0798006803,0.144837752,0.0416582711,0.2406427413,0.0066326791,0.288566947,0.2132270485,0.3364417851,0.2534502447,-0.5096054673,-0.1977286786,-0.0098068425,-0.0152435564,0.1644136608,-0.2644812167,-0.3185324371,-0.6420421004,-0.0458642617,0.3040487468,-0.3656052351,0.0190449692,0.1011226326,-0.079265222,-0.1980586052,0.0806414783,-0.1534317881,0.0853769258,0.4080639184,-0.0744123384,-0.2806637883,0.01849439,0.3022505641,0.1215462387,-0.2107042521,0.4307021797,-0.4703114927,-0.2629357576,-0.1035872251,0.2154363841,0.1363151819,0.1198757738,0.0926353335,-0.1318544745,-0.1832281351,0.095146209,0.2641550899,-0.1274956018,-0.1099078432,-0.1724202782,-0.0183412004,-0.3060437143,-0.0105907451,-0.0599312298,0.0916231722,0.0973536298,0.1349281818,-0.3185404241,0.1697503924,-0.2577683032,-0.1210310012,0.1259122044,0.1538813114,-0.2014111429,-0.1586651206,0.1208085716,0.0411622934,0.0997675955,0.2575886548,0.0350962169,-0.2286946177,-0.0834933743,0.1424899101,0.0466305017,-0.3604576588,0.0659405738,0.1358203888,-0.084547624,-0.5128016472,0.2053451836,0.171624288,0.0137767084,-0.4263342917,0.1967772543,-0.0045447494,-0.0069766436,0.1540150493,-0.0409199856,0.0195887871,0.109384492,0.216730088,0.120891653,-0.2039531469,0.0436549038,0.2604890168,0.3350314796,0.1718847007,0.0119804759,-0.0260333307,-0.0632942468,0.1637773514,-0.1699793339,0.2126780003,-0.2245703042,-0.4430892766,0.206537798,0.1477205604,0.0679504126,-0.2368669808,0.4112255871,-0.2698408365,0.1705465466,0.053706035,0.2234908342,0.2558570802,-0.2167446166,-0.1678139269,0.4580506384,-0.1856303513,0.0927564204,0.1185214594,-0.046437759,0.0143661359,0.3236362636,0.0896296725,0.1318290085,0.4207837284,0.2065015584,0.2570444345,0.2392932475,0.3731354475,-0.188723579,-0.2324151993,-0.0579184704,0.4835402966,-0.3375782371,0.1452248245,-0.0031014248,-0.0090717077,0.0150641687,-0.5577237606,-0.036189843,0.3808557689,-0.3738011122,-0.263463974,-0.1039748564,0.0000507043,-0.0444148481,0.3579735756,-0.1306927949,-0.0440516546,-0.3228889406,-0.1085643396,-0.0701361597,0.1116295084,-0.2539830804,-0.2345301658,0.2184887081,-0.2175605595,0.2340624332,-0.4240419269,0.1349432617,-0.0556653365,0.3740637004,0.3734024763,0.2763729393,-0.1379146725,0.0604414158,-0.0782147422,0.0910299569,0.0376118086,0.1411616951,-0.0731158033,-0.3212688565,0.1342654973,0.0041144425,-0.1798272431,-0.356651634,0.0329824649,-0.1090813056,0.1966264844,0.2209125012,0.0509850122,-0.1480569541,0.1166607812,0.0628015101,0.0814494565,-0.1601275951,0.1489979923,0.0793074965,0.1826460212,-0.0843046457,0.070610404,0.0378762856,0.4189379513,0.0274144709,-0.1120420173,-0.3786530197,-0.3307703435,-0.3441821039,-0.0883446038,-0.1933036447,0.1373802274,-0.0022712855,0.0905381069,0.1760569066,0.1948624253,0.207484588,0.1234851927,-0.2083262056,-0.0998503342,-0.2851068377,0.4816954136,-0.1726653874,0.0396555029,0.0064430176,-0.3506652713,0.4216940999,0.0160460547,0.1311934888,-0.1760375798,0.0523305871,-0.021239562,0.0931963772,0.6377805471,0.1688714474,0.0617193729,-0.0346783958,0.1621492952,0.154145211,0.1810178906,-0.2202599496,0.0676567033,-0.0157296471,-0.0018751186,0.0582159013,0.027588414,-0.0795356855,-0.1120131165,0.1741838008,-0.0966245085,-0.2279877067,-0.0935885981,0.0079850471,0.1224939153,-0.2413401306,0.3319844604,0.073415406,0.2080664486,-0.3973875642,-0.0614512227,0.4318276644,-0.6654880047,-0.1201498881,-0.2693093717,-0.0117832627,0.3286216855,0.1732205451,-0.1057517454,-0.2817881405,0.5291823149,0.0278505087,-0.2813952565,0.3174489439,-0.0696155056,-0.0076537505,-0.0337328911,0.3865725994,-0.2410537302,-0.1833971888,-0.0738726184,-0.0449592285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1827","title":"Regarding On-the-fly Data Loading","comments":"Possible duplicate\r\n\r\n#1776 https:\/\/github.com\/huggingface\/datasets\/issues\/\r\n\r\nreally looking PR for this feature","body":"Hi,\r\n\r\nI was wondering if it is possible to load images\/texts as a batch during the training process, without loading the entire dataset on the RAM at any given point.\r\n\r\nThanks,\r\nGunjan","comment_length":10,"text":"Regarding On-the-fly Data Loading\nHi,\r\n\r\nI was wondering if it is possible to load images\/texts as a batch during the training process, without loading the entire dataset on the RAM at any given point.\r\n\r\nThanks,\r\nGunjan\nPossible duplicate\r\n\r\n#1776 https:\/\/github.com\/huggingface\/datasets\/issues\/\r\n\r\nreally looking PR for this feature","embeddings":[-0.2632015049,-0.468085736,-0.099150002,0.2329204082,0.4669876099,0.179101631,0.565231204,0.0583881103,0.2498599589,0.1583630145,-0.0594898276,-0.1817819774,-0.1609693766,0.4356863797,0.2764420211,0.1032606736,-0.0345524214,0.1798989177,-0.0421933904,-0.0337472409,0.0279306266,-0.0682156011,-0.0187377706,-0.239893198,-0.373208344,-0.1227851585,-0.1700692922,-0.0776974857,-0.1643121094,-0.1333782524,0.1556618065,0.2134788036,0.2178059518,0.2081834972,-0.0001241276,0.0796857923,0.2969023883,-0.1872229576,-0.4020967782,0.0592345558,-0.0743686035,0.0721684024,0.1251225621,-0.1761431247,-0.0363043472,0.2858171761,0.2100158185,-0.239590764,0.5396622419,-0.1171553656,0.0442747474,0.1911271363,-0.3172139823,0.0706197247,0.0298111625,0.3693019152,-0.0029411553,0.1426291466,0.3727179468,-0.1059762016,-0.1091678515,0.3823476434,0.0409047529,-0.0612255298,0.3190648556,0.067197673,0.0276602004,-0.3847416341,-0.0674839765,0.1311467141,0.4330565333,-0.0180650707,-0.3342523873,-0.5180974603,0.035455998,-0.3701883256,-0.0524727143,0.5328271985,-0.4467960298,0.1105366573,-0.6460332274,-0.3950530291,-0.2335518003,-0.0873800963,-0.0671495274,-0.442022562,0.0005977821,-0.0426378995,0.1476648152,0.2015482038,0.174654901,-0.1860179901,-0.037905775,0.4009208381,-0.2582012415,-0.2551917732,-0.492732048,-0.1282131523,0.1414798945,-0.1195068583,0.1141027287,0.029873874,-0.1964045614,0.1493570209,0.4409528971,-0.0741296485,0.0847697556,-0.0252403803,0.1113113686,-0.0633233786,-0.1812695265,-0.0285773333,-0.3152810931,0.1739256531,-0.1849362552,-0.3493721485,-0.3294318914,-0.0895838067,0.1547180861,-0.2190453261,0.1389656365,-0.0060194549,-0.2436479181,0.2768961489,0.012109532,-0.0696285665,0.4212339222,-0.1436369866,-0.2328459322,-0.3099014461,-0.0507986024,-0.0215292405,-0.3196798861,0.2591525018,0.1467587352,-0.0589601547,0.1510422677,-0.0860447437,0.0238058604,0.011709813,-0.053695906,-0.3656508327,0.1678112447,0.1777882874,-0.0598882586,-0.1946383417,0.0532309227,0.2816330194,0.0044461749,0.4289036691,-0.2943580449,-0.0125583913,0.3345974684,-0.0453902371,0.0172899254,-0.1331051588,-0.5839412808,0.4442558587,0.1086929813,0.070657894,0.1635872722,0.0585746616,-0.3735124469,0.1365891993,0.2022443563,0.4743243754,-0.2946819961,-0.4846610427,0.4842073917,-0.1176198274,-0.2114589363,0.1230802983,-0.2634784281,0.4000502825,-0.081605114,0.0179379992,0.0766694918,-0.2290443182,-0.0554191619,0.409716934,-0.0449022278,-0.0450233109,-0.005558236,0.5518739223,0.1613066941,0.1941318959,0.3144373596,0.7723010182,0.0116892597,0.1241589189,-0.0246140826,-0.3357358873,-0.2950282097,0.3241791427,0.1790588796,0.3113326132,0.0584207214,-0.1575377584,0.2341070324,-0.2895471454,0.0965902358,0.2309385836,-0.0149621787,0.2984659374,-0.0548662543,0.129609108,-0.5174821019,0.2101181895,-0.0285347868,-0.0058884975,0.1040725857,-0.1106870994,0.3226855397,-0.2678308189,0.1078642085,0.1750758141,-0.1533358991,-0.2377489358,-0.1605292559,0.081466876,-0.5633355975,0.6861085892,-0.5166232586,0.3127548993,-0.2019627541,0.2403113842,0.5240774155,-0.1214324981,-0.0924950838,-0.1447692364,-0.3484589159,-0.071762532,0.0833752379,0.376370579,0.0715366974,0.2575036585,-0.0531341881,0.0497499369,0.225040704,0.069686152,0.0813606605,-0.4179385006,0.0607529841,-0.1953144819,-0.3148581386,0.5322809219,0.0590445325,0.2542864382,0.3217590153,-0.0874095932,0.024870066,-0.1640102416,-0.1151327193,0.1052150503,0.1659173369,0.0863488391,0.412758261,0.0240986608,-0.4029485285,-0.3493637443,-0.0218009036,-0.0529440083,-0.0215669777,0.1809576303,-0.0854632035,0.0761678368,0.02016248,-0.170184955,0.2598951757,0.1064254865,-0.1068832651,0.1787575036,0.2484385818,0.0461449437,0.0361877978,-0.0159671921,0.288403362,0.0429010652,0.0543049276,-0.0393765047,-0.4868087471,0.3358736634,0.1315147877,-0.2304329574,-0.16262725,0.0884732977,0.1243264452,-0.3359367549,-0.182815358,0.1607520729,-0.2581202686,0.3132431209,0.0406793356,0.10998822,0.1740043163,0.0598663427,0.0407362618,1.0897593498,-0.0739132017,-0.1228031442,-0.1387331188,0.0312326513,0.1049934253,-0.0709807053,0.2238693237,-0.1143729463,0.362175107,0.0346353799,-0.2760070264,0.018717397,0.0612593032,0.2455221564,0.1006108001,0.1230501905,-0.2007463872,0.1592671424,-0.1412606984,0.0781063288,-0.0417852998,-0.3350070715,-0.3548156917,-0.4182071686,0.4223145545,0.1785781682,-0.1953584999,0.1437067389,-0.4721197188,-0.0522263795,0.3451586962,0.2097574472,0.0564891994,0.4852254689,0.0613133311,-0.2291212827,-0.2508637309,-0.0687261745,-0.0296294987,-0.625186801,-0.1351030171,-0.0663869679,-0.232356593,-0.1371657103,0.0428573005,-0.1107785329,0.7102029324,-0.8807191253,-0.1831404269,-0.0774266198,-0.154075399,-0.2065131217,0.1595052481,-0.0012327847,-0.3217660189,0.0612027198,-0.0218219087,0.0747759193,0.0943019837,-0.0888756886,0.047978878,0.2255611718,0.1987421662,0.0226040762,0.8768633604,-0.1995837986,-0.186742425,0.2347092479,-0.0898168385,0.4740725458,-0.2455242574,0.0286847483,-0.0543515235,-0.1358460635,-0.2090152353,0.1068343967,0.0933724791,-0.199475348,-0.3706570566,0.2643801272,-0.0774580538,0.0059589101,0.1213975623,-0.1279480904,0.3137634397,0.0059520667,0.0219533294,-0.1004974321,-0.2415060401,0.0902695954,0.1403574646,0.4972282052,-0.155991137,-0.2538508475,0.3396686912,-0.3150261939,0.0378741883,-0.0028301768,0.0899938568,-0.0194625817,-0.0611519329,0.2567858696,0.1037866175,0.8126317263,0.0606344976,-0.0326968767,-0.0952810794,-0.5470388532,-0.1325907111,0.006890628,-0.072368212,-0.1761171669,0.0648725331,0.5520423651,-0.0303107873,-0.2136164606,0.0571383014,-0.0601296499,-0.4030556083,-0.3345966339,-0.2204543799,0.1529431343,-0.1573850811,0.1005616784,0.1574794948,-0.0093157934,-0.0428495519,-0.0628983527,-0.3317591846,0.3553682864,-0.0097741997,-0.2797499001,-0.0246873721,0.3208675086,0.2469788939,0.252558887,0.230952546,0.1478930861,0.3912725747,-0.2894180417,-0.1237775609,-0.1614850014,-0.061247021,0.3951776922,0.6219514608,0.3195205033,0.1118044853,0.1244126558,0.0870960355,-0.5540974736,0.1933850348,0.1316363364,-0.082442835,-0.2706854939,-0.5424346328,0.6589618325,0.1694328487,0.0924662501,0.3212853968,-0.2426027656,-0.2622658014,0.0953930244,0.158835575,0.8489964008,-0.4011645615,0.7009202838,-0.1176703647,0.3689952195,0.144116208,-0.1941661239,0.2351492792,-0.3395088315,-0.3197266757,-0.1459116787,-0.2254333198,-0.1632439345,0.2449162751,0.2051555067,0.1016823277,0.0175315347,0.0851339996,-0.1018371433,-0.0956442952,-0.0343017168,-0.341614753,-0.2774917483,0.0280192383,0.2076084018,0.2439269423,-0.2214866728,0.0987575501,0.1517828107,-0.1462423205,-0.0419627577,-0.1010665894,-0.3236864805,0.0407442935,0.0308299307,-0.1633769572,-0.2194108516,0.2420033962,0.2149681747,-0.0566315576,0.0442208983,0.0164423287,-0.0770586953,0.189093098,-0.1804019213,-0.0263507217,0.3680513203,0.1017417759,0.1164005175,0.0547250323,-0.1051351279,-0.4320550561,-0.105783999,0.0497202054,0.0989704952,-0.132891804,-0.1413413882,0.1455928385,0.3046974838,0.0807178617,-0.0013519984,0.2442611605,-0.1165174618,0.3285895884,0.1679162681,-0.1967667043,0.0654145032,0.5832528472,0.0080817174,-0.395077914,0.2882790864,0.0391305164,-0.0428459495,-0.1212581471,0.2048997134,0.4110337198,-0.2016686052,0.2541601956,0.1990087479,0.1658383608,-0.2656558752,0.2635586858,0.086360462,-0.198212117,-0.0797113106,-0.2356249988,-0.3706317246,0.2532756031,-0.0770234168,0.3554752469,0.2899661958,-0.0598163083,0.2330658287,0.1335668564,-0.1764582396,0.2434308976,-0.0335287191,0.1862058043,0.0985909849,0.1680076569,0.0055608209,-0.0927542225,-0.069449991,0.1562855542,-0.0718559101,-0.1370573193,0.0130129717,0.2523568273,0.0460185893,0.0859373361,0.0539905727,-0.2019209415,-0.1993930936,-0.116280064,0.2948442698,0.3078059554,-0.0843667388,-0.143144995,0.2277823985,0.4799517095,-0.1099091023,0.3482234776,-0.0971689969,0.3365724683,0.2355245948,-0.1513783038,-0.0935464799,0.1542483121,0.0231101699,0.0901674479,0.4358460903,-0.1175744683,0.0539021194,-0.3602250814,0.0847236887,-0.1166029125,0.3929209709,0.3516824841,-0.5932661891,0.1105789617,0.3839470744,0.1071924493,-0.0505431816,-0.1299688965,0.0718918368,-0.3392384648,-0.0544618033,0.324835211,0.2759063244,0.0457553677,-0.1920530051,-0.1333240271,0.2309329957,-0.1885624677,0.0445347391,0.2638915479,0.0225237738,0.0193033945,0.0289165974,0.5759753585,0.0313836522,0.1070239395,-0.1045103818,0.136479646,-0.2083183974,0.1582903564,-0.2502234578,-0.4476267099,0.0429151766,0.5544227958,0.0016198276,0.2327352017,-0.4011294842,0.2408512831,-0.0588861965,-0.3377067745,-0.1405555159,0.1517622918,-0.1581978947,0.0997976959,-0.1751717776,-0.0591877513,-0.1134904474,0.5091202855,-0.0160163119,-0.2844645679,0.1559485942,0.2685894966,0.0889920518,0.033908952,-0.1269688606,0.2816438377,0.1883296818,-0.127179116,0.0553340316,-0.4070908129,0.294917047,0.4714074433,0.4851033986,-0.1103820428,0.2921581864,0.1539163738,0.0673773885,-0.0036311897,0.1078241989,-0.09134534,0.2335291803,0.2347722799,0.1070419997,0.2495939136,-0.0253234953,-0.0622498058,0.2898011804,-0.0056397645,-0.1563783139,-0.5808227658,0.1683896333,0.0929522887,-0.0693560243,-0.1918279529,-0.0167591721,-0.2876630425,0.028080862,0.307395786,-0.4185906053,0.0441150181,0.5073511004,-0.0133742373,-0.2370623499,0.4195361733,0.4078361988,0.1850650758,-0.3134997189,-0.1833077669,-0.2968603969,-0.0198693741,-0.0718252733,0.0486336164,-0.0715597346,0.138413161,0.2294891328,0.0838927999,-0.0862424746,-0.3571361601,-0.0184370689,0.1077983007,-0.3906847239,0.0637060553,0.0060393447,0.4225224853,-0.1130787656,-0.2604913712,0.1225655228,0.0032463975,-0.2737011313,0.3567539454,0.1672005057,-0.2453481108,0.0007319158,0.3963027,-0.4080678225,0.3010903299,0.0301310215,-0.0118808867,0.0190909412,-0.2576216757,-0.3837069273,-0.1107645333,0.1670376807,0.0719521493,-0.2785293758,-0.1473815739,-0.1138533503,-0.2208196968,0.1253255308,-0.1841415018,-0.1502671391,-0.2784177363,-0.0265321396,-0.0040361332,0.1768232286,0.4018389583,-0.1601702422,-0.1389230043,-0.2340750694,0.0076760435,0.4682448208,-0.560978055,-0.0913820267,-0.2107741684,0.1629374176,0.0161941778,-0.0260398984,-0.5070508718,0.123903282,0.312407583,-0.0882936046,-0.1128436103,0.4889924824,0.1494767815,-0.0145034548,-0.0391127802,0.2132435441,0.0604408272,-0.0629916787,-0.4190783501,-0.390345186]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1827","title":"Regarding On-the-fly Data Loading","comments":"Hi @acul3 \r\n\r\nIssue #1776 talks about doing on-the-fly data pre-processing, which I think is solved in the next release as mentioned in the issue #1825. I also look forward to using this feature, though :)\r\n\r\nI wanted to ask about on-the-fly data loading from the cache (before pre-processing).","body":"Hi,\r\n\r\nI was wondering if it is possible to load images\/texts as a batch during the training process, without loading the entire dataset on the RAM at any given point.\r\n\r\nThanks,\r\nGunjan","comment_length":48,"text":"Regarding On-the-fly Data Loading\nHi,\r\n\r\nI was wondering if it is possible to load images\/texts as a batch during the training process, without loading the entire dataset on the RAM at any given point.\r\n\r\nThanks,\r\nGunjan\nHi @acul3 \r\n\r\nIssue #1776 talks about doing on-the-fly data pre-processing, which I think is solved in the next release as mentioned in the issue #1825. I also look forward to using this feature, though :)\r\n\r\nI wanted to ask about on-the-fly data loading from the cache (before pre-processing).","embeddings":[-0.1726847142,-0.2835986018,-0.1183969602,0.2190789431,0.3321972191,0.0917301774,0.5707204938,0.1317177862,0.1204758659,0.0670047775,0.0713878274,-0.0857139006,-0.0965600237,0.2116191685,0.2725099623,0.2309678495,0.0025834758,0.0690006167,0.0591212176,-0.0284568463,-0.0402434245,-0.0728218779,-0.0123700965,-0.362198323,-0.4274348319,-0.0935060233,-0.1793473214,0.0571449585,-0.153655991,-0.2890402675,0.1886149794,0.2540030479,0.2708257139,0.1172411516,-0.0001203638,-0.0124503793,0.3828080595,-0.1435648203,-0.3206498921,0.1220704764,-0.0993443206,0.1157685742,0.106421046,-0.1725357622,-0.0760356486,0.4242566228,0.1247473955,-0.3189669847,0.4757882953,-0.081970863,0.0456120484,-0.069350563,-0.3194617033,0.030737374,0.2106879205,0.1520231217,-0.0428887308,-0.0413794033,0.3662909865,-0.1179742813,-0.1022316962,0.3062210381,-0.0025014977,0.0041443375,0.2204100341,-0.0154390763,0.0600902811,-0.4758675396,-0.0239071082,0.1586290151,0.5854740143,0.0645013526,-0.3178856969,-0.3903407753,0.0198414288,-0.3572824895,0.0862496868,0.382142365,-0.3288044333,0.0932317525,-0.5378326178,-0.4087218344,-0.2208578736,-0.114024736,-0.090202786,-0.2912571132,0.0688989162,-0.0724284202,0.0858769715,0.2078725994,0.6285194159,-0.1119770184,-0.1109058037,0.4937479198,-0.272139281,-0.2085574567,-0.4864164889,-0.1967238635,-0.0057810238,-0.1826279014,0.2753618062,0.1797651201,-0.1919866353,0.0621586256,0.3562072217,0.0273071248,0.1653118432,-0.031149257,0.0064679109,-0.2958640456,-0.2373790443,-0.0634201765,-0.4045855105,0.136860624,-0.1152536124,-0.2681531608,-0.4467125535,-0.250644356,0.2545191944,-0.2527233064,0.1836283952,-0.0403485224,-0.2313980907,0.312717706,-0.028800603,0.0083063543,0.2975400686,-0.1674126387,-0.077529043,-0.2540169954,-0.0478608571,-0.0181819089,-0.2493307739,0.1330200434,0.2042916268,0.0589198172,0.1179031208,-0.1313331276,0.0041415063,0.0270401686,-0.0003011711,-0.3179264069,0.3636932969,0.2754960954,-0.0150987152,-0.2916888893,0.0312473699,0.333538115,-0.0201603994,0.5988048315,-0.2974314392,-0.0827195048,0.3635725379,0.0118181836,-0.0296235587,-0.2398855984,-0.2626531422,0.4184547961,0.2489563823,0.0859438553,0.2036076039,0.0772561058,-0.2851469517,-0.0212870687,0.1105753183,0.350160569,-0.4026871324,-0.3087047637,0.2794791758,0.0222667381,-0.1220834181,-0.0091872336,-0.248816818,0.3373199701,0.0116719166,0.037483491,0.1123319566,-0.2921566367,-0.0236151498,0.4355233014,0.0490387045,-0.0319646671,-0.009075881,0.4269596636,0.1020324528,0.1715197563,0.295953244,0.7635238171,0.0538489595,0.0748611689,-0.1958502978,-0.3414077461,-0.2395244092,0.3621118069,0.2415037155,0.2589144409,0.2517335117,0.0273116045,0.0837511122,-0.2263398767,-0.0170133896,0.1305214167,-0.0317249894,0.1669828594,-0.0416582264,0.2819933891,-0.4686065018,0.2828407884,-0.1840272546,-0.0459719226,0.1098968163,-0.0095281713,0.3888610899,-0.2076938301,0.0993757248,0.1737114936,-0.0955861062,-0.2128090113,-0.0997159183,-0.0099680647,-0.4909341931,0.8521465063,-0.4758826494,0.2114950716,-0.1044537723,0.107416369,0.4491933584,-0.1081207469,-0.19637914,-0.1719347388,-0.2991917133,0.0181719642,0.0061659301,0.4675185382,0.1323865056,0.2539481521,-0.0984479487,0.1048576757,0.2965150774,0.0731375143,0.2004057616,-0.3838969767,0.0407693163,-0.1957342029,-0.4062644839,0.6067577004,0.0608738475,0.1954904348,0.2967914045,-0.2305458933,-0.0120987436,-0.1654074788,-0.0860128924,0.1190435216,0.0079784598,0.106078364,0.5033367872,-0.0313715898,-0.1819934845,-0.2241875976,-0.039342802,-0.0225382932,-0.1010949016,0.1562383324,-0.0333746783,-0.0968439132,0.0403099582,-0.0376495793,0.3064706028,0.086568445,-0.0088048233,0.1440639645,0.2152051926,-0.0710984617,0.1003785059,-0.0256861597,0.4002476633,0.0196015462,0.066852279,0.054219991,-0.4141884744,0.3761070073,0.2402523011,-0.2068132311,-0.1194646209,0.0249521323,0.1802029461,-0.2877338529,0.0511291102,0.0785601512,-0.2679390609,0.2447863221,0.0626737773,-0.0354731083,0.2312708497,-0.1007864252,-0.0505560227,1.058915019,-0.0778698623,0.0396899469,-0.1933262646,-0.0724977404,0.2199304551,0.0177281313,0.1843344718,-0.11261338,0.2349985838,0.1298400313,-0.1127245799,0.2563572228,-0.1056977063,0.1894396096,0.219392702,0.1237749234,-0.3309439719,-0.0084988605,-0.2049747854,0.0421055704,-0.1971979886,-0.4081793427,-0.2879915237,-0.4751538336,0.3778953552,0.112087518,-0.1671583802,0.0250952411,-0.4167984128,-0.0401691794,0.2077399194,0.3384683132,0.1406713873,0.3773111403,0.0208095033,-0.1806992143,-0.1592862159,-0.1036747843,-0.2029184848,-0.6760934591,-0.093938835,-0.1416723728,-0.1950516105,-0.0356874391,0.0936200768,0.0362055525,0.6651348472,-0.8735558391,-0.0531983189,0.0262281261,-0.1240393519,-0.2042838335,-0.0786603913,0.1373702139,-0.2599746883,-0.0103772916,-0.0497465432,0.2246382236,0.1929009855,0.0336729549,0.0733998492,0.2787472606,0.0319000855,0.1225322708,0.8831708431,-0.371219635,-0.065941833,0.2678942978,0.1573865861,0.3801428676,-0.2386514395,0.2636007071,-0.1346945912,-0.1658603847,-0.1958541572,0.1565245241,0.0872743577,-0.5369459987,-0.4127477407,0.4158640206,-0.1284589767,0.116991736,0.2079683691,-0.322535038,0.2557944357,0.0267961081,0.0214499962,-0.1968212575,-0.2115892619,0.0191262551,-0.0576586686,0.6054168344,-0.2114352882,-0.1240902469,0.3219369948,-0.2747213542,0.0078739533,0.0306296926,0.2162823975,-0.0298612621,-0.1958163232,0.3248186707,-0.014948532,0.6353758574,-0.1768060029,0.0008968615,0.0656326041,-0.461817652,0.0665302128,0.0197554454,-0.0642730072,-0.1148599535,0.0466996841,0.2646927238,0.135175243,-0.117320247,-0.0877334327,-0.2665193975,-0.3192571998,-0.3963135183,-0.0701058283,0.1605161875,-0.212730065,0.0817422941,0.1631474048,-0.040869236,-0.2503907681,-0.0283840913,-0.3545411527,0.4247928262,-0.1319829673,-0.2478570491,-0.0554513037,0.2683358788,0.1844049543,0.2386129946,0.1415825933,-0.015265055,0.2034436315,-0.2024108022,0.0800784826,-0.1507410258,-0.1296240836,0.2492766082,0.6636570692,0.286067605,-0.0303023066,0.1506052613,-0.1189297289,-0.2937456369,0.1309518367,0.0313464515,-0.2929352522,-0.162003845,-0.5915787816,0.6851963997,0.2224940211,0.023062611,0.3544329107,-0.5506945848,-0.3202627599,0.2900858223,0.3368039727,0.9200327396,-0.3764573038,0.5159749389,-0.2994262576,0.309664011,0.0008142353,-0.3276216984,0.2093081474,-0.3369632661,-0.2832555175,-0.1473209113,-0.1753264219,-0.1173820049,0.3771118224,0.1250200719,-0.0722639784,0.1168524548,0.0181828532,0.0548080318,-0.0946767032,0.0608398691,-0.4045918882,-0.0636876598,0.0648535714,0.2162375897,0.2144351006,-0.2849906087,0.0548532307,0.1731054932,-0.1351608783,0.0026148583,0.062008094,-0.2733359337,0.0041463855,-0.2019814998,-0.2839854956,-0.3501133323,0.148651287,0.3003534675,-0.1588252038,0.0428688675,0.2137990892,0.2311343253,0.1083908454,-0.0447534807,-0.0045008669,0.3801799417,-0.0869496092,0.1371850669,0.1390541345,-0.245412305,-0.5426343679,-0.0188657548,0.1041588634,0.2103516459,-0.0124632837,-0.216229707,0.2761100531,0.0940301493,0.0988000557,0.073687546,0.2683965564,-0.1985176057,0.3451650143,0.0830182135,-0.1866431534,0.0497790463,0.6502137184,0.1749261916,-0.4405233562,0.178612411,0.1150550023,-0.0882909223,-0.0799116269,0.0166133437,0.2343143225,-0.2239985317,0.0934885964,0.0793305188,0.1605614126,-0.0854919031,0.2037151307,0.0990557894,-0.0556939244,-0.1945218593,-0.2388638407,-0.3176458478,0.166285798,-0.138989076,0.3290902674,0.15828453,-0.2370779663,0.2533075213,0.1814437807,-0.2273147851,0.1966858655,0.1132530719,0.2126767337,-0.0204177052,0.1259811819,-0.1217228472,-0.0782732666,-0.0608332194,0.0926747695,0.0013122847,-0.1889594644,0.08946722,0.2597551942,0.1029034778,-0.0615427978,-0.0641864687,-0.2494578809,-0.2111489028,-0.0103687551,0.2952494323,0.087061353,-0.1629939377,-0.1616751105,0.2854116261,0.5793213248,-0.1693763733,0.4471299052,0.0124735693,0.3442240059,0.0885529816,-0.0301330499,-0.2365308702,0.1098894179,0.2594616413,-0.091769591,0.2493804097,0.0170957074,0.1877783835,-0.4105764627,0.1509673297,-0.1571805179,0.1854480207,0.273578614,-0.4995346963,0.0643147528,0.3117897213,0.1444229484,0.0699683353,-0.1901301593,-0.1092532873,-0.3621828556,-0.1649497151,0.2804121375,0.5584884286,0.0302951578,0.000343113,0.0439927876,0.3751169741,-0.4146692157,0.0782157332,0.4015363753,-0.0463289991,0.0490774922,0.0858730823,0.5185032487,-0.0125632323,0.1787185818,0.0150603093,0.2180618942,0.0458537564,0.0936096162,-0.267770797,-0.4308397472,0.0587029345,0.6751397848,-0.0713496804,0.4519225955,-0.3695746064,-0.0558469296,-0.0169712268,-0.25552845,-0.1590129435,0.1286561936,-0.1327878237,0.0760777965,-0.1052349061,-0.1248513758,-0.1445704997,0.4197894633,-0.100756377,-0.1425279081,0.2564867139,0.2617850006,0.0130307963,0.1989991069,-0.1012509242,0.0968433619,0.2134023756,-0.2731975615,0.2259692997,-0.4411214888,0.1773986965,0.4834467769,0.347346276,-0.0813815147,0.3185938001,-0.1027416065,0.0960776657,0.096718058,0.069189325,-0.0256328229,0.3193223476,0.3588232696,0.0635117888,0.1042463854,0.0253968369,-0.0752459764,0.4104229808,-0.0105711445,-0.0307152849,-0.3659185469,0.1415156722,0.2978209853,-0.1230254024,0.0002620419,0.0063752267,-0.3958482444,0.0855415538,0.2274093032,-0.4957816005,0.0051388103,0.4473087788,0.016177712,-0.2025238425,0.5034615397,0.3643020391,0.1137145683,-0.2773648798,-0.2858846784,-0.3451362252,0.0241380911,0.0438249148,0.1658494174,-0.0062255012,0.234655425,0.1811325848,0.1554619819,-0.1374716908,-0.4772333205,-0.10181465,0.1482325196,-0.3347150981,0.228040427,0.1797850728,0.4630303085,-0.2251701504,-0.2288309336,0.1836902201,0.1424229294,-0.2046662271,0.3740409613,-0.087462604,0.006389962,0.073138617,0.4084907174,-0.4653453529,0.3128404915,0.0173085649,0.0075583765,-0.1158367917,-0.3511052132,-0.2685951889,-0.1359149069,0.1138921976,0.0028499062,-0.21716097,-0.0792466253,-0.0178184398,-0.2589983344,-0.0852824971,-0.1233858392,-0.1446760148,-0.4979203641,-0.0809314176,-0.0383524969,0.2050119936,0.3365011811,-0.0431852937,-0.2461984754,-0.3440424502,-0.0021832497,0.3099215925,-0.4316575229,0.0084865736,-0.2264376581,0.0324326567,-0.0082221283,0.1528892815,-0.4410767257,0.1689273566,0.2995244861,-0.1188209802,-0.0464240126,0.3388530016,0.1388069689,-0.0459358059,-0.1166079193,0.2619570494,0.0111126183,-0.0911463052,-0.6637380123,-0.4207083285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1827","title":"Regarding On-the-fly Data Loading","comments":"Hi ! Currently when you load a dataset via `load_dataset` for example, then the dataset is memory-mapped from an Arrow file on disk. Therefore there's almost no RAM usage even if your dataset contains TB of data.\r\nUsually at training time only one batch of data at a time is loaded in memory.\r\n\r\nDoes that answer your question or were you thinking about something else ?","body":"Hi,\r\n\r\nI was wondering if it is possible to load images\/texts as a batch during the training process, without loading the entire dataset on the RAM at any given point.\r\n\r\nThanks,\r\nGunjan","comment_length":66,"text":"Regarding On-the-fly Data Loading\nHi,\r\n\r\nI was wondering if it is possible to load images\/texts as a batch during the training process, without loading the entire dataset on the RAM at any given point.\r\n\r\nThanks,\r\nGunjan\nHi ! Currently when you load a dataset via `load_dataset` for example, then the dataset is memory-mapped from an Arrow file on disk. Therefore there's almost no RAM usage even if your dataset contains TB of data.\r\nUsually at training time only one batch of data at a time is loaded in memory.\r\n\r\nDoes that answer your question or were you thinking about something else ?","embeddings":[-0.286943078,-0.3630418181,-0.1495332122,0.2731824219,0.414362222,0.1709423512,0.4223019183,0.1292583793,0.1851403415,0.0802383274,0.121642299,-0.0679904073,-0.2427961379,0.1964000762,0.1893507391,0.2370749265,-0.0065636951,0.1383550018,-0.1860133559,0.0667945966,-0.0094696237,-0.2281021625,-0.1345320791,-0.3600880802,-0.3519837558,-0.1788701564,-0.2061012089,-0.081674464,-0.0664418563,-0.1807874441,0.0383012444,0.1454581469,0.2661631107,0.2168910652,-0.0001134483,0.0263619404,0.1878576428,-0.1615202576,-0.4358328581,0.1928040832,-0.0849268138,0.0534444526,0.1489462852,-0.2591665685,-0.0432943068,0.2636418343,0.1821040213,-0.3938759565,0.3936271966,-0.0597929992,0.1287946999,-0.0312944241,-0.3372347951,0.0115782944,0.0496058911,0.1317645311,0.0707556158,0.0703822523,0.2710404396,-0.0599783845,-0.1753790826,0.3019686937,0.004044468,0.04812067,0.3422016799,0.0536040962,0.0754288882,-0.3361243308,-0.0324160419,0.0897561833,0.5630343556,-0.0347455889,-0.155330807,-0.3663960993,0.1146519333,-0.2698690593,-0.0625268295,0.489333421,-0.4659104049,0.0407533422,-0.5683214068,-0.4256089032,-0.2526451349,-0.0436525494,-0.0367207378,-0.4736000896,0.132329464,0.0079848934,0.1394614279,0.1308273971,0.343752265,-0.231676504,-0.0460172929,0.4015500844,-0.2577864826,-0.0960031673,-0.6005839109,-0.2998289764,0.0383353271,-0.2760585845,0.1747638434,0.1581802815,-0.095134899,0.1061405912,0.3347076774,-0.019386448,0.0244126618,0.0254782867,-0.0001031627,-0.275926739,-0.2644038498,-0.0578746609,-0.4595869184,0.0949928835,0.0331917182,-0.3943001032,-0.4661898017,0.0087624053,0.1717686653,-0.2818394005,0.0412119217,-0.013791346,-0.2435890436,0.2964889705,0.0139826648,-0.0481421165,0.4604890943,-0.2862051129,-0.1249799207,-0.2746886313,-0.0637196377,0.0656898022,-0.3060081303,0.2646063566,0.1331712753,0.1531426758,0.0667293072,-0.1371395141,-0.0669994801,0.036404524,0.0199883021,-0.3356121778,0.2059192061,0.1551752388,-0.1178060398,-0.2792707682,-0.0381688736,0.3118372858,0.006095598,0.5027794838,-0.2631541789,-0.069812201,0.4293095469,0.0817824006,0.0833737627,-0.1546327919,-0.6008709669,0.3655919731,0.1369956434,0.1333419383,0.070576556,-0.0077770185,-0.3763471842,0.0051170397,0.2648199797,0.3271280229,-0.3991860747,-0.3566908538,0.4185029566,-0.0158250816,-0.1391628832,0.1062798724,-0.3086325228,0.4188513458,0.0300433338,-0.1001220644,0.0982156247,-0.2143739313,-0.0543934517,0.3533630073,0.0158604402,-0.1013925895,0.0247092582,0.572217226,0.0938572213,0.2492914796,0.329151988,0.7063392997,0.1441548765,0.1200050861,-0.1268433928,-0.2603088915,-0.3210431337,0.3683959842,0.164371863,0.1685014516,0.0841636136,0.0194546748,0.234720543,-0.2200764269,0.0599162243,0.2151897699,-0.0518270917,0.1326141506,-0.1009771004,0.2299596518,-0.4384998381,0.2532017827,0.0444661118,0.0821051747,0.1139745936,-0.0054874746,0.3631191552,-0.2487814277,0.1507816315,0.2011341751,-0.0444285497,-0.3101141155,-0.225204125,-0.0229294375,-0.4112915099,0.6960084438,-0.4796499908,0.2173445076,-0.1416359097,0.0989393145,0.5003316998,-0.0849593729,-0.0852534771,-0.2335910052,-0.4635132253,0.0353132933,0.0689441934,0.3818085194,0.1407142431,0.1782442778,0.0316990651,0.0492442623,0.1813828796,0.0268362239,0.1826771796,-0.3104831576,0.1619306803,-0.1521638334,-0.3586111367,0.4180773497,0.0945809186,0.2759921551,0.3086817265,-0.1523555964,0.0141109461,-0.1444022208,-0.0534743853,0.1024145037,0.0438667648,0.226858452,0.4621651173,0.1707957387,-0.2648866773,-0.345797509,0.0504017062,-0.1120194867,-0.0432916172,0.2249034345,-0.1269934475,0.0364626758,0.0329791345,-0.0982062221,0.2339931279,0.2031129748,-0.0120156836,0.1703813374,0.2188822925,0.0095251286,0.0882775038,0.062237855,0.4783914387,-0.1956154704,-0.0780261382,-0.0296746604,-0.2640933096,0.2508116364,0.1417489797,-0.1726943702,-0.1391683221,-0.0242390316,0.0047018984,-0.3065868616,-0.0998248756,0.3313088715,-0.1850053966,0.3601140678,-0.0126785412,0.0224999301,0.1696310043,0.0071446649,-0.0975894257,0.9094578028,-0.046964597,-0.2100371271,-0.021173507,-0.0736701265,-0.0035659403,0.0343072638,0.3430948257,-0.0483931974,0.3704320192,0.140648514,-0.1775661558,0.1492753774,0.2240161002,0.1529071629,0.2234522253,-0.0167874638,-0.3906622827,0.1024204642,-0.2542730272,0.0477848612,-0.0878183767,-0.338757962,-0.224340111,-0.3469252288,0.3597928286,0.0980084389,-0.0353977121,-0.007406882,-0.4434201419,-0.0958467573,0.1404809058,0.1994834989,0.1051898003,0.3841191828,0.1694537401,-0.1420047879,-0.1363412291,-0.0258121118,0.0546567142,-0.4763892889,-0.0898559168,0.0492235459,-0.1751328707,-0.0948596522,0.040258389,0.0185468849,0.8389362693,-0.8290775418,0.1393468529,-0.0379304364,-0.2141063213,-0.2819705307,0.0518494621,0.0399550274,-0.2387923598,-0.0956739113,0.0473068841,0.3101844788,0.1588173509,-0.102908656,0.0373466238,0.298878938,0.1454367489,0.1473434269,0.7117272019,-0.212352708,-0.1662727445,0.1837857813,-0.0557092316,0.2883814871,-0.1327329874,0.2319425642,-0.1628266573,-0.2090507299,-0.2608184516,0.1789029241,0.0665001944,-0.3049507141,-0.4226719141,0.2403942645,-0.0813541636,-0.0320503041,0.3473742604,-0.3740388453,0.3303478658,-0.0016484029,-0.0334107429,-0.0969609395,-0.235749796,-0.0053578382,-0.1043845564,0.4143750668,-0.2544370294,-0.2346518785,0.4172889292,-0.2592050731,0.0629912019,-0.0961620137,0.1816753745,0.0727681518,-0.140225932,0.2659955323,0.1421539783,0.5721114874,0.0801068097,-0.129056558,0.0482354835,-0.442424953,-0.2215891927,0.0046019987,-0.1377782822,-0.0947394297,-0.0602909289,0.3573643863,-0.0340163894,-0.2430554628,-0.0223230161,-0.0218736455,-0.406721741,-0.3591226935,-0.1792962849,0.1477944106,-0.0813553929,0.0700003058,0.1895980686,0.0160893481,0.0467331074,-0.1547943503,-0.3114152849,0.3588238358,0.0349277779,-0.290246129,0.0128696952,0.278367281,0.2972322702,0.1560090631,0.167665422,0.2824077308,0.2115089297,-0.2807257175,0.0307464115,-0.1402087212,-0.0448341072,0.3898746967,0.4764642417,0.2214212567,0.0638551936,0.1461298466,0.0189742614,-0.5502631068,0.1762140244,0.0959275067,-0.1931820959,-0.2183631361,-0.6674816012,0.70443362,0.2265467346,0.030438887,0.3714713156,-0.4017287791,-0.3358765244,0.1563834846,0.3137791157,0.7594898939,-0.5083619952,0.5862455964,-0.2089076787,0.3120793104,-0.0432063043,-0.2429618239,0.2859002948,-0.2782697678,-0.218647778,-0.0573672913,-0.2283290476,-0.1376565099,0.4003313482,0.1875219345,0.0805417746,0.1148524433,0.0512248054,-0.0216505695,-0.0426980183,0.0049441545,-0.363723278,-0.0700795352,0.1293797344,0.2027829289,0.0689090565,-0.2274380922,0.1318281144,0.056933675,-0.0522894636,0.1077696234,-0.0969800875,-0.325219512,0.0566040203,-0.2804447114,-0.1579392403,-0.2129371464,0.2018792927,-0.0126548642,-0.1298573762,0.109461084,0.1635957509,0.0348536894,0.2512392402,-0.0455206446,-0.021439055,0.389870882,-0.021895973,0.029210642,0.1299551427,-0.2188918293,-0.3678267598,-0.0035802564,0.0904565528,0.1948286444,-0.0676594004,-0.1820919663,0.1736174077,0.2327826917,0.206348449,0.1096125916,0.1880124807,-0.1446280777,0.4939877093,0.2086305916,-0.2386705577,0.0895971805,0.4949341416,0.1367442161,-0.3033528626,0.2160906196,-0.0414564535,-0.1133973077,-0.1351358443,0.2718051672,0.4831981063,-0.0381482579,0.1606023312,0.151920557,0.2268359214,-0.0794175193,0.1180792674,0.0783066228,-0.1671869904,-0.1144032627,-0.103643842,-0.4050602913,0.2820176184,0.1186199337,0.3430781662,0.1811318099,-0.2666763365,0.1819149405,0.1119837984,-0.2680925727,0.1936521381,0.086563535,0.283942759,-0.1136800721,0.1599310189,-0.0207436569,-0.0552258268,0.0529272631,0.0634664074,0.0021787353,-0.2107870132,0.1371091306,0.2034443021,-0.0163181834,0.0004696983,-0.0077615059,-0.4162228703,-0.149364382,-0.1376383305,0.1833476871,0.166113317,-0.0933712795,-0.1546976417,0.3541572094,0.4425534904,-0.2050752789,0.4429436624,-0.1198163629,0.39136675,0.1933415234,-0.078200981,0.0165948346,0.0673755333,-0.0023892517,-0.0114155561,0.3575136662,-0.1186276674,-0.0139956633,-0.3259765804,0.1605261415,-0.1186755151,0.288454622,0.3929074109,-0.5295219421,0.0210717805,0.2882755101,0.2199261785,0.041338075,-0.0999147072,0.001391249,-0.3731642067,-0.0823175535,0.2886896431,0.2444267273,0.0573416948,-0.2947581112,-0.1455802023,0.2715757489,-0.2606422901,-0.0618506037,0.2537121177,0.1017387733,-0.0880400985,-0.0667773783,0.5002996922,0.0994487032,0.2923493087,-0.1078582332,0.331366837,-0.0780323297,0.0673841909,-0.3112196922,-0.360411793,0.0117536066,0.6964067221,-0.0603914522,0.1763447076,-0.3629172146,0.1245694235,0.1141187251,-0.3701484799,-0.0549518056,-0.0252004284,-0.1251753569,0.2137401998,-0.1683695465,-0.1392388493,-0.1488051116,0.4257665873,-0.0214825552,-0.316121459,0.1841951311,0.3428277075,0.0629125312,0.2339558452,-0.0558647364,0.1977966279,0.092999652,-0.2929558158,-0.0018574389,-0.3005573452,0.2502619028,0.4217050672,0.3071512282,-0.0888091102,0.2305086255,-0.0098295659,0.06922777,0.1632234007,0.0619340651,-0.1380838156,0.113798745,0.3265502751,0.2447982132,0.1347542256,0.0764113665,-0.1088597104,0.2404532433,0.0029706727,-0.1757607758,-0.438847512,0.0652569309,0.2699503601,-0.1519249827,-0.0684697106,0.0186510589,-0.243613705,0.0455427282,0.3752802908,-0.5222757459,-0.0179570038,0.485555619,0.0530991033,-0.2187103033,0.5302197933,0.1026279405,0.2320314348,-0.31973207,-0.3163997233,-0.3911034167,-0.0813904628,-0.1839075387,0.1252804846,-0.0581018291,0.1672071666,0.2231078297,0.1155931056,-0.1143737137,-0.4481284022,-0.1608466357,0.2227238268,-0.3137449026,0.0800440311,0.0210040323,0.4221844375,-0.1229951382,-0.3117901087,0.064292185,-0.1346099079,-0.1530557126,0.3762382269,0.0729845688,-0.0030936473,0.2004393339,0.4600739181,-0.4780099988,0.3159677088,0.0674662888,0.0163115002,0.0193634052,-0.1592834592,-0.2847013474,0.0015848173,0.1615588963,-0.0372262932,-0.2762799561,-0.0611413009,-0.0320599191,-0.1948691159,-0.0641590357,-0.2005607784,-0.1224491522,-0.4066183567,0.0587397292,0.0516069494,0.1041534469,0.3399559557,-0.1439994425,-0.2960096598,-0.3409697711,0.0431045666,0.3233141303,-0.4750151038,0.0190587305,-0.2495691478,0.1382085234,-0.0614196099,0.1678857803,-0.3513655066,0.1177775115,0.4220013618,-0.0616196394,-0.1031270549,0.356587857,0.2294960469,-0.0286254995,-0.0584889539,0.2589511573,-0.2766314745,-0.154982537,-0.607260406,-0.4248478413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1827","title":"Regarding On-the-fly Data Loading","comments":"Hi @lhoestq,\r\n\r\nI apologize for the late response. This answers my question. Thanks a lot.","body":"Hi,\r\n\r\nI was wondering if it is possible to load images\/texts as a batch during the training process, without loading the entire dataset on the RAM at any given point.\r\n\r\nThanks,\r\nGunjan","comment_length":15,"text":"Regarding On-the-fly Data Loading\nHi,\r\n\r\nI was wondering if it is possible to load images\/texts as a batch during the training process, without loading the entire dataset on the RAM at any given point.\r\n\r\nThanks,\r\nGunjan\nHi @lhoestq,\r\n\r\nI apologize for the late response. This answers my question. Thanks a lot.","embeddings":[-0.2567608654,-0.4141984582,-0.1661193222,0.2483003736,0.4269746244,0.1720282137,0.536202848,0.0338745154,0.1606053114,0.1289488822,-0.001788524,-0.055735305,-0.1721334457,0.2820934355,0.1901776493,0.1909988374,-0.0595938563,0.1587921083,-0.138773784,0.0033268146,0.0890625492,-0.1568349302,-0.0424369425,-0.349450767,-0.3389955461,-0.1996597946,-0.2822291255,-0.0711162686,-0.1060825735,-0.1634207964,0.0533678718,0.1527322531,0.3181540668,0.2086392045,-0.0001150392,-0.0170390587,0.218017891,-0.1820169836,-0.4191125631,0.1408158094,-0.1051086709,0.0830257833,0.1061532572,-0.2735402286,-0.0624837689,0.4026037157,0.2268292755,-0.2588222027,0.5197231174,-0.1203156561,0.1135745198,0.0155396154,-0.251711905,0.0338300802,0.0180318486,0.1570058167,0.0664375499,0.1371039152,0.4230663776,-0.0291004572,-0.10009709,0.2690835893,0.0597923473,-0.0717447773,0.2697406709,0.0292260442,0.1847450733,-0.4544056952,-0.0178142637,0.0481160544,0.5479182601,0.064766027,-0.1644562185,-0.2980218828,0.062438149,-0.2903074324,-0.0737055093,0.5474870205,-0.4233376384,0.0662197396,-0.5753109455,-0.4353668988,-0.340839535,-0.1214880422,-0.1248081774,-0.4485819936,0.0945488811,-0.0744498372,0.0184976254,0.2043346018,0.3853936493,-0.1420106888,-0.0838415176,0.4815199077,-0.2250701636,-0.2192800641,-0.6950021386,-0.2093052566,-0.001318793,-0.2905165553,0.2047459334,0.0725843534,-0.1342184395,0.0970776677,0.3903063536,-0.1428330243,0.1169907674,0.0733061358,-0.0802501664,-0.2386540771,-0.3664359152,-0.0372660533,-0.4085919857,0.1482342631,-0.0325562544,-0.2751605213,-0.5268071294,-0.1560787857,0.2876810133,-0.2287418842,0.0406805575,0.0393366776,-0.2840093076,0.2459859252,-0.017318666,-0.0802402049,0.4309794009,-0.2289165705,-0.1490017921,-0.235943079,-0.0138066383,0.0192500744,-0.2652452886,0.2100819796,0.147348702,0.1339302659,0.0781344622,-0.1186170056,-0.0349199772,0.0416334309,-0.0278266557,-0.3587900996,0.252528131,0.1472424716,-0.0298219696,-0.2553009987,-0.0916276723,0.4009249806,0.0713001937,0.5136198997,-0.2414181232,0.0211137626,0.4230254292,0.0808039904,0.1357252449,-0.2786237895,-0.414991945,0.3116093874,0.0774988234,0.1131320149,0.0474183559,0.0078667179,-0.2752754688,0.0506170765,0.0981041342,0.2681327462,-0.3788866699,-0.2997926474,0.42508021,-0.0693949386,-0.164846018,0.0966724679,-0.2613754869,0.3951227367,0.0030149855,-0.0836805552,0.0809053853,-0.2302006483,0.0437117144,0.3659825623,-0.071990855,-0.117163159,0.0314558893,0.5221555829,0.2550575435,0.1516492367,0.2851321995,0.7101941109,0.0966283232,0.0855020657,-0.1768975407,-0.2066118121,-0.3096502721,0.3663668931,0.1815696657,0.2520350516,0.1426855922,-0.0168799404,0.2277274728,-0.2867417932,0.0230926145,0.2027809024,0.0167472698,0.2598630786,-0.0693213716,0.3091482222,-0.3652183414,0.1415251195,-0.0983895212,0.0797333494,0.1544086784,-0.0158718154,0.35297364,-0.349332124,0.2450022846,0.2438263446,-0.0364973098,-0.255867213,-0.2441063225,-0.0441226065,-0.4127935469,0.718601346,-0.5417463183,0.1966092139,-0.0512823276,0.09646824,0.4296171367,-0.1172874272,-0.0213219374,-0.2439219058,-0.3871411681,-0.0267093945,0.0685139149,0.3544848859,0.0704997256,0.2509620786,0.0586093552,0.0112091862,0.2117924243,0.0162897408,0.2168889344,-0.3982014358,0.0664826334,-0.1508395523,-0.2844488025,0.5160201192,0.02444911,0.204485178,0.4325642288,-0.1383248717,0.0436304659,-0.2370224893,-0.0575171895,0.1321759969,0.0815988034,0.2126390785,0.3622999787,0.0901333466,-0.3557208478,-0.2448659837,-0.1161270514,-0.0551739372,-0.0683919489,0.1254815459,-0.0497627743,0.1124672815,-0.0003965897,-0.0780056641,0.211232394,0.1831405461,-0.061360687,0.1921744645,0.1410925537,-0.0235297307,0.0833840445,-0.024731189,0.4433245957,-0.0583296083,-0.0206397194,-0.0817430243,-0.373829782,0.3031649292,0.1477950662,-0.2642938495,-0.0870580226,0.0057576727,0.1134869605,-0.359593302,-0.1776631624,0.2359383404,-0.1588464826,0.3218360543,0.0702675804,0.0451420806,0.1434844285,0.0360315219,0.0052467887,0.9871277213,-0.0839214697,0.0043523284,-0.0539533906,-0.0803548992,0.011555844,0.040823698,0.2158866972,-0.0427481979,0.3710187376,0.1277460009,-0.2476490289,0.0663629323,0.1459338814,0.1700533777,0.2001652122,-0.0303136166,-0.4320502877,0.1189188883,-0.2146365047,0.067000173,-0.1402360201,-0.4020684958,-0.2220141888,-0.4311283231,0.3902618289,0.079478018,-0.0892353877,-0.0681830496,-0.5006967783,-0.0115268929,0.1544539183,0.2714295387,0.1364927888,0.4148814082,0.0841204375,-0.2388205379,-0.2071796805,-0.0080502378,-0.0256947931,-0.5578286052,-0.1077885628,-0.0739578903,-0.256526053,-0.0775331184,0.08641316,0.0621785335,0.7273237109,-0.8379276395,0.005221582,0.0414835513,-0.2470873892,-0.3170511723,0.055178076,-0.0348134004,-0.3148235977,-0.0968041569,0.0100210374,0.3065188229,0.2028958052,-0.0523615852,0.1385281533,0.1653434932,0.0870225579,-0.0035858597,0.691917181,-0.2893444896,-0.1965042502,0.1555946767,-0.0085192574,0.2469947338,-0.1404084563,0.1644654274,-0.0867963508,-0.1428429782,-0.3402161002,0.1624065489,0.0717259496,-0.3956108987,-0.4134685397,0.3103095591,-0.1441555321,0.0130169354,0.2465601861,-0.2894836068,0.3667176366,-0.0432610065,-0.0108503103,-0.152601555,-0.2064619809,0.0337397866,-0.0447560437,0.472345531,-0.2427278906,-0.1626865715,0.3872154057,-0.2076838762,0.0265423078,0.0005745254,0.2030395269,0.0167508256,-0.0487754941,0.322622031,0.0882009938,0.6163678169,0.0475718528,-0.0284592789,0.0319012329,-0.4291597605,-0.0964821354,-0.0170418303,-0.2263426632,-0.1728131771,0.0690125301,0.3415278792,0.0883052424,-0.1003848389,0.017444741,-0.0781887695,-0.3951706588,-0.3393760324,-0.0738433748,0.2017858177,-0.0846781507,0.0518524274,0.1381412745,-0.0109708505,-0.1735318899,-0.1649075747,-0.2571209073,0.3716025949,-0.0570074506,-0.2998161912,-0.0576978028,0.313998282,0.2946043015,0.1446267962,0.2483860701,0.155156523,0.1991737187,-0.2490459681,0.0963363647,-0.1571870595,-0.1717410386,0.3730605841,0.5805916786,0.3411245644,0.0477187261,0.0438671671,-0.0235839114,-0.4088102579,0.0236451142,0.0836625919,-0.1996649951,-0.1460694671,-0.5474240184,0.6335026026,0.1967407465,0.116274178,0.2649513781,-0.4954720438,-0.3273021579,0.2346536815,0.3142284453,0.8436256051,-0.4508902431,0.5444676876,-0.3048494756,0.2473471165,-0.0201042779,-0.2719485462,0.2400155663,-0.3923582137,-0.2029294819,-0.076067239,-0.2129281163,-0.184211567,0.387517184,0.1719744205,0.0551500805,0.078948319,0.0265491828,-0.0386298746,-0.0457458086,-0.0428438671,-0.3182277083,-0.168866232,0.1138143465,0.2894886732,0.0939642414,-0.2768952847,0.1020570472,0.0402831361,-0.0021514636,0.0921297073,0.0109625356,-0.2738729715,0.0460078754,-0.163930431,-0.1411316395,-0.2457915843,0.2322496921,0.1718426794,-0.102959007,0.0817219615,0.1289931387,0.0468949415,0.1746356338,-0.0462113135,-0.0158720743,0.4099961519,-0.0640880689,0.0698834062,0.1037017852,-0.183783114,-0.3859792948,-0.1115392148,0.0252950564,0.1762059778,-0.0275518335,-0.1019827351,0.1552456766,0.1611894965,0.1687710136,0.1073132977,0.1551029235,-0.1421439946,0.4910888672,0.1882764697,-0.1859416515,0.0786794648,0.5953602791,0.0951155648,-0.3628373146,0.2864352465,0.0443352684,-0.081372872,-0.1686702371,0.1470764577,0.2825474143,-0.1079276577,0.1888983101,0.1545469016,0.2254856527,-0.1124739125,0.2628932893,0.1220366955,-0.1079726294,-0.2072371542,-0.2010175586,-0.3627651334,0.2102878094,0.0184412114,0.3180756867,0.1455856711,-0.1054202393,0.2449546903,0.1054375619,-0.2790642679,0.2624481916,0.0356138237,0.3540089726,-0.0419607721,0.1211530641,-0.1476963907,-0.0645010099,0.0516973734,0.1453121454,0.0060292115,-0.2100868821,0.160660997,0.2023804635,0.0588510819,0.0495348498,-0.0611523353,-0.325971663,-0.1818641424,0.0119775143,0.1785271615,0.1633697748,-0.076006338,-0.1281002164,0.253924638,0.4820946157,-0.1333113015,0.4049239755,-0.1489795744,0.3809505105,0.1598469168,-0.1109177545,-0.0979172289,0.0815809444,0.0814426169,-0.0511123911,0.3641611338,-0.0127803246,0.0560659617,-0.3446500003,0.1657738537,-0.1912276894,0.2810378373,0.3597773612,-0.5640115738,0.036028903,0.3473659158,0.2125970274,0.0948763192,-0.101464957,-0.0700953975,-0.3996508718,-0.1379435509,0.2918556631,0.3479622602,-0.0500913151,-0.208357349,-0.0818204656,0.2415781617,-0.2856408954,-0.0388377756,0.280144304,0.0372007079,-0.0301976595,-0.0816213191,0.457891196,0.0365955792,0.1703118086,-0.0699832588,0.2537415326,-0.0408559591,0.0956816077,-0.2326490134,-0.2841991782,0.0441901572,0.629134655,0.0013403427,0.2709088922,-0.3439078033,0.1267871559,0.058765009,-0.357845962,-0.0759032518,0.0577209629,-0.1390878111,0.2013517767,-0.3088398278,-0.1083235517,-0.1483544856,0.4615742564,-0.0202561487,-0.1977924556,0.0634487495,0.3826987445,0.0262583364,0.1879312694,-0.1432705075,0.1812732816,0.1256388724,-0.2444056571,0.0555977076,-0.4015748501,0.2927613556,0.5355116725,0.419773221,-0.0755835697,0.3085467517,-0.020947963,0.0496039689,0.1997850835,0.0527762771,-0.0215652231,0.1705199331,0.3493461609,0.1046233177,0.1931972653,0.0697023049,-0.1269367337,0.3257361948,0.0266522076,-0.1698644608,-0.4003460407,0.068459332,0.331235826,-0.1443014592,-0.1411231905,0.0112106577,-0.3646990955,0.027409751,0.1893028021,-0.5241971612,0.0429250672,0.5569411516,0.0431280285,-0.2500905395,0.5597043037,0.2698698044,0.2045705914,-0.3305613697,-0.3172106147,-0.2677657604,0.0178782139,-0.0788245797,0.0593425743,-0.104605481,0.2013200223,0.2316366583,0.1113851443,-0.0454274379,-0.4500358999,-0.1181713045,0.2458626181,-0.4012693167,0.1063436642,0.1258490384,0.4947145581,-0.1750806421,-0.2096334994,0.1867687255,-0.0789291933,-0.1214432344,0.4983829558,0.0803528726,-0.0804291368,0.0769530609,0.4124682248,-0.5119664073,0.3900464773,0.043737229,-0.0051686806,0.0340045467,-0.247576043,-0.2745242417,-0.1018904522,0.2332584858,-0.0129439188,-0.1918456256,-0.0371498987,0.0398194455,-0.2484873533,-0.0652522743,-0.1068968549,-0.1223924309,-0.4048467875,0.00660004,-0.001201319,0.0769994557,0.3385459781,-0.11030031,-0.2600475252,-0.3939266205,0.0083855502,0.3129310608,-0.4489578307,-0.0429107063,-0.305523932,0.1661120653,-0.0805962384,0.1965610683,-0.4190232754,0.1713126153,0.3454068303,-0.1482984126,-0.061926771,0.4145238996,0.3104002476,-0.0135199837,-0.0714200884,0.152829811,-0.119272016,-0.0014661039,-0.71905303,-0.3286589384]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1825","title":"Datasets library not suitable for huge text datasets.","comments":"Hi ! Looks related to #861 \r\n\r\nYou are right: tokenizing a dataset using map takes a lot of space since it can store `input_ids` but also `token_type_ids`, `attention_mask` and `special_tokens_mask`. Moreover if your tokenization function returns python integers then by default they'll be stored as int64 which can take a lot of space. Padding can also increase the size of the tokenized dataset.\r\n\r\nTo make things more convenient, we recently added a \"lazy map\" feature that allows to tokenize each batch at training time as you mentioned. For example you'll be able to do\r\n```python\r\nfrom transformers import BertTokenizer\r\n\r\ntokenizer = BertTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\ndef encode(batch):\r\n return tokenizer(batch[\"text\"], padding=\"longest\", truncation=True, max_length=512, return_tensors=\"pt\")\r\n\r\ndataset.set_transform(encode)\r\nprint(dataset.format)\r\n# {'type': 'custom', 'format_kwargs': {'transform': <function __main__.encode(batch)>}, 'columns': ['idx', 'label', 'sentence1', 'sentence2'], 'output_all_columns': False}\r\nprint(dataset[:2])\r\n# {'input_ids': tensor([[ 101, 2572, 3217, ... 102]]), 'token_type_ids': tensor([[0, 0, 0, ... 0]]), 'attention_mask': tensor([[1, 1, 1, ... 1]])}\r\n\r\n```\r\nIn this example the `encode` transform is applied on-the-fly on the \"text\" column.\r\n\r\nThis feature will be available in the next release 2.0 which will happen in a few days.\r\nYou can already play with it by installing `datasets` from source if you want :)\r\n\r\nHope that helps !","body":"Hi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??","comment_length":197,"text":"Datasets library not suitable for huge text datasets.\nHi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??\nHi ! Looks related to #861 \r\n\r\nYou are right: tokenizing a dataset using map takes a lot of space since it can store `input_ids` but also `token_type_ids`, `attention_mask` and `special_tokens_mask`. Moreover if your tokenization function returns python integers then by default they'll be stored as int64 which can take a lot of space. Padding can also increase the size of the tokenized dataset.\r\n\r\nTo make things more convenient, we recently added a \"lazy map\" feature that allows to tokenize each batch at training time as you mentioned. For example you'll be able to do\r\n```python\r\nfrom transformers import BertTokenizer\r\n\r\ntokenizer = BertTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\ndef encode(batch):\r\n return tokenizer(batch[\"text\"], padding=\"longest\", truncation=True, max_length=512, return_tensors=\"pt\")\r\n\r\ndataset.set_transform(encode)\r\nprint(dataset.format)\r\n# {'type': 'custom', 'format_kwargs': {'transform': <function __main__.encode(batch)>}, 'columns': ['idx', 'label', 'sentence1', 'sentence2'], 'output_all_columns': False}\r\nprint(dataset[:2])\r\n# {'input_ids': tensor([[ 101, 2572, 3217, ... 102]]), 'token_type_ids': tensor([[0, 0, 0, ... 0]]), 'attention_mask': tensor([[1, 1, 1, ... 1]])}\r\n\r\n```\r\nIn this example the `encode` transform is applied on-the-fly on the \"text\" column.\r\n\r\nThis feature will be available in the next release 2.0 which will happen in a few days.\r\nYou can already play with it by installing `datasets` from source if you want :)\r\n\r\nHope that helps !","embeddings":[-0.2644477785,0.0704934821,0.0091305831,0.1295920014,0.3082516789,0.0080511766,0.2901795208,0.3717779517,-0.2807389498,-0.0181723814,-0.048823636,0.1473779678,-0.2360719144,0.0108855832,0.0396098942,0.0288687628,0.2195769548,0.2025509775,0.1469030827,0.0644162074,-0.0421971455,-0.2423410267,-0.2289886475,-0.0132892057,-0.4302169383,-0.2289030105,-0.130100742,-0.2940586507,-0.3577356339,-0.4037176669,-0.0329179354,0.1460277736,0.4799691737,0.4330447912,-0.0001187578,-0.1787714958,0.1574260443,0.0388377905,-0.091521956,0.0657222122,0.2080893964,-0.3404701948,-0.1510676444,-0.2540647089,0.1679103673,-0.1874020994,0.0126407435,-0.1223173738,0.2558726966,-0.022112092,0.1069587767,0.1478157789,-0.1175837219,0.1175930575,0.2055493891,-0.0257249586,-0.067175068,-0.0461689681,0.3811223805,-0.0764790028,-0.187481612,0.3004252911,-0.1841914803,-0.0576366261,0.3920024931,-0.0049807709,-0.1302286685,-0.3428998888,0.2336874157,0.2656778097,0.3912026584,-0.3416002393,-0.2912170589,-0.4751839638,-0.1442976296,-0.1671044528,-0.1160393953,0.1333686113,0.1011420563,0.1327646375,-0.5343756676,-0.3300271034,-0.0942187086,0.2805031836,-0.1244559139,0.2533674836,-0.0309925396,0.0617218614,0.0674949214,-0.3011275232,0.1197618097,-0.3782680929,0.1251965016,0.4119563103,-0.2103306204,-0.1967040151,-0.162057966,-0.131968081,0.4441077709,-0.2146405429,-0.3352544904,0.0751733482,-0.2477160096,-0.0298868325,0.2738423347,0.2010719776,-0.1895847917,0.3804612458,0.1412707418,-0.2511979938,-0.0760981366,-0.1168308482,-0.5041647553,0.0023232726,0.242370829,-0.3414865732,-0.2537666857,-0.0109220063,0.0321966819,0.0937678665,-0.3200736344,0.2049039751,0.1570596844,0.2677529752,-0.1823857278,0.1182553396,-0.2504811287,0.1530012339,-0.1295621693,0.0136844954,-0.0348925665,-0.0809213817,-0.3118556142,0.0885422602,0.1323872507,0.0369310565,0.1483543813,-0.0311594773,0.0830833986,0.1199528426,0.2286500633,-0.3431074619,0.0696898922,0.0881681591,-0.0551424436,0.3614559472,-0.2426368594,0.0543855242,-0.2421161383,0.3458733559,-0.3575131595,-0.3633337617,0.1150057763,0.0956483856,-0.2124391645,-0.0192326438,-0.4771781266,0.3124535382,0.4131705165,0.0353512764,0.0308051649,-0.2233496308,-0.2891534865,-0.1119893193,0.2520446479,0.2172581702,-0.3746162057,-0.0462599099,-0.0060771499,0.2774538696,0.3547437489,0.7049410939,-0.2344234437,0.4682484269,-0.0217173658,0.5970696807,0.1308767945,-0.0120305633,-0.5504743457,0.306410104,0.1361112893,-0.0626283363,-0.0474245176,0.0487032607,0.5649160147,-0.1264642328,0.2421895266,0.3477282822,-0.1075709909,0.2397238165,-0.2524233758,-0.3894592524,0.2543738484,0.2126838565,0.2313421071,-0.4073577225,-0.1076531634,0.5631249547,0.1074653864,-0.2344982773,0.3512014151,0.0694597587,0.1932828426,0.1367788911,-0.0244933888,-0.1917746514,-0.1133494154,0.138760373,0.1284853667,-0.0340240113,-0.0177696273,-0.3649345636,0.1988357306,-0.0558937266,-0.0451731496,0.0023124169,-0.0087510115,-0.2987191975,-0.0523148142,0.0450563654,-0.1235791892,-0.2520162165,-0.1439156234,0.1190637127,-0.0510434993,0.0653499141,0.0642399043,-0.0879655033,-0.0818806812,0.2881655991,-0.1019601226,-0.1098917425,-0.0430198386,0.1155033335,0.1792636067,-0.1729746014,-0.1989774257,0.6291497946,0.3313725889,-0.0527769439,0.2620900869,0.0579496026,0.0874846652,-0.1010522097,-0.0608754344,0.1980132908,0.1038440242,0.1809868366,0.037133649,-0.2367585897,0.184265852,-0.0231454857,0.0577142648,-0.0625004098,0.0631765053,0.1399936825,0.4071207047,0.3443516493,-0.0991740823,-0.1226790249,0.3456887901,-0.2547423244,0.0944562778,0.3692593873,-0.4270763099,-0.370239526,0.0560605228,-0.2841128409,0.073677212,0.1215593964,0.0545074455,-0.0743528157,0.3321838975,0.004992832,0.2567667663,0.1768510193,0.0905410051,0.0599088483,0.1524800658,0.2476588041,-0.0831809565,-0.1484982371,0.2264687419,0.2812068462,-0.0641535372,0.0165543146,-0.1782043725,-0.147877723,0.0071278196,-0.0921132267,-0.1096905544,0.0529508367,0.1498169303,-0.0503600165,-0.0050369781,0.2096641213,0.0560787916,0.3908675611,0.0888429284,-0.2763009369,0.0869960859,-0.0899240077,-0.3656361103,-0.0193913653,0.2563130558,-0.3312307596,0.2955761552,0.1239056662,0.1111488491,-0.3529420793,-0.5111264586,0.055365812,-0.0807594806,0.0430213548,0.0051294304,0.0119368415,-0.3470669687,0.0450911373,-0.0829578936,-0.2204847634,-0.048679471,0.0377844833,-0.1488906741,-0.1909849197,0.099119328,-0.1178259403,-0.2849353254,-0.1407158971,0.1821797192,-0.1708399057,0.144347325,0.0215969048,0.058514379,0.0398076661,-0.0908531174,-0.216559723,0.0396113619,-0.3456558883,0.4540537,0.053776063,-0.1599427313,-0.0739158913,-0.0598627143,-0.0437619947,0.3011980355,-0.4520861208,0.2026535124,-0.2943244874,0.2322741151,-0.0625725538,0.1154094338,0.062830694,0.0229553003,0.0532675311,0.0854735672,-0.1402453333,-0.1986550242,-0.0442462415,0.1966251731,0.0458277091,0.1926076561,0.3319135308,0.8323061466,0.064997755,-0.1224136129,0.2346021682,0.1743752807,0.1243966073,-0.3345831931,-0.1610758156,0.1468512416,-0.0445961915,-0.1165466756,0.537445426,0.1378043741,-0.152831912,0.350389272,-0.3317063749,-0.0801909044,-0.1775426418,0.364733696,-0.2959564924,0.2762163579,0.0002361259,-0.0691211969,-0.2257589996,-0.3720051646,0.0028131222,-0.010876297,0.0650557578,0.1188461334,-0.335893333,0.1355585903,-0.5341358781,0.1895345002,0.0412229262,0.2819493413,0.1803769469,0.1240452379,0.1511198282,0.1265985966,0.5065208077,-0.0254912749,-0.1990229785,-0.1221660301,-0.2344269603,-0.0770513713,0.0300621986,-0.2233587652,0.1062349379,0.4797866642,0.6534510255,-0.3339403868,-0.2007192075,-0.3113008142,0.4322643578,-0.0292310752,-0.1336647719,0.0284199603,-0.0316572934,-0.567843914,0.223359108,0.1654522419,0.3397708237,0.2408814877,-0.1090207398,-0.202644065,-0.0134126404,0.2838229835,0.0837169737,0.1670539379,0.0003052847,0.2988259196,-0.0417262912,0.0744474456,0.431532234,-0.0293054413,-0.2243986428,-0.4143536389,-0.0252462681,0.0462160669,0.5054761767,0.0984765813,-0.0385241285,0.2553429306,-0.1259785444,0.2532570064,-0.2888217866,0.5742155313,0.2188751549,0.0965442434,-0.7376465201,-0.5411154628,0.0903887302,0.2293254435,-0.09818618,0.1740108728,-0.1532960385,-0.3754510283,0.5328889489,0.2531071901,0.906961441,-0.3994151652,0.4245498776,-0.2723042369,0.4491722882,0.5515657067,-0.5493831635,0.0299037974,-0.3276279271,0.097127445,0.1051325575,0.028546067,-0.0868091583,0.276642859,-0.0814039111,0.247018531,0.0972772539,0.3555195332,-0.1267232448,0.4976910949,0.1301132739,-0.5027886033,-0.1041884273,0.0738881007,-0.0947316065,-0.0715888366,-0.1766844243,0.0566981211,0.073020719,-0.0675735399,-0.2201564312,-0.1896804273,-0.1857761592,0.2202737182,-0.0419980139,0.0303264484,0.461209923,-0.0011070019,0.2244462669,0.5506526828,-0.1509454995,0.1174335033,-0.1730620116,-0.0176459458,0.0873551145,-0.242223233,0.1032318771,0.0627399832,0.1997383088,-0.1990748793,0.0809132233,-0.0542820171,-0.3179367483,-0.1111050248,0.6560403109,-0.3565067351,-0.1661115587,-0.1440653503,-0.0968542621,0.0061933077,0.0565191545,0.1002257019,-0.045260258,0.3752139807,0.034715578,-0.2648664117,-0.0726717785,0.0646326318,0.0507312156,-0.0524653718,0.5686107874,0.0318167768,-0.2139265388,-0.1469620466,0.385161221,0.2808520496,-0.1480039209,0.1107220724,0.1490308195,0.0435368828,-0.1176785231,0.1838830262,0.1670864373,-0.3886724412,-0.0599257499,-0.1476562917,-0.4500799179,0.117471166,-0.0276115574,0.2204982936,0.1344050616,0.2206279188,-0.1693745703,0.052947849,-0.2138481587,0.3104968667,0.007994147,0.3019279242,-0.3741823733,0.0043555931,-0.1641305983,-0.0982664675,-0.0976643562,0.2237660289,0.0310724024,-0.1915962547,-0.2536722124,0.1974970996,0.2984792292,-0.3040182292,0.0982727781,-0.2110511214,-0.0752807036,-0.4249739945,0.2353437543,0.1522119492,0.0180687122,-0.1894283593,0.2478582561,-0.1959649473,-0.2526566386,0.1285891235,0.0337722823,0.1977977008,0.0746600404,0.23859182,0.3486384749,-0.148400113,-0.0132973278,0.179926455,0.6943491101,-0.0666868985,0.0668378398,-0.0838264376,0.1475378722,0.3351517618,0.0089718541,0.3940816522,-0.2136284262,-0.3860997558,0.6833882332,0.1724226177,-0.0602168329,-0.1880691797,-0.0288697388,-0.1790680736,-0.1656182557,0.1121012568,0.1801537871,0.2911601067,-0.4418913424,-0.0985675156,0.3846249878,0.1857512593,0.0258366913,0.4675312936,0.0653676167,-0.0693073496,0.3300024569,0.3051339686,0.2250804603,0.4662766457,-0.0665648729,0.4718081653,0.0667065084,0.1253409386,-0.1344000399,-0.3910471499,-0.2368483245,0.425621897,0.1948470473,0.0373422056,-0.2140035033,0.2732304633,0.257464081,-0.2321343869,-0.0505207516,0.3097582459,-0.2751244903,-0.1667289883,-0.0587160625,-0.0501799099,0.0308765266,0.2555799186,-0.2814871669,-0.281313628,0.101858072,0.0614651553,0.0013059773,-0.2767422199,0.2092770636,0.0969454125,0.0506159626,-0.2989830375,0.2383568883,-0.3614274859,0.0396073237,-0.2037047148,0.2652468681,0.3167506456,0.4223527014,-0.1601740271,0.0402517878,-0.1901478916,0.1461488307,-0.1259520352,0.3894744217,0.0751186833,0.1377942413,0.1000468135,-0.0678023398,-0.0697038695,-0.1074380949,0.0438324772,-0.3280678689,-0.0221281666,0.0759527236,0.238201946,-0.2457774878,-0.0323611088,0.0971806869,-0.3176351786,0.1855528951,0.8031607866,-0.3024765253,0.1586753428,-0.1604958028,0.0480638593,-0.161602512,0.446498692,0.4236746132,-0.0030971593,-0.5891641378,-0.0724819079,-0.3910683393,0.1330499947,-0.4093894362,0.2987821102,-0.0047313226,0.0367052443,-0.1346133798,0.0945018157,-0.267228514,0.1531141847,-0.4629123807,0.0851813257,-0.0476909466,0.2911083996,-0.1900393367,0.2047682256,-0.0044686114,-0.370197922,0.3627031147,-0.1943225414,-0.1329682022,-0.1911631376,0.2968972623,0.178827405,-0.1411943585,0.4452565014,0.1750631928,0.1358044297,0.060782969,0.2444523871,-0.007435956,0.0730454549,-0.2477153093,0.0883704722,-0.0207429025,-0.0447110087,-0.2221280187,0.1344269067,-0.1506128907,-0.1317330152,-0.0069137975,-0.1570843905,-0.2271625251,-0.0132330116,-0.1807029098,0.1478371769,-0.0815054923,0.4159241021,-0.1348463595,-0.0150566185,-0.2692967057,-0.2827503681,0.2788333595,-0.6746147275,-0.4008940756,-0.148191154,-0.0568874367,0.5397025943,-0.1090696007,-0.6030365229,0.1040227488,0.4999953806,0.1065346077,-0.4595339894,0.1034715846,-0.0113228904,0.0178211574,-0.195600912,0.5466092229,-0.1942952573,-0.2376628965,0.0750037283,-0.2591937184]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1825","title":"Datasets library not suitable for huge text datasets.","comments":"How recently was `set_transform` added? I am actually trying to implement it and getting an error:\r\n\r\n`AttributeError: 'Dataset' object has no attribute 'set_transform'\r\n`\r\n\r\nI'm on v.1.2.1.\r\n\r\nEDIT: Oh, wait I see now it's in the v.2.0. Whoops! This should be really useful.","body":"Hi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??","comment_length":43,"text":"Datasets library not suitable for huge text datasets.\nHi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??\nHow recently was `set_transform` added? I am actually trying to implement it and getting an error:\r\n\r\n`AttributeError: 'Dataset' object has no attribute 'set_transform'\r\n`\r\n\r\nI'm on v.1.2.1.\r\n\r\nEDIT: Oh, wait I see now it's in the v.2.0. Whoops! This should be really useful.","embeddings":[-0.4678060412,0.1656083465,0.0304235257,0.1393032819,0.3698442876,-0.0799572915,0.1343442947,0.4200882018,-0.4338594377,-0.1087068021,0.1051200852,0.0978749618,-0.2596930265,0.1617353857,0.1377583444,0.0071735461,0.2880678475,0.1773830205,-0.0256084055,0.1219855323,-0.068343088,-0.1277218163,-0.2867692113,-0.0003999639,-0.5252425671,-0.2634716034,-0.0174378678,-0.2435948402,-0.4448638558,-0.3248973191,-0.1924744993,0.1364666373,0.4324246347,0.5125051141,-0.0001255967,-0.2210252136,0.2472047806,-0.0384248644,-0.2875906825,-0.0394756272,0.293053627,-0.3577700257,-0.1968452185,-0.2132425606,0.1938291043,-0.3745193481,-0.0414699279,-0.1099502668,0.1161912084,0.2161804438,0.0858882964,0.0541299358,0.0701664165,0.0395160988,0.1645486206,-0.0014555308,-0.1042268798,-0.0228799507,0.4538724422,0.0654921979,-0.1268360168,0.1399395615,-0.12165346,-0.0212641973,0.4577085376,-0.0739728734,-0.2455075085,-0.2150257528,0.1715577692,0.2889243364,0.5197036862,-0.4864006937,-0.3977509141,-0.5967864394,-0.1693623364,-0.2333569825,-0.0051260851,0.1050855666,0.064477928,0.1954610646,-0.5726397634,-0.4743477106,-0.1159548759,0.3932580948,-0.2551800013,0.1143131405,-0.1285423338,0.0937680304,0.1853441149,-0.1451212466,0.1337652802,-0.2565786839,0.0210567191,0.1875143945,-0.3344244659,-0.1271075159,-0.1622018218,-0.021164,0.4064775705,0.0946194306,-0.4326750338,0.0650007576,-0.1359995008,-0.0083531858,0.2543174624,0.4710545838,-0.2227854133,0.2742271125,0.2237857282,-0.2196394056,0.0120657673,-0.080084011,-0.4790991247,-0.1735170484,0.0792802125,-0.358417809,-0.120091565,-0.1453280896,-0.0396949798,-0.0014614257,-0.1404424459,0.1979501396,0.1161104217,0.1320023835,-0.0579987131,0.1802783459,-0.1072323769,0.2354936749,-0.1325275749,-0.2000354528,0.002107711,-0.020393068,-0.2910279334,0.1360286772,0.1617564857,-0.0009595223,0.0848516077,-0.1097623631,0.0947208032,0.1195635498,0.1449590027,-0.323037684,0.0003708185,0.1692297161,-0.1279814839,0.4077442586,-0.1729893684,-0.0469149798,-0.2752201855,0.3672422767,-0.2981247008,-0.3904721141,-0.0615494438,0.0727412924,-0.4090559185,-0.048892457,-0.5169223547,0.4131361842,0.4666951895,0.0010689226,0.0228556469,-0.2046528012,-0.1611898541,-0.0093320664,0.170317173,0.1939947903,-0.6206810474,-0.0736391172,-0.0242317617,0.1578421146,0.2399742454,0.6512376666,-0.1410071254,0.340269655,-0.0592872202,0.3595706522,0.2331751436,0.1929224283,-0.5321199298,0.2934281528,-0.0334726572,-0.2285967767,-0.0180567205,0.0428134017,0.3913392425,-0.1193694323,0.0282697044,0.4206727147,-0.0616831332,0.1134212613,-0.2282675803,-0.3263324499,0.1932424307,0.2130988836,0.2006414533,-0.1420846581,0.0438174717,0.4546659291,0.0379386954,-0.1514840275,0.4403600395,0.091993764,0.217300117,0.1510068029,0.0594997033,-0.2520761192,0.176645726,0.0390712395,0.1651211083,-0.0873571485,-0.0359869972,-0.2174773961,0.0089027276,-0.0221343096,0.0318895541,0.0726806074,-0.1267323494,-0.15514189,-0.0079077687,0.1740851104,-0.1608186811,-0.1197563335,-0.23613213,0.072443381,-0.1739453971,0.0709985644,0.0937701315,-0.1703380048,0.0002909967,0.1770595163,-0.1185424551,-0.1547638178,-0.0633910447,0.2232189924,-0.0343861319,-0.1681163013,-0.3214416504,0.6541765928,0.2707799375,-0.062552765,0.2410966605,-0.0750871524,0.0416625328,-0.1014060453,-0.0262598339,0.1974630952,0.0318035707,0.1712822914,0.0407972075,-0.21162799,0.0912056342,-0.0658040643,0.1283020824,-0.0092189023,0.1961651146,0.0282079093,0.3526709378,0.2581016421,-0.0891507044,-0.1441114247,0.5428144932,-0.2562134862,0.257979244,0.4638553858,-0.4305692017,-0.3219676912,0.0750782192,-0.1099267378,0.2708714306,0.1814125776,0.0384941138,-0.005050791,0.2958002687,-0.0585589409,0.226982072,0.129266277,0.2312542796,0.1655011028,0.1857072562,0.1758005172,-0.1652547121,-0.1809598655,0.2347026318,0.3232022822,-0.1956056803,0.0806342736,-0.2088699192,-0.1941362321,-0.0987124592,0.1454793811,-0.1387936622,0.0052878088,0.0636099949,-0.0917775184,0.1439724714,0.1453877687,0.1162511706,0.3910382688,0.0875247419,-0.3043805361,0.1152973697,-0.0589727871,-0.4323800802,-0.0351220965,0.2794806659,-0.4264618456,0.2161747068,0.1520200521,0.0763974115,-0.28428635,-0.5105540156,0.0153107001,0.0186661482,0.2145485431,-0.0979565829,0.020444002,-0.1042166203,0.1304889321,-0.0322462022,-0.1589455456,-0.1028219536,0.0985742286,-0.1815346777,-0.2435962707,0.1276798546,-0.1638975441,-0.2868719399,-0.2101882696,0.2809813917,-0.1480194032,0.0006387518,0.1361135989,0.1532779485,-0.0221642666,0.0062422557,-0.0760251433,0.1199329719,-0.2126324326,0.5063880682,0.155168891,-0.2401786298,-0.1168237105,0.0823830068,-0.1706950217,0.4046607614,-0.4827156067,0.2901175618,-0.3174150288,0.1822003871,-0.0747014508,0.1196025982,0.1363299936,-0.0662222058,0.091195114,0.0743866339,-0.2566071153,-0.2644274831,-0.0348302983,0.319036305,-0.0538880974,0.3580652475,0.5046021342,0.5707176924,-0.128400594,-0.2025967389,0.302220434,0.0581376217,0.1859385669,-0.3812960088,-0.2078484744,0.3076785803,-0.1130783707,-0.1500231773,0.4042946994,0.0733633414,-0.1256055385,0.1822137833,-0.2858064175,-0.0770426989,0.0022837936,0.1816804856,-0.2554910481,0.194970265,-0.0332255363,-0.147334069,-0.0552725829,-0.3543818593,0.0065026255,0.0854855403,-0.0398811661,0.2780982256,-0.5231354833,0.144545719,-0.3172294199,0.1839579791,-0.1198381186,0.3176474571,0.0893478021,0.0443348512,0.172850877,-0.0029821319,0.6573579311,0.1614986807,-0.2243918031,-0.0802065656,-0.184988752,-0.2815262377,0.0307366047,-0.1564119756,0.2526804805,0.3969074786,0.4458347857,-0.2846605182,-0.0874878466,-0.1563075781,0.599565804,-0.0287428573,-0.2090794146,0.0097266203,-0.193761006,-0.6171143651,0.141123727,0.1184751689,0.3723028302,0.0513239987,-0.0176972523,-0.1607232988,-0.148767978,0.3314806521,-0.0212381352,0.1076566428,-0.0759269893,0.2275631726,-0.1148314476,0.1661521643,0.5096878409,-0.0112086767,0.1589721888,-0.2057615966,0.0160308219,0.0860805139,0.4228656292,-0.0182476137,-0.1591086835,0.2727679014,-0.1593001485,0.3351101577,-0.2517122626,0.4993906319,0.1424314976,0.0247756094,-0.7850490808,-0.4419436157,0.3065380156,0.1834492683,-0.0020832019,0.2332212925,-0.0898343325,-0.3548890054,0.2965226173,0.1264454871,0.9342719316,-0.3664009571,0.2955781817,-0.3144507408,0.3217417598,0.5618305206,-0.6394956708,-0.0108736726,-0.3764114976,0.0519127883,0.0855140537,-0.054223191,0.1334859133,0.4011874199,-0.1039246544,0.2485841215,0.0921480656,0.0529664345,-0.0280639902,0.6024857759,-0.1124617681,-0.354742825,-0.1395692974,0.0302560478,-0.1978614628,-0.0176334456,-0.1501988024,0.1268520057,-0.048003003,0.0881531015,-0.3702218235,-0.0857852548,-0.0762250498,0.1309331954,-0.0710070431,-0.1586057842,0.5522637963,-0.0243855808,0.2939382195,0.490629524,-0.159932822,0.1951515228,-0.3952571154,0.0072428365,0.2270343751,-0.2535532415,0.1769521236,-0.0029716771,0.046727825,-0.227456972,0.1215277463,-0.0144595047,-0.186314851,-0.1406602412,0.5457144976,-0.2815807462,-0.1525867283,-0.1081250086,0.0739671364,-0.0937485844,0.0124913845,0.1646723747,-0.0203944277,0.3688541949,-0.0650466457,-0.2675851583,-0.1635231376,0.2240000516,0.0383487567,-0.0482339151,0.5090438128,0.1999021769,-0.1995310783,-0.0959967002,0.3914939463,0.2192799002,-0.2844052017,0.2249106169,0.1075528637,-0.0183575917,-0.1117914915,0.0926157087,0.1175076887,-0.443125397,-0.0679436848,-0.2272987664,-0.3753504753,0.1325864047,-0.0316812247,0.1135018021,0.0622485988,0.1203564778,0.0003968311,0.1120575964,-0.1463401765,0.1943098456,0.0917973444,0.2783542275,-0.156272009,0.047858309,-0.2649526298,-0.090774484,-0.1531646103,0.1001147702,0.1419309974,-0.1376613528,-0.206895858,0.2038150728,0.2516921759,-0.2625119686,0.148856014,-0.294152379,-0.1333899349,-0.3269467354,0.2188902795,0.2685706913,-0.0101765199,-0.0553013198,0.2593246102,-0.3359715641,-0.2528076172,0.0586050637,-0.0326535553,0.1197470948,0.1532147378,0.3857433796,0.5578308702,-0.0974076912,-0.3966265619,0.2364943773,0.4823636413,-0.1004022956,0.1028597057,-0.2452235967,0.1180023402,0.3210249841,0.1504739672,0.484167248,-0.1821766496,-0.3888131082,0.5509516001,0.0933211446,-0.1055201888,-0.1948432773,0.0620901063,-0.1472336799,-0.2354937494,0.1320221275,0.2544418275,0.1740469337,-0.4353885353,-0.0554228388,0.5197481513,0.2001790255,0.0402675532,0.4844231308,0.1650195271,-0.0563524291,0.3396508098,0.2591528594,0.2522543669,0.5600369573,-0.015649803,0.5746831894,0.2159358114,0.229124099,-0.1778336316,-0.4864786565,-0.1272131205,0.3517716825,0.2182939947,-0.0824787021,-0.2162065208,0.2877639234,0.3367443681,-0.1225604266,-0.0357289538,0.1663766801,-0.191600129,-0.121651195,-0.1463640183,-0.166008696,-0.0130390972,0.2312279046,-0.2712216377,-0.3700610101,0.0600817315,0.0320276245,-0.105799228,-0.4628716409,0.4335577488,0.0649348497,0.0241066553,-0.2738912106,0.2925997972,-0.2164276242,0.0736267194,0.0415441617,0.2999282181,0.4472155571,0.4638708234,-0.0596711598,0.0109046949,-0.2410855293,0.0245371293,-0.1439651996,0.4624995887,0.0349773653,0.2712832689,0.1345170736,-0.0633618906,-0.032446675,0.0050668335,0.1022722125,-0.3170353174,-0.1514536887,0.0370284058,0.1535603851,-0.3451302946,-0.1238740757,0.1821555942,-0.2598915398,0.0171437059,0.9408844709,-0.3457409739,0.0535711236,-0.1901853681,0.0102618393,-0.1116014421,0.4436601996,0.4570569694,0.0764182582,-0.3387860358,-0.0778188184,-0.3848974705,0.202005446,-0.2915293574,0.2598511875,0.0068153339,0.0922793373,-0.0870103985,0.0521694124,-0.2772401869,0.331789434,-0.3469387889,0.071258612,-0.076610975,0.0129409842,-0.1179920658,0.2422270179,-0.118453376,-0.365526557,0.3441645205,-0.1634904444,-0.1625525504,-0.2217511833,0.2344190925,0.1863702238,-0.124104917,0.5211454034,0.163992852,0.2844225168,0.0696416348,0.3814973831,-0.0657802224,0.029627176,-0.3679363132,0.2523971498,-0.0413009822,0.020654073,-0.2122358382,0.1473592669,-0.0459803045,0.0547392443,-0.0601740703,-0.0250843037,-0.2111238986,-0.0575815961,-0.1792091727,0.142264232,0.0848786682,0.2873156071,-0.056095615,0.0088380892,-0.3809773326,-0.2963832915,0.2829847038,-0.4146686792,-0.3906490505,-0.1051533893,-0.0266615264,0.534350276,-0.2405608445,-0.5345792174,0.0504025407,0.5844872594,0.0505463742,-0.3297665417,0.3089816272,-0.232164681,0.0101887556,-0.144369483,0.5861437917,-0.1608718336,-0.2709832788,0.0085744103,-0.1008406058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1825","title":"Datasets library not suitable for huge text datasets.","comments":"Yes indeed it was added a few days ago. The code is available on master\r\nWe'll do a release next week :)\r\n\r\nFeel free to install `datasets` from source to try it out though, I would love to have some feedbacks","body":"Hi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??","comment_length":41,"text":"Datasets library not suitable for huge text datasets.\nHi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??\nYes indeed it was added a few days ago. The code is available on master\r\nWe'll do a release next week :)\r\n\r\nFeel free to install `datasets` from source to try it out though, I would love to have some feedbacks","embeddings":[-0.4051040113,0.1390962005,0.0155028645,0.2904538512,0.2991148829,0.0074277432,0.0747686848,0.3566738963,-0.3289943933,-0.0648281872,0.0511006378,0.0604724623,-0.2490902841,0.1777262539,0.0443806089,0.0889961049,0.1888465285,0.1591734588,0.1126290709,0.1340800226,0.0361788236,-0.1229671836,-0.2318736166,0.0238546301,-0.5338650942,-0.2475266159,-0.0014842657,-0.2042943835,-0.4535925984,-0.316849649,-0.1865609139,0.1878174841,0.5089609027,0.4989701509,-0.0001241754,-0.2766101062,0.2280961722,-0.001814332,-0.2488396019,0.0375921987,0.4174685776,-0.2963128388,-0.2088589221,-0.1917603016,0.1633433104,-0.2845399678,-0.024078181,-0.0812101662,0.1066380516,0.0794834346,0.0776780844,0.0891764089,0.0327625982,0.1611440033,0.1430723965,0.0420889556,-0.0692556351,0.0685506612,0.4472152889,0.0410758667,-0.1649745107,0.1808107644,-0.1345980018,-0.003542365,0.3252602816,-0.0841970816,-0.1148669943,-0.1845379919,0.224188894,0.2748838365,0.5318579078,-0.4414384365,-0.2680335045,-0.5356898904,-0.1947643161,-0.299777329,-0.0366853178,0.1811596304,-0.0221885163,0.204956919,-0.4301466644,-0.4338746369,-0.1384976655,0.3633370101,-0.2156594247,0.1345005333,-0.1368915439,0.1114961281,0.1807055622,-0.2048989534,0.1003355011,-0.3374049067,0.0127526671,0.1920605749,-0.3882723451,-0.1360710561,-0.1989927888,-0.0134857958,0.4718787074,-0.0272693075,-0.4255781174,0.1566315889,-0.1229032055,0.0211920235,0.3621504605,0.2150782645,-0.1023104712,0.2536747158,0.1884473413,-0.1695374101,-0.0573535673,-0.073120825,-0.5236894488,-0.0749365762,0.108167015,-0.4771719277,-0.2528762817,-0.1971201003,-0.0683676526,0.0907616094,-0.1943468899,0.262750566,0.1845922023,0.1667827368,-0.0995505899,0.0795410201,-0.2099237293,0.1964030266,-0.1286838502,-0.1992129385,0.0066114264,-0.0878102556,-0.2628050148,0.1638805717,0.1636521667,-0.0238197427,0.1044936627,-0.0577499457,0.0845344812,-0.0026520302,0.1463346183,-0.3493956625,-0.0723238215,0.0937067717,-0.0408833437,0.3709303737,-0.2378408462,0.0360968411,-0.2047561109,0.3789343834,-0.2517656386,-0.36022681,-0.0587221161,0.072268188,-0.3575673699,0.0288752932,-0.5238496065,0.3691043258,0.4520553946,0.0841351524,0.0100024715,-0.107897155,-0.2290685177,-0.015236103,0.1661327779,0.1557083428,-0.6027988195,0.0375130214,-0.0748698413,0.1012729108,0.3303999007,0.7473345399,-0.1570432633,0.3759785295,-0.102838181,0.4475915432,0.1434928924,0.1725686342,-0.5138141513,0.4206911325,0.0045104902,-0.1445603967,-0.0063966457,0.0962307379,0.3777561784,-0.1191317439,0.1935288608,0.331550777,-0.0646060333,0.0937095582,-0.3203207552,-0.3219234645,0.1349286735,0.3081050217,0.1901600063,-0.2884480953,-0.0160415825,0.3337594867,0.150905028,-0.1702376455,0.3706077635,0.1314394325,0.2265491635,0.163771987,-0.008740874,-0.2258760333,0.1498987824,0.0636247545,0.1120895147,-0.1439846605,-0.0422225483,-0.2829776704,0.1273653656,-0.0255168118,0.0055882861,0.1112153679,-0.0991995335,-0.2185772508,-0.0950815231,0.1362289786,-0.0690708905,-0.1132749319,-0.3314360678,0.1292933226,-0.1011646986,0.0312824994,0.0371802822,-0.1410601437,0.0365892053,0.2281328738,-0.1317255199,-0.1346917003,0.0014941593,0.1887519509,0.0156657305,-0.2401512712,-0.2595067024,0.5918124318,0.2400096953,0.0207171254,0.2485296875,-0.0443755053,0.0471287817,-0.1107606813,0.0558445752,0.2861930132,0.0867251381,0.1239675432,0.0046203751,-0.2903572917,0.0834710971,-0.0330149233,0.1967634112,0.0264623966,0.2717000246,0.068381995,0.3077633977,0.3182150126,-0.1597676426,-0.1507178694,0.3479215205,-0.3244886994,0.1942491829,0.4423333108,-0.4264603555,-0.3297384679,0.0858769789,-0.2397742122,0.1328829825,0.1559302658,0.0524388589,-0.0497584604,0.366088599,-0.0712773949,0.1652619392,0.1384433508,0.1298765242,0.2031743526,0.1204415187,0.1302973181,-0.1725421101,-0.0647165999,0.1956180632,0.3540200889,-0.1527119279,0.0529190451,-0.2162585109,-0.307713896,-0.0110397991,0.1897613406,-0.1802602559,0.0459261052,0.1223604605,0.0164135657,0.1256352812,0.055649057,0.1618564427,0.3633149266,0.0336523354,-0.3247652054,0.1699574143,-0.0947100669,-0.4917177558,-0.0354666486,0.3435530066,-0.3346315324,0.2998607755,0.2190070003,0.0502540059,-0.3155395985,-0.4621923566,0.017577108,0.1138878241,0.2294166088,-0.1505284011,0.0415531918,-0.1999851763,0.162247166,-0.0843454227,-0.1278079152,-0.0805024207,0.0736151412,-0.1590747684,-0.218530193,0.1113522947,-0.0300942156,-0.2708160281,-0.1825365573,0.2240138501,-0.1710543334,0.0879482105,0.0159658473,0.0831059739,-0.014135113,-0.0620023534,-0.1454125792,0.0298042018,-0.3128145933,0.4159113467,0.2049835175,-0.2109167129,-0.0561014414,0.1267626286,-0.0717334598,0.3511763811,-0.543887496,0.3058475256,-0.1919425577,0.1967351437,-0.1046427116,0.044099953,0.0190379974,-0.155185923,0.1073558256,0.1218106598,-0.2267125845,-0.2391878963,-0.1558688432,0.3402346671,-0.0448819697,0.3125685453,0.4254991412,0.5209850073,-0.0903389901,-0.1613211632,0.3123972714,0.1599565744,0.2083926946,-0.3067101836,-0.1090814248,0.2963912189,-0.0646377429,-0.1498523057,0.49757725,0.1420865506,-0.1262381226,0.244457081,-0.3252066076,-0.0537730493,-0.0712484121,0.2029606402,-0.2755312622,0.2106801271,-0.0213302597,-0.1758273393,-0.0468518585,-0.3733199239,-0.0410228372,0.1587292254,-0.0383768789,0.2196475416,-0.5036861897,0.2241736352,-0.3506961465,0.1727131158,-0.1008414179,0.3289103806,0.1193172261,0.1923475564,0.2529752553,0.0295973029,0.5905706882,0.1286064684,-0.1712749302,-0.1472329646,-0.1064749435,-0.1970107257,0.0440700985,-0.1064045504,0.2871232033,0.4626609087,0.4678314626,-0.3716439307,-0.1510194689,-0.2931528986,0.55991292,-0.0111813573,-0.3234752417,-0.0211638138,-0.1401328146,-0.5439132452,0.1772209257,0.1554722786,0.3789795339,0.1098548025,-0.0748986006,-0.1285215765,-0.0583997034,0.3375995457,-0.0386172235,0.1393350214,-0.1615232527,0.2351150513,-0.2048264444,0.1606556028,0.5302140117,-0.0235485043,0.1311443597,-0.2705217898,0.0150603801,0.0480278581,0.4064321518,0.0143546509,-0.1317492127,0.3606282473,-0.2724704146,0.274716109,-0.186923489,0.5377910137,0.1321879625,0.0761110336,-0.8546541333,-0.3134076595,0.2389555424,0.2096378952,-0.0203944258,0.2407515645,-0.1654565632,-0.2896406651,0.2982034385,0.0951959863,0.8781989813,-0.4335549772,0.3147643507,-0.3794058859,0.341439724,0.5416302085,-0.6066262126,-0.048749119,-0.4083095193,0.161525026,0.1442632377,0.0018779256,0.0331327841,0.4243250787,0.0419178642,0.2461102009,0.1147083342,0.142683059,0.0231641904,0.6419767141,-0.0407159105,-0.3181240857,-0.1395403147,0.0397098288,-0.2120798528,-0.0449907742,-0.2248385549,0.0811411887,0.0633500218,0.0467655547,-0.2977199554,-0.1569368988,-0.106802687,0.1042602211,-0.0331255421,-0.1095273122,0.5250971317,-0.0911677927,0.249813661,0.52596277,-0.1830591857,0.1304571778,-0.3457993567,-0.1237147003,0.1785326004,-0.2479604781,0.0235177651,-0.0207599066,0.0566466525,-0.2709762156,0.1404962987,-0.0610768832,-0.1720446348,-0.0925641805,0.5309393406,-0.2246365696,-0.1855248064,-0.2013044655,0.0065347459,-0.058891207,0.0140479309,0.2020163089,-0.0909735486,0.3338695467,-0.0031674027,-0.2340922654,-0.1526648402,0.1435001045,0.0647662878,-0.0494586714,0.4163269401,0.2087792754,-0.16092439,-0.1365436763,0.3690233827,0.2042340189,-0.1333568096,0.1669920385,0.2157006413,0.0336738005,-0.0534231849,0.0314004421,0.16242145,-0.4770179093,-0.0215207357,-0.2094075233,-0.4017368257,0.0650749579,-0.0659414455,0.1400868595,0.0986139551,0.1832349449,-0.1213977188,0.077423051,-0.1426038742,0.3031952977,0.071724914,0.2567864358,-0.201208055,0.0669809803,-0.2717162073,-0.0734872967,-0.1367842257,0.181003809,0.0861925036,-0.1584144235,-0.1302120984,0.2053361386,0.2879722714,-0.3035103977,0.1958216429,-0.2923232913,-0.1571637392,-0.2866822481,0.2603016496,0.2527885139,0.0190024599,-0.2047920525,0.2475490868,-0.183520481,-0.3007825017,0.1086715683,-0.0695391148,0.1680461019,0.1572906822,0.3060738146,0.5086444616,-0.1263925433,-0.2461180538,0.2235292792,0.537969172,-0.1390584856,0.0469004735,-0.2240172625,0.0562618449,0.3431706429,0.1255344301,0.4919006228,-0.1794983596,-0.4391124547,0.5194427371,0.1138290241,-0.0627842173,-0.1875599772,0.0103217093,-0.2073718905,-0.2274604589,0.1133365855,0.2939917743,0.2318718433,-0.4301713705,-0.1073142663,0.4482294023,0.2221769392,-0.0320121124,0.3989517689,0.2036204487,-0.0728804171,0.3351424634,0.3068632483,0.2198588401,0.5268708467,-0.0581775159,0.4922595322,0.1662915349,0.2116738111,-0.2272545248,-0.5121085644,-0.21156542,0.4970754683,0.2400820255,-0.121946238,-0.2991296649,0.2617676556,0.283981204,-0.1622587293,-0.0132668819,0.1090702191,-0.2722469568,-0.1288956851,-0.1676670313,-0.1421465278,0.0024206976,0.2354127169,-0.2957063019,-0.3620262444,-0.1098732203,-0.0170904733,-0.1344923973,-0.4729762077,0.3292241096,0.1380198002,-0.0384657979,-0.2397853732,0.3410393,-0.2583078146,0.0637729689,0.008056391,0.159217,0.3800536394,0.4069366157,-0.044939395,-0.0394804478,-0.2551726103,0.1269799024,-0.1580903977,0.471390605,0.1255235076,0.2445816696,0.2040171474,-0.0757748187,-0.0273606125,0.0225054678,0.1525158882,-0.3925790191,-0.134425655,-0.0289971419,0.1252280027,-0.3083194494,-0.1817846,0.1292359233,-0.2639212906,0.0268078111,0.9157556295,-0.3334185183,0.0808728114,-0.1699113101,0.0257630255,-0.2719905674,0.4641678035,0.453031987,0.0959292948,-0.4936905503,-0.1008706838,-0.423951298,0.2393170446,-0.4164301753,0.2648034692,-0.032267604,0.0965504125,-0.0368506387,0.0834422037,-0.3513628542,0.4076350033,-0.3663589656,0.0291205551,-0.1237226874,0.1158565879,-0.1410600394,0.1594539583,-0.1152452379,-0.4063183963,0.3437701166,-0.2157748342,-0.1688446701,-0.1341076344,0.2802434862,0.1360771954,-0.1517363638,0.4600016475,0.1406886131,0.2341683954,0.1425192654,0.2970135212,-0.0326542892,0.1505418271,-0.3552688658,0.1838354319,0.027948929,-0.002260362,-0.1371672601,0.231285587,-0.0470789373,0.0380471162,-0.0372092314,-0.0808859468,-0.1327141076,-0.0746406242,-0.1198441163,0.1599339992,0.0416635424,0.3343163431,-0.0387916341,-0.0374467,-0.3385150433,-0.2272844464,0.2538482845,-0.414409548,-0.2750650346,-0.1406800896,0.0405467302,0.6282331944,-0.2390319556,-0.5446512699,0.0836587995,0.5542650819,0.0989967212,-0.2914943397,0.2658554316,-0.1875569969,0.0611147024,-0.1237359047,0.4827390611,-0.1363034248,-0.2857731879,0.078294158,-0.1773688942]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1825","title":"Datasets library not suitable for huge text datasets.","comments":"For information: it's now available in `datasets` 1.3.0.\r\nThe 2.0 is reserved for even cooler features ;)","body":"Hi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??","comment_length":17,"text":"Datasets library not suitable for huge text datasets.\nHi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??\nFor information: it's now available in `datasets` 1.3.0.\r\nThe 2.0 is reserved for even cooler features ;)","embeddings":[-0.4345221519,0.1332366019,0.0117715299,0.2648242116,0.2677350342,0.020276349,0.0753503516,0.3607457876,-0.3179919422,-0.0622065589,0.035968475,0.0512572899,-0.2549733818,0.1889113784,0.052126471,0.0839582682,0.1855184138,0.1725188196,0.1144125983,0.1579608023,0.0227796771,-0.1485536993,-0.2536755502,0.0014495688,-0.500429213,-0.2469414175,0.001542429,-0.2171645463,-0.4491963685,-0.3404015005,-0.1768149287,0.1993858516,0.5116481781,0.4627266228,-0.0001238177,-0.2813290656,0.2467759699,0.0012895344,-0.2723431289,0.0447508805,0.3876777589,-0.3245376348,-0.2008230388,-0.1749605089,0.1608986109,-0.2847053111,0.0148903253,-0.0699783564,0.0965193138,0.0494214557,0.0779271871,0.0929794982,0.0187179893,0.1376075447,0.1242062077,0.0346610248,-0.0822470486,0.0537681244,0.450540334,0.0497524925,-0.1552138776,0.1750814766,-0.1166119501,0.0095225954,0.3391453922,-0.1126097068,-0.1405123919,-0.1884115785,0.2254876196,0.2983859181,0.5142073631,-0.4272146523,-0.2893555462,-0.5464208126,-0.2044093609,-0.3196335435,-0.0669308677,0.1782543808,-0.0131109236,0.2042060941,-0.4558200836,-0.4421231747,-0.1168552041,0.3653897047,-0.2350143194,0.1488129497,-0.1368505657,0.1141696051,0.165915221,-0.2179446071,0.1126268208,-0.3288344741,0.0219105557,0.1951112449,-0.3774071336,-0.1529268175,-0.2038584799,-0.0359365866,0.4822560847,-0.0165515207,-0.4332740009,0.1530094594,-0.1389319003,0.0294946805,0.3459942341,0.2049816549,-0.1033098772,0.264049083,0.1982870251,-0.1752960384,-0.0434123948,-0.0784640163,-0.5217927098,-0.0583364256,0.1153693572,-0.467672497,-0.2794636488,-0.1709811538,-0.07688234,0.0893994644,-0.1661054492,0.3055479527,0.1728393286,0.1523613483,-0.1039775908,0.0859919488,-0.2191548795,0.1837836951,-0.0974596068,-0.1802397817,0.0022475657,-0.1083154082,-0.2347271442,0.1755829602,0.1531202048,-0.043590188,0.1302113384,-0.0576467887,0.0576003231,0.0121801309,0.1697938591,-0.3426538706,-0.0727010965,0.1154427007,-0.0238889977,0.3663884103,-0.2478738576,0.0542965382,-0.2058062255,0.3545786738,-0.2399349213,-0.3611765504,-0.0457572043,0.0719051436,-0.340780288,0.020672312,-0.5490440726,0.3761832118,0.4263476431,0.0905840993,0.011853788,-0.1186689585,-0.24796094,-0.016100876,0.172204569,0.1334883869,-0.5778154731,0.0366079248,-0.0681395009,0.1124098897,0.3034361899,0.754522562,-0.1512485296,0.3304150999,-0.0916665569,0.4423208535,0.145202443,0.159895286,-0.5134213567,0.4223712683,0.0371215418,-0.164030388,-0.0128785549,0.1075425148,0.3807855546,-0.1396206617,0.1872460544,0.3340221345,-0.0645646378,0.0848616734,-0.2775012851,-0.3438349664,0.1302629858,0.288086772,0.1969875246,-0.2810756862,-0.0097941812,0.3730165064,0.1329231858,-0.1553367674,0.3648222387,0.1345099509,0.1909362078,0.1730598807,-0.0348844379,-0.2371498346,0.1483793855,0.0740290731,0.1063018963,-0.1766885221,-0.0091572646,-0.3077905178,0.1440496892,-0.0243835747,0.0311875548,0.1103275791,-0.1044520289,-0.219715625,-0.0854507387,0.1351994723,-0.1043639556,-0.1191572174,-0.3198543191,0.1355397701,-0.0892160609,-0.0005721598,0.062965624,-0.1361074448,0.041191835,0.224611789,-0.1476466656,-0.1313588023,0.021907825,0.1872600168,0.0183313452,-0.2357047498,-0.2702253163,0.5758532882,0.2275518328,0.0480047278,0.2099347413,-0.0309427567,0.0352990106,-0.0920437574,0.0470425598,0.2821452618,0.1014203727,0.1170904115,0.0349614583,-0.2804763317,0.1036198661,-0.0223459899,0.1937521845,0.006490774,0.2591932118,0.0792194158,0.3156825602,0.3179970682,-0.1806907803,-0.1745746583,0.3581426442,-0.3210413158,0.2150854319,0.4250735343,-0.4366922379,-0.3382962644,0.0725767687,-0.2350004613,0.1056227908,0.142742306,0.0731131583,-0.0428087004,0.3631473482,-0.0525317565,0.1752878577,0.1317133904,0.1352782249,0.2091443986,0.1257627159,0.1298955083,-0.1594125479,-0.0640987232,0.1919549555,0.3366985321,-0.1090408564,0.0508374013,-0.2087420374,-0.3060903251,0.0004508442,0.2116795033,-0.1640042812,0.042693004,0.1251249313,0.0437285565,0.1306495816,0.0544447936,0.1263915896,0.3693228662,0.0503856726,-0.3098164201,0.1500760764,-0.0545791127,-0.4672508836,-0.0370833911,0.3112077713,-0.3574154675,0.2815753818,0.2022108436,0.0586500317,-0.3188081682,-0.4476560354,0.0214609336,0.1317866743,0.2077723891,-0.1410446167,0.0467339642,-0.1914004982,0.1635475755,-0.1014137268,-0.1061106697,-0.0787284598,0.0649727359,-0.1645175815,-0.1977962554,0.1205721647,-0.0024966493,-0.2704381347,-0.19324781,0.224264279,-0.1564694494,0.0722561777,0.0276173782,0.064383015,-0.0258462578,-0.0570879094,-0.1401272118,0.0512430221,-0.313062191,0.4259621501,0.2022450566,-0.1976532489,-0.0776709914,0.1161190644,-0.0912989229,0.3588635623,-0.5269432068,0.3114678562,-0.1990622729,0.1910693496,-0.0856881663,0.017711157,0.0093990229,-0.1782578677,0.1044141427,0.125222981,-0.2186260968,-0.255084604,-0.1566656381,0.3327920735,-0.034482915,0.3154123425,0.4613989294,0.5344964862,-0.0702422932,-0.1545268148,0.3131116629,0.1734436452,0.2241769284,-0.3325264156,-0.0860123485,0.2662095726,-0.0626249537,-0.1146854684,0.4754683971,0.1411808878,-0.1146157011,0.2510148883,-0.3503859043,-0.024832299,-0.0610997751,0.1744010895,-0.3004296124,0.2343150228,-0.0234205332,-0.1732247919,-0.0275207236,-0.3812793195,-0.0580097176,0.1766906679,-0.0298212059,0.238510415,-0.4791231751,0.2534342408,-0.3464235067,0.170521602,-0.1004620269,0.3617984354,0.1357777268,0.1863541901,0.2199883759,0.0214749742,0.5436990857,0.124103047,-0.1632372737,-0.1701993048,-0.125470981,-0.2008711398,0.051249411,-0.1064467132,0.2916578054,0.4566519558,0.4787528217,-0.3634493947,-0.1394793689,-0.2989522219,0.5648918748,-0.0116858352,-0.295166254,-0.01168619,-0.1437797546,-0.5237726569,0.1901263744,0.1533011347,0.3718514144,0.1237611175,-0.0931889564,-0.1831271797,-0.09198495,0.328253001,-0.030057231,0.0913557112,-0.1428197026,0.2336041182,-0.2097428888,0.14488478,0.5053817034,-0.0385293029,0.1447773278,-0.289925158,0.0055155815,0.0698986873,0.3870850801,-0.0314584114,-0.1435098052,0.3864057362,-0.2510170639,0.2598465979,-0.193682462,0.5484662056,0.1282880008,0.0551667176,-0.861471951,-0.3365226388,0.239101246,0.2278057933,-0.0184535738,0.2278820872,-0.1837542355,-0.2924677134,0.276453346,0.1032720804,0.8179162741,-0.4555811286,0.3244090378,-0.4055635333,0.3411645889,0.5398203135,-0.5751265287,-0.036534965,-0.3981420994,0.1355864257,0.1298666894,0.0066971155,0.0402744897,0.4150400162,0.0717967674,0.2767036855,0.1156628728,0.1328900903,0.0083337538,0.6496943235,-0.07788679,-0.3102951944,-0.134836778,0.0370053425,-0.1922157854,-0.0505622029,-0.2243538052,0.1070405245,0.1151734963,0.0607990958,-0.3009162545,-0.1705037653,-0.1133765504,0.1218414679,-0.0318334624,-0.1087736785,0.506159544,-0.0841539875,0.2461420894,0.5565766692,-0.1738191247,0.1275268346,-0.3588402271,-0.1274367422,0.1747511029,-0.2499728501,0.0264440086,-0.0169457048,0.0973863453,-0.2636167705,0.1535229385,-0.0709164739,-0.1601657867,-0.0865984708,0.5369499326,-0.2056362629,-0.1871614903,-0.2134485394,0.0121409399,-0.0626670122,0.0102580972,0.1915257573,-0.0572933815,0.2991551459,-0.0133971265,-0.2419014275,-0.1446450353,0.1430332214,0.0640015975,-0.0426904224,0.4207988381,0.1829645485,-0.1477912217,-0.1278966069,0.3998601139,0.2036781609,-0.1033186466,0.1846523434,0.2032349259,0.029056875,-0.0601439066,0.0495668277,0.1374868602,-0.498546809,-0.052676551,-0.2140190005,-0.3960346878,0.0668353811,-0.0665161386,0.1478191018,0.1198974028,0.1561048925,-0.1483896673,0.0702316388,-0.133462429,0.3090940118,0.0708299279,0.261269331,-0.2005297393,0.0824506059,-0.2894151807,-0.0595850013,-0.1407784522,0.173431918,0.0907121152,-0.1506279111,-0.1572410464,0.2040564865,0.3015046716,-0.2897494435,0.1962042302,-0.2798905969,-0.1721050739,-0.3013718426,0.2562596202,0.2545862496,0.0247845817,-0.2037382871,0.2178129405,-0.1880245209,-0.3108050227,0.1195622236,-0.0631842017,0.1413165778,0.1413948089,0.3235113621,0.5331244469,-0.1139003187,-0.2503632009,0.216642186,0.5411872268,-0.1625242084,0.0464541502,-0.1985036731,0.0562483557,0.3416445553,0.1208994985,0.4793934822,-0.1609943956,-0.4485669136,0.5291866064,0.101811856,-0.0579216816,-0.1864106357,-0.0204842985,-0.188623026,-0.2163289487,0.1032048017,0.289916873,0.2510035336,-0.4375896454,-0.1248647124,0.4320606589,0.2399139702,-0.0240661744,0.4210997224,0.2311270535,-0.0921539962,0.3365715146,0.3232385218,0.2409477532,0.5528314114,-0.0477683023,0.5267713666,0.1754154563,0.1959537417,-0.1960280389,-0.503600657,-0.2226984799,0.4708331823,0.228330195,-0.1109245643,-0.3145104945,0.2295250744,0.2637507617,-0.1477516145,-0.0191179179,0.1155165955,-0.2704103589,-0.1847075224,-0.1443496943,-0.1364813298,0.026638044,0.263538152,-0.2816627622,-0.3591853082,-0.1041412503,-0.033550635,-0.0890957192,-0.454290837,0.3771337569,0.1530843824,-0.0305491351,-0.237596184,0.3415534198,-0.2511643469,0.0685580373,0.0139335142,0.1642075032,0.3705232739,0.3742658794,-0.0429952815,-0.0434730053,-0.2629970312,0.1246314645,-0.1643608212,0.4631353617,0.1134570539,0.2455964237,0.2111252546,-0.0780475289,-0.0154689755,0.0147947306,0.1294232458,-0.403185904,-0.1628544927,-0.0139064798,0.1217602268,-0.2851449549,-0.1575857997,0.1719270349,-0.2431897074,0.0065637296,0.9193087816,-0.3252094984,0.0808188096,-0.1628735662,0.0249119401,-0.2839209735,0.4645775557,0.4519178867,0.0823256075,-0.445536375,-0.0914772302,-0.4034664929,0.2238804549,-0.404019624,0.2532674968,-0.0449525267,0.1031092852,-0.0548036769,0.0809808075,-0.3658023775,0.4317956567,-0.3566435874,0.0201066565,-0.0732471868,0.0947919711,-0.1340707839,0.1532260031,-0.1174076796,-0.4135877192,0.3353910148,-0.2376116961,-0.1686392128,-0.1295073777,0.2847554088,0.1521080583,-0.1714214832,0.4486646652,0.1417051852,0.2121012211,0.1436334103,0.287457794,-0.023393983,0.1706351042,-0.3489333689,0.1779577136,0.0055746161,-0.0166500397,-0.1431230307,0.2426627278,-0.0186138395,0.0469753742,-0.0603478327,-0.091692403,-0.1263712645,-0.0789710283,-0.1517826319,0.1540196389,0.0576653183,0.3408267498,-0.0482659265,-0.0625468343,-0.3105644882,-0.2342225015,0.2298222482,-0.4213120341,-0.2648097277,-0.1801561415,0.0154333189,0.6642018557,-0.2201962769,-0.5252726674,0.1011431143,0.5639672875,0.0911809802,-0.3091644049,0.2561299503,-0.1819330454,0.0629252866,-0.1353604794,0.4806586802,-0.149354592,-0.2951605618,0.0539431237,-0.1682290286]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1825","title":"Datasets library not suitable for huge text datasets.","comments":"Hi @alexvaca0 , we have optimized Datasets' disk usage in the latest release v1.5.\r\n\r\nFeel free to update your Datasets version\r\n```shell\r\npip install -U datasets\r\n```\r\nand see if it better suits your needs.","body":"Hi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??","comment_length":35,"text":"Datasets library not suitable for huge text datasets.\nHi,\r\n\r\nI'm trying to use datasets library to load a 187GB dataset of pure text, with the intention of building a Language Model. The problem is that from the 187GB it goes to some TB when processed by Datasets. First of all, I think the pre-tokenizing step (with tokenizer.map()) is not really thought for datasets this big, but for fine-tuning datasets, as this process alone takes so much time, usually in expensive machines (due to the need of tpus - gpus) which is not being used for training. It would possibly be more efficient in such cases to tokenize each batch at training time (receive batch - tokenize batch - train with batch), so that the whole time the machine is up it's being used for training. \r\nMoreover, the pyarrow objects created from a 187 GB datasets are huge, I mean, we always receive OOM, or No Space left on device errors when only 10-12% of the dataset has been processed, and only that part occupies 2.1TB in disk, which is so many times the disk usage of the pure text (and this doesn't make sense, as tokenized texts should be lighter than pure texts).\r\n\r\nAny suggestions??\nHi @alexvaca0 , we have optimized Datasets' disk usage in the latest release v1.5.\r\n\r\nFeel free to update your Datasets version\r\n```shell\r\npip install -U datasets\r\n```\r\nand see if it better suits your needs.","embeddings":[-0.4028155208,0.0771392435,0.0086002536,0.3193157315,0.3035772443,0.0233495664,0.0301555488,0.3559309542,-0.2499681413,-0.0241902824,0.0388909541,0.1070225537,-0.2629811168,0.1188514456,0.0347161479,0.1091509685,0.2107038796,0.1721984446,0.1067035645,0.146104753,0.0033639085,-0.0768989548,-0.2359556705,-0.0442245565,-0.4917674363,-0.2467965335,-0.0052548172,-0.2256905138,-0.416131705,-0.3108104765,-0.1463014185,0.2057922333,0.4900791645,0.4980648756,-0.0001231333,-0.269946456,0.2363723516,0.0186441652,-0.3187198043,0.0508903265,0.4693198502,-0.3577676415,-0.2028603107,-0.2081337422,0.1659426242,-0.2171029598,-0.0014259464,-0.1183160767,0.1075647473,0.0541204512,0.0970020145,0.130324468,0.0105504636,0.1590501219,0.101554513,0.0934823006,-0.0940500125,0.0596261993,0.4631817937,0.0898426846,-0.1676841974,0.2201166749,-0.1535094678,0.0275999047,0.3630459607,-0.0872998089,-0.0916152596,-0.1663148254,0.2746692002,0.2592946291,0.5688686371,-0.4413167238,-0.261215359,-0.545519948,-0.221926555,-0.2722447515,0.004418415,0.1677701473,-0.037591964,0.1915647388,-0.4412501752,-0.4712109268,-0.0885306075,0.3492920399,-0.2183879167,0.0799290761,-0.1663344651,0.102665931,0.2382712811,-0.2418371588,0.102812469,-0.3338505328,0.0022664245,0.1644580215,-0.3736112416,-0.1478253454,-0.2142490298,-0.057693895,0.4252462387,0.0029207745,-0.415071547,0.1444519013,-0.1237988397,0.0363089368,0.3516076505,0.2578930855,-0.099096708,0.2503686249,0.1979775131,-0.1737039536,-0.0160005689,-0.1105490923,-0.5208440423,-0.066299513,0.1376723647,-0.5249603987,-0.2645240426,-0.1763340384,-0.0862714052,0.0780038387,-0.0980500206,0.2459835708,0.1834304184,0.181846872,-0.0734423846,0.1200844571,-0.1878760457,0.1694710404,-0.1226170287,-0.1676702648,0.0043567512,-0.1066701487,-0.2287060469,0.1115951836,0.1700580269,-0.0129800774,0.1193298697,-0.0604935177,0.03899711,0.0096373605,0.1425361633,-0.3770751059,-0.0747798011,0.1250582784,-0.0162472688,0.3819413483,-0.2452010512,0.0438579172,-0.231059745,0.3684508801,-0.2714810669,-0.3721601069,-0.1170408577,0.0728527829,-0.3259705901,-0.0053129266,-0.609857142,0.3593026996,0.4154552519,0.0878379345,-0.0540280528,-0.102754645,-0.2495370954,-0.0512957126,0.1633385271,0.1536833495,-0.5558549166,0.0424050689,-0.105006054,0.1381257772,0.3448683023,0.7818447948,-0.1540688574,0.3443174958,-0.0837314427,0.3829895556,0.1397251934,0.0865029097,-0.5390255451,0.4260479808,0.0805559084,-0.146929875,0.0165783502,0.108947143,0.3508238196,-0.1401628107,0.1549426317,0.3197361827,-0.0168230794,0.0955384597,-0.3492801487,-0.3166877329,0.1213875636,0.3060807288,0.1666428149,-0.3364423513,-0.0267380495,0.3738551438,0.1500833482,-0.1257288009,0.3451298475,0.1695325971,0.2498047352,0.1797233522,-0.003174836,-0.2126871198,0.1286491156,0.0907980949,0.1697634459,-0.1913997829,-0.0275453646,-0.2772671282,0.1492590904,-0.0167024639,0.0196446124,0.1086308062,-0.0833172873,-0.2104953378,-0.0578190275,0.1369724125,-0.0665684193,-0.1515645534,-0.3148004711,0.1400073916,-0.0645669997,0.0235904809,0.0524303429,-0.116723448,0.0183351226,0.2091658115,-0.1672680229,-0.1363949031,0.0121231256,0.1948985308,0.0151766548,-0.2138117105,-0.1999004036,0.5545613766,0.2331965268,0.0591159351,0.2807889879,-0.0021756282,0.053805925,-0.0984087959,0.0032171234,0.2129262537,0.0822622254,0.150980249,-0.012264736,-0.2621985674,0.1246827915,0.018197218,0.2300263643,0.017934218,0.2678823471,0.0690589547,0.3252793849,0.3191105127,-0.1439495832,-0.1528828442,0.3700921535,-0.3132962286,0.2127896398,0.4723603427,-0.4098826945,-0.3531783223,0.0724826008,-0.1715939343,0.1418943852,0.1543644071,0.0579908714,-0.092488952,0.3867768645,-0.0452662185,0.2055207789,0.1396352947,0.1116908416,0.1776803732,0.1302106678,0.115786843,-0.1611314118,-0.0654386654,0.1719735414,0.3680223525,-0.1127545685,0.0396668427,-0.2015713602,-0.2762778997,0.0119576966,0.2238869667,-0.1521978378,0.0127018597,0.0925220028,0.0416020341,0.1437939256,0.0740731284,0.2089531124,0.3292385042,0.0085600037,-0.307706356,0.1153598651,-0.0328585468,-0.4963890314,-0.0283088312,0.3630423546,-0.341493994,0.2948131859,0.1539149731,0.076633364,-0.3204392493,-0.4033299088,-0.0133970762,0.1026637033,0.22246176,-0.1516816914,0.0797144473,-0.2505561113,0.1806444526,-0.0657986626,-0.1377865076,-0.0574445799,0.0679614693,-0.1512971371,-0.2519557178,0.0981158093,-0.0180098768,-0.3014727533,-0.1782745421,0.2398389727,-0.1504434198,0.1131775454,0.0142042711,0.0704071447,-0.0160824694,-0.0572495908,-0.1329592466,0.0318354033,-0.3583616316,0.4102056623,0.2523451746,-0.1754553914,-0.0017284822,0.1510429233,-0.086694181,0.2985329926,-0.536164403,0.2482324541,-0.2271320075,0.1707008481,-0.0989369527,0.0939537883,0.0085230693,-0.1408994347,0.1126374751,0.1357520074,-0.2377644032,-0.2494041175,-0.1299275756,0.3253559768,-0.0392635763,0.3188129663,0.5074319243,0.4777010679,-0.0744746029,-0.1131634191,0.3097053766,0.1841125339,0.2432103008,-0.3235909045,-0.1137646511,0.2541649938,-0.0705374926,-0.1344346404,0.4563224614,0.1818729639,-0.1035492867,0.2594854236,-0.3549362421,-0.0414930284,-0.0583156422,0.202291429,-0.24576208,0.2087108046,-0.0166273061,-0.1609211862,-0.0318124518,-0.3627414405,-0.0547643863,0.1677932441,-0.01100381,0.2217927426,-0.4764861465,0.2394002378,-0.3992719352,0.1984274685,-0.1730269939,0.3066082597,0.1419065595,0.2114693373,0.2595523596,0.017400749,0.5492854118,0.1528874636,-0.1475335509,-0.1698978841,-0.1421878785,-0.2579243183,0.0550858863,-0.1505862176,0.2799949646,0.4357599318,0.5038671494,-0.3647415638,-0.2012859285,-0.3397356272,0.5800454617,-0.0480564162,-0.2907161117,-0.0267969742,-0.2180372328,-0.5310125947,0.1750017703,0.1872734427,0.3673219085,0.1741114706,-0.0582792349,-0.1614657193,-0.0703887939,0.3420441747,-0.0464393683,0.184415862,-0.1968512237,0.2335072011,-0.2128456384,0.1189101189,0.511797905,-0.0424066484,0.0731494352,-0.2060356587,-0.0037427223,0.0500433184,0.3726010621,-0.0330478884,-0.1493559033,0.3144882321,-0.2959533632,0.2798677981,-0.2172093987,0.4888456166,0.1572149396,0.0978677943,-0.7616376877,-0.3631928563,0.2361015528,0.2465048283,-0.0222331639,0.1951895803,-0.1537142694,-0.2862220407,0.3004185855,0.1105560511,0.826004684,-0.4686470926,0.2963401675,-0.3631171882,0.3105923831,0.5010316968,-0.6091917753,-0.032178238,-0.3963600397,0.1395549327,0.1546203047,0.0086515918,0.0458436012,0.4068327844,0.0398775749,0.3078384101,0.1045456678,0.0989680812,-0.0132916458,0.6598169208,-0.1008167565,-0.317669034,-0.1528227627,0.0603333898,-0.2292146683,-0.0482931547,-0.1904260963,0.0984812379,0.1104760766,0.0445938669,-0.302424401,-0.2183114439,-0.1078225598,0.1403215379,-0.030979922,-0.0960663036,0.4986239374,-0.0021379446,0.20628272,0.5608209968,-0.1603024155,0.1573555022,-0.3864063323,-0.0873467848,0.1849469244,-0.1999988109,-0.0410522372,-0.0567885712,0.095838137,-0.3077464104,0.1138442755,-0.0688253865,-0.1552900821,-0.082139641,0.4945062399,-0.1761374623,-0.2396705151,-0.2425158471,-0.0239895247,-0.0590677857,0.0263952538,0.2346218079,-0.076795049,0.3747395277,-0.0302199777,-0.2205341905,-0.1249228865,0.1324593425,0.0872406811,-0.0343001001,0.4078709185,0.1404269487,-0.1649450511,-0.1530836672,0.4082039893,0.124707669,-0.1289007366,0.210568279,0.2344245166,0.0088986242,-0.0501609184,0.0030257197,0.1752154529,-0.5009049773,-0.0918842927,-0.1805161685,-0.3987695575,0.0736567453,-0.0459936298,0.1356541067,0.1200377569,0.0612377301,-0.1396028697,0.0608017184,-0.1573093981,0.3485962152,0.0797350407,0.2396012843,-0.2243048549,0.0398437306,-0.2172419578,-0.0972272381,-0.1267705858,0.1385208517,0.093940407,-0.1650671214,-0.1413872838,0.1879232079,0.2528239489,-0.2673346102,0.1879463643,-0.2861518562,-0.1580534875,-0.3490875661,0.2193027884,0.2547700405,0.0222536847,-0.1858412176,0.2162828892,-0.2088623196,-0.2859118283,0.112504676,-0.0678051263,0.1309940815,0.1480821371,0.3166667521,0.5489244461,-0.13662754,-0.2980544865,0.1985019296,0.5358867049,-0.1197691783,0.0347278826,-0.2065893412,0.0574977919,0.3748197854,0.1081883907,0.4812161922,-0.1683340073,-0.4468156099,0.4947699308,0.1289743632,-0.0571262576,-0.2476886958,0.002982473,-0.1806352586,-0.2533296645,0.0809936821,0.3137275875,0.2014209628,-0.406683892,-0.1012643874,0.5237576962,0.1653321385,0.0029217589,0.4272448719,0.2243710309,-0.1126737595,0.3166330159,0.2939412892,0.2337683886,0.5904647112,-0.0754012689,0.5070323944,0.2195574641,0.1999546587,-0.2541435659,-0.4688105881,-0.2125880122,0.4607968628,0.2010645419,-0.1069234684,-0.3281907141,0.2388104945,0.2232533544,-0.1692104638,-0.050106816,0.0795331001,-0.2972475588,-0.1967018247,-0.0497738831,-0.0938539729,0.0315491594,0.2399543971,-0.2930003405,-0.391002804,-0.1438047737,-0.0094451774,-0.1031277403,-0.3992467523,0.3458120227,0.1379220635,-0.0330604315,-0.2205663919,0.32623142,-0.2420054674,0.0778665692,-0.0303840972,0.141775623,0.3925178349,0.3610640764,-0.0422690809,-0.0485455953,-0.271145165,0.149219662,-0.1828677803,0.4642491043,0.0610441118,0.2757788599,0.181659475,-0.0559629165,-0.0337355025,-0.026884716,0.1235690862,-0.3364807665,-0.14848499,-0.0217156541,0.0802412331,-0.2911404669,-0.1484193653,0.1495282799,-0.2476088554,-0.0356494933,0.8947803974,-0.2939738333,0.0742620528,-0.1504418552,0.0362366438,-0.2652697265,0.4319400489,0.418150872,0.0045709051,-0.4741909802,-0.0920988172,-0.4654784203,0.1554789245,-0.3860726058,0.2582815886,-0.0038844277,0.1088590473,-0.0433051921,0.0417872183,-0.3536687791,0.4079421163,-0.3383268714,-0.0126243541,-0.1160093471,0.1326435655,-0.1304860562,0.1457920521,-0.0697081238,-0.4247712493,0.3027373254,-0.2206453681,-0.1616446376,-0.1389219463,0.3290127516,0.1359473914,-0.0847077891,0.4330039024,0.1464904398,0.1933506578,0.1440932155,0.230001539,-0.0101876697,0.1664080769,-0.3035311699,0.1640965641,0.0135092931,0.0317514054,-0.1838488579,0.2569767833,-0.0370001569,0.0674785972,-0.0618181415,-0.07447806,-0.1171975881,-0.0723747909,-0.1005485579,0.1552348584,0.0650890023,0.3275765479,-0.0324825384,-0.059436366,-0.3826101422,-0.2208423167,0.2281024009,-0.3616298735,-0.2914254963,-0.150273636,0.0420755893,0.6636847258,-0.1947220713,-0.5574444532,0.0941894203,0.536716342,0.0990346819,-0.3179399967,0.2728794217,-0.1779838055,0.0576391146,-0.0885170847,0.5082892776,-0.1583695561,-0.3004288673,0.0639650375,-0.1544060707]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1821","title":"Provide better exception message when one of many files results in an exception","comments":"Hi!\r\n\r\nThank you for reporting this issue. I agree that the information about the exception should be more clear and explicit.\r\n\r\nI could take on this issue.\r\n\r\nOn the meantime, as you can see from the exception stack trace, HF Datasets uses pandas to read the CSV files. You can pass arguments to `pandas.read_csv` by passing additional keyword arguments to `load_dataset`. For example, you may find useful this argument:\r\n- `error_bad_lines` : bool, default True\r\n Lines with too many fields (e.g. a csv line with too many commas) will by default cause an exception to be raised, and no DataFrame will be returned. If False, then these \u201cbad lines\u201d will be dropped from the DataFrame that is returned.\r\n\r\nYou could try:\r\n```python\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files), error_bad_lines=False)\r\n```\r\n","body":"I find when I process many files, i.e.\r\n\r\n```\r\ntrain_files = glob.glob('rain*.csv')\r\nvalidation_files = glob.glob(validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n```\r\n\r\nI sometimes encounter an error due to one of the files being misformed (i.e. no data, or a comma in a field that isn't quoted, etc).\r\n\r\nFor example, this is the tail of an exception which I suspect is due to a stray comma.\r\n\r\n> File \"pandas\/_libs\/parsers.pyx\", line 756, in pandas._libs.parsers.TextReader.read\r\n> File \"pandas\/_libs\/parsers.pyx\", line 783, in pandas._libs.parsers.TextReader._read_low_memory\r\n> File \"pandas\/_libs\/parsers.pyx\", line 827, in pandas._libs.parsers.TextReader._read_rows\r\n> File \"pandas\/_libs\/parsers.pyx\", line 814, in pandas._libs.parsers.TextReader._tokenize_rows\r\n> File \"pandas\/_libs\/parsers.pyx\", line 1951, in pandas._libs.parsers.raise_parser_error\r\n> pandas.errors.ParserError: Error tokenizing data. C error: Expected 2 fields in line 559, saw 3\r\n\r\nIt would be nice if the exception trace contained the name of the file being processed (I have 250 separate files!)","comment_length":129,"text":"Provide better exception message when one of many files results in an exception\nI find when I process many files, i.e.\r\n\r\n```\r\ntrain_files = glob.glob('rain*.csv')\r\nvalidation_files = glob.glob(validation*.csv')\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files))\r\n```\r\n\r\nI sometimes encounter an error due to one of the files being misformed (i.e. no data, or a comma in a field that isn't quoted, etc).\r\n\r\nFor example, this is the tail of an exception which I suspect is due to a stray comma.\r\n\r\n> File \"pandas\/_libs\/parsers.pyx\", line 756, in pandas._libs.parsers.TextReader.read\r\n> File \"pandas\/_libs\/parsers.pyx\", line 783, in pandas._libs.parsers.TextReader._read_low_memory\r\n> File \"pandas\/_libs\/parsers.pyx\", line 827, in pandas._libs.parsers.TextReader._read_rows\r\n> File \"pandas\/_libs\/parsers.pyx\", line 814, in pandas._libs.parsers.TextReader._tokenize_rows\r\n> File \"pandas\/_libs\/parsers.pyx\", line 1951, in pandas._libs.parsers.raise_parser_error\r\n> pandas.errors.ParserError: Error tokenizing data. C error: Expected 2 fields in line 559, saw 3\r\n\r\nIt would be nice if the exception trace contained the name of the file being processed (I have 250 separate files!)\nHi!\r\n\r\nThank you for reporting this issue. I agree that the information about the exception should be more clear and explicit.\r\n\r\nI could take on this issue.\r\n\r\nOn the meantime, as you can see from the exception stack trace, HF Datasets uses pandas to read the CSV files. You can pass arguments to `pandas.read_csv` by passing additional keyword arguments to `load_dataset`. For example, you may find useful this argument:\r\n- `error_bad_lines` : bool, default True\r\n Lines with too many fields (e.g. a csv line with too many commas) will by default cause an exception to be raised, and no DataFrame will be returned. If False, then these \u201cbad lines\u201d will be dropped from the DataFrame that is returned.\r\n\r\nYou could try:\r\n```python\r\ndatasets = load_dataset(\"csv\", data_files=dict(train=train_files, validation=validation_files), error_bad_lines=False)\r\n```\r\n","embeddings":[-0.1087328643,-0.4190312624,0.0075155268,0.2815199196,0.184535712,0.1635629386,0.1409228891,0.3934292495,0.2262881994,0.2882854939,0.2727559209,-0.2174746692,-0.0589207113,-0.2714683115,-0.2407846451,-0.1619681716,0.0860742107,0.1235782728,0.2567390203,-0.0407727547,-0.04355729,0.1729320735,0.1145248041,0.2511814535,-0.4897430241,-0.2769164145,-0.3172519505,0.3817558289,0.2166707367,-0.5677108765,0.000150643,0.0310577564,0.1003034189,0.5251937509,-0.0001157862,-0.1504169106,0.150186792,0.0271878988,-0.3778850734,-0.1116750389,-0.1856664419,-0.6877264977,-0.0241563059,-0.2813296616,0.358089596,-0.0366705395,0.2445521355,-0.1590948254,0.3761559427,0.3548002541,0.0947954953,-0.1653120667,0.2582828104,-0.0023046003,0.3424384892,0.1431551725,0.0669529811,0.2645777464,0.3179498017,-0.4284532666,-0.2257969826,0.0672468171,-0.2100593746,0.060488984,0.072225824,0.0547770523,0.3765808046,-0.1714562327,-0.2216404676,0.4012447298,0.2046240717,-0.1446513832,-0.5563559532,-0.6639549732,0.1868021935,-0.299066335,-0.0956107154,0.3052101433,-0.224656716,0.1164680719,0.3411990404,-0.1564714909,-0.3696013987,-0.0656503066,-0.0943443105,-0.1774515957,0.1163334325,-0.0665725842,0.0795393437,-0.0301015154,-0.2295916528,-0.0244011693,-0.1405241638,-0.0659567118,-0.5154403448,-0.0611237176,0.1398732513,-0.594663322,0.2344194651,-0.3872507811,0.3423087001,0.1268074214,-0.0876092389,0.3240608573,0.3673042655,-0.180264622,0.0446514003,-0.073589474,0.4588177502,0.097556524,-0.1286380738,-0.1786259413,0.0580481887,-0.2865338326,0.0131826662,0.491589874,0.0972695798,-0.2422264218,-0.56257689,0.1740452945,-0.2998229861,-0.129499957,-0.1019075364,0.3863109648,-0.1491400152,0.294147253,-0.071840927,0.0743074864,0.1058579758,-0.3490169644,0.0190026928,0.0955065712,-0.1309894323,0.1971978396,-0.1066791713,0.042652946,-0.1053275689,0.212528646,0.1612885594,-0.1958126575,-0.0266665928,-0.4223652482,-0.0123045053,0.1347548068,0.0360869244,0.2787174284,0.1180074811,-0.1997915804,-0.015529871,0.0390922278,-0.5052164793,-0.4055265486,-0.1516358107,-0.0190636199,0.2258602232,0.2026077211,-0.118459411,0.2543493807,-0.065301612,-0.1149035096,-0.0961175859,-0.1168498471,0.2591307163,-0.3146848679,0.0744724944,0.5779359937,-0.3668700159,0.2835256159,-0.1532492936,0.0071681733,0.3014096618,0.501670301,-0.2213502079,0.39205724,-0.3043756783,0.2214913666,0.0214416888,-0.2410392314,0.0696125552,0.5922524333,-0.0995081216,0.3384661973,0.4496501684,-0.0548262373,0.0299315285,-0.1319982409,0.2314151824,0.1058333963,-0.2101560682,0.1193098426,-0.1383868307,-0.0164880119,0.0586396866,-0.0880045891,-0.0457933284,-0.1698718518,0.3214453757,-0.1717794687,0.6836057305,-0.2008433342,-0.198603496,0.3270806074,0.3706170321,0.1801689714,-0.3365702331,0.1081396043,-0.1433928013,-0.0562939867,-0.0956453979,-0.0406308472,-0.3002382517,-0.3844315112,0.0482085273,-0.1880551279,0.1139161587,0.0421368554,0.0909382999,0.1164957806,-0.2879985869,-0.1191728339,0.1267445534,0.4215732217,-0.3311342895,0.2209107429,0.207485348,0.0408954434,-0.1284812987,-0.0950806364,0.1966287941,0.0105177909,0.389231056,-0.238566041,-0.0453906916,0.4332703054,0.0857673287,-0.0695130229,-0.0058437306,0.00813655,0.1619238406,0.1554265022,-0.2069426924,-0.1082888991,-0.2072972506,0.015207666,-0.160050258,0.1527825445,0.1803387702,0.1812221408,0.1355612725,0.0601340458,0.199920997,0.0841965824,0.2385478467,-0.018100366,0.0409563221,-0.1881004125,-0.1225963458,-0.0814209506,-0.5105944276,-0.0667410195,0.3919047415,-0.0359766707,0.1042282656,0.3644504845,0.2609383762,0.0450894274,0.424387455,0.200959444,0.302917093,0.2737497687,-0.166833967,-0.1221467257,-0.2403468937,-0.0530329049,0.1558333784,0.2600558996,-0.0753523186,0.2811388969,-0.0913240835,0.1730599105,-0.0010684173,-0.5670700073,-0.0049581947,0.1933502704,-0.413593024,0.1167614385,-0.3286731243,-0.4525234699,-0.3685514927,0.0141993221,-0.1077616587,-0.2883921564,-0.0081439633,0.1077218056,-0.1718173772,-0.0819019526,0.0299162623,0.1688558608,0.0349316746,0.2562255263,-0.0761338174,0.1002630889,-0.1177725792,-0.1330836266,0.3697405756,-0.139603287,0.0051866383,-0.0098292409,-0.2386259735,-0.3681987524,0.0867851526,0.0194472205,-0.0267849416,-0.1194075644,0.244999662,0.3923146427,-0.1222421974,-0.5848962665,0.1295443475,0.011956431,0.0188376699,0.0164076444,0.1567737907,0.0439011529,-0.0345854722,-0.1233541071,-0.1896684021,-0.4294611812,0.1302319914,-0.4879535735,0.3233160377,0.25190413,0.2206076831,0.1107888818,-0.0591552444,0.3506516814,0.2963022888,-0.2689560354,0.1074777544,-0.1403618306,-0.1288823932,0.0123325856,0.0599699281,-0.1671210527,0.1541715711,-0.4859209955,-0.2744222283,-0.0720594376,-0.0810689181,-0.1790816188,0.0421768464,-0.0532373823,0.0425296649,-0.1384543031,0.1540128589,-0.0546145439,-0.0860621706,0.0819201097,0.5511020422,-0.1832885444,0.380451262,0.2542597353,0.4371030927,0.0799585655,0.2348530889,0.348305285,0.1514683068,0.2243076265,0.4164159298,-0.0817876309,0.0672468469,-0.4160613716,-0.1437940747,0.0633477643,-0.0692745224,0.0354940519,0.0847490206,-0.1507839262,-0.3204994202,-0.4465426207,0.2643267214,-0.3387426138,0.095834665,-0.1978265792,-0.2251453996,0.1519889683,0.18213512,0.1994278133,0.3629790843,0.458083123,-0.0320933685,0.4049341679,-0.4784291387,-0.2203938216,0.2758816183,0.1289044321,0.4157720804,-0.2298939377,0.22357063,-0.130199492,0.018821422,0.8039573431,-0.2620460391,-0.3299178481,-0.0154313054,0.0681546777,0.1336505264,-0.1578281969,-0.1799954027,0.4455316365,0.2653980255,0.3116160035,-0.4177491963,-0.2524324954,-0.0584327988,0.4364773631,-0.0185253024,0.0429916158,-0.0998030901,0.0102289468,-0.2890790701,0.2386060953,0.2925159633,-0.0136920391,-0.016936399,-0.2078027129,-0.0076673734,-0.0479654558,0.4319797158,0.1046213731,0.1274832338,-0.3237047493,-0.0890049785,-0.0513638109,0.4259849191,0.8350525498,0.2568825185,-0.4463810623,-0.4004332721,-0.0267007742,-0.1182921603,0.6949506402,0.546037972,0.0011484664,0.3498132229,-0.2061717957,0.1629811972,-0.0017578322,0.2405006737,0.4419852793,0.2451389134,-0.5271652341,-0.2777045965,-0.0143188406,0.2728558481,-0.152777344,0.248819679,0.282242626,-0.3482763469,0.4481820464,-0.2180053145,0.9202817082,-0.1080999672,0.1769841015,-0.2334662229,-0.2355285287,0.4540643692,-0.4126852453,0.2460375428,-0.3492016792,-0.4742131829,-0.1100322381,0.0365054794,0.3672527075,0.3979192078,-0.0953564942,0.471098423,-0.0768573955,-0.0594707876,-0.1357295066,0.495211035,-0.0808083713,-0.2098915875,-0.46400249,-0.0342162028,0.0577560626,-0.1429520547,-0.0774757341,-0.2687586844,0.153649509,-0.3202186525,-0.5534124374,-0.2738958299,-0.1238663644,0.0520060509,-0.1065474674,-0.2890205681,0.2439143658,0.0601760931,0.0170922298,0.2996908426,0.074878484,0.0250021573,-0.1556888074,0.0086949505,-0.266174525,-0.2534126043,0.4243947566,0.2638280988,-0.1476567388,0.365254581,-0.1935147196,-0.3234770596,0.1794547439,0.0400719121,0.0007884057,0.1152526066,-0.038195122,-0.231710121,-0.2358821183,-0.2420250922,0.0861755162,0.2091705501,-0.181123659,0.3741346002,-0.0825570002,-0.200474143,-0.0525821596,-0.0787113085,-0.4278917015,-0.0917410403,0.2446846217,-0.0592811443,-0.1956137866,-0.1246199235,0.0944462344,0.1489368975,-0.0456994064,0.1569084674,-0.0389284976,0.0742425174,0.0140254283,-0.1312748641,0.4664576948,0.1307409555,-0.2354867607,-0.3894788623,-0.3047163188,-0.000416674,0.0312523469,0.3860535622,0.0768799707,0.1696014255,-0.0221420228,0.2160578072,-0.2095839828,-0.1860386878,-0.152814731,-0.0540534966,-0.1907169223,0.2986856103,0.0557188392,-0.0465523154,0.0703085363,0.2521032989,0.3689203262,-0.0473037437,-0.0336485282,0.1581770629,-0.029782299,0.1915013045,0.206806317,-0.257368505,0.2554954886,-0.2544526458,-0.006263989,0.2262755781,0.0535783693,0.1188180819,0.0879487395,0.0235852581,-0.186526373,0.0087525882,-0.1153787151,0.2938220799,0.3038095832,0.1885798872,-0.1978465617,-0.0713794753,0.1604602039,0.2074501812,-0.1234296784,-0.0735500976,0.111127764,0.1211486384,-0.4866571128,0.3509528637,0.2637128234,0.1089419499,-0.0653704852,0.09641321,0.3562037945,0.0854203254,-0.0353142619,-0.1407131851,0.048770722,-0.3552114964,-0.0006576293,0.2693496048,0.2658037841,-0.1856267154,-0.1421913058,-0.0903941542,0.3831519783,0.1608846188,0.0113519458,0.3325926065,0.0484102219,0.0151844388,0.1616109014,-0.0812548473,0.4215298295,0.3943688869,-0.1059536859,-0.1336726546,0.0422808453,-0.2566126883,-0.2990857661,-0.3335377276,0.3333734572,0.1233001277,-0.2008337826,-0.2057076544,0.1558199972,0.5819082856,-0.3255527616,-0.1698464453,0.0932434499,0.0151967481,-0.1296875775,0.135593906,0.2539349794,0.1188134924,-0.1187751368,-0.3039392829,0.1671482474,-0.1479536742,0.1235577464,-0.0113671971,0.2726219594,-0.0429066084,-0.1298954338,0.1385739446,0.586818397,-0.2482185364,-0.110995695,0.2599121332,-0.1228413284,0.0991812646,-0.1952550858,0.6805849671,0.3961875737,0.2024142146,-0.2139309049,-0.2139332145,0.1438143551,0.0880247578,-0.2117620111,-0.2278227657,-0.1733570695,0.2931328714,0.0596402586,-0.0600832514,0.1721616983,0.2266809642,0.0493763164,-0.1090073809,-0.0734014288,0.0593280904,-0.2405629456,-0.1325250268,-0.1790397912,-0.2729077637,-0.2147063464,0.4603151977,-0.2889192402,0.1828607768,0.1995469481,0.0663000122,0.1559032649,0.0591118671,0.4475522637,0.2244100571,-0.4179139435,0.0879035667,-0.8058602214,-0.1965546161,-0.1332134455,0.0353205092,0.1043250859,0.0707272664,-0.1179591939,-0.1067901477,0.3442202508,0.0314412117,0.1797849089,0.2814751267,-0.4411147237,-0.1865924895,-0.2060501277,-0.0418431275,-0.0726902336,-0.0307980813,0.2907048464,0.3199259937,0.0122961858,-0.1926455349,0.087967582,-0.0157397594,0.1223672852,0.1741416752,0.0154718766,0.4901055098,-0.2843928039,-0.243340686,0.1756736338,0.0526331253,-0.2411286235,0.221332863,0.0903312713,0.2468899488,-0.090796724,-0.3145160079,-0.296418488,0.3545731604,0.1674235016,-0.157546863,-0.2645468414,0.4259020388,0.0076894038,-0.0380671397,-0.011122901,-0.0175953992,0.0060380786,0.1518342048,-0.2819745541,-0.3354225755,0.3631479144,0.1472552717,-0.0215623397,-0.2779440284,0.5018212795,-0.1264557242,0.4052014947,-0.3974799216,-0.3426754773,0.3026511967,0.2451997101,-0.5436035395,-0.1363782287,0.0492596477,0.0748070329,-0.0160509944,0.1413267106,0.1578428745,-0.2588274479,0.2586653531,-0.137397334]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1818","title":"Loading local dataset raise requests.exceptions.ConnectTimeout","comments":"Hi ! Thanks for reporting. This was indeed a bug introduced when we moved the `json` dataset loader inside the `datasets` package (before that, the `json` loader was fetched online, as all the other dataset scripts).\r\n\r\nThis should be fixed on master now. Feel free to install `datasets` from source to try it out.\r\nThe fix will be available in the next release of `datasets` in a few days","body":"Load local dataset:\r\n```\r\ndataset = load_dataset('json', data_files=[\"..\/..\/data\/json.json\"])\r\ntrain = dataset[\"train\"]\r\nprint(train.features)\r\ntrain1 = train.map(lambda x: {\"labels\": 1})\r\nprint(train1[:2])\r\n```\r\n\r\nbut it raised requests.exceptions.ConnectTimeout:\r\n\r\n```\r\n\/Users\/littlely\/myvirtual\/tf2\/bin\/python3.7 \/Users\/littlely\/projects\/python_projects\/pytorch_learning\/nlp\/dataset\/transformers_datasets.py\r\nTraceback (most recent call last):\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connection.py\", line 160, in _new_conn\r\n (self._dns_host, self.port), self.timeout, **extra_kw\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/util\/connection.py\", line 84, in create_connection\r\n raise err\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/util\/connection.py\", line 74, in create_connection\r\n sock.connect(sa)\r\nsocket.timeout: timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connectionpool.py\", line 677, in urlopen\r\n chunked=chunked,\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connectionpool.py\", line 381, in _make_request\r\n self._validate_conn(conn)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connectionpool.py\", line 978, in _validate_conn\r\n conn.connect()\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connection.py\", line 309, in connect\r\n conn = self._new_conn()\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connection.py\", line 167, in _new_conn\r\n % (self.host, self.timeout),\r\nurllib3.exceptions.ConnectTimeoutError: (<urllib3.connection.HTTPSConnection object at 0x1181e9940>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)')\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/adapters.py\", line 449, in send\r\n timeout=timeout\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connectionpool.py\", line 727, in urlopen\r\n method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/util\/retry.py\", line 439, in increment\r\n raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\nurllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/json\/json.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x1181e9940>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/littlely\/projects\/python_projects\/pytorch_learning\/nlp\/dataset\/transformers_datasets.py\", line 12, in <module>\r\n dataset = load_dataset('json', data_files=[\"..\/..\/data\/json.json\"])\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 591, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset, max_retries=download_config.max_retries)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 232, in head_hf_s3\r\n max_retries=max_retries,\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 523, in http_head\r\n max_retries=max_retries,\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 458, in _request_with_retry\r\n raise err\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 454, in _request_with_retry\r\n response = requests.request(verb.upper(), url, **params)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/sessions.py\", line 530, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/sessions.py\", line 643, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/json\/json.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x1181e9940>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\nWhy it want to connect a remote url when I load local datasets, and how can I fix it?","comment_length":69,"text":"Loading local dataset raise requests.exceptions.ConnectTimeout\nLoad local dataset:\r\n```\r\ndataset = load_dataset('json', data_files=[\"..\/..\/data\/json.json\"])\r\ntrain = dataset[\"train\"]\r\nprint(train.features)\r\ntrain1 = train.map(lambda x: {\"labels\": 1})\r\nprint(train1[:2])\r\n```\r\n\r\nbut it raised requests.exceptions.ConnectTimeout:\r\n\r\n```\r\n\/Users\/littlely\/myvirtual\/tf2\/bin\/python3.7 \/Users\/littlely\/projects\/python_projects\/pytorch_learning\/nlp\/dataset\/transformers_datasets.py\r\nTraceback (most recent call last):\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connection.py\", line 160, in _new_conn\r\n (self._dns_host, self.port), self.timeout, **extra_kw\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/util\/connection.py\", line 84, in create_connection\r\n raise err\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/util\/connection.py\", line 74, in create_connection\r\n sock.connect(sa)\r\nsocket.timeout: timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connectionpool.py\", line 677, in urlopen\r\n chunked=chunked,\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connectionpool.py\", line 381, in _make_request\r\n self._validate_conn(conn)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connectionpool.py\", line 978, in _validate_conn\r\n conn.connect()\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connection.py\", line 309, in connect\r\n conn = self._new_conn()\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connection.py\", line 167, in _new_conn\r\n % (self.host, self.timeout),\r\nurllib3.exceptions.ConnectTimeoutError: (<urllib3.connection.HTTPSConnection object at 0x1181e9940>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)')\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/adapters.py\", line 449, in send\r\n timeout=timeout\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/connectionpool.py\", line 727, in urlopen\r\n method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/urllib3\/util\/retry.py\", line 439, in increment\r\n raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\nurllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/json\/json.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x1181e9940>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/littlely\/projects\/python_projects\/pytorch_learning\/nlp\/dataset\/transformers_datasets.py\", line 12, in <module>\r\n dataset = load_dataset('json', data_files=[\"..\/..\/data\/json.json\"])\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 591, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset, max_retries=download_config.max_retries)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 232, in head_hf_s3\r\n max_retries=max_retries,\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 523, in http_head\r\n max_retries=max_retries,\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 458, in _request_with_retry\r\n raise err\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 454, in _request_with_retry\r\n response = requests.request(verb.upper(), url, **params)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/sessions.py\", line 530, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/sessions.py\", line 643, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/Users\/littlely\/myvirtual\/tf2\/lib\/python3.7\/site-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/json\/json.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x1181e9940>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n\r\nProcess finished with exit code 1\r\n\r\n```\r\n\r\nWhy it want to connect a remote url when I load local datasets, and how can I fix it?\nHi ! Thanks for reporting. This was indeed a bug introduced when we moved the `json` dataset loader inside the `datasets` package (before that, the `json` loader was fetched online, as all the other dataset scripts).\r\n\r\nThis should be fixed on master now. Feel free to install `datasets` from source to try it out.\r\nThe fix will be available in the next release of `datasets` in a few days","embeddings":[-0.0876652002,-0.3096545637,-0.0747830272,0.2496923506,0.3151809573,0.0891386345,0.3149231672,0.2885445356,-0.1467543393,0.0568148792,-0.046975974,0.3553402722,0.0624657385,-0.072460711,-0.007479148,-0.357103914,-0.1442816854,0.0838932991,-0.047547698,-0.1543697864,-0.2529872954,0.0257254131,-0.187468946,0.2330839336,-0.252992183,-0.2066891938,-0.0868969336,0.1169291735,0.0257288907,-0.4841195643,0.1179933324,-0.3125517368,0.0555519424,0.528383553,-0.0001154436,0.1222111136,0.0945106819,0.0215760078,-0.132141009,-0.0498595685,-0.0674959645,-0.2368896157,0.2450839132,-0.3710689545,-0.1075161919,-0.3912537396,0.0075889351,-0.2066582143,0.5702975392,0.5039224625,0.145775646,0.5681708455,0.0515227467,0.0785590485,0.2667636871,0.1773377657,-0.206467852,0.2898198366,0.2549915314,-0.4005657732,0.0400110893,-0.0825460404,-0.0017520832,0.0780525357,0.0907534659,0.1272234172,-0.1359151006,-0.348036617,-0.1655737907,0.3539237678,0.3100979924,0.0707029179,-0.2927102745,-0.2270765603,-0.2015623301,-0.029326994,0.1347280443,0.1666531116,-0.3210844398,0.1775446981,-0.3006032407,0.0551207215,-0.4535173178,0.5489299893,0.0865125135,0.5155925155,0.083038494,0.3091308177,-0.0458222441,0.0133977653,-0.0089779766,-0.0843441561,0.0171154663,0.1844487637,-0.1963673532,-0.1544388831,0.1146633103,-0.4060586095,0.1601896584,0.0676575676,0.154643774,-0.0878824592,-0.169943139,0.3116516471,-0.0025950375,0.1281774193,-0.0578197986,0.3318803608,0.3709953129,0.0912054181,-0.1829870492,0.1557956785,0.108564809,-0.1445657611,0.0419754311,-0.0599971972,0.4929864407,0.0511100776,0.087548472,-0.0159910042,-0.586858511,-0.0332805514,0.1956097335,0.3493115306,-0.3422437608,0.0827540681,0.0123645719,-0.0544691868,-0.3163072765,0.0691055581,-0.2425964028,0.2482573241,0.0804514885,-0.059125293,0.1660154909,0.0636304095,0.4791110754,-0.1673445851,-0.0349101722,-0.0415872373,0.5842089653,-0.3233092725,0.1414042711,0.2345120758,-0.2718019187,0.0687621161,0.2806974947,0.1037614346,-0.1080797613,0.020145949,-0.245202899,-0.4628188312,-0.0414956026,0.2253668755,0.0443511568,-0.1087396145,0.0094914436,-0.163241595,0.4137928188,-0.3300018907,0.0524004549,-0.2123021483,-0.2179457247,-0.1649781466,0.2293939739,0.3468136191,-0.2586836219,0.0798254237,0.2683034539,-0.0013026508,0.2266992033,0.4093843997,-0.5121421218,0.4743845761,-0.3170554042,0.1415546685,0.5460799336,-0.4671511054,-0.2384578139,0.5044529438,-0.2510955334,-0.0711534768,0.0435655005,0.2766900659,0.1196675524,0.0741798505,0.4246086478,0.3509928286,-0.2929828167,0.1134737208,-0.2766141593,-0.2069235593,0.0072743786,0.0785306469,0.0622740686,0.0114246299,0.161780104,0.0338042267,0.2731345892,-0.0619287938,0.1070537642,0.3353098929,0.0764932334,-0.0913644955,0.0302906986,-0.0215979572,-0.2053330541,0.2261997312,-0.0307398029,0.2425401062,-0.267367363,-0.0202196706,-0.3861397505,0.0638860911,-0.3079503775,0.0851653814,0.0618030913,0.2712073624,0.1725235134,0.1565878391,0.0425035395,-0.0275297686,-0.257917434,0.1061814725,-0.3778765202,0.2617046833,-0.0599820949,-0.0379351117,-0.091981329,-0.1287303716,0.1428045034,-0.2689329088,-0.2771681845,0.1943447888,0.0190408062,0.1645133048,0.0183444358,0.0385957956,0.0624704137,-0.1664237082,0.0212311801,0.0946419165,0.114807874,-0.1870043427,-0.1732774675,0.2998911738,0.2156782746,0.3294457197,0.1182266697,-0.1258407235,0.2493576407,0.0717816204,-0.0189659242,-0.0102474336,0.0795538276,0.0412090905,0.1136682853,-0.0922211409,-0.2574563026,0.0602617338,0.4460497797,-0.1701764911,0.0450152308,0.0696924403,0.3061302006,-0.109502539,0.1763370931,0.1549542844,0.6341338754,0.2085809857,0.1621567309,0.1511496454,-0.1348972172,-0.2292595953,0.0644129962,-0.0266804118,0.247306034,0.3175880313,0.3057259321,0.024317883,-0.4263645113,-0.2775946259,0.1161995381,0.3775265217,-0.2742401958,0.2351563126,-0.1033193171,-0.1564782113,-0.0477512851,0.0394215733,-0.2301467061,-0.4215005338,0.0583361499,0.2158536613,-0.0731704235,0.2756845355,-0.0093073398,0.020146383,0.350088954,-0.438377589,-0.1064934731,-0.1761183441,-0.364154309,0.0252939407,0.152500093,-0.0963635519,0.4177519679,-0.0499243215,-0.4756876528,-0.1872623712,-0.2943217158,0.0119482232,-0.0652063191,0.247996226,0.1509259939,0.3515704572,-0.0174462646,-0.1918195188,0.422131151,-0.0898680165,-0.1238978207,-0.0267956257,0.0240708888,0.149292931,-0.0307472851,-0.6379082203,-0.4048389196,-0.406676501,0.2181975991,0.0466194525,0.277518779,0.0849544778,0.0358151533,0.3597218692,0.1112167463,0.1276403666,-0.0538620725,-0.3271385431,0.257460475,-0.2653145492,-0.2633663714,-0.014294737,-0.117288053,0.0885663331,0.0501000956,-0.5194664598,-0.4453582466,-0.2787286937,0.2407431006,-0.2776933014,-0.1170527413,0.2960995436,-0.1703329831,-0.1082185134,-0.0064746281,-0.0444768742,0.3415368199,0.2659059167,0.1737858653,0.0614486374,0.7596716285,-0.0332468897,0.5413731337,0.0296803378,0.0352777988,0.4892271757,-0.2829693258,-0.0174787603,-0.0779776052,-0.2684919238,0.0801827088,0.1409265995,-0.2075176537,-0.0933657289,0.1615559161,-0.1225306019,-0.2405091077,-0.2304320633,-0.2888832688,-0.2335736603,0.0097268336,-0.146550715,0.2895323336,-0.0222472865,0.0659928173,-0.1431170851,-0.0516231693,-0.1466680914,0.0693064407,-0.0167993568,-0.1624400467,-0.6813395619,-0.1749103814,-0.5136041045,0.2706353664,0.1567562521,0.5439186096,-0.2641364038,-0.0684387013,0.2611940801,-0.1415598541,0.8109989166,-0.2408433557,0.1052410826,0.0250692181,0.1355768293,-0.3428855538,0.0058179321,-0.0008368833,0.2688808143,0.2481940389,0.5743729472,-0.1280055195,-0.286565125,0.2069715858,0.2740257084,-0.23582232,-0.0762401,-0.2177331597,-0.6896690726,-0.196845293,0.0832699761,-0.0547159947,0.2028003186,0.0924303606,-0.4169505835,0.095230639,0.1664060354,-0.0523580834,0.07994885,-0.110845983,0.0158900302,0.1148235202,-0.2471914887,0.2845290601,0.2450394332,0.3522232175,-0.1490688026,-0.2630354166,0.14364779,0.1238204241,0.3318550587,0.2947627008,-0.0212467164,0.1408848017,0.1599983722,0.3401945829,-0.0828349888,0.387380749,0.532392323,-0.1223366559,-0.2031581253,-0.5039529204,0.1332383603,-0.0471888855,-0.0636864454,0.2192026675,-0.2862211466,-0.2772036195,0.3792078495,0.1395544708,0.9921345115,-0.0113061294,0.2558076978,0.6406010389,-0.0664974079,0.664960742,-0.2646169364,0.2580269575,-0.380402714,0.0429762192,-0.1054763645,-0.2358799875,0.0285094399,0.0944366902,-0.1649485528,0.1491956413,-0.1172832921,0.1959847063,-0.1463426799,0.5414198041,0.0756280869,-0.0598478727,-0.587449193,0.1064096466,-0.0261029843,0.0985835493,-0.0268351175,-0.1081690863,0.0200306233,-0.3731744885,0.1574330181,0.0943821445,-0.2812208235,0.5121895075,-0.1859298497,-0.0758693144,-0.0156010687,0.3076259196,0.0348666534,-0.0000248985,-0.1064446494,0.0362806693,-0.1130685508,0.1308701932,-0.1006316245,-0.009530602,0.1547370404,-0.1163203418,-0.4218951464,0.4718749821,-0.3494643569,-0.3757626414,-0.0709015355,0.1223008558,-0.059524402,-0.4409900308,-0.1029623821,-0.1443653703,-0.2691304088,-0.2003671825,0.0858176649,0.0195271447,-0.31881845,0.3204016984,0.0177516248,-0.3153619766,-0.0215747058,0.4522782564,0.1165136471,0.2216842175,0.7525095344,0.169973731,0.0249590855,-0.3292616606,0.461276263,-0.018030921,-0.2892801762,0.1046873182,0.2482049018,0.2518322766,0.0637088194,0.2772772014,0.2292512506,0.1389219761,0.0036961758,-0.3916372061,-0.4703882635,0.1692252606,0.0765442252,0.2141619921,0.2396151125,0.2498911619,-0.1893876046,0.2043239474,-0.2954698503,0.0561624914,-0.537104845,0.1856480539,-0.0375427864,-0.1155711934,0.3783704042,0.0621669926,0.1401452571,-0.0213010162,-0.3744773567,-0.166437313,-0.0505243689,0.1663369536,0.0718498603,-0.1864739805,-0.4386374056,-0.5062546134,-0.0181478839,0.0965658203,0.0535424091,0.0188472234,-0.0811803266,0.0175819211,0.0141579369,0.1220884621,0.0841595531,0.047267627,-0.1538110524,-0.0157427434,0.2301687747,0.077261366,0.1983973384,-0.1885783672,-0.0674609244,0.0197661463,0.0356259532,-0.1773955822,0.292669028,-0.284787029,-0.2742281854,-0.0748811737,0.2242606133,0.3945913613,-0.196132943,0.0283792987,0.1952040344,0.2080867738,-0.1833198667,-0.292060554,0.4560912848,-0.2019453049,0.1378970891,0.1103229895,-0.2127424031,0.0894703344,-0.2245317847,0.2033671886,0.2341226637,0.0223259181,0.0371878445,0.3005511463,-0.2989983261,-0.082812354,0.1474566162,0.1270948797,0.3212957978,0.0047756373,-0.0055455612,0.043327976,-0.3767881095,0.3542631269,-0.0860257968,-0.4847977459,0.0512528755,-0.1688597798,-0.298890233,-0.1300305724,-0.0237504598,0.4819837511,-0.236739561,-0.1881805211,-0.1995155662,0.3600148857,-0.1028243527,-0.1955419481,-0.229516238,-0.2858203053,0.1354685277,-0.1978503764,0.1609138101,0.0190450847,-0.0650869831,0.089473404,0.0383867063,-0.2019248754,-0.0201455243,0.0612040684,0.0640069693,-0.3245759308,0.1949737519,0.2483701259,-0.0607037582,-0.4171534777,0.0106875682,0.3459572494,0.4239967465,-0.0218450949,0.312499404,-0.0947102532,-0.1583963484,-0.0055409689,0.1134829745,0.2338083386,-0.0364379808,0.3514870107,0.0933529809,-0.1481024921,0.2936782837,-0.0237193424,0.1534293592,-0.221174106,0.509334147,0.012240869,-0.2928759456,0.0004254936,0.0211054422,-0.4169076085,0.100751169,0.4297308326,-0.1344196051,0.3220779896,-0.0988746658,0.0500603542,-0.3950724304,0.6884117126,0.2332841754,0.1067482457,-0.342554599,0.0008541515,-0.3407241106,0.0283508394,-0.4788582027,-0.0240355656,0.0740338713,0.1004003137,0.0750098228,-0.0460601151,-0.1270329803,0.0030353446,0.0218563005,0.1133119166,-0.2876620591,-0.2759694159,-0.004359541,0.2354534864,-0.0187375713,-0.4727282524,0.216354385,-0.2600308359,-0.0307254028,-0.2679136097,-0.0484618172,0.2021927536,-0.0579152964,0.4343681335,0.3496297598,0.4295275509,0.0720883831,-0.0606512651,0.0439471118,-0.4789818525,0.0459217206,-0.004792315,0.1576924324,0.5942357183,-0.1884735674,-0.0180719439,-0.339083463,0.1729625762,-0.1238193512,-0.3846410215,0.323759526,-0.0524617843,0.0261466578,0.0659149811,0.1001644731,0.1330552548,0.0819867998,0.0297979694,-0.2536607683,-0.2163209766,0.7234705687,-0.317905575,-0.1667278856,-0.0437353142,0.1857223809,0.0977983698,0.004968035,-0.1586291194,-0.0803990513,0.6721755862,0.0862089992,-0.3109281063,0.4904264212,0.1578738689,0.0310350414,0.0725700036,-0.0335415863,-0.1904409528,-0.2747128308,-0.0402268954,-0.3100720346]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1817","title":"pyarrow.lib.ArrowInvalid: Column 1 named input_ids expected length 599 but got length 1500","comments":"Hi !\r\nThe error you have is due to the `input_ids` column not having the same number of examples as the other columns.\r\nIndeed you're concatenating the `input_ids` at this line:\r\n\r\nhttps:\/\/github.com\/LuCeHe\/GenericTools\/blob\/431835d8e13ec24dceb5ee4dc4ae58f0e873b091\/KerasTools\/lm_preprocessing.py#L134\r\n\r\nHowever the other columns are kept unchanged, and therefore you end up with an `input_ids` column with 599 elements while the others columns like `attention_mask` have 1500.\r\n\r\nTo fix that you can instead concatenate them all using\r\n```python\r\nconcatenated_examples = {k: sum(examples[k], []) for k in examples.keys()}\r\n```\r\n\r\nAlso you may need to drop the \"text\" column before applying `group_texts` since strings can't be concatenated with lists. You can drop it at the tokenization step:\r\n```python\r\ndset = dset.map(\r\n tokenize_function,\r\n batched=True,\r\n remove_columns=[\"text\"]\r\n)\r\n```","body":"I am trying to preprocess any dataset in this package with GPT-2 tokenizer, so I need to structure the datasets as long sequences of text without padding. I've been following a couple of your tutorials and here you can find the script that is failing right at the end\r\n\r\nhttps:\/\/github.com\/LuCeHe\/GenericTools\/blob\/master\/KerasTools\/lm_preprocessing.py\r\n\r\nIn the last iteration of the last dset.map, it gives the error that I copied in the title. Another issue that I have, if I leave the batch_size set as 1000 in the last .map, I'm afraid it's going to lose most text, so I'm considering setting both writer_batch_size and batch_size to 300 K, but I'm not sure it's the best way to go.\r\n\r\nCan you help me?\r\nThanks!","comment_length":116,"text":"pyarrow.lib.ArrowInvalid: Column 1 named input_ids expected length 599 but got length 1500\nI am trying to preprocess any dataset in this package with GPT-2 tokenizer, so I need to structure the datasets as long sequences of text without padding. I've been following a couple of your tutorials and here you can find the script that is failing right at the end\r\n\r\nhttps:\/\/github.com\/LuCeHe\/GenericTools\/blob\/master\/KerasTools\/lm_preprocessing.py\r\n\r\nIn the last iteration of the last dset.map, it gives the error that I copied in the title. Another issue that I have, if I leave the batch_size set as 1000 in the last .map, I'm afraid it's going to lose most text, so I'm considering setting both writer_batch_size and batch_size to 300 K, but I'm not sure it's the best way to go.\r\n\r\nCan you help me?\r\nThanks!\nHi !\r\nThe error you have is due to the `input_ids` column not having the same number of examples as the other columns.\r\nIndeed you're concatenating the `input_ids` at this line:\r\n\r\nhttps:\/\/github.com\/LuCeHe\/GenericTools\/blob\/431835d8e13ec24dceb5ee4dc4ae58f0e873b091\/KerasTools\/lm_preprocessing.py#L134\r\n\r\nHowever the other columns are kept unchanged, and therefore you end up with an `input_ids` column with 599 elements while the others columns like `attention_mask` have 1500.\r\n\r\nTo fix that you can instead concatenate them all using\r\n```python\r\nconcatenated_examples = {k: sum(examples[k], []) for k in examples.keys()}\r\n```\r\n\r\nAlso you may need to drop the \"text\" column before applying `group_texts` since strings can't be concatenated with lists. You can drop it at the tokenization step:\r\n```python\r\ndset = dset.map(\r\n tokenize_function,\r\n batched=True,\r\n remove_columns=[\"text\"]\r\n)\r\n```","embeddings":[-0.155002296,-0.0703644529,-0.0086344574,0.1708289087,0.2552530169,0.0364475176,0.5036029816,0.5221663713,-0.524661541,0.0104543045,0.3807778358,0.2319069505,0.0748812631,-0.1745078415,0.1261400282,-0.2886483371,0.0141824838,0.2477580458,-0.0138815129,0.1944544166,-0.0676677376,0.3307289779,-0.3295415342,0.2716697752,-0.3865055442,-0.0922516733,-0.1422722787,-0.1698894501,-0.1691986173,-0.5931814313,0.0828088745,-0.228557989,0.0478052273,0.2794103026,-0.0001236288,-0.1168290973,0.4077006876,-0.1066744849,0.0701526999,-0.0044740783,0.0422801189,-0.2361371666,0.0165137015,-0.2340285033,0.0912679061,-0.5475600958,-0.2920790911,0.2607315779,0.2835215032,0.3589550257,0.1257847846,-0.0448946021,0.1790397018,-0.0412862003,0.5702495575,0.0536095574,0.0055336687,0.1160502806,0.1458778381,-0.2013936937,0.0534413718,0.2144300789,-0.2680252492,-0.1032517701,0.0580859445,-0.0039267489,0.0565320142,-0.2023723572,0.2105338126,0.1588908434,0.3037465513,-0.3780476153,-0.0483502857,-0.2628532946,0.1766878068,-0.3986361921,-0.0885518119,-0.060647849,-0.1943920553,0.1883023083,-0.1021537259,0.1745141596,-0.2574053705,0.0938389152,0.0956305861,0.2773377597,0.1587263346,0.2896916568,-0.0706556216,-0.170455128,0.0365101434,0.1347864866,-0.2240155935,0.2108975053,-0.2329891622,-0.03535638,-0.129837513,-0.2215498388,0.4244503975,-0.1432873756,0.2827808559,-0.1606893837,0.7112699151,-0.118645817,-0.0855468288,0.4598975778,0.2372592837,0.2067829072,-0.1270625293,-0.3863641322,0.2913980186,0.1470927745,-0.2602185905,-0.6117146611,0.2757360935,0.1679496765,0.0617795885,0.1802980602,-0.1971882433,0.4800275862,-0.49648121,-0.0251220632,-0.1530764401,0.0553499162,0.0796749368,0.0637729615,-0.0022886391,0.0737530217,0.0643044338,-0.2553144097,-0.102769427,0.2139763832,-0.3910429776,0.0601491854,0.108048901,0.0560679771,0.3017202616,0.0623576455,0.0746015981,0.0635029078,-0.0101668006,-0.0717614666,0.2443722337,0.2771908343,-0.1186136901,0.1828497052,0.2038530856,-0.1445943713,-0.301787138,0.2960030437,-0.327039808,-0.3279041946,-0.1085567176,0.0631789789,0.1062358767,0.1053576842,-0.2185850292,0.150951609,0.6920489073,-0.0988643765,0.1823328137,-0.4258134067,0.1980241686,-0.3123607934,0.1485884637,0.1613928229,-0.567311883,0.1217048019,0.3949619532,0.3733777106,0.3328751624,0.3813624978,-0.0831799135,0.5039740801,-0.090698503,0.7552221417,0.3342810869,-0.088839069,-0.2165766656,0.3936943412,-0.064330481,-0.2323006392,0.0295620374,-0.2319584787,0.2128834426,-0.0096776718,0.1547478586,-0.1016394943,-0.1074724048,-0.0029568034,-0.2887978852,-0.1346039921,0.1084654257,-0.0687080324,-0.1076743975,-0.2563621104,-0.2608839571,-0.410367161,0.1153546721,-0.1635286063,0.3597878218,0.3475106359,0.0196291078,0.026715342,0.3291780949,0.09509442,-0.1781709939,-0.0235901047,0.0442914218,-0.034959048,-0.2201670855,-0.3523425758,-0.2672821581,0.1990024149,0.0381783172,0.1776503325,0.0093461052,-0.2265123725,0.1835489571,-0.1633616388,-0.0336497128,-0.2718361914,0.0508424453,0.0964222252,-0.1822098196,-0.1217020303,-0.173158586,-0.4061906636,0.0997496471,-0.0075107748,0.2952218652,-0.0537830442,0.0143547142,0.3127261996,0.1311428398,-0.3347700834,-0.3954313695,0.2027165443,-0.0828412622,-0.1941531301,-0.1910981089,-0.1059259474,-0.0302319415,-0.0867645591,0.1134169549,0.3114247918,0.0878982991,0.3004992604,-0.2077561468,-0.031966202,0.0630754307,0.0686767474,-0.2348009795,0.010069998,0.0778633356,-0.0475963801,0.1326374859,0.0023546249,-0.3153679371,0.1728616804,0.1770055741,0.015508987,-0.0637862533,0.1986921877,-0.0801163837,-0.0147262001,0.1154872105,-0.1970091164,0.1011901796,0.1659036577,-0.0247780811,-0.0547355078,-0.206338048,-0.1961253583,0.2262758017,0.073609978,0.2035349458,0.5516645312,0.0526557639,0.1273688078,0.0501684658,-0.0934747234,0.2298325896,0.355068922,-0.3621885777,0.1933803111,-0.2526600957,0.0196315181,-0.0059756753,-0.6113991141,0.1719790548,-0.0306187682,0.0457914285,-0.1291826069,-0.1622223854,0.3853451014,0.1819876581,0.1849831939,0.5495913625,-0.167556107,-0.1204284579,-0.2368342578,-0.1843619943,-0.0954642966,0.2132757753,-0.1506299973,-0.0212398209,0.1077955365,-0.1174562797,-0.2242490947,-0.5022531152,0.3119056523,-0.3466986716,-0.1121642813,0.4024557173,0.1223672107,-0.2633654773,-0.2280546427,0.055442825,0.229954347,-0.279808104,0.3484917283,-0.0896362588,-0.2746524811,-0.1506217271,-0.1697364002,0.1745996326,-0.1914544106,0.1772340089,-0.1534639895,0.1468400061,0.0637494847,0.4653074741,0.0369056687,-0.0069357813,0.0424283706,0.0574550591,0.2105319202,0.4343683422,-0.2766295373,-0.2573385835,0.0865589529,-0.1728406549,0.3282194138,0.2733305097,-0.3643850088,0.1617097706,-0.0685935691,0.0936416462,-0.3578870595,0.0590331927,0.4605173469,0.1466504782,0.0051058815,-0.114196457,0.0346330069,0.3736735582,-0.2701492012,0.0551125109,0.0552536957,0.6357032061,-0.0250455793,0.7274159789,0.2814561427,-0.1197825223,0.1013612822,-0.2490906417,-0.1202093959,-0.2398970127,0.1625431031,-0.2318488359,-0.0949842706,0.0369534381,0.3688107729,-0.2972483039,-0.3330015242,0.2210148871,0.1925230473,-0.2142071277,-0.0720821917,0.3196913302,0.0360590927,0.3239009082,-0.0419200025,0.0790789574,-0.1929029673,-0.0333548822,0.0495355465,-0.3764074445,0.0325696096,0.064905487,-0.231773302,-0.4118334353,-0.6226688623,0.3684797883,0.3064644635,0.4693975151,0.1016400158,-0.3109788895,-0.0006427866,-0.2770530283,0.5176249743,-0.1439099312,-0.3334732652,0.3643978834,0.1254676878,-0.4200968146,-0.3367704451,-0.0527213141,0.1455676407,0.3767842948,0.1362761557,-0.5397258401,-0.0318675861,0.3549748361,-0.026121309,-0.0126672434,-0.1469438523,-0.2477856278,-0.3502267599,-0.2020339519,-0.0166321322,0.6145424247,0.3996125162,-0.0125640174,-0.0049707559,-0.0547665656,-0.2995449901,0.0977136195,0.0095575973,0.3902045786,0.0011863619,0.2208579332,-0.2235004902,0.3756492138,0.3819264472,0.1147094965,0.1507869363,-0.2386599481,0.2495274097,-0.273616761,0.4812792838,0.2760556638,-0.2233426273,0.2875681221,-0.444521755,0.2794278264,-0.0792393535,0.1409674585,0.396592021,-0.1731871814,-0.356267482,-0.1305025816,0.2430725396,0.0210243817,-0.0263376161,0.4567234218,0.067395404,-0.3701596558,0.3138682246,0.2621949911,1.1447813511,0.3166950047,0.2496214956,0.3733635843,0.0835130438,0.0062097469,0.0769708529,0.1055417806,-0.5544155836,-0.0051707979,0.0093712239,-0.167592451,-0.0138096958,0.2013982534,-0.3329415321,-0.1281974614,-0.3151694834,0.1709833592,-0.1320949644,0.1428222805,-0.0059347283,-0.0872802883,0.2379440516,-0.0208471566,-0.0131020397,-0.3700397611,-0.1652097255,-0.0691830441,-0.3673365414,-0.3569473624,-0.411229074,-0.0752253383,-0.225060463,0.0391518287,0.1121888384,-0.258179456,0.1797897816,0.0262509994,0.206953004,0.1652821302,-0.1984622329,0.0085508162,0.0443305187,-0.29094401,-0.1584410369,-0.1409945488,0.1220935211,-0.0171994977,0.0525759496,0.185058549,0.042504549,0.0694965646,-0.3031349182,-0.1627691835,0.2704518139,-0.4907824993,-0.092175968,0.0304886848,0.0029679146,-0.1565638781,-0.00674449,0.0492533632,-0.2209215909,0.2321986407,-0.1529700905,-0.2571035922,0.0601378046,0.0723258257,0.0023339693,-0.0311885308,0.607306242,0.3159707189,-0.1589477658,-0.106992051,0.1637086868,0.5035057664,-0.6606781483,0.0749177858,-0.1885086298,0.0594582148,0.0509862117,0.5169564486,0.2717175782,0.3059383631,0.1216227114,-0.0753786042,-0.2998271883,0.2028276175,0.2588067055,0.3612178564,-0.2054408193,0.0983741358,-0.162079677,0.1159508154,-0.1893984526,-0.0878791437,-0.0577512346,0.1876956671,-0.1030353606,-0.0182805229,-0.2642341852,0.0514644422,-0.0315391645,0.3885488212,-0.0102441115,-0.102779232,-0.2301065326,0.1527895331,0.0681776628,-0.0534089841,0.2742183208,-0.1448958367,-0.2985362411,-0.1450137049,0.0136728724,-0.0535814278,-0.0210790541,-0.0242110062,0.4298163354,0.0646908358,-0.0873849317,0.0276945569,0.002244011,0.145347774,0.1146919206,0.2281081825,0.0958378017,0.0262142979,0.1361534297,0.335101068,0.0527841188,-0.3552295864,0.1339013875,-0.4393780529,-0.1296576709,-0.0578085743,0.2475418448,0.3009122908,-0.3606338203,-0.1328799278,0.4503161013,0.1535473317,-0.286534518,-0.1164170429,-0.182986781,0.0641661212,-0.1379845589,0.200839147,0.1094480008,-0.0606062524,-0.2202546149,0.173743397,0.3542883098,0.1231874377,-0.3958587945,0.319421649,-0.1859204769,0.1239870638,0.3163275421,0.2035693526,0.2621989548,0.0297564585,-0.118768461,0.2451472878,0.2678400874,0.0912137777,-0.239334166,-0.2031654119,-0.0724440068,0.2762990594,0.088381879,0.0028185274,0.1790753752,0.3353696465,0.4524991512,-0.2289245874,0.027890414,0.3590196669,-0.1949386746,-0.0660153478,-0.7122880816,0.0824287459,-0.2180038691,-0.1749486029,-0.1023459584,0.0737308785,0.4343536496,0.0964008868,-0.0904745683,-0.1987380534,0.070728153,-0.0068393275,0.094507277,-0.323153764,0.4119297862,0.1004951298,-0.2849639058,-0.1137816682,0.3278914988,0.6728407145,0.4071210623,-0.1261104941,-0.3004499674,-0.2880660295,0.1028206497,0.1032603905,0.3540619314,0.3359224796,0.1313992739,0.2158366442,-0.0917939097,-0.0349796489,-0.0927517936,0.3317200541,-0.0684672743,0.2488201708,-0.0547331758,-0.1573168039,-0.4111024737,-0.0005268923,0.2015111744,-0.5054785609,0.2205101848,0.5505584478,0.1804078966,0.1515893191,-0.0997031927,0.0322801024,-0.0781506747,0.4440885782,0.3804039061,0.1010725498,-0.3919534683,-0.3986079395,-0.2410693318,0.1371295303,-0.4373247623,-0.1311998218,0.2652823329,0.1691052169,0.2406357378,0.019580042,0.1167047545,0.1667091399,-0.3454831243,0.5820026994,-0.2765940428,0.2108579278,0.148213774,0.0865802765,0.0272796769,-0.1982696652,0.426626116,-0.0151001485,-0.0107729807,-0.36559394,0.0269141831,0.1797393262,0.2164812833,0.2508496046,0.0789010152,0.2836628854,-0.0113754692,0.1057761908,-0.3002665639,-0.1518848687,-0.1774403602,0.3001046777,0.0960744023,0.0366607606,-0.1852332056,-0.2117047608,0.0342736356,-0.1169182956,-0.0780774727,-0.0607341118,-0.3419034183,0.1300679147,-0.2000257224,0.1868883818,-0.0354950689,0.1137792096,0.1058961675,0.2066833824,-0.3272295892,-0.6539340615,0.706984818,-0.4853319824,-0.4971392751,-0.0232494641,-0.006308917,0.3170477748,0.1148778051,-0.1392882466,-0.1747901142,0.2745012939,0.1119710803,-0.5378230214,0.3033047616,-0.0773033723,0.0450051688,-0.1753377616,0.2336113155,0.0831816718,-0.224484235,-0.4705750048,-0.6530278325]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1817","title":"pyarrow.lib.ArrowInvalid: Column 1 named input_ids expected length 599 but got length 1500","comments":"You saved my life.","body":"I am trying to preprocess any dataset in this package with GPT-2 tokenizer, so I need to structure the datasets as long sequences of text without padding. I've been following a couple of your tutorials and here you can find the script that is failing right at the end\r\n\r\nhttps:\/\/github.com\/LuCeHe\/GenericTools\/blob\/master\/KerasTools\/lm_preprocessing.py\r\n\r\nIn the last iteration of the last dset.map, it gives the error that I copied in the title. Another issue that I have, if I leave the batch_size set as 1000 in the last .map, I'm afraid it's going to lose most text, so I'm considering setting both writer_batch_size and batch_size to 300 K, but I'm not sure it's the best way to go.\r\n\r\nCan you help me?\r\nThanks!","comment_length":4,"text":"pyarrow.lib.ArrowInvalid: Column 1 named input_ids expected length 599 but got length 1500\nI am trying to preprocess any dataset in this package with GPT-2 tokenizer, so I need to structure the datasets as long sequences of text without padding. I've been following a couple of your tutorials and here you can find the script that is failing right at the end\r\n\r\nhttps:\/\/github.com\/LuCeHe\/GenericTools\/blob\/master\/KerasTools\/lm_preprocessing.py\r\n\r\nIn the last iteration of the last dset.map, it gives the error that I copied in the title. Another issue that I have, if I leave the batch_size set as 1000 in the last .map, I'm afraid it's going to lose most text, so I'm considering setting both writer_batch_size and batch_size to 300 K, but I'm not sure it's the best way to go.\r\n\r\nCan you help me?\r\nThanks!\nYou saved my life.","embeddings":[-0.1766743213,-0.1751580089,-0.0120626669,0.1091196015,0.3175965548,-0.006574675,0.4156205952,0.5791722536,-0.5055668354,0.0253004041,0.3509283066,0.1375004947,-0.0068367352,-0.1639076322,0.1266014278,-0.2992775738,0.0000711574,0.2100487649,-0.0985218659,0.2215590626,-0.0684654117,0.2611641288,-0.2544045448,0.2255976051,-0.3887857795,-0.0702716857,0.045475293,-0.1750828773,-0.2231029123,-0.686550796,0.1786231697,-0.252622813,-0.010065157,0.2528516352,-0.0001239286,-0.0422393754,0.4644704759,-0.0696046427,0.2090206891,0.0568971708,0.0232262984,-0.1618719846,0.074364908,-0.1935875267,0.00843878,-0.4933508039,-0.184875682,0.2745061219,0.3267763853,0.4529261589,0.1404154748,-0.0589657128,0.2371018678,-0.0632984489,0.6484586596,0.0222249869,-0.0017233441,0.1741036773,0.1621655226,-0.1080565155,-0.0385692753,0.2175615728,-0.2777198255,-0.1078605056,0.170630157,-0.1376401335,0.10995996,-0.2303932607,0.2175481766,0.2033149749,0.3571465015,-0.4360974431,-0.0213354193,-0.1335427314,0.1713027805,-0.3742949069,0.0205788668,-0.0473337732,-0.1990511417,0.2269994915,-0.1311343759,0.0312837623,-0.30449453,-0.0227091014,0.0977513492,0.2685827613,0.1502717286,0.2751024663,-0.0701526701,-0.0253173001,0.0821840465,0.2286211848,-0.1803701073,0.1720747948,-0.1954784691,-0.0633334517,-0.1049854755,-0.1209373549,0.3615576327,-0.1458522677,0.2639909387,-0.2336592823,0.6155974865,-0.1481486559,-0.0895676613,0.3550347388,0.2017287761,0.125750497,-0.0340002589,-0.3484611809,0.2136089653,0.135710761,-0.1356599778,-0.6533436775,0.3343852758,0.1916346848,0.1611168087,0.113746576,-0.1628893167,0.4585907757,-0.3949616849,0.006878424,-0.1347430646,0.0225452464,-0.026287321,0.0204855911,0.039051082,0.0553622097,-0.0230538137,-0.2747732997,-0.1197377592,0.25764516,-0.3635525703,0.0632380694,0.1693677306,0.0481600761,0.2586097419,-0.0222069323,0.0433288477,0.0088821063,-0.0319899097,-0.0844936892,0.1758182645,0.3211537004,-0.1968870163,0.093084313,0.1517216712,-0.0546062775,-0.3045717776,0.3608318567,-0.3830424249,-0.3270250559,-0.1236717403,0.0799890608,0.1369202435,0.1131557301,-0.2152214348,0.1073907018,0.6951628923,-0.2075892538,0.3110642135,-0.400796771,0.1735845208,-0.2786607742,0.1428898871,0.1039522588,-0.6154482365,0.1598875821,0.4791704714,0.2394615561,0.3859573603,0.369420439,0.0174723696,0.4900176823,-0.1017690897,0.6525345445,0.3150834441,-0.022846017,-0.2728411853,0.3525094092,-0.1050090343,-0.3631492853,0.0930587724,-0.1810545474,0.081278652,-0.0785440058,0.0826260746,-0.1187715605,-0.0770839304,-0.0343150534,-0.3666255474,-0.0871622413,0.0110869668,-0.0105367666,0.1010249183,-0.3291171789,-0.270427525,-0.3407223225,0.1471741349,-0.0077798758,0.3491672277,0.362059325,0.0038992888,-0.0411714725,0.311542213,-0.0564554632,-0.1470843256,-0.0479197577,-0.1021170691,-0.0111335209,-0.1966545135,-0.3338863254,-0.2228071243,0.214344874,0.0382019095,0.1738995314,0.0158853326,-0.2962860465,0.1601005346,-0.0276129767,0.0461437777,-0.3856557608,0.1273955256,0.0997276232,-0.1895538419,-0.1537516117,-0.1524100304,-0.3062798977,0.0521395616,-0.0019632916,0.1766707599,-0.0328448527,-0.0986349732,0.3591448367,0.090877451,-0.3078687489,-0.3498347104,0.1600912362,-0.1044498086,-0.4239540696,-0.2294103801,-0.0192958619,-0.0963710621,-0.0911491141,0.0373688117,0.2193321139,0.1287062168,0.2460327893,-0.2285935879,-0.1131187677,0.0191854704,0.0613525994,-0.2525315285,-0.0162887964,0.0573094152,-0.0769714639,0.1756968796,0.0557650626,-0.2835202813,0.0550352186,0.3329233527,-0.0211460385,0.0182632189,0.2018211335,-0.0097708153,-0.1104385853,0.085677065,-0.1878948808,0.0294344313,0.1711420566,-0.0557353981,-0.0718313083,-0.2079151273,-0.2234809995,0.21100685,-0.0006057002,0.3183003664,0.5299075246,0.0362303369,0.089062348,0.0513924323,-0.0819866285,0.1981581599,0.295923084,-0.2757933736,0.0981611386,-0.2761799991,-0.0003911599,-0.0843114108,-0.5018407106,0.0027251306,-0.0427186452,0.0703704432,-0.1671755612,-0.1303551346,0.4221889675,0.154376179,0.1639728248,0.5866324306,-0.19403404,-0.0850139633,-0.1995365322,-0.1129627451,-0.1089943349,0.2538188994,-0.2593076527,0.0669335276,0.1089839116,-0.1032804102,-0.1656167805,-0.471409142,0.3249374032,-0.3482506871,-0.147471875,0.4590153694,0.1146083176,-0.1316245645,0.0185283087,0.0845613256,0.1422238797,-0.2417942882,0.3187662363,-0.1716936678,-0.146170795,-0.1257910132,-0.1097240299,0.1849137545,-0.2022107542,0.1860544831,-0.2456447035,0.0458741635,0.1083343327,0.4638873637,0.0860138237,0.0178753752,0.0530858301,0.160767436,0.1790510416,0.3574560583,-0.2136683464,-0.184447974,0.0538714118,-0.0406699814,0.4061959982,0.2510784566,-0.4656319618,0.1633923203,-0.0841551051,0.0955613256,-0.3534997702,-0.0143504087,0.5115365386,0.140424639,-0.0295867529,-0.1409849226,0.0477325022,0.3430030942,-0.3678680062,0.057352487,0.0913917646,0.6893717051,-0.0604525916,0.7663541436,0.1285618097,-0.0304673631,0.1303575486,-0.2400027215,-0.1230951399,-0.3448050022,0.2416426241,-0.1796971262,-0.0781993791,-0.0171409007,0.3628409505,-0.2957040966,-0.3631355166,0.2251010537,0.0794157535,-0.1365576088,-0.0154466396,0.305051446,0.0376657993,0.2909702957,-0.0020399438,0.0775254071,-0.1576067209,-0.0791732967,0.0495776795,-0.3869661391,0.1057380512,0.0690137744,-0.1746228933,-0.421915859,-0.7352221012,0.3086659312,0.2800712287,0.3501635194,0.0359808095,-0.2450672537,0.0252134595,-0.260853678,0.4319431484,-0.0965122059,-0.1690684259,0.26377213,0.1750130951,-0.4439266026,-0.3080821931,-0.0262090806,0.280043155,0.332143873,0.1043924391,-0.4149194658,-0.0998589993,0.3628444672,-0.0661389977,-0.0267815385,-0.222957477,-0.2786859274,-0.3430936933,-0.1022591516,0.0690444559,0.6395099163,0.3173075616,-0.0057329996,-0.0279907752,-0.0215573832,-0.2640000582,0.0340355672,-0.1300528944,0.3444886506,-0.0740944147,0.1382949501,-0.2799984813,0.3604094386,0.3942457139,0.1816157252,0.0074219946,-0.1781119853,0.2269819677,-0.2421316653,0.5309952497,0.245687604,-0.260455966,0.238722235,-0.4028730392,0.3304236531,-0.0950586349,0.1878207773,0.3671247065,-0.1964538246,-0.4292609692,-0.123164773,0.326467067,0.0012799879,0.0075331409,0.456709981,0.1186578795,-0.4093414247,0.2840290666,0.3444000483,1.1037712097,0.273982197,0.3022606671,0.336812675,0.1839375347,0.0216514505,0.0854045823,0.1029959098,-0.5447431207,0.0261712465,-0.0074068545,-0.1002875641,-0.1005044505,0.1632960886,-0.3306119144,-0.0991431847,-0.1530728638,0.1741378009,-0.1226548478,0.1467044801,0.0195928458,-0.0646133274,0.2493568361,-0.0144018689,-0.0935970321,-0.3683240414,-0.1454389244,-0.0847295746,-0.2429815233,-0.2991004884,-0.3915582299,-0.0425271019,-0.3287634254,0.0147407781,0.1646643728,-0.1932461262,0.1983449757,0.0184516702,0.1213814542,0.3068561554,-0.3064801991,0.0321071632,0.0121905254,-0.2889996469,-0.1927521825,-0.0433327258,0.2234179527,-0.0851651281,0.0275739785,0.0425231196,-0.0654881746,0.0295835249,-0.2015702277,-0.2654810846,0.2433282882,-0.3833796084,-0.1726064682,-0.0976399779,0.0096989302,-0.1404843479,0.0095970072,0.0940764919,-0.1716021746,0.1099964753,-0.0868285149,-0.3297216296,0.0448900089,0.1107680351,-0.0338874906,0.0484085493,0.6296880245,0.3642184734,-0.103451699,-0.1754539609,0.143587634,0.607298851,-0.7667056918,0.0846407712,-0.1786691397,0.1148707569,-0.0018973954,0.4304591119,0.2286710739,0.241853565,0.1239862069,-0.0782374665,-0.3680878282,0.1974223852,0.2146090865,0.3931159675,-0.158664614,0.1341106147,-0.0365311913,0.1480956525,-0.1858034879,-0.0850804597,-0.0228041224,0.0827589706,-0.1048411131,-0.0102916732,-0.2202164829,0.110490635,-0.043548774,0.4414943457,-0.0819691271,-0.1062375233,-0.2078041732,0.1511459798,0.0197527725,-0.0183471981,0.2784864604,-0.2026415616,-0.3422279954,-0.0947529972,-0.0099083614,-0.0809777379,0.0533934794,0.0021808944,0.5833783746,0.1048225239,-0.0683768168,0.0161317103,0.0805827156,0.1256943941,0.0153927216,0.1978304684,0.0924047083,-0.0045503327,0.1276919544,0.3129022419,0.0772389621,-0.3021063209,0.1261066645,-0.4823364913,-0.1253781468,0.032714162,0.3143964708,0.2665525377,-0.3227415085,-0.1101516932,0.3455741405,0.152803272,-0.3106027246,-0.0527269095,-0.1118292883,0.0976012424,-0.1491708159,0.1245023087,0.0770155787,0.0302475858,-0.1623914093,0.243406564,0.2786101401,0.0166601315,-0.4866552651,0.2706183493,-0.1748635471,0.0920444652,0.3806656003,0.2204508632,0.3078964651,0.1325226277,0.0383822657,0.3428896368,0.215152204,0.090660654,-0.2492884994,-0.2394655496,-0.1770052165,0.281388402,0.0076776063,-0.053860642,0.127350226,0.3706337214,0.3963010609,-0.3153933883,0.032334514,0.4081681967,-0.1591241062,-0.1311394423,-0.6648288965,0.1229418144,-0.1763771772,-0.120293878,-0.1152255014,0.2453670353,0.3790635765,0.0875376314,-0.0881877914,-0.0923942402,0.0407584235,-0.0416295826,0.1176341772,-0.2801448703,0.490478754,0.1260679811,-0.2528735697,-0.1072693169,0.2888457477,0.6067458391,0.4359171391,-0.259075135,-0.316588521,-0.3236300349,0.0679125264,0.0511036813,0.3576410413,0.436863184,0.2641768157,0.2935290635,-0.1085991561,-0.0332184769,0.0575735793,0.2795381248,-0.0011349244,0.2670922875,-0.0508919246,-0.0784318671,-0.4263802469,-0.1448076218,0.2445245087,-0.5848464966,0.1394991875,0.6105624437,0.1111005247,0.1068616062,-0.1699482352,0.026216954,-0.018095769,0.5772604346,0.3929363787,0.2115440965,-0.3787790239,-0.365350157,-0.2455708086,0.1419493407,-0.3958915472,-0.0505119301,0.2763909996,0.0901922286,0.2381721288,-0.053793259,0.1813725829,0.1541706175,-0.2735191584,0.5109969378,-0.1615760773,0.1695412695,0.0681485981,0.0773710161,0.0465104133,-0.2270092964,0.3800117373,-0.0237980243,0.0021178876,-0.4039328992,0.0215680152,0.2113674879,0.2436238676,0.3078386486,0.1019394547,0.2654489279,-0.0610558279,0.0848264769,-0.3022169471,-0.1119553745,-0.2448924333,0.1990012378,0.0880130231,0.1634327918,-0.1998850107,-0.287722379,0.0200816523,0.011641358,-0.1629550606,-0.0194665175,-0.2809843421,-0.0098479846,-0.2064357847,0.0488041528,0.1343484223,0.0907203406,0.1281673163,0.2075342983,-0.4085485637,-0.6642430425,0.7121554017,-0.4121994972,-0.4753579199,-0.0339298733,-0.0253294911,0.2803119123,0.0952996612,-0.2498158216,-0.1265973598,0.2671216428,0.1794459224,-0.6065152884,0.3348789215,-0.0649587139,0.1311957389,-0.1390260905,0.3535680771,0.0549277775,-0.2231693715,-0.5452306271,-0.6800867319]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1811","title":"Unable to add Multi-label Datasets","comments":"Thanks for adding this dataset! As far as I know `supervised_keys` is mostly a holdover from TFDS, but isn't really used, so feel free to drop it (@lhoestq or @thomwolf correct me if I'm wrong). It definitely shouldn't be blocking :) ","body":"I am trying to add [CIFAR-100](https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html) dataset. The dataset contains two labels per image - `fine label` and `coarse label`. Using just one label in supervised keys as \r\n`supervised_keys=(\"img\", \"fine_label\")` raises no issue. But trying `supervised_keys=(\"img\", \"fine_label\",\"coarse_label\")` leads to this error : \r\n\r\n```python\r\nTraceback (most recent call last):\r\n File \"test_script.py\", line 2, in <module>\r\n d = load_dataset('.\/datasets\/cifar100')\r\n File \"~\/datasets\/src\/datasets\/load.py\", line 668, in load_dataset\r\n **config_kwargs,\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 896, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 247, in __init__\r\n info.update(self._info())\r\n File \"~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cifar100\/61d2489b2d4a4abc34201432541b7380984ec714e290817d9a1ee318e4b74e0f\/cifar100.py\", line 79, in _info\r\n citation=_CITATION,\r\n File \"<string>\", line 19, in __init__\r\n File \"~\/datasets\/src\/datasets\/info.py\", line 136, in __post_init__\r\n self.supervised_keys = SupervisedKeysData(*self.supervised_keys)\r\nTypeError: __init__() takes from 1 to 3 positional arguments but 4 were given\r\n```\r\nIs there a way I can fix this?\r\n\r\nAlso, what does adding `supervised_keys` do? Is it necessary? How would I specify `supervised_keys` for a multi-input, multi-label dataset?\r\n\r\nThanks,\r\nGunjan","comment_length":41,"text":"Unable to add Multi-label Datasets\nI am trying to add [CIFAR-100](https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html) dataset. The dataset contains two labels per image - `fine label` and `coarse label`. Using just one label in supervised keys as \r\n`supervised_keys=(\"img\", \"fine_label\")` raises no issue. But trying `supervised_keys=(\"img\", \"fine_label\",\"coarse_label\")` leads to this error : \r\n\r\n```python\r\nTraceback (most recent call last):\r\n File \"test_script.py\", line 2, in <module>\r\n d = load_dataset('.\/datasets\/cifar100')\r\n File \"~\/datasets\/src\/datasets\/load.py\", line 668, in load_dataset\r\n **config_kwargs,\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 896, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 247, in __init__\r\n info.update(self._info())\r\n File \"~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cifar100\/61d2489b2d4a4abc34201432541b7380984ec714e290817d9a1ee318e4b74e0f\/cifar100.py\", line 79, in _info\r\n citation=_CITATION,\r\n File \"<string>\", line 19, in __init__\r\n File \"~\/datasets\/src\/datasets\/info.py\", line 136, in __post_init__\r\n self.supervised_keys = SupervisedKeysData(*self.supervised_keys)\r\nTypeError: __init__() takes from 1 to 3 positional arguments but 4 were given\r\n```\r\nIs there a way I can fix this?\r\n\r\nAlso, what does adding `supervised_keys` do? Is it necessary? How would I specify `supervised_keys` for a multi-input, multi-label dataset?\r\n\r\nThanks,\r\nGunjan\nThanks for adding this dataset! As far as I know `supervised_keys` is mostly a holdover from TFDS, but isn't really used, so feel free to drop it (@lhoestq or @thomwolf correct me if I'm wrong). It definitely shouldn't be blocking :) ","embeddings":[0.0216598418,-0.2712222338,0.0181464944,0.1329571754,0.3712174594,0.3166499734,0.6164801717,-0.4762451947,0.1994253397,0.2660482228,-0.2177258879,0.1535530984,-0.2759885192,0.3479327261,0.1080327854,-0.1204760596,0.1780963987,-0.217401877,0.0256008487,-0.033268027,-0.2215413302,0.0074294899,-0.2346253097,-0.2840359509,-0.6864250302,-0.2522574961,-0.348762989,-0.1600370556,-0.2611855567,-0.6842919588,0.1742349118,0.3276168406,-0.1799799353,0.4868170321,-0.0001189698,0.053939525,0.0060154423,-0.0270139854,-0.1315755099,-0.0884083286,0.0338955149,-0.0997653976,-0.0230811313,-0.1876841038,-0.0205910355,-0.3216594756,-0.0478544869,-0.2678296268,0.2622791529,0.1708982736,0.1447003931,-0.1211824119,0.3119575679,-0.1777848601,-0.0340049379,0.2013507783,-0.1293978244,0.1572745889,0.1325554699,-0.1618613303,0.4810110629,0.2527094781,-0.0045298049,-0.1883502454,0.4128499031,-0.0407814495,-0.1479687542,-0.5755074024,0.1524925977,0.3097066581,0.6528835297,-0.0376973525,-0.6485276222,-0.144102335,0.3986090422,-0.1604739726,-0.1403426677,0.010431719,-0.0038597537,-0.2375007272,-0.3692445457,-0.0498816818,-0.2118247896,0.2764362693,0.333170861,0.3576138914,0.0522686765,0.35098809,0.2702969611,-0.0573835894,-0.0252628159,-0.1056556776,0.0451553315,0.2221504897,-0.2898245156,0.1726470292,-0.0475980975,0.1391517818,0.1451030672,-0.1027215347,0.0525578931,-0.3232679665,-0.211719349,0.1196939275,0.1027332097,-0.0196695887,-0.0022725062,0.4025920331,-0.0366642326,-0.0274585523,0.1102099642,-0.2352029085,-0.0683441162,-0.1296106279,0.3619584143,0.1870019734,-0.1424213946,0.1166465431,0.0462243594,-0.15703246,-0.061152637,-0.0322760865,0.3567192256,0.309689492,0.2346779108,0.0138949612,0.1058847159,0.4291902483,0.1078813076,0.0548968427,-0.1495378613,0.35435009,-0.1401570141,0.0218004473,0.2909840345,0.0014130031,0.1271663308,0.3206980526,0.2556933761,-0.1487238407,0.0474536568,-0.2202002108,0.1937618554,0.3569915891,-0.0255860426,0.2145913243,-0.0456783921,-0.3797814846,-0.1100404412,0.04526655,-0.3311825991,-0.0257803425,-0.052225735,0.0699773654,0.1953081638,-0.0917193443,-0.4771164656,0.2014684528,0.1235159412,0.0339203253,-0.0702257752,-0.2045563459,-0.2669656575,-0.2525112927,0.2751828134,0.401172787,-0.2575544119,-0.209166497,-0.1150707379,0.16939044,-0.2034942359,0.2128368467,-0.2620731592,0.4487377703,-0.0940307975,0.0628931224,0.042674195,-0.8922803998,-0.0666826069,-0.0147576723,0.0866596475,-0.128503859,0.1614068151,0.3256177604,0.1087990627,0.0289789066,0.1426028311,0.1137600467,-0.1924496591,-0.0482096337,0.1788373441,0.1565864831,-0.1048276648,0.3070019782,0.2102843821,-0.2103247494,0.1564217657,-0.1566660702,-0.0160315279,-0.1635690033,0.12073908,0.3795161843,0.2092370689,-0.0499869846,0.0313238017,0.0380041413,-0.4350197017,0.4339157343,0.2955799401,0.3049954176,-0.1049287096,-0.1757050157,-0.0116091687,0.0613495074,-0.4852899611,-0.1146177873,0.0656736344,0.2741549015,-0.0914811268,-0.0599610843,0.0205955021,0.1393007636,0.4397166073,0.1631942391,0.0898363367,0.0351097733,-0.1254134476,0.031797003,-0.0982277617,0.0179512184,0.4089220166,-0.0632736459,0.0348855965,0.4561132491,0.1855800599,-0.3028663397,-0.1581857651,0.0631943867,0.4092926979,-0.2760539651,-0.2940944433,-0.2007402331,0.2320427746,-0.3411729634,-0.155014053,0.0775262788,-0.3345649242,0.2867106795,-0.0063107894,0.1957471073,0.1347314268,0.1028221473,-0.1135910451,-0.2067493051,0.0379192606,-0.1822332293,-0.0228359047,-0.2354649007,0.0413877219,0.0655713156,0.0258155391,0.0400971286,-0.0610706173,-0.0535848029,0.0286562536,0.2431066483,0.1560307741,0.5032753944,0.2892154753,0.2665799856,-0.1650920212,-0.2013620734,-0.2896959782,0.0578938462,0.1002955586,-0.0812715143,0.1857181787,0.1985814273,0.1882681102,0.0334806368,-0.2237827629,-0.4800541997,0.3684946597,0.1311239153,-0.2911791503,0.0136999218,-0.3054610789,-0.209730491,-0.0478190146,-0.2719694078,-0.0329390801,-0.273193121,0.100284718,0.2303827852,0.1110883728,0.2187402695,0.0505732298,0.2245625407,-0.2533396184,0.0736529753,-0.0290039685,-0.1045127735,-0.002999285,-0.0265577007,0.3638055325,-0.3310750723,0.5024328232,0.0892854407,-0.1429113895,-0.204916358,-0.4361688793,0.0277465582,-0.1707712859,0.2560727,0.0792692378,-0.0453793332,-0.120553337,-0.2528287768,0.3231553733,0.1433298141,-0.067182377,-0.02225996,-0.2092654854,-0.3181765974,-0.3291820884,-0.0019011601,-0.2415915877,-0.0313346125,0.1783394068,0.0039089699,-0.0620984249,0.6229845881,0.1421978921,-0.3269357383,0.1114385575,0.1087919623,-0.0204633344,-0.3791366518,0.230792135,-0.3280832469,-0.245816052,0.1112751365,-0.2078543007,0.4561773539,-0.0236298926,-0.074241221,-0.0516700149,0.06004408,0.0250432417,-0.0242213272,0.2214048356,0.1233270913,0.1223821193,0.0012233047,-0.2397067994,-0.4332716763,0.3612817526,-0.230379194,0.2302520275,-0.3453046679,0.2405825853,0.0744778737,0.2453038096,0.0080904132,0.0415512472,0.4380138814,-0.2188450545,0.3824874759,-0.1967352331,-0.5095447898,0.2663565278,0.0068039051,0.0746747702,-0.0564309396,0.0025118168,-0.1534558535,-0.13697429,0.4596606493,-0.2982488871,-0.0714152157,-0.1395442188,-0.0614005066,0.1793229282,-0.1286798716,-0.2386004925,-0.4252116382,-0.0979251638,-0.3168427944,0.0837059692,0.030056743,-0.0836135522,-0.5794285536,-0.279966265,-0.2095399946,0.0962752774,0.2015597671,0.1998217553,-0.250143379,0.3073728979,0.0423248895,0.0750111043,0.6299776435,-0.5078380108,-0.207209006,0.4064024687,0.0823346972,-0.4420149326,-0.3596889377,-0.0637626573,0.2939634025,0.0598797761,0.3375253379,-0.3172761202,0.1752974391,0.1442773193,0.0290731676,-0.1157465428,-0.2602501512,-0.1650700122,-0.3463683724,-0.2542931139,0.0387581997,0.2610555291,0.2220253199,0.0756111965,0.05578788,0.0451995172,0.0471867733,0.1067099497,-0.0073393807,0.0891595408,0.5999578238,0.3521910608,0.0442494377,-0.223894015,0.5654521585,0.499717921,-0.3498901725,-0.2776920497,-0.0696075335,-0.2492358238,0.2417492121,-0.0740425736,0.2450897247,0.2074229419,0.2566080987,0.0326639861,0.0169424899,0.1901456118,0.5320228338,-0.0338639915,-0.5876461864,-0.3630560637,0.1882240474,0.0840374306,-0.0925073922,0.3905680776,0.035589207,-0.310058862,0.2069021463,0.152016446,0.7129688263,-0.5040864944,0.2006383538,-0.1281060278,-0.051407069,0.6691420674,0.0259150546,0.1060547009,-0.1650911123,-0.2414849997,0.107521005,0.0631438717,0.0530461222,0.3399398625,-0.1476923823,0.1040081382,-0.1996166557,0.1442142576,-0.2286387533,0.0942784548,0.2413909137,-0.3566122055,0.315527916,0.1148123816,-0.0379253142,-0.2212827057,0.0147405667,-0.0131456228,-0.0485539548,-0.2762643993,-0.4560342729,-0.1082490757,-0.3264353573,0.351482749,-0.0545667596,-0.082947515,0.0922916085,0.3642334342,0.6377292871,-0.0870183855,-0.0121600647,0.1931255907,-0.007270975,-0.1853410453,-0.4156341255,-0.2583948374,0.1583053172,-0.0967743844,-0.3491543531,0.0630329549,0.0483213104,-0.2924515903,-0.1737174392,0.1112410799,-0.1338459104,0.0843074471,0.0194527209,0.2658884227,0.2253153622,-0.3193776011,0.0304356217,0.0522682555,-0.0975166485,0.334641397,-0.1579447091,0.0292434208,-0.0027380958,0.3327620029,0.0285511427,0.2265633047,0.417263478,-0.0195385832,0.0441499874,-0.0707633495,-0.0785979331,0.2807226181,-0.3569045067,-0.0215917844,0.4509515166,0.0682092831,-0.0082824528,0.4070376158,0.0060866834,0.1032975987,0.0879296362,-0.1227887198,-0.4140322208,0.0571798645,-0.4007037878,0.2787562013,-0.0407934636,0.4472046494,0.1755109727,0.4107339382,-0.2010179609,-0.0609041527,-0.2943552136,-0.0041191275,0.4779884815,0.0559544377,-0.1362131238,0.0865932256,0.10840258,-0.1853463352,0.1712531149,-0.031114975,-0.1749444455,0.1246730909,0.2514070272,0.1749788076,0.0139208622,-0.4309507608,-0.1583899558,-0.1143202707,0.2133870125,0.2629471719,-0.0949871689,0.3122138679,-0.4964765906,0.0038022301,0.0238737185,0.1440283507,0.0420851298,-0.0607595481,0.3293673396,0.0790245682,0.127211228,-0.1499333233,-0.3488268852,0.0482538603,-0.4432691634,-0.1288997084,0.0841796026,-0.3632187247,0.4464629889,0.2567118406,0.2399899364,-0.1217627004,-0.2593407631,0.4296592474,0.0848011822,0.073372066,-0.2197782993,-0.4196273088,-0.0084309662,0.0238845516,0.1413663328,0.2538610399,0.2197544277,-0.0843564644,0.0170371383,-0.1205665097,0.0592124984,0.0718337819,0.1195045114,0.1281981468,-0.0726764426,0.0837647766,0.2819838822,0.2315364778,0.0600666255,0.194866702,-0.0722384155,0.3028277159,-0.1764678955,-0.0757141411,0.240550667,-0.3558896482,0.4976215661,-0.1813222617,-0.0065754028,-0.044587303,-0.0960169062,0.2008923292,-0.2110212594,-0.3887046874,-0.2711659968,0.3218339682,-0.1305858493,-0.1066607237,-0.182562381,-0.105645217,-0.0563145503,-0.0584490336,0.0690638646,-0.1914356351,0.3167058825,-0.1098625362,-0.1567377895,-0.1262349039,-0.4305464029,0.461884141,0.2834605277,-0.2831432223,0.2036371827,0.3541097939,-0.4410456121,0.1897425056,0.3322366178,0.6006937027,0.2195312977,0.2928979397,0.0838777795,-0.1760641783,-0.1078898758,0.1742133349,0.0496440083,0.4417835474,-0.0686274543,0.1369452924,0.0067783715,-0.0832349434,0.2980151176,0.1812421679,0.3393843174,-0.0167961605,-0.3750321269,-0.0341072753,-0.1780612916,-0.0763006732,0.3583650291,-0.3215570152,-0.0696086511,0.2845407724,-0.0804059282,0.0143419076,-0.3408436179,0.0184786692,0.1767684072,0.4923861325,0.1872951239,0.1653291136,-0.2372946292,-0.1239616424,-0.5642417669,-0.2042975426,-0.3196799457,-0.2337514162,-0.0577933416,-0.0994747952,0.3641085327,-0.2035193145,0.6210821867,-0.2856467068,0.384831816,0.3869747221,-0.2538062036,0.0657707751,0.201900661,0.0772044435,0.0396992639,-0.1581979841,-0.0872712657,0.0368853025,0.0007560396,-0.1040414795,0.1626861542,-0.2012913674,-0.2207449675,0.1988331974,0.0763103962,0.8045886159,-0.0558607131,0.064656727,-0.1496199071,-0.0560613871,-0.1025550291,0.0681211501,0.4282157123,0.586725831,-0.2910673916,-0.0661168844,-0.0673986897,-0.0161997508,-0.0980607271,-0.0475920923,-0.2880606055,0.3186068237,-0.0191521253,0.0055988389,0.262262702,0.2214763165,-0.0565426797,0.0405983664,-0.0261393134,-0.3694877028,0.5187733173,-0.1914700568,-0.1348089725,-0.1738880426,0.0139728244,0.0507746078,0.104873769,-0.3421268761,0.1284422427,0.2554286122,0.0750762448,-0.7180352211,0.4141148031,-0.2150640041,-0.0544044301,0.0822755918,0.0061151502,-0.1176430583,0.0204468705,-0.1353336126,-0.1060189977]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1811","title":"Unable to add Multi-label Datasets","comments":"I can confirm that it comes from TFDS and is not used at the moment.","body":"I am trying to add [CIFAR-100](https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html) dataset. The dataset contains two labels per image - `fine label` and `coarse label`. Using just one label in supervised keys as \r\n`supervised_keys=(\"img\", \"fine_label\")` raises no issue. But trying `supervised_keys=(\"img\", \"fine_label\",\"coarse_label\")` leads to this error : \r\n\r\n```python\r\nTraceback (most recent call last):\r\n File \"test_script.py\", line 2, in <module>\r\n d = load_dataset('.\/datasets\/cifar100')\r\n File \"~\/datasets\/src\/datasets\/load.py\", line 668, in load_dataset\r\n **config_kwargs,\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 896, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 247, in __init__\r\n info.update(self._info())\r\n File \"~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cifar100\/61d2489b2d4a4abc34201432541b7380984ec714e290817d9a1ee318e4b74e0f\/cifar100.py\", line 79, in _info\r\n citation=_CITATION,\r\n File \"<string>\", line 19, in __init__\r\n File \"~\/datasets\/src\/datasets\/info.py\", line 136, in __post_init__\r\n self.supervised_keys = SupervisedKeysData(*self.supervised_keys)\r\nTypeError: __init__() takes from 1 to 3 positional arguments but 4 were given\r\n```\r\nIs there a way I can fix this?\r\n\r\nAlso, what does adding `supervised_keys` do? Is it necessary? How would I specify `supervised_keys` for a multi-input, multi-label dataset?\r\n\r\nThanks,\r\nGunjan","comment_length":15,"text":"Unable to add Multi-label Datasets\nI am trying to add [CIFAR-100](https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html) dataset. The dataset contains two labels per image - `fine label` and `coarse label`. Using just one label in supervised keys as \r\n`supervised_keys=(\"img\", \"fine_label\")` raises no issue. But trying `supervised_keys=(\"img\", \"fine_label\",\"coarse_label\")` leads to this error : \r\n\r\n```python\r\nTraceback (most recent call last):\r\n File \"test_script.py\", line 2, in <module>\r\n d = load_dataset('.\/datasets\/cifar100')\r\n File \"~\/datasets\/src\/datasets\/load.py\", line 668, in load_dataset\r\n **config_kwargs,\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 896, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 247, in __init__\r\n info.update(self._info())\r\n File \"~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cifar100\/61d2489b2d4a4abc34201432541b7380984ec714e290817d9a1ee318e4b74e0f\/cifar100.py\", line 79, in _info\r\n citation=_CITATION,\r\n File \"<string>\", line 19, in __init__\r\n File \"~\/datasets\/src\/datasets\/info.py\", line 136, in __post_init__\r\n self.supervised_keys = SupervisedKeysData(*self.supervised_keys)\r\nTypeError: __init__() takes from 1 to 3 positional arguments but 4 were given\r\n```\r\nIs there a way I can fix this?\r\n\r\nAlso, what does adding `supervised_keys` do? Is it necessary? How would I specify `supervised_keys` for a multi-input, multi-label dataset?\r\n\r\nThanks,\r\nGunjan\nI can confirm that it comes from TFDS and is not used at the moment.","embeddings":[0.0216598418,-0.2712222338,0.0181464944,0.1329571754,0.3712174594,0.3166499734,0.6164801717,-0.4762451947,0.1994253397,0.2660482228,-0.2177258879,0.1535530984,-0.2759885192,0.3479327261,0.1080327854,-0.1204760596,0.1780963987,-0.217401877,0.0256008487,-0.033268027,-0.2215413302,0.0074294899,-0.2346253097,-0.2840359509,-0.6864250302,-0.2522574961,-0.348762989,-0.1600370556,-0.2611855567,-0.6842919588,0.1742349118,0.3276168406,-0.1799799353,0.4868170321,-0.0001189698,0.053939525,0.0060154423,-0.0270139854,-0.1315755099,-0.0884083286,0.0338955149,-0.0997653976,-0.0230811313,-0.1876841038,-0.0205910355,-0.3216594756,-0.0478544869,-0.2678296268,0.2622791529,0.1708982736,0.1447003931,-0.1211824119,0.3119575679,-0.1777848601,-0.0340049379,0.2013507783,-0.1293978244,0.1572745889,0.1325554699,-0.1618613303,0.4810110629,0.2527094781,-0.0045298049,-0.1883502454,0.4128499031,-0.0407814495,-0.1479687542,-0.5755074024,0.1524925977,0.3097066581,0.6528835297,-0.0376973525,-0.6485276222,-0.144102335,0.3986090422,-0.1604739726,-0.1403426677,0.010431719,-0.0038597537,-0.2375007272,-0.3692445457,-0.0498816818,-0.2118247896,0.2764362693,0.333170861,0.3576138914,0.0522686765,0.35098809,0.2702969611,-0.0573835894,-0.0252628159,-0.1056556776,0.0451553315,0.2221504897,-0.2898245156,0.1726470292,-0.0475980975,0.1391517818,0.1451030672,-0.1027215347,0.0525578931,-0.3232679665,-0.211719349,0.1196939275,0.1027332097,-0.0196695887,-0.0022725062,0.4025920331,-0.0366642326,-0.0274585523,0.1102099642,-0.2352029085,-0.0683441162,-0.1296106279,0.3619584143,0.1870019734,-0.1424213946,0.1166465431,0.0462243594,-0.15703246,-0.061152637,-0.0322760865,0.3567192256,0.309689492,0.2346779108,0.0138949612,0.1058847159,0.4291902483,0.1078813076,0.0548968427,-0.1495378613,0.35435009,-0.1401570141,0.0218004473,0.2909840345,0.0014130031,0.1271663308,0.3206980526,0.2556933761,-0.1487238407,0.0474536568,-0.2202002108,0.1937618554,0.3569915891,-0.0255860426,0.2145913243,-0.0456783921,-0.3797814846,-0.1100404412,0.04526655,-0.3311825991,-0.0257803425,-0.052225735,0.0699773654,0.1953081638,-0.0917193443,-0.4771164656,0.2014684528,0.1235159412,0.0339203253,-0.0702257752,-0.2045563459,-0.2669656575,-0.2525112927,0.2751828134,0.401172787,-0.2575544119,-0.209166497,-0.1150707379,0.16939044,-0.2034942359,0.2128368467,-0.2620731592,0.4487377703,-0.0940307975,0.0628931224,0.042674195,-0.8922803998,-0.0666826069,-0.0147576723,0.0866596475,-0.128503859,0.1614068151,0.3256177604,0.1087990627,0.0289789066,0.1426028311,0.1137600467,-0.1924496591,-0.0482096337,0.1788373441,0.1565864831,-0.1048276648,0.3070019782,0.2102843821,-0.2103247494,0.1564217657,-0.1566660702,-0.0160315279,-0.1635690033,0.12073908,0.3795161843,0.2092370689,-0.0499869846,0.0313238017,0.0380041413,-0.4350197017,0.4339157343,0.2955799401,0.3049954176,-0.1049287096,-0.1757050157,-0.0116091687,0.0613495074,-0.4852899611,-0.1146177873,0.0656736344,0.2741549015,-0.0914811268,-0.0599610843,0.0205955021,0.1393007636,0.4397166073,0.1631942391,0.0898363367,0.0351097733,-0.1254134476,0.031797003,-0.0982277617,0.0179512184,0.4089220166,-0.0632736459,0.0348855965,0.4561132491,0.1855800599,-0.3028663397,-0.1581857651,0.0631943867,0.4092926979,-0.2760539651,-0.2940944433,-0.2007402331,0.2320427746,-0.3411729634,-0.155014053,0.0775262788,-0.3345649242,0.2867106795,-0.0063107894,0.1957471073,0.1347314268,0.1028221473,-0.1135910451,-0.2067493051,0.0379192606,-0.1822332293,-0.0228359047,-0.2354649007,0.0413877219,0.0655713156,0.0258155391,0.0400971286,-0.0610706173,-0.0535848029,0.0286562536,0.2431066483,0.1560307741,0.5032753944,0.2892154753,0.2665799856,-0.1650920212,-0.2013620734,-0.2896959782,0.0578938462,0.1002955586,-0.0812715143,0.1857181787,0.1985814273,0.1882681102,0.0334806368,-0.2237827629,-0.4800541997,0.3684946597,0.1311239153,-0.2911791503,0.0136999218,-0.3054610789,-0.209730491,-0.0478190146,-0.2719694078,-0.0329390801,-0.273193121,0.100284718,0.2303827852,0.1110883728,0.2187402695,0.0505732298,0.2245625407,-0.2533396184,0.0736529753,-0.0290039685,-0.1045127735,-0.002999285,-0.0265577007,0.3638055325,-0.3310750723,0.5024328232,0.0892854407,-0.1429113895,-0.204916358,-0.4361688793,0.0277465582,-0.1707712859,0.2560727,0.0792692378,-0.0453793332,-0.120553337,-0.2528287768,0.3231553733,0.1433298141,-0.067182377,-0.02225996,-0.2092654854,-0.3181765974,-0.3291820884,-0.0019011601,-0.2415915877,-0.0313346125,0.1783394068,0.0039089699,-0.0620984249,0.6229845881,0.1421978921,-0.3269357383,0.1114385575,0.1087919623,-0.0204633344,-0.3791366518,0.230792135,-0.3280832469,-0.245816052,0.1112751365,-0.2078543007,0.4561773539,-0.0236298926,-0.074241221,-0.0516700149,0.06004408,0.0250432417,-0.0242213272,0.2214048356,0.1233270913,0.1223821193,0.0012233047,-0.2397067994,-0.4332716763,0.3612817526,-0.230379194,0.2302520275,-0.3453046679,0.2405825853,0.0744778737,0.2453038096,0.0080904132,0.0415512472,0.4380138814,-0.2188450545,0.3824874759,-0.1967352331,-0.5095447898,0.2663565278,0.0068039051,0.0746747702,-0.0564309396,0.0025118168,-0.1534558535,-0.13697429,0.4596606493,-0.2982488871,-0.0714152157,-0.1395442188,-0.0614005066,0.1793229282,-0.1286798716,-0.2386004925,-0.4252116382,-0.0979251638,-0.3168427944,0.0837059692,0.030056743,-0.0836135522,-0.5794285536,-0.279966265,-0.2095399946,0.0962752774,0.2015597671,0.1998217553,-0.250143379,0.3073728979,0.0423248895,0.0750111043,0.6299776435,-0.5078380108,-0.207209006,0.4064024687,0.0823346972,-0.4420149326,-0.3596889377,-0.0637626573,0.2939634025,0.0598797761,0.3375253379,-0.3172761202,0.1752974391,0.1442773193,0.0290731676,-0.1157465428,-0.2602501512,-0.1650700122,-0.3463683724,-0.2542931139,0.0387581997,0.2610555291,0.2220253199,0.0756111965,0.05578788,0.0451995172,0.0471867733,0.1067099497,-0.0073393807,0.0891595408,0.5999578238,0.3521910608,0.0442494377,-0.223894015,0.5654521585,0.499717921,-0.3498901725,-0.2776920497,-0.0696075335,-0.2492358238,0.2417492121,-0.0740425736,0.2450897247,0.2074229419,0.2566080987,0.0326639861,0.0169424899,0.1901456118,0.5320228338,-0.0338639915,-0.5876461864,-0.3630560637,0.1882240474,0.0840374306,-0.0925073922,0.3905680776,0.035589207,-0.310058862,0.2069021463,0.152016446,0.7129688263,-0.5040864944,0.2006383538,-0.1281060278,-0.051407069,0.6691420674,0.0259150546,0.1060547009,-0.1650911123,-0.2414849997,0.107521005,0.0631438717,0.0530461222,0.3399398625,-0.1476923823,0.1040081382,-0.1996166557,0.1442142576,-0.2286387533,0.0942784548,0.2413909137,-0.3566122055,0.315527916,0.1148123816,-0.0379253142,-0.2212827057,0.0147405667,-0.0131456228,-0.0485539548,-0.2762643993,-0.4560342729,-0.1082490757,-0.3264353573,0.351482749,-0.0545667596,-0.082947515,0.0922916085,0.3642334342,0.6377292871,-0.0870183855,-0.0121600647,0.1931255907,-0.007270975,-0.1853410453,-0.4156341255,-0.2583948374,0.1583053172,-0.0967743844,-0.3491543531,0.0630329549,0.0483213104,-0.2924515903,-0.1737174392,0.1112410799,-0.1338459104,0.0843074471,0.0194527209,0.2658884227,0.2253153622,-0.3193776011,0.0304356217,0.0522682555,-0.0975166485,0.334641397,-0.1579447091,0.0292434208,-0.0027380958,0.3327620029,0.0285511427,0.2265633047,0.417263478,-0.0195385832,0.0441499874,-0.0707633495,-0.0785979331,0.2807226181,-0.3569045067,-0.0215917844,0.4509515166,0.0682092831,-0.0082824528,0.4070376158,0.0060866834,0.1032975987,0.0879296362,-0.1227887198,-0.4140322208,0.0571798645,-0.4007037878,0.2787562013,-0.0407934636,0.4472046494,0.1755109727,0.4107339382,-0.2010179609,-0.0609041527,-0.2943552136,-0.0041191275,0.4779884815,0.0559544377,-0.1362131238,0.0865932256,0.10840258,-0.1853463352,0.1712531149,-0.031114975,-0.1749444455,0.1246730909,0.2514070272,0.1749788076,0.0139208622,-0.4309507608,-0.1583899558,-0.1143202707,0.2133870125,0.2629471719,-0.0949871689,0.3122138679,-0.4964765906,0.0038022301,0.0238737185,0.1440283507,0.0420851298,-0.0607595481,0.3293673396,0.0790245682,0.127211228,-0.1499333233,-0.3488268852,0.0482538603,-0.4432691634,-0.1288997084,0.0841796026,-0.3632187247,0.4464629889,0.2567118406,0.2399899364,-0.1217627004,-0.2593407631,0.4296592474,0.0848011822,0.073372066,-0.2197782993,-0.4196273088,-0.0084309662,0.0238845516,0.1413663328,0.2538610399,0.2197544277,-0.0843564644,0.0170371383,-0.1205665097,0.0592124984,0.0718337819,0.1195045114,0.1281981468,-0.0726764426,0.0837647766,0.2819838822,0.2315364778,0.0600666255,0.194866702,-0.0722384155,0.3028277159,-0.1764678955,-0.0757141411,0.240550667,-0.3558896482,0.4976215661,-0.1813222617,-0.0065754028,-0.044587303,-0.0960169062,0.2008923292,-0.2110212594,-0.3887046874,-0.2711659968,0.3218339682,-0.1305858493,-0.1066607237,-0.182562381,-0.105645217,-0.0563145503,-0.0584490336,0.0690638646,-0.1914356351,0.3167058825,-0.1098625362,-0.1567377895,-0.1262349039,-0.4305464029,0.461884141,0.2834605277,-0.2831432223,0.2036371827,0.3541097939,-0.4410456121,0.1897425056,0.3322366178,0.6006937027,0.2195312977,0.2928979397,0.0838777795,-0.1760641783,-0.1078898758,0.1742133349,0.0496440083,0.4417835474,-0.0686274543,0.1369452924,0.0067783715,-0.0832349434,0.2980151176,0.1812421679,0.3393843174,-0.0167961605,-0.3750321269,-0.0341072753,-0.1780612916,-0.0763006732,0.3583650291,-0.3215570152,-0.0696086511,0.2845407724,-0.0804059282,0.0143419076,-0.3408436179,0.0184786692,0.1767684072,0.4923861325,0.1872951239,0.1653291136,-0.2372946292,-0.1239616424,-0.5642417669,-0.2042975426,-0.3196799457,-0.2337514162,-0.0577933416,-0.0994747952,0.3641085327,-0.2035193145,0.6210821867,-0.2856467068,0.384831816,0.3869747221,-0.2538062036,0.0657707751,0.201900661,0.0772044435,0.0396992639,-0.1581979841,-0.0872712657,0.0368853025,0.0007560396,-0.1040414795,0.1626861542,-0.2012913674,-0.2207449675,0.1988331974,0.0763103962,0.8045886159,-0.0558607131,0.064656727,-0.1496199071,-0.0560613871,-0.1025550291,0.0681211501,0.4282157123,0.586725831,-0.2910673916,-0.0661168844,-0.0673986897,-0.0161997508,-0.0980607271,-0.0475920923,-0.2880606055,0.3186068237,-0.0191521253,0.0055988389,0.262262702,0.2214763165,-0.0565426797,0.0405983664,-0.0261393134,-0.3694877028,0.5187733173,-0.1914700568,-0.1348089725,-0.1738880426,0.0139728244,0.0507746078,0.104873769,-0.3421268761,0.1284422427,0.2554286122,0.0750762448,-0.7180352211,0.4141148031,-0.2150640041,-0.0544044301,0.0822755918,0.0061151502,-0.1176430583,0.0204468705,-0.1353336126,-0.1060189977]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1811","title":"Unable to add Multi-label Datasets","comments":"Thanks @yjernite @lhoestq \r\n\r\nThe template for new dataset makes it slightly confusing. I suppose the comment suggesting its update can be removed.","body":"I am trying to add [CIFAR-100](https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html) dataset. The dataset contains two labels per image - `fine label` and `coarse label`. Using just one label in supervised keys as \r\n`supervised_keys=(\"img\", \"fine_label\")` raises no issue. But trying `supervised_keys=(\"img\", \"fine_label\",\"coarse_label\")` leads to this error : \r\n\r\n```python\r\nTraceback (most recent call last):\r\n File \"test_script.py\", line 2, in <module>\r\n d = load_dataset('.\/datasets\/cifar100')\r\n File \"~\/datasets\/src\/datasets\/load.py\", line 668, in load_dataset\r\n **config_kwargs,\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 896, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 247, in __init__\r\n info.update(self._info())\r\n File \"~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cifar100\/61d2489b2d4a4abc34201432541b7380984ec714e290817d9a1ee318e4b74e0f\/cifar100.py\", line 79, in _info\r\n citation=_CITATION,\r\n File \"<string>\", line 19, in __init__\r\n File \"~\/datasets\/src\/datasets\/info.py\", line 136, in __post_init__\r\n self.supervised_keys = SupervisedKeysData(*self.supervised_keys)\r\nTypeError: __init__() takes from 1 to 3 positional arguments but 4 were given\r\n```\r\nIs there a way I can fix this?\r\n\r\nAlso, what does adding `supervised_keys` do? Is it necessary? How would I specify `supervised_keys` for a multi-input, multi-label dataset?\r\n\r\nThanks,\r\nGunjan","comment_length":22,"text":"Unable to add Multi-label Datasets\nI am trying to add [CIFAR-100](https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html) dataset. The dataset contains two labels per image - `fine label` and `coarse label`. Using just one label in supervised keys as \r\n`supervised_keys=(\"img\", \"fine_label\")` raises no issue. But trying `supervised_keys=(\"img\", \"fine_label\",\"coarse_label\")` leads to this error : \r\n\r\n```python\r\nTraceback (most recent call last):\r\n File \"test_script.py\", line 2, in <module>\r\n d = load_dataset('.\/datasets\/cifar100')\r\n File \"~\/datasets\/src\/datasets\/load.py\", line 668, in load_dataset\r\n **config_kwargs,\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 896, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 247, in __init__\r\n info.update(self._info())\r\n File \"~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cifar100\/61d2489b2d4a4abc34201432541b7380984ec714e290817d9a1ee318e4b74e0f\/cifar100.py\", line 79, in _info\r\n citation=_CITATION,\r\n File \"<string>\", line 19, in __init__\r\n File \"~\/datasets\/src\/datasets\/info.py\", line 136, in __post_init__\r\n self.supervised_keys = SupervisedKeysData(*self.supervised_keys)\r\nTypeError: __init__() takes from 1 to 3 positional arguments but 4 were given\r\n```\r\nIs there a way I can fix this?\r\n\r\nAlso, what does adding `supervised_keys` do? Is it necessary? How would I specify `supervised_keys` for a multi-input, multi-label dataset?\r\n\r\nThanks,\r\nGunjan\nThanks @yjernite @lhoestq \r\n\r\nThe template for new dataset makes it slightly confusing. I suppose the comment suggesting its update can be removed.","embeddings":[0.0216598418,-0.2712222338,0.0181464944,0.1329571754,0.3712174594,0.3166499734,0.6164801717,-0.4762451947,0.1994253397,0.2660482228,-0.2177258879,0.1535530984,-0.2759885192,0.3479327261,0.1080327854,-0.1204760596,0.1780963987,-0.217401877,0.0256008487,-0.033268027,-0.2215413302,0.0074294899,-0.2346253097,-0.2840359509,-0.6864250302,-0.2522574961,-0.348762989,-0.1600370556,-0.2611855567,-0.6842919588,0.1742349118,0.3276168406,-0.1799799353,0.4868170321,-0.0001189698,0.053939525,0.0060154423,-0.0270139854,-0.1315755099,-0.0884083286,0.0338955149,-0.0997653976,-0.0230811313,-0.1876841038,-0.0205910355,-0.3216594756,-0.0478544869,-0.2678296268,0.2622791529,0.1708982736,0.1447003931,-0.1211824119,0.3119575679,-0.1777848601,-0.0340049379,0.2013507783,-0.1293978244,0.1572745889,0.1325554699,-0.1618613303,0.4810110629,0.2527094781,-0.0045298049,-0.1883502454,0.4128499031,-0.0407814495,-0.1479687542,-0.5755074024,0.1524925977,0.3097066581,0.6528835297,-0.0376973525,-0.6485276222,-0.144102335,0.3986090422,-0.1604739726,-0.1403426677,0.010431719,-0.0038597537,-0.2375007272,-0.3692445457,-0.0498816818,-0.2118247896,0.2764362693,0.333170861,0.3576138914,0.0522686765,0.35098809,0.2702969611,-0.0573835894,-0.0252628159,-0.1056556776,0.0451553315,0.2221504897,-0.2898245156,0.1726470292,-0.0475980975,0.1391517818,0.1451030672,-0.1027215347,0.0525578931,-0.3232679665,-0.211719349,0.1196939275,0.1027332097,-0.0196695887,-0.0022725062,0.4025920331,-0.0366642326,-0.0274585523,0.1102099642,-0.2352029085,-0.0683441162,-0.1296106279,0.3619584143,0.1870019734,-0.1424213946,0.1166465431,0.0462243594,-0.15703246,-0.061152637,-0.0322760865,0.3567192256,0.309689492,0.2346779108,0.0138949612,0.1058847159,0.4291902483,0.1078813076,0.0548968427,-0.1495378613,0.35435009,-0.1401570141,0.0218004473,0.2909840345,0.0014130031,0.1271663308,0.3206980526,0.2556933761,-0.1487238407,0.0474536568,-0.2202002108,0.1937618554,0.3569915891,-0.0255860426,0.2145913243,-0.0456783921,-0.3797814846,-0.1100404412,0.04526655,-0.3311825991,-0.0257803425,-0.052225735,0.0699773654,0.1953081638,-0.0917193443,-0.4771164656,0.2014684528,0.1235159412,0.0339203253,-0.0702257752,-0.2045563459,-0.2669656575,-0.2525112927,0.2751828134,0.401172787,-0.2575544119,-0.209166497,-0.1150707379,0.16939044,-0.2034942359,0.2128368467,-0.2620731592,0.4487377703,-0.0940307975,0.0628931224,0.042674195,-0.8922803998,-0.0666826069,-0.0147576723,0.0866596475,-0.128503859,0.1614068151,0.3256177604,0.1087990627,0.0289789066,0.1426028311,0.1137600467,-0.1924496591,-0.0482096337,0.1788373441,0.1565864831,-0.1048276648,0.3070019782,0.2102843821,-0.2103247494,0.1564217657,-0.1566660702,-0.0160315279,-0.1635690033,0.12073908,0.3795161843,0.2092370689,-0.0499869846,0.0313238017,0.0380041413,-0.4350197017,0.4339157343,0.2955799401,0.3049954176,-0.1049287096,-0.1757050157,-0.0116091687,0.0613495074,-0.4852899611,-0.1146177873,0.0656736344,0.2741549015,-0.0914811268,-0.0599610843,0.0205955021,0.1393007636,0.4397166073,0.1631942391,0.0898363367,0.0351097733,-0.1254134476,0.031797003,-0.0982277617,0.0179512184,0.4089220166,-0.0632736459,0.0348855965,0.4561132491,0.1855800599,-0.3028663397,-0.1581857651,0.0631943867,0.4092926979,-0.2760539651,-0.2940944433,-0.2007402331,0.2320427746,-0.3411729634,-0.155014053,0.0775262788,-0.3345649242,0.2867106795,-0.0063107894,0.1957471073,0.1347314268,0.1028221473,-0.1135910451,-0.2067493051,0.0379192606,-0.1822332293,-0.0228359047,-0.2354649007,0.0413877219,0.0655713156,0.0258155391,0.0400971286,-0.0610706173,-0.0535848029,0.0286562536,0.2431066483,0.1560307741,0.5032753944,0.2892154753,0.2665799856,-0.1650920212,-0.2013620734,-0.2896959782,0.0578938462,0.1002955586,-0.0812715143,0.1857181787,0.1985814273,0.1882681102,0.0334806368,-0.2237827629,-0.4800541997,0.3684946597,0.1311239153,-0.2911791503,0.0136999218,-0.3054610789,-0.209730491,-0.0478190146,-0.2719694078,-0.0329390801,-0.273193121,0.100284718,0.2303827852,0.1110883728,0.2187402695,0.0505732298,0.2245625407,-0.2533396184,0.0736529753,-0.0290039685,-0.1045127735,-0.002999285,-0.0265577007,0.3638055325,-0.3310750723,0.5024328232,0.0892854407,-0.1429113895,-0.204916358,-0.4361688793,0.0277465582,-0.1707712859,0.2560727,0.0792692378,-0.0453793332,-0.120553337,-0.2528287768,0.3231553733,0.1433298141,-0.067182377,-0.02225996,-0.2092654854,-0.3181765974,-0.3291820884,-0.0019011601,-0.2415915877,-0.0313346125,0.1783394068,0.0039089699,-0.0620984249,0.6229845881,0.1421978921,-0.3269357383,0.1114385575,0.1087919623,-0.0204633344,-0.3791366518,0.230792135,-0.3280832469,-0.245816052,0.1112751365,-0.2078543007,0.4561773539,-0.0236298926,-0.074241221,-0.0516700149,0.06004408,0.0250432417,-0.0242213272,0.2214048356,0.1233270913,0.1223821193,0.0012233047,-0.2397067994,-0.4332716763,0.3612817526,-0.230379194,0.2302520275,-0.3453046679,0.2405825853,0.0744778737,0.2453038096,0.0080904132,0.0415512472,0.4380138814,-0.2188450545,0.3824874759,-0.1967352331,-0.5095447898,0.2663565278,0.0068039051,0.0746747702,-0.0564309396,0.0025118168,-0.1534558535,-0.13697429,0.4596606493,-0.2982488871,-0.0714152157,-0.1395442188,-0.0614005066,0.1793229282,-0.1286798716,-0.2386004925,-0.4252116382,-0.0979251638,-0.3168427944,0.0837059692,0.030056743,-0.0836135522,-0.5794285536,-0.279966265,-0.2095399946,0.0962752774,0.2015597671,0.1998217553,-0.250143379,0.3073728979,0.0423248895,0.0750111043,0.6299776435,-0.5078380108,-0.207209006,0.4064024687,0.0823346972,-0.4420149326,-0.3596889377,-0.0637626573,0.2939634025,0.0598797761,0.3375253379,-0.3172761202,0.1752974391,0.1442773193,0.0290731676,-0.1157465428,-0.2602501512,-0.1650700122,-0.3463683724,-0.2542931139,0.0387581997,0.2610555291,0.2220253199,0.0756111965,0.05578788,0.0451995172,0.0471867733,0.1067099497,-0.0073393807,0.0891595408,0.5999578238,0.3521910608,0.0442494377,-0.223894015,0.5654521585,0.499717921,-0.3498901725,-0.2776920497,-0.0696075335,-0.2492358238,0.2417492121,-0.0740425736,0.2450897247,0.2074229419,0.2566080987,0.0326639861,0.0169424899,0.1901456118,0.5320228338,-0.0338639915,-0.5876461864,-0.3630560637,0.1882240474,0.0840374306,-0.0925073922,0.3905680776,0.035589207,-0.310058862,0.2069021463,0.152016446,0.7129688263,-0.5040864944,0.2006383538,-0.1281060278,-0.051407069,0.6691420674,0.0259150546,0.1060547009,-0.1650911123,-0.2414849997,0.107521005,0.0631438717,0.0530461222,0.3399398625,-0.1476923823,0.1040081382,-0.1996166557,0.1442142576,-0.2286387533,0.0942784548,0.2413909137,-0.3566122055,0.315527916,0.1148123816,-0.0379253142,-0.2212827057,0.0147405667,-0.0131456228,-0.0485539548,-0.2762643993,-0.4560342729,-0.1082490757,-0.3264353573,0.351482749,-0.0545667596,-0.082947515,0.0922916085,0.3642334342,0.6377292871,-0.0870183855,-0.0121600647,0.1931255907,-0.007270975,-0.1853410453,-0.4156341255,-0.2583948374,0.1583053172,-0.0967743844,-0.3491543531,0.0630329549,0.0483213104,-0.2924515903,-0.1737174392,0.1112410799,-0.1338459104,0.0843074471,0.0194527209,0.2658884227,0.2253153622,-0.3193776011,0.0304356217,0.0522682555,-0.0975166485,0.334641397,-0.1579447091,0.0292434208,-0.0027380958,0.3327620029,0.0285511427,0.2265633047,0.417263478,-0.0195385832,0.0441499874,-0.0707633495,-0.0785979331,0.2807226181,-0.3569045067,-0.0215917844,0.4509515166,0.0682092831,-0.0082824528,0.4070376158,0.0060866834,0.1032975987,0.0879296362,-0.1227887198,-0.4140322208,0.0571798645,-0.4007037878,0.2787562013,-0.0407934636,0.4472046494,0.1755109727,0.4107339382,-0.2010179609,-0.0609041527,-0.2943552136,-0.0041191275,0.4779884815,0.0559544377,-0.1362131238,0.0865932256,0.10840258,-0.1853463352,0.1712531149,-0.031114975,-0.1749444455,0.1246730909,0.2514070272,0.1749788076,0.0139208622,-0.4309507608,-0.1583899558,-0.1143202707,0.2133870125,0.2629471719,-0.0949871689,0.3122138679,-0.4964765906,0.0038022301,0.0238737185,0.1440283507,0.0420851298,-0.0607595481,0.3293673396,0.0790245682,0.127211228,-0.1499333233,-0.3488268852,0.0482538603,-0.4432691634,-0.1288997084,0.0841796026,-0.3632187247,0.4464629889,0.2567118406,0.2399899364,-0.1217627004,-0.2593407631,0.4296592474,0.0848011822,0.073372066,-0.2197782993,-0.4196273088,-0.0084309662,0.0238845516,0.1413663328,0.2538610399,0.2197544277,-0.0843564644,0.0170371383,-0.1205665097,0.0592124984,0.0718337819,0.1195045114,0.1281981468,-0.0726764426,0.0837647766,0.2819838822,0.2315364778,0.0600666255,0.194866702,-0.0722384155,0.3028277159,-0.1764678955,-0.0757141411,0.240550667,-0.3558896482,0.4976215661,-0.1813222617,-0.0065754028,-0.044587303,-0.0960169062,0.2008923292,-0.2110212594,-0.3887046874,-0.2711659968,0.3218339682,-0.1305858493,-0.1066607237,-0.182562381,-0.105645217,-0.0563145503,-0.0584490336,0.0690638646,-0.1914356351,0.3167058825,-0.1098625362,-0.1567377895,-0.1262349039,-0.4305464029,0.461884141,0.2834605277,-0.2831432223,0.2036371827,0.3541097939,-0.4410456121,0.1897425056,0.3322366178,0.6006937027,0.2195312977,0.2928979397,0.0838777795,-0.1760641783,-0.1078898758,0.1742133349,0.0496440083,0.4417835474,-0.0686274543,0.1369452924,0.0067783715,-0.0832349434,0.2980151176,0.1812421679,0.3393843174,-0.0167961605,-0.3750321269,-0.0341072753,-0.1780612916,-0.0763006732,0.3583650291,-0.3215570152,-0.0696086511,0.2845407724,-0.0804059282,0.0143419076,-0.3408436179,0.0184786692,0.1767684072,0.4923861325,0.1872951239,0.1653291136,-0.2372946292,-0.1239616424,-0.5642417669,-0.2042975426,-0.3196799457,-0.2337514162,-0.0577933416,-0.0994747952,0.3641085327,-0.2035193145,0.6210821867,-0.2856467068,0.384831816,0.3869747221,-0.2538062036,0.0657707751,0.201900661,0.0772044435,0.0396992639,-0.1581979841,-0.0872712657,0.0368853025,0.0007560396,-0.1040414795,0.1626861542,-0.2012913674,-0.2207449675,0.1988331974,0.0763103962,0.8045886159,-0.0558607131,0.064656727,-0.1496199071,-0.0560613871,-0.1025550291,0.0681211501,0.4282157123,0.586725831,-0.2910673916,-0.0661168844,-0.0673986897,-0.0161997508,-0.0980607271,-0.0475920923,-0.2880606055,0.3186068237,-0.0191521253,0.0055988389,0.262262702,0.2214763165,-0.0565426797,0.0405983664,-0.0261393134,-0.3694877028,0.5187733173,-0.1914700568,-0.1348089725,-0.1738880426,0.0139728244,0.0507746078,0.104873769,-0.3421268761,0.1284422427,0.2554286122,0.0750762448,-0.7180352211,0.4141148031,-0.2150640041,-0.0544044301,0.0822755918,0.0061151502,-0.1176430583,0.0204468705,-0.1353336126,-0.1060189977]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1811","title":"Unable to add Multi-label Datasets","comments":"Closing this issue since it was answered.","body":"I am trying to add [CIFAR-100](https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html) dataset. The dataset contains two labels per image - `fine label` and `coarse label`. Using just one label in supervised keys as \r\n`supervised_keys=(\"img\", \"fine_label\")` raises no issue. But trying `supervised_keys=(\"img\", \"fine_label\",\"coarse_label\")` leads to this error : \r\n\r\n```python\r\nTraceback (most recent call last):\r\n File \"test_script.py\", line 2, in <module>\r\n d = load_dataset('.\/datasets\/cifar100')\r\n File \"~\/datasets\/src\/datasets\/load.py\", line 668, in load_dataset\r\n **config_kwargs,\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 896, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 247, in __init__\r\n info.update(self._info())\r\n File \"~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cifar100\/61d2489b2d4a4abc34201432541b7380984ec714e290817d9a1ee318e4b74e0f\/cifar100.py\", line 79, in _info\r\n citation=_CITATION,\r\n File \"<string>\", line 19, in __init__\r\n File \"~\/datasets\/src\/datasets\/info.py\", line 136, in __post_init__\r\n self.supervised_keys = SupervisedKeysData(*self.supervised_keys)\r\nTypeError: __init__() takes from 1 to 3 positional arguments but 4 were given\r\n```\r\nIs there a way I can fix this?\r\n\r\nAlso, what does adding `supervised_keys` do? Is it necessary? How would I specify `supervised_keys` for a multi-input, multi-label dataset?\r\n\r\nThanks,\r\nGunjan","comment_length":7,"text":"Unable to add Multi-label Datasets\nI am trying to add [CIFAR-100](https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html) dataset. The dataset contains two labels per image - `fine label` and `coarse label`. Using just one label in supervised keys as \r\n`supervised_keys=(\"img\", \"fine_label\")` raises no issue. But trying `supervised_keys=(\"img\", \"fine_label\",\"coarse_label\")` leads to this error : \r\n\r\n```python\r\nTraceback (most recent call last):\r\n File \"test_script.py\", line 2, in <module>\r\n d = load_dataset('.\/datasets\/cifar100')\r\n File \"~\/datasets\/src\/datasets\/load.py\", line 668, in load_dataset\r\n **config_kwargs,\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 896, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"~\/datasets\/src\/datasets\/builder.py\", line 247, in __init__\r\n info.update(self._info())\r\n File \"~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cifar100\/61d2489b2d4a4abc34201432541b7380984ec714e290817d9a1ee318e4b74e0f\/cifar100.py\", line 79, in _info\r\n citation=_CITATION,\r\n File \"<string>\", line 19, in __init__\r\n File \"~\/datasets\/src\/datasets\/info.py\", line 136, in __post_init__\r\n self.supervised_keys = SupervisedKeysData(*self.supervised_keys)\r\nTypeError: __init__() takes from 1 to 3 positional arguments but 4 were given\r\n```\r\nIs there a way I can fix this?\r\n\r\nAlso, what does adding `supervised_keys` do? Is it necessary? How would I specify `supervised_keys` for a multi-input, multi-label dataset?\r\n\r\nThanks,\r\nGunjan\nClosing this issue since it was answered.","embeddings":[0.0216598418,-0.2712222338,0.0181464944,0.1329571754,0.3712174594,0.3166499734,0.6164801717,-0.4762451947,0.1994253397,0.2660482228,-0.2177258879,0.1535530984,-0.2759885192,0.3479327261,0.1080327854,-0.1204760596,0.1780963987,-0.217401877,0.0256008487,-0.033268027,-0.2215413302,0.0074294899,-0.2346253097,-0.2840359509,-0.6864250302,-0.2522574961,-0.348762989,-0.1600370556,-0.2611855567,-0.6842919588,0.1742349118,0.3276168406,-0.1799799353,0.4868170321,-0.0001189698,0.053939525,0.0060154423,-0.0270139854,-0.1315755099,-0.0884083286,0.0338955149,-0.0997653976,-0.0230811313,-0.1876841038,-0.0205910355,-0.3216594756,-0.0478544869,-0.2678296268,0.2622791529,0.1708982736,0.1447003931,-0.1211824119,0.3119575679,-0.1777848601,-0.0340049379,0.2013507783,-0.1293978244,0.1572745889,0.1325554699,-0.1618613303,0.4810110629,0.2527094781,-0.0045298049,-0.1883502454,0.4128499031,-0.0407814495,-0.1479687542,-0.5755074024,0.1524925977,0.3097066581,0.6528835297,-0.0376973525,-0.6485276222,-0.144102335,0.3986090422,-0.1604739726,-0.1403426677,0.010431719,-0.0038597537,-0.2375007272,-0.3692445457,-0.0498816818,-0.2118247896,0.2764362693,0.333170861,0.3576138914,0.0522686765,0.35098809,0.2702969611,-0.0573835894,-0.0252628159,-0.1056556776,0.0451553315,0.2221504897,-0.2898245156,0.1726470292,-0.0475980975,0.1391517818,0.1451030672,-0.1027215347,0.0525578931,-0.3232679665,-0.211719349,0.1196939275,0.1027332097,-0.0196695887,-0.0022725062,0.4025920331,-0.0366642326,-0.0274585523,0.1102099642,-0.2352029085,-0.0683441162,-0.1296106279,0.3619584143,0.1870019734,-0.1424213946,0.1166465431,0.0462243594,-0.15703246,-0.061152637,-0.0322760865,0.3567192256,0.309689492,0.2346779108,0.0138949612,0.1058847159,0.4291902483,0.1078813076,0.0548968427,-0.1495378613,0.35435009,-0.1401570141,0.0218004473,0.2909840345,0.0014130031,0.1271663308,0.3206980526,0.2556933761,-0.1487238407,0.0474536568,-0.2202002108,0.1937618554,0.3569915891,-0.0255860426,0.2145913243,-0.0456783921,-0.3797814846,-0.1100404412,0.04526655,-0.3311825991,-0.0257803425,-0.052225735,0.0699773654,0.1953081638,-0.0917193443,-0.4771164656,0.2014684528,0.1235159412,0.0339203253,-0.0702257752,-0.2045563459,-0.2669656575,-0.2525112927,0.2751828134,0.401172787,-0.2575544119,-0.209166497,-0.1150707379,0.16939044,-0.2034942359,0.2128368467,-0.2620731592,0.4487377703,-0.0940307975,0.0628931224,0.042674195,-0.8922803998,-0.0666826069,-0.0147576723,0.0866596475,-0.128503859,0.1614068151,0.3256177604,0.1087990627,0.0289789066,0.1426028311,0.1137600467,-0.1924496591,-0.0482096337,0.1788373441,0.1565864831,-0.1048276648,0.3070019782,0.2102843821,-0.2103247494,0.1564217657,-0.1566660702,-0.0160315279,-0.1635690033,0.12073908,0.3795161843,0.2092370689,-0.0499869846,0.0313238017,0.0380041413,-0.4350197017,0.4339157343,0.2955799401,0.3049954176,-0.1049287096,-0.1757050157,-0.0116091687,0.0613495074,-0.4852899611,-0.1146177873,0.0656736344,0.2741549015,-0.0914811268,-0.0599610843,0.0205955021,0.1393007636,0.4397166073,0.1631942391,0.0898363367,0.0351097733,-0.1254134476,0.031797003,-0.0982277617,0.0179512184,0.4089220166,-0.0632736459,0.0348855965,0.4561132491,0.1855800599,-0.3028663397,-0.1581857651,0.0631943867,0.4092926979,-0.2760539651,-0.2940944433,-0.2007402331,0.2320427746,-0.3411729634,-0.155014053,0.0775262788,-0.3345649242,0.2867106795,-0.0063107894,0.1957471073,0.1347314268,0.1028221473,-0.1135910451,-0.2067493051,0.0379192606,-0.1822332293,-0.0228359047,-0.2354649007,0.0413877219,0.0655713156,0.0258155391,0.0400971286,-0.0610706173,-0.0535848029,0.0286562536,0.2431066483,0.1560307741,0.5032753944,0.2892154753,0.2665799856,-0.1650920212,-0.2013620734,-0.2896959782,0.0578938462,0.1002955586,-0.0812715143,0.1857181787,0.1985814273,0.1882681102,0.0334806368,-0.2237827629,-0.4800541997,0.3684946597,0.1311239153,-0.2911791503,0.0136999218,-0.3054610789,-0.209730491,-0.0478190146,-0.2719694078,-0.0329390801,-0.273193121,0.100284718,0.2303827852,0.1110883728,0.2187402695,0.0505732298,0.2245625407,-0.2533396184,0.0736529753,-0.0290039685,-0.1045127735,-0.002999285,-0.0265577007,0.3638055325,-0.3310750723,0.5024328232,0.0892854407,-0.1429113895,-0.204916358,-0.4361688793,0.0277465582,-0.1707712859,0.2560727,0.0792692378,-0.0453793332,-0.120553337,-0.2528287768,0.3231553733,0.1433298141,-0.067182377,-0.02225996,-0.2092654854,-0.3181765974,-0.3291820884,-0.0019011601,-0.2415915877,-0.0313346125,0.1783394068,0.0039089699,-0.0620984249,0.6229845881,0.1421978921,-0.3269357383,0.1114385575,0.1087919623,-0.0204633344,-0.3791366518,0.230792135,-0.3280832469,-0.245816052,0.1112751365,-0.2078543007,0.4561773539,-0.0236298926,-0.074241221,-0.0516700149,0.06004408,0.0250432417,-0.0242213272,0.2214048356,0.1233270913,0.1223821193,0.0012233047,-0.2397067994,-0.4332716763,0.3612817526,-0.230379194,0.2302520275,-0.3453046679,0.2405825853,0.0744778737,0.2453038096,0.0080904132,0.0415512472,0.4380138814,-0.2188450545,0.3824874759,-0.1967352331,-0.5095447898,0.2663565278,0.0068039051,0.0746747702,-0.0564309396,0.0025118168,-0.1534558535,-0.13697429,0.4596606493,-0.2982488871,-0.0714152157,-0.1395442188,-0.0614005066,0.1793229282,-0.1286798716,-0.2386004925,-0.4252116382,-0.0979251638,-0.3168427944,0.0837059692,0.030056743,-0.0836135522,-0.5794285536,-0.279966265,-0.2095399946,0.0962752774,0.2015597671,0.1998217553,-0.250143379,0.3073728979,0.0423248895,0.0750111043,0.6299776435,-0.5078380108,-0.207209006,0.4064024687,0.0823346972,-0.4420149326,-0.3596889377,-0.0637626573,0.2939634025,0.0598797761,0.3375253379,-0.3172761202,0.1752974391,0.1442773193,0.0290731676,-0.1157465428,-0.2602501512,-0.1650700122,-0.3463683724,-0.2542931139,0.0387581997,0.2610555291,0.2220253199,0.0756111965,0.05578788,0.0451995172,0.0471867733,0.1067099497,-0.0073393807,0.0891595408,0.5999578238,0.3521910608,0.0442494377,-0.223894015,0.5654521585,0.499717921,-0.3498901725,-0.2776920497,-0.0696075335,-0.2492358238,0.2417492121,-0.0740425736,0.2450897247,0.2074229419,0.2566080987,0.0326639861,0.0169424899,0.1901456118,0.5320228338,-0.0338639915,-0.5876461864,-0.3630560637,0.1882240474,0.0840374306,-0.0925073922,0.3905680776,0.035589207,-0.310058862,0.2069021463,0.152016446,0.7129688263,-0.5040864944,0.2006383538,-0.1281060278,-0.051407069,0.6691420674,0.0259150546,0.1060547009,-0.1650911123,-0.2414849997,0.107521005,0.0631438717,0.0530461222,0.3399398625,-0.1476923823,0.1040081382,-0.1996166557,0.1442142576,-0.2286387533,0.0942784548,0.2413909137,-0.3566122055,0.315527916,0.1148123816,-0.0379253142,-0.2212827057,0.0147405667,-0.0131456228,-0.0485539548,-0.2762643993,-0.4560342729,-0.1082490757,-0.3264353573,0.351482749,-0.0545667596,-0.082947515,0.0922916085,0.3642334342,0.6377292871,-0.0870183855,-0.0121600647,0.1931255907,-0.007270975,-0.1853410453,-0.4156341255,-0.2583948374,0.1583053172,-0.0967743844,-0.3491543531,0.0630329549,0.0483213104,-0.2924515903,-0.1737174392,0.1112410799,-0.1338459104,0.0843074471,0.0194527209,0.2658884227,0.2253153622,-0.3193776011,0.0304356217,0.0522682555,-0.0975166485,0.334641397,-0.1579447091,0.0292434208,-0.0027380958,0.3327620029,0.0285511427,0.2265633047,0.417263478,-0.0195385832,0.0441499874,-0.0707633495,-0.0785979331,0.2807226181,-0.3569045067,-0.0215917844,0.4509515166,0.0682092831,-0.0082824528,0.4070376158,0.0060866834,0.1032975987,0.0879296362,-0.1227887198,-0.4140322208,0.0571798645,-0.4007037878,0.2787562013,-0.0407934636,0.4472046494,0.1755109727,0.4107339382,-0.2010179609,-0.0609041527,-0.2943552136,-0.0041191275,0.4779884815,0.0559544377,-0.1362131238,0.0865932256,0.10840258,-0.1853463352,0.1712531149,-0.031114975,-0.1749444455,0.1246730909,0.2514070272,0.1749788076,0.0139208622,-0.4309507608,-0.1583899558,-0.1143202707,0.2133870125,0.2629471719,-0.0949871689,0.3122138679,-0.4964765906,0.0038022301,0.0238737185,0.1440283507,0.0420851298,-0.0607595481,0.3293673396,0.0790245682,0.127211228,-0.1499333233,-0.3488268852,0.0482538603,-0.4432691634,-0.1288997084,0.0841796026,-0.3632187247,0.4464629889,0.2567118406,0.2399899364,-0.1217627004,-0.2593407631,0.4296592474,0.0848011822,0.073372066,-0.2197782993,-0.4196273088,-0.0084309662,0.0238845516,0.1413663328,0.2538610399,0.2197544277,-0.0843564644,0.0170371383,-0.1205665097,0.0592124984,0.0718337819,0.1195045114,0.1281981468,-0.0726764426,0.0837647766,0.2819838822,0.2315364778,0.0600666255,0.194866702,-0.0722384155,0.3028277159,-0.1764678955,-0.0757141411,0.240550667,-0.3558896482,0.4976215661,-0.1813222617,-0.0065754028,-0.044587303,-0.0960169062,0.2008923292,-0.2110212594,-0.3887046874,-0.2711659968,0.3218339682,-0.1305858493,-0.1066607237,-0.182562381,-0.105645217,-0.0563145503,-0.0584490336,0.0690638646,-0.1914356351,0.3167058825,-0.1098625362,-0.1567377895,-0.1262349039,-0.4305464029,0.461884141,0.2834605277,-0.2831432223,0.2036371827,0.3541097939,-0.4410456121,0.1897425056,0.3322366178,0.6006937027,0.2195312977,0.2928979397,0.0838777795,-0.1760641783,-0.1078898758,0.1742133349,0.0496440083,0.4417835474,-0.0686274543,0.1369452924,0.0067783715,-0.0832349434,0.2980151176,0.1812421679,0.3393843174,-0.0167961605,-0.3750321269,-0.0341072753,-0.1780612916,-0.0763006732,0.3583650291,-0.3215570152,-0.0696086511,0.2845407724,-0.0804059282,0.0143419076,-0.3408436179,0.0184786692,0.1767684072,0.4923861325,0.1872951239,0.1653291136,-0.2372946292,-0.1239616424,-0.5642417669,-0.2042975426,-0.3196799457,-0.2337514162,-0.0577933416,-0.0994747952,0.3641085327,-0.2035193145,0.6210821867,-0.2856467068,0.384831816,0.3869747221,-0.2538062036,0.0657707751,0.201900661,0.0772044435,0.0396992639,-0.1581979841,-0.0872712657,0.0368853025,0.0007560396,-0.1040414795,0.1626861542,-0.2012913674,-0.2207449675,0.1988331974,0.0763103962,0.8045886159,-0.0558607131,0.064656727,-0.1496199071,-0.0560613871,-0.1025550291,0.0681211501,0.4282157123,0.586725831,-0.2910673916,-0.0661168844,-0.0673986897,-0.0161997508,-0.0980607271,-0.0475920923,-0.2880606055,0.3186068237,-0.0191521253,0.0055988389,0.262262702,0.2214763165,-0.0565426797,0.0405983664,-0.0261393134,-0.3694877028,0.5187733173,-0.1914700568,-0.1348089725,-0.1738880426,0.0139728244,0.0507746078,0.104873769,-0.3421268761,0.1284422427,0.2554286122,0.0750762448,-0.7180352211,0.4141148031,-0.2150640041,-0.0544044301,0.0822755918,0.0061151502,-0.1176430583,0.0204468705,-0.1353336126,-0.1060189977]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1810","title":"Add Hateful Memes Dataset","comments":"I am not sure, but would `datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value(\"int\")))` work?","body":"## Add Hateful Memes Dataset\r\n- **Name:** Hateful Memes\r\n- **Description:** [https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set]( https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set)\r\n- **Paper:** [https:\/\/arxiv.org\/pdf\/2005.04790.pdf](https:\/\/arxiv.org\/pdf\/2005.04790.pdf)\r\n- **Data:** [This link](https:\/\/drivendata-competition-fb-hateful-memes-data.s3.amazonaws.com\/XjiOc5ycDBRRNwbhRlgH.zip?AWSAccessKeyId=AKIARVBOBDCY4MWEDJKS&Signature=DaUuGgZWUgDHzEPPbyJ2PhSJ56Q%3D&Expires=1612816874)\r\n- **Motivation:** Including multi-modal datasets to \ud83e\udd17 datasets.\r\n\r\nI will be adding this dataset. It requires the user to sign an agreement on DrivenData. So, it will be used with a manual download.\r\n\r\nThe issue with this dataset is that the images are of different sizes. The image datasets added so far (CIFAR-10 and MNIST) have a uniform shape throughout.\r\nSo something like \r\n```python\r\n datasets.Array2D(shape=(28, 28), dtype=\"uint8\")\r\n```\r\nwon't work for the images. How would I add image features then? I checked `datasets\/features.py` but couldn't figure out the appropriate class for this. I'm assuming I would want to avoid re-sizing at all since we want the user to be able to access the original images.\r\n\r\nAlso, in case I want to load only a subset of the data, since the actual data is around 8.8GB, how would that be possible?\r\n\r\nThanks,\r\nGunjan","comment_length":8,"text":"Add Hateful Memes Dataset\n## Add Hateful Memes Dataset\r\n- **Name:** Hateful Memes\r\n- **Description:** [https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set]( https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set)\r\n- **Paper:** [https:\/\/arxiv.org\/pdf\/2005.04790.pdf](https:\/\/arxiv.org\/pdf\/2005.04790.pdf)\r\n- **Data:** [This link](https:\/\/drivendata-competition-fb-hateful-memes-data.s3.amazonaws.com\/XjiOc5ycDBRRNwbhRlgH.zip?AWSAccessKeyId=AKIARVBOBDCY4MWEDJKS&Signature=DaUuGgZWUgDHzEPPbyJ2PhSJ56Q%3D&Expires=1612816874)\r\n- **Motivation:** Including multi-modal datasets to \ud83e\udd17 datasets.\r\n\r\nI will be adding this dataset. It requires the user to sign an agreement on DrivenData. So, it will be used with a manual download.\r\n\r\nThe issue with this dataset is that the images are of different sizes. The image datasets added so far (CIFAR-10 and MNIST) have a uniform shape throughout.\r\nSo something like \r\n```python\r\n datasets.Array2D(shape=(28, 28), dtype=\"uint8\")\r\n```\r\nwon't work for the images. How would I add image features then? I checked `datasets\/features.py` but couldn't figure out the appropriate class for this. I'm assuming I would want to avoid re-sizing at all since we want the user to be able to access the original images.\r\n\r\nAlso, in case I want to load only a subset of the data, since the actual data is around 8.8GB, how would that be possible?\r\n\r\nThanks,\r\nGunjan\nI am not sure, but would `datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value(\"int\")))` work?","embeddings":[-0.2937175333,0.0855384767,-0.1159735844,0.0699359924,0.1269494891,0.2755765915,0.4390200377,-0.0566710681,-0.1771434247,-0.1746338904,0.0117157195,-0.2779462934,-0.413777858,0.2872680426,-0.0266066231,-0.1327530891,0.0206236802,0.0383341871,-0.1117254421,-0.1070943326,-0.0549840815,-0.2103049904,0.1676933616,-0.330945611,-0.5663618445,-0.0882714689,-0.3568055332,0.1963410378,-0.2136043459,-0.4603568614,0.0225622263,0.4092771709,0.1220521256,0.5035597086,-0.0001194587,0.0485528745,-0.0649021715,-0.205264017,0.0972615108,0.0906426758,0.0360072628,-0.5808857083,-0.1404824555,-0.2241454273,0.0812341794,-0.4965279102,0.183574751,-0.0835044533,0.6133990884,0.0107100774,0.1329786628,-0.0971789658,0.1576445997,-0.0082562743,0.0698476732,0.6771777868,-0.0687815547,-0.1844714582,0.1253792346,0.0879749432,0.218440339,0.3165776134,-0.1011788324,0.0878188461,0.4719903469,-0.1527953148,-0.0212135818,-0.4029989541,0.0196486376,0.352358073,0.3491420448,-0.0356927104,-0.3349613547,-0.4729738235,0.0084397532,-0.0069883284,-0.1372766644,0.2877499163,-0.0156178651,-0.1148332953,-0.3858343065,-0.0660942867,-0.3324632347,0.4430134594,-0.059737239,-0.0666999072,0.0409867801,0.3385580182,-0.0676361918,-0.3059031367,-0.0550892539,-0.0655018091,0.3048940599,0.3053776622,0.1959698498,0.0150656905,-0.205292061,-0.1631768197,0.4885697961,-0.1952335238,-0.0777987167,-0.0631569922,-0.5285086632,-0.1411130875,0.2447132319,-0.1947510391,-0.3624276221,0.371842891,0.099360764,0.065913938,-0.0906532258,-0.0027018685,-0.1000854149,0.0327966884,-0.0238929093,0.0080758072,0.2431573719,0.0254494287,-0.1853599697,-0.2103683054,0.1728921235,0.0585894994,-0.0892397985,0.1712954789,0.087433055,0.0176439993,-0.0761236921,0.2548806965,0.0201505329,-0.1551139057,-0.1470597982,-0.038162142,-0.11110273,0.2274316251,0.2764129639,0.263759315,0.3128427267,0.0027551625,0.0036961986,0.1025279835,0.175251618,-0.1694854647,0.1583035141,0.084591493,0.0169257093,-0.0464987196,-0.0098418258,0.1222071424,-0.1681594104,0.153349176,-0.2954181135,-0.0112610776,-0.0650695488,0.065855585,-0.2774297595,0.1206959784,-0.0711846948,0.3908601403,0.0377172939,-0.087345764,0.1390065551,-0.0971954837,-0.3133896589,-0.3689454794,0.0396298058,0.6324535608,-0.4300075471,-0.049863901,-0.3333335817,0.0106501337,-0.3215206563,0.1146233901,-0.0356093682,0.0472816005,-0.1277965456,-0.1053192914,0.2398156375,-0.1694946587,-0.343577832,0.1769169569,0.0491777584,0.2673987448,0.141945526,0.4753127098,0.1557541192,-0.0811164677,0.0108433468,0.5948139429,-0.2657892108,-0.0580342859,0.2125912607,-0.4148363769,0.0453303643,0.4032479525,0.3112516105,-0.2243260145,0.3586634398,-0.1759250164,0.1523033679,-0.2508983612,0.1786841452,0.1007788107,0.3106218576,0.0462820865,0.099262625,-0.1949589849,-0.6319137812,0.3118923306,-0.1677637398,-0.1596534103,-0.1802334189,-0.1253962368,-0.2578324378,-0.1533289552,-0.5462970734,-0.1065131426,0.0525083058,0.1168013513,0.0687274486,-0.0959997475,-0.0954144672,0.3298021853,0.0977478549,0.083773993,0.0377462879,0.3489416838,0.2924624085,-0.2212093771,-0.1156145483,0.1714469939,0.003044724,-0.2718423605,0.0240103211,0.4639286697,0.283934176,0.2443392873,-0.1027240232,0.47948125,0.4339742362,-0.3778837323,0.4103341401,0.2700510323,0.1511265486,-0.1987266392,-0.131663695,0.4820560813,-0.0989778191,0.0622776113,-0.1282143891,-0.1457249373,0.1021323577,0.2004790008,-0.119837828,-0.4945302308,-0.0719330087,0.4070829749,-0.0655244663,-0.0510894731,0.0183770694,-0.0630109906,0.1913587749,0.126072824,0.1482370645,0.1109215692,-0.2234340757,0.2331326455,0.1464485228,-0.0776320323,0.5119310021,0.2028425336,-0.256061703,0.2130887359,0.07395643,0.0798793435,-0.1104858592,0.0909354612,0.1221527979,0.2030837983,0.2969138622,-0.0957301632,-0.3028353453,-0.3606948853,-0.0209493004,-0.1488836706,0.0580920354,-0.1526671946,-0.1813776344,-0.0529471636,-0.2147160918,0.0161727499,-0.3431216776,0.2057517767,0.2428832501,-0.1148703098,-0.0036785163,0.1080254912,-0.1788789183,0.483250469,0.2085607946,-0.118267931,0.1848701537,0.253403008,-0.1923679709,0.0083490964,0.2781821191,0.0008617757,0.3136177063,-0.2379128933,0.3229171038,-0.3200471997,-0.5347151756,0.0125321979,-0.1872775406,0.4736200571,0.1230406761,0.0623102561,0.1620389819,0.0274044815,0.2825316489,0.1366673708,-0.0711084828,-0.3312490582,-0.3168213069,-0.049915053,0.0910056755,0.1572563946,-0.306170404,-0.2408108562,0.465616107,0.5908679366,0.1045459509,0.215658322,0.5232957006,-0.1983288229,0.176995635,0.2024472505,-0.0926145613,-0.072984837,0.4671340585,-0.2270825952,-0.3014446199,-0.0322095565,-0.2163986415,-0.2421966791,0.386390686,-0.5156235695,-0.1705159694,-0.1243104413,0.2059426755,0.0055515938,0.0254959185,0.0642079189,-0.0854989141,0.1152484119,-0.0361449085,-0.0992598087,-0.018799644,0.3751676679,0.1980702728,0.0457677953,-0.0310681351,0.2545377016,0.5387602448,0.0517804883,-0.1016361713,0.445482254,0.0193946511,0.2526952922,-0.0042883721,-0.3327640891,0.317035228,0.1641715318,-0.1010322049,0.1285212338,0.1007304788,0.0991789922,0.0195760839,0.3976241946,-0.128722012,-0.1870105714,0.2253072858,-0.0627800673,0.2216213495,-0.198886618,-0.2431411445,-0.7520788312,-0.2509350181,0.0254244134,0.3237931132,-0.2463395596,0.2239464968,-0.6154397726,0.0465481542,0.1328912228,0.3980661333,0.1483188719,0.070229955,-0.0629505217,-0.0562450886,-0.0330265835,0.0213278756,0.7140958309,0.0233327132,-0.3234889507,0.0020836294,0.2037481368,-0.3840944767,-0.2931174338,0.0934495106,-0.129262805,-0.1516548544,0.7130683661,-0.2277084142,0.0584715791,0.2418118566,0.3138315678,-0.1722397804,-0.2536334693,-0.1142407879,-0.3545643091,-0.4567254186,-0.106880337,0.1781568527,0.0292391684,0.1301032752,-0.4181121886,-0.0993959457,-0.1857047379,0.1741354018,0.0826415122,0.0471070483,0.2865478098,-0.0148238065,0.139746204,-0.1508102715,0.2505714297,0.4330878556,0.0506872199,-0.2388632894,0.058222726,-0.1679072082,0.0910295323,0.4374073148,0.2501261532,-0.1954241097,0.0759148598,-0.0400852636,-0.2038255781,0.4758559763,0.1709891409,-0.2010823786,-0.0993547961,-0.6172378659,0.2699255049,0.1844975948,0.0398372263,0.1084737331,-0.1405309141,-0.2386515588,0.3419980407,0.3702081442,0.9126947522,-0.1187946647,0.3275818527,-0.1839372963,0.2664664686,0.1573178917,-0.224052906,-0.1418024451,0.0873244405,-0.0931925476,-0.1199782938,-0.0762330219,0.3079785407,0.2201264799,0.0824639574,0.0638758838,0.1005969271,0.1047132388,-0.0967094451,0.2107244134,-0.3313531876,-0.3187658489,0.0116908401,0.0724847242,0.0189074688,0.0320192724,-0.1079508662,-0.105491899,-0.1694928706,-0.0939154401,-0.3186375201,-0.3952669799,-0.5887841582,0.2372635901,0.0982396975,-0.1717513055,-0.0190417711,0.3852882683,0.2168402225,-0.0411626063,-0.0882696211,0.2087244838,0.1865607649,-0.034412954,-0.4086063802,-0.4450063109,0.4324043691,0.0107570533,-0.1028217599,0.0216205176,0.1825669259,-0.051647041,-0.5810029507,-0.0868032351,0.3680060506,-0.2799919248,-0.3647779524,-0.057400398,0.1363157481,-0.0419157892,0.0368162394,0.1211728007,0.0540438592,0.1900109947,0.2775429785,-0.1119126007,0.0003479132,0.0473747253,0.2192946821,-0.2978385389,0.3861463368,0.0889807045,0.2237133086,-0.0926382542,0.3294116557,0.24729684,-0.2427513301,0.0427171029,0.0018465582,-0.0660758317,-0.2378780097,0.2711307108,0.2923834324,0.0730149001,0.2204766721,-0.4516609013,-0.3665493727,0.287950933,-0.0624154247,0.4061438143,-0.0885281786,0.2674668729,-0.1780355275,0.2894331217,-0.2288457006,0.2335332632,0.0150095504,0.061566364,0.1269450933,-0.0706922188,-0.3242351413,0.2337631285,0.066938512,0.0303459615,-0.4507034421,-0.0747660473,-0.1806904525,0.2190743387,0.1604650468,-0.1158462241,0.0515569113,-0.2943884432,-0.0646273643,-0.0885691941,0.217425406,-0.1590387523,-0.3185085356,-0.0027010243,-0.2785857618,0.0019581381,-0.030958388,0.1845767945,0.034449853,0.1251820624,0.4537071288,0.4655617774,-0.0770083815,-0.0418921933,0.0099931452,0.263631314,0.1702830642,0.0708016753,0.2408603281,-0.1613979936,0.1950865984,0.1346160322,-0.013130879,0.3477699459,-0.3733760715,0.1125284955,0.1391809583,0.1129346266,-0.2520384192,0.031234242,0.3882946074,-0.0807227343,0.1353029609,0.3263286352,0.3564777076,0.3188418448,-0.3227795064,0.0825773403,0.1910686195,0.0314009972,0.1358549595,0.1778287441,-0.2700680196,0.0838016495,0.2176420242,0.4763093293,0.1700156182,0.428093493,-0.309386313,0.339989543,0.0792570114,-0.1689702272,0.2990726233,-0.6059762239,0.3069840074,0.3864267468,0.080413565,0.4509662986,-0.1428806484,0.0776891336,-0.0797230825,-0.3499289453,-0.1455958635,0.1439571679,0.2603924274,-0.1001931056,-0.3823722303,-0.3482127488,-0.5616949797,-0.0275059883,-0.039100565,-0.5293336511,0.1426361203,-0.1042367145,-0.1121784076,-0.2762344778,0.0126541341,0.218073383,0.028202204,-0.3541238308,0.2240118831,0.4073631465,-0.0886428356,0.2655682564,-0.1395180076,0.2475689054,0.1413712204,-0.2297652215,0.2979266047,-0.0986753404,-0.2477451265,-0.1063093022,0.1104971394,0.0706892237,0.3479542136,0.1064423621,0.0733980611,-0.1141804531,-0.2000591308,-0.118723236,-0.0669871271,-0.6041572094,-0.0501908287,-0.1807584465,-0.2360827029,-0.0957510248,0.0517643169,-0.3780147135,-0.6041253805,0.5218055844,-0.4665619433,0.2061370909,-0.0659727901,0.0029283084,-0.4517360032,0.0474342741,0.3793910444,-0.0032305494,0.0205464344,-0.0647329167,-0.3843297362,0.1998124272,-0.2432140559,-0.3210808337,-0.0064151343,0.2512453496,0.2229406834,0.0421656743,0.3162599206,-0.1036214158,0.0805238858,0.2197497934,-0.2301976681,-0.0105260946,0.0730203465,0.542717278,0.0479770489,-0.1563195437,0.0096952431,-0.0218954366,-0.0290360246,-0.1323186755,-0.2471099943,0.104664512,-0.3426624835,0.359337002,0.0583583601,0.2325874716,-0.0230351314,-0.1040937677,-0.360812813,0.1507022679,-0.002870783,0.0321106464,0.1347031593,0.7245918512,-0.2765657902,0.2687219083,-0.0653663874,0.2870862484,0.1020112485,-0.3651659489,-0.0976069868,0.2973195016,-0.2652300596,0.0071447636,0.1769500077,0.3343554437,-0.0109333657,-0.1203945056,0.3280436099,-0.3239560127,0.4122439623,-0.2093634158,-0.4671034813,-0.037503276,-0.1104955822,0.065776214,0.4514725804,-0.2580550015,0.2579109371,0.3936809599,0.0531577058,-0.0908820406,0.3012433648,0.005796588,-0.4690389037,-0.0028442098,0.1708011627,-0.3031516671,-0.1876915842,-0.0274309516,-0.2910802364]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1810","title":"Add Hateful Memes Dataset","comments":"Also, I found the information for loading only subsets of the data [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/docs\/source\/splits.rst).","body":"## Add Hateful Memes Dataset\r\n- **Name:** Hateful Memes\r\n- **Description:** [https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set]( https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set)\r\n- **Paper:** [https:\/\/arxiv.org\/pdf\/2005.04790.pdf](https:\/\/arxiv.org\/pdf\/2005.04790.pdf)\r\n- **Data:** [This link](https:\/\/drivendata-competition-fb-hateful-memes-data.s3.amazonaws.com\/XjiOc5ycDBRRNwbhRlgH.zip?AWSAccessKeyId=AKIARVBOBDCY4MWEDJKS&Signature=DaUuGgZWUgDHzEPPbyJ2PhSJ56Q%3D&Expires=1612816874)\r\n- **Motivation:** Including multi-modal datasets to \ud83e\udd17 datasets.\r\n\r\nI will be adding this dataset. It requires the user to sign an agreement on DrivenData. So, it will be used with a manual download.\r\n\r\nThe issue with this dataset is that the images are of different sizes. The image datasets added so far (CIFAR-10 and MNIST) have a uniform shape throughout.\r\nSo something like \r\n```python\r\n datasets.Array2D(shape=(28, 28), dtype=\"uint8\")\r\n```\r\nwon't work for the images. How would I add image features then? I checked `datasets\/features.py` but couldn't figure out the appropriate class for this. I'm assuming I would want to avoid re-sizing at all since we want the user to be able to access the original images.\r\n\r\nAlso, in case I want to load only a subset of the data, since the actual data is around 8.8GB, how would that be possible?\r\n\r\nThanks,\r\nGunjan","comment_length":13,"text":"Add Hateful Memes Dataset\n## Add Hateful Memes Dataset\r\n- **Name:** Hateful Memes\r\n- **Description:** [https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set]( https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set)\r\n- **Paper:** [https:\/\/arxiv.org\/pdf\/2005.04790.pdf](https:\/\/arxiv.org\/pdf\/2005.04790.pdf)\r\n- **Data:** [This link](https:\/\/drivendata-competition-fb-hateful-memes-data.s3.amazonaws.com\/XjiOc5ycDBRRNwbhRlgH.zip?AWSAccessKeyId=AKIARVBOBDCY4MWEDJKS&Signature=DaUuGgZWUgDHzEPPbyJ2PhSJ56Q%3D&Expires=1612816874)\r\n- **Motivation:** Including multi-modal datasets to \ud83e\udd17 datasets.\r\n\r\nI will be adding this dataset. It requires the user to sign an agreement on DrivenData. So, it will be used with a manual download.\r\n\r\nThe issue with this dataset is that the images are of different sizes. The image datasets added so far (CIFAR-10 and MNIST) have a uniform shape throughout.\r\nSo something like \r\n```python\r\n datasets.Array2D(shape=(28, 28), dtype=\"uint8\")\r\n```\r\nwon't work for the images. How would I add image features then? I checked `datasets\/features.py` but couldn't figure out the appropriate class for this. I'm assuming I would want to avoid re-sizing at all since we want the user to be able to access the original images.\r\n\r\nAlso, in case I want to load only a subset of the data, since the actual data is around 8.8GB, how would that be possible?\r\n\r\nThanks,\r\nGunjan\nAlso, I found the information for loading only subsets of the data [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/docs\/source\/splits.rst).","embeddings":[-0.2698048651,0.0629496351,-0.1263260543,0.1009711325,0.100906752,0.2537287176,0.4158315361,-0.035687156,-0.1360063851,-0.1861694604,-0.0590821505,-0.3079329133,-0.366784662,0.3052083552,0.0037772416,-0.1220715865,0.0089078024,0.0843262523,-0.1353784353,-0.0887768269,-0.0366887674,-0.1925545484,0.1741472632,-0.3324965835,-0.5459133983,-0.0685199425,-0.3048211634,0.2468579113,-0.2426986843,-0.4581846595,0.0387089662,0.3708100617,0.1331256181,0.5253241658,-0.0001187238,0.0858247057,-0.0459977575,-0.1917188019,0.0577797778,0.0437298417,0.0595304705,-0.5751154423,-0.1188848987,-0.197877869,0.091283761,-0.519313395,0.1551680416,-0.0690651312,0.6851333976,0.0173047744,0.1306125522,-0.0588219278,0.1484666467,-0.0070960815,0.0677287504,0.6541165709,-0.0575960688,-0.1602327079,0.1243478507,0.081042856,0.1819947213,0.2907532752,-0.0883738399,0.1471549571,0.4553940296,-0.1656918377,-0.0537802614,-0.453276515,0.0438518561,0.4020507932,0.3492212594,-0.0264959559,-0.3448543847,-0.4710114896,-0.0147447223,-0.0447330624,-0.1292107403,0.3056822121,-0.0374915749,-0.0813319683,-0.4014496803,-0.042522721,-0.3317481577,0.4222877026,-0.0802344829,-0.0929647163,0.0044290777,0.3131304979,-0.0314225182,-0.2837106287,-0.070527859,-0.0367271788,0.2657312453,0.3052322268,0.1335976422,0.0182673559,-0.2217299342,-0.1685773581,0.5121042728,-0.2019746751,-0.0496922173,-0.0319277048,-0.5223115683,-0.1523485184,0.2879935205,-0.1743933707,-0.3375919461,0.3382920921,0.1158279479,0.0533986539,-0.0733788833,-0.0048812805,-0.1329013556,0.0515440106,-0.0987999365,-0.0289717242,0.2286360115,-0.0127898352,-0.2251685262,-0.2320984155,0.1927139759,0.0141397351,-0.0779580399,0.2157581896,0.0748248547,0.0020655559,-0.0824438408,0.243811816,0.0100877816,-0.1633301228,-0.1444556862,-0.0218721796,-0.1034576148,0.21706523,0.3236093521,0.227643013,0.3428680301,-0.0281400625,0.0130040888,0.09858381,0.1714543253,-0.1874876022,0.1456341147,0.0968333557,0.0700643212,-0.0444812886,0.0227530431,0.1597344726,-0.1522202194,0.1641412675,-0.2759128809,0.0063350326,-0.0898526758,0.0606805682,-0.3039343953,0.1466388106,-0.1236879081,0.4143945873,-0.0084347604,-0.0956233069,0.130501762,-0.0388024934,-0.3115499318,-0.3598717153,0.1218542904,0.6289481521,-0.4274500906,-0.0551858172,-0.3068075478,-0.0021861366,-0.3297739029,0.1625503302,-0.0495067164,0.0619581118,-0.1302774847,-0.0769171268,0.2154412568,-0.1167767495,-0.3651899397,0.2167813033,0.0395334847,0.275895834,0.1414383948,0.4193153083,0.1665357798,-0.0658455193,0.0135022802,0.6329682469,-0.2212257087,-0.0289444122,0.2220250964,-0.4107780159,0.0769129768,0.3960155845,0.2879933417,-0.1941358745,0.3847528398,-0.1724174023,0.1456891,-0.2317305505,0.1905541718,0.0884116292,0.2926544845,0.0370090678,0.1148018241,-0.2210638076,-0.6492828131,0.3376336992,-0.1619624794,-0.2315296084,-0.1529713571,-0.0864856318,-0.2750979066,-0.1616089642,-0.5329520702,-0.1073854789,0.0452989936,0.0790332258,0.0920447409,-0.1106664985,-0.1315653622,0.3177299798,0.0918055028,0.0767554864,0.0682508051,0.3630800843,0.2929937243,-0.2370770425,-0.0964434594,0.1645198911,-0.0168724451,-0.2954450548,0.0168063305,0.4636930525,0.271963805,0.226866886,-0.1062358469,0.4737734497,0.4125096798,-0.3598571718,0.4109470248,0.2707540393,0.1577754468,-0.1993264407,-0.1406831294,0.4624776542,-0.0940441638,0.0625553355,-0.1137635112,-0.1766258478,0.109888196,0.1920065135,-0.1043899283,-0.5065057874,-0.0549308546,0.4036423862,-0.0323884785,-0.0091973422,0.032885205,-0.1084142253,0.1602377295,0.1299989223,0.1300695837,0.1346637309,-0.2325537652,0.2007126063,0.1347484589,-0.1059595421,0.5092343688,0.1873653531,-0.2475457937,0.1931822151,0.140787527,0.0861924067,-0.0966427997,0.0820862129,0.1327365488,0.2175453752,0.2844741046,-0.094543092,-0.2937791646,-0.3159109354,-0.0147040635,-0.1390279531,0.0378580876,-0.172875002,-0.1929108351,-0.0706275254,-0.1786084771,0.0098447949,-0.3154638708,0.2524059117,0.2329145074,-0.1098884493,0.0034891285,0.0881311074,-0.1573385596,0.498716265,0.1915968508,-0.0330363922,0.1708768904,0.2570089996,-0.1838128567,0.0145647498,0.2902193964,0.0336794183,0.2983180881,-0.2392868549,0.2859788239,-0.3160988092,-0.5136060715,0.0349818207,-0.2051110864,0.4891817272,0.1125223935,0.0989633948,0.1205701903,0.0329959616,0.2862930298,0.0993616804,-0.1113846898,-0.3203083575,-0.311227262,-0.012170461,0.0915217176,0.1662195027,-0.3284063041,-0.2579277754,0.5169570446,0.5712695122,0.1315049529,0.2253345847,0.4949728549,-0.1967157871,0.1644772887,0.1997697651,-0.0919038579,-0.0985587165,0.4818865955,-0.2421328574,-0.3297350109,-0.0130916024,-0.174441129,-0.2733517885,0.3758663833,-0.5316542983,-0.1793955863,-0.1323262155,0.1897663176,0.0023909332,0.0386139825,0.0207745302,-0.1399775892,0.0989354029,-0.0549661815,-0.1139334068,-0.0388175137,0.377476871,0.169179216,0.0787936822,-0.0444812253,0.2574743927,0.5631974339,0.0556892008,-0.0976566821,0.4849176109,0.0171620995,0.2821604609,-0.034432359,-0.3481866419,0.3294689059,0.1434024423,-0.0794742405,0.1656247526,0.1079793796,0.0921540409,-0.0056859944,0.3321167529,-0.1806746125,-0.170360297,0.2465295494,-0.0373129882,0.1831419766,-0.184737727,-0.2375822961,-0.7183035612,-0.2435868979,0.0432192162,0.2982776761,-0.2102948278,0.2433882505,-0.6099787354,0.0044510956,0.1138031334,0.3698862493,0.1123623252,0.0934444144,-0.0701406598,-0.0570894405,-0.0247327555,0.0543083102,0.7179995179,0.0552390628,-0.2866594493,-0.0266754646,0.1539301127,-0.3867190778,-0.2658391297,0.1183848083,-0.0883328915,-0.1147576645,0.6786510944,-0.2637817264,0.0215875451,0.2330485284,0.3296975195,-0.1695246547,-0.2573253512,-0.136246115,-0.3492185175,-0.4881242514,-0.1156662256,0.18589589,0.0401393063,0.148933664,-0.4106874764,-0.0782124624,-0.1740305871,0.1618881673,0.1034357026,0.0639242902,0.3068849742,-0.0084033851,0.1840675771,-0.1334578991,0.2494772077,0.4584991336,0.0402311049,-0.2625992894,0.0605749711,-0.1325777471,0.0799630731,0.4260537624,0.2424503118,-0.1940700263,0.0907687545,-0.0451743342,-0.1737151593,0.4956526756,0.1084145382,-0.2056770921,-0.0616376214,-0.6093304157,0.270231694,0.1781580895,0.0563801825,0.1399827152,-0.1351262629,-0.2168394476,0.3453196883,0.3412940502,0.8837724328,-0.1166348606,0.3066751957,-0.1903944016,0.290443033,0.1443280578,-0.2256130725,-0.1689365655,0.1060050428,-0.1112610772,-0.1208685562,-0.0711652935,0.2940474153,0.2106103003,0.1386645734,0.0664766878,0.1077237725,0.1328757256,-0.0815856755,0.2378358096,-0.3102786541,-0.3429757059,-0.0317245722,0.068804644,0.0128027815,0.0409458764,-0.0688318163,-0.1117106229,-0.1718124598,-0.1071449891,-0.3116569221,-0.3785116971,-0.630314827,0.2572087944,0.1174918041,-0.1732238531,-0.0234634411,0.4163005948,0.2084572911,-0.0530585125,-0.1034444273,0.2119644731,0.1839005947,-0.0157125853,-0.432659775,-0.4276296198,0.3659292459,0.0070905215,-0.1030038148,0.0260977037,0.1720227748,-0.0189405978,-0.5434173346,-0.0373608135,0.3190795779,-0.2671540976,-0.3687651753,-0.0902485624,0.1615340859,-0.0545206368,0.0389946923,0.1653078645,-0.0085108606,0.1515002698,0.2937541306,-0.1488594562,0.0062973071,0.0661388785,0.1586945802,-0.3081633747,0.3967936933,0.105193153,0.2050501555,-0.0840641782,0.3106209636,0.2760282457,-0.2352599502,0.0393562168,0.0157346334,-0.066869393,-0.2480946034,0.3135075867,0.2914641798,0.0696792006,0.2083567232,-0.4812285006,-0.3398098946,0.3166283667,-0.0828945413,0.3896735907,-0.0731636807,0.2449611425,-0.1733464599,0.3347526789,-0.230848819,0.2501249909,0.0151179833,0.0752387717,0.0951796472,-0.1094321832,-0.3244827688,0.2187314034,0.0606943667,0.0490765199,-0.4533315897,-0.0843921825,-0.1837477535,0.218981877,0.1569612473,-0.1598668694,0.0279851034,-0.2815692723,-0.0500287898,-0.1155189201,0.250013262,-0.1798022985,-0.3040290177,-0.0474806651,-0.2357627749,-0.0228258427,-0.0549592003,0.2069177926,0.0280764531,0.115918234,0.4483347833,0.4594790339,-0.0680645034,-0.0378307626,0.013896279,0.2380716056,0.2289350033,0.0513937771,0.2313739508,-0.1708082408,0.2076285928,0.137510851,0.028159691,0.3660909534,-0.3685293198,0.1411535591,0.1458617896,0.1122729778,-0.229150027,0.0470962599,0.4093641937,-0.1366433203,0.1385902166,0.3364864588,0.3882443607,0.2981243432,-0.3183271885,0.0696678162,0.2068975717,0.0656154454,0.0791336149,0.1483144313,-0.2952793241,0.0485528596,0.1965855658,0.4680413604,0.1562707722,0.3953897655,-0.3202670515,0.331864506,0.0507706627,-0.1263484359,0.2346793711,-0.6594438553,0.3382435739,0.3807473481,0.0446405448,0.4135000706,-0.1509511918,0.1000894234,-0.1120816618,-0.3174078166,-0.130982399,0.147559911,0.2600157857,-0.1045529321,-0.4088055193,-0.3317551315,-0.5329182744,-0.0171641484,-0.0346194953,-0.5544852614,0.1135575026,-0.090222314,-0.1384177953,-0.293238759,-0.0163319036,0.2524356246,0.0160816405,-0.3610450327,0.2702174485,0.3793795109,-0.0878633633,0.2712516189,-0.1305067688,0.2399335504,0.1435425878,-0.2139380425,0.2792160511,-0.0674316287,-0.2528710663,-0.1248967126,0.1193890795,0.0793664902,0.3809874952,0.1066610664,0.0743967965,-0.1191804707,-0.2352820039,-0.1579051614,-0.0798048228,-0.6143346429,-0.1036970913,-0.1857888252,-0.2372473329,-0.0926254392,0.0233082697,-0.3891319931,-0.592033267,0.469664216,-0.4804220796,0.1938942522,-0.0724908635,0.0098216804,-0.4569826126,0.0606333092,0.3807204664,0.0307696629,-0.0016719782,-0.0664622933,-0.3786933422,0.1966575533,-0.2449967712,-0.3075909615,0.02759156,0.261680007,0.2132133842,0.061095573,0.292225033,-0.0425607599,0.0887635946,0.2222103924,-0.2329732478,0.0256214328,0.0684981123,0.580391109,0.0448312908,-0.1350613683,0.0359882452,0.006601362,-0.0352454856,-0.1482955664,-0.2382979244,0.1179489046,-0.3659752607,0.3714723587,0.0298532732,0.2374375761,-0.0330595076,-0.0913998634,-0.3701099157,0.1386655569,0.0109939994,0.0326292366,0.1610576212,0.7133926153,-0.2639420331,0.2280156314,-0.0864303261,0.272918582,0.0647511706,-0.3519095778,-0.0529554375,0.2847110331,-0.263043046,-0.0027439254,0.178145811,0.3194311857,-0.0273101479,-0.1437446922,0.3396831155,-0.3283259273,0.4062032998,-0.2416460067,-0.456864059,-0.051278051,-0.0832161084,0.0859006122,0.4241687953,-0.2856220901,0.2749409974,0.3766796291,0.0576435328,-0.1017420813,0.3008750081,-0.0254476126,-0.4870834053,-0.0039990461,0.1762327105,-0.2739369571,-0.1938409954,-0.0407273956,-0.2906008065]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1810","title":"Add Hateful Memes Dataset","comments":"Hi @lhoestq,\r\n\r\nRequest you to check this once.\r\n\r\nThanks,\r\nGunjan","body":"## Add Hateful Memes Dataset\r\n- **Name:** Hateful Memes\r\n- **Description:** [https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set]( https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set)\r\n- **Paper:** [https:\/\/arxiv.org\/pdf\/2005.04790.pdf](https:\/\/arxiv.org\/pdf\/2005.04790.pdf)\r\n- **Data:** [This link](https:\/\/drivendata-competition-fb-hateful-memes-data.s3.amazonaws.com\/XjiOc5ycDBRRNwbhRlgH.zip?AWSAccessKeyId=AKIARVBOBDCY4MWEDJKS&Signature=DaUuGgZWUgDHzEPPbyJ2PhSJ56Q%3D&Expires=1612816874)\r\n- **Motivation:** Including multi-modal datasets to \ud83e\udd17 datasets.\r\n\r\nI will be adding this dataset. It requires the user to sign an agreement on DrivenData. So, it will be used with a manual download.\r\n\r\nThe issue with this dataset is that the images are of different sizes. The image datasets added so far (CIFAR-10 and MNIST) have a uniform shape throughout.\r\nSo something like \r\n```python\r\n datasets.Array2D(shape=(28, 28), dtype=\"uint8\")\r\n```\r\nwon't work for the images. How would I add image features then? I checked `datasets\/features.py` but couldn't figure out the appropriate class for this. I'm assuming I would want to avoid re-sizing at all since we want the user to be able to access the original images.\r\n\r\nAlso, in case I want to load only a subset of the data, since the actual data is around 8.8GB, how would that be possible?\r\n\r\nThanks,\r\nGunjan","comment_length":10,"text":"Add Hateful Memes Dataset\n## Add Hateful Memes Dataset\r\n- **Name:** Hateful Memes\r\n- **Description:** [https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set]( https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set)\r\n- **Paper:** [https:\/\/arxiv.org\/pdf\/2005.04790.pdf](https:\/\/arxiv.org\/pdf\/2005.04790.pdf)\r\n- **Data:** [This link](https:\/\/drivendata-competition-fb-hateful-memes-data.s3.amazonaws.com\/XjiOc5ycDBRRNwbhRlgH.zip?AWSAccessKeyId=AKIARVBOBDCY4MWEDJKS&Signature=DaUuGgZWUgDHzEPPbyJ2PhSJ56Q%3D&Expires=1612816874)\r\n- **Motivation:** Including multi-modal datasets to \ud83e\udd17 datasets.\r\n\r\nI will be adding this dataset. It requires the user to sign an agreement on DrivenData. So, it will be used with a manual download.\r\n\r\nThe issue with this dataset is that the images are of different sizes. The image datasets added so far (CIFAR-10 and MNIST) have a uniform shape throughout.\r\nSo something like \r\n```python\r\n datasets.Array2D(shape=(28, 28), dtype=\"uint8\")\r\n```\r\nwon't work for the images. How would I add image features then? I checked `datasets\/features.py` but couldn't figure out the appropriate class for this. I'm assuming I would want to avoid re-sizing at all since we want the user to be able to access the original images.\r\n\r\nAlso, in case I want to load only a subset of the data, since the actual data is around 8.8GB, how would that be possible?\r\n\r\nThanks,\r\nGunjan\nHi @lhoestq,\r\n\r\nRequest you to check this once.\r\n\r\nThanks,\r\nGunjan","embeddings":[-0.2784939408,0.0814017504,-0.1171806604,0.0508640818,0.1218710616,0.276519537,0.4314359426,-0.0582628846,-0.167802453,-0.1871521175,-0.0217326749,-0.2919866145,-0.4143495858,0.2870904803,-0.033073999,-0.1282305717,0.0199136417,0.0501642153,-0.1291908473,-0.0971455202,-0.0427190736,-0.2187714428,0.173680082,-0.3291632533,-0.5686940551,-0.0909399018,-0.348364681,0.2094183564,-0.2482506782,-0.4358717501,0.039356865,0.3886331916,0.1061591655,0.4965772033,-0.000119284,0.0746740848,-0.0749180019,-0.2010511458,0.1137956455,0.0826198086,0.0753227621,-0.5928841233,-0.1401996017,-0.2102022022,0.1012603268,-0.510445416,0.1682695299,-0.0584354103,0.6137468815,0.0075588389,0.139414683,-0.1055491194,0.1667633355,-0.0106527926,0.0693166181,0.6924908757,-0.0665753558,-0.1911941618,0.1077478975,0.0660819113,0.2350669354,0.2974582613,-0.1187427714,0.1030942649,0.4525357783,-0.1753823161,-0.0354101732,-0.409506768,0.0342921503,0.3599967062,0.3359134495,-0.0396119319,-0.3183224499,-0.4553354383,0.0132362023,0.0306646936,-0.151593864,0.3057699203,-0.0138108954,-0.08980266,-0.3847868145,-0.0846828148,-0.3367912173,0.434112668,-0.0489613973,-0.0879441127,0.034902934,0.3317888677,-0.0592961982,-0.3138483763,-0.0603132844,-0.0252972059,0.2958116829,0.3043890297,0.1851316094,0.0283862501,-0.2041894048,-0.1559091508,0.4950287342,-0.2129442841,-0.0661781877,-0.0665557459,-0.5394417644,-0.1275184453,0.2359757423,-0.2005809098,-0.3806651533,0.3621714711,0.1018306985,0.0286470465,-0.113427043,-0.0025703583,-0.0937304348,0.0333879292,-0.0296275895,-0.0365022682,0.2669765055,0.0226449836,-0.1991434693,-0.2270829678,0.1668164134,0.0560742207,-0.0957354605,0.1779061407,0.0842486098,0.0024989485,-0.0769132003,0.2755938768,0.0097583178,-0.1479379237,-0.1440334618,-0.026918212,-0.1026793495,0.2190553993,0.287399292,0.2834021151,0.3031592369,0.0154273221,0.0191055983,0.1249133945,0.1775186211,-0.1737806499,0.1722967625,0.0936201215,0.0566389933,-0.0486712083,0.01132139,0.127223298,-0.1590442955,0.1474415362,-0.3003194034,0.0015120277,-0.0861698315,0.0702593997,-0.2898094654,0.1186713129,-0.0598903783,0.389084518,0.0154211037,-0.1127947569,0.1350811422,-0.0819616169,-0.2953154743,-0.3763979971,0.0435508676,0.6135712266,-0.4192544222,-0.0516589954,-0.3323787451,0.0061297016,-0.322586298,0.1256648153,-0.02663403,0.0554464012,-0.1314493269,-0.0971013457,0.237860173,-0.1488356143,-0.3436804414,0.1717283428,0.0583634228,0.279958576,0.144118309,0.4454452693,0.1607640833,-0.0458705537,-0.0028893275,0.6028018594,-0.2660286129,-0.050549943,0.2112982869,-0.4136244953,0.0722670704,0.3828107417,0.2895820737,-0.2180482894,0.381944716,-0.1611184776,0.1356556565,-0.2465438843,0.1778133065,0.0793891028,0.3348071873,0.0307530109,0.105018273,-0.2077444196,-0.6249386668,0.3057836294,-0.1715692431,-0.1472455561,-0.1707968712,-0.1145882532,-0.2591233552,-0.1613288522,-0.5535222292,-0.1132655889,0.0612551123,0.1008557901,0.0584220178,-0.0939020365,-0.086090751,0.3152758181,0.1091795936,0.0663261414,0.0743795186,0.3624193668,0.2816004753,-0.2374018133,-0.1371991783,0.1748521477,-0.0096447654,-0.2578362226,0.0109125013,0.4542725086,0.2580161989,0.2449667901,-0.0729105473,0.4916996062,0.4144461751,-0.4045648575,0.411591351,0.2688371241,0.1608158052,-0.2061175257,-0.1133594364,0.4534238279,-0.0806227922,0.0604933687,-0.1168305576,-0.1439986378,0.1038985103,0.2003336549,-0.0963575318,-0.5034181476,-0.0844112784,0.3765295446,-0.0532421768,-0.0430696383,0.0376173593,-0.0746209025,0.1867621094,0.1348380595,0.1705276221,0.1147266403,-0.2141565084,0.2376301736,0.1369387954,-0.0955501571,0.5169094801,0.2064987719,-0.269109875,0.2124691755,0.0892676339,0.0939551964,-0.1103274375,0.0852453858,0.1232639998,0.2039515078,0.3023958504,-0.1052987874,-0.307315439,-0.3477516472,-0.0219573546,-0.1481483132,0.0520602763,-0.1748444885,-0.19435139,-0.0518179461,-0.2235263288,0.0137507971,-0.3581023812,0.2197589874,0.251740545,-0.1229720488,0.004625516,0.1032064855,-0.1632249504,0.4921138287,0.2080518603,-0.1056273431,0.1926105022,0.2745219171,-0.1810999662,0.0173437241,0.2801067531,0.0080717914,0.3141483068,-0.2298336625,0.3383669257,-0.3133255541,-0.5422759056,0.0081714904,-0.1835669279,0.474847883,0.1163194478,0.0629600957,0.1557949632,0.0283132549,0.2824357152,0.111193791,-0.0868982002,-0.3114109039,-0.3310749233,-0.0448047481,0.0797403678,0.1611577719,-0.3245593905,-0.2356306314,0.4755323231,0.5631926656,0.1013915315,0.2133185416,0.5245591402,-0.2005772442,0.1874780804,0.2109723091,-0.0834284425,-0.0742029175,0.4681355655,-0.2427838147,-0.2984548807,-0.0182336364,-0.1934221536,-0.2464924008,0.3613251448,-0.4977895319,-0.1607376784,-0.1256053746,0.2048348039,0.0037161815,0.0290302671,0.0460298993,-0.0876278207,0.1066225171,-0.0536339059,-0.0972507074,-0.0121934554,0.3957065046,0.1893548518,0.0444887988,-0.0582862794,0.2635527253,0.5308299661,0.0392647833,-0.0762944594,0.469468534,0.0124964099,0.2690454423,-0.0065538986,-0.3644984663,0.3330453038,0.1659679562,-0.1023775861,0.1263072044,0.1117953286,0.0895957053,0.018387448,0.3785131574,-0.1399179548,-0.1693176776,0.234740451,-0.0377422944,0.1974469721,-0.2071605325,-0.2668409944,-0.742030561,-0.2528631091,0.0322723538,0.3284542561,-0.2526260316,0.236709848,-0.6186631918,0.0087223714,0.1558362842,0.3957827985,0.1414353997,0.0712128505,-0.070176512,-0.0717557967,-0.0236803088,0.0456275344,0.7164163589,0.0331753418,-0.3336513638,-0.010801509,0.216945678,-0.3836071491,-0.2909134328,0.1045939401,-0.1113911346,-0.133508563,0.6956555247,-0.2288880646,0.0639157593,0.245055899,0.3116856515,-0.1739060581,-0.2551371455,-0.1219469458,-0.3590189219,-0.449098438,-0.1257053614,0.1681126505,0.0371899195,0.1268436462,-0.4124236405,-0.0808386654,-0.1752023995,0.172907576,0.0824552998,0.0376102626,0.2756085396,-0.0311951134,0.1473056674,-0.1608462036,0.2567861974,0.4503074288,0.0308590941,-0.2251045555,0.0561484434,-0.1533210278,0.0972236246,0.4323202968,0.2592263818,-0.1872834712,0.0524288416,-0.057053607,-0.1778898835,0.4695497751,0.1623548418,-0.2195246369,-0.0707867444,-0.6017794013,0.2478765398,0.1750525236,0.0646978617,0.1027368903,-0.1319257915,-0.2219133675,0.3639348745,0.3374022543,0.9152295589,-0.1137349159,0.3171519339,-0.2031781375,0.2536638975,0.1205560043,-0.2464502603,-0.1673814654,0.1054662615,-0.0725407153,-0.1067248285,-0.061635308,0.3144361973,0.2085865587,0.0894151405,0.0426700264,0.10198576,0.1100287139,-0.0891751647,0.2073322684,-0.3218770027,-0.3185016215,0.0272090808,0.0739697069,0.0052120225,0.0301331282,-0.0887863785,-0.1200070679,-0.1743078232,-0.0962682962,-0.3293294013,-0.3941262066,-0.5755335093,0.2470410466,0.119115673,-0.15975371,-0.0178331081,0.3886505067,0.2010442168,-0.036690969,-0.1033431888,0.2095346898,0.207958281,-0.0193905383,-0.3969162107,-0.4398912191,0.4340967238,0.0057099145,-0.1313991398,0.0235536303,0.1801558733,-0.0436950177,-0.5931692123,-0.0816349238,0.3612248003,-0.2871821225,-0.3724399209,-0.0733290985,0.1464447975,-0.0567342602,0.0435459353,0.1270923167,0.0412792824,0.2108997256,0.2954786718,-0.1356407106,0.005618134,0.0417881869,0.1943236291,-0.2882255912,0.3771695793,0.070153974,0.2209883034,-0.0883200765,0.34707129,0.2490863949,-0.2342849523,0.0300331004,0.0238232557,-0.0543433316,-0.2480918616,0.2600035667,0.2851164639,0.090788886,0.1882717907,-0.4660252333,-0.3511835635,0.2919216752,-0.055611942,0.3810684085,-0.0734068006,0.2796344459,-0.1639407128,0.312592864,-0.2353004813,0.2512954772,0.0328093357,0.0496061146,0.0999261439,-0.0796863213,-0.3266205788,0.2219831645,0.0692587718,0.0209993999,-0.4290067255,-0.075580433,-0.1697085947,0.2219977677,0.1606767029,-0.113319248,0.0373120047,-0.2978625298,-0.0551404469,-0.0852779821,0.2179474086,-0.1587828547,-0.3339323401,-0.0077671893,-0.2903627455,-0.0191479865,-0.0194822438,0.1871196777,0.0290460326,0.1066834554,0.4522794187,0.4573124051,-0.0930792317,-0.0453871824,0.0382180586,0.2436520457,0.1809972078,0.0654570684,0.2425002754,-0.1657809019,0.2109259814,0.1212089881,-0.0005639871,0.3298000097,-0.38225317,0.1369211525,0.1501855701,0.1139659807,-0.2458815128,0.0535357557,0.4125860631,-0.091999799,0.1188107803,0.3403544128,0.368545413,0.3145181835,-0.3251480758,0.0898464695,0.1966227293,0.0430450998,0.1100468561,0.1641660184,-0.2905448377,0.0803458393,0.2107965946,0.447817266,0.1617069393,0.4417580068,-0.3243278861,0.3492729962,0.0514116585,-0.174683556,0.3023401797,-0.6140472293,0.3257373273,0.3558625579,0.0817890316,0.4533123076,-0.1464973688,0.107705839,-0.0812882856,-0.3499694467,-0.1276538372,0.1550839245,0.251401633,-0.092284672,-0.3808583617,-0.3438471258,-0.549279809,-0.0425754711,-0.0327404924,-0.5213015079,0.1346989423,-0.1156951189,-0.1454174072,-0.2550783157,-0.0056173769,0.2281589657,0.0242632721,-0.3506049216,0.2080070078,0.4110035002,-0.0920915976,0.2544584572,-0.1232949048,0.2627692223,0.1428430527,-0.2389777452,0.286975354,-0.0896528885,-0.2665422559,-0.1037312001,0.1139600053,0.0727670118,0.3466344178,0.0849254951,0.074311018,-0.1220170781,-0.2189469486,-0.1204492599,-0.0576301366,-0.5980290174,-0.0885819644,-0.1827259213,-0.236931175,-0.118357867,0.041256167,-0.3944420218,-0.6223079562,0.5188286901,-0.4719315767,0.1972251087,-0.0871896744,0.0061508683,-0.4353380799,0.0472745225,0.370457679,-0.0066959038,0.0316906124,-0.0603925288,-0.3696496487,0.2024715543,-0.2445918322,-0.3120775521,-0.0083426712,0.2451507598,0.2217405736,0.0361868478,0.3277430832,-0.1004993692,0.0983187854,0.2265036255,-0.2412224561,0.010816413,0.063288033,0.5402448773,0.0629119873,-0.1471966207,0.0201961584,-0.0133875031,-0.0232526641,-0.1428967863,-0.2542994022,0.116732195,-0.3374703825,0.3647206128,0.0576650724,0.2491661906,-0.0288138222,-0.0921088383,-0.3389089704,0.1473403126,0.0052462951,0.019022299,0.1440438628,0.7112110853,-0.2767018676,0.2400478274,-0.048737444,0.2905966938,0.0778291151,-0.3753023148,-0.0906603336,0.302398473,-0.2516911328,-0.003763987,0.1538396627,0.3128836453,-0.0214533992,-0.1289131343,0.3363265693,-0.3302836716,0.4167709947,-0.1775719225,-0.4630618393,-0.0356225818,-0.1016087011,0.0431656986,0.4428947866,-0.2830368578,0.2521986067,0.3789921999,0.0789407045,-0.0520297103,0.3073398471,-0.0033950503,-0.4619597793,0.0069293887,0.1848960519,-0.3040081859,-0.1815333813,-0.0119439997,-0.2782369554]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1810","title":"Add Hateful Memes Dataset","comments":"Hi @gchhablani since Array2D doesn't support images of different sizes, I would suggest to store in the dataset the paths to the image file instead of the image data. This has the advantage of not decompressing the data (images are often compressed using jpeg, png etc.). Users can still apply `.map` to load the images if they want to. Though it would en up being Sequences features.\r\n\r\nIn the future we'll add support for ragged tensors for this case and update the relevant dataset with this feature.","body":"## Add Hateful Memes Dataset\r\n- **Name:** Hateful Memes\r\n- **Description:** [https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set]( https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set)\r\n- **Paper:** [https:\/\/arxiv.org\/pdf\/2005.04790.pdf](https:\/\/arxiv.org\/pdf\/2005.04790.pdf)\r\n- **Data:** [This link](https:\/\/drivendata-competition-fb-hateful-memes-data.s3.amazonaws.com\/XjiOc5ycDBRRNwbhRlgH.zip?AWSAccessKeyId=AKIARVBOBDCY4MWEDJKS&Signature=DaUuGgZWUgDHzEPPbyJ2PhSJ56Q%3D&Expires=1612816874)\r\n- **Motivation:** Including multi-modal datasets to \ud83e\udd17 datasets.\r\n\r\nI will be adding this dataset. It requires the user to sign an agreement on DrivenData. So, it will be used with a manual download.\r\n\r\nThe issue with this dataset is that the images are of different sizes. The image datasets added so far (CIFAR-10 and MNIST) have a uniform shape throughout.\r\nSo something like \r\n```python\r\n datasets.Array2D(shape=(28, 28), dtype=\"uint8\")\r\n```\r\nwon't work for the images. How would I add image features then? I checked `datasets\/features.py` but couldn't figure out the appropriate class for this. I'm assuming I would want to avoid re-sizing at all since we want the user to be able to access the original images.\r\n\r\nAlso, in case I want to load only a subset of the data, since the actual data is around 8.8GB, how would that be possible?\r\n\r\nThanks,\r\nGunjan","comment_length":87,"text":"Add Hateful Memes Dataset\n## Add Hateful Memes Dataset\r\n- **Name:** Hateful Memes\r\n- **Description:** [https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set]( https:\/\/ai.facebook.com\/blog\/hateful-memes-challenge-and-data-set)\r\n- **Paper:** [https:\/\/arxiv.org\/pdf\/2005.04790.pdf](https:\/\/arxiv.org\/pdf\/2005.04790.pdf)\r\n- **Data:** [This link](https:\/\/drivendata-competition-fb-hateful-memes-data.s3.amazonaws.com\/XjiOc5ycDBRRNwbhRlgH.zip?AWSAccessKeyId=AKIARVBOBDCY4MWEDJKS&Signature=DaUuGgZWUgDHzEPPbyJ2PhSJ56Q%3D&Expires=1612816874)\r\n- **Motivation:** Including multi-modal datasets to \ud83e\udd17 datasets.\r\n\r\nI will be adding this dataset. It requires the user to sign an agreement on DrivenData. So, it will be used with a manual download.\r\n\r\nThe issue with this dataset is that the images are of different sizes. The image datasets added so far (CIFAR-10 and MNIST) have a uniform shape throughout.\r\nSo something like \r\n```python\r\n datasets.Array2D(shape=(28, 28), dtype=\"uint8\")\r\n```\r\nwon't work for the images. How would I add image features then? I checked `datasets\/features.py` but couldn't figure out the appropriate class for this. I'm assuming I would want to avoid re-sizing at all since we want the user to be able to access the original images.\r\n\r\nAlso, in case I want to load only a subset of the data, since the actual data is around 8.8GB, how would that be possible?\r\n\r\nThanks,\r\nGunjan\nHi @gchhablani since Array2D doesn't support images of different sizes, I would suggest to store in the dataset the paths to the image file instead of the image data. This has the advantage of not decompressing the data (images are often compressed using jpeg, png etc.). Users can still apply `.map` to load the images if they want to. Though it would en up being Sequences features.\r\n\r\nIn the future we'll add support for ragged tensors for this case and update the relevant dataset with this feature.","embeddings":[-0.2504210472,0.15920645,-0.1446140856,0.0486409999,0.1048059687,0.2664602399,0.444988817,-0.0732126981,-0.0832349584,-0.2253455967,-0.0230498239,-0.2821851373,-0.5021380186,0.2475005686,0.0156487133,-0.1245207638,0.0323905088,0.0315233096,-0.1846613884,-0.0556139722,-0.0390983522,-0.1718231291,0.1833574176,-0.2621174455,-0.5729023218,-0.0794128329,-0.2922713757,0.2743028998,-0.1746696681,-0.4757303596,0.0342119224,0.3668724895,0.1425356716,0.5114071369,-0.0001180374,0.0613285005,-0.0814565793,-0.1974319369,0.0460136309,0.0979811102,0.0788964629,-0.5317852497,-0.1205243692,-0.2593847811,0.1425022781,-0.4997737706,0.2076667249,-0.0505467318,0.6721628904,-0.0005184649,0.1403406113,-0.1678145826,0.2236776948,-0.029542353,0.0302443169,0.7310441732,-0.0535114408,-0.2448932976,0.0850695521,0.1053892821,0.2414755225,0.2661069632,-0.1824454665,0.0981596634,0.4686677754,-0.1495052427,-0.0850304961,-0.3651400805,0.0692548603,0.3533005714,0.3158091307,-0.1157739684,-0.2830895483,-0.4789533019,-0.0519871153,0.0470750518,-0.0973815098,0.2933705449,-0.0024280832,-0.0083623491,-0.4078708887,-0.1113797426,-0.3908492029,0.4031023383,-0.0128045585,-0.1113748103,0.0441607237,0.3313264251,-0.0470419489,-0.3176805675,-0.0503538102,-0.0321193412,0.3344113231,0.2942711115,0.2471242994,0.0357241593,-0.233834669,-0.2644758224,0.4943545759,-0.2746843398,-0.0365184322,-0.0611696504,-0.5127977729,-0.1091958955,0.1728673279,-0.2186565399,-0.4133867621,0.2884222865,0.142345041,0.0239077061,-0.0680184364,-0.0000656558,-0.1721003503,0.0852599442,-0.0206120834,-0.0413677394,0.2201980501,0.0065697813,-0.2225710154,-0.2236707956,0.1090388522,0.0613116771,-0.079758428,0.1821876615,0.0928377137,-0.0467927866,-0.069164969,0.2312929332,0.0642930046,-0.1113332435,-0.1213239431,-0.0465073399,-0.0835090652,0.2189407796,0.3173370957,0.4066416323,0.3729762435,-0.050235115,0.0361435004,0.1877490431,0.1703289747,-0.153628394,0.239740625,0.102562815,0.1239588559,-0.0443040691,-0.0259082131,0.1373515576,-0.1783611476,0.1234838665,-0.296526581,-0.0349515602,-0.0787504464,0.0607702471,-0.3073102534,0.1393744498,-0.0952757224,0.3325945437,0.04459906,-0.0998574942,0.0798218846,-0.0691168234,-0.2916133106,-0.4136890173,0.0731455311,0.5717698932,-0.4198009372,-0.0388678946,-0.2598000467,0.0048770937,-0.3514771163,0.1411675215,-0.0738463476,0.0804831609,-0.142295748,-0.0395541564,0.1971670538,-0.1646540314,-0.2917400301,0.1448032856,0.0503726937,0.2691359818,0.199737072,0.4102457762,0.1679878533,-0.0923593566,-0.0039954693,0.6235691309,-0.2270171046,-0.0232366286,0.216776222,-0.4549306631,0.090089567,0.3774319291,0.3031720817,-0.2699362338,0.366201371,-0.0529149659,0.1597317755,-0.2786393166,0.1843125671,0.0729182139,0.3280514777,0.0044640056,0.0812786371,-0.1909705251,-0.5832839608,0.2914872468,-0.1301626563,-0.2425613403,-0.1890538931,-0.0745365024,-0.2226565182,-0.1695448756,-0.5431597829,-0.1439875215,0.080965206,0.0673680156,0.1173334271,-0.0855121985,-0.1075575128,0.3108456731,0.037763834,0.0376631394,0.0759399161,0.3562743664,0.2526936233,-0.2590902746,-0.1550349891,0.193331033,-0.0770662054,-0.2776979208,0.0310585853,0.4657758176,0.2533091009,0.1999029815,-0.0602821521,0.5921089053,0.3864941299,-0.4306572974,0.4246513844,0.2981195748,0.1676470041,-0.1962485909,-0.1448151022,0.416193068,-0.1086042821,0.0944967568,-0.0901602134,-0.1568721086,0.1208517626,0.1988417357,-0.0697348937,-0.4860266745,-0.1375582665,0.3124346435,-0.0352833606,-0.0333198868,0.0816607848,-0.0311006028,0.1444080919,0.1584203094,0.1950880587,0.1174780205,-0.2045092583,0.2176683545,0.1160252914,-0.0820590928,0.5058944821,0.2438028753,-0.2884396315,0.1893194914,0.0441551432,0.070334971,-0.0591774695,0.0718808621,0.1592857093,0.1814645529,0.2923966646,-0.097211875,-0.2935050726,-0.3649154603,-0.0464099757,-0.1391712874,0.0285090171,-0.2159312814,-0.1760035157,-0.0752429143,-0.1771730036,0.0619004332,-0.273111999,0.3103006184,0.2432677001,-0.0844474509,-0.0076462994,0.0986702889,-0.1790386438,0.4522792995,0.1903005987,-0.1124393195,0.2018894255,0.2738096416,-0.1722243875,0.023474969,0.3414861858,0.0180947911,0.3103942275,-0.2085749805,0.2875601947,-0.3016797006,-0.5269958973,-0.0177559536,-0.127862975,0.4184361398,0.1583189368,0.056980744,0.1466142535,-0.0490887575,0.2545942366,0.0839701518,-0.1261508763,-0.3579851985,-0.3124555051,-0.1176280752,0.087633431,0.1875389665,-0.293767482,-0.2434359193,0.4645284414,0.5125316381,0.1423606277,0.2152908146,0.5227329731,-0.156318441,0.1688376963,0.2114794105,-0.0623731613,-0.0816042498,0.4824644625,-0.2617923617,-0.2787492871,-0.0011846425,-0.200517118,-0.2126512676,0.363460511,-0.4871141911,-0.1780950576,-0.1202250049,0.1878046393,0.0320536643,-0.0260774158,0.0824709162,-0.0281694625,0.1132401004,-0.0663222373,-0.0998316407,0.0764128268,0.3912824988,0.2211259156,0.0891547129,-0.0360578895,0.2834275961,0.4973535836,0.0504425019,-0.0778694227,0.4475955069,0.059290342,0.2628367841,-0.002623196,-0.3571556509,0.3998303413,0.1978180408,-0.1814906299,0.1413678378,0.1234125346,0.0140401004,0.0465474389,0.3846581578,-0.195073992,-0.1501945853,0.2425009012,0.0356559493,0.2080605179,-0.2231541127,-0.2774920166,-0.7103016376,-0.2136514038,0.0608367771,0.2419238091,-0.2063893527,0.2661414742,-0.6109212637,-0.0451589338,0.1607521027,0.3234853148,0.1610220671,0.0686568841,-0.1207006425,-0.0736890435,-0.0431587324,0.0290513206,0.7224555016,0.0184899177,-0.3272473812,-0.0201015491,0.2369354069,-0.3815586865,-0.2712821364,0.1092837974,-0.0826962888,-0.1491669863,0.6568533778,-0.257314682,0.0512714721,0.2450477034,0.3219650388,-0.1487639844,-0.2222961485,-0.0937421173,-0.3714405894,-0.474139452,-0.1325502247,0.1794004291,0.0563347228,0.1760863662,-0.4480550587,-0.0918842927,-0.1943971217,0.1649773419,0.1058010757,0.0951880813,0.2612834573,-0.0345723517,0.1756406724,-0.1584342718,0.2575211525,0.3741614521,0.0439132638,-0.14284724,0.0949466377,-0.1685846001,0.1172895208,0.4597485363,0.2344309688,-0.2063049525,0.0265828222,-0.0773551837,-0.1809724867,0.4501093626,0.1412698478,-0.2658033371,-0.024455009,-0.5977189541,0.1939571202,0.1625718623,0.0387332141,0.083334744,-0.1662767082,-0.2910216153,0.3688751757,0.4148890078,0.889053762,-0.1004764438,0.3690149486,-0.1744827181,0.1872283667,0.029185297,-0.2127092183,-0.2151515633,0.102306895,0.0166737586,-0.1287180483,-0.061799638,0.309414655,0.2182050198,0.1082589999,0.0370074846,0.1305918396,0.145158276,-0.0833373442,0.1909997612,-0.2910300493,-0.3754475117,-0.0224435013,0.0718095079,0.0025818758,-0.0975055993,-0.1028366387,-0.1750881374,-0.1285846829,-0.0624645837,-0.3139395714,-0.4170061946,-0.617782414,0.3116016388,0.0959149748,-0.1583430618,-0.0477428176,0.4364581406,0.2387737483,-0.0216691215,-0.0999172553,0.1813478023,0.1393334568,0.0159077458,-0.384326756,-0.4722924531,0.3890893757,-0.0097460113,-0.0883987695,0.0408188924,0.1852222383,-0.0622617975,-0.6213223934,-0.1009950936,0.3510938585,-0.23717314,-0.3729324937,-0.0460395366,0.084594138,-0.0351510532,0.0482117385,0.1355676055,0.0513295271,0.2593280673,0.2925769389,-0.1346592903,0.014493905,0.0537607074,0.141553238,-0.248570174,0.3979207873,0.0191123914,0.2050881237,-0.1009723097,0.4004041255,0.2464225441,-0.2275554389,0.030112084,0.0558883548,-0.0923078433,-0.3010914028,0.1867818087,0.3047510684,0.112344943,0.1786678433,-0.420599997,-0.3385640383,0.3300171494,-0.0448731929,0.3595152497,-0.0959174931,0.2453262806,-0.1896498948,0.2075492889,-0.2435894161,0.2797915041,0.0342172906,0.0165742319,0.013361685,-0.0864679068,-0.3129191399,0.1869139671,0.0548303537,-0.0084563084,-0.4545390308,-0.081984818,-0.1728125364,0.216376245,0.150824517,-0.1186473072,0.0461593196,-0.3085326254,-0.0792784989,-0.0771323666,0.1993965507,-0.121066682,-0.3016906977,-0.0212108046,-0.2817974985,-0.0210441761,-0.0050857202,0.1918309182,0.0055685565,0.0947469249,0.4647138417,0.4612266719,-0.0899199173,-0.0705758333,0.1159430668,0.2631657422,0.132107079,0.0130428076,0.2711670995,-0.2114795148,0.1537314057,0.1622492969,0.0216147434,0.3007538319,-0.3699859977,0.1090492532,0.2009070069,0.121393837,-0.2129845023,0.1097905487,0.4337896109,-0.0872262195,0.101593852,0.3715549111,0.3916659951,0.2633987367,-0.2922078669,0.1028849557,0.2551817,0.0022372187,0.0483148322,0.1876870543,-0.3026909828,0.0690411553,0.174618572,0.444933027,0.2387215346,0.5049601197,-0.3436600864,0.4133921862,0.1104386076,-0.1653073728,0.2926430106,-0.5142848492,0.2334800661,0.3220683932,0.1003449336,0.4405691922,-0.141152218,0.1429871917,-0.0843481421,-0.3823216856,-0.1536344737,0.1817495972,0.2389148027,-0.1287261248,-0.3973871768,-0.3295617402,-0.5255417824,-0.0513877273,-0.0321119577,-0.5779091716,0.1415560693,-0.1235465854,-0.1133022532,-0.3197180033,0.0532001704,0.2641002834,0.0214331392,-0.3121998906,0.2355457693,0.4401602149,-0.0962105617,0.2448836416,-0.0619920194,0.2725956738,0.171217531,-0.2464679927,0.2377240211,-0.102248542,-0.2724087238,-0.1546006203,0.105309628,0.0545735843,0.3160356283,0.1064424515,0.0642451793,-0.1034053341,-0.2114142329,-0.1423077434,-0.0897925571,-0.4951949716,-0.1446505636,-0.1938596964,-0.1958304048,-0.0647402406,0.0566847362,-0.4392277598,-0.6049891114,0.4906970561,-0.5163568854,0.2248381227,-0.0296893846,0.0092393449,-0.4217175841,0.0520844385,0.3018870354,-0.0143103721,0.034555614,-0.0447396934,-0.3840903342,0.1674565822,-0.2783689797,-0.257851392,-0.0539681651,0.2680950463,0.2295307815,0.0208568908,0.3160776496,-0.1001035869,0.0873313546,0.2617773414,-0.2485550642,0.0713350996,0.1462796479,0.5599984527,0.0346909128,-0.0936441943,0.040840406,0.0030868128,-0.0220302474,-0.1176195592,-0.1751277745,0.055096969,-0.337094903,0.3505986631,0.0402449146,0.1881272048,-0.0477155484,-0.0568387099,-0.367163837,0.1420492381,-0.020248726,-0.0100389021,0.1543936431,0.7235084772,-0.3094791472,0.226092577,-0.0420955159,0.3157183826,0.0397889577,-0.3485491276,-0.0814389512,0.3176930547,-0.2902961671,0.0474669822,0.1136442572,0.2330194712,-0.038627699,-0.1395374388,0.3328942657,-0.3573359549,0.4361351132,-0.1375585347,-0.4307523668,0.0136343492,-0.0708092898,0.0960577354,0.4268599451,-0.2791130841,0.2198582441,0.345535934,0.0700151771,-0.0514796041,0.3274357021,-0.0098308688,-0.4515477419,-0.0036132946,0.1303824782,-0.3480070531,-0.2129986584,-0.0162922125,-0.3059490621]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1808","title":"writing Datasets in a human readable format","comments":"AFAIK, there is currently no built-in method on the `Dataset` object to do this.\r\nHowever, a workaround is to directly use the Arrow table backing the dataset, **but it implies loading the whole dataset in memory** (correct me if I'm mistaken @lhoestq).\r\n\r\nYou can convert the Arrow table to a pandas dataframe to save the data as csv as follows:\r\n```python\r\narrow_table = dataset.data\r\ndataframe = arrow_table.to_pandas()\r\ndataframe.to_csv(\"\/path\/to\/file.csv\")\r\n```\r\n\r\nSimilarly, you can convert the dataset to a Python dict and save it as JSON:\r\n```python\r\nimport json\r\narrow_table = dataset.data\r\npy_dict = arrow_table.to_pydict()\r\nwith open(\"\/path\/to\/file.json\", \"w+\") as f:\r\n json.dump(py_dict, f)\r\n```","body":"Hi\r\nI see there is a save_to_disk function to save data, but this is not human readable format, is there a way I could save a Dataset object in a human readable format to a file like json? thanks @lhoestq ","comment_length":101,"text":"writing Datasets in a human readable format\nHi\r\nI see there is a save_to_disk function to save data, but this is not human readable format, is there a way I could save a Dataset object in a human readable format to a file like json? thanks @lhoestq \nAFAIK, there is currently no built-in method on the `Dataset` object to do this.\r\nHowever, a workaround is to directly use the Arrow table backing the dataset, **but it implies loading the whole dataset in memory** (correct me if I'm mistaken @lhoestq).\r\n\r\nYou can convert the Arrow table to a pandas dataframe to save the data as csv as follows:\r\n```python\r\narrow_table = dataset.data\r\ndataframe = arrow_table.to_pandas()\r\ndataframe.to_csv(\"\/path\/to\/file.csv\")\r\n```\r\n\r\nSimilarly, you can convert the dataset to a Python dict and save it as JSON:\r\n```python\r\nimport json\r\narrow_table = dataset.data\r\npy_dict = arrow_table.to_pydict()\r\nwith open(\"\/path\/to\/file.json\", \"w+\") as f:\r\n json.dump(py_dict, f)\r\n```","embeddings":[-0.1064517573,0.2575984001,-0.0731424093,0.124281235,0.4906324446,0.2974198163,-0.1625235528,0.3496721387,0.0443107188,0.0258972384,-0.1533560306,0.3309133351,-0.4147045612,0.2707338035,0.2647757828,-0.1661130488,0.0718676373,-0.0046492941,0.0115184709,0.1498278081,-0.1541504264,-0.05089375,0.2294982523,-0.0256731845,-0.0389475413,-0.3820806444,-0.2350030839,-0.1319584846,-0.2951259911,-0.2115113586,-0.087023221,0.0582865924,0.1421895921,0.0619914234,-0.000117436,-0.090346083,0.0592741333,-0.1883002818,-0.4185785353,-0.1192436665,-0.257265389,-0.4672527611,0.2826541662,-0.388561368,0.0572322868,-0.5000993609,0.1672141254,-0.3590875864,0.4838615954,0.2706120014,0.2117938548,-0.1439215094,0.3098011613,0.2972798944,0.0524505749,0.6159294248,-0.3162021935,0.1544303447,-0.0757002309,0.0732607916,0.3559266031,-0.1744703054,-0.1545521617,-0.3159635663,0.3756185472,0.0782597363,-0.2154064029,-0.0293780938,-0.0379322022,0.2223854065,0.756688714,-0.7057831287,-0.5702959299,-0.0148927746,0.0362194851,-0.0766560137,-0.1443181634,0.5130928755,0.0262532253,0.2297064811,-0.1699939519,-0.5853190422,-0.3661932349,0.0704637915,0.3859952092,-0.1131257564,-0.1840336323,-0.0402981117,0.0515578464,0.0164258908,0.31370911,-0.176554203,-0.0960149541,0.1227475703,-0.150960952,-0.2741802037,-0.5675682425,-0.2052541971,0.0101159187,0.0844311565,0.4611463845,-0.0475819334,-0.0811262503,0.0940760672,0.3766885102,0.0881229118,0.188190043,0.0747268647,0.1809167713,-0.2470789105,0.0940365717,-0.2720762789,0.0188911464,0.1625400782,0.4624987841,-0.0868978351,0.0454677418,-0.2265895158,-0.0494800098,0.0183054637,-0.1780694127,0.1294333339,-0.3326537907,0.1597582102,0.0246283002,0.1252074242,0.1071007475,0.2388100177,0.1295621246,-0.0150892157,0.1851650476,0.1889404058,0.0421168618,0.1680559516,-0.0994377881,0.0223751292,0.0488014072,-0.0798074752,0.0742932633,-0.4312080145,0.3465843201,-0.1667003334,0.2300370783,0.2840815485,0.1565832645,0.0880093202,-0.0418890566,-0.5203971863,-0.2893023491,0.2648203373,0.0096722208,-0.1103153452,-0.3392949998,0.0516568497,-0.1047599241,-0.1274159104,-0.7118952274,0.0507395752,0.0460144728,0.0264461786,0.1481796503,-0.0401071124,-0.1091865525,-0.5510830879,0.1095230356,0.0285975505,-0.576321125,0.1701294929,0.0707032606,-0.0269818287,0.0793612823,0.2944976091,0.1584192812,0.3940740824,0.0567225516,0.1613829583,0.3168116212,-0.1142696813,-0.0339379646,0.3281467259,0.1536334455,0.0527911782,-0.2635242939,0.4900129139,0.4048589766,0.0110457633,-0.1125622839,0.0533574186,0.1434189975,0.0167620722,-0.1362846196,-0.3738783896,0.1573848575,0.0359619595,-0.3810765445,-0.0902455226,0.3390375078,-0.0026245373,0.302003473,-0.2979596555,0.261669904,0.5488736629,0.3298592567,0.0892501846,-0.1869285256,-0.2176185399,-0.2303956449,-0.1483852565,0.473742485,-0.1687740237,-0.1578415185,-0.3411399424,0.1417446733,0.0717261359,-0.1318061352,0.2105413675,0.0646364987,-0.2472129762,-0.1737648696,0.1570231616,-0.2253973186,0.3347906768,-0.0315161124,0.1102832705,-0.2123126388,0.1414784044,-0.004306261,-0.2985953689,0.1239468232,0.0093764998,0.0077000391,0.0071598659,-0.0397443697,0.3718758523,0.171789065,0.0548245013,0.0678471997,0.5374658704,0.0692658946,-0.217672348,0.1381903291,0.2048177421,0.218248263,-0.0328975357,-0.686059773,0.5157960057,0.0122280223,0.0096767591,-0.0234441217,0.2769929171,0.1521539092,0.1961560845,-0.1398269832,0.0576458164,0.0091584511,0.0627285913,-0.1090610996,-0.228632018,-0.4677338898,0.0491144024,0.5178409815,-0.3079947531,0.5596216321,0.0245202035,-0.0674514472,-0.119786121,0.139173612,0.1694332957,0.0988084748,0.0131214242,0.0084054992,-0.0152110709,-0.1322947294,0.0793197006,0.1636240035,-0.175905332,0.2189535797,-0.1926093251,-0.0880049616,-0.1375405788,-0.2838563323,0.2554503083,-0.1744093746,0.0286313854,-0.2344365269,0.0963485539,-0.1480449885,0.0690904409,-0.0803010464,0.0835794732,-0.1047159508,-0.266687721,-0.3607384562,0.1366749853,0.101683639,-0.1860567033,-0.068729274,0.2181332707,-0.1224072799,-0.0292222481,-0.2400660366,-0.0789850354,0.0528297499,0.0032243826,0.1677333415,0.2358342856,0.4595133662,-0.1864616275,0.3139979541,-0.3245364726,0.0425703526,-0.2352767736,-0.0733075216,0.2968516648,-0.0385017358,0.3675038218,-0.4057822526,0.0554117076,-0.0578178652,0.068049483,-0.1202371269,0.0670581311,0.1310899109,-0.1453340948,0.0033633357,0.1358358115,-0.331548214,-0.0918807089,0.5880922675,-0.0212842803,0.0851841718,-0.2015008628,0.1822723746,0.0390201882,-0.0594496615,-0.1134716719,0.1022591069,-0.4799207747,0.3875850141,-0.2450955063,-0.3918482363,0.2173522115,0.3804702163,-0.1191383526,-0.0605500191,-0.1390841752,0.0108353263,-0.1433646083,-0.1700054705,-0.053838037,0.1813925952,0.0168658532,0.2060139924,0.1637835056,0.064199768,-0.2167957276,0.1224522814,0.3177887499,0.0929094329,-0.0520332344,0.3663929701,0.0803732425,-0.0182100609,0.3591733277,0.2471932918,0.2278045565,-0.0486975871,0.4381061792,-0.1248669848,0.0292693451,0.0408791304,-0.3164154887,-0.1899912655,-0.102308929,0.1691313684,0.0362691209,0.2131179422,-0.0089108944,-0.2511673272,-0.2523018718,0.2345303595,-0.242293939,0.0265634898,-0.117832467,0.110016413,-0.0175662525,-0.1718044877,0.0806341395,0.3837936521,0.2663038373,0.0322684795,-0.5151160359,-0.2940559983,-0.5803810358,0.0998745188,-0.0295965411,0.3618458211,0.1022810638,0.1812073439,0.2147047371,-0.1119509041,0.3909142613,-0.1343432367,-0.2980668247,0.0561339483,0.162156418,-0.367141366,0.2348611057,-0.0473510027,0.2741597593,-0.2787133455,0.3191592395,-0.1746150106,-0.1206107587,-0.2288211733,0.3668919504,-0.4481759667,-0.1200767979,0.08216919,-0.421810776,-0.1823729724,-0.3053171933,0.151322335,-0.1954495758,0.0067475787,-0.1221501827,-0.334416151,0.0702108294,0.1866705716,-0.2903488874,0.2243078053,-0.0119879479,-0.1412697434,0.3285467625,0.3046223521,0.1024217159,0.305498153,-0.0795759335,-0.3287593722,-0.2732930481,-0.1670098007,0.3981633186,-0.0980879515,0.0966515094,-0.0320462249,0.2533678412,-0.242379114,-0.3861219585,-0.1294434518,0.0660814196,0.1474064738,-0.5840488076,-0.6087498665,0.3970062435,0.1133667529,0.0549411736,0.0138993915,0.1954510361,-0.1718693674,0.4643928707,0.3657189608,0.9020754695,-0.6044504046,0.5735934973,-0.0385324284,-0.6457203031,0.2766192555,-0.4204300344,-0.0273249447,-0.3712976873,0.1174391508,-0.118944034,-0.009325997,0.3070980906,0.2328091413,-0.2221037596,0.179843843,-0.1137190759,-0.1695739478,0.0838306844,0.2791963816,-0.1214254424,-0.1039378345,-0.1601458639,0.1506625563,-0.1781390458,-0.359647274,0.0895338506,-0.3730091751,-0.0610664524,0.2402313501,-0.2900500894,-0.0070625083,-0.1904812157,0.1023721695,-0.3404386938,-0.2851460278,0.1627357453,0.2513294816,0.2861890197,0.2359692454,-0.0347433127,-0.0771179423,-0.2271253765,0.0297578387,0.0466886163,-0.0615851544,0.1257041097,0.2127915025,-0.0325120725,0.2309788018,-0.0446460135,-0.3499729335,0.1107666269,-0.0269317683,-0.0427048244,-0.2084444463,-0.2369850725,-0.117841363,-0.1680610031,-0.1451389343,0.096828647,0.4093709588,0.314499557,0.4892159998,-0.047016155,-0.0696015805,0.0739859343,0.0777783021,0.0151667381,0.065084897,0.2172195464,-0.0387640595,-0.2438251078,-0.2203436792,0.6501064301,0.1738114506,0.005395324,0.2061900347,-0.2736006379,0.1468953341,-0.2225253135,0.0149073824,0.0925174952,-0.2587207556,-0.2671196163,-0.4189402461,-0.3676987886,0.1884374022,0.3738772571,0.2795002759,0.1300150305,-0.2986118197,-0.251231432,-0.108519569,-0.1967222691,0.0870152041,-0.0538908318,-0.0562731326,0.0787377432,0.2680081427,0.2168200463,-0.3074116409,0.0350421295,-0.0519607589,0.1876037866,-0.0996490866,-0.1900952905,0.130559355,0.0866737515,-0.1253448129,-0.1687789559,-0.4526858926,-0.0305360891,-0.152185142,-0.3153108954,0.3631752431,-0.0966382399,0.3008654118,0.01259128,0.1385028213,-0.0113984188,-0.0335729718,0.0614380687,0.3263465464,-0.1698084325,0.0586511493,0.2863140106,-0.0419736058,-0.118875809,0.1244864464,0.1370639354,-0.1231820211,0.3291630447,-0.1266667843,-0.0109349396,-0.0754383057,0.3813678622,0.3408978879,-0.1769037545,0.2275378555,0.2048679441,0.0786914229,0.1729987711,-0.1341187656,0.2587120533,-0.0603931099,-0.1156147867,0.1173647568,-0.1337970495,0.0944371596,-0.3814754784,-0.3169582188,0.0581980906,-0.1218342483,-0.1312621087,0.2519333661,0.1693173647,-0.1154477075,0.0012946581,0.2567517459,0.2531091869,0.6264767647,0.1446099281,0.111755088,0.5608848929,0.1677169204,0.0637161434,0.0563953333,0.0664289594,-0.1512662023,0.0879573599,0.3625525832,0.4243084192,0.0104288794,-0.002009758,-0.2219341099,0.0830376893,0.131235972,-0.0474900827,-0.0755933523,-0.0296517387,-0.0509904549,0.1968562305,-0.1094434485,-0.2953989506,-0.4354917705,0.0387592912,0.1948023289,0.4805167019,0.0781077594,0.3756712675,0.1052036136,0.1889464259,-0.2450515181,-0.2562160194,-0.0034827194,0.0284074657,-0.3446308076,0.1636253893,0.2867845297,0.0134552764,-0.1108854264,0.4924834669,-0.101163432,-0.1818975359,0.1060706154,0.0341399163,0.1640357077,-0.1392957717,0.3932145834,0.1226921231,0.1094272062,0.2109396905,0.0213190764,-0.0144721409,-0.0988774374,0.0922459513,-0.0188539084,-0.501326859,-0.127931878,-0.1039540693,-0.1271952689,-0.2156641036,0.5050705075,-0.0352518857,0.112487264,-0.0289675687,0.0482331738,0.0474047549,0.3016920388,0.4544300139,0.0897114053,-0.069518432,0.0308586452,-0.1595074236,-0.1504456252,-0.0818226933,-0.0607455336,-0.0279468354,0.0946064293,0.3062109351,0.3246448636,-0.0154573126,0.3501591384,-0.0966228917,0.4746274948,-0.1529736072,-0.1095174477,0.3460095823,0.396283567,0.0240954254,-0.4002616704,0.4055837989,0.2499224842,0.0269376691,-0.3491510153,0.3835625648,-0.3014108241,-0.0613452718,0.3309569657,0.069627896,-0.0606319234,-0.0214393083,-0.2752467692,0.2633811533,0.3521860242,0.0958076343,0.3177362978,-0.0724049211,0.3416817784,-0.2775278687,0.1618970186,-0.0880663171,-0.2432600409,-0.5080469847,-0.0426239483,-0.2744732499,0.1040451452,-0.1105254665,-0.0234463084,0.1629510522,0.2395677716,-0.2278751731,0.2146221697,-0.1298564821,0.3445456326,-0.0600248314,-0.0869837105,-0.2443918884,0.38534078,-0.126838997,-0.0440095477,-0.0251780134,-0.4408193827,-0.0154450042,0.330283016,-0.2197275907,0.0105376393,0.2405091524,-0.0074133771,0.2687164545,-0.0039453404,0.6607905626,-0.0080851214,-0.0002099035,-0.2892734706,-0.3808015883]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1808","title":"writing Datasets in a human readable format","comments":"Indeed this works as long as you have enough memory.\r\nIt would be amazing to have export options like csv, json etc. !\r\n\r\nIt should be doable to implement something that iterates through the dataset batch by batch to write to csv for example.\r\nThere is already an `export` method but currently the only export type that is supported is `tfrecords`.","body":"Hi\r\nI see there is a save_to_disk function to save data, but this is not human readable format, is there a way I could save a Dataset object in a human readable format to a file like json? thanks @lhoestq ","comment_length":61,"text":"writing Datasets in a human readable format\nHi\r\nI see there is a save_to_disk function to save data, but this is not human readable format, is there a way I could save a Dataset object in a human readable format to a file like json? thanks @lhoestq \nIndeed this works as long as you have enough memory.\r\nIt would be amazing to have export options like csv, json etc. !\r\n\r\nIt should be doable to implement something that iterates through the dataset batch by batch to write to csv for example.\r\nThere is already an `export` method but currently the only export type that is supported is `tfrecords`.","embeddings":[-0.2098388523,0.0492693856,-0.1076665297,0.0664265603,0.5001397729,0.2327695489,-0.1172598526,0.3574723303,0.0007521105,0.149274841,-0.1676979065,0.0527242981,-0.3711419106,0.4290185571,0.3048388958,-0.1157784685,0.0749913231,-0.1399861872,0.0092282267,0.1033091769,-0.1731934994,0.0031017275,0.185680151,-0.104470022,-0.0076001934,-0.4232765138,-0.2167166322,-0.2166220844,-0.1934353411,-0.1114805341,-0.1856017858,0.0026591034,0.3307574987,0.0372143053,-0.0001181302,-0.1858900487,-0.0230949037,-0.1615507007,-0.3787545562,-0.088904269,-0.2690251172,-0.2977554798,0.0627992153,-0.402382642,0.0958815962,-0.2471900433,0.4078348875,-0.3925501704,0.2636525631,0.2429079562,0.1875607669,-0.1289436668,0.2089093029,0.2194775045,0.0918987915,0.6853557825,-0.3673555851,0.0276128408,0.0916349739,0.187260136,0.4378953874,-0.2874055207,-0.1316048354,-0.3443317115,0.343981564,0.0561668612,-0.0894858018,-0.0855567008,-0.011304005,0.2284470648,0.8819463253,-0.5850656629,-0.4965889454,0.0562340692,-0.0367571339,-0.0845804811,-0.1327416897,0.4962075353,0.0031539435,0.2550381124,-0.3079487681,-0.6020662785,-0.4352656603,0.0158189349,0.4432986379,-0.0723043829,-0.1896705925,-0.04866229,0.0651546344,0.1475459933,0.3959632516,-0.1175903231,-0.1520476192,0.1304979324,-0.1174586117,-0.3553515971,-0.6162095666,-0.0790424421,0.0551156811,0.0091142301,0.3531559408,-0.081389986,0.0763224065,0.0985915363,0.4005744457,0.0226747524,0.3162567914,-0.0479763411,0.1818596721,-0.4053877592,0.1173192486,-0.2435553223,0.0092502451,0.1748861074,0.4574065506,-0.0047270153,-0.0768976063,-0.2648996413,0.0425273515,-0.0964032635,-0.0780492052,-0.0268151686,-0.4872848094,0.0815216228,0.0194435436,-0.1135362983,0.0742608681,0.0766617134,-0.0136129288,-0.1499937922,0.2075235695,0.1393625587,0.0309044737,0.1772511303,-0.1824786216,-0.0466998853,-0.0887516439,-0.0720064715,0.1316601634,-0.3586089313,0.3845713437,-0.2493529618,0.2336791158,0.0582666993,0.1374220252,0.1378741711,-0.1352057457,-0.5077313185,-0.3294370472,0.201381892,-0.1224831715,-0.1518190801,-0.3533756435,0.0441593826,-0.0617387444,-0.2341740727,-0.6611484289,0.1238480136,0.1232445017,-0.0381360948,0.0982439071,-0.0732536614,-0.1500760615,-0.4955060482,0.0272695906,-0.0362048224,-0.6494598389,0.188305527,0.1602357626,-0.0922067091,0.1351405829,0.3854272962,0.2524668276,0.5349962711,-0.0050936947,0.19084993,0.279283464,0.0055641709,0.1569122523,0.3973152041,0.1233955696,-0.1115937009,-0.2646249235,0.4624730647,0.5034453273,-0.1858539432,-0.1742169112,0.0487023927,0.1154660359,-0.0147593711,-0.1799330264,-0.349065721,0.0157871991,-0.0664079338,-0.2319363058,-0.0439234041,0.5080922246,0.0998144001,0.2476150542,-0.3435503244,0.3596270382,0.5124644637,0.3237612247,0.1429360509,-0.2747886479,-0.2920857668,-0.2649782598,-0.2304356694,0.3505921364,-0.2540549636,0.2577134073,-0.3342514634,0.2862891853,0.0189476907,0.0523889288,0.2952235043,0.0199203696,-0.0535012223,-0.1708802879,0.1500766724,-0.3343756497,0.2304334193,-0.0469821878,-0.0708581209,-0.1806437671,0.1290433705,0.006797886,-0.2635652423,0.0730489269,-0.1115209684,-0.0074959802,-0.1651372313,0.0253623724,0.305508852,0.0756190792,0.0746358484,0.1550768763,0.6029985547,0.0291009638,-0.0924293026,0.061812751,0.0457080118,0.0842041448,-0.0397253223,-0.732578516,0.4708710015,-0.0299039949,-0.0231565628,-0.0380999707,0.1662543118,0.0391301513,0.0894585475,0.0318249017,0.0888793096,0.1403706819,0.0402228944,-0.2641543448,-0.2821545303,-0.4200903773,0.0944668204,0.6116841435,-0.3029912412,0.4891398549,-0.0324115194,0.1629302353,-0.0735493004,0.1590601355,0.0599536747,0.1638482362,0.0063388962,0.1523171514,-0.0200324561,-0.100285612,0.0414518863,0.1809062213,-0.3330414295,0.1086975262,-0.1366849989,-0.273260951,-0.026308069,-0.2412550002,0.2050079405,-0.1510562897,0.0028075317,-0.253469497,-0.0020915945,-0.1439948231,0.0682080537,-0.0892927051,0.2924490869,-0.0737616271,-0.2302346081,-0.2821521163,0.2217075378,0.0110720415,-0.2715150714,0.0190669838,0.39490363,-0.2087052763,0.1366722435,-0.2805602252,-0.0104772337,0.0772219449,0.0013887953,0.1501663029,0.2121299207,0.5001707673,-0.2471012175,0.2344092876,-0.3675446808,-0.0073752617,-0.1600474566,-0.0063728634,0.3519592583,-0.101243414,0.3395303786,-0.4360646307,0.0627961382,-0.1589544564,0.0349303111,-0.0650752261,-0.074305594,0.1325968355,-0.2086403817,-0.0840452462,0.1942927539,-0.3629709482,-0.1273270398,0.5436404347,-0.0702461004,0.1053248942,-0.2141474187,0.0746374577,-0.0660374984,-0.1418339759,-0.1646412462,0.074482359,-0.6360775232,0.2971100807,-0.3602563441,-0.4670175016,0.2232372016,0.5383545756,-0.1780404299,0.0747439638,-0.2141808271,0.1051238477,0.0104011782,-0.2355577946,-0.0313845314,0.2207895517,0.0310314111,0.084916316,0.172157526,0.0622602627,-0.1802071482,0.0317772888,0.3285797238,0.2549753785,-0.0333706029,0.2822730243,0.2257814258,0.0145830298,0.2588456273,0.2226624489,0.1427477747,0.0221684892,0.4139010906,-0.0327833705,0.0022067367,0.347582072,-0.3031911254,-0.2985248268,-0.0196204558,0.0720028356,0.0848768726,0.1824378371,0.0174716525,-0.0904310346,-0.2348344028,0.3269669414,0.0030278773,0.1502865553,-0.1556913108,0.1154052839,-0.1132466793,-0.133651644,0.0767766312,0.3554423153,0.3367665708,0.1070488244,-0.6552957892,-0.2315314859,-0.485299468,0.190707311,0.0501582287,0.2119172215,-0.1267174929,0.1747836322,0.1891009808,-0.0203106459,0.409702152,-0.2429734468,-0.2298627645,0.0398616455,0.1058927327,-0.1899779886,0.1949257404,-0.0307499394,0.2844811082,-0.0784664601,0.3948440254,-0.1224789098,-0.1046812385,-0.183570683,0.261767298,-0.45138219,-0.2274119109,0.2144451886,-0.2858097851,-0.0793606043,-0.2414368391,0.1410573572,-0.2363556623,-0.0683966801,-0.1851714551,-0.2665023804,0.135086,0.1919200718,-0.3439517915,-0.0101495301,0.0290646441,-0.223308444,0.2827488482,0.3313415051,0.0968825147,0.0708976164,-0.1467025131,-0.229076162,-0.0741189569,-0.3246363401,0.5238562226,-0.1108002067,0.186069563,-0.070848152,0.3306650817,-0.1715370268,-0.3715516925,-0.2041895539,0.0513592027,0.1517584473,-0.7371711135,-0.4786507785,0.1942820549,0.016983524,0.0677904114,-0.0319631957,0.1035449654,-0.1954326332,0.2718692124,0.5431913137,0.9837471843,-0.6078290343,0.4128284156,-0.1637287587,-0.5005873442,0.328420043,-0.4307344854,0.0627746657,-0.3551385403,-0.0276474766,-0.1078198925,0.0088213412,0.2886626124,0.2476262748,-0.2568733394,0.1114459336,0.0429644138,-0.0276019517,0.0065248511,0.2042028308,-0.0899184123,-0.0616594627,-0.0040143942,0.1460142285,-0.1574330479,-0.3205996752,-0.0201734789,-0.3327824175,0.1395350248,0.4025574923,-0.202018708,-0.0209280886,-0.1463060528,0.0150448224,-0.3082629442,-0.2517386675,0.1647444218,0.2179530859,0.534599483,0.2676829696,-0.0082500223,-0.2146235108,-0.2167550772,0.0655893683,-0.0220923331,-0.0354238711,0.0939725041,0.1627648622,-0.0731025562,0.306450218,0.0288082194,-0.3392605186,0.0228164773,-0.079897359,-0.0847570822,-0.2946383655,-0.1190949678,-0.0714354366,-0.0903848186,-0.0838376954,0.0825991854,0.4026633501,0.3868673742,0.6266878843,-0.1138431877,-0.0243492872,-0.0264971647,0.073580578,0.0862438828,-0.1006961018,0.0623360313,-0.0058064307,-0.2779268324,-0.2418328524,0.6459102631,0.0221287515,-0.0729259402,0.2438969016,-0.3161091506,0.1931054741,-0.1895979494,0.1027759239,0.0620111898,-0.3043517768,-0.3431846201,-0.2934036255,-0.1822025925,0.1060878336,0.351362437,0.2742486894,0.1964376867,-0.2088546455,-0.1980812699,-0.071591191,-0.1946109086,-0.0148154441,-0.0851608515,0.0202354565,0.1642276198,0.2659725249,0.069085598,-0.2713037431,-0.0321456045,-0.0136918966,0.1524148732,-0.0729749054,-0.2203808427,0.144368276,0.0652039722,0.0645052269,-0.0626737624,-0.3862687349,0.0028053212,-0.1917511374,-0.3020140231,0.4181834161,-0.1112125218,0.3120131195,0.0823349506,0.1562836021,-0.0018334045,-0.0398665182,0.2709319293,0.4351328611,-0.2582455575,0.1065220907,0.2574400902,-0.0351615027,-0.120685406,0.1419337243,0.3585576117,-0.0054716305,0.291890204,-0.1320079565,0.021015333,0.0817457736,0.1922919303,0.223026976,-0.1513839811,0.2095308751,0.2021977603,0.0894815996,0.3666161299,-0.0980457217,0.280993849,-0.0219849534,-0.1846565604,0.174745664,-0.0888469741,0.0154188415,-0.0415839255,-0.3821218014,0.0678361133,-0.205319494,-0.1655510366,0.2034166008,0.1272655576,-0.0663162842,-0.0298959296,0.1465521753,0.2336004972,0.4929717183,0.2062358409,0.0732564479,0.4920814335,0.2002183646,0.0824682489,-0.0982073024,0.1632049829,-0.148412928,0.1188445315,0.4701641202,0.4780254066,-0.2022809088,-0.1375740469,-0.2050095946,0.060681425,0.0712559372,-0.0733529776,0.0407070853,-0.1687211543,-0.0187055729,0.2555471659,0.0180926602,-0.3080140054,-0.328976512,0.0334218554,0.2333346754,0.5413273573,0.1409212649,0.3044002056,-0.0077628242,0.1501619965,-0.2206456959,-0.1622949988,-0.1543229222,0.0149485208,-0.3174990714,0.3099464476,0.1703502685,0.058846198,0.0753085017,0.4118942022,-0.1623411179,-0.0831413269,0.0934861079,0.1500549614,0.071775429,-0.0524237156,0.4647702873,0.0988010019,0.1652533412,0.3295050263,0.0059086024,-0.1236521229,-0.134171769,-0.0283534881,0.1266488582,-0.463858515,-0.1727948934,-0.1395136416,-0.104898423,-0.1019242555,0.4397470653,-0.1383690536,0.1056913659,0.0689348057,0.0451065451,0.0810732469,0.3107047975,0.5162873864,0.031330213,-0.0684718639,0.0654620826,-0.0065769991,0.0252946429,0.0163489021,0.0962507278,-0.1207719147,0.1353666782,0.3248269856,0.348624438,0.0382465795,0.1583917141,-0.0753917992,0.5430313349,-0.1199696511,-0.1642115265,0.330296576,0.3895040154,0.0072048609,-0.3909868598,0.5046799779,0.283726722,-0.0130073493,-0.4221143425,0.4101163447,-0.2834379673,-0.2210488915,0.3415217698,0.068348892,-0.1018354595,-0.0187306963,-0.2133675963,0.2306762636,0.3925397992,0.1220635623,0.2381232679,0.0703169629,0.2417998463,-0.3717209101,0.0908149779,-0.1417985111,-0.1396602243,-0.4831582308,0.1779390574,-0.1455750465,-0.0111704264,-0.1442710608,-0.1653169841,0.3295235038,0.1953500658,-0.205749467,0.2056866586,-0.1991421282,0.5085456371,-0.1622638851,-0.093535237,-0.2657583952,0.2492876947,-0.1632487625,-0.0005450865,-0.0438026153,-0.3620735407,-0.0830063745,0.3840695024,-0.2108114958,0.1024110615,0.2996649444,-0.009946486,0.2332293987,0.0598658845,0.5324202776,0.1018796563,-0.051843103,-0.2351112813,-0.3200591207]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1805","title":"can't pickle SwigPyObject objects when calling dataset.get_nearest_examples from FAISS index","comments":"Hi ! Indeed we used to require mapping functions to be picklable with `pickle` or `dill` in order to cache the resulting datasets. And FAISS indexes are not picklable unfortunately.\r\n\r\nBut since #1703 this is no longer required (the caching will simply be disabled). This change will be available in the next release of `datasets`, or you can also install `datasets` from source.","body":"So, I have the following instances in my dataset\r\n\r\n```\r\n{'question': 'An astronomer observes that a planet rotates faster after a meteorite impact. Which is the most likely effect of \r\nthis increase in rotation?', \r\n'answer': 'C', \r\n'example_id': 'ARCCH_Mercury_7175875', \r\n'options':[{'option_context': 'One effect of increased amperage in the planetary world (..)', 'option_id': 'A', 'option_text': 'Planetary density will decrease.'},\r\n (...)]}\r\n```\r\n\r\nThe `options` value is always an list with 4 options, each one is a dict with `option_context`; `option_id` and `option_text`.\r\n\r\nI would like to overwrite the `option_context` of each instance of my dataset for a dpr result that I am developing. Then, I trained a model already and save it in a FAISS index\r\n```\r\ndpr_dataset = load_dataset(\r\n \"text\",\r\n data_files=ARC_CORPUS_TEXT,\r\n cache_dir=CACHE_DIR,\r\n split=\"train[:100%]\",\r\n )\r\ndpr_dataset.load_faiss_index(\"embeddings\", f\"{ARC_CORPUS_FAISS}\")\r\ntorch.set_grad_enabled(False)\r\n```\r\n\r\nThen, as a processor of my dataset, I created a map function that calls the `dpr_dataset` for each _option_\r\n\r\n```\r\ndef generate_context(example):\r\n question_text = example['question']\r\n for option in example['options']:\r\n question_with_option = question_text + \" \" + option['option_text']\r\n tokenize_text = question_tokenizer(question_with_option, return_tensors=\"pt\").to(device)\r\n question_embed = (\r\n question_encoder(**tokenize_text)\r\n )[0][0].cpu().numpy()\r\n _, retrieved_examples = dpr_dataset.get_nearest_examples(\r\n \"embeddings\", question_embed, k=10\r\n )\r\n # option[\"option_context\"] = retrieved_examples[\"text\"]\r\n # option[\"option_context\"] = \" \".join(option[\"option_context\"]).strip()\r\n #result_dict = {\r\n # 'example_id': example['example_id'],\r\n # 'answer': example['answer'],\r\n # 'question': question_text,\r\n #options': example['options']\r\n # }\r\n return example\r\n```\r\n\r\nI intentionally commented on this portion of the code.\r\n\r\nBut when I call the `map` method, `ds_with_context = dataset.map(generate_context,load_from_cache_file=False)`\r\n\r\nIt calls the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-55-75a458ce205c> in <module>\r\n----> 1 ds_with_context = dataset.map(generate_context,load_from_cache_file=False)\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in <dictcomp>(.0)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1257 fn_kwargs=fn_kwargs,\r\n 1258 new_fingerprint=new_fingerprint,\r\n-> 1259 update_data=update_data,\r\n 1260 )\r\n 1261 else:\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 155 }\r\n 156 # apply actual function\r\n--> 157 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 158 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 159 # re-apply format to the output\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 387 file = StringIO()\r\n 388 with _no_cache_fields(obj):\r\n--> 389 dump(obj, file)\r\n 390 return file.getvalue()\r\n 391 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 359 def dump(obj, file):\r\n 360 \"\"\"pickle an object to a file\"\"\"\r\n--> 361 Pickler(file, recurse=True).dump(obj)\r\n 362 return\r\n 363 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 452 raise PicklingError(msg)\r\n 453 else:\r\n--> 454 StockPickler.dump(self, obj)\r\n 455 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 456 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in save_function(pickler, obj)\r\n 554 dill._dill._create_function,\r\n 555 (obj.__code__, globs, obj.__name__, obj.__defaults__, obj.__closure__, obj.__dict__, fkwdefaults),\r\n--> 556 obj=obj,\r\n 557 )\r\n 558 else:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 784 write(MARK)\r\n 785 for element in obj:\r\n--> 786 save(element)\r\n 787 \r\n 788 if id(obj) in memo:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 885 k, v = tmp[0]\r\n 886 save(k)\r\n--> 887 save(v)\r\n 888 write(SETITEM)\r\n 889 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 885 k, v = tmp[0]\r\n 886 save(k)\r\n--> 887 save(v)\r\n 888 write(SETITEM)\r\n 889 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle SwigPyObject objects\r\n```\r\n\r\nWhich I have no idea how to solve\/deal with it\r\n\r\n","comment_length":63,"text":"can't pickle SwigPyObject objects when calling dataset.get_nearest_examples from FAISS index\nSo, I have the following instances in my dataset\r\n\r\n```\r\n{'question': 'An astronomer observes that a planet rotates faster after a meteorite impact. Which is the most likely effect of \r\nthis increase in rotation?', \r\n'answer': 'C', \r\n'example_id': 'ARCCH_Mercury_7175875', \r\n'options':[{'option_context': 'One effect of increased amperage in the planetary world (..)', 'option_id': 'A', 'option_text': 'Planetary density will decrease.'},\r\n (...)]}\r\n```\r\n\r\nThe `options` value is always an list with 4 options, each one is a dict with `option_context`; `option_id` and `option_text`.\r\n\r\nI would like to overwrite the `option_context` of each instance of my dataset for a dpr result that I am developing. Then, I trained a model already and save it in a FAISS index\r\n```\r\ndpr_dataset = load_dataset(\r\n \"text\",\r\n data_files=ARC_CORPUS_TEXT,\r\n cache_dir=CACHE_DIR,\r\n split=\"train[:100%]\",\r\n )\r\ndpr_dataset.load_faiss_index(\"embeddings\", f\"{ARC_CORPUS_FAISS}\")\r\ntorch.set_grad_enabled(False)\r\n```\r\n\r\nThen, as a processor of my dataset, I created a map function that calls the `dpr_dataset` for each _option_\r\n\r\n```\r\ndef generate_context(example):\r\n question_text = example['question']\r\n for option in example['options']:\r\n question_with_option = question_text + \" \" + option['option_text']\r\n tokenize_text = question_tokenizer(question_with_option, return_tensors=\"pt\").to(device)\r\n question_embed = (\r\n question_encoder(**tokenize_text)\r\n )[0][0].cpu().numpy()\r\n _, retrieved_examples = dpr_dataset.get_nearest_examples(\r\n \"embeddings\", question_embed, k=10\r\n )\r\n # option[\"option_context\"] = retrieved_examples[\"text\"]\r\n # option[\"option_context\"] = \" \".join(option[\"option_context\"]).strip()\r\n #result_dict = {\r\n # 'example_id': example['example_id'],\r\n # 'answer': example['answer'],\r\n # 'question': question_text,\r\n #options': example['options']\r\n # }\r\n return example\r\n```\r\n\r\nI intentionally commented on this portion of the code.\r\n\r\nBut when I call the `map` method, `ds_with_context = dataset.map(generate_context,load_from_cache_file=False)`\r\n\r\nIt calls the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-55-75a458ce205c> in <module>\r\n----> 1 ds_with_context = dataset.map(generate_context,load_from_cache_file=False)\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in <dictcomp>(.0)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1257 fn_kwargs=fn_kwargs,\r\n 1258 new_fingerprint=new_fingerprint,\r\n-> 1259 update_data=update_data,\r\n 1260 )\r\n 1261 else:\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 155 }\r\n 156 # apply actual function\r\n--> 157 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 158 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 159 # re-apply format to the output\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 387 file = StringIO()\r\n 388 with _no_cache_fields(obj):\r\n--> 389 dump(obj, file)\r\n 390 return file.getvalue()\r\n 391 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 359 def dump(obj, file):\r\n 360 \"\"\"pickle an object to a file\"\"\"\r\n--> 361 Pickler(file, recurse=True).dump(obj)\r\n 362 return\r\n 363 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 452 raise PicklingError(msg)\r\n 453 else:\r\n--> 454 StockPickler.dump(self, obj)\r\n 455 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 456 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in save_function(pickler, obj)\r\n 554 dill._dill._create_function,\r\n 555 (obj.__code__, globs, obj.__name__, obj.__defaults__, obj.__closure__, obj.__dict__, fkwdefaults),\r\n--> 556 obj=obj,\r\n 557 )\r\n 558 else:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 784 write(MARK)\r\n 785 for element in obj:\r\n--> 786 save(element)\r\n 787 \r\n 788 if id(obj) in memo:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 885 k, v = tmp[0]\r\n 886 save(k)\r\n--> 887 save(v)\r\n 888 write(SETITEM)\r\n 889 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 885 k, v = tmp[0]\r\n 886 save(k)\r\n--> 887 save(v)\r\n 888 write(SETITEM)\r\n 889 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle SwigPyObject objects\r\n```\r\n\r\nWhich I have no idea how to solve\/deal with it\r\n\r\n\nHi ! Indeed we used to require mapping functions to be picklable with `pickle` or `dill` in order to cache the resulting datasets. And FAISS indexes are not picklable unfortunately.\r\n\r\nBut since #1703 this is no longer required (the caching will simply be disabled). This change will be available in the next release of `datasets`, or you can also install `datasets` from source.","embeddings":[0.06904158,-0.3887840509,-0.0696357936,0.2468255013,0.1642265022,0.0277648028,0.1784276962,0.0541120283,0.4089650214,0.5476052165,-0.0500990711,0.4959974289,0.115793407,-0.1273343563,-0.267626375,0.0515106916,0.0849947408,0.0134227211,0.0216204002,-0.2425854504,-0.5357055068,-0.0484026559,-0.1984261274,0.0493396372,-0.0398413986,0.0180335399,-0.0332831852,0.0954476371,-0.0584461205,-0.4539258182,0.4045750499,0.257684648,0.2116607726,-0.0134229129,-0.0001266745,0.0679022521,0.2053468674,-0.021252377,-0.1646225452,0.086750403,-0.001575422,0.3058409989,0.2944130301,-0.354249239,-0.2179245949,-0.2550703585,-0.0504558645,-0.134555921,0.0680654272,0.1915952861,0.0670926794,0.1456792355,-0.1130726263,-0.1586890072,0.2215849906,-0.0109133879,-0.0620617904,0.1621502042,0.2719885111,0.3066604137,0.2544573843,0.2748863101,-0.2460707575,0.0368728079,-0.0725996867,0.0926401168,0.3378049135,0.0920075402,0.155671373,0.0969465524,0.2498126328,-0.2737818956,-0.4702080786,-0.4137730896,-0.061396569,-0.1567487866,-0.2627948523,-0.5509217381,-0.0172370858,0.2482920587,0.1254633218,0.1021858975,-0.0639564916,0.5212545395,0.4118613303,0.0359554738,0.0144240391,0.210933432,0.1066410914,-0.0576482825,0.1366034746,-0.1332527101,-0.0283684339,-0.1321200579,-0.2131488323,-0.1242142096,0.4846388996,-0.2354142964,0.245058015,0.0846258178,-0.1290584356,0.401452601,0.0691412911,0.107549645,0.1615174562,-0.118626982,-0.0983864889,0.0395724252,-0.117117323,0.0870361477,-0.1678486615,0.1142573133,0.0268994607,-0.0553063974,-0.631529808,-0.0129407663,0.3205948472,-0.1449027658,-0.1014321595,0.2506060302,-0.6014733911,0.34290874,0.1318702847,0.5309945941,0.0145976041,0.1166362464,0.1949887425,0.2470773607,-0.2960154414,0.4217503667,-0.2266984284,-0.0420274772,0.0346693248,-0.0588204525,-0.1218073964,0.4785673618,0.2849897146,-0.2606017888,0.097831212,-0.0784839243,0.0227258429,-0.1106626913,0.3640687764,0.1970125288,-0.0533642881,0.0772830248,0.1710085273,-0.0954404548,-0.1744895428,0.1174802035,0.1384007186,-0.1021388397,0.1903932989,0.0709636062,-0.1389862746,-0.1317988187,-0.2130447179,0.1211962253,-0.0878358111,0.2369573712,0.187057212,-0.1878823042,-0.0876030996,-0.1486045569,0.2090401649,0.5161315799,-0.2350449562,-0.0758556351,0.0244935937,0.1951778829,0.3501953185,0.2010139525,-0.326797992,0.256172508,0.0371250957,0.232217297,0.3163653016,-0.4490192533,-0.6639236212,-0.0901995748,-0.0756988227,0.0397732891,0.3317385614,0.0975898504,0.3703524768,0.2930629551,-0.1459473521,0.517279923,0.1696157157,0.0434674174,-0.2609554231,-0.344840616,0.0231414568,0.1270146668,-0.0733682066,0.1632385254,-0.0440482982,-0.2676261961,0.3614017367,-0.1503375173,-0.1021282971,0.2690655887,0.5216957331,0.2159873843,0.2130895555,-0.2032638788,-0.0844261721,0.1282851845,-0.1522228718,0.0228898693,-0.0782172158,-0.2090477645,-0.0135675445,0.1977491081,-0.3135472238,-0.1740483791,-0.0746521577,0.1529769897,0.2881989479,-0.1318628043,-0.0989452079,0.049343016,-0.339966625,0.2049800456,-0.5214464068,0.4270998836,-0.0935477465,-0.3137369156,-0.187201947,0.2583495975,0.0134848785,-0.1509953737,-0.158976227,0.1067028642,0.4748960435,-0.4063611031,0.1756881326,-0.3079205453,-0.0195666682,-0.1098886207,0.1134148315,0.0090821665,0.1443973482,0.2156585157,-0.009515278,0.6227203012,0.3724325597,0.1507984847,-0.0765272379,0.1891476214,0.1473806798,-0.1591730714,-0.052828595,-0.5078774691,-0.1828075796,0.3740829527,0.1135972589,0.0283298828,-0.4408773184,0.4054227173,0.2401168942,0.0603241175,-0.1184272468,0.1541838497,-0.0568144694,-0.0367468372,0.0292891059,-0.2139268368,0.105295524,0.3675740957,-0.2779199481,-0.3208730817,0.0087620532,-0.0421885885,0.0650617704,0.3736092448,-0.042365253,-0.2475410551,0.0935092494,-0.0054279529,-0.3840718567,-0.3060503006,-0.03811929,0.0701655,-0.1114427149,0.1743900478,-0.3749520183,-0.149213627,0.0872083902,-0.4578383565,0.191630736,-0.0981231406,0.1739199758,0.225857988,-0.2370437384,0.4466663599,0.1191981435,0.1598459333,0.0529770367,-0.1998583525,-0.1200764999,-0.1755337268,-0.4576171041,-0.0922895148,-0.2286255509,0.4021472335,0.1773949713,0.2158044875,-0.0975214317,-0.2927103043,0.0211023055,0.0287384707,0.1406073123,0.2020925283,0.1879706383,0.0467985831,-0.2393500805,-0.1135700494,0.0707941055,0.166129753,-0.2033124864,-0.1762379706,-0.043042507,-0.0814640969,0.1018627435,-0.2113613635,-0.2890201807,-0.2259377986,0.1515955329,0.0919152722,0.1095537916,0.0218211338,0.3725827038,0.0551367439,-0.2080986202,-0.108014144,-0.272910744,0.0094556408,-0.0735264048,-0.0414995626,-0.2861770391,-0.5065011382,-0.3398023546,-0.2370948046,0.1956369877,-0.2353993952,-0.3361153305,-0.3440832198,0.159289673,0.1556718647,0.140961498,0.376039356,0.2247824073,-0.0875745639,-0.1666593105,-0.3055527508,0.4445741475,0.1400023252,0.0882999972,0.1077333391,0.4067178071,-0.1840823442,0.5277171135,0.1247858405,-0.5209571123,0.3717174232,-0.0140295494,0.0158724152,-0.0675409809,-0.2851006985,0.1656641513,0.1925272942,-0.1816551536,0.1626509279,-0.198401615,0.0591691248,0.339425534,0.3197869658,-0.3819201887,-0.27683568,-0.0537330657,-0.15952003,-0.3160973191,0.0252677295,0.3339828253,-0.5367691517,0.2368258089,0.0828461349,0.1489067078,0.1879779398,-0.0566982739,-0.0769120902,-0.106807448,-0.3594061434,0.4149492979,0.0144542353,0.2401235402,0.08461456,-0.0340125449,0.3434586823,0.0684322938,0.9571897984,-0.0614788868,-0.0239519682,0.3976439536,-0.2214727849,-0.3634823263,0.0256643128,-0.1771548539,0.0797582045,0.1019296348,0.2299428284,-0.1601284593,-0.3377129734,0.0265321676,0.0167390537,-0.1257558465,0.0394157954,-0.0708352476,-0.4167459607,-0.4663290679,0.1146943346,-0.0202144012,0.1422989368,0.0958665907,-0.2113025934,-0.5041607618,-0.1101821139,-0.1364934295,0.1517459899,0.3936243057,0.1454341859,0.1311239004,0.142593503,0.3486032188,0.6238431334,0.2257177234,-0.0752178133,-0.2151195705,-0.0062146247,0.1642845273,0.0945564359,0.428067714,-0.0638237447,0.4363097847,-0.0415078662,0.1246345341,0.0061394032,0.2227660418,0.2262482643,0.2113569379,0.0564512536,-0.4379893243,0.563046515,0.0210876353,-0.1794916093,0.3136003017,0.1618500501,-0.2578877807,1.0739566088,0.2233299315,1.0657092333,-0.0835493803,0.2263788283,0.1986545175,0.0644554123,0.4328969419,0.1228055507,0.1599913239,-0.331441164,0.2574236691,0.0078828093,-0.3100057542,0.1413200647,0.2411961406,-0.0830923244,-0.0947633758,-0.2193005085,-0.2675542235,-0.0694857761,0.0392062962,0.3647851646,-0.3496559858,-0.3084585667,-0.0315425247,0.2854220867,0.2928269207,0.2488317192,-0.1491137296,-0.4724773765,-0.1838760078,-0.2159600109,0.0052999565,-0.2775577009,0.2701747417,0.1710419804,-0.3057253659,-0.1693383455,0.2380591631,0.1891984493,-0.147720024,-0.1746553034,0.1169824526,0.1554353386,-0.0974491313,-0.00898444,-0.2751391828,0.3802004158,0.1556450427,-0.3932548463,0.0153315347,0.1589360088,-0.4387569726,-0.090444915,-0.0801719949,-0.2309862673,-0.2791677117,-0.351500541,-0.0657418296,-0.139007166,-0.4286187887,-0.0218832772,0.0130575607,-0.2057515979,0.415224731,-0.1201144308,-0.3386862576,-0.1725048423,0.5766255856,0.2049619704,0.0953757986,0.4988758564,0.2045176476,-0.1476264596,-0.186162293,-0.4698058963,0.4244079888,0.0736841559,0.303047359,-0.0764291734,-0.3145205379,-0.109878324,-0.0109322704,0.0322496183,-0.1338242143,-0.5415477753,-0.2227721959,-0.545301199,0.1993275881,0.0189190842,0.4727430642,-0.2054039091,0.5167491436,-0.5072888136,-0.2756152451,-0.1670345366,-0.0221071858,-0.46642524,0.0234727189,-0.3302716315,-0.1691894233,0.0772074834,0.0515028238,0.0324487314,-0.0511224791,-0.0658903122,-0.1055541858,-0.1573596746,0.1804270297,0.0931681171,-0.1917532533,-0.1235965416,-0.0418644771,-0.14710702,-0.1790007949,0.1822984815,0.2753224671,-0.0650036708,0.1734464914,0.0464400053,0.0953359529,0.2519491315,0.0938533098,-0.0296612848,0.3096496165,0.0970651284,0.1175345853,-0.1214079484,-0.0311483145,-0.0639295802,0.0233555976,-0.0015800457,-0.0470110551,-0.085018307,-0.0344834104,-0.3594438434,0.3289599419,0.2795447111,0.5147302747,0.0854526088,-0.2399228215,0.2423895448,0.046331685,-0.211993292,-0.2285026461,0.353567183,-0.2673204243,0.2338826358,0.2101583034,-0.0903712437,0.1981293559,-0.3148201108,-0.0522817336,0.2038480788,0.3505214155,0.1475191712,0.2091761082,0.1317100525,0.051564198,0.1366330981,0.485137105,0.7115123868,0.5667834282,-0.0109820254,0.6284346581,-0.2339820266,0.0387725495,0.033012677,-0.5802499056,0.5933232307,0.0816221684,-0.0160705559,0.0552336425,0.0665338337,0.1697085947,0.1220055372,-0.2425166517,-0.0767806843,0.2856157124,-0.2219860852,-0.5239715576,-0.1637636423,0.0134422285,0.0549093671,-0.1030844599,-0.0041454178,-0.2838759422,0.1822272092,-0.1851773113,-0.2403619289,-0.2812503278,-0.2857099771,-0.0640084222,0.1365903914,-0.0753130168,-0.1024112701,-0.1006365418,-0.1416368634,-0.1163166463,0.0588549003,0.0947042778,0.141690731,-0.5192406774,-0.0378288999,-0.0573733412,-0.1766133904,-0.2065089643,0.0233231988,-0.1637978405,-0.0212642699,0.1977410316,-0.0061104987,-0.0758739635,-0.3445657194,0.2321965992,0.3549711108,0.0381556414,0.0939724445,0.0618268326,0.078183949,0.0891682655,-0.0918711051,-0.2969636023,-0.097644791,0.3326547444,-0.0406777188,0.1927183717,0.1647498906,0.0005437313,-0.2479170412,0.39039433,0.3924219906,0.324752897,-0.2164477706,0.3476310372,-0.036235489,0.2104081511,-0.0479440689,-0.3186001778,0.377913177,0.1847470999,0.0428155512,-0.0657226592,0.0145698935,0.3071994483,0.3281786442,0.6772935987,-0.2702371776,-0.1875334829,-0.3799323738,0.043086946,0.2865407765,-0.3073634505,-0.0778112784,0.0472268462,-0.126560539,-0.3398209512,-0.0070954086,0.1033463553,-0.169827804,0.2588123083,0.316236943,0.1465733498,-0.074398078,-0.1218501255,-0.1294409335,-0.2518594861,-0.2222988158,-0.2623869181,-0.1795754731,0.1798290312,-0.0964840204,-0.0973355696,-0.1671266407,0.0176442452,0.3991200924,-0.2093284279,-0.1668925881,0.3177430928,-0.2315447927,0.0597222187,0.0446629524,0.2218232751,-0.0057720412,0.4940701723,-0.2855693996,-0.0785241425,0.7860209942,-0.391851306,-0.0228455327,-0.2092846036,0.2463771105,-0.084290266,-0.1669448614,-0.2345831096,0.0083817672,0.2062714994,0.0018958861,-0.3033872545,0.0600997768,0.4667137265,-0.0710082427,-0.1460602582,0.3939974904,0.1948147714,0.1731724292,0.1124255583,-0.2069988251]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1805","title":"can't pickle SwigPyObject objects when calling dataset.get_nearest_examples from FAISS index","comments":"I totally forgot to answer this issue, I'm so sorry. \r\n\r\nI was able to get it working by installing `datasets` from source. Huge thanks!","body":"So, I have the following instances in my dataset\r\n\r\n```\r\n{'question': 'An astronomer observes that a planet rotates faster after a meteorite impact. Which is the most likely effect of \r\nthis increase in rotation?', \r\n'answer': 'C', \r\n'example_id': 'ARCCH_Mercury_7175875', \r\n'options':[{'option_context': 'One effect of increased amperage in the planetary world (..)', 'option_id': 'A', 'option_text': 'Planetary density will decrease.'},\r\n (...)]}\r\n```\r\n\r\nThe `options` value is always an list with 4 options, each one is a dict with `option_context`; `option_id` and `option_text`.\r\n\r\nI would like to overwrite the `option_context` of each instance of my dataset for a dpr result that I am developing. Then, I trained a model already and save it in a FAISS index\r\n```\r\ndpr_dataset = load_dataset(\r\n \"text\",\r\n data_files=ARC_CORPUS_TEXT,\r\n cache_dir=CACHE_DIR,\r\n split=\"train[:100%]\",\r\n )\r\ndpr_dataset.load_faiss_index(\"embeddings\", f\"{ARC_CORPUS_FAISS}\")\r\ntorch.set_grad_enabled(False)\r\n```\r\n\r\nThen, as a processor of my dataset, I created a map function that calls the `dpr_dataset` for each _option_\r\n\r\n```\r\ndef generate_context(example):\r\n question_text = example['question']\r\n for option in example['options']:\r\n question_with_option = question_text + \" \" + option['option_text']\r\n tokenize_text = question_tokenizer(question_with_option, return_tensors=\"pt\").to(device)\r\n question_embed = (\r\n question_encoder(**tokenize_text)\r\n )[0][0].cpu().numpy()\r\n _, retrieved_examples = dpr_dataset.get_nearest_examples(\r\n \"embeddings\", question_embed, k=10\r\n )\r\n # option[\"option_context\"] = retrieved_examples[\"text\"]\r\n # option[\"option_context\"] = \" \".join(option[\"option_context\"]).strip()\r\n #result_dict = {\r\n # 'example_id': example['example_id'],\r\n # 'answer': example['answer'],\r\n # 'question': question_text,\r\n #options': example['options']\r\n # }\r\n return example\r\n```\r\n\r\nI intentionally commented on this portion of the code.\r\n\r\nBut when I call the `map` method, `ds_with_context = dataset.map(generate_context,load_from_cache_file=False)`\r\n\r\nIt calls the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-55-75a458ce205c> in <module>\r\n----> 1 ds_with_context = dataset.map(generate_context,load_from_cache_file=False)\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in <dictcomp>(.0)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1257 fn_kwargs=fn_kwargs,\r\n 1258 new_fingerprint=new_fingerprint,\r\n-> 1259 update_data=update_data,\r\n 1260 )\r\n 1261 else:\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 155 }\r\n 156 # apply actual function\r\n--> 157 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 158 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 159 # re-apply format to the output\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 387 file = StringIO()\r\n 388 with _no_cache_fields(obj):\r\n--> 389 dump(obj, file)\r\n 390 return file.getvalue()\r\n 391 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 359 def dump(obj, file):\r\n 360 \"\"\"pickle an object to a file\"\"\"\r\n--> 361 Pickler(file, recurse=True).dump(obj)\r\n 362 return\r\n 363 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 452 raise PicklingError(msg)\r\n 453 else:\r\n--> 454 StockPickler.dump(self, obj)\r\n 455 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 456 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in save_function(pickler, obj)\r\n 554 dill._dill._create_function,\r\n 555 (obj.__code__, globs, obj.__name__, obj.__defaults__, obj.__closure__, obj.__dict__, fkwdefaults),\r\n--> 556 obj=obj,\r\n 557 )\r\n 558 else:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 784 write(MARK)\r\n 785 for element in obj:\r\n--> 786 save(element)\r\n 787 \r\n 788 if id(obj) in memo:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 885 k, v = tmp[0]\r\n 886 save(k)\r\n--> 887 save(v)\r\n 888 write(SETITEM)\r\n 889 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 885 k, v = tmp[0]\r\n 886 save(k)\r\n--> 887 save(v)\r\n 888 write(SETITEM)\r\n 889 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle SwigPyObject objects\r\n```\r\n\r\nWhich I have no idea how to solve\/deal with it\r\n\r\n","comment_length":24,"text":"can't pickle SwigPyObject objects when calling dataset.get_nearest_examples from FAISS index\nSo, I have the following instances in my dataset\r\n\r\n```\r\n{'question': 'An astronomer observes that a planet rotates faster after a meteorite impact. Which is the most likely effect of \r\nthis increase in rotation?', \r\n'answer': 'C', \r\n'example_id': 'ARCCH_Mercury_7175875', \r\n'options':[{'option_context': 'One effect of increased amperage in the planetary world (..)', 'option_id': 'A', 'option_text': 'Planetary density will decrease.'},\r\n (...)]}\r\n```\r\n\r\nThe `options` value is always an list with 4 options, each one is a dict with `option_context`; `option_id` and `option_text`.\r\n\r\nI would like to overwrite the `option_context` of each instance of my dataset for a dpr result that I am developing. Then, I trained a model already and save it in a FAISS index\r\n```\r\ndpr_dataset = load_dataset(\r\n \"text\",\r\n data_files=ARC_CORPUS_TEXT,\r\n cache_dir=CACHE_DIR,\r\n split=\"train[:100%]\",\r\n )\r\ndpr_dataset.load_faiss_index(\"embeddings\", f\"{ARC_CORPUS_FAISS}\")\r\ntorch.set_grad_enabled(False)\r\n```\r\n\r\nThen, as a processor of my dataset, I created a map function that calls the `dpr_dataset` for each _option_\r\n\r\n```\r\ndef generate_context(example):\r\n question_text = example['question']\r\n for option in example['options']:\r\n question_with_option = question_text + \" \" + option['option_text']\r\n tokenize_text = question_tokenizer(question_with_option, return_tensors=\"pt\").to(device)\r\n question_embed = (\r\n question_encoder(**tokenize_text)\r\n )[0][0].cpu().numpy()\r\n _, retrieved_examples = dpr_dataset.get_nearest_examples(\r\n \"embeddings\", question_embed, k=10\r\n )\r\n # option[\"option_context\"] = retrieved_examples[\"text\"]\r\n # option[\"option_context\"] = \" \".join(option[\"option_context\"]).strip()\r\n #result_dict = {\r\n # 'example_id': example['example_id'],\r\n # 'answer': example['answer'],\r\n # 'question': question_text,\r\n #options': example['options']\r\n # }\r\n return example\r\n```\r\n\r\nI intentionally commented on this portion of the code.\r\n\r\nBut when I call the `map` method, `ds_with_context = dataset.map(generate_context,load_from_cache_file=False)`\r\n\r\nIt calls the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-55-75a458ce205c> in <module>\r\n----> 1 ds_with_context = dataset.map(generate_context,load_from_cache_file=False)\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in <dictcomp>(.0)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1257 fn_kwargs=fn_kwargs,\r\n 1258 new_fingerprint=new_fingerprint,\r\n-> 1259 update_data=update_data,\r\n 1260 )\r\n 1261 else:\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 155 }\r\n 156 # apply actual function\r\n--> 157 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 158 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 159 # re-apply format to the output\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 387 file = StringIO()\r\n 388 with _no_cache_fields(obj):\r\n--> 389 dump(obj, file)\r\n 390 return file.getvalue()\r\n 391 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 359 def dump(obj, file):\r\n 360 \"\"\"pickle an object to a file\"\"\"\r\n--> 361 Pickler(file, recurse=True).dump(obj)\r\n 362 return\r\n 363 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 452 raise PicklingError(msg)\r\n 453 else:\r\n--> 454 StockPickler.dump(self, obj)\r\n 455 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 456 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in save_function(pickler, obj)\r\n 554 dill._dill._create_function,\r\n 555 (obj.__code__, globs, obj.__name__, obj.__defaults__, obj.__closure__, obj.__dict__, fkwdefaults),\r\n--> 556 obj=obj,\r\n 557 )\r\n 558 else:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 784 write(MARK)\r\n 785 for element in obj:\r\n--> 786 save(element)\r\n 787 \r\n 788 if id(obj) in memo:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 885 k, v = tmp[0]\r\n 886 save(k)\r\n--> 887 save(v)\r\n 888 write(SETITEM)\r\n 889 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 880 for k, v in tmp:\r\n 881 save(k)\r\n--> 882 save(v)\r\n 883 write(SETITEMS)\r\n 884 elif n:\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 939 # we only care about session the first pass thru\r\n 940 pickler._session = False\r\n--> 941 StockPickler.save_dict(pickler, obj)\r\n 942 log.info(\"# D2\")\r\n 943 return\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 854 \r\n 855 self.memoize(obj)\r\n--> 856 self._batch_setitems(obj.items())\r\n 857 \r\n 858 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 885 k, v = tmp[0]\r\n 886 save(k)\r\n--> 887 save(v)\r\n 888 write(SETITEM)\r\n 889 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle SwigPyObject objects\r\n```\r\n\r\nWhich I have no idea how to solve\/deal with it\r\n\r\n\nI totally forgot to answer this issue, I'm so sorry. \r\n\r\nI was able to get it working by installing `datasets` from source. Huge thanks!","embeddings":[0.06904158,-0.3887840509,-0.0696357936,0.2468255013,0.1642265022,0.0277648028,0.1784276962,0.0541120283,0.4089650214,0.5476052165,-0.0500990711,0.4959974289,0.115793407,-0.1273343563,-0.267626375,0.0515106916,0.0849947408,0.0134227211,0.0216204002,-0.2425854504,-0.5357055068,-0.0484026559,-0.1984261274,0.0493396372,-0.0398413986,0.0180335399,-0.0332831852,0.0954476371,-0.0584461205,-0.4539258182,0.4045750499,0.257684648,0.2116607726,-0.0134229129,-0.0001266745,0.0679022521,0.2053468674,-0.021252377,-0.1646225452,0.086750403,-0.001575422,0.3058409989,0.2944130301,-0.354249239,-0.2179245949,-0.2550703585,-0.0504558645,-0.134555921,0.0680654272,0.1915952861,0.0670926794,0.1456792355,-0.1130726263,-0.1586890072,0.2215849906,-0.0109133879,-0.0620617904,0.1621502042,0.2719885111,0.3066604137,0.2544573843,0.2748863101,-0.2460707575,0.0368728079,-0.0725996867,0.0926401168,0.3378049135,0.0920075402,0.155671373,0.0969465524,0.2498126328,-0.2737818956,-0.4702080786,-0.4137730896,-0.061396569,-0.1567487866,-0.2627948523,-0.5509217381,-0.0172370858,0.2482920587,0.1254633218,0.1021858975,-0.0639564916,0.5212545395,0.4118613303,0.0359554738,0.0144240391,0.210933432,0.1066410914,-0.0576482825,0.1366034746,-0.1332527101,-0.0283684339,-0.1321200579,-0.2131488323,-0.1242142096,0.4846388996,-0.2354142964,0.245058015,0.0846258178,-0.1290584356,0.401452601,0.0691412911,0.107549645,0.1615174562,-0.118626982,-0.0983864889,0.0395724252,-0.117117323,0.0870361477,-0.1678486615,0.1142573133,0.0268994607,-0.0553063974,-0.631529808,-0.0129407663,0.3205948472,-0.1449027658,-0.1014321595,0.2506060302,-0.6014733911,0.34290874,0.1318702847,0.5309945941,0.0145976041,0.1166362464,0.1949887425,0.2470773607,-0.2960154414,0.4217503667,-0.2266984284,-0.0420274772,0.0346693248,-0.0588204525,-0.1218073964,0.4785673618,0.2849897146,-0.2606017888,0.097831212,-0.0784839243,0.0227258429,-0.1106626913,0.3640687764,0.1970125288,-0.0533642881,0.0772830248,0.1710085273,-0.0954404548,-0.1744895428,0.1174802035,0.1384007186,-0.1021388397,0.1903932989,0.0709636062,-0.1389862746,-0.1317988187,-0.2130447179,0.1211962253,-0.0878358111,0.2369573712,0.187057212,-0.1878823042,-0.0876030996,-0.1486045569,0.2090401649,0.5161315799,-0.2350449562,-0.0758556351,0.0244935937,0.1951778829,0.3501953185,0.2010139525,-0.326797992,0.256172508,0.0371250957,0.232217297,0.3163653016,-0.4490192533,-0.6639236212,-0.0901995748,-0.0756988227,0.0397732891,0.3317385614,0.0975898504,0.3703524768,0.2930629551,-0.1459473521,0.517279923,0.1696157157,0.0434674174,-0.2609554231,-0.344840616,0.0231414568,0.1270146668,-0.0733682066,0.1632385254,-0.0440482982,-0.2676261961,0.3614017367,-0.1503375173,-0.1021282971,0.2690655887,0.5216957331,0.2159873843,0.2130895555,-0.2032638788,-0.0844261721,0.1282851845,-0.1522228718,0.0228898693,-0.0782172158,-0.2090477645,-0.0135675445,0.1977491081,-0.3135472238,-0.1740483791,-0.0746521577,0.1529769897,0.2881989479,-0.1318628043,-0.0989452079,0.049343016,-0.339966625,0.2049800456,-0.5214464068,0.4270998836,-0.0935477465,-0.3137369156,-0.187201947,0.2583495975,0.0134848785,-0.1509953737,-0.158976227,0.1067028642,0.4748960435,-0.4063611031,0.1756881326,-0.3079205453,-0.0195666682,-0.1098886207,0.1134148315,0.0090821665,0.1443973482,0.2156585157,-0.009515278,0.6227203012,0.3724325597,0.1507984847,-0.0765272379,0.1891476214,0.1473806798,-0.1591730714,-0.052828595,-0.5078774691,-0.1828075796,0.3740829527,0.1135972589,0.0283298828,-0.4408773184,0.4054227173,0.2401168942,0.0603241175,-0.1184272468,0.1541838497,-0.0568144694,-0.0367468372,0.0292891059,-0.2139268368,0.105295524,0.3675740957,-0.2779199481,-0.3208730817,0.0087620532,-0.0421885885,0.0650617704,0.3736092448,-0.042365253,-0.2475410551,0.0935092494,-0.0054279529,-0.3840718567,-0.3060503006,-0.03811929,0.0701655,-0.1114427149,0.1743900478,-0.3749520183,-0.149213627,0.0872083902,-0.4578383565,0.191630736,-0.0981231406,0.1739199758,0.225857988,-0.2370437384,0.4466663599,0.1191981435,0.1598459333,0.0529770367,-0.1998583525,-0.1200764999,-0.1755337268,-0.4576171041,-0.0922895148,-0.2286255509,0.4021472335,0.1773949713,0.2158044875,-0.0975214317,-0.2927103043,0.0211023055,0.0287384707,0.1406073123,0.2020925283,0.1879706383,0.0467985831,-0.2393500805,-0.1135700494,0.0707941055,0.166129753,-0.2033124864,-0.1762379706,-0.043042507,-0.0814640969,0.1018627435,-0.2113613635,-0.2890201807,-0.2259377986,0.1515955329,0.0919152722,0.1095537916,0.0218211338,0.3725827038,0.0551367439,-0.2080986202,-0.108014144,-0.272910744,0.0094556408,-0.0735264048,-0.0414995626,-0.2861770391,-0.5065011382,-0.3398023546,-0.2370948046,0.1956369877,-0.2353993952,-0.3361153305,-0.3440832198,0.159289673,0.1556718647,0.140961498,0.376039356,0.2247824073,-0.0875745639,-0.1666593105,-0.3055527508,0.4445741475,0.1400023252,0.0882999972,0.1077333391,0.4067178071,-0.1840823442,0.5277171135,0.1247858405,-0.5209571123,0.3717174232,-0.0140295494,0.0158724152,-0.0675409809,-0.2851006985,0.1656641513,0.1925272942,-0.1816551536,0.1626509279,-0.198401615,0.0591691248,0.339425534,0.3197869658,-0.3819201887,-0.27683568,-0.0537330657,-0.15952003,-0.3160973191,0.0252677295,0.3339828253,-0.5367691517,0.2368258089,0.0828461349,0.1489067078,0.1879779398,-0.0566982739,-0.0769120902,-0.106807448,-0.3594061434,0.4149492979,0.0144542353,0.2401235402,0.08461456,-0.0340125449,0.3434586823,0.0684322938,0.9571897984,-0.0614788868,-0.0239519682,0.3976439536,-0.2214727849,-0.3634823263,0.0256643128,-0.1771548539,0.0797582045,0.1019296348,0.2299428284,-0.1601284593,-0.3377129734,0.0265321676,0.0167390537,-0.1257558465,0.0394157954,-0.0708352476,-0.4167459607,-0.4663290679,0.1146943346,-0.0202144012,0.1422989368,0.0958665907,-0.2113025934,-0.5041607618,-0.1101821139,-0.1364934295,0.1517459899,0.3936243057,0.1454341859,0.1311239004,0.142593503,0.3486032188,0.6238431334,0.2257177234,-0.0752178133,-0.2151195705,-0.0062146247,0.1642845273,0.0945564359,0.428067714,-0.0638237447,0.4363097847,-0.0415078662,0.1246345341,0.0061394032,0.2227660418,0.2262482643,0.2113569379,0.0564512536,-0.4379893243,0.563046515,0.0210876353,-0.1794916093,0.3136003017,0.1618500501,-0.2578877807,1.0739566088,0.2233299315,1.0657092333,-0.0835493803,0.2263788283,0.1986545175,0.0644554123,0.4328969419,0.1228055507,0.1599913239,-0.331441164,0.2574236691,0.0078828093,-0.3100057542,0.1413200647,0.2411961406,-0.0830923244,-0.0947633758,-0.2193005085,-0.2675542235,-0.0694857761,0.0392062962,0.3647851646,-0.3496559858,-0.3084585667,-0.0315425247,0.2854220867,0.2928269207,0.2488317192,-0.1491137296,-0.4724773765,-0.1838760078,-0.2159600109,0.0052999565,-0.2775577009,0.2701747417,0.1710419804,-0.3057253659,-0.1693383455,0.2380591631,0.1891984493,-0.147720024,-0.1746553034,0.1169824526,0.1554353386,-0.0974491313,-0.00898444,-0.2751391828,0.3802004158,0.1556450427,-0.3932548463,0.0153315347,0.1589360088,-0.4387569726,-0.090444915,-0.0801719949,-0.2309862673,-0.2791677117,-0.351500541,-0.0657418296,-0.139007166,-0.4286187887,-0.0218832772,0.0130575607,-0.2057515979,0.415224731,-0.1201144308,-0.3386862576,-0.1725048423,0.5766255856,0.2049619704,0.0953757986,0.4988758564,0.2045176476,-0.1476264596,-0.186162293,-0.4698058963,0.4244079888,0.0736841559,0.303047359,-0.0764291734,-0.3145205379,-0.109878324,-0.0109322704,0.0322496183,-0.1338242143,-0.5415477753,-0.2227721959,-0.545301199,0.1993275881,0.0189190842,0.4727430642,-0.2054039091,0.5167491436,-0.5072888136,-0.2756152451,-0.1670345366,-0.0221071858,-0.46642524,0.0234727189,-0.3302716315,-0.1691894233,0.0772074834,0.0515028238,0.0324487314,-0.0511224791,-0.0658903122,-0.1055541858,-0.1573596746,0.1804270297,0.0931681171,-0.1917532533,-0.1235965416,-0.0418644771,-0.14710702,-0.1790007949,0.1822984815,0.2753224671,-0.0650036708,0.1734464914,0.0464400053,0.0953359529,0.2519491315,0.0938533098,-0.0296612848,0.3096496165,0.0970651284,0.1175345853,-0.1214079484,-0.0311483145,-0.0639295802,0.0233555976,-0.0015800457,-0.0470110551,-0.085018307,-0.0344834104,-0.3594438434,0.3289599419,0.2795447111,0.5147302747,0.0854526088,-0.2399228215,0.2423895448,0.046331685,-0.211993292,-0.2285026461,0.353567183,-0.2673204243,0.2338826358,0.2101583034,-0.0903712437,0.1981293559,-0.3148201108,-0.0522817336,0.2038480788,0.3505214155,0.1475191712,0.2091761082,0.1317100525,0.051564198,0.1366330981,0.485137105,0.7115123868,0.5667834282,-0.0109820254,0.6284346581,-0.2339820266,0.0387725495,0.033012677,-0.5802499056,0.5933232307,0.0816221684,-0.0160705559,0.0552336425,0.0665338337,0.1697085947,0.1220055372,-0.2425166517,-0.0767806843,0.2856157124,-0.2219860852,-0.5239715576,-0.1637636423,0.0134422285,0.0549093671,-0.1030844599,-0.0041454178,-0.2838759422,0.1822272092,-0.1851773113,-0.2403619289,-0.2812503278,-0.2857099771,-0.0640084222,0.1365903914,-0.0753130168,-0.1024112701,-0.1006365418,-0.1416368634,-0.1163166463,0.0588549003,0.0947042778,0.141690731,-0.5192406774,-0.0378288999,-0.0573733412,-0.1766133904,-0.2065089643,0.0233231988,-0.1637978405,-0.0212642699,0.1977410316,-0.0061104987,-0.0758739635,-0.3445657194,0.2321965992,0.3549711108,0.0381556414,0.0939724445,0.0618268326,0.078183949,0.0891682655,-0.0918711051,-0.2969636023,-0.097644791,0.3326547444,-0.0406777188,0.1927183717,0.1647498906,0.0005437313,-0.2479170412,0.39039433,0.3924219906,0.324752897,-0.2164477706,0.3476310372,-0.036235489,0.2104081511,-0.0479440689,-0.3186001778,0.377913177,0.1847470999,0.0428155512,-0.0657226592,0.0145698935,0.3071994483,0.3281786442,0.6772935987,-0.2702371776,-0.1875334829,-0.3799323738,0.043086946,0.2865407765,-0.3073634505,-0.0778112784,0.0472268462,-0.126560539,-0.3398209512,-0.0070954086,0.1033463553,-0.169827804,0.2588123083,0.316236943,0.1465733498,-0.074398078,-0.1218501255,-0.1294409335,-0.2518594861,-0.2222988158,-0.2623869181,-0.1795754731,0.1798290312,-0.0964840204,-0.0973355696,-0.1671266407,0.0176442452,0.3991200924,-0.2093284279,-0.1668925881,0.3177430928,-0.2315447927,0.0597222187,0.0446629524,0.2218232751,-0.0057720412,0.4940701723,-0.2855693996,-0.0785241425,0.7860209942,-0.391851306,-0.0228455327,-0.2092846036,0.2463771105,-0.084290266,-0.1669448614,-0.2345831096,0.0083817672,0.2062714994,0.0018958861,-0.3033872545,0.0600997768,0.4667137265,-0.0710082427,-0.1460602582,0.3939974904,0.1948147714,0.1731724292,0.1124255583,-0.2069988251]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1803","title":"Querying examples from big datasets is slower than small datasets","comments":"Hello, @lhoestq \/ @gaceladri : We have been seeing similar behavior with bigger datasets, where querying time increases. Are you folks aware of any solution that fixes this problem yet? ","body":"After some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.","comment_length":30,"text":"Querying examples from big datasets is slower than small datasets\nAfter some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.\nHello, @lhoestq \/ @gaceladri : We have been seeing similar behavior with bigger datasets, where querying time increases. Are you folks aware of any solution that fixes this problem yet? ","embeddings":[-0.4054936171,0.0586443171,-0.0799714625,0.1315324455,-0.0796937943,-0.1776915044,0.2441749275,0.4918096364,-0.231398806,0.1801942289,0.0814933404,0.055089511,0.1250858605,-0.1007566378,0.1307256222,-0.273406297,0.0807416663,0.1248459145,-0.0751322359,-0.0345290378,0.0240299944,-0.1902869195,-0.3486269414,0.0184702277,-0.0663269982,-0.2240722179,-0.2307065576,0.1072061211,-0.1658504605,-0.417394191,-0.0595365614,-0.1528194398,0.2299299836,0.4270310104,-0.0001130733,0.0164052267,0.2279306203,0.0894177258,-0.1997470111,0.1829340011,0.1748079509,-0.2838544846,0.0868743137,-0.4097666144,-0.047154028,-0.2971978486,0.0934281796,0.0189011656,-0.0412533842,0.3089759946,0.1403642744,0.5131544471,-0.1917047799,-0.1336632371,0.623067379,0.2051521689,-0.3506302536,0.5299629569,0.5806369781,-0.0487995744,-0.3053155243,0.1361406893,-0.0814444497,0.2701411545,0.0370220914,0.1979730874,-0.1128048003,0.1795920134,0.1238104329,0.4088888764,0.3083902597,-0.0893917903,-0.314886421,-0.2920183241,-0.0569284223,-0.2332935333,0.0203074124,0.3264197111,-0.258379221,0.2713423073,-0.3137764633,0.1547778845,-0.105650574,0.1795520484,-0.1511591077,-0.0686156824,0.1150508747,0.1985104829,0.0937960669,0.0133130373,0.151192978,-0.1696210504,0.0284059383,-0.0521108173,-0.2883044779,-0.0881545097,0.1556635648,0.1755786091,0.2685056031,0.1593101472,0.4223206639,0.1182835624,0.2651936114,0.0943621695,0.0451356918,-0.0090215756,-0.2695650458,-0.000978097,0.3478802443,0.2753004134,-0.2945160568,0.0208778251,-0.2383211255,-0.1844542176,0.1815629154,-0.1973295957,-0.0274535324,-0.1337165982,0.0043231687,-0.0873047113,-0.3421146274,-0.1278225183,-0.0265576821,0.4784684777,-0.1834655702,-0.0265724137,-0.2126669139,0.122223787,-0.3353427052,-0.2540251613,-0.1243124604,-0.1883357316,-0.1600804329,0.2303074747,0.1097953022,0.0560976602,0.3020737767,0.058540564,-0.1985150278,0.2419624776,0.3923716843,-0.180865705,0.1044564322,0.0197549015,0.0112557681,0.0635319874,-0.1637397707,0.2447643578,-0.3910897076,0.1972706318,-0.1953568757,-0.3384783268,0.0306392331,0.0919930637,-0.1056957245,-0.0997598916,-0.2270807326,0.2462454289,-0.0694222674,-0.1997463107,-0.0774374455,-0.0941937268,0.0402900465,-0.1531330943,0.1413004994,0.0776268393,-0.2506591082,0.2018505782,-0.0451865606,0.3970346451,0.2543899119,0.5308368206,-0.5031036139,0.1623610705,-0.132778585,0.2966162264,0.2320235521,-0.0281323362,-0.6667606235,0.4973528087,-0.004258499,0.0241891816,0.2724842727,0.1617120504,0.3516778648,0.1078481972,-0.0490624495,0.3837120235,0.1839479655,0.2279311866,-0.3400998116,-0.2303100824,-0.0487276353,0.347974956,-0.3244735301,-0.1571216136,0.3112509251,-0.3645975888,0.4906598926,-0.2065642923,-0.1836265028,0.2509784997,0.0870239511,-0.2777459025,0.1183484569,0.0964874178,0.0859434977,0.242176488,-0.1894870698,-0.0184647031,0.0450902246,-0.1888088584,-0.0722152367,0.0619630329,-0.0552603528,-0.0473853871,-0.0448970497,-0.0918301344,0.3306871057,-0.0035315091,0.164733693,-0.1632062942,-0.2693369985,-0.0094458498,-0.2626185715,0.0792928189,0.0950683355,-0.1213107258,0.0093064588,-0.2492527664,-0.0884235278,-0.0017328389,-0.1519371569,0.3275276423,0.1502290517,0.1502473801,0.223530978,0.1522480249,0.1419565082,-0.0849310458,0.3879742324,0.2968659997,0.075315088,0.0365262404,-0.4053477347,0.5263332129,0.3933131695,0.3160547614,0.2904030383,-0.3733905554,0.1644445658,0.0483795069,0.1105019897,0.4106391072,0.2937985361,0.2421562374,0.5489552617,0.1663255692,-0.3161974847,0.1155823469,0.4506140351,-0.2496014982,0.0180420037,0.1873041689,-0.4217130542,-0.1856174171,0.3702158034,-0.150969401,0.3265801966,0.3312532008,0.1606444567,-0.1239553019,0.0274431203,-0.2904406786,0.2625498772,0.0967946574,0.1414081901,-0.2201083004,0.4368430972,-0.0920058116,-0.2460109293,-0.3815978765,-0.0392074697,0.192131713,-0.1264010072,0.3072988987,-0.0982985049,0.053618893,-0.0324512795,0.1308004558,0.0790725499,-0.0865352377,0.0450386032,0.1935409456,-0.181210503,-0.1583380401,-0.0551914535,0.2278954685,0.2617880702,-0.2087610364,-0.0017746408,-0.0716872513,-0.3983874917,0.0803131238,0.2775991857,0.1493917406,0.1450458467,-0.0469177924,-0.1854478121,-0.2801742256,-0.3152382076,-0.0065734186,0.0799065456,0.2866188288,-0.0709002763,0.0763942152,-0.324452132,-0.1447660178,0.0637101158,-0.232910037,-0.1513257623,0.2760073543,-0.1807471663,0.0008984554,0.088783294,-0.3163755536,-0.1673076302,-0.3001769483,0.3784309924,0.1727609485,0.2083455771,0.1075554192,0.3943434358,0.1223714128,0.0096602468,0.1228192747,0.0379166566,-0.4149514139,0.5190879107,-0.180263564,-0.4500640631,-0.4888216555,-0.2429303229,-0.1641484052,0.4524995983,-0.5341605544,-0.0311471894,-0.5311310291,0.4127760828,0.1235397533,0.1021677181,0.049603384,-0.2131572515,-0.2089558542,0.1292297095,0.0826174915,-0.0064191199,-0.087850906,0.2021380067,-0.0615213774,0.0337112471,-0.0098683694,0.7692419887,0.0689440593,-0.012904753,0.2635357678,-0.0949033275,0.051906649,-0.1647451371,0.1458878964,0.0905036107,-0.2608803511,-0.2699279487,0.4183898568,0.164147079,-0.308024317,0.0591343604,-0.1851301342,-0.0976590216,-0.278183639,0.2541961968,0.1588247716,0.1743984371,0.0698609799,0.3926577866,-0.3018998206,-0.1359266788,0.070709303,-0.2578145564,0.0201729052,-0.1169836447,-0.3373412788,-0.0597153492,-0.6845930219,0.2021396309,-0.0698044449,0.0391389653,0.2540457547,-0.3021757603,0.2348047197,-0.0662641674,0.5487117767,0.1905718446,-0.199644044,-0.1152183488,-0.2325007915,-0.4247890115,-0.1554235816,-0.4956260026,0.1136515513,-0.0005391352,0.3345813751,-0.3185559511,0.1118363068,0.00006073,0.3112286925,0.1916495711,-0.0442787148,-0.167654708,-0.282341063,-0.1325233579,0.162377581,-0.1995147467,0.0504497364,-0.2694344521,-0.1696106941,-0.1472620368,0.1772927642,0.1168199182,0.4819425046,0.2846929729,-0.438256681,0.3633345962,0.0272247978,0.1518049091,0.3828196526,0.4000194669,0.3224876225,-0.1235773861,0.1998887509,0.1022288054,0.0747962669,0.5333119631,-0.0012650972,0.1972725838,0.06110771,0.2763022184,-0.056449011,0.0823665708,0.3274206221,-0.0823386386,-0.0624974258,-0.3168290257,0.7024027109,0.0998988524,0.0085827122,0.2355283499,-0.1901876628,0.1304431111,0.3013944626,-0.1022066027,0.9143377542,-0.1250879019,0.2525978684,0.3494732976,0.0667307153,0.0075898813,-0.0959177613,0.1510162801,-0.3135029376,-0.2316493988,0.1761055142,-0.2805035412,0.0849209949,0.2128594816,-0.131274581,0.0862881094,-0.053731624,0.1990427524,-0.2258604914,0.239251107,-0.2529598475,-0.0510925241,-0.435341537,0.028821053,0.0972892493,-0.0646518022,0.0886060148,-0.1453693062,-0.5541805625,0.0102535067,-0.0938770473,0.0389147289,-0.2355801612,0.1632978916,0.0593763478,-0.4771617949,-0.0842538849,0.1212539822,-0.4953687489,0.0789547786,-0.2458116859,0.4174092412,0.0435457937,0.2301589698,0.530808866,-0.3330542445,-0.0087194387,-0.0292189121,-0.2165679038,0.2495411932,-0.153989464,-0.2520101964,0.0252853706,0.1437309831,0.3611815572,-0.1675023139,-0.2776458263,-0.1240047589,-0.4408734739,-0.2431267202,0.0841377825,-0.0427369662,0.0855582356,0.2314765006,0.3174693286,-0.3549927771,-0.2242870182,0.3709997535,0.0943125486,0.2603205144,0.3130868971,0.0880956575,-0.2163755447,-0.1274372339,0.2965531647,0.4111424685,-0.4757104218,0.0690818205,0.0139511302,0.1246268898,0.2461802214,0.0182096381,0.0741977692,-0.2569902539,-0.3386793733,-0.0690864176,0.009944234,0.3278678656,0.1207619831,0.1357962489,0.0576914959,-0.0145963011,-0.3653215766,0.0142486626,-0.2701688409,0.1390895247,-0.0243229065,0.1828631014,-0.4315895736,0.0174094886,0.0081375437,0.1431654245,0.109162569,0.1063816175,-0.1234385744,-0.2166474015,0.0291855633,0.129072383,-0.1361879706,-0.2797192037,0.2394430935,-0.2063068748,-0.050865896,-0.3758653998,0.1421239972,-0.1364207715,-0.0316604786,-0.4254381359,0.2596428692,-0.1692032218,-0.3019402027,0.3180857599,-0.4044679403,0.1779984385,0.0607639216,0.1787560433,0.00400561,-0.1559256464,0.0720915347,0.2029667348,0.1648203284,0.1396783441,0.3797296882,-0.0262120664,-0.0972278863,-0.1073902994,0.3312670588,0.3493904769,0.0380295813,-0.1387915611,0.1757908612,0.2256359309,-0.3251369894,-0.282820791,-0.1296371967,-0.2570061684,0.2418893427,0.0095792804,0.2093761712,0.1670227349,-0.0789549351,-0.2068968713,0.0986438468,0.550412178,-0.1451756805,0.6776704788,0.3272514343,-0.1993043572,0.2702796161,0.0020866466,0.3459143341,0.3618881702,-0.1196030676,0.5368241668,0.2320750654,0.2449539006,-0.1339675486,-0.3090492785,0.1352112889,-0.001517202,-0.0136430366,0.1391916424,0.0621288307,0.449042201,0.005227657,0.0771015361,-0.3266300857,-0.1267494112,0.0290437471,-0.0826167762,-0.0433856323,-0.1177394092,-0.0748514682,0.1068687961,-0.3724662662,-0.1312437803,0.3648230731,0.0338448547,0.1812083423,-0.4446079135,0.2035410553,-0.1569180191,0.2180987746,-0.2398335487,0.1747612059,0.0500846319,-0.06679409,0.0713923797,0.101684019,0.3679461181,0.1852422655,-0.1066009551,-0.173092857,-0.0566401109,-0.0385056213,-0.0139069818,-0.0869699642,0.0414151885,-0.0986003429,0.3017621934,0.0308141168,-0.155720979,-0.1387580335,-0.0228548311,-0.2242797017,-0.3263448179,-0.3453311622,-0.2006980777,-0.328232646,-0.1295249164,-0.18956846,-0.2536431551,0.1835207492,0.227934584,-0.313332051,-0.067558758,0.000671117,0.0845006332,-0.5105262399,0.6805168986,0.4650323391,0.5556749701,-0.3209068477,-0.1160446405,-0.4356546998,0.0629211441,-0.6595268846,-0.0497882031,0.2942712009,0.3931102157,0.1516689956,0.2153778374,0.1646079421,0.1359258145,-0.1916753799,0.3211123943,-0.5305005908,-0.0809501186,-0.0275071263,0.0169766285,-0.2176148295,-0.6412060857,0.4145414233,-0.4362512231,0.0196193382,-0.179539904,-0.1805547178,0.3991856277,0.2217820287,0.5155704021,0.0022797305,0.1430400163,0.0240666997,-0.1811050922,-0.1920024306,-0.0376321599,0.0861413702,0.176272735,-0.0916480869,0.0870476067,0.1293841302,-0.1807276309,0.0944081545,0.6064497828,-0.1033742651,0.0085011916,0.1139353737,-0.1410082132,-0.2004245371,0.1103874967,-0.1858650744,-0.0338981897,-0.0985818803,-0.2373308688,-0.1588731855,-0.1999848634,0.4235751927,-0.4598163664,0.0573060848,-0.177360788,0.2722935677,0.2497566789,0.3505442739,-0.1962876171,0.029730957,0.3197360933,0.0932119787,-0.0431473553,0.2311461419,0.1514271349,0.1145225316,-0.074937284,0.39183864,-0.3036595583,-0.3066390157,-0.2293879539,-0.3594545424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1803","title":"Querying examples from big datasets is slower than small datasets","comments":"Hi ! I'm pretty sure that it can be fixed by using the Arrow IPC file format instead of the raw streaming format but I haven't tested yet.\r\nI'll take a look at it soon and let you know","body":"After some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.","comment_length":39,"text":"Querying examples from big datasets is slower than small datasets\nAfter some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.\nHi ! I'm pretty sure that it can be fixed by using the Arrow IPC file format instead of the raw streaming format but I haven't tested yet.\r\nI'll take a look at it soon and let you know","embeddings":[-0.4054936171,0.0586443171,-0.0799714625,0.1315324455,-0.0796937943,-0.1776915044,0.2441749275,0.4918096364,-0.231398806,0.1801942289,0.0814933404,0.055089511,0.1250858605,-0.1007566378,0.1307256222,-0.273406297,0.0807416663,0.1248459145,-0.0751322359,-0.0345290378,0.0240299944,-0.1902869195,-0.3486269414,0.0184702277,-0.0663269982,-0.2240722179,-0.2307065576,0.1072061211,-0.1658504605,-0.417394191,-0.0595365614,-0.1528194398,0.2299299836,0.4270310104,-0.0001130733,0.0164052267,0.2279306203,0.0894177258,-0.1997470111,0.1829340011,0.1748079509,-0.2838544846,0.0868743137,-0.4097666144,-0.047154028,-0.2971978486,0.0934281796,0.0189011656,-0.0412533842,0.3089759946,0.1403642744,0.5131544471,-0.1917047799,-0.1336632371,0.623067379,0.2051521689,-0.3506302536,0.5299629569,0.5806369781,-0.0487995744,-0.3053155243,0.1361406893,-0.0814444497,0.2701411545,0.0370220914,0.1979730874,-0.1128048003,0.1795920134,0.1238104329,0.4088888764,0.3083902597,-0.0893917903,-0.314886421,-0.2920183241,-0.0569284223,-0.2332935333,0.0203074124,0.3264197111,-0.258379221,0.2713423073,-0.3137764633,0.1547778845,-0.105650574,0.1795520484,-0.1511591077,-0.0686156824,0.1150508747,0.1985104829,0.0937960669,0.0133130373,0.151192978,-0.1696210504,0.0284059383,-0.0521108173,-0.2883044779,-0.0881545097,0.1556635648,0.1755786091,0.2685056031,0.1593101472,0.4223206639,0.1182835624,0.2651936114,0.0943621695,0.0451356918,-0.0090215756,-0.2695650458,-0.000978097,0.3478802443,0.2753004134,-0.2945160568,0.0208778251,-0.2383211255,-0.1844542176,0.1815629154,-0.1973295957,-0.0274535324,-0.1337165982,0.0043231687,-0.0873047113,-0.3421146274,-0.1278225183,-0.0265576821,0.4784684777,-0.1834655702,-0.0265724137,-0.2126669139,0.122223787,-0.3353427052,-0.2540251613,-0.1243124604,-0.1883357316,-0.1600804329,0.2303074747,0.1097953022,0.0560976602,0.3020737767,0.058540564,-0.1985150278,0.2419624776,0.3923716843,-0.180865705,0.1044564322,0.0197549015,0.0112557681,0.0635319874,-0.1637397707,0.2447643578,-0.3910897076,0.1972706318,-0.1953568757,-0.3384783268,0.0306392331,0.0919930637,-0.1056957245,-0.0997598916,-0.2270807326,0.2462454289,-0.0694222674,-0.1997463107,-0.0774374455,-0.0941937268,0.0402900465,-0.1531330943,0.1413004994,0.0776268393,-0.2506591082,0.2018505782,-0.0451865606,0.3970346451,0.2543899119,0.5308368206,-0.5031036139,0.1623610705,-0.132778585,0.2966162264,0.2320235521,-0.0281323362,-0.6667606235,0.4973528087,-0.004258499,0.0241891816,0.2724842727,0.1617120504,0.3516778648,0.1078481972,-0.0490624495,0.3837120235,0.1839479655,0.2279311866,-0.3400998116,-0.2303100824,-0.0487276353,0.347974956,-0.3244735301,-0.1571216136,0.3112509251,-0.3645975888,0.4906598926,-0.2065642923,-0.1836265028,0.2509784997,0.0870239511,-0.2777459025,0.1183484569,0.0964874178,0.0859434977,0.242176488,-0.1894870698,-0.0184647031,0.0450902246,-0.1888088584,-0.0722152367,0.0619630329,-0.0552603528,-0.0473853871,-0.0448970497,-0.0918301344,0.3306871057,-0.0035315091,0.164733693,-0.1632062942,-0.2693369985,-0.0094458498,-0.2626185715,0.0792928189,0.0950683355,-0.1213107258,0.0093064588,-0.2492527664,-0.0884235278,-0.0017328389,-0.1519371569,0.3275276423,0.1502290517,0.1502473801,0.223530978,0.1522480249,0.1419565082,-0.0849310458,0.3879742324,0.2968659997,0.075315088,0.0365262404,-0.4053477347,0.5263332129,0.3933131695,0.3160547614,0.2904030383,-0.3733905554,0.1644445658,0.0483795069,0.1105019897,0.4106391072,0.2937985361,0.2421562374,0.5489552617,0.1663255692,-0.3161974847,0.1155823469,0.4506140351,-0.2496014982,0.0180420037,0.1873041689,-0.4217130542,-0.1856174171,0.3702158034,-0.150969401,0.3265801966,0.3312532008,0.1606444567,-0.1239553019,0.0274431203,-0.2904406786,0.2625498772,0.0967946574,0.1414081901,-0.2201083004,0.4368430972,-0.0920058116,-0.2460109293,-0.3815978765,-0.0392074697,0.192131713,-0.1264010072,0.3072988987,-0.0982985049,0.053618893,-0.0324512795,0.1308004558,0.0790725499,-0.0865352377,0.0450386032,0.1935409456,-0.181210503,-0.1583380401,-0.0551914535,0.2278954685,0.2617880702,-0.2087610364,-0.0017746408,-0.0716872513,-0.3983874917,0.0803131238,0.2775991857,0.1493917406,0.1450458467,-0.0469177924,-0.1854478121,-0.2801742256,-0.3152382076,-0.0065734186,0.0799065456,0.2866188288,-0.0709002763,0.0763942152,-0.324452132,-0.1447660178,0.0637101158,-0.232910037,-0.1513257623,0.2760073543,-0.1807471663,0.0008984554,0.088783294,-0.3163755536,-0.1673076302,-0.3001769483,0.3784309924,0.1727609485,0.2083455771,0.1075554192,0.3943434358,0.1223714128,0.0096602468,0.1228192747,0.0379166566,-0.4149514139,0.5190879107,-0.180263564,-0.4500640631,-0.4888216555,-0.2429303229,-0.1641484052,0.4524995983,-0.5341605544,-0.0311471894,-0.5311310291,0.4127760828,0.1235397533,0.1021677181,0.049603384,-0.2131572515,-0.2089558542,0.1292297095,0.0826174915,-0.0064191199,-0.087850906,0.2021380067,-0.0615213774,0.0337112471,-0.0098683694,0.7692419887,0.0689440593,-0.012904753,0.2635357678,-0.0949033275,0.051906649,-0.1647451371,0.1458878964,0.0905036107,-0.2608803511,-0.2699279487,0.4183898568,0.164147079,-0.308024317,0.0591343604,-0.1851301342,-0.0976590216,-0.278183639,0.2541961968,0.1588247716,0.1743984371,0.0698609799,0.3926577866,-0.3018998206,-0.1359266788,0.070709303,-0.2578145564,0.0201729052,-0.1169836447,-0.3373412788,-0.0597153492,-0.6845930219,0.2021396309,-0.0698044449,0.0391389653,0.2540457547,-0.3021757603,0.2348047197,-0.0662641674,0.5487117767,0.1905718446,-0.199644044,-0.1152183488,-0.2325007915,-0.4247890115,-0.1554235816,-0.4956260026,0.1136515513,-0.0005391352,0.3345813751,-0.3185559511,0.1118363068,0.00006073,0.3112286925,0.1916495711,-0.0442787148,-0.167654708,-0.282341063,-0.1325233579,0.162377581,-0.1995147467,0.0504497364,-0.2694344521,-0.1696106941,-0.1472620368,0.1772927642,0.1168199182,0.4819425046,0.2846929729,-0.438256681,0.3633345962,0.0272247978,0.1518049091,0.3828196526,0.4000194669,0.3224876225,-0.1235773861,0.1998887509,0.1022288054,0.0747962669,0.5333119631,-0.0012650972,0.1972725838,0.06110771,0.2763022184,-0.056449011,0.0823665708,0.3274206221,-0.0823386386,-0.0624974258,-0.3168290257,0.7024027109,0.0998988524,0.0085827122,0.2355283499,-0.1901876628,0.1304431111,0.3013944626,-0.1022066027,0.9143377542,-0.1250879019,0.2525978684,0.3494732976,0.0667307153,0.0075898813,-0.0959177613,0.1510162801,-0.3135029376,-0.2316493988,0.1761055142,-0.2805035412,0.0849209949,0.2128594816,-0.131274581,0.0862881094,-0.053731624,0.1990427524,-0.2258604914,0.239251107,-0.2529598475,-0.0510925241,-0.435341537,0.028821053,0.0972892493,-0.0646518022,0.0886060148,-0.1453693062,-0.5541805625,0.0102535067,-0.0938770473,0.0389147289,-0.2355801612,0.1632978916,0.0593763478,-0.4771617949,-0.0842538849,0.1212539822,-0.4953687489,0.0789547786,-0.2458116859,0.4174092412,0.0435457937,0.2301589698,0.530808866,-0.3330542445,-0.0087194387,-0.0292189121,-0.2165679038,0.2495411932,-0.153989464,-0.2520101964,0.0252853706,0.1437309831,0.3611815572,-0.1675023139,-0.2776458263,-0.1240047589,-0.4408734739,-0.2431267202,0.0841377825,-0.0427369662,0.0855582356,0.2314765006,0.3174693286,-0.3549927771,-0.2242870182,0.3709997535,0.0943125486,0.2603205144,0.3130868971,0.0880956575,-0.2163755447,-0.1274372339,0.2965531647,0.4111424685,-0.4757104218,0.0690818205,0.0139511302,0.1246268898,0.2461802214,0.0182096381,0.0741977692,-0.2569902539,-0.3386793733,-0.0690864176,0.009944234,0.3278678656,0.1207619831,0.1357962489,0.0576914959,-0.0145963011,-0.3653215766,0.0142486626,-0.2701688409,0.1390895247,-0.0243229065,0.1828631014,-0.4315895736,0.0174094886,0.0081375437,0.1431654245,0.109162569,0.1063816175,-0.1234385744,-0.2166474015,0.0291855633,0.129072383,-0.1361879706,-0.2797192037,0.2394430935,-0.2063068748,-0.050865896,-0.3758653998,0.1421239972,-0.1364207715,-0.0316604786,-0.4254381359,0.2596428692,-0.1692032218,-0.3019402027,0.3180857599,-0.4044679403,0.1779984385,0.0607639216,0.1787560433,0.00400561,-0.1559256464,0.0720915347,0.2029667348,0.1648203284,0.1396783441,0.3797296882,-0.0262120664,-0.0972278863,-0.1073902994,0.3312670588,0.3493904769,0.0380295813,-0.1387915611,0.1757908612,0.2256359309,-0.3251369894,-0.282820791,-0.1296371967,-0.2570061684,0.2418893427,0.0095792804,0.2093761712,0.1670227349,-0.0789549351,-0.2068968713,0.0986438468,0.550412178,-0.1451756805,0.6776704788,0.3272514343,-0.1993043572,0.2702796161,0.0020866466,0.3459143341,0.3618881702,-0.1196030676,0.5368241668,0.2320750654,0.2449539006,-0.1339675486,-0.3090492785,0.1352112889,-0.001517202,-0.0136430366,0.1391916424,0.0621288307,0.449042201,0.005227657,0.0771015361,-0.3266300857,-0.1267494112,0.0290437471,-0.0826167762,-0.0433856323,-0.1177394092,-0.0748514682,0.1068687961,-0.3724662662,-0.1312437803,0.3648230731,0.0338448547,0.1812083423,-0.4446079135,0.2035410553,-0.1569180191,0.2180987746,-0.2398335487,0.1747612059,0.0500846319,-0.06679409,0.0713923797,0.101684019,0.3679461181,0.1852422655,-0.1066009551,-0.173092857,-0.0566401109,-0.0385056213,-0.0139069818,-0.0869699642,0.0414151885,-0.0986003429,0.3017621934,0.0308141168,-0.155720979,-0.1387580335,-0.0228548311,-0.2242797017,-0.3263448179,-0.3453311622,-0.2006980777,-0.328232646,-0.1295249164,-0.18956846,-0.2536431551,0.1835207492,0.227934584,-0.313332051,-0.067558758,0.000671117,0.0845006332,-0.5105262399,0.6805168986,0.4650323391,0.5556749701,-0.3209068477,-0.1160446405,-0.4356546998,0.0629211441,-0.6595268846,-0.0497882031,0.2942712009,0.3931102157,0.1516689956,0.2153778374,0.1646079421,0.1359258145,-0.1916753799,0.3211123943,-0.5305005908,-0.0809501186,-0.0275071263,0.0169766285,-0.2176148295,-0.6412060857,0.4145414233,-0.4362512231,0.0196193382,-0.179539904,-0.1805547178,0.3991856277,0.2217820287,0.5155704021,0.0022797305,0.1430400163,0.0240666997,-0.1811050922,-0.1920024306,-0.0376321599,0.0861413702,0.176272735,-0.0916480869,0.0870476067,0.1293841302,-0.1807276309,0.0944081545,0.6064497828,-0.1033742651,0.0085011916,0.1139353737,-0.1410082132,-0.2004245371,0.1103874967,-0.1858650744,-0.0338981897,-0.0985818803,-0.2373308688,-0.1588731855,-0.1999848634,0.4235751927,-0.4598163664,0.0573060848,-0.177360788,0.2722935677,0.2497566789,0.3505442739,-0.1962876171,0.029730957,0.3197360933,0.0932119787,-0.0431473553,0.2311461419,0.1514271349,0.1145225316,-0.074937284,0.39183864,-0.3036595583,-0.3066390157,-0.2293879539,-0.3594545424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1803","title":"Querying examples from big datasets is slower than small datasets","comments":"My workaround is to shard the dataset into splits in my ssd disk and feed the data in different training sessions. But it is a bit of a pain when we need to reload the last training session with the rest of the split with the Trainer in transformers.\r\n\r\nI mean, when I split the training and then reloads the model and optimizer, it not gets the correct global_status of the optimizer, so I need to hardcode some things. I'm planning to open an issue in transformers and think about it.\r\n```\r\nfrom datasets import load_dataset\r\n\r\nbook_corpus = load_dataset(\"bookcorpus\", split=\"train[:25%]\")\r\nwikicorpus = load_dataset(\"wikicorpus\", split=\"train[:25%]\")\r\nopenwebtext = load_dataset(\"openwebtext\", split=\"train[:25%]\")\r\n\r\nbig_dataset = datasets.concatenate_datasets([wikicorpus, openwebtext, book_corpus])\r\nbig_dataset.shuffle(seed=42)\r\nbig_dataset = big_dataset.map(encode, batched=True, num_proc=20, load_from_cache_file=True, writer_batch_size=5000)\r\nbig_dataset.set_format(type='torch', columns=[\"text\", \"input_ids\", \"attention_mask\", \"token_type_ids\"])\r\n\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/linear_bert\",\r\n overwrite_output_dir=True,\r\n per_device_train_batch_size=71,\r\n save_steps=500,\r\n save_total_limit=10,\r\n logging_first_step=True,\r\n logging_steps=100,\r\n gradient_accumulation_steps=9,\r\n fp16=True,\r\n dataloader_num_workers=20,\r\n warmup_steps=24000,\r\n learning_rate=0.000545205002870214,\r\n adam_epsilon=1e-6,\r\n adam_beta2=0.98,\r\n weight_decay=0.01,\r\n max_steps=138974, # the total number of steps after concatenating 100% datasets\r\n max_grad_norm=1.0,\r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=big_dataset,\r\n tokenizer=tokenizer))\r\n```\r\n\r\nI do one training pass with the total steps of this shard and I use len(bbig)\/batchsize to stop the training (hardcoded in the trainer.py) when I pass over all the examples in this split.\r\n\r\nNow Im working, I will edit the comment with a more elaborated answer when I left the work.","body":"After some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.","comment_length":218,"text":"Querying examples from big datasets is slower than small datasets\nAfter some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.\nMy workaround is to shard the dataset into splits in my ssd disk and feed the data in different training sessions. But it is a bit of a pain when we need to reload the last training session with the rest of the split with the Trainer in transformers.\r\n\r\nI mean, when I split the training and then reloads the model and optimizer, it not gets the correct global_status of the optimizer, so I need to hardcode some things. I'm planning to open an issue in transformers and think about it.\r\n```\r\nfrom datasets import load_dataset\r\n\r\nbook_corpus = load_dataset(\"bookcorpus\", split=\"train[:25%]\")\r\nwikicorpus = load_dataset(\"wikicorpus\", split=\"train[:25%]\")\r\nopenwebtext = load_dataset(\"openwebtext\", split=\"train[:25%]\")\r\n\r\nbig_dataset = datasets.concatenate_datasets([wikicorpus, openwebtext, book_corpus])\r\nbig_dataset.shuffle(seed=42)\r\nbig_dataset = big_dataset.map(encode, batched=True, num_proc=20, load_from_cache_file=True, writer_batch_size=5000)\r\nbig_dataset.set_format(type='torch', columns=[\"text\", \"input_ids\", \"attention_mask\", \"token_type_ids\"])\r\n\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/linear_bert\",\r\n overwrite_output_dir=True,\r\n per_device_train_batch_size=71,\r\n save_steps=500,\r\n save_total_limit=10,\r\n logging_first_step=True,\r\n logging_steps=100,\r\n gradient_accumulation_steps=9,\r\n fp16=True,\r\n dataloader_num_workers=20,\r\n warmup_steps=24000,\r\n learning_rate=0.000545205002870214,\r\n adam_epsilon=1e-6,\r\n adam_beta2=0.98,\r\n weight_decay=0.01,\r\n max_steps=138974, # the total number of steps after concatenating 100% datasets\r\n max_grad_norm=1.0,\r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=big_dataset,\r\n tokenizer=tokenizer))\r\n```\r\n\r\nI do one training pass with the total steps of this shard and I use len(bbig)\/batchsize to stop the training (hardcoded in the trainer.py) when I pass over all the examples in this split.\r\n\r\nNow Im working, I will edit the comment with a more elaborated answer when I left the work.","embeddings":[-0.4054936171,0.0586443171,-0.0799714625,0.1315324455,-0.0796937943,-0.1776915044,0.2441749275,0.4918096364,-0.231398806,0.1801942289,0.0814933404,0.055089511,0.1250858605,-0.1007566378,0.1307256222,-0.273406297,0.0807416663,0.1248459145,-0.0751322359,-0.0345290378,0.0240299944,-0.1902869195,-0.3486269414,0.0184702277,-0.0663269982,-0.2240722179,-0.2307065576,0.1072061211,-0.1658504605,-0.417394191,-0.0595365614,-0.1528194398,0.2299299836,0.4270310104,-0.0001130733,0.0164052267,0.2279306203,0.0894177258,-0.1997470111,0.1829340011,0.1748079509,-0.2838544846,0.0868743137,-0.4097666144,-0.047154028,-0.2971978486,0.0934281796,0.0189011656,-0.0412533842,0.3089759946,0.1403642744,0.5131544471,-0.1917047799,-0.1336632371,0.623067379,0.2051521689,-0.3506302536,0.5299629569,0.5806369781,-0.0487995744,-0.3053155243,0.1361406893,-0.0814444497,0.2701411545,0.0370220914,0.1979730874,-0.1128048003,0.1795920134,0.1238104329,0.4088888764,0.3083902597,-0.0893917903,-0.314886421,-0.2920183241,-0.0569284223,-0.2332935333,0.0203074124,0.3264197111,-0.258379221,0.2713423073,-0.3137764633,0.1547778845,-0.105650574,0.1795520484,-0.1511591077,-0.0686156824,0.1150508747,0.1985104829,0.0937960669,0.0133130373,0.151192978,-0.1696210504,0.0284059383,-0.0521108173,-0.2883044779,-0.0881545097,0.1556635648,0.1755786091,0.2685056031,0.1593101472,0.4223206639,0.1182835624,0.2651936114,0.0943621695,0.0451356918,-0.0090215756,-0.2695650458,-0.000978097,0.3478802443,0.2753004134,-0.2945160568,0.0208778251,-0.2383211255,-0.1844542176,0.1815629154,-0.1973295957,-0.0274535324,-0.1337165982,0.0043231687,-0.0873047113,-0.3421146274,-0.1278225183,-0.0265576821,0.4784684777,-0.1834655702,-0.0265724137,-0.2126669139,0.122223787,-0.3353427052,-0.2540251613,-0.1243124604,-0.1883357316,-0.1600804329,0.2303074747,0.1097953022,0.0560976602,0.3020737767,0.058540564,-0.1985150278,0.2419624776,0.3923716843,-0.180865705,0.1044564322,0.0197549015,0.0112557681,0.0635319874,-0.1637397707,0.2447643578,-0.3910897076,0.1972706318,-0.1953568757,-0.3384783268,0.0306392331,0.0919930637,-0.1056957245,-0.0997598916,-0.2270807326,0.2462454289,-0.0694222674,-0.1997463107,-0.0774374455,-0.0941937268,0.0402900465,-0.1531330943,0.1413004994,0.0776268393,-0.2506591082,0.2018505782,-0.0451865606,0.3970346451,0.2543899119,0.5308368206,-0.5031036139,0.1623610705,-0.132778585,0.2966162264,0.2320235521,-0.0281323362,-0.6667606235,0.4973528087,-0.004258499,0.0241891816,0.2724842727,0.1617120504,0.3516778648,0.1078481972,-0.0490624495,0.3837120235,0.1839479655,0.2279311866,-0.3400998116,-0.2303100824,-0.0487276353,0.347974956,-0.3244735301,-0.1571216136,0.3112509251,-0.3645975888,0.4906598926,-0.2065642923,-0.1836265028,0.2509784997,0.0870239511,-0.2777459025,0.1183484569,0.0964874178,0.0859434977,0.242176488,-0.1894870698,-0.0184647031,0.0450902246,-0.1888088584,-0.0722152367,0.0619630329,-0.0552603528,-0.0473853871,-0.0448970497,-0.0918301344,0.3306871057,-0.0035315091,0.164733693,-0.1632062942,-0.2693369985,-0.0094458498,-0.2626185715,0.0792928189,0.0950683355,-0.1213107258,0.0093064588,-0.2492527664,-0.0884235278,-0.0017328389,-0.1519371569,0.3275276423,0.1502290517,0.1502473801,0.223530978,0.1522480249,0.1419565082,-0.0849310458,0.3879742324,0.2968659997,0.075315088,0.0365262404,-0.4053477347,0.5263332129,0.3933131695,0.3160547614,0.2904030383,-0.3733905554,0.1644445658,0.0483795069,0.1105019897,0.4106391072,0.2937985361,0.2421562374,0.5489552617,0.1663255692,-0.3161974847,0.1155823469,0.4506140351,-0.2496014982,0.0180420037,0.1873041689,-0.4217130542,-0.1856174171,0.3702158034,-0.150969401,0.3265801966,0.3312532008,0.1606444567,-0.1239553019,0.0274431203,-0.2904406786,0.2625498772,0.0967946574,0.1414081901,-0.2201083004,0.4368430972,-0.0920058116,-0.2460109293,-0.3815978765,-0.0392074697,0.192131713,-0.1264010072,0.3072988987,-0.0982985049,0.053618893,-0.0324512795,0.1308004558,0.0790725499,-0.0865352377,0.0450386032,0.1935409456,-0.181210503,-0.1583380401,-0.0551914535,0.2278954685,0.2617880702,-0.2087610364,-0.0017746408,-0.0716872513,-0.3983874917,0.0803131238,0.2775991857,0.1493917406,0.1450458467,-0.0469177924,-0.1854478121,-0.2801742256,-0.3152382076,-0.0065734186,0.0799065456,0.2866188288,-0.0709002763,0.0763942152,-0.324452132,-0.1447660178,0.0637101158,-0.232910037,-0.1513257623,0.2760073543,-0.1807471663,0.0008984554,0.088783294,-0.3163755536,-0.1673076302,-0.3001769483,0.3784309924,0.1727609485,0.2083455771,0.1075554192,0.3943434358,0.1223714128,0.0096602468,0.1228192747,0.0379166566,-0.4149514139,0.5190879107,-0.180263564,-0.4500640631,-0.4888216555,-0.2429303229,-0.1641484052,0.4524995983,-0.5341605544,-0.0311471894,-0.5311310291,0.4127760828,0.1235397533,0.1021677181,0.049603384,-0.2131572515,-0.2089558542,0.1292297095,0.0826174915,-0.0064191199,-0.087850906,0.2021380067,-0.0615213774,0.0337112471,-0.0098683694,0.7692419887,0.0689440593,-0.012904753,0.2635357678,-0.0949033275,0.051906649,-0.1647451371,0.1458878964,0.0905036107,-0.2608803511,-0.2699279487,0.4183898568,0.164147079,-0.308024317,0.0591343604,-0.1851301342,-0.0976590216,-0.278183639,0.2541961968,0.1588247716,0.1743984371,0.0698609799,0.3926577866,-0.3018998206,-0.1359266788,0.070709303,-0.2578145564,0.0201729052,-0.1169836447,-0.3373412788,-0.0597153492,-0.6845930219,0.2021396309,-0.0698044449,0.0391389653,0.2540457547,-0.3021757603,0.2348047197,-0.0662641674,0.5487117767,0.1905718446,-0.199644044,-0.1152183488,-0.2325007915,-0.4247890115,-0.1554235816,-0.4956260026,0.1136515513,-0.0005391352,0.3345813751,-0.3185559511,0.1118363068,0.00006073,0.3112286925,0.1916495711,-0.0442787148,-0.167654708,-0.282341063,-0.1325233579,0.162377581,-0.1995147467,0.0504497364,-0.2694344521,-0.1696106941,-0.1472620368,0.1772927642,0.1168199182,0.4819425046,0.2846929729,-0.438256681,0.3633345962,0.0272247978,0.1518049091,0.3828196526,0.4000194669,0.3224876225,-0.1235773861,0.1998887509,0.1022288054,0.0747962669,0.5333119631,-0.0012650972,0.1972725838,0.06110771,0.2763022184,-0.056449011,0.0823665708,0.3274206221,-0.0823386386,-0.0624974258,-0.3168290257,0.7024027109,0.0998988524,0.0085827122,0.2355283499,-0.1901876628,0.1304431111,0.3013944626,-0.1022066027,0.9143377542,-0.1250879019,0.2525978684,0.3494732976,0.0667307153,0.0075898813,-0.0959177613,0.1510162801,-0.3135029376,-0.2316493988,0.1761055142,-0.2805035412,0.0849209949,0.2128594816,-0.131274581,0.0862881094,-0.053731624,0.1990427524,-0.2258604914,0.239251107,-0.2529598475,-0.0510925241,-0.435341537,0.028821053,0.0972892493,-0.0646518022,0.0886060148,-0.1453693062,-0.5541805625,0.0102535067,-0.0938770473,0.0389147289,-0.2355801612,0.1632978916,0.0593763478,-0.4771617949,-0.0842538849,0.1212539822,-0.4953687489,0.0789547786,-0.2458116859,0.4174092412,0.0435457937,0.2301589698,0.530808866,-0.3330542445,-0.0087194387,-0.0292189121,-0.2165679038,0.2495411932,-0.153989464,-0.2520101964,0.0252853706,0.1437309831,0.3611815572,-0.1675023139,-0.2776458263,-0.1240047589,-0.4408734739,-0.2431267202,0.0841377825,-0.0427369662,0.0855582356,0.2314765006,0.3174693286,-0.3549927771,-0.2242870182,0.3709997535,0.0943125486,0.2603205144,0.3130868971,0.0880956575,-0.2163755447,-0.1274372339,0.2965531647,0.4111424685,-0.4757104218,0.0690818205,0.0139511302,0.1246268898,0.2461802214,0.0182096381,0.0741977692,-0.2569902539,-0.3386793733,-0.0690864176,0.009944234,0.3278678656,0.1207619831,0.1357962489,0.0576914959,-0.0145963011,-0.3653215766,0.0142486626,-0.2701688409,0.1390895247,-0.0243229065,0.1828631014,-0.4315895736,0.0174094886,0.0081375437,0.1431654245,0.109162569,0.1063816175,-0.1234385744,-0.2166474015,0.0291855633,0.129072383,-0.1361879706,-0.2797192037,0.2394430935,-0.2063068748,-0.050865896,-0.3758653998,0.1421239972,-0.1364207715,-0.0316604786,-0.4254381359,0.2596428692,-0.1692032218,-0.3019402027,0.3180857599,-0.4044679403,0.1779984385,0.0607639216,0.1787560433,0.00400561,-0.1559256464,0.0720915347,0.2029667348,0.1648203284,0.1396783441,0.3797296882,-0.0262120664,-0.0972278863,-0.1073902994,0.3312670588,0.3493904769,0.0380295813,-0.1387915611,0.1757908612,0.2256359309,-0.3251369894,-0.282820791,-0.1296371967,-0.2570061684,0.2418893427,0.0095792804,0.2093761712,0.1670227349,-0.0789549351,-0.2068968713,0.0986438468,0.550412178,-0.1451756805,0.6776704788,0.3272514343,-0.1993043572,0.2702796161,0.0020866466,0.3459143341,0.3618881702,-0.1196030676,0.5368241668,0.2320750654,0.2449539006,-0.1339675486,-0.3090492785,0.1352112889,-0.001517202,-0.0136430366,0.1391916424,0.0621288307,0.449042201,0.005227657,0.0771015361,-0.3266300857,-0.1267494112,0.0290437471,-0.0826167762,-0.0433856323,-0.1177394092,-0.0748514682,0.1068687961,-0.3724662662,-0.1312437803,0.3648230731,0.0338448547,0.1812083423,-0.4446079135,0.2035410553,-0.1569180191,0.2180987746,-0.2398335487,0.1747612059,0.0500846319,-0.06679409,0.0713923797,0.101684019,0.3679461181,0.1852422655,-0.1066009551,-0.173092857,-0.0566401109,-0.0385056213,-0.0139069818,-0.0869699642,0.0414151885,-0.0986003429,0.3017621934,0.0308141168,-0.155720979,-0.1387580335,-0.0228548311,-0.2242797017,-0.3263448179,-0.3453311622,-0.2006980777,-0.328232646,-0.1295249164,-0.18956846,-0.2536431551,0.1835207492,0.227934584,-0.313332051,-0.067558758,0.000671117,0.0845006332,-0.5105262399,0.6805168986,0.4650323391,0.5556749701,-0.3209068477,-0.1160446405,-0.4356546998,0.0629211441,-0.6595268846,-0.0497882031,0.2942712009,0.3931102157,0.1516689956,0.2153778374,0.1646079421,0.1359258145,-0.1916753799,0.3211123943,-0.5305005908,-0.0809501186,-0.0275071263,0.0169766285,-0.2176148295,-0.6412060857,0.4145414233,-0.4362512231,0.0196193382,-0.179539904,-0.1805547178,0.3991856277,0.2217820287,0.5155704021,0.0022797305,0.1430400163,0.0240666997,-0.1811050922,-0.1920024306,-0.0376321599,0.0861413702,0.176272735,-0.0916480869,0.0870476067,0.1293841302,-0.1807276309,0.0944081545,0.6064497828,-0.1033742651,0.0085011916,0.1139353737,-0.1410082132,-0.2004245371,0.1103874967,-0.1858650744,-0.0338981897,-0.0985818803,-0.2373308688,-0.1588731855,-0.1999848634,0.4235751927,-0.4598163664,0.0573060848,-0.177360788,0.2722935677,0.2497566789,0.3505442739,-0.1962876171,0.029730957,0.3197360933,0.0932119787,-0.0431473553,0.2311461419,0.1514271349,0.1145225316,-0.074937284,0.39183864,-0.3036595583,-0.3066390157,-0.2293879539,-0.3594545424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1803","title":"Querying examples from big datasets is slower than small datasets","comments":"I just tested and using the Arrow File format doesn't improve the speed... This will need further investigation.\r\n\r\nMy guess is that it has to iterate over the record batches or chunks of a ChunkedArray in order to retrieve elements.\r\n\r\nHowever if we know in advance in which chunk the element is, and at what index it is, then we can access it instantaneously. But this requires dealing with the chunked arrays instead of the pyarrow Table directly which is not practical.","body":"After some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.","comment_length":82,"text":"Querying examples from big datasets is slower than small datasets\nAfter some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.\nI just tested and using the Arrow File format doesn't improve the speed... This will need further investigation.\r\n\r\nMy guess is that it has to iterate over the record batches or chunks of a ChunkedArray in order to retrieve elements.\r\n\r\nHowever if we know in advance in which chunk the element is, and at what index it is, then we can access it instantaneously. But this requires dealing with the chunked arrays instead of the pyarrow Table directly which is not practical.","embeddings":[-0.4054936171,0.0586443171,-0.0799714625,0.1315324455,-0.0796937943,-0.1776915044,0.2441749275,0.4918096364,-0.231398806,0.1801942289,0.0814933404,0.055089511,0.1250858605,-0.1007566378,0.1307256222,-0.273406297,0.0807416663,0.1248459145,-0.0751322359,-0.0345290378,0.0240299944,-0.1902869195,-0.3486269414,0.0184702277,-0.0663269982,-0.2240722179,-0.2307065576,0.1072061211,-0.1658504605,-0.417394191,-0.0595365614,-0.1528194398,0.2299299836,0.4270310104,-0.0001130733,0.0164052267,0.2279306203,0.0894177258,-0.1997470111,0.1829340011,0.1748079509,-0.2838544846,0.0868743137,-0.4097666144,-0.047154028,-0.2971978486,0.0934281796,0.0189011656,-0.0412533842,0.3089759946,0.1403642744,0.5131544471,-0.1917047799,-0.1336632371,0.623067379,0.2051521689,-0.3506302536,0.5299629569,0.5806369781,-0.0487995744,-0.3053155243,0.1361406893,-0.0814444497,0.2701411545,0.0370220914,0.1979730874,-0.1128048003,0.1795920134,0.1238104329,0.4088888764,0.3083902597,-0.0893917903,-0.314886421,-0.2920183241,-0.0569284223,-0.2332935333,0.0203074124,0.3264197111,-0.258379221,0.2713423073,-0.3137764633,0.1547778845,-0.105650574,0.1795520484,-0.1511591077,-0.0686156824,0.1150508747,0.1985104829,0.0937960669,0.0133130373,0.151192978,-0.1696210504,0.0284059383,-0.0521108173,-0.2883044779,-0.0881545097,0.1556635648,0.1755786091,0.2685056031,0.1593101472,0.4223206639,0.1182835624,0.2651936114,0.0943621695,0.0451356918,-0.0090215756,-0.2695650458,-0.000978097,0.3478802443,0.2753004134,-0.2945160568,0.0208778251,-0.2383211255,-0.1844542176,0.1815629154,-0.1973295957,-0.0274535324,-0.1337165982,0.0043231687,-0.0873047113,-0.3421146274,-0.1278225183,-0.0265576821,0.4784684777,-0.1834655702,-0.0265724137,-0.2126669139,0.122223787,-0.3353427052,-0.2540251613,-0.1243124604,-0.1883357316,-0.1600804329,0.2303074747,0.1097953022,0.0560976602,0.3020737767,0.058540564,-0.1985150278,0.2419624776,0.3923716843,-0.180865705,0.1044564322,0.0197549015,0.0112557681,0.0635319874,-0.1637397707,0.2447643578,-0.3910897076,0.1972706318,-0.1953568757,-0.3384783268,0.0306392331,0.0919930637,-0.1056957245,-0.0997598916,-0.2270807326,0.2462454289,-0.0694222674,-0.1997463107,-0.0774374455,-0.0941937268,0.0402900465,-0.1531330943,0.1413004994,0.0776268393,-0.2506591082,0.2018505782,-0.0451865606,0.3970346451,0.2543899119,0.5308368206,-0.5031036139,0.1623610705,-0.132778585,0.2966162264,0.2320235521,-0.0281323362,-0.6667606235,0.4973528087,-0.004258499,0.0241891816,0.2724842727,0.1617120504,0.3516778648,0.1078481972,-0.0490624495,0.3837120235,0.1839479655,0.2279311866,-0.3400998116,-0.2303100824,-0.0487276353,0.347974956,-0.3244735301,-0.1571216136,0.3112509251,-0.3645975888,0.4906598926,-0.2065642923,-0.1836265028,0.2509784997,0.0870239511,-0.2777459025,0.1183484569,0.0964874178,0.0859434977,0.242176488,-0.1894870698,-0.0184647031,0.0450902246,-0.1888088584,-0.0722152367,0.0619630329,-0.0552603528,-0.0473853871,-0.0448970497,-0.0918301344,0.3306871057,-0.0035315091,0.164733693,-0.1632062942,-0.2693369985,-0.0094458498,-0.2626185715,0.0792928189,0.0950683355,-0.1213107258,0.0093064588,-0.2492527664,-0.0884235278,-0.0017328389,-0.1519371569,0.3275276423,0.1502290517,0.1502473801,0.223530978,0.1522480249,0.1419565082,-0.0849310458,0.3879742324,0.2968659997,0.075315088,0.0365262404,-0.4053477347,0.5263332129,0.3933131695,0.3160547614,0.2904030383,-0.3733905554,0.1644445658,0.0483795069,0.1105019897,0.4106391072,0.2937985361,0.2421562374,0.5489552617,0.1663255692,-0.3161974847,0.1155823469,0.4506140351,-0.2496014982,0.0180420037,0.1873041689,-0.4217130542,-0.1856174171,0.3702158034,-0.150969401,0.3265801966,0.3312532008,0.1606444567,-0.1239553019,0.0274431203,-0.2904406786,0.2625498772,0.0967946574,0.1414081901,-0.2201083004,0.4368430972,-0.0920058116,-0.2460109293,-0.3815978765,-0.0392074697,0.192131713,-0.1264010072,0.3072988987,-0.0982985049,0.053618893,-0.0324512795,0.1308004558,0.0790725499,-0.0865352377,0.0450386032,0.1935409456,-0.181210503,-0.1583380401,-0.0551914535,0.2278954685,0.2617880702,-0.2087610364,-0.0017746408,-0.0716872513,-0.3983874917,0.0803131238,0.2775991857,0.1493917406,0.1450458467,-0.0469177924,-0.1854478121,-0.2801742256,-0.3152382076,-0.0065734186,0.0799065456,0.2866188288,-0.0709002763,0.0763942152,-0.324452132,-0.1447660178,0.0637101158,-0.232910037,-0.1513257623,0.2760073543,-0.1807471663,0.0008984554,0.088783294,-0.3163755536,-0.1673076302,-0.3001769483,0.3784309924,0.1727609485,0.2083455771,0.1075554192,0.3943434358,0.1223714128,0.0096602468,0.1228192747,0.0379166566,-0.4149514139,0.5190879107,-0.180263564,-0.4500640631,-0.4888216555,-0.2429303229,-0.1641484052,0.4524995983,-0.5341605544,-0.0311471894,-0.5311310291,0.4127760828,0.1235397533,0.1021677181,0.049603384,-0.2131572515,-0.2089558542,0.1292297095,0.0826174915,-0.0064191199,-0.087850906,0.2021380067,-0.0615213774,0.0337112471,-0.0098683694,0.7692419887,0.0689440593,-0.012904753,0.2635357678,-0.0949033275,0.051906649,-0.1647451371,0.1458878964,0.0905036107,-0.2608803511,-0.2699279487,0.4183898568,0.164147079,-0.308024317,0.0591343604,-0.1851301342,-0.0976590216,-0.278183639,0.2541961968,0.1588247716,0.1743984371,0.0698609799,0.3926577866,-0.3018998206,-0.1359266788,0.070709303,-0.2578145564,0.0201729052,-0.1169836447,-0.3373412788,-0.0597153492,-0.6845930219,0.2021396309,-0.0698044449,0.0391389653,0.2540457547,-0.3021757603,0.2348047197,-0.0662641674,0.5487117767,0.1905718446,-0.199644044,-0.1152183488,-0.2325007915,-0.4247890115,-0.1554235816,-0.4956260026,0.1136515513,-0.0005391352,0.3345813751,-0.3185559511,0.1118363068,0.00006073,0.3112286925,0.1916495711,-0.0442787148,-0.167654708,-0.282341063,-0.1325233579,0.162377581,-0.1995147467,0.0504497364,-0.2694344521,-0.1696106941,-0.1472620368,0.1772927642,0.1168199182,0.4819425046,0.2846929729,-0.438256681,0.3633345962,0.0272247978,0.1518049091,0.3828196526,0.4000194669,0.3224876225,-0.1235773861,0.1998887509,0.1022288054,0.0747962669,0.5333119631,-0.0012650972,0.1972725838,0.06110771,0.2763022184,-0.056449011,0.0823665708,0.3274206221,-0.0823386386,-0.0624974258,-0.3168290257,0.7024027109,0.0998988524,0.0085827122,0.2355283499,-0.1901876628,0.1304431111,0.3013944626,-0.1022066027,0.9143377542,-0.1250879019,0.2525978684,0.3494732976,0.0667307153,0.0075898813,-0.0959177613,0.1510162801,-0.3135029376,-0.2316493988,0.1761055142,-0.2805035412,0.0849209949,0.2128594816,-0.131274581,0.0862881094,-0.053731624,0.1990427524,-0.2258604914,0.239251107,-0.2529598475,-0.0510925241,-0.435341537,0.028821053,0.0972892493,-0.0646518022,0.0886060148,-0.1453693062,-0.5541805625,0.0102535067,-0.0938770473,0.0389147289,-0.2355801612,0.1632978916,0.0593763478,-0.4771617949,-0.0842538849,0.1212539822,-0.4953687489,0.0789547786,-0.2458116859,0.4174092412,0.0435457937,0.2301589698,0.530808866,-0.3330542445,-0.0087194387,-0.0292189121,-0.2165679038,0.2495411932,-0.153989464,-0.2520101964,0.0252853706,0.1437309831,0.3611815572,-0.1675023139,-0.2776458263,-0.1240047589,-0.4408734739,-0.2431267202,0.0841377825,-0.0427369662,0.0855582356,0.2314765006,0.3174693286,-0.3549927771,-0.2242870182,0.3709997535,0.0943125486,0.2603205144,0.3130868971,0.0880956575,-0.2163755447,-0.1274372339,0.2965531647,0.4111424685,-0.4757104218,0.0690818205,0.0139511302,0.1246268898,0.2461802214,0.0182096381,0.0741977692,-0.2569902539,-0.3386793733,-0.0690864176,0.009944234,0.3278678656,0.1207619831,0.1357962489,0.0576914959,-0.0145963011,-0.3653215766,0.0142486626,-0.2701688409,0.1390895247,-0.0243229065,0.1828631014,-0.4315895736,0.0174094886,0.0081375437,0.1431654245,0.109162569,0.1063816175,-0.1234385744,-0.2166474015,0.0291855633,0.129072383,-0.1361879706,-0.2797192037,0.2394430935,-0.2063068748,-0.050865896,-0.3758653998,0.1421239972,-0.1364207715,-0.0316604786,-0.4254381359,0.2596428692,-0.1692032218,-0.3019402027,0.3180857599,-0.4044679403,0.1779984385,0.0607639216,0.1787560433,0.00400561,-0.1559256464,0.0720915347,0.2029667348,0.1648203284,0.1396783441,0.3797296882,-0.0262120664,-0.0972278863,-0.1073902994,0.3312670588,0.3493904769,0.0380295813,-0.1387915611,0.1757908612,0.2256359309,-0.3251369894,-0.282820791,-0.1296371967,-0.2570061684,0.2418893427,0.0095792804,0.2093761712,0.1670227349,-0.0789549351,-0.2068968713,0.0986438468,0.550412178,-0.1451756805,0.6776704788,0.3272514343,-0.1993043572,0.2702796161,0.0020866466,0.3459143341,0.3618881702,-0.1196030676,0.5368241668,0.2320750654,0.2449539006,-0.1339675486,-0.3090492785,0.1352112889,-0.001517202,-0.0136430366,0.1391916424,0.0621288307,0.449042201,0.005227657,0.0771015361,-0.3266300857,-0.1267494112,0.0290437471,-0.0826167762,-0.0433856323,-0.1177394092,-0.0748514682,0.1068687961,-0.3724662662,-0.1312437803,0.3648230731,0.0338448547,0.1812083423,-0.4446079135,0.2035410553,-0.1569180191,0.2180987746,-0.2398335487,0.1747612059,0.0500846319,-0.06679409,0.0713923797,0.101684019,0.3679461181,0.1852422655,-0.1066009551,-0.173092857,-0.0566401109,-0.0385056213,-0.0139069818,-0.0869699642,0.0414151885,-0.0986003429,0.3017621934,0.0308141168,-0.155720979,-0.1387580335,-0.0228548311,-0.2242797017,-0.3263448179,-0.3453311622,-0.2006980777,-0.328232646,-0.1295249164,-0.18956846,-0.2536431551,0.1835207492,0.227934584,-0.313332051,-0.067558758,0.000671117,0.0845006332,-0.5105262399,0.6805168986,0.4650323391,0.5556749701,-0.3209068477,-0.1160446405,-0.4356546998,0.0629211441,-0.6595268846,-0.0497882031,0.2942712009,0.3931102157,0.1516689956,0.2153778374,0.1646079421,0.1359258145,-0.1916753799,0.3211123943,-0.5305005908,-0.0809501186,-0.0275071263,0.0169766285,-0.2176148295,-0.6412060857,0.4145414233,-0.4362512231,0.0196193382,-0.179539904,-0.1805547178,0.3991856277,0.2217820287,0.5155704021,0.0022797305,0.1430400163,0.0240666997,-0.1811050922,-0.1920024306,-0.0376321599,0.0861413702,0.176272735,-0.0916480869,0.0870476067,0.1293841302,-0.1807276309,0.0944081545,0.6064497828,-0.1033742651,0.0085011916,0.1139353737,-0.1410082132,-0.2004245371,0.1103874967,-0.1858650744,-0.0338981897,-0.0985818803,-0.2373308688,-0.1588731855,-0.1999848634,0.4235751927,-0.4598163664,0.0573060848,-0.177360788,0.2722935677,0.2497566789,0.3505442739,-0.1962876171,0.029730957,0.3197360933,0.0932119787,-0.0431473553,0.2311461419,0.1514271349,0.1145225316,-0.074937284,0.39183864,-0.3036595583,-0.3066390157,-0.2293879539,-0.3594545424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1803","title":"Querying examples from big datasets is slower than small datasets","comments":"I have a dataset with about 2.7 million rows (which I'm loading via `load_from_disk`), and I need to fetch around 300k (particular) rows of it, by index. Currently this is taking a really long time (~8 hours). I tried sharding the large dataset but overall it doesn't change how long it takes to fetch the desired rows.\r\n\r\nI actually have enough RAM that I could fit the large dataset in memory. Would having the large dataset in memory speed up querying? To find out, I tried to load (a column of) the large dataset into memory like this:\r\n```\r\ncolumn_data = large_ds['column_name']\r\n```\r\nbut in itself this takes a really long time.\r\n\r\nI'm pretty stuck - do you have any ideas what I should do? ","body":"After some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.","comment_length":125,"text":"Querying examples from big datasets is slower than small datasets\nAfter some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.\nI have a dataset with about 2.7 million rows (which I'm loading via `load_from_disk`), and I need to fetch around 300k (particular) rows of it, by index. Currently this is taking a really long time (~8 hours). I tried sharding the large dataset but overall it doesn't change how long it takes to fetch the desired rows.\r\n\r\nI actually have enough RAM that I could fit the large dataset in memory. Would having the large dataset in memory speed up querying? To find out, I tried to load (a column of) the large dataset into memory like this:\r\n```\r\ncolumn_data = large_ds['column_name']\r\n```\r\nbut in itself this takes a really long time.\r\n\r\nI'm pretty stuck - do you have any ideas what I should do? ","embeddings":[-0.4054936171,0.0586443171,-0.0799714625,0.1315324455,-0.0796937943,-0.1776915044,0.2441749275,0.4918096364,-0.231398806,0.1801942289,0.0814933404,0.055089511,0.1250858605,-0.1007566378,0.1307256222,-0.273406297,0.0807416663,0.1248459145,-0.0751322359,-0.0345290378,0.0240299944,-0.1902869195,-0.3486269414,0.0184702277,-0.0663269982,-0.2240722179,-0.2307065576,0.1072061211,-0.1658504605,-0.417394191,-0.0595365614,-0.1528194398,0.2299299836,0.4270310104,-0.0001130733,0.0164052267,0.2279306203,0.0894177258,-0.1997470111,0.1829340011,0.1748079509,-0.2838544846,0.0868743137,-0.4097666144,-0.047154028,-0.2971978486,0.0934281796,0.0189011656,-0.0412533842,0.3089759946,0.1403642744,0.5131544471,-0.1917047799,-0.1336632371,0.623067379,0.2051521689,-0.3506302536,0.5299629569,0.5806369781,-0.0487995744,-0.3053155243,0.1361406893,-0.0814444497,0.2701411545,0.0370220914,0.1979730874,-0.1128048003,0.1795920134,0.1238104329,0.4088888764,0.3083902597,-0.0893917903,-0.314886421,-0.2920183241,-0.0569284223,-0.2332935333,0.0203074124,0.3264197111,-0.258379221,0.2713423073,-0.3137764633,0.1547778845,-0.105650574,0.1795520484,-0.1511591077,-0.0686156824,0.1150508747,0.1985104829,0.0937960669,0.0133130373,0.151192978,-0.1696210504,0.0284059383,-0.0521108173,-0.2883044779,-0.0881545097,0.1556635648,0.1755786091,0.2685056031,0.1593101472,0.4223206639,0.1182835624,0.2651936114,0.0943621695,0.0451356918,-0.0090215756,-0.2695650458,-0.000978097,0.3478802443,0.2753004134,-0.2945160568,0.0208778251,-0.2383211255,-0.1844542176,0.1815629154,-0.1973295957,-0.0274535324,-0.1337165982,0.0043231687,-0.0873047113,-0.3421146274,-0.1278225183,-0.0265576821,0.4784684777,-0.1834655702,-0.0265724137,-0.2126669139,0.122223787,-0.3353427052,-0.2540251613,-0.1243124604,-0.1883357316,-0.1600804329,0.2303074747,0.1097953022,0.0560976602,0.3020737767,0.058540564,-0.1985150278,0.2419624776,0.3923716843,-0.180865705,0.1044564322,0.0197549015,0.0112557681,0.0635319874,-0.1637397707,0.2447643578,-0.3910897076,0.1972706318,-0.1953568757,-0.3384783268,0.0306392331,0.0919930637,-0.1056957245,-0.0997598916,-0.2270807326,0.2462454289,-0.0694222674,-0.1997463107,-0.0774374455,-0.0941937268,0.0402900465,-0.1531330943,0.1413004994,0.0776268393,-0.2506591082,0.2018505782,-0.0451865606,0.3970346451,0.2543899119,0.5308368206,-0.5031036139,0.1623610705,-0.132778585,0.2966162264,0.2320235521,-0.0281323362,-0.6667606235,0.4973528087,-0.004258499,0.0241891816,0.2724842727,0.1617120504,0.3516778648,0.1078481972,-0.0490624495,0.3837120235,0.1839479655,0.2279311866,-0.3400998116,-0.2303100824,-0.0487276353,0.347974956,-0.3244735301,-0.1571216136,0.3112509251,-0.3645975888,0.4906598926,-0.2065642923,-0.1836265028,0.2509784997,0.0870239511,-0.2777459025,0.1183484569,0.0964874178,0.0859434977,0.242176488,-0.1894870698,-0.0184647031,0.0450902246,-0.1888088584,-0.0722152367,0.0619630329,-0.0552603528,-0.0473853871,-0.0448970497,-0.0918301344,0.3306871057,-0.0035315091,0.164733693,-0.1632062942,-0.2693369985,-0.0094458498,-0.2626185715,0.0792928189,0.0950683355,-0.1213107258,0.0093064588,-0.2492527664,-0.0884235278,-0.0017328389,-0.1519371569,0.3275276423,0.1502290517,0.1502473801,0.223530978,0.1522480249,0.1419565082,-0.0849310458,0.3879742324,0.2968659997,0.075315088,0.0365262404,-0.4053477347,0.5263332129,0.3933131695,0.3160547614,0.2904030383,-0.3733905554,0.1644445658,0.0483795069,0.1105019897,0.4106391072,0.2937985361,0.2421562374,0.5489552617,0.1663255692,-0.3161974847,0.1155823469,0.4506140351,-0.2496014982,0.0180420037,0.1873041689,-0.4217130542,-0.1856174171,0.3702158034,-0.150969401,0.3265801966,0.3312532008,0.1606444567,-0.1239553019,0.0274431203,-0.2904406786,0.2625498772,0.0967946574,0.1414081901,-0.2201083004,0.4368430972,-0.0920058116,-0.2460109293,-0.3815978765,-0.0392074697,0.192131713,-0.1264010072,0.3072988987,-0.0982985049,0.053618893,-0.0324512795,0.1308004558,0.0790725499,-0.0865352377,0.0450386032,0.1935409456,-0.181210503,-0.1583380401,-0.0551914535,0.2278954685,0.2617880702,-0.2087610364,-0.0017746408,-0.0716872513,-0.3983874917,0.0803131238,0.2775991857,0.1493917406,0.1450458467,-0.0469177924,-0.1854478121,-0.2801742256,-0.3152382076,-0.0065734186,0.0799065456,0.2866188288,-0.0709002763,0.0763942152,-0.324452132,-0.1447660178,0.0637101158,-0.232910037,-0.1513257623,0.2760073543,-0.1807471663,0.0008984554,0.088783294,-0.3163755536,-0.1673076302,-0.3001769483,0.3784309924,0.1727609485,0.2083455771,0.1075554192,0.3943434358,0.1223714128,0.0096602468,0.1228192747,0.0379166566,-0.4149514139,0.5190879107,-0.180263564,-0.4500640631,-0.4888216555,-0.2429303229,-0.1641484052,0.4524995983,-0.5341605544,-0.0311471894,-0.5311310291,0.4127760828,0.1235397533,0.1021677181,0.049603384,-0.2131572515,-0.2089558542,0.1292297095,0.0826174915,-0.0064191199,-0.087850906,0.2021380067,-0.0615213774,0.0337112471,-0.0098683694,0.7692419887,0.0689440593,-0.012904753,0.2635357678,-0.0949033275,0.051906649,-0.1647451371,0.1458878964,0.0905036107,-0.2608803511,-0.2699279487,0.4183898568,0.164147079,-0.308024317,0.0591343604,-0.1851301342,-0.0976590216,-0.278183639,0.2541961968,0.1588247716,0.1743984371,0.0698609799,0.3926577866,-0.3018998206,-0.1359266788,0.070709303,-0.2578145564,0.0201729052,-0.1169836447,-0.3373412788,-0.0597153492,-0.6845930219,0.2021396309,-0.0698044449,0.0391389653,0.2540457547,-0.3021757603,0.2348047197,-0.0662641674,0.5487117767,0.1905718446,-0.199644044,-0.1152183488,-0.2325007915,-0.4247890115,-0.1554235816,-0.4956260026,0.1136515513,-0.0005391352,0.3345813751,-0.3185559511,0.1118363068,0.00006073,0.3112286925,0.1916495711,-0.0442787148,-0.167654708,-0.282341063,-0.1325233579,0.162377581,-0.1995147467,0.0504497364,-0.2694344521,-0.1696106941,-0.1472620368,0.1772927642,0.1168199182,0.4819425046,0.2846929729,-0.438256681,0.3633345962,0.0272247978,0.1518049091,0.3828196526,0.4000194669,0.3224876225,-0.1235773861,0.1998887509,0.1022288054,0.0747962669,0.5333119631,-0.0012650972,0.1972725838,0.06110771,0.2763022184,-0.056449011,0.0823665708,0.3274206221,-0.0823386386,-0.0624974258,-0.3168290257,0.7024027109,0.0998988524,0.0085827122,0.2355283499,-0.1901876628,0.1304431111,0.3013944626,-0.1022066027,0.9143377542,-0.1250879019,0.2525978684,0.3494732976,0.0667307153,0.0075898813,-0.0959177613,0.1510162801,-0.3135029376,-0.2316493988,0.1761055142,-0.2805035412,0.0849209949,0.2128594816,-0.131274581,0.0862881094,-0.053731624,0.1990427524,-0.2258604914,0.239251107,-0.2529598475,-0.0510925241,-0.435341537,0.028821053,0.0972892493,-0.0646518022,0.0886060148,-0.1453693062,-0.5541805625,0.0102535067,-0.0938770473,0.0389147289,-0.2355801612,0.1632978916,0.0593763478,-0.4771617949,-0.0842538849,0.1212539822,-0.4953687489,0.0789547786,-0.2458116859,0.4174092412,0.0435457937,0.2301589698,0.530808866,-0.3330542445,-0.0087194387,-0.0292189121,-0.2165679038,0.2495411932,-0.153989464,-0.2520101964,0.0252853706,0.1437309831,0.3611815572,-0.1675023139,-0.2776458263,-0.1240047589,-0.4408734739,-0.2431267202,0.0841377825,-0.0427369662,0.0855582356,0.2314765006,0.3174693286,-0.3549927771,-0.2242870182,0.3709997535,0.0943125486,0.2603205144,0.3130868971,0.0880956575,-0.2163755447,-0.1274372339,0.2965531647,0.4111424685,-0.4757104218,0.0690818205,0.0139511302,0.1246268898,0.2461802214,0.0182096381,0.0741977692,-0.2569902539,-0.3386793733,-0.0690864176,0.009944234,0.3278678656,0.1207619831,0.1357962489,0.0576914959,-0.0145963011,-0.3653215766,0.0142486626,-0.2701688409,0.1390895247,-0.0243229065,0.1828631014,-0.4315895736,0.0174094886,0.0081375437,0.1431654245,0.109162569,0.1063816175,-0.1234385744,-0.2166474015,0.0291855633,0.129072383,-0.1361879706,-0.2797192037,0.2394430935,-0.2063068748,-0.050865896,-0.3758653998,0.1421239972,-0.1364207715,-0.0316604786,-0.4254381359,0.2596428692,-0.1692032218,-0.3019402027,0.3180857599,-0.4044679403,0.1779984385,0.0607639216,0.1787560433,0.00400561,-0.1559256464,0.0720915347,0.2029667348,0.1648203284,0.1396783441,0.3797296882,-0.0262120664,-0.0972278863,-0.1073902994,0.3312670588,0.3493904769,0.0380295813,-0.1387915611,0.1757908612,0.2256359309,-0.3251369894,-0.282820791,-0.1296371967,-0.2570061684,0.2418893427,0.0095792804,0.2093761712,0.1670227349,-0.0789549351,-0.2068968713,0.0986438468,0.550412178,-0.1451756805,0.6776704788,0.3272514343,-0.1993043572,0.2702796161,0.0020866466,0.3459143341,0.3618881702,-0.1196030676,0.5368241668,0.2320750654,0.2449539006,-0.1339675486,-0.3090492785,0.1352112889,-0.001517202,-0.0136430366,0.1391916424,0.0621288307,0.449042201,0.005227657,0.0771015361,-0.3266300857,-0.1267494112,0.0290437471,-0.0826167762,-0.0433856323,-0.1177394092,-0.0748514682,0.1068687961,-0.3724662662,-0.1312437803,0.3648230731,0.0338448547,0.1812083423,-0.4446079135,0.2035410553,-0.1569180191,0.2180987746,-0.2398335487,0.1747612059,0.0500846319,-0.06679409,0.0713923797,0.101684019,0.3679461181,0.1852422655,-0.1066009551,-0.173092857,-0.0566401109,-0.0385056213,-0.0139069818,-0.0869699642,0.0414151885,-0.0986003429,0.3017621934,0.0308141168,-0.155720979,-0.1387580335,-0.0228548311,-0.2242797017,-0.3263448179,-0.3453311622,-0.2006980777,-0.328232646,-0.1295249164,-0.18956846,-0.2536431551,0.1835207492,0.227934584,-0.313332051,-0.067558758,0.000671117,0.0845006332,-0.5105262399,0.6805168986,0.4650323391,0.5556749701,-0.3209068477,-0.1160446405,-0.4356546998,0.0629211441,-0.6595268846,-0.0497882031,0.2942712009,0.3931102157,0.1516689956,0.2153778374,0.1646079421,0.1359258145,-0.1916753799,0.3211123943,-0.5305005908,-0.0809501186,-0.0275071263,0.0169766285,-0.2176148295,-0.6412060857,0.4145414233,-0.4362512231,0.0196193382,-0.179539904,-0.1805547178,0.3991856277,0.2217820287,0.5155704021,0.0022797305,0.1430400163,0.0240666997,-0.1811050922,-0.1920024306,-0.0376321599,0.0861413702,0.176272735,-0.0916480869,0.0870476067,0.1293841302,-0.1807276309,0.0944081545,0.6064497828,-0.1033742651,0.0085011916,0.1139353737,-0.1410082132,-0.2004245371,0.1103874967,-0.1858650744,-0.0338981897,-0.0985818803,-0.2373308688,-0.1588731855,-0.1999848634,0.4235751927,-0.4598163664,0.0573060848,-0.177360788,0.2722935677,0.2497566789,0.3505442739,-0.1962876171,0.029730957,0.3197360933,0.0932119787,-0.0431473553,0.2311461419,0.1514271349,0.1145225316,-0.074937284,0.39183864,-0.3036595583,-0.3066390157,-0.2293879539,-0.3594545424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1803","title":"Querying examples from big datasets is slower than small datasets","comments":"Hi ! Feel free to post a message on the [forum](https:\/\/discuss.huggingface.co\/c\/datasets\/10). I'd be happy to help you with this.\r\n\r\nIn your post on the forum, feel free to add more details about your setup:\r\nWhat are column names and types of your dataset ?\r\nHow was the dataset constructed ?\r\nIs the dataset shuffled ?\r\nIs the dataset tokenized ?\r\nAre you on a SSD or an HDD ?\r\n\r\nI'm sure we can figure something out.\r\nFor example on my laptop I can access the 6 millions articles from wikipedia in less than a minute.","body":"After some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.","comment_length":95,"text":"Querying examples from big datasets is slower than small datasets\nAfter some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.\nHi ! Feel free to post a message on the [forum](https:\/\/discuss.huggingface.co\/c\/datasets\/10). I'd be happy to help you with this.\r\n\r\nIn your post on the forum, feel free to add more details about your setup:\r\nWhat are column names and types of your dataset ?\r\nHow was the dataset constructed ?\r\nIs the dataset shuffled ?\r\nIs the dataset tokenized ?\r\nAre you on a SSD or an HDD ?\r\n\r\nI'm sure we can figure something out.\r\nFor example on my laptop I can access the 6 millions articles from wikipedia in less than a minute.","embeddings":[-0.4054936171,0.0586443171,-0.0799714625,0.1315324455,-0.0796937943,-0.1776915044,0.2441749275,0.4918096364,-0.231398806,0.1801942289,0.0814933404,0.055089511,0.1250858605,-0.1007566378,0.1307256222,-0.273406297,0.0807416663,0.1248459145,-0.0751322359,-0.0345290378,0.0240299944,-0.1902869195,-0.3486269414,0.0184702277,-0.0663269982,-0.2240722179,-0.2307065576,0.1072061211,-0.1658504605,-0.417394191,-0.0595365614,-0.1528194398,0.2299299836,0.4270310104,-0.0001130733,0.0164052267,0.2279306203,0.0894177258,-0.1997470111,0.1829340011,0.1748079509,-0.2838544846,0.0868743137,-0.4097666144,-0.047154028,-0.2971978486,0.0934281796,0.0189011656,-0.0412533842,0.3089759946,0.1403642744,0.5131544471,-0.1917047799,-0.1336632371,0.623067379,0.2051521689,-0.3506302536,0.5299629569,0.5806369781,-0.0487995744,-0.3053155243,0.1361406893,-0.0814444497,0.2701411545,0.0370220914,0.1979730874,-0.1128048003,0.1795920134,0.1238104329,0.4088888764,0.3083902597,-0.0893917903,-0.314886421,-0.2920183241,-0.0569284223,-0.2332935333,0.0203074124,0.3264197111,-0.258379221,0.2713423073,-0.3137764633,0.1547778845,-0.105650574,0.1795520484,-0.1511591077,-0.0686156824,0.1150508747,0.1985104829,0.0937960669,0.0133130373,0.151192978,-0.1696210504,0.0284059383,-0.0521108173,-0.2883044779,-0.0881545097,0.1556635648,0.1755786091,0.2685056031,0.1593101472,0.4223206639,0.1182835624,0.2651936114,0.0943621695,0.0451356918,-0.0090215756,-0.2695650458,-0.000978097,0.3478802443,0.2753004134,-0.2945160568,0.0208778251,-0.2383211255,-0.1844542176,0.1815629154,-0.1973295957,-0.0274535324,-0.1337165982,0.0043231687,-0.0873047113,-0.3421146274,-0.1278225183,-0.0265576821,0.4784684777,-0.1834655702,-0.0265724137,-0.2126669139,0.122223787,-0.3353427052,-0.2540251613,-0.1243124604,-0.1883357316,-0.1600804329,0.2303074747,0.1097953022,0.0560976602,0.3020737767,0.058540564,-0.1985150278,0.2419624776,0.3923716843,-0.180865705,0.1044564322,0.0197549015,0.0112557681,0.0635319874,-0.1637397707,0.2447643578,-0.3910897076,0.1972706318,-0.1953568757,-0.3384783268,0.0306392331,0.0919930637,-0.1056957245,-0.0997598916,-0.2270807326,0.2462454289,-0.0694222674,-0.1997463107,-0.0774374455,-0.0941937268,0.0402900465,-0.1531330943,0.1413004994,0.0776268393,-0.2506591082,0.2018505782,-0.0451865606,0.3970346451,0.2543899119,0.5308368206,-0.5031036139,0.1623610705,-0.132778585,0.2966162264,0.2320235521,-0.0281323362,-0.6667606235,0.4973528087,-0.004258499,0.0241891816,0.2724842727,0.1617120504,0.3516778648,0.1078481972,-0.0490624495,0.3837120235,0.1839479655,0.2279311866,-0.3400998116,-0.2303100824,-0.0487276353,0.347974956,-0.3244735301,-0.1571216136,0.3112509251,-0.3645975888,0.4906598926,-0.2065642923,-0.1836265028,0.2509784997,0.0870239511,-0.2777459025,0.1183484569,0.0964874178,0.0859434977,0.242176488,-0.1894870698,-0.0184647031,0.0450902246,-0.1888088584,-0.0722152367,0.0619630329,-0.0552603528,-0.0473853871,-0.0448970497,-0.0918301344,0.3306871057,-0.0035315091,0.164733693,-0.1632062942,-0.2693369985,-0.0094458498,-0.2626185715,0.0792928189,0.0950683355,-0.1213107258,0.0093064588,-0.2492527664,-0.0884235278,-0.0017328389,-0.1519371569,0.3275276423,0.1502290517,0.1502473801,0.223530978,0.1522480249,0.1419565082,-0.0849310458,0.3879742324,0.2968659997,0.075315088,0.0365262404,-0.4053477347,0.5263332129,0.3933131695,0.3160547614,0.2904030383,-0.3733905554,0.1644445658,0.0483795069,0.1105019897,0.4106391072,0.2937985361,0.2421562374,0.5489552617,0.1663255692,-0.3161974847,0.1155823469,0.4506140351,-0.2496014982,0.0180420037,0.1873041689,-0.4217130542,-0.1856174171,0.3702158034,-0.150969401,0.3265801966,0.3312532008,0.1606444567,-0.1239553019,0.0274431203,-0.2904406786,0.2625498772,0.0967946574,0.1414081901,-0.2201083004,0.4368430972,-0.0920058116,-0.2460109293,-0.3815978765,-0.0392074697,0.192131713,-0.1264010072,0.3072988987,-0.0982985049,0.053618893,-0.0324512795,0.1308004558,0.0790725499,-0.0865352377,0.0450386032,0.1935409456,-0.181210503,-0.1583380401,-0.0551914535,0.2278954685,0.2617880702,-0.2087610364,-0.0017746408,-0.0716872513,-0.3983874917,0.0803131238,0.2775991857,0.1493917406,0.1450458467,-0.0469177924,-0.1854478121,-0.2801742256,-0.3152382076,-0.0065734186,0.0799065456,0.2866188288,-0.0709002763,0.0763942152,-0.324452132,-0.1447660178,0.0637101158,-0.232910037,-0.1513257623,0.2760073543,-0.1807471663,0.0008984554,0.088783294,-0.3163755536,-0.1673076302,-0.3001769483,0.3784309924,0.1727609485,0.2083455771,0.1075554192,0.3943434358,0.1223714128,0.0096602468,0.1228192747,0.0379166566,-0.4149514139,0.5190879107,-0.180263564,-0.4500640631,-0.4888216555,-0.2429303229,-0.1641484052,0.4524995983,-0.5341605544,-0.0311471894,-0.5311310291,0.4127760828,0.1235397533,0.1021677181,0.049603384,-0.2131572515,-0.2089558542,0.1292297095,0.0826174915,-0.0064191199,-0.087850906,0.2021380067,-0.0615213774,0.0337112471,-0.0098683694,0.7692419887,0.0689440593,-0.012904753,0.2635357678,-0.0949033275,0.051906649,-0.1647451371,0.1458878964,0.0905036107,-0.2608803511,-0.2699279487,0.4183898568,0.164147079,-0.308024317,0.0591343604,-0.1851301342,-0.0976590216,-0.278183639,0.2541961968,0.1588247716,0.1743984371,0.0698609799,0.3926577866,-0.3018998206,-0.1359266788,0.070709303,-0.2578145564,0.0201729052,-0.1169836447,-0.3373412788,-0.0597153492,-0.6845930219,0.2021396309,-0.0698044449,0.0391389653,0.2540457547,-0.3021757603,0.2348047197,-0.0662641674,0.5487117767,0.1905718446,-0.199644044,-0.1152183488,-0.2325007915,-0.4247890115,-0.1554235816,-0.4956260026,0.1136515513,-0.0005391352,0.3345813751,-0.3185559511,0.1118363068,0.00006073,0.3112286925,0.1916495711,-0.0442787148,-0.167654708,-0.282341063,-0.1325233579,0.162377581,-0.1995147467,0.0504497364,-0.2694344521,-0.1696106941,-0.1472620368,0.1772927642,0.1168199182,0.4819425046,0.2846929729,-0.438256681,0.3633345962,0.0272247978,0.1518049091,0.3828196526,0.4000194669,0.3224876225,-0.1235773861,0.1998887509,0.1022288054,0.0747962669,0.5333119631,-0.0012650972,0.1972725838,0.06110771,0.2763022184,-0.056449011,0.0823665708,0.3274206221,-0.0823386386,-0.0624974258,-0.3168290257,0.7024027109,0.0998988524,0.0085827122,0.2355283499,-0.1901876628,0.1304431111,0.3013944626,-0.1022066027,0.9143377542,-0.1250879019,0.2525978684,0.3494732976,0.0667307153,0.0075898813,-0.0959177613,0.1510162801,-0.3135029376,-0.2316493988,0.1761055142,-0.2805035412,0.0849209949,0.2128594816,-0.131274581,0.0862881094,-0.053731624,0.1990427524,-0.2258604914,0.239251107,-0.2529598475,-0.0510925241,-0.435341537,0.028821053,0.0972892493,-0.0646518022,0.0886060148,-0.1453693062,-0.5541805625,0.0102535067,-0.0938770473,0.0389147289,-0.2355801612,0.1632978916,0.0593763478,-0.4771617949,-0.0842538849,0.1212539822,-0.4953687489,0.0789547786,-0.2458116859,0.4174092412,0.0435457937,0.2301589698,0.530808866,-0.3330542445,-0.0087194387,-0.0292189121,-0.2165679038,0.2495411932,-0.153989464,-0.2520101964,0.0252853706,0.1437309831,0.3611815572,-0.1675023139,-0.2776458263,-0.1240047589,-0.4408734739,-0.2431267202,0.0841377825,-0.0427369662,0.0855582356,0.2314765006,0.3174693286,-0.3549927771,-0.2242870182,0.3709997535,0.0943125486,0.2603205144,0.3130868971,0.0880956575,-0.2163755447,-0.1274372339,0.2965531647,0.4111424685,-0.4757104218,0.0690818205,0.0139511302,0.1246268898,0.2461802214,0.0182096381,0.0741977692,-0.2569902539,-0.3386793733,-0.0690864176,0.009944234,0.3278678656,0.1207619831,0.1357962489,0.0576914959,-0.0145963011,-0.3653215766,0.0142486626,-0.2701688409,0.1390895247,-0.0243229065,0.1828631014,-0.4315895736,0.0174094886,0.0081375437,0.1431654245,0.109162569,0.1063816175,-0.1234385744,-0.2166474015,0.0291855633,0.129072383,-0.1361879706,-0.2797192037,0.2394430935,-0.2063068748,-0.050865896,-0.3758653998,0.1421239972,-0.1364207715,-0.0316604786,-0.4254381359,0.2596428692,-0.1692032218,-0.3019402027,0.3180857599,-0.4044679403,0.1779984385,0.0607639216,0.1787560433,0.00400561,-0.1559256464,0.0720915347,0.2029667348,0.1648203284,0.1396783441,0.3797296882,-0.0262120664,-0.0972278863,-0.1073902994,0.3312670588,0.3493904769,0.0380295813,-0.1387915611,0.1757908612,0.2256359309,-0.3251369894,-0.282820791,-0.1296371967,-0.2570061684,0.2418893427,0.0095792804,0.2093761712,0.1670227349,-0.0789549351,-0.2068968713,0.0986438468,0.550412178,-0.1451756805,0.6776704788,0.3272514343,-0.1993043572,0.2702796161,0.0020866466,0.3459143341,0.3618881702,-0.1196030676,0.5368241668,0.2320750654,0.2449539006,-0.1339675486,-0.3090492785,0.1352112889,-0.001517202,-0.0136430366,0.1391916424,0.0621288307,0.449042201,0.005227657,0.0771015361,-0.3266300857,-0.1267494112,0.0290437471,-0.0826167762,-0.0433856323,-0.1177394092,-0.0748514682,0.1068687961,-0.3724662662,-0.1312437803,0.3648230731,0.0338448547,0.1812083423,-0.4446079135,0.2035410553,-0.1569180191,0.2180987746,-0.2398335487,0.1747612059,0.0500846319,-0.06679409,0.0713923797,0.101684019,0.3679461181,0.1852422655,-0.1066009551,-0.173092857,-0.0566401109,-0.0385056213,-0.0139069818,-0.0869699642,0.0414151885,-0.0986003429,0.3017621934,0.0308141168,-0.155720979,-0.1387580335,-0.0228548311,-0.2242797017,-0.3263448179,-0.3453311622,-0.2006980777,-0.328232646,-0.1295249164,-0.18956846,-0.2536431551,0.1835207492,0.227934584,-0.313332051,-0.067558758,0.000671117,0.0845006332,-0.5105262399,0.6805168986,0.4650323391,0.5556749701,-0.3209068477,-0.1160446405,-0.4356546998,0.0629211441,-0.6595268846,-0.0497882031,0.2942712009,0.3931102157,0.1516689956,0.2153778374,0.1646079421,0.1359258145,-0.1916753799,0.3211123943,-0.5305005908,-0.0809501186,-0.0275071263,0.0169766285,-0.2176148295,-0.6412060857,0.4145414233,-0.4362512231,0.0196193382,-0.179539904,-0.1805547178,0.3991856277,0.2217820287,0.5155704021,0.0022797305,0.1430400163,0.0240666997,-0.1811050922,-0.1920024306,-0.0376321599,0.0861413702,0.176272735,-0.0916480869,0.0870476067,0.1293841302,-0.1807276309,0.0944081545,0.6064497828,-0.1033742651,0.0085011916,0.1139353737,-0.1410082132,-0.2004245371,0.1103874967,-0.1858650744,-0.0338981897,-0.0985818803,-0.2373308688,-0.1588731855,-0.1999848634,0.4235751927,-0.4598163664,0.0573060848,-0.177360788,0.2722935677,0.2497566789,0.3505442739,-0.1962876171,0.029730957,0.3197360933,0.0932119787,-0.0431473553,0.2311461419,0.1514271349,0.1145225316,-0.074937284,0.39183864,-0.3036595583,-0.3066390157,-0.2293879539,-0.3594545424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1803","title":"Querying examples from big datasets is slower than small datasets","comments":"Thanks @lhoestq, I've [posted on the forum](https:\/\/discuss.huggingface.co\/t\/fetching-rows-of-a-large-dataset-by-index\/4271?u=abisee).","body":"After some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.","comment_length":7,"text":"Querying examples from big datasets is slower than small datasets\nAfter some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.\nThanks @lhoestq, I've [posted on the forum](https:\/\/discuss.huggingface.co\/t\/fetching-rows-of-a-large-dataset-by-index\/4271?u=abisee).","embeddings":[-0.4054936171,0.0586443171,-0.0799714625,0.1315324455,-0.0796937943,-0.1776915044,0.2441749275,0.4918096364,-0.231398806,0.1801942289,0.0814933404,0.055089511,0.1250858605,-0.1007566378,0.1307256222,-0.273406297,0.0807416663,0.1248459145,-0.0751322359,-0.0345290378,0.0240299944,-0.1902869195,-0.3486269414,0.0184702277,-0.0663269982,-0.2240722179,-0.2307065576,0.1072061211,-0.1658504605,-0.417394191,-0.0595365614,-0.1528194398,0.2299299836,0.4270310104,-0.0001130733,0.0164052267,0.2279306203,0.0894177258,-0.1997470111,0.1829340011,0.1748079509,-0.2838544846,0.0868743137,-0.4097666144,-0.047154028,-0.2971978486,0.0934281796,0.0189011656,-0.0412533842,0.3089759946,0.1403642744,0.5131544471,-0.1917047799,-0.1336632371,0.623067379,0.2051521689,-0.3506302536,0.5299629569,0.5806369781,-0.0487995744,-0.3053155243,0.1361406893,-0.0814444497,0.2701411545,0.0370220914,0.1979730874,-0.1128048003,0.1795920134,0.1238104329,0.4088888764,0.3083902597,-0.0893917903,-0.314886421,-0.2920183241,-0.0569284223,-0.2332935333,0.0203074124,0.3264197111,-0.258379221,0.2713423073,-0.3137764633,0.1547778845,-0.105650574,0.1795520484,-0.1511591077,-0.0686156824,0.1150508747,0.1985104829,0.0937960669,0.0133130373,0.151192978,-0.1696210504,0.0284059383,-0.0521108173,-0.2883044779,-0.0881545097,0.1556635648,0.1755786091,0.2685056031,0.1593101472,0.4223206639,0.1182835624,0.2651936114,0.0943621695,0.0451356918,-0.0090215756,-0.2695650458,-0.000978097,0.3478802443,0.2753004134,-0.2945160568,0.0208778251,-0.2383211255,-0.1844542176,0.1815629154,-0.1973295957,-0.0274535324,-0.1337165982,0.0043231687,-0.0873047113,-0.3421146274,-0.1278225183,-0.0265576821,0.4784684777,-0.1834655702,-0.0265724137,-0.2126669139,0.122223787,-0.3353427052,-0.2540251613,-0.1243124604,-0.1883357316,-0.1600804329,0.2303074747,0.1097953022,0.0560976602,0.3020737767,0.058540564,-0.1985150278,0.2419624776,0.3923716843,-0.180865705,0.1044564322,0.0197549015,0.0112557681,0.0635319874,-0.1637397707,0.2447643578,-0.3910897076,0.1972706318,-0.1953568757,-0.3384783268,0.0306392331,0.0919930637,-0.1056957245,-0.0997598916,-0.2270807326,0.2462454289,-0.0694222674,-0.1997463107,-0.0774374455,-0.0941937268,0.0402900465,-0.1531330943,0.1413004994,0.0776268393,-0.2506591082,0.2018505782,-0.0451865606,0.3970346451,0.2543899119,0.5308368206,-0.5031036139,0.1623610705,-0.132778585,0.2966162264,0.2320235521,-0.0281323362,-0.6667606235,0.4973528087,-0.004258499,0.0241891816,0.2724842727,0.1617120504,0.3516778648,0.1078481972,-0.0490624495,0.3837120235,0.1839479655,0.2279311866,-0.3400998116,-0.2303100824,-0.0487276353,0.347974956,-0.3244735301,-0.1571216136,0.3112509251,-0.3645975888,0.4906598926,-0.2065642923,-0.1836265028,0.2509784997,0.0870239511,-0.2777459025,0.1183484569,0.0964874178,0.0859434977,0.242176488,-0.1894870698,-0.0184647031,0.0450902246,-0.1888088584,-0.0722152367,0.0619630329,-0.0552603528,-0.0473853871,-0.0448970497,-0.0918301344,0.3306871057,-0.0035315091,0.164733693,-0.1632062942,-0.2693369985,-0.0094458498,-0.2626185715,0.0792928189,0.0950683355,-0.1213107258,0.0093064588,-0.2492527664,-0.0884235278,-0.0017328389,-0.1519371569,0.3275276423,0.1502290517,0.1502473801,0.223530978,0.1522480249,0.1419565082,-0.0849310458,0.3879742324,0.2968659997,0.075315088,0.0365262404,-0.4053477347,0.5263332129,0.3933131695,0.3160547614,0.2904030383,-0.3733905554,0.1644445658,0.0483795069,0.1105019897,0.4106391072,0.2937985361,0.2421562374,0.5489552617,0.1663255692,-0.3161974847,0.1155823469,0.4506140351,-0.2496014982,0.0180420037,0.1873041689,-0.4217130542,-0.1856174171,0.3702158034,-0.150969401,0.3265801966,0.3312532008,0.1606444567,-0.1239553019,0.0274431203,-0.2904406786,0.2625498772,0.0967946574,0.1414081901,-0.2201083004,0.4368430972,-0.0920058116,-0.2460109293,-0.3815978765,-0.0392074697,0.192131713,-0.1264010072,0.3072988987,-0.0982985049,0.053618893,-0.0324512795,0.1308004558,0.0790725499,-0.0865352377,0.0450386032,0.1935409456,-0.181210503,-0.1583380401,-0.0551914535,0.2278954685,0.2617880702,-0.2087610364,-0.0017746408,-0.0716872513,-0.3983874917,0.0803131238,0.2775991857,0.1493917406,0.1450458467,-0.0469177924,-0.1854478121,-0.2801742256,-0.3152382076,-0.0065734186,0.0799065456,0.2866188288,-0.0709002763,0.0763942152,-0.324452132,-0.1447660178,0.0637101158,-0.232910037,-0.1513257623,0.2760073543,-0.1807471663,0.0008984554,0.088783294,-0.3163755536,-0.1673076302,-0.3001769483,0.3784309924,0.1727609485,0.2083455771,0.1075554192,0.3943434358,0.1223714128,0.0096602468,0.1228192747,0.0379166566,-0.4149514139,0.5190879107,-0.180263564,-0.4500640631,-0.4888216555,-0.2429303229,-0.1641484052,0.4524995983,-0.5341605544,-0.0311471894,-0.5311310291,0.4127760828,0.1235397533,0.1021677181,0.049603384,-0.2131572515,-0.2089558542,0.1292297095,0.0826174915,-0.0064191199,-0.087850906,0.2021380067,-0.0615213774,0.0337112471,-0.0098683694,0.7692419887,0.0689440593,-0.012904753,0.2635357678,-0.0949033275,0.051906649,-0.1647451371,0.1458878964,0.0905036107,-0.2608803511,-0.2699279487,0.4183898568,0.164147079,-0.308024317,0.0591343604,-0.1851301342,-0.0976590216,-0.278183639,0.2541961968,0.1588247716,0.1743984371,0.0698609799,0.3926577866,-0.3018998206,-0.1359266788,0.070709303,-0.2578145564,0.0201729052,-0.1169836447,-0.3373412788,-0.0597153492,-0.6845930219,0.2021396309,-0.0698044449,0.0391389653,0.2540457547,-0.3021757603,0.2348047197,-0.0662641674,0.5487117767,0.1905718446,-0.199644044,-0.1152183488,-0.2325007915,-0.4247890115,-0.1554235816,-0.4956260026,0.1136515513,-0.0005391352,0.3345813751,-0.3185559511,0.1118363068,0.00006073,0.3112286925,0.1916495711,-0.0442787148,-0.167654708,-0.282341063,-0.1325233579,0.162377581,-0.1995147467,0.0504497364,-0.2694344521,-0.1696106941,-0.1472620368,0.1772927642,0.1168199182,0.4819425046,0.2846929729,-0.438256681,0.3633345962,0.0272247978,0.1518049091,0.3828196526,0.4000194669,0.3224876225,-0.1235773861,0.1998887509,0.1022288054,0.0747962669,0.5333119631,-0.0012650972,0.1972725838,0.06110771,0.2763022184,-0.056449011,0.0823665708,0.3274206221,-0.0823386386,-0.0624974258,-0.3168290257,0.7024027109,0.0998988524,0.0085827122,0.2355283499,-0.1901876628,0.1304431111,0.3013944626,-0.1022066027,0.9143377542,-0.1250879019,0.2525978684,0.3494732976,0.0667307153,0.0075898813,-0.0959177613,0.1510162801,-0.3135029376,-0.2316493988,0.1761055142,-0.2805035412,0.0849209949,0.2128594816,-0.131274581,0.0862881094,-0.053731624,0.1990427524,-0.2258604914,0.239251107,-0.2529598475,-0.0510925241,-0.435341537,0.028821053,0.0972892493,-0.0646518022,0.0886060148,-0.1453693062,-0.5541805625,0.0102535067,-0.0938770473,0.0389147289,-0.2355801612,0.1632978916,0.0593763478,-0.4771617949,-0.0842538849,0.1212539822,-0.4953687489,0.0789547786,-0.2458116859,0.4174092412,0.0435457937,0.2301589698,0.530808866,-0.3330542445,-0.0087194387,-0.0292189121,-0.2165679038,0.2495411932,-0.153989464,-0.2520101964,0.0252853706,0.1437309831,0.3611815572,-0.1675023139,-0.2776458263,-0.1240047589,-0.4408734739,-0.2431267202,0.0841377825,-0.0427369662,0.0855582356,0.2314765006,0.3174693286,-0.3549927771,-0.2242870182,0.3709997535,0.0943125486,0.2603205144,0.3130868971,0.0880956575,-0.2163755447,-0.1274372339,0.2965531647,0.4111424685,-0.4757104218,0.0690818205,0.0139511302,0.1246268898,0.2461802214,0.0182096381,0.0741977692,-0.2569902539,-0.3386793733,-0.0690864176,0.009944234,0.3278678656,0.1207619831,0.1357962489,0.0576914959,-0.0145963011,-0.3653215766,0.0142486626,-0.2701688409,0.1390895247,-0.0243229065,0.1828631014,-0.4315895736,0.0174094886,0.0081375437,0.1431654245,0.109162569,0.1063816175,-0.1234385744,-0.2166474015,0.0291855633,0.129072383,-0.1361879706,-0.2797192037,0.2394430935,-0.2063068748,-0.050865896,-0.3758653998,0.1421239972,-0.1364207715,-0.0316604786,-0.4254381359,0.2596428692,-0.1692032218,-0.3019402027,0.3180857599,-0.4044679403,0.1779984385,0.0607639216,0.1787560433,0.00400561,-0.1559256464,0.0720915347,0.2029667348,0.1648203284,0.1396783441,0.3797296882,-0.0262120664,-0.0972278863,-0.1073902994,0.3312670588,0.3493904769,0.0380295813,-0.1387915611,0.1757908612,0.2256359309,-0.3251369894,-0.282820791,-0.1296371967,-0.2570061684,0.2418893427,0.0095792804,0.2093761712,0.1670227349,-0.0789549351,-0.2068968713,0.0986438468,0.550412178,-0.1451756805,0.6776704788,0.3272514343,-0.1993043572,0.2702796161,0.0020866466,0.3459143341,0.3618881702,-0.1196030676,0.5368241668,0.2320750654,0.2449539006,-0.1339675486,-0.3090492785,0.1352112889,-0.001517202,-0.0136430366,0.1391916424,0.0621288307,0.449042201,0.005227657,0.0771015361,-0.3266300857,-0.1267494112,0.0290437471,-0.0826167762,-0.0433856323,-0.1177394092,-0.0748514682,0.1068687961,-0.3724662662,-0.1312437803,0.3648230731,0.0338448547,0.1812083423,-0.4446079135,0.2035410553,-0.1569180191,0.2180987746,-0.2398335487,0.1747612059,0.0500846319,-0.06679409,0.0713923797,0.101684019,0.3679461181,0.1852422655,-0.1066009551,-0.173092857,-0.0566401109,-0.0385056213,-0.0139069818,-0.0869699642,0.0414151885,-0.0986003429,0.3017621934,0.0308141168,-0.155720979,-0.1387580335,-0.0228548311,-0.2242797017,-0.3263448179,-0.3453311622,-0.2006980777,-0.328232646,-0.1295249164,-0.18956846,-0.2536431551,0.1835207492,0.227934584,-0.313332051,-0.067558758,0.000671117,0.0845006332,-0.5105262399,0.6805168986,0.4650323391,0.5556749701,-0.3209068477,-0.1160446405,-0.4356546998,0.0629211441,-0.6595268846,-0.0497882031,0.2942712009,0.3931102157,0.1516689956,0.2153778374,0.1646079421,0.1359258145,-0.1916753799,0.3211123943,-0.5305005908,-0.0809501186,-0.0275071263,0.0169766285,-0.2176148295,-0.6412060857,0.4145414233,-0.4362512231,0.0196193382,-0.179539904,-0.1805547178,0.3991856277,0.2217820287,0.5155704021,0.0022797305,0.1430400163,0.0240666997,-0.1811050922,-0.1920024306,-0.0376321599,0.0861413702,0.176272735,-0.0916480869,0.0870476067,0.1293841302,-0.1807276309,0.0944081545,0.6064497828,-0.1033742651,0.0085011916,0.1139353737,-0.1410082132,-0.2004245371,0.1103874967,-0.1858650744,-0.0338981897,-0.0985818803,-0.2373308688,-0.1588731855,-0.1999848634,0.4235751927,-0.4598163664,0.0573060848,-0.177360788,0.2722935677,0.2497566789,0.3505442739,-0.1962876171,0.029730957,0.3197360933,0.0932119787,-0.0431473553,0.2311461419,0.1514271349,0.1145225316,-0.074937284,0.39183864,-0.3036595583,-0.3066390157,-0.2293879539,-0.3594545424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1803","title":"Querying examples from big datasets is slower than small datasets","comments":"Fixed by #2122.","body":"After some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.","comment_length":3,"text":"Querying examples from big datasets is slower than small datasets\nAfter some experiments with bookcorpus I noticed that querying examples from big datasets is slower than small datasets.\r\nFor example\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nb1 = load_dataset(\"bookcorpus\", split=\"train[:1%]\")\r\nb50 = load_dataset(\"bookcorpus\", split=\"train[:50%]\")\r\nb100 = load_dataset(\"bookcorpus\", split=\"train[:100%]\")\r\n\r\n%timeit _ = b1[-1] \r\n# 12.2 \u00b5s \u00b1 70.4 ns per loop (mean \u00b1 std. dev. of 7 runs, 100000 loops each)\r\n\r\n%timeit _ = b50[-1] \r\n# 92.5 \u00b5s \u00b1 1.24 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n%timeit _ = b100[-1] \r\n# 177 \u00b5s \u00b1 3.13 \u00b5s per loop (mean \u00b1 std. dev. of 7 runs, 10000 loops each)\r\n\r\n```\r\n\r\nIt looks like the time to fetch the example increases with the size of the dataset.\r\n\r\nThis is maybe due to the use of the Arrow streaming format to store the data on disk. I guess pyarrow needs to iterate through the file as a stream to find the queried sample.\r\n\r\nMaybe switching to the Arrow IPC file format could help fixing this issue.\r\n\r\nIndeed according to the [documentation](https:\/\/arrow.apache.org\/docs\/format\/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample, which could fix the issue:\r\n> We define a \u201cfile format\u201d supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer.\r\n\r\ncc @gaceladri since it can help speed up your training when this one is fixed.\nFixed by #2122.","embeddings":[-0.4054936171,0.0586443171,-0.0799714625,0.1315324455,-0.0796937943,-0.1776915044,0.2441749275,0.4918096364,-0.231398806,0.1801942289,0.0814933404,0.055089511,0.1250858605,-0.1007566378,0.1307256222,-0.273406297,0.0807416663,0.1248459145,-0.0751322359,-0.0345290378,0.0240299944,-0.1902869195,-0.3486269414,0.0184702277,-0.0663269982,-0.2240722179,-0.2307065576,0.1072061211,-0.1658504605,-0.417394191,-0.0595365614,-0.1528194398,0.2299299836,0.4270310104,-0.0001130733,0.0164052267,0.2279306203,0.0894177258,-0.1997470111,0.1829340011,0.1748079509,-0.2838544846,0.0868743137,-0.4097666144,-0.047154028,-0.2971978486,0.0934281796,0.0189011656,-0.0412533842,0.3089759946,0.1403642744,0.5131544471,-0.1917047799,-0.1336632371,0.623067379,0.2051521689,-0.3506302536,0.5299629569,0.5806369781,-0.0487995744,-0.3053155243,0.1361406893,-0.0814444497,0.2701411545,0.0370220914,0.1979730874,-0.1128048003,0.1795920134,0.1238104329,0.4088888764,0.3083902597,-0.0893917903,-0.314886421,-0.2920183241,-0.0569284223,-0.2332935333,0.0203074124,0.3264197111,-0.258379221,0.2713423073,-0.3137764633,0.1547778845,-0.105650574,0.1795520484,-0.1511591077,-0.0686156824,0.1150508747,0.1985104829,0.0937960669,0.0133130373,0.151192978,-0.1696210504,0.0284059383,-0.0521108173,-0.2883044779,-0.0881545097,0.1556635648,0.1755786091,0.2685056031,0.1593101472,0.4223206639,0.1182835624,0.2651936114,0.0943621695,0.0451356918,-0.0090215756,-0.2695650458,-0.000978097,0.3478802443,0.2753004134,-0.2945160568,0.0208778251,-0.2383211255,-0.1844542176,0.1815629154,-0.1973295957,-0.0274535324,-0.1337165982,0.0043231687,-0.0873047113,-0.3421146274,-0.1278225183,-0.0265576821,0.4784684777,-0.1834655702,-0.0265724137,-0.2126669139,0.122223787,-0.3353427052,-0.2540251613,-0.1243124604,-0.1883357316,-0.1600804329,0.2303074747,0.1097953022,0.0560976602,0.3020737767,0.058540564,-0.1985150278,0.2419624776,0.3923716843,-0.180865705,0.1044564322,0.0197549015,0.0112557681,0.0635319874,-0.1637397707,0.2447643578,-0.3910897076,0.1972706318,-0.1953568757,-0.3384783268,0.0306392331,0.0919930637,-0.1056957245,-0.0997598916,-0.2270807326,0.2462454289,-0.0694222674,-0.1997463107,-0.0774374455,-0.0941937268,0.0402900465,-0.1531330943,0.1413004994,0.0776268393,-0.2506591082,0.2018505782,-0.0451865606,0.3970346451,0.2543899119,0.5308368206,-0.5031036139,0.1623610705,-0.132778585,0.2966162264,0.2320235521,-0.0281323362,-0.6667606235,0.4973528087,-0.004258499,0.0241891816,0.2724842727,0.1617120504,0.3516778648,0.1078481972,-0.0490624495,0.3837120235,0.1839479655,0.2279311866,-0.3400998116,-0.2303100824,-0.0487276353,0.347974956,-0.3244735301,-0.1571216136,0.3112509251,-0.3645975888,0.4906598926,-0.2065642923,-0.1836265028,0.2509784997,0.0870239511,-0.2777459025,0.1183484569,0.0964874178,0.0859434977,0.242176488,-0.1894870698,-0.0184647031,0.0450902246,-0.1888088584,-0.0722152367,0.0619630329,-0.0552603528,-0.0473853871,-0.0448970497,-0.0918301344,0.3306871057,-0.0035315091,0.164733693,-0.1632062942,-0.2693369985,-0.0094458498,-0.2626185715,0.0792928189,0.0950683355,-0.1213107258,0.0093064588,-0.2492527664,-0.0884235278,-0.0017328389,-0.1519371569,0.3275276423,0.1502290517,0.1502473801,0.223530978,0.1522480249,0.1419565082,-0.0849310458,0.3879742324,0.2968659997,0.075315088,0.0365262404,-0.4053477347,0.5263332129,0.3933131695,0.3160547614,0.2904030383,-0.3733905554,0.1644445658,0.0483795069,0.1105019897,0.4106391072,0.2937985361,0.2421562374,0.5489552617,0.1663255692,-0.3161974847,0.1155823469,0.4506140351,-0.2496014982,0.0180420037,0.1873041689,-0.4217130542,-0.1856174171,0.3702158034,-0.150969401,0.3265801966,0.3312532008,0.1606444567,-0.1239553019,0.0274431203,-0.2904406786,0.2625498772,0.0967946574,0.1414081901,-0.2201083004,0.4368430972,-0.0920058116,-0.2460109293,-0.3815978765,-0.0392074697,0.192131713,-0.1264010072,0.3072988987,-0.0982985049,0.053618893,-0.0324512795,0.1308004558,0.0790725499,-0.0865352377,0.0450386032,0.1935409456,-0.181210503,-0.1583380401,-0.0551914535,0.2278954685,0.2617880702,-0.2087610364,-0.0017746408,-0.0716872513,-0.3983874917,0.0803131238,0.2775991857,0.1493917406,0.1450458467,-0.0469177924,-0.1854478121,-0.2801742256,-0.3152382076,-0.0065734186,0.0799065456,0.2866188288,-0.0709002763,0.0763942152,-0.324452132,-0.1447660178,0.0637101158,-0.232910037,-0.1513257623,0.2760073543,-0.1807471663,0.0008984554,0.088783294,-0.3163755536,-0.1673076302,-0.3001769483,0.3784309924,0.1727609485,0.2083455771,0.1075554192,0.3943434358,0.1223714128,0.0096602468,0.1228192747,0.0379166566,-0.4149514139,0.5190879107,-0.180263564,-0.4500640631,-0.4888216555,-0.2429303229,-0.1641484052,0.4524995983,-0.5341605544,-0.0311471894,-0.5311310291,0.4127760828,0.1235397533,0.1021677181,0.049603384,-0.2131572515,-0.2089558542,0.1292297095,0.0826174915,-0.0064191199,-0.087850906,0.2021380067,-0.0615213774,0.0337112471,-0.0098683694,0.7692419887,0.0689440593,-0.012904753,0.2635357678,-0.0949033275,0.051906649,-0.1647451371,0.1458878964,0.0905036107,-0.2608803511,-0.2699279487,0.4183898568,0.164147079,-0.308024317,0.0591343604,-0.1851301342,-0.0976590216,-0.278183639,0.2541961968,0.1588247716,0.1743984371,0.0698609799,0.3926577866,-0.3018998206,-0.1359266788,0.070709303,-0.2578145564,0.0201729052,-0.1169836447,-0.3373412788,-0.0597153492,-0.6845930219,0.2021396309,-0.0698044449,0.0391389653,0.2540457547,-0.3021757603,0.2348047197,-0.0662641674,0.5487117767,0.1905718446,-0.199644044,-0.1152183488,-0.2325007915,-0.4247890115,-0.1554235816,-0.4956260026,0.1136515513,-0.0005391352,0.3345813751,-0.3185559511,0.1118363068,0.00006073,0.3112286925,0.1916495711,-0.0442787148,-0.167654708,-0.282341063,-0.1325233579,0.162377581,-0.1995147467,0.0504497364,-0.2694344521,-0.1696106941,-0.1472620368,0.1772927642,0.1168199182,0.4819425046,0.2846929729,-0.438256681,0.3633345962,0.0272247978,0.1518049091,0.3828196526,0.4000194669,0.3224876225,-0.1235773861,0.1998887509,0.1022288054,0.0747962669,0.5333119631,-0.0012650972,0.1972725838,0.06110771,0.2763022184,-0.056449011,0.0823665708,0.3274206221,-0.0823386386,-0.0624974258,-0.3168290257,0.7024027109,0.0998988524,0.0085827122,0.2355283499,-0.1901876628,0.1304431111,0.3013944626,-0.1022066027,0.9143377542,-0.1250879019,0.2525978684,0.3494732976,0.0667307153,0.0075898813,-0.0959177613,0.1510162801,-0.3135029376,-0.2316493988,0.1761055142,-0.2805035412,0.0849209949,0.2128594816,-0.131274581,0.0862881094,-0.053731624,0.1990427524,-0.2258604914,0.239251107,-0.2529598475,-0.0510925241,-0.435341537,0.028821053,0.0972892493,-0.0646518022,0.0886060148,-0.1453693062,-0.5541805625,0.0102535067,-0.0938770473,0.0389147289,-0.2355801612,0.1632978916,0.0593763478,-0.4771617949,-0.0842538849,0.1212539822,-0.4953687489,0.0789547786,-0.2458116859,0.4174092412,0.0435457937,0.2301589698,0.530808866,-0.3330542445,-0.0087194387,-0.0292189121,-0.2165679038,0.2495411932,-0.153989464,-0.2520101964,0.0252853706,0.1437309831,0.3611815572,-0.1675023139,-0.2776458263,-0.1240047589,-0.4408734739,-0.2431267202,0.0841377825,-0.0427369662,0.0855582356,0.2314765006,0.3174693286,-0.3549927771,-0.2242870182,0.3709997535,0.0943125486,0.2603205144,0.3130868971,0.0880956575,-0.2163755447,-0.1274372339,0.2965531647,0.4111424685,-0.4757104218,0.0690818205,0.0139511302,0.1246268898,0.2461802214,0.0182096381,0.0741977692,-0.2569902539,-0.3386793733,-0.0690864176,0.009944234,0.3278678656,0.1207619831,0.1357962489,0.0576914959,-0.0145963011,-0.3653215766,0.0142486626,-0.2701688409,0.1390895247,-0.0243229065,0.1828631014,-0.4315895736,0.0174094886,0.0081375437,0.1431654245,0.109162569,0.1063816175,-0.1234385744,-0.2166474015,0.0291855633,0.129072383,-0.1361879706,-0.2797192037,0.2394430935,-0.2063068748,-0.050865896,-0.3758653998,0.1421239972,-0.1364207715,-0.0316604786,-0.4254381359,0.2596428692,-0.1692032218,-0.3019402027,0.3180857599,-0.4044679403,0.1779984385,0.0607639216,0.1787560433,0.00400561,-0.1559256464,0.0720915347,0.2029667348,0.1648203284,0.1396783441,0.3797296882,-0.0262120664,-0.0972278863,-0.1073902994,0.3312670588,0.3493904769,0.0380295813,-0.1387915611,0.1757908612,0.2256359309,-0.3251369894,-0.282820791,-0.1296371967,-0.2570061684,0.2418893427,0.0095792804,0.2093761712,0.1670227349,-0.0789549351,-0.2068968713,0.0986438468,0.550412178,-0.1451756805,0.6776704788,0.3272514343,-0.1993043572,0.2702796161,0.0020866466,0.3459143341,0.3618881702,-0.1196030676,0.5368241668,0.2320750654,0.2449539006,-0.1339675486,-0.3090492785,0.1352112889,-0.001517202,-0.0136430366,0.1391916424,0.0621288307,0.449042201,0.005227657,0.0771015361,-0.3266300857,-0.1267494112,0.0290437471,-0.0826167762,-0.0433856323,-0.1177394092,-0.0748514682,0.1068687961,-0.3724662662,-0.1312437803,0.3648230731,0.0338448547,0.1812083423,-0.4446079135,0.2035410553,-0.1569180191,0.2180987746,-0.2398335487,0.1747612059,0.0500846319,-0.06679409,0.0713923797,0.101684019,0.3679461181,0.1852422655,-0.1066009551,-0.173092857,-0.0566401109,-0.0385056213,-0.0139069818,-0.0869699642,0.0414151885,-0.0986003429,0.3017621934,0.0308141168,-0.155720979,-0.1387580335,-0.0228548311,-0.2242797017,-0.3263448179,-0.3453311622,-0.2006980777,-0.328232646,-0.1295249164,-0.18956846,-0.2536431551,0.1835207492,0.227934584,-0.313332051,-0.067558758,0.000671117,0.0845006332,-0.5105262399,0.6805168986,0.4650323391,0.5556749701,-0.3209068477,-0.1160446405,-0.4356546998,0.0629211441,-0.6595268846,-0.0497882031,0.2942712009,0.3931102157,0.1516689956,0.2153778374,0.1646079421,0.1359258145,-0.1916753799,0.3211123943,-0.5305005908,-0.0809501186,-0.0275071263,0.0169766285,-0.2176148295,-0.6412060857,0.4145414233,-0.4362512231,0.0196193382,-0.179539904,-0.1805547178,0.3991856277,0.2217820287,0.5155704021,0.0022797305,0.1430400163,0.0240666997,-0.1811050922,-0.1920024306,-0.0376321599,0.0861413702,0.176272735,-0.0916480869,0.0870476067,0.1293841302,-0.1807276309,0.0944081545,0.6064497828,-0.1033742651,0.0085011916,0.1139353737,-0.1410082132,-0.2004245371,0.1103874967,-0.1858650744,-0.0338981897,-0.0985818803,-0.2373308688,-0.1588731855,-0.1999848634,0.4235751927,-0.4598163664,0.0573060848,-0.177360788,0.2722935677,0.2497566789,0.3505442739,-0.1962876171,0.029730957,0.3197360933,0.0932119787,-0.0431473553,0.2311461419,0.1514271349,0.1145225316,-0.074937284,0.39183864,-0.3036595583,-0.3066390157,-0.2293879539,-0.3594545424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1797","title":"Connection error","comments":"Hi ! For future references let me add a link to our discussion here : https:\/\/github.com\/huggingface\/datasets\/issues\/759#issuecomment-770684693\r\n\r\nLet me know if you manage to fix your proxy issue or if we can do something on our end to help you :)","body":"Hi\r\nI am hitting to the error, help me and thanks.\r\n\r\n`train_data = datasets.load_dataset(\"xsum\", split=\"train\")`\r\n`ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/xsum\/xsum.py`","comment_length":40,"text":"Connection error\nHi\r\nI am hitting to the error, help me and thanks.\r\n\r\n`train_data = datasets.load_dataset(\"xsum\", split=\"train\")`\r\n`ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/xsum\/xsum.py`\nHi ! For future references let me add a link to our discussion here : https:\/\/github.com\/huggingface\/datasets\/issues\/759#issuecomment-770684693\r\n\r\nLet me know if you manage to fix your proxy issue or if we can do something on our end to help you :)","embeddings":[-0.2624680996,-0.4733125567,-0.1055454686,0.1787157357,0.562163651,-0.1030881405,-0.0170716885,0.2268313915,0.1645879149,0.2273807228,-0.3096103966,-0.0045985654,0.2044959962,0.443228662,0.2047396898,-0.2917329371,-0.0440437049,-0.0842047855,-0.4558957517,0.0675014555,0.0823032334,0.0975515321,-0.1073436663,0.0312042311,-0.3415580392,-0.0755456164,-0.1273664385,0.2207470983,-0.2686606348,-0.1070599109,0.4869900942,0.1393877864,0.0232104808,0.5769346356,-0.0001075649,0.1149471328,0.3164846003,0.0242445804,-0.223999083,-0.33114025,-0.1180704832,-0.0296365749,0.2868143916,-0.0743958578,-0.3245800436,0.0551903173,0.1663982123,-0.2252040654,0.597260654,0.1780353189,0.2932913601,0.4454437494,0.3417463601,-0.1741364151,-0.0087845353,-0.2359611988,-0.0033456399,0.4921544492,0.1815267652,0.0471874774,0.213283807,0.233723402,0.0923611745,-0.1159159169,0.2341250628,-0.1357224286,0.0822611749,-0.3459514081,0.034813907,0.3595282733,0.557164669,-0.1973253042,-0.2897788584,0.0463784263,-0.0473954342,-0.1539386511,0.2277335525,0.0708233267,-0.0782245174,0.1751246154,-0.2249851078,-0.214958176,-0.2798596323,0.2543148696,-0.0623600334,0.0234312005,-0.2333235741,0.1285450757,0.066239208,-0.132538259,-0.12612997,0.157585308,0.0271808878,0.1737780869,-0.4677267075,-0.0263176393,-0.1510633975,0.230041042,0.2147616893,0.1480178237,-0.1089281887,0.176799491,-0.1645568162,0.2396333069,0.2342701405,0.0991906896,-0.0924919024,-0.080631882,0.2583271563,0.2595568299,-0.1094023585,-0.1085266024,-0.0784777477,-0.2059405595,-0.3137497902,-0.062031053,0.2355893999,-0.2644431889,-0.2093426883,0.0777385458,0.1508231759,-0.0532449372,0.1676166505,0.4941645861,-0.2439726591,0.1843062937,0.0398412868,0.1510654837,-0.2680298984,-0.0373108126,-0.3129376173,0.1806302667,-0.1964211762,-0.0615591407,-0.0442417897,-0.1772384048,0.2160775959,-0.0785509795,0.3434203267,-0.1640195251,0.0547124855,-0.1044938639,-0.3085640371,0.2629107237,0.1635641903,-0.0282108132,0.2825821638,-0.0607512631,-0.0348773003,-0.2365302444,-0.3340730667,-0.3741346598,-0.0552138425,0.2754744887,-0.0566122793,-0.0085337404,-0.3372317255,0.0182861052,-0.0396474004,0.1264602989,-0.1866016686,-0.0632073134,-0.1109022647,-0.1256732792,0.4100495577,0.2269438207,-0.0682605281,-0.2711590827,0.3355163634,-0.3717980385,0.1890808195,0.1869220287,-0.1268949658,-0.0194468331,-0.1061921418,0.1100192741,0.1655587703,-0.3452140987,-0.3684436679,0.1967104226,-0.2838747203,-0.0734609514,0.0429037847,0.0068499376,0.0907800794,0.1455781758,0.2033013552,0.2076736391,-0.0318414196,-0.043676544,-0.1337381154,-0.2656517625,0.0427703001,0.2702589035,0.0227740202,0.0227792356,0.1320153177,-0.0515481979,0.2843168676,0.0502961464,0.0927804336,0.2592829466,0.2045941204,0.2270036787,0.0053051566,-0.3561691344,0.0282296278,0.0911857635,0.0471842065,0.2919330895,-0.0791223496,-0.1105626151,-0.3629341125,-0.1520826221,-0.1835043579,-0.0659433082,0.1536897123,-0.2198696434,0.1973173767,0.2385952175,-0.1191703454,0.2829247415,-0.1795577705,0.2405000925,-0.4697937071,0.4009589255,-0.1937504411,0.0068577528,0.0877900422,0.093488425,0.1160451248,-0.2712267935,-0.1282043308,0.4408646822,-0.2160945684,0.3192709088,0.389751941,0.0405472703,0.256257534,-0.458319962,-0.0476714186,-0.0960997418,0.0659594163,0.0526324585,0.2186976671,0.2048560381,0.0553214327,0.1382392198,0.135387525,0.1918755323,0.3227006793,0.0606966577,-0.0891448855,0.0702666044,0.2163235098,-0.1981064826,0.0355753079,-0.1962534189,-0.2209892124,-0.0054439199,0.0077681267,-0.0190759338,-0.0522501655,0.0720913336,-0.1022571996,-0.0005349888,0.072940357,0.0690679401,0.2663550675,0.2399249077,-0.0378363021,0.2792978883,0.0580011383,-0.1021743342,0.185502544,0.0684306175,0.0449506678,-0.0382117555,-0.1211706698,-0.1182554737,-0.2871050835,-0.0011871299,-0.2210567147,0.179521665,-0.247075066,-0.0214799866,-0.0927760676,-0.3123773038,0.0658893958,-0.1617256105,-0.2731583118,-0.3480960429,-0.0534439236,0.1365808845,0.0331345461,0.1988117844,-0.3494465649,0.1007373407,0.155825004,0.0168126933,-0.1158866808,-0.1174212024,-0.1286856532,0.1565780342,0.0093747275,0.0626327097,0.3976941109,-0.2893912494,-0.1418686509,-0.105853945,-0.2319445759,0.0629834086,-0.1083935797,0.3044180572,0.277816087,0.4564883411,-0.086418204,-0.0520351678,0.605652988,-0.1802677661,-0.0275544822,-0.0341372415,0.0227979887,-0.1331533045,0.1109787449,-0.299169749,-0.5104176998,-0.2131534219,0.1405189931,-0.0693684369,0.1047188267,0.1160707101,0.0068443767,0.1172981635,-0.1809821874,0.1372247487,-0.3094083667,-0.8145682812,0.1136590019,-0.0933670551,-0.4647526443,0.1229601279,0.1953829229,0.2569904029,0.1191847622,-0.3289041221,-0.3935401738,-0.033409223,0.09856911,0.1863355637,0.140938893,0.1660903245,-0.2663914859,-0.1748818904,-0.0119509129,0.026689576,-0.0368766636,-0.2150575817,0.4406511188,-0.0084777055,0.4517733157,-0.0168994535,0.4030779302,0.2012053132,0.0466530882,0.4095164835,-0.3087844551,0.2212135047,-0.0356283337,-0.603844583,0.2042991966,-0.1897949129,0.0618310198,0.0096048834,0.1967375427,0.2806479037,-0.4943080246,-0.2354852557,-0.2183479965,-0.271551609,-0.2354330122,-0.0054231891,0.1608945727,0.0405542143,-0.1433330327,-0.0249078274,-0.0667842552,-0.0164611228,0.5720892549,-0.0693982393,0.1114212349,-0.6191424727,0.0422341228,-0.6355404854,0.3562630117,-0.1448629498,0.1651325673,-0.2958746552,0.0792203099,0.0938228518,-0.0150768524,0.7718532085,0.0018984537,-0.0179304443,-0.1105164513,-0.1907024384,-0.2759567201,0.1091517061,-0.008985199,0.4083504081,0.4919059575,0.2473710179,-0.3470739722,-0.1145486012,0.0872872248,-0.0210563932,-0.2440116405,-0.209492445,-0.3656122088,-0.2786716223,-0.1774246395,0.0564010181,-0.0437285006,0.3815519214,0.0810395256,-0.2656276524,0.0091093658,-0.0542900674,0.2826078832,0.03852598,0.1447791159,0.0094427085,0.1691836566,0.3119984567,0.42080459,0.1319475323,0.4850691557,-0.2108382434,-0.2717648447,-0.0984516889,0.1013869494,0.1231099442,0.2183836401,-0.194126308,0.2023682445,0.2547457516,0.4601501226,-0.170676291,0.2522602975,0.2773421109,0.2792245746,-0.3641133606,-0.4865392447,0.1555067152,-0.0917388946,0.0943062603,0.2631676793,0.2964454889,-0.035594862,-0.1662316024,-0.2842913568,1.0478123426,-0.0956733227,0.0067074695,0.1944355071,0.1370845586,0.4507586956,-0.0921948999,0.14174968,-0.2144227773,-0.2949659526,-0.0541798845,-0.1743752211,0.1680574566,-0.2362244576,-0.0537432209,0.3484443128,-0.0203507338,0.0292559173,-0.1320039481,0.2688501477,-0.296713531,0.0120330295,-0.3091320992,0.255362004,-0.1343014538,0.4604220986,-0.0302182641,-0.1518836021,-0.1035428718,-0.3361431658,-0.0636127293,0.1191884577,-0.2714743018,0.1095891446,0.1017429084,-0.2224692702,0.0017539592,0.1140809506,0.0938335285,0.0761765838,-0.2716825604,0.2065556496,-0.5015836954,-0.1528689414,0.1270386875,0.1177479699,0.1461581588,-0.2006833106,-0.493542552,0.1187355667,0.1249828935,-0.1905112714,-0.051073011,-0.1371843964,0.0943688229,-0.470569849,-0.1956136078,-0.3145422935,0.0358889513,-0.196957618,0.2031294554,0.1291759014,-0.2606883049,0.093567498,0.2006009668,-0.1175684035,-0.0630932078,0.5965725183,-0.0924626812,0.0054062055,0.4578071237,0.1677633226,-0.138016507,-0.1807493865,-0.0779089853,0.2252724767,-0.2601616085,-0.044652354,0.0442487113,0.1300584376,-0.208308205,0.0384569429,0.2260815352,0.009782684,0.2754867077,-0.5902446508,-0.4207306802,0.1600779593,0.0696798712,0.1743370593,0.2952528894,0.3555336595,0.1188822165,0.0842632279,-0.3726804256,0.01419417,-0.2579594254,-0.0338925496,0.3824023306,-0.3016345799,0.3503540754,-0.1016359329,0.2228459418,-0.064188242,-0.3282786906,-0.228940174,-0.0984089747,0.0898085982,-0.0183257833,-0.38398844,-0.0003593418,-0.119276382,-0.06859079,0.0058197044,0.1239214838,0.1958062053,0.0229750182,-0.1963895261,0.0738457814,-0.0365415439,-0.0993711799,0.155978024,0.1055222079,0.2584663928,0.0365741067,-0.0811827406,-0.1045225933,-0.2432992756,-0.2736442089,-0.0624735951,0.2513032854,-0.1278461963,0.2676116228,-0.1688304991,-0.0725648403,0.1447963268,0.5498507619,0.3072592318,-0.2752390504,0.3285318911,0.1525736898,0.2810035348,-0.2495319843,-0.0088571878,0.2969357967,-0.0427344367,-0.081926398,-0.1622228324,0.0722938776,0.1996451318,0.0240235087,0.0446379818,0.5042868257,-0.0238164719,0.178891018,-0.0350746326,-0.0562042035,0.0050123557,0.1521483213,0.232099399,-0.0585869029,0.3224658072,-0.3049378693,0.3366059065,-0.2997869253,0.2508377135,-0.1354233623,-0.3039555252,-0.1447141916,0.1380612552,-0.2659606636,-0.108838059,-0.2559172809,0.7806465626,-0.3334018588,-0.0139065627,-0.2232389003,0.3534821868,-0.0758548379,-0.3060179055,-0.0491390042,-0.2112845927,-0.0329262763,-0.1981339455,0.1309959143,-0.2458434403,0.0111714173,0.034395922,-0.0802687705,-0.5813345909,-0.1702587008,0.1976888478,0.1456969976,-0.0387902036,0.2339258343,0.2654058933,0.1201949269,0.0710984915,0.3345949054,0.4603468776,0.2285658568,0.1212587059,0.102303572,0.1440177858,-0.0438807085,0.0165887736,0.0583188646,0.1513800472,0.0268123019,0.279459089,0.2609655559,-0.2058109492,0.0252850801,-0.0916614309,0.3647478521,-0.2482298464,0.3037289381,-0.1616885513,0.0187716037,-0.4195710123,0.0384396873,-0.7279619575,0.3941763341,0.1401265562,-0.0890669376,0.1646356136,-0.3102861345,0.1091733202,-0.1697917283,0.5486783385,0.2740631998,0.262169838,-0.2371128798,-0.1858242601,-0.5177457333,0.1752017587,-0.0475243665,-0.2880723476,-0.0966670141,0.0466473252,-0.1142756939,0.0965362266,0.1236988977,0.0150240939,0.1897215247,-0.0559559613,-0.1848658919,0.0535531268,-0.3134872615,0.1869217306,0.1249472201,-0.1335354596,0.2388921231,0.1444466859,0.0627690181,-0.0373318307,0.2129230648,-0.267439425,-0.2997754216,0.2467645258,0.1838709563,0.5290333629,-0.0266213939,0.0692393482,-0.1084940135,-0.4502314925,-0.2254646271,0.0992140248,0.1485068947,0.1336231828,-0.0102896458,-0.2486870438,-0.2791526318,0.4303966761,-0.0519702509,-0.3362524509,-0.0394177437,-0.0630524457,0.1515261978,-0.0076329978,0.1565960646,0.3153595328,-0.0226594172,0.0860468894,-0.098981753,-0.2700575888,0.6556456685,-0.2379179895,-0.0130219683,-0.251904875,0.2108675092,0.3994065821,-0.1213897839,-0.4444725513,0.253652662,0.2380932868,0.0883379802,-0.0961514339,0.4788861275,-0.0321511626,0.1870010048,0.0243745428,0.0740242377,0.0143951904,-0.1268232465,0.1934874058,-0.0743513182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1796","title":"Filter on dataset too much slowww","comments":"When I use the filter on the arrow table directly, it works like butter. But I can't find a way to update the table in `Dataset` object.\r\n\r\n```\r\nds_table = dataset.data.filter(mask=dataset['flag'])\r\n```","body":"I have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n","comment_length":32,"text":"Filter on dataset too much slowww\nI have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n\nWhen I use the filter on the arrow table directly, it works like butter. But I can't find a way to update the table in `Dataset` object.\r\n\r\n```\r\nds_table = dataset.data.filter(mask=dataset['flag'])\r\n```","embeddings":[-0.2712090611,-0.059731666,-0.114763163,-0.2779481411,-0.170650363,-0.2504285872,0.1707617491,0.2090248764,0.2598769367,-0.1163968369,-0.0861833245,0.2958269715,-0.0557140596,0.4060816467,-0.180425927,0.2453580648,0.1307907403,0.297056675,0.1960131675,0.0333571322,-0.0224025566,-0.1882677227,-0.2929413021,-0.0703883544,-0.0096254768,-0.2621649802,0.1859868318,-0.3086514473,-0.2785188854,-0.2126028389,-0.0173911992,0.5965904593,-0.0182769466,0.5038164854,-0.0001307184,-0.2338380963,0.2439537793,0.0724606216,0.0831183866,0.1498079598,-0.2113010138,-0.1286040395,-0.014037692,-0.0004979747,-0.1250068545,0.002729736,-0.3317998648,-0.2216515541,0.3499701023,0.0944688246,-0.0277750231,-0.1412250102,-0.2549342811,0.3521134555,0.2052817196,0.3541748226,0.1411342919,-0.1513690203,0.5533320904,-0.3358145058,0.0539450236,0.6059803367,-0.4179320335,-0.1313287318,0.2100652605,-0.2898597717,0.0583834425,-0.1515574008,0.2810197771,0.084834516,0.1335704476,-0.1620167792,-0.2986915112,-0.2322742343,-0.0849645361,-0.4459281862,-0.0143870069,-0.093405813,-0.2579454482,0.0345795825,-0.5128167868,-0.3484371901,0.0861519277,0.1266560107,-0.3261454105,0.3600158095,-0.006306435,0.2096572816,0.1311976016,-0.2452675253,0.06995029,-0.2122394294,0.1913420558,0.3051820099,-0.6571980119,-0.1016245335,-0.096286118,0.1605431437,-0.0209174547,0.0603329986,-0.184784472,0.5039317608,0.2157584131,-0.0072758566,0.0003438363,0.2503774464,0.1227045953,0.4484069049,0.280266434,-0.145333305,-0.2991339564,0.1851134747,-0.0733834952,0.0487983041,0.0876706019,-0.057178542,-0.3887727559,-0.1275837868,-0.3137276471,-0.0533896498,-0.2973858416,-0.059795659,0.6444664001,0.5041763783,-0.1133061051,-0.0650423318,-0.0732961521,-0.2993195355,-0.1293625981,0.017078558,0.032341212,-0.1580843776,-0.1048117131,0.2392276824,0.0378134847,-0.2345514894,-0.1911473572,-0.0324770808,-0.1686941683,0.1806276292,0.0712083355,-0.3621056974,0.2372625619,0.413911283,-0.2685701549,0.3007687032,0.0842118859,0.055668354,-0.1265523136,0.1775274575,-0.3156157434,-0.3334103525,0.2214105278,-0.0167185534,-0.0847421587,0.1368683875,-0.1805811822,0.4090796113,0.2862333059,-0.2051102072,-0.1643706262,-0.1768735349,-0.595266521,-0.1606847793,0.2633840442,0.2223779261,-0.5520042181,0.0523661226,-0.3266305327,0.2614666224,0.4948300123,0.3584709466,-0.0671725869,0.2171266228,0.1002386138,-0.070361577,0.2984155416,-0.0785397142,-0.4758496284,0.5441429615,-0.0614160672,0.012077041,0.1328438818,0.155942142,0.5201342106,-0.2442356944,0.3975709081,0.337546438,-0.1997921616,0.0735185072,0.0002533132,-0.0091579854,0.2259215713,0.1488590688,-0.3431073427,-0.022170566,-0.0321321823,-0.426648289,0.1524595916,0.1380381584,-0.0138798021,0.2082129866,0.2469478995,0.1468655616,0.0114980666,-0.1173818856,-0.002903319,-0.0856793895,0.0895480141,-0.2041616291,-0.2678201497,-0.2276494056,-0.2681233287,0.1229263097,0.0763175488,0.2027755231,-0.1417338103,-0.1732809395,0.1918347627,-0.2997716069,-0.0245427042,0.4423805177,-0.1199391857,-0.0766720176,0.1876756102,0.0041585388,0.3773688972,-0.1478652209,0.1558626592,-0.0638166666,0.2566291094,0.0597918741,-0.0512209646,0.2228671908,0.4853754342,-0.2909352779,-0.3249587119,0.1911716908,0.0907264799,0.3223301768,0.1594703197,-0.0299946852,0.2172129899,-0.0669633523,-0.19752644,0.5967999101,-0.0842720866,0.4515283704,0.0961709395,-0.0534643419,0.3551806509,0.2579828501,-0.1379966438,-0.224004209,0.2483920008,-0.0168593042,0.1414318383,0.191679582,-0.1041448414,0.3914271593,-0.0069532744,-0.1684072316,-0.0578750633,0.2389459014,0.1703025848,0.0307574421,0.1890267134,0.0051107556,0.1146775112,0.1844976097,-0.1427407414,-0.0982926339,0.081593819,-0.0547583327,0.2547175586,0.1561012119,0.3075129986,-0.2542977333,0.3530647159,-0.1137326285,-0.1732011735,-0.1776498109,0.2034720778,0.333828032,-0.2167353928,0.0699750334,-0.3757014275,-0.535559535,0.2271120101,-0.1121494323,0.22655797,0.0880119279,0.2230781019,0.0783291161,0.1746549904,0.285586983,0.19949992,0.1593718231,0.166126743,-0.4469213784,-0.1344804317,-0.1596938968,-0.0713537633,-0.1074967459,0.2684340775,0.3947562873,0.3735952377,0.2810976207,0.048986651,-0.1861354113,-0.5580239296,-0.2156649977,-0.0151027096,0.0744738132,0.2174893618,0.1304199845,-0.3905109465,-0.050193429,0.0361119956,-0.1326581687,-0.1809353977,0.0968123153,-0.264318347,0.0712426528,0.1325804293,0.1171383932,-0.23426494,-0.0533800982,-0.1480298936,-0.0901711583,0.289006263,-0.2378833294,0.0384709165,-0.3095988929,-0.2900487483,-0.0745963454,0.062701799,-0.1542517245,0.2361339927,0.2845228016,-0.0103315674,-0.1767285317,-0.0545509867,-0.2922154367,0.5495020747,-0.2610016763,0.1806714833,-0.4031961858,-0.286221981,0.1927287281,0.0184681565,0.3718920946,-0.1154165938,0.1365425885,-0.0016428275,-0.4650854468,-0.1635533124,-0.1761951447,0.0963357612,0.2124688625,0.8827466965,0.0940158293,0.3070503473,0.0628353581,-0.0016844188,0.0701517612,0.2428095192,-0.1153798103,-0.2179842293,-0.336771965,-0.2030336112,0.1458653063,0.0642521009,0.0387095511,0.1361506879,-0.1435865462,-0.0339635089,-0.2271154225,-0.0657965764,-0.1354943961,0.094266504,-0.2203656882,0.2903557718,0.3504481614,-0.0555741489,-0.5428822041,-0.3475463092,-0.2544793189,0.0263413955,0.1311549246,-0.1841544956,-0.6282449961,-0.318633914,-0.3433012664,0.4391598403,0.1865683198,0.441232115,0.1740358919,0.0723440945,0.1996858865,-0.0993004516,0.7873359919,-0.179536134,-0.0000881626,0.0532541461,-0.5388451219,-0.4195431769,0.0940796807,-0.0101871816,0.2481472194,0.2991225421,0.3818729818,-0.277079612,-0.3589819074,-0.040796265,0.12422245,-0.1010713056,-0.1960698962,-0.2369445711,-0.2199009657,-0.3587256372,0.2803162932,-0.0494059362,0.2188858092,0.3615349233,-0.3112304509,-0.0476938225,-0.1954779029,0.1896836758,0.049327001,0.0551218987,0.3121325672,0.3041161597,-0.0797566697,0.1950198859,0.0565646924,-0.1036900729,-0.1127292663,0.0270732157,-0.009185832,0.1956090629,0.3815271556,0.5263146758,-0.1246369407,0.5532614589,0.0629926994,0.424169302,0.0678064376,0.2134750485,-0.1809375435,0.1748359203,-0.5886353254,-0.7435212135,0.1466280073,0.4004863203,-0.3319710493,0.5015232563,0.1510134637,-0.0640742481,0.4643263519,0.0759486109,0.8864929676,-0.3235602081,-0.1276146919,-0.0528230146,0.0650778264,0.3599280715,-0.1648576856,0.119332701,0.0267442744,-0.3669936955,0.0609795153,-0.0934690461,0.0710566044,0.0802966654,0.0669524968,0.3530903459,-0.1266107112,0.0602889732,0.0606924333,0.2552217543,0.2188882381,-0.3514135182,-0.244919464,-0.0771519989,0.1735336781,-0.1433549523,0.044794146,0.0445488356,0.2399152517,-0.3018660247,-0.21136114,-0.5126231909,-0.4052533805,-0.0081724087,-0.1805772632,-0.1766787618,0.1086182147,-0.2058490068,-0.2409984916,0.4316933751,0.1569402069,0.2249771804,0.2834919095,0.3169202209,-0.0349664651,-0.3678517044,0.2215467989,-0.1324400604,-0.1802753508,0.0131076518,-0.0311694592,-0.5412837863,0.0419516042,0.1744588464,0.4846715033,0.0853864551,-0.1789310426,0.0405986309,0.0019731526,-0.2304074764,-0.0539916381,0.0655844659,0.0918431357,0.6818690896,0.0109314295,-0.3466662467,-0.0823412761,0.2244322002,0.2488310188,-0.1049904451,0.3481358588,0.1715012938,-0.3952924907,-0.1170046553,0.0336576402,0.2506342828,0.1648853719,0.1414195746,-0.1345140785,-0.2305841446,0.4424909949,0.0179605354,-0.1558717489,0.0163455568,-0.1148778126,-0.6028130054,-0.1591872275,0.0887798294,-0.033945933,-0.2836894691,0.0792492852,0.1788769811,-0.2286403626,0.4104103446,-0.1429885924,-0.1034457162,0.0149187781,0.1952461302,-0.2978616953,0.3553290963,-0.1092388034,-0.0254916232,-0.058435794,0.0818289667,0.0158344395,-0.1095690057,0.1582012177,0.1992669404,-0.1821382791,-0.1482774913,0.1193575934,-0.319293648,-0.2785274684,-0.1667119265,0.0863983855,-0.0945423469,0.0135389436,-0.6226950288,0.0796450451,-0.3068782389,-0.0308547691,0.1479804963,-0.0689831153,-0.2993613183,0.0393475667,0.1808097512,0.2547493577,-0.0701490268,0.0305705033,-0.0425859764,0.008794371,-0.1373348385,0.1609221697,-0.0131394239,-0.0361738354,0.1334585845,0.1339655668,0.3425039649,0.1542834938,-0.2189279944,0.4608941674,-0.0892685205,-0.324628979,-0.1629707515,0.1868814528,0.1323443204,0.191449821,0.1589532644,-0.143294692,0.4014587402,-0.1859279275,0.1897576451,0.2408556938,0.0335241593,-0.1840328276,0.3610528111,0.3969342113,-0.1160279214,0.402451843,0.3945572972,-0.22622706,0.6142657399,0.222492829,0.4195976853,0.4634606838,0.4451642334,-0.3034519851,-0.3380827308,-0.2056551278,0.62243855,-0.1491276473,-0.2401643991,0.1059998348,-0.2050468922,0.4029743373,-0.4408563673,0.1021219566,0.3491744399,0.2446169108,-0.1234745681,-0.4045478404,0.0168199521,0.1725481451,-0.0365309231,-0.0128714629,-0.0016640153,0.2551496327,0.0111501999,0.2619935274,-0.3760182261,-0.2330495417,-0.1158901826,0.2028701007,0.0230053049,0.1552746892,-0.2527198195,0.3291514218,0.3357107341,-0.3138510883,0.1728478223,0.424010694,-0.0268687941,0.157084778,-0.3043255508,-0.1973879635,-0.0152562112,0.3792417943,0.1525566727,0.1451670378,-0.0693076923,-0.1082880422,-0.0344024077,-0.025302615,0.2767591476,0.1954934895,0.0847867429,0.3311166465,0.2029563189,0.0142709622,-0.0384322219,0.0576427802,-0.1150836051,0.2869706154,0.2904757559,-0.335318327,0.168660894,0.1297983378,-0.0057099378,-0.2572619617,0.1182305142,-0.0713957176,0.3901482224,-0.2078109235,-0.3351299167,-0.4416581094,-0.3361499608,-0.3414925337,-0.0613200851,0.144658491,0.0870982632,0.2598431706,0.0411545373,0.1491354406,0.0392584801,-0.2425340116,0.0631310195,-0.1808049679,0.3376826942,-0.1651826203,0.3906686306,-0.0743537247,-0.3516141772,0.3682199121,-0.2127816975,-0.0578486584,-0.0417921394,0.0814252123,0.2041239291,0.1451190114,0.0432647876,-0.014580491,0.360032618,0.1521870494,0.5289028287,-0.0460165404,0.0723427013,-0.3264321387,0.259090811,0.2396050543,0.1181286126,-0.0328038968,0.0460817628,0.1615455598,0.0981580392,-0.1329068989,-0.1898031533,-0.0765680596,0.1568205208,0.1963785589,0.3022772968,-0.3197207451,0.108046934,0.0558899306,0.0469930656,-0.305877924,-0.0662635937,0.5844329,-0.5551506281,-0.2744622231,-0.3311441243,0.226670295,0.400195837,0.0762991533,0.053293407,-0.1094825566,0.1079561487,-0.1184766963,-0.3128961027,0.3040601611,0.0193845332,0.0772828385,-0.2103271931,-0.134958446,0.0183361601,-0.1712665856,-0.4492625296,-0.187703073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1796","title":"Filter on dataset too much slowww","comments":"@thomwolf @lhoestq can you guys please take a look and recommend some solution.","body":"I have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n","comment_length":13,"text":"Filter on dataset too much slowww\nI have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n\n@thomwolf @lhoestq can you guys please take a look and recommend some solution.","embeddings":[-0.2553203404,-0.2769626379,-0.1042277515,-0.191494599,-0.1148009077,-0.2562538683,0.2120214105,0.2069259435,0.2685796916,-0.1292301714,0.0151763875,0.1192787886,-0.1280844361,0.4037200809,-0.2452518791,0.1354870647,0.0983364955,0.2222633064,0.3897942603,0.0320132673,-0.0664342046,-0.0890462995,-0.3774906397,-0.012524101,-0.0734490305,-0.2800956368,0.2880972922,-0.2499666512,-0.2650186121,-0.1744286716,-0.1335566938,0.4177105129,0.0282326434,0.5295689106,-0.0001289329,-0.3831353188,0.2667166889,0.0997707769,0.2290613353,0.1209235042,-0.1769671738,-0.0017297317,-0.0306571405,-0.0622256584,-0.2244420052,0.0726448819,-0.284576565,-0.2326079458,0.4449988008,0.1445551217,-0.024221966,-0.2159018815,-0.3314349353,0.3760126233,0.1414524913,0.3206852973,0.1831285655,-0.2321004272,0.6945299506,-0.3379615545,0.0822385177,0.4718360603,-0.4729018509,-0.1759243459,0.1419285983,-0.3444545567,0.1215578541,-0.3156278729,0.2314595431,0.1786624044,0.0671810508,-0.0407974496,-0.1858372092,-0.2302809805,-0.0464883149,-0.3982253373,0.0256667845,0.0463561863,-0.3080676198,0.0106547475,-0.578540206,-0.215866968,0.0843120888,0.1577578038,-0.3494929373,0.380787909,0.0395921133,0.1336714476,0.1742640436,-0.2076579481,0.0233311616,-0.2517266572,0.2680599391,0.3421780169,-0.6742807031,-0.1620527059,-0.1782374829,0.2304182351,0.0111018466,0.0055427994,-0.3759352863,0.4775307178,0.2454158515,-0.0296255331,-0.0725487471,0.2293362767,0.0591038577,0.4937187135,0.3134313524,-0.2362648547,-0.3152812719,0.2554736733,-0.0456468873,0.0689831749,0.1086279526,-0.0607486889,-0.5013031363,-0.1777885109,-0.2818199694,-0.1122500971,-0.4077599943,-0.124199681,0.6695059538,0.487077266,-0.105280146,-0.1086201072,-0.1288596243,-0.2565720677,-0.1508439481,0.0154243726,0.0222908258,-0.0478031971,-0.0522766858,0.1973616183,0.0584618859,-0.3054198623,-0.1705264151,0.0589537546,-0.1982086599,0.1115731001,0.2333503366,-0.3865981102,0.109832935,0.2916238904,-0.2720002234,0.3009663522,0.0748446956,0.1162619516,-0.1135890335,0.1238972843,-0.4159632027,-0.3373242915,0.219429478,0.0112007624,-0.0937117413,0.1789090633,-0.0857739076,0.3419914544,0.3026427925,-0.0791412517,-0.2131774276,-0.1863799542,-0.6024045944,-0.1636570096,0.1744667888,0.1082698256,-0.4512965679,0.0203973465,-0.2661448717,0.2019932121,0.6189799905,0.4511259496,-0.0482484587,0.2244680524,-0.052831579,0.0186868589,0.2602067292,-0.1181961894,-0.3548624218,0.6506745815,-0.0423517451,-0.0107837347,0.1948289126,0.1154088825,0.5478641987,-0.2544558644,0.5040104985,0.3734447658,-0.1935945004,0.1335345209,-0.0208086967,-0.0970563069,0.1705562025,0.2245894819,-0.2910606265,-0.0645191371,-0.0523105077,-0.3646586835,0.1849197894,0.0254861619,0.0193281751,0.1882749349,0.2404324114,0.0543331727,0.0189543217,-0.1881296486,0.1105268523,-0.0767451748,-0.0561440103,-0.1633724272,-0.1641264558,-0.1743017882,-0.2498283535,-0.0143377809,0.0437901914,0.2335148305,-0.1188744232,-0.1754950732,0.1294328123,-0.2272019684,0.052913446,0.4961780906,-0.0836527571,-0.0426002964,0.1938756555,-0.0381508395,0.3340575993,-0.1073297709,0.1675846875,0.0102449618,0.3772317171,0.00358121,0.0375690721,0.1452276856,0.5068955421,-0.29597947,-0.3014855385,0.166558072,0.1545439214,0.3242525756,0.1153020933,0.0267268866,0.146146521,-0.1152107716,-0.1317740083,0.5105996132,-0.0455397628,0.4169857204,0.0637420267,-0.0999688953,0.317958504,0.1632632166,-0.0611112751,-0.2041215152,0.4905211329,0.029856937,0.1302554011,0.2602405846,-0.1035686955,0.2969912887,-0.0130712297,-0.2937028408,-0.0522404909,0.2735585272,0.1774273515,0.0913492367,0.1303706765,-0.0362082422,0.0530435033,0.2039971203,-0.2109454423,-0.1462622732,0.0972649679,-0.0480939038,0.2681371868,0.0893004313,0.2402095348,-0.2201187611,0.2573031783,-0.1293806583,-0.1028938442,-0.2857192755,0.2317534685,0.3478753269,-0.2277269214,0.0071077249,-0.3442222178,-0.5610300303,0.2319212109,-0.0821667314,0.2094505429,0.0411813334,0.1571984291,0.0568297841,0.2992652953,0.1247127801,0.1410211027,0.170296818,0.2411869466,-0.4621629119,-0.1151067093,-0.1283653229,-0.0506344028,-0.0647403821,0.2661579251,0.3184905648,0.3828890026,0.2978678942,-0.1160141528,-0.125516668,-0.519257009,-0.1500781327,0.0090796556,0.1495730728,0.1240779981,0.1400052309,-0.318559885,-0.0992041007,-0.028929919,0.0027084923,-0.1763915122,0.0693364665,-0.2148922831,0.063613154,0.1176343635,0.1331204176,-0.2881636918,-0.0214367304,-0.2121097893,-0.0829125121,0.2898828089,-0.317124933,-0.1363794655,-0.2373524904,-0.2826256454,-0.0529266149,0.0135930181,-0.2037442476,0.285695076,0.2899574935,-0.0284799822,-0.1294320822,-0.0698765814,-0.3206011057,0.5137388706,-0.2495615333,0.1332132667,-0.4222961664,-0.2307976037,0.1691938937,-0.0362476259,0.2602393627,-0.1547907591,0.1570783406,0.0515004583,-0.2939304709,-0.1805712134,-0.1884307563,0.2711237073,0.2823275626,0.7504600286,0.0293998681,0.3123113513,0.1782704443,0.0840730891,0.0272313301,0.2509970367,-0.3198777139,-0.1651544124,-0.3269864619,-0.0338167436,0.2712780535,0.0651869401,0.0715376213,0.1515585929,-0.1511166692,-0.0522267818,-0.2421470731,-0.0907476842,-0.1178933159,0.0886146128,-0.1492008567,0.2979524136,0.216075182,-0.0486029126,-0.4446672499,-0.3853903711,-0.2848783135,0.0764872432,0.1589096785,-0.1960804611,-0.5030329823,-0.2133108824,-0.3699952066,0.3729581237,0.2139894366,0.2745116949,0.0586863197,0.1613556445,0.1647468805,0.0047145477,0.8605701923,-0.1942650974,0.0318701565,0.0511978157,-0.5594775081,-0.264868319,0.1042214483,-0.0529429093,0.0757328942,0.4168176055,0.4154501259,-0.3104287386,-0.3421818316,-0.0478844345,0.1159402877,0.0012862341,-0.4004428387,-0.1730293483,-0.139406234,-0.3682291806,0.2941549718,-0.0434696004,0.2855573297,0.3116886616,-0.4380500019,0.0190492272,-0.2388848066,0.3014058173,-0.0229730289,-0.0472739711,0.144923687,0.2574340403,-0.0859378576,0.1564519256,-0.0109785302,-0.2500292957,-0.0183493793,0.0396656804,0.0891390666,0.0518802516,0.4200533032,0.4532893002,-0.1689916104,0.5998983383,0.0403594188,0.5063835979,0.0143164564,0.2121613622,-0.084114857,0.2228858471,-0.691970408,-0.6111631393,0.0634212643,0.3974199295,-0.3400244713,0.3578102887,0.0184896141,-0.0761035159,0.4325414598,0.0748185739,0.9730539918,-0.4304570556,-0.2325305641,-0.09901198,0.2407799512,0.4447362721,-0.1475916803,0.1138999164,0.0426894054,-0.3545999527,0.0240212101,-0.1576638967,0.0243099164,0.114530839,0.0388531201,0.3901435733,-0.0953421593,0.0345560908,0.0191774033,0.3376794159,0.293872267,-0.3435220718,-0.2051745802,-0.0613561496,0.0799668208,-0.0446943603,0.0180588439,0.0209743232,0.2605160773,-0.2045141906,-0.1898500174,-0.4529131949,-0.3691295385,-0.0306968242,-0.1786549389,-0.1451853216,0.2305676937,-0.1406867653,-0.1659258008,0.485332638,0.0449901298,0.2002005726,0.2249986529,0.2166668326,-0.0578862242,-0.4099061191,0.1719933748,-0.1755543351,-0.2186050117,0.0660814345,-0.0169639587,-0.4442357123,-0.050805144,0.1321567297,0.4896734953,0.0044777784,-0.2451889068,0.0018459624,-0.0141961947,-0.1714234203,-0.0362212509,0.1454568207,0.0642974526,0.7065261602,-0.0276592989,-0.2937303185,-0.1383033246,0.1196821854,0.3053269684,-0.2594481707,0.3081386685,0.2062759548,-0.4034247994,-0.1169759333,0.0236447528,0.1297991574,0.3514896929,0.0731555372,0.0232037324,-0.2572634518,0.4650072157,0.0398884676,-0.1640136391,0.0142626418,0.0099276425,-0.6089185476,-0.2262092382,-0.0055128615,-0.130278185,-0.2593255937,0.1010954604,0.2391535789,-0.2330726683,0.4070627689,-0.152532801,-0.1324643195,-0.0878071412,0.1976806968,-0.1722856611,0.1762806773,-0.1119215265,0.0883968249,-0.009074402,0.1083813161,0.0364237055,-0.1164651364,0.1347115934,0.1902955472,-0.1586977839,-0.2304452211,0.2693232,-0.3119254112,-0.271101445,-0.1457385719,0.0929621756,-0.0236081462,0.0525742099,-0.6441964507,0.0777320713,-0.5125806928,-0.1270813346,0.1197555289,-0.1508543193,-0.3093314469,0.0482873917,0.1999247968,0.1688140482,-0.0575977489,-0.0025253403,-0.0772159323,0.1448428035,-0.0844456181,0.2535162866,0.0307662692,-0.0843798816,0.2528878152,0.0997548476,0.1823928803,0.0614197515,-0.1485625803,0.4528777003,-0.0757407993,-0.29144153,-0.180291146,0.1859092861,0.1895304918,0.1498289704,0.1759894937,-0.2377126664,0.3685729504,-0.1848174036,0.0938247591,0.4147260487,-0.056066718,-0.237264961,0.2962656021,0.4328493774,-0.2046612203,0.4470288754,0.3614481986,-0.2110696137,0.5686612725,0.1945058852,0.3315460682,0.3485913873,0.5459007025,-0.2337538153,-0.3589886129,-0.1036473885,0.7670474052,-0.2649130523,-0.2053324431,0.0522936098,-0.2011773735,0.34039855,-0.3432568908,0.1169123724,0.4432993233,0.1409277469,-0.0483132638,-0.3500481248,0.0381282344,0.2415860295,-0.0281742755,0.0161570851,0.0177308675,0.1521712989,0.2019066811,0.2738354802,-0.2713567019,-0.2422729284,-0.1068615988,0.2050922066,-0.0688312352,0.1399031132,-0.2763532102,0.2509287596,0.3116754591,-0.3540735543,0.2119383961,0.4107790589,-0.0183114391,0.10551586,-0.3356378078,-0.1403146088,0.038651865,0.4328116179,0.1977692395,0.2181322575,-0.0489257984,-0.0724422559,-0.0693418756,-0.0925554186,0.2242997885,0.0879141465,0.1486904919,0.2475807071,0.2807063758,0.0485129803,-0.1412747055,0.0697876588,-0.1676377952,0.3969396353,0.2083111554,-0.2617925107,0.2280635983,0.2208370566,0.0063132322,-0.2435274273,0.0196674578,-0.0698949844,0.4340594709,-0.2946918607,-0.2999195457,-0.3372758627,-0.3162536323,-0.3227834702,-0.0271526314,0.0930402651,-0.0134436451,0.1941464245,0.0112112947,0.0260258205,0.1143032759,-0.2291910648,0.0571984425,-0.1319423169,0.3624452949,-0.0400894545,0.3499548137,-0.0773162022,-0.3215970099,0.5012843013,-0.2401418388,-0.0492975935,0.0412211455,0.1038544178,0.1294599622,0.0616872348,-0.1093721241,0.0433288701,0.4214927554,0.1092717573,0.5975204706,0.0873569548,0.1483055204,-0.2256640494,0.1781498045,0.333874464,0.0935279727,-0.009332831,0.0747174025,0.0847846046,0.2207491398,-0.1401593983,-0.2919771671,-0.0987366289,0.0724401921,0.1583242416,0.2948013544,-0.2533888221,0.1279874593,-0.0210516322,-0.0051564374,-0.2319864333,-0.1232576966,0.5373458266,-0.5094869733,-0.2460173517,-0.3483477235,0.2679461837,0.2730356455,0.2020679861,0.1286165118,-0.0493422188,0.169872418,-0.1289496869,-0.2752562165,0.3222987652,0.0246729031,0.0310956016,-0.2027093917,-0.2627892494,0.0881824568,-0.1332430393,-0.3150141537,-0.147654742]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1796","title":"Filter on dataset too much slowww","comments":"Hi ! Currently the filter method reads the dataset batch by batch to write a new, filtered, arrow file on disk. Therefore all the reading + writing can take some time.\r\nUsing a mask directly on the arrow table doesn't do any read or write operation therefore it's way quicker.\r\n\r\nReplacing the old table by the new one should do the job:\r\n```python\r\ndataset._data = dataset._data.filter(...)\r\n```\r\n\r\nNote: this is a **workaround** and in general users shouldn't have to do that. In particular if you did some `shuffle` or `select` before that then it would not work correctly since the indices mapping (index from `__getitem__` -> index in the table) would not be valid anymore. But if you haven't done any `shuffle`, `select`, `shard`, `train_test_split` etc. then it should work.\r\n\r\nIdeally it would be awesome to update the filter function to allow masking this way !\r\nIf you would like to give it a shot I will be happy to help :) ","body":"I have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n","comment_length":162,"text":"Filter on dataset too much slowww\nI have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n\nHi ! Currently the filter method reads the dataset batch by batch to write a new, filtered, arrow file on disk. Therefore all the reading + writing can take some time.\r\nUsing a mask directly on the arrow table doesn't do any read or write operation therefore it's way quicker.\r\n\r\nReplacing the old table by the new one should do the job:\r\n```python\r\ndataset._data = dataset._data.filter(...)\r\n```\r\n\r\nNote: this is a **workaround** and in general users shouldn't have to do that. In particular if you did some `shuffle` or `select` before that then it would not work correctly since the indices mapping (index from `__getitem__` -> index in the table) would not be valid anymore. But if you haven't done any `shuffle`, `select`, `shard`, `train_test_split` etc. then it should work.\r\n\r\nIdeally it would be awesome to update the filter function to allow masking this way !\r\nIf you would like to give it a shot I will be happy to help :) ","embeddings":[-0.2580144703,-0.1224426553,-0.086424768,-0.2601470947,-0.1652652174,-0.245699659,0.259375453,0.2613680065,0.3136883378,-0.1532164067,0.0066367225,0.2217891663,-0.1138353497,0.3481482267,-0.2551463842,0.2076891959,0.0726486444,0.2243813276,0.3868038952,0.048585169,-0.0817239657,-0.1438356787,-0.2755509317,-0.0574542768,-0.0197119191,-0.3511652946,0.2738279402,-0.1885665953,-0.2887648642,-0.2391116172,-0.0230933558,0.4542039335,-0.0112250363,0.4893327951,-0.0001260598,-0.3033389747,0.2190801948,0.0967456177,0.0969073474,0.1311701089,-0.1555372328,-0.0342102721,-0.0015716637,-0.0732351691,-0.138098225,-0.027774537,-0.2938837707,-0.3395695984,0.4358003438,0.0762713328,-0.0212181713,-0.1640912294,-0.3436227441,0.3906527758,0.148646459,0.3552731276,0.1600951999,-0.1355320364,0.6643034816,-0.3910477161,-0.0594231226,0.5160925984,-0.5235146284,-0.0935743675,0.2468020022,-0.2897862196,0.1473238766,-0.2044605464,0.225132063,0.1434737444,0.041435644,-0.1305047274,-0.2387340814,-0.311427623,-0.0841018185,-0.4326247275,0.0584406741,0.0501062013,-0.3467150629,0.0126457047,-0.5090168118,-0.243846342,0.0958384648,0.1211493388,-0.3476194143,0.3207414746,0.0449049138,0.1487958878,0.2462175488,-0.1956782043,0.0892603323,-0.2859349549,0.2442054302,0.3371780515,-0.5745357275,-0.1418312788,-0.1381054968,0.1576311886,-0.0300550237,0.0824053586,-0.2531941235,0.4543230832,0.2301663309,-0.0332320482,-0.0178977232,0.2066557109,0.0794800445,0.5078425407,0.2514458597,-0.1950930953,-0.1972020417,0.2502977848,-0.0304293092,0.0641797781,0.1051938608,-0.0808328241,-0.4626242816,-0.1382437795,-0.2418837398,-0.094039619,-0.3737149537,-0.094264105,0.6346338987,0.4418670237,-0.0783586502,-0.1341266185,-0.0508993678,-0.279555887,-0.1667898744,0.0838874578,0.0413886756,-0.1199249551,-0.0650092289,0.2181845307,0.0996672586,-0.2918973565,-0.2306297421,0.0494003966,-0.1871843338,0.1988064498,0.2259200066,-0.3643764853,0.2283469439,0.3430494964,-0.3088803291,0.3217532635,0.1475788355,0.0212170351,-0.1408422887,0.1872083843,-0.347283721,-0.3065002561,0.2215021998,0.0037915038,-0.1534255296,0.1652536243,-0.1067522243,0.3195845187,0.2384779453,-0.1845529974,-0.2099701464,-0.2296187431,-0.5389133096,-0.1934917569,0.2326156646,0.1322148442,-0.5897626281,0.0171470419,-0.3143188059,0.2596709132,0.5231363773,0.4517054558,-0.0743229315,0.2232606262,-0.0143170971,0.0126739703,0.2559303045,-0.08512678,-0.4165561199,0.6068706512,-0.0548085533,0.090655826,0.1787693948,0.1575143337,0.5775741935,-0.2714684606,0.4816095233,0.329998821,-0.2129823864,0.0879459083,-0.0300120395,-0.1323307902,0.1419948488,0.1850213259,-0.3507846892,-0.1207064763,-0.0437385701,-0.4507984221,0.2112187296,0.0466443896,0.0334007889,0.1372991502,0.262842685,0.0677381158,0.002427065,-0.1319780201,0.0437037013,-0.0079642255,0.0262256209,-0.1991896182,-0.2843853831,-0.2172747701,-0.2058331966,-0.0102893366,0.050219547,0.2281267643,-0.1286116242,-0.2074558884,0.1340384334,-0.2319640815,-0.0181153938,0.4373145103,-0.1152948588,-0.0412792563,0.1415910572,-0.0606077351,0.3979136944,-0.0976240933,0.1006910503,-0.0314055234,0.3001841605,0.0400555059,0.0454706252,0.1611023545,0.4899815917,-0.2906467915,-0.2897323072,0.1746491492,0.112086229,0.3210213482,0.1585340053,0.0342961885,0.1842804849,-0.0922882929,-0.2971901596,0.5812556744,-0.1030587703,0.4783247709,0.1044817641,-0.0394966491,0.3194071352,0.2093010694,-0.1225639656,-0.1995681375,0.4308719933,-0.0314352959,0.1005753577,0.2806020975,-0.0277568065,0.3283559084,-0.0127498945,-0.2614408433,-0.0546488613,0.2450743467,0.1661770493,0.0988611728,0.1766045243,0.0861362442,0.0534123145,0.2262300998,-0.1610709727,-0.1625951827,0.004000071,-0.0623852015,0.287443161,0.1224418506,0.2521753013,-0.2539429665,0.2011266053,-0.110584788,-0.116606459,-0.2918695211,0.1945299208,0.300825417,-0.230190888,0.0814738572,-0.343275249,-0.5400549769,0.2144392282,-0.0921164081,0.2033649981,0.0089937691,0.1777583212,0.0631354377,0.1822884083,0.1455850303,0.0892181769,0.2276866436,0.2094933838,-0.5467377901,-0.1483846158,-0.2128753811,-0.0616178475,-0.0649384856,0.3395971954,0.3524471521,0.391884625,0.3119646013,-0.1020379886,-0.1546333283,-0.4452131987,-0.1585076451,-0.0286889486,0.1104630008,0.1032709405,0.1338426769,-0.3872923255,-0.1545075178,-0.0147920363,-0.0254872162,-0.1784635335,0.1289550215,-0.1389271319,0.1105797812,0.0965419263,0.1577357054,-0.2108279318,-0.0657868311,-0.1922358423,-0.0803296119,0.3034641147,-0.1891367733,-0.0917611942,-0.22786735,-0.3064533472,-0.0987778306,0.0939662009,-0.1424891353,0.2768471539,0.3038761318,-0.0370685533,-0.0938280672,-0.1172878146,-0.3516655862,0.6226114035,-0.2873896062,0.1780132949,-0.4116136432,-0.2254093885,0.1542645395,0.032483656,0.2691565752,-0.0742154196,0.1376710683,0.0273184665,-0.2938502431,-0.18134664,-0.191617772,0.2125632316,0.2483040243,0.7651829123,0.127829209,0.3413033783,0.1974574775,0.1569409519,-0.0195175558,0.2924416661,-0.1555918157,-0.180123955,-0.3459396362,-0.1539196521,0.1461711973,0.0725648105,0.033955086,0.1557670832,-0.1621886194,-0.0385272577,-0.140811637,-0.047446534,-0.2200168818,0.1565882266,-0.2120187879,0.2928079069,0.2550818026,-0.0010025838,-0.5004399419,-0.3563649058,-0.2340479791,0.0245728288,0.1940565854,-0.1826559901,-0.4920963049,-0.1426581591,-0.4179569185,0.3404889405,0.1994544864,0.2822418213,0.0868157372,0.0365652554,0.1148640141,0.0004402943,0.8182596564,-0.1471986175,0.0154000679,0.1322244555,-0.4685988426,-0.460277915,0.1328984201,-0.0798726976,0.1653358191,0.2958440185,0.3902944326,-0.2847401798,-0.3065438867,-0.0625816956,0.1826979071,-0.0667350069,-0.2874946594,-0.1903697401,-0.2101447582,-0.3705480397,0.2802362442,-0.0175919645,0.2808855474,0.3762192428,-0.3433228731,0.0080510527,-0.2611861527,0.3136887848,0.0138876978,0.0069035119,0.2061003894,0.303176254,0.0193466637,0.1766962558,0.0116156125,-0.1248456389,-0.0295882262,0.0242842566,0.0136292987,0.0800657496,0.36168769,0.5575121045,-0.1202879548,0.5586543679,0.0533808768,0.4472558498,-0.0033800791,0.1796393394,-0.1985721886,0.1545091867,-0.6475355029,-0.6661444306,0.0821385235,0.3528614938,-0.3209252954,0.49585253,0.1206011549,-0.0689841062,0.4795326293,0.0689442679,0.9122279286,-0.4223338068,-0.1293202341,-0.0037964662,0.188713178,0.4288327694,-0.0892130882,0.1283966899,0.0789816901,-0.3048508763,0.0442273095,-0.1779508293,0.0511391908,0.0752148479,0.0846467242,0.3383791447,-0.1061581746,0.0073740035,0.0275460239,0.3325157464,0.1898484528,-0.3797965348,-0.146201387,-0.055665452,0.108428359,-0.1734426469,0.0630043149,0.0359469689,0.1965924948,-0.2196014822,-0.1999838054,-0.4815897346,-0.355247587,0.0192448664,-0.2322669327,-0.1909116656,0.1824404448,-0.213192597,-0.2439898252,0.4488253593,0.0917862654,0.1708309054,0.2580387592,0.2994173169,-0.0158793814,-0.4601495862,0.1972998232,-0.1218351871,-0.1465219259,0.0549800955,-0.0540582985,-0.5098172426,0.0483080707,0.2030020058,0.4885250628,0.0099243578,-0.2172056735,-0.0251093935,-0.0074489517,-0.1663379222,-0.0355488993,0.1535983682,0.0961094946,0.6945757866,-0.0525581688,-0.316873014,-0.0998816565,0.1070697606,0.1990881711,-0.2146319896,0.299444139,0.1840758324,-0.4404842556,-0.1097445413,0.0722016841,0.2474161088,0.2998215556,0.1541187465,-0.0664503276,-0.215815261,0.4328508675,0.0284718256,-0.0817720667,0.0312947072,-0.0928012356,-0.5693457723,-0.2018580139,0.114426598,-0.0439683013,-0.2117531002,0.0614688806,0.1806743443,-0.2705644369,0.361346215,-0.1649131626,-0.0791563094,-0.0128516546,0.2124870718,-0.254698813,0.2238381654,-0.0781839117,-0.0072289738,-0.0163495187,0.0788105577,-0.0027591824,-0.1311949044,0.1103281155,0.1862534732,-0.1803196222,-0.2964834273,0.2364366651,-0.3541324437,-0.2470614463,-0.2154276371,0.0686766133,0.0040949797,0.0105029419,-0.683689177,0.0963220745,-0.4543239474,-0.1039836258,0.1455973238,-0.1669253707,-0.2937225103,0.0120160719,0.2149537504,0.1863130033,-0.0142357256,0.0982929468,-0.0320920944,0.0350415781,-0.0960116908,0.2655916512,0.0278560892,-0.044338841,0.1940906942,0.1606880128,0.3102345169,0.0813849196,-0.1666775793,0.48343876,-0.059180662,-0.2837421894,-0.2047854066,0.1989349127,0.1791320592,0.1836616248,0.1498323828,-0.2726891637,0.3341100514,-0.2075674534,0.1090354547,0.3740884662,-0.0703384504,-0.2158269584,0.3936489224,0.4317412674,-0.2141314447,0.4373224676,0.3448053896,-0.1633817703,0.6248002052,0.1625697315,0.3235560656,0.4254294038,0.5249886513,-0.3500039577,-0.4269688427,-0.0866610184,0.6189131737,-0.234491244,-0.183445394,0.0928305909,-0.2442388535,0.2608452737,-0.3488543034,0.1344257444,0.3292419612,0.1828830242,-0.0258944016,-0.3440887034,-0.0044808853,0.2810173929,0.0080984188,-0.044400502,-0.0167210773,0.15775989,0.1425016969,0.2987141013,-0.29461357,-0.192516908,-0.1364805549,0.2268216312,-0.0764760748,0.0530295372,-0.2720610499,0.2915358841,0.289827615,-0.3670114875,0.2098889649,0.4030947387,-0.0381555259,0.1471684724,-0.3606916666,-0.1813366413,-0.0150076253,0.4001683891,0.1676852107,0.1575707048,-0.0220812783,-0.0604053847,-0.0522032827,-0.1389400959,0.2176390141,0.0740095079,0.1071447954,0.3026019037,0.2705262899,0.0071112001,-0.0615187064,0.0523054972,-0.1224310994,0.3401498795,0.2895957828,-0.2651776671,0.2117100507,0.2369523048,0.0192529131,-0.2257686853,0.0452480838,-0.0455717407,0.3996000886,-0.2582890987,-0.247811377,-0.4215651453,-0.4059768319,-0.4042250216,-0.0365700312,0.1455504447,0.0629473478,0.2698787749,0.0955258086,0.0435491055,0.129807964,-0.2022348493,0.0265410077,-0.1372742057,0.3220697641,0.0594363883,0.3153896928,-0.1090594009,-0.3862200081,0.472835362,-0.2299935073,-0.0308477357,0.1162138209,0.0425747558,0.2271540314,0.0803696737,-0.0552264042,0.0678233653,0.3736478686,0.191531226,0.5193092227,0.0005668932,0.197870627,-0.2546800971,0.2054695338,0.2659472823,0.1792629212,-0.1063973233,-0.0012778378,0.0430169478,0.21058698,-0.1673601121,-0.1914366931,-0.1458096355,0.11354132,0.1374347359,0.3232653141,-0.3512881994,0.1273343265,-0.0530495606,0.0673583448,-0.2710545063,-0.0772756264,0.5529332161,-0.6079835296,-0.2792535722,-0.2729260325,0.2432747483,0.3418397605,0.1827060431,0.1242088303,-0.0995533243,0.1359942853,-0.1705013663,-0.3291724324,0.220258832,-0.0508968532,0.0331555977,-0.2044649124,-0.1727465838,0.0590475611,-0.1852393448,-0.3525728881,-0.1822319627]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1796","title":"Filter on dataset too much slowww","comments":"Yes, would be happy to contribute. Thanks","body":"I have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n","comment_length":7,"text":"Filter on dataset too much slowww\nI have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n\nYes, would be happy to contribute. Thanks","embeddings":[-0.2402773499,-0.2545272112,-0.1003498808,-0.2198818475,-0.1227757335,-0.2799421847,0.2207996249,0.2127729505,0.2623277307,-0.1228026971,0.0232397821,0.1090588048,-0.1218431965,0.3967724144,-0.2421645969,0.143516019,0.1019549593,0.2192169577,0.3881216943,0.0340878554,-0.0634120256,-0.0847114697,-0.3591637313,-0.0311744381,-0.071901612,-0.268815726,0.2820260227,-0.2297161222,-0.2636346519,-0.1719828397,-0.1159737855,0.4346099198,0.0226275586,0.5013754964,-0.0001285152,-0.3718090653,0.2548577189,0.0986165106,0.2319883108,0.1068777293,-0.1647320688,-0.0090217944,-0.0346706584,-0.0474451557,-0.221072793,0.0650268197,-0.2899990082,-0.2305462807,0.4283931851,0.1384053379,-0.0294803493,-0.200680539,-0.343520999,0.3633515537,0.1421847492,0.327050209,0.1868698299,-0.2343711257,0.692976594,-0.3630414903,0.0680446699,0.4875717759,-0.4760681689,-0.1712096632,0.1626761407,-0.3426041305,0.1024522036,-0.2979903519,0.2309891284,0.1791175902,0.0460553765,-0.0543734804,-0.191929087,-0.2528147399,-0.0438370444,-0.3954439163,0.0222031642,0.0338812619,-0.3197771311,-0.0062435637,-0.5776154995,-0.2062618732,0.0934391171,0.152448073,-0.3777622283,0.3774753213,0.0536322258,0.1188140139,0.1939512044,-0.2126911581,-0.0006158605,-0.2551400661,0.290050298,0.3536432981,-0.6685741544,-0.1517433375,-0.1613438576,0.2193557769,0.0082470374,0.0210214816,-0.3716861904,0.4654060006,0.2460116446,-0.0317276008,-0.0718037784,0.2296689451,0.0633137673,0.4889689088,0.2996124625,-0.2187758684,-0.3104633093,0.2654385865,-0.0527737774,0.0733158514,0.0868184119,-0.0593863726,-0.4943645597,-0.1513371319,-0.2783291936,-0.1036018431,-0.4129793942,-0.1236302853,0.678194046,0.4656475186,-0.1048275977,-0.1236505508,-0.1331508458,-0.2447561622,-0.1634279639,0.0168486796,0.0198041014,-0.041431617,-0.0534564741,0.1859338731,0.0823490024,-0.2942584753,-0.183765769,0.0437498577,-0.181648761,0.1273708791,0.231054917,-0.377917707,0.1286219358,0.2947853506,-0.297169745,0.305773586,0.0923086628,0.1021089628,-0.1134793237,0.1242509857,-0.3762093484,-0.3510634005,0.2301907837,0.00839194,-0.1149160862,0.1762961894,-0.0672422796,0.3427844346,0.2987914681,-0.0885774866,-0.2213286459,-0.2007402033,-0.5890280008,-0.1697222441,0.1975255758,0.1148280874,-0.4494901001,0.0093025193,-0.277892977,0.2113212198,0.6057111621,0.4472838342,-0.0363148153,0.2120772302,-0.0296171866,0.0407206602,0.2302911133,-0.1217505932,-0.3529136479,0.6208583117,-0.0441591553,0.0115740448,0.19527632,0.0968347043,0.5464392304,-0.2624704838,0.5159488916,0.3757754862,-0.2025654912,0.1274270713,-0.0133818658,-0.1021562815,0.1809444129,0.233576104,-0.2888844609,-0.0743050203,-0.0667566285,-0.3699983954,0.1762480289,0.0301596448,0.0160606895,0.1599272192,0.2311047316,0.0476133227,0.0113886474,-0.187184453,0.1131599993,-0.0571304746,-0.0464743227,-0.144448638,-0.1806621403,-0.1779678315,-0.2504740059,-0.0154906437,0.0667887852,0.2274637818,-0.1145353466,-0.1763185561,0.1198732555,-0.2212261409,0.0342432782,0.4862348437,-0.0723452717,-0.0304809529,0.1984912008,-0.0565858819,0.3468163311,-0.1130706817,0.1707818657,0.0338062756,0.3783162832,0.0169337559,0.0511227064,0.1511913687,0.5040012598,-0.2822240293,-0.312784344,0.1669359058,0.1463855803,0.3556897342,0.1092720479,0.0143428687,0.135141477,-0.1066777408,-0.159839958,0.5335882902,-0.0392724574,0.4193562269,0.0919087082,-0.092758745,0.3149051964,0.1702481657,-0.0907343924,-0.2195444554,0.4995703399,0.0184987243,0.1332165599,0.2572388351,-0.1153756678,0.3184193075,-0.0351820402,-0.3102843761,-0.0585473217,0.25789994,0.1867654771,0.0894528627,0.1449461579,-0.0057679066,0.0393899903,0.2028030753,-0.1860540211,-0.1476627141,0.0684218854,-0.0514374077,0.2672276497,0.0921070576,0.2305079699,-0.2248898447,0.2542558312,-0.1293564737,-0.1186855063,-0.2937058508,0.2206137776,0.3259740174,-0.2178397179,0.0190115552,-0.3540261388,-0.5721030235,0.2373391539,-0.102088578,0.2034955174,0.044429522,0.1631038189,0.0528447516,0.2842854261,0.1165380552,0.1276731193,0.1860585362,0.2293321192,-0.4912076294,-0.1064471304,-0.1062731072,-0.0397748537,-0.0589306988,0.2472022921,0.3200701773,0.3800521493,0.3068467975,-0.0916228443,-0.1028950959,-0.5269814134,-0.1500266492,-0.0121124117,0.130134657,0.1340956092,0.1357788891,-0.3142942786,-0.0991318077,-0.0189685598,0.009551581,-0.1610018611,0.0767915323,-0.1984455436,0.0737340525,0.1075765938,0.1551740468,-0.2808564305,-0.0295734666,-0.2238487601,-0.078518793,0.2837266922,-0.3108154237,-0.121591568,-0.2424649149,-0.2767058015,-0.0567089058,0.0251376703,-0.1826880574,0.298006922,0.2971807122,-0.0100440625,-0.1379388869,-0.0770876557,-0.3311392665,0.5520771146,-0.2370014042,0.127017498,-0.4133720398,-0.2401777804,0.1909674108,-0.0332648642,0.2755605578,-0.1376259029,0.1482989043,0.0406105258,-0.3078706861,-0.1919053346,-0.1842205077,0.2584030926,0.2677793801,0.7330991626,0.0600228608,0.3119849563,0.1707664728,0.0942510441,0.0160442796,0.2694071531,-0.3109174073,-0.1672167331,-0.3166559339,-0.0431081429,0.2585339844,0.0540937372,0.0704235211,0.1722522229,-0.1491773129,-0.0471089222,-0.2261487246,-0.0861322954,-0.1424787343,0.1036641896,-0.1508199722,0.312290132,0.2363678664,-0.0298635736,-0.4631287158,-0.3722175956,-0.2966814339,0.0670058429,0.1720416993,-0.1870609522,-0.4909403026,-0.2102657259,-0.3990023732,0.3647806048,0.219676882,0.2691723108,0.0582103245,0.1220634803,0.1519637406,0.0027125247,0.8666927814,-0.19458206,0.0310044661,0.0749542043,-0.5635308623,-0.2839429379,0.1034801453,-0.0465713404,0.0737081245,0.4109230042,0.4042406678,-0.304920882,-0.3569970131,-0.0543464944,0.1248578131,0.0158592798,-0.3737433255,-0.1587765366,-0.1617502868,-0.3721355796,0.3056936562,-0.0357633941,0.2938616872,0.3135254383,-0.4067285061,0.0206651017,-0.2386655957,0.3298971355,-0.015798999,-0.0448291413,0.1455546319,0.2480178326,-0.0656703785,0.1364159882,-0.0136284996,-0.2521076798,-0.0221387241,0.0333807729,0.0766707435,0.0559487566,0.3974484205,0.4912239313,-0.1657132655,0.5896593332,0.0532020405,0.5023969412,0.0112070506,0.211962536,-0.0865836293,0.2070833147,-0.6752635241,-0.6326030493,0.0606436431,0.3875877261,-0.3515584469,0.3487812579,0.0265598968,-0.0807151496,0.4408506155,0.0658437684,0.9560610652,-0.4236285686,-0.2176095098,-0.0705679655,0.2506343424,0.4501132667,-0.1479608268,0.1193736419,0.0631927624,-0.3605845869,0.0251184944,-0.159092173,0.0233628973,0.1161083505,0.0641688183,0.3675503433,-0.100732699,0.0420409553,0.0238025263,0.3231610954,0.292742908,-0.3619939387,-0.2005821019,-0.0633553565,0.0982885435,-0.0486692525,0.0233311523,0.027804723,0.2614167035,-0.2168837786,-0.2026834488,-0.4592892528,-0.3521227837,-0.0459212139,-0.1731692255,-0.1588933319,0.2026453018,-0.1423448771,-0.1666289419,0.4743825197,0.0542227663,0.1998710632,0.2317743003,0.2300251126,-0.0567542352,-0.4213398099,0.1946234256,-0.17065005,-0.2002880871,0.036260251,-0.0150356553,-0.4477983713,-0.0481721498,0.1366474479,0.4839239717,-0.0005009735,-0.2309638709,0.0015668145,-0.0064275311,-0.17772156,-0.0360894687,0.1439639181,0.070677571,0.6924601197,-0.0410527997,-0.2799558938,-0.1369671673,0.1190591902,0.3040412664,-0.2659932971,0.3038587272,0.2087016702,-0.4074697793,-0.1131529436,0.0062353169,0.1431041807,0.3565794826,0.0875349268,-0.0030965407,-0.2762711644,0.4604900479,0.0453895442,-0.1534852684,0.0090200212,0.0092839077,-0.6108727455,-0.2154678702,0.0144704282,-0.1084281579,-0.2433734536,0.0940173939,0.2439099699,-0.2326947302,0.3987815082,-0.1554259956,-0.1337856501,-0.0761049613,0.1933466941,-0.1788554788,0.1816151738,-0.0979172215,0.0836511925,-0.0088605443,0.094000116,0.0140442839,-0.1186867878,0.1272984445,0.1898169369,-0.1683938801,-0.2326007783,0.2912493646,-0.2927501202,-0.2596657574,-0.173383981,0.1088823378,-0.0262057781,0.0458075404,-0.6508744359,0.0548537262,-0.4824454486,-0.1160866097,0.1010411307,-0.1323594898,-0.3136637807,0.0636213124,0.1940173805,0.159207955,-0.0374435671,0.0203365181,-0.0668501109,0.1291176528,-0.0841172561,0.2444594502,0.044735495,-0.0845451429,0.2222917676,0.1211464107,0.1864476204,0.0525336117,-0.144560799,0.461396873,-0.0729820877,-0.2859913707,-0.1894619316,0.2004592419,0.1980032176,0.161650002,0.1860414147,-0.2322217226,0.3707811236,-0.1899377704,0.0946531966,0.4308142662,-0.059957888,-0.2244689018,0.3165216148,0.4238734245,-0.2095214576,0.4479428828,0.3593586087,-0.2119558454,0.5873918533,0.2064040154,0.3306261897,0.3587342203,0.5392187238,-0.2690783739,-0.3759899735,-0.1000151709,0.7628919482,-0.2536074519,-0.2111105025,0.0407190323,-0.1972560734,0.3258876503,-0.3352958858,0.1111131608,0.4303644001,0.1520309597,-0.0608436204,-0.3509233594,0.0244581457,0.2392033041,-0.0287273824,0.0072551654,0.01542194,0.1657253951,0.1774057001,0.2984442711,-0.2788741887,-0.2288468927,-0.140765354,0.1943701357,-0.0613881238,0.1345091313,-0.2762652338,0.2588126957,0.3280677497,-0.3723088503,0.2147519886,0.4029502869,-0.0219061971,0.1061898619,-0.3592620492,-0.1484079361,0.0223414879,0.4353690743,0.2051828504,0.2221914381,-0.0444084033,-0.0738318637,-0.0716140643,-0.0723766088,0.2281835526,0.0683034509,0.156188339,0.270205766,0.2790248692,0.0632243827,-0.1275682598,0.0609319359,-0.1570318341,0.396886915,0.1848491877,-0.2483234853,0.2444537431,0.2394408435,0.0090627391,-0.2402894944,0.004651871,-0.0892789364,0.4429185688,-0.2784524262,-0.3015824258,-0.3396236002,-0.3340788186,-0.3357553482,-0.0350053161,0.1032398343,0.0005810225,0.2125556469,0.019730337,0.0188017115,0.1103447452,-0.2328928113,0.0520926639,-0.1187780946,0.3637828827,-0.0315222591,0.3439206183,-0.0768297836,-0.3153738976,0.4809239805,-0.2549056411,-0.0406331718,0.0488817655,0.0844896734,0.1436852366,0.050615076,-0.1031228378,0.0445615426,0.4123341739,0.1149136126,0.5713286996,0.0512083694,0.1683560163,-0.2397980839,0.1752071083,0.322236985,0.1194870844,-0.0151624866,0.0767793655,0.0808279142,0.2362082005,-0.1285609007,-0.2848425508,-0.1072772443,0.0826447085,0.1503406316,0.2956114113,-0.2677984834,0.1417738348,-0.0339076109,0.0148020284,-0.232893914,-0.1246406436,0.5511728525,-0.5317599773,-0.2671302259,-0.3625561893,0.2553189099,0.2855137885,0.2134870142,0.1370442659,-0.0518262722,0.145610258,-0.1377858818,-0.3060030639,0.3175981343,0.0139027359,-0.0001584848,-0.2049388885,-0.2677419484,0.0847920254,-0.1360785961,-0.3229294121,-0.1577924043]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1796","title":"Filter on dataset too much slowww","comments":"Hi @lhoestq @ayubSubhaniya,\r\n\r\nIf there's no progress on this one, can I try working on it?\r\n\r\nThanks,\r\nGunjan","body":"I have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n","comment_length":18,"text":"Filter on dataset too much slowww\nI have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n\nHi @lhoestq @ayubSubhaniya,\r\n\r\nIf there's no progress on this one, can I try working on it?\r\n\r\nThanks,\r\nGunjan","embeddings":[-0.2252546996,-0.2595492601,-0.1091805696,-0.2148160636,-0.1230366156,-0.269241631,0.2215615809,0.2114835232,0.2634298503,-0.123314254,0.0279618911,0.090880774,-0.116068624,0.4082371593,-0.2278894037,0.1516521722,0.1011282429,0.213109985,0.4031715691,0.0390016884,-0.0577567853,-0.0807915404,-0.3609131277,-0.0298266672,-0.084706299,-0.2762124538,0.2782402933,-0.2476797551,-0.2799648941,-0.1902088523,-0.117655322,0.4328626096,0.0416844748,0.5227159262,-0.000128576,-0.384228617,0.2686328292,0.1016895473,0.2251769751,0.1034961194,-0.1696645468,-0.0080827391,-0.0367199033,-0.0586646497,-0.2172694355,0.0669343695,-0.2896471322,-0.2367753685,0.4300411642,0.1496510208,-0.0269949771,-0.2186436504,-0.3358837664,0.3563361466,0.1525760591,0.3359158635,0.1869012415,-0.2261263877,0.7080863714,-0.3562733829,0.0696439669,0.472147882,-0.4560546279,-0.1615430415,0.1487727165,-0.3453471661,0.1096845344,-0.3132957518,0.2409458607,0.176920563,0.0656614229,-0.0571096055,-0.2008105218,-0.2538344562,-0.045438882,-0.4103205204,0.0202633683,0.0456011929,-0.3242428005,-0.0042452426,-0.5788224936,-0.2189129293,0.0876276642,0.158389762,-0.3853456378,0.3735460639,0.0321886279,0.1135237738,0.1993607432,-0.215529263,0.0075017055,-0.2545441091,0.2832427919,0.3554543555,-0.674097538,-0.1508339942,-0.1581910998,0.2203153521,0.0110632014,0.0257151257,-0.3736515939,0.46273458,0.2508269846,-0.040211305,-0.0583137199,0.2461997569,0.0641968772,0.4818011224,0.2978991568,-0.219860673,-0.2987447381,0.2547263801,-0.0507870615,0.0679056197,0.0823750198,-0.0540153347,-0.5015363097,-0.170279786,-0.2727939785,-0.10285642,-0.4168030322,-0.120347321,0.668340981,0.4683458507,-0.1007532477,-0.1219638139,-0.1264315546,-0.2415731996,-0.1646133959,0.0298383143,0.0279094558,-0.0445403717,-0.0681126043,0.1991347671,0.0899436623,-0.2929748595,-0.1792654544,0.0473016202,-0.1829548031,0.1261359155,0.2369165123,-0.3831785619,0.13238132,0.292304039,-0.2854670584,0.3071512878,0.0823027566,0.1024769023,-0.1182383671,0.1186317205,-0.3777612448,-0.3367712498,0.2411236316,0.0114388652,-0.1120516658,0.1816656739,-0.0739685744,0.3442537487,0.306984067,-0.0880860984,-0.2214004248,-0.1915690303,-0.5816186666,-0.158752054,0.2008324116,0.0921893939,-0.4657712579,0.0186849385,-0.2798204422,0.210533604,0.5880839229,0.4387430549,-0.03694015,0.224987179,-0.0430899151,0.0483519398,0.2181489319,-0.112259388,-0.3453876376,0.623346746,-0.039500691,0.0077329222,0.1885050386,0.0785486177,0.5663010478,-0.2632850707,0.5140783787,0.3756265938,-0.2153287977,0.1178362891,-0.0218422543,-0.10364829,0.1589710116,0.2371780723,-0.2816350162,-0.0733494461,-0.0703740865,-0.3758479059,0.1747685671,0.0260584168,0.030902924,0.1709758937,0.2361922264,0.062734805,0.0111601958,-0.1820666343,0.1184111238,-0.0608216114,-0.050430499,-0.1652128398,-0.1754475087,-0.1815262139,-0.2594470084,-0.009087028,0.0641421527,0.2332099229,-0.1163491011,-0.1710831076,0.1199571043,-0.2128815055,0.0389979929,0.487057656,-0.0661606491,-0.0465142801,0.1828293353,-0.0481044054,0.3280166984,-0.1143477932,0.1694704741,0.0258435123,0.3714915812,0.0142312702,0.0571064577,0.1386633962,0.5027495027,-0.2991530001,-0.3137467802,0.1767044514,0.1542363316,0.3389374912,0.1094481423,0.016436182,0.1302595288,-0.1143233329,-0.1493640542,0.5312942863,-0.0404825136,0.4238991737,0.0877424106,-0.0993584171,0.3138713539,0.1738130003,-0.0924373791,-0.2166422606,0.5034923553,0.0138215674,0.1286027282,0.2449540347,-0.1035164595,0.3022047579,-0.04750029,-0.3135300577,-0.059206225,0.2680288255,0.1906663477,0.0849729031,0.1419332027,-0.0174565259,0.0390058644,0.1978618503,-0.1883351803,-0.1547813118,0.0823922828,-0.0621349774,0.258713901,0.1034958884,0.2353141904,-0.2182342857,0.2533356845,-0.1286801845,-0.0980202109,-0.2866779268,0.2284122109,0.3222565651,-0.2280865908,0.0149425017,-0.3553006351,-0.5684863329,0.2344612628,-0.1010885537,0.2024694681,0.0458728783,0.1592308432,0.0498124883,0.28446123,0.1175139099,0.1196427196,0.2055429518,0.2214313596,-0.4647490382,-0.1081777811,-0.1323721856,-0.041707769,-0.0571200885,0.2584053576,0.3142126203,0.3845154941,0.306491226,-0.0974844322,-0.1131294072,-0.5374432802,-0.1481490433,0.0105012637,0.1324184984,0.1226705164,0.1452452987,-0.314394474,-0.1035704985,-0.0251875464,0.0196898859,-0.1629319787,0.0662417337,-0.1998528242,0.0767853707,0.1098999381,0.1567242444,-0.2886810005,-0.0323659219,-0.2201781124,-0.0711628944,0.2759138644,-0.3124832213,-0.117224656,-0.2404733896,-0.2843366861,-0.063058123,0.0203151796,-0.1998870969,0.2910321951,0.2740531266,-0.01750651,-0.1282386631,-0.0829764009,-0.3217385113,0.5405508876,-0.2413505465,0.1424674392,-0.4174325466,-0.2339668721,0.1934642047,-0.0433743484,0.2607778609,-0.1583853513,0.1518045962,0.0461019576,-0.3172047436,-0.186086297,-0.1872590035,0.2764711976,0.2739782035,0.7350676656,0.0479093269,0.3134377599,0.1614057422,0.0978119075,0.0234633312,0.2742680311,-0.2951162159,-0.1754238904,-0.3233992755,-0.0374852084,0.2471006811,0.050765384,0.077100873,0.1735542119,-0.1551261097,-0.0521914028,-0.2132920772,-0.0872249156,-0.1421200633,0.1017120406,-0.1648368239,0.306812197,0.2345013916,-0.039152693,-0.458776921,-0.3636745214,-0.3014778197,0.0613552816,0.1693892479,-0.1827860624,-0.5065771937,-0.2059262842,-0.3918886781,0.3594691753,0.2159307003,0.279833436,0.0462322123,0.1300626397,0.1529965401,-0.0007642696,0.8587453961,-0.1964030713,0.0236833524,0.0773434713,-0.5784474015,-0.2750440836,0.1098078266,-0.0400722995,0.0650057942,0.4197523296,0.3936816752,-0.3051287234,-0.3487202525,-0.0523087531,0.1268809885,0.0077716731,-0.3749650717,-0.1559260041,-0.1641940624,-0.3765984476,0.3039278686,-0.0235089045,0.3030177951,0.3059747517,-0.4037964046,0.0278203525,-0.2339780331,0.3217283189,-0.0272874441,-0.0450220518,0.1439339221,0.2509563863,-0.0557608716,0.1492969096,-0.0016955561,-0.2590816915,-0.0302044824,0.0440486856,0.0972662866,0.0444831736,0.3970540762,0.4827743769,-0.1612476259,0.5882353783,0.0505079404,0.4886027277,0.011629683,0.2173449993,-0.0917056054,0.2079638541,-0.6904321313,-0.6199195385,0.0585212782,0.3905729055,-0.346159339,0.3499727249,0.0200455803,-0.0846653879,0.4471519291,0.0663013384,0.9581962824,-0.4153416753,-0.2108306587,-0.0812214315,0.2532030642,0.4408656657,-0.1588865221,0.1169549003,0.0642557144,-0.3563610613,0.0335152484,-0.1627473682,0.0319045931,0.1154624,0.0592168868,0.3585591614,-0.099434875,0.0464691371,0.0220796969,0.332105875,0.303319931,-0.3577495813,-0.1986604184,-0.0548402965,0.1066495702,-0.044585634,0.0151255345,0.0325653479,0.2558647096,-0.2082432359,-0.1969842017,-0.4552558959,-0.34328264,-0.0408173613,-0.1839427501,-0.17076689,0.2155314833,-0.145385623,-0.1489435285,0.4777180254,0.0395529382,0.2064716369,0.2317768931,0.2153300643,-0.0568481311,-0.4187892973,0.1855115891,-0.1786766052,-0.1958904117,0.0447833873,-0.0129556833,-0.4329488575,-0.0632161945,0.140157789,0.4846388996,-0.0020972302,-0.2411370873,0.0008224576,-0.0012010918,-0.170816794,-0.0320651941,0.147512868,0.0614322945,0.7015863657,-0.0356989689,-0.2789462209,-0.1368632764,0.1221278757,0.2942487001,-0.2797678709,0.3097616732,0.2331241071,-0.413025111,-0.1083008572,0.0172183178,0.1657754034,0.3465482891,0.0790231824,0.0083406046,-0.2764845192,0.4665234685,0.0402587838,-0.1502177715,-0.006530548,0.0210220087,-0.608069241,-0.2222399414,0.0133887064,-0.1192979217,-0.2351454049,0.0938155279,0.2466761172,-0.2269974798,0.4047358334,-0.1552992314,-0.1191569269,-0.0896128193,0.2054951489,-0.1751523018,0.1732694209,-0.1048567444,0.0844784975,-0.0092603043,0.0975957289,0.0234885365,-0.1187349334,0.1299327612,0.1904266775,-0.1591579318,-0.2345916629,0.2908121943,-0.2887116969,-0.256633848,-0.1597928256,0.0984030589,-0.0179285165,0.0577174798,-0.6547785401,0.0694828033,-0.4861409664,-0.1320483387,0.113509208,-0.1249681637,-0.3016546965,0.0651138499,0.1955117434,0.1558886021,-0.0396988355,0.0083106728,-0.059449736,0.1363196075,-0.0981676877,0.2446375191,0.0385851189,-0.0546671562,0.2382380366,0.120196417,0.1842215359,0.0489126854,-0.1412813067,0.4671067894,-0.0722419992,-0.2741684616,-0.1641023606,0.1977159828,0.1868246794,0.1511606425,0.179021135,-0.2410615534,0.3629327416,-0.1833376437,0.0998070911,0.4350858629,-0.0554787032,-0.2217369378,0.3130109608,0.4332221746,-0.2051404119,0.4461446106,0.3638696671,-0.2131380588,0.5768561363,0.186554417,0.3318859637,0.3495144546,0.5341730714,-0.2553034127,-0.3887854815,-0.1109636799,0.7703387141,-0.2449410409,-0.2136731595,0.0333937705,-0.2079736292,0.3349086344,-0.3433472216,0.1132228523,0.4351329505,0.1481791437,-0.0408240445,-0.3570049405,0.0204015728,0.248273164,-0.036998298,0.0036891499,0.018553948,0.1613125503,0.1794333309,0.2775251269,-0.2858623862,-0.2290665209,-0.1520702392,0.1804335862,-0.0591078624,0.15228571,-0.2897800207,0.2591445446,0.3215397,-0.3563667536,0.2093613148,0.415440172,-0.0153534878,0.1073969826,-0.3542917669,-0.155683741,0.0286301486,0.4394167364,0.2068323493,0.2189008743,-0.0373840034,-0.0680969879,-0.0727639496,-0.0777654797,0.2325563133,0.0656391233,0.157876119,0.2685889006,0.2677966058,0.052647952,-0.1324347109,0.0658370033,-0.1772332191,0.3990635276,0.1876475811,-0.2525111735,0.2420499623,0.2378875166,0.0081937537,-0.2458667904,0.012394188,-0.0878853276,0.4355199635,-0.2969923317,-0.3003424704,-0.348913908,-0.3365279138,-0.3431505263,-0.0251454189,0.0969923437,0.0010713384,0.2146293819,0.0203985926,0.008032836,0.1167682111,-0.2403487712,0.0444591418,-0.1369325817,0.3749761879,-0.0254141651,0.3593270481,-0.0807765126,-0.3088649213,0.4820462465,-0.2655583918,-0.0422151722,0.0557011291,0.0875636116,0.1483191401,0.0405413993,-0.108242929,0.0369232856,0.4210778177,0.1204250902,0.57259202,0.0505463928,0.1496593058,-0.226986751,0.1850090176,0.323787719,0.1223639175,-0.0324243158,0.0748083666,0.0803684592,0.2249293625,-0.1496294588,-0.2773569226,-0.0969022959,0.0771478489,0.1535182446,0.2936276793,-0.2447824478,0.1370540112,-0.0368246585,0.0204693507,-0.2410460413,-0.1227077544,0.5522251129,-0.5193554163,-0.2587075233,-0.3608834445,0.2557354569,0.282985419,0.2068324834,0.1441439092,-0.0422927365,0.1479316056,-0.1221395358,-0.2939148247,0.3119493425,0.0089959074,0.0091708573,-0.2084155232,-0.2674075663,0.095736824,-0.1376583129,-0.3251849711,-0.1448193491]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1796","title":"Filter on dataset too much slowww","comments":"Sure @gchhablani feel free to start working on it, this would be very appreciated :)\r\nThis feature is would be really awesome, especially since arrow allows to mask really quickly and without having to rewrite the dataset on disk","body":"I have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n","comment_length":39,"text":"Filter on dataset too much slowww\nI have a dataset with 50M rows.\r\nFor pre-processing, I need to tokenize this and filter rows with the large sequence.\r\n\r\nMy tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.\r\n\r\nWhen I applied the `filter()` function it is taking too much time. I need to filter sequences based on a boolean column.\r\nBelow are the variants I tried.\r\n1. filter() with batch size 1024, single process (takes roughly 3 hr)\r\n2. filter() with batch size 1024, 96 processes (takes 5-6 hrs \u00af\\\\\\_(\u30c4)\\_\/\u00af)\r\n3. filter() with loading all data in memory, only a single boolean column (never ends).\r\n\r\nCan someone please help?\r\n\r\nBelow is a sample code for small dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.map(lambda x: {'flag': random.randint(0,1)==1})\r\n\r\ndef _amplify(data):\r\n return data\r\n\r\ndataset = dataset.filter(_amplify, batch_size=1024, keep_in_memory=False, input_columns=['flag'])\r\n```\r\n\nSure @gchhablani feel free to start working on it, this would be very appreciated :)\r\nThis feature is would be really awesome, especially since arrow allows to mask really quickly and without having to rewrite the dataset on disk","embeddings":[-0.2547401488,-0.2351185828,-0.1014519706,-0.2257349342,-0.1428643167,-0.2871686816,0.2794651985,0.2359782606,0.31276685,-0.1190766096,0.0621099062,0.130415678,-0.1263469011,0.4224111736,-0.2473777384,0.1691477746,0.0948600248,0.2099197805,0.3651643395,0.0446351208,-0.0494469479,-0.1138269603,-0.3327174187,-0.0442434326,-0.0404985696,-0.2961439788,0.293048203,-0.266183883,-0.2599883974,-0.1789186001,-0.0893380716,0.4252875447,0.0437732004,0.4724216461,-0.0001284532,-0.3640066385,0.2625659704,0.1131875664,0.1909705698,0.1135952622,-0.1848612279,-0.0163564384,-0.0218522791,-0.0593499728,-0.2138731629,0.0530097894,-0.2731750607,-0.3023944199,0.4287187159,0.1006162092,-0.0289606843,-0.1832480431,-0.3738381565,0.3723735809,0.1458556354,0.3561851382,0.1437241137,-0.2502307594,0.6577313542,-0.373789221,0.0461209491,0.5352955461,-0.4705422223,-0.1681782454,0.1968522519,-0.3184746802,0.1125075445,-0.2822371125,0.2291529626,0.1690669507,0.0149930781,-0.077709727,-0.214353025,-0.2915641069,-0.061559841,-0.4139195085,0.0252865311,0.0286112968,-0.3154198229,0.0164226741,-0.5580670238,-0.2037364841,0.0965511426,0.140373975,-0.3547916114,0.3611058593,0.0346127339,0.1313696802,0.2142255008,-0.2130946219,0.0226945058,-0.2870528996,0.2648449838,0.3623133004,-0.6492136121,-0.1423451453,-0.1743784249,0.175046578,0.0069197412,0.078633137,-0.3529721797,0.468601197,0.2220184952,-0.0259510782,-0.0626640618,0.2397354394,0.0740979016,0.4572773874,0.287361294,-0.2147802562,-0.2366798669,0.2544233799,-0.0684944168,0.0798996687,0.080058217,-0.0319134593,-0.493953228,-0.1388382912,-0.2648066878,-0.1102305353,-0.3951866329,-0.1192370504,0.6373211145,0.425801307,-0.0795858949,-0.1344533414,-0.1196869686,-0.238056615,-0.1651940793,0.0285185836,0.0324384682,-0.0963223055,-0.0695135668,0.1784169823,0.0781029537,-0.2720378637,-0.19776164,0.032622695,-0.177771166,0.173016876,0.255194664,-0.3473986089,0.1557765603,0.3013786674,-0.3203011453,0.3183274567,0.1355057508,0.0904031172,-0.1295000464,0.1477259398,-0.3459281623,-0.3141595125,0.213075757,-0.0019208555,-0.1204015985,0.1689202189,-0.1210167781,0.3417942226,0.2751415074,-0.1006942317,-0.2016447186,-0.1915017962,-0.5472353101,-0.1620805562,0.20939821,0.1171443835,-0.4860534966,-0.0119034592,-0.2796179056,0.2233322412,0.6011432409,0.4344096482,-0.0460530594,0.1472395062,-0.029907085,0.0564316921,0.2358345389,-0.1227665097,-0.3312916458,0.6238361597,-0.0638757572,0.0400916561,0.2126489878,0.105339013,0.5331244469,-0.2556838095,0.4627811611,0.3944644928,-0.1878147721,0.1091042608,0.0033416434,-0.1312590688,0.1949516237,0.2179062665,-0.3320954442,-0.0757945925,-0.06967213,-0.44811064,0.1965843141,0.0232221596,0.0347890407,0.1189482808,0.2403095067,0.0163257774,0.0206499789,-0.1357768178,0.088432923,-0.051119294,-0.0140721612,-0.1671031862,-0.2494914681,-0.1959039867,-0.2026102096,-0.0224119183,0.0633288771,0.2025333345,-0.1286606789,-0.1770589203,0.0898274183,-0.2292308807,0.0088760043,0.4807582796,-0.0409268737,-0.0277851075,0.1705214977,-0.0746284947,0.3705918193,-0.0782507658,0.1451504827,0.016999267,0.3502371907,0.018661961,0.0410104617,0.1407466233,0.4962439537,-0.289190799,-0.2919785082,0.1777757704,0.145196721,0.3355655074,0.1232620925,0.0453377999,0.1189533919,-0.1018491685,-0.2242897451,0.5475719571,-0.0494961962,0.4638311863,0.0708090588,-0.0753633305,0.3280289769,0.1719776988,-0.1158203855,-0.2379233837,0.460211277,-0.0143444501,0.1340287924,0.2429938912,-0.0889173448,0.3365106881,-0.0756700039,-0.2729927003,-0.0627097338,0.2376005501,0.1726811677,0.1039543226,0.1899481267,0.0109502422,0.0056815515,0.2170878053,-0.1980079561,-0.1666433364,0.0511952043,-0.0437555015,0.2654491365,0.0814944804,0.2596529424,-0.2470981181,0.257267952,-0.1374319047,-0.117131114,-0.3422016203,0.201179564,0.2727069855,-0.2090350688,0.0312487297,-0.3480932415,-0.5751774311,0.254105866,-0.0879080147,0.2226923853,0.0729397461,0.1610177904,0.0471449569,0.245237574,0.1408628225,0.0855317712,0.2216345668,0.2345061153,-0.5197231174,-0.1105312631,-0.1512378454,-0.0394730493,-0.0571239963,0.2479060888,0.3156698942,0.3780960441,0.341663897,-0.0667107627,-0.1172677502,-0.5213433504,-0.1506472975,-0.0124812704,0.1178539246,0.1023150012,0.1462201476,-0.3354575336,-0.1229352877,-0.0298256818,0.0491522551,-0.1772672236,0.072665222,-0.1669165492,0.0829740614,0.0800649673,0.1700888574,-0.2357677966,-0.0381800681,-0.2073810995,-0.0565625541,0.2702920735,-0.2656601965,-0.0730579272,-0.278113842,-0.2712409794,-0.1031879261,0.0364703238,-0.2044027001,0.2926090062,0.2826998532,0.0196530279,-0.1517456025,-0.0983475074,-0.3492052555,0.5780216455,-0.2696869671,0.1584291905,-0.4038267732,-0.2313586622,0.2019394934,-0.0334451422,0.2865878642,-0.1035000458,0.1401912421,0.0275014229,-0.3251667321,-0.194995448,-0.2094114274,0.2677611411,0.2775139809,0.7451361418,0.0741668865,0.3454885781,0.1802725196,0.1070337966,0.0168004595,0.2816306055,-0.2569590807,-0.1691365987,-0.3194511533,-0.0622099191,0.2351968884,0.0858149529,0.0257593412,0.1553872973,-0.177125603,-0.0389002822,-0.2113517225,-0.0666952059,-0.159169808,0.1071524248,-0.203222841,0.2908296883,0.2369368374,-0.0089528793,-0.4920891821,-0.352634877,-0.2461751699,0.0338327512,0.2247825414,-0.1775359213,-0.5154926777,-0.1536707282,-0.3968797028,0.3571224511,0.210564062,0.2496906817,0.0733297765,0.0918404981,0.1379288137,0.0048201163,0.8831474185,-0.1925903857,0.047824014,0.0855068862,-0.5386545658,-0.3299160898,0.1213033423,-0.0266265254,0.062753357,0.3690216243,0.40054515,-0.3076560795,-0.336124748,-0.059670344,0.1310092509,-0.0072173565,-0.3539107144,-0.1504345834,-0.1287596077,-0.3918291926,0.3001624346,-0.0020009412,0.2927742302,0.3270325661,-0.3938694,0.0023757105,-0.2537275255,0.310267508,-0.0174858812,-0.0331122279,0.1579238176,0.2585356236,0.009452288,0.1730175614,-0.0365031622,-0.2073551863,-0.0223435946,-0.001539191,0.0565205142,0.048267331,0.3998790681,0.4951084554,-0.1465217769,0.5836417079,0.0694032833,0.4908259511,-0.0113032712,0.227251187,-0.1186985001,0.1846880466,-0.6643241644,-0.6317513585,0.0898283496,0.3616834581,-0.3592296243,0.4149486125,0.053905122,-0.0925548673,0.484637022,0.0796499103,0.9722087383,-0.4101478457,-0.1564437747,-0.1024759188,0.2348394096,0.4491517246,-0.1452313513,0.1185023487,0.0526879989,-0.3127822876,0.0187684689,-0.1789846122,0.0131357061,0.1035683602,0.1015665308,0.3528617322,-0.1354784518,0.0532929227,0.0173897129,0.3033020794,0.2745910287,-0.3907105625,-0.1281688958,-0.0674056113,0.1484354585,-0.0544957779,0.0417410508,0.0231089871,0.2586682737,-0.2181496322,-0.1786284149,-0.4309237003,-0.3346201181,-0.029031707,-0.1861093938,-0.2180696726,0.185608387,-0.1499716938,-0.1546514481,0.4800975323,0.0550310984,0.187208131,0.2562003434,0.2629372478,-0.0710088462,-0.4394882023,0.1941317022,-0.1478517652,-0.1893982142,0.0399271995,-0.0202837195,-0.4675250053,-0.0285941549,0.166480124,0.476583004,0.016073307,-0.2211516201,-0.0090734558,0.0133570014,-0.1800496727,-0.0390072986,0.1601356417,0.0872107595,0.6803513169,-0.0118059255,-0.271247685,-0.1171461344,0.1355697066,0.2716130316,-0.254437834,0.3072373569,0.2026948035,-0.4166202545,-0.1117487848,0.0054928735,0.2209941596,0.3390146792,0.1124830619,-0.017363416,-0.2973784506,0.4311403632,0.0399324,-0.1429939121,0.0226015579,-0.0035556429,-0.6040536761,-0.1978469491,0.044560086,-0.1624215543,-0.2262354046,0.0922241807,0.2347444147,-0.2562168241,0.370243758,-0.1560823321,-0.1145806164,-0.0422539301,0.2249818444,-0.1870007366,0.2397274673,-0.0696390346,0.0781731009,-0.0215617903,0.0846773461,0.0055006947,-0.1249881387,0.0938567519,0.1914307028,-0.169272691,-0.2661699951,0.2854486704,-0.3040853441,-0.2480561882,-0.2004140019,0.1181157231,-0.0329118706,0.0581699833,-0.6831446886,0.0294047892,-0.4843916893,-0.146528095,0.0845035985,-0.1279477328,-0.3025523424,0.0446089022,0.1762179732,0.1422257274,-0.027261626,0.0484138951,-0.0653512478,0.1146876216,-0.0626347959,0.247288391,0.0431434289,-0.0827456564,0.2133442909,0.1262270361,0.1769425571,0.0563259348,-0.1567684561,0.4866526425,-0.0668555796,-0.286857307,-0.1953933984,0.1671299785,0.2077479511,0.1736731082,0.1619257629,-0.2451898903,0.3579375446,-0.1853931099,0.0889693052,0.4020831585,-0.0821812078,-0.2352085859,0.338804245,0.4077502489,-0.2249969691,0.4805638492,0.3468933105,-0.1938000917,0.6008088589,0.1963528991,0.3269647956,0.4239937961,0.5279600024,-0.3396155834,-0.3884891272,-0.1107790843,0.7172064185,-0.2031840235,-0.1873832047,0.0708058849,-0.2135522664,0.3435969651,-0.3242046237,0.1368947923,0.399792552,0.168961376,-0.0689866617,-0.3674280643,0.0155459922,0.2693645656,-0.0258648619,0.0015569801,0.0039907065,0.2068347633,0.1559245884,0.3189149797,-0.2765103579,-0.2405990511,-0.1286069453,0.2134978324,-0.0510533191,0.1259865463,-0.2663528323,0.2794303894,0.3113368154,-0.3693409264,0.2065236419,0.369775027,-0.0400802977,0.0977606848,-0.3742189705,-0.1569204628,0.0034288182,0.4361085892,0.2118588835,0.1915192604,-0.0341015868,-0.073379688,-0.0547663756,-0.0629169941,0.2353907079,0.0497872643,0.161933437,0.2595260143,0.2800493538,0.0348718017,-0.1228137761,0.0269114859,-0.1231809035,0.3983190358,0.2354516536,-0.2709564865,0.2363279164,0.2429218292,0.0075737471,-0.2389284074,-0.0133874547,-0.0855355188,0.4185398519,-0.2886800468,-0.2669920623,-0.3452297747,-0.3425404727,-0.3784074485,-0.0410966389,0.0931633189,0.0125888037,0.2585785389,0.0449906029,0.0296727791,0.1020231992,-0.1909074038,0.0283049326,-0.1394940168,0.3953852654,0.009559935,0.3836015761,-0.0542448536,-0.3450880647,0.4803273082,-0.2367374152,-0.0424051769,0.0815749168,0.0867153406,0.1680946052,0.0282796379,-0.0564314313,0.0356681123,0.3896331191,0.1423925906,0.5392435789,0.0255773012,0.1664277315,-0.2396234721,0.1745081544,0.3232299984,0.1623243243,-0.0563945174,0.0658027753,0.065319337,0.2356828898,-0.1420142353,-0.2512159348,-0.1008307263,0.1202272624,0.1188263446,0.2824155986,-0.3085885644,0.1336143166,-0.0352308489,0.0329057164,-0.226747632,-0.1435497254,0.5653973818,-0.5540592074,-0.3062086999,-0.3145028353,0.2423718721,0.3347381055,0.1990547329,0.1113849729,-0.058614973,0.1164706796,-0.1394054294,-0.3094705045,0.2706581652,0.0268352889,0.0200900957,-0.2097317129,-0.2281597257,0.0892246813,-0.1314508319,-0.3454653919,-0.1742684096]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1790","title":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.","comments":"Hi !\r\n\r\nApache Beam is a framework used to define data transformation pipelines. These pipeline can then be run in many runtimes: DataFlow, Spark, Flink, etc. There also exist a local runner called the DirectRunner.\r\nWikipedia is a dataset that requires some parsing, so to allow the processing to be run on this kind of runtime we're using Apache Beam.\r\n\r\nAt Hugging Face we've already processed certain versions of wikipedia (the `20200501.en` one for example) so that users can directly download the processed version instead of using Apache Beam to process it.\r\nHowever for the japanese language we haven't processed it so you'll have to run the processing on your side.\r\nSo you do need Apache Beam to process `20200501.ja`.\r\n\r\nYou can install Apache Beam with\r\n```\r\npip install apache-beam\r\n```\r\n\r\nI think we can probably improve the error message to let users know of this subtlety.\r\nWhat #498 implied is that Apache Beam is not needed when you process a dataset that doesn't use Apache Beam.","body":"```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?","comment_length":167,"text":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.\n```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?\nHi !\r\n\r\nApache Beam is a framework used to define data transformation pipelines. These pipeline can then be run in many runtimes: DataFlow, Spark, Flink, etc. There also exist a local runner called the DirectRunner.\r\nWikipedia is a dataset that requires some parsing, so to allow the processing to be run on this kind of runtime we're using Apache Beam.\r\n\r\nAt Hugging Face we've already processed certain versions of wikipedia (the `20200501.en` one for example) so that users can directly download the processed version instead of using Apache Beam to process it.\r\nHowever for the japanese language we haven't processed it so you'll have to run the processing on your side.\r\nSo you do need Apache Beam to process `20200501.ja`.\r\n\r\nYou can install Apache Beam with\r\n```\r\npip install apache-beam\r\n```\r\n\r\nI think we can probably improve the error message to let users know of this subtlety.\r\nWhat #498 implied is that Apache Beam is not needed when you process a dataset that doesn't use Apache Beam.","embeddings":[0.1911676675,0.0017940417,-0.0041889772,0.1869834214,0.271382004,0.1396206915,0.4213475585,0.3950362802,0.2294802517,0.178105548,0.1076992676,-0.091416508,0.0533491559,-0.1092457175,0.1017360389,-0.4603553414,0.3123897612,0.0755011961,-0.1031829417,-0.1862583458,-0.3473697007,0.0844961405,-0.3394604623,0.2037167251,-0.2264511585,0.1729814857,-0.1223797128,0.0795244798,0.0796693712,-0.3102597296,0.0695770979,-0.2526939511,-0.0015343654,0.3789234757,-0.0001123201,0.1725937277,0.4942116439,-0.1868751496,-0.3528478742,-0.2001891136,-0.0440738685,-0.5268995762,0.0270454362,-0.499565959,-0.1185732931,-0.3200562298,0.1248418018,-0.3306967914,0.6608893275,0.2097997069,0.1821456999,0.3544296324,0.0479206964,-0.0867460966,0.4274948239,0.0936123654,-0.149034515,-0.0419644639,-0.0831065923,0.0119523332,-0.0941681638,0.5700821877,-0.3236299157,-0.0393849798,0.2292389423,-0.2227157354,0.0566582233,-0.6094756126,0.1776165217,0.3224822283,0.7865334153,-0.1351768821,-0.2221815437,-0.1530813575,-0.1098430529,0.2351433337,0.1595033854,0.1091495752,-0.2148615271,0.0463166982,0.4381077886,-0.270696491,-0.1476813704,0.0733524188,0.1302394271,0.4315976799,0.0706527233,0.3176282346,-0.0545347892,-0.0860568136,-0.1191400737,-0.3226896524,0.2525667548,0.3874098063,-0.1137113497,0.0848223418,0.1017933562,0.150898084,-0.0367009826,-0.2810484469,-0.1921423376,0.1158809811,0.1053410843,0.1082061157,0.1243435964,0.1274571866,0.3880292475,-0.0257628113,0.3348400295,-0.2162707746,0.1098147631,0.0442705415,0.0299781673,-0.2803065479,-0.3338729441,0.22260122,0.3565318882,-0.0905370265,-0.0576841198,0.0482546426,-0.4567850232,-0.3259344697,-0.2221905589,0.3378888369,0.1325860918,0.4152707756,0.1352861971,0.2345647663,-0.2575584352,-0.454652369,-0.0323016606,0.34243536,-0.4076470733,0.1976642311,-0.0461325906,0.014478053,0.219536826,-0.155263111,0.1326275021,0.1820785701,-0.0900048688,-0.0856828094,-0.0457458831,0.450314641,0.0657334551,0.4466489851,0.2975469828,-0.127250582,-0.2159790546,0.2281534076,-0.2719882131,-0.2765989006,0.1818429679,0.0594422966,-0.3565073907,0.2080850005,-0.0104246242,-0.0740670934,0.0480912887,-0.0854229033,0.2088577151,0.0514473468,-0.0277333166,-0.2898869216,-0.0214107037,0.7498963475,-0.2269245684,-0.0638563037,-0.0097224638,-0.143658042,-0.1310082823,-0.0040014111,-0.1556620598,0.2832137346,-0.2314025909,-0.0215374921,0.5308304429,-0.5257617235,0.0584362261,0.1944158971,-0.0982031152,0.2102937549,0.2612053752,0.0026731209,-0.4192942679,-0.2195367217,-0.2881307304,0.0506402589,0.1025504246,-0.0246719941,-0.256819725,-0.0081708189,0.1157977879,0.1537024528,0.1938264966,0.0787255466,0.1328656375,0.4823886752,0.2445164323,-0.1990591437,0.2263817489,0.4565427899,-0.0051692198,0.1091252044,0.2236133516,0.0644488558,-0.062983647,0.2775996923,-0.0855787322,0.4019979537,-0.0905072242,-0.1003317833,-0.147992745,-0.0885240436,-0.2833200097,-0.3008551598,0.0780689865,0.0352994539,-0.0723938271,0.3316417038,-0.284832269,0.4456302524,-0.1219724715,0.0501287058,-0.5256504416,-0.0278601795,-0.2387652099,0.1458378732,-0.0339003801,0.0709542856,0.2233174741,-0.1908288151,-0.0455725528,0.1484087855,0.1055572927,0.1109236106,-0.0744641945,-0.0624695234,0.1449841708,-0.3980558217,0.3778229952,0.135535866,0.1264114082,-0.0347073674,0.113877058,0.2636615336,0.1505611986,0.5216038227,0.0231506247,-0.0100320475,0.0495372601,0.0356119163,-0.343896091,-0.1951042712,0.2003543377,0.2368834317,0.2999696732,0.0473005436,0.117445074,-0.3627568185,0.5069276094,-0.0543003976,0.175747022,-0.0242701843,-0.3246910274,0.1475041211,0.1711483002,-0.2194782048,0.2410337776,0.1912721097,-0.1260139197,0.0883795917,0.2026082873,-0.1247962117,0.0700958893,0.224393934,0.4852194488,-0.0354694054,-0.2386367619,0.0484007522,-0.1771229506,-0.0702499673,-0.2258546054,-0.0329044536,-0.2475070804,0.1603958607,-0.1956047118,-0.6308538318,-0.3146678805,0.049406942,-0.5606020093,-0.284016192,-0.0563855954,-0.2035318166,-0.2767908573,0.2280882597,0.1178847477,-0.0494031869,0.0303965993,-0.4305200875,-0.3263076246,-0.3265705407,-0.3022225499,0.0115317823,0.2494443655,0.0808152035,0.1863838136,-0.400044173,0.0187860429,-0.2099525928,-0.2723637819,0.2021510154,-0.1723351181,-0.001047937,-0.0152867716,0.5179567933,-0.0564169735,-0.0476004295,0.4222886264,-0.1399167329,-0.1939951777,0.1210300326,0.1419619769,0.0817631334,0.0471344739,-0.5049287081,-0.1154193878,-0.4123275876,-0.0873092562,0.0832739323,0.1298006624,0.2418057621,0.1741642505,0.2212671041,0.0440381467,0.0194630306,0.0029390538,-0.0079011405,0.2857334018,-0.133402586,-0.3504652083,0.3051893711,-0.1546028107,0.1144625098,0.1189529523,-0.2039324045,-0.1464451849,0.2433487475,0.0775506347,-0.0539503545,0.3316069841,0.4205645323,-0.1318393499,0.00018201,-0.1190083399,-0.0543212742,-0.1399504989,-0.1769160479,0.325774461,0.2829321623,0.317104876,-0.0185836926,0.8535458446,-0.0275243409,0.4695922434,0.4642806649,0.1470199525,0.3714669645,-0.4112051725,-0.1172696128,-0.0916099772,0.0645114928,-0.1340083629,0.1620410383,-0.0790208057,0.0574375354,-0.6090327501,-0.3949504793,-0.3404390514,-0.2219667882,0.15752846,-0.1700629145,0.2305217832,0.026509935,0.0417210273,0.1240234748,0.1018324047,0.250807941,0.3875487745,0.1745309979,0.0880678073,-0.0849715099,-0.0875831619,-0.2883145511,0.3871926963,-0.0084190676,0.1123481989,-0.0652785376,-0.3105088472,0.0889986306,0.0730067492,0.4744766355,-0.2369825691,-0.2140560895,0.2444828004,-0.0033569643,-0.6545328498,-0.0518580638,0.0534754731,0.0339649469,-0.0903271288,0.1465218216,-0.2105579823,-0.3132273555,0.1991851777,0.4235745966,-0.1190467328,-0.2092808485,-0.587559402,-0.0709562749,-0.4787138104,-0.0949800313,-0.0751945227,-0.0542525426,0.071302183,0.3650435805,0.1260063797,-0.0556114614,0.0407625958,-0.028509805,0.0701580048,-0.1987766325,-0.3843260705,0.1629834622,0.2838365436,-0.1293620467,0.3129348159,-0.0545833483,-0.4171644151,-0.2030954957,-0.0646217987,0.1295168996,0.4219566286,0.0074734087,-0.1024843231,0.009375059,-0.2213627994,0.0771461725,-0.0407827012,0.1367144734,-0.0255442094,-0.4656984508,-0.4260053635,0.7275453806,0.0573015697,-0.1102519482,0.2195152044,0.3865062892,-0.0961203277,0.4131186903,0.1205051467,1.0968141556,-0.0403658897,0.2054995149,0.1531868875,0.1145942286,0.7433243394,-0.6203227043,0.1198608875,-0.1833266467,0.1576327235,-0.1322313547,-0.0103877289,0.6084547043,0.0741045773,0.0390936956,0.5708758235,0.163899973,0.2875731289,-0.0295884348,0.1656965166,-0.2634520233,-0.0626116395,-0.4114671052,0.0360437371,-0.1083326042,0.7435284257,-0.1365130991,-0.1755887717,-0.3005131781,-0.4897763729,-0.0995775238,-0.0115369791,-0.2393388301,0.1199973449,-0.2641693354,-0.3201241493,-0.0009482363,0.3039798141,0.3844959736,-0.0567821562,-0.2318536341,0.1234674826,-0.1061983854,-0.0880051479,-0.2654598355,-0.0403252952,0.4411506653,-0.1480442882,-0.1679426432,0.1714525372,-0.4244392216,-0.0879567936,-0.1084509119,0.1357586086,0.5168765783,0.2983564138,-0.1830006242,0.2779881358,0.0278041139,0.0789502114,0.122702606,-0.1928294897,-0.069757551,-0.1583988518,-0.2607504427,-0.1299844533,-0.0765771121,0.2389581949,0.2075750083,-0.1365613043,0.2965829074,0.0048327763,-0.2153781503,-0.1935855597,0.2620070577,0.1968804896,-0.1408138126,-0.2442782819,0.1170089692,0.0080767199,-0.2089564651,0.067118898,0.0679156333,-0.1068510935,-0.0382123776,-0.2769324481,-0.2390612513,0.3213487267,0.0108338911,-0.2085025012,0.0373120941,0.1167564988,0.5416676998,-0.0854343325,-0.2512436211,-0.23627159,0.1338641346,0.2141526639,0.587795198,0.3229282498,0.0615964159,0.0207409281,0.0353414975,-0.007323816,-0.3004595935,-0.1186441779,0.0404926986,0.2043307871,0.1712900102,0.0379314348,0.2153557092,-0.2638367712,0.2779995203,-0.1914842427,0.0485908091,0.2978192866,-0.1865315437,0.2497089207,0.1298921704,0.3748373687,-0.3129646778,0.1826845109,-0.2493490279,0.1914865822,0.2297151834,-0.0310409069,0.3727429211,0.1146909297,-0.136060372,-0.1334819645,-0.1840704083,-0.0747601911,-0.001825369,-0.1629542857,0.0507639945,-0.0449644476,0.126365304,0.1798333079,-0.2456463724,-0.0594854131,0.2569033206,0.1356156766,-0.3840177059,0.1147181466,0.4462440014,-0.2306538969,-0.1444597393,0.3152908683,0.2557352483,0.0696520507,0.286488533,0.0793633461,0.1090921238,-0.0833654851,0.3604850173,0.354011029,0.1339815259,0.1699279249,0.119309701,-0.0394221283,0.123560898,0.6183962822,-0.3597970307,0.5293438435,0.1383294463,0.223486647,-0.2095681727,-0.1719873846,0.4043190479,0.3491841853,0.0307415258,-0.063723281,-0.0098115513,-0.0151653588,-0.0383445397,-0.0439844429,-0.1558847874,0.1526424289,-0.0238103755,0.1840652823,0.0606990382,-0.2756785452,-0.0898628905,-0.0042633289,0.0583703183,-0.3857102394,0.200837329,0.1657795757,-0.0824829489,-0.0335788727,0.2867563963,-0.0610284582,0.1891627014,-0.0892436728,-0.156272456,0.0612951703,-0.2288975716,0.0749252737,0.0788586736,0.2476976961,0.2621339858,-0.1203019619,-0.021571368,-0.013732655,-0.0387167707,0.1328176111,0.1091329232,-0.0945224315,0.2317098826,0.2226289213,0.0645292401,-0.1721493006,-0.0432614386,0.1390096247,0.003363126,-0.4188861251,0.0442829207,-0.4233117104,-0.1423737407,-0.4599448442,0.1326027811,-0.4896370769,0.0648630932,0.142006591,0.2148461789,-0.1376210749,-0.3698993921,0.0368540995,-0.1593538821,0.256894052,0.3579009175,0.1144520268,-0.2953969836,-0.1813544631,-0.6622396111,0.0893708318,-0.3208755553,0.180541873,-0.1669025123,-0.1049254164,0.2438538074,0.3968717754,-0.0135987466,0.0720950067,-0.0804831684,0.0321504548,-0.3542997241,-0.1191497892,-0.0947066173,0.125385046,-0.1326467395,-0.2455040067,0.085980624,0.0227902029,-0.0070100315,-0.1017634869,-0.3187882304,0.4922682643,-0.0574580431,0.5561671257,0.0317356922,0.3317186832,0.0810964555,-0.1596710533,0.0483924896,0.0461773016,-0.2101214528,0.2767443955,0.0403250046,0.2443181574,-0.2023214251,-0.1389499754,-0.3106030822,0.5070671439,-0.0894303396,0.0841185972,-0.2534946501,-0.0817406476,-0.0257490184,0.1667099744,-0.1138295457,0.0313684382,-0.0296894144,-0.1530405879,-0.5479568839,-0.0543605424,0.401042074,-0.6405128837,-0.3518935144,0.0424926765,0.2713645697,-0.0068686795,-0.0719659328,-0.4508554041,0.0561415404,0.2868243456,0.2098804712,0.0925857127,-0.0562560558,-0.239913404,-0.1344611645,-0.0398824438,-0.4349634051,-0.1276827306,-0.0872320086,-0.0755798146,-0.3943339586]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1790","title":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.","comments":"Thanks for your reply! \r\nI understood.\r\n\r\nI tried again with installing apache-beam, add ` beam_runner=\"DirectRunner\"` and an anther `mwparserfromhell` is also required so I installed it.\r\nbut, it also failed. It exited 1 without error message.\r\n\r\n```py\r\nimport datasets\r\n# BTW, 20200501.ja doesn't exist at wikipedia, so I specified date argument\r\nwiki = datasets.load_dataset(\"wikipedia\", language=\"ja\", date=\"20210120\", cache_dir=\".\/datasets\", beam_runner=\"DirectRunner\")\r\nprint(wiki)\r\n```\r\nand its log is below\r\n```\r\nUsing custom data configuration 20210120.ja\r\nDownloading and preparing dataset wikipedia\/20210120.ja-date=20210120,language=ja (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to .\/datasets\/wikipedia\/20210120.ja-date=20210120,language=ja\/0.0.0\/4021357e28509391eab2f8300d9b689e7e8f3a877ebb3d354b01577d497ebc63...\r\nKilled\r\n```\r\n\r\nI also tried on another machine because it may caused by insufficient resources.\r\n```\r\n$ python main.py\r\nUsing custom data configuration 20210120.ja\r\nDownloading and preparing dataset wikipedia\/20210120.ja-date=20210120,language=ja (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to .\/datasets\/wikipedia\/20210120.ja-date=20210120,language=ja\/0.0.0\/4021357e28509391eab2f8300d9b689e7e8f3a877ebb3d354b01577d497ebc63...\r\n\r\nTraceback (most recent call last):\r\n File \"main.py\", line 3, in <module>\r\n wiki = datasets.load_dataset(\"wikipedia\", language=\"ja\", date=\"20210120\", cache_dir=\".\/datasets\", beam_runner=\"DirectRunner\")\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 609, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 526, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1069, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/pipeline.py\", line 561, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/direct\/direct_runner.py\", line 126, in run_pipeline\r\n return runner.run_pipeline(pipeline, options)\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 182, in run_pipeline\r\n self._latest_run_result = self.run_via_runner_api(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 193, in run_via_runner_api\r\n return self.run_stages(stage_context, stages)\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 358, in run_stages\r\n stage_results = self._run_stage(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 549, in _run_stage\r\n last_result, deferred_inputs, fired_timers = self._run_bundle(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 595, in _run_bundle\r\n result, splits = bundle_manager.process_bundle(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 888, in process_bundle\r\n self._send_input_to_worker(process_bundle_id, transform_id, elements)\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 765, in _send_input_to_worker\r\n data_out.write(byte_stream)\r\n File \"apache_beam\/coders\/stream.pyx\", line 42, in apache_beam.coders.stream.OutputStream.write\r\n File \"apache_beam\/coders\/stream.pyx\", line 47, in apache_beam.coders.stream.OutputStream.write\r\n File \"apache_beam\/coders\/stream.pyx\", line 109, in apache_beam.coders.stream.OutputStream.extend\r\nAssertionError: OutputStream realloc failed.\r\n```\r\n\r\n","body":"```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?","comment_length":279,"text":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.\n```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?\nThanks for your reply! \r\nI understood.\r\n\r\nI tried again with installing apache-beam, add ` beam_runner=\"DirectRunner\"` and an anther `mwparserfromhell` is also required so I installed it.\r\nbut, it also failed. It exited 1 without error message.\r\n\r\n```py\r\nimport datasets\r\n# BTW, 20200501.ja doesn't exist at wikipedia, so I specified date argument\r\nwiki = datasets.load_dataset(\"wikipedia\", language=\"ja\", date=\"20210120\", cache_dir=\".\/datasets\", beam_runner=\"DirectRunner\")\r\nprint(wiki)\r\n```\r\nand its log is below\r\n```\r\nUsing custom data configuration 20210120.ja\r\nDownloading and preparing dataset wikipedia\/20210120.ja-date=20210120,language=ja (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to .\/datasets\/wikipedia\/20210120.ja-date=20210120,language=ja\/0.0.0\/4021357e28509391eab2f8300d9b689e7e8f3a877ebb3d354b01577d497ebc63...\r\nKilled\r\n```\r\n\r\nI also tried on another machine because it may caused by insufficient resources.\r\n```\r\n$ python main.py\r\nUsing custom data configuration 20210120.ja\r\nDownloading and preparing dataset wikipedia\/20210120.ja-date=20210120,language=ja (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to .\/datasets\/wikipedia\/20210120.ja-date=20210120,language=ja\/0.0.0\/4021357e28509391eab2f8300d9b689e7e8f3a877ebb3d354b01577d497ebc63...\r\n\r\nTraceback (most recent call last):\r\n File \"main.py\", line 3, in <module>\r\n wiki = datasets.load_dataset(\"wikipedia\", language=\"ja\", date=\"20210120\", cache_dir=\".\/datasets\", beam_runner=\"DirectRunner\")\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 609, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 526, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 1069, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/pipeline.py\", line 561, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/direct\/direct_runner.py\", line 126, in run_pipeline\r\n return runner.run_pipeline(pipeline, options)\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 182, in run_pipeline\r\n self._latest_run_result = self.run_via_runner_api(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 193, in run_via_runner_api\r\n return self.run_stages(stage_context, stages)\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 358, in run_stages\r\n stage_results = self._run_stage(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 549, in _run_stage\r\n last_result, deferred_inputs, fired_timers = self._run_bundle(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 595, in _run_bundle\r\n result, splits = bundle_manager.process_bundle(\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 888, in process_bundle\r\n self._send_input_to_worker(process_bundle_id, transform_id, elements)\r\n File \"\/home\/miyamonz\/.cache\/pypoetry\/virtualenvs\/try-datasets-4t4JWXxu-py3.8\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 765, in _send_input_to_worker\r\n data_out.write(byte_stream)\r\n File \"apache_beam\/coders\/stream.pyx\", line 42, in apache_beam.coders.stream.OutputStream.write\r\n File \"apache_beam\/coders\/stream.pyx\", line 47, in apache_beam.coders.stream.OutputStream.write\r\n File \"apache_beam\/coders\/stream.pyx\", line 109, in apache_beam.coders.stream.OutputStream.extend\r\nAssertionError: OutputStream realloc failed.\r\n```\r\n\r\n","embeddings":[0.0917942896,0.0745991319,0.0195861496,0.251845777,0.2570836842,0.0409060419,0.3200161457,0.3480713069,0.1667267084,0.1959524751,0.2174150348,0.0488914922,-0.0974617451,-0.2695732117,0.0972716436,-0.3023030162,0.2464139014,0.0408723839,-0.1644416153,-0.2330404669,-0.3931181431,0.1340927631,-0.3705989122,0.0350908488,-0.1296226531,0.3397006094,-0.0780813843,0.0638152286,0.0173252411,-0.2700100541,0.0453570858,-0.324616313,0.1215118393,0.3739317358,-0.0001173204,0.1343809068,0.4923765659,-0.2628591955,-0.3836726844,-0.2569671571,-0.1403349489,-0.4439759254,0.0448291525,-0.4475933611,-0.1867608875,-0.2697493434,0.0532009304,-0.3586476445,0.395362854,0.2671947777,0.1607774794,0.2647910118,0.1899022758,-0.041800715,0.418223381,0.0558920279,-0.2807168067,0.001108119,-0.0778786838,0.1042615697,0.0328844711,0.3811944425,-0.2652494013,0.0351823829,0.219607994,-0.2989069521,0.0275765453,-0.5322272182,0.1891094148,0.1666529626,0.7942336202,-0.3169869483,-0.1359947622,0.0134905996,-0.0399667807,0.0834234431,0.2887854874,0.1149210781,-0.1483090073,-0.0395828001,0.4454575777,-0.2758440971,-0.1213793457,0.3104602993,0.0970906764,0.4800923765,0.0280743092,0.285240382,-0.2220673412,-0.1609825641,-0.0193839762,-0.3305304348,0.2255716175,0.3868534863,-0.2711615562,0.0091108838,0.0303918943,0.1701530665,0.0020982851,-0.2076063156,-0.2619356215,0.1003814191,0.1692396998,-0.0123883486,0.0890526325,0.1528998315,0.3734167516,0.1121669933,0.2206748128,-0.1710113436,0.0056283185,0.0340125859,-0.0521680787,-0.3492521644,-0.3236457109,0.1878946424,0.3179504871,-0.1621291339,-0.0892107636,0.0642646402,-0.6158861518,-0.2876563966,-0.2265936136,0.3501834869,0.0944344401,0.3287703395,0.1527711451,0.2902515829,-0.2045259774,-0.3957129121,-0.0542582273,0.4753558338,-0.3645497561,0.2409627885,0.087393932,0.1367838532,0.1506344676,-0.1451434493,-0.0452821516,0.0375527404,-0.0371271111,0.0036610053,-0.1230324283,0.3735723197,0.1352759451,0.5192132592,0.2212901413,-0.0427361988,-0.0936333016,0.1430326551,-0.182173118,-0.2547640502,0.0409881771,0.0198266562,-0.3785459697,0.1690728813,0.0418639407,-0.1529215425,-0.0193206761,-0.01373877,0.1795716882,0.0659218878,-0.0775272995,-0.24829337,0.1928215921,0.7632588148,-0.1837839335,-0.0071699829,-0.0386874974,-0.0945175886,-0.1337481886,-0.1851740181,-0.0107638203,0.3875401318,-0.2644916773,-0.06522277,0.6910915971,-0.5287438035,-0.0048345681,0.1956304312,-0.0111865513,0.0338991694,0.2492822856,-0.1067635268,-0.2923358977,-0.1760918498,-0.259635061,0.0050270432,0.130420655,-0.0345540307,-0.3318171799,-0.0714585781,0.0876164213,0.2745759189,0.3723239303,0.0291089024,0.2889192104,0.6823456287,0.1276972145,-0.0668428093,0.2065407038,0.5193422437,-0.0871524811,-0.02277302,0.1958383918,-0.0088050636,-0.235602349,0.2555713058,-0.1881491244,0.2531144321,0.0075532491,0.0816442966,-0.1720115095,-0.1806830317,-0.3440953791,-0.2144478112,0.0655369386,0.0168310143,0.0882710665,0.3130510449,-0.3127648532,0.439604193,-0.197692737,0.013464028,-0.4467652142,0.0812010467,-0.1344239563,-0.0401019268,-0.0064320094,0.0986566991,0.3700562716,-0.225478217,-0.1251113862,0.1368416697,0.0944711342,-0.0113428039,-0.0501034856,-0.110739477,0.1207010895,-0.3321649134,0.4237888455,0.1321918815,0.1378662437,-0.1063949019,0.1518630087,0.087618798,0.3021607697,0.4313708246,0.0128362458,-0.0354976282,-0.0844971165,-0.0023270249,-0.2098860145,-0.2393537313,0.2860515416,0.362285763,0.2126557231,-0.0158595368,0.1649516076,-0.3541213274,0.4365091324,0.0414417125,0.2351138145,-0.0608063415,-0.450204879,0.2783480287,0.0984192416,-0.1064570919,0.3670893908,0.1384744048,-0.2372380793,0.1293950081,0.2279010564,-0.0575799905,0.092207022,0.1565444469,0.4523301721,0.1082719192,-0.2713344693,0.0434037186,-0.0382356793,-0.1230822653,-0.0780402496,0.115962863,-0.2822549045,-0.05891186,-0.3727226555,-0.4914531708,-0.3270525634,0.0718242377,-0.4088910222,-0.2334057838,-0.1688344926,-0.1222827435,-0.1346062422,0.3930952847,0.2123800069,-0.2270422429,-0.0681200176,-0.3959479034,-0.2092330903,-0.2032681257,-0.2683122754,-0.0390979871,0.2976367176,0.1895361543,0.192569524,-0.2675949335,-0.0959035605,-0.20994398,-0.1003018096,0.1910176873,0.0083410237,-0.099483192,-0.0512177721,0.456911087,-0.0361206084,0.0497274473,0.3699610233,-0.0084841112,-0.1452653557,0.1777404249,0.0052207164,0.1532685161,-0.0275911447,-0.5211430788,-0.2453733236,-0.392180413,-0.2898659408,0.2246519476,0.0522166938,0.0134093035,0.1566919237,0.1527421176,-0.0422593057,0.0668442696,0.0891847908,0.0916851461,0.4245998859,-0.1491640508,-0.2796567082,0.3423870802,-0.1902858913,0.0289856307,0.0970074683,-0.1730143279,-0.0193443969,0.3139023185,-0.1163840815,-0.0652336404,0.2874525487,0.5190108418,-0.1943499744,0.0928247049,-0.0690189078,0.0083449809,-0.1444002986,-0.2441487312,0.4245213866,0.297408253,0.3056080043,-0.043531213,0.7822398543,0.0108604189,0.2556969523,0.4260956943,0.0762988031,0.2692808211,-0.348682791,-0.0499942712,0.0104912361,0.2544057369,-0.0323169902,0.1112649217,-0.0598254576,0.0762266666,-0.6192234159,-0.369808495,-0.2994987965,-0.2159063518,0.1090988219,-0.274140358,0.191042304,-0.048707325,0.0950107202,0.1237663031,0.0207048543,0.2188324481,0.4647522271,0.0035707247,0.0639350638,-0.0234036613,-0.0922177061,-0.2643969953,0.3273771703,-0.0421187393,0.1358178556,-0.0947875232,-0.2361926585,0.0894806683,0.0631990135,0.3092834055,-0.4035589099,-0.1606223583,0.3869546354,0.0780745149,-0.6535663605,-0.175849244,0.0604872108,-0.0153945163,-0.0175038707,0.0102821812,-0.1933236718,-0.244700402,0.2949996293,0.4285174906,-0.142317906,-0.3749653697,-0.4727918506,0.0281872731,-0.489436239,-0.1883932054,-0.1379600912,0.0788534284,0.2421522439,0.2184175551,0.0651045889,-0.0924429372,0.0236946382,-0.1127033904,0.1126640663,-0.1607255936,-0.161054194,-0.0255069006,0.3260902762,-0.1186316162,0.0631388873,0.104670994,-0.2547037005,-0.0499393418,-0.0169840101,0.0706428662,0.2264755666,-0.0151435975,-0.0195699651,-0.1449170858,-0.1970620304,0.1059784219,-0.1217359379,0.0783259794,-0.2491557896,-0.3910131752,-0.3480043709,0.7426955104,0.0271398667,-0.0159643684,0.3429035246,0.201701194,-0.1514662802,0.3397690058,0.0474293232,0.9796898365,0.014183769,0.0064000688,0.0223962534,0.1022481918,0.8341006041,-0.6507053375,0.066703774,-0.2045768648,0.228632614,-0.0545771271,-0.0850019753,0.5718107224,0.1132340729,0.0078878105,0.6712852716,0.140336588,0.4805466831,-0.0816391334,0.206282571,-0.163616091,-0.0353388563,-0.3509262502,0.0182056464,-0.1721029431,0.8090721369,-0.1634285152,-0.0957965851,-0.2525319159,-0.328134656,-0.2581402957,0.1180371493,-0.2528877258,0.1364771724,-0.3172075748,-0.3505349457,0.0723372251,0.3597848713,0.4233637452,0.0166258439,-0.1673576683,0.1833550185,-0.0774707794,-0.158569783,-0.2163073719,0.1075420007,0.2258039415,-0.1334332675,-0.1807707995,0.1690982282,-0.3425919116,-0.119356297,-0.0113395564,0.0227902774,0.668975234,0.2682593465,-0.2767080963,0.3553762436,0.1283432841,0.0693111718,0.1103161499,-0.2117784917,-0.0266514197,0.0278240386,-0.3508480489,-0.0341709852,-0.1292220354,0.3872739673,0.2311330736,-0.113041617,0.3309112191,0.1251431555,-0.1233675107,-0.2332203388,0.2853122652,0.1719541252,-0.2076363862,-0.2752786875,0.2128836513,0.1703671515,-0.098397024,0.0386644863,0.0335196741,-0.0663089529,0.0395466201,-0.2681114376,-0.1334896982,0.2705458403,-0.038836997,-0.1897752434,-0.0689270571,0.2141501904,0.4571789503,-0.2815128863,-0.1937303692,-0.2539089322,0.0339952707,0.1256140023,0.4958905578,0.3753301203,-0.0051880861,0.0660261139,0.0335872993,-0.0018729224,-0.1946708709,-0.1027396694,0.0161383096,0.2142414749,0.2061053813,0.163252309,0.1095705479,-0.4472341239,0.2173190117,-0.159572795,-0.0223159194,0.2547009587,-0.1646430194,0.2865309417,0.1736465991,0.321105063,-0.529632926,0.2142919749,-0.1817493737,0.3453174531,0.1947263926,0.0170994103,0.4853265285,0.1384412795,-0.4599797726,-0.1773356497,-0.3498986661,-0.0312797911,0.032364998,-0.1172965318,0.0418919921,-0.0186135769,0.0521198623,0.183457613,-0.2758882642,-0.0591307618,0.2848194838,0.1028100103,-0.2974634767,0.1754073054,0.4890995026,-0.0981934592,-0.140495941,0.233376503,0.1426938772,-0.0916714072,0.3412108123,0.1548683792,0.163151294,-0.1787314713,0.2053101659,0.1840970963,0.2948670983,0.1575997621,0.0475279838,-0.0538380109,0.0880151913,0.6539548039,-0.2858646512,0.4924992025,0.1011866555,0.1677298248,-0.2116198093,-0.2416178584,0.3595826328,0.4214579463,0.0517491624,-0.1443926692,-0.1237956882,-0.0773716345,-0.0147103518,-0.0196921397,-0.2302414626,0.1290175915,0.0224811919,0.2884569466,-0.1276941299,-0.2530802488,-0.0551311821,0.029763747,0.0555024967,-0.3724519312,0.153400898,0.2548291087,-0.11216674,-0.0339415595,0.3947484791,0.0025756797,0.0506305136,-0.1041762978,-0.0266681835,-0.0051421253,-0.1853563637,-0.0603177175,0.1111040488,0.2266120762,0.3590020537,-0.1547209918,-0.0779511333,0.1268206984,-0.0255476218,0.1880700737,0.0604324304,-0.155550316,0.2081418484,0.2435003072,0.0489713326,-0.1524852514,-0.0412455387,0.2414973974,0.0703484491,-0.2723352909,0.0677771941,-0.3573551774,0.0387389883,-0.4263155758,0.2583354414,-0.5642676353,0.0693702325,0.0122655304,0.3249205351,-0.1382800937,-0.3196537197,0.0110139456,-0.2432870716,0.1348072886,0.2248660773,0.1185955331,-0.4263069034,-0.3554022014,-0.4867426753,0.0426193364,-0.3128663003,0.1308723539,-0.2051989287,0.0502482057,0.0146705592,0.3552877903,0.046179492,0.2269870043,-0.1186083183,0.2433991879,-0.3425807655,-0.1701988876,-0.0914433673,0.110093914,-0.0911925882,-0.2694914639,-0.0306064449,-0.226007238,-0.0242983829,-0.1080486402,-0.4202312231,0.5537158251,-0.0947477743,0.5825892687,-0.1459338665,0.3008624017,0.0713767335,-0.1396201402,-0.0046735518,-0.0106659625,-0.0691249445,0.154435724,-0.053748101,0.2252952009,-0.2020622343,-0.0179008245,-0.3485690057,0.4514452517,-0.249937132,0.1921251714,-0.1374956071,0.0063455044,0.0350949019,0.2750198245,-0.0353063643,-0.075401336,-0.1648888439,0.0065836059,-0.5357224941,-0.0180694219,0.2722668052,-0.4996759593,-0.217219308,-0.0332432128,0.3454782665,-0.14042826,-0.0262269378,-0.4189309478,0.1473215967,0.3714236319,0.1580149233,0.1390686482,-0.011535665,-0.158378914,-0.0784908608,0.0356109813,-0.5600673556,-0.1644401252,-0.0257759057,-0.0694186836,-0.2995336652]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1790","title":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.","comments":"Hi @miyamonz,\r\n\r\nI tried replicating this issue using the same snippet used by you. I am able to download the dataset without any issues, although I stopped it in the middle because the dataset is huge.\r\n\r\nBased on a similar issue [here](https:\/\/github.com\/google-research\/fixmatch\/issues\/23), it could be related to your environment setup, although I am just guessing here. Can you share these details?","body":"```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?","comment_length":61,"text":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.\n```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?\nHi @miyamonz,\r\n\r\nI tried replicating this issue using the same snippet used by you. I am able to download the dataset without any issues, although I stopped it in the middle because the dataset is huge.\r\n\r\nBased on a similar issue [here](https:\/\/github.com\/google-research\/fixmatch\/issues\/23), it could be related to your environment setup, although I am just guessing here. Can you share these details?","embeddings":[-0.0868910477,0.1544118226,0.0349213034,0.2511788607,0.3533329964,0.1083717346,0.3212845027,0.4499050975,0.3734617233,0.0088102743,0.102325432,0.0063319677,-0.0099073453,-0.0225594807,-0.0207931511,-0.2547829747,0.1721131206,0.0012150535,-0.2130487263,-0.2989025116,-0.3872275054,0.0866747573,-0.3275980651,0.0159582961,-0.0521656424,0.155601874,-0.0279715098,0.1574117094,0.0401064716,-0.1749257743,0.2103829086,-0.1249948144,-0.0505453423,0.3043377995,-0.000107928,0.0701222792,0.4934644401,-0.2194035649,-0.3645669818,-0.4068615735,-0.0887664631,-0.347299546,0.0214760397,-0.410813123,-0.1609840095,-0.1692377776,0.027584387,-0.424421072,0.2876330912,0.2971323729,0.2558148801,0.3765882254,0.1853017509,-0.0902392045,0.4627014101,-0.2508678436,-0.1734852195,0.152273953,0.1026163325,-0.0133816032,0.0926948115,0.4647519588,-0.3321602941,0.1229287907,0.3978904486,-0.2104093581,-0.1099664122,-0.5223609805,0.1911682338,0.2853013873,0.8053057194,-0.3187530339,-0.2136244327,-0.0639347881,-0.0466213264,0.0935501829,0.1669068485,0.1960045397,-0.1385709345,0.0132206343,0.2564665973,-0.27760759,-0.1129507571,0.2547970414,0.0206995755,0.434542954,0.1130512282,0.290637821,-0.1213328838,-0.1309855431,0.0601092987,-0.3038204908,0.1790289581,0.4227370322,-0.393653661,0.0983674303,0.0753656924,0.0314873047,0.2047479749,-0.0097002788,-0.1731485426,0.2701100409,0.1781290025,0.0558500886,0.1880670786,0.2425366193,0.2048566341,0.2180304527,0.3791832328,-0.0918194652,-0.0715203583,0.02139025,0.0087919412,-0.303060472,-0.2551554441,0.2065252811,0.3416895568,-0.2393628508,-0.2113465816,0.1680399925,-0.3933001757,-0.1952902377,-0.1890354902,0.4464722872,0.0903828144,0.3386282921,0.1270755529,0.3620221615,-0.2277982086,-0.4340928197,-0.1698250324,0.2832837701,-0.2518642247,0.1006729528,0.1977664679,0.1170082912,0.1428786665,-0.1512924433,0.0498563163,0.0672362298,0.1198379323,0.0282127038,-0.0978283435,0.3394678831,0.1440699846,0.5026990175,0.1971620619,-0.0773207396,-0.1979626715,0.0690227002,-0.3356408179,-0.0876067355,0.0300218817,0.1454830319,-0.4106610715,0.1392475069,0.0276989378,-0.1609510332,-0.0260570478,-0.1307491064,0.0702846199,-0.0336305313,-0.0718866885,-0.3641683161,0.1907553971,0.7568348646,-0.1171793342,-0.0507880189,0.1492203921,-0.0326972418,-0.0346767493,0.0621211827,-0.0179945864,0.0897547454,-0.2658486664,0.0493914075,0.5301237106,-0.4971586764,-0.3745914698,0.1492233872,-0.1303744167,0.0674136579,0.193470031,0.0981658325,-0.223392114,-0.0491434783,-0.0737069398,0.32162112,0.1220701709,0.1127692088,-0.3805288672,-0.0859911293,0.206578061,0.1555681378,0.2180370688,0.0378918611,0.2462634146,0.1764459312,0.1007474288,-0.2408952415,0.2300170809,0.4026339352,0.0822453201,-0.0768727884,0.2185535282,-0.0472873859,-0.3427559733,0.3855517209,-0.2262893766,0.2532649636,-0.1266787201,0.0323280282,-0.2458299547,-0.0815848038,-0.3950968087,-0.3153595924,0.1326069832,-0.0861049891,0.1729106605,0.4195102155,-0.0904738083,0.5479971766,-0.0806826502,0.0869766548,-0.3802219331,0.3531749547,-0.1568661332,-0.0223212987,0.0899974629,0.1627476811,0.2166815847,-0.2049010694,-0.147184819,0.2453877628,0.0316126756,0.175381422,0.0656094626,-0.1363481283,0.1981263161,-0.417429477,0.2811176479,0.3059633076,0.2019191384,0.0005689365,-0.0248592496,0.31941396,0.2762654126,0.3443269432,0.1001222581,0.1419816464,0.1740662158,-0.115653336,-0.1184245646,-0.2050980926,0.3828374743,0.2041463405,0.2647698522,-0.0008264494,0.0527264811,-0.192902118,0.4928575754,-0.0328532271,0.3465650678,-0.0434932075,-0.4553969204,0.1410555542,0.0728564709,0.0346844718,0.3330875039,0.1398335248,-0.1654711217,0.1523960531,0.1527922004,-0.0811330453,0.070357874,0.0401980095,0.2829311192,0.1309210509,-0.0361876376,0.0502580702,-0.1252750605,-0.0067715566,-0.2078783512,0.1457998157,-0.2919570506,0.0663333833,-0.3126637042,-0.3919141293,-0.1884225458,0.2504111826,-0.47329247,-0.3105266392,-0.1453045309,-0.0280135628,-0.1986221969,0.2472381592,-0.0033547888,-0.192531988,0.1633858979,-0.4583185315,-0.1651168615,-0.3762441278,-0.3328979611,0.0173073206,0.3358540535,0.133705914,0.2784568369,-0.2997746468,0.0515952557,-0.4177368283,-0.2721670866,0.1581750214,-0.0346399918,0.0360914879,0.0005482019,0.4427484274,-0.0048957947,0.0365115628,0.3516401052,-0.2354400307,-0.1929251105,0.3119670451,0.0282668155,-0.0133800702,0.0411232896,-0.6374550462,-0.3070060909,-0.4252651632,-0.1391487718,0.0489384271,-0.0108339433,0.0516505577,0.1416082233,-0.0006669456,0.0320225321,-0.0082557173,-0.0140781803,0.1327501237,0.5062490106,-0.1456731111,-0.5656459928,0.3194148839,-0.1603342742,0.0624311306,0.1956934184,-0.1955649406,-0.0620265044,0.1375658214,-0.0365846641,0.0259969458,0.2926805317,0.3939027488,-0.257181108,0.0382182263,-0.1365457326,0.0062151481,-0.0437446237,-0.0114895953,0.2950164974,0.2645323575,0.1776743829,-0.0364728346,0.7936730981,0.0073540937,0.2100244761,0.4813813567,0.0773186609,0.2936622798,-0.3421624899,-0.1942974031,0.008329886,-0.0027579318,0.0149708977,0.1873188317,-0.0818461403,-0.1355629265,-0.6442763805,-0.3588293791,-0.2766448557,-0.3226227462,0.0453205034,-0.2160212994,0.1843608171,0.2023991197,0.0938611031,0.0738990903,-0.1065479144,0.23401618,0.2605069876,-0.1759538949,0.0929550007,-0.1118317395,-0.1769476682,-0.3282276094,0.4158121943,0.02262423,0.2022850066,0.003037744,-0.2317768782,0.0856690928,0.0189486016,0.4692161977,-0.3120108247,-0.0437582582,0.1216033921,0.2594099641,-0.5550561547,-0.1870956272,-0.0322666131,0.0321294591,-0.0438287407,0.1810924262,-0.2550391257,-0.0802217051,0.0668696836,0.4545513988,-0.0828830004,-0.1274315566,-0.4367995262,-0.0502201021,-0.5733696818,-0.0828643218,-0.0977893919,0.1422573626,0.0972928479,0.162354812,0.0961017013,-0.014274084,-0.0320801698,-0.0085723111,0.3404173553,-0.2043996751,-0.035457246,0.0853673145,0.1748306006,0.0460118912,0.4567006528,0.1034712195,-0.4493019283,-0.0529026166,-0.2239479721,-0.045778729,0.1897377074,-0.0751508027,-0.204116106,-0.1439588964,-0.0854967684,0.0563913062,-0.0305148829,0.0876147449,-0.0236539915,-0.4352927506,-0.3729906678,0.5270693898,-0.0330390967,-0.0570130944,0.3110024929,0.1127648056,-0.1070813164,0.4222750068,0.0187406596,0.8645877838,-0.0188705325,0.2101957649,0.0582884662,-0.0400821939,0.7248646617,-0.5399271846,0.0771638453,-0.2701845765,0.0008098632,-0.0879065245,-0.137129128,0.3018958867,0.0046131369,-0.0912819952,0.7011968493,0.1136356965,0.3072912693,-0.0284123514,0.1268320531,-0.2231795043,0.0247500334,-0.2159255594,0.0936157033,-0.2517371774,0.7930037379,-0.1146824881,-0.1685597599,-0.3997270167,-0.4847644269,-0.2735730708,0.2067976296,-0.3265916705,0.1057648584,-0.136581108,-0.2755322158,-0.016013613,0.1685109884,0.3396864831,0.0835617483,-0.3524114192,0.19915694,-0.0432496965,-0.0633542687,-0.0457835458,0.003071571,0.2717383206,-0.0967706144,-0.2318739295,0.1485337466,-0.2545036077,-0.1019379273,-0.115730606,0.0610876568,0.3317450881,0.2247243822,-0.2792199254,0.2246249765,0.153365016,-0.0630287006,0.1800225079,-0.0494341403,-0.0303298812,-0.0032992174,-0.1809817702,-0.2474683821,-0.1228529289,0.5205882788,0.1839739233,-0.2424342185,0.4066540599,0.1723616421,-0.2104174197,-0.2411000133,0.1621385366,0.1905707419,-0.2650225759,-0.1460513622,-0.0318621732,0.0957714096,-0.1386876106,0.1677304804,-0.0113472678,-0.1096298322,0.095337458,-0.4442583919,-0.0326412171,0.1852376461,-0.0854347944,-0.1010634154,0.0033143249,0.0053180354,0.4032241702,-0.29716295,-0.2888949215,-0.1376590133,-0.1059079096,0.0224616602,0.5605177879,0.2935122848,0.0023231616,-0.0581455044,0.0937645435,-0.1268181205,-0.1937803775,-0.2027932554,-0.0149474302,0.1385528892,0.0434804,0.1630503237,0.0703656152,-0.4823302329,0.2021599561,-0.1527929306,0.0567928627,0.2536279559,-0.1994744837,0.1771403253,0.1085895598,0.184301123,-0.3525300324,0.2674998641,-0.3646936119,0.1721933931,0.1304795146,0.0686091557,0.3867251277,0.1516537815,-0.1735510379,-0.1742301583,-0.3743775487,-0.0482581146,0.2108212411,-0.1414894164,-0.0068628881,0.1340981871,0.1062337533,0.2562635243,-0.3503274024,-0.1075855568,0.1836726367,0.1969269961,-0.4555110633,-0.0751978159,0.403758049,-0.13587825,-0.0214961302,0.1843521893,0.2636778057,-0.24109523,0.3059701324,0.1916551143,0.3945595324,-0.1170290634,0.3011955023,0.3527545333,0.1779989898,0.1396125406,-0.0304129999,0.0474552736,0.135385856,0.648486793,-0.2091656029,0.411473006,0.0674022064,0.1030037999,-0.262419194,-0.3011828363,0.2992061973,0.2204039246,-0.0695167631,0.0292887799,-0.0240158923,-0.0019979517,0.0860410631,-0.0565788373,-0.1585500687,0.1666674316,0.0012737131,0.2253455818,0.0179712363,-0.3519010246,-0.072908178,0.1331725866,0.0312036555,-0.2385888994,0.2281278372,0.1682943255,-0.1240408048,-0.1239727885,0.3742433488,-0.1189199314,0.013631844,-0.137655437,0.0254037883,0.0463537127,-0.1598244458,0.0069211512,0.1635757387,0.2606301308,0.3488022089,-0.2866474688,-0.0238143392,0.1784340441,-0.0675253943,-0.0155993411,0.0920844749,-0.0621978082,-0.0524863042,0.2890604436,0.1136477664,-0.1780467331,0.0953820497,0.0433127917,-0.0369702466,-0.2705107033,-0.0533740595,-0.3792496622,0.0871791542,-0.3402257562,0.0389809646,-0.6098284125,-0.0060964464,0.1512973607,0.1895903945,-0.0063755512,-0.3605642915,0.0605465062,-0.1073406115,0.3097571135,0.4319081008,0.1203090325,-0.3997267187,-0.231517911,-0.7111353278,0.2788400352,-0.4106044471,0.145000428,-0.1495794952,0.0707299113,0.098414056,0.2337602228,0.1786883026,0.3277378082,-0.2366164923,0.1039685011,-0.3699017167,-0.3021474481,-0.2403442413,-0.00115719,-0.1039976999,-0.3225291073,0.1565545499,-0.1339518726,0.0949152112,-0.140835017,-0.3069468439,0.3905301988,-0.0223495811,0.5410960317,0.0151774548,0.4371804893,0.0846307799,0.0127384728,-0.0869757831,-0.0852244869,-0.1114194095,0.1609831303,-0.0790185109,0.0597908497,-0.131600365,-0.1634125412,-0.2565716803,0.4323997796,0.1219811961,0.1696543247,-0.1993857026,0.0667921379,0.0294767153,0.1640390754,-0.1357732862,-0.0447176471,-0.203832984,-0.0330543034,-0.3078811765,-0.2646474838,0.2891611457,-0.4281413555,-0.3605467081,-0.0024637305,0.1913995743,-0.2226697356,0.0112384493,-0.3648004234,0.2167717814,0.3255018592,0.1010457277,-0.1234168783,-0.012208649,-0.1454690546,-0.0881905779,0.0029760017,-0.3081102073,0.0867416263,-0.129656896,0.0083834277,-0.2291346192]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1790","title":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.","comments":"thanks for your reply and sorry for my late response.\r\n\r\n## environment\r\nmy local machine environment info\r\n- Ubuntu on WSL2\r\n\r\n`lsb_release -a`\r\n```\r\nNo LSB modules are available.\r\nDistributor ID: Ubuntu\r\nDescription: Ubuntu 20.04.2 LTS\r\nRelease: 20.04\r\nCodename: focal\r\n```\r\n\r\nRTX 2070 super\r\nInside WSL, there is no nvidia-msi command. I don't know why.\r\nBut, `torch.cuda.is_available()` is true and when I start something ML training code GPU usage is growing up, so I think it works.\r\n\r\nFrom PowerShell, there is nvidia-smi.exe and result is below.\r\n```\r\n+-----------------------------------------------------------------------------+\r\n| NVIDIA-SMI 470.05 Driver Version: 470.05 CUDA Version: 11.3 |\r\n|-------------------------------+----------------------+----------------------+\r\n| GPU Name TCC\/WDDM | Bus-Id Disp.A | Volatile Uncorr. ECC |\r\n| Fan Temp Perf Pwr:Usage\/Cap| Memory-Usage | GPU-Util Compute M. |\r\n| | | MIG M. |\r\n|===============================+======================+======================|\r\n| 0 NVIDIA GeForce ... WDDM | 00000000:09:00.0 On | N\/A |\r\n| 0% 30C P8 19W \/ 175W | 523MiB \/ 8192MiB | 3% Default |\r\n| | | N\/A |\r\n+-------------------------------+----------------------+----------------------+\r\n\r\n+-----------------------------------------------------------------------------+\r\n| Processes: |\r\n| GPU GI CI PID Type Process name GPU Memory |\r\n| ID ID Usage |\r\n|=============================================================================|\r\n| 0 N\/A N\/A 1728 C+G Insufficient Permissions N\/A |\r\n| 0 N\/A N\/A 3672 C+G ...ekyb3d8bbwe\\YourPhone.exe N\/A |\r\n| 0 N\/A N\/A 6304 C+G ...2txyewy\\TextInputHost.exe N\/A |\r\n| 0 N\/A N\/A 8648 C+G C:\\Windows\\explorer.exe N\/A |\r\n| 0 N\/A N\/A 9536 C+G ...y\\ShellExperienceHost.exe N\/A |\r\n| 0 N\/A N\/A 10668 C+G ...5n1h2txyewy\\SearchApp.exe N\/A |\r\n| 0 N\/A N\/A 10948 C+G ...artMenuExperienceHost.exe N\/A |\r\n| 0 N\/A N\/A 11988 C+G ...8wekyb3d8bbwe\\Cortana.exe N\/A |\r\n| 0 N\/A N\/A 12464 C+G ...cw5n1h2txyewy\\LockApp.exe N\/A |\r\n| 0 N\/A N\/A 13280 C+G ...upport\\CEF\\Max Helper.exe N\/A |\r\n| 0 N\/A N\/A 15948 C+G ...t\\GoogleIMEJaRenderer.exe N\/A |\r\n| 0 N\/A N\/A 16128 C+G ...ram Files\\Slack\\Slack.exe N\/A |\r\n| 0 N\/A N\/A 19096 C+G ...8bbwe\\WindowsTerminal.exe N\/A |\r\n+-----------------------------------------------------------------------------+\r\n```\r\n\r\nI don't know what should I show in such a case. If it's not enough, please tell me some commands.\r\n\r\n---\r\n## what I did\r\nI surveyed more and I found 2 issues.\r\n\r\nAbout the first one, I wrote it as a new issue.\r\nhttps:\/\/github.com\/huggingface\/datasets\/issues\/2031\r\n\r\nThe error I mentioned in the previous comment above, which occurred on my local machine, is no longer occurring.\r\n\r\nBut, it still failed. In the previous comment, I wrote `AssertionError: OutputStream realloc failed.` happen on another machine. It also happens on my local machine.\r\n\r\nHere's what I've tried.\r\n\r\nthe wikipedia.py downloads these xml.bz2 files based on dumpstatus.json\r\nIn Japanese Wikipedia dataset that I specified, it will download these 6 files.\r\n\r\n\r\n`https:\/\/dumps.wikimedia.org\/jawiki\/20210120\/dumpstatus.json`\r\nand filtered json based on wikipedia.py is below.\r\n```json\r\n {\r\n \"jobs\": {\r\n \"articlesmultistreamdump\": {\r\n \"files\": {\r\n \"jawiki-20210120-pages-articles-multistream1.xml-p1p114794.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream1.xml-p1p114794.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream2.xml-p114795p390428.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream2.xml-p114795p390428.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream3.xml-p390429p902407.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream3.xml-p390429p902407.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream4.xml-p902408p1721646.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream4.xml-p902408p1721646.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream5.xml-p1721647p2807947.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream5.xml-p1721647p2807947.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream6.xml-p2807948p4290013.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream6.xml-p2807948p4290013.bz2\"\r\n }\r\n }\r\n }\r\n }\r\n }\r\n```\r\n\r\nSo, I tried running with fewer resources by modifying this line.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L524\r\nI changed it like this. just change filepaths list.\r\n` | \"Initialize\" >> beam.Create(filepaths[:1])`\r\n\r\nand I added a print line inside for the loop of _extract_content.\r\nlike this `if(i % 100000 == 0): print(i)`\r\n\r\nfirst, without modification, it always stops after all _extract_content is done.\r\n\r\n- `filepaths[:1]` then it succeeded.\r\n- `filepaths[:2]` then it failed.\r\nI don't try all patterns because each pattern takes a long time.\r\n\r\n### my opinion\r\nIt seems it's successful when the entire file size is small.\r\n \r\nso, at least it doesn't file-specific issue.\r\n\r\n\r\nI don't know it's true but I think when beam_writter writes into a file, it consumes memory depends on its entire file.\r\nbut It's correct Apache Beam's behavior? I'm not familiar with this library.\r\n","body":"```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?","comment_length":606,"text":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.\n```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?\nthanks for your reply and sorry for my late response.\r\n\r\n## environment\r\nmy local machine environment info\r\n- Ubuntu on WSL2\r\n\r\n`lsb_release -a`\r\n```\r\nNo LSB modules are available.\r\nDistributor ID: Ubuntu\r\nDescription: Ubuntu 20.04.2 LTS\r\nRelease: 20.04\r\nCodename: focal\r\n```\r\n\r\nRTX 2070 super\r\nInside WSL, there is no nvidia-msi command. I don't know why.\r\nBut, `torch.cuda.is_available()` is true and when I start something ML training code GPU usage is growing up, so I think it works.\r\n\r\nFrom PowerShell, there is nvidia-smi.exe and result is below.\r\n```\r\n+-----------------------------------------------------------------------------+\r\n| NVIDIA-SMI 470.05 Driver Version: 470.05 CUDA Version: 11.3 |\r\n|-------------------------------+----------------------+----------------------+\r\n| GPU Name TCC\/WDDM | Bus-Id Disp.A | Volatile Uncorr. ECC |\r\n| Fan Temp Perf Pwr:Usage\/Cap| Memory-Usage | GPU-Util Compute M. |\r\n| | | MIG M. |\r\n|===============================+======================+======================|\r\n| 0 NVIDIA GeForce ... WDDM | 00000000:09:00.0 On | N\/A |\r\n| 0% 30C P8 19W \/ 175W | 523MiB \/ 8192MiB | 3% Default |\r\n| | | N\/A |\r\n+-------------------------------+----------------------+----------------------+\r\n\r\n+-----------------------------------------------------------------------------+\r\n| Processes: |\r\n| GPU GI CI PID Type Process name GPU Memory |\r\n| ID ID Usage |\r\n|=============================================================================|\r\n| 0 N\/A N\/A 1728 C+G Insufficient Permissions N\/A |\r\n| 0 N\/A N\/A 3672 C+G ...ekyb3d8bbwe\\YourPhone.exe N\/A |\r\n| 0 N\/A N\/A 6304 C+G ...2txyewy\\TextInputHost.exe N\/A |\r\n| 0 N\/A N\/A 8648 C+G C:\\Windows\\explorer.exe N\/A |\r\n| 0 N\/A N\/A 9536 C+G ...y\\ShellExperienceHost.exe N\/A |\r\n| 0 N\/A N\/A 10668 C+G ...5n1h2txyewy\\SearchApp.exe N\/A |\r\n| 0 N\/A N\/A 10948 C+G ...artMenuExperienceHost.exe N\/A |\r\n| 0 N\/A N\/A 11988 C+G ...8wekyb3d8bbwe\\Cortana.exe N\/A |\r\n| 0 N\/A N\/A 12464 C+G ...cw5n1h2txyewy\\LockApp.exe N\/A |\r\n| 0 N\/A N\/A 13280 C+G ...upport\\CEF\\Max Helper.exe N\/A |\r\n| 0 N\/A N\/A 15948 C+G ...t\\GoogleIMEJaRenderer.exe N\/A |\r\n| 0 N\/A N\/A 16128 C+G ...ram Files\\Slack\\Slack.exe N\/A |\r\n| 0 N\/A N\/A 19096 C+G ...8bbwe\\WindowsTerminal.exe N\/A |\r\n+-----------------------------------------------------------------------------+\r\n```\r\n\r\nI don't know what should I show in such a case. If it's not enough, please tell me some commands.\r\n\r\n---\r\n## what I did\r\nI surveyed more and I found 2 issues.\r\n\r\nAbout the first one, I wrote it as a new issue.\r\nhttps:\/\/github.com\/huggingface\/datasets\/issues\/2031\r\n\r\nThe error I mentioned in the previous comment above, which occurred on my local machine, is no longer occurring.\r\n\r\nBut, it still failed. In the previous comment, I wrote `AssertionError: OutputStream realloc failed.` happen on another machine. It also happens on my local machine.\r\n\r\nHere's what I've tried.\r\n\r\nthe wikipedia.py downloads these xml.bz2 files based on dumpstatus.json\r\nIn Japanese Wikipedia dataset that I specified, it will download these 6 files.\r\n\r\n\r\n`https:\/\/dumps.wikimedia.org\/jawiki\/20210120\/dumpstatus.json`\r\nand filtered json based on wikipedia.py is below.\r\n```json\r\n {\r\n \"jobs\": {\r\n \"articlesmultistreamdump\": {\r\n \"files\": {\r\n \"jawiki-20210120-pages-articles-multistream1.xml-p1p114794.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream1.xml-p1p114794.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream2.xml-p114795p390428.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream2.xml-p114795p390428.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream3.xml-p390429p902407.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream3.xml-p390429p902407.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream4.xml-p902408p1721646.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream4.xml-p902408p1721646.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream5.xml-p1721647p2807947.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream5.xml-p1721647p2807947.bz2\"\r\n },\r\n \"jawiki-20210120-pages-articles-multistream6.xml-p2807948p4290013.bz2\": {\r\n \"url\": \"\/jawiki\/20210120\/jawiki-20210120-pages-articles-multistream6.xml-p2807948p4290013.bz2\"\r\n }\r\n }\r\n }\r\n }\r\n }\r\n```\r\n\r\nSo, I tried running with fewer resources by modifying this line.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/13a5b7db992ad5cf77895e4c0f76595314390418\/datasets\/wikipedia\/wikipedia.py#L524\r\nI changed it like this. just change filepaths list.\r\n` | \"Initialize\" >> beam.Create(filepaths[:1])`\r\n\r\nand I added a print line inside for the loop of _extract_content.\r\nlike this `if(i % 100000 == 0): print(i)`\r\n\r\nfirst, without modification, it always stops after all _extract_content is done.\r\n\r\n- `filepaths[:1]` then it succeeded.\r\n- `filepaths[:2]` then it failed.\r\nI don't try all patterns because each pattern takes a long time.\r\n\r\n### my opinion\r\nIt seems it's successful when the entire file size is small.\r\n \r\nso, at least it doesn't file-specific issue.\r\n\r\n\r\nI don't know it's true but I think when beam_writter writes into a file, it consumes memory depends on its entire file.\r\nbut It's correct Apache Beam's behavior? I'm not familiar with this library.\r\n","embeddings":[-0.0164622311,-0.1208429337,-0.0449261479,0.1289527118,0.3602375388,0.0117902383,0.4635161161,0.3990748227,0.2657029033,0.2212447822,0.3692156672,-0.0926491097,0.0788542926,-0.0503802449,0.3171323836,-0.3475853801,0.3893397152,-0.1023074314,-0.0446148254,-0.2853878438,-0.3970679343,0.0972740725,-0.327434361,0.204333216,-0.2763909101,0.3834485114,0.0855949596,0.0011822655,0.2684019208,-0.0742643848,0.0483055077,-0.3112154603,0.2795861363,0.1183508188,-0.0001173796,0.1539295018,0.278711319,-0.3796845078,-0.2545412481,-0.3259486854,0.1180826053,-0.5497899652,0.0536575466,-0.4102146924,-0.3081610799,-0.3167407513,-0.0167241152,-0.3641595542,0.2351645082,0.2283264399,0.1696380079,0.2152368277,-0.0553050004,0.0463573895,0.4933214486,0.0602145866,-0.3058097661,-0.0181186721,0.069760412,0.0744795501,0.1465839148,0.6420863867,-0.2314606458,-0.0152312061,0.0673134327,-0.0167856645,0.1663643569,-0.3263391256,0.19459562,0.0899543688,0.4828100801,-0.2662925422,-0.0162716396,0.0410083421,-0.1090239137,-0.1496426016,0.2806896865,0.2335036099,-0.2153159529,0.0255010594,0.446775943,-0.075121589,0.0193944462,0.1508821994,-0.0414782725,0.6642962694,0.0698077157,0.3139586449,-0.1830180436,-0.0387369879,-0.2058107555,-0.1763622612,0.2024045289,0.1869180501,-0.4122222066,0.0578316189,0.1730617136,-0.1539803296,-0.1882955134,0.0296372641,-0.2198068798,0.022954246,0.5886098742,0.0437284037,-0.1492434144,0.1641064882,0.2641566396,-0.1119217053,0.2892181575,-0.2116505355,-0.0529462434,-0.0375723504,-0.1176789403,-0.3208047152,-0.6114045382,0.2161325961,0.3799573481,-0.1107203215,-0.3758541346,0.0015660009,-0.7844842076,-0.1767554581,-0.0009919176,0.5293400884,0.2928486764,0.418664515,0.2891373038,0.2036619633,-0.2172861695,-0.3478887975,-0.0793721676,0.2945079505,-0.3448713422,0.2523286641,0.1170190796,0.1859697104,0.2121743113,-0.2013958693,-0.0995041579,0.0343454182,0.0177220199,-0.0544664115,0.1690147072,0.4799576998,0.0484015569,0.4080634713,0.2257784307,0.0116720451,-0.1172764748,0.2429509014,-0.154858768,-0.3944120705,-0.1468821168,0.0511061251,-0.016755797,0.1968870014,0.3628740013,0.0426825397,0.1228851378,0.012357777,0.253030926,0.0178576149,-0.1054161936,-0.212425977,0.2507418394,0.5459201932,-0.303904146,-0.1106563732,0.3961590827,-0.105383873,-0.078605324,0.1152247265,-0.097856611,0.163681522,-0.2283571512,-0.0972187445,0.4880371392,-0.5070934892,-0.0110670151,0.0804193318,-0.1673003882,-0.1454685032,0.4160169363,-0.0158101302,-0.215586096,-0.1384606659,-0.2177503705,0.0179399848,0.0955032632,-0.0426162966,-0.2979273498,-0.0848434418,0.2965357006,0.21431683,0.236806035,-0.1862458289,0.1066011116,0.7018893957,0.180423528,-0.1468919516,0.1232184842,0.1206345037,0.2017994672,-0.0314943567,0.2057330459,0.081673488,-0.319511503,0.1778331995,-0.1611637324,0.2450395823,-0.0391570479,0.0096766958,0.1222385839,-0.0296207815,-0.1899358928,-0.4091172814,0.0764830336,0.1182895973,0.0033131961,0.221707359,-0.3516422212,0.2491336018,-0.6650169492,0.0230833925,-0.1228884459,0.0653185621,-0.2248590589,-0.1393245459,-0.2083543092,0.2041692883,0.1631704569,-0.2657233477,-0.0550111122,-0.0124198413,0.0863434672,-0.1019486636,-0.1688940376,0.1001053229,0.1083895713,-0.1080817431,0.3500638306,0.0759367421,0.2190008312,0.0002390871,-0.0466164351,0.1902189553,0.3913536072,0.4031084478,0.0441157036,-0.0826886818,0.1671350598,-0.0675747916,-0.2320238948,-0.2165590972,0.1896586716,0.316590935,0.0288632903,-0.0392221026,0.0378770046,-0.1052472144,0.592294991,0.1246693283,0.297246784,-0.0710373074,-0.3637948334,0.1717136204,0.2669872046,-0.136199981,0.3765210509,0.1118713319,-0.2514681816,0.0740888342,0.1832255572,-0.1975178421,0.1446842402,0.0935384706,0.1617263407,0.1095689833,-0.1530915946,-0.00496019,-0.1756981611,0.0117729921,-0.0014159677,-0.1109336689,-0.2594310939,0.1029188037,-0.3478597403,-0.4859148264,-0.2813389003,0.0404133014,-0.2785394788,-0.0732485875,-0.1160379052,-0.0294321608,-0.2187058777,0.4304548502,0.2571213245,0.021932533,-0.0411654711,-0.3306412995,-0.1026054248,-0.2048814148,-0.2449599057,-0.003624555,0.3105489612,0.1761790067,0.2550486326,-0.3488143682,0.0870179385,-0.108633928,-0.3360551894,0.1465450972,0.0418365188,-0.2310839593,-0.1000549942,0.1832405925,0.0651180074,-0.0239215121,0.4265317917,-0.1253953725,-0.1427242905,-0.0013663948,-0.049899675,-0.0958694741,-0.1020317972,-0.4964431524,-0.2645772099,-0.4758715034,-0.4259921908,0.2412468791,0.0810968503,-0.1281322837,0.16065301,0.292352736,-0.1703210622,0.0371363312,0.0311211087,0.2079810202,0.2317314893,-0.0090702567,-0.3604699969,0.4372163713,-0.0842040181,0.0385988802,0.1158174276,-0.2603441477,0.013745781,0.2989229858,-0.0371980071,-0.0009432279,0.2632432282,0.7203304768,-0.225091055,0.0127088632,-0.0141631505,-0.3290919363,0.0510527492,0.0073770634,0.5198993087,0.1697766036,0.4387988746,0.0848533288,0.8618529439,-0.0845359787,-0.0700841323,0.3741458356,0.0939900205,0.207183376,-0.316272378,-0.0314853527,0.2195483893,0.0864551291,-0.0109632937,0.1871255636,-0.0883607939,0.1706031412,-0.5178010464,-0.2898615897,-0.0111297648,-0.1790732592,0.3663791716,-0.2854356468,0.1987072527,0.0410362408,0.1937968582,0.0469065607,0.1392715424,0.1639201194,0.6295225024,-0.0126570882,0.0139247552,-0.1779278219,-0.12026342,-0.1199805811,0.2195675224,0.0725238547,0.218949765,-0.1290306747,-0.4745143056,0.185698837,0.0349481441,0.321356833,-0.0358856581,-0.2871356905,0.3316368461,-0.2321506441,-0.3497001231,-0.3062481582,-0.0189685114,-0.1487324387,-0.0570861809,0.0013778289,-0.0317172557,-0.4585867524,0.37618348,0.2340424508,-0.2524055243,-0.2274864763,-0.4287629426,-0.046967864,-0.5095894933,0.0084903194,-0.1324009895,-0.129297778,0.0264418628,0.1764395088,-0.0690718293,-0.2017700672,-0.0562061295,-0.062993452,0.0851252154,-0.069971092,-0.2636384964,-0.2326473892,0.7186853886,-0.1507540494,-0.0402487777,0.1925615221,-0.0018099736,0.119908385,-0.1079837009,0.2137888372,0.4486494362,-0.176803872,0.1594310552,-0.2837937772,-0.1207114384,-0.1063013822,-0.1808039695,0.3044555485,-0.0509129241,-0.0430611707,-0.3201408684,0.6241831779,-0.1073555723,0.0044305995,0.1471897513,0.131854251,-0.1044127643,0.3319960833,0.2336579412,1.0358109474,0.1642733216,0.1324460208,0.1799951345,0.1077856719,0.8329393864,-0.5984461904,0.2173622996,-0.0851069242,0.1365096718,-0.1082874462,-0.0818740129,0.594587028,-0.0067094401,0.0705297813,0.3790996969,0.2664603591,0.6037870646,0.0037625036,-0.1099149436,0.0355485082,-0.1497260779,-0.5489433408,-0.0168923866,-0.0744285733,0.7894800305,-0.2071013004,-0.1128803119,-0.228025794,-0.3641822636,-0.0988957807,0.0411763303,-0.2909246385,0.1858052611,-0.3828192949,-0.3268052936,-0.0402576178,0.1437874734,0.3549681902,-0.0868406072,-0.1235941052,0.0426940024,0.0076065077,0.0000147992,-0.2081599981,0.036792852,0.3847236335,-0.2481488436,-0.2196934074,0.1634670049,-0.1109325215,-0.0494442247,0.0160127971,0.0696564838,0.7141798139,0.2927417457,-0.2967756987,0.3870972097,0.0980729312,0.0052476195,0.0951581746,-0.2514205873,-0.1515534967,-0.0723019838,-0.2172103077,-0.1010128856,-0.2248369306,0.3276063502,0.1999413967,-0.0972818658,0.2712373734,0.1732693613,-0.1741064936,-0.2119039744,0.1623615474,0.0368800163,-0.4369900525,-0.2890703082,0.302382648,-0.1709166914,-0.1562230736,0.1783195585,0.0681604445,-0.1861292869,-0.0853959098,-0.2221893221,-0.2340110093,0.2098848224,-0.1233102903,-0.2150214314,-0.0631809756,0.2806889117,0.4156888127,-0.1536999792,-0.2261552811,-0.2628735602,-0.1230095252,0.1928089857,0.4427232742,0.2965599597,-0.0127700642,0.1895495951,0.0074948054,-0.0380019471,-0.2186236978,-0.1312117279,-0.0150048006,0.1765696704,0.2285003364,0.1959930509,0.1880875826,-0.2676889896,0.0865137875,-0.0725882575,0.0992606804,0.1864968836,-0.0894795656,0.35080266,-0.0306736249,0.3316594958,-0.4477111399,0.237273857,-0.2499564886,0.3688063622,0.2676358223,-0.0803975984,0.4092100859,0.2389479578,-0.3746153712,-0.1523714364,-0.3645432591,0.0104731806,-0.0546281151,-0.0719809979,-0.1021174043,-0.1859799176,-0.0245211534,0.1607095152,-0.3545147777,-0.0365306512,0.2327327132,0.1266964227,-0.4543902874,0.3731452227,0.242768243,-0.2154417485,-0.0772716627,0.3060437143,0.4231770039,-0.0923661664,0.3880966902,0.1781988144,0.0841993615,-0.2030155361,0.2066869289,0.3618157208,0.2963565588,0.3598019779,0.0202724896,0.1120370626,0.195010975,0.5736836195,-0.1265514642,0.5901022553,0.1574232727,0.3495480716,-0.096941486,-0.1490585208,0.3120417595,0.3322589397,-0.0146958921,-0.3303588927,-0.2264349908,-0.3042713404,0.045434121,-0.194999367,-0.2628131509,0.1808093339,-0.1972092986,0.0793892741,-0.1102149859,-0.1240668893,-0.0704076588,0.0157016125,0.0519338213,-0.1951790154,-0.0949360952,0.1215049922,-0.1331934333,0.0533881448,0.1281686276,-0.0643568784,0.030802127,-0.086005643,0.0869326964,0.1445930749,-0.1378281265,-0.0600599311,0.0844450593,0.2197665721,0.4621062577,-0.0828695297,-0.0257999152,0.1449400634,0.0030351239,0.0541763157,0.1591967195,-0.0613807403,0.2768367529,0.2779884636,0.0565829128,-0.1591497511,0.1218319982,0.1582957506,-0.0883103088,-0.2958942652,0.2388644218,-0.4307894111,0.0510210767,-0.5330103636,0.2233333439,-0.3449479342,-0.0787925944,0.0157657973,0.1041164994,-0.0151449498,-0.2899124622,0.017829068,-0.2573353946,0.2379448563,-0.0647530779,0.0380463265,-0.2512810826,-0.2304193974,-0.3195031881,0.1034676284,-0.2553522289,0.2922728956,-0.2727172375,0.1275989562,0.1763634533,0.2395235002,-0.0413038209,0.1959237009,0.075729385,0.0428490825,-0.3069564402,-0.2197024077,-0.3401431441,0.116020143,-0.1083828062,-0.2545761764,-0.0424585976,-0.2332197279,-0.0129375588,-0.3310772479,-0.2316154689,0.4752678573,0.0814733356,0.548107326,-0.0275500752,0.2614423633,0.1008105427,-0.0696966425,-0.0400298126,0.0391201563,-0.2041771859,0.19289428,0.0422978103,0.0197436195,-0.0921124145,-0.1000735909,-0.2449546307,0.3905465901,-0.0151442718,-0.0077196048,-0.1053684428,0.0947116464,-0.0885487348,0.0687782541,0.2280658782,-0.0909727067,-0.140173167,-0.1901365966,-0.4491505027,0.0603900403,0.2830468714,-0.3310953379,-0.1412520111,-0.0922324136,0.3614767194,-0.2765082121,0.0967674777,-0.3255517483,0.0648278221,0.2081926763,0.0999976322,0.0435724258,0.1568867564,0.0078008212,-0.2377977222,-0.0828835964,-0.7149184942,-0.1067865193,-0.001387068,-0.1434371173,-0.4423049986]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1790","title":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.","comments":"I don't know if this is related, but there is this issue on the wikipedia processing that you reported at #2031 (open PR is at #2037 ) .\r\nDoes the fix your proposed at #2037 helps in your case ?\r\n\r\nAnd for information, the DirectRunner of Apache Beam is not optimized for memory intensive tasks, so you must be right when you say that it uses the memory for the entire file.","body":"```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?","comment_length":72,"text":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.\n```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?\nI don't know if this is related, but there is this issue on the wikipedia processing that you reported at #2031 (open PR is at #2037 ) .\r\nDoes the fix your proposed at #2037 helps in your case ?\r\n\r\nAnd for information, the DirectRunner of Apache Beam is not optimized for memory intensive tasks, so you must be right when you say that it uses the memory for the entire file.","embeddings":[0.1117349043,-0.0064410353,0.008770884,0.3851280212,0.3766983449,0.0904633775,0.1866939515,0.4343985915,0.2380293459,0.1273683906,0.1705141813,0.0224920474,-0.0121388175,-0.2175724953,0.0293770581,-0.3352380395,0.2749054134,-0.0193757229,-0.170393005,-0.2335792184,-0.3629105985,0.1133486554,-0.4148781896,-0.1094507873,-0.1601733416,0.1928723603,-0.0480315238,0.0954622552,0.1394944191,-0.3413199782,-0.1171517968,-0.1067331508,-0.0180490203,0.3060228229,-0.0001087955,0.0912562832,0.5745626688,-0.1888058186,-0.2846512198,0.0406400599,-0.0371731855,-0.495447129,-0.0139747849,-0.4732246399,-0.0351132341,-0.2119355202,0.1526018977,-0.3777803779,0.3918661177,0.1425276399,0.2509561777,0.2651480734,0.0663075224,-0.0078427643,0.5856633186,-0.0968204886,-0.0510738976,-0.1122035012,-0.2888480127,-0.0498995297,-0.2235361189,0.599576354,-0.2624768615,0.0702701062,0.3599029779,-0.2300790399,-0.149204731,-0.4521154761,0.3267112672,0.1715229154,0.8122545481,-0.2932122052,-0.1780870855,-0.1161256284,-0.1816053092,0.0567492656,0.2139713913,0.3344722688,-0.1897802502,-0.0612507463,0.4706632793,-0.227505818,-0.0409827977,0.0693851858,0.020294413,0.3031225502,0.0945488587,0.1988317519,0.1080079526,-0.0590918399,-0.0815237164,-0.2535690367,0.0567316376,0.2874174416,-0.3248102665,0.1121491566,0.1160264686,0.062517181,0.1234364212,-0.2639259398,-0.2927086353,0.2417604625,0.2671032548,0.0355527736,0.2241144478,0.1502585113,0.0286995266,0.1377189308,0.5333328247,-0.1961988807,-0.1896426678,-0.0155713782,0.0257251393,-0.2397148013,-0.083557196,-0.070114553,0.1421355158,-0.1068671793,-0.0019677756,0.1800702661,-0.2986363471,-0.1659310758,-0.2292383164,0.5316998959,0.0575341061,0.5937864184,0.2013931125,0.126219213,-0.343328476,-0.2992616296,-0.1430035681,0.3405376673,-0.3400125802,0.2080896795,0.0598624051,0.2209750414,0.2671977878,-0.1440650076,0.0279361475,-0.0987526625,0.1566368192,-0.1712189764,-0.0322682187,0.4470678568,0.0875977725,0.4486974478,0.1538452506,-0.0098935515,-0.0720622018,0.2632384896,-0.2432984412,-0.3493455946,0.2551338673,0.1173766479,-0.287961781,0.2902911603,0.0025140997,-0.0284827761,0.1678847522,-0.0079855034,0.0465840176,-0.0180220101,-0.2205018252,-0.3830631375,0.2032209784,0.796346128,-0.069289431,0.0477906503,0.0506921969,-0.0662884638,0.1584432423,0.1949176788,-0.0704009309,0.2086960524,-0.1676570475,-0.0316931754,0.5124144554,-0.5858023763,-0.1874885708,0.2188697606,0.0223700199,0.033020895,0.266482085,0.107522808,-0.1900941581,-0.1260578781,-0.0653749332,0.1762634218,0.1519229859,0.3302550614,-0.5582460761,0.0229572281,0.0813492984,0.1652372479,0.3261908293,-0.0923689976,-0.013221873,0.5644573569,0.2360059023,-0.2773537338,0.2775585949,0.4496910274,-0.1457103342,-0.021472482,0.3341239393,-0.0170702338,0.0314774066,0.2832883596,-0.2546029091,0.2395661175,0.017152505,-0.1440320611,-0.0935746357,-0.009036338,-0.217204392,-0.3565832376,0.1877317429,-0.0370384529,-0.0251046866,0.2928380966,-0.090611212,0.3070314527,-0.198390156,-0.0440984257,-0.3090109825,0.0575743206,-0.2160296738,-0.009151265,0.0731722489,0.0985509753,0.1550627649,-0.039197538,-0.1397496313,0.2085858583,0.2723045051,0.0015527311,0.064366959,-0.0344755128,0.1558855325,-0.2128017694,0.3646449447,0.2092391551,0.2507326901,-0.1380883306,-0.0111312671,0.1207710877,0.2412226498,0.3808264136,0.1354600191,-0.0667670816,0.0642810762,0.0613384545,-0.0879729241,-0.2283333838,0.1964172274,0.3356743753,0.0988572016,0.0624628328,0.0435915925,-0.2446837723,0.5019389987,0.0910699517,0.1994129121,0.0351188853,-0.5036807656,-0.0295866113,0.1220312491,-0.123221539,0.2044374496,0.1894410253,-0.0085466793,-0.0047064982,0.3041694462,-0.0965253934,0.2194781303,0.1881991476,0.3777441382,0.021214461,0.0054696547,-0.0665690377,-0.1015988737,0.0596394688,-0.1179074049,0.1842221469,-0.1934273839,-0.0259490423,-0.286480695,-0.4760195315,-0.1996368617,0.3563892543,-0.3619391918,-0.2971886098,-0.0733863041,-0.0943983123,-0.251352638,0.2048183233,0.1730594784,-0.0069653168,0.0783338472,-0.2935427427,-0.1500487328,-0.5236815214,-0.2300798893,0.0003816478,0.3928463161,0.05052628,0.2879631221,-0.0820050389,-0.0886784419,-0.1884259135,-0.0557332896,0.097169362,0.0015285301,-0.0784612224,-0.1432334036,0.4003280103,0.0184107292,0.0845485553,0.346652329,-0.1144120619,-0.0997573361,0.1798037142,0.1147885472,0.1361715198,0.0201299712,-0.407571137,-0.1568488032,-0.5240032673,-0.036450047,0.1353272945,0.14698717,-0.0409758911,0.1697188765,0.0780954957,0.1428804249,0.141445592,-0.0451132134,0.101059787,0.3315563202,-0.1958372146,-0.4916899204,0.2615267634,-0.1101638898,0.0772579312,0.2027608901,-0.3696198761,-0.1323046833,0.0450658016,-0.0470005199,-0.0215638541,0.3600570261,0.4021160603,-0.1137850657,-0.0945901349,-0.0554751232,-0.0545101985,-0.1878399998,-0.3114802241,0.126268521,0.1768587083,0.2660343051,0.124463588,0.8918671608,0.061335057,0.528016448,0.4090193808,0.1065370813,0.2370456606,-0.3116820157,-0.0100682499,0.0871226564,-0.021293534,-0.1122227311,0.2411689162,0.0693531185,-0.255313158,-0.4276477396,-0.4160268307,-0.2520229816,-0.3027328253,0.2996943593,-0.1315898448,0.2387937456,0.1150137931,0.1127486452,-0.1054300889,-0.0923343822,0.2437108755,0.1362127215,0.1016861126,-0.0783150345,0.1050666869,-0.0967992991,-0.4224517047,0.3081125617,-0.0670120418,0.0607825108,0.0079725627,-0.3941713572,0.1360591948,-0.1513961554,0.5284689069,-0.1971025318,-0.1498373151,0.210680604,-0.0545678586,-0.6518242359,-0.0845868811,0.0267251879,0.0095116794,0.0669475496,-0.0528255999,-0.3164817989,-0.1907039732,0.1585877687,0.4411771595,-0.1934053749,-0.1890026331,-0.5053489804,-0.0205361899,-0.5622540116,-0.0558291227,-0.106750235,0.0558152385,0.1581384689,0.236536175,0.1242459863,-0.0472104289,-0.0412538014,0.0051669753,0.3255873024,-0.2354885042,-0.1083526388,-0.0681345165,0.1213603392,-0.1879368871,0.2294452339,0.0457890853,-0.1827738285,-0.0938026458,-0.0796409771,0.027724864,0.3150016069,-0.0404921286,-0.0408106968,-0.1707971692,-0.119384937,-0.1208507866,0.0131126624,0.1002247632,-0.1311949193,-0.2984963357,-0.483643949,0.617990315,0.0960463434,-0.0949290842,0.4127902389,0.0930462107,-0.2057657391,0.577888906,0.1268905997,0.9856177568,-0.0980039537,0.2745046616,0.0798085481,0.0544569865,0.6281552315,-0.7983608246,0.2477318645,-0.295081377,0.0890977457,0.0915507153,-0.023363078,0.2411326766,-0.0934451967,0.0697581768,0.4806116819,0.1804445386,0.2832387388,-0.1868930906,0.2330121696,-0.1742579937,-0.1990815997,-0.2856556475,0.1044498384,-0.065561302,0.4440692663,-0.1535347104,-0.0071035875,-0.1861335486,-0.2392897755,-0.3046659231,-0.0099200318,-0.1477642655,0.1759129018,-0.3943840265,-0.3973295987,-0.1162180007,0.3625176549,0.2008932233,0.1639960408,-0.2189245671,0.2082365453,-0.1592372805,-0.0306057166,-0.122702904,0.1668754965,0.2763910592,-0.2980985045,-0.1943889707,0.0629001185,-0.3802568913,-0.1166154593,-0.1380203813,0.1541883945,0.4423926175,0.2177787423,-0.304763943,0.2423822731,-0.1472385973,-0.0508678481,0.1976381987,-0.084267512,-0.1481400728,0.1261873096,-0.1883113533,-0.1309754401,-0.0109779034,0.3976472616,0.3433181643,-0.0041510006,0.4232973456,0.052104719,-0.3017267883,-0.2160416096,0.1823889315,-0.0104284268,-0.118382737,0.0265465043,0.1244933978,-0.0301535539,-0.2032164186,-0.0371115506,-0.0755902454,-0.1407431215,-0.0820235908,-0.2634491026,-0.2033625841,0.2270507812,-0.1080821157,-0.1015240252,-0.0274177529,0.046967376,0.2720535994,-0.0508882627,-0.3056166768,-0.1709724665,-0.0949804261,0.2063294798,0.3415115178,0.2306925505,0.0489353091,0.1019946858,0.076737605,0.1110586002,-0.2087029666,-0.2281427979,0.0516656302,0.1599779129,0.0309397746,0.0177851431,0.1988802254,-0.391528815,0.182553485,-0.3639070392,0.1421436369,0.1092841849,-0.2722728252,0.2204234302,0.1361768097,0.2705431581,-0.1716288328,0.2488495111,-0.2288345546,0.3137668073,0.2543608248,0.0372395143,0.4439241588,0.00920597,-0.3684693575,-0.1206266135,-0.2661694586,-0.0666710883,0.1150611192,-0.1013009399,-0.1882531941,0.0767025501,0.0809456259,0.2771137059,-0.1883517057,0.0011662986,0.2110152543,0.2167643458,-0.2984398603,0.0378491879,0.3729888797,-0.3067619205,-0.073912397,0.1783727258,0.2772955894,-0.1161824614,0.1965162158,0.2012996078,0.1917830706,-0.3311785161,0.1913967729,0.0333522595,0.2625003457,0.0091253761,-0.0425488651,-0.0267030615,0.2035576701,0.6850838065,-0.1094501615,0.5020678639,0.2098357677,0.1165139824,-0.3393754065,-0.3047336638,0.362703681,0.3669022918,-0.0226041153,0.0368798524,-0.0835043415,-0.0752135888,0.2278695852,-0.1050911024,-0.1897900403,0.2455680817,-0.087779507,0.0067217676,0.0493445694,-0.2445988655,0.0870347321,0.2050684392,0.1063699797,-0.2664584816,0.1961647421,0.254200995,-0.1388517022,-0.0565979034,0.4479157329,-0.0961103961,0.297023952,-0.2762940526,-0.1655033529,0.0654663816,-0.0799275413,-0.0222030673,0.0861077756,0.3277603686,0.3482866883,-0.3159306645,-0.0767402798,0.0297199916,-0.0269048754,-0.0978560597,0.1472106129,-0.0032557347,0.3576393127,0.1961808056,0.0948207751,-0.225752905,-0.1506104022,0.2103355527,-0.1046937704,-0.2842313051,-0.0389176384,-0.1862660497,-0.2125804871,-0.3217767775,0.0565657727,-0.4617812932,0.1619290709,0.0885244533,0.1317564696,-0.1158986017,-0.2898355722,0.0823644623,-0.1569626778,0.3282449841,0.1139904857,0.1221565604,-0.3864789009,-0.2317990959,-0.5559620261,0.14743568,-0.4531601965,0.1083783582,-0.1252708733,0.2878430784,0.0056685354,0.195599094,-0.0177467316,-0.0319790803,-0.0608994924,0.2072953433,-0.4182611406,-0.1043717861,-0.2250997275,-0.0161748324,-0.0684762076,-0.3324565291,0.0818646252,-0.1801074296,0.0814111903,-0.0909273848,-0.3238874972,0.4874147475,0.0769171268,0.6690852642,0.0324620344,0.2532437146,0.0384891182,-0.1536326557,-0.1856962442,-0.0469212458,-0.145180732,-0.0742813796,0.0275628958,0.2506354749,-0.2115619779,-0.2013020515,-0.2911404967,0.594409287,-0.173104316,0.2295299619,0.0483158045,-0.0574910045,0.1654849648,0.2090781778,-0.1006283462,0.097866185,-0.1423152983,-0.1428383738,-0.592769444,-0.2566542327,0.3185786903,-0.6523000002,-0.3354024887,0.0270409342,0.3586628139,-0.2010263801,-0.0247714575,-0.4774939418,0.1134856343,0.2376119047,0.2094261795,-0.1371398419,-0.0500030555,-0.2336895764,-0.0083020888,-0.063583672,-0.3490174413,-0.1845492125,-0.2370746732,-0.0594089217,-0.3310347795]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1790","title":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.","comments":"the #2037 doesn't solve my problem directly, but I found the point!\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/datasets\/wikipedia\/wikipedia.py#L523\r\nthis `beam.transforms.Reshuffle()` cause the memory error.\r\n\r\nit makes sense if I consider the shuffle means. Beam's reshuffle seems need put all data in memory.\r\nPreviously I doubt that this line causes error, but at that time another bug showed in #2037 made error, so I can't found it.\r\n\r\nAnyway, I comment out this line, and run load_dataset, then it works!\r\n\r\n```python\r\nwiki = datasets.load_dataset(\r\n \".\/wikipedia.py\",\r\n cache_dir=\".\/datasets\",\r\n beam_runner=\"DirectRunner\",\r\n language=\"ja\",\r\n date=\"20210120\",\r\n)[\"train\"]\r\n```\r\n![image](https:\/\/user-images.githubusercontent.com\/6331508\/112283369-6a9f3300-8ccb-11eb-82e5-827bf7fddfb9.png)\r\n\r\nDataset has already shuffle function. https:\/\/github.com\/huggingface\/datasets\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/src\/datasets\/arrow_dataset.py#L2069\r\nSo, though I don't know it's difference correctly, but I think Beam's reshuffle isn't be needed. How do you think?","body":"```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?","comment_length":111,"text":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.\n```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?\nthe #2037 doesn't solve my problem directly, but I found the point!\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/datasets\/wikipedia\/wikipedia.py#L523\r\nthis `beam.transforms.Reshuffle()` cause the memory error.\r\n\r\nit makes sense if I consider the shuffle means. Beam's reshuffle seems need put all data in memory.\r\nPreviously I doubt that this line causes error, but at that time another bug showed in #2037 made error, so I can't found it.\r\n\r\nAnyway, I comment out this line, and run load_dataset, then it works!\r\n\r\n```python\r\nwiki = datasets.load_dataset(\r\n \".\/wikipedia.py\",\r\n cache_dir=\".\/datasets\",\r\n beam_runner=\"DirectRunner\",\r\n language=\"ja\",\r\n date=\"20210120\",\r\n)[\"train\"]\r\n```\r\n![image](https:\/\/user-images.githubusercontent.com\/6331508\/112283369-6a9f3300-8ccb-11eb-82e5-827bf7fddfb9.png)\r\n\r\nDataset has already shuffle function. https:\/\/github.com\/huggingface\/datasets\/blob\/349ac4398a3bcae6356f14c5754483383a60e8a4\/src\/datasets\/arrow_dataset.py#L2069\r\nSo, though I don't know it's difference correctly, but I think Beam's reshuffle isn't be needed. How do you think?","embeddings":[0.119434841,0.110886395,0.0094729532,0.4295069575,0.5010867715,0.0540743396,0.1819471717,0.4008020759,-0.0547771603,0.316262126,0.1112686098,0.0270881485,-0.1956253052,-0.3097158074,0.0873773694,-0.3430968225,0.269775033,-0.0960953832,-0.2489880174,-0.2321131676,-0.3576426208,0.0421388857,-0.4852919579,-0.1931643486,-0.2590049207,0.3534104824,-0.1904881597,0.2451300025,-0.0117135011,-0.17570813,-0.0820924789,-0.1067039967,-0.1121078283,0.3131394386,-0.0001133347,0.1047418267,0.3758174777,-0.2779262066,-0.3259073198,-0.1989419013,-0.0553082637,-0.3000113964,-0.1625952423,-0.3917806149,-0.0717182159,-0.2675777078,0.1196648106,-0.3548972011,0.422327131,0.201214999,0.2053963393,0.225887537,0.0281292796,-0.0164729133,0.6700220704,0.0871259719,-0.1116377413,0.0683365986,-0.1488498002,0.0098396195,-0.1615006328,0.549500823,-0.2070959359,0.074329488,0.2397671491,-0.2371826023,-0.2322136909,-0.3541304469,0.2216895223,0.1782290488,0.6323610544,-0.3489941955,-0.124347426,-0.1607485861,-0.0539470986,0.1443855762,0.2540586889,0.1864249408,-0.1200918108,-0.1221161336,0.1319267005,-0.1353091449,0.0127926497,0.2098914981,0.230622232,0.424031496,0.0276736021,0.2996921539,0.1630720496,-0.1836214811,0.1571726948,-0.2461177707,0.0766194835,0.357293427,-0.3937429488,0.0347965136,0.0281328484,0.0014923022,0.108503826,-0.1232893765,-0.1755147129,0.1744697839,0.3144144118,-0.0094594676,0.1199689955,0.4047933221,0.0989428982,0.26471892,0.1185763478,-0.1872927994,-0.063955538,-0.0435254909,-0.0904999748,-0.4336491525,-0.262034744,0.2131782025,0.1339182705,-0.0305909198,-0.1540098786,0.056545902,-0.4272989929,-0.1620934904,-0.0612861179,0.4848382175,0.1439730823,0.500371933,0.2122715861,0.1486690044,-0.2858898342,-0.2623577416,-0.2568903267,0.4522441626,-0.481220603,0.1976672858,0.0383407511,0.2481758595,0.2800936699,-0.1724491417,0.1591964513,0.0334207118,0.0613640845,-0.1664860398,0.048756659,0.4103954434,0.0119138006,0.3502828777,0.1928022653,0.1393462718,-0.0928634256,0.2679367065,-0.3535985053,-0.3601426184,0.2299660593,0.0713830814,-0.3113717437,0.1920525134,-0.1264992952,-0.0647272021,0.1432967186,0.0446434766,0.2136843801,-0.0210933238,-0.1541174054,-0.3666602075,0.1856174916,0.777557373,-0.1078999266,-0.0855644569,0.033053983,-0.2180709094,0.1770857871,0.06840837,-0.07541731,0.0966386795,-0.271671325,0.0151542276,0.569583416,-0.4580412805,-0.1668783426,-0.0221467726,-0.0789949968,0.0365227982,0.189515695,0.1257303208,-0.2894612551,-0.104720287,0.117308259,0.1043076292,0.0383791775,0.2566961646,-0.4343810081,-0.1676717252,0.0616063438,0.2503466308,0.2828432024,-0.0925019979,0.0311437473,0.5866729021,0.1889612526,-0.1102685556,0.2881116867,0.4524781406,-0.0726473853,0.0778378174,0.16384238,-0.0598281808,-0.2577006221,0.29026106,-0.1085456833,0.2222910076,0.050758332,-0.0135211656,-0.0683132559,-0.2290741056,-0.297770381,-0.282966882,0.1336334944,0.0285771936,0.0880181268,0.1543253958,-0.2626301944,0.3814667165,-0.3416258693,0.0295177028,-0.5024129152,0.2508628964,-0.1954639256,-0.1462162435,0.0308585577,0.08732526,0.1917511374,-0.1371607929,-0.1381656677,0.391384244,0.2522928417,-0.0563739017,-0.0376479924,-0.0895878524,0.1325113177,-0.2032930702,0.3407712281,0.0584762096,0.1820220649,-0.1298588067,0.1848866343,0.1325737238,0.0625198483,0.3970299959,0.1156466007,-0.0693381056,-0.0677482262,0.1494114399,-0.1136081815,-0.3245390654,0.3509789705,0.332865864,0.1574705243,0.08859317,0.0377293639,-0.2971400619,0.6881403923,0.1426716894,0.1136762351,-0.0574944168,-0.4095008373,0.0693988726,0.1513948888,0.0344957672,0.4079182148,0.1365347356,-0.1149398386,0.1409748793,0.2153791785,-0.0987564772,0.1317031085,0.2377418131,0.4221460521,0.0527265966,-0.0896576345,-0.0112266522,-0.1930359602,-0.1728314012,0.0667530224,0.1590051949,-0.2549693286,-0.1209206656,-0.4197220504,-0.517444849,-0.275336206,0.1492746472,-0.3977880478,-0.245301798,-0.1070484668,-0.0726701394,-0.3224294484,0.3194853365,0.2790144384,-0.0490602814,0.0168626588,-0.3235235512,-0.1718972772,-0.2900437415,-0.2571320832,0.0216138382,0.3672423363,0.0608446002,0.2379075438,-0.2665831745,-0.1685305834,-0.1087961793,-0.1508057266,0.1511537284,0.0263009146,-0.1442512125,-0.1114368215,0.351028502,-0.1820258945,-0.1018859893,0.4775199294,-0.0571261868,-0.2166934162,0.2488492876,0.183147043,0.0790500864,-0.1002003998,-0.4466011226,-0.1296705604,-0.5109665394,-0.0740841255,0.0898636878,0.0974504128,0.1583665609,0.2631686926,0.1819608361,-0.1276183277,-0.0301727708,-0.1160964295,-0.0521410853,0.2722156942,-0.1465057135,-0.3368284702,0.2014806122,-0.1044794023,-0.0021723937,0.3633410931,-0.1813987643,-0.097296685,0.0007995882,-0.2872744203,-0.0532515272,0.3520984352,0.5604915023,-0.1533249021,-0.0316332802,-0.0331121609,-0.0858613476,-0.1202542633,0.0027140847,0.3534732461,0.0590062961,0.2731181085,0.1735912561,0.932516098,-0.0268989336,0.4621212482,0.4056243598,0.049927447,0.3116858602,-0.3958239853,-0.0405504853,0.2475835532,0.1420385987,-0.0782605857,0.1939635575,-0.0112836547,-0.0358982719,-0.5152168274,-0.3167847693,-0.2198868692,-0.3144631684,0.3542356193,-0.1514079422,0.0098828468,-0.0006797466,0.0473474637,0.0278676935,-0.0633758679,0.0390424877,0.2465269268,-0.0347133055,0.0417484902,-0.229319185,-0.1894801855,-0.3416495025,0.2945893109,0.0248188097,0.170450449,-0.0590169132,-0.3400983512,0.1011070609,0.0709243417,0.6148952246,-0.2257058322,-0.2467264682,0.286055088,0.0270744115,-0.6423119307,-0.1949949563,-0.0242505763,-0.1003393158,-0.1818844825,0.0819021016,-0.2344076037,-0.197913453,0.1702038646,0.347284168,-0.0462823659,-0.1886787266,-0.4585424066,-0.0359802246,-0.5773245692,-0.137450695,-0.0305477902,0.0302876681,0.0899485424,0.3163578212,0.1898748428,-0.1917277277,-0.0215156488,-0.1119429022,0.341362685,-0.1046372354,-0.171059981,0.1090289354,0.042470403,0.0303462986,0.2900733054,0.1259759367,-0.1707874537,0.0434146523,0.0294685159,0.0508783087,0.4066846967,-0.0721736923,-0.0411551557,-0.0437263362,-0.0693141371,-0.0208204892,0.0298812464,0.1085256115,-0.2762964964,-0.2065789551,-0.6122494936,0.6164750457,0.0779390559,-0.0070472327,0.4423122704,0.1935300082,-0.2312546819,0.5099320412,0.1472665817,1.1177163124,-0.1503802538,0.2815023363,0.1789483428,0.134201929,0.6970809102,-0.6101930141,0.3381558955,-0.1902341545,0.0593845583,0.0498276912,-0.1147465408,0.3421474397,0.074226439,0.0520732999,0.5547217727,0.0944094732,0.3872458637,-0.3301387131,-0.0382318906,-0.4087442458,-0.0995562449,-0.2862854898,0.0773014426,-0.1073559523,0.5356621742,-0.1105025336,0.0492262319,-0.3678701818,-0.2697418928,-0.2712926269,0.0676360056,-0.2752242088,0.2253467739,-0.438857466,-0.1411998272,-0.1835291237,0.2323127836,0.2649606168,-0.0146416826,-0.0779357925,0.2345309407,0.0545259602,0.0144431507,-0.161246255,0.0704640374,0.3199256659,-0.1493270397,-0.2023090422,-0.0072201653,-0.2528013885,-0.1998404562,-0.270863384,0.1244182736,0.5775037408,0.2529714406,-0.3201369047,0.3392543793,-0.0437237695,0.0853199586,0.1442815065,-0.1620114446,-0.0689978078,0.0958442241,-0.3166086376,-0.0318893231,-0.0409998931,0.4114470184,0.3425064385,-0.0451265834,0.3743166029,0.02680411,-0.1699328125,-0.1892459095,0.1815708727,0.2453127354,-0.2064424306,0.0405977741,0.053003855,-0.0656869635,-0.1135319322,-0.0082211802,-0.0316907018,0.0257291794,0.0861477256,-0.2623572648,-0.2688733339,0.4009700418,0.0285682399,0.0962442532,-0.1838034391,0.1682519615,0.2581397891,-0.2804543078,-0.2636615634,-0.1664469838,-0.1379230767,0.2131751329,0.1611237377,0.2940632999,0.0711482018,0.0260661133,0.0308838803,0.1700147092,-0.3656999767,-0.1506891847,0.026994843,0.1648257524,0.1837409586,0.0149642564,0.162931487,-0.474291265,0.1331102997,-0.3103115261,0.1451510936,0.3282083273,-0.1741094142,0.3330098391,0.3119806647,0.1528908312,-0.1263669133,0.2810007632,-0.208089307,0.1811099648,0.2060035169,-0.0202517901,0.3747199178,0.1296892762,-0.4093849957,-0.1814788878,-0.3265288174,-0.0267219115,-0.0540514179,-0.0679625869,-0.0941458941,-0.0720657632,0.1391068995,0.3085330129,-0.2720419168,-0.1431088448,0.2774803042,0.1624221057,-0.2192168087,0.0750611052,0.3696585298,-0.0732685104,-0.0315308087,0.1624495238,0.0274782535,-0.0455584191,0.1803496331,0.1916210055,0.2042515129,-0.2743170857,0.3263809383,0.123056151,0.2487036437,-0.0368460566,0.01430503,-0.0407124572,0.2321396321,0.6695385575,-0.1790863723,0.5875691175,0.0621428862,0.1620479375,-0.2189346105,-0.398702234,0.2998220921,0.6812435389,-0.0985095724,0.1004625559,-0.1580052674,-0.0540077649,0.1763891727,-0.0204280112,-0.0960498899,0.1124386564,-0.0607522465,0.1433696151,0.0214042813,-0.2690525055,-0.1828234345,0.1030614227,0.0882200152,-0.3828249276,0.3714802265,0.1974015683,-0.1855776161,-0.0804095343,0.3417043984,0.1561096311,0.0718711689,-0.2412338257,-0.1007751524,0.1273665428,-0.1581710428,0.1788203567,0.1164054349,0.2657682598,0.4134484529,-0.0676163509,-0.084783107,0.1543543786,-0.0182007272,-0.0186742675,0.1784266829,0.0545294285,0.1362746507,0.1742999554,0.0870502591,-0.2107060999,-0.1594396383,0.2782296836,0.0892288461,-0.2767755091,-0.0333555005,-0.3387450278,0.0231567286,-0.2729213536,0.0958562642,-0.6517592669,0.0046341377,-0.0973900557,0.1509311199,0.0231203865,-0.1643293202,0.0431653634,-0.1716599017,0.258797586,0.1433948725,-0.1367458552,-0.3192219436,-0.3952594399,-0.5910947919,0.0117549431,-0.3184987605,0.3399732411,-0.1448810101,0.2900011837,0.0270659588,0.1840905398,-0.0946335196,0.0659247637,-0.0728727132,0.1686778814,-0.4083973467,-0.1804000437,-0.1847619712,0.0844625756,0.0828011408,-0.3851663172,-0.0301346797,-0.0941679627,-0.0328182243,-0.0980993509,-0.2634514868,0.4284462035,-0.1124792695,0.5186240077,-0.1062076911,0.2381480336,0.0832460523,-0.1408236921,-0.263381958,-0.1576009691,-0.0788619891,-0.0940654576,0.0313772559,0.4429895878,-0.2951670289,-0.0622901022,-0.3243883848,0.4698706567,-0.2132890522,0.2390983701,-0.1712668687,0.151472047,0.1471274048,0.3593338728,-0.1664117426,0.0064460696,-0.0772218555,-0.0631920993,-0.5480507612,-0.2193656415,0.528793335,-0.4788553119,-0.2819333076,-0.0606862605,0.2528176606,-0.1767332703,-0.0653752163,-0.5018324256,0.1924413443,0.2456646562,0.1923528016,-0.002793866,0.0548757054,-0.2297174186,-0.0328225866,-0.0410186686,-0.364145726,-0.2055200636,-0.1534763277,0.1743489504,-0.2991899848]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1790","title":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.","comments":"The reshuffle is needed when you use parallelism.\r\nThe objective is to redistribute the articles evenly on the workers, since the `_extract_content` step generated many articles per file. By using reshuffle, we can split the processing of the articles of one file into several workers. Without reshuffle, all the articles of one file would be processed on the same worker that read the file, making the whole process take a very long time.","body":"```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?","comment_length":73,"text":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.\n```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?\nThe reshuffle is needed when you use parallelism.\r\nThe objective is to redistribute the articles evenly on the workers, since the `_extract_content` step generated many articles per file. By using reshuffle, we can split the processing of the articles of one file into several workers. Without reshuffle, all the articles of one file would be processed on the same worker that read the file, making the whole process take a very long time.","embeddings":[0.0401644371,0.0101211676,-0.0058726757,0.1293456107,-0.0119499862,0.0596711785,0.3661439419,0.3480103612,0.0229750033,0.279761821,0.1604572237,0.0736099333,0.1980750263,-0.195365712,-0.0845636874,-0.2648461461,0.3423892856,-0.0130630629,-0.0566421784,-0.3558188379,-0.4479026794,0.2000831217,-0.503382206,-0.1780140251,-0.1696120054,0.2718338072,-0.1433468461,0.3892425597,0.1860793829,-0.0491476916,-0.0400718972,-0.0870492756,-0.08698605,0.256269455,-0.0001181821,0.0990975425,0.2742292285,-0.2264608443,-0.2354054004,-0.1355381012,-0.0001665064,-0.5018975139,-0.0932022035,-0.4231704175,0.0232972745,-0.0400476493,0.0391367041,-0.2083500624,0.5846295357,0.1001808941,0.103046,0.2666604519,0.0476952866,-0.0251141619,0.320020169,-0.0474892519,-0.0405142456,0.1522981375,-0.002084804,0.2132861316,-0.1075008661,0.4607252777,-0.2558431923,0.1357323825,-0.0586365238,-0.1971606612,-0.1175447851,-0.4867197275,0.3078394234,0.2610002756,0.4405623972,0.0164707396,-0.1753842384,-0.2519320548,-0.1330917925,0.114490591,0.1225961372,0.1937907636,0.1123032272,0.0998713523,0.571342051,-0.3144812286,0.0563262366,0.0551692396,0.2326830029,0.5872220397,0.110598579,0.4362630248,0.0444842242,0.1545606107,0.2064007372,-0.2861570418,0.2010479569,0.2365494519,-0.2430971265,-0.0052309814,-0.037382368,-0.2237178087,-0.0731735379,-0.1004010513,-0.5007820725,0.2464473844,0.4761475623,0.1704075336,0.1579053849,0.1818657815,0.1793934554,0.219405219,0.2643758059,-0.2684376836,-0.0538066588,-0.0212680344,0.133273229,-0.3763225377,-0.4731953144,0.2896768451,0.0331057906,-0.1172800213,-0.1124831215,-0.2394607365,-0.2518795133,-0.3447423279,-0.2870922685,0.2796641588,0.3040516078,0.5849859118,0.0309539326,0.2881352603,-0.5063956976,-0.1919756085,-0.1429306567,0.1993660778,-0.2677182257,0.1640225947,-0.0752119794,0.1458439678,0.1551924944,-0.1586454362,0.2210362107,0.0329928398,0.0926920846,-0.0229866132,0.029406013,0.4956110716,0.1296814382,0.3343687356,0.381850183,0.0441212393,-0.3361369967,0.2725721598,-0.5118633509,-0.4985116124,0.1436737925,0.0310975295,-0.1891684532,0.1958587915,-0.0187793169,-0.0286519136,0.286187619,-0.1459286064,0.2613285482,-0.0038731967,-0.0005569244,-0.3166024387,-0.042889379,0.637907207,-0.1172443628,0.0082371486,-0.0107365968,-0.1041538864,0.2127786577,0.1939695179,-0.1101893634,0.0763258785,-0.2047847956,0.0210151263,0.5991586447,-0.5733520985,-0.1203276813,0.0884790793,-0.2460712343,0.1072986573,0.4706140757,-0.0097041624,-0.0795367435,-0.2263072878,0.0292198919,0.1647599488,0.0946116149,0.1012687683,-0.4208419323,0.0282597113,0.1501709223,0.1666884273,0.193199411,0.1797627509,0.0612409264,0.119194299,0.2179642171,-0.030802751,0.2628243566,0.2563932836,-0.1955601871,0.242650196,0.3924926221,0.2315067053,-0.0732826814,0.281163007,-0.1374687999,0.3438961208,0.0679792091,-0.1475525796,-0.0213426966,-0.4191904664,-0.3652015626,-0.5023626685,-0.0390964411,-0.0010685757,0.2320947945,0.2806372344,-0.4111865163,0.360601604,-0.3701222539,0.0576596595,-0.5175859332,0.2484353781,-0.1954278946,-0.0150761884,-0.0902091339,0.116037406,-0.0062022605,-0.241061911,-0.0278798696,0.3686268628,0.2898938656,0.1484854817,0.1862214357,-0.227882266,0.1058301106,-0.2935655117,0.3605100811,0.216786176,0.1143658459,-0.1836787909,0.2624615133,0.3334526718,0.0331231393,0.4286051691,0.0929735154,-0.0501370542,0.1583972275,-0.0014180471,-0.1741600186,-0.1964772195,0.2689024508,0.2168021053,0.0400736295,-0.084064737,0.0163175445,-0.1216809228,0.6155574322,0.0400267616,0.1203417853,-0.0023303851,-0.2098282576,-0.0071058688,0.0358044952,0.0477509946,0.4985297322,0.1664860547,0.0284510814,0.0910320133,0.3129204214,-0.2214421332,-0.0211745407,0.1403507143,0.3321332037,0.22385405,-0.2455611825,0.0303164311,-0.3016816676,-0.1041558981,-0.1987624466,0.0637186244,-0.0474325418,-0.0213095788,-0.0229489952,-0.3624950647,-0.4416368902,0.055863332,-0.3911738098,-0.3372012675,0.0788687915,-0.0377292261,-0.2612815797,0.3961455524,0.3479470015,-0.0649439842,-0.0434287265,-0.126715824,-0.2274429649,-0.3022428155,-0.2869898379,-0.041445639,0.1613826901,0.2441117764,0.050085023,-0.3857806027,-0.1624459624,-0.172975868,-0.1353340149,0.0865205452,-0.0582750365,-0.1021668613,-0.1136889085,0.3374117315,-0.0435153618,-0.1603801996,0.362831533,-0.0497339293,-0.1989869773,0.2759115398,0.100740537,0.0926558822,0.0612153411,-0.6890731454,-0.3914910555,-0.4526816308,0.0283048656,0.1488553137,-0.002345246,-0.1121013537,0.1457041651,-0.186211586,0.0437317267,-0.1842689663,-0.141288355,-0.1482243389,0.1839481294,-0.1587505788,-0.317183584,0.1166820228,-0.1749303043,-0.1934995651,0.1352559626,-0.0955842957,-0.0055908985,0.1183072999,-0.2816363573,-0.1481215507,0.6191095114,0.6890797019,-0.1661870927,0.0601065345,-0.0240419414,0.1223906353,-0.0701789036,0.0990803391,0.3527521193,-0.0878814906,0.0569446497,0.0799825862,0.8017918468,0.1350184232,0.621397078,0.2583411932,0.1399969459,0.2320947647,-0.4885397255,-0.0433358736,0.1662050486,0.0175423678,-0.1670777053,0.2452231348,0.0493522137,0.0112376139,-0.4611929953,-0.4745738208,-0.3980816901,-0.2800992727,0.2716125548,-0.0590544119,0.1806712449,0.0098430039,-0.1435567737,0.1308434606,0.1081400886,0.0183616579,0.5039035082,0.0818250254,0.0498275235,-0.3906681538,-0.2206053883,-0.2547490001,0.390404582,0.1038838923,0.0071674245,0.2192421257,-0.4376813173,0.1756004691,0.0366980098,0.7510885,-0.2891374826,-0.2462833226,0.121405229,0.0734254345,-0.4148050249,-0.2408629507,-0.0132125197,0.001745541,-0.0753208622,0.1231319085,-0.3026685119,-0.0831562728,0.1161784977,0.280875057,-0.2045972496,-0.2897413671,-0.6348304152,-0.1960388869,-0.7667278051,-0.1009179726,-0.2245741785,-0.0269016363,-0.0457418784,0.2402990162,0.266446501,-0.1687164158,-0.1306377798,0.0495673865,0.0420691334,-0.2761183977,-0.3636623025,0.1034521908,-0.1921042055,0.2361412793,0.2059913725,-0.0251834597,-0.1817453057,-0.0661988035,-0.153883338,0.2551400065,0.3604367673,-0.0031678542,0.0669379383,-0.0211364254,-0.0070202849,-0.0431657843,-0.4192327559,-0.00459876,-0.0795827061,-0.2701924741,-0.5315517187,0.4778116047,-0.1089426801,-0.0097237788,0.2204697132,0.1652796119,-0.0740981549,0.5455192327,-0.0370404609,1.1979786158,0.0697285309,0.2619315386,0.0200011637,0.067367278,0.8408936858,-0.7155609131,0.3852576315,-0.178808704,0.0753538087,-0.1027157903,-0.0964369923,0.5330148935,0.1523684859,-0.111549437,0.6340992451,0.089798063,0.1107655615,-0.382946223,0.0491497554,-0.5872061253,-0.0020174573,-0.0955374986,-0.1017634273,0.1078922078,0.5783392191,-0.1541689187,0.1405357271,-0.4126497507,-0.4174630344,-0.0435713977,-0.0402046219,0.1563469321,0.3821920156,-0.2015430331,-0.2772367597,-0.2275538296,0.0341191031,0.3032435179,-0.245038569,0.1195626408,0.0719767287,0.1027842984,-0.1334927976,-0.085118942,0.0266555585,0.3381581008,-0.1617871374,-0.2683914006,-0.1418526024,-0.2589023113,-0.2996107638,-0.1787232608,0.2115920186,0.3893196583,0.158839643,-0.3468773067,0.6813212633,-0.1538217813,-0.0783033818,0.0517501049,-0.0718403906,-0.3309112787,0.2146337479,-0.4696040452,-0.0516263172,-0.1445708573,0.2845223546,0.5598363876,-0.0852776393,0.2096586376,0.0481222793,-0.1559476703,-0.1406456381,0.206129387,0.0663564727,-0.1167549863,-0.0088436073,-0.0003366791,-0.0602786765,-0.1183355227,0.1223803014,0.0960986316,0.17564556,-0.1687562019,-0.508020997,0.0360033661,0.368311435,0.0885486901,-0.1269354671,-0.0267042518,0.2383686155,0.3466283381,-0.0447961353,-0.1975692064,-0.0443151034,-0.0991519094,0.3087215126,0.1973850578,0.1657436341,0.0202005971,0.0983670279,-0.0506111123,0.082061328,-0.2794934213,-0.089670755,0.1034569815,0.2129496187,0.1301282644,0.083939448,0.104017131,-0.2892776728,0.1351737231,-0.1096886471,0.0343006812,0.1533063352,-0.2824153006,0.4188383818,0.3241230249,0.2274622321,-0.415766716,0.302682668,-0.3377509117,0.0373846926,0.0953711569,0.1226106957,0.3036298752,0.1579971761,-0.2214698792,-0.112548694,-0.1013616771,0.0859857798,0.1019834429,-0.0053940611,0.0509434268,-0.3437861204,0.2998645008,0.5285134912,-0.000493506,-0.1901673526,0.1603437811,0.0536319129,-0.3772275746,-0.000357947,0.6144313812,-0.116039589,-0.0262058284,0.2547378242,0.2971385717,-0.1633602828,0.1214412823,0.1717962325,0.3075534701,-0.3022236824,0.3114132285,0.2198165208,0.2022478729,0.1023765653,0.0438005477,-0.0593718588,0.2315729856,0.4631899297,0.0525063202,0.7212814689,0.2852931917,0.1726116985,-0.264349848,-0.3424153924,0.3993784487,0.4465332925,0.044677157,-0.0783046857,0.0100714341,-0.1394176036,0.0058779898,0.1516019553,0.0343411416,0.2824628353,-0.0321251601,0.1338588893,0.1411106735,-0.2624067366,0.018719662,-0.0741360411,0.1088449135,-0.0734167099,0.4331256151,0.122741051,-0.0267004836,0.1284412593,0.2447285354,-0.2824225724,0.1447903812,-0.0760190189,-0.2901899219,0.2440315485,0.0011969076,0.164797768,0.2264438421,0.2207713276,0.1477451026,-0.1869587749,-0.0110664936,0.1191351935,-0.0742229894,-0.0400303714,0.2172244042,-0.0867466703,0.1863644123,0.0443030559,0.0615513325,-0.1333114207,-0.0444112718,0.1823524535,-0.1542583108,-0.1811796129,0.0117924977,-0.1430220455,-0.1555945575,-0.1952771395,0.2217181474,-0.4802034199,-0.0314695276,-0.0326493792,-0.1143594906,-0.1712022573,-0.1373384446,0.0024435122,-0.1196265221,0.2270769924,0.4151248038,0.0142982211,-0.1361878365,-0.3554025292,-0.3997882903,0.1810676903,-0.4627872705,0.1659277678,-0.1951676309,0.3386707902,0.1467349976,0.350442946,-0.0512540601,0.1434316486,-0.2581034303,0.2402316779,-0.1519417316,-0.1650690436,-0.1928881258,-0.0824905112,-0.1049787477,-0.3492977023,-0.0071600461,0.2173494846,-0.0368054211,0.0735219568,-0.2756356597,0.5410722494,-0.1676594466,0.3580463231,0.1330910474,0.2596685588,0.0470487215,-0.087751314,0.2461746931,-0.0556871593,-0.2559540868,0.1827918738,-0.012111621,0.21329768,-0.4219235778,-0.0811932683,-0.214940697,0.4899175167,-0.024941016,0.1334857494,-0.332462728,0.1876621097,0.1471927613,0.132583335,-0.0528517589,0.3269251883,0.0415285192,-0.1088424474,-0.4738565385,-0.0035289603,0.2111254334,-0.4024795592,-0.4467402399,-0.0865475014,0.2853616774,-0.1469683051,0.0271623265,-0.4059700668,0.0087410221,0.2635793686,0.1890694648,0.0547970571,0.0043810862,-0.1645897776,-0.367469579,0.0327628627,-0.4655790627,-0.1391334534,-0.1570225656,0.0946325362,-0.4148332179]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1790","title":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.","comments":"Maybe the reshuffle step can be added only if the runner is not a DirectRunner ?","body":"```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?","comment_length":16,"text":"ModuleNotFoundError: No module named 'apache_beam', when specific languages.\n```py\r\nimport datasets\r\nwiki = datasets.load_dataset('wikipedia', '20200501.ja', cache_dir='.\/datasets')\r\n```\r\nthen `ModuleNotFoundError: No module named 'apache_beam'` happend.\r\n\r\nThe error doesn't appear when it's '20200501.en'.\r\nI don't know Apache Beam, but according to #498 it isn't necessary when it's saved to local. is it correct?\nMaybe the reshuffle step can be added only if the runner is not a DirectRunner ?","embeddings":[-0.0378297865,0.0517897308,0.045922339,0.1618950516,0.2217684835,0.0495494045,0.2877710462,0.3032413721,0.0259332228,0.179933846,0.2910352945,-0.0192492269,0.1045147106,-0.0232278313,0.1100320294,-0.3740752339,0.3071031868,-0.0370492227,-0.0959916338,-0.281830281,-0.442635417,0.1054399982,-0.379992038,-0.1579352915,-0.0031889877,0.2331445068,-0.05683605,0.2654174268,0.2003732473,-0.1437357068,-0.0345408544,-0.2057222873,-0.0999974385,0.3504476547,-0.000123889,0.2205034047,0.3199234605,-0.293764621,-0.2419773638,-0.1562591195,0.0231217556,-0.3215471506,-0.1388468146,-0.3631034791,-0.101666145,-0.0253451038,0.0665042177,-0.3819905818,0.6048766375,0.1928147525,0.1098594815,0.4210021794,-0.020772513,-0.0147780627,0.3787573278,-0.0865190327,-0.1000084653,-0.0779461712,-0.2010744512,0.2293732613,-0.171431005,0.3713362515,-0.2388273776,0.0087697785,0.0032124519,-0.2786224782,-0.0851293057,-0.4114531279,0.1897355914,0.2687947154,0.4893785715,-0.2717169523,-0.2248738855,-0.2171139568,0.0023349314,-0.0665183142,0.2809661031,0.0015106902,0.0660388917,0.0086567244,0.5118035674,-0.3065735996,0.1025713012,-0.0175952259,-0.1090355292,0.6667248011,0.2183995992,0.357163012,-0.0904260501,0.1953649521,0.332996577,-0.2005532831,0.0652890056,0.1246236116,-0.3468623161,-0.0545837805,0.0373448953,0.1232199669,-0.1677461118,-0.009198511,-0.2472332567,0.1628329456,0.3050271571,0.063653037,0.0405441672,0.2389967591,0.1641890854,0.3014983833,0.2608810365,-0.3187677264,0.1340390891,0.0162616652,0.1202592701,-0.5691347718,-0.5156914592,0.3740118742,0.0190277621,-0.0261211637,-0.1588702649,-0.0768229663,-0.3505404294,-0.2740905285,-0.0533224158,0.3207163513,0.1593187302,0.3862256706,0.137473613,0.304372102,-0.2998504937,-0.1806768328,-0.0933372676,0.1023257375,-0.2912924588,0.0453901216,-0.0807791501,0.0810583979,0.1793289036,-0.1545542479,0.1975855678,0.2324463874,-0.0161839221,0.0808045417,-0.0513106994,0.4848632216,0.2038983852,0.3424087167,0.4302805364,0.1734014153,-0.2173183113,0.3705198169,-0.4540519714,-0.3109500706,0.1928400844,-0.0105659207,-0.1869196445,0.0678293332,-0.0408775695,-0.0447849706,0.1960520595,-0.1627370268,0.2673732936,-0.1260741204,-0.1356477141,-0.2988694012,0.0402593277,0.6308399439,-0.1694732308,-0.0710991621,0.0578628331,-0.0961322412,0.1474748254,-0.0768611059,-0.0319539048,0.2058146298,-0.2901695669,-0.0351816416,0.9126730561,-0.5441179276,0.046596624,-0.0909166038,-0.3012412786,0.0395175368,0.5204709768,-0.0860053599,-0.2685622573,-0.2502494156,-0.0669794902,0.280220747,0.0710942894,0.1019854546,-0.2831763625,0.1158957928,0.1996939331,0.2040532976,0.347771883,0.1337680519,-0.0263225343,0.3089297712,0.1329003274,-0.043742165,0.210990116,0.3810606003,-0.3102937639,0.0953511894,0.3288868964,0.1389881968,-0.0363512002,0.273640424,-0.2340336293,0.1261845827,0.2200858444,-0.0863326341,-0.0491343476,-0.3538447917,-0.447189182,-0.1766669452,-0.017715564,0.0479140356,0.100673534,0.2592098713,-0.4386266768,0.3504111767,-0.2382588834,0.1331081688,-0.3313281834,0.3779862523,-0.1291907132,-0.0094744954,-0.0357464924,0.1551730484,0.1023946106,-0.2608760893,-0.0117665604,0.3821970522,0.1847783774,0.0009458666,0.0661202073,-0.2647804916,0.1108395979,-0.0088176252,0.3106170297,0.2222774923,-0.0423321202,-0.1995381862,0.1582317352,0.38154006,-0.0263559707,0.5533995628,0.0767531693,0.0388492607,0.0933386534,0.0327161737,-0.1753463149,-0.0997213349,0.3674673438,0.2749312222,0.1782632768,-0.1733285636,-0.0387798697,-0.1163836643,0.6952420473,0.1830939502,0.1338067204,0.0048368541,-0.2690501511,0.1514980644,-0.0059789564,0.0746214315,0.3360258937,0.202152282,0.0381283127,0.1766224205,0.2083055079,-0.1654423028,0.0427262001,0.1187444925,0.3286023736,0.3047626317,-0.2040672302,-0.0914360657,-0.3562949598,-0.0728420764,-0.0272835717,0.0737777576,-0.199899435,-0.0119142653,-0.2538814247,-0.4885871112,-0.4647909403,-0.0220778082,-0.5278829932,-0.3266161382,0.1899416745,0.0178199653,-0.1534126103,0.5467302203,0.3333817422,-0.1433529854,-0.0060545877,-0.419012785,-0.2095358819,-0.26168859,-0.2897976339,0.0098996479,0.3442018926,-0.0150851095,0.064774394,-0.3819509745,-0.0901692286,-0.1610023528,-0.2000038922,0.0520419963,-0.0357182696,-0.3044560552,-0.0490274988,0.3831196725,-0.1667331606,0.0269654505,0.3886791468,0.1256408095,-0.2855384648,0.2123602331,0.1414213926,0.0906042233,0.0217540469,-0.5343981385,-0.2619627714,-0.4425385892,-0.1127432287,0.3377592862,-0.0179360975,-0.2148833126,0.2542978227,-0.177093938,0.0141038094,-0.2195478529,-0.0476564765,-0.0188096277,0.2171597034,-0.0823240355,-0.2377469391,0.1454093605,-0.1483948231,-0.0491365977,0.1627832651,-0.1970965117,-0.2041375041,0.2359644622,-0.2119003385,-0.155411616,0.6291600466,0.791995585,-0.058462251,0.0516695939,-0.0522565953,-0.0450738333,-0.1296370327,-0.0457174741,0.4874377847,-0.0442974418,0.3777843416,-0.0016312762,0.8330232501,-0.0795279145,0.5151055455,0.2486043721,0.0314508714,0.357662946,-0.5895619988,0.0750974342,0.1400813758,0.1810805351,-0.0809005871,0.1826249808,0.0848843977,0.2279676944,-0.425232619,-0.2832069993,-0.1590494961,-0.3067361712,0.24245435,-0.2019101977,0.0492009707,-0.020922564,-0.0076369033,-0.045968309,0.0837611482,0.130399406,0.4588710666,-0.0162471272,0.114908047,-0.3714070916,-0.1982638091,-0.2697690427,0.4062987864,-0.0375498645,0.0089863213,-0.0109702451,-0.3440268636,0.0577891693,0.0095753074,0.7247846723,-0.2223731875,-0.2159993798,0.1361850649,0.0658366978,-0.3969592452,-0.2671532929,0.1685899347,0.1104610339,-0.1129765734,-0.0603253059,-0.1990754157,0.0028695238,0.2013628483,0.350261122,-0.1858436465,-0.2565714121,-0.6710650921,-0.1283436716,-0.6659097075,0.0345693268,-0.0235290602,-0.0380815119,0.1507719308,0.1970325112,0.1336765289,-0.410520494,0.0060413284,-0.033990629,0.1067852378,-0.1963941306,-0.393101573,0.083872512,0.1498659104,0.0119267842,0.1818652898,0.0489210226,-0.1927951723,-0.0224620234,-0.0940004811,0.3057221174,0.3254758418,-0.1210918948,0.2026845217,-0.0988028497,-0.1369830817,0.0581120253,-0.2772009969,0.0825003013,0.0212357324,-0.2918603718,-0.5804404616,0.5959303379,0.0238303598,-0.0611400865,0.4912584722,0.2390518636,-0.210700199,0.5178577304,-0.1179182157,1.1283652782,-0.0416321978,0.2160153836,0.1365785897,-0.0306075495,0.652921617,-0.7152821422,0.3328053057,-0.24997136,0.1918945611,-0.10732124,-0.1381448805,0.4551345408,0.0230791327,-0.0000857469,0.6919450164,-0.069671452,0.2239047289,-0.2821063995,-0.1195758432,-0.2298959792,0.0141806696,-0.2873310745,-0.1048291326,0.0490019768,0.5750827193,-0.2609502673,0.1015172675,-0.2526404262,-0.34399876,-0.0964746773,-0.0538978018,-0.1134000123,0.100381121,-0.250674516,-0.3562538922,-0.1733186841,0.1902553439,0.4215758145,-0.1046226844,0.063127324,0.2058884799,0.0568377487,0.001773979,-0.1125627905,0.1085177585,0.2183764726,-0.1717536151,-0.1444943398,-0.0625027716,-0.2645857334,-0.3730089068,-0.1124175638,0.0927063972,0.7888134718,0.2436784506,-0.1366640627,0.5964682698,0.0447729491,-0.0597375147,0.0402124971,-0.1973062307,-0.052092988,0.0244716369,-0.3202798963,-0.0879467949,-0.1904217601,0.3382103443,0.5054776073,-0.109008871,0.3684285581,0.0020968537,-0.064161323,-0.102273263,0.1413153112,0.0933740586,-0.1211449206,0.0321340822,0.1755928397,-0.1490768194,-0.1830691099,-0.0648159608,0.0359758548,0.1165765226,-0.1092981473,-0.3471044898,0.0097767832,0.1867243797,0.0818004161,-0.07499028,-0.1192834303,0.4146800935,0.311011225,-0.2838917971,-0.1570914984,-0.1663307399,0.0561406016,0.1389807016,0.0468083285,0.3287314475,-0.0659128353,0.2154006511,0.001971619,-0.0859812647,-0.2071032673,-0.0754699558,0.0507456698,0.2360820174,0.1155925915,0.1225415468,0.2319329083,-0.2529885769,0.1190035641,-0.1333070546,-0.0471670255,0.2049925625,-0.1743321568,0.2479425073,0.4722281396,0.1558709294,-0.4540858865,0.1151994616,-0.2590802908,-0.0721191168,0.017685499,0.2242804915,0.2557978332,0.1642473638,-0.3291210234,-0.0661544353,-0.0070633874,0.0523623414,0.0173699949,-0.1378448904,0.0108228121,-0.2007793486,0.2597387433,0.4037540257,-0.1165418476,-0.1918813735,0.2356770188,0.0097520165,-0.614834547,0.1283179522,0.5827642679,-0.043628905,0.0342312232,0.2916868329,0.2184493095,-0.2475515753,0.3061931431,0.1987028867,0.1265827417,-0.3680979013,0.395807445,0.2837421894,0.1989347488,0.1737141907,0.3414800465,-0.1861775964,0.1620644778,0.6131902933,-0.0779114142,0.5723170638,0.2195226252,-0.0528066009,-0.3409297168,-0.344013989,0.3430022597,0.4752599597,0.2266085297,-0.2065160871,0.154410556,-0.1298751384,-0.0306211822,-0.0297720898,-0.1927209347,0.0874814987,-0.1007989049,0.0834336877,0.0292317718,-0.3153875768,0.0002237243,-0.0326267257,0.1068402007,-0.0736549944,0.2326966524,-0.0119542265,-0.0000042149,0.0383541435,0.2689740956,-0.250752151,0.0448512658,-0.0744486004,-0.1020018831,0.2303335369,-0.0982782319,0.2454404235,0.1394628733,0.0990584791,0.0319892727,-0.0642367527,-0.1891172379,0.0995636955,0.0698037371,-0.0910216719,0.0897794738,-0.1794643104,0.1652263999,0.1663865745,0.0541426167,-0.1633561105,0.1356814355,0.0526213497,-0.1210123897,-0.1623201668,0.2607012093,-0.1448886096,-0.0855366588,-0.2999683022,0.1323325634,-0.4391359985,-0.03618012,0.069201678,0.1096901745,-0.208939001,-0.2250740528,-0.0050088889,0.0577385947,0.1765886992,0.3686086833,-0.1502882838,-0.1989369392,-0.2875460386,-0.4894713163,0.1540007144,-0.4054336548,0.2183384448,-0.2065917253,0.2182586938,0.1433670074,0.2793420255,0.1631580293,-0.0037899173,-0.1784085184,0.1113899723,-0.2246072441,-0.1581054032,-0.1996611059,0.0984460115,-0.1441681385,-0.3681927919,-0.0305782482,0.0476009324,-0.0159632973,0.0230947323,-0.2837213576,0.5991744995,-0.2503395379,0.2883057296,0.0000726791,0.2158013433,0.1250244975,-0.0017143375,0.150681898,-0.1211901009,-0.4373408556,0.1146835089,-0.0757653266,0.2590178549,-0.3916987479,-0.1559026092,-0.3126984537,0.4246472716,0.0334188342,0.3534763157,-0.2941413224,0.2738617957,0.1248201802,0.2421265095,-0.0782948658,0.24047333,0.0372114256,-0.0417716801,-0.5601404905,-0.090134725,0.2070991248,-0.4825275242,-0.3584974408,-0.0474843383,0.1790843904,-0.0680190101,0.0958560705,-0.4054121673,0.1350020915,0.3061630428,0.0924020559,0.1401241273,0.0371975787,-0.4364372492,-0.1897845715,-0.0181784499,-0.6344162822,-0.1510044038,-0.1968302131,0.1020775661,-0.3524161577]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1786","title":"How to use split dataset ","comments":"By default, all 3 splits will be loaded if you run the following:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"lambada\")\r\nprint(dataset[\"train\"])\r\nprint(dataset[\"valid\"])\r\n\r\n```\r\n\r\nIf you wanted to do load this manually, you could do this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndata_files = {\r\n \"train\": \"data\/lambada\/train.txt\",\r\n \"valid\": \"data\/lambada\/valid.txt\",\r\n \"test\": \"data\/lambada\/test.txt\",\r\n}\r\nds = load_dataset(\"text\", data_files=data_files)\r\n```","body":"![Capture1](https:\/\/user-images.githubusercontent.com\/78090287\/106057436-cb6a1f00-6111-11eb-8c9c-3658065b1fdf.PNG)\r\n\r\nHey,\r\nI want to split the lambada dataset into corpus, test, train and valid txt files (like penn treebank) but I am not able to achieve this. What I am doing is, executing the lambada.py file in my project but its not giving desired results. Any help will be appreciated!","comment_length":56,"text":"How to use split dataset \n![Capture1](https:\/\/user-images.githubusercontent.com\/78090287\/106057436-cb6a1f00-6111-11eb-8c9c-3658065b1fdf.PNG)\r\n\r\nHey,\r\nI want to split the lambada dataset into corpus, test, train and valid txt files (like penn treebank) but I am not able to achieve this. What I am doing is, executing the lambada.py file in my project but its not giving desired results. Any help will be appreciated!\nBy default, all 3 splits will be loaded if you run the following:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"lambada\")\r\nprint(dataset[\"train\"])\r\nprint(dataset[\"valid\"])\r\n\r\n```\r\n\r\nIf you wanted to do load this manually, you could do this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndata_files = {\r\n \"train\": \"data\/lambada\/train.txt\",\r\n \"valid\": \"data\/lambada\/valid.txt\",\r\n \"test\": \"data\/lambada\/test.txt\",\r\n}\r\nds = load_dataset(\"text\", data_files=data_files)\r\n```","embeddings":[-0.3074781597,-0.046431113,-0.0625262633,0.4651071429,0.1305702031,0.2037909329,0.2765963376,0.5807616711,-0.4295654595,0.0150459204,-0.3925747275,0.254095912,0.1838493347,0.3340367377,-0.1944946945,-0.4850153923,0.0906330496,0.2447201163,0.0513657033,-0.1060134172,-0.07859727,0.1636883616,-0.4311268926,-0.1009087116,-0.5181128979,-0.1812312454,-0.2913366556,0.1126133949,-0.3164643347,0.0334359258,0.0643469393,0.1915279776,0.2770765722,0.4318369627,-0.00012366,0.0267235059,-0.0548695512,-0.251575917,-0.145966813,-0.0903863758,0.0458019525,0.0903777853,0.1970305294,-0.3328788877,-0.1814205348,-0.4163512588,-0.0681711361,0.0015483622,0.6935947537,0.4342878461,0.0081108157,-0.4164196253,-0.005543787,0.1647204608,0.367272526,0.3402585387,-0.0229089577,-0.0657090917,0.1916342378,0.0812231302,0.0823046118,0.0832441226,-0.0158376154,0.3440508246,0.0033670156,0.185282737,-0.2075952888,-0.4350948334,0.0201418605,0.1973951459,0.6681491733,-0.3729459643,-0.1618387252,-0.2038089186,-0.1345122904,-0.4180262387,-0.0779000297,0.4423206747,-0.1205714867,0.1378150731,-0.2383987606,0.1256262809,-0.2445560098,0.3950924575,-0.0653971806,0.2446008474,-0.0851913691,0.1580072045,0.1469615698,0.2207309157,0.179539606,-0.238950327,-0.1969656497,0.0609343536,-0.33901757,-0.1809932292,-0.4194630682,-0.2334724963,0.2693914771,0.4270803034,-0.1869285405,-0.1305606067,-0.1368005723,-0.1355264187,0.511703968,0.0301423278,0.2042330801,0.3433418572,0.0389579013,-0.0347348452,-0.2185376137,-0.1680016369,-0.1378880292,-0.2395493239,-0.0726651475,0.3465918005,0.2915106714,-0.2763409317,-0.1029529944,-0.5062012672,-0.5687624812,-0.2831332684,0.0132070417,0.226178661,0.0955337211,-0.087401852,-0.0726813674,0.3877472281,-0.3162638843,-0.1819904596,0.0231045298,0.0718567744,-0.2617369294,-0.0702702254,0.3142620325,-0.4424532652,0.1238225624,-0.0749288872,0.0452285521,-0.3328416646,0.1651387364,-0.3937927485,0.0545788966,0.2088656276,0.2188968211,-0.0664284006,0.1271425337,0.1447395831,-0.2334046513,0.104900308,-0.1808667034,-0.0801759586,0.1558273733,-0.0196957737,0.0611532554,0.1351347417,-0.4866174161,0.2976647019,0.0896814466,0.1010288969,0.0367785692,0.1594928354,-0.1407074034,-0.0847328827,0.2250472009,0.134390071,-0.2925608456,0.0847757012,0.0228861179,-0.2819042206,0.2453172356,0.110860236,-0.1238232777,0.5956167579,-0.03305142,0.0815179124,0.4777341783,-0.7511567473,-0.1438726634,0.375092417,0.2684719861,-0.267881453,0.0732861385,-0.1370616108,0.5824891329,-0.0600705929,-0.1912472248,0.5537322164,-0.2710953951,0.0165227335,0.0668279082,-0.2530251443,-0.0888290927,0.2614716589,0.0190685932,-0.0778126791,0.2262897342,0.1465780139,0.366556555,-0.1042811647,0.308786273,-0.0434692241,0.242729038,0.0071516386,0.1786873341,-0.3716312647,-0.2358853668,0.0932078063,0.0062300847,-0.0141476244,0.0756585151,-0.2274846584,-0.2097719461,-0.3526072502,-0.378256321,-0.1538071483,0.0048068264,-0.196291253,0.2621349692,0.0109238541,-0.2893612683,0.0339712389,0.0975752473,0.4323088825,0.0469637886,0.4479081035,0.1255497336,-0.0886189491,0.0299216919,0.0520881601,0.0634901151,-0.1969065517,-0.1304466724,0.4200092852,0.387339741,-0.0858752206,-0.0276352838,-0.2635817528,0.2716654539,-0.2346802205,0.0363780484,-0.0131422896,0.0668626577,-0.2364047766,-0.3931208253,0.151324302,-0.2954005301,0.2507550716,-0.0045375316,0.1367302239,0.1403153241,-0.2662443519,0.1603447199,0.0253539756,-0.1090277135,-0.2627272606,0.2647825778,-0.3021570444,0.1317908764,-0.0627249852,0.4427015781,-0.1344834119,-0.0271012913,0.0019468812,-0.014633242,0.3422632217,0.1277846545,0.4383707643,0.360081315,0.1405371428,0.116945684,0.1188058928,0.2640932798,-0.0330177024,0.1776481122,0.1477965117,0.213997826,0.1553716511,-0.2412400693,-0.1930330247,-0.0302048922,-0.0356371067,0.2965058386,0.3059887588,-0.088938944,0.1264159381,-0.2307716161,-0.4963020682,-0.3396852314,-0.3279790878,0.1289585531,0.0233278181,0.0331730172,0.0473307297,0.2906074822,0.1235113218,0.1898518354,0.0364795402,-0.0740121007,0.0600012466,-0.0685784966,-0.1514945626,-0.1349993348,0.0668269694,0.3735645115,0.7918115854,0.2371259183,-0.318859309,0.0106522609,-0.2567154169,0.0796766654,0.2232891917,-0.3098639846,0.049828466,0.1674468517,-0.0184569322,0.0096676741,-0.099256672,0.1296456605,0.0814134553,-0.1913520694,0.007976234,-0.0498731807,0.0759786069,-0.0945540741,-0.5739889741,-0.1470645517,-0.2587658465,0.4122140408,0.238113448,0.1185838729,-0.0971847847,0.0823044553,0.0910289809,0.0020560918,0.0315576866,0.1899357438,-0.1119549945,0.1132454425,-0.2350589633,-0.2499632984,0.3675474226,0.1742352545,-0.057356324,-0.0786823556,-0.2121851593,0.2335443199,-0.0393019952,-0.0227115043,0.0712002218,0.1150887311,0.2429207265,-0.2319124639,0.1688636541,-0.090231359,0.0654569939,0.1491007954,-0.2139852494,0.2626654208,-0.1342073679,0.3909491599,0.0201898441,0.6240466833,0.4335097671,-0.0120473206,0.0362733565,-0.1909749806,-0.3649744689,-0.078871794,-0.2182035893,0.0700497329,-0.1186102927,0.0122603374,0.4296044111,0.2938132882,-0.0210730266,-0.3732335865,-0.3243964314,-0.3882584274,-0.3626786768,0.0344237275,-0.1725732386,-0.0155389681,-0.2830408812,0.0333824418,0.2826636434,-0.1496884376,-0.3008393049,0.2882889211,0.0686449334,-0.0268583652,-0.4724097848,0.0764361396,-0.2553853989,0.3712052405,-0.0017333549,-0.164752692,-0.0501848944,0.1439056993,0.3047733903,0.0562230498,0.6318733692,0.0120184366,-0.0175541788,0.0538035855,-0.0395117588,0.1283358037,0.0321726501,-0.3007615507,0.180201441,0.2179471701,0.0696545914,-0.4603878558,-0.3480935395,0.3474244773,0.0905086249,-0.1073826477,-0.0542227924,-0.2201905251,-0.1384026557,-0.4269974828,-0.0372310765,0.2381214947,0.0470557883,-0.1664711833,-0.2025655359,-0.0576917753,-0.1789474338,0.2900052965,0.083204776,0.2673639953,-0.1505029947,0.1634496748,0.0319860876,-0.0256315488,0.5261566639,0.5509867668,-0.2388633639,-0.6641385555,-0.0823727474,-0.2544230223,0.4143175781,0.01887675,-0.1984589398,-0.0576789379,-0.0474886,-0.0492348745,-0.0493975468,0.0503164567,0.2752569318,-0.0799268261,-0.4217450917,-0.3295519352,0.2977847457,-0.049042888,0.1013770252,0.3222388029,-0.0535087325,-0.3448227048,0.4046643376,0.1755125076,0.8032982945,-0.3154443502,-0.0005889569,-0.0163416024,0.0220427178,0.4005249739,-0.6602150798,0.0886543915,-0.2385678142,-0.1123942286,-0.1394269615,-0.0939427614,0.2245839685,0.4019753635,-0.4566559196,0.1663924009,0.1698404253,-0.0922531039,-0.2540281713,0.1580015272,-0.1684266627,-0.1770733595,-0.1891635209,-0.0005676431,-0.17978324,0.124297522,0.088383764,-0.0329314731,-0.1870020032,-0.019758936,-0.0014657561,0.0808525532,-0.1125742719,-0.1086507216,-0.1998767406,-0.2284726351,0.2189382613,0.6059162021,0.2206418216,-0.1742937863,-0.1218375713,0.1038349196,-0.0908457488,-0.1631279588,-0.098410517,-0.0659004301,0.1742140353,-0.0648382902,-0.3266788423,0.1432164013,0.1472800821,-0.1852090955,0.2426713109,0.1704329699,0.103017576,-0.5554352999,-0.0098945517,0.0667537153,-0.0916172042,-0.1178815216,0.0122631602,0.0074496823,-0.2226940393,0.3630220294,-0.3571643829,-0.254780978,-0.0981250554,-0.0154844178,0.0817958415,-0.0431545414,0.2503519058,0.4610136449,0.080297254,-0.2130610347,0.3313158154,-0.0468947552,-0.5658300519,0.192509532,-0.0715362877,0.2121951729,0.2029460967,0.6457028985,-0.123753272,-0.0773306265,-0.1040319055,-0.3982209563,-0.1800463945,0.1881074905,0.0979687795,0.2443711013,0.1031343415,0.4572248459,0.1946162283,0.1644676477,-0.0912946165,0.4208087921,0.0160167255,0.2344453335,-0.2441108078,0.0202300716,0.1939492077,0.0831433833,-0.0726449192,0.1258039773,-0.0110147772,-0.1045785174,-0.1326719224,0.2093546987,0.0392485484,-0.0429210998,-0.1682123542,-0.204652369,-0.2691370845,-0.2062194049,0.1293133944,-0.0536734015,0.0404619873,0.3889114857,0.5246465802,0.0588931553,-0.552372694,-0.0831440985,-0.0314932317,0.0178719815,0.126294747,0.1004039124,-0.0244424623,0.0050222278,-0.5351830125,-0.0438055024,0.4880593717,-0.0485445671,0.32296139,0.0124438936,0.3669821322,-0.1774817854,0.4626227915,0.3782722652,-0.0179461502,0.1993534267,0.1697976291,0.0539515428,0.0176978894,0.1849734187,0.2205457836,-0.0006582027,0.0160363317,0.4903227687,-0.2690656781,0.1666280329,-0.371758163,-0.0421494357,0.3210780621,0.1542692184,0.303319186,0.9101154804,-0.0595885254,0.0866482779,0.1254602671,0.0057804156,0.0839022249,0.4529548585,-0.0634018555,0.3624615967,-0.2818384469,0.1066175476,0.208330974,-0.135065034,0.1179995835,0.1469717026,-0.1605196744,-0.2424382865,-0.0634819269,0.6651053429,-0.0648372844,-0.177545175,-0.2696160078,0.3691827655,0.2023305148,0.0846262202,-0.3215957284,-0.083563529,0.2510619164,-0.107363984,-0.1984586865,-0.2229514122,0.0666629821,0.2417345494,0.0216236245,-0.0559882447,0.0558457337,-0.0804671943,0.1826959848,-0.0439481065,0.0037152092,0.1432080567,-0.0718044564,0.3600772321,0.1603082716,0.0938417614,0.0030690506,-0.1667497903,0.0194929782,-0.0423204042,-0.07010331,0.3184689879,0.2982739508,0.1052557826,0.4170581102,0.2447344214,0.0421320871,-0.023673581,-0.1360382438,-0.0143915266,-0.039927315,-0.2733765244,0.3864547014,0.0119799003,-0.093482241,-0.0182078779,-0.0559803657,-0.3887083232,-0.1473826468,0.0218490548,-0.2804156542,0.0626523942,-0.0745948553,-0.0414689034,-0.0644684508,0.2426128387,-0.0561755709,0.2943323255,-0.3316037059,-0.1498588473,-0.1963465512,-0.1389787197,-0.4514510036,-0.5549052954,-0.07209941,0.067326799,0.0214780755,0.2326613516,0.0926978067,-0.1600025147,0.1497936994,0.5700589418,-0.1629796475,-0.2597334683,-0.279067874,0.3490466475,-0.1956105679,-0.3895210326,0.0840132236,0.3195920289,-0.2047347724,0.1459270418,-0.1514397264,0.4545051754,0.0031884925,-0.3598990738,0.062917389,0.5531312823,0.2346211076,0.3608319759,0.1640867293,0.0573130175,-0.0704560727,0.2012005001,0.1174852401,0.2913325131,-0.3149191737,0.2003580183,-0.1085395515,0.0894612297,-0.0003400925,-0.014992225,-0.3560351431,0.0652306899,-0.0707268715,0.1926894039,0.5887848735,0.5346587896,-0.2799522877,-0.1900985092,-0.2888097167,-0.1398529261,0.1622958779,-0.5503791571,-0.1900802553,0.1941524893,0.0497574881,0.3609685004,-0.0554295443,-0.575209856,-0.0025689879,0.2839442194,-0.1510566026,-0.1649816036,0.2527161539,0.3656668663,-0.132168293,0.0594671555,0.0527022257,0.1399403214,-0.2519494891,-0.014925777,-0.237212196]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1786","title":"How to use split dataset ","comments":"Thank you for the quick response! ","body":"![Capture1](https:\/\/user-images.githubusercontent.com\/78090287\/106057436-cb6a1f00-6111-11eb-8c9c-3658065b1fdf.PNG)\r\n\r\nHey,\r\nI want to split the lambada dataset into corpus, test, train and valid txt files (like penn treebank) but I am not able to achieve this. What I am doing is, executing the lambada.py file in my project but its not giving desired results. Any help will be appreciated!","comment_length":6,"text":"How to use split dataset \n![Capture1](https:\/\/user-images.githubusercontent.com\/78090287\/106057436-cb6a1f00-6111-11eb-8c9c-3658065b1fdf.PNG)\r\n\r\nHey,\r\nI want to split the lambada dataset into corpus, test, train and valid txt files (like penn treebank) but I am not able to achieve this. What I am doing is, executing the lambada.py file in my project but its not giving desired results. Any help will be appreciated!\nThank you for the quick response! ","embeddings":[-0.3412870169,-0.0799432471,-0.0685725808,0.349974215,0.1454317421,0.1740137935,0.1837530434,0.6036044955,-0.6203961372,-0.0244175252,-0.3259925544,0.1631098837,0.1784003228,0.432602942,-0.1338737607,-0.5312819481,0.0257801749,0.2501095831,0.0046424032,-0.0312587917,0.0128678642,0.1358522475,-0.4370960295,-0.0161432661,-0.5081188083,-0.2810676694,-0.245587945,0.0590517372,-0.3167053163,0.0402501151,0.0628129393,0.0608134829,0.2171420902,0.3915943503,-0.0001233874,0.0134182991,-0.1004704461,-0.2378047258,-0.0310616381,-0.0771494731,0.0590621457,0.2294310927,0.0516587682,-0.3398338854,-0.1824589968,-0.3610413373,-0.0067142509,0.0461786576,0.7706207633,0.5139290094,0.0182019286,-0.4751038849,0.0718030557,0.1701760739,0.3042525649,0.3989392817,-0.0575498641,-0.1091056988,0.2467321754,0.1147959605,0.0628187358,0.1368769556,-0.0161815006,0.2843043208,-0.0921097174,0.1690429151,-0.3129303157,-0.4284793139,0.1040589958,0.1714149117,0.6745436192,-0.3828538656,-0.0835382864,-0.0640795305,-0.1180870011,-0.3642676771,-0.1081123352,0.4440303445,-0.1635897607,0.2505545914,-0.2879108489,0.0520938076,-0.2852061987,0.2168703973,-0.0473031886,0.0857849419,-0.0542302728,0.143323943,0.1107383743,0.2740936279,0.1569014192,-0.1182341352,-0.3153847158,0.0390467346,-0.2710516155,-0.291503638,-0.4799602926,-0.3520372212,0.2520760298,0.3927045465,-0.1812564731,-0.1816564947,-0.1273738295,-0.1462503523,0.4497423768,0.0111506525,0.2036968768,0.3838519752,0.0904665738,-0.1436502934,-0.3426269591,-0.1334594339,-0.0532299057,-0.2257864177,-0.1189009026,0.327347666,0.2876941264,-0.3580709398,-0.1007483453,-0.4790122509,-0.6107210517,-0.2408696264,-0.0259959344,0.1746536642,0.1120347232,-0.2165992707,-0.0505473875,0.2864291966,-0.184345156,-0.236867398,0.0532409027,0.1471124589,-0.2647788227,-0.0543289818,0.2152874917,-0.4236468971,0.082757622,-0.117901966,0.0290842447,-0.2374394387,0.2904924154,-0.448728472,0.1004683301,0.1952213347,0.1179946139,-0.1317019463,0.0421434306,0.2241588831,-0.2162721753,0.1586589664,-0.1244096756,-0.0719370693,0.0525421277,-0.0295818746,0.1066658497,0.1239422858,-0.3108245432,0.3195697665,0.0727334246,0.0270067062,0.0414937995,0.2215272188,-0.148016721,-0.1161062196,0.1867319345,0.0034741086,-0.3836002946,0.1168498918,0.0751962438,-0.26960361,0.314257443,0.1641217768,-0.1040325463,0.5912819505,-0.002399534,0.119707495,0.4815702736,-0.8136312962,-0.1425164491,0.4133327305,0.0926362649,-0.4205981493,0.1316864938,-0.1161067784,0.6529287696,-0.0510529689,-0.2187437266,0.5653647184,-0.3385531306,-0.009376809,0.0118933916,-0.29294011,-0.2422184199,0.2132003307,-0.0102953771,-0.0486640111,0.2599250376,0.1177611426,0.3270382583,-0.167957589,0.3416075408,-0.1084914804,0.3565301299,-0.1306574196,0.1857822537,-0.3822259903,-0.1191646755,0.0077486662,-0.0177501105,0.0293442681,0.0799758285,-0.3227486014,-0.2012592256,-0.3401516378,-0.2113644779,-0.2709956467,0.0041566188,-0.3592329919,0.2525541782,0.0639936104,-0.2425792664,0.0931711867,0.0843450874,0.4002971053,0.1633850038,0.4176060855,0.0490712076,-0.0985489488,0.040993128,0.0197151154,-0.071846664,-0.2087424695,-0.1000001431,0.354948014,0.27046597,-0.0992372558,0.1993361115,-0.287119031,0.211729154,-0.3471509814,-0.0203319229,-0.0991915092,0.0457102135,-0.1817341596,-0.386898458,0.0867545009,-0.2627896369,0.1620510817,-0.0037506595,0.1285013855,0.2118420899,-0.3406466544,0.2315531969,0.0886403844,-0.1681699157,-0.2361675352,0.2056315094,-0.2902216017,0.0272908416,0.0079610907,0.4947623312,-0.0828408226,0.0283045582,0.0659766793,-0.0118348757,0.3243176937,0.1642234474,0.4708817303,0.4061581194,0.1481096447,0.1465270519,0.1855263114,0.269852221,0.0120187961,0.2387895435,0.0526437014,0.1858041435,0.0678152516,-0.2935587466,-0.1744145006,0.0106424028,-0.0701475963,0.2697226703,0.3438750207,0.020521732,0.1035661623,-0.2422216982,-0.5310816765,-0.3095686138,-0.1885725409,0.226672709,0.0810666382,0.0954315662,0.0391234234,0.3135066032,0.1322568506,0.1333393902,0.1618953347,-0.0394292735,0.1689239591,-0.0353386924,-0.1626445502,-0.1129955202,0.1347439438,0.374281615,0.7596662045,0.3099465072,-0.264331609,0.0922044143,-0.2511067688,-0.0032567331,0.2742128968,-0.2537922859,0.0166831166,0.1502750814,0.0066910712,-0.0380294248,-0.0361231118,0.0421510786,0.0254507381,-0.2478704602,-0.0648215935,-0.1519205868,0.0460383594,-0.0980083421,-0.5685401559,-0.1580942869,-0.2246876806,0.4766479731,0.1845575422,0.1511991322,-0.1798882782,0.0980534554,0.1556721032,-0.1310633421,-0.0143612837,0.2172050923,-0.1622860432,0.0337425508,-0.2204699516,-0.2618497312,0.428989172,0.1598299295,-0.000461064,-0.088800922,-0.1947084814,0.1381116211,-0.0175799448,-0.0352487192,0.0773293152,0.109044157,0.2270869911,-0.2193086445,0.1035125703,-0.0457754992,0.1493418068,0.2263105214,-0.1658373624,0.2691793442,-0.0952862725,0.3757542372,0.0794058889,0.6872524619,0.4494308233,-0.0682159886,-0.0382768027,-0.2141839713,-0.4060540497,-0.0639838651,-0.1788194925,0.1684097052,-0.0879952163,0.0202713273,0.3799196482,0.2927087247,0.0043632137,-0.3776004612,-0.362285018,-0.3850240409,-0.3425160944,0.1037357524,-0.0457158089,0.0202040169,-0.2412830442,-0.0833217204,0.3255710602,-0.1477571875,-0.268528074,0.1777317226,0.0510158986,-0.0150206489,-0.4692912698,0.043174468,-0.2547307611,0.4140937626,0.070269987,-0.1667288393,-0.098073557,0.1991733462,0.3187235594,0.0339002684,0.6328437924,-0.0278080758,-0.0721464604,-0.0998541564,-0.03235282,0.3173188865,0.2011755556,-0.2700578272,0.2667041719,0.1675983965,-0.0657636002,-0.3857210875,-0.2866610885,0.4017634988,0.0099195084,-0.1200281456,0.0250968896,-0.0856484696,-0.1649098694,-0.4397808909,-0.0132701183,0.2968487144,-0.0506978594,-0.0980949923,-0.2292717099,-0.0459274836,-0.0964483023,0.2398406416,0.0895429328,0.2376999259,-0.2755097747,0.1518415213,-0.015583857,-0.0602166727,0.5236555338,0.4734737873,-0.1936201304,-0.5399323702,-0.1204984784,-0.3393959701,0.4939302504,0.0790755078,-0.2228964269,0.0563639998,-0.0654479936,-0.0594410449,0.0012701597,0.0286191721,0.2609579861,0.0365631245,-0.4303353429,-0.2778626382,0.1143459678,-0.0258388985,0.0768352523,0.3664779961,-0.050155364,-0.2985242903,0.3836781979,0.215575844,0.8984950185,-0.2670555115,0.0516851805,-0.1310025007,-0.0673351213,0.3540557921,-0.6070682406,0.0993032679,-0.2494407743,-0.1488836408,-0.1642919183,-0.1083495989,0.2983589172,0.4156801403,-0.4870944023,0.0962137431,0.28772825,-0.1300332993,-0.2491708249,0.0601512194,-0.0455813855,-0.0901467204,-0.1045327187,-0.0128122363,-0.2053568065,0.0406714901,0.1285536438,-0.1105055138,-0.1803112924,0.0400205366,-0.0690497234,0.0546162836,-0.0893366262,-0.1351336539,-0.1384771466,-0.1690186262,0.2772354186,0.5255891681,0.2602678239,-0.1125219241,-0.0926981196,0.1174438372,-0.0762847438,-0.1298855394,-0.0153240124,0.0097333565,0.2654401362,-0.157581374,-0.355178237,0.2676064074,0.1751703471,-0.153740108,0.2649909556,0.0968487412,0.0607411973,-0.4905615747,-0.0607813708,0.0796049535,-0.2336194515,-0.1261756718,0.0251946133,0.0635167062,-0.2735058963,0.4032372832,-0.3881626427,-0.2195112705,-0.0492839776,0.0043390607,0.1321712583,-0.0095257461,0.1493086666,0.4276222289,0.0976478904,-0.2100987136,0.4277068079,-0.0428814813,-0.6376228333,0.226936236,-0.0544251278,0.1581470966,0.2063679248,0.6451449394,-0.0944721922,-0.0571228415,-0.1611975133,-0.3491708934,-0.3195790052,0.1717916876,0.0500082672,0.1492188573,0.1127881855,0.4720616043,0.2735511661,0.2178096771,-0.1125197411,0.4396477938,-0.0428453162,0.280988574,-0.1485965252,0.0124487896,0.2475562394,0.0576121435,-0.0762279034,0.1837333739,-0.0334775783,-0.1164168343,-0.104507871,0.2092128992,0.0016918796,-0.0924521089,-0.128635481,-0.2226386964,-0.310967207,-0.0389710851,0.0807748884,-0.0496468544,0.1115948409,0.3513354957,0.4245022833,-0.0610656627,-0.5661671162,-0.0922847241,-0.0330580361,0.0268770307,0.1007753611,0.0874022469,-0.0532972403,-0.0938121751,-0.4067375064,-0.0328399986,0.4548531473,-0.0035389569,0.2949209213,0.0943752527,0.2953438163,-0.2158225179,0.4336222112,0.2432436347,-0.0243521761,0.1777664274,0.2082341611,0.0340792686,-0.0052507408,0.1725883335,0.2087209672,-0.1127172634,0.0158610307,0.4235452116,-0.277895093,0.2743535936,-0.3673897982,-0.0182217248,0.2329045534,0.0434179232,0.1946105808,0.8824445605,-0.0935789421,0.1296719462,0.1452106237,0.027012201,0.1281421483,0.4959869981,-0.0475635715,0.3892989457,-0.2761987448,0.0793285221,0.3144405782,-0.0541826077,0.1602943689,0.1383519918,-0.1892952174,-0.2271596044,-0.0091981161,0.6125055552,0.0727149546,-0.1992131621,-0.1326942593,0.4126821756,0.2037737519,0.1450643688,-0.2834800184,-0.0332651474,0.2825853229,-0.1120804399,-0.1856741011,-0.1130842716,0.1278814226,0.2284614444,0.0670312941,0.0726940259,0.0186935496,-0.1556070745,0.1835703254,0.0146058174,0.0372922122,0.1122811139,-0.0090948222,0.3606677055,0.2238004953,0.068304494,-0.0512555242,-0.1356157362,0.0591671728,-0.0262194779,-0.0889776126,0.3058076203,0.304042846,0.1837462038,0.3142757118,0.1790761054,0.0273450073,-0.0250797216,-0.0591706261,-0.0135948388,-0.1653836221,-0.1864324808,0.3514402807,0.0500029996,-0.2264912874,-0.0215758011,-0.0362885669,-0.3530631661,-0.159569934,0.0349678285,-0.4097780585,0.0305207428,-0.100355193,-0.0365952514,-0.035588108,0.203265816,-0.0891476125,0.2655451,-0.2558456957,-0.1364368349,-0.1858063638,-0.0544737391,-0.4267215729,-0.5385267735,-0.1588670462,0.0061498447,0.1041746214,0.2403407097,0.1390342563,-0.1295043975,0.1221356764,0.5584467649,-0.1827678531,-0.2230422497,-0.2455352843,0.3307960629,-0.229023695,-0.3796153069,0.2300110161,0.3885210454,-0.1512034982,0.1117974073,-0.1674849391,0.4065766633,0.0002189285,-0.3880093694,0.0202651992,0.5218768716,0.189355433,0.3825166821,0.3454397321,0.1059826314,-0.1123188213,0.1640266031,0.226789996,0.226925537,-0.3194072843,0.1486837417,-0.0444208831,0.0122295674,-0.0218801163,-0.0039130952,-0.3503672481,0.0551125333,-0.071021907,0.202111572,0.5935384035,0.5260388851,-0.2595942616,-0.327555716,-0.3404728472,-0.170149833,0.1707124114,-0.4786860645,-0.1561175734,0.26225546,-0.0047110673,0.3354539573,-0.1103468537,-0.5609514117,-0.0788547322,0.2107683122,-0.1550815552,-0.1358770281,0.2721975446,0.3421288431,-0.0530602708,0.0734361783,0.084164843,0.1036844701,-0.2119469196,-0.0540144406,-0.2015880346]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1785","title":"Not enough disk space (Needed: Unknown size) when caching on a cluster","comments":"Hi ! \r\n\r\nWhat do you mean by \"disk_usage(\".\").free` can't compute on the cluster's shared disk\" exactly ?\r\nDoes it return 0 ?","body":"I'm running some experiments where I'm caching datasets on a cluster and accessing it through multiple compute nodes. However, I get an error when loading the cached dataset from the shared disk.\r\n\r\nThe exact error thrown:\r\n\r\n```bash\r\n>>> load_dataset(dataset, cache_dir=\"\/path\/to\/cluster\/shared\/path\")\r\nOSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)\r\n```\r\n\r\n\r\n[`utils.has_sufficient_disk_space`](https:\/\/github.com\/huggingface\/datasets\/blob\/8a03ab7d123a76ee744304f21ce868c75f411214\/src\/datasets\/utils\/py_utils.py#L332) fails on each job because of how the cluster system is designed (`disk_usage(\".\").free` can't compute on the cluster's shared disk).\r\n\r\n\r\nThis is exactly where the error gets thrown:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L502\r\n\r\n```python\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {}, post-processed: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n utils.size_str(self.info.dataset_size or 0),\r\n utils.size_str(self.info.post_processing_size or 0),\r\n )\r\n )\r\n\r\n```\r\n\r\nWhat would be a good way to circumvent this? my current fix is to manually comment out that part, but that is not ideal. \r\nWould it be possible to pass a flag to skip this check on disk space?","comment_length":22,"text":"Not enough disk space (Needed: Unknown size) when caching on a cluster\nI'm running some experiments where I'm caching datasets on a cluster and accessing it through multiple compute nodes. However, I get an error when loading the cached dataset from the shared disk.\r\n\r\nThe exact error thrown:\r\n\r\n```bash\r\n>>> load_dataset(dataset, cache_dir=\"\/path\/to\/cluster\/shared\/path\")\r\nOSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)\r\n```\r\n\r\n\r\n[`utils.has_sufficient_disk_space`](https:\/\/github.com\/huggingface\/datasets\/blob\/8a03ab7d123a76ee744304f21ce868c75f411214\/src\/datasets\/utils\/py_utils.py#L332) fails on each job because of how the cluster system is designed (`disk_usage(\".\").free` can't compute on the cluster's shared disk).\r\n\r\n\r\nThis is exactly where the error gets thrown:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L502\r\n\r\n```python\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {}, post-processed: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n utils.size_str(self.info.dataset_size or 0),\r\n utils.size_str(self.info.post_processing_size or 0),\r\n )\r\n )\r\n\r\n```\r\n\r\nWhat would be a good way to circumvent this? my current fix is to manually comment out that part, but that is not ideal. \r\nWould it be possible to pass a flag to skip this check on disk space?\nHi ! \r\n\r\nWhat do you mean by \"disk_usage(\".\").free` can't compute on the cluster's shared disk\" exactly ?\r\nDoes it return 0 ?","embeddings":[-0.0777432546,-0.4187171161,-0.0846218243,0.2431606501,-0.0970096961,0.1506603807,0.0705348626,0.3401208818,0.3962355852,0.3933159113,0.4183927476,-0.1389313489,-0.0912498608,0.0038917698,-0.1739582866,0.2417320162,-0.0082299486,-0.0606454238,0.3455670178,0.1274867654,-0.405728817,0.0130443303,0.0944502652,-0.3736138344,-0.542624414,-0.1880920529,-0.1662807167,0.1566853374,0.2269620299,-0.2247435004,0.3221060932,-0.0054948069,-0.1011393741,0.5720781088,-0.000115944,0.0363123342,0.1715248972,-0.1809052825,-0.4399374723,-0.0386959203,0.2173043787,-0.4069226086,-0.2233585268,-0.3619658649,-0.0103093972,0.16035676,-0.0511811189,-0.4436121881,0.1765249521,-0.0691841096,0.1679693311,0.3530050814,-0.2203017622,-0.0926141366,-0.265940696,0.1767231971,-0.046913106,0.5180592537,0.2992629409,0.0753154457,-0.278730154,0.220809415,-0.1333248466,0.1601159424,0.4138147831,-0.1359979659,-0.0745279565,-0.1845136285,0.196237877,-0.0268672947,0.4277347624,-0.1950327456,-0.5273439288,-0.5711030364,0.0500526614,-0.1597716212,0.439658612,0.0364257619,-0.4096871614,0.1226714104,-0.3018596768,-0.3372609317,-0.2010516077,-0.2099774629,0.0532090776,0.1641300172,0.0016833327,0.083085537,0.3270326853,0.0175013132,-0.0256257802,0.0264363643,0.1729609668,0.2536420822,-0.2865471244,0.0151761649,-0.0948409587,0.2101237476,0.101119183,0.2106985748,0.0610056035,-0.1125408858,0.0735730827,0.2540668547,0.0136580197,0.2466122806,-0.2650673985,-0.0978883803,0.2030567378,-0.0030463671,0.1735390574,-0.1133197173,0.0481721535,-0.3508072197,0.1162030324,-0.0180879608,0.2040126473,-0.0714850947,-0.3821848929,0.1272478998,0.2050054371,0.3049914837,0.2122824043,0.2848548591,0.0810501873,0.2399788499,0.2188417315,0.1994525045,-0.0902340636,-0.0968263894,-0.1159810424,-0.0147075355,0.0885305181,0.1693156362,0.3448050916,0.0367706008,0.2115975916,-0.0179041158,0.4590659142,-0.1054459512,-0.0721758902,-0.2352618277,-0.1191628948,0.4233517647,0.1197756976,0.0117573794,0.2519826591,0.1500506103,-0.1510386318,0.2677905858,-0.3836438954,-0.4553592801,0.0440193452,0.0724626184,-0.3528414071,0.2011191398,-0.4066096246,-0.0384400897,0.3097243011,-0.1824393868,0.1011645198,0.0928381905,-0.3887099922,-0.2462706715,0.1381942779,0.3473314941,-0.1142510027,0.0037163277,-0.3107164502,0.1621557027,0.0525881574,0.4145958126,0.131162703,0.1386437565,-0.4277253449,0.138010487,0.1127065942,-0.3308421373,-0.5418027639,0.3061300516,-0.1718518436,0.2516016066,0.537789166,0.2614465952,-0.0163331181,-0.2336914986,0.2327867597,0.271453768,-0.2196908742,0.0563449599,-0.3533278108,-0.2634562254,0.0278719384,0.0614294,0.3454007506,0.1085645556,0.0109477919,-0.0060027656,0.1818708628,0.0549793318,0.2010059208,0.3193238974,0.3190516531,-0.3444590569,-0.0738658607,0.0217245314,-0.5194780231,0.3491626382,0.1930718422,-0.2100150585,-0.0646497682,-0.443295449,-0.2809630632,0.0622078627,0.2482933402,0.3274293542,0.0486466065,-0.0373882763,0.249466747,-0.306935966,-0.0596500039,0.1668305844,-0.2572431266,-0.0578986779,0.1028762013,-0.0371632501,-0.0423917137,-0.140039593,0.0494481213,0.0646104813,0.0650270507,0.1349681914,-0.0143349981,0.4134909809,-0.1063747108,0.2247718275,-0.0903512463,0.1346000135,0.1523135304,0.3512614369,0.0020521295,0.1752402633,0.0069469591,-0.1697661281,-0.1110437587,0.2095975876,-0.0455697291,0.2853596509,-0.082658127,0.0640584752,0.0624382496,0.0560384765,0.1237755492,-0.0592896119,0.2901299,-0.1966567934,0.4879351854,0.0813175663,0.0228467751,-0.0144203082,0.069439292,0.1592282355,0.2913535535,0.1565068215,0.2970966697,-0.0345214605,-0.0058793798,0.2817410529,0.2664145827,0.1560027152,0.015309358,-0.1753182411,0.0451875925,-0.2055585682,0.0528198555,-0.0503015332,0.1580183506,0.2922110558,0.0067300084,-0.2964170873,0.0960466042,-0.309045732,0.2155113816,0.0772978142,-0.2874605954,-0.0579072982,0.0557797439,0.1828185171,0.0615936369,0.1889596432,-0.45891729,-0.314848125,0.1746363044,0.6444194317,-0.2373899072,-0.0952426791,-0.3729638457,0.47324121,0.0780528784,-0.4257940054,-0.0265632216,-0.2685465217,-0.1055517048,0.0170707349,0.3254656196,-0.2578275502,0.3435428143,0.0094133504,0.0713445768,-0.4250703454,0.1581226587,-0.1697877198,0.0267442688,0.1604891568,-0.1307636648,0.3355614543,-0.1221082732,-0.1611018181,0.1600524932,-0.0386517271,-0.0417478047,0.0091869533,0.1570615619,0.1753278226,-0.3232641518,0.0315406322,-0.1793826818,-0.5495167375,0.1743905246,-0.1816838086,0.2716465294,0.103640452,0.1733868867,-0.1049473435,0.7099817991,-0.1731761396,-0.2496192306,-0.6116562486,0.2382910401,-0.0803573653,-0.029552035,-0.0471714363,0.0296280216,0.2122024596,0.4242192209,-0.4782416821,-0.4002625942,-0.1671660841,0.2294674516,-0.2420354337,-0.0321569964,0.179103747,-0.1486543715,-0.0734726936,-0.1409638077,-0.0281644203,0.052829124,0.0094645806,0.2762105465,0.2927938402,0.2192623615,0.3207839429,0.7062381506,-0.1594575793,0.2500558496,0.2755180597,0.1817433685,0.6162175536,0.0905259475,-0.0536038391,0.0400996394,0.0080741094,-0.6150441766,0.0189749803,-0.0332708806,0.1380777508,-0.028210273,-0.3949748576,0.0797362924,-0.2690038085,0.3064286411,-0.4407605529,0.2631149292,-0.2624957263,0.0036537724,-0.3661702573,-0.0888018832,-0.0278780982,0.0957245156,0.5163893104,0.1993711442,-0.2578812838,-0.2660910189,-0.5647937059,0.3253159523,-0.0060486109,0.124461174,-0.3100081384,0.0745289549,-0.0808814466,0.0101274392,0.7869898081,-0.0905365571,0.2198172063,0.0822387934,0.0257656258,-0.5359348059,-0.0108491303,0.1368561983,0.3804580867,-0.0719698519,0.4385934174,-0.1767694503,0.1249716505,-0.4394182563,0.3606899679,-0.3306917846,-0.2415882647,-0.3854801953,-0.381474793,-0.4833005369,0.4473903179,0.4565170407,0.2527860105,-0.0483525693,0.3299098015,-0.1702632755,-0.0149607062,0.1736515164,-0.0075801178,0.4432084858,-0.1366670579,0.0413887985,0.1019200087,-0.0565784462,0.4277549982,0.4132584035,0.1096961051,0.1532112807,0.2755295932,0.0156118423,0.3339522779,0.3349297941,-0.0720830709,0.0272417497,0.1159468964,0.2456974536,-0.4392106235,-0.0649360716,-0.3937533498,-0.0534106717,-0.0902632251,-0.0206971467,0.3726735115,0.3526023328,0.0177913494,0.0702154413,0.3179870248,-0.312716037,0.2607080042,-0.0260276664,0.8725862503,-0.0851960108,0.0190446638,0.0668715686,0.300827086,0.3401433825,-0.4726655781,0.3613077104,0.044084318,-0.2942003012,0.032581687,-0.0121935271,-0.021077238,0.2959340811,-0.1688451171,0.4016993642,-0.5068138838,0.0121941203,-0.0954152122,0.1365531236,-0.6846148968,-0.3742953539,-0.2389675081,0.1206893325,-0.1377955973,0.2103001922,-0.2451041192,0.1086563542,0.0337137021,-0.3128852248,-0.3662663102,-0.1701539159,0.0267177485,0.30356884,-0.1789850444,0.023642892,-0.0893730149,0.1346775889,0.215277493,0.3010421693,-0.1435466856,0.1024801061,-0.1572918892,-0.0557976328,0.0648063496,0.0752229244,0.1123975962,0.0016258261,-0.3096863627,0.1574651748,-0.2591036558,-0.48992908,0.0290627331,-0.0438140891,-0.2942894101,-0.0338733867,-0.2902983725,-0.4566001594,-0.0054634041,-0.0876832604,0.066506885,0.0191741381,-0.1189481691,0.1476351321,0.023525605,-0.2044091672,0.1164215803,0.1907236129,0.1715484858,-0.3462996483,0.3859966099,-0.3731985688,-0.0873677656,-0.1736939251,-0.1504089236,-0.1342989355,0.0950825363,0.3430252671,-0.0794039071,-0.128579393,-0.2427508533,-0.2310128659,0.6502943039,0.4262381494,-0.1440830678,-0.4783772826,-0.3731835485,0.2676646113,0.0803002119,0.2381606102,-0.0641208217,-0.0984878987,0.0306084044,0.1129468232,-0.2559910119,-0.0557872616,-0.4407341778,0.0071715922,0.2117857784,0.1263764203,0.2143028975,-0.0947062522,-0.0827281326,0.1265942156,-0.0557577945,-0.1126358956,0.0449620858,0.1506276131,0.0469185263,0.1951721609,0.1735437512,-0.4588237107,0.2376364022,-0.2655067444,0.1753760129,0.41826877,-0.0454813391,-0.3415568471,0.0746578425,0.4974749386,-0.1709237695,0.1422317028,0.1285944879,0.154649049,-0.2296802551,0.184286952,0.1251479387,0.1417401433,0.236686185,-0.1120783314,0.2201440632,0.0385219157,-0.1545711309,0.0684712753,-0.0448022932,0.2478228807,0.2652956843,-0.0790446252,-0.0293991826,0.0338054039,0.0698838606,0.2173097134,-0.2394904494,-0.1172485054,0.2396288067,-0.0299394131,0.1197568104,0.5013854504,0.3876144588,-0.3748737276,0.0745419413,0.0923809335,0.5648558736,-0.2601763606,0.0413093716,0.133695066,0.0506068207,0.0030422173,0.1490101218,-0.0633012503,0.3891547918,0.3070395887,-0.1713535339,0.2159722894,0.2408117056,-0.1208608598,-0.1497628689,-0.4051087797,0.0863526687,0.0817287043,0.0425829925,0.1304463148,-0.3192768693,-0.0476001762,-0.3641909659,-0.1965409964,-0.22491844,-0.1132412404,-0.4139758348,-0.010649167,-0.0181167759,-0.0879318193,-0.0273674056,-0.1237477139,0.3569250107,-0.385168761,0.3071930408,0.0091478787,0.0622749887,0.2990060449,-0.2994197011,0.4684779346,0.1641292274,-0.3684591353,0.037796393,0.2732798159,-0.2538939714,0.299712956,0.171287179,0.678319037,0.2227468789,-0.0918961391,0.0675157979,0.0342152268,0.0763492212,-0.3394201696,-0.0230132863,0.0669182613,-0.1332622319,0.1145668924,0.1374943554,-0.0944960862,0.2236462384,0.0117416503,0.2606374323,-0.2572070956,-0.1200354174,-0.1719239354,-0.1995496005,-0.1337638348,0.0917495266,-0.2527796328,-0.2110942006,0.3524641693,-0.0509155765,0.183152765,0.1991884857,0.067212522,-0.0960669443,0.271378696,0.3381310403,-0.5006517172,-0.3073892891,0.0890709236,-0.5300481319,0.2179004252,-0.2879717052,-0.0791888162,-0.0316296592,0.216260314,-0.0353243575,-0.1100046337,0.0923833996,-0.0869553164,-0.0684573874,-0.1544490755,-0.2633790076,-0.2788788676,-0.0878155828,-0.2979598939,0.2715916932,-0.5973241329,-0.0476713628,0.1096985787,0.0626047924,0.0808959603,-0.0069344779,0.2312024087,0.4421150684,0.3236790001,-0.0014010575,0.4533215761,0.0691781789,-0.1578182876,-0.3555991352,-0.0512277186,-0.1567316502,0.0503353998,0.1229820624,0.2397244871,-0.6420225501,0.0890365317,-0.2371957451,0.2477700412,0.0678071231,-0.0797343701,0.0980477482,-0.0238934793,-0.3431888819,0.1982951909,0.2360748947,0.2301772684,-0.0658628494,0.3270435035,-0.1933472306,-0.1563149393,0.4829874635,-0.1396929771,-0.3495692015,-0.2246772647,0.4760475159,0.0577694811,-0.040422637,-0.1970507354,-0.0889563486,0.185353756,0.0791050568,0.063200742,-0.1794749498,-0.2648894787,-0.2492375225,-0.0139695071,0.2091803253,-0.312114656,-0.3135881722,0.0350733325,-0.2094392478]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1785","title":"Not enough disk space (Needed: Unknown size) when caching on a cluster","comments":"Yes, that's right. It shows 0 free space even though there is. I suspect it might have to do with permissions on the shared disk.\r\n\r\n```python\r\n>>> disk_usage(\".\")\r\nusage(total=999999, used=999999, free=0)\r\n```","body":"I'm running some experiments where I'm caching datasets on a cluster and accessing it through multiple compute nodes. However, I get an error when loading the cached dataset from the shared disk.\r\n\r\nThe exact error thrown:\r\n\r\n```bash\r\n>>> load_dataset(dataset, cache_dir=\"\/path\/to\/cluster\/shared\/path\")\r\nOSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)\r\n```\r\n\r\n\r\n[`utils.has_sufficient_disk_space`](https:\/\/github.com\/huggingface\/datasets\/blob\/8a03ab7d123a76ee744304f21ce868c75f411214\/src\/datasets\/utils\/py_utils.py#L332) fails on each job because of how the cluster system is designed (`disk_usage(\".\").free` can't compute on the cluster's shared disk).\r\n\r\n\r\nThis is exactly where the error gets thrown:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L502\r\n\r\n```python\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {}, post-processed: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n utils.size_str(self.info.dataset_size or 0),\r\n utils.size_str(self.info.post_processing_size or 0),\r\n )\r\n )\r\n\r\n```\r\n\r\nWhat would be a good way to circumvent this? my current fix is to manually comment out that part, but that is not ideal. \r\nWould it be possible to pass a flag to skip this check on disk space?","comment_length":32,"text":"Not enough disk space (Needed: Unknown size) when caching on a cluster\nI'm running some experiments where I'm caching datasets on a cluster and accessing it through multiple compute nodes. However, I get an error when loading the cached dataset from the shared disk.\r\n\r\nThe exact error thrown:\r\n\r\n```bash\r\n>>> load_dataset(dataset, cache_dir=\"\/path\/to\/cluster\/shared\/path\")\r\nOSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)\r\n```\r\n\r\n\r\n[`utils.has_sufficient_disk_space`](https:\/\/github.com\/huggingface\/datasets\/blob\/8a03ab7d123a76ee744304f21ce868c75f411214\/src\/datasets\/utils\/py_utils.py#L332) fails on each job because of how the cluster system is designed (`disk_usage(\".\").free` can't compute on the cluster's shared disk).\r\n\r\n\r\nThis is exactly where the error gets thrown:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L502\r\n\r\n```python\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {}, post-processed: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n utils.size_str(self.info.dataset_size or 0),\r\n utils.size_str(self.info.post_processing_size or 0),\r\n )\r\n )\r\n\r\n```\r\n\r\nWhat would be a good way to circumvent this? my current fix is to manually comment out that part, but that is not ideal. \r\nWould it be possible to pass a flag to skip this check on disk space?\nYes, that's right. It shows 0 free space even though there is. I suspect it might have to do with permissions on the shared disk.\r\n\r\n```python\r\n>>> disk_usage(\".\")\r\nusage(total=999999, used=999999, free=0)\r\n```","embeddings":[-0.0777432546,-0.4187171161,-0.0846218243,0.2431606501,-0.0970096961,0.1506603807,0.0705348626,0.3401208818,0.3962355852,0.3933159113,0.4183927476,-0.1389313489,-0.0912498608,0.0038917698,-0.1739582866,0.2417320162,-0.0082299486,-0.0606454238,0.3455670178,0.1274867654,-0.405728817,0.0130443303,0.0944502652,-0.3736138344,-0.542624414,-0.1880920529,-0.1662807167,0.1566853374,0.2269620299,-0.2247435004,0.3221060932,-0.0054948069,-0.1011393741,0.5720781088,-0.000115944,0.0363123342,0.1715248972,-0.1809052825,-0.4399374723,-0.0386959203,0.2173043787,-0.4069226086,-0.2233585268,-0.3619658649,-0.0103093972,0.16035676,-0.0511811189,-0.4436121881,0.1765249521,-0.0691841096,0.1679693311,0.3530050814,-0.2203017622,-0.0926141366,-0.265940696,0.1767231971,-0.046913106,0.5180592537,0.2992629409,0.0753154457,-0.278730154,0.220809415,-0.1333248466,0.1601159424,0.4138147831,-0.1359979659,-0.0745279565,-0.1845136285,0.196237877,-0.0268672947,0.4277347624,-0.1950327456,-0.5273439288,-0.5711030364,0.0500526614,-0.1597716212,0.439658612,0.0364257619,-0.4096871614,0.1226714104,-0.3018596768,-0.3372609317,-0.2010516077,-0.2099774629,0.0532090776,0.1641300172,0.0016833327,0.083085537,0.3270326853,0.0175013132,-0.0256257802,0.0264363643,0.1729609668,0.2536420822,-0.2865471244,0.0151761649,-0.0948409587,0.2101237476,0.101119183,0.2106985748,0.0610056035,-0.1125408858,0.0735730827,0.2540668547,0.0136580197,0.2466122806,-0.2650673985,-0.0978883803,0.2030567378,-0.0030463671,0.1735390574,-0.1133197173,0.0481721535,-0.3508072197,0.1162030324,-0.0180879608,0.2040126473,-0.0714850947,-0.3821848929,0.1272478998,0.2050054371,0.3049914837,0.2122824043,0.2848548591,0.0810501873,0.2399788499,0.2188417315,0.1994525045,-0.0902340636,-0.0968263894,-0.1159810424,-0.0147075355,0.0885305181,0.1693156362,0.3448050916,0.0367706008,0.2115975916,-0.0179041158,0.4590659142,-0.1054459512,-0.0721758902,-0.2352618277,-0.1191628948,0.4233517647,0.1197756976,0.0117573794,0.2519826591,0.1500506103,-0.1510386318,0.2677905858,-0.3836438954,-0.4553592801,0.0440193452,0.0724626184,-0.3528414071,0.2011191398,-0.4066096246,-0.0384400897,0.3097243011,-0.1824393868,0.1011645198,0.0928381905,-0.3887099922,-0.2462706715,0.1381942779,0.3473314941,-0.1142510027,0.0037163277,-0.3107164502,0.1621557027,0.0525881574,0.4145958126,0.131162703,0.1386437565,-0.4277253449,0.138010487,0.1127065942,-0.3308421373,-0.5418027639,0.3061300516,-0.1718518436,0.2516016066,0.537789166,0.2614465952,-0.0163331181,-0.2336914986,0.2327867597,0.271453768,-0.2196908742,0.0563449599,-0.3533278108,-0.2634562254,0.0278719384,0.0614294,0.3454007506,0.1085645556,0.0109477919,-0.0060027656,0.1818708628,0.0549793318,0.2010059208,0.3193238974,0.3190516531,-0.3444590569,-0.0738658607,0.0217245314,-0.5194780231,0.3491626382,0.1930718422,-0.2100150585,-0.0646497682,-0.443295449,-0.2809630632,0.0622078627,0.2482933402,0.3274293542,0.0486466065,-0.0373882763,0.249466747,-0.306935966,-0.0596500039,0.1668305844,-0.2572431266,-0.0578986779,0.1028762013,-0.0371632501,-0.0423917137,-0.140039593,0.0494481213,0.0646104813,0.0650270507,0.1349681914,-0.0143349981,0.4134909809,-0.1063747108,0.2247718275,-0.0903512463,0.1346000135,0.1523135304,0.3512614369,0.0020521295,0.1752402633,0.0069469591,-0.1697661281,-0.1110437587,0.2095975876,-0.0455697291,0.2853596509,-0.082658127,0.0640584752,0.0624382496,0.0560384765,0.1237755492,-0.0592896119,0.2901299,-0.1966567934,0.4879351854,0.0813175663,0.0228467751,-0.0144203082,0.069439292,0.1592282355,0.2913535535,0.1565068215,0.2970966697,-0.0345214605,-0.0058793798,0.2817410529,0.2664145827,0.1560027152,0.015309358,-0.1753182411,0.0451875925,-0.2055585682,0.0528198555,-0.0503015332,0.1580183506,0.2922110558,0.0067300084,-0.2964170873,0.0960466042,-0.309045732,0.2155113816,0.0772978142,-0.2874605954,-0.0579072982,0.0557797439,0.1828185171,0.0615936369,0.1889596432,-0.45891729,-0.314848125,0.1746363044,0.6444194317,-0.2373899072,-0.0952426791,-0.3729638457,0.47324121,0.0780528784,-0.4257940054,-0.0265632216,-0.2685465217,-0.1055517048,0.0170707349,0.3254656196,-0.2578275502,0.3435428143,0.0094133504,0.0713445768,-0.4250703454,0.1581226587,-0.1697877198,0.0267442688,0.1604891568,-0.1307636648,0.3355614543,-0.1221082732,-0.1611018181,0.1600524932,-0.0386517271,-0.0417478047,0.0091869533,0.1570615619,0.1753278226,-0.3232641518,0.0315406322,-0.1793826818,-0.5495167375,0.1743905246,-0.1816838086,0.2716465294,0.103640452,0.1733868867,-0.1049473435,0.7099817991,-0.1731761396,-0.2496192306,-0.6116562486,0.2382910401,-0.0803573653,-0.029552035,-0.0471714363,0.0296280216,0.2122024596,0.4242192209,-0.4782416821,-0.4002625942,-0.1671660841,0.2294674516,-0.2420354337,-0.0321569964,0.179103747,-0.1486543715,-0.0734726936,-0.1409638077,-0.0281644203,0.052829124,0.0094645806,0.2762105465,0.2927938402,0.2192623615,0.3207839429,0.7062381506,-0.1594575793,0.2500558496,0.2755180597,0.1817433685,0.6162175536,0.0905259475,-0.0536038391,0.0400996394,0.0080741094,-0.6150441766,0.0189749803,-0.0332708806,0.1380777508,-0.028210273,-0.3949748576,0.0797362924,-0.2690038085,0.3064286411,-0.4407605529,0.2631149292,-0.2624957263,0.0036537724,-0.3661702573,-0.0888018832,-0.0278780982,0.0957245156,0.5163893104,0.1993711442,-0.2578812838,-0.2660910189,-0.5647937059,0.3253159523,-0.0060486109,0.124461174,-0.3100081384,0.0745289549,-0.0808814466,0.0101274392,0.7869898081,-0.0905365571,0.2198172063,0.0822387934,0.0257656258,-0.5359348059,-0.0108491303,0.1368561983,0.3804580867,-0.0719698519,0.4385934174,-0.1767694503,0.1249716505,-0.4394182563,0.3606899679,-0.3306917846,-0.2415882647,-0.3854801953,-0.381474793,-0.4833005369,0.4473903179,0.4565170407,0.2527860105,-0.0483525693,0.3299098015,-0.1702632755,-0.0149607062,0.1736515164,-0.0075801178,0.4432084858,-0.1366670579,0.0413887985,0.1019200087,-0.0565784462,0.4277549982,0.4132584035,0.1096961051,0.1532112807,0.2755295932,0.0156118423,0.3339522779,0.3349297941,-0.0720830709,0.0272417497,0.1159468964,0.2456974536,-0.4392106235,-0.0649360716,-0.3937533498,-0.0534106717,-0.0902632251,-0.0206971467,0.3726735115,0.3526023328,0.0177913494,0.0702154413,0.3179870248,-0.312716037,0.2607080042,-0.0260276664,0.8725862503,-0.0851960108,0.0190446638,0.0668715686,0.300827086,0.3401433825,-0.4726655781,0.3613077104,0.044084318,-0.2942003012,0.032581687,-0.0121935271,-0.021077238,0.2959340811,-0.1688451171,0.4016993642,-0.5068138838,0.0121941203,-0.0954152122,0.1365531236,-0.6846148968,-0.3742953539,-0.2389675081,0.1206893325,-0.1377955973,0.2103001922,-0.2451041192,0.1086563542,0.0337137021,-0.3128852248,-0.3662663102,-0.1701539159,0.0267177485,0.30356884,-0.1789850444,0.023642892,-0.0893730149,0.1346775889,0.215277493,0.3010421693,-0.1435466856,0.1024801061,-0.1572918892,-0.0557976328,0.0648063496,0.0752229244,0.1123975962,0.0016258261,-0.3096863627,0.1574651748,-0.2591036558,-0.48992908,0.0290627331,-0.0438140891,-0.2942894101,-0.0338733867,-0.2902983725,-0.4566001594,-0.0054634041,-0.0876832604,0.066506885,0.0191741381,-0.1189481691,0.1476351321,0.023525605,-0.2044091672,0.1164215803,0.1907236129,0.1715484858,-0.3462996483,0.3859966099,-0.3731985688,-0.0873677656,-0.1736939251,-0.1504089236,-0.1342989355,0.0950825363,0.3430252671,-0.0794039071,-0.128579393,-0.2427508533,-0.2310128659,0.6502943039,0.4262381494,-0.1440830678,-0.4783772826,-0.3731835485,0.2676646113,0.0803002119,0.2381606102,-0.0641208217,-0.0984878987,0.0306084044,0.1129468232,-0.2559910119,-0.0557872616,-0.4407341778,0.0071715922,0.2117857784,0.1263764203,0.2143028975,-0.0947062522,-0.0827281326,0.1265942156,-0.0557577945,-0.1126358956,0.0449620858,0.1506276131,0.0469185263,0.1951721609,0.1735437512,-0.4588237107,0.2376364022,-0.2655067444,0.1753760129,0.41826877,-0.0454813391,-0.3415568471,0.0746578425,0.4974749386,-0.1709237695,0.1422317028,0.1285944879,0.154649049,-0.2296802551,0.184286952,0.1251479387,0.1417401433,0.236686185,-0.1120783314,0.2201440632,0.0385219157,-0.1545711309,0.0684712753,-0.0448022932,0.2478228807,0.2652956843,-0.0790446252,-0.0293991826,0.0338054039,0.0698838606,0.2173097134,-0.2394904494,-0.1172485054,0.2396288067,-0.0299394131,0.1197568104,0.5013854504,0.3876144588,-0.3748737276,0.0745419413,0.0923809335,0.5648558736,-0.2601763606,0.0413093716,0.133695066,0.0506068207,0.0030422173,0.1490101218,-0.0633012503,0.3891547918,0.3070395887,-0.1713535339,0.2159722894,0.2408117056,-0.1208608598,-0.1497628689,-0.4051087797,0.0863526687,0.0817287043,0.0425829925,0.1304463148,-0.3192768693,-0.0476001762,-0.3641909659,-0.1965409964,-0.22491844,-0.1132412404,-0.4139758348,-0.010649167,-0.0181167759,-0.0879318193,-0.0273674056,-0.1237477139,0.3569250107,-0.385168761,0.3071930408,0.0091478787,0.0622749887,0.2990060449,-0.2994197011,0.4684779346,0.1641292274,-0.3684591353,0.037796393,0.2732798159,-0.2538939714,0.299712956,0.171287179,0.678319037,0.2227468789,-0.0918961391,0.0675157979,0.0342152268,0.0763492212,-0.3394201696,-0.0230132863,0.0669182613,-0.1332622319,0.1145668924,0.1374943554,-0.0944960862,0.2236462384,0.0117416503,0.2606374323,-0.2572070956,-0.1200354174,-0.1719239354,-0.1995496005,-0.1337638348,0.0917495266,-0.2527796328,-0.2110942006,0.3524641693,-0.0509155765,0.183152765,0.1991884857,0.067212522,-0.0960669443,0.271378696,0.3381310403,-0.5006517172,-0.3073892891,0.0890709236,-0.5300481319,0.2179004252,-0.2879717052,-0.0791888162,-0.0316296592,0.216260314,-0.0353243575,-0.1100046337,0.0923833996,-0.0869553164,-0.0684573874,-0.1544490755,-0.2633790076,-0.2788788676,-0.0878155828,-0.2979598939,0.2715916932,-0.5973241329,-0.0476713628,0.1096985787,0.0626047924,0.0808959603,-0.0069344779,0.2312024087,0.4421150684,0.3236790001,-0.0014010575,0.4533215761,0.0691781789,-0.1578182876,-0.3555991352,-0.0512277186,-0.1567316502,0.0503353998,0.1229820624,0.2397244871,-0.6420225501,0.0890365317,-0.2371957451,0.2477700412,0.0678071231,-0.0797343701,0.0980477482,-0.0238934793,-0.3431888819,0.1982951909,0.2360748947,0.2301772684,-0.0658628494,0.3270435035,-0.1933472306,-0.1563149393,0.4829874635,-0.1396929771,-0.3495692015,-0.2246772647,0.4760475159,0.0577694811,-0.040422637,-0.1970507354,-0.0889563486,0.185353756,0.0791050568,0.063200742,-0.1794749498,-0.2648894787,-0.2492375225,-0.0139695071,0.2091803253,-0.312114656,-0.3135881722,0.0350733325,-0.2094392478]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1785","title":"Not enough disk space (Needed: Unknown size) when caching on a cluster","comments":"That's an interesting behavior...\r\nDo you know any other way to get the free space that works in your case ?\r\nAlso if it's a permission issue could you try fix the permissions and let mus know if that helped ?","body":"I'm running some experiments where I'm caching datasets on a cluster and accessing it through multiple compute nodes. However, I get an error when loading the cached dataset from the shared disk.\r\n\r\nThe exact error thrown:\r\n\r\n```bash\r\n>>> load_dataset(dataset, cache_dir=\"\/path\/to\/cluster\/shared\/path\")\r\nOSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)\r\n```\r\n\r\n\r\n[`utils.has_sufficient_disk_space`](https:\/\/github.com\/huggingface\/datasets\/blob\/8a03ab7d123a76ee744304f21ce868c75f411214\/src\/datasets\/utils\/py_utils.py#L332) fails on each job because of how the cluster system is designed (`disk_usage(\".\").free` can't compute on the cluster's shared disk).\r\n\r\n\r\nThis is exactly where the error gets thrown:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L502\r\n\r\n```python\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {}, post-processed: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n utils.size_str(self.info.dataset_size or 0),\r\n utils.size_str(self.info.post_processing_size or 0),\r\n )\r\n )\r\n\r\n```\r\n\r\nWhat would be a good way to circumvent this? my current fix is to manually comment out that part, but that is not ideal. \r\nWould it be possible to pass a flag to skip this check on disk space?","comment_length":41,"text":"Not enough disk space (Needed: Unknown size) when caching on a cluster\nI'm running some experiments where I'm caching datasets on a cluster and accessing it through multiple compute nodes. However, I get an error when loading the cached dataset from the shared disk.\r\n\r\nThe exact error thrown:\r\n\r\n```bash\r\n>>> load_dataset(dataset, cache_dir=\"\/path\/to\/cluster\/shared\/path\")\r\nOSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)\r\n```\r\n\r\n\r\n[`utils.has_sufficient_disk_space`](https:\/\/github.com\/huggingface\/datasets\/blob\/8a03ab7d123a76ee744304f21ce868c75f411214\/src\/datasets\/utils\/py_utils.py#L332) fails on each job because of how the cluster system is designed (`disk_usage(\".\").free` can't compute on the cluster's shared disk).\r\n\r\n\r\nThis is exactly where the error gets thrown:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L502\r\n\r\n```python\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {}, post-processed: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n utils.size_str(self.info.dataset_size or 0),\r\n utils.size_str(self.info.post_processing_size or 0),\r\n )\r\n )\r\n\r\n```\r\n\r\nWhat would be a good way to circumvent this? my current fix is to manually comment out that part, but that is not ideal. \r\nWould it be possible to pass a flag to skip this check on disk space?\nThat's an interesting behavior...\r\nDo you know any other way to get the free space that works in your case ?\r\nAlso if it's a permission issue could you try fix the permissions and let mus know if that helped ?","embeddings":[-0.0777432546,-0.4187171161,-0.0846218243,0.2431606501,-0.0970096961,0.1506603807,0.0705348626,0.3401208818,0.3962355852,0.3933159113,0.4183927476,-0.1389313489,-0.0912498608,0.0038917698,-0.1739582866,0.2417320162,-0.0082299486,-0.0606454238,0.3455670178,0.1274867654,-0.405728817,0.0130443303,0.0944502652,-0.3736138344,-0.542624414,-0.1880920529,-0.1662807167,0.1566853374,0.2269620299,-0.2247435004,0.3221060932,-0.0054948069,-0.1011393741,0.5720781088,-0.000115944,0.0363123342,0.1715248972,-0.1809052825,-0.4399374723,-0.0386959203,0.2173043787,-0.4069226086,-0.2233585268,-0.3619658649,-0.0103093972,0.16035676,-0.0511811189,-0.4436121881,0.1765249521,-0.0691841096,0.1679693311,0.3530050814,-0.2203017622,-0.0926141366,-0.265940696,0.1767231971,-0.046913106,0.5180592537,0.2992629409,0.0753154457,-0.278730154,0.220809415,-0.1333248466,0.1601159424,0.4138147831,-0.1359979659,-0.0745279565,-0.1845136285,0.196237877,-0.0268672947,0.4277347624,-0.1950327456,-0.5273439288,-0.5711030364,0.0500526614,-0.1597716212,0.439658612,0.0364257619,-0.4096871614,0.1226714104,-0.3018596768,-0.3372609317,-0.2010516077,-0.2099774629,0.0532090776,0.1641300172,0.0016833327,0.083085537,0.3270326853,0.0175013132,-0.0256257802,0.0264363643,0.1729609668,0.2536420822,-0.2865471244,0.0151761649,-0.0948409587,0.2101237476,0.101119183,0.2106985748,0.0610056035,-0.1125408858,0.0735730827,0.2540668547,0.0136580197,0.2466122806,-0.2650673985,-0.0978883803,0.2030567378,-0.0030463671,0.1735390574,-0.1133197173,0.0481721535,-0.3508072197,0.1162030324,-0.0180879608,0.2040126473,-0.0714850947,-0.3821848929,0.1272478998,0.2050054371,0.3049914837,0.2122824043,0.2848548591,0.0810501873,0.2399788499,0.2188417315,0.1994525045,-0.0902340636,-0.0968263894,-0.1159810424,-0.0147075355,0.0885305181,0.1693156362,0.3448050916,0.0367706008,0.2115975916,-0.0179041158,0.4590659142,-0.1054459512,-0.0721758902,-0.2352618277,-0.1191628948,0.4233517647,0.1197756976,0.0117573794,0.2519826591,0.1500506103,-0.1510386318,0.2677905858,-0.3836438954,-0.4553592801,0.0440193452,0.0724626184,-0.3528414071,0.2011191398,-0.4066096246,-0.0384400897,0.3097243011,-0.1824393868,0.1011645198,0.0928381905,-0.3887099922,-0.2462706715,0.1381942779,0.3473314941,-0.1142510027,0.0037163277,-0.3107164502,0.1621557027,0.0525881574,0.4145958126,0.131162703,0.1386437565,-0.4277253449,0.138010487,0.1127065942,-0.3308421373,-0.5418027639,0.3061300516,-0.1718518436,0.2516016066,0.537789166,0.2614465952,-0.0163331181,-0.2336914986,0.2327867597,0.271453768,-0.2196908742,0.0563449599,-0.3533278108,-0.2634562254,0.0278719384,0.0614294,0.3454007506,0.1085645556,0.0109477919,-0.0060027656,0.1818708628,0.0549793318,0.2010059208,0.3193238974,0.3190516531,-0.3444590569,-0.0738658607,0.0217245314,-0.5194780231,0.3491626382,0.1930718422,-0.2100150585,-0.0646497682,-0.443295449,-0.2809630632,0.0622078627,0.2482933402,0.3274293542,0.0486466065,-0.0373882763,0.249466747,-0.306935966,-0.0596500039,0.1668305844,-0.2572431266,-0.0578986779,0.1028762013,-0.0371632501,-0.0423917137,-0.140039593,0.0494481213,0.0646104813,0.0650270507,0.1349681914,-0.0143349981,0.4134909809,-0.1063747108,0.2247718275,-0.0903512463,0.1346000135,0.1523135304,0.3512614369,0.0020521295,0.1752402633,0.0069469591,-0.1697661281,-0.1110437587,0.2095975876,-0.0455697291,0.2853596509,-0.082658127,0.0640584752,0.0624382496,0.0560384765,0.1237755492,-0.0592896119,0.2901299,-0.1966567934,0.4879351854,0.0813175663,0.0228467751,-0.0144203082,0.069439292,0.1592282355,0.2913535535,0.1565068215,0.2970966697,-0.0345214605,-0.0058793798,0.2817410529,0.2664145827,0.1560027152,0.015309358,-0.1753182411,0.0451875925,-0.2055585682,0.0528198555,-0.0503015332,0.1580183506,0.2922110558,0.0067300084,-0.2964170873,0.0960466042,-0.309045732,0.2155113816,0.0772978142,-0.2874605954,-0.0579072982,0.0557797439,0.1828185171,0.0615936369,0.1889596432,-0.45891729,-0.314848125,0.1746363044,0.6444194317,-0.2373899072,-0.0952426791,-0.3729638457,0.47324121,0.0780528784,-0.4257940054,-0.0265632216,-0.2685465217,-0.1055517048,0.0170707349,0.3254656196,-0.2578275502,0.3435428143,0.0094133504,0.0713445768,-0.4250703454,0.1581226587,-0.1697877198,0.0267442688,0.1604891568,-0.1307636648,0.3355614543,-0.1221082732,-0.1611018181,0.1600524932,-0.0386517271,-0.0417478047,0.0091869533,0.1570615619,0.1753278226,-0.3232641518,0.0315406322,-0.1793826818,-0.5495167375,0.1743905246,-0.1816838086,0.2716465294,0.103640452,0.1733868867,-0.1049473435,0.7099817991,-0.1731761396,-0.2496192306,-0.6116562486,0.2382910401,-0.0803573653,-0.029552035,-0.0471714363,0.0296280216,0.2122024596,0.4242192209,-0.4782416821,-0.4002625942,-0.1671660841,0.2294674516,-0.2420354337,-0.0321569964,0.179103747,-0.1486543715,-0.0734726936,-0.1409638077,-0.0281644203,0.052829124,0.0094645806,0.2762105465,0.2927938402,0.2192623615,0.3207839429,0.7062381506,-0.1594575793,0.2500558496,0.2755180597,0.1817433685,0.6162175536,0.0905259475,-0.0536038391,0.0400996394,0.0080741094,-0.6150441766,0.0189749803,-0.0332708806,0.1380777508,-0.028210273,-0.3949748576,0.0797362924,-0.2690038085,0.3064286411,-0.4407605529,0.2631149292,-0.2624957263,0.0036537724,-0.3661702573,-0.0888018832,-0.0278780982,0.0957245156,0.5163893104,0.1993711442,-0.2578812838,-0.2660910189,-0.5647937059,0.3253159523,-0.0060486109,0.124461174,-0.3100081384,0.0745289549,-0.0808814466,0.0101274392,0.7869898081,-0.0905365571,0.2198172063,0.0822387934,0.0257656258,-0.5359348059,-0.0108491303,0.1368561983,0.3804580867,-0.0719698519,0.4385934174,-0.1767694503,0.1249716505,-0.4394182563,0.3606899679,-0.3306917846,-0.2415882647,-0.3854801953,-0.381474793,-0.4833005369,0.4473903179,0.4565170407,0.2527860105,-0.0483525693,0.3299098015,-0.1702632755,-0.0149607062,0.1736515164,-0.0075801178,0.4432084858,-0.1366670579,0.0413887985,0.1019200087,-0.0565784462,0.4277549982,0.4132584035,0.1096961051,0.1532112807,0.2755295932,0.0156118423,0.3339522779,0.3349297941,-0.0720830709,0.0272417497,0.1159468964,0.2456974536,-0.4392106235,-0.0649360716,-0.3937533498,-0.0534106717,-0.0902632251,-0.0206971467,0.3726735115,0.3526023328,0.0177913494,0.0702154413,0.3179870248,-0.312716037,0.2607080042,-0.0260276664,0.8725862503,-0.0851960108,0.0190446638,0.0668715686,0.300827086,0.3401433825,-0.4726655781,0.3613077104,0.044084318,-0.2942003012,0.032581687,-0.0121935271,-0.021077238,0.2959340811,-0.1688451171,0.4016993642,-0.5068138838,0.0121941203,-0.0954152122,0.1365531236,-0.6846148968,-0.3742953539,-0.2389675081,0.1206893325,-0.1377955973,0.2103001922,-0.2451041192,0.1086563542,0.0337137021,-0.3128852248,-0.3662663102,-0.1701539159,0.0267177485,0.30356884,-0.1789850444,0.023642892,-0.0893730149,0.1346775889,0.215277493,0.3010421693,-0.1435466856,0.1024801061,-0.1572918892,-0.0557976328,0.0648063496,0.0752229244,0.1123975962,0.0016258261,-0.3096863627,0.1574651748,-0.2591036558,-0.48992908,0.0290627331,-0.0438140891,-0.2942894101,-0.0338733867,-0.2902983725,-0.4566001594,-0.0054634041,-0.0876832604,0.066506885,0.0191741381,-0.1189481691,0.1476351321,0.023525605,-0.2044091672,0.1164215803,0.1907236129,0.1715484858,-0.3462996483,0.3859966099,-0.3731985688,-0.0873677656,-0.1736939251,-0.1504089236,-0.1342989355,0.0950825363,0.3430252671,-0.0794039071,-0.128579393,-0.2427508533,-0.2310128659,0.6502943039,0.4262381494,-0.1440830678,-0.4783772826,-0.3731835485,0.2676646113,0.0803002119,0.2381606102,-0.0641208217,-0.0984878987,0.0306084044,0.1129468232,-0.2559910119,-0.0557872616,-0.4407341778,0.0071715922,0.2117857784,0.1263764203,0.2143028975,-0.0947062522,-0.0827281326,0.1265942156,-0.0557577945,-0.1126358956,0.0449620858,0.1506276131,0.0469185263,0.1951721609,0.1735437512,-0.4588237107,0.2376364022,-0.2655067444,0.1753760129,0.41826877,-0.0454813391,-0.3415568471,0.0746578425,0.4974749386,-0.1709237695,0.1422317028,0.1285944879,0.154649049,-0.2296802551,0.184286952,0.1251479387,0.1417401433,0.236686185,-0.1120783314,0.2201440632,0.0385219157,-0.1545711309,0.0684712753,-0.0448022932,0.2478228807,0.2652956843,-0.0790446252,-0.0293991826,0.0338054039,0.0698838606,0.2173097134,-0.2394904494,-0.1172485054,0.2396288067,-0.0299394131,0.1197568104,0.5013854504,0.3876144588,-0.3748737276,0.0745419413,0.0923809335,0.5648558736,-0.2601763606,0.0413093716,0.133695066,0.0506068207,0.0030422173,0.1490101218,-0.0633012503,0.3891547918,0.3070395887,-0.1713535339,0.2159722894,0.2408117056,-0.1208608598,-0.1497628689,-0.4051087797,0.0863526687,0.0817287043,0.0425829925,0.1304463148,-0.3192768693,-0.0476001762,-0.3641909659,-0.1965409964,-0.22491844,-0.1132412404,-0.4139758348,-0.010649167,-0.0181167759,-0.0879318193,-0.0273674056,-0.1237477139,0.3569250107,-0.385168761,0.3071930408,0.0091478787,0.0622749887,0.2990060449,-0.2994197011,0.4684779346,0.1641292274,-0.3684591353,0.037796393,0.2732798159,-0.2538939714,0.299712956,0.171287179,0.678319037,0.2227468789,-0.0918961391,0.0675157979,0.0342152268,0.0763492212,-0.3394201696,-0.0230132863,0.0669182613,-0.1332622319,0.1145668924,0.1374943554,-0.0944960862,0.2236462384,0.0117416503,0.2606374323,-0.2572070956,-0.1200354174,-0.1719239354,-0.1995496005,-0.1337638348,0.0917495266,-0.2527796328,-0.2110942006,0.3524641693,-0.0509155765,0.183152765,0.1991884857,0.067212522,-0.0960669443,0.271378696,0.3381310403,-0.5006517172,-0.3073892891,0.0890709236,-0.5300481319,0.2179004252,-0.2879717052,-0.0791888162,-0.0316296592,0.216260314,-0.0353243575,-0.1100046337,0.0923833996,-0.0869553164,-0.0684573874,-0.1544490755,-0.2633790076,-0.2788788676,-0.0878155828,-0.2979598939,0.2715916932,-0.5973241329,-0.0476713628,0.1096985787,0.0626047924,0.0808959603,-0.0069344779,0.2312024087,0.4421150684,0.3236790001,-0.0014010575,0.4533215761,0.0691781789,-0.1578182876,-0.3555991352,-0.0512277186,-0.1567316502,0.0503353998,0.1229820624,0.2397244871,-0.6420225501,0.0890365317,-0.2371957451,0.2477700412,0.0678071231,-0.0797343701,0.0980477482,-0.0238934793,-0.3431888819,0.1982951909,0.2360748947,0.2301772684,-0.0658628494,0.3270435035,-0.1933472306,-0.1563149393,0.4829874635,-0.1396929771,-0.3495692015,-0.2246772647,0.4760475159,0.0577694811,-0.040422637,-0.1970507354,-0.0889563486,0.185353756,0.0791050568,0.063200742,-0.1794749498,-0.2648894787,-0.2492375225,-0.0139695071,0.2091803253,-0.312114656,-0.3135881722,0.0350733325,-0.2094392478]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1785","title":"Not enough disk space (Needed: Unknown size) when caching on a cluster","comments":"I think its an issue on the clusters end (unclear exactly why -- maybe something with docker containers?), will close the issue","body":"I'm running some experiments where I'm caching datasets on a cluster and accessing it through multiple compute nodes. However, I get an error when loading the cached dataset from the shared disk.\r\n\r\nThe exact error thrown:\r\n\r\n```bash\r\n>>> load_dataset(dataset, cache_dir=\"\/path\/to\/cluster\/shared\/path\")\r\nOSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)\r\n```\r\n\r\n\r\n[`utils.has_sufficient_disk_space`](https:\/\/github.com\/huggingface\/datasets\/blob\/8a03ab7d123a76ee744304f21ce868c75f411214\/src\/datasets\/utils\/py_utils.py#L332) fails on each job because of how the cluster system is designed (`disk_usage(\".\").free` can't compute on the cluster's shared disk).\r\n\r\n\r\nThis is exactly where the error gets thrown:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L502\r\n\r\n```python\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {}, post-processed: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n utils.size_str(self.info.dataset_size or 0),\r\n utils.size_str(self.info.post_processing_size or 0),\r\n )\r\n )\r\n\r\n```\r\n\r\nWhat would be a good way to circumvent this? my current fix is to manually comment out that part, but that is not ideal. \r\nWould it be possible to pass a flag to skip this check on disk space?","comment_length":22,"text":"Not enough disk space (Needed: Unknown size) when caching on a cluster\nI'm running some experiments where I'm caching datasets on a cluster and accessing it through multiple compute nodes. However, I get an error when loading the cached dataset from the shared disk.\r\n\r\nThe exact error thrown:\r\n\r\n```bash\r\n>>> load_dataset(dataset, cache_dir=\"\/path\/to\/cluster\/shared\/path\")\r\nOSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)\r\n```\r\n\r\n\r\n[`utils.has_sufficient_disk_space`](https:\/\/github.com\/huggingface\/datasets\/blob\/8a03ab7d123a76ee744304f21ce868c75f411214\/src\/datasets\/utils\/py_utils.py#L332) fails on each job because of how the cluster system is designed (`disk_usage(\".\").free` can't compute on the cluster's shared disk).\r\n\r\n\r\nThis is exactly where the error gets thrown:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/builder.py#L502\r\n\r\n```python\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {}, post-processed: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n utils.size_str(self.info.dataset_size or 0),\r\n utils.size_str(self.info.post_processing_size or 0),\r\n )\r\n )\r\n\r\n```\r\n\r\nWhat would be a good way to circumvent this? my current fix is to manually comment out that part, but that is not ideal. \r\nWould it be possible to pass a flag to skip this check on disk space?\nI think its an issue on the clusters end (unclear exactly why -- maybe something with docker containers?), will close the issue","embeddings":[-0.0777432546,-0.4187171161,-0.0846218243,0.2431606501,-0.0970096961,0.1506603807,0.0705348626,0.3401208818,0.3962355852,0.3933159113,0.4183927476,-0.1389313489,-0.0912498608,0.0038917698,-0.1739582866,0.2417320162,-0.0082299486,-0.0606454238,0.3455670178,0.1274867654,-0.405728817,0.0130443303,0.0944502652,-0.3736138344,-0.542624414,-0.1880920529,-0.1662807167,0.1566853374,0.2269620299,-0.2247435004,0.3221060932,-0.0054948069,-0.1011393741,0.5720781088,-0.000115944,0.0363123342,0.1715248972,-0.1809052825,-0.4399374723,-0.0386959203,0.2173043787,-0.4069226086,-0.2233585268,-0.3619658649,-0.0103093972,0.16035676,-0.0511811189,-0.4436121881,0.1765249521,-0.0691841096,0.1679693311,0.3530050814,-0.2203017622,-0.0926141366,-0.265940696,0.1767231971,-0.046913106,0.5180592537,0.2992629409,0.0753154457,-0.278730154,0.220809415,-0.1333248466,0.1601159424,0.4138147831,-0.1359979659,-0.0745279565,-0.1845136285,0.196237877,-0.0268672947,0.4277347624,-0.1950327456,-0.5273439288,-0.5711030364,0.0500526614,-0.1597716212,0.439658612,0.0364257619,-0.4096871614,0.1226714104,-0.3018596768,-0.3372609317,-0.2010516077,-0.2099774629,0.0532090776,0.1641300172,0.0016833327,0.083085537,0.3270326853,0.0175013132,-0.0256257802,0.0264363643,0.1729609668,0.2536420822,-0.2865471244,0.0151761649,-0.0948409587,0.2101237476,0.101119183,0.2106985748,0.0610056035,-0.1125408858,0.0735730827,0.2540668547,0.0136580197,0.2466122806,-0.2650673985,-0.0978883803,0.2030567378,-0.0030463671,0.1735390574,-0.1133197173,0.0481721535,-0.3508072197,0.1162030324,-0.0180879608,0.2040126473,-0.0714850947,-0.3821848929,0.1272478998,0.2050054371,0.3049914837,0.2122824043,0.2848548591,0.0810501873,0.2399788499,0.2188417315,0.1994525045,-0.0902340636,-0.0968263894,-0.1159810424,-0.0147075355,0.0885305181,0.1693156362,0.3448050916,0.0367706008,0.2115975916,-0.0179041158,0.4590659142,-0.1054459512,-0.0721758902,-0.2352618277,-0.1191628948,0.4233517647,0.1197756976,0.0117573794,0.2519826591,0.1500506103,-0.1510386318,0.2677905858,-0.3836438954,-0.4553592801,0.0440193452,0.0724626184,-0.3528414071,0.2011191398,-0.4066096246,-0.0384400897,0.3097243011,-0.1824393868,0.1011645198,0.0928381905,-0.3887099922,-0.2462706715,0.1381942779,0.3473314941,-0.1142510027,0.0037163277,-0.3107164502,0.1621557027,0.0525881574,0.4145958126,0.131162703,0.1386437565,-0.4277253449,0.138010487,0.1127065942,-0.3308421373,-0.5418027639,0.3061300516,-0.1718518436,0.2516016066,0.537789166,0.2614465952,-0.0163331181,-0.2336914986,0.2327867597,0.271453768,-0.2196908742,0.0563449599,-0.3533278108,-0.2634562254,0.0278719384,0.0614294,0.3454007506,0.1085645556,0.0109477919,-0.0060027656,0.1818708628,0.0549793318,0.2010059208,0.3193238974,0.3190516531,-0.3444590569,-0.0738658607,0.0217245314,-0.5194780231,0.3491626382,0.1930718422,-0.2100150585,-0.0646497682,-0.443295449,-0.2809630632,0.0622078627,0.2482933402,0.3274293542,0.0486466065,-0.0373882763,0.249466747,-0.306935966,-0.0596500039,0.1668305844,-0.2572431266,-0.0578986779,0.1028762013,-0.0371632501,-0.0423917137,-0.140039593,0.0494481213,0.0646104813,0.0650270507,0.1349681914,-0.0143349981,0.4134909809,-0.1063747108,0.2247718275,-0.0903512463,0.1346000135,0.1523135304,0.3512614369,0.0020521295,0.1752402633,0.0069469591,-0.1697661281,-0.1110437587,0.2095975876,-0.0455697291,0.2853596509,-0.082658127,0.0640584752,0.0624382496,0.0560384765,0.1237755492,-0.0592896119,0.2901299,-0.1966567934,0.4879351854,0.0813175663,0.0228467751,-0.0144203082,0.069439292,0.1592282355,0.2913535535,0.1565068215,0.2970966697,-0.0345214605,-0.0058793798,0.2817410529,0.2664145827,0.1560027152,0.015309358,-0.1753182411,0.0451875925,-0.2055585682,0.0528198555,-0.0503015332,0.1580183506,0.2922110558,0.0067300084,-0.2964170873,0.0960466042,-0.309045732,0.2155113816,0.0772978142,-0.2874605954,-0.0579072982,0.0557797439,0.1828185171,0.0615936369,0.1889596432,-0.45891729,-0.314848125,0.1746363044,0.6444194317,-0.2373899072,-0.0952426791,-0.3729638457,0.47324121,0.0780528784,-0.4257940054,-0.0265632216,-0.2685465217,-0.1055517048,0.0170707349,0.3254656196,-0.2578275502,0.3435428143,0.0094133504,0.0713445768,-0.4250703454,0.1581226587,-0.1697877198,0.0267442688,0.1604891568,-0.1307636648,0.3355614543,-0.1221082732,-0.1611018181,0.1600524932,-0.0386517271,-0.0417478047,0.0091869533,0.1570615619,0.1753278226,-0.3232641518,0.0315406322,-0.1793826818,-0.5495167375,0.1743905246,-0.1816838086,0.2716465294,0.103640452,0.1733868867,-0.1049473435,0.7099817991,-0.1731761396,-0.2496192306,-0.6116562486,0.2382910401,-0.0803573653,-0.029552035,-0.0471714363,0.0296280216,0.2122024596,0.4242192209,-0.4782416821,-0.4002625942,-0.1671660841,0.2294674516,-0.2420354337,-0.0321569964,0.179103747,-0.1486543715,-0.0734726936,-0.1409638077,-0.0281644203,0.052829124,0.0094645806,0.2762105465,0.2927938402,0.2192623615,0.3207839429,0.7062381506,-0.1594575793,0.2500558496,0.2755180597,0.1817433685,0.6162175536,0.0905259475,-0.0536038391,0.0400996394,0.0080741094,-0.6150441766,0.0189749803,-0.0332708806,0.1380777508,-0.028210273,-0.3949748576,0.0797362924,-0.2690038085,0.3064286411,-0.4407605529,0.2631149292,-0.2624957263,0.0036537724,-0.3661702573,-0.0888018832,-0.0278780982,0.0957245156,0.5163893104,0.1993711442,-0.2578812838,-0.2660910189,-0.5647937059,0.3253159523,-0.0060486109,0.124461174,-0.3100081384,0.0745289549,-0.0808814466,0.0101274392,0.7869898081,-0.0905365571,0.2198172063,0.0822387934,0.0257656258,-0.5359348059,-0.0108491303,0.1368561983,0.3804580867,-0.0719698519,0.4385934174,-0.1767694503,0.1249716505,-0.4394182563,0.3606899679,-0.3306917846,-0.2415882647,-0.3854801953,-0.381474793,-0.4833005369,0.4473903179,0.4565170407,0.2527860105,-0.0483525693,0.3299098015,-0.1702632755,-0.0149607062,0.1736515164,-0.0075801178,0.4432084858,-0.1366670579,0.0413887985,0.1019200087,-0.0565784462,0.4277549982,0.4132584035,0.1096961051,0.1532112807,0.2755295932,0.0156118423,0.3339522779,0.3349297941,-0.0720830709,0.0272417497,0.1159468964,0.2456974536,-0.4392106235,-0.0649360716,-0.3937533498,-0.0534106717,-0.0902632251,-0.0206971467,0.3726735115,0.3526023328,0.0177913494,0.0702154413,0.3179870248,-0.312716037,0.2607080042,-0.0260276664,0.8725862503,-0.0851960108,0.0190446638,0.0668715686,0.300827086,0.3401433825,-0.4726655781,0.3613077104,0.044084318,-0.2942003012,0.032581687,-0.0121935271,-0.021077238,0.2959340811,-0.1688451171,0.4016993642,-0.5068138838,0.0121941203,-0.0954152122,0.1365531236,-0.6846148968,-0.3742953539,-0.2389675081,0.1206893325,-0.1377955973,0.2103001922,-0.2451041192,0.1086563542,0.0337137021,-0.3128852248,-0.3662663102,-0.1701539159,0.0267177485,0.30356884,-0.1789850444,0.023642892,-0.0893730149,0.1346775889,0.215277493,0.3010421693,-0.1435466856,0.1024801061,-0.1572918892,-0.0557976328,0.0648063496,0.0752229244,0.1123975962,0.0016258261,-0.3096863627,0.1574651748,-0.2591036558,-0.48992908,0.0290627331,-0.0438140891,-0.2942894101,-0.0338733867,-0.2902983725,-0.4566001594,-0.0054634041,-0.0876832604,0.066506885,0.0191741381,-0.1189481691,0.1476351321,0.023525605,-0.2044091672,0.1164215803,0.1907236129,0.1715484858,-0.3462996483,0.3859966099,-0.3731985688,-0.0873677656,-0.1736939251,-0.1504089236,-0.1342989355,0.0950825363,0.3430252671,-0.0794039071,-0.128579393,-0.2427508533,-0.2310128659,0.6502943039,0.4262381494,-0.1440830678,-0.4783772826,-0.3731835485,0.2676646113,0.0803002119,0.2381606102,-0.0641208217,-0.0984878987,0.0306084044,0.1129468232,-0.2559910119,-0.0557872616,-0.4407341778,0.0071715922,0.2117857784,0.1263764203,0.2143028975,-0.0947062522,-0.0827281326,0.1265942156,-0.0557577945,-0.1126358956,0.0449620858,0.1506276131,0.0469185263,0.1951721609,0.1735437512,-0.4588237107,0.2376364022,-0.2655067444,0.1753760129,0.41826877,-0.0454813391,-0.3415568471,0.0746578425,0.4974749386,-0.1709237695,0.1422317028,0.1285944879,0.154649049,-0.2296802551,0.184286952,0.1251479387,0.1417401433,0.236686185,-0.1120783314,0.2201440632,0.0385219157,-0.1545711309,0.0684712753,-0.0448022932,0.2478228807,0.2652956843,-0.0790446252,-0.0293991826,0.0338054039,0.0698838606,0.2173097134,-0.2394904494,-0.1172485054,0.2396288067,-0.0299394131,0.1197568104,0.5013854504,0.3876144588,-0.3748737276,0.0745419413,0.0923809335,0.5648558736,-0.2601763606,0.0413093716,0.133695066,0.0506068207,0.0030422173,0.1490101218,-0.0633012503,0.3891547918,0.3070395887,-0.1713535339,0.2159722894,0.2408117056,-0.1208608598,-0.1497628689,-0.4051087797,0.0863526687,0.0817287043,0.0425829925,0.1304463148,-0.3192768693,-0.0476001762,-0.3641909659,-0.1965409964,-0.22491844,-0.1132412404,-0.4139758348,-0.010649167,-0.0181167759,-0.0879318193,-0.0273674056,-0.1237477139,0.3569250107,-0.385168761,0.3071930408,0.0091478787,0.0622749887,0.2990060449,-0.2994197011,0.4684779346,0.1641292274,-0.3684591353,0.037796393,0.2732798159,-0.2538939714,0.299712956,0.171287179,0.678319037,0.2227468789,-0.0918961391,0.0675157979,0.0342152268,0.0763492212,-0.3394201696,-0.0230132863,0.0669182613,-0.1332622319,0.1145668924,0.1374943554,-0.0944960862,0.2236462384,0.0117416503,0.2606374323,-0.2572070956,-0.1200354174,-0.1719239354,-0.1995496005,-0.1337638348,0.0917495266,-0.2527796328,-0.2110942006,0.3524641693,-0.0509155765,0.183152765,0.1991884857,0.067212522,-0.0960669443,0.271378696,0.3381310403,-0.5006517172,-0.3073892891,0.0890709236,-0.5300481319,0.2179004252,-0.2879717052,-0.0791888162,-0.0316296592,0.216260314,-0.0353243575,-0.1100046337,0.0923833996,-0.0869553164,-0.0684573874,-0.1544490755,-0.2633790076,-0.2788788676,-0.0878155828,-0.2979598939,0.2715916932,-0.5973241329,-0.0476713628,0.1096985787,0.0626047924,0.0808959603,-0.0069344779,0.2312024087,0.4421150684,0.3236790001,-0.0014010575,0.4533215761,0.0691781789,-0.1578182876,-0.3555991352,-0.0512277186,-0.1567316502,0.0503353998,0.1229820624,0.2397244871,-0.6420225501,0.0890365317,-0.2371957451,0.2477700412,0.0678071231,-0.0797343701,0.0980477482,-0.0238934793,-0.3431888819,0.1982951909,0.2360748947,0.2301772684,-0.0658628494,0.3270435035,-0.1933472306,-0.1563149393,0.4829874635,-0.1396929771,-0.3495692015,-0.2246772647,0.4760475159,0.0577694811,-0.040422637,-0.1970507354,-0.0889563486,0.185353756,0.0791050568,0.063200742,-0.1794749498,-0.2648894787,-0.2492375225,-0.0139695071,0.2091803253,-0.312114656,-0.3135881722,0.0350733325,-0.2094392478]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1784","title":"JSONDecodeError on JSON with multiple lines","comments":"Hi !\r\n\r\nThe `json` dataset script does support this format. For example loading a dataset with this format works on my side:\r\n```json\r\n{\"key1\":11, \"key2\":12, \"key3\":13}\r\n{\"key1\":21, \"key2\":22, \"key3\":23}\r\n```\r\n\r\nCan you show the full stacktrace please ? Also which version of datasets and pyarrow are you using ?\r\n\r\n","body":"Hello :),\r\n\r\nI have been trying to load data using a JSON file. Based on the [docs](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files), the following format is supported:\r\n\r\n```json\r\n{\"key1\":11, \"key2\":12, \"key3\":13}\r\n{\"key1\":21, \"key2\":22, \"key3\":23}\r\n```\r\n But, when I try loading a dataset with the same format, I get a JSONDecodeError : `JSONDecodeError: Extra data: line 2 column 1 (char 7142)`. Now, this is expected when using `json` to load a JSON file. But I was wondering if there are any special arguments to pass when using `load_dataset` as the docs suggest that this format is supported.\r\n\r\nWhen I convert the JSON file to a list of dictionaries format, I get AttributeError: `AttributeError: 'list' object has no attribute 'keys'`. So, I can't convert them to list of dictionaries either.\r\n\r\nPlease let me know :)\r\n\r\nThanks,\r\nGunjan","comment_length":49,"text":"JSONDecodeError on JSON with multiple lines\nHello :),\r\n\r\nI have been trying to load data using a JSON file. Based on the [docs](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files), the following format is supported:\r\n\r\n```json\r\n{\"key1\":11, \"key2\":12, \"key3\":13}\r\n{\"key1\":21, \"key2\":22, \"key3\":23}\r\n```\r\n But, when I try loading a dataset with the same format, I get a JSONDecodeError : `JSONDecodeError: Extra data: line 2 column 1 (char 7142)`. Now, this is expected when using `json` to load a JSON file. But I was wondering if there are any special arguments to pass when using `load_dataset` as the docs suggest that this format is supported.\r\n\r\nWhen I convert the JSON file to a list of dictionaries format, I get AttributeError: `AttributeError: 'list' object has no attribute 'keys'`. So, I can't convert them to list of dictionaries either.\r\n\r\nPlease let me know :)\r\n\r\nThanks,\r\nGunjan\nHi !\r\n\r\nThe `json` dataset script does support this format. For example loading a dataset with this format works on my side:\r\n```json\r\n{\"key1\":11, \"key2\":12, \"key3\":13}\r\n{\"key1\":21, \"key2\":22, \"key3\":23}\r\n```\r\n\r\nCan you show the full stacktrace please ? Also which version of datasets and pyarrow are you using ?\r\n\r\n","embeddings":[0.0756989047,-0.004834638,-0.0292827431,0.4874006808,0.3547834456,0.1495195478,0.4615655243,0.0634569228,0.1587526202,0.0187912937,0.1641427726,0.2806328535,-0.0464724079,0.2131021619,0.077341415,-0.2579198778,0.0545295775,0.2539742589,0.2304694355,0.093923308,-0.1753567457,0.2821831107,-0.1558834761,-0.0859780163,-0.1164932847,0.1821443141,-0.1478194445,0.2230660915,0.1152197272,-0.6116786003,0.199168548,0.1072017103,0.1027609408,0.4384293556,-0.0001222827,0.1784515083,0.2798017859,-0.2406275123,-0.437752068,-0.2415917069,-0.320045948,-0.2146712393,0.3941478133,-0.216466099,-0.0171037875,-0.5522251725,-0.1071552709,-0.2231709659,0.8473547697,0.3154249489,0.0736973509,-0.0370574035,0.1545831263,0.0937757269,0.1382793635,0.3976793885,0.1040199548,0.3903293014,0.1572732627,0.3830772638,0.0803718418,-0.1940742433,-0.0791405886,0.1125906631,0.2857994139,0.1097656786,-0.0717052966,-0.0807544291,0.0151098603,0.4550840259,0.6541770697,-0.1046296731,-0.4009039104,-0.4307039976,-0.1288366765,-0.1373524815,0.2610408366,0.1544523239,-0.1068272069,0.1755381823,-0.2287271768,-0.1890083104,-0.253189832,0.258387804,-0.0928343758,-0.3173473179,-0.2160383761,0.2146330774,0.1909279376,-0.1639520824,-0.1476037353,-0.2164838314,-0.3134977221,0.1610168219,-0.0158154685,0.0244269148,-0.1692642719,-0.2996096313,0.3776483238,0.0606462993,0.2912250161,0.0090573318,-0.3051211834,0.1320208758,0.4233737886,-0.0412974209,0.4679245651,-0.2059043348,0.0228297338,-0.0509501658,0.2377184778,-0.2316978425,-0.1745301634,-0.0246609468,0.0667581111,-0.2337434739,0.1367664486,-0.1931565702,0.0409572013,-0.1141456068,-0.3299007714,-0.2379553616,-0.0778570026,0.2825756967,0.3413724899,0.1305530369,0.0085191252,0.4388341904,0.151241526,-0.1779143363,0.0612650961,0.03361414,0.1091153249,0.1203234196,0.2002600133,-0.1847383827,0.1546734571,0.1673329771,-0.031202672,-0.1802987605,-0.22481516,0.0954530612,0.0053398209,0.2034656852,0.1514878124,0.0381725281,0.0465457104,-0.3754882812,-0.1229787245,0.2614931464,-0.3664222658,-0.1086113676,-0.2158961743,0.0045943297,-0.2165379226,-0.042754326,-0.7504179478,0.2429645956,-0.0581944734,-0.3593727946,0.0278873239,-0.1414280534,0.1401708722,-0.289802134,0.0539408103,0.484359175,-0.5632103682,-0.3126644492,0.290502429,-0.1905960143,-0.1055358797,0.3626500368,-0.410725832,0.2188340575,-0.2073307037,0.3010171056,0.3687204421,-0.3092291951,-0.2301794887,0.5485827327,0.1091068685,0.3371363878,0.2294214517,-0.166246891,0.2583563626,-0.0088671762,0.2358898371,0.1111461073,0.139396593,0.0451238118,0.1017302871,-0.0208238997,0.3151668608,0.2690217197,-0.3838232458,0.0342161693,0.0319788903,0.0770256668,0.3040084541,-0.2865880132,-0.1244976372,0.300216943,-0.2268988639,0.0445505753,0.0317665003,0.1740537286,-0.7827896476,0.0318350345,0.1511203051,-0.1864999384,-0.3102796376,-0.0598722361,-0.0096396944,0.018044956,-0.459142983,-0.0520047881,-0.043289952,0.0529062413,0.1861691922,0.0698304251,-0.1512107551,0.26028198,0.0843929574,0.1272289455,-0.2944525778,0.3275438249,0.113003619,-0.1274397075,0.0821474716,-0.0147800939,0.1743901968,-0.3622354865,-0.0237970762,0.3017584383,0.4365182817,-0.2380350083,-0.1748333722,0.0855652019,0.2663771808,0.2200504243,-0.0744529739,0.2505381703,0.223716557,-0.2906742692,-0.3921757638,0.2784978747,-0.088645637,0.4363918304,0.0388432629,-0.1501077861,0.423946172,0.1045752615,-0.2950928807,-0.1024177745,0.1572768241,0.1999422461,0.0250990391,0.0371618345,-0.3032545745,-0.193817094,0.5734389424,-0.1982163042,0.1366867721,0.3101391792,-0.1643854976,-0.122413367,0.0132789621,-0.023327589,0.3320555389,0.1498828232,-0.3159901798,0.0376023762,0.2749466002,-0.2184944153,0.1245727986,-0.1084819809,0.2380923778,0.0401965491,0.1783394068,-0.0797343478,-0.330512464,-0.4385953248,0.1077532545,0.0704088658,-0.5264180303,0.113750726,-0.4291846454,-0.0921923444,-0.0660933629,-0.2417521179,-0.1806002408,-0.2090834528,-0.1459433138,0.068922624,0.0243189838,-0.0922085494,0.0943390876,0.1706511229,0.1414880604,-0.1998651624,-0.1468555778,-0.1468743831,-0.2379731387,-0.0691627041,0.2437769026,-0.0099238567,0.007075578,-0.3262991309,-0.1477711946,-0.1586072445,-0.0661571175,0.109212473,-0.1067405641,0.1586824656,0.146321699,0.482786715,-0.0861451924,-0.0333815441,0.3210743368,0.2586284578,-0.2111027539,0.2427027076,0.1239626259,-0.0303509329,-0.0638520271,-0.0726980343,0.0486168303,-0.1663968116,0.6842475533,-0.0452208295,0.1938864738,-0.0541440547,0.1299834996,0.0972091407,-0.2045085877,0.0939908549,-0.1179483831,-0.2161597461,0.2748553157,-0.1578396261,-0.4564421773,0.0866717771,0.1376078129,-0.1101882309,-0.0661053807,-0.4292573929,-0.1951761991,-0.0504839197,0.1136903092,-0.1206954196,-0.0474421643,0.2318428457,-0.0019965582,0.0809704736,-0.0558322445,-0.3087209761,0.1214718446,0.0186988898,0.0840990543,0.2554667294,0.7511204481,-0.1643750668,0.1048239917,0.4095941186,-0.0908322558,0.3358144164,-0.2318022549,0.2796764076,-0.2613416314,-0.0558581464,-0.1485669166,-0.1008694842,0.0553968213,0.2300320268,-0.0044680801,0.101144515,-0.1548943818,0.0414446406,-0.4394150376,-0.1142980829,0.1158775911,-0.0837005824,-0.2346474528,-0.2505991459,0.0827206075,0.0823536217,0.1297953874,-0.2238012403,0.4012429416,0.290909797,0.01087943,-0.482331574,-0.178270936,-0.2993755341,0.3115851879,0.0724723563,0.3987670541,-0.0525590926,0.3056712747,0.1513424516,-0.1640331447,0.8771930933,-0.0720194355,-0.1486297697,0.1351096779,0.0790693536,-0.4990383685,0.0231587868,-0.0832152814,0.3365105987,0.1128194183,0.1889238507,-0.5262200236,-0.1411768049,-0.1502993256,0.1219519526,-0.1306488812,-0.3131032884,-0.1105352268,-0.4094940722,-0.548758924,-0.2788081467,0.0503897518,0.2496143579,0.0547449179,-0.1173462719,0.0764079392,-0.2830381691,0.0264884885,0.100623697,0.1723232418,0.1050178856,-0.0677216053,0.116044201,0.0810785294,0.6050027609,0.6168935895,0.220040977,-0.5725251436,-0.1515591443,-0.3779136837,0.3386909664,0.1236374751,0.0089898407,0.0958796665,-0.1069691181,-0.0281355903,0.0381140485,-0.2126393914,0.4485839605,-0.1523486674,-0.3894649744,-0.4445132613,0.5677691698,0.022304466,-0.0151191819,0.0311268214,0.2863840461,-0.4327487051,0.3331761658,-0.2572877705,0.5948890448,-0.0670593306,0.3598081172,0.4278510809,-0.2802689672,0.5248782635,-0.0857867151,-0.0531439446,-0.1445716321,0.1677118093,-0.1333737969,-0.1318366379,0.2990582287,0.4635649621,-0.2111042142,0.19945696,-0.3796827793,-0.0257816426,0.021922674,0.3363604248,-0.0447592475,-0.2403125614,-0.6766855121,0.01428736,-0.0731552765,-0.2198832184,-0.0554788075,-0.176281333,0.0918210149,-0.3764566481,-0.3643436134,-0.0236470904,0.0180888176,-0.0852237269,0.2905055583,-0.2800568938,0.3040168285,0.2946938872,0.0030285858,-0.1377632618,-0.1765669286,0.0029854106,-0.3240697384,-0.1500299424,-0.3393163681,-0.0613905005,0.189865917,-0.0163223036,-0.2327295393,0.1667875946,-0.1638559103,-0.4918404222,-0.0593733005,0.250664264,-0.5110679269,-0.4745010734,-0.3947389126,-0.1168439314,0.0519515201,-0.0442528576,-0.0072201532,0.2400055677,-0.0497964136,-0.1048772112,0.0629913658,-0.0384448655,-0.019588111,0.1853263229,0.0224057492,-0.1815966368,0.6452671885,0.4009434283,0.0333054475,-0.1279838532,0.3129695058,0.6787834764,-0.1460890174,0.3107436597,0.1308354437,0.4295876622,-0.070420742,0.3472555578,-0.0749998465,0.1801513135,0.3063793778,-0.6065922976,-0.0971379876,0.259836942,0.0545560569,0.1040454805,0.2366100699,-0.0283830743,0.0236271508,-0.0122389933,-0.1246076226,0.3266866207,0.0775351077,0.0731420591,0.1894127727,0.1207098216,0.2873913348,-0.0857672468,-0.0112722963,0.086157687,0.005568298,-0.0671929941,-0.0116523504,0.1908575296,0.2449630201,-0.1065154001,-0.2095552087,-0.1429029256,-0.0313249081,-0.1739067882,0.3243989348,0.064442955,-0.1287902147,0.210099265,0.3157557249,0.1497428268,-0.3055717349,0.1047445163,-0.1872396171,0.5721406341,-0.021531228,0.0584521331,-0.1145423427,0.0733322054,-0.2476494312,0.3973836899,0.1267005354,0.0646819025,0.2737924755,-0.4780191481,0.2461462468,0.0544703379,0.1764316261,0.5079083443,-0.4632913768,0.2767193317,-0.087151587,-0.0323948003,0.0746794268,-0.1712875962,0.3691149354,0.1350059211,-0.1295898706,0.1843113452,-0.0573189184,-0.1345751882,-0.0280409791,-0.1097541302,0.2922796309,0.1591729969,-0.0371067114,0.4146482348,-0.1724990308,-0.078438133,0.0897987559,0.3331101835,0.2535051107,0.3189611435,-0.2833828628,-0.0099925529,0.2608397901,0.1258873343,-0.0723958984,-0.38774845,-0.0550847948,0.3519828022,0.0916911811,0.1211039051,0.1613054723,0.1335014999,0.0547608621,0.1395141333,-0.1904909909,0.0833459422,0.0102805197,-0.2840882242,-0.0998743251,-0.1132306606,-0.1254787296,0.0288863722,-0.2928874493,-0.2805137634,0.1353381425,0.0819415823,0.0951075181,-0.4350188971,0.3131980002,-0.0405780859,-0.0153842112,-0.2048435807,0.2946709991,0.3339356482,-0.159625113,0.576992929,0.009086905,0.5483865738,0.2124696821,0.1156956106,0.0132708391,-0.4909027815,-0.0946568251,0.2071464807,0.3553981781,0.0508531258,0.0046821646,0.2090478837,0.0558940209,0.086975947,0.3290416002,0.0325114727,0.1568682194,-0.3156081736,0.0302328486,-0.3788301349,-0.0873625278,-0.1211289763,-0.1011806726,-0.1344500184,-0.4340040982,0.1330700219,0.1707014591,0.0793333352,0.0714916661,-0.0028354893,-0.1310885549,0.4665918648,0.3894787431,0.2655844092,0.0237389542,-0.0335917398,-0.165628776,-0.2036831975,-0.2785916924,-0.1370723695,0.2476901412,0.0628393441,0.2403697371,0.257714361,-0.0058911131,0.2844989598,-0.3355717063,0.0001502599,-0.2361418903,-0.2020061165,0.3589678407,0.2450478375,-0.1921957135,0.0815471932,0.3909071386,-0.0826663747,-0.0857479945,-0.218266204,0.3969542086,-0.2062812746,-0.1763969362,0.237464726,0.2302735001,0.5443312526,0.1605999321,-0.0869177058,-0.0404684059,-0.1670814157,0.0135559496,0.6156164408,-0.0370027572,0.4068597853,-0.2951594591,0.0592503101,-0.1242647097,0.0251256414,-0.4160119891,-0.0127061838,-0.3439686894,0.2034494728,-0.0015565152,-0.1828735471,0.0542489365,0.0863015279,-0.1540883034,0.0811340362,-0.0142950974,-0.1092115417,0.2355830371,-0.3649301231,-0.1855000108,0.1340004206,-0.0453587808,-0.1485257149,-0.0898531377,-0.539095819,0.1714611948,0.4368406832,-0.0269895159,-0.5867701173,0.1004967093,-0.0533172488,-0.175578922,-0.152948007,0.3230693936,0.0704886764,-0.2000647038,-0.1911253631,-0.3086665571]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1784","title":"JSONDecodeError on JSON with multiple lines","comments":"Hi Quentin!\r\n\r\nI apologize for bothering you. There was some issue with my pyarrow version as far as I understand. I don't remember the exact version I was using as I didn't check it.\r\n\r\nI repeated it with `datasets 1.2.1` and `pyarrow 2.0.0` and it worked.\r\n\r\nClosing this issue. Again, sorry for the bother.\r\n\r\nThanks,\r\nGunjan","body":"Hello :),\r\n\r\nI have been trying to load data using a JSON file. Based on the [docs](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files), the following format is supported:\r\n\r\n```json\r\n{\"key1\":11, \"key2\":12, \"key3\":13}\r\n{\"key1\":21, \"key2\":22, \"key3\":23}\r\n```\r\n But, when I try loading a dataset with the same format, I get a JSONDecodeError : `JSONDecodeError: Extra data: line 2 column 1 (char 7142)`. Now, this is expected when using `json` to load a JSON file. But I was wondering if there are any special arguments to pass when using `load_dataset` as the docs suggest that this format is supported.\r\n\r\nWhen I convert the JSON file to a list of dictionaries format, I get AttributeError: `AttributeError: 'list' object has no attribute 'keys'`. So, I can't convert them to list of dictionaries either.\r\n\r\nPlease let me know :)\r\n\r\nThanks,\r\nGunjan","comment_length":56,"text":"JSONDecodeError on JSON with multiple lines\nHello :),\r\n\r\nI have been trying to load data using a JSON file. Based on the [docs](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files), the following format is supported:\r\n\r\n```json\r\n{\"key1\":11, \"key2\":12, \"key3\":13}\r\n{\"key1\":21, \"key2\":22, \"key3\":23}\r\n```\r\n But, when I try loading a dataset with the same format, I get a JSONDecodeError : `JSONDecodeError: Extra data: line 2 column 1 (char 7142)`. Now, this is expected when using `json` to load a JSON file. But I was wondering if there are any special arguments to pass when using `load_dataset` as the docs suggest that this format is supported.\r\n\r\nWhen I convert the JSON file to a list of dictionaries format, I get AttributeError: `AttributeError: 'list' object has no attribute 'keys'`. So, I can't convert them to list of dictionaries either.\r\n\r\nPlease let me know :)\r\n\r\nThanks,\r\nGunjan\nHi Quentin!\r\n\r\nI apologize for bothering you. There was some issue with my pyarrow version as far as I understand. I don't remember the exact version I was using as I didn't check it.\r\n\r\nI repeated it with `datasets 1.2.1` and `pyarrow 2.0.0` and it worked.\r\n\r\nClosing this issue. Again, sorry for the bother.\r\n\r\nThanks,\r\nGunjan","embeddings":[0.1118792072,-0.0649543628,-0.0127836056,0.509495914,0.3668510318,0.1452076733,0.4116669297,0.0697856471,0.116488494,0.0775828362,0.1227522716,0.3139997125,-0.0133123985,0.1743163913,0.1052531078,-0.3321564198,0.056305293,0.242220372,0.2812244892,0.0448313579,-0.1883261949,0.3222274482,-0.1211521253,-0.021525478,-0.1003587916,0.2009462863,-0.1490634829,0.2098397315,0.0983382314,-0.5983390808,0.1811593324,0.0787486359,0.0896519274,0.4094790816,-0.0001244392,0.1713657826,0.2335566431,-0.2385098338,-0.4193005562,-0.2048569024,-0.2344136983,-0.1837661862,0.4074413478,-0.2153499424,-0.0434500948,-0.6132315397,-0.1171405688,-0.057215374,0.8382027149,0.2899007499,0.0557004213,0.0221735276,0.2437990606,0.0981308147,0.1929415911,0.3184738457,0.1145073026,0.4430990219,0.2087321728,0.338168025,0.0679094046,-0.1586386412,-0.0711870641,0.0835569054,0.280957669,0.1195681542,-0.0219780076,-0.1146162972,-0.0106241079,0.4550395012,0.6182894111,-0.1244438142,-0.3540066779,-0.4397033155,-0.1298505664,-0.2184529155,0.2876081169,0.1772159338,-0.1519283801,0.1779494435,-0.1595244706,-0.1754688919,-0.3075227141,0.3047269881,-0.0470938049,-0.2925513089,-0.1445957422,0.2423642874,0.2514731884,-0.1574359983,-0.1731507629,-0.1707074642,-0.3005772233,0.1942362934,-0.0374578722,0.0332614407,-0.1622178853,-0.2602837682,0.3593869805,0.0219991133,0.2777704895,-0.0156081272,-0.2210023552,0.1750971228,0.3867213726,-0.0310707763,0.3983153701,-0.2118357122,0.0357383192,-0.0887702927,0.1991060227,-0.1957683861,-0.1575291157,-0.0030412157,0.0626828149,-0.2527766824,0.1424385309,-0.2047821879,0.0941172689,-0.128095746,-0.3872186244,-0.2571391463,-0.1042992398,0.3170131147,0.3308335245,0.0327450186,0.0213842969,0.4348160028,0.1162052304,-0.2527853251,0.0686020181,0.0260868575,0.1068151295,0.1247438118,0.1768325269,-0.2131353468,0.1577491909,0.1446985006,-0.020176921,-0.1401219964,-0.2365128398,0.1040509418,-0.0314321332,0.2174638808,0.0933527872,0.0051284656,0.065684773,-0.3857875168,-0.1371872276,0.2650327086,-0.357152909,-0.1853625178,-0.2478067875,-0.0130293518,-0.2101794779,-0.0484983437,-0.7054870129,0.2720554173,-0.0497150049,-0.3916545808,0.0422807261,-0.1151323467,0.2001676857,-0.2836830914,0.0393819921,0.4265245795,-0.5727180839,-0.3258422911,0.3022897542,-0.201630488,-0.0620368347,0.3925494552,-0.4634450674,0.2692922056,-0.1763111651,0.3346166909,0.3625286222,-0.3054618537,-0.2864094377,0.476051569,0.0608116463,0.3308361769,0.2555824816,-0.1510689557,0.2409437746,0.0393104143,0.2559550405,0.0936086625,0.1579339653,0.0332766436,0.0208977256,-0.0401273556,0.3830362558,0.2622578144,-0.3467324078,0.0336253196,0.0298387222,0.0210787412,0.3198065162,-0.3416233063,-0.1486139148,0.3134537637,-0.1775080264,0.0725343972,0.0418675579,0.1931006163,-0.7275482416,0.0468464419,0.0616339333,-0.0923423618,-0.328135699,-0.0458588116,-0.0145272473,0.0376030318,-0.4042335451,-0.0969732851,-0.0637844577,0.0427362137,0.2181390673,0.1092857569,-0.1538801491,0.2307792157,0.0312642269,0.1312134564,-0.3026619554,0.372982949,0.0696923211,-0.1885002553,0.0212982036,-0.0272981878,0.206064865,-0.3279631436,-0.0248574186,0.2660458088,0.3668588102,-0.2670189738,-0.1989662945,0.0652980506,0.2587857246,0.0962710753,-0.0828846321,0.2987167537,0.2130886018,-0.2320663184,-0.3045404255,0.3062797487,0.0069992328,0.4205735922,0.0265317839,-0.1826451123,0.3832916319,0.1058386415,-0.2332844585,-0.0060083168,0.2103117108,0.2088834792,0.0312987082,0.0615092516,-0.3478973806,-0.1770059019,0.5678311586,-0.2046071887,0.112140663,0.3307668567,-0.1569916606,-0.0826685578,0.0503760092,-0.1266833395,0.3211077452,0.1670219898,-0.3652741611,0.0371902436,0.2347430587,-0.2011923045,0.1160394996,-0.0974501669,0.2182054222,0.0531891771,0.2291597575,-0.0788060874,-0.4170311987,-0.4203803241,0.1339421272,0.110830456,-0.5512763858,0.1252056658,-0.3722783625,-0.0822068155,-0.0405491106,-0.3005830944,-0.2753337622,-0.1956277639,-0.0834283903,0.0990590006,0.0312615968,-0.0518527552,0.1712452471,0.1700271368,0.1119886413,-0.245759815,-0.2071046233,-0.2123643309,-0.2687690854,-0.0905148014,0.2157162875,-0.0490054823,-0.0251578577,-0.3031786978,-0.2045316249,-0.1507426798,-0.1372845322,0.0808488876,-0.170614928,0.2122220993,0.0848781765,0.4887607396,-0.0839299336,-0.1283809543,0.3590871394,0.2363953292,-0.2614621818,0.2470536977,0.0831591785,0.018430192,-0.0464081392,-0.0690745115,0.005278151,-0.1544251442,0.7430199981,-0.0833438262,0.2016305178,-0.0415605605,0.1645389646,0.1299038827,-0.2662785053,0.100053072,-0.1210284382,-0.2139341831,0.2681325674,-0.1243895292,-0.4402165413,0.072198689,0.1469231844,-0.0704936609,-0.0599183403,-0.4526904821,-0.2472373396,-0.038853351,0.2286962718,-0.1373573691,-0.0575727709,0.2596032619,-0.0101235891,0.0912241712,-0.0265845638,-0.334239006,0.117329061,0.0551325083,0.1163617969,0.1643093973,0.6633324623,-0.2180379182,0.2359458506,0.3052803576,-0.0470109694,0.3674233854,-0.2426069677,0.259614557,-0.2432775795,-0.0913032293,-0.126805529,-0.1253752559,0.0130448202,0.1845621169,-0.0201930013,0.0507537499,-0.1736587435,-0.0452298857,-0.4609677792,-0.1012778506,0.1402494162,0.0856636167,-0.2287009656,-0.221210748,0.0866021067,0.0337409265,0.1150739267,-0.1839328706,0.402884692,0.3039779365,0.0090632774,-0.4590246379,-0.2491078377,-0.2739679515,0.3141450286,0.0809369534,0.3002000749,-0.0620871447,0.2909625769,0.2164378166,-0.1976756155,0.9261081219,0.0026831655,-0.1659577191,0.1430734098,0.0866749138,-0.4900355339,0.0012997066,-0.051439967,0.254128933,0.1277290136,0.20146285,-0.6062143445,-0.1465935856,-0.1256859303,0.1696174294,-0.1325752288,-0.3024128377,-0.1702273339,-0.4666592777,-0.5862988234,-0.218121469,0.0285962988,0.2952228785,0.0815350264,-0.0359525979,0.0951209217,-0.2656361461,0.0024546061,0.1268087327,0.1854035556,0.0804401934,-0.0903359056,0.0820388198,0.0542672873,0.5888195038,0.6108232141,0.2106247991,-0.6033201218,-0.1698660254,-0.3187557757,0.4119630456,0.1917922348,-0.0333697535,0.1753222346,-0.1057091728,-0.0074270633,0.0734502673,-0.2139061838,0.4310311675,-0.13500534,-0.3312546909,-0.3607971072,0.5874239802,0.003359681,-0.0127398232,0.0707016736,0.3162629306,-0.3953839242,0.3670030236,-0.2552439868,0.716647923,-0.0982211009,0.3945844769,0.4653698504,-0.2684081495,0.5387393832,-0.1318394095,0.0256851576,-0.1567891538,0.1508012861,-0.1145974025,-0.1021996066,0.286812067,0.4203557074,-0.1900787652,0.1530015618,-0.3553855717,-0.0051560802,0.0449537039,0.2930894196,0.0029418324,-0.2236181945,-0.8049345613,-0.0073047471,-0.0906870589,-0.1672313362,-0.0486390032,-0.2482415438,0.0377590433,-0.4397956133,-0.3008874059,-0.028953433,-0.0417473875,-0.0845264047,0.3456767499,-0.3072955906,0.2221971899,0.3093556464,-0.0634282455,-0.1697472483,-0.2068005651,0.0339385532,-0.2961464822,-0.2027090341,-0.3475338519,-0.0442654453,0.1643640995,-0.0564825833,-0.2733549476,0.2610814571,-0.2105809301,-0.5443157554,-0.0701415762,0.2332809269,-0.4967075288,-0.4386483133,-0.312734127,-0.1415804923,0.035083469,-0.08484824,-0.0190849788,0.198903814,-0.1074971408,-0.1280622184,0.0554453284,-0.078976512,-0.0511300936,0.2231666744,0.0287353508,-0.1254391223,0.685685873,0.3832620084,0.0582152978,-0.1036356091,0.2032696754,0.6720959544,-0.2489135712,0.3087472618,0.1828707308,0.431840241,-0.1328354925,0.4236624241,0.0219976138,0.1273991913,0.3115240633,-0.5580520034,-0.1535034925,0.3016931117,0.0815705508,0.0776998475,0.2658526003,0.0633564964,0.0652709529,0.0579958558,-0.1272093505,0.335377723,0.0415443741,0.0779568925,0.2087285519,0.0826460272,0.3267770112,-0.0425710157,-0.0193542968,0.1076976508,-0.0010706006,-0.0558862127,-0.049148839,0.201855734,0.241983518,-0.0960251912,-0.1949386597,-0.153912425,0.0119558908,-0.1337751895,0.3213222623,0.0384281762,-0.0873596221,0.1955226511,0.3714819252,0.1643387377,-0.3251362145,0.1141154245,-0.163846001,0.6059881449,0.0469971374,-0.0119016943,-0.1791574955,0.0803646967,-0.2306517661,0.3374820054,0.0988278911,0.000154498,0.2705645859,-0.4505706429,0.1796679795,-0.0524169654,0.2017089576,0.4938368499,-0.458525598,0.3138849437,-0.0284848958,-0.0570586324,0.0034637474,-0.1709434986,0.4224455357,0.0875827447,-0.1638118476,0.1700978279,0.0028253705,-0.133695513,0.0171125215,-0.1054552719,0.2873849869,0.200653106,-0.0894223154,0.431250155,-0.1834622324,-0.0569862686,0.1248561591,0.3617696762,0.2412105054,0.2141115516,-0.3609218299,-0.0355872661,0.1423890442,0.1633582115,-0.1051959246,-0.3548905253,-0.0279864352,0.3401579857,0.20293805,0.0616271123,0.1194150522,0.2362374663,0.1331353635,0.1293346882,-0.1833067387,0.2270361036,0.030860398,-0.3004808724,-0.0711387247,-0.1158107221,-0.0918033943,0.0500390008,-0.2572751045,-0.3297420442,0.1884557456,0.1104962304,0.0763768852,-0.4972600043,0.1957602501,0.0194555651,-0.0014934689,-0.2102148086,0.3131484091,0.3145588338,-0.2211554646,0.5350205302,0.0256423634,0.5477107167,0.2180265337,0.0665232912,0.0182383731,-0.5012834072,-0.044291418,0.1924670488,0.3591532111,0.1176426411,0.0810038745,0.210943684,0.0274811834,0.057717558,0.2730031013,0.0655142441,0.1863233447,-0.3028700948,-0.0369334146,-0.3501702845,-0.1001455784,-0.1512101889,-0.1170870364,-0.1513699442,-0.3333923817,0.0944498107,0.2090520263,0.067927517,0.1047090963,-0.0146007389,-0.1393784583,0.5112262368,0.4039310813,0.3075303435,-0.0957833901,-0.0280130319,-0.2047959417,-0.1363404244,-0.3467086256,-0.1190747693,0.2723594904,0.086754337,0.263931185,0.2851451039,-0.0107749617,0.2542973459,-0.3200688362,0.0292569585,-0.3025755286,-0.1702691019,0.3497337401,0.2315204442,-0.1659980863,0.1436178982,0.4026559293,-0.0710459575,-0.0964014307,-0.2188056558,0.3641434312,-0.1985043585,-0.0947704315,0.310470283,0.1703802794,0.5703426003,0.1612477899,-0.1028225273,0.0325654224,-0.1898340434,-0.0179046541,0.5546193719,0.0107954908,0.3781962395,-0.2969704568,0.0308743138,-0.180038929,0.0773290396,-0.3476775587,-0.0699006841,-0.2934139073,0.2021338195,0.0524849258,-0.1814599186,0.0146753555,0.0893942416,-0.1310040653,0.0649524406,-0.1094191447,-0.1019570976,0.3342763782,-0.4352563918,-0.1552677155,0.1767790467,-0.0317634158,-0.1027403548,-0.1141519099,-0.5677464008,0.1682778001,0.446346432,-0.014063824,-0.5691127777,0.1102258936,-0.0618531294,-0.1592490822,-0.147410661,0.267237246,0.0738587379,-0.1737381816,-0.2226580679,-0.3186435699]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1783","title":"Dataset Examples Explorer","comments":"Hi @ChewKokWah,\r\n\r\nWe're working on it! In the meantime, you can still find the dataset explorer at the following URL: https:\/\/huggingface.co\/datasets\/viewer\/","body":"In the Older version of the Dataset, there are a useful Dataset Explorer that allow user to visualize the examples (training, test and validation) of a particular dataset, it is no longer there in current version.\r\n\r\nHope HuggingFace can re-enable the feature that at least allow viewing of the first 20 examples of a particular dataset, or alternatively can extract 20 examples for each datasets and make those part of the Dataset Card Documentation.","comment_length":21,"text":"Dataset Examples Explorer\nIn the Older version of the Dataset, there are a useful Dataset Explorer that allow user to visualize the examples (training, test and validation) of a particular dataset, it is no longer there in current version.\r\n\r\nHope HuggingFace can re-enable the feature that at least allow viewing of the first 20 examples of a particular dataset, or alternatively can extract 20 examples for each datasets and make those part of the Dataset Card Documentation.\nHi @ChewKokWah,\r\n\r\nWe're working on it! In the meantime, you can still find the dataset explorer at the following URL: https:\/\/huggingface.co\/datasets\/viewer\/","embeddings":[-0.3861365914,-0.3873980641,-0.0618844815,0.2974122167,0.1567154527,0.2989473045,0.0844833627,0.3552836776,-0.0643111393,0.3763421774,0.0037855837,0.2994420528,-0.2900573909,0.5288341045,0.3926576674,-0.2897698879,0.0115508502,0.2083534747,-0.1008172706,0.1148869991,0.0792255774,0.0855689645,-0.1743282527,0.1514346898,-0.4029159546,-0.1880390346,-0.3037720025,-0.0746549815,-0.2611444592,-0.5259741545,0.0975277424,-0.096980527,0.3257598579,0.0875246227,-0.0001094514,-0.1372068524,0.2247065306,0.0796680152,-0.4892878234,-0.059252806,-0.3846616447,-0.0151574807,0.3065688312,-0.160370186,-0.2138927281,-0.4064326286,0.2569763362,0.060766831,0.3277089596,-0.0574039184,0.1862078011,0.4664745033,-0.101199545,-0.309589535,-0.0577799752,0.5932620168,-0.3776036501,0.1418806612,0.4056650996,0.1719203442,-0.055873353,0.3942483366,0.4434017837,0.1074619591,0.0666836277,0.1224726513,-0.0140311671,-0.1412602812,0.035226807,0.6123383641,0.4841229916,0.1082843915,-0.3699077666,-0.4363870919,-0.0279569943,-0.1658657789,-0.1121710092,0.382439822,0.070182316,0.2538605034,-0.3750244081,0.1134655699,-0.1830792427,-0.0366972871,-0.0914128274,-0.0198506881,-0.3480017781,-0.071208328,0.4680240452,-0.0432211272,0.0942932814,0.2341705263,-0.0011738002,0.1652157605,-0.0263169371,-0.2085904032,0.1860803217,0.1761122644,0.1291079372,0.2292695642,0.1167370975,0.3246678412,-0.0312280841,0.1755796075,0.2116531134,-0.1055743247,0.4409830868,0.2119702101,0.1845704615,0.3149191439,0.0476228483,-0.042671524,-0.381049186,0.157037273,-0.3530599475,-0.2361382991,-0.0104282359,-0.169717446,-0.0877128765,-0.1154186949,-0.2391984165,-0.0636856928,0.1094625071,0.4093828499,-0.0947325677,-0.3719478548,-0.3213818669,-0.1324314177,0.0660622194,-0.3266577125,-0.1745932847,0.0009161785,-0.282733351,0.030079376,0.144810617,-0.5260006785,0.4474388063,-0.0701306015,-0.0689212009,0.0031333715,0.334326148,0.1553655118,0.3408134878,0.1037811339,-0.0048323544,-0.2238629311,0.155713439,0.0739415511,-0.2017610073,-0.0292448308,-0.1053102538,-0.3409497738,-0.3423970044,0.1137071028,-0.2802437544,0.0596916303,-0.3804872632,0.4780719876,-0.4661561549,-0.2010666579,-0.0501739606,0.1169522181,-0.255343765,0.0600068644,0.3718675971,0.2557033598,-0.3894895315,-0.3772592247,0.0182705354,-0.2124765813,-0.3901623189,-0.037462756,-0.344284147,0.3350663483,-0.5363284945,0.235228464,0.3367296159,-0.2228100151,-0.0518802553,0.0936631188,0.1573923677,-0.0549474545,0.0014534025,-0.1481589973,0.4459006786,-0.0305543151,-0.3026451766,-0.0790591314,0.0599405281,-0.0368748344,0.0009243836,-0.1719297171,-0.0129102254,0.0563326664,-0.1486153454,0.1010532603,0.1191323847,0.0383442864,0.4723883271,-0.2317597419,-0.1125413254,0.1484722197,0.2192695141,0.2432035655,0.0111545511,-0.4020672739,-0.1583313048,0.0949400812,0.0038213802,-0.180739522,0.30763188,-0.3185966313,-0.113110669,-0.0341547132,-0.2596442401,-0.2621262372,0.1584979594,-0.1365663111,0.1611819267,0.1251431406,-0.1892434061,0.08858978,-0.3987357616,0.187005043,-0.3901411891,0.2090323716,0.1105593592,0.0536062457,0.0102647813,0.0267262701,-0.0408343114,-0.2687215507,0.0686031654,0.221715048,0.1641524285,-0.1090132892,0.1630971432,0.4362353384,0.208985284,-0.3917674422,-0.0104676588,-0.1331777424,-0.1332426518,0.2064635158,-0.7598805428,0.3216645122,0.2521365583,0.1578047723,0.317338407,0.1152514145,0.023177091,0.111087881,-0.2549541593,-0.2837450504,0.1308498681,0.0769312978,0.1748405248,-0.0528696813,-0.6406166553,0.1478694379,0.2648247778,-0.0942745283,-0.0380185768,0.0963488072,-0.3271735609,-0.0201585311,0.4219538271,0.0524048917,0.0221044496,0.2686966658,0.2439734191,0.0857682675,-0.1344521195,0.1146751642,0.1895128489,0.0684709623,0.0004230356,-0.3579305708,-0.173208192,0.0783771053,-0.2876156867,-0.0603781566,0.1106496826,0.1175418198,-0.2738903463,0.0905591995,-0.3630374968,-0.4837638736,-0.069794178,0.0704646781,0.1552468687,-0.3996530175,0.0803671479,0.2161303461,-0.1143753901,0.0583186224,-0.162574172,0.3223586679,-0.297540158,0.7413788438,-0.1173184812,0.1161013916,-0.1537316442,0.1364964247,0.2291521132,0.3248565793,0.3808384836,-0.223133713,0.1128322482,-0.5781584382,-0.5961934328,0.4128218889,0.0946190879,0.2726972699,0.3713106215,0.1063176394,-0.4431185722,-0.0808980688,0.0142680882,-0.3306340277,-0.2247574329,-0.1559310108,-0.0645111352,0.1442426592,-0.0447174907,-0.1804035902,-0.1693991423,-0.1982709318,0.375235945,-0.0849383101,0.1998657435,0.1477505267,0.0412277952,0.375274986,-0.5501070023,-0.0154579962,-0.1189945638,-0.4080892801,0.0377525948,-0.6036459804,-0.3449813128,0.1304653734,-0.1481451243,-0.0367349349,-0.118149586,-0.2437891811,-0.2489361763,-0.2817431092,-0.046446465,-0.0438788831,0.0582123101,0.336622715,-0.0722353309,-0.0918443277,0.0653492138,-0.1439099461,-0.0627382174,-0.0116690723,0.1149907112,-0.1781760901,-0.0343371555,0.1868714988,0.3779286742,0.2188417315,-0.1248089671,0.0270542298,-0.4515987635,0.2974494398,-0.0454939976,-0.0707407594,0.1005067378,-0.1144086495,0.14883627,0.1586331278,0.22254619,0.3080058694,0.3026154339,0.28153494,-0.3162564337,-0.1943415105,0.054528363,-0.1732912362,0.2260578126,-0.1356072128,0.1330337077,-0.051331114,0.1063690931,0.0045788237,0.0369431861,0.2384206504,0.025827134,-0.3806378841,0.0398965068,0.0520673357,-0.0444802232,-0.0089074112,-0.1164747849,-0.2867601216,-0.1160019785,0.0673911273,0.2316898257,0.6677560806,-0.0412938185,-0.0051503093,0.0473208204,-0.471927762,-0.3972479701,-0.0349904597,-0.2830821276,0.0984149948,0.1577619165,0.4498451948,0.1115661711,-0.2531613708,0.6400966644,0.1472349018,-0.1924183667,0.1099449247,0.1065796241,0.1087524816,-0.1604260951,0.003370787,0.1266916245,-0.0600127839,-0.071454078,-0.2331012189,-0.3549361527,0.0225643478,0.0138484435,0.4331592619,0.3815614283,0.090909861,0.2579393387,0.5203465819,0.3753033578,0.3056423366,0.3549167812,0.1044133827,-0.120968163,0.3064055145,-0.0617571473,0.3310107887,0.1565058529,0.0389383174,0.1024665758,-0.0421241634,0.1390384734,-0.331756264,0.3228484988,0.2974020839,0.0894573107,-0.3194744587,-0.2341733724,0.560567975,0.2237866074,-0.1851388514,0.4620555043,0.4765681624,0.1422177106,-0.2373576313,0.0535312705,0.6876664162,0.0365229622,0.0843452886,0.2455760837,-0.171654135,0.5021948218,0.1209287122,-0.1125573441,-0.2956690192,-0.4778721929,-0.1147814915,-0.2216896266,0.2821293771,0.1816496402,-0.1988605559,-0.0246849786,-0.0100786444,0.1824443787,-0.0426988788,0.0637452751,0.015780583,-0.0726718009,-0.3819892108,0.1237992719,0.1878641993,-0.0032552474,0.009338107,-0.1006219611,0.0058429101,-0.0603996217,-0.4639254212,-0.2156301588,-0.1343782693,0.0016824683,0.2792566419,-0.2689447105,0.3096630573,0.3147808611,0.0312166549,0.0395608507,-0.2702919841,0.0169410035,-0.4660326838,-0.045832742,-0.1562336236,-0.0017328738,0.26501894,-0.1244602427,-0.2057224065,0.2733051181,-0.0144140208,0.0176642984,-0.0818944722,0.1602960825,-0.2555712759,-0.1735799611,-0.0668572262,-0.1034816056,0.0298794508,-0.2667164207,0.1227919012,0.2154372633,-0.1349743158,-0.0131823206,0.386621505,0.180819422,0.0047535067,0.1999901384,-0.1486938745,-0.00186888,0.2752651572,0.1902955621,0.0427181944,-0.2142770141,0.2983701229,0.4879491627,-0.218692109,-0.0052247737,-0.128059417,0.0626682714,0.1973385364,0.3569821417,0.199488759,-0.1900323182,-0.2914748788,-0.0581808761,-0.3802822828,0.2114995569,-0.118348226,0.4987813234,-0.179766342,0.0155375246,0.0540673248,0.1239064783,-0.319607228,0.2481544763,-0.3238674104,0.0638504475,-0.0027058793,0.299105227,0.1134036556,-0.3290360272,-0.0358923301,0.1201806739,-0.1838930696,-0.231904611,-0.3390771747,0.1232621446,0.0479894988,-0.001572477,0.2810545564,0.3564519286,-0.0047399714,-0.2098503858,0.1643420309,0.0381746069,0.2374487519,0.2311757207,0.0434191972,-0.1417311728,0.2041552216,-0.0501084067,0.2943194509,0.1833570451,0.3143186271,-0.0947325677,-0.1675237864,-0.0726325884,-0.0399611406,0.1879500002,-0.0444883145,-0.345372498,0.0505961776,0.0246744324,0.1378045529,-0.0245431215,0.3533416986,0.0892766118,0.0256678388,0.1503741592,0.3059256375,0.2060716152,-0.1718243212,-0.1820332259,0.0941530541,-0.0817900822,0.2399889678,-0.0283042174,0.1882384121,0.3200186193,0.1430033594,-0.1456785649,-0.019348938,-0.0278929994,0.4929859638,0.3488085866,-0.2583788633,0.0763230845,0.1167522967,0.0699775815,0.3328178227,-0.0869752988,-0.0373555347,0.363217324,-0.2329984903,0.0822153836,0.5594725609,-0.2854761183,-0.0367401056,0.095780775,-0.1388897151,0.0861750469,-0.203425765,0.5114117265,-0.1582648456,-0.3456502557,-0.2892788649,-0.0635989904,0.0663787723,-0.0988219678,-0.0952317938,-0.2246950716,0.0775165856,0.100346148,0.1074533015,-0.4275765419,0.1521026492,-0.3320240676,-0.0672014058,-0.4065212309,-0.0077387472,0.2845515609,0.2409390956,0.0388464667,-0.0567140058,-0.1350168139,0.034503065,0.0069476962,0.6226800084,0.1937186122,-0.2719529271,0.4494187236,-0.0038368623,-0.2786713839,-0.0097364467,0.0881031454,-0.093806468,0.0963916704,-0.3321146965,0.2460118085,0.1230233759,-0.1396175623,0.1429747045,-0.1674233228,0.2251501083,-0.4204905629,-0.1419075578,-0.1662168056,-0.2215796709,0.1382443011,-0.2380772978,-0.243644461,0.1490854323,0.3808838129,0.0073080109,-0.0184756722,0.106279768,0.0704990774,0.0176711548,0.3879520297,0.5070174932,0.0912010223,0.0155644929,-0.4338004589,-0.4961233735,-0.0477703921,-0.2083726823,-0.0755692422,0.2671868205,0.0514980368,0.0792371705,0.1202501506,0.1934061646,-0.1729039103,0.1122848094,0.1488929838,-0.3747888803,-0.2444615811,0.1333277076,0.1647691578,-0.1115220562,0.0973017961,0.1833662391,-0.2671873569,-0.0605350435,-0.0539660342,0.4409012794,-0.3858111203,-0.2134543061,0.2742444575,-0.0893195421,0.2211431712,-0.0542889796,-0.057926584,-0.1133795753,0.1059224457,0.0567836128,0.3305473626,0.1192980185,0.3175492287,-0.0366035327,-0.271453619,-0.0694233328,-0.1528828889,0.4136188626,0.1176905483,-0.2102363706,0.026281707,0.0310724061,-0.0865952522,0.5606865883,0.3495244086,-0.1818980277,-0.2859419286,-0.0456779487,-0.0751353055,0.2033851147,-0.4242960215,0.0203814339,-0.0072014169,0.0730981529,0.2851403654,-0.1397035569,-0.5321121812,-0.0779317319,0.3862553239,-0.2537599504,0.081772387,0.4235799015,0.0683877245,0.0249759108,-0.0616192855,0.7284914851,0.2601014078,-0.1255476624,0.0781496093,-0.2238434553]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1783","title":"Dataset Examples Explorer","comments":"Glad to see that it still exist, this existing one is more than good enough for me, it is feature rich, simple to use and concise. \r\nHope similar feature can be retain in the future version.","body":"In the Older version of the Dataset, there are a useful Dataset Explorer that allow user to visualize the examples (training, test and validation) of a particular dataset, it is no longer there in current version.\r\n\r\nHope HuggingFace can re-enable the feature that at least allow viewing of the first 20 examples of a particular dataset, or alternatively can extract 20 examples for each datasets and make those part of the Dataset Card Documentation.","comment_length":36,"text":"Dataset Examples Explorer\nIn the Older version of the Dataset, there are a useful Dataset Explorer that allow user to visualize the examples (training, test and validation) of a particular dataset, it is no longer there in current version.\r\n\r\nHope HuggingFace can re-enable the feature that at least allow viewing of the first 20 examples of a particular dataset, or alternatively can extract 20 examples for each datasets and make those part of the Dataset Card Documentation.\nGlad to see that it still exist, this existing one is more than good enough for me, it is feature rich, simple to use and concise. \r\nHope similar feature can be retain in the future version.","embeddings":[-0.3628720045,-0.4238744974,-0.0469127297,0.3083789945,0.1272349954,0.2346735299,0.1710146368,0.3707437217,-0.0460052155,0.3754509687,0.0373594649,0.2407734096,-0.3106932044,0.5867730379,0.425044924,-0.371111691,0.0176204182,0.1814071983,-0.0853752345,0.0868427306,0.0558809489,0.0799584836,-0.1806430072,0.1946162581,-0.4375636578,-0.2032096237,-0.2906790078,-0.0181118343,-0.2234526873,-0.5350555182,0.0933511034,-0.1131595075,0.318122983,0.0995311812,-0.0001116613,-0.1911393255,0.2427989393,0.1175147444,-0.5374827981,-0.017024342,-0.3509198427,-0.0632480681,0.3206753433,-0.1844154298,-0.1859279424,-0.3335364759,0.2009563446,0.0871430859,0.3625756204,-0.1000136361,0.1540151387,0.4514330626,-0.1147692949,-0.2746319771,0.021525532,0.6503049135,-0.4158145189,0.1027740315,0.4872650206,0.1220444217,-0.0628314987,0.436434269,0.3849204779,0.1084414795,0.082786575,0.1359821707,0.0367397927,-0.1285455823,-0.0026759566,0.6877962947,0.4381465912,0.0931990221,-0.3632138669,-0.5371533632,-0.094430171,-0.2077556103,-0.123566322,0.3460617065,0.045661848,0.2922644615,-0.3274420202,0.1238086373,-0.2014064342,-0.0435199961,-0.031609647,-0.0419529602,-0.3505464196,-0.0388570651,0.4434104264,-0.0454552062,0.0321435891,0.1650788635,0.0274939202,0.1676371545,-0.0362644419,-0.1789749712,0.2143942416,0.1790531129,0.2006743848,0.1832567006,0.0436740443,0.2886774242,-0.0051313275,0.1875259876,0.2100379318,-0.1485780478,0.4455875456,0.1579916477,0.198757574,0.290583849,0.1111951321,-0.0247339904,-0.3412183225,0.1907035261,-0.3514389098,-0.2364153862,0.0480780192,-0.1570365429,-0.0745462552,-0.1104142219,-0.1966985762,-0.1001117453,0.0987718701,0.3787342906,-0.0820647627,-0.3699345589,-0.3043927848,-0.132314533,0.0953299776,-0.3404927254,-0.1633529812,0.0134238387,-0.2573073506,0.07651674,0.1408937424,-0.4891138673,0.4031631052,-0.0637202039,-0.0823209435,0.0182059221,0.3176814616,0.1376454979,0.3231967986,0.0846775249,-0.0698207393,-0.2001847029,0.1453597397,0.0466810837,-0.2402201742,-0.0332044587,-0.1150539815,-0.4029281437,-0.3197678924,0.0904974714,-0.1786216497,0.1010476425,-0.5091859698,0.5675261617,-0.4719489217,-0.1423956603,-0.052287437,0.1688278466,-0.2888948023,0.0808207989,0.3657828867,0.3202440441,-0.3815145195,-0.402302593,0.0040496085,-0.2024496198,-0.354172647,-0.0102657899,-0.3748866022,0.3842636049,-0.5060606003,0.1932112724,0.3482882082,-0.2705029547,-0.0268902443,0.0732418075,0.1342188269,-0.0153008252,0.0547206663,-0.1656402797,0.5171090961,-0.0580826513,-0.3266589046,-0.0058833905,0.0500524789,0.0045681633,-0.0057807006,-0.1436333656,-0.0126427999,0.0987416208,-0.1880908757,0.082975395,0.1046871617,-0.0291380771,0.5324881077,-0.2547031343,-0.062380895,0.1240162924,0.2447225451,0.1867270172,0.0332897976,-0.3907249272,-0.2663160264,0.0874274001,0.0173915755,-0.1704370826,0.3375398517,-0.3268709481,-0.1081181169,-0.0301380176,-0.264152348,-0.2458115965,0.1156627536,-0.1632660478,0.1129676774,0.1194049194,-0.2301286161,0.1028624848,-0.4876904786,0.1936997324,-0.3489359319,0.1925223768,0.1154591665,0.08967641,-0.0020306529,0.0805486366,-0.0232968833,-0.2826875448,0.052323807,0.1806110144,0.1651832312,-0.1275327951,0.2021798342,0.4183845222,0.2501938939,-0.3330259621,-0.0042097811,-0.2055593133,-0.071884647,0.1867924929,-0.7351933122,0.3116651475,0.2892594337,0.1737939566,0.276227951,0.0559668019,0.0272191186,0.0849289671,-0.2509565651,-0.2338690907,0.1624337435,0.0834244862,0.1351511478,0.0118411081,-0.6773616076,0.179995209,0.3371834755,-0.1255377829,-0.0636920333,0.1597159505,-0.283005178,-0.0718507692,0.422870338,0.0580269322,0.0078619411,0.2213018984,0.1757678092,0.0611357428,-0.1028247252,0.1389880627,0.2311141044,0.059535902,-0.0213477742,-0.362592876,-0.1599088758,0.1234794483,-0.2691359818,-0.0851850137,0.1261221319,0.067501232,-0.2492976934,0.0882234201,-0.312413156,-0.4699494541,-0.0558281988,0.0536702834,0.1064356714,-0.3449403346,0.0961487666,0.1833516359,-0.1305019855,0.0808016434,-0.0739755332,0.4409167767,-0.2626431882,0.6817628145,-0.1589347124,0.1216822714,-0.1466044635,0.0929053277,0.1964097917,0.2997838557,0.395537138,-0.1922944933,0.1253749579,-0.6362916231,-0.5710225701,0.4147366881,0.114051111,0.2819750309,0.3237912357,0.1212586164,-0.480971992,-0.0920611545,0.0238983538,-0.3108506799,-0.1952959597,-0.1606523693,-0.0696496442,0.0866157934,-0.0036351357,-0.1760458946,-0.2337187082,-0.1928891242,0.3730760217,-0.1350295991,0.2156967968,0.1801916361,0.022220606,0.3586769998,-0.5773912072,0.0656881705,-0.1230051145,-0.4196787179,-0.034191139,-0.539200604,-0.3599954844,0.0488108844,-0.1813386828,-0.1222140566,-0.0861305147,-0.2249406278,-0.2832895219,-0.2888917923,-0.0127118332,-0.0550171845,0.1428121626,0.3644660413,-0.0129617769,-0.0480523892,0.0823384747,-0.2271732837,-0.0411864668,-0.0195229668,0.1034803912,-0.1157057807,0.0113717662,0.1424403787,0.4282975197,0.1344205439,-0.0786102414,0.0477127843,-0.4147192836,0.2543253899,-0.0843585953,-0.1063033864,0.1534173042,-0.0989839807,0.171402961,0.1308541298,0.2102694511,0.3452121615,0.2885038257,0.2634059787,-0.2960059643,-0.2334072441,0.0258640796,-0.0685663298,0.2710050941,-0.0581062995,0.0561370216,-0.0502934344,0.114802517,0.001715407,0.1360842437,0.2629626095,0.0426918678,-0.4515696168,0.0488931015,0.0635928586,-0.0651577339,0.0423951,-0.1487871706,-0.2744347751,-0.103025049,0.0133102573,0.2753252983,0.6601600647,0.0227461625,-0.0210919306,0.0015430604,-0.5326405764,-0.38593027,-0.0235722344,-0.2493938506,0.0746174902,0.1191318855,0.459521085,0.0635116026,-0.3181056678,0.6373790503,0.2131972164,-0.1942306459,0.1262274981,0.1192458645,0.1060936302,-0.2007537484,0.0158857182,0.167734772,-0.0721866861,-0.0692640319,-0.2534737885,-0.369161576,0.0623059608,-0.0160835367,0.4453268647,0.388785243,0.0738943517,0.3272370398,0.4302355945,0.4262765348,0.3560321331,0.2934803665,0.0179014914,-0.1467515528,0.259575665,-0.1083230153,0.4040625393,0.2071855366,-0.0402671322,0.1380898952,-0.1056490913,0.1593226194,-0.3352407217,0.290412724,0.3205354214,0.1410128027,-0.3666876554,-0.2220756859,0.56498456,0.2509381473,-0.1964270324,0.4854005873,0.5616508722,0.1654265225,-0.233763665,-0.0134838708,0.748855114,0.011356473,0.1794234514,0.2811605334,-0.0862902179,0.5181322694,0.2136398107,-0.0785983652,-0.3509382308,-0.4129319489,-0.1072376892,-0.2359375358,0.2514556646,0.2131433636,-0.1995893121,-0.0309330244,0.03396146,0.1847021431,-0.0806807503,0.0858266652,0.0671873316,-0.1996517628,-0.4594395161,0.0668153688,0.1864132285,-0.0313172564,-0.0020040085,-0.145494014,0.0337265916,-0.0769511908,-0.416177243,-0.2480276078,-0.1253672838,-0.0278667584,0.373272568,-0.3237701654,0.324667275,0.2412865162,-0.0232411996,0.0666115135,-0.20861274,-0.0316211805,-0.5383197665,0.0119531937,-0.2288863957,-0.0497355685,0.2615383565,-0.1323259473,-0.2374707013,0.1870981157,-0.0615191646,0.0161329079,-0.0706648678,0.1741616726,-0.1979465038,-0.1789819896,-0.0728757456,-0.1146976128,0.0056844298,-0.2668553889,0.0995808095,0.2071620673,-0.1186468303,-0.0114446003,0.3965777159,0.1650594175,0.0238777604,0.1108318418,-0.1763143837,-0.016128961,0.2718544006,0.1713452637,0.0496565625,-0.1989776492,0.2984876931,0.5128170848,-0.2056119442,0.0405337214,-0.0790155455,-0.0047814026,0.1420996338,0.4264116883,0.2334615588,-0.1881693155,-0.2181652933,-0.042689167,-0.3736994565,0.2281521261,-0.1707381308,0.4515784681,-0.1238324121,-0.0085978499,0.0644105449,0.1072041318,-0.2979231477,0.2373313606,-0.2675721049,0.0675865933,0.1066540703,0.2853392363,0.1486874819,-0.2966265678,-0.0560372174,0.0898982659,-0.1692582667,-0.2235216796,-0.3569561243,0.1449780911,0.0643262416,-0.0398164392,0.2931802571,0.3415266573,-0.0312435273,-0.2159667462,0.214490816,0.0533115119,0.262531817,0.2650318146,0.0118767377,-0.1454114616,0.2332726121,0.0268369894,0.2896949053,0.1986461282,0.3081661463,-0.1180163845,-0.135485068,-0.0996299013,0.0139544401,0.1951918602,-0.048649665,-0.3119460642,0.0135786571,0.0048814137,0.0742327794,0.0398384668,0.3551677465,0.1148306951,0.0165186822,0.1727311164,0.2973673046,0.1809267253,-0.1903831512,-0.209531948,0.1714634299,-0.1108704507,0.276352495,0.0186879691,0.2120939642,0.3231056333,0.16003111,-0.1867732108,0.0366658941,-0.0178786553,0.473691076,0.281827718,-0.251159519,0.0975119025,0.1115472913,0.1697711647,0.3672658503,-0.1209312379,-0.0234210249,0.3076204956,-0.1960165501,0.0861851797,0.558036387,-0.3268481493,0.033643309,0.0866759047,-0.1684750021,0.0443012565,-0.209053576,0.5345990062,-0.1278414726,-0.4057995677,-0.251426965,-0.0618171766,0.0548415817,-0.1127424166,-0.0393093787,-0.2156902999,0.0880423188,0.0932185352,0.1074286401,-0.372996062,0.2144721597,-0.3236888051,-0.0083959326,-0.4254332185,0.0057018623,0.3477914333,0.2657504082,0.0909907222,-0.0972994193,-0.0993013978,0.0328818858,-0.0346272998,0.6206565499,0.1180765405,-0.2262898684,0.4333232045,0.0406723134,-0.3129925132,0.002246202,0.0734697282,-0.1207942143,0.10678716,-0.340085417,0.2328075171,0.0746830851,-0.1075115353,0.1573762,-0.1454617828,0.0912463143,-0.4371271729,-0.1730298847,-0.188923195,-0.2234083563,0.0394577123,-0.230926007,-0.2144008875,0.0947491527,0.5052252412,0.0336165503,0.0103738178,0.0725554228,0.055790823,0.007101214,0.329657793,0.5122724175,0.1003587246,0.0307941046,-0.3707309663,-0.4566681683,-0.0553625114,-0.2726143003,-0.1133474335,0.3075873852,0.0595557541,0.020858217,0.1412316114,0.2326322049,-0.1926282942,0.149674654,0.0902005285,-0.4039148986,-0.2468186319,0.1658665985,0.1431381553,-0.1139870957,0.0197614338,0.1975493133,-0.21652776,-0.0819167867,-0.09057419,0.4436953366,-0.3907444775,-0.1234756783,0.2300353348,-0.0508332215,0.1791044921,-0.0547770448,-0.1305190027,-0.1095266789,0.1321536899,0.0601376593,0.321605444,0.1370901167,0.2485487461,-0.0404412746,-0.2820129097,-0.1084609926,-0.1037034839,0.4561384022,0.0615927354,-0.2457846552,0.0784033164,0.0684582964,-0.0974804536,0.5709724426,0.358867377,-0.2080979049,-0.2477644831,-0.1020753533,-0.1012575924,0.2366830111,-0.3834226429,-0.0257164594,-0.0323445499,0.0827093869,0.2758225203,-0.1074641719,-0.5303416252,-0.1059396639,0.3487037122,-0.2784738243,0.0735507756,0.4093360901,0.0977711827,-0.0054469951,-0.068658635,0.6830686331,0.2762746215,-0.1177213266,0.0996516719,-0.2393695861]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1781","title":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import ","comments":"Hi ! I'm not able to reproduce the issue. Can you try restarting your runtime ?\r\n\r\nThe PyExtensionType is available in pyarrow starting 0.17.1 iirc. If restarting your runtime doesn't fix this, can you try updating pyarrow ?\r\n```\r\npip install pyarrow --upgrade\r\n```","body":"I'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n","comment_length":44,"text":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import \nI'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n\nHi ! I'm not able to reproduce the issue. Can you try restarting your runtime ?\r\n\r\nThe PyExtensionType is available in pyarrow starting 0.17.1 iirc. If restarting your runtime doesn't fix this, can you try updating pyarrow ?\r\n```\r\npip install pyarrow --upgrade\r\n```","embeddings":[-0.1790662259,-0.1585770994,-0.0443940833,0.2205853313,0.1801975965,-0.0639269799,0.1799789518,0.2682418525,-0.311421901,0.1932341158,-0.0361046977,0.4329881668,-0.1667326093,-0.1118558422,0.08164116,-0.067264691,0.2507495582,0.3972359598,0.2283162028,0.0152990064,-0.1987970024,0.2214333564,-0.3317900002,0.2289108038,-0.3630690873,0.2572512329,0.0558958426,-0.1920558065,-0.2301744372,-0.5764158368,0.4622827768,-0.457690388,0.1837862581,0.0198195726,-0.0001144319,-0.0085836183,0.3469851613,0.0369327553,-0.0782001764,-0.179927215,0.0146695673,-0.0796009824,0.3153556585,-0.0688825548,0.1376845986,-0.049098216,-0.0455625169,0.3671183288,-0.0516867824,0.2796535194,0.2375448793,-0.0251752865,0.5402342081,0.0167230759,0.5227000713,-0.1609021574,-0.2256377041,0.6437642574,0.3970014751,-0.3397271037,0.3303521574,-0.0623879507,-0.3161487877,0.1467274725,0.3413374126,0.0093860058,0.4785271287,-0.2968401313,-0.0349214338,0.0639048517,0.1993095577,-0.106132485,-0.0536232851,0.1321716458,0.2737747133,-0.4418088794,0.1780537218,0.3744233847,-0.0651033148,0.1674182564,0.3446937799,-0.0282741636,-0.1547164768,0.1121773794,-0.3299468458,0.4400127232,-0.0608913675,-0.0126784639,0.1606597602,-0.014353618,0.4384037852,0.1720194966,-0.146457091,0.1746094525,-0.1346684247,0.1695789099,0.0346121825,0.3614486456,-0.016898077,-0.3039343059,-0.1646404862,-0.2202556729,0.185238108,0.4585883021,-0.1547237784,0.296910435,0.0425983295,-0.1898492277,0.1939644963,0.0431142598,-0.1044290736,0.1421603113,-0.1413403302,-0.3777513504,0.2381351292,-0.0861328915,0.4219927192,-0.1524940878,-0.4551401138,-0.3460401297,-0.5161126256,0.0715425164,-0.0819104239,0.2317310572,-0.1786213517,0.1398227513,0.0135104461,0.2279352993,-0.1297752708,-0.2341376245,-0.0415705703,0.3183738291,-0.1867560446,-0.2240166664,-0.0605837144,-0.1977543384,-0.03448724,0.0647213534,0.0029570737,-0.1728652567,-0.3638984561,-0.231403172,-0.2383738458,0.3250190616,-0.1241195798,-0.0144523475,0.1709778607,-0.5575419664,-0.1424941123,0.3194749057,-0.5617517233,-0.4043883681,-0.2904806435,0.1659997851,-0.1723260581,-0.1548972428,-0.3065091968,-0.1803507656,-0.0339657068,-0.2712192237,-0.134447813,-0.4794248044,-0.0160602592,-0.4012240171,-0.0779944062,0.0237126052,-0.4536841214,-0.1713841707,-0.0673347339,-0.0593353547,0.3455157578,0.0348362587,-0.1741034687,0.0599535629,-0.08434508,0.1987013072,0.0848146528,-0.1136741415,-0.2233837992,-0.2362246662,-0.0785076916,-0.0386033542,-0.0020760684,-0.2877239287,-0.0197688248,-0.0431312881,0.4431330562,0.2906011641,-0.1093343198,0.2179859579,-0.3110338748,-0.1981665194,0.3580446541,0.1255996376,0.0955358669,-0.1678241789,0.0892780572,-0.2828473449,0.0909027904,-0.1441042423,-0.0333426073,-0.0247795042,0.5660248995,0.0915235952,-0.1175733432,-0.0555668958,0.0603022575,0.1770724803,-0.1671856344,0.1522559077,-0.3447423875,-0.0380014926,-0.3247757852,0.3926260173,-0.2453923523,-0.0746597871,0.0887517035,-0.1059178561,-0.1791073531,0.4000236094,-0.0704847723,0.0581471808,0.1778243631,0.1856284589,0.0744912401,0.4301854074,-0.270832032,-0.4979071617,-0.1084052026,0.0034129983,0.2324229479,0.0094567342,-0.1076803729,-0.0670001134,-0.0296544079,0.1820956469,-0.2738328874,0.1543227136,0.0993480384,-0.4678512514,-0.0098041911,-0.2234818041,0.0692760423,-0.0187455416,0.8860567808,0.1675926,0.254861325,0.1322639734,-0.2607864141,0.1686573178,0.0046349959,0.092463553,0.2096389085,0.1399048716,0.0681173578,0.2411949486,-0.181121707,-0.1170749813,-0.025532376,0.0542542972,0.2834241688,0.2026453614,0.1976955384,0.0019476041,-0.4756842554,0.0932431668,0.236591965,-0.173293978,0.3250072002,0.0668008924,-0.3597570062,0.0757200792,0.0793354735,-0.0889667869,0.2429632545,0.1491228789,0.5926089883,-0.0030886792,0.265087992,0.0802495033,-0.1685391068,-0.227558732,-0.0866515636,0.2750094235,-0.2485767454,0.2147495747,-0.3447979093,-0.0234028008,-0.1641614139,-0.5490457416,-0.1036438271,-0.126457274,-0.1215843335,0.2856437564,0.1915252656,0.0521730147,0.2701273263,-0.0963661373,0.0707283691,-0.4580160081,-0.0769520402,-0.2203986943,-0.3743089139,0.0262891296,0.125177443,-0.0590720661,-0.0304245558,-0.3601930737,0.1301649362,0.0701581091,-0.3697500825,0.1027367115,-0.2174154371,0.3510778546,0.1288847178,0.138935551,-0.2283433229,-0.2265936732,0.3410273492,-0.2279568166,-0.2316577584,0.1729710698,-0.2530034184,-0.1466672271,-0.0388243198,-0.0024364509,-0.4087469876,-0.135675177,-0.0016151909,0.0617810786,0.0647763386,0.2473357022,0.3024420142,0.2300611138,0.0246636327,-0.0055180066,0.1005854756,0.1662604958,0.4291490912,-0.1283849478,-0.1799859107,0.3432487845,-0.2596538365,-0.0454661138,-0.1025901437,-0.2102573067,-0.4968859553,-0.0208501443,0.4942528009,-0.1182472929,-0.0494846925,0.195320338,0.1326886714,0.0062993877,0.0963878632,-0.0298702251,-0.4402624667,0.3675146997,0.2336600572,0.0597360283,0.0384743735,-0.1328354329,0.5266131163,-0.3150179982,-0.0349395014,0.3293399513,-0.2715450227,-0.0616858229,0.0273055118,-0.3756971359,-0.0280947611,-0.1555030942,-0.2060226351,0.1761333048,-0.2777555287,-0.1640523523,-0.1638626456,-0.0774843618,-0.2383016944,-0.211811766,0.157907486,0.320115298,0.2225836366,-0.1609209478,-0.1791943163,-0.0132112838,-0.1428373307,0.1238498911,0.2051279247,-0.2716382444,0.070098713,0.4383113682,-0.6262353659,-0.0596388616,0.2981524169,0.3296098411,-0.0452162959,0.1585372835,0.0653334036,0.0440482982,-0.0163499322,0.3518734276,0.2703382075,-0.0755230933,0.1412079483,0.0766282231,-0.1490995139,-0.4287262857,-0.1191016287,-0.2681034505,0.0243277531,-0.4682481289,-0.1296655834,0.2163131386,0.3303775787,0.4147616923,-0.1004247591,0.293302983,-0.3792907894,-0.4569360912,-0.1829234958,-0.1471691579,0.1107234657,0.5425566435,0.0442091785,0.2326034904,-0.1123693436,-0.1195011735,0.0481864773,0.2808307111,0.0800946355,-0.0887425244,-0.0581214353,-0.2451636791,-0.1497919708,0.4851491153,0.3575556874,0.3432013988,-0.5243742466,-0.2735806704,0.1761348397,0.2826775312,-0.0850728601,-0.3356333375,0.1472049057,-0.6066705585,-0.1070991307,0.2576452196,-0.055997856,0.2430212349,0.1222137883,-0.0310080741,0.1505090445,0.5593098998,0.0524693616,0.0837615132,-0.1632101685,0.4474861622,-0.1040820181,0.1998447031,0.0331470855,0.8620254397,0.096216701,-0.1062677801,0.5184580684,-0.0881283954,0.187896505,-0.0081638461,0.1899982393,-0.3045163751,-0.0799034685,-0.0812915638,-0.2375046611,0.3007557392,-0.3801081777,-0.185572058,0.0321749002,-0.1073666438,0.0412659794,0.2691592276,-0.3244300783,0.1020285934,-0.0270945523,-0.1815610975,0.1279204041,-0.3390734196,0.3881623149,-0.2331756949,0.0897140428,-0.2383809537,-0.0345177911,-0.356508702,0.1594756246,-0.474730432,0.0953373015,0.4025704861,-0.2607304454,0.2180155218,0.2544823587,-0.0881718323,-0.2051602453,-0.1190965176,0.1531813145,0.210489586,-0.2932991683,-0.1400910914,-0.0148190809,0.3789052665,-0.266335696,-0.2053687274,0.3243813813,-0.1200553551,0.0190121774,0.207923457,0.2172119021,0.2174226642,-0.1397690326,-0.1159708947,-0.1548206359,0.4396516383,-0.1793833077,0.0974545404,0.1022866741,-0.1480679661,0.2389145344,0.0597929209,-0.1215438694,-0.0676660612,0.0342872739,0.0530832857,0.1256922781,0.4378902018,0.0360092968,-0.1338087171,-0.0892588124,0.0835771114,-0.245250836,-0.3923930228,0.1614235938,0.0413192809,0.0945864543,0.1541693658,0.6016369462,0.2200914919,0.1959430873,-0.0029780893,-0.393584162,0.0153974351,0.1952128708,0.223405689,-0.1829378307,0.1482762694,0.4871101081,0.2310220599,-0.2154344916,-0.2355452329,0.1699389964,-0.0202158522,0.1654955,-0.20813106,-0.0237617251,0.1376885176,-0.0255881511,0.1706531644,0.1360502839,0.0790043622,-0.2413068563,-0.2038363963,0.1567434072,0.0473498441,0.1087637916,0.1444313526,0.0323912241,0.1363891959,0.107562758,-0.0144930696,-0.0115745785,-0.2664196193,0.02062767,0.5048086643,0.0116718393,-0.1032412574,0.139886573,-0.1736040264,0.4591966271,0.1729912311,0.0851141736,0.0556668043,0.1222586036,0.0067542689,-0.1525090188,-0.3688284457,-0.082253091,0.2527532279,-0.5157036781,-0.0102973552,-0.3664902747,-0.0804439932,0.368168503,-0.2823773623,0.1550220698,0.3580296338,0.1487148255,-0.2343886197,0.0802348927,0.1561585665,0.0121045457,-0.065727964,0.2085698545,0.096271567,0.1325168312,0.0135209747,0.1941374689,0.112574257,0.3337906599,0.0733123422,0.1991668493,-0.1907485276,0.3111235499,0.3012802005,-0.0447270758,0.0997254848,0.1071460694,-0.2848716974,0.1915014982,-0.1223457381,-0.0366430134,0.3447068334,-0.2727971673,0.0545439571,-0.0499814786,0.2050121874,-0.1866604686,0.2229485512,0.4996074736,-0.03812664,-0.0852463171,0.0450395942,0.4715332091,-0.0909212008,0.2189595699,0.2121885717,0.0400196053,-0.1102890149,0.0615603216,0.0899292976,-0.1647674292,0.1405571699,0.3287213445,-0.2781499326,-0.168334499,0.134196192,0.1551510543,0.2446388006,-0.151016742,-0.029192524,-0.0126024392,-0.2535862923,-0.3425185382,0.3309091926,0.5272144675,0.3617966473,-0.2228709012,-0.350569576,-0.262776345,-0.0863151848,0.1195718646,-0.1643646806,0.1972000301,0.5600979328,0.0767566711,0.1942419708,-0.170449689,0.1397040784,-0.0916650072,-0.0431894138,-0.0491229817,0.1625778377,-0.0082927756,-0.2567868233,-0.5138967633,0.0966709107,-0.0504026636,0.3500345945,0.112573199,0.1148154289,-0.1855021268,0.0442882888,0.0609000511,0.166542992,0.3051973581,0.3875773847,0.475540489,-0.2138941735,-0.1814345717,-0.2916370332,0.3696193695,-0.2389659733,0.0414594263,0.1105279401,-0.0062889932,-0.1999723762,0.3702288866,-0.011494047,0.1765713245,0.0734883472,0.0298025087,-0.3209836781,-0.2221903056,0.1444744915,-0.3279424608,0.0714496151,-0.1231062338,0.23197487,-0.6399855614,0.1101722121,-0.3500314057,-0.2746466994,0.0027535872,-0.0029036941,0.6594691873,-0.0845692158,0.4161297679,0.1328029633,0.0950930789,0.0378227122,-0.3472763002,-0.0520948917,0.2472882867,-0.0002780956,0.2236509174,0.1149878949,0.0075579491,-0.3633066714,0.3691530526,0.0104589136,-0.1032638997,-0.1003397927,0.0807716623,0.0182328187,0.161239177,-0.1669697762,0.0177487005,-0.1094033271,0.0470383093,-0.3611803651,-0.5371569395,0.4243586361,-0.3146577477,-0.3414463699,0.0673392788,0.2100278288,-0.1327817142,-0.1468293816,-0.4087649286,0.0332328007,0.2969633639,0.1331399381,-0.2754375041,0.2085576653,0.0338269994,0.0557053797,-0.0335556269,0.3838145435,-0.0997876003,0.1347156912,-0.2518532574,-0.0629429668]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1781","title":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import ","comments":"We should bump up the version test of pyarrow maybe no?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/__init__.py#L60","body":"I'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n","comment_length":12,"text":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import \nI'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n\nWe should bump up the version test of pyarrow maybe no?\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/__init__.py#L60","embeddings":[-0.2523330152,-0.0487640463,-0.0525005311,0.1863226593,0.1662422568,-0.0158157963,0.2124302089,0.3041544259,-0.2492188811,0.1005084142,-0.0795976073,0.2190067768,-0.1763952523,0.1348823607,0.1778921634,-0.0839820057,0.3172160089,0.3425171673,0.170229733,0.0034572429,-0.1986777186,0.2041054517,-0.3929193616,0.1429923326,-0.4562448859,0.1120674536,-0.1004237384,-0.1659390181,-0.3091564775,-0.545675993,0.4575783014,-0.1959267259,0.2351273745,0.2184674442,-0.0001175922,-0.0124965468,0.3385895193,0.0366292931,-0.1208419204,-0.1643351912,0.0028457558,-0.1162932366,0.3002604842,0.0155366054,0.1306965649,-0.1173732802,-0.131010741,0.2819643021,0.0104620466,0.3382070959,0.2187791616,0.1117322221,0.4797917306,-0.03051465,0.4269357324,0.0394579247,-0.1841093302,0.6450079083,0.2746907175,-0.3599128723,0.3146471083,0.0252605733,-0.1951122731,0.1803992391,0.3484561145,0.0466831587,0.3071748316,-0.2623764277,-0.0181796644,0.166992113,0.266253233,-0.2084515393,-0.2049445361,0.0037891234,0.2382532507,-0.4549330473,0.2376501709,0.2463242114,-0.0878767967,0.206163913,0.1515924931,-0.1400679946,-0.1458550692,0.093880266,-0.3951752484,0.3318157494,-0.1226086617,0.0173219126,0.2952832282,-0.0221125688,0.4374909103,0.1551593989,-0.1368001401,0.1820469499,-0.1769174486,0.0529863201,0.0432354845,0.3180353343,0.1049897671,-0.2260015458,-0.2836251557,-0.2488674223,0.1070189774,0.360394001,-0.1391819865,0.3542323112,0.0992075205,-0.0572002903,0.2610767782,0.1545376927,-0.1755642742,0.163325116,-0.1762329787,-0.5038961768,0.1768726259,-0.0958001614,0.5071957707,-0.2843331695,-0.3498052955,-0.2900445163,-0.4053878188,-0.073068209,-0.1290082037,0.2994889319,-0.1866181642,0.1249050274,-0.0805247277,0.21753636,-0.2008436173,-0.1438865215,-0.0682213604,0.2669624686,-0.1834172606,-0.0570335276,-0.0626304001,-0.271076262,0.0411303379,-0.0582130253,0.0407617912,-0.1001184881,-0.3670749366,-0.2836510241,-0.1496762037,0.3451512158,-0.1858060956,0.0214838702,0.0956160128,-0.4512593448,-0.2020385116,0.2469689846,-0.4720640779,-0.4343170524,-0.1926783472,0.1236828119,-0.2943957448,-0.2721839845,-0.4825967848,-0.0777638406,0.0236567073,-0.2987986207,-0.1069213971,-0.4969184399,-0.0210132021,-0.2885314822,0.0636576489,0.1813222319,-0.5285366774,-0.1953869462,0.0388196558,-0.1559575945,0.2169226855,0.122774519,-0.1058243811,0.06658452,-0.1272356063,0.1722142249,0.0678254217,-0.1390699148,-0.3685930371,-0.165362522,0.0014286933,-0.0888137966,0.0626112521,-0.2822947204,0.0318400972,-0.1196343973,0.3919803202,0.2424140424,-0.0739543438,0.1438730061,-0.2321498096,-0.2062594742,0.3833554983,0.1533295661,0.0426843837,-0.1974776387,0.0777683035,-0.2716159523,0.172383666,-0.1321299523,-0.1104113683,-0.0025512481,0.5634419322,-0.0041639465,-0.1342942417,-0.1427400708,-0.0695081428,0.1323201656,-0.0282430872,0.1216867715,-0.4317907095,-0.1166170016,-0.3871613145,0.3886778355,-0.2334272861,-0.2031684965,0.0357394218,-0.0862233266,-0.1130234376,0.4462723434,-0.1010146812,0.1027990282,0.1067141294,0.1957783103,-0.0168711562,0.5140176415,-0.2333605438,-0.4203687608,0.0148832239,0.063688755,0.2070460618,-0.122683011,-0.1610477865,-0.0151304184,-0.0619876944,0.2789939046,-0.0767858326,0.1227469593,0.1486809999,-0.4519362748,0.0198201947,-0.2628512979,0.015533193,-0.0407112911,0.7290278077,0.1976096779,0.2279250771,0.1615702063,-0.0953812748,0.0725404024,0.0677554086,0.166843012,0.213824302,0.1273829043,0.0644412041,0.1757564098,-0.0243866388,-0.0287701897,-0.0390456878,0.0340319164,0.4066469371,0.0276176669,0.1781419516,0.0748027563,-0.4287624061,-0.0395352952,0.3294918537,-0.1510696709,0.3953806758,0.0655234605,-0.2726128697,0.1707897633,0.0891469046,-0.0429267138,0.1947553456,0.0876765251,0.5724298358,-0.0105736926,0.1980861574,0.0523582101,-0.2156609148,-0.2612199783,0.0148652764,0.3684641123,-0.3555316627,0.168821618,-0.3422127664,-0.0564826354,-0.0804246962,-0.5909098983,-0.1934214681,-0.1574142128,-0.1811639071,0.270045191,0.2141496986,0.1339780241,0.1576889157,-0.0733083636,0.1067483202,-0.4863282442,-0.0738995895,-0.2282581329,-0.2883931398,0.0446725264,0.1919703782,0.0290280655,0.0329237916,-0.3296040893,0.0806200206,-0.0470587313,-0.4120879769,0.1427522451,-0.3298730254,0.4225692749,0.2567951679,0.1710402668,-0.1535221189,-0.1787311435,0.3516642451,-0.3453472257,-0.273106724,0.0799568295,-0.1811098307,-0.1049610376,0.0768165588,-0.1474284083,-0.4768425226,-0.1534945667,0.0598899312,0.0830638185,0.0527623147,0.423609674,0.312040329,0.2915709019,-0.0322347879,-0.0764632151,0.0810081139,-0.0490612164,0.4150721729,-0.0922160372,-0.2377426922,0.2039554417,-0.1771888733,-0.1088011116,-0.0711606964,-0.3573568761,-0.5601979494,-0.0418603197,0.4927075803,0.010578312,0.0575984083,0.2255036086,0.0393240824,0.0180021767,0.1091768891,-0.1545198709,-0.4140290022,0.3165152371,0.2181731313,-0.0145038012,0.1435663551,-0.1518925726,0.5895211101,-0.144222483,-0.0825711414,0.3251060247,-0.3352972865,0.0957802907,-0.0571888573,-0.436619401,0.0928663239,-0.1862511188,-0.1871150732,0.1749067456,-0.1497739255,-0.0794539526,-0.1770402789,-0.1446828246,-0.2073644102,-0.228625074,0.0970658138,0.2553845644,0.166098997,-0.1231866032,-0.1512463093,-0.1692303419,-0.141112566,0.1167834029,0.2507198453,-0.2396346927,0.0616483875,0.2172185332,-0.4265663326,-0.3909418583,0.3710407317,0.3012891412,0.0259770304,0.1834537834,0.0809392184,-0.0030561611,-0.0431584418,0.5169432163,0.2112692446,-0.0681010634,0.1293731332,-0.0629754737,-0.1789999902,-0.313759774,-0.1536003947,-0.2273901403,0.0419161431,-0.2458281219,-0.1203752831,0.0496405847,0.4875429273,0.3996710479,0.002337249,0.1753500104,-0.3314769566,-0.5423309803,-0.1761453152,-0.1423760504,0.0889781341,0.3972504437,-0.0267481524,0.2037088722,-0.1242250428,-0.243479073,0.0697272494,0.3628192842,0.0644105449,-0.0068066614,-0.0483323596,-0.10679432,-0.1511835456,0.5850639939,0.4819463491,0.2271704823,-0.5583181381,-0.286367923,0.1815141737,0.4056475759,-0.0193757918,-0.3025690019,0.1438621581,-0.4708253443,-0.0668404996,0.2128702402,0.0905152708,0.3326124251,0.2842575908,-0.2486423105,0.0388180315,0.6382364035,0.0988051891,0.0343161672,-0.0892922953,0.49299106,-0.1428381652,0.1206919178,-0.1030422077,0.8876590133,0.0681363791,-0.0077908449,0.5693282485,-0.0614268482,0.2771974504,0.068101339,0.0622132197,-0.2855435908,-0.2603215873,-0.0983155668,-0.2728047669,0.4164667428,-0.4336304963,-0.221946165,0.155780077,-0.026932653,0.066119954,0.3067183793,-0.2498595566,0.0961832851,0.013390149,-0.3934651911,0.1107999235,-0.2122752666,0.4279294908,-0.2234969735,0.101549767,-0.1754229516,-0.0497286431,-0.2396419942,0.1175374761,-0.430814147,0.0331923291,0.3709737659,-0.317253679,0.2288956344,0.2577436566,0.0889182091,-0.141722247,-0.1916982234,0.0845987424,0.0329199918,-0.1811101884,-0.0875386521,-0.068687737,0.452534914,-0.2162842602,-0.3488132358,0.2185436636,-0.1042521298,-0.101282239,0.1989808828,0.1156762838,0.1828416735,-0.1929689944,-0.1359177232,-0.2483337224,0.3307644725,-0.2326059341,0.0540718846,0.1086548418,-0.2120224983,0.1648385674,0.1059635356,-0.1092541367,-0.1594353765,0.1256254166,0.0241577402,0.1654987782,0.4907960892,0.0987457111,-0.0840590298,-0.1092972606,0.1311759055,-0.0803507939,-0.363050431,0.1574039161,0.0485974811,0.0473388284,0.0904609561,0.6905888319,0.2953507602,0.0312895663,0.0890228674,-0.4370048642,-0.1233450845,0.2039526552,0.1348643601,-0.0916294232,0.1951624304,0.475140214,0.224166438,-0.0929956585,-0.2081570029,0.1727321446,0.0033662173,0.1069729254,-0.0771828964,-0.0189059824,0.232889846,-0.0772680119,0.1376776546,0.1503412873,-0.0015319613,-0.2224681377,-0.1826071292,0.1586173475,0.0367654748,0.1103428006,0.2400776446,0.0778755695,0.0360419936,0.0310729817,0.0403478816,0.0142794624,-0.2006367147,0.0482726395,0.4718226492,-0.1344973445,-0.1459968537,0.1610676497,-0.1127182469,0.3590018749,0.1598260999,0.1207844317,0.0390725471,0.1166166961,-0.0071600676,-0.0766648501,-0.2912123501,-0.0944376364,0.3002471626,-0.5431557298,-0.0121316956,-0.2362562269,0.1075000241,0.4157030582,-0.2435488254,0.197552368,0.4006819725,0.096564889,-0.3491300941,0.0937689617,0.3165521622,0.0085670557,0.0116302986,0.2309058309,0.1136868969,0.2488333881,0.0472894199,0.1845332831,0.2099066824,0.3159948885,0.1457403004,0.249399364,-0.2320206165,0.2925786674,0.4821704626,0.0805382729,0.0750810653,0.1849027723,-0.2852540612,0.3341079354,-0.1796644181,0.0386552773,0.3141566515,-0.3337796032,0.2156296223,0.0320255458,0.1587243527,-0.2431035489,0.2359022349,0.6398796439,-0.1011789069,-0.1626476645,-0.0635517165,0.4580010772,-0.0519554913,0.0675268844,0.2264327407,-0.0090872282,-0.0393096656,-0.0127110826,0.021764908,-0.3342083991,0.2747444212,0.2877911627,-0.2475378215,-0.3277218342,0.1648161709,0.1407870352,0.199048534,-0.0586869754,0.0995972008,-0.0684360638,-0.2094880939,-0.2392916977,0.4428699911,0.6097870469,0.2719297111,-0.1265610904,-0.3305949271,-0.2185559869,-0.0713566467,0.1073850095,-0.1323079616,0.1383729726,0.6073840857,0.1334795654,0.1943719536,-0.1200854182,0.2023511231,-0.1061750576,-0.062720798,-0.235678032,0.2892154455,-0.1120434031,-0.3041085303,-0.4342931807,0.0971997753,-0.0791025981,0.3896041811,0.1614879966,-0.0466888547,-0.0932931229,0.0263552647,0.0339113064,0.1216713712,0.4294709861,0.4924502969,0.4564655125,-0.1882781386,-0.0716709644,-0.5296524167,0.3033544123,-0.1564306617,0.0040733321,0.0894771218,-0.0616853014,-0.1133640036,0.4180998206,-0.0234917551,0.1760894954,0.0287360642,-0.1552024484,-0.4042856395,-0.237808004,-0.055605758,-0.2099163234,0.0312488209,-0.1989637166,0.2864398062,-0.6396810412,0.0399436317,-0.2259148806,-0.1114634946,-0.0466876291,-0.1536789685,0.6727237105,-0.0054675024,0.5921102762,0.1088128164,0.0799226239,-0.108301878,-0.3806100488,-0.1008622944,0.2871838808,-0.0343180299,0.246243611,0.0259384718,-0.0946852192,-0.3128486872,0.3983432651,0.138027519,-0.03853634,-0.1736392379,0.1281903535,0.0662966594,0.0477795191,-0.0080333147,0.0903383568,-0.2160423845,0.0488568209,-0.3286257088,-0.4875940681,0.4248382449,-0.3976857364,-0.3032748997,-0.0264454316,0.2357451022,-0.0130851129,-0.2893612385,-0.5277013779,0.159249723,0.2845777273,0.0891317129,-0.2934297621,0.3356145322,-0.0103861308,0.0543596335,-0.0265037157,0.4959978461,-0.208951503,0.003418762,-0.1168434545,-0.0803293511]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1781","title":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import ","comments":"Yes indeed.\r\n\r\nAlso it looks like Pyarrow 3.0.0 got released on pypi 10 hours ago. This might be related to the bug, I'll investigate\r\nEDIT: looks like the 3.0.0 release doesn't have unexpected breaking changes for us, so I don't think the issue comes from that","body":"I'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n","comment_length":46,"text":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import \nI'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n\nYes indeed.\r\n\r\nAlso it looks like Pyarrow 3.0.0 got released on pypi 10 hours ago. This might be related to the bug, I'll investigate\r\nEDIT: looks like the 3.0.0 release doesn't have unexpected breaking changes for us, so I don't think the issue comes from that","embeddings":[-0.0131391417,-0.0668326616,-0.0331190228,0.1483884603,0.1552635431,-0.1366585791,0.2972118258,0.3255247176,-0.3937358558,0.1865732521,-0.0172689985,0.3017130196,-0.0978723019,-0.1359524727,0.1570970267,-0.0766468048,0.369223088,0.3322661221,0.2101120651,0.0341323018,-0.2047906369,0.2435746193,-0.3605642617,0.2816300988,-0.3349714279,0.2964655757,-0.050194364,-0.2114924639,-0.172048673,-0.5515245199,0.46529755,-0.3716568947,0.171802789,0.0862077251,-0.0001152871,0.04787625,0.3382734656,0.0108991088,-0.0602638423,-0.0971442983,0.0056828819,-0.0494792387,0.2477095872,-0.0170459449,0.1799539924,-0.0465855226,-0.1243836135,0.3421866596,-0.0906015411,0.3016500771,0.2360054255,0.0828581676,0.4873179197,0.0309320204,0.6077615023,-0.0773163289,-0.2693599463,0.5535903573,0.3471966088,-0.3623815775,0.3086406291,-0.0397058986,-0.2690623701,0.1196532249,0.2864121497,0.0299369562,0.4240362048,-0.2688582838,-0.0992740989,0.1516013741,0.1677065045,-0.1728445292,-0.1481738836,0.0579551607,0.2575570345,-0.4599102437,0.2317678928,0.3112021685,-0.1151481494,0.156272009,0.3843080699,0.0039483965,-0.2121084481,0.0566879921,-0.3574636579,0.4364005625,0.0000958723,-0.0321236737,0.180415839,-0.0024627035,0.4826472402,0.2176441848,-0.225059852,0.1874581873,-0.1264522374,0.1205261424,0.0969840586,0.2131741941,0.0326534286,-0.2434328943,-0.1718937308,-0.2731764019,0.2743750513,0.3453637064,-0.1747222692,0.2405582964,0.0511558503,-0.0985387638,0.2141356468,-0.0170268845,0.0022324608,0.1582268327,-0.132488519,-0.3309861124,0.159512043,-0.1174669117,0.3887009919,-0.1071969867,-0.4032920003,-0.3672438562,-0.4772162735,0.0425831191,-0.2066436559,0.1653785855,-0.0999653637,0.1763216704,-0.0519443937,0.2494451553,-0.2004780173,-0.135060966,-0.0154274553,0.3069825768,-0.1381405443,-0.1166273504,-0.15019539,-0.1173732951,-0.1203548834,-0.028892681,0.0037054943,-0.0938859358,-0.3903512955,-0.2269083858,-0.1927281469,0.4000966549,-0.2094475627,0.0106314486,0.0361664966,-0.4328805804,-0.1751117557,0.3084089458,-0.510522902,-0.4189543128,-0.3386988342,0.1622046232,-0.2392550856,-0.1590326726,-0.2101456225,-0.1539177746,-0.0257825088,-0.297717005,-0.0679370239,-0.4593394995,0.066123262,-0.3761543632,-0.0889949277,0.1172000244,-0.46005705,-0.208294034,-0.0396745801,-0.1002497524,0.2123406976,0.0334528834,-0.1382900774,0.0476342961,-0.1362772286,0.2307432145,-0.0242855586,-0.0594620593,-0.2722768188,-0.2320833504,-0.0442512855,-0.0388597511,0.0928495303,-0.3169508278,0.0343441851,-0.0749509707,0.404643178,0.2313245982,-0.1184696704,0.2528594136,-0.2151905745,-0.225058198,0.4088801742,0.066841647,0.097303316,-0.239780426,0.0692021325,-0.2193102688,0.1449562311,-0.1628499776,-0.0161281396,-0.0041746185,0.5557934642,0.0114316503,-0.0856287479,-0.0505718328,0.0189566277,0.2229174525,-0.1853208244,0.1959264427,-0.377017051,0.0508204363,-0.2732280195,0.439093709,-0.2399704158,-0.2291372567,0.069143638,-0.0709552988,-0.1686139107,0.4575914443,-0.1253303289,0.0845669806,0.2847536504,0.1448570341,0.1470871568,0.4999439716,-0.2506097257,-0.4530481994,-0.1584197134,0.0097183008,0.186177671,-0.0356305502,-0.0880455971,-0.0568735711,0.0108953062,0.265304774,-0.2652036548,0.1235499009,0.1738428026,-0.3929579556,-0.0181139112,-0.1742797196,0.087743938,-0.033560656,0.8921657801,0.1687954217,0.4038028121,0.1602562368,-0.1517969668,0.131570518,-0.0135127008,0.1269462109,0.1719961166,0.1579272151,0.035810519,0.3080631196,-0.1634715497,-0.0871258974,0.0040982887,0.0815083906,0.3480783403,0.0909014121,0.1321678311,0.0951317251,-0.4533497989,0.1085838899,0.2875907123,-0.2044256032,0.3006251752,-0.0027225383,-0.3478746414,0.0968064591,0.0687792227,-0.1036333442,0.2312446088,0.1464685798,0.6274519563,-0.0539617874,0.2600464523,0.1208144203,-0.1754732877,-0.1735768914,-0.0345090143,0.1754893064,-0.3216808736,0.2472686321,-0.2865734696,-0.0330845788,-0.1837985516,-0.6172889471,-0.2467290163,-0.0972245112,-0.086363785,0.2397946715,0.1439485401,0.0494822413,0.1944263726,-0.1363151819,0.1318318397,-0.3939236104,-0.0659630224,-0.1660009921,-0.4110151231,0.0342337117,0.0789111704,-0.1280239671,-0.0826188922,-0.3824841082,0.0980388969,0.0232926235,-0.4178722501,0.1835727692,-0.2367060035,0.2517473996,0.1566994041,0.0464332849,-0.1301650405,-0.1829271317,0.2820117474,-0.2733283043,-0.3171103001,0.2481305152,-0.1968504637,-0.0255361069,-0.0423546024,0.001637207,-0.4315136969,-0.1270964444,0.0063964021,0.0411540382,0.0813728124,0.4008952379,0.3015103936,0.2564197183,-0.0694132,-0.0516866073,0.2051507384,0.1729595214,0.401971221,-0.0453340113,-0.246250391,0.1988833547,-0.2110725939,-0.1946014017,-0.0431774817,-0.2507724762,-0.4114255905,0.0060346811,0.5550536513,-0.0497529954,-0.0697415993,0.3415048718,0.0640967414,0.0203942601,0.1028010398,-0.057550367,-0.4246288538,0.3636845946,0.1950739473,0.0209975354,-0.0470951125,-0.0910297558,0.5344064236,-0.3050165176,0.0196943264,0.3806328475,-0.3165952563,0.0085187927,-0.0075039733,-0.3508319855,0.0020642059,-0.2164058238,-0.2069074363,0.1151534244,-0.2021021098,-0.199947238,-0.2324251384,-0.0081792064,-0.1488262713,-0.2450309694,0.1605930328,0.1723382026,0.2058907151,-0.1472098231,-0.0359695964,-0.0803489611,-0.1351228207,0.1414121091,0.0991822407,-0.2391814739,0.0221705828,0.3737478554,-0.4715344012,-0.1865340173,0.2090768665,0.3911676109,-0.0005617907,0.2155106962,-0.1019380316,0.0302234422,-0.0597649887,0.4634606838,0.332999438,-0.076412946,0.2497055829,0.106395483,-0.1925335079,-0.4660725296,-0.1530637443,-0.3820136786,0.0467143022,-0.3956745267,-0.0991568044,0.1917426884,0.3078096211,0.354278475,-0.0524161458,0.2326931357,-0.3907874525,-0.5037890673,-0.1083224267,-0.1856617481,0.0735420585,0.4067328274,-0.0511603653,0.2021992505,-0.052249711,-0.1719228327,-0.055118531,0.2742959857,-0.0263556708,-0.111403659,-0.0432220139,-0.2032582015,-0.1201069206,0.4974148571,0.3016121089,0.4307509661,-0.5198167562,-0.2346646637,0.1539543122,0.2733048797,0.0282205511,-0.2632871866,0.1659512222,-0.559669435,-0.1724910587,0.241160959,-0.1367806643,0.358876586,0.0973062068,-0.0596351102,0.2025464922,0.5289083719,0.0248945672,0.0749860704,-0.245906949,0.4537625611,-0.0983142629,0.0699341223,0.0436736681,0.8350268602,0.1248506531,0.0305072814,0.641688168,-0.2158602327,0.2537650764,0.0945035517,0.1731382161,-0.2961376309,-0.1071064249,-0.1075477973,-0.2412971407,0.3092472851,-0.5395001769,-0.1265194714,-0.0160753522,-0.0628821999,0.1311442703,0.3468616009,-0.4278757274,0.0938196033,-0.0330335349,-0.2038721889,0.1121441424,-0.2138210088,0.3966593444,-0.2033079416,0.0648427308,-0.2075214088,-0.0953239948,-0.3254086077,0.2572106719,-0.3843839467,0.0723648667,0.4309488535,-0.3136196733,0.0897597671,0.1476743668,-0.1038964465,-0.3308497369,-0.1381653249,0.1379954666,0.2811755538,-0.1602493376,-0.0330578946,-0.0758945122,0.4511496127,-0.2567560375,-0.2407038212,0.3248722851,-0.1724440306,-0.0252042748,0.1689875126,0.2126978189,0.312954098,-0.063405782,0.0474485643,-0.1949811578,0.4568570256,-0.1830202341,0.0838970616,0.0512905419,-0.1537175924,0.1318116933,-0.0672405511,-0.0820612311,-0.093393974,0.142264545,0.1050036475,0.167774722,0.4702965021,0.0916435495,-0.0756587312,-0.0608483553,0.1444073915,-0.1836295128,-0.4542887807,0.0926684514,-0.05914177,0.0339045003,0.0791583657,0.5761159658,0.2180941701,0.0987731293,-0.0183402821,-0.3346140981,-0.0040292363,0.1269530505,0.144427374,-0.2800020874,0.1589998007,0.4735196233,0.3088252842,-0.2921720743,-0.2072615474,0.1344074607,0.0139316851,0.111437425,-0.1984357983,-0.0344798304,0.1641917825,-0.0530163385,0.1802767664,0.1762223244,0.0299032107,-0.1960486174,-0.2763878107,0.1646424532,0.0262452029,0.0828449279,0.2277648747,-0.0142278057,0.1288473457,0.1400894076,-0.0335574076,-0.0232177731,-0.31327492,0.0749504119,0.4686215222,-0.0035445965,-0.0383273959,0.1672213972,-0.182165131,0.4613619447,0.1279820949,0.1962504685,-0.004873964,0.1413667798,-0.0738826543,-0.0500204302,-0.5202831626,-0.0969450846,0.300270617,-0.5328297615,0.0007334877,-0.4318061173,-0.0064554857,0.4072599709,-0.3208839893,0.0898788646,0.402859807,0.1345389634,-0.2542098165,0.068184115,0.0502434708,-0.0042676083,-0.1172659546,0.2047824711,0.1054388806,0.0983835459,0.1125990227,0.2177937478,0.0951183811,0.4087663591,0.0784460381,0.3108332753,-0.2199385017,0.3216120005,0.3078154027,-0.0826826096,0.070075348,0.049185019,-0.248876825,0.1971558034,-0.14394252,-0.0132074812,0.3123978674,-0.2327858806,0.0940205976,-0.0964866802,0.1734163165,-0.2089202404,0.3062186837,0.4925253093,-0.0597414225,-0.0954320654,-0.0726951435,0.4616529644,-0.0985453501,0.2110271603,0.2701740861,-0.0040653273,-0.0381481051,0.0233471505,0.0364448279,-0.1881675571,0.2913278043,0.2702759802,-0.25668782,-0.2534961998,0.1303738803,0.0896568969,0.2583937347,-0.1812179387,0.0337640569,-0.0694994405,-0.2266219258,-0.3217248917,0.3388970196,0.5313584805,0.3796490431,-0.1952034235,-0.3217888772,-0.1860153973,-0.0729911476,0.1196313277,-0.130895853,0.2220329791,0.6603367925,0.1193673611,0.1698138267,-0.1256334782,0.1839614958,-0.0307588447,0.0563009828,-0.1492141783,0.2286603898,-0.1467947811,-0.2453062236,-0.4656848907,0.1232056618,-0.0555101708,0.3984955847,0.0743235126,0.0825470909,-0.1017229483,0.0300504062,0.0563483536,0.1842734516,0.3445960283,0.4053149223,0.3979044557,-0.1108897701,-0.104443498,-0.2590641677,0.442345947,-0.2495420873,0.0895811617,0.1316608042,-0.0088868095,-0.1571729183,0.3125434816,-0.0154559417,0.247641623,-0.0797139108,-0.0851190761,-0.3925023377,-0.2460024953,0.107151553,-0.205327794,-0.0182974227,-0.1419058144,0.3042200506,-0.6519694328,0.1224004105,-0.2782100141,-0.2564264834,0.0777498186,0.0453955829,0.6552137136,-0.0219034348,0.3961552978,0.1705237776,-0.027098123,-0.0799271762,-0.3554822505,-0.0862280354,0.2046666741,-0.0640858933,0.1238133758,0.1020912752,-0.0947146192,-0.3230880499,0.4325676262,0.0517914146,-0.0371951237,-0.1337218583,0.1391983181,0.0638655946,0.0863116235,-0.117784597,-0.0463273525,-0.1643606871,-0.0378949754,-0.327434063,-0.5092481375,0.3003089428,-0.4232924581,-0.3363003731,0.0668110475,0.184718594,-0.2212514281,-0.1529459208,-0.3810843825,0.0995872021,0.311221242,0.131325841,-0.3076262772,0.2013766319,-0.0241783131,0.0009544167,-0.0122423396,0.4474137723,-0.1461007297,0.0542349778,-0.1215178072,-0.0445530042]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1781","title":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import ","comments":"Maybe colab moved to pyarrow 0.16 by default (instead of 0.14 before)?","body":"I'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n","comment_length":12,"text":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import \nI'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n\nMaybe colab moved to pyarrow 0.16 by default (instead of 0.14 before)?","embeddings":[-0.1645145416,-0.0720891654,-0.0750411004,0.1178448275,0.1584094018,-0.1295015663,0.2293434441,0.2779433429,-0.3452712893,0.1958843917,-0.0898593739,0.3200353086,-0.2103375047,-0.0961442664,0.098879993,-0.0435224138,0.3291444778,0.3521497846,0.1217684373,-0.0262893159,-0.2683095038,0.2418088913,-0.3221585155,0.2537873983,-0.3110311329,0.2383453846,0.0224266928,-0.2041633278,-0.2748620808,-0.5592814684,0.4930470884,-0.3512657583,0.1850106269,0.035177473,-0.0001165204,-0.0361651815,0.3318021297,0.0065788408,-0.0864367709,-0.053093791,-0.0051921196,-0.0487096682,0.2911211252,0.048387263,0.1807696223,-0.0986343995,-0.028642619,0.2645857334,-0.1578989625,0.3618157804,0.2438595295,-0.0714592785,0.4829968512,0.0073237382,0.4845308959,-0.1620030105,-0.1685479879,0.6063100696,0.3403577805,-0.4351582527,0.2559783459,-0.0469060726,-0.3364959657,0.125969559,0.3174851835,-0.0463345051,0.4476968646,-0.2834878862,0.0438661575,0.0775088295,0.1891506314,-0.0342324264,-0.0471382886,0.0933558643,0.2436902076,-0.4170853198,0.1674913913,0.2858343422,-0.0423188657,0.1504784226,0.3508407772,-0.0897436738,-0.1016161665,0.0680699423,-0.4133269489,0.5771668553,-0.0279990882,-0.0376078263,0.2959961891,-0.0085381679,0.5448233485,0.2162516415,-0.1618939787,0.222033605,-0.0583505556,0.1888456941,0.1605658531,0.4117749631,-0.0708716735,-0.2710757256,-0.2219688892,-0.2425484657,0.179050982,0.3104254007,-0.2410864681,0.2709750533,0.0869703963,-0.1587501466,0.2051494271,0.0565389805,-0.0376949273,0.2294587791,-0.2030162066,-0.4689817131,0.1330039501,-0.1200959012,0.4012034535,-0.1132640466,-0.3020500541,-0.3518182933,-0.4912129939,0.0280248094,-0.2047780305,0.1603088379,-0.1871050894,0.1799094379,-0.0155371735,0.1593592465,-0.1768813133,-0.1288703084,-0.0001186189,0.2986554205,-0.1199791506,-0.1555567086,-0.1402951479,-0.2143132389,-0.0508267619,0.0276488625,-0.0248200111,-0.0860241503,-0.4432329535,-0.2307029217,-0.1609699726,0.3282334208,-0.214696005,0.0410073102,0.1438892037,-0.5303807259,-0.1504500061,0.348110646,-0.5548686981,-0.5257501006,-0.2722967565,0.1597962677,-0.1449406892,-0.225527972,-0.3228105605,-0.1240791157,-0.0706610084,-0.3261315227,-0.0970707089,-0.5662430525,-0.0398145169,-0.3641531467,-0.1322649121,0.0701281428,-0.4691922069,-0.1757370234,-0.1136256307,-0.0610620715,0.2524588704,0.0115915807,-0.19016698,0.0091432752,-0.1501307189,0.1888745725,-0.010618398,-0.0925845876,-0.269382745,-0.233519122,-0.0171880126,-0.0711429715,-0.0104374159,-0.3427701294,0.0541403294,-0.0934526473,0.5391680002,0.1510398537,-0.0919262692,0.2365180105,-0.227668792,-0.3093230128,0.3867648244,0.0412511453,0.1334745288,-0.1325006187,0.0893353,-0.2360306382,0.1637835205,-0.1645265371,-0.147960797,-0.0704261884,0.5804980397,-0.0270812791,-0.0900834426,-0.003732942,0.029752139,0.1618443578,-0.107944034,0.1411731392,-0.3013646007,-0.0108052874,-0.3366024494,0.3884294331,-0.1777534038,-0.2031802833,0.0466581509,-0.0819423199,-0.1706157625,0.4494848549,-0.0259675253,0.0750542656,0.2065198272,0.126603812,0.0992998332,0.4277631044,-0.2595004737,-0.3804759681,-0.0804548562,-0.0207439642,0.1841564029,0.0090517616,-0.109522976,-0.0390950181,0.0093241138,0.334451884,-0.2505614758,0.2435173243,0.1193105206,-0.4310282767,0.049314592,-0.1776409149,0.0438721403,0.0354819298,0.9647497535,0.1740974635,0.3293619454,0.1306004822,-0.1022682339,0.1376475096,0.0047586132,0.2163223624,0.1980262548,0.1379887015,0.0545954853,0.2922383845,-0.024054274,-0.0614387952,-0.0388668478,0.1256191283,0.3483796716,0.1695744395,0.1889870912,0.0469826162,-0.4144162536,0.0259662904,0.2890537381,-0.143115893,0.3453141451,0.0142147625,-0.3387793601,0.1451757252,0.050350368,-0.1385845244,0.3333581388,0.0864160806,0.648167491,-0.0729345158,0.261235863,0.1200967655,-0.2495345622,-0.2060934007,-0.1258096695,0.2833065987,-0.2414136827,0.2103032619,-0.3985242248,-0.0551562645,-0.254514128,-0.5509348512,-0.1973472089,-0.1725238115,-0.1139546484,0.2784939408,0.1574025154,0.0771273896,0.1897511184,-0.054373499,0.0486665368,-0.3254787028,-0.0577231571,-0.2021348476,-0.3555716872,0.0475757234,0.0170963146,-0.1840283871,-0.1104161218,-0.4024191201,0.070791252,0.0826471001,-0.4530937672,0.1303321123,-0.184988454,0.2781078815,0.2048248202,0.0560009889,-0.0858312398,-0.2264006138,0.2063110173,-0.2688762844,-0.2403257489,0.0881053209,-0.2671324611,-0.0355247445,-0.0089013074,0.0303281061,-0.456562072,-0.1765736938,-0.0519555397,0.0684290305,0.0857642367,0.443420589,0.3146209121,0.2522636056,0.0910166949,-0.0427558683,0.1634216905,0.1564957201,0.415851593,-0.0812081918,-0.1460250467,0.1678746045,-0.2114142925,-0.1200488582,-0.1240275875,-0.2911477089,-0.456800878,-0.0699446052,0.4132165611,0.0348858796,-0.0707118884,0.3015021086,0.172379896,-0.0039202622,0.1330003887,-0.1222164705,-0.5223017931,0.3581734896,0.1809668392,0.0018428396,0.02258664,-0.1016334444,0.4859682918,-0.3755908906,-0.0019701815,0.2666122019,-0.296900779,0.0369150117,-0.0639351457,-0.3988137245,0.0626056641,-0.1733404398,-0.2596685588,0.189924866,-0.2547605634,-0.1864781678,-0.1708739698,-0.0320366211,-0.2070043981,-0.2131971568,0.1703486741,0.169105351,0.1547159255,-0.0949320793,-0.1183284447,-0.1442456543,-0.1638818085,0.1148489565,0.1288942546,-0.2713195682,0.0515943356,0.4013968408,-0.5058498979,-0.1804200858,0.3568184376,0.3850039542,-0.0096379565,0.2406649888,-0.0666891113,-0.0454880968,-0.0516025759,0.3270597756,0.2610031962,-0.0815387219,0.2016935945,0.1343266368,-0.1634489596,-0.4492335916,-0.0890724212,-0.3102241158,0.0476914644,-0.4662247896,0.021020567,0.2243335694,0.4201574326,0.3731260598,-0.0159937255,0.421843797,-0.3733371496,-0.5288704038,-0.1109474078,-0.1720677018,0.0216839891,0.4257792234,-0.1639861614,0.2517544627,-0.0238130614,-0.2701671124,-0.0161001384,0.2492891699,-0.0038485397,-0.0502117649,-0.1989298612,-0.2487881333,-0.0994682014,0.5532004833,0.323592782,0.3339138925,-0.4476071,-0.1881608367,0.1667434871,0.3165079653,0.0154609904,-0.2832193077,0.138192907,-0.49151963,-0.1377989501,0.2868101299,-0.0673117861,0.3222185373,0.1671177894,-0.0857474282,0.1504503787,0.5431862473,0.0558925383,0.0305470955,-0.137306869,0.4470208883,-0.1607065946,0.0755336657,0.0513578802,0.8093291521,0.1972099841,-0.0713917837,0.580876708,-0.1713638008,0.2217870802,0.083036378,0.1935179532,-0.2672306597,-0.1059105992,-0.0959535837,-0.2172439098,0.3515286744,-0.4754901826,-0.154272005,-0.0269255489,-0.0830905437,0.0914276391,0.3505363762,-0.3953585625,0.1052301601,0.0428789333,-0.2826353312,0.1095426753,-0.2515359521,0.3893062472,-0.2843337357,0.1771463901,-0.1845501065,-0.0952226892,-0.3193155527,0.2185672075,-0.4056266248,0.1615051478,0.3519607782,-0.2274359167,0.0974299312,0.2358255237,0.0349524058,-0.253505528,-0.1072574407,0.1067542955,0.1945320666,-0.1750321388,-0.047618378,-0.0574366078,0.3682474494,-0.2682172954,-0.225393042,0.2945694327,-0.1152777523,-0.1393019557,0.1371951848,0.1879735589,0.2379244715,0.0066986391,-0.0812988579,-0.1269926578,0.4297590852,-0.1974842995,0.076378122,-0.0354041979,-0.061693307,0.1408201009,0.0325393714,-0.1347824931,-0.1178279966,0.1311802119,0.0809430704,0.1294774115,0.4460105002,0.0234846137,-0.1296998858,-0.0721730515,0.1058906019,-0.2622688711,-0.3670963347,0.1818706095,-0.1171014011,-0.0236315019,0.1838139296,0.7160930634,0.1958960891,0.1661732942,-0.0121193109,-0.3122233748,-0.0168309379,0.2320231497,0.0953228697,-0.2704334855,0.194060415,0.4931548238,0.296168685,-0.1812768131,-0.2069212943,0.2426740676,0.0556959733,0.1666335762,-0.1505984068,-0.0460640565,0.1222340465,-0.0153711624,0.1434961557,0.1723550409,0.0071765222,-0.2187000811,-0.3212631047,0.1763138622,0.0901571289,0.072399728,0.249682799,-0.0267795902,0.0720805302,0.1492034793,0.007686941,-0.0220131706,-0.2190498412,0.0418791734,0.4887138903,0.0051506301,-0.0255327541,0.1734695286,-0.1277794093,0.5053605437,0.1628074348,0.1249473765,0.029544916,0.0718356073,0.0116689689,-0.0542907231,-0.4572419226,-0.0116545726,0.2815454602,-0.5142031312,0.0324298963,-0.4295419157,-0.0797766671,0.3085459173,-0.2143480629,0.0243989564,0.3098647892,0.1103944108,-0.3148319423,0.0323338583,0.1988336891,0.0396808386,-0.081750527,0.2109107226,0.1648621559,0.1701039523,0.1953655779,0.1832239628,0.1144109592,0.3284378946,0.1693155169,0.2213565409,-0.1877636462,0.368157357,0.379311651,-0.0877244398,0.138752237,0.1206348464,-0.3023947775,0.2204647362,-0.0349951647,-0.1071091443,0.3906280398,-0.3080329597,0.1927025616,-0.0338147953,0.1121931747,-0.220267728,0.2605233788,0.4402599037,-0.0888254717,-0.1076513752,-0.0281267818,0.4338366687,-0.1158691347,0.085378021,0.2411759496,0.0129895052,-0.0778769851,0.0744092539,-0.0093867732,-0.195149824,0.2820471823,0.2406565994,-0.29490152,-0.2643832564,0.0936382636,0.0406804495,0.2444376796,-0.1022985876,0.0063968007,-0.103298232,-0.2316746116,-0.2962755561,0.4479668736,0.5408077836,0.3880316317,-0.1276355833,-0.359662801,-0.1721631587,-0.095367685,0.0811316594,-0.220889926,0.3042082191,0.5926116109,0.0316482969,0.2035863698,-0.1521673352,0.2760887444,-0.0707886666,-0.030653147,-0.0974317566,0.1334453672,0.0361874849,-0.2450038344,-0.4429077804,0.1917596608,-0.0117918029,0.4361614287,0.0876753554,0.0308965873,-0.15337722,0.038955722,0.0542021431,0.2537040412,0.3131489158,0.4380608499,0.4988467097,-0.0414690599,-0.1986476928,-0.3172761798,0.4085575938,-0.1357423514,0.1763887256,0.0518952943,0.0108982977,-0.0829534382,0.3229416907,0.0146292876,0.1371672601,0.0077765407,-0.2083167881,-0.2862458527,-0.1632740051,0.0788174495,-0.2364974022,-0.0498275012,-0.156491667,0.2980408669,-0.7075690031,0.0886761174,-0.2970641255,-0.2789492905,-0.0010191388,-0.0639748946,0.6198538542,-0.0459320247,0.4054369628,0.1122917011,0.0199534725,-0.0881473273,-0.2967690527,-0.0725334287,0.2547629178,-0.0204425976,0.2260514051,0.098268114,-0.0276957098,-0.2933853567,0.4293963015,0.1147017255,0.0199770741,-0.2061464936,0.1117843464,0.0711825192,0.0128484387,-0.132445544,0.056291379,-0.1074809656,0.0188223179,-0.2853554189,-0.5000659227,0.3295197487,-0.2990777194,-0.3212665021,-0.0214895159,0.2275518477,-0.1140553281,-0.1692272127,-0.4221136868,0.0542082675,0.2937287688,0.1044339836,-0.272325933,0.2133916169,-0.0256383084,0.0405002572,-0.0573666841,0.4282317758,-0.1077278331,0.0799712315,-0.1702286452,-0.1052582487]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1781","title":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import ","comments":"Installing datasets installs pyarrow>=0.17.1 so in theory it doesn't matter which version of pyarrow colab has by default (which is currently pyarrow 0.14.1).\r\n\r\nAlso now the colab runtime refresh the pyarrow version automatically after the update from pip (previously you needed to restart your runtime).\r\n\r\nI guess what happened is that Colab didn't refresh pyarrow for some reason, and the AttributeError was raised *before* the pyarrow version check from `datasets` at https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/__init__.py#L60","body":"I'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n","comment_length":72,"text":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import \nI'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n\nInstalling datasets installs pyarrow>=0.17.1 so in theory it doesn't matter which version of pyarrow colab has by default (which is currently pyarrow 0.14.1).\r\n\r\nAlso now the colab runtime refresh the pyarrow version automatically after the update from pip (previously you needed to restart your runtime).\r\n\r\nI guess what happened is that Colab didn't refresh pyarrow for some reason, and the AttributeError was raised *before* the pyarrow version check from `datasets` at https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/__init__.py#L60","embeddings":[-0.4181626439,0.150439024,-0.019076623,0.1533482373,0.1342275739,-0.0236932579,0.2542464137,0.2454890162,-0.3241707385,0.1116399616,0.0381206125,0.3774304986,-0.0813133493,-0.0478629395,-0.0118858498,0.0418873914,0.3027802706,0.3409284055,0.2071904242,0.0259101894,-0.2342766374,0.0909717605,-0.3210463226,0.2035171092,-0.4118345976,0.2203688473,-0.0174811408,-0.1700021774,-0.3172970712,-0.6792119145,0.5830789208,-0.2323222309,0.2768806219,0.1425857395,-0.0001209328,-0.0209162738,0.3580555916,0.1024649292,-0.2893368304,-0.1401554644,-0.1511045843,-0.1866816431,0.3294048905,0.0138392476,0.2098614872,-0.2561298907,-0.0134524936,0.2181422859,-0.080727227,0.2746795416,0.1697658896,0.0567860603,0.4970565736,0.0424643382,0.3831528127,-0.071018897,-0.1261998862,0.5384056568,0.2573763132,-0.3175978959,0.3273626864,0.0955689847,-0.3002936244,0.2453082353,0.3384176791,0.0261107702,0.3156374991,-0.1950557381,-0.0080030542,0.2326883525,0.4724449515,-0.2055785507,-0.3742718995,-0.1450867057,0.2462435961,-0.3504833579,0.2955636084,0.1823577136,0.0134681668,0.1489953399,0.1497543305,-0.0184530206,-0.066981405,0.0797984451,-0.3698368967,0.4304915071,-0.1151573136,0.0948726386,0.1171312258,-0.066792652,0.5531333089,-0.0353317596,-0.2178254575,0.2489418089,-0.2337918431,0.1079049408,0.0849388614,0.383694768,0.0191436578,-0.1634806246,-0.2799125612,-0.3150515556,0.0434298851,0.3021691144,-0.0314206555,0.2702073753,0.0722960085,-0.0455675982,0.2369808257,0.0130285798,-0.0343974493,0.1162366569,-0.1001426578,-0.3603296578,0.2520412803,-0.0574507527,0.6233623028,-0.2348875254,-0.420814544,-0.2791058421,-0.4959393144,-0.1293876171,-0.1401707828,0.1667962372,-0.2977566421,0.2598457634,-0.1432746053,0.1083350405,-0.25399068,-0.1861823797,-0.0646512508,0.1568896621,-0.1410632133,-0.1038216278,-0.0783214942,-0.1958213747,-0.0038999112,0.0601154342,-0.1246229187,-0.0684761256,-0.2699031234,-0.3182118237,-0.1636870354,0.4307076037,-0.2343376428,0.1612659991,0.1176150292,-0.5139939785,-0.1687964052,0.2821596265,-0.5118224025,-0.4434715807,-0.2082923055,0.0998681337,-0.3005480468,-0.1936513484,-0.4257184565,-0.0990353525,0.0882480592,-0.481126368,-0.0242845472,-0.5677361488,0.0094444519,-0.3748410344,-0.0311756358,0.2954781353,-0.6035419106,-0.1913348883,-0.1196309403,-0.143869862,0.1422236115,-0.0320601985,-0.1355499476,0.0877708271,-0.0691845268,0.0052972008,0.0978054181,-0.201942578,-0.4289782643,-0.12813209,0.0391551964,0.0890583992,0.0629899874,-0.1787740737,0.0733823925,-0.0738649815,0.3456241786,0.2120342851,-0.1600933373,0.0766813233,-0.094631426,-0.1960214227,0.3323121667,0.1353757977,0.0791467503,-0.0963715017,0.1051179245,-0.1528168619,0.1042477936,-0.0824322104,-0.0228677914,0.0151800336,0.5127726793,0.0129785007,-0.1164855957,-0.1611536592,-0.1859289855,0.2242620438,0.0850028098,-0.0598823577,-0.4205748737,-0.1836570352,-0.2416515052,0.2941110432,-0.1884559542,-0.2594415545,-0.0057253679,0.1202157661,-0.1577471048,0.3912128806,-0.1406080425,0.3141974211,0.0699943304,0.2449356467,-0.0793026537,0.3649939895,-0.1673131436,-0.3892548084,-0.1534980088,0.0461947657,0.1388100386,-0.0449939333,-0.1133975834,0.1098341197,-0.0093778027,0.2966071069,-0.2214619964,0.1578570604,0.1264435202,-0.1821916997,0.0310812052,-0.3086495996,-0.0047076247,0.0026354652,0.6934741139,0.1561753899,0.2425137311,0.1873312294,-0.1939786822,0.0710175335,0.0452383682,0.1769446135,0.0338250846,0.0847995728,-0.0266789943,0.2097433507,0.024180308,0.0152949709,-0.124628,0.1560415477,0.3652108312,0.1491890252,0.1532502174,-0.0186699107,-0.4601625204,0.0689493641,0.383128494,0.1304515749,0.4296892881,0.0333592482,-0.3057473898,0.0427863896,-0.0728984252,0.0013502085,0.0672047064,0.1850849092,0.5060129166,-0.0201589353,0.2183589935,0.2089821696,-0.1588149071,-0.3102003336,0.0411301889,0.3739798367,-0.3386669755,0.2943220735,-0.3655646145,-0.111781247,-0.1900116056,-0.6143972874,-0.2298163772,-0.139765203,-0.1837114692,0.3203368485,0.2036804259,0.1484705657,0.0623821504,-0.135220319,0.0419023074,-0.4303118587,0.069035545,-0.1785099655,-0.2635135353,-0.0004695377,0.0800385773,-0.1433688253,0.0108111631,-0.4225106835,0.0828151554,-0.0152948219,-0.3495781124,0.0856365636,-0.3275427222,0.4431426823,0.2561699748,0.1518147588,-0.2441195846,-0.1336629242,0.3365137875,-0.217584908,-0.2805583775,-0.0804267302,-0.1813363135,-0.0858778656,0.0960069001,-0.0889935195,-0.4324345589,-0.1147843525,0.0239354949,0.1166306436,0.0384806171,0.3249641657,0.2670057714,0.1228089854,-0.0381520726,0.0272235703,0.0604037233,0.0145734949,0.3432318568,-0.043967396,-0.0917213112,0.2418819517,-0.1861955971,-0.1578913778,0.0988411903,-0.3921024799,-0.606420517,-0.1058822721,0.4493834376,-0.080487445,0.0285757277,0.3183552325,0.210398972,0.05888227,-0.0014327982,-0.1303637624,-0.5016965866,0.2687036991,0.1951235384,0.0392837301,0.1592850536,-0.11933209,0.5885620713,-0.1180830076,-0.1307075024,0.1967376322,-0.2780768573,0.2577849329,0.0037795489,-0.3999598026,0.0001089875,-0.1838212609,-0.17223607,0.0561543852,-0.0889739618,0.0206062961,-0.1776569784,-0.1040868312,-0.204578504,-0.1649920344,0.0485991724,0.1484631896,0.2544666827,0.0124209905,0.0146217225,-0.19893004,-0.1270336807,-0.0133111048,0.3179433346,-0.1033573076,0.0975216329,0.2111834139,-0.3801303506,-0.1920144409,0.3192766607,0.3189559877,0.0780694336,0.2148048133,0.0115080494,0.0117196571,-0.0136026982,0.3773003817,0.1215167716,-0.0888893828,0.1494613141,0.0307090636,-0.2940443158,-0.3276789188,-0.0779706612,-0.199932009,-0.0321918949,-0.2915962338,-0.1149149239,0.0800374448,0.478174299,0.486528337,0.0007785781,0.2729388773,-0.2663759291,-0.4203901589,-0.1045850813,-0.1371212602,-0.0841327086,0.2471646369,-0.009054631,0.3496675789,-0.2689824402,-0.1279772967,0.0009084608,0.2837114036,0.1002658755,-0.0145211462,0.0328851901,-0.0677738264,-0.0991422981,0.6560024619,0.5683662295,0.2228669822,-0.5871340632,-0.3114333153,0.2615457177,0.3081309199,0.0385411754,-0.3654317558,0.1598126888,-0.4349407256,-0.2039362043,0.1181143224,-0.0032237123,0.3611794114,0.2568581104,-0.2989949882,-0.1709065437,0.6227719784,0.1870379746,-0.0497827269,-0.0120430794,0.4989397526,-0.1086793914,0.0867680311,-0.0708632842,0.7722873688,0.2161619812,-0.0164658278,0.5215468407,-0.0375140533,0.3984513283,0.0963153616,0.2034585625,-0.3084239364,-0.2251468152,-0.1771802604,-0.2383138239,0.4370369613,-0.3925249875,-0.1204816774,0.1257025748,0.1099550799,0.1674297601,0.2934601605,-0.2357234806,-0.097084485,0.0407844186,-0.2799022496,0.0703093782,-0.2062651217,0.4521021247,-0.2394184768,0.19165501,-0.17444022,-0.1181278899,-0.3269029856,0.0272019077,-0.4223391116,0.0469849966,0.3487402201,-0.3465306163,0.1412333846,0.3082630038,0.1490027905,-0.1105612591,-0.1429972351,0.1925393939,0.1203573793,-0.0830731764,-0.0876866803,-0.1431604922,0.4585089386,-0.0869705677,-0.2914060056,0.1181493849,-0.1757516265,-0.1169962585,0.1927832216,0.1780388504,0.1803470403,-0.1606610864,-0.0084349588,-0.2092546821,0.4160994589,-0.206287384,0.0337353349,0.0291549563,-0.0308183823,0.2589293718,-0.0415641554,-0.2023432404,-0.2383624166,0.22516945,0.0538900271,0.2207307965,0.5160393119,0.0045142476,-0.1424064785,-0.0259941109,0.0368429199,-0.2040504664,-0.4193035662,0.0943090916,-0.0472629741,0.1737484336,0.1484474689,0.5579470992,0.1887614578,0.0739826337,-0.0120352181,-0.4998326004,-0.0175013449,0.2965069711,0.1705465466,-0.1402316391,0.2619215548,0.3161825538,0.1462996006,-0.2080896497,-0.1621001214,0.1114125103,0.0875496641,0.2187371105,-0.1014429331,0.1857637912,0.2712578177,-0.1186808199,0.0882631838,0.1107569858,0.0232381821,-0.1679696739,-0.1551795453,0.195230931,0.0790806487,-0.0309851542,0.144366771,-0.0506839193,0.0514103062,-0.0166523345,0.1229135022,0.1435354501,-0.2517283559,0.0204631966,0.3447132409,0.1018136889,-0.1064436361,0.2578164041,-0.1646614969,0.3178764582,0.1187177896,0.2142812312,0.0441243276,0.1659112424,-0.0493739881,-0.036911957,-0.3980924189,-0.0909594297,0.2025741488,-0.5558549166,0.0694073588,-0.2835442722,0.2389627993,0.4619525075,-0.2067473382,0.0361855105,0.487744838,0.074722603,-0.2718813717,0.1139427871,0.2718523443,0.1040509194,0.0084565748,0.3226695657,0.1910163462,0.2318346053,0.0050497185,0.2871256471,0.2534640431,0.3062468171,0.2237193882,0.3408560455,-0.1183861345,0.3203251064,0.4664037526,0.1085290462,-0.0223922078,0.2567122579,-0.2122326642,0.2128281295,0.0245729517,-0.0816997215,0.19816643,-0.4361230731,0.3141304255,-0.0638900772,0.1463705301,-0.2770724297,0.359344691,0.420617342,-0.174863562,-0.0972498283,-0.1061295345,0.4280222356,-0.1015812457,0.0380421691,0.2254812419,-0.0710921362,0.076577805,0.0310191046,-0.0099998079,-0.2566748559,0.3153357506,0.1951400787,-0.3157757819,-0.4515143931,0.1660030484,0.1537614465,0.2545733452,-0.0800383165,0.0687818453,-0.0559652969,-0.1540690958,-0.1322260201,0.5119328499,0.6078239083,0.332393378,-0.0730936229,-0.2313455641,-0.3119511306,-0.0285724327,0.0980184972,-0.0754058436,-0.0635234937,0.3634943366,0.1849327087,0.1777666807,-0.1057398394,0.2683843374,-0.1154611781,-0.0876863226,-0.3277676702,0.3378293216,-0.077192232,-0.1504781097,-0.2973752916,0.266648978,0.006382762,0.275913775,0.2482229769,0.0690190196,-0.1253392249,-0.0256373081,0.0088808397,0.1134221777,0.3014828861,0.5731633306,0.3423219025,-0.0488622375,-0.032264486,-0.5246586204,0.2737059891,-0.3799618781,-0.012691767,0.1221082658,-0.1102099866,-0.1972216219,0.4066541195,0.1094076186,0.1486454755,0.0448764712,-0.0550349168,-0.3251521289,-0.3328723013,-0.0518054552,-0.2736149132,-0.0470761508,-0.284877032,0.2699871361,-0.6724924445,-0.0098381685,-0.1648218036,-0.1948522031,0.1339521259,-0.2329448462,0.6932852864,-0.0060903192,0.5324743986,0.1539001316,0.0189525448,-0.2397966236,-0.3865697384,-0.2289653569,0.3694833219,-0.1253312379,0.3657640815,-0.1208417341,-0.009807094,-0.2979019284,0.4230306745,0.0462107919,0.0518479608,-0.1854896545,0.1832358539,-0.0671849251,0.0773551315,-0.1123984531,0.0336336121,-0.0960216299,0.2061692029,-0.3234835267,-0.5417073369,0.4071372747,-0.4472004175,-0.2611281276,-0.1382684261,0.1575761437,-0.0201770458,-0.2910699546,-0.5523152947,0.1713495851,0.3971259296,0.087034665,-0.3133409023,0.1738685966,-0.010651988,0.0655374676,-0.0747493729,0.504085958,-0.3182087839,-0.0150989536,-0.0719003677,-0.0107887303]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1781","title":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import ","comments":"Yes colab doesn\u2019t reload preloaded library unless you restart the instance. Maybe we should move the check on top of the init ","body":"I'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n","comment_length":22,"text":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import \nI'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n\nYes colab doesn\u2019t reload preloaded library unless you restart the instance. Maybe we should move the check on top of the init ","embeddings":[-0.1401676834,-0.144001618,-0.038694378,0.2877447307,0.2086876184,-0.1575326025,0.1334634572,0.1926029772,-0.2747627497,0.1211347207,-0.0618401542,0.4179868102,-0.149158448,0.0373287387,0.1723472774,0.2285047174,0.2047673911,0.4507710338,0.1933162212,-0.0358727388,-0.0563337989,0.0911345482,-0.4517784119,-0.0227694139,-0.3740769923,0.1728886664,0.0168113057,-0.0385997444,-0.1438812613,-0.5761191249,0.5109921694,-0.2018954903,0.1176529378,0.0456297174,-0.0001169598,-0.0215618145,0.2957280576,-0.0815277025,-0.1828682423,-0.0304470025,0.103999503,0.0859086588,0.4524633884,-0.0387615412,0.1377262771,0.0620980412,-0.1428721398,0.2268146276,0.0426715799,0.2654359639,0.2162775993,-0.0245499201,0.2785424888,0.2467481047,0.3330410719,-0.0338498317,-0.1099255979,0.5772957802,0.3482242525,-0.3117033243,0.2219970673,0.0770806745,-0.3472619951,0.1299000978,0.5009498596,-0.1533364505,0.2511973977,-0.1369098276,-0.0713904575,0.0851653963,0.1308462173,-0.1807099581,-0.0385490544,-0.0226831529,0.3367497027,-0.4357969165,0.2943397164,0.1536544412,0.0482858643,0.1406814009,0.403832972,-0.1159274876,-0.0492080003,0.0561702885,-0.2977418602,0.5725084543,-0.0990440398,0.0298043918,0.2147823274,0.0776145682,0.5299271345,0.1352569312,-0.3125365674,0.2130876333,-0.1809976697,0.0572746024,0.1509382278,0.3558631241,-0.2009800971,-0.0605282858,0.0495058037,-0.1632780135,0.1300141066,0.3676849902,-0.142496407,0.3776366711,0.154883489,-0.0959658325,0.3594748974,0.1658726186,-0.2740063965,0.1589539945,-0.234902218,-0.3367799819,0.4507613182,-0.0824937448,0.3774913549,-0.2482677698,-0.3130382597,-0.2947343588,-0.5301507711,-0.1343030632,0.1129386052,0.2356717139,-0.2533465922,0.2633489072,0.0208705775,0.0620075278,-0.332431078,0.0141465422,-0.0426387079,0.3480690122,-0.2938275933,-0.0135670295,0.0325777046,-0.2584100962,0.1664141566,-0.1595478654,0.2574126422,-0.3906576633,-0.243710652,-0.2300013453,-0.1325031221,0.480109483,-0.2656368911,-0.1174909025,0.2300435603,-0.55451864,-0.2062151581,0.3420908153,-0.4835709333,-0.4480426311,-0.0692128018,0.1318643838,-0.2776029408,-0.1435279548,-0.1711693406,-0.3030305803,0.0746000558,-0.4376575351,-0.0581505597,-0.4439565241,0.0284816865,-0.3985586762,0.0237916056,0.1919599175,-0.4507101178,-0.1807584763,-0.0612381808,0.0278929882,0.1731995344,-0.0307787582,-0.2933194637,0.0139535349,-0.2074744701,-0.0857498795,0.082256712,-0.1066971645,-0.2593953907,-0.2776183486,-0.0503491536,-0.0017578917,0.0750196427,-0.2023127526,-0.0372993797,-0.1282960624,0.3937426805,0.2658712566,-0.0960765332,0.1671169847,-0.3055381477,-0.1047835797,0.235751316,-0.0243394673,0.1718826741,-0.0098852199,0.2645913064,-0.058047384,0.0703272074,-0.0737958774,-0.2017795295,-0.0384665318,0.5287053585,-0.0025241051,-0.0461851656,0.0231026895,0.0909546167,0.1952457875,-0.0939868689,0.0149220079,-0.3915699124,-0.012315142,-0.3587919772,0.3929842114,-0.2013744712,-0.226988703,0.0655969009,0.0814307183,-0.1832759678,0.4519284666,-0.1334312409,0.108577311,-0.0099228984,0.0554277226,0.027166307,0.4176918566,-0.1555547565,-0.4423778057,-0.2066223025,0.0164660048,0.0704902336,-0.0276796855,-0.2374277413,0.0117331594,0.0611475073,0.3287153542,-0.2297953218,0.1134170145,0.2255122662,-0.3239575624,-0.0130361123,-0.3053932786,0.2059909403,-0.0975382552,0.8549458385,0.2274992317,0.1324662119,0.1277952045,-0.1672811955,-0.0512829274,0.0174078308,0.0339290015,0.0438178144,0.1433740109,0.0099968398,0.3170184791,0.015631998,0.0575691387,-0.0104984194,-0.0532746278,0.2539075911,0.1852033436,0.1830993444,-0.039284151,-0.3725682795,0.0153125338,0.3927857578,-0.1042217761,0.3537482023,-0.0030459079,-0.4114221931,0.0008033758,0.1020718738,-0.1002085507,0.145644635,0.1736979932,0.8331612349,0.0913991332,0.2913973629,0.0990439206,-0.3745927811,-0.2292360365,0.0848100334,0.1412866861,-0.3006951213,0.1962329894,-0.2893141806,0.0214098971,-0.1627856046,-0.5309426188,-0.0738258511,-0.1034965515,-0.1243858039,0.4379892051,0.2560116649,0.1048152149,0.2743670344,-0.0679702982,-0.0091138976,-0.6351595521,-0.0303658582,-0.2195123583,-0.3502953649,-0.0034262158,0.2406198382,0.0253645908,0.0885530934,-0.2230242491,0.1152781844,0.196949169,-0.4630786479,0.0964266434,-0.1084730998,0.3668431044,0.034061335,0.1044906154,-0.1882028133,-0.2383419573,0.2162095755,-0.2671554089,-0.2033492029,0.0590681843,-0.0798357204,0.0052186437,-0.0440485701,-0.113650851,-0.2847565413,-0.1694559604,-0.1283544153,0.1254850924,0.0338844024,0.3986454904,0.3249465227,0.2000761181,0.1735307872,0.0766688958,0.122238405,-0.0927146524,0.323150456,-0.0346933454,0.0231620576,0.2990019023,-0.3058594465,0.0286030192,-0.1258528382,-0.4488914609,-0.5858923197,0.0146212727,0.2705054581,-0.0449821949,0.0228953138,0.3292979002,0.1701287329,-0.0099394657,0.1031591445,-0.1817325503,-0.393458873,0.2843803763,0.2490620166,-0.0437380001,-0.0391653478,-0.2207206339,0.5412877798,-0.4730611145,-0.034414988,0.3007108867,-0.0463123769,0.1220678017,-0.0938377604,-0.3473147452,0.0026158891,-0.1770345271,-0.3872644603,0.1603215337,-0.1921954304,-0.1894459426,-0.2503657043,0.0279520322,-0.2932752669,-0.2187057585,0.1310602725,0.0105107771,0.0252701472,-0.0920439363,-0.0690289885,-0.0432195477,-0.1032589749,0.0107467845,0.2959595323,-0.2220540047,0.0204843152,0.331979394,-0.6371978521,0.0114338864,0.078569144,0.2924874723,0.0161098074,0.2801786363,-0.1238994449,0.0581365898,-0.1154007241,0.2588289082,0.2669907808,-0.1222725585,0.187941879,-0.1569963992,0.0872341916,-0.4683198929,0.0024570308,-0.3910137713,-0.0371957645,-0.5108362436,-0.0007461057,0.3016482592,0.3158557117,0.2830547988,-0.0747000203,0.1845620424,-0.4773102403,-0.492888689,-0.4238076806,-0.1448927671,-0.0340278484,0.2792028785,0.0137782879,0.1574703753,-0.1167866588,-0.242143482,0.0762997046,0.3607854247,0.1209139079,0.1257701665,-0.049343843,-0.2595839798,-0.135041073,0.487064749,0.4426611662,0.293805182,-0.3357536197,-0.3080013096,0.2927727401,0.3107409477,0.043500457,-0.3941925466,0.1987721324,-0.4934760928,-0.1911955774,0.173110947,0.0440081432,0.1127988175,0.2055023462,-0.2596436441,-0.0505586155,0.5910872221,0.0034476607,0.0358360857,0.061752893,0.3198788166,-0.1341251433,0.2430290729,0.2231681645,0.9656525254,0.012014945,-0.054653462,0.4602909386,0.0327180661,0.2713984251,-0.0514922775,0.3049483299,-0.3573488891,-0.245804593,-0.0611770824,-0.2653256655,0.356057018,-0.6037139297,-0.2224780172,-0.0300631523,0.0151700657,-0.0567898862,0.1870259792,-0.2615447342,-0.0575116165,-0.0694346949,-0.2246145159,0.0461080559,-0.2375421077,0.553184092,-0.3224280477,0.2120822072,-0.0156269725,-0.102476567,-0.2567564547,0.1074157134,-0.3843137622,0.0819493011,0.1826650202,-0.1906039119,-0.0020575665,0.2088581324,-0.0480971336,0.0258954186,-0.0038115382,0.205335483,0.2875348032,-0.25091061,-0.2406321317,-0.0412313268,0.4544915855,-0.0739618838,-0.2232007235,0.1632983387,-0.1486967504,-0.1203098968,0.2079336643,0.2491942942,0.2404622883,-0.0666804984,-0.1194640771,0.0503870361,0.1910235286,-0.2349574715,0.0667046458,0.051996164,-0.1581678987,0.1981551796,0.1511411518,-0.3012104928,-0.1404926032,0.096822843,0.1954908967,0.1271107197,0.4786173999,0.0046555661,-0.0343614928,-0.1182939038,0.1016350761,-0.4566182196,-0.2994999886,0.2755730748,0.185428828,-0.0628446192,0.2035854459,0.5781240463,0.1484231502,0.2833882272,0.0016783556,-0.413729459,-0.0685865656,0.1447931677,0.1112540811,-0.0606706552,0.2027845681,0.5849716663,0.1415169686,-0.0208866224,-0.2094156146,0.0760336667,-0.0795700178,0.1131141484,-0.3074975312,0.1867481172,0.0430931225,-0.0073712924,0.1148912534,0.2219358534,0.0825929865,-0.2004008442,-0.0969476849,0.1931075305,0.039728988,0.3005503714,0.2324893326,-0.0833238363,0.0887920931,0.0431920886,0.0707473457,-0.045478195,-0.2720468342,0.0044179391,0.524897337,-0.0293157734,-0.1394841373,0.2138797641,-0.263438791,0.3821101189,0.0388786905,-0.0040826704,-0.0215801895,0.1026027501,0.0395444036,-0.0718550235,-0.4443147779,0.0025520136,0.1375374943,-0.4270157814,-0.0080808131,-0.4568271935,-0.0880281106,0.4660222828,-0.1164558679,0.1230644286,0.3373887837,0.090979293,-0.3088119328,0.1949314177,-0.0460893363,-0.0582606196,0.0389129929,0.4427160919,0.059505377,0.107446827,0.2294200212,0.2066303641,-0.0231666025,0.2925180197,0.1275272071,0.1681661606,-0.2863566875,0.2210294455,0.3486667573,-0.0868724361,0.0117008025,-0.0556897856,-0.1806059629,0.4193442166,-0.1166278273,0.1621092856,0.0819762573,-0.5037825704,0.3588400185,0.093225278,0.189332962,-0.1108092517,0.3356678784,0.411401093,0.0175601244,-0.1149590313,0.036513254,0.3873960376,-0.1941831112,0.2069128901,0.2968497872,-0.0104498966,-0.0987941027,0.1528921723,0.2635689378,-0.1121147797,0.1424230486,0.2765941918,-0.3965967596,-0.3444944918,-0.0058121542,0.0751990899,0.1885026544,-0.1558033973,0.0897364244,-0.0908199921,-0.1286638826,-0.2432130873,0.329436183,0.4504807889,0.4281196594,-0.079504773,-0.3229372501,-0.1604532003,0.0746457502,0.1047828794,-0.1451224685,0.2769164741,0.6709734201,0.0522963516,0.1748087257,-0.1952821016,0.3402732015,-0.0967914239,-0.3059833348,-0.1856523901,0.1936623901,0.2526377738,-0.1094963849,-0.4630556405,0.2653677166,-0.0621708632,0.3748278618,0.0647197589,-0.0029489961,-0.1486265212,0.0214040093,0.038571842,0.2800284326,0.3117568195,0.5116200447,0.3169768155,-0.1398620158,-0.1157506853,-0.5177516937,0.4010747671,-0.1786647737,0.1842114627,0.0220363103,-0.026788529,-0.0997658521,0.3125271499,-0.0283734109,0.119455874,0.0742438212,-0.1069629639,-0.2954262793,-0.3318929672,-0.0963867083,-0.3195107281,-0.0000720553,-0.2787986994,0.1629818231,-0.6408566236,0.0828795284,-0.180221051,-0.3284862638,0.1035502851,-0.0689845681,0.7786416411,-0.1724223644,0.3721227944,0.1072878167,0.0892624632,0.1509101242,-0.4144458473,-0.1447090656,0.0910985693,0.1098464057,0.3065454662,0.0231364537,-0.0199301895,-0.3485963047,0.352478981,0.1307270825,-0.0824345648,-0.2510973513,-0.006151896,0.0335112996,0.1983196139,-0.1538299471,0.3151010573,-0.1176366955,0.0559946969,-0.2864589095,-0.4610404968,0.4683490992,-0.223179847,-0.2698687315,0.0057845041,0.1729331315,-0.2487136275,-0.2186249495,-0.5525951385,0.0779676139,0.2595119178,0.1068583727,-0.2170872837,0.249126181,0.0512747765,0.0937593728,-0.1307453215,0.2270705402,-0.1303285807,0.1138667539,-0.193269521,-0.0613108762]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1781","title":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import ","comments":"Yes I'll do that :)","body":"I'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n","comment_length":5,"text":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import \nI'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n\nYes I'll do that :)","embeddings":[-0.1288739294,-0.0892390683,-0.0967089087,0.2205693722,0.1913802922,-0.1033518314,0.2272486687,0.2181507647,-0.2696265876,0.1800048053,-0.0922490433,0.2725971639,-0.2191487402,0.0275556836,0.1896382123,-0.0217199326,0.2833055854,0.3113927245,0.2577956617,-0.0183085483,-0.2165312916,0.1959256083,-0.2382953614,0.1642822474,-0.4343675375,0.2114052474,-0.0027451487,-0.1277509183,-0.2201936841,-0.4696079195,0.3564158678,-0.304569602,0.1827454865,0.0947647765,-0.0001128653,0.0161038339,0.281755954,-0.0371608101,-0.033738032,-0.0249360446,-0.0445993692,-0.0110739842,0.1847331375,-0.0703956634,0.1842019707,0.0013300687,-0.1211556867,0.2091176957,-0.0515668988,0.3512316048,0.2887635827,-0.0400661826,0.4468090534,0.0682536513,0.5097162724,-0.14429304,-0.1366010755,0.5307875276,0.3413736522,-0.3102930784,0.346450448,0.0656790808,-0.3242434263,0.1000508443,0.4597298205,-0.0064206305,0.5129779577,-0.3387315869,0.0352846347,-0.0021154471,0.1367950737,-0.0441388711,-0.0925378874,0.1108954921,0.244526282,-0.4143862128,0.1307378709,0.3320655525,-0.0706906468,0.1635658443,0.3650648594,-0.207807079,-0.1725714356,-0.0392548852,-0.405786097,0.4829260409,-0.0869214311,-0.045593027,0.2673347592,0.0188234225,0.4691580236,0.168049112,-0.1691512316,0.1330942512,-0.0828557909,0.1281613857,0.0113842208,0.3415560424,-0.0680243671,-0.2635437846,-0.1795231849,-0.2351904064,0.244913891,0.3850977719,-0.1887755543,0.276027292,0.0089208251,-0.079914853,0.1754148155,0.0024417781,-0.1411463916,0.132406801,-0.1357144564,-0.5255402327,0.1619138122,-0.0891619921,0.4122585952,-0.2526783943,-0.4017069042,-0.2738364637,-0.4161018431,0.0254066419,-0.2010581344,0.2346715629,-0.1521458328,0.196854338,0.0210320726,0.2119026482,-0.1628260165,-0.1110595837,-0.0223237798,0.4171533883,-0.1842987537,-0.1039751694,-0.1695175171,-0.1038749069,-0.0750958249,-0.0814103633,0.1466445774,-0.1573602408,-0.3418109715,-0.2419295162,-0.1943428963,0.3673195839,-0.1648342758,-0.1063691154,0.1157744974,-0.4986496866,-0.1563068777,0.3138448298,-0.5095929503,-0.3259477615,-0.1899779737,0.2202623039,-0.1889689863,-0.219527185,-0.301692009,-0.2053524852,0.0289197061,-0.2866484821,-0.0860308111,-0.4762461185,-0.0212961435,-0.3798209727,-0.1119621918,0.0878108293,-0.4314419031,-0.1856024563,0.0339449495,-0.0055372058,0.3142158687,0.121200867,-0.1310380995,0.0544683114,-0.1106324941,0.1835374236,0.0586586595,-0.1096459627,-0.2002599686,-0.2984018326,-0.1429006457,-0.1652211249,0.0872971192,-0.2449048012,0.0694185719,-0.1523195505,0.5518490672,0.3144082129,-0.1410886943,0.198401466,-0.3490147591,-0.1699159741,0.347142458,0.1022918671,0.2163587064,-0.1800457686,0.0717474967,-0.1888221055,0.0625150353,-0.2070928216,-0.1120619774,0.0246413685,0.558780849,0.0334417634,-0.1447115391,-0.1495828331,0.102219373,0.1159008369,-0.0898844078,0.1785422713,-0.3981632292,-0.0863922462,-0.368196547,0.4149836898,-0.3042022884,-0.1710462719,0.1252979338,-0.10727676,-0.1508001685,0.3812382519,-0.0998253524,0.1284338087,0.1798637807,0.1097147912,0.0207922868,0.456349045,-0.3122194111,-0.3427392244,-0.0634144843,0.0046370407,0.175701499,-0.0588342734,-0.0855528265,-0.1320468485,-0.1167385429,0.2944754064,-0.1268905252,0.2139746696,0.2071737051,-0.5042848587,0.0481904261,-0.2399586588,0.1187071651,-0.1444853097,0.8278719783,0.2016857862,0.2281517088,0.1707098186,-0.1674889773,0.0738516971,-0.0322455242,0.1515463889,0.1431533694,0.1254112571,0.0683337077,0.313442111,-0.1844415963,-0.1439933777,0.0325272419,0.0937814042,0.2697514594,0.1583288461,0.1357518584,0.0169057269,-0.360702157,0.0822106674,0.3359976411,-0.1546699852,0.3511427343,0.0612426102,-0.3667969406,0.1276914477,0.1346846223,-0.0992993563,0.2475949228,0.1242643297,0.6549097896,-0.0827990025,0.2092818171,0.0802429914,-0.2006506771,-0.2900007963,-0.1183514968,0.2302942872,-0.321746856,0.2533904612,-0.29092592,-0.0256971866,-0.1533488035,-0.5618238449,-0.1246444359,-0.1464746296,-0.133392334,0.2190171927,0.1132509932,-0.0107296929,0.1471323818,-0.0475394316,0.0562688299,-0.3297478259,-0.0881586373,-0.2183518708,-0.2283635437,0.0845119283,0.1724345684,-0.0343185924,0.0739752725,-0.4179580212,0.1444265395,0.1028703675,-0.3723767102,0.1391796917,-0.2421471477,0.2731698453,0.1248976663,0.1650901437,-0.0815000534,-0.2417911887,0.2466869056,-0.2149168998,-0.2069196403,0.0929302052,-0.1716102958,-0.119896777,-0.0359272473,0.0234336797,-0.4479819536,-0.150362879,-0.0476774909,0.1717286557,0.0451156273,0.4087423682,0.2735316157,0.2703090012,0.1169290766,0.0107265403,0.1480535716,0.0916589946,0.3843226731,-0.1301646829,-0.1981365383,0.2383732051,-0.2260391563,-0.1011064574,-0.1527102143,-0.2538123429,-0.4915043116,-0.0120012173,0.3054753244,-0.0686874613,0.0623601042,0.1255710423,0.2043825686,-0.0548562445,0.1552562863,-0.0187602546,-0.4737403095,0.3680259585,0.2172100544,-0.0406897739,-0.0262292139,-0.1755801886,0.4815886617,-0.3753927648,0.020472791,0.2633134723,-0.2355789989,-0.0068035712,-0.0528293476,-0.4611094594,0.0036397723,-0.1847560257,-0.3041336238,0.1817416251,-0.2408041656,-0.1257539392,-0.2307342142,-0.0366391391,-0.17575939,-0.2328239083,0.2032376975,0.1861503571,0.0798726752,-0.1830538958,-0.1921985745,-0.1006597355,-0.0913051143,0.1361495256,0.2216224223,-0.2793317735,0.0155516956,0.3415165544,-0.4479674995,-0.1989312768,0.341755271,0.2854674757,-0.0886623412,0.1616114229,0.0808900893,-0.0696092322,-0.1376150846,0.3729710877,0.2816265225,-0.0757552832,0.2678307295,0.026554361,-0.0564356148,-0.3571333587,-0.217158556,-0.3384555876,0.0057417117,-0.4966936111,-0.045504462,0.2713519335,0.3164219558,0.3757125735,-0.1122342274,0.2704490125,-0.3465849459,-0.5519669056,-0.3517625332,-0.1407109201,0.1530865878,0.5028290153,-0.0168275442,0.1946056932,-0.0392419845,-0.2785259485,0.0967339054,0.2477916181,-0.0157323163,-0.0035838066,-0.1822852492,-0.2473189682,-0.0830554068,0.5164919496,0.3500697315,0.3221660554,-0.4429982603,-0.3156476319,0.1012359709,0.3974390328,0.0882873386,-0.2729812562,0.1453929991,-0.5090759993,-0.1499265283,0.3440805078,0.0061723427,0.1730276942,0.2660875022,-0.12506634,0.083803378,0.5289405584,0.0912651494,0.0901364461,-0.1546895653,0.4772411287,-0.1450432688,0.1398287714,0.1510213614,1.0787631273,0.1814513654,0.0077457032,0.4770992696,-0.1532650143,0.2228242159,-0.019487882,0.1378747523,-0.2480952591,-0.1071970612,-0.0761737078,-0.2383585274,0.3729977906,-0.4476438761,-0.234717682,0.0159725435,-0.1639995873,-0.0122661302,0.2495940626,-0.3516530693,0.1477900594,-0.0657474622,-0.2961906195,0.1360822171,-0.250891,0.4241493344,-0.3590578437,0.0559557192,-0.1995578557,-0.0336894095,-0.317160219,0.1243306249,-0.3967086375,0.1247496977,0.1494258642,-0.2415790111,0.1310437769,0.2121975273,0.064492777,-0.2136448324,-0.0762004405,0.0279127453,0.23157309,-0.2378038466,-0.1434338838,-0.0753222778,0.3855234683,-0.3502840996,-0.2142904848,0.3173222542,-0.0825129077,-0.0908808261,0.1725524664,0.2091875821,0.2769806087,-0.0550583862,-0.1885495037,-0.0894352645,0.3013722301,-0.1852095127,0.1110983491,0.0666016117,-0.2075998187,0.1949197501,0.1747109741,-0.0737398714,-0.1099790633,0.0175029542,0.1461649835,0.1302606314,0.3898948431,0.0290484298,-0.1551445723,-0.1268409789,0.145129323,-0.3027576804,-0.2951273322,0.1759530902,-0.0258383472,-0.0668789372,0.0280269701,0.6873705983,0.3259896636,0.1552599967,0.0550707392,-0.4291833341,-0.0774787962,0.1243566051,0.152766481,-0.1249969304,0.1157025918,0.5038741827,0.1741125435,-0.1016042754,-0.2748785913,0.1354853511,-0.0593032427,0.1030383334,-0.1105343923,-0.0648717433,0.0845016316,-0.0123631386,0.1646894217,0.2588499188,0.0595654137,-0.2579057515,-0.1283561438,0.1562783718,0.0239581857,0.1828053296,0.3103038967,0.0552141517,0.0644356608,0.2268367857,-0.0016735588,-0.0094346432,-0.1668878645,0.0558828525,0.532106638,-0.0258575846,-0.0889189169,0.1958430707,-0.0979389176,0.5363427401,0.1408674419,0.1017613783,0.0171538219,0.1174359843,0.0843005031,-0.0718393698,-0.3412329555,-0.0513494685,0.2846089005,-0.4904822409,0.0625645593,-0.3364526629,-0.1324691176,0.451361239,-0.286696136,0.2231879532,0.3102892637,0.1463077068,-0.2944112122,0.0809044093,0.1782269776,-0.1134465486,-0.030985035,0.3269980252,0.0480658934,0.1815331876,0.1428014189,0.2542449832,0.0582327992,0.3335330188,0.0865662768,0.1174786687,-0.3116099834,0.3210150003,0.3409734368,0.0276753921,0.0891450867,0.076928243,-0.3156929612,0.3016649187,-0.0614565648,0.0225104336,0.416898936,-0.2689742148,0.2537066936,0.0008674453,0.1846020222,-0.1069063693,0.3282034397,0.4171300828,-0.0523750335,-0.2081702054,0.1060012355,0.5490252972,-0.1221178696,0.2584014535,0.2681068778,-0.0565662682,-0.0693685859,0.0551592596,0.1086182594,-0.2574960589,0.0817353055,0.2580649555,-0.2532090247,-0.0839340836,0.1810257733,0.0074315607,0.2367499769,-0.1356083006,-0.0013656012,-0.059763521,-0.1610699594,-0.2568966746,0.3806136549,0.5217980146,0.3818217218,-0.2359579355,-0.3423900902,-0.1962793767,-0.0874526948,0.0738692656,-0.1809852719,0.2881011367,0.5520424843,0.2105224729,0.215358451,-0.234961465,0.2406452447,-0.084272854,-0.1514885575,-0.060168054,0.223252669,-0.0355659351,-0.2803725302,-0.5386535525,0.0837789848,-0.1358059645,0.2805465162,0.0189821571,-0.0822638199,-0.1417829245,0.119693175,0.0653760582,0.1721200943,0.3914944232,0.4023909271,0.5184406042,-0.1240526512,-0.0729528815,-0.3672166765,0.40309605,-0.1775859892,0.0986978337,0.0125101237,-0.0803737044,-0.0398988985,0.4504226446,-0.0612568706,0.2097421139,0.0034019505,-0.0715811253,-0.3556345403,-0.156913653,0.0931287482,-0.2545481324,0.0195333157,-0.1280136257,0.2437775135,-0.5917360187,0.1725881845,-0.2705285549,-0.1944616884,-0.0958156362,-0.0983714536,0.706305325,-0.0965525135,0.5003581047,0.0930401385,-0.0305546708,0.1179192141,-0.2922755182,-0.0464042574,0.2140001506,0.1180317104,0.155904144,0.1021036804,-0.1047888175,-0.3668814301,0.4173870385,0.1349469423,-0.0432441607,-0.1881148964,0.092941381,0.0799035206,0.0618053228,-0.1675795168,0.0241242982,-0.1527570337,-0.0659463704,-0.358762145,-0.5053910613,0.5118796229,-0.211914435,-0.4171209335,0.1004086435,0.1857523471,-0.1575042605,-0.1519348472,-0.4806273282,0.095879592,0.2085960209,0.0151501102,-0.2411969304,0.2493701428,0.0141245928,-0.0595726818,-0.0591214895,0.3708215952,-0.2053608596,0.1289943606,-0.2856554091,-0.1129257381]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1781","title":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import ","comments":"I updated the pyarrow version check in #1782","body":"I'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n","comment_length":8,"text":"AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import \nI'm using Colab. And suddenly this morning, there is this error. Have a look below!\r\n\r\n![screenshot-colab research google com-2021 01 26-08-15-36](https:\/\/user-images.githubusercontent.com\/45964869\/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)\r\n\nI updated the pyarrow version check in #1782","embeddings":[-0.1124686375,-0.0378036723,-0.0735252127,0.164127171,0.1544367224,-0.0958330259,0.2061190307,0.282099098,-0.3557239175,0.1526601166,-0.0846707374,0.3128411174,-0.1389750689,-0.0631880686,0.1409622729,-0.0942642838,0.3295861185,0.3606531918,0.2008519471,0.0204993598,-0.1710345149,0.2157241851,-0.3541971445,0.1867910028,-0.3741960526,0.2079935968,-0.0320674069,-0.2501802444,-0.2437812239,-0.5753995776,0.4486568272,-0.3361332417,0.1935667843,0.0710114241,-0.0001155161,-0.0160723832,0.3138076365,0.0019260962,-0.0547400452,-0.0776998848,0.0906295478,-0.0390862003,0.2458619475,0.0013490829,0.1946565807,-0.1191044301,-0.1037618294,0.354454577,-0.1228308454,0.2993268669,0.2384106368,-0.0697568879,0.5609894991,0.0383027196,0.5071519017,-0.130071789,-0.1689492315,0.6112796068,0.3375304639,-0.3840689361,0.2591115832,-0.0245374143,-0.2751604319,0.1576609015,0.3581281304,-0.0082474761,0.4751442075,-0.2434910089,-0.0197683256,0.0426393561,0.1899731457,-0.0679698661,-0.0622953214,0.0837467685,0.2524551749,-0.427141577,0.1876240969,0.2963083386,-0.083512716,0.1303890347,0.346353054,-0.1518793404,-0.169567883,0.0842554867,-0.3452576697,0.4248476326,-0.0305372365,0.0136884665,0.2424239218,-0.0076305154,0.5105876327,0.2294441164,-0.1592216939,0.2011142522,-0.1121024415,0.136520952,0.0597119629,0.351762861,-0.0152089261,-0.3056841493,-0.1907361895,-0.2672472894,0.2012715042,0.3683919609,-0.2163178474,0.3055542409,0.03416989,-0.1275661737,0.1872810572,0.0205095932,-0.1288060993,0.1963112652,-0.1803538501,-0.4741946459,0.2242390215,-0.1051282436,0.4093228281,-0.1850850731,-0.3368154168,-0.2970516682,-0.5312214494,-0.0009133238,-0.1789848953,0.1765611917,-0.196219638,0.1751807332,0.0098732011,0.2177179456,-0.1557417661,-0.1737435311,-0.015508418,0.3761232495,-0.168391645,-0.1346339136,-0.0991862193,-0.1879915446,-0.0344835781,-0.0116438419,0.0181776155,-0.1456897557,-0.4630939364,-0.2223874032,-0.2246121019,0.3160358965,-0.1776612848,-0.0363333859,0.0879849344,-0.5194098353,-0.1641188711,0.3128075898,-0.5522886515,-0.4386613369,-0.30300951,0.1607139111,-0.2019747347,-0.2211219519,-0.3484857678,-0.173454091,0.0369870588,-0.284514606,-0.1278896034,-0.5290755033,0.0261631235,-0.3962886333,-0.0754592195,0.0291221589,-0.4789656699,-0.1463800818,0.0013629723,-0.0530551635,0.2237424552,0.0469873697,-0.1539220065,0.0591673963,-0.0956592262,0.2041077316,-0.0042096465,-0.1021972895,-0.2656530142,-0.260989368,-0.021172205,-0.0882820338,0.0206084568,-0.3344439566,0.0525296256,-0.1012532711,0.4490200281,0.2826895416,-0.099680312,0.2058138549,-0.2889840901,-0.2170895785,0.4060925543,0.1129837334,0.1512932479,-0.2087875009,0.1085989773,-0.223853156,0.1350322515,-0.1140996516,-0.0860951021,-0.0214143246,0.4984558225,0.0726376399,-0.0838330239,-0.026180096,0.0225686245,0.1296487153,-0.1658645272,0.1710989624,-0.3791067302,-0.0251812972,-0.336607188,0.4128983617,-0.2210571021,-0.1960130483,0.0793145224,-0.0936159119,-0.1132225096,0.4208649397,-0.0313710682,0.0848802626,0.2234196514,0.1708586216,0.0787442103,0.4909259379,-0.3156828582,-0.4774178267,-0.0248798225,0.00549311,0.2481825501,-0.0405309014,-0.1291113794,-0.0200868249,-0.0562747605,0.2581667006,-0.2205220014,0.1656321138,0.1114037186,-0.4775620103,0.0358232483,-0.2195965797,0.0346976705,-0.0296700317,0.9173343182,0.1330341548,0.2958163619,0.1296980679,-0.1814094186,0.1339263022,-0.0125214458,0.1761591285,0.2564432919,0.1994247586,0.0100244051,0.2853644192,-0.1856587678,-0.1025533602,0.007062335,0.0744660869,0.325815171,0.1651006341,0.1231619194,0.0126922838,-0.4272812009,0.0212625656,0.3283031881,-0.2026514709,0.3039994538,0.0446234941,-0.3709373176,0.1085254401,0.0793547034,-0.0695158839,0.2275386751,0.1490172744,0.6420378089,-0.0414076336,0.2208136618,0.1209487766,-0.1660213023,-0.2487543672,-0.0732486099,0.3189907074,-0.2671295702,0.1916843653,-0.3418708742,-0.0256536789,-0.1284380406,-0.5866007805,-0.1755900085,-0.1328154951,-0.106823869,0.2877445817,0.2056490928,0.0908265114,0.2154552788,-0.1168931797,0.1013856083,-0.4101772606,-0.0520619936,-0.2066058367,-0.3605291843,0.0443742014,0.0935711712,-0.0692228526,-0.0919254571,-0.3627234101,0.1216171458,0.105400078,-0.4222922921,0.1178361177,-0.2779344022,0.2962707877,0.1820938885,0.1284395009,-0.1607793719,-0.2292151153,0.2884704173,-0.3069826365,-0.2447897792,0.1384818703,-0.2012093216,-0.1120298356,0.0271423627,0.0091784792,-0.4106236696,-0.1296188682,-0.0303596184,0.1086964011,0.0868661776,0.3450619876,0.3306105137,0.2713086903,0.0568980537,-0.009848346,0.155839771,0.0944446474,0.4344382584,-0.0707263276,-0.1634395123,0.2532993555,-0.1979409456,-0.1411481798,-0.1138304099,-0.2590644062,-0.4392936826,0.0223470777,0.4394187927,-0.0512276664,-0.0456402004,0.2601254284,0.1525090188,0.0137657681,0.1438574642,-0.0559290983,-0.453830421,0.3718741238,0.2538721561,-0.0128349429,-0.0538343415,-0.1033134013,0.5017088056,-0.3493603766,-0.0123524638,0.3065973818,-0.2327265739,-0.0338437632,0.0290860459,-0.3761235178,0.019552866,-0.2041319609,-0.2193300873,0.1572101563,-0.2137958258,-0.1272414923,-0.183895126,-0.0980646238,-0.2190408409,-0.2001003325,0.1550676823,0.2022349238,0.1531186104,-0.1872555017,-0.1602838039,-0.0624665506,-0.1208174601,0.1384869218,0.1665459424,-0.2808387876,0.0872497708,0.4063445628,-0.4979240894,-0.1757446676,0.3400971889,0.314589262,-0.0342004374,0.2395341992,0.0448649414,0.0162530784,-0.0483375154,0.3306387961,0.2264914364,-0.075172931,0.1958135813,0.1094259322,-0.1259836853,-0.4531097114,-0.1010990366,-0.3232712448,0.0232770815,-0.4254502356,-0.0852847695,0.1865972877,0.373585701,0.3594915271,-0.0650618598,0.2643875778,-0.3291902244,-0.4780862629,-0.185909912,-0.1692800373,0.0830284655,0.4888724089,-0.0162027813,0.2461379766,-0.0330923572,-0.2215816528,-0.0064832955,0.2747358084,0.0167299863,-0.0466583855,-0.1293275654,-0.249110952,-0.1481095701,0.5636306405,0.3008914292,0.3132924139,-0.479536742,-0.3219971061,0.1857724339,0.3259756565,-0.0885010213,-0.3116883039,0.1394634992,-0.5710722208,-0.1419472992,0.3130743206,-0.0488311537,0.2397875041,0.1352001578,-0.0947647318,0.1530192494,0.5696870685,0.0295478869,0.0315455049,-0.1354991943,0.4589650631,-0.131165117,0.1742155999,0.0413142815,0.8475158811,0.1511539519,-0.0923378617,0.5495868921,-0.1872691065,0.2076961994,0.0248969328,0.1365446597,-0.2657337785,-0.1115643829,-0.0859363973,-0.2354570031,0.336886555,-0.4944016933,-0.2005314827,-0.0017704756,-0.0786479637,0.0086832065,0.3116438985,-0.3489542603,0.1584766805,0.016670689,-0.2096280605,0.1313531697,-0.3303144574,0.4047878087,-0.2559388578,0.0984016061,-0.197747916,-0.0650303587,-0.3222208321,0.1911070943,-0.420547843,0.0861370265,0.3772366047,-0.2550759315,0.1468473375,0.2089601755,-0.0545334555,-0.2543567717,-0.112523824,0.1249736473,0.2168727964,-0.240950346,-0.0872569084,-0.0533671565,0.3795850575,-0.2667252719,-0.227329284,0.3284482658,-0.1222152263,-0.113557294,0.1618291438,0.1367326528,0.2624084651,-0.058163695,-0.0215844177,-0.1723210216,0.3871839643,-0.2176901996,0.0844264552,0.038870763,-0.1573741883,0.1678261161,0.0529235862,-0.0607902221,-0.0849956498,0.0624172911,0.1089151129,0.1041437313,0.4062592387,0.0937165543,-0.1455744654,-0.0764998719,0.0887247324,-0.2459032983,-0.3700652421,0.1563571543,-0.0559189804,0.0592486747,0.1214181557,0.6714067459,0.2794709206,0.1797622293,0.0764037818,-0.3774918616,-0.0556785241,0.1717635691,0.1413030773,-0.1993961483,0.1882619113,0.4810976982,0.2617723346,-0.1202934459,-0.2182924151,0.177065447,0.0298970304,0.1353892237,-0.1896177977,-0.0635510013,0.1047136262,-0.0503076464,0.1441191882,0.1863055527,0.0487374291,-0.209386155,-0.2565219104,0.158509627,0.0577464551,0.1249838099,0.2117958665,0.0154075343,0.1196243763,0.1461400837,-0.0200746506,-0.0093905348,-0.2532601058,0.0638394132,0.5592533946,-0.0509715229,-0.0587743744,0.1307712644,-0.1519883871,0.4583652914,0.2210689485,0.0836993381,0.0426480658,0.1050319746,0.0113619817,-0.0747863427,-0.4212161303,-0.0642229691,0.2996359169,-0.52564466,0.00958117,-0.4049111605,-0.042428188,0.3681215048,-0.2417972237,0.1392770261,0.3892607987,0.1161824018,-0.268333137,0.1061104238,0.1938464046,0.0283378661,-0.0950894654,0.1908836067,0.0829091147,0.1054414734,0.0950886905,0.2395569235,0.1353023499,0.3289608657,0.1008621752,0.2628572285,-0.2566657364,0.315885514,0.3654891849,-0.0449549146,0.0665886402,0.1109428331,-0.2747679055,0.2293279767,-0.0926469415,-0.0462779477,0.3561130762,-0.2217226923,0.1003416106,-0.0024455988,0.2226579487,-0.2116025388,0.2902748287,0.4970498979,0.0101334164,-0.1630514115,0.0231590178,0.4626983404,-0.0969028994,0.2003806084,0.2060602456,0.0182420146,-0.0898845047,0.0480001196,0.0700539425,-0.2608355582,0.2380621284,0.3056742847,-0.3201716244,-0.2440646738,0.1266938746,0.0636194274,0.2373266518,-0.1233720183,0.0020383811,-0.0492514595,-0.2567702234,-0.2933326066,0.3305199444,0.5262302756,0.3295855224,-0.190962553,-0.3734183013,-0.2212866098,-0.0560507216,0.144514963,-0.1780757308,0.2303151488,0.5802273154,0.0901426151,0.1861439943,-0.1545344442,0.2537144125,-0.0646504983,-0.0655586198,-0.0693410635,0.1709649563,-0.0193924196,-0.283208251,-0.4628162682,0.1249742433,-0.0338218696,0.3867750168,0.0504111424,0.0623179451,-0.1748026758,0.0344868973,0.0518705212,0.1936522126,0.342024833,0.4349009991,0.4891680479,-0.1340840608,-0.1615569741,-0.3317535222,0.4099685848,-0.2027936429,0.1009103209,0.0943947583,0.0079338113,-0.1505703926,0.4006496966,0.041659832,0.2520709038,0.012970455,-0.0760929883,-0.3443528116,-0.1569252908,0.1188235953,-0.2753844857,0.0392375477,-0.1003034189,0.2659503818,-0.6327345967,0.1201640442,-0.3119339943,-0.2667694688,0.0166767705,-0.0190861672,0.6783794165,-0.1068060249,0.4570755363,0.1489378661,0.0270859282,-0.0808219239,-0.3019062579,-0.0793080851,0.26098153,-0.0071211718,0.1951284409,0.1068722531,-0.0455032215,-0.3352682889,0.3715782166,0.0432847962,-0.0578927435,-0.1473907679,0.1107622087,0.1017372757,0.1083230898,-0.1432146877,-0.0583370589,-0.1319215596,-0.0084357643,-0.3947144151,-0.4863258302,0.3831761181,-0.2993132174,-0.3470004201,0.0528298356,0.1867039353,-0.1514095664,-0.1566532403,-0.4346128106,0.0862068683,0.2748096883,0.1471645683,-0.27368626,0.2126863301,0.0297748279,0.0297784172,-0.0231617521,0.4084547162,-0.1660972089,0.103372559,-0.2379494011,-0.0992112979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1776","title":"[Question & Bug Report] Can we preprocess a dataset on the fly?","comments":"We are very actively working on this. How does your dataset look like in practice (number\/size\/type of files)?","body":"I know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532","comment_length":18,"text":"[Question & Bug Report] Can we preprocess a dataset on the fly?\nI know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532\nWe are very actively working on this. How does your dataset look like in practice (number\/size\/type of files)?","embeddings":[-0.3903958201,-0.1971480101,-0.07739079,-0.0038245281,0.3977741301,0.3356423378,0.2610038817,0.2968967557,-0.084641166,0.1327671707,0.0693483576,0.1286533326,-0.1356009245,0.2288652807,0.2126400322,0.1509196609,0.2128424346,0.0664261952,0.0356727354,0.2065104991,-0.2539820969,-0.1531421393,-0.1100583076,-0.212891683,-0.4570014477,-0.2720251679,-0.1605053544,0.0205966514,-0.2341964692,-0.1767743677,0.3146640956,0.3214798272,0.0866012499,0.5502457023,-0.0001117338,0.008893053,0.4019140005,-0.0934246853,-0.1076299772,-0.1060749441,0.1285001636,-0.2643048763,-0.1480513215,-0.2470785379,-0.1567154676,-0.0504478477,0.1196824685,-0.6373549104,0.3856018484,0.1254341304,0.2115146071,0.128656894,-0.0810716078,0.1382953972,0.3179634213,0.3585823476,-0.3582457304,-0.0789366886,0.2596543431,-0.0795592964,0.0480676144,0.1202432886,-0.3685803413,-0.0111847473,0.3391450942,0.1131046638,-0.2082394958,-0.6446156502,0.1036311015,0.276486367,0.49214077,-0.5251832008,-0.6757846475,-0.2348736078,0.1164427474,-0.3433661759,0.0445728451,0.4104189277,-0.1506752223,0.0462534279,-0.4445935488,-0.3685857356,0.007384201,-0.3361414373,0.2708496153,-0.1538696289,-0.2351511568,0.0023274606,0.0621331371,0.1466435045,0.3794543743,-0.2408935875,-0.3575361371,0.5096227527,-0.2625169456,-0.2431843877,-0.2934468687,0.1031612903,0.2890485227,0.1290298402,0.0322178863,-0.1347647905,-0.0453485623,-0.0331239849,0.1808281094,0.0479214564,-0.0593688935,-0.1961773932,0.3188611269,-0.2196499556,0.0754331648,-0.0793797448,0.0598629154,-0.3965128958,0.3271101117,0.0377091952,0.3040319681,-0.32267344,0.4087907672,-0.0180785656,0.0922285989,-0.0135145253,-0.4056245685,0.1021770164,-0.1372481585,-0.0178536512,0.1294896901,0.0807888061,0.0287715979,-0.173416838,-0.1040252373,0.0153777571,0.0140405511,0.2762532234,0.1291560978,-0.225557074,0.0598062761,0.1925711781,-0.0866798386,-0.1647258699,0.1988245845,-0.4564754963,0.1030178517,0.3438337743,0.0327928253,0.0540712178,-0.06752152,0.3300350904,-0.1948367059,0.4849701524,-0.2016842663,-0.2776722312,0.0362526886,0.1042887717,-0.0199730638,0.0556141399,-0.6019009948,0.1712470353,0.5740996003,0.2262290269,0.2237465233,0.2339284569,-0.2646639049,-0.1679983735,0.1193476245,0.4449639022,-0.189895004,0.2122202665,0.2736875415,0.0231916141,0.1523585618,0.1042374074,0.0275389515,0.2540634573,-0.063007541,-0.1503755003,0.3098082244,-0.1066606641,0.1157015264,0.4008814991,-0.0550089292,-0.2682392001,0.0742101893,0.1855351776,-0.0364895724,-0.1308399588,0.0000576921,0.3976077735,0.0050129583,0.1950966269,-0.1847909838,-0.4482399523,-0.3217773736,0.0058107013,-0.0378195792,0.1062782556,-0.1856068522,-0.3014105856,0.0877304524,-0.1978133023,0.1949757338,0.3726005554,-0.0443876237,0.0690668374,0.0505192615,-0.1592517346,-0.366135776,0.2145513147,-0.1230579466,0.0163033884,0.18168284,-0.4576365054,0.1819765866,0.0651791766,0.0800102875,0.1132603362,0.0453959852,-0.3407961726,0.3032753468,-0.0114895562,-0.4556630552,0.2294508666,-0.191767633,0.1052213982,-0.3946021497,-0.0814101025,0.1480355561,-0.06047507,-0.0355989411,-0.1678518206,-0.2173678726,-0.00864325,-0.0828095376,0.3009221852,-0.0961251631,0.1768975705,-0.0134273153,0.168675676,-0.1474278718,-0.1476063132,-0.0205906946,-0.1321052313,0.1468913108,-0.1920770556,-0.2768490016,0.2378067821,-0.1526382416,0.4071863294,-0.0565594919,-0.1704910994,0.2015391737,-0.0626928955,-0.046720691,0.1874064505,0.011347251,-0.0184700694,0.7148051858,-0.1235886514,-0.0870060027,-0.1725437194,-0.0114093842,-0.0265265126,-0.0448284857,0.2636304498,-0.0986315608,-0.1139736176,-0.228943184,0.0929387957,0.2143616378,0.1423552632,0.0484319404,0.1799986213,0.0954068303,-0.0961784422,0.2947970629,0.010229283,0.4094163775,-0.0306621846,0.217375055,0.1128181815,-0.1889411211,0.171087265,0.0641639605,0.0664077103,-0.1352333426,0.1167892814,0.1735371798,-0.1462999731,0.1443755031,0.0346317515,-0.2236531228,0.1055327132,-0.1247566193,0.0065264278,0.0684499368,-0.0014727175,-0.4323998988,0.4077481329,0.0714962184,-0.1607204676,-0.1685139388,-0.1798398942,0.2189041227,0.0133905103,0.2648711503,-0.1633476466,0.4653586149,-0.051857233,0.0478228517,-0.3106508553,0.188993901,0.0980869606,0.0368409231,0.3631496131,-0.0228803158,0.0573131777,0.2028739601,0.1895760149,-0.0003710148,-0.1338364035,-0.1611996889,-0.2465218157,-0.0330792442,0.0012594801,-0.2042115182,0.0467956923,-0.1116845608,-0.2437698245,0.1459947973,-0.000191208,0.0341539495,0.24361597,0.0226549488,-0.1487399489,-0.1455842257,0.079873994,-0.0226015337,-0.7236916423,0.2888316512,-0.0164141487,-0.3574145734,0.0216522235,0.0053142523,-0.0362320058,0.3411466777,-0.605407238,0.0128710484,-0.3743886352,0.0355547108,-0.1713722199,0.1643203348,0.2667718232,0.0743670687,-0.071507223,0.0220159795,-0.1048546284,0.0018189943,-0.1901122034,0.0152298519,0.2821198702,0.4061176777,0.0109551968,0.6870184541,0.1741272062,0.0132037643,0.3461167216,0.0442665368,0.5539386272,-0.277326405,-0.0958695188,-0.1771088541,-0.2428266555,-0.1766515821,0.1236018613,0.2735069394,-0.3021138906,-0.1938185394,0.0253618974,-0.0795211717,-0.0992978662,0.1347560883,-0.2884258926,0.2887288034,-0.0055785999,0.0322169103,-0.2510035932,-0.4088840783,0.1039896905,0.0649311841,0.698807776,-0.1419106424,-0.1913323253,0.2181936651,-0.6576189399,0.206913963,-0.0033274896,0.2896779478,0.1281390935,0.0671626925,0.0759672448,-0.30187428,0.5862106085,-0.2058389634,0.0302286595,0.4009722471,-0.0075551919,-0.238610357,0.1260057986,0.0773038641,0.5740273595,-0.1925812662,0.3010319471,-0.0307211485,-0.1716298908,0.0872897133,0.3777670562,-0.2151721716,-0.5332978964,-0.1703979075,-0.179748252,-0.1750022024,0.150467366,0.054904215,-0.1973290741,-0.1321411282,-0.1062016785,-0.3844801486,0.3052115142,-0.0211571623,-0.1930192113,0.1160207838,-0.0337628983,0.4853582084,0.0394346826,-0.1699628234,-0.0728725046,0.5232403278,-0.2595112324,-0.1893886328,-0.1474987119,-0.0363619067,0.3864379227,0.2437516898,0.0483536087,0.2093797624,-0.0149337165,0.0861013159,-0.5349301696,0.1091596186,0.1268619597,-0.0836400464,-0.6452922821,-0.6370645761,0.610353291,0.4527736306,0.0547731034,0.4409403205,0.0547217354,-0.1575469822,0.2837332487,0.2144525796,0.8360589743,-0.3474932313,0.2523153722,-0.3638066053,0.1622582674,0.2669763565,-0.3113697469,0.0150995106,-0.3076028228,-0.4002587795,0.04246841,-0.1455601305,0.1053687558,0.3223255873,0.0580709726,0.298222512,-0.0511149392,-0.0921743512,-0.118870765,-0.074338451,-0.0784387365,-0.5844621658,-0.1622669846,0.1747305989,-0.0435513481,0.1200110391,-0.1409794092,0.2036852986,0.256159246,-0.1187517792,-0.3911299109,-0.016768638,-0.2120415866,-0.280205965,-0.1674090624,-0.1748959869,-0.1033788398,0.0724101737,0.4791938961,0.5132244229,-0.0546384938,-0.0849703029,0.1039533615,0.115241155,-0.098663196,0.1121684238,0.1111837327,-0.0754236206,0.0722719207,-0.0286082793,-0.3929864168,-0.1395228803,0.1768446714,-0.0854211375,0.0860090852,0.0270311441,-0.2994519472,0.0694888085,0.2213982642,-0.0579576939,0.1077322066,0.3272394538,0.0183360036,0.3171113729,-0.1506401151,-0.3271249533,0.0125231463,0.2531433702,0.1449327022,-0.3179357648,0.2900613546,0.199264884,-0.2374034226,-0.1379886121,0.2061214596,-0.0179354046,-0.3303303421,0.2150023878,0.1320089698,0.2330530882,-0.0694905221,0.1627016813,0.0822028071,-0.2622517943,0.138838172,-0.1855163127,-0.5902348161,-0.0965857282,0.1428097039,0.2233102471,0.5382516384,-0.373632133,0.1490962654,0.3163086772,-0.2533814907,-0.0101814922,0.0427116714,0.1136294827,0.1111271158,0.2753556371,0.0298255738,-0.178919822,-0.005251151,0.0539140739,0.0332929604,-0.2794554532,-0.0477462113,0.1669956595,-0.142047435,-0.0874356255,0.2233615518,-0.212816447,-0.2577115595,-0.3818079531,0.0933895782,0.341283232,-0.0700268373,-0.2541680932,0.3046003878,0.145024702,-0.3586770594,0.3340472281,0.0700193867,0.1435259134,-0.1232577115,0.1237586141,0.166476965,-0.0555976853,0.1118846014,-0.1866806298,0.2464661598,-0.0238259509,0.1362109929,-0.1701822281,0.0610307343,0.2766332924,0.2868178487,0.1773668379,-0.0933022499,0.1268343031,0.2990505695,0.1772527099,-0.2163124382,-0.2040119767,0.1293659657,-0.0591139831,0.1035018787,0.2595470846,0.2191545218,0.1358405501,-0.1556211114,0.2655941844,0.2433726341,-0.4232631922,0.04237229,0.2177234292,0.1372024417,0.1452836841,0.2677956522,0.3036687076,0.1623480469,0.3597693443,0.2897263169,0.5725321174,0.3371704817,0.054758437,-0.2753030062,-0.6611911058,0.3065600991,0.0854211748,0.3103347421,0.3788242042,-0.0135851065,0.2490010709,0.0256536007,-0.1785312891,-0.2163430005,0.1090382338,-0.1140523329,-0.1785174161,0.3778686225,0.0201884862,0.3296587765,0.1227003708,-0.0472564287,-0.1081706882,0.2652043998,0.0449241772,0.1607597619,0.1678289324,0.1748287976,0.0563541241,0.3255894482,-0.1390040815,0.1670777053,0.0338007435,0.0520103276,-0.0451621301,0.5297795534,0.2772931159,0.3780847192,-0.0987553895,0.1468119025,-0.0556049496,0.0012173828,-0.0299473815,0.2641468048,-0.1150572672,-0.1527374834,0.1634347886,0.0694898367,-0.0818065926,0.243288815,-0.1549185067,0.3400703669,-0.1396979243,0.2530255914,0.0932917297,-0.3390542567,-0.2198701501,0.1748737693,-0.2293969989,0.1370789707,0.6883219481,-0.3523358703,-0.0466697663,-0.0549992882,0.0778949335,0.1893008947,0.4720003009,0.3703411222,0.0690937862,-0.1510737687,0.0763527825,-0.4328628778,-0.153794229,-0.1426944733,0.2503023148,-0.2072927505,-0.1127906814,0.1592163593,0.2857863009,-0.0219077729,-0.3598353863,0.0739048049,0.4348348379,-0.312769115,-0.1881666481,-0.18507348,0.0751492381,0.0557162277,-0.5030342937,0.1237587035,0.1257848293,-0.074268885,0.0443494879,0.0941213295,0.1537397951,-0.0111306533,0.4049521983,0.0043677324,0.1726037711,-0.1656093597,-0.0145584084,-0.3031440675,-0.0436842814,-0.2267341316,0.1296515316,0.0770133361,-0.0606910773,-0.5568870902,-0.2705358863,-0.0729453787,-0.2407102585,-0.2368827909,0.060051728,-0.3554624617,-0.2455931157,-0.1167488098,-0.0597560257,0.1881877929,0.2010663748,-0.0200573858,0.0895658657,-0.3076409996,-0.2897495925,0.5246585011,-0.1734453887,-0.2983180285,-0.2442420125,0.0616433807,0.0614122376,-0.1606086195,-0.3795977235,0.2405109555,0.2741991282,-0.1191730499,0.1551201791,0.3109645545,-0.3516446948,0.0646962672,-0.2036039233,0.6356161833,-0.1921976358,0.0002170089,-0.402925849,-0.3338172138]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1776","title":"[Question & Bug Report] Can we preprocess a dataset on the fly?","comments":"It's a text file with many lines (about 1B) of Chinese sentences. I use it to train language model using https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py","body":"I know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532","comment_length":21,"text":"[Question & Bug Report] Can we preprocess a dataset on the fly?\nI know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532\nIt's a text file with many lines (about 1B) of Chinese sentences. I use it to train language model using https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py","embeddings":[-0.3749497831,-0.3105449677,-0.025356181,0.0781464875,0.4190768003,0.3276045918,0.2523874938,0.2493899167,-0.1120243371,0.0602596849,-0.0421746373,0.0807497799,-0.1185980663,0.1677815318,0.3189421296,0.0842123479,0.2063732594,0.0334929265,-0.0025367262,0.1758535057,-0.2277178764,-0.165063709,-0.0601746254,-0.1133634001,-0.5667706132,-0.2768295705,-0.2453327775,-0.0490853488,-0.1583475769,-0.2284193635,0.275698185,0.234082222,0.0893836394,0.5178625584,-0.0001183977,0.0513982028,0.3754434586,-0.0806588978,-0.1265788823,-0.1006739661,0.1375207007,-0.2211702019,-0.1770978719,-0.1919109672,-0.1763421595,-0.0012004221,0.1539304107,-0.4922321737,0.5047305226,0.1946835071,0.1448628455,0.2098596096,-0.0517701283,0.2219605446,0.2185227722,0.3186280429,-0.2461316735,0.144497931,0.2155944854,-0.016662417,-0.0376082994,0.1634667963,-0.3166071773,-0.0151862083,0.2694856524,0.1486750096,-0.1537855864,-0.5454908013,0.12963368,0.3487482965,0.3943709731,-0.56211555,-0.674554944,-0.2740337849,0.1808799356,-0.3332337737,0.0932866931,0.4548129737,-0.1409767717,0.2020023465,-0.4392782152,-0.4125903845,0.0781616643,-0.3117173016,0.2973233759,-0.0474813022,-0.26030761,0.0327828899,0.0513391271,0.1321082711,0.3066859841,-0.2835433185,-0.3123467267,0.5733699799,-0.248072654,-0.2415750623,-0.3126248419,0.1712335199,0.2013997138,-0.0665585697,0.0682614893,-0.1289230883,0.0041906396,-0.0502716117,0.1260250211,0.039293427,-0.0833768249,-0.2117564976,0.2936162949,-0.1786146164,0.0200481806,-0.0229564458,0.0223846883,-0.325699985,0.2305509746,0.0885231346,0.2718850076,-0.1417565346,0.3139848411,-0.0136686498,-0.0189775229,-0.0311282482,-0.3354018331,0.1593687981,-0.1914641559,-0.0439447723,0.1445102543,0.1041965783,-0.030840667,-0.2627187371,-0.0954937488,0.0899274126,-0.0553904288,0.3115380704,0.1160631478,-0.1680899262,0.0952999592,0.18655321,-0.1242667958,-0.2649587095,0.0811405852,-0.4365721643,0.0148296906,0.2506143749,-0.0197152831,0.0720883459,-0.0367933847,0.3380571902,-0.1779045165,0.4492940307,-0.2043383867,-0.2573890686,0.1642541438,0.0534856468,-0.0083526773,0.0555140488,-0.4702357352,0.3273878396,0.6150686145,0.2876583338,0.2803550065,0.3247433007,-0.3547723889,-0.0543346629,0.111244522,0.5088639259,-0.2824386358,0.0445280075,0.3910319507,-0.0378996059,0.1227306202,0.2004511058,-0.0386202559,0.3778627515,-0.0348490328,-0.1377107948,0.254900068,-0.0279567167,-0.004930364,0.4297841489,-0.1329644769,-0.2408827692,0.0994630307,0.2273842692,-0.1922977716,-0.0534686111,0.0641369522,0.3943487704,0.0507263318,0.1347920001,-0.1889541596,-0.5016103983,-0.2405148745,-0.0254901741,-0.051484853,0.1133788601,-0.1985218972,-0.0784508139,0.1419783831,-0.1865409166,0.176210016,0.3741005957,-0.1776565462,0.104966253,-0.0246022623,-0.239148438,-0.3564998209,0.2378043532,-0.1851833761,0.0727882981,0.1977110654,-0.4412870407,0.2564163506,0.0928590894,0.0211978368,0.0967225656,-0.0041481513,-0.3817107081,0.327762574,0.0220820736,-0.4443692565,0.2501779795,-0.3590976596,0.0738288611,-0.4026620686,-0.1143499687,0.1062477306,0.0095006414,-0.1003314927,-0.094527714,-0.2604745924,0.0071637603,-0.1008772627,0.2901645899,-0.1616960764,0.227222383,-0.1203622669,0.0991880819,-0.0737567991,-0.2234307379,-0.0107732685,-0.1301418245,0.1044430733,-0.2295478284,-0.2490978539,0.2669839561,-0.0294239223,0.4033820629,0.0256114807,-0.2489114255,0.2217794508,-0.0411257967,-0.0857890472,0.0846273154,0.1030280665,-0.0495610014,0.7756674886,-0.0289043859,-0.1323899776,-0.2946957648,0.1223669425,-0.032211598,0.050943207,0.2729354799,-0.1622239351,-0.1328100115,-0.2435219139,0.0138938036,0.2328962684,0.0966448411,0.0882694572,0.3098804355,0.0511683114,-0.1263473928,0.2850105166,0.0187537316,0.4396485686,-0.0250515621,0.1615845263,0.1324831098,-0.2792648971,0.2479462773,0.0971257538,0.0697617605,-0.2154580653,0.1413437873,0.0990858078,-0.1793445945,0.0310041215,0.0360496268,-0.2645478845,0.1805561334,-0.2056177706,-0.1350847632,0.1120666489,0.0225045737,-0.3421093524,0.3792150319,0.103795521,-0.0551779829,-0.0645874441,-0.1764418185,0.1415843517,-0.0590537488,0.3099807799,-0.2132001966,0.3635677993,-0.0960223675,-0.0447247028,-0.2309784889,0.1858299822,0.1836382151,0.0575980917,0.3153477907,0.0150038954,0.0612768978,0.1458824575,0.1730890423,0.03150969,-0.1268611252,-0.2293548435,-0.2512013316,0.0078260573,-0.002918558,-0.1471859664,0.0929360017,-0.1853148192,-0.1976004988,0.2195747793,-0.0661494881,0.0594113842,0.3334246874,-0.0181256551,0.0040230104,-0.1771635711,0.0646656826,-0.0010531198,-0.7291204333,0.3107971847,0.0556921028,-0.3598349392,-0.0095338486,-0.0833324417,0.0246738512,0.3075171709,-0.6023690104,0.0590275042,-0.3641526997,0.022292247,-0.1750625819,0.1446501315,0.1694180518,-0.0068356548,0.0020276711,0.0335193649,-0.0329024866,0.0712472722,-0.1892556846,0.0448544733,0.2905829847,0.6020951271,0.001918658,0.677452445,0.2382281572,-0.010103276,0.3234529197,0.1135637835,0.5247787833,-0.2154264897,-0.0926540792,-0.167107597,-0.2162948102,-0.1027712598,0.258295536,0.2151998132,-0.2334208637,-0.1804961711,-0.0443670526,-0.0072844303,-0.0935489759,0.197103411,-0.3860430717,0.3524753749,0.0431217179,0.0404772647,-0.2245044708,-0.3934687674,0.1024748757,0.0265667196,0.6047847867,-0.0617151111,-0.2506677508,0.1356464028,-0.7349545956,0.1575907171,0.0608926304,0.3999400735,0.1161181405,0.1423440576,0.1530472487,-0.2671493292,0.6611443162,-0.1820887923,0.0236953739,0.2840817869,-0.0700310543,-0.2258528769,0.1690517217,0.0049059754,0.514867723,0.019225264,0.3401063085,-0.0085865268,-0.2024806887,0.09084896,0.3764360845,-0.2770526111,-0.5819290876,-0.2186643928,-0.1276070923,-0.2294227034,0.1933106482,0.1035650223,-0.1981812268,-0.0617471002,-0.0676765889,-0.3200576305,0.2900869548,0.0510933958,-0.1866369247,0.0040524015,0.0837670639,0.4016826451,0.0518748686,-0.0788614452,-0.1687557101,0.5229564309,-0.2597317398,-0.26516819,-0.1819013506,-0.0811790302,0.3944948912,0.3223751187,0.0941209048,0.2992992401,0.1786826849,0.113853097,-0.5120099187,0.1892691404,0.154981181,0.0719947889,-0.7229226232,-0.6222501993,0.6525927186,0.3519832492,0.1259389073,0.2537674606,0.1220532805,-0.2655967772,0.2247297615,0.2336197197,0.8952514529,-0.3194475472,0.2945694625,-0.2977174222,0.1884461194,0.4412761033,-0.3313085735,0.0870243385,-0.1864237338,-0.3148833513,-0.0036829179,-0.1161767617,0.1716416478,0.3708029985,0.0952131748,0.3008534312,-0.0179842636,-0.0516084097,-0.0166100729,0.029302327,-0.0820476189,-0.585824728,-0.2183347642,0.1013282835,-0.0855746418,0.1453358531,-0.1920997798,0.1840758473,0.2510245442,-0.0993478,-0.4174712896,-0.0366447754,-0.3467862606,-0.1779520214,-0.1979885995,-0.1622329503,0.0526023954,0.0865078941,0.5049939156,0.5158249736,-0.0750930235,-0.0995385572,0.019285921,0.0976047739,-0.1561747193,0.0864427686,0.1889238954,-0.0573588647,0.0544863082,-0.0292992778,-0.3847137094,-0.1260056496,0.1087398827,-0.1294967979,0.0852848887,0.1332561523,-0.410595119,0.0430438444,0.188116163,-0.0803622305,0.0460288785,0.2664834559,-0.0843106732,0.1766819358,-0.1003565714,-0.3924740851,-0.0170777012,0.3325716257,0.1509129703,-0.2541078627,0.4268797338,0.3013948202,-0.278175056,-0.073016651,0.1619438231,0.0966813862,-0.4329403043,0.1042791232,0.1879659742,0.245978713,-0.2142018378,0.0548299924,0.1220951751,-0.2156116515,0.1215832755,-0.2027389854,-0.6988819242,-0.1157463416,0.193743363,0.0963307768,0.5451959968,-0.095300138,0.1898050904,0.4475330412,-0.1852530986,-0.034150295,-0.0051073274,0.0938498527,0.1589479297,0.2476893812,0.0246211383,-0.135083437,-0.0457717553,0.0844506025,0.0626788735,-0.237390548,-0.0988688096,0.200025931,-0.057025861,-0.1207513511,0.2125354856,-0.1871806979,-0.1679422259,-0.4030842185,0.033921916,0.3511114717,-0.0069714403,-0.2097284943,0.2676523328,0.2040484101,-0.3131277859,0.2779680192,0.0748594254,0.1064923629,-0.068685405,0.0969453827,0.2907650769,0.0461495705,0.0658223853,-0.1529538035,0.2815493941,-0.127648145,0.0869903937,-0.2359276712,0.0880420133,0.2308562547,0.2229379863,0.188907966,-0.1196663529,0.0540845543,0.3452784419,0.0965868533,-0.3158839345,-0.1373806,0.2207077891,-0.1279360503,0.1054342762,0.2710894942,0.208310768,0.2108288109,-0.1556241065,0.2213840485,0.1554814279,-0.3679103851,-0.0255755149,0.3143434525,0.2194734812,0.2891559303,0.3471426666,0.3937501311,0.1932381094,0.3604700863,0.296665132,0.5566427708,0.1156123206,0.1460451037,-0.276104033,-0.6948469281,0.1686465144,0.1119164228,0.1437652856,0.3430640101,-0.1404846758,0.2499821782,0.0134365726,-0.2159703523,-0.1870099902,0.1724847704,-0.2066755593,-0.2646560967,0.3597432375,0.0299956799,0.3080589771,0.1755875647,-0.0062246555,-0.1758528203,0.1199119762,0.0362682752,0.105983153,0.0291435048,0.1209735796,0.0765616298,0.1871217638,-0.2097547501,0.157423839,0.0188931357,0.1319532543,-0.1634799689,0.428920418,0.1874071509,0.3955987096,-0.075728707,0.1927971095,-0.0569285005,0.0233517606,-0.0576291122,0.1993986815,0.002304693,-0.074015297,0.1786220968,0.0053612082,-0.0582531728,0.1853408366,-0.0793755054,0.3324712217,-0.22972247,0.2897548378,-0.038106259,-0.3007814586,-0.2321134061,0.2577091157,-0.3346066177,0.3339974284,0.7271051407,-0.4429167509,-0.0402122997,-0.0623483248,0.0351394117,0.1675348431,0.512786746,0.3862255216,0.1302076578,-0.2132239938,0.1055066958,-0.5249496698,-0.1979750842,-0.1035253182,0.3607322574,-0.2679536641,-0.1362632364,0.1628433019,0.2141897827,-0.1237194166,-0.2999058068,-0.0559153073,0.3310471177,-0.2579765916,-0.2018976808,-0.3481370211,0.1867280453,0.0171720367,-0.4353329837,0.1584533155,0.2004406154,-0.1738946438,-0.0889285803,0.1638792902,0.0304905381,-0.0180219114,0.3460170627,0.0734669268,0.2587005794,-0.1941834986,0.0096100606,-0.3058735728,0.0232038144,-0.2586739361,0.208764866,0.0514140837,0.0079051135,-0.5748571157,-0.3172647655,-0.1032143608,-0.2164853662,-0.1374967843,-0.0587229952,-0.3331210613,-0.2301083207,-0.1784718484,0.002949351,0.323813498,0.2766809464,-0.0085442914,0.1011760309,-0.2328116,-0.1817187518,0.4950383902,-0.3164907098,-0.3039825261,-0.2405492663,0.0564855225,0.123800084,-0.2566739917,-0.4457372725,0.1504116952,0.2864790261,-0.1394608766,0.058869902,0.3648118973,-0.3023817539,0.0923664942,-0.21126917,0.62631917,-0.1821257174,-0.0996520519,-0.3672665656,-0.4044740796]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1776","title":"[Question & Bug Report] Can we preprocess a dataset on the fly?","comments":"Indeed I will submit a PR in a fez days to enable processing on-the-fly :)\r\nThis can be useful in language modeling for tokenization, padding etc.\r\n","body":"I know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532","comment_length":26,"text":"[Question & Bug Report] Can we preprocess a dataset on the fly?\nI know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532\nIndeed I will submit a PR in a fez days to enable processing on-the-fly :)\r\nThis can be useful in language modeling for tokenization, padding etc.\r\n","embeddings":[-0.3806301355,-0.1969681829,-0.0746243596,-0.0068640667,0.4007776082,0.3399396241,0.2778782547,0.263427645,-0.0995364562,0.1461998075,0.1342773885,0.1849375814,-0.1421174705,0.2875471711,0.2689658701,0.1431531012,0.1967978925,0.0332212746,0.0725330934,0.23776564,-0.2820772529,-0.1595617831,-0.1298619509,-0.1796594113,-0.4086602628,-0.3066292107,-0.1547997147,-0.0939180106,-0.2736937702,-0.2001533508,0.2956175804,0.2963145673,0.0653250068,0.465472281,-0.0001154027,0.0105220759,0.4278814197,-0.0547123067,-0.0995460376,-0.1435340941,0.1376840323,-0.2483471483,-0.1386116147,-0.2409489602,-0.1121577248,-0.0338170566,0.0784299448,-0.6800120473,0.3494402766,0.1158155948,0.1876232773,0.1196270287,-0.0841414332,0.1558762938,0.2843365967,0.3790517151,-0.3493456244,-0.1307842582,0.1972600967,-0.0772146508,0.0815445036,0.0771728605,-0.359899044,-0.0484260358,0.2894827425,0.0892697796,-0.2145828605,-0.6604301333,0.1388300955,0.2746991515,0.4689072967,-0.5630953312,-0.7029758692,-0.2193475068,0.1402742267,-0.3208349347,0.0382909775,0.4482021928,-0.1715643257,0.0731241927,-0.4151315689,-0.4202463031,0.0413668342,-0.3086306751,0.2753442824,-0.0720978752,-0.2119609118,0.0252016485,0.034201704,0.1810992807,0.4665647149,-0.2718546689,-0.3090370893,0.4835742712,-0.2577773631,-0.1970091015,-0.2645561993,0.0677481517,0.2692093849,0.0593970232,0.0312879607,-0.1208760664,-0.0593945198,-0.0603432357,0.1857628673,0.0467237532,-0.0205967501,-0.188860476,0.2885476351,-0.2618692517,0.1094831303,-0.0513226241,0.0204735938,-0.3383103311,0.3059313595,0.0955198556,0.2437417209,-0.2711470127,0.3885354102,-0.0272215083,0.0765126199,-0.0235701893,-0.4008276463,0.0876809806,-0.1159519106,-0.0047159297,0.1546853483,0.096196264,0.0672433153,-0.2288593948,-0.0770790279,0.0140275694,0.0041487277,0.2325498909,0.072771199,-0.1973730028,0.0311659817,0.2105323821,-0.0917334408,-0.2058473527,0.1751339138,-0.4416748583,0.0907105356,0.2982267439,0.0351358987,0.0468139425,-0.0299171954,0.3525289893,-0.236672923,0.5580145121,-0.1571931541,-0.2958838046,0.046776969,0.0742287561,-0.0204167459,0.0069830599,-0.5607430935,0.234284237,0.6066606045,0.2160427123,0.2605373263,0.2320741117,-0.3053673208,-0.1547010988,0.1365791261,0.4368867874,-0.2761692107,0.1823856533,0.2781616449,0.0397016071,0.2247021049,0.028512964,-0.0242422968,0.2706718147,-0.0475754179,-0.1463966668,0.3211702108,-0.0999133587,0.1367200911,0.3996351957,-0.0590499789,-0.2813992202,0.081068255,0.1655155271,-0.0632078126,-0.0565833561,-0.0004899409,0.4461781085,0.0238893647,0.1589461416,-0.1804377586,-0.4752100408,-0.2243528366,-0.0135994283,0.0118641201,0.1263299286,-0.1286602467,-0.2577133477,0.0810363814,-0.2028314918,0.199029237,0.3393393457,-0.0400335081,0.0700502843,0.0501671359,-0.2194067091,-0.3255533576,0.2392324805,-0.1290400326,0.0725405589,0.1684771925,-0.4414997995,0.2057801932,0.1176253408,0.0929606333,0.1650901884,0.0184887648,-0.3471727371,0.2571828365,0.0080345338,-0.4888250828,0.2847564518,-0.2818238735,0.0605553575,-0.3192244768,-0.0891122371,0.1544453949,-0.026418956,-0.0729771405,-0.117057994,-0.2385203838,0.0320946388,-0.1021132097,0.2757893503,-0.0663902313,0.1682732105,-0.0866347328,0.1256416887,-0.1447451264,-0.1290698946,-0.0076086516,-0.1717793643,0.1499705762,-0.1908589005,-0.1879913211,0.2441369593,-0.0906544477,0.4551089704,-0.0186686572,-0.1764632612,0.1805130243,0.0065298025,-0.0552366637,0.1250545233,0.0155817894,-0.0882805586,0.7388350368,-0.1510331631,-0.0979687944,-0.1166276708,0.0265188515,0.0043680449,-0.0117035331,0.2582092881,-0.1133449152,-0.1405486763,-0.2205936909,0.0856269002,0.2107994109,0.1212838516,0.0395467691,0.2407395691,0.070957236,-0.0643182918,0.3207211494,0.0227348041,0.4563131034,-0.046853222,0.2302242517,0.157525897,-0.2121620625,0.147954464,0.0611933805,0.0646841079,-0.13868846,0.118262507,0.1323515773,-0.1580020636,0.1392466724,-0.028587671,-0.2757461071,0.0956975296,-0.1378237307,-0.1189022735,0.1435193419,-0.0278891902,-0.4474339783,0.468906939,0.05117134,-0.1783297509,-0.1720498204,-0.2506058514,0.2144922763,-0.0087660635,0.233284533,-0.1438086778,0.4303097427,-0.0180496797,0.0589523837,-0.2585681081,0.1374772489,0.1043521389,0.0309754312,0.3830716908,-0.0061089993,0.0010167625,0.208538115,0.2050906122,-0.001941492,-0.1447562575,-0.2002999038,-0.2609392405,-0.0236170497,-0.0058719674,-0.2088480294,0.0928851366,-0.1521229595,-0.2390484661,0.1361910552,-0.0025109316,0.0520357974,0.1913918257,-0.0341625996,-0.0971090719,-0.1601897925,0.0680320486,-0.0249221716,-0.6827620268,0.2437421829,0.017617479,-0.3386533856,-0.023872152,0.0088692727,0.0437363163,0.3255078793,-0.5940142274,0.0640683994,-0.4057153761,0.0709718466,-0.2209239602,0.1615331024,0.3194848299,0.1000271887,-0.0216584522,0.0339365676,-0.1603554636,-0.0201173481,-0.168243289,-0.0235502236,0.3433021307,0.4535853863,0.0314815566,0.6971470714,0.1354644448,-0.0072680274,0.3471180201,0.037520349,0.4989186227,-0.3026138246,-0.1154119,-0.1929935366,-0.2688210905,-0.1742166132,0.0781138763,0.2036479115,-0.2888621986,-0.2173533738,0.0515052043,-0.1161352992,-0.040354006,0.0860865042,-0.3470912278,0.3549282849,0.0275551509,0.0275918599,-0.2921555638,-0.3815273643,0.1133362651,0.0050596469,0.7152578831,-0.1412914395,-0.1982315779,0.1818397194,-0.6152505875,0.18129237,0.0240406524,0.3918983042,0.1309017241,0.0447119065,0.1315587163,-0.2936672866,0.5625488758,-0.2596140504,0.0218204409,0.3942946792,-0.0488944761,-0.2484005243,0.150346294,0.080365628,0.5243753791,-0.195470646,0.2850285769,-0.0203666985,-0.1341270059,0.04742378,0.3616240621,-0.2078993917,-0.5691261888,-0.2100686133,-0.1369137913,-0.1960627735,0.1672897786,0.0506859384,-0.1484781057,-0.0989842787,-0.0793981999,-0.4353748858,0.3044626415,-0.0492163561,-0.2059293091,0.0309125502,-0.0521687567,0.4565043151,0.0107045714,-0.2476540655,-0.1679905951,0.4879389107,-0.1882152259,-0.2456014007,-0.1580664814,-0.005101915,0.4487923384,0.2274577022,0.0468065403,0.2421487272,0.0288116448,0.0678722039,-0.5434034467,0.0580658875,0.1293226033,-0.0768534169,-0.6611810923,-0.6796532273,0.5924068093,0.4728298783,0.0690123886,0.4679059684,0.0373828076,-0.1862157732,0.3434485793,0.2481505424,0.8985002041,-0.3707823157,0.212494567,-0.3760999143,0.116207689,0.2998495102,-0.3049748838,0.0756964982,-0.309253782,-0.3410373926,0.0316975415,-0.1626029909,0.1744961143,0.32478863,0.0713598207,0.2498449087,-0.0499544144,-0.084162578,-0.0228537899,-0.1055415794,-0.0549489036,-0.5814414024,-0.1364497393,0.1379918307,-0.0684122592,0.1358137578,-0.1396816224,0.2095086724,0.2830487788,-0.1300168484,-0.4388834536,-0.0014405549,-0.2389284521,-0.2830254436,-0.1929852813,-0.1482586116,-0.0205200799,0.0289844777,0.5220213532,0.5669858456,0.0151478164,-0.1017068699,0.144987151,0.1029068753,-0.0549154952,0.1144180968,0.0950661227,-0.0876574293,0.0644190609,-0.0018640027,-0.3893852532,-0.1696494073,0.1456438005,-0.0497541279,0.1266869754,0.0535050705,-0.3307958841,0.1241155714,0.2743766308,-0.0897077993,0.0835441872,0.3032914698,-0.0005669278,0.2926302254,-0.1984924823,-0.3364444077,0.0016812606,0.2342017293,0.1770995259,-0.2752555013,0.2237219214,0.2338781357,-0.2575758994,-0.1063044816,0.1724000126,-0.0763132125,-0.3368683755,0.1599050462,0.1280328035,0.2558526993,-0.1217140853,0.1438892633,0.0312527306,-0.186990425,0.0978013575,-0.2196779549,-0.5614849329,-0.0840027481,0.0762471929,0.1519244313,0.5885524154,-0.3070678115,0.1264452785,0.3701990545,-0.2179435045,-0.0109183034,0.0840466097,0.0907296687,0.1064969152,0.2765712738,-0.0261680055,-0.1470153034,-0.023552537,0.0801510811,0.0814554766,-0.2511892617,-0.0210669953,0.1889359057,-0.1097426489,-0.0278197862,0.2378380299,-0.1992598027,-0.3205328584,-0.3503112495,0.0754176751,0.3556346297,-0.1250248551,-0.2205460966,0.3059661388,0.1924342066,-0.394723922,0.3645472229,0.1125156134,0.1303454638,-0.1061447188,0.175512135,0.209929958,-0.0145890955,0.1223226786,-0.1652099788,0.2344039828,-0.0214193035,0.1516185999,-0.2099452019,0.1074152738,0.1961651146,0.2712395191,0.1962304562,-0.0628978312,0.0937997997,0.3183845282,0.1273361146,-0.2558273971,-0.1813257784,0.1610768288,-0.0221007057,0.1044314131,0.2958099246,0.2726112008,0.1973697245,-0.1681461036,0.2517396212,0.2223005444,-0.4465553164,0.047888808,0.2298188508,0.1601936817,0.1691002697,0.2685855329,0.3256409764,0.0999970511,0.3341900408,0.3340494633,0.5820022821,0.398899734,0.0812578127,-0.2637985349,-0.6308768392,0.3162618279,0.062525399,0.2412298471,0.3569069207,-0.0555262752,0.2297913879,0.0353861563,-0.170692265,-0.1988729239,0.121228531,-0.137563929,-0.1665373296,0.3772591949,0.0165350977,0.3176463842,0.1049214676,-0.0418565907,-0.099271901,0.2685160637,0.0013307203,0.1823670566,0.1183916479,0.1652448475,0.0600968488,0.285771668,-0.183423996,0.176075995,-0.0436940975,0.0471842587,-0.1334072053,0.5387243032,0.2672435641,0.3809131086,-0.1241645217,0.1615903676,-0.1029652953,0.010155878,-0.012483729,0.2942136228,-0.0430893786,-0.0743365139,0.1786453575,0.0455708355,-0.0442043841,0.2325785905,-0.1303555965,0.385547936,-0.1641396731,0.3359889686,0.0904144496,-0.2982102633,-0.1875431836,0.2061236203,-0.2436428517,0.1571838856,0.7409971952,-0.359680444,-0.1059645861,-0.0848393887,0.0538699552,0.1989263743,0.4421608746,0.3787406385,0.0970761478,-0.1298151314,0.0690915138,-0.3757626116,-0.156197831,-0.1044114158,0.2584383488,-0.2027385682,-0.1442376524,0.1973621398,0.3004540503,-0.032004647,-0.2971875072,-0.0014055131,0.4319021702,-0.2911933959,-0.1324740499,-0.2300363332,0.0653185993,0.0283206031,-0.4767260849,0.1169424728,0.1917362064,-0.1106866375,0.0352426171,0.0704239681,0.1664677709,-0.060374178,0.4525508881,0.0339960158,0.1825510412,-0.1279176623,0.0058649853,-0.3119247556,-0.0759388804,-0.2287695259,0.2396232486,0.0511361435,-0.1284503341,-0.5494626164,-0.306665659,-0.1070034355,-0.2587788403,-0.2332361192,0.03912168,-0.3262362778,-0.2682570219,-0.1688866615,-0.0569063127,0.2241002172,0.1854241639,0.0212510806,0.0982697904,-0.2764485478,-0.2764080763,0.4450547397,-0.1166559458,-0.3241908252,-0.2345124185,0.057857085,0.0800354704,-0.2177993506,-0.385401547,0.1555528045,0.2995453179,-0.1309604496,0.2152191401,0.3185412586,-0.389082998,0.1024350002,-0.2206639946,0.6346171498,-0.2192077488,0.0336236805,-0.4068103731,-0.3738469779]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1776","title":"[Question & Bug Report] Can we preprocess a dataset on the fly?","comments":"any update on this issue? ...really look forward to use it ","body":"I know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532","comment_length":11,"text":"[Question & Bug Report] Can we preprocess a dataset on the fly?\nI know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532\nany update on this issue? ...really look forward to use it ","embeddings":[-0.3913038969,-0.2353689075,-0.0527780466,0.0122349374,0.4117043018,0.3109975457,0.279240489,0.2528895736,-0.1153569892,0.1204007193,0.1142090559,0.1698933691,-0.1199369878,0.2365342081,0.2234720588,0.1351319104,0.228430301,0.0495311692,0.0354344249,0.2275026292,-0.2491151839,-0.1349677294,-0.1753234863,-0.1825249046,-0.452917397,-0.2570207417,-0.1514614671,-0.0518691167,-0.2057731897,-0.2073408663,0.358797729,0.2701020837,0.0376464538,0.5408548117,-0.0001153744,0.007334983,0.4726910591,-0.056421116,-0.0887129158,-0.1314571947,0.0896476209,-0.2646889389,-0.0938433632,-0.2365348786,-0.137411505,-0.0467506163,0.0895897895,-0.6576628685,0.3294092715,0.1216332018,0.1973014623,0.1569767594,-0.0774318054,0.1003536433,0.2930387259,0.3533273935,-0.3769573271,-0.0924066827,0.264330864,-0.0517449267,0.0647785291,0.1207653657,-0.3051045239,0.0212722123,0.3152576685,0.1052179262,-0.2040335685,-0.6549833417,0.1530664712,0.2641944885,0.558565259,-0.5264739394,-0.6729978323,-0.1945065856,0.1294228584,-0.2966519594,0.073373653,0.408529371,-0.1787392497,0.037317805,-0.4725541174,-0.4117541313,-0.0172946546,-0.3276662529,0.2335697562,-0.114817515,-0.2629806101,0.0320257843,0.0165751465,0.1384833157,0.4328875542,-0.2503905892,-0.3554623723,0.4991389215,-0.2634126246,-0.2540122271,-0.2696558535,0.115881443,0.2948218584,0.1163263917,0.0184439383,-0.1187776029,-0.0177972652,-0.064432852,0.2256644517,0.046178475,-0.009879767,-0.1726369709,0.3565645218,-0.2238591164,0.1035009697,-0.1024409831,0.0683846027,-0.3795362711,0.357316196,0.0406235419,0.3261400759,-0.3509826064,0.415307492,0.0279082954,0.1009939983,-0.0196721163,-0.3956863582,0.1123018265,-0.1445615888,-0.0104062371,0.1232793182,0.0706248358,0.0549062639,-0.1477013379,-0.0923748463,-0.0523049124,-0.020611953,0.2570596039,0.1110924259,-0.2079761475,0.0653103366,0.1909388006,-0.0740979165,-0.1667143852,0.1608237326,-0.4625171125,0.1319248378,0.3501077592,0.0261418838,0.0926619247,-0.1034870073,0.3811652362,-0.192211628,0.5381245613,-0.1831321865,-0.2894229889,-0.0015653953,0.0817874521,-0.0463460647,0.0349835232,-0.5741597414,0.1643382758,0.5721703768,0.2163049877,0.2478599548,0.2325249761,-0.2978604138,-0.1778895259,0.1230793446,0.4145370424,-0.2231506407,0.1886384487,0.2470085472,0.0120800249,0.1614609361,0.0838068351,-0.0180595815,0.2311852574,-0.0474749096,-0.1800072789,0.3255036473,-0.0955241323,0.0736123919,0.4199148715,-0.0316171832,-0.2556696832,0.0243797489,0.1627577543,-0.0624335371,-0.09476237,-0.0121791521,0.394631803,-0.0306684859,0.2024058849,-0.1922035068,-0.452773124,-0.279821068,-0.0152163953,0.0337564051,0.0826677084,-0.1573309749,-0.2997128069,0.0944190621,-0.1527411044,0.2118885964,0.389387548,-0.0006934989,0.0627926961,0.0750386342,-0.1637516469,-0.3758691847,0.2164489925,-0.1051182002,-0.0012016,0.1774052978,-0.4462644458,0.1634746641,0.0894154906,0.072372146,0.0988966674,0.0078595849,-0.3602010012,0.3024236858,0.0196501985,-0.4540774226,0.2376606613,-0.1761718839,0.0812804699,-0.3632971346,-0.0849552751,0.1491796374,-0.0614651963,-0.0784816891,-0.15054138,-0.2213386297,0.0195640363,-0.1189282686,0.2937612236,-0.1254694909,0.1648638994,-0.0345056802,0.1126922816,-0.1117504761,-0.2055585831,0.0007554385,-0.1504150778,0.1543410271,-0.2200318128,-0.2653051913,0.2172542065,-0.1522412002,0.4260235131,-0.0510548763,-0.1547423601,0.186885491,-0.0446710438,-0.0564770252,0.1605170369,-0.0081768474,-0.0796433315,0.7069702744,-0.15518336,-0.0976715758,-0.1529772431,0.0276536904,0.0087977275,-0.0326276496,0.2751893699,-0.1313533932,-0.1178480089,-0.2040018588,0.0539022312,0.2588995397,0.1215402782,0.0613836795,0.2126242369,0.1112519056,-0.095969148,0.3399933577,0.0046546627,0.4254179895,-0.0333614089,0.2357476652,0.0941486061,-0.1927354932,0.2057366371,0.1071105674,0.0663398504,-0.1491219699,0.0915315002,0.1724846363,-0.1535189599,0.1638742536,0.0062128552,-0.23649472,0.0918823034,-0.1472282708,-0.0090793669,0.0899707675,0.0256715901,-0.44226107,0.4169695079,0.0887494609,-0.1201641187,-0.1878690571,-0.2017769814,0.2220894247,0.0049728304,0.2315044999,-0.2034287602,0.4496265352,-0.0345227234,0.069017902,-0.3456406891,0.1482603252,0.0874725059,0.0415443331,0.3881587684,0.0080670305,0.0107696755,0.1772739887,0.2333422303,0.0170115083,-0.1792710871,-0.1828169078,-0.2464339584,-0.0463258065,0.0018820865,-0.1816310138,0.0544060655,-0.12669833,-0.2562914193,0.1467564255,-0.0590143539,0.0438780524,0.2283809483,-0.0014999121,-0.138409391,-0.146624133,0.0436156653,-0.0500294566,-0.7102841139,0.2915790081,-0.0328898728,-0.3869550228,0.0240184516,0.0275736414,0.0163417775,0.321099788,-0.6315894723,0.0486708619,-0.3961657584,0.0781179592,-0.1475029588,0.1844329089,0.2822653353,0.0541504323,-0.0521070622,0.022240568,-0.167364791,-0.0096296864,-0.2110014409,0.0416741557,0.332813859,0.4177029729,0.0074993642,0.7122834325,0.1780762374,-0.0019232178,0.3445574343,0.028452076,0.5579954982,-0.2772135139,-0.1211384162,-0.1946613342,-0.2142383754,-0.1413664818,0.0702723339,0.2519640923,-0.3330572546,-0.2177139074,0.0203945264,-0.0499702841,-0.0831304416,0.0984789282,-0.3412390649,0.3190704584,0.0102853896,0.0272235274,-0.2585068941,-0.4105863571,0.1096911132,0.023843471,0.6739200354,-0.1839284897,-0.2142762542,0.1710340977,-0.6742998958,0.204590708,-0.013280374,0.3620601594,0.1162158698,0.0672743395,0.1101106331,-0.2730483711,0.5663537979,-0.2505612671,0.0123232016,0.414418757,-0.043241702,-0.2596127093,0.1303728968,0.0578912832,0.6084027886,-0.2039531767,0.3074113429,-0.002355983,-0.1870510578,0.1331274807,0.3073420525,-0.23216407,-0.5661637187,-0.1956929564,-0.1524101645,-0.1292851865,0.1758700013,0.0299043171,-0.184614867,-0.1327011138,-0.1156175658,-0.3998001814,0.3003884852,-0.0679833442,-0.1839969456,0.0989982784,-0.0616357885,0.5166421533,0.0397935212,-0.1879315227,-0.0712105781,0.5346564054,-0.2102050781,-0.1800593585,-0.0907104015,-0.0292669535,0.4326894283,0.2707434297,0.02303694,0.2422832102,-0.0352169946,0.0973325446,-0.5302481055,0.0582021065,0.1276598573,-0.0769894272,-0.6150191426,-0.5785787702,0.5974004269,0.4587287605,0.0350323096,0.4718873799,0.0902521163,-0.1446983218,0.2456972897,0.1922079772,0.8586695194,-0.3126828372,0.2264028192,-0.3521909416,0.10484083,0.2628362179,-0.2384893894,0.0360334925,-0.3349104524,-0.3620525599,0.0524588898,-0.178657338,0.1440884024,0.2543674707,0.0418512151,0.2796762586,-0.1008177102,-0.0681455284,-0.120461233,-0.1016058773,-0.0876769498,-0.5523704886,-0.1459232271,0.1451415271,-0.0558545589,0.144073233,-0.1816010028,0.2091806382,0.2613277137,-0.1393196285,-0.4386136234,0.004906022,-0.2140115052,-0.2438669056,-0.1354427189,-0.1973509938,-0.0141814081,0.0350982696,0.511588335,0.514570415,-0.0757501572,-0.0598519556,0.1307942569,0.1262363046,-0.0812429041,0.1213025749,0.1036277115,-0.0802472606,0.0264716484,0.0093169129,-0.381049633,-0.1335030198,0.1447110027,-0.0712471381,0.1022098437,-0.0101593025,-0.2962921858,0.0573792569,0.2628182471,-0.0605856553,0.0894616395,0.3292609453,0.0108017223,0.3653940856,-0.1192257479,-0.3350674808,0.0266561825,0.297480464,0.156204164,-0.3033962548,0.286462307,0.1983084679,-0.2543386221,-0.1288330257,0.1981595606,0.0063798176,-0.3610734046,0.2128325105,0.1307283491,0.2465961128,-0.0845350623,0.13892591,0.0415939353,-0.2070361376,0.1120714769,-0.1751885116,-0.5877973437,-0.0705102906,0.0973943472,0.2225599587,0.5030190945,-0.3990536332,0.1596841663,0.3320441246,-0.2281685621,0.0169599783,0.043273624,0.1134160459,0.0740878507,0.2638677657,0.0392143726,-0.2002221048,-0.0199869331,0.0728044063,0.0167906582,-0.2414200604,-0.0421738401,0.1874605864,-0.1520764083,-0.0990252867,0.2055355012,-0.2096877992,-0.2729857564,-0.3134709299,0.1171407327,0.3692091107,-0.071194239,-0.2368951887,0.3172477484,0.1279520094,-0.3666725457,0.3836712837,0.0973880962,0.1629292369,-0.1350254714,0.1355592161,0.1831191182,-0.0544473119,0.1036327109,-0.1795713156,0.2208047062,-0.0127160568,0.1711082011,-0.20655334,0.0752992854,0.2629701495,0.2708016336,0.2031385005,-0.1154550835,0.0901802257,0.2642630935,0.150938049,-0.21744214,-0.1999921501,0.0922266096,-0.014361741,0.0580570437,0.2394213676,0.2373979241,0.1460178941,-0.1399421543,0.300352037,0.2328730375,-0.403298527,0.0513429753,0.2246432155,0.1163928732,0.1405850649,0.3162820637,0.3329310417,0.1493568718,0.3646498024,0.3178604841,0.5474416614,0.3969348967,0.0817120597,-0.2365936935,-0.6340230107,0.3120298386,0.0662089512,0.2535705268,0.3112818897,-0.0143228173,0.2013297379,0.0026699002,-0.2013033926,-0.2248368859,0.1003767923,-0.1543974876,-0.1874814332,0.3944633305,0.025604913,0.3287373781,0.1398015469,-0.0750307143,-0.0762116462,0.3307416737,0.007709689,0.1387692243,0.1203110442,0.1984256804,0.0815224051,0.3295721412,-0.1398652494,0.2043153644,-0.0408565588,0.0399612486,-0.0827811584,0.6058395505,0.2702246606,0.4063007832,-0.0962741971,0.1391275972,-0.0157346763,-0.0129840113,0.0126822907,0.2949560881,-0.0441889577,-0.1370933205,0.152249068,0.0454923883,-0.0456068143,0.2766233981,-0.1446354091,0.3570408821,-0.1281332225,0.276389122,0.049251467,-0.3372701705,-0.2017888129,0.1880870461,-0.2160469294,0.1191871539,0.7250255346,-0.3446101844,-0.0458472967,-0.1178604811,0.059643615,0.1917717308,0.4887149036,0.389513135,0.0561827682,-0.1661458611,0.0460257269,-0.4128605723,-0.1396569908,-0.1102350876,0.3076603115,-0.185302034,-0.0820882171,0.1092477143,0.2605444789,0.0179862641,-0.3600306809,0.0547201522,0.4241559803,-0.3278152347,-0.138120234,-0.1829429418,0.0658180416,0.0492774397,-0.500090003,0.1386627108,0.1277717501,-0.1108404621,0.0280090552,0.0821854472,0.1479584724,0.013644387,0.4552467763,0.0333944634,0.1702421308,-0.0950658843,-0.0009280501,-0.3013450801,-0.1201582253,-0.1929363757,0.1922675371,0.1027781665,-0.119512938,-0.5236124396,-0.310749799,-0.0750729442,-0.266427964,-0.2603972554,0.1137377396,-0.320558995,-0.2474634796,-0.1452060193,-0.0577137657,0.1940708309,0.1950250268,-0.0433248095,0.108327888,-0.3000654578,-0.3078199327,0.5511478186,-0.1447297037,-0.3217539489,-0.2479684204,0.0935083926,0.0809068307,-0.2549147308,-0.4340294302,0.1881868988,0.3084205389,-0.1326379627,0.1661923677,0.3081932068,-0.3657455444,0.0797850564,-0.2153674364,0.6930431724,-0.2152214348,0.0158531591,-0.398742944,-0.3405056894]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1776","title":"[Question & Bug Report] Can we preprocess a dataset on the fly?","comments":"Hi @acul3,\r\n\r\nPlease look at the discussion on a related Issue #1825. I think using `set_transform` after building from source should do.","body":"I know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532","comment_length":22,"text":"[Question & Bug Report] Can we preprocess a dataset on the fly?\nI know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532\nHi @acul3,\r\n\r\nPlease look at the discussion on a related Issue #1825. I think using `set_transform` after building from source should do.","embeddings":[-0.4329630136,-0.149762705,-0.0210898854,0.0117295887,0.4850175381,0.2730880976,0.3065044582,0.2762750387,-0.1747130752,0.1026615873,0.0982053503,0.1963543445,-0.0868418068,0.2460448146,0.2638710439,0.1422843635,0.2081043124,0.08988031,-0.096630387,0.1875966489,-0.295986861,-0.1425745487,-0.0633271337,-0.1839763373,-0.46941787,-0.2427725941,-0.1292227507,0.001916178,-0.181621775,-0.203106001,0.2927230895,0.2324055135,-0.0148586556,0.4824579358,-0.0001141936,0.0488191023,0.412786454,-0.1013961732,-0.1420305073,-0.1252627522,0.1022013649,-0.3030607104,-0.1348371804,-0.2586301267,-0.1555446684,-0.1020514742,0.0397011638,-0.6355341673,0.4321356118,0.1466097385,0.2200481743,0.1474144459,-0.0451228656,0.1245692372,0.2611001432,0.3140592873,-0.3821599185,-0.1072558537,0.180836916,0.0386675075,0.0378094502,0.1519301087,-0.326919049,0.0053006238,0.3591570258,0.1101513281,-0.2363422066,-0.5635491014,0.1477532685,0.2872587442,0.4784781337,-0.560829699,-0.7194266319,-0.2661271691,0.1451830268,-0.3385506868,0.1005583555,0.4154743552,-0.170515433,0.0658453032,-0.4523686171,-0.4309619963,0.0252115261,-0.3683648109,0.2405146509,-0.144568041,-0.2335285097,0.0204998776,0.0057092081,0.1899290532,0.4107083678,-0.2188436091,-0.3381609023,0.4355342686,-0.2185551375,-0.2187634856,-0.2661304176,0.0814255625,0.2457327992,0.1398300827,0.0155080305,-0.1161653027,-0.0325907953,-0.0893719718,0.1360760033,0.1203241125,-0.0844900087,-0.1408303976,0.2896365523,-0.2251360416,0.0950003862,-0.098081395,0.0825609788,-0.4091348052,0.3306641877,0.1045778021,0.3203131855,-0.3002013862,0.3989455402,0.0140089439,0.1325963587,0.0150183141,-0.3814519644,0.145734176,-0.1104523465,0.0512313768,0.1466380209,0.142272979,0.0613337867,-0.1713287979,-0.1215861589,-0.0579744615,-0.0176576432,0.3037647605,0.1017586887,-0.2035060525,0.0960330591,0.1968317479,-0.069305785,-0.1474777311,0.1593564004,-0.4074446559,0.1773698032,0.3492155075,-0.0258463807,0.0895796493,-0.0435275286,0.3530143797,-0.1998040676,0.5373976827,-0.1354406178,-0.3122022748,0.0356287472,0.0945883542,-0.0531800017,0.0373736508,-0.5619783998,0.1426612884,0.6161594391,0.2149135768,0.2612337172,0.1844769716,-0.2494729757,-0.1610206962,0.1391241848,0.4578093588,-0.2424228489,0.1542511582,0.3046616018,0.0511818081,0.1385352463,0.0961020291,-0.0289497059,0.2064516246,-0.0387377702,-0.1997522414,0.3259778321,-0.0721958131,0.0467531718,0.3394372165,-0.0908213481,-0.2645249963,0.0823537037,0.1384712756,-0.0276599843,-0.0962675363,-0.0548363104,0.4266054332,-0.0195711553,0.2360433042,-0.1810713112,-0.4425275922,-0.2827946544,-0.0746699944,0.0176095404,0.137852028,-0.196992442,-0.2738490999,0.0570900589,-0.1821342707,0.2325855792,0.3486989141,-0.0456053466,0.0465126596,0.0585953332,-0.1447848231,-0.3847301304,0.2210926712,-0.10427057,0.0422115959,0.1136271581,-0.4316208363,0.1256343722,0.1094216779,0.0725463331,0.1065265983,0.0396175422,-0.3019027114,0.294749856,-0.0410910696,-0.4444380403,0.2486053556,-0.1694128513,0.112823531,-0.4200748801,-0.0761258751,0.1506418884,-0.0647382587,-0.0618649907,-0.1236242577,-0.1879749298,-0.0074727675,-0.1552285254,0.3102422953,-0.1509397626,0.2376552671,-0.0833659768,0.1479143947,-0.1031506434,-0.2216746509,0.0134715196,-0.1461161673,0.1204976365,-0.1898351312,-0.2374291122,0.2405589223,-0.125441432,0.4528132379,-0.0244471598,-0.1067399904,0.19655931,-0.0574267358,-0.0650350153,0.098939389,-0.0469948873,-0.0435777158,0.7283758521,-0.0893419087,-0.0646382645,-0.0855534598,0.0716886148,0.0146110766,-0.0450502597,0.2503600121,-0.1430051625,-0.1239451095,-0.2058190554,0.1123894826,0.2271235883,0.1460506618,0.0281832051,0.2094620317,0.0540075414,-0.1065891832,0.331243515,0.0195442215,0.4602996409,-0.0285264142,0.2461500764,0.1056514382,-0.2152439803,0.2044366449,0.0851661637,0.0561343469,-0.1728527695,0.1126303226,0.1550612152,-0.069464922,0.0361564942,0.0005299171,-0.224599123,0.0833590478,-0.1207841113,-0.0309183504,0.0240624063,0.0583936274,-0.4411794543,0.3388668597,0.0531037003,-0.1855516434,-0.1853674948,-0.2100474685,0.2090856135,0.0030894969,0.191182062,-0.1702626795,0.4499090612,-0.0654062256,0.0290908013,-0.2961729169,0.1470841169,0.0948989466,-0.0052774232,0.3585245013,0.0148649886,0.0273944866,0.2091139704,0.1998988837,0.0595944412,-0.1688087285,-0.1806282252,-0.2080806196,-0.0564350933,0.0179906879,-0.1513291448,-0.006899714,-0.0297326576,-0.2902205884,0.1317759156,-0.027170239,0.022543136,0.2314047217,0.0408415608,-0.1508438587,-0.0609039739,0.0651347861,-0.0411670655,-0.6186643839,0.301677078,-0.0587813146,-0.3465429544,-0.0441343002,-0.027905494,0.0178735331,0.3690830171,-0.603469789,0.0759540871,-0.4336880147,0.0787584335,-0.1546194106,0.1942055672,0.3209499419,0.1343019754,-0.095620364,-0.0347259119,-0.1735876054,0.0148214698,-0.1583309025,-0.0087433467,0.2583470941,0.4254817069,-0.0141152879,0.7667742968,0.1563379019,-0.0408220962,0.3145268261,0.017723931,0.587883234,-0.3564191163,-0.1258499771,-0.181350708,-0.2166433185,-0.149137333,0.0718437955,0.2537595928,-0.2575016022,-0.2045564353,0.0701103061,-0.1112936661,-0.1172044575,0.1579772085,-0.3516556323,0.2993707061,0.0467632636,0.1009597033,-0.2630337477,-0.3549570143,0.1259233654,-0.0599192791,0.685069859,-0.1539006084,-0.1800369918,0.1375827044,-0.6912074685,0.2251766771,0.0109912232,0.2751890421,0.1694918573,0.012461815,0.0676198974,-0.3236368597,0.5596726537,-0.2098633349,-0.0185781121,0.3916489482,-0.0410803296,-0.3185816705,0.1438379139,0.0832032189,0.5215532184,-0.1703884453,0.2730243504,-0.0799725577,-0.1630998701,0.1364084929,0.408241868,-0.2205162942,-0.4579713941,-0.2055647522,-0.2104367465,-0.2233043611,0.1625536978,-0.000497964,-0.1777331531,-0.1616404504,-0.11082232,-0.3879926503,0.2419279665,-0.1016708985,-0.1816687584,0.1594799757,-0.0070234649,0.502897203,0.0914542004,-0.1354880929,-0.046701353,0.6026400924,-0.1695320606,-0.1621041894,-0.1636356711,0.0054741069,0.3667346537,0.3100365698,0.0287972149,0.140553534,0.0074562761,0.105636321,-0.5073130727,0.0600987114,0.1315062791,-0.0806751698,-0.5603205562,-0.6287379861,0.6531296968,0.3827987611,0.0311174095,0.4319263995,0.1165407673,-0.16844064,0.3139211237,0.266528517,0.8644783497,-0.3468036056,0.2490417063,-0.2838608325,0.0972332209,0.3066899776,-0.2182277143,0.0343566276,-0.3303461969,-0.3620761037,0.0688374639,-0.1924998909,0.1141512543,0.2670470178,0.0511521213,0.2359419912,-0.095877476,-0.1219236627,-0.1197623834,-0.0817142874,-0.1360372156,-0.60080266,-0.1783622503,0.1515986174,-0.046802856,0.1319120675,-0.1145199165,0.2092532218,0.1811388284,-0.1507108063,-0.4558628201,0.0270152427,-0.2561564445,-0.2469817698,-0.1386923492,-0.2268927544,-0.0680658296,0.0406020768,0.5350571871,0.4937665462,-0.0453580506,-0.0554353073,0.1498515606,0.1393456608,-0.0006824836,0.1040051207,0.1361522228,-0.0553039536,0.0626219213,0.0052337088,-0.3632959425,-0.1164649948,0.1880636364,-0.1644742489,0.0754111111,-0.0557452105,-0.2545494139,0.0733781755,0.2466436476,-0.1198310256,0.095756188,0.275167048,0.0035112286,0.2970243096,-0.1572024971,-0.3641905487,-0.0224305168,0.2522280514,0.1832285821,-0.2786245346,0.3542678356,0.1972767413,-0.2224344015,-0.1341266036,0.1842357516,0.0182401501,-0.4513961077,0.2480618507,0.1262967288,0.1829118878,-0.1139216051,0.1716469228,0.0206693616,-0.1863154322,0.1057582349,-0.1607483327,-0.5908619165,-0.0775445551,0.1379878372,0.2136132568,0.4876232147,-0.3452980518,0.1147185117,0.3632794321,-0.2515030503,-0.052747082,0.0716380849,0.13769871,0.081459403,0.2629783452,0.0363624282,-0.1899966151,-0.0067937863,0.0413637497,0.0050946139,-0.2564018667,-0.065159373,0.1668859571,-0.1567249596,-0.0877524093,0.1447249353,-0.1997513771,-0.2308963537,-0.4138133824,0.1614041179,0.3609610796,-0.1076853722,-0.2222438455,0.2847194076,0.1251553744,-0.3574776053,0.2958545685,0.0576341338,0.1430909187,-0.1466937363,0.1799290925,0.1828742623,-0.0703944713,0.0689473972,-0.1754450649,0.2120843679,0.0560050718,0.1681276411,-0.2220020592,0.0552012175,0.2648106515,0.3314078748,0.191004619,-0.0749554485,0.0767024904,0.2916676104,0.1731113195,-0.2603944242,-0.1648773998,0.0726814196,-0.0823378488,0.1157803833,0.2786782682,0.2319999039,0.1014923528,-0.108148776,0.2868593633,0.1727774739,-0.4057714343,0.0897596329,0.25126791,0.1278176308,0.1547182053,0.3263249397,0.3106876612,0.188029483,0.3593755066,0.2982604504,0.5835420489,0.3375901282,0.0854640827,-0.2475432158,-0.6166616082,0.2919543087,0.012683413,0.253706038,0.3417038321,0.0011655862,0.2255642563,0.0410305597,-0.2557652593,-0.2489865571,0.1445866972,-0.1577445865,-0.1740818322,0.3542081714,0.0349883586,0.3517551422,0.134575218,-0.068427667,-0.0899404362,0.320682466,0.0430112407,0.1703484207,0.1047036126,0.2247071862,0.007458515,0.3076194227,-0.1570518911,0.1943330169,-0.0012040465,0.0391936488,-0.0729845092,0.5906715989,0.312756151,0.3928544223,-0.1074756756,0.1859054416,-0.0150375506,-0.0197711308,-0.0328730009,0.2742159367,-0.0866408944,-0.1349046379,0.1934501827,0.0645885691,-0.0940357596,0.1881127059,-0.1571763456,0.3570961952,-0.107585445,0.3226363063,0.096985139,-0.3717995286,-0.2190296054,0.1861296296,-0.1853238344,0.1298693568,0.7191513777,-0.3829756081,-0.0661422238,-0.1163975,0.0697057173,0.2253584117,0.5203012228,0.385986805,0.0675291345,-0.170445919,0.0798051506,-0.3966825306,-0.1173795313,-0.1092893258,0.2891944945,-0.1593012214,-0.0939625502,0.1794192195,0.2901896238,-0.002313575,-0.3146266937,0.0482032597,0.4311459363,-0.2662266791,-0.2024322599,-0.1907080114,0.0574135184,0.023413511,-0.5342944264,0.1379930675,0.1352623701,-0.0791158602,0.0726100951,0.1056387872,0.1728839278,-0.0099298116,0.4256455302,0.0713874698,0.164506495,-0.1374816597,-0.0105165383,-0.300101161,-0.1640689373,-0.2427419871,0.1930494159,0.0539636463,-0.0371851325,-0.4960953295,-0.2699705958,-0.1496705562,-0.2413807064,-0.1611332446,0.1025703251,-0.3433539271,-0.1769263148,-0.1870300025,-0.0430507474,0.1983023286,0.212907508,-0.0182533581,0.1253940165,-0.3085353971,-0.3872483373,0.5898017287,-0.1565357894,-0.3682809174,-0.2112172097,0.1110893041,0.0912112743,-0.2258026004,-0.39421615,0.2338773161,0.2611788213,-0.1221783385,0.121526137,0.2710575461,-0.3783168793,0.0823223293,-0.2273855507,0.6979739666,-0.2254201323,-0.0171109997,-0.4264149368,-0.2989898324]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1776","title":"[Question & Bug Report] Can we preprocess a dataset on the fly?","comments":"@gchhablani thank you so much\r\n\r\nwill try look at it","body":"I know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532","comment_length":10,"text":"[Question & Bug Report] Can we preprocess a dataset on the fly?\nI know we can use `Datasets.map` to preprocess a dataset, but I'm using it with very large corpus which generates huge cache file (several TB cache from a 400 GB text file). I have no disk large enough to save it. Can we preprocess a dataset on the fly without generating cache?\r\n\r\nBTW, I tried raising `writer_batch_size`. Seems that argument doesn't have any effect when it's larger than `batch_size`, because you are saving all the batch instantly after it's processed. Please check the following code:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/0281f9d881f3a55c89aeaa642f1ba23444b64083\/src\/datasets\/arrow_dataset.py#L1532\n@gchhablani thank you so much\r\n\r\nwill try look at it","embeddings":[-0.3963780701,-0.2699373364,-0.0596869923,0.0206832662,0.4002688825,0.3706586957,0.2639387846,0.269467473,-0.0940201357,0.1387309432,0.0946395919,0.1482467651,-0.109260194,0.2208978534,0.2471573204,0.1389282197,0.1943968087,0.0875917152,0.047151763,0.2221857309,-0.2532025278,-0.1831166595,-0.1161879897,-0.1799758673,-0.4234603643,-0.2745017111,-0.1588207036,-0.0409512073,-0.2042390108,-0.1730399281,0.3297152519,0.2718337476,0.0821759254,0.5273167491,-0.0001156312,0.0177009013,0.4247830808,-0.0858545452,-0.0670243129,-0.0979572684,0.076722838,-0.2295341492,-0.1374184042,-0.2835773826,-0.1494528353,-0.0165846162,0.0951016098,-0.6653779745,0.3692826331,0.1179353297,0.1935073435,0.1317332983,-0.1170118302,0.1430439353,0.2809810638,0.3523539901,-0.3516600728,-0.0980626941,0.2035966069,0.0042198258,0.0924571306,0.1137757301,-0.3742207885,0.0119807506,0.2897333503,0.1418007761,-0.1840186864,-0.6168151498,0.1538583487,0.2618382275,0.5053151846,-0.5414749384,-0.6362533569,-0.178010121,0.1262356788,-0.3183552027,0.0406002887,0.4248283505,-0.1589541286,0.0529149845,-0.4602016509,-0.4128107429,-0.0084663192,-0.3262822032,0.2603620887,-0.1245802939,-0.2414903045,0.0438629389,0.0038920504,0.1647661626,0.4145130813,-0.2774051726,-0.3580848873,0.4815643132,-0.2550781071,-0.2545881867,-0.3270524144,0.1024664789,0.2833714485,0.1010250747,0.0106914686,-0.1597763747,0.011401643,-0.069197081,0.1725240052,0.0124330334,-0.0339139141,-0.1596907675,0.3042763174,-0.2450617105,0.0966729224,-0.094588235,0.0687236041,-0.3778977692,0.3818573952,0.0600836873,0.3315643668,-0.3201246858,0.407138437,0.0057293535,0.0749928802,-0.0029031248,-0.4126420617,0.1289419681,-0.1611051857,-0.056387607,0.1322657615,0.0709234327,0.0735525489,-0.1838813573,-0.0987066403,-0.0035577954,0.0240282826,0.2789648473,0.0929639041,-0.2167078108,0.0735405684,0.1932341605,-0.1089172289,-0.2046993375,0.1609139144,-0.4740567505,0.1035928726,0.3211582005,0.0623421408,0.0749852285,-0.072175391,0.3756803572,-0.19468759,0.5445746183,-0.1761361957,-0.2782934606,0.0281914696,0.0877456367,-0.0143471593,0.0382439047,-0.616710484,0.1448685676,0.6152248383,0.2389906198,0.2742651403,0.2696401477,-0.3014744222,-0.1472842693,0.1165877581,0.4581457973,-0.2180867642,0.1855753809,0.265050739,0.0295995902,0.1920804828,0.0726340935,0.0126505969,0.2786057591,-0.070591189,-0.164181143,0.3590724766,-0.111126855,0.0952586308,0.4140898585,-0.0725266859,-0.3050426841,0.0721720606,0.189704597,-0.0387923978,-0.083546944,-0.0109552667,0.3855225742,-0.0024445113,0.2042468339,-0.1645145118,-0.4521510303,-0.2830235958,-0.0342520438,0.0087059978,0.1173039228,-0.1761152446,-0.2904418707,0.0768326744,-0.1656025052,0.2180761993,0.3960106373,-0.0515473895,0.0486526154,0.0353408977,-0.1786418557,-0.3569026887,0.2030697763,-0.1022339165,-0.0103905033,0.1921082586,-0.4693473279,0.1381455064,0.0791683495,0.0823971331,0.1584845036,0.0186091941,-0.3526956141,0.3002455235,-0.0174253602,-0.4302161932,0.2570854723,-0.2065984011,0.0748430863,-0.3623390496,-0.0741557255,0.1469956487,-0.0532944687,-0.0577958971,-0.1696858704,-0.2115025371,0.0247006845,-0.1076619998,0.2977830172,-0.1406491846,0.1519868225,-0.0469310433,0.1044423431,-0.160523355,-0.1940455884,-0.0211881232,-0.1532703042,0.1506917328,-0.2010622174,-0.2277744859,0.1913166195,-0.1015321389,0.4339427948,-0.0729039535,-0.1610773802,0.1817876846,-0.058995381,-0.0238613263,0.142760247,-0.008132631,-0.0341391638,0.7227954268,-0.1329113096,-0.0832818523,-0.1854902357,0.0078930482,-0.0147355748,-0.0043914169,0.2552455962,-0.1400571167,-0.1072903499,-0.2198107988,0.0503911115,0.2506258786,0.1367792487,0.0286954865,0.2149310857,0.1224011257,-0.0681637302,0.3098418415,0.0079067303,0.4381504059,-0.0612254664,0.2153004706,0.0904530212,-0.1856395751,0.189775005,0.105524838,0.0814538747,-0.1488022804,0.1117699221,0.1801401675,-0.1417727172,0.1401697993,0.026288053,-0.21260041,0.1233665496,-0.1488449872,-0.0324932858,0.1233258769,0.000990169,-0.4470859468,0.3995273411,0.102027528,-0.1726061553,-0.185608834,-0.1972857863,0.2070890516,-0.0032057913,0.2164865583,-0.1666543037,0.4623328149,-0.0587492511,0.0527450368,-0.3228973746,0.2259444445,0.0856503993,0.0460502058,0.3504273593,-0.0038370339,0.080384098,0.1728916466,0.2337858677,-0.0027987319,-0.1318110228,-0.1935571134,-0.2584287822,-0.0503347367,0.011449296,-0.1641427577,0.0754138902,-0.112194784,-0.2269431651,0.1008694619,-0.0180411637,0.0418887287,0.1864375025,-0.0167260207,-0.1509581357,-0.1289444417,0.0715154707,-0.0453077257,-0.7272841334,0.2738019824,-0.0102228569,-0.3445872366,0.020957727,0.0065342942,0.0175025854,0.3197810054,-0.624661684,0.0556191877,-0.3844240308,0.0562049374,-0.2016119808,0.1677177548,0.2831380665,0.0766935498,-0.0495812371,0.0402441174,-0.1356847882,0.0009445982,-0.1947954744,0.0014875508,0.3052511215,0.439802736,-0.0341784917,0.6765698791,0.1703455597,-0.0199353043,0.3351294696,0.0083743213,0.5390661955,-0.2626231313,-0.0840448439,-0.1922981441,-0.1794324815,-0.1563004404,0.0932097733,0.2128505409,-0.2869414687,-0.1965849996,0.0054249545,-0.082695134,-0.0803015679,0.0753440484,-0.3362888992,0.3011340201,0.004298544,0.0367643572,-0.2865436673,-0.4156557322,0.1141539812,0.0365693681,0.6741656661,-0.1642793864,-0.1973815858,0.1919383854,-0.6622305512,0.2190826535,-0.0013860567,0.3367394507,0.1499789208,0.1046953648,0.1207845584,-0.2787526846,0.5854111314,-0.2556190491,0.0332361832,0.3738041222,-0.0157192368,-0.2635178566,0.1385295689,0.0767919421,0.5807870626,-0.1941949874,0.2824613452,-0.0301897246,-0.1678883284,0.1253339797,0.3494532406,-0.242079556,-0.5930480957,-0.2017902434,-0.1658680588,-0.154409647,0.176927954,0.0194244981,-0.1666681617,-0.101552546,-0.135882169,-0.4560400844,0.3039650321,-0.043136131,-0.1854304522,0.104696013,-0.0273313411,0.4903509617,0.0114568081,-0.1669923216,-0.0872022733,0.5035831332,-0.2067304701,-0.1839146763,-0.1352506876,-0.0107418979,0.4203826487,0.1981285959,0.0478276722,0.2497997135,0.0035542059,0.1064971462,-0.5146233439,0.1092473492,0.1029704288,-0.0582070127,-0.6595358253,-0.6095448732,0.6146941781,0.4636205137,0.0617325641,0.4299257696,0.0722428039,-0.162802279,0.2521599531,0.2452404797,0.851074934,-0.3635547757,0.2139333934,-0.3555929065,0.1609016657,0.284591496,-0.253357619,0.0356628038,-0.3193680048,-0.3936892748,0.0308126733,-0.1695743054,0.119142592,0.2974778414,0.0452376716,0.2854276597,-0.089777045,-0.0442909226,-0.106685929,-0.1099286005,-0.0637402609,-0.5736804008,-0.1546770781,0.1408970058,-0.0900180936,0.1395102739,-0.1684435606,0.2277221084,0.2642521858,-0.1477137059,-0.4050736427,-0.0057600425,-0.2360583991,-0.263766259,-0.1967220157,-0.1726365834,-0.0631604344,0.0627696961,0.5011712313,0.5586279035,-0.0384362116,-0.0962997228,0.1222491488,0.0822097287,-0.0921281055,0.11366532,0.048995357,-0.0705260709,0.0323858932,0.0551415756,-0.3723394871,-0.1409771442,0.1966385692,-0.0793292448,0.0941231251,0.0126143647,-0.3300968409,0.0599682927,0.2588136792,-0.0466126241,0.0756142139,0.3048135042,0.0078794528,0.3260024488,-0.1256938875,-0.3668502271,0.0119121177,0.2663677335,0.1927668601,-0.2832932174,0.2959339619,0.1611096859,-0.2260062397,-0.1335148811,0.192396611,-0.0246341322,-0.3375849426,0.1945447028,0.1413900554,0.2771757841,-0.0884485468,0.1486727744,0.0392480195,-0.2032376528,0.1317270845,-0.1917875409,-0.6049103141,-0.1059483364,0.1207636818,0.2113292515,0.5734285712,-0.3531159759,0.1093381718,0.3397013545,-0.2206737548,-0.0200568885,0.0187319573,0.1167706028,0.10144566,0.3031184375,0.0308624543,-0.1713505089,-0.0166470502,0.0871150345,0.0642003343,-0.2588780224,-0.0171687193,0.1852422804,-0.1452804655,-0.0597721562,0.1921737492,-0.2002736479,-0.25148651,-0.3492163122,0.1132150143,0.3617615998,-0.0750097185,-0.2592402399,0.3350555003,0.1302541047,-0.427513063,0.3443126082,0.0675306171,0.1481648386,-0.1586336941,0.1525290459,0.20849666,-0.0475451164,0.0949062631,-0.2019593865,0.2784481347,-0.0272214357,0.1119530797,-0.1810279489,0.0709397346,0.2770771086,0.2376763672,0.1869467348,-0.0966609865,0.0962112844,0.2751815617,0.1495196372,-0.2243721038,-0.1857292801,0.1110737473,-0.0592785366,0.102580294,0.2821657658,0.1856367588,0.1485946178,-0.1757582873,0.2734597027,0.2169089019,-0.3978438675,0.0255459957,0.23856318,0.1591196805,0.1473368555,0.2942468226,0.3509289622,0.1511117071,0.343945682,0.3224249184,0.5895383954,0.3599142134,0.061332792,-0.2748076916,-0.640658319,0.3119876981,0.0801077113,0.2743702233,0.3254688978,-0.0078062089,0.17057693,0.0450963192,-0.1949580014,-0.240472883,0.100600034,-0.1441679895,-0.1669578403,0.371622771,0.0627990514,0.3609543443,0.1256708801,-0.0302469917,-0.1105441228,0.2873829007,0.0173406079,0.1557271332,0.1642630398,0.1954055876,0.1178635508,0.2684461474,-0.1667502224,0.1649568528,-0.0336192101,0.0618477538,-0.0925557911,0.5549312234,0.2401469648,0.397333622,-0.0653524771,0.1567449719,-0.0000533363,-0.0222292077,0.0011139432,0.2448000163,-0.0732206181,-0.141274631,0.1562197655,0.0521969795,-0.0540892817,0.2557226717,-0.1299265772,0.3677465916,-0.1152041778,0.2744580805,0.0681562647,-0.3476177156,-0.1988572031,0.1962357014,-0.2259154916,0.1520846933,0.696107626,-0.3788110614,-0.0482192934,-0.082105957,0.0517675765,0.1710101515,0.468888998,0.3844719529,0.0778481141,-0.195096314,0.063689217,-0.3709209561,-0.1412638873,-0.1386759132,0.2402573228,-0.1926654875,-0.1190546602,0.1255681366,0.2708852887,0.0039026032,-0.3255684376,0.0397704616,0.4542101026,-0.3056806326,-0.1823941022,-0.1887935698,0.0603202023,0.0501537509,-0.526171267,0.1363077611,0.1394505501,-0.1042183265,0.0344852023,0.1121865362,0.1522815824,-0.0129106222,0.443475008,0.0094382679,0.1839092523,-0.1544112265,0.0165696815,-0.251486361,-0.0649147481,-0.2189585865,0.1807660162,0.1128549054,-0.1151954383,-0.5404207706,-0.2619356215,-0.0631869957,-0.2738595009,-0.2108430862,0.0929942057,-0.3149286807,-0.2479957342,-0.1468553543,-0.0455220714,0.2192450464,0.1568731964,-0.0099949259,0.1248348579,-0.28435269,-0.2873159945,0.5400215387,-0.145314917,-0.3084610105,-0.2682121992,0.0971677899,0.0644994155,-0.2183614522,-0.3997899592,0.2150957733,0.329382062,-0.1258538663,0.1823340654,0.3122740686,-0.3683247268,0.1105987057,-0.1963896155,0.6848792434,-0.2417512387,0.0237465315,-0.4157450199,-0.3622170985]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1775","title":"Efficient ways to iterate the dataset","comments":"It seems that selecting a subset of colums directly from the dataset, i.e., dataset[\"column\"], is slow.","body":"For a large dataset that does not fits the memory, how can I select only a subset of features from each example?\r\n\r\nIf I iterate over the dataset and then select the subset of features one by one, the resulted memory usage will be huge. Any ways to solve this?\r\n\r\nThanks","comment_length":16,"text":"Efficient ways to iterate the dataset\nFor a large dataset that does not fits the memory, how can I select only a subset of features from each example?\r\n\r\nIf I iterate over the dataset and then select the subset of features one by one, the resulted memory usage will be huge. Any ways to solve this?\r\n\r\nThanks\nIt seems that selecting a subset of colums directly from the dataset, i.e., dataset[\"column\"], is slow.","embeddings":[-0.272621274,-0.3842424452,-0.0938193724,0.3328414559,-0.064883776,0.0687170699,-0.0869265497,0.2765877545,0.1568154693,0.1853520125,0.1041573808,0.0968727618,0.075772956,0.1713013947,0.0594490469,0.02440935,-0.1179833263,0.2368059754,-0.1938435286,-0.0943645537,-0.3235397935,-0.3371296525,-0.3414105773,-0.413921237,-0.0448212847,-0.2763502598,-0.0030553236,-0.0701165721,-0.1004956514,-0.2022732198,0.0649098828,-0.1059204489,0.1723704934,0.1313776672,-0.0001110769,-0.3096601963,-0.0061720056,-0.0167973097,-0.2181179821,0.2127902359,-0.4324012399,-0.1536879987,-0.209673211,-0.4574201703,0.0789363161,-0.193544507,0.0418520048,-0.3434140682,-0.3945272565,0.0019652992,0.1722190231,0.3632744551,-0.4281520247,-0.2193821967,0.1636476517,0.2020667642,-0.1860643774,-0.1688093543,0.3972516656,-0.0317177549,0.1105101332,0.1986813247,0.0027408882,0.0643330738,0.4581536949,-0.0933370367,-0.2695295811,-0.0912525281,0.2907192707,0.3570749164,0.7055131197,0.0142396716,-0.0204161704,-0.2548285127,-0.1146911904,-0.4821657538,-0.0554719344,0.2097758949,-0.1812402755,0.1794072241,-0.3754857481,-0.282171011,-0.1214113608,0.1009940654,0.2251813561,-0.1510582417,0.0702474043,0.0095484257,0.5192238688,-0.222860828,0.2806361318,-0.1883521229,0.042395819,0.1281944066,-0.5794634223,-0.3097940385,-0.1118838117,-0.0624917336,0.3452713192,0.107684426,0.08151225,0.0904842541,0.2307037711,0.0643644556,0.3840838373,-0.1338301599,-0.1883253604,0.1345576942,0.1456446201,-0.160663262,-0.0020948441,-0.0907806307,-0.2182660252,0.0937924832,0.2085401565,-0.4199842215,-0.1569328755,-0.1290027648,-0.149421066,-0.1261437833,-0.1378621161,-0.0162505209,-0.0246443041,0.4202540517,0.0619461909,-0.1889308989,-0.4265065789,-0.0111111114,-0.2233894914,-0.2640571594,-0.2986492515,0.2844229043,-0.0498715825,0.0189673565,0.0652078763,0.1994225979,0.1540557295,-0.1386214793,-0.1547173858,-0.1392389834,0.0547332019,-0.3360626101,0.2683524489,-0.0082828514,-0.1358440518,-0.0077023222,-0.0597725362,0.336125195,-0.3245050013,0.2191703767,0.1463289112,-0.5682111382,0.183412537,0.0710723624,0.0017380278,-0.0667319819,-0.3420160115,0.4201233983,-0.0989892408,0.4257798791,-0.0995337963,0.0706477538,-0.3074536622,-0.1526209116,0.372374475,0.4698694348,-0.3293994665,0.051087141,0.1202649251,0.13932015,0.3856441379,0.4193019271,-0.220395565,0.1525849849,-0.1224395558,0.0959463716,0.1902549267,0.1820035726,-0.5025990009,0.3416081369,0.108068198,0.0722895935,0.2504362464,0.5023416877,0.0730348751,0.0638115257,-0.1298099905,0.3935356736,-0.2775791883,0.1868896633,-0.0036721453,-0.3479519784,0.2195587009,0.0396700948,0.2890775204,0.2412097603,-0.008328557,0.044327341,0.2857303619,-0.3043045998,-0.1506115645,0.1334234476,0.5146899819,-0.0039331494,-0.0409264006,-0.3039368987,-0.3744559288,0.2273036391,0.4574954808,-0.5265105367,0.3239097595,-0.1340290457,0.1073588058,0.018804796,0.1776627302,0.1412481368,0.0175361764,-0.0444856323,0.2947908342,-0.4913444817,-0.2310359031,0.0906002671,-0.3048676252,-0.1007343754,-0.2995914519,-0.0163098555,0.5265932083,0.0019104267,0.2137762606,-0.0286798533,-0.0328834429,-0.2907178104,0.14983356,0.1483504176,0.3771882653,-0.0658315644,0.1169122458,0.3529287875,0.0880182907,-0.1726694256,0.0474601462,-0.3250952065,0.2133404762,0.031034952,-0.210539028,0.1571555734,0.0987472236,0.5199263692,0.2963037193,-0.0036353387,0.027996894,-0.0309289023,0.2764967978,0.0714425743,0.1936671436,0.1736863703,0.1350107789,0.35304901,-0.6447686553,0.0454972126,0.199426353,0.0886545852,-0.0918916166,0.0251979101,-0.1220782474,0.0939576253,0.1217482761,-0.3537807465,-0.0296607222,0.0877936333,0.0354814269,-0.1350253373,0.0879065171,0.10482876,0.1275731772,0.2802185714,0.3064313829,-0.3647532761,-0.0424480997,-0.2706400454,-0.1003278941,0.0671235397,0.0494892485,0.1145843267,0.0018403609,0.0724698827,-0.355209887,-0.1168477088,0.2057601959,0.3313178122,0.2393414825,-0.1104752049,-0.1124653295,0.5158164501,-0.0581075549,-0.1062864065,-0.1323304027,0.5061990619,-0.0120506072,0.0168139953,0.1103373021,-0.0603423305,-0.069999136,0.0045810724,0.3546584249,0.1378186643,0.3736892641,0.12142355,-0.1678830236,-0.4728028476,0.1906893551,0.1277216971,0.2162524611,0.0835691318,-0.1990766376,0.1286506951,-0.3519138992,0.1047185734,-0.1672060639,0.1980487108,-0.0278880168,0.0853997767,-0.0343408547,-0.0196806658,0.3902205229,-0.3208624721,-0.2228164524,-0.292335242,0.3292638063,0.0246352442,0.2442338169,-0.0116648702,0.0537039451,-0.2610080242,-0.0664389506,0.0236434694,0.0586672686,-0.1673668623,0.2899850905,-0.0720577836,-0.1938501447,-0.0757572055,0.0666714013,-0.028663991,0.5141797066,-0.4075911641,0.0645164996,-0.5274323821,-0.0467714816,-0.1802237332,0.0347189903,-0.0563657843,-0.2087927014,-0.0364183746,0.0583287887,-0.0273865499,0.2015375346,0.1771997958,0.054727219,0.3681704998,0.1776552051,0.2260527313,0.2913785875,0.3049386442,0.2829522789,0.2341083735,-0.0918738768,-0.132628575,0.0755493417,-0.0428376831,0.2239385396,-0.1186179668,-0.2447644919,0.2743477821,-0.1057319716,-0.2245058864,-0.0041373861,-0.061805103,0.0859005377,-0.1929441988,0.2523665428,0.0101647312,0.2669517994,0.15922755,-0.0574014224,-0.3361461759,0.0413213521,0.0789017305,-0.3824475408,0.300942421,0.0097929593,-0.6226494908,-0.1496895403,-0.3150054812,-0.004921651,-0.1470125318,0.3232965469,0.1405627578,-0.331343472,0.0358268395,0.2714028358,0.5188843608,0.2771598399,-0.1971921772,-0.0996634737,-0.6063248515,-0.0046596788,0.1985527873,-0.3383186758,0.4118952155,-0.0864855424,0.1271384805,-0.1311614066,-0.3343245685,0.3492299616,0.5830070972,-0.1878553927,-0.1890680939,0.2267914712,-0.2329859734,-0.0603498593,0.2388966829,0.2808482051,0.0524784103,0.2656178474,-0.0532157123,-0.3900204301,-0.0699562356,0.4186623991,0.1585326195,0.0948848426,-0.0974436551,0.3590675592,0.0423909351,-0.0083671138,0.423368603,0.5041764975,0.2458576709,-0.4154836535,0.0452452973,-0.1006383002,0.7665141225,0.2695271969,0.0186782256,0.2186739147,-0.2198276371,0.1789371371,-0.3878577054,-0.1302852035,0.0250329357,-0.1545523703,-0.2564932704,-0.3809866905,0.4144178927,0.0153406234,-0.0022170111,0.2992088497,-0.0444113053,0.1037926599,0.0025705597,0.2479048222,0.7095934749,-0.1336078197,-0.1503455341,-0.1070073247,0.3820200264,0.2121770233,-0.0353323929,0.3061558306,-0.1831867844,-0.8363991976,0.0446951613,-0.1112045646,-0.0182258356,0.2377367169,0.1937702298,0.4545472562,-0.0720031783,0.107569091,0.065919891,-0.1468530148,-0.0741967037,-0.2159574479,0.0487664081,0.0783154592,0.1801817715,-0.0230883677,0.0946116075,0.1520499438,0.1482019424,-0.0146763073,-0.2205895633,-0.0444679633,-0.3808138967,0.0430257283,0.0243885685,-0.3118437231,0.0591275096,-0.024910707,0.003538142,0.2777443826,0.1099119931,0.0352247879,0.250372529,0.1068689451,-0.1262620687,0.0785343647,-0.0459505916,0.0317995064,-0.1842186898,0.1145003885,-0.0049385801,0.052902557,-0.0910719633,-0.0822486058,0.2172928154,0.0548579767,-0.4014465213,0.1169022918,-0.1205481738,-0.1110370308,0.1151477471,0.0813661143,-0.0654046461,0.4636640549,0.0919792503,-0.1415277421,-0.1286737621,-0.1442203671,0.2252931893,0.0965717882,0.375638783,0.1870127469,-0.2087515891,-0.2522000074,0.1631618738,0.3087825179,0.1714212894,-0.0227549952,-0.3325639069,0.1336074322,-0.0184555426,0.1659062058,0.0621697307,-0.209488675,-0.2350201011,-0.0221911911,-0.0307509042,0.3404422998,0.2304009944,0.2985278666,-0.2309800237,-0.3417856097,-0.4882922471,0.43654719,-0.2348097414,-0.0507626683,-0.3153342903,0.2696285546,0.1566262841,0.3287502229,-0.1983394623,-0.0383740477,0.1348455101,0.175371021,0.0400101654,-0.1948371977,-0.0821788758,0.1031888053,0.0268351343,-0.2245436162,0.0337737091,-0.23298572,-0.3304444849,-0.4621060193,0.2427948862,-0.0296162926,0.149723947,-0.0282516871,0.2559702396,-0.4835476875,-0.2318070531,0.373015821,0.1059627309,0.1376313567,0.1328078657,0.2342755198,0.3057880998,-0.1516671926,-0.1500672847,0.0125362277,0.0622963496,-0.1901885271,-0.127241388,0.1784268171,0.0779715702,0.322512269,-0.1324900985,0.6161224842,0.2012063116,0.1732307523,0.3271650374,0.2236049771,-0.1429850757,-0.4484677315,0.01364116,-0.2504253387,0.1873440444,0.564892292,0.2223634571,0.2721704245,-0.1308403462,-0.3672899604,0.1098507419,-0.0540547036,0.0232716668,0.2495568097,0.2116692066,-0.4596951306,0.0172507651,0.5931332707,0.4760163128,0.5107029676,-0.084830679,0.3243482113,0.2221823633,0.1353186518,-0.1216244027,-0.6607636213,-0.0843937695,0.4037324786,-0.0318579264,0.015597946,-0.1882157773,-0.4264801443,0.0212306157,-0.1018531844,-0.2572141588,-0.0009442301,0.0771638602,-0.1222023591,-0.0241955668,-0.2490437329,-0.1489127427,0.2084702551,0.0189621188,-0.4960002005,0.3610440791,0.0067266626,0.3690362275,-0.2016271502,0.5517862439,-0.0104258209,0.2209771276,-0.3973642886,0.0490523875,-0.2184408009,0.0852004811,-0.0450373553,0.0911324769,-0.2329302132,0.1148798391,-0.0150885917,-0.0266086664,-0.1786059141,-0.2313803285,-0.1345183402,-0.2221688032,-0.2319947332,0.1745178699,0.3362204134,-0.0219492409,0.018803075,0.1220083535,0.0644322857,-0.238884598,-0.6822423339,-0.2142414898,0.1744690984,-0.2910794616,0.1559342742,0.0221108887,0.0655490682,-0.2823663056,0.420737654,-0.466041714,0.1919424683,0.5151318908,0.0926827639,-0.528908968,0.4705024958,0.0161607563,0.0309730414,-0.2906499803,-0.2319734991,-0.084987849,-0.0696396381,-0.4132067263,-0.1349157095,0.571947217,0.4682939053,0.009673872,0.1252076924,-0.0798149556,0.0434404127,-0.0545615181,0.4297470152,-0.2400334179,-0.4153181016,0.0612085164,0.2196917832,0.0498348512,-0.3647942245,0.4084433615,-0.2142224759,0.0282030012,-0.1376089007,0.0682933703,0.2423673123,0.1012122259,0.2061181217,-0.1186033264,-0.0502854101,0.0660860389,-0.0480228886,0.0322182812,-0.0795101449,0.0072773066,0.1419922858,-0.0371353962,0.0833230168,-0.2190542817,0.391569078,0.0881103277,0.1482821256,-0.0699859858,-0.1448043585,0.0704681426,0.0794524699,-0.1418895274,0.3347402513,0.2606513798,0.1003755406,-0.2480081022,-0.0457481556,-0.0300719384,0.0121020572,0.1946262121,0.0748130158,-0.3734347224,-0.2700416148,0.1632227153,0.2970550358,-0.0165146906,-0.1167205051,-0.085226886,0.3182207644,-0.2520097196,-0.0121923853,0.200053066,0.0126722204,-0.2384376526,-0.0195359979,0.423286289,-0.3153413832,-0.2699567676,-0.2308864594,-0.3392125964]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1775","title":"Efficient ways to iterate the dataset","comments":"I was wrong, ```dataset[\"column\"]``` is fast.","body":"For a large dataset that does not fits the memory, how can I select only a subset of features from each example?\r\n\r\nIf I iterate over the dataset and then select the subset of features one by one, the resulted memory usage will be huge. Any ways to solve this?\r\n\r\nThanks","comment_length":6,"text":"Efficient ways to iterate the dataset\nFor a large dataset that does not fits the memory, how can I select only a subset of features from each example?\r\n\r\nIf I iterate over the dataset and then select the subset of features one by one, the resulted memory usage will be huge. Any ways to solve this?\r\n\r\nThanks\nI was wrong, ```dataset[\"column\"]``` is fast.","embeddings":[-0.2756459415,-0.468128562,-0.1285640895,0.3369154334,-0.0494370535,0.1071939021,-0.0851009116,0.3245358765,0.1788128614,0.1340755224,0.1120866016,0.1552499533,0.0256234277,0.2160820216,0.1022985354,-0.0614758879,-0.1157237142,0.2750163376,-0.1747321934,-0.0828541592,-0.2483522147,-0.3539268374,-0.3403426409,-0.3118047714,-0.0629606992,-0.3190019727,-0.03566489,-0.1483367085,-0.087984927,-0.2470878661,-0.0042620795,-0.0973600969,0.1467313915,0.1337923259,-0.0001115046,-0.2680115104,0.0112004662,-0.0473098643,-0.1700433046,0.1948329061,-0.5286120176,-0.190762803,-0.2098984718,-0.4612517953,0.1409865618,-0.2614294291,0.0430405512,-0.3923400342,-0.2784847319,0.0330316015,0.1861806512,0.2858508825,-0.4228091836,-0.1398807317,0.1154910624,0.2330436856,-0.1779843271,-0.2480240166,0.3662460148,-0.0191673413,0.1461574882,0.2179497182,-0.0582532063,0.0492375083,0.4238063395,-0.0895723552,-0.2033211291,-0.107842125,0.2946236432,0.3583475947,0.7589269876,-0.0007338207,-0.0226003118,-0.2744782865,-0.1172167733,-0.3964936733,-0.0817822292,0.1964516193,-0.131008476,0.175528869,-0.3859832287,-0.3547475934,-0.1166309416,0.077793844,0.2933899164,-0.1559051573,0.0513846651,0.028930245,0.4416371286,-0.266557157,0.3287567794,-0.2520357072,-0.0214381702,0.1252031177,-0.5075810552,-0.3412575126,-0.1267409027,-0.108331345,0.3563973308,0.0648758411,0.0619939119,0.0782034844,0.2330467254,0.0327255502,0.3822102547,-0.0917548016,-0.1510809213,0.1156526506,0.1408085525,-0.2089566737,-0.0909204483,-0.0973784998,-0.2143757194,0.0766778067,0.2558801174,-0.4136761129,-0.1271271259,-0.0838174447,-0.133028537,-0.1502370536,-0.1501519829,-0.0843517259,-0.0384561308,0.4479604661,0.1092413962,-0.157318905,-0.4641813338,-0.0441050231,-0.1826267987,-0.2609206736,-0.3089893758,0.3232057095,-0.0122926328,0.0158528704,0.0104168877,0.1980885416,0.1477535516,-0.126598537,-0.208325848,-0.159625724,0.098186411,-0.3591650724,0.2458127588,-0.0497209728,-0.1247465536,0.0136445891,-0.0749892369,0.3802159727,-0.3098506629,0.2297316194,0.0817483217,-0.5885893106,0.2191138268,0.0950704068,-0.0453339182,-0.0902551115,-0.3563770652,0.4660998881,0.0112344213,0.3650879562,-0.0762406886,0.1315849274,-0.3080327213,-0.1251855493,0.2918272614,0.4726677537,-0.2780974507,0.0316949189,0.1101208627,0.1702732742,0.3362140954,0.3550053239,-0.1930657029,0.2712133825,-0.1606461704,0.0966095328,0.2968051434,0.1896467209,-0.5020601749,0.4072839916,0.0822531506,0.0077122618,0.2725686431,0.4952426553,0.0412165448,0.0854645073,-0.1591809839,0.3467791378,-0.2650964856,0.2100361288,0.042354472,-0.3713879883,0.2071609646,-0.0146867037,0.2690726817,0.1971260756,-0.0643590018,0.1475664377,0.3248860538,-0.2832591832,-0.1355053186,0.18438676,0.4730816185,-0.0226837303,-0.0044273608,-0.4083937407,-0.3184022009,0.1793997288,0.4156848788,-0.5426943302,0.2938497365,-0.0836397633,0.02710733,0.0440009311,0.1039854214,0.1737812757,0.0382325612,-0.0397475362,0.2970722616,-0.441701442,-0.245862335,0.1432183534,-0.2575132549,-0.1216950268,-0.2598105371,-0.0338755175,0.5089336038,0.0125312237,0.1991798282,-0.0215717852,-0.0167062934,-0.2904073298,0.143721953,0.0831850991,0.3416111171,-0.033221215,0.1527047306,0.3046149611,0.0708261132,-0.2280883938,0.0962578356,-0.2558506131,0.2431627959,0.0168756228,-0.2830992043,0.127917856,-0.0008946864,0.5591956973,0.2271181345,-0.019137837,0.0197035391,0.0065320111,0.25179106,0.0558263175,0.2372860461,0.0990369245,0.1728402376,0.3244842887,-0.5848324299,0.0402641483,0.3017566204,0.0783885717,-0.034101665,0.0496365577,-0.1501238346,0.060169708,0.1351944953,-0.3370412588,0.0118509736,0.1079054326,0.0048922682,-0.1307452172,0.1182266697,0.1270122677,0.1688838303,0.2691067457,0.2818232775,-0.3338803649,-0.0004583986,-0.3070983589,-0.1175564602,0.0359553881,0.1611351967,0.1834457815,-0.0181777477,0.1056287661,-0.3602513671,-0.1492560208,0.1878812611,0.3253984749,0.240638271,-0.1253408641,-0.1094811186,0.4451908171,0.0109764412,-0.0465853103,-0.1755550057,0.4631407857,0.0289773419,-0.0874760747,0.0576041304,-0.0822373778,-0.0904426202,0.0104409279,0.3635648191,0.1286925226,0.3956053257,0.1068232134,-0.212830618,-0.4510279298,0.2162103504,0.0895716995,0.199814111,0.0688231364,-0.1811410785,0.1712427139,-0.323982954,0.1802396923,-0.1776747257,0.1928985268,-0.0719630569,0.1135234013,-0.0303480718,0.0078113852,0.4022991657,-0.401373446,-0.2021050006,-0.295422256,0.324678123,-0.0248926654,0.26991418,-0.0271444786,0.1140312627,-0.2368994206,-0.1127898246,0.0450431369,0.0508648194,-0.1520570219,0.3118434548,-0.0181564242,-0.1807760298,-0.0571576096,0.0424831249,-0.0204257146,0.3884344101,-0.4264629781,0.0336549431,-0.5360410213,-0.1155939102,-0.1871710569,0.0852321386,-0.0712207481,-0.1136996597,-0.0321681015,0.0399160832,-0.022014305,0.1874552965,0.161032185,0.0206706058,0.4287727177,0.240977034,0.2217412144,0.2823570073,0.3755410314,0.2936742604,0.2390128821,-0.1184868962,-0.183112219,0.061783161,-0.117343016,0.2019872069,-0.101970382,-0.2038857341,0.2292532325,-0.1289141178,-0.1371975243,0.0025055096,-0.0647156611,0.1282723844,-0.1488136351,0.25501284,-0.1494284719,0.23947528,0.1832550019,-0.0623036437,-0.3915431201,0.0174777918,0.0675699115,-0.342903614,0.2937984169,0.0123679787,-0.6645640731,-0.1237499863,-0.2904980183,0.0515156016,-0.1389027685,0.3438188136,0.1261108071,-0.2209736407,0.0524074845,0.2123644948,0.5732184052,0.2809284031,-0.1696149856,-0.1337233782,-0.6243345141,-0.0387505107,0.2437192053,-0.3127666712,0.4972465038,-0.1580832899,0.1312089264,-0.1504437923,-0.3318823278,0.3335514963,0.5679478049,-0.2044339627,-0.2807412148,0.1946450472,-0.197580725,-0.0479741953,0.2354754806,0.2642273903,0.0104654934,0.2850100398,-0.0511160195,-0.3827923834,-0.1125059649,0.4171054363,0.1450327784,0.0896375701,-0.1031606868,0.3642239869,0.0273763947,-0.0155208195,0.277202338,0.4541700184,0.2191331536,-0.4764958024,-0.0159060843,-0.1074760929,0.7704519033,0.2415875942,0.0136292251,0.2466727346,-0.1592480689,0.1932881474,-0.337408632,-0.1102342755,0.0447289683,-0.0958149806,-0.3036480546,-0.5038030744,0.4738487005,0.0508098081,-0.0182812102,0.3184924424,0.0004121611,0.0851076841,0.0087803183,0.1656087339,0.7456817627,-0.1109873354,-0.1556107551,-0.1109809428,0.3423111737,0.3189383447,-0.0853250176,0.3201678395,-0.1476583332,-0.7197934985,0.0537051,-0.1049712151,-0.0920710713,0.2528153658,0.1445315331,0.5074355006,-0.0505534597,0.0201833844,0.0432202853,-0.0872353166,-0.0829357132,-0.1900123656,0.0061628809,0.0929215029,0.1247329563,-0.0341638178,0.0985962823,0.1160214916,0.171291098,-0.0193932913,-0.2576333284,-0.0690780059,-0.3049025536,0.0906772837,-0.0516124628,-0.3482960165,0.094557859,-0.0231657829,-0.0101120183,0.2209986746,0.1112087518,-0.0094130868,0.1648990214,0.0817017034,-0.2010678947,0.0770613551,-0.0210167821,0.0536835976,-0.2400761098,0.1424462795,0.0059671206,0.0273190532,-0.1334502995,-0.0597768202,0.2170165479,0.0082862694,-0.3971117139,0.099902533,-0.1488437206,-0.1101091281,0.1137365624,0.1449168026,-0.0715828538,0.5055673122,0.1361587793,-0.1434043795,-0.1115418747,-0.1531309038,0.2454178929,0.0904511809,0.429649204,0.1805190295,-0.2179347575,-0.2847964764,0.2068274915,0.2729477882,0.1886984706,-0.0217532124,-0.3239086568,0.1305256337,-0.0498729348,0.1226890162,-0.0020387808,-0.1178313196,-0.2064535916,-0.0164852142,-0.1172146201,0.3544329107,0.2046742737,0.2993151546,-0.1831563413,-0.3605982661,-0.4341520965,0.4585973918,-0.2362541407,-0.0416199788,-0.3104836047,0.2460698336,0.1878909767,0.2965840995,-0.2493751198,-0.0436595604,0.1469246745,0.1566726416,0.0110842679,-0.2039675266,-0.0717095435,0.1129910946,0.0139092701,-0.1946008801,-0.0402144194,-0.2676291466,-0.3129578829,-0.4612752199,0.1980216652,-0.0067859553,0.1534802765,0.0324700139,0.193611905,-0.5211185813,-0.2211908996,0.3254092336,0.0888623223,0.0828876719,0.1504346132,0.2445865721,0.3968782723,-0.1459202319,-0.2359990776,-0.0220919345,0.1250583231,-0.2100081593,-0.1205854416,0.1070540547,0.1101795658,0.3190266788,-0.136783123,0.6434050202,0.1724091321,0.2079329789,0.2978217006,0.2212184668,-0.1962184161,-0.4696377516,0.0322953016,-0.1834685653,0.1439217627,0.5204063654,0.1565587372,0.2813718617,-0.1152098924,-0.2834275365,0.1628157943,-0.1229280084,0.0101267193,0.2447569668,0.1635740846,-0.4603108466,0.0426903553,0.6100426912,0.4475295246,0.6188147068,-0.0721175522,0.3401937187,0.1916792244,0.1722177118,-0.0782313123,-0.542773366,-0.1417859048,0.3737064898,-0.0233582091,0.0066320701,-0.1313119978,-0.4239557683,0.0858635381,-0.1192884296,-0.1890651584,-0.0112337349,0.0343203135,-0.1298368871,0.0183838457,-0.243926838,-0.1170877144,0.1620659828,0.0609368049,-0.4717728198,0.4436312616,0.0489567816,0.3452139795,-0.1933306307,0.5855688453,0.0978477895,0.2186326236,-0.423832655,0.0288958549,-0.1480866522,0.0593836866,-0.1044474766,0.1164273173,-0.1757689267,0.0770280436,-0.0338349864,0.0107002789,-0.216771692,-0.2417327762,-0.1653363705,-0.2066511065,-0.2030605674,0.1571859717,0.3236780465,-0.017148206,0.0413056463,0.1900364459,0.1390839964,-0.2242521793,-0.6899138093,-0.1792503148,0.1529851556,-0.3470785916,0.1300542504,0.0484861247,0.0491406396,-0.2291258126,0.4288969338,-0.4303441346,0.1594502479,0.4490227997,0.0879296139,-0.5533277392,0.4545185268,0.0262530446,0.0431634933,-0.3106143773,-0.1304688752,-0.0664952025,-0.1291138679,-0.4531156123,-0.1200546026,0.5035319328,0.4184785783,0.0219653845,0.1377376914,-0.0428548381,0.0957817361,-0.1144883409,0.4972445369,-0.2626009583,-0.424790144,0.0899235308,0.2996076941,0.1047627553,-0.3518970907,0.4243508279,-0.2447713166,0.0302406438,-0.1764573306,0.0689745843,0.2741497159,0.1521196067,0.2262912691,-0.0503752604,-0.0099153379,0.0438341945,0.0200413279,0.0071843755,-0.0935844779,0.0051971623,0.1885405928,0.0516263582,0.0307641476,-0.2972728014,0.363322109,0.1206555441,0.0664594397,-0.1425762177,-0.0754781961,0.0452264845,0.0807641968,-0.1512036026,0.4012962282,0.2304543406,0.1098022461,-0.254309684,-0.0863752514,-0.0496965945,0.0366004705,0.1940952092,0.0874390602,-0.4353280067,-0.2480521053,0.1641818583,0.3280241787,-0.0407472141,-0.1194627807,-0.111470066,0.3689447939,-0.2382484078,0.0187442824,0.1837414205,0.0969224796,-0.1700257808,-0.0237376876,0.4961647093,-0.3349610865,-0.2581802905,-0.1735898703,-0.3611704707]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1774","title":"is it possible to make slice to be more compatible like python list and numpy?","comments":"Hi ! Thanks for reporting.\r\nI am working on changes in the way data are sliced from arrow. I can probably fix your issue with the changes I'm doing.\r\nIf you have some code to reproduce the issue it would be nice so I can make sure that this case will be supported :)\r\nI'll make a PR in a few days ","body":"Hi,\r\nsee below error:\r\n```\r\nAssertionError: Requested slice [:10000000000000000] incompatible with 20 examples.\r\n```","comment_length":62,"text":"is it possible to make slice to be more compatible like python list and numpy?\nHi,\r\nsee below error:\r\n```\r\nAssertionError: Requested slice [:10000000000000000] incompatible with 20 examples.\r\n```\nHi ! Thanks for reporting.\r\nI am working on changes in the way data are sliced from arrow. I can probably fix your issue with the changes I'm doing.\r\nIf you have some code to reproduce the issue it would be nice so I can make sure that this case will be supported :)\r\nI'll make a PR in a few days ","embeddings":[0.0045192172,-0.06874533,-0.24153319,0.0569718927,0.4289458394,-0.3012439907,0.1267030835,0.3713265359,-0.1317171156,0.4324041903,-0.0149546824,0.7657305598,-0.0799334198,0.2857099771,-0.3763241768,-0.4880154133,-0.0210640207,0.1064821929,0.2514834702,0.1202106699,0.0552423708,0.0996666849,-0.3396672308,0.1170674637,-0.1888773739,-0.179093495,0.2490351796,-0.0215795748,-0.0056372839,-0.2691525221,-0.1261418462,-0.4162088037,0.0354158655,0.1907055676,-0.0001093078,0.1138291806,0.5290846229,0.042696517,-0.2076610029,-0.0507463217,0.2870479524,-0.487957716,0.2421301901,-0.2416446954,0.0444297679,0.0554977059,-0.3735236526,0.010754359,0.4677677155,-0.0264655594,0.2545258105,0.0062989909,0.3490653336,0.0459645353,0.0709797516,0.3102607727,-0.2072952092,-0.1372469664,0.5058743954,0.4029471874,0.2683990598,0.1032428667,-0.0779396668,0.040543437,-0.4270553887,0.0044288989,-0.0684800744,-0.3739693165,-0.1177275181,0.573800385,0.1728356332,0.1928459704,-0.2151761502,-0.0106723728,-0.3015369475,-0.2650514543,-0.1554951519,0.1012288854,-0.0972411335,0.1049901024,0.0690572336,0.0057898541,-0.1316060871,0.3030082881,0.2846271992,0.4896726608,-0.035527613,0.1082353294,0.1635646969,-0.3503347039,0.2268579453,0.1243717968,-0.0745611191,-0.0966618955,-0.1592775136,-0.0053485418,-0.5801075101,0.1688659042,0.1484343112,-0.0063061547,0.323872745,0.4233803749,-0.1502631009,0.1502199322,0.234686926,-0.0638783276,-0.0861096159,0.0048741279,0.1310970485,-0.1369228065,0.1043624878,-0.0289418213,0.2619714141,-0.3201720119,-0.2399541736,0.0034458842,-0.0258917715,-0.0790593252,-0.0733677894,-0.2654659152,-0.3296071589,-0.0969409198,0.3133118153,0.1284629852,-0.0212011002,0.096530579,-0.1779138446,0.2430861592,-0.1994260103,-0.1251432896,0.0082124947,0.0385231972,0.1689667255,-0.483379364,-0.0331377424,-0.2485744357,-0.0058779297,-0.0048196171,0.3314142227,0.0867033154,-0.1492839754,-0.0148875462,0.062819384,0.0198610872,0.0747855604,-0.2142450064,0.1739360392,0.067209661,-0.0024390426,0.4090491831,-0.5157783031,-0.1685249209,-0.147864148,0.1307652444,0.4457748234,0.2731143236,0.2207438052,-0.1539462507,0.0057326038,-0.3226754069,-0.0811057612,-0.1813516766,0.2451282144,-0.309543252,0.2209000736,-0.3931464255,-0.3490618169,0.0853376165,-0.4180666208,0.2275340259,0.2932496369,0.7705193758,-0.1812043488,-0.2930043042,-0.0307123773,0.3370466828,0.4182480574,0.1050323322,-0.2453847677,0.2037253678,0.0509392545,-0.1719106883,-0.3283394575,0.2673069537,0.15334481,-0.2094565779,-0.0209724996,0.1696024537,0.0461627729,-0.0351623185,-0.2500964701,-0.4086686373,0.2898615003,-0.0888149589,-0.0698139966,-0.1406375021,0.205048278,-0.3768809438,0.2609752119,-0.0919979513,0.0247562341,0.0110767772,0.4300157726,-0.4044119716,0.1026900262,-0.0949773118,0.1602244973,-0.1923354417,-0.1811702996,-0.0422615074,-0.1741902679,0.0141710266,0.1139547676,0.2723791003,0.2118366957,0.1726776212,0.14381814,-0.4210804403,0.3586548269,-0.1922202557,0.0335088223,-0.033112105,-0.1949820071,0.0527647324,-0.1643980145,0.3395628035,-0.3022041619,-0.3105169535,-0.1181154996,0.1678342372,0.23984465,-0.189041242,-0.2328326702,0.2748931646,0.101476267,-0.1820404828,-0.268628031,0.0532165542,0.161516875,-0.4648517668,0.233528465,0.2280954868,0.0137516083,-0.0216453671,0.0080416631,0.2652312815,-0.542475462,0.3630341291,0.0904720724,0.2836489975,0.1975425482,0.1266445369,0.0652361587,-0.0316096395,0.3338298798,0.1231466234,-0.1631327569,-0.1990456134,-0.1483731121,0.012621426,-0.003789986,-0.0754330158,0.2589483261,0.282145679,0.0397437662,0.1176440716,-0.0154709909,-0.4251084328,0.1360585392,0.2354196757,-0.0837447271,-0.1584738642,0.2301815897,-0.33596313,0.2332283109,0.452904433,0.028897604,-0.1256339699,0.0725977942,-0.0046532042,0.0529952198,0.2206350118,-0.296528697,0.0165035054,-0.0617351197,0.1354552358,-0.3268111944,0.2384214401,0.0465466157,-0.8412544727,-0.1915154159,-0.3362053633,0.0675206706,0.0370615907,-0.210921824,0.307638973,0.1330839097,0.1019628569,-0.0271361265,0.324450016,-0.1016389579,-0.1711580306,-0.3002125025,0.1642317623,0.1031269506,0.3488458097,-0.0832882375,0.1877930909,0.0227539055,0.0295580793,-0.4092308581,0.0001540455,-0.0346478075,0.556363225,0.1088159159,-0.0025661278,-0.1718282551,-0.3161525428,0.1827595979,-0.4125513434,-0.0562067293,0.3178083599,0.0368712023,-0.2372612953,0.010753314,0.0160018001,-0.6098569632,-0.3351960778,0.3220384717,0.2363469899,0.1585870832,-0.1259809881,0.1917729527,-0.1530326903,0.273591727,-0.074640967,-0.0318100564,0.0687904805,0.1247193143,-0.1566830128,-0.1934564412,-0.1348921657,-0.1770818084,-0.1167836115,-0.0765605867,0.0415161811,0.1929940581,-0.3029778302,-0.0505099446,-0.1226354912,0.1321017146,0.3967476487,0.1560400128,-0.1027520299,-0.0566878691,-0.1804210395,0.1074578986,0.0232549608,-0.1332282424,0.4811725616,0.3465818465,0.1279286891,0.368840009,0.2477864623,-0.1196170077,0.2847773731,0.2797137797,0.0390914492,-0.1042084172,-0.2776113749,0.2276642919,-0.0037570784,-0.2275750935,0.0122455861,0.0141047416,0.0418597832,0.0377201214,-0.2182125002,-0.4395800233,-0.1761209071,0.2133845985,0.0921330601,-0.0486647934,-0.1101056859,-0.1205698177,-0.2722116113,-0.2261462957,0.105827406,-0.1237039641,0.4476037621,-0.2446353883,-0.3568989933,-0.5452691317,0.044170633,0.5647826195,0.3321603537,0.2490008324,-0.0278467909,-0.1814689189,0.0598043986,0.059417367,0.6105985045,-0.2074550688,-0.3850255311,-0.1052481681,0.0917274281,-0.1567699611,0.2288676649,-0.1590213776,0.3620495498,0.2798972428,-0.1927829087,-0.084524028,0.0735425949,0.2832459509,0.2585216165,-0.0507099554,-0.0452483706,0.2243757248,0.2511662245,-0.1911723465,-0.0640020743,-0.0868367776,0.0999160334,-0.1130231842,-0.2351840883,-0.4449388683,-0.4831213951,-0.1476335824,-0.2586736679,0.3886404037,-0.2742809951,-0.0184731744,-0.3252069652,0.2851640582,0.2168972939,0.259950757,-0.1063733697,0.1940310448,-0.2688569725,0.0768449828,0.2581610084,0.0847632289,0.0405711569,0.1307294965,0.0482750572,0.156104669,0.1198096871,-0.5726235509,-0.0314530097,-0.1110070199,-0.2129098624,0.3697690666,0.4043018222,0.3005575836,-0.032587707,0.4440711439,-0.198614046,0.0789543092,0.1998645365,0.4521450996,0.9886347651,0.1013303027,-0.1530082226,-0.0159239974,-0.0597201958,0.4266940355,0.1008190513,0.166397348,-0.3995174468,-0.1502482742,-0.0319391862,-0.2249175608,-0.1185266003,0.0561291389,-0.3110295534,-0.0563965999,-0.5337849259,-0.0403964818,-0.2572265267,0.0648105592,0.0796505287,-0.1968759,-0.3503139913,0.1315542758,-0.1524081975,-0.3069136441,-0.018340487,-0.3018190563,-0.2096472383,-0.2561691403,-0.1240016297,0.0877362266,-0.3059896827,0.4181001186,0.2314329147,-0.4586016834,0.2111605704,0.0515432209,0.2225201577,0.0978386998,-0.1563038975,-0.0360193476,0.1318300962,-0.0493436046,0.0660979301,0.2049187124,0.1444841772,0.0021690144,-0.1125631481,0.1495084465,0.0704611093,-0.596678257,-0.2847770452,0.1402310282,0.1145455167,-0.2185788006,-0.2229589522,-0.2625519037,-0.402854383,-0.1619104743,0.146035254,0.2562430799,-0.0379576646,0.3883275092,0.542960465,0.1784557551,-0.040491771,-0.032743074,0.1441873014,0.0696646348,0.1716309637,0.3212611675,-0.2490287572,-0.1890686154,0.1801009029,-0.181672439,-0.0541249588,0.0256879479,-0.0807537064,0.0155019592,0.0634423941,0.3246165514,0.1424079239,0.4862164557,-0.4675274491,-0.3633792996,-0.2418723255,0.3637576699,0.0507297665,0.1370358914,-0.2009012103,-0.0810729861,-0.2109217048,0.2938998938,-0.2792305052,0.2067285031,0.0603095479,0.1292989552,-0.3833136261,-0.1093000695,-0.0650033355,0.0169279594,0.0719959661,0.2888394892,-0.2490579486,-0.2088914216,0.1214211658,0.095768638,0.0369802564,-0.5805706978,-0.3033237457,-0.1584573388,0.0030398099,-0.1595594436,-0.4319700003,-0.2644386888,-0.0336927697,-0.1819983274,0.4959421456,-0.0086945537,0.1645433158,0.2707722485,-0.17684111,0.3435405493,0.1144160032,-0.0168846231,0.3338825405,-0.0011246402,-0.0050521074,0.0724369064,0.202782169,0.0538878776,0.0777539536,-0.2313015014,-0.305388391,-0.0528203361,0.4625085294,0.2455825657,0.1443072557,0.1524994671,0.2244813144,0.1997745335,-0.054832723,-0.2919039428,0.2320183069,0.201173082,0.0303215049,0.3115656376,0.0192217659,0.1655310541,0.3544034362,-0.0071893614,0.1710476875,-0.6634289026,0.0769563764,0.3633344173,0.0860723481,-0.1353579611,-0.2520190179,-0.3793860674,0.2799762785,-0.016505558,-0.2989173532,0.1296316683,0.1696752608,-0.0758236274,0.3128222823,0.0548952259,-0.129425481,0.0404874161,-0.0090988716,0.0540237539,0.3673407733,0.3564632237,-0.0172517411,-0.1925057024,-0.226139769,0.1992288232,0.0159932375,-0.128341496,-0.164208129,-0.1144075245,0.0476979166,-0.3337884843,0.0281411093,0.2662841678,0.4338027239,0.2356037945,0.0121936118,-0.2688331604,0.0040649716,0.2407898754,0.1526182294,-0.2270524949,-0.0184957553,0.0678812414,-0.0465256907,-0.1567807794,-0.0096866116,0.58917588,0.2141220719,-0.4062794447,-0.1657733321,-0.1119140238,-0.1672048569,0.0304751527,0.1676184982,0.1938883066,0.1115805879,0.1253652424,0.2348316908,-0.1369703412,0.0620135553,0.1314628273,0.2517552078,-0.2453313321,0.3672529459,0.1214343458,-0.2679489553,-0.3027443886,-0.360424459,-0.1830506474,0.1741008162,-0.2602975965,-0.3829799891,-0.0166715421,0.2625739872,0.0917008221,-0.2118695527,0.3728059232,0.1517221034,0.155277729,-0.220861569,0.0982097983,0.1039775163,0.0218629204,-0.2331648469,0.1711068749,0.2647582293,-0.0954849795,0.3805432022,0.4274761975,0.2128989249,0.0471872203,0.0323063992,0.1043119505,-0.5206044912,0.1669880748,0.3593800366,-0.1979935914,0.0848283768,-0.0323782116,0.086491853,0.0843960047,0.038853623,-0.2193262428,0.0182626266,0.0208914354,0.175385043,0.1576270908,0.453590095,-0.0117714107,-0.168004781,-0.0772649199,0.1454778463,-0.0514577925,-0.0009285238,0.6229221821,0.2327602059,0.2851266861,-0.2586369216,-0.4273087978,-0.3670192361,-0.284576267,-0.3651994467,0.335776329,0.1669154614,0.1830334961,-0.0947477594,0.0311849266,0.138692826,-0.0341283865,0.1467219293,0.2278098762,-0.0313041285,-0.2833384573,0.2253668606,-0.0052036773,-0.0835770965,-0.0154669946,-0.0148501955,0.0944099203,-0.1263691187,-0.0704232156,-0.287409842,0.2488090247,-0.1169525087,0.3575585186,0.0694596097,0.3240207136,-0.1032735407,-0.0204295274,0.2113588601,0.3053734899,0.0561128147,-0.180769667,-0.0801790506]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1774","title":"is it possible to make slice to be more compatible like python list and numpy?","comments":"Good if you can take care at your side.\r\nHere is the [colab notebook](https:\/\/colab.research.google.com\/drive\/19c-abm87RTRYgW9G1D8ktfwRW95zDYBZ?usp=sharing)","body":"Hi,\r\nsee below error:\r\n```\r\nAssertionError: Requested slice [:10000000000000000] incompatible with 20 examples.\r\n```","comment_length":14,"text":"is it possible to make slice to be more compatible like python list and numpy?\nHi,\r\nsee below error:\r\n```\r\nAssertionError: Requested slice [:10000000000000000] incompatible with 20 examples.\r\n```\nGood if you can take care at your side.\r\nHere is the [colab notebook](https:\/\/colab.research.google.com\/drive\/19c-abm87RTRYgW9G1D8ktfwRW95zDYBZ?usp=sharing)","embeddings":[-0.1463743001,-0.1338042319,-0.2212999612,0.1068373248,0.3687288761,-0.1109129563,0.109775424,0.3639496267,-0.1976847947,0.5654138923,-0.1734055281,0.5952394605,-0.1914872378,0.4860168993,-0.309690088,-0.4220104814,-0.0067528505,0.0748366341,0.3716909587,0.2013477683,0.0220909435,0.1627366096,-0.418005079,-0.0126569895,-0.2699395716,-0.1749576777,0.0461755544,-0.0543190837,0.0242151339,-0.1029120386,-0.0421707444,-0.2449236065,0.1113489345,0.3501066864,-0.0001145465,-0.0247005783,0.339586705,0.010610004,-0.2888609767,-0.1177478805,0.2399788499,-0.4081625342,0.1575608701,-0.2614780962,-0.0630315915,0.4026615024,-0.1513607502,0.2490580231,0.3858807683,0.0084134517,0.2191824317,-0.0313964635,0.2811878324,0.055539865,0.0803095624,0.2945855558,-0.2348261178,0.0489668474,0.5267019868,0.3894968927,0.5000545979,-0.0610772111,-0.1321926564,-0.0364569202,-0.4367593229,-0.0362099186,-0.0115188695,-0.4340872765,-0.13329795,0.5013420582,-0.0326160602,0.2550181746,-0.1691958606,-0.0432528183,-0.2910272479,-0.3340623677,-0.1880221218,0.1574760526,-0.1480476707,0.1315200478,-0.0714794546,0.0132060731,-0.0102240369,0.2094409168,0.3894726336,0.6710613966,-0.039924264,-0.007296877,0.2856900692,-0.2653096914,0.2987681627,0.0942021832,-0.0016524378,-0.1533067226,-0.0776806325,-0.0756215975,-0.6117354631,0.442617774,0.0987176523,-0.360416621,0.2154320627,0.3568885028,-0.0852467939,0.2414955795,0.0811055824,-0.0763418004,-0.1495347321,-0.0185216181,0.1358646303,-0.0247627851,0.0174296033,0.1131078154,0.2455212176,-0.1601316631,-0.3452510834,0.0263187662,-0.1573827863,-0.0044661197,-0.0451889262,-0.3908392787,-0.4171558917,-0.0688735768,0.3215846717,0.0865321532,-0.2047594041,-0.1327313036,-0.2364898324,0.0616065189,-0.30271703,-0.0799520686,0.0563741401,0.2219841927,0.1285314262,-0.4259631038,-0.0836235806,-0.4230267406,-0.0125950733,-0.0065456801,0.4350409508,0.1751700342,-0.2162342668,-0.1704368144,-0.0893775374,0.0078044836,0.1843194664,-0.3660707176,0.1210878044,-0.1292431206,-0.0307376459,0.3225240409,-0.436124891,-0.1370525211,-0.0243028812,0.0858358666,0.3735511899,0.3180369735,-0.013290938,-0.0595596619,0.0642000735,-0.3805703819,-0.2227147669,-0.2955046296,-0.0694119632,-0.3025777042,0.2005238086,-0.3970080018,-0.2273574919,0.046645619,-0.2396271229,0.2638906538,0.3528092802,0.679526031,-0.0261853971,-0.2558016777,-0.1436401606,0.2174029797,0.3169987798,0.2011861056,-0.1319438815,0.1897441894,0.1700671911,-0.1082977057,-0.2624375522,0.2824304998,0.2040960789,-0.3303208947,0.2032666802,0.1576140523,-0.0534996577,0.0295291878,-0.2332039624,-0.476062268,0.3124851584,-0.1663630456,0.0187875535,0.0178823397,0.1365362108,-0.1638207585,0.2113603354,-0.1654578596,-0.2080750912,-0.1482063532,0.5208625793,-0.2875786722,0.00043412,-0.2785627544,0.2702053785,-0.1709327698,-0.0957396999,0.0559412763,0.0333458148,-0.0233912505,0.1229145154,0.1661356241,0.2767054439,0.054522343,0.0939501598,-0.3221396208,0.4544162154,-0.2085350901,0.0982955992,0.2128653228,-0.2247031033,-0.0074440162,-0.2440985441,0.3196657896,-0.4030084908,-0.3283036947,-0.0175588019,0.1813261211,0.3346835971,-0.2934066653,-0.1605484039,0.2109126449,0.1109216288,-0.363273412,-0.2697387934,0.1716515869,0.1714918762,-0.2443142384,0.1357454211,0.0862366036,-0.0246116631,0.0350704901,0.1413362473,0.3297730386,-0.583122313,0.2946592867,0.1858439147,0.1203041673,0.1297107786,0.1908291429,0.1535271853,0.0454845056,0.3824050725,0.2474129796,-0.387920022,-0.1192771792,-0.1163953543,0.030097628,-0.0735844374,-0.1121861711,0.2329208553,0.3125766218,0.1758454293,0.1065289453,-0.091101408,-0.4737367928,0.2227592468,0.1644078195,-0.1580740213,0.0582000054,0.1763469726,-0.2879417241,0.1933528632,0.4694195986,-0.1322252899,-0.2721380889,0.1136654317,0.0522442684,0.087894693,0.1069818959,-0.2936097085,0.0278367065,-0.0563769899,0.193500638,-0.1756854206,0.0417759605,0.0861763284,-0.751080215,-0.1703931242,-0.3278562725,0.0511887334,0.2992703617,-0.0046241605,0.2684949636,0.3187942505,0.1605608612,-0.0577719249,0.458362788,-0.0267480798,-0.0815882608,-0.2886000276,0.1724373698,0.1225785017,0.3139460683,-0.0457468294,-0.109230116,-0.2445595115,0.1312621385,-0.4563375115,0.0307505317,-0.0793864429,0.7301079035,0.1524665803,-0.0310618542,-0.280272305,-0.423491627,0.1138059869,-0.3169486821,0.0032560795,0.1429301202,0.0472586751,-0.2728820443,0.0277557299,0.1826931834,-0.6647526026,-0.2458188832,0.1978778243,0.2386526018,0.2435221672,-0.0160124227,0.0484660342,-0.1727282256,0.3658221364,-0.0096655013,-0.0585262775,-0.1023457274,0.0801431239,-0.1780867577,-0.1429953277,-0.1062780097,-0.2739801109,-0.2292208523,-0.1349261403,0.1151523665,0.0306721888,-0.2502322495,-0.1430070847,-0.0057319165,0.185367316,0.187475428,0.0366277397,-0.023884166,0.0902384967,-0.3642365932,-0.0207110811,0.1285443902,-0.0605802648,0.4990599155,0.2828123868,0.1165540367,0.2350545973,0.2284584194,-0.1006909534,0.2455296814,0.2318499684,0.0380302556,-0.0583351925,-0.4208358824,0.4038537741,0.0741876364,-0.4299725592,0.1540854871,-0.00984265,0.1599344909,-0.0111499736,-0.2722673118,-0.4967731833,-0.1554601341,0.2019174546,0.3252015114,-0.0063658734,-0.1080916971,-0.2127013057,-0.2294374555,-0.2309923917,0.0103609757,0.1993940026,0.3873068988,-0.2424591631,-0.1221472099,-0.6457659006,0.019528443,0.5191545486,0.4653425217,0.1261240244,-0.0654686317,-0.0529938638,0.0080380198,0.1542042494,0.7793080807,-0.2064718604,-0.4124643803,-0.0496790484,-0.1429058611,0.1279723644,0.1272413433,-0.1593178213,0.2851988077,0.4812598825,-0.1364738792,0.0087929657,0.279468298,0.5064293742,0.2913021743,-0.0872535408,-0.0823250264,0.1255923212,0.1622004211,-0.0978699848,-0.0827136487,-0.1546410024,0.0996441394,-0.0912678689,-0.1731019467,-0.4023498297,-0.5565629005,0.0368810073,-0.1568412632,0.1440297067,-0.0547214374,-0.098411411,-0.5561718941,0.1781596541,0.2501731217,-0.0186579768,-0.2040574998,0.2948747277,-0.1993360221,0.1569866389,0.3013067544,-0.0375666134,-0.0332259312,0.0205224715,-0.0159482192,0.2079016864,0.1993665248,-0.3967541158,-0.0649238676,0.0498089716,-0.4072391987,0.2141667604,0.2668101192,0.3590014577,0.0145494835,0.0433491841,-0.3059431016,0.02386971,0.1121305898,0.375315696,0.9474918246,-0.0537927896,-0.2969853282,-0.061782252,0.1691085547,0.5492321253,0.0564334206,0.3624795675,-0.223995775,-0.4255598485,-0.0971408561,-0.1897029579,-0.097704649,-0.1846909076,-0.4777362645,0.1157285273,-0.4101392031,-0.0285277162,-0.240632996,-0.075177744,0.1478288025,-0.0241769478,-0.580037415,0.1023633033,-0.183354944,-0.2246444523,-0.0958522409,-0.1340884715,-0.118856445,-0.115453206,-0.1938373148,0.0211600903,-0.3172887862,0.3400797248,0.1398476213,-0.3153769076,0.2293872535,0.1678365022,0.3695214987,0.2817682028,-0.0888051391,-0.1700528711,0.0351114348,-0.1588913947,-0.0216623042,0.2172594517,0.1610785425,0.0193350855,0.007804798,0.0867518634,0.2018770427,-0.5095809102,-0.307323277,0.1025562361,0.2089280635,-0.2341076881,-0.2316348404,-0.2674970925,-0.2764848173,-0.1658926904,0.08020138,0.1150451973,-0.1143171191,0.4435449243,0.4437083602,0.1331102699,-0.1199532449,-0.0542628281,0.1993964463,0.0484062508,0.0239368118,0.3131355643,-0.24142012,-0.0859249234,0.0969798639,-0.5355571508,0.3427474499,0.1518926173,-0.0109916255,0.0299484693,0.033120282,0.5520337224,0.1676724106,0.4237661958,-0.3793120682,-0.4066291749,-0.1593941897,0.2191098928,0.095309332,0.1028448567,0.0464040413,0.0889924988,-0.2049644589,0.2996847034,-0.2251147777,0.0946149081,-0.1162916049,0.1517087519,-0.263234973,-0.0638693795,-0.0827467293,0.0778135285,0.0413728729,0.356713593,-0.1973663419,-0.1920624226,0.0588949732,0.0954183713,-0.0078372499,-0.3989557624,-0.3400700688,-0.0039469106,0.021713553,-0.2590763271,-0.3779963553,-0.1954835802,0.0731319189,-0.2213238776,0.4910060167,-0.0648906827,0.2258819938,0.2642232478,-0.1982290894,0.3222678602,0.2569621205,0.0224477574,0.289021492,0.0142172556,0.0128868325,0.0256441087,0.1698478907,0.0425182693,0.009162005,-0.0704619959,-0.2680034935,0.0639675632,0.352152586,0.1886855811,0.2749020457,0.1645492613,0.3072308898,0.0790228099,-0.041484382,-0.193380326,0.4618855119,0.1535438895,0.0616551451,0.5074474812,-0.0691564009,0.2653869092,0.2966341376,-0.1180020347,0.2931157947,-0.4557472169,0.1440646946,0.3809446096,0.1706268936,0.0753597915,-0.3124566972,-0.3792649806,0.2548114657,-0.0632116646,-0.3033097386,0.2088728547,-0.0773473084,0.0026387055,0.4684697986,0.0708988607,0.055583261,0.2465233803,0.100368537,0.0494825244,0.3290079534,0.2209452987,-0.2734415531,-0.0545464084,-0.1708994508,0.2969811857,0.0030361661,-0.0955383554,0.1110130101,-0.0981729254,-0.0052091549,-0.2267052382,0.1651898921,0.2053192407,0.3363941014,0.1473001987,0.0474562086,-0.0272284057,-0.0933490545,0.2225243747,0.0823052153,-0.2312779129,-0.0641931668,-0.1484693736,0.0388695598,-0.2890441716,0.0570908338,0.6059703827,0.2163904011,-0.3049992621,-0.2829700708,-0.15239802,-0.0483227894,-0.0835250244,0.0688271374,0.1021189392,0.3400026858,0.0000839449,0.2096627355,-0.0860867128,0.1552154869,0.0660764128,0.0911285952,-0.2876434028,0.1895565838,0.2176987231,-0.1329372227,-0.3089644015,-0.2442031652,-0.1386733502,0.2537459433,-0.4294860065,-0.4314926267,-0.0080328183,0.4642736614,0.0790166929,-0.1345050335,0.2964865267,0.1871024966,0.056385275,-0.2996495366,0.1304316074,0.23662889,-0.0071852542,-0.1420755982,0.166452527,0.1449316591,-0.2256847918,0.1908170581,0.4170701802,0.0068298206,0.0653894022,-0.0003176824,0.0811484754,-0.4300746918,0.0797775313,0.192116648,-0.333814919,0.1080795825,0.0755399838,0.1054851785,-0.0621442609,-0.034833774,-0.2734720111,0.0737167671,-0.0844757333,0.0121249156,-0.0370758213,0.490521729,0.1257591099,-0.2869726717,-0.049620185,0.2761175632,0.1011201143,0.0470702685,0.5464611053,0.2503136694,0.1432879865,-0.3280180991,-0.2789608538,-0.3519434035,-0.1168932095,-0.1701781899,0.2994925976,0.216628015,0.0836088061,-0.1176443994,0.0804130659,0.2080838084,-0.0472951122,0.0699110329,0.2768038511,-0.0011916182,-0.133992821,0.2390550375,-0.0087671578,-0.0861539021,-0.226119861,-0.0728408471,-0.0425347723,-0.1131279469,-0.015629679,-0.320743829,0.2360308021,-0.0923968032,0.2989393175,0.3014296591,0.3135289252,-0.1116926596,-0.0261242259,-0.0051412848,0.229879424,-0.0028920362,-0.1710802019,-0.0605657734]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1773","title":"bug in loading datasets ","comments":"Looks like an issue with your csv file. Did you use the right delimiter ?\r\nApparently at line 37 the CSV reader from pandas reads 2 fields instead of 1.","body":"Hi,\r\nI need to load a dataset, I use these commands:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files={'train': 'sick\/train.csv',\r\n 'test': 'sick\/test.csv',\r\n 'validation': 'sick\/validation.csv'})\r\nprint(dataset['validation'])\r\n```\r\nthe dataset in sick\/train.csv are simple csv files representing the data. I am getting this error, do you have an idea how I can solve this? thank you @lhoestq \r\n\r\n \r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-61468fc71a743ec1 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nTraceback (most recent call last):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 485, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 604, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 959, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/tqdm-4.49.0-py3.7.egg\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/julia\/cache_home_2\/modules\/datasets_modules\/datasets\/csv\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2\/csv.py\", line 129, in _generate_tables\r\n for batch_idx, df in enumerate(csv_file_reader):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1029, in __next__\r\n return self.get_chunk()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1079, in get_chunk\r\n return self.read(nrows=size)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1052, in read\r\n index, columns, col_dict = self._engine.read(nrows)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 2056, in read\r\n data = self._reader.read(nrows)\r\n File \"pandas\/_libs\/parsers.pyx\", line 756, in pandas._libs.parsers.TextReader.read\r\n File \"pandas\/_libs\/parsers.pyx\", line 783, in pandas._libs.parsers.TextReader._read_low_memory\r\n File \"pandas\/_libs\/parsers.pyx\", line 827, in pandas._libs.parsers.TextReader._read_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 814, in pandas._libs.parsers.TextReader._tokenize_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 1951, in pandas._libs.parsers.raise_parser_error\r\npandas.errors.ParserError: Error tokenizing data. C error: Expected 1 fields in line 37, saw 2\r\n\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"write_sick.py\", line 19, in <module>\r\n 'validation': 'sick\/validation.csv'})\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 534, in download_and_prepare\r\n self._save_info()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/contextlib.py\", line 130, in __exit__\r\n self.gen.throw(type, value, traceback)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 491, in incomplete_dir\r\n shutil.rmtree(tmp_dir)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 498, in rmtree\r\n onerror(os.rmdir, path, sys.exc_info())\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 496, in rmtree\r\n os.rmdir(path)\r\nOSError: [Errno 39] Directory not empty: '\/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2.incomplete'\r\n```\r\n\r\n","comment_length":30,"text":"bug in loading datasets \nHi,\r\nI need to load a dataset, I use these commands:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files={'train': 'sick\/train.csv',\r\n 'test': 'sick\/test.csv',\r\n 'validation': 'sick\/validation.csv'})\r\nprint(dataset['validation'])\r\n```\r\nthe dataset in sick\/train.csv are simple csv files representing the data. I am getting this error, do you have an idea how I can solve this? thank you @lhoestq \r\n\r\n \r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-61468fc71a743ec1 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nTraceback (most recent call last):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 485, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 604, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 959, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/tqdm-4.49.0-py3.7.egg\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/julia\/cache_home_2\/modules\/datasets_modules\/datasets\/csv\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2\/csv.py\", line 129, in _generate_tables\r\n for batch_idx, df in enumerate(csv_file_reader):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1029, in __next__\r\n return self.get_chunk()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1079, in get_chunk\r\n return self.read(nrows=size)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1052, in read\r\n index, columns, col_dict = self._engine.read(nrows)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 2056, in read\r\n data = self._reader.read(nrows)\r\n File \"pandas\/_libs\/parsers.pyx\", line 756, in pandas._libs.parsers.TextReader.read\r\n File \"pandas\/_libs\/parsers.pyx\", line 783, in pandas._libs.parsers.TextReader._read_low_memory\r\n File \"pandas\/_libs\/parsers.pyx\", line 827, in pandas._libs.parsers.TextReader._read_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 814, in pandas._libs.parsers.TextReader._tokenize_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 1951, in pandas._libs.parsers.raise_parser_error\r\npandas.errors.ParserError: Error tokenizing data. C error: Expected 1 fields in line 37, saw 2\r\n\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"write_sick.py\", line 19, in <module>\r\n 'validation': 'sick\/validation.csv'})\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 534, in download_and_prepare\r\n self._save_info()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/contextlib.py\", line 130, in __exit__\r\n self.gen.throw(type, value, traceback)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 491, in incomplete_dir\r\n shutil.rmtree(tmp_dir)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 498, in rmtree\r\n onerror(os.rmdir, path, sys.exc_info())\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 496, in rmtree\r\n os.rmdir(path)\r\nOSError: [Errno 39] Directory not empty: '\/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2.incomplete'\r\n```\r\n\r\n\nLooks like an issue with your csv file. Did you use the right delimiter ?\r\nApparently at line 37 the CSV reader from pandas reads 2 fields instead of 1.","embeddings":[-0.2842670083,-0.2714768052,-0.1463154405,0.4608483613,0.3019686937,0.2408371568,0.0834566206,0.5121120214,0.0764491484,0.0585810058,-0.0004540544,-0.2665965557,-0.10810256,0.2536355257,-0.0337027758,0.0352051184,-0.0331644155,0.1680267006,-0.0946208239,0.019386515,-0.2684040964,0.2590358853,-0.0957386568,0.0894601047,-0.0035116654,-0.1219344884,0.0992638245,0.2603849769,-0.2975808978,-0.3687266707,0.3527499437,-0.1765222102,0.3671296537,0.2056512088,-0.0001100329,0.109722428,0.4414513707,-0.222237736,-0.2273045331,-0.2782835364,-0.3186444938,-0.5255689621,-0.0424169973,-0.2519200146,-0.0720899403,-0.1986101866,-0.1186174303,-0.358505249,0.1968753785,0.4839046001,0.3277643323,-0.2024365067,-0.0772970542,-0.2182596326,0.1736675203,0.0825364217,0.0177823696,0.0985001624,0.0383216366,0.1259890795,0.5457251668,0.22481516,-0.0871027187,0.062276613,0.1328184605,-0.0650291443,0.2201796621,-0.3268524706,0.1852118075,0.1893909127,0.8423666954,-0.1272012442,-0.1620606333,-0.14171803,0.1230335012,-0.309578687,0.0035668784,0.3891044259,-0.1592620909,0.1557286382,-0.1961616576,0.022634821,-0.056327071,0.1439458579,0.1991666853,-0.1276440918,-0.256634742,0.2265536189,-0.066450499,-0.0968799889,0.0019838086,-0.0634838715,-0.1318117976,0.220077008,-0.3061932921,-0.0059892815,-0.0865267888,-0.0474983081,0.1305490732,0.0191281587,0.2057310492,-0.3622546494,0.1416555792,0.1029283404,0.3085215986,0.3533758223,-0.0966351852,0.0186193027,0.0468764491,0.1750154495,0.0544107519,-0.2243265808,-0.1458094567,-0.4689761698,0.2325953245,0.2488867044,0.3317735791,-0.1506784409,-0.2495945245,0.1341956556,-0.0551047288,0.2364847958,0.1205831841,0.5153110623,-0.307118088,0.1684069484,-0.062956959,0.02560132,-0.0268063359,-0.2216439694,-0.295609206,0.0240087099,-0.1846616417,-0.1395942122,0.2067635804,-0.2110509127,0.2445815951,0.079535611,0.0408378132,-0.3071087897,0.0538251624,-0.3537349403,-0.023229761,0.4643990993,0.0900522098,0.1324354857,0.2597099245,-0.0839310288,-0.218993932,0.1771682501,-0.3590294123,-0.5951446295,-0.0110709053,0.26149863,-0.1140651032,0.1193498001,-0.4246470928,0.0700019151,0.160954386,-0.0954999104,-0.0518162772,-0.1787497252,-0.2614167333,-0.3476524353,0.1499703825,0.3107527196,-0.5323678851,0.1687188745,-0.3266535997,-0.2732284069,0.297688961,0.2120352238,-0.1209672764,0.3310961425,-0.2531271875,0.0592158996,0.3152398765,-0.3005893826,-0.3999977708,0.3908524215,0.016317809,0.1309871674,0.0034007411,0.0098110065,0.3307875693,0.0078690965,0.1102341935,0.212696299,-0.1486558169,-0.0314865559,-0.262057066,-0.1641774923,-0.0509140231,0.156482935,0.1396570951,0.1109460518,0.1691527665,-0.3006534874,-0.0518827923,-0.065568164,0.2696269155,0.3614275157,0.2116627544,0.1624421924,0.1707906276,0.0812941045,-0.535584569,0.3256810606,0.0115941083,-0.1998816878,-0.0919436142,-0.1241044924,-0.5685514212,-0.0164452605,-0.1286688447,-0.00269557,0.1094136685,-0.105446212,-0.0243037492,-0.0921541229,-0.0302179251,0.5924558043,-0.1262022704,-0.1524438113,-0.1371084154,0.2782709002,0.0124592735,0.055331625,0.0212132409,0.0994580835,0.1813473701,-0.1014740244,-0.3069878519,0.3099374771,0.0669195652,-0.0598004498,-0.0391001031,-0.0881186426,-0.1291055232,-0.1401408911,-0.0137780439,0.157358557,0.2477229089,-0.0469300412,0.1126644611,0.1488340795,-0.2980319262,0.0036978342,-0.1184609905,0.0627791062,0.5067969561,0.0400967784,0.0703584403,-0.3693780303,0.2232140005,-0.1662302166,0.3108948171,-0.0223938841,-0.2173395753,-0.3514190018,0.1450280249,0.0782161579,-0.0974682942,0.4313172102,-0.0909135789,-0.1065372974,-0.0840541273,0.3009824455,0.5254251361,0.0834561214,-0.1675480157,0.2054767311,0.2845359147,-0.0061824503,0.2890945375,-0.1024241671,0.3282971978,0.2726708055,-0.0958034098,-0.1561962217,-0.1069378108,-0.1512015164,0.1809896976,0.480276674,-0.296161443,0.0076901154,-0.0392973498,-0.1293672323,-0.0263259336,-0.125998646,0.0485745929,-0.0580259934,-0.1010541618,0.1505691707,0.0410420112,-0.1340404749,-0.1771068275,0.2878244519,0.1359101534,-0.2367094755,-0.1092019901,-0.1559366435,-0.2683488727,0.1195478737,0.0350314416,0.01967692,0.2585455179,-0.1997407377,0.0914035216,-0.1030231938,-0.049859073,-0.0447158515,-0.001954356,0.3377043307,0.1962086111,0.2400903255,0.0899587795,0.0516655073,0.2236879468,-0.0502688475,-0.0530633293,0.3498876691,-0.1486122906,-0.0987843126,-0.0612148754,-0.4378128946,-0.5983160138,-0.3591504395,0.0012708708,0.0437375307,0.0452357754,0.2158264667,0.2280218601,-0.1275926828,0.3982062936,-0.0774030164,-0.3031054139,-0.5205017328,0.2376762778,-0.3503072262,-0.1679317504,0.0081502553,-0.0291733481,0.1761376858,0.0675562024,-0.5610889792,0.3567280471,-0.23838,0.0649895445,-0.2308612466,-0.3174830675,0.220419839,-0.1381314099,-0.0619107075,-0.248017773,-0.1541718543,-0.0304518882,0.0611269623,0.1230327561,-0.0677224621,0.5385031104,0.0502661429,0.5127761364,0.3299145997,-0.0016039216,0.4850444198,0.0683038458,0.3586151302,-0.2322539389,-0.3839135766,-0.0448417515,-0.0905613229,-0.2724584341,0.1639659405,-0.0662225708,-0.1001958027,-0.1565478146,-0.0477787927,-0.3332802057,0.0539398119,0.1047426835,-0.1628108025,0.3125785887,0.0887057185,-0.1207629219,-0.3875431418,-0.1475298703,-0.2114031762,0.3583055437,-0.0312432665,0.2129054815,-0.3576206565,0.0854414776,-0.3756477237,0.4200848341,0.1093478948,0.4523644447,-0.1290504336,0.2127711326,0.0378855318,0.0534792878,0.7103981376,-0.4645594358,0.0358819552,0.1659949124,0.0275682602,-0.3042504787,-0.285187155,-0.0084803542,0.3178522289,0.31413275,0.5803300142,-0.0361209065,0.0007447536,-0.0464967526,0.1986033916,-0.2785790265,-0.3796182871,-0.3453670144,-0.2140879482,-0.241249308,-0.0628155991,0.1035616323,0.4664235115,-0.0524130017,-0.11790663,-0.1577501893,0.0153487744,0.1427042633,0.0880366713,0.3629487157,-0.014422928,0.364948988,0.0744690523,0.0885612592,0.1970508397,0.6999536157,0.1273981035,-0.5392558575,-0.126157403,0.0696689337,0.0954375714,-0.0811017454,-0.2950466573,-0.1615704447,-0.0361088924,0.2540148199,-0.1686358005,0.2842821777,0.2741622627,-0.0936936513,-0.5076526403,-0.4867087603,0.2544363439,0.0747539699,-0.0489867106,0.2786591649,-0.213526547,-0.2284800708,0.0430670418,0.0376372449,0.6544899344,-0.0589199401,-0.1074389592,0.1674174219,-0.2733144164,0.2432312965,-0.4013942778,0.0802245364,-0.3252433836,-0.4311294854,-0.0189837143,-0.2221660614,0.0923741683,0.1845080107,-0.1360657662,0.3313300312,-0.2476435453,0.3380993307,-0.0432728752,0.0570543483,0.0132916989,-0.1616701931,-0.1127723157,0.2004107237,-0.2213427126,0.2725203037,-0.2305128127,-0.0070641725,0.0349441878,-0.2116746902,-0.3580960631,0.0895582661,-0.0012524695,0.332059145,-0.1337656528,-0.2141943872,0.0264652427,0.3820576966,0.2773966491,0.274762094,-0.1680817753,0.1762898564,0.0078900587,-0.0227795355,-0.0677925348,0.1592404395,0.057432916,-0.0700750053,-0.2695395648,0.1729298085,-0.1512601525,-0.1990173906,0.0742826536,-0.055200696,-0.0439750627,-0.3788851798,-0.4944426119,-0.2080943584,0.1806277335,-0.0699635819,0.1461118907,-0.0117006134,-0.0562958606,0.2779733837,0.0098770913,-0.4934690893,-0.0934577808,0.2358596474,0.1504454613,0.0423076935,0.577039063,-0.0227455944,-0.1267643571,-0.2060649842,0.1569144279,0.1524967104,-0.4153022766,0.2984954417,0.0065650269,0.2282925844,-0.0564338081,0.1067050397,0.2867481112,-0.0137914447,0.2696353793,-0.6932899952,-0.2438646704,0.0520335473,-0.2252726853,0.1804030538,0.1651545316,0.0428266078,-0.0882505253,-0.0495004915,-0.3103482425,0.1951560676,-0.2436397225,-0.0127718588,0.1171145812,0.0246869922,0.1024996936,-0.0507664867,0.1229806319,0.056868609,-0.2471610159,-0.2958017588,-0.1120594591,0.129828617,-0.0144994557,-0.1231976897,0.0595683232,-0.2221883237,-0.263602823,-0.0755089,0.0633006692,0.2292277962,0.0596436448,0.150957793,-0.1756052971,0.0745841935,-0.300188601,0.320567131,0.067909196,0.2082365453,-0.0075415117,0.2315397412,0.1362181306,-0.0870125145,-0.4957450926,-0.1165431663,0.1486171037,0.0311064757,0.3345166743,-0.2841587663,0.4677669406,0.185994938,0.177285254,0.2053661793,-0.1843367815,0.136656478,0.2792657316,0.2459561676,-0.341280371,-0.0360048562,-0.0271832012,0.2645361125,-0.151609987,0.0192744341,0.3053024113,-0.0335899964,-0.0239056367,0.0817202777,0.3608446419,0.1403882802,0.1985132247,0.4787999988,0.0550891384,0.1043441519,0.1473935395,0.1529682577,0.1486538649,0.8432779908,-0.0870846435,0.047505524,0.2190115452,0.1941461414,-0.2537889481,-0.7259444594,0.2099428028,0.2318012267,-0.1292050779,0.0636873543,0.0515897572,0.0154766347,-0.056893263,0.0287104771,-0.3996512592,-0.112338677,-0.1974150389,-0.0439040847,-0.1145654693,-0.0935092345,-0.095047079,0.0906768739,-0.1025839821,-0.0517633855,-0.0532695316,0.0804565623,-0.287350297,-0.3159598708,0.1776091307,0.4647827148,-0.0634765103,-0.3196803629,0.2313063741,0.1958958656,0.0570135266,0.1599377394,0.1180126965,0.5215353966,0.5515784621,-0.0306069646,0.106946446,0.1577190608,-0.1133501381,-0.0408964641,0.1973387003,0.3230094612,0.3432394862,0.2013502866,0.2322909236,-0.1237463504,0.3629772961,0.1353282332,0.2386198044,-0.158225745,0.0627807826,-0.1321887374,0.0853877291,-0.2960370183,0.1692750156,-0.2322701067,0.0723846555,0.537407279,0.173791036,0.1843811572,0.0281956922,0.0819050521,-0.1794721484,0.3207796812,0.1872993857,-0.0046041477,-0.4473469555,-0.1420500726,-0.4908000827,0.15742594,-0.3643202782,-0.1044761837,0.1451110989,0.1176997945,-0.0711466447,0.0115686674,0.0331624448,-0.0390997864,-0.1048875004,0.2687349617,-0.1599753946,-0.3986561298,-0.0576156043,0.1079313755,0.1156313792,-0.5605536699,-0.015596332,-0.1182360575,0.1090889573,-0.2412688285,-0.155986473,0.071618177,-0.0598636344,0.4797489047,0.0273879673,0.4527481496,-0.0294720344,0.1498543769,-0.3443867862,-0.3039680719,-0.0468843468,0.3253214657,0.1939056218,0.2710701227,-0.2828695178,0.1123746485,-0.1649846733,-0.002180147,0.0596382879,0.1852112263,-0.0965949371,-0.0817463994,-0.1364638507,0.1857875586,0.1331339628,0.2467502058,-0.1912277788,0.072395578,-0.0053284904,-0.3337082267,0.5975276828,-0.0790108368,-0.2732951343,0.0007333648,0.0964822322,0.2757702172,-0.1867084354,-0.2042184472,0.1991377473,0.4787468016,-0.1332235187,-0.0488588773,0.171931237,0.0861932039,0.111177519,-0.0059969663,0.1856501997,0.0163134839,-0.1944563091,0.1022795588,-0.3659286499]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1773","title":"bug in loading datasets ","comments":"Note that you can pass any argument you would pass to `pandas.read_csv` as kwargs to `load_dataset`. For example you can do\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files=data_files, sep=\"\\t\")\r\n```\r\n\r\nfor example to use a tab separator.\r\n\r\nYou can see the full list of arguments here: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/packaged_modules\/csv\/csv.py\r\n\r\n(I've not found the list in the documentation though, we definitely must add them !)","body":"Hi,\r\nI need to load a dataset, I use these commands:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files={'train': 'sick\/train.csv',\r\n 'test': 'sick\/test.csv',\r\n 'validation': 'sick\/validation.csv'})\r\nprint(dataset['validation'])\r\n```\r\nthe dataset in sick\/train.csv are simple csv files representing the data. I am getting this error, do you have an idea how I can solve this? thank you @lhoestq \r\n\r\n \r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-61468fc71a743ec1 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nTraceback (most recent call last):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 485, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 604, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 959, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/tqdm-4.49.0-py3.7.egg\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/julia\/cache_home_2\/modules\/datasets_modules\/datasets\/csv\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2\/csv.py\", line 129, in _generate_tables\r\n for batch_idx, df in enumerate(csv_file_reader):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1029, in __next__\r\n return self.get_chunk()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1079, in get_chunk\r\n return self.read(nrows=size)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1052, in read\r\n index, columns, col_dict = self._engine.read(nrows)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 2056, in read\r\n data = self._reader.read(nrows)\r\n File \"pandas\/_libs\/parsers.pyx\", line 756, in pandas._libs.parsers.TextReader.read\r\n File \"pandas\/_libs\/parsers.pyx\", line 783, in pandas._libs.parsers.TextReader._read_low_memory\r\n File \"pandas\/_libs\/parsers.pyx\", line 827, in pandas._libs.parsers.TextReader._read_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 814, in pandas._libs.parsers.TextReader._tokenize_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 1951, in pandas._libs.parsers.raise_parser_error\r\npandas.errors.ParserError: Error tokenizing data. C error: Expected 1 fields in line 37, saw 2\r\n\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"write_sick.py\", line 19, in <module>\r\n 'validation': 'sick\/validation.csv'})\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 534, in download_and_prepare\r\n self._save_info()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/contextlib.py\", line 130, in __exit__\r\n self.gen.throw(type, value, traceback)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 491, in incomplete_dir\r\n shutil.rmtree(tmp_dir)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 498, in rmtree\r\n onerror(os.rmdir, path, sys.exc_info())\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 496, in rmtree\r\n os.rmdir(path)\r\nOSError: [Errno 39] Directory not empty: '\/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2.incomplete'\r\n```\r\n\r\n","comment_length":64,"text":"bug in loading datasets \nHi,\r\nI need to load a dataset, I use these commands:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files={'train': 'sick\/train.csv',\r\n 'test': 'sick\/test.csv',\r\n 'validation': 'sick\/validation.csv'})\r\nprint(dataset['validation'])\r\n```\r\nthe dataset in sick\/train.csv are simple csv files representing the data. I am getting this error, do you have an idea how I can solve this? thank you @lhoestq \r\n\r\n \r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-61468fc71a743ec1 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nTraceback (most recent call last):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 485, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 604, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 959, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/tqdm-4.49.0-py3.7.egg\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/julia\/cache_home_2\/modules\/datasets_modules\/datasets\/csv\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2\/csv.py\", line 129, in _generate_tables\r\n for batch_idx, df in enumerate(csv_file_reader):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1029, in __next__\r\n return self.get_chunk()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1079, in get_chunk\r\n return self.read(nrows=size)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1052, in read\r\n index, columns, col_dict = self._engine.read(nrows)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 2056, in read\r\n data = self._reader.read(nrows)\r\n File \"pandas\/_libs\/parsers.pyx\", line 756, in pandas._libs.parsers.TextReader.read\r\n File \"pandas\/_libs\/parsers.pyx\", line 783, in pandas._libs.parsers.TextReader._read_low_memory\r\n File \"pandas\/_libs\/parsers.pyx\", line 827, in pandas._libs.parsers.TextReader._read_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 814, in pandas._libs.parsers.TextReader._tokenize_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 1951, in pandas._libs.parsers.raise_parser_error\r\npandas.errors.ParserError: Error tokenizing data. C error: Expected 1 fields in line 37, saw 2\r\n\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"write_sick.py\", line 19, in <module>\r\n 'validation': 'sick\/validation.csv'})\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 534, in download_and_prepare\r\n self._save_info()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/contextlib.py\", line 130, in __exit__\r\n self.gen.throw(type, value, traceback)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 491, in incomplete_dir\r\n shutil.rmtree(tmp_dir)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 498, in rmtree\r\n onerror(os.rmdir, path, sys.exc_info())\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 496, in rmtree\r\n os.rmdir(path)\r\nOSError: [Errno 39] Directory not empty: '\/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2.incomplete'\r\n```\r\n\r\n\nNote that you can pass any argument you would pass to `pandas.read_csv` as kwargs to `load_dataset`. For example you can do\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files=data_files, sep=\"\\t\")\r\n```\r\n\r\nfor example to use a tab separator.\r\n\r\nYou can see the full list of arguments here: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/packaged_modules\/csv\/csv.py\r\n\r\n(I've not found the list in the documentation though, we definitely must add them !)","embeddings":[-0.2842670083,-0.2714768052,-0.1463154405,0.4608483613,0.3019686937,0.2408371568,0.0834566206,0.5121120214,0.0764491484,0.0585810058,-0.0004540544,-0.2665965557,-0.10810256,0.2536355257,-0.0337027758,0.0352051184,-0.0331644155,0.1680267006,-0.0946208239,0.019386515,-0.2684040964,0.2590358853,-0.0957386568,0.0894601047,-0.0035116654,-0.1219344884,0.0992638245,0.2603849769,-0.2975808978,-0.3687266707,0.3527499437,-0.1765222102,0.3671296537,0.2056512088,-0.0001100329,0.109722428,0.4414513707,-0.222237736,-0.2273045331,-0.2782835364,-0.3186444938,-0.5255689621,-0.0424169973,-0.2519200146,-0.0720899403,-0.1986101866,-0.1186174303,-0.358505249,0.1968753785,0.4839046001,0.3277643323,-0.2024365067,-0.0772970542,-0.2182596326,0.1736675203,0.0825364217,0.0177823696,0.0985001624,0.0383216366,0.1259890795,0.5457251668,0.22481516,-0.0871027187,0.062276613,0.1328184605,-0.0650291443,0.2201796621,-0.3268524706,0.1852118075,0.1893909127,0.8423666954,-0.1272012442,-0.1620606333,-0.14171803,0.1230335012,-0.309578687,0.0035668784,0.3891044259,-0.1592620909,0.1557286382,-0.1961616576,0.022634821,-0.056327071,0.1439458579,0.1991666853,-0.1276440918,-0.256634742,0.2265536189,-0.066450499,-0.0968799889,0.0019838086,-0.0634838715,-0.1318117976,0.220077008,-0.3061932921,-0.0059892815,-0.0865267888,-0.0474983081,0.1305490732,0.0191281587,0.2057310492,-0.3622546494,0.1416555792,0.1029283404,0.3085215986,0.3533758223,-0.0966351852,0.0186193027,0.0468764491,0.1750154495,0.0544107519,-0.2243265808,-0.1458094567,-0.4689761698,0.2325953245,0.2488867044,0.3317735791,-0.1506784409,-0.2495945245,0.1341956556,-0.0551047288,0.2364847958,0.1205831841,0.5153110623,-0.307118088,0.1684069484,-0.062956959,0.02560132,-0.0268063359,-0.2216439694,-0.295609206,0.0240087099,-0.1846616417,-0.1395942122,0.2067635804,-0.2110509127,0.2445815951,0.079535611,0.0408378132,-0.3071087897,0.0538251624,-0.3537349403,-0.023229761,0.4643990993,0.0900522098,0.1324354857,0.2597099245,-0.0839310288,-0.218993932,0.1771682501,-0.3590294123,-0.5951446295,-0.0110709053,0.26149863,-0.1140651032,0.1193498001,-0.4246470928,0.0700019151,0.160954386,-0.0954999104,-0.0518162772,-0.1787497252,-0.2614167333,-0.3476524353,0.1499703825,0.3107527196,-0.5323678851,0.1687188745,-0.3266535997,-0.2732284069,0.297688961,0.2120352238,-0.1209672764,0.3310961425,-0.2531271875,0.0592158996,0.3152398765,-0.3005893826,-0.3999977708,0.3908524215,0.016317809,0.1309871674,0.0034007411,0.0098110065,0.3307875693,0.0078690965,0.1102341935,0.212696299,-0.1486558169,-0.0314865559,-0.262057066,-0.1641774923,-0.0509140231,0.156482935,0.1396570951,0.1109460518,0.1691527665,-0.3006534874,-0.0518827923,-0.065568164,0.2696269155,0.3614275157,0.2116627544,0.1624421924,0.1707906276,0.0812941045,-0.535584569,0.3256810606,0.0115941083,-0.1998816878,-0.0919436142,-0.1241044924,-0.5685514212,-0.0164452605,-0.1286688447,-0.00269557,0.1094136685,-0.105446212,-0.0243037492,-0.0921541229,-0.0302179251,0.5924558043,-0.1262022704,-0.1524438113,-0.1371084154,0.2782709002,0.0124592735,0.055331625,0.0212132409,0.0994580835,0.1813473701,-0.1014740244,-0.3069878519,0.3099374771,0.0669195652,-0.0598004498,-0.0391001031,-0.0881186426,-0.1291055232,-0.1401408911,-0.0137780439,0.157358557,0.2477229089,-0.0469300412,0.1126644611,0.1488340795,-0.2980319262,0.0036978342,-0.1184609905,0.0627791062,0.5067969561,0.0400967784,0.0703584403,-0.3693780303,0.2232140005,-0.1662302166,0.3108948171,-0.0223938841,-0.2173395753,-0.3514190018,0.1450280249,0.0782161579,-0.0974682942,0.4313172102,-0.0909135789,-0.1065372974,-0.0840541273,0.3009824455,0.5254251361,0.0834561214,-0.1675480157,0.2054767311,0.2845359147,-0.0061824503,0.2890945375,-0.1024241671,0.3282971978,0.2726708055,-0.0958034098,-0.1561962217,-0.1069378108,-0.1512015164,0.1809896976,0.480276674,-0.296161443,0.0076901154,-0.0392973498,-0.1293672323,-0.0263259336,-0.125998646,0.0485745929,-0.0580259934,-0.1010541618,0.1505691707,0.0410420112,-0.1340404749,-0.1771068275,0.2878244519,0.1359101534,-0.2367094755,-0.1092019901,-0.1559366435,-0.2683488727,0.1195478737,0.0350314416,0.01967692,0.2585455179,-0.1997407377,0.0914035216,-0.1030231938,-0.049859073,-0.0447158515,-0.001954356,0.3377043307,0.1962086111,0.2400903255,0.0899587795,0.0516655073,0.2236879468,-0.0502688475,-0.0530633293,0.3498876691,-0.1486122906,-0.0987843126,-0.0612148754,-0.4378128946,-0.5983160138,-0.3591504395,0.0012708708,0.0437375307,0.0452357754,0.2158264667,0.2280218601,-0.1275926828,0.3982062936,-0.0774030164,-0.3031054139,-0.5205017328,0.2376762778,-0.3503072262,-0.1679317504,0.0081502553,-0.0291733481,0.1761376858,0.0675562024,-0.5610889792,0.3567280471,-0.23838,0.0649895445,-0.2308612466,-0.3174830675,0.220419839,-0.1381314099,-0.0619107075,-0.248017773,-0.1541718543,-0.0304518882,0.0611269623,0.1230327561,-0.0677224621,0.5385031104,0.0502661429,0.5127761364,0.3299145997,-0.0016039216,0.4850444198,0.0683038458,0.3586151302,-0.2322539389,-0.3839135766,-0.0448417515,-0.0905613229,-0.2724584341,0.1639659405,-0.0662225708,-0.1001958027,-0.1565478146,-0.0477787927,-0.3332802057,0.0539398119,0.1047426835,-0.1628108025,0.3125785887,0.0887057185,-0.1207629219,-0.3875431418,-0.1475298703,-0.2114031762,0.3583055437,-0.0312432665,0.2129054815,-0.3576206565,0.0854414776,-0.3756477237,0.4200848341,0.1093478948,0.4523644447,-0.1290504336,0.2127711326,0.0378855318,0.0534792878,0.7103981376,-0.4645594358,0.0358819552,0.1659949124,0.0275682602,-0.3042504787,-0.285187155,-0.0084803542,0.3178522289,0.31413275,0.5803300142,-0.0361209065,0.0007447536,-0.0464967526,0.1986033916,-0.2785790265,-0.3796182871,-0.3453670144,-0.2140879482,-0.241249308,-0.0628155991,0.1035616323,0.4664235115,-0.0524130017,-0.11790663,-0.1577501893,0.0153487744,0.1427042633,0.0880366713,0.3629487157,-0.014422928,0.364948988,0.0744690523,0.0885612592,0.1970508397,0.6999536157,0.1273981035,-0.5392558575,-0.126157403,0.0696689337,0.0954375714,-0.0811017454,-0.2950466573,-0.1615704447,-0.0361088924,0.2540148199,-0.1686358005,0.2842821777,0.2741622627,-0.0936936513,-0.5076526403,-0.4867087603,0.2544363439,0.0747539699,-0.0489867106,0.2786591649,-0.213526547,-0.2284800708,0.0430670418,0.0376372449,0.6544899344,-0.0589199401,-0.1074389592,0.1674174219,-0.2733144164,0.2432312965,-0.4013942778,0.0802245364,-0.3252433836,-0.4311294854,-0.0189837143,-0.2221660614,0.0923741683,0.1845080107,-0.1360657662,0.3313300312,-0.2476435453,0.3380993307,-0.0432728752,0.0570543483,0.0132916989,-0.1616701931,-0.1127723157,0.2004107237,-0.2213427126,0.2725203037,-0.2305128127,-0.0070641725,0.0349441878,-0.2116746902,-0.3580960631,0.0895582661,-0.0012524695,0.332059145,-0.1337656528,-0.2141943872,0.0264652427,0.3820576966,0.2773966491,0.274762094,-0.1680817753,0.1762898564,0.0078900587,-0.0227795355,-0.0677925348,0.1592404395,0.057432916,-0.0700750053,-0.2695395648,0.1729298085,-0.1512601525,-0.1990173906,0.0742826536,-0.055200696,-0.0439750627,-0.3788851798,-0.4944426119,-0.2080943584,0.1806277335,-0.0699635819,0.1461118907,-0.0117006134,-0.0562958606,0.2779733837,0.0098770913,-0.4934690893,-0.0934577808,0.2358596474,0.1504454613,0.0423076935,0.577039063,-0.0227455944,-0.1267643571,-0.2060649842,0.1569144279,0.1524967104,-0.4153022766,0.2984954417,0.0065650269,0.2282925844,-0.0564338081,0.1067050397,0.2867481112,-0.0137914447,0.2696353793,-0.6932899952,-0.2438646704,0.0520335473,-0.2252726853,0.1804030538,0.1651545316,0.0428266078,-0.0882505253,-0.0495004915,-0.3103482425,0.1951560676,-0.2436397225,-0.0127718588,0.1171145812,0.0246869922,0.1024996936,-0.0507664867,0.1229806319,0.056868609,-0.2471610159,-0.2958017588,-0.1120594591,0.129828617,-0.0144994557,-0.1231976897,0.0595683232,-0.2221883237,-0.263602823,-0.0755089,0.0633006692,0.2292277962,0.0596436448,0.150957793,-0.1756052971,0.0745841935,-0.300188601,0.320567131,0.067909196,0.2082365453,-0.0075415117,0.2315397412,0.1362181306,-0.0870125145,-0.4957450926,-0.1165431663,0.1486171037,0.0311064757,0.3345166743,-0.2841587663,0.4677669406,0.185994938,0.177285254,0.2053661793,-0.1843367815,0.136656478,0.2792657316,0.2459561676,-0.341280371,-0.0360048562,-0.0271832012,0.2645361125,-0.151609987,0.0192744341,0.3053024113,-0.0335899964,-0.0239056367,0.0817202777,0.3608446419,0.1403882802,0.1985132247,0.4787999988,0.0550891384,0.1043441519,0.1473935395,0.1529682577,0.1486538649,0.8432779908,-0.0870846435,0.047505524,0.2190115452,0.1941461414,-0.2537889481,-0.7259444594,0.2099428028,0.2318012267,-0.1292050779,0.0636873543,0.0515897572,0.0154766347,-0.056893263,0.0287104771,-0.3996512592,-0.112338677,-0.1974150389,-0.0439040847,-0.1145654693,-0.0935092345,-0.095047079,0.0906768739,-0.1025839821,-0.0517633855,-0.0532695316,0.0804565623,-0.287350297,-0.3159598708,0.1776091307,0.4647827148,-0.0634765103,-0.3196803629,0.2313063741,0.1958958656,0.0570135266,0.1599377394,0.1180126965,0.5215353966,0.5515784621,-0.0306069646,0.106946446,0.1577190608,-0.1133501381,-0.0408964641,0.1973387003,0.3230094612,0.3432394862,0.2013502866,0.2322909236,-0.1237463504,0.3629772961,0.1353282332,0.2386198044,-0.158225745,0.0627807826,-0.1321887374,0.0853877291,-0.2960370183,0.1692750156,-0.2322701067,0.0723846555,0.537407279,0.173791036,0.1843811572,0.0281956922,0.0819050521,-0.1794721484,0.3207796812,0.1872993857,-0.0046041477,-0.4473469555,-0.1420500726,-0.4908000827,0.15742594,-0.3643202782,-0.1044761837,0.1451110989,0.1176997945,-0.0711466447,0.0115686674,0.0331624448,-0.0390997864,-0.1048875004,0.2687349617,-0.1599753946,-0.3986561298,-0.0576156043,0.1079313755,0.1156313792,-0.5605536699,-0.015596332,-0.1182360575,0.1090889573,-0.2412688285,-0.155986473,0.071618177,-0.0598636344,0.4797489047,0.0273879673,0.4527481496,-0.0294720344,0.1498543769,-0.3443867862,-0.3039680719,-0.0468843468,0.3253214657,0.1939056218,0.2710701227,-0.2828695178,0.1123746485,-0.1649846733,-0.002180147,0.0596382879,0.1852112263,-0.0965949371,-0.0817463994,-0.1364638507,0.1857875586,0.1331339628,0.2467502058,-0.1912277788,0.072395578,-0.0053284904,-0.3337082267,0.5975276828,-0.0790108368,-0.2732951343,0.0007333648,0.0964822322,0.2757702172,-0.1867084354,-0.2042184472,0.1991377473,0.4787468016,-0.1332235187,-0.0488588773,0.171931237,0.0861932039,0.111177519,-0.0059969663,0.1856501997,0.0163134839,-0.1944563091,0.1022795588,-0.3659286499]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1773","title":"bug in loading datasets ","comments":"You can try to convert the file to (CSV UTF-8)","body":"Hi,\r\nI need to load a dataset, I use these commands:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files={'train': 'sick\/train.csv',\r\n 'test': 'sick\/test.csv',\r\n 'validation': 'sick\/validation.csv'})\r\nprint(dataset['validation'])\r\n```\r\nthe dataset in sick\/train.csv are simple csv files representing the data. I am getting this error, do you have an idea how I can solve this? thank you @lhoestq \r\n\r\n \r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-61468fc71a743ec1 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nTraceback (most recent call last):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 485, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 604, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 959, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/tqdm-4.49.0-py3.7.egg\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/julia\/cache_home_2\/modules\/datasets_modules\/datasets\/csv\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2\/csv.py\", line 129, in _generate_tables\r\n for batch_idx, df in enumerate(csv_file_reader):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1029, in __next__\r\n return self.get_chunk()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1079, in get_chunk\r\n return self.read(nrows=size)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1052, in read\r\n index, columns, col_dict = self._engine.read(nrows)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 2056, in read\r\n data = self._reader.read(nrows)\r\n File \"pandas\/_libs\/parsers.pyx\", line 756, in pandas._libs.parsers.TextReader.read\r\n File \"pandas\/_libs\/parsers.pyx\", line 783, in pandas._libs.parsers.TextReader._read_low_memory\r\n File \"pandas\/_libs\/parsers.pyx\", line 827, in pandas._libs.parsers.TextReader._read_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 814, in pandas._libs.parsers.TextReader._tokenize_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 1951, in pandas._libs.parsers.raise_parser_error\r\npandas.errors.ParserError: Error tokenizing data. C error: Expected 1 fields in line 37, saw 2\r\n\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"write_sick.py\", line 19, in <module>\r\n 'validation': 'sick\/validation.csv'})\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 534, in download_and_prepare\r\n self._save_info()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/contextlib.py\", line 130, in __exit__\r\n self.gen.throw(type, value, traceback)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 491, in incomplete_dir\r\n shutil.rmtree(tmp_dir)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 498, in rmtree\r\n onerror(os.rmdir, path, sys.exc_info())\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 496, in rmtree\r\n os.rmdir(path)\r\nOSError: [Errno 39] Directory not empty: '\/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2.incomplete'\r\n```\r\n\r\n","comment_length":10,"text":"bug in loading datasets \nHi,\r\nI need to load a dataset, I use these commands:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files={'train': 'sick\/train.csv',\r\n 'test': 'sick\/test.csv',\r\n 'validation': 'sick\/validation.csv'})\r\nprint(dataset['validation'])\r\n```\r\nthe dataset in sick\/train.csv are simple csv files representing the data. I am getting this error, do you have an idea how I can solve this? thank you @lhoestq \r\n\r\n \r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-61468fc71a743ec1 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nTraceback (most recent call last):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 485, in incomplete_dir\r\n yield tmp_dir\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 604, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 959, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/tqdm-4.49.0-py3.7.egg\/tqdm\/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"\/julia\/cache_home_2\/modules\/datasets_modules\/datasets\/csv\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2\/csv.py\", line 129, in _generate_tables\r\n for batch_idx, df in enumerate(csv_file_reader):\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1029, in __next__\r\n return self.get_chunk()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1079, in get_chunk\r\n return self.read(nrows=size)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 1052, in read\r\n index, columns, col_dict = self._engine.read(nrows)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/pandas-1.2.0-py3.7-linux-x86_64.egg\/pandas\/io\/parsers.py\", line 2056, in read\r\n data = self._reader.read(nrows)\r\n File \"pandas\/_libs\/parsers.pyx\", line 756, in pandas._libs.parsers.TextReader.read\r\n File \"pandas\/_libs\/parsers.pyx\", line 783, in pandas._libs.parsers.TextReader._read_low_memory\r\n File \"pandas\/_libs\/parsers.pyx\", line 827, in pandas._libs.parsers.TextReader._read_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 814, in pandas._libs.parsers.TextReader._tokenize_rows\r\n File \"pandas\/_libs\/parsers.pyx\", line 1951, in pandas._libs.parsers.raise_parser_error\r\npandas.errors.ParserError: Error tokenizing data. C error: Expected 1 fields in line 37, saw 2\r\n\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"write_sick.py\", line 19, in <module>\r\n 'validation': 'sick\/validation.csv'})\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 534, in download_and_prepare\r\n self._save_info()\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/contextlib.py\", line 130, in __exit__\r\n self.gen.throw(type, value, traceback)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/site-packages\/datasets-1.2.0-py3.7.egg\/datasets\/builder.py\", line 491, in incomplete_dir\r\n shutil.rmtree(tmp_dir)\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 498, in rmtree\r\n onerror(os.rmdir, path, sys.exc_info())\r\n File \"\/julia\/libs\/anaconda3\/envs\/success\/lib\/python3.7\/shutil.py\", line 496, in rmtree\r\n os.rmdir(path)\r\nOSError: [Errno 39] Directory not empty: '\/julia\/cache_home_2\/datasets\/csv\/default-61468fc71a743ec1\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2.incomplete'\r\n```\r\n\r\n\nYou can try to convert the file to (CSV UTF-8)","embeddings":[-0.2842670083,-0.2714768052,-0.1463154405,0.4608483613,0.3019686937,0.2408371568,0.0834566206,0.5121120214,0.0764491484,0.0585810058,-0.0004540544,-0.2665965557,-0.10810256,0.2536355257,-0.0337027758,0.0352051184,-0.0331644155,0.1680267006,-0.0946208239,0.019386515,-0.2684040964,0.2590358853,-0.0957386568,0.0894601047,-0.0035116654,-0.1219344884,0.0992638245,0.2603849769,-0.2975808978,-0.3687266707,0.3527499437,-0.1765222102,0.3671296537,0.2056512088,-0.0001100329,0.109722428,0.4414513707,-0.222237736,-0.2273045331,-0.2782835364,-0.3186444938,-0.5255689621,-0.0424169973,-0.2519200146,-0.0720899403,-0.1986101866,-0.1186174303,-0.358505249,0.1968753785,0.4839046001,0.3277643323,-0.2024365067,-0.0772970542,-0.2182596326,0.1736675203,0.0825364217,0.0177823696,0.0985001624,0.0383216366,0.1259890795,0.5457251668,0.22481516,-0.0871027187,0.062276613,0.1328184605,-0.0650291443,0.2201796621,-0.3268524706,0.1852118075,0.1893909127,0.8423666954,-0.1272012442,-0.1620606333,-0.14171803,0.1230335012,-0.309578687,0.0035668784,0.3891044259,-0.1592620909,0.1557286382,-0.1961616576,0.022634821,-0.056327071,0.1439458579,0.1991666853,-0.1276440918,-0.256634742,0.2265536189,-0.066450499,-0.0968799889,0.0019838086,-0.0634838715,-0.1318117976,0.220077008,-0.3061932921,-0.0059892815,-0.0865267888,-0.0474983081,0.1305490732,0.0191281587,0.2057310492,-0.3622546494,0.1416555792,0.1029283404,0.3085215986,0.3533758223,-0.0966351852,0.0186193027,0.0468764491,0.1750154495,0.0544107519,-0.2243265808,-0.1458094567,-0.4689761698,0.2325953245,0.2488867044,0.3317735791,-0.1506784409,-0.2495945245,0.1341956556,-0.0551047288,0.2364847958,0.1205831841,0.5153110623,-0.307118088,0.1684069484,-0.062956959,0.02560132,-0.0268063359,-0.2216439694,-0.295609206,0.0240087099,-0.1846616417,-0.1395942122,0.2067635804,-0.2110509127,0.2445815951,0.079535611,0.0408378132,-0.3071087897,0.0538251624,-0.3537349403,-0.023229761,0.4643990993,0.0900522098,0.1324354857,0.2597099245,-0.0839310288,-0.218993932,0.1771682501,-0.3590294123,-0.5951446295,-0.0110709053,0.26149863,-0.1140651032,0.1193498001,-0.4246470928,0.0700019151,0.160954386,-0.0954999104,-0.0518162772,-0.1787497252,-0.2614167333,-0.3476524353,0.1499703825,0.3107527196,-0.5323678851,0.1687188745,-0.3266535997,-0.2732284069,0.297688961,0.2120352238,-0.1209672764,0.3310961425,-0.2531271875,0.0592158996,0.3152398765,-0.3005893826,-0.3999977708,0.3908524215,0.016317809,0.1309871674,0.0034007411,0.0098110065,0.3307875693,0.0078690965,0.1102341935,0.212696299,-0.1486558169,-0.0314865559,-0.262057066,-0.1641774923,-0.0509140231,0.156482935,0.1396570951,0.1109460518,0.1691527665,-0.3006534874,-0.0518827923,-0.065568164,0.2696269155,0.3614275157,0.2116627544,0.1624421924,0.1707906276,0.0812941045,-0.535584569,0.3256810606,0.0115941083,-0.1998816878,-0.0919436142,-0.1241044924,-0.5685514212,-0.0164452605,-0.1286688447,-0.00269557,0.1094136685,-0.105446212,-0.0243037492,-0.0921541229,-0.0302179251,0.5924558043,-0.1262022704,-0.1524438113,-0.1371084154,0.2782709002,0.0124592735,0.055331625,0.0212132409,0.0994580835,0.1813473701,-0.1014740244,-0.3069878519,0.3099374771,0.0669195652,-0.0598004498,-0.0391001031,-0.0881186426,-0.1291055232,-0.1401408911,-0.0137780439,0.157358557,0.2477229089,-0.0469300412,0.1126644611,0.1488340795,-0.2980319262,0.0036978342,-0.1184609905,0.0627791062,0.5067969561,0.0400967784,0.0703584403,-0.3693780303,0.2232140005,-0.1662302166,0.3108948171,-0.0223938841,-0.2173395753,-0.3514190018,0.1450280249,0.0782161579,-0.0974682942,0.4313172102,-0.0909135789,-0.1065372974,-0.0840541273,0.3009824455,0.5254251361,0.0834561214,-0.1675480157,0.2054767311,0.2845359147,-0.0061824503,0.2890945375,-0.1024241671,0.3282971978,0.2726708055,-0.0958034098,-0.1561962217,-0.1069378108,-0.1512015164,0.1809896976,0.480276674,-0.296161443,0.0076901154,-0.0392973498,-0.1293672323,-0.0263259336,-0.125998646,0.0485745929,-0.0580259934,-0.1010541618,0.1505691707,0.0410420112,-0.1340404749,-0.1771068275,0.2878244519,0.1359101534,-0.2367094755,-0.1092019901,-0.1559366435,-0.2683488727,0.1195478737,0.0350314416,0.01967692,0.2585455179,-0.1997407377,0.0914035216,-0.1030231938,-0.049859073,-0.0447158515,-0.001954356,0.3377043307,0.1962086111,0.2400903255,0.0899587795,0.0516655073,0.2236879468,-0.0502688475,-0.0530633293,0.3498876691,-0.1486122906,-0.0987843126,-0.0612148754,-0.4378128946,-0.5983160138,-0.3591504395,0.0012708708,0.0437375307,0.0452357754,0.2158264667,0.2280218601,-0.1275926828,0.3982062936,-0.0774030164,-0.3031054139,-0.5205017328,0.2376762778,-0.3503072262,-0.1679317504,0.0081502553,-0.0291733481,0.1761376858,0.0675562024,-0.5610889792,0.3567280471,-0.23838,0.0649895445,-0.2308612466,-0.3174830675,0.220419839,-0.1381314099,-0.0619107075,-0.248017773,-0.1541718543,-0.0304518882,0.0611269623,0.1230327561,-0.0677224621,0.5385031104,0.0502661429,0.5127761364,0.3299145997,-0.0016039216,0.4850444198,0.0683038458,0.3586151302,-0.2322539389,-0.3839135766,-0.0448417515,-0.0905613229,-0.2724584341,0.1639659405,-0.0662225708,-0.1001958027,-0.1565478146,-0.0477787927,-0.3332802057,0.0539398119,0.1047426835,-0.1628108025,0.3125785887,0.0887057185,-0.1207629219,-0.3875431418,-0.1475298703,-0.2114031762,0.3583055437,-0.0312432665,0.2129054815,-0.3576206565,0.0854414776,-0.3756477237,0.4200848341,0.1093478948,0.4523644447,-0.1290504336,0.2127711326,0.0378855318,0.0534792878,0.7103981376,-0.4645594358,0.0358819552,0.1659949124,0.0275682602,-0.3042504787,-0.285187155,-0.0084803542,0.3178522289,0.31413275,0.5803300142,-0.0361209065,0.0007447536,-0.0464967526,0.1986033916,-0.2785790265,-0.3796182871,-0.3453670144,-0.2140879482,-0.241249308,-0.0628155991,0.1035616323,0.4664235115,-0.0524130017,-0.11790663,-0.1577501893,0.0153487744,0.1427042633,0.0880366713,0.3629487157,-0.014422928,0.364948988,0.0744690523,0.0885612592,0.1970508397,0.6999536157,0.1273981035,-0.5392558575,-0.126157403,0.0696689337,0.0954375714,-0.0811017454,-0.2950466573,-0.1615704447,-0.0361088924,0.2540148199,-0.1686358005,0.2842821777,0.2741622627,-0.0936936513,-0.5076526403,-0.4867087603,0.2544363439,0.0747539699,-0.0489867106,0.2786591649,-0.213526547,-0.2284800708,0.0430670418,0.0376372449,0.6544899344,-0.0589199401,-0.1074389592,0.1674174219,-0.2733144164,0.2432312965,-0.4013942778,0.0802245364,-0.3252433836,-0.4311294854,-0.0189837143,-0.2221660614,0.0923741683,0.1845080107,-0.1360657662,0.3313300312,-0.2476435453,0.3380993307,-0.0432728752,0.0570543483,0.0132916989,-0.1616701931,-0.1127723157,0.2004107237,-0.2213427126,0.2725203037,-0.2305128127,-0.0070641725,0.0349441878,-0.2116746902,-0.3580960631,0.0895582661,-0.0012524695,0.332059145,-0.1337656528,-0.2141943872,0.0264652427,0.3820576966,0.2773966491,0.274762094,-0.1680817753,0.1762898564,0.0078900587,-0.0227795355,-0.0677925348,0.1592404395,0.057432916,-0.0700750053,-0.2695395648,0.1729298085,-0.1512601525,-0.1990173906,0.0742826536,-0.055200696,-0.0439750627,-0.3788851798,-0.4944426119,-0.2080943584,0.1806277335,-0.0699635819,0.1461118907,-0.0117006134,-0.0562958606,0.2779733837,0.0098770913,-0.4934690893,-0.0934577808,0.2358596474,0.1504454613,0.0423076935,0.577039063,-0.0227455944,-0.1267643571,-0.2060649842,0.1569144279,0.1524967104,-0.4153022766,0.2984954417,0.0065650269,0.2282925844,-0.0564338081,0.1067050397,0.2867481112,-0.0137914447,0.2696353793,-0.6932899952,-0.2438646704,0.0520335473,-0.2252726853,0.1804030538,0.1651545316,0.0428266078,-0.0882505253,-0.0495004915,-0.3103482425,0.1951560676,-0.2436397225,-0.0127718588,0.1171145812,0.0246869922,0.1024996936,-0.0507664867,0.1229806319,0.056868609,-0.2471610159,-0.2958017588,-0.1120594591,0.129828617,-0.0144994557,-0.1231976897,0.0595683232,-0.2221883237,-0.263602823,-0.0755089,0.0633006692,0.2292277962,0.0596436448,0.150957793,-0.1756052971,0.0745841935,-0.300188601,0.320567131,0.067909196,0.2082365453,-0.0075415117,0.2315397412,0.1362181306,-0.0870125145,-0.4957450926,-0.1165431663,0.1486171037,0.0311064757,0.3345166743,-0.2841587663,0.4677669406,0.185994938,0.177285254,0.2053661793,-0.1843367815,0.136656478,0.2792657316,0.2459561676,-0.341280371,-0.0360048562,-0.0271832012,0.2645361125,-0.151609987,0.0192744341,0.3053024113,-0.0335899964,-0.0239056367,0.0817202777,0.3608446419,0.1403882802,0.1985132247,0.4787999988,0.0550891384,0.1043441519,0.1473935395,0.1529682577,0.1486538649,0.8432779908,-0.0870846435,0.047505524,0.2190115452,0.1941461414,-0.2537889481,-0.7259444594,0.2099428028,0.2318012267,-0.1292050779,0.0636873543,0.0515897572,0.0154766347,-0.056893263,0.0287104771,-0.3996512592,-0.112338677,-0.1974150389,-0.0439040847,-0.1145654693,-0.0935092345,-0.095047079,0.0906768739,-0.1025839821,-0.0517633855,-0.0532695316,0.0804565623,-0.287350297,-0.3159598708,0.1776091307,0.4647827148,-0.0634765103,-0.3196803629,0.2313063741,0.1958958656,0.0570135266,0.1599377394,0.1180126965,0.5215353966,0.5515784621,-0.0306069646,0.106946446,0.1577190608,-0.1133501381,-0.0408964641,0.1973387003,0.3230094612,0.3432394862,0.2013502866,0.2322909236,-0.1237463504,0.3629772961,0.1353282332,0.2386198044,-0.158225745,0.0627807826,-0.1321887374,0.0853877291,-0.2960370183,0.1692750156,-0.2322701067,0.0723846555,0.537407279,0.173791036,0.1843811572,0.0281956922,0.0819050521,-0.1794721484,0.3207796812,0.1872993857,-0.0046041477,-0.4473469555,-0.1420500726,-0.4908000827,0.15742594,-0.3643202782,-0.1044761837,0.1451110989,0.1176997945,-0.0711466447,0.0115686674,0.0331624448,-0.0390997864,-0.1048875004,0.2687349617,-0.1599753946,-0.3986561298,-0.0576156043,0.1079313755,0.1156313792,-0.5605536699,-0.015596332,-0.1182360575,0.1090889573,-0.2412688285,-0.155986473,0.071618177,-0.0598636344,0.4797489047,0.0273879673,0.4527481496,-0.0294720344,0.1498543769,-0.3443867862,-0.3039680719,-0.0468843468,0.3253214657,0.1939056218,0.2710701227,-0.2828695178,0.1123746485,-0.1649846733,-0.002180147,0.0596382879,0.1852112263,-0.0965949371,-0.0817463994,-0.1364638507,0.1857875586,0.1331339628,0.2467502058,-0.1912277788,0.072395578,-0.0053284904,-0.3337082267,0.5975276828,-0.0790108368,-0.2732951343,0.0007333648,0.0964822322,0.2757702172,-0.1867084354,-0.2042184472,0.1991377473,0.4787468016,-0.1332235187,-0.0488588773,0.171931237,0.0861932039,0.111177519,-0.0059969663,0.1856501997,0.0163134839,-0.1944563091,0.1022795588,-0.3659286499]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1771","title":"Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py","comments":"I temporary manually download csv.py as custom dataset loading script","body":"Hi,\r\nWhen I load_dataset from local csv files, below error happened, looks raw.githubusercontent.com was blocked by the chinese government. But why it need to download csv.py? should it include when pip install the dataset?\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\r\n\r\n```","comment_length":10,"text":"Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\nHi,\r\nWhen I load_dataset from local csv files, below error happened, looks raw.githubusercontent.com was blocked by the chinese government. But why it need to download csv.py? should it include when pip install the dataset?\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\r\n\r\n```\nI temporary manually download csv.py as custom dataset loading script","embeddings":[-0.1969490349,-0.2061192393,-0.1841538996,0.0926497653,0.242069602,-0.0245270859,0.3229030371,0.167100504,0.3168790936,0.2279049158,0.0691505224,-0.2167990357,0.2441080958,0.1349897683,0.2439681739,-0.0011481892,-0.0132608032,0.1838507503,0.0296760909,0.0927444175,-0.2004780322,0.0951571018,-0.0548260324,0.2160592973,0.178629145,-0.0194923338,0.0142805213,0.2578603923,-0.1424791813,-0.6663218737,0.5473432541,-0.0680860654,0.108065933,0.3600767255,-0.0001195253,0.222969979,0.2120305896,0.0153513057,-0.3275021315,-0.336509496,-0.3290456235,-0.4228983521,0.1852897853,-0.1483903378,-0.1757824868,0.0847378001,0.0263391156,-0.048021853,0.530028522,0.3286390305,0.1390079409,0.3927007914,0.1799832731,-0.0434313975,0.0464960337,-0.2990507782,0.0332890414,0.6693676114,0.260388881,0.0929999799,0.2027494311,0.0247815754,0.0037122911,0.0965218917,-0.0101203127,-0.2113113254,-0.0100434087,-0.3796847761,0.1437087953,0.1905751526,0.413977474,-0.0284750815,-0.3938136995,-0.1276945025,-0.0775336921,-0.2828590572,0.1863932014,0.2586400807,-0.2035839558,0.3870331943,-0.4109651744,-0.3659065962,-0.2299625576,0.0873101726,0.1113186181,-0.0789797828,-0.2431789935,0.255256474,-0.1502258033,0.2065335363,-0.0669561923,-0.2008752227,0.1962405145,0.2579600513,-0.1724251211,0.0778473541,0.0051618111,0.1608538628,0.0449185632,-0.0015215022,-0.1057638228,-0.0222983621,0.1151894629,0.175844878,0.1899362057,0.0757733136,-0.2369034439,0.2602403462,0.3778322041,0.1409008652,0.17660743,0.0005423555,-0.1458932459,-0.1562752724,-0.225397557,0.0767948553,0.317147553,-0.2315861881,-0.3172631264,0.0771652833,0.0661692247,0.0122996634,-0.1852307767,0.274007678,-0.2581250072,-0.1248394325,-0.3193017542,0.2256391495,0.0514941961,0.2318936288,-0.133154735,-0.0360298716,0.0109040886,-0.0071723727,0.0111601045,-0.3203340769,0.381485045,0.0312625058,0.0025324523,-0.080840379,0.1458582729,-0.0565574579,-0.0852696225,0.6523807049,0.2277386636,0.1405158937,0.2869548202,-0.25658077,-0.1525477916,-0.1312748492,-0.3703847229,-0.6194493175,0.1211815476,0.1407828927,-0.2159942091,0.0934283361,-0.3589367867,-0.2712599635,0.1370488107,-0.0490073226,0.0336411633,-0.0994313061,0.0871587172,-0.2522758245,0.1896377355,0.3966382444,-0.3020039499,-0.0141839851,-0.0752836242,-0.2241472602,0.2021617293,0.359000802,-0.3288212121,0.1613723338,-0.4275642335,-0.1517564654,0.3188758194,-0.5589490533,-0.5085228682,0.6391569376,-0.5307084918,0.117152527,0.3265010118,-0.0959639177,0.1010153517,-0.1588915735,0.1061029807,0.2603435218,-0.1678221673,-0.0501223356,-0.0577370003,-0.2285492122,0.1111978516,0.096326679,0.1533869207,0.2432200015,0.4428294301,-0.3547056615,0.1842449158,0.0853876472,0.0310546085,0.3478429019,0.0577566586,0.1285423785,-0.0171269048,-0.0126856742,-0.1075840667,0.1851170361,0.0289487317,0.1169029623,-0.4799278378,-0.2232474834,-0.4874551594,0.0722390264,-0.3540833592,-0.1803608984,-0.0124380998,0.1420348287,0.1989736855,0.0502013266,-0.0877555758,0.3285340071,-0.2672290504,0.1020130515,-0.3349591792,0.0118947392,-0.3210260272,0.2469958216,0.0775369257,0.2022463977,0.1452335566,-0.3354509771,-0.0258199163,0.3236062527,-0.2244082391,0.2379588932,0.4347685874,-0.12846452,0.1049637049,-0.3255903721,0.0193629563,0.004960421,0.1618990898,-0.0023356853,0.3094712496,0.3137034774,-0.1618857533,0.3303669691,0.1678130627,0.0788311809,0.541618526,0.1128791943,-0.1281749606,-0.0825494826,0.1197677776,0.3287342489,0.2806787491,0.0780456215,-0.2254360765,0.0097166328,0.3544604778,-0.1873847991,-0.0471538603,0.2012195736,-0.0489390306,-0.1449252665,0.2258712947,0.2960305214,0.2489401698,0.1078833863,0.0089538842,0.0222342573,0.0606422462,-0.2888617218,0.2401614487,-0.0199254267,-0.0761572123,-0.0271123908,-0.201496914,-0.0138481846,-0.179246068,-0.3153145909,-0.0205278713,0.1443105787,-0.3195095062,0.0127183637,-0.1578916013,-0.2295783311,0.0445241742,-0.1906919777,-0.0278932136,-0.0245805848,-0.2674894035,-0.143237412,0.0271293335,0.082984671,-0.1433550715,0.0039001242,0.075038515,-0.3300081789,0.0627219751,-0.2363189161,-0.0938873813,0.102665551,0.071163103,-0.200143829,0.3981681168,-0.3580671549,-0.1487505734,-0.2538259327,-0.1190376952,-0.0312095843,-0.2487812638,0.2058324665,0.1862164587,0.6411461234,-0.0142793534,0.0861860365,0.1527985036,0.2094501406,-0.061699979,0.0976204053,-0.2243711948,-0.2914520502,0.0157160182,-0.3471868336,-0.6579802632,-0.3063192368,0.4253968596,0.0751204044,0.1681449562,0.0974504128,0.1366984844,0.170243755,0.021240633,0.1342751235,-0.1110953391,-0.7418760657,0.1805948317,-0.1922915131,-0.3806400001,0.124672845,0.0956399962,0.1782004237,0.1391222328,-0.455801338,0.0437652282,-0.2643215358,0.5890811086,0.191258654,0.2088419646,0.1974751204,-0.1290023029,-0.0293897595,0.0723281354,-0.0648053363,-0.1240729913,-0.0918305814,0.1225583851,0.2331567109,0.6345149875,-0.0644344389,0.8888664246,0.1238419637,0.2087316364,0.3815770745,-0.1937227398,0.6180426478,-0.1105554998,-0.397752434,0.1487150937,-0.1252796948,-0.0219610389,-0.0636867732,-0.0451626889,0.1941672564,-0.3416737318,-0.2147370726,-0.452645272,-0.0254468173,-0.2672270238,-0.0725701228,0.2829962671,0.3063677251,-0.2915183008,-0.058019869,-0.0145247094,-0.1119689122,0.4785342515,0.1663359106,0.0870080814,-0.2340773493,-0.3577452302,-0.4465774,0.2167133093,0.1869516075,0.2902840972,-0.1718058437,0.2110651433,0.0799061283,-0.0002365416,0.4243252575,-0.1842354983,0.3335191011,-0.1632293761,-0.11790815,-0.2420277148,0.0204548314,0.1819484681,0.4381572604,0.4159663916,0.3367570937,-0.2441447675,-0.0197231155,0.2383439541,0.3959396183,-0.0120175686,0.0064537711,-0.4187787175,-0.4474771023,-0.2672109902,0.0418556742,0.0570945255,-0.0033314188,-0.0825629756,-0.3460530937,0.0836957693,0.0891808718,-0.1467718929,0.0045115268,-0.156282261,-0.1160831079,-0.0725863352,-0.0489228889,-0.0366211534,0.2328026593,0.8784669042,-0.0367666036,-0.3641440868,0.0762484819,-0.0915661156,0.1247389689,0.349604249,-0.2525467575,0.1978998929,0.2932415605,0.2166502327,0.0705646351,0.0144462809,0.1108205691,0.181975022,-0.3123062253,-0.5749318004,0.2332706898,-0.2268923074,-0.0649113804,0.1612473279,0.2398845106,-0.0607111976,0.0789008886,-0.3032289445,0.8494951725,-0.0796539634,0.1058655605,0.2865784168,-0.1564764082,0.556936264,-0.0793382749,-0.0895024762,-0.0222863741,0.0149788717,-0.1874904633,-0.2024975419,0.262771666,-0.0671204329,0.0847243592,0.5718619227,0.1050211564,0.2835291624,0.3183987737,0.4354213774,-0.109913215,0.1757588089,-0.2358414978,0.034255676,-0.0014446668,0.3249995708,-0.1311359107,-0.1622727811,-0.216626808,-0.4397167265,-0.3321402669,0.1527711749,-0.2070112079,0.2972948253,-0.3188688755,-0.4012084603,0.0343611613,0.2150567025,0.3635442257,0.1499481201,-0.3475130796,0.1354264468,0.0054530906,-0.1969682276,0.0650330335,0.2139675915,0.0590792373,-0.2406923175,-0.2243006676,0.1297974288,-0.3712370992,-0.4614556432,0.1245114207,0.1781748086,-0.0643798783,-0.0170395002,-0.280030489,-0.4104151726,-0.0652670488,-0.1919267774,0.0378308371,0.0086867344,0.0773302168,-0.1473588347,0.1153303236,-0.4852019548,-0.277310729,0.3103234768,0.0850670561,0.0743961781,0.3574180007,-0.0405433029,-0.215330258,-0.1177147031,-0.1126495078,0.1368799657,-0.3542805612,-0.087951377,-0.0229838919,0.1834204346,-0.2589744925,0.1256423146,0.3888766766,-0.0845077634,-0.0176193416,-0.7460857034,-0.2841875255,-0.104866311,-0.0947547778,-0.0329577252,0.2270518988,0.1898005158,0.2289595902,-0.079624176,-0.214805305,0.0327754878,-0.2392866313,0.1819066405,0.3572003543,0.1054020971,0.4205701351,-0.0290725231,0.1385042369,-0.252941817,-0.2880213559,-0.1412551254,0.1099326387,0.1709262282,-0.0248100404,-0.2389231324,-0.180346176,-0.2072596103,0.0145605914,0.1589036584,0.1578312367,0.0765953511,0.0955281258,-0.1852499694,-0.2424688488,0.0518973544,-0.2450108081,0.2854524851,-0.4171980619,0.3418105543,0.0214747414,0.164472878,-0.0528839454,-0.0467721671,-0.2161142081,-0.2290865481,0.0731772333,-0.0019150729,0.5194383264,-0.3279051483,0.1509516835,0.1485845596,0.3437284231,0.2286242098,-0.176237613,0.0948157161,0.0820655972,0.096350722,-0.3052568734,0.2010356933,0.0976682007,-0.044383809,-0.0630739927,0.0548753217,0.3990854919,-0.1064848676,0.2111064643,-0.0113045545,0.7854106426,0.0472548045,0.0184044987,0.2689115703,0.182999298,0.238925308,-0.0601860248,0.361345619,-0.0778857991,0.282156229,-0.2769093513,-0.1138646677,-0.0239833053,0.2433207929,-0.2244490385,-0.5417872071,0.5455887914,0.1188922897,-0.4663682878,-0.2258938402,0.2632601261,0.3340494335,-0.3103322089,0.221301496,0.0187868942,0.1200946718,0.0218066201,0.0674126372,0.0473525934,-0.0405448265,0.0573115014,-0.0434440002,0.091910176,-0.1448794603,-0.0184122361,0.1064101234,-0.1854784787,-0.4029099941,0.1841347516,0.0276183821,0.0018913304,-0.175048545,0.2575090826,0.1014605761,-0.1440904289,-0.0302648488,0.1879362911,0.5893906355,0.2722703218,0.0379848219,0.0894556865,0.2444490939,-0.2660178542,-0.0742401853,0.1641228795,-0.0156824272,0.3121240139,0.2503847182,0.1828242242,-0.088684164,0.2843287587,0.0588281043,0.012683684,-0.2496294081,0.2258969396,-0.2286780328,-0.4112636447,-0.1332216263,0.1796362549,-0.4994313717,0.1605785042,0.2956433296,-0.1921279132,0.085652031,-0.4153088629,0.0427462272,-0.0306188688,0.3671235144,0.7350178361,0.4757154286,-0.3649742305,-0.1150976047,-0.5359549522,0.1801822037,-0.2674511373,-0.2595320046,0.1197268739,0.0435542986,-0.1343605816,0.3162280321,-0.0603701435,0.1648322642,0.0761959255,0.1368903667,0.1278596669,-0.0908898339,-0.1075619608,0.2263334543,-0.1748804003,-0.3929465711,0.4082396328,-0.3835279644,0.0061996123,-0.0869504586,0.1704077423,0.2376867235,-0.2941945493,0.4731257856,0.066882439,0.5033974648,-0.1132855639,-0.0654449984,-0.0997038707,-0.0664136037,-0.0765052661,0.2450500429,0.0318754278,0.017956702,-0.1341199428,-0.1952099949,-0.2646332681,0.2651206255,0.0130420085,0.0834508613,0.2525874376,-0.1687511355,-0.0551952906,0.0306308363,0.2216985971,-0.0220255032,0.0938702077,0.0434185937,-0.1375422329,-0.2712616026,0.3674415946,-0.3559832871,-0.0741982609,0.1429470628,0.3438723683,0.1076587513,-0.1932466477,-0.1395086646,0.2420581877,0.4609507024,0.0662625581,-0.3439002931,0.2159630507,-0.1776047796,0.2343275845,-0.0233168453,0.2712427676,-0.2130893171,-0.2832908332,-0.0797162876,-0.2947262526]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1771","title":"Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py","comments":"Indeed in 1.2.1 the script to process csv file is downloaded. Starting from the next release though we include the csv processing directly in the library.\r\nSee PR #1726 \r\nWe'll do a new release soon :)","body":"Hi,\r\nWhen I load_dataset from local csv files, below error happened, looks raw.githubusercontent.com was blocked by the chinese government. But why it need to download csv.py? should it include when pip install the dataset?\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\r\n\r\n```","comment_length":36,"text":"Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\nHi,\r\nWhen I load_dataset from local csv files, below error happened, looks raw.githubusercontent.com was blocked by the chinese government. But why it need to download csv.py? should it include when pip install the dataset?\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\r\n\r\n```\nIndeed in 1.2.1 the script to process csv file is downloaded. Starting from the next release though we include the csv processing directly in the library.\r\nSee PR #1726 \r\nWe'll do a new release soon :)","embeddings":[-0.0955533609,-0.2019754499,-0.1823532134,0.099574469,0.1923654974,-0.0114183808,0.2810024917,0.1859909147,0.2537339628,0.2542534173,0.0892188251,-0.1967477947,0.238546744,0.0940238312,0.2848533392,0.0054628174,0.0613934845,0.1827716827,-0.0548004583,0.073835142,-0.2219042629,0.1612168401,-0.0770464987,0.2406686842,0.1830085814,-0.0438245535,-0.0183390863,0.2499673516,-0.1553301513,-0.6863773465,0.5335741043,-0.0877207369,0.0872390419,0.3809689581,-0.0001219624,0.178487286,0.2266519368,0.0600899719,-0.3041380644,-0.4303004444,-0.2043637484,-0.4646512866,0.2093785554,-0.0881472453,-0.1116306409,0.0846512914,0.0986055508,-0.0679887012,0.5056246519,0.257109493,0.1253383458,0.4090678096,0.1611114889,-0.0237109307,0.0797307566,-0.2780303657,0.0383103751,0.6396939754,0.2506096959,0.0791779608,0.186325267,0.06411881,0.0218927301,0.0369007252,0.0472989157,-0.1882143319,-0.1258438975,-0.4115513861,0.1586073488,0.1841291189,0.4187727571,-0.0393403769,-0.4462423921,-0.1533686668,-0.1058287397,-0.2536841631,0.2397563457,0.2356263846,-0.1876695901,0.3894172907,-0.3782468438,-0.4112767875,-0.2539040744,0.0796013549,0.1502715051,-0.1225686893,-0.273391664,0.2718202472,-0.0534091145,0.2579774857,-0.1241595596,-0.2064068913,0.238528043,0.2558237016,-0.1334979534,0.0767889321,0.0650765002,0.1773647964,0.0727835,0.0202904772,-0.1572663337,-0.068401888,0.0870650709,0.2111350596,0.2821736336,0.0334909111,-0.2121037543,0.2013408542,0.4550292194,0.0691013485,0.2622926235,-0.0102665219,-0.102917783,-0.2107793689,-0.2703860998,0.0307912007,0.3211868107,-0.2815464139,-0.3093490899,0.0659012422,0.0390477702,-0.0104021057,-0.1799643934,0.2581602931,-0.2859198153,-0.0678930432,-0.2721945941,0.2000262737,0.0214113835,0.2753878236,-0.1119688749,-0.0004646548,0.0060164444,-0.0009940049,0.0340764895,-0.4372750223,0.3914862871,0.0934399664,0.0480776541,-0.0987748578,0.1128349006,-0.0641058385,-0.0430664085,0.6849476695,0.174444288,0.1752889305,0.2168378532,-0.1809906363,-0.1927255094,-0.0909492597,-0.3817342222,-0.5966464877,0.1252205223,0.091446735,-0.2169649005,0.0871368274,-0.2984280586,-0.2251374125,0.1321952343,-0.1171602979,0.0286836047,-0.0901559293,0.1095235199,-0.2676126361,0.1917374581,0.3740238845,-0.3153310716,0.0494485535,-0.0734258741,-0.2135497779,0.1762744784,0.3705996275,-0.3056227863,0.0591847859,-0.3774471581,-0.136213392,0.2382946163,-0.5621118546,-0.4215043485,0.5992712975,-0.5369255543,0.128500253,0.3748978674,-0.0943396538,0.0501166098,-0.1642115712,-0.0010971364,0.2005169243,-0.1767111272,-0.0859080553,-0.0647001714,-0.2305059731,0.1496224552,0.0512239337,0.1470538676,0.2475226521,0.4015806615,-0.3437528312,0.2249791175,0.079838559,0.0593375489,0.3031328619,0.1127032191,0.1874863952,0.0576475039,-0.067809023,-0.0180609412,0.1433225274,0.0969163179,0.1331735998,-0.4544298351,-0.2827016711,-0.3851602674,0.0597076491,-0.2557530999,-0.1976463795,-0.0729224384,0.1620244682,0.198942095,0.0629702955,-0.1394384503,0.2614642382,-0.2417074144,0.1346646249,-0.3927445412,-0.0052875746,-0.3166234195,0.2521175444,0.0628735945,0.2250802964,0.1026251689,-0.3936483264,0.0388784222,0.3471484482,-0.1636752337,0.2459466904,0.4471339285,-0.0768294036,0.1481189728,-0.2535534203,-0.0019223186,0.0207380857,0.1652927101,0.0995788425,0.2445020974,0.4189689457,-0.1381774992,0.3534336388,0.2430633456,0.1018153727,0.4970537424,0.1846902519,-0.1638279408,-0.054910671,0.1627466977,0.2569490969,0.2440239191,0.0510584787,-0.3063215315,0.0358605981,0.3714100122,-0.206137538,-0.1045365036,0.2198976129,-0.0375454277,-0.1541121304,0.268955797,0.325432837,0.2397065014,0.1287347227,0.0073773777,0.008486758,0.0245217718,-0.2758022547,0.2178353369,0.0225247871,-0.0848600641,-0.0997317284,-0.1577929407,0.0239043329,-0.2462339252,-0.2599892616,-0.0818862766,0.1003525481,-0.2951285243,0.0956832096,-0.1791385412,-0.2106065005,0.03055094,-0.1278257221,-0.0708777234,-0.0244658906,-0.2541636229,-0.0768772364,0.0034975037,0.0543405041,-0.1177082434,0.0808912441,-0.0000808615,-0.2572306097,0.0149193574,-0.1822189689,-0.116703853,0.0802343041,0.0735346079,-0.2460044324,0.386069119,-0.3160516322,-0.1374778301,-0.2647013366,-0.2660467327,-0.0025329294,-0.2927845716,0.1916858107,0.1971205026,0.570371151,-0.0727637038,0.1112002879,0.1009879559,0.1587015539,-0.0927196443,0.0947242975,-0.1793780178,-0.3052485585,-0.02936331,-0.4112699926,-0.6069230437,-0.2414167225,0.5490335226,0.0594549216,0.1699184328,0.1289374679,0.1287506521,0.2001085728,-0.0949767977,0.189691335,-0.1098501757,-0.7366933227,0.1865401417,-0.2187229693,-0.4208389223,0.0948563516,0.1742120683,0.1524999738,0.1145450398,-0.4824326932,-0.0435416549,-0.2732465565,0.5928384662,0.228684023,0.2681688368,0.2591039538,-0.1083404943,-0.0190776773,0.0581811406,-0.1092537642,-0.1205402464,-0.0330048017,0.0799576566,0.2373790443,0.6281406879,0.0022583799,0.8966236711,0.1433395147,0.2809495032,0.3366571963,-0.1672466844,0.6787809134,-0.1122263819,-0.3695983291,0.1828486174,-0.1843002141,-0.0341747403,-0.0349531062,0.0113930423,0.1826461703,-0.3542369902,-0.2172434181,-0.4547280669,-0.013186791,-0.2512990534,-0.0439465977,0.3033425808,0.3213041425,-0.2542965412,-0.0445461087,-0.0465730838,-0.1392973959,0.5175519586,0.3043706119,0.1067072079,-0.1937413365,-0.4147574008,-0.4655102789,0.1850390285,0.1652392745,0.30336532,-0.2039225847,0.1496023834,0.1451834291,-0.0264938027,0.4634089768,-0.0855552405,0.2608711123,-0.2503010035,-0.2143211365,-0.2421769053,0.0859914348,0.1358066648,0.3931620419,0.4021693766,0.3111833632,-0.2517565489,-0.0634317398,0.240328759,0.4169409871,0.0360831544,0.0322946906,-0.4389885664,-0.4671419263,-0.3043808341,0.0443476848,0.0750485733,-0.0327414572,-0.0796365961,-0.3801339865,0.139790684,0.0495755635,-0.0849317536,0.0108677456,-0.17567195,-0.1639826,-0.0881423429,0.0292921886,-0.0119901523,0.2336805314,0.8473033905,-0.054331217,-0.4416351616,0.1069408506,-0.0873418301,0.2426243573,0.4287479222,-0.3434511423,0.2248141617,0.2938658297,0.1941403598,0.0278723352,-0.0889987126,0.2278602123,0.2560952306,-0.2686398923,-0.5812897086,0.23685202,-0.1923528761,-0.0798066482,0.1616540253,0.3225770295,-0.0231527556,0.0887571871,-0.3266094327,0.9755817056,-0.061897736,0.1599707752,0.2736922204,-0.2049049884,0.6905869842,-0.075292334,-0.0260114335,-0.0312329996,0.0121632572,-0.2138667107,-0.1836461872,0.3075832129,-0.1075575203,0.1308748871,0.5412428975,0.1051567569,0.279381603,0.2818588018,0.4345101714,-0.1370307803,0.1518081725,-0.2523875535,0.0112934811,0.0890046135,0.349209249,-0.0967983752,-0.1812431663,-0.2169582546,-0.4635491073,-0.3159100115,0.0770953447,-0.1700168401,0.2835496962,-0.2618564367,-0.3819081485,0.0394770242,0.2156237513,0.3661672473,0.1271671951,-0.3483726382,0.1041695178,-0.0325814858,-0.2253560573,0.1193992123,0.2196259797,0.0585099868,-0.2676690221,-0.1631458402,0.0912333727,-0.3837051392,-0.4753275514,0.1165167615,0.160683006,-0.0886342973,0.0285576992,-0.2594716549,-0.3641318083,-0.1353180259,-0.2234558761,0.0221921038,0.0332573652,0.0855412334,-0.1595683992,0.1000858173,-0.4792576432,-0.3083316088,0.2961961627,0.101111196,0.0086646555,0.3008598089,-0.0254838429,-0.2978412211,-0.0730056614,-0.1070542336,0.1128735244,-0.3831183314,-0.0554381534,0.0093055787,0.1903038025,-0.2904722691,0.121399492,0.3454098701,-0.1291727871,-0.1401872784,-0.7301217318,-0.2574651539,-0.0440690145,-0.0860196203,-0.046787858,0.2605698705,0.1884119809,0.2191005498,-0.1121384948,-0.1931420416,-0.0352065638,-0.118907012,0.1893437952,0.3981718719,0.128067717,0.404263705,-0.0418101028,0.0963353664,-0.2398322672,-0.2778914571,-0.0938006341,0.1272538602,0.1937197298,-0.015985949,-0.2113343179,-0.1374362707,-0.1894382089,-0.0302952975,0.1373148561,0.1910872906,0.0885146037,0.0768461451,-0.2128359973,-0.3246434629,0.0326537304,-0.1514021456,0.2506736219,-0.4052933455,0.2931370735,0.0419883616,0.1846660674,-0.054700207,-0.062847279,-0.1790301353,-0.1747977883,-0.0010290015,-0.0156443026,0.4725161195,-0.3057200909,0.121736154,0.1238189489,0.3870134652,0.3176066279,-0.0764718577,0.1636552811,0.0920121521,0.0528599098,-0.2982033491,0.1763188392,0.0198563635,-0.0124956919,-0.0869376212,0.0837000757,0.396900773,-0.0511396006,0.2406800836,-0.0074045281,0.7359580398,-0.03519978,0.098652795,0.3116627634,0.2028191686,0.2901833057,-0.0074153668,0.3918151259,-0.0961792171,0.2838306725,-0.2792786062,-0.0649529919,-0.0066281441,0.2520179152,-0.2160855383,-0.4989340603,0.5282220244,0.1405059099,-0.4128902256,-0.222704038,0.2649854422,0.4716642201,-0.3868049383,0.1617426574,0.0028840969,0.1396621913,0.0008569877,0.0548565499,0.0867649987,-0.0528747812,0.0536498837,-0.0138530126,0.0497173369,-0.1499880403,0.073226057,0.0829765499,-0.1660098881,-0.4901471734,0.0985208154,-0.0611774027,0.0640603229,-0.1050400138,0.2438178062,0.0441352986,-0.1199348047,-0.0564587675,0.2529461682,0.6069973111,0.2839480042,0.1196365505,0.0438911021,0.1174877509,-0.2431918979,-0.0309548173,0.238539353,-0.0226951689,0.3810318708,0.2282349914,0.136320129,-0.0333245024,0.2987473905,0.0323636793,-0.0079726204,-0.2900753021,0.2597250938,-0.2596825361,-0.3766297698,-0.1175566614,0.1649689674,-0.512032032,0.1394943446,0.2924079299,-0.1979348361,0.0157302674,-0.4259531796,0.034939602,0.0278183352,0.2674633265,0.7155258656,0.4606139362,-0.3348760307,-0.1098515168,-0.5381643176,0.1705892384,-0.2221709341,-0.1982749254,0.0511964038,0.0114330174,-0.1057403609,0.3257549405,-0.0462971963,0.1087363064,0.0759241581,0.0574222989,0.1789706051,-0.0208211113,-0.1078761593,0.1753511429,-0.1996838301,-0.3156131208,0.4249691665,-0.3957900405,-0.047964476,-0.1112158149,0.2341019511,0.1720631868,-0.3405556977,0.4727347493,0.0987477079,0.4923421443,-0.1560232192,-0.1178598329,-0.1519267112,-0.0938045755,-0.1615957469,0.2455444038,-0.0167906769,0.006897511,-0.1859281063,-0.2319275737,-0.2419971824,0.3104193509,0.0172113627,0.0564077869,0.2882194221,-0.1637437344,-0.0938537121,0.0140685169,0.2250897884,0.0133415051,0.0627988055,-0.0355007648,-0.2181967199,-0.3019966185,0.4021445811,-0.4057481289,-0.039643649,0.1320572793,0.3493552208,0.1719122678,-0.2641636431,-0.2059079558,0.180251345,0.3741888106,0.119712837,-0.3547089696,0.2156614065,-0.2529556751,0.1879665852,-0.0445609763,0.3003157377,-0.1705958843,-0.3227390945,0.0059781978,-0.2893488407]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1771","title":"Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py","comments":"Thanks.","body":"Hi,\r\nWhen I load_dataset from local csv files, below error happened, looks raw.githubusercontent.com was blocked by the chinese government. But why it need to download csv.py? should it include when pip install the dataset?\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\r\n\r\n```","comment_length":1,"text":"Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\nHi,\r\nWhen I load_dataset from local csv files, below error happened, looks raw.githubusercontent.com was blocked by the chinese government. But why it need to download csv.py? should it include when pip install the dataset?\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/home\/tom\/pyenv\/pystory\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/csv\/csv.py\r\n\r\n```\nThanks.","embeddings":[-0.1563563347,-0.2410446554,-0.1811293811,0.1131328642,0.2398847938,-0.0027395028,0.2845861614,0.1787883043,0.2420271933,0.2321637124,0.0800503269,-0.2211632282,0.2691062391,0.1082019135,0.2388585359,-0.0329398215,0.0347485319,0.2014846504,0.0314682983,0.0701691285,-0.1909445524,0.1178895384,-0.062379837,0.2506891787,0.1569260508,-0.0474651195,-0.0108445846,0.2422563434,-0.1564468443,-0.6826015711,0.5021808147,-0.124328807,0.0664196163,0.3840642869,-0.0001212761,0.195008561,0.2078873366,0.0367582925,-0.3090150058,-0.3778352141,-0.3193042874,-0.4471861124,0.1608800441,-0.1218490154,-0.1489934027,0.0815711245,0.0441984534,-0.015303025,0.5730929375,0.3271841407,0.1240221262,0.405208528,0.2140662372,-0.0336832218,0.03783454,-0.318513751,0.0403370783,0.6891165376,0.2261292934,0.1207953691,0.2535718083,0.0732692182,-0.0057731983,0.0760470852,-0.0393702462,-0.2209563404,-0.0236276872,-0.3864572346,0.1667194217,0.1775799543,0.4139237702,-0.0088118305,-0.3937250674,-0.1248834357,-0.0870620161,-0.2423586398,0.2039305568,0.2825602889,-0.2109403461,0.3968256414,-0.4112616181,-0.4344572127,-0.2178604305,0.0771312639,0.1007027552,-0.0784156322,-0.2600668669,0.2547100484,-0.1269769371,0.2399259955,-0.1094031855,-0.2277435362,0.197788924,0.2276975662,-0.1303789318,0.0439136401,-0.0026103593,0.1807803363,0.0433765352,-0.0226208866,-0.1742221117,-0.0707899854,0.1041015163,0.2132114917,0.2166526765,0.0389494337,-0.2151042819,0.274382025,0.3818828464,0.0667400062,0.2020057291,-0.0196833424,-0.1291785836,-0.1601676047,-0.2559386194,0.0283328537,0.3095073402,-0.2692318261,-0.314899236,0.0758490339,0.0159332938,0.0065445798,-0.1900276393,0.2580598891,-0.2814117372,-0.1277806312,-0.2878868878,0.2453986257,0.0354341567,0.22096847,-0.1330271512,-0.0435664877,0.0221700128,-0.0178786963,-0.0350486934,-0.3842558563,0.3779615164,0.0493890569,0.0001141679,-0.0616199449,0.1814273298,-0.0579120964,-0.0680320263,0.6628061533,0.2091829032,0.1510518044,0.2957091928,-0.2643514872,-0.1658450663,-0.1173788831,-0.3967662454,-0.6227531433,0.1096610278,0.1191057339,-0.2118610442,0.1097154543,-0.3408189416,-0.2474473268,0.1491195112,-0.0562659353,0.0393288136,-0.0932592526,0.1317590922,-0.2449247539,0.1913319975,0.3774902225,-0.2372602224,-0.0483660959,-0.0483065657,-0.2149384469,0.1883459538,0.3624522388,-0.3096221089,0.1295927465,-0.3965184391,-0.154607892,0.322142303,-0.5501464605,-0.4503738284,0.6174302697,-0.5499893427,0.0764865428,0.3262780011,-0.1211191341,0.0318005569,-0.1472374052,0.0938146636,0.2516174614,-0.1886323243,-0.0795569792,-0.0843984336,-0.2182158232,0.1246293113,0.0720992461,0.1635981202,0.2408949137,0.4116427302,-0.343909055,0.193838805,0.0990219712,0.0327936932,0.3425993621,0.0766472146,0.1573987156,0.0013732112,-0.0199630708,-0.0014700809,0.1453058124,0.0476295836,0.1239325479,-0.4886389971,-0.2294130325,-0.4388480783,0.0706830174,-0.2971651554,-0.1853473336,-0.0316834711,0.1286958605,0.2091641128,0.0917337686,-0.0563899912,0.3039813638,-0.2573672235,0.1003708616,-0.3347401917,0.0564479046,-0.3561032712,0.2625222206,0.0472210646,0.2086967677,0.1212686002,-0.3286180198,0.0078503061,0.3308479786,-0.2512446344,0.2226369679,0.4563417137,-0.1022335961,0.1160324514,-0.3269078434,0.0241927896,-0.0094890576,0.1327425838,0.045856867,0.3202397525,0.3244204819,-0.1152649373,0.3238616884,0.1870807409,0.1242238879,0.5205080509,0.1482996792,-0.1322966516,-0.0713457018,0.1571677774,0.2967028618,0.2378788292,0.0494597852,-0.2401417792,0.0199237596,0.3322926462,-0.1808426529,-0.0578296967,0.195104301,-0.0556584112,-0.1309705228,0.236399129,0.2932893336,0.2368123084,0.1139227375,0.0404382013,0.0090050576,0.0557731055,-0.2825939357,0.240822956,-0.0126956552,-0.1010536551,-0.0682009384,-0.1943629384,-0.0208165124,-0.1804379821,-0.29311046,-0.0407089666,0.123277083,-0.2971437275,0.0508204997,-0.1407281905,-0.2402473539,0.0279242676,-0.1628529876,-0.0814508274,-0.0338485166,-0.230400607,-0.1537219137,0.0521970727,0.1083320528,-0.1496403962,0.0409547873,0.0414725393,-0.2712599635,0.0695844367,-0.232189104,-0.1159310192,0.0930081978,0.0540635251,-0.223592788,0.4110139906,-0.3542580605,-0.1335813254,-0.2553883195,-0.1683229953,-0.0274576023,-0.286891073,0.2002361119,0.1455159336,0.6579937339,-0.0453043394,0.1040818319,0.1839255095,0.2074201107,-0.0697040409,0.0897600874,-0.1727247238,-0.3257085383,0.0379322805,-0.3272730708,-0.6531930566,-0.2733005285,0.4685191214,0.0573473424,0.1684782356,0.1253144741,0.155997619,0.1898667514,-0.0465967357,0.1742492765,-0.1183921993,-0.7758178711,0.1586947441,-0.178802669,-0.3863557875,0.1101233661,0.1376292259,0.1536136121,0.1221759245,-0.4742228389,0.002496423,-0.2639468014,0.5874404907,0.1861316115,0.2543068826,0.2282167822,-0.1291827857,-0.0166110843,0.073265858,-0.0586226545,-0.1323651373,-0.0738500655,0.137080133,0.2050496489,0.5787892342,-0.0285734311,0.8792691231,0.1427954584,0.2481355667,0.3327168226,-0.1632292569,0.6667252183,-0.0800235271,-0.392901063,0.1755370498,-0.1255291849,-0.0296625216,-0.0709263384,0.0012178855,0.2009396106,-0.3691460788,-0.2303314358,-0.4523937404,-0.0058089215,-0.2615118623,-0.0720785931,0.3117090464,0.2974015474,-0.311224997,-0.0659328923,-0.0178491883,-0.1294194311,0.4773781598,0.1977323592,0.0943106636,-0.2346531898,-0.4133125544,-0.4272956252,0.2215266228,0.1954937279,0.2756316364,-0.1981257498,0.2051556259,0.099942565,0.0110830097,0.4422719777,-0.1378234029,0.2928468585,-0.214884907,-0.120629549,-0.2277940214,0.0448108017,0.1538145095,0.4195612669,0.4078160524,0.312140882,-0.272636205,-0.0283689667,0.2224621773,0.4071061015,-0.0126211084,0.04576464,-0.4362190068,-0.4283210635,-0.2755347192,0.0634057298,0.0723451972,-0.0211935733,-0.0469348244,-0.3466608524,0.1102662086,0.1094178706,-0.1287613213,-0.009941346,-0.1559693366,-0.1381973475,-0.119793877,-0.0318543762,-0.0295707956,0.2504213154,0.8786793351,-0.0580577403,-0.3424161375,0.0741127282,-0.0912261605,0.1670088321,0.3967500329,-0.2763954699,0.1841987818,0.2979236841,0.2527126074,0.0663305894,-0.0245936401,0.1768866032,0.2306191623,-0.2901761234,-0.5492662787,0.2404350191,-0.2380925715,-0.0536359996,0.2040012777,0.2998479903,-0.0269165616,0.0895176604,-0.2952456176,1.0041339397,-0.065616481,0.136513412,0.2704317868,-0.1611944586,0.6128191352,-0.0829902887,-0.0637702644,0.0171539132,-0.0045962944,-0.1944845617,-0.2233084887,0.2694252729,-0.0990077481,0.0800690576,0.5033233762,0.0916633978,0.2775499523,0.308816582,0.3981868327,-0.1333500892,0.2132301182,-0.2574107647,0.0288125388,0.0425834768,0.3165706694,-0.1440594941,-0.1596969217,-0.239719376,-0.458750844,-0.3418045342,0.1379304677,-0.1841132343,0.3175750673,-0.2792279422,-0.4069951475,-0.0007253564,0.22824426,0.3756903708,0.1482485831,-0.3653871119,0.129101634,-0.0175604112,-0.2429400384,0.1046423465,0.2334127426,0.0613537505,-0.2549017072,-0.2159293741,0.1547511965,-0.3576308489,-0.4669434726,0.1084725931,0.170192048,-0.0914606154,-0.0021193172,-0.2638569176,-0.3461528122,-0.0879211053,-0.1799425334,0.030639654,0.0392214991,0.0520851202,-0.1712442487,0.1317398101,-0.4931274354,-0.2746449709,0.3062804043,0.1070964932,0.0410499014,0.3522435427,-0.056983579,-0.2602390945,-0.0950848088,-0.1091675982,0.1072692052,-0.3487247527,-0.0788603723,-0.0476717688,0.1918364316,-0.3320611417,0.0986317545,0.4022350907,-0.0879579633,-0.024114307,-0.7451732159,-0.2453397214,-0.0692292526,-0.107754305,-0.0442331582,0.2149512321,0.2017235756,0.2420507669,-0.0646556169,-0.2029934824,0.0219058692,-0.2457155138,0.2196038514,0.3621726334,0.1098583043,0.436467737,-0.0684668794,0.122667335,-0.2876788974,-0.2927344143,-0.1256465316,0.1290528774,0.1825916171,0.0202701371,-0.245195955,-0.1479105502,-0.2126059383,0.0409952439,0.1643772721,0.143314302,0.085078083,0.0773667321,-0.1881577075,-0.248686403,0.0473729223,-0.2275074869,0.2529829741,-0.3885369599,0.3332722485,0.027886413,0.146790728,-0.033614058,-0.0522441603,-0.2226280868,-0.1914927512,0.0350773893,-0.0075981705,0.521386385,-0.3365167677,0.1519975215,0.1285645664,0.3702539802,0.2712720931,-0.1577590108,0.1529960185,0.1182811409,0.0741926879,-0.2926197648,0.1714193523,0.0904668048,-0.0471933372,-0.0958996639,0.0352554321,0.3674223721,-0.0960816666,0.2164404541,0.0093872966,0.7377848625,0.0232635811,0.0603610426,0.2858811915,0.2133389264,0.2517284751,-0.0630956218,0.394911617,-0.08854305,0.2199291587,-0.2709970474,-0.1198975816,-0.0254829023,0.2224649042,-0.2280920893,-0.4984095097,0.5330371261,0.1039429381,-0.431684345,-0.2525219917,0.2415466756,0.3706515729,-0.3368387818,0.1551490575,0.0302791651,0.1291109771,-0.0002062176,0.0916635916,0.0673651621,-0.0523679852,0.0203774814,-0.053384725,0.0914296359,-0.1737630367,0.0167778693,0.1263026744,-0.1612961888,-0.4427237213,0.1621576995,0.0353332944,-0.010664722,-0.1689449549,0.2359220833,0.0982304215,-0.1622467041,-0.0456437506,0.2388789505,0.6160343885,0.2606177032,0.0690085664,0.0808382779,0.2045813203,-0.285531491,-0.0439752601,0.1923816949,-0.0101106316,0.3008985817,0.2120005488,0.1704494655,-0.0708175004,0.2995329201,0.0359134525,0.0130427368,-0.2442528456,0.2476310879,-0.2403681874,-0.4320491552,-0.1248059422,0.1215320304,-0.5245003104,0.2075121999,0.2927386761,-0.217170462,0.0492431,-0.4117721915,0.0348943062,-0.0044279615,0.3315515518,0.7685950994,0.4923595786,-0.3622516096,-0.123109214,-0.4990955889,0.1866999418,-0.2880333066,-0.223513186,0.0937663838,0.0183279924,-0.1059880257,0.3493594527,-0.0651488081,0.166561082,0.0607317165,0.089238964,0.1272797585,-0.0864477232,-0.0640148371,0.2273728997,-0.1788778007,-0.3374184668,0.4312672019,-0.4067108333,-0.0141495634,-0.0619964413,0.1611232162,0.1978188157,-0.3219062984,0.4529937506,0.1023837775,0.5117699504,-0.1101443097,-0.0798865184,-0.0701899007,-0.093353115,-0.1002574489,0.2441064417,0.0222982727,0.0008530268,-0.1944595724,-0.228142038,-0.2637334466,0.2944278717,0.0139155667,0.0475700758,0.2765705585,-0.1999108046,-0.04669879,0.0304141212,0.2370665073,-0.0042228336,0.0702696964,0.0541142672,-0.1693628728,-0.2568135262,0.3855921328,-0.3769752681,-0.0582786202,0.1506350636,0.3428703845,0.1617583781,-0.2087308913,-0.1579333544,0.2158634514,0.4425744116,0.0939289853,-0.383921355,0.2166905999,-0.1837735623,0.2404706031,-0.0177746899,0.2965270877,-0.2342969775,-0.2982748747,-0.0258561857,-0.2747751474]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1770","title":"how can I combine 2 dataset with different\/same features?","comments":"Hi ! Currently we don't have a way to `zip` datasets but we plan to add this soon :)\r\nFor now you'll need to use `map` to add the fields from one dataset to the other. See the comment here for more info : https:\/\/github.com\/huggingface\/datasets\/issues\/853#issuecomment-727872188","body":"to combine 2 dataset by one-one map like ds = zip(ds1, ds2):\r\nds1: {'text'}, ds2: {'text'}, combine ds:{'src', 'tgt'} \r\nor different feature:\r\nds1: {'src'}, ds2: {'tgt'}, combine ds:{'src', 'tgt'}","comment_length":45,"text":"how can I combine 2 dataset with different\/same features?\nto combine 2 dataset by one-one map like ds = zip(ds1, ds2):\r\nds1: {'text'}, ds2: {'text'}, combine ds:{'src', 'tgt'} \r\nor different feature:\r\nds1: {'src'}, ds2: {'tgt'}, combine ds:{'src', 'tgt'}\nHi ! Currently we don't have a way to `zip` datasets but we plan to add this soon :)\r\nFor now you'll need to use `map` to add the fields from one dataset to the other. See the comment here for more info : https:\/\/github.com\/huggingface\/datasets\/issues\/853#issuecomment-727872188","embeddings":[-0.3618090153,-0.4964470565,-0.0687218681,0.1217603683,0.0697285309,0.3581112623,-0.092245914,0.2210621834,-0.0352847092,0.0138700567,-0.3246749341,0.4096489549,0.0693914667,0.7398912311,0.1613286585,-0.2478816509,0.078949064,0.129455775,-0.4716953337,0.1984139681,0.0128847482,0.0042593097,-0.0246831551,-0.0849265978,-0.0984388143,0.2673225105,-0.2330110967,-0.1992989331,-0.1573341936,0.0235670432,0.3827166855,0.2693840265,-0.0531156473,0.2027809918,-0.0001115768,0.0722315609,-0.0460229665,-0.1208759993,0.1077825353,-0.4201931953,-0.0392919108,-0.5601700544,-0.1074662507,-0.0181255024,-0.1781563014,-0.2882148921,-0.2893857956,-0.2174081355,0.3418446779,-0.1690038443,0.1079963818,0.1024667546,0.1639198214,-0.0635811985,0.2932220101,0.4440595806,-0.0876196325,-0.069648549,0.4562326968,0.0492846109,0.5520344973,0.1969706416,-0.019789217,0.0005702212,0.2822974622,0.1590328515,-0.2205993831,-0.20029594,-0.0121477181,0.3940691054,0.4385496378,-0.2857182324,-0.3185835481,-0.189653784,0.1420851201,-0.1715218872,-0.1700829864,0.3175992072,0.0211353395,0.02320919,-0.4666415751,-0.3988069892,0.0761492923,0.1548091769,-0.0889578015,0.1069054008,-0.0845471174,0.1540905237,-0.0578114167,-0.2493236959,0.1112819239,-0.4102554917,-0.089173317,0.2716374397,-0.2897037268,-0.3540632129,-0.2580055594,-0.2053414881,0.7211526036,0.1921920478,-0.1244395897,-0.131403178,-0.4426646233,0.2014518231,0.2647921741,0.2856196463,0.2340371311,0.0185884591,0.0427907445,-0.2273828536,-0.131199345,-0.0583620593,0.4423027337,-0.1362012029,-0.1892872453,-0.2544163465,0.4779630005,-0.2223193645,0.1482369304,-0.4064995944,0.0526720248,-0.2132237703,-0.1734143347,-0.0509308726,0.0698622018,-0.0080112331,-0.2046570927,0.2878879905,0.2749117911,-0.2954862416,-0.1059869528,-0.0737898126,0.1206281409,0.1911255121,-0.012508153,-0.3862589896,0.1450714916,0.6239692569,-0.1406674236,0.2539289892,0.3124404848,-0.1426234841,-0.1376311481,-0.1937665194,0.3240775466,0.1543080658,0.0961580873,0.1441503167,-0.2133418322,-0.0007746207,-0.3544187844,-0.1190918982,0.0511283167,0.1720260233,-0.1338262707,-0.1185326949,-0.3662887216,0.669705689,-0.0450399779,-0.1963850111,-0.1601148397,0.4825178385,-0.3460585773,-0.1729307175,0.1248904914,0.0466939658,-0.2356795967,-0.4060325921,0.1128408611,-0.0408670418,-0.2736009657,0.4258361459,-0.2547849417,0.0148463221,-0.2920529544,0.0254394338,0.4448811114,-0.0706946105,-0.2393836975,0.4166375697,-0.0338701159,-0.0262855366,-0.0796054527,0.1589786559,0.1578690559,0.1973277479,0.0528209619,0.5103905201,-0.2451841384,-0.2101137489,0.0171056483,-0.4156886935,0.4063560963,-0.0769312903,-0.0820079893,0.0215606131,0.0277162995,-0.4218937457,0.1445216238,-0.15991126,0.2335322052,0.1152569503,0.0342923552,0.0504973084,-0.0890547559,-0.4413529634,-0.4822226763,-0.0164514408,-0.1436119974,-0.1910775304,0.1615774632,-0.4283260703,-0.0080508031,-0.3666362762,-0.0625188798,-0.1186320409,0.0343080238,0.0187574625,0.1252595782,-0.1757806689,-0.2722326815,0.2241860926,0.5275174975,0.0491098091,-0.3530474305,0.5951489806,0.240128696,0.0126929348,-0.0470200293,0.1135402992,0.0824418664,-0.0154712535,0.1004678532,0.0125805149,-0.5362690091,0.3304072917,0.4451425374,-0.1900831908,0.2659787536,-0.2425027937,0.1363273859,-0.0252018459,0.0164538473,-0.2492333651,-0.3427091539,0.3071554005,0.0362624452,0.2554853857,0.1160895601,0.0826603174,0.1574560553,-0.1364686638,0.047816705,0.0063440022,-0.011840228,0.0121240411,0.3719067574,0.2906228304,-0.1891371459,0.1635454446,0.254083246,0.0006374083,0.0234006979,0.0033061143,-0.3710803688,0.0487012006,0.0733456686,-0.0457924716,0.5803026557,0.3118932247,-0.0138465557,0.1654967964,0.1007250622,0.1608840525,-0.0108852163,-0.0204664636,-0.0809135586,0.1754918993,0.307376802,0.1203590706,-0.0418767892,0.2153899223,0.1384509355,-0.1983055174,-0.0339096487,-0.2077614367,-0.2305031568,-0.2206584811,-0.3354868293,-0.1023770049,-0.4055423439,0.0058899112,-0.0967234746,0.1889790297,0.067526117,0.0023474752,0.2830423713,0.2402458936,-0.1767149866,-0.1294568777,-0.0370173678,0.2874001563,0.2032704949,0.1446807086,0.154216826,0.101480864,0.38425529,-0.0276632365,-0.0555931553,-0.6561787724,-0.3905450106,0.1056405306,-0.1866333336,0.16265136,-0.0132696228,0.0612567216,0.0317173265,-0.2150374055,0.2902244925,0.3680696785,-0.1453747004,-0.1525584161,-0.0711419731,0.0424222611,-0.153492406,-0.2626842856,-0.2550150454,-0.130769223,0.4938812852,-0.2122130543,0.0253909901,-0.3388878107,0.0590585954,-0.145503521,-0.0381393991,-0.0373349972,-0.2105336636,-0.4417386055,0.3633479774,-0.1019067094,-0.1232514456,-0.0622192658,-0.1185194403,0.0129733961,0.0582186133,-0.0850687996,-0.1113943607,-0.3798117936,0.2991844714,-0.0581444874,0.3092561066,0.2840961516,0.0622070655,0.1154543906,0.028177442,-0.4375109076,0.0236828197,0.0776255801,0.2953591049,0.0861718655,0.0952155665,-0.1847838014,0.5392614007,0.4855138361,0.2069980949,0.314810276,-0.0207610782,0.2014724761,-0.1138301343,-0.3359416127,-0.15461725,0.0052298256,0.2537072599,0.0774524957,0.1430764347,0.1139905229,-0.3639834821,-0.3012901545,-0.0190328397,-0.1052176505,-0.0961650908,0.0764964744,0.0234005917,-0.1047778428,-0.4167267382,-0.2212542742,0.0701889694,0.1113704666,0.3068456948,-0.0301218871,0.076307334,-0.4565531909,-0.1959426403,-0.2196298242,0.243543312,0.1135359928,-0.3452657759,0.1435208321,0.0839312077,-0.2891019881,-0.0461752899,0.4688130319,0.0044267145,-0.1657568216,-0.2415521592,-0.3337163329,-0.1596201658,-0.1541397721,0.2101828158,-0.1223801523,-0.0428109579,0.7450358868,-0.4150353074,-0.1063113213,0.2144472152,0.3093689978,-0.0340210944,-0.2858569622,0.2248763889,-0.1265195608,-0.2038751096,-0.3736388385,-0.0918770432,-0.0946601108,-0.0005786291,-0.3123131692,-0.225540176,-0.1106936112,0.4122420847,0.093133375,0.1979771405,0.0016144878,0.3067940474,0.1092078537,0.1054374576,-0.1184780896,0.7144662142,0.092373468,-0.5906254649,-0.3595565557,-0.1414742768,0.4458451867,-0.154121086,-0.0330142453,0.497327745,-0.3980501592,-0.031287834,-0.1536016613,0.0443198197,0.2229009122,0.3378208578,-0.3514523804,-0.2078111768,0.616351068,0.4150248468,-0.0853637531,0.2741283774,-0.1544028223,-0.0853123888,-0.0610344745,-0.4421529472,0.7151219249,0.0019197379,0.2235155106,-0.1618157327,0.0517005399,0.4223560095,-0.0141938003,-0.1407146156,0.0445346572,-0.4924365282,-0.1345448047,-0.0360832624,-0.0127558419,0.2731688321,-0.3648453653,0.387542367,0.0279473346,0.3813567758,-0.1812819242,0.1914039701,0.0356788486,0.0019292096,-0.2923057079,0.0774030462,-0.3082001209,0.2018806636,0.0318570435,-0.1815778613,-0.0393772833,-0.2110555619,-0.1716475487,0.115020819,0.0215130877,0.1849931031,0.2158883959,-0.1307343543,0.086160928,0.1218130514,0.1773827076,0.1479754746,-0.1574471444,-0.1854998171,-0.0523377918,0.1065877303,-0.3663006425,-0.1125913113,-0.0478987508,0.0185200833,-0.1067312434,0.2390482128,0.2548415065,0.0043043257,-0.1485947222,-0.0203753635,-0.0757183731,-0.3721718788,-0.306765765,0.4104664028,0.2765319347,-0.1722232103,0.0149113936,0.1202457175,-0.1559036523,0.1447322667,0.0585457683,-0.1185307056,-0.1087363362,0.0966595188,0.2115072012,0.0077756401,0.1847671419,-0.0656261072,0.2368045896,0.0324388109,0.1078701168,0.1853032261,-0.0826606974,0.0941469967,-0.0528572276,0.0864880383,-0.0487322845,0.2020090818,0.1860230714,0.3130823374,0.242110163,-0.1659176797,-0.3828411102,0.1942708492,0.1468085945,0.2859948575,0.0524921678,0.2371541709,-0.0772296265,0.5865795016,-0.2435095608,0.1217966899,0.2938934267,0.1696381867,0.4679046273,0.092002973,-0.1409837008,-0.0514323637,0.0716399625,-0.111687474,-0.0237845443,-0.2098345459,-0.1993460208,0.0840243995,0.13052167,-0.036022298,-0.0042906413,-0.0485143214,0.3371495306,-0.1892850995,0.4083261788,0.0316528268,0.0519720465,0.0441098064,0.1356363893,-0.1600202173,-0.4557378292,0.3019766808,0.0825198814,-0.0855898187,0.0340582281,-0.0427307598,-0.1179133654,-0.1373353004,-0.2232819349,-0.1955260485,0.4275312722,-0.0626697093,0.0048006107,0.2037334442,-0.0406900868,0.1561904252,0.1611381173,-0.0074360375,-0.1330078542,0.4312365055,0.0191973597,0.1314160377,-0.2810477316,-0.1722833514,0.2121521384,0.0001842297,0.1162934825,0.0445541479,-0.0169120729,0.3816171587,-0.1187453568,-0.024992479,0.0680543408,0.2002497762,0.1851522923,0.4368991256,0.1196716651,-0.0775462091,0.0855028108,0.3447013199,0.1528349221,-0.0628436208,-0.2231145352,0.0950922891,-0.0355563089,0.009563107,0.6225680113,-0.1861011386,0.0067541641,0.1668475717,0.0266050436,-0.0159043148,0.1713163406,0.3761461377,0.0212113392,-0.4307376146,0.1032129303,0.5190813541,0.1420276612,-0.2155852467,0.1240864322,-0.2345855534,-0.1258781105,0.1448838711,0.1114491895,-0.5025008321,0.7270304561,-0.0855778232,0.0586719327,0.0205249507,0.3626752496,0.3423205316,0.0683147684,0.0319002979,0.1234925613,0.2584229708,-0.3591663837,-0.0128001655,0.4288056791,0.1311598718,0.0650766343,0.1156101003,0.3647965193,-0.0392975435,-0.1133828536,0.152416721,-0.1405665129,-0.1483445317,-0.2890030146,0.1610236168,0.1670106053,-0.0702814609,0.0520771705,-0.5179771781,0.1358426511,-0.6229063272,0.2271407098,0.0051208539,-0.2663627863,-0.123784028,-0.1238271371,-0.2615607083,-0.4799641371,0.5373873115,0.1291896403,-0.0313791744,-0.0803097636,0.0965203643,-0.0629199892,0.0792929828,0.3690905571,-0.0724233016,-0.2462380826,-0.1202219501,-0.4041544497,-0.0201995857,0.0593257137,-0.0565922745,-0.2689474225,0.0749184713,0.1075350046,0.0945207328,0.1419161856,0.0593484081,0.3613251746,0.0193935912,-0.3074978888,-0.1817618161,0.0881310031,0.0771058351,0.243537128,-0.1472337097,0.3143064082,0.1537710726,-0.0558653064,-0.1373118162,0.245180577,0.0664099008,-0.3503929079,0.1685199291,0.1277627349,0.3754206598,-0.3541483581,0.1319274455,0.4446251392,0.0321799926,-0.3205437362,0.2490745783,0.1502970904,0.0839509889,-0.4384068251,0.1390394717,0.1322807372,-0.0707605034,-0.0463987216,-0.4495652318,-0.2335360497,0.185085997,0.0243403874,-0.0596530922,0.1690426618,0.52406919,-0.0794886053,-0.0281582773,0.1459885538,-0.1959228814,0.4356800914,0.0627722144,-0.1813947856,-0.1825110316,-0.2807190418,0.2014479041,-0.1435723901,-0.4898808897,-0.1636502892,0.2929660976,-0.0049858312,0.3040722609,0.3011699021,0.0274858084,-0.1143818274,-0.2004224211,0.3541581035,0.2150261104,-0.2349356562,0.0430080853,-0.2074888647]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1770","title":"how can I combine 2 dataset with different\/same features?","comments":"Good to hear.\r\nCurrently I did not use map , just fetch src and tgt from the 2 dataset and merge them.\r\nIt will be a release if you can deal with it at the backend.\r\nThanks.","body":"to combine 2 dataset by one-one map like ds = zip(ds1, ds2):\r\nds1: {'text'}, ds2: {'text'}, combine ds:{'src', 'tgt'} \r\nor different feature:\r\nds1: {'src'}, ds2: {'tgt'}, combine ds:{'src', 'tgt'}","comment_length":37,"text":"how can I combine 2 dataset with different\/same features?\nto combine 2 dataset by one-one map like ds = zip(ds1, ds2):\r\nds1: {'text'}, ds2: {'text'}, combine ds:{'src', 'tgt'} \r\nor different feature:\r\nds1: {'src'}, ds2: {'tgt'}, combine ds:{'src', 'tgt'}\nGood to hear.\r\nCurrently I did not use map , just fetch src and tgt from the 2 dataset and merge them.\r\nIt will be a release if you can deal with it at the backend.\r\nThanks.","embeddings":[-0.3885659575,-0.4940942526,-0.1406932473,0.0185404997,-0.0224656239,0.2798508108,-0.1746791303,0.3226885796,-0.0510883927,0.0088932943,-0.1925193518,0.6096795201,0.0999665633,0.5465123653,0.0698018521,-0.11272192,0.0745652765,0.1688348651,-0.4552721381,0.1778494567,0.0620454885,-0.0594893433,-0.0334581845,-0.0982747376,-0.0798539296,0.2818122506,-0.1274135411,-0.2967349291,-0.2358781695,0.0145080378,0.2774324417,0.2345999032,-0.1273669004,0.2523550987,-0.0001082497,0.1177333593,0.0513774976,-0.1531057805,0.2523333728,-0.43814677,-0.1970332414,-0.4797204137,-0.1601139903,-0.1272794157,-0.1080754325,-0.1871521175,-0.2989039123,-0.3243273795,0.2503189445,-0.1271040291,0.1325851083,-0.0880205035,0.1223926917,-0.0137145016,0.2999750674,0.363771081,-0.0602158383,-0.1029037163,0.5676066875,0.0092757046,0.4451414049,0.2175048441,-0.1230507195,-0.0073250057,0.3124162853,0.048622299,-0.0380412228,-0.2631063163,0.0710004494,0.2478632182,0.5259116888,-0.1586421877,-0.0991821215,0.0621924251,0.0316509567,-0.0059683854,-0.1653365195,0.2786548734,0.0410577916,0.0446243025,-0.5148202777,-0.44294855,0.0948954821,0.1656849831,-0.0767568573,0.2319708616,0.0322263762,0.1849222928,-0.2162769139,-0.2405795306,0.2670075893,-0.4301855862,-0.0571950935,0.3326957822,-0.212325722,-0.3501248658,-0.2511488795,-0.4439242482,0.6490111947,-0.0200989395,0.0990228429,-0.0340400115,-0.380594492,0.280482769,0.2375020683,0.2917183042,0.3173268139,0.189862594,-0.0087112775,-0.4177415371,-0.3030484021,0.0329587683,0.4209108949,-0.2382072657,-0.1569514126,-0.1717440933,0.4447456598,-0.3677548468,0.2103084624,-0.5202794075,-0.0881638229,-0.1420803666,-0.2262052298,-0.0958222449,0.175053224,0.0489226729,-0.2405053079,0.1676711142,0.3340087831,-0.2938676476,-0.0370179676,-0.0422511697,0.1603026092,0.04718237,-0.0015432743,-0.1771240234,0.0949376151,0.5621643066,-0.3058844805,0.1829916388,0.3147560656,-0.1040003076,-0.0195233803,-0.2901396155,0.3635196686,0.0882699341,0.0345524587,0.2711079717,-0.2068246007,0.1299701035,-0.4372511208,-0.1448739618,0.1334382892,0.2379768342,0.0191794056,-0.0671926066,-0.2039432526,0.5709993839,0.0468189046,-0.2714456916,-0.1705272645,0.5460534692,-0.2374549955,-0.2488662302,-0.0948712006,-0.0343842693,-0.3212696016,-0.3335896134,0.0642851144,-0.0274077561,-0.2287088484,0.3375684023,-0.3624535203,0.0571283549,-0.1392395794,0.0410392694,0.552892983,-0.0943096653,-0.3914399147,0.511829257,-0.007198059,-0.146579057,-0.115699023,0.2643761337,0.2004470825,0.1718614548,0.1076660454,0.5094723701,-0.2747680545,-0.0966094509,-0.0189080052,-0.418153435,0.4239629507,-0.1348892599,-0.0575222224,-0.0474160314,0.1543736905,-0.4866879582,0.1541965455,-0.1080196425,0.2645322084,0.1089467481,0.017614454,-0.1733415723,-0.0762553662,-0.3590360582,-0.3984617591,-0.1078947857,-0.3227751553,-0.1823046356,0.0847267658,-0.3686960638,-0.0110871708,-0.4826446474,-0.0400576256,0.0088729654,0.0902565941,-0.0263376534,0.0050436859,-0.2365912497,-0.156092003,0.1332933009,0.5647348762,-0.0498310141,-0.2055795044,0.5626952052,0.2106457651,0.0336607695,0.0225735214,-0.0472389124,0.1355435103,-0.0224668141,-0.0492446311,0.0592307821,-0.4545077682,0.3014397323,0.3840034008,-0.2035782039,0.2649064362,-0.3029321134,0.142938152,0.0698782802,0.0661109239,-0.2787417173,-0.2938680649,0.2300539315,0.1171721816,0.2025708258,0.0860554427,0.0840895623,0.1174589396,-0.176944375,0.0689538345,-0.047432363,-0.1171294376,-0.030280469,0.2370960265,0.4099989533,0.0255227983,0.2372204363,0.247135371,0.0381263606,0.1396588087,-0.0745834336,-0.3314553499,-0.0346246697,-0.0187520646,-0.0686756447,0.5103157759,0.254093796,-0.1636122167,0.1103900447,0.2492556572,0.2375882268,-0.0140336351,-0.109353289,0.0952235311,0.1363039315,0.3759461641,0.16348809,0.102835685,0.1159216315,0.2855808139,-0.1664953828,0.113533631,-0.3573138118,-0.2230840921,-0.1297074258,-0.2990342677,-0.0198958907,-0.2856923044,-0.0178767089,-0.0423254296,0.0760572925,0.0820795819,0.0532877035,0.3012829721,0.0908265039,-0.1270110607,-0.1823029965,0.045443207,0.1271011978,0.1881293952,0.1879554391,0.1435558945,0.158248812,0.3156916201,0.0445809476,-0.0184111632,-0.48882249,-0.3923448324,0.1243815124,-0.0000142496,-0.0214338861,-0.0514348447,-0.0697956756,0.0403116681,-0.1614388824,0.1152889431,0.2696072161,-0.1647182107,-0.2373158485,-0.1565071195,-0.0114669735,-0.1103405803,-0.3104776144,-0.2459207475,-0.2052778006,0.182214126,-0.2187639028,0.0829838812,-0.5726172328,0.0895796716,-0.2523791194,0.0419618227,-0.0752221644,-0.1315330416,-0.368395716,0.3846784532,-0.1310067475,-0.1836591363,-0.063963972,-0.0795925856,0.1381900012,0.1339152902,-0.0233511105,0.0152277648,-0.2839080989,0.2396766245,-0.1025734767,0.1269342154,0.2458895147,0.2217749208,0.0532566346,0.0195346735,-0.3257350028,-0.0101708891,0.1243051738,0.2586684823,0.1649142951,-0.0023148581,-0.1266412586,0.4759716094,0.41021052,0.0864422619,0.2461429238,0.0690271258,-0.0310459081,-0.0720844269,-0.265080452,-0.1932799369,-0.0270625558,0.2228746265,0.007795494,-0.0200222805,-0.2028658688,-0.4187778831,-0.3887877762,-0.1032177731,-0.090183422,-0.0248359852,0.022597637,0.1172929481,-0.084455058,-0.4240666032,-0.2681210339,0.0643287599,0.178718701,0.1185797006,-0.0168630816,0.0371456742,-0.4187373519,-0.1615968943,-0.0375836417,0.2943128943,0.2055929154,-0.3450768292,0.1880765259,0.0988820195,-0.3641401827,-0.0752881393,0.3639464378,-0.0067338701,-0.0272207148,-0.2934643924,-0.1550566256,-0.0181934983,-0.1162452474,0.2589886487,0.0121607939,-0.0418913811,0.6412338018,-0.3157121539,-0.039496731,0.1368405223,0.1856145412,-0.0850628614,-0.2468770891,0.2709826529,-0.0777583495,-0.2421099693,-0.4657281041,-0.0692933127,-0.1497351229,0.06149235,-0.4079434872,-0.211431101,-0.0460082032,0.3354440629,0.1244231537,0.1765254885,-0.1636610478,0.1690299213,-0.0456600338,0.1065554172,-0.143304199,0.627271831,0.267029494,-0.4620840251,-0.2423051596,-0.2583221495,0.4511785507,-0.1259610504,0.0580201633,0.4277215302,-0.4641078711,-0.1543427259,0.0512974933,-0.0932699889,0.0950442776,0.254609108,-0.4261369705,-0.1574268937,0.5728214383,0.3493755758,-0.1567673087,0.3115528524,-0.4469200075,-0.0828493014,0.1440853924,-0.3016005456,0.8176675439,0.03916765,0.0679656938,-0.3305321038,0.105415374,0.1836758405,-0.1053980812,-0.0440056324,0.0527380221,-0.3161380589,-0.1374404132,0.0265801456,-0.0556862988,0.4730741382,-0.4407370985,0.3485810757,0.0469170213,0.3959118426,-0.1265681386,0.2334693968,0.1905294508,0.1281436384,-0.1501764655,0.0895575061,-0.3499027491,0.0966256261,0.0604889877,-0.3020166457,-0.0989628136,-0.12326435,-0.0856965333,0.1203918457,0.0427156538,0.3616164029,0.0299584251,-0.066143252,0.047894299,0.0981603116,-0.1054163203,0.1261378378,-0.0819061473,-0.1535652429,0.1932197511,0.1397257,-0.2461336255,-0.051955048,-0.0711550564,-0.0834356993,-0.1991076171,0.346242696,0.132000342,-0.0813195184,-0.1592241079,-0.0524581969,0.0770971179,-0.2919894457,-0.297978878,0.5363789797,0.0895888656,-0.0724028796,0.0567635037,0.0446158499,-0.1053930074,0.2349029332,0.0937959701,-0.2018083036,-0.0544941239,0.1176370233,0.4531309307,0.0468373522,0.0828655064,-0.1654901654,0.2697144449,-0.0222481694,0.1073923483,0.0532890968,-0.0576490089,0.0545428097,-0.2332496047,0.0694817305,0.0910306722,0.1202282086,0.1606689095,0.5604791045,0.1323271692,-0.1157587469,-0.3453596532,0.1237276345,0.2395761311,0.2470714152,-0.03657455,0.198519513,-0.0793556795,0.6873180866,-0.2969149053,0.1318293512,0.3975134492,0.3426426053,0.3198750615,0.179624632,-0.3751250803,-0.0741952658,0.1279970109,-0.0654600039,-0.0295179076,-0.2420735806,-0.0677833483,0.0826019347,0.0908955187,0.0578958616,-0.1134132221,-0.226594463,0.2689422369,-0.0860190392,0.3376496136,0.0107867019,0.0591254234,0.0466690622,0.1443540007,-0.2734315693,-0.4316164851,0.3497260511,-0.0207211636,-0.0865008906,-0.0390621722,-0.01090666,-0.0280412678,-0.1253492683,-0.1689086407,-0.2858787775,0.3722502291,0.0713134557,0.0568391681,0.1690735817,0.0196758658,0.1826631576,-0.0616477616,-0.0913354307,-0.2369185686,0.3942173123,-0.1218975261,0.2006997615,-0.2924503684,-0.2348063886,0.2931777537,-0.0259799697,0.145014599,-0.0636136308,0.0103197871,0.3602033854,-0.0889462903,0.0661943629,0.1304408163,0.0912622958,0.1841406375,0.4665001631,0.1780335009,-0.0550872423,0.0136528574,0.2035761923,0.1864255965,-0.0048272456,-0.1700942367,0.1611706913,0.1710534394,-0.0510014333,0.72757864,-0.0116877193,-0.0057894071,0.2497446239,-0.0912280232,0.0694127157,0.4447436929,0.1390617341,0.0992985964,-0.4390279651,0.1337791681,0.4352565408,0.1393812746,-0.0699817538,-0.0423329808,-0.2399267703,-0.1427197605,0.1366895884,0.0697481781,-0.2833135724,0.8152461052,0.0431853794,0.1011746973,0.2025991827,0.4424374998,0.3067195117,0.2169554532,0.000686874,0.0079771345,0.2020893842,-0.420309335,-0.0674838573,0.3493855298,0.0268330444,0.182602644,0.0041168123,0.3164588511,0.0658694133,-0.1819423437,0.110670656,-0.1567562968,-0.0701432824,-0.3488503695,0.0923336968,0.2150824219,-0.0796745494,0.0480164625,-0.3853045404,0.1449500024,-0.366671592,0.1461028457,0.0895379484,-0.3051271439,-0.0472588725,-0.1326808035,-0.1727474928,-0.6433356404,0.5588120818,0.1113869324,-0.0329862274,-0.11048574,0.1103616729,-0.159553051,0.0911312848,0.279338032,-0.1291183531,-0.2231062204,-0.1268815696,-0.3720630705,0.115521431,0.0454919003,-0.022807816,-0.2274467647,0.1429172605,0.2188262045,0.1887916476,0.1171537414,-0.0265935287,0.3292481005,0.1265394539,-0.3300289512,-0.0987650231,0.1161152124,0.0971552953,0.2759951353,-0.2498864681,0.3398551941,0.0256446358,0.025994828,-0.2597183883,0.044050511,0.2548987567,-0.2090631872,0.2395812869,0.0702373013,0.2490377426,-0.4570767283,0.136301294,0.5659573078,0.1062537283,-0.2446699888,0.2806098759,0.3050479889,-0.031650044,-0.418245852,0.3034989238,0.1939335465,-0.1193651482,-0.2073642015,-0.3572675288,-0.2085309178,0.038777288,-0.0009309047,-0.0102528436,0.0332772657,0.3955827355,-0.015533994,-0.0519191064,0.1367719471,-0.1302718967,0.353479892,0.1321196407,-0.2396323979,-0.1321877986,-0.2882088423,0.1360863,-0.0368303582,-0.3484397233,-0.2428430617,0.345930934,-0.0116855651,0.3151527643,0.2930339575,0.0231120009,-0.0914197862,-0.2050624341,0.1921301335,0.1376333684,-0.2124323398,-0.124657996,-0.259170562]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1769","title":"_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union when calling datasets.map with num_proc=2","comments":"More information: `run_mlm.py` will raise same error when `data_args.line_by_line==True`\r\n\r\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/9152f16023b59d262b51573714b40325c8e49370\/examples\/language-modeling\/run_mlm.py#L300\r\n","body":"It may be a bug of multiprocessing with Datasets, when I disable the multiprocessing by set num_proc to None, everything works fine.\r\n\r\nThe script I use is https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py\r\n\r\nScript args:\r\n\r\n```\r\n--model_name_or_path\r\n..\/..\/..\/model\/chinese-roberta-wwm-ext\r\n--train_file\r\n\/nfs\/volume-377-2\/bert\/data\/test\/train.txt\r\n--output_dir\r\ntest\r\n--do_train\r\n--per_device_train_batch_size\r\n2\r\n--gradient_accumulation_steps\r\n2\r\n--learning_rate\r\n1e-4\r\n--max_steps\r\n1000\r\n--warmup_steps\r\n10\r\n--save_steps\r\n1000\r\n--save_total_limit\r\n1\r\n--seed\r\n23333\r\n--max_seq_length\r\n512\r\n--preprocessing_num_workers\r\n2\r\n--cache_dir\r\n\/nfs\/volume-377-2\/bert\/data\/test\/cache\r\n```\r\n\r\nWhere the `\/nfs\/volume-377-2\/bert\/data\/test\/train.txt` is just a toy example with 10000 lines of random string, you should be able to reproduce this error esaily.\r\n\r\nFull Traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 398, in <module>\r\n main()\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 325, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\r\n","comment_length":10,"text":"_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union when calling datasets.map with num_proc=2\nIt may be a bug of multiprocessing with Datasets, when I disable the multiprocessing by set num_proc to None, everything works fine.\r\n\r\nThe script I use is https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py\r\n\r\nScript args:\r\n\r\n```\r\n--model_name_or_path\r\n..\/..\/..\/model\/chinese-roberta-wwm-ext\r\n--train_file\r\n\/nfs\/volume-377-2\/bert\/data\/test\/train.txt\r\n--output_dir\r\ntest\r\n--do_train\r\n--per_device_train_batch_size\r\n2\r\n--gradient_accumulation_steps\r\n2\r\n--learning_rate\r\n1e-4\r\n--max_steps\r\n1000\r\n--warmup_steps\r\n10\r\n--save_steps\r\n1000\r\n--save_total_limit\r\n1\r\n--seed\r\n23333\r\n--max_seq_length\r\n512\r\n--preprocessing_num_workers\r\n2\r\n--cache_dir\r\n\/nfs\/volume-377-2\/bert\/data\/test\/cache\r\n```\r\n\r\nWhere the `\/nfs\/volume-377-2\/bert\/data\/test\/train.txt` is just a toy example with 10000 lines of random string, you should be able to reproduce this error esaily.\r\n\r\nFull Traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 398, in <module>\r\n main()\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 325, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\r\n\nMore information: `run_mlm.py` will raise same error when `data_args.line_by_line==True`\r\n\r\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/9152f16023b59d262b51573714b40325c8e49370\/examples\/language-modeling\/run_mlm.py#L300\r\n","embeddings":[-0.2903597653,-0.2974283993,0.1132445857,0.1903711855,0.166158393,-0.0522105806,0.3321136534,0.2303635925,0.29939574,0.2801507711,0.0167170409,0.3376862705,-0.1110442802,0.2174146324,0.0041834014,-0.1060649753,-0.1158019528,-0.1348113567,-0.3830081224,0.0106051518,-0.2922798097,0.3295297623,-0.3364185989,0.178581655,-0.6447761059,-0.0980493426,-0.0480723754,0.2914254069,0.0867962465,-0.0838413835,0.091296576,-0.0355660208,0.0124873333,0.495975703,-0.0001100203,0.0433959849,0.2030376494,-0.0829199255,-0.0667256936,-0.4173735082,0.0713781044,-0.0390120484,0.070758529,-0.0548069626,0.1504491121,0.1799487174,0.1406906098,-0.0804709494,0.2613464594,0.1155149564,0.1944084316,0.5662001371,0.1144360006,-0.076135695,-0.1252603531,0.2157489657,-0.0322175771,0.2660860121,0.1158297211,-0.1664748639,-0.2567045093,0.2081440091,-0.0091969082,0.2851698399,-0.1553526968,0.3262346983,0.0900292844,-0.2589209974,0.1284050196,0.3979728222,0.0784635693,-0.232056573,-0.3231726289,-0.0680227727,-0.0854364112,-0.1713582575,0.0877538249,0.1660903543,-0.1176943481,0.0851684958,-0.2029177397,0.1319692582,0.2207043022,0.1039848626,0.2320752591,0.3622331023,-0.031342186,0.3686209619,0.2795686424,-0.1373411715,-0.2073924094,-0.1868893355,0.0720700994,0.0632183775,-0.0458192751,-0.1408384144,0.045508083,-0.0947569758,0.073656261,-0.1785828322,-0.1766469479,0.1712978035,-0.2389756292,0.3193747997,0.3000181019,-0.1272128671,-0.0990271121,0.2027037889,0.4027116001,-0.0739725083,-0.2251318693,0.1159959584,0.3312569559,-0.3447957933,-0.1041944399,0.1564332843,0.329223603,-0.1115081087,-0.149697721,-0.1945511699,-0.1339909732,-0.1725845933,-0.010302064,0.2903743684,0.0592746921,0.3461214006,0.0566079207,0.1878037006,-0.5467142463,-0.0829292685,-0.3398105502,0.218926996,-0.1916940212,0.0626075044,0.2220834941,0.2673146129,0.3550154865,0.1994703859,-0.0371265337,-0.1523097605,0.2351196259,-0.2494036108,-0.1287895143,0.1925622076,0.1652551144,0.2886993587,0.2680864036,-0.3080254495,-0.0635940209,0.0783043131,-0.2370546311,-0.2890685201,-0.0729562566,0.139649719,-0.0142909233,0.1257887632,-0.1699093133,0.0447562262,0.4315924942,-0.0497591682,-0.1882438213,-0.2977280915,-0.3939255178,-0.2207495868,-0.0378487483,0.3762208223,-0.2341665477,-0.1520403773,0.1977583915,-0.1235371456,0.5048832893,0.5401949883,-0.2475932837,-0.0822768435,-0.1492327303,0.6606734991,0.0360711515,0.1096854657,-0.2572190762,0.202608645,-0.0988949314,-0.1114991456,-0.0841935053,0.0370499715,-0.3572835028,-0.0326678194,0.4061434865,0.1351237893,0.0573740341,0.192356661,-0.271171689,-0.3141673207,0.1697528809,-0.0544769354,0.0685291588,-0.2757891119,-0.1262007654,-0.0896317139,0.4380550981,-0.2241077721,0.0792603567,-0.0038885754,0.1627585292,-0.0515407771,-0.0201107692,-0.2338376492,-0.5796083808,0.2878185511,-0.2233290672,0.2118987888,-0.064044036,-0.1344355494,0.2768068016,-0.0169793814,-0.3259687424,-0.0686179921,0.0756605044,0.0722182989,0.0509032123,-0.1822725981,-0.0842780322,-0.1165589243,-0.1049529836,-0.0149970902,-0.4686059952,0.0425728932,-0.4037323892,-0.4181522727,-0.189516753,0.2076245993,0.1067516208,-0.0068623866,-0.1171695143,0.1369666457,0.1502617449,-0.0461106859,-0.0495142005,0.0562658869,-0.0190868974,-0.0961284563,0.021799013,0.219856739,0.1103201881,-0.1780940443,-0.2147070765,0.4751463234,0.0284074377,0.2698736787,0.0416779704,0.0957993492,0.2242285013,0.0413596444,0.1322552115,-0.1709280461,0.1677090824,0.1714064628,0.1670323014,0.0913200006,-0.239271,0.039523486,0.5964310765,0.2025493234,0.4119851291,0.0081310831,-0.2059065104,-0.1373883784,-0.1942453682,0.0325800702,0.6064072251,0.1613005549,0.0643860176,0.001883429,-0.1250701845,-0.1186368167,0.0121865347,-0.0157128964,0.0383681133,0.1391063333,0.110177435,-0.0765566528,-0.303314656,-0.1031182781,0.01030437,0.3184674084,-0.3177983165,0.0145384418,-0.3326050341,0.4729814231,-0.2347498834,0.0097457618,-0.1918038428,-0.2922730744,-0.1737028062,0.3688523471,-0.2211097777,0.1546901911,0.1278085411,-0.2074619383,-0.1040067896,0.2358911335,0.0728950351,-0.0771441162,-0.2809194922,-0.0019136857,0.4048869908,-0.0430667438,0.0471233055,0.1536123306,-0.2586254776,-0.0553867705,-0.13959831,0.0791322961,-0.0761685297,0.2284352332,0.2975741327,0.2531296015,-0.0203261487,-0.3027900457,0.278198719,-0.2991173267,-0.1905634552,-0.0421640575,0.0779534876,-0.1181903258,-0.3645498157,-0.3952235579,-0.4591894448,-0.3345877528,0.4066636562,-0.0599519573,0.1627806872,0.2841716111,0.174965024,0.2511926591,0.1270663142,-0.1767324805,-0.2883633971,0.1619478464,0.22296682,0.0186432451,-0.2608917952,-0.0488841124,-0.1683621109,-0.0411670841,-0.2099181116,-0.1619539857,0.0679083541,-0.4185375571,0.250623107,0.1012951508,0.2987335026,0.6218903065,0.2733477652,-0.1571446955,-0.0458586663,-0.2676429152,0.1766682565,0.2072681785,0.2516293228,-0.0044127959,0.1994465888,-0.0090489136,0.7935594916,0.2588758171,-0.2274359912,0.2747074664,0.0024317801,-0.1775773615,-0.0838076547,-0.157329604,-0.063519977,-0.323271215,0.1069680005,0.2216331661,-0.1139478162,-0.1790085435,0.0517217256,0.1503136307,-0.0914623663,-0.2932106853,0.4917781651,0.2086156756,-0.1075986624,0.0574083813,-0.0566314757,-0.1271755695,0.0405725762,-0.0875590742,0.0143404966,0.0110158855,-0.1077002361,-0.176575616,-0.2057750374,-0.507540524,0.503590703,0.1288685203,0.3068889678,-0.1078222767,-0.2095698863,-0.0616324544,-0.0828850344,0.8502633572,-0.3850911558,0.0100383013,0.119369559,-0.2391629219,-0.2514419258,-0.2370387465,-0.2766326666,0.3788052201,0.3212099075,0.4916625619,-0.2372591496,-0.1024899855,0.3546540141,0.0173880365,-0.1381927431,-0.3041334152,-0.3489311337,-0.2444817573,-0.4569969475,-0.0768626481,0.2236945331,0.1462452263,-0.1781131774,0.0569178686,-0.0031951673,-0.249013409,0.2492820919,0.025196746,0.1551803052,0.2063275874,0.0066008242,-0.0015961076,0.2242492288,-0.1629241556,0.2497969717,-0.1691099554,0.0070374175,0.294501245,-0.2529675364,0.3081232309,0.1772142202,0.0588466898,0.3214899004,0.2185170501,0.3157062232,-0.0787473843,0.1128968671,0.321444124,0.2550255358,-0.1521228403,-0.2701735795,0.5763700008,0.2069915682,0.0711066574,0.2796461582,-0.2895084321,-0.2001235336,0.195239231,-0.0853109583,0.8592373729,-0.0334537104,0.1365751326,0.1039628759,0.0401422009,0.3963153958,0.188486591,0.1458002776,-0.42859146,0.0199723877,0.0063629807,-0.1561582386,0.0699793026,0.1950140893,-0.4718109071,0.3989383578,-0.1540239602,-0.1372059584,-0.0209685918,0.2605495751,0.0444882326,-0.267472446,0.0987746865,0.1068307459,-0.0973727331,0.2509978712,0.0981686786,-0.0154077979,0.076826185,-0.3066434562,-0.3827336729,0.1194452196,-0.2119664103,0.5681127906,0.2654306591,0.1176436022,0.3349756002,0.1474426836,0.0042988327,0.1244659871,-0.1284873635,-0.0136926407,0.0300500039,0.1223837361,-0.2138327956,-0.0197317768,0.2193788439,0.0380547084,-0.393679142,-0.161359027,-0.0826720819,0.0640948936,-0.1460669786,0.0284341387,0.0239791553,-0.2233829945,-0.467312336,0.0605952144,-0.3597423434,-0.2393892854,0.1097849011,-0.025140468,-0.4206591249,0.2320714295,0.0984065682,-0.2648565769,0.097448267,0.4654233754,0.1427570283,0.3938289285,0.6578114629,0.4816328287,-0.1619143486,-0.2177283466,-0.1378337741,0.2559019625,-0.5192023516,0.2636250556,0.1261709332,-0.1565431207,-0.024752453,0.3667325675,0.0422928073,0.2563848495,-0.2222355306,-0.284660548,-0.4323421121,0.0412186794,-0.0124555873,0.1584905982,0.0377329327,0.4392094314,0.0725223646,0.344676882,-0.3166029155,0.1015893668,-0.4592546225,0.127050668,0.0869547948,-0.3884992898,0.00718047,0.1028387994,0.1086364761,0.3089999557,-0.2994246483,-0.2348469496,-0.2264022082,0.0722892433,-0.0692547038,-0.0594015829,0.2124275714,0.2266760767,-0.0329504758,-0.1760732979,-0.0075875646,0.052118022,-0.0104456767,0.1474796236,0.2194553912,0.0195672642,-0.1188956872,0.0047815116,-0.240739122,0.0302963369,0.1427271068,0.0124206934,0.1459383219,0.0674338713,-0.1345796138,-0.1771808118,0.0809547603,0.1359450519,-0.190765202,-0.3552043736,-0.2111646235,-0.1962959766,0.2878009677,0.1750219911,-0.1338817179,0.0425298698,0.0711072236,0.229493916,-0.3894515038,0.1009596959,0.3159351945,-0.0812982693,0.2417636067,0.2346811593,-0.1067333743,0.1655505002,-0.2935241759,0.126708895,0.0022715835,-0.4911993146,-0.0092727225,0.0905084088,-0.1272123307,0.1393156946,0.3981228471,-0.0438104644,0.2180084288,0.1734299213,-0.0295288283,0.2861862779,-0.3371940255,0.1603489518,0.0428447984,-0.6167395711,0.2229512334,0.3042711914,-0.1545279771,0.2104459554,-0.0307244956,0.3675802946,-0.5919659734,-0.1959854066,-0.0851378292,0.2790286541,-0.3977425098,-0.419826895,-0.0324660949,-0.0547241345,-0.2179806381,0.0574031733,-0.071031034,0.1367316544,0.3471319675,0.0215960611,-0.1445557326,-0.3487926722,-0.1260674745,-0.0126963854,0.1340655684,-0.2956922054,0.0682361051,0.2749158442,-0.0459633209,-0.5047225356,0.2228365391,0.5591908097,0.3988352418,-0.3692156076,0.0725747719,-0.039594613,0.0618858784,-0.1044700444,0.0768813267,0.0960938558,0.1491709203,0.2230308801,0.1165978387,-0.2096038312,-0.0821885839,0.1226791218,0.3338048458,0.0517791621,0.4257694781,-0.0936316997,-0.2233311534,-0.0471887253,0.1381967813,-0.4184854329,0.2153057754,0.5088744164,-0.2176353484,0.2733813822,-0.0391260795,0.0981447101,-0.0012670336,0.6340998411,0.1495898217,0.0237321984,-0.4357896745,0.2468439192,-0.5064620376,0.0341480039,-0.2610830367,0.143806532,-0.3708487749,0.347661078,0.1202931255,-0.0631867871,-0.0096283695,0.3259037733,0.2164503187,0.3538523316,-0.3152367771,-0.1542017907,-0.2725798786,-0.211403653,0.3093431592,-0.4370567203,0.0129666105,0.1691231281,0.0520184338,-0.3621643782,0.0353428237,0.0055553382,0.1597293019,-0.0082195206,0.2578552365,0.4893649518,-0.2150480598,-0.0147047071,-0.0939578488,-0.0348746069,-0.165054366,0.024586929,0.0982365906,0.5375682116,-0.2332443744,0.3349668682,-0.4291658401,0.1068600491,0.0017594325,0.0730224401,-0.1125725955,0.1844571829,0.144515112,0.0153929098,0.0043698065,0.3746371269,0.012435195,-0.0591237508,-0.196019277,-0.2834496796,0.6401267648,-0.2923009396,-0.4539929032,-0.4232752621,0.1518884748,-0.2287297398,-0.0712872669,-0.5009861588,0.0414506793,0.1367105395,-0.0241665188,-0.4582684636,0.238963291,0.0122072957,-0.2194100767,-0.0912206173,0.3244309127,0.1423429251,-0.2926914096,0.1541839242,-0.3422243297]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1769","title":"_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union when calling datasets.map with num_proc=2","comments":"Hi ! What version of python and datasets do you have ? And also what version of dill and pickle ?","body":"It may be a bug of multiprocessing with Datasets, when I disable the multiprocessing by set num_proc to None, everything works fine.\r\n\r\nThe script I use is https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py\r\n\r\nScript args:\r\n\r\n```\r\n--model_name_or_path\r\n..\/..\/..\/model\/chinese-roberta-wwm-ext\r\n--train_file\r\n\/nfs\/volume-377-2\/bert\/data\/test\/train.txt\r\n--output_dir\r\ntest\r\n--do_train\r\n--per_device_train_batch_size\r\n2\r\n--gradient_accumulation_steps\r\n2\r\n--learning_rate\r\n1e-4\r\n--max_steps\r\n1000\r\n--warmup_steps\r\n10\r\n--save_steps\r\n1000\r\n--save_total_limit\r\n1\r\n--seed\r\n23333\r\n--max_seq_length\r\n512\r\n--preprocessing_num_workers\r\n2\r\n--cache_dir\r\n\/nfs\/volume-377-2\/bert\/data\/test\/cache\r\n```\r\n\r\nWhere the `\/nfs\/volume-377-2\/bert\/data\/test\/train.txt` is just a toy example with 10000 lines of random string, you should be able to reproduce this error esaily.\r\n\r\nFull Traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 398, in <module>\r\n main()\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 325, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\r\n","comment_length":21,"text":"_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union when calling datasets.map with num_proc=2\nIt may be a bug of multiprocessing with Datasets, when I disable the multiprocessing by set num_proc to None, everything works fine.\r\n\r\nThe script I use is https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py\r\n\r\nScript args:\r\n\r\n```\r\n--model_name_or_path\r\n..\/..\/..\/model\/chinese-roberta-wwm-ext\r\n--train_file\r\n\/nfs\/volume-377-2\/bert\/data\/test\/train.txt\r\n--output_dir\r\ntest\r\n--do_train\r\n--per_device_train_batch_size\r\n2\r\n--gradient_accumulation_steps\r\n2\r\n--learning_rate\r\n1e-4\r\n--max_steps\r\n1000\r\n--warmup_steps\r\n10\r\n--save_steps\r\n1000\r\n--save_total_limit\r\n1\r\n--seed\r\n23333\r\n--max_seq_length\r\n512\r\n--preprocessing_num_workers\r\n2\r\n--cache_dir\r\n\/nfs\/volume-377-2\/bert\/data\/test\/cache\r\n```\r\n\r\nWhere the `\/nfs\/volume-377-2\/bert\/data\/test\/train.txt` is just a toy example with 10000 lines of random string, you should be able to reproduce this error esaily.\r\n\r\nFull Traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 398, in <module>\r\n main()\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 325, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\r\n\nHi ! What version of python and datasets do you have ? And also what version of dill and pickle ?","embeddings":[-0.2903597653,-0.2974283993,0.1132445857,0.1903711855,0.166158393,-0.0522105806,0.3321136534,0.2303635925,0.29939574,0.2801507711,0.0167170409,0.3376862705,-0.1110442802,0.2174146324,0.0041834014,-0.1060649753,-0.1158019528,-0.1348113567,-0.3830081224,0.0106051518,-0.2922798097,0.3295297623,-0.3364185989,0.178581655,-0.6447761059,-0.0980493426,-0.0480723754,0.2914254069,0.0867962465,-0.0838413835,0.091296576,-0.0355660208,0.0124873333,0.495975703,-0.0001100203,0.0433959849,0.2030376494,-0.0829199255,-0.0667256936,-0.4173735082,0.0713781044,-0.0390120484,0.070758529,-0.0548069626,0.1504491121,0.1799487174,0.1406906098,-0.0804709494,0.2613464594,0.1155149564,0.1944084316,0.5662001371,0.1144360006,-0.076135695,-0.1252603531,0.2157489657,-0.0322175771,0.2660860121,0.1158297211,-0.1664748639,-0.2567045093,0.2081440091,-0.0091969082,0.2851698399,-0.1553526968,0.3262346983,0.0900292844,-0.2589209974,0.1284050196,0.3979728222,0.0784635693,-0.232056573,-0.3231726289,-0.0680227727,-0.0854364112,-0.1713582575,0.0877538249,0.1660903543,-0.1176943481,0.0851684958,-0.2029177397,0.1319692582,0.2207043022,0.1039848626,0.2320752591,0.3622331023,-0.031342186,0.3686209619,0.2795686424,-0.1373411715,-0.2073924094,-0.1868893355,0.0720700994,0.0632183775,-0.0458192751,-0.1408384144,0.045508083,-0.0947569758,0.073656261,-0.1785828322,-0.1766469479,0.1712978035,-0.2389756292,0.3193747997,0.3000181019,-0.1272128671,-0.0990271121,0.2027037889,0.4027116001,-0.0739725083,-0.2251318693,0.1159959584,0.3312569559,-0.3447957933,-0.1041944399,0.1564332843,0.329223603,-0.1115081087,-0.149697721,-0.1945511699,-0.1339909732,-0.1725845933,-0.010302064,0.2903743684,0.0592746921,0.3461214006,0.0566079207,0.1878037006,-0.5467142463,-0.0829292685,-0.3398105502,0.218926996,-0.1916940212,0.0626075044,0.2220834941,0.2673146129,0.3550154865,0.1994703859,-0.0371265337,-0.1523097605,0.2351196259,-0.2494036108,-0.1287895143,0.1925622076,0.1652551144,0.2886993587,0.2680864036,-0.3080254495,-0.0635940209,0.0783043131,-0.2370546311,-0.2890685201,-0.0729562566,0.139649719,-0.0142909233,0.1257887632,-0.1699093133,0.0447562262,0.4315924942,-0.0497591682,-0.1882438213,-0.2977280915,-0.3939255178,-0.2207495868,-0.0378487483,0.3762208223,-0.2341665477,-0.1520403773,0.1977583915,-0.1235371456,0.5048832893,0.5401949883,-0.2475932837,-0.0822768435,-0.1492327303,0.6606734991,0.0360711515,0.1096854657,-0.2572190762,0.202608645,-0.0988949314,-0.1114991456,-0.0841935053,0.0370499715,-0.3572835028,-0.0326678194,0.4061434865,0.1351237893,0.0573740341,0.192356661,-0.271171689,-0.3141673207,0.1697528809,-0.0544769354,0.0685291588,-0.2757891119,-0.1262007654,-0.0896317139,0.4380550981,-0.2241077721,0.0792603567,-0.0038885754,0.1627585292,-0.0515407771,-0.0201107692,-0.2338376492,-0.5796083808,0.2878185511,-0.2233290672,0.2118987888,-0.064044036,-0.1344355494,0.2768068016,-0.0169793814,-0.3259687424,-0.0686179921,0.0756605044,0.0722182989,0.0509032123,-0.1822725981,-0.0842780322,-0.1165589243,-0.1049529836,-0.0149970902,-0.4686059952,0.0425728932,-0.4037323892,-0.4181522727,-0.189516753,0.2076245993,0.1067516208,-0.0068623866,-0.1171695143,0.1369666457,0.1502617449,-0.0461106859,-0.0495142005,0.0562658869,-0.0190868974,-0.0961284563,0.021799013,0.219856739,0.1103201881,-0.1780940443,-0.2147070765,0.4751463234,0.0284074377,0.2698736787,0.0416779704,0.0957993492,0.2242285013,0.0413596444,0.1322552115,-0.1709280461,0.1677090824,0.1714064628,0.1670323014,0.0913200006,-0.239271,0.039523486,0.5964310765,0.2025493234,0.4119851291,0.0081310831,-0.2059065104,-0.1373883784,-0.1942453682,0.0325800702,0.6064072251,0.1613005549,0.0643860176,0.001883429,-0.1250701845,-0.1186368167,0.0121865347,-0.0157128964,0.0383681133,0.1391063333,0.110177435,-0.0765566528,-0.303314656,-0.1031182781,0.01030437,0.3184674084,-0.3177983165,0.0145384418,-0.3326050341,0.4729814231,-0.2347498834,0.0097457618,-0.1918038428,-0.2922730744,-0.1737028062,0.3688523471,-0.2211097777,0.1546901911,0.1278085411,-0.2074619383,-0.1040067896,0.2358911335,0.0728950351,-0.0771441162,-0.2809194922,-0.0019136857,0.4048869908,-0.0430667438,0.0471233055,0.1536123306,-0.2586254776,-0.0553867705,-0.13959831,0.0791322961,-0.0761685297,0.2284352332,0.2975741327,0.2531296015,-0.0203261487,-0.3027900457,0.278198719,-0.2991173267,-0.1905634552,-0.0421640575,0.0779534876,-0.1181903258,-0.3645498157,-0.3952235579,-0.4591894448,-0.3345877528,0.4066636562,-0.0599519573,0.1627806872,0.2841716111,0.174965024,0.2511926591,0.1270663142,-0.1767324805,-0.2883633971,0.1619478464,0.22296682,0.0186432451,-0.2608917952,-0.0488841124,-0.1683621109,-0.0411670841,-0.2099181116,-0.1619539857,0.0679083541,-0.4185375571,0.250623107,0.1012951508,0.2987335026,0.6218903065,0.2733477652,-0.1571446955,-0.0458586663,-0.2676429152,0.1766682565,0.2072681785,0.2516293228,-0.0044127959,0.1994465888,-0.0090489136,0.7935594916,0.2588758171,-0.2274359912,0.2747074664,0.0024317801,-0.1775773615,-0.0838076547,-0.157329604,-0.063519977,-0.323271215,0.1069680005,0.2216331661,-0.1139478162,-0.1790085435,0.0517217256,0.1503136307,-0.0914623663,-0.2932106853,0.4917781651,0.2086156756,-0.1075986624,0.0574083813,-0.0566314757,-0.1271755695,0.0405725762,-0.0875590742,0.0143404966,0.0110158855,-0.1077002361,-0.176575616,-0.2057750374,-0.507540524,0.503590703,0.1288685203,0.3068889678,-0.1078222767,-0.2095698863,-0.0616324544,-0.0828850344,0.8502633572,-0.3850911558,0.0100383013,0.119369559,-0.2391629219,-0.2514419258,-0.2370387465,-0.2766326666,0.3788052201,0.3212099075,0.4916625619,-0.2372591496,-0.1024899855,0.3546540141,0.0173880365,-0.1381927431,-0.3041334152,-0.3489311337,-0.2444817573,-0.4569969475,-0.0768626481,0.2236945331,0.1462452263,-0.1781131774,0.0569178686,-0.0031951673,-0.249013409,0.2492820919,0.025196746,0.1551803052,0.2063275874,0.0066008242,-0.0015961076,0.2242492288,-0.1629241556,0.2497969717,-0.1691099554,0.0070374175,0.294501245,-0.2529675364,0.3081232309,0.1772142202,0.0588466898,0.3214899004,0.2185170501,0.3157062232,-0.0787473843,0.1128968671,0.321444124,0.2550255358,-0.1521228403,-0.2701735795,0.5763700008,0.2069915682,0.0711066574,0.2796461582,-0.2895084321,-0.2001235336,0.195239231,-0.0853109583,0.8592373729,-0.0334537104,0.1365751326,0.1039628759,0.0401422009,0.3963153958,0.188486591,0.1458002776,-0.42859146,0.0199723877,0.0063629807,-0.1561582386,0.0699793026,0.1950140893,-0.4718109071,0.3989383578,-0.1540239602,-0.1372059584,-0.0209685918,0.2605495751,0.0444882326,-0.267472446,0.0987746865,0.1068307459,-0.0973727331,0.2509978712,0.0981686786,-0.0154077979,0.076826185,-0.3066434562,-0.3827336729,0.1194452196,-0.2119664103,0.5681127906,0.2654306591,0.1176436022,0.3349756002,0.1474426836,0.0042988327,0.1244659871,-0.1284873635,-0.0136926407,0.0300500039,0.1223837361,-0.2138327956,-0.0197317768,0.2193788439,0.0380547084,-0.393679142,-0.161359027,-0.0826720819,0.0640948936,-0.1460669786,0.0284341387,0.0239791553,-0.2233829945,-0.467312336,0.0605952144,-0.3597423434,-0.2393892854,0.1097849011,-0.025140468,-0.4206591249,0.2320714295,0.0984065682,-0.2648565769,0.097448267,0.4654233754,0.1427570283,0.3938289285,0.6578114629,0.4816328287,-0.1619143486,-0.2177283466,-0.1378337741,0.2559019625,-0.5192023516,0.2636250556,0.1261709332,-0.1565431207,-0.024752453,0.3667325675,0.0422928073,0.2563848495,-0.2222355306,-0.284660548,-0.4323421121,0.0412186794,-0.0124555873,0.1584905982,0.0377329327,0.4392094314,0.0725223646,0.344676882,-0.3166029155,0.1015893668,-0.4592546225,0.127050668,0.0869547948,-0.3884992898,0.00718047,0.1028387994,0.1086364761,0.3089999557,-0.2994246483,-0.2348469496,-0.2264022082,0.0722892433,-0.0692547038,-0.0594015829,0.2124275714,0.2266760767,-0.0329504758,-0.1760732979,-0.0075875646,0.052118022,-0.0104456767,0.1474796236,0.2194553912,0.0195672642,-0.1188956872,0.0047815116,-0.240739122,0.0302963369,0.1427271068,0.0124206934,0.1459383219,0.0674338713,-0.1345796138,-0.1771808118,0.0809547603,0.1359450519,-0.190765202,-0.3552043736,-0.2111646235,-0.1962959766,0.2878009677,0.1750219911,-0.1338817179,0.0425298698,0.0711072236,0.229493916,-0.3894515038,0.1009596959,0.3159351945,-0.0812982693,0.2417636067,0.2346811593,-0.1067333743,0.1655505002,-0.2935241759,0.126708895,0.0022715835,-0.4911993146,-0.0092727225,0.0905084088,-0.1272123307,0.1393156946,0.3981228471,-0.0438104644,0.2180084288,0.1734299213,-0.0295288283,0.2861862779,-0.3371940255,0.1603489518,0.0428447984,-0.6167395711,0.2229512334,0.3042711914,-0.1545279771,0.2104459554,-0.0307244956,0.3675802946,-0.5919659734,-0.1959854066,-0.0851378292,0.2790286541,-0.3977425098,-0.419826895,-0.0324660949,-0.0547241345,-0.2179806381,0.0574031733,-0.071031034,0.1367316544,0.3471319675,0.0215960611,-0.1445557326,-0.3487926722,-0.1260674745,-0.0126963854,0.1340655684,-0.2956922054,0.0682361051,0.2749158442,-0.0459633209,-0.5047225356,0.2228365391,0.5591908097,0.3988352418,-0.3692156076,0.0725747719,-0.039594613,0.0618858784,-0.1044700444,0.0768813267,0.0960938558,0.1491709203,0.2230308801,0.1165978387,-0.2096038312,-0.0821885839,0.1226791218,0.3338048458,0.0517791621,0.4257694781,-0.0936316997,-0.2233311534,-0.0471887253,0.1381967813,-0.4184854329,0.2153057754,0.5088744164,-0.2176353484,0.2733813822,-0.0391260795,0.0981447101,-0.0012670336,0.6340998411,0.1495898217,0.0237321984,-0.4357896745,0.2468439192,-0.5064620376,0.0341480039,-0.2610830367,0.143806532,-0.3708487749,0.347661078,0.1202931255,-0.0631867871,-0.0096283695,0.3259037733,0.2164503187,0.3538523316,-0.3152367771,-0.1542017907,-0.2725798786,-0.211403653,0.3093431592,-0.4370567203,0.0129666105,0.1691231281,0.0520184338,-0.3621643782,0.0353428237,0.0055553382,0.1597293019,-0.0082195206,0.2578552365,0.4893649518,-0.2150480598,-0.0147047071,-0.0939578488,-0.0348746069,-0.165054366,0.024586929,0.0982365906,0.5375682116,-0.2332443744,0.3349668682,-0.4291658401,0.1068600491,0.0017594325,0.0730224401,-0.1125725955,0.1844571829,0.144515112,0.0153929098,0.0043698065,0.3746371269,0.012435195,-0.0591237508,-0.196019277,-0.2834496796,0.6401267648,-0.2923009396,-0.4539929032,-0.4232752621,0.1518884748,-0.2287297398,-0.0712872669,-0.5009861588,0.0414506793,0.1367105395,-0.0241665188,-0.4582684636,0.238963291,0.0122072957,-0.2194100767,-0.0912206173,0.3244309127,0.1423429251,-0.2926914096,0.1541839242,-0.3422243297]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1769","title":"_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union when calling datasets.map with num_proc=2","comments":"> Hi ! What version of python and datasets do you have ? And also what version of dill and pickle ?\r\n\r\npython==3.6.10\r\ndatasets==1.2.1\r\ndill==0.3.2\r\npickle.format_version==4.0","body":"It may be a bug of multiprocessing with Datasets, when I disable the multiprocessing by set num_proc to None, everything works fine.\r\n\r\nThe script I use is https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py\r\n\r\nScript args:\r\n\r\n```\r\n--model_name_or_path\r\n..\/..\/..\/model\/chinese-roberta-wwm-ext\r\n--train_file\r\n\/nfs\/volume-377-2\/bert\/data\/test\/train.txt\r\n--output_dir\r\ntest\r\n--do_train\r\n--per_device_train_batch_size\r\n2\r\n--gradient_accumulation_steps\r\n2\r\n--learning_rate\r\n1e-4\r\n--max_steps\r\n1000\r\n--warmup_steps\r\n10\r\n--save_steps\r\n1000\r\n--save_total_limit\r\n1\r\n--seed\r\n23333\r\n--max_seq_length\r\n512\r\n--preprocessing_num_workers\r\n2\r\n--cache_dir\r\n\/nfs\/volume-377-2\/bert\/data\/test\/cache\r\n```\r\n\r\nWhere the `\/nfs\/volume-377-2\/bert\/data\/test\/train.txt` is just a toy example with 10000 lines of random string, you should be able to reproduce this error esaily.\r\n\r\nFull Traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 398, in <module>\r\n main()\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 325, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\r\n","comment_length":26,"text":"_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union when calling datasets.map with num_proc=2\nIt may be a bug of multiprocessing with Datasets, when I disable the multiprocessing by set num_proc to None, everything works fine.\r\n\r\nThe script I use is https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py\r\n\r\nScript args:\r\n\r\n```\r\n--model_name_or_path\r\n..\/..\/..\/model\/chinese-roberta-wwm-ext\r\n--train_file\r\n\/nfs\/volume-377-2\/bert\/data\/test\/train.txt\r\n--output_dir\r\ntest\r\n--do_train\r\n--per_device_train_batch_size\r\n2\r\n--gradient_accumulation_steps\r\n2\r\n--learning_rate\r\n1e-4\r\n--max_steps\r\n1000\r\n--warmup_steps\r\n10\r\n--save_steps\r\n1000\r\n--save_total_limit\r\n1\r\n--seed\r\n23333\r\n--max_seq_length\r\n512\r\n--preprocessing_num_workers\r\n2\r\n--cache_dir\r\n\/nfs\/volume-377-2\/bert\/data\/test\/cache\r\n```\r\n\r\nWhere the `\/nfs\/volume-377-2\/bert\/data\/test\/train.txt` is just a toy example with 10000 lines of random string, you should be able to reproduce this error esaily.\r\n\r\nFull Traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 398, in <module>\r\n main()\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 325, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\r\n\n> Hi ! What version of python and datasets do you have ? And also what version of dill and pickle ?\r\n\r\npython==3.6.10\r\ndatasets==1.2.1\r\ndill==0.3.2\r\npickle.format_version==4.0","embeddings":[-0.2903597653,-0.2974283993,0.1132445857,0.1903711855,0.166158393,-0.0522105806,0.3321136534,0.2303635925,0.29939574,0.2801507711,0.0167170409,0.3376862705,-0.1110442802,0.2174146324,0.0041834014,-0.1060649753,-0.1158019528,-0.1348113567,-0.3830081224,0.0106051518,-0.2922798097,0.3295297623,-0.3364185989,0.178581655,-0.6447761059,-0.0980493426,-0.0480723754,0.2914254069,0.0867962465,-0.0838413835,0.091296576,-0.0355660208,0.0124873333,0.495975703,-0.0001100203,0.0433959849,0.2030376494,-0.0829199255,-0.0667256936,-0.4173735082,0.0713781044,-0.0390120484,0.070758529,-0.0548069626,0.1504491121,0.1799487174,0.1406906098,-0.0804709494,0.2613464594,0.1155149564,0.1944084316,0.5662001371,0.1144360006,-0.076135695,-0.1252603531,0.2157489657,-0.0322175771,0.2660860121,0.1158297211,-0.1664748639,-0.2567045093,0.2081440091,-0.0091969082,0.2851698399,-0.1553526968,0.3262346983,0.0900292844,-0.2589209974,0.1284050196,0.3979728222,0.0784635693,-0.232056573,-0.3231726289,-0.0680227727,-0.0854364112,-0.1713582575,0.0877538249,0.1660903543,-0.1176943481,0.0851684958,-0.2029177397,0.1319692582,0.2207043022,0.1039848626,0.2320752591,0.3622331023,-0.031342186,0.3686209619,0.2795686424,-0.1373411715,-0.2073924094,-0.1868893355,0.0720700994,0.0632183775,-0.0458192751,-0.1408384144,0.045508083,-0.0947569758,0.073656261,-0.1785828322,-0.1766469479,0.1712978035,-0.2389756292,0.3193747997,0.3000181019,-0.1272128671,-0.0990271121,0.2027037889,0.4027116001,-0.0739725083,-0.2251318693,0.1159959584,0.3312569559,-0.3447957933,-0.1041944399,0.1564332843,0.329223603,-0.1115081087,-0.149697721,-0.1945511699,-0.1339909732,-0.1725845933,-0.010302064,0.2903743684,0.0592746921,0.3461214006,0.0566079207,0.1878037006,-0.5467142463,-0.0829292685,-0.3398105502,0.218926996,-0.1916940212,0.0626075044,0.2220834941,0.2673146129,0.3550154865,0.1994703859,-0.0371265337,-0.1523097605,0.2351196259,-0.2494036108,-0.1287895143,0.1925622076,0.1652551144,0.2886993587,0.2680864036,-0.3080254495,-0.0635940209,0.0783043131,-0.2370546311,-0.2890685201,-0.0729562566,0.139649719,-0.0142909233,0.1257887632,-0.1699093133,0.0447562262,0.4315924942,-0.0497591682,-0.1882438213,-0.2977280915,-0.3939255178,-0.2207495868,-0.0378487483,0.3762208223,-0.2341665477,-0.1520403773,0.1977583915,-0.1235371456,0.5048832893,0.5401949883,-0.2475932837,-0.0822768435,-0.1492327303,0.6606734991,0.0360711515,0.1096854657,-0.2572190762,0.202608645,-0.0988949314,-0.1114991456,-0.0841935053,0.0370499715,-0.3572835028,-0.0326678194,0.4061434865,0.1351237893,0.0573740341,0.192356661,-0.271171689,-0.3141673207,0.1697528809,-0.0544769354,0.0685291588,-0.2757891119,-0.1262007654,-0.0896317139,0.4380550981,-0.2241077721,0.0792603567,-0.0038885754,0.1627585292,-0.0515407771,-0.0201107692,-0.2338376492,-0.5796083808,0.2878185511,-0.2233290672,0.2118987888,-0.064044036,-0.1344355494,0.2768068016,-0.0169793814,-0.3259687424,-0.0686179921,0.0756605044,0.0722182989,0.0509032123,-0.1822725981,-0.0842780322,-0.1165589243,-0.1049529836,-0.0149970902,-0.4686059952,0.0425728932,-0.4037323892,-0.4181522727,-0.189516753,0.2076245993,0.1067516208,-0.0068623866,-0.1171695143,0.1369666457,0.1502617449,-0.0461106859,-0.0495142005,0.0562658869,-0.0190868974,-0.0961284563,0.021799013,0.219856739,0.1103201881,-0.1780940443,-0.2147070765,0.4751463234,0.0284074377,0.2698736787,0.0416779704,0.0957993492,0.2242285013,0.0413596444,0.1322552115,-0.1709280461,0.1677090824,0.1714064628,0.1670323014,0.0913200006,-0.239271,0.039523486,0.5964310765,0.2025493234,0.4119851291,0.0081310831,-0.2059065104,-0.1373883784,-0.1942453682,0.0325800702,0.6064072251,0.1613005549,0.0643860176,0.001883429,-0.1250701845,-0.1186368167,0.0121865347,-0.0157128964,0.0383681133,0.1391063333,0.110177435,-0.0765566528,-0.303314656,-0.1031182781,0.01030437,0.3184674084,-0.3177983165,0.0145384418,-0.3326050341,0.4729814231,-0.2347498834,0.0097457618,-0.1918038428,-0.2922730744,-0.1737028062,0.3688523471,-0.2211097777,0.1546901911,0.1278085411,-0.2074619383,-0.1040067896,0.2358911335,0.0728950351,-0.0771441162,-0.2809194922,-0.0019136857,0.4048869908,-0.0430667438,0.0471233055,0.1536123306,-0.2586254776,-0.0553867705,-0.13959831,0.0791322961,-0.0761685297,0.2284352332,0.2975741327,0.2531296015,-0.0203261487,-0.3027900457,0.278198719,-0.2991173267,-0.1905634552,-0.0421640575,0.0779534876,-0.1181903258,-0.3645498157,-0.3952235579,-0.4591894448,-0.3345877528,0.4066636562,-0.0599519573,0.1627806872,0.2841716111,0.174965024,0.2511926591,0.1270663142,-0.1767324805,-0.2883633971,0.1619478464,0.22296682,0.0186432451,-0.2608917952,-0.0488841124,-0.1683621109,-0.0411670841,-0.2099181116,-0.1619539857,0.0679083541,-0.4185375571,0.250623107,0.1012951508,0.2987335026,0.6218903065,0.2733477652,-0.1571446955,-0.0458586663,-0.2676429152,0.1766682565,0.2072681785,0.2516293228,-0.0044127959,0.1994465888,-0.0090489136,0.7935594916,0.2588758171,-0.2274359912,0.2747074664,0.0024317801,-0.1775773615,-0.0838076547,-0.157329604,-0.063519977,-0.323271215,0.1069680005,0.2216331661,-0.1139478162,-0.1790085435,0.0517217256,0.1503136307,-0.0914623663,-0.2932106853,0.4917781651,0.2086156756,-0.1075986624,0.0574083813,-0.0566314757,-0.1271755695,0.0405725762,-0.0875590742,0.0143404966,0.0110158855,-0.1077002361,-0.176575616,-0.2057750374,-0.507540524,0.503590703,0.1288685203,0.3068889678,-0.1078222767,-0.2095698863,-0.0616324544,-0.0828850344,0.8502633572,-0.3850911558,0.0100383013,0.119369559,-0.2391629219,-0.2514419258,-0.2370387465,-0.2766326666,0.3788052201,0.3212099075,0.4916625619,-0.2372591496,-0.1024899855,0.3546540141,0.0173880365,-0.1381927431,-0.3041334152,-0.3489311337,-0.2444817573,-0.4569969475,-0.0768626481,0.2236945331,0.1462452263,-0.1781131774,0.0569178686,-0.0031951673,-0.249013409,0.2492820919,0.025196746,0.1551803052,0.2063275874,0.0066008242,-0.0015961076,0.2242492288,-0.1629241556,0.2497969717,-0.1691099554,0.0070374175,0.294501245,-0.2529675364,0.3081232309,0.1772142202,0.0588466898,0.3214899004,0.2185170501,0.3157062232,-0.0787473843,0.1128968671,0.321444124,0.2550255358,-0.1521228403,-0.2701735795,0.5763700008,0.2069915682,0.0711066574,0.2796461582,-0.2895084321,-0.2001235336,0.195239231,-0.0853109583,0.8592373729,-0.0334537104,0.1365751326,0.1039628759,0.0401422009,0.3963153958,0.188486591,0.1458002776,-0.42859146,0.0199723877,0.0063629807,-0.1561582386,0.0699793026,0.1950140893,-0.4718109071,0.3989383578,-0.1540239602,-0.1372059584,-0.0209685918,0.2605495751,0.0444882326,-0.267472446,0.0987746865,0.1068307459,-0.0973727331,0.2509978712,0.0981686786,-0.0154077979,0.076826185,-0.3066434562,-0.3827336729,0.1194452196,-0.2119664103,0.5681127906,0.2654306591,0.1176436022,0.3349756002,0.1474426836,0.0042988327,0.1244659871,-0.1284873635,-0.0136926407,0.0300500039,0.1223837361,-0.2138327956,-0.0197317768,0.2193788439,0.0380547084,-0.393679142,-0.161359027,-0.0826720819,0.0640948936,-0.1460669786,0.0284341387,0.0239791553,-0.2233829945,-0.467312336,0.0605952144,-0.3597423434,-0.2393892854,0.1097849011,-0.025140468,-0.4206591249,0.2320714295,0.0984065682,-0.2648565769,0.097448267,0.4654233754,0.1427570283,0.3938289285,0.6578114629,0.4816328287,-0.1619143486,-0.2177283466,-0.1378337741,0.2559019625,-0.5192023516,0.2636250556,0.1261709332,-0.1565431207,-0.024752453,0.3667325675,0.0422928073,0.2563848495,-0.2222355306,-0.284660548,-0.4323421121,0.0412186794,-0.0124555873,0.1584905982,0.0377329327,0.4392094314,0.0725223646,0.344676882,-0.3166029155,0.1015893668,-0.4592546225,0.127050668,0.0869547948,-0.3884992898,0.00718047,0.1028387994,0.1086364761,0.3089999557,-0.2994246483,-0.2348469496,-0.2264022082,0.0722892433,-0.0692547038,-0.0594015829,0.2124275714,0.2266760767,-0.0329504758,-0.1760732979,-0.0075875646,0.052118022,-0.0104456767,0.1474796236,0.2194553912,0.0195672642,-0.1188956872,0.0047815116,-0.240739122,0.0302963369,0.1427271068,0.0124206934,0.1459383219,0.0674338713,-0.1345796138,-0.1771808118,0.0809547603,0.1359450519,-0.190765202,-0.3552043736,-0.2111646235,-0.1962959766,0.2878009677,0.1750219911,-0.1338817179,0.0425298698,0.0711072236,0.229493916,-0.3894515038,0.1009596959,0.3159351945,-0.0812982693,0.2417636067,0.2346811593,-0.1067333743,0.1655505002,-0.2935241759,0.126708895,0.0022715835,-0.4911993146,-0.0092727225,0.0905084088,-0.1272123307,0.1393156946,0.3981228471,-0.0438104644,0.2180084288,0.1734299213,-0.0295288283,0.2861862779,-0.3371940255,0.1603489518,0.0428447984,-0.6167395711,0.2229512334,0.3042711914,-0.1545279771,0.2104459554,-0.0307244956,0.3675802946,-0.5919659734,-0.1959854066,-0.0851378292,0.2790286541,-0.3977425098,-0.419826895,-0.0324660949,-0.0547241345,-0.2179806381,0.0574031733,-0.071031034,0.1367316544,0.3471319675,0.0215960611,-0.1445557326,-0.3487926722,-0.1260674745,-0.0126963854,0.1340655684,-0.2956922054,0.0682361051,0.2749158442,-0.0459633209,-0.5047225356,0.2228365391,0.5591908097,0.3988352418,-0.3692156076,0.0725747719,-0.039594613,0.0618858784,-0.1044700444,0.0768813267,0.0960938558,0.1491709203,0.2230308801,0.1165978387,-0.2096038312,-0.0821885839,0.1226791218,0.3338048458,0.0517791621,0.4257694781,-0.0936316997,-0.2233311534,-0.0471887253,0.1381967813,-0.4184854329,0.2153057754,0.5088744164,-0.2176353484,0.2733813822,-0.0391260795,0.0981447101,-0.0012670336,0.6340998411,0.1495898217,0.0237321984,-0.4357896745,0.2468439192,-0.5064620376,0.0341480039,-0.2610830367,0.143806532,-0.3708487749,0.347661078,0.1202931255,-0.0631867871,-0.0096283695,0.3259037733,0.2164503187,0.3538523316,-0.3152367771,-0.1542017907,-0.2725798786,-0.211403653,0.3093431592,-0.4370567203,0.0129666105,0.1691231281,0.0520184338,-0.3621643782,0.0353428237,0.0055553382,0.1597293019,-0.0082195206,0.2578552365,0.4893649518,-0.2150480598,-0.0147047071,-0.0939578488,-0.0348746069,-0.165054366,0.024586929,0.0982365906,0.5375682116,-0.2332443744,0.3349668682,-0.4291658401,0.1068600491,0.0017594325,0.0730224401,-0.1125725955,0.1844571829,0.144515112,0.0153929098,0.0043698065,0.3746371269,0.012435195,-0.0591237508,-0.196019277,-0.2834496796,0.6401267648,-0.2923009396,-0.4539929032,-0.4232752621,0.1518884748,-0.2287297398,-0.0712872669,-0.5009861588,0.0414506793,0.1367105395,-0.0241665188,-0.4582684636,0.238963291,0.0122072957,-0.2194100767,-0.0912206173,0.3244309127,0.1423429251,-0.2926914096,0.1541839242,-0.3422243297]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1769","title":"_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union when calling datasets.map with num_proc=2","comments":"Multiprocessing in python require all the functions to be picklable. More specifically, functions need to be picklable with `dill`.\r\n\r\nHowever objects like `typing.Union[str, NoneType]` are not picklable in python <3.7.\r\nCan you try to update your python version to python>=3.7 ?\r\n","body":"It may be a bug of multiprocessing with Datasets, when I disable the multiprocessing by set num_proc to None, everything works fine.\r\n\r\nThe script I use is https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py\r\n\r\nScript args:\r\n\r\n```\r\n--model_name_or_path\r\n..\/..\/..\/model\/chinese-roberta-wwm-ext\r\n--train_file\r\n\/nfs\/volume-377-2\/bert\/data\/test\/train.txt\r\n--output_dir\r\ntest\r\n--do_train\r\n--per_device_train_batch_size\r\n2\r\n--gradient_accumulation_steps\r\n2\r\n--learning_rate\r\n1e-4\r\n--max_steps\r\n1000\r\n--warmup_steps\r\n10\r\n--save_steps\r\n1000\r\n--save_total_limit\r\n1\r\n--seed\r\n23333\r\n--max_seq_length\r\n512\r\n--preprocessing_num_workers\r\n2\r\n--cache_dir\r\n\/nfs\/volume-377-2\/bert\/data\/test\/cache\r\n```\r\n\r\nWhere the `\/nfs\/volume-377-2\/bert\/data\/test\/train.txt` is just a toy example with 10000 lines of random string, you should be able to reproduce this error esaily.\r\n\r\nFull Traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 398, in <module>\r\n main()\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 325, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\r\n","comment_length":41,"text":"_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union when calling datasets.map with num_proc=2\nIt may be a bug of multiprocessing with Datasets, when I disable the multiprocessing by set num_proc to None, everything works fine.\r\n\r\nThe script I use is https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm_wwm.py\r\n\r\nScript args:\r\n\r\n```\r\n--model_name_or_path\r\n..\/..\/..\/model\/chinese-roberta-wwm-ext\r\n--train_file\r\n\/nfs\/volume-377-2\/bert\/data\/test\/train.txt\r\n--output_dir\r\ntest\r\n--do_train\r\n--per_device_train_batch_size\r\n2\r\n--gradient_accumulation_steps\r\n2\r\n--learning_rate\r\n1e-4\r\n--max_steps\r\n1000\r\n--warmup_steps\r\n10\r\n--save_steps\r\n1000\r\n--save_total_limit\r\n1\r\n--seed\r\n23333\r\n--max_seq_length\r\n512\r\n--preprocessing_num_workers\r\n2\r\n--cache_dir\r\n\/nfs\/volume-377-2\/bert\/data\/test\/cache\r\n```\r\n\r\nWhere the `\/nfs\/volume-377-2\/bert\/data\/test\/train.txt` is just a toy example with 10000 lines of random string, you should be able to reproduce this error esaily.\r\n\r\nFull Traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 398, in <module>\r\n main()\r\n File \"\/nfs\/volume-377-2\/bert\/transformers\/examples\/language-modeling\/run_mlm_wwm.py\", line 325, in main\r\n load_from_cache_file=not data_args.overwrite_cache,\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1318, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/connection.py\", line 209, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/multiprocess\/reduction.py\", line 54, in dumps\r\n cls(buf, protocol, *args, **kwds).dump(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/home\/luban\/miniconda3\/envs\/py36\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\r\n\nMultiprocessing in python require all the functions to be picklable. More specifically, functions need to be picklable with `dill`.\r\n\r\nHowever objects like `typing.Union[str, NoneType]` are not picklable in python <3.7.\r\nCan you try to update your python version to python>=3.7 ?\r\n","embeddings":[-0.2903597653,-0.2974283993,0.1132445857,0.1903711855,0.166158393,-0.0522105806,0.3321136534,0.2303635925,0.29939574,0.2801507711,0.0167170409,0.3376862705,-0.1110442802,0.2174146324,0.0041834014,-0.1060649753,-0.1158019528,-0.1348113567,-0.3830081224,0.0106051518,-0.2922798097,0.3295297623,-0.3364185989,0.178581655,-0.6447761059,-0.0980493426,-0.0480723754,0.2914254069,0.0867962465,-0.0838413835,0.091296576,-0.0355660208,0.0124873333,0.495975703,-0.0001100203,0.0433959849,0.2030376494,-0.0829199255,-0.0667256936,-0.4173735082,0.0713781044,-0.0390120484,0.070758529,-0.0548069626,0.1504491121,0.1799487174,0.1406906098,-0.0804709494,0.2613464594,0.1155149564,0.1944084316,0.5662001371,0.1144360006,-0.076135695,-0.1252603531,0.2157489657,-0.0322175771,0.2660860121,0.1158297211,-0.1664748639,-0.2567045093,0.2081440091,-0.0091969082,0.2851698399,-0.1553526968,0.3262346983,0.0900292844,-0.2589209974,0.1284050196,0.3979728222,0.0784635693,-0.232056573,-0.3231726289,-0.0680227727,-0.0854364112,-0.1713582575,0.0877538249,0.1660903543,-0.1176943481,0.0851684958,-0.2029177397,0.1319692582,0.2207043022,0.1039848626,0.2320752591,0.3622331023,-0.031342186,0.3686209619,0.2795686424,-0.1373411715,-0.2073924094,-0.1868893355,0.0720700994,0.0632183775,-0.0458192751,-0.1408384144,0.045508083,-0.0947569758,0.073656261,-0.1785828322,-0.1766469479,0.1712978035,-0.2389756292,0.3193747997,0.3000181019,-0.1272128671,-0.0990271121,0.2027037889,0.4027116001,-0.0739725083,-0.2251318693,0.1159959584,0.3312569559,-0.3447957933,-0.1041944399,0.1564332843,0.329223603,-0.1115081087,-0.149697721,-0.1945511699,-0.1339909732,-0.1725845933,-0.010302064,0.2903743684,0.0592746921,0.3461214006,0.0566079207,0.1878037006,-0.5467142463,-0.0829292685,-0.3398105502,0.218926996,-0.1916940212,0.0626075044,0.2220834941,0.2673146129,0.3550154865,0.1994703859,-0.0371265337,-0.1523097605,0.2351196259,-0.2494036108,-0.1287895143,0.1925622076,0.1652551144,0.2886993587,0.2680864036,-0.3080254495,-0.0635940209,0.0783043131,-0.2370546311,-0.2890685201,-0.0729562566,0.139649719,-0.0142909233,0.1257887632,-0.1699093133,0.0447562262,0.4315924942,-0.0497591682,-0.1882438213,-0.2977280915,-0.3939255178,-0.2207495868,-0.0378487483,0.3762208223,-0.2341665477,-0.1520403773,0.1977583915,-0.1235371456,0.5048832893,0.5401949883,-0.2475932837,-0.0822768435,-0.1492327303,0.6606734991,0.0360711515,0.1096854657,-0.2572190762,0.202608645,-0.0988949314,-0.1114991456,-0.0841935053,0.0370499715,-0.3572835028,-0.0326678194,0.4061434865,0.1351237893,0.0573740341,0.192356661,-0.271171689,-0.3141673207,0.1697528809,-0.0544769354,0.0685291588,-0.2757891119,-0.1262007654,-0.0896317139,0.4380550981,-0.2241077721,0.0792603567,-0.0038885754,0.1627585292,-0.0515407771,-0.0201107692,-0.2338376492,-0.5796083808,0.2878185511,-0.2233290672,0.2118987888,-0.064044036,-0.1344355494,0.2768068016,-0.0169793814,-0.3259687424,-0.0686179921,0.0756605044,0.0722182989,0.0509032123,-0.1822725981,-0.0842780322,-0.1165589243,-0.1049529836,-0.0149970902,-0.4686059952,0.0425728932,-0.4037323892,-0.4181522727,-0.189516753,0.2076245993,0.1067516208,-0.0068623866,-0.1171695143,0.1369666457,0.1502617449,-0.0461106859,-0.0495142005,0.0562658869,-0.0190868974,-0.0961284563,0.021799013,0.219856739,0.1103201881,-0.1780940443,-0.2147070765,0.4751463234,0.0284074377,0.2698736787,0.0416779704,0.0957993492,0.2242285013,0.0413596444,0.1322552115,-0.1709280461,0.1677090824,0.1714064628,0.1670323014,0.0913200006,-0.239271,0.039523486,0.5964310765,0.2025493234,0.4119851291,0.0081310831,-0.2059065104,-0.1373883784,-0.1942453682,0.0325800702,0.6064072251,0.1613005549,0.0643860176,0.001883429,-0.1250701845,-0.1186368167,0.0121865347,-0.0157128964,0.0383681133,0.1391063333,0.110177435,-0.0765566528,-0.303314656,-0.1031182781,0.01030437,0.3184674084,-0.3177983165,0.0145384418,-0.3326050341,0.4729814231,-0.2347498834,0.0097457618,-0.1918038428,-0.2922730744,-0.1737028062,0.3688523471,-0.2211097777,0.1546901911,0.1278085411,-0.2074619383,-0.1040067896,0.2358911335,0.0728950351,-0.0771441162,-0.2809194922,-0.0019136857,0.4048869908,-0.0430667438,0.0471233055,0.1536123306,-0.2586254776,-0.0553867705,-0.13959831,0.0791322961,-0.0761685297,0.2284352332,0.2975741327,0.2531296015,-0.0203261487,-0.3027900457,0.278198719,-0.2991173267,-0.1905634552,-0.0421640575,0.0779534876,-0.1181903258,-0.3645498157,-0.3952235579,-0.4591894448,-0.3345877528,0.4066636562,-0.0599519573,0.1627806872,0.2841716111,0.174965024,0.2511926591,0.1270663142,-0.1767324805,-0.2883633971,0.1619478464,0.22296682,0.0186432451,-0.2608917952,-0.0488841124,-0.1683621109,-0.0411670841,-0.2099181116,-0.1619539857,0.0679083541,-0.4185375571,0.250623107,0.1012951508,0.2987335026,0.6218903065,0.2733477652,-0.1571446955,-0.0458586663,-0.2676429152,0.1766682565,0.2072681785,0.2516293228,-0.0044127959,0.1994465888,-0.0090489136,0.7935594916,0.2588758171,-0.2274359912,0.2747074664,0.0024317801,-0.1775773615,-0.0838076547,-0.157329604,-0.063519977,-0.323271215,0.1069680005,0.2216331661,-0.1139478162,-0.1790085435,0.0517217256,0.1503136307,-0.0914623663,-0.2932106853,0.4917781651,0.2086156756,-0.1075986624,0.0574083813,-0.0566314757,-0.1271755695,0.0405725762,-0.0875590742,0.0143404966,0.0110158855,-0.1077002361,-0.176575616,-0.2057750374,-0.507540524,0.503590703,0.1288685203,0.3068889678,-0.1078222767,-0.2095698863,-0.0616324544,-0.0828850344,0.8502633572,-0.3850911558,0.0100383013,0.119369559,-0.2391629219,-0.2514419258,-0.2370387465,-0.2766326666,0.3788052201,0.3212099075,0.4916625619,-0.2372591496,-0.1024899855,0.3546540141,0.0173880365,-0.1381927431,-0.3041334152,-0.3489311337,-0.2444817573,-0.4569969475,-0.0768626481,0.2236945331,0.1462452263,-0.1781131774,0.0569178686,-0.0031951673,-0.249013409,0.2492820919,0.025196746,0.1551803052,0.2063275874,0.0066008242,-0.0015961076,0.2242492288,-0.1629241556,0.2497969717,-0.1691099554,0.0070374175,0.294501245,-0.2529675364,0.3081232309,0.1772142202,0.0588466898,0.3214899004,0.2185170501,0.3157062232,-0.0787473843,0.1128968671,0.321444124,0.2550255358,-0.1521228403,-0.2701735795,0.5763700008,0.2069915682,0.0711066574,0.2796461582,-0.2895084321,-0.2001235336,0.195239231,-0.0853109583,0.8592373729,-0.0334537104,0.1365751326,0.1039628759,0.0401422009,0.3963153958,0.188486591,0.1458002776,-0.42859146,0.0199723877,0.0063629807,-0.1561582386,0.0699793026,0.1950140893,-0.4718109071,0.3989383578,-0.1540239602,-0.1372059584,-0.0209685918,0.2605495751,0.0444882326,-0.267472446,0.0987746865,0.1068307459,-0.0973727331,0.2509978712,0.0981686786,-0.0154077979,0.076826185,-0.3066434562,-0.3827336729,0.1194452196,-0.2119664103,0.5681127906,0.2654306591,0.1176436022,0.3349756002,0.1474426836,0.0042988327,0.1244659871,-0.1284873635,-0.0136926407,0.0300500039,0.1223837361,-0.2138327956,-0.0197317768,0.2193788439,0.0380547084,-0.393679142,-0.161359027,-0.0826720819,0.0640948936,-0.1460669786,0.0284341387,0.0239791553,-0.2233829945,-0.467312336,0.0605952144,-0.3597423434,-0.2393892854,0.1097849011,-0.025140468,-0.4206591249,0.2320714295,0.0984065682,-0.2648565769,0.097448267,0.4654233754,0.1427570283,0.3938289285,0.6578114629,0.4816328287,-0.1619143486,-0.2177283466,-0.1378337741,0.2559019625,-0.5192023516,0.2636250556,0.1261709332,-0.1565431207,-0.024752453,0.3667325675,0.0422928073,0.2563848495,-0.2222355306,-0.284660548,-0.4323421121,0.0412186794,-0.0124555873,0.1584905982,0.0377329327,0.4392094314,0.0725223646,0.344676882,-0.3166029155,0.1015893668,-0.4592546225,0.127050668,0.0869547948,-0.3884992898,0.00718047,0.1028387994,0.1086364761,0.3089999557,-0.2994246483,-0.2348469496,-0.2264022082,0.0722892433,-0.0692547038,-0.0594015829,0.2124275714,0.2266760767,-0.0329504758,-0.1760732979,-0.0075875646,0.052118022,-0.0104456767,0.1474796236,0.2194553912,0.0195672642,-0.1188956872,0.0047815116,-0.240739122,0.0302963369,0.1427271068,0.0124206934,0.1459383219,0.0674338713,-0.1345796138,-0.1771808118,0.0809547603,0.1359450519,-0.190765202,-0.3552043736,-0.2111646235,-0.1962959766,0.2878009677,0.1750219911,-0.1338817179,0.0425298698,0.0711072236,0.229493916,-0.3894515038,0.1009596959,0.3159351945,-0.0812982693,0.2417636067,0.2346811593,-0.1067333743,0.1655505002,-0.2935241759,0.126708895,0.0022715835,-0.4911993146,-0.0092727225,0.0905084088,-0.1272123307,0.1393156946,0.3981228471,-0.0438104644,0.2180084288,0.1734299213,-0.0295288283,0.2861862779,-0.3371940255,0.1603489518,0.0428447984,-0.6167395711,0.2229512334,0.3042711914,-0.1545279771,0.2104459554,-0.0307244956,0.3675802946,-0.5919659734,-0.1959854066,-0.0851378292,0.2790286541,-0.3977425098,-0.419826895,-0.0324660949,-0.0547241345,-0.2179806381,0.0574031733,-0.071031034,0.1367316544,0.3471319675,0.0215960611,-0.1445557326,-0.3487926722,-0.1260674745,-0.0126963854,0.1340655684,-0.2956922054,0.0682361051,0.2749158442,-0.0459633209,-0.5047225356,0.2228365391,0.5591908097,0.3988352418,-0.3692156076,0.0725747719,-0.039594613,0.0618858784,-0.1044700444,0.0768813267,0.0960938558,0.1491709203,0.2230308801,0.1165978387,-0.2096038312,-0.0821885839,0.1226791218,0.3338048458,0.0517791621,0.4257694781,-0.0936316997,-0.2233311534,-0.0471887253,0.1381967813,-0.4184854329,0.2153057754,0.5088744164,-0.2176353484,0.2733813822,-0.0391260795,0.0981447101,-0.0012670336,0.6340998411,0.1495898217,0.0237321984,-0.4357896745,0.2468439192,-0.5064620376,0.0341480039,-0.2610830367,0.143806532,-0.3708487749,0.347661078,0.1202931255,-0.0631867871,-0.0096283695,0.3259037733,0.2164503187,0.3538523316,-0.3152367771,-0.1542017907,-0.2725798786,-0.211403653,0.3093431592,-0.4370567203,0.0129666105,0.1691231281,0.0520184338,-0.3621643782,0.0353428237,0.0055553382,0.1597293019,-0.0082195206,0.2578552365,0.4893649518,-0.2150480598,-0.0147047071,-0.0939578488,-0.0348746069,-0.165054366,0.024586929,0.0982365906,0.5375682116,-0.2332443744,0.3349668682,-0.4291658401,0.1068600491,0.0017594325,0.0730224401,-0.1125725955,0.1844571829,0.144515112,0.0153929098,0.0043698065,0.3746371269,0.012435195,-0.0591237508,-0.196019277,-0.2834496796,0.6401267648,-0.2923009396,-0.4539929032,-0.4232752621,0.1518884748,-0.2287297398,-0.0712872669,-0.5009861588,0.0414506793,0.1367105395,-0.0241665188,-0.4582684636,0.238963291,0.0122072957,-0.2194100767,-0.0912206173,0.3244309127,0.1423429251,-0.2926914096,0.1541839242,-0.3422243297]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1766","title":"Issues when run two programs compute the same metrics","comments":"Hi ! To avoid collisions you can specify a `experiment_id` when instantiating your metric using `load_metric`. It will replace \"default_experiment\" with the experiment id that you provide in the arrow filename. \r\n\r\nAlso when two `experiment_id` collide we're supposed to detect it using our locking mechanism. Not sure why it didn't work in your case. Could you share some code that reproduces the issue ? This would help us investigate.","body":"I got the following error when running two different programs that both compute sacreblue metrics. It seems that both read\/and\/write to the same location (.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow) where it caches the batches:\r\n\r\n```\r\nFile \"train_matching_min.py\", line 160, in <module>ch_9_label\r\n avg_loss = valid(epoch, args.batch, args.validation, args.with_label)\r\n File \"train_matching_min.py\", line 93, in valid\r\n bleu += eval.compute()\r\n File \"\/u\/tlhoang\/projects\/seal\/match\/models\/eval.py\", line 23, in compute\r\n return self.metric.compute()['score']\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/metric.py\", line 387, in compute\r\n self._finalize()\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/metric.py\", line 355, in _finalize\r\n self.data = Dataset(**reader.read_files([{\"filename\": f} for f in file_paths]))\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 231, in read_files\r\n pa_table = self._read_files(files)\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 170, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 299, in _get_dataset_from_filename\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 481, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Expected to read 1819307375 metadata bytes, but only read 454396\r\n``` ","comment_length":69,"text":"Issues when run two programs compute the same metrics\nI got the following error when running two different programs that both compute sacreblue metrics. It seems that both read\/and\/write to the same location (.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow) where it caches the batches:\r\n\r\n```\r\nFile \"train_matching_min.py\", line 160, in <module>ch_9_label\r\n avg_loss = valid(epoch, args.batch, args.validation, args.with_label)\r\n File \"train_matching_min.py\", line 93, in valid\r\n bleu += eval.compute()\r\n File \"\/u\/tlhoang\/projects\/seal\/match\/models\/eval.py\", line 23, in compute\r\n return self.metric.compute()['score']\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/metric.py\", line 387, in compute\r\n self._finalize()\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/metric.py\", line 355, in _finalize\r\n self.data = Dataset(**reader.read_files([{\"filename\": f} for f in file_paths]))\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 231, in read_files\r\n pa_table = self._read_files(files)\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 170, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 299, in _get_dataset_from_filename\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 481, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Expected to read 1819307375 metadata bytes, but only read 454396\r\n``` \nHi ! To avoid collisions you can specify a `experiment_id` when instantiating your metric using `load_metric`. It will replace \"default_experiment\" with the experiment id that you provide in the arrow filename. \r\n\r\nAlso when two `experiment_id` collide we're supposed to detect it using our locking mechanism. Not sure why it didn't work in your case. Could you share some code that reproduces the issue ? This would help us investigate.","embeddings":[-0.3611746132,-0.1505349725,-0.0431874171,0.3872220218,0.267059505,-0.1068402603,0.0815510601,0.3045549393,-0.1409017295,0.2166149169,-0.3760763705,-0.0134595446,0.0852669552,0.0051951916,-0.1859996617,-0.0460625142,0.0034708965,-0.0094137192,0.0772821903,-0.0010643345,-0.2850643992,0.227998957,0.0163463801,0.2554321587,-0.1016236469,-0.1028563008,-0.013551767,-0.1215876862,-0.0857034922,-0.4742900133,0.435684979,0.0281861611,-0.1604654789,0.7840995193,-0.0001188638,0.099064149,0.2483720183,0.0258432329,-0.0703957081,-0.3716357052,0.0305294432,-0.0560124815,0.1552931517,-0.2752086818,0.0879628584,-0.1430897117,0.0955486447,-0.5427463651,0.3801245987,0.164743185,0.1546675563,0.5314354897,-0.4868223071,-0.1824205071,0.1283859909,-0.0579498075,-0.0503404476,0.5445207357,0.0008056638,0.2044068873,0.0235877763,0.4144544601,-0.1186394989,0.2631291151,0.0570788719,0.0971547663,0.4060327411,-0.015863508,0.1656940132,-0.1240436584,0.257058531,-0.3838199377,-0.2255284339,0.1382777393,-0.198722437,-0.5372884274,0.1851033419,0.1369577646,-0.1783016771,-0.1763565391,-0.2779467702,0.0720578209,-0.2121041268,-0.1963711381,0.0815001726,-0.2687269449,-0.1574116647,0.164332062,0.1694015861,0.3121425509,0.1240120903,0.0018443464,-0.1553128809,-0.008267967,-0.4027771652,0.054256916,-0.3037120998,-0.3437744081,-0.0953906998,0.2882916033,0.378970623,-0.0451941267,0.2840242088,0.2224460244,0.0309231561,0.7004981041,0.1498498172,0.2902023792,0.2009002417,0.1810625941,-0.3100050688,-0.1913619936,0.1990121007,-0.5310234427,0.2414128631,0.2200143933,-0.0132561866,-0.3235704899,-0.4480689466,0.0615485795,0.1938808113,-0.0467135496,0.1330566704,0.3195029795,0.0339151844,0.0123235732,0.1278748959,0.4262725711,-0.2209419608,0.2100048512,-0.2910306752,0.1002703831,-0.4435735643,0.3646081686,0.0474664383,-0.0839140639,0.2243039757,-0.0320464298,0.403193891,-0.3089308441,0.3652119339,-0.4421754479,-0.0029980617,0.0954085141,0.0782008916,0.225120455,0.2995713949,-0.098140642,-0.081104748,-0.0738897845,-0.187210083,-0.1491118819,0.5365419388,0.1563584208,-0.013178803,0.4289540946,-0.3462406993,-0.0732095689,0.3937607408,-0.2173101753,0.0699690506,0.0065712999,-0.4017104805,-0.3470825553,0.1470744908,0.3551437259,-0.2117386758,-0.0738259852,0.2929493189,0.129828155,-0.0102830911,0.4685421288,-0.2201554477,0.1626340151,-0.2866150141,-0.024593858,0.5620930791,-0.8298754692,-0.3216198087,0.1176252291,-0.163472876,-0.3332692385,0.0825219378,0.0429988578,0.3266257942,-0.0876047909,0.0890730992,-0.2243258208,0.1395228654,-0.1509531438,-0.3216227889,-0.3151447773,-0.0638329834,-0.2678596377,0.2974461913,-0.047140453,0.2434967458,-0.2063741088,-0.0600684881,-0.1215572208,0.0200320091,0.2438928783,0.1884081215,-0.028605096,0.099135004,-0.2850906551,0.2925028503,0.1448740959,-0.5383432508,-0.2961361706,0.1691065133,-0.2005111128,-0.1927509755,-0.0857767761,-0.0706390068,-0.1958038062,0.0659025609,0.136058867,0.2209204584,0.278021425,-0.0576278158,0.3118051291,-0.0584468171,0.0707678422,-0.0973505303,0.0738313496,-0.1232286021,-0.2663716674,-0.0980604962,0.0787088349,0.3763604462,-0.0226036236,-0.1681235582,0.3127592504,-0.1168536097,0.2160870582,0.1792309731,0.1188419238,0.1721316278,-0.2322574258,0.0048484607,-0.0386074148,0.1945935041,-0.0987797379,0.4015487432,0.2953327596,-0.0228908621,0.1296619624,0.2755026519,0.0170342363,-0.1007524133,0.0285557341,0.1727861464,-0.190131247,0.3023916483,-0.0971518382,0.0521629862,-0.1124897897,0.0968029946,-0.1043636352,0.1355577111,0.1627224982,-0.0012208633,-0.1695545763,-0.0369496532,0.073288016,-0.2711120844,0.2888995409,0.7128472328,0.0229041651,0.1526313871,0.1330121607,0.0188372079,-0.1760327816,0.1453155577,0.209442988,-0.1075597852,0.0993787572,0.1662219763,-0.0816874504,-0.0364570096,0.0903407335,0.0381227024,-0.0213953033,-0.1647393852,0.3923781514,-0.0219715536,0.0753559023,-0.5359921455,0.0141770039,-0.0288164504,-0.1689181924,-0.1003929675,0.3601456285,0.0499683656,0.0707454607,-0.2301941812,0.5003135204,-0.275437355,0.2575498521,0.0213136356,-0.1551043391,-0.3020187914,-0.0712804124,0.1685049534,0.0605503395,0.3412746191,-0.1754945517,-0.0699492022,-0.2088632137,-0.1869240254,-0.0449571498,0.2885992825,0.4888947904,-0.0525448881,-0.1393486857,-0.3070409596,-0.3469005823,0.3138327301,-0.4048361182,-0.2079191208,-0.1324315965,0.0150443558,-0.2289836705,-0.1872957051,-0.1068752706,-0.0245066341,-0.2213677019,0.208386302,-0.2001251131,0.0360720046,0.0712476075,-0.1740555614,0.1142761707,0.0748567805,0.1112008393,-0.2778829634,-0.1941749454,0.0222590528,-0.4008953869,-0.1950280666,-0.0420212522,0.1363772899,0.4599738121,0.2368850708,-0.4259492457,-0.4368488193,-0.1682187617,-0.1173759624,0.102015242,-0.0418617651,0.0551192351,-0.0682302192,-0.1335440278,-0.1503509581,-0.1933322251,0.502068758,-0.2520491481,0.06857346,-0.1603963971,-0.1222267151,0.0443004556,0.8054549098,0.4146428406,0.0838827565,0.3142866492,0.2035724372,0.3654498756,0.0039692703,-0.3886853456,0.1884925067,0.0200654455,-0.2714687884,0.12796399,-0.085899882,0.1054664776,-0.1474000365,-0.1158738136,-0.2137659043,-0.0955248252,-0.1489702314,-0.2166232914,0.2357062399,0.0829566717,0.0911733583,-0.3216699958,0.0277164523,0.2208309472,0.2361793518,-0.2803712785,0.0243198089,-0.1450072229,-0.1817682534,-0.222554937,0.5148034096,0.089162387,0.2446999252,-0.2091472745,0.1408783346,0.0523314923,-0.0709715262,0.663815558,-0.0739477873,-0.0868660435,-0.1109795943,-0.0193758514,-0.4094045758,-0.1419025809,0.0550740398,0.1366434842,0.1798198223,0.4839473069,-0.1208600104,-0.0051794075,-0.0153063145,0.0120943002,-0.185464561,-0.2006284297,-0.3357676864,-0.383115381,0.0732771754,0.0737477392,-0.0195125453,0.0452137813,-0.0937864855,0.0940699279,-0.1731128097,0.0474347621,0.2436570674,-0.0133806868,0.5679039359,-0.36991027,0.1197623536,0.2774435878,0.0813160092,0.2905486822,0.2618488073,-0.1487618387,-0.4146742821,0.1154792905,0.0839681029,-0.0068028928,0.4959119856,-0.1706465036,0.1998108923,-0.0143724568,-0.0525066815,0.1222909689,0.0716386065,0.2608548403,0.301256597,-0.0817963481,-0.1803607643,0.0001942078,0.4088627696,-0.1110965312,0.4448748529,-0.3718874454,-0.1304301322,0.1831740588,0.2075397968,1.1415235996,0.1096199825,0.0940427482,0.3654060662,-0.3788008392,0.1448805183,-0.1338844001,0.3521232009,-0.2913300693,-0.426278621,0.0443983078,-0.4234814942,0.0784708038,-0.2582255006,-0.0987935886,0.2994014621,-0.2575712502,0.0228546616,-0.1675747186,-0.0094809374,-0.1764464378,0.1637750417,0.2784411311,0.1187395155,-0.1347374171,0.1918252707,-0.1957799196,-0.20621939,-0.2727337778,-0.1241284832,-0.1999096721,0.1254761368,-0.3591457605,0.5477301478,-0.0289820638,-0.1873505116,-0.0564383268,0.3179163635,1.0567492247,-0.0337516703,-0.0532283708,0.1115203053,0.1322745979,0.0769814104,0.0788669288,-0.4823551774,0.1216792762,-0.1621990651,-0.2046578676,-0.0102409078,0.1712741256,0.0545476303,0.1436229497,0.0065041431,-0.0442720428,-0.1453600824,0.0713083297,0.053824462,-0.038230624,-0.1429397464,0.0365857854,-0.0503247641,-0.2049170136,0.1293086261,-0.2136397809,-0.3586632311,0.1601792574,0.239181906,0.3221595585,-0.3406072855,0.5207618475,-0.4181826711,-0.1647209823,-0.122928001,-0.4281492531,-0.1048446074,-0.4858562946,0.3519490659,-0.3284840286,-0.0230383947,-0.1290413588,0.4235875905,0.3696026504,0.2082636952,-0.0514331646,-0.2575737536,-0.3079732656,0.0665430054,-0.2560645044,0.2094359845,-0.0667820275,0.203317374,-0.1151048541,0.2468768656,-0.2290828228,0.14910914,-0.2235806584,-0.057861153,-0.0239928402,-0.3083412349,0.27396819,0.0583666898,-0.0829618722,0.2449745387,-0.4150684774,-0.1025744602,-0.1703071892,0.1670022011,-0.201087594,-0.166345492,-0.1639753282,-0.0849377215,0.2917098105,-0.0966839641,0.0385239124,0.4952764511,-0.0047269403,-0.1825706512,-0.0801014304,0.3579129875,0.076042302,0.3112907708,0.0722698942,-0.0891168639,0.0455706678,0.2249445617,-0.1751985699,-0.1544536948,-0.2151746005,-0.1035947949,0.1023805737,0.1173027754,-0.2384694964,-0.3789310157,-0.2773679495,0.049262587,0.2394450158,0.1593903005,-0.0221532211,-0.167626828,0.0452718884,0.1217806414,0.0035554131,0.0100833569,0.0712792799,-0.066570878,-0.2697435319,0.2501280606,0.4367051721,-0.0595394559,0.2153592706,0.0835738331,0.3153458536,-0.3591237962,0.4809872806,-0.3079811037,0.2562136054,0.2291851789,0.2896816432,0.1402166486,0.1642458439,-0.1615498662,-0.2578151822,0.3288826048,0.100914456,-0.0656246021,-0.1758930087,-0.183143571,-0.1239003688,-0.1155415699,-0.1164165363,0.1536553055,-0.0742524192,0.465230763,-0.2960209846,-0.3396763504,-0.2351730466,0.0977833644,-0.2720488012,-0.1077874452,0.0314898901,-0.3171126246,0.2278341055,0.0468038023,0.2442206293,0.2121112049,0.2209787518,0.0465209521,-0.2491003424,0.0607201047,-0.2737984657,0.6104670763,0.0576187,-0.1199758649,0.3298225999,0.2280463725,-0.0417494178,0.0415111743,0.6623183489,0.522105813,0.3617999554,0.0910812095,-0.1146947071,0.2342058271,0.1228307337,-0.0147621948,0.1736107171,0.2521833777,-0.1960238665,0.2471695989,0.1507108063,-0.1146790832,0.363612622,-0.0644530505,0.2004002631,-0.2909461558,0.2844320238,-0.1778313518,0.0205781367,-0.1967296153,-0.0421836972,-0.5328439474,0.1024001166,0.2991532087,-0.1790359914,0.2391099483,-0.0970063284,0.071795851,0.1840809584,0.2452759594,0.478064239,0.1257411689,-0.5874685049,0.0541213676,-0.4772076607,0.47644189,0.0592098981,0.416478008,-0.2925382257,0.2583390176,-0.2482527941,-0.1623592228,0.4818668962,0.1885393709,-0.030882854,0.0682155117,-0.5626353025,0.1788057834,0.0410292447,-0.3143729568,0.253207624,-0.3407095075,0.0291234013,0.0009255062,0.0330469795,0.229257077,0.0543095283,0.0794539079,-0.0988799408,0.339712888,0.1724283099,0.3838354051,0.0864274278,0.0152714383,-0.0633856282,-0.1108324975,-0.2412479222,-0.1147983074,0.0356716476,0.2226445526,-0.5101049542,0.0544219427,-0.3817653358,0.2033056915,0.2425129563,-0.2194537669,0.0528791621,-0.0975426659,0.0198445246,0.0163124464,0.452906549,0.4042087197,0.0936904326,0.4554382563,-0.3360853791,-0.129624784,0.5763473511,-0.523494482,-0.2780394554,0.119004555,0.4433909655,0.3935000896,0.0013246259,-0.6530673504,-0.0659773424,0.2789849043,-0.1601079553,-0.0528023168,0.1729199141,-0.3497534692,-0.0094572362,0.1064969003,0.014876266,0.0260494947,-0.3596184254,0.0831908807,-0.2089135349]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1766","title":"Issues when run two programs compute the same metrics","comments":"Thank you for your response. I fixed the issue by set \"keep_in_memory=True\" when load_metric. \r\nI cannot share the entire source code but below is the wrapper I wrote:\r\n\r\n```python\r\nclass Evaluation:\r\n def __init__(self, metric='sacrebleu'):\r\n # self.metric = load_metric(metric, keep_in_memory=True)\r\n self.metric = load_metric(metric)\r\n\r\n def add(self, predictions, references):\r\n self.metric.add_batch(predictions=predictions, references=references)\r\n\r\n def compute(self):\r\n return self.metric.compute()['score']\r\n```\r\n\r\nThen call the given wrapper as follows:\r\n\r\n```python\r\neval = Evaluation(metric='sacrebleu')\r\nfor query, candidates, labels in tqdm(dataset):\r\n predictions = net.generate(query)\r\n references = [[s] for s in labels]\r\n eval.add(predictions, references)\r\n if n % 100 == 0:\r\n bleu += eval.compute()\r\n eval = Evaluation(metric='sacrebleu')","body":"I got the following error when running two different programs that both compute sacreblue metrics. It seems that both read\/and\/write to the same location (.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow) where it caches the batches:\r\n\r\n```\r\nFile \"train_matching_min.py\", line 160, in <module>ch_9_label\r\n avg_loss = valid(epoch, args.batch, args.validation, args.with_label)\r\n File \"train_matching_min.py\", line 93, in valid\r\n bleu += eval.compute()\r\n File \"\/u\/tlhoang\/projects\/seal\/match\/models\/eval.py\", line 23, in compute\r\n return self.metric.compute()['score']\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/metric.py\", line 387, in compute\r\n self._finalize()\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/metric.py\", line 355, in _finalize\r\n self.data = Dataset(**reader.read_files([{\"filename\": f} for f in file_paths]))\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 231, in read_files\r\n pa_table = self._read_files(files)\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 170, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 299, in _get_dataset_from_filename\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 481, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Expected to read 1819307375 metadata bytes, but only read 454396\r\n``` ","comment_length":94,"text":"Issues when run two programs compute the same metrics\nI got the following error when running two different programs that both compute sacreblue metrics. It seems that both read\/and\/write to the same location (.cache\/huggingface\/metrics\/sacrebleu\/default\/default_experiment-1-0.arrow) where it caches the batches:\r\n\r\n```\r\nFile \"train_matching_min.py\", line 160, in <module>ch_9_label\r\n avg_loss = valid(epoch, args.batch, args.validation, args.with_label)\r\n File \"train_matching_min.py\", line 93, in valid\r\n bleu += eval.compute()\r\n File \"\/u\/tlhoang\/projects\/seal\/match\/models\/eval.py\", line 23, in compute\r\n return self.metric.compute()['score']\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/metric.py\", line 387, in compute\r\n self._finalize()\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/metric.py\", line 355, in _finalize\r\n self.data = Dataset(**reader.read_files([{\"filename\": f} for f in file_paths]))\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 231, in read_files\r\n pa_table = self._read_files(files)\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 170, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/dccstor\/know\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_reader.py\", line 299, in _get_dataset_from_filename\r\n pa_table = f.read_all()\r\n File \"pyarrow\/ipc.pxi\", line 481, in pyarrow.lib.RecordBatchReader.read_all\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Expected to read 1819307375 metadata bytes, but only read 454396\r\n``` \nThank you for your response. I fixed the issue by set \"keep_in_memory=True\" when load_metric. \r\nI cannot share the entire source code but below is the wrapper I wrote:\r\n\r\n```python\r\nclass Evaluation:\r\n def __init__(self, metric='sacrebleu'):\r\n # self.metric = load_metric(metric, keep_in_memory=True)\r\n self.metric = load_metric(metric)\r\n\r\n def add(self, predictions, references):\r\n self.metric.add_batch(predictions=predictions, references=references)\r\n\r\n def compute(self):\r\n return self.metric.compute()['score']\r\n```\r\n\r\nThen call the given wrapper as follows:\r\n\r\n```python\r\neval = Evaluation(metric='sacrebleu')\r\nfor query, candidates, labels in tqdm(dataset):\r\n predictions = net.generate(query)\r\n references = [[s] for s in labels]\r\n eval.add(predictions, references)\r\n if n % 100 == 0:\r\n bleu += eval.compute()\r\n eval = Evaluation(metric='sacrebleu')","embeddings":[-0.3611746132,-0.1505349725,-0.0431874171,0.3872220218,0.267059505,-0.1068402603,0.0815510601,0.3045549393,-0.1409017295,0.2166149169,-0.3760763705,-0.0134595446,0.0852669552,0.0051951916,-0.1859996617,-0.0460625142,0.0034708965,-0.0094137192,0.0772821903,-0.0010643345,-0.2850643992,0.227998957,0.0163463801,0.2554321587,-0.1016236469,-0.1028563008,-0.013551767,-0.1215876862,-0.0857034922,-0.4742900133,0.435684979,0.0281861611,-0.1604654789,0.7840995193,-0.0001188638,0.099064149,0.2483720183,0.0258432329,-0.0703957081,-0.3716357052,0.0305294432,-0.0560124815,0.1552931517,-0.2752086818,0.0879628584,-0.1430897117,0.0955486447,-0.5427463651,0.3801245987,0.164743185,0.1546675563,0.5314354897,-0.4868223071,-0.1824205071,0.1283859909,-0.0579498075,-0.0503404476,0.5445207357,0.0008056638,0.2044068873,0.0235877763,0.4144544601,-0.1186394989,0.2631291151,0.0570788719,0.0971547663,0.4060327411,-0.015863508,0.1656940132,-0.1240436584,0.257058531,-0.3838199377,-0.2255284339,0.1382777393,-0.198722437,-0.5372884274,0.1851033419,0.1369577646,-0.1783016771,-0.1763565391,-0.2779467702,0.0720578209,-0.2121041268,-0.1963711381,0.0815001726,-0.2687269449,-0.1574116647,0.164332062,0.1694015861,0.3121425509,0.1240120903,0.0018443464,-0.1553128809,-0.008267967,-0.4027771652,0.054256916,-0.3037120998,-0.3437744081,-0.0953906998,0.2882916033,0.378970623,-0.0451941267,0.2840242088,0.2224460244,0.0309231561,0.7004981041,0.1498498172,0.2902023792,0.2009002417,0.1810625941,-0.3100050688,-0.1913619936,0.1990121007,-0.5310234427,0.2414128631,0.2200143933,-0.0132561866,-0.3235704899,-0.4480689466,0.0615485795,0.1938808113,-0.0467135496,0.1330566704,0.3195029795,0.0339151844,0.0123235732,0.1278748959,0.4262725711,-0.2209419608,0.2100048512,-0.2910306752,0.1002703831,-0.4435735643,0.3646081686,0.0474664383,-0.0839140639,0.2243039757,-0.0320464298,0.403193891,-0.3089308441,0.3652119339,-0.4421754479,-0.0029980617,0.0954085141,0.0782008916,0.225120455,0.2995713949,-0.098140642,-0.081104748,-0.0738897845,-0.187210083,-0.1491118819,0.5365419388,0.1563584208,-0.013178803,0.4289540946,-0.3462406993,-0.0732095689,0.3937607408,-0.2173101753,0.0699690506,0.0065712999,-0.4017104805,-0.3470825553,0.1470744908,0.3551437259,-0.2117386758,-0.0738259852,0.2929493189,0.129828155,-0.0102830911,0.4685421288,-0.2201554477,0.1626340151,-0.2866150141,-0.024593858,0.5620930791,-0.8298754692,-0.3216198087,0.1176252291,-0.163472876,-0.3332692385,0.0825219378,0.0429988578,0.3266257942,-0.0876047909,0.0890730992,-0.2243258208,0.1395228654,-0.1509531438,-0.3216227889,-0.3151447773,-0.0638329834,-0.2678596377,0.2974461913,-0.047140453,0.2434967458,-0.2063741088,-0.0600684881,-0.1215572208,0.0200320091,0.2438928783,0.1884081215,-0.028605096,0.099135004,-0.2850906551,0.2925028503,0.1448740959,-0.5383432508,-0.2961361706,0.1691065133,-0.2005111128,-0.1927509755,-0.0857767761,-0.0706390068,-0.1958038062,0.0659025609,0.136058867,0.2209204584,0.278021425,-0.0576278158,0.3118051291,-0.0584468171,0.0707678422,-0.0973505303,0.0738313496,-0.1232286021,-0.2663716674,-0.0980604962,0.0787088349,0.3763604462,-0.0226036236,-0.1681235582,0.3127592504,-0.1168536097,0.2160870582,0.1792309731,0.1188419238,0.1721316278,-0.2322574258,0.0048484607,-0.0386074148,0.1945935041,-0.0987797379,0.4015487432,0.2953327596,-0.0228908621,0.1296619624,0.2755026519,0.0170342363,-0.1007524133,0.0285557341,0.1727861464,-0.190131247,0.3023916483,-0.0971518382,0.0521629862,-0.1124897897,0.0968029946,-0.1043636352,0.1355577111,0.1627224982,-0.0012208633,-0.1695545763,-0.0369496532,0.073288016,-0.2711120844,0.2888995409,0.7128472328,0.0229041651,0.1526313871,0.1330121607,0.0188372079,-0.1760327816,0.1453155577,0.209442988,-0.1075597852,0.0993787572,0.1662219763,-0.0816874504,-0.0364570096,0.0903407335,0.0381227024,-0.0213953033,-0.1647393852,0.3923781514,-0.0219715536,0.0753559023,-0.5359921455,0.0141770039,-0.0288164504,-0.1689181924,-0.1003929675,0.3601456285,0.0499683656,0.0707454607,-0.2301941812,0.5003135204,-0.275437355,0.2575498521,0.0213136356,-0.1551043391,-0.3020187914,-0.0712804124,0.1685049534,0.0605503395,0.3412746191,-0.1754945517,-0.0699492022,-0.2088632137,-0.1869240254,-0.0449571498,0.2885992825,0.4888947904,-0.0525448881,-0.1393486857,-0.3070409596,-0.3469005823,0.3138327301,-0.4048361182,-0.2079191208,-0.1324315965,0.0150443558,-0.2289836705,-0.1872957051,-0.1068752706,-0.0245066341,-0.2213677019,0.208386302,-0.2001251131,0.0360720046,0.0712476075,-0.1740555614,0.1142761707,0.0748567805,0.1112008393,-0.2778829634,-0.1941749454,0.0222590528,-0.4008953869,-0.1950280666,-0.0420212522,0.1363772899,0.4599738121,0.2368850708,-0.4259492457,-0.4368488193,-0.1682187617,-0.1173759624,0.102015242,-0.0418617651,0.0551192351,-0.0682302192,-0.1335440278,-0.1503509581,-0.1933322251,0.502068758,-0.2520491481,0.06857346,-0.1603963971,-0.1222267151,0.0443004556,0.8054549098,0.4146428406,0.0838827565,0.3142866492,0.2035724372,0.3654498756,0.0039692703,-0.3886853456,0.1884925067,0.0200654455,-0.2714687884,0.12796399,-0.085899882,0.1054664776,-0.1474000365,-0.1158738136,-0.2137659043,-0.0955248252,-0.1489702314,-0.2166232914,0.2357062399,0.0829566717,0.0911733583,-0.3216699958,0.0277164523,0.2208309472,0.2361793518,-0.2803712785,0.0243198089,-0.1450072229,-0.1817682534,-0.222554937,0.5148034096,0.089162387,0.2446999252,-0.2091472745,0.1408783346,0.0523314923,-0.0709715262,0.663815558,-0.0739477873,-0.0868660435,-0.1109795943,-0.0193758514,-0.4094045758,-0.1419025809,0.0550740398,0.1366434842,0.1798198223,0.4839473069,-0.1208600104,-0.0051794075,-0.0153063145,0.0120943002,-0.185464561,-0.2006284297,-0.3357676864,-0.383115381,0.0732771754,0.0737477392,-0.0195125453,0.0452137813,-0.0937864855,0.0940699279,-0.1731128097,0.0474347621,0.2436570674,-0.0133806868,0.5679039359,-0.36991027,0.1197623536,0.2774435878,0.0813160092,0.2905486822,0.2618488073,-0.1487618387,-0.4146742821,0.1154792905,0.0839681029,-0.0068028928,0.4959119856,-0.1706465036,0.1998108923,-0.0143724568,-0.0525066815,0.1222909689,0.0716386065,0.2608548403,0.301256597,-0.0817963481,-0.1803607643,0.0001942078,0.4088627696,-0.1110965312,0.4448748529,-0.3718874454,-0.1304301322,0.1831740588,0.2075397968,1.1415235996,0.1096199825,0.0940427482,0.3654060662,-0.3788008392,0.1448805183,-0.1338844001,0.3521232009,-0.2913300693,-0.426278621,0.0443983078,-0.4234814942,0.0784708038,-0.2582255006,-0.0987935886,0.2994014621,-0.2575712502,0.0228546616,-0.1675747186,-0.0094809374,-0.1764464378,0.1637750417,0.2784411311,0.1187395155,-0.1347374171,0.1918252707,-0.1957799196,-0.20621939,-0.2727337778,-0.1241284832,-0.1999096721,0.1254761368,-0.3591457605,0.5477301478,-0.0289820638,-0.1873505116,-0.0564383268,0.3179163635,1.0567492247,-0.0337516703,-0.0532283708,0.1115203053,0.1322745979,0.0769814104,0.0788669288,-0.4823551774,0.1216792762,-0.1621990651,-0.2046578676,-0.0102409078,0.1712741256,0.0545476303,0.1436229497,0.0065041431,-0.0442720428,-0.1453600824,0.0713083297,0.053824462,-0.038230624,-0.1429397464,0.0365857854,-0.0503247641,-0.2049170136,0.1293086261,-0.2136397809,-0.3586632311,0.1601792574,0.239181906,0.3221595585,-0.3406072855,0.5207618475,-0.4181826711,-0.1647209823,-0.122928001,-0.4281492531,-0.1048446074,-0.4858562946,0.3519490659,-0.3284840286,-0.0230383947,-0.1290413588,0.4235875905,0.3696026504,0.2082636952,-0.0514331646,-0.2575737536,-0.3079732656,0.0665430054,-0.2560645044,0.2094359845,-0.0667820275,0.203317374,-0.1151048541,0.2468768656,-0.2290828228,0.14910914,-0.2235806584,-0.057861153,-0.0239928402,-0.3083412349,0.27396819,0.0583666898,-0.0829618722,0.2449745387,-0.4150684774,-0.1025744602,-0.1703071892,0.1670022011,-0.201087594,-0.166345492,-0.1639753282,-0.0849377215,0.2917098105,-0.0966839641,0.0385239124,0.4952764511,-0.0047269403,-0.1825706512,-0.0801014304,0.3579129875,0.076042302,0.3112907708,0.0722698942,-0.0891168639,0.0455706678,0.2249445617,-0.1751985699,-0.1544536948,-0.2151746005,-0.1035947949,0.1023805737,0.1173027754,-0.2384694964,-0.3789310157,-0.2773679495,0.049262587,0.2394450158,0.1593903005,-0.0221532211,-0.167626828,0.0452718884,0.1217806414,0.0035554131,0.0100833569,0.0712792799,-0.066570878,-0.2697435319,0.2501280606,0.4367051721,-0.0595394559,0.2153592706,0.0835738331,0.3153458536,-0.3591237962,0.4809872806,-0.3079811037,0.2562136054,0.2291851789,0.2896816432,0.1402166486,0.1642458439,-0.1615498662,-0.2578151822,0.3288826048,0.100914456,-0.0656246021,-0.1758930087,-0.183143571,-0.1239003688,-0.1155415699,-0.1164165363,0.1536553055,-0.0742524192,0.465230763,-0.2960209846,-0.3396763504,-0.2351730466,0.0977833644,-0.2720488012,-0.1077874452,0.0314898901,-0.3171126246,0.2278341055,0.0468038023,0.2442206293,0.2121112049,0.2209787518,0.0465209521,-0.2491003424,0.0607201047,-0.2737984657,0.6104670763,0.0576187,-0.1199758649,0.3298225999,0.2280463725,-0.0417494178,0.0415111743,0.6623183489,0.522105813,0.3617999554,0.0910812095,-0.1146947071,0.2342058271,0.1228307337,-0.0147621948,0.1736107171,0.2521833777,-0.1960238665,0.2471695989,0.1507108063,-0.1146790832,0.363612622,-0.0644530505,0.2004002631,-0.2909461558,0.2844320238,-0.1778313518,0.0205781367,-0.1967296153,-0.0421836972,-0.5328439474,0.1024001166,0.2991532087,-0.1790359914,0.2391099483,-0.0970063284,0.071795851,0.1840809584,0.2452759594,0.478064239,0.1257411689,-0.5874685049,0.0541213676,-0.4772076607,0.47644189,0.0592098981,0.416478008,-0.2925382257,0.2583390176,-0.2482527941,-0.1623592228,0.4818668962,0.1885393709,-0.030882854,0.0682155117,-0.5626353025,0.1788057834,0.0410292447,-0.3143729568,0.253207624,-0.3407095075,0.0291234013,0.0009255062,0.0330469795,0.229257077,0.0543095283,0.0794539079,-0.0988799408,0.339712888,0.1724283099,0.3838354051,0.0864274278,0.0152714383,-0.0633856282,-0.1108324975,-0.2412479222,-0.1147983074,0.0356716476,0.2226445526,-0.5101049542,0.0544219427,-0.3817653358,0.2033056915,0.2425129563,-0.2194537669,0.0528791621,-0.0975426659,0.0198445246,0.0163124464,0.452906549,0.4042087197,0.0936904326,0.4554382563,-0.3360853791,-0.129624784,0.5763473511,-0.523494482,-0.2780394554,0.119004555,0.4433909655,0.3935000896,0.0013246259,-0.6530673504,-0.0659773424,0.2789849043,-0.1601079553,-0.0528023168,0.1729199141,-0.3497534692,-0.0094572362,0.1064969003,0.014876266,0.0260494947,-0.3596184254,0.0831908807,-0.2089135349]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1765","title":"Error iterating over Dataset with DataLoader","comments":"Instead of:\r\n```python\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n```\r\nIt should be:\r\n```python\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_size=32)\r\n```\r\n\r\n`batch_sampler` accepts a Sampler object or an Iterable, so you get an error.","body":"I have a Dataset that I've mapped a tokenizer over:\r\n\r\n```\r\nencoded_dataset.set_format(type='torch',columns=['attention_mask','input_ids','token_type_ids'])\r\nencoded_dataset[:1]\r\n```\r\n```\r\n{'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]),\r\n 'input_ids': tensor([[ 101, 178, 1198, 1400, 1714, 22233, 21365, 4515, 8618, 1113,\r\n 102]]),\r\n 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]])}\r\n```\r\n\r\nWhen I try to iterate as in the docs, I get errors:\r\n\r\n```\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\nnext(iter(dataloader))\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-45-05180ba8aa35> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataloader.py in __init__(self, loader)\r\n 411 self._timeout = loader.timeout\r\n 412 self._collate_fn = loader.collate_fn\r\n--> 413 self._sampler_iter = iter(self._index_sampler)\r\n 414 self._base_seed = torch.empty((), dtype=torch.int64).random_(generator=loader.generator).item()\r\n 415 self._persistent_workers = loader.persistent_workers\r\n\r\nTypeError: 'int' object is not iterable\r\n\r\n\r\n```","comment_length":30,"text":"Error iterating over Dataset with DataLoader\nI have a Dataset that I've mapped a tokenizer over:\r\n\r\n```\r\nencoded_dataset.set_format(type='torch',columns=['attention_mask','input_ids','token_type_ids'])\r\nencoded_dataset[:1]\r\n```\r\n```\r\n{'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]),\r\n 'input_ids': tensor([[ 101, 178, 1198, 1400, 1714, 22233, 21365, 4515, 8618, 1113,\r\n 102]]),\r\n 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]])}\r\n```\r\n\r\nWhen I try to iterate as in the docs, I get errors:\r\n\r\n```\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\nnext(iter(dataloader))\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-45-05180ba8aa35> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataloader.py in __init__(self, loader)\r\n 411 self._timeout = loader.timeout\r\n 412 self._collate_fn = loader.collate_fn\r\n--> 413 self._sampler_iter = iter(self._index_sampler)\r\n 414 self._base_seed = torch.empty((), dtype=torch.int64).random_(generator=loader.generator).item()\r\n 415 self._persistent_workers = loader.persistent_workers\r\n\r\nTypeError: 'int' object is not iterable\r\n\r\n\r\n```\nInstead of:\r\n```python\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n```\r\nIt should be:\r\n```python\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_size=32)\r\n```\r\n\r\n`batch_sampler` accepts a Sampler object or an Iterable, so you get an error.","embeddings":[-0.2379294932,0.0554564521,-0.0456708856,0.2057333887,0.1531800032,0.00014107,0.7921316028,0.3068522215,0.0211306885,0.1527129412,0.0467958041,0.2304260731,-0.2809546292,-0.2416442186,-0.0902263075,-0.0604503341,-0.0764540955,0.0790703148,-0.1224215627,0.0328939222,-0.1435404718,-0.086729899,-0.1925974041,0.0319360159,-0.2509384155,-0.1322708279,-0.1035913602,-0.0947072953,-0.3341249228,-0.5228338242,0.2897110581,0.1179489493,0.4238479137,0.6812455058,-0.0001170856,0.0532652996,0.1204966977,-0.1216139197,-0.2641813755,-0.232062906,0.1392339319,-0.0516221598,0.0336749442,-0.3547690809,0.2605599761,-0.331272155,-0.017206395,-0.2235834897,0.1987376958,0.2407110631,0.1273105145,0.4878684282,-0.0130664473,0.0686062798,0.04524187,0.2505432963,-0.019511655,0.1971431524,0.1916248351,0.0102274111,-0.3597660363,0.1168679073,-0.1690709442,0.0475237295,0.135968551,-0.1127389744,-0.217760846,-0.1320601255,-0.0326831453,-0.0796288997,0.5732844472,-0.2473982126,-0.3338905275,-0.1655160934,-0.2215303928,-0.2673027515,0.319212079,-0.1650728881,-0.1586042792,-0.2139902264,0.1157573313,0.2405707538,-0.2040264159,0.1367142946,-0.25916031,0.2086893469,-0.0498498268,0.304499954,0.2167702019,-0.2556965649,0.3407104909,0.1005515233,0.0440492071,0.2752284706,-0.5098053217,-0.2515716553,0.0933656469,-0.4813280404,-0.0889511406,0.1817471981,0.2512769699,0.0659232885,-0.0267228652,0.2420649678,0.4608971775,0.1867473274,0.1686100066,0.3430075943,0.0560028069,-0.1008399948,0.1115823984,-0.0628378764,-0.3309622705,-0.304035604,0.2297867239,0.1007610634,0.2355149388,0.0638043508,-0.24075225,-0.0010412553,-0.2854396999,-0.1408834308,0.1115429774,0.1466972083,0.2154821754,0.0784652904,-0.0211962853,0.0494476669,-0.0294592027,-0.1010558978,-0.0631030872,0.1225908026,-0.2360044122,0.0062334132,0.2692131698,-0.0449522883,-0.0005978197,0.1379322708,-0.0163492449,0.1064936593,0.1838307977,-0.1677454114,0.2751210034,-0.066830501,0.0526376925,0.3259077668,0.2427886128,0.0527041815,-0.1517042071,0.0064420728,-0.4252358377,-0.0854079202,-0.0711294115,0.1343848705,-0.0657273754,-0.0596553646,-0.4029016495,-0.1930606961,0.3476107717,-0.1592839658,0.0990208313,-0.2618944049,0.0679650232,-0.0949438438,0.0716987848,0.3785112798,-0.4588417411,-0.1389319897,0.1558684111,0.100578852,0.268353045,0.2800276279,-0.2775180936,0.5750991702,-0.4041410387,0.3146153092,0.3382745087,-0.4575732052,-0.3010725677,0.30232355,-0.1728587449,0.1828440428,0.1311056763,0.0220962148,0.4072544873,-0.2117989957,0.1262145638,0.1312219352,-0.1754143536,-0.084498167,-0.0188895371,0.40735659,0.5204417706,0.0579210706,0.1899082512,0.0911744684,-0.2563124001,0.1866555065,0.0407245047,-0.1624023467,0.0752803087,0.1461818963,0.08729247,0.1154016927,0.2047132254,-0.1541337818,-0.2414141595,0.0174530409,0.08473593,0.1103395596,-0.1569616348,0.1011083499,0.0156297926,0.1833895743,-0.4093043208,-0.1476449072,0.1057991534,0.3391777277,-0.1085056514,-0.2262801528,-0.278257817,0.0101970183,-0.1502237916,0.1229054332,-0.3177072704,0.0756461397,0.0731091648,-0.1621202379,-0.1109917238,0.0012578749,0.2185538709,-0.2305950075,-0.0198753402,0.152743876,0.0733504519,0.0837663114,-0.1409202218,0.0840372369,0.2027758658,-0.0805915147,-0.0350414291,0.0263240449,0.0633541942,-0.2463466823,0.2860279679,0.3557419777,-0.0415121466,0.5438959002,-0.1846143901,0.1865570545,0.0854834467,0.1429632455,-0.2297751456,0.1214375868,0.2915955782,0.165278703,0.2727060616,-0.0742381364,-0.1061879173,-0.0162541084,0.135785684,0.034656506,-0.1672892421,0.0645060986,-0.2023294568,0.078733027,0.1431891769,-0.3583348989,0.2936249375,-0.0419645049,0.1355626434,-0.2297315598,0.1093874052,0.0548776314,-0.0068245805,0.1373674572,0.0501099378,0.1365749091,-0.0179888159,0.032990437,-0.1993452013,-0.2619042993,-0.0321639255,0.4030472338,-0.2017506361,0.420080632,-0.2251237631,-0.0806137398,-0.1357363909,-0.4637432098,-0.1244161502,-0.488583684,-0.2934887409,0.4377196431,-0.0523557365,0.2884217203,0.1593024582,0.0659951046,0.6007106304,-0.1201817542,-0.1299342215,-0.1651609242,-0.1434867382,0.002023285,0.3916653693,-0.3905928731,0.3192386329,-0.14857319,-0.0364399292,-0.4931823909,-0.1713028252,0.23434636,-0.2515105605,-0.1061097458,0.250123322,0.160312131,-0.1276741028,-0.3462354839,0.2360218316,-0.0971406028,-0.0598608814,0.3538429141,-0.0178579725,0.164885059,0.1942387223,-0.1618052274,-0.1143744439,-0.2210423201,0.1488950402,-0.024225384,0.3098029792,0.1279512346,0.0508767813,0.0688194558,0.0538865477,-0.0885351226,-0.0529175512,-0.4608557522,0.4211562574,-0.3188976645,-0.2935510576,-0.0610978752,-0.005195491,0.3583402038,0.374879539,-0.2954478264,-0.0019789448,0.1968028694,0.0960602537,-0.1780664027,0.0133007132,0.1617016345,0.1165711582,-0.0526863299,-0.0528675951,0.1458159238,0.1710808575,-0.1876334697,0.2580990195,0.159302935,0.4765362144,0.2614045143,0.6606869102,0.0389066897,-0.3858367205,0.3321016431,-0.1400644928,-0.0007825629,0.086658068,-0.6534711719,0.1830661446,-0.2897427082,-0.02632441,-0.1086072251,-0.3637517095,-0.2020133138,-0.1655701101,0.1691449881,-0.0271603651,-0.3952352703,0.4657104313,-0.430401504,0.2838457525,-0.0792960823,0.136812672,-0.6179904342,-0.0159705244,0.134028852,0.0014620165,0.2550830841,-0.0820336565,-0.394413799,0.0771882385,-0.4658885002,0.2320282906,0.0853058323,0.7238261104,0.1851679087,-0.2038863599,-0.1605457962,-0.0338941254,0.726137042,0.2351621836,-0.0131854638,0.0982150808,-0.3114118576,-0.4381252229,-0.0794281214,-0.2774267197,0.6424908042,0.1563335955,0.2806633115,-0.2328459471,-0.1301856488,0.1159901023,0.1067452282,-0.2158435285,-0.2748224735,-0.0760231242,-0.0680433884,-0.353223443,-0.014091325,0.0438390039,0.2820525467,0.0923700854,0.0220608246,-0.3040832281,-0.0276663322,0.3623539209,-0.0427861214,0.0660697371,-0.1614816934,0.1583146006,0.1936412007,0.2400769889,0.6411280036,0.3070535362,0.2450909913,-0.4524520338,0.3747632802,0.0567932725,0.2789521813,0.126682505,-0.1120252907,-0.0463845544,-0.3257368505,0.0500912033,-0.4352755249,0.5216366649,0.5002289414,0.1470032632,-0.5732986331,-0.3403778374,0.2020383924,0.1137958467,0.1821293831,0.7328404784,-0.1325454563,-0.1383227408,0.1305194199,0.1368960142,0.7372894287,0.0599588566,0.1241550595,0.4110813141,0.5742589831,0.2874032259,-0.5326961279,0.0999787003,-0.2278511822,-0.0477599874,-0.1629463881,-0.2553313375,0.0387578979,0.1704581827,-0.0596478768,0.1802617759,-0.0350546502,0.5183291435,0.0984523073,0.0610763133,-0.3110624254,-0.2922029495,0.0958925858,0.1132938266,-0.1642979085,0.1464618146,-0.0775095448,-0.0749971867,0.1822136045,-0.2826287746,-0.2507403791,-0.0257558469,-0.2285092771,0.1024926826,0.0329171196,-0.4766200781,0.1010436937,0.0133420769,0.050727874,-0.090359576,0.0723627657,-0.2212266326,-0.0407183506,-0.1306499988,-0.132861495,-0.0846425891,0.4096860588,0.12493743,0.130896911,0.281706512,0.2302460819,-0.3494544327,0.0834648982,0.222866118,0.2381524444,-0.5538777709,0.1067495197,-0.1169768646,-0.200172022,-0.2225140333,0.0751270652,0.0035707487,-0.3264312446,0.00617181,-0.1581954062,-0.0941699371,0.0912848115,0.4193665385,0.0990061983,-0.4401395023,0.7563033104,0.1833338439,-0.2378757447,-0.1601602882,0.1304143965,0.4554119706,-0.2698962986,0.1051627994,-0.3681175113,0.1937119961,-0.2056511492,0.0954531431,0.0749833733,0.0011241743,-0.0847064331,-0.3185295165,-0.376385808,0.0641730577,-0.127597034,0.2709617615,0.0657135397,0.1442403048,-0.2424863875,0.1084674597,-0.2284084707,-0.0599610582,-0.1304063201,0.2170205265,0.0687656924,-0.1155499667,-0.029965708,0.1643112749,0.0875988752,0.3649315536,0.0772954896,-0.2171725333,-0.1124109924,0.1081866995,0.1269553751,-0.2696897686,0.0701325387,-0.0216650311,-0.0106119867,-0.3168860674,0.0920091048,0.0593076125,-0.0337995403,0.2827904522,0.3835873008,0.0453121625,0.1048733965,0.0202995613,0.244404465,0.264428556,0.1860617399,0.3221852183,-0.034697488,-0.0614788644,-0.2507389784,0.3639924824,0.3781985939,-0.1171507537,0.1562584788,-0.4842172861,0.1358973086,-0.1309468895,0.2535530627,0.6113288403,-0.3544901311,-0.1553913802,0.1691128314,0.1126620173,-0.0045485063,0.0386662707,0.0609535091,-0.0578211956,-0.0058272169,0.1932299137,-0.0571368709,-0.0611666217,0.1354420781,-0.0829459876,0.2889599502,-0.1634664685,0.1883116364,0.0641768351,-0.4278978705,-0.0858652741,0.3677484095,0.2824953794,0.1013357863,0.2784427404,-0.1078804582,0.427321136,0.0708391145,-0.2493039072,-0.0165874641,-0.4929739535,-0.3299542665,0.1111973599,0.2463370413,0.1948013008,0.0489695594,0.320019722,-0.0799887404,-0.1117729396,-0.0213638823,0.1956691891,-0.277502358,0.0963109508,-0.3862141073,-0.2213355303,-0.2077884525,-0.1047486588,-0.077270098,-0.1806131005,0.0019157402,0.118787311,-0.1874989718,-0.4298039377,0.0627504662,-0.0045477967,-0.112351805,-0.3342280388,0.2179592699,0.1602630466,0.1170849726,0.0349093527,0.2651892602,0.5087574124,0.4610842764,-0.1494466215,0.1769150198,-0.5010870695,-0.1911614835,-0.0631125867,0.1125183627,-0.4148905277,-0.1014754996,0.3867393434,0.0840830281,-0.1143442765,-0.2153678983,0.1340745538,0.0735940412,0.0001483769,0.1247917414,-0.0815566033,0.0414281264,-0.0814748928,-0.0441691838,-0.4062025249,-0.1182344258,0.4678344429,0.1166635156,0.1968592107,0.0207538828,0.0589621589,-0.2503395975,0.4257987738,0.4754786193,0.2222500294,-0.409289211,-0.2357503623,-0.2958708107,0.3727261722,-0.1213889048,-0.1949660182,-0.0196231641,0.1651484668,0.3228288591,0.2691094577,0.243530497,-0.145322457,-0.0710379779,0.3360802829,-0.3235668242,-0.3320512772,0.035634283,0.2394293398,-0.0671317875,-0.2778089046,0.2334064543,-0.1008600667,0.1332826614,-0.2046568245,-0.3008014262,0.0800730586,0.0444457643,0.4428635836,0.1897007227,0.4272838831,0.0945996493,0.0628773049,-0.278421998,-0.2361575365,-0.1914744675,0.0452262163,0.0448537469,0.2142349631,-0.0324527472,-0.1344611943,-0.5011916161,0.2168956101,-0.1091530472,-0.2911404371,-0.3574871719,0.2795459926,0.0037162364,0.070227012,-0.0714731738,0.2456234843,0.2120578587,0.537260592,-0.3672761619,-0.2944155633,0.4216322899,-0.6915383935,-0.4907265902,0.0250352118,-0.1989057362,0.1786561608,0.27032426,-0.5640615821,0.0031023284,0.317204386,-0.1694213003,-0.1327279061,0.0034326748,0.0969765931,0.0240144264,-0.1689353287,0.3889947236,0.1035356969,-0.0350189246,-0.0177480802,-0.2236268073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1765","title":"Error iterating over Dataset with DataLoader","comments":"@mariosasko I thought that would fix it, but now I'm getting a different error:\r\n\r\n```\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py:851: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at \/pytorch\/torch\/csrc\/utils\/tensor_numpy.cpp:141.)\r\n return torch.tensor(x, **format_kwargs)\r\n---------------------------------------------------------------------------\r\nRuntimeError Traceback (most recent call last)\r\n<ipython-input-20-3af1d82bf93a> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_size=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/_utils\/collate.py in default_collate(batch)\r\n 53 storage = elem.storage()._new_shared(numel)\r\n 54 out = elem.new(storage)\r\n---> 55 return torch.stack(batch, 0, out=out)\r\n 56 elif elem_type.__module__ == 'numpy' and elem_type.__name__ != 'str_' \\\r\n 57 and elem_type.__name__ != 'string_':\r\n\r\nRuntimeError: stack expects each tensor to be equal size, but got [7] at entry 0 and [10] at entry 1\r\n```\r\n\r\nAny thoughts what this means?I Do I need padding?","body":"I have a Dataset that I've mapped a tokenizer over:\r\n\r\n```\r\nencoded_dataset.set_format(type='torch',columns=['attention_mask','input_ids','token_type_ids'])\r\nencoded_dataset[:1]\r\n```\r\n```\r\n{'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]),\r\n 'input_ids': tensor([[ 101, 178, 1198, 1400, 1714, 22233, 21365, 4515, 8618, 1113,\r\n 102]]),\r\n 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]])}\r\n```\r\n\r\nWhen I try to iterate as in the docs, I get errors:\r\n\r\n```\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\nnext(iter(dataloader))\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-45-05180ba8aa35> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataloader.py in __init__(self, loader)\r\n 411 self._timeout = loader.timeout\r\n 412 self._collate_fn = loader.collate_fn\r\n--> 413 self._sampler_iter = iter(self._index_sampler)\r\n 414 self._base_seed = torch.empty((), dtype=torch.int64).random_(generator=loader.generator).item()\r\n 415 self._persistent_workers = loader.persistent_workers\r\n\r\nTypeError: 'int' object is not iterable\r\n\r\n\r\n```","comment_length":169,"text":"Error iterating over Dataset with DataLoader\nI have a Dataset that I've mapped a tokenizer over:\r\n\r\n```\r\nencoded_dataset.set_format(type='torch',columns=['attention_mask','input_ids','token_type_ids'])\r\nencoded_dataset[:1]\r\n```\r\n```\r\n{'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]),\r\n 'input_ids': tensor([[ 101, 178, 1198, 1400, 1714, 22233, 21365, 4515, 8618, 1113,\r\n 102]]),\r\n 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]])}\r\n```\r\n\r\nWhen I try to iterate as in the docs, I get errors:\r\n\r\n```\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\nnext(iter(dataloader))\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-45-05180ba8aa35> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataloader.py in __init__(self, loader)\r\n 411 self._timeout = loader.timeout\r\n 412 self._collate_fn = loader.collate_fn\r\n--> 413 self._sampler_iter = iter(self._index_sampler)\r\n 414 self._base_seed = torch.empty((), dtype=torch.int64).random_(generator=loader.generator).item()\r\n 415 self._persistent_workers = loader.persistent_workers\r\n\r\nTypeError: 'int' object is not iterable\r\n\r\n\r\n```\n@mariosasko I thought that would fix it, but now I'm getting a different error:\r\n\r\n```\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py:851: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at \/pytorch\/torch\/csrc\/utils\/tensor_numpy.cpp:141.)\r\n return torch.tensor(x, **format_kwargs)\r\n---------------------------------------------------------------------------\r\nRuntimeError Traceback (most recent call last)\r\n<ipython-input-20-3af1d82bf93a> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_size=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/_utils\/collate.py in default_collate(batch)\r\n 53 storage = elem.storage()._new_shared(numel)\r\n 54 out = elem.new(storage)\r\n---> 55 return torch.stack(batch, 0, out=out)\r\n 56 elif elem_type.__module__ == 'numpy' and elem_type.__name__ != 'str_' \\\r\n 57 and elem_type.__name__ != 'string_':\r\n\r\nRuntimeError: stack expects each tensor to be equal size, but got [7] at entry 0 and [10] at entry 1\r\n```\r\n\r\nAny thoughts what this means?I Do I need padding?","embeddings":[-0.2379294932,0.0554564521,-0.0456708856,0.2057333887,0.1531800032,0.00014107,0.7921316028,0.3068522215,0.0211306885,0.1527129412,0.0467958041,0.2304260731,-0.2809546292,-0.2416442186,-0.0902263075,-0.0604503341,-0.0764540955,0.0790703148,-0.1224215627,0.0328939222,-0.1435404718,-0.086729899,-0.1925974041,0.0319360159,-0.2509384155,-0.1322708279,-0.1035913602,-0.0947072953,-0.3341249228,-0.5228338242,0.2897110581,0.1179489493,0.4238479137,0.6812455058,-0.0001170856,0.0532652996,0.1204966977,-0.1216139197,-0.2641813755,-0.232062906,0.1392339319,-0.0516221598,0.0336749442,-0.3547690809,0.2605599761,-0.331272155,-0.017206395,-0.2235834897,0.1987376958,0.2407110631,0.1273105145,0.4878684282,-0.0130664473,0.0686062798,0.04524187,0.2505432963,-0.019511655,0.1971431524,0.1916248351,0.0102274111,-0.3597660363,0.1168679073,-0.1690709442,0.0475237295,0.135968551,-0.1127389744,-0.217760846,-0.1320601255,-0.0326831453,-0.0796288997,0.5732844472,-0.2473982126,-0.3338905275,-0.1655160934,-0.2215303928,-0.2673027515,0.319212079,-0.1650728881,-0.1586042792,-0.2139902264,0.1157573313,0.2405707538,-0.2040264159,0.1367142946,-0.25916031,0.2086893469,-0.0498498268,0.304499954,0.2167702019,-0.2556965649,0.3407104909,0.1005515233,0.0440492071,0.2752284706,-0.5098053217,-0.2515716553,0.0933656469,-0.4813280404,-0.0889511406,0.1817471981,0.2512769699,0.0659232885,-0.0267228652,0.2420649678,0.4608971775,0.1867473274,0.1686100066,0.3430075943,0.0560028069,-0.1008399948,0.1115823984,-0.0628378764,-0.3309622705,-0.304035604,0.2297867239,0.1007610634,0.2355149388,0.0638043508,-0.24075225,-0.0010412553,-0.2854396999,-0.1408834308,0.1115429774,0.1466972083,0.2154821754,0.0784652904,-0.0211962853,0.0494476669,-0.0294592027,-0.1010558978,-0.0631030872,0.1225908026,-0.2360044122,0.0062334132,0.2692131698,-0.0449522883,-0.0005978197,0.1379322708,-0.0163492449,0.1064936593,0.1838307977,-0.1677454114,0.2751210034,-0.066830501,0.0526376925,0.3259077668,0.2427886128,0.0527041815,-0.1517042071,0.0064420728,-0.4252358377,-0.0854079202,-0.0711294115,0.1343848705,-0.0657273754,-0.0596553646,-0.4029016495,-0.1930606961,0.3476107717,-0.1592839658,0.0990208313,-0.2618944049,0.0679650232,-0.0949438438,0.0716987848,0.3785112798,-0.4588417411,-0.1389319897,0.1558684111,0.100578852,0.268353045,0.2800276279,-0.2775180936,0.5750991702,-0.4041410387,0.3146153092,0.3382745087,-0.4575732052,-0.3010725677,0.30232355,-0.1728587449,0.1828440428,0.1311056763,0.0220962148,0.4072544873,-0.2117989957,0.1262145638,0.1312219352,-0.1754143536,-0.084498167,-0.0188895371,0.40735659,0.5204417706,0.0579210706,0.1899082512,0.0911744684,-0.2563124001,0.1866555065,0.0407245047,-0.1624023467,0.0752803087,0.1461818963,0.08729247,0.1154016927,0.2047132254,-0.1541337818,-0.2414141595,0.0174530409,0.08473593,0.1103395596,-0.1569616348,0.1011083499,0.0156297926,0.1833895743,-0.4093043208,-0.1476449072,0.1057991534,0.3391777277,-0.1085056514,-0.2262801528,-0.278257817,0.0101970183,-0.1502237916,0.1229054332,-0.3177072704,0.0756461397,0.0731091648,-0.1621202379,-0.1109917238,0.0012578749,0.2185538709,-0.2305950075,-0.0198753402,0.152743876,0.0733504519,0.0837663114,-0.1409202218,0.0840372369,0.2027758658,-0.0805915147,-0.0350414291,0.0263240449,0.0633541942,-0.2463466823,0.2860279679,0.3557419777,-0.0415121466,0.5438959002,-0.1846143901,0.1865570545,0.0854834467,0.1429632455,-0.2297751456,0.1214375868,0.2915955782,0.165278703,0.2727060616,-0.0742381364,-0.1061879173,-0.0162541084,0.135785684,0.034656506,-0.1672892421,0.0645060986,-0.2023294568,0.078733027,0.1431891769,-0.3583348989,0.2936249375,-0.0419645049,0.1355626434,-0.2297315598,0.1093874052,0.0548776314,-0.0068245805,0.1373674572,0.0501099378,0.1365749091,-0.0179888159,0.032990437,-0.1993452013,-0.2619042993,-0.0321639255,0.4030472338,-0.2017506361,0.420080632,-0.2251237631,-0.0806137398,-0.1357363909,-0.4637432098,-0.1244161502,-0.488583684,-0.2934887409,0.4377196431,-0.0523557365,0.2884217203,0.1593024582,0.0659951046,0.6007106304,-0.1201817542,-0.1299342215,-0.1651609242,-0.1434867382,0.002023285,0.3916653693,-0.3905928731,0.3192386329,-0.14857319,-0.0364399292,-0.4931823909,-0.1713028252,0.23434636,-0.2515105605,-0.1061097458,0.250123322,0.160312131,-0.1276741028,-0.3462354839,0.2360218316,-0.0971406028,-0.0598608814,0.3538429141,-0.0178579725,0.164885059,0.1942387223,-0.1618052274,-0.1143744439,-0.2210423201,0.1488950402,-0.024225384,0.3098029792,0.1279512346,0.0508767813,0.0688194558,0.0538865477,-0.0885351226,-0.0529175512,-0.4608557522,0.4211562574,-0.3188976645,-0.2935510576,-0.0610978752,-0.005195491,0.3583402038,0.374879539,-0.2954478264,-0.0019789448,0.1968028694,0.0960602537,-0.1780664027,0.0133007132,0.1617016345,0.1165711582,-0.0526863299,-0.0528675951,0.1458159238,0.1710808575,-0.1876334697,0.2580990195,0.159302935,0.4765362144,0.2614045143,0.6606869102,0.0389066897,-0.3858367205,0.3321016431,-0.1400644928,-0.0007825629,0.086658068,-0.6534711719,0.1830661446,-0.2897427082,-0.02632441,-0.1086072251,-0.3637517095,-0.2020133138,-0.1655701101,0.1691449881,-0.0271603651,-0.3952352703,0.4657104313,-0.430401504,0.2838457525,-0.0792960823,0.136812672,-0.6179904342,-0.0159705244,0.134028852,0.0014620165,0.2550830841,-0.0820336565,-0.394413799,0.0771882385,-0.4658885002,0.2320282906,0.0853058323,0.7238261104,0.1851679087,-0.2038863599,-0.1605457962,-0.0338941254,0.726137042,0.2351621836,-0.0131854638,0.0982150808,-0.3114118576,-0.4381252229,-0.0794281214,-0.2774267197,0.6424908042,0.1563335955,0.2806633115,-0.2328459471,-0.1301856488,0.1159901023,0.1067452282,-0.2158435285,-0.2748224735,-0.0760231242,-0.0680433884,-0.353223443,-0.014091325,0.0438390039,0.2820525467,0.0923700854,0.0220608246,-0.3040832281,-0.0276663322,0.3623539209,-0.0427861214,0.0660697371,-0.1614816934,0.1583146006,0.1936412007,0.2400769889,0.6411280036,0.3070535362,0.2450909913,-0.4524520338,0.3747632802,0.0567932725,0.2789521813,0.126682505,-0.1120252907,-0.0463845544,-0.3257368505,0.0500912033,-0.4352755249,0.5216366649,0.5002289414,0.1470032632,-0.5732986331,-0.3403778374,0.2020383924,0.1137958467,0.1821293831,0.7328404784,-0.1325454563,-0.1383227408,0.1305194199,0.1368960142,0.7372894287,0.0599588566,0.1241550595,0.4110813141,0.5742589831,0.2874032259,-0.5326961279,0.0999787003,-0.2278511822,-0.0477599874,-0.1629463881,-0.2553313375,0.0387578979,0.1704581827,-0.0596478768,0.1802617759,-0.0350546502,0.5183291435,0.0984523073,0.0610763133,-0.3110624254,-0.2922029495,0.0958925858,0.1132938266,-0.1642979085,0.1464618146,-0.0775095448,-0.0749971867,0.1822136045,-0.2826287746,-0.2507403791,-0.0257558469,-0.2285092771,0.1024926826,0.0329171196,-0.4766200781,0.1010436937,0.0133420769,0.050727874,-0.090359576,0.0723627657,-0.2212266326,-0.0407183506,-0.1306499988,-0.132861495,-0.0846425891,0.4096860588,0.12493743,0.130896911,0.281706512,0.2302460819,-0.3494544327,0.0834648982,0.222866118,0.2381524444,-0.5538777709,0.1067495197,-0.1169768646,-0.200172022,-0.2225140333,0.0751270652,0.0035707487,-0.3264312446,0.00617181,-0.1581954062,-0.0941699371,0.0912848115,0.4193665385,0.0990061983,-0.4401395023,0.7563033104,0.1833338439,-0.2378757447,-0.1601602882,0.1304143965,0.4554119706,-0.2698962986,0.1051627994,-0.3681175113,0.1937119961,-0.2056511492,0.0954531431,0.0749833733,0.0011241743,-0.0847064331,-0.3185295165,-0.376385808,0.0641730577,-0.127597034,0.2709617615,0.0657135397,0.1442403048,-0.2424863875,0.1084674597,-0.2284084707,-0.0599610582,-0.1304063201,0.2170205265,0.0687656924,-0.1155499667,-0.029965708,0.1643112749,0.0875988752,0.3649315536,0.0772954896,-0.2171725333,-0.1124109924,0.1081866995,0.1269553751,-0.2696897686,0.0701325387,-0.0216650311,-0.0106119867,-0.3168860674,0.0920091048,0.0593076125,-0.0337995403,0.2827904522,0.3835873008,0.0453121625,0.1048733965,0.0202995613,0.244404465,0.264428556,0.1860617399,0.3221852183,-0.034697488,-0.0614788644,-0.2507389784,0.3639924824,0.3781985939,-0.1171507537,0.1562584788,-0.4842172861,0.1358973086,-0.1309468895,0.2535530627,0.6113288403,-0.3544901311,-0.1553913802,0.1691128314,0.1126620173,-0.0045485063,0.0386662707,0.0609535091,-0.0578211956,-0.0058272169,0.1932299137,-0.0571368709,-0.0611666217,0.1354420781,-0.0829459876,0.2889599502,-0.1634664685,0.1883116364,0.0641768351,-0.4278978705,-0.0858652741,0.3677484095,0.2824953794,0.1013357863,0.2784427404,-0.1078804582,0.427321136,0.0708391145,-0.2493039072,-0.0165874641,-0.4929739535,-0.3299542665,0.1111973599,0.2463370413,0.1948013008,0.0489695594,0.320019722,-0.0799887404,-0.1117729396,-0.0213638823,0.1956691891,-0.277502358,0.0963109508,-0.3862141073,-0.2213355303,-0.2077884525,-0.1047486588,-0.077270098,-0.1806131005,0.0019157402,0.118787311,-0.1874989718,-0.4298039377,0.0627504662,-0.0045477967,-0.112351805,-0.3342280388,0.2179592699,0.1602630466,0.1170849726,0.0349093527,0.2651892602,0.5087574124,0.4610842764,-0.1494466215,0.1769150198,-0.5010870695,-0.1911614835,-0.0631125867,0.1125183627,-0.4148905277,-0.1014754996,0.3867393434,0.0840830281,-0.1143442765,-0.2153678983,0.1340745538,0.0735940412,0.0001483769,0.1247917414,-0.0815566033,0.0414281264,-0.0814748928,-0.0441691838,-0.4062025249,-0.1182344258,0.4678344429,0.1166635156,0.1968592107,0.0207538828,0.0589621589,-0.2503395975,0.4257987738,0.4754786193,0.2222500294,-0.409289211,-0.2357503623,-0.2958708107,0.3727261722,-0.1213889048,-0.1949660182,-0.0196231641,0.1651484668,0.3228288591,0.2691094577,0.243530497,-0.145322457,-0.0710379779,0.3360802829,-0.3235668242,-0.3320512772,0.035634283,0.2394293398,-0.0671317875,-0.2778089046,0.2334064543,-0.1008600667,0.1332826614,-0.2046568245,-0.3008014262,0.0800730586,0.0444457643,0.4428635836,0.1897007227,0.4272838831,0.0945996493,0.0628773049,-0.278421998,-0.2361575365,-0.1914744675,0.0452262163,0.0448537469,0.2142349631,-0.0324527472,-0.1344611943,-0.5011916161,0.2168956101,-0.1091530472,-0.2911404371,-0.3574871719,0.2795459926,0.0037162364,0.070227012,-0.0714731738,0.2456234843,0.2120578587,0.537260592,-0.3672761619,-0.2944155633,0.4216322899,-0.6915383935,-0.4907265902,0.0250352118,-0.1989057362,0.1786561608,0.27032426,-0.5640615821,0.0031023284,0.317204386,-0.1694213003,-0.1327279061,0.0034326748,0.0969765931,0.0240144264,-0.1689353287,0.3889947236,0.1035356969,-0.0350189246,-0.0177480802,-0.2236268073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1765","title":"Error iterating over Dataset with DataLoader","comments":"Yes, padding is an answer. \r\n\r\nThis can be solved easily by passing a callable to the collate_fn arg of DataLoader that adds padding. ","body":"I have a Dataset that I've mapped a tokenizer over:\r\n\r\n```\r\nencoded_dataset.set_format(type='torch',columns=['attention_mask','input_ids','token_type_ids'])\r\nencoded_dataset[:1]\r\n```\r\n```\r\n{'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]),\r\n 'input_ids': tensor([[ 101, 178, 1198, 1400, 1714, 22233, 21365, 4515, 8618, 1113,\r\n 102]]),\r\n 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]])}\r\n```\r\n\r\nWhen I try to iterate as in the docs, I get errors:\r\n\r\n```\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\nnext(iter(dataloader))\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-45-05180ba8aa35> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataloader.py in __init__(self, loader)\r\n 411 self._timeout = loader.timeout\r\n 412 self._collate_fn = loader.collate_fn\r\n--> 413 self._sampler_iter = iter(self._index_sampler)\r\n 414 self._base_seed = torch.empty((), dtype=torch.int64).random_(generator=loader.generator).item()\r\n 415 self._persistent_workers = loader.persistent_workers\r\n\r\nTypeError: 'int' object is not iterable\r\n\r\n\r\n```","comment_length":23,"text":"Error iterating over Dataset with DataLoader\nI have a Dataset that I've mapped a tokenizer over:\r\n\r\n```\r\nencoded_dataset.set_format(type='torch',columns=['attention_mask','input_ids','token_type_ids'])\r\nencoded_dataset[:1]\r\n```\r\n```\r\n{'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]),\r\n 'input_ids': tensor([[ 101, 178, 1198, 1400, 1714, 22233, 21365, 4515, 8618, 1113,\r\n 102]]),\r\n 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]])}\r\n```\r\n\r\nWhen I try to iterate as in the docs, I get errors:\r\n\r\n```\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\nnext(iter(dataloader))\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-45-05180ba8aa35> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataloader.py in __init__(self, loader)\r\n 411 self._timeout = loader.timeout\r\n 412 self._collate_fn = loader.collate_fn\r\n--> 413 self._sampler_iter = iter(self._index_sampler)\r\n 414 self._base_seed = torch.empty((), dtype=torch.int64).random_(generator=loader.generator).item()\r\n 415 self._persistent_workers = loader.persistent_workers\r\n\r\nTypeError: 'int' object is not iterable\r\n\r\n\r\n```\nYes, padding is an answer. \r\n\r\nThis can be solved easily by passing a callable to the collate_fn arg of DataLoader that adds padding. ","embeddings":[-0.2379294932,0.0554564521,-0.0456708856,0.2057333887,0.1531800032,0.00014107,0.7921316028,0.3068522215,0.0211306885,0.1527129412,0.0467958041,0.2304260731,-0.2809546292,-0.2416442186,-0.0902263075,-0.0604503341,-0.0764540955,0.0790703148,-0.1224215627,0.0328939222,-0.1435404718,-0.086729899,-0.1925974041,0.0319360159,-0.2509384155,-0.1322708279,-0.1035913602,-0.0947072953,-0.3341249228,-0.5228338242,0.2897110581,0.1179489493,0.4238479137,0.6812455058,-0.0001170856,0.0532652996,0.1204966977,-0.1216139197,-0.2641813755,-0.232062906,0.1392339319,-0.0516221598,0.0336749442,-0.3547690809,0.2605599761,-0.331272155,-0.017206395,-0.2235834897,0.1987376958,0.2407110631,0.1273105145,0.4878684282,-0.0130664473,0.0686062798,0.04524187,0.2505432963,-0.019511655,0.1971431524,0.1916248351,0.0102274111,-0.3597660363,0.1168679073,-0.1690709442,0.0475237295,0.135968551,-0.1127389744,-0.217760846,-0.1320601255,-0.0326831453,-0.0796288997,0.5732844472,-0.2473982126,-0.3338905275,-0.1655160934,-0.2215303928,-0.2673027515,0.319212079,-0.1650728881,-0.1586042792,-0.2139902264,0.1157573313,0.2405707538,-0.2040264159,0.1367142946,-0.25916031,0.2086893469,-0.0498498268,0.304499954,0.2167702019,-0.2556965649,0.3407104909,0.1005515233,0.0440492071,0.2752284706,-0.5098053217,-0.2515716553,0.0933656469,-0.4813280404,-0.0889511406,0.1817471981,0.2512769699,0.0659232885,-0.0267228652,0.2420649678,0.4608971775,0.1867473274,0.1686100066,0.3430075943,0.0560028069,-0.1008399948,0.1115823984,-0.0628378764,-0.3309622705,-0.304035604,0.2297867239,0.1007610634,0.2355149388,0.0638043508,-0.24075225,-0.0010412553,-0.2854396999,-0.1408834308,0.1115429774,0.1466972083,0.2154821754,0.0784652904,-0.0211962853,0.0494476669,-0.0294592027,-0.1010558978,-0.0631030872,0.1225908026,-0.2360044122,0.0062334132,0.2692131698,-0.0449522883,-0.0005978197,0.1379322708,-0.0163492449,0.1064936593,0.1838307977,-0.1677454114,0.2751210034,-0.066830501,0.0526376925,0.3259077668,0.2427886128,0.0527041815,-0.1517042071,0.0064420728,-0.4252358377,-0.0854079202,-0.0711294115,0.1343848705,-0.0657273754,-0.0596553646,-0.4029016495,-0.1930606961,0.3476107717,-0.1592839658,0.0990208313,-0.2618944049,0.0679650232,-0.0949438438,0.0716987848,0.3785112798,-0.4588417411,-0.1389319897,0.1558684111,0.100578852,0.268353045,0.2800276279,-0.2775180936,0.5750991702,-0.4041410387,0.3146153092,0.3382745087,-0.4575732052,-0.3010725677,0.30232355,-0.1728587449,0.1828440428,0.1311056763,0.0220962148,0.4072544873,-0.2117989957,0.1262145638,0.1312219352,-0.1754143536,-0.084498167,-0.0188895371,0.40735659,0.5204417706,0.0579210706,0.1899082512,0.0911744684,-0.2563124001,0.1866555065,0.0407245047,-0.1624023467,0.0752803087,0.1461818963,0.08729247,0.1154016927,0.2047132254,-0.1541337818,-0.2414141595,0.0174530409,0.08473593,0.1103395596,-0.1569616348,0.1011083499,0.0156297926,0.1833895743,-0.4093043208,-0.1476449072,0.1057991534,0.3391777277,-0.1085056514,-0.2262801528,-0.278257817,0.0101970183,-0.1502237916,0.1229054332,-0.3177072704,0.0756461397,0.0731091648,-0.1621202379,-0.1109917238,0.0012578749,0.2185538709,-0.2305950075,-0.0198753402,0.152743876,0.0733504519,0.0837663114,-0.1409202218,0.0840372369,0.2027758658,-0.0805915147,-0.0350414291,0.0263240449,0.0633541942,-0.2463466823,0.2860279679,0.3557419777,-0.0415121466,0.5438959002,-0.1846143901,0.1865570545,0.0854834467,0.1429632455,-0.2297751456,0.1214375868,0.2915955782,0.165278703,0.2727060616,-0.0742381364,-0.1061879173,-0.0162541084,0.135785684,0.034656506,-0.1672892421,0.0645060986,-0.2023294568,0.078733027,0.1431891769,-0.3583348989,0.2936249375,-0.0419645049,0.1355626434,-0.2297315598,0.1093874052,0.0548776314,-0.0068245805,0.1373674572,0.0501099378,0.1365749091,-0.0179888159,0.032990437,-0.1993452013,-0.2619042993,-0.0321639255,0.4030472338,-0.2017506361,0.420080632,-0.2251237631,-0.0806137398,-0.1357363909,-0.4637432098,-0.1244161502,-0.488583684,-0.2934887409,0.4377196431,-0.0523557365,0.2884217203,0.1593024582,0.0659951046,0.6007106304,-0.1201817542,-0.1299342215,-0.1651609242,-0.1434867382,0.002023285,0.3916653693,-0.3905928731,0.3192386329,-0.14857319,-0.0364399292,-0.4931823909,-0.1713028252,0.23434636,-0.2515105605,-0.1061097458,0.250123322,0.160312131,-0.1276741028,-0.3462354839,0.2360218316,-0.0971406028,-0.0598608814,0.3538429141,-0.0178579725,0.164885059,0.1942387223,-0.1618052274,-0.1143744439,-0.2210423201,0.1488950402,-0.024225384,0.3098029792,0.1279512346,0.0508767813,0.0688194558,0.0538865477,-0.0885351226,-0.0529175512,-0.4608557522,0.4211562574,-0.3188976645,-0.2935510576,-0.0610978752,-0.005195491,0.3583402038,0.374879539,-0.2954478264,-0.0019789448,0.1968028694,0.0960602537,-0.1780664027,0.0133007132,0.1617016345,0.1165711582,-0.0526863299,-0.0528675951,0.1458159238,0.1710808575,-0.1876334697,0.2580990195,0.159302935,0.4765362144,0.2614045143,0.6606869102,0.0389066897,-0.3858367205,0.3321016431,-0.1400644928,-0.0007825629,0.086658068,-0.6534711719,0.1830661446,-0.2897427082,-0.02632441,-0.1086072251,-0.3637517095,-0.2020133138,-0.1655701101,0.1691449881,-0.0271603651,-0.3952352703,0.4657104313,-0.430401504,0.2838457525,-0.0792960823,0.136812672,-0.6179904342,-0.0159705244,0.134028852,0.0014620165,0.2550830841,-0.0820336565,-0.394413799,0.0771882385,-0.4658885002,0.2320282906,0.0853058323,0.7238261104,0.1851679087,-0.2038863599,-0.1605457962,-0.0338941254,0.726137042,0.2351621836,-0.0131854638,0.0982150808,-0.3114118576,-0.4381252229,-0.0794281214,-0.2774267197,0.6424908042,0.1563335955,0.2806633115,-0.2328459471,-0.1301856488,0.1159901023,0.1067452282,-0.2158435285,-0.2748224735,-0.0760231242,-0.0680433884,-0.353223443,-0.014091325,0.0438390039,0.2820525467,0.0923700854,0.0220608246,-0.3040832281,-0.0276663322,0.3623539209,-0.0427861214,0.0660697371,-0.1614816934,0.1583146006,0.1936412007,0.2400769889,0.6411280036,0.3070535362,0.2450909913,-0.4524520338,0.3747632802,0.0567932725,0.2789521813,0.126682505,-0.1120252907,-0.0463845544,-0.3257368505,0.0500912033,-0.4352755249,0.5216366649,0.5002289414,0.1470032632,-0.5732986331,-0.3403778374,0.2020383924,0.1137958467,0.1821293831,0.7328404784,-0.1325454563,-0.1383227408,0.1305194199,0.1368960142,0.7372894287,0.0599588566,0.1241550595,0.4110813141,0.5742589831,0.2874032259,-0.5326961279,0.0999787003,-0.2278511822,-0.0477599874,-0.1629463881,-0.2553313375,0.0387578979,0.1704581827,-0.0596478768,0.1802617759,-0.0350546502,0.5183291435,0.0984523073,0.0610763133,-0.3110624254,-0.2922029495,0.0958925858,0.1132938266,-0.1642979085,0.1464618146,-0.0775095448,-0.0749971867,0.1822136045,-0.2826287746,-0.2507403791,-0.0257558469,-0.2285092771,0.1024926826,0.0329171196,-0.4766200781,0.1010436937,0.0133420769,0.050727874,-0.090359576,0.0723627657,-0.2212266326,-0.0407183506,-0.1306499988,-0.132861495,-0.0846425891,0.4096860588,0.12493743,0.130896911,0.281706512,0.2302460819,-0.3494544327,0.0834648982,0.222866118,0.2381524444,-0.5538777709,0.1067495197,-0.1169768646,-0.200172022,-0.2225140333,0.0751270652,0.0035707487,-0.3264312446,0.00617181,-0.1581954062,-0.0941699371,0.0912848115,0.4193665385,0.0990061983,-0.4401395023,0.7563033104,0.1833338439,-0.2378757447,-0.1601602882,0.1304143965,0.4554119706,-0.2698962986,0.1051627994,-0.3681175113,0.1937119961,-0.2056511492,0.0954531431,0.0749833733,0.0011241743,-0.0847064331,-0.3185295165,-0.376385808,0.0641730577,-0.127597034,0.2709617615,0.0657135397,0.1442403048,-0.2424863875,0.1084674597,-0.2284084707,-0.0599610582,-0.1304063201,0.2170205265,0.0687656924,-0.1155499667,-0.029965708,0.1643112749,0.0875988752,0.3649315536,0.0772954896,-0.2171725333,-0.1124109924,0.1081866995,0.1269553751,-0.2696897686,0.0701325387,-0.0216650311,-0.0106119867,-0.3168860674,0.0920091048,0.0593076125,-0.0337995403,0.2827904522,0.3835873008,0.0453121625,0.1048733965,0.0202995613,0.244404465,0.264428556,0.1860617399,0.3221852183,-0.034697488,-0.0614788644,-0.2507389784,0.3639924824,0.3781985939,-0.1171507537,0.1562584788,-0.4842172861,0.1358973086,-0.1309468895,0.2535530627,0.6113288403,-0.3544901311,-0.1553913802,0.1691128314,0.1126620173,-0.0045485063,0.0386662707,0.0609535091,-0.0578211956,-0.0058272169,0.1932299137,-0.0571368709,-0.0611666217,0.1354420781,-0.0829459876,0.2889599502,-0.1634664685,0.1883116364,0.0641768351,-0.4278978705,-0.0858652741,0.3677484095,0.2824953794,0.1013357863,0.2784427404,-0.1078804582,0.427321136,0.0708391145,-0.2493039072,-0.0165874641,-0.4929739535,-0.3299542665,0.1111973599,0.2463370413,0.1948013008,0.0489695594,0.320019722,-0.0799887404,-0.1117729396,-0.0213638823,0.1956691891,-0.277502358,0.0963109508,-0.3862141073,-0.2213355303,-0.2077884525,-0.1047486588,-0.077270098,-0.1806131005,0.0019157402,0.118787311,-0.1874989718,-0.4298039377,0.0627504662,-0.0045477967,-0.112351805,-0.3342280388,0.2179592699,0.1602630466,0.1170849726,0.0349093527,0.2651892602,0.5087574124,0.4610842764,-0.1494466215,0.1769150198,-0.5010870695,-0.1911614835,-0.0631125867,0.1125183627,-0.4148905277,-0.1014754996,0.3867393434,0.0840830281,-0.1143442765,-0.2153678983,0.1340745538,0.0735940412,0.0001483769,0.1247917414,-0.0815566033,0.0414281264,-0.0814748928,-0.0441691838,-0.4062025249,-0.1182344258,0.4678344429,0.1166635156,0.1968592107,0.0207538828,0.0589621589,-0.2503395975,0.4257987738,0.4754786193,0.2222500294,-0.409289211,-0.2357503623,-0.2958708107,0.3727261722,-0.1213889048,-0.1949660182,-0.0196231641,0.1651484668,0.3228288591,0.2691094577,0.243530497,-0.145322457,-0.0710379779,0.3360802829,-0.3235668242,-0.3320512772,0.035634283,0.2394293398,-0.0671317875,-0.2778089046,0.2334064543,-0.1008600667,0.1332826614,-0.2046568245,-0.3008014262,0.0800730586,0.0444457643,0.4428635836,0.1897007227,0.4272838831,0.0945996493,0.0628773049,-0.278421998,-0.2361575365,-0.1914744675,0.0452262163,0.0448537469,0.2142349631,-0.0324527472,-0.1344611943,-0.5011916161,0.2168956101,-0.1091530472,-0.2911404371,-0.3574871719,0.2795459926,0.0037162364,0.070227012,-0.0714731738,0.2456234843,0.2120578587,0.537260592,-0.3672761619,-0.2944155633,0.4216322899,-0.6915383935,-0.4907265902,0.0250352118,-0.1989057362,0.1786561608,0.27032426,-0.5640615821,0.0031023284,0.317204386,-0.1694213003,-0.1327279061,0.0034326748,0.0969765931,0.0240144264,-0.1689353287,0.3889947236,0.1035356969,-0.0350189246,-0.0177480802,-0.2236268073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1765","title":"Error iterating over Dataset with DataLoader","comments":"Padding was the fix, thanks!","body":"I have a Dataset that I've mapped a tokenizer over:\r\n\r\n```\r\nencoded_dataset.set_format(type='torch',columns=['attention_mask','input_ids','token_type_ids'])\r\nencoded_dataset[:1]\r\n```\r\n```\r\n{'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]),\r\n 'input_ids': tensor([[ 101, 178, 1198, 1400, 1714, 22233, 21365, 4515, 8618, 1113,\r\n 102]]),\r\n 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]])}\r\n```\r\n\r\nWhen I try to iterate as in the docs, I get errors:\r\n\r\n```\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\nnext(iter(dataloader))\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-45-05180ba8aa35> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataloader.py in __init__(self, loader)\r\n 411 self._timeout = loader.timeout\r\n 412 self._collate_fn = loader.collate_fn\r\n--> 413 self._sampler_iter = iter(self._index_sampler)\r\n 414 self._base_seed = torch.empty((), dtype=torch.int64).random_(generator=loader.generator).item()\r\n 415 self._persistent_workers = loader.persistent_workers\r\n\r\nTypeError: 'int' object is not iterable\r\n\r\n\r\n```","comment_length":5,"text":"Error iterating over Dataset with DataLoader\nI have a Dataset that I've mapped a tokenizer over:\r\n\r\n```\r\nencoded_dataset.set_format(type='torch',columns=['attention_mask','input_ids','token_type_ids'])\r\nencoded_dataset[:1]\r\n```\r\n```\r\n{'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]),\r\n 'input_ids': tensor([[ 101, 178, 1198, 1400, 1714, 22233, 21365, 4515, 8618, 1113,\r\n 102]]),\r\n 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]])}\r\n```\r\n\r\nWhen I try to iterate as in the docs, I get errors:\r\n\r\n```\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\nnext(iter(dataloader))\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-45-05180ba8aa35> in <module>()\r\n 1 dataloader = torch.utils.data.DataLoader(encoded_dataset, batch_sampler=32)\r\n----> 2 next(iter(dataloader))\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataloader.py in __init__(self, loader)\r\n 411 self._timeout = loader.timeout\r\n 412 self._collate_fn = loader.collate_fn\r\n--> 413 self._sampler_iter = iter(self._index_sampler)\r\n 414 self._base_seed = torch.empty((), dtype=torch.int64).random_(generator=loader.generator).item()\r\n 415 self._persistent_workers = loader.persistent_workers\r\n\r\nTypeError: 'int' object is not iterable\r\n\r\n\r\n```\nPadding was the fix, thanks!","embeddings":[-0.2379294932,0.0554564521,-0.0456708856,0.2057333887,0.1531800032,0.00014107,0.7921316028,0.3068522215,0.0211306885,0.1527129412,0.0467958041,0.2304260731,-0.2809546292,-0.2416442186,-0.0902263075,-0.0604503341,-0.0764540955,0.0790703148,-0.1224215627,0.0328939222,-0.1435404718,-0.086729899,-0.1925974041,0.0319360159,-0.2509384155,-0.1322708279,-0.1035913602,-0.0947072953,-0.3341249228,-0.5228338242,0.2897110581,0.1179489493,0.4238479137,0.6812455058,-0.0001170856,0.0532652996,0.1204966977,-0.1216139197,-0.2641813755,-0.232062906,0.1392339319,-0.0516221598,0.0336749442,-0.3547690809,0.2605599761,-0.331272155,-0.017206395,-0.2235834897,0.1987376958,0.2407110631,0.1273105145,0.4878684282,-0.0130664473,0.0686062798,0.04524187,0.2505432963,-0.019511655,0.1971431524,0.1916248351,0.0102274111,-0.3597660363,0.1168679073,-0.1690709442,0.0475237295,0.135968551,-0.1127389744,-0.217760846,-0.1320601255,-0.0326831453,-0.0796288997,0.5732844472,-0.2473982126,-0.3338905275,-0.1655160934,-0.2215303928,-0.2673027515,0.319212079,-0.1650728881,-0.1586042792,-0.2139902264,0.1157573313,0.2405707538,-0.2040264159,0.1367142946,-0.25916031,0.2086893469,-0.0498498268,0.304499954,0.2167702019,-0.2556965649,0.3407104909,0.1005515233,0.0440492071,0.2752284706,-0.5098053217,-0.2515716553,0.0933656469,-0.4813280404,-0.0889511406,0.1817471981,0.2512769699,0.0659232885,-0.0267228652,0.2420649678,0.4608971775,0.1867473274,0.1686100066,0.3430075943,0.0560028069,-0.1008399948,0.1115823984,-0.0628378764,-0.3309622705,-0.304035604,0.2297867239,0.1007610634,0.2355149388,0.0638043508,-0.24075225,-0.0010412553,-0.2854396999,-0.1408834308,0.1115429774,0.1466972083,0.2154821754,0.0784652904,-0.0211962853,0.0494476669,-0.0294592027,-0.1010558978,-0.0631030872,0.1225908026,-0.2360044122,0.0062334132,0.2692131698,-0.0449522883,-0.0005978197,0.1379322708,-0.0163492449,0.1064936593,0.1838307977,-0.1677454114,0.2751210034,-0.066830501,0.0526376925,0.3259077668,0.2427886128,0.0527041815,-0.1517042071,0.0064420728,-0.4252358377,-0.0854079202,-0.0711294115,0.1343848705,-0.0657273754,-0.0596553646,-0.4029016495,-0.1930606961,0.3476107717,-0.1592839658,0.0990208313,-0.2618944049,0.0679650232,-0.0949438438,0.0716987848,0.3785112798,-0.4588417411,-0.1389319897,0.1558684111,0.100578852,0.268353045,0.2800276279,-0.2775180936,0.5750991702,-0.4041410387,0.3146153092,0.3382745087,-0.4575732052,-0.3010725677,0.30232355,-0.1728587449,0.1828440428,0.1311056763,0.0220962148,0.4072544873,-0.2117989957,0.1262145638,0.1312219352,-0.1754143536,-0.084498167,-0.0188895371,0.40735659,0.5204417706,0.0579210706,0.1899082512,0.0911744684,-0.2563124001,0.1866555065,0.0407245047,-0.1624023467,0.0752803087,0.1461818963,0.08729247,0.1154016927,0.2047132254,-0.1541337818,-0.2414141595,0.0174530409,0.08473593,0.1103395596,-0.1569616348,0.1011083499,0.0156297926,0.1833895743,-0.4093043208,-0.1476449072,0.1057991534,0.3391777277,-0.1085056514,-0.2262801528,-0.278257817,0.0101970183,-0.1502237916,0.1229054332,-0.3177072704,0.0756461397,0.0731091648,-0.1621202379,-0.1109917238,0.0012578749,0.2185538709,-0.2305950075,-0.0198753402,0.152743876,0.0733504519,0.0837663114,-0.1409202218,0.0840372369,0.2027758658,-0.0805915147,-0.0350414291,0.0263240449,0.0633541942,-0.2463466823,0.2860279679,0.3557419777,-0.0415121466,0.5438959002,-0.1846143901,0.1865570545,0.0854834467,0.1429632455,-0.2297751456,0.1214375868,0.2915955782,0.165278703,0.2727060616,-0.0742381364,-0.1061879173,-0.0162541084,0.135785684,0.034656506,-0.1672892421,0.0645060986,-0.2023294568,0.078733027,0.1431891769,-0.3583348989,0.2936249375,-0.0419645049,0.1355626434,-0.2297315598,0.1093874052,0.0548776314,-0.0068245805,0.1373674572,0.0501099378,0.1365749091,-0.0179888159,0.032990437,-0.1993452013,-0.2619042993,-0.0321639255,0.4030472338,-0.2017506361,0.420080632,-0.2251237631,-0.0806137398,-0.1357363909,-0.4637432098,-0.1244161502,-0.488583684,-0.2934887409,0.4377196431,-0.0523557365,0.2884217203,0.1593024582,0.0659951046,0.6007106304,-0.1201817542,-0.1299342215,-0.1651609242,-0.1434867382,0.002023285,0.3916653693,-0.3905928731,0.3192386329,-0.14857319,-0.0364399292,-0.4931823909,-0.1713028252,0.23434636,-0.2515105605,-0.1061097458,0.250123322,0.160312131,-0.1276741028,-0.3462354839,0.2360218316,-0.0971406028,-0.0598608814,0.3538429141,-0.0178579725,0.164885059,0.1942387223,-0.1618052274,-0.1143744439,-0.2210423201,0.1488950402,-0.024225384,0.3098029792,0.1279512346,0.0508767813,0.0688194558,0.0538865477,-0.0885351226,-0.0529175512,-0.4608557522,0.4211562574,-0.3188976645,-0.2935510576,-0.0610978752,-0.005195491,0.3583402038,0.374879539,-0.2954478264,-0.0019789448,0.1968028694,0.0960602537,-0.1780664027,0.0133007132,0.1617016345,0.1165711582,-0.0526863299,-0.0528675951,0.1458159238,0.1710808575,-0.1876334697,0.2580990195,0.159302935,0.4765362144,0.2614045143,0.6606869102,0.0389066897,-0.3858367205,0.3321016431,-0.1400644928,-0.0007825629,0.086658068,-0.6534711719,0.1830661446,-0.2897427082,-0.02632441,-0.1086072251,-0.3637517095,-0.2020133138,-0.1655701101,0.1691449881,-0.0271603651,-0.3952352703,0.4657104313,-0.430401504,0.2838457525,-0.0792960823,0.136812672,-0.6179904342,-0.0159705244,0.134028852,0.0014620165,0.2550830841,-0.0820336565,-0.394413799,0.0771882385,-0.4658885002,0.2320282906,0.0853058323,0.7238261104,0.1851679087,-0.2038863599,-0.1605457962,-0.0338941254,0.726137042,0.2351621836,-0.0131854638,0.0982150808,-0.3114118576,-0.4381252229,-0.0794281214,-0.2774267197,0.6424908042,0.1563335955,0.2806633115,-0.2328459471,-0.1301856488,0.1159901023,0.1067452282,-0.2158435285,-0.2748224735,-0.0760231242,-0.0680433884,-0.353223443,-0.014091325,0.0438390039,0.2820525467,0.0923700854,0.0220608246,-0.3040832281,-0.0276663322,0.3623539209,-0.0427861214,0.0660697371,-0.1614816934,0.1583146006,0.1936412007,0.2400769889,0.6411280036,0.3070535362,0.2450909913,-0.4524520338,0.3747632802,0.0567932725,0.2789521813,0.126682505,-0.1120252907,-0.0463845544,-0.3257368505,0.0500912033,-0.4352755249,0.5216366649,0.5002289414,0.1470032632,-0.5732986331,-0.3403778374,0.2020383924,0.1137958467,0.1821293831,0.7328404784,-0.1325454563,-0.1383227408,0.1305194199,0.1368960142,0.7372894287,0.0599588566,0.1241550595,0.4110813141,0.5742589831,0.2874032259,-0.5326961279,0.0999787003,-0.2278511822,-0.0477599874,-0.1629463881,-0.2553313375,0.0387578979,0.1704581827,-0.0596478768,0.1802617759,-0.0350546502,0.5183291435,0.0984523073,0.0610763133,-0.3110624254,-0.2922029495,0.0958925858,0.1132938266,-0.1642979085,0.1464618146,-0.0775095448,-0.0749971867,0.1822136045,-0.2826287746,-0.2507403791,-0.0257558469,-0.2285092771,0.1024926826,0.0329171196,-0.4766200781,0.1010436937,0.0133420769,0.050727874,-0.090359576,0.0723627657,-0.2212266326,-0.0407183506,-0.1306499988,-0.132861495,-0.0846425891,0.4096860588,0.12493743,0.130896911,0.281706512,0.2302460819,-0.3494544327,0.0834648982,0.222866118,0.2381524444,-0.5538777709,0.1067495197,-0.1169768646,-0.200172022,-0.2225140333,0.0751270652,0.0035707487,-0.3264312446,0.00617181,-0.1581954062,-0.0941699371,0.0912848115,0.4193665385,0.0990061983,-0.4401395023,0.7563033104,0.1833338439,-0.2378757447,-0.1601602882,0.1304143965,0.4554119706,-0.2698962986,0.1051627994,-0.3681175113,0.1937119961,-0.2056511492,0.0954531431,0.0749833733,0.0011241743,-0.0847064331,-0.3185295165,-0.376385808,0.0641730577,-0.127597034,0.2709617615,0.0657135397,0.1442403048,-0.2424863875,0.1084674597,-0.2284084707,-0.0599610582,-0.1304063201,0.2170205265,0.0687656924,-0.1155499667,-0.029965708,0.1643112749,0.0875988752,0.3649315536,0.0772954896,-0.2171725333,-0.1124109924,0.1081866995,0.1269553751,-0.2696897686,0.0701325387,-0.0216650311,-0.0106119867,-0.3168860674,0.0920091048,0.0593076125,-0.0337995403,0.2827904522,0.3835873008,0.0453121625,0.1048733965,0.0202995613,0.244404465,0.264428556,0.1860617399,0.3221852183,-0.034697488,-0.0614788644,-0.2507389784,0.3639924824,0.3781985939,-0.1171507537,0.1562584788,-0.4842172861,0.1358973086,-0.1309468895,0.2535530627,0.6113288403,-0.3544901311,-0.1553913802,0.1691128314,0.1126620173,-0.0045485063,0.0386662707,0.0609535091,-0.0578211956,-0.0058272169,0.1932299137,-0.0571368709,-0.0611666217,0.1354420781,-0.0829459876,0.2889599502,-0.1634664685,0.1883116364,0.0641768351,-0.4278978705,-0.0858652741,0.3677484095,0.2824953794,0.1013357863,0.2784427404,-0.1078804582,0.427321136,0.0708391145,-0.2493039072,-0.0165874641,-0.4929739535,-0.3299542665,0.1111973599,0.2463370413,0.1948013008,0.0489695594,0.320019722,-0.0799887404,-0.1117729396,-0.0213638823,0.1956691891,-0.277502358,0.0963109508,-0.3862141073,-0.2213355303,-0.2077884525,-0.1047486588,-0.077270098,-0.1806131005,0.0019157402,0.118787311,-0.1874989718,-0.4298039377,0.0627504662,-0.0045477967,-0.112351805,-0.3342280388,0.2179592699,0.1602630466,0.1170849726,0.0349093527,0.2651892602,0.5087574124,0.4610842764,-0.1494466215,0.1769150198,-0.5010870695,-0.1911614835,-0.0631125867,0.1125183627,-0.4148905277,-0.1014754996,0.3867393434,0.0840830281,-0.1143442765,-0.2153678983,0.1340745538,0.0735940412,0.0001483769,0.1247917414,-0.0815566033,0.0414281264,-0.0814748928,-0.0441691838,-0.4062025249,-0.1182344258,0.4678344429,0.1166635156,0.1968592107,0.0207538828,0.0589621589,-0.2503395975,0.4257987738,0.4754786193,0.2222500294,-0.409289211,-0.2357503623,-0.2958708107,0.3727261722,-0.1213889048,-0.1949660182,-0.0196231641,0.1651484668,0.3228288591,0.2691094577,0.243530497,-0.145322457,-0.0710379779,0.3360802829,-0.3235668242,-0.3320512772,0.035634283,0.2394293398,-0.0671317875,-0.2778089046,0.2334064543,-0.1008600667,0.1332826614,-0.2046568245,-0.3008014262,0.0800730586,0.0444457643,0.4428635836,0.1897007227,0.4272838831,0.0945996493,0.0628773049,-0.278421998,-0.2361575365,-0.1914744675,0.0452262163,0.0448537469,0.2142349631,-0.0324527472,-0.1344611943,-0.5011916161,0.2168956101,-0.1091530472,-0.2911404371,-0.3574871719,0.2795459926,0.0037162364,0.070227012,-0.0714731738,0.2456234843,0.2120578587,0.537260592,-0.3672761619,-0.2944155633,0.4216322899,-0.6915383935,-0.4907265902,0.0250352118,-0.1989057362,0.1786561608,0.27032426,-0.5640615821,0.0031023284,0.317204386,-0.1694213003,-0.1327279061,0.0034326748,0.0969765931,0.0240144264,-0.1689353287,0.3889947236,0.1035356969,-0.0350189246,-0.0177480802,-0.2236268073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1764","title":"Connection Issues","comments":"Academic WIFI was blocking.","body":"Today, I am getting connection issues while loading a dataset and the metric.\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/train.py\", line 180, in <module>\r\n train_dataset, dev_dataset, test_dataset = create_race_dataset()\r\n File \"src\/train.py\", line 130, in create_race_dataset\r\n train_dataset = load_dataset(\"race\", \"all\", split=\"train\")\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 591, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/race\/race.py\r\n```\r\n\r\nOr\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/train.py\", line 105, in <module>\r\n rouge = datasets.load_metric(\"rouge\")\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 500, in load_metric\r\n dataset=False,\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/metrics\/rouge\/rouge.py\r\n```","comment_length":4,"text":"Connection Issues\nToday, I am getting connection issues while loading a dataset and the metric.\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/train.py\", line 180, in <module>\r\n train_dataset, dev_dataset, test_dataset = create_race_dataset()\r\n File \"src\/train.py\", line 130, in create_race_dataset\r\n train_dataset = load_dataset(\"race\", \"all\", split=\"train\")\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 591, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/datasets\/race\/race.py\r\n```\r\n\r\nOr\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/train.py\", line 105, in <module>\r\n rouge = datasets.load_metric(\"rouge\")\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 500, in load_metric\r\n dataset=False,\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n max_retries=download_config.max_retries,\r\n File \"\/Users\/saeed\/Desktop\/codes\/repos\/dreamscape-qa\/env\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 617, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.1\/metrics\/rouge\/rouge.py\r\n```\nAcademic WIFI was blocking.","embeddings":[-0.5180004239,0.1518898457,-0.1293317974,0.4499788284,0.3969016969,-0.1457725763,0.1320408732,0.1762890667,-0.0140063707,0.1082949191,-0.124875173,-0.1481359303,0.2510062456,-0.0198108274,0.0466071069,-0.1421063393,-0.2230492383,-0.1000881568,-0.1379762888,0.1928783953,-0.2462978661,-0.0685145631,-0.0341915675,0.1003587842,-0.3674422801,-0.1130034477,0.2118816078,0.1025795341,-0.1906882524,-0.5526769161,0.2381566316,0.3073061109,0.0368950665,0.4877927005,-0.0001093358,0.1064935774,0.4667753279,0.129618749,-0.4098181725,-0.3776007593,-0.4410849512,-0.2135694176,0.2583916783,-0.0508425497,-0.1517905593,-0.0361567736,-0.0947532207,-0.3747791648,0.3922136724,0.3934460282,0.2814341187,0.2624728084,0.065754123,-0.1848604083,0.0056370199,-0.1325550824,0.0420695096,0.5977998972,0.0363061987,-0.1302544624,-0.141337201,0.0864764899,0.0450211279,0.0783322975,0.3826260865,-0.0640693456,0.2347893417,-0.1510821879,0.1824484766,0.2609403729,0.4098980129,-0.1440190226,-0.4422341585,-0.0506780334,-0.1268638521,-0.6068960428,0.3543641567,-0.0385129526,-0.2514629364,-0.0125298472,-0.4555767775,-0.1677048653,-0.3311396241,0.2458788753,-0.2871293128,0.0626944751,-0.0343677849,0.0893413574,0.0814110413,-0.1123764366,0.2095321119,-0.152984336,-0.0734671429,0.0678390339,-0.3960903585,0.2580829561,-0.0856753066,-0.0510927178,0.0402808115,0.1119916141,0.067824997,0.1628222764,-0.0665859431,0.1818760186,0.2862825096,0.3040936589,0.0728456527,0.053481888,0.4194645584,0.3111786842,-0.0673014671,-0.1241603941,-0.0464264639,-0.3569214344,0.0936919823,0.2271097004,0.058944948,-0.1966579556,-0.3724307418,0.1134745255,0.1127974093,-0.1751014441,0.2640088797,0.4180767834,-0.3457605541,0.0988804325,0.0030513967,0.23961775,-0.2246599346,0.0092489691,-0.3212332129,0.0002274964,-0.2881858647,0.1446988136,0.2070100605,0.1083842814,0.1364062577,-0.0763500035,0.1232288554,-0.1994099766,0.1277836412,-0.3439017832,-0.2384491563,0.2847863138,0.0215479527,-0.0208668895,0.058784768,-0.1627436876,0.0912695453,-0.1208806485,-0.3458520174,-0.1684644818,0.1809292436,0.2656354308,-0.2325540185,-0.0200075395,-0.2725390792,0.0030653828,0.0226195194,-0.2275676727,-0.2533537149,-0.1253038347,-0.1990160197,-0.2155647278,0.5127736926,0.2836515009,-0.0721952543,-0.2389689535,0.0373438932,-0.234428212,0.0385681987,0.1375327408,-0.2000637054,0.1705142111,-0.0953032002,-0.1522197127,0.6915007234,-0.5640895367,-0.4621166289,0.1385969669,-0.0843587965,0.0335559584,-0.0349465199,0.0333490483,0.279222995,-0.0332770161,0.5665246248,0.3344761729,0.0067832801,-0.2098933905,-0.0717394277,-0.1832251847,0.0452091321,0.2948998809,0.2426053733,0.0199132506,0.1093176752,-0.200646311,0.0794821158,0.2632910609,0.065764375,0.2559763491,-0.0651138425,-0.0256962478,0.0096352305,-0.0317879394,-0.1829829216,0.2434766591,-0.1602118462,-0.0526641347,-0.1938071251,0.0897260159,-0.3435393274,-0.0050576413,-0.1328041703,0.2188961953,0.1965782046,-0.0028584837,-0.0889944881,0.2512772083,-0.0429172181,0.3508363962,-0.0777853131,0.0879404396,-0.018431345,0.3717056513,0.0571826734,-0.1394953132,0.195872739,0.2675303221,0.2873638272,-0.0805832148,-0.1392545402,0.3354116976,-0.0118376445,0.3655145466,0.2171221524,0.2878853977,0.3624401987,-0.0323363878,0.068549633,-0.1366793811,0.1014979556,-0.1343502104,0.132679984,0.2943197489,0.0459469594,0.2977102697,0.2369897813,0.0020798559,0.2968909144,0.0162052345,0.0289728902,0.065606378,0.2752691805,-0.1127328426,0.1391452551,-0.1350894272,0.0917014927,0.0747376308,0.245801121,-0.084207207,0.0636247098,0.1820786297,-0.1050997078,-0.2456618398,-0.1677815616,0.1384423077,0.4333441854,0.1354047507,0.1188822985,0.2923322022,-0.0504026897,-0.2828835249,0.3180322051,0.1809640527,-0.1714705825,0.3529526591,0.0919610262,-0.1024634987,-0.1411800534,-0.256169647,-0.0036928128,0.2967860699,-0.3918045163,0.1140519604,-0.2953872681,-0.2041802108,0.1480870992,-0.1455181986,-0.303791672,-0.5441190004,-0.0656884089,0.2260402143,0.0353956074,0.1959815025,-0.0241571125,0.2882223129,0.0196514856,0.1021520421,-0.122720167,-0.5162810683,-0.2213449478,0.0693113357,0.1434305906,-0.1634460539,0.3588271737,-0.2839470804,0.0697342157,-0.0082697412,-0.0323561728,0.0227198116,0.209939152,0.5121235847,0.0811059698,0.2875458598,0.0987255201,0.0801546499,0.5440696478,0.0003572651,0.0770383403,0.0656586736,-0.0336723104,-0.0217504464,0.130291149,-0.2625752091,-0.4195988476,-0.3231443763,-0.045309186,-0.0251582451,0.087330319,0.2898463905,-0.0126671391,0.168619141,-0.0685789436,0.1992180943,-0.3329153955,-0.5175252557,0.0786200985,-0.138180092,-0.325517118,0.0788639262,0.1935058385,0.1995464712,0.2311193347,-0.5495778322,-0.46843943,-0.0112880385,0.2621096671,0.0702006891,-0.0666346923,0.011004963,-0.1054924056,-0.1849503815,-0.0116431834,-0.1358875483,0.0574334785,-0.0450732447,0.3892758787,-0.0469928309,0.4414332509,-0.2201838791,0.6145742536,0.1591567844,-0.0850710869,0.2463884056,-0.1970653683,0.417937696,-0.0510557108,-0.3118383884,0.0699205846,-0.0783850998,0.0103273457,-0.0137200924,0.1247215122,0.1249918863,-0.5026952624,-0.1441327482,-0.2501017153,-0.3724626601,-0.1490376592,-0.1566585898,0.2366256714,0.1054019555,0.1368264407,-0.2089195997,-0.1385321468,-0.0593031868,0.5529293418,0.0150394449,-0.1041865572,-0.3137789667,-0.0078846151,-0.3424931467,0.3515614867,0.018089829,0.4888181686,-0.5512746572,0.1804420352,0.0588590056,-0.022565946,0.6089797616,-0.383361578,0.1292365342,0.1308057159,0.100957036,-0.2785435617,0.1102558002,0.0965493768,0.2002999038,0.1314139962,0.5278847218,-0.1751534641,-0.233010754,-0.0173436627,0.1403595805,-0.3152048588,-0.1836376041,-0.265227437,-0.2938791811,0.0327158049,-0.0905557424,-0.2927690744,0.2864076197,0.0369237177,-0.1782617718,0.0081807002,-0.0971032381,0.2378230095,0.0132800415,0.136157915,-0.2250260264,0.2668131292,0.4904593825,0.1157298312,0.2346478552,0.2966921628,-0.0060179983,-0.2871244252,0.1238165572,0.1528791636,0.1624113917,0.4545560479,-0.1278036684,0.1800590158,0.1465393305,0.1617208272,-0.2975646257,0.2920679748,0.2351577133,0.0805737302,-0.4495462179,-0.4700047672,0.2869491577,-0.0073681711,-0.1976431459,0.214736864,-0.3197669983,-0.17394045,-0.1591699868,-0.1480980963,1.0096501112,0.1101532057,-0.001333268,0.2283272594,-0.3349340558,0.1861586273,-0.3709096611,0.1963088959,-0.2709699273,-0.1568680406,-0.0951144993,-0.1690188646,0.1716848165,-0.3043259084,0.0161116254,0.4394074082,-0.3312294483,0.131805405,-0.0577941462,0.0837191418,-0.1117463857,0.0298463069,-0.2551341355,0.2423286438,-0.2050527483,0.096756652,-0.1573448926,-0.138948679,-0.0931161195,-0.1594887823,-0.1129400879,-0.0035910795,-0.339664042,0.2274913192,-0.0622317009,-0.2498008758,-0.2273176014,0.0083596753,0.1508934051,0.0225397721,-0.1874333322,0.3485206664,-0.4078956842,0.1641660035,0.1739226133,0.1460708678,-0.0257057175,-0.2236062735,-0.3374369442,0.1803673506,-0.0557486862,-0.0708359256,0.0580772571,-0.1462438852,0.0796597525,-0.1679439098,0.1262485236,-0.3142540455,0.112913318,-0.0794019401,0.200857386,0.165986076,0.0443957075,0.1982750595,0.195293799,-0.2185045332,-0.0695661083,0.5943551064,-0.0098748608,-0.2306684554,0.3861602545,0.0034320049,-0.2417515963,-0.2354781032,0.0770583004,0.0864983425,-0.3671213686,0.0052016657,0.0830532685,0.1468408704,-0.090529874,0.2467571497,0.2510602474,-0.1001930162,0.1711987406,-0.6785686016,-0.1609284282,0.1726964414,-0.1347057223,0.2189988196,-0.1965313256,-0.0490297973,-0.2157632411,0.0791828558,-0.366022259,-0.1438466907,-0.366878897,0.0080018137,0.1696390063,-0.3241266906,0.3054667115,-0.1216413006,0.2169777304,-0.3472321033,-0.4371874928,-0.2340248227,-0.0546389483,0.1067061052,-0.1566357315,-0.4206061065,-0.1477061361,-0.249249056,-0.0003488412,0.0059260023,0.0265645031,0.0408577099,-0.0810761824,-0.2151926011,-0.1263347119,0.2753791511,0.0055557149,0.2244072407,0.1932219118,0.4230403602,-0.1433608681,-0.0314203426,-0.1545742005,-0.1150324196,-0.1540711969,-0.0465239994,0.1225582138,-0.1530347914,0.4913654923,-0.1824718118,0.0609684736,0.2395864129,0.4115927815,0.4023323059,-0.1757397503,0.0352920555,-0.1276525706,0.2720545828,-0.0263614077,-0.1966059953,0.1881255656,0.1308602393,-0.0286859013,-0.0831238329,0.1664058268,0.0732532814,-0.0741375387,-0.0046867854,0.6448647976,-0.1004172191,0.1160551831,0.0793245509,0.1257136315,0.0466591679,0.2852467,0.2716386914,-0.0523384102,0.2038525194,-0.2319123894,0.1513451487,0.1917323321,0.1966912895,-0.3012568653,-0.4229506552,0.0707033351,0.0439982787,-0.2500997782,-0.1726221293,-0.1768520921,0.3001233637,-0.3496533632,0.0208788849,-0.2471484095,0.0178893935,-0.1155513898,-0.2320354283,-0.1394762397,-0.4418420196,-0.1096185595,-0.2276967168,0.0848134384,0.0933011696,-0.0008790619,0.0800682232,0.0756236538,-0.2542055547,-0.0091140689,0.2536837757,0.071937196,-0.0537668429,0.3708094358,0.2544518709,0.2320619076,0.1515146792,0.408672303,0.5366947651,0.3852316439,-0.1667702496,0.0145119177,0.1743244082,-0.0680128038,-0.0188565236,0.2316375971,0.39757967,-0.1177132577,0.2967394888,0.2895070314,-0.1911889464,0.5396581888,-0.1190728545,0.2044420987,-0.1884616762,0.5268084407,-0.1599613577,-0.0473599806,-0.2523485422,-0.2365204692,-0.3550065458,-0.0325435884,0.428232044,-0.0210278276,0.1585393548,-0.2133675665,0.0857639536,-0.244665727,0.6634171605,0.2506338954,0.2151750624,-0.3808695376,0.0831404477,-0.4975030422,0.061674919,0.0648095682,0.022216361,0.0716279075,0.0240385626,-0.2558310032,0.2419308275,0.2305718213,0.0606478602,0.2361894846,-0.1064861491,-0.381438911,-0.0854349062,0.1560914069,-0.0470716208,-0.13941589,-0.2070657909,0.2237917036,-0.296984762,0.0945704132,0.0424843393,-0.1760328561,0.071815297,0.077920571,0.3377152979,0.1458941996,0.7701454163,0.1694014072,0.1123799086,-0.1923610568,-0.6222954988,0.1627147645,0.0653213784,-0.0065368605,0.1548682451,0.0150915235,-0.0413514785,-0.2868538201,0.4293663204,-0.1574685574,-0.1150914058,0.003608891,-0.1416710019,0.1350250095,0.00124935,0.1174905449,-0.0005213026,0.0664803833,0.3123764098,-0.0750659034,-0.2616921365,0.5130414367,-0.185765177,-0.0862645954,-0.12546359,0.4602627754,0.2010318488,0.0668696389,-0.4015524089,0.0826882049,0.5066984892,-0.143796578,-0.3005382717,0.3413335979,-0.2313749641,0.0919228643,0.009105999,-0.0085856998,-0.1377776563,-0.2656106949,-0.0723011047,0.086391069]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1762","title":"Unable to format dataset to CUDA Tensors","comments":"Hi ! You can get CUDA tensors with\r\n\r\n```python\r\ndataset.set_format(\"torch\", columns=columns, device=\"cuda\")\r\n```\r\n\r\nIndeed `set_format` passes the `**kwargs` to `torch.tensor`","body":"Hi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan","comment_length":20,"text":"Unable to format dataset to CUDA Tensors\nHi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan\nHi ! You can get CUDA tensors with\r\n\r\n```python\r\ndataset.set_format(\"torch\", columns=columns, device=\"cuda\")\r\n```\r\n\r\nIndeed `set_format` passes the `**kwargs` to `torch.tensor`","embeddings":[-0.2256218046,-0.4427936971,-0.0796787292,0.1839720011,0.5433498621,0.3885726929,0.5336815119,0.3345811665,-0.0096712606,0.1427577734,-0.0833449513,0.2423237115,-0.2235853821,0.0973638371,0.1539933234,-0.3738492131,0.2507269979,-0.1545419246,-0.0661568865,0.0754857361,-0.1834034473,0.0132907908,-0.1419782192,-0.2985790372,-0.2301402688,-0.1756631881,-0.0042094737,-0.361751169,-0.1628382355,-0.0704989508,0.2917990685,0.0672327131,0.2707696855,0.4780969024,-0.0001229248,-0.0246179085,-0.0642263368,-0.0974277705,-0.122957699,-0.3244974613,0.0951584503,-0.3088740706,0.3264772296,-0.2337548733,-0.4170435369,0.1638752818,0.0245886184,-0.3956411183,0.231423676,0.5253636837,0.119892478,0.1924467087,0.2394107431,0.0240939595,-0.1740587056,0.376606375,-0.3542357385,0.0254851878,0.1695797294,0.1583283544,0.2881679833,0.1371888369,-0.1417466849,0.0376533344,0.280163765,0.0131864483,-0.3805172145,-0.3432404995,0.0215503462,0.1126778498,0.6594465971,-0.3422845304,-0.3564411998,0.095874913,0.0935508758,-0.3833390176,-0.2534759343,0.6357634664,-0.3619587719,0.0637376755,-0.3649029136,-0.0060929279,-0.245016858,0.1199359074,-0.2295886874,0.0090733347,-0.1717318147,0.0108391419,0.0507865362,0.156946525,0.3423224092,0.1247882247,0.2225869298,0.1706423312,-0.4807884991,-0.0802845806,-0.5487517118,-0.1374809742,0.1584186256,0.0792689025,0.1561627537,0.0094163744,-0.1356783211,0.0310788862,-0.2366456538,0.2848207355,0.0388161391,0.2237691134,-0.0503091887,-0.2578214407,0.345012635,-0.0694997683,-0.2909878194,-0.5439584255,-0.3113088906,0.2484921366,0.0647554994,-0.0952001512,-0.2748989463,-0.0943066329,-0.0721347257,0.165994361,-0.2177155465,0.2759567201,0.0138929682,0.1429402828,0.4675445557,0.1958003789,0.0387668163,0.0351162851,-0.0183568466,0.1494888365,-0.2646305561,-0.1256720126,0.0721442774,-0.4210522473,-0.0233483706,0.1139881462,0.0096137766,0.0315900445,-0.1051627249,0.1368781775,0.4796121716,0.1230212525,-0.0158100594,0.3791803718,0.2303938419,0.2479830235,-0.1812108457,0.1148489639,-0.4754990935,-0.1314103752,-0.4951535165,0.0048490674,0.2056559473,0.01648595,-0.2150599808,0.1820430756,0.210121274,0.1832931787,0.0954422355,-0.1238449365,-0.1386018097,-0.3152528703,0.4940919578,0.089436546,-0.3502280414,0.0345808007,0.4406894445,-0.0463685356,0.2261004299,0.2932239175,0.14811939,-0.0855086967,-0.1110272482,-0.3296705186,0.4267115891,-0.4239212275,-0.0523386598,-0.1957856268,0.13238208,-0.26577124,-0.0977082402,0.1303257346,0.2611166537,0.0910996199,0.1769356579,0.2144690305,-0.0166977216,-0.0255447607,-0.0110306442,-0.0166317895,0.5035648942,-0.0718852207,-0.1488172412,0.3139469922,-0.2549298406,-0.055885464,0.1944796145,-0.0787514821,-0.0455782451,0.0698695853,0.2203968912,0.0215008222,-0.1393847913,0.0578471944,-0.4551055431,-0.0518703535,0.3765419722,0.2090652883,0.0017812524,-0.1541997045,-0.0092121791,0.0174240246,-0.0316410325,-0.0423277281,-0.0057317647,-0.3412948549,-0.010529248,-0.1243202612,-0.2345956713,0.283226341,-0.2481101602,0.18150644,-0.335683465,0.0683290213,0.0299714506,-0.0916065276,0.1092939675,0.214572385,0.0308763776,-0.2610089779,0.0671415254,0.1018130183,0.1645316035,-0.0581313781,-0.4233418405,0.2127869427,0.2228583843,-0.03677525,-0.0745977685,0.1234505102,-0.0332943983,-0.1430667788,-0.361646682,0.3625277579,-0.0108390218,0.1866386086,-0.3714398444,0.4350891113,0.1689334214,0.1201382354,-0.0524894856,0.0288176443,0.0178159252,0.1589596868,0.1490103751,-0.0830552503,-0.8168122768,-0.0437467247,0.0440922417,0.0056203199,0.0737127885,0.0615217909,-0.1806369424,0.1289304495,0.2762049437,-0.3633997142,-0.1357697397,0.0835299119,0.1757229865,0.0601648912,0.0710736066,-0.1641990393,0.0282698665,-0.0812102109,-0.066002138,-0.1247668564,0.1195551977,-0.0269373935,0.0500964075,0.2913753092,-0.029888507,0.1166295782,-0.4848844111,-0.0065647871,-0.3681723475,0.0670638904,-0.4132386148,-0.3664826453,0.0107416939,-0.0310969166,-0.157729879,0.1503382027,0.346829474,0.2500931621,0.2051338553,0.1608518958,0.2119716108,-0.1426743418,-0.1758129746,-0.2458103001,0.1105065942,0.0328496732,0.3166613877,-0.1641018987,0.2430425733,-0.2375469357,0.1929851919,-0.3415489197,0.0782576501,-0.1242570505,-0.1230966151,-0.1152369902,0.0981335044,0.3823597729,0.0002366745,-0.0251259524,0.1249269471,0.1794395,-0.0476490892,0.0180805139,-0.3153509498,-0.041939497,-0.1957779378,-0.292768985,-0.2447785586,-0.2122534066,0.2192921638,-0.017944729,0.1260715425,-0.318810761,0.3760603666,-0.1312127858,0.0854474083,-0.0370538048,0.0366186723,-0.521058023,0.6136804819,-0.2563226521,-0.5021974444,0.1569042206,0.1786887497,0.4287458658,0.4022276402,-0.0092980806,0.1788422316,0.1736887842,0.0001520957,0.0375470333,0.0406192504,0.2938455045,0.0051613618,0.1208508536,0.1078773066,-0.110772036,-0.063993305,0.0457474627,0.3260577619,0.1296520233,0.5017365217,0.0955557004,0.7273526788,0.1096755788,-0.5754067302,0.0250867624,-0.2393105775,0.3088155091,-0.1056425571,-0.3586204946,0.2695134282,-0.0499895699,0.1267050207,-0.0085066753,-0.1291822493,-0.0067318748,-0.1740820408,0.186376825,-0.3261072338,-0.1349490881,0.138260603,-0.3389184177,0.2386583239,-0.3227366805,0.3202269971,-0.0496731065,-0.1763435006,0.2526324391,0.3094443381,0.2343845367,0.1302158386,-0.077821292,0.0152144935,-0.6871008873,0.2909292579,-0.1383936554,0.4108787477,-0.0845230743,0.2319002897,0.1422141641,0.2238976359,0.7626750469,0.1783870012,-0.4628430009,-0.005534647,-0.0901670009,-0.488987416,-0.1416971833,0.0379627831,0.1972917169,-0.0523719154,0.0690905228,-0.1246237457,-0.2433526218,-0.0321111232,0.2817805111,-0.2991705835,0.0557020605,-0.1103805229,-0.0708338395,-0.1124794558,0.0760320723,0.2284829319,0.2087957114,0.0493771061,-0.2428570688,-0.2562504411,0.1293086112,0.4290926754,-0.1435289085,0.1916427165,0.2431151718,-0.1228983104,0.1585057229,0.2474368811,0.2909747064,-0.0748076811,-0.1656505764,-0.1268557757,-0.0157538522,-0.0794741139,0.4617824852,-0.0818618909,0.0343346782,-0.1174079105,-0.2418864667,-0.059698537,-0.33348912,0.2495240867,0.2493163794,0.2935783267,-0.7081381679,-0.5079486966,0.2671420872,0.1704100817,0.2011548579,0.5503252745,0.2755131125,-0.0937255844,0.1817551404,0.5460436344,0.7262740135,-0.437369287,0.0883292481,0.1695214957,-0.1852004081,0.4915032089,-0.4952092767,0.2678431571,-0.3490416408,-0.0129417721,-0.343545109,-0.1704233885,0.0923678204,0.154735595,0.0693901554,0.2467311621,-0.4792620838,0.1329328269,-0.076567471,-0.0640296564,0.1686973125,-0.2746075094,0.0459035784,0.0735660046,-0.221301496,0.1633025408,0.0387776531,0.0078238705,-0.0820803344,0.2580882907,-0.2444155365,-0.1008852422,-0.339466542,0.0662844405,0.0043605776,-0.9245516062,0.3079785407,0.6653313041,0.6603671908,-0.0115602259,-0.1696527898,-0.0574398153,-0.1245386079,-0.1532781273,-0.168695733,0.0241653323,0.2793971002,0.0762407929,0.0564429201,0.3012823164,0.2693251371,-0.0793761089,0.0156172542,0.0329604819,0.0515445359,-0.3542846739,-0.5156933665,0.0778118819,0.314876616,0.0347288735,0.0476152971,0.1336491108,-0.1509747505,0.2008598894,-0.0868731588,-0.1367921233,0.0985727906,0.2253587097,-0.0048616841,-0.2825213969,0.3255103827,0.4667453766,-0.2335277796,-0.0720068291,-0.0452550724,0.5412320495,-0.1509995311,0.0699636266,-0.1486242265,0.3061976433,-0.1462108493,0.252373606,-0.2261338234,0.0204297826,-0.0249018185,-0.1099753082,-0.287463665,-0.1830470264,-0.221732989,0.3471129239,0.2211426049,0.1099042222,0.0347807854,-0.2116915286,-0.1549738199,0.0499274768,0.029291952,0.2112976164,0.1265248805,0.3123990297,-0.0772057101,-0.2509689033,0.013780104,0.0757981688,0.0760443807,-0.135506168,-0.1599954218,0.1292594671,0.0249467902,0.1285496056,0.1513337493,-0.0317271911,-0.2759664953,-0.2957114279,0.2763574123,0.5281389952,-0.0334190093,0.5064694881,-0.0431852713,0.1148932874,-0.1124215126,-0.1123214066,0.2780424058,-0.0100905644,0.1051439643,0.0377379879,0.0768366233,-0.0667946413,-0.3280538619,0.2337832898,0.1334264874,-0.2346501648,0.1191409454,-0.0016483364,0.2164837271,-0.048249092,0.0902827606,-0.0337738134,-0.081768468,0.3144367039,0.2636373043,0.0512534864,-0.0015748254,0.0905373618,-0.0152504034,-0.0281641651,-0.1318352371,0.0769977346,0.1474354714,-0.0191241149,-0.2113027275,-0.1767431796,0.1221657693,-0.3572856486,0.2631485164,0.0600495562,0.2629470825,0.0300521106,0.2148787677,0.01083236,-0.0188162494,-0.1008993909,0.1697933972,0.4328995347,0.3964439631,-0.1226719022,0.2822855115,-0.1244772896,-0.2815710306,-0.3143926561,0.0735902414,0.237586841,-0.0060708001,0.3808269799,-0.1185416058,-0.3919467926,-0.1549634337,0.173368305,-0.3100358844,0.0316709317,-0.1686130613,0.1180938184,0.2738740742,0.0992390662,-0.0267529469,-0.4498534203,0.2340622395,0.118274428,-0.0248378329,0.2466272116,0.1177191362,0.2841738462,0.0447553098,0.0399526991,-0.1534443796,-0.2346263975,-0.1117723957,-0.0916308761,0.2085826397,0.3270852864,0.1454712451,0.1321723014,0.1781019717,-0.0921996459,-0.076067023,0.3173444271,0.202947244,-0.0109718172,-0.0393635295,0.097282581,0.1139340401,0.0900125504,-0.078960374,-0.013252099,0.269638747,-0.2736498117,-0.1661757529,-0.3909124732,-0.3425942659,-0.4184507132,-0.0018124385,-0.1765225381,0.0578070469,0.3780952394,0.0690374747,-0.1483121365,0.0484112613,-0.0011280694,0.0168216657,0.3757729828,0.0264281463,0.3644811511,-0.1024349257,-0.3040473163,-0.1069056466,0.077959694,0.2708779275,0.1656139195,-0.1299972534,-0.0976870954,0.1329159886,0.0638891086,-0.2048265487,-0.1444834173,0.1746788323,0.4526612759,-0.0779576525,-0.0537578613,-0.4576847255,0.0474800766,0.0657894984,0.0188359562,0.0524390638,-0.0736980215,-0.0982628912,-0.2549460232,0.4785960019,-0.0774046257,0.2852930129,0.1914908439,-0.0014076034,0.339348644,0.2105212063,0.2949731648,-0.0881440639,0.1394772828,-0.0894132331,0.3484451175,-0.1745031923,0.1316315979,-0.1218910962,-0.4343792796,-0.1767448634,0.5815299749,0.1364261061,-0.1919686496,-0.1335175782,0.2131027728,-0.2032683194,-0.2481188923,0.6295922399,0.3496640325,0.0629956946,0.177890718,-0.2997425795,-0.0053987647,0.2633985579,-0.2851155102,-0.1775717437,0.0267934278,-0.2837469578,0.3643314242,0.1285240352,-1.0042157173,-0.0508150682,0.0431104377,-0.2969077528,-0.1156271398,0.2367577255,0.0655907094,0.156638816,-0.1750954241,0.3621439338,0.1089465693,0.0532211103,-0.4503191113,-0.3442657888]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1762","title":"Unable to format dataset to CUDA Tensors","comments":"Hi @lhoestq,\r\n\r\nThanks a lot. Is this true for all format types?\r\n\r\nAs in, for 'torch', I can have `**kwargs` to `torch.tensor` and for 'tf' those args are passed to `tf.Tensor`, and the same for 'numpy' and 'pandas'?","body":"Hi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan","comment_length":38,"text":"Unable to format dataset to CUDA Tensors\nHi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan\nHi @lhoestq,\r\n\r\nThanks a lot. Is this true for all format types?\r\n\r\nAs in, for 'torch', I can have `**kwargs` to `torch.tensor` and for 'tf' those args are passed to `tf.Tensor`, and the same for 'numpy' and 'pandas'?","embeddings":[-0.1914244592,-0.4953543544,-0.0635152757,0.1430512965,0.5737578273,0.3578240275,0.6109963059,0.3751015067,-0.0064690905,0.093313165,-0.1843090802,0.238564074,-0.1888935864,0.2082742453,0.1855475008,-0.3550991416,0.2348218411,-0.1543388367,-0.0376627818,0.0704863742,-0.2085145414,-0.0241091959,-0.1562636644,-0.2900007665,-0.1949911267,-0.1617193669,0.0201715548,-0.4400323927,-0.1024810895,-0.0641930923,0.2913292944,0.104888238,0.3066851795,0.4878616333,-0.0001256564,-0.0178643558,-0.0497946441,-0.1510568112,-0.1228252649,-0.2663969994,0.0131856529,-0.2795284688,0.3273176849,-0.2697499394,-0.3616902232,0.1351331323,0.0503888875,-0.3767498434,0.200864166,0.5606645942,0.0873320624,0.1567419022,0.2239812762,0.0313126706,-0.1401109546,0.4742643237,-0.3869950175,0.0321357176,0.1239138469,0.2313414663,0.2390954942,0.121947214,-0.1170619056,0.0366932191,0.2436490506,0.0068774805,-0.3590634465,-0.2845258713,0.0415751301,0.2316781729,0.7356418967,-0.305837065,-0.4198352993,0.0801290721,0.0384078212,-0.3957471251,-0.1968512535,0.6596462727,-0.3205678761,0.0670806766,-0.3185796142,0.1147153229,-0.2370989323,0.1259754151,-0.2630985379,0.0228494518,-0.2386255264,-0.00528268,0.0134357577,0.1373595893,0.2915566266,0.1084817946,0.1864098608,0.1729704142,-0.4968365729,-0.1123953313,-0.5637843609,-0.1750736088,0.145513311,0.0750934631,0.1889846027,0.0470552519,-0.1527156979,0.0456818715,-0.1689350009,0.224168852,0.0323081762,0.1593714356,-0.052119717,-0.2683012784,0.339374572,-0.0639518201,-0.2907772958,-0.5000951886,-0.294485569,0.1828467697,0.0936030969,-0.0746787786,-0.2550733685,-0.1188581735,-0.0248166993,0.1076056436,-0.2666125596,0.249533087,0.0253214706,0.3020795584,0.3945814669,0.1859082431,0.0624374934,0.0885022879,0.0459994301,0.0976752639,-0.2249604762,-0.2046596855,0.0503197499,-0.3494428098,-0.0520413965,0.1319437474,-0.02984442,-0.0050222706,-0.1101642549,0.0849643946,0.4755417109,0.1198236719,-0.1078406125,0.4093901813,0.2108049691,0.2150567323,-0.2697773278,0.1633630097,-0.4828787744,-0.050340388,-0.4977714121,-0.0458525456,0.2782033086,-0.0052368548,-0.2089643925,0.2063405514,0.2674427032,0.1263454705,0.1163300574,-0.1863862425,-0.1187278032,-0.2959930599,0.4839713573,0.0943365097,-0.346241802,-0.0135652162,0.5384857059,-0.0493879654,0.2286550105,0.3445949554,0.1999177635,-0.1558273286,-0.1161526293,-0.3076292574,0.514144659,-0.3327646852,0.0620910712,-0.2046283185,0.238594085,-0.2397407144,-0.0306929629,0.0901034251,0.2652629912,0.0417132713,0.1719368994,0.2446942329,-0.0256077517,0.0062519168,-0.020202009,0.0564026684,0.459436506,-0.0907293856,-0.1993737668,0.3107618392,-0.2633801699,-0.1151411682,0.162151441,-0.0987520665,-0.0622772612,0.012384885,0.2049012482,0.0305001959,-0.1359109133,0.0358943231,-0.4950877428,-0.1287979037,0.3969790339,0.1687712073,0.1217771024,-0.1747484952,-0.0030755748,-0.0099151405,0.0079434421,0.0448172875,-0.0464094318,-0.2493266314,-0.0874044448,-0.0615764409,-0.3723646998,0.1812834293,-0.2245939821,0.1582474411,-0.2071206272,-0.1247370765,0.0592541248,-0.0716325566,0.0630290061,0.2352627367,0.0131812859,-0.2124841213,0.0284208134,0.0819359645,0.2183371037,-0.0558925457,-0.5166804194,0.2893258929,0.2828788459,0.104551442,-0.1155059114,0.0311221592,0.0055087954,-0.1629833579,-0.3709156811,0.4067362845,0.0599326901,0.230954513,-0.3554264009,0.3759162426,0.1866152138,0.1095321998,-0.0815121084,0.0628495738,-0.0306959264,0.154332459,0.1242306978,-0.0269195717,-0.8723095059,-0.0826110169,0.0893359408,-0.0524719991,0.1247086152,0.0851079971,-0.1595858186,0.1669843048,0.2630163133,-0.3978264332,-0.1376062632,0.0414259285,0.1940864325,-0.048570931,0.0565142669,-0.233563751,0.0186636318,-0.0982431695,-0.0915792584,-0.0727772489,0.107151702,-0.0184024405,0.0337337218,0.3695034683,-0.0305227917,0.0190546066,-0.5362110138,-0.0158574786,-0.3519634902,0.0977476537,-0.3490992188,-0.3521323204,0.1291348189,-0.0189528149,-0.1611277312,0.0534757115,0.3019583523,0.2789478302,0.1680697799,0.2721075416,0.1798411161,-0.1224969178,-0.2118988186,-0.2402275205,0.1241373718,0.000944814,0.4400053918,-0.2165757865,0.2178514898,-0.2075906545,0.1460314691,-0.2733972967,0.0451167375,-0.1020593345,-0.1112595499,-0.1189441085,0.08324451,0.3903753459,0.0218295809,-0.0288239792,0.135019049,0.1823802292,-0.0211636052,0.0235007703,-0.3299600482,-0.0373772122,-0.2545598149,-0.2003005892,-0.232990548,-0.2443696707,0.2478479892,-0.0416196585,0.1478679627,-0.3123798072,0.3225725591,-0.1367617995,-0.0515976213,0.0416276306,0.0545709059,-0.458408922,0.6442713141,-0.2775747478,-0.4975637794,0.1916207522,0.1685626656,0.3659411967,0.448874265,0.0472360961,0.1143314838,0.1473894864,-0.1106012315,0.0562096797,0.0760409087,0.2353334278,-0.0226625111,0.157155171,0.1175529584,-0.0740970597,-0.0103896093,0.0545890033,0.2211627066,0.2090297937,0.4589605927,0.0763145536,0.6128472686,0.0670986921,-0.6483486295,0.0579884239,-0.2046810538,0.3649874926,-0.1180188134,-0.2335178256,0.3266437054,-0.0169422813,0.1533859968,0.0291706175,-0.150801301,-0.0952664688,-0.1970416456,0.2728224099,-0.3071236014,-0.1540924609,0.2696659863,-0.3257266283,0.2577631474,-0.3124789298,0.2771010399,-0.0672111511,-0.1610768288,0.2538989782,0.3399350941,0.2657427192,0.1425737441,-0.1545014232,-0.0029822981,-0.5467552543,0.2598885298,-0.0741064101,0.4347230792,-0.0868902579,0.1399039477,0.0934223235,0.2414610088,0.772678256,0.2027324289,-0.5556976199,-0.011864909,-0.0835212022,-0.4828517437,-0.1213544607,0.0065440647,0.1386806667,-0.1353883445,0.1908851117,-0.2186716795,-0.2697869837,-0.0511344559,0.2668797374,-0.2181890309,0.0268845186,-0.0829345882,-0.0512741245,-0.0673806071,0.0994904563,0.2984827757,0.1287569404,0.0215685517,-0.2902406454,-0.3583566248,0.1438719183,0.4408706427,-0.1962933987,0.1931112558,0.2548776269,-0.2116420567,0.0877540782,0.2692282498,0.2353906929,-0.180731222,-0.168445617,-0.091791831,0.0657554567,-0.0601532012,0.4541960955,-0.0447351299,0.0373751931,-0.1145983115,-0.2269802988,-0.1090813056,-0.3234998882,0.2126027644,0.2521570623,0.3025761843,-0.7884063721,-0.5442990661,0.2448840737,0.1512766927,0.2036586404,0.4967103899,0.3014929891,-0.0893107131,0.1331835836,0.5167665482,0.7235730886,-0.4403663874,0.2087450922,0.1528609097,-0.135205552,0.486353159,-0.468397826,0.2737192512,-0.3209706247,-0.0333502591,-0.407974273,-0.240257591,0.0910375938,0.1453151405,0.1042102352,0.1746630073,-0.4771811366,0.2962515354,-0.0020691252,-0.105137147,0.0622182898,-0.3006749153,0.0242569912,0.0447032563,-0.2382386923,0.0619882569,0.0513072163,-0.0498453714,-0.0427678376,0.2735278606,-0.2546387315,-0.1487122029,-0.3635269105,-0.0240821913,0.0802915096,-0.9106853008,0.3314269483,0.6718658209,0.6400459409,-0.0299200155,-0.1160826981,-0.123391442,-0.0824023113,-0.1175472364,-0.1499608606,0.0598965026,0.2463067919,0.1009033322,0.1222558618,0.3658819795,0.2592406273,-0.0899253041,-0.0198987406,0.008575717,-0.0084149148,-0.3936410248,-0.6244149208,0.0060454998,0.3228117526,0.089100346,0.0279464293,0.1268392354,-0.1233137697,0.2705704272,-0.1016560048,-0.1050185785,0.1336261332,0.1531156749,0.0065718633,-0.2178822756,0.3395912349,0.3548999131,-0.2539772391,-0.0367240608,-0.0327969566,0.4872442186,-0.1175085977,0.0898446441,-0.121131748,0.1823830903,-0.1915600151,0.314158231,-0.214355424,0.0506001003,-0.0304672681,-0.1168168113,-0.1696013808,-0.0866609514,-0.2223856002,0.3939132094,0.1930728704,0.147480607,0.0428572297,-0.2638688087,-0.1246192381,-0.0007995364,0.0246535596,0.2186475992,0.0373856314,0.342235446,-0.074458234,-0.1857156605,0.0047265305,0.1263846755,0.0988324732,-0.1255887896,-0.1860200912,0.1571830809,0.0885666013,0.1546735018,0.1922037452,-0.0290545225,-0.3014992476,-0.2998075187,0.2054164559,0.4749247432,-0.082278803,0.5092017651,-0.0395168886,0.1536429375,-0.037647862,-0.0112228906,0.2795646787,-0.0268449131,0.1563315094,0.0051210797,0.0481975786,-0.0066693509,-0.3012356758,0.3035320938,0.220890373,-0.241597712,0.0347440653,-0.0162920486,0.1615001559,0.0111292358,0.0818254352,0.0272702891,-0.0798231363,0.3388363421,0.2370690405,-0.0012147756,0.0374929756,0.1251108944,0.0685419291,-0.0216491614,-0.0595895015,0.1306432486,0.1558703035,-0.1192541942,-0.108807683,-0.2447170466,0.0351114869,-0.3608444333,0.277954936,0.0782328248,0.2904632688,0.0649636984,0.1931928992,0.0119620841,-0.0322425589,-0.2728925347,0.2033378482,0.337523669,0.3924202025,-0.1142239645,0.3086024821,-0.2124437839,-0.2520955801,-0.295330584,0.0957645774,0.3000318706,0.0520972461,0.4198810756,-0.0845129117,-0.4007326961,-0.1573777646,0.1022641063,-0.30235973,0.0060130982,-0.2600737214,0.1281616539,0.2903257012,0.0516579375,-0.0867060423,-0.3957056701,0.2572455406,0.1843475401,0.0914098769,0.2649308145,0.1433956325,0.2192559838,0.0652457848,0.0689195171,-0.0540590547,-0.2141500711,-0.0996983573,-0.0414924249,0.2261923403,0.2643825412,0.2051022649,0.1496020705,0.1185393259,-0.1286414415,-0.0508817583,0.3582123518,0.1805955917,0.0865351111,-0.066182889,0.1045112982,0.0872217119,0.1204346716,-0.0297857597,-0.0625433847,0.1956659704,-0.2986263633,-0.1839805394,-0.3820001185,-0.2986164391,-0.3454685807,-0.0250276551,-0.195041284,0.0397908315,0.3429348469,0.0748338252,-0.1223473698,0.1172995195,-0.0189204793,0.0689859763,0.3252040446,-0.0286588725,0.3553131819,-0.0247253627,-0.2297629416,-0.1227491051,0.0783759728,0.3053512275,0.2091906071,-0.1296176016,-0.0933266357,0.1991824061,0.0890193656,-0.1796441674,-0.1754827946,0.1452627629,0.437238723,-0.0852479339,-0.1261150837,-0.4263877869,0.0997572988,0.0162860565,-0.0552185066,0.0447842404,-0.0466197468,-0.1109094322,-0.2146396637,0.4659695625,-0.0741572604,0.3106869459,0.1252733171,0.0046409545,0.2990760803,0.2199093401,0.2212869674,-0.0824818611,0.1476169527,-0.136006102,0.2829578519,-0.2025212049,0.1033237502,-0.2090316117,-0.3423644602,-0.1338417828,0.5338133574,0.0748816952,-0.2111455947,-0.0728997737,0.1870834082,-0.199167639,-0.3507272899,0.6488537788,0.3711220324,0.0672060475,0.1906409413,-0.2452268153,0.0379550532,0.2480581105,-0.3406535089,-0.189187482,0.084176138,-0.3319997191,0.3432599902,0.2048108727,-0.949313581,-0.0206992961,0.0416903459,-0.2652865648,-0.0787372366,0.2311904877,0.0780442208,0.0643083751,-0.2288236022,0.3666216135,0.1664108783,0.0501810499,-0.5354930162,-0.3855128288]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1762","title":"Unable to format dataset to CUDA Tensors","comments":"Yes the keywords arguments are passed to the convert function like `np.array`, `torch.tensor` or `tensorflow.ragged.constant`.\r\nWe don't support the kwargs for pandas on the other hand.","body":"Hi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan","comment_length":26,"text":"Unable to format dataset to CUDA Tensors\nHi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan\nYes the keywords arguments are passed to the convert function like `np.array`, `torch.tensor` or `tensorflow.ragged.constant`.\r\nWe don't support the kwargs for pandas on the other hand.","embeddings":[-0.2339969426,-0.4713830352,-0.0587259531,0.1454277188,0.5945415497,0.3813997209,0.5536247492,0.3486228287,0.0204021242,0.0847379118,-0.1387395561,0.3407218456,-0.1757797003,0.1476910412,0.1723191291,-0.3778468668,0.3084686399,-0.0880436674,-0.0487243161,0.0729149655,-0.2402442247,-0.0089956438,-0.1322969198,-0.2712605894,-0.173953414,-0.2240069062,-0.0373401642,-0.3639823198,-0.1242673844,-0.0815783963,0.280126363,0.0425559804,0.2684721947,0.479721576,-0.0001250429,-0.0380646735,-0.0376146324,-0.1130806655,-0.0815804303,-0.3308831453,0.143385753,-0.3101426661,0.3145223856,-0.2309604138,-0.4044010341,0.1008605212,0.0041844072,-0.3171213269,0.2870211303,0.5416358709,0.103670679,0.1582948267,0.2337694168,0.0570037179,-0.2884039283,0.3508221805,-0.3622895777,0.029085191,0.1062355489,0.0989784002,0.2423032969,0.1262860745,-0.1447199285,-0.009933468,0.3419010639,-0.0117045538,-0.3300116956,-0.343046397,0.0289544445,0.1749110073,0.5755484104,-0.3697309792,-0.3937684298,0.0476243757,0.1024229005,-0.4620788097,-0.2309355736,0.6167895794,-0.3567797244,0.1083152667,-0.3492470682,0.0367215872,-0.2219844311,0.1147983447,-0.2369441837,-0.0129355257,-0.1936635822,-0.0190675873,0.111353457,0.1237546578,0.3501030505,0.1134016663,0.2030448914,0.1549072415,-0.4486976564,-0.0980172381,-0.5485565066,-0.1681699306,0.1375159323,0.0402063243,0.1304198951,0.0079594404,-0.1278871298,0.0056559616,-0.2242074311,0.3471833169,0.0669080615,0.1643438339,-0.0432013609,-0.2695736289,0.3038643897,-0.0564141087,-0.2316732109,-0.5269571543,-0.316483289,0.1881118268,0.0535000786,-0.0537037998,-0.2678466141,-0.1250574142,-0.0247980822,0.2073383927,-0.2367475629,0.3064680099,0.0266959406,0.175603658,0.4172639251,0.2402886152,0.0884256884,0.1147113591,0.0082564084,0.1830558926,-0.2545364499,-0.1619744003,0.0182341728,-0.408423394,-0.0000098696,0.114477843,-0.0146229621,0.0369337648,-0.1712637842,0.1097175553,0.4806713462,0.1500868648,-0.0555732399,0.3168988526,0.2587451339,0.171579197,-0.1793723553,0.1278153509,-0.4333970845,-0.0524579324,-0.4958718717,-0.0392540209,0.1906520277,0.0205783788,-0.2681034207,0.2509735525,0.1977762133,0.1804834902,0.1070223823,-0.1601639986,-0.1691135317,-0.3387620151,0.4698261321,0.0842390805,-0.3453785181,0.0297357813,0.434987694,0.0004083192,0.2666268945,0.3180944026,0.174704507,-0.1362379342,-0.0582676083,-0.3067197502,0.5030120611,-0.3811711371,-0.002110458,-0.1177432239,0.1807863861,-0.2046895921,-0.1161084548,0.0791502222,0.2161780149,0.0598128401,0.1382626891,0.2432869822,-0.0262607913,-0.0294425562,-0.0116635039,-0.0437627174,0.555408299,-0.1015026867,-0.1470306069,0.3318081796,-0.2507635951,-0.113105461,0.1760647446,-0.0809093043,-0.0702735409,0.063646093,0.2080050111,0.0232507326,-0.1183819696,0.0695905983,-0.4207933247,-0.0812511146,0.3631971478,0.1965861619,-0.0006355069,-0.1281881928,-0.0671002492,0.0156985279,0.0038450852,-0.0556031615,-0.0155436303,-0.388495326,-0.0255307052,-0.0298726223,-0.2764238715,0.2545250058,-0.2664058805,0.172347039,-0.2724768817,0.0402159765,0.0574771687,-0.1314865947,0.1593995988,0.2320960909,0.0638310835,-0.2490208894,0.048407007,0.1410157382,0.1752935052,-0.0210816301,-0.4079723358,0.2039381415,0.2236028016,0.0174957719,-0.0682967976,0.141394645,-0.0304877721,-0.1648557484,-0.4090501666,0.3824997842,-0.0257203504,0.2120332271,-0.3623810112,0.4157853723,0.20953843,0.0852198824,-0.0370515585,0.0753452778,-0.0228061192,0.0823445246,0.1773440689,-0.0755217895,-0.8429781199,-0.0377741903,0.0881043673,0.0032723898,0.09828493,0.1012567803,-0.1349601299,0.0992460772,0.293107599,-0.3925173283,-0.1315825284,0.0586476363,0.1630243063,0.0702608079,0.0551951453,-0.1624016315,0.0284021366,-0.0688166469,-0.0366205983,-0.172326535,0.131505087,-0.0544202663,0.0410273485,0.3336190283,-0.0218048412,0.1382029355,-0.4633518457,0.0450826026,-0.4018615484,-0.0179238003,-0.3308508396,-0.3282232881,0.0491726622,-0.0365940444,-0.2419452816,0.1314435303,0.3399285078,0.3103050292,0.2322828323,0.1405813694,0.2255209386,-0.0986832827,-0.2828424573,-0.2550128996,0.1266607791,0.0233456176,0.3692157269,-0.1241345257,0.1733714044,-0.195393905,0.2128217965,-0.3078772426,0.0772547796,-0.1389772892,-0.1688556969,-0.0307458434,0.1255291551,0.3994468749,-0.1166265905,-0.0213176664,0.1272454411,0.1648911834,0.0273488835,0.0256968681,-0.2978483438,-0.0583029315,-0.1860079616,-0.2147425562,-0.2217249423,-0.2360036522,0.253924638,-0.0015282225,0.1012187228,-0.2703938782,0.428529501,-0.1214859039,0.0936646014,-0.0723860636,0.094597362,-0.4932852089,0.6376007199,-0.2505540252,-0.5051738024,0.1750522554,0.1669685543,0.4039465785,0.3599305749,-0.0157587193,0.1210851297,0.1616235971,-0.0264710654,0.0623376891,0.0371967033,0.2368351221,-0.0230726916,0.131292358,0.1076188013,-0.0747549087,-0.0488992855,0.099597685,0.3001369536,0.1771588475,0.559242785,0.0930162221,0.6866155267,0.1286579221,-0.5674318075,0.010748093,-0.2740852237,0.3683100939,-0.1180503145,-0.3410993516,0.2362449467,-0.0375306904,0.1796848476,0.019972777,-0.1156372651,-0.0244520456,-0.1877105981,0.1029733196,-0.3049916923,-0.1181727648,0.1908495873,-0.3305314481,0.165181309,-0.315515697,0.3517726958,-0.0692252591,-0.1802682281,0.2391376346,0.3620740175,0.275984019,0.1826495975,-0.0466220826,-0.006591517,-0.6597186327,0.253644079,-0.148320809,0.4231063724,-0.0879114568,0.2414022982,0.1323263794,0.2363472283,0.8280133605,0.1968530864,-0.5567103028,0.0190378185,-0.1075336039,-0.5107793808,-0.1107200384,0.0723911375,0.1427491158,-0.077439189,0.0941019729,-0.135724172,-0.2569351196,-0.0863170028,0.3119142652,-0.2843702137,0.0239941198,-0.1211042553,-0.1073609963,-0.1028807685,0.0914705619,0.2308840007,0.1489974856,0.0312933959,-0.2396965474,-0.2657473981,0.093497768,0.4447605312,-0.1039798707,0.1937540323,0.2402018756,-0.1588641405,0.0563891791,0.1761802137,0.2523498833,-0.1300628334,-0.2216259539,-0.1375313103,-0.0766494945,-0.1014094949,0.484934032,-0.0676988587,-0.0314504765,-0.1294441819,-0.2315019518,-0.05989106,-0.2750641704,0.1962395608,0.2176714242,0.2921273708,-0.6761106253,-0.5005911589,0.308051616,0.1360964924,0.2173369974,0.5419599414,0.3116105199,-0.0733251125,0.2043450624,0.4828853309,0.7467299104,-0.4484685063,0.0871140137,0.2286560088,-0.2027852684,0.5286083221,-0.4583533704,0.261725843,-0.3106412888,-0.0468288586,-0.3594837785,-0.1660862416,0.1121765748,0.1560602486,0.0301767159,0.2758594453,-0.4692665637,0.1199285686,-0.0416723602,-0.0396449305,0.1339377314,-0.2912847102,-0.0638097152,0.0522450209,-0.1875170171,0.1136719361,0.0789010599,-0.0623909049,-0.1171060726,0.280148983,-0.2094515115,-0.0915093869,-0.2913132906,0.0351174995,0.0444838293,-0.9460867047,0.2878387272,0.6920858622,0.6729732156,-0.0539445505,-0.1622531712,-0.0928835273,-0.1192679107,-0.1274364442,-0.1747278571,0.0548550636,0.3042757511,0.0911561027,0.1132734939,0.2836304605,0.2406707406,-0.0439344235,-0.0002022966,0.0242049024,-0.0465829894,-0.3549320698,-0.5426561832,0.0033772455,0.2708292902,0.0374263823,0.028977884,0.1265203506,-0.0738216937,0.2401249409,-0.1189150065,-0.134951055,0.0792446285,0.1422627866,-0.0135104759,-0.2353617996,0.3112528026,0.4385105669,-0.2563995421,-0.0221350584,-0.0889918506,0.4913953841,-0.0746981874,0.1575949937,-0.162722826,0.2984327972,-0.167040363,0.2334007323,-0.2054908574,0.0543498918,-0.0556946285,-0.1090748087,-0.2791038752,-0.1723313928,-0.1887349486,0.3204328716,0.2089359909,0.1702379882,0.0825770572,-0.1930774301,-0.1443174928,0.0172904469,0.0009180605,0.2042666525,0.1081136316,0.3062471151,-0.1017630473,-0.2382640839,0.000216053,0.0964022353,0.0771968663,-0.1164920777,-0.1810879707,0.1445774734,0.0485258922,0.1201601103,0.1659923047,-0.0196213946,-0.2690389156,-0.3319353163,0.221333608,0.5063858032,-0.0520913862,0.5310232043,0.0156911816,0.1246673539,-0.0467447415,-0.1051017717,0.228362307,0.000969684,0.0690018535,0.0398609452,0.1266223192,-0.0482469872,-0.2575536072,0.1838673204,0.1103387251,-0.2583168745,0.1548948139,0.0688125268,0.1719314605,-0.0313290916,0.024329396,0.0608489402,-0.0913805962,0.3637606502,0.2778079808,0.0094633456,-0.0219558273,0.1275048107,0.0253048912,-0.0299194548,-0.1048790514,0.095012337,0.1831591278,-0.1008542702,-0.1703537852,-0.2146994323,0.1838196665,-0.4327237904,0.1902825236,0.1170277819,0.2538706958,0.0556347892,0.2116870284,0.0399436504,-0.0420731269,-0.0849588141,0.1783195138,0.3768408298,0.3827632964,-0.1203703731,0.3099438548,-0.1055229977,-0.2833268046,-0.2708876729,0.0052581918,0.2114012539,0.0384193137,0.3883092701,-0.0909429193,-0.3526311517,-0.1037718803,0.2078643739,-0.2998798788,0.0348779969,-0.1480338275,0.1560299098,0.3107509017,0.0497475937,-0.0431060493,-0.4737075567,0.1968720853,0.1356637329,0.0340192243,0.3772858977,0.0854439139,0.236226812,0.082014136,0.0096494639,-0.2035853118,-0.1631763428,-0.1027629003,-0.1156549156,0.1711797714,0.374936372,0.1542989016,0.112427026,0.1536953449,-0.0506163687,-0.0966267511,0.3804787099,0.2438126206,0.0128500247,0.0483910739,0.0619366542,0.0990793854,0.1020308435,-0.0945561007,-0.0530292653,0.2216544896,-0.2926778793,-0.1310749799,-0.3994072378,-0.3589651585,-0.3850874007,-0.0181527529,-0.1514041424,0.102150254,0.3458804786,0.0406321138,-0.147754699,0.067339696,-0.016040612,0.0564908609,0.3583323061,0.0135487458,0.3651579618,-0.06246797,-0.2486884445,-0.1333655417,0.0395069495,0.2420811653,0.1670306623,-0.1253892481,-0.1192721426,0.113257505,0.0601062402,-0.2492971569,-0.1358513087,0.1473100781,0.4346064627,-0.0511639453,-0.1123150066,-0.4668065608,0.0582287796,0.0485129096,0.0328894481,0.1192582473,-0.0798923671,-0.1070711389,-0.2286001742,0.5114039183,-0.0823114961,0.3495625556,0.118604511,0.0059268624,0.3164976537,0.2320826203,0.2943645418,-0.0598478355,0.1566628814,-0.0823221728,0.3606934249,-0.1621673852,0.1024591252,-0.1270540804,-0.4881169498,-0.1648622751,0.5384293795,0.1112362742,-0.1753000021,-0.0873459354,0.2083647698,-0.2033679485,-0.284755677,0.6637989879,0.3430191576,0.0469379276,0.1581352055,-0.2644059956,-0.0194741003,0.2774275541,-0.3332233429,-0.2580160201,0.0428676903,-0.3855658174,0.3643036187,0.1272259951,-0.9844715595,-0.0503295101,0.0165103581,-0.2429480553,-0.1102323905,0.2102184594,0.0143804923,0.1292782128,-0.1888148487,0.2801195383,0.1004270986,0.0505047217,-0.5019500852,-0.3692333698]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1762","title":"Unable to format dataset to CUDA Tensors","comments":"Thanks @lhoestq,\r\nWould it be okay if I added this to the docs and made a PR?","body":"Hi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan","comment_length":17,"text":"Unable to format dataset to CUDA Tensors\nHi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan\nThanks @lhoestq,\r\nWould it be okay if I added this to the docs and made a PR?","embeddings":[-0.2076083124,-0.4493276179,-0.0808781087,0.142959103,0.5672462583,0.3796422184,0.5310692787,0.3309663832,-0.0181646831,0.1409579068,-0.0788681507,0.2365308106,-0.2230986804,0.1325985044,0.1703114212,-0.3306939304,0.2573104203,-0.0979152545,-0.0379618742,0.0790335238,-0.1359184682,-0.0128564648,-0.1050756574,-0.3105601072,-0.2359376848,-0.1902676523,0.0561589748,-0.3857475519,-0.1967529207,-0.0576163493,0.1955188215,0.1002451479,0.29746297,0.4419077337,-0.0001257112,-0.0405382775,-0.0763713568,-0.1063870415,-0.080232136,-0.2797963619,0.1008523479,-0.2619596422,0.3520948589,-0.1931514144,-0.4167786539,0.1588604003,0.0376700573,-0.3366959691,0.2211165577,0.5201233625,0.1055461466,0.201981917,0.2504799366,0.024018513,-0.1411984116,0.4127913415,-0.4059521258,0.0593819618,0.1868831068,0.1419476122,0.2361409068,0.1956163645,-0.1018036827,0.0367225483,0.2884670198,-0.0079526799,-0.3925335705,-0.3323218226,0.0225080829,0.1464354545,0.7426348925,-0.3099430799,-0.3527896404,0.1035459638,0.0786293224,-0.3819108605,-0.2360118926,0.6073791981,-0.3357323706,0.1058309004,-0.3175977767,-0.0252773352,-0.2114556581,0.1135248095,-0.2010389268,-0.0210831333,-0.2168882936,-0.0134135429,0.0800354704,0.1509759128,0.3454316556,0.212830469,0.2072887421,0.146277532,-0.457313329,-0.1075192839,-0.5549705625,-0.1373179555,0.1715201437,0.0919843167,0.197463721,-0.002513587,-0.1347372383,0.0147332316,-0.2574051321,0.2582004666,0.0727009997,0.2164323777,-0.0381977893,-0.2039446384,0.3264054954,-0.0522569455,-0.3161253333,-0.5621752739,-0.3139206767,0.2240673155,0.0709721297,-0.1325156987,-0.2899289429,-0.0630590469,-0.0203232691,0.1519998461,-0.195232898,0.3259002268,0.0424672514,0.1394878179,0.4517869651,0.1938624978,0.0105005354,0.0338832811,0.0050432449,0.1696887314,-0.3044994771,-0.0738299713,0.0310795661,-0.3389312625,-0.0471709073,0.0743234605,-0.0217698961,-0.007042604,-0.1374233067,0.1592934132,0.528816402,0.1485944837,-0.0566933826,0.3177837431,0.2165926993,0.2340798825,-0.2023090124,0.1190707684,-0.4427163005,-0.1329214722,-0.5485044718,-0.0207916033,0.226234585,0.0080709346,-0.1792021841,0.2097907364,0.2465562373,0.1304863691,0.1013705879,-0.0826847628,-0.1323800981,-0.3094055057,0.4982712269,0.1535867304,-0.3233247101,0.0316362903,0.4733214974,-0.0435247943,0.2070112228,0.3417697847,0.1974048018,-0.050364919,-0.1204217672,-0.3095732033,0.4586009383,-0.4009733498,-0.024447076,-0.1857907623,0.1177423745,-0.2924159765,-0.0787742212,0.0899528787,0.2781101465,0.0961890295,0.1539414227,0.1623061746,-0.0119309342,0.0019589656,-0.0272789262,-0.0119462311,0.4723087251,-0.0996463746,-0.1939267218,0.3507114649,-0.2216272205,-0.0664068758,0.1824704558,-0.0920220986,-0.0442125946,0.0883365497,0.2601041198,0.0278739631,-0.1492639631,0.0622367114,-0.4646399319,-0.1267797053,0.434522599,0.2147192508,0.0734812245,-0.1725203544,-0.0275027994,0.0007657374,-0.0201715771,-0.0845566466,-0.028565878,-0.3560156226,-0.0588525571,-0.1167134345,-0.2934119999,0.2282381803,-0.3179704547,0.2113440931,-0.3106459379,0.0084493449,0.05475666,-0.0987833515,0.0593863875,0.2388004661,0.0139859179,-0.255692184,0.0610446706,0.0880655795,0.1665399373,-0.0616914034,-0.3724156618,0.2719942629,0.2509323359,-0.0638868287,-0.1255123764,0.0395871699,-0.0599653982,-0.1379622072,-0.4441963732,0.3333820403,0.009532907,0.1833406538,-0.3780225515,0.3914254606,0.1607538313,0.1127121523,-0.0941014662,0.019278327,0.0129382582,0.1788825095,0.1063875332,-0.0939074531,-0.8347510695,-0.0346029177,0.0099747889,-0.009352969,0.1038298309,0.0902130902,-0.0889422148,0.1563893557,0.2987086177,-0.3835813105,-0.1463311315,0.0697986335,0.1932675242,0.0536920652,0.0900284275,-0.1724131256,-0.0065649915,-0.1495097131,-0.0629123747,-0.1108783409,0.1053052619,-0.0268622749,0.0198155977,0.3037309647,-0.0329217874,0.1207721233,-0.5268215537,-0.0821063891,-0.3464764059,0.0970350206,-0.3644326031,-0.3428700268,0.0165922493,0.0103524057,-0.1444885582,0.1534781605,0.3193197548,0.2531527877,0.2624109089,0.1782203615,0.1932383478,-0.0999086872,-0.1598995328,-0.2299977094,0.1815827489,0.0119407261,0.3474968672,-0.1560917199,0.204447493,-0.3022927344,0.1655821502,-0.3841092288,0.0206680726,-0.0932870582,-0.104485698,-0.1126242355,0.1318238974,0.4242872596,-0.0067447647,0.0133429272,0.1077385843,0.1408372372,-0.0286890734,-0.0282626059,-0.339951098,-0.0449544862,-0.2176354229,-0.281904906,-0.1910872608,-0.2567911148,0.256862998,-0.0152868703,0.1554702967,-0.3396693468,0.3959318995,-0.1299618036,0.0046285097,-0.0234765373,0.0288503002,-0.5347702503,0.6057141423,-0.2924804091,-0.5323602557,0.1957694292,0.1842218935,0.3844261765,0.392991513,0.0005280214,0.1903716028,0.1710219234,-0.0711646229,0.0433815904,0.0737030581,0.3059945703,-0.0140764154,0.1499236375,0.1225319579,-0.1269413978,-0.0466792248,0.0888376534,0.2946559489,0.1153137833,0.4837462604,0.0614362136,0.6757726073,0.0886685252,-0.5966697335,0.074806653,-0.2757519186,0.3746644557,-0.1389728934,-0.2873623073,0.295643419,-0.031816341,0.1092242301,0.0229961742,-0.1211837381,0.0397660546,-0.14700903,0.2281887829,-0.3481969535,-0.1224866956,0.2093814909,-0.3385240138,0.1809763014,-0.3129068017,0.3045587242,-0.0584182851,-0.1516751796,0.1967682391,0.3891717494,0.3139397204,0.1189284995,-0.1016838029,0.0703647956,-0.6821168661,0.2481828928,-0.1574200392,0.3805269599,-0.1512129456,0.2029242069,0.1421333551,0.2285236865,0.7871838808,0.1639446169,-0.5181646347,-0.060919106,-0.0635137931,-0.4389618337,-0.1175813302,0.0592319034,0.173729524,-0.1141628325,0.1568437517,-0.1653032452,-0.2418538034,-0.0125079947,0.2245450616,-0.2485740334,0.0428930745,-0.1151407361,-0.0653065145,-0.1427768469,0.1266492158,0.3015598357,0.1439415812,0.0740274042,-0.2782035172,-0.2106913328,0.1488359123,0.4324990809,-0.1354568601,0.191882059,0.2367688268,-0.1541232318,0.1587654203,0.2291404009,0.2965294421,-0.0599356219,-0.2078420222,-0.1069347039,0.0463185534,-0.0780026838,0.4774352312,-0.0192241706,0.0863265842,-0.1339597106,-0.1964454502,-0.0529191904,-0.3171695769,0.2225009203,0.2268725038,0.3045936823,-0.7227424979,-0.4659357965,0.2787935436,0.1624550968,0.2051001638,0.5931316614,0.3560985029,-0.0579251386,0.147376895,0.5321173072,0.7870185375,-0.4854322374,0.1115230247,0.169010818,-0.2026635408,0.4901743233,-0.5520546436,0.2140350342,-0.3513668776,-0.0185123459,-0.3378240168,-0.2004266232,0.1204777062,0.1363215894,0.072450608,0.2074892521,-0.4844915569,0.2136542946,-0.0438520163,-0.0662566945,0.1448486447,-0.3037320375,0.0800590366,0.0359201841,-0.222435236,0.1497515589,0.0071135149,0.0018467073,-0.090789564,0.2409534156,-0.2458492517,-0.1143798083,-0.3347643018,-0.0400816351,0.0403618924,-0.8884328008,0.3341011703,0.6728857756,0.6103256941,-0.0293981545,-0.1831840426,-0.075850226,-0.1074196175,-0.1476211697,-0.2166453749,0.0443934202,0.2339739949,0.0690468699,0.0551419891,0.2488982826,0.2439636886,-0.0811582357,-0.019442711,0.0190799683,0.0443793163,-0.4064261615,-0.5095794201,0.0706914291,0.2918235958,0.042824883,0.0334734432,0.1449056268,-0.169449687,0.1561665535,-0.0355580933,-0.1128104627,0.08772479,0.2027309686,-0.0460673124,-0.2352349609,0.2552710176,0.4344182909,-0.2322795242,-0.0719461218,-0.0610614866,0.6061143279,-0.138958469,0.0365319625,-0.1724819094,0.1928213537,-0.1596536636,0.2387522757,-0.1975675225,0.0277715065,0.00536869,-0.0947651565,-0.2900633812,-0.1597348154,-0.269921869,0.3686087132,0.1735462248,0.1552822441,0.0558239259,-0.2245473713,-0.1342835128,0.0632727593,0.0518477261,0.2201613635,0.1218952537,0.2995178103,-0.0751096308,-0.262899369,-0.0004000829,0.1126528904,0.0660861805,-0.1165143028,-0.156600073,0.1453291327,0.0634857193,0.1447235346,0.1579525918,0.0045555336,-0.2916793525,-0.32062307,0.2610510588,0.5018334389,-0.0261957776,0.5026631951,0.0075254901,0.0791921094,-0.0773143768,-0.096601963,0.2939926088,0.000271245,0.0600576662,0.0017516485,0.0430061519,-0.0644944087,-0.297593981,0.2500182092,0.2100893706,-0.2192618102,0.0757016093,0.0438815281,0.2049108446,-0.0795012936,0.1120113209,-0.0127637144,-0.0753770769,0.3390367925,0.2564381659,0.0205129609,0.0415288061,0.1302811801,0.034827102,-0.0434780642,-0.1071994379,0.0841318741,0.1625445336,-0.0236622076,-0.19653216,-0.2139080912,0.0881646276,-0.3737869263,0.2155651599,0.0374126397,0.2139494568,0.0640223622,0.2510870993,0.0048766243,-0.0291462112,-0.1208115816,0.1924336106,0.4192002416,0.3636755943,-0.0985710397,0.2845052183,-0.1430582702,-0.2455260009,-0.3108027577,0.1504472643,0.2094430774,-0.0165157095,0.3741379976,-0.0452180617,-0.4020013809,-0.1395904869,0.1565153152,-0.3509288728,0.0147745982,-0.1717998236,0.129199326,0.3239371181,0.0625625849,-0.0667169467,-0.4656114578,0.2416135967,0.1107540652,0.0372458138,0.278429091,0.1098633632,0.2368880659,0.0604932867,0.0329405516,-0.0933829471,-0.2792635262,-0.0987304375,-0.0773724541,0.2119199038,0.3134056628,0.1710140407,0.1951716244,0.1725251228,-0.1121418923,-0.0668845698,0.3330702186,0.2009197325,0.0205651075,-0.0387231261,0.0629111156,0.0913388208,0.0867865533,-0.0760749355,-0.0480678193,0.2358455658,-0.2489839047,-0.2699623406,-0.3898544312,-0.3758755922,-0.380815357,-0.0015238931,-0.1787271798,0.1003260687,0.365806967,0.0336299799,-0.1460151523,0.0823104531,-0.0213113166,0.0583301,0.3585035205,-0.0015690379,0.3383734822,-0.0702756643,-0.292306602,-0.1447414905,0.1255771965,0.2702936232,0.1637944132,-0.1062755808,-0.0835557804,0.1526221931,0.098030746,-0.1812796593,-0.207480222,0.1495788097,0.4486217499,-0.0912011042,-0.0646641031,-0.4672204554,0.0905518234,0.0718428269,0.0319588482,0.0767272636,-0.0218784027,-0.1141312569,-0.250600189,0.5121714473,-0.1104845256,0.3139121532,0.1758183092,-0.0079048201,0.3140559196,0.1937715262,0.2147172242,-0.0985924974,0.1259611398,-0.0971218497,0.3895472884,-0.1474761218,0.0831446201,-0.1647632271,-0.3862561882,-0.18342489,0.529003799,0.1201298162,-0.1372963935,-0.1372113526,0.2273372263,-0.1324647814,-0.2655061185,0.6308995485,0.3584332764,0.0822625831,0.1822302788,-0.3089642823,-0.0003804027,0.3028071821,-0.2566522658,-0.1460650563,0.1207099184,-0.2959152162,0.4274281859,0.096721977,-1.0665234327,-0.0314471237,-0.0032904292,-0.3078857064,-0.1359232664,0.248935461,0.0394866057,0.1263052374,-0.1835102439,0.3833095133,0.1133232489,0.0715671554,-0.491666317,-0.3911873102]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1762","title":"Unable to format dataset to CUDA Tensors","comments":"Sure ! Feel free to open a PR to improve the documentation :) ","body":"Hi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan","comment_length":13,"text":"Unable to format dataset to CUDA Tensors\nHi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan\nSure ! Feel free to open a PR to improve the documentation :) ","embeddings":[-0.2288715243,-0.4802087247,-0.0844359696,0.1721917093,0.5675295591,0.3805964291,0.5245794654,0.3204935193,0.003594059,0.147289589,-0.1087801307,0.2554025948,-0.2011275589,0.1395729184,0.1657762676,-0.3607392013,0.2556073964,-0.1187702045,-0.0838356838,0.0844582319,-0.1578695029,0.0023357901,-0.1268036216,-0.2954472899,-0.2199977934,-0.1965823025,0.0365882739,-0.3819654584,-0.1764319092,-0.0776844621,0.1973728687,0.0829123631,0.285859108,0.4521905482,-0.0001260059,-0.0448658019,-0.0646484345,-0.1125686765,-0.0739720687,-0.3027530015,0.1040062979,-0.2533049881,0.3640278876,-0.1920741796,-0.4160137773,0.1546325088,0.0679140612,-0.3342970312,0.2437143773,0.4998241365,0.1042142957,0.1926234365,0.2438696027,0.0121915843,-0.1476804167,0.4034995437,-0.3707418144,0.0349337682,0.1954595149,0.1534217894,0.2251276672,0.1764603406,-0.1011284962,0.0375858396,0.301676929,-0.0235859677,-0.3981577158,-0.3445564806,0.0277977418,0.1560963392,0.731861949,-0.3105591834,-0.319675535,0.1307534426,0.082830742,-0.373901099,-0.2377060503,0.6231278777,-0.3422445953,0.0951245055,-0.3435271382,-0.0033716718,-0.2236743122,0.1041303426,-0.1996159106,-0.0001410445,-0.2178576142,0.013069516,0.0949735716,0.1669413894,0.3296225369,0.2150331289,0.1877894253,0.154527992,-0.4967325628,-0.0993126705,-0.5587373376,-0.1489122063,0.1730109155,0.0950459465,0.1955509782,0.0085788425,-0.1246602908,0.0222840812,-0.2437526882,0.2601795197,0.0801400244,0.1986269355,-0.0341226421,-0.2200173289,0.3203794956,-0.0650249794,-0.3352735043,-0.5642104149,-0.3135152161,0.2495964617,0.0512503013,-0.106578514,-0.302306056,-0.0730680302,-0.043126639,0.1408097297,-0.1885933876,0.3293704093,0.0385617577,0.1043371931,0.4375833869,0.1968247443,0.0357058831,0.0276488438,0.0060686679,0.1532099694,-0.2927617431,-0.0971303284,0.0368484706,-0.3716046214,-0.0427599438,0.0812155977,-0.0180780664,-0.0085331257,-0.1103718802,0.122493051,0.5160408616,0.158349067,-0.035537865,0.3239786923,0.2154832333,0.2569059134,-0.208425805,0.1437076479,-0.4595963359,-0.1416771561,-0.5339333415,-0.0296210311,0.2446335703,0.0267624035,-0.1960794628,0.2254499197,0.222346887,0.1542760283,0.1117447317,-0.0938683003,-0.1417992413,-0.3187061548,0.5051116347,0.1350945979,-0.3392821848,0.0450046472,0.4575935304,-0.0659089833,0.2130127102,0.3107723296,0.1806404889,-0.0668025017,-0.1093781888,-0.3113625646,0.4436480403,-0.3786649108,-0.0132012982,-0.1796414703,0.1334445477,-0.3039022088,-0.1036363319,0.1409557164,0.2722385824,0.0988592133,0.1322395355,0.191530332,-0.0086429138,-0.0144088129,-0.011755228,0.003429472,0.4920734763,-0.0853182152,-0.2033623606,0.3254646659,-0.2161832154,-0.0831352919,0.1789198816,-0.0849051923,-0.035846252,0.1015492976,0.269249171,0.0358976424,-0.158461839,0.0637344271,-0.4423708618,-0.114528127,0.4026538432,0.2028377354,0.0509695783,-0.1455996782,-0.0156371091,0.0119439429,-0.0197966844,-0.0706403032,-0.0373642035,-0.3497921824,-0.078933619,-0.1461343616,-0.2886095941,0.2604892254,-0.3170469403,0.1917237192,-0.3252305388,0.010373543,0.0497323014,-0.0802721605,0.0636571273,0.2137546986,0.0271192826,-0.2774088681,0.0688324124,0.127556026,0.1680942029,-0.0766832829,-0.3814060688,0.236418426,0.2375000268,-0.078528583,-0.1187825799,0.0508392192,-0.044064112,-0.1358036548,-0.4227778614,0.3392410278,-0.0185545832,0.1712910086,-0.4011728764,0.3845160306,0.163198784,0.1035624221,-0.0725907981,0.0282325707,0.019993661,0.1763267666,0.1297071129,-0.0772219673,-0.8642541766,-0.0446533151,0.0173683893,-0.0145036364,0.0942104235,0.0937300697,-0.1138121635,0.1540994942,0.2809925675,-0.3742114305,-0.1248788014,0.0600883961,0.1915304214,0.0558067001,0.0785323381,-0.1725849807,-0.0019194552,-0.1405514628,-0.0542193986,-0.1218241677,0.1021987274,-0.0282792393,0.0298904199,0.2870012522,-0.0230940375,0.1318757385,-0.506714344,-0.0598632954,-0.3499054909,0.0745012015,-0.3779944181,-0.3327980936,0.0056978236,0.0002319866,-0.1474156678,0.1390368789,0.3259263635,0.2553574145,0.2740023732,0.178290993,0.2093762457,-0.0894555375,-0.1511279792,-0.2331205606,0.1793906689,0.0096819932,0.3343394697,-0.1336868703,0.1831010878,-0.3087966442,0.1678065658,-0.358128041,0.0397036746,-0.0962247998,-0.1092099622,-0.1100610942,0.125346154,0.4134386778,-0.0267386474,0.0162605364,0.1130377278,0.1552252769,-0.0322608091,0.0142386882,-0.3431274593,-0.0585855022,-0.2182042748,-0.2698281109,-0.2100970447,-0.2586115003,0.2549442053,-0.0434230492,0.1756465137,-0.3248331249,0.3785178661,-0.1216871217,-0.0069342372,-0.0233828668,0.0498085245,-0.5527608991,0.608029604,-0.2885236144,-0.5282217264,0.1869076788,0.2099668086,0.3856399357,0.4054643214,-0.0048607918,0.1723938584,0.1806278378,-0.0821590126,0.0190059636,0.0695919171,0.2991285622,-0.026162168,0.1475078911,0.1193841919,-0.1277427673,-0.0288999714,0.1160598993,0.3303607404,0.1306238621,0.4815924764,0.069869101,0.6686831117,0.1180547252,-0.5832336545,0.0697396994,-0.2470056713,0.3555250764,-0.1401072145,-0.2779200077,0.2883781493,-0.0300972518,0.137008518,0.0144983279,-0.1160594597,0.0389495008,-0.1474920213,0.1987340897,-0.3409946561,-0.1107419357,0.2138722092,-0.3422921002,0.2099233717,-0.3356548548,0.2881712019,-0.0529127978,-0.1520639211,0.2071471661,0.37130934,0.2994513512,0.1181944534,-0.1029910073,0.0554410145,-0.6865428686,0.2591077685,-0.1716932505,0.387001425,-0.1518035978,0.2179108858,0.1623198986,0.2451640368,0.7754688263,0.1713530123,-0.5120682716,-0.0586222224,-0.0620195679,-0.4560672343,-0.1232617125,0.0506965481,0.1833432764,-0.1068986803,0.1365982443,-0.1476931423,-0.2700020671,-0.0232487414,0.2453421205,-0.2488987595,0.0450332053,-0.1186777204,-0.0803631172,-0.1111255065,0.1194067672,0.2976031899,0.1594004333,0.0869313329,-0.2699478269,-0.2166794688,0.1569167972,0.4386842847,-0.1229710206,0.1967141628,0.2452185005,-0.1684728414,0.154166162,0.2236249447,0.2823885977,-0.0683295354,-0.2002563626,-0.1000171229,0.046857357,-0.0887464285,0.4912897348,-0.0467777736,0.0702937245,-0.124567695,-0.2337275594,-0.0457159653,-0.309718281,0.2173767388,0.2414478064,0.3156200647,-0.7100848556,-0.4672840238,0.2528426349,0.1454380453,0.1961791664,0.5986383557,0.3061929047,-0.0424038135,0.1689738929,0.5314579606,0.7695659399,-0.4810445607,0.0892674327,0.1500231773,-0.2053514719,0.4758703411,-0.5433450341,0.255282104,-0.3517926633,-0.0541873686,-0.3534318209,-0.2013027519,0.1299349815,0.1461840272,0.0816704929,0.2066008598,-0.4868344665,0.1845462322,-0.0564268343,-0.0848696008,0.1565693617,-0.2866553962,0.0896237046,0.0336609781,-0.2479168922,0.1473848671,0.017827712,-0.0032141362,-0.0722338334,0.2480958253,-0.2485729456,-0.1123464406,-0.3396209776,-0.0190128032,0.0556374788,-0.8800778985,0.3488007784,0.6586976051,0.613142848,-0.0381848216,-0.1379984468,-0.0718865693,-0.1027464196,-0.1328598559,-0.2179480642,0.0351581089,0.2362666577,0.0694005638,0.0338397883,0.2625417113,0.2630541027,-0.0861025527,-0.0006579669,0.0404890552,0.0334314033,-0.3790766299,-0.5223232508,0.0647535026,0.3002345264,0.0347869694,0.0276760105,0.1358090788,-0.1819043308,0.1699932665,-0.0693064854,-0.1406235546,0.0930179805,0.191881761,-0.0405231714,-0.231095776,0.2622154951,0.4611985981,-0.2192883641,-0.0603858419,-0.0428490601,0.6009038091,-0.1553411186,0.0293174721,-0.1939466447,0.2130491734,-0.160150215,0.2577592731,-0.2058668286,0.0571036749,-0.0150284115,-0.1102908701,-0.299826473,-0.1650536209,-0.249616757,0.3645400107,0.1786942184,0.1433877647,0.0817774311,-0.1894597113,-0.1295279711,0.0603471287,0.033927314,0.2250240296,0.127394408,0.2893979549,-0.0653692484,-0.2565428019,-0.0078776246,0.1212037429,0.039741274,-0.1074400619,-0.1547309607,0.1498655826,0.0438249484,0.1307385415,0.1386780292,-0.0158683043,-0.2822112441,-0.304160744,0.2611734569,0.5082019567,-0.0397356115,0.5142461061,-0.0103787296,0.0660090744,-0.0980998948,-0.1142917424,0.3090555072,0.0081296517,0.0793780759,0.002305679,0.0604133867,-0.0788879246,-0.3261667192,0.2492771298,0.1950280219,-0.2433332354,0.0673770905,0.0507590622,0.1912713945,-0.0865079761,0.108710438,-0.0325506181,-0.0571709312,0.3080493212,0.2472234815,0.0172571167,0.0245551988,0.1252396107,0.0352210402,-0.0128042987,-0.1228445321,0.0816163719,0.1710969657,-0.0127721801,-0.2103459388,-0.2040442675,0.1053732708,-0.3663757443,0.2266833484,0.0343406312,0.2284637839,0.0406926125,0.2507325113,0.0212162156,-0.0256423522,-0.1127337739,0.2017885298,0.4175475836,0.3744438887,-0.1063304693,0.2815154791,-0.1458006948,-0.2503876388,-0.2792277038,0.124622941,0.1981898993,-0.0165696219,0.4093690813,-0.0571329072,-0.4099903703,-0.1346208453,0.1890886575,-0.3417005539,0.0056615667,-0.1806069314,0.1382787824,0.2773360908,0.0771563277,-0.0429011248,-0.4543261826,0.2789701819,0.1134296656,0.0232971292,0.2682377994,0.0861109421,0.2667704225,0.0535436943,0.0294160228,-0.0853847414,-0.2236506641,-0.0863118097,-0.0708210394,0.2133782655,0.3165902197,0.1659243107,0.1741801202,0.1982381642,-0.1099406257,-0.0684124976,0.3535695076,0.1852478236,0.036763072,-0.0208836552,0.0654197261,0.0855360925,0.1083986834,-0.059804406,-0.0443224609,0.2770190835,-0.2520701289,-0.2388517261,-0.4032147229,-0.3666342497,-0.3939366341,-0.0227339752,-0.1990679651,0.0893132463,0.3723177314,0.0265237633,-0.149915576,0.0841838643,-0.0254461989,0.0385239683,0.3687800169,0.0069995429,0.3398016989,-0.0927751064,-0.326112777,-0.1338278651,0.1365613192,0.2683086991,0.1732540131,-0.107614927,-0.082139425,0.1330524832,0.0756585747,-0.1730948091,-0.1880433261,0.1464419216,0.454346478,-0.0901201144,-0.04506341,-0.4783560336,0.0925782844,0.0509988554,0.0609681457,0.0826300383,-0.0113423532,-0.1199315339,-0.2422452271,0.4829610884,-0.087214388,0.3065360785,0.176562652,-0.0165961925,0.301422894,0.2011658996,0.2600897849,-0.0683838576,0.1287356168,-0.0718158334,0.3784384131,-0.1279515624,0.1027736887,-0.1572600007,-0.3924399912,-0.1826824099,0.5346992016,0.0955273882,-0.1711713374,-0.1418129802,0.2095173597,-0.1723583639,-0.261582911,0.6477419138,0.3629731536,0.0919918045,0.1663662046,-0.315946579,0.0168237798,0.2798039913,-0.2589093149,-0.1701331437,0.1113684922,-0.3115831912,0.4258958399,0.1228700876,-1.0639640093,-0.0381746627,-0.0060496638,-0.3022233248,-0.1268765181,0.2341136485,0.0402341746,0.1162606627,-0.1751023531,0.3711378276,0.1280292869,0.0534102134,-0.4879975617,-0.3834216893]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1762","title":"Unable to format dataset to CUDA Tensors","comments":"Closing this issue as it has been resolved.","body":"Hi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan","comment_length":8,"text":"Unable to format dataset to CUDA Tensors\nHi,\r\n\r\nI came across this [link](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html) where the docs show show to convert a dataset to a particular format. I see that there is an option to convert it to tensors, but I don't see any option to convert it to CUDA tensors.\r\n\r\nI tried this, but Dataset doesn't support assignment:\r\n```\r\n columns=['input_ids', 'token_type_ids', 'attention_mask', 'start_positions','end_positions']\r\n\r\n samples.set_format(type='torch', columns = columns)\r\n for column in columns:\r\n samples[column].to(torch.device(self.config.device))\r\n```\r\nThere should be an option to do so, or if there is already a way to do this, please let me know.\r\n\r\nThanks,\r\nGunjan\nClosing this issue as it has been resolved.","embeddings":[-0.2375893891,-0.4483739436,-0.0885980576,0.1816409826,0.5787226558,0.3644931614,0.5476467609,0.3623672724,0.0102678742,0.134642005,-0.1012924835,0.2628093064,-0.1983363777,0.1372024715,0.1499369591,-0.3600692451,0.256236732,-0.1195168197,-0.0809422284,0.0762076899,-0.1736076474,0.0028658097,-0.1414910257,-0.3121667504,-0.2349396348,-0.1797040403,0.0392391421,-0.3853328824,-0.1568331569,-0.0758314207,0.2119638771,0.0630116686,0.292247355,0.4778642058,-0.0001242314,-0.0291838348,-0.0823853537,-0.1070434898,-0.0898339674,-0.3069269359,0.0846840665,-0.2492355555,0.3412580192,-0.2101157904,-0.4239158332,0.1496180445,0.0416121371,-0.352959156,0.2465307266,0.5070274472,0.123470448,0.1822905689,0.2446448058,0.0021269165,-0.1648390144,0.3884168863,-0.3797696233,0.0132393055,0.1874618828,0.1441068947,0.2541210353,0.1794229299,-0.1040809676,0.0279992744,0.2855726779,-0.0483351499,-0.3815896511,-0.333732456,0.0501029752,0.1318610907,0.7174282074,-0.3106859028,-0.317792356,0.1412312537,0.0916379467,-0.360042572,-0.2155566663,0.6271401644,-0.3200551867,0.0983200446,-0.3294781744,0.0077016763,-0.2229304761,0.1230573356,-0.2153621465,-0.0209520999,-0.2209809124,0.0170583483,0.0449809246,0.1555359364,0.3350451291,0.1921585202,0.1761767566,0.1368158758,-0.5329499245,-0.0824100226,-0.5532200336,-0.1736866683,0.1560022831,0.0995607302,0.1890237927,0.0270055551,-0.1084626764,0.0050919331,-0.2400688976,0.2790406942,0.0856925175,0.2152073383,-0.0607450008,-0.2133326828,0.2881924808,-0.0655160993,-0.3206279278,-0.588087678,-0.2924034595,0.2624446452,0.0477402695,-0.1191018149,-0.297803998,-0.0489501618,-0.0474618115,0.1687667221,-0.1817503124,0.3322706521,0.0334163569,0.132657975,0.4480935335,0.2234765887,0.0530755781,0.0581414253,-0.0032714494,0.1463077813,-0.3063915968,-0.1368328482,0.0657398924,-0.3601771891,-0.0389277078,0.0815239772,-0.0411760844,-0.0282433201,-0.1065335348,0.1002707332,0.541192174,0.1534416527,-0.0291934088,0.3263790607,0.207870543,0.2458609194,-0.1678130329,0.1443224847,-0.4603359699,-0.1247509122,-0.5376306176,-0.0103090713,0.2468352765,0.0277045295,-0.2038491964,0.1990199238,0.2299572974,0.160319522,0.0883871391,-0.1168586984,-0.143741399,-0.3160966635,0.4875549376,0.1199611872,-0.3395973742,0.0499131605,0.4342158437,-0.0600435473,0.2318320274,0.3130021095,0.1959756166,-0.0824262649,-0.1056741327,-0.3150985837,0.4431183338,-0.3783392012,-0.0513606444,-0.1604445875,0.1272756457,-0.30766505,-0.0978430659,0.1213182434,0.2919287086,0.0892929137,0.1620818526,0.2030938715,0.0049770307,0.0002666751,-0.0458814763,0.0009682931,0.4882958531,-0.0590909682,-0.15539141,0.321320802,-0.2302984893,-0.0869943798,0.1591812968,-0.079582721,-0.0242635589,0.0976813063,0.2957291007,0.0137696732,-0.148352921,0.0542566888,-0.4426160753,-0.1016912013,0.3751659393,0.1915365905,0.0371990651,-0.1123904437,-0.0390572473,0.0120633962,-0.0391848087,-0.0413975231,-0.0150508415,-0.3422026336,-0.0895865336,-0.1347229332,-0.2688769996,0.2529266179,-0.3014670014,0.1745847166,-0.2956706583,0.0623148158,0.0357746184,-0.102827698,0.090200454,0.2148805559,0.0454323329,-0.2794165313,0.0563692227,0.1190419942,0.1546673626,-0.0945865661,-0.3863792419,0.2032359987,0.2227258235,-0.0836639255,-0.1236764118,0.0459192432,-0.0294678323,-0.148808226,-0.3925617933,0.3377266228,-0.0036144834,0.1691078991,-0.4055010676,0.3964381516,0.1860992908,0.0735486299,-0.05387206,0.01925743,0.0485874228,0.1952502728,0.1215644255,-0.0902605355,-0.8605141044,-0.0427471995,0.024615813,0.0300191306,0.0934330374,0.0746768564,-0.119470872,0.1554549187,0.2725260556,-0.3638429344,-0.1309846044,0.0631689057,0.1592711806,0.0492728241,0.0871154144,-0.1787365079,-0.0160161685,-0.1384293288,-0.0337781459,-0.0878509507,0.1451888084,-0.0314324722,0.0375834033,0.3031577766,0.0063749393,0.1452458203,-0.5007763505,-0.0597175919,-0.3759431541,0.087236613,-0.3742347956,-0.3352739215,0.012929542,0.007675082,-0.1507548243,0.1497073621,0.3394322097,0.278368175,0.2974734008,0.1554911435,0.2045487314,-0.0919997394,-0.1737803519,-0.2261909544,0.1744595021,0.0316806473,0.3647947907,-0.1365307271,0.1995019466,-0.2885458171,0.1725625843,-0.3671796918,0.0555090867,-0.126253888,-0.0989760682,-0.093025066,0.0974017382,0.4034965634,-0.0142400526,0.0245523751,0.1227749586,0.1622834951,-0.0030686795,0.0492334887,-0.3506352901,-0.078372851,-0.2145982534,-0.2646438479,-0.2129010707,-0.2526014149,0.2300569266,-0.0351566635,0.1591614783,-0.3411643803,0.3767536581,-0.1427225918,0.0160024073,-0.0517533086,0.0464953668,-0.5363053679,0.6017537117,-0.2873581648,-0.5250380635,0.1919486076,0.2063644677,0.3911952376,0.4117250741,-0.0075867306,0.1890725344,0.1789425313,-0.0797186941,0.0049560256,0.031494841,0.3122873902,-0.0058111977,0.1348391175,0.1165259257,-0.1225534379,-0.0143829882,0.1269454658,0.3356842995,0.1094821021,0.4952766001,0.0924998596,0.6629980206,0.1178714707,-0.5851995945,0.0810188502,-0.2373187244,0.3131901622,-0.1511008292,-0.3076869845,0.3003976941,-0.0449149869,0.1403128505,0.0108647496,-0.136392042,0.0026270584,-0.1701179743,0.2059934884,-0.340711385,-0.1109894663,0.2166767865,-0.3350530863,0.2207233459,-0.3075783253,0.299764663,-0.0588003322,-0.1559160948,0.2156946659,0.3658542931,0.2736421227,0.1019290313,-0.113046214,0.0557663254,-0.6825041771,0.2692524195,-0.1834531873,0.4328612387,-0.1640120447,0.2414411306,0.1485601068,0.2397718281,0.7876378298,0.1561892331,-0.4851257503,-0.0159995165,-0.0566351786,-0.4586112797,-0.150565654,0.0608015954,0.185745731,-0.0940818787,0.139406085,-0.1436646432,-0.2510990202,-0.0355194509,0.2559412718,-0.2553305328,0.0180301871,-0.1150585935,-0.0665665045,-0.1245570555,0.1017233059,0.2766370177,0.173524037,0.0796038359,-0.276149869,-0.2419409454,0.1362827122,0.4326419234,-0.1267287731,0.20213826,0.2505373955,-0.1256114691,0.110398531,0.2172439992,0.3173766434,-0.0697826892,-0.1862637252,-0.0951294005,0.0489434302,-0.1158978716,0.47286129,-0.0532700159,0.0291050021,-0.1227480471,-0.2426380962,-0.0459114537,-0.2882446647,0.2011073679,0.2495508045,0.3081305921,-0.7178777456,-0.4591623247,0.2174122781,0.1149548739,0.2155702114,0.6018154025,0.2733131349,-0.0263794307,0.201996401,0.505268693,0.7581421137,-0.4771946669,0.0524910279,0.1630739868,-0.1922908276,0.4470733404,-0.5176221728,0.2740055323,-0.3725457489,-0.0576100275,-0.3400777578,-0.2029639184,0.1413699985,0.1534932703,0.070329234,0.1955522001,-0.4875113368,0.1673123538,-0.0594812967,-0.0770410821,0.1761496812,-0.2883060575,0.0682273358,0.0529891402,-0.2719006538,0.1375550777,0.0161941852,-0.0190454628,-0.0833563656,0.2634969056,-0.2370057404,-0.0786042958,-0.3273302913,0.0064754314,0.0514007695,-0.8868618608,0.3686649203,0.6418622136,0.6291354299,-0.0260357372,-0.1373171061,-0.0561328903,-0.0863459781,-0.1268831491,-0.1964540631,0.0283581056,0.2599595785,0.0478932858,0.0176009387,0.2584154606,0.2780303657,-0.0814436376,0.0266748071,0.0296181496,0.0443567224,-0.3882834613,-0.5376062393,0.0542694852,0.282204777,0.0149420183,0.046565,0.13329494,-0.1625612974,0.1859102696,-0.0713918954,-0.1391692311,0.1039612293,0.2065836936,-0.0537579469,-0.2128810436,0.301009208,0.4811123013,-0.2195220143,-0.0777235329,-0.0601975992,0.5427533984,-0.1569192708,0.0494201034,-0.2090230584,0.2305812687,-0.1479835361,0.24875018,-0.2153117359,0.0468847491,0.008039197,-0.1249898523,-0.2963287532,-0.1958699077,-0.2590619326,0.3440905213,0.1849785,0.1613717079,0.0727085173,-0.1899766028,-0.1517030597,0.0485554598,0.0032726584,0.2275100648,0.0748395175,0.2518543005,-0.0831569657,-0.2690027356,0.0142565416,0.1184042469,0.0350177288,-0.1315378994,-0.1440579146,0.1410901994,0.0573416315,0.1338343471,0.1377143413,-0.0405497067,-0.3026919961,-0.2950114608,0.2386890799,0.5016495585,-0.0607511252,0.5176263452,-0.0196407884,0.0493889861,-0.0988557115,-0.0999217331,0.2922369242,-0.019440623,0.0709382445,0.0138830291,0.0798229277,-0.0785250664,-0.3572257161,0.2337915748,0.1554924548,-0.2210101485,0.0871996358,0.0056669731,0.2124741524,-0.0555118322,0.1025184616,-0.0295097604,-0.0638903901,0.3100940287,0.233479172,0.0419685394,0.0179990791,0.144989118,0.0548322573,0.0011754726,-0.1085845977,0.0722613707,0.1959192604,-0.0343017988,-0.2097225636,-0.1811386943,0.1074631736,-0.3802779317,0.2394264191,0.0258419011,0.2563048601,0.0508138463,0.2312509269,0.0136100613,-0.0414777584,-0.1009409875,0.2242845595,0.4040648341,0.3835083842,-0.1225366667,0.2586270571,-0.1762560159,-0.2399117798,-0.3014161289,0.0899583772,0.202880308,-0.0064896243,0.3980153799,-0.0716086775,-0.4104303718,-0.1430524588,0.1931355596,-0.3549000323,0.0339183062,-0.189240858,0.1357515454,0.280156523,0.0751614422,-0.0530047938,-0.4034483731,0.2476573735,0.1323425472,-0.0121737979,0.2571706772,0.067605719,0.2764949799,0.0597573258,0.0106140571,-0.0936482325,-0.2267163694,-0.0899447352,-0.070154056,0.2053866833,0.3263841867,0.1711070538,0.167917639,0.1936169118,-0.1055845693,-0.0797946304,0.3474060297,0.2118050009,0.0486787111,0.0001270126,0.0797135606,0.0990158767,0.0926653519,-0.0720631331,-0.0324623846,0.2917531431,-0.2435730249,-0.2605279386,-0.4191078544,-0.3687749505,-0.4071875513,-0.0402488336,-0.1955939829,0.0866985694,0.3754505515,0.0189767405,-0.1343004107,0.0667638332,-0.0149110984,0.0146407979,0.3751571178,-0.0069766212,0.3218850493,-0.1089349538,-0.3320273757,-0.1215645298,0.1427429616,0.2498386055,0.1965273619,-0.0923735797,-0.0875530913,0.1394822299,0.0414763801,-0.1793994009,-0.1802498102,0.1573174894,0.4731461406,-0.1109294444,-0.026715301,-0.4576218724,0.0653583631,0.0597301312,0.0375643075,0.0905910656,-0.0084432177,-0.0993150473,-0.2478233278,0.4620330036,-0.0962469131,0.2984600663,0.1958611161,-0.0183874313,0.3062129617,0.2034220397,0.2659821212,-0.0840001479,0.1039821655,-0.0563907437,0.3841333687,-0.1116324589,0.1404178441,-0.1217777878,-0.3792555928,-0.1891881973,0.5256233811,0.0874458328,-0.1486507356,-0.1475752145,0.2207490355,-0.1605129838,-0.2437911779,0.6473529935,0.3609123826,0.0806836486,0.1681814492,-0.3382867277,0.0075387633,0.2723206878,-0.2402586043,-0.2146826386,0.1032225043,-0.3078205585,0.3820512593,0.1378306448,-1.0423010588,-0.0538462773,0.0063442495,-0.3110330701,-0.1178985611,0.2328580171,0.0295657683,0.1381897479,-0.1682040691,0.3512328565,0.1171904728,0.062326178,-0.4857842028,-0.3697026968]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1759","title":"wikipedia dataset incomplete","comments":"Hi !\r\nFrom what pickle file fo you get this ?\r\nI guess you mean the dataset loaded using `load_dataset` ?","body":"Hey guys,\r\n\r\nI am using the https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wikipedia dataset.\r\nUnfortunately, I found out that there is an incompleteness for the German dataset.\r\n For reasons unknown to me, the number of inhabitants has been removed from many pages:\r\nThorey-sur-Ouche has 128 inhabitants according to the webpage (https:\/\/de.wikipedia.org\/wiki\/Thorey-sur-Ouche).\r\nThe pickle file however shows: franz\u00f6sische Gemeinde mit Einwohnern (Stand).\r\n Is it possible to fix this?\r\n\r\nBest regards \r\nChris\r\n","comment_length":21,"text":"wikipedia dataset incomplete\nHey guys,\r\n\r\nI am using the https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wikipedia dataset.\r\nUnfortunately, I found out that there is an incompleteness for the German dataset.\r\n For reasons unknown to me, the number of inhabitants has been removed from many pages:\r\nThorey-sur-Ouche has 128 inhabitants according to the webpage (https:\/\/de.wikipedia.org\/wiki\/Thorey-sur-Ouche).\r\nThe pickle file however shows: franz\u00f6sische Gemeinde mit Einwohnern (Stand).\r\n Is it possible to fix this?\r\n\r\nBest regards \r\nChris\r\n\nHi !\r\nFrom what pickle file fo you get this ?\r\nI guess you mean the dataset loaded using `load_dataset` ?","embeddings":[0.1241123378,0.0906564444,-0.0648524389,0.4390091002,-0.1451850832,0.1351810396,0.1400446743,-0.0993287042,0.3315919638,0.1742715687,0.1878574491,-0.0555863045,0.3833341599,-0.3714283407,0.0213763621,-0.2188196033,0.1357136071,0.1867661327,-0.353669852,-0.2992884517,-0.1205291748,0.4303146303,-0.2972927988,-0.3738275468,0.0041588228,0.19129242,0.0423987173,-0.0265655871,-0.0398979373,-0.2545049489,0.0718956441,-0.1289307326,0.0232157223,0.2944527268,-0.0001212063,-0.2076835334,0.1829906702,0.0327180289,-0.5496045947,-0.0704218149,-0.318674922,-0.3378354311,-0.0553493388,-0.3048529327,0.3356400728,0.1256951839,0.2353796661,0.0285403319,-0.0430291593,-0.1988056302,0.109937042,-0.3088945448,0.1718199402,0.0193357505,0.3870480955,0.5492720008,0.1299690902,0.418032676,-0.0086954348,-0.1311780214,0.0025488234,0.5580912828,0.1151680574,-0.0844878107,0.1515882313,-0.0853332505,0.0486637726,-0.3207388222,0.4474427402,0.4326772094,0.7783439755,0.0387785062,-0.2873822451,-0.2301689982,-0.025564,0.0607123598,0.4423192143,0.0954175889,-0.0011091381,0.0861819386,0.0174800642,-0.3801380992,0.0836794153,0.2865539491,-0.2406740189,0.6932728887,-0.0522080064,0.1574092954,-0.1112748161,0.0383665152,-0.5061631799,-0.1751054674,-0.1195133105,0.6439186335,-0.012000083,0.1531442702,0.1128866076,0.0546699315,0.3744640648,-0.1240142137,-0.3182374239,-0.1055599824,-0.4076496661,0.0868473873,0.4266243279,-0.0020243784,0.3739506006,-0.2473478317,0.2485174686,0.0315778702,0.0509780347,0.253721118,0.1115050092,-0.2443264574,-0.0742328838,-0.0230276305,0.0975892469,-0.2007698417,0.0104226209,0.4637053013,-0.2206739187,-0.2525127232,-0.185787797,0.0229531098,-0.1924729943,0.314414233,0.0461675338,-0.0150527935,-0.3274431229,-0.3958059847,0.0411142036,0.0460367687,-0.0717362761,-0.1020293087,-0.1323872358,-0.227997154,0.3639759719,0.0687114224,-0.0924394056,-0.1216051579,0.0250204597,-0.3007547855,0.0401272103,0.1244297475,0.2084525526,0.2604510486,-0.1159794927,-0.6150658131,0.0189969148,0.0594440587,-0.3399091065,-0.1113068163,-0.714305222,0.1314143091,-0.1911712885,-0.0315862261,-0.2961347997,0.1198506355,0.2013316005,-0.4084000289,0.1167671159,0.0557347946,-0.1401756108,-0.0329447761,0.3443155587,0.5783535838,-0.1809437871,-0.2524649501,-0.2087493092,0.0897938833,0.353200525,0.4228056967,-0.0118461531,0.3736517727,-0.1751294583,0.1606291533,0.0374035127,0.0762989819,-0.4493754506,0.0295907371,0.127416417,0.012836772,-0.3067496419,0.0343672782,0.0258893091,0.3042875528,0.0671055391,0.1469595134,0.3602668941,-0.0720262006,-0.3002051413,-0.1694299728,0.5182990432,-0.0393483154,0.1673227102,0.0278055016,0.2274997383,0.1395635605,0.6033773422,0.1070422903,0.3978745043,0.789696157,0.0262635704,0.2337703109,0.2118063271,-0.0106214238,-0.2573165596,-0.0086753704,-0.0164199304,0.3641389012,0.1543468535,-0.0990670845,-0.1662692875,-0.0136340894,0.0195926055,-0.1372879893,0.0141813159,0.1456440389,-0.0317268781,0.0626532882,0.1838206798,-0.3371160626,-0.254968375,-0.0783704221,-0.6629975438,0.3918353915,-0.1417583674,0.0992441401,-0.3523250222,0.1621439308,-0.0117894178,-0.142051518,-0.1318246126,0.0079831313,0.3144644499,-0.078461282,0.4909600019,0.0006882547,0.1212997511,-0.3434089422,-0.0658256337,0.1046491191,0.0298574697,0.0752246752,-0.6009052396,-0.1981950849,0.2485599816,0.226744771,-0.2269640267,0.2283167094,0.5967853069,0.1458377838,0.0836337805,-0.4401924014,0.38652426,0.2097622454,0.1051055044,0.069303669,-0.261615634,0.0292556901,-0.0195674468,0.1131461412,0.3585663438,0.3546279967,-0.29358688,-0.3039520681,0.2648385465,-0.2347448021,-0.166177392,0.1754751652,0.3176705539,-0.1919918358,0.3871274889,0.0223829336,0.1180270687,0.2003120035,0.1394442618,0.1772623956,0.1525942832,0.1820432395,-0.330262363,-0.1329103708,-0.071037814,0.3756258488,-0.0305539984,-0.0649510026,-0.0540692508,-0.0113120107,0.0340101533,0.0396412872,-0.8664521575,-0.1790496707,0.030365875,-0.0061965608,-0.1516415775,0.2164192051,-0.0783867091,-0.1252055615,-0.2134806961,0.1640907377,-0.1619288772,-0.1873651743,-0.2594039142,-0.0083929719,0.0020934211,-0.0685687736,0.1145582572,-0.467007041,-0.4988545179,-0.6564518809,-0.38599962,0.2722924054,0.000618317,0.0336333402,0.1947737634,0.4568445086,-0.42540887,0.1016558111,0.131040588,-0.0450766347,0.0273406692,0.0357402973,-0.2880314589,0.0095882779,-0.071097441,-0.1275347173,-0.23204723,0.0671186075,0.1465851963,0.3558940887,0.2403901964,0.3179811239,0.042869024,-0.2407635301,-0.0785083696,-0.1428986192,-0.373886168,-0.4261963964,0.0893008336,-0.2573982179,-0.2928767502,0.0356932953,-0.0971708,-0.0374456979,-0.115373455,-0.5714575648,0.0845571607,-0.1595594436,0.1899251789,-0.3796055019,0.1761974394,0.4080471694,-0.0137708457,0.1159530953,-0.0897347257,-0.1603077054,-0.0024278515,0.3095690608,0.6256231666,-0.1449142396,0.1137359738,0.0443239957,0.631234169,0.4565292895,0.4432752132,0.1186548099,0.0140662948,0.2230178714,-0.1847384274,-0.0118874479,0.1425248533,-0.0815662965,-0.0956149399,0.4853443503,0.1688071489,0.0298475958,-0.0498667806,-0.118943572,-0.5683521032,-0.1573637575,0.1096540689,0.2222973555,0.4297432601,0.07246916,0.0042687282,-0.4546712935,-0.3010433614,-0.0173498355,0.09974996,0.4043809772,0.2055342644,-0.255972743,-0.0552056804,-0.3435917497,0.2963940501,0.1967509687,-0.1161651239,-0.0431463718,0.2480863482,0.1920321286,0.0599629618,0.2783235312,-0.3235580921,0.3307548463,0.0630649552,0.3061447144,-0.1952378452,0.1243378073,-0.0076755444,0.0780840069,0.383510381,-0.1459000409,-0.3941358626,0.0068214675,0.1317758113,0.3869905174,-0.4159858823,-0.1628044695,-0.1502656937,-0.0955924764,-0.0743830577,-0.2400771827,0.1050749272,0.1407226622,0.4571057558,0.2124797404,0.2755673826,0.3150890172,-0.0278869346,0.1234167367,0.3730911613,0.5226802826,-0.1958729178,0.0950234085,0.1847807765,0.2306091785,0.6488135457,-0.0811292902,-0.2041320801,0.2236938924,-0.227650404,0.2851317525,0.048197329,0.1541280597,0.1433591098,0.314725101,0.0176212806,-0.2084538043,-0.001911256,0.0730575919,0.219182536,-0.1945897788,-0.2165082693,0.4007437825,-0.1713006198,-0.2189119309,0.1593727916,0.3846010864,-0.2809464931,0.2617731392,0.1700270772,1.0994677544,0.3411224484,0.096565567,-0.0495123044,0.0070651858,0.3117257655,-0.1347501278,0.1021702737,-0.3982689381,0.1721234769,0.0311942268,0.1471392512,-0.0875843912,0.1679605395,-0.0116095375,0.0379859433,-0.3027431071,0.1452645212,-0.1494725943,0.5097946525,0.2185716927,-0.0646300316,0.0943343341,0.0636624843,-0.0737537891,0.5269140601,-0.3921823204,0.1309323013,-0.0216824915,-0.0639726147,-0.462346077,-0.1099231467,0.1634226441,0.2898712754,0.0740599483,-0.3698223829,0.0629415959,0.3448385298,-0.0424821712,0.279594332,-0.1665314585,0.2678977847,-0.1225324944,0.0890750587,0.2536838055,0.256084919,-0.095719479,-0.1574682444,-0.2887415588,0.0607150868,-0.2990496457,-0.1858643144,-0.2878057659,-0.2203775644,-0.2034488767,-0.2111725956,0.2374621928,0.0578254312,0.0327009037,-0.2778077126,0.0405841246,0.3265536427,-0.0773851648,-0.0328522921,0.040015664,-0.2750094831,-0.0557426251,0.2483488321,0.0966591835,0.043319568,0.543592155,0.3383606672,-0.2557899058,-0.0880203173,0.0423483029,-0.2499470562,-0.3204173744,-0.1307281107,-0.1373304427,0.3106864393,-0.5650644302,0.0534807257,0.2527605593,-0.0783643797,-0.2076030225,-0.4911599159,-0.2866072059,0.1836094707,0.0466200113,-0.0691136867,0.0986911803,-0.3289675117,0.1177261546,0.1879459471,-0.2075761706,-0.2144394964,-0.1628053337,0.2485565096,0.0800275058,0.0706924722,-0.2035686076,-0.1712407768,0.0082639446,-0.4229908884,0.2090296447,-0.1092143953,0.0070721707,0.216400817,0.0040403912,-0.2575333118,0.1536416411,-0.1462586373,0.0072741592,-0.2359383255,-0.1799969524,0.0627090037,0.1607578397,0.3790535629,-0.0113615599,0.2351302952,-0.3436379433,0.2361392379,0.208863616,0.1010834575,0.0122874379,0.2078628093,0.4228385985,-0.0371350534,-0.0647029579,0.0304246638,-0.0875318572,-0.0714623928,0.0140394438,-0.2621015608,0.4172373414,0.4105350971,0.1569612473,0.019444162,-0.1367607713,0.1276397556,0.1322270483,0.1944712251,-0.5206779838,-0.0753168836,0.3623286188,0.1386783123,-0.0603638701,0.027306946,-0.0727728978,-0.0995892659,0.0091199419,0.0946719795,-0.0029428147,-0.0711898506,0.0443444476,0.2746727765,-0.0135109266,0.1159088761,-0.2780139744,0.2060650438,0.029139448,0.288610369,0.2312504202,-0.0053351391,-0.0740163103,-0.016195843,0.0289745629,-0.1542554498,0.220347777,0.214186579,-0.0943239108,0.1877345443,0.2088026553,-0.0468163751,0.1082425267,-0.0126323625,-0.3118581474,0.1510059237,-0.1738506854,-0.0970462784,-0.5045150518,-0.1496814042,0.1221644878,-0.021241961,-0.0276513752,-0.3495687544,0.0585049018,0.1448530555,-0.3765434623,-0.5085454583,0.0045265188,0.17144759,0.2790907025,-0.2493131608,0.0415736474,0.1924792081,-0.1547212601,-0.2463207692,0.5353747606,0.2218730301,-0.0933490545,-0.0847058147,-0.0881296769,0.0008303075,-0.0353853963,-0.2191946507,0.1410607547,0.0154904956,-0.2104367465,0.0953918472,0.0374166965,-0.1108055785,-0.0878579468,0.2928644717,0.1717206687,0.0481252819,-0.3466182649,-0.1448609084,-0.2381432652,-0.1054682955,-0.0719589666,0.0435567796,-0.061015252,0.3149779737,0.3373503089,0.0707162321,-0.1007737368,0.0237521864,-0.0149044171,0.1300119311,0.2392907888,0.0557317063,-0.0892165974,-0.4169371128,-0.4850449264,0.2806759477,-0.4443153441,-0.435885489,0.157036528,0.1774456203,0.0588555112,0.1095639244,0.1656245887,0.3773581684,-0.2004365921,0.5021958947,-0.1642214209,-0.069922097,0.0462781824,0.195947051,-0.1062896103,-0.2104584277,0.3261578977,0.2263188809,-0.1130203083,-0.2087387145,0.014973741,0.0139917685,0.3081994057,0.2971160412,0.2244362533,0.2409468293,-0.3251003921,-0.2193528563,-0.082947731,-0.2932718992,-0.0631550029,0.7308732867,0.0041922056,0.0557394996,-0.1740665436,-0.1332311928,0.0522397123,0.3003104627,0.1083897129,0.193181321,-0.5772894621,-0.0473172925,0.0324312486,0.1926531196,0.0499002971,-0.07506603,0.1249439493,-0.2492323071,-0.4154958129,-0.2398275286,0.023705354,-0.4101731479,-0.2318981588,-0.6320625544,0.0406712331,0.0872410238,-0.3288663924,-0.2046055943,-0.1694362611,0.414149642,0.1858736873,-0.028733369,-0.3691380322,-0.1478294432,0.049878709,-0.1312876642,0.3941732943,0.0958981737,-0.4242753983,0.0694527626,-0.2967579663]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1759","title":"wikipedia dataset incomplete","comments":"yes sorry, I used the `load_dataset`function and saved the data to a pickle file so I don't always have to reload it and are able to work offline. ","body":"Hey guys,\r\n\r\nI am using the https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wikipedia dataset.\r\nUnfortunately, I found out that there is an incompleteness for the German dataset.\r\n For reasons unknown to me, the number of inhabitants has been removed from many pages:\r\nThorey-sur-Ouche has 128 inhabitants according to the webpage (https:\/\/de.wikipedia.org\/wiki\/Thorey-sur-Ouche).\r\nThe pickle file however shows: franz\u00f6sische Gemeinde mit Einwohnern (Stand).\r\n Is it possible to fix this?\r\n\r\nBest regards \r\nChris\r\n","comment_length":28,"text":"wikipedia dataset incomplete\nHey guys,\r\n\r\nI am using the https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wikipedia dataset.\r\nUnfortunately, I found out that there is an incompleteness for the German dataset.\r\n For reasons unknown to me, the number of inhabitants has been removed from many pages:\r\nThorey-sur-Ouche has 128 inhabitants according to the webpage (https:\/\/de.wikipedia.org\/wiki\/Thorey-sur-Ouche).\r\nThe pickle file however shows: franz\u00f6sische Gemeinde mit Einwohnern (Stand).\r\n Is it possible to fix this?\r\n\r\nBest regards \r\nChris\r\n\nyes sorry, I used the `load_dataset`function and saved the data to a pickle file so I don't always have to reload it and are able to work offline. ","embeddings":[0.0643151104,0.0894950703,-0.0597212315,0.4419097006,-0.1323364824,0.1083998829,0.1577170193,-0.1221325919,0.2750071883,0.1589432061,0.1738340408,-0.1358434111,0.4038769603,-0.4070718586,0.0144945011,-0.2074819207,0.1505155265,0.152774632,-0.3166952133,-0.3021358252,-0.11335399,0.4350934923,-0.3123499155,-0.3774802685,0.0437858999,0.2163709998,0.0335686095,-0.0815825909,-0.0119257383,-0.2183104753,0.0533051379,-0.1486276239,0.0443650149,0.2822428346,-0.0001218747,-0.2125065178,0.1440680623,0.0306985285,-0.5327082872,-0.0756023824,-0.320605576,-0.3069421053,-0.0897938013,-0.3360282779,0.3274797201,0.1198892295,0.2581579089,0.0344284698,-0.0294537786,-0.1684188247,0.1105972081,-0.3508931994,0.1684127003,0.0006384781,0.356484592,0.5651245713,0.1221901104,0.3544335961,-0.011749737,-0.179120943,0.0342154093,0.5692538619,0.0489683226,-0.0831356049,0.1574838907,-0.0839778334,0.0501949899,-0.3444976211,0.5438238978,0.4237298369,0.7482103109,0.00764461,-0.2656746507,-0.2140215188,-0.0473108031,0.0873438418,0.4619984329,0.1152036563,-0.0019732735,0.1031655148,-0.0077709132,-0.3689866662,0.0687636361,0.3272413313,-0.3060223758,0.6822800636,-0.0861405656,0.1421229839,-0.1508950889,0.0387474075,-0.4797349572,-0.2150186151,-0.1107819304,0.6301547289,-0.0117602758,0.1536504626,0.1211947575,0.0564348139,0.3472965062,-0.1567848623,-0.3279043138,-0.1055673584,-0.3150283694,0.085141547,0.4395576417,0.0223501641,0.3921347857,-0.258122474,0.2694416344,0.0293090865,0.0256403182,0.2339448333,0.1045444384,-0.2102553248,-0.0032201924,-0.0132923527,0.144697383,-0.2322029769,0.0543964766,0.4595737755,-0.2129658014,-0.2590139508,-0.1943206042,0.0714289621,-0.1848265082,0.3223711848,0.048870001,0.0114345308,-0.3538573384,-0.3810112774,0.0648380071,0.0386596434,-0.0400859155,-0.096080564,-0.1224542782,-0.1822863668,0.2952705324,0.0197274294,-0.0703630149,-0.1389851272,0.0344718695,-0.3196364939,0.099314779,0.118077226,0.2433472276,0.2680312693,-0.1300933659,-0.6249844432,0.0362713188,0.0963543504,-0.2753038406,-0.135342747,-0.7036629319,0.1318925023,-0.2303450108,-0.0175632741,-0.2686011791,0.1542358696,0.2390289605,-0.3774132133,0.1194948927,0.0705103576,-0.1370931864,-0.0309503023,0.3859170675,0.5811342001,-0.1503598988,-0.2410179228,-0.1784706712,0.0995239317,0.3083503842,0.3931098878,-0.0097178603,0.3673390746,-0.123205103,0.0689337254,0.0478423163,0.1104082093,-0.4387206733,0.0502612069,0.1064193025,-0.0024918374,-0.3263646662,0.0607176013,0.068100512,0.3225802183,0.0500689112,0.1457122117,0.3562041819,-0.0380589738,-0.3022038639,-0.166160956,0.5067033172,-0.0243811589,0.2399932891,0.0606650114,0.2152320296,0.1761258841,0.5853968859,0.0900954232,0.4497523606,0.8086163402,0.0410142392,0.2599589825,0.2317345142,-0.0186162889,-0.2249160856,-0.0196688753,0.0085124895,0.3689703643,0.1535931975,-0.1252012253,-0.1137937829,-0.0342600048,0.0428101011,-0.1327238232,0.0137876468,0.1244750693,0.013955065,0.103322871,0.195651561,-0.3357091248,-0.1871322095,-0.1004585102,-0.628423512,0.4010466635,-0.1259019822,0.1213175729,-0.351793617,0.1536542922,-0.0062448056,-0.0991640389,-0.1660703272,0.0231198594,0.2443483472,-0.0934233889,0.4642220438,0.0366498344,0.1356393397,-0.3289432526,-0.0140460925,0.0952283517,0.0065336777,0.1135988459,-0.6195914745,-0.2180344015,0.303324312,0.2615132332,-0.2023946941,0.2310920209,0.6155297756,0.1041673198,0.1185883731,-0.4379398823,0.3753535748,0.1786495596,0.0308943242,0.0456814729,-0.2653453052,0.0838880837,-0.0787140429,0.1410297155,0.3548550606,0.3276737332,-0.3154002428,-0.2769112289,0.2495277673,-0.2054148912,-0.1864516139,0.171800524,0.2963220179,-0.1858100146,0.4059666693,0.0281814244,0.1268699914,0.2485369146,0.1192484424,0.1118604243,0.1804561615,0.1863763034,-0.3563382626,-0.1338765919,-0.0549654141,0.3730303943,-0.0320130326,-0.0760621801,-0.0214119777,-0.0800322667,0.0657873228,0.081108883,-0.8917042017,-0.1627465636,0.0567930527,-0.0055605825,-0.1735197604,0.221001327,-0.0586623028,-0.1721410453,-0.234770909,0.1137968749,-0.1291861087,-0.2169316858,-0.2443751395,-0.0020234282,-0.0328176767,-0.05973747,0.1563057303,-0.4442386031,-0.4988531172,-0.6673246026,-0.3995530903,0.2493881136,0.0330022871,-0.0058581671,0.1518760622,0.4192475975,-0.4625677168,0.1145005971,0.1680914313,-0.0909884647,0.0793017671,0.0523316003,-0.2788633406,-0.0128775975,-0.1035127789,-0.1117368937,-0.2085937411,0.0770719126,0.1515254378,0.3591684997,0.2418566346,0.2694685757,0.0467882194,-0.2720733285,-0.1089181527,-0.1376685202,-0.3673323989,-0.4052069783,0.1023066193,-0.2623264194,-0.3153405786,0.0601709597,-0.0712210611,-0.0488249511,-0.0601378009,-0.6084823608,0.0356831625,-0.129826054,0.122962445,-0.3658171296,0.1698866487,0.3927969038,-0.0339411125,0.1076795906,-0.0673577413,-0.1049568504,0.0132308835,0.3028675616,0.598723352,-0.1778348237,0.119486928,0.051991865,0.5267832279,0.4268625081,0.5182902217,0.0735583007,-0.0113334209,0.2064493299,-0.1893418282,-0.0263113696,0.1609650403,-0.0731869116,-0.1525501162,0.5049317479,0.1899757683,-0.0368412435,-0.0840222538,-0.1401457936,-0.5720708966,-0.1805714518,0.1014308929,0.2487937212,0.5083112121,0.0969913676,0.0378734842,-0.497371912,-0.3354012072,-0.0251508076,0.0841027051,0.3870459199,0.1760148108,-0.2663965225,-0.061365217,-0.4050893784,0.3238611817,0.1768577546,-0.1495721787,-0.0401820876,0.2509098649,0.1589121968,0.0247878283,0.2293138355,-0.3580600321,0.3446396887,0.0843645334,0.2656255662,-0.2265178263,0.1304504424,-0.0130930627,-0.0009669185,0.4135722518,-0.160813421,-0.3760305941,-0.0049715899,0.1477148235,0.3667692542,-0.4045117199,-0.131616056,-0.128435269,-0.1067699194,-0.1258197129,-0.2567431927,0.1178541481,0.1275595427,0.4597366452,0.2020027041,0.309612304,0.375074476,-0.0222707056,0.10822092,0.3880188465,0.5211099982,-0.1680448204,0.0485164747,0.1964561194,0.2341528833,0.5824503899,-0.0569688827,-0.2218944132,0.2266015112,-0.2282979637,0.3059276938,0.0471702218,0.0986551791,0.1458796114,0.3456745148,0.1033604294,-0.2108806521,-0.0053259037,0.0337197743,0.213078171,-0.1709383875,-0.2221275419,0.3496203125,-0.1516682059,-0.1745558828,0.1300293654,0.3755857646,-0.274025023,0.2951129079,0.1486352384,1.0607903004,0.3447882533,0.0788980275,-0.123635605,0.0092162723,0.2861762941,-0.1518213153,0.1172869802,-0.3929425478,0.1485294253,0.0549722724,0.1671450585,-0.0804553628,0.1859728843,-0.0198749043,0.0259391665,-0.2554929554,0.2764765024,-0.162109226,0.4707971811,0.2257468253,-0.0783592463,0.0818223804,0.0511321425,-0.0212818347,0.488083899,-0.3742939234,0.1539584398,-0.007319205,-0.0668250471,-0.4324820638,-0.1401273012,0.1286010593,0.2918867767,0.0101489956,-0.353690207,0.0581821166,0.3371959329,0.0042057899,0.288140893,-0.2000993639,0.2276308835,-0.1306284517,0.1145418286,0.2848286927,0.2474770695,-0.1515913308,-0.1420457512,-0.2853090167,0.0626695752,-0.2895980179,-0.1410525739,-0.3069053888,-0.255382359,-0.2306338847,-0.2116296738,0.3574074209,0.0723324791,0.0303959921,-0.3096146584,0.0463452004,0.3127027452,-0.0626065657,0.0086875847,-0.0190824568,-0.258113116,-0.0561034717,0.2872316837,0.0907894075,-0.0236315262,0.4977655113,0.3514389992,-0.2264496982,-0.0814527571,0.0067697088,-0.3021357954,-0.3260261714,-0.1122288257,-0.1603142619,0.300599277,-0.5272343159,0.0222011805,0.2055333853,-0.0706898868,-0.1759598553,-0.4852510095,-0.274053961,0.2114831507,0.1044387519,-0.1064962,0.1058685035,-0.3244018257,0.1323316395,0.1528977156,-0.2119720876,-0.2077441365,-0.1508572847,0.2941684723,-0.0118108615,0.050461784,-0.2046804428,-0.1677406877,0.0085928757,-0.4053913951,0.2234886438,-0.1143293902,0.0449705943,0.2320536226,-0.0507527143,-0.2713606358,0.1506614387,-0.1844345778,-0.0352683961,-0.2412580252,-0.1594513208,0.0748617873,0.1108355075,0.2966975868,-0.0648251176,0.222153753,-0.3317199647,0.2220244557,0.2099198997,0.1152129024,-0.0242196973,0.243647784,0.4147491157,-0.0241700541,-0.0524010137,0.0270160418,-0.1087437868,-0.0767961591,0.056904681,-0.2534982264,0.3784090877,0.4685874581,0.1953288168,0.0135253584,-0.1067930385,0.1757264882,0.0882330164,0.1909665167,-0.534011662,-0.0693596974,0.3859979808,0.1083145365,-0.040389061,0.0816605985,-0.0318532549,-0.1537616253,0.0482423529,0.1064629927,0.0540366657,-0.0889346823,0.0465966947,0.2780443132,-0.020541491,0.1160141528,-0.3238047957,0.2555984855,0.024991069,0.3017582595,0.288700074,-0.0545161739,-0.0528357662,0.0232010372,0.0257221963,-0.1221859083,0.1985286921,0.1686081886,-0.0738741979,0.1862123758,0.1652341336,-0.1427025795,0.1433381438,0.0247069728,-0.3094670177,0.1690118909,-0.1509307921,-0.0258259699,-0.5115837455,-0.1569619477,0.1301718801,-0.0323113948,-0.0002201961,-0.3970230818,0.0464692041,0.1543808281,-0.3306342363,-0.4434216022,0.0411055423,0.1818974465,0.2838747203,-0.2849944234,-0.0035640812,0.1924327016,-0.1421485841,-0.2672965527,0.557753861,0.2276382446,-0.0958473384,-0.1166276336,-0.053111203,-0.0016043421,-0.0355980806,-0.2380159795,0.1267373711,0.0315539576,-0.1447692364,0.1087078229,0.0258535072,-0.0964853615,-0.1042926833,0.270960927,0.1727023572,0.0668981448,-0.3142918348,-0.1021554619,-0.2569535375,-0.0656749979,-0.1126782969,0.0564748347,-0.1030590832,0.3124278784,0.326205492,0.0918547213,-0.1263726503,0.0192983672,-0.0242653657,0.110448204,0.2006641179,0.0578148663,-0.0627017692,-0.4903538823,-0.5188625455,0.2414766848,-0.4477035403,-0.3777931929,0.1942809671,0.1893029362,0.0575312413,0.1087449491,0.1562176794,0.4024175406,-0.2073469758,0.508155942,-0.1827882081,-0.0603978187,0.0534946322,0.2580348849,-0.0862741172,-0.2108464986,0.3570457101,0.1798704118,-0.1149010211,-0.1822195947,-0.0029852458,-0.0081472527,0.2827708721,0.2995383441,0.204402104,0.171300292,-0.308740139,-0.2129374146,-0.0485414974,-0.2621197402,-0.0270408075,0.7471531034,0.0397932604,0.0338847898,-0.1692429036,-0.1632170379,0.0909647197,0.2456413805,0.1131736487,0.1881642044,-0.5572510958,-0.0396219939,0.0552947298,0.2511416078,0.0060987161,-0.0872540399,0.123847045,-0.2110015154,-0.4428655803,-0.2427507937,0.0140888328,-0.398443222,-0.2658976316,-0.7156354785,0.1114247516,0.0852635652,-0.3103724718,-0.200728178,-0.1691340059,0.4191295207,0.1735121459,-0.0139667047,-0.3795804679,-0.1692909151,0.0258111041,-0.1268443316,0.35200876,0.1013602167,-0.4374337196,0.0498603657,-0.2752310038]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1759","title":"wikipedia dataset incomplete","comments":"The wikipedia articles are processed using the `mwparserfromhell` library. Even if it works well in most cases, such issues can happen unfortunately. You can find the repo here: https:\/\/github.com\/earwig\/mwparserfromhell\r\n\r\nThere also exist other datasets based on wikipedia that were processed differently (and are often cleaner) such as `wiki40b`.\r\n\r\n","body":"Hey guys,\r\n\r\nI am using the https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wikipedia dataset.\r\nUnfortunately, I found out that there is an incompleteness for the German dataset.\r\n For reasons unknown to me, the number of inhabitants has been removed from many pages:\r\nThorey-sur-Ouche has 128 inhabitants according to the webpage (https:\/\/de.wikipedia.org\/wiki\/Thorey-sur-Ouche).\r\nThe pickle file however shows: franz\u00f6sische Gemeinde mit Einwohnern (Stand).\r\n Is it possible to fix this?\r\n\r\nBest regards \r\nChris\r\n","comment_length":48,"text":"wikipedia dataset incomplete\nHey guys,\r\n\r\nI am using the https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wikipedia dataset.\r\nUnfortunately, I found out that there is an incompleteness for the German dataset.\r\n For reasons unknown to me, the number of inhabitants has been removed from many pages:\r\nThorey-sur-Ouche has 128 inhabitants according to the webpage (https:\/\/de.wikipedia.org\/wiki\/Thorey-sur-Ouche).\r\nThe pickle file however shows: franz\u00f6sische Gemeinde mit Einwohnern (Stand).\r\n Is it possible to fix this?\r\n\r\nBest regards \r\nChris\r\n\nThe wikipedia articles are processed using the `mwparserfromhell` library. Even if it works well in most cases, such issues can happen unfortunately. You can find the repo here: https:\/\/github.com\/earwig\/mwparserfromhell\r\n\r\nThere also exist other datasets based on wikipedia that were processed differently (and are often cleaner) such as `wiki40b`.\r\n\r\n","embeddings":[0.145783633,0.1651307046,-0.0604214706,0.4512960911,-0.1779640317,0.1422897428,0.1325682402,-0.0599784143,0.2211329937,0.1373941153,0.1483723372,-0.0644864812,0.4038364589,-0.4707553387,-0.0247412473,-0.1964903921,0.1687633395,0.1441214383,-0.3371717632,-0.2833011746,-0.150470674,0.4101845622,-0.3018541932,-0.3771216571,0.0009892298,0.1730692834,0.0294008255,-0.1099493206,0.0388543122,-0.255102247,0.0370923541,-0.1234474033,0.0171350595,0.2694829702,-0.0001211163,-0.1979540586,0.1709197611,0.0784731358,-0.4775889814,-0.0129319895,-0.2857728004,-0.3218949437,-0.064013049,-0.3470281959,0.3659273982,0.1441548318,0.2030157298,0.0585192963,-0.0170048568,-0.1380969882,0.1241210103,-0.3913951814,0.2005907446,0.0655082017,0.4258055389,0.5858446956,0.1269980669,0.2719853222,-0.0048444187,-0.1762056053,0.0176583491,0.5391578078,0.0057500992,-0.0616088994,0.1903233975,-0.0977577493,0.0417255089,-0.375428915,0.5308150649,0.4282439351,0.6989920139,0.0609150454,-0.2956066132,-0.2108039111,-0.0446788296,0.1271712184,0.4841116369,0.1465952694,0.0158949103,0.0581623539,-0.0173052158,-0.3101015389,0.0548868589,0.2982166708,-0.2402050644,0.7044153214,-0.0301209353,0.1586293578,-0.1816008389,0.0438450165,-0.3793136477,-0.2826716602,-0.0948735401,0.6344087124,0.0443101265,0.1643259674,0.0751382336,0.0772433355,0.3611153364,-0.2755673528,-0.3678851128,-0.1502043605,-0.3200321496,0.0574990436,0.4369731843,-0.0295739174,0.3606520295,-0.200948894,0.2603800297,0.0274022203,0.0076454156,0.280877471,0.1464324296,-0.1832085103,0.0047896113,-0.0615609214,0.0689780787,-0.2187531292,0.0432224944,0.5090514421,-0.2550533116,-0.2354683429,-0.2305583358,-0.0571923517,-0.1718482971,0.358758539,0.0155743603,0.0492518358,-0.3456348777,-0.3288669586,0.0548176505,0.0684341863,-0.0149544394,-0.076012373,-0.1218015626,-0.1956166625,0.3033376038,0.0875600502,-0.1614001989,-0.143944785,0.005945567,-0.3173173964,0.0513227582,0.1171248853,0.2250036299,0.2758552134,-0.1570949256,-0.6372517347,0.0443666168,0.1000973582,-0.3013814688,-0.0364339538,-0.6779907942,0.1429633498,-0.2141105086,-0.0380949564,-0.2095004022,0.1293669045,0.2918624282,-0.3391788304,0.0900603682,0.0538967252,-0.1480180025,-0.0453009531,0.3423948586,0.5277428627,-0.0726070404,-0.2131535262,-0.240062505,0.185655728,0.3504770994,0.4081870615,0.0336732157,0.3773176968,-0.0445005782,0.0923473015,-0.0536393002,0.1296361834,-0.4094777107,0.0275954809,0.1863747835,0.0908563286,-0.3522014916,0.0342830569,0.0083405105,0.2894448042,0.028009614,0.1449129134,0.3542480171,-0.0304772258,-0.3212764263,-0.1869341582,0.486083746,-0.0571016073,0.2496021241,0.0195017811,0.1994181126,0.1431911886,0.5697168112,0.0964566693,0.4354119599,0.7665311694,0.0510880984,0.2489742935,0.2217193395,-0.0409097932,-0.140591085,-0.0486743562,0.0215399005,0.3681215644,0.1580456942,-0.1656516045,-0.1193706095,0.0043172794,0.0545834191,-0.1634843349,0.0601624586,0.1263922453,0.0036240914,0.1342677474,0.2368512005,-0.4029847682,-0.2450559586,-0.1003006026,-0.5714448094,0.4301270843,-0.1574401259,0.0866298378,-0.3112455606,0.1544338614,-0.046983514,-0.0862774476,-0.159728229,-0.0192471482,0.2688753307,-0.090005435,0.5272328258,0.0586883686,0.1302475035,-0.3522474766,-0.0622084849,0.1057255641,0.0245693717,0.1013292223,-0.6246506572,-0.2231550813,0.1938699633,0.2869718969,-0.2390442342,0.2374240458,0.5918765068,0.1011636853,0.1432625353,-0.4077341259,0.3371963203,0.1860398948,-0.0232663974,0.0664843842,-0.1998917907,0.0710164085,-0.0652072728,0.1524333358,0.3409911096,0.3338623643,-0.294464767,-0.3108476102,0.2318411916,-0.229502514,-0.2038976252,0.1474936157,0.2784415781,-0.2145889252,0.3905881345,0.0154753011,0.1806613505,0.2805579901,0.1175009906,0.1110365018,0.1559869647,0.2408441752,-0.3074939251,-0.0712040141,-0.0985700563,0.3704074919,-0.0236846171,-0.0726825222,-0.0377204977,-0.0324173644,0.03895678,0.0751195252,-0.8128618002,-0.2056323886,0.0372341052,-0.0130473068,-0.1882798523,0.1903685778,-0.0759144574,-0.1974064708,-0.1964819431,0.2040979415,-0.1688502282,-0.2201683372,-0.2426234633,0.014934645,0.0277103335,-0.0396465696,0.0803911239,-0.4445254803,-0.5066559911,-0.6288619041,-0.4411588609,0.2037249506,0.012816513,0.0592472665,0.0856286436,0.3786281347,-0.4475949109,0.0401363708,0.1387779415,-0.0454167165,0.1066057235,0.0998874083,-0.2520989776,-0.1221778393,-0.0513879843,-0.1038913876,-0.1540354043,0.1118792892,0.0949263647,0.3399053216,0.2324995548,0.2452100962,0.0753850043,-0.2840758264,-0.0840278044,-0.1073355228,-0.3288804889,-0.2740887702,0.1363540739,-0.2322098166,-0.3054614365,0.0712974221,-0.0898418278,-0.0707059354,-0.0583204627,-0.5821250677,0.0886529684,-0.1367853135,0.1188964024,-0.3265385032,0.1683312356,0.3140134215,0.0313094854,0.1092464998,-0.0234486405,-0.097319819,-0.0231414456,0.2916723788,0.5615307689,-0.15699628,0.1271216422,0.0614072233,0.5506908894,0.5221112967,0.4804535806,0.0566064045,0.0543397851,0.1798391342,-0.1528558284,0.0092240963,0.1997855902,-0.1374715716,-0.1388491392,0.5862267613,0.1349093467,-0.0819989443,-0.0399247035,-0.1424108297,-0.6010252237,-0.1871091872,0.0840949565,0.2459484637,0.5408473611,0.1182598695,0.0256258305,-0.4206124842,-0.3877400756,-0.0325785279,0.046978239,0.4023585618,0.2233219147,-0.2932401597,-0.1286980361,-0.3533387482,0.2833379209,0.1879225522,-0.1200781986,0.0142864874,0.2271016985,0.1421093196,0.0559371747,0.1480295658,-0.3797412813,0.2764495611,0.0346303992,0.2937158346,-0.2467794418,0.1407268941,-0.0480145216,0.0527354516,0.3260178566,-0.1593944579,-0.3618428111,0.0406280942,0.1157185808,0.3878132999,-0.3619176745,-0.1815863699,-0.0694167912,-0.0828781277,-0.0923163891,-0.2579603791,0.097644873,0.1278556436,0.352101624,0.154668048,0.2665950954,0.3562868237,-0.0003054901,0.134926036,0.3300655186,0.4820422232,-0.1621252,0.0101773059,0.1432143301,0.2127494961,0.5548613667,-0.0642780289,-0.1784170419,0.1429616362,-0.2956863046,0.2526402473,0.0168636441,0.1204579622,0.1171630546,0.3544507623,0.0122522963,-0.1734018773,-0.0570187457,0.0792110041,0.195433557,-0.0970923826,-0.1670973003,0.3800335228,-0.1346752644,-0.17459777,0.1131977886,0.3719496727,-0.2953276634,0.2891617715,0.2699331939,1.0011174679,0.3788383603,0.0121556353,-0.1793261766,-0.0529594459,0.2972718477,-0.1071492955,0.1084455624,-0.3848905861,0.1584665328,0.068654485,0.191152662,-0.0917754099,0.2376818806,0.0016857316,0.037453983,-0.286705792,0.2256292403,-0.1718363315,0.5127314925,0.2711779773,-0.0895369053,0.1535363644,0.0653078407,-0.0509207211,0.4138062894,-0.3755304515,0.1747429371,-0.0412141122,0.0088907015,-0.4454413354,-0.1418692917,0.1368909478,0.2824143767,0.0549767688,-0.3239550591,0.10390044,0.3498434126,-0.0466885343,0.2900590301,-0.1569158137,0.2296323776,-0.1612826437,0.0859037936,0.3205883503,0.2394997627,-0.1280836463,-0.1320456862,-0.2758923173,-0.0044747014,-0.2783293724,-0.1323282719,-0.2475752085,-0.2759231925,-0.1543503702,-0.1918700486,0.3122684956,0.0656158701,-0.0326024927,-0.3162341118,0.0795125291,0.2867476642,-0.042628292,0.1013093591,-0.0385174789,-0.2451887429,-0.0434151366,0.2008703053,0.0943084136,0.0440776125,0.4918823242,0.3106234968,-0.3043175638,-0.0675351992,0.007316248,-0.3844090998,-0.3108671904,-0.1408266276,-0.1906195581,0.2942609787,-0.5391882658,0.0431110486,0.209644258,-0.0622674301,-0.161613673,-0.4557008743,-0.2316180021,0.1932724267,0.1517165452,-0.1242123246,0.0442088693,-0.3824693859,0.1184570417,0.2389193028,-0.2095684558,-0.2484270334,-0.1406660229,0.2769829631,0.0125341825,0.047386121,-0.3052440882,-0.1483774632,0.0197530836,-0.3925029039,0.3026607931,-0.1368869841,0.0554951951,0.2028739899,-0.0349629074,-0.2961297333,0.1907842457,-0.1808453202,-0.028047815,-0.2980057299,-0.2192565352,-0.0351183638,0.0956452116,0.3528510034,-0.0468889549,0.1906128973,-0.3021704853,0.2236239016,0.1446101516,0.0907531306,-0.012356502,0.2586515546,0.520139575,-0.0052804677,-0.0617355518,0.0004938229,-0.1591590792,-0.0417903699,0.1035340503,-0.2429112494,0.3560665548,0.5593594313,0.1313528568,0.0726648197,-0.118219085,0.1067297682,0.0877885148,0.2283999473,-0.5155929923,-0.0981054232,0.3583297133,0.1445646733,-0.0694257915,0.1060203686,-0.0033220376,-0.2009487897,0.0152547657,0.1081549078,0.057489872,-0.0894693807,0.0416972786,0.2708217502,0.0354921259,0.1084888726,-0.3412257433,0.1712488383,0.029204756,0.296240896,0.3516981304,-0.0876998529,-0.0002045053,-0.0156003349,0.0786094293,0.0215557832,0.2241318077,0.1309172064,-0.0751387626,0.2060133517,0.2162433416,-0.1399489939,0.1883498877,0.0065366114,-0.3409818113,0.1340706199,-0.130917713,-0.0184517857,-0.509867847,-0.1693205237,0.1387696564,-0.0741069317,-0.0382837728,-0.3801883459,0.0856136382,0.1762547046,-0.3443518579,-0.4727272689,0.1152074262,0.1771207452,0.3541853428,-0.3139342368,-0.0319401808,0.2008035332,-0.1550044119,-0.3127721846,0.5225102305,0.2265105397,-0.0742238685,-0.1609318405,-0.0873167962,-0.0264170412,-0.0357025154,-0.1778579056,0.1859539151,-0.0164363384,-0.1689247191,0.0997921377,0.0280410834,-0.1042821109,-0.1733643711,0.2708337009,0.1868908405,0.0621004961,-0.3633446395,-0.1073137298,-0.2540304661,-0.0255058259,-0.1426617056,0.1547555774,-0.0709241331,0.2788181901,0.277556926,0.0554703958,-0.1205735877,0.027389586,-0.063549161,0.1031074673,0.2089660019,0.0098679317,-0.0618140101,-0.4642460346,-0.4811490774,0.2121981084,-0.4761243761,-0.4154522419,0.2206808776,0.180460304,-0.0116343629,0.1107136384,0.0977721512,0.3260559142,-0.233854413,0.5221579075,-0.1762695909,-0.0663128421,0.0272838213,0.1712913662,-0.0882775113,-0.1848082542,0.3778120279,0.1932524741,-0.1162320971,-0.1612396687,-0.0358335078,0.0842573345,0.371042043,0.2509284616,0.2829445004,0.1493945867,-0.3115492463,-0.1744605452,-0.0440336876,-0.2770478725,-0.0549364015,0.7693365216,-0.0266972724,0.0448488295,-0.1694971174,-0.142961964,0.1333087683,0.2583148479,0.0921495855,0.1760729402,-0.5699791312,-0.0860629529,0.0051493039,0.2407623976,0.0549502969,-0.0580487847,0.1391150653,-0.2182603478,-0.4384664297,-0.3041389585,-0.0382790416,-0.3999017477,-0.285128504,-0.6288002133,0.0630685091,0.1092441306,-0.3254574835,-0.1430913359,-0.2490007579,0.4475740194,0.1568648964,0.0031935051,-0.4028256238,-0.1381462067,0.0274075251,-0.0907366127,0.2588889599,0.0797083974,-0.4124220014,0.0294043403,-0.2853695452]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1759","title":"wikipedia dataset incomplete","comments":"ok great. Thank you, @lhoestq. ","body":"Hey guys,\r\n\r\nI am using the https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wikipedia dataset.\r\nUnfortunately, I found out that there is an incompleteness for the German dataset.\r\n For reasons unknown to me, the number of inhabitants has been removed from many pages:\r\nThorey-sur-Ouche has 128 inhabitants according to the webpage (https:\/\/de.wikipedia.org\/wiki\/Thorey-sur-Ouche).\r\nThe pickle file however shows: franz\u00f6sische Gemeinde mit Einwohnern (Stand).\r\n Is it possible to fix this?\r\n\r\nBest regards \r\nChris\r\n","comment_length":5,"text":"wikipedia dataset incomplete\nHey guys,\r\n\r\nI am using the https:\/\/github.com\/huggingface\/datasets\/tree\/master\/datasets\/wikipedia dataset.\r\nUnfortunately, I found out that there is an incompleteness for the German dataset.\r\n For reasons unknown to me, the number of inhabitants has been removed from many pages:\r\nThorey-sur-Ouche has 128 inhabitants according to the webpage (https:\/\/de.wikipedia.org\/wiki\/Thorey-sur-Ouche).\r\nThe pickle file however shows: franz\u00f6sische Gemeinde mit Einwohnern (Stand).\r\n Is it possible to fix this?\r\n\r\nBest regards \r\nChris\r\n\nok great. Thank you, @lhoestq. ","embeddings":[0.1494655162,0.0706044808,-0.0663508698,0.4164565504,-0.1602319479,0.0935590938,0.1459739208,-0.1077719182,0.2492056191,0.1976480633,0.1634980738,-0.0981558561,0.380328089,-0.4207990468,0.0191584919,-0.2423884273,0.1626209319,0.1680546403,-0.3314844668,-0.3050085008,-0.1061897948,0.4273441434,-0.3013417423,-0.3642159104,0.0172304753,0.2017402798,0.0547052696,-0.0847612768,-0.0282020792,-0.226417616,0.0105246911,-0.1829527467,-0.0021669313,0.2998417318,-0.0001233001,-0.1916974932,0.1219362244,0.0505164005,-0.5027077794,-0.0618752837,-0.3043524623,-0.3383553922,-0.1022527143,-0.309727639,0.3452602327,0.1279518008,0.254889518,0.0647387654,-0.0436499082,-0.1867807657,0.0955038145,-0.3581036329,0.1965487748,0.0421855077,0.3626025319,0.5887580514,0.0921911076,0.3782013953,0.0203296542,-0.1261488646,0.018579701,0.5866979957,0.1082650051,-0.0886563584,0.1041596159,-0.082581155,0.0219712481,-0.3331803977,0.5067540407,0.4272122383,0.724974215,0.0271046571,-0.3094195426,-0.2088791877,-0.0718200579,0.0624142326,0.4730197191,0.1082817614,0.0386906639,0.0997731239,0.0100503052,-0.3517794013,0.0768127665,0.2967782915,-0.2372471094,0.7121506929,-0.05430701,0.1550558656,-0.1327212751,0.015079611,-0.5119352937,-0.1699641943,-0.1095346957,0.6541705132,0.0130833928,0.105940409,0.1146239266,0.0330824032,0.3728637397,-0.1533526033,-0.3495481908,-0.1318877637,-0.4036119878,0.0688479394,0.4210546017,0.0072299391,0.3820686042,-0.2286904305,0.2568836808,0.0140734958,0.0404748283,0.2806217074,0.138831839,-0.2412751019,-0.045472797,-0.0317694172,0.0998177603,-0.1990041137,0.0812444761,0.5035442114,-0.2494067848,-0.2648427784,-0.1956779063,0.0300555043,-0.1613406241,0.3050386906,-0.0126664024,0.0163539127,-0.3298102021,-0.3855405748,0.0616821125,0.069535464,-0.0492942706,-0.1366926581,-0.1379124075,-0.1696882397,0.3215602338,0.0345139541,-0.0899899304,-0.0998475999,0.0402483121,-0.344584167,0.0805721954,0.1235455498,0.1913522184,0.2518861592,-0.1273199618,-0.6027302146,0.0420993418,0.0698459148,-0.2904687822,-0.0978349298,-0.7131019235,0.1144629344,-0.1701044291,-0.0282119699,-0.2434897572,0.103803359,0.2306281477,-0.3896898031,0.1277629435,0.0830944404,-0.1229059994,-0.0237571523,0.3061232865,0.5219880939,-0.1061266586,-0.2559818029,-0.1664964259,0.1272678524,0.370945394,0.4558882415,0.024674885,0.3128069043,-0.1095658615,0.146970287,0.0038493467,0.1244560704,-0.4556643367,0.0180528406,0.1146694869,-0.0003662386,-0.2926528454,0.0384463742,0.0037418117,0.303645879,0.0716817752,0.0848797783,0.3793571293,-0.054300122,-0.3175931275,-0.1785656065,0.5057629347,-0.068870768,0.2158600837,0.0400844924,0.2097590268,0.1134237424,0.5782539248,0.1134409681,0.4517666399,0.8089392185,0.0777946934,0.2191382647,0.2220048457,-0.0216583703,-0.183515951,-0.0465118662,-0.0308381133,0.3702415824,0.1916066557,-0.1181045994,-0.1086121351,-0.0116819572,0.067113474,-0.1444303989,0.0059149321,0.1087328047,0.0175287165,0.0972392112,0.1916528046,-0.3964972794,-0.1964006424,-0.0728731304,-0.6157619953,0.3751233816,-0.1691828072,0.0880898684,-0.3779814243,0.1792717874,-0.0318757482,-0.1313417554,-0.1357645839,-0.0327622183,0.2751289904,-0.0878044367,0.5135022402,0.0344569683,0.1361728311,-0.3318755329,-0.0577017441,0.0670385286,-0.0192962121,0.1049707159,-0.6084150076,-0.226970762,0.2949789166,0.2435998917,-0.2416025996,0.2568028569,0.6374652386,0.129885599,0.1212308854,-0.4281192422,0.3754766583,0.1618727744,0.0000308034,0.0633220449,-0.3035465777,0.0756941065,-0.0902865306,0.1434658915,0.3769550323,0.364908129,-0.2834347486,-0.2941161096,0.2769707441,-0.2332340032,-0.1897920668,0.1583859921,0.2997081876,-0.2168618739,0.4149315059,0.0399565436,0.1296894252,0.2067985386,0.085848771,0.1321330518,0.1715526283,0.1921866983,-0.3032917976,-0.1344324946,-0.0677670091,0.3647299707,-0.0073799486,-0.0756511539,-0.0415285118,-0.0204801504,0.0814558417,0.0451209582,-0.858087182,-0.1348477453,0.0853558108,-0.0062730401,-0.1900239587,0.2442240864,-0.0599858947,-0.1765693128,-0.2479281425,0.1878914982,-0.174268499,-0.193786934,-0.2183891833,-0.0070924889,-0.0368130766,-0.1065592393,0.1138503328,-0.4442293942,-0.5218816996,-0.657042861,-0.4152415693,0.2412212044,0.0186329726,-0.0102269175,0.1785593778,0.4393773377,-0.4719170034,0.0972793698,0.1517515182,-0.0482171029,0.0587891378,0.0506960116,-0.2837360203,-0.023016382,-0.0870412737,-0.0822001398,-0.2107230723,0.1028375477,0.2132871598,0.3688127995,0.2573288083,0.3098420799,0.0679798201,-0.2803899348,-0.1206318364,-0.1356301755,-0.3655002415,-0.4015804231,0.0789600313,-0.2595092952,-0.2707222998,0.0510584787,-0.1095310673,-0.0658383146,-0.1273333579,-0.5478389263,0.0535377823,-0.1593538076,0.1355970055,-0.3408167958,0.1864566505,0.4116995037,0.0075190947,0.116665788,-0.0701728016,-0.1439398378,-0.0038761441,0.3544436395,0.649083674,-0.1533523798,0.1145021692,0.0831826031,0.5699127316,0.4673727155,0.507337451,0.0842147171,0.0258921906,0.172245115,-0.1568309069,-0.0115847522,0.2017399073,-0.0688435733,-0.1355171204,0.5082337856,0.1855338812,0.0333817862,-0.0301663559,-0.1194596067,-0.5655457377,-0.1634268463,0.1308961809,0.2550827861,0.4977904558,0.0997067913,-0.027953947,-0.4799258411,-0.29371351,-0.0726270303,0.0819391012,0.4050561786,0.2130585462,-0.2602979541,-0.0909087509,-0.3730365932,0.305734098,0.2109376341,-0.1316435337,-0.0655269995,0.2466438562,0.1643230468,0.0689327344,0.222637862,-0.3617118895,0.3289341629,0.0197836123,0.2890817523,-0.1573143601,0.1359351873,-0.0059154532,0.0346767157,0.4368779063,-0.1392279267,-0.3898054659,0.015060639,0.1346267909,0.3307594061,-0.3887256384,-0.1359630525,-0.1058500707,-0.0921204016,-0.0986278653,-0.2457529455,0.1483512521,0.110716112,0.4927206337,0.195542872,0.3101830184,0.3442770243,-0.0057043526,0.0960618556,0.3692839146,0.5197784901,-0.2038056403,0.0179625824,0.2195785493,0.2201221734,0.5805825591,-0.0606369115,-0.1745736748,0.1819626242,-0.2396120876,0.3247269392,0.0353011526,0.1238964051,0.1939641535,0.3268432915,0.081919305,-0.177889213,-0.0378471464,0.0951139331,0.2298914194,-0.126197502,-0.1594513208,0.3597323298,-0.1754661202,-0.2055506408,0.1203244999,0.4261186719,-0.2522003055,0.2502575815,0.1772893518,1.0977163315,0.4059517086,0.0602680296,-0.1154259369,-0.0478114597,0.2524145544,-0.1225333586,0.1055976525,-0.3736310899,0.1699964851,0.0474714376,0.1580686718,-0.0903487653,0.1825789213,-0.0067458153,0.0045471806,-0.2917152643,0.1807055473,-0.1651063859,0.4311937392,0.2305698842,-0.0741698891,0.139810428,0.0391577594,-0.0587360822,0.5056097507,-0.3516316712,0.1518404782,-0.038988065,-0.0821933895,-0.4649471343,-0.1692867577,0.1762503088,0.2888658345,0.1055493578,-0.3555231988,0.0923186541,0.3211099207,-0.0174863283,0.2903752029,-0.1501544267,0.2240755111,-0.1003122032,0.1110984311,0.2981273532,0.273691833,-0.1647111326,-0.1629944146,-0.3047099113,0.0610551797,-0.3126923442,-0.1455809027,-0.3066946566,-0.2378225327,-0.2488594502,-0.1927986592,0.3018719554,0.0696106255,-0.0379487947,-0.3062077463,0.0374170803,0.3288734853,-0.0929889455,0.0362999849,0.0038022578,-0.2412882447,-0.0341259055,0.2406121492,0.1006708592,0.0250878036,0.5093710423,0.3235272467,-0.2564942539,-0.0807046369,0.0206303801,-0.3155845404,-0.3041211367,-0.1279859543,-0.1303137839,0.2597368658,-0.5728529096,0.0120951068,0.2549056411,-0.032789465,-0.2140283287,-0.4407693446,-0.2896041274,0.214335829,0.042238459,-0.102171272,0.0679010078,-0.2971904278,0.1235837415,0.1914823651,-0.1948782653,-0.2229435146,-0.160087347,0.2576498985,-0.0036972922,0.0438148715,-0.2344537079,-0.1755087376,-0.0156976245,-0.4353678524,0.2481606007,-0.1050295308,0.0581548326,0.227149874,-0.0275080465,-0.2833813429,0.1768395007,-0.1282167882,-0.0090940613,-0.2218873054,-0.2202049196,0.0425518081,0.139929831,0.359765619,-0.07952106,0.2139827758,-0.2932629585,0.2142839879,0.2180496752,0.1021947563,-0.0011756473,0.2216404527,0.4247913659,-0.0452241264,-0.0576987416,-0.0123133436,-0.1402626783,-0.0766221881,0.0315202922,-0.274346143,0.3881453872,0.4635351002,0.1806302816,-0.0117187612,-0.1035494581,0.1423003823,0.1038509756,0.1857289225,-0.5448027849,-0.0728768632,0.3530917764,0.1297983229,-0.0490863919,0.0691981092,-0.0699785948,-0.1212417707,0.0338392071,0.1162699461,-0.0181806665,-0.085747242,0.0462374203,0.286110729,-0.0224584024,0.1184852049,-0.2898319364,0.1985766888,0.0605493188,0.2652162313,0.2633119822,-0.0540226661,-0.075807035,0.0021873426,0.0762173757,-0.074939914,0.1747277528,0.1822749972,-0.0651559979,0.1910926998,0.2084066123,-0.0856816545,0.1386852264,0.0106399879,-0.3088919222,0.164646998,-0.1476364732,-0.0629633963,-0.4984716773,-0.1198295951,0.1245121509,-0.0716402307,-0.0154950116,-0.3555729985,0.0483831279,0.1455497146,-0.320043385,-0.4837768376,0.0074156434,0.1797401309,0.3251796067,-0.274207741,0.0386787169,0.211342901,-0.163493678,-0.3133459687,0.5458288789,0.2076703161,-0.1157763153,-0.0868338645,-0.0667060092,-0.068741709,-0.0447488613,-0.2260116041,0.1262060106,0.0486164391,-0.2162435949,0.0971777365,0.0126061253,-0.0941422954,-0.0966992751,0.2670751512,0.1821808368,0.1034435406,-0.3281660974,-0.1242137626,-0.2716138065,-0.0563558005,-0.1171224862,0.0862840414,-0.0880736113,0.2588320374,0.3267031014,0.0708474219,-0.1347447783,0.0107464101,-0.0083457557,0.0857804567,0.1943074465,0.0160734579,-0.0738075301,-0.4712224901,-0.4080421925,0.2467099726,-0.4893448949,-0.4233098924,0.1663795114,0.1897788346,0.034765292,0.0862413719,0.1811811179,0.3993905783,-0.2336624265,0.5507082343,-0.1646545231,-0.0404907912,0.0913634002,0.250512749,-0.0649272949,-0.1720971465,0.358397454,0.2280311435,-0.1315845549,-0.2021318227,0.0291024894,-0.0179026518,0.3012357354,0.2703741789,0.2395481616,0.1640055031,-0.3368478715,-0.2078860849,-0.0418729559,-0.2984218001,-0.0489740744,0.7161167264,0.0315184779,0.0427515581,-0.15528813,-0.1428253651,0.1137514934,0.266433537,0.1220509931,0.1852339953,-0.5251134038,-0.015414957,0.0700627565,0.2107520849,0.0238406695,-0.100646019,0.1387717724,-0.242653504,-0.4708910584,-0.2898930609,0.0100338841,-0.3746478856,-0.2635761201,-0.6739988327,0.0702043995,0.0797035992,-0.3590925336,-0.1669872403,-0.2018074989,0.3833510876,0.2012369335,0.0034308215,-0.4006696045,-0.1434845775,0.0191001333,-0.1163679808,0.3921944797,0.1127796024,-0.4063971341,0.071939826,-0.2563809454]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1758","title":"dataset.search() (elastic) cannot reliably retrieve search results","comments":"Hi !\r\nI tried your code on my side and I was able to workaround this issue by waiting a few seconds before querying the index.\r\nMaybe this is because the index is not updated yet on the ElasticSearch side ?","body":"I am trying to use elastic search to retrieve the indices of items in the dataset in their precise order, given shuffled training indices.\r\n\r\nThe problem I have is that I cannot retrieve reliable results with my data on my first search. I have to run the search **twice** to get the right answer.\r\n\r\nI am indexing data that looks like the following from the HF SQuAD 2.0 data set:\r\n\r\n```\r\n['57318658e6313a140071d02b',\r\n '56f7165e3d8e2e1400e3733a',\r\n '570e2f6e0b85d914000d7d21',\r\n '5727e58aff5b5019007d97d0',\r\n '5a3b5a503ff257001ab8441f',\r\n '57262fab271a42140099d725']\r\n```\r\n\r\n\r\n\r\nTo reproduce the issue, try:\r\n\r\n```\r\nfrom datasets import load_dataset, load_metric\r\nfrom transformers import BertTokenizerFast, BertForQuestionAnswering\r\nfrom elasticsearch import Elasticsearch\r\nimport numpy as np\r\nimport collections\r\nfrom tqdm.auto import tqdm\r\nimport torch\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-uncased')\r\nmax_length = 384 # The maximum length of a feature (question and context)\r\ndoc_stride = 128 # The authorized overlap between two part of the context when splitting it is needed.\r\npad_on_right = tokenizer.padding_side == \"right\"\r\nsquad_v2 = True\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef prepare_validation_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n # We keep the example_id that gave us this feature and we will store the offset mappings.\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token\r\n # position is part of the context or not.\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (list(o) if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n\r\n\r\n\r\n\r\n# build base examples, features set of training data\r\nshuffled_idx = pd.read_csv('https:\/\/raw.githubusercontent.com\/afogarty85\/temp\/main\/idx.csv')['idx'].to_list()\r\nexamples = load_dataset(\"squad_v2\").shuffle(seed=1)['train']\r\nfeatures = load_dataset(\"squad_v2\").shuffle(seed=1)['train'].map(\r\n prepare_validation_features,\r\n batched=True,\r\n remove_columns=['answers', 'context', 'id', 'question', 'title'])\r\n# reorder features by the training process\r\nfeatures = features.select(indices=shuffled_idx)\r\n# get the example ids to match with the \"example\" data; get unique entries\r\nid_list = list(dict.fromkeys(features['example_id']))\r\n# now search for their index positions in the examples data set; load elastic search\r\nes = Elasticsearch([{'host': 'localhost'}]).ping()\r\n# add an index to the id column for the examples\r\nexamples.add_elasticsearch_index(column='id')\r\n# retrieve the example index\r\nexample_idx_k1 = [examples.search(index_name='id', query=i, k=1).indices for i in id_list]\r\nexample_idx_k1 = [item for sublist in example_idx_k1 for item in sublist]\r\n\r\nexample_idx_k2 = [examples.search(index_name='id', query=i, k=3).indices for i in id_list]\r\nexample_idx_k2 = [item for sublist in example_idx_k2 for item in sublist]\r\n\r\nlen(example_idx_k1) # should be 130319\r\nlen(example_idx_k2) # should be 130319\r\n\r\n#trial 1 lengths:\r\n# k=1: 130314\r\n# k=3: 130319\r\n\r\n# trial 2:\r\n# just run k=3 first: 130310\r\n# try k=1 after k=3: 130319\r\n```\r\n\r\n","comment_length":41,"text":"dataset.search() (elastic) cannot reliably retrieve search results\nI am trying to use elastic search to retrieve the indices of items in the dataset in their precise order, given shuffled training indices.\r\n\r\nThe problem I have is that I cannot retrieve reliable results with my data on my first search. I have to run the search **twice** to get the right answer.\r\n\r\nI am indexing data that looks like the following from the HF SQuAD 2.0 data set:\r\n\r\n```\r\n['57318658e6313a140071d02b',\r\n '56f7165e3d8e2e1400e3733a',\r\n '570e2f6e0b85d914000d7d21',\r\n '5727e58aff5b5019007d97d0',\r\n '5a3b5a503ff257001ab8441f',\r\n '57262fab271a42140099d725']\r\n```\r\n\r\n\r\n\r\nTo reproduce the issue, try:\r\n\r\n```\r\nfrom datasets import load_dataset, load_metric\r\nfrom transformers import BertTokenizerFast, BertForQuestionAnswering\r\nfrom elasticsearch import Elasticsearch\r\nimport numpy as np\r\nimport collections\r\nfrom tqdm.auto import tqdm\r\nimport torch\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-uncased')\r\nmax_length = 384 # The maximum length of a feature (question and context)\r\ndoc_stride = 128 # The authorized overlap between two part of the context when splitting it is needed.\r\npad_on_right = tokenizer.padding_side == \"right\"\r\nsquad_v2 = True\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef prepare_validation_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n # We keep the example_id that gave us this feature and we will store the offset mappings.\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token\r\n # position is part of the context or not.\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (list(o) if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n\r\n\r\n\r\n\r\n# build base examples, features set of training data\r\nshuffled_idx = pd.read_csv('https:\/\/raw.githubusercontent.com\/afogarty85\/temp\/main\/idx.csv')['idx'].to_list()\r\nexamples = load_dataset(\"squad_v2\").shuffle(seed=1)['train']\r\nfeatures = load_dataset(\"squad_v2\").shuffle(seed=1)['train'].map(\r\n prepare_validation_features,\r\n batched=True,\r\n remove_columns=['answers', 'context', 'id', 'question', 'title'])\r\n# reorder features by the training process\r\nfeatures = features.select(indices=shuffled_idx)\r\n# get the example ids to match with the \"example\" data; get unique entries\r\nid_list = list(dict.fromkeys(features['example_id']))\r\n# now search for their index positions in the examples data set; load elastic search\r\nes = Elasticsearch([{'host': 'localhost'}]).ping()\r\n# add an index to the id column for the examples\r\nexamples.add_elasticsearch_index(column='id')\r\n# retrieve the example index\r\nexample_idx_k1 = [examples.search(index_name='id', query=i, k=1).indices for i in id_list]\r\nexample_idx_k1 = [item for sublist in example_idx_k1 for item in sublist]\r\n\r\nexample_idx_k2 = [examples.search(index_name='id', query=i, k=3).indices for i in id_list]\r\nexample_idx_k2 = [item for sublist in example_idx_k2 for item in sublist]\r\n\r\nlen(example_idx_k1) # should be 130319\r\nlen(example_idx_k2) # should be 130319\r\n\r\n#trial 1 lengths:\r\n# k=1: 130314\r\n# k=3: 130319\r\n\r\n# trial 2:\r\n# just run k=3 first: 130310\r\n# try k=1 after k=3: 130319\r\n```\r\n\r\n\nHi !\r\nI tried your code on my side and I was able to workaround this issue by waiting a few seconds before querying the index.\r\nMaybe this is because the index is not updated yet on the ElasticSearch side ?","embeddings":[0.2319182903,-0.0181096476,-0.1212057546,0.1662061661,0.1539811343,-0.3275492191,-0.1240265667,0.1151942611,-0.2562256157,0.223544389,-0.0919988006,-0.0185625739,-0.0620763004,-0.5503269434,-0.2119903266,0.047404211,-0.1246098429,0.1536926925,0.032147754,-0.3201370835,-0.0627460554,0.0043781945,-0.2288940251,0.0211867131,-0.4899494946,-0.0326173268,-0.0106744906,0.0597604848,-0.2692687809,-0.2286838889,0.5223035216,0.0061602998,-0.0448917672,0.3566158414,-0.0001214736,0.0496643707,0.1778374165,-0.1051905826,-0.1969797164,-0.3112073839,-0.1039012298,0.0208721925,0.2502502203,-0.1207282096,-0.2994047105,-0.0313754082,0.0162320342,-0.4722240865,0.4968915582,0.1408977509,0.1539392173,0.0794691965,-0.2473961115,-0.0978696719,0.531675458,0.1491986066,0.0635458529,-0.5371038318,0.3572324812,0.04666676,-0.1673317105,0.2593125403,-0.0130433552,0.1279951185,0.0269338954,0.3127305806,-0.4903326035,-0.1085786521,0.101237677,0.0566994436,0.2948672771,0.1065647304,-0.5151871443,-0.1737106889,0.0258475281,0.1346712261,0.0453901738,-0.0322916657,0.0384985842,-0.0453942157,-0.3651883304,0.0745277181,-0.1190353557,0.2235465646,-0.3281975389,0.358646065,0.0217059273,-0.0825723931,0.15646182,-0.2332246006,0.3354169726,0.1929093152,0.0828036293,0.2111808211,-0.3941687047,0.014127735,0.1393809021,-0.3425761163,0.1961891055,-0.1744431704,0.1355975121,0.215233162,-0.1107876822,-0.132972002,0.1402100027,0.3549906313,0.071560964,0.6889232397,0.0562935285,-0.3481680155,-0.4805217087,-0.1533411592,0.0640205517,0.1087686941,0.3282479942,-0.1395204663,-0.1512972265,-0.2715940475,-0.3633507192,0.2554884851,-0.7213668227,-0.2695736289,0.1087374762,0.2101597488,0.0075201592,-0.2276466191,0.2753213048,0.1703205109,-0.0937597007,-0.0165956244,-0.2273170501,-0.2755911946,-0.1744347513,0.1264992654,0.29306373,-0.3172920346,0.3059695065,-0.0172565505,0.3794793189,-0.2384904027,0.227548793,-0.1642407924,0.5017402172,-0.0497204028,-0.0622424074,0.1019674763,-0.2309207618,-0.0008214309,0.0803700611,-0.0531149879,-0.3773986101,-0.1401287466,0.0950885341,0.1630227715,0.0189064592,0.0635307655,0.0570505597,0.1438114196,-0.0320245512,0.2286981344,0.031667266,-0.2421645373,-0.0283871461,0.1248354539,0.2620407939,0.0838135779,0.0766112506,0.0478054099,-0.2097933292,0.5392237306,-0.1937483549,0.5320905447,-0.3519343138,0.3588716388,-0.1959837526,0.453209579,0.3438785076,-0.3066939116,-0.4178410172,-0.0070617716,0.0701471195,0.0818505734,-0.1956540048,-0.0772005543,0.5850794315,0.0584456325,0.1684946418,-0.1058119684,-0.0147512406,-0.179843232,-0.2057993412,0.0793086067,-0.132401973,0.122537069,0.3650589585,-0.3190761507,0.2059546113,0.200201124,0.103379041,0.1383926272,0.1309669018,0.1993986517,0.1290268302,0.245124653,0.1009816676,-0.0176461246,-0.1045701206,0.3578746319,-0.1139287204,-0.1758185178,-0.1187240481,0.0789473206,-0.0378453694,-0.1796457022,-0.3683208823,0.1106185243,0.0799578205,-0.0667589158,-0.1576201916,0.1252178252,0.0373865589,0.09110789,-0.1036002338,-0.0239481311,-0.1346371621,0.5873185992,0.1401750892,0.0618472695,-0.128961727,0.0682050437,0.3323591053,0.14089486,-0.0288241003,0.2554713786,-0.0950421542,-0.0953587368,0.242970109,0.0621274449,0.0064191972,-0.4569573104,0.0902888328,0.1820322573,-0.0453461483,-0.1043459103,-0.1893620938,0.2133301646,-0.10961283,0.297929436,-0.1006127596,-0.3357075751,0.0486052558,0.0263248794,-0.3719770312,-0.1223485693,0.1710480154,-0.2374024838,0.0927953869,-0.162245214,-0.2991945148,-0.1154294983,0.3506642878,0.0298579559,0.0070000105,0.2359534949,0.1895967275,0.1017938852,-0.2224409282,-0.0980562493,0.1962488294,0.1883406043,0.161173895,-0.1316046864,-0.1498198658,0.0884239078,-0.1883002371,0.18649593,0.0751937702,0.3247421086,0.3557135463,0.1353528649,-0.1519439965,-0.0239975769,0.0758682191,-0.1002445817,-0.0602428578,0.0886291414,-0.5449981689,0.1838710755,-0.2592953444,-0.0696897656,-0.0694548115,-0.0686879158,0.0790016353,0.0417687818,0.1969467252,-0.0094868578,-0.1913135052,-0.0404250994,0.3600957394,-0.1786870658,0.020867141,0.0950416252,-0.1861397028,-0.0053194547,-0.2531757653,0.1226850599,0.2429987639,0.1191356257,0.0263542347,-0.3971422613,-0.2332561314,0.21915932,0.2508250475,-0.0781759471,-0.0317705013,-0.3412439525,-0.7546483874,-0.2112495005,0.1506475955,-0.0891418234,-0.227286309,0.1767146289,0.0478027537,0.159980908,0.2062618881,-0.2788503766,-0.0083804009,0.159673214,-0.1973229349,-0.1182057559,-0.1904225796,0.2377604246,-0.213734448,0.0460078493,-0.3016382754,-0.0945709199,-0.3605505824,-0.230144605,0.3662579954,-0.0772942156,-0.0925500989,-0.3487813771,-0.2428897172,-0.3498476744,0.6523901224,-0.3380325437,0.1325749904,0.1887347996,-0.0874462873,-0.3462240398,0.1892597526,0.212760672,-0.2267203629,-0.0007457668,-0.1576300263,-0.0001613272,0.2392983139,0.4229534566,0.4874313176,-0.1483017951,0.1631302536,0.3141477704,0.9544945359,0.1182295457,-0.0469501838,0.0025962624,0.2337536961,0.0497342944,-0.1628332883,-0.2231667638,-0.0578453988,-0.0823696181,-0.0777804852,0.136577487,0.0194904916,-0.0144602666,-0.0024008469,0.2149406374,-0.4264387488,-0.2490727156,-0.0195305925,-0.0126173571,0.3440343142,0.2669635713,0.6866026521,0.2155230194,-0.2718796134,-0.4153939784,-0.4057601094,0.0981000513,0.0687077343,0.2006708384,-0.2229739875,-0.3778162599,0.1756176502,0.1501041502,0.4039020538,0.3343907595,0.1257066131,0.2573717535,0.3225034177,0.466912657,0.0720628947,0.0438119099,-0.1328052282,-0.334066391,-0.0256501567,-0.1807622612,-0.3394303024,0.2815589607,-0.1100582555,0.4821234941,-0.255040437,-0.0481168292,-0.2062539607,-0.0797446147,-0.1733807027,-0.1570751518,-0.1262797117,-0.0224325582,0.4037593007,0.1792227328,0.1237856224,0.1246491224,-0.0701937005,-0.1975953728,-0.2033241242,0.0158267729,0.1126544029,-0.0686699748,0.571734786,-0.0629523024,0.3072263002,0.7875511646,-0.0241908692,0.1152012572,0.3683220148,-0.2191747129,0.4279131293,-0.0050387783,-0.1256173849,-0.0697352961,0.3411696255,-0.0148837045,0.0142013542,-0.0078938296,0.0300535709,0.0397817902,-0.0877641067,0.5422750711,-0.3086483181,-0.3892702758,-0.0221062507,0.3651223481,-0.3544043601,-0.0099978996,0.0689876452,0.5028800368,-0.3959294856,0.2949846387,0.292887181,0.5569580197,-0.0255054012,-0.1178328022,-0.1432171017,0.1455268413,0.3744358122,-0.2167769969,0.4011741579,-0.179035753,-0.4535493553,-0.1155648604,-0.0676670671,-0.2743688524,0.4145802855,-0.1395337582,0.3018550873,0.2184752077,0.1198493168,-0.110799782,0.3054760993,0.3888858259,-0.048405014,-0.0716838464,0.097501792,0.0452106409,0.3173329234,0.0045349058,0.0331101082,0.1601284444,-0.2046093941,-0.2156311721,0.071578145,-0.33221066,0.3397744596,0.5011976361,-0.4189105332,0.1553664058,-0.3061593175,0.0958516002,0.0976354778,-0.1372382194,0.1370802373,0.3023985326,0.133386597,0.3347493112,-0.1124004498,0.1055382043,0.1609679908,-0.2286613137,-0.1176993847,0.2183936536,-0.4788444936,-0.2822570503,0.1897849292,-0.1204191223,-0.1570332497,0.0703280792,0.2056447715,-0.0054782266,-0.0223848298,0.0564024039,0.3085350096,-0.012983093,0.5014365911,0.4594114721,-0.2057445347,-0.0384186469,0.6241492033,-0.0392480828,0.0801608562,0.1555704325,-0.0122021614,-0.2864847183,-0.0284059066,0.2435368001,0.298992902,-0.1358129978,-0.248971507,-0.5415066481,0.334010303,0.4406498969,0.1384904683,0.1985381246,-0.3680636287,-0.3013118804,-0.0576692335,0.2907001078,0.2671514153,0.4164357185,-0.0426881686,-0.2128512114,0.069795616,-0.0651044101,0.105298318,-0.2458114326,0.1288781315,-0.066374898,0.1809228808,-0.6103871465,-0.1602731794,0.3117671609,-0.0190640353,0.053106714,0.1648203284,-0.1333865523,-0.1044773608,0.0219222754,0.1273991317,0.0057218219,0.2300191373,-0.236671418,-0.3809963167,-0.0191565342,-0.2873553038,-0.172226876,0.5963318944,0.33916834,0.1285149753,0.1607302427,-0.2684528232,0.1595602632,-0.0520422943,0.0400286056,-0.0280343611,0.0801938996,0.2192452848,0.2671088576,-0.0457937084,-0.2076483071,0.0715227574,-0.0080309268,0.0070218602,0.3336579502,-0.3741815686,-0.0541522168,0.2127549201,0.5488230586,0.4308081865,-0.2640267313,-0.5410853028,0.1240553632,0.1717007458,-0.1271345019,-0.3978877366,0.0317804702,0.341398716,-0.08805044,0.2219772488,-0.1419967711,-0.2504144907,0.1384326667,0.2344424427,0.3853815496,-0.2593209445,0.0194244348,-0.0569816642,0.1952126473,-0.2900905311,0.202322945,-0.1742208451,0.1853569597,0.3222083151,0.2221049517,0.4281585813,-0.1703189611,0.1965885609,0.1464491189,-0.171550259,-0.0959243774,0.3868349791,-0.1653012633,0.1305285841,-0.2231541574,0.3798935115,-0.0463735498,-0.0868945494,-0.3670883477,-0.1713474095,-0.0548603907,0.2007000297,0.2769956589,-0.3592944443,-0.2113573402,-0.2130507231,-0.1868733615,0.1379600763,-0.1687571108,0.0883286595,0.014661368,-0.1809162199,-0.182818681,-0.1264399588,0.2019327432,-0.2738080323,0.3066410124,0.1707685143,0.012430273,0.2588489056,0.1947530806,-0.1559538841,0.2856386304,-0.433825165,0.2635906935,-0.0902262181,-0.2458146662,-0.0114255082,0.0880511478,-0.0440075845,-0.7379398942,0.2653433084,0.0217674114,-0.0454740003,0.0760260001,0.170929268,0.3233115375,-0.2415288836,-0.074650228,0.0779387504,0.3614705801,0.0754794553,-0.1379213929,-0.3765068352,-0.0245188121,-0.1823033094,-0.0379734971,0.1886950582,0.08247336,0.0594121553,-0.220564425,0.1172611713,0.1997216493,-0.0821480826,-0.180922851,-0.1568582952,-0.4860467911,0.0663298741,-0.3449667096,0.2042013705,0.4468431175,0.5287171602,-0.1400733292,-0.1025343612,0.3878425658,-0.4869925976,-0.1354864687,0.2283742279,-0.2258522213,-0.166597262,-0.2700258493,-0.0003964539,-0.0444552675,-0.0698095858,0.043266736,-0.1307406425,0.0677453652,0.0380585827,0.0204470102,0.0603121854,-0.0367266461,0.2781249285,0.1148319542,-0.0658263341,0.1099122912,0.1881407052,0.070943214,-0.210319519,-0.0628601313,-0.0054699923,0.0735880211,0.2926050127,-0.0658663064,-0.3773882091,-0.1498064399,0.0059065097,-0.0139134098,-0.0626696423,-0.4696812034,0.009010816,-0.1425819546,0.0877599418,-0.3129696846,0.3146182597,-0.0055828989,0.3284255862,-0.1333074421,-0.2391646802,0.7692409754,-0.3602598906,-0.0566504039,-0.3632564247,0.2906856239,0.5245938897,0.0564526506,-0.3426836729,-0.136487633,0.4094573259,-0.0500620455,0.0035277961,-0.2560936213,0.164354682,0.0586028844,-0.1332415193,-0.0374308452,0.2646919489,-0.0371248461,-0.0444123894,-0.0341039598]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1758","title":"dataset.search() (elastic) cannot reliably retrieve search results","comments":"Thanks for the feedback! I added a 30 second \"sleep\" and that seemed to work well!","body":"I am trying to use elastic search to retrieve the indices of items in the dataset in their precise order, given shuffled training indices.\r\n\r\nThe problem I have is that I cannot retrieve reliable results with my data on my first search. I have to run the search **twice** to get the right answer.\r\n\r\nI am indexing data that looks like the following from the HF SQuAD 2.0 data set:\r\n\r\n```\r\n['57318658e6313a140071d02b',\r\n '56f7165e3d8e2e1400e3733a',\r\n '570e2f6e0b85d914000d7d21',\r\n '5727e58aff5b5019007d97d0',\r\n '5a3b5a503ff257001ab8441f',\r\n '57262fab271a42140099d725']\r\n```\r\n\r\n\r\n\r\nTo reproduce the issue, try:\r\n\r\n```\r\nfrom datasets import load_dataset, load_metric\r\nfrom transformers import BertTokenizerFast, BertForQuestionAnswering\r\nfrom elasticsearch import Elasticsearch\r\nimport numpy as np\r\nimport collections\r\nfrom tqdm.auto import tqdm\r\nimport torch\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-uncased')\r\nmax_length = 384 # The maximum length of a feature (question and context)\r\ndoc_stride = 128 # The authorized overlap between two part of the context when splitting it is needed.\r\npad_on_right = tokenizer.padding_side == \"right\"\r\nsquad_v2 = True\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef prepare_validation_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n # We keep the example_id that gave us this feature and we will store the offset mappings.\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token\r\n # position is part of the context or not.\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (list(o) if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n\r\n\r\n\r\n\r\n# build base examples, features set of training data\r\nshuffled_idx = pd.read_csv('https:\/\/raw.githubusercontent.com\/afogarty85\/temp\/main\/idx.csv')['idx'].to_list()\r\nexamples = load_dataset(\"squad_v2\").shuffle(seed=1)['train']\r\nfeatures = load_dataset(\"squad_v2\").shuffle(seed=1)['train'].map(\r\n prepare_validation_features,\r\n batched=True,\r\n remove_columns=['answers', 'context', 'id', 'question', 'title'])\r\n# reorder features by the training process\r\nfeatures = features.select(indices=shuffled_idx)\r\n# get the example ids to match with the \"example\" data; get unique entries\r\nid_list = list(dict.fromkeys(features['example_id']))\r\n# now search for their index positions in the examples data set; load elastic search\r\nes = Elasticsearch([{'host': 'localhost'}]).ping()\r\n# add an index to the id column for the examples\r\nexamples.add_elasticsearch_index(column='id')\r\n# retrieve the example index\r\nexample_idx_k1 = [examples.search(index_name='id', query=i, k=1).indices for i in id_list]\r\nexample_idx_k1 = [item for sublist in example_idx_k1 for item in sublist]\r\n\r\nexample_idx_k2 = [examples.search(index_name='id', query=i, k=3).indices for i in id_list]\r\nexample_idx_k2 = [item for sublist in example_idx_k2 for item in sublist]\r\n\r\nlen(example_idx_k1) # should be 130319\r\nlen(example_idx_k2) # should be 130319\r\n\r\n#trial 1 lengths:\r\n# k=1: 130314\r\n# k=3: 130319\r\n\r\n# trial 2:\r\n# just run k=3 first: 130310\r\n# try k=1 after k=3: 130319\r\n```\r\n\r\n","comment_length":16,"text":"dataset.search() (elastic) cannot reliably retrieve search results\nI am trying to use elastic search to retrieve the indices of items in the dataset in their precise order, given shuffled training indices.\r\n\r\nThe problem I have is that I cannot retrieve reliable results with my data on my first search. I have to run the search **twice** to get the right answer.\r\n\r\nI am indexing data that looks like the following from the HF SQuAD 2.0 data set:\r\n\r\n```\r\n['57318658e6313a140071d02b',\r\n '56f7165e3d8e2e1400e3733a',\r\n '570e2f6e0b85d914000d7d21',\r\n '5727e58aff5b5019007d97d0',\r\n '5a3b5a503ff257001ab8441f',\r\n '57262fab271a42140099d725']\r\n```\r\n\r\n\r\n\r\nTo reproduce the issue, try:\r\n\r\n```\r\nfrom datasets import load_dataset, load_metric\r\nfrom transformers import BertTokenizerFast, BertForQuestionAnswering\r\nfrom elasticsearch import Elasticsearch\r\nimport numpy as np\r\nimport collections\r\nfrom tqdm.auto import tqdm\r\nimport torch\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-uncased')\r\nmax_length = 384 # The maximum length of a feature (question and context)\r\ndoc_stride = 128 # The authorized overlap between two part of the context when splitting it is needed.\r\npad_on_right = tokenizer.padding_side == \"right\"\r\nsquad_v2 = True\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef prepare_validation_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n # We keep the example_id that gave us this feature and we will store the offset mappings.\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token\r\n # position is part of the context or not.\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (list(o) if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n\r\n\r\n\r\n\r\n# build base examples, features set of training data\r\nshuffled_idx = pd.read_csv('https:\/\/raw.githubusercontent.com\/afogarty85\/temp\/main\/idx.csv')['idx'].to_list()\r\nexamples = load_dataset(\"squad_v2\").shuffle(seed=1)['train']\r\nfeatures = load_dataset(\"squad_v2\").shuffle(seed=1)['train'].map(\r\n prepare_validation_features,\r\n batched=True,\r\n remove_columns=['answers', 'context', 'id', 'question', 'title'])\r\n# reorder features by the training process\r\nfeatures = features.select(indices=shuffled_idx)\r\n# get the example ids to match with the \"example\" data; get unique entries\r\nid_list = list(dict.fromkeys(features['example_id']))\r\n# now search for their index positions in the examples data set; load elastic search\r\nes = Elasticsearch([{'host': 'localhost'}]).ping()\r\n# add an index to the id column for the examples\r\nexamples.add_elasticsearch_index(column='id')\r\n# retrieve the example index\r\nexample_idx_k1 = [examples.search(index_name='id', query=i, k=1).indices for i in id_list]\r\nexample_idx_k1 = [item for sublist in example_idx_k1 for item in sublist]\r\n\r\nexample_idx_k2 = [examples.search(index_name='id', query=i, k=3).indices for i in id_list]\r\nexample_idx_k2 = [item for sublist in example_idx_k2 for item in sublist]\r\n\r\nlen(example_idx_k1) # should be 130319\r\nlen(example_idx_k2) # should be 130319\r\n\r\n#trial 1 lengths:\r\n# k=1: 130314\r\n# k=3: 130319\r\n\r\n# trial 2:\r\n# just run k=3 first: 130310\r\n# try k=1 after k=3: 130319\r\n```\r\n\r\n\nThanks for the feedback! I added a 30 second \"sleep\" and that seemed to work well!","embeddings":[0.2319182903,-0.0181096476,-0.1212057546,0.1662061661,0.1539811343,-0.3275492191,-0.1240265667,0.1151942611,-0.2562256157,0.223544389,-0.0919988006,-0.0185625739,-0.0620763004,-0.5503269434,-0.2119903266,0.047404211,-0.1246098429,0.1536926925,0.032147754,-0.3201370835,-0.0627460554,0.0043781945,-0.2288940251,0.0211867131,-0.4899494946,-0.0326173268,-0.0106744906,0.0597604848,-0.2692687809,-0.2286838889,0.5223035216,0.0061602998,-0.0448917672,0.3566158414,-0.0001214736,0.0496643707,0.1778374165,-0.1051905826,-0.1969797164,-0.3112073839,-0.1039012298,0.0208721925,0.2502502203,-0.1207282096,-0.2994047105,-0.0313754082,0.0162320342,-0.4722240865,0.4968915582,0.1408977509,0.1539392173,0.0794691965,-0.2473961115,-0.0978696719,0.531675458,0.1491986066,0.0635458529,-0.5371038318,0.3572324812,0.04666676,-0.1673317105,0.2593125403,-0.0130433552,0.1279951185,0.0269338954,0.3127305806,-0.4903326035,-0.1085786521,0.101237677,0.0566994436,0.2948672771,0.1065647304,-0.5151871443,-0.1737106889,0.0258475281,0.1346712261,0.0453901738,-0.0322916657,0.0384985842,-0.0453942157,-0.3651883304,0.0745277181,-0.1190353557,0.2235465646,-0.3281975389,0.358646065,0.0217059273,-0.0825723931,0.15646182,-0.2332246006,0.3354169726,0.1929093152,0.0828036293,0.2111808211,-0.3941687047,0.014127735,0.1393809021,-0.3425761163,0.1961891055,-0.1744431704,0.1355975121,0.215233162,-0.1107876822,-0.132972002,0.1402100027,0.3549906313,0.071560964,0.6889232397,0.0562935285,-0.3481680155,-0.4805217087,-0.1533411592,0.0640205517,0.1087686941,0.3282479942,-0.1395204663,-0.1512972265,-0.2715940475,-0.3633507192,0.2554884851,-0.7213668227,-0.2695736289,0.1087374762,0.2101597488,0.0075201592,-0.2276466191,0.2753213048,0.1703205109,-0.0937597007,-0.0165956244,-0.2273170501,-0.2755911946,-0.1744347513,0.1264992654,0.29306373,-0.3172920346,0.3059695065,-0.0172565505,0.3794793189,-0.2384904027,0.227548793,-0.1642407924,0.5017402172,-0.0497204028,-0.0622424074,0.1019674763,-0.2309207618,-0.0008214309,0.0803700611,-0.0531149879,-0.3773986101,-0.1401287466,0.0950885341,0.1630227715,0.0189064592,0.0635307655,0.0570505597,0.1438114196,-0.0320245512,0.2286981344,0.031667266,-0.2421645373,-0.0283871461,0.1248354539,0.2620407939,0.0838135779,0.0766112506,0.0478054099,-0.2097933292,0.5392237306,-0.1937483549,0.5320905447,-0.3519343138,0.3588716388,-0.1959837526,0.453209579,0.3438785076,-0.3066939116,-0.4178410172,-0.0070617716,0.0701471195,0.0818505734,-0.1956540048,-0.0772005543,0.5850794315,0.0584456325,0.1684946418,-0.1058119684,-0.0147512406,-0.179843232,-0.2057993412,0.0793086067,-0.132401973,0.122537069,0.3650589585,-0.3190761507,0.2059546113,0.200201124,0.103379041,0.1383926272,0.1309669018,0.1993986517,0.1290268302,0.245124653,0.1009816676,-0.0176461246,-0.1045701206,0.3578746319,-0.1139287204,-0.1758185178,-0.1187240481,0.0789473206,-0.0378453694,-0.1796457022,-0.3683208823,0.1106185243,0.0799578205,-0.0667589158,-0.1576201916,0.1252178252,0.0373865589,0.09110789,-0.1036002338,-0.0239481311,-0.1346371621,0.5873185992,0.1401750892,0.0618472695,-0.128961727,0.0682050437,0.3323591053,0.14089486,-0.0288241003,0.2554713786,-0.0950421542,-0.0953587368,0.242970109,0.0621274449,0.0064191972,-0.4569573104,0.0902888328,0.1820322573,-0.0453461483,-0.1043459103,-0.1893620938,0.2133301646,-0.10961283,0.297929436,-0.1006127596,-0.3357075751,0.0486052558,0.0263248794,-0.3719770312,-0.1223485693,0.1710480154,-0.2374024838,0.0927953869,-0.162245214,-0.2991945148,-0.1154294983,0.3506642878,0.0298579559,0.0070000105,0.2359534949,0.1895967275,0.1017938852,-0.2224409282,-0.0980562493,0.1962488294,0.1883406043,0.161173895,-0.1316046864,-0.1498198658,0.0884239078,-0.1883002371,0.18649593,0.0751937702,0.3247421086,0.3557135463,0.1353528649,-0.1519439965,-0.0239975769,0.0758682191,-0.1002445817,-0.0602428578,0.0886291414,-0.5449981689,0.1838710755,-0.2592953444,-0.0696897656,-0.0694548115,-0.0686879158,0.0790016353,0.0417687818,0.1969467252,-0.0094868578,-0.1913135052,-0.0404250994,0.3600957394,-0.1786870658,0.020867141,0.0950416252,-0.1861397028,-0.0053194547,-0.2531757653,0.1226850599,0.2429987639,0.1191356257,0.0263542347,-0.3971422613,-0.2332561314,0.21915932,0.2508250475,-0.0781759471,-0.0317705013,-0.3412439525,-0.7546483874,-0.2112495005,0.1506475955,-0.0891418234,-0.227286309,0.1767146289,0.0478027537,0.159980908,0.2062618881,-0.2788503766,-0.0083804009,0.159673214,-0.1973229349,-0.1182057559,-0.1904225796,0.2377604246,-0.213734448,0.0460078493,-0.3016382754,-0.0945709199,-0.3605505824,-0.230144605,0.3662579954,-0.0772942156,-0.0925500989,-0.3487813771,-0.2428897172,-0.3498476744,0.6523901224,-0.3380325437,0.1325749904,0.1887347996,-0.0874462873,-0.3462240398,0.1892597526,0.212760672,-0.2267203629,-0.0007457668,-0.1576300263,-0.0001613272,0.2392983139,0.4229534566,0.4874313176,-0.1483017951,0.1631302536,0.3141477704,0.9544945359,0.1182295457,-0.0469501838,0.0025962624,0.2337536961,0.0497342944,-0.1628332883,-0.2231667638,-0.0578453988,-0.0823696181,-0.0777804852,0.136577487,0.0194904916,-0.0144602666,-0.0024008469,0.2149406374,-0.4264387488,-0.2490727156,-0.0195305925,-0.0126173571,0.3440343142,0.2669635713,0.6866026521,0.2155230194,-0.2718796134,-0.4153939784,-0.4057601094,0.0981000513,0.0687077343,0.2006708384,-0.2229739875,-0.3778162599,0.1756176502,0.1501041502,0.4039020538,0.3343907595,0.1257066131,0.2573717535,0.3225034177,0.466912657,0.0720628947,0.0438119099,-0.1328052282,-0.334066391,-0.0256501567,-0.1807622612,-0.3394303024,0.2815589607,-0.1100582555,0.4821234941,-0.255040437,-0.0481168292,-0.2062539607,-0.0797446147,-0.1733807027,-0.1570751518,-0.1262797117,-0.0224325582,0.4037593007,0.1792227328,0.1237856224,0.1246491224,-0.0701937005,-0.1975953728,-0.2033241242,0.0158267729,0.1126544029,-0.0686699748,0.571734786,-0.0629523024,0.3072263002,0.7875511646,-0.0241908692,0.1152012572,0.3683220148,-0.2191747129,0.4279131293,-0.0050387783,-0.1256173849,-0.0697352961,0.3411696255,-0.0148837045,0.0142013542,-0.0078938296,0.0300535709,0.0397817902,-0.0877641067,0.5422750711,-0.3086483181,-0.3892702758,-0.0221062507,0.3651223481,-0.3544043601,-0.0099978996,0.0689876452,0.5028800368,-0.3959294856,0.2949846387,0.292887181,0.5569580197,-0.0255054012,-0.1178328022,-0.1432171017,0.1455268413,0.3744358122,-0.2167769969,0.4011741579,-0.179035753,-0.4535493553,-0.1155648604,-0.0676670671,-0.2743688524,0.4145802855,-0.1395337582,0.3018550873,0.2184752077,0.1198493168,-0.110799782,0.3054760993,0.3888858259,-0.048405014,-0.0716838464,0.097501792,0.0452106409,0.3173329234,0.0045349058,0.0331101082,0.1601284444,-0.2046093941,-0.2156311721,0.071578145,-0.33221066,0.3397744596,0.5011976361,-0.4189105332,0.1553664058,-0.3061593175,0.0958516002,0.0976354778,-0.1372382194,0.1370802373,0.3023985326,0.133386597,0.3347493112,-0.1124004498,0.1055382043,0.1609679908,-0.2286613137,-0.1176993847,0.2183936536,-0.4788444936,-0.2822570503,0.1897849292,-0.1204191223,-0.1570332497,0.0703280792,0.2056447715,-0.0054782266,-0.0223848298,0.0564024039,0.3085350096,-0.012983093,0.5014365911,0.4594114721,-0.2057445347,-0.0384186469,0.6241492033,-0.0392480828,0.0801608562,0.1555704325,-0.0122021614,-0.2864847183,-0.0284059066,0.2435368001,0.298992902,-0.1358129978,-0.248971507,-0.5415066481,0.334010303,0.4406498969,0.1384904683,0.1985381246,-0.3680636287,-0.3013118804,-0.0576692335,0.2907001078,0.2671514153,0.4164357185,-0.0426881686,-0.2128512114,0.069795616,-0.0651044101,0.105298318,-0.2458114326,0.1288781315,-0.066374898,0.1809228808,-0.6103871465,-0.1602731794,0.3117671609,-0.0190640353,0.053106714,0.1648203284,-0.1333865523,-0.1044773608,0.0219222754,0.1273991317,0.0057218219,0.2300191373,-0.236671418,-0.3809963167,-0.0191565342,-0.2873553038,-0.172226876,0.5963318944,0.33916834,0.1285149753,0.1607302427,-0.2684528232,0.1595602632,-0.0520422943,0.0400286056,-0.0280343611,0.0801938996,0.2192452848,0.2671088576,-0.0457937084,-0.2076483071,0.0715227574,-0.0080309268,0.0070218602,0.3336579502,-0.3741815686,-0.0541522168,0.2127549201,0.5488230586,0.4308081865,-0.2640267313,-0.5410853028,0.1240553632,0.1717007458,-0.1271345019,-0.3978877366,0.0317804702,0.341398716,-0.08805044,0.2219772488,-0.1419967711,-0.2504144907,0.1384326667,0.2344424427,0.3853815496,-0.2593209445,0.0194244348,-0.0569816642,0.1952126473,-0.2900905311,0.202322945,-0.1742208451,0.1853569597,0.3222083151,0.2221049517,0.4281585813,-0.1703189611,0.1965885609,0.1464491189,-0.171550259,-0.0959243774,0.3868349791,-0.1653012633,0.1305285841,-0.2231541574,0.3798935115,-0.0463735498,-0.0868945494,-0.3670883477,-0.1713474095,-0.0548603907,0.2007000297,0.2769956589,-0.3592944443,-0.2113573402,-0.2130507231,-0.1868733615,0.1379600763,-0.1687571108,0.0883286595,0.014661368,-0.1809162199,-0.182818681,-0.1264399588,0.2019327432,-0.2738080323,0.3066410124,0.1707685143,0.012430273,0.2588489056,0.1947530806,-0.1559538841,0.2856386304,-0.433825165,0.2635906935,-0.0902262181,-0.2458146662,-0.0114255082,0.0880511478,-0.0440075845,-0.7379398942,0.2653433084,0.0217674114,-0.0454740003,0.0760260001,0.170929268,0.3233115375,-0.2415288836,-0.074650228,0.0779387504,0.3614705801,0.0754794553,-0.1379213929,-0.3765068352,-0.0245188121,-0.1823033094,-0.0379734971,0.1886950582,0.08247336,0.0594121553,-0.220564425,0.1172611713,0.1997216493,-0.0821480826,-0.180922851,-0.1568582952,-0.4860467911,0.0663298741,-0.3449667096,0.2042013705,0.4468431175,0.5287171602,-0.1400733292,-0.1025343612,0.3878425658,-0.4869925976,-0.1354864687,0.2283742279,-0.2258522213,-0.166597262,-0.2700258493,-0.0003964539,-0.0444552675,-0.0698095858,0.043266736,-0.1307406425,0.0677453652,0.0380585827,0.0204470102,0.0603121854,-0.0367266461,0.2781249285,0.1148319542,-0.0658263341,0.1099122912,0.1881407052,0.070943214,-0.210319519,-0.0628601313,-0.0054699923,0.0735880211,0.2926050127,-0.0658663064,-0.3773882091,-0.1498064399,0.0059065097,-0.0139134098,-0.0626696423,-0.4696812034,0.009010816,-0.1425819546,0.0877599418,-0.3129696846,0.3146182597,-0.0055828989,0.3284255862,-0.1333074421,-0.2391646802,0.7692409754,-0.3602598906,-0.0566504039,-0.3632564247,0.2906856239,0.5245938897,0.0564526506,-0.3426836729,-0.136487633,0.4094573259,-0.0500620455,0.0035277961,-0.2560936213,0.164354682,0.0586028844,-0.1332415193,-0.0374308452,0.2646919489,-0.0371248461,-0.0444123894,-0.0341039598]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1757","title":"FewRel","comments":"+1","body":"## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":1,"text":"FewRel\n## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n+1","embeddings":[-0.2723810077,-0.0474956818,-0.1295970231,0.1361517161,0.0515405349,-0.0986004546,0.3482960761,0.2295035869,0.1168564036,0.2561810911,-0.3586153686,0.0880709961,-0.025458388,-0.1972101331,0.0731428638,-0.0231830683,-0.1357664615,0.280015558,0.0567615069,-0.1696253717,-0.224564001,0.1390003711,0.2075006813,-0.2117629647,-0.2893826663,-0.3589718044,-0.1765655577,0.2898524404,-0.2312083095,-0.2850160301,0.1688060313,0.3549354076,0.1335845292,0.3499802649,-0.0000998085,-0.107164748,0.1092357561,-0.0159993432,0.0814512298,0.1611503214,0.2071500272,-0.4546532333,-0.0617862716,-0.3060390353,-0.179855898,-0.2504825294,-0.1459655315,-0.1279021949,0.3483549953,0.315450877,0.3548479974,0.0648872554,0.1455044448,-0.4766403735,0.0516733825,0.1029927284,-0.1532540768,0.3063488603,0.3852418363,-0.056742806,-0.0791858584,0.102679044,0.0712254196,-0.1380878538,0.3735221028,-0.0062974822,-0.0131681645,-0.3446175158,0.0751930177,0.4047895968,0.1932726651,0.0706644207,-0.2176492363,-0.0034761161,0.0144925714,-0.1767764091,-0.1907782108,-0.1005330458,0.0044545773,0.1950269341,0.038833227,-0.1407747716,-0.318015486,0.2383707315,-0.017287856,0.1576220244,-0.0778219178,-0.0833898038,0.1144268438,0.0347687826,-0.3539718688,0.0780153722,-0.1082817242,0.0992972255,-0.0819370151,-0.2823030353,0.143725127,0.2002937496,0.469699502,0.2748092711,-0.0579027385,0.0022586707,0.0814069957,0.2448265553,-0.0338758603,-0.0943446234,0.1001181081,-0.0559764802,0.2081864476,0.0035050062,0.0313422568,0.0185803268,-0.1832010299,0.1954883188,-0.0050955876,-0.2814794481,-0.1402692497,-0.1495223343,-0.0453648381,0.2443869561,-0.0404771902,0.1732062399,-0.0769709051,0.4379487038,-0.0827760622,0.2845282555,-0.1268897653,0.09875606,0.1187595055,0.0294372421,-0.3026388288,0.1851992309,-0.0758290291,-0.1724320799,0.125623256,0.1234712973,0.2385923415,-0.2841027975,0.0681708828,-0.0074865008,0.0720570609,-0.1261488348,0.3382704258,0.1465108544,-0.1104760543,-0.1108860373,0.1458017081,-0.3001419902,-0.2421143204,-0.1544867158,-0.0454800501,-0.0732265934,-0.4262217283,0.3386029899,-0.0368868932,-0.1143028438,0.1336560398,0.3419979215,-0.024974145,0.0103514167,-0.0094824051,0.1752921045,-0.2733846605,-0.0821040049,0.0848003775,0.2279431522,-0.3432430029,-0.0742347613,-0.1842399985,-0.0217244513,0.1806471944,0.1776655018,-0.203580305,0.0974218398,0.0341970064,0.0712723136,0.4290264845,-0.128845796,-0.3529996872,0.4893999398,-0.3543075621,-0.0096303113,0.1579389721,0.4430541694,0.3485955894,-0.0414313525,0.099876672,0.2972692847,-0.2314456552,-0.0148434937,-0.202800706,-0.3034564555,-0.3649443984,-0.0510698855,0.4742319286,0.242831558,0.3011213839,-0.1258327216,0.3285387754,0.0340307653,0.2394602895,0.194770813,0.4801593125,0.1955171824,-0.0671554729,-0.2915151119,-0.1091032922,0.0184481982,-0.0238409955,0.1545566022,0.1109762415,-0.329803586,-0.2452510744,-0.0305077322,-0.0474557653,-0.2287196219,0.2706514597,0.2206039578,0.0610881038,-0.1946129799,-0.0777981356,0.2591598332,0.0020680584,-0.0275173802,-0.0679885298,0.1385726631,-0.3876523674,0.055996567,0.0744221583,0.2788981795,-0.0532980263,-0.1365576237,0.3236396611,0.3102966845,-0.1826233417,-0.1106800959,0.4110580087,-0.0074724103,0.023338642,-0.2138737589,-0.0300314128,-0.1156932041,0.107517533,0.1626448333,-0.0687326789,0.5069717765,0.3131361902,0.0874507502,0.0782807544,-0.0121233724,0.0882398412,-0.0815941021,0.1784182042,-0.390923351,0.1382941753,0.2548006475,-0.043861974,-0.0135686593,-0.3718822002,0.2456205487,-0.0072169942,0.0523163639,0.1062720492,0.146378547,0.0602528043,-0.0509959497,-0.1085523367,0.2038955241,0.2197173238,0.4907991886,-0.0484483764,-0.1089370325,0.0001978601,-0.1065166593,0.035927508,0.1983653903,0.246375367,0.0468878374,0.3290911317,0.1750499606,-0.3131481111,-0.198431313,-0.1340469122,0.2116525471,-0.0537346788,0.2190837264,0.2256359309,0.2507269382,-0.0574509054,0.14073053,0.0136249578,-0.2204598933,0.2318588942,0.0184594244,0.0222259723,0.2333923876,-0.4184170663,0.1802811027,-0.2936785519,0.2766708434,0.0204110015,-0.2107327878,-0.1375115663,0.3010191023,0.116143696,-0.0970225632,0.6850885153,0.1483391672,-0.0804833174,-0.4678079784,-0.6208962798,0.1674873233,-0.2433518916,0.366704464,0.1497223079,0.0269530118,-0.0113773393,-0.1790282577,0.1154012382,-0.4033825994,-0.2595626414,-0.1784507334,-0.362780422,0.0219537448,-0.0083235269,-0.3697687089,-0.2058861256,-0.3482291698,0.2662261128,0.0554956757,0.0978035331,0.201822564,0.0265343096,-0.0153567074,-0.0791330263,0.1327895373,-0.1919411719,-0.2148700356,0.1940324754,-0.3287808299,-0.4206026196,-0.0196463764,-0.2009013146,-0.050642021,-0.0735343248,-0.1291987151,-0.4173081517,-0.0381721146,0.2563761771,0.0991548896,-0.1816834807,0.2768501937,-0.2002765834,-0.3644879162,0.0066862763,0.2063759565,0.3696003854,-0.3823933601,0.1022265255,-0.2566468716,0.4167087078,-0.1742833108,0.2577904463,-0.0549586974,-0.2926952839,0.0685772523,-0.1366004795,0.3622861207,0.0360212363,-0.3021107018,0.201799944,-0.0804442763,0.0395421572,0.4250045121,-0.1616125107,-0.3883517683,-0.0617161505,0.083626315,-0.1899982244,0.0019632478,-0.0724996254,-0.0597609468,0.1825122982,-0.1149198711,-0.3152500093,-0.0604751259,-0.0914177671,0.0134468526,0.2907371223,-0.1248390898,-0.087923564,-0.6833950281,0.0914546251,-0.3592763841,0.2719997764,0.3110381663,-0.0901622474,-0.1674197316,-0.3240658045,0.0012107899,-0.1201352775,0.4118987322,-0.4165804088,-0.2485966384,-0.0040682494,0.1932500005,-0.1533976346,0.3438490033,-0.3342775702,0.0326903984,0.0971661061,0.2097039968,-0.2262218744,-0.19983311,0.583434701,0.0139627904,-0.0727253109,-0.0854116455,-0.2484313995,-0.2137704641,-0.1810698509,0.079146944,-0.0898152888,0.1363684386,0.1230330542,-0.0782305151,-0.0866723731,0.0350973643,0.38443923,0.0350436941,-0.1073866934,0.2390797883,0.2157514393,-0.0644966364,0.2282500267,0.5054059029,0.0707656667,-0.3933893144,0.010401533,0.0759763643,-0.2684245706,0.5016145706,0.1671835333,0.1170265526,0.1052339748,0.222084105,0.2233614773,0.028504204,0.2634692192,0.2797597647,-0.1548331976,-0.3211748004,-0.1394168139,0.2258566767,0.0497771017,0.0055510243,-0.2329281718,0.2749688625,-0.0001545254,-0.1475843638,0.1661816239,0.8515752554,-0.0014680212,0.035974931,0.0777421072,-0.3888760805,0.4805100262,-0.110165149,-0.0517684333,-0.3495276868,-0.1430525184,-0.1151111722,0.0593954399,0.1323983371,0.096863389,0.0596476533,-0.1094693244,-0.0322190188,-0.0532136112,-0.0015770765,0.2365860641,0.2945954204,-0.2658879161,-0.1132267937,0.3172930181,0.1457504183,-0.1490950733,0.001244447,-0.1172972843,-0.3955017626,-0.0528282933,-0.0229663178,-0.1564025283,-0.1929864138,0.1787601262,0.2646698356,-0.6266624928,-0.0071302322,-0.0988363996,0.154685691,0.0075736823,-0.0497147664,0.0577469692,-0.2278069109,-0.289085716,0.1700262874,-0.1544319838,0.2131282389,-0.0992112309,-0.1317117512,0.37120381,-0.0037532421,-0.0696427897,-0.2267644405,0.0561236143,0.0051111453,-0.0241528265,-0.0765325576,0.1326310486,-0.2103683203,-0.3432041407,0.2321896702,0.0849546418,-0.1594446301,-0.1240248978,0.0624400862,-0.0966024548,-0.1307185888,0.3253189921,0.2150614411,-0.0397289656,0.2022734284,0.1277732253,-0.2043286264,-0.4198526144,-0.1257780939,0.5014775991,-0.4762550294,-0.1439402699,0.0964367688,-0.3850698173,0.248270452,0.4557594061,0.1856952608,-0.1874591261,-0.1599625945,-0.1298168898,-0.501149416,0.1876544207,-0.3430403769,0.2204102129,-0.0192865431,0.3628471792,-0.19742091,0.2527810335,-0.5277906656,-0.0658432245,-0.1952628791,0.0871723518,-0.1400826573,-0.2639394701,0.1550604552,-0.041004803,0.231236577,-0.0237385519,-0.2236000001,-0.3356051743,-0.1954995394,0.0462245196,-0.1002096161,0.2498924136,0.1766132116,0.2250261009,-0.4136322141,0.0750662386,0.2403122634,-0.1813688129,0.056636963,-0.1424637735,-0.1609215587,-0.1682237685,-0.09279304,-0.1466750205,0.0795906931,0.0231203735,0.0376150161,-0.0573175065,-0.1340045929,-0.2186542153,0.0693505928,0.4233386517,0.2141890377,-0.0608506985,0.3303694427,-0.0663586482,-0.1572984606,0.7045446038,0.4063594341,-0.0191574041,0.0782544911,0.1158925667,-0.18353948,0.3995436728,-0.4160231352,-0.1049283445,-0.1928290874,-0.3978354335,0.330935806,0.4267205,0.1894477606,0.0123812817,0.0676311404,0.1397064179,0.4031330049,-0.2668554485,-0.1697755456,0.0634762794,0.2048854232,0.1522899419,0.4291450679,0.2694552243,0.4880066514,0.0632757992,0.0594102852,0.3599497974,-0.0249257684,0.3479652107,0.300598681,-0.0800448656,0.2232841998,0.0939563662,0.0965925679,-0.152059406,-0.1025091708,0.5751723051,0.0913282782,0.0740051791,-0.0311988574,0.2090848982,-0.1217218712,-0.0084340731,-0.0821769908,-0.261125952,-0.274946779,-0.231377095,0.0763359666,0.027395688,0.3030739725,-0.1421354115,-0.0536867045,-0.341796279,0.0289644394,0.0734242573,0.2890705168,-0.2764775455,0.1092654094,-0.0667820424,-0.1454632878,0.0810140744,0.2873893678,0.3881834745,-0.0681109875,0.2446983606,0.1042976603,-0.0690940022,-0.3485617638,-0.2870359719,0.0395615324,0.2181217074,-0.152624011,0.1170468926,0.2739205658,-0.3469251692,0.0507217795,0.2084899545,0.130910486,0.0826918706,0.262680769,0.1001416072,-0.3131832182,-0.1573269218,0.1491151303,-0.2949496508,0.1631710082,0.0587805435,-0.14241907,0.0180475693,-0.1322912276,0.1801296473,-0.0399511307,0.2762738168,0.1741073728,-0.0565305948,-0.2005084753,-0.2025280297,-0.525700748,0.2454850674,-0.0402686112,-0.1514442414,-0.0739259496,0.2410929054,0.2303050905,-0.0016204425,-0.0461590327,-0.1198723614,-0.0423525982,-0.1291580051,-0.3138329387,0.0253813192,0.0850664228,0.0812751651,0.2108699828,-0.1418056339,-0.1623129547,-0.0581402257,0.1706409305,-0.1928032041,-0.19386141,-0.0377724096,0.1040056273,0.2357244492,0.0275448337,0.4338408709,-0.2156859785,-0.2029349357,-0.1230352372,-0.1169038191,-0.3271255195,0.3687591851,0.2970775962,0.3706576526,-0.0417720415,-0.0277277827,0.1421367973,-0.0182252023,0.3234283626,0.1128680483,-0.3505707681,0.1866266578,0.0429438017,-0.1117237583,0.2118037343,0.2971824706,0.2717915475,-0.0301723517,-0.2220501006,-0.4679828882,0.4597323537,0.231015414,-0.0692674667,-0.0676039532,0.3143661916,0.0641673505,-0.1488430053,-0.4054516256,0.0866857246,0.403873086,-0.0122255953,-0.2630368769,0.2761768699,-0.0361607336,-0.0413668714,-0.0592905059,-0.1058662087,0.021452887,-0.2013196647,-0.0596649982,-0.0837045088]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1757","title":"FewRel","comments":"@dspoka Please check the following link : https:\/\/github.com\/thunlp\/FewRel\r\nThis link mentions two versions of the datasets. Also, this one seems to be the official link.\r\n\r\nI am assuming this is the correct link and implementing based on the same.","body":"## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":39,"text":"FewRel\n## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\n@dspoka Please check the following link : https:\/\/github.com\/thunlp\/FewRel\r\nThis link mentions two versions of the datasets. Also, this one seems to be the official link.\r\n\r\nI am assuming this is the correct link and implementing based on the same.","embeddings":[-0.1948349923,-0.0344234481,-0.1016143709,0.1268530041,-0.0582779087,-0.1250493228,0.4315415025,0.1708205491,0.1060709804,0.1873810887,-0.4312252402,0.0766163617,-0.0305741187,-0.3551288247,0.2079174072,-0.1194455251,-0.0137296366,0.2213974893,0.0144300517,-0.2777475715,-0.208515957,0.2417953908,0.0616172217,-0.1108966023,-0.2568879128,-0.3504311442,-0.2791405618,0.2126297504,-0.2346280068,-0.3258255422,0.3127015233,0.3423486054,0.236085698,0.4270271361,-0.000102641,-0.1168433502,0.1504649669,-0.0227653738,0.2300339341,0.065228872,0.2461844981,-0.581530571,0.0220020264,-0.270445466,-0.0693214983,-0.3054866791,-0.0593082197,-0.0659846961,0.3989920914,0.2640653849,0.3045372069,-0.0256966613,0.2329767197,-0.415032059,0.0251856335,0.0579345711,-0.1375132352,0.3970343471,0.406915307,-0.048483558,-0.0412055813,-0.0503174476,0.1264362484,-0.1019397005,0.3068552613,-0.0098813819,-0.073044695,-0.5875650048,0.0355400294,0.4702458978,0.3035934567,0.1698890477,-0.2683259547,0.104188703,-0.0309116635,-0.1546928436,-0.1621081382,-0.0901775435,-0.0814457089,0.0900968611,-0.0095678195,-0.1053602919,-0.3569051623,0.2582896054,-0.1199415326,0.1873029619,-0.1128008142,-0.0755117759,0.1000802442,-0.0299591348,-0.2914909124,0.0547142662,-0.0248714257,0.226285696,-0.1493637562,-0.2985912263,0.04510976,0.2989329398,0.5806908607,0.2920951247,-0.1349260658,-0.0020116165,-0.0244322624,0.1738010645,-0.0050677597,0.0620811656,0.1071795821,-0.0976178795,0.2153973579,0.0976598486,0.1153569296,0.1360460967,-0.3167721629,0.0317346342,-0.0946736336,-0.2656799853,-0.2686154544,-0.1873129308,0.0141022345,0.3091271818,-0.1353947818,0.1039839685,-0.104213044,0.3705091476,-0.1415278912,0.3739591241,-0.1348130405,0.0459486619,0.1294860989,0.0374454074,-0.2365318686,0.2692069113,-0.0996681228,-0.1687610447,0.1625194252,0.0585832931,0.2478317171,-0.2568924725,0.0748605505,0.0678149462,-0.0504781604,-0.1628346443,0.3042094409,0.1541785151,-0.1740553379,-0.0858008787,0.0715481341,-0.2308282107,-0.3051958382,-0.1649234742,-0.0587764792,0.0251882523,-0.2654542327,0.2546764314,-0.0919789299,-0.1681697071,0.1078646034,0.4329473972,-0.1162305996,0.0310744662,-0.0215807576,0.1758451015,-0.2366054505,-0.0912673846,0.2016696483,0.2709169686,-0.407050699,-0.0816617608,-0.2495727241,-0.0352467522,0.1647792161,0.1570712179,-0.1957983375,0.1144887954,0.020055797,0.075720638,0.4357150793,-0.135609597,-0.3292346895,0.5616938472,-0.1428675652,0.043967139,0.1381468922,0.3059666753,0.3618828356,-0.0729374439,0.0767347664,0.2616869211,-0.2014493793,-0.1164130569,-0.1320562363,-0.3784996271,-0.3847476542,-0.0944835693,0.5402752757,0.1610286087,0.3405368626,-0.1514064372,0.3001446724,0.1266346425,0.2024376541,0.1396225393,0.5302109122,0.1900881976,-0.0358964689,-0.2318319231,-0.1146017015,0.0589498766,-0.0395759419,0.1955925524,0.1803219616,-0.1537195146,-0.2410766631,-0.135657832,-0.0562960096,-0.3382447362,0.1935910583,0.2385710925,0.1584198624,-0.2077828795,-0.1073490232,0.299267143,0.1788813472,-0.1213110536,-0.0368953981,0.2745821774,-0.4515448213,0.177214548,0.0068607121,0.253030926,0.0016299685,-0.230342865,0.3806056082,0.3563605845,-0.1053834483,-0.2045086622,0.428384304,-0.0621716045,0.1458102465,-0.2260694653,0.1204099432,-0.1315041184,0.0909258649,0.244899556,-0.0000739391,0.5810728073,0.2931474447,0.0285525266,0.1140239462,-0.0070644561,0.1190366223,0.0095453598,0.1411252618,-0.3736830056,0.087313883,0.3077257872,-0.0203570556,-0.1098162085,-0.467192024,0.1483647823,-0.1070786417,-0.0307745114,0.0113055492,0.1671342403,0.1091932207,-0.0787724257,-0.1423222721,0.3321632445,0.2356979251,0.4182392061,-0.0649807304,-0.0916898847,-0.0043933867,-0.0287387259,0.0683248788,0.2646042705,0.302741468,0.0663110316,0.2482176572,0.2664308846,-0.2229568809,-0.2070318311,-0.0658157542,0.155260846,0.0077507147,0.1695458591,0.1440908909,0.2144911736,-0.261707902,0.0228633825,-0.0659873858,-0.2564628124,0.2115870416,-0.1428839266,0.2090875357,0.3292405903,-0.4815833867,0.227349177,-0.4066334963,0.3680802286,0.0983323455,-0.0784763694,-0.3023244739,0.2722123861,0.1302017719,-0.2210624665,0.5940274596,0.0488381162,-0.1143660098,-0.4036473036,-0.734056592,0.3076915443,-0.2309062332,0.4652511179,0.1293065846,-0.0113285072,-0.0347547941,-0.1014852226,0.0613072664,-0.5267003775,-0.2977417111,-0.2505534589,-0.3947771788,0.0163212717,0.0256120227,-0.434248358,-0.3470573127,-0.242942065,0.2471621633,0.1174121872,0.1606267691,0.2522360981,-0.076719664,-0.0131175816,0.0112109128,0.0919826925,-0.188360095,-0.1630577445,0.2885794342,-0.4790608883,-0.5077258348,0.1003755257,-0.1816789955,-0.048965916,-0.0948415548,-0.1655977517,-0.480766207,-0.075268954,0.3006833494,0.2141378671,-0.089953132,0.2338048667,-0.3037539721,-0.3084408641,0.0085154437,0.2025715411,0.4028569758,-0.4210573733,0.1648031622,-0.0959573165,0.5690150857,-0.1091206297,0.3569287062,0.0350179411,-0.3291341066,-0.0378538817,-0.1166755855,0.4221532643,0.0258365534,-0.1519756317,0.2515668273,-0.1604850441,0.0971379355,0.3257716,-0.1527766287,-0.4096334279,-0.1608093977,0.1710219383,-0.2514334321,0.1008221954,-0.1981503069,-0.1485136747,0.1992205977,-0.1356605738,-0.3192770183,-0.0273625553,-0.1692511886,-0.0095627606,0.3584057093,-0.0911248997,0.0196216218,-0.6412318349,0.1002479494,-0.3555569351,0.350853622,0.3263651133,0.0846559852,-0.1404084712,-0.3038271368,-0.0280989613,-0.1446859837,0.3904827237,-0.4032163322,-0.2489654869,-0.0755449012,0.2410813719,-0.2321610749,0.1924747825,-0.2888331115,0.028145181,0.0378839038,0.1959316283,-0.176436916,-0.2767350376,0.5746406913,-0.0118691949,-0.0618813746,0.0060386425,-0.1643147171,-0.1177976727,-0.0228785742,-0.0069779153,-0.1040864959,0.0514113903,0.0060760151,0.0381991677,0.0050985892,0.0457031354,0.3633357882,0.0900890306,-0.0893752649,0.2335476279,0.2089573145,0.0448447615,0.1686370522,0.5208560824,0.0773428828,-0.4342058897,0.0092290919,0.2245377749,-0.1899528503,0.4233028889,0.1524060518,0.0886672363,-0.0919308811,0.2043445557,0.1073888168,0.1087801009,0.1067835689,0.3649703264,-0.1767453402,-0.2914912701,-0.1994679868,0.2086274028,0.1136725098,-0.0447599478,-0.234474346,0.2381378561,0.0613950379,-0.201606974,0.1958928406,0.7938673496,0.0535370521,0.0564456135,0.0717636645,-0.5559239388,0.4628688693,-0.1370230913,-0.0152129345,-0.3494098783,-0.1274490505,-0.1913981885,0.0301337652,0.1231654808,0.121002771,0.0159668475,0.0021429711,-0.04956742,-0.0844683573,0.0293208603,0.20752877,0.3125416338,-0.1744141132,-0.0345437676,0.2706574202,0.2149135619,-0.0426106304,-0.0016283541,-0.0616666414,-0.5263208747,-0.0678347051,-0.0818144679,-0.0919504985,-0.1133589223,0.2590901852,0.3719120622,-0.6455821991,0.0334825814,-0.0275622662,0.2099269629,-0.1769394577,-0.0650682226,0.0897771567,-0.1560286283,-0.3026634157,0.2249397486,-0.1417625844,0.2554256022,-0.1679507345,-0.0738022327,0.3642205,-0.0728304982,0.0657315329,-0.2658986747,0.0691035688,-0.0296811704,0.1194776222,0.0231785439,0.222653538,-0.0533249229,-0.2640281618,0.1794907749,0.0596045032,-0.1977024078,-0.1584542692,-0.0430640206,-0.0344803967,-0.0806506202,0.4147302806,0.1429398656,0.0305213444,0.2401253134,0.1684053093,-0.2359341681,-0.3452755511,-0.1083543003,0.5156996846,-0.5914090872,-0.2685299814,-0.0390959121,-0.2759586573,0.3413716257,0.5564902425,0.1793883443,-0.1912810504,-0.1530119628,-0.2031973153,-0.541502893,0.147339493,-0.5056758523,0.2357789427,-0.0727167651,0.3771274686,-0.0197018441,0.2641434371,-0.4471136928,0.0071390527,-0.1296503246,-0.004519267,-0.1992209256,-0.2223643214,0.1353749186,0.0397876576,0.1324140579,-0.0320298038,-0.3486899137,-0.217328459,-0.1669685245,0.0621907488,-0.068445839,0.1250793934,0.1526536942,0.1905920357,-0.4427495897,0.2042124569,0.2229616642,-0.1053894609,0.0292233769,-0.1503476948,-0.1081607342,-0.2333945781,-0.0313133188,-0.1561748236,0.2293699682,0.0695945099,0.2426783741,0.0235761926,-0.147792697,-0.2319273502,-0.0494957641,0.4965571165,0.1277211159,-0.1563830376,0.3315269649,-0.1826723665,-0.1078139842,0.6040503383,0.4646678865,0.0868366361,0.1145821363,0.1285393983,-0.0904756039,0.2801995575,-0.4079580307,-0.1464688033,-0.2841692865,-0.310656637,0.1986572444,0.3563961983,0.376666069,0.0781208873,0.111849919,0.1665924042,0.4697527587,-0.2706463039,-0.0113694873,0.113182418,0.2222956419,0.2097015083,0.4822472036,0.13309443,0.3767611086,-0.0205600392,0.0293649957,0.212193042,0.0055864505,0.4002189338,0.4163548648,-0.1765593141,0.2393303961,0.0699880868,0.0705054328,-0.2296879292,-0.1267513037,0.5943972468,0.1086574495,0.1331958473,-0.1361577511,0.208336696,0.0184066817,-0.0711425692,-0.0612211339,-0.3688546121,-0.2815545499,-0.1877405494,0.0114332587,-0.0446936041,0.3664059937,-0.1655047983,-0.1093176603,-0.3379770517,-0.0985407978,0.0970790312,0.3179985285,-0.1744046211,0.2670804858,-0.1424382925,-0.2200054079,0.1298749596,0.3507328928,0.3833196461,-0.0370867774,0.2647227943,-0.0580474362,0.010272908,-0.319122225,-0.1735744625,0.084566161,0.2251166999,-0.2941142321,0.0996502116,0.2031763047,-0.2700675726,0.1751092523,0.1986063272,0.1831952333,0.0721767023,0.2444214523,0.0410772413,-0.2169228792,-0.0434683189,0.1895227581,-0.3223095238,0.1835825741,-0.0219365507,-0.0963081867,-0.056744013,-0.0955544561,0.1596498489,-0.105700992,0.2242289633,0.1842171252,-0.0698265806,-0.1372910142,-0.2968896031,-0.4947137535,0.1266514212,0.0586738661,-0.1929433048,-0.0974274948,0.2545585632,0.1445436776,-0.0740208551,0.0113139628,-0.1953830868,0.0658583939,-0.0477682278,-0.2914222181,0.0688739568,0.0407692418,0.2265983373,0.1290673912,-0.1045505032,-0.2551732063,-0.1046205461,0.0858756229,-0.0853225589,-0.2774783373,-0.0243859813,0.1511465758,0.2302410156,0.0873774439,0.4935260415,-0.1358513683,-0.2504100502,-0.2734026909,-0.1217617989,-0.3229370415,0.3654846251,0.2327882051,0.2945095003,-0.1016464531,-0.0075412393,0.1279548556,-0.1188662499,0.2819202244,0.1379439682,-0.3423040807,0.0486828201,-0.0289279278,-0.1913620532,0.384523958,0.330060482,0.2339158803,-0.088885501,-0.2191726863,-0.3465105593,0.4386180639,0.131814301,0.0666038767,-0.1635780632,0.2013248354,0.0277908556,-0.1799482405,-0.3909984529,0.0589271896,0.4154228568,-0.0094552301,-0.2802794278,0.2262725234,-0.0470433347,-0.1045427546,-0.0862362385,-0.0681231543,0.0298431721,-0.1870165169,-0.1349453181,-0.0159287937]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1757","title":"FewRel","comments":"Hi @lhoestq,\r\n\r\nThis issue can be closed, I guess.","body":"## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":9,"text":"FewRel\n## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nHi @lhoestq,\r\n\r\nThis issue can be closed, I guess.","embeddings":[-0.1808029711,-0.0181143489,-0.0780677646,0.120809719,0.1676013321,-0.1296756268,0.4108457565,0.2106958926,0.0621083528,0.2879484892,-0.2904240787,0.0767703801,-0.0169591755,-0.1308662146,0.0116936071,-0.0286398884,-0.1885499209,0.2275907248,0.0777176097,-0.1060961485,-0.1975560188,0.2301272154,0.1895740181,-0.0882273689,-0.2996605337,-0.309656173,-0.0765421987,0.2386603355,-0.2245041132,-0.3065410852,0.2668050826,0.2183933407,0.0175551753,0.3460306525,-0.0000974733,0.018956095,0.2162040621,-0.0207171626,0.1066471934,0.1169297248,0.2235879302,-0.4069417417,-0.0293102562,-0.2288516015,-0.1753050685,-0.2412506044,-0.1184017807,-0.0784119144,0.2979819477,0.3750507534,0.401792556,0.0852036774,0.1948197335,-0.5101656914,0.0663812086,0.0178114139,-0.2355064899,0.3359563947,0.4162049592,-0.0507415012,-0.0465715826,0.1132877693,0.1174258441,-0.066364035,0.2888450623,-0.0625035986,0.0316069648,-0.3124871552,0.1570777744,0.3329906762,0.2418317348,0.0481344052,-0.2077130377,0.0328542925,0.0902677327,-0.1357441992,0.0187647492,-0.0416836254,-0.030480681,0.1987448931,0.0923806205,-0.1858100295,-0.3178921938,0.2012419999,0.044709675,0.0984826162,-0.0919002444,-0.0593509115,0.1076175943,-0.0214385707,-0.2554181218,0.1773334891,-0.1140585691,0.0937752351,-0.1122472808,-0.1759533882,0.0943795517,0.162274763,0.4482193589,0.2046452463,-0.0552310571,-0.0106870737,0.0290232599,0.2852057219,0.0631905422,-0.0088036889,0.0584984757,-0.0263165664,0.2566437423,0.1109607071,0.0313489102,0.0392951816,-0.1897882074,0.0825290903,-0.0327060781,-0.2553116381,0.006225599,-0.146185562,-0.0316742323,0.3297106028,-0.0191867091,0.2478522658,-0.1020207033,0.4516358078,0.0000712528,0.2034422159,-0.0209179986,0.1370739639,0.0579067841,0.010663298,-0.2604020834,0.1974891573,-0.1758312285,-0.1350475997,0.1569526643,0.0796194598,0.2915763259,-0.2635476589,0.0671823248,-0.0632347465,0.0173636079,-0.0462771207,0.2194437236,0.273419559,-0.0401356854,0.0162940435,0.1226025075,-0.3373734951,-0.1078729257,-0.1142849997,-0.1314961463,-0.1381130069,-0.5157940984,0.3525086939,0.0308464747,-0.1046556681,0.114300929,0.2216270119,0.0478558168,-0.0600191057,-0.0960441455,0.1303197145,-0.3066619039,-0.0990547985,0.0206880867,0.212213397,-0.3391809464,-0.1122600809,-0.1902308017,-0.0382862277,0.2116434276,0.3116362989,-0.3064493835,0.078679204,-0.0801115483,0.1044395417,0.3001908362,-0.0898460299,-0.388969034,0.4361030161,-0.3649218082,0.0027906715,0.079398647,0.3739340603,0.2675077319,-0.0252506901,0.119905062,0.2300880104,-0.1341723353,-0.0019038641,-0.302396208,-0.3006786704,-0.3238678575,-0.0742134601,0.4525180757,0.2742343843,0.3122784793,-0.0970218852,0.2162890434,0.0165893827,0.238463372,0.3067974746,0.5038230419,0.2314114869,-0.030106103,-0.2044205368,-0.0811566114,0.0518614538,-0.1124265641,0.1997338831,0.0985166281,-0.2959753871,-0.3114283681,-0.0066559264,-0.0703927875,-0.2894371748,0.2840496898,0.1759010851,-0.0940946341,-0.1097225994,-0.0521645099,0.2695731819,-0.0055281147,-0.0538732521,-0.0671928823,0.0875445455,-0.491230756,-0.006064069,0.0625972003,0.2264440954,0.0320327915,-0.09513437,0.1727654487,0.3258150518,-0.1495260596,-0.0811321661,0.3624196649,-0.0430062898,0.0389935561,-0.2586916983,-0.0873662233,-0.0699494556,0.1100273207,0.1219020337,0.0107524395,0.4187420905,0.2516456842,0.0200031158,0.1170992851,0.0416396931,0.2519804537,-0.142164588,0.2165128291,-0.3450019956,0.1627526581,0.2314021885,-0.0652139112,-0.0580293834,-0.3667825162,0.1272725165,-0.0029019797,0.1323487759,0.0670004636,0.1729195565,0.067348361,-0.0543640703,-0.101155445,0.2557019889,0.2808769643,0.3990395665,-0.1252444685,0.0487149507,0.019615747,-0.1634215713,0.0975825414,0.1984104514,0.2704538107,0.1052128673,0.375715524,0.08317931,-0.3901143074,-0.0981280208,-0.0110338964,0.2425990999,-0.2126902491,0.1900671721,0.1261487305,0.2175794691,-0.0601832978,0.1364499331,-0.0085577294,-0.2400442958,0.2732390165,0.0862797573,-0.0382658318,0.2846345603,-0.3472651839,0.1481615305,-0.2647798359,0.36599347,0.0834739208,-0.2566330433,-0.1778127849,0.2710965276,0.154545337,-0.1231421605,0.635571897,0.1507196277,-0.1489630193,-0.5046024919,-0.6275627613,0.1832464188,-0.136450693,0.4536368847,0.1753200144,0.0766571909,-0.0131043503,-0.20239833,0.1722240895,-0.4359531105,-0.2899014652,-0.0873359665,-0.3042447269,-0.0161519721,-0.1024261713,-0.3714097738,-0.2258157879,-0.3922270536,0.2071761936,-0.0616301112,0.1119166017,0.3182543516,0.0308360942,-0.0708208978,-0.0565743707,0.0180277843,-0.2954258025,-0.191549927,0.2193383127,-0.4090587497,-0.5498838425,0.1575867087,-0.0728152692,0.0117168259,-0.1231562644,-0.1363469064,-0.3830792904,-0.0102424882,0.1844962984,0.0779681653,-0.2018749118,0.2609051764,-0.2392719537,-0.3688221872,-0.084757112,0.1115631163,0.3843595386,-0.3873572052,0.0980261192,-0.230054602,0.4816046655,-0.1098193228,0.2507067323,-0.0373265631,-0.2182802111,0.1951286644,-0.1093723774,0.4279085994,-0.039712131,-0.3216721714,0.1915164292,-0.1016862616,0.0220176615,0.4184542298,-0.1851148307,-0.3701322079,-0.0510330796,0.0779800043,-0.2020585686,-0.0665655658,-0.1119051352,-0.0255076922,0.2047600299,0.0064605814,-0.305162847,-0.0777821243,-0.1031171829,0.0693221837,0.3196751773,-0.1103210822,-0.0941891596,-0.6322010159,0.0613824166,-0.3712113202,0.2117922455,0.2575876117,0.0801220834,-0.268029511,-0.2475098372,0.1118543074,-0.1345658004,0.4778987765,-0.4532021284,-0.235138163,0.0690477043,0.1874253601,-0.1655575931,0.1844262332,-0.256325841,0.1586071551,0.1365531087,0.1540069133,-0.1881376952,-0.1216735616,0.491253227,-0.0786711425,-0.1226390302,-0.0775036514,-0.2745089233,-0.230557397,-0.2186880559,0.0097854007,-0.004955933,0.1337018907,0.103072688,-0.0729262009,-0.0698084757,0.0262816455,0.2885278761,-0.029149022,-0.0175474845,0.2378564328,0.2739335597,-0.0581260622,0.2452708483,0.483696878,0.1778495759,-0.3261319697,-0.0471012704,0.1769392639,-0.2911952436,0.3951834738,0.1574407965,0.1013020128,0.0871591792,0.1808570623,0.2406641096,0.0803166181,0.2237750739,0.3007832468,-0.1695039868,-0.2671209872,-0.0972461104,0.1445599794,-0.0069934176,0.012474848,-0.1525265723,0.278006196,-0.0446117595,-0.1008845046,0.0154108172,0.8127753735,-0.0103392461,-0.0285805166,0.0762713179,-0.4063682258,0.4561861157,-0.0778924301,-0.0302499235,-0.3827077746,-0.2142033279,-0.0388347916,0.0491030328,0.1128981113,-0.0009273285,0.0538455844,-0.1655240953,-0.0906514674,-0.0853361413,-0.0128459297,0.2685771883,0.2995320559,-0.2275349051,-0.1281597912,0.3530899882,0.0991489142,-0.0796749666,-0.0346979611,-0.1329343468,-0.3955854177,-0.1244655401,-0.0940723643,-0.0767636821,-0.1380714774,0.2331942618,0.2770351768,-0.4989718795,0.1298664063,-0.1043008566,0.2008329183,-0.0024036174,-0.1944067031,0.1969820112,-0.1992521584,-0.3588381708,0.0911505148,-0.0750241727,0.2421191931,-0.1770430654,-0.1016751081,0.3314073086,-0.0384770781,-0.1298699528,-0.2597381771,-0.0137260668,-0.0139463907,-0.0585114509,-0.145024538,0.0000325897,-0.1234035641,-0.411118567,0.2466455102,0.0497920103,-0.2435127497,-0.168440029,0.0950877368,-0.0897551551,-0.0649034306,0.3532231748,0.1842102706,-0.027140867,0.3283738196,0.1764295399,-0.2269107252,-0.3749859631,-0.1826583445,0.4069662094,-0.5545350909,-0.0856506154,0.0857353359,-0.368786633,0.2555191815,0.4795072079,0.1169917434,-0.1802443117,-0.125709787,-0.1733247936,-0.4848690629,0.1199774966,-0.4167416692,0.1511498392,-0.0163115673,0.4133132994,-0.1413415968,0.2034909725,-0.5242086053,0.0168003105,-0.2935023904,0.0785383061,-0.1044522524,-0.2765339017,0.1294825375,-0.0846299753,0.2080330402,0.0194306783,-0.1794977933,-0.3288038075,-0.1829348058,0.055635117,-0.0800641701,0.2264829874,0.1472465694,0.1100631058,-0.409932375,0.1848565638,0.2468568087,-0.0748329908,-0.022060018,-0.200030461,-0.1497261375,-0.153646782,-0.0012866595,-0.1294607222,0.0566793233,0.1003654152,0.026135914,-0.0841583535,-0.168185249,-0.1754586697,-0.0180477276,0.3504442573,0.1725651473,0.0133981789,0.3302252293,-0.1944218576,-0.0957019776,0.6609928012,0.3718491793,-0.0646085888,-0.0441681035,0.1088987067,-0.0897361115,0.3961893916,-0.473642081,-0.1916470081,-0.1482535899,-0.3870127499,0.1807162911,0.2670252323,0.2844669521,0.017499866,0.0985448807,0.2442566752,0.3551357985,-0.2688136101,-0.1049297303,0.0175603926,0.0864282176,0.1230958924,0.3854641616,0.1508794576,0.4262494147,0.1618199944,0.0555315167,0.3419499397,0.0433966815,0.2872472405,0.2619726956,-0.1629120857,0.1712902039,0.0178294014,0.1604354084,-0.1213101968,-0.0615904033,0.5345629454,0.0566303767,0.0610487051,-0.0490589663,0.259888351,-0.2023025155,-0.0173291732,-0.0954884067,-0.2550853789,-0.2432696223,-0.1292509586,0.0156160537,0.1270459592,0.3617632985,-0.1353558451,-0.1795092076,-0.3323293328,-0.0260956883,0.0787938684,0.2364375293,-0.2689930201,0.1665545553,-0.1160538048,-0.149691835,0.0671646222,0.3805427849,0.4659148157,0.0434172638,0.2461191118,0.020060908,0.0019618338,-0.3012499213,-0.213534072,0.059526857,0.3002118468,-0.046627786,0.1194072738,0.2427444458,-0.3386581838,0.0542123392,0.2088083029,0.2124068886,0.1142008156,0.1936499178,0.0244508479,-0.2626784444,-0.2064420283,0.0702904984,-0.2665739357,0.1733293533,0.0899973214,-0.1161904484,0.0383968875,-0.1808621138,0.1974032521,-0.0951185077,0.2997194529,0.1457478702,-0.1463412941,-0.1897750348,-0.2256280184,-0.6411647797,0.3347455561,-0.1065040901,-0.121358268,-0.0441799201,0.2051998824,0.1695377082,-0.0853292942,0.0131835649,-0.1106608436,-0.0496621132,-0.1308104843,-0.3578247726,0.0725002289,0.1493583173,-0.0091124866,0.2044561058,-0.1869358718,-0.0985556394,-0.0188493393,0.1715780199,-0.1563979089,-0.1734339297,-0.0990466774,0.1176955998,0.2990611792,0.0750098452,0.4551798701,-0.1850959808,-0.1769665927,-0.2278187573,-0.2734071612,-0.2725765109,0.3996143341,0.3662178516,0.4007105827,-0.0433427468,-0.0541247316,0.0414936207,-0.0183664858,0.2173535675,0.1563581675,-0.438266933,0.2726967335,0.0129066678,-0.0042252471,0.2568922341,0.1996421367,0.199921295,-0.0348591432,-0.3074877858,-0.4889596403,0.5142807364,0.1757254153,-0.1467980891,0.0420523174,0.3661520779,0.0241865311,-0.2214048803,-0.4769831002,0.0908577442,0.3528246582,0.0643206909,-0.2513669729,0.2589532733,-0.14803119,-0.0580652356,-0.0208256692,-0.0525892079,0.1139867455,-0.2292681634,0.0108076558,-0.0203334484]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1757","title":"FewRel","comments":"Yes :) closing\r\nThanks again for adding FewRel !","body":"## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":9,"text":"FewRel\n## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nYes :) closing\r\nThanks again for adding FewRel !","embeddings":[-0.2875337303,0.0106587615,-0.1523834467,0.0971100256,-0.010653452,-0.1158419102,0.3578749597,0.1888900995,0.0925165564,0.1230087727,-0.3634783328,0.0637451485,-0.1328780949,-0.0399035066,0.1875283271,-0.0803906098,-0.0900181085,0.2024358362,0.0415915772,-0.1434133798,-0.1888293028,0.1651819199,0.2043955028,-0.1686096191,-0.2479612231,-0.349935174,-0.2151769698,0.2594509125,-0.2728521824,-0.2859751284,0.1656404138,0.3750579953,0.1401642263,0.3853712082,-0.0000992832,-0.1298501641,0.0777483433,-0.0264195856,0.0907977819,0.1485745758,0.1842165589,-0.4958061278,-0.0975043699,-0.3213952482,-0.1789551079,-0.2443127185,-0.1557503939,-0.1080068722,0.3411913514,0.2636165619,0.365937382,0.0218074862,0.2224309742,-0.4490585327,0.0538558587,0.2442874163,-0.198841542,0.3173871338,0.4268197715,-0.0863176063,-0.0196411423,0.0991021469,0.0838247836,-0.1965164095,0.4442656636,0.0426513739,0.0545721538,-0.3263798058,0.0288198274,0.4053670764,0.2394439578,0.0274675768,-0.3262938857,-0.0484689809,0.0865741819,-0.1634723544,-0.2504361868,-0.0326608606,-0.0718135759,0.1650737822,0.0192598738,-0.248087436,-0.3146440983,0.1446080953,0.0101069016,0.1805828959,-0.0703910291,-0.1390487254,0.04267212,0.0194811132,-0.2575359046,0.0105512897,-0.1312971115,0.1188223511,-0.0865580887,-0.3111782074,0.0613264777,0.2149315923,0.4685500264,0.2276465595,-0.0438331179,-0.0610441454,-0.0086562876,0.2153106928,0.1316839755,-0.0943963081,0.041289106,-0.001553926,0.1242503375,-0.0268379599,0.0688538253,0.075582549,-0.155159831,0.1695082933,-0.04885922,-0.2177114487,-0.2435970753,-0.1682212651,0.0081367781,0.2771607637,-0.0034208114,0.1095888838,-0.1201867089,0.3874194324,-0.0621997081,0.2311492711,-0.0918801874,0.1214367524,0.1913250983,-0.0068128631,-0.2269782722,0.2282889038,-0.0596470721,-0.1287112832,0.1234245896,0.147028178,0.1211746186,-0.3174200058,0.1126171872,0.0374625064,0.0856075436,-0.0586406291,0.3905527294,0.13288863,-0.1296921372,-0.1929733008,0.1450287551,-0.2876396477,-0.2656193674,-0.0908796191,0.0073934831,0.0233111922,-0.4176879227,0.3243764043,-0.0816099569,-0.1985695213,0.0992235765,0.4074063897,-0.0406472571,-0.0271635745,0.0338275395,0.2425031215,-0.2383504063,-0.0632192791,0.0560784899,0.2210788876,-0.3295474946,-0.0912756398,-0.1150328964,-0.0768974274,0.1189088598,0.128370136,-0.1998441517,0.0438596383,0.020069737,0.1230496168,0.3791143894,-0.0951874182,-0.2231141478,0.4523551166,-0.3013658226,-0.0187369622,0.1364471763,0.4479233921,0.3015094697,-0.0625758693,0.0572183356,0.2925126553,-0.2566671669,-0.0248197094,-0.192551598,-0.3302466273,-0.3025145829,0.0199945047,0.4915995598,0.219728902,0.245598048,-0.1333459914,0.392593056,-0.0014048787,0.191536516,0.1148372218,0.5021218061,0.1998799145,-0.0559194461,-0.2939603031,-0.0203084033,-0.0084900539,-0.0038652192,0.2254637182,0.1468742937,-0.345802635,-0.1599151045,-0.0447860882,-0.0239097551,-0.2488329262,0.283947438,0.1079255939,0.0160202328,-0.1739485413,-0.1250492483,0.2801417112,0.0118879229,-0.0743173286,-0.0695930347,0.2170924991,-0.3353437781,0.1267438978,0.0726153105,0.2572835684,-0.1895456165,-0.1395813674,0.3312996626,0.2357709259,-0.1704182625,-0.0671296194,0.4358615279,0.0427999645,0.0882419571,-0.175455451,0.0328115337,-0.2251827866,0.0565402135,0.0974495709,-0.0380427837,0.6018109918,0.2903515399,0.0660116896,0.1310124397,0.0054880432,0.12049824,-0.0631572977,0.112692751,-0.3268459141,0.131019637,0.2469531,-0.0868785307,-0.1658140421,-0.301618427,0.2086547464,-0.0966262668,0.0448407121,0.1644018292,0.1092361212,0.0403834991,-0.0174589679,-0.0582844988,0.2347857952,0.2777389288,0.4795808196,-0.0856320262,-0.0817909986,0.0480001867,-0.0602510944,0.0823124722,0.1881574541,0.2941642404,0.0826357529,0.2501282096,0.1838366687,-0.3123256564,-0.1778454632,-0.1523909569,0.1517473012,-0.0506782867,0.2062710375,0.3601318896,0.1550801992,0.0005383271,0.1676723063,0.0171711482,-0.2274632454,0.2411015481,-0.0704209656,0.0084083872,0.2138962895,-0.4384093285,0.2927947044,-0.3772347569,0.2788537741,0.0246485993,-0.2675840557,-0.0362920463,0.3432736695,0.1893237084,-0.1330326051,0.7287749648,0.229046911,-0.0321698226,-0.4866200984,-0.6355057955,0.2339186668,-0.2664433122,0.3015739918,0.2073710412,-0.0075737876,0.0477184691,-0.1361570209,0.1382267475,-0.3964841962,-0.3172879815,-0.2671272159,-0.2999072373,-0.0293843746,-0.0793872774,-0.263420403,-0.3177826703,-0.2964352071,0.2829226851,0.1315188706,0.0790157765,0.2388917059,-0.0349725299,-0.0175608248,-0.0840111151,0.0458833352,-0.21545811,-0.2385615706,0.1429241002,-0.3682644069,-0.3774743974,-0.0232931431,-0.2360658944,0.0341162272,-0.1301502883,-0.0974801034,-0.4307121336,0.083976984,0.2789283097,0.1392573267,-0.1590548903,0.2591384649,-0.1670362651,-0.3652663827,0.0370687321,0.2513431013,0.3511079252,-0.3416073024,0.0957665667,-0.1931426823,0.3641133308,-0.1266853064,0.2588871419,-0.0505500957,-0.2579142451,0.0323480591,-0.1520160288,0.3124376237,0.0855914801,-0.2974049747,0.1654810607,-0.1653755009,-0.0325751305,0.4339203835,-0.1342047155,-0.4452147782,-0.1238721311,0.0728034824,-0.2326901704,-0.0034945905,-0.0634067878,-0.0677720308,0.1786368191,-0.1646080315,-0.4210709333,-0.1100380048,-0.1302788258,0.0510389656,0.297696054,-0.057168413,-0.1678046286,-0.6591905951,0.1126190349,-0.4337726831,0.2263770103,0.2729229629,-0.1818242669,-0.1617194563,-0.3182863593,-0.0720301345,-0.1466973275,0.4162207246,-0.4531994164,-0.3186271787,0.0070719868,0.1742385477,-0.0861137733,0.3215677738,-0.3044390678,0.0013389824,0.0829149336,0.2028677613,-0.1973909885,-0.1216507703,0.5747256279,-0.079643473,-0.1001477763,-0.1787333786,-0.164250344,-0.1398013681,-0.1391204894,0.0734041408,-0.0387465507,0.1271094084,0.1352702677,-0.030318059,-0.0368143171,0.1033981144,0.4114249945,0.0282781087,-0.142258808,0.2676340938,0.2297679931,-0.0564101525,0.1865285635,0.459900111,0.1264956146,-0.3995288908,0.0163274966,0.059076298,-0.35529989,0.4931574762,0.2073812187,0.1378112733,0.0625204891,0.1836926639,0.1478473693,-0.0527431034,0.1196395233,0.2224983871,-0.1567843407,-0.3134070933,-0.1756420583,0.2643992603,0.0717277378,0.0164301433,-0.239122197,0.3413561881,-0.0806607828,-0.1542971432,0.2193665951,0.859398067,-0.0250372179,0.148004815,0.0772601366,-0.3167913854,0.4993755221,-0.1293971837,-0.0646160021,-0.3548464775,-0.1686112583,-0.1730317473,0.0097987494,0.1495941728,0.1340570897,0.0342640355,-0.1679648906,-0.0086201048,-0.0258637145,-0.0508285984,0.2731270194,0.3305746317,-0.2595179975,-0.1435028017,0.3277001977,0.1946571618,-0.1617003083,-0.0191793777,-0.1429377049,-0.4701788425,0.0181076489,-0.1404737681,-0.171665132,-0.1542669237,0.1221645325,0.1744614542,-0.592680037,0.0816790462,-0.1070447415,0.1642124802,-0.0896489024,-0.0724602491,0.0523984656,-0.2021716088,-0.200508967,0.2021634281,-0.0964225754,0.2286461294,-0.1274392903,-0.0349879228,0.4232826233,0.0107882535,-0.139517352,-0.2284707576,0.0720746294,0.0236559808,0.0028935969,-0.1003718898,0.2432017773,-0.204118371,-0.2718681693,0.2371293306,0.1507667601,-0.1413346529,-0.1356479079,0.0228810273,-0.0144340303,-0.1180677637,0.3046966791,0.2383089215,-0.0531435274,0.1642585993,0.16848737,-0.1920378804,-0.4270596802,-0.110406585,0.4733380079,-0.453532964,-0.1759237349,0.1451991647,-0.3562276959,0.2171920687,0.4309535921,0.2760737836,-0.1646876186,-0.1243498623,-0.1129826903,-0.4427353442,0.1748243123,-0.3682540655,0.2715051174,-0.0644526631,0.3512436152,-0.1981609464,0.2010611892,-0.5286085606,-0.1205844581,-0.105429031,-0.0295944735,-0.0884403735,-0.1676302701,0.1176245287,-0.1192450449,0.2149566561,-0.0414175391,-0.1926350892,-0.3285610378,-0.1362056732,0.0490976907,-0.0357553065,0.1555749029,0.2108418792,0.2674060762,-0.4335003793,0.1327947974,0.2127736211,-0.1310081333,0.0270164032,-0.1038974673,-0.1093416885,-0.2464048713,-0.1110697165,-0.0843382552,0.1767394841,0.0718250871,0.1079543903,0.0005587236,-0.0973356739,-0.166267246,0.1166290417,0.5086703897,0.1509511471,-0.0721791759,0.3508801758,-0.0710726231,-0.1193414927,0.6591043472,0.4402605891,0.0469150655,0.0344090872,0.2066114992,-0.2083340436,0.4050641954,-0.4362843037,-0.118632786,-0.2199586332,-0.3662112057,0.2784168422,0.400996536,0.237154305,0.0817925408,0.0515142269,0.122528702,0.4015545249,-0.234258458,-0.0619484372,0.0594778508,0.2377640754,0.1692575514,0.3713253736,0.2718601525,0.4022030234,0.0338509493,0.0601159967,0.328792423,-0.011272816,0.3472571373,0.3563884795,-0.0138716307,0.2213044763,0.0917461738,0.0315638818,-0.1003786474,-0.0436645076,0.5973936915,0.0814030617,0.0300579127,-0.0778026506,0.1792795211,-0.0860808343,0.0415826812,-0.0819956511,-0.3031034768,-0.2964605987,-0.2328409404,-0.023607865,-0.0412154831,0.2524965107,-0.1484970003,0.0276750047,-0.2894982696,0.0475467741,0.0520153195,0.3001571596,-0.2969777286,0.0705281049,-0.0851714984,-0.1610518098,0.1087612212,0.3480525613,0.3604705334,-0.0651714355,0.2429694235,0.0712956935,-0.1482897252,-0.2688575387,-0.2261680514,-0.005673958,0.2815137804,-0.2588606179,0.1549865752,0.2756815255,-0.2951668799,0.1053909883,0.219185397,0.089402616,0.0560530461,0.2085376829,0.0821752474,-0.2664248943,-0.1503493935,0.1433918178,-0.2735259831,0.1293117851,0.0650363341,-0.1314902455,0.0139298374,-0.1149972975,0.1834139675,-0.1123321801,0.2601918876,0.2314534038,0.0036756841,-0.1994983554,-0.1940575838,-0.5071986914,0.2205884308,-0.0549170077,-0.1557895094,-0.0919940174,0.204753086,0.3347509503,0.054913301,-0.0795720518,-0.1874669641,-0.0878361538,-0.1474990398,-0.3655314147,0.004848849,0.1099773943,0.1546253264,0.1906917542,-0.0301785357,-0.2167176306,-0.0848456174,0.1510771811,-0.2482943237,-0.1648330539,-0.0443035327,0.1106970832,0.2309981287,-0.0372029692,0.4769468307,-0.1794149727,-0.2553845048,-0.0641819239,-0.1565840095,-0.3128892183,0.4231953323,0.3031691611,0.357368499,-0.0536675714,-0.043716576,0.1868401617,-0.0788539574,0.2340695411,0.0629871786,-0.4088656008,0.1562540829,0.1336901337,-0.1340944618,0.1494118571,0.2744581401,0.2179469019,-0.0025269687,-0.196327731,-0.3722653091,0.4109042585,0.2210918069,0.009732658,-0.0144434012,0.2379879057,0.0206063837,-0.1558743119,-0.4271237552,0.054615479,0.3901998997,0.0177917052,-0.1722000241,0.2606582046,-0.0190175343,-0.06081938,-0.1168080941,-0.0765928999,-0.0628617108,-0.1698685139,-0.15774858,-0.1085227057]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1757","title":"FewRel","comments":"Thanks for adding this @gchhablani ! Sorry didn't see the email notifications sooner!","body":"## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n","comment_length":13,"text":"FewRel\n## Adding a Dataset\r\n- **Name:** FewRel\r\n- **Description:** Large-Scale Supervised Few-Shot Relation Classification Dataset\r\n- **Paper:** @inproceedings{han2018fewrel,\r\n title={FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation},\r\n author={Han, Xu and Zhu, Hao and Yu, Pengfei and Wang, Ziyun and Yao, Yuan and Liu, Zhiyuan and Sun, Maosong},\r\n booktitle={EMNLP},\r\n year={2018}}\r\n- **Data:** https:\/\/github.com\/ProKil\/FewRel\r\n- **Motivation:** relationship extraction dataset that's been used by some state of the art systems that should be incorporated.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\nThanks for adding this @gchhablani ! Sorry didn't see the email notifications sooner!","embeddings":[-0.1900702268,-0.0144992452,-0.1562916636,0.161405459,0.0706780776,-0.0939077362,0.384331882,0.212866351,0.1212483421,0.2625028789,-0.3554014862,0.0460107252,-0.0803114548,-0.2084320188,0.1336005032,-0.0399165042,-0.076281324,0.1871867627,0.0871007517,-0.2520676553,-0.1885856688,0.1232660413,0.242095232,-0.1866009086,-0.3009221554,-0.3609582484,-0.2366044223,0.3326798975,-0.2185071111,-0.288333714,0.1857167482,0.3460225165,0.0858053714,0.3593059182,-0.0001005667,-0.1380589902,0.1750151068,0.0188197549,0.04335225,0.1519100219,0.2614310682,-0.5399442911,-0.0816475451,-0.2640249133,-0.1600672752,-0.2487351,-0.1214047819,-0.1083766818,0.3144036531,0.3036656976,0.3376188576,0.0304618981,0.2102383226,-0.4834238887,0.054740712,0.1356033832,-0.1364352554,0.2815606296,0.4307636917,-0.058318723,-0.0241041407,0.0310208257,0.0360192023,-0.1632511467,0.379432708,0.0236519594,-0.0523140579,-0.4109354019,0.054597538,0.4084942043,0.2652336061,0.0865260214,-0.207544744,-0.0101573607,0.04328667,-0.1706671268,-0.1937091947,-0.0567571633,-0.0229944512,0.1541974694,0.0049401373,-0.1573597938,-0.3674313426,0.242808789,-0.034800712,0.1242666468,-0.0677570552,-0.0981712043,0.0309496429,0.041715648,-0.3721035719,0.0435342193,-0.0990321413,0.1791328192,-0.0562249571,-0.3267281055,0.1533728689,0.1808540374,0.4816202521,0.2858572006,-0.0851259157,-0.0373482928,0.028596472,0.2077587992,0.0037744751,-0.1080968156,0.1201503426,-0.0599145293,0.1684412807,0.0156695228,0.1089125872,0.0364900604,-0.2193184197,0.2600757182,-0.0785576254,-0.2503230572,-0.1884501576,-0.1731484234,-0.060636282,0.3098232448,-0.034238752,0.1600990295,-0.1446526945,0.3885330856,-0.1218172386,0.267016381,-0.1527222246,0.0802130774,0.1670618355,0.0431302823,-0.2954867184,0.187254563,-0.0746470988,-0.1688214839,0.11724893,0.1048883796,0.1633947045,-0.2846150398,0.1160283685,0.0059228181,0.0090924837,-0.1061220169,0.3364074826,0.1781630665,-0.1329759359,-0.0448362939,0.1398036331,-0.2845572531,-0.2663898766,-0.1273352802,-0.0629763827,-0.0615012869,-0.4380453229,0.3229442835,-0.0911532268,-0.1581274569,0.1619058549,0.3866597116,-0.1029952541,0.0840035975,0.0004980047,0.2565163374,-0.2645188272,-0.0941264108,0.0617985129,0.2278361768,-0.3654570878,-0.114468798,-0.2475518137,-0.0974167436,0.2141191512,0.1823307872,-0.2099311352,0.0691155419,0.0376858562,0.0485897399,0.3898592889,-0.1015208215,-0.3593335748,0.4715607166,-0.3996663392,0.0059052943,0.1693788469,0.4188311398,0.3293089569,-0.069239445,0.077889055,0.2983881235,-0.2525953352,-0.042433165,-0.1998378485,-0.3389650881,-0.3431213498,-0.0419055969,0.5109502673,0.2115339041,0.3549255431,-0.1506305635,0.354680717,0.0829691291,0.2154671252,0.1893285513,0.481400162,0.2599649131,-0.0833791271,-0.2496428043,-0.0893777013,0.0335073769,0.0450351536,0.2059235871,0.1000142172,-0.2705875039,-0.2358311564,-0.0177404918,-0.0310977064,-0.2938491404,0.2493413538,0.1910058409,0.0510167181,-0.1072163507,-0.1244808882,0.2835575342,-0.0177356955,-0.048654791,-0.0244627539,0.1559522003,-0.3150945008,0.0947036147,0.002792709,0.2931680083,-0.0653572977,-0.1159678176,0.3695694506,0.2470180839,-0.1943728924,-0.1078828275,0.47724998,0.0235700458,0.0562779121,-0.1653675586,0.0340931639,-0.1504672617,0.09811984,0.118870832,-0.0381768085,0.5581243634,0.2744426727,0.0628850013,0.0449655354,-0.009465253,0.0558745414,-0.0433647297,0.161623776,-0.3240496814,0.1517037898,0.2595786154,-0.0845185369,-0.0927616879,-0.3839275837,0.2305375338,-0.0068245628,-0.0108318189,0.056660451,0.1574317366,0.0422499254,-0.0298721213,-0.1258571595,0.1956394464,0.2419599295,0.4526284933,-0.0378080606,-0.0898369253,0.0509364568,-0.0897590071,0.057606373,0.2135776281,0.2511664331,0.0213473067,0.3273831606,0.2114893645,-0.3109033704,-0.1465100497,-0.0894625038,0.162602976,-0.0390056521,0.1824658513,0.2884674966,0.2221601754,-0.1079074219,0.0726324394,-0.0501018167,-0.2391324341,0.2216691673,-0.0203706864,0.0873163044,0.2146704048,-0.3983554542,0.2782430649,-0.3043729365,0.297711134,0.026918361,-0.1578306109,-0.1325371712,0.3151072562,0.1054279134,-0.1702257246,0.7093059421,0.1851036102,-0.0610455498,-0.4948991537,-0.6101561785,0.1465800703,-0.252751261,0.3824195266,0.1277246773,0.0432365946,-0.0334133878,-0.170096606,0.0884748176,-0.3831410408,-0.3016197681,-0.2199166566,-0.3487504125,0.0569637902,-0.0039320649,-0.327652961,-0.2480899245,-0.2774356008,0.296877563,0.0330968425,0.1092961356,0.2283270806,-0.0240853932,-0.0029272011,-0.1338694543,0.0425414518,-0.2131209821,-0.2525308728,0.2263199538,-0.3307808638,-0.4148762524,0.0053341347,-0.1654051542,-0.107954219,-0.1699311584,-0.1671255529,-0.4542716444,-0.0601006523,0.2725510001,0.1361134201,-0.1978038698,0.24731341,-0.2358742207,-0.3502654433,0.0206142925,0.233303085,0.2701229751,-0.346455574,0.1270588785,-0.2727606595,0.4163656533,-0.1734109968,0.3034305274,-0.0132335713,-0.2822539806,0.0374224633,-0.113537088,0.3599158227,0.0470846072,-0.3589015305,0.2434345782,-0.0822774991,0.0136074536,0.402077198,-0.1739495993,-0.4085637331,-0.0783561021,0.0624613278,-0.1147775128,0.0256424621,-0.0314944163,-0.0834057927,0.1475787163,-0.1348057538,-0.3798795342,-0.0190615188,-0.1298001409,0.0030910887,0.3476361036,-0.1312892288,-0.0565833226,-0.6572769284,0.1057868525,-0.3817918003,0.2583281696,0.3123549521,-0.1255730391,-0.1988281757,-0.3021894693,0.0497129858,-0.0537781976,0.4213858247,-0.3975445032,-0.2893410623,0.0124575803,0.2310200185,-0.1031678542,0.2925239503,-0.3843233287,0.0349347331,0.0735548511,0.2050803751,-0.1165636554,-0.1865635365,0.5847241879,0.0028078151,-0.0771579295,-0.0467848033,-0.2452976555,-0.1711891294,-0.1777749509,0.137879625,-0.0371395834,0.1770626158,0.0487743579,-0.0572646819,-0.0136062568,0.1333001852,0.3394661844,-0.0653958544,-0.1489088833,0.2968918979,0.2098667324,0.0002459521,0.1981272697,0.5366670489,0.0771075338,-0.3925707936,-0.0280152932,0.0859426707,-0.2446949035,0.554420054,0.2504076064,0.1611799598,0.1057203189,0.2297707647,0.2516904175,0.0141212121,0.1945316643,0.2897646427,-0.1663683057,-0.3487012982,-0.1758626699,0.1841583252,0.0250898618,0.0122620249,-0.2254076749,0.2477867603,0.0433266833,-0.1938146651,0.2017298192,0.899856925,0.0118910791,0.065429166,0.0180119686,-0.4050844312,0.4948390424,-0.1011541337,-0.0127323763,-0.3347481489,-0.1197046563,-0.12264397,0.0274798851,0.1839458495,0.1274670362,0.0921378434,-0.1220777035,-0.0397133082,-0.0519053042,0.0041756872,0.2217673659,0.2907906175,-0.2299280614,-0.0618937761,0.323839426,0.1815609634,-0.0782326832,0.0144520774,-0.1412954926,-0.4487269819,-0.0469432212,-0.1071408764,-0.1663660407,-0.1920526624,0.1564887613,0.2727521956,-0.6254184246,0.0333614834,-0.1068415791,0.1485752463,-0.0291720405,-0.0389604606,0.0818330944,-0.2286265045,-0.2887102664,0.1722696424,-0.1986713707,0.2604533732,-0.098526828,-0.1170639843,0.3516009748,-0.0340218991,-0.0481850207,-0.316488117,0.1070092544,0.0128224269,0.0053871316,-0.0069221859,0.1413509697,-0.1581256241,-0.2788006067,0.22966896,0.1568029076,-0.1733350456,-0.1728959829,0.0406007729,-0.1150752455,-0.1173463613,0.3746272326,0.2017480582,-0.1037524641,0.2482941002,0.1537083834,-0.1898086816,-0.4193211496,-0.0646734238,0.5095945597,-0.5036644936,-0.1673623919,0.1142972037,-0.3388318419,0.2425370067,0.4636287391,0.165961504,-0.1558223218,-0.1936192811,-0.1742767245,-0.4288534224,0.1586353183,-0.4121941328,0.2065007836,-0.1194808558,0.3374957144,-0.1235231981,0.1794218272,-0.5173113942,-0.0299134236,-0.1870832294,0.0370240696,-0.0783389956,-0.2721550167,0.1812527329,-0.0699473247,0.2127100974,-0.0780235305,-0.2446213663,-0.329860419,-0.1667360365,0.0571284592,-0.073303178,0.2571614683,0.1744246781,0.2426323593,-0.3846096396,0.1428587586,0.2565899193,-0.1536628455,0.1053477228,-0.1176211908,-0.1242390051,-0.1483405828,-0.1051617488,-0.1772517264,0.1101925299,0.0871923864,0.0727777332,-0.0860333145,-0.1192615405,-0.1880307645,0.1019799188,0.4901584983,0.2104071975,-0.0629505888,0.3338208795,-0.0626747385,-0.1608938277,0.719874382,0.3552597761,-0.0252219141,0.0222609509,0.0760734305,-0.1219270602,0.383620739,-0.3694924712,-0.1482564211,-0.2198963463,-0.3824380636,0.2054023147,0.3845688105,0.2405638546,0.0585425049,0.0908453986,0.1397282481,0.4585956335,-0.2880103886,-0.1141364947,0.0668843985,0.1892652363,0.1522824317,0.4338438809,0.2446785867,0.4025593996,0.0815968364,0.1192693487,0.3037881553,-0.003767259,0.3395798504,0.2956933379,-0.1156295463,0.2456303537,0.0946558192,0.0218983237,-0.1363718361,-0.0927898884,0.6032199264,0.0590658225,0.094837606,-0.0614040084,0.1985834837,-0.133763954,0.049109485,-0.015471465,-0.2994345427,-0.3184757829,-0.1988072097,0.0736999586,0.0113245165,0.2655567825,-0.1855341792,-0.110948965,-0.3270041943,-0.0185571294,0.1115501598,0.3338699639,-0.2689291239,0.131902948,-0.1015625298,-0.1693012863,0.0745362863,0.3017919064,0.3978284895,-0.1152765006,0.2894566953,0.1163969785,-0.0573481731,-0.3456091881,-0.1934802234,0.0934320018,0.2574136853,-0.2218544632,0.052542448,0.2620060146,-0.325091213,0.0877832994,0.2161863148,0.1428124309,0.0476507023,0.2727470398,0.0272572227,-0.3008250892,-0.2494058609,0.1975773871,-0.3319015503,0.1632455438,0.0998378098,-0.1236869916,0.0189595483,-0.1818652451,0.1735938638,-0.0718908831,0.307025969,0.2222327739,-0.0490059257,-0.1651391536,-0.2170878351,-0.4863786399,0.2450276613,-0.0432994477,-0.0557960309,-0.1198467612,0.223275736,0.2528907061,-0.0337761454,-0.0673178807,-0.1585988551,-0.0740739331,-0.1220541969,-0.3301768899,0.0547077097,0.1542684287,0.1224835739,0.1682266891,-0.0547266826,-0.1594328135,-0.0507572554,0.150415808,-0.1974629015,-0.1916597635,-0.0777233317,0.1278348565,0.1954806596,0.0223775934,0.4594004452,-0.1684740186,-0.2510571778,-0.1394349486,-0.1120170206,-0.321967423,0.4179719687,0.2911700606,0.349086374,-0.0309617519,-0.0613577217,0.1796352118,-0.0416466296,0.2634922862,0.1286901534,-0.3708580434,0.1178338826,0.009356712,-0.1464712918,0.1834002733,0.2475122809,0.2790436447,-0.0473069996,-0.2027941942,-0.4436736703,0.4713127315,0.2330410779,-0.0729955807,-0.1081022024,0.3135801554,0.008469779,-0.1128596291,-0.4314526916,0.0632082671,0.4026740193,-0.0075977538,-0.2829242051,0.3086669743,-0.0057997676,-0.0738224015,-0.0647769123,-0.0931840613,-0.0147878453,-0.1707258821,-0.0851893425,-0.102797322]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1755","title":"Using select\/reordering datasets slows operations down immensely","comments":"You can use `Dataset.flatten_indices()` to make it fast after a select or shuffle.","body":"I am using portions of HF's helpful work in preparing \/ scoring the SQuAD 2.0 data. The problem I have is that after using `select` to re-ordering the dataset, computations slow down immensely where the total scoring process on 131k training examples would take maybe 3 minutes, now take over an hour.\r\n\r\nThe below example should be reproducible and I have ran myself down this path because I want to use HF's scoring functions and helpful data preparation, but use my own trainer. The training process uses shuffle and therefore the order I trained on no longer matches the original data set order. So, to score my results correctly, the original data set needs to match the order of the training. This requires that I: (1) collect the index for each row of data emitted during training, and (2) use this index information to re-order the datasets correctly so the orders match when I go to score.\r\n\r\n\r\nThe problem is, the dataset class starts performing very poorly as soon as you start manipulating its order by immense magnitudes.\r\n\r\n\r\n\r\n```\r\nfrom datasets import load_dataset, load_metric\r\nfrom transformers import BertTokenizerFast, BertForQuestionAnswering\r\nfrom elasticsearch import Elasticsearch\r\nimport numpy as np\r\nimport collections\r\nfrom tqdm.auto import tqdm\r\nimport torch\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-uncased')\r\nmax_length = 384 # The maximum length of a feature (question and context)\r\ndoc_stride = 128 # The authorized overlap between two part of the context when splitting it is needed.\r\npad_on_right = tokenizer.padding_side == \"right\"\r\nsquad_v2 = True\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef prepare_validation_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n # We keep the example_id that gave us this feature and we will store the offset mappings.\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token\r\n # position is part of the context or not.\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (list(o) if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef postprocess_qa_predictions(examples, features, starting_logits, ending_logits, n_best_size = 20, max_answer_length = 30):\r\n all_start_logits, all_end_logits = starting_logits, ending_logits\r\n # Build a map example to its corresponding features.\r\n example_id_to_index = {k: i for i, k in enumerate(examples[\"id\"])}\r\n features_per_example = collections.defaultdict(list)\r\n\r\n for i, feature in enumerate(features):\r\n features_per_example[example_id_to_index[feature[\"example_id\"]]].append(i)\r\n\r\n # The dictionaries we have to fill.\r\n predictions = collections.OrderedDict()\r\n\r\n # Logging.\r\n print(f\"Post-processing {len(examples)} example predictions split into {len(features)} features.\")\r\n\r\n # Let's loop over all the examples!\r\n for example_index, example in enumerate(tqdm(examples)):\r\n # Those are the indices of the features associated to the current example.\r\n feature_indices = features_per_example[example_index]\r\n\r\n min_null_score = None # Only used if squad_v2 is True.\r\n valid_answers = []\r\n\r\n context = example[\"context\"]\r\n # Looping through all the features associated to the current example.\r\n for feature_index in feature_indices:\r\n\r\n # We grab the predictions of the model for this feature.\r\n start_logits = all_start_logits[feature_index]\r\n end_logits = all_end_logits[feature_index]\r\n # This is what will allow us to map some the positions in our logits to span of texts in the original\r\n # context.\r\n offset_mapping = features[feature_index][\"offset_mapping\"]\r\n\r\n # Update minimum null prediction.\r\n cls_index = features[feature_index][\"input_ids\"].index(tokenizer.cls_token_id)\r\n feature_null_score = start_logits[cls_index] + end_logits[cls_index]\r\n if min_null_score is None or min_null_score < feature_null_score:\r\n min_null_score = feature_null_score\r\n\r\n # Go through all possibilities for the `n_best_size` greater start and end logits.\r\n start_indexes = np.argsort(start_logits)[-1 : -n_best_size - 1 : -1].tolist()\r\n end_indexes = np.argsort(end_logits)[-1 : -n_best_size - 1 : -1].tolist()\r\n for start_index in start_indexes:\r\n for end_index in end_indexes:\r\n # Don't consider out-of-scope answers, either because the indices are out of bounds or correspond\r\n # to part of the input_ids that are not in the context.\r\n if (\r\n start_index >= len(offset_mapping)\r\n or end_index >= len(offset_mapping)\r\n or offset_mapping[start_index] is None\r\n or offset_mapping[end_index] is None\r\n ):\r\n continue\r\n # Don't consider answers with a length that is either < 0 or > max_answer_length.\r\n if end_index < start_index or end_index - start_index + 1 > max_answer_length:\r\n continue\r\n\r\n start_char = offset_mapping[start_index][0]\r\n end_char = offset_mapping[end_index][1]\r\n valid_answers.append(\r\n {\r\n \"score\": start_logits[start_index] + end_logits[end_index],\r\n \"text\": context[start_char: end_char]\r\n }\r\n )\r\n\r\n\r\n if len(valid_answers) > 0:\r\n best_answer = sorted(valid_answers, key=lambda x: x[\"score\"], reverse=True)[0]\r\n else:\r\n # In the very rare edge case we have not a single non-null prediction, we create a fake prediction to avoid\r\n # failure.\r\n best_answer = {\"text\": \"\", \"score\": 0.0}\r\n\r\n # Let's pick our final answer: the best one or the null answer (only for squad_v2)\r\n if not squad_v2:\r\n predictions[example[\"id\"]] = best_answer[\"text\"]\r\n else:\r\n answer = best_answer[\"text\"] if best_answer[\"score\"] > min_null_score else \"\"\r\n predictions[example[\"id\"]] = answer\r\n\r\n return predictions\r\n\r\n\r\n\r\n# build base examples, features from training data\r\nexamples = load_dataset(\"squad_v2\").shuffle(seed=5)['train']\r\nfeatures = load_dataset(\"squad_v2\").shuffle(seed=5)['train'].map(\r\n prepare_validation_features,\r\n batched=True,\r\n remove_columns=['answers', 'context', 'id', 'question', 'title'])\r\n\r\n# sim some shuffled training indices that we want to use to re-order the data to compare how we did\r\nshuffle_idx = np.arange(0, 131754)\r\nnp.random.shuffle(shuffle_idx)\r\n# create a new dataset with rows selected following the training shuffle\r\nfeatures = features.select(indices=shuffle_idx)\r\n# get unique example ids to match with the \"example\" data\r\nid_list = list(dict.fromkeys(features['example_id']))\r\n# now search for their index positions; load elastic search\r\nes = Elasticsearch([{'host': 'localhost'}]).ping()\r\n# add an index to the id column for the examples\r\nexamples.add_elasticsearch_index(column='id')\r\n# search the examples for their index position\r\nexample_idx = [examples.search(index_name='id', query=i, k=1).indices for i in id_list]\r\n# drop the elastic search\r\nexamples.drop_index(index_name='id')\r\n# put examples in the right order\r\nexamples = examples.select(indices=example_idx)\r\n\r\n# generate some fake data\r\nlogits = {'starting_logits': torch.randn(131754, 384), 'ending_logits': torch.randn(131754, 384)}\r\n\r\n\r\ndef score_squad(logits, n_best_size, max_answer):\r\n # proceed with QA calculation\r\n final_predictions = postprocess_qa_predictions(examples=examples,\r\n features=features,\r\n starting_logits=logits['starting_logits'],\r\n ending_logits=logits['ending_logits'],\r\n n_best_size=20,\r\n max_answer_length=30)\r\n metric = load_metric(\"squad_v2\")\r\n formatted_predictions = [{\"id\": k, \"prediction_text\": v, \"no_answer_probability\": 0.0} for k, v in final_predictions.items()]\r\n references = [{\"id\": ex[\"id\"], \"answers\": ex[\"answers\"]} for ex in examples]\r\n metrics = metric.compute(predictions=formatted_predictions, references=references)\r\n return metrics\r\n\r\nmetrics = score_squad(logits, n_best_size=20, max_answer=30)\r\n```\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":13,"text":"Using select\/reordering datasets slows operations down immensely\nI am using portions of HF's helpful work in preparing \/ scoring the SQuAD 2.0 data. The problem I have is that after using `select` to re-ordering the dataset, computations slow down immensely where the total scoring process on 131k training examples would take maybe 3 minutes, now take over an hour.\r\n\r\nThe below example should be reproducible and I have ran myself down this path because I want to use HF's scoring functions and helpful data preparation, but use my own trainer. The training process uses shuffle and therefore the order I trained on no longer matches the original data set order. So, to score my results correctly, the original data set needs to match the order of the training. This requires that I: (1) collect the index for each row of data emitted during training, and (2) use this index information to re-order the datasets correctly so the orders match when I go to score.\r\n\r\n\r\nThe problem is, the dataset class starts performing very poorly as soon as you start manipulating its order by immense magnitudes.\r\n\r\n\r\n\r\n```\r\nfrom datasets import load_dataset, load_metric\r\nfrom transformers import BertTokenizerFast, BertForQuestionAnswering\r\nfrom elasticsearch import Elasticsearch\r\nimport numpy as np\r\nimport collections\r\nfrom tqdm.auto import tqdm\r\nimport torch\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-uncased')\r\nmax_length = 384 # The maximum length of a feature (question and context)\r\ndoc_stride = 128 # The authorized overlap between two part of the context when splitting it is needed.\r\npad_on_right = tokenizer.padding_side == \"right\"\r\nsquad_v2 = True\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef prepare_validation_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n # We keep the example_id that gave us this feature and we will store the offset mappings.\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token\r\n # position is part of the context or not.\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (list(o) if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef postprocess_qa_predictions(examples, features, starting_logits, ending_logits, n_best_size = 20, max_answer_length = 30):\r\n all_start_logits, all_end_logits = starting_logits, ending_logits\r\n # Build a map example to its corresponding features.\r\n example_id_to_index = {k: i for i, k in enumerate(examples[\"id\"])}\r\n features_per_example = collections.defaultdict(list)\r\n\r\n for i, feature in enumerate(features):\r\n features_per_example[example_id_to_index[feature[\"example_id\"]]].append(i)\r\n\r\n # The dictionaries we have to fill.\r\n predictions = collections.OrderedDict()\r\n\r\n # Logging.\r\n print(f\"Post-processing {len(examples)} example predictions split into {len(features)} features.\")\r\n\r\n # Let's loop over all the examples!\r\n for example_index, example in enumerate(tqdm(examples)):\r\n # Those are the indices of the features associated to the current example.\r\n feature_indices = features_per_example[example_index]\r\n\r\n min_null_score = None # Only used if squad_v2 is True.\r\n valid_answers = []\r\n\r\n context = example[\"context\"]\r\n # Looping through all the features associated to the current example.\r\n for feature_index in feature_indices:\r\n\r\n # We grab the predictions of the model for this feature.\r\n start_logits = all_start_logits[feature_index]\r\n end_logits = all_end_logits[feature_index]\r\n # This is what will allow us to map some the positions in our logits to span of texts in the original\r\n # context.\r\n offset_mapping = features[feature_index][\"offset_mapping\"]\r\n\r\n # Update minimum null prediction.\r\n cls_index = features[feature_index][\"input_ids\"].index(tokenizer.cls_token_id)\r\n feature_null_score = start_logits[cls_index] + end_logits[cls_index]\r\n if min_null_score is None or min_null_score < feature_null_score:\r\n min_null_score = feature_null_score\r\n\r\n # Go through all possibilities for the `n_best_size` greater start and end logits.\r\n start_indexes = np.argsort(start_logits)[-1 : -n_best_size - 1 : -1].tolist()\r\n end_indexes = np.argsort(end_logits)[-1 : -n_best_size - 1 : -1].tolist()\r\n for start_index in start_indexes:\r\n for end_index in end_indexes:\r\n # Don't consider out-of-scope answers, either because the indices are out of bounds or correspond\r\n # to part of the input_ids that are not in the context.\r\n if (\r\n start_index >= len(offset_mapping)\r\n or end_index >= len(offset_mapping)\r\n or offset_mapping[start_index] is None\r\n or offset_mapping[end_index] is None\r\n ):\r\n continue\r\n # Don't consider answers with a length that is either < 0 or > max_answer_length.\r\n if end_index < start_index or end_index - start_index + 1 > max_answer_length:\r\n continue\r\n\r\n start_char = offset_mapping[start_index][0]\r\n end_char = offset_mapping[end_index][1]\r\n valid_answers.append(\r\n {\r\n \"score\": start_logits[start_index] + end_logits[end_index],\r\n \"text\": context[start_char: end_char]\r\n }\r\n )\r\n\r\n\r\n if len(valid_answers) > 0:\r\n best_answer = sorted(valid_answers, key=lambda x: x[\"score\"], reverse=True)[0]\r\n else:\r\n # In the very rare edge case we have not a single non-null prediction, we create a fake prediction to avoid\r\n # failure.\r\n best_answer = {\"text\": \"\", \"score\": 0.0}\r\n\r\n # Let's pick our final answer: the best one or the null answer (only for squad_v2)\r\n if not squad_v2:\r\n predictions[example[\"id\"]] = best_answer[\"text\"]\r\n else:\r\n answer = best_answer[\"text\"] if best_answer[\"score\"] > min_null_score else \"\"\r\n predictions[example[\"id\"]] = answer\r\n\r\n return predictions\r\n\r\n\r\n\r\n# build base examples, features from training data\r\nexamples = load_dataset(\"squad_v2\").shuffle(seed=5)['train']\r\nfeatures = load_dataset(\"squad_v2\").shuffle(seed=5)['train'].map(\r\n prepare_validation_features,\r\n batched=True,\r\n remove_columns=['answers', 'context', 'id', 'question', 'title'])\r\n\r\n# sim some shuffled training indices that we want to use to re-order the data to compare how we did\r\nshuffle_idx = np.arange(0, 131754)\r\nnp.random.shuffle(shuffle_idx)\r\n# create a new dataset with rows selected following the training shuffle\r\nfeatures = features.select(indices=shuffle_idx)\r\n# get unique example ids to match with the \"example\" data\r\nid_list = list(dict.fromkeys(features['example_id']))\r\n# now search for their index positions; load elastic search\r\nes = Elasticsearch([{'host': 'localhost'}]).ping()\r\n# add an index to the id column for the examples\r\nexamples.add_elasticsearch_index(column='id')\r\n# search the examples for their index position\r\nexample_idx = [examples.search(index_name='id', query=i, k=1).indices for i in id_list]\r\n# drop the elastic search\r\nexamples.drop_index(index_name='id')\r\n# put examples in the right order\r\nexamples = examples.select(indices=example_idx)\r\n\r\n# generate some fake data\r\nlogits = {'starting_logits': torch.randn(131754, 384), 'ending_logits': torch.randn(131754, 384)}\r\n\r\n\r\ndef score_squad(logits, n_best_size, max_answer):\r\n # proceed with QA calculation\r\n final_predictions = postprocess_qa_predictions(examples=examples,\r\n features=features,\r\n starting_logits=logits['starting_logits'],\r\n ending_logits=logits['ending_logits'],\r\n n_best_size=20,\r\n max_answer_length=30)\r\n metric = load_metric(\"squad_v2\")\r\n formatted_predictions = [{\"id\": k, \"prediction_text\": v, \"no_answer_probability\": 0.0} for k, v in final_predictions.items()]\r\n references = [{\"id\": ex[\"id\"], \"answers\": ex[\"answers\"]} for ex in examples]\r\n metrics = metric.compute(predictions=formatted_predictions, references=references)\r\n return metrics\r\n\r\nmetrics = score_squad(logits, n_best_size=20, max_answer=30)\r\n```\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\nYou can use `Dataset.flatten_indices()` to make it fast after a select or shuffle.","embeddings":[-0.2229693085,0.2477027923,-0.0138683552,-0.1000716835,-0.003328807,-0.1465102434,-0.0916016847,0.1460768282,-0.3497589529,0.1094219536,-0.2728040516,0.3873890042,0.2126539499,-0.1992182434,-0.0613091849,-0.0526460782,-0.1135229096,0.0000033445,-0.0314249322,-0.0937863365,-0.1005084962,-0.3484925032,-0.3477401137,-0.145543471,-0.2218254358,-0.1603693813,0.0549748018,0.3512459695,-0.1260461956,-0.3404070437,0.0977792069,0.5284638405,-0.0601430126,0.3460890353,-0.0001203739,-0.1140628681,-0.301356703,-0.0590985864,-0.4936899543,0.0832120925,-0.0697157905,-0.0438395105,-0.1616155952,0.0779754221,-0.1216855645,-0.1129818335,-0.0549107939,-0.3639614284,0.4560742378,-0.0008517363,0.1208517849,0.220481649,-0.5714429021,-0.2556709349,0.1670394242,0.0162417125,-0.0953832343,-0.0365809016,0.4183206856,0.1212784946,-0.1140871421,0.3431174755,-0.0950435996,-0.1506789178,0.2451233715,0.1294504702,-0.2699114084,0.1256278604,0.005174641,0.0222797319,0.2026703358,-0.1132044941,-0.5826923847,-0.3719269037,0.341755867,-0.0127385156,0.2049164623,-0.2810357511,0.4607771039,-0.0416236669,-0.306301266,0.2016853541,0.0358903036,0.0214370992,0.1613190323,0.2261903286,0.0236478597,-0.0106987497,-0.0192314014,-0.1237981394,0.1931634396,-0.187747255,0.0458304808,0.3385604024,-0.485106051,-0.2102170885,-0.1109896228,-0.2480468452,0.030924825,0.1953477114,0.0991410986,0.4193668067,0.0793582574,-0.3105548918,0.1926971674,0.2943916023,-0.0932467952,0.3354053497,0.0661656633,-0.0558295101,-0.0670939237,-0.027262589,0.05751976,-0.3548845947,-0.1153057069,0.0621315576,-0.416890353,-0.1396674961,-0.4675750732,0.0101275723,-0.2367890626,0.026424244,0.3597140312,0.3569377363,-0.1339115798,-0.1230955496,0.0204742644,0.1510910392,-0.3827776909,0.1194274426,-0.1549317241,-0.1807512939,-0.1282603741,0.4981126487,0.0682006031,0.0105687119,-0.0445278138,0.1479117125,0.6504570246,0.1346221715,0.1789738238,-0.0625602156,0.3049418628,0.1330021918,-0.1644210666,0.0284394585,-0.0693844259,-0.0828607082,-0.2640684247,0.2404960692,-0.4224526882,-0.1730779409,0.2738597393,0.1426498741,0.053241685,-0.1324371696,-0.0261703134,0.1152576506,0.0700723827,-0.3271812797,0.0334452242,-0.2640089989,0.0580386445,0.0417545848,0.1055177599,0.264410466,-0.4662193656,-0.0805213153,0.0947195888,0.5643754601,0.2235662043,0.3344970942,-0.2951166928,0.0774924085,0.0623126552,0.2190618217,0.4485758841,-0.1682099849,-0.4064477086,0.0793168619,-0.3333152533,0.1301809698,0.1100042984,0.2987503409,0.4087423086,-0.1437912285,-0.0917459428,0.2847900689,0.0783123896,-0.0005592417,-0.2430839092,-0.1494386792,-0.1460276693,0.2493930757,-0.0854728371,0.0004539371,0.0658750907,0.2544437647,0.0936773643,-0.212437734,-0.1569136828,0.1440961361,-0.0487086475,0.2361472845,0.009082377,-0.2017503828,-0.2624258995,0.2542325854,-0.1482537538,-0.0874456838,0.1352851391,-0.2490602881,-0.2674162686,-0.2243238986,-0.3613672256,-0.0841277689,-0.0326935686,-0.0000497185,0.2648528516,-0.374301374,-0.1358233392,0.2545201778,-0.29800722,0.1125548556,0.1867353618,0.3353077769,0.0942640156,-0.0836718678,-0.1753809452,0.223179847,0.0413799025,-0.0102596274,-0.0425318852,0.4816074669,-0.0305584166,-0.2010594904,0.300344646,0.1919036806,0.1189256087,-0.0048329839,0.1848534644,-0.0567650683,-0.1038584858,-0.0857224688,-0.4112776518,0.7307934761,-0.6390340924,0.3906633854,-0.0081895739,-0.3914916515,-0.1364453137,0.2339695692,-0.0671845451,0.028089473,0.1946619898,0.0692243576,0.3068402112,0.0219235756,-0.1708946526,0.1148176268,0.3807358742,0.1116098538,-0.116190806,0.0593309514,0.1933869421,0.0581158064,-0.1526626647,-0.4381038547,0.2674148977,0.2144753784,-0.0098131234,-0.0738034323,-0.0876859352,-0.0899261758,-0.0617198683,-0.0450996533,-0.0257454459,0.3969434798,0.3169572949,0.2558146119,-0.2715595961,-0.0461911708,0.0497649349,-0.0269536227,-0.3187946081,0.3653787971,-0.2655484378,0.3262257278,-0.0995716602,0.0995328948,-0.2339483351,-0.1894762069,0.2998906076,-0.0532458164,0.0420344025,0.2280463427,0.0082486607,0.2510022223,0.193622604,-0.4015026987,0.0486457944,0.1283143908,-0.0853393897,-0.0386849381,0.1792995185,0.2340388447,0.299641192,0.1631267071,-0.2064593881,-0.0706808493,-0.3488852978,-0.0585122183,0.0901514664,0.1498829275,-0.0947868451,-0.1986880302,-0.505982399,-0.336107105,-0.0967887044,0.1074076146,-0.189230904,0.0047539468,-0.2101000994,0.0690367818,0.1734790504,-0.4781151414,0.0168381706,0.0162553135,0.0344494693,0.0098374942,0.066954568,0.1476120055,-0.1610152125,-0.1116630211,-0.1796435267,0.0880680829,-0.3538632095,-0.2686766982,0.2319982201,-0.1055770144,-0.008286356,-0.4196908772,-0.0765763521,-0.5765689611,0.2538782358,0.0160375256,-0.1266462505,-0.0915807411,-0.0984586105,-0.3653621674,0.2959348261,0.3322464824,-0.2109398097,0.032138925,0.110281691,-0.0774022862,0.0623431467,0.5486826301,0.0462288484,-0.4003959298,0.1239522099,0.1109858006,1.1333417892,-0.0141810263,0.1114459336,-0.0653473362,0.1698570549,0.0690061226,-0.349622041,-0.1671998799,0.2396910191,-0.006079861,0.0631310791,0.3585751355,0.0024383776,0.2197257131,0.3613546193,0.4713778198,-0.1210378334,-0.1453615576,0.1049261838,0.1241666153,0.1314238906,0.2344905585,0.330524236,-0.254545182,-0.1899589747,-0.0217920169,-0.2138453871,0.2415392846,-0.1783572584,-0.4820561707,0.0176173411,-0.0887342021,0.1038669348,0.1211590916,0.0310366135,0.3065715432,-0.0894950405,0.4382134676,-0.1552442014,0.6033061147,0.010327599,-0.0748458281,0.2094712108,-0.3021306992,0.031763982,-0.0028447087,-0.3103168607,0.1381147504,0.0280377362,0.2992950976,-0.3955968618,-0.146859467,0.2374914289,0.0136861531,0.0182699207,0.0168006681,-0.157424897,0.045749519,0.0769337416,0.4116093814,-0.276858598,0.1689107865,0.0499895476,0.1534054875,-0.2058267593,0.0772600695,0.108990714,0.0978837684,0.232063368,0.3968159556,0.2968211472,0.5759822726,-0.0820016488,0.2667078078,0.3654349148,-0.0405631289,-0.0320802666,0.1777066886,0.0213753227,0.1245086268,0.7451142669,0.1889926642,0.1802190989,-0.0091554085,0.3382091224,-0.2709101439,-0.0570228845,0.3755310476,-0.1951373965,-0.1541287303,-0.2498485744,0.3288463056,0.0089340182,-0.0365891941,-0.1135575995,0.5571148992,-0.1169481799,0.369922787,0.6643725634,1.153963685,-0.3158586621,0.2839170396,0.0857243538,-0.0627596825,0.2783313394,-0.4571999311,0.3110564649,-0.3608015478,-0.5183977485,0.1271019131,-0.2768598795,0.1780613959,0.3513542414,0.0487249717,0.2352935076,-0.3827722371,0.0779589638,-0.3664308488,0.2086066455,0.1052867845,-0.5642211437,-0.0995811895,0.0342080556,0.0851489305,-0.161505565,-0.1026572734,0.1778127551,-0.0776009262,0.2319298089,-0.0566128828,-0.1408687979,-0.3118817806,-0.0234600436,-0.049573943,-0.3439301848,0.6047793627,-0.0415184759,0.1774502844,0.1372202486,0.2033912241,0.1746311635,0.170843333,0.4651058614,0.4565700293,-0.1060808674,0.0508966371,0.1657200456,0.1711342186,-0.1148238555,0.1943956017,-0.431830138,-0.3854596615,0.1150658503,0.0980385244,-0.1946828216,-0.0203516781,0.2804602087,0.1253672391,-0.2001907676,0.0503734462,0.0382371284,-0.0283566099,0.5486469865,0.1635395885,-0.2518385649,-0.2608955801,0.6345748305,0.0965518653,-0.0834737271,0.4297213554,-0.0487744696,-0.3911808431,-0.1019709036,0.2128725201,0.2624303102,-0.3412107825,-0.0872746333,-0.1879292428,-0.279442668,0.1587748379,0.2018291652,0.1933795959,-0.0701102465,-0.2999260128,0.1244067475,0.3185898662,0.0201228522,0.6046158075,0.0398607478,-0.3336846828,0.2227038443,-0.5289216042,0.3563667834,-0.2616191208,-0.1069983616,-0.248258993,0.2899232805,-0.5045489073,-0.1483784765,0.2586853504,-0.1163726002,0.0048723738,-0.0601766892,-0.1983085126,-0.1448252201,-0.1970014423,0.0980262831,-0.1997065097,0.1243187636,-0.0188031849,-0.2507639229,-0.2704802155,-0.4096262753,0.2014555335,0.3396630883,0.2517517805,-0.079448767,0.2015297562,-0.2823911607,0.1224734187,-0.0024231735,0.1681148112,-0.1667636335,0.0411642306,0.245342955,-0.1037383303,-0.1357138753,-0.0697518066,0.2789188623,0.5007079244,0.0211990867,0.2153330296,-0.2984635234,-0.0545401052,0.4007716179,0.4186480641,0.6963021159,0.2969825566,-0.6313238144,0.0892218128,0.0435882285,-0.0628520176,-0.3881573379,0.021162672,-0.1007334739,0.0100013483,0.1811839938,-0.186611712,0.3341404498,0.1153280437,0.1421733499,-0.0810378119,-0.0633160174,0.1096307114,0.221096471,0.215869844,-0.1135800108,0.3845514953,-0.1954146177,0.1096801981,0.0798891932,0.1793094426,0.4664360583,-0.0411300138,0.3082848787,-0.3080215752,-0.2969990373,-0.2421714216,0.4442578256,0.1990970224,-0.0469324812,0.0693915188,0.1385658681,-0.0666203275,-0.5457808375,-0.0916194022,0.1064575464,-0.084910579,-0.119918935,-0.0566291884,-0.1938597113,-0.165054068,-0.1094780266,-0.3108012974,0.129476279,0.039264407,-0.1247126535,0.4548836946,-0.3548500538,-0.07787738,-0.2523677051,0.3607309163,-0.1750518978,-0.0198753662,-0.0690099299,0.3045783639,0.3067370653,0.0552976057,-0.1740990728,0.1050492302,-0.3952314556,-0.0766809285,-0.0727442056,0.0688731596,0.2560726404,-0.0455823876,-0.2837682664,-0.3933627307,0.4206233621,-0.018373549,-0.1651886702,-0.6061081886,0.2266858965,0.0555941425,-0.2718512416,0.0316607803,0.2860613763,0.0470649041,0.4065903723,-0.0323086232,0.061348442,-0.000361293,-0.2930675149,-0.1541664749,0.0500857383,0.2389614582,0.0626749992,0.1489731073,0.5292367339,0.1847597957,-0.0745077059,-0.3575538099,0.0056569469,-0.515768826,-0.0746005774,-0.0146510173,0.0519744009,0.4381356239,0.4754748344,0.1372810155,0.0549258851,0.1292781681,-0.59433496,0.0565181337,0.08937224,-0.1599140167,-0.0063516195,-0.2795810997,-0.092877619,-0.2524853945,-0.2728768587,0.1100642011,0.3300592899,-0.0536877029,0.1298539042,0.0280912239,0.0561371706,0.2114116848,0.045230113,-0.0933138877,-0.0210877396,0.0484026857,0.1092743725,0.1135340929,-0.2138202935,-0.2178073376,0.0998906493,-0.2642327845,0.3201899827,-0.1137272641,-0.1714500338,-0.1460872442,0.1175538823,-0.0223562326,0.1953673959,-0.5066771507,0.1262384802,0.1293935478,0.517236352,-0.2342355847,0.4483053386,0.0020684863,0.4348350167,-0.2447377443,-0.0530665629,0.4661418796,0.0571614131,-0.1567380875,-0.0178665519,0.3398124576,0.2781736553,0.0795370042,0.1954004467,-0.1007859632,0.3035782278,-0.0931361243,-0.0184627101,0.1331180185,-0.0543180071,-0.1423792094,-0.2693759203,-0.0274897944,-0.0685155839,-0.1467158496,-0.2096290737,-0.269469887]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1755","title":"Using select\/reordering datasets slows operations down immensely","comments":"Thanks for the input! I gave that a try by adding this after my selection \/ reordering operations, but before the big computation task of `score_squad`\r\n\r\n```\r\nexamples = examples.flatten_indices()\r\nfeatures = features.flatten_indices()\r\n```\r\n\r\nThat helped quite a bit!","body":"I am using portions of HF's helpful work in preparing \/ scoring the SQuAD 2.0 data. The problem I have is that after using `select` to re-ordering the dataset, computations slow down immensely where the total scoring process on 131k training examples would take maybe 3 minutes, now take over an hour.\r\n\r\nThe below example should be reproducible and I have ran myself down this path because I want to use HF's scoring functions and helpful data preparation, but use my own trainer. The training process uses shuffle and therefore the order I trained on no longer matches the original data set order. So, to score my results correctly, the original data set needs to match the order of the training. This requires that I: (1) collect the index for each row of data emitted during training, and (2) use this index information to re-order the datasets correctly so the orders match when I go to score.\r\n\r\n\r\nThe problem is, the dataset class starts performing very poorly as soon as you start manipulating its order by immense magnitudes.\r\n\r\n\r\n\r\n```\r\nfrom datasets import load_dataset, load_metric\r\nfrom transformers import BertTokenizerFast, BertForQuestionAnswering\r\nfrom elasticsearch import Elasticsearch\r\nimport numpy as np\r\nimport collections\r\nfrom tqdm.auto import tqdm\r\nimport torch\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-uncased')\r\nmax_length = 384 # The maximum length of a feature (question and context)\r\ndoc_stride = 128 # The authorized overlap between two part of the context when splitting it is needed.\r\npad_on_right = tokenizer.padding_side == \"right\"\r\nsquad_v2 = True\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef prepare_validation_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n # We keep the example_id that gave us this feature and we will store the offset mappings.\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token\r\n # position is part of the context or not.\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (list(o) if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef postprocess_qa_predictions(examples, features, starting_logits, ending_logits, n_best_size = 20, max_answer_length = 30):\r\n all_start_logits, all_end_logits = starting_logits, ending_logits\r\n # Build a map example to its corresponding features.\r\n example_id_to_index = {k: i for i, k in enumerate(examples[\"id\"])}\r\n features_per_example = collections.defaultdict(list)\r\n\r\n for i, feature in enumerate(features):\r\n features_per_example[example_id_to_index[feature[\"example_id\"]]].append(i)\r\n\r\n # The dictionaries we have to fill.\r\n predictions = collections.OrderedDict()\r\n\r\n # Logging.\r\n print(f\"Post-processing {len(examples)} example predictions split into {len(features)} features.\")\r\n\r\n # Let's loop over all the examples!\r\n for example_index, example in enumerate(tqdm(examples)):\r\n # Those are the indices of the features associated to the current example.\r\n feature_indices = features_per_example[example_index]\r\n\r\n min_null_score = None # Only used if squad_v2 is True.\r\n valid_answers = []\r\n\r\n context = example[\"context\"]\r\n # Looping through all the features associated to the current example.\r\n for feature_index in feature_indices:\r\n\r\n # We grab the predictions of the model for this feature.\r\n start_logits = all_start_logits[feature_index]\r\n end_logits = all_end_logits[feature_index]\r\n # This is what will allow us to map some the positions in our logits to span of texts in the original\r\n # context.\r\n offset_mapping = features[feature_index][\"offset_mapping\"]\r\n\r\n # Update minimum null prediction.\r\n cls_index = features[feature_index][\"input_ids\"].index(tokenizer.cls_token_id)\r\n feature_null_score = start_logits[cls_index] + end_logits[cls_index]\r\n if min_null_score is None or min_null_score < feature_null_score:\r\n min_null_score = feature_null_score\r\n\r\n # Go through all possibilities for the `n_best_size` greater start and end logits.\r\n start_indexes = np.argsort(start_logits)[-1 : -n_best_size - 1 : -1].tolist()\r\n end_indexes = np.argsort(end_logits)[-1 : -n_best_size - 1 : -1].tolist()\r\n for start_index in start_indexes:\r\n for end_index in end_indexes:\r\n # Don't consider out-of-scope answers, either because the indices are out of bounds or correspond\r\n # to part of the input_ids that are not in the context.\r\n if (\r\n start_index >= len(offset_mapping)\r\n or end_index >= len(offset_mapping)\r\n or offset_mapping[start_index] is None\r\n or offset_mapping[end_index] is None\r\n ):\r\n continue\r\n # Don't consider answers with a length that is either < 0 or > max_answer_length.\r\n if end_index < start_index or end_index - start_index + 1 > max_answer_length:\r\n continue\r\n\r\n start_char = offset_mapping[start_index][0]\r\n end_char = offset_mapping[end_index][1]\r\n valid_answers.append(\r\n {\r\n \"score\": start_logits[start_index] + end_logits[end_index],\r\n \"text\": context[start_char: end_char]\r\n }\r\n )\r\n\r\n\r\n if len(valid_answers) > 0:\r\n best_answer = sorted(valid_answers, key=lambda x: x[\"score\"], reverse=True)[0]\r\n else:\r\n # In the very rare edge case we have not a single non-null prediction, we create a fake prediction to avoid\r\n # failure.\r\n best_answer = {\"text\": \"\", \"score\": 0.0}\r\n\r\n # Let's pick our final answer: the best one or the null answer (only for squad_v2)\r\n if not squad_v2:\r\n predictions[example[\"id\"]] = best_answer[\"text\"]\r\n else:\r\n answer = best_answer[\"text\"] if best_answer[\"score\"] > min_null_score else \"\"\r\n predictions[example[\"id\"]] = answer\r\n\r\n return predictions\r\n\r\n\r\n\r\n# build base examples, features from training data\r\nexamples = load_dataset(\"squad_v2\").shuffle(seed=5)['train']\r\nfeatures = load_dataset(\"squad_v2\").shuffle(seed=5)['train'].map(\r\n prepare_validation_features,\r\n batched=True,\r\n remove_columns=['answers', 'context', 'id', 'question', 'title'])\r\n\r\n# sim some shuffled training indices that we want to use to re-order the data to compare how we did\r\nshuffle_idx = np.arange(0, 131754)\r\nnp.random.shuffle(shuffle_idx)\r\n# create a new dataset with rows selected following the training shuffle\r\nfeatures = features.select(indices=shuffle_idx)\r\n# get unique example ids to match with the \"example\" data\r\nid_list = list(dict.fromkeys(features['example_id']))\r\n# now search for their index positions; load elastic search\r\nes = Elasticsearch([{'host': 'localhost'}]).ping()\r\n# add an index to the id column for the examples\r\nexamples.add_elasticsearch_index(column='id')\r\n# search the examples for their index position\r\nexample_idx = [examples.search(index_name='id', query=i, k=1).indices for i in id_list]\r\n# drop the elastic search\r\nexamples.drop_index(index_name='id')\r\n# put examples in the right order\r\nexamples = examples.select(indices=example_idx)\r\n\r\n# generate some fake data\r\nlogits = {'starting_logits': torch.randn(131754, 384), 'ending_logits': torch.randn(131754, 384)}\r\n\r\n\r\ndef score_squad(logits, n_best_size, max_answer):\r\n # proceed with QA calculation\r\n final_predictions = postprocess_qa_predictions(examples=examples,\r\n features=features,\r\n starting_logits=logits['starting_logits'],\r\n ending_logits=logits['ending_logits'],\r\n n_best_size=20,\r\n max_answer_length=30)\r\n metric = load_metric(\"squad_v2\")\r\n formatted_predictions = [{\"id\": k, \"prediction_text\": v, \"no_answer_probability\": 0.0} for k, v in final_predictions.items()]\r\n references = [{\"id\": ex[\"id\"], \"answers\": ex[\"answers\"]} for ex in examples]\r\n metrics = metric.compute(predictions=formatted_predictions, references=references)\r\n return metrics\r\n\r\nmetrics = score_squad(logits, n_best_size=20, max_answer=30)\r\n```\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n","comment_length":39,"text":"Using select\/reordering datasets slows operations down immensely\nI am using portions of HF's helpful work in preparing \/ scoring the SQuAD 2.0 data. The problem I have is that after using `select` to re-ordering the dataset, computations slow down immensely where the total scoring process on 131k training examples would take maybe 3 minutes, now take over an hour.\r\n\r\nThe below example should be reproducible and I have ran myself down this path because I want to use HF's scoring functions and helpful data preparation, but use my own trainer. The training process uses shuffle and therefore the order I trained on no longer matches the original data set order. So, to score my results correctly, the original data set needs to match the order of the training. This requires that I: (1) collect the index for each row of data emitted during training, and (2) use this index information to re-order the datasets correctly so the orders match when I go to score.\r\n\r\n\r\nThe problem is, the dataset class starts performing very poorly as soon as you start manipulating its order by immense magnitudes.\r\n\r\n\r\n\r\n```\r\nfrom datasets import load_dataset, load_metric\r\nfrom transformers import BertTokenizerFast, BertForQuestionAnswering\r\nfrom elasticsearch import Elasticsearch\r\nimport numpy as np\r\nimport collections\r\nfrom tqdm.auto import tqdm\r\nimport torch\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-uncased')\r\nmax_length = 384 # The maximum length of a feature (question and context)\r\ndoc_stride = 128 # The authorized overlap between two part of the context when splitting it is needed.\r\npad_on_right = tokenizer.padding_side == \"right\"\r\nsquad_v2 = True\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef prepare_validation_features(examples):\r\n # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results\r\n # in one example possible giving several features when a context is long, each of those features having a\r\n # context that overlaps a bit the context of the previous feature.\r\n tokenized_examples = tokenizer(\r\n examples[\"question\" if pad_on_right else \"context\"],\r\n examples[\"context\" if pad_on_right else \"question\"],\r\n truncation=\"only_second\" if pad_on_right else \"only_first\",\r\n max_length=max_length,\r\n stride=doc_stride,\r\n return_overflowing_tokens=True,\r\n return_offsets_mapping=True,\r\n padding=\"max_length\",\r\n )\r\n\r\n # Since one example might give us several features if it has a long context, we need a map from a feature to\r\n # its corresponding example. This key gives us just that.\r\n sample_mapping = tokenized_examples.pop(\"overflow_to_sample_mapping\")\r\n\r\n # We keep the example_id that gave us this feature and we will store the offset mappings.\r\n tokenized_examples[\"example_id\"] = []\r\n\r\n for i in range(len(tokenized_examples[\"input_ids\"])):\r\n # Grab the sequence corresponding to that example (to know what is the context and what is the question).\r\n sequence_ids = tokenized_examples.sequence_ids(i)\r\n context_index = 1 if pad_on_right else 0\r\n\r\n # One example can give several spans, this is the index of the example containing this span of text.\r\n sample_index = sample_mapping[i]\r\n tokenized_examples[\"example_id\"].append(examples[\"id\"][sample_index])\r\n\r\n # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token\r\n # position is part of the context or not.\r\n tokenized_examples[\"offset_mapping\"][i] = [\r\n (list(o) if sequence_ids[k] == context_index else None)\r\n for k, o in enumerate(tokenized_examples[\"offset_mapping\"][i])\r\n ]\r\n\r\n return tokenized_examples\r\n\r\n# from https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/question_answering.ipynb#scrollTo=941LPhDWeYv-\r\ndef postprocess_qa_predictions(examples, features, starting_logits, ending_logits, n_best_size = 20, max_answer_length = 30):\r\n all_start_logits, all_end_logits = starting_logits, ending_logits\r\n # Build a map example to its corresponding features.\r\n example_id_to_index = {k: i for i, k in enumerate(examples[\"id\"])}\r\n features_per_example = collections.defaultdict(list)\r\n\r\n for i, feature in enumerate(features):\r\n features_per_example[example_id_to_index[feature[\"example_id\"]]].append(i)\r\n\r\n # The dictionaries we have to fill.\r\n predictions = collections.OrderedDict()\r\n\r\n # Logging.\r\n print(f\"Post-processing {len(examples)} example predictions split into {len(features)} features.\")\r\n\r\n # Let's loop over all the examples!\r\n for example_index, example in enumerate(tqdm(examples)):\r\n # Those are the indices of the features associated to the current example.\r\n feature_indices = features_per_example[example_index]\r\n\r\n min_null_score = None # Only used if squad_v2 is True.\r\n valid_answers = []\r\n\r\n context = example[\"context\"]\r\n # Looping through all the features associated to the current example.\r\n for feature_index in feature_indices:\r\n\r\n # We grab the predictions of the model for this feature.\r\n start_logits = all_start_logits[feature_index]\r\n end_logits = all_end_logits[feature_index]\r\n # This is what will allow us to map some the positions in our logits to span of texts in the original\r\n # context.\r\n offset_mapping = features[feature_index][\"offset_mapping\"]\r\n\r\n # Update minimum null prediction.\r\n cls_index = features[feature_index][\"input_ids\"].index(tokenizer.cls_token_id)\r\n feature_null_score = start_logits[cls_index] + end_logits[cls_index]\r\n if min_null_score is None or min_null_score < feature_null_score:\r\n min_null_score = feature_null_score\r\n\r\n # Go through all possibilities for the `n_best_size` greater start and end logits.\r\n start_indexes = np.argsort(start_logits)[-1 : -n_best_size - 1 : -1].tolist()\r\n end_indexes = np.argsort(end_logits)[-1 : -n_best_size - 1 : -1].tolist()\r\n for start_index in start_indexes:\r\n for end_index in end_indexes:\r\n # Don't consider out-of-scope answers, either because the indices are out of bounds or correspond\r\n # to part of the input_ids that are not in the context.\r\n if (\r\n start_index >= len(offset_mapping)\r\n or end_index >= len(offset_mapping)\r\n or offset_mapping[start_index] is None\r\n or offset_mapping[end_index] is None\r\n ):\r\n continue\r\n # Don't consider answers with a length that is either < 0 or > max_answer_length.\r\n if end_index < start_index or end_index - start_index + 1 > max_answer_length:\r\n continue\r\n\r\n start_char = offset_mapping[start_index][0]\r\n end_char = offset_mapping[end_index][1]\r\n valid_answers.append(\r\n {\r\n \"score\": start_logits[start_index] + end_logits[end_index],\r\n \"text\": context[start_char: end_char]\r\n }\r\n )\r\n\r\n\r\n if len(valid_answers) > 0:\r\n best_answer = sorted(valid_answers, key=lambda x: x[\"score\"], reverse=True)[0]\r\n else:\r\n # In the very rare edge case we have not a single non-null prediction, we create a fake prediction to avoid\r\n # failure.\r\n best_answer = {\"text\": \"\", \"score\": 0.0}\r\n\r\n # Let's pick our final answer: the best one or the null answer (only for squad_v2)\r\n if not squad_v2:\r\n predictions[example[\"id\"]] = best_answer[\"text\"]\r\n else:\r\n answer = best_answer[\"text\"] if best_answer[\"score\"] > min_null_score else \"\"\r\n predictions[example[\"id\"]] = answer\r\n\r\n return predictions\r\n\r\n\r\n\r\n# build base examples, features from training data\r\nexamples = load_dataset(\"squad_v2\").shuffle(seed=5)['train']\r\nfeatures = load_dataset(\"squad_v2\").shuffle(seed=5)['train'].map(\r\n prepare_validation_features,\r\n batched=True,\r\n remove_columns=['answers', 'context', 'id', 'question', 'title'])\r\n\r\n# sim some shuffled training indices that we want to use to re-order the data to compare how we did\r\nshuffle_idx = np.arange(0, 131754)\r\nnp.random.shuffle(shuffle_idx)\r\n# create a new dataset with rows selected following the training shuffle\r\nfeatures = features.select(indices=shuffle_idx)\r\n# get unique example ids to match with the \"example\" data\r\nid_list = list(dict.fromkeys(features['example_id']))\r\n# now search for their index positions; load elastic search\r\nes = Elasticsearch([{'host': 'localhost'}]).ping()\r\n# add an index to the id column for the examples\r\nexamples.add_elasticsearch_index(column='id')\r\n# search the examples for their index position\r\nexample_idx = [examples.search(index_name='id', query=i, k=1).indices for i in id_list]\r\n# drop the elastic search\r\nexamples.drop_index(index_name='id')\r\n# put examples in the right order\r\nexamples = examples.select(indices=example_idx)\r\n\r\n# generate some fake data\r\nlogits = {'starting_logits': torch.randn(131754, 384), 'ending_logits': torch.randn(131754, 384)}\r\n\r\n\r\ndef score_squad(logits, n_best_size, max_answer):\r\n # proceed with QA calculation\r\n final_predictions = postprocess_qa_predictions(examples=examples,\r\n features=features,\r\n starting_logits=logits['starting_logits'],\r\n ending_logits=logits['ending_logits'],\r\n n_best_size=20,\r\n max_answer_length=30)\r\n metric = load_metric(\"squad_v2\")\r\n formatted_predictions = [{\"id\": k, \"prediction_text\": v, \"no_answer_probability\": 0.0} for k, v in final_predictions.items()]\r\n references = [{\"id\": ex[\"id\"], \"answers\": ex[\"answers\"]} for ex in examples]\r\n metrics = metric.compute(predictions=formatted_predictions, references=references)\r\n return metrics\r\n\r\nmetrics = score_squad(logits, n_best_size=20, max_answer=30)\r\n```\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\nThanks for the input! I gave that a try by adding this after my selection \/ reordering operations, but before the big computation task of `score_squad`\r\n\r\n```\r\nexamples = examples.flatten_indices()\r\nfeatures = features.flatten_indices()\r\n```\r\n\r\nThat helped quite a bit!","embeddings":[-0.2229693085,0.2477027923,-0.0138683552,-0.1000716835,-0.003328807,-0.1465102434,-0.0916016847,0.1460768282,-0.3497589529,0.1094219536,-0.2728040516,0.3873890042,0.2126539499,-0.1992182434,-0.0613091849,-0.0526460782,-0.1135229096,0.0000033445,-0.0314249322,-0.0937863365,-0.1005084962,-0.3484925032,-0.3477401137,-0.145543471,-0.2218254358,-0.1603693813,0.0549748018,0.3512459695,-0.1260461956,-0.3404070437,0.0977792069,0.5284638405,-0.0601430126,0.3460890353,-0.0001203739,-0.1140628681,-0.301356703,-0.0590985864,-0.4936899543,0.0832120925,-0.0697157905,-0.0438395105,-0.1616155952,0.0779754221,-0.1216855645,-0.1129818335,-0.0549107939,-0.3639614284,0.4560742378,-0.0008517363,0.1208517849,0.220481649,-0.5714429021,-0.2556709349,0.1670394242,0.0162417125,-0.0953832343,-0.0365809016,0.4183206856,0.1212784946,-0.1140871421,0.3431174755,-0.0950435996,-0.1506789178,0.2451233715,0.1294504702,-0.2699114084,0.1256278604,0.005174641,0.0222797319,0.2026703358,-0.1132044941,-0.5826923847,-0.3719269037,0.341755867,-0.0127385156,0.2049164623,-0.2810357511,0.4607771039,-0.0416236669,-0.306301266,0.2016853541,0.0358903036,0.0214370992,0.1613190323,0.2261903286,0.0236478597,-0.0106987497,-0.0192314014,-0.1237981394,0.1931634396,-0.187747255,0.0458304808,0.3385604024,-0.485106051,-0.2102170885,-0.1109896228,-0.2480468452,0.030924825,0.1953477114,0.0991410986,0.4193668067,0.0793582574,-0.3105548918,0.1926971674,0.2943916023,-0.0932467952,0.3354053497,0.0661656633,-0.0558295101,-0.0670939237,-0.027262589,0.05751976,-0.3548845947,-0.1153057069,0.0621315576,-0.416890353,-0.1396674961,-0.4675750732,0.0101275723,-0.2367890626,0.026424244,0.3597140312,0.3569377363,-0.1339115798,-0.1230955496,0.0204742644,0.1510910392,-0.3827776909,0.1194274426,-0.1549317241,-0.1807512939,-0.1282603741,0.4981126487,0.0682006031,0.0105687119,-0.0445278138,0.1479117125,0.6504570246,0.1346221715,0.1789738238,-0.0625602156,0.3049418628,0.1330021918,-0.1644210666,0.0284394585,-0.0693844259,-0.0828607082,-0.2640684247,0.2404960692,-0.4224526882,-0.1730779409,0.2738597393,0.1426498741,0.053241685,-0.1324371696,-0.0261703134,0.1152576506,0.0700723827,-0.3271812797,0.0334452242,-0.2640089989,0.0580386445,0.0417545848,0.1055177599,0.264410466,-0.4662193656,-0.0805213153,0.0947195888,0.5643754601,0.2235662043,0.3344970942,-0.2951166928,0.0774924085,0.0623126552,0.2190618217,0.4485758841,-0.1682099849,-0.4064477086,0.0793168619,-0.3333152533,0.1301809698,0.1100042984,0.2987503409,0.4087423086,-0.1437912285,-0.0917459428,0.2847900689,0.0783123896,-0.0005592417,-0.2430839092,-0.1494386792,-0.1460276693,0.2493930757,-0.0854728371,0.0004539371,0.0658750907,0.2544437647,0.0936773643,-0.212437734,-0.1569136828,0.1440961361,-0.0487086475,0.2361472845,0.009082377,-0.2017503828,-0.2624258995,0.2542325854,-0.1482537538,-0.0874456838,0.1352851391,-0.2490602881,-0.2674162686,-0.2243238986,-0.3613672256,-0.0841277689,-0.0326935686,-0.0000497185,0.2648528516,-0.374301374,-0.1358233392,0.2545201778,-0.29800722,0.1125548556,0.1867353618,0.3353077769,0.0942640156,-0.0836718678,-0.1753809452,0.223179847,0.0413799025,-0.0102596274,-0.0425318852,0.4816074669,-0.0305584166,-0.2010594904,0.300344646,0.1919036806,0.1189256087,-0.0048329839,0.1848534644,-0.0567650683,-0.1038584858,-0.0857224688,-0.4112776518,0.7307934761,-0.6390340924,0.3906633854,-0.0081895739,-0.3914916515,-0.1364453137,0.2339695692,-0.0671845451,0.028089473,0.1946619898,0.0692243576,0.3068402112,0.0219235756,-0.1708946526,0.1148176268,0.3807358742,0.1116098538,-0.116190806,0.0593309514,0.1933869421,0.0581158064,-0.1526626647,-0.4381038547,0.2674148977,0.2144753784,-0.0098131234,-0.0738034323,-0.0876859352,-0.0899261758,-0.0617198683,-0.0450996533,-0.0257454459,0.3969434798,0.3169572949,0.2558146119,-0.2715595961,-0.0461911708,0.0497649349,-0.0269536227,-0.3187946081,0.3653787971,-0.2655484378,0.3262257278,-0.0995716602,0.0995328948,-0.2339483351,-0.1894762069,0.2998906076,-0.0532458164,0.0420344025,0.2280463427,0.0082486607,0.2510022223,0.193622604,-0.4015026987,0.0486457944,0.1283143908,-0.0853393897,-0.0386849381,0.1792995185,0.2340388447,0.299641192,0.1631267071,-0.2064593881,-0.0706808493,-0.3488852978,-0.0585122183,0.0901514664,0.1498829275,-0.0947868451,-0.1986880302,-0.505982399,-0.336107105,-0.0967887044,0.1074076146,-0.189230904,0.0047539468,-0.2101000994,0.0690367818,0.1734790504,-0.4781151414,0.0168381706,0.0162553135,0.0344494693,0.0098374942,0.066954568,0.1476120055,-0.1610152125,-0.1116630211,-0.1796435267,0.0880680829,-0.3538632095,-0.2686766982,0.2319982201,-0.1055770144,-0.008286356,-0.4196908772,-0.0765763521,-0.5765689611,0.2538782358,0.0160375256,-0.1266462505,-0.0915807411,-0.0984586105,-0.3653621674,0.2959348261,0.3322464824,-0.2109398097,0.032138925,0.110281691,-0.0774022862,0.0623431467,0.5486826301,0.0462288484,-0.4003959298,0.1239522099,0.1109858006,1.1333417892,-0.0141810263,0.1114459336,-0.0653473362,0.1698570549,0.0690061226,-0.349622041,-0.1671998799,0.2396910191,-0.006079861,0.0631310791,0.3585751355,0.0024383776,0.2197257131,0.3613546193,0.4713778198,-0.1210378334,-0.1453615576,0.1049261838,0.1241666153,0.1314238906,0.2344905585,0.330524236,-0.254545182,-0.1899589747,-0.0217920169,-0.2138453871,0.2415392846,-0.1783572584,-0.4820561707,0.0176173411,-0.0887342021,0.1038669348,0.1211590916,0.0310366135,0.3065715432,-0.0894950405,0.4382134676,-0.1552442014,0.6033061147,0.010327599,-0.0748458281,0.2094712108,-0.3021306992,0.031763982,-0.0028447087,-0.3103168607,0.1381147504,0.0280377362,0.2992950976,-0.3955968618,-0.146859467,0.2374914289,0.0136861531,0.0182699207,0.0168006681,-0.157424897,0.045749519,0.0769337416,0.4116093814,-0.276858598,0.1689107865,0.0499895476,0.1534054875,-0.2058267593,0.0772600695,0.108990714,0.0978837684,0.232063368,0.3968159556,0.2968211472,0.5759822726,-0.0820016488,0.2667078078,0.3654349148,-0.0405631289,-0.0320802666,0.1777066886,0.0213753227,0.1245086268,0.7451142669,0.1889926642,0.1802190989,-0.0091554085,0.3382091224,-0.2709101439,-0.0570228845,0.3755310476,-0.1951373965,-0.1541287303,-0.2498485744,0.3288463056,0.0089340182,-0.0365891941,-0.1135575995,0.5571148992,-0.1169481799,0.369922787,0.6643725634,1.153963685,-0.3158586621,0.2839170396,0.0857243538,-0.0627596825,0.2783313394,-0.4571999311,0.3110564649,-0.3608015478,-0.5183977485,0.1271019131,-0.2768598795,0.1780613959,0.3513542414,0.0487249717,0.2352935076,-0.3827722371,0.0779589638,-0.3664308488,0.2086066455,0.1052867845,-0.5642211437,-0.0995811895,0.0342080556,0.0851489305,-0.161505565,-0.1026572734,0.1778127551,-0.0776009262,0.2319298089,-0.0566128828,-0.1408687979,-0.3118817806,-0.0234600436,-0.049573943,-0.3439301848,0.6047793627,-0.0415184759,0.1774502844,0.1372202486,0.2033912241,0.1746311635,0.170843333,0.4651058614,0.4565700293,-0.1060808674,0.0508966371,0.1657200456,0.1711342186,-0.1148238555,0.1943956017,-0.431830138,-0.3854596615,0.1150658503,0.0980385244,-0.1946828216,-0.0203516781,0.2804602087,0.1253672391,-0.2001907676,0.0503734462,0.0382371284,-0.0283566099,0.5486469865,0.1635395885,-0.2518385649,-0.2608955801,0.6345748305,0.0965518653,-0.0834737271,0.4297213554,-0.0487744696,-0.3911808431,-0.1019709036,0.2128725201,0.2624303102,-0.3412107825,-0.0872746333,-0.1879292428,-0.279442668,0.1587748379,0.2018291652,0.1933795959,-0.0701102465,-0.2999260128,0.1244067475,0.3185898662,0.0201228522,0.6046158075,0.0398607478,-0.3336846828,0.2227038443,-0.5289216042,0.3563667834,-0.2616191208,-0.1069983616,-0.248258993,0.2899232805,-0.5045489073,-0.1483784765,0.2586853504,-0.1163726002,0.0048723738,-0.0601766892,-0.1983085126,-0.1448252201,-0.1970014423,0.0980262831,-0.1997065097,0.1243187636,-0.0188031849,-0.2507639229,-0.2704802155,-0.4096262753,0.2014555335,0.3396630883,0.2517517805,-0.079448767,0.2015297562,-0.2823911607,0.1224734187,-0.0024231735,0.1681148112,-0.1667636335,0.0411642306,0.245342955,-0.1037383303,-0.1357138753,-0.0697518066,0.2789188623,0.5007079244,0.0211990867,0.2153330296,-0.2984635234,-0.0545401052,0.4007716179,0.4186480641,0.6963021159,0.2969825566,-0.6313238144,0.0892218128,0.0435882285,-0.0628520176,-0.3881573379,0.021162672,-0.1007334739,0.0100013483,0.1811839938,-0.186611712,0.3341404498,0.1153280437,0.1421733499,-0.0810378119,-0.0633160174,0.1096307114,0.221096471,0.215869844,-0.1135800108,0.3845514953,-0.1954146177,0.1096801981,0.0798891932,0.1793094426,0.4664360583,-0.0411300138,0.3082848787,-0.3080215752,-0.2969990373,-0.2421714216,0.4442578256,0.1990970224,-0.0469324812,0.0693915188,0.1385658681,-0.0666203275,-0.5457808375,-0.0916194022,0.1064575464,-0.084910579,-0.119918935,-0.0566291884,-0.1938597113,-0.165054068,-0.1094780266,-0.3108012974,0.129476279,0.039264407,-0.1247126535,0.4548836946,-0.3548500538,-0.07787738,-0.2523677051,0.3607309163,-0.1750518978,-0.0198753662,-0.0690099299,0.3045783639,0.3067370653,0.0552976057,-0.1740990728,0.1050492302,-0.3952314556,-0.0766809285,-0.0727442056,0.0688731596,0.2560726404,-0.0455823876,-0.2837682664,-0.3933627307,0.4206233621,-0.018373549,-0.1651886702,-0.6061081886,0.2266858965,0.0555941425,-0.2718512416,0.0316607803,0.2860613763,0.0470649041,0.4065903723,-0.0323086232,0.061348442,-0.000361293,-0.2930675149,-0.1541664749,0.0500857383,0.2389614582,0.0626749992,0.1489731073,0.5292367339,0.1847597957,-0.0745077059,-0.3575538099,0.0056569469,-0.515768826,-0.0746005774,-0.0146510173,0.0519744009,0.4381356239,0.4754748344,0.1372810155,0.0549258851,0.1292781681,-0.59433496,0.0565181337,0.08937224,-0.1599140167,-0.0063516195,-0.2795810997,-0.092877619,-0.2524853945,-0.2728768587,0.1100642011,0.3300592899,-0.0536877029,0.1298539042,0.0280912239,0.0561371706,0.2114116848,0.045230113,-0.0933138877,-0.0210877396,0.0484026857,0.1092743725,0.1135340929,-0.2138202935,-0.2178073376,0.0998906493,-0.2642327845,0.3201899827,-0.1137272641,-0.1714500338,-0.1460872442,0.1175538823,-0.0223562326,0.1953673959,-0.5066771507,0.1262384802,0.1293935478,0.517236352,-0.2342355847,0.4483053386,0.0020684863,0.4348350167,-0.2447377443,-0.0530665629,0.4661418796,0.0571614131,-0.1567380875,-0.0178665519,0.3398124576,0.2781736553,0.0795370042,0.1954004467,-0.1007859632,0.3035782278,-0.0931361243,-0.0184627101,0.1331180185,-0.0543180071,-0.1423792094,-0.2693759203,-0.0274897944,-0.0685155839,-0.1467158496,-0.2096290737,-0.269469887]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1747","title":"datasets slicing with seed ","comments":"Hi :) \r\nThe slicing API from https:\/\/huggingface.co\/docs\/datasets\/splits.html doesn't shuffle the data.\r\nYou can shuffle and then take a subset of your dataset with\r\n```python\r\n# shuffle and take the first 100 examples\r\ndataset = dataset.shuffle(seed=42).select(range(100))\r\n```\r\n\r\nYou can find more information about shuffling and selecting rows in the documentation: https:\/\/huggingface.co\/docs\/datasets\/processing.html#selecting-sorting-shuffling-splitting-rows","body":"Hi\r\nI need to slice a dataset with random seed, I looked into documentation here https:\/\/huggingface.co\/docs\/datasets\/splits.html \r\nI could not find a seed option, could you assist me please how I can get a slice for different seeds?\r\nthank you.\r\n@lhoestq ","comment_length":50,"text":"datasets slicing with seed \nHi\r\nI need to slice a dataset with random seed, I looked into documentation here https:\/\/huggingface.co\/docs\/datasets\/splits.html \r\nI could not find a seed option, could you assist me please how I can get a slice for different seeds?\r\nthank you.\r\n@lhoestq \nHi :) \r\nThe slicing API from https:\/\/huggingface.co\/docs\/datasets\/splits.html doesn't shuffle the data.\r\nYou can shuffle and then take a subset of your dataset with\r\n```python\r\n# shuffle and take the first 100 examples\r\ndataset = dataset.shuffle(seed=42).select(range(100))\r\n```\r\n\r\nYou can find more information about shuffling and selecting rows in the documentation: https:\/\/huggingface.co\/docs\/datasets\/processing.html#selecting-sorting-shuffling-splitting-rows","embeddings":[0.0095471777,-0.6955253482,-0.1016979888,0.1060836539,0.2887978554,-0.0554467142,0.0833864436,0.0634743124,-0.1392816752,0.6284068227,-0.0467900895,0.0993473008,-0.1830736846,0.7351151109,0.1561356187,-0.3828530908,0.1237090528,-0.2180424631,0.3468203247,-0.1670893282,-0.1487395614,-0.0430829674,-0.2222317755,-0.1483155638,-0.2410769165,-0.2185825855,-0.1768480539,0.1884933859,-0.1170748547,-0.0675891563,-0.0007688868,0.0190950315,0.2182830274,0.2653706074,-0.0001224413,-0.1700753421,0.0218278132,0.0203418434,0.0527785569,-0.0495617092,-0.263219744,0.2918643057,-0.1174263209,-0.1212330386,-0.4069699645,-0.0068350136,0.0658833534,-0.1543095261,0.0397260301,-0.1973049641,0.0544659793,0.0421159044,-0.0192037076,-0.1066148803,0.3257082701,0.6462308764,-0.0514416285,-0.179392159,0.0944860727,0.4545609355,0.3125537038,-0.0050214338,0.1066776589,0.1581705064,0.0530147627,0.0890873894,-0.3917905986,-0.4616832733,0.0310127381,0.4272669256,0.4352067411,0.0732965246,-0.3272964358,-0.2401065528,-0.013761837,-0.3003647625,-0.2683162391,0.5809765458,-0.1436819434,-0.0845444798,-0.043180503,-0.0514528193,-0.1303498,0.1159753129,0.2786247432,0.6345117688,-0.124746412,0.1643834412,0.4493525624,-0.0622664839,-0.0786297396,-0.000780512,-0.0412161089,0.2373900414,0.0242677052,-0.2484133542,-0.4149946868,0.7269174457,0.464211762,0.1487490088,-0.1918441206,0.0850190818,-0.0767042637,-0.0869185925,0.339769125,-0.3440186083,0.1506422758,0.3478377163,0.0303422026,-0.0713103116,0.1579862535,0.0249996018,0.1558550596,0.1213516891,-0.2965348363,-0.2008727342,-0.0705739856,0.0100824274,0.019276714,-0.5418998599,-0.1341821849,-0.3639208972,-0.0333030857,0.2046196312,-0.0193076245,-0.3753099442,-0.2521540821,0.0529920906,-0.1427578181,-0.3852556348,-0.2759348452,0.107563898,0.0617525131,0.0967912748,-0.124631159,-0.2775677741,0.182314992,0.1278393865,0.1188352108,-0.0221498236,-0.3988849819,0.2535947561,-0.1487498581,0.1161133423,0.145652011,-0.1353644133,0.1568274647,0.1440234035,-0.1639458984,-0.027623279,-0.1059765741,-0.1007242799,0.1096103266,-0.0228938423,0.1034320071,0.0206286386,-0.5214861631,0.2818064988,-0.1603294015,0.0818920955,-0.0345447585,0.246831134,-0.3699512482,-0.2161037922,0.3508655429,0.0122972112,-0.1378553808,-0.3832705617,0.0940629765,0.0147692179,0.3416873515,0.3521430194,-0.1484612674,-0.1812331825,-0.1329239309,0.2817608714,0.0835436583,0.3524158597,-0.1219630316,0.1441714466,-0.0311125945,-0.113511391,0.1805301309,0.1857650578,0.1282109916,0.0182743352,-0.0982355401,0.1558003575,-0.4006550014,0.1039377451,-0.1521548629,-0.2310302854,0.1609902233,0.2237419188,0.0479384176,-0.0844174102,0.2544508874,-0.2822081447,0.4949577749,0.0642020553,0.1098062694,-0.0205259956,0.1160778031,0.3239250183,0.0567650534,-0.3433822393,-0.1044875756,0.0508574657,0.0214895979,-0.0372122079,0.1731483936,-0.2949849665,-0.0790552348,-0.0670879632,-0.0073476825,-0.0415524766,-0.115839906,-0.0344084203,0.5063342452,0.117753841,-0.183581695,-0.048893556,-0.1026113555,0.1379621923,-0.742970407,0.396138519,0.0598833114,-0.2251468003,0.0123458039,-0.0442432314,0.0640271679,-0.1030629054,0.0469687246,0.3983740509,0.2964779735,-0.0860215053,0.1261243969,0.1368528903,0.2491817027,-0.2031776011,0.0765056089,-0.2056942582,0.113884151,-0.0848917216,-0.1547457576,0.3523127139,-0.490917474,0.0638669729,-0.1260856092,-0.1007662341,-0.0347461365,0.1185103804,-0.154418841,-0.0178789776,0.1284217536,-0.2224131078,0.0787673891,-0.0738221705,-0.5414638519,0.1156985983,0.0632875934,-0.373178184,0.0503697358,0.090750128,0.0763203651,0.265542239,0.0472899266,0.1305420846,0.144749403,0.335288763,0.1010301337,-0.1944216043,0.3914379776,-0.1287425011,0.054144796,0.1907889992,-0.0312002115,-0.1542370468,-0.2791148126,-0.1743432134,-0.0398011729,0.2993864119,-0.1720972657,0.1142061576,-0.1102464646,-0.2132905126,-0.3585127592,-0.0358970091,-0.1769143939,-0.5306732059,-0.2908897102,-0.1259763986,0.1057344377,0.3068057597,0.1082262248,0.1551924497,0.0926231742,0.0965646058,-0.1541959792,0.4554950595,-0.224445805,0.0699029118,0.0342397951,0.0505521446,-0.0491161048,0.6051309109,0.2825826108,-0.1961036623,-0.210027054,-0.133538574,0.0447517522,0.028057795,-0.2764391601,0.653911531,0.3125838339,0.211489439,-0.32417202,-0.1267129928,0.342420876,-0.1622316986,0.0195052382,-0.0833903775,0.2167857438,-0.1330220103,0.2517122626,-0.012123323,-0.4640728831,-0.0953936502,0.2303369194,-0.0390129015,-0.1246972233,-0.0943552628,0.0383832194,-0.3634671569,-0.0090797469,-0.3386926055,-0.0542810857,-0.5055660009,0.0504824854,-0.190212518,-0.1662492752,0.0919757709,-0.0106832124,-0.2968222201,0.0837210938,0.0697625428,0.5544331074,-0.0950103998,0.0956510231,-0.1617533118,0.1615988761,0.5859101415,-0.2568829656,0.2266479582,-0.0340305865,-0.071851328,-0.0921675786,-0.3189784884,0.2888515592,0.3687152267,0.3478620052,-0.0704156011,0.7969763875,0.0661664233,0.2158705741,-0.0724332184,-0.0264472477,-0.0002844717,0.0541496687,-0.3065511584,0.3096850216,0.4177287817,-0.332727015,0.3756397963,0.1726528257,0.2435272932,-0.0054203151,-0.3924669623,-0.1362071931,-0.1476922631,-0.0267193075,0.1089382768,-0.0608603209,-0.2926632464,-0.1087171286,-0.0969127938,-0.0877781212,-0.3429720402,0.3228102624,0.6068434119,0.0465488583,-0.5338969827,-0.3214606047,-0.3645895422,0.1100873724,0.370456934,-0.1880289465,-0.0531865358,-0.1404211521,0.4958545566,0.2846363485,0.8806203008,-0.0834171027,-0.1512144059,-0.2613041699,-0.3685552478,-0.1362062544,0.2458343655,-0.1690740436,-0.1273730695,-0.087545611,0.0294915792,-0.5456354022,-0.381265074,0.2368127406,0.0162888411,-0.0401974954,-0.1685617417,-0.0815595239,0.1758322865,-0.0546018034,0.1231881678,0.0644009858,0.0445666425,-0.1355675608,0.0218296517,-0.0120067075,-0.2656212449,0.0618258044,-0.1280231327,0.1829459667,0.1233976409,0.0773131028,0.3261013329,0.0946138501,0.2468687743,0.0424086787,-0.0906762257,-0.06036691,-0.1720471382,0.0105895251,0.4419589937,0.0912356079,0.0083019026,0.079551734,-0.0138765275,0.2005517632,-0.342718333,-0.367657274,0.2233267426,-0.1688141525,-0.3760852218,-0.329367429,0.3520815074,0.0732192919,-0.1938610822,0.5608253479,0.4099510312,-0.056971889,0.10359478,0.0406196676,0.7752846479,-0.010373367,-0.0919960067,0.1127670482,0.2219066471,0.2127144337,0.5584771633,0.2137182653,-0.3080766201,-0.4992981255,-0.2018874735,0.158010602,0.0638273805,0.1913168728,-0.0572685823,0.1850073189,-0.0547650717,-0.1005603746,-0.3885397911,-0.1939564645,-0.2248799652,-0.0032211011,-0.1430080384,-0.0336680003,-0.2274885327,0.2590165436,0.0589292645,0.189016223,-0.0609454811,-0.2277991921,-0.2339081913,-0.075994648,-0.3006744981,-0.1416209787,0.0431919135,-0.1364132017,0.002610483,0.3202061951,0.2756868303,0.0858456567,0.013740221,0.2838686705,-0.0923782215,-0.2239915282,-0.2897014022,0.2096967399,0.0199785326,0.1456448436,-0.4162992537,-0.1392607987,0.0684591383,-0.447493434,-0.8403379917,0.0033898395,0.1823561192,-0.5605269074,-0.1671974957,-0.1617654711,0.2584699988,-0.0225807782,-0.0110908393,0.1377934515,0.0547356717,0.2562351525,0.1947604269,0.0849729255,-0.2678752244,0.2098754346,0.0878481641,0.4105135798,0.1986329854,0.1149100214,-0.164253667,-0.1973213553,0.5678084493,0.1755661219,0.098354429,0.0772537142,0.0783279762,0.3218278587,0.1400384307,0.3459123969,0.0977364108,-0.0591797307,0.1046874151,-0.3396751881,0.0934101567,0.3857947588,0.124416329,0.3020279706,0.1752879769,0.2373279929,-0.2597450316,0.2616390586,-0.1473516673,0.2337994576,0.2653442323,0.0466202572,0.1707775593,0.3170280457,0.3241878748,0.1005756706,0.0166592263,0.0941780955,0.1193482503,0.0067495913,-0.1122626066,0.1422743499,-0.0904710144,-0.0279867519,-0.0473504551,-0.0587896407,-0.187356934,-0.0922980905,0.1419070363,0.3153980076,0.270765245,0.086141482,1.2160904408,0.0939448252,-0.1745872051,0.3296565115,-0.153055802,0.2653256357,-0.1119081005,-0.0509430803,-0.0297368113,-0.1117943376,-0.1885169446,0.0322829336,0.5101653337,0.0154172666,0.1380629689,0.0736907274,-0.015725553,0.0804562792,0.2490064055,0.0058054486,0.2402932495,0.2699508071,0.1849491,0.0672851503,0.0092551326,-0.2049562633,0.0901462063,0.2932851017,0.0805137828,0.5057779551,-0.4518731833,0.2890124321,0.0591696128,-0.0799052715,0.4291380048,-0.181535691,0.2624655962,0.2134267539,-0.1114846766,-0.3146134615,0.0464107245,0.0968530998,0.1788669974,-0.0478864014,-0.3192085624,0.4151837528,-0.4028055668,-0.0699219406,0.1929900944,-0.1915121824,-0.2740603387,0.4229440093,-0.1345815212,0.1036738008,0.0598585755,0.5918381214,-0.46133551,0.0373026282,-0.1354266703,0.4274469316,0.2274970114,-0.078880243,-0.4201159775,-0.2546609044,-0.0546231717,-0.0765407756,0.0722230226,-0.1821109504,0.4113156796,0.0849754214,-0.0048287716,-0.2381505519,-0.0982733965,0.1170517355,0.1821959913,-0.0969475955,0.0293952581,-0.3142323494,-0.0825222582,0.4190383852,-0.0195564032,-0.0495142154,-0.2980828285,0.08160083,-0.1910801977,-0.3140588105,-0.1701983064,0.1534588337,0.2465098798,-0.1668257564,0.1121322438,0.1433655471,0.0645341426,-0.027092183,-0.0833774433,0.13514705,0.4787383974,-0.4653786421,0.2377957404,-0.4024116099,0.0649119765,-0.1906553209,-0.0835268572,-0.2828076482,0.4337541163,-0.7622525692,-0.3180724978,-0.1321804672,0.1753839254,0.017974047,-0.0986813009,0.2209614664,0.1497925073,-0.4015921354,-0.1154054701,-0.3428226113,0.18600142,-0.2979780436,-0.1306388676,-0.0382650569,-0.0205505844,-0.0626406148,0.0933579654,-0.0205482841,-0.0371427648,0.1172766834,-0.0490413904,0.5686763525,0.0801249444,-0.2503062189,-0.3237126768,0.2560372055,0.0651095957,0.0279074777,0.0750872195,0.2733128965,-0.2028953135,-0.1034882367,0.2581923008,-0.0324737094,-0.5140293837,-0.1780578643,0.2874120176,0.0611245371,-0.1079595312,-0.009131914,0.0712559074,-0.1119281948,-0.379004091,0.2771791816,0.0665910169,0.6011492014,-0.0342454016,-0.4451884925,-0.2468308806,0.0783799291,-0.101594992,0.1680883616,0.0300931111,0.2269437462,-0.1531110853,0.0279315487,0.2849901021,0.2729871273,-0.2187121063,0.0370078646,0.3146010935,-0.0239031762,0.4214176238,-0.3195383251,0.0510875136,-0.2341621667,-0.1059206054,0.1615003198,-0.4090160728,0.0091464836,-0.0598208159,0.1622515023,0.1562801301,0.4829617739,0.0694247708,0.3323283195,-0.2731132507,0.1071471125,0.5657522082,0.3125078082,-0.0431901366,0.370120883,-0.1503661722]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1747","title":"datasets slicing with seed ","comments":"thank you so much\n\nOn Mon, Jan 18, 2021 at 3:17 PM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> Hi :)\n> The slicing API doesn't shuffle the data.\n> You can shuffle and then take a subset of your dataset with\n>\n> # shuffle and take the first 100 examplesdataset = dataset.shuffle(seed=42).select(range(100))\n>\n> You can find more information about shuffling and selecting rows in the\n> documentation:\n> https:\/\/huggingface.co\/docs\/datasets\/processing.html#selecting-sorting-shuffling-splitting-rows\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/1747#issuecomment-762278134>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/AM3GZM5D5MDPLJGI4IG3UADS2Q7GPANCNFSM4WHLOZJQ>\n> .\n>\n","body":"Hi\r\nI need to slice a dataset with random seed, I looked into documentation here https:\/\/huggingface.co\/docs\/datasets\/splits.html \r\nI could not find a seed option, could you assist me please how I can get a slice for different seeds?\r\nthank you.\r\n@lhoestq ","comment_length":103,"text":"datasets slicing with seed \nHi\r\nI need to slice a dataset with random seed, I looked into documentation here https:\/\/huggingface.co\/docs\/datasets\/splits.html \r\nI could not find a seed option, could you assist me please how I can get a slice for different seeds?\r\nthank you.\r\n@lhoestq \nthank you so much\n\nOn Mon, Jan 18, 2021 at 3:17 PM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> Hi :)\n> The slicing API doesn't shuffle the data.\n> You can shuffle and then take a subset of your dataset with\n>\n> # shuffle and take the first 100 examplesdataset = dataset.shuffle(seed=42).select(range(100))\n>\n> You can find more information about shuffling and selecting rows in the\n> documentation:\n> https:\/\/huggingface.co\/docs\/datasets\/processing.html#selecting-sorting-shuffling-splitting-rows\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/1747#issuecomment-762278134>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/AM3GZM5D5MDPLJGI4IG3UADS2Q7GPANCNFSM4WHLOZJQ>\n> .\n>\n","embeddings":[-0.0242704786,-0.5800485015,-0.1025025472,0.0901766643,0.3337330222,-0.0313543528,0.0862427354,0.0778434575,-0.1555906981,0.6615405679,-0.0357807092,0.1159344763,-0.1422756314,0.6808801293,0.1332575381,-0.3228478432,0.0889167264,-0.1975939274,0.2540716231,-0.149235338,-0.0809885114,-0.0049858391,-0.1872225553,-0.1387491673,-0.2912373245,-0.2559998631,-0.171070084,0.2226489484,-0.1419475228,-0.009559893,0.0738407299,0.0576182567,0.2052414566,0.1588170379,-0.000118139,-0.139104709,0.0257986262,0.0376942456,0.0179532841,-0.0529340282,-0.2801009417,0.3696419299,-0.1711483747,-0.1028147414,-0.4033963084,0.0392841026,0.1133556515,-0.2076660842,0.0146942846,-0.1915639639,0.098517783,0.0506312661,0.0284842215,-0.1905965805,0.2813759744,0.6558921337,-0.0894705504,-0.1163896844,0.063131474,0.5073189139,0.2269356847,0.0002389428,0.1760697961,0.1714318693,0.0334674008,0.0956163257,-0.3726251721,-0.4360311031,0.0508839004,0.4526301324,0.4963739514,0.1073247269,-0.3182222843,-0.1937035471,-0.0015614852,-0.2984308898,-0.2344824076,0.5080899,-0.1129811406,-0.0207737647,-0.0342491008,0.0239949524,-0.108331494,0.0960868746,0.329854846,0.455275774,-0.1264501214,0.1400949061,0.4577620924,-0.0861060843,-0.0909518972,0.0599522032,-0.1416305751,0.2137380093,-0.0388232544,-0.2371636629,-0.3499301076,0.7325920463,0.5226100087,0.1343268007,-0.1152156293,0.1385985911,-0.0751083791,-0.1247148514,0.3147113919,-0.3212530613,0.1867842674,0.2914988101,-0.0277669262,-0.0767461881,0.1510375887,0.0350168422,0.1655572802,0.1850872934,-0.3834688365,-0.1994805187,-0.085770756,-0.014803418,-0.0018438157,-0.5842655301,-0.1403471678,-0.3088261187,-0.0198216457,0.1910836399,0.0416997895,-0.4040729702,-0.2972889841,-0.0018221317,-0.1055932119,-0.4199671745,-0.3290487826,0.1051816568,0.0158047527,0.111004144,-0.0712986141,-0.3014401495,0.1724160761,0.1104671434,0.1535440087,0.0028018493,-0.4736948013,0.2348037064,-0.1354741454,0.1330885291,0.1170670539,-0.1257555932,0.1066314131,0.1307789534,-0.1033917591,-0.0394009054,-0.1314308643,-0.0652163774,0.1241779104,0.0127374995,0.0695370585,0.0033158404,-0.4873548746,0.311891675,-0.2089270502,0.047704719,-0.0773766264,0.3150740862,-0.3417730033,-0.2146972865,0.3613391519,0.013794262,-0.2427788675,-0.3567782044,0.0290716179,0.0296073928,0.3583237231,0.353829205,-0.1359320581,-0.2312394232,-0.1669854671,0.3993638158,0.00805155,0.3812057674,-0.1486324817,0.1341962516,-0.1041427925,-0.1667921692,0.1973541826,0.1176743805,0.1574713439,-0.0099299653,-0.0779606253,0.050833337,-0.3968096673,0.0489632078,-0.1658023894,-0.2540096045,0.055411458,0.2026689202,0.0343664251,-0.0920500979,0.2964076698,-0.2191883326,0.480355233,0.0400295779,0.1040017679,0.0423704088,0.1516966969,0.2454891801,0.0378708206,-0.2893220782,-0.1509354711,0.073101446,0.0028931748,-0.0156618543,0.2078377008,-0.2937504351,-0.061011225,-0.0298586376,0.0571585149,-0.0456082039,-0.0513974652,0.0000218835,0.4871122539,0.1023581848,-0.1233881861,-0.0375184454,-0.1614747941,0.0974078253,-0.7664036751,0.3914408386,0.0388752148,-0.2232470065,0.0756954402,-0.0344887823,0.0620724857,-0.1102513894,0.0592324026,0.3625947833,0.2247779071,-0.0178070217,0.1429630071,0.1382005662,0.2544497848,-0.1579792053,0.007788376,-0.2979333401,0.1091153696,-0.0585157201,-0.2237074971,0.3307854235,-0.4219641089,0.0152340652,-0.1238243431,-0.1087066308,-0.0145641025,0.0868039727,-0.1273791194,-0.065780662,0.1646885425,-0.2167832404,0.0568321496,-0.0496253073,-0.5082008243,0.2045126259,0.1152578741,-0.3140591383,-0.0271635558,0.0599233545,0.0868315101,0.2323121727,0.0578234829,0.136695683,0.1389480382,0.3809364736,0.1306705177,-0.2710238397,0.395421505,-0.1604272723,0.0631251782,0.1573937833,0.0463623814,-0.1512766331,-0.2365282029,-0.1928135753,-0.1225320101,0.2646406293,-0.0871590152,0.127707094,-0.0990575776,-0.2676284015,-0.4211467803,-0.0128322178,-0.1435071826,-0.4676619768,-0.2551902235,-0.1288243532,0.1196715757,0.3570855558,0.0584838837,0.1540054828,0.1172048226,0.1035313904,-0.192457214,0.4856258035,-0.3060836196,0.095652096,0.0159236994,0.1082333624,0.0052157082,0.5828412771,0.2883780897,-0.1631916016,-0.2167949378,-0.1514893323,-0.0317186564,0.0742449462,-0.2648353279,0.6869398952,0.2674885988,0.174036786,-0.3678111136,-0.1260756701,0.2882689238,-0.1811948866,-0.0253332872,-0.1145413443,0.1908085048,-0.1281627119,0.200338617,-0.0837968886,-0.3759827912,-0.1047151312,0.2560347319,-0.0970777571,-0.1402120143,-0.0218529887,0.0314281546,-0.3185782731,-0.044652123,-0.3788219094,-0.1389826685,-0.5120393634,0.0317750424,-0.2471351922,-0.2191503346,0.0938065648,-0.0099170664,-0.2453134358,0.0794136897,0.0780087784,0.5433710217,-0.0363137834,0.1234697104,-0.1115003079,0.0138579523,0.6331720352,-0.3010635376,0.1640507132,-0.0399133042,-0.1221870705,-0.0517044738,-0.2836710811,0.3046175241,0.3274873197,0.312335223,-0.0451563373,0.768820107,0.032335408,0.2276027501,-0.0585773066,-0.0526451245,0.0181535911,0.0496937372,-0.2660487592,0.3946563601,0.3597624302,-0.364284575,0.3904456496,0.173783049,0.1833560467,0.0529088639,-0.325843811,-0.1548325717,-0.1775919199,0.0076405667,0.1233656257,-0.1138254255,-0.191002205,-0.1076810658,-0.0872894004,-0.0686334819,-0.3474594653,0.2767553329,0.6086893082,0.0353085287,-0.5521413684,-0.2775803506,-0.410900861,0.0925830528,0.3612717688,-0.1448525786,-0.1316083819,-0.0899693891,0.4575630724,0.2928394377,0.9118690491,-0.0480404235,-0.0764623061,-0.2678030133,-0.3367661834,-0.0698805004,0.2615612447,-0.1675386578,-0.1604408175,-0.0977822095,0.008009797,-0.562923491,-0.368833214,0.2234751582,-0.0423948765,-0.0119401747,-0.164633736,-0.0375978276,0.1303614676,-0.0818555132,0.1148414835,0.0997272953,0.1194921657,-0.1351350844,-0.0092963157,0.0160996914,-0.3069129586,0.0547068082,-0.1615728587,0.1988475323,0.1703484207,0.0921558812,0.3174723387,0.0863786787,0.3073706925,0.0350333788,0.0105655072,-0.0355812088,-0.1031763926,-0.0827477202,0.4001293778,0.1128337756,-0.0206638444,0.0576550998,0.0313649625,0.1760915667,-0.3059206307,-0.3526521325,0.2256846726,-0.204052791,-0.3664292991,-0.2355872542,0.3317005932,0.1174637005,-0.169529736,0.5695177317,0.3859548569,-0.0499190129,0.0382433459,0.0924312472,0.7290905714,0.0294325221,-0.1381930262,0.1464158297,0.2581764758,0.2535150945,0.5658347011,0.2663366199,-0.3213190436,-0.560905993,-0.1801814437,0.1945436597,0.0701376423,0.2883497477,-0.0299655795,0.1047845632,0.0862353221,-0.0319912285,-0.3092242479,-0.2146296352,-0.2465344071,-0.0768426806,-0.1547740996,0.0205518305,-0.2051934451,0.1990475655,0.0473394804,0.1271197349,-0.0826354921,-0.2482586652,-0.2205244303,-0.0395421162,-0.3270930946,-0.1997731477,0.0493603088,-0.093818292,-0.0173346698,0.2839570642,0.2538844943,0.1005021632,-0.0461255424,0.2969195545,-0.1870472878,-0.2214701623,-0.3022742271,0.1434091926,0.0238753129,0.0939171091,-0.3886086345,-0.2284817249,0.051929079,-0.44028759,-0.7552902699,-0.0343232937,0.1560227126,-0.4987485111,-0.1597283036,-0.1754286736,0.203316018,-0.0585872792,0.0171044543,0.1049088761,0.0587450899,0.1825892478,0.2684274614,0.1431346983,-0.2712363303,0.2703589201,0.088266924,0.4458450675,0.168955043,0.1485593617,-0.1471713036,-0.2188889086,0.5108020306,0.1886906177,0.0121514341,0.0519048199,0.1168077439,0.2932375073,0.1533007622,0.341440469,0.0736494139,-0.0661104023,0.0901014432,-0.3865050972,0.0707790479,0.413765341,0.0008949341,0.3082986474,0.1829466671,0.2645627558,-0.2461036146,0.2580798864,-0.1961766481,0.2095335722,0.2055710256,0.0104881367,0.1557118148,0.2658082843,0.3691785932,0.0534163155,0.0538712665,0.2029087991,0.0868858844,-0.0415695421,-0.0816562623,0.1096634641,-0.0930949003,0.08008302,-0.014889719,-0.0118027451,-0.2303005159,-0.0752516091,0.189769119,0.280862242,0.2509468794,0.0298431814,1.1711047888,0.1738696545,-0.1717866212,0.3409661651,-0.1159722507,0.2363992929,-0.1439261734,-0.0527301393,-0.1090842038,-0.1064584777,-0.166085884,0.0044810819,0.4357903898,0.0557054095,0.130350545,0.0545395389,-0.0873418376,0.103743732,0.210366115,-0.0572845675,0.2755919397,0.257665962,0.1125284135,0.1159226671,0.0014913054,-0.1673841476,0.0919455886,0.2776227295,0.0923940837,0.4643509686,-0.4044194818,0.2489083558,0.1218207106,-0.0440364331,0.4120545983,-0.1648295075,0.2424529046,0.1935707182,-0.1634542942,-0.2682673037,0.1122432351,0.0482799709,0.1798921525,-0.0163139421,-0.3425213099,0.4168281555,-0.4191645384,0.0038346509,0.2048422843,-0.2509158552,-0.2212838829,0.4851027727,-0.1065087244,0.1006896272,0.0395357497,0.5624908209,-0.4088311195,0.0537931435,-0.1628302336,0.3841965199,0.1737348735,-0.0578384958,-0.4703173339,-0.2601882219,-0.0356058553,-0.0807705298,0.0447900742,-0.1512200385,0.4077131748,0.0227883682,-0.0748030096,-0.2502104044,-0.1570412368,0.1276269406,0.1296439767,-0.0733407065,0.0684465393,-0.3397677541,-0.0903370827,0.4200606942,-0.0090483744,-0.0451190844,-0.3887082636,0.2030234039,-0.2416937798,-0.2977421582,-0.195083946,0.1429152787,0.2987177074,-0.1367261112,0.0560592823,0.1678356081,0.1176630855,-0.0504579097,-0.0840740204,0.1555877477,0.5019218326,-0.4701641202,0.1706538349,-0.3687229753,0.0971892253,-0.1885647029,-0.0868102759,-0.212815091,0.4047724903,-0.7692498565,-0.3894668818,-0.1174425408,0.1612358391,0.0490509868,-0.1257901192,0.2492704988,0.1120211557,-0.4836636484,-0.1140960008,-0.3321504295,0.189454034,-0.2385615259,-0.1081532463,0.0517783128,-0.0050429651,-0.0495233685,0.1552315801,-0.0473092347,-0.0576285906,0.0768438727,-0.0536823869,0.540569365,0.016732296,-0.2378337532,-0.2922285795,0.2026412487,0.0832480118,0.0292578712,0.1403606534,0.3057791293,-0.1653121412,-0.1385951489,0.2969428301,-0.0595102906,-0.5467324257,-0.1643439233,0.3374169469,0.0138294632,-0.1412120014,-0.0477770492,-0.0052816505,-0.1105587631,-0.3598602116,0.2565427423,0.0893586203,0.6455032825,-0.0133886514,-0.4323683679,-0.2270382792,0.0555317551,-0.0681263655,0.2774097025,-0.0389790423,0.3033760488,-0.1828287095,0.0322628543,0.2802876532,0.2584705055,-0.1828203648,0.0236726571,0.314050138,-0.0918687433,0.4399707317,-0.2999430299,0.1169457212,-0.207585752,-0.1133516505,0.1766426116,-0.4431574941,0.0033018661,-0.031394843,0.1449700296,0.1314010769,0.5045564175,0.0795728192,0.2805849612,-0.2539165318,0.1061659604,0.5807652473,0.3482684195,-0.0909174457,0.3747271001,-0.1533652544]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1745","title":"difference between wsc and wsc.fixed for superglue","comments":"From the description given in the dataset script for `wsc.fixed`:\r\n```\r\nThis version fixes issues where the spans are not actually substrings of the text.\r\n```","body":"Hi\r\nI see two versions of wsc in superglue, and I am not sure what is the differences and which one is the original one. could you help to discuss the differences? thanks @lhoestq ","comment_length":26,"text":"difference between wsc and wsc.fixed for superglue\nHi\r\nI see two versions of wsc in superglue, and I am not sure what is the differences and which one is the original one. could you help to discuss the differences? thanks @lhoestq \nFrom the description given in the dataset script for `wsc.fixed`:\r\n```\r\nThis version fixes issues where the spans are not actually substrings of the text.\r\n```","embeddings":[-0.1165046319,-0.4438869059,0.0426136553,-0.3368446231,-0.0602852553,-0.3824757338,0.4176901281,-0.2519408464,-0.1031499952,0.054486338,0.1969245225,-0.146696955,0.1097188964,0.1931587756,0.098360531,-0.2450374216,0.3370181918,-0.0297049787,0.2183911204,-0.0973014235,-0.2882769108,-0.0660792813,-0.4690131247,-0.0129365521,-0.6201684475,-0.0405249335,-0.0461752936,-0.0209756084,-0.202447027,-0.2518570423,0.5749289989,0.339104265,0.3738506138,-0.0397943296,-0.0001158088,-0.199822247,0.5318863392,0.0798769891,0.3641519845,0.4048228562,-0.4408574104,-0.4927159548,-0.0975277573,-0.0858862773,0.1681796461,0.249312371,-0.1676397026,-0.3478982449,0.1804486066,-0.1944323033,0.1312809587,-0.0228099246,0.0137679586,-0.1547960639,-0.0083533507,-0.2238358408,-0.1393196881,0.2698023319,0.3193906546,0.0427220277,0.1924658567,0.4580579102,-0.2261031866,-0.0896212086,0.1591403782,0.0118364505,-0.5829805136,-0.0270515736,0.0613174327,0.4081568122,0.4068083763,0.0251670796,-0.4480766952,-0.0524281301,-0.0332935452,0.1166869998,0.1952300221,0.3218772411,0.3227186799,0.1619338691,-0.4081544876,-0.0240215193,0.1407888979,-0.4149757624,-0.3853112757,0.3058308959,-0.1230528504,0.2891016304,-0.1223356426,0.0770807117,0.2488511652,-0.0006772426,-0.1661952585,-0.0300209261,-0.4951646328,-0.2110798806,-0.217123419,0.2030080557,-0.1595705152,0.2685799599,0.1453615278,0.0540380478,0.1727085859,-0.074678205,-0.0876556784,0.1890463382,0.0804159418,-0.0715229735,0.1665004641,-0.1243544444,-0.2451315224,0.0617137402,0.4321227968,-0.3441905081,-0.2333057225,-0.0550082698,0.3404221833,-0.1166873649,-0.2876285017,0.1818105727,-0.0015121206,-0.4694730639,-0.0372020528,0.2117013484,0.0783709735,0.2736050189,-0.0158583168,0.2528206408,-0.0467464849,-0.2208691835,-0.1682070345,0.0275320709,0.0647709519,0.2110201567,-0.1216972619,-0.4539420903,-0.033943221,0.4554688334,-0.5253096223,-0.139845401,-0.4471267462,-0.1749224514,0.0238407571,-0.1458441168,-0.2342798114,0.164782539,-0.0649746954,0.0967292264,0.082372129,0.0846462697,-0.0599702783,-0.082592912,-0.1808074117,0.0868717879,-0.0978944078,-0.2788570821,0.4834412634,0.2546778023,-0.1537862718,0.0886731967,0.2153271735,-0.0972260833,-0.1082553267,-0.2127085626,0.0479025878,0.1230947673,-0.6595558524,-0.0180035178,0.1797147393,-0.0761959553,0.3189204335,0.0976558775,0.1346985549,-0.2406017184,-0.0522274561,0.2427656651,-0.0070213135,0.0770991668,-0.5214035511,0.2309281379,0.4672337174,0.0627568886,0.0758894533,0.0268515404,0.2087683827,-0.0016146232,-0.3322051764,-0.2706946433,-0.0721659437,-0.2226541042,-0.2867246866,0.0091016293,0.2818875313,-0.1519309282,0.0136288907,0.0160741434,0.0633374453,-0.0027107908,0.2329680026,0.1287259907,0.01502789,-0.1269881874,-0.0684637502,0.1541464031,-0.046953436,-0.3897000253,-0.3581689298,-0.107981436,-0.2921330929,-0.0466267727,0.2770764828,-0.0608710386,-0.0595487058,0.1147930995,0.2504432797,-0.1975950599,-0.014917572,0.1125035286,-0.0262016356,-0.4090225101,-0.4172938466,0.0371238329,-0.2061022967,0.1091822609,0.230171442,0.271671474,-0.0824755356,-0.330368042,0.1588196456,0.3281594813,0.1995610446,0.1326562017,-0.271525979,0.6245141029,-0.3346571624,0.2080749422,-0.198858887,-0.1412716061,0.1792337149,-0.5478566885,0.3255382776,0.1101532727,0.1726595014,-0.0353356414,-0.2805935144,0.2657681108,0.1534332484,-0.0480337329,-0.0009915096,-0.0549673736,-0.0492547303,-0.2937425673,-0.0547298752,-0.0695325956,-0.0577465519,-0.0149122719,0.2772341371,0.0630943179,-0.5504789948,0.1491101682,0.7143086195,-0.2740681171,0.0435123034,0.1654292047,-0.3776516914,-0.2486041486,-0.1394638866,0.280959934,0.3068526089,0.0137192775,0.0569086932,0.0276713036,-0.0188716482,-0.092522718,0.4488794804,-0.1777626872,-0.4682577252,-0.0699948221,-0.1501349658,0.0450520515,0.0252174456,0.3341430426,-0.545355916,-0.2040148079,0.0825195983,-0.448271215,-0.3298107088,-0.1052339152,-0.4454060793,-0.0635501072,-0.2789914906,-0.0894788355,-0.131238386,-0.4173159301,-0.1389562041,0.2662662268,-0.0660323128,0.5702053308,-0.2510118783,0.7024792433,-0.2471657544,0.0324246362,-0.2356300354,0.0417775102,-0.0437815748,-0.158128202,-0.0723070428,0.1260567456,0.3512503207,-0.1640392393,-0.3056465089,0.0262424797,-0.0177334249,0.1230627596,0.3480786681,0.1450388432,0.1119326949,-0.1225036532,0.1085024998,-0.5787023306,0.3448266387,-0.2222723812,0.2646610737,0.0796502307,-0.019484248,-0.2624861896,-0.2568075359,-0.1245551482,-0.117643185,-0.3266426027,-0.1079664975,-0.1618291885,-0.16266267,0.0219603814,0.4068734944,0.1268577129,-0.118953228,-0.3631409109,0.2663448155,0.2002221197,-0.3214403391,0.0324532986,-0.2388673872,-0.2139357924,-0.2120586634,-0.050197307,0.3093805015,-0.3111266792,-0.4103981555,0.3222432137,0.2700643837,0.2628370821,-0.1162113175,-0.0704913288,-0.3537479937,-0.3785714507,-0.2534999549,-0.0969455615,0.214597255,0.3455377221,-0.0914842859,-0.0452797115,0.1181723922,0.2031222731,0.0522397496,0.269436121,0.3256165981,-0.1411495954,0.0911589116,-0.1026595831,0.3240303099,-0.0613539889,0.3414685726,0.334343642,0.2504622638,0.2113295197,0.3287374377,0.0730561167,-0.2365051657,-0.0412861854,-0.2008492947,0.1837061346,0.1138679683,0.1185923293,0.1139983386,0.1917091012,0.0320331156,0.089736864,0.3430525064,0.3720321059,-0.1509701461,-0.5998120308,-0.0014664643,-0.051022213,0.544395268,0.1676471382,0.1514751762,0.1254289597,-0.2022349238,-0.0010726636,0.0742148235,0.2895819843,-0.2946631908,-0.2194706351,-0.2821369469,0.1496335864,-0.1376449913,-0.3736897707,0.0785979405,-0.0116492212,0.3005962372,0.0199010018,0.1665444672,-0.0905204117,0.3712961376,0.0989195406,-0.2228335887,-0.1065160781,0.1200200915,-0.1837931126,0.2333819866,-0.1639436185,-0.3374067843,0.1049899086,-0.1425287426,-0.0196296386,-0.485422194,-0.1621075422,0.1448292583,-0.07154385,0.4643679857,-0.0747644454,-0.1323195547,-0.04175473,0.3151393831,-0.1964721829,0.3328539729,-0.0415208489,0.1277202368,0.0566015504,-0.1474510133,0.376609087,0.0727247968,-0.0516804792,0.0949762687,0.2785257697,0.0665106401,-0.733427465,-0.1142558753,0.0152771324,0.3676266968,-0.4343618751,-0.4931998253,0.4126250744,0.2666194141,-0.1380537152,0.0734548643,0.2209329754,-0.0491462499,0.2081402689,0.0505729616,0.9365344048,0.1980689466,-0.2469336838,-0.1019617245,-0.2741835415,0.3697303534,0.0405952856,0.0432205722,-0.300244391,-0.4480784833,-0.085269399,-0.1426514089,0.1508601457,-0.0020692414,-0.3891645372,0.2941819727,0.220184505,0.1418271363,0.0852306932,0.0325533524,-0.0698265657,0.2176839411,0.3613609672,0.0449696518,-0.0370145403,-0.0552394129,-0.1562980562,0.1887682527,0.247321412,-0.3652153015,-0.5746682882,0.0253224857,-0.0232664421,-0.2587082386,0.3016466498,-0.3251586854,0.3217347562,0.3093059957,0.1426471919,-0.0370717123,0.0989445224,0.2027401775,0.0756096765,0.4827195406,0.1381400079,0.2285456806,0.1422356069,0.0402427875,-0.1778756231,-0.1407414377,-0.1174150333,-0.1111765355,-0.378349334,0.1505167037,0.0274444651,-0.3826553226,0.0916949362,0.0696369782,0.4628220797,-0.3492791653,0.0851812735,-0.1707525551,0.1437137574,0.3847947419,-0.0324678607,-0.1817149818,-0.1388542652,0.0127842603,0.2796833515,0.0065390454,-0.0110842418,0.1339948475,-0.2759569883,-0.1533311754,-0.0854334533,-0.1520001143,-0.2181312889,0.178847909,-0.3036330044,-0.1310917884,0.1493566632,0.2421166301,0.0134803476,0.4598396719,-0.2012213171,-0.0795493871,-0.2525122166,-0.1291654259,0.04688894,0.380740881,-0.1464852393,0.0773127973,0.1458464414,0.4178494513,-0.2300201356,0.1093067825,0.0891270712,-0.0381934196,-0.2304638773,0.1413809359,-0.0614194795,-0.0160041172,0.04631247,0.1518714577,0.1652251482,-0.1143795028,-0.0602277108,0.1549896151,-0.0431237668,-0.4361815751,-0.0579770394,0.3159973919,0.5292282701,-0.0179573391,0.185237959,0.1282753646,0.0576702841,0.4169333875,-0.0482062399,0.0880253166,-0.2638797462,0.0522679985,-0.0435411446,0.7487371564,-0.001794092,0.1936087608,0.2278986871,-0.0316660851,0.1052272916,0.0779993162,0.2765851319,0.3018429577,-0.0159572773,0.0317952558,-0.0017216717,-0.0121679213,0.3045690656,0.0512974076,-0.0264315438,-0.1751222759,-0.0587978214,0.1952408552,-0.1160698086,0.1136824265,0.0659986809,-0.0015763738,-0.0387681089,0.1221804321,0.2216412574,-0.0462209433,-0.2587590814,-0.0778970122,0.8659674525,-0.0654850453,0.4197645783,0.2804835737,0.2366857827,0.0050816271,0.3150863349,-0.1764658391,-0.1408130825,-0.0079362784,0.3237074614,0.1673202962,0.2659126222,0.1955088377,0.1513551921,0.0168421995,-0.0060149124,0.3988684714,-0.4364010394,0.044103168,-0.0471872166,0.0466686375,-0.0800793245,-0.3324149549,0.0883189291,0.1127825975,-0.0003434786,-0.6063635945,-0.2023245841,-0.083075285,0.0823903009,0.1436345726,0.0192310568,0.2930012047,0.3953756392,-0.1128228828,0.0609161928,0.0618284158,0.1219065264,-0.0333817154,0.1890121102,-0.1018297821,0.0895104855,0.4069797099,0.510878861,0.6194599271,0.6993117929,0.304499507,0.1465719938,-0.1528843045,0.1024471447,0.064024739,0.1051127538,0.08578033,0.0151709449,-0.2231172621,-0.0410407297,0.2151520699,0.1484856755,-0.0540320575,0.4727928638,-0.2869014144,0.0217142086,-0.3567307591,0.142577216,-0.0946089178,-0.0245343149,0.157575205,0.1210760176,-0.2551406622,-0.1617352366,0.2509945035,-0.2055130303,0.0300442092,-0.0569228791,0.0575967804,0.1609440744,0.4223007858,0.3717046678,0.3308536708,0.1262884289,-0.3960304856,0.1680608243,-0.2094755918,0.4219665825,0.1094523966,-0.1022148952,0.0109411506,-0.2573257387,-0.4651069343,0.3291637301,0.3728501797,-0.0181742888,0.3111791015,-0.0353110507,0.2168151587,0.4290097654,0.1149638072,-0.1367866695,-0.0433994457,-0.0255732574,-0.1515861899,-0.0207638219,-0.0036236565,0.0147147076,0.2758736908,-0.2911996841,0.5122100711,0.3826971054,0.0351491421,-0.2888607979,0.0521693751,-0.2241282612,-0.1043329984,-0.2918564379,0.1459976137,-0.1513597816,0.2586990297,0.0715013221,0.0758012384,-0.0526814051,-0.0911571532,-0.1767743528,0.5641891956,-0.3511331379,-0.09152814,-0.0921094641,-0.357047081,0.2390084863,0.2044112384,-0.0453023501,0.455142796,-0.1789672226,-0.2361582071,0.0921661109,0.1019842625,-0.0491840579,-0.1732167602,0.0368489996,0.0346574932,0.4055052102,-0.6999351382,-0.3297367692,0.1721246392,0.1971317381,-0.023020396,0.0124482932,0.4854852557,-0.2080432326,0.0844949484,-0.2390807718,0.328764677,-0.0839252025,0.0110598393,-0.1954714954]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1743","title":"Issue while Creating Custom Metric","comments":"Currently it's only possible to define the features for the two columns `references` and `predictions`.\r\nThe data for these columns can then be passed to `metric.add_batch` and `metric.compute`.\r\nInstead of defining more columns `text`, `offset_mapping` and `ground` you must include them in either references and predictions.\r\n\r\nFor example \r\n```python\r\nfeatures = datasets.Features({\r\n 'predictions':datasets.Sequence(datasets.Value(\"int32\")),\r\n \"references\": datasets.Sequence({\r\n \"references_ids\": datasets.Value(\"int32\"),\r\n \"offset_mapping\": datasets.Value(\"int32\"),\r\n 'text': datasets.Value('string'),\r\n \"ground\": datasets.Value(\"int32\")\r\n }),\r\n})\r\n```\r\n\r\nAnother option would be to simply have the two features like \r\n```python\r\nfeatures = datasets.Features({\r\n 'predictions':datasets.Sequence(datasets.Value(\"int32\")),\r\n \"references\": datasets.Sequence(datasets.Value(\"int32\")),\r\n})\r\n```\r\nand keep `offset_mapping`, `text` and `ground` as as parameters for the computation (i.e. kwargs when calling `metric.compute`).\r\n\r\n\r\nWhat is the metric you would like to implement ?\r\n\r\nI'm asking since we consider allowing additional fields as requested in the `Comet` metric (see PR and discussion [here](https:\/\/github.com\/huggingface\/datasets\/pull\/1577)) and I'd like to know if it's something that can be interesting for users.\r\n\r\nWhat do you think ?","body":"Hi Team,\r\n\r\nI am trying to create a custom metric for my training as follows, where f1 is my own metric:\r\n\r\n```python\r\n def _info(self):\r\n # TODO: Specifies the datasets.MetricInfo object\r\n return datasets.MetricInfo(\r\n # This is the description that will appear on the metrics page.\r\n description=_DESCRIPTION,\r\n citation=_CITATION,\r\n inputs_description=_KWARGS_DESCRIPTION,\r\n # This defines the format of each prediction and reference\r\n features = datasets.Features({'predictions':datasets.Sequence(datasets.Value(\"int32\")), \"references\": datasets.Sequence(datasets.Value(\"int32\")),\"offset_mapping\":datasets.Sequence(datasets.Value(\"int32\")),'text':datasets.Sequence(datasets.Value('string')),\"ground\":datasets.Sequence(datasets.Value(\"int32\")),}),\r\n # Homepage of the metric for documentation\r\n homepage=\"http:\/\/metric.homepage\",\r\n # Additional links to the codebase or references\r\n codebase_urls=[\"http:\/\/github.com\/path\/to\/codebase\/of\/new_metric\"],\r\n reference_urls=[\"http:\/\/path.to.reference.url\/new_metric\"]\r\n )\r\n\r\n def _compute(self,predictions,references,text,offset_mapping,spans):\r\n\r\n pred_spans = []\r\n\r\n for i,preds in enumerate(predictions):\r\n current_preds = []\r\n for j,token_preds in enumerate(preds):\r\n if (preds>0.5):\r\n current_preds+=list(range(offset_mapping[i][j][0],offset_mapping[i][j][1]))\r\n pred_spans.append(current_spans)\r\n \r\n return {\r\n \"Token Wise F1\": f1_score(references,predictions,labels=[0,1]),\r\n \"Offset Wise F1\": np.mean([f1(preds,gold) for preds,fold in zip(pred_spans,ground)])\r\n }\r\n\r\n```\r\n\r\nI believe this is not correct. But that's not the issue I am facing right now. I get this error :\r\n```python\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-144-ed7349b50821> in <module>()\r\n----> 1 new_metric.compute(predictions=inputs[\"labels\"],references=inputs[\"labels\"], text=inputs[\"text\"], offset_mapping=inputs[\"offset_mapping\"],ground=inputs[\"ground\"] )\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/features.py in encode_batch(self, batch)\r\n 802 encoded_batch = {}\r\n 803 if set(batch) != set(self):\r\n--> 804 print(batch)\r\n 805 print(self)\r\n 806 raise ValueError(\"Column mismatch between batch {} and features {}\".format(set(batch), set(self)))\r\n\r\nValueError: Column mismatch between batch {'references', 'predictions'} and features {'ground', 'predictions', 'offset_mapping', 'text', 'references'}\r\n```\r\nOn checking the features.py file, I see the call is made from add_batch() in metrics.py which only takes in predictions and references.\r\n\r\nHow do I make my custom metric work? Will it work with a trainer even if I am able to make this metric work?\r\n\r\nThanks,\r\nGunjan","comment_length":151,"text":"Issue while Creating Custom Metric\nHi Team,\r\n\r\nI am trying to create a custom metric for my training as follows, where f1 is my own metric:\r\n\r\n```python\r\n def _info(self):\r\n # TODO: Specifies the datasets.MetricInfo object\r\n return datasets.MetricInfo(\r\n # This is the description that will appear on the metrics page.\r\n description=_DESCRIPTION,\r\n citation=_CITATION,\r\n inputs_description=_KWARGS_DESCRIPTION,\r\n # This defines the format of each prediction and reference\r\n features = datasets.Features({'predictions':datasets.Sequence(datasets.Value(\"int32\")), \"references\": datasets.Sequence(datasets.Value(\"int32\")),\"offset_mapping\":datasets.Sequence(datasets.Value(\"int32\")),'text':datasets.Sequence(datasets.Value('string')),\"ground\":datasets.Sequence(datasets.Value(\"int32\")),}),\r\n # Homepage of the metric for documentation\r\n homepage=\"http:\/\/metric.homepage\",\r\n # Additional links to the codebase or references\r\n codebase_urls=[\"http:\/\/github.com\/path\/to\/codebase\/of\/new_metric\"],\r\n reference_urls=[\"http:\/\/path.to.reference.url\/new_metric\"]\r\n )\r\n\r\n def _compute(self,predictions,references,text,offset_mapping,spans):\r\n\r\n pred_spans = []\r\n\r\n for i,preds in enumerate(predictions):\r\n current_preds = []\r\n for j,token_preds in enumerate(preds):\r\n if (preds>0.5):\r\n current_preds+=list(range(offset_mapping[i][j][0],offset_mapping[i][j][1]))\r\n pred_spans.append(current_spans)\r\n \r\n return {\r\n \"Token Wise F1\": f1_score(references,predictions,labels=[0,1]),\r\n \"Offset Wise F1\": np.mean([f1(preds,gold) for preds,fold in zip(pred_spans,ground)])\r\n }\r\n\r\n```\r\n\r\nI believe this is not correct. But that's not the issue I am facing right now. I get this error :\r\n```python\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-144-ed7349b50821> in <module>()\r\n----> 1 new_metric.compute(predictions=inputs[\"labels\"],references=inputs[\"labels\"], text=inputs[\"text\"], offset_mapping=inputs[\"offset_mapping\"],ground=inputs[\"ground\"] )\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/features.py in encode_batch(self, batch)\r\n 802 encoded_batch = {}\r\n 803 if set(batch) != set(self):\r\n--> 804 print(batch)\r\n 805 print(self)\r\n 806 raise ValueError(\"Column mismatch between batch {} and features {}\".format(set(batch), set(self)))\r\n\r\nValueError: Column mismatch between batch {'references', 'predictions'} and features {'ground', 'predictions', 'offset_mapping', 'text', 'references'}\r\n```\r\nOn checking the features.py file, I see the call is made from add_batch() in metrics.py which only takes in predictions and references.\r\n\r\nHow do I make my custom metric work? Will it work with a trainer even if I am able to make this metric work?\r\n\r\nThanks,\r\nGunjan\nCurrently it's only possible to define the features for the two columns `references` and `predictions`.\r\nThe data for these columns can then be passed to `metric.add_batch` and `metric.compute`.\r\nInstead of defining more columns `text`, `offset_mapping` and `ground` you must include them in either references and predictions.\r\n\r\nFor example \r\n```python\r\nfeatures = datasets.Features({\r\n 'predictions':datasets.Sequence(datasets.Value(\"int32\")),\r\n \"references\": datasets.Sequence({\r\n \"references_ids\": datasets.Value(\"int32\"),\r\n \"offset_mapping\": datasets.Value(\"int32\"),\r\n 'text': datasets.Value('string'),\r\n \"ground\": datasets.Value(\"int32\")\r\n }),\r\n})\r\n```\r\n\r\nAnother option would be to simply have the two features like \r\n```python\r\nfeatures = datasets.Features({\r\n 'predictions':datasets.Sequence(datasets.Value(\"int32\")),\r\n \"references\": datasets.Sequence(datasets.Value(\"int32\")),\r\n})\r\n```\r\nand keep `offset_mapping`, `text` and `ground` as as parameters for the computation (i.e. kwargs when calling `metric.compute`).\r\n\r\n\r\nWhat is the metric you would like to implement ?\r\n\r\nI'm asking since we consider allowing additional fields as requested in the `Comet` metric (see PR and discussion [here](https:\/\/github.com\/huggingface\/datasets\/pull\/1577)) and I'd like to know if it's something that can be interesting for users.\r\n\r\nWhat do you think ?","embeddings":[-0.2388352007,-0.3020202518,-0.1457836926,0.1864611059,0.4193514287,-0.0384185873,0.106930159,0.19828628,0.0518294647,0.2997829318,-0.0557398088,0.2383317798,-0.1781434864,0.1406398267,-0.0143401977,-0.1415147632,-0.3350710273,0.1471369863,0.3345439434,-0.0345635414,-0.3107739687,0.1606325507,-0.069587566,0.1216962636,-0.2355847806,-0.0933226347,0.2366884053,-0.0159126148,-0.4124858677,-0.3041528165,0.0990642533,0.0513419546,-0.149600476,0.5381695032,-0.0000985569,0.0695141777,0.2994537354,-0.0466109328,-0.0341314264,-0.1648347974,0.0899349824,-0.0930585712,-0.0194854271,-0.3661278486,-0.3656154275,-0.2527798414,-0.3794954121,-0.2056539357,0.1416984946,0.5694257021,0.3465053737,0.2177312523,-0.0514513589,-0.4145726264,-0.0320219919,-0.2802481055,-0.2366340607,0.1862058491,-0.3302085102,-0.1422023922,-0.0290144905,0.2373677641,0.0977393314,0.2280803323,0.4603567123,0.0747326687,0.4640995562,-0.0288445782,-0.0639058128,0.0993279815,-0.0348311998,-0.2291294634,-0.138928622,-0.086506106,-0.0435681529,-0.6063225269,0.000866775,0.041297853,-0.1401290447,0.0708582774,-0.213723883,0.2448111624,-0.242516309,0.1692857444,-0.2068609595,0.2759689689,-0.1333458424,-0.0129320482,0.2290190905,0.0137924077,0.088157095,-0.1253055483,0.1343980581,-0.030086752,-0.0864596441,-0.102865167,0.0322972387,-0.2499298751,0.071289286,-0.0608593412,0.1571666747,-0.0135507304,0.0931812674,-0.0254779793,-0.2397663593,0.410728544,-0.1583964229,0.1640750319,0.0718855336,0.2834461629,-0.1001683697,-0.076656796,0.2461405247,-0.2452431917,0.0941842049,0.1683179736,0.0826832131,0.0549368151,-0.4134446383,0.1224458292,0.017110968,0.0622441471,0.363204807,0.1792094707,-0.0475458242,-0.0578017496,-0.0745770186,0.1806987673,-0.3036069572,0.1323247403,-0.3576906025,0.186811924,-0.2822223604,0.2094503194,0.102309145,0.3251958489,0.4162543118,-0.0829675421,0.286404103,-0.4359333515,0.0148107437,-0.1164849699,0.0146783115,-0.037698701,-0.116946809,-0.0906927958,0.2322407961,-0.32481879,-0.0325562023,-0.0127761299,-0.0772035941,-0.2612805963,-0.0464587547,0.3655208945,-0.1033221483,0.0883795992,0.1066239253,0.2783166766,0.1253259778,0.1891436875,0.0639014095,0.0877345353,-0.6182097197,-0.1497565061,0.2072361559,-0.0406126045,-0.0707487762,0.0159722082,0.0374544375,0.0529674031,0.0853644535,0.1257972121,0.0015913615,0.5011470914,-0.1231235564,0.0715176761,0.4473967552,-0.3901565671,-0.2748812735,-0.0695148408,-0.2604046464,-0.182681933,-0.0402026139,0.1134250984,0.2639745474,-0.0010217377,0.3351184428,-0.0332489051,0.0672770068,-0.0819639713,-0.336643368,-0.0550780781,-0.110137932,0.0128997751,0.2926904559,0.1166745052,0.115680851,0.0590457879,-0.0790484622,-0.069792673,-0.0462030694,0.2923014164,0.1704942733,-0.1382119209,-0.1373235732,-0.3439494669,0.0773457363,0.1063652709,-0.148277089,0.1029536799,0.1586437076,-0.1641625613,-0.7732864022,-0.0339807272,-0.1352519244,-0.1356810331,0.3323923349,-0.0662694052,0.0512315929,-0.0233430769,-0.1197001338,-0.0865130946,-0.4374274611,-0.0387742259,0.0409117527,0.1444449127,-0.077667743,-0.1523495615,0.2320803255,0.6334596276,0.3142646849,0.2431090474,-0.1536217779,0.4181961417,0.1928747147,-0.0966607258,0.0709495097,0.2647994161,0.1443881691,-0.4176365137,0.0823048502,0.0835839361,0.0380928293,-0.1041174009,-0.1718855947,0.4496607482,-0.0595270917,-0.0188563298,-0.0873197019,-0.0377625525,0.0356232151,-0.0133336354,-0.2262201905,-0.360584408,0.1040816531,-0.2649138272,-0.074418202,-0.1512782425,0.0140222004,-0.0469461717,0.5437553525,0.0859436989,0.020997908,-0.0018737854,0.0927261636,-0.0672308877,-0.40878883,-0.1390052438,0.3537845314,0.2374161929,0.0789087862,0.3056024313,-0.181652993,-0.2899184227,0.1207427308,0.1682116985,-0.2001344264,0.1640179753,0.1174229234,-0.1201038957,-0.0836910084,0.0853296295,-0.136768803,0.1860298961,-0.3435234129,-0.063168399,-0.1156024113,0.1328385919,-0.0778240561,-0.3552176952,0.0478062369,-0.3057578206,0.0757077262,0.1828735173,0.1641632915,0.2744426429,0.1596274674,0.149750337,0.2517192066,-0.0102645345,0.0825760588,-0.2220394164,0.0131280273,0.2060865313,-0.1667333543,0.0321329646,0.3787149489,0.0630213395,0.2129551619,-0.1641378999,-0.3228057027,0.0167389028,-0.0855586529,0.3682919741,0.2106132954,-0.0868878365,-0.1201217026,0.0202160794,0.4172611237,-0.0720684901,-0.059700042,0.0572300069,-0.1489229798,-0.0393495448,-0.151238665,-0.0986143425,0.1086128727,-0.3211775124,-0.1853629351,0.211649254,-0.0162883531,0.240393132,0.1772322059,0.2700444758,0.3833012283,-0.119669266,-0.4149218202,-0.27283144,0.3139650822,-0.49654001,-0.1606483907,0.0105524594,-0.2040606439,0.3531615138,-0.065951407,-0.4769181907,-0.4896347821,0.0200474542,0.2011553943,-0.067957513,-0.1374583393,0.1257937998,-0.1389677227,-0.2248542309,-0.2828626931,-0.3725372851,0.0969550386,-0.1559907794,0.0767360777,-0.1679982394,0.2963856757,0.0796193704,0.5379875898,0.3091501296,-0.0959010199,0.1537613869,-0.1665041894,0.1702274978,-0.2196134925,-0.2937709689,0.0400277041,0.2761957645,-0.0092067979,0.3007904887,0.0125199882,0.0655873716,-0.0821676329,-0.0163452234,-0.1726568043,-0.0903281942,-0.2720501125,0.0268463567,0.2756066024,-0.0166566204,0.2943208218,-0.307978332,-0.0173058566,0.1736261696,0.1240034923,-0.2654491961,-0.2161292285,-0.2026868165,0.0979779065,-0.1469545513,0.1126972288,-0.0043462636,0.0319835022,0.0253151543,0.278198123,0.0260556694,0.0583922602,0.30260095,-0.2847106755,-0.0721262693,0.1879707724,0.0299128983,-0.2612213492,0.0207335632,0.0607772171,0.2170864344,-0.0766194388,0.7068268657,-0.3140385151,-0.2015741765,0.3677611947,-0.0866843835,-0.115678139,-0.2455518395,-0.2079696506,-0.4916193485,0.0866601914,0.1118829921,0.0345400684,0.4062201977,0.2088234723,-0.1182316616,0.0021238416,-0.2255106717,0.1432759166,-0.0509579405,0.2105502933,0.1123033017,0.0916013271,0.0428008996,-0.1372723281,-0.1102942154,0.0896127,-0.0622721612,-0.2546942532,-0.0094315782,-0.1743390709,0.1378944367,0.1431899667,-0.1240522042,-0.008678495,-0.00855924,0.3078177571,-0.1836660057,0.2394238561,0.2416365147,-0.0930792317,-0.0456482843,-0.3316988051,0.4245659411,0.0737655386,0.0202902872,0.31581375,-0.0291648246,-0.222344175,-0.0226509105,0.1994321793,0.6062893271,0.1603306979,-0.3175835609,0.525252223,-0.146967575,0.0887881666,-0.275362432,0.0864622518,-0.2138996571,-0.1080952808,0.0103716105,-0.0206056125,-0.0015153845,-0.0870497376,-0.0405262671,0.1582216471,0.0064192554,0.1279787868,0.0123804454,0.1309275478,0.2566299736,-0.224310264,-0.2564955354,0.3084740043,-0.1585789621,-0.0529763773,-0.0328532793,-0.0789762884,-0.1387616545,-0.1088169813,0.0663616359,-0.2599680424,-0.3353091478,-0.0554852188,0.0048328764,-0.2521205246,0.0037994706,0.0339758918,0.6332516074,0.1763828248,-0.193376258,0.1461806744,-0.214881584,0.0211608037,-0.1543854326,0.035420917,-0.028258374,-0.1656349301,-0.1136269346,0.0832109228,0.1465190351,-0.1040131673,-0.0523916408,-0.2039106488,0.0592292584,-0.2386748046,0.0388639979,-0.1454927176,-0.0332794972,-0.2638089061,0.2938853502,-0.1383424997,-0.1521502435,0.1905476153,0.162686348,-0.3084410429,-0.0262026787,0.0629230291,0.0559829138,0.0739613026,0.1647737175,-0.1561599374,-0.1534170806,-0.3211123347,-0.0512611084,-0.1514146477,0.0145030292,0.360753864,0.0004530298,-0.0166291855,0.231844306,0.4143939912,0.1999179721,0.2255794704,0.1316360086,-0.1300473511,-0.3161061108,-0.1467150897,-0.0892666727,0.2483056039,-0.2668385804,0.6535883546,-0.1102593988,-0.0258174557,-0.5168286562,0.0210987758,-0.3219112158,0.0016517855,-0.0877803415,-0.2357414067,-0.0257900115,0.2199912518,0.2234495133,0.200025171,-0.4137070477,-0.3525777757,-0.2297541946,0.0474800989,0.0519215129,-0.0544137917,0.0205131005,0.1132489145,-0.4150606692,-0.0894882157,-0.1021690816,0.3201111853,0.1548327506,0.1541000754,-0.2030623406,0.3304716945,0.1944779903,-0.2278048545,0.1355090439,0.3945113122,0.0560095757,-0.0594261251,-0.0535650402,-0.0257706027,-0.0678887293,-0.0250606071,0.117646724,0.0966678858,0.2362035215,-0.0858821496,-0.1094243973,0.3337861896,0.2333496064,-0.0174512081,-0.269690901,-0.1854153275,-0.0912891626,0.3808512986,-0.1617774367,0.11387036,0.1146671325,0.0822917968,0.4596897662,0.3710055649,0.1696257591,-0.0328819528,-0.1256029159,0.0606863275,0.3149350286,-0.25513798,-0.0310555454,0.0309937522,0.0848925859,0.0113152061,0.4787457585,0.1104654744,0.1877098382,0.4650905132,-0.0248465464,0.2866050303,0.113560006,0.0256369915,-0.1688521206,0.2463981807,-0.1097464561,0.1416637003,-0.0437036417,-0.1601411402,-0.2172695994,-0.0553399995,-0.1343214661,0.1474780291,-0.1634102315,-0.0035561712,-0.0300314482,-0.0843847319,-0.4169696271,-0.1988135129,-0.0874230415,-0.1785982251,0.1360294372,0.2287361622,0.0039017063,-0.2996532321,0.0245342739,0.1027036309,0.0521601215,0.2422717959,0.283642292,-0.2555299103,0.2037386745,0.2507945597,0.3229714632,-0.0833271444,0.3195424676,0.1042081118,0.1733374745,-0.3330545723,-0.2613289952,0.0299530402,-0.2656732202,0.0219598822,0.1441774219,0.3015913963,-0.1541232169,0.402978003,0.2882746756,-0.4048208594,0.2984380126,0.1974197179,-0.0133801494,-0.432481885,0.3722925484,-0.0042624511,-0.0333424062,-0.0184285324,-0.0564747192,-0.1463487595,-0.0918448195,0.1887684017,0.0247337054,0.3445622325,-0.119789429,0.1697074324,-0.0821185708,0.3426369429,-0.0398551635,0.0763787478,-0.4161768258,-0.1355724335,-0.4027681947,0.1609249711,-0.0098998137,0.1158459559,0.0341134854,0.2473656088,-0.2458713949,0.2662317157,0.1309258491,0.0437190644,-0.0359608755,0.0830051824,-0.34445104,0.0456631258,-0.2424387187,-0.1139571071,0.2453490049,-0.136760816,-0.1752684265,-0.1287543178,0.2528265417,-0.1321383864,0.1615892947,0.1247489154,0.4439778626,0.2961933315,0.0682060793,0.1773564517,-0.1089350432,0.1017327607,-0.0991781652,0.0317045599,-0.0573241711,0.2549757659,0.1116755232,0.3584685922,-0.0195242874,-0.1864133328,-0.222733885,0.2448714525,0.3298005462,-0.2393908203,-0.0345268995,0.2742847502,-0.1462465376,-0.1039075032,0.0426047146,0.2757709026,0.0453789011,0.4797607362,-0.1038704515,-0.4449886084,0.5055987239,0.0036407665,-0.3741857111,0.2025415748,0.315913856,0.3365118504,-0.1287272274,-0.4725493789,0.1250617206,0.4936556518,-0.1135326996,-0.0947745293,0.2826546729,-0.2521481812,0.0280762855,0.1982813627,-0.1538530141,0.0279016066,-0.0774058402,0.0954035819,-0.0694282576]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1743","title":"Issue while Creating Custom Metric","comments":"Hi @lhoestq,\r\n\r\nI am doing text segmentation and the metric is effectively dice score on character offsets. So I need to pass the actual spans and I want to be able to get the spans based on predictions using offset_mapping.\r\n\r\nIncluding them in references seems like a good idea. I'll try it out and get back to you. If there's a better way to write a metric function for the same, please let me know.","body":"Hi Team,\r\n\r\nI am trying to create a custom metric for my training as follows, where f1 is my own metric:\r\n\r\n```python\r\n def _info(self):\r\n # TODO: Specifies the datasets.MetricInfo object\r\n return datasets.MetricInfo(\r\n # This is the description that will appear on the metrics page.\r\n description=_DESCRIPTION,\r\n citation=_CITATION,\r\n inputs_description=_KWARGS_DESCRIPTION,\r\n # This defines the format of each prediction and reference\r\n features = datasets.Features({'predictions':datasets.Sequence(datasets.Value(\"int32\")), \"references\": datasets.Sequence(datasets.Value(\"int32\")),\"offset_mapping\":datasets.Sequence(datasets.Value(\"int32\")),'text':datasets.Sequence(datasets.Value('string')),\"ground\":datasets.Sequence(datasets.Value(\"int32\")),}),\r\n # Homepage of the metric for documentation\r\n homepage=\"http:\/\/metric.homepage\",\r\n # Additional links to the codebase or references\r\n codebase_urls=[\"http:\/\/github.com\/path\/to\/codebase\/of\/new_metric\"],\r\n reference_urls=[\"http:\/\/path.to.reference.url\/new_metric\"]\r\n )\r\n\r\n def _compute(self,predictions,references,text,offset_mapping,spans):\r\n\r\n pred_spans = []\r\n\r\n for i,preds in enumerate(predictions):\r\n current_preds = []\r\n for j,token_preds in enumerate(preds):\r\n if (preds>0.5):\r\n current_preds+=list(range(offset_mapping[i][j][0],offset_mapping[i][j][1]))\r\n pred_spans.append(current_spans)\r\n \r\n return {\r\n \"Token Wise F1\": f1_score(references,predictions,labels=[0,1]),\r\n \"Offset Wise F1\": np.mean([f1(preds,gold) for preds,fold in zip(pred_spans,ground)])\r\n }\r\n\r\n```\r\n\r\nI believe this is not correct. But that's not the issue I am facing right now. I get this error :\r\n```python\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-144-ed7349b50821> in <module>()\r\n----> 1 new_metric.compute(predictions=inputs[\"labels\"],references=inputs[\"labels\"], text=inputs[\"text\"], offset_mapping=inputs[\"offset_mapping\"],ground=inputs[\"ground\"] )\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/features.py in encode_batch(self, batch)\r\n 802 encoded_batch = {}\r\n 803 if set(batch) != set(self):\r\n--> 804 print(batch)\r\n 805 print(self)\r\n 806 raise ValueError(\"Column mismatch between batch {} and features {}\".format(set(batch), set(self)))\r\n\r\nValueError: Column mismatch between batch {'references', 'predictions'} and features {'ground', 'predictions', 'offset_mapping', 'text', 'references'}\r\n```\r\nOn checking the features.py file, I see the call is made from add_batch() in metrics.py which only takes in predictions and references.\r\n\r\nHow do I make my custom metric work? Will it work with a trainer even if I am able to make this metric work?\r\n\r\nThanks,\r\nGunjan","comment_length":75,"text":"Issue while Creating Custom Metric\nHi Team,\r\n\r\nI am trying to create a custom metric for my training as follows, where f1 is my own metric:\r\n\r\n```python\r\n def _info(self):\r\n # TODO: Specifies the datasets.MetricInfo object\r\n return datasets.MetricInfo(\r\n # This is the description that will appear on the metrics page.\r\n description=_DESCRIPTION,\r\n citation=_CITATION,\r\n inputs_description=_KWARGS_DESCRIPTION,\r\n # This defines the format of each prediction and reference\r\n features = datasets.Features({'predictions':datasets.Sequence(datasets.Value(\"int32\")), \"references\": datasets.Sequence(datasets.Value(\"int32\")),\"offset_mapping\":datasets.Sequence(datasets.Value(\"int32\")),'text':datasets.Sequence(datasets.Value('string')),\"ground\":datasets.Sequence(datasets.Value(\"int32\")),}),\r\n # Homepage of the metric for documentation\r\n homepage=\"http:\/\/metric.homepage\",\r\n # Additional links to the codebase or references\r\n codebase_urls=[\"http:\/\/github.com\/path\/to\/codebase\/of\/new_metric\"],\r\n reference_urls=[\"http:\/\/path.to.reference.url\/new_metric\"]\r\n )\r\n\r\n def _compute(self,predictions,references,text,offset_mapping,spans):\r\n\r\n pred_spans = []\r\n\r\n for i,preds in enumerate(predictions):\r\n current_preds = []\r\n for j,token_preds in enumerate(preds):\r\n if (preds>0.5):\r\n current_preds+=list(range(offset_mapping[i][j][0],offset_mapping[i][j][1]))\r\n pred_spans.append(current_spans)\r\n \r\n return {\r\n \"Token Wise F1\": f1_score(references,predictions,labels=[0,1]),\r\n \"Offset Wise F1\": np.mean([f1(preds,gold) for preds,fold in zip(pred_spans,ground)])\r\n }\r\n\r\n```\r\n\r\nI believe this is not correct. But that's not the issue I am facing right now. I get this error :\r\n```python\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-144-ed7349b50821> in <module>()\r\n----> 1 new_metric.compute(predictions=inputs[\"labels\"],references=inputs[\"labels\"], text=inputs[\"text\"], offset_mapping=inputs[\"offset_mapping\"],ground=inputs[\"ground\"] )\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/features.py in encode_batch(self, batch)\r\n 802 encoded_batch = {}\r\n 803 if set(batch) != set(self):\r\n--> 804 print(batch)\r\n 805 print(self)\r\n 806 raise ValueError(\"Column mismatch between batch {} and features {}\".format(set(batch), set(self)))\r\n\r\nValueError: Column mismatch between batch {'references', 'predictions'} and features {'ground', 'predictions', 'offset_mapping', 'text', 'references'}\r\n```\r\nOn checking the features.py file, I see the call is made from add_batch() in metrics.py which only takes in predictions and references.\r\n\r\nHow do I make my custom metric work? Will it work with a trainer even if I am able to make this metric work?\r\n\r\nThanks,\r\nGunjan\nHi @lhoestq,\r\n\r\nI am doing text segmentation and the metric is effectively dice score on character offsets. So I need to pass the actual spans and I want to be able to get the spans based on predictions using offset_mapping.\r\n\r\nIncluding them in references seems like a good idea. I'll try it out and get back to you. If there's a better way to write a metric function for the same, please let me know.","embeddings":[-0.2388352007,-0.3020202518,-0.1457836926,0.1864611059,0.4193514287,-0.0384185873,0.106930159,0.19828628,0.0518294647,0.2997829318,-0.0557398088,0.2383317798,-0.1781434864,0.1406398267,-0.0143401977,-0.1415147632,-0.3350710273,0.1471369863,0.3345439434,-0.0345635414,-0.3107739687,0.1606325507,-0.069587566,0.1216962636,-0.2355847806,-0.0933226347,0.2366884053,-0.0159126148,-0.4124858677,-0.3041528165,0.0990642533,0.0513419546,-0.149600476,0.5381695032,-0.0000985569,0.0695141777,0.2994537354,-0.0466109328,-0.0341314264,-0.1648347974,0.0899349824,-0.0930585712,-0.0194854271,-0.3661278486,-0.3656154275,-0.2527798414,-0.3794954121,-0.2056539357,0.1416984946,0.5694257021,0.3465053737,0.2177312523,-0.0514513589,-0.4145726264,-0.0320219919,-0.2802481055,-0.2366340607,0.1862058491,-0.3302085102,-0.1422023922,-0.0290144905,0.2373677641,0.0977393314,0.2280803323,0.4603567123,0.0747326687,0.4640995562,-0.0288445782,-0.0639058128,0.0993279815,-0.0348311998,-0.2291294634,-0.138928622,-0.086506106,-0.0435681529,-0.6063225269,0.000866775,0.041297853,-0.1401290447,0.0708582774,-0.213723883,0.2448111624,-0.242516309,0.1692857444,-0.2068609595,0.2759689689,-0.1333458424,-0.0129320482,0.2290190905,0.0137924077,0.088157095,-0.1253055483,0.1343980581,-0.030086752,-0.0864596441,-0.102865167,0.0322972387,-0.2499298751,0.071289286,-0.0608593412,0.1571666747,-0.0135507304,0.0931812674,-0.0254779793,-0.2397663593,0.410728544,-0.1583964229,0.1640750319,0.0718855336,0.2834461629,-0.1001683697,-0.076656796,0.2461405247,-0.2452431917,0.0941842049,0.1683179736,0.0826832131,0.0549368151,-0.4134446383,0.1224458292,0.017110968,0.0622441471,0.363204807,0.1792094707,-0.0475458242,-0.0578017496,-0.0745770186,0.1806987673,-0.3036069572,0.1323247403,-0.3576906025,0.186811924,-0.2822223604,0.2094503194,0.102309145,0.3251958489,0.4162543118,-0.0829675421,0.286404103,-0.4359333515,0.0148107437,-0.1164849699,0.0146783115,-0.037698701,-0.116946809,-0.0906927958,0.2322407961,-0.32481879,-0.0325562023,-0.0127761299,-0.0772035941,-0.2612805963,-0.0464587547,0.3655208945,-0.1033221483,0.0883795992,0.1066239253,0.2783166766,0.1253259778,0.1891436875,0.0639014095,0.0877345353,-0.6182097197,-0.1497565061,0.2072361559,-0.0406126045,-0.0707487762,0.0159722082,0.0374544375,0.0529674031,0.0853644535,0.1257972121,0.0015913615,0.5011470914,-0.1231235564,0.0715176761,0.4473967552,-0.3901565671,-0.2748812735,-0.0695148408,-0.2604046464,-0.182681933,-0.0402026139,0.1134250984,0.2639745474,-0.0010217377,0.3351184428,-0.0332489051,0.0672770068,-0.0819639713,-0.336643368,-0.0550780781,-0.110137932,0.0128997751,0.2926904559,0.1166745052,0.115680851,0.0590457879,-0.0790484622,-0.069792673,-0.0462030694,0.2923014164,0.1704942733,-0.1382119209,-0.1373235732,-0.3439494669,0.0773457363,0.1063652709,-0.148277089,0.1029536799,0.1586437076,-0.1641625613,-0.7732864022,-0.0339807272,-0.1352519244,-0.1356810331,0.3323923349,-0.0662694052,0.0512315929,-0.0233430769,-0.1197001338,-0.0865130946,-0.4374274611,-0.0387742259,0.0409117527,0.1444449127,-0.077667743,-0.1523495615,0.2320803255,0.6334596276,0.3142646849,0.2431090474,-0.1536217779,0.4181961417,0.1928747147,-0.0966607258,0.0709495097,0.2647994161,0.1443881691,-0.4176365137,0.0823048502,0.0835839361,0.0380928293,-0.1041174009,-0.1718855947,0.4496607482,-0.0595270917,-0.0188563298,-0.0873197019,-0.0377625525,0.0356232151,-0.0133336354,-0.2262201905,-0.360584408,0.1040816531,-0.2649138272,-0.074418202,-0.1512782425,0.0140222004,-0.0469461717,0.5437553525,0.0859436989,0.020997908,-0.0018737854,0.0927261636,-0.0672308877,-0.40878883,-0.1390052438,0.3537845314,0.2374161929,0.0789087862,0.3056024313,-0.181652993,-0.2899184227,0.1207427308,0.1682116985,-0.2001344264,0.1640179753,0.1174229234,-0.1201038957,-0.0836910084,0.0853296295,-0.136768803,0.1860298961,-0.3435234129,-0.063168399,-0.1156024113,0.1328385919,-0.0778240561,-0.3552176952,0.0478062369,-0.3057578206,0.0757077262,0.1828735173,0.1641632915,0.2744426429,0.1596274674,0.149750337,0.2517192066,-0.0102645345,0.0825760588,-0.2220394164,0.0131280273,0.2060865313,-0.1667333543,0.0321329646,0.3787149489,0.0630213395,0.2129551619,-0.1641378999,-0.3228057027,0.0167389028,-0.0855586529,0.3682919741,0.2106132954,-0.0868878365,-0.1201217026,0.0202160794,0.4172611237,-0.0720684901,-0.059700042,0.0572300069,-0.1489229798,-0.0393495448,-0.151238665,-0.0986143425,0.1086128727,-0.3211775124,-0.1853629351,0.211649254,-0.0162883531,0.240393132,0.1772322059,0.2700444758,0.3833012283,-0.119669266,-0.4149218202,-0.27283144,0.3139650822,-0.49654001,-0.1606483907,0.0105524594,-0.2040606439,0.3531615138,-0.065951407,-0.4769181907,-0.4896347821,0.0200474542,0.2011553943,-0.067957513,-0.1374583393,0.1257937998,-0.1389677227,-0.2248542309,-0.2828626931,-0.3725372851,0.0969550386,-0.1559907794,0.0767360777,-0.1679982394,0.2963856757,0.0796193704,0.5379875898,0.3091501296,-0.0959010199,0.1537613869,-0.1665041894,0.1702274978,-0.2196134925,-0.2937709689,0.0400277041,0.2761957645,-0.0092067979,0.3007904887,0.0125199882,0.0655873716,-0.0821676329,-0.0163452234,-0.1726568043,-0.0903281942,-0.2720501125,0.0268463567,0.2756066024,-0.0166566204,0.2943208218,-0.307978332,-0.0173058566,0.1736261696,0.1240034923,-0.2654491961,-0.2161292285,-0.2026868165,0.0979779065,-0.1469545513,0.1126972288,-0.0043462636,0.0319835022,0.0253151543,0.278198123,0.0260556694,0.0583922602,0.30260095,-0.2847106755,-0.0721262693,0.1879707724,0.0299128983,-0.2612213492,0.0207335632,0.0607772171,0.2170864344,-0.0766194388,0.7068268657,-0.3140385151,-0.2015741765,0.3677611947,-0.0866843835,-0.115678139,-0.2455518395,-0.2079696506,-0.4916193485,0.0866601914,0.1118829921,0.0345400684,0.4062201977,0.2088234723,-0.1182316616,0.0021238416,-0.2255106717,0.1432759166,-0.0509579405,0.2105502933,0.1123033017,0.0916013271,0.0428008996,-0.1372723281,-0.1102942154,0.0896127,-0.0622721612,-0.2546942532,-0.0094315782,-0.1743390709,0.1378944367,0.1431899667,-0.1240522042,-0.008678495,-0.00855924,0.3078177571,-0.1836660057,0.2394238561,0.2416365147,-0.0930792317,-0.0456482843,-0.3316988051,0.4245659411,0.0737655386,0.0202902872,0.31581375,-0.0291648246,-0.222344175,-0.0226509105,0.1994321793,0.6062893271,0.1603306979,-0.3175835609,0.525252223,-0.146967575,0.0887881666,-0.275362432,0.0864622518,-0.2138996571,-0.1080952808,0.0103716105,-0.0206056125,-0.0015153845,-0.0870497376,-0.0405262671,0.1582216471,0.0064192554,0.1279787868,0.0123804454,0.1309275478,0.2566299736,-0.224310264,-0.2564955354,0.3084740043,-0.1585789621,-0.0529763773,-0.0328532793,-0.0789762884,-0.1387616545,-0.1088169813,0.0663616359,-0.2599680424,-0.3353091478,-0.0554852188,0.0048328764,-0.2521205246,0.0037994706,0.0339758918,0.6332516074,0.1763828248,-0.193376258,0.1461806744,-0.214881584,0.0211608037,-0.1543854326,0.035420917,-0.028258374,-0.1656349301,-0.1136269346,0.0832109228,0.1465190351,-0.1040131673,-0.0523916408,-0.2039106488,0.0592292584,-0.2386748046,0.0388639979,-0.1454927176,-0.0332794972,-0.2638089061,0.2938853502,-0.1383424997,-0.1521502435,0.1905476153,0.162686348,-0.3084410429,-0.0262026787,0.0629230291,0.0559829138,0.0739613026,0.1647737175,-0.1561599374,-0.1534170806,-0.3211123347,-0.0512611084,-0.1514146477,0.0145030292,0.360753864,0.0004530298,-0.0166291855,0.231844306,0.4143939912,0.1999179721,0.2255794704,0.1316360086,-0.1300473511,-0.3161061108,-0.1467150897,-0.0892666727,0.2483056039,-0.2668385804,0.6535883546,-0.1102593988,-0.0258174557,-0.5168286562,0.0210987758,-0.3219112158,0.0016517855,-0.0877803415,-0.2357414067,-0.0257900115,0.2199912518,0.2234495133,0.200025171,-0.4137070477,-0.3525777757,-0.2297541946,0.0474800989,0.0519215129,-0.0544137917,0.0205131005,0.1132489145,-0.4150606692,-0.0894882157,-0.1021690816,0.3201111853,0.1548327506,0.1541000754,-0.2030623406,0.3304716945,0.1944779903,-0.2278048545,0.1355090439,0.3945113122,0.0560095757,-0.0594261251,-0.0535650402,-0.0257706027,-0.0678887293,-0.0250606071,0.117646724,0.0966678858,0.2362035215,-0.0858821496,-0.1094243973,0.3337861896,0.2333496064,-0.0174512081,-0.269690901,-0.1854153275,-0.0912891626,0.3808512986,-0.1617774367,0.11387036,0.1146671325,0.0822917968,0.4596897662,0.3710055649,0.1696257591,-0.0328819528,-0.1256029159,0.0606863275,0.3149350286,-0.25513798,-0.0310555454,0.0309937522,0.0848925859,0.0113152061,0.4787457585,0.1104654744,0.1877098382,0.4650905132,-0.0248465464,0.2866050303,0.113560006,0.0256369915,-0.1688521206,0.2463981807,-0.1097464561,0.1416637003,-0.0437036417,-0.1601411402,-0.2172695994,-0.0553399995,-0.1343214661,0.1474780291,-0.1634102315,-0.0035561712,-0.0300314482,-0.0843847319,-0.4169696271,-0.1988135129,-0.0874230415,-0.1785982251,0.1360294372,0.2287361622,0.0039017063,-0.2996532321,0.0245342739,0.1027036309,0.0521601215,0.2422717959,0.283642292,-0.2555299103,0.2037386745,0.2507945597,0.3229714632,-0.0833271444,0.3195424676,0.1042081118,0.1733374745,-0.3330545723,-0.2613289952,0.0299530402,-0.2656732202,0.0219598822,0.1441774219,0.3015913963,-0.1541232169,0.402978003,0.2882746756,-0.4048208594,0.2984380126,0.1974197179,-0.0133801494,-0.432481885,0.3722925484,-0.0042624511,-0.0333424062,-0.0184285324,-0.0564747192,-0.1463487595,-0.0918448195,0.1887684017,0.0247337054,0.3445622325,-0.119789429,0.1697074324,-0.0821185708,0.3426369429,-0.0398551635,0.0763787478,-0.4161768258,-0.1355724335,-0.4027681947,0.1609249711,-0.0098998137,0.1158459559,0.0341134854,0.2473656088,-0.2458713949,0.2662317157,0.1309258491,0.0437190644,-0.0359608755,0.0830051824,-0.34445104,0.0456631258,-0.2424387187,-0.1139571071,0.2453490049,-0.136760816,-0.1752684265,-0.1287543178,0.2528265417,-0.1321383864,0.1615892947,0.1247489154,0.4439778626,0.2961933315,0.0682060793,0.1773564517,-0.1089350432,0.1017327607,-0.0991781652,0.0317045599,-0.0573241711,0.2549757659,0.1116755232,0.3584685922,-0.0195242874,-0.1864133328,-0.222733885,0.2448714525,0.3298005462,-0.2393908203,-0.0345268995,0.2742847502,-0.1462465376,-0.1039075032,0.0426047146,0.2757709026,0.0453789011,0.4797607362,-0.1038704515,-0.4449886084,0.5055987239,0.0036407665,-0.3741857111,0.2025415748,0.315913856,0.3365118504,-0.1287272274,-0.4725493789,0.1250617206,0.4936556518,-0.1135326996,-0.0947745293,0.2826546729,-0.2521481812,0.0280762855,0.1982813627,-0.1538530141,0.0279016066,-0.0774058402,0.0954035819,-0.0694282576]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1741","title":"error when run fine_tuning on text_classification","comments":"none","body":"dataset:sem_eval_2014_task_1\r\npretrained_model:bert-base-uncased\r\n\r\nerror description:\r\nwhen i use these resoruce to train fine_tuning a text_classification on sem_eval_2014_task_1,there always be some problem(when i use other dataset ,there exist the error too). And i followed the colab code (url:https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/text_classification.ipynb#scrollTo=TlqNaB8jIrJW).\r\n\r\n\r\nthe error is like this :\r\n`File \"train.py\", line 69, in <module>\r\n trainer.train()\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/transformers\/trainer.py\", line 784, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 435, in __next__\r\n data = self._next_data()\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 475, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\nKeyError: 2`\r\n\r\nthis is my code :\r\n```dataset_name = 'sem_eval_2014_task_1'\r\nnum_labels_size = 3\r\nbatch_size = 4\r\nmodel_checkpoint = 'bert-base-uncased'\r\nnumber_train_epoch = 5\r\n\r\ndef tokenize(batch):\r\nreturn tokenizer(batch['premise'], batch['hypothesis'], truncation=True, )\r\n\r\ndef compute_metrics(pred):\r\nlabels = pred.label_ids\r\npreds = pred.predictions.argmax(-1)\r\nprecision, recall, f1, _ = precision_recall_fscore_support(labels, preds, average='micro')\r\nacc = accuracy_score(labels, preds)\r\nreturn {\r\n'accuracy': acc,\r\n'f1': f1,\r\n'precision': precision,\r\n'recall': recall\r\n}\r\n\r\nmodel = BertForSequenceClassification.from_pretrained(model_checkpoint, num_labels=num_labels_size)\r\ntokenizer = BertTokenizerFast.from_pretrained(model_checkpoint, use_fast=True)\r\n\r\ntrain_dataset = load_dataset(dataset_name, split='train')\r\ntest_dataset = load_dataset(dataset_name, split='test')\r\n\r\ntrain_encoded_dataset = train_dataset.map(tokenize, batched=True)\r\ntest_encoded_dataset = test_dataset.map(tokenize, batched=True)\r\n\r\nargs = TrainingArguments(\r\noutput_dir='.\/results',\r\nevaluation_strategy=\"epoch\",\r\nlearning_rate=2e-5,\r\nper_device_train_batch_size=batch_size,\r\nper_device_eval_batch_size=batch_size,\r\nnum_train_epochs=number_train_epoch,\r\nweight_decay=0.01,\r\ndo_predict=True,\r\n)\r\ntrainer = Trainer(\r\nmodel=model,\r\nargs=args,\r\ncompute_metrics=compute_metrics,\r\ntrain_dataset=train_encoded_dataset,\r\neval_dataset=test_encoded_dataset,\r\ntokenizer=tokenizer\r\n)\r\n\r\ntrainer.train()\r\ntrainer.evaluate()\r\n\r\n","comment_length":1,"text":"error when run fine_tuning on text_classification\ndataset:sem_eval_2014_task_1\r\npretrained_model:bert-base-uncased\r\n\r\nerror description:\r\nwhen i use these resoruce to train fine_tuning a text_classification on sem_eval_2014_task_1,there always be some problem(when i use other dataset ,there exist the error too). And i followed the colab code (url:https:\/\/colab.research.google.com\/github\/huggingface\/notebooks\/blob\/master\/examples\/text_classification.ipynb#scrollTo=TlqNaB8jIrJW).\r\n\r\n\r\nthe error is like this :\r\n`File \"train.py\", line 69, in <module>\r\n trainer.train()\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/transformers\/trainer.py\", line 784, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 435, in __next__\r\n data = self._next_data()\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 475, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/projects\/anaconda3\/envs\/calibration\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\nKeyError: 2`\r\n\r\nthis is my code :\r\n```dataset_name = 'sem_eval_2014_task_1'\r\nnum_labels_size = 3\r\nbatch_size = 4\r\nmodel_checkpoint = 'bert-base-uncased'\r\nnumber_train_epoch = 5\r\n\r\ndef tokenize(batch):\r\nreturn tokenizer(batch['premise'], batch['hypothesis'], truncation=True, )\r\n\r\ndef compute_metrics(pred):\r\nlabels = pred.label_ids\r\npreds = pred.predictions.argmax(-1)\r\nprecision, recall, f1, _ = precision_recall_fscore_support(labels, preds, average='micro')\r\nacc = accuracy_score(labels, preds)\r\nreturn {\r\n'accuracy': acc,\r\n'f1': f1,\r\n'precision': precision,\r\n'recall': recall\r\n}\r\n\r\nmodel = BertForSequenceClassification.from_pretrained(model_checkpoint, num_labels=num_labels_size)\r\ntokenizer = BertTokenizerFast.from_pretrained(model_checkpoint, use_fast=True)\r\n\r\ntrain_dataset = load_dataset(dataset_name, split='train')\r\ntest_dataset = load_dataset(dataset_name, split='test')\r\n\r\ntrain_encoded_dataset = train_dataset.map(tokenize, batched=True)\r\ntest_encoded_dataset = test_dataset.map(tokenize, batched=True)\r\n\r\nargs = TrainingArguments(\r\noutput_dir='.\/results',\r\nevaluation_strategy=\"epoch\",\r\nlearning_rate=2e-5,\r\nper_device_train_batch_size=batch_size,\r\nper_device_eval_batch_size=batch_size,\r\nnum_train_epochs=number_train_epoch,\r\nweight_decay=0.01,\r\ndo_predict=True,\r\n)\r\ntrainer = Trainer(\r\nmodel=model,\r\nargs=args,\r\ncompute_metrics=compute_metrics,\r\ntrain_dataset=train_encoded_dataset,\r\neval_dataset=test_encoded_dataset,\r\ntokenizer=tokenizer\r\n)\r\n\r\ntrainer.train()\r\ntrainer.evaluate()\r\n\r\n\nnone","embeddings":[-0.2144676298,-0.55945158,0.0576395318,0.3607929647,0.4646721482,0.1070740595,0.2590162754,0.2739262581,-0.5177255273,0.1579576582,-0.1494843513,0.1323395371,-0.251417309,0.2249938846,-0.1775562912,-0.1774537563,0.0042477539,-0.0007083664,0.157754302,-0.1244561151,-0.1295443475,0.1497635096,-0.5183944702,0.0840891525,-0.2092587501,-0.227935791,-0.1029543355,0.0204488896,-0.2851721048,-0.3128148019,0.3561379313,-0.0759477094,0.3352320492,0.2108807415,-0.0001218343,-0.0217116103,0.0561814606,-0.1555660516,-0.2418771684,-0.2812211215,-0.3933752477,-0.239844203,0.171027109,-0.1465002447,-0.1228254065,0.1302608699,0.0469782837,0.1569697112,0.6134641171,0.3661015928,0.0856523588,-0.1938480735,-0.0240622014,-0.1578443199,-0.2037757933,-0.110580869,-0.2583837807,0.3704072535,-0.0928423554,0.068268992,-0.0323592462,0.3426347375,-0.1620995253,0.1414745152,0.1527857184,0.1371457428,-0.1212285534,-0.3955920339,0.3103570938,0.3197604418,0.3017079532,-0.0959337205,-0.4639147818,0.0200480018,0.4723022878,-0.4898726046,0.08417321,0.0531602651,-0.0719418004,0.3566849828,-0.1625478566,-0.2320619822,-0.2285587341,-0.0690730214,-0.0013151723,0.2779161036,-0.2424880564,-0.02225958,0.3171557486,-0.0054102601,0.1033282354,0.1265070438,0.2395551056,0.1882403195,-0.5437376499,-0.1652665585,-0.3719879985,0.2643744051,-0.0967103615,0.3280752897,0.0326840766,0.0275556725,0.201085493,-0.0614743829,-0.1405315697,0.1795584708,-0.0529032536,0.3850485384,0.3579415977,0.1519238204,-0.0868584737,-0.0435880236,-0.4803898633,-0.0554622002,-0.0868302509,-0.0521250032,0.0820594281,-0.3345047534,-0.5001711845,0.0764316097,-0.3954467177,0.0277458355,0.0426664278,0.511560142,-0.0886732936,-0.3290576935,0.1648975462,0.2891061306,-0.119461976,-0.1834414154,0.0543629825,0.1438683718,-0.0908901915,0.0162098184,0.2309962064,-0.4651764631,0.1514147222,0.005590044,0.1147035435,-0.3081361055,-0.0353146903,-0.3461569846,-0.1585471332,0.271283567,0.0778261125,0.1396289468,0.0754875094,0.1985090524,0.0102728857,0.0041879728,-0.4581688046,-0.4856680334,0.0112793129,0.0634451509,-0.2260698974,0.0156436618,0.0008685407,0.3366593122,0.2206213325,0.1662385017,-0.0254106894,-0.2825123668,-0.521553278,0.1517682374,0.3867296278,0.1500691622,-0.1783468127,-0.1375587881,-0.3402281404,0.2585060596,0.2651514411,0.3910776973,-0.0236571282,0.3891070783,-0.0070875054,0.1945904791,0.2778442204,0.0799790397,-0.3496067226,0.0684830844,0.1459902674,0.1372625083,0.1278059632,-0.4341709912,0.223381415,0.0116829947,0.1824317724,0.1283530146,-0.1260718703,-0.3800552487,-0.4106817842,0.0656786561,0.1390874535,0.0233124793,0.2530212998,-0.0189502146,-0.1890088767,0.4210623503,-0.1502419263,-0.2328864932,0.2234416157,0.1652293205,0.1307042092,0.0001885965,0.1775277108,0.0703620985,-0.2100604624,0.0006586665,0.321819365,0.5056551695,-0.2823263705,-0.0844112635,-0.2416575998,-0.056435544,-0.2238537222,-0.3163668811,0.0479145274,-0.1444974989,-0.2657391429,-0.1667229533,-0.1563924253,0.3217224479,-0.3533496857,0.2459317744,0.2166710049,0.2062280327,-0.2813818455,-0.2946025431,-0.0675818697,0.2301955521,0.0187874939,-0.152391538,-0.1970032007,0.3193896711,-0.0893802643,-0.2433020771,0.0288903061,-0.0629306734,0.0077072866,-0.6136330962,-0.1548435241,0.0385748222,-0.0925097242,0.0317475088,0.2056771517,0.542509675,0.3495855927,0.0600222275,0.0675695539,-0.0318625979,0.0530409142,0.0929042175,-0.0474784598,-0.3042319715,0.3361219764,-0.1383003294,0.4412943125,-0.1156934649,-0.4338339567,-0.1927289963,0.2280964106,0.0142987473,0.1445791125,0.3752095699,0.0421452522,-0.0882832557,-0.1993863881,-0.2938042581,0.2348694801,-0.0722848475,-0.1888148785,0.4041289687,-0.026635984,-0.0812387243,0.1776303947,0.216355443,0.0874854624,0.1493326128,0.0637782663,0.0669434965,-0.0733328983,-0.2942897081,0.0454320237,0.5072415471,-0.4115504026,0.1506951004,-0.2992470562,-0.0570116714,-0.3729400635,-0.1500742286,-0.3120444119,0.0872932076,0.2193260193,0.0241922215,0.2109887302,0.25467816,0.0690871328,0.3314626217,0.1407992542,0.3808372617,-0.1311869621,-0.0354064628,-0.2320403159,0.0089550558,0.198794663,-0.0593902469,-0.130157575,-0.1958557814,0.0380235612,0.3792105317,-0.4469455779,0.2396788746,-0.218787536,0.3500809968,0.2582999468,0.1284429729,-0.6027834415,-0.2113817632,0.138061896,0.1216313243,-0.1181183085,0.07745599,0.0051881871,0.1855076402,-0.1114366949,-0.3393699229,-0.3153971136,-0.0294418279,-0.0597359464,-0.165209949,-0.0031229369,0.3289567232,0.0415355414,0.1801893264,0.1810934991,0.198592484,-0.0398441516,-0.1235805154,0.3934524655,0.0093801301,-0.1322584301,-0.1372332722,-0.1251955926,0.2386852652,-0.0744422823,-0.227809906,-0.0286319368,0.2337598205,-0.043935582,-0.2324396968,-0.2886560261,0.0443252176,0.042999696,0.0943720117,-0.1531774998,0.0536484458,0.4128887057,-0.1876622587,0.2014043778,-0.2260108292,0.325843662,-0.0197635163,0.6415804029,0.2971327007,0.0336703695,0.0647119209,0.1820463985,0.0518957786,-0.1074064076,-0.1801529974,0.125554949,0.1071860716,-0.333615303,0.2742439508,-0.0081405155,0.2723384202,-0.0822928846,0.022218233,0.0217604432,-0.1920213252,0.1448641717,0.0788819045,0.3479619622,0.1071721241,0.1647070199,0.2482372075,-0.0710881576,-0.0702566057,0.4527952969,-0.0725552142,0.1274129003,0.2135470361,-0.4892085791,-0.2318516076,0.3728888035,0.3455658853,0.2752574682,-0.0803047791,0.1931768507,0.3021771014,0.1722811908,0.0999407098,-0.2889327109,-0.2922037244,-0.04837396,-0.2737625837,0.2650038302,-0.0925008506,-0.3705333173,-0.3581210971,0.455072701,0.3515622318,-0.0559598021,-0.1951598078,0.388897717,-0.0360779203,-0.2446646243,0.0775556713,-0.1626805812,-0.2375956178,-0.1646864116,0.2968493104,0.220888108,0.256090641,0.0106645906,0.4308206737,-0.0708304718,0.0674488097,0.1045571938,0.3445489109,0.0730936155,-0.1427156031,0.1227377206,0.2122327983,-0.2894381881,0.5933100581,0.2986900806,-0.2552807927,-0.4227852523,0.1464388669,-0.0146914413,0.2030423135,0.1039180756,-0.2396741211,0.1750097126,0.010736661,0.016125856,0.1451515555,0.1978329122,0.3457220793,0.311463356,-0.6275060177,-0.4609931111,0.0915696695,0.0754156932,0.0328380279,-0.2523558438,0.2001631707,-0.2753789723,0.429196924,0.0997262225,0.9185284376,0.0339833647,0.0632902607,0.1496915519,0.3176858723,0.1395189464,-0.2998842895,0.3164820075,-0.102861546,-0.5518776178,-0.0974635929,0.1286051422,0.2069117725,-0.1046038195,-0.4925997555,0.3480652869,0.203514412,-0.1167211086,-0.0075149196,0.039330896,0.5583392978,-0.0051785214,-0.482289046,0.0100889467,-0.1556378305,0.4814767241,-0.0938074887,0.0214937367,0.1038863808,-0.1593244225,-0.39688465,-0.0020872494,-0.1270958632,0.0382475853,0.1476843059,-0.468251735,0.5285185575,-0.0543910787,0.5922079086,0.1566279382,-0.2321147472,0.0754158199,0.07199727,-0.0018886165,0.0476917662,0.0257460978,0.5697031617,-0.2891815007,0.0543860234,-0.0406563766,0.0060357731,-0.0050418279,0.187775299,-0.3678410053,0.0068217744,-0.381577462,0.2264733016,-0.0489138626,0.1850270778,-0.4350206554,0.0717098787,-0.1046975031,-0.2740537524,-0.0096312594,0.3100106716,-0.4267202616,-0.1290108413,0.4233057201,0.1633353233,0.098512359,0.6959379315,-0.0396995507,-0.0932731256,-0.1243054196,-0.0982386917,0.1026488394,-0.335694164,0.3115037084,-0.0741347373,0.094585456,0.1008643359,0.6131538749,0.1983929574,0.1453306377,-0.1436981112,-0.2602827251,-0.4799219668,-0.0724010393,0.1088487655,-0.0735350326,-0.0127816489,0.5952906013,0.2521891296,-0.0170807522,-0.2138652205,-0.0498230346,-0.151790157,0.169791773,-0.0434500612,-0.18072173,0.2263089269,0.377504617,-0.0011016966,0.3562057912,-0.247344926,-0.0537674762,-0.2055985034,0.175970003,0.0304148402,-0.2934740782,0.0037564756,0.3029086292,-0.0495062284,-0.0995158926,-0.0880216286,-0.0844683871,0.1070486531,0.1282210499,-0.2037877887,0.2722803652,0.0088060433,-0.2573166192,0.0027208531,0.1445866823,0.2534755468,0.0744739175,0.0849617645,-0.0861943588,-0.1005175933,-0.0114090247,0.1722123325,-0.0684689805,0.0436763689,-0.3248052597,0.0799093843,-0.1603556424,0.2519840002,0.252560854,-0.0786894262,-0.2915099859,0.3951742053,0.0120956209,-0.4366323352,0.3250692189,0.064941451,-0.0407732502,-0.2584431171,0.4039795399,0.0954512581,0.1506994516,0.0968993157,0.1145276353,-0.003678224,0.0418784432,0.2462931573,0.4967960417,0.1020970047,0.4097110331,0.3586777449,-0.2521685958,0.1725097746,0.2228470594,-0.4518596232,0.4035180509,-0.0400413796,0.1152072623,0.2191981822,-0.4616671503,0.0348906554,0.2105672657,-0.0401436649,0.0503163338,-0.2401469797,0.4135707915,-0.2128786445,-0.3604338467,-0.0706168711,0.2925843894,-0.4412465692,-0.2755232751,0.1876989603,-0.0479518548,-0.2804298401,0.1016741842,-0.1401610225,0.0510668494,0.2977919877,0.0440342054,-0.1373703629,-0.1902676225,-0.3093777299,0.1785401553,0.1054386273,-0.2609920502,0.408141613,-0.0728892535,-0.2416422218,0.1030794233,0.2154896855,0.3861976564,0.2246137261,0.2517972291,-0.2542098761,0.0756266639,0.088005662,0.0639370009,0.1138616055,0.293387413,0.5774318576,0.1765705049,-0.0918676481,-0.1493518949,0.4720678926,-0.0396476164,0.1358252764,-0.2823147476,-0.1547411531,0.2001651078,0.0334112458,-0.20860295,-0.0652208179,-0.0797021762,0.3930701911,0.170544371,0.1467795223,-0.0333734751,0.0189959612,0.0030013351,0.0724484324,0.575732708,0.4242975116,-0.4088569283,-0.2542712092,-0.2262933254,-0.2775295973,0.1452726275,0.0166662727,0.0582860634,0.0343176909,-0.0575019605,-0.1634641588,0.0148090906,0.0571044683,-0.2960196137,-0.0618949793,0.2778194845,-0.3522389233,0.0264310371,-0.2927132547,0.0632290617,0.0521417335,-0.0333995149,-0.0402370542,-0.038822148,-0.0914316922,0.09862829,-0.2090255022,-0.1474615037,-0.2267259657,0.2743208408,-0.0407853015,0.2383604497,-0.0761104897,0.0737821832,-0.1771598607,-0.2301980853,-0.2703691423,0.4310872257,0.3395461738,0.3627293706,-0.0534431376,0.155490309,-0.2907877266,0.0862981379,0.2800109088,-0.1542855203,-0.2129245847,0.2927227616,-0.1480735242,-0.0181017239,0.0301921777,0.1824199706,0.1302612573,-0.0902800187,-0.6067728996,-0.4277603328,0.7808181047,-0.3343407214,-0.2089636177,-0.3239196539,0.1138707623,0.4230691493,-0.305116266,-1.1453722715,0.0158926938,0.2261588126,0.0545447618,-0.2847104073,0.1985793412,0.1125663146,0.0890891477,0.0105343163,0.2650995255,0.0275371764,-0.3521079123,0.113341786,-0.3723094165]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1733","title":"connection issue with glue, what is the data url for glue? ","comments":"Hello @juliahane, which config of GLUE causes you trouble?\r\nThe URLs are defined in the dataset script source code: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/glue\/glue.py","body":"Hi\r\nmy codes sometimes fails due to connection issue with glue, could you tell me how I can have the URL datasets library is trying to read GLUE from to test the machines I am working on if there is an issue on my side or not\r\nthanks ","comment_length":20,"text":"connection issue with glue, what is the data url for glue? \nHi\r\nmy codes sometimes fails due to connection issue with glue, could you tell me how I can have the URL datasets library is trying to read GLUE from to test the machines I am working on if there is an issue on my side or not\r\nthanks \nHello @juliahane, which config of GLUE causes you trouble?\r\nThe URLs are defined in the dataset script source code: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/glue\/glue.py","embeddings":[-0.0552913994,0.0118692834,-0.0427740142,0.2993855476,0.3010600507,-0.2684803307,0.2035407424,0.0815812498,0.1122899055,0.126336351,0.0642895028,-0.0648157224,0.2632441223,0.1945556253,0.2005842328,-0.0129128126,-0.1107104123,-0.0857953802,-0.3790926933,0.0695378631,-0.2274179757,0.1339347214,-0.2569567859,0.1540927887,-0.2331764847,-0.0182273686,-0.2408938557,0.1416757405,-0.1169200763,-0.3114697337,0.2458859384,0.2722536027,0.0157695338,0.2451800853,-0.0001094538,-0.085432075,0.4476234019,0.0518655442,-0.0857244059,-0.0963327363,-0.4144301116,-0.1583484858,0.1298841983,-0.2039457262,-0.172820285,0.4333641827,0.1408487856,-0.3225314021,0.1861434132,0.1474582404,0.2255823165,0.3084511757,0.2373654991,0.037909925,0.3510044217,-0.2436030507,0.1521846205,0.4593857527,0.3264584243,-0.1734827757,0.113360852,0.1072214246,0.1378021985,-0.0038308683,-0.0707272291,0.0960997865,-0.2217198759,-0.3329151571,0.0700250566,0.3006030917,0.3522767723,-0.145694539,-0.2870568335,0.0623582155,-0.1634307504,0.3614005446,0.3465101719,0.2372853011,0.0861449018,0.2820421755,-0.1781607717,-0.2360922396,-0.2536133528,0.2970753312,0.0505382232,-0.0472006425,-0.2656107247,0.2072946429,-0.0061237258,-0.2143399715,-0.0764592513,-0.1481080204,-0.0374755003,0.1016432494,-0.1890322417,-0.1231699586,-0.1298278272,0.539232254,0.3787512183,0.0442047007,0.0274433456,0.0363213941,-0.1351932734,0.2267657816,0.4534760714,-0.1265631914,0.1274647862,-0.0749848261,0.558650732,0.1327300072,-0.1617051959,-0.2315634042,0.0721815377,-0.2383840233,-0.1969603598,-0.1852637827,0.4247693121,-0.3179559112,-0.0180465151,-0.0207218789,-0.2300951034,0.0216086395,0.0870369077,0.3594588339,-0.3293738961,0.0328834988,-0.0079367915,-0.1125487909,-0.1614854783,-0.1475437433,-0.1700424999,0.0387516581,-0.2326640636,0.0795596838,0.0292003993,-0.2172808349,0.0885176808,0.0719203278,-0.1632151157,-0.085661903,0.0172570609,-0.1295012534,0.1328124851,0.1723618656,0.1386316717,0.2162532657,-0.1134451553,-0.3098357618,-0.039017044,-0.1024172157,-0.3364962339,-0.0733487457,-0.014145284,0.169818759,-0.4461430907,0.0614763573,-0.0504904576,-0.1832020879,-0.3620137572,0.0507001504,0.0676263869,0.0361977257,0.0771713182,-0.4720363319,0.328597039,0.4581777453,-0.2045367658,0.0299887862,0.2047373801,-0.2899807096,-0.1610530615,0.2446179986,-0.0275420807,-0.2317216247,-0.3104969561,0.3071649373,0.2097474933,-0.4537995756,-0.5420528054,0.5876682401,-0.1791019887,-0.2067996562,-0.0125934463,0.1852538139,0.149897024,0.0511213839,0.0089088762,0.5321074724,-0.0603240207,-0.1297146678,-0.2670769393,-0.2808828354,0.0112549728,0.0750852898,-0.1924652606,-0.1969628036,0.3012685776,-0.3112698793,0.2584023774,0.2315500081,-0.0003224949,0.2314703912,0.4709127843,0.1136943847,0.191098243,0.0146156047,-0.3011435568,0.0670153126,0.0895378441,0.0161551274,-0.1345145851,-0.1280486584,-0.040525768,0.0316529498,-0.0519643165,-0.0366878211,0.1640718132,-0.0981393754,0.1427839249,0.2040522397,-0.095654048,0.5069954395,0.0449108481,0.2054917663,-0.3476516604,0.7957617044,-0.0628449991,-0.0831364542,0.1679748446,0.1581304371,0.3577303886,-0.2406215221,-0.0793361738,0.5479306579,0.0030467724,0.2087198049,0.4648137987,0.3619560897,0.2616003752,-0.2679555714,0.176691547,0.1234835014,0.1318592578,0.0528849438,0.2697487473,0.0565306731,-0.0929097086,0.3933546543,0.2558864653,0.2535201013,-0.0885492265,-0.2306993604,-0.3375173211,0.0257895011,0.1821230799,-0.0102817575,0.3197614849,0.0665562302,-0.1968088746,0.0343665332,0.2283052355,-0.2139407694,0.1477039605,-0.1218866035,-0.242770493,-0.0533956476,0.0949436128,0.4247909486,0.397308141,0.1556915045,-0.1571696997,0.1278558224,-0.1030102223,-0.1715970635,0.1571452916,0.0120760379,-0.1970631778,-0.2106176764,-0.1756360382,-0.1759063154,-0.062820144,-0.0401063003,-0.2609094083,0.0049221944,-0.4890078306,-0.1162075996,-0.42560938,-0.4683541358,-0.2294389606,-0.2413150668,-0.4593434036,-0.4235491157,-0.2111191154,-0.1315097809,0.0404940136,0.0419472121,-0.0451764576,0.2296126485,-0.1644717008,0.0167804677,-0.0442774966,-0.1598864049,-0.3016820252,0.0908692554,0.0384635665,0.3077384531,0.3564952016,-0.131058529,0.2043779343,-0.3110781014,-0.1825407743,0.0033888323,0.1127296984,0.5722730756,0.4303124547,0.3833498955,0.3812004626,0.0102159157,0.386446327,-0.0884319842,0.1346813887,-0.1440852284,0.1837702245,-0.1304980963,0.1290677786,-0.1468420625,-0.3778319657,-0.0945281982,0.2201990485,-0.3047661185,0.088039957,-0.1557004452,-0.0484125949,0.2364871055,-0.383141309,-0.1040025577,-0.2310832292,-0.5189390182,0.1645469666,-0.0906724259,-0.4740213156,-0.0076211207,0.0513649695,0.1766807586,0.0559087172,-0.3993262053,-0.5216785073,-0.0929370522,0.0221441817,0.3070882559,0.3270330429,0.1475023031,-0.2089836299,0.0304434504,-0.1695893109,-0.2473968714,0.1664075851,0.0020562306,0.2740554512,0.5102769732,0.2605576813,0.0484128483,0.1950868219,0.1677782387,-0.1330775619,0.2533980012,-0.1815490574,0.4770168066,0.3129312694,-0.3273966908,-0.0628941208,-0.0154644242,-0.0486290902,0.0327871889,0.2938314676,0.1585819572,-0.3641500175,-0.111448355,-0.3802973032,-0.3738039732,-0.1508466005,0.0350634605,0.1327314228,0.0921060517,-0.0486222543,0.141505003,-0.1445543468,0.0976966619,0.409224242,0.0653709322,-0.1441711187,-0.3791982532,-0.4537935853,-0.3430480063,0.4417433739,-0.0403177775,0.5260638595,-0.2394114733,0.2421417087,0.1949622482,0.022884462,0.6257149577,-0.8164910674,-0.0070504872,0.0928349122,0.0328910872,-0.3561477661,0.128403917,-0.1157548875,-0.0272787437,0.3937647343,0.2405095845,-0.2447447628,-0.1883088499,0.1386578679,0.0054346272,-0.1128375456,-0.2477955967,0.0030020692,-0.2437552214,-0.0549951382,-0.1995415986,-0.169307366,-0.0894130245,0.2424606234,-0.0916549563,-0.0216374714,-0.0464587063,0.0621432476,-0.1363262832,0.1105245277,-0.15540649,0.2256638408,0.4375859201,0.2917253077,-0.1947759986,0.2232237458,-0.2668697536,0.1280412376,-0.0454734154,-0.022897087,0.5653005242,0.3047223389,-0.1275051534,0.2242003679,0.0120504135,0.0939892754,-0.2646261752,0.0565565787,0.3686784804,0.1756885648,-0.2953342199,-0.378960669,0.2947137654,0.0836953595,-0.2495907098,0.1924502105,0.3177802265,-0.1595832705,0.2043216825,-0.1447807401,1.0169451237,0.0349386558,-0.084538132,-0.0615023151,-0.4995140135,0.3972772658,0.0536320247,-0.0996960625,-0.4945423901,-0.2277206928,-0.2082972676,-0.2642942965,0.2048657238,-0.0383235663,-0.1943507195,0.5626404285,-0.1556049436,0.254873544,-0.0557968467,0.3983104229,-0.0434016548,0.1283490062,-0.2900544107,0.2098005116,-0.1040625423,0.3654922843,-0.0500560217,-0.1420089155,0.1566396207,-0.2494006902,-0.3668981194,0.0266251359,-0.7568395734,-0.04917375,0.238357693,0.0457224697,-0.1346175522,0.05372357,0.1496385783,0.4605714977,-0.2692959905,0.1116435453,-0.0248332843,-0.1819412708,0.2858279049,0.2463254482,0.0267923195,-0.1077813953,-0.329649955,0.0096886512,0.1551782489,-0.1951411068,-0.0431731753,-0.0618030652,-0.0139537919,-0.3604128063,-0.1586524397,-0.1269038022,0.213603273,0.0761332884,0.1427555531,0.2127289176,0.0423799567,0.2980400026,0.2540828288,-0.1287964433,-0.0947677568,0.4355383813,-0.3086751997,0.0063694157,0.0378936455,-0.0339656994,-0.2220364809,-0.1452679783,0.2257869244,0.2803980708,-0.1841592938,-0.1570624262,-0.1335282773,0.218123734,-0.2223347127,-0.2338033915,0.2138862908,0.1519240737,-0.0702140704,-0.5190091729,0.0507641546,0.3575649858,0.1958250552,0.1384168565,-0.0674845129,-0.2428609878,0.1987520754,-0.13099356,-0.3228833079,0.2070924342,-0.2509309053,0.034299504,0.5042775869,-0.127954483,0.3872488737,-0.1507943124,0.1516044885,-0.2329693884,-0.1897889674,-0.1367394477,-0.091343224,0.1242059842,-0.0518360883,-0.1987394542,-0.1399902701,-0.5320920348,0.0867602155,0.0653605834,-0.0412718058,0.1310142428,-0.0736202672,-0.1985262632,-0.1111866683,-0.1988460422,-0.2343993187,0.5295158625,-0.0517399982,0.428868413,0.0220480859,-0.0075596943,-0.1291714758,-0.2968685031,-0.0157876126,0.1475239843,-0.1545960456,-0.0495596193,0.2362096608,-0.0682055503,-0.1005271226,0.2418270558,0.4660553038,0.0894192904,-0.2641600966,0.111151129,0.1586784422,0.1903086752,0.0985202342,0.0805367306,0.2481308579,0.0509819947,0.0627993271,0.0710276887,0.110489592,0.1608980596,0.0799778476,-0.0279995669,0.6018794179,-0.01247501,-0.1534199715,-0.0428467095,-0.1130861714,-0.0245188326,0.1392493546,0.159964487,-0.0297674462,-0.1557417661,0.0245952643,0.0581391193,0.3666993976,0.1528993845,0.2531642318,0.0034270026,-0.1608128101,0.1576141715,-0.2578915954,-0.1013397351,-0.1850724071,0.7514363527,-0.3354577124,-0.0197870452,0.0422042347,0.2192431241,0.1011179164,-0.1890845895,0.0355265215,-0.0211011935,-0.4360969067,-0.2370218188,0.0235330015,-0.2358254641,0.2895056307,0.0476795547,0.2070529312,-0.3042061627,0.0768790916,0.067093946,0.1367952973,-0.16746144,0.2207103968,-0.2710696757,0.5232746005,0.2478224039,0.4361562133,0.4109049439,0.247524187,-0.0898542255,0.0824168399,0.0135765271,-0.0878052264,0.0734698847,0.1821487546,-0.0129206488,0.054478012,0.4246553481,0.208585754,-0.0741776302,0.0050215912,-0.2670271993,0.2647578418,-0.1759053171,0.4376753867,-0.3028048873,-0.0440963358,-0.4799283147,-0.1268560886,-0.5758994222,-0.1265254617,0.3075518608,-0.3345972896,0.1155702621,-0.2402162701,0.0843651816,-0.0755840838,0.4613205791,0.492344439,0.6064280868,-0.3703455627,-0.3879162371,-0.5264345407,-0.117371209,-0.1129512265,-0.1075243801,-0.0976995379,0.0565412864,-0.2380265146,0.2443152517,0.1563062072,0.3593222201,-0.1874040514,-0.0764413252,-0.1741304249,-0.0336344577,-0.1541620195,0.0143577391,0.0307837315,-0.2010679394,0.2596473694,-0.0500959344,0.0287054181,0.0635404363,0.0763314441,-0.069521673,-0.3577857018,0.2602365613,0.1772950888,0.4283646643,0.17209436,0.0787238404,-0.3032388389,-0.3302256465,0.0490187295,-0.057802137,-0.1350289285,-0.1704171002,0.1871255338,-0.190078035,-0.2916152775,0.0989706218,-0.1606833488,0.5082817674,0.2039367855,-0.2726267874,0.1324838996,0.0322695188,0.141062811,-0.0009934891,-0.1701037437,0.1290653348,0.1595219672,-0.2460220456,0.3463923931,-0.264411509,0.0783219934,-0.1526682377,0.4158907533,-0.0216395333,-0.0334009938,-0.333814621,-0.1911590397,0.4394245744,0.0418402664,-0.1830800921,0.3604086936,0.0428948738,0.2880257368,0.1544221789,0.1137441024,0.1443602741,-0.1935795099,0.0694854483,-0.0632313043]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1731","title":"Couldn't reach swda.py","comments":"Hi @yangp725,\r\nThe SWDA has been added very recently and has not been released yet, thus it is not available in the `1.2.0` version of \ud83e\udd17`datasets`.\r\nYou can still access it by installing the latest version of the library (master branch), by following instructions in [this issue](https:\/\/github.com\/huggingface\/datasets\/issues\/1641#issuecomment-751571471).\r\nLet me know if this helps !","body":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.0\/datasets\/swda\/swda.py\r\n","comment_length":54,"text":"Couldn't reach swda.py\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.0\/datasets\/swda\/swda.py\r\n\nHi @yangp725,\r\nThe SWDA has been added very recently and has not been released yet, thus it is not available in the `1.2.0` version of \ud83e\udd17`datasets`.\r\nYou can still access it by installing the latest version of the library (master branch), by following instructions in [this issue](https:\/\/github.com\/huggingface\/datasets\/issues\/1641#issuecomment-751571471).\r\nLet me know if this helps !","embeddings":[-0.2362142503,-0.3874821961,-0.1583183408,-0.0046996926,0.3615036309,-0.0226995479,-0.0685163513,0.2251031548,0.0263329204,0.218404904,-0.1205207929,-0.007295799,0.0098906029,0.3339485228,-0.0140376193,-0.2254329771,0.0509144589,0.0313128009,-0.256431818,0.0775595978,0.0202031508,0.325912118,-0.1554065049,0.2297449112,-0.1134689227,-0.0685833097,-0.0201578867,-0.0775326192,-0.175789386,-0.4362722039,0.5638242364,0.1927216649,0.1256830394,0.3472210169,-0.0001007638,-0.0771251395,0.4186194837,-0.0503711626,-0.1755157858,-0.309874028,0.0497583449,-0.1840837002,0.212349236,0.1146646366,-0.2763558328,-0.3036899567,0.1272522956,-0.0963162184,0.4207452238,0.0446711183,0.3689472973,0.2689450681,0.483820796,-0.4246512055,-0.0924544185,0.1234186217,-0.082154952,0.2442761213,0.0605740286,-0.0762190372,0.139697358,0.352983892,0.1305269897,-0.0450913236,0.301171571,-0.045279853,-0.1961440891,-0.2342961282,0.239060387,0.1528021693,0.5129566789,-0.2608686388,-0.4449080527,-0.0117749497,0.06313245,0.0231450815,0.1606127024,-0.0162839703,-0.1544131637,0.1218490675,-0.0956054926,-0.0866331384,-0.1289826334,0.1462022811,0.115401946,0.0305994917,-0.2691115737,0.0719906092,0.2066189647,-0.3192819357,0.0528106168,0.1959657371,-0.1088816002,0.2442480624,-0.178972438,0.0687169731,0.0610098392,-0.0150783351,0.3950170875,0.0805762783,-0.1344826221,0.0119044837,0.0585580021,0.0515343063,0.1297624707,0.0434685461,0.0165331922,-0.3237161636,0.2677385807,0.2697731853,0.1876418144,-0.2395910472,-0.0104302671,-0.2571239173,-0.3182106912,-0.3037604988,0.2617524564,-0.2514619529,-0.1170674264,0.0519806109,-0.0632575229,-0.0098854918,0.0848204046,0.4707405865,0.0258671548,0.0295214802,0.1808131039,0.2489558458,0.0418968089,0.253698796,-0.3355825245,-0.0228681862,-0.0963060781,-0.0400750414,0.1465428323,-0.277179122,0.3220079243,-0.0852720737,0.0394149013,-0.0062831119,-0.2069952339,0.1721775383,-0.063818261,0.36161533,0.1384340525,0.1034318581,0.1152971238,0.0605930574,-0.0969779938,-0.2377071381,-0.1938384175,-0.2642819285,-0.2458227128,0.2662176192,-0.0662668422,0.0833410248,-0.101832442,0.0710770041,-0.1863034964,-0.1065629497,-0.0815844238,0.0334060416,0.0264020786,-0.1412682682,0.3396386802,0.3164157271,-0.2243431211,-0.1330294162,0.2190386653,-0.3956375718,-0.0767425522,0.202038303,-0.2344354987,-0.0362960026,-0.2745136619,0.0762544349,0.2386868,-0.2935951948,-0.4345347583,0.0180809591,-0.0682203546,-0.145633921,0.0590169169,-0.0120171206,-0.1774288416,0.1605425775,0.0308523513,0.0477609262,-0.0305401757,-0.0809975117,0.0124896606,-0.2905766666,-0.1029108539,0.0454880446,-0.008517215,0.1289531887,0.1767852157,-0.0447493047,0.2707518935,0.0093451925,0.0446858406,0.21540685,0.2782873809,0.3302980363,-0.0016547708,-0.2491439879,-0.2228259295,0.0563065112,0.1059809774,0.1817086935,-0.2402734309,-0.1508601457,-0.2314825356,0.0776703805,0.0272419248,-0.2084968984,0.2749223411,0.0467883609,0.4856116474,0.2913158834,-0.2161174268,0.2894093096,-0.0706809461,0.2030579001,-0.3141874373,0.3132790923,-0.1704610288,-0.0810309798,0.2489076704,0.3151202798,-0.006789349,-0.1510169357,-0.0285708997,0.2379435152,-0.3230615854,0.067503646,0.2759997547,0.0608956926,0.0068592695,-0.3545849323,0.0661717653,-0.0347228311,0.1070296019,0.075018242,-0.006090587,0.2346146703,-0.0405365936,0.1180555373,0.0864085704,0.2939911485,0.5806789994,0.1721523851,-0.1451497078,-0.0839139149,0.0711118951,-0.3425208628,0.2790347636,-0.2193441242,-0.1918277144,0.0295599606,0.2969753146,-0.0106134815,-0.0616576895,0.2379881293,-0.2565304041,0.0416484326,0.0365801416,0.2131392211,0.280184716,0.3061872125,-0.014422928,0.1521448642,-0.0207854547,-0.1786594838,0.1268230081,0.0331972949,0.0013121923,-0.0468086451,0.0002218561,-0.0261392631,-0.4086170495,-0.1547626555,-0.4498315752,0.1796073914,-0.0121430214,-0.0576352663,-0.2008185834,-0.42852965,-0.0639552996,-0.3473838568,-0.4321454465,-0.3055025935,-0.0869297832,0.123040095,0.0512889028,0.2503718138,-0.0344964713,0.2276600301,0.0568357036,0.1265101135,-0.2285927087,-0.0626287162,-0.0975805745,0.1578879505,0.0749975666,0.1486072987,0.4697656333,-0.1726647019,0.2427733392,-0.3078085482,-0.3838958144,-0.0681613684,-0.1170026734,0.135069266,0.4359375536,0.2606773376,0.1073088795,-0.3065447509,0.3773047626,-0.3965027034,-0.2221936435,-0.1747205704,-0.1398114562,-0.1904285103,-0.1356692612,-0.0715870336,-0.3290016651,-0.3216324747,0.4432864785,0.0032814273,0.0138701145,-0.1150308773,0.1491563767,0.354808718,-0.3191218972,0.0954770744,-0.2415756434,-0.5166891217,0.1373102367,-0.2456361055,-0.5283465385,0.293569535,0.0682730377,0.3508424461,-0.3473393917,-0.3184232116,-0.2784529626,-0.2883039713,0.0639864281,0.0537601076,0.1304268092,0.367338419,-0.055444818,-0.158501029,-0.0401856452,-0.367146343,0.0235889722,-0.2736007273,0.2128954083,-0.0115489615,0.5309565663,0.0039977925,0.3552599847,0.2560673058,-0.0179176331,0.2243462652,-0.085212253,0.3694252968,-0.1641628295,-0.4451016188,0.145262748,-0.0163218156,0.3346743584,0.1585913301,0.2155842185,0.2309572548,-0.2864500582,-0.2261157036,-0.2300143391,-0.1561430544,-0.268058598,0.0996603668,-0.102397725,-0.0105200987,-0.0012376304,-0.1558271945,-0.0800504014,0.0561879911,0.4081967473,-0.0166640077,0.2063531578,-0.4561300278,0.0164850205,-0.6440913677,0.2902688086,-0.1089194492,0.116914399,-0.1910178214,-0.0648416281,0.1081865728,-0.114250727,0.7474418879,0.2149777412,-0.045833528,-0.037172772,-0.2370029241,-0.2169665992,0.017552983,0.1022649407,0.0018339846,0.1910143793,0.2641950846,-0.1633614749,-0.2779791951,0.2271229774,-0.063348271,-0.0807683095,-0.1196158752,-0.1826631427,-0.3779200912,-0.2293116599,-0.0106896199,-0.080282189,0.3667839766,0.0375742093,-0.1185253188,-0.0972491577,-0.2570383847,0.0907770768,0.2076945305,0.0661019161,0.0247233734,0.1202237606,0.3205242157,0.2657814324,0.2866654396,0.3943685293,0.1303093135,-0.376000911,-0.0277026184,0.0071175811,0.2320281416,0.1390189081,-0.2231498212,0.2771843672,0.0751577988,0.3167052567,-0.3357445896,-0.0248058513,0.3693125248,0.2216687799,-0.1189089194,-0.1189015806,0.2103301883,0.0107553815,0.0994332209,0.1974934787,0.5864677429,-0.0503746085,0.140793696,-0.1289325655,0.8587399125,-0.1729924679,0.0596930832,0.1039238051,-0.1804524064,0.7909548283,0.0750596672,-0.0105049144,-0.1966325492,-0.1860005111,0.0389674716,-0.0827530846,0.337451309,-0.2728669643,-0.0060355803,0.148443684,0.0743965134,0.2465984672,-0.090353176,0.1225633919,-0.1150699034,-0.1350624263,-0.4238072336,0.2955787182,0.0525461957,0.5171592832,-0.0467538051,-0.2380643934,-0.1474396735,-0.319245398,-0.237786606,0.0677691549,0.027591018,-0.0119501268,0.2533754408,-0.1183496267,0.0417808704,0.3251807094,0.3742039502,0.1307641417,-0.3935614824,0.0735672191,-0.2684689462,-0.2233149558,-0.1534226388,0.2687022686,0.090008758,-0.214021638,-0.1347144097,0.1969278902,0.1610653549,-0.1661099792,-0.1285467595,-0.1134868935,-0.3286352754,-0.1834145784,-0.1063453332,-0.1068623364,0.0609456822,-0.2337657064,0.2167895436,0.175925687,-0.1900973469,-0.1561124772,0.2755554616,-0.0521448851,-0.0395325907,0.5391178131,-0.1666146517,-0.1935068816,0.3695310056,0.1841918528,-0.1976189315,-0.2024125159,-0.2359915376,0.3632256687,-0.4715962112,-0.0329788551,0.1850077808,-0.0187846944,0.0037766255,0.1230331063,0.2124695629,-0.2078505605,0.1133642346,-0.3853943348,-0.5158317685,0.2861068845,-0.0498593226,0.2116559148,-0.0426580794,0.4333708584,0.0629358664,0.0844092369,-0.4175316989,0.0381263308,-0.2708222568,-0.160228163,-0.0243168864,-0.2146621495,0.3230582178,0.043704018,0.2889895141,-0.0086582378,-0.4189891517,-0.2767114639,-0.0424882397,0.0254526362,-0.0507049859,-0.1335799694,0.1350361705,-0.1641880125,-0.1690312773,-0.0202440415,0.1385892183,0.2416800261,0.0261365697,-0.0839853063,0.1395798922,-0.0528468676,-0.0023196598,0.0482759178,0.1449997723,0.1412596256,-0.0748835728,-0.0746162087,-0.3047424257,-0.0873736963,0.0242527816,0.1072815731,0.0247854497,-0.0389524549,0.2696347535,-0.0875188261,-0.1188457608,0.1732929945,0.6467870474,0.0991852283,-0.1012160778,0.3181520402,0.258559525,0.2481140643,-0.4696616232,-0.051695589,0.0433524661,0.065845333,-0.072409898,0.0034567669,0.0157447793,-0.0013170396,0.0847943649,0.0394551046,0.3680606186,-0.1525165141,0.2127614468,0.2981660664,0.0615514033,-0.0217195544,0.2573702335,0.2060713023,0.1557853669,0.4488225579,-0.3339229822,0.209837541,-0.3700535595,0.2186111957,0.004590217,-0.2674551904,-0.1844625324,0.0444293357,-0.0307833608,-0.2394087911,-0.249041602,0.4403609931,-0.236470893,-0.1967810094,-0.3007943034,0.2721996307,-0.0936316848,-0.3123007119,0.0833825916,-0.2114217728,0.1632499248,-0.1342036277,0.0923008695,-0.2154898494,0.2986693084,-0.1910533309,0.0306890141,-0.3790186942,-0.3417529762,0.1006175652,0.0817705691,0.102799125,0.1398153305,0.1021648124,-0.0774198771,0.1495229304,0.4495654106,0.4510248005,0.1356544495,0.2426744401,0.1698467582,-0.2272647172,-0.1259876192,-0.0107404599,0.3142699897,0.2791320086,0.0453250706,0.2673902214,0.2546359301,-0.2156300694,0.070892714,-0.1534287781,0.368231982,-0.1705945134,0.1954407394,-0.4465487003,-0.0082836729,-0.3041037619,0.2065588832,-0.3922937214,0.0303477924,0.0115931267,-0.1388445199,0.0324864499,-0.3460269868,0.1592811793,-0.045587264,0.5311887264,0.2424556017,0.3454270065,0.0757067502,-0.3340868354,-0.4518872201,0.0233649593,-0.0387929529,0.1890481412,-0.2042272836,-0.0423683226,-0.0280791335,-0.0444067344,0.1455202252,-0.0560373478,0.0207110979,-0.2269634306,-0.1657605916,-0.07350225,-0.2578937113,-0.0830431059,0.1116268486,0.0362980291,0.1515350342,-0.2637988031,0.1011532545,0.0703667998,0.5182725787,-0.1915942729,-0.1213172227,0.1917880177,0.1084795371,0.2277489156,0.0284312032,0.0128329908,-0.3024169803,-0.4064079225,-0.1753892899,0.2050896585,0.0517958812,0.3251955509,-0.1819045991,-0.25654459,-0.1585440785,0.1949697733,0.0109829539,0.2420244813,-0.2220569849,0.3535389006,-0.0386137739,0.0018601188,0.3870484531,0.1296756864,-0.0972554386,0.1025543064,-0.125904873,-0.2123662084,0.6435747743,-0.2412498742,0.0517188311,-0.0005433178,0.0938082263,0.2518316209,-0.4103262722,-0.4838001728,0.2961052656,0.1420930773,0.1774173379,-0.0935305953,0.3457148373,-0.2220901251,-0.1139118075,0.000839346,0.3805652261,0.0373097919,-0.0207490437,0.2242957503,-0.0514386222]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1731","title":"Couldn't reach swda.py","comments":"Thanks @SBrandeis ,\r\nProblem solved by downloading and installing the latest version datasets.","body":"ConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.0\/datasets\/swda\/swda.py\r\n","comment_length":13,"text":"Couldn't reach swda.py\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.2.0\/datasets\/swda\/swda.py\r\n\nThanks @SBrandeis ,\r\nProblem solved by downloading and installing the latest version datasets.","embeddings":[-0.2237073928,-0.3480649889,-0.1958538741,0.0301490873,0.3838005066,-0.1013383642,0.1176114306,0.2514494359,0.0031281312,0.2458541691,-0.0599451996,-0.0130991284,0.1201226264,0.4173997641,-0.0604807921,-0.1694166511,-0.0063660392,0.059304215,-0.2082902342,0.0732943714,0.0003459323,0.2906225622,-0.1776577532,0.1730582565,-0.069383733,-0.03725053,0.0128069902,-0.0443978645,-0.2260177881,-0.4467227757,0.5055350661,0.1106965169,0.0043174354,0.4085598886,-0.0000958232,-0.0176837556,0.4025201797,-0.0343439542,-0.0386779457,-0.1168419793,0.0518369898,-0.1610255688,0.1759162247,0.0516615883,-0.3230692446,-0.3362303078,0.0614052042,-0.0396133289,0.4658401012,0.0454016998,0.4033230245,0.3231393099,0.3288650811,-0.3758232296,-0.070174396,0.0270721409,-0.1071203202,0.1872022152,0.1504050195,-0.1935249567,0.0550176129,0.4520253539,0.0986679569,0.023204498,0.2190804929,-0.0845589787,-0.1063639894,-0.2072822899,0.2562870085,0.1309455484,0.4486600161,-0.2088181823,-0.3486409485,0.0065828264,0.0109828971,-0.0470570624,0.1181083694,-0.0081320191,-0.1311015487,0.1200608313,-0.1040950119,-0.0273672733,-0.1083353162,0.0441836379,0.0161812063,0.0416294262,-0.2690986693,0.0045129284,0.1158902049,-0.3186094463,0.0445776843,0.2222859859,-0.0966080651,0.1381345391,-0.2381033301,0.0527907014,0.1070350036,-0.0908451676,0.4084692597,0.0276609473,-0.1147017553,0.1112391129,0.0733841583,0.087586619,0.1015325859,0.036568854,0.0174838658,-0.2870851755,0.258682698,0.254648,0.1308070123,-0.1768621057,0.0409475416,-0.2853093147,-0.2758462131,-0.2254685909,0.3028025627,-0.2556474209,-0.0894260854,0.0593868271,-0.0212555919,0.037580844,0.0237744153,0.5315715671,0.0317593068,0.0222246498,0.2021054924,0.3529034853,-0.0452097356,0.2537178397,-0.3451942503,-0.1072879136,-0.1020604819,-0.1388879716,0.1516920179,-0.1635595113,0.3526792228,-0.0160501283,-0.0066734427,0.0340723284,-0.0700819418,0.2124215215,-0.054955326,0.36746943,0.1745770276,0.0955137983,0.1014315113,0.0650428385,-0.0215570759,-0.1907879412,-0.1663581878,-0.2212535143,-0.2233618349,0.3332802355,0.032949239,0.0473898537,-0.0249781869,-0.0184603482,-0.0268135183,-0.0800871849,-0.1040141508,-0.0693608373,0.0553311259,-0.1027573794,0.3160770833,0.3010647893,-0.1802834868,-0.1319434643,0.2435503602,-0.4488347471,-0.0103973038,0.3142732084,-0.2056035697,-0.111313045,-0.1752173603,0.13655819,0.2521481216,-0.2705993652,-0.4501510561,-0.0099197626,-0.035526026,-0.1799046248,0.0339156836,-0.0731449649,-0.0439064838,0.2060275823,-0.0185976774,0.186592266,-0.0241721068,0.0725809783,-0.1159266904,-0.3167658448,-0.0236832853,0.1115791202,0.0598049164,0.0518154949,0.1092076078,-0.1558590531,0.3098129928,-0.0238866434,0.0374705158,0.1183782369,0.3224298358,0.2222387791,0.0256831422,-0.1601081043,-0.1528914869,0.0084992424,0.0363159813,0.2529557347,-0.188946858,-0.1422561705,-0.2147476822,0.0849788785,-0.0599863715,-0.2240587622,0.3443843722,0.0636983141,0.2576440275,0.2026868165,-0.2208631039,0.2093292624,-0.1175311655,0.2760332525,-0.2431588173,0.3099604845,-0.2670265138,-0.1470736116,0.3171859086,0.1844234467,0.0374808498,-0.1859867424,-0.0554432683,0.1859377027,-0.2177495807,0.0492481105,0.3298153579,0.0063328003,0.0005208623,-0.3167520761,0.0543908998,-0.0374392308,0.090480566,0.0325394385,0.003125516,0.1644380391,-0.0157709494,0.1092662364,0.0702299401,0.2263121456,0.601495862,0.0286186356,-0.047295671,-0.028145127,0.0594027489,-0.2560562789,0.1949958652,-0.2203732431,-0.2374462336,0.1203284413,0.3395717144,-0.0146109536,-0.0008410247,0.1500323862,-0.1850072742,-0.009140661,0.0106693637,0.2297254354,0.1747024953,0.2444141507,-0.0270585306,0.1174805686,-0.0241532177,-0.1709869653,0.1227389798,-0.0083993515,-0.0145723792,0.0663056672,0.0614345409,-0.0985732824,-0.4992950261,-0.1872356683,-0.3637148738,0.1166317463,-0.0261346642,-0.0418631881,-0.1803800315,-0.4079031646,0.0480167605,-0.3840751946,-0.4299280047,-0.3029088676,0.0251880474,0.1242448911,-0.0175693836,0.3162029684,0.0413392931,0.2474262416,0.1695605516,0.1161370799,-0.228589654,-0.1101185009,-0.0741423666,0.2087287158,0.0652748421,0.1763483286,0.4623808861,-0.2344217598,0.2084547728,-0.1831941903,-0.355735153,-0.0533320047,-0.0704678297,0.1732048094,0.3384903669,0.3902335465,0.184249565,-0.2252496481,0.3779099584,-0.3994681239,-0.1792486459,-0.1731818169,-0.0382149108,-0.2918351889,-0.1414560378,-0.0988975614,-0.2372414023,-0.3274066448,0.2493046969,0.0401149131,0.0115506016,-0.1116099954,0.2158566713,0.3016341031,-0.2477919459,0.0707853287,-0.2331909388,-0.462490648,0.1558843255,-0.2214986086,-0.540417552,0.2353655249,0.0842346251,0.3923362494,-0.3117666245,-0.3387042582,-0.3028455377,-0.1963200122,0.1462963074,0.0958551392,0.1788327545,0.3043561578,0.0308042746,-0.2045103312,-0.0966171697,-0.4155587852,0.0287843924,-0.2849397957,0.1875354648,-0.0860957131,0.4413440228,-0.0677208379,0.4245086312,0.1249109656,-0.0597864762,0.3453612924,-0.0630246177,0.3977819383,-0.1648960561,-0.3485438824,0.0861045048,0.0767505914,0.3037887216,0.1046007946,0.1423106939,0.1448972225,-0.2477249503,-0.2163062543,-0.2698568404,-0.1952343136,-0.205737263,0.2404360026,-0.0965181738,0.0306778792,0.0845896676,-0.1186807528,-0.0043348982,0.1029314026,0.437325269,-0.0649217442,0.1085842401,-0.4740851521,0.0840153918,-0.5896564722,0.2854155898,-0.1550413817,0.0890628025,-0.1627488434,-0.0675736666,0.0547079742,-0.207852304,0.7362417579,0.1252894551,-0.028404478,0.0332185291,-0.1700384915,-0.1345591396,-0.0253943056,0.0319995098,-0.0048048357,0.1599938571,0.2012716532,-0.1854706854,-0.211891681,0.1110846698,0.0323880538,-0.1088961437,-0.0182215348,-0.1991821826,-0.3933493197,-0.3056825697,-0.0035645745,-0.0864514783,0.2454596311,0.0300333183,-0.1688382775,-0.1668361723,-0.2721361816,0.0322674811,0.3463979065,0.1131633818,-0.0589518175,0.1658125073,0.1212047562,0.2196816802,0.2678529024,0.4225251973,0.0826835558,-0.3466485143,-0.0440437049,-0.0510351174,0.1864994317,0.1746421158,-0.1592644602,0.2024264485,0.0222158171,0.3755927384,-0.1855009049,0.0349181034,0.3832139373,0.2496268898,-0.0680847391,-0.121510379,0.1727585047,-0.1242199987,0.0730291754,0.2110857815,0.4789377451,-0.0610238761,0.2760619521,-0.1475715339,0.8998943567,-0.1372552514,-0.0165878106,0.1478447616,-0.1376612335,0.5555967093,0.0605298169,0.0488479771,-0.2951392531,-0.2339006066,0.04125949,-0.0968674198,0.2897569537,-0.2320279926,-0.0190889668,0.1129524931,0.083715193,0.0923237652,-0.1521187723,0.0765728131,-0.0143517023,-0.1194212213,-0.4534569681,0.3536142111,0.045751188,0.456519872,-0.0897771567,-0.2375231534,-0.101438418,-0.3424305022,-0.1264860928,0.1802719235,0.040610563,-0.0009572923,0.2184436172,-0.1141499579,-0.0372316577,0.2982881069,0.2889577746,0.1430098414,-0.3365095258,0.1038593948,-0.1538208574,-0.2299453616,-0.1514301002,0.2322813123,0.0788302049,-0.2506790161,-0.1948549896,0.1700565964,0.1475422531,-0.1174620464,-0.0755148903,-0.1373512,-0.2405716926,-0.3061335385,-0.0345093198,-0.075353168,0.0424312018,-0.2232241631,0.2642589211,0.159869805,-0.1834793389,-0.0325671099,0.2535817623,-0.031601917,-0.0951563418,0.4308345616,-0.1423818171,-0.0940936506,0.3149313629,0.1940956861,-0.2156932205,-0.28103441,-0.2892254889,0.323271662,-0.4338692129,0.0158047844,0.1344123185,-0.1910823733,0.0439144932,0.2045893669,0.1567863524,-0.1507425606,0.1555756778,-0.3843191266,-0.4521442354,0.2055763006,-0.074151963,0.2282615602,-0.0656726584,0.3630532026,0.0766687468,0.0090960292,-0.4710079432,0.0621886291,-0.276506424,-0.0875012353,-0.0731592923,-0.3039706349,0.3073917627,-0.0067864386,0.335299015,0.0033540623,-0.3884941936,-0.3612569571,-0.0294882394,0.0101547483,-0.0012780579,-0.130098328,0.14160119,-0.0726487339,-0.2279578894,-0.0489006415,0.1447808892,0.1235317886,0.007594584,-0.0926329643,0.1093209982,-0.0381786861,0.0258404277,0.0265945625,0.1238553897,0.1084834561,-0.1318737864,-0.1904687434,-0.2845087051,-0.0758663192,0.0703906938,0.046485927,0.0160569511,0.0439271443,0.3059985638,-0.0949720442,-0.1369228065,0.2580768466,0.6514681578,0.2214015722,-0.2058162689,0.2538484931,0.2273949087,0.3457742035,-0.5252274871,-0.0803516135,0.0098729869,0.0112815481,-0.0310088936,0.0063842623,-0.039665442,-0.0386931859,0.0358022973,0.1238021329,0.3864209652,-0.1507483125,0.1780587137,0.1790352911,-0.1085982025,-0.0367865376,0.3031566143,0.2047501802,0.1030755416,0.464964509,-0.2963241935,0.2282878458,-0.2592403591,0.1282867193,-0.0029636635,-0.2127178311,-0.0580143817,0.0361766852,0.0141403954,-0.1852599382,-0.1556016207,0.3336200416,-0.2587509453,-0.2459969819,-0.3449460864,0.2378118783,-0.1864787489,-0.203838706,0.113955237,-0.160468623,0.2044878304,-0.09022706,0.073280327,-0.1483730227,0.3094863594,-0.1234255582,-0.0318928882,-0.3718895614,-0.2998238206,0.0744455382,0.1174013168,0.0186332241,0.1538233906,0.1778643876,-0.1001720726,0.0480838306,0.4236602485,0.4834029377,0.1792708039,0.1607542783,0.1229665056,-0.1466842741,-0.1413237453,-0.0672002807,0.2507365346,0.2196327597,0.0681441128,0.2791547477,0.2924017608,-0.2768961191,0.081859678,-0.1501794308,0.263443172,-0.055801861,0.14018327,-0.3709548414,-0.0089941481,-0.357649982,0.0546993911,-0.3880121112,0.0321585014,0.0436079279,-0.1101503,0.0904063731,-0.3892801404,0.1873441637,-0.078604497,0.6130607128,0.2102432549,0.2450865805,0.0728555694,-0.2311115116,-0.4742298424,0.0834987238,-0.0369439013,0.06069877,-0.2378930598,-0.0285096411,0.016132066,0.0186930001,0.1845157892,-0.0226252209,0.054503601,-0.2129639536,-0.2606075108,-0.001710589,-0.213533029,-0.1165165082,0.1787106246,-0.1366167068,0.1701220423,-0.3507857323,0.2016436309,0.0883023441,0.375310421,-0.1649176031,-0.1537006348,0.1622485965,0.1201182455,0.2483728081,-0.0289951954,-0.0789699405,-0.2424039394,-0.4497918189,-0.1818383336,0.2393960804,0.1576492488,0.2583964765,-0.0405076109,-0.1400389522,-0.2154565006,0.1812776476,0.0935894996,0.2792016566,-0.1506288201,0.3175647557,0.0640001148,-0.0551515371,0.3476883173,0.0730157644,-0.0923939049,0.0800852031,-0.1955385953,-0.2940383255,0.5606746078,-0.253572464,-0.0164848473,-0.0570261441,0.0678327158,0.2090951055,-0.3254485726,-0.5532847643,0.3700197935,0.0775925741,0.1322588772,-0.1565454155,0.3490916789,-0.1839167923,-0.1236282438,-0.0239523239,0.3364186287,0.0933053419,-0.0758094192,0.1799940765,-0.0278393179]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1729","title":"Is there support for Deep learning datasets?","comments":"Hi @ZurMaD!\r\nThanks for your interest in \ud83e\udd17 `datasets`. Support for image datasets is at an early stage, with CIFAR-10 added in #1617 \r\nMNIST is also on the way: #1730 \r\n\r\nIf you feel like adding another image dataset, I would advise starting by reading the [ADD_NEW_DATASET.md](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md) guide. New datasets are always very much appreciated \ud83d\ude80\r\n","body":"I looked around this repository and looking the datasets I think that there's no support for images-datasets. Or am I missing something? For example to add a repo like this https:\/\/github.com\/DZPeru\/fish-datasets","comment_length":55,"text":"Is there support for Deep learning datasets?\nI looked around this repository and looking the datasets I think that there's no support for images-datasets. Or am I missing something? For example to add a repo like this https:\/\/github.com\/DZPeru\/fish-datasets\nHi @ZurMaD!\r\nThanks for your interest in \ud83e\udd17 `datasets`. Support for image datasets is at an early stage, with CIFAR-10 added in #1617 \r\nMNIST is also on the way: #1730 \r\n\r\nIf you feel like adding another image dataset, I would advise starting by reading the [ADD_NEW_DATASET.md](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md) guide. New datasets are always very much appreciated \ud83d\ude80\r\n","embeddings":[-0.2464449555,-0.0732379109,-0.2349893898,-0.0977336168,0.2804594338,0.0302638169,0.2250186205,0.0864046291,0.0842465088,0.2471515536,-0.0170774795,0.0281189773,-0.3046905696,0.3419502676,0.2932601869,-0.1668531895,0.3030381501,0.029771775,0.1124463677,-0.1883488297,-0.1011667475,-0.0620983243,-0.1791171879,-0.3200937212,-0.1909748465,-0.0778028518,-0.2675436139,-0.2924896181,-0.2347606122,-0.2387118638,0.3272100389,0.0372786708,0.2575370669,0.6088250875,-0.0001094966,0.0936006978,0.1888516992,0.0344999172,0.1902192682,-0.2489059865,-0.4143947959,-0.3501292765,-0.0992010087,-0.1630510539,-0.1417375803,-0.1659959406,0.3405531943,-0.0920914561,0.2136177272,0.1095001474,0.1909378469,0.0286527984,0.1526862234,-0.0124512631,0.1259145886,0.4637610614,-0.4192467928,0.0787267014,0.3635125458,-0.0840877965,0.3061642051,0.3910604715,-0.0415558629,-0.0781772956,0.0574290119,0.1204648539,0.0095714182,-0.6005347967,-0.0693713874,0.3455680311,0.4273528755,0.16948919,-0.4737809896,-0.2742058039,0.1135304719,-0.3056729138,-0.363910675,0.3375839591,0.0854369625,-0.0274839327,-0.5430440903,-0.550498426,-0.1716226637,0.3525954187,0.0071944292,0.1653153002,0.0343288891,0.0596226752,0.0275784601,-0.0464016572,-0.0350121148,-0.1218505725,-0.0354858451,0.1890714467,-0.3749040365,-0.3790487349,-0.3840422332,-0.0489446558,0.378639549,0.1953843385,0.0268187691,0.0727604032,-0.5005358458,0.198321715,0.2776097357,-0.0487883911,-0.526247859,0.23842673,0.4992908835,-0.10259673,0.2686316669,0.0182736013,-0.2126104385,0.0345907696,-0.0959782973,-0.3556536436,0.082033582,0.0555136763,0.2450485229,-0.3079968393,-0.2041275054,-0.1634489,-0.0028695005,-0.002901478,-0.0299655441,-0.0549219325,-0.0771602616,0.3101503849,0.0049461052,-0.2031582743,-0.0077231703,0.1446855962,-0.3078582287,0.1587817967,0.1561163962,-0.0348398201,0.1697469801,-0.357203275,0.2749171555,0.1219056994,0.1673989743,-0.1888694167,0.0184282605,0.4009966254,-0.1780541092,-0.3972993791,-0.0999811292,0.1693808138,-0.108338654,0.0679147467,-0.4045160413,0.072082065,-0.2208457291,0.1123012602,-0.0265457053,-0.310079217,-0.5300796628,0.4608233571,-0.1473255903,-0.1077617556,0.148238942,-0.051006414,-0.3330506384,-0.1678077281,0.3007892966,-0.0945542753,-0.1657098085,-0.1141156182,-0.3624179661,0.2976293266,-0.0095219007,0.1803171337,-0.1027760282,-0.1470511556,-0.0156495795,0.1419146657,0.4793444276,-0.1160755679,-0.1822009832,-0.097726211,0.3983662724,-0.3402421772,-0.2129779458,0.4864846766,0.2005528659,0.0731717274,-0.27010113,0.3275510967,-0.1403850019,-0.18258591,0.1647537947,-0.2831168771,-0.2380251735,0.4602395594,0.4331450462,-0.012694926,0.3475661874,-0.2191439718,-0.1625747234,-0.006864273,0.0831413418,-0.0199107304,0.662574172,0.1291887015,0.1037046164,-0.3360137641,-0.0976158008,0.2703981996,0.2004428953,0.4144092798,0.091187194,0.0119064162,0.27502653,0.202672556,-0.0872692689,-0.1572868228,0.0078640534,-0.1088456959,0.1064412743,0.0308883358,-0.353969723,0.0372102968,-0.0923408195,0.2923875153,-0.2196456492,0.6330333352,-0.1056521088,-0.0154705551,0.0786022842,0.196305722,-0.1879234016,-0.1049830243,0.2716382444,0.0510785058,-0.0152541846,-0.0681521595,0.3380039632,0.4709867537,0.5233405232,-0.609627068,0.1475699246,0.012997875,0.0229643155,0.0183824226,-0.223199755,0.2636472285,-0.0607782267,0.2634519637,-0.0805983469,0.0552256256,0.2424298376,-0.0176500455,-0.1863025874,0.0340176076,-0.1354718506,-0.2317546308,0.0777997449,-0.3133499324,-0.3291055858,-0.4214858711,-0.0325010531,-0.0755662173,0.1321131587,0.2652238905,-0.1466513723,0.1360957026,0.2810122669,0.4580815434,0.2789267302,0.2799292207,-0.1267132908,-0.2005002499,-0.0899405181,0.0111509459,-0.0068551637,0.2319177091,0.4414569736,-0.09220732,0.2556333244,0.0863187909,-0.2216324508,-0.1761960536,-0.0946418494,-0.0566557385,0.2319621593,-0.0603158511,-0.0935873911,-0.4871141613,0.1330734491,0.1424444616,-0.3022167385,-0.0720267668,0.0389214419,-0.2811459601,0.3109104633,0.2504755557,-0.1876751184,0.5262498856,-0.1141057462,0.1686607003,-0.3002625406,0.0219657943,0.0994812846,0.1673241258,0.1026539505,-0.3192421794,0.3941636682,0.4315930009,-0.0025457474,-0.2596161664,-0.6218697429,0.3442952335,-0.1961560547,0.1873702854,-0.0038358874,0.038285736,0.1127832979,0.31848225,0.1826595515,-0.2116437852,-0.1564266086,-0.3393561542,0.0348302945,-0.1265832186,-0.1262586266,0.2830400467,-0.4299687743,-0.019820502,0.1755001247,0.2003930062,-0.0757167786,0.3603032231,0.4419185817,-0.1968563944,0.0153541872,-0.1385998279,-0.014558116,-0.2842707038,0.4448160231,-0.1399222314,-0.2531705797,0.3671397865,-0.0729259104,0.082773231,0.2562768161,-0.6753509641,-0.1112541035,-0.207938835,0.0338670239,0.2140515745,0.0628792346,-0.0092199938,-0.2110792994,-0.0579751506,-0.0709017888,-0.2626710832,-0.1191503331,0.2377869338,0.3308423758,0.3337474167,0.3474175334,-0.0508259721,0.2955156863,0.0418461822,-0.3043087125,0.5080927014,0.170845598,0.3578542769,-0.0822675452,-0.0730764344,0.1549730152,-0.01399785,-0.1258409917,-0.1660049707,0.3011251688,-0.2082641572,-0.3810061812,0.1994307935,-0.2339512408,0.2077827752,-0.2149605453,-0.3136372566,0.3180966079,-0.2900024056,-0.2464130223,-0.0132356929,-0.0552353226,-0.2524007261,0.3302025497,0.399436146,-0.0875256732,-0.4125699103,-0.0641661063,-0.3355792165,0.1341335326,-0.1440152973,0.1293898374,-0.2495233566,0.5356863737,0.048614379,0.1078265235,0.4328628778,-0.4323401749,-0.4585812688,-0.2489062846,-0.284065783,0.2490467578,0.0137831224,0.0322509706,-0.0804286003,-0.1285142004,0.4602340758,0.1161310151,-0.0563445725,0.5523471236,-0.2404560298,-0.2236586213,-0.2206918597,0.0754097924,-0.0589251257,-0.1352549344,-0.2169573456,-0.0127181411,-0.2157991379,0.1938751936,-0.3250811994,-0.2172662467,-0.2820443511,0.0835776627,-0.1445715874,-0.0860546678,0.2058267593,0.0775972828,-0.0432379507,0.2464959323,0.4906102419,0.3285128772,-0.2036202997,-0.3507887125,0.1024306193,0.051309932,0.3266136944,-0.0016988264,0.1138846129,-0.0288494714,0.066454336,0.0831645131,-0.2991631329,0.0079621403,0.5065001249,0.2099130452,-0.3440674543,-0.411103636,0.3037992716,-0.0803539902,0.1457876116,0.0603419654,0.2356189191,-0.0985826254,0.1996971816,0.2182704657,0.7865880728,-0.3661304712,0.0217233002,-0.1699560881,-0.0276449081,0.6699262857,-0.1017897651,0.100491412,0.1861003786,-0.06974902,-0.1985665411,-0.0281916801,0.0019197032,0.0711047053,0.1636197269,0.2225092351,0.5093091726,0.0615024678,-0.0986593142,0.1390783936,0.2285276949,-0.1587290168,-0.1139915586,0.232211858,-0.117317304,-0.0147218313,-0.1691866666,-0.1279064119,-0.1667530984,0.3324788511,-0.4615353346,-0.1730368137,0.0875830501,0.1277988702,0.1244047955,-0.4775340855,0.3379827142,0.3541281521,0.3576821089,-0.0059586149,-0.2773002386,0.2542754412,-0.1821150035,-0.2067136317,-0.2546038926,-0.1713719517,0.1231673583,-0.0283395592,-0.2673963904,-0.0030286587,0.1565459967,-0.5180782676,-0.4821407795,0.0698830858,0.282066375,0.3480670452,0.035098806,0.1954478174,-0.0900082141,-0.0028620902,0.1746328175,0.2746788859,0.2125209272,0.1904276609,0.2809981108,0.0064610806,-0.2811050713,0.1441840529,0.1661398262,-0.0400572605,0.2079942971,0.0924574584,-0.0652017146,-0.1144262254,0.2407436669,0.3130511642,-0.0680657327,-0.1212749481,-0.1804793477,0.1700887829,0.0368520021,-0.0250448529,0.1844711453,-0.3268205523,-0.1332448423,-0.3180103302,0.1126062945,0.2135604173,-0.4756162167,0.4511298835,-0.1247861534,-0.0726483241,0.1455396861,0.0709099025,-0.2572884262,-0.0732545853,-0.1358036846,-0.1226661429,0.1898765415,0.1770219803,-0.0777919516,-0.28032884,0.0019590659,-0.1329588592,-0.1806692034,-0.1255865991,0.3036170304,0.1159869805,0.2453401238,-0.1540643275,-0.1141804233,-0.0983618125,-0.3798813224,-0.029168291,-0.0869643167,-0.0318116248,-0.0399777293,0.1403693706,0.2658537626,-0.0729502738,0.1383341104,0.2148325443,0.0143395448,0.4625805616,0.3712101281,-0.0650178865,0.4889165759,-0.0754884407,-0.1743277013,0.0902900696,-0.134933129,-0.2563120127,0.7665195465,-0.1000459567,0.2560340762,0.094625093,0.3729102612,-0.0319293104,-0.2949265242,0.3698741496,0.2496504039,0.165274024,-0.1704831421,-0.1848823279,0.0436836109,0.2250023484,-0.0009228729,0.1780534685,0.2021030039,0.3823187649,-0.1252901256,0.1218773276,0.4000947773,-0.1958667785,0.3291005492,-0.1569529176,0.1640087217,0.1983255595,0.0127402786,0.2278691232,0.0634799525,0.2766064405,-0.0325784683,0.5885313749,0.1290863156,0.0567678995,0.5713597536,-0.1318250746,0.0687382296,0.1127164662,-0.0274141226,0.0210784394,-0.1410956681,0.1185602024,-0.0948638991,-0.109277375,-0.1550277174,0.0704739392,0.1049080566,0.12729159,-0.5323839188,-0.0167651083,-0.1656365097,0.1219404414,-0.002403911,-0.2331485599,0.0134638082,0.097171478,-0.1814601123,0.0835476294,0.0296989325,0.2149153501,0.079379864,-0.003422094,0.2248190492,0.049023997,-0.2759844959,0.0976132676,0.489670068,0.1253914982,0.051601693,0.0371329784,0.0226301942,0.0143884905,-0.1855718046,0.1290099621,0.0461362377,-0.0168129839,-0.1357356012,-0.0582036451,0.1124752164,0.060541179,0.2560666203,-0.1724523455,0.1230979636,-0.3439575136,-0.1288021803,-0.169756338,-0.0752941146,-0.107288748,0.1784673035,-0.343359828,0.049137909,0.3856395483,-0.4856764972,-0.188718468,0.1320596039,0.0575545914,-0.4405553639,0.2636921108,0.4373852611,0.0780700222,-0.0032713274,-0.3652270734,-0.1010644287,-0.1627756059,-0.2862777114,0.0384077914,-0.0583501309,0.0766448453,0.2414325774,0.1381833553,0.1278271526,-0.3685331345,0.1556268334,0.0460254028,-0.254547298,-0.1693393439,-0.0619372278,0.2793893218,0.0035317368,0.2072129548,0.2506818771,-0.2790717185,-0.0912709385,0.0267616175,0.1174780652,-0.246588856,0.0616710372,0.3384707868,0.0565450117,0.1937884539,-0.1457109302,0.0391254649,0.0145709338,-0.2286934555,-0.0986883491,0.1057048067,-0.1536144316,0.2978627384,-0.2515074313,-0.177813828,0.0429250151,-0.1701200902,-0.1572858989,-0.2763397396,-0.0353284962,-0.1203954741,-0.0496347994,-0.0983002335,0.2921787202,-0.102278024,-0.0385119244,-0.3568793237,-0.0475923792,0.0385048501,0.1455025822,-0.2158138454,0.160553053,-0.1748198271,-0.3451704979,-0.0188897792,0.2482510209,-0.4402969778,-0.1968695968,0.2989643514,0.2130810767,-0.1564776897,0.4072247744,0.5054078698,-0.1011883691,0.0181343276,0.142723918,-0.1118072122,0.0107195694,-0.2021875829,-0.1147718281]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1728","title":"Add an entry to an arrow dataset","comments":"Hi @ameet-1997,\r\nI think what you are looking for is the `concatenate_datasets` function: https:\/\/huggingface.co\/docs\/datasets\/processing.html?highlight=concatenate#concatenate-several-datasets\r\n\r\nFor your use case, I would use the [`map` method](https:\/\/huggingface.co\/docs\/datasets\/processing.html?highlight=concatenate#processing-data-with-map) to transform the SQuAD sentences and the `concatenate` the original and mapped dataset.\r\n\r\nLet me know If this helps!","body":"Is it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n","comment_length":43,"text":"Add an entry to an arrow dataset\nIs it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n\nHi @ameet-1997,\r\nI think what you are looking for is the `concatenate_datasets` function: https:\/\/huggingface.co\/docs\/datasets\/processing.html?highlight=concatenate#concatenate-several-datasets\r\n\r\nFor your use case, I would use the [`map` method](https:\/\/huggingface.co\/docs\/datasets\/processing.html?highlight=concatenate#processing-data-with-map) to transform the SQuAD sentences and the `concatenate` the original and mapped dataset.\r\n\r\nLet me know If this helps!","embeddings":[0.1183504239,0.1227682754,-0.0548609123,-0.0218543373,0.1973935813,0.3098824024,0.170818463,-0.1640379131,-0.0837013274,-0.0505276099,0.2646122575,0.5609719157,-0.2010443658,0.0916972011,0.3351857066,-0.2962880433,0.1142137274,0.0437108278,-0.2922752798,-0.0279929936,0.0021185945,-0.1372906566,0.2007560581,-0.0317320488,-0.2767151892,-0.1182214543,-0.1379704624,-0.2123791128,-0.138488844,-0.5027133822,0.008782669,-0.0272243004,0.0590722486,0.3376502395,-0.000117042,0.0920339152,0.2727537155,0.1627476662,-0.5811193585,-0.3001923859,-0.3158358037,-0.5750210881,0.1763750762,-0.1296510398,0.1810124815,-0.340570271,0.1682196707,-0.516061008,0.3616964817,0.1440825015,0.1551565081,0.0048572132,0.3422509134,-0.1587884426,0.1316577643,0.2934283912,-0.1651616395,0.0660046414,0.0775927529,0.1870577186,0.210771367,0.1433064193,0.2321269512,-0.1839473099,0.5127686858,0.2322122306,-0.0962173641,-0.1791201383,-0.1925228685,0.1984659582,0.64759022,-0.415719837,-0.7800760269,-0.3988769352,0.5506299734,-0.16825293,-0.2445785254,0.2328118235,0.0239017699,0.1169849485,0.0696600899,-0.4716575444,-0.1921496391,0.2013756186,0.0445766747,0.2364700288,0.0233941972,0.0998460054,0.0355715565,-0.1198580414,0.2599288821,0.0791534111,-0.1234588847,0.2387579829,-0.309689939,0.0351709463,-0.2527564764,-0.1268173605,0.0663558617,-0.0297921486,0.4854209423,-0.1098813489,-0.3336729109,0.0790681094,0.2000619918,-0.0342736021,0.1430096328,0.0687184557,-0.3912691176,-0.3791761696,-0.1096804142,-0.2170026153,-0.1550947279,-0.1713347435,0.2424238622,-0.1673100144,0.1473714411,0.313683629,0.0629964322,-0.0935476422,-0.302016288,0.059125375,-0.2725579143,0.0612748526,0.4182109535,0.109153524,0.1732089669,0.4408244491,0.3726042807,-0.1566441506,0.0253166705,0.0513319708,-0.0543777086,0.3669013083,-0.2423496246,-0.0477180257,0.0371122248,-0.1206810251,0.1456056237,-0.1919995546,0.0828311518,-0.0404513627,0.2058393657,0.0076227454,-0.0446058661,0.1492536664,0.1150151417,-0.1696118563,-0.0430958346,0.2046017498,-0.0826668441,-0.0649525002,-0.4545913041,0.088045232,-0.0275905952,-0.3335460424,-0.0333478004,0.4336046576,-0.0151358461,0.0004981061,0.1414936632,0.009133257,-0.2703191638,-0.1619907767,-0.0206436161,0.1551439911,-0.512889266,-0.4108776152,-0.2430013418,0.0905044153,-0.5492098331,-0.0854313895,-0.1658569276,0.4677471817,0.005220708,0.047719121,0.5233315825,-0.1115105674,-0.167239517,-0.1299366355,-0.3428252637,0.0833971202,-0.0064481604,0.4439302087,0.1647202224,0.2340108603,0.0283424053,0.2201442719,0.0826974437,-0.3359245062,0.278637141,0.0015661336,0.2065098733,-0.1058211625,-0.3012523949,-0.0254255347,0.1261385232,-0.1821459681,0.1216382831,-0.2194676548,0.2375243455,0.306014955,0.16865848,0.2052364349,-0.1291643679,-0.1226723939,-0.2524817586,-0.0185622871,0.1810260266,0.0529476367,-0.0552381873,-0.2419681251,-0.0322534144,0.3225970864,-0.0610444359,0.0148529364,0.0318089277,0.0758635402,-0.0819452256,-0.2782067358,-0.3902855515,0.4074019194,-0.0087648733,0.2047669888,-0.2296218425,0.1660542935,0.2363984734,0.0036824329,-0.3240821064,0.2122142911,-0.1640319079,0.0895017013,0.0789454207,0.1134709269,-0.1195055768,0.2341029197,0.1487384588,0.0327494219,0.1080177873,-0.3662585318,0.0763284415,-0.0375731699,0.0954278409,-0.0913993344,-0.3540546596,0.2417161614,0.1802269071,0.358039856,-0.1495783925,0.2765904665,0.1072116494,-0.0300822426,-0.4279772937,-0.2050343454,-0.1618532091,0.0159433894,0.279984206,-0.1748908162,-0.1389584094,0.0067604636,-0.0259438623,-0.1193043813,0.1013053954,0.1819251925,-0.5684722662,0.289631784,-0.0735917762,0.3036424816,0.3615439832,0.2742063403,-0.0376004875,0.1356413215,0.1420550942,0.0517570712,0.0522803739,0.1145410091,0.4417301118,0.0077748694,0.4800372124,0.1265679449,-0.0580536574,-0.0096648177,-0.0327755064,-0.0302462541,-0.3340159059,-0.0168455876,-0.0518505089,-0.1416765749,-0.5122766495,-0.1967370212,-0.0311987214,-0.3337020576,-0.053238906,-0.2285900712,-0.0717634708,0.1076778322,0.1049499661,0.2497825176,-0.076599963,-0.0839089006,0.1579049975,-0.1696772426,0.258208096,-0.0121931871,-0.0431305915,0.0414119065,0.3619517684,0.0993797183,0.1087101474,-0.3433416784,-0.1797894388,0.0348750688,-0.0459054634,0.1033415049,0.0131804319,0.0726373568,-0.1497794539,-0.2909355164,0.081686914,0.023718087,-0.1779968739,-0.1926491857,-0.0532523021,-0.0540067218,0.045900017,-0.0440836027,-0.5674846768,-0.1944701672,0.4599407911,0.1678510457,-0.0582460165,0.1613591909,0.2073466778,0.2443645895,-0.0269283876,-0.0540948398,0.0441003218,-0.1064452827,0.4240250885,-0.1946118176,-0.1915540844,-0.030255856,-0.3298680484,0.0344480462,0.0987454131,-0.2150246352,0.0016386468,-0.1454131454,0.2108080983,-0.1521884203,0.1087907255,0.3900870383,0.049431283,-0.0747972503,-0.1380809098,-0.4104962945,0.1876413077,0.2697172761,0.1381595582,-0.0397371389,0.263351649,0.0126135089,0.3805563748,0.1919939965,-0.304358542,0.0460212901,-0.3049993813,0.2573345602,-0.0679783672,-0.3093288839,-0.2927863598,-0.3105886281,-0.1926646233,-0.2148533016,0.142957896,0.0314567238,0.1148939133,0.05997926,-0.2940514684,-0.1667509675,0.2652803957,-0.5815134645,0.2634644508,-0.1513064653,-0.2137211561,-0.4562422633,0.0570243485,0.0172461346,0.0119302254,0.2096430659,-0.1302230656,-0.5189207792,-0.1052691638,-0.1825606376,0.1870447844,0.2848891318,0.0409294963,0.2111172676,-0.1281001419,-0.054487206,0.1286325902,0.4015107453,-0.2101173699,-0.5021834373,0.3108868301,0.1300798953,-0.5276559591,0.1357315779,0.0001150665,0.3417448401,0.0279724468,0.1077675596,-0.2061166167,0.3304412365,0.2148548216,0.2868593335,-0.2674980462,-0.1786042899,0.0959056541,0.0080993436,-0.3620217144,-0.2194253653,0.033234857,0.0782863945,0.4012641907,-0.1648359895,-0.3028608263,0.0789414421,0.1901785582,-0.1019092202,0.16427508,0.3630395532,0.0984761119,0.2240504324,0.1412564218,0.0841497332,0.579823494,-0.0055795987,-0.5680826902,-0.3137509525,-0.2384532541,0.6173958778,0.1011935323,0.3702192008,0.5125343204,0.0721113309,-0.1607180834,-0.0868123844,0.0316265337,0.1669401973,0.1239069328,-0.4113565385,-0.4321609735,0.5005094409,0.0362601727,-0.0168649536,0.2640680373,0.2928214371,-0.4001209736,0.4860458672,0.1219498217,0.9126859307,-0.1145880893,0.3797572851,0.0232584868,-0.3634345233,0.6220881939,0.1060946584,-0.1080737412,-0.1013952047,-0.1234064177,-0.0526012331,-0.2997168601,0.0754948258,0.1275458336,-0.4212502837,-0.1573472172,-0.3843897283,0.0647300184,0.0013178341,0.2870388627,0.0207655448,-0.344722122,-0.2162036598,0.1235494837,0.2255604714,-0.1749017537,-0.1564715058,-0.2046266794,-0.3168256283,0.0599343106,-0.6070039868,-0.2896137834,-0.1706378311,-0.1501219422,0.2276558876,-0.5194820762,0.298553586,0.4806924164,0.242214486,-0.1505767703,0.0100359013,0.2047694921,-0.0675514415,0.2147722691,0.1130865589,-0.3405281305,0.2628997266,0.2204297632,-0.2200184464,0.3842211962,0.1054163873,-0.5912263393,0.2402300835,0.0444692858,0.0296441354,-0.2897059321,-0.275745362,0.4203806221,0.022381451,-0.1884147078,0.0896284506,0.1498454213,-0.1582795531,-0.0848680586,0.3082499504,-0.2128809243,0.1340120733,0.0668442547,-0.0161557235,0.0898440182,0.6236746311,0.1486747265,0.0013381779,-0.0669525638,0.185101971,0.3109865785,-0.3935835958,-0.0033083835,0.1122306809,0.0195170045,-0.3946285844,-0.1075062677,0.1711089313,0.2316144705,-0.3769903481,-0.2538278401,-0.3830397129,0.3205639124,-0.057648398,0.2032012939,-0.2092088312,0.2275728881,-0.0503256992,0.4206245542,-0.2381623387,-0.0382595807,0.1344927847,0.012355363,0.2349124849,0.1758813411,-0.1389648169,-0.2987452745,0.0955364257,0.1382970512,0.3021369874,-0.1650759429,-0.2145964652,0.1436089873,0.1084745079,-0.0549199209,-0.0708807483,-0.1440371871,0.3611432314,-0.3057288527,-0.2458442152,0.2038039863,-0.0754249543,0.1035798118,0.4034571946,0.2424578667,-0.1424088627,0.0249776803,0.2437241822,0.19488585,0.0925836042,0.2005091161,0.16766949,0.1246202439,-0.1107254326,0.3039937019,0.4095856845,0.0034355263,0.0307547804,-0.214852035,0.1017245352,-0.0100514805,0.2032630444,0.1487461478,-0.3235373497,0.4601846933,0.0889041275,0.183409825,-0.2215225548,0.0492056236,0.2862812281,-0.157129094,-0.0184970535,0.1438932419,-0.0952477753,0.3241392374,-0.3016679883,-0.0336547345,-0.28221187,0.067814216,0.201326102,0.1810992658,0.2624467015,-0.0867181942,0.1089945212,0.1661396325,0.2174193561,0.1549303085,-0.1797844768,0.232144773,0.4300880134,-0.0374771878,0.2956375182,0.2165901959,0.2032027245,0.0660922825,0.351562053,0.0029344738,0.3370348811,0.4048221409,0.3064149022,-0.2324177474,0.0884142146,0.2514267266,0.1240265965,-0.2480943799,-0.0932433158,0.0365998037,-0.0284840111,-0.1121460274,0.0026526377,-0.5771657825,0.1557986438,-0.0457028262,0.1306901127,-0.3807972968,0.3548991978,0.2698544264,0.082900539,-0.1574144363,-0.0642105937,-0.1329989433,0.2097125202,0.2874039412,0.4605228901,0.1455061287,0.0901300013,-0.0937163681,0.6465041637,-0.2929738462,-0.1388495415,0.014124617,-0.1115552858,-0.0682274029,-0.1355383247,0.2433322668,0.1671343297,-0.0767860338,0.1482755095,0.0829102471,0.294255197,-0.5951100588,0.1925399005,-0.0541328117,-0.1810987294,-0.1505552679,0.2217446268,-0.1482458711,0.0335925967,0.5801295638,-0.2220921069,-0.029115634,0.1627810746,0.0306218285,0.0370203927,0.315616399,0.4071694016,0.5021685362,-0.0857448205,-0.1438606381,-0.5734360218,-0.065940991,-0.0792130977,-0.1524409205,-0.0374573916,0.0093771955,0.3101043403,0.3368249834,0.3097115159,-0.1413752437,-0.3536221385,0.171035111,-0.1950477809,-0.342509836,0.3337602317,0.3418866098,0.0802903473,0.1111700162,-0.2285287976,0.1602797359,0.0043956591,-0.0351680331,-0.0819215477,-0.1506333351,-0.0712445974,0.5276620984,-0.1153399199,0.3695677519,-0.0902272984,0.0822120011,-0.1922505647,-0.451382786,-0.4487344325,0.4448750317,-0.0723153576,0.3471491933,-0.3197867572,0.0114935068,0.1620357633,-0.2662712932,-0.079500787,-0.142075032,-0.4082279205,0.3657561541,-0.1145641357,-0.0250445735,-0.1423920393,0.1617483348,0.017603036,0.2556729913,0.0724254325,-0.15030922,0.2750244737,-0.3098266423,-0.1333382279,0.2292268574,-0.0305895116,0.1441351622,-0.1657263041,-0.7632929683,-0.1766393185,0.3928314149,-0.0794504061,0.0721662045,0.1734795719,0.0661078915,0.0923299789,-0.1821919233,0.5891467929,-0.1789297312,0.0542211533,-0.4637468159,-0.2602521777]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1728","title":"Add an entry to an arrow dataset","comments":"That's a great idea! Thank you so much!\r\n\r\nWhen I try that solution, I get the following error when I try to concatenate `datasets` and `modified_dataset`. I have also attached the output I get when I print out those two variables. Am I missing something?\r\n\r\nCode:\r\n``` python\r\ncombined_dataset = concatenate_datasets([datasets, modified_dataset])\r\n```\r\n\r\nError:\r\n```\r\nAttributeError: 'DatasetDict' object has no attribute 'features'\r\n```\r\n\r\nOutput:\r\n```\r\n(Pdb) datasets\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['attention_mask', 'input_ids', 'special_tokens_mask'],\r\n num_rows: 493\r\n })\r\n})\r\n(Pdb) modified_dataset\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['attention_mask', 'input_ids', 'special_tokens_mask'],\r\n num_rows: 493\r\n })\r\n})\r\n```\r\n\r\nThe error is stemming from the fact that the attribute `datasets.features` does not exist. Would it not be possible to use `concatenate_datasets` in such a case? Is there an alternate solution?","body":"Is it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n","comment_length":123,"text":"Add an entry to an arrow dataset\nIs it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n\nThat's a great idea! Thank you so much!\r\n\r\nWhen I try that solution, I get the following error when I try to concatenate `datasets` and `modified_dataset`. I have also attached the output I get when I print out those two variables. Am I missing something?\r\n\r\nCode:\r\n``` python\r\ncombined_dataset = concatenate_datasets([datasets, modified_dataset])\r\n```\r\n\r\nError:\r\n```\r\nAttributeError: 'DatasetDict' object has no attribute 'features'\r\n```\r\n\r\nOutput:\r\n```\r\n(Pdb) datasets\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['attention_mask', 'input_ids', 'special_tokens_mask'],\r\n num_rows: 493\r\n })\r\n})\r\n(Pdb) modified_dataset\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['attention_mask', 'input_ids', 'special_tokens_mask'],\r\n num_rows: 493\r\n })\r\n})\r\n```\r\n\r\nThe error is stemming from the fact that the attribute `datasets.features` does not exist. Would it not be possible to use `concatenate_datasets` in such a case? Is there an alternate solution?","embeddings":[0.0949398354,0.2781065404,-0.0392040946,0.0230574142,0.2436094135,0.377887696,0.2803983092,-0.1585282087,-0.2577433586,-0.0745178536,0.3132960498,0.5345349312,-0.1596789658,0.0007422063,0.2160830647,-0.2442814261,0.1367064565,0.1000794992,-0.1950359046,0.0196850672,-0.0471322238,-0.147789523,0.0882382169,-0.0374560989,-0.3350251019,-0.0334434696,-0.1999395043,-0.150008589,-0.1133730188,-0.4348908663,-0.0019859537,-0.0370974801,0.0035896413,0.3410022855,-0.0001201627,0.1059936211,0.2928904295,0.1239432693,-0.5508271456,-0.197150439,-0.3457706273,-0.5728435516,0.1122911274,-0.1652757376,0.2542717159,-0.5107655525,0.0592758618,-0.5218868852,0.1629167795,0.2337290794,0.1391426623,0.0411460921,0.4349201918,-0.1688048542,0.1174908951,0.2363585532,-0.1814585477,0.0147102326,-0.017161198,0.1729206741,0.3916795552,0.0469979793,0.2433085591,-0.226460889,0.4740366638,0.2806395292,-0.0606916212,-0.1978556514,-0.0475819409,0.1038512513,0.6565675735,-0.4195333123,-0.7511697412,-0.3529868424,0.607022047,-0.1742109507,-0.1856439859,0.1261354387,0.0625533611,0.0905173272,0.0979989022,-0.524441421,-0.1920197457,0.1756156236,0.0487164557,0.2053502351,0.0891902447,0.1311822236,-0.085890241,-0.1179474518,0.3495790362,-0.0492359847,-0.1029369533,0.1775587648,-0.2410248965,0.0938502774,-0.2655631006,-0.3777485192,-0.0426801331,-0.0045626462,0.4142281413,-0.1584382951,-0.1647912413,0.1298494637,0.2361717075,0.0060687154,0.0351451933,0.2194169313,-0.4275453985,-0.4441646636,-0.0518340208,-0.1291937083,-0.2086056918,-0.2535239458,0.299210161,-0.1239159852,0.2640658617,0.312490344,0.0177036822,-0.0637345761,-0.3315076232,0.0947255418,-0.1798247397,0.0061582313,0.3212430775,0.0917390585,0.1746656746,0.3785211444,0.4458310902,-0.0889534429,0.0020019137,0.0982939675,-0.0852296799,0.1905242056,-0.2227180898,-0.0200647805,-0.0167526156,-0.1474682242,0.1622805297,-0.1375227273,0.0615988709,-0.0430382676,0.1672564,0.0139296632,0.0231758077,0.1048490703,0.0947606191,-0.2394231856,-0.0634574741,0.2614063621,-0.0971488655,-0.1494695097,-0.5945218205,0.095838055,-0.0279962942,-0.2157822102,-0.04640048,0.3512548804,0.1539960653,-0.0271742549,0.0557841212,-0.0656976476,-0.1824213564,-0.1591274291,0.0550409444,0.1193372458,-0.4866577089,-0.3338994384,-0.1635727882,0.148028791,-0.4304989874,-0.1139238328,-0.1592982709,0.4569547772,-0.0372045673,-0.085643366,0.6153522134,-0.0849262625,-0.1500373632,-0.14382644,-0.268727392,0.0542884022,0.0565822311,0.3473592401,0.0686083585,0.1872704029,0.1601715982,0.1712306142,0.0413600057,-0.4098122418,0.2447949052,-0.0735128671,0.3654520512,-0.1491919607,-0.1708804071,0.0740852505,0.1003545225,-0.1355796456,0.0207677837,-0.2471004277,0.2322204262,0.3281669617,0.2038031369,0.1248062477,-0.1473128796,-0.0609295592,-0.2525945008,0.0210459102,0.296795249,0.0060208403,-0.0078766178,-0.2195046544,-0.0810929015,0.3016640544,-0.0502284095,0.0538662747,0.0400150456,0.1648097187,-0.1211740598,-0.3005161881,-0.417498976,0.4620872438,-0.0052939602,0.138076514,-0.2055999786,0.179153502,0.101815477,-0.0454691276,-0.2537805438,0.2562515736,-0.0762678459,0.1740576327,0.0129561126,0.0900579691,-0.2205075771,0.1277780533,0.1649122387,-0.0100810695,0.0791926607,-0.2740038633,0.0128129758,-0.011099346,0.1449087113,-0.1138377786,-0.2944462001,0.1484531611,0.1360001266,0.4129469097,-0.0269872714,0.2385406345,0.1496558785,-0.0078585558,-0.4193473756,-0.2293293476,-0.2127715796,0.0128889373,0.3095213473,-0.1552267075,-0.1388232708,0.0976166502,-0.0091856793,-0.0095406324,0.0620600805,0.1718090177,-0.5074893236,0.2328690886,-0.0809808671,0.2883297205,0.4779847264,0.262775898,-0.0419580713,0.205390811,0.0421055667,0.0893631727,0.0840999261,0.1871212721,0.3860476911,0.1431772411,0.4335136712,0.1157339886,0.0573308244,-0.1672387421,0.062663883,0.0576609597,-0.3890411258,-0.0108102709,-0.0527421124,-0.0096305991,-0.3886736035,-0.2951336205,0.0230066255,-0.3966552317,-0.0737865195,-0.1356207132,-0.0671552122,0.0733814016,-0.087303184,0.1971832663,0.0449831039,-0.1264404804,0.1811292171,-0.081291005,0.1692827642,-0.0278002322,-0.0546722002,-0.0206074547,0.2912627757,0.1787186861,0.1999196857,-0.2610251904,-0.166877225,0.0133868167,0.0031255102,0.1212544516,0.0114523582,-0.0469485819,-0.108213529,-0.4225558639,0.054674644,0.1960376054,-0.1094812378,-0.099623166,0.024214182,-0.2206699252,0.0593017489,-0.1258939952,-0.5891417265,-0.2157818824,0.2873737812,0.174544692,-0.0803018361,0.2116658837,0.3083162904,0.3107210696,0.1965717971,-0.0475443266,-0.0099163782,-0.0303186551,0.5247279406,-0.1951566786,-0.2251553386,0.0467643179,-0.4135626256,0.1101681888,0.1235316694,-0.1574349552,0.1387554854,-0.0580013208,0.176375553,-0.2004720867,0.0811580941,0.3056042194,0.1505975723,-0.084144868,-0.0847527161,-0.4099744558,0.1357345283,0.3004318774,0.227889806,-0.1091424003,0.1362154335,-0.0673065856,0.4504676163,0.106894657,-0.2593741715,0.04322071,-0.2751227021,0.3826049566,-0.0402342752,-0.4448498189,-0.3242741525,-0.3273374438,-0.2452388853,-0.2040431947,0.0792932585,0.0237939954,0.0419279076,0.1391580403,-0.1569458693,-0.1752624512,0.2127830982,-0.6013404131,0.2956725955,-0.1307006925,-0.2297112644,-0.4818823934,0.0993064344,-0.0449011475,0.076996021,0.0479558595,-0.1311235428,-0.5606611371,-0.1555871069,-0.188018918,0.1097496301,0.2204019129,0.265927732,0.2145656645,-0.0978004634,-0.147044763,0.1006336212,0.384247303,-0.2125471532,-0.5697062612,0.4248259962,0.2341233939,-0.4442066252,0.0142440861,0.0099963136,0.3804396093,-0.0114084361,0.1019915342,-0.1336503178,0.3616288602,0.2096711099,0.2413605601,-0.3181265593,-0.1124855727,0.0307473633,-0.0567151867,-0.3592161536,-0.3129605651,0.0205045547,0.1432089806,0.3045032322,-0.1618649513,-0.4555437863,-0.0219117757,0.2387917191,-0.1175484657,0.1498011798,0.3454281986,0.1717688739,0.0334147513,0.1279961318,0.1813500524,0.5413866639,0.1100350916,-0.589615643,-0.2115219831,-0.3717561066,0.6812923551,0.0252607055,0.1749813855,0.4369099438,-0.0457267836,-0.1110180989,0.0084868213,0.0501815639,0.2271531075,0.0482963063,-0.4718443155,-0.4662940204,0.3843891919,0.0067953807,0.0540866032,0.2895106077,0.2082000375,-0.4573581815,0.4454078972,0.076407142,0.9311355352,0.0458281562,0.3057080805,0.0718974397,-0.3916465044,0.5683711767,0.1181867123,-0.0761592016,-0.1081329584,-0.259432584,-0.0413507037,-0.2849692404,0.1493323147,0.1107188687,-0.47170645,-0.1215753853,-0.4297173321,0.0861568302,0.0397169366,0.3061490953,0.1586368531,-0.2875172794,-0.209175393,0.0849498957,0.2206139416,-0.2384736389,-0.0867979527,-0.170708999,-0.481887728,0.0509659909,-0.5248125792,-0.1866619587,-0.1544187367,-0.0219025686,0.1781320274,-0.5238455534,0.3283817768,0.3898057044,0.2787604928,-0.1245286688,-0.0041835662,0.1162559018,-0.035959512,0.1550441086,0.0513915271,-0.3494859338,0.267832756,0.1964034438,-0.2033124864,0.4784384668,0.0678569078,-0.5774133205,0.2164763361,0.0650452971,0.130557999,-0.3340484202,-0.3530471325,0.4331821203,0.0111246863,-0.2080414742,0.0632218868,0.0332115702,-0.2652063072,-0.0405545309,0.244328931,-0.1819822341,0.1550503522,0.0502986759,0.0892612934,0.1533144265,0.6899826527,0.0169990063,0.0386291146,0.0280532967,0.1391943097,0.2770433128,-0.3112417758,-0.0852896273,0.0653472096,-0.0030739233,-0.380889535,-0.132685408,0.1694084257,0.3222751617,-0.3416877687,-0.2842265666,-0.430342406,0.2670712173,0.0678488314,0.2315497994,-0.2564610243,0.1963721067,-0.0829160959,0.3826929331,-0.205866158,-0.1467097104,0.0724110156,0.0601887666,0.226130724,0.1346983463,-0.173304975,-0.2768847942,0.0939946547,0.0403851531,0.3214462399,-0.1358165741,-0.1624219269,0.1488988101,0.2249884158,-0.0464235768,-0.1000562161,-0.1643284112,0.2909638882,-0.2527961433,-0.312428385,0.2480216026,-0.0747717023,0.109594591,0.2240988165,0.2588258684,-0.1933639944,0.0363575034,0.259811461,0.2524392903,0.1345205605,0.2513072491,0.1465764642,0.1558569372,-0.2523503006,0.2744453251,0.1775446534,0.0416721553,0.0549325198,-0.290753603,0.1783407927,-0.097404249,0.0951529518,0.2716449499,-0.359631598,0.4081236422,0.0837373883,0.1162089482,-0.2869894207,0.0472663641,0.2409876585,-0.1543050557,-0.0344919339,0.1801659614,-0.093560271,0.221919924,-0.4223341644,0.0280247703,-0.2977149785,0.1174357086,0.2466383427,0.2686145604,0.2032658607,0.065966256,0.0549957678,0.2347799093,0.1374090761,0.2698480487,-0.0922985375,0.2126544267,0.634062469,-0.088708818,0.322445184,0.2316927165,0.2331198603,-0.0167096481,0.4093011916,-0.0334461145,0.352548629,0.2604314387,0.2266249657,-0.1808939129,0.1143442467,0.1710251421,0.0969557613,-0.1943643391,0.0699338913,0.0238582939,-0.1063879058,-0.1568580866,0.0902361646,-0.5656858683,0.0634934977,-0.1466249228,0.0556055717,-0.2716339827,0.5452595353,0.3046424687,-0.0458474196,-0.1410917491,-0.0680309087,-0.0762583837,0.1992262602,0.3087108433,0.4064922929,0.2343642414,0.2340598106,-0.0894151554,0.4733196199,-0.1198834851,-0.2696188986,0.0461901873,-0.0820768103,-0.0833963752,-0.1651276648,0.2309200317,0.1366723776,-0.0681317374,0.1952935457,0.1487997174,0.2619003654,-0.5672976375,0.3657277822,-0.026823299,-0.2480604351,-0.0181576684,0.2150847316,-0.1077647731,-0.2042974681,0.6744141579,-0.1407830119,0.0337296836,0.1602958441,0.0252887197,0.0353338867,0.3453664184,0.3773519397,0.4509749115,-0.0888511688,-0.0016574096,-0.5564728379,-0.0392563716,-0.0373838879,-0.2161786407,-0.0139367403,0.0744098797,0.207366392,0.2814114392,0.3518777192,-0.0497970842,-0.2626155615,0.1464737207,-0.2191754431,-0.4201034009,0.2965903282,0.3462172151,0.1036058813,-0.0007780138,-0.1439506412,0.198411867,0.000938962,-0.03382973,-0.1928084791,-0.0611810759,0.0750955939,0.4989222288,-0.1543782353,0.4688833356,-0.0769367218,0.0583548322,-0.1721830666,-0.4284749925,-0.4607102871,0.5446799994,0.04581514,0.218941167,-0.3439549804,0.0508864298,0.1071242467,-0.2285876423,-0.0111197783,-0.1229112297,-0.4091597795,0.4311248362,-0.0631328523,0.0971574858,-0.132509768,0.0272268616,-0.0314446539,0.3439383209,0.06241237,-0.1612472981,0.3624554873,-0.1864136606,-0.2118339539,0.202044189,-0.0647424906,0.2963013053,-0.0687342584,-0.7376129031,-0.2276840657,0.4214466214,-0.1764869094,0.0461305082,0.1430074871,-0.1273700595,0.131813094,-0.120021522,0.5443986654,-0.3239011168,-0.020789938,-0.4493511021,-0.1988045424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1728","title":"Add an entry to an arrow dataset","comments":"You should do `combined_dataset = concatenate_datasets([datasets['train'], modified_dataset['train']])`\r\n\r\nDidn't we talk about returning a Dataset instead of a DatasetDict with load_dataset and no split provided @lhoestq? Not sure it's the way to go but I'm wondering if it's not simpler for some use-cases.","body":"Is it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n","comment_length":42,"text":"Add an entry to an arrow dataset\nIs it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n\nYou should do `combined_dataset = concatenate_datasets([datasets['train'], modified_dataset['train']])`\r\n\r\nDidn't we talk about returning a Dataset instead of a DatasetDict with load_dataset and no split provided @lhoestq? Not sure it's the way to go but I'm wondering if it's not simpler for some use-cases.","embeddings":[0.1145721227,0.2276223749,-0.0684058368,-0.0251238011,0.1537322551,0.2710112035,0.226799041,-0.1038071588,0.0163868833,-0.0527591147,0.3032632768,0.5584480166,-0.2891287804,0.047076717,0.3657113016,-0.280523628,0.0873283595,0.0396179557,-0.2587966621,-0.0790688768,-0.0180866886,-0.2019383013,0.2285862714,-0.0326016434,-0.228646785,-0.1232241467,-0.0993934721,-0.262957871,-0.1477606595,-0.4609306753,0.0618369095,-0.0937316194,0.0361965336,0.251676023,-0.0001185569,0.1118827239,0.2736075521,0.1426105648,-0.5661526322,-0.2433639467,-0.259247005,-0.5845739841,0.2074297965,-0.1865363866,0.2207546234,-0.3495488167,0.1405095756,-0.5783141255,0.2560048699,0.2211306393,0.1515809,-0.0092613129,0.2489777058,-0.1223886013,0.2697330415,0.2970865071,-0.1688393652,0.0161133036,0.0952636451,0.2036064416,0.1866354793,0.0433867425,0.1943320781,-0.1995443255,0.4199509025,0.2216569334,-0.11024376,-0.1197424829,-0.120614931,0.1585541219,0.7166870832,-0.3376507759,-0.7637710571,-0.3833065331,0.4716689289,-0.2002213597,-0.2719372511,0.2291770875,0.0322279632,0.100159511,0.0578414537,-0.5695364475,-0.1689431518,0.1344406754,-0.0713209584,0.2662101686,0.1282932311,0.1429059654,-0.0294441059,0.0496059395,0.4449470341,0.0718639791,-0.0850041285,0.2023862749,-0.275421679,-0.0111991679,-0.2009694129,-0.1555013657,0.022890646,-0.0264489166,0.4322582483,-0.0977271944,-0.349150151,0.0062519102,0.2315907329,-0.0902674645,0.0625158623,0.1639588773,-0.3666475415,-0.5172889829,-0.0959439203,-0.2136398405,-0.1333695799,-0.108360745,0.274302274,-0.1353407055,0.1114093438,0.2889271677,0.0096789235,-0.1018484384,-0.34245345,0.010887946,-0.2589613795,-0.0176445786,0.4288860261,0.1240437105,0.1030766889,0.4695477188,0.3152875304,-0.2198577672,0.0048457375,0.0921725333,-0.0695167407,0.2748762071,-0.1877287477,0.0636667535,-0.053088326,-0.2365311384,0.0488272458,-0.1673048288,0.1473578811,-0.0812282339,0.1897466034,0.0458911844,-0.0276269838,0.1726948172,0.1745205075,-0.1622058004,-0.1053395048,0.2671445906,-0.0765427202,-0.2106023729,-0.3816097677,0.0947101936,-0.0319601744,-0.3137810528,-0.0676094815,0.3853579462,0.0281780735,0.0744016841,0.141911298,-0.1503307819,-0.2972721457,-0.1496002823,-0.0254141912,0.0646295771,-0.6223896146,-0.330209285,-0.3623220026,0.0471313372,-0.4713605046,-0.0428867377,-0.1817537397,0.6620752811,-0.0138257304,-0.0914267153,0.5936508775,-0.0985190496,-0.2818264663,-0.192326799,-0.29875049,0.0530211516,0.0030412513,0.5245484114,0.2708130181,0.2221010625,0.0133398613,0.1885728538,0.0027442609,-0.2894113958,0.2583531141,-0.0393458456,0.2655002177,-0.0190553162,-0.1200523749,0.0276346356,0.2107279897,-0.0663183555,0.1436348855,-0.2056606561,0.2251610458,0.0847937167,0.2002291232,0.1562264711,-0.1784159839,-0.1116406843,-0.2688981593,0.0021332121,0.2222335339,0.0003280949,0.0635620952,-0.2276721299,-0.0282906704,0.3634390831,0.0011588238,0.0436765365,0.0339295343,-0.01935184,-0.0979097486,-0.2692063749,-0.3994077742,0.4055222869,-0.0663043708,0.1545792073,-0.2101407498,0.2363847941,0.2581991553,0.0021447032,-0.3161668181,0.2485579401,-0.2150732726,0.072397843,0.091633305,0.003637125,-0.042057164,0.2055780143,0.1146879494,0.0066276598,0.065740779,-0.3981017172,0.2017666698,0.0086334189,0.0744551942,-0.1684645116,-0.3035768569,0.2037469894,0.0876132548,0.3461208045,-0.1305732578,0.2628031671,0.0295077022,0.0066060149,-0.4481200874,-0.169487536,-0.1674040854,-0.0128956717,0.3151218295,-0.2133638263,-0.0394492522,-0.0212720353,0.0690554455,-0.1229185611,0.164433077,0.1388319582,-0.4902573526,0.208188042,-0.0712753013,0.3739586771,0.3999239802,0.2587515712,0.0306262709,0.0953699797,0.1411890537,0.017081609,0.1153032929,0.1724141389,0.3782645166,0.0006368248,0.3976308405,0.1009947807,-0.0365315303,-0.0323385783,-0.0555147193,0.0155075481,-0.3034598231,-0.077748172,-0.0351036154,-0.0424127616,-0.4622249007,-0.2190282345,0.0756054074,-0.4277582169,-0.0240824334,-0.2205747962,-0.0749537423,0.0449578203,-0.040751107,0.2646930516,-0.1134312823,-0.0750159249,0.1873975843,-0.2155476958,0.2505766451,0.0112599647,-0.072178781,0.0076076663,0.4727571011,0.1681151837,0.1057905555,-0.3163026273,-0.1212039888,0.0094633382,-0.0259412955,0.1020844951,0.0664756149,0.0212311354,-0.0882798508,-0.2201028615,0.0009320721,-0.0085922442,-0.1276937723,-0.2165890336,-0.0300458875,0.0397780314,0.0676835105,-0.1183193401,-0.6376333833,-0.1809846014,0.3180422783,0.1766311377,-0.0548113473,0.0939994603,0.1549415439,0.3043794334,0.0776009187,-0.1957989633,0.0346550234,-0.0302676409,0.424775809,-0.1688498259,-0.1342032403,-0.0297146272,-0.3406842351,0.0908480659,0.1222209781,-0.2838962674,0.1603948474,-0.1505679488,0.2198680341,-0.127745986,0.0273664091,0.3573589027,0.0602986179,-0.0297792889,-0.1033775434,-0.3401471674,0.1137283072,0.2512102723,0.0733656585,-0.0979011804,0.2904837728,0.0094959065,0.4994259775,0.0853195935,-0.2391103804,0.014282614,-0.2247809619,0.2057155222,-0.1449179798,-0.3179531395,-0.3026532233,-0.3183403909,-0.3310902119,-0.1800659597,0.1134015098,-0.0285721943,0.1375082433,0.0903949291,-0.1987198591,-0.2026927173,0.2708093524,-0.6740223765,0.1813392192,-0.0891495049,-0.1258283406,-0.404538393,0.0984961838,-0.0701917708,-0.0384590775,0.2655133903,-0.1265854239,-0.5840800405,-0.111919798,-0.2278465927,0.1193516329,0.2090334296,-0.0272874292,0.2653865516,-0.1220546514,-0.0045997296,0.206783548,0.3725515306,-0.2080981135,-0.4580442607,0.285202533,0.1473425925,-0.4616538882,0.1151888818,-0.0796240941,0.4444980323,-0.0613804646,0.0051466841,-0.1397317499,0.4023553729,0.2305787653,0.2341134548,-0.3104236722,-0.1449581534,0.131154865,0.0012798422,-0.2486538291,-0.1349311769,-0.0098688491,0.0213395171,0.2452444583,-0.2326613218,-0.3676378131,0.029967241,0.2006902844,-0.1398627609,0.1253411025,0.2870236337,0.1219745353,0.1349612325,0.1399237663,0.1121450812,0.5952422619,-0.0202477463,-0.479912281,-0.2391503751,-0.2915229201,0.6086901426,0.1036718637,0.3361125886,0.4594851732,0.0016220398,-0.2352100611,-0.0120350001,-0.0177960824,0.0919251889,0.0123249087,-0.4333645105,-0.5076603889,0.4729131162,-0.076935485,-0.030795807,0.2208504081,0.2614873648,-0.4657917619,0.6054710746,0.1998542696,0.8932750821,-0.1469056755,0.3738089204,0.0348119587,-0.3326680064,0.6814112067,0.077867642,-0.1354804039,-0.1288094223,-0.1333549023,-0.0370165072,-0.2810210884,0.0225821566,0.2465423793,-0.3567438424,-0.102960512,-0.25158602,0.0648593754,0.0022847245,0.3220764399,0.0512555167,-0.260430485,-0.1749311537,0.112795718,0.1939959675,-0.2546724379,-0.1819509715,-0.0688524917,-0.3532896638,0.0265283976,-0.5213059187,-0.3650172949,-0.0701386854,-0.1139116436,0.1487181783,-0.6611782312,0.346783489,0.4616463482,0.1879403442,-0.128595084,-0.003723809,0.176964283,-0.0068006986,0.3089605868,0.1885764003,-0.3346206248,0.2705011964,0.273540467,-0.2388724834,0.3515390456,0.0438984819,-0.6558929682,0.1967584044,0.0641772449,0.2177383304,-0.3088622987,-0.2596672773,0.4255558252,0.0035273561,-0.1987577528,0.0807658955,0.1387988329,-0.0753990561,-0.0320843533,0.2848678529,-0.2318786234,0.0608326197,-0.0147259617,0.0693530068,0.1450016797,0.6772645712,0.1283527613,0.0261388868,-0.0661325231,0.2625180781,0.3511837721,-0.406136632,-0.0821773261,0.0408539772,0.0779559687,-0.3869836032,-0.2224243879,0.1610711515,0.2294378579,-0.4493671656,-0.2168513238,-0.3177274764,0.3808581233,-0.0504068211,0.23924914,-0.18727763,0.1589281559,-0.1343093514,0.4010458589,-0.2158996165,-0.0859098732,0.2108703554,0.0759401321,0.1640669852,0.1852780581,-0.2107566446,-0.2444543839,0.0836763158,0.0423154235,0.3284806311,-0.1688713133,-0.1297980398,0.1542046815,0.1382047534,-0.0173474122,-0.0715978369,-0.1414534599,0.303001076,-0.2538660765,-0.2127749622,0.2620382905,-0.0331278667,0.1737251282,0.5312664509,0.337606281,-0.2378075421,0.116462037,0.2220294178,0.2771136463,0.0760804191,0.2996048033,0.2779506445,0.1226189211,-0.1188042834,0.2885484397,0.3634527326,0.1376227736,0.0904041529,-0.1753167212,0.1930162013,-0.0609081611,0.1542157382,0.1333895475,-0.291692704,0.3697533906,0.1069101468,0.179209128,-0.1814869195,0.1208542511,0.2839511037,-0.1111589149,-0.0474189483,0.2033611238,-0.1261960566,0.3426189721,-0.298107326,-0.0993819907,-0.137164399,0.0946204141,0.1641959697,0.3173690438,0.187322706,-0.0866383538,0.0626525432,0.0638852119,0.1450577527,0.3319681883,-0.1406487823,0.3146107495,0.464848578,-0.0208337922,0.350861311,0.2495752424,0.1409587115,0.030789258,0.317615658,-0.0233212374,0.3664501309,0.2210206836,0.2842615545,-0.1086848155,-0.0091051832,0.107525751,0.1083621606,-0.2118569762,-0.122474663,0.0364987701,-0.0205463879,-0.0741009563,-0.1425177753,-0.5123056769,-0.0341583304,-0.0853708684,0.1404239237,-0.2805654705,0.4928234816,0.1721782237,-0.0094985384,-0.1633637398,-0.0768961832,-0.2077182233,0.2828519344,0.3363240659,0.3815351427,-0.0006949767,0.1333020031,-0.1226529106,0.5888957381,-0.2367308289,-0.1762885451,-0.0659889504,-0.045477055,-0.1597679108,-0.0615835823,0.2744336426,0.1691193283,-0.0449253023,0.186847955,0.1161122471,0.2091293335,-0.6766968369,0.2686004043,0.0323620178,-0.2752904594,-0.0638600215,0.279037118,-0.0605890192,-0.0382368937,0.5366311669,-0.2183659971,-0.0128683774,0.1525652707,0.0214357711,-0.063797012,0.3111459315,0.3449728787,0.5670549273,-0.0785512403,-0.0329227075,-0.5537803173,0.0127063245,-0.1254498214,-0.0787407532,-0.1023719236,0.0203919243,0.3542886078,0.3510195613,0.3150633574,-0.1362245828,-0.374083519,0.1765363216,-0.0343620703,-0.4440455735,0.2825790048,0.336189568,-0.0131964078,0.042288959,-0.1669196784,0.2616893053,-0.0259233043,-0.0873204693,-0.1772266328,-0.0064237369,0.0534231365,0.5239661932,-0.0151322978,0.3918854594,-0.0064399615,0.1270287633,-0.2696122825,-0.4526206255,-0.4414311349,0.4491162896,-0.118187733,0.3053471446,-0.2289825827,-0.0073042377,0.1779317111,-0.2373286188,-0.1189952642,-0.0734003633,-0.4311752319,0.3442469835,-0.2255412787,-0.0031985899,-0.1325526237,0.1185994148,0.0393945351,0.2314158231,0.0691797584,-0.0613231324,0.1286036521,-0.2695285082,-0.0482555032,0.2009219825,0.0318991281,0.1656073928,-0.1682991832,-0.703436017,-0.2480917722,0.4426995218,-0.173402071,0.1599380076,0.0933861658,0.1548041403,0.1076184288,-0.203557685,0.5826995969,-0.2855380177,0.065538682,-0.4006927013,-0.2868953347]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1728","title":"Add an entry to an arrow dataset","comments":"> Didn't we talk about returning a Dataset instead of a DatasetDict with load_dataset and no split provided @lhoestq? Not sure it's the way to go but I'm wondering if it's not simpler for some use-cases.\r\n\r\nMy opinion is that users should always know in advance what type of objects they're going to get. Otherwise the development workflow on their side is going to be pretty chaotic with sometimes unexpected behaviors.\r\nFor instance is `split=` is not specified it's currently always returning a DatasetDict. And if `split=\"train\"` is given for example it's always returning a Dataset.","body":"Is it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n","comment_length":96,"text":"Add an entry to an arrow dataset\nIs it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n\n> Didn't we talk about returning a Dataset instead of a DatasetDict with load_dataset and no split provided @lhoestq? Not sure it's the way to go but I'm wondering if it's not simpler for some use-cases.\r\n\r\nMy opinion is that users should always know in advance what type of objects they're going to get. Otherwise the development workflow on their side is going to be pretty chaotic with sometimes unexpected behaviors.\r\nFor instance is `split=` is not specified it's currently always returning a DatasetDict. And if `split=\"train\"` is given for example it's always returning a Dataset.","embeddings":[0.148923099,0.2032505423,-0.053670004,0.0051242253,0.1111596525,0.0731266066,0.273730725,-0.0879939646,0.0876799151,0.0217209589,0.3423003554,0.4724981785,-0.3235454559,0.074565649,0.3323621154,-0.2733576596,0.0625657812,0.0564744137,-0.2627886534,-0.1407849193,-0.162413016,-0.1628737152,0.2530605793,0.0563627779,-0.2190205157,-0.1572354585,-0.0339656062,-0.2609067261,-0.1684876382,-0.4851483405,0.1064505875,-0.07251513,0.1209618077,0.1816625446,-0.0001208031,0.1351278275,0.2979466915,0.1087860689,-0.649518013,-0.3481771946,-0.2126668096,-0.5955251455,0.3207849562,-0.2333696336,0.1756075919,-0.3336721063,0.2961094379,-0.5603634119,0.1842574328,0.1638450176,0.1235807464,0.0644707233,0.1811811775,-0.0514905676,0.3474474251,0.2932673395,-0.2006346732,-0.0563739426,0.1106201485,0.2912874222,0.2085744292,0.0371348225,0.2039234936,-0.1755607873,0.4447499216,0.2522757351,-0.1014347598,-0.2049790025,-0.2078369558,0.2939831913,0.6473250389,-0.2176996022,-0.748470366,-0.4845074415,0.4366750419,-0.1822227538,-0.1971818507,0.2327401936,-0.0499619879,0.1590688676,-0.0214077383,-0.5370777845,-0.1212952062,0.1997885555,-0.0781413838,0.2104478776,0.0671586096,0.1875111312,-0.0743544549,0.0980985984,0.4603864253,-0.1218010858,0.0015314725,0.09527383,-0.2415177673,-0.0359287411,-0.167870149,-0.0899103358,-0.0488909744,-0.011518646,0.4208772779,-0.0647685379,-0.3213697076,-0.0447110087,0.31207937,-0.1362739801,0.0155989397,0.1233666912,-0.3104012311,-0.56049788,-0.03761952,-0.1776010096,-0.07526806,-0.0327271484,0.2470703423,-0.0488372482,0.2522363067,0.2627443969,0.0140607636,-0.1079403013,-0.3875339627,-0.0179843139,-0.2659661174,-0.0414333045,0.3194258809,0.2228212059,0.1218316555,0.5000299215,0.2188774645,-0.2792986333,0.0607988089,0.0001286661,-0.031761229,0.2036891878,-0.1720125675,0.1425707936,-0.123877883,-0.2613933682,-0.0565125048,-0.1344753504,0.1910654306,-0.1047654226,0.2506734431,0.0635320246,-0.1267582029,0.2028165162,0.0935413092,-0.1517158747,-0.1709172428,0.2163983136,-0.1591831744,-0.3000670075,-0.265322715,0.0598790832,-0.0181106552,-0.1914280057,-0.0541082658,0.3932879865,-0.052761402,0.1217951626,0.0635990202,-0.2271443307,-0.3002089262,-0.2059619725,-0.1810673028,0.0652472973,-0.5810016394,-0.3230880797,-0.3965204656,-0.0007034457,-0.4486553371,-0.030775208,-0.2260406911,0.7877774239,0.038224753,-0.2032297701,0.6214304566,-0.0511002392,-0.282556653,-0.1758330166,-0.271505028,0.1707303524,0.1053495109,0.5825029016,0.3470585942,0.179490909,-0.0493706241,0.1778184026,-0.1028186306,-0.2526222169,0.2119360715,-0.2013582438,0.2867557108,-0.044158116,-0.0076798014,0.1050690487,0.1689320952,-0.0399998538,0.1117276773,-0.1499639452,0.1267689615,-0.0233246386,0.2841738462,0.1198530048,-0.1577977091,-0.111959964,-0.431922704,-0.0166914091,0.1539265066,-0.0038995047,0.0790131316,-0.2508654594,0.0055507189,0.3583600521,-0.0269216802,0.1621829569,-0.0060971617,0.0200405102,-0.0497466438,-0.2303821594,-0.3420437276,0.3552744687,-0.1019709408,0.2033836246,-0.2506042123,0.2522015274,0.299546808,0.1008187234,-0.287771523,0.1638204604,-0.2433373034,-0.0385141447,0.1070636883,0.0011131002,0.0557104275,0.1099162698,0.0411931388,0.1264181137,0.0943575725,-0.2818855047,0.1626383066,0.0383653939,0.1223354638,-0.0892992765,-0.2213629186,0.3601261079,0.1634077877,0.3806601763,-0.1537361145,0.2221234143,-0.0061401171,0.1120590568,-0.4190380275,-0.1833490431,-0.2857301235,0.0351908281,0.3241585493,-0.1198494509,-0.0863564461,-0.0843106508,0.3114069998,-0.0787733942,0.2204767168,0.0828586593,-0.3383288682,0.1602659971,-0.0215830114,0.3791629076,0.4297638535,0.2466750592,-0.0312965326,-0.0034333984,0.0568315089,-0.0030326303,0.1057171896,0.2794118226,0.4154753387,-0.08432208,0.2271009684,0.0517538227,-0.0090965657,-0.0256557651,-0.0598952584,-0.0686763376,-0.2929493785,-0.0567642637,-0.0812907591,0.0581328794,-0.4557873905,-0.2034327239,0.0852361917,-0.5229114294,0.0822524056,-0.2457781732,-0.1431063712,0.045576591,0.0284416284,0.3708215356,-0.2270681709,-0.0604219846,0.2260522991,-0.1569586396,0.1129825786,-0.0194963254,-0.0490853898,-0.0899859145,0.4903205633,0.0772578642,0.0964477882,-0.3023446798,-0.2243509591,0.104368031,0.0484220237,0.1497203708,0.1132983267,0.0604018979,0.026830852,-0.2328117788,-0.0854227766,-0.027918892,-0.1692318171,-0.2233995944,-0.0223854464,0.0732144117,0.0249797795,-0.2803646326,-0.5676974058,-0.1390215158,0.3047088981,0.1716894954,-0.0562261678,-0.0078882314,0.1825787872,0.1792978495,0.0547783487,-0.1193318143,0.046669282,0.013216503,0.3442359567,-0.1703812033,-0.1163793281,-0.1266447455,-0.4445589483,0.0552888215,0.2123786509,-0.3078700006,0.1286229789,-0.1602936536,0.2679947913,-0.0304954033,0.0031080828,0.3798833787,0.1025995612,0.0253196005,-0.0556906946,-0.2710007429,0.1926504821,0.1774702519,0.085074231,0.0125161121,0.3000861704,-0.0044603921,0.6486833692,-0.0794266164,-0.3330708146,-0.0649883971,-0.1147667617,0.326797843,-0.2337396145,-0.2221952677,-0.3267285228,-0.3186516166,-0.3132205904,-0.0835192576,0.005292763,-0.077703543,0.0887069702,0.2114212513,-0.2691901624,-0.2678107321,0.3111074567,-0.6298688054,0.0771703422,-0.0743313134,-0.0013970719,-0.3483353853,0.129980281,-0.159095943,0.0514903776,0.2891385257,-0.0769694671,-0.6699850559,-0.08706972,-0.1398099512,0.1890466958,0.1725387126,-0.0104783718,0.2996673882,-0.1478800029,0.0120929033,0.2610084414,0.2844207287,-0.2166789472,-0.5602386594,0.1395817697,0.0543415248,-0.3661279976,0.0453470834,-0.169952035,0.4356377125,-0.0861514881,-0.0046418458,-0.2474862039,0.3057092726,0.2288710475,0.1961987615,-0.3520723283,-0.0805228874,0.0824758634,0.0904961601,-0.236997366,-0.0938414857,-0.0167402308,-0.0108907688,0.0594586916,-0.3298030496,-0.3788066506,0.0147902658,0.31976071,-0.0916461125,0.1416493058,0.2713501751,0.0501978397,0.1301242113,0.30782637,0.1400090754,0.5922588706,-0.0919274762,-0.4606248438,-0.1038496196,-0.2106000483,0.5864296556,0.3168060184,0.2484680712,0.3790532053,0.0353511795,-0.1912751347,-0.0631447285,-0.0208831392,0.2025591433,-0.0373814367,-0.4521546066,-0.5612297654,0.4618368149,-0.1597911417,-0.1130278185,0.1908476353,0.333807379,-0.4775661826,0.6409643888,0.2579957843,0.9893674254,-0.211165145,0.3801468611,0.0350422636,-0.457918793,0.7528198957,0.006091428,-0.1959100068,-0.1499024481,-0.0239044651,-0.0728196651,-0.2540908158,0.1052178293,0.409646064,-0.2787844241,-0.0575833879,-0.1519394666,0.0742922425,-0.0488014333,0.4891758859,0.0362976566,-0.260941714,-0.1551221013,0.0495852008,0.1489238888,-0.1676212102,-0.2056401074,-0.0364191979,-0.3538440466,-0.0177908614,-0.3853800595,-0.4576152563,-0.0545438901,-0.0621016212,0.1572807729,-0.631254077,0.2773555815,0.4316859245,0.1770157367,-0.0818972811,0.0612282902,0.1142999008,0.0093822666,0.3079133928,0.2511932552,-0.2919282019,0.3507091105,0.2755832076,-0.2182187885,0.3583336473,-0.0312861204,-0.5524540544,0.0790699199,0.0269332249,0.2787884772,-0.4025583565,-0.280854255,0.3540420234,-0.0073138773,-0.1434085518,0.0453060493,0.1206349656,0.0537955165,-0.0208730344,0.3707684875,-0.3586654067,-0.0569192767,-0.1460547745,0.1236387789,0.1889640093,0.6302036047,-0.0363749824,0.0173717141,-0.0663973391,0.3725510538,0.4095952511,-0.4745312631,-0.0575968698,0.0810798034,0.007069278,-0.3730997741,-0.1483187675,0.161997363,0.155482024,-0.4441173673,-0.1573477536,-0.1814362556,0.4117986262,-0.0880282819,0.2797031403,-0.1598335057,0.1193644926,-0.120648697,0.3268199563,-0.1923899353,-0.2208414823,0.2083687782,0.1398629844,0.0472863503,0.2672201991,-0.1941837221,-0.1489673704,0.0188519713,-0.1273257285,0.1814620793,-0.1143285707,-0.0884692669,0.1907345653,0.088232398,0.0903362408,-0.0910239741,-0.1595667452,0.2966213524,-0.2590958774,-0.2267874479,0.2751898766,-0.0027090902,0.2392907739,0.5310438871,0.3237861097,-0.2116095573,0.2079462111,0.1457070708,0.3464228809,0.0278846938,0.301366806,0.2141531706,0.0851070657,-0.0797819123,0.2289365977,0.335493505,0.1972949803,0.069767423,-0.171671465,0.3238380849,-0.0686435178,0.0310595389,0.2195026577,-0.2505013049,0.2719677985,0.0904528648,0.1206587553,-0.0914857909,0.1981366724,0.298103869,-0.1589946151,-0.0850248933,0.3534744382,-0.0201178156,0.4279463887,-0.1790750176,-0.1807137281,-0.1604134738,0.0289342757,0.2364527732,0.2993360758,0.0981521532,-0.0064234324,0.1560715735,-0.0072660539,0.1173323542,0.3177246153,-0.0918362588,0.2884958386,0.407904774,0.0130472276,0.4421755672,0.2073028237,0.1801927388,0.0164598394,0.1949005723,0.0604798645,0.4181532562,0.1689252853,0.1771315038,-0.0797941238,-0.1518691778,-0.0339651927,0.065409556,-0.1403416842,-0.1685167253,0.0192501899,-0.0469795987,-0.1040003821,-0.2351780981,-0.5446068645,-0.0315430984,-0.0346724913,0.2435895652,-0.2146822214,0.5122625232,-0.0096062338,0.0798562169,-0.1552592516,-0.0565939136,-0.3192983866,0.3565749526,0.4216636121,0.3080934882,-0.0680493265,0.2205003649,-0.1089674383,0.4040494263,-0.1560076475,-0.1215821877,-0.1482998133,-0.0475272015,-0.0810331106,-0.0306339543,0.3937811852,0.133745715,-0.0135326935,0.1561074406,0.1313361973,0.1306078583,-0.6683197021,0.3725160658,0.0266257282,-0.245339483,-0.017381005,0.3169880807,-0.0643003285,-0.0783241466,0.4493919313,-0.2633306384,-0.0176755022,0.0469757877,0.0154293049,-0.033552818,0.2938138843,0.2739835382,0.5760567784,-0.0559357218,0.1109661162,-0.5531082153,-0.0354328044,-0.1621958017,-0.0557524376,-0.0669010133,0.065895766,0.3969287574,0.4664539695,0.2242666483,-0.1635936052,-0.3849122524,0.1385299414,-0.0976768136,-0.4929006696,0.2218852043,0.3181905746,-0.0132462699,-0.0267989654,-0.1311637759,0.269962132,-0.0999356955,-0.0804139599,-0.0880935863,0.0824149251,0.1394577324,0.5118872523,0.0602499843,0.4006276131,0.0086113848,0.0406881832,-0.2300785929,-0.4782167077,-0.4323364794,0.504268527,-0.1593205929,0.3590171635,-0.2211955935,0.0551279522,0.1431582421,-0.2344006002,-0.0636246651,0.0150138289,-0.4317902029,0.2810560167,-0.2680620551,0.0578377284,-0.0730624422,0.1771153361,0.0814535245,0.1846217811,0.1128995717,-0.0336081199,0.1386211812,-0.1476259679,-0.0028349629,0.178612113,-0.0001056461,0.198628962,-0.160692215,-0.7023429275,-0.2793782353,0.5140505433,-0.2919913232,0.2940156162,0.0106887436,0.1839267164,0.0735346526,-0.2659797966,0.5728676915,-0.3577988446,0.0743611455,-0.3251976669,-0.327375263]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1728","title":"Add an entry to an arrow dataset","comments":"Thanks @thomwolf. Your solution worked!","body":"Is it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n","comment_length":5,"text":"Add an entry to an arrow dataset\nIs it possible to add an entry to a dataset object?\r\n\r\n**Motivation: I want to transform the sentences in the dataset and add them to the original dataset**\r\n\r\nFor example, say we have the following code:\r\n\r\n``` python\r\nfrom datasets import load_dataset\r\n\r\n# Load a dataset and print the first examples in the training set\r\nsquad_dataset = load_dataset('squad')\r\nprint(squad_dataset['train'][0])\r\n```\r\n\r\nIs it possible to add an entry to `squad_dataset`? Something like the following?\r\n\r\n``` python\r\nsquad_dataset.append({'text': \"This is a new sentence\"})\r\n```\r\n\r\nThe motivation for doing this is that I want to transform the sentences in the squad dataset and add them to the original dataset.\r\n\r\nIf the above doesn't work, is there any other way of achieving the motivation mentioned above? Perhaps by creating a new arrow dataset by using the older one and the transformer sentences?\r\n\nThanks @thomwolf. Your solution worked!","embeddings":[0.1719388515,0.1842069328,-0.0813796669,-0.0647207573,0.2561539114,0.3051029146,0.2442452908,-0.1456887126,-0.1036642194,-0.0502436534,0.3738768995,0.5951262116,-0.2461642623,0.0165975615,0.3777017891,-0.2667967379,0.1133872792,0.0447983071,-0.2504124343,-0.0691238418,0.0446947627,-0.2058667243,0.1913676262,-0.0647286698,-0.2802021503,-0.1174940988,-0.0864465088,-0.2907868624,-0.1284998804,-0.5253661871,0.0066226744,-0.0905576423,0.0825037062,0.3375755548,-0.0001176547,0.149991557,0.3458918333,0.1861445159,-0.5488336086,-0.2188324779,-0.3215689957,-0.5417425036,0.1834633797,-0.1978980005,0.1780669987,-0.326338172,0.1672555506,-0.5153521299,0.281978935,0.2529530525,0.1619930565,0.0061959298,0.3172542453,-0.1124840528,0.2491490543,0.271894455,-0.2256937474,0.0735901818,0.0727753043,0.2279538661,0.2208415121,0.1071285382,0.2217745483,-0.138267383,0.462936908,0.1437303126,-0.0724119321,-0.1300264448,-0.1583129168,0.173115626,0.7781195045,-0.3917584419,-0.7442389727,-0.2926286757,0.5203200579,-0.1109493896,-0.2232936472,0.2196496427,0.0293874294,0.1139742285,0.1267449558,-0.5327539444,-0.2255693525,0.1891495138,0.0314486176,0.3134854734,0.0635810345,0.0910144001,-0.0674679801,-0.023824079,0.4265984595,0.1639313549,-0.1148058772,0.2213632464,-0.3010472655,0.0189895164,-0.2151296735,-0.0984511971,-0.0729873702,-0.056642957,0.5149729848,-0.1260854751,-0.2866313756,0.0601715185,0.1936084926,-0.0800912753,0.1067703366,0.0892983302,-0.3686448634,-0.4356934428,-0.1487978399,-0.2314954847,-0.1668333411,-0.159984827,0.3759285212,-0.1407966912,0.1801189631,0.279600352,0.0820889995,-0.0699917674,-0.3405244052,0.0817987323,-0.250977546,0.0512987822,0.4520310462,0.0728476718,0.1301232129,0.5077289939,0.311390698,-0.1262048334,0.0279042479,0.1321007609,-0.051018063,0.2443618178,-0.1978809237,0.076860182,0.0534870252,-0.1877687573,0.0570274331,-0.1994450241,0.1603489071,-0.039147839,0.143252179,0.0518695302,-0.0863494202,0.2035090178,0.1152211204,-0.1456092894,-0.0298990533,0.2803729773,-0.0814466774,-0.0684952885,-0.4108831286,0.1053436249,-0.0723612756,-0.3927749395,-0.0068884846,0.3194470406,0.002484659,-0.0232584272,0.1735840738,-0.0419901237,-0.238129884,-0.1037280783,-0.0774059445,0.0736762807,-0.6103615165,-0.3345002234,-0.304064393,0.1155218184,-0.4890156388,-0.1554301381,-0.1317970902,0.5391035676,0.0578253381,-0.0070139314,0.5192616582,-0.1101896018,-0.2425990552,-0.1723958105,-0.352450639,-0.0039072642,-0.0077383253,0.5090691447,0.196208626,0.1991958916,-0.0495190322,0.1515126228,0.1105714589,-0.3049331307,0.2003921866,0.0473437794,0.1925040781,-0.0156276468,-0.2112497091,-0.0052674529,0.1987493336,-0.040482454,0.1175077856,-0.1971632689,0.1980590224,0.3129967451,0.2969707549,0.149919644,-0.1098316163,-0.1596617997,-0.150403887,-0.0247115567,0.2253180593,0.066613175,0.0111957956,-0.2086413503,-0.0932321027,0.3418779075,-0.0403696857,0.0188218672,0.0530666262,0.0343400724,-0.1273206621,-0.2279298156,-0.382360369,0.3692679703,-0.09092246,0.1472356468,-0.1031986549,0.1236291826,0.1876509935,-0.0479932651,-0.3328092396,0.2053543776,-0.1746655256,0.0654909536,0.0077670161,0.046256341,-0.0281263664,0.2172822803,0.1229071617,-0.0092899147,0.0838603228,-0.4570809901,0.1547608972,0.0999025628,0.1048716679,-0.0698370561,-0.3615306318,0.1758290529,0.1518366784,0.296645999,-0.1527711749,0.2653492689,0.1135731861,-0.0241109449,-0.4201289713,-0.1620526463,-0.1688970476,0.0356223248,0.2442015707,-0.1985565722,-0.1162944809,-0.0325121023,-0.0440252014,-0.1027366146,0.1665565073,0.1536232233,-0.523127079,0.2834532261,-0.0143895457,0.3447696865,0.3222021163,0.2558426559,-0.0103448573,0.0977636874,0.1558148265,0.0026228528,0.0536827259,0.0834592208,0.4460922182,-0.0296926256,0.4354908168,0.1223078817,-0.0878283903,-0.0030405198,-0.0197004713,0.0811052322,-0.3767969906,-0.0591277219,-0.0537132844,-0.0576136224,-0.4549122751,-0.1399757117,0.0684682652,-0.3603794575,0.0097241495,-0.225445658,-0.0552390069,0.087163493,0.0402042791,0.1982860565,-0.0418281555,-0.059205126,0.1350327432,-0.2375448048,0.2279473096,0.0029936128,-0.0613170341,0.0737963766,0.3656819165,0.1559689641,0.1212822944,-0.3203746378,-0.1758863181,-0.0213611517,0.0310834181,0.1787581593,-0.0321631916,0.1098207682,-0.1562097222,-0.1968761832,0.0006911614,-0.0690231025,-0.1483791918,-0.1523826569,-0.0657513216,-0.0570726059,0.0616139174,-0.0882838294,-0.5746464729,-0.1591787785,0.3237910271,0.2182342112,-0.0455938764,0.1472127438,0.228694886,0.2522748709,0.1068869308,-0.1234072298,0.0891510621,-0.1149877459,0.4776900411,-0.1619498432,-0.1984499991,0.0418431088,-0.3387674391,0.0570990667,0.0605427064,-0.2635363638,0.0706474483,-0.186005488,0.1246801764,-0.1006116867,0.0522980988,0.3314876854,0.0001342068,-0.0654500723,-0.1060611233,-0.3827452064,0.1325768679,0.2730799913,0.094255656,-0.0721605048,0.2734757662,-0.0049929591,0.3655053377,0.0405944996,-0.2727563083,0.0932064876,-0.2783531249,0.2287055701,-0.089059785,-0.3324878812,-0.2098475695,-0.2627031505,-0.3536171913,-0.2394239902,0.1302055568,-0.0916768238,0.1748020947,0.0288217422,-0.2705383897,-0.1210890189,0.2459214777,-0.6368916631,0.2142460793,-0.0901653841,-0.1685890108,-0.47543329,0.0104446607,-0.0203892104,-0.0104679754,0.1778462231,-0.1473812312,-0.5058400035,-0.0890733674,-0.1417291164,0.0971920863,0.1788366735,0.0928651839,0.1911506802,-0.1009556651,0.0474686064,0.1242197454,0.3182453215,-0.307927072,-0.3710151017,0.3187039196,0.2091121674,-0.5123502612,0.1053686067,-0.0693475232,0.3809458911,-0.0015650303,-0.0017230365,-0.0960475355,0.4088679254,0.2514020801,0.1933528036,-0.2791011631,-0.2130685598,0.0733656958,-0.0122357579,-0.3298392892,-0.1472469568,0.0394107066,-0.0208506044,0.3758946359,-0.2129125446,-0.3571538329,0.1124042049,0.196012333,-0.1046269089,0.1180082858,0.3463539779,0.0619434118,0.1143854335,0.1151861027,0.0408041626,0.5161113739,0.0766911209,-0.4703429639,-0.2708627284,-0.356028229,0.5750693679,0.0921342671,0.3431915641,0.4254712164,0.0184825286,-0.2070478201,0.0206674561,-0.0136401467,0.1295748204,0.1047321185,-0.3397843242,-0.4892902672,0.4690124691,-0.0449720919,-0.0108460533,0.1762008071,0.2733536661,-0.3791273534,0.5777012706,0.2149410993,0.866897583,-0.1518758088,0.3083608747,0.0470425822,-0.378335923,0.5958612561,0.0396064557,-0.1277167052,-0.1168490499,-0.1160946861,-0.0435985848,-0.2897390723,0.1061693877,0.2030832767,-0.4062409103,-0.2245836556,-0.3787296712,0.0866175517,0.0579393134,0.3001862168,0.0164268184,-0.226706773,-0.1714736819,0.1521448195,0.2192448378,-0.1961070448,-0.1795079857,-0.1361663789,-0.3413030803,0.0633094907,-0.5717216134,-0.2676184177,-0.1015270576,-0.1424398422,0.1454800963,-0.5662549138,0.3562009633,0.528852582,0.2525895238,-0.1280064881,-0.0195608083,0.2514640987,0.0231880546,0.2359195948,0.1595983505,-0.2926623225,0.279768765,0.1877827495,-0.2957254052,0.3967508972,0.063576892,-0.6503013372,0.248610571,0.0362917855,0.1129599735,-0.2827554047,-0.3385262191,0.3785719275,-0.0502605699,-0.195577383,0.1144839376,0.1156725511,-0.1456046849,-0.0162064154,0.3338679969,-0.1767818779,0.1445916891,0.0302563943,0.0161568839,0.1985996813,0.706063509,0.1213697717,-0.0205398723,-0.1263540685,0.2361733913,0.310175091,-0.3659886122,-0.0702539384,0.069875516,0.0132976612,-0.3425194323,-0.2112476379,0.1443569064,0.2309429646,-0.4441606998,-0.2757880688,-0.3808216155,0.2852993309,-0.0495437272,0.1992142648,-0.2619334459,0.158517316,-0.0094011007,0.3731145561,-0.2377436608,-0.0334239863,0.113980487,0.0260032602,0.1696209759,0.1975983381,-0.2241990268,-0.361325115,0.1184922159,0.041540131,0.3197012246,-0.1789211184,-0.1547964364,0.1386212707,0.1180244386,-0.0409549139,-0.097826831,-0.18638587,0.2861250043,-0.2191686481,-0.2844661176,0.2255234867,-0.0820446983,0.1917773336,0.4782927334,0.2847206891,-0.1711396724,0.0166201387,0.2026724368,0.197935611,0.0872522146,0.1659371555,0.2891697288,0.1681928188,-0.0910965353,0.2641099691,0.2767882049,0.0998461321,0.0641755089,-0.2948086858,0.2581684887,-0.0094566206,0.1558409929,0.1042607054,-0.3042669594,0.42751652,0.1061842144,0.1864607781,-0.2465356439,0.1255349517,0.3298350871,-0.127292037,-0.0430618562,0.1382620186,-0.1103333905,0.3325312436,-0.2928093374,-0.0052608736,-0.2533000112,0.004104543,0.1637609005,0.2135332078,0.2075354308,-0.0864716619,0.0852766186,0.0808367506,0.144641608,0.3118271232,-0.1084227785,0.2488341928,0.4958634973,-0.0068766987,0.2311778069,0.2511972487,0.1811705679,0.0191296302,0.3594857454,-0.0330701731,0.4265044332,0.2942458093,0.3572005332,-0.1569593549,0.0160835665,0.1415014714,0.0971480012,-0.2262900919,-0.1465116739,0.0212189034,0.0098107653,-0.1080295667,-0.0778558403,-0.5076557994,0.0307906661,-0.0170208048,0.0776496902,-0.3670449555,0.4362710118,0.195774287,0.0163112376,-0.1397008896,-0.0526153967,-0.2624590099,0.2389401495,0.26947245,0.4228457808,0.0740513504,0.1553857028,-0.1275697798,0.6526129246,-0.2754353285,-0.2100571543,-0.0105165038,-0.0669029951,-0.0384923108,-0.0340929925,0.2049478143,0.1791396439,-0.0776749402,0.2123188227,0.1460525542,0.2674693465,-0.5587687492,0.2314593047,0.0021623161,-0.2117541283,-0.157181114,0.2106578946,-0.116741389,0.0214068498,0.5861936212,-0.2147695273,-0.0337464511,0.143483758,0.0199660994,-0.0496112779,0.280816853,0.3375513554,0.5006226301,-0.082156457,-0.0510418527,-0.6495748758,-0.0069146818,-0.1721888036,-0.1586251557,0.001999815,-0.0037193031,0.3159676492,0.2970845103,0.3495329618,-0.1261740327,-0.3820070326,0.1954391301,-0.1784556061,-0.3322654665,0.2693359554,0.4139780402,0.0084623983,0.0787584111,-0.2471406907,0.172330156,0.0198475514,-0.0488781855,-0.1806687862,-0.1075082123,-0.0021482788,0.5569047332,-0.197632581,0.3813206851,-0.0292551573,0.1484609842,-0.2773640752,-0.4599594176,-0.4529805779,0.4605460465,-0.0358174294,0.3831977844,-0.2575508952,0.0412691608,0.1643142402,-0.2953552604,-0.1911395192,-0.1149239689,-0.3915191293,0.382829845,-0.1595135331,0.0779716,-0.1854016781,0.0157711897,0.0178229138,0.2458890527,0.0194092561,-0.1361711323,0.2527491152,-0.2755021155,-0.1147704944,0.2317389548,0.0218710229,0.0993177891,-0.1704912633,-0.6921935678,-0.1676823199,0.3868511915,-0.1198872998,0.1533678025,0.1369834244,0.1111879721,0.1747711301,-0.1596159488,0.6029738188,-0.2624224722,0.155985713,-0.5471912026,-0.255864352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"Upgrading tensorflow to version 2.4.0 solved the issue.","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":8,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\nUpgrading tensorflow to version 2.4.0 solved the issue.","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"I still have the same error even with TF 2.4.0.","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":10,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\nI still have the same error even with TF 2.4.0.","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"And I have the same error with TF 2.4.1. I believe this issue should be reopened. Any ideas?!","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":18,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\nAnd I have the same error with TF 2.4.1. I believe this issue should be reopened. Any ideas?!","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"I'm seeing the same issue with TF 2.4.1 when running the following in https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb:\r\n```\r\n!pip install git+https:\/\/github.com\/google-research\/bleurt.git\r\nreferences = [\"foo bar baz\", \"one two three\"]\r\nbleurt_metric = load_metric('bleurt')\r\npredictions = [\"foo bar\", \"four five six\"]\r\nbleurt_metric.compute(predictions=predictions, references=references)\r\n```","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":39,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\nI'm seeing the same issue with TF 2.4.1 when running the following in https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb:\r\n```\r\n!pip install git+https:\/\/github.com\/google-research\/bleurt.git\r\nreferences = [\"foo bar baz\", \"one two three\"]\r\nbleurt_metric = load_metric('bleurt')\r\npredictions = [\"foo bar\", \"four five six\"]\r\nbleurt_metric.compute(predictions=predictions, references=references)\r\n```","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"@aleSuglia @oscartackstrom - Are you getting the error when running your code in a Jupyter notebook ?\r\n\r\nI tried reproducing this error again, and was unable to do so from the python command line console in a virtual environment similar to the one I originally used (and unfortunately no longer have access to) when I first got the error. \r\nHowever, I've managed to reproduce the error by running the same code in a Jupyter notebook running a kernel from the same virtual environment.\r\nThis made me suspect that the problem is somehow related to the Jupyter notebook.\r\n\r\nMore environment details:\r\n```\r\nOS: Ubuntu Linux 18.04\r\nconda==4.8.3\r\npython==3.8.5\r\ndatasets==1.3.0\r\ntensorflow==2.4.0\r\nBLEURT==0.0.1\r\nnotebook==6.2.0\r\n```","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":112,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\n@aleSuglia @oscartackstrom - Are you getting the error when running your code in a Jupyter notebook ?\r\n\r\nI tried reproducing this error again, and was unable to do so from the python command line console in a virtual environment similar to the one I originally used (and unfortunately no longer have access to) when I first got the error. \r\nHowever, I've managed to reproduce the error by running the same code in a Jupyter notebook running a kernel from the same virtual environment.\r\nThis made me suspect that the problem is somehow related to the Jupyter notebook.\r\n\r\nMore environment details:\r\n```\r\nOS: Ubuntu Linux 18.04\r\nconda==4.8.3\r\npython==3.8.5\r\ndatasets==1.3.0\r\ntensorflow==2.4.0\r\nBLEURT==0.0.1\r\nnotebook==6.2.0\r\n```","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"This happens when running the notebook on colab. The issue seems to be that colab populates sys.argv with arguments not handled by bleurt.\r\n\r\nRunning this before calling bleurt fixes it:\r\n```\r\nimport sys\r\nsys.argv = sys.argv[:1]\r\n```\r\n\r\nNot the most elegant solution. Perhaps it needs to be fixed in the bleurt code itself rather than huggingface?\r\n\r\nThis is the output of `print(sys.argv)` when running on colab:\r\n```\r\n['\/usr\/local\/lib\/python3.7\/dist-packages\/ipykernel_launcher.py', '-f', '\/root\/.local\/share\/jupyter\/runtime\/kernel-a857a78c-44d6-4b9d-b18a-030b858ee327.json']\r\n```","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":71,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\nThis happens when running the notebook on colab. The issue seems to be that colab populates sys.argv with arguments not handled by bleurt.\r\n\r\nRunning this before calling bleurt fixes it:\r\n```\r\nimport sys\r\nsys.argv = sys.argv[:1]\r\n```\r\n\r\nNot the most elegant solution. Perhaps it needs to be fixed in the bleurt code itself rather than huggingface?\r\n\r\nThis is the output of `print(sys.argv)` when running on colab:\r\n```\r\n['\/usr\/local\/lib\/python3.7\/dist-packages\/ipykernel_launcher.py', '-f', '\/root\/.local\/share\/jupyter\/runtime\/kernel-a857a78c-44d6-4b9d-b18a-030b858ee327.json']\r\n```","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"I got the error when running it from the command line. It looks more like an error that should be fixed in the BLEURT codebase.","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":25,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\nI got the error when running it from the command line. It looks more like an error that should be fixed in the BLEURT codebase.","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"Seems to be a known issue in the bleurt codebase: https:\/\/github.com\/google-research\/bleurt\/issues\/24.","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":11,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\nSeems to be a known issue in the bleurt codebase: https:\/\/github.com\/google-research\/bleurt\/issues\/24.","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1727","title":"BLEURT score calculation raises UnrecognizedFlagError","comments":"Hi, the problem should be solved now.","body":"Calling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```","comment_length":7,"text":"BLEURT score calculation raises UnrecognizedFlagError\nCalling the `compute` method for **bleurt** metric fails with an `UnrecognizedFlagError` for `FLAGS.bleurt_batch_size`. \r\n\r\nMy environment:\r\n```\r\npython==3.8.5\r\ndatasets==1.2.0\r\ntensorflow==2.3.1\r\ncudatoolkit==11.0.221\r\n```\r\n\r\nTest code for reproducing the error:\r\n```\r\nfrom datasets import load_metric\r\nbleurt = load_metric('bleurt')\r\ngen_text = \"I am walking on the promenade today\"\r\nref_text = \"I am walking along the promenade on this sunny day\"\r\nbleurt.compute(predictions=[test_text], references=[test_text])\r\n```\r\n\r\nError Output:\r\n```\r\nUsing default BLEURT-Base checkpoint for sequence maximum length 128. You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').\r\nINFO:tensorflow:Reading checkpoint \/home\/ubuntu\/.cache\/huggingface\/metrics\/bleurt\/default\/downloads\/extracted\/9aee35580225730ac5422599f35c4986e4c49cafd08082123342b1019720dac4\/bleurt-base-128.\r\nINFO:tensorflow:Config file found, reading.\r\nINFO:tensorflow:Will load checkpoint bert_custom\r\nINFO:tensorflow:Performs basic checks...\r\nINFO:tensorflow:... name:bert_custom\r\nINFO:tensorflow:... vocab_file:vocab.txt\r\nINFO:tensorflow:... bert_config_file:bert_config.json\r\nINFO:tensorflow:... do_lower_case:True\r\nINFO:tensorflow:... max_seq_length:128\r\nINFO:tensorflow:Creating BLEURT scorer.\r\nINFO:tensorflow:Loading model...\r\nINFO:tensorflow:BLEURT initialized.\r\n---------------------------------------------------------------------------\r\nUnrecognizedFlagError Traceback (most recent call last)\r\n<ipython-input-12-8b3f4322318a> in <module>\r\n 2 gen_text = \"I am walking on the promenade today\"\r\n 3 ref_text = \"I am walking along the promenade on this sunny day\"\r\n----> 4 bleurt.compute(predictions=[gen_text], references=[ref_text])\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/datasets\/metric.py in compute(self, *args, **kwargs)\r\n 396 references = self.data[\"references\"]\r\n 397 with temp_seed(self.seed):\r\n--> 398 output = self._compute(predictions=predictions, references=references, **kwargs)\r\n 399 \r\n 400 if self.buf_writer is not None:\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bleurt\/b1de33e1cbbcb1dbe276c887efa1fad68c6aff913885108078fa1ad408908778\/bleurt.py in _compute(self, predictions, references)\r\n 103 \r\n 104 def _compute(self, predictions, references):\r\n--> 105 scores = self.scorer.score(references=references, candidates=predictions)\r\n 106 return {\"scores\": scores}\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/bleurt\/score.py in score(self, references, candidates, batch_size)\r\n 164 \"\"\"\r\n 165 if not batch_size:\r\n--> 166 batch_size = FLAGS.bleurt_batch_size\r\n 167 \r\n 168 candidates, references = list(candidates), list(references)\r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/tensorflow\/python\/platform\/flags.py in __getattr__(self, name)\r\n 83 # a flag.\r\n 84 if not wrapped.is_parsed():\r\n---> 85 wrapped(_sys.argv)\r\n 86 return wrapped.__getattr__(name)\r\n 87 \r\n\r\n~\/anaconda3\/envs\/noved\/lib\/python3.8\/site-packages\/absl\/flags\/_flagvalues.py in __call__(self, argv, known_only)\r\n 643 for name, value in unknown_flags:\r\n 644 suggestions = _helpers.get_flag_suggestions(name, list(self))\r\n--> 645 raise _exceptions.UnrecognizedFlagError(\r\n 646 name, value, suggestions=suggestions)\r\n 647 \r\n\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n\r\nPossible Fix:\r\nModify `_compute` method https:\/\/github.com\/huggingface\/datasets\/blob\/7e64851a12263dc74d41c668167918484c8000ab\/metrics\/bleurt\/bleurt.py#L104\r\nto receive a `batch_size` argument, for example:\r\n```\r\ndef _compute(self, predictions, references, batch_size=1):\r\n scores = self.scorer.score(references=references, candidates=predictions, batch_size=batch_size)\r\n return {\"scores\": scores}\r\n```\nHi, the problem should be solved now.","embeddings":[-0.3299125135,-0.3983678818,0.0414073691,0.4000958502,0.3001840711,-0.2275916785,0.2741967142,0.2562865913,0.0054997536,0.2942811847,-0.0122730052,0.0224692263,-0.0850862861,0.0461418107,-0.2817866504,0.1284505874,-0.209513098,-0.0770389214,0.4217458963,0.0855357051,-0.3209675848,0.1901822984,-0.1129608974,-0.1827558875,-0.3120544553,0.0189661421,0.3288890123,-0.0016792981,-0.1310292333,-0.2663407624,0.1180293337,-0.1967550069,0.1121654436,0.5339958668,-0.0001142144,-0.1661914289,0.1390840411,-0.1814889759,-0.1773090065,-0.1401195824,-0.0342725813,-0.2571440339,-0.1122626364,-0.2426050603,0.0441503674,0.1876826435,-0.1606330574,-0.0122861192,0.1843592823,0.3979115784,0.1449992061,0.1509114951,0.0781051069,0.0778455064,-0.1381148547,-0.3425809443,-0.1521806866,0.3253555,-0.0669247732,0.0316266343,-0.209743306,0.221054256,0.1617935449,0.214526251,0.2830664814,-0.1113334671,0.3766462803,-0.0479228683,0.2515899241,0.1977617294,-0.0995697305,-0.2013901174,-0.2185860127,0.1886230856,0.1277872771,-0.6358430386,-0.0538012832,-0.0609726347,-0.1685434878,-0.3811208308,-0.3757064044,-0.1014486775,-0.1010174826,-0.1352425069,-0.1513607055,0.1787613928,-0.3350163996,0.1010795608,0.0664991736,0.037248496,0.2274393886,0.0955426246,0.1029124409,0.2206829786,-0.4990333319,-0.1053616852,0.0160875209,-0.1888549775,-0.0162346847,-0.0325364172,0.2466262728,-0.0158617459,-0.0560192503,0.1580464989,-0.1570668668,0.5400379896,0.1700801402,0.0842136517,0.3933697343,0.5855426192,0.0929541662,-0.0040577799,-0.0056776549,-0.5385399461,0.2322417945,0.0733196586,-0.1784122437,-0.1894301772,-0.4930152595,0.0014649347,-0.0609924719,0.0312800631,0.2731887996,0.4516296387,-0.3899906874,-0.2442983538,0.2423392832,-0.0518201552,-0.0440993942,-0.0485079139,-0.278729707,0.0635427013,-0.0296433624,0.1907825172,0.0879457742,-0.6552608609,0.3337691426,-0.1572265476,0.256819278,-0.2171640247,0.0846604779,-0.2342718989,-0.0140116513,0.0292475577,-0.055336684,-0.3133440316,0.3256939352,-0.2782121897,0.0963667035,0.0749984458,-0.0507273413,-0.4088027775,0.0596363954,0.1178674996,-0.1385251433,0.0586275198,-0.1454676092,0.205530867,0.3106637597,-0.0517417118,0.0607756674,-0.1468340755,-0.1884095222,-0.1321550906,0.5624065995,-0.02965164,-0.2255628407,-0.0499477759,-0.0330220945,0.0214515924,0.360920608,0.3322623372,0.0474140309,0.1374679953,-0.1043959707,-0.1068220213,0.2952225506,-0.3557236493,-0.1823655665,-0.1604952812,0.0411742367,-0.1827295721,0.0532729402,-0.14260526,-0.0261133779,-0.1544633806,0.4189432859,0.0851458907,-0.1766564548,-0.1190728992,-0.4814083576,-0.3502425551,-0.091270633,0.0589542314,0.5071213245,-0.0762533695,0.1716108322,0.2642099559,0.0554546081,-0.0777342767,0.0007423353,0.1357578784,0.2604104877,-0.2453193814,0.1736738533,-0.1839156747,0.19437778,0.2270919085,0.1480082572,0.1326930821,0.0645504296,-0.0957873538,-0.3693134785,0.0117412331,0.0408739001,-0.0906421617,0.1942002624,-0.0236327965,0.020072462,-0.0462615639,-0.1779586524,0.0190616455,-0.6478570104,0.0577800311,0.2421558648,-0.0119237667,0.0095317839,-0.1117237285,0.0777575672,0.3475860655,0.1864814013,0.0795118362,-0.1525591016,0.3241609931,0.1469948739,-0.2106697559,0.0685734451,-0.1213834658,-0.0259960033,0.3158133626,-0.0452680923,0.0832680464,0.0991952717,-0.1209597886,0.2441323251,0.0936465636,-0.0794048607,0.1794829667,0.2946690023,-0.0805851817,-0.0727350041,0.1661313027,0.1495787203,-0.1780600399,0.1724616438,-0.1364564002,0.0664393827,-0.2640105784,0.0271362178,-0.2498497814,0.4172476232,0.0708090067,0.0645807236,0.1937785298,-0.025959827,-0.0589471012,-0.0350064449,0.0832065344,0.2039912492,-0.0499570891,0.2260341942,0.0716900676,-0.166817978,-0.1371464133,0.0104705878,0.0662441626,0.2401139736,0.3439278305,0.1682401597,-0.1599610299,-0.2710843086,-0.161084801,0.2464816123,0.2385196686,-0.1506380886,-0.0568524748,0.0756938085,0.1517330408,-0.2876555622,-0.2880500853,-0.253351301,-0.1597195268,0.0883836225,0.2925890684,0.257884413,-0.0014518611,0.3952098191,0.4617116153,0.0593588874,-0.0513374992,0.0502368845,-0.2153857648,-0.3148036599,0.088668257,0.0649697036,-0.101613991,0.2333034426,-0.1021954417,-0.194899261,-0.3437841535,-0.4103054404,0.0100563504,-0.2103194147,0.4719873071,0.1816637367,-0.004250186,-0.3588891327,-0.1189134493,0.1462681592,0.1211081371,-0.039833039,-0.019685775,-0.1493832171,0.0391473398,-0.0708142668,-0.2220977545,0.1640029252,-0.3041469455,0.240951851,0.0763138384,0.0306515042,-0.0612248778,0.0523096174,0.5045964718,0.5048177242,-0.0379235446,-0.2633743584,0.0904082954,0.1380197257,-0.336689651,-0.3102604747,0.0216101333,0.0162846167,0.5390122533,0.1647621095,-0.4579927325,-0.31728971,-0.0815850124,-0.2054210007,-0.2434681058,0.1044477895,0.1045603976,0.0453711972,-0.0166840106,-0.1219532117,-0.1212315559,0.0335266553,-0.2295121402,0.6499086022,-0.2761822343,0.2803352475,0.3270616829,0.863853395,0.1714121252,-0.0271327849,0.0819885209,0.1208694801,0.0416551083,-0.0073013082,-0.0545647815,0.627938509,-0.294133991,-0.1778904647,0.149763301,-0.1594050229,0.2452199757,-0.0572215132,-0.0218926091,0.1148412153,-0.1691703349,-0.1525198966,0.1420355588,0.2261671424,0.0844101235,0.0116431694,-0.2870468199,-0.0439699218,0.0260114204,0.2201598287,-0.0691902041,-0.0089998869,0.2485221773,0.0041998685,-0.5446138382,0.4457467496,0.2141600996,0.1316138059,0.0335004516,0.1320725083,-0.0986372977,-0.1617611498,0.2219135165,-0.2070658654,0.0413911864,-0.0388574675,-0.0824468508,-0.5960401893,0.031793341,-0.0619527288,-0.2877740264,0.4565431774,0.3751793206,-0.2594050169,0.1372561604,0.2299483269,0.2474470586,-0.0048910626,0.0177282467,-0.5006985664,-0.4096999466,-0.1569185108,0.1279593259,0.1380255818,0.2855314016,-0.0317522436,0.0416396111,-0.1645378172,-0.1756142974,0.0485698655,0.2861667275,0.0373402871,-0.098820813,0.1486673206,-0.193886742,-0.3395846784,0.2005548924,0.1217000261,-0.0412944481,-0.9286247492,0.0397423431,0.02721956,0.2492156178,0.1350886226,-0.1588994712,0.2074218392,-0.1697260588,0.4346753657,0.0744932592,0.2771122754,0.1609030366,0.3366521001,-0.0105164517,-0.1253431886,0.3105500042,0.2530325353,0.2215178162,0.5469456911,-0.0915648192,-0.1180723384,0.1247878373,0.615729928,0.6351805329,0.0195631701,-0.2783786058,0.3620405197,-0.0484780371,0.4466259181,-0.4077373147,0.4081354737,-0.3001525402,-0.5378539562,0.0986024886,0.0166891702,0.0294555742,-0.2506218851,-0.1074840575,0.3304620385,-0.1585827768,-0.2208053917,0.0330511443,0.0757257715,-0.3909953833,-0.081369929,-0.0159061309,0.1809707731,-0.1984898597,0.1530699432,-0.2391128689,0.2261513174,-0.0269033071,-0.0777362436,-0.2188996971,0.000869909,-0.4559265375,0.1823392808,-0.0668027401,-0.0695174485,0.0376175493,0.2896275818,0.8918440938,0.1306747496,-0.0776182339,0.1159057617,-0.0147266211,-0.0918246731,-0.0855847299,0.0643571392,0.2934022248,-0.261760354,-0.1687386483,-0.0811507106,0.092575945,-0.1607918888,0.3625639081,-0.0467103682,-0.0748684406,0.1506672204,0.0202170517,0.0266314764,-0.0678415149,-0.1730990261,0.1156159267,-0.1631360203,-0.2514810562,-0.0116507262,0.0139606968,-0.3538233936,0.0769196376,0.1848153025,0.0963559747,-0.1526756734,0.4362220168,-0.1270215362,-0.0689057633,-0.0826517642,-0.4828763008,0.0940786079,-0.0615820624,0.3277035356,-0.039193999,-0.1508189291,0.1088475659,0.8712249398,0.2959722877,0.1533786952,-0.2297155708,-0.3328505754,-0.3247037232,0.0144759631,0.0052403528,0.0989413634,-0.1632369906,0.3649547994,-0.1145974398,0.3088021278,-0.24895899,0.1180856973,-0.4077550471,0.1190058738,-0.147776112,-0.1600040644,0.300770551,0.3664518595,0.0113631813,0.4424424171,-0.2432103455,-0.1815048158,-0.1888818741,0.1199714541,0.2367853373,-0.0373019725,-0.1552311927,-0.0459340885,-0.3102086782,-0.1767927557,0.2119247913,0.2571959496,-0.0661092475,0.0637439564,-0.2494707853,0.4223911464,0.2251355052,-0.2963390052,0.3116740584,0.2022940814,0.1305121332,0.0995247811,0.1436862946,0.0720312074,0.0804097354,-0.2597728968,-0.0098550497,0.0085550481,0.1967402846,-0.2309919894,-0.1561208069,-0.3708268404,0.0517182872,0.3635759354,0.0074519305,-0.2741087675,0.1914914697,0.1322606653,-0.3251188695,0.1190333739,-0.2570672929,0.0539284423,0.0856841654,0.2737969458,0.004499577,0.0231398568,0.0370614938,-0.0404764377,-0.1941719055,-0.5982773304,-0.0423341915,-0.1481177658,-0.2313742042,0.2417520583,0.3802965879,-0.1006906033,0.3059274256,0.0606965087,-0.1094735339,0.2570567131,-0.6360206604,0.0924597085,0.0907637253,-0.057219103,0.309350878,0.0586567447,-0.0854963511,-0.0184701551,-0.2389745414,0.340297401,-0.3041678369,0.09762685,-0.2271204591,0.0330096297,-0.1078648418,0.00520493,0.1080260873,-0.0795881748,0.1374003738,-0.0734620392,0.2036412358,0.1140370071,-0.2236913145,0.2299929857,-0.1168181971,0.0737469122,-0.6453632116,0.5514611602,0.3634344935,-0.0982691944,0.0167626441,0.2653759122,0.0152231241,-0.187647596,0.3254674971,0.2012493908,0.4787555933,-0.0432809703,-0.0736334696,0.1241817325,0.0740473345,-0.091107212,0.2807722092,0.5041821599,0.0719717294,0.4548557997,0.0556848533,-0.2469397932,0.1096700281,-0.0671044439,0.0015848979,-0.1018987224,0.4422678351,-0.0822585002,0.0086626401,-0.3223491013,-0.1104552746,-0.3370456696,0.3438332379,0.0558212548,-0.0479788855,0.2515256107,0.0112792794,0.0834922791,-0.0346290544,0.2525554299,0.1349637508,-0.1671020687,-0.3239260018,-0.1811329424,-0.5610326529,0.2268654704,0.1335673332,0.1841498613,-0.031348072,0.1695928723,-0.2529273927,-0.1192011461,-0.0463252701,-0.1963619739,0.2509229183,0.0248075183,-0.2768289745,-0.3529919684,-0.2416246235,-0.4227740169,0.1708838493,-0.2491306812,0.009029299,-0.1536934972,-0.0019237419,-0.0455349497,-0.036319796,0.2611357868,0.3332713842,0.2374525368,0.0346155614,0.564497292,-0.0378533602,-0.0734489858,0.0158986207,0.1402173787,-0.2252645195,0.2352957428,0.1453663111,0.3947979212,-0.0356721021,-0.1783235669,-0.4293422997,0.7531892657,0.287335366,-0.226571098,-0.0793083459,-0.0411108993,0.1083262786,0.0716211423,0.3129256666,0.3638014793,-0.1406292021,0.3308031857,-0.358816117,-0.2075114995,0.5828334689,-0.4029491246,-0.2971513867,-0.0082997093,0.3402396441,0.5178588033,-0.4353549778,-0.8319962621,-0.0066721682,0.2214436829,0.1113310754,-0.2993386388,0.0849801749,-0.1665294319,0.2673640847,-0.0818054974,0.343968004,-0.0756888092,-0.121486716,0.0391110182,-0.242582947]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1725","title":"load the local dataset","comments":"You should rephrase your question or give more examples and details on what you want to do.\r\n\r\nit\u2019s not possible to understand it and help you with only this information.","body":"your guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!","comment_length":30,"text":"load the local dataset\nyour guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!\nYou should rephrase your question or give more examples and details on what you want to do.\r\n\r\nit\u2019s not possible to understand it and help you with only this information.","embeddings":[-0.1804081053,0.013499002,-0.1037141457,-0.0130341612,0.2134609818,0.1470807493,0.2700249851,0.2880049646,0.4363234639,0.0544830486,0.1162025779,0.4947062135,-0.0745658726,0.2974509597,0.2197879851,-0.1235246137,0.0534672737,0.2126086503,-0.2095779181,-0.2524624467,-0.1248621941,0.0600582771,-0.0063393991,0.115239121,0.0288788117,-0.1423294246,0.0913052633,0.5297982097,-0.1131520942,-0.3507753909,0.1977621168,-0.1289591193,0.4495811462,0.2374496162,-0.0000996339,0.2585187554,0.2192779928,-0.2029905617,-0.1397664696,-0.4289161265,-0.3890913725,-0.0182180032,0.4359698594,-0.3825985193,-0.2413374037,-0.193669349,0.1122729182,-0.1140697002,0.4548487067,0.5852871537,0.2613310814,-0.0851335749,0.0730208233,-0.016471386,-0.1200057715,0.2349961251,0.1261286438,0.5693177581,-0.0351350754,0.0236935075,0.1668702662,0.1679306328,-0.0048863706,0.1744319946,0.2951236069,0.1507116258,0.2086285353,-0.1883335412,0.0194293633,-0.0166477263,0.543004334,-0.1054640636,-0.1357060075,0.1112841442,-0.1266473085,-0.0491798073,-0.0290388782,0.2329448611,-0.1778219789,0.4197865427,0.042490419,-0.1539365053,-0.2715227902,0.3597122133,0.0952370167,0.1745562851,-0.2555059195,0.1659582555,0.1680125743,0.0616418421,-0.1496705115,-0.064267613,0.1697380245,0.283551991,-0.0429452881,0.0360108837,0.1783190072,-0.0266477074,0.0429214053,-0.0336224735,0.0597616546,0.1215231717,0.0151272807,0.4068186581,0.14132303,-0.0603359379,0.1202225015,0.1193656474,0.0246319901,0.1288490593,-0.2984278798,-0.1590562016,-0.3900526166,-0.237193644,-0.1974733919,0.0538755544,0.2985183895,-0.3917785287,0.0706147999,-0.1277719289,-0.0662080869,-0.2095964402,0.1045247093,0.459592551,-0.2042670101,0.0449056178,0.0241914894,0.183881402,-0.1363729239,-0.0699679703,-0.2046466321,0.201271683,-0.1179721281,-0.2125794739,0.0378543437,0.1141893715,0.6578570008,-0.1471209079,-0.1344776899,-0.1377892792,0.4352629185,0.025790425,0.0769679695,0.3557407558,0.177524507,-0.0939830616,0.2335483432,-0.2354285419,-0.3238989711,0.253567338,-0.4369358718,-0.3963064551,0.0372236446,0.2864255309,-0.0015867417,-0.0293876156,-0.3729645312,0.1193521321,-0.2197752297,0.0926272422,-0.2116462588,-0.0017827078,0.0263524521,-0.2408631146,0.4416387379,0.2379226983,-0.5031754375,-0.0197728407,0.0899934694,-0.2318455428,0.1753769815,0.033525534,-0.6356109381,0.5764732957,-0.0785183683,0.1443059295,0.6426224709,-0.278113395,0.0568954162,0.4205006361,-0.2087589502,-0.1774463952,0.0441889018,-0.0274734329,-0.4110879302,0.1960454583,0.2484508157,0.5545178056,0.0291259028,0.0272961203,-0.1241591722,-0.0672422424,0.008020726,0.2939467132,-0.1867227405,0.213648811,0.130880639,0.005861748,0.3009951413,-0.2669842839,-0.1808281392,0.3929591179,0.1529488564,-0.127654016,0.0471435897,0.2223428935,-0.2559236288,0.0447063372,0.0038740032,-0.227761969,0.0014641505,-0.0632902384,-0.2188643068,-0.0520210117,-0.3524464965,-0.0175709967,0.1918036044,0.0984412953,0.2418265939,-0.0412261561,-0.3753814101,0.0178711563,-0.4167240858,0.1365870684,-0.3241452575,0.0080959955,-0.1709156185,-0.0025115889,0.0867375135,-0.1733758152,-0.0086334897,-0.2545939088,-0.1253509372,0.3699713945,0.2719967663,0.1170046329,0.1235614866,0.1868063658,-0.1348803937,-0.0490514226,0.0882867724,0.1166790873,0.2536686659,0.0933782309,-0.3188493848,0.3371166587,0.0640420765,0.0670153797,0.1694770306,-0.1314997822,0.431586951,-0.1802596897,-0.2370909601,0.0997731388,0.0120466938,0.0793500617,0.1840080321,0.0215452556,-0.4251138568,-0.1393190771,0.2735140622,-0.1181033999,0.00432431,0.040310692,-0.1925407499,-0.0169023257,0.1137784496,0.2351687253,0.3781880438,0.2797128558,0.0750256702,0.0664956123,-0.0970142484,-0.1405507475,-0.0004030906,-0.0310391337,0.1065654233,-0.054551322,-0.1856284589,-0.3148079515,-0.1685457677,-0.0794031024,-0.0456368439,0.2858555913,-0.1693475246,0.1779532135,-0.1955839247,-0.0627032518,-0.338786602,-0.0970545188,0.1272835732,0.1369191408,-0.1404073983,-0.0968635455,-0.1612037122,0.0268023051,0.0185635537,0.0378132313,-0.1304678619,-0.1197356433,0.1198437437,-0.2314549387,-0.3636472523,0.1404199749,0.3530862927,0.2672512829,0.083815001,-0.4258140326,-0.137434274,0.3108015656,0.027889112,0.1481626481,0.1081307456,0.4255303741,0.0291996431,0.4161866009,-0.090715766,-0.156022191,0.3424634933,0.0791388154,-0.052951511,-0.1151214242,-0.0137034375,-0.0771543831,-0.169735685,-0.8360049129,-0.4267987311,-0.2437295765,0.2515364289,0.3813188672,0.203326866,0.1878329366,0.4112026989,0.3307701647,-0.0392426513,0.2694143057,-0.0949388668,-0.5751318932,0.2998142242,-0.3364127874,-0.3522518277,0.2509348392,-0.0576538593,0.3363949656,-0.3029782474,-0.3647485673,-0.2921612859,0.0185072403,-0.0856117755,-0.1314375699,0.1504500508,0.2429223061,-0.1419465095,-0.1863694042,-0.1153180823,-0.0638701022,0.3799186647,0.1802685857,0.0839897841,0.2681705952,0.4132495522,-0.2054721415,0.4035097361,-0.1192240492,-0.173917383,0.1824161857,-0.1665041745,0.0020543232,-0.1057960913,-0.0438924916,-0.0701341406,0.0149789806,-0.2904994488,0.2512622178,0.139680475,-0.2773466706,-0.3615431488,-0.2635118365,-0.4432038665,-0.1137110293,0.3095838726,0.0748802572,0.0761646852,-0.0310602896,-0.1522819996,-0.1269502044,0.2882063985,0.0985445008,0.4352890551,0.0914191231,0.1312288344,-0.530172646,-0.3288079202,-0.1166613623,0.1134802923,-0.1647601277,0.0812484324,-0.0170536246,-0.1916356832,0.1134064645,-0.1743624806,0.3603603244,-0.0253002364,0.1332012266,-0.0133771403,-0.1236724406,-0.1057799533,0.0122461664,-0.1068152413,0.2005384117,-0.0631161705,0.5251100659,-0.007377164,-0.292470932,0.0335680246,0.1961867511,-0.2615385056,0.0514181405,-0.3343849778,-0.2454885244,-0.3385724127,-0.037036933,-0.1053107604,0.0318158753,0.3588357568,-0.0161740016,0.0064366446,0.0948756486,-0.0071068471,0.272162497,0.2297204137,0.0950549543,-0.2394701093,0.2546639144,0.2845193744,0.103252396,0.5296412706,-0.0492236465,-0.1167807728,-0.1272141039,-0.0102473851,0.4566897154,0.0296899229,-0.1004598364,-0.128896594,0.307619065,-0.1433250159,-0.0862257034,0.4661803544,0.0271893721,-0.0027226547,0.1241117343,-0.6534739733,0.326177001,-0.0758992955,-0.0648348555,0.1800938249,-0.3001715839,-0.1833595783,0.2977046072,0.0765392035,0.6713842154,0.0618688241,0.1501227468,0.1704759747,-0.2820648551,-0.0261183307,-0.3874349892,0.002536397,-0.0941130668,0.1491040289,-0.0417420045,-0.2041797042,0.2626315057,0.4218592942,-0.1820925921,0.1042776331,-0.1022796556,0.1531568468,-0.0893079937,0.2941600084,0.2905026376,-0.1251192838,-0.4410365522,0.1859495044,-0.0732610226,0.0673742071,-0.038112428,-0.2764884531,-0.0485030636,-0.160739392,0.4327569902,0.1349504292,-0.2871324718,0.0973232239,0.0212923251,-0.1603081673,-0.3687705994,0.2899450958,0.0548296496,0.1663361788,-0.2173327804,0.055934757,-0.228734374,-0.0576282963,-0.3444230258,-0.0400699228,0.282550782,-0.177753225,-0.2714877427,0.0965823978,-0.2840976417,-0.2838330269,0.0528945439,-0.0436369479,0.1202059388,-0.041388154,-0.6282684207,0.190787226,0.043590188,0.0114523079,0.2560285628,0.1455115676,-0.3519009352,0.0299045686,0.2597539127,-0.2880240083,-0.0560317822,0.2329965681,0.0536738671,0.1473553181,0.4500331283,0.0706721246,0.0957442448,-0.3191044927,0.1130556166,0.5872125626,0.0368304998,-0.0056532822,0.1907361746,0.1515697539,0.0789794996,0.1843736768,0.1533559412,-0.171047166,-0.1717030853,-0.3371007442,-0.3974804282,0.1489287168,0.1309201121,0.2080318034,0.252841711,0.0744114891,0.1296781451,-0.0242938101,-0.4086298048,0.1230992377,-0.3641242087,0.1800055504,0.0982023329,0.1851014197,0.1832218021,0.0067085116,0.2019731104,-0.2810566425,-0.1896120161,-0.3669373393,-0.0410260595,0.0467770025,0.0708796084,0.0467658825,0.0611163154,-0.1980156898,-0.1256959587,0.1120119616,-0.0587953776,-0.1911488473,-0.0956037939,0.3177486658,0.0393730029,0.0039508766,-0.1159153059,-0.1371204555,-0.3044009209,0.0185533483,0.0795816928,-0.0725392401,-0.1318907887,-0.1469395012,-0.0299244169,-0.0011662041,0.0021769151,0.1928350925,0.0667676553,-0.2016334683,0.126426369,-0.1851962209,0.1918070465,0.0581220463,-0.1781561822,0.448210299,0.40930897,0.2715812624,-0.1289393455,-0.118810825,0.3240143657,-0.2621675134,0.116381757,0.0616545081,0.1073400825,0.1584142894,0.0267345104,-0.0146917142,0.2206787616,0.107825309,0.0270260293,0.3091077805,-0.2741140425,0.0281624496,-0.0942392722,0.3863438368,0.4005352557,0.2741338909,-0.2054328173,0.1652839631,0.1451986432,0.0937589109,-0.1912653744,-0.390320152,0.054081399,0.0497934669,-0.1352187395,0.1128084362,-0.0478174575,0.2078143656,0.0467929319,-0.1704923064,0.0072862972,0.105847016,-0.0636516809,-0.26173383,-0.6350268126,-0.0890803039,-0.1265394241,-0.014527631,0.0859211236,-0.0864754468,0.0613241978,-0.0898331478,0.0252500251,0.0654369667,0.1226252764,0.0148234135,-0.1891258657,-0.0333362557,0.1516667455,-0.0910480022,-0.0217187926,-0.0060534743,-0.100865677,0.1498211026,0.3079253137,-0.0177200064,0.0066295341,0.1473597884,-0.1967263818,-0.1637203842,-0.0085991202,0.0047013764,0.1841702163,0.1915691197,0.3222596943,-0.2432380468,0.3756039739,-0.0618229918,0.0091701541,-0.1349721402,-0.0211994983,0.3499359488,-0.0902890041,-0.110616371,0.0844852105,-0.6342903376,-0.107316643,0.3021229208,0.0595541857,0.0979305133,-0.0712772682,0.1423605978,-0.1184403822,0.6424095631,-0.0229897033,0.2567496002,-0.3226080537,-0.1697990596,-0.3066847622,-0.0373114683,-0.1120181233,-0.0475168526,0.1425908357,0.0505203605,0.2933207154,0.3095522523,-0.1405724883,-0.0703697652,-0.2856652439,0.2319448739,-0.024950318,-0.2608463466,-0.2296885848,0.407409668,0.0233912487,-0.1877525598,-0.054589171,-0.1773392558,0.1422916204,-0.1538172811,-0.1475447714,0.1166759804,0.0102822259,0.4713776112,-0.1223145202,0.2215839177,-0.1648267508,0.1500891447,0.0391686708,0.0510802679,-0.2566117048,0.0432795547,-0.2373982817,0.2677327991,-0.3237299919,0.0017964607,-0.3764790297,-0.2181725949,0.0686685666,-0.3193565607,0.0365227498,0.0752138421,-0.0215513036,-0.1368738562,0.1834613532,0.236317113,-0.0669829696,-0.1554829776,-0.2172460407,0.1010730043,0.4586862028,-0.3350116909,-0.1781712472,0.0916138813,-0.0368952453,-0.0838731155,0.030441761,-0.2697399557,0.2142856866,0.1981936842,-0.0256940313,-0.3089805543,0.4218195677,0.2658833563,0.0984549895,-0.1044294089,0.0271276087,0.0050161933,-0.0953024402,-0.4929320514,-0.40970397]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1725","title":"load the local dataset","comments":"sorry for that.\r\ni want to know how could i load the train set and the test set from the local ,which api or function should i use .\r\n","body":"your guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!","comment_length":29,"text":"load the local dataset\nyour guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!\nsorry for that.\r\ni want to know how could i load the train set and the test set from the local ,which api or function should i use .\r\n","embeddings":[-0.3174705207,0.0638200939,-0.0952593461,0.0737695917,0.119674623,0.1557237059,0.1555574834,0.3187198937,0.5008644462,0.0614467226,0.1642027497,0.4374155104,-0.1208815649,0.4828317761,0.3047558069,-0.1635474712,0.0707763433,0.2294304967,-0.2788066268,-0.2441236973,-0.1344443262,0.0736518875,-0.0855388269,0.1153904349,0.1242739707,-0.0878494158,0.068395339,0.4703999162,-0.1047726199,-0.3025941849,0.2307722569,-0.1746587753,0.4277434945,0.2090977281,-0.0001064755,0.2021758258,0.1668529958,-0.2235588431,-0.0543571301,-0.6375690103,-0.339815259,0.0122430706,0.5199456215,-0.239844352,-0.2735789418,-0.3229026794,0.0742032379,-0.2420977503,0.5297538042,0.6811374426,0.1637198776,-0.1159812957,-0.0274794437,0.0788089857,-0.1082483754,0.1993105263,0.0425917953,0.4848556221,0.2161073983,-0.1329426765,0.1207038686,0.0659026802,0.0594503693,0.1458467692,0.3084445596,0.1042212099,0.1001513749,-0.3671630621,-0.03228781,-0.0068275719,0.6595649719,-0.1519090384,-0.0672272295,0.1932344437,-0.1363980472,0.0188089032,0.0514709093,0.20400998,-0.1985253841,0.3418989778,-0.139582634,-0.303314209,-0.2920081019,0.3929930031,0.1436104327,0.0630981103,-0.1361532956,0.191981867,0.1232660413,0.0773561895,-0.0476346426,-0.1123594195,0.1810999364,0.4176240861,-0.147015661,0.0318899527,0.1204608008,-0.1583307683,-0.0212166645,0.1425637603,-0.0351912454,0.2076310515,-0.1413223445,0.2797667384,0.1994033903,0.0905409157,0.2510216236,0.2301160544,-0.0413734615,0.0451927371,-0.2475229204,-0.0928240344,-0.384180069,-0.146848157,-0.2581519186,0.1096848547,0.1985315084,-0.489693284,0.2011567205,-0.1036792174,-0.1590282768,-0.3006119728,0.1027131304,0.3576223552,-0.2637363672,0.1756728739,0.026358461,0.1759695411,-0.1818370372,-0.0355486237,-0.1697660983,0.2390910387,-0.087726891,-0.0803422928,0.0881424919,0.0737368092,0.6081708074,-0.1651269495,-0.0182265453,-0.071615696,0.5714467168,-0.0449614152,0.104512237,0.3682502806,0.1844315976,-0.1087694168,0.1584704071,-0.1697958857,-0.3091650605,0.266433239,-0.4438110292,-0.4999341667,0.0858391151,0.2154586464,0.0015287043,-0.0579149798,-0.4173059165,0.1919698566,-0.2347786576,0.0992010087,-0.2122064978,-0.0167671833,-0.0949358791,-0.1602852494,0.4525190592,0.1626813114,-0.5647460818,-0.0626615286,0.232469365,-0.1813011169,0.1460630894,0.077684857,-0.6236317754,0.5268677473,-0.0807110667,0.1067872569,0.7347121239,-0.3785433471,0.156506896,0.4080643952,-0.1055254489,-0.289902091,-0.1411564201,-0.0831240788,-0.4013299942,0.1834557205,0.2094933242,0.7176532745,-0.0399099477,-0.0396732092,-0.0849481151,-0.1355524361,0.0854758024,0.3924347162,0.0022945537,0.1123969182,0.0790324137,-0.056993667,0.2535614967,-0.2241216749,-0.2327947915,0.3431114852,0.2093584985,-0.1896333694,0.0995650142,0.1646358073,-0.2271852791,0.0807110891,-0.0943072587,-0.0907400176,0.0667656437,0.0093658278,-0.2246161699,-0.1069226936,-0.3677223027,0.0222594719,0.0560126193,-0.0122846644,0.2122127116,0.0821349919,-0.4414839447,-0.0597798154,-0.3856554329,0.1207780838,-0.3741035461,0.0258612614,-0.0800241232,0.0126402108,0.0898481682,-0.156071499,-0.0130083757,-0.3751966953,-0.1884447336,0.370516032,0.3448432982,0.1940324008,0.1215705201,0.2672254443,0.0299345255,-0.047066398,-0.0173276402,0.1069503278,0.2393458188,0.0013541033,-0.349444896,0.3390460908,0.0386656746,0.0980397537,0.2118876725,-0.1019683778,0.3982222378,-0.102011919,-0.1447942406,0.1466412693,0.0330731422,-0.0230385568,0.2643129826,-0.0116129369,-0.3277983963,-0.1617969871,0.1831050217,-0.1951005757,-0.0211493224,-0.0869825035,-0.1502008587,-0.1252111942,0.2040324658,0.1477635354,0.3505549431,0.1726505905,0.0883328915,0.1220305637,-0.0217351578,-0.0865062401,-0.0081358254,-0.0589315258,0.1279908419,-0.0403368361,-0.2419560105,-0.2958584428,-0.1434846222,-0.0562508292,0.0435062498,0.2657514215,-0.1682803035,0.1302164346,-0.1576895416,-0.0896330699,-0.339509517,-0.1717470586,0.1737675518,0.1312695295,-0.0780021846,-0.2171972394,-0.1090754047,0.017932415,0.0351296104,0.1275705546,-0.121084936,-0.1230857223,0.1192730814,-0.2470790148,-0.4121358395,0.1210517138,0.4731254578,0.3146317899,0.2221260965,-0.3679429889,-0.164971143,0.3988380432,0.0432445966,0.2520950139,0.1601738334,0.4491714537,-0.0479058065,0.2613448501,-0.074271664,-0.1335169673,0.2895405889,-0.0687779784,-0.0948221982,-0.1374575496,-0.0300958734,-0.0145067871,-0.1382264942,-0.8957135677,-0.4849814177,-0.097554639,0.2685419023,0.3610462844,0.2054995298,0.0800819024,0.3286765218,0.3222788572,0.0620690249,0.1969749331,-0.0913054943,-0.6318541169,0.1565038562,-0.3055170178,-0.3713766932,0.1807253212,0.0271573234,0.2557729483,-0.338681221,-0.4009631574,-0.4390980601,0.0541658811,0.0394946374,-0.1106428504,0.0456797965,0.1686145812,-0.2539508045,-0.0385424905,-0.0096106809,-0.0296596251,0.3398182988,0.2781490982,0.016129164,0.2597625554,0.3580315113,-0.2566970289,0.532569468,-0.0517986044,-0.1453032941,0.1072095037,-0.1643764824,-0.1630872488,-0.1720696837,-0.0392161906,-0.032288935,0.0151129691,-0.2939655483,0.2059080899,0.1479049176,-0.324929446,-0.4695692658,-0.2655549943,-0.4256272614,-0.0692084879,0.2413306981,-0.0929224715,0.0891217068,-0.0838542134,-0.1526905298,-0.0741877407,0.2159678489,0.2643196583,0.2657872736,0.047366444,0.1414777189,-0.5233097672,-0.3210381866,-0.2115696669,0.1446567774,-0.0736160651,0.1446073353,-0.0245966744,-0.1750077754,0.0957818702,-0.1868209988,0.3129841387,-0.0624143444,0.0044385269,0.0050760945,-0.2586843967,-0.0336844996,0.007562045,-0.1549841613,0.1454947144,-0.0189592391,0.5691066384,0.1021944657,-0.373019129,0.0102386316,0.1659719497,-0.329425782,0.060611289,-0.1722368449,-0.2334903479,-0.1779260933,-0.0441489927,-0.1816343665,0.0774300098,0.2729201317,-0.0770235062,0.0954923406,0.1084554493,0.058858078,0.200607866,0.1579819024,0.0900379941,-0.1235239282,0.2795608044,0.2263959348,0.0995098352,0.5160395503,-0.1208221763,-0.0932747424,-0.0980211422,0.0133100711,0.4504297376,-0.044062119,-0.0691114664,-0.1843322963,0.1564916521,-0.1846903414,-0.1419919133,0.5305557251,-0.0593716092,0.0890268311,0.1828617752,-0.7420043349,0.3443668187,0.0258721542,-0.126205951,0.3025677204,-0.395306468,-0.1399767697,0.3523108661,0.1404715478,0.672097981,-0.1475032419,0.0649837926,0.1114402711,-0.2425683588,-0.0275101215,-0.4764209986,-0.0962170586,-0.1661290675,0.147413522,-0.1081641391,-0.1775084734,0.2052083015,0.2944385111,-0.0788648576,0.211218372,0.064258948,0.1740528643,-0.1527910084,0.3718277514,0.2599757314,-0.0680958256,-0.4120949507,0.1482232064,-0.0565021075,0.0387307666,-0.0474210754,-0.175858885,0.0722910687,-0.1177069247,0.4393653274,0.1304596364,-0.3135420978,0.2166678905,0.0189224686,-0.205626905,-0.2958341241,0.4419995844,0.0425953902,0.0507640243,-0.3014822006,0.0820120499,-0.209375307,-0.0163177811,-0.2445131838,-0.0493104532,0.3662244081,-0.2017053217,-0.2798491716,0.1274431497,-0.3033840656,-0.1805817485,-0.0149253141,-0.0624018386,0.2127467245,-0.0225910973,-0.6232087612,0.2336686701,0.1806158125,0.0767596588,0.179189384,0.2113498151,-0.3314853013,0.1345971823,0.2630288899,-0.3129191399,0.0186923705,0.1660471708,0.0899796635,0.0476433188,0.4621698856,0.1369717717,0.224015668,-0.2455445379,0.1923388839,0.7654072642,-0.0014511299,-0.0823999643,0.1178187951,0.1518153399,0.1626159102,0.341876179,0.1667464226,-0.2668483853,-0.0990421325,-0.3496147096,-0.3968908787,0.1771052778,0.1989317685,0.2853882015,0.2766464949,-0.1020391658,0.1606216431,0.1117013544,-0.3113422096,0.1378338188,-0.4248327315,0.18765378,0.0617885254,0.2419247627,0.1575873047,-0.0595356785,0.0913118199,-0.3470112085,-0.2024553865,-0.255535394,0.015452005,0.0763110742,0.0362307057,-0.030080732,-0.0765472651,-0.2830148935,-0.0939805731,0.1103298813,-0.0177911241,-0.246388942,-0.130389899,0.3935235143,0.0644613653,-0.0738264099,-0.2512656748,0.0819174647,-0.2246737629,-0.0700796247,0.0667812154,-0.0086286496,-0.02282211,-0.1806571931,0.1397463381,-0.0232921597,0.0865432546,0.1199713945,0.0443628654,-0.1638139188,0.1342011839,-0.2023623437,0.1800617278,0.1675395817,-0.1533308774,0.5118133426,0.35458529,0.2145790756,-0.0705340654,-0.1711406559,0.2912952304,-0.1217342243,-0.0248769,-0.0000001922,0.1463106871,0.1166583821,0.0346997194,0.0099092629,0.105828777,0.1709627658,-0.0028781453,0.2126764059,-0.3698403537,-0.0562259033,-0.2348611355,0.3301040232,0.2876147032,0.1682388932,-0.1426960379,0.2786853015,0.1448792517,0.1635974795,-0.1563027352,-0.5004987121,0.0598285384,-0.030809423,-0.1984286755,0.0915015265,0.0527279936,0.366659075,0.1219547242,-0.2985166013,0.0444092788,0.0961449593,0.042935174,-0.2359731048,-0.6581156254,-0.0446993634,-0.08066535,0.0528188385,0.1473628134,-0.1456526518,0.053944055,0.0032465269,-0.0297366343,0.200325489,-0.0222262833,0.1361423433,-0.2343276143,0.0038962404,0.2026712149,-0.0249816123,-0.0790950805,-0.0592225716,-0.1384842694,0.0697316676,0.274800688,-0.2765672207,0.1134473309,0.2999899983,-0.1086355969,-0.0660689175,0.0128892027,-0.0467650928,0.1701362282,0.1916871816,0.2317947,-0.1066219658,0.3731818199,-0.1648183763,0.1058882624,-0.1258361936,0.0950122699,0.4501927793,-0.1058308035,-0.1437921971,0.1536426246,-0.5023991466,-0.0623067878,0.3093746006,0.057024993,0.0898620412,-0.0182288587,0.0892705098,-0.1230484322,0.7096775174,0.1012094542,0.3367557824,-0.2881675661,-0.2534313798,-0.3108585477,-0.030583268,0.0201083217,0.014214295,0.157341972,-0.0195713062,0.3890890181,0.3380630612,-0.2163644433,0.0089401882,-0.2175746262,0.2482001036,-0.1180365384,-0.1989140064,-0.2724533677,0.3911563158,-0.0053832894,-0.1439172626,0.0063834745,-0.2344978452,0.0228953231,-0.0816471055,-0.1802420318,0.1973019242,-0.1116226315,0.5349857807,-0.1301765144,0.3435780406,-0.1005773768,0.3551087677,0.2129983306,0.0634916052,-0.13225317,0.0219805837,-0.3497194946,0.2116931081,-0.2870812118,0.0299692899,-0.3312669694,-0.1925327778,0.011209622,-0.2979670465,0.0358242691,-0.0981777683,-0.1259828955,-0.1631266326,0.2314461917,0.3694643676,-0.0616620742,-0.198059842,-0.1590631455,0.2030383945,0.3930237293,-0.3736936152,-0.097924076,-0.0567450598,-0.1206712648,-0.0557391495,-0.0354730152,-0.4181663096,0.1017340496,0.2013045698,-0.0511194617,-0.2396304607,0.4971449673,0.3414482772,0.2352102846,-0.1638088524,0.1114085615,-0.039151527,-0.0278404858,-0.5796435475,-0.40022403]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1725","title":"load the local dataset","comments":"Did you try to follow the instructions in the documentation?\r\nHere: https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#from-local-files","body":"your guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!","comment_length":12,"text":"load the local dataset\nyour guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!\nDid you try to follow the instructions in the documentation?\r\nHere: https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#from-local-files","embeddings":[-0.1453462094,-0.3470217288,0.022234248,0.1911160946,0.2607083619,0.1140020341,0.1933437884,0.1934901327,0.466264993,0.1144998446,-0.0479661524,0.2722287476,-0.034057986,0.5225479603,0.3441549838,-0.1946159154,0.0359846279,0.2158504128,-0.2649086118,-0.2357825786,-0.1470328569,0.244349122,0.0113172587,0.1472212523,-0.1752158999,-0.0418445282,0.0911264494,0.5804826021,-0.166978687,-0.280000627,0.2809688747,-0.0868198872,0.396371007,0.2903326452,-0.000109439,0.2398896962,0.233813107,-0.2708652914,-0.2277002633,-0.4597376883,-0.2061599791,0.0304177664,0.4435914755,-0.1756745428,-0.3197599649,-0.1558015049,0.0680667907,-0.0785403103,0.5432766676,0.4273916185,0.1822501123,0.2234005928,0.1105717346,-0.0797806531,-0.2054338902,0.4039157927,0.100451313,0.6277890205,0.0081645306,0.0039712503,0.1498622,0.2108730525,0.0980664641,0.2172358185,0.4214543998,0.2261766642,0.0184625294,-0.0458845459,0.0527738817,0.152033329,0.3569372594,-0.0462160967,-0.1936454028,-0.1396294832,-0.1327332109,-0.1926796287,0.1483157426,0.2192565948,-0.1969624162,0.4014127553,-0.1555713266,-0.2772049308,-0.1823046654,0.3683601916,0.1637546569,0.0509629995,-0.3642439544,0.1491610706,0.3733197451,-0.0087094586,-0.3839467466,0.0012047609,0.0473694317,0.3499691784,-0.0186558198,0.1015383527,0.2201512605,0.1926757395,0.2650075257,-0.0173205864,0.024597168,0.1043767259,-0.0718293935,0.3011573255,0.2146097124,0.119421564,0.1520464569,0.0359592363,0.1485102177,0.3246612251,-0.1091430187,-0.2137137651,-0.246964708,-0.1684232652,-0.3444881141,-0.1293779165,0.3293631971,-0.3820783496,0.0323192738,-0.1587743163,-0.0208543036,-0.2698319852,0.1419846267,0.5227510929,-0.1640101969,-0.0393162444,0.0863955691,0.2580112517,-0.192248702,-0.1100848764,-0.2268826514,0.1251911074,-0.1289055049,0.0144166239,0.072680153,-0.1454752088,0.6424365044,-0.1277283877,0.0500531532,-0.1998793334,0.0547050871,0.1299834847,0.0114816139,0.2818613648,0.1127066016,-0.0273118559,0.2634924054,-0.3025127649,-0.2963536382,0.0712348297,-0.3521436453,-0.4426988363,-0.0100415358,0.1204129085,-0.1282891184,-0.0226432364,-0.5823689103,0.212364316,-0.2177937478,0.1371238381,-0.191983968,0.1538591832,-0.204830125,-0.0787404776,0.5899089575,0.4426854849,-0.3950515091,-0.2615448833,0.2469842732,-0.2704396248,0.0460510142,0.2403369844,-0.5539281964,0.5161659718,-0.2365138084,0.1269730628,0.4211646318,-0.2930454612,0.0604894347,0.340433836,-0.2194057554,-0.058250878,-0.0250457693,-0.152640596,-0.2432838976,0.2552218735,0.1957429796,0.554099381,0.0346990004,0.0850930139,-0.1420802921,-0.1411299109,0.0718344748,0.2870236933,-0.2639866173,0.2100425214,0.0244196691,-0.0783973783,0.2911156416,-0.2232508659,-0.1258099228,0.405749917,0.1453983039,0.140332669,0.0015908768,0.2107576132,-0.430023402,0.1077738479,0.0806182697,-0.0413833335,-0.0102151781,-0.0829547942,-0.2337580025,-0.0193435811,-0.3916427195,-0.1238774583,0.0134249255,0.1155083254,0.3146415055,0.1521456391,-0.468041569,0.1737809181,-0.3747709394,0.2681185603,-0.5279142261,0.136931479,-0.0221593622,-0.067413412,0.0568270497,-0.0333951116,-0.0221041627,-0.3703499436,-0.0725660026,0.3628904521,0.1158201173,0.1557946801,0.111487098,0.0911929458,0.1085848212,-0.1205694675,-0.0553201586,-0.0216714125,0.1545681953,0.0570079274,-0.3006932139,0.3810212314,-0.0049264031,0.092738539,0.092989862,-0.1207052395,0.400478214,-0.1091704592,-0.2197211087,0.0411348678,0.1276531965,-0.0278721042,0.278883189,0.0445352048,-0.5119220018,-0.1741515845,0.319742322,-0.1400787085,-0.0627249703,0.1279806048,-0.2202716768,-0.0032301974,0.1164680123,0.0056570265,0.3897715509,0.1671112031,0.0209351685,0.1221181229,-0.0496256128,-0.1274874359,-0.0617330596,-0.1287638545,0.075555034,-0.0605109483,-0.2211132944,-0.3215992153,-0.4150203168,0.0716864765,-0.0819521546,0.2935137451,-0.3145927489,0.1522689909,-0.1523322463,-0.1410442144,-0.3415911198,-0.1535202265,-0.1542386115,0.1353666037,-0.1478597224,0.0648861527,-0.0186048467,0.057029739,0.1806038618,0.1607768238,-0.1263403893,-0.1343875527,-0.0507793501,-0.1319593638,-0.3666535318,0.0101123108,0.3015169799,0.1980042905,0.1240464374,-0.5420706868,-0.1298703402,0.1758740544,0.0395017751,0.2042912841,-0.0816068351,0.4857823253,0.0944137648,0.4809831083,-0.1660915762,-0.0953786522,0.5108160973,-0.0624086149,-0.204104811,-0.1076453105,0.0424318984,-0.0233213995,-0.203183338,-0.5090081692,-0.3252232969,-0.2274017483,0.4473901987,0.2764441967,0.0773983896,0.1500327587,0.3345221579,0.3666883409,-0.3492247462,0.2255443633,-0.1548617333,-0.7672072053,0.2952209115,-0.3071957529,-0.4336892664,0.2243738025,0.0956966951,0.18426162,-0.3938052058,-0.3623451293,-0.3909476399,-0.0120475572,0.0237870254,-0.0786601752,0.2895843387,0.2603645325,-0.2523155212,-0.0274639633,-0.0257374253,-0.2899149656,0.3447400033,0.1922237128,0.1082056314,0.2604761422,0.4943681359,-0.1869831234,0.5733231306,0.010439558,-0.1266411543,0.3176485896,-0.2601931691,0.1515416652,-0.2062468827,-0.2334638238,-0.0035042807,0.0819673017,-0.1310916841,0.2624558806,0.3411488831,0.1210638583,-0.3790666163,-0.2581587732,-0.3491064608,-0.170662716,0.1357677281,0.285528332,0.0042417217,-0.0690998882,-0.1017621383,-0.0751715228,0.1689835787,0.1194176748,0.555336535,0.22153081,0.2733727098,-0.5179000497,-0.2235816866,-0.2837498486,0.1995506734,-0.1615705192,0.003184878,-0.0307731535,-0.1474719346,0.2521977425,-0.0443693101,0.5556643009,0.2218683064,0.060778264,-0.2066012919,-0.4433993399,-0.1838087142,0.0496083014,0.030931551,0.0510961376,-0.0302867871,0.6536297202,-0.200737372,-0.4559596181,0.0190527085,0.2032952607,-0.284339875,-0.0316281281,-0.4557517171,-0.2571879029,-0.3925388753,-0.0415029228,-0.0691491365,0.1719130129,0.4030075073,0.0809027255,0.1059521586,0.0777115002,-0.0177932344,0.1583301127,0.2793427408,0.1406480521,-0.2168688327,0.5062263012,0.3529477417,-0.0099582886,0.6595923901,-0.0706750527,-0.3051676154,-0.1617464572,0.0911218822,0.4607324898,0.0482373945,-0.096582517,-0.0412311554,0.3654061854,0.0280257594,-0.3211870492,0.4184384644,0.0825174376,0.0892678797,0.0919140279,-0.527887702,0.3945811987,-0.0978538319,0.0635764003,0.2180872709,0.112043485,-0.1424653232,0.1704485416,-0.1316114068,0.6071215272,-0.1382737011,0.2214168757,0.3111868501,-0.1309792399,0.2683862448,-0.3711282313,0.0009705237,-0.0779132545,0.1819912195,-0.0808159709,-0.1415221393,0.1963390261,0.1580586582,-0.0245329905,0.1989405304,-0.0089214854,0.2169980258,-0.1365683377,0.2786263227,0.2268244177,-0.2685544491,-0.549117744,0.1049085781,-0.156612888,0.357322365,0.0034738658,-0.2391430885,-0.0187522359,-0.2577875555,0.3165684342,-0.0396370627,-0.3076092601,-0.0067072171,0.2887598872,-0.1780775636,-0.3117537796,0.4048936367,0.1305091828,0.0649142712,-0.2982255518,-0.045327168,-0.3615796864,-0.1440385133,-0.4613817334,-0.0183729082,0.2284081876,-0.1081633344,-0.2046048343,0.048833359,-0.1856689006,-0.3117564917,-0.0522256009,-0.0205251276,-0.1512332112,-0.1230936423,-0.59363693,0.1033198908,0.2119564116,0.0057710926,0.1331538707,0.2483436763,-0.3972241282,-0.1600959599,0.3897393346,-0.3120038807,-0.1183027327,0.4089417458,-0.2191824913,0.1043455973,0.3885578513,0.2111161053,0.0430123731,-0.1595385522,0.1500599533,0.7332829833,-0.2098918855,0.0127887251,0.3062122464,0.1612653732,-0.136511147,0.359790206,0.2143021673,-0.3800876737,0.0545267053,-0.3379921615,-0.4831695557,0.2997219265,0.0406364091,0.1000338122,0.3465123177,0.1781018972,0.1067534462,0.0048626182,-0.2843385935,0.2122462243,-0.2949517071,0.0184027683,0.3599720895,0.121231474,0.337693572,-0.0283540431,0.1033997536,-0.2076489925,-0.2724737823,-0.2405893505,-0.0856820568,0.1085614488,0.0836709887,-0.0284878388,0.0322016552,-0.1020244509,-0.105441168,-0.0361499749,0.1331737787,-0.0200679824,-0.0452434719,0.2705665231,0.0427377932,-0.0725848526,-0.0658800453,-0.1249182001,-0.1446237266,0.078696616,0.0928228945,-0.2452651262,-0.1898887902,-0.0901352391,0.002807356,0.1001383513,0.1445733458,0.0435095951,0.0588259473,-0.2417633384,0.0531949289,-0.145588547,0.375410229,0.104545854,-0.2136198133,0.4729065597,0.4415872395,0.1291360259,-0.1275567114,-0.1362484396,0.3809289932,-0.1390684843,0.0454579107,0.1535132527,0.0608339347,0.182679072,0.0420935303,-0.0250845104,0.1790314019,0.1763170809,0.0027925125,0.1263610274,-0.2622361779,0.0683934689,0.0795104802,0.4236169457,0.2734794319,0.2003069818,-0.214183107,0.1145165265,-0.2470863312,0.2649225593,-0.0992758945,-0.5357871056,-0.0206572842,0.0300977714,-0.1087223366,-0.049743887,-0.2507770658,0.6559771895,-0.1222760454,-0.2105812579,-0.1127796322,0.2564949095,-0.0913335457,-0.2916359901,-0.3873142898,-0.1165347025,-0.1082920134,0.0892765522,0.0339680053,-0.2669191957,0.1163126603,-0.0897280797,0.0277857259,-0.0541113429,-0.0747942477,0.0937999338,-0.1379360855,0.0325337984,0.2412899435,-0.0684690326,0.0272041317,0.0190420765,0.0438681543,0.0469786488,0.2050960213,0.0593110025,0.0277937055,-0.005310256,-0.1001337022,-0.1118503064,0.0753195509,-0.04458379,0.1161602885,0.2296295166,0.1812197417,-0.165260002,0.2346564084,-0.2212786973,0.1488019824,-0.2179717124,0.0610593222,0.0287822653,0.0555315353,-0.252081275,0.0363155492,-0.643212676,0.1082834825,0.1892987341,0.0866571739,0.1252983063,-0.1138160005,0.0787985548,-0.0820136815,0.5496039987,0.2376953512,0.2364974767,-0.3544034958,-0.2063646466,-0.3934609592,-0.0461337492,-0.0007497586,-0.0020715336,0.0986177102,0.0681926087,0.2558981478,0.3108035624,-0.1359741688,-0.1372091621,-0.0843766928,0.089869976,0.0082932869,-0.2326359153,-0.3417146802,0.3371090889,0.0699764788,-0.0518912002,0.0929391906,-0.209678486,-0.0123569677,-0.115368776,0.2041124254,-0.2135308832,-0.165798068,0.3280217648,-0.0699301735,0.2198837698,-0.0970939621,0.1408244222,-0.0059448769,-0.0209306683,-0.2299098819,0.0376855433,-0.295909524,0.289952904,-0.3802650571,-0.1956190318,-0.4983512759,-0.0645446703,0.1687581241,-0.2782407403,-0.0586112738,0.0925109312,-0.0721440241,-0.2072758675,0.400375694,0.3387805223,-0.1711653024,-0.1341584474,-0.1463594288,0.1639472842,0.6138671041,-0.3879590631,-0.0770295188,0.1096496508,-0.071619451,0.2233008593,-0.3001941144,-0.5353482366,0.258812964,0.163815394,-0.0011104099,-0.2581618726,0.4627589285,0.08688391,0.0383631624,-0.0694983602,0.2210990041,0.1281414777,-0.0565458983,-0.2089571506,-0.3570669591]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1725","title":"load the local dataset","comments":"thanks a lot \r\ni find that the problem is i dont use vpn...\r\nso i have to keep my net work even if i want to load the local data ?","body":"your guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!","comment_length":31,"text":"load the local dataset\nyour guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!\nthanks a lot \r\ni find that the problem is i dont use vpn...\r\nso i have to keep my net work even if i want to load the local data ?","embeddings":[-0.2262580544,0.0726394951,-0.0507601015,0.0298312698,0.1185123622,0.063802816,0.2788845003,0.2106349766,0.4118477702,0.1336882561,0.1261633039,0.4746668637,0.0714066476,0.4720021486,0.3170851171,-0.0984342843,0.1031512767,0.2740089297,-0.1861927062,-0.2366859019,-0.1793768257,0.0070895357,-0.1479733437,0.0982294828,0.0665092617,-0.1281439811,0.1553800553,0.4630153775,-0.0324669965,-0.3530551195,0.3163314164,-0.1135789379,0.5149089098,0.2220309228,-0.0001110482,0.3156653047,0.2686825693,-0.1798223108,-0.176581502,-0.4398923516,-0.3317746818,-0.0825264752,0.5880652666,-0.3125265539,-0.3187110126,-0.1652936339,0.1679867506,-0.1342593282,0.4851852059,0.5206210613,0.1277266443,-0.0649732426,-0.0465647876,0.0319498666,-0.2079118788,0.0409767069,0.0914772376,0.5332658291,-0.0358720645,-0.0306603406,0.2445849627,0.2169679403,0.0101462267,0.1108647063,0.4078392684,0.2312994301,0.1855111569,-0.2728904784,-0.0108615942,0.029412169,0.6502358913,-0.0345303304,-0.0615688451,0.1497109979,-0.1731130332,-0.0593592636,0.0818984658,0.1594633311,-0.1583940238,0.4750199914,-0.0652493611,-0.267205447,-0.3562138081,0.5089408159,0.1050527841,0.1424237788,-0.2099211216,0.1786520481,0.2246071398,0.0834832639,-0.0239068996,-0.1044606417,0.2395540923,0.3202842772,-0.0881080478,0.0494631305,0.1082440168,-0.0218074657,-0.0792380944,-0.1370923668,0.0832048506,0.1110032201,-0.0430995561,0.3450957537,0.2440096289,-0.0805026665,0.0778989792,0.114690356,0.0827255473,0.0343719237,-0.1902605742,-0.1788253337,-0.4520168006,-0.2575865984,-0.2210620344,0.0694191158,0.388713181,-0.4943304062,0.2136765569,-0.099729538,-0.0802950338,-0.1920658797,0.1916211247,0.3994500339,-0.3089072704,0.1041301861,-0.0222692955,0.1574257612,-0.2376615405,-0.0765827894,-0.1493881643,0.0304349363,-0.0418887474,-0.1859297305,0.0700737759,0.1622484475,0.4741714597,-0.1978305131,-0.0110634416,-0.0682446361,0.398191303,-0.0363114998,0.0524886884,0.4357510805,0.1996600628,-0.1175832897,0.1381771415,-0.2458844185,-0.3373399973,0.3195278347,-0.4019109905,-0.4656979144,0.1041049361,0.1509398669,0.041095823,-0.115384005,-0.3019272387,0.1050076112,-0.2668136954,0.031799905,-0.2557733655,0.0546940342,-0.1697489619,-0.1358542591,0.5634297729,0.3151148856,-0.5412475467,0.0201918539,0.0491371863,-0.2760152817,0.2094730139,0.1329900771,-0.6482594013,0.5789049268,-0.0216294937,0.0538205467,0.6857836843,-0.3355177939,-0.0480752476,0.5132650733,-0.2959933281,-0.3279221356,0.0262847431,0.0608985461,-0.3787731826,0.1456359327,0.2567851543,0.7998130918,-0.0743965954,-0.0323773958,-0.1378634274,-0.0969179943,0.146775648,0.3137151003,-0.12660034,0.2838522494,0.1752076298,-0.0933167636,0.1890191734,-0.1494246274,-0.2203900069,0.3022560179,0.1621784121,-0.0572287962,0.0590971932,0.350907445,-0.3783249855,0.0031324769,-0.0467180498,-0.3301528394,-0.0039017866,-0.110073477,-0.1614359766,-0.1613136828,-0.3629112542,0.1327986866,0.0395569615,0.0078608813,0.2077229023,-0.0180349424,-0.3729109764,0.0122258915,-0.3724631071,0.1111291572,-0.3929052353,0.0575090796,-0.1297760606,0.0845783204,0.0042424733,-0.115063116,0.0015265071,-0.3098142147,-0.1387270093,0.3397436142,0.1583430916,0.2285282612,0.1951051652,0.1558132321,-0.0385021828,-0.2769066691,0.0877685845,0.0464889258,0.2177213877,-0.0373213068,-0.3309574127,0.3192175925,-0.1494605243,0.1390111595,0.0939441696,-0.1213916913,0.4528521299,-0.1952369064,-0.2817572057,0.1229007468,0.0119883381,0.126000613,0.2017009407,0.0835389122,-0.4605388343,-0.172970295,0.1462798417,-0.1816798747,-0.0034767515,0.0519896038,-0.0585378073,-0.0531862155,0.1458414048,0.1503778994,0.4096986651,0.2233097106,0.1376297027,0.2141470611,-0.0171461683,-0.1415855438,0.055437699,-0.0227767769,0.1109442338,-0.0638661087,-0.2919035256,-0.2469857484,-0.118993789,-0.0623462833,0.0299576428,0.2335624248,-0.187639758,0.2312341779,-0.2249925286,-0.2441207021,-0.2344409078,-0.1185711324,0.111547336,0.2049376816,-0.1599368602,-0.1388336271,-0.1216986924,0.089702487,0.0070075262,0.1787019074,-0.1082115397,-0.2284968793,0.1359652579,-0.3116207421,-0.4153101146,0.0083473269,0.3807973564,0.2456946671,0.1344306916,-0.4152359366,-0.1526008099,0.3851387501,0.0807947367,0.170309484,0.2366387248,0.4177155495,-0.0022947465,0.4468155801,-0.1556139141,-0.0613436252,0.3115589321,0.0588511601,-0.1194642633,-0.0135620506,-0.0880732611,-0.0143259335,-0.1511420012,-0.8512554765,-0.63345927,-0.2101437151,0.3240525723,0.3335136175,0.2899512649,0.2690646052,0.348797977,0.3145422935,-0.1079902053,0.1839396507,-0.0700643212,-0.6561444402,0.261235863,-0.2274521887,-0.3461762965,0.2240776718,-0.048554372,0.3382542431,-0.3104378283,-0.4571043551,-0.2013507783,-0.0710650608,0.0937449485,-0.1810588092,0.2116857171,0.2588582635,-0.2723754942,-0.0588330254,0.0396798402,0.0404402278,0.3224676549,0.2134728134,-0.0062858616,0.3955478966,0.4657087028,-0.2823897898,0.4909546971,-0.1352150589,-0.0588798895,0.2291333973,-0.1228797361,0.0602015406,-0.1580145359,-0.0853770375,-0.0694195405,0.0933696479,-0.3961996436,0.2484788299,0.1691972613,-0.3091883957,-0.4658645391,-0.4788145721,-0.3670511544,-0.0959727243,0.2645775378,0.0205749739,0.1205359474,-0.0846663564,-0.2391330302,-0.1046194211,0.1549605876,0.0071256524,0.4453780353,0.1494165808,0.1911899596,-0.5017588735,-0.2629395425,-0.1663333476,0.0994419158,-0.2479006499,0.2519576848,-0.0161753241,-0.1006996408,0.1460195631,0.0045239101,0.4230776429,-0.0603621155,0.1707173586,0.0363348685,-0.19721964,-0.0440889597,0.0459354781,-0.1049662381,0.1431607455,-0.0270016994,0.6064021587,0.0833073556,-0.3359900713,0.0913642943,0.2119274586,-0.3059129119,-0.0439770482,-0.3621866107,-0.2126410753,-0.2512966096,0.0367479697,-0.0700359717,-0.0012102801,0.3518469036,-0.0404223539,0.0216866359,0.0369929671,-0.127925396,0.1977523416,0.1579345912,0.109594129,-0.151370123,0.3164835274,0.2592551112,0.0167331267,0.544908464,-0.0467437729,-0.1096124649,-0.1179965734,-0.0111859627,0.6491324902,0.0015017504,-0.1544339061,-0.0976325572,0.4172551334,-0.0801223889,-0.1650693715,0.347222954,-0.1255068481,-0.0769255832,0.1121483296,-0.6469174623,0.4748933911,-0.0439277515,-0.0730186477,0.2882025242,-0.3884681165,-0.1008055508,0.2121140063,0.0590956174,0.8134894371,0.0333576016,0.1971678883,0.055508025,-0.2451761663,0.0460708104,-0.4619621336,-0.0093027642,0.0029381029,0.1612439007,-0.075104177,-0.2174424529,0.2566370964,0.267046541,-0.268812567,0.1819356382,-0.1068791896,0.053467378,-0.0974062905,0.3243716657,0.3428744972,-0.1853105426,-0.3735664785,0.0440542176,-0.0843409747,0.1216725186,-0.0979245678,-0.2312125713,0.0627225041,-0.1918807328,0.6113891602,0.1241856366,-0.2540839911,0.0761305913,0.0886384174,-0.133761704,-0.3852085471,0.4123330414,0.102430962,0.1150168851,-0.2677460313,-0.03418836,-0.2237814218,-0.0154444193,-0.3711474538,-0.0428795852,0.1807942092,-0.2243052572,-0.3251643479,0.0511719026,-0.3033109307,-0.2515758574,0.0120191798,0.0173002779,0.1422309577,0.0446703397,-0.4860934615,0.1974079162,0.0500841737,0.1121004149,0.1477911472,0.274415791,-0.4298262298,0.0734211206,0.3985386193,-0.4108062983,-0.0074868756,0.3197149038,0.038926024,0.2434494495,0.4458518624,0.077825442,0.0871667713,-0.2342288345,0.2417250574,0.5168827772,0.0547065474,-0.0988067687,0.1443186402,0.2257346362,-0.0260421708,0.1810661703,0.1050979346,-0.2007462084,-0.1377864927,-0.4303857684,-0.4617102146,0.1570798606,0.1193920746,0.1880943179,0.3091334999,0.0087855561,0.1888036281,0.0080669969,-0.2785654068,0.1335347593,-0.4279068112,0.3117873073,0.08136978,0.1686210185,0.107332617,-0.0807094052,0.0980130285,-0.3277256191,-0.1721661091,-0.2063762993,-0.0310379472,0.1371995956,0.029014878,-0.0520431623,0.0060063819,-0.2138794661,-0.1051650345,0.1937743276,-0.1081062183,-0.2283717096,-0.0520228073,0.3322049379,0.0071706413,0.0336239114,-0.1595478952,-0.0090892529,-0.3278469145,0.0831670612,-0.0065641594,-0.0498455763,-0.0616948865,-0.15670003,0.0218350869,-0.1117295995,0.0842223763,0.0012764682,0.0120315133,-0.2264371216,0.2103295028,-0.2362059057,0.121560961,0.1910015196,-0.1345961243,0.4035406113,0.5213238597,0.1573079079,-0.1257504225,-0.1589361578,0.4078941941,-0.1581019461,-0.0142410034,0.0566852167,0.1140329018,0.1502876878,0.0129460162,-0.0665266439,0.1348983496,0.081034191,0.0565081015,0.3000145555,-0.2890316844,0.0230228044,-0.242492944,0.4674586952,0.2498479486,0.3675689101,-0.2711752951,0.0885182396,0.1509307623,0.2402298152,-0.255625248,-0.5581215024,0.0258302148,0.0739894286,-0.2537554502,0.0096960953,-0.1121890172,0.1194773167,0.0159299653,-0.1037271321,0.0927841291,0.1385781914,-0.063193351,-0.2561330497,-0.7088008523,-0.0451977775,-0.0946444646,0.0427377112,0.1373205185,-0.1139260978,0.0077885101,0.0951222777,0.056524381,0.1058130041,0.2204571962,0.0432744101,-0.2128089815,-0.0445867106,0.2721945345,-0.168821916,0.0205907412,-0.0004447714,-0.0594225489,-0.0186071862,0.3385345042,-0.1025634259,0.094907783,0.2701540291,-0.23610349,-0.1904638112,-0.0318167731,0.045788765,0.3163700402,0.1674360931,0.170038864,-0.1183956787,0.5580307841,-0.1314406991,-0.0622243658,-0.1964294165,0.0710359886,0.389937371,-0.1037722081,-0.1489881277,0.1841045469,-0.6802805662,-0.0585346632,0.3712248504,0.0439656079,0.0981783494,-0.0462779514,0.0590234548,-0.1247636601,0.783561945,0.0875961632,0.4212257266,-0.3404266238,-0.1409499943,-0.392806083,-0.0432755277,-0.1832234263,0.0332546085,0.2428207695,0.040797133,0.2119902074,0.5221179724,-0.1135595888,-0.0735933557,-0.2089071274,0.1800690591,0.1102128699,-0.189069435,-0.1641605794,0.5195657015,-0.039792534,-0.2209168077,0.0574469082,-0.1535334885,0.0074699675,-0.0161445849,-0.1233069897,0.0022896843,0.0641649961,0.5476078987,-0.0957919881,0.3385377824,-0.1360266805,0.2483476847,0.1215864941,0.0980734825,-0.2691836655,-0.0086150356,-0.2382548153,0.0336122774,-0.4011253715,-0.0235947538,-0.404345274,-0.221629709,-0.0262479279,-0.1667847484,0.1563258916,-0.0904121995,-0.0331187434,-0.0355156437,0.1822905242,0.0938132703,-0.1104054302,-0.2038424015,-0.2395140827,0.1443880349,0.4357283413,-0.2469320148,-0.071121119,-0.072988987,-0.0379525274,0.0096267946,0.0815483332,-0.1749923527,0.2481606156,0.2966869771,-0.0883735046,-0.1968812197,0.4825163484,0.3504326344,0.0868383721,-0.1436398029,0.0646011233,-0.0482754372,-0.0431322493,-0.5779155493,-0.4731293619]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1725","title":"load the local dataset","comments":"We will solve this soon (cf #1724)","body":"your guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!","comment_length":7,"text":"load the local dataset\nyour guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!\nWe will solve this soon (cf #1724)","embeddings":[-0.2038414031,0.054225672,-0.0868986323,-0.0048990678,0.2707055807,0.1515009254,0.3097408116,0.3507884145,0.4961802959,0.0991699249,0.1321932226,0.5409706235,-0.0459115766,0.3431528509,0.2221464366,-0.119718343,0.0668618232,0.2321132272,-0.2257097512,-0.2464870661,-0.1728466749,0.1221223697,-0.0469453558,0.0859737471,0.0184646286,-0.079463318,0.1419752836,0.4853382707,-0.1007504091,-0.3718459308,0.2451318949,-0.1296850741,0.4341570735,0.2154562771,-0.0001032796,0.2595534623,0.2810080349,-0.2127646804,-0.1753618866,-0.5067890286,-0.3740185499,0.002149045,0.4697888494,-0.3277897835,-0.2752194405,-0.221541509,0.1267615706,-0.1522159725,0.404866904,0.5725696087,0.2307304591,-0.0948033258,0.0629865527,-0.0251607131,-0.1344604641,0.2429228872,0.1088789552,0.5441483259,0.0189863816,-0.0465283096,0.1213323176,0.1916235983,0.0398011357,0.128527835,0.3260287941,0.1380524188,0.2127304077,-0.1408893019,0.0300060883,-0.02472624,0.5713528395,-0.1203400716,-0.152121976,0.1333305687,-0.1228951663,-0.0570823103,0.0448737629,0.2152130753,-0.2135710716,0.4171665907,0.0277283173,-0.1988464296,-0.2933235168,0.3913790286,0.1576211154,0.1868701279,-0.247442767,0.1577453315,0.2338227928,0.0708932877,-0.1047490165,-0.0402806625,0.142419219,0.3195075691,-0.0655132756,0.0563573167,0.1807840765,-0.0181447919,0.0283090826,-0.0811148658,0.0722220764,0.0854332149,0.0089147696,0.3851694465,0.1938206851,-0.0346281454,0.1227956638,0.1291253418,0.038576521,0.1550219804,-0.3040722013,-0.1692777574,-0.4135531187,-0.27854684,-0.2163794935,0.1076581031,0.2579717338,-0.4231984019,0.0917861983,-0.0964370444,-0.0353688449,-0.223607108,0.1285439879,0.4163063467,-0.1913971901,0.100823313,0.012033104,0.1949502826,-0.149505794,-0.0897627324,-0.2076356411,0.1631419212,-0.1458088309,-0.2780834734,0.083736904,0.1153863817,0.6433872581,-0.1550775915,-0.0447119363,-0.1236936599,0.3996960521,0.0079993019,0.0571934581,0.4125249088,0.1676218808,-0.0634000599,0.2430067062,-0.266287297,-0.3288575709,0.3021516204,-0.4176858962,-0.4239239097,0.033505097,0.2404674292,0.0038966991,-0.081285812,-0.4189507961,0.1061361283,-0.1785523593,0.1271775663,-0.2576479912,0.0062469519,-0.0462573878,-0.2221367359,0.4579757154,0.253277272,-0.5322844386,-0.0052877381,0.0773317367,-0.23335962,0.1927871555,0.0743155926,-0.6883653402,0.546947062,-0.1176820472,0.1255856156,0.619220078,-0.3028550446,0.0487546958,0.4784238935,-0.2154595852,-0.2085251659,0.0838377252,-0.0315419286,-0.4227083027,0.1602734923,0.2484181523,0.6455643773,0.0064891339,0.0423668586,-0.1437334865,-0.0929332376,0.1141686738,0.292298615,-0.1115932316,0.1936552823,0.1422335654,-0.023821136,0.3135109842,-0.2579699159,-0.2359773517,0.3482638597,0.1719633788,-0.1334350109,0.1006586328,0.2484445274,-0.2809293866,0.0213922337,0.0377992764,-0.2154787332,0.0188979935,-0.0588977896,-0.2102387249,-0.0241554342,-0.3754697442,0.0134433135,0.1433742344,0.0293895397,0.187476173,0.0000744961,-0.3977323771,0.0566640422,-0.4260914326,0.1257853806,-0.3508147299,-0.0351571068,-0.1598045826,0.0058882777,0.0465393104,-0.1656379253,-0.0063481703,-0.2488674521,-0.1850138605,0.3777573705,0.3126480281,0.1778545529,0.1257358342,0.1030340493,-0.1205982864,-0.0932772383,0.0209136456,0.0953179151,0.2259384841,0.0047897967,-0.3222871423,0.3324903548,-0.0396894552,0.1056471914,0.1445301175,-0.1152757257,0.4289567471,-0.1286872774,-0.2580963969,0.1180540621,0.0501259938,0.0780298933,0.1757314354,0.0502567105,-0.3885141909,-0.1630135775,0.2182337344,-0.0954037681,-0.0425747298,0.0630544052,-0.147015959,-0.0353831723,0.1507817209,0.2238850743,0.4021844864,0.2660897374,0.0621867217,0.0766374394,-0.1053584442,-0.1352955997,0.0387991928,-0.0461766012,0.1010964587,-0.0000637722,-0.1990274191,-0.3297381401,-0.257271558,-0.0424044766,0.0123295616,0.3108564913,-0.212078318,0.1597821862,-0.188847363,-0.12341699,-0.3404437006,-0.1532117873,0.132637009,0.1488185227,-0.1489721984,-0.0779765248,-0.132916525,0.0634744391,0.000492533,0.0537675694,-0.1296979338,-0.1478182226,0.0866450444,-0.2321622372,-0.352599442,0.0918227583,0.3911116421,0.202605471,0.055628106,-0.4311253726,-0.1603966802,0.3588344455,-0.0060349968,0.1356502026,0.1339308172,0.448702693,0.0904960185,0.3838487864,-0.0727589801,-0.1690413207,0.3616460264,0.1086265445,-0.0640557408,-0.1000129804,-0.0533317551,-0.0666963607,-0.2131697834,-0.8376142383,-0.4101791084,-0.2348393351,0.2209612578,0.3578776419,0.2371746153,0.2337986082,0.4085538983,0.2799689472,-0.0703039914,0.1954716444,-0.0768314824,-0.6186457872,0.2933449447,-0.320679456,-0.3477152288,0.2101002783,-0.0449490845,0.3457720578,-0.2890545726,-0.3395163715,-0.2928256691,-0.0160343647,-0.0458671376,-0.1613378078,0.178904444,0.1843146086,-0.1773489118,-0.1488181502,-0.1167989001,-0.0611700416,0.3771713376,0.2154487967,0.1056000665,0.2663561106,0.4507005811,-0.1610111892,0.4484513998,-0.1202754006,-0.1176436543,0.2108702362,-0.1700270921,0.0251418725,-0.1546661109,-0.0941027775,-0.0410306565,0.0634596124,-0.3532579541,0.216684714,0.1849603653,-0.3004968762,-0.3767761886,-0.2858856022,-0.3655713201,-0.0993708447,0.280489713,0.0539712496,0.0468545929,-0.0018063145,-0.1617916077,-0.1484783888,0.2882610857,0.1247719228,0.4322036505,0.1305557936,0.2059181482,-0.4905440211,-0.3296356797,-0.1492179632,0.0658879802,-0.1716968268,0.1188288108,0.0256461687,-0.2193738073,0.1035777926,-0.1266862303,0.4727346599,-0.0080042426,0.1325413436,-0.0033645772,-0.1467818916,-0.0893709958,0.0086134588,-0.0392391831,0.2079758793,-0.0934347808,0.5755704045,0.0579348765,-0.3226610422,0.0632555857,0.1721105278,-0.2629438937,0.0206975341,-0.3118627667,-0.2444629818,-0.2881639302,-0.0599713512,-0.0906306803,0.0426516756,0.3664307296,-0.0007576831,0.0363235734,0.0688679293,-0.040127825,0.2502267063,0.2321983278,0.0964546278,-0.1756259203,0.2605656087,0.2501484752,0.0638677552,0.5466179848,-0.0754972771,-0.1488693506,-0.1541844159,-0.0397921987,0.5744270682,-0.0230330769,-0.1545447111,-0.1499999613,0.2956715524,-0.1610809267,-0.1151894405,0.486277014,-0.0176073741,-0.0232728478,0.1363483667,-0.6981270313,0.3369233906,-0.0824290663,-0.066082418,0.2374458015,-0.3365598619,-0.1676529646,0.3236937523,0.0417070761,0.5782731771,0.0130545925,0.1445972919,0.1630621105,-0.2943026125,-0.0251119547,-0.3526707888,0.0256288387,-0.0968696997,0.1616255641,-0.0291939229,-0.214430511,0.2995776236,0.4118421376,-0.1609011739,0.1110091433,-0.089988783,0.1152082905,-0.1007005051,0.3086189628,0.3027780354,-0.1549457014,-0.3797754049,0.1499544531,-0.1234992146,0.1057706028,-0.0359967873,-0.2387593836,0.0204088259,-0.2353230268,0.4254660308,0.1404586285,-0.3112785518,0.1000383422,0.0670136213,-0.1681686789,-0.3328871429,0.3359849155,0.0382624045,0.1788569242,-0.2101939321,0.0756085217,-0.2140391618,-0.0297992174,-0.3432735503,-0.020124495,0.2419758141,-0.1615270078,-0.2487514764,0.0408063829,-0.2678837776,-0.3075323105,0.0101277502,-0.0258030593,0.1332224607,-0.0156717021,-0.6258201599,0.1567719877,0.090741165,0.040910162,0.2283738554,0.1717324108,-0.4015977681,0.069032006,0.2562243938,-0.338231355,-0.0359869339,0.283351779,0.0415517166,0.1264769435,0.4599137604,0.1090164557,0.100396283,-0.2999937832,0.1381928474,0.5790197849,-0.000608289,-0.0249112304,0.1976605356,0.1512736678,0.0377138592,0.16466932,0.1181291342,-0.2004817426,-0.1146844998,-0.3619681597,-0.4642760158,0.1035424247,0.069512479,0.187638998,0.3002492189,0.0161022712,0.1098736376,-0.0192163009,-0.3566132188,0.166702956,-0.3273673356,0.1851952672,0.0767071024,0.2190682292,0.1896383613,-0.0197085384,0.1846546382,-0.3165109456,-0.1580404639,-0.3185802102,-0.0672405809,0.063308835,0.0599951968,0.0558411218,0.0501285084,-0.1809337288,-0.1516857147,0.1249868721,-0.0598185472,-0.1373273432,-0.1129853576,0.3895634711,0.0220059752,-0.0249770246,-0.0656459406,-0.1200542226,-0.2383259833,0.0554542579,0.0981822386,-0.0490631647,-0.0865986496,-0.1134066433,0.0118327281,-0.0301534403,-0.0344063565,0.1578231156,0.0442098863,-0.2433008552,0.1405386031,-0.1707751602,0.2100744247,0.0534385331,-0.2004861981,0.461101383,0.4685071707,0.2462804466,-0.1675431132,-0.1882216185,0.3377271593,-0.1866745353,0.0478704832,0.0473371297,0.1426435411,0.1285919845,0.0543893613,-0.0404531583,0.2452160269,0.1106287614,0.0654461384,0.3248256445,-0.3055736125,0.0024308297,-0.1600624174,0.4238993824,0.3468108773,0.3849779367,-0.201871261,0.1566724032,0.184373036,0.0694309771,-0.207024008,-0.4491119385,0.0352660678,0.0469827689,-0.141660586,0.1049719006,-0.0326282494,0.2967346311,0.0839540735,-0.1669298857,0.0400735028,0.051792711,-0.0951498821,-0.2713370025,-0.6343138218,-0.0623518676,-0.1365952939,0.017016029,0.0592823699,-0.0841420144,0.0687728301,-0.0405268818,0.0654214174,0.0366580859,0.1109847799,0.0379140154,-0.1960859299,-0.0622511469,0.193502605,-0.128499642,-0.0437832028,0.0271099769,-0.1230853796,0.1171548218,0.3918814659,-0.023893699,-0.0342886299,0.1321674436,-0.1865619719,-0.2004938424,0.0018577098,-0.0218701158,0.2369251251,0.2202726007,0.2662012577,-0.1969264299,0.4059151113,-0.039241679,-0.007386026,-0.1391392946,0.0339628309,0.3148831427,-0.077297017,-0.0733366385,0.1244457588,-0.6063254476,-0.1259919107,0.3455734253,0.0992525667,0.1208669394,-0.092030555,0.1194716692,-0.1117258742,0.6558456421,0.0145945167,0.2494899184,-0.370770812,-0.1614123434,-0.3598178923,-0.0307780616,-0.0948671103,0.012279382,0.1825878024,0.0394251123,0.3164454103,0.318031013,-0.098215498,-0.1018864587,-0.2226641327,0.2503739297,-0.0091635035,-0.286693573,-0.1881708801,0.4086690843,0.0068215164,-0.2163504809,-0.018706616,-0.1843521148,0.1166062504,-0.1867649555,-0.0884525776,0.1031448618,0.036735408,0.4444125593,-0.1080511808,0.2516133487,-0.1472679973,0.2087511569,0.007563679,0.0037640319,-0.2719208896,0.0042841756,-0.2768445611,0.2529257834,-0.3601029217,-0.0534271188,-0.4244095683,-0.269500196,0.0419913046,-0.2898386717,0.0046215253,0.0638095513,-0.0760403872,-0.1117296219,0.155055359,0.2075773478,-0.0719845369,-0.137203902,-0.202423647,0.1047250405,0.4759759605,-0.3504706621,-0.2336904854,0.0416402481,-0.0452635363,-0.0943783671,-0.0086405827,-0.2874787152,0.1983168274,0.2188694477,-0.0076761525,-0.2868458331,0.4722772241,0.2810596526,0.121908091,-0.0831746981,0.0446498506,0.0256251879,-0.1064612344,-0.5304174423,-0.4613021314]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1725","title":"load the local dataset","comments":"thanks a lot","body":"your guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!","comment_length":3,"text":"load the local dataset\nyour guidebook's example is like\r\n>>>from datasets import load_dataset\r\n>>> dataset = load_dataset('json', data_files='my_file.json')\r\nbut the first arg is path...\r\nso how should i do if i want to load the local dataset for model training?\r\ni will be grateful if you can help me handle this problem!\r\nthanks a lot!\nthanks a lot","embeddings":[-0.2283395827,0.0291790497,-0.0767193586,0.0087452633,0.2179478556,0.1823533624,0.2907708287,0.31472826,0.4999287128,0.0756818876,0.1266063601,0.5287618041,-0.0578976683,0.3532990515,0.2550475299,-0.1098131612,0.0864253044,0.2560073137,-0.2038420886,-0.2668070793,-0.1949939728,0.0900898427,-0.0482833609,0.0950423926,0.0199891739,-0.0691284761,0.1281027049,0.4747527838,-0.0835428387,-0.3660683334,0.2602765858,-0.12783283,0.4612749815,0.2016450316,-0.000107371,0.2648452818,0.2534545958,-0.2268603146,-0.1487689465,-0.4949115217,-0.414601028,0.0009282554,0.4915297329,-0.3300980628,-0.2760000229,-0.209382847,0.1501269639,-0.1469760835,0.3892984986,0.580933094,0.186288178,-0.1390552819,0.0283018686,0.0058570304,-0.1284695268,0.2449051291,0.1391579062,0.5586284995,0.0090502547,-0.0238615815,0.1820466816,0.1785779446,-0.0027933957,0.1521695405,0.3366523683,0.1589894444,0.2292604297,-0.1413157582,0.0349478088,-0.0050848103,0.6021072268,-0.0967661217,-0.1239715219,0.1475595385,-0.1407317668,-0.0587935522,0.0277418904,0.2194793224,-0.1925540864,0.4285127521,0.0167299919,-0.2369164675,-0.2942166626,0.4479216635,0.1530097574,0.2119295597,-0.2489111423,0.1748100519,0.2557197511,0.0840142816,-0.0954623446,-0.0674636364,0.15076603,0.3615655899,-0.0831231847,0.0652424619,0.126394406,0.0208613165,0.0086405873,-0.1022748575,0.0494472049,0.0827725455,-0.0131904548,0.3943919837,0.1838697344,-0.0680813044,0.1410136223,0.1372413188,0.0216919761,0.109345071,-0.283095479,-0.1716454625,-0.4414388537,-0.2753280997,-0.2016864419,0.0744310096,0.2739488482,-0.4476769567,0.1249362156,-0.1304544955,-0.039474234,-0.2718436718,0.1360902637,0.4320298731,-0.2206375748,0.1034047157,-0.0207658336,0.1734394282,-0.1465231031,-0.0843232498,-0.1804359406,0.1631583869,-0.1071894318,-0.2330270261,0.0138387186,0.0833855048,0.6599410176,-0.156448245,-0.0898493454,-0.1430161148,0.4122404158,-0.0004692763,0.0516811498,0.4134290218,0.1660397351,-0.049957145,0.2346653491,-0.25923419,-0.330558002,0.3299708664,-0.4560597837,-0.4290554225,0.0249206591,0.2086442113,0.0099139418,-0.0902151689,-0.442469269,0.1161517426,-0.1889464259,0.118337594,-0.2359248847,0.0020380851,-0.0358762853,-0.2102108598,0.4841822982,0.2572189569,-0.5429471135,-0.021651933,0.0539686643,-0.2295052707,0.2032945007,0.0594915375,-0.6892175078,0.5835615993,-0.1062024608,0.1151389107,0.6744263768,-0.2971440852,0.0881508738,0.4880120754,-0.1844873279,-0.2362808883,0.0734461397,0.0234509595,-0.4372363091,0.1649764329,0.2362773418,0.6585780978,0.0108496221,0.0198611636,-0.0944839865,-0.0760805309,0.1342888921,0.31760028,-0.1235312074,0.2221731246,0.1382845342,-0.0009638968,0.3094764054,-0.2508766651,-0.2579938769,0.3565575778,0.1682749093,-0.1319555342,0.0947332308,0.2348657697,-0.2650983036,-0.0073467693,0.0521685742,-0.2927366793,0.0560956113,-0.0642486289,-0.2230369747,-0.0335023291,-0.3998247087,0.0537178181,0.0787681639,0.0316913836,0.1998359412,0.0007813546,-0.434904784,0.0442707278,-0.4435225427,0.1416538954,-0.3511810601,-0.0519312732,-0.1129341349,0.0166392084,0.0164242275,-0.1745299101,-0.0151798716,-0.2775452733,-0.1569091827,0.3938217163,0.3296565711,0.1746874601,0.1160082892,0.113631323,-0.0903482884,-0.1051432639,0.0337304659,0.0745236501,0.2417838126,0.0105324201,-0.3060017824,0.3151185513,-0.0530107766,0.1048526764,0.1186652333,-0.1367208362,0.4167094529,-0.1289580315,-0.2502816319,0.1133312285,0.0497606955,0.0708375201,0.210302487,0.0559354499,-0.3936013579,-0.1987313628,0.2162954211,-0.1228908822,-0.0326557234,0.062088646,-0.1708476692,-0.0178804193,0.1535255015,0.2012703419,0.4056282043,0.242888242,0.0641242638,0.0910420194,-0.0868446603,-0.1262092441,0.028197078,-0.040561974,0.0944973007,-0.0516165346,-0.2665902972,-0.3179437816,-0.2195241302,-0.0367991477,0.0037238973,0.3020810485,-0.208317548,0.1847840995,-0.1742232591,-0.1345957071,-0.3567132652,-0.182818681,0.1332014799,0.1981337816,-0.1917529553,-0.0875983983,-0.0727710649,0.0534504354,-0.0029598086,0.0366748199,-0.1462019235,-0.1620545089,0.1049748436,-0.2403212041,-0.3486929536,0.0542201549,0.3781071007,0.2253789753,0.0625184625,-0.4634776413,-0.1527397633,0.3880110085,0.0512048937,0.1396060735,0.1635353863,0.4360892773,0.0569723099,0.4024225175,-0.0844703987,-0.1651235819,0.3381739557,0.1358755678,-0.0476903282,-0.1197248772,-0.0618852153,-0.062312305,-0.1672937125,-0.8097906113,-0.4461222291,-0.2020238936,0.2055359632,0.3720391393,0.2369582057,0.1987403482,0.3999963999,0.3182941973,-0.0713810846,0.2079224139,-0.0552428216,-0.6166783571,0.3143173456,-0.2996945381,-0.300214678,0.2289743274,-0.0657142252,0.3336193264,-0.2920608819,-0.3348841369,-0.2964292169,-0.029057866,-0.0858225822,-0.1590665728,0.2217456549,0.1778475642,-0.206146583,-0.095853202,-0.0630748123,-0.0521583743,0.3619034886,0.2254378498,0.1045744419,0.3117385209,0.451846987,-0.2187470645,0.4559381902,-0.1140441,-0.1335558146,0.1985559613,-0.1834183484,0.0112726921,-0.1112362817,-0.0806511343,-0.0686599612,0.0844301879,-0.3323993385,0.2137224227,0.1707668602,-0.2857590318,-0.3984694183,-0.3272979558,-0.4090162814,-0.0659364164,0.279607743,0.023494361,0.0653739944,-0.0489494093,-0.1837068796,-0.1261299402,0.2756949961,0.1048420817,0.474955976,0.1459796131,0.2128988057,-0.5024760962,-0.3555282652,-0.0973366275,0.0941037908,-0.1871669292,0.1097113639,0.028684523,-0.2008769661,0.1131751984,-0.1113595515,0.4726494253,-0.0083884243,0.1279718578,-0.0288969558,-0.1354225278,-0.072912395,0.0010732909,-0.0538038202,0.225714162,-0.1125470623,0.5830498934,0.0672822595,-0.3254365623,0.0878929645,0.1999477148,-0.295479238,0.0222189222,-0.3318901658,-0.2400677651,-0.251701802,-0.0724585205,-0.1140824333,0.0196717083,0.4166971147,-0.0147119388,0.0131033016,0.058861427,-0.0260060988,0.2729395926,0.2241485119,0.1321263909,-0.1963882446,0.2876906097,0.2450471371,0.0539298914,0.5335607529,-0.0898738727,-0.161133498,-0.1662638932,-0.0392477699,0.6345285773,-0.0743478164,-0.1544995159,-0.119377397,0.3086175919,-0.1935121268,-0.1161682978,0.5175310969,-0.0206119344,0.0111668557,0.0982153416,-0.7415834069,0.3886553645,-0.0543809012,-0.056328766,0.2470629662,-0.3321386278,-0.1689243019,0.2899752557,0.0389188305,0.5940825939,-0.0116521418,0.1615994275,0.10658288,-0.2887981236,-0.0160909817,-0.3400627673,-0.0125712827,-0.067738995,0.1976224929,-0.0572846159,-0.2162340879,0.3102315664,0.4089546502,-0.1726314425,0.1408755481,-0.1230738461,0.1172374263,-0.0972812325,0.3101018369,0.3409784436,-0.1403201371,-0.3735281825,0.1079845652,-0.1553070843,0.1092594266,-0.0336161964,-0.2145355344,0.0484894663,-0.1986335665,0.4605781734,0.1002123728,-0.3113683462,0.0808963329,0.0509640463,-0.1867669374,-0.3361840248,0.382619828,0.0334519334,0.1616123319,-0.1994613558,0.0482797027,-0.2316723913,-0.0661372989,-0.3713699877,-0.0025269371,0.2409732789,-0.1608939469,-0.2613609135,0.0891186669,-0.2632648647,-0.299851805,0.0371492021,-0.0206249785,0.1428000778,-0.0095463833,-0.6632546186,0.1606486589,0.1038798913,0.0849905461,0.1832184643,0.1760919243,-0.4258709848,0.0764898434,0.2692532837,-0.360858947,-0.0122256791,0.2499247193,0.0408112705,0.1195889488,0.4617339373,0.079851836,0.1377932727,-0.2611419857,0.1744645983,0.5844294429,0.0421008132,-0.0510740466,0.2108992189,0.1881650388,0.0343813002,0.167251572,0.1389062256,-0.2126132995,-0.1412642002,-0.3850998878,-0.4365974367,0.1185185984,0.1002893969,0.2003492713,0.3081592321,-0.008306413,0.1233827248,-0.0024009994,-0.3011685908,0.1756212413,-0.3389260471,0.2089117616,0.1021696404,0.2561727464,0.1671501398,-0.0171112157,0.1429240257,-0.3487143517,-0.1454358697,-0.2787145078,-0.0795659944,0.0865531564,0.0570237823,0.0488975421,0.0596931912,-0.2183733135,-0.1115412489,0.1296479106,-0.063860096,-0.1731916219,-0.1041993871,0.4058419466,0.0517119914,-0.0456275456,-0.1333404481,-0.1024316549,-0.2593421936,0.0306609571,0.1044150293,-0.0297763366,-0.06277439,-0.1275870502,0.0063946103,-0.0521377362,0.021282319,0.1075159982,0.001385498,-0.2332456261,0.1676182747,-0.1941765994,0.1908759475,0.0931243896,-0.1766443998,0.497053355,0.4854447544,0.196796298,-0.144108966,-0.199904412,0.3618409932,-0.1944564581,0.0296996757,0.0570221767,0.1287029237,0.1857436001,0.0412029549,-0.0682528093,0.2324700803,0.1271361411,0.0670964345,0.3251563609,-0.2963039279,0.0065597198,-0.1531554908,0.4519001245,0.3417812586,0.3640360236,-0.2134097517,0.1590839624,0.1665029079,0.0762831047,-0.230274573,-0.4680967331,0.037524309,0.0430713259,-0.168380931,0.0956245735,-0.0257235412,0.3139333129,0.0996081978,-0.1833685488,0.0588411801,0.0577335544,-0.0446902886,-0.2893649638,-0.6464657784,-0.0550435036,-0.1372771561,0.0113920653,0.0810646266,-0.1345765144,0.075623326,-0.03488639,0.0919368416,0.0670766756,0.1374000311,0.0582092814,-0.2311389148,-0.0434882864,0.1899002343,-0.1516076177,-0.0306602269,0.0325093493,-0.1559869796,0.0627746806,0.3777853549,-0.0496725775,0.0050198552,0.1422237903,-0.1832716465,-0.154491365,-0.0363092385,-0.0381373055,0.2365215719,0.1768685579,0.2382216305,-0.158636868,0.4373874962,-0.0783557743,-0.017008407,-0.1592418849,0.0485620014,0.3178426325,-0.0862189233,-0.0779492706,0.1570954174,-0.6088183522,-0.1027718857,0.3507997394,0.0869786814,0.0950287282,-0.0254723821,0.0885225907,-0.1084133312,0.6535915136,0.0393105373,0.3207547665,-0.3691971302,-0.1743567139,-0.3436467648,-0.0518978536,-0.0848182589,-0.0192866586,0.2070866525,0.0123313358,0.3064210117,0.3621981144,-0.0961820409,-0.0810821503,-0.236599341,0.2411995977,0.0577492453,-0.3107813895,-0.2239123434,0.4482027292,-0.0229922961,-0.2116918713,-0.018929854,-0.1974861473,0.0536205135,-0.1837992519,-0.0897568986,0.1165640056,0.0145104723,0.4337092936,-0.1391922385,0.2675824761,-0.1473603547,0.2854857147,0.0650100634,0.0395629033,-0.2587375641,0.0142589388,-0.2803115845,0.2147778869,-0.3989946544,-0.0479578786,-0.4154091179,-0.2609857321,0.064824447,-0.2922825217,0.0569756217,0.0311851129,-0.0695860907,-0.1465891153,0.1929966807,0.1952983588,-0.0894407034,-0.1499374658,-0.190624088,0.1784253269,0.4314338565,-0.3552111685,-0.1906356364,0.0374732278,-0.0491149016,-0.0894508213,0.0214471705,-0.2834303677,0.2047121078,0.2579639554,-0.0355291441,-0.3045892715,0.5088086724,0.3199113905,0.1194591001,-0.0840294808,0.0225351918,0.0015101158,-0.1047733575,-0.5650414824,-0.4851825237]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1724","title":"could not run models on a offline server successfully","comments":"Transferred to `datasets` based on the stack trace.","body":"Hi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...","comment_length":8,"text":"could not run models on a offline server successfully\nHi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...\nTransferred to `datasets` based on the stack trace.","embeddings":[-0.0556910373,0.0243539196,-0.0579144545,0.3918629289,0.4269246757,-0.0151701393,0.3904104233,0.298786968,-0.1572268605,0.0592139885,0.1667313725,0.1254510283,-0.2340093553,0.0872182623,-0.0685639232,-0.3129381537,-0.0851620659,0.1403034776,-0.1411686391,-0.2622511685,0.1505435705,-0.0554317608,-0.0592292212,0.2202765644,-0.2302864045,-0.467972964,0.012530284,0.22390531,-0.5310454965,-0.1212290451,0.1194959059,-0.2809810638,0.4198203981,0.331756413,-0.0001226552,0.121100843,-0.0373298712,-0.1700540781,-0.1388303638,-0.1060684323,0.2123339921,-0.1155663505,0.2027080208,-0.365437001,0.187757507,0.2866430581,0.106699422,-0.1474490166,0.2152355909,0.2083147615,0.096568726,0.3499038815,-0.0251559485,-0.0589950122,-0.0772551671,-0.0404997207,-0.2430715561,0.3625583351,-0.2368530184,-0.169762671,0.3926005661,0.2588818073,-0.4203540385,0.2263971865,0.0279945862,-0.0400136933,0.2902400196,-0.2790037394,0.3710434437,0.4563606083,0.0158991758,0.1809749454,0.0201989301,-0.0331147797,0.1325574666,-0.3270993531,0.1356494129,0.5363458991,-0.219579041,0.3261992931,-0.452675283,-0.2438669652,-0.2808463871,0.041620668,-0.0044476911,0.3476590812,-0.0909933895,0.107159324,0.00503584,0.0487714075,-0.2216824144,-0.0720266327,0.0066398787,-0.1334008723,-0.0579671152,0.0798984319,-0.1234315559,0.434030205,-0.059826009,0.1686680913,0.1699040532,-0.1022081003,0.1268263757,0.1049730182,-0.0293027125,0.2262643874,-0.1835754812,-0.0728955939,0.326536864,0.3820692301,-0.30159536,-0.1337359548,-0.212770015,-0.0064038988,-0.3399181962,-0.1110610291,0.0906261206,-0.2743965983,0.0870430097,-0.0716111138,-0.112247549,0.0825341716,-0.0103024701,0.2949566245,-0.2500039339,-0.142714411,0.0096915299,0.135981679,-0.2525166571,-0.5636779666,0.1639352143,-0.1194168627,-0.1961861551,-0.000851351,0.1160331517,-0.1008527577,0.3347654641,-0.2758070827,0.2286782414,-0.1271293461,0.2295681685,-0.4415831566,0.2442725152,0.4925825894,0.1176769659,-0.1013374999,0.3818972707,0.0472970828,0.1900496781,0.1768841594,0.0076119164,-0.4210786819,0.3419170082,0.0150749413,-0.2129486948,-0.1885931939,-0.2548158169,0.1262318939,0.0575956218,-0.1938211024,0.1010104045,0.1167105436,-0.1039247364,-0.1128132567,0.5225958824,0.4673410058,-0.5930619836,0.3458200395,-0.3274499178,0.1220240742,0.2482403964,0.2817395926,-0.1824966818,0.4848084748,-0.2521621287,-0.3565782607,0.2180324942,0.070909977,-0.0620911978,0.2216520458,-0.1266279668,-0.3236060143,0.0560269691,-0.1604461223,0.3786844313,0.3516846001,-0.2182253897,0.1235101819,-0.0861168057,-0.1722643375,-0.330997467,-0.4467284083,0.1974660456,0.2069846839,0.3538249433,0.2508027852,0.0228695888,-0.485794723,0.0755461603,0.0873737037,-0.0862534195,-0.0893326923,0.1970318556,0.3123746812,-0.0630127043,-0.0014758225,-0.0186751969,0.1764492691,-0.1035211384,0.4966407418,-0.3425385058,0.2824816406,0.065020822,-0.0137199918,-0.1149609387,0.0583074316,0.0031410451,-0.2536644638,0.0123441527,0.1011200324,-0.1845243722,0.0288939308,0.0283862483,0.1154108495,-0.0568373837,0.3425052762,-0.2459877729,-0.0970470905,-0.2218426019,-0.3343949318,0.0853183866,-0.0218305103,-0.1819233745,0.2291695923,-0.2060632706,-0.0633745044,0.4894827008,0.1975284368,0.1157759652,-0.1499468237,0.5101521015,0.2767942548,0.1160899624,0.1643984318,-0.1113232374,0.5298624039,0.1353057027,0.3993644118,-0.0000582945,0.2212205976,0.2773774266,0.099238582,-0.0566073023,-0.1517509222,0.282958746,0.6375872493,0.0014365148,-0.3186997175,-0.1699869335,-0.1082374007,0.0984605104,0.1735091507,0.2821090221,-0.1888431758,-0.1124163941,0.1306475848,0.12439996,0.1769031584,0.1901476532,0.1301747113,-0.2143428475,0.3231209219,-0.1432169676,-0.1994461417,0.3286115229,0.3372875452,0.4366584718,-0.0239492003,0.092135556,-0.2578811944,-0.3374707997,-0.2699581683,0.0768600553,0.2462205887,-0.0987670347,0.3499799371,-0.5501237512,0.0798486471,-0.1187707335,0.1581935436,-0.7781599164,0.1997314394,0.1282280087,-0.1347550899,-0.0750995576,0.2298358083,-0.1176602617,0.4333102107,-0.1308853477,-0.0731838346,0.112260133,-0.3121380806,-0.3242466152,-0.0654900745,0.254668057,0.0612685978,0.4147370458,-0.0186793152,-0.2687003613,0.2001266629,-0.2751914263,0.0360614657,0.2371213287,0.0337984599,0.1785872877,0.3941558599,-0.1467324197,0.2776374221,0.4634224176,-0.3128438592,0.1838412881,-0.1652416438,-0.1743357033,0.1320197284,-0.1794538349,-0.2701441348,-0.4321778119,-0.1197767258,0.2209528983,0.327852428,0.1659439057,0.017990971,0.4319823086,-0.100827232,-0.259832859,0.1577003747,0.0737441853,-0.6973856688,0.3142058551,0.0719881505,-0.2081975788,0.3505109847,0.1935385317,-0.0703935251,-0.094707787,-0.6351255774,-0.0410377234,0.246533215,0.0732079446,0.1392654628,0.0780638754,0.4643867314,-0.1559262127,0.0455896594,-0.1225955859,-0.0656497777,0.1947244406,-0.1538506299,0.1946520358,-0.0939563587,0.6978850961,0.0475096069,0.2954368591,-0.3793720007,0.0560107306,0.1645643115,-0.0508089326,0.3439247906,-0.1796972156,-0.0134810722,-0.0217834208,0.0197543725,-0.7506904006,0.199707225,0.1223117113,-0.075871788,-0.4359531403,0.1131858155,0.1183793992,-0.1222617999,0.1176453456,0.2155300081,0.2556089759,0.318125546,0.1536725461,-0.2834195197,0.1153100953,0.1161882281,0.284293294,-0.0127209863,0.0940765366,-0.4686962664,-0.3218895495,0.1981167644,0.1136100367,-0.1842117757,0.1963742077,-0.1388379782,0.0366767906,0.2931828499,-0.0652546585,0.3130239546,-0.1083297655,0.0187967271,0.0375945941,-0.0417327881,-0.1770829707,-0.3699327707,0.1069414094,-0.0241607912,0.3871088028,0.1003656313,0.0544311665,0.1893159449,0.0227276757,0.1017211303,-0.3488131762,-0.1346779168,-0.636464417,-0.1117157638,-0.3126108944,-0.0087792845,-0.2526462972,-0.1724986583,-0.0894229561,0.1714596152,0.0491584726,0.2851911485,-0.2657012343,0.3015719354,0.1358818561,-0.3329971731,-0.1194319874,0.3101801872,0.3462340832,0.3224855661,0.2065996379,-0.3498119414,0.1480214596,0.1245699599,-0.0175921302,0.0337931775,0.3753103018,-0.0399800129,-0.3540718257,0.5565075278,0.2603226304,0.0144442068,-0.1225248203,-0.0537162125,-0.1825929284,0.2271774262,-0.31820786,0.3446363807,-0.1109971255,0.1575921178,-0.4492437243,0.0954926983,-0.0062034456,0.4035402536,0.0785589814,1.1776641607,0.0057296664,0.1529058218,0.005792351,-0.0936253965,0.4443847835,-0.5557780266,0.2182833254,-0.0988342017,0.024896333,-0.0504663885,0.003223255,0.3420741856,0.2894691229,-0.0730464309,0.2678490877,-0.1473528296,0.1614359915,0.007276583,-0.3365943134,0.3745903075,-0.2453681976,-0.9768887758,0.0456513688,0.187405929,0.0503585599,-0.3977455795,-0.288856864,0.1081410348,0.0954561234,-0.0225523468,-0.1320454031,-0.4224470258,0.0526942238,-0.1701724678,0.0413230434,-0.1953404993,0.0393911414,0.3178823292,0.1899033487,-0.1605306715,-0.0130184731,0.1074195802,-0.3594020903,-0.1007682458,0.2840637267,-0.1081393808,-0.307202071,-0.2423688918,-0.1749455184,-0.2249063998,-0.0325257927,0.3698732555,-0.2749661505,0.1358049065,0.1904898882,0.0654078797,-0.0162686501,0.2144324929,-0.1002895012,0.0289376862,0.1336513013,-0.1156365275,0.2179370821,0.2374548018,-0.0716690943,-0.1591771841,0.3463511765,-0.1012130901,0.2501929402,0.2619578838,0.3471616805,-0.1770067066,-0.2192958146,0.2467852235,-0.3433137834,-0.3631329834,0.2657961547,0.0975587592,0.1374494731,-0.1518628448,-0.065241918,0.2909279764,-0.1356179267,-0.167367503,-0.4936048985,-0.4773641825,0.079505533,0.038550321,-0.2481338382,0.2096689343,0.0709410608,0.2675348222,-0.140865162,-0.1759779304,0.1034182534,-0.1362705231,0.0290326606,0.2021521479,-0.0831901431,0.177807048,0.1579467654,-0.0400682203,0.0343825556,-0.1864973754,-0.0896561742,0.0181710478,0.266422987,0.1412601769,-0.4759919345,-0.0859898031,-0.44411394,-0.4063995481,0.0596760847,-0.1562878042,-0.0322698802,0.0326008797,-0.2125333548,0.2263936549,0.177562803,-0.4640947282,0.0068693492,0.0326221175,0.3147843182,0.2061523348,-0.1331223994,0.0894116759,-0.0567561388,-0.0086215464,-0.0915202722,-0.1575757414,0.1651647538,0.7159166336,-0.2913449407,0.2967382669,0.157624945,0.2130130082,0.2392484397,0.1218625605,-0.0504366048,0.3800207078,0.0561329387,-0.4624739587,0.0723747388,0.2003559768,0.0562256873,-0.0152013637,0.1639230847,0.2127391547,0.1301943064,-0.1533505023,0.221894592,0.3049509823,-0.0625053868,0.0151256602,-0.0960415676,-0.1090026498,0.2183822691,-0.1546232551,-0.0222871397,0.0709303543,0.2865228057,-0.18200095,0.112842761,0.138242051,-0.15833278,0.0604613386,-0.3185128868,-0.0390988328,0.3115638494,0.3970400393,-0.036231149,-0.43405357,0.1901130825,-0.19650653,-0.221236676,-0.2295344323,0.1677957773,0.0132759623,0.0976172909,-0.4903243184,-0.1886346638,-0.0379164517,0.3139840066,0.2132030129,0.1624592543,0.2035952657,0.3732839525,-0.1633323729,-0.3366264403,-0.4683103561,0.2358145267,0.0004386642,-0.0367361866,0.0675347149,0.4776980281,-0.2368071824,0.0490208864,0.0811308324,0.2835336626,0.0905881748,-0.1026512012,-0.0017712403,-0.0093693826,0.1093934998,-0.1809483767,-0.1770328879,0.4764706194,0.4360701144,-0.1206160709,-0.0569084249,-0.0408326499,0.6312876344,-0.2066013664,0.1816823334,0.0463429987,0.0936120525,0.2018330693,-0.1837957203,-0.3696830273,-0.1462641805,-0.3108263314,0.2939442992,0.4017177522,-0.014365064,-0.1338933855,-0.099462837,-0.0257302504,-0.2020753324,0.7074982524,0.1911570728,0.0595854372,-0.2065489441,-0.1123592556,-0.2836446762,-0.1368738562,-0.3590146899,-0.0171253383,-0.0667060092,0.06566035,0.3729098439,0.1597434282,0.3143861592,-0.3018169403,0.0341651589,0.2146001607,-0.1377696544,0.01014442,-0.2611502707,0.2335319966,-0.005106105,-0.1859284341,0.1435077786,-0.2543183565,-0.054862462,-0.0206437316,-0.1313333064,-0.3680714071,0.2590581775,0.3205949366,0.0995570347,0.4386128187,0.3547783792,-0.2626567781,-0.2313314825,0.2360306978,-0.1624908149,0.0671287701,0.2840961218,0.3771425188,-0.1786534637,0.0831171647,-0.4249831736,0.2092973888,0.132719785,-0.1113799065,-0.2545430362,-0.3838586211,0.2512685359,0.068888016,0.142711848,0.0514683239,0.1254795641,-0.0540810227,-0.6072198749,0.1804524064,0.4583279192,-0.2692609429,-0.2131065279,-0.5148097873,0.1890654266,-0.0210679509,0.1861857921,-0.6983623505,-0.0555562973,0.0995267034,0.0280560311,-0.1064276099,0.2126747817,0.1422933042,-0.0793740377,0.0178700686,0.1982595474,-0.4368927777,-0.3895274401,-0.4552981257,-0.054636322]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1724","title":"could not run models on a offline server successfully","comments":"Hi @lkcao !\r\nYour issue is indeed related to `datasets`. In addition to installing the package manually, you will need to download the `text.py` script on your server. You'll find it (under `datasets\/datasets\/text`: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/text\/text.py.\r\nThen you can change the line 221 of `run_mlm_new.py` into:\r\n```python\r\n datasets = load_dataset('\/path\/to\/text.py', data_files=data_files)\r\n```\r\nWhere `\/path\/to\/text.py` is the path on the server where you saved the `text.py` script.","body":"Hi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...","comment_length":64,"text":"could not run models on a offline server successfully\nHi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...\nHi @lkcao !\r\nYour issue is indeed related to `datasets`. In addition to installing the package manually, you will need to download the `text.py` script on your server. You'll find it (under `datasets\/datasets\/text`: https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/text\/text.py.\r\nThen you can change the line 221 of `run_mlm_new.py` into:\r\n```python\r\n datasets = load_dataset('\/path\/to\/text.py', data_files=data_files)\r\n```\r\nWhere `\/path\/to\/text.py` is the path on the server where you saved the `text.py` script.","embeddings":[-0.0568937883,0.0447787531,-0.0331752077,0.3332995474,0.3864117861,-0.0045954473,0.4649333358,0.2692718804,-0.1353352517,0.0317711979,0.192270115,0.1354355365,-0.1981076449,0.275693953,0.0291390829,-0.2823957503,-0.0642578453,0.1513577849,-0.160916388,-0.2264636755,0.1386750937,-0.0304774791,-0.0809588507,0.2365331948,-0.1869858503,-0.4619706869,0.0225121267,0.2390140295,-0.5007817149,-0.2214997411,0.1716638207,-0.2094939798,0.4189849794,0.3880804479,-0.0001210613,0.0953017846,-0.0045674457,-0.1595673114,-0.2970446944,-0.2103157192,0.2237516046,-0.2177058011,0.2277451754,-0.3331462145,0.1763512492,0.2350689471,0.0499492511,-0.2093142718,0.1599961668,0.2205237001,0.1110998243,0.3654481769,0.0041520968,-0.0529920086,-0.1416849643,0.033652205,-0.2613532245,0.3367996216,-0.2496024221,-0.1932765543,0.3615841269,0.2267213762,-0.3846462369,0.2377173752,0.1071204618,-0.0059323949,0.1734903902,-0.2360102087,0.2626617253,0.4554595947,0.1333780289,0.0602807403,-0.1196677536,-0.1701345146,0.1185573488,-0.3823783696,0.1855008155,0.4157039523,-0.1865484118,0.2766376734,-0.4178329408,-0.2676378787,-0.2441746145,0.1326269209,0.1015365571,0.2954365909,-0.1405875832,0.1240573004,0.0271607693,-0.0107553462,-0.2150408328,-0.070469968,0.0458315536,-0.0930377543,-0.1145909354,0.0824120045,-0.0527197085,0.4119786918,-0.0158199985,0.1477682143,0.1082998812,-0.1033835709,0.0469201058,0.0857243612,0.0226649493,0.2896146476,-0.0809616446,-0.0761770383,0.3698170185,0.4719712734,-0.1749238074,-0.1006729305,-0.1715146005,-0.0430684276,-0.3674766123,-0.1204360053,0.193653807,-0.2874996662,0.0668749884,-0.0899016708,-0.0507406071,0.0551261604,0.046324987,0.2603744864,-0.2619176209,-0.0300933458,0.0608170331,0.1793389469,-0.2896103561,-0.5131764412,0.1465680152,-0.1178623065,-0.2210033834,0.1116654426,0.1693404913,-0.1891278774,0.3809357285,-0.2327386886,0.2336417735,-0.109286122,0.1384670883,-0.3832319081,0.0996208116,0.3830470741,0.1516096592,0.0212223399,0.360397011,0.0293014739,0.1617751271,0.1189214662,-0.0127126779,-0.4248858094,0.3130065203,0.0168757252,-0.2212020755,-0.2422118038,-0.3803575337,0.1152495071,0.0694254935,-0.2000061423,0.074297145,0.118258208,-0.1640274376,-0.1023323461,0.5407943726,0.5818747282,-0.6021817327,0.2777576149,-0.2972584367,0.0902728662,0.2060284019,0.2522566319,-0.2069410384,0.5162286758,-0.1943635345,-0.2867471576,0.2468045205,0.0532551073,-0.0808427185,0.269138366,-0.1503320485,-0.2583183348,0.0646226853,-0.1896719038,0.3683045208,0.3258492947,-0.1560758799,0.1533566564,-0.0751300678,-0.147002548,-0.2901532948,-0.4389777184,0.176802516,0.2234634906,0.2603990734,0.2487623394,-0.006258266,-0.4477118552,0.0863196999,0.0298736766,-0.0239641406,0.0434700362,0.2015668899,0.4506844878,-0.0705122352,-0.0583301298,-0.1100971997,0.2461984158,-0.0667301863,0.5235767961,-0.3406287432,0.2355118096,0.0231517348,-0.0720983073,-0.1216815412,0.0767612606,-0.0096623376,-0.1745710075,0.0196085032,0.1260846108,-0.236101225,0.0986645594,0.0495572574,0.1481173933,-0.1525630355,0.2786816657,-0.2784129381,-0.1017347127,-0.1776435226,-0.2753492892,0.1279512942,-0.0899026319,-0.1917274296,0.315594852,-0.2771892846,-0.0355192237,0.446508348,0.1324356794,0.1414646655,-0.02056171,0.5296019316,0.2241675109,0.1720950156,0.1947706044,-0.050362695,0.5095844865,0.0626396164,0.403229773,0.0088892309,0.2083589882,0.3123408556,0.12541233,-0.0796352327,-0.1443969756,0.3125241399,0.4856480658,0.0297941454,-0.3327068985,-0.1742797345,-0.1118212119,0.1797510087,0.1433899254,0.3035912216,-0.1278378218,-0.1522689611,0.177632764,0.0633262843,0.153775081,0.274887234,0.0991967842,-0.223033756,0.3310138583,-0.1183784232,-0.1956309229,0.3147284687,0.2974816561,0.3986822367,-0.0363951325,0.0783895478,-0.225951165,-0.3413812518,-0.2408806235,0.0990349054,0.2719106078,-0.159142077,0.3608199358,-0.5341259241,0.1025403887,-0.1960663795,0.0632085428,-0.7216086388,0.1818989068,0.0750070065,-0.0777050555,-0.0632227063,0.2351086438,-0.0519084558,0.383091718,-0.1081383303,-0.1599704027,0.1140702367,-0.2853776217,-0.3427647352,-0.0876118019,0.2408557385,0.1433359534,0.3394796252,-0.0341618992,-0.2927483022,0.112728551,-0.2475780249,0.024004437,0.1149387583,0.088655822,0.194939509,0.4221848547,-0.0921320915,0.2511415184,0.4791585505,-0.358211875,0.1556310952,-0.1405010968,-0.1662369817,0.104041554,-0.2029569894,-0.4261443317,-0.4214285612,-0.152221173,0.3104702234,0.325891912,0.1697570831,0.0576725043,0.3748372197,-0.0834805593,-0.2663264275,0.1694471985,0.0455105118,-0.6586930752,0.2608668208,0.0511809848,-0.2268145233,0.3487376869,0.233999446,-0.1156647652,-0.0868986025,-0.576264143,-0.060054943,0.2335471958,0.1256029606,0.1397918761,0.1781140417,0.497764349,-0.107529901,0.0600537732,-0.1068507731,-0.2000727057,0.1969374716,-0.1299939007,0.1676265746,-0.0121158911,0.7902722359,0.0092348503,0.3142922521,-0.2903054655,0.0084185665,0.2128259838,-0.0871247724,0.3944884837,-0.1626713574,-0.1247040257,-0.0546581894,-0.0717778876,-0.5755215287,0.2132688165,0.0728450045,-0.0311929584,-0.4914730489,0.0303768534,0.044148881,-0.1480859965,0.0771165639,0.1659271419,0.3032755256,0.29891783,0.154123798,-0.2825708389,0.1377645135,0.0948115438,0.2894389629,0.0204421077,0.1212162748,-0.4598534405,-0.2257892787,0.1209856644,0.1304297745,-0.1828575581,0.318651408,-0.1515053809,0.0459565632,0.2558956146,-0.0397475809,0.3702270091,-0.1033918411,-0.0566630252,-0.0389979295,-0.0555194803,-0.208042562,-0.3152329326,0.0562949069,0.05639318,0.354060173,0.1808839291,-0.0648783445,0.10829629,0.0892377347,0.1999564469,-0.332205236,-0.1189402193,-0.6464838386,-0.1652052999,-0.401922971,-0.0064550443,-0.3250182569,-0.0830190107,-0.0548292436,0.213272959,0.0939214528,0.2819795012,-0.2759363651,0.2893920839,0.1434932053,-0.3002434969,-0.0288374703,0.3691074848,0.3564686179,0.3488225341,0.3006007075,-0.3558163643,0.0463298745,0.0748423338,0.0357962139,0.0512646139,0.3899364769,-0.1118722707,-0.2598865032,0.5852513909,0.2602341473,-0.0252087936,-0.1555958092,0.0152275972,-0.0960639045,0.1373549253,-0.4063613415,0.3523404598,-0.1227376908,0.1576822847,-0.4403464794,0.2064340115,-0.0335903578,0.3793543577,-0.0625405833,1.1074099541,-0.0092536146,0.1446345747,0.0692001432,-0.0059328824,0.4799922705,-0.5333252549,0.2535618842,-0.12944372,-0.0046211644,-0.0461915843,-0.0050679184,0.339992851,0.2836647928,-0.0960370675,0.4485304952,-0.07013648,0.1649612486,-0.0255470891,-0.2019960433,0.3052366078,-0.310603261,-0.9818566442,0.0478712656,0.1462788582,0.1332314163,-0.3802545369,-0.2694807053,0.0687722787,-0.020563731,-0.02705759,-0.1034884751,-0.4376792908,0.0515956096,-0.1494438946,0.0368107632,-0.0721609965,0.0294300858,0.4105336368,0.2096689045,-0.1813260168,-0.0883182883,0.00335224,-0.3626656234,-0.1218233854,0.2802727818,-0.0562629215,-0.274786979,-0.3198837936,-0.2276834846,-0.231062308,-0.0266148727,0.3604471684,-0.2786826789,0.1338134259,0.0680290908,0.0352665149,-0.0155466748,0.2715950012,-0.1527604908,0.0309093148,0.1501651108,-0.138130188,0.2089790702,0.1909314543,-0.0390963629,-0.2122273892,0.364325732,-0.0944558829,0.2175115645,0.3239703774,0.3774923384,-0.2118942589,-0.204738006,0.2006216049,-0.3164903224,-0.3938567638,0.1992643923,0.1818438023,0.1145847291,-0.1365828514,0.0054974766,0.2334147096,-0.0853170082,-0.1376295984,-0.505522728,-0.4226487875,0.1015447453,0.0501287356,-0.2350702286,0.3599445224,0.112142019,0.2510254383,-0.1264880896,-0.1807378083,0.094090417,-0.1220680252,0.0205024686,0.3346769214,-0.0894712508,0.2907010317,0.1360314935,-0.0294913389,0.006374198,-0.1348231882,-0.0721661597,0.0219442714,0.2543153465,0.1180179417,-0.463937521,-0.1080319136,-0.4621662498,-0.448428154,-0.0026294836,-0.1175844371,0.0384513885,0.0495541878,-0.1855060309,0.1831013411,0.1789461523,-0.4746150672,0.0617504008,-0.0389922112,0.1563979685,0.2187201977,-0.0739007965,0.0583332218,0.0015175481,-0.1295061857,-0.0923409536,-0.0293779522,0.1943140328,0.6456201077,-0.3700994253,0.2425409257,0.2151115239,0.2719478905,0.2969449461,0.1480138749,-0.0341502093,0.4043141901,0.0608062744,-0.5158955455,0.0311036017,0.2101082951,-0.0471313074,0.0149566112,0.1986990273,0.2471060306,0.0987571627,-0.1906706393,0.1931796819,0.3085995615,0.0115039591,0.0893054307,-0.0543377511,-0.0926649421,0.1870611608,-0.0917176679,0.0334026814,0.0498373248,0.3700146675,-0.2356915474,0.087171264,0.052316349,-0.1669750661,-0.0295132436,-0.3545897305,0.0348820351,0.2810544074,0.4033113122,-0.1178088933,-0.4476167858,0.2899415195,-0.3564798236,-0.1696411371,-0.2801992893,0.1834975332,-0.0051846304,0.0378137901,-0.4128588736,-0.1591304839,-0.0348882042,0.213010788,0.1806744188,0.0613004006,0.2449627519,0.3368614912,-0.1566214412,-0.4440951347,-0.4630818367,0.236345157,-0.0368836895,0.0164533034,0.0976201221,0.4217366576,-0.2259596437,-0.0861220136,0.1288255453,0.3297294378,0.0920788944,-0.0299750995,0.0365841463,-0.0084360717,0.1087547839,-0.09681876,-0.230280906,0.4081574082,0.4481625855,-0.0940641314,-0.0531834923,-0.0469577536,0.5351571441,-0.1724585593,0.1449895799,-0.0111898519,0.1382080019,0.1681204289,-0.1573055685,-0.4443290234,-0.0726281703,-0.3360543847,0.3278777301,0.4997357428,0.0441538915,-0.1408452541,-0.1766206622,-0.0241370592,-0.1380175948,0.6954714656,0.2196353078,0.0189850107,-0.2244930714,-0.059395723,-0.3687849045,-0.1776572764,-0.328856647,-0.0072521726,-0.0982662365,0.0550447293,0.3538908362,0.1690914482,0.2330797166,-0.2691388726,0.0178086571,0.0921897143,-0.1448839158,-0.0735687688,-0.3350901604,0.2141494155,0.0445821397,-0.1982061863,0.0902882665,-0.244534567,-0.0779767111,-0.0633493736,0.0008508303,-0.3760929108,0.0904644951,0.2897028625,0.0729008168,0.4466968775,0.3902719021,-0.3010517061,-0.294626385,0.1509020627,-0.2258808464,0.1717267483,0.2802489698,0.3936597407,-0.2018096447,0.0658832118,-0.5258012414,0.2210103571,0.1593195051,-0.1110807285,-0.2012768239,-0.3043133318,0.2239813805,0.0633029491,0.1959609538,0.0439173579,0.068954505,0.0123973796,-0.5844638348,0.122751765,0.4918978214,-0.2912522554,-0.1594193429,-0.5328425169,0.2430768758,0.0117936581,0.0775991455,-0.7704250216,0.0368389748,0.0950353816,0.0739345029,-0.1435618103,0.2238374501,0.0500981621,-0.0602890588,0.0142887942,0.2476168871,-0.4136193395,-0.4147634208,-0.3414915502,-0.0328724161]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1724","title":"could not run models on a offline server successfully","comments":"We're working on including the local dataset builders (csv, text, json etc.) directly in the `datasets` package so that they can be used offline","body":"Hi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...","comment_length":24,"text":"could not run models on a offline server successfully\nHi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...\nWe're working on including the local dataset builders (csv, text, json etc.) directly in the `datasets` package so that they can be used offline","embeddings":[-0.1060371175,0.0922159106,-0.046883624,0.3895358443,0.3865782022,0.0059097745,0.4341858625,0.3095848262,-0.1331249028,0.0817197114,0.2425644696,0.1654364914,-0.2283213139,0.1497392505,-0.0170877595,-0.2990975976,-0.0592587888,0.1828980446,-0.1974499673,-0.2429979891,0.1269834936,-0.0581292957,-0.0441667959,0.2301857471,-0.21554856,-0.4256712496,0.0423487984,0.1900561005,-0.5258668661,-0.1492341012,0.1743561178,-0.2595254183,0.4095986187,0.2606842816,-0.0001232795,0.1577787846,-0.0261516292,-0.1618656665,-0.1889204234,-0.168033421,0.2139279693,-0.1500169039,0.2031965107,-0.343084842,0.1663936973,0.2027682215,0.0882515311,-0.2262526304,0.1819858998,0.2225345969,0.0811528936,0.3456409574,-0.0080614006,-0.0516863428,-0.0894696116,-0.0270259585,-0.2824190557,0.3596595228,-0.1925052106,-0.1691604704,0.3959687054,0.2600797713,-0.3926531672,0.2729837298,0.0724465623,-0.0229853652,0.2645563185,-0.2598523796,0.3517810106,0.4191648364,0.1268899292,0.1497762948,-0.0386599675,-0.0704283938,0.1231043637,-0.3096102774,0.1817725003,0.5108624697,-0.2465870827,0.3363565207,-0.4833446145,-0.2510819137,-0.3149570823,0.1009634361,0.0626166612,0.3504547477,-0.0984806642,0.1237138882,0.0024207146,0.0537846945,-0.1307026446,-0.1126635671,0.0507372916,-0.0992250144,0.0065360749,0.1016821787,-0.0649581552,0.4636526704,-0.0791180581,0.1569879055,0.0828815699,-0.1005582884,0.1395375133,0.1461672485,0.0042678369,0.2160400152,-0.0977964327,-0.0820842907,0.326407969,0.3431952,-0.2357829809,-0.127774775,-0.2011373192,-0.0050393478,-0.3550642133,-0.1234038919,0.0834385082,-0.2681816816,0.1247945651,-0.0364502594,-0.0826351121,0.0757178068,-0.0164595135,0.2262731045,-0.2433173805,-0.1361799836,0.0051484997,0.1221495718,-0.262883991,-0.5288695097,0.1647808999,-0.1130513996,-0.2026740164,0.0002788581,0.1247887239,-0.1336740702,0.3475115001,-0.2760000825,0.2243671715,-0.0759025514,0.2380285263,-0.4224008024,0.2049113214,0.4813667536,0.0758747384,-0.081188418,0.3487046957,-0.0194177367,0.1453605145,0.1722142249,0.0393870845,-0.4563670456,0.2807501256,0.0020650353,-0.2078278512,-0.1681116074,-0.2502406538,0.0999007374,0.0361761861,-0.1831386685,0.0544170551,0.1322178394,-0.1089423746,-0.0522018746,0.5060938001,0.4959051311,-0.6791664958,0.3387345374,-0.3162170053,0.1019221842,0.214804396,0.2621174455,-0.2264838517,0.6079880595,-0.2353295088,-0.3891132474,0.2655009329,0.0433162041,-0.0551420934,0.305403173,-0.1246066391,-0.3638512194,0.0518736392,-0.1435494274,0.336173296,0.3802532256,-0.1806554645,0.1694301963,-0.1315045059,-0.1806259453,-0.2857104838,-0.4504421651,0.1794899851,0.2346063703,0.3742853403,0.2641890645,0.012556795,-0.5038218498,0.0280284155,0.0363710709,-0.0293478053,-0.0439176522,0.1664084047,0.3082428873,-0.0919730291,-0.0644746497,-0.112760365,0.1963846534,-0.1144931987,0.5098642111,-0.3852114975,0.2672775686,0.0571122691,-0.0079921652,-0.1296545863,0.076863572,-0.0258728843,-0.2126410007,0.016152421,0.1104285046,-0.2051447779,0.0521740802,0.0527118742,0.1214864329,-0.0695323795,0.2951959074,-0.2672995031,-0.1024870872,-0.2109735608,-0.2992351651,0.0634127408,-0.0716551915,-0.1867232323,0.2357448488,-0.2127280831,-0.080583334,0.4685858488,0.1320838034,0.1254627407,-0.0921301693,0.4863225222,0.2074817717,0.0990616009,0.1751621068,-0.1444108486,0.5445819497,0.1018009633,0.4248337746,0.0000196064,0.2099057138,0.3192688525,0.0931873694,-0.052761782,-0.1990206391,0.2491931766,0.5976501107,0.0451320335,-0.2945423126,-0.1576206088,-0.1352266371,0.1207931861,0.1714498401,0.2262714505,-0.2169025987,-0.080658704,0.1490546018,0.1672179699,0.1669870466,0.2008853853,0.1179870218,-0.2305525541,0.3015746772,-0.1642576158,-0.2037895769,0.2730614841,0.2957656085,0.3667021692,0.0228051282,0.0161165912,-0.2569947243,-0.3741302192,-0.2441598177,0.0841519907,0.2269873023,-0.0501484536,0.3766278028,-0.5777348876,0.0765169188,-0.1849752218,0.1236771196,-0.7295866013,0.2035398781,0.1181023344,-0.1463213414,-0.1182480976,0.2264245749,-0.0453697629,0.4255085588,-0.1549142152,-0.1408883929,0.1204851642,-0.3105560541,-0.3584547341,-0.0728731602,0.2951801717,0.1145786867,0.4369487762,-0.071857594,-0.26792413,0.1672137678,-0.2344250977,0.0772113055,0.2278923094,0.1281882823,0.1719551533,0.3319923282,-0.1141428053,0.2252657264,0.4916746616,-0.2578901947,0.2110649496,-0.1713347137,-0.1653002799,0.0724964589,-0.220841065,-0.3486708403,-0.393298775,-0.1253815293,0.2265319824,0.3537360132,0.1589218378,-0.0445102453,0.4573685229,-0.0929059312,-0.2593960166,0.1491314024,0.1031404436,-0.7283962965,0.2829964459,0.0349384099,-0.2215238065,0.3270794153,0.1820893884,-0.0318351723,-0.0948087871,-0.5988149047,0.0097073819,0.2486181408,0.1019084901,0.1571266204,0.1113996878,0.5097570419,-0.1424682736,0.0674903095,-0.1166140363,-0.1063201278,0.1910106838,-0.1679154336,0.1807020009,-0.0488632061,0.6888198853,0.0007349898,0.3086834848,-0.390596211,0.0166347846,0.17128326,-0.0299936868,0.40847224,-0.1971964985,-0.0439686812,0.014222241,0.0347858369,-0.7397351861,0.1826052368,0.0957731381,-0.0742822587,-0.4723063111,0.1330207288,0.0896026194,-0.115159601,0.0915839896,0.1975882649,0.2886815071,0.303163439,0.1400658041,-0.2730112076,0.1631091237,0.1285757273,0.2771193981,0.0189319998,0.1096610129,-0.4963159859,-0.3320191801,0.2083682269,0.0710559413,-0.1494296491,0.2202667147,-0.1420293897,0.0498666838,0.2850314379,-0.0757779479,0.2722777426,-0.16470626,-0.0134768616,0.0775523335,-0.0239908248,-0.1855079681,-0.3541506827,0.1425025463,0.0040041162,0.3386434317,0.1648259461,0.0326472186,0.1159757376,0.0680910498,0.1314382702,-0.3629449904,-0.1577282995,-0.6153283715,-0.0834550411,-0.3465621173,-0.0133139975,-0.3082726002,-0.1270842105,-0.0039522261,0.130096823,0.0682387725,0.3179319203,-0.3118037283,0.2789821029,0.1137669906,-0.3539096713,-0.0815211684,0.3305835724,0.3893921375,0.328368485,0.2185277343,-0.3805767,0.1686003059,0.0959164649,-0.057903368,0.0689773932,0.3452143371,-0.0702002943,-0.4040652812,0.5173208117,0.2632911801,-0.0330146775,-0.0983414054,-0.0295500942,-0.1979849488,0.2581474781,-0.3464402258,0.3385020792,-0.1365176886,0.1339896619,-0.4609634578,0.1261379421,0.0039555221,0.4226443768,0.1108060852,1.1008733511,-0.0302267466,0.151302591,-0.0305160359,-0.0831148624,0.4423339665,-0.5608237386,0.2042626441,-0.0627675354,0.0994587392,-0.0665462241,-0.0084088231,0.3909048438,0.3260176182,-0.0431140698,0.297632426,-0.1120742559,0.2329374701,-0.0292063318,-0.2955103517,0.3363870382,-0.2708414793,-0.9013369679,0.0446568057,0.2037181556,0.054089766,-0.4390063584,-0.2603447139,0.1509891301,0.0731734261,-0.0094026299,-0.1440803856,-0.3819884658,0.0011635107,-0.1484069079,0.087262623,-0.1668409556,0.1029596478,0.3616217375,0.2291163355,-0.1945030689,-0.0927305445,0.0261328984,-0.3629099727,-0.1293322444,0.265946269,-0.0840732306,-0.3029451966,-0.2341602743,-0.2372385859,-0.2614692152,-0.0060253278,0.3813209832,-0.259647727,0.1571882814,0.1574899405,0.0479959212,0.0001243685,0.2620933056,-0.0874706879,0.0241285432,0.112803936,-0.1245092601,0.2401937991,0.1556711346,-0.0731699988,-0.1726798117,0.3328465819,-0.0587854758,0.2105105966,0.2397674769,0.3009365201,-0.1843081117,-0.2276007086,0.2428056449,-0.2906495035,-0.333481878,0.2115782797,0.1371348202,0.0888892114,-0.1580158621,-0.0295714606,0.3003534079,-0.1638622135,-0.1792953312,-0.488000989,-0.5016860962,0.0750242919,0.0533122197,-0.2410890758,0.238759473,0.1236289814,0.3061273694,-0.0918692425,-0.1614235938,0.0825772956,-0.1390286088,0.0727649182,0.2325455099,-0.0116850818,0.1808242947,0.140403986,-0.0536127575,0.0056340918,-0.2128901035,-0.0784872547,0.000878223,0.2617081404,0.1803006828,-0.4546940029,-0.089555718,-0.4941138327,-0.4025227129,0.0643893182,-0.1324833632,-0.0453479625,0.0152403116,-0.182668373,0.2745902538,0.208201617,-0.4284770787,0.0357549675,0.0305662509,0.3380967081,0.1776779592,-0.1456724256,0.1034392938,-0.0515710339,0.0552823469,-0.088270627,-0.1582063138,0.1931501478,0.7184301615,-0.3177259564,0.3302538395,0.1819321513,0.2421544641,0.2712643743,0.1515278816,-0.0539722666,0.4231826961,0.0505090207,-0.4565162063,-0.0195251424,0.1728264689,0.0098730158,-0.0091750072,0.1225518808,0.2617377341,0.0980711058,-0.1515580714,0.1995248646,0.2672557831,-0.0535317659,-0.0175356641,-0.0352311432,-0.1184032485,0.2347486317,-0.1525400877,0.0071960767,0.0602438375,0.3606817722,-0.1660752892,0.0655263513,0.1131355837,-0.1095704287,0.0064582806,-0.3195029199,0.0528095737,0.3226977885,0.444642961,-0.0697607994,-0.4415719807,0.2010569572,-0.2087524086,-0.2013075948,-0.208667174,0.1809003055,-0.0036266393,0.0665345043,-0.5100694895,-0.1663980186,-0.0279705562,0.3207267821,0.2205475718,0.1367192864,0.1971613914,0.3423575461,-0.1319387108,-0.2946231365,-0.4448850751,0.1799904108,-0.0756607205,-0.0390101597,0.0898576304,0.4078085423,-0.2082224339,-0.0387638211,0.0797020569,0.2462125868,0.122876361,-0.1356429607,-0.0056636212,-0.054616455,0.1134184673,-0.2144869715,-0.1570132822,0.3972935081,0.4349235296,-0.0639509633,-0.0727436468,-0.0289393961,0.6289818287,-0.2289360613,0.164932847,0.0454148166,0.1317918003,0.2369409353,-0.2130007595,-0.3524011672,-0.1187879369,-0.3009316027,0.29174155,0.4879913032,0.0057118456,-0.1250534654,-0.1439050436,-0.0292713605,-0.1707938761,0.7764104605,0.197144866,0.0538190641,-0.2216299623,-0.1533848494,-0.3016816676,-0.1651780158,-0.365228951,0.0504463054,-0.0464380458,0.0698894486,0.3945067823,0.2195347548,0.2461516708,-0.2578766346,0.0086982502,0.1628670245,-0.1257894486,-0.0179674979,-0.2908248007,0.227585569,0.0085609481,-0.2548984289,0.1419038773,-0.2346863151,-0.0755825564,-0.0317225568,-0.0817788541,-0.3402712345,0.1852445304,0.3405696452,0.1134485602,0.4145264626,0.3741554618,-0.2367184311,-0.2411896288,0.2530457377,-0.2121548653,0.071350649,0.2649787068,0.3803726435,-0.2436315268,0.1585649699,-0.4597676992,0.1670285761,0.1543709338,-0.0887939259,-0.2322809994,-0.385086,0.1865187436,0.0526648238,0.1540886164,0.0594634898,0.127595976,-0.0384555794,-0.5950487256,0.185004279,0.4575264454,-0.2326516062,-0.2262018621,-0.4928091168,0.152681157,0.01217335,0.1016606987,-0.6858365536,-0.0126666613,0.1080217734,0.0367454253,-0.0920476392,0.2051384896,0.1761123985,-0.0472821593,-0.019661352,0.2132984996,-0.4040830433,-0.3638872206,-0.4775109589,-0.0922587439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1724","title":"could not run models on a offline server successfully","comments":"The local dataset builders (csv, text , json and pandas) are now part of the `datasets` package since #1726 :)\r\nYou can now use them offline\r\n```python\r\ndatasets = load_dataset('text', data_files=data_files)\r\n```\r\n\r\nWe'll do a new release soon","body":"Hi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...","comment_length":38,"text":"could not run models on a offline server successfully\nHi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...\nThe local dataset builders (csv, text , json and pandas) are now part of the `datasets` package since #1726 :)\r\nYou can now use them offline\r\n```python\r\ndatasets = load_dataset('text', data_files=data_files)\r\n```\r\n\r\nWe'll do a new release soon","embeddings":[-0.0688206926,0.1015685052,-0.0669888109,0.3511100113,0.3767763674,-0.0086315172,0.4361560941,0.2894013524,-0.1399609298,0.0760931149,0.2346403003,0.1608860195,-0.2274725735,0.1778730303,-0.0325524807,-0.3421509564,-0.0454738364,0.1792569011,-0.1910052598,-0.2585516274,0.1118428409,-0.0674841031,-0.0758616179,0.2248041332,-0.172747612,-0.4493913054,0.028594479,0.1861250699,-0.537869215,-0.1589070708,0.1751311123,-0.2850415707,0.4374857247,0.256544441,-0.0001215812,0.1360872984,-0.0246050414,-0.1551845372,-0.2021266073,-0.2065086514,0.2451994121,-0.2061143517,0.2791469991,-0.3414956331,0.1687799245,0.1580099016,0.0705215707,-0.1988772452,0.1477961838,0.2522169054,0.1006830335,0.3377572,-0.0068500517,-0.0385962427,-0.159553498,-0.0426585227,-0.2651070654,0.3764236867,-0.1992182434,-0.216839686,0.3203285038,0.1805513203,-0.4037184715,0.2859338522,0.0476996899,-0.0072088926,0.3059214056,-0.2814565301,0.328284651,0.3897051513,0.0928827003,0.1497496516,-0.0353211053,-0.0996459574,0.0985234976,-0.3189719617,0.1626932472,0.4870989025,-0.2380461097,0.3007386625,-0.4002816379,-0.2382664233,-0.2836399674,0.1429473609,0.0697753504,0.4067317247,-0.1066973731,0.0930780992,-0.0250147469,0.0395161957,-0.1332188845,-0.1188887358,0.0877983421,-0.0954038501,-0.0036215484,0.1037962288,-0.0244902447,0.4370110035,-0.117531769,0.1676568538,0.1017598063,-0.1033442393,0.1169741079,0.0955589935,0.0173146091,0.2371708453,-0.1497069895,-0.0764175728,0.3308113515,0.3549759686,-0.2280217707,-0.1040030345,-0.1588992029,-0.0280609783,-0.3715130389,-0.1064284965,0.1462471485,-0.223219052,0.1276468933,-0.0170747302,-0.1228610501,0.0449146479,0.0043858583,0.201846227,-0.2599536777,-0.0614691675,0.0148380194,0.1398490369,-0.2758599818,-0.5037819743,0.1586385071,-0.1060098559,-0.1690450311,-0.0141743161,0.1528880447,-0.1425388455,0.3483355939,-0.2321776599,0.2416369766,-0.0999860615,0.2083473653,-0.4021793902,0.1685494781,0.4523255825,0.1266523749,-0.0599456392,0.3684250414,-0.046919357,0.1518625915,0.1702974886,0.0549549162,-0.4641089439,0.286265403,0.0226029456,-0.1837539524,-0.2005116493,-0.2791651785,0.1201579869,0.0192346498,-0.1471369714,0.0545253977,0.1439160705,-0.101535067,-0.0637280494,0.5197229981,0.4969885349,-0.6513744593,0.343490839,-0.3135584295,0.1347789317,0.2270342559,0.2421960384,-0.2208391577,0.5864211917,-0.2081828713,-0.4126344323,0.2668750286,0.0500167273,-0.0629508495,0.2916077375,-0.1212403476,-0.3211032152,0.0613924712,-0.2034472823,0.3311758339,0.3757843673,-0.1575743258,0.1672545075,-0.1323935539,-0.1724302769,-0.3133817315,-0.4422492683,0.1823100299,0.233255595,0.395018816,0.2591549158,-0.0004789559,-0.4904565215,0.0822837427,0.0750812963,-0.030376263,-0.0373420939,0.1735895276,0.322699964,-0.0871328339,-0.050399676,-0.1006901041,0.1844821274,-0.0768969432,0.5632579923,-0.3706647158,0.2933447659,0.0315240994,-0.0384810418,-0.1228797659,0.0736027732,0.0064319707,-0.2515653074,0.0297923647,0.1520210207,-0.2159970999,0.0400627218,0.0401393212,0.1142153144,-0.0814347714,0.32415694,-0.2643056214,-0.1023516133,-0.1939538717,-0.3051472306,0.0810594186,-0.0574015453,-0.2148802876,0.2094394714,-0.2124065161,-0.0432941429,0.4440833032,0.1399232447,0.1407678127,-0.0312370732,0.478983283,0.2490859181,0.1185441688,0.1844966561,-0.0986160487,0.5424450636,0.0634376928,0.4350113571,0.0048439461,0.2102136314,0.326266259,0.1305159032,-0.0781854615,-0.2040623426,0.282686621,0.5560109019,0.0814245418,-0.2923345864,-0.1697802097,-0.1641284972,0.096555829,0.1423974931,0.2526114881,-0.2213065624,-0.0806182474,0.1542587131,0.1293653995,0.1689746082,0.2208225578,0.0985001624,-0.2407869399,0.326325655,-0.1443661749,-0.2146850675,0.2925775349,0.3399430215,0.393356353,0.0169404782,0.0489334017,-0.2314456105,-0.3556240201,-0.2272410244,0.0943486243,0.2546875477,-0.0677916482,0.3628233075,-0.5687308908,0.1182087138,-0.1478933543,0.0290822964,-0.7140902877,0.193494603,0.0739201009,-0.1115890965,-0.1312538981,0.2420375943,-0.0791335031,0.3908633292,-0.1166283637,-0.1547572762,0.119524844,-0.2907871306,-0.3274575174,-0.0611502193,0.2906631529,0.1177891865,0.3798543811,-0.0413909331,-0.2935419381,0.1742339581,-0.2367729396,0.0490828827,0.1882865876,0.1058585718,0.1678858548,0.3347955346,-0.1121794954,0.23501499,0.4916298389,-0.2610308826,0.2256456465,-0.1642932147,-0.1681237668,0.1103269085,-0.1871712953,-0.3800619245,-0.4347546101,-0.1456546932,0.2535462379,0.376981616,0.1690071374,-0.0515177213,0.4320072532,-0.1171561554,-0.2062694579,0.1546186507,0.1036776751,-0.7016678452,0.2701812387,0.0347405747,-0.2287964374,0.29579404,0.1756510884,-0.0413084105,-0.1743876487,-0.5842974186,-0.0135889994,0.25187096,0.1386651695,0.1520708799,0.0982301682,0.4951812625,-0.1266094446,0.0680923313,-0.0909656584,-0.0951695889,0.1854781955,-0.1201190427,0.1145274788,-0.0571163371,0.742993176,-0.0356152765,0.2881908119,-0.3934966922,0.0458831191,0.1955237687,-0.0219292026,0.3647749722,-0.2171272635,-0.0775315166,-0.0422316827,-0.0208114944,-0.7154508233,0.1765983552,0.0569561236,-0.0636512935,-0.495248735,0.0771271288,0.0867157727,-0.1323036551,0.0827839971,0.1421041042,0.2872105539,0.3212448061,0.1748225093,-0.2716975808,0.1677949131,0.0896409899,0.2322819829,0.0098250788,0.090733923,-0.4119895697,-0.2953493297,0.1611486822,0.0700788647,-0.160579443,0.2710830569,-0.0938439593,0.0386294797,0.2771812677,-0.0449423939,0.2524920702,-0.1391098648,-0.0380916782,0.0602886267,-0.0353126116,-0.2176900208,-0.3449375033,0.1183990836,0.0200136565,0.3473455012,0.185046196,0.0676729679,0.096241802,0.0682872012,0.1352413595,-0.3547175825,-0.1549387574,-0.6495368481,-0.0956030488,-0.3129642904,0.0145580079,-0.3000569344,-0.107872352,-0.0351734012,0.1679344475,0.0900232941,0.2815142572,-0.3234457374,0.2786961794,0.0911788717,-0.3535845876,-0.0731417835,0.300555855,0.3768797815,0.3257950842,0.2044167668,-0.4117551744,0.115269132,0.1214464381,-0.0201979484,0.0688486025,0.3274901807,-0.0772834644,-0.3651017547,0.54949826,0.2760508657,-0.0480095185,-0.1442995518,-0.0186418481,-0.1724534631,0.2175670564,-0.3638169169,0.3758825958,-0.1624684632,0.1553263366,-0.4925996065,0.1520401537,-0.0101998169,0.4426673055,0.0694619045,1.0599328279,0.022854019,0.1284062862,0.0047172201,-0.0944832638,0.4703812897,-0.5697631836,0.2320479751,-0.0539572239,0.0694855899,-0.0571762882,-0.0019675265,0.3732961416,0.2930673957,-0.0484439693,0.3198630214,-0.110016562,0.2009478509,-0.0233043376,-0.2344212681,0.3120016754,-0.2653403878,-0.9842761159,0.0483068526,0.1925529391,0.0627847537,-0.4223129451,-0.2704314888,0.1238486022,0.0586050637,-0.0010960313,-0.1663839817,-0.3592179418,0.0524136908,-0.1638941169,0.0338289551,-0.1758542806,0.0736074597,0.3699012697,0.1915730685,-0.1955907047,-0.0991091952,0.0215019938,-0.3495211303,-0.1159206629,0.282584101,-0.0724923834,-0.2928802669,-0.2865163684,-0.2055083066,-0.2601913214,-0.0066365334,0.3399687111,-0.2427624166,0.1476844102,0.1373726577,0.0744900778,-0.0076543242,0.2890259624,-0.1079538837,0.0415004306,0.1084112898,-0.1026229709,0.2344977707,0.1501000971,-0.0739880726,-0.2278772593,0.3479820788,-0.0426664166,0.2574891448,0.2616275251,0.3537609279,-0.2005519122,-0.2120904624,0.2622404695,-0.3185271621,-0.3099796176,0.2088734955,0.145812884,0.1254112571,-0.1561202258,-0.0193799883,0.3010338843,-0.1232164651,-0.1636246145,-0.4854910076,-0.5012033582,0.1014195681,0.0857826322,-0.2803422213,0.2758418024,0.1405427009,0.3010770679,-0.0994813815,-0.1697589308,0.0891413093,-0.1296363026,0.054707475,0.2539983988,-0.0599360764,0.2103337049,0.1526263654,-0.0253661908,-0.0054237503,-0.1418764442,-0.0956316665,0.0087537458,0.2517922819,0.1856032461,-0.4747941196,-0.1154040098,-0.4813809693,-0.4180763066,0.0528796501,-0.1577677429,-0.0088734441,0.0365303941,-0.1716648787,0.2572540939,0.2017696202,-0.4307262599,0.0397006124,0.0049727703,0.3044846952,0.1815281808,-0.1100684702,0.1526843607,0.0032962465,0.0171785653,-0.1092249155,-0.1570153981,0.1787300557,0.7193294168,-0.3075017631,0.3063839674,0.1931765825,0.2308468819,0.3015109301,0.1803452373,-0.0536856242,0.4138587713,0.0702543184,-0.4882838726,0.0069469521,0.176931411,0.047249686,0.0104862917,0.1354908496,0.2199312449,0.0552910864,-0.1720401049,0.2299408764,0.2734171748,-0.0337720439,0.0319669954,-0.0166451465,-0.0856810361,0.2249730676,-0.1628892124,0.0085393945,0.0474330336,0.3475340903,-0.1718399227,0.0406697355,0.0825721174,-0.1506287903,-0.0046278602,-0.2909693718,0.0662904233,0.2990258932,0.3514059186,-0.1013189778,-0.4393189251,0.2247025818,-0.2592667937,-0.1647417843,-0.2474612743,0.181633383,-0.0260366295,0.0554256104,-0.4503201246,-0.1686243117,-0.0023025586,0.290969044,0.2373080701,0.1173041165,0.1643869281,0.3563939631,-0.1417886168,-0.3237810135,-0.4595843554,0.1912478209,-0.0613956079,-0.0222264454,0.0422744974,0.4373606443,-0.2647348642,-0.1146643162,0.064167887,0.2666687667,0.1226724163,-0.1760981381,0.0033139484,0.0050314735,0.0997149721,-0.1578692347,-0.1710403711,0.4175082147,0.4818113744,-0.0837444067,-0.0513478704,-0.0397050083,0.6139253378,-0.1812516451,0.1471345872,-0.0052047637,0.1477607489,0.2392620444,-0.2017848045,-0.3352897167,-0.1234086156,-0.2997220755,0.2940615416,0.4473364949,0.0295346044,-0.1219905317,-0.1494264007,-0.0233373512,-0.1579224616,0.7504980564,0.1971173137,0.0482381061,-0.2123737335,-0.1313573122,-0.295815587,-0.1818986535,-0.3605087101,0.035326954,-0.0797302425,0.0501419529,0.3883093596,0.2088676691,0.218861863,-0.2428185046,-0.0423410945,0.1709816307,-0.1217677519,-0.0260697845,-0.3417568207,0.2257138044,0.0036651115,-0.2411002219,0.1120152771,-0.2478396297,-0.0512706935,-0.0587820187,-0.092659317,-0.31890136,0.2473709285,0.3093870878,0.1269901991,0.4091373086,0.3795032799,-0.2697545886,-0.2558471859,0.2380370647,-0.2149432003,0.1211886406,0.2240929306,0.3407705426,-0.2189384699,0.0882967561,-0.4573328197,0.190909043,0.1478906572,-0.1057857946,-0.2481844276,-0.3945682347,0.2460967898,0.0343816765,0.1491891742,0.0501419418,0.110535197,0.0170140043,-0.5854977369,0.195134297,0.4119105935,-0.2618524432,-0.1984443516,-0.5416531563,0.1560494602,-0.0023718451,0.1555011123,-0.6582355499,-0.025205465,0.1295283884,0.0451100506,-0.089271538,0.1891094595,0.183068797,-0.0569461286,-0.0276941061,0.1851194203,-0.3997740149,-0.3908595145,-0.4278211892,-0.0841174573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1724","title":"could not run models on a offline server successfully","comments":"> The local dataset builders (csv, text , json and pandas) are now part of the `datasets` package since #1726 :)\r\n> You can now use them offline\r\n> \r\n> ```python\r\n> datasets = load_dataset('text', data_files=data_files)\r\n> ```\r\n> \r\n> We'll do a new release soon\r\n\r\nso the new version release now?","body":"Hi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...","comment_length":52,"text":"could not run models on a offline server successfully\nHi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...\n> The local dataset builders (csv, text , json and pandas) are now part of the `datasets` package since #1726 :)\r\n> You can now use them offline\r\n> \r\n> ```python\r\n> datasets = load_dataset('text', data_files=data_files)\r\n> ```\r\n> \r\n> We'll do a new release soon\r\n\r\nso the new version release now?","embeddings":[-0.0705997646,0.0818422586,-0.067387253,0.3484895825,0.3655336201,-0.0059052366,0.4193034768,0.2942718565,-0.1489263922,0.0716177747,0.2294794321,0.1583710313,-0.2269311994,0.1734987497,-0.0371109247,-0.3468999863,-0.0394554771,0.1774479449,-0.1939214319,-0.2578321099,0.0990721583,-0.061965622,-0.0807371214,0.2230645716,-0.1615162343,-0.4670543373,0.0276674759,0.1842142045,-0.5382243991,-0.1700838804,0.1658654809,-0.283714354,0.4375509918,0.2521188557,-0.0001213945,0.1277884841,-0.0248076674,-0.1500540227,-0.2117671371,-0.2167317718,0.2522087097,-0.2124804705,0.2840510011,-0.3407836258,0.1697348654,0.1742061079,0.0756895021,-0.1966519952,0.1607891172,0.2545123696,0.1027426422,0.3456666768,-0.0017496607,-0.0426849723,-0.1672260165,-0.0622239225,-0.2551248968,0.3821850121,-0.2034241557,-0.2088387758,0.3286895454,0.1911961883,-0.3957904279,0.28680107,0.0446274206,0.0000419508,0.3129147887,-0.2795209289,0.323861897,0.3837933242,0.0983648077,0.1457062662,-0.050395526,-0.0933907926,0.0946790725,-0.3246632218,0.1679434627,0.4823390245,-0.2290003151,0.3012810349,-0.3965063095,-0.2433443069,-0.2756930888,0.1433299333,0.0731505752,0.4021547139,-0.1145825982,0.0903980434,-0.0285767373,0.0393344574,-0.1420435309,-0.1136839837,0.0838163346,-0.0864124745,-0.00361577,0.1033836454,-0.0319475718,0.4332543015,-0.1196015477,0.1626177728,0.1086184308,-0.1116893217,0.1194215491,0.0874299109,0.0294075161,0.2210605145,-0.1583593786,-0.0785427988,0.3308566213,0.3641965091,-0.2270265073,-0.1102049425,-0.149982363,-0.0261561871,-0.349898994,-0.1013975069,0.1574403048,-0.2232752442,0.1326773018,-0.0291154329,-0.1286501884,0.0402707495,-0.0050060549,0.2086393684,-0.2704685628,-0.0478030406,0.0200174898,0.1302569807,-0.2841922641,-0.5011151433,0.1586946696,-0.1097302362,-0.1684495211,-0.0036691439,0.1457222402,-0.1474144161,0.341573745,-0.2266682535,0.2293043286,-0.1161577702,0.2034096867,-0.3967208862,0.1668656617,0.4541451335,0.1340209246,-0.0559192151,0.3632282615,-0.0435790904,0.1427352577,0.1752673388,0.0568508878,-0.4620818496,0.2975709438,0.0243984107,-0.1721846163,-0.2028435469,-0.2606344223,0.1229296178,0.0196395274,-0.1636729091,0.054811351,0.1466071308,-0.1068672538,-0.0756872892,0.518643558,0.5003830194,-0.6551429629,0.3357151449,-0.3109006882,0.1295129806,0.2178948075,0.2354239374,-0.2067971826,0.5655161142,-0.2079125047,-0.4268912673,0.2711883783,0.0439608134,-0.0732552111,0.2840721905,-0.1325544715,-0.3194464147,0.0735415891,-0.198613748,0.327812463,0.3712581098,-0.1727527082,0.1563497484,-0.1345729679,-0.1739404052,-0.318761766,-0.4410797954,0.1834992021,0.2193382531,0.3940840065,0.2665600181,0.0074794665,-0.4645365477,0.0909483805,0.0913659409,-0.0380537249,-0.0364861004,0.1778343767,0.3347792625,-0.0872904882,-0.0518357754,-0.1034801453,0.1783381253,-0.0663980618,0.5493116379,-0.3502076864,0.2871763706,0.0241941735,-0.0407498144,-0.1121701077,0.0731939226,0.0052383589,-0.2501824796,0.0480812117,0.1526526511,-0.2161738873,0.0541430302,0.042806752,0.1116733849,-0.0890432224,0.3269085288,-0.2754415572,-0.1051766723,-0.1878639907,-0.3095560968,0.0890194774,-0.0565584823,-0.212663427,0.2083296329,-0.2054125369,-0.0481308885,0.4253845215,0.1450350136,0.1370151192,-0.017994618,0.4820729792,0.2605344057,0.1178944111,0.1872793287,-0.0823040977,0.5436932445,0.0677114725,0.4257641137,0.0182650015,0.2127682716,0.3230121434,0.1354770511,-0.0674882308,-0.2129700482,0.2741913795,0.5513848066,0.091666162,-0.3043964803,-0.1712460369,-0.1530237496,0.1117714942,0.1362790465,0.254383862,-0.2206159383,-0.0868165791,0.1520422846,0.1127185449,0.1753354073,0.2232252359,0.0989453644,-0.2311029881,0.3421707749,-0.1296846271,-0.2052721232,0.296341002,0.3406378329,0.3999171257,0.0022243706,0.0476659685,-0.2282091081,-0.3398965895,-0.2150229812,0.0960685983,0.2548360229,-0.071879141,0.3595644534,-0.5689226389,0.1225346774,-0.1531618088,0.0271621924,-0.7147392631,0.1970946491,0.0718186721,-0.090321742,-0.117038615,0.24285689,-0.0787803605,0.3732526004,-0.117195487,-0.1490920186,0.1266219169,-0.2835884392,-0.3315928578,-0.0591550991,0.2754873037,0.1071223989,0.3786298633,-0.0468066819,-0.2891301811,0.17163077,-0.2428634912,0.0421195552,0.1895648539,0.1023742184,0.177362144,0.3391303718,-0.1183856875,0.2422783822,0.4870125949,-0.2876998186,0.2076154053,-0.1738831252,-0.1704434901,0.117872335,-0.1792330891,-0.3962234259,-0.4508716464,-0.151104331,0.2607577443,0.3716067374,0.1726300865,-0.0400325842,0.4205482006,-0.1183398589,-0.2045203298,0.1578614414,0.0993773267,-0.696569562,0.2649095654,0.040320579,-0.2331422865,0.3088751137,0.1726256162,-0.0531161949,-0.1818626076,-0.5868797302,-0.0214212835,0.256139636,0.1477552205,0.1491210312,0.0979867652,0.5046492219,-0.1231771484,0.0652968809,-0.0895049423,-0.0938149244,0.1784211695,-0.1206737459,0.1079897583,-0.0565727279,0.76003474,-0.0328369215,0.2919978499,-0.3816135228,0.0486582704,0.1910071373,-0.0224883631,0.3683617413,-0.2188501209,-0.0778906867,-0.0568119809,-0.0303967763,-0.7145969868,0.1775298268,0.059114892,-0.0627833083,-0.4945368171,0.0594473258,0.0791763738,-0.1280043721,0.07861536,0.1506466568,0.2830786109,0.3220365644,0.1692655534,-0.2706499398,0.1640763283,0.0838266909,0.2446197122,0.007387904,0.1024084464,-0.4074581563,-0.2934479117,0.1693006605,0.077125527,-0.1476099938,0.276824832,-0.0925848261,0.0361406356,0.2758313417,-0.043746192,0.2513382733,-0.1297478229,-0.0251546316,0.0629387796,-0.0492124856,-0.2024413347,-0.3432178795,0.1095079705,0.017467957,0.3631527722,0.1710249484,0.0696342215,0.0981349424,0.0642424822,0.1448426396,-0.3554892838,-0.1485963613,-0.6650857925,-0.0909764469,-0.3191138804,0.0148422476,-0.2876193523,-0.1067289859,-0.0378259644,0.1699755937,0.0927127153,0.2727770805,-0.3351663649,0.2794829309,0.0989113376,-0.3488745689,-0.0690423697,0.2968818247,0.3751766086,0.3203035891,0.2065782547,-0.4071089923,0.1150906384,0.1151577011,-0.0066588116,0.0705892146,0.3163350523,-0.0844659507,-0.3517905176,0.5517035127,0.2774515748,-0.0388038531,-0.1616735607,-0.0258169882,-0.1697948426,0.2110009044,-0.3693974912,0.370582819,-0.1526387483,0.1521832347,-0.4862380624,0.1542181671,-0.0045062765,0.4191434979,0.0522385649,1.0550314188,0.0257998165,0.1309543848,0.0051356493,-0.1059452966,0.479355067,-0.5588315129,0.2374621481,-0.0598274879,0.0512066521,-0.0542153306,-0.000292077,0.3729058504,0.2897644937,-0.0662990287,0.3334056735,-0.1074233949,0.1965206712,-0.0245396588,-0.2288120389,0.3103707433,-0.2573020458,-1.011077404,0.0481126904,0.2008790374,0.0685821027,-0.4254418314,-0.2625392377,0.1260493398,0.0517849028,-0.0037790886,-0.1718616486,-0.3710801899,0.0607311204,-0.1598017514,0.0400771573,-0.1811505556,0.0683309287,0.3599056602,0.1948922724,-0.1812684834,-0.1069085747,0.0269829705,-0.3582991064,-0.1135511026,0.2985422015,-0.0603783838,-0.3018140197,-0.2927597463,-0.1994758397,-0.2634775341,-0.0027642876,0.3410246074,-0.2400836051,0.1370644122,0.1434681863,0.0743674487,-0.0086697536,0.2809583545,-0.1125468537,0.0409221016,0.1026950628,-0.1081101224,0.2284501493,0.1589966863,-0.078796804,-0.2242346704,0.3580927551,-0.0433431417,0.2658289075,0.2737148404,0.3521732986,-0.2044046968,-0.2140547037,0.2711829841,-0.3227516115,-0.3032827079,0.2143535614,0.1450777203,0.1297001839,-0.1516969949,-0.0194952916,0.3032157123,-0.115248464,-0.1638211608,-0.4914746583,-0.5047595501,0.103902556,0.0834252238,-0.2941827774,0.2828801572,0.1400387734,0.309204191,-0.0938187465,-0.1720235348,0.0793830827,-0.1266270727,0.0522203483,0.253800869,-0.0700948089,0.207263127,0.1504088193,-0.0250734538,-0.007144358,-0.1429828405,-0.0962937549,0.0138003072,0.2527059913,0.1756492257,-0.4706896544,-0.1229942217,-0.4690316021,-0.4169414043,0.0578682832,-0.1629226655,-0.0034416427,0.0561903864,-0.1688565761,0.2602782845,0.2026125342,-0.4410692751,0.0401295982,-0.0016458052,0.3058702052,0.1766481847,-0.1100077629,0.1643206626,0.0053441939,-0.0037486171,-0.1174772605,-0.1487119794,0.1736080498,0.717658639,-0.3057673872,0.2996409833,0.1940703094,0.2246846557,0.2971052229,0.1905901134,-0.0541578569,0.4123796523,0.0704125464,-0.4994432628,0.0125956936,0.183672443,0.048441954,0.0121509023,0.1410067081,0.2081937194,0.057200741,-0.1602872759,0.2355940342,0.2667345405,-0.042387981,0.0478238799,-0.0055512046,-0.0702365562,0.2259914279,-0.1606919765,0.0124248108,0.0358020291,0.336899817,-0.176181674,0.0426650047,0.0880716294,-0.1517412215,-0.0000069223,-0.2870091498,0.0492887944,0.2996186018,0.3487866223,-0.0995432138,-0.4406186938,0.2084199935,-0.2626591623,-0.1658480465,-0.2455715239,0.176463753,-0.0306634884,0.0524702556,-0.426879704,-0.1589352638,-0.0060811653,0.2915115654,0.2397077978,0.1118709818,0.1691294611,0.360047698,-0.1350090504,-0.3335427344,-0.459785223,0.1924862564,-0.0544725657,-0.0243362822,0.0418315791,0.4390508831,-0.2516793013,-0.111335814,0.0650954172,0.2630604208,0.121608451,-0.1737196445,0.0108979205,0.0098242303,0.0950754955,-0.1498254389,-0.1764335036,0.417257905,0.4848113656,-0.0907847881,-0.0514540449,-0.0366767682,0.626896143,-0.1798520386,0.1380153447,-0.0175957847,0.1584319919,0.2334702909,-0.2039184719,-0.3364567757,-0.1063284501,-0.309235841,0.3028317392,0.4466798306,0.0231800955,-0.1234376729,-0.14738217,-0.0228737816,-0.1504836529,0.7381174564,0.197390005,0.0520588867,-0.2032816708,-0.1200924143,-0.2982132733,-0.1874388307,-0.3620199859,0.0337895416,-0.0823804215,0.0505466312,0.3819795549,0.2184433043,0.2313506901,-0.2517598569,-0.0496514067,0.1727845222,-0.1262926012,-0.03470378,-0.3399157226,0.2288373411,0.0046911896,-0.2396191508,0.1062944606,-0.2555373609,-0.0517653376,-0.0667810589,-0.0910547897,-0.321120739,0.2661161721,0.3059483469,0.1287091076,0.4071596563,0.3806912303,-0.2751890421,-0.2592385411,0.2345474362,-0.2205582559,0.1360311657,0.2204751372,0.33912009,-0.2149956524,0.0869021639,-0.4555115998,0.197889924,0.1485187262,-0.1088388339,-0.2396107465,-0.3907483816,0.2567366362,0.0355462171,0.1516326964,0.0520592667,0.1054296643,0.0214451384,-0.5895581841,0.1879620254,0.403672874,-0.2699086666,-0.1935707182,-0.5413028598,0.1640541106,-0.0048655118,0.1507671773,-0.6646758318,-0.017838059,0.1382295787,0.0461634062,-0.085822016,0.1881661266,0.1773776859,-0.0622323602,-0.0267473515,0.2104129046,-0.4070430696,-0.3941547573,-0.4229817986,-0.0887774006]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1724","title":"could not run models on a offline server successfully","comments":"Yes it's been available since datasets 1.3.0 !","body":"Hi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...","comment_length":8,"text":"could not run models on a offline server successfully\nHi, I really need your help about this.\r\nI am trying to fine-tuning a RoBERTa on a remote server, which is strictly banning internet. I try to install all the packages by hand and try to run run_mlm.py on the server. It works well on colab, but when I try to run it on this offline server, it shows:\r\n![image](https:\/\/user-images.githubusercontent.com\/49967236\/104276256-25a88600-546a-11eb-9776-8ec695dfa24e.png)\r\n\r\nis there anything I can do? Is it possible to download all the things in cache and upload it to the server? Please help me out...\nYes it's been available since datasets 1.3.0 !","embeddings":[-0.0941142291,0.0776090473,-0.0432079397,0.3600873947,0.3943435848,0.0129289189,0.4342003763,0.2662603855,-0.2084600329,0.0606552958,0.1521141827,0.0999783278,-0.2405201346,0.1086934954,-0.0267537162,-0.2937995791,-0.0702854618,0.2014566064,-0.1709294468,-0.2459276766,0.1499604136,-0.0695106983,-0.1060980707,0.2256887406,-0.2538661063,-0.4708199203,-0.0334440768,0.1971526295,-0.5598735809,-0.1613823622,0.1991155446,-0.2210645229,0.4581335187,0.3183014095,-0.0001236394,0.0941371024,-0.009975357,-0.1528132111,-0.1889753789,-0.0895487368,0.1458666176,-0.1315437555,0.2037657797,-0.3578217626,0.1922030002,0.2491594702,0.1115848869,-0.1717204899,0.1772670597,0.2240559161,0.0890417546,0.3521714807,-0.0313603468,-0.1053388491,-0.0708592162,-0.0410514809,-0.2386084944,0.3336293399,-0.2261980027,-0.1959319562,0.4061599374,0.2701774538,-0.40167135,0.2572107911,0.0228233822,-0.0151926419,0.2296027094,-0.2379100174,0.3653410375,0.462118566,0.1784139723,0.14204669,-0.0556448512,-0.0951534584,0.1466457695,-0.339572072,0.1861754805,0.4954666495,-0.1958934516,0.358080864,-0.4739685357,-0.2540710568,-0.2734104693,0.0690282732,0.028380692,0.3460478485,-0.1016759202,0.1193129122,0.0102082295,0.032624308,-0.1265606731,-0.0673760027,-0.0094350083,-0.0988257378,-0.0828611553,0.0532662831,-0.126505971,0.4474729896,-0.0340845548,0.1615380794,0.1217615083,-0.1269039214,0.0810280815,0.0732716843,0.0175083149,0.2052370757,-0.136899516,-0.0400558859,0.3541574478,0.3913927376,-0.2557724416,-0.1321205646,-0.2081843466,-0.0170975104,-0.3028604984,-0.0998328403,0.0921024457,-0.2819167376,0.0785094351,-0.0745651498,-0.1136770323,0.0476786345,-0.0230495241,0.2666521966,-0.2662287951,-0.1275903285,0.0081583355,0.1064950526,-0.2686209381,-0.5850533843,0.1320649981,-0.129142195,-0.1970578283,-0.004506825,0.133388266,-0.1124979779,0.3380158246,-0.2341366261,0.2066597342,-0.0604510568,0.1943563819,-0.434540987,0.202411592,0.5040166378,0.0996874794,-0.0519486926,0.3407735229,0.033178784,0.1871485412,0.1822204143,0.0326520875,-0.4391786456,0.3339382708,0.0034271909,-0.2215426713,-0.2111340016,-0.2619789541,0.1576838046,0.051075276,-0.2386934608,0.0821101218,0.1078816056,-0.1146372035,-0.0918230712,0.5263264775,0.5031126738,-0.658403933,0.3725267351,-0.3807784915,0.0957422182,0.2089616358,0.2561107576,-0.188240096,0.4663667381,-0.2856932878,-0.3418555558,0.2451175898,0.0563192405,-0.119434379,0.2214620411,-0.0822608024,-0.3452470005,0.0466268808,-0.1610066146,0.3952846825,0.3230416179,-0.2277513146,0.112780571,-0.1153735742,-0.2114453614,-0.2925620079,-0.4429679811,0.170912087,0.2356825769,0.3546543419,0.2689455748,0.0460968018,-0.4505620897,0.1018867046,0.0806470215,-0.0635531396,-0.0904437751,0.158730343,0.2996402383,-0.0586504005,-0.0684935004,-0.0865686536,0.1993212104,-0.0321777537,0.4749881625,-0.339101553,0.2488710135,0.0532758571,-0.0042396099,-0.1202668473,0.0265205875,-0.0224010758,-0.2484909743,0.0265161581,0.0801827088,-0.2355669141,0.0550657846,0.0488048121,0.1152842492,-0.0728716105,0.3316210508,-0.2279637009,-0.1074149087,-0.1865680218,-0.3211731017,0.0881181583,-0.0085913362,-0.2082027346,0.2802807093,-0.1660204679,-0.0725163221,0.4680640697,0.1557986885,0.1135814264,-0.1198147312,0.4982288182,0.2440752238,0.09654589,0.1768866777,-0.1533230394,0.5068509579,0.0963157937,0.3492538631,0.0082206698,0.2312940806,0.313329339,0.1182263121,-0.0656548962,-0.2025182098,0.2544747889,0.6261815429,0.0219052229,-0.3110753,-0.1960979104,-0.078338556,0.1324680597,0.1386784762,0.2783958614,-0.2056775391,-0.1225402579,0.118358627,0.1908654273,0.2186316103,0.2269233167,0.1525943875,-0.1572724879,0.3139649034,-0.1621211916,-0.1859517097,0.3026895225,0.2959276736,0.436075598,-0.0372676961,0.0906640738,-0.2599167228,-0.3684539199,-0.2604209483,0.0877945647,0.2827498019,-0.0999044478,0.320057869,-0.5990613103,0.0740047991,-0.086475201,0.1329036206,-0.7361245155,0.1990461051,0.0987612903,-0.0914353281,-0.0624203458,0.2633652091,-0.1001046151,0.4184595942,-0.1342913359,-0.0351187885,0.1046896279,-0.2882633209,-0.3265175521,-0.0662842169,0.2094566375,0.0464855619,0.3992811441,-0.0280120764,-0.2311965823,0.1436872184,-0.28703475,0.031749893,0.2411896288,0.0689458698,0.2410507202,0.3435763121,-0.1529985815,0.3361578584,0.4536802173,-0.3012375832,0.1776579022,-0.1949688643,-0.1609923989,0.1227254495,-0.1847421974,-0.2920963466,-0.434391439,-0.1193293706,0.2087549269,0.319786638,0.1471485645,0.0037015441,0.4214481115,-0.0830111504,-0.2302472442,0.0969806537,0.0433746949,-0.7001066804,0.3550785184,0.0561823696,-0.2334762514,0.3480248153,0.1813387424,-0.0748430565,-0.0651671588,-0.6168518662,-0.0272768866,0.2354542613,0.1050933003,0.17873317,0.1106930077,0.5212643743,-0.1439025104,0.0714739189,-0.1422813982,-0.1106195971,0.1686727256,-0.1658249199,0.1760618687,-0.0987560004,0.712022543,0.0619273186,0.3173947334,-0.3562927544,0.0514531955,0.1627572328,-0.0712339804,0.4017713666,-0.1772922873,-0.0645442456,-0.0147004547,0.0430491045,-0.7054393291,0.1997813731,0.1270674914,-0.0676418245,-0.4708611369,0.1038971543,0.1338897198,-0.097622931,0.0405800492,0.168668136,0.2709746957,0.372754693,0.1826903075,-0.2779745758,0.1128133759,0.0824330747,0.2683256269,0.0449219942,0.1109259725,-0.4913359582,-0.2687451243,0.1865476221,0.1091290787,-0.1887910217,0.241161108,-0.1256823987,0.0570518635,0.3168942928,-0.0500776246,0.3645352423,-0.1754654646,0.0082860542,0.0252648015,-0.0073348354,-0.2162702084,-0.3182400167,0.1187983751,0.0130534302,0.3281352222,0.134835735,0.048181545,0.1431266218,0.0911048502,0.1136414334,-0.3559308946,-0.1556683481,-0.6100108624,-0.1267460585,-0.2938163579,-0.0131862545,-0.2828558981,-0.1885072291,-0.072817266,0.1417305768,0.0128387064,0.2810969353,-0.2712682188,0.3250031769,0.1147596836,-0.316768527,-0.0589620396,0.3639217317,0.3146078885,0.3214507699,0.2555961609,-0.3659088612,0.1041106433,0.1411313266,-0.0083219027,0.0716832206,0.3456017077,-0.0602267683,-0.3701354861,0.5397409797,0.2206541002,-0.0093958946,-0.1270764768,-0.0359305963,-0.1572833359,0.1759513915,-0.4208221436,0.3437187374,-0.0874977112,0.1074816957,-0.4229723513,0.1282842755,-0.0138739748,0.3649594486,0.050703574,1.1294602156,0.0166295469,0.1691722721,0.053311009,-0.1007464305,0.4897571206,-0.4986493886,0.2317555398,-0.1039595306,0.0283547807,-0.0860179886,0.0094236936,0.3769072592,0.3425270319,-0.0618906952,0.2874660194,-0.1084730476,0.1625212878,0.0297233183,-0.2965026498,0.3452199101,-0.2518599033,-0.992141068,0.0383210443,0.1907066405,0.0246467162,-0.3914364278,-0.2337587029,0.1407593638,0.0814214945,-0.0312540717,-0.1332419813,-0.4179923534,0.0558086671,-0.187783435,-0.0095422054,-0.158883065,0.0980611071,0.3387007713,0.2117675394,-0.1709176302,0.0010137371,0.075367339,-0.3472282588,-0.0849491134,0.2685735524,-0.0800981596,-0.2946327627,-0.2551591396,-0.1719483584,-0.1951818168,-0.0388709009,0.360760361,-0.2744850218,0.107859768,0.144332096,0.0492657945,-0.004867929,0.2705987096,-0.1275309324,0.0176746137,0.1386919022,-0.0968632624,0.2189687192,0.1783683151,-0.0450276881,-0.1658144295,0.4075103998,-0.100190796,0.2825689018,0.2648198307,0.364828974,-0.1806016564,-0.2273273468,0.2695282698,-0.289545238,-0.3742438257,0.2360688746,0.0836438835,0.1607474387,-0.1298590451,-0.0656962991,0.2780749202,-0.1770270765,-0.152379781,-0.5111489892,-0.4836502373,0.0654865578,0.0075664818,-0.2577106357,0.2662904561,0.0555381812,0.2505366206,-0.1265422255,-0.1603222787,0.1130369455,-0.0674086064,0.0086566238,0.2567279041,-0.0472387001,0.2122627646,0.1008411795,-0.0568953566,0.0167604387,-0.1915498376,-0.0727068484,0.0024880164,0.2759255171,0.156119734,-0.4878827631,-0.0943853408,-0.481946677,-0.45654127,0.0461665131,-0.1424895972,-0.0063157971,0.0326725841,-0.1793730706,0.2197674811,0.1509070396,-0.4406446517,0.0178226475,0.0497179963,0.2805906236,0.2041791826,-0.1201151162,0.1267191023,-0.0590188541,0.0239112973,-0.0872384533,-0.1700592041,0.148934707,0.73741436,-0.3320868909,0.3225134909,0.1644751281,0.2714930475,0.2393150628,0.1344933659,-0.0655556917,0.4088085592,0.0487751886,-0.4963769317,0.0584817566,0.209338963,0.1011372805,-0.0292812902,0.14138785,0.1916953623,0.1390230507,-0.1810312271,0.2433232963,0.3300098777,-0.0818552822,0.0326420777,-0.0250584856,-0.1311312467,0.2053777575,-0.0847030729,-0.0120882895,0.0498908237,0.3922174275,-0.1928302646,0.1041701734,0.1835770756,-0.1519281268,0.0311038401,-0.3583935201,-0.0097895786,0.2967861891,0.3949942887,-0.0818708986,-0.4067247808,0.2211034,-0.263240099,-0.1854548454,-0.2435574234,0.13678132,0.003847254,0.0715304315,-0.4648035765,-0.1992740184,-0.0004306941,0.2927929461,0.1717375219,0.121154286,0.2722755373,0.335285306,-0.1305299699,-0.4125495851,-0.420543462,0.2396733463,-0.0105639538,-0.0322136097,0.0959290415,0.4178256094,-0.2064186037,0.0482645445,0.1110376343,0.2752015889,0.088925004,-0.0810531303,-0.0023128244,-0.0427520499,0.0857499689,-0.1825574636,-0.1587188989,0.4183456004,0.4359100163,-0.1016077995,-0.0668420568,-0.0173558053,0.6537755728,-0.2010898143,0.1366944909,-0.0242347699,0.0818175972,0.214002043,-0.1987288743,-0.3147359192,-0.1018864587,-0.2664402723,0.2768242657,0.4659196138,-0.0311907697,-0.135937795,-0.1580657363,-0.0333126895,-0.1840001345,0.6862232089,0.2411717623,0.0595503859,-0.1684007794,-0.0978973955,-0.3647304177,-0.1682173461,-0.3416318297,-0.0529919006,-0.0686264038,0.0492494367,0.3839150965,0.1726431251,0.340975672,-0.2794887424,0.0800529122,0.1920221746,-0.1259761602,-0.0314094499,-0.2799803019,0.2520031631,-0.0326594077,-0.2365086824,0.1536411941,-0.261832267,-0.0925014094,-0.0346201956,-0.0980858803,-0.3278577626,0.2436831743,0.3398918509,0.1507662088,0.4461238682,0.372544229,-0.2172729671,-0.3056395054,0.2131997496,-0.1934088767,0.0873510987,0.2707689106,0.3878370523,-0.2208674252,0.1105461344,-0.4202245474,0.2055381984,0.1343578696,-0.0596520938,-0.2636010051,-0.3232772052,0.2133378834,0.0741864219,0.1324398071,0.0411358066,0.1048027873,-0.0451271944,-0.5618829131,0.1636462063,0.4735971689,-0.31179443,-0.2124941349,-0.5306079388,0.1716252118,0.0508395657,0.1060438976,-0.7215576172,-0.0574320965,0.1280388087,0.0513115339,-0.1207855344,0.2367852628,0.1600075513,-0.045071885,0.004435278,0.2954466641,-0.4146332741,-0.4464059174,-0.4346722662,-0.0830519497]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"Thanks for reporting !\r\nI was able to reproduce thanks to your code and find the origin of the bug.\r\nThe cache was not reusing the same file because one object was not deterministic. It comes from a conversion from `set` to `list` in the `datasets.arrrow_dataset.transmit_format` function, where the resulting list would not always be in the same order and therefore the function that computes the hash used by the cache would not always return the same result.\r\nI'm opening a PR to fix this.\r\n\r\nAlso we plan to do a new release in the coming days so you can expect the fix to be available soon.\r\nNote that you can still specify `cache_file_name=` in the second `map()` call to name the cache file yourself if you want to.","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":129,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nThanks for reporting !\r\nI was able to reproduce thanks to your code and find the origin of the bug.\r\nThe cache was not reusing the same file because one object was not deterministic. It comes from a conversion from `set` to `list` in the `datasets.arrrow_dataset.transmit_format` function, where the resulting list would not always be in the same order and therefore the function that computes the hash used by the cache would not always return the same result.\r\nI'm opening a PR to fix this.\r\n\r\nAlso we plan to do a new release in the coming days so you can expect the fix to be available soon.\r\nNote that you can still specify `cache_file_name=` in the second `map()` call to name the cache file yourself if you want to.","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"Thanks for the fast reply, waiting for the fix :)\r\n\r\nI tried to use `cache_file_names` and wasn't sure how, I tried to give it the following:\r\n```\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n cache_file_names={k: f'.cache\/{str(k)}' for k in tokenized_datasets}\r\n)\r\n```\r\n\r\nand got an error:\r\n```\r\nmultiprocess.pool.RemoteTraceback:\r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/venv\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 157, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1491, in _map_single\r\n tmp_file = tempfile.NamedTemporaryFile(\"wb\", dir=os.path.dirname(cache_file_name), delete=False)\r\n File \"\/usr\/lib\/python3.6\/tempfile.py\", line 690, in NamedTemporaryFile\r\n (fd, name) = _mkstemp_inner(dir, prefix, suffix, flags, output_type)\r\n File \"\/usr\/lib\/python3.6\/tempfile.py\", line 401, in _mkstemp_inner\r\n fd = _os.open(file, flags, 0o600)\r\nFileNotFoundError: [Errno 2] No such file or directory: '_00000_of_00060.cache\/tmpsvszxtop'\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 48, in <module>\r\n cache_file_names={k: f'.cache\/{str(k)}' for k in tokenized_datasets}\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1317, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1317, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/venv\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\nFileNotFoundError: [Errno 2] No such file or directory: '_00000_of_00060.cache\/tmpsvszxtop'\r\n```\r\n","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":229,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nThanks for the fast reply, waiting for the fix :)\r\n\r\nI tried to use `cache_file_names` and wasn't sure how, I tried to give it the following:\r\n```\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n cache_file_names={k: f'.cache\/{str(k)}' for k in tokenized_datasets}\r\n)\r\n```\r\n\r\nand got an error:\r\n```\r\nmultiprocess.pool.RemoteTraceback:\r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/venv\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 157, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1491, in _map_single\r\n tmp_file = tempfile.NamedTemporaryFile(\"wb\", dir=os.path.dirname(cache_file_name), delete=False)\r\n File \"\/usr\/lib\/python3.6\/tempfile.py\", line 690, in NamedTemporaryFile\r\n (fd, name) = _mkstemp_inner(dir, prefix, suffix, flags, output_type)\r\n File \"\/usr\/lib\/python3.6\/tempfile.py\", line 401, in _mkstemp_inner\r\n fd = _os.open(file, flags, 0o600)\r\nFileNotFoundError: [Errno 2] No such file or directory: '_00000_of_00060.cache\/tmpsvszxtop'\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 48, in <module>\r\n cache_file_names={k: f'.cache\/{str(k)}' for k in tokenized_datasets}\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in map\r\n for k, dataset in self.items()\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 303, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1317, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/venv\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1317, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/venv\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 644, in get\r\n raise self._value\r\nFileNotFoundError: [Errno 2] No such file or directory: '_00000_of_00060.cache\/tmpsvszxtop'\r\n```\r\n","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"The documentation says\r\n```\r\ncache_file_names (`Optional[Dict[str, str]]`, defaults to `None`): Provide the name of a cache file to use to store the\r\n results of the computation instead of the automatically generated cache file name.\r\n You have to provide one :obj:`cache_file_name` per dataset in the dataset dictionary.\r\n```\r\nWhat is expected is simply the name of a file, not a path. The file will be located in the cache directory of the `wikitext` dataset. You can try again with something like\r\n```python\r\ncache_file_names = {k: f'tokenized_and_grouped_{str(k)}' for k in tokenized_datasets}\r\n```","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":90,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nThe documentation says\r\n```\r\ncache_file_names (`Optional[Dict[str, str]]`, defaults to `None`): Provide the name of a cache file to use to store the\r\n results of the computation instead of the automatically generated cache file name.\r\n You have to provide one :obj:`cache_file_name` per dataset in the dataset dictionary.\r\n```\r\nWhat is expected is simply the name of a file, not a path. The file will be located in the cache directory of the `wikitext` dataset. You can try again with something like\r\n```python\r\ncache_file_names = {k: f'tokenized_and_grouped_{str(k)}' for k in tokenized_datasets}\r\n```","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"Managed to get `cache_file_names` working and caching works well with it\r\nHad to make a small modification for it to work:\r\n```\r\ncache_file_names = {k: f'tokenized_and_grouped_{str(k)}.arrow' for k in tokenized_datasets}\r\n```","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":31,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nManaged to get `cache_file_names` working and caching works well with it\r\nHad to make a small modification for it to work:\r\n```\r\ncache_file_names = {k: f'tokenized_and_grouped_{str(k)}.arrow' for k in tokenized_datasets}\r\n```","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"Another comment on `cache_file_names`, it doesn't save the produced cached files in the dataset's cache folder, it requires to give a path to an existing directory for it to work.\r\nI can confirm that this is how it works in `datasets==1.1.3`","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":41,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nAnother comment on `cache_file_names`, it doesn't save the produced cached files in the dataset's cache folder, it requires to give a path to an existing directory for it to work.\r\nI can confirm that this is how it works in `datasets==1.1.3`","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"Oh yes indeed ! Maybe we need to update the docstring to mention that it is a path","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":18,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nOh yes indeed ! Maybe we need to update the docstring to mention that it is a path","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"I fixed the docstring. Hopefully this is less confusing now: https:\/\/github.com\/huggingface\/datasets\/commit\/42ccc0012ba8864e6db1392430100f350236183a","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":11,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nI fixed the docstring. Hopefully this is less confusing now: https:\/\/github.com\/huggingface\/datasets\/commit\/42ccc0012ba8864e6db1392430100f350236183a","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"I upgraded to the latest version and I encountered some strange behaviour, the script I posted in the OP doesn't trigger recalculation, however, if I add the following change it does trigger partial recalculation, I am not sure if its something wrong on my machine or a bug:\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n# CHANGE\r\nprint('hello')\r\n# CHANGE\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n...\r\n```\r\nI am using datasets in the `run_mlm.py` script in the transformers examples and I found that if I change the script without touching any of the preprocessing. it still triggers recalculation which is very weird\r\n\r\nEdit: accidently clicked the close issue button ","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":136,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nI upgraded to the latest version and I encountered some strange behaviour, the script I posted in the OP doesn't trigger recalculation, however, if I add the following change it does trigger partial recalculation, I am not sure if its something wrong on my machine or a bug:\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n# CHANGE\r\nprint('hello')\r\n# CHANGE\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n...\r\n```\r\nI am using datasets in the `run_mlm.py` script in the transformers examples and I found that if I change the script without touching any of the preprocessing. it still triggers recalculation which is very weird\r\n\r\nEdit: accidently clicked the close issue button ","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"This is because the `group_texts` line definition changes (it is defined 3 lines later than in the previous call). Currently if a function is moved elsewhere in a script we consider it to be different.\r\n\r\nNot sure this is actually a good idea to keep this behavior though. We had this as a security in the early development of the lib but now the recursive hashing of objects is robust so we can probably remove that.\r\nMoreover we're already ignoring the line definition for lambda functions.","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":86,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nThis is because the `group_texts` line definition changes (it is defined 3 lines later than in the previous call). Currently if a function is moved elsewhere in a script we consider it to be different.\r\n\r\nNot sure this is actually a good idea to keep this behavior though. We had this as a security in the early development of the lib but now the recursive hashing of objects is robust so we can probably remove that.\r\nMoreover we're already ignoring the line definition for lambda functions.","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"I opened a PR to change this, let me know what you think.","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":13,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nI opened a PR to change this, let me know what you think.","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"Sounds great, thank you for your quick responses and help! Looking forward for the next release.","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":16,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nSounds great, thank you for your quick responses and help! Looking forward for the next release.","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"I am having a similar issue where only the grouped files are loaded from cache while the tokenized ones aren't. I can confirm both datasets are being stored to file, but only the grouped version is loaded from cache. Not sure what might be going on. But I've tried to remove all kinds of non deterministic behaviour, but still no luck. Thanks for the help!\r\n\r\n\r\n```python\r\n # Datasets\r\n train = sorted(glob(args.data_dir + '*.{}'.format(args.ext)))\r\n if args.dev_split >= len(train):\r\n raise ValueError(\"Not enough dev files\")\r\n dev = []\r\n state = random.Random(1001)\r\n for _ in range(args.dev_split):\r\n dev.append(train.pop(state.randint(0, len(train) - 1)))\r\n\r\n max_seq_length = min(args.max_seq_length, tokenizer.model_max_length)\r\n\r\n def tokenize_function(examples):\r\n return tokenizer(examples['text'], return_special_tokens_mask=True)\r\n\r\n def group_texts(examples):\r\n # Concatenate all texts from our dataset and generate chunks of max_seq_length\r\n concatenated_examples = {k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # Truncate (not implementing padding)\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_seq_length\r\n result = {\r\n k: [t[i : i + max_seq_length] for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\n datasets = load_dataset(\r\n 'text', name='DBNL', data_files={'train': train[:10], 'dev': dev[:5]}, \r\n cache_dir=args.data_cache_dir)\r\n datasets = datasets.map(tokenize_function, \r\n batched=True, remove_columns=['text'], \r\n cache_file_names={k: os.path.join(args.data_cache_dir, f'{k}-tokenized') for k in datasets},\r\n load_from_cache_file=not args.overwrite_cache)\r\n datasets = datasets.map(group_texts, \r\n batched=True,\r\n cache_file_names={k: os.path.join(args.data_cache_dir, f'{k}-grouped') for k in datasets},\r\n load_from_cache_file=not args.overwrite_cache)\r\n```\r\n\r\nAnd this is the log\r\n\r\n```\r\n04\/26\/2021 10:26:59 - WARNING - datasets.builder - Using custom data configuration DBNL-f8d988ad33ccf2c1\r\n04\/26\/2021 10:26:59 - WARNING - datasets.builder - Reusing dataset text (\/home\/manjavacasema\/data\/.cache\/text\/DBNL-f8d988ad33ccf2c1\/0.0.0\/e16f44aa1b321ece1f87b07977cc5d70be93d69b20486d6dacd62e12cf25c9a5)\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 13\/13 [00:00<00:00, 21.07ba\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 40\/40 [00:01<00:00, 24.28ba\/s]\r\n04\/26\/2021 10:27:01 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at \/home\/manjavacasema\/data\/.cache\/train-grouped\r\n04\/26\/2021 10:27:01 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at \/home\/manjavacasema\/data\/.cache\/dev-grouped\r\n```\r\n","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":274,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nI am having a similar issue where only the grouped files are loaded from cache while the tokenized ones aren't. I can confirm both datasets are being stored to file, but only the grouped version is loaded from cache. Not sure what might be going on. But I've tried to remove all kinds of non deterministic behaviour, but still no luck. Thanks for the help!\r\n\r\n\r\n```python\r\n # Datasets\r\n train = sorted(glob(args.data_dir + '*.{}'.format(args.ext)))\r\n if args.dev_split >= len(train):\r\n raise ValueError(\"Not enough dev files\")\r\n dev = []\r\n state = random.Random(1001)\r\n for _ in range(args.dev_split):\r\n dev.append(train.pop(state.randint(0, len(train) - 1)))\r\n\r\n max_seq_length = min(args.max_seq_length, tokenizer.model_max_length)\r\n\r\n def tokenize_function(examples):\r\n return tokenizer(examples['text'], return_special_tokens_mask=True)\r\n\r\n def group_texts(examples):\r\n # Concatenate all texts from our dataset and generate chunks of max_seq_length\r\n concatenated_examples = {k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # Truncate (not implementing padding)\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_seq_length\r\n result = {\r\n k: [t[i : i + max_seq_length] for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\n datasets = load_dataset(\r\n 'text', name='DBNL', data_files={'train': train[:10], 'dev': dev[:5]}, \r\n cache_dir=args.data_cache_dir)\r\n datasets = datasets.map(tokenize_function, \r\n batched=True, remove_columns=['text'], \r\n cache_file_names={k: os.path.join(args.data_cache_dir, f'{k}-tokenized') for k in datasets},\r\n load_from_cache_file=not args.overwrite_cache)\r\n datasets = datasets.map(group_texts, \r\n batched=True,\r\n cache_file_names={k: os.path.join(args.data_cache_dir, f'{k}-grouped') for k in datasets},\r\n load_from_cache_file=not args.overwrite_cache)\r\n```\r\n\r\nAnd this is the log\r\n\r\n```\r\n04\/26\/2021 10:26:59 - WARNING - datasets.builder - Using custom data configuration DBNL-f8d988ad33ccf2c1\r\n04\/26\/2021 10:26:59 - WARNING - datasets.builder - Reusing dataset text (\/home\/manjavacasema\/data\/.cache\/text\/DBNL-f8d988ad33ccf2c1\/0.0.0\/e16f44aa1b321ece1f87b07977cc5d70be93d69b20486d6dacd62e12cf25c9a5)\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 13\/13 [00:00<00:00, 21.07ba\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 40\/40 [00:01<00:00, 24.28ba\/s]\r\n04\/26\/2021 10:27:01 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at \/home\/manjavacasema\/data\/.cache\/train-grouped\r\n04\/26\/2021 10:27:01 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at \/home\/manjavacasema\/data\/.cache\/dev-grouped\r\n```\r\n","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"Hi ! What tokenizer are you using ?","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":8,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nHi ! What tokenizer are you using ?","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1718","title":"Possible cache miss in datasets","comments":"It's the ByteLevelBPETokenizer","body":"Hi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```","comment_length":3,"text":"Possible cache miss in datasets\nHi,\r\n\r\nI am using the datasets package and even though I run the same data processing functions, datasets always recomputes the function instead of using cache.\r\nI have attached an example script that for me reproduces the problem.\r\nIn the attached example the second map function always recomputes instead of loading from cache.\r\nIs this a bug or am I doing something wrong?\r\nIs there a way for fix this and avoid all the recomputation?\r\n\r\nThanks\r\n\r\nEdit:\r\ntransformers==3.5.1\r\ndatasets==1.2.0\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\n\r\ndatasets = load_dataset('wikitext', 'wikitext-103-raw-v1')\r\ntokenizer = AutoTokenizer.from_pretrained('bert-base-uncased', use_fast=True)\r\n\r\n\r\ncolumn_names = datasets[\"train\"].column_names\r\ntext_column_name = \"text\" if \"text\" in column_names else column_names[0]\r\ndef tokenize_function(examples):\r\n return tokenizer(examples[text_column_name], return_special_tokens_mask=True)\r\n\r\ntokenized_datasets = datasets.map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=60,\r\n remove_columns=[text_column_name],\r\n load_from_cache_file=True,\r\n)\r\nmax_seq_length = tokenizer.model_max_length\r\ndef group_texts(examples):\r\n # Concatenate all texts.\r\n concatenated_examples = {\r\n k: sum(examples[k], []) for k in examples.keys()}\r\n total_length = len(concatenated_examples[list(examples.keys())[0]])\r\n # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can\r\n # customize this part to your needs.\r\n total_length = (total_length \/\/ max_seq_length) * max_seq_length\r\n # Split by chunks of max_len.\r\n result = {\r\n k: [t[i: i + max_seq_length]\r\n for i in range(0, total_length, max_seq_length)]\r\n for k, t in concatenated_examples.items()\r\n }\r\n return result\r\n\r\ntokenized_datasets = tokenized_datasets.map(\r\n group_texts,\r\n batched=True,\r\n num_proc=60,\r\n load_from_cache_file=True,\r\n)\r\nprint(tokenized_datasets)\r\n\r\nprint('finished')\r\n```\nIt's the ByteLevelBPETokenizer","embeddings":[-0.1930393726,0.1210535541,0.0500952005,0.1900934726,0.0076969219,0.0369213931,-0.0168548934,0.2303498536,0.0167332795,-0.1084433347,0.1029985696,0.4139710963,0.2249117643,-0.1117237061,-0.0011100402,0.2446131408,0.3074613512,0.3216940165,-0.1533999145,-0.2088759243,-0.2891754806,0.1171123162,-0.3328206539,-0.092429027,-0.4170501232,0.2123280317,-0.0168774482,-0.4208398163,0.1122606322,-0.2555866539,0.3177160025,0.0858283564,0.1992189288,0.5570235252,-0.0001258203,-0.2676019073,0.3540959358,-0.0694738552,-0.2052473277,0.0768211856,-0.2844775021,0.0380443037,-0.1126935929,0.0233671591,0.0584392846,0.0587090738,0.0379663296,-0.6514899731,0.3060197532,0.1629111767,0.102592662,0.033276476,-0.0953430757,0.2345797122,0.0502028316,-0.0120253302,-0.0387438312,-0.3330449462,0.1630381346,-0.3523213267,0.0621471033,0.5163400769,-0.1757435054,-0.1419030875,0.3870675266,-0.0117806941,0.0179351512,-0.2646991313,0.3020609319,-0.0644705594,0.5929020047,-0.3732129037,-0.4453589916,-0.2218152285,-0.2183965743,-0.1797114164,0.3478957415,-0.2814225554,0.115124248,0.1367774606,-0.5952916145,-0.3717092276,0.2665030062,0.1268505901,-0.0126040196,0.118829295,-0.0442643128,0.1275966167,-0.2583603263,0.0305633396,0.278693229,-0.5107082129,-0.020411972,0.4761810005,-0.4573765695,-0.114198938,-0.0813957304,0.0444577448,0.031320218,0.2521778643,-0.1078691259,0.0848287344,-0.1045756191,-0.0248278677,0.1718008816,0.3819052577,0.1599488705,0.3959278464,-0.0466323309,-0.2572212815,-0.4361515045,-0.1423511654,0.1015104204,0.0623166673,0.537632525,0.0291995946,-0.2670723796,-0.1527193785,0.1350922734,0.1104742214,-0.1131508872,-0.154272154,0.020687161,0.1861227155,-0.2187203914,0.5657741427,-0.0489024818,0.0113220746,-0.2269070297,-0.0754960403,-0.2520525157,-0.1949410737,-0.2797417641,0.4240118265,0.107980758,-0.0087579098,0.1931643337,0.1137457117,0.1412912309,-0.2485303879,0.2450801283,-0.1952346861,0.1446888447,0.0313405395,-0.122426711,0.5588548183,0.0238276497,0.0820711404,-0.0578298308,0.2793508768,-0.3909225166,-0.0032497863,0.2678503692,0.0766011178,-0.1049885824,-0.051326897,-0.1365459859,-0.0477274917,0.6230661273,-0.3534024954,0.0028586881,-0.1369546503,-0.2181598991,-0.3552233577,0.1964023113,0.4151437283,-0.2203010619,-0.1271440536,-0.2892490625,0.4262260199,0.3456280828,0.1904119849,-0.0997599214,0.2238963842,-0.2436235696,0.2533464134,0.3463941514,-0.3264125288,-0.7522772551,-0.0623180978,0.2395538688,0.2138240188,-0.2641494572,0.0091178091,0.0585763827,0.0220128987,-0.0473067649,0.1247596368,0.0710947663,0.0479407869,-0.1934205443,-0.3237839639,0.3307866454,-0.1534485519,0.373434782,0.2039108723,-0.2786331475,-0.0654283091,0.1897790283,-0.0733121186,0.204567641,0.2111120969,0.0859391093,0.0963291302,0.3425368965,-0.144525975,-0.2208454311,0.2361599505,-0.0074976902,-0.4080512822,0.05238121,-0.0693818182,0.0342106409,-0.1751439422,-0.2368721217,-0.250767231,0.0469596423,0.2982925773,0.2158326656,0.0493254699,-0.0628557652,0.4346975684,0.2247962803,-0.0474488139,-0.1369978189,0.2314382493,0.1378540844,-0.0723532364,-0.2622561157,-0.007789142,0.3865086436,0.2050782442,-0.1728083491,0.1999227405,0.2821433246,0.1305397302,-0.1580862552,0.0377752185,-0.0964976549,0.076044932,-0.1803530306,0.0189615674,0.0912419036,-0.2024737746,0.3389697969,-0.0982242227,0.0435390286,0.1728527993,-0.1738719493,-0.1164959967,-0.0194160827,-0.0575414672,0.1215716228,-0.2605112791,-0.1267734617,-0.1007431298,0.4018777013,0.1480001211,0.3170870245,0.1431048512,0.5319442153,0.1638038456,0.0702496767,-0.0754961893,-0.595046699,-0.153305307,0.0420683324,0.1067333892,0.4031660557,0.0140680997,0.1372346729,-0.1065353453,0.1262332499,0.0536733977,0.0839905441,0.0515747443,0.0034557909,0.1623288542,0.0819173902,0.1239183173,-0.1544849575,0.3739548624,0.2223696411,0.0224084184,-0.1201334149,-0.0047830353,-0.4614660442,0.2616464496,-0.0094154151,-0.1877422035,-0.1089543253,-0.3240405321,-0.0293718316,0.1750336438,0.2323505729,0.1702650785,-0.1067733094,0.1637891084,-0.0965859368,-0.0634079278,-0.2861728072,-0.0442264751,-0.3516021371,-0.1043687165,0.0276290718,-0.3332431018,0.1092685685,-0.1978000402,0.0448775329,-0.3304952085,-0.2496137768,0.0687255412,0.1424214393,-0.191361472,-0.1454960704,-0.1331452131,-0.3406224549,-0.0777615905,0.1052851975,-0.3755034506,-0.0589957796,-0.1942424774,0.226897493,-0.0732832849,0.0613905266,-0.0987810716,0.0450629443,0.0523568131,-0.0922801569,-0.1205202043,-0.0991960391,-0.1073513255,-0.1778948605,-0.1473686546,-0.2940423191,-0.1225034297,-0.5240326524,-0.2998071015,0.536532402,-0.1247007027,-0.0793028027,-0.1053363308,-0.0294937789,0.3053979874,0.4665106535,-0.5257647634,-0.0645799488,-0.2027752846,-0.18987225,-0.1475555003,0.3041585684,0.3952634335,0.1911331415,0.1284653395,-0.4285719097,-0.2033012211,0.0207050201,0.001100692,0.4422153831,-0.0327308625,0.2021819651,-0.0818895251,0.8030892611,0.6181532145,-0.2177960426,0.2308448106,0.1147788912,0.3255800009,-0.1409276873,-0.197888568,-0.1367323548,-0.3422304988,0.0949928164,0.1833507419,0.1425249279,-0.3120464981,-0.1570544988,0.0919301361,-0.3112705052,-0.3297131956,-0.080523394,-0.248529613,0.3068063259,0.3510835767,0.3141096532,-0.7173701525,-0.1220384836,0.2528284788,-0.2144832611,0.4689786434,0.1147963554,-0.2109618038,-0.1388645768,-0.3443192542,0.3209452033,0.2275709808,0.366902113,0.2177835703,0.0174156725,-0.1508575678,0.1088577211,0.4315701127,-0.7976599932,-0.3380011618,-0.1114928275,-0.0554741435,-0.0210326798,-0.0406019278,0.3510410786,0.2257121801,0.0413003154,0.2744194567,0.0385528579,0.0075800805,-0.3233849108,0.1729340404,-0.2346344888,-0.3152335286,-0.098660022,0.2701252401,-0.0900599062,-0.2996428013,0.0286339186,-0.1689186394,0.0574233979,-0.1737730801,-0.3051756024,0.057996951,0.1066477746,0.0292743519,0.5556049347,-0.0282388832,0.3208510578,0.1760175824,0.0183290076,0.0711162463,0.2807700634,-0.1428882778,-0.2337372899,-0.0897049531,-0.0923274234,0.2570895851,-0.0319633856,-0.2056109309,-0.0013512475,-0.2553197443,0.0258840714,-0.2113858163,0.1346880794,0.3652509749,0.3597345948,-0.4407249689,-0.4705250859,0.1731275171,0.2737513483,-0.1833476275,0.5526421666,-0.0829963014,-0.2301497012,0.1558222175,0.0446283519,1.0017770529,0.0292428322,0.0918867812,-0.2499911636,0.0564776845,0.6650125384,-0.3729462028,0.3306760788,-0.2858806849,-0.2105312198,-0.1887110025,-0.2268438637,-0.0835621208,0.2917122841,-0.1994392872,0.4963091612,0.0540511943,0.5679221153,0.2249383479,0.0703188479,0.1906241328,-0.1199669912,0.2616660595,0.0731541961,0.1753248572,0.3935571313,-0.0909895673,0.111892134,0.0907490999,-0.0987972692,-0.2807958722,0.1061755121,-0.3655889034,0.253131032,-0.0505574681,0.1743299663,0.1004234403,0.3575809598,0.5060386062,0.2082125247,-0.1885528415,0.1358557492,0.1558084786,0.2617042661,-0.1411559433,-0.2263830453,0.0280737951,0.0002943175,-0.2079171538,-0.0109651573,0.0326269791,-0.4142291546,-0.3528572023,-0.0270258524,0.3227232099,-0.2036621124,0.0705916733,0.1881310195,0.0630478859,-0.0437566787,0.0380565785,-0.0221824739,0.0738399476,0.5688581467,-0.2717235088,-0.1392876804,-0.0132387551,0.4597799182,0.3361432254,-0.1777528524,0.5680013299,-0.2200241834,-0.2402041703,-0.0472222604,-0.4798617959,0.0584012866,-0.3148997128,-0.0378465839,-0.396669507,0.0605309904,0.1406712085,-0.0575221926,-0.0307454877,-0.1993784159,-0.1216959804,-0.2299207151,-0.2037149817,0.0421950184,0.0436034948,0.1631635129,0.2255757749,0.064470306,-0.0405305289,0.4189855158,-0.1535955071,0.0444021709,0.1507005692,0.1221961603,-0.2109900862,-0.0877635479,-0.0586923547,-0.1046089008,-0.1137624606,0.3114514947,-0.0639020652,-0.0521665886,0.0988849849,0.2019437551,0.2077478021,-0.0239151306,-0.0457028449,-0.018330561,0.0787012056,-0.3652362227,0.0914696753,0.0596851856,-0.0229968298,0.2353386879,0.3357806802,-0.0833123475,-0.2787672877,0.1575027704,-0.0676100925,0.0496588871,-0.0292313974,0.2691749036,0.363668263,0.0198298581,-0.0800043941,-0.0291804746,0.1131558567,0.0327559784,0.3520662785,-0.4142053127,0.0171418078,0.2618717551,-0.0939997137,0.3780742884,-0.125548631,-0.3512196541,0.2893672585,0.1102726683,-0.267742157,-0.3457253277,0.2865933478,0.0986227393,-0.1312385052,0.0857274085,0.0336204618,-0.1756941974,0.3264606595,0.10745015,0.4046726227,-0.435103327,0.3502946198,0.2615408897,0.2842255533,0.1460161656,0.1785676181,0.2187537551,-0.2382124811,0.16903162,0.1195477247,0.2626834214,0.1509537399,-0.0861337408,0.0670148283,-0.2940175235,-0.0199439023,0.3258394301,-0.104535833,0.3778481185,-0.0565288328,-0.126149103,0.1844799668,-0.0648226738,-0.2893035114,0.1713653952,-0.2577165961,-0.1373938471,0.2168833017,-0.160634011,-0.0198615715,-0.1444623172,0.1380321831,0.001731375,0.5208857059,0.0190923531,-0.1504765153,-0.11493963,-0.2739804089,0.1826743633,0.574524641,-0.1868020147,0.0926649496,-0.2138551474,-0.0239140857,-0.090123266,0.3141936362,0.1792688966,0.4027068317,-0.1473160088,0.0805873051,0.2573625743,0.0996952057,-0.036728967,0.4875122309,-0.1763319671,-0.3967134953,0.0521719418,0.0690327659,-0.0318287313,-0.1598775983,-0.2711258829,0.3679016531,-0.4220822752,0.302328229,-0.0954641998,0.0734807774,-0.0674493536,0.1810982376,-0.3221398592,0.1976938546,0.5176382661,-0.11355488,0.1017572209,-0.2707715631,0.0281984452,-0.0405584574,0.2850448489,0.4347339571,-0.2526247501,-0.207654193,-0.2799325585,-0.4617501199,0.3735391498,-0.1497468054,0.2867358029,0.0239861738,0.1044591516,-0.0519853346,-0.049172461,0.3006149828,0.1457595974,-0.2279871255,0.3076718748,-0.3471156657,-0.0405948274,-0.3000037372,0.0067512854,0.1716319621,-0.3455072641,0.4444896281,0.1955612153,0.027293751,-0.0233994164,0.0678267553,0.2191312015,-0.0610966384,0.545042932,0.3465767205,0.2369408756,-0.0209325254,-0.0934214145,-0.0949543193,-0.0210047159,-0.1207422614,0.0470128544,-0.1210353673,0.08204256,-0.4738941491,0.0651586577,-0.3022509515,-0.1956375837,0.2856243849,-0.0026543764,-0.1738129705,-0.1556400657,-0.0610377043,0.1276529729,0.1668199152,0.4713115692,-0.2290990055,0.4036515653,-0.2490559518,-0.1976542473,0.3329885304,-0.6163414717,-0.4550292492,-0.2239357829,0.1375330836,0.1176633537,-0.3350002766,-0.6715468764,-0.1810437441,0.3174129426,0.0883680359,-0.0337375924,0.1405457258,0.044409316,0.2483877391,-0.1456893086,0.5623004436,0.0550513268,0.0295160655,0.2277147025,-0.1859031618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1717","title":"SciFact dataset - minor changes","comments":"Hi Dave,\r\nYou are more than welcome to open a PR to make these changes! \ud83e\udd17\r\nYou will find the relevant information about opening a PR in the [contributing guide](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/CONTRIBUTING.md) and in the [dataset addition guide](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nPinging also @lhoestq for the Google cloud matter.","body":"Hi,\r\n\r\nSciFact dataset creator here. First of all, thanks for adding the dataset to Huggingface, much appreciated!\r\n\r\nI'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nIt also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nThanks,\r\n\r\nDave","comment_length":44,"text":"SciFact dataset - minor changes\nHi,\r\n\r\nSciFact dataset creator here. First of all, thanks for adding the dataset to Huggingface, much appreciated!\r\n\r\nI'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nIt also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nThanks,\r\n\r\nDave\nHi Dave,\r\nYou are more than welcome to open a PR to make these changes! \ud83e\udd17\r\nYou will find the relevant information about opening a PR in the [contributing guide](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/CONTRIBUTING.md) and in the [dataset addition guide](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nPinging also @lhoestq for the Google cloud matter.","embeddings":[0.1588933617,-0.1973490417,-0.0676539242,0.1470401138,0.1484657675,-0.1071610376,-0.2045498639,-0.0401365869,0.0873415917,-0.0903791636,-0.0821836367,0.0158945657,-0.0085386354,0.4948686659,0.1395404339,-0.2466709316,0.2298517078,-0.0996816456,-0.026662264,-0.2833099961,-0.0561362617,0.2272369713,0.1043683589,-0.1867742091,-0.0700056925,-0.2488317639,-0.3025515676,0.3253338635,-0.4019877315,-0.3336291313,-0.0090463301,0.4374462068,0.0153007675,0.3517856002,-0.0001106163,-0.1575024277,0.3256404996,-0.0009351738,-0.2085137516,-0.1126065403,-0.162084192,-0.096408911,0.0218258165,0.071520403,-0.0710576251,-0.0761889145,-0.1483071595,-0.124201335,0.3208521307,0.0993056372,0.2183046639,0.2332518697,0.179612264,-0.2070899904,-0.0511534885,0.2912590504,-0.0473824404,0.3453986347,0.1674585044,0.3302257061,0.025699567,0.4653010368,-0.0388631821,-0.1789093316,0.6320386529,0.0213729311,-0.3761971891,-0.4856018424,0.1486368775,0.2742673159,0.5723049641,-0.3517041504,-0.6798822284,-0.2079389691,0.2144465744,-0.073569268,-0.0122222416,0.316181004,0.0381379463,0.0511892289,0.0710301921,-0.4668088555,-0.3095673621,-0.1419863254,0.0620473512,0.0905595496,-0.1406484842,0.0051066061,0.1124816686,-0.0500589721,-0.0696471035,-0.2305473238,-0.1012077853,0.1105057597,-0.0996387154,-0.2565585971,-0.2176599503,0.297652036,0.0186595768,0.3842715025,0.0801841542,-0.0068234601,-0.2027323991,-0.0759955645,0.326421082,-0.0812568069,0.0278393105,0.1425428092,0.2957104146,0.1814731359,0.4126073122,-0.1415103227,0.1716012657,-0.1521790922,-0.4780347049,-0.1535412073,0.1730055064,-0.1835856438,0.1002775803,-0.0033972061,0.1850010604,-0.1852900684,-0.1225436181,0.2218345851,-0.1419371516,-0.0247634184,-0.0591177754,0.1648083031,0.0859883577,-0.2317878753,-0.2415515184,0.1570417434,0.2464354932,0.1684089154,0.0972013846,0.043922592,0.330231756,-0.0416658968,0.1220899448,0.186925441,0.0519484989,0.0592010245,-0.1077931076,0.3586030602,-0.0418989845,0.1905272752,0.0802967474,0.1038988978,-0.3220552206,0.0591778941,0.0409943685,-0.1317871213,-0.4527026713,0.1883636117,-0.31850788,-0.3556963503,-0.1726116985,0.1786646396,-0.2416288853,0.0080965115,0.2086093724,0.0757346749,-0.086124897,-0.2084441036,0.329996109,0.5944766998,-0.1610799581,-0.0999741703,-0.2164887786,-0.1343189031,-0.0417121276,0.2109253705,-0.0506248809,0.0558876321,-0.0204849038,0.1024541408,0.0101120761,-0.3500709236,-0.2602942586,-0.046818655,-0.1536080092,-0.0353873558,0.0790935531,0.1806231141,-0.107394807,-0.0908078924,-0.0359525941,0.0308668986,0.0005205862,-0.0852445215,-0.3125625253,-0.3891200125,0.0260638129,0.0143070379,-0.0739083663,0.3470021784,0.1622290313,-0.2902589142,0.2583916783,-0.1702040136,-0.0315410011,0.4124395549,0.5787875056,0.0700063407,0.0150519069,0.0433175191,-0.4493978024,0.0813691691,-0.2138365656,0.3823771477,-0.2618270516,-0.3103593588,-0.4002998769,-0.0186415166,0.1500386447,-0.4743153453,0.0841954798,-0.1355000138,0.2801226974,0.1153332219,-0.1389507949,0.299204886,0.0006324258,0.2137969732,-0.42919752,0.1492035091,-0.0950798541,-0.0213581398,0.3690761924,0.2145181447,0.118515335,-0.1379200667,0.0254460499,0.5660966039,-0.0693006068,0.4441081882,0.563893795,0.4257847667,0.347258687,-0.0107713165,0.2729820907,-0.1193465143,0.0613526553,0.0304935239,-0.5344094038,0.371244818,-0.2114973217,0.1566208303,0.1573420316,0.2463967204,0.0629866719,0.2265214771,-0.2491347343,-0.003994999,-0.1361677498,0.1006820276,0.2312166393,0.0416550674,-0.5127981305,0.2248272449,0.0237503927,-0.2264443189,-0.2222786099,0.3270247281,-0.4630303979,-0.1636605263,0.4058668911,0.3480612934,0.1047010869,0.2272708565,-0.0109121948,0.1717864126,0.2524057031,-0.1232980117,0.3218128979,0.2035265863,-0.1904883534,-0.2553881705,0.1692811251,0.2428085357,-0.2287008762,0.0455408357,-0.1582133174,0.1512369812,-0.2233811766,-0.0805480331,0.0692929626,-0.4180279374,-0.256213516,0.1239786372,-0.4829692245,-0.3185680211,-0.1371294707,0.0616204441,-0.1055617705,0.0678856298,-0.024173988,0.3342973292,-0.1543316543,-0.0404443443,-0.2764347792,-0.0694546849,-0.0403007083,0.0962014496,0.1196943521,0.1254648715,0.4261551201,-0.0977291167,0.126293689,-0.486353606,-0.5364232063,0.0431924872,-0.1406095922,0.35860762,0.1961200982,0.3100593686,-0.0980533808,0.0648044944,0.0551871285,-0.1514209062,0.0159155279,-0.3629391789,-0.1526401937,-0.0399961472,-0.0153281828,-0.2927731574,-0.160278365,-0.0898766443,0.5506942868,0.0937382802,0.0564837791,0.2059240788,0.1169883013,-0.090377219,-0.1289117485,-0.0748206973,-0.1009100527,-0.3535017967,0.318916142,-0.1554074287,-0.4775128365,0.0284705739,0.1739633381,0.1253253073,-0.2889413238,-0.3194074929,-0.3141020238,-0.4386745095,0.1385427564,0.1424367577,0.376802057,0.3760669231,-0.0236699786,-0.1428286135,-0.096005477,-0.4281664789,-0.2592428327,0.2307000607,0.1189916432,-0.1817500144,-0.2812539935,-0.0279712435,0.929833591,0.0400967859,-0.0338755585,-0.095644705,-0.1299847811,0.2708821297,0.0201997235,-0.3125040829,0.337089479,-0.1742410958,0.0979834124,0.3018555641,0.2431259453,0.3818461001,-0.0008534834,0.0643427819,-0.3905079961,-0.1964021474,-0.1641322076,0.0110532735,0.1059082597,0.1631605327,-0.1477033347,-0.2493380755,-0.0688695833,0.169011727,0.3093601763,0.181208387,0.0856421143,-0.3147589266,0.0345930159,-0.6594232321,0.3933965266,-0.1651834995,-0.1826473624,-0.0425819084,-0.075396724,0.1312235296,0.0651427582,0.6219670177,-0.2234705985,-0.2187649906,-0.0119555155,-0.00794644,-0.2581506371,0.1627407223,-0.1135927588,0.0928789377,0.4785260558,0.2214787751,-0.2654496431,-0.31110093,0.119549863,0.1183936521,0.0916562974,0.2331947982,-0.1035419703,-0.290073365,-0.1447432339,-0.2621444166,-0.0168712381,-0.036868576,-0.1000212133,0.1903064847,0.1350828856,0.0020784941,0.2289063036,0.0460974202,0.2442632467,0.1829409003,-0.004651357,0.3893855512,0.0968130454,0.128074497,0.4356867671,-0.1251814961,-0.263038069,-0.0409790687,-0.0978346244,0.2172549069,0.3517825007,0.0537206382,0.0856248662,0.4536276162,0.3400260508,-0.2403524071,0.2815037966,0.0527688153,0.0325788409,-0.264241159,-0.4618489742,0.5030096173,0.3260634243,-0.37012586,0.2631376386,0.5139727592,-0.0565770641,0.0787323341,0.2433065921,1.0592781305,-0.2437282056,0.1810193062,0.0664194971,-0.51930511,0.7468007207,-0.0798201934,0.0026949139,-0.2765857279,-0.0661222339,-0.0315877907,0.0158078335,0.3336234987,-0.1907820553,-0.221291244,0.3060911,0.086713098,-0.0921864286,0.1333425939,0.6834868789,-0.0637684762,-0.2451443374,0.0288843643,0.1543766111,0.0677054971,0.2733620405,-0.0469341911,-0.3334123194,-0.3033264577,-0.278408587,-0.4236513674,-0.1375841051,-0.3581519425,-0.3926289678,0.1499046981,-0.4280253649,0.2272787094,0.3937420547,0.4124483764,0.018031802,-0.3948340118,0.3754694164,-0.2419387102,0.0223252159,0.1005975306,0.0764286593,-0.0017417312,-0.082153976,-0.0730194896,-0.2072698325,-0.0593104921,-0.0520381704,-0.2748453021,-0.2732432485,-0.1914602667,-0.214211598,0.1176779121,-0.0190688632,-0.0299960002,-0.3094288111,0.1417072713,0.0812371746,0.1857295483,-0.0910323784,0.2000012845,0.0047739758,-0.2591331303,0.389046967,0.0951727107,-0.2510616183,0.3386106193,0.224584505,0.0897254273,-0.2786566913,-0.2519180775,-0.3440026343,-0.4056321084,0.2000599951,-0.0404256247,-0.0463341326,-0.3048221469,0.4211056828,0.1648300439,0.1100365892,0.2215452641,-0.1731854528,-0.1047953218,0.1888487786,-0.2557897866,0.1407646388,-0.1645471305,0.2194356918,-0.0177372061,0.0833306015,-0.3677552342,0.0820879489,0.0823843107,-0.2298934907,0.3981005251,0.0130685,0.2936739922,-0.3772652447,0.0988096073,-0.2523197234,-0.1576174349,-0.1331223845,-0.1655596048,0.1214986593,-0.0729313269,-0.1102866828,-0.0245817732,0.1727442443,-0.0578338243,0.0335256942,0.2165139467,0.1390505135,0.0039022728,0.140680775,-0.3608777523,0.2295107096,-0.0658484995,-0.1927149296,0.159619078,0.1456418186,0.2425497621,0.1228320673,0.0370227844,-0.041362606,0.1882186532,0.1895299554,0.0168707445,0.0166580845,0.3114473522,0.0529844426,-0.2144714892,0.2577016652,0.4860501289,0.0278367121,-0.1735053509,-0.0748665035,0.1374054998,0.2162800729,-0.1905764639,0.058216989,0.065217793,-0.0255746059,-0.0386462212,0.2975444496,0.3883385658,0.1256416887,-0.0352905653,0.1612505168,0.4712631404,0.0164629575,0.4930728078,-0.0657235309,-0.0159014501,-0.0677991286,0.2639963329,0.3475111723,-0.0513452701,-0.0434785895,0.1219754666,-0.0611155219,-0.1185129806,0.087373741,0.0106928684,-0.2702272832,0.4897191226,0.1818745881,0.1325346231,0.0878744721,-0.0756923258,0.6670544744,-0.1076184958,-0.2506597638,-0.1232699007,0.1698817313,0.2341444641,-0.1249593943,0.0393227935,-0.2366906404,0.1775383949,-0.0599157885,0.0254246648,-0.2656185329,0.2926729023,-0.1754809618,-0.1949750781,-0.5111839175,-0.1031386405,0.0241306275,0.4765481651,0.1010928228,-0.0011815842,0.096839644,-0.2140393257,0.1716049612,0.332195431,0.3361046016,-0.0068269549,-0.0473041497,0.2766436338,-0.093570061,-0.083333984,0.2441325039,-0.0137742497,0.1523180157,-0.0012727208,0.1586197168,0.1750176847,-0.0863667354,0.093616426,-0.0689672604,0.4385336339,-0.3567752838,0.2980392575,-0.3026965261,-0.3025602996,-0.0153618855,0.0154188741,-0.2900404632,-0.0211054124,-0.1927954853,-0.0551195517,0.0292357653,-0.1150280386,0.0889521688,0.1201709583,0.4455469549,0.442009449,0.1910865605,-0.1213687137,-0.2247120291,-0.3587116301,-0.2266031653,0.1407592893,-0.030728329,-0.0411870144,-0.0035501104,0.0035218655,-0.1079033017,-0.0036673783,-0.2120238841,-0.3354949057,-0.1128469706,-0.1514199227,-0.0116738956,-0.4705739617,0.1753574163,0.1398100257,0.2095280141,0.0997639745,-0.0692753494,0.0049730842,-0.1092126742,0.1780130863,-0.015484523,-0.2124408931,0.3247260749,0.2799308598,0.1539987177,-0.1100384369,-0.4077778757,-0.1591808647,-0.1423752159,-0.2861742675,0.238703832,-0.0979567766,0.4708804786,0.0885981992,-0.430031538,-0.3081097305,0.4030492902,0.2242702693,-0.1181935892,-0.3933402598,0.3088780642,0.0192105658,-0.0555651635,0.0542383566,0.1077390388,-0.0422649682,0.048699569,-0.3787397146,-0.219064787,0.6983258724,-0.2918163538,0.0275560003,0.0461210087,0.3747472167,-0.0657112151,-0.1171064377,-0.6519762278,0.2062691003,0.1633516848,0.130972892,-0.2118215859,0.2152027041,-0.1069553643,-0.2190624923,-0.0384439714,0.6719532013,-0.1442336887,-0.252387315,0.0540390126,-0.1074284613]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1717","title":"SciFact dataset - minor changes","comments":"> I'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nSure ! Also feel free to ping us for reviews or if we can help :)\r\n\r\n> It also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nWhat makes you think that ?\r\nAfaik there's no scifact on our google storage\r\n","body":"Hi,\r\n\r\nSciFact dataset creator here. First of all, thanks for adding the dataset to Huggingface, much appreciated!\r\n\r\nI'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nIt also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nThanks,\r\n\r\nDave","comment_length":91,"text":"SciFact dataset - minor changes\nHi,\r\n\r\nSciFact dataset creator here. First of all, thanks for adding the dataset to Huggingface, much appreciated!\r\n\r\nI'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nIt also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nThanks,\r\n\r\nDave\n> I'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nSure ! Also feel free to ping us for reviews or if we can help :)\r\n\r\n> It also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nWhat makes you think that ?\r\nAfaik there's no scifact on our google storage\r\n","embeddings":[0.1750965118,-0.1627209932,-0.1382708549,0.1695231795,0.171601519,-0.1054838747,-0.0107192928,0.0339706019,0.2187184989,0.0415754393,-0.1746508926,0.0221550968,-0.0037148951,0.5154721737,0.166182369,-0.2211896926,0.2049407661,-0.1498279423,-0.1490776688,-0.3085084558,-0.0934292749,0.2177199125,0.1105605587,-0.1985752881,-0.053899508,-0.1465506405,-0.2876536846,0.3299089968,-0.3410831094,-0.2185066938,0.0059862831,0.2872893214,-0.0213820469,0.315020144,-0.000102045,-0.108007133,0.3075104058,-0.070631668,-0.2162955105,-0.0264658704,-0.1062400192,-0.0436003692,0.0415757969,0.0017974518,-0.1029068604,0.0601079315,-0.0751407295,-0.2309908122,0.3554613888,-0.0319922231,0.3132331073,0.1185379252,0.0779164582,-0.1776697338,0.0726440698,0.2887015343,-0.133516103,0.2008469105,0.1898788363,0.3068340719,-0.1447353363,0.6204882264,-0.0216451697,-0.1450505853,0.5957466364,-0.0318081863,-0.3159663975,-0.4047898054,0.1408154964,0.3081873655,0.4457279444,-0.2545996606,-0.6352859139,-0.227434054,0.1143576726,-0.1083268151,-0.0184525438,0.2890632153,0.0686166883,0.0535121337,0.0243590213,-0.4525232911,-0.3507419825,-0.1081755832,-0.047113698,0.108327046,-0.1657474786,-0.0440968312,0.091824986,-0.0021018211,-0.0841014311,-0.182271868,-0.0529444069,0.094959788,-0.1866751313,-0.2585341632,-0.1515841186,0.2619608939,0.0508794934,0.362362653,0.1547906846,0.0821071193,-0.1197000667,-0.1157878116,0.2962263525,-0.0366734713,0.0230399873,0.0158712436,0.2667606771,0.060267631,0.3479256332,-0.0917803794,0.1576174349,-0.130873695,-0.5323855877,-0.1689876467,0.0656358749,-0.170612976,-0.0131803751,-0.0349683277,0.1527695209,-0.1324877739,-0.1000986397,0.2938219309,-0.0978160575,-0.029558083,-0.0576104075,0.2187986672,0.0609581918,-0.2394988388,-0.3341899514,0.202619046,0.1900565177,0.0827171877,0.1475896984,0.0390844308,0.2758976519,-0.0857573748,0.0215234868,0.2337734401,0.1375716627,-0.0556045473,-0.021000633,0.3844462335,-0.0098000346,0.1058368757,0.0518009514,0.1202451736,-0.2956322134,-0.0409873463,0.0416215993,-0.110791035,-0.4523085058,0.2638582289,-0.3114067316,-0.3310674727,-0.1912941188,0.1648497432,-0.1809655577,0.0624513067,0.1114881709,0.1220203713,-0.1212252602,-0.2155525088,0.2229269296,0.418141067,-0.1954847425,-0.1882674992,-0.315848738,-0.151160419,0.0142155308,0.2107599825,0.0462556072,0.0258883312,-0.1377400458,0.1400977671,-0.0074031297,-0.3021460176,-0.2463244349,0.0408311337,-0.1063932553,0.0342386626,0.110600546,0.1690256596,0.0097783562,-0.0350154676,-0.0310879257,0.1615830511,0.0088427402,-0.0714623332,-0.2766363025,-0.490786314,0.0404329002,0.0260183979,0.063353844,0.1510100067,0.182021156,-0.2856173217,0.3023248017,-0.1926104575,-0.018312104,0.4099454582,0.5346345305,0.0429560654,0.0295976475,0.0317672789,-0.5193303823,0.147583589,-0.218706429,0.4071362615,-0.2894701958,-0.2436176836,-0.2690857053,-0.0541456528,0.0588006489,-0.4067020416,0.2099795192,-0.1151897609,0.2097765058,0.0788095817,-0.0493905619,0.2961065471,0.0036831291,0.2046887577,-0.4347661138,0.1959660649,-0.135314554,0.0174518488,0.2929261923,0.1585515887,0.0795653015,-0.1501151025,0.0340366326,0.501691401,-0.0295417011,0.4781923294,0.502761662,0.3911731243,0.4086720645,-0.0107059507,0.2504283488,-0.0324828364,0.0246457253,-0.0572975911,-0.4910120368,0.2947308719,-0.1037092805,0.1660445929,0.1834676117,0.2979111373,0.0981002077,0.2008259445,-0.2598776221,-0.156166926,-0.0991061777,0.0865385458,0.208753854,0.0503789261,-0.4136379361,0.2915058732,0.077395916,-0.2357574701,-0.0906447694,0.2380052805,-0.4635380507,-0.2474510372,0.4250988364,0.3064067662,0.0493452474,0.3404064775,0.0635357201,0.0494972691,0.1891489476,-0.1588693857,0.4017777145,0.1743432432,-0.0399962589,-0.1824591309,0.1818552315,0.0983664542,-0.3777604997,-0.045032084,-0.2103866786,0.0755845532,-0.2193056345,-0.1137075126,0.0073696109,-0.3583326936,-0.1370738447,0.0920670331,-0.4101584256,-0.4434770644,-0.0507215597,0.0759968832,-0.248207137,0.0382062197,-0.1408405602,0.3447827101,-0.0321776532,0.0043004225,-0.3078268468,-0.0855194852,0.0033639602,0.1974467486,0.1433514804,0.1119345725,0.4356259704,-0.0864042863,0.1355852783,-0.4129348695,-0.6301935911,0.1747861952,-0.2031736374,0.3550866246,0.2258835435,0.2690949738,-0.0462635644,0.0270503275,-0.0165732019,-0.1791456193,-0.0409112498,-0.3050620556,-0.1499483287,0.0350809097,-0.0846811458,-0.2805574238,-0.1690338403,-0.0637585372,0.4855988324,0.095969528,0.148746863,0.2344144583,0.1089552194,-0.0266425125,-0.1086443737,0.011763663,-0.2309374064,-0.3808627725,0.2564818263,-0.1707243174,-0.4585993886,-0.0391090438,0.0907628611,0.1902610511,-0.2478321195,-0.329780221,-0.2076347321,-0.4169572294,0.0906143486,0.1164740473,0.273201406,0.3255670369,0.0061794859,-0.1889618039,-0.135603711,-0.2109445333,-0.142507568,0.1865995675,0.0911460444,-0.2945891321,-0.3575014472,0.0876167491,0.822694838,-0.160870105,0.0603462793,-0.1181040779,-0.0738643333,0.2442931831,-0.0199389625,-0.2384810448,0.4008345306,-0.218167901,0.0416644141,0.2813808322,0.2424054295,0.234934181,0.0366948955,0.1570050269,-0.409606576,-0.1984178871,-0.2051500976,0.0737980008,-0.0160843395,0.1428353041,-0.0753043368,-0.102574937,-0.0952806547,0.1816660166,0.1702075005,0.1958914995,0.0054958435,-0.298899889,0.0274680201,-0.6712693572,0.3566029966,-0.182219848,-0.3079404831,-0.0450305939,-0.199129194,0.0733796433,0.0478542708,0.5412956476,-0.1563112885,-0.1667944342,0.0207336787,0.0084109595,-0.3620074987,0.1612844169,-0.2475473881,-0.0220260601,0.5003551245,0.2489691377,-0.1638910621,-0.2980360091,0.126844421,0.1487523615,0.040381182,0.2064268589,-0.0352390483,-0.293926537,-0.1500555128,-0.2607956827,-0.0450857617,-0.0082864612,-0.1854827851,0.118946135,0.118496418,-0.0191089492,0.2695919573,0.0130161438,0.2740361691,0.0582065694,0.0147293136,0.3183832467,0.2254603356,0.1026195586,0.4502186179,-0.0342097469,-0.1215227172,0.0615303516,-0.1934173107,0.151912868,0.4016211927,0.0532924719,-0.0343920775,0.2892886996,0.2728689611,-0.3131509721,0.2336546034,0.0171414968,0.0099437488,-0.1338871717,-0.4116716683,0.5648608804,0.2755479515,-0.3339821994,0.2637113035,0.4595813751,-0.166881904,0.2503581047,0.3432019949,1.0519461632,-0.182316348,0.1915661693,0.2723131776,-0.3975040317,0.5669394732,-0.1354913861,0.1201680079,-0.2444373518,-0.111329712,-0.0198984835,0.0294673741,0.2397435158,-0.0675934628,-0.1836671233,0.2452342212,0.0635967255,-0.0749720111,0.0485109016,0.5296371579,-0.1810363084,-0.2435669601,0.1027202681,0.2484957725,0.1063237265,0.24014467,-0.0482961796,-0.2501688898,-0.2191303223,-0.2106993347,-0.3076816499,-0.0465486348,-0.3181011081,-0.1696781665,0.0081070252,-0.3784960508,0.0651180595,0.2864228189,0.3279699385,0.0127872834,-0.3710659146,0.2980848849,-0.1320471019,-0.0058049089,0.1807524413,0.1145462319,0.1004910171,-0.0866208822,-0.076833576,-0.2526573837,-0.0226571336,-0.0901272297,-0.2980693877,-0.1858934611,-0.2280789018,-0.1989988387,0.0638995096,0.0138010057,-0.0903901383,-0.2950102389,0.2060499936,0.0922402367,0.2022002339,0.0057085077,0.1439734548,-0.1062403247,-0.3203229904,0.4124445319,0.2102407813,-0.2354560941,0.3347992599,0.2515662909,0.0121593252,-0.382006377,-0.1919419467,-0.2965377569,-0.3045251071,0.2305505425,-0.1346245706,-0.137967512,-0.3149487078,0.4261558056,0.123160854,0.1007918268,0.1444395483,-0.1644696742,-0.0644600391,0.3394781649,-0.2732061148,0.2789137661,-0.1416908652,0.1691854447,-0.0875366703,0.0055957823,-0.4733804464,0.0885532275,0.1170244068,-0.2836105227,0.3447381854,-0.0236670114,0.2896568179,-0.3583015501,0.2171766013,-0.2909926176,-0.1677574962,-0.2391606867,-0.0864441022,0.0779039785,-0.05982811,-0.0788131431,0.0405377634,0.184905991,-0.0928234309,0.0087925373,0.3291770816,0.2893912196,0.0751079619,0.2261412293,-0.4084892869,0.1449999958,0.1052685753,-0.1523462683,0.1549372524,0.1519583166,0.1298219711,0.1377008408,0.0583403856,-0.0605939403,0.2330098152,0.1923187524,-0.1366099417,0.1516266614,0.3314452767,0.039477855,-0.1939131469,0.3143909276,0.4291400015,-0.0537076481,-0.1632532328,-0.1534480453,-0.0181990955,0.3565852642,-0.2972002625,-0.0816849247,-0.1046852246,0.0187936388,0.0593879223,0.2791618705,0.2986148894,0.0432249866,-0.0128565561,0.2199676931,0.4392496943,0.0287508667,0.4842642844,-0.0041235406,-0.0900336429,-0.1107513979,0.2043793201,0.2356896698,0.05027375,0.0556201935,0.0724823847,-0.0543045886,-0.1226926371,0.1164688542,0.1226573288,-0.2403240949,0.3531138003,0.2654886246,-0.0431747921,0.1660890132,-0.0641694069,0.6414595246,-0.1258918643,-0.2579583228,-0.1385197341,0.1955437213,0.1499342471,-0.1286888123,-0.0021005333,-0.2624365985,0.1095241681,-0.0681486726,-0.007744167,-0.2621625364,0.2859225571,-0.1656338125,-0.1700389981,-0.3744331598,0.0053643677,0.0191868339,0.4165748954,0.0813730434,0.0558629669,0.2220390886,-0.145696938,0.1433161646,0.3310251832,0.2914970219,-0.020554224,-0.1091671363,0.1421899647,-0.10040503,-0.152254954,0.0915052667,-0.0382965505,0.1382108927,-0.0918598026,0.3246209621,0.2497988045,-0.1418579072,0.1387271136,0.0831676126,0.4113074839,-0.220800221,0.3107289672,-0.2487463057,-0.2101234347,0.0166688152,-0.0460826717,-0.4195713401,-0.1452141553,-0.1362058669,-0.0967141837,0.1004754156,-0.1105058715,0.1494554132,0.0561530888,0.4306778312,0.4445018768,0.0664393976,-0.0852897614,-0.1931337565,-0.2694976032,-0.2304938734,0.0691065714,0.0502773151,-0.1349869519,0.2231627107,0.0894186273,-0.0683903769,0.0751369223,-0.1929585189,-0.3643417954,-0.1695173681,-0.2320842743,0.0505088791,-0.300121069,0.1985495389,0.0823197961,0.054489661,0.1448315382,-0.1597454548,0.0827784836,-0.0586887561,0.1942728311,-0.0465263315,-0.2598098218,0.3093500733,0.1788987815,0.0617698729,-0.2087698579,-0.4297063649,-0.1614361703,-0.2023562789,-0.2559978068,0.2328845561,-0.0271438845,0.5041917562,0.086025767,-0.3409916461,-0.14156726,0.4344553947,0.2316703796,-0.1149545237,-0.477539748,0.3527237773,-0.01117275,-0.040336635,-0.0579485148,0.2095296681,-0.0992563739,0.0490898862,-0.2914091647,-0.379943192,0.6491066217,-0.2303994,-0.0401021801,-0.0236902256,0.289357841,0.032320261,0.0021480708,-0.6299197078,0.2145919204,0.1657151282,0.1649064273,-0.1398124695,0.1787533611,0.0340929516,-0.1925086379,-0.1009703055,0.513053596,-0.0183232054,-0.2909580171,-0.0379199833,-0.1867756993]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1717","title":"SciFact dataset - minor changes","comments":"\r\n\r\n> > I'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n> \r\n> Sure ! Also feel free to ping us for reviews or if we can help :)\r\n> \r\nOK! We're organizing a [shared task](https:\/\/sdproc.org\/2021\/sharedtasks.html#sciver) based on the dataset, and I made some updates and changed the download URL - so the current code points to a dead URL. I'll update appropriately once the task is finalized and make a PR.\r\n\r\n> > It also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n> \r\n> What makes you think that ?\r\n> Afaik there's no scifact on our google storage\r\n\r\nYou're right, I had the data cached on my machine somewhere. \r\n\r\n","body":"Hi,\r\n\r\nSciFact dataset creator here. First of all, thanks for adding the dataset to Huggingface, much appreciated!\r\n\r\nI'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nIt also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nThanks,\r\n\r\nDave","comment_length":152,"text":"SciFact dataset - minor changes\nHi,\r\n\r\nSciFact dataset creator here. First of all, thanks for adding the dataset to Huggingface, much appreciated!\r\n\r\nI'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nIt also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nThanks,\r\n\r\nDave\n\r\n\r\n> > I'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n> \r\n> Sure ! Also feel free to ping us for reviews or if we can help :)\r\n> \r\nOK! We're organizing a [shared task](https:\/\/sdproc.org\/2021\/sharedtasks.html#sciver) based on the dataset, and I made some updates and changed the download URL - so the current code points to a dead URL. I'll update appropriately once the task is finalized and make a PR.\r\n\r\n> > It also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n> \r\n> What makes you think that ?\r\n> Afaik there's no scifact on our google storage\r\n\r\nYou're right, I had the data cached on my machine somewhere. \r\n\r\n","embeddings":[0.1659245044,-0.1632084399,-0.1014101356,0.2528709471,0.1667679101,-0.1169615015,-0.0117663369,0.065581575,0.25607723,0.0539326929,-0.158615768,0.0316010676,0.0221038926,0.4208246469,0.1964264661,-0.1602756232,0.1690869331,-0.1912660003,-0.1692505032,-0.264374882,-0.0958833322,0.2250898182,0.1093872935,-0.2362832278,-0.0988471434,-0.1291609406,-0.339111805,0.3356045187,-0.244578734,-0.2094162554,0.0959263667,0.2644577324,0.0055237049,0.3104948103,-0.0001009035,-0.1070521474,0.3121140897,-0.0670214966,-0.2641534805,0.0139368484,-0.1189585775,-0.0307925176,0.0353437439,-0.0341879725,-0.1467750221,0.0920753106,-0.0842793286,-0.3005348444,0.3577445447,-0.0076566124,0.3215502203,0.1390474141,0.0338772982,-0.1082980558,0.024564825,0.2060159594,-0.0999378189,0.2146789432,0.1796554327,0.3415538669,-0.1511780471,0.6248745918,-0.0134906257,-0.0615412742,0.6073548794,-0.002498172,-0.3729184568,-0.3723898828,0.1324513108,0.2662905455,0.4439245164,-0.2553659678,-0.5880433321,-0.2767368853,0.0677962527,-0.1180537641,0.0802191198,0.2208799869,0.1082304344,0.110559918,-0.002932146,-0.3719548583,-0.2812481225,-0.0933684111,-0.0572902411,0.0883339271,-0.2347088307,-0.00691836,-0.0219213124,0.0505925417,-0.0879138708,-0.1871252507,-0.0515399575,0.0686974823,-0.217071563,-0.2231781334,-0.1066396981,0.2603141367,0.0473404825,0.3677626848,0.2138269097,0.042710539,-0.1064621136,-0.0991134122,0.3071835637,0.0602344982,-0.0334405936,0.0095238779,0.2098219842,0.0712004378,0.2411014885,-0.1147147343,0.1489451379,-0.0809174329,-0.4109897316,-0.1142466441,0.127658844,-0.2160331458,-0.0206933301,-0.030133903,0.1604787707,-0.1487119645,-0.1054576263,0.2967299223,-0.1989014149,0.021017991,-0.0608125776,0.1347529143,0.0275204405,-0.2565941811,-0.3591398001,0.136519134,0.1830496192,0.1658910364,0.1521867216,0.0264011286,0.266823709,-0.1001813486,-0.058599297,0.0916243792,0.1977423429,-0.1109998897,0.0347128324,0.4107367992,0.0203840286,0.1604322344,0.1007263362,0.1393442154,-0.3201084435,0.0282548442,-0.0381403118,-0.1832457185,-0.3506273627,0.2678078711,-0.3660208285,-0.3243698478,-0.2245863825,0.1378593892,-0.0728247836,0.0547398925,0.1099484488,0.0779945999,-0.1805955619,-0.2324555814,0.2896608114,0.4897944629,-0.1547354311,-0.197051689,-0.3170634508,-0.130992502,-0.0137209417,0.2298933417,0.0137206456,0.087102063,-0.1730643213,0.1507139355,-0.0737685487,-0.3053477705,-0.3150310516,0.0309275649,-0.1864068657,0.0680138022,0.0683805346,0.164967671,0.0605829395,-0.1109465957,0.0058717839,0.2109098434,0.0223009847,-0.029438667,-0.3631426692,-0.5218738914,0.0028266865,0.0355954692,0.0160767641,0.1326296628,0.1774075329,-0.2517201006,0.2485509813,-0.1364070326,0.0255250335,0.4775681496,0.4186940491,0.0962246954,0.0142955771,0.0225130152,-0.5479564071,0.1748357117,-0.1704976857,0.2987992167,-0.2126060724,-0.2755693793,-0.2794927061,-0.0767840818,0.0091918409,-0.4557008743,0.2026423365,-0.0396115072,0.2555353642,0.0474681519,-0.0561455935,0.4210202992,-0.0011723654,0.1512536407,-0.4600059688,0.2111089528,-0.126643315,-0.0034927027,0.2701086402,0.1262537092,0.1248256117,-0.1292121261,0.0162339862,0.5227890015,-0.045363795,0.494094342,0.4585257769,0.3518596292,0.3367432356,-0.0483895428,0.3075618148,0.0482788421,0.0640711039,-0.0326159708,-0.4407789409,0.313038677,-0.1468823403,0.159747377,0.1760475785,0.2420877218,0.0994074196,0.1470552385,-0.2510013282,-0.151469782,-0.0545384549,0.0868230388,0.1815574765,0.1376991272,-0.4012317955,0.3901750445,0.1800296158,-0.1799881905,-0.0727726445,0.2597621083,-0.4883365035,-0.2888489366,0.3766419888,0.3128044903,0.1028181091,0.307931006,0.096668452,0.0593261868,0.1894824356,-0.152934283,0.3611404598,0.125602439,0.0013814741,-0.2054732144,0.1876898557,0.1487430036,-0.2898752391,-0.001274313,-0.1462234259,0.0650820658,-0.282189399,-0.0283022709,-0.0673037767,-0.332326591,-0.2045745254,0.1434539258,-0.4108769,-0.4312398434,-0.0501125082,0.0695999563,-0.2341261655,0.0100773368,-0.1702439785,0.3444338143,-0.0842498988,-0.0888793617,-0.3108499348,-0.1166760102,-0.0730242804,0.1652176976,0.1239604801,0.1223328114,0.4448599517,-0.1131160408,0.1181747317,-0.3981671631,-0.5399107337,0.1104154885,-0.1731895208,0.3696745038,0.0853897631,0.2747528851,-0.0652669221,-0.0163569152,0.0184384398,-0.2034553289,-0.0422918536,-0.314124763,-0.0759491026,0.0965252444,-0.0852784067,-0.3394380808,-0.1758095771,-0.0871695429,0.4697408974,0.112992011,0.152296558,0.2350764275,0.0947654322,0.0154346488,-0.1238915026,-0.0211453214,-0.3138516247,-0.4614569843,0.2333492786,-0.1357364506,-0.4435072541,-0.0291079804,0.0915530324,0.1781039983,-0.2261597663,-0.3716216385,-0.2547497749,-0.4176287651,0.0993888825,0.1467260122,0.1682907045,0.4198363721,-0.0480393395,-0.2037337869,-0.13063173,-0.2296764404,-0.0689652786,0.1544959992,0.1245271116,-0.2486196458,-0.32275334,0.1143443286,0.9150562882,-0.1271840185,0.0333031453,-0.0384288728,-0.0034311935,0.2881451249,-0.0543049984,-0.1876615286,0.4380381405,-0.2377439588,-0.0139574902,0.2795522511,0.270663023,0.2290080041,0.0003981642,0.0620314255,-0.455226481,-0.2427370846,-0.1421176791,0.0553785414,0.0102332272,0.1922933459,-0.0975208357,-0.1200738996,-0.1699835509,0.1681482345,0.2204692066,0.2204895318,0.0611265041,-0.2955599129,0.0381026417,-0.643066287,0.3647662699,-0.181944415,-0.2396472096,-0.1097279415,-0.2054857463,0.1155089512,-0.0303194001,0.5701812506,-0.1514645964,-0.073446624,0.0137191713,-0.0291011464,-0.3352385759,0.1464283913,-0.1687944382,-0.0581755601,0.4786731601,0.1987055391,-0.1654240638,-0.2544857264,0.0897849351,0.1746033281,-0.0013532102,0.2343285084,-0.0881912783,-0.3738437593,-0.172971189,-0.2486247867,-0.0005544615,-0.0744138733,-0.1833431721,0.1432867497,0.0403203592,0.0168457907,0.1693707705,-0.0254959203,0.2929743826,0.0588333979,0.0862281397,0.3918727636,0.2154993415,0.0629100278,0.4492222071,-0.0386003628,-0.0804870874,0.0785091817,-0.1523718685,0.0721799582,0.4211024344,0.0437518135,-0.038887091,0.3254598677,0.2381462455,-0.3317847252,0.266802907,-0.0310908966,0.0507719927,-0.1206861585,-0.4809465408,0.5317932963,0.3323586881,-0.3395459652,0.3411290348,0.3344012499,-0.1751316488,0.1791659445,0.330755502,1.1044721603,-0.2396254539,0.2183927596,0.2713886201,-0.3675172627,0.5915099978,-0.1588971466,0.1215533987,-0.2350538224,-0.1532121152,-0.0178208854,0.0288633369,0.2204356939,-0.1250164509,-0.1721253395,0.2953158617,0.0236440729,-0.0586719103,0.0611169748,0.5449422598,-0.2023491412,-0.2332599461,0.0303317513,0.2792079151,0.0246794838,0.3005772531,-0.048964709,-0.188594982,-0.2500043213,-0.1683728397,-0.2956948578,0.0222447086,-0.2925338745,-0.1654115915,-0.0340940692,-0.3455653489,-0.0326834954,0.2827169597,0.3265953362,0.0111176614,-0.3677639067,0.3616234958,-0.1643996984,0.0491185784,0.238903299,0.0498263128,0.2165528834,-0.0773442611,-0.058522258,-0.2473369092,-0.0336185917,-0.1177684665,-0.2782156467,-0.1453955323,-0.1803478003,-0.1917414069,0.0384255983,0.016758414,-0.0860861689,-0.2852520645,0.2113433033,0.1021882445,0.2083405405,0.0009636696,0.087617062,-0.1539215893,-0.3013564944,0.4673447609,0.1977534294,-0.1946942061,0.3502906263,0.2066466957,0.0243975054,-0.3506257534,-0.1857432723,-0.3186751008,-0.3479624987,0.2132835537,-0.1885840744,-0.0866947323,-0.2987436056,0.448269248,0.1046101674,0.075892061,0.1190066561,-0.2171260118,-0.0584230497,0.2973365188,-0.2570818067,0.2073879838,-0.0703190342,0.1772554815,-0.0477305874,0.0109267104,-0.4741080701,0.1144640669,0.0171371214,-0.2385254055,0.3569376171,-0.0719135702,0.3654998541,-0.3403912485,0.1926848888,-0.2375713438,-0.1610366553,-0.2635545433,-0.1117866114,0.0725864843,-0.0622511581,-0.0850168318,-0.0649640337,0.0751676112,-0.0295933932,-0.0352814645,0.2782481015,0.2639962733,0.0875070766,0.1205091998,-0.434566617,0.1550924778,0.0651685297,-0.0855757669,0.0699234605,0.1493201852,0.0538699217,0.1294628084,0.0240773037,-0.0103441784,0.1673239619,0.121666804,-0.0864703581,0.1596730947,0.3795135617,0.0134694669,-0.2088517547,0.2468449473,0.4186221957,-0.0502963252,-0.1933386326,-0.2112444043,-0.0423813276,0.3510423601,-0.1897952259,0.0118211815,-0.0755440742,0.0166237429,0.0203553233,0.2811971605,0.2679562271,0.0852024108,0.0483864062,0.2590622306,0.5885541439,0.0260475464,0.4186683595,-0.0134794144,-0.0387239642,-0.0310651157,0.2673403621,0.280025214,0.0678585693,0.0605228841,0.124110207,-0.0415211283,-0.1150122732,0.150773257,0.0175779331,-0.3053456545,0.3549687266,0.234958142,-0.0947242677,0.1808713078,-0.0853522792,0.6152272224,-0.1651923805,-0.2626313865,-0.1313143522,0.1952122301,0.1257756948,-0.1131003127,0.0571509898,-0.2251837105,0.1447263956,-0.009965959,0.0000540009,-0.3066886067,0.3339623511,-0.119556576,-0.2409220934,-0.4610393047,0.0194275603,0.0507977121,0.4381589592,0.033906769,0.1344057918,0.1502237469,-0.1041343659,0.1403133571,0.303306967,0.3024393916,-0.0267978683,-0.1026207581,0.1678671837,-0.004272473,-0.1526672244,0.0363046601,-0.0225429013,0.1587714404,-0.1440173388,0.3165237904,0.2696392238,-0.1581303328,0.1510713995,0.0617024601,0.4713743925,-0.2484210283,0.324806273,-0.286881566,-0.1645697355,-0.0274211746,-0.0088625494,-0.4613204002,-0.1099575534,-0.1738194823,-0.1194102913,0.1611708403,-0.0961388275,0.1478843689,0.058398068,0.4296531975,0.4756164253,0.0601961501,-0.1230809912,-0.1693604589,-0.4051965475,-0.2141819,0.0667780787,-0.0001475057,-0.1294726729,0.2560307086,0.0394413248,-0.0751807913,0.0125174904,-0.129863292,-0.2912452221,-0.171221301,-0.2054769695,0.0174035653,-0.3714542389,0.2298817784,0.0886320248,-0.0377576798,0.1739328206,-0.1348356009,0.0882593989,-0.0656104311,0.0863380432,0.004493875,-0.2129162401,0.3346933424,0.1569328606,0.0603172928,-0.1478258073,-0.3977351189,-0.2031396329,-0.1829536259,-0.2368999422,0.2022664398,-0.0810160786,0.4406937361,0.0277289581,-0.2790139318,-0.2162279487,0.4099099338,0.2423655093,-0.049670659,-0.4071410298,0.3173881471,-0.0522003844,-0.0149546769,0.0446260944,0.2617966533,-0.0912173688,0.0775285363,-0.3626358807,-0.2951004207,0.5989770293,-0.2397185415,-0.0193430576,-0.0318006016,0.3708543777,0.0363018624,-0.0763150454,-0.593342185,0.2024229914,0.2399945557,0.1632973403,-0.1635065079,0.1674386859,-0.0080948872,-0.1446020454,-0.0883523002,0.4926125407,-0.0399095304,-0.2402552068,0.0053293696,-0.13683559]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1717","title":"SciFact dataset - minor changes","comments":"I opened a PR about this: https:\/\/github.com\/huggingface\/datasets\/pull\/1780. Closing this issue, will continue there.","body":"Hi,\r\n\r\nSciFact dataset creator here. First of all, thanks for adding the dataset to Huggingface, much appreciated!\r\n\r\nI'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nIt also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nThanks,\r\n\r\nDave","comment_length":13,"text":"SciFact dataset - minor changes\nHi,\r\n\r\nSciFact dataset creator here. First of all, thanks for adding the dataset to Huggingface, much appreciated!\r\n\r\nI'd like to make a few minor changes, including the citation information and the `_URL` from which to download the dataset. Can I submit a PR for this?\r\n\r\nIt also looks like the dataset is being downloaded directly from Huggingface's Google cloud account rather than via the `_URL` in [scifact.py](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/scifact\/scifact.py). Can you help me update the version on gcloud?\r\n\r\nThanks,\r\n\r\nDave\nI opened a PR about this: https:\/\/github.com\/huggingface\/datasets\/pull\/1780. Closing this issue, will continue there.","embeddings":[0.2070191652,-0.1952877641,-0.0556955114,0.1916944385,0.1510135233,-0.0969563797,-0.1887090057,0.0132451551,0.1114652008,-0.1106241271,-0.0804411694,0.0390700772,-0.0309824105,0.4415709078,0.1327659637,-0.2253412455,0.2328605056,-0.1081742421,-0.0802788734,-0.3198138177,-0.0585840866,0.2774597406,0.084239468,-0.1581545025,-0.1156414822,-0.2201213986,-0.2490102053,0.282435298,-0.3966191113,-0.3390519917,0.0749481544,0.4013519883,-0.0054334258,0.3338243365,-0.0001107199,-0.1531109363,0.3292019367,-0.0188297499,-0.1685743034,-0.1079659238,-0.1067180783,-0.1256477088,0.0606663525,0.0402373634,-0.0752016827,-0.0516205356,-0.1372546852,-0.0850492418,0.3527564108,0.0786363333,0.2283425331,0.1906204373,0.1001526788,-0.135064736,-0.0405513532,0.3495544195,-0.0457147807,0.2671400607,0.1023499817,0.315893054,0.000382663,0.4947763085,-0.0672822222,-0.1492548585,0.6343651414,0.0145163042,-0.4003590345,-0.4666485488,0.1596461385,0.3020663261,0.4869964719,-0.3107042313,-0.6470098495,-0.2400811613,0.1919306815,-0.0511184409,-0.0298365578,0.302153796,0.036962647,0.0273130741,0.0690210462,-0.4438324869,-0.316488862,-0.1295174658,0.0195685979,0.0876298845,-0.1777819246,0.0094542261,0.0648106486,-0.0586316176,-0.0604658686,-0.2352398336,-0.0457774661,0.1913601309,-0.1431263834,-0.2342828959,-0.2097836435,0.2687039673,0.0557322241,0.3615639508,0.0375724137,-0.0109213339,-0.225568831,-0.0876842886,0.3284628391,-0.0704780668,0.0510836393,0.1353430301,0.3051563799,0.202102527,0.3839263022,-0.1280380338,0.1617750823,-0.1142572016,-0.5027332902,-0.1850789934,0.2186969072,-0.2425589412,0.0676189736,0.0058245063,0.1678392738,-0.2062970251,-0.1262403131,0.2235925943,-0.1493153721,0.0065665091,-0.0527852848,0.1825798601,0.0872851834,-0.2132056952,-0.2776604891,0.1402275413,0.2569067478,0.1490970403,0.0959714502,-0.023112908,0.350933075,-0.0504302979,0.0705474913,0.1772742122,0.0308573395,0.0587919764,-0.0954022631,0.3496402204,-0.0652345791,0.2427414805,0.1167885661,0.1509667337,-0.341209501,0.0152328685,-0.021364022,-0.1405693442,-0.3903029859,0.1898880303,-0.3791836202,-0.3769913316,-0.1463137418,0.1321174353,-0.2062134743,0.0369031876,0.1935020983,0.0655302331,-0.1391403079,-0.2095229179,0.3284426928,0.6121068597,-0.1302760094,-0.1741629988,-0.1957480907,-0.1379784048,0.0109711764,0.2549556196,-0.0411042832,0.0785599425,-0.0280161593,0.1287118644,0.0231467132,-0.297421962,-0.2788325548,0.0035027915,-0.1244615316,0.0135792857,0.0550977997,0.1509934068,-0.0967097655,-0.0758175775,-0.0095526027,0.0306180231,0.0524740592,-0.0716395751,-0.3122313917,-0.4258087873,0.0451739505,-0.0043241079,-0.0266949553,0.3387380838,0.117719762,-0.2871601582,0.3122203648,-0.1075337604,-0.02100127,0.4440083504,0.5577414632,0.1210379377,0.036575038,0.0288606752,-0.3876316845,0.0880558938,-0.18855609,0.3668851554,-0.2676448524,-0.2944400311,-0.4066998065,-0.0074863764,0.1258080155,-0.489100188,0.0920979008,-0.1875522584,0.2675869465,0.1676946878,-0.1234956011,0.3412328959,-0.0174870435,0.2197583467,-0.463185668,0.1848306954,-0.0333649442,-0.0172710214,0.3356550932,0.2273078412,0.1679156721,-0.1664358228,-0.0025723672,0.5319182873,-0.0290265512,0.4115120471,0.5352915525,0.3856110871,0.3490709066,0.0243665315,0.2629445195,-0.0507102273,0.0569168925,0.0331427939,-0.488432318,0.3442982137,-0.1941537857,0.1856945008,0.1313343793,0.2844697833,0.0896188393,0.2511502802,-0.2634794116,-0.0075975293,-0.0744415298,0.046916239,0.2119475156,0.0702517033,-0.5023847818,0.1928487271,0.070814386,-0.2133611888,-0.2556990385,0.3749161661,-0.5015886426,-0.1707426608,0.4135684967,0.3836425245,0.1131189018,0.2219924033,-0.0338508077,0.1779744178,0.2460436672,-0.1032361686,0.3417837024,0.2227689028,-0.1900671124,-0.2582217157,0.1246231571,0.2516068816,-0.2767295539,0.0307028573,-0.2205592394,0.1547273248,-0.2305187434,-0.078153111,0.021648936,-0.4352829754,-0.2883395851,0.1370334476,-0.4874366224,-0.3557218611,-0.1290208697,0.0273596644,-0.0999385417,0.0775208324,-0.089522399,0.3354848623,-0.1197094545,-0.034006834,-0.2902546823,-0.0490799285,-0.0595328435,0.0901371092,0.0752273351,0.1126498803,0.4013534784,-0.1377457082,0.1226167083,-0.4738873243,-0.5415538549,0.0559395887,-0.155504182,0.3217745125,0.2456494123,0.3079081774,-0.0582820699,0.0602172092,0.0365762599,-0.2366561592,-0.0032715315,-0.2931038737,-0.1645913273,-0.0173591636,0.0064813443,-0.2559606135,-0.1434765756,-0.0640920997,0.546205163,0.0941027403,0.0281814802,0.2077195197,0.048434563,-0.0829972923,-0.1918280572,-0.0673200861,-0.1505270302,-0.3879559934,0.3204443455,-0.1355008632,-0.4881453514,0.0360278785,0.1514080167,0.1265871972,-0.3022559583,-0.3435955346,-0.2593034804,-0.4519959688,0.1102210507,0.1121056676,0.3877241611,0.3676057458,-0.046609439,-0.1190801561,-0.1043439284,-0.3825244308,-0.2500043213,0.223347187,0.110671863,-0.2130008787,-0.2568894923,0.0168817379,0.9401144981,0.0877413154,-0.0151052326,-0.0617807619,-0.1486410052,0.2814576626,0.0382941477,-0.3285822272,0.3123357594,-0.1561918706,0.1079776585,0.3016971946,0.2578416169,0.3800107837,0.0283882637,0.0640436709,-0.3895127177,-0.2053598911,-0.1983939856,0.0231155287,0.1023291275,0.1775554419,-0.0978277475,-0.2141439766,-0.1075397059,0.1834543049,0.265666157,0.1858589947,0.0994038731,-0.3308607936,-0.0153337149,-0.6988524199,0.3644624352,-0.1859220713,-0.2121392787,-0.0337331071,-0.1135885343,0.1289971918,0.1067451239,0.6380493641,-0.1594728082,-0.2135081589,-0.0252164751,-0.0285395756,-0.3753391206,0.1385586262,-0.0885341614,0.077423133,0.5255035758,0.2285206765,-0.2473279387,-0.3309162855,0.1204197407,0.1124706864,0.0611604974,0.2027508169,-0.119944483,-0.2648466229,-0.1423734128,-0.2703365088,-0.0175430104,-0.0294683482,-0.1023098826,0.1695195436,0.1063460782,0.0093374401,0.2051511556,0.0681892037,0.2678184211,0.146043852,0.0274786949,0.4487716258,0.1139915213,0.086744301,0.4714268148,-0.1032421365,-0.3057269156,-0.0600751154,-0.1014952362,0.206653446,0.3470588028,-0.0293838661,0.0787232071,0.4206154048,0.3170092702,-0.2627899945,0.3010169864,0.0994896144,0.0302880406,-0.2148532718,-0.476849705,0.5081979632,0.3416750431,-0.3433972597,0.2625451684,0.4927096963,-0.0531527512,0.0967917442,0.2523823082,1.034614563,-0.1930854172,0.2237912565,0.0852304101,-0.4800776839,0.7641957402,-0.0895376503,-0.0198587757,-0.2831858695,-0.0798301846,-0.0351600945,0.0226884149,0.2910289466,-0.2171241343,-0.1762429923,0.348484993,0.0904807448,-0.1114586145,0.0834456757,0.6317034364,-0.0750549957,-0.2384598702,0.0673251972,0.1589185596,0.0471775495,0.3382358253,-0.0451856293,-0.2935827672,-0.2769960761,-0.2852519751,-0.3940753937,-0.1385869235,-0.3047077656,-0.3636064529,0.1867910773,-0.3848995566,0.1989908367,0.3855169415,0.3840226531,0.0221816059,-0.4268711507,0.3251349628,-0.2453621924,0.0101173529,0.1136080548,0.0822718814,-0.0001878395,-0.0778039023,-0.0501943007,-0.2651414871,-0.0375524014,-0.0012476017,-0.2398385406,-0.2649158537,-0.2237215191,-0.2702524066,0.0799018517,-0.0053486526,0.027197266,-0.3168895841,0.1434590816,0.1498686522,0.1210499704,-0.100949131,0.1754241139,-0.0171348248,-0.2336532176,0.3949896395,0.0987022966,-0.2478135228,0.3631983697,0.1876827627,0.0891630277,-0.275011003,-0.1759792268,-0.331581533,-0.4360620677,0.196702823,-0.0578706861,-0.0183790084,-0.331179738,0.4550181329,0.1125785038,0.0920193344,0.2105617076,-0.1774556339,-0.1088132635,0.2028252184,-0.2699633241,0.1475846171,-0.1648818254,0.2170677334,-0.0061571435,0.0703929663,-0.3570075929,0.1066355631,0.071439147,-0.2577008605,0.413856864,0.0059668054,0.2974036634,-0.3473897576,0.0965173766,-0.2296427786,-0.149203226,-0.1460534334,-0.1978464127,0.1179259568,-0.0511802696,-0.103520982,0.008437492,0.1760369092,-0.0851682872,0.0256757177,0.1940364838,0.1707769632,0.0258151367,0.2418554425,-0.3848331571,0.1925778687,-0.0843412504,-0.1526546925,0.1264733225,0.1326934844,0.2507071495,0.128470242,0.0455969609,-0.034471903,0.1687145233,0.1592748314,-0.0283781216,0.0540949143,0.3192533255,0.0139429783,-0.1999294013,0.2815964818,0.4696665704,0.0512596406,-0.1840696186,-0.098266907,0.1318157613,0.2298227847,-0.2421250045,0.0049383095,0.0643519461,0.0109462244,-0.0479023792,0.2834136784,0.3438400626,0.1482663155,0.0320158154,0.1607735157,0.4740704,0.0195488092,0.5068218112,-0.0261885654,-0.0095009143,-0.0519745946,0.2411175072,0.329505831,-0.0639387742,0.0194633957,0.2065877318,-0.0402674228,-0.1416194141,0.0597732402,0.0137489559,-0.2748443484,0.4676835835,0.1454845369,0.0223825444,0.0605650432,-0.1305901706,0.6777181625,-0.1140309274,-0.2497959882,-0.1537914276,0.1929056942,0.2039208561,-0.1425682455,0.0669996738,-0.2376960367,0.2320131212,-0.0627016202,0.045043271,-0.2563023269,0.274399817,-0.1350105107,-0.2028084099,-0.550888896,-0.0921725109,0.0362579338,0.4914512932,0.1174814254,-0.0317193568,0.0606376044,-0.2071910501,0.1589747518,0.3202861547,0.3550580144,-0.0209754221,-0.0295272637,0.2543513477,-0.0984321162,-0.0688225627,0.223331064,0.0171338767,0.1622956097,0.0043224622,0.1920106262,0.1699302793,-0.089463301,0.0900943205,-0.062912412,0.4421292841,-0.3466348946,0.3665519655,-0.3184464574,-0.2987056375,-0.0443125963,0.0109018488,-0.2880377769,0.0050370833,-0.1852010339,-0.0236290898,0.0454036668,-0.173949495,0.0897116959,0.0912044793,0.4617469311,0.4285912812,0.2257491797,-0.1157429218,-0.2243168354,-0.3858318925,-0.2265107781,0.1120965257,-0.037691731,-0.0244556274,0.0045210924,-0.0206315219,-0.1284615695,-0.0057523567,-0.2075674236,-0.3327474892,-0.1275262237,-0.1211985946,-0.0396401919,-0.4357021749,0.1472252607,0.1295672357,0.2058864087,0.0804684758,-0.1151735634,-0.004390528,-0.0741336346,0.1391678154,-0.0669323802,-0.1967879683,0.2871537209,0.2980929017,0.098992452,-0.1036701649,-0.3567239642,-0.1601319909,-0.1744984984,-0.3144659996,0.215510875,-0.1157386079,0.4652203918,0.0523505881,-0.4345726669,-0.3290249407,0.4114646018,0.2819351256,-0.1494358927,-0.4482078254,0.3392561376,-0.0214199051,-0.0412177853,0.0471051149,0.1506477743,-0.0640787184,0.0207512304,-0.3129019737,-0.2184680849,0.7305489779,-0.2740422189,-0.0206853542,0.023259243,0.3610419333,-0.0429601185,-0.1385860741,-0.6399489641,0.20878914,0.1995987892,0.1078687981,-0.2172808349,0.2350383252,-0.0770466924,-0.226508975,-0.0259282645,0.6818796396,-0.1068787724,-0.2430654913,0.1494418234,-0.1299390048]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1713","title":"Installation using conda","comments":"Yes indeed the idea is to have the next release on conda cc @LysandreJik ","body":"Will a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.","comment_length":14,"text":"Installation using conda\nWill a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.\nYes indeed the idea is to have the next release on conda cc @LysandreJik ","embeddings":[0.0085202865,-0.0874278247,-0.1298305392,0.1607175767,0.1904288083,-0.0082974453,0.1289227903,-0.120206818,-0.3218152523,-0.2392597646,-0.1998030692,0.0481313542,-0.1469176114,0.6642385125,0.4709976912,-0.1980905682,0.230057776,0.2197075635,-0.6107321978,-0.0725707859,0.0960880816,0.2435639203,-0.1832856983,-0.1401586682,-0.3814275563,-0.0527646206,-0.104694128,-0.1983786821,-0.0408273041,-0.2789980769,0.5686342716,0.1479544044,0.260192275,0.5703943968,-0.0001209382,-0.0918949246,0.1459111273,-0.0566010065,-0.3119974136,0.0266024042,-0.3773542941,-0.3670848012,-0.0368924737,0.0755847692,-0.0071028648,0.0574524254,-0.0343534462,-0.076880157,0.3324482739,-0.0622958764,0.1488010287,0.3692223132,0.2050758153,-0.0566151403,-0.2078104019,0.4338385463,-0.2863973975,-0.091802448,0.1146281511,0.1374934018,0.5829331875,0.0337610506,0.0347317234,-0.1659874767,0.1788112372,0.2152952552,-0.1929988414,-0.2725946903,-0.1435567737,0.2182329893,0.7750683427,-0.307957083,-0.6988129616,-0.2318817973,0.0060414858,-0.0501833111,-0.2015631795,-0.0944940075,-0.1846046597,0.2438854873,-0.2366686016,-0.6971508861,-0.228910014,0.091693148,-0.0884662643,0.5169668198,-0.1048987061,0.0600620918,0.1694575846,0.0143656963,0.2921688855,-0.1023892462,-0.0459958836,0.1542890072,-0.0756627694,-0.2829062045,-0.2668038607,0.3594132364,0.3089677989,0.1351480633,-0.3849028945,-0.1767903268,-0.2125577927,-0.2086182833,0.3127329051,0.0509150699,0.0518738292,0.4709951878,0.1055534855,-0.1470064223,0.5414571762,-0.0272678621,-0.0051570735,0.0826286152,-0.1837621331,-0.3932794333,-0.1267949492,-0.2734049261,0.045885317,-0.295308888,0.5333511829,-0.0954295173,0.0666823238,0.1532618552,0.0248378571,0.1387384683,-0.0387794375,0.2895154059,0.3521422148,0.0472706668,-0.067489855,-0.1862839162,-0.0341925956,0.2237980962,0.1861188263,-0.2063782662,0.144145146,-0.2449947894,0.3378726542,0.3289126158,-0.0870191231,0.2222128212,-0.181200847,0.4924482107,-0.4183914363,0.1912828088,-0.2499211133,0.2207408398,-0.2804434299,0.0703476816,0.0575644709,-0.0521668531,-0.1277970225,-0.0447617806,-0.1064504758,-0.3651297092,-0.2972142398,-0.0880346596,-0.2488771081,-0.1462747157,-0.0571463406,0.2271388024,0.1470898986,-0.0133511927,0.1056775004,0.3727695048,-0.4077263176,-0.4022584856,0.1309997886,-0.4224534035,-0.1437251717,0.0361965857,-0.0565248467,-0.079014346,-0.1103129834,-0.3406487703,0.4560738802,-0.5110286474,-0.2671888769,-0.0041431659,0.0054020197,0.2130565792,-0.1102920473,0.3521629572,0.2619479597,0.0216317158,-0.0374313183,0.2783651948,0.0247833654,-0.3653498292,0.2051915079,-0.218692556,0.1438448578,0.2253856808,0.2280387133,0.3477080166,-0.1348447055,-0.1070685461,0.140359357,0.1645570546,0.134008348,0.2068226635,0.3694903255,0.2735621035,0.1462167501,-0.2572901249,-0.4193299711,0.0440350659,0.0888983607,-0.0234750453,-0.1043612733,-0.2254040837,-0.0504649431,0.1623305827,0.0464891531,-0.1542392224,-0.0922342315,-0.1917153299,0.4502688646,-0.2130046338,-0.4134638608,0.9463129044,0.4027641118,0.3167490065,0.0801152289,0.2621009648,-0.0706618354,0.0921906456,-0.0597778708,0.2460314482,-0.0395743288,-0.275991112,0.2488142103,0.1629540324,-0.1728492081,-0.1435367465,0.3759490252,0.1687811762,0.2409506887,0.052501779,0.1792611033,-0.1747156829,-0.1221080199,0.0679095313,0.0111609008,-0.2055443227,-0.1157267764,-0.0640380755,0.0332675986,0.271150589,0.3413239717,0.2234383523,-0.1836045235,-0.1828825474,-0.1400717497,0.0091281356,0.234816283,-0.0404988676,-0.1199398339,-0.1763008237,-0.0914558023,-0.29995206,0.0044715083,0.3118252754,-0.2202157974,0.484329015,0.3101904392,0.2540208995,0.3320215642,0.1858151555,0.0840536952,-0.1446708441,-0.0229077432,0.0156310126,-0.0154685359,0.1626010686,0.0732462332,-0.2071736604,-0.2880797684,0.044233758,0.0726649761,-0.2506463528,-0.3325447738,-0.1028768495,-0.0879379809,-0.1117214784,-0.1566414386,-0.0298271682,-0.2989642322,-0.305665344,-0.2093284577,0.1059355885,-0.0180992857,-0.2336950004,0.343898654,0.1656990945,-0.2175828516,0.4524609447,-0.2636607587,-0.3406255841,-0.0961653739,0.2751854062,-0.0651593953,0.0721218139,0.041515559,-0.1478576213,0.1369083524,-0.1630835086,0.4213823974,-0.2479301989,-0.455111295,0.1519783437,-0.156507805,0.3002279699,-0.0268424079,0.1401988864,0.0872947946,-0.0059737219,-0.0734129176,-0.0122986231,0.1417942941,-0.2846796513,0.0029020037,-0.2556284368,-0.1349144876,-0.1055772454,-0.5397431254,-0.0586173609,0.5203054547,0.1558901072,-0.2158519179,-0.0522303581,0.012047329,0.0226878859,-0.1801690608,0.2152175158,0.0278938469,-0.1045605838,0.4464301467,-0.1267271638,-0.1552211791,-0.0613864213,-0.1491264701,0.3774144351,0.0875329524,-0.4490641356,0.2091958821,0.1380717605,0.117044501,0.1780235618,0.1425401568,0.5938482881,0.0237502009,0.2346215397,-0.0156898405,-0.3752746284,-0.1947610825,-0.018025063,0.1994974911,-0.0346568003,0.0453283414,-0.3425238132,0.2803937793,0.2046310753,-0.051887054,0.0646582469,0.252779752,0.6765316129,0.097477071,-0.1431703568,-0.0606764518,0.0969110876,0.1166679263,-0.1141615883,0.3138398528,-0.1560320705,-0.3966214955,0.0334995128,-0.2284776717,0.1340972483,-0.2234156132,0.0317327455,0.2431064844,0.0715662092,-0.0985236168,-0.0427849703,0.0628194436,-0.0106372954,0.3579872549,0.134325996,-0.0129335718,-0.2725042403,-0.0945824757,-0.3640406132,0.0164169595,-0.029701665,0.2139147669,-0.100987047,0.147721827,-0.1112243533,0.1459421962,0.8324124813,-0.0663237125,-0.6520237327,-0.240878582,0.1058156639,-0.2911931276,0.1072535515,-0.0308731608,-0.0780027956,0.16113846,0.3508254886,-0.2116964906,-0.1406017393,0.3701604903,0.1236857474,-0.0333372802,-0.3117132783,-0.100771755,-0.0906478241,-0.123091273,-0.1674504429,-0.0809931234,-0.1940903664,-0.0866348073,-0.3931413293,-0.0681917742,0.0035111078,0.2547830045,0.0546308532,0.1403577924,-0.1257375479,0.2682710588,0.3918375373,0.0260893032,0.1793254614,0.1645226628,-0.0973760933,-0.2702353597,-0.0864564255,0.0589680672,0.4315900207,0.0463718735,-0.0491094738,-0.2618038356,0.0228431448,-0.0241531506,-0.1073154062,-0.3294062614,0.3950648308,0.3321527243,-0.6212932467,-0.3546619415,0.4287908673,0.3164614737,-0.0517385788,0.3716968,0.4259189665,-0.1007000059,-0.1443474442,-0.0634440184,0.7754275799,-0.3642304242,0.3443341255,0.0010900191,-0.317951411,0.5111575127,-0.3539283872,-0.0583094731,-0.0281146653,-0.042351108,-0.2844640315,-0.0935558602,0.2859669924,0.3045707047,-0.4057635665,0.421018362,-0.0691119805,0.0714865103,0.0006459579,0.0398006588,0.2048077434,-0.2726687491,-0.075877279,0.1067397669,0.1615178734,0.411939621,-0.3725754619,0.0457134098,-0.2267694473,-0.1528674811,-0.2398363203,-0.0245642662,0.2051364481,0.0593271591,0.5930737257,-0.0603905581,0.0639143512,0.3770187795,0.3817611635,0.091209881,-0.3693625033,0.1199471503,-0.4232884347,-0.1382489651,0.0799581632,0.2128832042,-0.0328753963,-0.0657821745,-0.3385077417,0.0119134858,-0.0092015555,-0.3080902696,0.054795213,0.0397876278,0.0163753945,-0.1475831568,0.0032314307,0.4584843814,0.2371670455,0.0125820329,0.0837934986,0.2311681956,-0.092266947,0.3662792742,0.1484466195,0.0415792726,-0.2984237671,0.1511873454,0.0873988271,-0.3732914627,0.0199618824,-0.1275822818,-0.3670713902,-0.1025198624,-0.2821676433,-0.203011319,-0.0579882078,-0.2515562177,0.2057210356,-0.24725613,-0.2396308631,-0.0535000116,0.1482738107,-0.2274663448,-0.2963647246,-0.1730904579,-0.3435681462,0.2195361853,-0.1208001897,0.2619924247,0.0669923201,0.2094813436,0.1628261656,-0.019299604,-0.1150235087,-0.2076951116,0.134908095,-0.1625852883,0.5410006046,0.2909416258,0.2216081321,-0.0392708518,-0.0343529806,-0.0144761652,0.0361536592,0.0020892466,0.0660442635,0.1675362289,0.1078027561,-0.2773772478,-0.0586978532,-0.0067117279,-0.1210240722,-0.0267747846,0.0694738999,0.0729787946,0.1895100623,0.0758325458,0.2169328779,-0.0251715481,-0.1536463648,0.3392036855,0.4335432649,0.064753823,0.28091079,0.2138392478,-0.0119305942,0.1044758335,-0.1561672837,0.3959755599,0.2121253014,0.0695423856,0.3432674706,-0.2011945844,0.2424633056,0.3531263471,0.2570663095,0.0575112514,0.0721162632,0.0940453857,0.4507888556,0.0408944115,-0.164367184,-0.3985135555,0.1965371072,0.2867138386,-0.1898738444,0.5020639896,-0.0151404934,0.1791162342,0.6316930056,0.1223914623,0.070824042,-0.0072132759,0.5201701522,0.1738851517,0.262594521,0.143482849,0.3128729165,0.3060312867,0.0291832592,0.1994736493,-0.4889533818,-0.0130655384,-0.3160983622,0.1448079795,0.1736328751,-0.103915751,-0.2165749669,-0.2020706832,0.2571739256,-0.0357197821,-0.105275251,0.3954840004,-0.1940326989,-0.1363813579,-0.4379023314,-0.0508810617,-0.2005168945,-0.1115949675,0.6171352863,-0.1486629695,0.2920290828,-0.1185842156,0.226302281,-0.5018606186,0.3688622713,-0.1104899719,0.1503191292,-0.1705832779,0.3132563531,0.125017792,-0.202464968,0.2969697118,0.1460269541,-0.0792468414,-0.0335114859,0.0042616534,0.1864615232,0.1341087222,0.0572237633,0.3626267016,0.1552016437,-0.1667081267,-0.0003055366,-0.0535846874,0.3011091352,-0.0096564218,0.0876167938,-0.1303964555,0.1128876507,-0.000345749,-0.0169417858,0.01302501,0.3422029018,-0.3872022033,0.3435955942,-0.3002128601,-0.1911319196,-0.2993788123,0.1556406766,-0.2684454024,0.1538393199,0.3637365103,-0.2725538909,-0.1981485337,0.0051452885,-0.0035622884,0.0310944598,0.2757793367,0.0610266849,0.5044983625,-0.0357892364,-0.2397720069,-0.3093130291,-0.1879573762,-0.1205047071,0.0118444124,-0.1336364448,-0.1546019316,-0.0175601523,0.1118064448,0.1053544879,-0.1272592545,0.0390902497,-0.4134025872,-0.1335238367,-0.0739803463,-0.2479567379,0.3897983134,0.1449882835,-0.1267722249,-0.0601378903,0.1271502525,-0.2355958819,-0.0232169721,0.3331075609,-0.2177167386,-0.168154493,0.2821484208,-0.2181236446,0.5252028704,0.0692683384,-0.0858030915,-0.1336628348,-0.4320104122,-0.1106515899,0.0313918367,-0.0632825717,-0.1273757517,-0.5843636394,-0.130471468,-0.1481877714,0.0381195135,0.0350018628,-0.0198153276,-0.1269294769,0.0190878902,-0.2350707203,-0.0517153405,0.2394971848,0.1602450162,-0.1531550586,0.0384499505,0.0658078864,0.2183180302,0.4845194221,0.0190102011,-0.0198731776,0.00507581,-0.2939732075,0.0793381184,-0.4572471082,-0.7549019456,0.2284853458,0.1402245313,0.1586521119,0.0918826461,0.2804740071,0.1132152602,-0.1678627729,-0.1245678961,0.5668264627,-0.1809016615,-0.0259100161,0.1181611791,0.0431374907]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1713","title":"Installation using conda","comments":"Great! Did you guys have a timeframe in mind for the next release?\r\n\r\nThank you for all the great work in developing this library.","body":"Will a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.","comment_length":24,"text":"Installation using conda\nWill a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.\nGreat! Did you guys have a timeframe in mind for the next release?\r\n\r\nThank you for all the great work in developing this library.","embeddings":[-0.0539567955,-0.0376583301,-0.1801617593,0.1761816591,0.191117093,-0.1741129607,0.1625533998,-0.0670141056,-0.3270974159,-0.221177429,-0.2294666469,0.0597516261,-0.188989535,0.5104839802,0.3297406733,-0.277556479,0.2336804867,0.186041683,-0.6403389573,-0.0971227363,0.116015628,0.3619967103,-0.2395424843,-0.1564272493,-0.3490127921,-0.0934232697,-0.13800475,-0.1415150762,-0.1676210165,-0.2464525402,0.5433064103,0.1990732551,0.213730216,0.6924842,-0.0001173334,-0.2475258708,0.2480924875,-0.0030501266,-0.2502144873,-0.039988324,-0.2281609029,-0.3681676686,0.0377836451,0.0528491288,-0.019863531,0.0676723197,-0.057624232,-0.0261221956,0.4012193978,-0.0857416764,0.1677254438,0.3654803336,0.2736262083,-0.0947252661,-0.1673788875,0.433835417,-0.3844279945,-0.170255363,0.2569201887,0.0808812082,0.4432737231,0.0667961538,0.0478788279,-0.2124155611,0.3321673274,0.1987416595,-0.2067472786,-0.3278177381,-0.2350192666,0.2532398403,0.8008554578,-0.2084442675,-0.6760145426,-0.2062558681,-0.1349032521,-0.0220964476,-0.0845387205,-0.1256479323,-0.1271618903,0.1594615728,-0.3282053173,-0.6029955745,-0.3283930123,0.0234305058,-0.0905868635,0.5560168028,-0.140727669,-0.0073349727,0.2014150769,-0.0811844841,0.3116420507,0.0117755476,-0.0864330903,0.2956623435,-0.0735695139,-0.3632636368,-0.1712514162,0.3323934078,0.2996618748,0.2440695465,-0.374333024,-0.0830928087,-0.1668815166,-0.1910725981,0.3662338853,0.1367796212,0.0134646604,0.391607374,0.2731012702,-0.0827826038,0.5417615175,-0.0868362039,0.03892814,0.0535690747,-0.3300046027,-0.3209067285,-0.0340300053,-0.3715228438,0.1857255846,-0.3058177829,0.6140985489,-0.3056350946,0.0181265455,0.2475005239,-0.1649832129,0.2572250962,0.0021929587,0.1915689558,0.2798495591,0.2018961459,-0.0706252903,-0.0543984249,-0.1438066661,0.1062562168,0.243094191,-0.1890078038,0.1164429337,-0.1818544716,0.408199966,0.3598951101,-0.0997912437,0.1833704263,-0.1032745168,0.4986630678,-0.4198851883,0.1544222683,-0.3225454092,0.1515449882,-0.2846422195,0.061768163,0.1592708081,-0.0116853099,-0.2567635477,0.0046030814,-0.2204418778,-0.2602276802,-0.2112771273,-0.0464954861,-0.2655153573,-0.0555664301,-0.0740792006,0.2384800613,0.2092833966,-0.0630939603,0.1236935332,0.2759695053,-0.350055486,-0.2852534354,0.1542486548,-0.4082544446,-0.1795652658,0.1058377326,-0.1141242832,-0.086944364,-0.0579333343,-0.3439196646,0.4250664413,-0.507025063,-0.1991416514,0.013025946,0.0835282132,0.123656258,-0.0307742599,0.2835633159,0.299123466,-0.0579781793,-0.1886734515,0.3549758792,0.073148571,-0.3144322336,0.0118684927,-0.2986495495,0.1209744439,0.1903316081,0.292054981,0.293209523,-0.0972980186,-0.0304683801,0.1906959713,0.1425542086,0.0983592644,0.2162647396,0.442346096,0.3299351037,0.1988466084,-0.1986918747,-0.4077663124,0.0472801737,0.1282979548,0.0997155905,-0.0727272481,-0.188121438,0.0218185782,0.1400581002,0.0449401364,-0.1936963052,-0.047278136,-0.1908100098,0.4457260668,-0.1792605668,-0.4518477917,0.8676214218,0.4596750736,0.3349504173,-0.1556449234,0.2536652386,-0.1142603531,0.0273682885,-0.1045836806,0.2058419287,-0.0777752623,-0.3490960002,0.1965986937,0.1790527552,-0.1173755378,-0.0738114268,0.3485862017,0.2091695219,0.4061680138,-0.0318085738,0.2890359461,-0.2030059993,-0.0077361902,0.1084442064,-0.1507164687,-0.1262257099,-0.2082740813,-0.0573269241,0.1094209552,0.133650139,0.3417794704,0.1257269382,-0.2899388969,0.0026717156,-0.0403357483,-0.0039170687,0.1421415359,-0.0493624061,-0.1159399748,-0.1446531266,0.0648135394,-0.2677936852,-0.1272663176,0.2468546778,-0.1077841297,0.379997313,0.3027243614,0.2017099708,0.3015313148,0.155334264,0.0641661361,-0.1555909514,-0.0644165501,-0.0047028014,0.0192373674,0.1764005274,0.0269128364,-0.1974040121,-0.2620178461,0.1294233948,0.0004806252,-0.2759349048,-0.3955468833,-0.1232315898,0.0388639607,-0.0889190361,-0.0142150717,-0.025792053,-0.1741257757,-0.4335142672,-0.2850687206,0.0618094578,-0.0128492685,-0.2115723789,0.3294836879,0.1963211745,-0.0935901105,0.400668174,-0.2807404697,-0.1898596734,-0.2390385568,0.2319550216,-0.0324195772,0.1155240387,0.0665080845,-0.1564191431,0.1868631393,-0.0507902727,0.3589324653,-0.1714442968,-0.5654632449,0.1169636771,-0.1695115119,0.3379515707,-0.0229033399,0.153260693,0.115314506,-0.1332169771,-0.0962347761,-0.1714169979,0.0874570385,-0.3237069249,-0.0788585097,-0.2840440869,-0.174318552,-0.1860374808,-0.4251682162,-0.1508772224,0.6054794192,0.193875134,-0.1933217794,0.0701252818,0.0246518441,0.0297382157,-0.2609931827,0.2217420042,0.0211420208,-0.20427683,0.4779252708,-0.1709124744,-0.2292917818,0.0040923553,-0.0835775211,0.3345167637,-0.0463145748,-0.3969550431,0.1388846189,0.0042131525,0.0415259153,0.1889078319,0.249149248,0.547753334,0.0593655482,0.1654090732,-0.0135443173,-0.3746863902,-0.238566041,0.0132685238,0.2313839197,-0.0185508989,0.0202506669,-0.2800436616,0.3675199449,0.2129456699,-0.0010246282,0.0391097777,0.3218259513,0.5508820415,0.0800627172,-0.2101369053,-0.0043514953,0.0992707834,0.1164324433,-0.0525781251,0.3405482173,-0.225272879,-0.5008480549,0.1265865713,-0.2486708313,0.0993179977,-0.1432224661,0.0597230941,0.1643359512,0.0505350791,-0.01243304,0.0149086015,0.0444039777,0.0273146406,0.29391855,0.1508589387,-0.0181824118,-0.2938577533,-0.0367759578,-0.4289734662,0.0541446842,-0.1224865839,0.0330711678,-0.0582186617,0.0772222653,-0.0672420263,0.0093627488,0.8861726522,0.0370968319,-0.625944972,-0.1034100354,-0.0453878418,-0.2085832059,0.1288173646,-0.0719757006,-0.0969725251,0.2236471921,0.3609751761,-0.1424887329,-0.2381036282,0.3552711308,0.1013857126,-0.0492369905,-0.2747233808,-0.0755080357,-0.1256340742,-0.2078076899,-0.1424304992,-0.0431453884,-0.1783043146,-0.1376971751,-0.3707946241,-0.0091292895,-0.0258179139,0.2059514225,0.0582744963,0.1896784455,-0.1677902341,0.3184922338,0.4929773211,0.0084955432,0.0350259691,0.1288127005,-0.1933210045,-0.1548799127,0.0342035741,0.0520271249,0.439458698,0.2427354306,0.0045822933,-0.2955552638,0.0223401058,0.0767623335,-0.1349578798,-0.3483427763,0.3332134783,0.2800482213,-0.566963017,-0.4329418838,0.351996243,0.3097464144,-0.0446384214,0.4278448224,0.4177296162,-0.0480137989,-0.0220433623,-0.0204607863,1.0245965719,-0.3411810994,0.2225030214,-0.0526758879,-0.2719838619,0.503892839,-0.4234574437,-0.0485702902,-0.1772782952,-0.0335520282,-0.2366062254,-0.1937131137,0.3874120116,0.1829225421,-0.4559367299,0.3584805131,-0.0175021011,0.0055543883,-0.0883878842,0.1998120546,0.0786669701,-0.3554798067,-0.1525533348,0.1492495686,0.1797525138,0.396699965,-0.3227911294,-0.0522865281,-0.2029496282,-0.1512458771,-0.2440911084,0.0148503883,0.2328116745,0.0305459034,0.5970540643,-0.0776579008,0.123431094,0.3457081914,0.4243237972,0.1735762656,-0.3065683842,0.2344344556,-0.3720159829,-0.1590754241,0.0074737389,0.2029972076,0.0571399443,-0.1579746157,-0.2544444799,-0.0921692625,-0.0892060921,-0.2849508226,-0.0085862987,0.0215221848,0.0655013695,-0.1775936484,0.1331942528,0.3783177137,0.1634024382,0.044039987,0.1453079283,0.4319881499,-0.1645714641,0.4580816627,0.2118549347,0.0654024109,-0.288013041,0.169289574,0.0601291843,-0.5256131291,-0.0620498583,0.0509218983,-0.4061178863,-0.1486798227,-0.2537130415,-0.1426845044,-0.1699756831,-0.1892681271,0.1566538364,-0.2482797801,-0.1302248836,0.095634222,0.1859773844,-0.2149008214,-0.3124377131,-0.1637180001,-0.3465153873,0.2990145385,-0.0815379545,0.3616744876,0.0277609229,0.0751556382,0.3091571033,-0.0129086412,-0.2013931423,-0.1603127867,0.1159133166,-0.124592036,0.4493319094,0.0799322128,0.2086569667,-0.0896913409,-0.0049086851,0.0538300201,-0.1456175297,-0.0520288311,-0.0660970956,0.1519003361,0.0779103339,-0.2837515771,-0.0983742252,0.0157785192,-0.1296202242,0.0332930908,0.0458136685,0.0502206534,0.1737287045,0.096241191,0.244290933,-0.0218371097,-0.0550006591,0.2851760089,0.4928340316,0.1117971018,0.3702398837,0.2209517509,-0.0318531394,-0.0180223808,-0.063819401,0.3279290795,0.1335070133,0.1007733792,0.3453576863,-0.1437817216,0.195451051,0.2943042219,0.4065571129,0.1574709266,0.0500612594,0.1322301179,0.3768377006,0.1130056158,-0.1438959837,-0.4009777904,0.0534145869,0.2222620547,-0.2158798575,0.3926509619,0.0579326563,0.1295057684,0.6797934771,0.194378078,0.22262685,-0.0924272016,0.5110803843,0.1530337781,-0.0122200809,0.0218144972,0.3314880133,0.3703776002,0.0434366316,0.3197548687,-0.3981615901,0.0058647669,-0.3019353747,0.2494276613,0.1640136242,-0.0950122625,-0.1212752908,-0.2490774542,0.1495940983,-0.1024499461,-0.0966309607,0.3726489544,-0.200670287,-0.135712564,-0.4122929871,0.0226082411,-0.1376242936,-0.0967173129,0.5791195631,-0.1789011359,0.2259738594,-0.1329737455,0.1426041424,-0.3491810262,0.3758728504,-0.0531374924,0.1682198048,-0.1773201972,0.3210996687,0.032514412,-0.0688769817,0.238220647,0.2318128794,0.010832659,-0.0422431193,-0.1797049791,0.1846616864,0.2092236131,0.027041927,0.2010369599,0.2089030892,-0.2073328793,-0.0391226448,0.0013758822,0.4193413556,0.0021673976,0.1157809123,-0.125527665,0.121538341,-0.0640331283,0.03905892,-0.1313105822,0.2916077077,-0.3792625964,0.4246207178,-0.1983202249,-0.2484906167,-0.2704584002,0.082819052,-0.3937572837,-0.0076762699,0.3037269413,-0.2559802234,-0.1656371504,-0.1114124507,0.0256098546,0.1360265911,0.2756669521,0.0475115031,0.6115580797,-0.0579745248,-0.3320987225,-0.3333335221,-0.2172275037,0.0463396497,0.2531864643,-0.1766962111,-0.0611169226,-0.0339315943,0.1957744062,0.1287039667,-0.2622005641,0.0566341877,-0.3625936508,-0.1561058611,0.0738629177,-0.1369580179,0.3464947641,0.1373437196,-0.0883231461,-0.0758523047,0.0845620483,-0.1794372648,-0.0995976403,0.2675313354,-0.0915983915,-0.1305022538,0.1973101944,-0.1003046706,0.4405755103,0.1412246376,-0.1319754869,-0.1559740454,-0.4620077908,0.0151223792,-0.032638412,0.0345832147,-0.0583523065,-0.4577644467,-0.1961345375,-0.1882942468,-0.0073617343,-0.0135914441,0.0246843677,-0.1915992796,-0.069600746,-0.2155485004,-0.0228579622,0.1813710183,0.231633842,-0.1442910582,0.0106616877,-0.0916071832,0.1752523035,0.5658098459,0.0055523249,-0.1327789873,-0.0999864563,-0.378987819,0.0443546399,-0.354832083,-0.7771007419,0.2319705635,0.0748231485,0.0685372576,0.0235214587,0.307708472,0.2624305487,-0.2073735595,-0.1677978188,0.5492648482,-0.0581691265,0.0257235318,0.1097707227,0.0490880087]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1713","title":"Installation using conda","comments":"I think we can have `datasets` on conda by next week. Will see what I can do!","body":"Will a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.","comment_length":17,"text":"Installation using conda\nWill a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.\nI think we can have `datasets` on conda by next week. Will see what I can do!","embeddings":[-0.0149538312,-0.0426130258,-0.1236476302,0.1562701762,0.2540784776,-0.009232902,0.1276855618,-0.1303180307,-0.2962277532,-0.2111263871,-0.2347903401,0.0809221417,-0.1403378844,0.6276564002,0.4649843276,-0.1856292635,0.1998481005,0.2295516729,-0.5902295113,-0.0625070333,0.1372841001,0.2634812891,-0.1885326356,-0.1960845441,-0.359808445,-0.0598923415,-0.113546215,-0.1968571693,-0.032453455,-0.2604272366,0.5524133444,0.159466356,0.2478712946,0.5726000667,-0.000118327,-0.0796176419,0.1754110008,-0.0648775473,-0.3543756604,0.0248022713,-0.3400885463,-0.3828353286,-0.0146474903,0.0758211985,-0.0000120852,0.0436181724,0.0059324247,-0.0815959498,0.3228554428,-0.0134683615,0.1626114994,0.3992315233,0.2094788998,-0.1016364098,-0.1831054389,0.4471345246,-0.3012720346,-0.1141189635,0.1077052578,0.0791413113,0.5704078674,0.023409361,0.0221342593,-0.1581074744,0.1740729958,0.2197768241,-0.2278297395,-0.2767653167,-0.1618104279,0.2351419032,0.7575784326,-0.3252034187,-0.6390735507,-0.2428076267,0.0055800704,-0.0127070127,-0.1900894791,-0.0598195679,-0.1736702919,0.2446870804,-0.2471012622,-0.6649011374,-0.2245127559,0.1040985435,-0.0616721809,0.4285608232,-0.1106449515,0.0321358144,0.2054019421,0.0325856656,0.2649052143,-0.0849942341,-0.0678414628,0.140592739,-0.0855201185,-0.2367231995,-0.2231902629,0.3559328318,0.3252642155,0.1699939519,-0.3703601956,-0.1642525643,-0.202173695,-0.1812936813,0.2401769608,0.0896565914,0.0804530829,0.4821060002,0.1383990645,-0.1284577549,0.4942330122,-0.062205296,-0.0256393421,0.0839998052,-0.2074997127,-0.3596906662,-0.1000820696,-0.2758504748,0.050233081,-0.3096505404,0.555665195,-0.149875164,0.1206117347,0.2145168632,-0.0139045715,0.1138209775,-0.0119964546,0.3075073063,0.3615592122,0.0330429077,-0.0756502971,-0.1224210635,-0.0381389149,0.1636590958,0.2613064647,-0.2416059524,0.143258661,-0.2239934653,0.3553830385,0.3323233426,-0.0616959222,0.2217128873,-0.1798949689,0.4900628626,-0.3739978671,0.2030829042,-0.2409490049,0.2028486878,-0.2840315104,0.0101918885,0.0449461192,-0.0374019295,-0.1617485583,-0.0304193981,-0.1271844953,-0.344838053,-0.3084508181,-0.055344943,-0.2809699476,-0.1488776058,-0.0635116473,0.2492283732,0.1657638401,-0.0337959081,0.1868308038,0.3834117651,-0.4151777923,-0.371511966,0.1071176976,-0.4512655139,-0.1253461689,0.0196368117,-0.0484802127,-0.0370420888,-0.1207711175,-0.3446337283,0.4303440452,-0.4813029468,-0.2709692717,-0.0350110382,0.0457312837,0.1971788406,-0.1197522283,0.3125359416,0.2328032553,0.0341631547,-0.0303027425,0.3453214765,0.0070707104,-0.3574683368,0.1885863394,-0.2112373561,0.1206962764,0.2460269928,0.2226640284,0.3194247484,-0.1441022456,-0.1359203756,0.1034693643,0.1362401992,0.134831354,0.2435184717,0.3739115,0.3233315051,0.162702471,-0.295645535,-0.4294299781,0.0684101731,0.1442991346,0.0211063884,-0.1536229104,-0.2209580541,-0.0352288149,0.1679532677,0.0712398142,-0.1438247263,-0.0619867183,-0.1933827847,0.4818265736,-0.1869964451,-0.383025825,0.9310396314,0.4141563773,0.3214601278,0.0335685685,0.2666040957,-0.0673376247,0.0461876914,-0.0373503491,0.2509399652,-0.0717819929,-0.2659291327,0.2164591104,0.1663508117,-0.170663029,-0.1484289467,0.3806863427,0.1441078484,0.2186952382,0.0165106896,0.1651927531,-0.1783978641,-0.0853505284,0.0703583658,-0.0803250521,-0.2433673143,-0.1189964935,-0.0555518232,0.0645686015,0.2935171723,0.3672555387,0.2279906422,-0.1893626451,-0.1606578827,-0.0933784917,-0.01783875,0.2225294858,-0.0340419337,-0.1180749536,-0.179772079,-0.0553299226,-0.2879695892,0.011367199,0.3001757264,-0.2451165468,0.5096589327,0.2987688184,0.2240209579,0.3043924868,0.2122796476,0.1009572968,-0.2046419382,-0.0767455772,-0.0063806823,-0.0253976416,0.1729675084,0.0567822009,-0.2454538196,-0.2457572669,0.0504124723,0.0275614504,-0.2938050032,-0.3847328424,-0.0910779312,-0.1032777503,-0.1019808576,-0.1778522134,-0.0333461575,-0.211353451,-0.2985105813,-0.1774642169,0.0904571563,-0.027546538,-0.2283671796,0.3074950874,0.1781060547,-0.1851152033,0.4231058061,-0.2911998034,-0.3212713599,-0.1236142144,0.2300423235,-0.0725589693,0.0876067504,0.0645356327,-0.1275738031,0.1365977824,-0.1487146616,0.429533869,-0.219564572,-0.4638290703,0.1157692447,-0.1211456135,0.3696663976,-0.0518055744,0.1717076451,0.0728386343,-0.0547171868,-0.0372060575,-0.0015676746,0.1516359895,-0.2317011207,0.0019128555,-0.30086115,-0.2050743699,-0.1011213213,-0.5442736745,-0.1078187376,0.5183225274,0.130866304,-0.2203021199,-0.0608644225,0.0415272638,0.0609666854,-0.2225181013,0.2633149922,0.0544292666,-0.1433713883,0.5334441662,-0.1355813891,-0.1888212711,-0.0515728369,-0.1765880883,0.3594616652,0.0756084546,-0.420894444,0.2093579918,0.0704013109,0.1192298308,0.1845211834,0.1898522824,0.5759529471,0.0368786715,0.2318338305,-0.0334946625,-0.3571576476,-0.1830795854,-0.0663716644,0.2051074952,-0.0745304078,0.0747422427,-0.3267486691,0.2876844704,0.1867843121,-0.0531546175,0.1046712101,0.2102616727,0.6634022593,0.1246430129,-0.1961597055,-0.0709323436,0.1021140367,0.1374605,-0.082986936,0.3483999372,-0.1684519798,-0.4113773406,0.0630350411,-0.2537770271,0.1128223985,-0.1911642551,-0.0000814393,0.1922035962,0.0669834837,-0.0701112151,-0.067476511,0.0527963229,-0.0163120478,0.328846395,0.0769025534,-0.0315424986,-0.3007464111,-0.1023317575,-0.3768817186,0.0127387345,-0.0685328096,0.1878643185,-0.1366524398,0.1729701608,-0.1371849179,0.1110708937,0.8732082248,-0.0608148947,-0.6852087379,-0.2119358629,0.1207319051,-0.3259773552,0.0769197047,-0.0545213483,-0.077271305,0.1502979845,0.3744634092,-0.2502126694,-0.1153742224,0.3510664105,0.1523091197,-0.0310954154,-0.3121961057,-0.0844720975,-0.126628533,-0.1468068212,-0.1661560088,-0.0606567413,-0.2059601396,-0.0625759363,-0.3869854808,-0.0785545111,0.0194121432,0.2817112207,0.0650488883,0.1591905951,-0.0888443515,0.3104571402,0.4042408168,0.0072452403,0.1504322737,0.1684994102,-0.1428222656,-0.297886014,-0.1026319042,0.0500545688,0.4498151541,0.0156628918,-0.0586882159,-0.262507081,0.0350013115,0.0037447326,-0.1459736377,-0.2563108802,0.35638237,0.305609405,-0.5959278345,-0.4213732183,0.406611383,0.3207868636,-0.0349358656,0.3828692734,0.3980771601,-0.0952873155,-0.1334001124,-0.1130599752,0.7541900277,-0.3667335808,0.2954513729,0.0371130072,-0.2931689322,0.5418283939,-0.3917771876,-0.0565686449,0.0122728208,-0.0946855843,-0.2348124385,-0.115928553,0.290025264,0.3145033121,-0.433935374,0.3982245922,-0.0898617357,0.057785403,-0.0038346681,0.061300315,0.1651471853,-0.3159349263,-0.0855446756,0.1321151704,0.1346388459,0.4040126204,-0.3301079571,0.0501513071,-0.2412516922,-0.1461052448,-0.2597301304,0.0420535356,0.1950344741,0.0809166282,0.609246552,-0.0078313239,0.0422807559,0.4076487124,0.3599236012,0.1351239085,-0.3579594493,0.1633001417,-0.3922894597,-0.1684909314,0.0549592189,0.2180157602,-0.0096596889,-0.0550150946,-0.3342308998,0.0008060107,-0.0294922888,-0.2681893706,0.053992305,0.0239078272,0.0236015655,-0.116948843,-0.0165982507,0.4251253307,0.2843789756,0.0253885277,0.1049015224,0.2831371129,-0.1061252952,0.3961710632,0.1556447148,0.0899262428,-0.3006476164,0.1609143913,0.0359721258,-0.4082327187,0.0516072474,-0.042329628,-0.3699823618,-0.1215574592,-0.2851475477,-0.1596379578,-0.0446700975,-0.2476648092,0.1840982884,-0.2189014703,-0.2216067016,-0.1138110757,0.1385852695,-0.2929534018,-0.2962226272,-0.2043326199,-0.3674982488,0.2221288085,-0.0678888559,0.2909976244,0.0843052045,0.1516161263,0.1719044447,0.0139418924,-0.1381873339,-0.2102385312,0.0888931975,-0.1440108716,0.5395697951,0.2788912356,0.2511493862,-0.0777839497,-0.0063120378,-0.0443625897,0.0527859256,-0.0109287584,0.0345649868,0.1527853012,0.0817854628,-0.253554523,-0.040444456,-0.0091410214,-0.1237372681,-0.0865024328,0.0515380763,0.0620674491,0.1488169134,0.1053896993,0.182298243,0.0029109782,-0.1166348979,0.3144982159,0.3794343174,0.0370800048,0.2952955067,0.1985396296,-0.0022336585,0.0792814046,-0.1164224595,0.3607226312,0.2053340971,0.0779614449,0.3609716892,-0.1540057808,0.2633631229,0.3500869572,0.2061882615,0.0167166032,0.0520868897,0.167712912,0.4607205093,0.0702986121,-0.1447550505,-0.3850226402,0.2137396485,0.289293319,-0.1552709341,0.4751921892,-0.0596183799,0.1794528514,0.577933073,0.1542572677,0.1014450118,0.0052223695,0.5116177201,0.1466652155,0.1921701282,0.096149832,0.2661855817,0.2873876393,0.0408827886,0.2661752403,-0.4669782817,0.0106267007,-0.3258081973,0.1498448402,0.1935580969,-0.1626450568,-0.1730595827,-0.2216648906,0.2443475872,-0.0009263232,-0.1346707642,0.4199364483,-0.2446666956,-0.0825810134,-0.4358354807,-0.0276466794,-0.1879342347,-0.1162275523,0.5522987843,-0.152867347,0.2951289117,-0.1290255636,0.2824722528,-0.4773631692,0.3376292288,-0.1106033102,0.1252856553,-0.1594455987,0.355797708,0.1588895768,-0.1523593813,0.2873933017,0.1230040118,-0.078565836,-0.0796688646,-0.0236223415,0.160233438,0.1859245896,0.0633263066,0.3489517272,0.1519738734,-0.1436861902,-0.0284351129,-0.082129173,0.3014133275,-0.0404597335,0.1227050647,-0.1453840733,0.1388833821,-0.0328169987,-0.055873327,-0.0331647545,0.3500949442,-0.355713129,0.3076306283,-0.3255294561,-0.2090114057,-0.2750300169,0.0853058174,-0.2894616723,0.1206173003,0.3848700225,-0.2447065413,-0.1913401634,-0.0527657159,0.014120304,0.0658660904,0.3058436215,0.0839056596,0.4952223003,-0.047329668,-0.2340963483,-0.3490592837,-0.2356703579,-0.1590826213,0.0272624865,-0.1302268803,-0.1499909908,-0.0111665362,0.1191718578,0.0777608231,-0.0803851634,0.03923456,-0.4372060299,-0.1903401315,-0.1089536026,-0.2377548218,0.3689498603,0.1963133514,-0.1196065322,-0.0647311509,0.096416235,-0.2048497647,-0.0412827395,0.3441520333,-0.2105277181,-0.1769166291,0.2914231122,-0.2044671625,0.5206035972,0.0656481758,-0.0573587529,-0.1465383917,-0.4169756472,-0.0528390668,0.0370214134,-0.0583939143,-0.089574188,-0.6124735475,-0.1727396995,-0.1507421434,0.0367892049,0.042364981,-0.0201984961,-0.1375287622,0.0041374904,-0.2497769147,-0.0135194911,0.2383354604,0.1118284985,-0.1766231954,0.0285090972,0.0809981301,0.2065770328,0.4909518361,-0.0105533656,-0.0322655924,-0.0155693861,-0.3327308297,0.0778910667,-0.4009783864,-0.6961501241,0.2136878371,0.1357451081,0.1262793392,0.0578861274,0.3085645735,0.1707357913,-0.1396750212,-0.1316262335,0.5406727791,-0.1637871563,-0.0049681473,0.1489530951,0.0853500515]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1713","title":"Installation using conda","comments":"Thank you. Looking forward to it.","body":"Will a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.","comment_length":6,"text":"Installation using conda\nWill a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.\nThank you. Looking forward to it.","embeddings":[0.0125354817,-0.081606552,-0.1193038076,0.1855625361,0.2079510987,0.0054804287,0.1137925163,-0.1881555021,-0.3080613911,-0.2556343079,-0.2470690012,0.0861123875,-0.1178962141,0.6067861915,0.4650484622,-0.2231969386,0.1940625161,0.2537831366,-0.608874619,-0.0715976283,0.1274662465,0.2532376051,-0.1835470796,-0.175617829,-0.3618765771,-0.0669761896,-0.0874827653,-0.1488614976,0.0118146483,-0.2646155655,0.5374698043,0.1614263207,0.2267152667,0.5546335578,-0.0001211785,-0.0556681193,0.1389309317,-0.0838581622,-0.3374754786,0.0350692347,-0.3743389547,-0.3764214814,-0.0150388526,0.0484451652,0.0233408343,0.0718753487,0.0005612299,-0.022672005,0.3994502425,-0.0466815569,0.1441554278,0.4026224613,0.2167762816,-0.0778971389,-0.2092528194,0.4406556785,-0.2638239264,-0.0726828352,0.1030707583,0.1366295964,0.5750223994,0.0072486997,0.0114433281,-0.1658126563,0.2338562906,0.1846917421,-0.2037954628,-0.2903779149,-0.144644931,0.2168079615,0.7535230517,-0.323772639,-0.6563157439,-0.2329462916,-0.0098256804,-0.0007064727,-0.2109329849,-0.0483572558,-0.1992621124,0.2231380939,-0.2649360597,-0.6493370533,-0.2234024107,0.0656702816,-0.0725194961,0.4327156842,-0.1113969684,0.046234183,0.1720936,0.0013134369,0.2390988469,-0.1125550568,-0.0429384373,0.1214704812,-0.0526222661,-0.2269072682,-0.2630386949,0.3514353931,0.3189166784,0.166984573,-0.3581607044,-0.1460025162,-0.2093607038,-0.1682951748,0.2378141433,0.0899299234,0.0587951206,0.4984940588,0.0788858905,-0.1654214412,0.4501239359,-0.0491757542,-0.0012716297,0.0920984447,-0.1893507093,-0.3971526027,-0.1531642824,-0.2910890281,0.0290743839,-0.2973873317,0.5416656137,-0.1144154593,0.1378297061,0.226386413,0.0170140062,0.0971467569,-0.0251805931,0.3411816061,0.3804722726,0.0613007098,-0.0606324077,-0.1152681783,-0.0202125497,0.1715049446,0.2304812223,-0.2155995369,0.2042556256,-0.2377731651,0.3690708876,0.3353696465,-0.0671855882,0.2223309875,-0.1624340862,0.4984617829,-0.3498467207,0.1900006533,-0.2589432895,0.2210944444,-0.2877407074,0.0348105319,0.0612666123,-0.0369257368,-0.1362108737,-0.0515530817,-0.1139925867,-0.3341027796,-0.280097127,-0.1064280197,-0.2677466571,-0.1432271004,-0.0521250367,0.228046149,0.167490676,-0.043708086,0.1542112529,0.4301126599,-0.3924926221,-0.392983526,0.1514794677,-0.4778983295,-0.1057393402,0.0124568436,-0.0650203675,-0.0454616584,-0.1062743142,-0.3308986723,0.4898227453,-0.4983777404,-0.2881285548,-0.0257658809,0.0337177664,0.2262957543,-0.1065872684,0.3409521282,0.2208638638,0.0255081132,-0.0270083621,0.344175756,0.0385390371,-0.3662377894,0.2036792934,-0.1772708595,0.1381119192,0.2551626563,0.2272174507,0.3773772418,-0.1365192533,-0.1296124756,0.1070476696,0.1268686652,0.1351827532,0.2634905875,0.3632307947,0.329544574,0.1270755231,-0.2820808887,-0.411981523,0.0637414008,0.082991451,-0.0109414449,-0.145162493,-0.2219063789,-0.0620639436,0.1404965967,0.0358999595,-0.1318959594,-0.0882180929,-0.1523270309,0.4941341579,-0.2210671753,-0.3513723314,0.9467833638,0.3983051181,0.3368335664,0.0717412159,0.2413452864,-0.056240771,0.0481727645,-0.0474054962,0.2498137802,-0.049321156,-0.2854135633,0.2381013185,0.1829729378,-0.1744476259,-0.1423763931,0.3645887971,0.1448051333,0.235680908,0.0378902927,0.1589725167,-0.1878440231,-0.0774122402,0.0612289794,-0.0200115554,-0.2335395962,-0.1079274341,-0.0728193223,0.0764946938,0.2444912046,0.3821183443,0.2005127966,-0.2000923753,-0.1650013924,-0.1113443971,0.0346763469,0.2406150848,-0.0434673429,-0.1245885268,-0.2363511175,-0.1042675525,-0.3033113182,-0.0109939342,0.2993929088,-0.2500708699,0.558968842,0.2868793607,0.2197446227,0.3270853162,0.2159544975,0.0840474367,-0.1527633965,-0.0499432832,0.001036511,-0.0489118248,0.1701762825,0.0530876219,-0.2280403823,-0.2711663842,0.0509931557,0.0399411507,-0.3092765808,-0.3894552886,-0.1162587777,-0.1105172113,-0.114002265,-0.1356248409,-0.0549023338,-0.2611543238,-0.3125962317,-0.209806785,0.1314287484,-0.0382491127,-0.2347847819,0.3171177506,0.2030922323,-0.2080367059,0.4187585711,-0.276773572,-0.4140197337,-0.0862174258,0.2753605843,-0.0795193836,0.0627378002,0.0784587339,-0.0813185722,0.1528854966,-0.1527926624,0.431135118,-0.2031306177,-0.4213750064,0.1201371849,-0.1229304001,0.3379208446,-0.059376847,0.1705462486,0.0921265408,-0.0646743849,-0.0452374332,0.0046241921,0.1662673056,-0.2720828354,0.0206898656,-0.3108380139,-0.19009085,-0.0872930959,-0.5379658937,-0.1065201312,0.5368382335,0.1936306804,-0.2255929857,-0.0617932416,0.0111323642,0.0548367575,-0.197877273,0.3018988073,0.0543993302,-0.0929409787,0.5265144706,-0.1223947108,-0.1754604131,-0.052941408,-0.1974130422,0.3788736165,0.106966719,-0.4118284285,0.1908218116,0.0759470239,0.1025479659,0.1897405535,0.2197789997,0.5848507881,0.0296384748,0.2418729514,-0.025607666,-0.3644325435,-0.1421122253,-0.0326175466,0.1986079514,-0.0360369571,0.0408601612,-0.3676706254,0.3265113235,0.1966288388,-0.0804853514,0.1005844101,0.2101529688,0.6478099227,0.1063755378,-0.1575641781,-0.0989586189,0.098999843,0.142893672,-0.0925440639,0.30904001,-0.1441699266,-0.4113992155,0.016843766,-0.2790069878,0.1049873084,-0.1868219376,-0.0090052122,0.2628933489,0.0677142739,-0.1049920991,-0.0633964539,0.097431846,-0.0062688752,0.3627237678,0.0684464723,-0.0100124199,-0.2978586853,-0.1031890363,-0.3477042317,0.024746025,-0.0675138831,0.1919189394,-0.1367273927,0.1563542038,-0.1571874619,0.1180881932,0.8750144839,-0.0463553071,-0.6675370336,-0.234750554,0.1103048846,-0.30512321,0.0943309441,-0.0854472965,-0.0774557889,0.1738235652,0.3756970763,-0.2528294921,-0.108944878,0.3284611106,0.1395583004,-0.0272754021,-0.3440355361,-0.1211376861,-0.1594705582,-0.1654295176,-0.1205394417,-0.0446190536,-0.2055437416,-0.0155847855,-0.3963989317,-0.0924196541,0.0029271371,0.3095357716,0.0837070346,0.1785472482,-0.0767477751,0.2740626633,0.3710874021,0.0349616148,0.1678017974,0.1558083147,-0.1300619692,-0.281493336,-0.1530754715,0.0663612485,0.4816920161,0.027137205,-0.0432159267,-0.2878828645,0.0092745721,-0.0068296827,-0.1015968844,-0.2731994987,0.3206413686,0.2860986292,-0.5582033396,-0.39208287,0.4232932627,0.3247115612,-0.0306646843,0.3255999386,0.4408338666,-0.1179009676,-0.1323714852,-0.0730321258,0.8423376083,-0.3395858705,0.3163875639,0.0185803212,-0.3069078028,0.5018220544,-0.398110956,-0.0546696894,0.0135429399,-0.0624521859,-0.260210067,-0.1515945494,0.3064657748,0.3013101816,-0.4307981431,0.4004119337,-0.1082953289,0.0276416037,0.0050625997,0.0611565821,0.1772366166,-0.338183105,-0.1260104775,0.0943790749,0.1905067414,0.4033849835,-0.3439296782,0.0088120317,-0.2259817868,-0.1661720276,-0.2146671563,0.0166941769,0.1666430831,0.0599226542,0.606978476,-0.0218883511,0.0804415345,0.4288466573,0.3316944838,0.1064805463,-0.3243187666,0.1651843339,-0.3943386674,-0.1713136733,0.0522910468,0.2141287029,-0.0399192981,-0.0722302943,-0.3412376344,0.0128685711,-0.0528601557,-0.2699426413,0.0907764286,0.0037692965,0.0046351673,-0.1425936073,-0.0599488243,0.4706614017,0.2478717864,0.0292851944,0.0832810625,0.2332385182,-0.1398199648,0.3994798064,0.1445847899,0.0665718988,-0.2973617613,0.1612494141,0.0669586584,-0.3972334266,0.0494261943,-0.1360638142,-0.3516342938,-0.0984837487,-0.3309950829,-0.2119076401,-0.0277386717,-0.2684343159,0.2341371477,-0.2504578233,-0.2659105062,-0.0678391308,0.1507382244,-0.2360898554,-0.2879275084,-0.1930928379,-0.3755355775,0.2025972754,-0.0856982768,0.2909670472,0.0856310353,0.1880367398,0.1853874773,0.0323667973,-0.1202351302,-0.2265191525,0.080720365,-0.1489807516,0.5370168686,0.2832778096,0.2279922962,-0.0295395628,-0.0290308576,-0.0018280161,0.0259889662,0.0155129461,0.065937914,0.1773602962,0.0920890123,-0.2864089906,-0.0723772049,0.0262418315,-0.0922721103,-0.0568600893,0.075903073,0.0475660414,0.1142017245,0.0846892446,0.1213959381,0.0199013203,-0.1808382422,0.3134968281,0.3787615895,0.0336062163,0.305275172,0.1969464719,-0.0262746196,0.0694924295,-0.0995683223,0.3839229941,0.2635221779,0.0560924672,0.3092076182,-0.1706852764,0.2403122485,0.3471409976,0.2176717669,0.0587536767,0.0519193374,0.1733650863,0.4526016414,0.0497723818,-0.1637506485,-0.3709401488,0.2373080701,0.2273483276,-0.1312070936,0.5030232668,-0.0248338301,0.1548070461,0.6223063469,0.1297173351,0.0930040479,0.0171456262,0.4952861071,0.1694941074,0.2058981359,0.1259293109,0.2770324647,0.3417434096,0.0705008507,0.256573081,-0.4401835501,0.0376261622,-0.3780774176,0.1473159939,0.1542909294,-0.1495671123,-0.1775618941,-0.2020445764,0.2412399948,0.0009571525,-0.1416774094,0.3818827868,-0.2030324936,-0.1247029901,-0.4468919635,-0.0251392964,-0.1960000396,-0.1125228554,0.5519209504,-0.1703492552,0.2792989314,-0.1201909259,0.2959105372,-0.4879186451,0.3072283566,-0.1277850121,0.1697865278,-0.1338504702,0.3452062309,0.1726908386,-0.185292393,0.278003484,0.1347157806,-0.063126564,-0.0564114302,-0.0301760174,0.1610761881,0.1770987064,0.0372644402,0.373513639,0.1492981911,-0.1438915581,0.0124737676,-0.0639894083,0.291896373,-0.018491881,0.1231768578,-0.118522428,0.1214974001,-0.0378730036,-0.0496009141,-0.0313288458,0.3292010725,-0.3697253466,0.3260979056,-0.3267105222,-0.2223031223,-0.2761810422,0.0653223172,-0.2794983685,0.1510552019,0.380246073,-0.2416643947,-0.1619774252,-0.0023294645,-0.0051036477,0.0297172125,0.3195904791,0.0679202676,0.4906852841,-0.0486300588,-0.2626934052,-0.3486282229,-0.2323553711,-0.1521580219,-0.0079041058,-0.1133720055,-0.1833284646,-0.01112659,0.0941321254,0.0599643178,-0.1079914197,0.0411909446,-0.4305486381,-0.1380660832,-0.0885200575,-0.260471195,0.3724285066,0.1814865619,-0.1211744919,-0.0957889557,0.1225397587,-0.2033199966,-0.0326380245,0.3436830342,-0.2231088132,-0.1587463617,0.2783975005,-0.2101497203,0.5602531433,0.0646077022,-0.0580364913,-0.0576319806,-0.3887792826,-0.059993621,-0.0030556337,-0.0392137282,-0.0962668359,-0.614939034,-0.1233389974,-0.1512817144,0.0165466499,0.0731476247,-0.0427972637,-0.1234554499,0.0107002947,-0.2694033384,-0.004770352,0.235677585,0.1630309224,-0.1814401448,0.0389194675,0.0521429814,0.2393892407,0.495574832,0.0200425368,-0.0543314628,-0.0079396991,-0.3150820136,0.0305594821,-0.4145951867,-0.7024553418,0.2409145385,0.1297398955,0.1162867844,0.0351375081,0.3079443276,0.1236673519,-0.1610228866,-0.1047233865,0.5230149031,-0.1716244966,-0.0175624564,0.1288028657,0.0628326535]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1713","title":"Installation using conda","comments":"`datasets` has been added to the huggingface channel thanks to @LysandreJik :)\r\nIt depends on conda-forge though\r\n\r\n```\r\nconda install -c huggingface -c conda-forge datasets\r\n```","body":"Will a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.","comment_length":26,"text":"Installation using conda\nWill a conda package for installing datasets be added to the huggingface conda channel? I have installed transformers using conda and would like to use the datasets library to use some of the scripts in the transformers\/examples folder but am unable to do so at the moment as datasets can only be installed using pip and using pip in a conda environment is generally a bad idea in my experience.\n`datasets` has been added to the huggingface channel thanks to @LysandreJik :)\r\nIt depends on conda-forge though\r\n\r\n```\r\nconda install -c huggingface -c conda-forge datasets\r\n```","embeddings":[-0.0296039023,-0.0399743132,-0.1089975238,0.1860422194,0.2394067347,-0.0445190966,0.1131713986,-0.1059008911,-0.3314240277,-0.302870661,-0.2650924623,0.1040904149,-0.1357483417,0.5306632519,0.3688568771,-0.0758271888,0.2095017582,0.2115162164,-0.6172249317,-0.0298800245,0.1432552785,0.2838039994,-0.1970897764,-0.1503042281,-0.3642676175,0.056296628,-0.0738267377,-0.1842348874,0.0266973823,-0.2115781009,0.5577281117,0.1635947675,0.2185151726,0.520349741,-0.0001208874,-0.0354604088,0.213015303,-0.0774974898,-0.3361149728,-0.0101666823,-0.389367491,-0.2792269886,-0.0301751755,0.0580167435,0.0097176535,0.0661545321,0.0211060625,-0.1383633763,0.3403495848,0.058080785,0.1515555829,0.3557485044,0.2534594536,-0.0241507553,-0.1675552875,0.4902023077,-0.2422238588,-0.0713035986,0.0869594291,0.0984996483,0.5578453541,0.0065273545,-0.0347136557,-0.100368768,0.2996935248,0.1745139956,-0.3399006724,-0.2819991112,-0.0511117317,0.2685140967,0.7587833405,-0.3713587523,-0.6890044808,-0.3122138679,-0.0279159416,0.1089591235,-0.1525970846,-0.1076399982,-0.1356344223,0.2149150968,-0.290166676,-0.6198328733,-0.1621965766,0.1124493405,-0.0005172484,0.3889229596,-0.1620222777,0.0777498335,0.2113397717,-0.0557194799,0.171326831,-0.0997946039,-0.0325682126,0.1761730015,-0.1066201851,-0.170584619,-0.2545810044,0.4082921445,0.2948686481,0.2181316167,-0.4194543958,-0.135757193,-0.2369313389,-0.183127597,0.1929246485,0.0634049773,0.1054866239,0.4957329333,0.0893854424,-0.1445681304,0.43280828,-0.0399534181,-0.0460183807,0.0161799416,-0.2737437487,-0.3993260562,-0.0891322643,-0.2870233953,0.0064034625,-0.2886557579,0.4500457346,-0.1929966509,0.1821888983,0.2403699905,0.0320765451,0.1322605014,-0.0038805385,0.2970759571,0.3454447389,0.0610121265,-0.0838139877,-0.0881427154,-0.0199419558,0.1140609384,0.2309093475,-0.3153795004,0.1855082214,-0.2256006896,0.4308143556,0.3417765498,-0.0638408288,0.2185708135,-0.0955144539,0.4455742538,-0.3512631357,0.2501911819,-0.1407224238,0.1761865169,-0.2922187448,-0.0109037859,0.0705821589,-0.004938202,-0.0855336487,-0.0491808765,-0.1708822995,-0.2739294171,-0.2653611898,-0.1015576869,-0.2199379206,-0.1350846589,0.0281362738,0.2281555235,0.1965715736,-0.0241776705,0.2415091395,0.4308135211,-0.3204354048,-0.4266450703,0.0783582628,-0.4344707727,-0.1541642249,-0.032543283,-0.0673796013,-0.1104503348,-0.1799430251,-0.366201818,0.3999581635,-0.5274589062,-0.3011988699,0.0217517093,0.1033757403,0.2396851778,-0.113210164,0.2434560806,0.1589307636,0.0802019313,0.0272397157,0.2447111458,0.0145050157,-0.3280427456,0.2043245286,-0.2349990308,0.106342718,0.2239687443,0.1952615529,0.3022698164,-0.1420414597,-0.0918333381,0.15574871,0.138288483,0.1555514336,0.267722249,0.3016324341,0.2884352803,0.0832171291,-0.3093565702,-0.3711173534,0.1130677611,0.1507466882,0.0016270534,-0.1291172206,-0.2351766676,-0.0462832563,0.1461788416,0.0964464843,-0.1716506332,-0.0775404125,-0.0991508588,0.4647377729,-0.2203253806,-0.3421970606,0.9912434816,0.4688918889,0.3277896345,-0.0909310356,0.2403007448,-0.047502812,0.0740584955,-0.0532221198,0.2777820528,-0.0386894457,-0.2761130631,0.2325638831,0.2086801678,-0.1758487821,-0.0812153965,0.277384609,0.1935042143,0.1843614578,-0.0035291989,0.1218149364,-0.2264946103,-0.0739820972,0.0571250021,0.0318089128,-0.2204486132,-0.1256756932,0.0195601881,0.0984865129,0.311791122,0.3137337565,0.2109018266,-0.2605006397,-0.2747134566,-0.0778359175,-0.0633036643,0.2634271085,-0.0131231667,-0.0589004792,-0.2266942859,-0.1191508621,-0.1861940324,-0.0210871138,0.3124918044,-0.2640961111,0.4855315685,0.2817257345,0.2035993785,0.3454295695,0.1963165849,0.0788591057,-0.1851339936,-0.0295420066,0.0008035467,0.0157449134,0.2013032436,0.0155260162,-0.3022401035,-0.3382844627,0.0539068803,0.0779838264,-0.2943549752,-0.4199885726,-0.0840438083,-0.1885014921,-0.0586015619,-0.2047686279,-0.0325729325,-0.2585756481,-0.3729180396,-0.2393331975,0.0493080653,-0.1605867594,-0.2397894412,0.2936691642,0.1699875742,-0.2903734148,0.3241269588,-0.304095149,-0.4081711769,-0.1580365151,0.2665275037,-0.0613298677,0.0668960959,0.0522693023,-0.0840248019,0.1296420544,-0.1698678583,0.4550246,-0.169860363,-0.4443704486,0.102739431,-0.1414861381,0.3694041967,0.0276789796,0.1013228446,0.0970082283,-0.0839463845,-0.0104794595,-0.0044471305,0.1619687676,-0.2443733662,0.0076658605,-0.290904969,-0.1808828712,-0.0214433428,-0.5090830922,-0.1175971702,0.5027697682,0.1387522668,-0.2172674537,-0.0292484071,0.0144085074,0.1492461264,-0.1999724209,0.2633713484,0.0334010907,-0.0873203203,0.5326375365,-0.105519928,-0.1174501628,-0.0264560059,-0.1532169133,0.4213302135,0.0937683359,-0.3301803768,0.1915220618,-0.0114307301,0.0560151376,0.1847115308,0.1781056523,0.5187995434,0.0392140485,0.2465070337,-0.0237926338,-0.3398368657,-0.1718513966,-0.160054341,0.2083532512,-0.0710422099,0.0630536154,-0.271345973,0.2838387489,0.2142466307,-0.1129646823,0.0421449058,0.1831898391,0.6972621083,0.1781626195,-0.281950295,0.0130610764,0.1370985657,0.1438920796,-0.0548726842,0.3215620816,-0.0295275096,-0.4412065148,0.0963349566,-0.2744323611,0.0528127886,-0.2521033883,-0.1082898304,0.1960715353,0.0879887789,-0.0457800291,-0.0036043215,0.0560310297,-0.0137688052,0.3663768172,0.0802091211,-0.0026431086,-0.329909265,-0.0756879747,-0.4199191034,0.0691907629,-0.057795193,0.2198608667,-0.1103693098,0.1133418903,-0.1646767855,0.1027797684,0.8495227695,0.003357132,-0.6551269889,-0.1627009362,0.1493963003,-0.4111643434,0.0671083778,-0.0011655081,-0.0602128804,0.211553216,0.4190407693,-0.2209342867,-0.1101189777,0.2696772218,0.122188732,0.0397893116,-0.2275602967,-0.175015986,-0.140286237,-0.1596504599,-0.1898905486,-0.0425572805,-0.1642307937,-0.0389067829,-0.3055097759,-0.0876214877,-0.0432187095,0.3834273517,0.0878402442,0.1816128343,-0.1050078273,0.3220688701,0.4807137549,-0.0310695712,0.1652893275,0.2242659479,-0.1692228466,-0.3239909112,-0.1658754945,0.1007019505,0.4438791275,-0.0365244262,-0.0940727219,-0.2365945131,0.0343465954,-0.0287235864,-0.1314878315,-0.2029210478,0.3194976151,0.2787730396,-0.5626211762,-0.4346798658,0.3367549777,0.2931680381,-0.0557135195,0.384572655,0.3917360008,-0.1271844953,-0.1787935644,-0.05765545,0.737872839,-0.2793720067,0.3199300766,-0.0150676407,-0.3507258892,0.5631979108,-0.3242426813,-0.0928973183,-0.0035475728,-0.0991473049,-0.2283880264,-0.1782834232,0.345828563,0.3576596975,-0.4025153816,0.3377335072,-0.0428243093,0.0883684158,0.0270023439,0.0059714825,0.1837412715,-0.2762798667,-0.0068649864,0.1197315454,0.1188795492,0.4403016269,-0.2937505543,0.0785686001,-0.2915751934,-0.1945421696,-0.3028696477,0.0219620876,0.1288807839,0.0314290151,0.7062219381,-0.007783473,0.1350425184,0.4751844704,0.4399010241,0.1391885579,-0.375146091,0.1966901273,-0.3600428998,-0.1497925967,0.05894899,0.1702253968,-0.0635526404,0.0184888039,-0.3094584644,0.0296926517,-0.0227294192,-0.2771973312,0.0550500378,0.0595797449,0.0061594988,-0.0839721784,-0.079402633,0.4485958517,0.278192997,0.0187076349,0.0748483986,0.2459592372,-0.0943628773,0.3475068212,0.0646574497,0.0322736315,-0.2667603195,0.2942609489,-0.0898806751,-0.3779761791,0.0836177096,-0.1084476784,-0.3593873978,-0.0669938624,-0.2827809751,-0.1874517798,-0.0743290633,-0.3029493392,0.2483583987,-0.2350662947,-0.1800590605,-0.1511821747,0.0805201009,-0.2166673243,-0.2099399269,-0.1750634164,-0.4139847159,0.210433051,-0.1120136455,0.2709073126,0.0844535083,0.1268007159,0.19094944,0.0739690065,-0.1119386777,-0.219850421,0.1207056642,-0.2033244073,0.5632286072,0.2070676833,0.2890428901,-0.1198424324,-0.0163390301,0.0201455355,0.0519647896,0.0155745735,-0.0080818152,0.1726922542,0.1215387806,-0.3104853928,-0.0491249636,-0.0010917662,-0.0993835628,-0.1420761049,0.1140270755,0.1778161228,0.1341089755,0.0922243148,0.0815035924,-0.0318919234,-0.1864519417,0.2320295125,0.3301813602,0.0387175232,0.3300394118,0.2174202204,-0.0276840944,0.0536675639,-0.1436593533,0.3325992823,0.2357628495,0.0720478818,0.3016224802,-0.1189977601,0.2152449787,0.3758436739,0.2486320883,0.0218294114,0.0531374216,0.1425356716,0.4835516214,0.0487559661,-0.1327562332,-0.3614990413,0.1282333583,0.2527441978,-0.1329939663,0.4360303283,-0.0652421638,0.1486799419,0.5753985047,0.1046218351,0.139727205,0.0066676973,0.5112197995,0.1781446934,0.2238909453,0.0852620378,0.3125483096,0.2670511603,0.1331119984,0.2372649461,-0.5011426806,0.0169533603,-0.3879602551,0.1621542871,0.183434546,-0.2091598511,-0.2089090645,-0.1860565692,0.1895498633,0.0291958712,-0.1704071313,0.4248936176,-0.2297881246,-0.0788780525,-0.3636819422,-0.0146221872,-0.1618776023,-0.1984182447,0.6090636253,-0.1854047477,0.2382386327,-0.1670907289,0.2437241673,-0.5002306104,0.3145496547,-0.138951689,0.1449721307,-0.2091700137,0.2995737195,0.146502167,-0.1955224574,0.2323250622,0.0675776228,-0.0633548945,-0.0797553584,0.0453097895,0.1263641715,0.2239430249,0.0417084843,0.405430764,0.1305998266,-0.1768981665,0.0312569179,-0.1092964485,0.3131628633,-0.0671945438,0.0879467502,-0.0878504962,0.1241916046,-0.046728719,-0.0821821317,-0.0391657948,0.3811219931,-0.4036622047,0.3750119805,-0.3935606778,-0.2029144466,-0.2238678485,0.1519770771,-0.2924770117,0.1369490325,0.3565064073,-0.2550810874,-0.1790159792,-0.0788415372,-0.0024690954,0.1163853928,0.3421878815,0.0514913052,0.4953832626,-0.007196066,-0.2586133182,-0.4099025726,-0.2329143137,-0.1203396544,0.0546787605,-0.1133131087,-0.2007817924,0.0185477342,0.0036698089,0.0510914959,-0.1076933146,0.0608738437,-0.4340240657,-0.069437325,-0.0979375839,-0.3268860579,0.3197919726,0.2493607253,-0.1238715872,-0.0955599323,0.1234005243,-0.203843534,-0.090766795,0.2887510359,-0.1459331959,-0.1377938837,0.3056755662,-0.1873536706,0.5309541225,0.0637580007,-0.0290032439,-0.1358949542,-0.3469368219,-0.0228982642,0.055427976,-0.1585592031,-0.0567212477,-0.6374249458,-0.1820965111,-0.1938541234,0.0409127884,0.075297229,-0.0810215399,-0.1524124295,0.0300287288,-0.311401844,0.0213716011,0.2045868039,0.1756633818,-0.1656179726,0.055998221,0.0426124483,0.1859712154,0.5106599331,-0.0949865207,0.0254287161,-0.0068091112,-0.2424941063,0.0563279353,-0.4501660466,-0.6654915214,0.2103219777,0.2145251334,0.1554281414,0.0828936547,0.2265338451,0.1486229897,-0.1002158225,-0.1532569826,0.5758805275,-0.1383210272,0.0286220964,0.2653759718,0.1104924679]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1710","title":"IsADirectoryError when trying to download C4","comments":"I haven't tested C4 on my side so there so there may be a few bugs in the code\/adjustments to make.\r\nHere it looks like in c4.py, line 190 one of the `files_to_download` is `'\/'` which is invalid.\r\nValid files are paths to local files or URLs to remote files.","body":"**TLDR**:\r\n\r\nI fail to download C4 and see a stacktrace originating in `IsADirectoryError` as an explanation for failure.\r\n\r\nHow can the problem be fixed? \r\n\r\n**VERBOSE**:\r\n\r\nI use Python version 3.7 and have the following dependencies listed in my project:\r\n\r\n```\r\ndatasets==1.2.0\r\napache-beam==2.26.0\r\n```\r\n\r\nWhen running the following code, where `\/data\/huggingface\/unpacked\/` contains a single unzipped `wet.paths` file manually downloaded as per the instructions for C4:\r\n\r\n```\r\nfrom datasets import load_dataset\r\n\r\nload_dataset(\"c4\", \"en\", data_dir=\"\/data\/huggingface\/unpacked\", beam_runner='DirectRunner')\r\n```\r\n\r\nI get the following stacktrace:\r\n\r\n```\r\n\/Users\/fredriko\/venv\/misc\/bin\/python \/Users\/fredriko\/source\/misc\/main.py\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/Users\/fredriko\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0\/8304cf264cc42bdebcb13fca4b9cb36368a96f557d36f9dc969bebbe2568b283...\r\nTraceback (most recent call last):\r\n File \"\/Users\/fredriko\/source\/misc\/main.py\", line 3, in <module>\r\n load_dataset(\"c4\", \"en\", data_dir=\"\/data\/huggingface\/unpacked\", beam_runner='DirectRunner')\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1066, in _download_and_prepare\r\n pipeline=pipeline,\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 582, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/Users\/fredriko\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/c4\/8304cf264cc42bdebcb13fca4b9cb36368a96f557d36f9dc969bebbe2568b283\/c4.py\", line 190, in _split_generators\r\n file_paths = dl_manager.download_and_extract(files_to_download)\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 258, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 189, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 117, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[str(url)] = get_size_checksum_dict(path)\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 80, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/'\r\n\r\nProcess finished with exit code 1\r\n```","comment_length":50,"text":"IsADirectoryError when trying to download C4\n**TLDR**:\r\n\r\nI fail to download C4 and see a stacktrace originating in `IsADirectoryError` as an explanation for failure.\r\n\r\nHow can the problem be fixed? \r\n\r\n**VERBOSE**:\r\n\r\nI use Python version 3.7 and have the following dependencies listed in my project:\r\n\r\n```\r\ndatasets==1.2.0\r\napache-beam==2.26.0\r\n```\r\n\r\nWhen running the following code, where `\/data\/huggingface\/unpacked\/` contains a single unzipped `wet.paths` file manually downloaded as per the instructions for C4:\r\n\r\n```\r\nfrom datasets import load_dataset\r\n\r\nload_dataset(\"c4\", \"en\", data_dir=\"\/data\/huggingface\/unpacked\", beam_runner='DirectRunner')\r\n```\r\n\r\nI get the following stacktrace:\r\n\r\n```\r\n\/Users\/fredriko\/venv\/misc\/bin\/python \/Users\/fredriko\/source\/misc\/main.py\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/Users\/fredriko\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0\/8304cf264cc42bdebcb13fca4b9cb36368a96f557d36f9dc969bebbe2568b283...\r\nTraceback (most recent call last):\r\n File \"\/Users\/fredriko\/source\/misc\/main.py\", line 3, in <module>\r\n load_dataset(\"c4\", \"en\", data_dir=\"\/data\/huggingface\/unpacked\", beam_runner='DirectRunner')\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 612, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 527, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1066, in _download_and_prepare\r\n pipeline=pipeline,\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 582, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/Users\/fredriko\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/c4\/8304cf264cc42bdebcb13fca4b9cb36368a96f557d36f9dc969bebbe2568b283\/c4.py\", line 190, in _split_generators\r\n file_paths = dl_manager.download_and_extract(files_to_download)\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 258, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 189, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 117, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[str(url)] = get_size_checksum_dict(path)\r\n File \"\/Users\/fredriko\/venv\/misc\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 80, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/'\r\n\r\nProcess finished with exit code 1\r\n```\nI haven't tested C4 on my side so there so there may be a few bugs in the code\/adjustments to make.\r\nHere it looks like in c4.py, line 190 one of the `files_to_download` is `'\/'` which is invalid.\r\nValid files are paths to local files or URLs to remote files.","embeddings":[-0.2110013068,-0.0166831054,0.002013599,0.2388121039,0.3003631532,0.0115945125,0.0879322886,0.2422638535,-0.1057231128,0.0968610942,-0.1910670251,-0.0968224183,-0.3083791435,-0.0378795564,0.0463685207,-0.0706733167,-0.1771388352,0.241326645,-0.4923979044,0.1420897543,-0.0148674138,0.4154529572,-0.1527776271,0.0248496011,-0.094913356,-0.105762437,-0.0448619463,0.2398068756,-0.0043044961,-0.431681782,0.6017952561,0.2295489162,0.1245357692,0.4874401093,-0.0001212964,0.2981967628,0.317119956,-0.1303761899,-0.3978928626,-0.2918590605,-0.0663455352,-0.0388181917,0.3356330693,-0.1701166928,0.1311804503,0.035082899,0.0841413587,-0.3720006645,0.2927715778,0.2490392476,0.1561473906,0.407823503,0.2956752777,0.0461126007,0.1961436719,-0.0739455968,0.0258721635,0.3925033808,-0.3268428147,0.2190594375,-0.1065383852,0.1551535875,0.061704576,0.0032203929,0.5413600206,-0.2076385319,0.1880460829,-0.0309579633,0.1410353035,0.0395150706,0.6169292331,-0.3673050702,-0.2273660153,0.0226472281,0.0485328585,-0.0480972268,0.5799092054,0.0286890045,-0.1914091855,0.1599879265,-0.3129753172,0.0555075221,-0.2554164231,0.180970639,-0.2790956795,-0.2163836211,-0.121321924,0.1668666899,0.3539424539,0.1593830884,0.0721893609,-0.1522883922,0.0950876772,0.1023896933,-0.3253191411,0.1324953139,-0.079841733,0.4449476302,-0.0588747934,-0.135213688,-0.0323127359,-0.0894959718,-0.1571263075,0.0771519467,-0.0998350158,0.5191348791,-0.1745387763,-0.0579019561,0.1916341782,0.1299021542,0.0251296032,0.0934778899,-0.0617643595,-0.2957741618,-0.3996480107,0.1983498484,-0.0308842752,-0.1465162635,-0.5061504245,-0.051147934,-0.6028206944,0.0617649481,-0.0486882254,0.4410617352,-0.2156828791,0.2982066274,0.427164048,0.1172027513,-0.0777522549,-0.1600181907,-0.0464469083,0.1821557879,-0.3387562037,-0.2175408304,0.146545738,-0.3067020178,0.0565164387,-0.057250414,0.1242894232,0.1198472083,-0.1313167661,0.083299838,-0.3962429762,0.3249585629,-0.0569178984,0.3533375263,0.141952157,-0.2476044744,-0.0419186801,-0.2503644526,-0.0288922265,-0.1689238399,-0.0059046121,0.0625003278,-0.2621732056,-0.0186396968,-0.0239851754,-0.3702903986,-0.0880559757,-0.0250878558,-0.3069533706,-0.3325378299,0.0403667279,-0.387948066,0.2867855728,0.7887271047,-0.37308231,-0.062373139,-0.3939866424,-0.3207274377,-0.088637799,0.1597679108,0.0083917333,0.4683425426,-0.3715884387,-0.0817715451,0.0349983349,-0.430870533,-0.3826598227,0.4031050205,-0.2026176751,0.181424588,0.2387690842,-0.1593855619,-0.0379303358,-0.0491600074,0.0392383374,0.073790893,0.1009896919,-0.2146940827,-0.3035178781,-0.0859721005,0.0347923078,0.1836927235,0.1889918745,0.1237996444,0.2890424132,0.0238364823,0.4252669811,-0.1544864923,0.1230125576,0.435883224,0.3003919125,0.0804158077,-0.2277423143,0.2824690938,-0.7021694779,0.1369891167,-0.0688533261,-0.0525572635,-0.188316077,0.0565886833,-0.3794920743,0.1098476052,-0.2681518793,-0.0508044325,0.0427405573,0.2153688669,0.0251013692,0.2253341377,0.0545547009,0.9237930775,-0.0291264597,0.2412460148,-0.1177422404,0.2889007926,-0.1239601597,-0.0006952902,0.2360188663,0.059098497,0.2810580134,-0.2603555918,-0.0164311752,0.4496192038,-0.3032152057,0.2257034332,-0.2356710434,0.2374099791,0.1409540772,-0.4850714803,0.2437880486,0.1826347411,0.0155507112,0.0794943646,0.3024570942,0.0373970866,-0.0111021306,0.294472754,0.261115849,0.2077235281,0.1051314324,-0.1697059274,0.186281085,0.0056127477,0.4510717988,-0.085326232,0.6482319832,-0.2358091921,-0.0486219153,0.0710991621,-0.0294509474,-0.0853036344,-0.2401970625,0.1699893326,-0.1940018684,0.1319160014,0.3667606413,0.2274281085,0.3210941851,0.1386129111,-0.0640742704,0.177634269,-0.0354291461,-0.0930108204,0.1974637359,0.0143876793,-0.1937520951,0.1127503589,-0.0337735526,0.071177125,-0.2204743177,-0.3273619413,-0.1083015278,0.0916903019,-0.5798996091,0.0877991691,-0.3344514966,-0.1501529217,-0.4275587499,0.1584721953,-0.1326646209,-0.3841993511,-0.2535524666,0.3956749141,-0.1553426683,0.0745387822,-0.0524655245,-0.3114846051,-0.0429090299,-0.3456931412,-0.2296126336,0.1166532636,-0.3646844923,-0.0188131277,0.2823852599,-0.163425982,0.1477771252,-0.3827365935,0.1993676722,-0.389424324,-0.193093136,0.3145964742,-0.1097067967,-0.1174798235,0.2981440127,0.2071418166,-0.1809550226,-0.3646033704,0.0009787113,0.0255349614,-0.1015324593,0.1881582737,0.0071540694,-0.0530275106,-0.2206927985,-0.323942095,-0.3468903303,-0.4373090863,0.2280209959,-0.2074750364,-0.0289415903,0.3682849109,0.1105989814,0.2621778846,-0.1250937283,0.0355542637,0.2126612961,0.0951536819,0.1580765694,0.0309461709,-0.29599002,0.1105617136,-0.1883838326,0.1125982702,0.4232776761,-0.6604568958,0.3621883392,-0.1609885246,0.2337641418,-0.1173595935,0.2308192253,0.3478416502,0.0552455708,0.0476462953,-0.1626119465,0.0329933874,0.0554185025,-0.2411595434,0.352789551,0.0200572927,0.1407480538,0.1069205329,0.7421910763,0.1976411641,-0.3773144782,0.5262420177,0.2205745727,0.5429071188,-0.0232981835,-0.1862657517,-0.0334116183,0.0498204976,0.085694328,0.1664590091,-0.0440137088,-0.0601375215,-0.2190629095,0.150667429,-0.3889398873,-0.2893818319,-0.1051248536,-0.0154780159,0.3131743073,0.0800834,-0.0028583203,0.2476548851,-0.1807789654,0.1971668154,0.4520085454,-0.0893389881,0.072692968,0.3228593469,-0.2642275989,-0.3816050589,0.4794580042,0.1105572358,0.072109364,0.0569231547,-0.0117975231,-0.0023615917,0.0984232798,0.5568631291,0.1064317524,0.1237618476,0.1287567019,-0.1111195534,-0.5451614261,-0.0821274742,-0.1311380416,-0.0492538065,0.1802386343,0.0957430154,-0.5184886456,-0.1800267696,0.0625547916,0.3143153489,-0.0285279956,-0.1551057398,-0.0741231441,-0.3765134811,-0.0848246962,-0.1324586123,-0.1140458733,0.1428080201,0.1670542955,0.1341439784,0.0762933642,-0.3344878256,0.15706411,-0.1501314342,0.2978230715,-0.0240870491,0.2143878639,0.226332128,0.33322227,0.1846717298,0.2643730938,-0.1563521177,0.2481510639,0.1476454139,0.0853166208,0.0780667886,0.2647037506,-0.1514919847,-0.1244874671,-0.1531338692,0.1062278599,0.0026878128,0.4056752026,0.1180930659,-0.2031940371,-0.169320181,-0.0508999974,0.2750612497,-0.3377221227,-0.1445810348,0.1453642696,0.4561205506,-0.1714355201,-0.0592920966,-0.0010156495,0.9429741502,-0.0749188513,0.0996004641,0.4034031928,-0.2098904848,-0.0325758457,-0.6176768541,-0.0989142805,-0.1444785446,0.2607507706,-0.0039601871,-0.1388367712,-0.0205169637,0.0532414168,-0.2829584777,0.3608724177,0.1214129627,0.0540410429,0.2724373043,0.0729144663,-0.0614798106,-0.0484999195,-0.273412019,0.0939456373,0.1057456061,0.4314227104,-0.3022381663,-0.1629446894,0.0097238123,-0.3524848521,-0.2409736961,0.1356851012,-0.4610415995,0.1396102458,-0.0170984026,-0.1849695444,0.2507363558,0.1617297083,0.1663229018,0.2726708055,-0.1978047192,0.0739849433,-0.3188921213,-0.1478724629,0.1963866502,0.0669668913,0.1357768178,-0.1648753583,0.0687193125,0.015435854,-0.2353783846,-0.2115257233,0.2892059088,-0.049784556,-0.0397084244,-0.1274456233,0.2722110748,0.1378647238,0.1189762279,-0.0447080843,0.0513011739,-0.0827126056,-0.0593067072,-0.0037943805,-0.1242143884,-0.2437945902,0.0163898114,0.5655068159,-0.0692046136,-0.1749799848,0.2817901075,0.1025268286,-0.3069685102,-0.178380996,-0.0649238378,0.0574099682,-0.5497649908,0.091249615,-0.24430269,0.0484156609,-0.183619231,0.2062886953,0.2045208812,0.0974803045,0.0302485544,-0.18795228,-0.3507925272,0.1664500833,-0.2784709036,0.1419964433,-0.0813893601,0.3665918708,0.1949106157,-0.1860466897,-0.2032552212,-0.0742215514,-0.1585238874,-0.0874830708,0.441133976,0.0149970464,0.571311295,0.1639360636,0.1571798027,0.1715365797,-0.3253118098,-0.0492531359,-0.0736198351,0.1873324066,0.0940412357,0.0791304111,0.3753319979,-0.007754737,-0.1374605149,0.047367394,0.2224427611,0.1146265566,-0.199139595,0.0036505118,-0.074226141,0.3938066959,-0.3645932376,0.0275477394,0.0569291562,0.2376100272,0.0823052377,-0.0054570553,-0.1402682215,-0.1004243717,-0.0089008333,0.0277920682,0.0635296106,-0.0590940528,0.1621925235,-0.1362021863,-0.0809091181,-0.226388827,0.1827614307,0.021526454,-0.1295414567,-0.0724511668,0.3625157475,0.0996334627,-0.0510757677,0.3101758659,0.135667786,-0.0271303337,-0.0272524972,0.0672805533,-0.0906958207,-0.2450662553,0.4972268045,0.2003454268,0.3116389215,0.2572151423,-0.1320090145,0.2532206476,-0.0358282104,0.0929638967,0.1072631553,-0.150806129,0.2952391505,0.4780038595,-0.6502730846,-0.0195513591,-0.4387215674,-0.0871149525,0.1716143638,-0.3194366097,0.021031674,0.2744723856,0.248983264,0.1790737212,-0.0345382988,0.5936598182,-0.1786314547,0.0288169216,-0.380507201,0.1556317359,-0.0463432595,0.1890696734,0.0329897627,-0.3242682815,-0.3419363797,0.1487344056,-0.0666352659,-0.2571842372,0.5585412383,0.3594275415,-0.0097634476,-0.4167915881,-0.0047229752,0.0774707347,0.0778319463,-0.145245716,0.1495110393,0.2501826286,0.0041692969,0.0649393648,0.1351977289,0.3599378765,0.0488775223,-0.0236336868,-0.1653164476,0.053755872,0.017595239,-0.110109888,0.2040827721,-0.0455690771,0.4231415987,0.3586812615,0.0661218166,-0.2273059487,-0.0137108145,-0.0868732184,0.2572360039,-0.2923269868,0.2525443733,-0.3696068525,0.0226278342,-0.0929619744,0.0115318075,-0.56722188,0.1695986837,0.14680852,-0.0183600541,-0.0049482007,-0.0286021624,0.0467447937,-0.1530151665,0.4417681694,0.4194592834,0.2856597602,-0.3929919899,-0.3780041933,-0.5556571484,0.2937075496,-0.1813764572,0.1456129253,-0.2025690377,0.0601713061,0.1234764382,0.1100146249,0.3431590796,0.1959430873,0.1913858801,-0.0996114016,-0.1689574271,-0.109852083,0.2226736248,-0.2226105332,0.0552140698,-0.4527213275,-0.029664183,-0.0800073743,0.0721194968,-0.0483369753,0.0872060135,0.1102650762,-0.4262734652,0.4338591397,0.0266978834,0.5156461,0.0675260648,-0.4733674228,-0.2356517911,-0.1582622379,-0.0280754473,-0.2784138322,-0.0248006601,0.5825551748,-0.1524595171,-0.071348384,-0.3878982365,0.5314272046,0.372400403,0.231578216,-0.2952343822,0.2969204485,-0.5836755037,0.2754730582,0.0480794087,0.2640866041,0.1151255518,0.1423322856,-0.640568912,-0.5447107553,0.5251246095,-0.6823359728,-0.1140259132,0.2898886502,0.2478706539,0.0345028378,-0.1271548718,-0.5278391242,0.1267159581,0.1762584299,-0.1000838727,-0.3245918155,0.0069526397,-0.1820833683,-0.1220137328,0.1465669423,0.217528522,0.2876183987,-0.2172274143,0.0693417862,-0.2737216651]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1706","title":"Error when downloading a large dataset on slow connection.","comments":"Hi ! Is this an issue you have with `openwebtext` specifically or also with other datasets ?\r\n\r\nIt looks like the downloaded file is corrupted and can't be extracted using `tarfile`.\r\nCould you try loading it again with \r\n```python\r\nimport datasets\r\ndatasets.load_dataset(\"openwebtext\", download_mode=\"force_redownload\")\r\n```","body":"I receive the following error after about an hour trying to download the `openwebtext` dataset.\r\n\r\nThe code used is:\r\n```python\r\nimport datasets\r\ndatasets.load_dataset(\"openwebtext\")\r\n```\r\n\r\n> Traceback (most recent call last): [4\/28]\r\n> File \"<stdin>\", line 1, in <module>\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 610, in load_dataset\r\n> ignore_verifications=ignore_verifications,\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 515, in download_and_prepare\r\n> dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 570, in _download_and_prepare\r\n> split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n> File \"\/home\/lucadiliello\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/openwebtext\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02\/openwebtext.py\", line 62, in _split_generators\r\n> dl_dir = dl_manager.download_and_extract(_URL)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n> return self.extract(self.download(url_or_urls))\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 235, in extract\r\n> num_proc=num_proc,\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n> return function(data_struct)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n> tar_file.extractall(output_path_extracted)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 2000, in extractall\r\n> numeric_owner=numeric_owner)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 2042, in extract\r\n> numeric_owner=numeric_owner)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 2112, in _extract_member\r\n> self.makefile(tarinfo, targetpath)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 2161, in makefile\r\n> copyfileobj(source, target, tarinfo.size, ReadError, bufsize)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 253, in copyfileobj\r\n> buf = src.read(remainder)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/lzma.py\", line 200, in read\r\n> return self._buffer.read(size)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/_compression.py\", line 68, in readinto\r\n> data = self.read(len(byte_view))\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/_compression.py\", line 99, in read\r\n> raise EOFError(\"Compressed file ended before the \"\r\n> EOFError: Compressed file ended before the end-of-stream marker was reached\r\n","comment_length":44,"text":"Error when downloading a large dataset on slow connection.\nI receive the following error after about an hour trying to download the `openwebtext` dataset.\r\n\r\nThe code used is:\r\n```python\r\nimport datasets\r\ndatasets.load_dataset(\"openwebtext\")\r\n```\r\n\r\n> Traceback (most recent call last): [4\/28]\r\n> File \"<stdin>\", line 1, in <module>\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 610, in load_dataset\r\n> ignore_verifications=ignore_verifications,\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 515, in download_and_prepare\r\n> dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 570, in _download_and_prepare\r\n> split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n> File \"\/home\/lucadiliello\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/openwebtext\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02\/openwebtext.py\", line 62, in _split_generators\r\n> dl_dir = dl_manager.download_and_extract(_URL)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n> return self.extract(self.download(url_or_urls))\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 235, in extract\r\n> num_proc=num_proc,\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n> return function(data_struct)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 343, in cached_path\r\n> tar_file.extractall(output_path_extracted)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 2000, in extractall\r\n> numeric_owner=numeric_owner)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 2042, in extract\r\n> numeric_owner=numeric_owner)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 2112, in _extract_member\r\n> self.makefile(tarinfo, targetpath)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 2161, in makefile\r\n> copyfileobj(source, target, tarinfo.size, ReadError, bufsize)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/tarfile.py\", line 253, in copyfileobj\r\n> buf = src.read(remainder)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/lzma.py\", line 200, in read\r\n> return self._buffer.read(size)\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/_compression.py\", line 68, in readinto\r\n> data = self.read(len(byte_view))\r\n> File \"\/home\/lucadiliello\/anaconda3\/envs\/nlp\/lib\/python3.7\/_compression.py\", line 99, in read\r\n> raise EOFError(\"Compressed file ended before the \"\r\n> EOFError: Compressed file ended before the end-of-stream marker was reached\r\n\nHi ! Is this an issue you have with `openwebtext` specifically or also with other datasets ?\r\n\r\nIt looks like the downloaded file is corrupted and can't be extracted using `tarfile`.\r\nCould you try loading it again with \r\n```python\r\nimport datasets\r\ndatasets.load_dataset(\"openwebtext\", download_mode=\"force_redownload\")\r\n```","embeddings":[-0.4910078049,0.0278396662,-0.1026933491,0.2193090171,0.2197147459,0.110001415,0.118274942,0.4798275232,0.0537419096,0.0058270204,-0.1662466526,-0.1080515906,-0.0206885375,0.09193407,-0.2218508869,-0.1143092886,-0.3376919031,0.2518778145,-0.1198240146,-0.0132287266,-0.2608829439,0.1014634371,-0.145431608,0.090435937,0.0286519937,-0.1973609328,0.2813675404,0.0957638249,-0.3298174739,-0.3242034912,0.1182493195,-0.0770616531,-0.1571705341,0.1978920549,-0.0001039065,0.1097812057,0.3197879195,-0.0127757546,-0.2622531056,-0.0126416013,-0.5175206661,-0.4060268998,-0.1065353081,-0.2328351736,0.1429639906,-0.3205480576,-0.0369036645,-0.4010324776,0.202750966,0.3491779864,0.3192969263,0.4104012847,0.2654625177,-0.1568136513,0.2949921787,-0.3527448177,-0.1403333992,0.0242081843,0.0966816396,0.1845524609,0.0158708096,-0.0254063047,-0.0977822691,0.3270840347,0.0858378485,-0.1159041077,0.1312582642,-0.1467746347,0.233209461,0.3168919384,0.6235703826,0.1186791435,-0.20031479,-0.0985546932,-0.1308957934,-0.0556169972,0.151860401,0.349847585,-0.263335675,0.1681705117,-0.0905272663,-0.1982218623,-0.1184892729,0.1373558342,-0.2389908135,0.209534958,-0.1198301837,0.1606208235,0.0837471113,-0.0340166725,0.1818950921,-0.2236408591,0.0057414072,-0.0193282627,-0.3181525171,0.0871534273,0.1587081403,0.0071564391,0.1576059014,0.1561466604,0.1806851625,-0.0814055875,-0.1455614567,0.0343292467,0.2485306412,0.1881300658,-0.2394278347,-0.1837043315,0.3369806409,0.0097603854,0.086356461,-0.0245990139,-0.0561116189,-0.4480530322,0.0691169053,-0.0665593967,0.4185087979,-0.0500821881,-0.6267598867,0.0847598687,0.0171588864,0.0917018354,0.0974617749,0.134782359,-0.3348250687,0.1765510738,0.2755548954,-0.0854988396,-0.2178014815,-0.2556179464,-0.287570864,-0.0248287618,-0.000398937,-0.1589295864,0.0877899602,-0.053948693,0.1952300072,-0.1422249824,-0.0910601318,-0.0516946539,-0.0116248503,-0.2797981799,-0.0472413972,0.4205085039,-0.117083542,0.217013225,0.0714032352,0.0680958405,-0.1618653089,0.0930211097,-0.148138687,-0.5448227525,-0.204013586,0.2917512655,-0.1607732773,-0.1163366884,0.0674725622,-0.1778175682,0.06017204,-0.0622055978,-0.2368813455,-0.1359258592,-0.3001925945,-0.0643514469,0.1700168997,0.4760253727,-0.1476195902,0.1128111631,-0.1583244354,-0.0900966674,0.1099371091,0.3871322274,-0.2834582925,0.3439683616,-0.1608299464,-0.0327681415,0.4121798277,-0.2027587891,-0.4505096972,0.4190320075,-0.2133525163,-0.0175491925,-0.0150948204,-0.0155986957,0.247559458,-0.2489022017,0.0600092858,0.4046271145,-0.2269219607,0.1832318604,-0.2877351344,-0.3469765484,0.0445903689,0.314044416,0.1543741524,-0.0975275189,0.073860243,0.3041230142,0.1083477139,0.2421858162,0.1819605082,0.2064588368,-0.0133030163,0.0129645886,-0.2803225815,0.0975571424,-0.2467361987,0.2262407392,0.2003411502,0.1243642643,-0.1202566773,0.0499298647,-0.5512596369,-0.0064605637,-0.1038976386,0.2097591162,0.1960365325,-0.0458704643,0.1451184601,0.2851141691,0.1329975277,0.2275888622,0.0520906821,-0.0502490923,-0.0580034144,0.0837767497,-0.0531922542,0.065541923,0.3550397754,-0.2205490768,-0.0292083267,-0.0798258036,-0.2772213221,0.2578870058,-0.1733716875,0.0356222764,-0.0232283715,0.1471143663,0.1160377115,-0.1642075181,0.0679106638,0.1062736958,0.3960361779,-0.0606126487,0.1969256699,-0.0338592194,0.1103016064,-0.0591172874,0.0435753763,0.0712376535,0.3167003989,-0.1312634945,0.2426616699,0.0766423568,0.2193253189,0.179530412,0.1072681695,-0.054201562,0.0830672681,0.0046871221,0.4841132164,-0.0464976095,0.0474111699,0.2676337063,0.0586583279,-0.1077326462,-0.2526250184,0.4030234814,0.4827991724,0.0838006437,0.2591084838,0.0453576967,0.1262770891,-0.2797545493,0.3492495418,0.3568131924,0.2874477804,0.2896438539,0.0301173292,-0.1230559945,-0.2845939696,-0.3535315096,0.3476915658,0.608286202,-0.1231075972,-0.0709432438,-0.1247352287,-0.0788890198,0.1208879426,0.0440233424,-0.0377581157,-0.4296837747,-0.0154182985,0.0485089384,-0.143442899,0.1439752132,-0.1638388634,0.2240302116,0.4131111503,0.0995168015,-0.1292496473,-0.1925107539,-0.1532576531,0.1184123158,0.3695813417,-0.0189591553,0.2307117283,-0.2334080637,-0.071258001,-0.3312543631,0.123114273,-0.018692553,0.0426615924,0.0100765182,0.2998990119,0.4455482364,0.1532676369,-0.2030158043,0.2708804011,-0.2101938725,0.0325599946,0.1356427372,0.0261241738,0.0646527857,0.0309957042,-0.2389444709,-0.3499540687,-0.6167093515,0.1416400969,-0.1712803692,0.1445546895,0.0156337209,-0.1197377443,0.1246792153,0.0642575175,0.0857084394,-0.193493858,-0.5720279813,0.3628085554,-0.154495433,-0.347192198,-0.0274207089,0.1902089417,0.1565325856,-0.022583887,-0.5953269601,0.0783892646,-0.446057111,0.0887931585,-0.2679098248,0.0684628487,0.1131035984,-0.0909366831,-0.2421576828,-0.0365531445,0.0605466366,0.103685841,-0.0470477194,0.1688963324,-0.1197036132,0.3804698586,0.1618496031,0.0838429481,0.2322383076,-0.0005098601,0.3763803244,0.1972851753,0.1473808438,-0.062704578,-0.1118450463,-0.0637829825,-0.1193949208,0.0731587335,0.2853550017,0.0205624364,-0.0401655324,-0.3821673393,-0.1634704769,-0.0830927491,-0.4175002277,-0.0579316579,-0.1260106266,0.1006861627,0.0835848674,-0.0498085804,-0.1936892122,-0.1125148833,-0.0042553185,0.0742680058,-0.0459685139,-0.0927464738,0.1516088843,0.3976322114,-0.7130667567,0.1737930477,-0.0037281583,0.4163466394,0.0052010105,0.077688776,0.1122407392,-0.0666665509,0.3629688919,-0.0977147296,0.3858887553,0.0432833135,0.145528093,-0.3521086276,-0.0236998405,-0.0561037585,0.1375778615,0.4362489283,-0.2392549068,-0.5605276227,-0.0402185582,0.3201287091,0.1238596961,-0.1307424009,-0.2306485176,-0.0287750103,-0.6097215414,-0.2158107013,-0.1519479603,0.2367378324,0.2336744219,0.137049526,-0.0550982244,-0.0979823247,0.0121197952,0.0348035246,-0.2681641579,-0.0824740678,-0.0190520473,0.5001140833,-0.0127166649,0.0137654217,0.162599206,0.6605308056,0.1451497823,-0.3008754253,0.0413114987,0.2078367919,-0.1393837929,0.2565187514,-0.0801176131,-0.0304289982,-0.086013943,0.46448645,-0.024819538,0.3133134246,0.3565271497,-0.1573584825,-0.3542947173,-0.322861135,0.4364982247,-0.0754619986,0.0638308376,0.4427607358,0.0964241251,0.0524069518,0.0602001585,-0.1492577195,0.7682144046,-0.1116572097,-0.0047828043,0.1478163004,-0.2076024115,0.3053106666,-0.1913778037,0.0256938953,-0.1335925162,-0.432204783,0.0787982866,-0.0426881872,0.0228358489,0.1223483756,-0.1477642804,0.0169648249,0.0007749416,-0.2142682672,0.0314741656,0.4907328486,-0.5565981269,-0.1166282222,-0.1270233989,0.2256366611,-0.2551411688,0.3092598915,-0.1285843551,-0.0900940076,0.1220004484,-0.2126545608,-0.4187890589,0.0347330905,-0.2291086018,0.4083173871,-0.2946042717,0.0652202666,0.3361534476,-0.0699621439,0.0371867307,0.1932360381,-0.2972599864,0.3335825503,-0.2697312534,-0.0872478858,0.1852784157,0.0979921743,0.1624974161,-0.2026276141,-0.1656112373,0.094204627,-0.1651604176,-0.2059892863,0.3227593899,0.0229297522,0.1906511337,-0.2575149536,-0.2088268548,-0.0784809068,-0.153033793,-0.1943639666,0.1732716709,0.3055021763,-0.1499928534,0.3735973537,-0.0456563905,-0.4248388708,-0.0651774108,0.3925415277,-0.2178729475,0.1265368313,0.4035223722,0.1035200655,-0.3173649311,-0.1918321252,0.2314175218,-0.1202641949,-0.5844944119,0.3571233749,-0.262491405,0.2845506668,0.0306754485,0.4045993984,0.172669813,0.1268849373,0.1824621409,-0.4925782681,-0.0616446249,-0.0679735765,-0.0823444203,0.1801594347,-0.0252833311,0.2049983442,0.0119211879,0.0358079486,-0.3645217419,0.3533165157,-0.3521827757,0.0474166833,0.0871924236,-0.3461465538,0.3602366149,0.1056677252,0.1546500474,0.0580749027,-0.3006545901,-0.2725975811,-0.0807265043,0.1427759677,0.0178488381,-0.4366220832,0.247548759,-0.3979848623,-0.2433427721,0.1706397533,0.0391605087,0.0720063671,-0.100295186,-0.2405495644,0.0105662877,-0.0252357442,-0.0187582225,0.2366704494,-0.1987348199,0.3995851576,0.1308743805,0.2940323353,0.0873842239,-0.1530359238,-0.2370465845,0.074462682,0.0038517423,-0.1078542173,0.5733604431,-0.3086607754,-0.019715257,0.1899845451,0.3839217722,0.3524338901,-0.0118960384,-0.140550524,0.1700798571,0.2697324753,-0.2485198975,-0.1776453704,0.0630617887,0.1271475554,0.1100936458,0.0188021287,-0.0161136463,-0.0989336893,-0.3694089353,-0.001929822,0.4117057621,0.3302065432,-0.0287569966,0.5262479186,0.0035627221,-0.0639847443,0.0594775639,0.117833063,0.3483618796,0.3944522738,-0.2472985834,0.3183898032,-0.0796353072,0.1970584989,-0.2897847891,-0.5639091134,0.1555327177,0.1142828092,-0.1137361079,0.0518601127,0.0750764832,0.1444234103,-0.2696656287,0.0964280218,-0.3930126727,0.5118334889,0.0461307839,-0.0731805339,0.207194075,-0.0755226836,-0.1088180244,0.1645210534,-0.0137956338,-0.0383344255,0.0655359849,0.1750602871,0.1009360552,-0.1891844422,-0.0304680541,-0.1590442061,0.0775338784,-0.3614949882,0.035659574,0.3301439881,-0.1794326305,-0.1093373001,0.5390256047,0.420320034,0.3193342686,-0.005500949,0.0822536871,0.0205139779,0.0272418987,-0.177460894,0.0628499389,0.1928868592,0.2628744841,0.4981499016,0.0966405496,-0.2112579942,-0.21730344,-0.0118186874,-0.1849784553,-0.0412873812,0.188111797,-0.0130401654,-0.0595851801,-0.3763494492,0.0329054333,-0.0846481547,0.0820523798,-0.0008932696,-0.3523808718,0.205358088,-0.1216223463,0.1146763638,-0.3505585492,0.6454350948,0.2102538049,0.1561595649,-0.23876369,-0.3794487417,-0.4723809063,0.1177992076,-0.2668834925,-0.1449926049,0.1356087625,0.2898333967,-0.225463897,0.1501580626,0.0957171842,0.4053815603,-0.1734325141,0.0557582788,-0.3363560736,-0.2513551712,0.1509876698,-0.095838353,-0.0231815111,-0.4900802374,0.2447085232,-0.3484030664,0.1998137236,-0.2401984483,-0.0076636416,-0.125893712,0.0079817884,0.3705799878,0.1578872651,0.2045782655,0.0700187758,-0.1450943649,-0.0836943537,-0.259566009,-0.129193306,0.1942569613,0.1567402035,0.0164395887,0.0662736595,0.1653201878,-0.1236556247,0.3453650773,-0.1953725964,0.3909793198,-0.2161419839,-0.1557773352,-0.1425409168,0.2317115217,0.2600620985,0.0118149053,-0.1051554009,0.0664021298,-0.3719685376,-0.4535042048,0.5555471778,0.1017124131,-0.2595461011,-0.1664861739,0.2619670033,0.1153146625,0.0494315363,-0.2518612444,0.2396019846,0.3347976506,-0.0465144441,-0.4082004428,0.2329713851,-0.0145888226,-0.1679085195,0.1108665466,0.2559424937,-0.0674548447,-0.4843255579,0.0090418532,-0.0340458453]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1701","title":"Some datasets miss dataset_infos.json or dummy_data.zip","comments":"Thanks for reporting.\r\nWe should indeed add all the missing dummy_data.zip and also the dataset_infos.json at least for lm1b, reclor and wikihow.\r\n\r\nFor c4 I haven't tested the script and I think we'll require some optimizations regarding beam datasets before processing it.\r\n","body":"While working on dataset REAME generation script at https:\/\/github.com\/madlag\/datasets_readme_generator , I noticed that some datasets miss a dataset_infos.json : \r\n\r\n```\r\nc4\r\nlm1b\r\nreclor\r\nwikihow\r\n```\r\n\r\nAnd some does not have a dummy_data.zip : \r\n\r\n```\r\nkor_nli\r\nmath_dataset\r\nmlqa\r\nms_marco\r\nnewsgroup\r\nqa4mre\r\nqangaroo\r\nreddit_tifu\r\nsuper_glue\r\ntrivia_qa\r\nweb_of_science\r\nwmt14\r\nwmt15\r\nwmt16\r\nwmt17\r\nwmt18\r\nwmt19\r\nxtreme\r\n```\r\n\r\nBut it seems that some of those last do have a \"dummy\" directory .\r\n\r\n","comment_length":42,"text":"Some datasets miss dataset_infos.json or dummy_data.zip\nWhile working on dataset REAME generation script at https:\/\/github.com\/madlag\/datasets_readme_generator , I noticed that some datasets miss a dataset_infos.json : \r\n\r\n```\r\nc4\r\nlm1b\r\nreclor\r\nwikihow\r\n```\r\n\r\nAnd some does not have a dummy_data.zip : \r\n\r\n```\r\nkor_nli\r\nmath_dataset\r\nmlqa\r\nms_marco\r\nnewsgroup\r\nqa4mre\r\nqangaroo\r\nreddit_tifu\r\nsuper_glue\r\ntrivia_qa\r\nweb_of_science\r\nwmt14\r\nwmt15\r\nwmt16\r\nwmt17\r\nwmt18\r\nwmt19\r\nxtreme\r\n```\r\n\r\nBut it seems that some of those last do have a \"dummy\" directory .\r\n\r\n\nThanks for reporting.\r\nWe should indeed add all the missing dummy_data.zip and also the dataset_infos.json at least for lm1b, reclor and wikihow.\r\n\r\nFor c4 I haven't tested the script and I think we'll require some optimizations regarding beam datasets before processing it.\r\n","embeddings":[0.1359894276,0.2731491327,-0.046184022,0.2401537597,0.2858270407,0.3439723551,0.2180934101,0.0033975567,-0.046791587,0.0312941,0.1941788048,0.0641290918,-0.0075305644,0.125732705,0.1229167581,-0.0037490961,0.3180769682,0.1551839411,0.0810289979,-0.3036448956,-0.6437272429,0.289752394,-0.2662017941,-0.2682595551,-0.3690704703,0.2852796316,-0.1664090157,0.1071849167,-0.076620549,-0.258440882,0.1300564408,0.0621439666,0.1241826564,0.2200273722,-0.0001218869,-0.1835252196,0.3358968496,-0.318447113,-0.6375688314,0.1391728073,-0.2461787611,-0.5139753819,-0.2116565853,-0.0337942131,0.0850708559,-0.3283745348,0.2760594189,-0.4367847443,0.0252547059,0.2878524363,0.1385961175,0.0108711906,0.0054563372,0.019353617,0.4706355333,0.5881893635,-0.0361154415,0.0234698635,-0.0308680721,0.1876572967,0.16875723,0.3128627837,-0.000031814,-0.0429410897,0.2564952374,0.0667167678,-0.0803639218,-0.313919574,0.0929040536,0.0367850214,0.5615602732,-0.3702829778,-0.57642138,-0.4374137819,0.2442951351,-0.0141523341,0.2991722524,0.1897802353,0.2458381951,-0.0496902652,-0.113386102,-0.2367790937,0.230628103,0.1129053608,0.2745579779,0.3560731709,-0.013902422,0.1380821466,-0.3792663813,0.0509369262,0.5013102889,-0.6761803031,-0.0721880421,-0.0376387462,-0.0408893935,0.0017078262,-0.2982947528,-0.0323235504,0.0126889264,-0.0287912749,-0.0522561073,-0.0006028205,-0.0576110333,0.0159178562,-0.0656326339,0.0639261454,0.2777886391,0.4664542973,0.0937818959,-0.2195849121,0.2740783989,-0.0936787277,-0.1424447447,-0.0849497914,-0.1278651655,0.3348123133,-0.2248756737,0.0716191381,-0.3074820936,0.0207017157,-0.1901276708,-0.1387444437,-0.1389830709,0.0208207294,-0.0931715593,0.5326836705,0.0541503876,0.1507941484,0.0578553416,-0.4711914361,-0.1495305598,0.046051167,-0.2453964204,0.2048980445,0.2601968944,0.268576175,0.5244526863,-0.0994325578,-0.2592103779,0.0434183478,0.3255158961,0.1851344109,-0.0268367305,0.2875718772,0.0094899181,0.3180415332,-0.1978770792,0.0012944319,0.0012153551,0.2754144669,-0.2969889641,-0.189390108,-0.1847658455,0.1388822645,-0.2081683278,0.0584152006,-0.4075929523,0.0272928681,0.0436878465,-0.2390128523,0.0437335782,-0.1227899194,0.019875573,-0.2703769505,0.1394422948,0.4934485555,-0.3240852356,-0.0644108057,-0.3538264334,0.1564725339,0.3093661368,-0.1211022809,0.0391098224,0.1585644186,-0.5670543313,-0.1100330055,0.3649954796,-0.4209957123,0.0515694283,-0.0461734235,0.169695273,0.0378894769,0.2327091843,0.2033316046,-0.212942481,-0.0323861837,-0.1897785962,-0.0280067399,-0.0973539576,-0.0145749338,-0.2613099515,-0.1372375637,0.201224342,0.3041262329,0.0936595052,0.0770851448,-0.0792812929,-0.1432590336,0.0890172422,-0.035811048,0.1986543834,0.3075972795,-0.3370240629,0.119671613,0.3427751362,0.2289583981,-0.914237082,0.1389713287,-0.0475766622,-0.1395034045,0.0272238664,-0.3427241743,-0.0693246722,-0.2794781625,-0.0491127633,-0.2503945231,0.0087548112,0.4585579336,0.2593406141,-0.1326010227,-0.2570742667,0.0305096209,-0.1728773117,0.2510372996,-0.5224624872,0.3718422353,0.004179833,0.1032609865,0.1918188781,0.101075165,0.1200629473,-0.1771908551,0.0113642123,0.3651761711,0.1379337758,0.0009049419,0.2633714676,0.2109721452,0.0876649991,-0.070728153,0.1987474412,0.117308192,0.0558156595,-0.1434938908,-0.0086810878,0.2717199326,-0.1113152727,0.0177729391,-0.1500590295,-0.0267463587,0.2357837856,0.0685027167,0.1884488165,-0.2720340192,0.057404846,0.3031097651,0.3498010933,0.0565030687,-0.1295443624,0.0957012847,0.4859601259,-0.3183689415,-0.0276597552,-0.0212369431,-0.3904747367,-0.044519186,0.2228826284,0.2744318545,0.4791937768,0.2456193566,-0.2215274125,-0.180290997,0.1141359061,-0.364847064,0.1915976554,0.0979761034,-0.1553996801,0.3336868286,0.0521303564,0.0374680012,-0.407613039,-0.0300660804,0.3921745718,0.1541529596,-0.3328437209,-0.2906156182,-0.0937220603,-0.2301835567,-0.0803006068,0.2770461738,-0.1640733331,-0.5771505237,0.2277849019,0.0148212621,-0.3227028847,0.3482872844,0.2542717159,0.0687833577,-0.3534528017,0.0768535957,-0.6138683558,-0.003677343,-0.2684159875,0.0301392935,0.1393002123,0.104258962,0.1292711943,-0.7210150361,0.1649117023,-0.5698032975,-0.2802166939,0.0580845736,-0.0072912183,0.1889294237,0.2599308491,0.0986851677,-0.0037966648,-0.1208011806,0.0712174177,0.1265302747,-0.1768732667,0.0677923188,-0.0114057483,-0.0121208653,-0.1999921799,-0.3856360316,-0.2595992088,-0.22407718,-0.1673861146,-0.136222899,0.0212258995,0.0338353962,-0.1363120079,-0.033357203,0.0263157953,-0.1303356439,-0.1572937518,-0.4554775953,0.5669626594,-0.1628836989,-0.0462445617,0.209676072,-0.3875752985,0.0092359427,-0.0341530442,-0.6006202102,-0.2145354152,0.0365673415,0.0606159568,-0.0902967826,0.0497137159,0.3813123703,0.3126040399,0.0970507562,-0.2882058322,-0.1956235915,-0.067770645,-0.1374189556,0.4222305417,-0.279386729,-0.0224152543,-0.3571079373,0.398136586,0.2477406859,-0.1056586355,0.1083179116,0.1153872833,0.4375533164,-0.0171891749,0.0605185069,0.0567205064,-0.0126334252,0.2005238533,0.3110232055,0.0785471201,0.1835682541,-0.0256749429,0.128479287,-0.3604442179,-0.1820427924,0.0232638344,-0.1196426526,0.0911342353,0.200458318,0.0888178051,-0.232808724,-0.196194455,0.0946549699,0.2079752833,0.5585255623,0.0503572933,-0.4174316227,-0.1866532415,-0.0107086571,0.4025134742,0.150041014,0.2676314116,0.2401701361,0.2874740362,-0.2447521836,0.2170765996,0.5383940339,-1.0103065968,-0.2003462762,0.0466930009,0.6165086627,-0.2542077899,-0.1170921698,0.0256243441,0.0829237774,0.013454183,0.1802129894,-0.21288158,-0.1615298688,0.2655406892,0.4586063027,0.0179915503,-0.2447164655,-0.078106299,-0.0554669462,-0.2801615596,-0.4896795452,-0.0010654279,-0.2761235535,-0.0090524917,-0.2039987445,0.0277847685,-0.2288344204,0.1820269972,-0.0100149456,0.227618888,0.303801775,0.316937983,0.1177252531,0.2444379181,0.5586234331,0.5635813475,-0.3033978045,-0.3029626906,0.1503861547,-0.1343718767,0.3772443831,0.214542821,0.068375349,-0.0693167895,-0.2652711272,-0.0362347178,-0.2845917046,0.177072078,0.1645657718,0.053407833,-0.3091245294,-0.4064051807,0.2014542073,0.0541184247,-0.1248434559,0.3389101923,0.1346417814,-0.3746578097,0.3007406592,-0.1152910069,1.002707243,0.1277339458,0.0836676061,-0.09475822,0.0115905199,0.6359570622,-0.315410763,0.1233190373,-0.1249523014,0.0711188614,-0.0100005697,-0.0093773603,0.1471477449,0.3722653985,-0.0328728817,0.5601644516,0.0938758329,0.4175255001,-0.0544891208,0.2500444651,-0.2537615895,-0.1809325963,0.2739742398,0.0162031241,-0.0426957197,0.0829455554,0.0595351309,-0.0245522987,-0.1429384053,-0.0078436201,-0.2475022227,-0.0282252766,-0.3308624327,-0.1277550608,-0.0289097559,-0.0725110248,-0.0067758625,0.1689165533,0.5375509858,0.2703065574,-0.0621441603,0.2709712982,-0.081920214,0.0922540426,0.041169472,-0.0740426257,-0.1028685048,0.0562778115,-0.0744242668,-0.0079889977,-0.412124902,0.0052622026,-0.4038646519,-0.4667079151,0.2069479525,-0.1776050627,0.2608372271,0.0941044912,-0.2116099149,-0.3247471452,0.0687936321,-0.0242491309,0.2117007822,-0.0209021065,-0.3105004728,-0.1551169008,-0.4754319787,0.1253658682,0.1770388782,-0.1009802148,0.4011859,-0.1709603816,-0.1162766516,-0.1660059988,0.2657420933,0.1273713857,-0.2822157443,0.0717786923,-0.0516419224,-0.2686140239,-0.0492221378,0.060232047,0.2299056798,-0.0214541797,0.1600276232,-0.4466021955,-0.2635631263,0.1815007627,0.1603439897,0.1072432771,-0.1957227588,-0.1532741785,-0.0334213898,0.0235113911,-0.203729704,-0.0890061632,-0.0180811901,0.316798687,0.0886947662,0.2731158137,0.1522659361,-0.134078458,0.0058693914,0.0242392551,0.0209878162,-0.0047082002,0.0320959464,0.2049372047,0.1247791201,-0.0957618207,0.2437929213,-0.0666902736,-0.1484735906,-0.2659059465,-0.2623092532,-0.1123137996,-0.0013641655,0.497284323,0.1143818349,0.0448696092,-0.0763872713,0.0579337589,-0.0657782108,0.5421603918,-0.1733845025,0.2078074962,0.3188311458,0.0422801115,-0.1831746995,-0.0930963233,-0.23731637,0.1044136807,0.4356819093,-0.2738849521,0.1221450716,0.2756364942,-0.007941206,0.3427302837,-0.2410666645,-0.0617780499,0.2770422697,0.1177262366,-0.3703524172,-0.1771057844,0.2478829026,0.3092760146,0.0648836419,0.1592766196,0.1337839216,-0.2467018217,0.2583225667,0.1362845898,0.3094519377,-0.0671507642,0.3148542643,0.2230911702,0.1964631826,0.2141834944,-0.0433703698,0.0141072664,-0.0717177466,0.555901289,0.1422373801,0.6019521952,0.1985079944,0.0706248358,0.0754957944,-0.3192096055,0.34393695,0.5413245559,-0.0692233145,0.2922509611,0.1845745891,-0.0180414766,-0.1218026951,0.1467634887,-0.1806455553,-0.0021914146,-0.1744205654,0.0432000533,-0.2166258991,-0.0661893487,-0.2614702582,0.0260347314,0.0454846807,-0.2277484238,0.2086768597,0.0064153364,-0.1540877223,-0.1541655362,0.4206681252,0.1175651029,0.3643223643,-0.2271979302,0.2231368273,0.3812369406,-0.31726107,0.4052366316,0.3339423537,0.2221366912,0.0724444315,-0.0196903106,-0.1172952577,0.0145675298,-0.0787317008,0.0801946968,0.476305902,-0.0863164291,-0.3060057759,0.1655384451,0.1167199612,-0.1239079684,-0.1751880497,0.2372307181,-0.1447286308,-0.273327589,0.3208118975,-0.2983164489,-0.0790087879,-0.1084705815,-0.0175694451,-0.3955311179,-0.3162437081,0.3800291419,0.0296805855,-0.0507141761,-0.2118146271,0.0316882469,-0.2614919543,0.444788754,0.4153662622,-0.0951008648,-0.3322181702,-0.295155853,-0.2513737977,0.041765742,-0.0512599386,0.0052002468,-0.105270505,-0.1297661513,0.1619136631,0.0121689113,-0.108023867,0.3831099272,0.1030439734,0.2312244326,-0.4300259948,-0.3357351422,-0.0854236782,0.0987180695,0.0812413916,-0.469142735,0.2824628353,0.1235402822,-0.1003422961,0.0333906487,0.0006551085,0.4899043441,0.3778561056,0.4759448767,0.1656396538,0.4670661092,-0.2572008967,-0.2580906749,0.0803517327,-0.2468325198,-0.0421415418,0.3837555051,0.0160487834,0.3203470111,-0.3290711343,0.1257861406,-0.1664905846,-0.204188168,-0.059520334,0.3903376758,-0.5664530396,0.3764694631,-0.1835462153,0.203862235,0.1211839765,-0.1135696843,0.0219943449,0.103354536,-0.2140759826,-0.2290601581,0.3075905442,-0.2217805535,-0.0855523199,-0.008398057,0.1346280277,-0.1375499815,0.2028888166,-0.5929297209,-0.0211003087,0.516874969,0.2246752232,0.0559395738,-0.0197068918,0.1601094007,-0.1229050085,-0.0073446259,0.1320776194,0.0531569868,-0.3395538032,0.163769275,-0.2539936602]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1696","title":"Unable to install datasets","comments":"Maybe try to create a virtual env with python 3.8 or 3.7","body":"** Edit **\r\nI believe there's a bug with the package when you're installing it with Python 3.9. I recommend sticking with previous versions. Thanks, @thomwolf for the insight! \r\n\r\n**Short description**\r\n\r\nI followed the instructions for installing datasets (https:\/\/huggingface.co\/docs\/datasets\/installation.html). However, while I tried to download datasets using `pip install datasets` I got a massive error message after getting stuck at \"Installing build dependencies...\" \r\n\r\nI was wondering if this problem can be fixed by creating a virtual environment, but it didn't help. Can anyone offer some advice on how to fix this issue? \r\n\r\nHere's an error message: \r\n\r\n`(env) Gas-MacBook-Pro:Downloads destiny$ pip install datasets\r\nCollecting datasets\r\n Using cached datasets-1.2.0-py3-none-any.whl (159 kB)\r\nCollecting numpy>=1.17\r\n Using cached numpy-1.19.5-cp39-cp39-macosx_10_9_x86_64.whl (15.6 MB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-2.0.0.tar.gz (58.9 MB)\r\n....\r\n\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceilf' [-Wincompatible-library-redeclaration]\r\n int ceilf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceilf' is a builtin with type 'float (float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintf' [-Wincompatible-library-redeclaration]\r\n int rintf (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintf' is a builtin with type 'float (float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncf' [-Wincompatible-library-redeclaration]\r\n int truncf (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncf' is a builtin with type 'float (float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtf' [-Wincompatible-library-redeclaration]\r\n int sqrtf (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtf' is a builtin with type 'float (float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10f' [-Wincompatible-library-redeclaration]\r\n int log10f (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10f' is a builtin with type 'float (float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logf' [-Wincompatible-library-redeclaration]\r\n int logf (void);\r\n ^\r\n _configtest.c:14:5: note: 'logf' is a builtin with type 'float (float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pf' [-Wincompatible-library-redeclaration]\r\n int log1pf (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pf' is a builtin with type 'float (float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expf' [-Wincompatible-library-redeclaration]\r\n int expf (void);\r\n ^\r\n _configtest.c:16:5: note: 'expf' is a builtin with type 'float (float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1f' [-Wincompatible-library-redeclaration]\r\n int expm1f (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1f' is a builtin with type 'float (float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinf' [-Wincompatible-library-redeclaration]\r\n int asinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinf' is a builtin with type 'float (float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosf' [-Wincompatible-library-redeclaration]\r\n int acosf (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosf' is a builtin with type 'float (float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanf' [-Wincompatible-library-redeclaration]\r\n int atanf (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanf' is a builtin with type 'float (float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhf' [-Wincompatible-library-redeclaration]\r\n int asinhf (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhf' is a builtin with type 'float (float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshf' [-Wincompatible-library-redeclaration]\r\n int acoshf (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshf' is a builtin with type 'float (float)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhf' [-Wincompatible-library-redeclaration]\r\n int atanhf (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhf' is a builtin with type 'float (float)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotf' [-Wincompatible-library-redeclaration]\r\n int hypotf (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotf' is a builtin with type 'float (float, float)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2f' [-Wincompatible-library-redeclaration]\r\n int atan2f (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2f' is a builtin with type 'float (float, float)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powf' [-Wincompatible-library-redeclaration]\r\n int powf (void);\r\n ^\r\n _configtest.c:26:5: note: 'powf' is a builtin with type 'float (float, float)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodf' [-Wincompatible-library-redeclaration]\r\n int fmodf (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodf' is a builtin with type 'float (float, float)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modff' [-Wincompatible-library-redeclaration]\r\n int modff (void);\r\n ^\r\n _configtest.c:28:5: note: 'modff' is a builtin with type 'float (float, float *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpf' [-Wincompatible-library-redeclaration]\r\n int frexpf (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpf' is a builtin with type 'float (float, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpf' [-Wincompatible-library-redeclaration]\r\n int ldexpf (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpf' is a builtin with type 'float (float, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2f' [-Wincompatible-library-redeclaration]\r\n int exp2f (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2f' is a builtin with type 'float (float)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2f' [-Wincompatible-library-redeclaration]\r\n int log2f (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2f' is a builtin with type 'float (float)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignf' [-Wincompatible-library-redeclaration]\r\n int copysignf (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignf' is a builtin with type 'float (float, float)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterf' [-Wincompatible-library-redeclaration]\r\n int nextafterf (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterf' is a builtin with type 'float (float, float)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtf' [-Wincompatible-library-redeclaration]\r\n int cbrtf (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtf' is a builtin with type 'float (float)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'sinl' [-Wincompatible-library-redeclaration]\r\n int sinl (void);\r\n ^\r\n _configtest.c:1:5: note: 'sinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cosl' [-Wincompatible-library-redeclaration]\r\n int cosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'tanl' [-Wincompatible-library-redeclaration]\r\n int tanl (void);\r\n ^\r\n _configtest.c:3:5: note: 'tanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'sinhl' [-Wincompatible-library-redeclaration]\r\n int sinhl (void);\r\n ^\r\n _configtest.c:4:5: note: 'sinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'coshl' [-Wincompatible-library-redeclaration]\r\n int coshl (void);\r\n ^\r\n _configtest.c:5:5: note: 'coshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'tanhl' [-Wincompatible-library-redeclaration]\r\n int tanhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'tanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'fabsl' [-Wincompatible-library-redeclaration]\r\n int fabsl (void);\r\n ^\r\n _configtest.c:7:5: note: 'fabsl' is a builtin with type 'long double (long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'floorl' [-Wincompatible-library-redeclaration]\r\n int floorl (void);\r\n ^\r\n _configtest.c:8:5: note: 'floorl' is a builtin with type 'long double (long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceill' [-Wincompatible-library-redeclaration]\r\n int ceill (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceill' is a builtin with type 'long double (long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintl' [-Wincompatible-library-redeclaration]\r\n int rintl (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintl' is a builtin with type 'long double (long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncl' [-Wincompatible-library-redeclaration]\r\n int truncl (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncl' is a builtin with type 'long double (long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtl' [-Wincompatible-library-redeclaration]\r\n int sqrtl (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtl' is a builtin with type 'long double (long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10l' [-Wincompatible-library-redeclaration]\r\n int log10l (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10l' is a builtin with type 'long double (long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logl' [-Wincompatible-library-redeclaration]\r\n int logl (void);\r\n ^\r\n _configtest.c:14:5: note: 'logl' is a builtin with type 'long double (long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pl' [-Wincompatible-library-redeclaration]\r\n int log1pl (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pl' is a builtin with type 'long double (long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expl' [-Wincompatible-library-redeclaration]\r\n int expl (void);\r\n ^\r\n _configtest.c:16:5: note: 'expl' is a builtin with type 'long double (long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1l' [-Wincompatible-library-redeclaration]\r\n int expm1l (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1l' is a builtin with type 'long double (long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinl' [-Wincompatible-library-redeclaration]\r\n int asinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosl' [-Wincompatible-library-redeclaration]\r\n int acosl (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanl' [-Wincompatible-library-redeclaration]\r\n int atanl (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhl' [-Wincompatible-library-redeclaration]\r\n int asinhl (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshl' [-Wincompatible-library-redeclaration]\r\n int acoshl (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhl' [-Wincompatible-library-redeclaration]\r\n int atanhl (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotl' [-Wincompatible-library-redeclaration]\r\n int hypotl (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2l' [-Wincompatible-library-redeclaration]\r\n int atan2l (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2l' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powl' [-Wincompatible-library-redeclaration]\r\n int powl (void);\r\n ^\r\n _configtest.c:26:5: note: 'powl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodl' [-Wincompatible-library-redeclaration]\r\n int fmodl (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modfl' [-Wincompatible-library-redeclaration]\r\n int modfl (void);\r\n ^\r\n _configtest.c:28:5: note: 'modfl' is a builtin with type 'long double (long double, long double *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpl' [-Wincompatible-library-redeclaration]\r\n int frexpl (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpl' is a builtin with type 'long double (long double, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpl' [-Wincompatible-library-redeclaration]\r\n int ldexpl (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpl' is a builtin with type 'long double (long double, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2l' [-Wincompatible-library-redeclaration]\r\n int exp2l (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2l' [-Wincompatible-library-redeclaration]\r\n int log2l (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignl' [-Wincompatible-library-redeclaration]\r\n int copysignl (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterl' [-Wincompatible-library-redeclaration]\r\n int nextafterl (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtl' [-Wincompatible-library-redeclaration]\r\n int cbrtl (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtl' is a builtin with type 'long double (long double)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:8:12: error: use of undeclared identifier 'HAVE_DECL_SIGNBIT'\r\n (void) HAVE_DECL_SIGNBIT;\r\n ^\r\n 1 error generated.\r\n failure.\r\n removing: _configtest.c _configtest.o\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabs' [-Wincompatible-library-redeclaration]\r\n int cabs (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabs' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacos' [-Wincompatible-library-redeclaration]\r\n int cacos (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacosh' [-Wincompatible-library-redeclaration]\r\n int cacosh (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'carg' [-Wincompatible-library-redeclaration]\r\n int carg (void);\r\n ^\r\n _configtest.c:4:5: note: 'carg' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casin' [-Wincompatible-library-redeclaration]\r\n int casin (void);\r\n ^\r\n _configtest.c:5:5: note: 'casin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinh' [-Wincompatible-library-redeclaration]\r\n int casinh (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catan' [-Wincompatible-library-redeclaration]\r\n int catan (void);\r\n ^\r\n _configtest.c:7:5: note: 'catan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanh' [-Wincompatible-library-redeclaration]\r\n int catanh (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccos' [-Wincompatible-library-redeclaration]\r\n int ccos (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccosh' [-Wincompatible-library-redeclaration]\r\n int ccosh (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexp' [-Wincompatible-library-redeclaration]\r\n int cexp (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexp' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimag' [-Wincompatible-library-redeclaration]\r\n int cimag (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimag' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clog' [-Wincompatible-library-redeclaration]\r\n int clog (void);\r\n ^\r\n _configtest.c:13:5: note: 'clog' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conj' [-Wincompatible-library-redeclaration]\r\n int conj (void);\r\n ^\r\n _configtest.c:14:5: note: 'conj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpow' [-Wincompatible-library-redeclaration]\r\n int cpow (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpow' is a builtin with type '_Complex double (_Complex double, _Complex double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cproj' [-Wincompatible-library-redeclaration]\r\n int cproj (void);\r\n ^\r\n _configtest.c:16:5: note: 'cproj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creal' [-Wincompatible-library-redeclaration]\r\n int creal (void);\r\n ^\r\n _configtest.c:17:5: note: 'creal' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csin' [-Wincompatible-library-redeclaration]\r\n int csin (void);\r\n ^\r\n _configtest.c:18:5: note: 'csin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinh' [-Wincompatible-library-redeclaration]\r\n int csinh (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrt' [-Wincompatible-library-redeclaration]\r\n int csqrt (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrt' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctan' [-Wincompatible-library-redeclaration]\r\n int ctan (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanh' [-Wincompatible-library-redeclaration]\r\n int ctanh (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanh' is a builtin with type '_Complex double (_Complex double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsf' [-Wincompatible-library-redeclaration]\r\n int cabsf (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosf' [-Wincompatible-library-redeclaration]\r\n int cacosf (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshf' [-Wincompatible-library-redeclaration]\r\n int cacoshf (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargf' [-Wincompatible-library-redeclaration]\r\n int cargf (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinf' [-Wincompatible-library-redeclaration]\r\n int casinf (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhf' [-Wincompatible-library-redeclaration]\r\n int casinhf (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanf' [-Wincompatible-library-redeclaration]\r\n int catanf (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhf' [-Wincompatible-library-redeclaration]\r\n int catanhf (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosf' [-Wincompatible-library-redeclaration]\r\n int ccosf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshf' [-Wincompatible-library-redeclaration]\r\n int ccoshf (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpf' [-Wincompatible-library-redeclaration]\r\n int cexpf (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagf' [-Wincompatible-library-redeclaration]\r\n int cimagf (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogf' [-Wincompatible-library-redeclaration]\r\n int clogf (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjf' [-Wincompatible-library-redeclaration]\r\n int conjf (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowf' [-Wincompatible-library-redeclaration]\r\n int cpowf (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowf' is a builtin with type '_Complex float (_Complex float, _Complex float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojf' [-Wincompatible-library-redeclaration]\r\n int cprojf (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'crealf' [-Wincompatible-library-redeclaration]\r\n int crealf (void);\r\n ^\r\n _configtest.c:17:5: note: 'crealf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinf' [-Wincompatible-library-redeclaration]\r\n int csinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhf' [-Wincompatible-library-redeclaration]\r\n int csinhf (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtf' [-Wincompatible-library-redeclaration]\r\n int csqrtf (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanf' [-Wincompatible-library-redeclaration]\r\n int ctanf (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhf' [-Wincompatible-library-redeclaration]\r\n int ctanhf (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhf' is a builtin with type '_Complex float (_Complex float)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsl' [-Wincompatible-library-redeclaration]\r\n int cabsl (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosl' [-Wincompatible-library-redeclaration]\r\n int cacosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshl' [-Wincompatible-library-redeclaration]\r\n int cacoshl (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargl' [-Wincompatible-library-redeclaration]\r\n int cargl (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinl' [-Wincompatible-library-redeclaration]\r\n int casinl (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhl' [-Wincompatible-library-redeclaration]\r\n int casinhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanl' [-Wincompatible-library-redeclaration]\r\n int catanl (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhl' [-Wincompatible-library-redeclaration]\r\n int catanhl (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosl' [-Wincompatible-library-redeclaration]\r\n int ccosl (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshl' [-Wincompatible-library-redeclaration]\r\n int ccoshl (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpl' [-Wincompatible-library-redeclaration]\r\n int cexpl (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagl' [-Wincompatible-library-redeclaration]\r\n int cimagl (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogl' [-Wincompatible-library-redeclaration]\r\n int clogl (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjl' [-Wincompatible-library-redeclaration]\r\n int conjl (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowl' [-Wincompatible-library-redeclaration]\r\n int cpowl (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowl' is a builtin with type '_Complex long double (_Complex long double, _Complex long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojl' [-Wincompatible-library-redeclaration]\r\n int cprojl (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creall' [-Wincompatible-library-redeclaration]\r\n int creall (void);\r\n ^\r\n _configtest.c:17:5: note: 'creall' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinl' [-Wincompatible-library-redeclaration]\r\n int csinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhl' [-Wincompatible-library-redeclaration]\r\n int csinhl (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtl' [-Wincompatible-library-redeclaration]\r\n int csqrtl (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanl' [-Wincompatible-library-redeclaration]\r\n int ctanl (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhl' [-Wincompatible-library-redeclaration]\r\n int ctanhl (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:2:12: warning: unused function 'static_func' [-Wunused-function]\r\n static int static_func (char * restrict a)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:3:19: warning: unused function 'static_func' [-Wunused-function]\r\n static inline int static_func (void)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h\r\n #define SIZEOF_PY_INTPTR_T 8\r\n #define SIZEOF_OFF_T 8\r\n #define SIZEOF_PY_LONG_LONG 8\r\n #define MATHLIB\r\n #define HAVE_SIN 1\r\n #define HAVE_COS 1\r\n #define HAVE_TAN 1\r\n #define HAVE_SINH 1\r\n #define HAVE_COSH 1\r\n #define HAVE_TANH 1\r\n #define HAVE_FABS 1\r\n #define HAVE_FLOOR 1\r\n #define HAVE_CEIL 1\r\n #define HAVE_SQRT 1\r\n #define HAVE_LOG10 1\r\n #define HAVE_LOG 1\r\n #define HAVE_EXP 1\r\n #define HAVE_ASIN 1\r\n #define HAVE_ACOS 1\r\n #define HAVE_ATAN 1\r\n #define HAVE_FMOD 1\r\n #define HAVE_MODF 1\r\n #define HAVE_FREXP 1\r\n #define HAVE_LDEXP 1\r\n #define HAVE_RINT 1\r\n #define HAVE_TRUNC 1\r\n #define HAVE_EXP2 1\r\n #define HAVE_LOG2 1\r\n #define HAVE_ATAN2 1\r\n #define HAVE_POW 1\r\n #define HAVE_NEXTAFTER 1\r\n #define HAVE_STRTOLL 1\r\n #define HAVE_STRTOULL 1\r\n #define HAVE_CBRT 1\r\n #define HAVE_STRTOLD_L 1\r\n #define HAVE_BACKTRACE 1\r\n #define HAVE_MADVISE 1\r\n #define HAVE_XMMINTRIN_H 1\r\n #define HAVE_EMMINTRIN_H 1\r\n #define HAVE_XLOCALE_H 1\r\n #define HAVE_DLFCN_H 1\r\n #define HAVE_SYS_MMAN_H 1\r\n #define HAVE___BUILTIN_ISNAN 1\r\n #define HAVE___BUILTIN_ISINF 1\r\n #define HAVE___BUILTIN_ISFINITE 1\r\n #define HAVE___BUILTIN_BSWAP32 1\r\n #define HAVE___BUILTIN_BSWAP64 1\r\n #define HAVE___BUILTIN_EXPECT 1\r\n #define HAVE___BUILTIN_MUL_OVERFLOW 1\r\n #define HAVE___BUILTIN_CPU_SUPPORTS 1\r\n #define HAVE__M_FROM_INT64 1\r\n #define HAVE__MM_LOAD_PS 1\r\n #define HAVE__MM_PREFETCH 1\r\n #define HAVE__MM_LOAD_PD 1\r\n #define HAVE___BUILTIN_PREFETCH 1\r\n #define HAVE_LINK_AVX 1\r\n #define HAVE_LINK_AVX2 1\r\n #define HAVE_XGETBV 1\r\n #define HAVE_ATTRIBUTE_NONNULL 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX2 1\r\n #define HAVE___THREAD 1\r\n #define HAVE_SINF 1\r\n #define HAVE_COSF 1\r\n #define HAVE_TANF 1\r\n #define HAVE_SINHF 1\r\n #define HAVE_COSHF 1\r\n #define HAVE_TANHF 1\r\n #define HAVE_FABSF 1\r\n #define HAVE_FLOORF 1\r\n #define HAVE_CEILF 1\r\n #define HAVE_RINTF 1\r\n #define HAVE_TRUNCF 1\r\n #define HAVE_SQRTF 1\r\n #define HAVE_LOG10F 1\r\n #define HAVE_LOGF 1\r\n #define HAVE_LOG1PF 1\r\n #define HAVE_EXPF 1\r\n #define HAVE_EXPM1F 1\r\n #define HAVE_ASINF 1\r\n #define HAVE_ACOSF 1\r\n #define HAVE_ATANF 1\r\n #define HAVE_ASINHF 1\r\n #define HAVE_ACOSHF 1\r\n #define HAVE_ATANHF 1\r\n #define HAVE_HYPOTF 1\r\n #define HAVE_ATAN2F 1\r\n #define HAVE_POWF 1\r\n #define HAVE_FMODF 1\r\n #define HAVE_MODFF 1\r\n #define HAVE_FREXPF 1\r\n #define HAVE_LDEXPF 1\r\n #define HAVE_EXP2F 1\r\n #define HAVE_LOG2F 1\r\n #define HAVE_COPYSIGNF 1\r\n #define HAVE_NEXTAFTERF 1\r\n #define HAVE_CBRTF 1\r\n #define HAVE_SINL 1\r\n #define HAVE_COSL 1\r\n #define HAVE_TANL 1\r\n #define HAVE_SINHL 1\r\n #define HAVE_COSHL 1\r\n #define HAVE_TANHL 1\r\n #define HAVE_FABSL 1\r\n #define HAVE_FLOORL 1\r\n #define HAVE_CEILL 1\r\n #define HAVE_RINTL 1\r\n #define HAVE_TRUNCL 1\r\n #define HAVE_SQRTL 1\r\n #define HAVE_LOG10L 1\r\n #define HAVE_LOGL 1\r\n #define HAVE_LOG1PL 1\r\n #define HAVE_EXPL 1\r\n #define HAVE_EXPM1L 1\r\n #define HAVE_ASINL 1\r\n #define HAVE_ACOSL 1\r\n #define HAVE_ATANL 1\r\n #define HAVE_ASINHL 1\r\n #define HAVE_ACOSHL 1\r\n #define HAVE_ATANHL 1\r\n #define HAVE_HYPOTL 1\r\n #define HAVE_ATAN2L 1\r\n #define HAVE_POWL 1\r\n #define HAVE_FMODL 1\r\n #define HAVE_MODFL 1\r\n #define HAVE_FREXPL 1\r\n #define HAVE_LDEXPL 1\r\n #define HAVE_EXP2L 1\r\n #define HAVE_LOG2L 1\r\n #define HAVE_COPYSIGNL 1\r\n #define HAVE_NEXTAFTERL 1\r\n #define HAVE_CBRTL 1\r\n #define HAVE_DECL_SIGNBIT\r\n #define HAVE_COMPLEX_H 1\r\n #define HAVE_CABS 1\r\n #define HAVE_CACOS 1\r\n #define HAVE_CACOSH 1\r\n #define HAVE_CARG 1\r\n #define HAVE_CASIN 1\r\n #define HAVE_CASINH 1\r\n #define HAVE_CATAN 1\r\n #define HAVE_CATANH 1\r\n #define HAVE_CCOS 1\r\n #define HAVE_CCOSH 1\r\n #define HAVE_CEXP 1\r\n #define HAVE_CIMAG 1\r\n #define HAVE_CLOG 1\r\n #define HAVE_CONJ 1\r\n #define HAVE_CPOW 1\r\n #define HAVE_CPROJ 1\r\n #define HAVE_CREAL 1\r\n #define HAVE_CSIN 1\r\n #define HAVE_CSINH 1\r\n #define HAVE_CSQRT 1\r\n #define HAVE_CTAN 1\r\n #define HAVE_CTANH 1\r\n #define HAVE_CABSF 1\r\n #define HAVE_CACOSF 1\r\n #define HAVE_CACOSHF 1\r\n #define HAVE_CARGF 1\r\n #define HAVE_CASINF 1\r\n #define HAVE_CASINHF 1\r\n #define HAVE_CATANF 1\r\n #define HAVE_CATANHF 1\r\n #define HAVE_CCOSF 1\r\n #define HAVE_CCOSHF 1\r\n #define HAVE_CEXPF 1\r\n #define HAVE_CIMAGF 1\r\n #define HAVE_CLOGF 1\r\n #define HAVE_CONJF 1\r\n #define HAVE_CPOWF 1\r\n #define HAVE_CPROJF 1\r\n #define HAVE_CREALF 1\r\n #define HAVE_CSINF 1\r\n #define HAVE_CSINHF 1\r\n #define HAVE_CSQRTF 1\r\n #define HAVE_CTANF 1\r\n #define HAVE_CTANHF 1\r\n #define HAVE_CABSL 1\r\n #define HAVE_CACOSL 1\r\n #define HAVE_CACOSHL 1\r\n #define HAVE_CARGL 1\r\n #define HAVE_CASINL 1\r\n #define HAVE_CASINHL 1\r\n #define HAVE_CATANL 1\r\n #define HAVE_CATANHL 1\r\n #define HAVE_CCOSL 1\r\n #define HAVE_CCOSHL 1\r\n #define HAVE_CEXPL 1\r\n #define HAVE_CIMAGL 1\r\n #define HAVE_CLOGL 1\r\n #define HAVE_CONJL 1\r\n #define HAVE_CPOWL 1\r\n #define HAVE_CPROJL 1\r\n #define HAVE_CREALL 1\r\n #define HAVE_CSINL 1\r\n #define HAVE_CSINHL 1\r\n #define HAVE_CSQRTL 1\r\n #define HAVE_CTANL 1\r\n #define HAVE_CTANHL 1\r\n #define NPY_RESTRICT restrict\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define HAVE_LDOUBLE_INTEL_EXTENDED_16_BYTES_LE 1\r\n #define NPY_PY3K 1\r\n #ifndef __cplusplus\r\n \/* #undef inline *\/\r\n #endif\r\n \r\n #ifndef _NPY_NPY_CONFIG_H_\r\n #error config.h should never be included directly, include npy_config.h instead\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n Generating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'exp' [-Wincompatible-library-redeclaration]\r\n int exp (void);\r\n ^\r\n _configtest.c:1:5: note: 'exp' is a builtin with type 'double (double)'\r\n 1 warning generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n #define NPY_SIZEOF_SHORT SIZEOF_SHORT\r\n #define NPY_SIZEOF_INT SIZEOF_INT\r\n #define NPY_SIZEOF_LONG SIZEOF_LONG\r\n #define NPY_SIZEOF_FLOAT 4\r\n #define NPY_SIZEOF_COMPLEX_FLOAT 8\r\n #define NPY_SIZEOF_DOUBLE 8\r\n #define NPY_SIZEOF_COMPLEX_DOUBLE 16\r\n #define NPY_SIZEOF_LONGDOUBLE 16\r\n #define NPY_SIZEOF_COMPLEX_LONGDOUBLE 32\r\n #define NPY_SIZEOF_PY_INTPTR_T 8\r\n #define NPY_SIZEOF_OFF_T 8\r\n #define NPY_SIZEOF_PY_LONG_LONG 8\r\n #define NPY_SIZEOF_LONGLONG 8\r\n #define NPY_NO_SMP 0\r\n #define NPY_HAVE_DECL_ISNAN\r\n #define NPY_HAVE_DECL_ISINF\r\n #define NPY_HAVE_DECL_ISFINITE\r\n #define NPY_HAVE_DECL_SIGNBIT\r\n #define NPY_USE_C99_COMPLEX 1\r\n #define NPY_HAVE_COMPLEX_DOUBLE 1\r\n #define NPY_HAVE_COMPLEX_FLOAT 1\r\n #define NPY_HAVE_COMPLEX_LONG_DOUBLE 1\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define NPY_USE_C99_FORMATS 1\r\n #define NPY_VISIBILITY_HIDDEN __attribute__((visibility(\"hidden\")))\r\n #define NPY_ABI_VERSION 0x01000009\r\n #define NPY_API_VERSION 0x0000000D\r\n \r\n #ifndef __STDC_FORMAT_MACROS\r\n #define __STDC_FORMAT_MACROS 1\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h']\r\n building extension \"numpy.core._multiarray_tests\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n building extension \"numpy.core._multiarray_umath\" sources\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_ufunc_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common' to include_dirs.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_internal.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h']\r\n building extension \"numpy.core._umath_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n building extension \"numpy.core._rational_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n building extension \"numpy.core._struct_ufunc_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n building extension \"numpy.core._operand_flag_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n building extension \"numpy.fft.fftpack_lite\" sources\r\n building extension \"numpy.linalg.lapack_lite\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n building extension \"numpy.linalg._umath_linalg\" sources\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n building extension \"numpy.random.mtrand\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/random\r\n building data_files sources\r\n build_src: building npy-pkg config files\r\n running build_py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/conftest.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_globals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/dual.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_distributor_init.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/ctypeslib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/matlib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_pytesttester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/py3k.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/_inspect.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/fromnumeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_add_newdocs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_methods.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_internal.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_string_helpers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/multiarray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/records.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup_common.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_aliased_types.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/memmap.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/overrides.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/getlimits.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype_ctypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/defchararray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/machar.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/einsumfunc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath_tests.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numerictypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_type_aliases.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/cversions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/arrayprint.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/code_generators\/generate_numpy_api.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/unixccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/numpy_distribution.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/conv_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/cpuinfo.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvc9compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/npy_pkg_config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/compat.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/misc_util.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/log.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/line_endings.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/lib2def.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/pathccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/system_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/exec_command.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/from_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/mingw32ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/extension.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/intelccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/distutils\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config_compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_ext.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_headers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_py.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_src.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/sdist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_scripts.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/bdist_rpm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/autodist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/egg_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/develop.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_data.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/gnu.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/compaq.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/intel.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/none.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/nag.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/ibm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/sun.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/lahey.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/g95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/mips.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/hpux.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/environment.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pathf95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/absoft.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/vast.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/misc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/internals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/creation.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/constants.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/ufuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/broadcasting.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/basics.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/subclassing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/indexing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/byteswapping.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/structured_arrays.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/glossary.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/common_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/crackfortran.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cb_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py2e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/func2subr.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/diagnose.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/capi_maps.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f90mod_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py_testing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/use_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/auxfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__main__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/helper.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/fftpack.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_iotools.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/mixins.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/nanfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/recfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/histograms.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/scimath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/user_array.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/format.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/twodim_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/financial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/index_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/npyio.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/stride_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arrayterator.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraysetops.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraypad.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/type_check.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_datasource.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/ufunclike.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/linalg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/extras.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/testutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/bench.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/timer_comparison.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/mrecords.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/defmatrix.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/laguerre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/_polybase.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polyutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite_e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/chebyshev.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/legendre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/print_coercion_tables.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/parameterized.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n running build_clib\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_clib\r\n building 'npymath' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n ar: adding 4 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n building 'npysort' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/quicksort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/mergesort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/heapsort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/selection.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/binsearch.c\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n 22 warnings generated.\r\n ar: adding 5 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n running build_ext\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_ext\r\n building 'numpy.core._dummy' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/dummymodule.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/dummymodule.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_dummy.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_tests.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_umath' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/core\/src\/multiarray\/alloc.c\r\n clang: numpy\/core\/src\/multiarray\/calculation.cclang: numpy\/core\/src\/multiarray\/array_assign_scalar.c\r\n clang: numpy\/core\/src\/multiarray\/convert.c\r\n \r\n clang: numpy\/core\/src\/multiarray\/ctors.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busday.c\r\n clang: numpy\/core\/src\/multiarray\/dragon4.cclang: numpy\/core\/src\/multiarray\/flagsobject.c\r\n \r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/arrayobject.c\r\n clang: numpy\/core\/src\/multiarray\/array_assign_array.c\r\n clang: numpy\/core\/src\/multiarray\/convert_datatype.c\r\n clang: numpy\/core\/src\/multiarray\/getset.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busdaycal.c\r\n clang: numpy\/core\/src\/multiarray\/buffer.c\r\n clang: numpy\/core\/src\/multiarray\/compiled_base.c\r\n clang: numpy\/core\/src\/multiarray\/hashdescr.c\r\n clang: numpy\/core\/src\/multiarray\/descriptor.c\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/conversion_utils.c\r\n clang: numpy\/core\/src\/multiarray\/item_selection.c\r\n clang: numpy\/core\/src\/multiarray\/dtype_transfer.c\r\n clang: numpy\/core\/src\/multiarray\/mapping.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n 3 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/datetime.c\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:477:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ptr = PyUnicode_AS_UNICODE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/common.c\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 6 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/nditer_pywrap.c\r\n 9 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/sequence.c\r\n clang: numpy\/core\/src\/multiarray\/shape.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n clang: numpy\/core\/src\/multiarray\/methods.c\r\n clang: numpy\/core\/src\/multiarray\/iterators.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_strings.c\r\n clang: numpy\/core\/src\/multiarray\/number.c\r\n clang: numpy\/core\/src\/multiarray\/scalarapi.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n numpy\/core\/src\/multiarray\/scalarapi.c:74:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:135:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1849:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buffer = PyUnicode_AS_DATA(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 5 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/typeinfo.c\r\n clang: numpy\/core\/src\/multiarray\/refcount.c\r\n clang: numpy\/core\/src\/multiarray\/usertypes.c\r\n clang: numpy\/core\/src\/multiarray\/multiarraymodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n clang: numpy\/core\/src\/multiarray\/vdot.c\r\n clang: numpy\/core\/src\/umath\/umathmodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n clang: numpy\/core\/src\/umath\/reduction.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n clang: numpy\/core\/src\/multiarray\/nditer_api.c\r\n 14 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/strfuncs.c\r\n numpy\/core\/src\/umath\/loops.c.src:655:18: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n result = PyEval_CallObject(tocall, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:178:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_ReprFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:195:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_StrFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 2 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/temp_elide.c\r\n clang: numpy\/core\/src\/umath\/cpuid.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n clang: numpy\/core\/src\/umath\/ufunc_object.c\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'byte_long' [-Wunused-function]\r\n byte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ubyte_long' [-Wunused-function]\r\n ubyte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'short_long' [-Wunused-function]\r\n short_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ushort_long' [-Wunused-function]\r\n ushort_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'int_long' [-Wunused-function]\r\n int_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'uint_long' [-Wunused-function]\r\n uint_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'long_long' [-Wunused-function]\r\n long_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulong_long' [-Wunused-function]\r\n ulong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longlong_long' [-Wunused-function]\r\n longlong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulonglong_long' [-Wunused-function]\r\n ulonglong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'half_long' [-Wunused-function]\r\n half_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'float_long' [-Wunused-function]\r\n float_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'double_long' [-Wunused-function]\r\n double_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longdouble_long' [-Wunused-function]\r\n longdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cfloat_long' [-Wunused-function]\r\n cfloat_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cdouble_long' [-Wunused-function]\r\n cdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'clongdouble_long' [-Wunused-function]\r\n clongdouble_long(PyObject *obj)\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/nditer_constr.c\r\n numpy\/core\/src\/umath\/ufunc_object.c:657:19: warning: comparison of integers of different signs: 'int' and 'size_t' (aka 'unsigned long') [-Wsign-compare]\r\n for (i = 0; i < len; i++) {\r\n ~ ^ ~~~\r\n clang: numpy\/core\/src\/umath\/override.c\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n clang: numpy\/core\/src\/common\/array_assign.c\r\n clang: numpy\/core\/src\/common\/ufunc_override.c\r\n clang: numpy\/core\/src\/common\/npy_longdouble.c\r\n clang: numpy\/core\/src\/common\/numpyos.c\r\n clang: numpy\/core\/src\/common\/ucsnarrow.c\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/umath\/extobj.c\r\n numpy\/core\/src\/common\/ucsnarrow.c:139:34: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n ret = (PyUnicodeObject *)PyUnicode_FromUnicode((Py_UNICODE*)buf,\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/common\/python_xerbla.c\r\n clang: numpy\/core\/src\/common\/cblasfuncs.c\r\n clang: \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c\r\n In file included from \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c:26:\r\n In file included from numpy\/core\/include\/numpy\/arrayobject.h:4:\r\n In file included from numpy\/core\/include\/numpy\/ndarrayobject.h:21:\r\n build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h:1463:1: warning: unused function '_import_array' [-Wunused-function]\r\n _import_array(void)\r\n ^\r\n 1 warning generated.\r\n 17 warnings generated.\r\n clang: numpy\/core\/src\/umath\/ufunc_type_resolution.c\r\n 4 warnings generated.\r\n 4 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/alloc.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arrayobject.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_scalar.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_array.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/buffer.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/calculation.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/compiled_base.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/common.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert_datatype.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/conversion_utils.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/ctors.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_strings.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busday.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busdaycal.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/descriptor.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dragon4.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dtype_transfer.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/flagsobject.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/getset.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/hashdescr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/item_selection.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/iterators.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/mapping.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/methods.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/multiarraymodule.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_api.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_constr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_pywrap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/number.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/refcount.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/sequence.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/shape.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalarapi.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/strfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/temp_elide.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/typeinfo.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/usertypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/vdot.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/umathmodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/reduction.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_object.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/extobj.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/cpuid.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_type_resolution.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/halffloat.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/array_assign.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/npy_longdouble.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ucsnarrow.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ufunc_override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/numpyos.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/cblasfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/python_xerbla.o build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -lnpysort -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_umath.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.core._umath_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_umath_tests.cpython-39-darwin.so\r\n building 'numpy.core._rational_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_rational_tests.cpython-39-darwin.so\r\n building 'numpy.core._struct_ufunc_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_struct_ufunc_tests.cpython-39-darwin.so\r\n building 'numpy.core._operand_flag_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_operand_flag_tests.cpython-39-darwin.so\r\n building 'numpy.fft.fftpack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n compile options: '-Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/fft\/fftpack_litemodule.c\r\n clang: numpy\/fft\/fftpack.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_lite.cpython-39-darwin.so\r\n building 'numpy.linalg.lapack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/linalg\/lapack_litemodule.c\r\n clang: numpy\/linalg\/lapack_lite\/python_xerbla.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.linalg._umath_linalg' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n numpy\/linalg\/umath_linalg.c.src:735:32: warning: unknown warning group '-Wmaybe-uninitialized', ignored [-Wunknown-warning-option]\r\n #pragma GCC diagnostic ignored \"-Wmaybe-uninitialized\"\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:541:1: warning: unused function 'dump_ufunc_object' [-Wunused-function]\r\n dump_ufunc_object(PyUFuncObject* ufunc)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:566:1: warning: unused function 'dump_linearize_data' [-Wunused-function]\r\n dump_linearize_data(const char* name, const LINEARIZE_DATA_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_FLOAT_matrix' [-Wunused-function]\r\n dump_FLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_DOUBLE_matrix' [-Wunused-function]\r\n dump_DOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CFLOAT_matrix' [-Wunused-function]\r\n dump_CFLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CDOUBLE_matrix' [-Wunused-function]\r\n dump_CDOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_FLOAT_matrix' [-Wunused-function]\r\n zero_FLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_DOUBLE_matrix' [-Wunused-function]\r\n zero_DOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CFLOAT_matrix' [-Wunused-function]\r\n zero_CFLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CDOUBLE_matrix' [-Wunused-function]\r\n zero_CDOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:1862:1: warning: unused function 'dump_geev_params' [-Wunused-function]\r\n dump_geev_params(const char *name, GEEV_PARAMS_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2132:1: warning: unused function 'init_cgeev' [-Wunused-function]\r\n init_cgeev(GEEV_PARAMS_t* params,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2213:1: warning: unused function 'process_cgeev_results' [-Wunused-function]\r\n process_cgeev_results(GEEV_PARAMS_t *NPY_UNUSED(params))\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2376:1: warning: unused function 'dump_gesdd_params' [-Wunused-function]\r\n dump_gesdd_params(const char *name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2864:1: warning: unused function 'dump_gelsd_params' [-Wunused-function]\r\n dump_gelsd_params(const char *name,\r\n ^\r\n 16 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/_umath_linalg.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.random.mtrand' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\r\n compile options: '-D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/random\/mtrand\/mtrand.c\r\n clang: numpy\/random\/mtrand\/initarray.cclang: numpy\/random\/mtrand\/randomkit.c\r\n \r\n clang: numpy\/random\/mtrand\/distributions.c\r\n numpy\/random\/mtrand\/mtrand.c:40400:34: error: no member named 'tp_print' in 'struct _typeobject'\r\n __pyx_type_6mtrand_RandomState.tp_print = 0;\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 12 warnings and 1 error generated.\r\n error: Command \"clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c numpy\/random\/mtrand\/mtrand.c -o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o -MMD -MF build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o.d\" failed with exit status 1","comment_length":12,"text":"Unable to install datasets\n** Edit **\r\nI believe there's a bug with the package when you're installing it with Python 3.9. I recommend sticking with previous versions. Thanks, @thomwolf for the insight! \r\n\r\n**Short description**\r\n\r\nI followed the instructions for installing datasets (https:\/\/huggingface.co\/docs\/datasets\/installation.html). However, while I tried to download datasets using `pip install datasets` I got a massive error message after getting stuck at \"Installing build dependencies...\" \r\n\r\nI was wondering if this problem can be fixed by creating a virtual environment, but it didn't help. Can anyone offer some advice on how to fix this issue? \r\n\r\nHere's an error message: \r\n\r\n`(env) Gas-MacBook-Pro:Downloads destiny$ pip install datasets\r\nCollecting datasets\r\n Using cached datasets-1.2.0-py3-none-any.whl (159 kB)\r\nCollecting numpy>=1.17\r\n Using cached numpy-1.19.5-cp39-cp39-macosx_10_9_x86_64.whl (15.6 MB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-2.0.0.tar.gz (58.9 MB)\r\n....\r\n\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceilf' [-Wincompatible-library-redeclaration]\r\n int ceilf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceilf' is a builtin with type 'float (float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintf' [-Wincompatible-library-redeclaration]\r\n int rintf (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintf' is a builtin with type 'float (float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncf' [-Wincompatible-library-redeclaration]\r\n int truncf (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncf' is a builtin with type 'float (float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtf' [-Wincompatible-library-redeclaration]\r\n int sqrtf (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtf' is a builtin with type 'float (float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10f' [-Wincompatible-library-redeclaration]\r\n int log10f (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10f' is a builtin with type 'float (float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logf' [-Wincompatible-library-redeclaration]\r\n int logf (void);\r\n ^\r\n _configtest.c:14:5: note: 'logf' is a builtin with type 'float (float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pf' [-Wincompatible-library-redeclaration]\r\n int log1pf (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pf' is a builtin with type 'float (float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expf' [-Wincompatible-library-redeclaration]\r\n int expf (void);\r\n ^\r\n _configtest.c:16:5: note: 'expf' is a builtin with type 'float (float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1f' [-Wincompatible-library-redeclaration]\r\n int expm1f (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1f' is a builtin with type 'float (float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinf' [-Wincompatible-library-redeclaration]\r\n int asinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinf' is a builtin with type 'float (float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosf' [-Wincompatible-library-redeclaration]\r\n int acosf (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosf' is a builtin with type 'float (float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanf' [-Wincompatible-library-redeclaration]\r\n int atanf (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanf' is a builtin with type 'float (float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhf' [-Wincompatible-library-redeclaration]\r\n int asinhf (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhf' is a builtin with type 'float (float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshf' [-Wincompatible-library-redeclaration]\r\n int acoshf (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshf' is a builtin with type 'float (float)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhf' [-Wincompatible-library-redeclaration]\r\n int atanhf (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhf' is a builtin with type 'float (float)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotf' [-Wincompatible-library-redeclaration]\r\n int hypotf (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotf' is a builtin with type 'float (float, float)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2f' [-Wincompatible-library-redeclaration]\r\n int atan2f (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2f' is a builtin with type 'float (float, float)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powf' [-Wincompatible-library-redeclaration]\r\n int powf (void);\r\n ^\r\n _configtest.c:26:5: note: 'powf' is a builtin with type 'float (float, float)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodf' [-Wincompatible-library-redeclaration]\r\n int fmodf (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodf' is a builtin with type 'float (float, float)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modff' [-Wincompatible-library-redeclaration]\r\n int modff (void);\r\n ^\r\n _configtest.c:28:5: note: 'modff' is a builtin with type 'float (float, float *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpf' [-Wincompatible-library-redeclaration]\r\n int frexpf (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpf' is a builtin with type 'float (float, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpf' [-Wincompatible-library-redeclaration]\r\n int ldexpf (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpf' is a builtin with type 'float (float, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2f' [-Wincompatible-library-redeclaration]\r\n int exp2f (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2f' is a builtin with type 'float (float)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2f' [-Wincompatible-library-redeclaration]\r\n int log2f (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2f' is a builtin with type 'float (float)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignf' [-Wincompatible-library-redeclaration]\r\n int copysignf (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignf' is a builtin with type 'float (float, float)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterf' [-Wincompatible-library-redeclaration]\r\n int nextafterf (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterf' is a builtin with type 'float (float, float)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtf' [-Wincompatible-library-redeclaration]\r\n int cbrtf (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtf' is a builtin with type 'float (float)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'sinl' [-Wincompatible-library-redeclaration]\r\n int sinl (void);\r\n ^\r\n _configtest.c:1:5: note: 'sinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cosl' [-Wincompatible-library-redeclaration]\r\n int cosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'tanl' [-Wincompatible-library-redeclaration]\r\n int tanl (void);\r\n ^\r\n _configtest.c:3:5: note: 'tanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'sinhl' [-Wincompatible-library-redeclaration]\r\n int sinhl (void);\r\n ^\r\n _configtest.c:4:5: note: 'sinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'coshl' [-Wincompatible-library-redeclaration]\r\n int coshl (void);\r\n ^\r\n _configtest.c:5:5: note: 'coshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'tanhl' [-Wincompatible-library-redeclaration]\r\n int tanhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'tanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'fabsl' [-Wincompatible-library-redeclaration]\r\n int fabsl (void);\r\n ^\r\n _configtest.c:7:5: note: 'fabsl' is a builtin with type 'long double (long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'floorl' [-Wincompatible-library-redeclaration]\r\n int floorl (void);\r\n ^\r\n _configtest.c:8:5: note: 'floorl' is a builtin with type 'long double (long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceill' [-Wincompatible-library-redeclaration]\r\n int ceill (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceill' is a builtin with type 'long double (long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintl' [-Wincompatible-library-redeclaration]\r\n int rintl (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintl' is a builtin with type 'long double (long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncl' [-Wincompatible-library-redeclaration]\r\n int truncl (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncl' is a builtin with type 'long double (long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtl' [-Wincompatible-library-redeclaration]\r\n int sqrtl (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtl' is a builtin with type 'long double (long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10l' [-Wincompatible-library-redeclaration]\r\n int log10l (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10l' is a builtin with type 'long double (long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logl' [-Wincompatible-library-redeclaration]\r\n int logl (void);\r\n ^\r\n _configtest.c:14:5: note: 'logl' is a builtin with type 'long double (long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pl' [-Wincompatible-library-redeclaration]\r\n int log1pl (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pl' is a builtin with type 'long double (long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expl' [-Wincompatible-library-redeclaration]\r\n int expl (void);\r\n ^\r\n _configtest.c:16:5: note: 'expl' is a builtin with type 'long double (long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1l' [-Wincompatible-library-redeclaration]\r\n int expm1l (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1l' is a builtin with type 'long double (long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinl' [-Wincompatible-library-redeclaration]\r\n int asinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosl' [-Wincompatible-library-redeclaration]\r\n int acosl (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanl' [-Wincompatible-library-redeclaration]\r\n int atanl (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhl' [-Wincompatible-library-redeclaration]\r\n int asinhl (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshl' [-Wincompatible-library-redeclaration]\r\n int acoshl (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhl' [-Wincompatible-library-redeclaration]\r\n int atanhl (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotl' [-Wincompatible-library-redeclaration]\r\n int hypotl (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2l' [-Wincompatible-library-redeclaration]\r\n int atan2l (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2l' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powl' [-Wincompatible-library-redeclaration]\r\n int powl (void);\r\n ^\r\n _configtest.c:26:5: note: 'powl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodl' [-Wincompatible-library-redeclaration]\r\n int fmodl (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modfl' [-Wincompatible-library-redeclaration]\r\n int modfl (void);\r\n ^\r\n _configtest.c:28:5: note: 'modfl' is a builtin with type 'long double (long double, long double *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpl' [-Wincompatible-library-redeclaration]\r\n int frexpl (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpl' is a builtin with type 'long double (long double, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpl' [-Wincompatible-library-redeclaration]\r\n int ldexpl (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpl' is a builtin with type 'long double (long double, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2l' [-Wincompatible-library-redeclaration]\r\n int exp2l (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2l' [-Wincompatible-library-redeclaration]\r\n int log2l (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignl' [-Wincompatible-library-redeclaration]\r\n int copysignl (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterl' [-Wincompatible-library-redeclaration]\r\n int nextafterl (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtl' [-Wincompatible-library-redeclaration]\r\n int cbrtl (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtl' is a builtin with type 'long double (long double)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:8:12: error: use of undeclared identifier 'HAVE_DECL_SIGNBIT'\r\n (void) HAVE_DECL_SIGNBIT;\r\n ^\r\n 1 error generated.\r\n failure.\r\n removing: _configtest.c _configtest.o\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabs' [-Wincompatible-library-redeclaration]\r\n int cabs (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabs' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacos' [-Wincompatible-library-redeclaration]\r\n int cacos (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacosh' [-Wincompatible-library-redeclaration]\r\n int cacosh (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'carg' [-Wincompatible-library-redeclaration]\r\n int carg (void);\r\n ^\r\n _configtest.c:4:5: note: 'carg' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casin' [-Wincompatible-library-redeclaration]\r\n int casin (void);\r\n ^\r\n _configtest.c:5:5: note: 'casin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinh' [-Wincompatible-library-redeclaration]\r\n int casinh (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catan' [-Wincompatible-library-redeclaration]\r\n int catan (void);\r\n ^\r\n _configtest.c:7:5: note: 'catan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanh' [-Wincompatible-library-redeclaration]\r\n int catanh (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccos' [-Wincompatible-library-redeclaration]\r\n int ccos (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccosh' [-Wincompatible-library-redeclaration]\r\n int ccosh (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexp' [-Wincompatible-library-redeclaration]\r\n int cexp (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexp' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimag' [-Wincompatible-library-redeclaration]\r\n int cimag (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimag' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clog' [-Wincompatible-library-redeclaration]\r\n int clog (void);\r\n ^\r\n _configtest.c:13:5: note: 'clog' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conj' [-Wincompatible-library-redeclaration]\r\n int conj (void);\r\n ^\r\n _configtest.c:14:5: note: 'conj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpow' [-Wincompatible-library-redeclaration]\r\n int cpow (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpow' is a builtin with type '_Complex double (_Complex double, _Complex double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cproj' [-Wincompatible-library-redeclaration]\r\n int cproj (void);\r\n ^\r\n _configtest.c:16:5: note: 'cproj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creal' [-Wincompatible-library-redeclaration]\r\n int creal (void);\r\n ^\r\n _configtest.c:17:5: note: 'creal' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csin' [-Wincompatible-library-redeclaration]\r\n int csin (void);\r\n ^\r\n _configtest.c:18:5: note: 'csin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinh' [-Wincompatible-library-redeclaration]\r\n int csinh (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrt' [-Wincompatible-library-redeclaration]\r\n int csqrt (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrt' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctan' [-Wincompatible-library-redeclaration]\r\n int ctan (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanh' [-Wincompatible-library-redeclaration]\r\n int ctanh (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanh' is a builtin with type '_Complex double (_Complex double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsf' [-Wincompatible-library-redeclaration]\r\n int cabsf (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosf' [-Wincompatible-library-redeclaration]\r\n int cacosf (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshf' [-Wincompatible-library-redeclaration]\r\n int cacoshf (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargf' [-Wincompatible-library-redeclaration]\r\n int cargf (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinf' [-Wincompatible-library-redeclaration]\r\n int casinf (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhf' [-Wincompatible-library-redeclaration]\r\n int casinhf (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanf' [-Wincompatible-library-redeclaration]\r\n int catanf (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhf' [-Wincompatible-library-redeclaration]\r\n int catanhf (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosf' [-Wincompatible-library-redeclaration]\r\n int ccosf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshf' [-Wincompatible-library-redeclaration]\r\n int ccoshf (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpf' [-Wincompatible-library-redeclaration]\r\n int cexpf (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagf' [-Wincompatible-library-redeclaration]\r\n int cimagf (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogf' [-Wincompatible-library-redeclaration]\r\n int clogf (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjf' [-Wincompatible-library-redeclaration]\r\n int conjf (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowf' [-Wincompatible-library-redeclaration]\r\n int cpowf (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowf' is a builtin with type '_Complex float (_Complex float, _Complex float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojf' [-Wincompatible-library-redeclaration]\r\n int cprojf (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'crealf' [-Wincompatible-library-redeclaration]\r\n int crealf (void);\r\n ^\r\n _configtest.c:17:5: note: 'crealf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinf' [-Wincompatible-library-redeclaration]\r\n int csinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhf' [-Wincompatible-library-redeclaration]\r\n int csinhf (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtf' [-Wincompatible-library-redeclaration]\r\n int csqrtf (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanf' [-Wincompatible-library-redeclaration]\r\n int ctanf (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhf' [-Wincompatible-library-redeclaration]\r\n int ctanhf (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhf' is a builtin with type '_Complex float (_Complex float)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsl' [-Wincompatible-library-redeclaration]\r\n int cabsl (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosl' [-Wincompatible-library-redeclaration]\r\n int cacosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshl' [-Wincompatible-library-redeclaration]\r\n int cacoshl (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargl' [-Wincompatible-library-redeclaration]\r\n int cargl (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinl' [-Wincompatible-library-redeclaration]\r\n int casinl (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhl' [-Wincompatible-library-redeclaration]\r\n int casinhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanl' [-Wincompatible-library-redeclaration]\r\n int catanl (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhl' [-Wincompatible-library-redeclaration]\r\n int catanhl (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosl' [-Wincompatible-library-redeclaration]\r\n int ccosl (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshl' [-Wincompatible-library-redeclaration]\r\n int ccoshl (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpl' [-Wincompatible-library-redeclaration]\r\n int cexpl (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagl' [-Wincompatible-library-redeclaration]\r\n int cimagl (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogl' [-Wincompatible-library-redeclaration]\r\n int clogl (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjl' [-Wincompatible-library-redeclaration]\r\n int conjl (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowl' [-Wincompatible-library-redeclaration]\r\n int cpowl (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowl' is a builtin with type '_Complex long double (_Complex long double, _Complex long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojl' [-Wincompatible-library-redeclaration]\r\n int cprojl (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creall' [-Wincompatible-library-redeclaration]\r\n int creall (void);\r\n ^\r\n _configtest.c:17:5: note: 'creall' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinl' [-Wincompatible-library-redeclaration]\r\n int csinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhl' [-Wincompatible-library-redeclaration]\r\n int csinhl (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtl' [-Wincompatible-library-redeclaration]\r\n int csqrtl (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanl' [-Wincompatible-library-redeclaration]\r\n int ctanl (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhl' [-Wincompatible-library-redeclaration]\r\n int ctanhl (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:2:12: warning: unused function 'static_func' [-Wunused-function]\r\n static int static_func (char * restrict a)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:3:19: warning: unused function 'static_func' [-Wunused-function]\r\n static inline int static_func (void)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h\r\n #define SIZEOF_PY_INTPTR_T 8\r\n #define SIZEOF_OFF_T 8\r\n #define SIZEOF_PY_LONG_LONG 8\r\n #define MATHLIB\r\n #define HAVE_SIN 1\r\n #define HAVE_COS 1\r\n #define HAVE_TAN 1\r\n #define HAVE_SINH 1\r\n #define HAVE_COSH 1\r\n #define HAVE_TANH 1\r\n #define HAVE_FABS 1\r\n #define HAVE_FLOOR 1\r\n #define HAVE_CEIL 1\r\n #define HAVE_SQRT 1\r\n #define HAVE_LOG10 1\r\n #define HAVE_LOG 1\r\n #define HAVE_EXP 1\r\n #define HAVE_ASIN 1\r\n #define HAVE_ACOS 1\r\n #define HAVE_ATAN 1\r\n #define HAVE_FMOD 1\r\n #define HAVE_MODF 1\r\n #define HAVE_FREXP 1\r\n #define HAVE_LDEXP 1\r\n #define HAVE_RINT 1\r\n #define HAVE_TRUNC 1\r\n #define HAVE_EXP2 1\r\n #define HAVE_LOG2 1\r\n #define HAVE_ATAN2 1\r\n #define HAVE_POW 1\r\n #define HAVE_NEXTAFTER 1\r\n #define HAVE_STRTOLL 1\r\n #define HAVE_STRTOULL 1\r\n #define HAVE_CBRT 1\r\n #define HAVE_STRTOLD_L 1\r\n #define HAVE_BACKTRACE 1\r\n #define HAVE_MADVISE 1\r\n #define HAVE_XMMINTRIN_H 1\r\n #define HAVE_EMMINTRIN_H 1\r\n #define HAVE_XLOCALE_H 1\r\n #define HAVE_DLFCN_H 1\r\n #define HAVE_SYS_MMAN_H 1\r\n #define HAVE___BUILTIN_ISNAN 1\r\n #define HAVE___BUILTIN_ISINF 1\r\n #define HAVE___BUILTIN_ISFINITE 1\r\n #define HAVE___BUILTIN_BSWAP32 1\r\n #define HAVE___BUILTIN_BSWAP64 1\r\n #define HAVE___BUILTIN_EXPECT 1\r\n #define HAVE___BUILTIN_MUL_OVERFLOW 1\r\n #define HAVE___BUILTIN_CPU_SUPPORTS 1\r\n #define HAVE__M_FROM_INT64 1\r\n #define HAVE__MM_LOAD_PS 1\r\n #define HAVE__MM_PREFETCH 1\r\n #define HAVE__MM_LOAD_PD 1\r\n #define HAVE___BUILTIN_PREFETCH 1\r\n #define HAVE_LINK_AVX 1\r\n #define HAVE_LINK_AVX2 1\r\n #define HAVE_XGETBV 1\r\n #define HAVE_ATTRIBUTE_NONNULL 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX2 1\r\n #define HAVE___THREAD 1\r\n #define HAVE_SINF 1\r\n #define HAVE_COSF 1\r\n #define HAVE_TANF 1\r\n #define HAVE_SINHF 1\r\n #define HAVE_COSHF 1\r\n #define HAVE_TANHF 1\r\n #define HAVE_FABSF 1\r\n #define HAVE_FLOORF 1\r\n #define HAVE_CEILF 1\r\n #define HAVE_RINTF 1\r\n #define HAVE_TRUNCF 1\r\n #define HAVE_SQRTF 1\r\n #define HAVE_LOG10F 1\r\n #define HAVE_LOGF 1\r\n #define HAVE_LOG1PF 1\r\n #define HAVE_EXPF 1\r\n #define HAVE_EXPM1F 1\r\n #define HAVE_ASINF 1\r\n #define HAVE_ACOSF 1\r\n #define HAVE_ATANF 1\r\n #define HAVE_ASINHF 1\r\n #define HAVE_ACOSHF 1\r\n #define HAVE_ATANHF 1\r\n #define HAVE_HYPOTF 1\r\n #define HAVE_ATAN2F 1\r\n #define HAVE_POWF 1\r\n #define HAVE_FMODF 1\r\n #define HAVE_MODFF 1\r\n #define HAVE_FREXPF 1\r\n #define HAVE_LDEXPF 1\r\n #define HAVE_EXP2F 1\r\n #define HAVE_LOG2F 1\r\n #define HAVE_COPYSIGNF 1\r\n #define HAVE_NEXTAFTERF 1\r\n #define HAVE_CBRTF 1\r\n #define HAVE_SINL 1\r\n #define HAVE_COSL 1\r\n #define HAVE_TANL 1\r\n #define HAVE_SINHL 1\r\n #define HAVE_COSHL 1\r\n #define HAVE_TANHL 1\r\n #define HAVE_FABSL 1\r\n #define HAVE_FLOORL 1\r\n #define HAVE_CEILL 1\r\n #define HAVE_RINTL 1\r\n #define HAVE_TRUNCL 1\r\n #define HAVE_SQRTL 1\r\n #define HAVE_LOG10L 1\r\n #define HAVE_LOGL 1\r\n #define HAVE_LOG1PL 1\r\n #define HAVE_EXPL 1\r\n #define HAVE_EXPM1L 1\r\n #define HAVE_ASINL 1\r\n #define HAVE_ACOSL 1\r\n #define HAVE_ATANL 1\r\n #define HAVE_ASINHL 1\r\n #define HAVE_ACOSHL 1\r\n #define HAVE_ATANHL 1\r\n #define HAVE_HYPOTL 1\r\n #define HAVE_ATAN2L 1\r\n #define HAVE_POWL 1\r\n #define HAVE_FMODL 1\r\n #define HAVE_MODFL 1\r\n #define HAVE_FREXPL 1\r\n #define HAVE_LDEXPL 1\r\n #define HAVE_EXP2L 1\r\n #define HAVE_LOG2L 1\r\n #define HAVE_COPYSIGNL 1\r\n #define HAVE_NEXTAFTERL 1\r\n #define HAVE_CBRTL 1\r\n #define HAVE_DECL_SIGNBIT\r\n #define HAVE_COMPLEX_H 1\r\n #define HAVE_CABS 1\r\n #define HAVE_CACOS 1\r\n #define HAVE_CACOSH 1\r\n #define HAVE_CARG 1\r\n #define HAVE_CASIN 1\r\n #define HAVE_CASINH 1\r\n #define HAVE_CATAN 1\r\n #define HAVE_CATANH 1\r\n #define HAVE_CCOS 1\r\n #define HAVE_CCOSH 1\r\n #define HAVE_CEXP 1\r\n #define HAVE_CIMAG 1\r\n #define HAVE_CLOG 1\r\n #define HAVE_CONJ 1\r\n #define HAVE_CPOW 1\r\n #define HAVE_CPROJ 1\r\n #define HAVE_CREAL 1\r\n #define HAVE_CSIN 1\r\n #define HAVE_CSINH 1\r\n #define HAVE_CSQRT 1\r\n #define HAVE_CTAN 1\r\n #define HAVE_CTANH 1\r\n #define HAVE_CABSF 1\r\n #define HAVE_CACOSF 1\r\n #define HAVE_CACOSHF 1\r\n #define HAVE_CARGF 1\r\n #define HAVE_CASINF 1\r\n #define HAVE_CASINHF 1\r\n #define HAVE_CATANF 1\r\n #define HAVE_CATANHF 1\r\n #define HAVE_CCOSF 1\r\n #define HAVE_CCOSHF 1\r\n #define HAVE_CEXPF 1\r\n #define HAVE_CIMAGF 1\r\n #define HAVE_CLOGF 1\r\n #define HAVE_CONJF 1\r\n #define HAVE_CPOWF 1\r\n #define HAVE_CPROJF 1\r\n #define HAVE_CREALF 1\r\n #define HAVE_CSINF 1\r\n #define HAVE_CSINHF 1\r\n #define HAVE_CSQRTF 1\r\n #define HAVE_CTANF 1\r\n #define HAVE_CTANHF 1\r\n #define HAVE_CABSL 1\r\n #define HAVE_CACOSL 1\r\n #define HAVE_CACOSHL 1\r\n #define HAVE_CARGL 1\r\n #define HAVE_CASINL 1\r\n #define HAVE_CASINHL 1\r\n #define HAVE_CATANL 1\r\n #define HAVE_CATANHL 1\r\n #define HAVE_CCOSL 1\r\n #define HAVE_CCOSHL 1\r\n #define HAVE_CEXPL 1\r\n #define HAVE_CIMAGL 1\r\n #define HAVE_CLOGL 1\r\n #define HAVE_CONJL 1\r\n #define HAVE_CPOWL 1\r\n #define HAVE_CPROJL 1\r\n #define HAVE_CREALL 1\r\n #define HAVE_CSINL 1\r\n #define HAVE_CSINHL 1\r\n #define HAVE_CSQRTL 1\r\n #define HAVE_CTANL 1\r\n #define HAVE_CTANHL 1\r\n #define NPY_RESTRICT restrict\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define HAVE_LDOUBLE_INTEL_EXTENDED_16_BYTES_LE 1\r\n #define NPY_PY3K 1\r\n #ifndef __cplusplus\r\n \/* #undef inline *\/\r\n #endif\r\n \r\n #ifndef _NPY_NPY_CONFIG_H_\r\n #error config.h should never be included directly, include npy_config.h instead\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n Generating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'exp' [-Wincompatible-library-redeclaration]\r\n int exp (void);\r\n ^\r\n _configtest.c:1:5: note: 'exp' is a builtin with type 'double (double)'\r\n 1 warning generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n #define NPY_SIZEOF_SHORT SIZEOF_SHORT\r\n #define NPY_SIZEOF_INT SIZEOF_INT\r\n #define NPY_SIZEOF_LONG SIZEOF_LONG\r\n #define NPY_SIZEOF_FLOAT 4\r\n #define NPY_SIZEOF_COMPLEX_FLOAT 8\r\n #define NPY_SIZEOF_DOUBLE 8\r\n #define NPY_SIZEOF_COMPLEX_DOUBLE 16\r\n #define NPY_SIZEOF_LONGDOUBLE 16\r\n #define NPY_SIZEOF_COMPLEX_LONGDOUBLE 32\r\n #define NPY_SIZEOF_PY_INTPTR_T 8\r\n #define NPY_SIZEOF_OFF_T 8\r\n #define NPY_SIZEOF_PY_LONG_LONG 8\r\n #define NPY_SIZEOF_LONGLONG 8\r\n #define NPY_NO_SMP 0\r\n #define NPY_HAVE_DECL_ISNAN\r\n #define NPY_HAVE_DECL_ISINF\r\n #define NPY_HAVE_DECL_ISFINITE\r\n #define NPY_HAVE_DECL_SIGNBIT\r\n #define NPY_USE_C99_COMPLEX 1\r\n #define NPY_HAVE_COMPLEX_DOUBLE 1\r\n #define NPY_HAVE_COMPLEX_FLOAT 1\r\n #define NPY_HAVE_COMPLEX_LONG_DOUBLE 1\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define NPY_USE_C99_FORMATS 1\r\n #define NPY_VISIBILITY_HIDDEN __attribute__((visibility(\"hidden\")))\r\n #define NPY_ABI_VERSION 0x01000009\r\n #define NPY_API_VERSION 0x0000000D\r\n \r\n #ifndef __STDC_FORMAT_MACROS\r\n #define __STDC_FORMAT_MACROS 1\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h']\r\n building extension \"numpy.core._multiarray_tests\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n building extension \"numpy.core._multiarray_umath\" sources\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_ufunc_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common' to include_dirs.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_internal.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h']\r\n building extension \"numpy.core._umath_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n building extension \"numpy.core._rational_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n building extension \"numpy.core._struct_ufunc_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n building extension \"numpy.core._operand_flag_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n building extension \"numpy.fft.fftpack_lite\" sources\r\n building extension \"numpy.linalg.lapack_lite\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n building extension \"numpy.linalg._umath_linalg\" sources\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n building extension \"numpy.random.mtrand\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/random\r\n building data_files sources\r\n build_src: building npy-pkg config files\r\n running build_py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/conftest.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_globals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/dual.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_distributor_init.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/ctypeslib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/matlib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_pytesttester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/py3k.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/_inspect.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/fromnumeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_add_newdocs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_methods.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_internal.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_string_helpers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/multiarray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/records.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup_common.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_aliased_types.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/memmap.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/overrides.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/getlimits.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype_ctypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/defchararray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/machar.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/einsumfunc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath_tests.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numerictypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_type_aliases.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/cversions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/arrayprint.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/code_generators\/generate_numpy_api.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/unixccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/numpy_distribution.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/conv_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/cpuinfo.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvc9compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/npy_pkg_config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/compat.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/misc_util.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/log.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/line_endings.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/lib2def.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/pathccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/system_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/exec_command.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/from_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/mingw32ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/extension.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/intelccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/distutils\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config_compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_ext.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_headers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_py.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_src.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/sdist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_scripts.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/bdist_rpm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/autodist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/egg_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/develop.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_data.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/gnu.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/compaq.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/intel.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/none.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/nag.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/ibm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/sun.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/lahey.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/g95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/mips.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/hpux.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/environment.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pathf95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/absoft.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/vast.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/misc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/internals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/creation.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/constants.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/ufuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/broadcasting.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/basics.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/subclassing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/indexing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/byteswapping.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/structured_arrays.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/glossary.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/common_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/crackfortran.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cb_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py2e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/func2subr.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/diagnose.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/capi_maps.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f90mod_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py_testing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/use_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/auxfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__main__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/helper.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/fftpack.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_iotools.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/mixins.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/nanfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/recfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/histograms.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/scimath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/user_array.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/format.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/twodim_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/financial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/index_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/npyio.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/stride_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arrayterator.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraysetops.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraypad.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/type_check.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_datasource.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/ufunclike.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/linalg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/extras.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/testutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/bench.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/timer_comparison.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/mrecords.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/defmatrix.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/laguerre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/_polybase.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polyutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite_e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/chebyshev.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/legendre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/print_coercion_tables.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/parameterized.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n running build_clib\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_clib\r\n building 'npymath' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n ar: adding 4 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n building 'npysort' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/quicksort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/mergesort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/heapsort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/selection.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/binsearch.c\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n 22 warnings generated.\r\n ar: adding 5 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n running build_ext\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_ext\r\n building 'numpy.core._dummy' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/dummymodule.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/dummymodule.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_dummy.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_tests.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_umath' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/core\/src\/multiarray\/alloc.c\r\n clang: numpy\/core\/src\/multiarray\/calculation.cclang: numpy\/core\/src\/multiarray\/array_assign_scalar.c\r\n clang: numpy\/core\/src\/multiarray\/convert.c\r\n \r\n clang: numpy\/core\/src\/multiarray\/ctors.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busday.c\r\n clang: numpy\/core\/src\/multiarray\/dragon4.cclang: numpy\/core\/src\/multiarray\/flagsobject.c\r\n \r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/arrayobject.c\r\n clang: numpy\/core\/src\/multiarray\/array_assign_array.c\r\n clang: numpy\/core\/src\/multiarray\/convert_datatype.c\r\n clang: numpy\/core\/src\/multiarray\/getset.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busdaycal.c\r\n clang: numpy\/core\/src\/multiarray\/buffer.c\r\n clang: numpy\/core\/src\/multiarray\/compiled_base.c\r\n clang: numpy\/core\/src\/multiarray\/hashdescr.c\r\n clang: numpy\/core\/src\/multiarray\/descriptor.c\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/conversion_utils.c\r\n clang: numpy\/core\/src\/multiarray\/item_selection.c\r\n clang: numpy\/core\/src\/multiarray\/dtype_transfer.c\r\n clang: numpy\/core\/src\/multiarray\/mapping.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n 3 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/datetime.c\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:477:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ptr = PyUnicode_AS_UNICODE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/common.c\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 6 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/nditer_pywrap.c\r\n 9 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/sequence.c\r\n clang: numpy\/core\/src\/multiarray\/shape.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n clang: numpy\/core\/src\/multiarray\/methods.c\r\n clang: numpy\/core\/src\/multiarray\/iterators.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_strings.c\r\n clang: numpy\/core\/src\/multiarray\/number.c\r\n clang: numpy\/core\/src\/multiarray\/scalarapi.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n numpy\/core\/src\/multiarray\/scalarapi.c:74:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:135:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1849:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buffer = PyUnicode_AS_DATA(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 5 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/typeinfo.c\r\n clang: numpy\/core\/src\/multiarray\/refcount.c\r\n clang: numpy\/core\/src\/multiarray\/usertypes.c\r\n clang: numpy\/core\/src\/multiarray\/multiarraymodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n clang: numpy\/core\/src\/multiarray\/vdot.c\r\n clang: numpy\/core\/src\/umath\/umathmodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n clang: numpy\/core\/src\/umath\/reduction.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n clang: numpy\/core\/src\/multiarray\/nditer_api.c\r\n 14 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/strfuncs.c\r\n numpy\/core\/src\/umath\/loops.c.src:655:18: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n result = PyEval_CallObject(tocall, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:178:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_ReprFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:195:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_StrFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 2 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/temp_elide.c\r\n clang: numpy\/core\/src\/umath\/cpuid.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n clang: numpy\/core\/src\/umath\/ufunc_object.c\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'byte_long' [-Wunused-function]\r\n byte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ubyte_long' [-Wunused-function]\r\n ubyte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'short_long' [-Wunused-function]\r\n short_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ushort_long' [-Wunused-function]\r\n ushort_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'int_long' [-Wunused-function]\r\n int_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'uint_long' [-Wunused-function]\r\n uint_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'long_long' [-Wunused-function]\r\n long_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulong_long' [-Wunused-function]\r\n ulong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longlong_long' [-Wunused-function]\r\n longlong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulonglong_long' [-Wunused-function]\r\n ulonglong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'half_long' [-Wunused-function]\r\n half_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'float_long' [-Wunused-function]\r\n float_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'double_long' [-Wunused-function]\r\n double_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longdouble_long' [-Wunused-function]\r\n longdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cfloat_long' [-Wunused-function]\r\n cfloat_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cdouble_long' [-Wunused-function]\r\n cdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'clongdouble_long' [-Wunused-function]\r\n clongdouble_long(PyObject *obj)\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/nditer_constr.c\r\n numpy\/core\/src\/umath\/ufunc_object.c:657:19: warning: comparison of integers of different signs: 'int' and 'size_t' (aka 'unsigned long') [-Wsign-compare]\r\n for (i = 0; i < len; i++) {\r\n ~ ^ ~~~\r\n clang: numpy\/core\/src\/umath\/override.c\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n clang: numpy\/core\/src\/common\/array_assign.c\r\n clang: numpy\/core\/src\/common\/ufunc_override.c\r\n clang: numpy\/core\/src\/common\/npy_longdouble.c\r\n clang: numpy\/core\/src\/common\/numpyos.c\r\n clang: numpy\/core\/src\/common\/ucsnarrow.c\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/umath\/extobj.c\r\n numpy\/core\/src\/common\/ucsnarrow.c:139:34: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n ret = (PyUnicodeObject *)PyUnicode_FromUnicode((Py_UNICODE*)buf,\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/common\/python_xerbla.c\r\n clang: numpy\/core\/src\/common\/cblasfuncs.c\r\n clang: \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c\r\n In file included from \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c:26:\r\n In file included from numpy\/core\/include\/numpy\/arrayobject.h:4:\r\n In file included from numpy\/core\/include\/numpy\/ndarrayobject.h:21:\r\n build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h:1463:1: warning: unused function '_import_array' [-Wunused-function]\r\n _import_array(void)\r\n ^\r\n 1 warning generated.\r\n 17 warnings generated.\r\n clang: numpy\/core\/src\/umath\/ufunc_type_resolution.c\r\n 4 warnings generated.\r\n 4 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/alloc.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arrayobject.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_scalar.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_array.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/buffer.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/calculation.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/compiled_base.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/common.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert_datatype.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/conversion_utils.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/ctors.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_strings.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busday.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busdaycal.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/descriptor.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dragon4.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dtype_transfer.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/flagsobject.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/getset.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/hashdescr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/item_selection.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/iterators.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/mapping.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/methods.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/multiarraymodule.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_api.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_constr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_pywrap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/number.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/refcount.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/sequence.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/shape.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalarapi.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/strfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/temp_elide.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/typeinfo.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/usertypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/vdot.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/umathmodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/reduction.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_object.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/extobj.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/cpuid.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_type_resolution.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/halffloat.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/array_assign.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/npy_longdouble.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ucsnarrow.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ufunc_override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/numpyos.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/cblasfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/python_xerbla.o build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -lnpysort -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_umath.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.core._umath_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_umath_tests.cpython-39-darwin.so\r\n building 'numpy.core._rational_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_rational_tests.cpython-39-darwin.so\r\n building 'numpy.core._struct_ufunc_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_struct_ufunc_tests.cpython-39-darwin.so\r\n building 'numpy.core._operand_flag_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_operand_flag_tests.cpython-39-darwin.so\r\n building 'numpy.fft.fftpack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n compile options: '-Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/fft\/fftpack_litemodule.c\r\n clang: numpy\/fft\/fftpack.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_lite.cpython-39-darwin.so\r\n building 'numpy.linalg.lapack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/linalg\/lapack_litemodule.c\r\n clang: numpy\/linalg\/lapack_lite\/python_xerbla.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.linalg._umath_linalg' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n numpy\/linalg\/umath_linalg.c.src:735:32: warning: unknown warning group '-Wmaybe-uninitialized', ignored [-Wunknown-warning-option]\r\n #pragma GCC diagnostic ignored \"-Wmaybe-uninitialized\"\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:541:1: warning: unused function 'dump_ufunc_object' [-Wunused-function]\r\n dump_ufunc_object(PyUFuncObject* ufunc)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:566:1: warning: unused function 'dump_linearize_data' [-Wunused-function]\r\n dump_linearize_data(const char* name, const LINEARIZE_DATA_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_FLOAT_matrix' [-Wunused-function]\r\n dump_FLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_DOUBLE_matrix' [-Wunused-function]\r\n dump_DOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CFLOAT_matrix' [-Wunused-function]\r\n dump_CFLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CDOUBLE_matrix' [-Wunused-function]\r\n dump_CDOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_FLOAT_matrix' [-Wunused-function]\r\n zero_FLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_DOUBLE_matrix' [-Wunused-function]\r\n zero_DOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CFLOAT_matrix' [-Wunused-function]\r\n zero_CFLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CDOUBLE_matrix' [-Wunused-function]\r\n zero_CDOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:1862:1: warning: unused function 'dump_geev_params' [-Wunused-function]\r\n dump_geev_params(const char *name, GEEV_PARAMS_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2132:1: warning: unused function 'init_cgeev' [-Wunused-function]\r\n init_cgeev(GEEV_PARAMS_t* params,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2213:1: warning: unused function 'process_cgeev_results' [-Wunused-function]\r\n process_cgeev_results(GEEV_PARAMS_t *NPY_UNUSED(params))\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2376:1: warning: unused function 'dump_gesdd_params' [-Wunused-function]\r\n dump_gesdd_params(const char *name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2864:1: warning: unused function 'dump_gelsd_params' [-Wunused-function]\r\n dump_gelsd_params(const char *name,\r\n ^\r\n 16 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/_umath_linalg.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.random.mtrand' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\r\n compile options: '-D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/random\/mtrand\/mtrand.c\r\n clang: numpy\/random\/mtrand\/initarray.cclang: numpy\/random\/mtrand\/randomkit.c\r\n \r\n clang: numpy\/random\/mtrand\/distributions.c\r\n numpy\/random\/mtrand\/mtrand.c:40400:34: error: no member named 'tp_print' in 'struct _typeobject'\r\n __pyx_type_6mtrand_RandomState.tp_print = 0;\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 12 warnings and 1 error generated.\r\n error: Command \"clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c numpy\/random\/mtrand\/mtrand.c -o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o -MMD -MF build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o.d\" failed with exit status 1\nMaybe try to create a virtual env with python 3.8 or 3.7","embeddings":[-0.372736305,-0.0714854077,-0.0845708922,0.2079993039,0.2768215835,0.0631752834,-0.2505984604,0.0786435232,-0.2070377618,-0.0340085514,0.0305880085,0.3349942863,-0.1231444255,0.0895890892,0.0807427391,0.0722528622,0.189591229,0.2151816189,-0.3760947287,0.0339735448,0.0351168886,0.1739135981,-0.1176334247,-0.0769590884,-0.1111136004,-0.2832199335,0.0382908024,0.1094798371,-0.532479167,-0.3132808805,0.291446507,0.0252360553,-0.0163556207,0.5489337444,-0.0001225818,0.1088191941,0.1684155613,0.018390229,-0.3556295037,-0.4525223374,-0.2284404039,-0.4977667034,0.1029686555,0.0328003541,0.2697275281,-0.2269828916,-0.1955371499,-0.3135961592,0.7321167588,0.186695531,0.1548049748,0.3139568865,0.1280289441,0.0492491126,0.1173379719,0.3006310165,-0.3177905381,0.3576257825,0.2570489943,-0.1687269062,0.2200257182,0.1589555889,-0.2279057503,0.2082943469,0.5441326499,-0.1704113036,-0.0771712363,-0.4264783263,-0.2593541741,0.2732714415,0.5033100843,-0.2703568041,-0.6291657686,-0.0928478241,0.0548009649,-0.2381634116,0.3260372877,0.223864004,-0.2062843293,0.0802830532,-0.4288931787,-0.280636698,-0.1864999384,0.2670033574,-0.0238301586,0.164580822,-0.1459159702,0.1176900715,0.2985162139,-0.0034400814,0.2251620293,0.0791203454,-0.0705096424,0.0032631624,-0.2801685035,0.0655268356,-0.2036794722,0.2356178463,0.1992911696,0.1688992977,-0.1797399819,-0.2005181015,-0.2243610919,0.0616602823,0.066762954,0.2027508169,-0.1514637321,0.1150361076,0.3352227509,0.2663400173,0.4560080171,0.0843001902,0.0707917213,-0.1902450472,-0.209478572,-0.1111337766,0.1524448395,-0.3028428555,-0.4012164772,0.1194737256,0.0135395043,0.1056581885,0.0688990727,0.1510053128,-0.1464695483,0.1697680354,0.2481278181,0.3707648218,-0.232209444,0.0304407589,-0.0021012723,-0.0312458072,-0.2525970638,0.0168092307,0.1831142902,-0.1020761356,0.3657963574,0.2121508121,0.1536038369,0.1180837601,0.2853684127,0.0168460198,-0.191518262,0.4379527569,-0.2559596002,0.2257401049,0.0813604146,0.3052143753,0.0385886207,-0.0087099858,0.1200957373,-0.0894460753,-0.1464142799,0.0493404903,-0.2953626513,0.051196415,-0.16042234,-0.0320122093,0.0422822423,-0.0837064162,0.015641503,-0.2536138892,0.0998989642,-0.1036708131,0.3004768193,0.414516747,-0.51960361,-0.038707234,-0.1943658739,-0.1891062409,-0.0495771132,0.1207961589,-0.1165530235,0.1363425553,-0.164071247,-0.4210309684,0.2341786772,-0.3646256328,-0.367303431,0.1608417332,-0.1646376401,-0.086746715,0.0553342775,0.0455297008,-0.0180610958,0.0614581928,-0.0477216803,0.326702714,-0.1066503674,-0.2653565109,0.0084401397,-0.5503122211,0.0675886497,0.0629741848,-0.0602851324,0.0150617491,-0.0644270927,0.139841333,0.0511045381,0.0582192577,-0.0014922149,0.6988541484,0.416780889,0.0951713175,-0.2119343877,-0.2936756313,-0.3661433756,0.1583058536,0.0255879425,0.0643135011,-0.3344601095,-0.0186114982,-0.1848731637,0.128934145,-0.0381260812,-0.0527399741,-0.0203619599,0.0027848945,0.2437096834,-0.2807203233,-0.233386904,0.5044182539,0.1574977785,0.2694315314,-0.4862544239,0.0366727933,-0.3219184875,-0.1439168453,0.1915305704,0.2339345664,0.0553197712,-0.2622070611,-0.1443571001,0.3139041066,-0.1654947251,-0.1707349569,-0.0076072491,0.1440196335,0.2182890773,-0.2099311203,0.2413927317,-0.0720721483,-0.1406363845,0.0983242542,-0.0743637085,-0.1697815061,0.0582351796,0.3382672966,-0.0074652079,0.3404535055,0.3141412735,0.0584230796,-0.1450206935,-0.4074391723,0.396992296,-0.0027272592,0.4638343155,-0.1445932984,0.0697035044,0.053041961,-0.1363226324,0.0916765481,-0.0233560391,0.0049676136,-0.072430402,0.1785400212,0.185349673,0.0348233506,0.5852743983,0.0619082898,-0.1347470284,0.1598725617,-0.0658946782,-0.0729199722,0.0373055674,0.1743800044,-0.0199271459,-0.0330145024,0.0282345489,-0.0624522939,-0.0412191078,-0.207671389,0.1253645271,0.2953568995,-0.3419783413,0.1114570275,-0.5203249454,-0.0051726159,-0.0906571373,-0.0719850361,-0.4134817719,-0.2235808372,-0.1058893874,0.1500547081,0.2082414627,0.1049688607,0.1166992337,0.0297720432,0.001696527,-0.209155798,-0.1628497243,-0.1249261945,-0.0730912238,-0.0193224195,0.1679847986,0.1203907058,0.4714294374,-0.0997886732,-0.1613499224,-0.4218867421,-0.261598736,0.107161954,-0.398339957,0.742423296,0.469727993,0.202538535,0.0647644475,0.1159555018,0.1875865012,-0.1887231469,-0.1085140929,-0.0309973564,-0.213460058,-0.2505242229,-0.1225516498,-0.3344918489,-0.32494241,-0.2550111711,0.1577945501,-0.0034383023,-0.1047476232,0.3142209351,0.2670063674,0.0404987186,-0.165805921,0.0468914136,-0.0268509146,-0.2246979922,0.177915588,-0.201831609,-0.2568696737,0.3019668162,0.2744387388,0.1829818487,0.2670661211,-0.7614741325,-0.2670167983,-0.3597413003,0.2398702502,-0.1860815585,0.2179733366,0.1911988556,0.0769287571,0.0779739991,-0.2079752088,-0.4608823061,-0.0097571854,0.0752782673,0.4969078898,0.0363830738,0.3437930942,-0.0585201345,0.6186058521,0.397529155,0.0103019681,0.1783744097,0.2334723771,0.5895819068,0.0778009444,-0.4482364953,0.2397813201,-0.0084547177,0.1363245547,-0.1942522824,0.2197874635,0.2414545864,-0.277361989,0.2657617033,0.0042451476,0.0055124569,-0.2210909873,-0.13537018,0.3050734699,0.2228150517,0.0064602434,0.0430852436,-0.1302471906,0.0280230325,0.1745655835,0.010982167,0.3000519872,-0.1954792738,-0.1373809576,-0.183105126,0.2909622788,-0.0428895615,0.4082678556,-0.2521476746,0.4757547975,0.0345090851,0.0108251274,0.475733906,-0.3083002269,-0.1645100862,0.1024655625,0.1262150258,-0.6377895474,-0.1250775605,-0.1298962086,0.1647143066,0.248098284,0.2099779695,-0.4318394959,-0.1485227644,0.2467826158,0.5641607046,0.0371086225,-0.1298632622,-0.3288927078,-0.3882852793,-0.2767420411,-0.1371272653,-0.058102645,-0.1284619868,0.2001442164,-0.0588850789,0.1972738355,0.1628406644,0.1069483757,-0.0801087394,0.2060672939,-0.1776001751,0.2210780978,0.3728102148,-0.1271787584,0.1659846008,0.737244606,-0.2061559558,-0.3858459294,-0.0367117412,0.1224328429,0.1555299163,0.3913010359,-0.2864465714,-0.1263913363,0.2399162054,0.2055450827,0.1049313173,0.14691028,0.289960444,0.1443811804,-0.195566237,-0.4653308094,0.1165174395,0.4374177158,0.1289001107,0.4145717323,0.0137373563,0.0171636641,-0.0216146,0.2067595273,0.8317614794,-0.0005798346,0.0876606479,0.3389310539,-0.0641187727,0.5443438888,-0.4247684479,0.04840637,-0.0802192986,-0.143647477,-0.1364471316,-0.3701676726,0.3992536366,0.1148540974,-0.2142028362,0.3585865796,-0.1218098328,0.29375422,0.0482256673,0.1957940161,-0.0632860139,-0.2802743912,-0.1753268093,0.1845365763,-0.233880952,0.1757169068,-0.1705885977,0.083735086,-0.3188177347,-0.0867874846,-0.5056916475,0.0665702522,-0.443862617,0.2258304805,-0.0344874859,-0.2310214341,0.3094379306,0.1276277453,0.3448578715,0.4515912831,-0.3112368286,0.1110853925,-0.2714110315,-0.2467177361,0.3035934865,0.0439759605,0.0467805564,-0.0107217925,-0.1727659255,-0.0625354499,-0.2353877276,0.0169170164,0.0544819459,-0.0357779078,-0.1322258413,-0.1305123866,-0.1653528064,-0.1025140062,0.2076852918,-0.2839880288,0.0506421104,0.07781367,-0.0196289588,0.4598361552,-0.1750662029,-0.2117223293,-0.1011548117,0.212752223,-0.1656555086,-0.2982017696,0.234419629,0.2739279866,-0.3369333148,-0.1571041346,-0.094751589,-0.1557620764,-0.4031636715,0.1162526757,0.1045675799,-0.1815831959,-0.1145593971,0.2249464691,0.2387459278,-0.2340987474,0.0011506572,-0.4424839914,-0.2236991078,0.2046369612,-0.1303168684,0.0251450893,0.1048556641,0.0480054915,-0.0482010245,0.1132196635,-0.1702207923,-0.1554876715,0.0242763478,-0.0713960975,0.3412455618,-0.0761460811,0.0959786624,-0.2396223247,-0.03813833,0.1586219817,-0.2529083788,-0.076279223,-0.0355480611,0.137926355,0.0933924913,-0.3160541952,0.212218523,-0.4301118255,-0.1320781857,-0.1007387489,0.183366701,0.2039175779,0.0625451133,-0.2080432922,-0.4835402668,0.0814061463,0.0820645839,0.2181892693,0.0904808417,0.0420457087,0.2254475355,0.0915967301,0.1900683045,-0.09184926,-0.3334983885,0.0691129938,0.0046931463,0.2152244151,0.1156246215,-0.3056086898,0.2381537408,0.2715603411,0.4740514159,0.5606653094,-0.0505268052,0.2009599954,0.3780045509,0.0619235337,-0.3095085323,-0.1179374307,0.0978347436,0.211131677,-0.0669629127,0.294595927,0.2904630005,-0.1662479341,0.1414123923,0.2137935907,0.2448209524,0.0442248173,0.2401469797,0.1740344614,0.1456380934,-0.0191844087,0.2478772849,0.2596318722,0.1486105621,0.2685900927,-0.0940031186,-0.0129717235,0.1056639478,0.2145727575,-0.1893776506,-0.3154191971,-0.0778527632,-0.14081873,0.3885463476,0.0168743245,-0.0760412812,0.2765650153,-0.3946425617,-0.3519563675,-0.1730092168,0.2614917159,0.0503016636,0.2280521244,0.3215574324,0.0809561014,0.1780232787,0.0048405635,0.0397591293,-0.4191242456,0.2413220108,0.1764912158,-0.1236993223,-0.770821631,0.2125341445,0.1664997637,-0.1692295074,-0.0538866408,0.3296135664,0.1637779623,-0.1454275995,-0.2812110484,0.5068164468,0.5802744627,0.3409990072,-0.0469977558,0.1705626249,-0.2142198682,0.0777670294,-0.0155832991,0.2700774372,-0.1369200051,0.154964745,0.236733079,0.0225825254,-0.1113553271,-0.1749951243,0.05340923,0.4135027826,-0.1955527812,0.1946346015,-0.1500745118,-0.0271252468,-0.0973940268,-0.0908994824,-0.4811028838,0.2690776885,0.5110985637,0.0316910855,-0.0370604061,-0.1930583566,0.0272696707,-0.0880471319,0.3524516523,0.495470494,0.191963926,-0.4115162492,-0.0745142847,-0.6475214958,-0.2043262273,-0.3776461184,-0.2580498755,0.1304243505,-0.0521419086,0.0602565371,0.3123220801,0.2321897298,0.2287330478,0.1469683051,-0.0183709953,-0.3618494868,-0.1180011779,0.0987984166,0.1109670922,0.0805131122,-0.2171806246,0.0052382955,0.2312532216,-0.2043973804,-0.2440494448,-0.0672755912,0.0961156338,0.0278290417,0.3226872683,0.0853667632,0.3677543998,0.2596828043,-0.2449216247,-0.4981413782,-0.2963164449,-0.1827402264,0.3557329476,-0.046792753,0.2464546263,-0.5367340446,0.0496459119,-0.5474626422,0.097678706,0.0588760599,-0.0952013209,-0.2168268859,0.2665975988,-0.215844512,0.3078436255,0.0800647959,0.1987073421,-0.0550648682,0.2909039557,-0.4305544794,-0.322242856,0.3843740821,0.168199867,-0.3109259009,0.3755890727,0.0602345243,0.2313626111,-0.4108432531,-0.5497607589,0.1153397188,0.0295620337,-0.0591647848,-0.2287070155,0.1518161595,-0.1175127327,-0.0596970692,0.009724074,0.5295980573,0.0082971165,-0.254521817,0.3482492864,0.1984189898]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1696","title":"Unable to install datasets","comments":"Thanks, @thomwolf! I fixed the issue by downgrading python to 3.7. ","body":"** Edit **\r\nI believe there's a bug with the package when you're installing it with Python 3.9. I recommend sticking with previous versions. Thanks, @thomwolf for the insight! \r\n\r\n**Short description**\r\n\r\nI followed the instructions for installing datasets (https:\/\/huggingface.co\/docs\/datasets\/installation.html). However, while I tried to download datasets using `pip install datasets` I got a massive error message after getting stuck at \"Installing build dependencies...\" \r\n\r\nI was wondering if this problem can be fixed by creating a virtual environment, but it didn't help. Can anyone offer some advice on how to fix this issue? \r\n\r\nHere's an error message: \r\n\r\n`(env) Gas-MacBook-Pro:Downloads destiny$ pip install datasets\r\nCollecting datasets\r\n Using cached datasets-1.2.0-py3-none-any.whl (159 kB)\r\nCollecting numpy>=1.17\r\n Using cached numpy-1.19.5-cp39-cp39-macosx_10_9_x86_64.whl (15.6 MB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-2.0.0.tar.gz (58.9 MB)\r\n....\r\n\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceilf' [-Wincompatible-library-redeclaration]\r\n int ceilf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceilf' is a builtin with type 'float (float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintf' [-Wincompatible-library-redeclaration]\r\n int rintf (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintf' is a builtin with type 'float (float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncf' [-Wincompatible-library-redeclaration]\r\n int truncf (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncf' is a builtin with type 'float (float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtf' [-Wincompatible-library-redeclaration]\r\n int sqrtf (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtf' is a builtin with type 'float (float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10f' [-Wincompatible-library-redeclaration]\r\n int log10f (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10f' is a builtin with type 'float (float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logf' [-Wincompatible-library-redeclaration]\r\n int logf (void);\r\n ^\r\n _configtest.c:14:5: note: 'logf' is a builtin with type 'float (float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pf' [-Wincompatible-library-redeclaration]\r\n int log1pf (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pf' is a builtin with type 'float (float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expf' [-Wincompatible-library-redeclaration]\r\n int expf (void);\r\n ^\r\n _configtest.c:16:5: note: 'expf' is a builtin with type 'float (float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1f' [-Wincompatible-library-redeclaration]\r\n int expm1f (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1f' is a builtin with type 'float (float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinf' [-Wincompatible-library-redeclaration]\r\n int asinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinf' is a builtin with type 'float (float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosf' [-Wincompatible-library-redeclaration]\r\n int acosf (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosf' is a builtin with type 'float (float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanf' [-Wincompatible-library-redeclaration]\r\n int atanf (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanf' is a builtin with type 'float (float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhf' [-Wincompatible-library-redeclaration]\r\n int asinhf (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhf' is a builtin with type 'float (float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshf' [-Wincompatible-library-redeclaration]\r\n int acoshf (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshf' is a builtin with type 'float (float)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhf' [-Wincompatible-library-redeclaration]\r\n int atanhf (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhf' is a builtin with type 'float (float)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotf' [-Wincompatible-library-redeclaration]\r\n int hypotf (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotf' is a builtin with type 'float (float, float)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2f' [-Wincompatible-library-redeclaration]\r\n int atan2f (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2f' is a builtin with type 'float (float, float)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powf' [-Wincompatible-library-redeclaration]\r\n int powf (void);\r\n ^\r\n _configtest.c:26:5: note: 'powf' is a builtin with type 'float (float, float)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodf' [-Wincompatible-library-redeclaration]\r\n int fmodf (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodf' is a builtin with type 'float (float, float)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modff' [-Wincompatible-library-redeclaration]\r\n int modff (void);\r\n ^\r\n _configtest.c:28:5: note: 'modff' is a builtin with type 'float (float, float *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpf' [-Wincompatible-library-redeclaration]\r\n int frexpf (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpf' is a builtin with type 'float (float, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpf' [-Wincompatible-library-redeclaration]\r\n int ldexpf (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpf' is a builtin with type 'float (float, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2f' [-Wincompatible-library-redeclaration]\r\n int exp2f (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2f' is a builtin with type 'float (float)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2f' [-Wincompatible-library-redeclaration]\r\n int log2f (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2f' is a builtin with type 'float (float)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignf' [-Wincompatible-library-redeclaration]\r\n int copysignf (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignf' is a builtin with type 'float (float, float)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterf' [-Wincompatible-library-redeclaration]\r\n int nextafterf (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterf' is a builtin with type 'float (float, float)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtf' [-Wincompatible-library-redeclaration]\r\n int cbrtf (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtf' is a builtin with type 'float (float)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'sinl' [-Wincompatible-library-redeclaration]\r\n int sinl (void);\r\n ^\r\n _configtest.c:1:5: note: 'sinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cosl' [-Wincompatible-library-redeclaration]\r\n int cosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'tanl' [-Wincompatible-library-redeclaration]\r\n int tanl (void);\r\n ^\r\n _configtest.c:3:5: note: 'tanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'sinhl' [-Wincompatible-library-redeclaration]\r\n int sinhl (void);\r\n ^\r\n _configtest.c:4:5: note: 'sinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'coshl' [-Wincompatible-library-redeclaration]\r\n int coshl (void);\r\n ^\r\n _configtest.c:5:5: note: 'coshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'tanhl' [-Wincompatible-library-redeclaration]\r\n int tanhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'tanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'fabsl' [-Wincompatible-library-redeclaration]\r\n int fabsl (void);\r\n ^\r\n _configtest.c:7:5: note: 'fabsl' is a builtin with type 'long double (long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'floorl' [-Wincompatible-library-redeclaration]\r\n int floorl (void);\r\n ^\r\n _configtest.c:8:5: note: 'floorl' is a builtin with type 'long double (long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceill' [-Wincompatible-library-redeclaration]\r\n int ceill (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceill' is a builtin with type 'long double (long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintl' [-Wincompatible-library-redeclaration]\r\n int rintl (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintl' is a builtin with type 'long double (long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncl' [-Wincompatible-library-redeclaration]\r\n int truncl (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncl' is a builtin with type 'long double (long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtl' [-Wincompatible-library-redeclaration]\r\n int sqrtl (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtl' is a builtin with type 'long double (long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10l' [-Wincompatible-library-redeclaration]\r\n int log10l (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10l' is a builtin with type 'long double (long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logl' [-Wincompatible-library-redeclaration]\r\n int logl (void);\r\n ^\r\n _configtest.c:14:5: note: 'logl' is a builtin with type 'long double (long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pl' [-Wincompatible-library-redeclaration]\r\n int log1pl (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pl' is a builtin with type 'long double (long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expl' [-Wincompatible-library-redeclaration]\r\n int expl (void);\r\n ^\r\n _configtest.c:16:5: note: 'expl' is a builtin with type 'long double (long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1l' [-Wincompatible-library-redeclaration]\r\n int expm1l (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1l' is a builtin with type 'long double (long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinl' [-Wincompatible-library-redeclaration]\r\n int asinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosl' [-Wincompatible-library-redeclaration]\r\n int acosl (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanl' [-Wincompatible-library-redeclaration]\r\n int atanl (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhl' [-Wincompatible-library-redeclaration]\r\n int asinhl (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshl' [-Wincompatible-library-redeclaration]\r\n int acoshl (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhl' [-Wincompatible-library-redeclaration]\r\n int atanhl (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotl' [-Wincompatible-library-redeclaration]\r\n int hypotl (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2l' [-Wincompatible-library-redeclaration]\r\n int atan2l (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2l' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powl' [-Wincompatible-library-redeclaration]\r\n int powl (void);\r\n ^\r\n _configtest.c:26:5: note: 'powl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodl' [-Wincompatible-library-redeclaration]\r\n int fmodl (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modfl' [-Wincompatible-library-redeclaration]\r\n int modfl (void);\r\n ^\r\n _configtest.c:28:5: note: 'modfl' is a builtin with type 'long double (long double, long double *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpl' [-Wincompatible-library-redeclaration]\r\n int frexpl (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpl' is a builtin with type 'long double (long double, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpl' [-Wincompatible-library-redeclaration]\r\n int ldexpl (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpl' is a builtin with type 'long double (long double, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2l' [-Wincompatible-library-redeclaration]\r\n int exp2l (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2l' [-Wincompatible-library-redeclaration]\r\n int log2l (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignl' [-Wincompatible-library-redeclaration]\r\n int copysignl (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterl' [-Wincompatible-library-redeclaration]\r\n int nextafterl (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtl' [-Wincompatible-library-redeclaration]\r\n int cbrtl (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtl' is a builtin with type 'long double (long double)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:8:12: error: use of undeclared identifier 'HAVE_DECL_SIGNBIT'\r\n (void) HAVE_DECL_SIGNBIT;\r\n ^\r\n 1 error generated.\r\n failure.\r\n removing: _configtest.c _configtest.o\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabs' [-Wincompatible-library-redeclaration]\r\n int cabs (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabs' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacos' [-Wincompatible-library-redeclaration]\r\n int cacos (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacosh' [-Wincompatible-library-redeclaration]\r\n int cacosh (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'carg' [-Wincompatible-library-redeclaration]\r\n int carg (void);\r\n ^\r\n _configtest.c:4:5: note: 'carg' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casin' [-Wincompatible-library-redeclaration]\r\n int casin (void);\r\n ^\r\n _configtest.c:5:5: note: 'casin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinh' [-Wincompatible-library-redeclaration]\r\n int casinh (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catan' [-Wincompatible-library-redeclaration]\r\n int catan (void);\r\n ^\r\n _configtest.c:7:5: note: 'catan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanh' [-Wincompatible-library-redeclaration]\r\n int catanh (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccos' [-Wincompatible-library-redeclaration]\r\n int ccos (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccosh' [-Wincompatible-library-redeclaration]\r\n int ccosh (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexp' [-Wincompatible-library-redeclaration]\r\n int cexp (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexp' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimag' [-Wincompatible-library-redeclaration]\r\n int cimag (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimag' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clog' [-Wincompatible-library-redeclaration]\r\n int clog (void);\r\n ^\r\n _configtest.c:13:5: note: 'clog' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conj' [-Wincompatible-library-redeclaration]\r\n int conj (void);\r\n ^\r\n _configtest.c:14:5: note: 'conj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpow' [-Wincompatible-library-redeclaration]\r\n int cpow (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpow' is a builtin with type '_Complex double (_Complex double, _Complex double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cproj' [-Wincompatible-library-redeclaration]\r\n int cproj (void);\r\n ^\r\n _configtest.c:16:5: note: 'cproj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creal' [-Wincompatible-library-redeclaration]\r\n int creal (void);\r\n ^\r\n _configtest.c:17:5: note: 'creal' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csin' [-Wincompatible-library-redeclaration]\r\n int csin (void);\r\n ^\r\n _configtest.c:18:5: note: 'csin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinh' [-Wincompatible-library-redeclaration]\r\n int csinh (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrt' [-Wincompatible-library-redeclaration]\r\n int csqrt (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrt' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctan' [-Wincompatible-library-redeclaration]\r\n int ctan (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanh' [-Wincompatible-library-redeclaration]\r\n int ctanh (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanh' is a builtin with type '_Complex double (_Complex double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsf' [-Wincompatible-library-redeclaration]\r\n int cabsf (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosf' [-Wincompatible-library-redeclaration]\r\n int cacosf (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshf' [-Wincompatible-library-redeclaration]\r\n int cacoshf (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargf' [-Wincompatible-library-redeclaration]\r\n int cargf (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinf' [-Wincompatible-library-redeclaration]\r\n int casinf (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhf' [-Wincompatible-library-redeclaration]\r\n int casinhf (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanf' [-Wincompatible-library-redeclaration]\r\n int catanf (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhf' [-Wincompatible-library-redeclaration]\r\n int catanhf (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosf' [-Wincompatible-library-redeclaration]\r\n int ccosf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshf' [-Wincompatible-library-redeclaration]\r\n int ccoshf (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpf' [-Wincompatible-library-redeclaration]\r\n int cexpf (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagf' [-Wincompatible-library-redeclaration]\r\n int cimagf (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogf' [-Wincompatible-library-redeclaration]\r\n int clogf (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjf' [-Wincompatible-library-redeclaration]\r\n int conjf (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowf' [-Wincompatible-library-redeclaration]\r\n int cpowf (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowf' is a builtin with type '_Complex float (_Complex float, _Complex float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojf' [-Wincompatible-library-redeclaration]\r\n int cprojf (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'crealf' [-Wincompatible-library-redeclaration]\r\n int crealf (void);\r\n ^\r\n _configtest.c:17:5: note: 'crealf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinf' [-Wincompatible-library-redeclaration]\r\n int csinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhf' [-Wincompatible-library-redeclaration]\r\n int csinhf (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtf' [-Wincompatible-library-redeclaration]\r\n int csqrtf (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanf' [-Wincompatible-library-redeclaration]\r\n int ctanf (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhf' [-Wincompatible-library-redeclaration]\r\n int ctanhf (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhf' is a builtin with type '_Complex float (_Complex float)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsl' [-Wincompatible-library-redeclaration]\r\n int cabsl (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosl' [-Wincompatible-library-redeclaration]\r\n int cacosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshl' [-Wincompatible-library-redeclaration]\r\n int cacoshl (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargl' [-Wincompatible-library-redeclaration]\r\n int cargl (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinl' [-Wincompatible-library-redeclaration]\r\n int casinl (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhl' [-Wincompatible-library-redeclaration]\r\n int casinhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanl' [-Wincompatible-library-redeclaration]\r\n int catanl (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhl' [-Wincompatible-library-redeclaration]\r\n int catanhl (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosl' [-Wincompatible-library-redeclaration]\r\n int ccosl (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshl' [-Wincompatible-library-redeclaration]\r\n int ccoshl (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpl' [-Wincompatible-library-redeclaration]\r\n int cexpl (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagl' [-Wincompatible-library-redeclaration]\r\n int cimagl (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogl' [-Wincompatible-library-redeclaration]\r\n int clogl (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjl' [-Wincompatible-library-redeclaration]\r\n int conjl (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowl' [-Wincompatible-library-redeclaration]\r\n int cpowl (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowl' is a builtin with type '_Complex long double (_Complex long double, _Complex long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojl' [-Wincompatible-library-redeclaration]\r\n int cprojl (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creall' [-Wincompatible-library-redeclaration]\r\n int creall (void);\r\n ^\r\n _configtest.c:17:5: note: 'creall' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinl' [-Wincompatible-library-redeclaration]\r\n int csinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhl' [-Wincompatible-library-redeclaration]\r\n int csinhl (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtl' [-Wincompatible-library-redeclaration]\r\n int csqrtl (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanl' [-Wincompatible-library-redeclaration]\r\n int ctanl (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhl' [-Wincompatible-library-redeclaration]\r\n int ctanhl (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:2:12: warning: unused function 'static_func' [-Wunused-function]\r\n static int static_func (char * restrict a)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:3:19: warning: unused function 'static_func' [-Wunused-function]\r\n static inline int static_func (void)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h\r\n #define SIZEOF_PY_INTPTR_T 8\r\n #define SIZEOF_OFF_T 8\r\n #define SIZEOF_PY_LONG_LONG 8\r\n #define MATHLIB\r\n #define HAVE_SIN 1\r\n #define HAVE_COS 1\r\n #define HAVE_TAN 1\r\n #define HAVE_SINH 1\r\n #define HAVE_COSH 1\r\n #define HAVE_TANH 1\r\n #define HAVE_FABS 1\r\n #define HAVE_FLOOR 1\r\n #define HAVE_CEIL 1\r\n #define HAVE_SQRT 1\r\n #define HAVE_LOG10 1\r\n #define HAVE_LOG 1\r\n #define HAVE_EXP 1\r\n #define HAVE_ASIN 1\r\n #define HAVE_ACOS 1\r\n #define HAVE_ATAN 1\r\n #define HAVE_FMOD 1\r\n #define HAVE_MODF 1\r\n #define HAVE_FREXP 1\r\n #define HAVE_LDEXP 1\r\n #define HAVE_RINT 1\r\n #define HAVE_TRUNC 1\r\n #define HAVE_EXP2 1\r\n #define HAVE_LOG2 1\r\n #define HAVE_ATAN2 1\r\n #define HAVE_POW 1\r\n #define HAVE_NEXTAFTER 1\r\n #define HAVE_STRTOLL 1\r\n #define HAVE_STRTOULL 1\r\n #define HAVE_CBRT 1\r\n #define HAVE_STRTOLD_L 1\r\n #define HAVE_BACKTRACE 1\r\n #define HAVE_MADVISE 1\r\n #define HAVE_XMMINTRIN_H 1\r\n #define HAVE_EMMINTRIN_H 1\r\n #define HAVE_XLOCALE_H 1\r\n #define HAVE_DLFCN_H 1\r\n #define HAVE_SYS_MMAN_H 1\r\n #define HAVE___BUILTIN_ISNAN 1\r\n #define HAVE___BUILTIN_ISINF 1\r\n #define HAVE___BUILTIN_ISFINITE 1\r\n #define HAVE___BUILTIN_BSWAP32 1\r\n #define HAVE___BUILTIN_BSWAP64 1\r\n #define HAVE___BUILTIN_EXPECT 1\r\n #define HAVE___BUILTIN_MUL_OVERFLOW 1\r\n #define HAVE___BUILTIN_CPU_SUPPORTS 1\r\n #define HAVE__M_FROM_INT64 1\r\n #define HAVE__MM_LOAD_PS 1\r\n #define HAVE__MM_PREFETCH 1\r\n #define HAVE__MM_LOAD_PD 1\r\n #define HAVE___BUILTIN_PREFETCH 1\r\n #define HAVE_LINK_AVX 1\r\n #define HAVE_LINK_AVX2 1\r\n #define HAVE_XGETBV 1\r\n #define HAVE_ATTRIBUTE_NONNULL 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX2 1\r\n #define HAVE___THREAD 1\r\n #define HAVE_SINF 1\r\n #define HAVE_COSF 1\r\n #define HAVE_TANF 1\r\n #define HAVE_SINHF 1\r\n #define HAVE_COSHF 1\r\n #define HAVE_TANHF 1\r\n #define HAVE_FABSF 1\r\n #define HAVE_FLOORF 1\r\n #define HAVE_CEILF 1\r\n #define HAVE_RINTF 1\r\n #define HAVE_TRUNCF 1\r\n #define HAVE_SQRTF 1\r\n #define HAVE_LOG10F 1\r\n #define HAVE_LOGF 1\r\n #define HAVE_LOG1PF 1\r\n #define HAVE_EXPF 1\r\n #define HAVE_EXPM1F 1\r\n #define HAVE_ASINF 1\r\n #define HAVE_ACOSF 1\r\n #define HAVE_ATANF 1\r\n #define HAVE_ASINHF 1\r\n #define HAVE_ACOSHF 1\r\n #define HAVE_ATANHF 1\r\n #define HAVE_HYPOTF 1\r\n #define HAVE_ATAN2F 1\r\n #define HAVE_POWF 1\r\n #define HAVE_FMODF 1\r\n #define HAVE_MODFF 1\r\n #define HAVE_FREXPF 1\r\n #define HAVE_LDEXPF 1\r\n #define HAVE_EXP2F 1\r\n #define HAVE_LOG2F 1\r\n #define HAVE_COPYSIGNF 1\r\n #define HAVE_NEXTAFTERF 1\r\n #define HAVE_CBRTF 1\r\n #define HAVE_SINL 1\r\n #define HAVE_COSL 1\r\n #define HAVE_TANL 1\r\n #define HAVE_SINHL 1\r\n #define HAVE_COSHL 1\r\n #define HAVE_TANHL 1\r\n #define HAVE_FABSL 1\r\n #define HAVE_FLOORL 1\r\n #define HAVE_CEILL 1\r\n #define HAVE_RINTL 1\r\n #define HAVE_TRUNCL 1\r\n #define HAVE_SQRTL 1\r\n #define HAVE_LOG10L 1\r\n #define HAVE_LOGL 1\r\n #define HAVE_LOG1PL 1\r\n #define HAVE_EXPL 1\r\n #define HAVE_EXPM1L 1\r\n #define HAVE_ASINL 1\r\n #define HAVE_ACOSL 1\r\n #define HAVE_ATANL 1\r\n #define HAVE_ASINHL 1\r\n #define HAVE_ACOSHL 1\r\n #define HAVE_ATANHL 1\r\n #define HAVE_HYPOTL 1\r\n #define HAVE_ATAN2L 1\r\n #define HAVE_POWL 1\r\n #define HAVE_FMODL 1\r\n #define HAVE_MODFL 1\r\n #define HAVE_FREXPL 1\r\n #define HAVE_LDEXPL 1\r\n #define HAVE_EXP2L 1\r\n #define HAVE_LOG2L 1\r\n #define HAVE_COPYSIGNL 1\r\n #define HAVE_NEXTAFTERL 1\r\n #define HAVE_CBRTL 1\r\n #define HAVE_DECL_SIGNBIT\r\n #define HAVE_COMPLEX_H 1\r\n #define HAVE_CABS 1\r\n #define HAVE_CACOS 1\r\n #define HAVE_CACOSH 1\r\n #define HAVE_CARG 1\r\n #define HAVE_CASIN 1\r\n #define HAVE_CASINH 1\r\n #define HAVE_CATAN 1\r\n #define HAVE_CATANH 1\r\n #define HAVE_CCOS 1\r\n #define HAVE_CCOSH 1\r\n #define HAVE_CEXP 1\r\n #define HAVE_CIMAG 1\r\n #define HAVE_CLOG 1\r\n #define HAVE_CONJ 1\r\n #define HAVE_CPOW 1\r\n #define HAVE_CPROJ 1\r\n #define HAVE_CREAL 1\r\n #define HAVE_CSIN 1\r\n #define HAVE_CSINH 1\r\n #define HAVE_CSQRT 1\r\n #define HAVE_CTAN 1\r\n #define HAVE_CTANH 1\r\n #define HAVE_CABSF 1\r\n #define HAVE_CACOSF 1\r\n #define HAVE_CACOSHF 1\r\n #define HAVE_CARGF 1\r\n #define HAVE_CASINF 1\r\n #define HAVE_CASINHF 1\r\n #define HAVE_CATANF 1\r\n #define HAVE_CATANHF 1\r\n #define HAVE_CCOSF 1\r\n #define HAVE_CCOSHF 1\r\n #define HAVE_CEXPF 1\r\n #define HAVE_CIMAGF 1\r\n #define HAVE_CLOGF 1\r\n #define HAVE_CONJF 1\r\n #define HAVE_CPOWF 1\r\n #define HAVE_CPROJF 1\r\n #define HAVE_CREALF 1\r\n #define HAVE_CSINF 1\r\n #define HAVE_CSINHF 1\r\n #define HAVE_CSQRTF 1\r\n #define HAVE_CTANF 1\r\n #define HAVE_CTANHF 1\r\n #define HAVE_CABSL 1\r\n #define HAVE_CACOSL 1\r\n #define HAVE_CACOSHL 1\r\n #define HAVE_CARGL 1\r\n #define HAVE_CASINL 1\r\n #define HAVE_CASINHL 1\r\n #define HAVE_CATANL 1\r\n #define HAVE_CATANHL 1\r\n #define HAVE_CCOSL 1\r\n #define HAVE_CCOSHL 1\r\n #define HAVE_CEXPL 1\r\n #define HAVE_CIMAGL 1\r\n #define HAVE_CLOGL 1\r\n #define HAVE_CONJL 1\r\n #define HAVE_CPOWL 1\r\n #define HAVE_CPROJL 1\r\n #define HAVE_CREALL 1\r\n #define HAVE_CSINL 1\r\n #define HAVE_CSINHL 1\r\n #define HAVE_CSQRTL 1\r\n #define HAVE_CTANL 1\r\n #define HAVE_CTANHL 1\r\n #define NPY_RESTRICT restrict\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define HAVE_LDOUBLE_INTEL_EXTENDED_16_BYTES_LE 1\r\n #define NPY_PY3K 1\r\n #ifndef __cplusplus\r\n \/* #undef inline *\/\r\n #endif\r\n \r\n #ifndef _NPY_NPY_CONFIG_H_\r\n #error config.h should never be included directly, include npy_config.h instead\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n Generating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'exp' [-Wincompatible-library-redeclaration]\r\n int exp (void);\r\n ^\r\n _configtest.c:1:5: note: 'exp' is a builtin with type 'double (double)'\r\n 1 warning generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n #define NPY_SIZEOF_SHORT SIZEOF_SHORT\r\n #define NPY_SIZEOF_INT SIZEOF_INT\r\n #define NPY_SIZEOF_LONG SIZEOF_LONG\r\n #define NPY_SIZEOF_FLOAT 4\r\n #define NPY_SIZEOF_COMPLEX_FLOAT 8\r\n #define NPY_SIZEOF_DOUBLE 8\r\n #define NPY_SIZEOF_COMPLEX_DOUBLE 16\r\n #define NPY_SIZEOF_LONGDOUBLE 16\r\n #define NPY_SIZEOF_COMPLEX_LONGDOUBLE 32\r\n #define NPY_SIZEOF_PY_INTPTR_T 8\r\n #define NPY_SIZEOF_OFF_T 8\r\n #define NPY_SIZEOF_PY_LONG_LONG 8\r\n #define NPY_SIZEOF_LONGLONG 8\r\n #define NPY_NO_SMP 0\r\n #define NPY_HAVE_DECL_ISNAN\r\n #define NPY_HAVE_DECL_ISINF\r\n #define NPY_HAVE_DECL_ISFINITE\r\n #define NPY_HAVE_DECL_SIGNBIT\r\n #define NPY_USE_C99_COMPLEX 1\r\n #define NPY_HAVE_COMPLEX_DOUBLE 1\r\n #define NPY_HAVE_COMPLEX_FLOAT 1\r\n #define NPY_HAVE_COMPLEX_LONG_DOUBLE 1\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define NPY_USE_C99_FORMATS 1\r\n #define NPY_VISIBILITY_HIDDEN __attribute__((visibility(\"hidden\")))\r\n #define NPY_ABI_VERSION 0x01000009\r\n #define NPY_API_VERSION 0x0000000D\r\n \r\n #ifndef __STDC_FORMAT_MACROS\r\n #define __STDC_FORMAT_MACROS 1\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h']\r\n building extension \"numpy.core._multiarray_tests\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n building extension \"numpy.core._multiarray_umath\" sources\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_ufunc_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common' to include_dirs.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_internal.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h']\r\n building extension \"numpy.core._umath_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n building extension \"numpy.core._rational_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n building extension \"numpy.core._struct_ufunc_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n building extension \"numpy.core._operand_flag_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n building extension \"numpy.fft.fftpack_lite\" sources\r\n building extension \"numpy.linalg.lapack_lite\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n building extension \"numpy.linalg._umath_linalg\" sources\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n building extension \"numpy.random.mtrand\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/random\r\n building data_files sources\r\n build_src: building npy-pkg config files\r\n running build_py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/conftest.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_globals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/dual.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_distributor_init.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/ctypeslib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/matlib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_pytesttester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/py3k.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/_inspect.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/fromnumeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_add_newdocs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_methods.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_internal.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_string_helpers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/multiarray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/records.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup_common.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_aliased_types.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/memmap.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/overrides.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/getlimits.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype_ctypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/defchararray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/machar.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/einsumfunc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath_tests.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numerictypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_type_aliases.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/cversions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/arrayprint.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/code_generators\/generate_numpy_api.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/unixccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/numpy_distribution.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/conv_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/cpuinfo.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvc9compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/npy_pkg_config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/compat.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/misc_util.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/log.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/line_endings.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/lib2def.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/pathccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/system_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/exec_command.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/from_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/mingw32ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/extension.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/intelccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/distutils\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config_compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_ext.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_headers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_py.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_src.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/sdist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_scripts.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/bdist_rpm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/autodist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/egg_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/develop.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_data.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/gnu.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/compaq.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/intel.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/none.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/nag.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/ibm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/sun.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/lahey.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/g95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/mips.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/hpux.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/environment.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pathf95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/absoft.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/vast.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/misc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/internals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/creation.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/constants.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/ufuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/broadcasting.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/basics.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/subclassing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/indexing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/byteswapping.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/structured_arrays.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/glossary.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/common_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/crackfortran.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cb_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py2e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/func2subr.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/diagnose.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/capi_maps.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f90mod_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py_testing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/use_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/auxfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__main__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/helper.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/fftpack.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_iotools.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/mixins.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/nanfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/recfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/histograms.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/scimath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/user_array.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/format.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/twodim_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/financial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/index_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/npyio.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/stride_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arrayterator.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraysetops.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraypad.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/type_check.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_datasource.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/ufunclike.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/linalg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/extras.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/testutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/bench.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/timer_comparison.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/mrecords.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/defmatrix.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/laguerre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/_polybase.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polyutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite_e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/chebyshev.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/legendre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/print_coercion_tables.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/parameterized.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n running build_clib\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_clib\r\n building 'npymath' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n ar: adding 4 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n building 'npysort' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/quicksort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/mergesort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/heapsort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/selection.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/binsearch.c\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n 22 warnings generated.\r\n ar: adding 5 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n running build_ext\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_ext\r\n building 'numpy.core._dummy' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/dummymodule.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/dummymodule.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_dummy.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_tests.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_umath' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/core\/src\/multiarray\/alloc.c\r\n clang: numpy\/core\/src\/multiarray\/calculation.cclang: numpy\/core\/src\/multiarray\/array_assign_scalar.c\r\n clang: numpy\/core\/src\/multiarray\/convert.c\r\n \r\n clang: numpy\/core\/src\/multiarray\/ctors.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busday.c\r\n clang: numpy\/core\/src\/multiarray\/dragon4.cclang: numpy\/core\/src\/multiarray\/flagsobject.c\r\n \r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/arrayobject.c\r\n clang: numpy\/core\/src\/multiarray\/array_assign_array.c\r\n clang: numpy\/core\/src\/multiarray\/convert_datatype.c\r\n clang: numpy\/core\/src\/multiarray\/getset.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busdaycal.c\r\n clang: numpy\/core\/src\/multiarray\/buffer.c\r\n clang: numpy\/core\/src\/multiarray\/compiled_base.c\r\n clang: numpy\/core\/src\/multiarray\/hashdescr.c\r\n clang: numpy\/core\/src\/multiarray\/descriptor.c\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/conversion_utils.c\r\n clang: numpy\/core\/src\/multiarray\/item_selection.c\r\n clang: numpy\/core\/src\/multiarray\/dtype_transfer.c\r\n clang: numpy\/core\/src\/multiarray\/mapping.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n 3 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/datetime.c\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:477:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ptr = PyUnicode_AS_UNICODE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/common.c\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 6 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/nditer_pywrap.c\r\n 9 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/sequence.c\r\n clang: numpy\/core\/src\/multiarray\/shape.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n clang: numpy\/core\/src\/multiarray\/methods.c\r\n clang: numpy\/core\/src\/multiarray\/iterators.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_strings.c\r\n clang: numpy\/core\/src\/multiarray\/number.c\r\n clang: numpy\/core\/src\/multiarray\/scalarapi.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n numpy\/core\/src\/multiarray\/scalarapi.c:74:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:135:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1849:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buffer = PyUnicode_AS_DATA(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 5 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/typeinfo.c\r\n clang: numpy\/core\/src\/multiarray\/refcount.c\r\n clang: numpy\/core\/src\/multiarray\/usertypes.c\r\n clang: numpy\/core\/src\/multiarray\/multiarraymodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n clang: numpy\/core\/src\/multiarray\/vdot.c\r\n clang: numpy\/core\/src\/umath\/umathmodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n clang: numpy\/core\/src\/umath\/reduction.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n clang: numpy\/core\/src\/multiarray\/nditer_api.c\r\n 14 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/strfuncs.c\r\n numpy\/core\/src\/umath\/loops.c.src:655:18: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n result = PyEval_CallObject(tocall, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:178:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_ReprFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:195:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_StrFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 2 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/temp_elide.c\r\n clang: numpy\/core\/src\/umath\/cpuid.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n clang: numpy\/core\/src\/umath\/ufunc_object.c\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'byte_long' [-Wunused-function]\r\n byte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ubyte_long' [-Wunused-function]\r\n ubyte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'short_long' [-Wunused-function]\r\n short_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ushort_long' [-Wunused-function]\r\n ushort_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'int_long' [-Wunused-function]\r\n int_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'uint_long' [-Wunused-function]\r\n uint_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'long_long' [-Wunused-function]\r\n long_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulong_long' [-Wunused-function]\r\n ulong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longlong_long' [-Wunused-function]\r\n longlong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulonglong_long' [-Wunused-function]\r\n ulonglong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'half_long' [-Wunused-function]\r\n half_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'float_long' [-Wunused-function]\r\n float_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'double_long' [-Wunused-function]\r\n double_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longdouble_long' [-Wunused-function]\r\n longdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cfloat_long' [-Wunused-function]\r\n cfloat_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cdouble_long' [-Wunused-function]\r\n cdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'clongdouble_long' [-Wunused-function]\r\n clongdouble_long(PyObject *obj)\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/nditer_constr.c\r\n numpy\/core\/src\/umath\/ufunc_object.c:657:19: warning: comparison of integers of different signs: 'int' and 'size_t' (aka 'unsigned long') [-Wsign-compare]\r\n for (i = 0; i < len; i++) {\r\n ~ ^ ~~~\r\n clang: numpy\/core\/src\/umath\/override.c\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n clang: numpy\/core\/src\/common\/array_assign.c\r\n clang: numpy\/core\/src\/common\/ufunc_override.c\r\n clang: numpy\/core\/src\/common\/npy_longdouble.c\r\n clang: numpy\/core\/src\/common\/numpyos.c\r\n clang: numpy\/core\/src\/common\/ucsnarrow.c\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/umath\/extobj.c\r\n numpy\/core\/src\/common\/ucsnarrow.c:139:34: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n ret = (PyUnicodeObject *)PyUnicode_FromUnicode((Py_UNICODE*)buf,\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/common\/python_xerbla.c\r\n clang: numpy\/core\/src\/common\/cblasfuncs.c\r\n clang: \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c\r\n In file included from \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c:26:\r\n In file included from numpy\/core\/include\/numpy\/arrayobject.h:4:\r\n In file included from numpy\/core\/include\/numpy\/ndarrayobject.h:21:\r\n build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h:1463:1: warning: unused function '_import_array' [-Wunused-function]\r\n _import_array(void)\r\n ^\r\n 1 warning generated.\r\n 17 warnings generated.\r\n clang: numpy\/core\/src\/umath\/ufunc_type_resolution.c\r\n 4 warnings generated.\r\n 4 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/alloc.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arrayobject.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_scalar.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_array.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/buffer.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/calculation.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/compiled_base.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/common.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert_datatype.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/conversion_utils.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/ctors.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_strings.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busday.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busdaycal.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/descriptor.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dragon4.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dtype_transfer.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/flagsobject.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/getset.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/hashdescr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/item_selection.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/iterators.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/mapping.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/methods.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/multiarraymodule.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_api.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_constr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_pywrap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/number.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/refcount.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/sequence.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/shape.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalarapi.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/strfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/temp_elide.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/typeinfo.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/usertypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/vdot.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/umathmodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/reduction.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_object.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/extobj.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/cpuid.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_type_resolution.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/halffloat.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/array_assign.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/npy_longdouble.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ucsnarrow.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ufunc_override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/numpyos.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/cblasfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/python_xerbla.o build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -lnpysort -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_umath.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.core._umath_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_umath_tests.cpython-39-darwin.so\r\n building 'numpy.core._rational_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_rational_tests.cpython-39-darwin.so\r\n building 'numpy.core._struct_ufunc_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_struct_ufunc_tests.cpython-39-darwin.so\r\n building 'numpy.core._operand_flag_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_operand_flag_tests.cpython-39-darwin.so\r\n building 'numpy.fft.fftpack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n compile options: '-Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/fft\/fftpack_litemodule.c\r\n clang: numpy\/fft\/fftpack.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_lite.cpython-39-darwin.so\r\n building 'numpy.linalg.lapack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/linalg\/lapack_litemodule.c\r\n clang: numpy\/linalg\/lapack_lite\/python_xerbla.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.linalg._umath_linalg' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n numpy\/linalg\/umath_linalg.c.src:735:32: warning: unknown warning group '-Wmaybe-uninitialized', ignored [-Wunknown-warning-option]\r\n #pragma GCC diagnostic ignored \"-Wmaybe-uninitialized\"\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:541:1: warning: unused function 'dump_ufunc_object' [-Wunused-function]\r\n dump_ufunc_object(PyUFuncObject* ufunc)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:566:1: warning: unused function 'dump_linearize_data' [-Wunused-function]\r\n dump_linearize_data(const char* name, const LINEARIZE_DATA_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_FLOAT_matrix' [-Wunused-function]\r\n dump_FLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_DOUBLE_matrix' [-Wunused-function]\r\n dump_DOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CFLOAT_matrix' [-Wunused-function]\r\n dump_CFLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CDOUBLE_matrix' [-Wunused-function]\r\n dump_CDOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_FLOAT_matrix' [-Wunused-function]\r\n zero_FLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_DOUBLE_matrix' [-Wunused-function]\r\n zero_DOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CFLOAT_matrix' [-Wunused-function]\r\n zero_CFLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CDOUBLE_matrix' [-Wunused-function]\r\n zero_CDOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:1862:1: warning: unused function 'dump_geev_params' [-Wunused-function]\r\n dump_geev_params(const char *name, GEEV_PARAMS_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2132:1: warning: unused function 'init_cgeev' [-Wunused-function]\r\n init_cgeev(GEEV_PARAMS_t* params,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2213:1: warning: unused function 'process_cgeev_results' [-Wunused-function]\r\n process_cgeev_results(GEEV_PARAMS_t *NPY_UNUSED(params))\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2376:1: warning: unused function 'dump_gesdd_params' [-Wunused-function]\r\n dump_gesdd_params(const char *name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2864:1: warning: unused function 'dump_gelsd_params' [-Wunused-function]\r\n dump_gelsd_params(const char *name,\r\n ^\r\n 16 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/_umath_linalg.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.random.mtrand' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\r\n compile options: '-D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/random\/mtrand\/mtrand.c\r\n clang: numpy\/random\/mtrand\/initarray.cclang: numpy\/random\/mtrand\/randomkit.c\r\n \r\n clang: numpy\/random\/mtrand\/distributions.c\r\n numpy\/random\/mtrand\/mtrand.c:40400:34: error: no member named 'tp_print' in 'struct _typeobject'\r\n __pyx_type_6mtrand_RandomState.tp_print = 0;\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 12 warnings and 1 error generated.\r\n error: Command \"clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c numpy\/random\/mtrand\/mtrand.c -o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o -MMD -MF build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o.d\" failed with exit status 1","comment_length":11,"text":"Unable to install datasets\n** Edit **\r\nI believe there's a bug with the package when you're installing it with Python 3.9. I recommend sticking with previous versions. Thanks, @thomwolf for the insight! \r\n\r\n**Short description**\r\n\r\nI followed the instructions for installing datasets (https:\/\/huggingface.co\/docs\/datasets\/installation.html). However, while I tried to download datasets using `pip install datasets` I got a massive error message after getting stuck at \"Installing build dependencies...\" \r\n\r\nI was wondering if this problem can be fixed by creating a virtual environment, but it didn't help. Can anyone offer some advice on how to fix this issue? \r\n\r\nHere's an error message: \r\n\r\n`(env) Gas-MacBook-Pro:Downloads destiny$ pip install datasets\r\nCollecting datasets\r\n Using cached datasets-1.2.0-py3-none-any.whl (159 kB)\r\nCollecting numpy>=1.17\r\n Using cached numpy-1.19.5-cp39-cp39-macosx_10_9_x86_64.whl (15.6 MB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-2.0.0.tar.gz (58.9 MB)\r\n....\r\n\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceilf' [-Wincompatible-library-redeclaration]\r\n int ceilf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceilf' is a builtin with type 'float (float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintf' [-Wincompatible-library-redeclaration]\r\n int rintf (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintf' is a builtin with type 'float (float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncf' [-Wincompatible-library-redeclaration]\r\n int truncf (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncf' is a builtin with type 'float (float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtf' [-Wincompatible-library-redeclaration]\r\n int sqrtf (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtf' is a builtin with type 'float (float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10f' [-Wincompatible-library-redeclaration]\r\n int log10f (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10f' is a builtin with type 'float (float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logf' [-Wincompatible-library-redeclaration]\r\n int logf (void);\r\n ^\r\n _configtest.c:14:5: note: 'logf' is a builtin with type 'float (float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pf' [-Wincompatible-library-redeclaration]\r\n int log1pf (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pf' is a builtin with type 'float (float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expf' [-Wincompatible-library-redeclaration]\r\n int expf (void);\r\n ^\r\n _configtest.c:16:5: note: 'expf' is a builtin with type 'float (float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1f' [-Wincompatible-library-redeclaration]\r\n int expm1f (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1f' is a builtin with type 'float (float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinf' [-Wincompatible-library-redeclaration]\r\n int asinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinf' is a builtin with type 'float (float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosf' [-Wincompatible-library-redeclaration]\r\n int acosf (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosf' is a builtin with type 'float (float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanf' [-Wincompatible-library-redeclaration]\r\n int atanf (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanf' is a builtin with type 'float (float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhf' [-Wincompatible-library-redeclaration]\r\n int asinhf (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhf' is a builtin with type 'float (float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshf' [-Wincompatible-library-redeclaration]\r\n int acoshf (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshf' is a builtin with type 'float (float)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhf' [-Wincompatible-library-redeclaration]\r\n int atanhf (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhf' is a builtin with type 'float (float)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotf' [-Wincompatible-library-redeclaration]\r\n int hypotf (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotf' is a builtin with type 'float (float, float)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2f' [-Wincompatible-library-redeclaration]\r\n int atan2f (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2f' is a builtin with type 'float (float, float)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powf' [-Wincompatible-library-redeclaration]\r\n int powf (void);\r\n ^\r\n _configtest.c:26:5: note: 'powf' is a builtin with type 'float (float, float)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodf' [-Wincompatible-library-redeclaration]\r\n int fmodf (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodf' is a builtin with type 'float (float, float)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modff' [-Wincompatible-library-redeclaration]\r\n int modff (void);\r\n ^\r\n _configtest.c:28:5: note: 'modff' is a builtin with type 'float (float, float *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpf' [-Wincompatible-library-redeclaration]\r\n int frexpf (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpf' is a builtin with type 'float (float, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpf' [-Wincompatible-library-redeclaration]\r\n int ldexpf (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpf' is a builtin with type 'float (float, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2f' [-Wincompatible-library-redeclaration]\r\n int exp2f (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2f' is a builtin with type 'float (float)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2f' [-Wincompatible-library-redeclaration]\r\n int log2f (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2f' is a builtin with type 'float (float)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignf' [-Wincompatible-library-redeclaration]\r\n int copysignf (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignf' is a builtin with type 'float (float, float)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterf' [-Wincompatible-library-redeclaration]\r\n int nextafterf (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterf' is a builtin with type 'float (float, float)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtf' [-Wincompatible-library-redeclaration]\r\n int cbrtf (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtf' is a builtin with type 'float (float)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'sinl' [-Wincompatible-library-redeclaration]\r\n int sinl (void);\r\n ^\r\n _configtest.c:1:5: note: 'sinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cosl' [-Wincompatible-library-redeclaration]\r\n int cosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'tanl' [-Wincompatible-library-redeclaration]\r\n int tanl (void);\r\n ^\r\n _configtest.c:3:5: note: 'tanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'sinhl' [-Wincompatible-library-redeclaration]\r\n int sinhl (void);\r\n ^\r\n _configtest.c:4:5: note: 'sinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'coshl' [-Wincompatible-library-redeclaration]\r\n int coshl (void);\r\n ^\r\n _configtest.c:5:5: note: 'coshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'tanhl' [-Wincompatible-library-redeclaration]\r\n int tanhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'tanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'fabsl' [-Wincompatible-library-redeclaration]\r\n int fabsl (void);\r\n ^\r\n _configtest.c:7:5: note: 'fabsl' is a builtin with type 'long double (long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'floorl' [-Wincompatible-library-redeclaration]\r\n int floorl (void);\r\n ^\r\n _configtest.c:8:5: note: 'floorl' is a builtin with type 'long double (long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceill' [-Wincompatible-library-redeclaration]\r\n int ceill (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceill' is a builtin with type 'long double (long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintl' [-Wincompatible-library-redeclaration]\r\n int rintl (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintl' is a builtin with type 'long double (long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncl' [-Wincompatible-library-redeclaration]\r\n int truncl (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncl' is a builtin with type 'long double (long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtl' [-Wincompatible-library-redeclaration]\r\n int sqrtl (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtl' is a builtin with type 'long double (long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10l' [-Wincompatible-library-redeclaration]\r\n int log10l (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10l' is a builtin with type 'long double (long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logl' [-Wincompatible-library-redeclaration]\r\n int logl (void);\r\n ^\r\n _configtest.c:14:5: note: 'logl' is a builtin with type 'long double (long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pl' [-Wincompatible-library-redeclaration]\r\n int log1pl (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pl' is a builtin with type 'long double (long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expl' [-Wincompatible-library-redeclaration]\r\n int expl (void);\r\n ^\r\n _configtest.c:16:5: note: 'expl' is a builtin with type 'long double (long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1l' [-Wincompatible-library-redeclaration]\r\n int expm1l (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1l' is a builtin with type 'long double (long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinl' [-Wincompatible-library-redeclaration]\r\n int asinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosl' [-Wincompatible-library-redeclaration]\r\n int acosl (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanl' [-Wincompatible-library-redeclaration]\r\n int atanl (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhl' [-Wincompatible-library-redeclaration]\r\n int asinhl (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshl' [-Wincompatible-library-redeclaration]\r\n int acoshl (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhl' [-Wincompatible-library-redeclaration]\r\n int atanhl (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotl' [-Wincompatible-library-redeclaration]\r\n int hypotl (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2l' [-Wincompatible-library-redeclaration]\r\n int atan2l (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2l' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powl' [-Wincompatible-library-redeclaration]\r\n int powl (void);\r\n ^\r\n _configtest.c:26:5: note: 'powl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodl' [-Wincompatible-library-redeclaration]\r\n int fmodl (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modfl' [-Wincompatible-library-redeclaration]\r\n int modfl (void);\r\n ^\r\n _configtest.c:28:5: note: 'modfl' is a builtin with type 'long double (long double, long double *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpl' [-Wincompatible-library-redeclaration]\r\n int frexpl (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpl' is a builtin with type 'long double (long double, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpl' [-Wincompatible-library-redeclaration]\r\n int ldexpl (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpl' is a builtin with type 'long double (long double, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2l' [-Wincompatible-library-redeclaration]\r\n int exp2l (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2l' [-Wincompatible-library-redeclaration]\r\n int log2l (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignl' [-Wincompatible-library-redeclaration]\r\n int copysignl (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterl' [-Wincompatible-library-redeclaration]\r\n int nextafterl (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtl' [-Wincompatible-library-redeclaration]\r\n int cbrtl (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtl' is a builtin with type 'long double (long double)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:8:12: error: use of undeclared identifier 'HAVE_DECL_SIGNBIT'\r\n (void) HAVE_DECL_SIGNBIT;\r\n ^\r\n 1 error generated.\r\n failure.\r\n removing: _configtest.c _configtest.o\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabs' [-Wincompatible-library-redeclaration]\r\n int cabs (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabs' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacos' [-Wincompatible-library-redeclaration]\r\n int cacos (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacosh' [-Wincompatible-library-redeclaration]\r\n int cacosh (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'carg' [-Wincompatible-library-redeclaration]\r\n int carg (void);\r\n ^\r\n _configtest.c:4:5: note: 'carg' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casin' [-Wincompatible-library-redeclaration]\r\n int casin (void);\r\n ^\r\n _configtest.c:5:5: note: 'casin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinh' [-Wincompatible-library-redeclaration]\r\n int casinh (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catan' [-Wincompatible-library-redeclaration]\r\n int catan (void);\r\n ^\r\n _configtest.c:7:5: note: 'catan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanh' [-Wincompatible-library-redeclaration]\r\n int catanh (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccos' [-Wincompatible-library-redeclaration]\r\n int ccos (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccosh' [-Wincompatible-library-redeclaration]\r\n int ccosh (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexp' [-Wincompatible-library-redeclaration]\r\n int cexp (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexp' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimag' [-Wincompatible-library-redeclaration]\r\n int cimag (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimag' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clog' [-Wincompatible-library-redeclaration]\r\n int clog (void);\r\n ^\r\n _configtest.c:13:5: note: 'clog' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conj' [-Wincompatible-library-redeclaration]\r\n int conj (void);\r\n ^\r\n _configtest.c:14:5: note: 'conj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpow' [-Wincompatible-library-redeclaration]\r\n int cpow (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpow' is a builtin with type '_Complex double (_Complex double, _Complex double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cproj' [-Wincompatible-library-redeclaration]\r\n int cproj (void);\r\n ^\r\n _configtest.c:16:5: note: 'cproj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creal' [-Wincompatible-library-redeclaration]\r\n int creal (void);\r\n ^\r\n _configtest.c:17:5: note: 'creal' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csin' [-Wincompatible-library-redeclaration]\r\n int csin (void);\r\n ^\r\n _configtest.c:18:5: note: 'csin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinh' [-Wincompatible-library-redeclaration]\r\n int csinh (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrt' [-Wincompatible-library-redeclaration]\r\n int csqrt (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrt' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctan' [-Wincompatible-library-redeclaration]\r\n int ctan (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanh' [-Wincompatible-library-redeclaration]\r\n int ctanh (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanh' is a builtin with type '_Complex double (_Complex double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsf' [-Wincompatible-library-redeclaration]\r\n int cabsf (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosf' [-Wincompatible-library-redeclaration]\r\n int cacosf (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshf' [-Wincompatible-library-redeclaration]\r\n int cacoshf (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargf' [-Wincompatible-library-redeclaration]\r\n int cargf (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinf' [-Wincompatible-library-redeclaration]\r\n int casinf (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhf' [-Wincompatible-library-redeclaration]\r\n int casinhf (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanf' [-Wincompatible-library-redeclaration]\r\n int catanf (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhf' [-Wincompatible-library-redeclaration]\r\n int catanhf (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosf' [-Wincompatible-library-redeclaration]\r\n int ccosf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshf' [-Wincompatible-library-redeclaration]\r\n int ccoshf (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpf' [-Wincompatible-library-redeclaration]\r\n int cexpf (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagf' [-Wincompatible-library-redeclaration]\r\n int cimagf (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogf' [-Wincompatible-library-redeclaration]\r\n int clogf (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjf' [-Wincompatible-library-redeclaration]\r\n int conjf (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowf' [-Wincompatible-library-redeclaration]\r\n int cpowf (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowf' is a builtin with type '_Complex float (_Complex float, _Complex float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojf' [-Wincompatible-library-redeclaration]\r\n int cprojf (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'crealf' [-Wincompatible-library-redeclaration]\r\n int crealf (void);\r\n ^\r\n _configtest.c:17:5: note: 'crealf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinf' [-Wincompatible-library-redeclaration]\r\n int csinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhf' [-Wincompatible-library-redeclaration]\r\n int csinhf (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtf' [-Wincompatible-library-redeclaration]\r\n int csqrtf (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanf' [-Wincompatible-library-redeclaration]\r\n int ctanf (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhf' [-Wincompatible-library-redeclaration]\r\n int ctanhf (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhf' is a builtin with type '_Complex float (_Complex float)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsl' [-Wincompatible-library-redeclaration]\r\n int cabsl (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosl' [-Wincompatible-library-redeclaration]\r\n int cacosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshl' [-Wincompatible-library-redeclaration]\r\n int cacoshl (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargl' [-Wincompatible-library-redeclaration]\r\n int cargl (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinl' [-Wincompatible-library-redeclaration]\r\n int casinl (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhl' [-Wincompatible-library-redeclaration]\r\n int casinhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanl' [-Wincompatible-library-redeclaration]\r\n int catanl (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhl' [-Wincompatible-library-redeclaration]\r\n int catanhl (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosl' [-Wincompatible-library-redeclaration]\r\n int ccosl (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshl' [-Wincompatible-library-redeclaration]\r\n int ccoshl (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpl' [-Wincompatible-library-redeclaration]\r\n int cexpl (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagl' [-Wincompatible-library-redeclaration]\r\n int cimagl (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogl' [-Wincompatible-library-redeclaration]\r\n int clogl (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjl' [-Wincompatible-library-redeclaration]\r\n int conjl (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowl' [-Wincompatible-library-redeclaration]\r\n int cpowl (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowl' is a builtin with type '_Complex long double (_Complex long double, _Complex long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojl' [-Wincompatible-library-redeclaration]\r\n int cprojl (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creall' [-Wincompatible-library-redeclaration]\r\n int creall (void);\r\n ^\r\n _configtest.c:17:5: note: 'creall' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinl' [-Wincompatible-library-redeclaration]\r\n int csinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhl' [-Wincompatible-library-redeclaration]\r\n int csinhl (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtl' [-Wincompatible-library-redeclaration]\r\n int csqrtl (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanl' [-Wincompatible-library-redeclaration]\r\n int ctanl (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhl' [-Wincompatible-library-redeclaration]\r\n int ctanhl (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:2:12: warning: unused function 'static_func' [-Wunused-function]\r\n static int static_func (char * restrict a)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:3:19: warning: unused function 'static_func' [-Wunused-function]\r\n static inline int static_func (void)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h\r\n #define SIZEOF_PY_INTPTR_T 8\r\n #define SIZEOF_OFF_T 8\r\n #define SIZEOF_PY_LONG_LONG 8\r\n #define MATHLIB\r\n #define HAVE_SIN 1\r\n #define HAVE_COS 1\r\n #define HAVE_TAN 1\r\n #define HAVE_SINH 1\r\n #define HAVE_COSH 1\r\n #define HAVE_TANH 1\r\n #define HAVE_FABS 1\r\n #define HAVE_FLOOR 1\r\n #define HAVE_CEIL 1\r\n #define HAVE_SQRT 1\r\n #define HAVE_LOG10 1\r\n #define HAVE_LOG 1\r\n #define HAVE_EXP 1\r\n #define HAVE_ASIN 1\r\n #define HAVE_ACOS 1\r\n #define HAVE_ATAN 1\r\n #define HAVE_FMOD 1\r\n #define HAVE_MODF 1\r\n #define HAVE_FREXP 1\r\n #define HAVE_LDEXP 1\r\n #define HAVE_RINT 1\r\n #define HAVE_TRUNC 1\r\n #define HAVE_EXP2 1\r\n #define HAVE_LOG2 1\r\n #define HAVE_ATAN2 1\r\n #define HAVE_POW 1\r\n #define HAVE_NEXTAFTER 1\r\n #define HAVE_STRTOLL 1\r\n #define HAVE_STRTOULL 1\r\n #define HAVE_CBRT 1\r\n #define HAVE_STRTOLD_L 1\r\n #define HAVE_BACKTRACE 1\r\n #define HAVE_MADVISE 1\r\n #define HAVE_XMMINTRIN_H 1\r\n #define HAVE_EMMINTRIN_H 1\r\n #define HAVE_XLOCALE_H 1\r\n #define HAVE_DLFCN_H 1\r\n #define HAVE_SYS_MMAN_H 1\r\n #define HAVE___BUILTIN_ISNAN 1\r\n #define HAVE___BUILTIN_ISINF 1\r\n #define HAVE___BUILTIN_ISFINITE 1\r\n #define HAVE___BUILTIN_BSWAP32 1\r\n #define HAVE___BUILTIN_BSWAP64 1\r\n #define HAVE___BUILTIN_EXPECT 1\r\n #define HAVE___BUILTIN_MUL_OVERFLOW 1\r\n #define HAVE___BUILTIN_CPU_SUPPORTS 1\r\n #define HAVE__M_FROM_INT64 1\r\n #define HAVE__MM_LOAD_PS 1\r\n #define HAVE__MM_PREFETCH 1\r\n #define HAVE__MM_LOAD_PD 1\r\n #define HAVE___BUILTIN_PREFETCH 1\r\n #define HAVE_LINK_AVX 1\r\n #define HAVE_LINK_AVX2 1\r\n #define HAVE_XGETBV 1\r\n #define HAVE_ATTRIBUTE_NONNULL 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX2 1\r\n #define HAVE___THREAD 1\r\n #define HAVE_SINF 1\r\n #define HAVE_COSF 1\r\n #define HAVE_TANF 1\r\n #define HAVE_SINHF 1\r\n #define HAVE_COSHF 1\r\n #define HAVE_TANHF 1\r\n #define HAVE_FABSF 1\r\n #define HAVE_FLOORF 1\r\n #define HAVE_CEILF 1\r\n #define HAVE_RINTF 1\r\n #define HAVE_TRUNCF 1\r\n #define HAVE_SQRTF 1\r\n #define HAVE_LOG10F 1\r\n #define HAVE_LOGF 1\r\n #define HAVE_LOG1PF 1\r\n #define HAVE_EXPF 1\r\n #define HAVE_EXPM1F 1\r\n #define HAVE_ASINF 1\r\n #define HAVE_ACOSF 1\r\n #define HAVE_ATANF 1\r\n #define HAVE_ASINHF 1\r\n #define HAVE_ACOSHF 1\r\n #define HAVE_ATANHF 1\r\n #define HAVE_HYPOTF 1\r\n #define HAVE_ATAN2F 1\r\n #define HAVE_POWF 1\r\n #define HAVE_FMODF 1\r\n #define HAVE_MODFF 1\r\n #define HAVE_FREXPF 1\r\n #define HAVE_LDEXPF 1\r\n #define HAVE_EXP2F 1\r\n #define HAVE_LOG2F 1\r\n #define HAVE_COPYSIGNF 1\r\n #define HAVE_NEXTAFTERF 1\r\n #define HAVE_CBRTF 1\r\n #define HAVE_SINL 1\r\n #define HAVE_COSL 1\r\n #define HAVE_TANL 1\r\n #define HAVE_SINHL 1\r\n #define HAVE_COSHL 1\r\n #define HAVE_TANHL 1\r\n #define HAVE_FABSL 1\r\n #define HAVE_FLOORL 1\r\n #define HAVE_CEILL 1\r\n #define HAVE_RINTL 1\r\n #define HAVE_TRUNCL 1\r\n #define HAVE_SQRTL 1\r\n #define HAVE_LOG10L 1\r\n #define HAVE_LOGL 1\r\n #define HAVE_LOG1PL 1\r\n #define HAVE_EXPL 1\r\n #define HAVE_EXPM1L 1\r\n #define HAVE_ASINL 1\r\n #define HAVE_ACOSL 1\r\n #define HAVE_ATANL 1\r\n #define HAVE_ASINHL 1\r\n #define HAVE_ACOSHL 1\r\n #define HAVE_ATANHL 1\r\n #define HAVE_HYPOTL 1\r\n #define HAVE_ATAN2L 1\r\n #define HAVE_POWL 1\r\n #define HAVE_FMODL 1\r\n #define HAVE_MODFL 1\r\n #define HAVE_FREXPL 1\r\n #define HAVE_LDEXPL 1\r\n #define HAVE_EXP2L 1\r\n #define HAVE_LOG2L 1\r\n #define HAVE_COPYSIGNL 1\r\n #define HAVE_NEXTAFTERL 1\r\n #define HAVE_CBRTL 1\r\n #define HAVE_DECL_SIGNBIT\r\n #define HAVE_COMPLEX_H 1\r\n #define HAVE_CABS 1\r\n #define HAVE_CACOS 1\r\n #define HAVE_CACOSH 1\r\n #define HAVE_CARG 1\r\n #define HAVE_CASIN 1\r\n #define HAVE_CASINH 1\r\n #define HAVE_CATAN 1\r\n #define HAVE_CATANH 1\r\n #define HAVE_CCOS 1\r\n #define HAVE_CCOSH 1\r\n #define HAVE_CEXP 1\r\n #define HAVE_CIMAG 1\r\n #define HAVE_CLOG 1\r\n #define HAVE_CONJ 1\r\n #define HAVE_CPOW 1\r\n #define HAVE_CPROJ 1\r\n #define HAVE_CREAL 1\r\n #define HAVE_CSIN 1\r\n #define HAVE_CSINH 1\r\n #define HAVE_CSQRT 1\r\n #define HAVE_CTAN 1\r\n #define HAVE_CTANH 1\r\n #define HAVE_CABSF 1\r\n #define HAVE_CACOSF 1\r\n #define HAVE_CACOSHF 1\r\n #define HAVE_CARGF 1\r\n #define HAVE_CASINF 1\r\n #define HAVE_CASINHF 1\r\n #define HAVE_CATANF 1\r\n #define HAVE_CATANHF 1\r\n #define HAVE_CCOSF 1\r\n #define HAVE_CCOSHF 1\r\n #define HAVE_CEXPF 1\r\n #define HAVE_CIMAGF 1\r\n #define HAVE_CLOGF 1\r\n #define HAVE_CONJF 1\r\n #define HAVE_CPOWF 1\r\n #define HAVE_CPROJF 1\r\n #define HAVE_CREALF 1\r\n #define HAVE_CSINF 1\r\n #define HAVE_CSINHF 1\r\n #define HAVE_CSQRTF 1\r\n #define HAVE_CTANF 1\r\n #define HAVE_CTANHF 1\r\n #define HAVE_CABSL 1\r\n #define HAVE_CACOSL 1\r\n #define HAVE_CACOSHL 1\r\n #define HAVE_CARGL 1\r\n #define HAVE_CASINL 1\r\n #define HAVE_CASINHL 1\r\n #define HAVE_CATANL 1\r\n #define HAVE_CATANHL 1\r\n #define HAVE_CCOSL 1\r\n #define HAVE_CCOSHL 1\r\n #define HAVE_CEXPL 1\r\n #define HAVE_CIMAGL 1\r\n #define HAVE_CLOGL 1\r\n #define HAVE_CONJL 1\r\n #define HAVE_CPOWL 1\r\n #define HAVE_CPROJL 1\r\n #define HAVE_CREALL 1\r\n #define HAVE_CSINL 1\r\n #define HAVE_CSINHL 1\r\n #define HAVE_CSQRTL 1\r\n #define HAVE_CTANL 1\r\n #define HAVE_CTANHL 1\r\n #define NPY_RESTRICT restrict\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define HAVE_LDOUBLE_INTEL_EXTENDED_16_BYTES_LE 1\r\n #define NPY_PY3K 1\r\n #ifndef __cplusplus\r\n \/* #undef inline *\/\r\n #endif\r\n \r\n #ifndef _NPY_NPY_CONFIG_H_\r\n #error config.h should never be included directly, include npy_config.h instead\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n Generating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'exp' [-Wincompatible-library-redeclaration]\r\n int exp (void);\r\n ^\r\n _configtest.c:1:5: note: 'exp' is a builtin with type 'double (double)'\r\n 1 warning generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n #define NPY_SIZEOF_SHORT SIZEOF_SHORT\r\n #define NPY_SIZEOF_INT SIZEOF_INT\r\n #define NPY_SIZEOF_LONG SIZEOF_LONG\r\n #define NPY_SIZEOF_FLOAT 4\r\n #define NPY_SIZEOF_COMPLEX_FLOAT 8\r\n #define NPY_SIZEOF_DOUBLE 8\r\n #define NPY_SIZEOF_COMPLEX_DOUBLE 16\r\n #define NPY_SIZEOF_LONGDOUBLE 16\r\n #define NPY_SIZEOF_COMPLEX_LONGDOUBLE 32\r\n #define NPY_SIZEOF_PY_INTPTR_T 8\r\n #define NPY_SIZEOF_OFF_T 8\r\n #define NPY_SIZEOF_PY_LONG_LONG 8\r\n #define NPY_SIZEOF_LONGLONG 8\r\n #define NPY_NO_SMP 0\r\n #define NPY_HAVE_DECL_ISNAN\r\n #define NPY_HAVE_DECL_ISINF\r\n #define NPY_HAVE_DECL_ISFINITE\r\n #define NPY_HAVE_DECL_SIGNBIT\r\n #define NPY_USE_C99_COMPLEX 1\r\n #define NPY_HAVE_COMPLEX_DOUBLE 1\r\n #define NPY_HAVE_COMPLEX_FLOAT 1\r\n #define NPY_HAVE_COMPLEX_LONG_DOUBLE 1\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define NPY_USE_C99_FORMATS 1\r\n #define NPY_VISIBILITY_HIDDEN __attribute__((visibility(\"hidden\")))\r\n #define NPY_ABI_VERSION 0x01000009\r\n #define NPY_API_VERSION 0x0000000D\r\n \r\n #ifndef __STDC_FORMAT_MACROS\r\n #define __STDC_FORMAT_MACROS 1\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h']\r\n building extension \"numpy.core._multiarray_tests\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n building extension \"numpy.core._multiarray_umath\" sources\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_ufunc_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common' to include_dirs.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_internal.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h']\r\n building extension \"numpy.core._umath_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n building extension \"numpy.core._rational_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n building extension \"numpy.core._struct_ufunc_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n building extension \"numpy.core._operand_flag_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n building extension \"numpy.fft.fftpack_lite\" sources\r\n building extension \"numpy.linalg.lapack_lite\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n building extension \"numpy.linalg._umath_linalg\" sources\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n building extension \"numpy.random.mtrand\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/random\r\n building data_files sources\r\n build_src: building npy-pkg config files\r\n running build_py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/conftest.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_globals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/dual.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_distributor_init.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/ctypeslib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/matlib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_pytesttester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/py3k.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/_inspect.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/fromnumeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_add_newdocs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_methods.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_internal.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_string_helpers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/multiarray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/records.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup_common.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_aliased_types.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/memmap.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/overrides.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/getlimits.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype_ctypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/defchararray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/machar.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/einsumfunc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath_tests.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numerictypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_type_aliases.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/cversions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/arrayprint.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/code_generators\/generate_numpy_api.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/unixccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/numpy_distribution.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/conv_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/cpuinfo.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvc9compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/npy_pkg_config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/compat.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/misc_util.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/log.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/line_endings.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/lib2def.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/pathccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/system_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/exec_command.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/from_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/mingw32ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/extension.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/intelccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/distutils\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config_compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_ext.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_headers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_py.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_src.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/sdist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_scripts.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/bdist_rpm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/autodist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/egg_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/develop.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_data.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/gnu.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/compaq.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/intel.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/none.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/nag.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/ibm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/sun.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/lahey.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/g95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/mips.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/hpux.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/environment.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pathf95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/absoft.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/vast.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/misc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/internals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/creation.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/constants.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/ufuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/broadcasting.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/basics.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/subclassing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/indexing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/byteswapping.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/structured_arrays.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/glossary.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/common_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/crackfortran.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cb_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py2e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/func2subr.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/diagnose.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/capi_maps.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f90mod_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py_testing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/use_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/auxfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__main__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/helper.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/fftpack.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_iotools.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/mixins.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/nanfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/recfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/histograms.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/scimath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/user_array.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/format.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/twodim_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/financial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/index_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/npyio.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/stride_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arrayterator.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraysetops.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraypad.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/type_check.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_datasource.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/ufunclike.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/linalg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/extras.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/testutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/bench.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/timer_comparison.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/mrecords.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/defmatrix.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/laguerre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/_polybase.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polyutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite_e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/chebyshev.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/legendre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/print_coercion_tables.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/parameterized.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n running build_clib\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_clib\r\n building 'npymath' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n ar: adding 4 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n building 'npysort' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/quicksort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/mergesort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/heapsort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/selection.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/binsearch.c\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n 22 warnings generated.\r\n ar: adding 5 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n running build_ext\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_ext\r\n building 'numpy.core._dummy' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/dummymodule.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/dummymodule.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_dummy.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_tests.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_umath' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/core\/src\/multiarray\/alloc.c\r\n clang: numpy\/core\/src\/multiarray\/calculation.cclang: numpy\/core\/src\/multiarray\/array_assign_scalar.c\r\n clang: numpy\/core\/src\/multiarray\/convert.c\r\n \r\n clang: numpy\/core\/src\/multiarray\/ctors.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busday.c\r\n clang: numpy\/core\/src\/multiarray\/dragon4.cclang: numpy\/core\/src\/multiarray\/flagsobject.c\r\n \r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/arrayobject.c\r\n clang: numpy\/core\/src\/multiarray\/array_assign_array.c\r\n clang: numpy\/core\/src\/multiarray\/convert_datatype.c\r\n clang: numpy\/core\/src\/multiarray\/getset.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busdaycal.c\r\n clang: numpy\/core\/src\/multiarray\/buffer.c\r\n clang: numpy\/core\/src\/multiarray\/compiled_base.c\r\n clang: numpy\/core\/src\/multiarray\/hashdescr.c\r\n clang: numpy\/core\/src\/multiarray\/descriptor.c\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/conversion_utils.c\r\n clang: numpy\/core\/src\/multiarray\/item_selection.c\r\n clang: numpy\/core\/src\/multiarray\/dtype_transfer.c\r\n clang: numpy\/core\/src\/multiarray\/mapping.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n 3 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/datetime.c\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:477:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ptr = PyUnicode_AS_UNICODE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/common.c\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 6 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/nditer_pywrap.c\r\n 9 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/sequence.c\r\n clang: numpy\/core\/src\/multiarray\/shape.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n clang: numpy\/core\/src\/multiarray\/methods.c\r\n clang: numpy\/core\/src\/multiarray\/iterators.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_strings.c\r\n clang: numpy\/core\/src\/multiarray\/number.c\r\n clang: numpy\/core\/src\/multiarray\/scalarapi.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n numpy\/core\/src\/multiarray\/scalarapi.c:74:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:135:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1849:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buffer = PyUnicode_AS_DATA(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 5 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/typeinfo.c\r\n clang: numpy\/core\/src\/multiarray\/refcount.c\r\n clang: numpy\/core\/src\/multiarray\/usertypes.c\r\n clang: numpy\/core\/src\/multiarray\/multiarraymodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n clang: numpy\/core\/src\/multiarray\/vdot.c\r\n clang: numpy\/core\/src\/umath\/umathmodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n clang: numpy\/core\/src\/umath\/reduction.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n clang: numpy\/core\/src\/multiarray\/nditer_api.c\r\n 14 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/strfuncs.c\r\n numpy\/core\/src\/umath\/loops.c.src:655:18: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n result = PyEval_CallObject(tocall, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:178:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_ReprFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:195:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_StrFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 2 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/temp_elide.c\r\n clang: numpy\/core\/src\/umath\/cpuid.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n clang: numpy\/core\/src\/umath\/ufunc_object.c\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'byte_long' [-Wunused-function]\r\n byte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ubyte_long' [-Wunused-function]\r\n ubyte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'short_long' [-Wunused-function]\r\n short_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ushort_long' [-Wunused-function]\r\n ushort_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'int_long' [-Wunused-function]\r\n int_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'uint_long' [-Wunused-function]\r\n uint_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'long_long' [-Wunused-function]\r\n long_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulong_long' [-Wunused-function]\r\n ulong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longlong_long' [-Wunused-function]\r\n longlong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulonglong_long' [-Wunused-function]\r\n ulonglong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'half_long' [-Wunused-function]\r\n half_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'float_long' [-Wunused-function]\r\n float_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'double_long' [-Wunused-function]\r\n double_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longdouble_long' [-Wunused-function]\r\n longdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cfloat_long' [-Wunused-function]\r\n cfloat_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cdouble_long' [-Wunused-function]\r\n cdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'clongdouble_long' [-Wunused-function]\r\n clongdouble_long(PyObject *obj)\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/nditer_constr.c\r\n numpy\/core\/src\/umath\/ufunc_object.c:657:19: warning: comparison of integers of different signs: 'int' and 'size_t' (aka 'unsigned long') [-Wsign-compare]\r\n for (i = 0; i < len; i++) {\r\n ~ ^ ~~~\r\n clang: numpy\/core\/src\/umath\/override.c\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n clang: numpy\/core\/src\/common\/array_assign.c\r\n clang: numpy\/core\/src\/common\/ufunc_override.c\r\n clang: numpy\/core\/src\/common\/npy_longdouble.c\r\n clang: numpy\/core\/src\/common\/numpyos.c\r\n clang: numpy\/core\/src\/common\/ucsnarrow.c\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/umath\/extobj.c\r\n numpy\/core\/src\/common\/ucsnarrow.c:139:34: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n ret = (PyUnicodeObject *)PyUnicode_FromUnicode((Py_UNICODE*)buf,\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/common\/python_xerbla.c\r\n clang: numpy\/core\/src\/common\/cblasfuncs.c\r\n clang: \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c\r\n In file included from \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c:26:\r\n In file included from numpy\/core\/include\/numpy\/arrayobject.h:4:\r\n In file included from numpy\/core\/include\/numpy\/ndarrayobject.h:21:\r\n build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h:1463:1: warning: unused function '_import_array' [-Wunused-function]\r\n _import_array(void)\r\n ^\r\n 1 warning generated.\r\n 17 warnings generated.\r\n clang: numpy\/core\/src\/umath\/ufunc_type_resolution.c\r\n 4 warnings generated.\r\n 4 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/alloc.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arrayobject.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_scalar.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_array.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/buffer.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/calculation.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/compiled_base.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/common.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert_datatype.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/conversion_utils.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/ctors.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_strings.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busday.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busdaycal.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/descriptor.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dragon4.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dtype_transfer.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/flagsobject.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/getset.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/hashdescr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/item_selection.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/iterators.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/mapping.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/methods.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/multiarraymodule.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_api.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_constr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_pywrap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/number.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/refcount.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/sequence.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/shape.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalarapi.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/strfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/temp_elide.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/typeinfo.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/usertypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/vdot.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/umathmodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/reduction.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_object.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/extobj.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/cpuid.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_type_resolution.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/halffloat.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/array_assign.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/npy_longdouble.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ucsnarrow.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ufunc_override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/numpyos.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/cblasfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/python_xerbla.o build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -lnpysort -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_umath.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.core._umath_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_umath_tests.cpython-39-darwin.so\r\n building 'numpy.core._rational_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_rational_tests.cpython-39-darwin.so\r\n building 'numpy.core._struct_ufunc_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_struct_ufunc_tests.cpython-39-darwin.so\r\n building 'numpy.core._operand_flag_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_operand_flag_tests.cpython-39-darwin.so\r\n building 'numpy.fft.fftpack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n compile options: '-Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/fft\/fftpack_litemodule.c\r\n clang: numpy\/fft\/fftpack.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_lite.cpython-39-darwin.so\r\n building 'numpy.linalg.lapack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/linalg\/lapack_litemodule.c\r\n clang: numpy\/linalg\/lapack_lite\/python_xerbla.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.linalg._umath_linalg' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n numpy\/linalg\/umath_linalg.c.src:735:32: warning: unknown warning group '-Wmaybe-uninitialized', ignored [-Wunknown-warning-option]\r\n #pragma GCC diagnostic ignored \"-Wmaybe-uninitialized\"\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:541:1: warning: unused function 'dump_ufunc_object' [-Wunused-function]\r\n dump_ufunc_object(PyUFuncObject* ufunc)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:566:1: warning: unused function 'dump_linearize_data' [-Wunused-function]\r\n dump_linearize_data(const char* name, const LINEARIZE_DATA_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_FLOAT_matrix' [-Wunused-function]\r\n dump_FLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_DOUBLE_matrix' [-Wunused-function]\r\n dump_DOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CFLOAT_matrix' [-Wunused-function]\r\n dump_CFLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CDOUBLE_matrix' [-Wunused-function]\r\n dump_CDOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_FLOAT_matrix' [-Wunused-function]\r\n zero_FLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_DOUBLE_matrix' [-Wunused-function]\r\n zero_DOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CFLOAT_matrix' [-Wunused-function]\r\n zero_CFLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CDOUBLE_matrix' [-Wunused-function]\r\n zero_CDOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:1862:1: warning: unused function 'dump_geev_params' [-Wunused-function]\r\n dump_geev_params(const char *name, GEEV_PARAMS_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2132:1: warning: unused function 'init_cgeev' [-Wunused-function]\r\n init_cgeev(GEEV_PARAMS_t* params,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2213:1: warning: unused function 'process_cgeev_results' [-Wunused-function]\r\n process_cgeev_results(GEEV_PARAMS_t *NPY_UNUSED(params))\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2376:1: warning: unused function 'dump_gesdd_params' [-Wunused-function]\r\n dump_gesdd_params(const char *name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2864:1: warning: unused function 'dump_gelsd_params' [-Wunused-function]\r\n dump_gelsd_params(const char *name,\r\n ^\r\n 16 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/_umath_linalg.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.random.mtrand' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\r\n compile options: '-D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/random\/mtrand\/mtrand.c\r\n clang: numpy\/random\/mtrand\/initarray.cclang: numpy\/random\/mtrand\/randomkit.c\r\n \r\n clang: numpy\/random\/mtrand\/distributions.c\r\n numpy\/random\/mtrand\/mtrand.c:40400:34: error: no member named 'tp_print' in 'struct _typeobject'\r\n __pyx_type_6mtrand_RandomState.tp_print = 0;\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 12 warnings and 1 error generated.\r\n error: Command \"clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c numpy\/random\/mtrand\/mtrand.c -o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o -MMD -MF build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o.d\" failed with exit status 1\nThanks, @thomwolf! I fixed the issue by downgrading python to 3.7. ","embeddings":[-0.372736305,-0.0714854077,-0.0845708922,0.2079993039,0.2768215835,0.0631752834,-0.2505984604,0.0786435232,-0.2070377618,-0.0340085514,0.0305880085,0.3349942863,-0.1231444255,0.0895890892,0.0807427391,0.0722528622,0.189591229,0.2151816189,-0.3760947287,0.0339735448,0.0351168886,0.1739135981,-0.1176334247,-0.0769590884,-0.1111136004,-0.2832199335,0.0382908024,0.1094798371,-0.532479167,-0.3132808805,0.291446507,0.0252360553,-0.0163556207,0.5489337444,-0.0001225818,0.1088191941,0.1684155613,0.018390229,-0.3556295037,-0.4525223374,-0.2284404039,-0.4977667034,0.1029686555,0.0328003541,0.2697275281,-0.2269828916,-0.1955371499,-0.3135961592,0.7321167588,0.186695531,0.1548049748,0.3139568865,0.1280289441,0.0492491126,0.1173379719,0.3006310165,-0.3177905381,0.3576257825,0.2570489943,-0.1687269062,0.2200257182,0.1589555889,-0.2279057503,0.2082943469,0.5441326499,-0.1704113036,-0.0771712363,-0.4264783263,-0.2593541741,0.2732714415,0.5033100843,-0.2703568041,-0.6291657686,-0.0928478241,0.0548009649,-0.2381634116,0.3260372877,0.223864004,-0.2062843293,0.0802830532,-0.4288931787,-0.280636698,-0.1864999384,0.2670033574,-0.0238301586,0.164580822,-0.1459159702,0.1176900715,0.2985162139,-0.0034400814,0.2251620293,0.0791203454,-0.0705096424,0.0032631624,-0.2801685035,0.0655268356,-0.2036794722,0.2356178463,0.1992911696,0.1688992977,-0.1797399819,-0.2005181015,-0.2243610919,0.0616602823,0.066762954,0.2027508169,-0.1514637321,0.1150361076,0.3352227509,0.2663400173,0.4560080171,0.0843001902,0.0707917213,-0.1902450472,-0.209478572,-0.1111337766,0.1524448395,-0.3028428555,-0.4012164772,0.1194737256,0.0135395043,0.1056581885,0.0688990727,0.1510053128,-0.1464695483,0.1697680354,0.2481278181,0.3707648218,-0.232209444,0.0304407589,-0.0021012723,-0.0312458072,-0.2525970638,0.0168092307,0.1831142902,-0.1020761356,0.3657963574,0.2121508121,0.1536038369,0.1180837601,0.2853684127,0.0168460198,-0.191518262,0.4379527569,-0.2559596002,0.2257401049,0.0813604146,0.3052143753,0.0385886207,-0.0087099858,0.1200957373,-0.0894460753,-0.1464142799,0.0493404903,-0.2953626513,0.051196415,-0.16042234,-0.0320122093,0.0422822423,-0.0837064162,0.015641503,-0.2536138892,0.0998989642,-0.1036708131,0.3004768193,0.414516747,-0.51960361,-0.038707234,-0.1943658739,-0.1891062409,-0.0495771132,0.1207961589,-0.1165530235,0.1363425553,-0.164071247,-0.4210309684,0.2341786772,-0.3646256328,-0.367303431,0.1608417332,-0.1646376401,-0.086746715,0.0553342775,0.0455297008,-0.0180610958,0.0614581928,-0.0477216803,0.326702714,-0.1066503674,-0.2653565109,0.0084401397,-0.5503122211,0.0675886497,0.0629741848,-0.0602851324,0.0150617491,-0.0644270927,0.139841333,0.0511045381,0.0582192577,-0.0014922149,0.6988541484,0.416780889,0.0951713175,-0.2119343877,-0.2936756313,-0.3661433756,0.1583058536,0.0255879425,0.0643135011,-0.3344601095,-0.0186114982,-0.1848731637,0.128934145,-0.0381260812,-0.0527399741,-0.0203619599,0.0027848945,0.2437096834,-0.2807203233,-0.233386904,0.5044182539,0.1574977785,0.2694315314,-0.4862544239,0.0366727933,-0.3219184875,-0.1439168453,0.1915305704,0.2339345664,0.0553197712,-0.2622070611,-0.1443571001,0.3139041066,-0.1654947251,-0.1707349569,-0.0076072491,0.1440196335,0.2182890773,-0.2099311203,0.2413927317,-0.0720721483,-0.1406363845,0.0983242542,-0.0743637085,-0.1697815061,0.0582351796,0.3382672966,-0.0074652079,0.3404535055,0.3141412735,0.0584230796,-0.1450206935,-0.4074391723,0.396992296,-0.0027272592,0.4638343155,-0.1445932984,0.0697035044,0.053041961,-0.1363226324,0.0916765481,-0.0233560391,0.0049676136,-0.072430402,0.1785400212,0.185349673,0.0348233506,0.5852743983,0.0619082898,-0.1347470284,0.1598725617,-0.0658946782,-0.0729199722,0.0373055674,0.1743800044,-0.0199271459,-0.0330145024,0.0282345489,-0.0624522939,-0.0412191078,-0.207671389,0.1253645271,0.2953568995,-0.3419783413,0.1114570275,-0.5203249454,-0.0051726159,-0.0906571373,-0.0719850361,-0.4134817719,-0.2235808372,-0.1058893874,0.1500547081,0.2082414627,0.1049688607,0.1166992337,0.0297720432,0.001696527,-0.209155798,-0.1628497243,-0.1249261945,-0.0730912238,-0.0193224195,0.1679847986,0.1203907058,0.4714294374,-0.0997886732,-0.1613499224,-0.4218867421,-0.261598736,0.107161954,-0.398339957,0.742423296,0.469727993,0.202538535,0.0647644475,0.1159555018,0.1875865012,-0.1887231469,-0.1085140929,-0.0309973564,-0.213460058,-0.2505242229,-0.1225516498,-0.3344918489,-0.32494241,-0.2550111711,0.1577945501,-0.0034383023,-0.1047476232,0.3142209351,0.2670063674,0.0404987186,-0.165805921,0.0468914136,-0.0268509146,-0.2246979922,0.177915588,-0.201831609,-0.2568696737,0.3019668162,0.2744387388,0.1829818487,0.2670661211,-0.7614741325,-0.2670167983,-0.3597413003,0.2398702502,-0.1860815585,0.2179733366,0.1911988556,0.0769287571,0.0779739991,-0.2079752088,-0.4608823061,-0.0097571854,0.0752782673,0.4969078898,0.0363830738,0.3437930942,-0.0585201345,0.6186058521,0.397529155,0.0103019681,0.1783744097,0.2334723771,0.5895819068,0.0778009444,-0.4482364953,0.2397813201,-0.0084547177,0.1363245547,-0.1942522824,0.2197874635,0.2414545864,-0.277361989,0.2657617033,0.0042451476,0.0055124569,-0.2210909873,-0.13537018,0.3050734699,0.2228150517,0.0064602434,0.0430852436,-0.1302471906,0.0280230325,0.1745655835,0.010982167,0.3000519872,-0.1954792738,-0.1373809576,-0.183105126,0.2909622788,-0.0428895615,0.4082678556,-0.2521476746,0.4757547975,0.0345090851,0.0108251274,0.475733906,-0.3083002269,-0.1645100862,0.1024655625,0.1262150258,-0.6377895474,-0.1250775605,-0.1298962086,0.1647143066,0.248098284,0.2099779695,-0.4318394959,-0.1485227644,0.2467826158,0.5641607046,0.0371086225,-0.1298632622,-0.3288927078,-0.3882852793,-0.2767420411,-0.1371272653,-0.058102645,-0.1284619868,0.2001442164,-0.0588850789,0.1972738355,0.1628406644,0.1069483757,-0.0801087394,0.2060672939,-0.1776001751,0.2210780978,0.3728102148,-0.1271787584,0.1659846008,0.737244606,-0.2061559558,-0.3858459294,-0.0367117412,0.1224328429,0.1555299163,0.3913010359,-0.2864465714,-0.1263913363,0.2399162054,0.2055450827,0.1049313173,0.14691028,0.289960444,0.1443811804,-0.195566237,-0.4653308094,0.1165174395,0.4374177158,0.1289001107,0.4145717323,0.0137373563,0.0171636641,-0.0216146,0.2067595273,0.8317614794,-0.0005798346,0.0876606479,0.3389310539,-0.0641187727,0.5443438888,-0.4247684479,0.04840637,-0.0802192986,-0.143647477,-0.1364471316,-0.3701676726,0.3992536366,0.1148540974,-0.2142028362,0.3585865796,-0.1218098328,0.29375422,0.0482256673,0.1957940161,-0.0632860139,-0.2802743912,-0.1753268093,0.1845365763,-0.233880952,0.1757169068,-0.1705885977,0.083735086,-0.3188177347,-0.0867874846,-0.5056916475,0.0665702522,-0.443862617,0.2258304805,-0.0344874859,-0.2310214341,0.3094379306,0.1276277453,0.3448578715,0.4515912831,-0.3112368286,0.1110853925,-0.2714110315,-0.2467177361,0.3035934865,0.0439759605,0.0467805564,-0.0107217925,-0.1727659255,-0.0625354499,-0.2353877276,0.0169170164,0.0544819459,-0.0357779078,-0.1322258413,-0.1305123866,-0.1653528064,-0.1025140062,0.2076852918,-0.2839880288,0.0506421104,0.07781367,-0.0196289588,0.4598361552,-0.1750662029,-0.2117223293,-0.1011548117,0.212752223,-0.1656555086,-0.2982017696,0.234419629,0.2739279866,-0.3369333148,-0.1571041346,-0.094751589,-0.1557620764,-0.4031636715,0.1162526757,0.1045675799,-0.1815831959,-0.1145593971,0.2249464691,0.2387459278,-0.2340987474,0.0011506572,-0.4424839914,-0.2236991078,0.2046369612,-0.1303168684,0.0251450893,0.1048556641,0.0480054915,-0.0482010245,0.1132196635,-0.1702207923,-0.1554876715,0.0242763478,-0.0713960975,0.3412455618,-0.0761460811,0.0959786624,-0.2396223247,-0.03813833,0.1586219817,-0.2529083788,-0.076279223,-0.0355480611,0.137926355,0.0933924913,-0.3160541952,0.212218523,-0.4301118255,-0.1320781857,-0.1007387489,0.183366701,0.2039175779,0.0625451133,-0.2080432922,-0.4835402668,0.0814061463,0.0820645839,0.2181892693,0.0904808417,0.0420457087,0.2254475355,0.0915967301,0.1900683045,-0.09184926,-0.3334983885,0.0691129938,0.0046931463,0.2152244151,0.1156246215,-0.3056086898,0.2381537408,0.2715603411,0.4740514159,0.5606653094,-0.0505268052,0.2009599954,0.3780045509,0.0619235337,-0.3095085323,-0.1179374307,0.0978347436,0.211131677,-0.0669629127,0.294595927,0.2904630005,-0.1662479341,0.1414123923,0.2137935907,0.2448209524,0.0442248173,0.2401469797,0.1740344614,0.1456380934,-0.0191844087,0.2478772849,0.2596318722,0.1486105621,0.2685900927,-0.0940031186,-0.0129717235,0.1056639478,0.2145727575,-0.1893776506,-0.3154191971,-0.0778527632,-0.14081873,0.3885463476,0.0168743245,-0.0760412812,0.2765650153,-0.3946425617,-0.3519563675,-0.1730092168,0.2614917159,0.0503016636,0.2280521244,0.3215574324,0.0809561014,0.1780232787,0.0048405635,0.0397591293,-0.4191242456,0.2413220108,0.1764912158,-0.1236993223,-0.770821631,0.2125341445,0.1664997637,-0.1692295074,-0.0538866408,0.3296135664,0.1637779623,-0.1454275995,-0.2812110484,0.5068164468,0.5802744627,0.3409990072,-0.0469977558,0.1705626249,-0.2142198682,0.0777670294,-0.0155832991,0.2700774372,-0.1369200051,0.154964745,0.236733079,0.0225825254,-0.1113553271,-0.1749951243,0.05340923,0.4135027826,-0.1955527812,0.1946346015,-0.1500745118,-0.0271252468,-0.0973940268,-0.0908994824,-0.4811028838,0.2690776885,0.5110985637,0.0316910855,-0.0370604061,-0.1930583566,0.0272696707,-0.0880471319,0.3524516523,0.495470494,0.191963926,-0.4115162492,-0.0745142847,-0.6475214958,-0.2043262273,-0.3776461184,-0.2580498755,0.1304243505,-0.0521419086,0.0602565371,0.3123220801,0.2321897298,0.2287330478,0.1469683051,-0.0183709953,-0.3618494868,-0.1180011779,0.0987984166,0.1109670922,0.0805131122,-0.2171806246,0.0052382955,0.2312532216,-0.2043973804,-0.2440494448,-0.0672755912,0.0961156338,0.0278290417,0.3226872683,0.0853667632,0.3677543998,0.2596828043,-0.2449216247,-0.4981413782,-0.2963164449,-0.1827402264,0.3557329476,-0.046792753,0.2464546263,-0.5367340446,0.0496459119,-0.5474626422,0.097678706,0.0588760599,-0.0952013209,-0.2168268859,0.2665975988,-0.215844512,0.3078436255,0.0800647959,0.1987073421,-0.0550648682,0.2909039557,-0.4305544794,-0.322242856,0.3843740821,0.168199867,-0.3109259009,0.3755890727,0.0602345243,0.2313626111,-0.4108432531,-0.5497607589,0.1153397188,0.0295620337,-0.0591647848,-0.2287070155,0.1518161595,-0.1175127327,-0.0596970692,0.009724074,0.5295980573,0.0082971165,-0.254521817,0.3482492864,0.1984189898]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1696","title":"Unable to install datasets","comments":"Damn sorry","body":"** Edit **\r\nI believe there's a bug with the package when you're installing it with Python 3.9. I recommend sticking with previous versions. Thanks, @thomwolf for the insight! \r\n\r\n**Short description**\r\n\r\nI followed the instructions for installing datasets (https:\/\/huggingface.co\/docs\/datasets\/installation.html). However, while I tried to download datasets using `pip install datasets` I got a massive error message after getting stuck at \"Installing build dependencies...\" \r\n\r\nI was wondering if this problem can be fixed by creating a virtual environment, but it didn't help. Can anyone offer some advice on how to fix this issue? \r\n\r\nHere's an error message: \r\n\r\n`(env) Gas-MacBook-Pro:Downloads destiny$ pip install datasets\r\nCollecting datasets\r\n Using cached datasets-1.2.0-py3-none-any.whl (159 kB)\r\nCollecting numpy>=1.17\r\n Using cached numpy-1.19.5-cp39-cp39-macosx_10_9_x86_64.whl (15.6 MB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-2.0.0.tar.gz (58.9 MB)\r\n....\r\n\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceilf' [-Wincompatible-library-redeclaration]\r\n int ceilf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceilf' is a builtin with type 'float (float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintf' [-Wincompatible-library-redeclaration]\r\n int rintf (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintf' is a builtin with type 'float (float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncf' [-Wincompatible-library-redeclaration]\r\n int truncf (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncf' is a builtin with type 'float (float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtf' [-Wincompatible-library-redeclaration]\r\n int sqrtf (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtf' is a builtin with type 'float (float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10f' [-Wincompatible-library-redeclaration]\r\n int log10f (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10f' is a builtin with type 'float (float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logf' [-Wincompatible-library-redeclaration]\r\n int logf (void);\r\n ^\r\n _configtest.c:14:5: note: 'logf' is a builtin with type 'float (float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pf' [-Wincompatible-library-redeclaration]\r\n int log1pf (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pf' is a builtin with type 'float (float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expf' [-Wincompatible-library-redeclaration]\r\n int expf (void);\r\n ^\r\n _configtest.c:16:5: note: 'expf' is a builtin with type 'float (float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1f' [-Wincompatible-library-redeclaration]\r\n int expm1f (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1f' is a builtin with type 'float (float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinf' [-Wincompatible-library-redeclaration]\r\n int asinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinf' is a builtin with type 'float (float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosf' [-Wincompatible-library-redeclaration]\r\n int acosf (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosf' is a builtin with type 'float (float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanf' [-Wincompatible-library-redeclaration]\r\n int atanf (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanf' is a builtin with type 'float (float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhf' [-Wincompatible-library-redeclaration]\r\n int asinhf (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhf' is a builtin with type 'float (float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshf' [-Wincompatible-library-redeclaration]\r\n int acoshf (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshf' is a builtin with type 'float (float)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhf' [-Wincompatible-library-redeclaration]\r\n int atanhf (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhf' is a builtin with type 'float (float)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotf' [-Wincompatible-library-redeclaration]\r\n int hypotf (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotf' is a builtin with type 'float (float, float)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2f' [-Wincompatible-library-redeclaration]\r\n int atan2f (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2f' is a builtin with type 'float (float, float)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powf' [-Wincompatible-library-redeclaration]\r\n int powf (void);\r\n ^\r\n _configtest.c:26:5: note: 'powf' is a builtin with type 'float (float, float)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodf' [-Wincompatible-library-redeclaration]\r\n int fmodf (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodf' is a builtin with type 'float (float, float)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modff' [-Wincompatible-library-redeclaration]\r\n int modff (void);\r\n ^\r\n _configtest.c:28:5: note: 'modff' is a builtin with type 'float (float, float *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpf' [-Wincompatible-library-redeclaration]\r\n int frexpf (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpf' is a builtin with type 'float (float, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpf' [-Wincompatible-library-redeclaration]\r\n int ldexpf (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpf' is a builtin with type 'float (float, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2f' [-Wincompatible-library-redeclaration]\r\n int exp2f (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2f' is a builtin with type 'float (float)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2f' [-Wincompatible-library-redeclaration]\r\n int log2f (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2f' is a builtin with type 'float (float)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignf' [-Wincompatible-library-redeclaration]\r\n int copysignf (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignf' is a builtin with type 'float (float, float)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterf' [-Wincompatible-library-redeclaration]\r\n int nextafterf (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterf' is a builtin with type 'float (float, float)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtf' [-Wincompatible-library-redeclaration]\r\n int cbrtf (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtf' is a builtin with type 'float (float)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'sinl' [-Wincompatible-library-redeclaration]\r\n int sinl (void);\r\n ^\r\n _configtest.c:1:5: note: 'sinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cosl' [-Wincompatible-library-redeclaration]\r\n int cosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'tanl' [-Wincompatible-library-redeclaration]\r\n int tanl (void);\r\n ^\r\n _configtest.c:3:5: note: 'tanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'sinhl' [-Wincompatible-library-redeclaration]\r\n int sinhl (void);\r\n ^\r\n _configtest.c:4:5: note: 'sinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'coshl' [-Wincompatible-library-redeclaration]\r\n int coshl (void);\r\n ^\r\n _configtest.c:5:5: note: 'coshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'tanhl' [-Wincompatible-library-redeclaration]\r\n int tanhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'tanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'fabsl' [-Wincompatible-library-redeclaration]\r\n int fabsl (void);\r\n ^\r\n _configtest.c:7:5: note: 'fabsl' is a builtin with type 'long double (long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'floorl' [-Wincompatible-library-redeclaration]\r\n int floorl (void);\r\n ^\r\n _configtest.c:8:5: note: 'floorl' is a builtin with type 'long double (long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceill' [-Wincompatible-library-redeclaration]\r\n int ceill (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceill' is a builtin with type 'long double (long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintl' [-Wincompatible-library-redeclaration]\r\n int rintl (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintl' is a builtin with type 'long double (long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncl' [-Wincompatible-library-redeclaration]\r\n int truncl (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncl' is a builtin with type 'long double (long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtl' [-Wincompatible-library-redeclaration]\r\n int sqrtl (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtl' is a builtin with type 'long double (long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10l' [-Wincompatible-library-redeclaration]\r\n int log10l (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10l' is a builtin with type 'long double (long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logl' [-Wincompatible-library-redeclaration]\r\n int logl (void);\r\n ^\r\n _configtest.c:14:5: note: 'logl' is a builtin with type 'long double (long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pl' [-Wincompatible-library-redeclaration]\r\n int log1pl (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pl' is a builtin with type 'long double (long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expl' [-Wincompatible-library-redeclaration]\r\n int expl (void);\r\n ^\r\n _configtest.c:16:5: note: 'expl' is a builtin with type 'long double (long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1l' [-Wincompatible-library-redeclaration]\r\n int expm1l (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1l' is a builtin with type 'long double (long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinl' [-Wincompatible-library-redeclaration]\r\n int asinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosl' [-Wincompatible-library-redeclaration]\r\n int acosl (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanl' [-Wincompatible-library-redeclaration]\r\n int atanl (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhl' [-Wincompatible-library-redeclaration]\r\n int asinhl (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshl' [-Wincompatible-library-redeclaration]\r\n int acoshl (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhl' [-Wincompatible-library-redeclaration]\r\n int atanhl (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotl' [-Wincompatible-library-redeclaration]\r\n int hypotl (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2l' [-Wincompatible-library-redeclaration]\r\n int atan2l (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2l' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powl' [-Wincompatible-library-redeclaration]\r\n int powl (void);\r\n ^\r\n _configtest.c:26:5: note: 'powl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodl' [-Wincompatible-library-redeclaration]\r\n int fmodl (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modfl' [-Wincompatible-library-redeclaration]\r\n int modfl (void);\r\n ^\r\n _configtest.c:28:5: note: 'modfl' is a builtin with type 'long double (long double, long double *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpl' [-Wincompatible-library-redeclaration]\r\n int frexpl (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpl' is a builtin with type 'long double (long double, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpl' [-Wincompatible-library-redeclaration]\r\n int ldexpl (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpl' is a builtin with type 'long double (long double, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2l' [-Wincompatible-library-redeclaration]\r\n int exp2l (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2l' [-Wincompatible-library-redeclaration]\r\n int log2l (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignl' [-Wincompatible-library-redeclaration]\r\n int copysignl (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterl' [-Wincompatible-library-redeclaration]\r\n int nextafterl (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtl' [-Wincompatible-library-redeclaration]\r\n int cbrtl (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtl' is a builtin with type 'long double (long double)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:8:12: error: use of undeclared identifier 'HAVE_DECL_SIGNBIT'\r\n (void) HAVE_DECL_SIGNBIT;\r\n ^\r\n 1 error generated.\r\n failure.\r\n removing: _configtest.c _configtest.o\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabs' [-Wincompatible-library-redeclaration]\r\n int cabs (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabs' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacos' [-Wincompatible-library-redeclaration]\r\n int cacos (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacosh' [-Wincompatible-library-redeclaration]\r\n int cacosh (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'carg' [-Wincompatible-library-redeclaration]\r\n int carg (void);\r\n ^\r\n _configtest.c:4:5: note: 'carg' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casin' [-Wincompatible-library-redeclaration]\r\n int casin (void);\r\n ^\r\n _configtest.c:5:5: note: 'casin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinh' [-Wincompatible-library-redeclaration]\r\n int casinh (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catan' [-Wincompatible-library-redeclaration]\r\n int catan (void);\r\n ^\r\n _configtest.c:7:5: note: 'catan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanh' [-Wincompatible-library-redeclaration]\r\n int catanh (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccos' [-Wincompatible-library-redeclaration]\r\n int ccos (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccosh' [-Wincompatible-library-redeclaration]\r\n int ccosh (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexp' [-Wincompatible-library-redeclaration]\r\n int cexp (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexp' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimag' [-Wincompatible-library-redeclaration]\r\n int cimag (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimag' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clog' [-Wincompatible-library-redeclaration]\r\n int clog (void);\r\n ^\r\n _configtest.c:13:5: note: 'clog' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conj' [-Wincompatible-library-redeclaration]\r\n int conj (void);\r\n ^\r\n _configtest.c:14:5: note: 'conj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpow' [-Wincompatible-library-redeclaration]\r\n int cpow (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpow' is a builtin with type '_Complex double (_Complex double, _Complex double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cproj' [-Wincompatible-library-redeclaration]\r\n int cproj (void);\r\n ^\r\n _configtest.c:16:5: note: 'cproj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creal' [-Wincompatible-library-redeclaration]\r\n int creal (void);\r\n ^\r\n _configtest.c:17:5: note: 'creal' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csin' [-Wincompatible-library-redeclaration]\r\n int csin (void);\r\n ^\r\n _configtest.c:18:5: note: 'csin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinh' [-Wincompatible-library-redeclaration]\r\n int csinh (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrt' [-Wincompatible-library-redeclaration]\r\n int csqrt (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrt' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctan' [-Wincompatible-library-redeclaration]\r\n int ctan (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanh' [-Wincompatible-library-redeclaration]\r\n int ctanh (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanh' is a builtin with type '_Complex double (_Complex double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsf' [-Wincompatible-library-redeclaration]\r\n int cabsf (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosf' [-Wincompatible-library-redeclaration]\r\n int cacosf (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshf' [-Wincompatible-library-redeclaration]\r\n int cacoshf (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargf' [-Wincompatible-library-redeclaration]\r\n int cargf (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinf' [-Wincompatible-library-redeclaration]\r\n int casinf (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhf' [-Wincompatible-library-redeclaration]\r\n int casinhf (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanf' [-Wincompatible-library-redeclaration]\r\n int catanf (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhf' [-Wincompatible-library-redeclaration]\r\n int catanhf (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosf' [-Wincompatible-library-redeclaration]\r\n int ccosf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshf' [-Wincompatible-library-redeclaration]\r\n int ccoshf (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpf' [-Wincompatible-library-redeclaration]\r\n int cexpf (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagf' [-Wincompatible-library-redeclaration]\r\n int cimagf (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogf' [-Wincompatible-library-redeclaration]\r\n int clogf (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjf' [-Wincompatible-library-redeclaration]\r\n int conjf (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowf' [-Wincompatible-library-redeclaration]\r\n int cpowf (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowf' is a builtin with type '_Complex float (_Complex float, _Complex float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojf' [-Wincompatible-library-redeclaration]\r\n int cprojf (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'crealf' [-Wincompatible-library-redeclaration]\r\n int crealf (void);\r\n ^\r\n _configtest.c:17:5: note: 'crealf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinf' [-Wincompatible-library-redeclaration]\r\n int csinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhf' [-Wincompatible-library-redeclaration]\r\n int csinhf (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtf' [-Wincompatible-library-redeclaration]\r\n int csqrtf (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanf' [-Wincompatible-library-redeclaration]\r\n int ctanf (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhf' [-Wincompatible-library-redeclaration]\r\n int ctanhf (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhf' is a builtin with type '_Complex float (_Complex float)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsl' [-Wincompatible-library-redeclaration]\r\n int cabsl (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosl' [-Wincompatible-library-redeclaration]\r\n int cacosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshl' [-Wincompatible-library-redeclaration]\r\n int cacoshl (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargl' [-Wincompatible-library-redeclaration]\r\n int cargl (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinl' [-Wincompatible-library-redeclaration]\r\n int casinl (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhl' [-Wincompatible-library-redeclaration]\r\n int casinhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanl' [-Wincompatible-library-redeclaration]\r\n int catanl (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhl' [-Wincompatible-library-redeclaration]\r\n int catanhl (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosl' [-Wincompatible-library-redeclaration]\r\n int ccosl (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshl' [-Wincompatible-library-redeclaration]\r\n int ccoshl (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpl' [-Wincompatible-library-redeclaration]\r\n int cexpl (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagl' [-Wincompatible-library-redeclaration]\r\n int cimagl (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogl' [-Wincompatible-library-redeclaration]\r\n int clogl (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjl' [-Wincompatible-library-redeclaration]\r\n int conjl (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowl' [-Wincompatible-library-redeclaration]\r\n int cpowl (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowl' is a builtin with type '_Complex long double (_Complex long double, _Complex long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojl' [-Wincompatible-library-redeclaration]\r\n int cprojl (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creall' [-Wincompatible-library-redeclaration]\r\n int creall (void);\r\n ^\r\n _configtest.c:17:5: note: 'creall' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinl' [-Wincompatible-library-redeclaration]\r\n int csinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhl' [-Wincompatible-library-redeclaration]\r\n int csinhl (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtl' [-Wincompatible-library-redeclaration]\r\n int csqrtl (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanl' [-Wincompatible-library-redeclaration]\r\n int ctanl (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhl' [-Wincompatible-library-redeclaration]\r\n int ctanhl (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:2:12: warning: unused function 'static_func' [-Wunused-function]\r\n static int static_func (char * restrict a)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:3:19: warning: unused function 'static_func' [-Wunused-function]\r\n static inline int static_func (void)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h\r\n #define SIZEOF_PY_INTPTR_T 8\r\n #define SIZEOF_OFF_T 8\r\n #define SIZEOF_PY_LONG_LONG 8\r\n #define MATHLIB\r\n #define HAVE_SIN 1\r\n #define HAVE_COS 1\r\n #define HAVE_TAN 1\r\n #define HAVE_SINH 1\r\n #define HAVE_COSH 1\r\n #define HAVE_TANH 1\r\n #define HAVE_FABS 1\r\n #define HAVE_FLOOR 1\r\n #define HAVE_CEIL 1\r\n #define HAVE_SQRT 1\r\n #define HAVE_LOG10 1\r\n #define HAVE_LOG 1\r\n #define HAVE_EXP 1\r\n #define HAVE_ASIN 1\r\n #define HAVE_ACOS 1\r\n #define HAVE_ATAN 1\r\n #define HAVE_FMOD 1\r\n #define HAVE_MODF 1\r\n #define HAVE_FREXP 1\r\n #define HAVE_LDEXP 1\r\n #define HAVE_RINT 1\r\n #define HAVE_TRUNC 1\r\n #define HAVE_EXP2 1\r\n #define HAVE_LOG2 1\r\n #define HAVE_ATAN2 1\r\n #define HAVE_POW 1\r\n #define HAVE_NEXTAFTER 1\r\n #define HAVE_STRTOLL 1\r\n #define HAVE_STRTOULL 1\r\n #define HAVE_CBRT 1\r\n #define HAVE_STRTOLD_L 1\r\n #define HAVE_BACKTRACE 1\r\n #define HAVE_MADVISE 1\r\n #define HAVE_XMMINTRIN_H 1\r\n #define HAVE_EMMINTRIN_H 1\r\n #define HAVE_XLOCALE_H 1\r\n #define HAVE_DLFCN_H 1\r\n #define HAVE_SYS_MMAN_H 1\r\n #define HAVE___BUILTIN_ISNAN 1\r\n #define HAVE___BUILTIN_ISINF 1\r\n #define HAVE___BUILTIN_ISFINITE 1\r\n #define HAVE___BUILTIN_BSWAP32 1\r\n #define HAVE___BUILTIN_BSWAP64 1\r\n #define HAVE___BUILTIN_EXPECT 1\r\n #define HAVE___BUILTIN_MUL_OVERFLOW 1\r\n #define HAVE___BUILTIN_CPU_SUPPORTS 1\r\n #define HAVE__M_FROM_INT64 1\r\n #define HAVE__MM_LOAD_PS 1\r\n #define HAVE__MM_PREFETCH 1\r\n #define HAVE__MM_LOAD_PD 1\r\n #define HAVE___BUILTIN_PREFETCH 1\r\n #define HAVE_LINK_AVX 1\r\n #define HAVE_LINK_AVX2 1\r\n #define HAVE_XGETBV 1\r\n #define HAVE_ATTRIBUTE_NONNULL 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX2 1\r\n #define HAVE___THREAD 1\r\n #define HAVE_SINF 1\r\n #define HAVE_COSF 1\r\n #define HAVE_TANF 1\r\n #define HAVE_SINHF 1\r\n #define HAVE_COSHF 1\r\n #define HAVE_TANHF 1\r\n #define HAVE_FABSF 1\r\n #define HAVE_FLOORF 1\r\n #define HAVE_CEILF 1\r\n #define HAVE_RINTF 1\r\n #define HAVE_TRUNCF 1\r\n #define HAVE_SQRTF 1\r\n #define HAVE_LOG10F 1\r\n #define HAVE_LOGF 1\r\n #define HAVE_LOG1PF 1\r\n #define HAVE_EXPF 1\r\n #define HAVE_EXPM1F 1\r\n #define HAVE_ASINF 1\r\n #define HAVE_ACOSF 1\r\n #define HAVE_ATANF 1\r\n #define HAVE_ASINHF 1\r\n #define HAVE_ACOSHF 1\r\n #define HAVE_ATANHF 1\r\n #define HAVE_HYPOTF 1\r\n #define HAVE_ATAN2F 1\r\n #define HAVE_POWF 1\r\n #define HAVE_FMODF 1\r\n #define HAVE_MODFF 1\r\n #define HAVE_FREXPF 1\r\n #define HAVE_LDEXPF 1\r\n #define HAVE_EXP2F 1\r\n #define HAVE_LOG2F 1\r\n #define HAVE_COPYSIGNF 1\r\n #define HAVE_NEXTAFTERF 1\r\n #define HAVE_CBRTF 1\r\n #define HAVE_SINL 1\r\n #define HAVE_COSL 1\r\n #define HAVE_TANL 1\r\n #define HAVE_SINHL 1\r\n #define HAVE_COSHL 1\r\n #define HAVE_TANHL 1\r\n #define HAVE_FABSL 1\r\n #define HAVE_FLOORL 1\r\n #define HAVE_CEILL 1\r\n #define HAVE_RINTL 1\r\n #define HAVE_TRUNCL 1\r\n #define HAVE_SQRTL 1\r\n #define HAVE_LOG10L 1\r\n #define HAVE_LOGL 1\r\n #define HAVE_LOG1PL 1\r\n #define HAVE_EXPL 1\r\n #define HAVE_EXPM1L 1\r\n #define HAVE_ASINL 1\r\n #define HAVE_ACOSL 1\r\n #define HAVE_ATANL 1\r\n #define HAVE_ASINHL 1\r\n #define HAVE_ACOSHL 1\r\n #define HAVE_ATANHL 1\r\n #define HAVE_HYPOTL 1\r\n #define HAVE_ATAN2L 1\r\n #define HAVE_POWL 1\r\n #define HAVE_FMODL 1\r\n #define HAVE_MODFL 1\r\n #define HAVE_FREXPL 1\r\n #define HAVE_LDEXPL 1\r\n #define HAVE_EXP2L 1\r\n #define HAVE_LOG2L 1\r\n #define HAVE_COPYSIGNL 1\r\n #define HAVE_NEXTAFTERL 1\r\n #define HAVE_CBRTL 1\r\n #define HAVE_DECL_SIGNBIT\r\n #define HAVE_COMPLEX_H 1\r\n #define HAVE_CABS 1\r\n #define HAVE_CACOS 1\r\n #define HAVE_CACOSH 1\r\n #define HAVE_CARG 1\r\n #define HAVE_CASIN 1\r\n #define HAVE_CASINH 1\r\n #define HAVE_CATAN 1\r\n #define HAVE_CATANH 1\r\n #define HAVE_CCOS 1\r\n #define HAVE_CCOSH 1\r\n #define HAVE_CEXP 1\r\n #define HAVE_CIMAG 1\r\n #define HAVE_CLOG 1\r\n #define HAVE_CONJ 1\r\n #define HAVE_CPOW 1\r\n #define HAVE_CPROJ 1\r\n #define HAVE_CREAL 1\r\n #define HAVE_CSIN 1\r\n #define HAVE_CSINH 1\r\n #define HAVE_CSQRT 1\r\n #define HAVE_CTAN 1\r\n #define HAVE_CTANH 1\r\n #define HAVE_CABSF 1\r\n #define HAVE_CACOSF 1\r\n #define HAVE_CACOSHF 1\r\n #define HAVE_CARGF 1\r\n #define HAVE_CASINF 1\r\n #define HAVE_CASINHF 1\r\n #define HAVE_CATANF 1\r\n #define HAVE_CATANHF 1\r\n #define HAVE_CCOSF 1\r\n #define HAVE_CCOSHF 1\r\n #define HAVE_CEXPF 1\r\n #define HAVE_CIMAGF 1\r\n #define HAVE_CLOGF 1\r\n #define HAVE_CONJF 1\r\n #define HAVE_CPOWF 1\r\n #define HAVE_CPROJF 1\r\n #define HAVE_CREALF 1\r\n #define HAVE_CSINF 1\r\n #define HAVE_CSINHF 1\r\n #define HAVE_CSQRTF 1\r\n #define HAVE_CTANF 1\r\n #define HAVE_CTANHF 1\r\n #define HAVE_CABSL 1\r\n #define HAVE_CACOSL 1\r\n #define HAVE_CACOSHL 1\r\n #define HAVE_CARGL 1\r\n #define HAVE_CASINL 1\r\n #define HAVE_CASINHL 1\r\n #define HAVE_CATANL 1\r\n #define HAVE_CATANHL 1\r\n #define HAVE_CCOSL 1\r\n #define HAVE_CCOSHL 1\r\n #define HAVE_CEXPL 1\r\n #define HAVE_CIMAGL 1\r\n #define HAVE_CLOGL 1\r\n #define HAVE_CONJL 1\r\n #define HAVE_CPOWL 1\r\n #define HAVE_CPROJL 1\r\n #define HAVE_CREALL 1\r\n #define HAVE_CSINL 1\r\n #define HAVE_CSINHL 1\r\n #define HAVE_CSQRTL 1\r\n #define HAVE_CTANL 1\r\n #define HAVE_CTANHL 1\r\n #define NPY_RESTRICT restrict\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define HAVE_LDOUBLE_INTEL_EXTENDED_16_BYTES_LE 1\r\n #define NPY_PY3K 1\r\n #ifndef __cplusplus\r\n \/* #undef inline *\/\r\n #endif\r\n \r\n #ifndef _NPY_NPY_CONFIG_H_\r\n #error config.h should never be included directly, include npy_config.h instead\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n Generating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'exp' [-Wincompatible-library-redeclaration]\r\n int exp (void);\r\n ^\r\n _configtest.c:1:5: note: 'exp' is a builtin with type 'double (double)'\r\n 1 warning generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n #define NPY_SIZEOF_SHORT SIZEOF_SHORT\r\n #define NPY_SIZEOF_INT SIZEOF_INT\r\n #define NPY_SIZEOF_LONG SIZEOF_LONG\r\n #define NPY_SIZEOF_FLOAT 4\r\n #define NPY_SIZEOF_COMPLEX_FLOAT 8\r\n #define NPY_SIZEOF_DOUBLE 8\r\n #define NPY_SIZEOF_COMPLEX_DOUBLE 16\r\n #define NPY_SIZEOF_LONGDOUBLE 16\r\n #define NPY_SIZEOF_COMPLEX_LONGDOUBLE 32\r\n #define NPY_SIZEOF_PY_INTPTR_T 8\r\n #define NPY_SIZEOF_OFF_T 8\r\n #define NPY_SIZEOF_PY_LONG_LONG 8\r\n #define NPY_SIZEOF_LONGLONG 8\r\n #define NPY_NO_SMP 0\r\n #define NPY_HAVE_DECL_ISNAN\r\n #define NPY_HAVE_DECL_ISINF\r\n #define NPY_HAVE_DECL_ISFINITE\r\n #define NPY_HAVE_DECL_SIGNBIT\r\n #define NPY_USE_C99_COMPLEX 1\r\n #define NPY_HAVE_COMPLEX_DOUBLE 1\r\n #define NPY_HAVE_COMPLEX_FLOAT 1\r\n #define NPY_HAVE_COMPLEX_LONG_DOUBLE 1\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define NPY_USE_C99_FORMATS 1\r\n #define NPY_VISIBILITY_HIDDEN __attribute__((visibility(\"hidden\")))\r\n #define NPY_ABI_VERSION 0x01000009\r\n #define NPY_API_VERSION 0x0000000D\r\n \r\n #ifndef __STDC_FORMAT_MACROS\r\n #define __STDC_FORMAT_MACROS 1\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h']\r\n building extension \"numpy.core._multiarray_tests\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n building extension \"numpy.core._multiarray_umath\" sources\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_ufunc_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common' to include_dirs.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_internal.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h']\r\n building extension \"numpy.core._umath_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n building extension \"numpy.core._rational_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n building extension \"numpy.core._struct_ufunc_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n building extension \"numpy.core._operand_flag_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n building extension \"numpy.fft.fftpack_lite\" sources\r\n building extension \"numpy.linalg.lapack_lite\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n building extension \"numpy.linalg._umath_linalg\" sources\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n building extension \"numpy.random.mtrand\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/random\r\n building data_files sources\r\n build_src: building npy-pkg config files\r\n running build_py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/conftest.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_globals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/dual.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_distributor_init.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/ctypeslib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/matlib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_pytesttester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/py3k.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/_inspect.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/fromnumeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_add_newdocs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_methods.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_internal.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_string_helpers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/multiarray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/records.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup_common.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_aliased_types.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/memmap.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/overrides.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/getlimits.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype_ctypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/defchararray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/machar.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/einsumfunc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath_tests.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numerictypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_type_aliases.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/cversions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/arrayprint.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/code_generators\/generate_numpy_api.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/unixccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/numpy_distribution.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/conv_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/cpuinfo.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvc9compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/npy_pkg_config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/compat.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/misc_util.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/log.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/line_endings.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/lib2def.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/pathccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/system_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/exec_command.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/from_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/mingw32ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/extension.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/intelccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/distutils\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config_compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_ext.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_headers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_py.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_src.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/sdist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_scripts.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/bdist_rpm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/autodist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/egg_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/develop.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_data.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/gnu.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/compaq.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/intel.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/none.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/nag.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/ibm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/sun.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/lahey.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/g95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/mips.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/hpux.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/environment.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pathf95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/absoft.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/vast.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/misc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/internals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/creation.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/constants.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/ufuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/broadcasting.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/basics.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/subclassing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/indexing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/byteswapping.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/structured_arrays.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/glossary.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/common_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/crackfortran.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cb_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py2e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/func2subr.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/diagnose.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/capi_maps.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f90mod_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py_testing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/use_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/auxfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__main__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/helper.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/fftpack.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_iotools.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/mixins.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/nanfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/recfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/histograms.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/scimath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/user_array.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/format.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/twodim_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/financial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/index_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/npyio.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/stride_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arrayterator.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraysetops.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraypad.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/type_check.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_datasource.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/ufunclike.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/linalg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/extras.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/testutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/bench.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/timer_comparison.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/mrecords.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/defmatrix.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/laguerre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/_polybase.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polyutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite_e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/chebyshev.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/legendre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/print_coercion_tables.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/parameterized.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n running build_clib\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_clib\r\n building 'npymath' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n ar: adding 4 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n building 'npysort' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/quicksort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/mergesort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/heapsort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/selection.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/binsearch.c\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n 22 warnings generated.\r\n ar: adding 5 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n running build_ext\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_ext\r\n building 'numpy.core._dummy' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/dummymodule.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/dummymodule.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_dummy.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_tests.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_umath' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/core\/src\/multiarray\/alloc.c\r\n clang: numpy\/core\/src\/multiarray\/calculation.cclang: numpy\/core\/src\/multiarray\/array_assign_scalar.c\r\n clang: numpy\/core\/src\/multiarray\/convert.c\r\n \r\n clang: numpy\/core\/src\/multiarray\/ctors.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busday.c\r\n clang: numpy\/core\/src\/multiarray\/dragon4.cclang: numpy\/core\/src\/multiarray\/flagsobject.c\r\n \r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/arrayobject.c\r\n clang: numpy\/core\/src\/multiarray\/array_assign_array.c\r\n clang: numpy\/core\/src\/multiarray\/convert_datatype.c\r\n clang: numpy\/core\/src\/multiarray\/getset.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busdaycal.c\r\n clang: numpy\/core\/src\/multiarray\/buffer.c\r\n clang: numpy\/core\/src\/multiarray\/compiled_base.c\r\n clang: numpy\/core\/src\/multiarray\/hashdescr.c\r\n clang: numpy\/core\/src\/multiarray\/descriptor.c\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/conversion_utils.c\r\n clang: numpy\/core\/src\/multiarray\/item_selection.c\r\n clang: numpy\/core\/src\/multiarray\/dtype_transfer.c\r\n clang: numpy\/core\/src\/multiarray\/mapping.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n 3 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/datetime.c\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:477:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ptr = PyUnicode_AS_UNICODE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/common.c\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 6 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/nditer_pywrap.c\r\n 9 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/sequence.c\r\n clang: numpy\/core\/src\/multiarray\/shape.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n clang: numpy\/core\/src\/multiarray\/methods.c\r\n clang: numpy\/core\/src\/multiarray\/iterators.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_strings.c\r\n clang: numpy\/core\/src\/multiarray\/number.c\r\n clang: numpy\/core\/src\/multiarray\/scalarapi.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n numpy\/core\/src\/multiarray\/scalarapi.c:74:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:135:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1849:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buffer = PyUnicode_AS_DATA(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 5 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/typeinfo.c\r\n clang: numpy\/core\/src\/multiarray\/refcount.c\r\n clang: numpy\/core\/src\/multiarray\/usertypes.c\r\n clang: numpy\/core\/src\/multiarray\/multiarraymodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n clang: numpy\/core\/src\/multiarray\/vdot.c\r\n clang: numpy\/core\/src\/umath\/umathmodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n clang: numpy\/core\/src\/umath\/reduction.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n clang: numpy\/core\/src\/multiarray\/nditer_api.c\r\n 14 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/strfuncs.c\r\n numpy\/core\/src\/umath\/loops.c.src:655:18: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n result = PyEval_CallObject(tocall, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:178:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_ReprFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:195:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_StrFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 2 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/temp_elide.c\r\n clang: numpy\/core\/src\/umath\/cpuid.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n clang: numpy\/core\/src\/umath\/ufunc_object.c\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'byte_long' [-Wunused-function]\r\n byte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ubyte_long' [-Wunused-function]\r\n ubyte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'short_long' [-Wunused-function]\r\n short_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ushort_long' [-Wunused-function]\r\n ushort_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'int_long' [-Wunused-function]\r\n int_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'uint_long' [-Wunused-function]\r\n uint_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'long_long' [-Wunused-function]\r\n long_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulong_long' [-Wunused-function]\r\n ulong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longlong_long' [-Wunused-function]\r\n longlong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulonglong_long' [-Wunused-function]\r\n ulonglong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'half_long' [-Wunused-function]\r\n half_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'float_long' [-Wunused-function]\r\n float_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'double_long' [-Wunused-function]\r\n double_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longdouble_long' [-Wunused-function]\r\n longdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cfloat_long' [-Wunused-function]\r\n cfloat_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cdouble_long' [-Wunused-function]\r\n cdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'clongdouble_long' [-Wunused-function]\r\n clongdouble_long(PyObject *obj)\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/nditer_constr.c\r\n numpy\/core\/src\/umath\/ufunc_object.c:657:19: warning: comparison of integers of different signs: 'int' and 'size_t' (aka 'unsigned long') [-Wsign-compare]\r\n for (i = 0; i < len; i++) {\r\n ~ ^ ~~~\r\n clang: numpy\/core\/src\/umath\/override.c\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n clang: numpy\/core\/src\/common\/array_assign.c\r\n clang: numpy\/core\/src\/common\/ufunc_override.c\r\n clang: numpy\/core\/src\/common\/npy_longdouble.c\r\n clang: numpy\/core\/src\/common\/numpyos.c\r\n clang: numpy\/core\/src\/common\/ucsnarrow.c\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/umath\/extobj.c\r\n numpy\/core\/src\/common\/ucsnarrow.c:139:34: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n ret = (PyUnicodeObject *)PyUnicode_FromUnicode((Py_UNICODE*)buf,\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/common\/python_xerbla.c\r\n clang: numpy\/core\/src\/common\/cblasfuncs.c\r\n clang: \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c\r\n In file included from \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c:26:\r\n In file included from numpy\/core\/include\/numpy\/arrayobject.h:4:\r\n In file included from numpy\/core\/include\/numpy\/ndarrayobject.h:21:\r\n build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h:1463:1: warning: unused function '_import_array' [-Wunused-function]\r\n _import_array(void)\r\n ^\r\n 1 warning generated.\r\n 17 warnings generated.\r\n clang: numpy\/core\/src\/umath\/ufunc_type_resolution.c\r\n 4 warnings generated.\r\n 4 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/alloc.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arrayobject.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_scalar.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_array.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/buffer.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/calculation.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/compiled_base.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/common.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert_datatype.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/conversion_utils.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/ctors.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_strings.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busday.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busdaycal.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/descriptor.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dragon4.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dtype_transfer.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/flagsobject.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/getset.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/hashdescr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/item_selection.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/iterators.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/mapping.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/methods.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/multiarraymodule.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_api.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_constr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_pywrap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/number.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/refcount.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/sequence.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/shape.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalarapi.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/strfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/temp_elide.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/typeinfo.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/usertypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/vdot.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/umathmodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/reduction.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_object.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/extobj.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/cpuid.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_type_resolution.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/halffloat.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/array_assign.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/npy_longdouble.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ucsnarrow.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ufunc_override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/numpyos.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/cblasfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/python_xerbla.o build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -lnpysort -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_umath.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.core._umath_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_umath_tests.cpython-39-darwin.so\r\n building 'numpy.core._rational_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_rational_tests.cpython-39-darwin.so\r\n building 'numpy.core._struct_ufunc_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_struct_ufunc_tests.cpython-39-darwin.so\r\n building 'numpy.core._operand_flag_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_operand_flag_tests.cpython-39-darwin.so\r\n building 'numpy.fft.fftpack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n compile options: '-Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/fft\/fftpack_litemodule.c\r\n clang: numpy\/fft\/fftpack.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_lite.cpython-39-darwin.so\r\n building 'numpy.linalg.lapack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/linalg\/lapack_litemodule.c\r\n clang: numpy\/linalg\/lapack_lite\/python_xerbla.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.linalg._umath_linalg' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n numpy\/linalg\/umath_linalg.c.src:735:32: warning: unknown warning group '-Wmaybe-uninitialized', ignored [-Wunknown-warning-option]\r\n #pragma GCC diagnostic ignored \"-Wmaybe-uninitialized\"\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:541:1: warning: unused function 'dump_ufunc_object' [-Wunused-function]\r\n dump_ufunc_object(PyUFuncObject* ufunc)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:566:1: warning: unused function 'dump_linearize_data' [-Wunused-function]\r\n dump_linearize_data(const char* name, const LINEARIZE_DATA_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_FLOAT_matrix' [-Wunused-function]\r\n dump_FLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_DOUBLE_matrix' [-Wunused-function]\r\n dump_DOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CFLOAT_matrix' [-Wunused-function]\r\n dump_CFLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CDOUBLE_matrix' [-Wunused-function]\r\n dump_CDOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_FLOAT_matrix' [-Wunused-function]\r\n zero_FLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_DOUBLE_matrix' [-Wunused-function]\r\n zero_DOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CFLOAT_matrix' [-Wunused-function]\r\n zero_CFLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CDOUBLE_matrix' [-Wunused-function]\r\n zero_CDOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:1862:1: warning: unused function 'dump_geev_params' [-Wunused-function]\r\n dump_geev_params(const char *name, GEEV_PARAMS_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2132:1: warning: unused function 'init_cgeev' [-Wunused-function]\r\n init_cgeev(GEEV_PARAMS_t* params,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2213:1: warning: unused function 'process_cgeev_results' [-Wunused-function]\r\n process_cgeev_results(GEEV_PARAMS_t *NPY_UNUSED(params))\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2376:1: warning: unused function 'dump_gesdd_params' [-Wunused-function]\r\n dump_gesdd_params(const char *name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2864:1: warning: unused function 'dump_gelsd_params' [-Wunused-function]\r\n dump_gelsd_params(const char *name,\r\n ^\r\n 16 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/_umath_linalg.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.random.mtrand' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\r\n compile options: '-D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/random\/mtrand\/mtrand.c\r\n clang: numpy\/random\/mtrand\/initarray.cclang: numpy\/random\/mtrand\/randomkit.c\r\n \r\n clang: numpy\/random\/mtrand\/distributions.c\r\n numpy\/random\/mtrand\/mtrand.c:40400:34: error: no member named 'tp_print' in 'struct _typeobject'\r\n __pyx_type_6mtrand_RandomState.tp_print = 0;\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 12 warnings and 1 error generated.\r\n error: Command \"clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c numpy\/random\/mtrand\/mtrand.c -o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o -MMD -MF build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o.d\" failed with exit status 1","comment_length":2,"text":"Unable to install datasets\n** Edit **\r\nI believe there's a bug with the package when you're installing it with Python 3.9. I recommend sticking with previous versions. Thanks, @thomwolf for the insight! \r\n\r\n**Short description**\r\n\r\nI followed the instructions for installing datasets (https:\/\/huggingface.co\/docs\/datasets\/installation.html). However, while I tried to download datasets using `pip install datasets` I got a massive error message after getting stuck at \"Installing build dependencies...\" \r\n\r\nI was wondering if this problem can be fixed by creating a virtual environment, but it didn't help. Can anyone offer some advice on how to fix this issue? \r\n\r\nHere's an error message: \r\n\r\n`(env) Gas-MacBook-Pro:Downloads destiny$ pip install datasets\r\nCollecting datasets\r\n Using cached datasets-1.2.0-py3-none-any.whl (159 kB)\r\nCollecting numpy>=1.17\r\n Using cached numpy-1.19.5-cp39-cp39-macosx_10_9_x86_64.whl (15.6 MB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-2.0.0.tar.gz (58.9 MB)\r\n....\r\n\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceilf' [-Wincompatible-library-redeclaration]\r\n int ceilf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceilf' is a builtin with type 'float (float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintf' [-Wincompatible-library-redeclaration]\r\n int rintf (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintf' is a builtin with type 'float (float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncf' [-Wincompatible-library-redeclaration]\r\n int truncf (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncf' is a builtin with type 'float (float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtf' [-Wincompatible-library-redeclaration]\r\n int sqrtf (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtf' is a builtin with type 'float (float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10f' [-Wincompatible-library-redeclaration]\r\n int log10f (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10f' is a builtin with type 'float (float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logf' [-Wincompatible-library-redeclaration]\r\n int logf (void);\r\n ^\r\n _configtest.c:14:5: note: 'logf' is a builtin with type 'float (float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pf' [-Wincompatible-library-redeclaration]\r\n int log1pf (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pf' is a builtin with type 'float (float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expf' [-Wincompatible-library-redeclaration]\r\n int expf (void);\r\n ^\r\n _configtest.c:16:5: note: 'expf' is a builtin with type 'float (float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1f' [-Wincompatible-library-redeclaration]\r\n int expm1f (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1f' is a builtin with type 'float (float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinf' [-Wincompatible-library-redeclaration]\r\n int asinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinf' is a builtin with type 'float (float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosf' [-Wincompatible-library-redeclaration]\r\n int acosf (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosf' is a builtin with type 'float (float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanf' [-Wincompatible-library-redeclaration]\r\n int atanf (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanf' is a builtin with type 'float (float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhf' [-Wincompatible-library-redeclaration]\r\n int asinhf (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhf' is a builtin with type 'float (float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshf' [-Wincompatible-library-redeclaration]\r\n int acoshf (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshf' is a builtin with type 'float (float)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhf' [-Wincompatible-library-redeclaration]\r\n int atanhf (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhf' is a builtin with type 'float (float)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotf' [-Wincompatible-library-redeclaration]\r\n int hypotf (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotf' is a builtin with type 'float (float, float)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2f' [-Wincompatible-library-redeclaration]\r\n int atan2f (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2f' is a builtin with type 'float (float, float)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powf' [-Wincompatible-library-redeclaration]\r\n int powf (void);\r\n ^\r\n _configtest.c:26:5: note: 'powf' is a builtin with type 'float (float, float)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodf' [-Wincompatible-library-redeclaration]\r\n int fmodf (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodf' is a builtin with type 'float (float, float)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modff' [-Wincompatible-library-redeclaration]\r\n int modff (void);\r\n ^\r\n _configtest.c:28:5: note: 'modff' is a builtin with type 'float (float, float *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpf' [-Wincompatible-library-redeclaration]\r\n int frexpf (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpf' is a builtin with type 'float (float, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpf' [-Wincompatible-library-redeclaration]\r\n int ldexpf (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpf' is a builtin with type 'float (float, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2f' [-Wincompatible-library-redeclaration]\r\n int exp2f (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2f' is a builtin with type 'float (float)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2f' [-Wincompatible-library-redeclaration]\r\n int log2f (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2f' is a builtin with type 'float (float)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignf' [-Wincompatible-library-redeclaration]\r\n int copysignf (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignf' is a builtin with type 'float (float, float)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterf' [-Wincompatible-library-redeclaration]\r\n int nextafterf (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterf' is a builtin with type 'float (float, float)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtf' [-Wincompatible-library-redeclaration]\r\n int cbrtf (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtf' is a builtin with type 'float (float)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'sinl' [-Wincompatible-library-redeclaration]\r\n int sinl (void);\r\n ^\r\n _configtest.c:1:5: note: 'sinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cosl' [-Wincompatible-library-redeclaration]\r\n int cosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'tanl' [-Wincompatible-library-redeclaration]\r\n int tanl (void);\r\n ^\r\n _configtest.c:3:5: note: 'tanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'sinhl' [-Wincompatible-library-redeclaration]\r\n int sinhl (void);\r\n ^\r\n _configtest.c:4:5: note: 'sinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'coshl' [-Wincompatible-library-redeclaration]\r\n int coshl (void);\r\n ^\r\n _configtest.c:5:5: note: 'coshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'tanhl' [-Wincompatible-library-redeclaration]\r\n int tanhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'tanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'fabsl' [-Wincompatible-library-redeclaration]\r\n int fabsl (void);\r\n ^\r\n _configtest.c:7:5: note: 'fabsl' is a builtin with type 'long double (long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'floorl' [-Wincompatible-library-redeclaration]\r\n int floorl (void);\r\n ^\r\n _configtest.c:8:5: note: 'floorl' is a builtin with type 'long double (long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceill' [-Wincompatible-library-redeclaration]\r\n int ceill (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceill' is a builtin with type 'long double (long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintl' [-Wincompatible-library-redeclaration]\r\n int rintl (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintl' is a builtin with type 'long double (long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncl' [-Wincompatible-library-redeclaration]\r\n int truncl (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncl' is a builtin with type 'long double (long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtl' [-Wincompatible-library-redeclaration]\r\n int sqrtl (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtl' is a builtin with type 'long double (long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10l' [-Wincompatible-library-redeclaration]\r\n int log10l (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10l' is a builtin with type 'long double (long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logl' [-Wincompatible-library-redeclaration]\r\n int logl (void);\r\n ^\r\n _configtest.c:14:5: note: 'logl' is a builtin with type 'long double (long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pl' [-Wincompatible-library-redeclaration]\r\n int log1pl (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pl' is a builtin with type 'long double (long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expl' [-Wincompatible-library-redeclaration]\r\n int expl (void);\r\n ^\r\n _configtest.c:16:5: note: 'expl' is a builtin with type 'long double (long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1l' [-Wincompatible-library-redeclaration]\r\n int expm1l (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1l' is a builtin with type 'long double (long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinl' [-Wincompatible-library-redeclaration]\r\n int asinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosl' [-Wincompatible-library-redeclaration]\r\n int acosl (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanl' [-Wincompatible-library-redeclaration]\r\n int atanl (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhl' [-Wincompatible-library-redeclaration]\r\n int asinhl (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshl' [-Wincompatible-library-redeclaration]\r\n int acoshl (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhl' [-Wincompatible-library-redeclaration]\r\n int atanhl (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotl' [-Wincompatible-library-redeclaration]\r\n int hypotl (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2l' [-Wincompatible-library-redeclaration]\r\n int atan2l (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2l' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powl' [-Wincompatible-library-redeclaration]\r\n int powl (void);\r\n ^\r\n _configtest.c:26:5: note: 'powl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodl' [-Wincompatible-library-redeclaration]\r\n int fmodl (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modfl' [-Wincompatible-library-redeclaration]\r\n int modfl (void);\r\n ^\r\n _configtest.c:28:5: note: 'modfl' is a builtin with type 'long double (long double, long double *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpl' [-Wincompatible-library-redeclaration]\r\n int frexpl (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpl' is a builtin with type 'long double (long double, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpl' [-Wincompatible-library-redeclaration]\r\n int ldexpl (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpl' is a builtin with type 'long double (long double, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2l' [-Wincompatible-library-redeclaration]\r\n int exp2l (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2l' [-Wincompatible-library-redeclaration]\r\n int log2l (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignl' [-Wincompatible-library-redeclaration]\r\n int copysignl (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterl' [-Wincompatible-library-redeclaration]\r\n int nextafterl (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtl' [-Wincompatible-library-redeclaration]\r\n int cbrtl (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtl' is a builtin with type 'long double (long double)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:8:12: error: use of undeclared identifier 'HAVE_DECL_SIGNBIT'\r\n (void) HAVE_DECL_SIGNBIT;\r\n ^\r\n 1 error generated.\r\n failure.\r\n removing: _configtest.c _configtest.o\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabs' [-Wincompatible-library-redeclaration]\r\n int cabs (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabs' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacos' [-Wincompatible-library-redeclaration]\r\n int cacos (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacosh' [-Wincompatible-library-redeclaration]\r\n int cacosh (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'carg' [-Wincompatible-library-redeclaration]\r\n int carg (void);\r\n ^\r\n _configtest.c:4:5: note: 'carg' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casin' [-Wincompatible-library-redeclaration]\r\n int casin (void);\r\n ^\r\n _configtest.c:5:5: note: 'casin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinh' [-Wincompatible-library-redeclaration]\r\n int casinh (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catan' [-Wincompatible-library-redeclaration]\r\n int catan (void);\r\n ^\r\n _configtest.c:7:5: note: 'catan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanh' [-Wincompatible-library-redeclaration]\r\n int catanh (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccos' [-Wincompatible-library-redeclaration]\r\n int ccos (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccosh' [-Wincompatible-library-redeclaration]\r\n int ccosh (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexp' [-Wincompatible-library-redeclaration]\r\n int cexp (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexp' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimag' [-Wincompatible-library-redeclaration]\r\n int cimag (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimag' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clog' [-Wincompatible-library-redeclaration]\r\n int clog (void);\r\n ^\r\n _configtest.c:13:5: note: 'clog' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conj' [-Wincompatible-library-redeclaration]\r\n int conj (void);\r\n ^\r\n _configtest.c:14:5: note: 'conj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpow' [-Wincompatible-library-redeclaration]\r\n int cpow (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpow' is a builtin with type '_Complex double (_Complex double, _Complex double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cproj' [-Wincompatible-library-redeclaration]\r\n int cproj (void);\r\n ^\r\n _configtest.c:16:5: note: 'cproj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creal' [-Wincompatible-library-redeclaration]\r\n int creal (void);\r\n ^\r\n _configtest.c:17:5: note: 'creal' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csin' [-Wincompatible-library-redeclaration]\r\n int csin (void);\r\n ^\r\n _configtest.c:18:5: note: 'csin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinh' [-Wincompatible-library-redeclaration]\r\n int csinh (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrt' [-Wincompatible-library-redeclaration]\r\n int csqrt (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrt' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctan' [-Wincompatible-library-redeclaration]\r\n int ctan (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanh' [-Wincompatible-library-redeclaration]\r\n int ctanh (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanh' is a builtin with type '_Complex double (_Complex double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsf' [-Wincompatible-library-redeclaration]\r\n int cabsf (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosf' [-Wincompatible-library-redeclaration]\r\n int cacosf (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshf' [-Wincompatible-library-redeclaration]\r\n int cacoshf (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargf' [-Wincompatible-library-redeclaration]\r\n int cargf (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinf' [-Wincompatible-library-redeclaration]\r\n int casinf (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhf' [-Wincompatible-library-redeclaration]\r\n int casinhf (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanf' [-Wincompatible-library-redeclaration]\r\n int catanf (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhf' [-Wincompatible-library-redeclaration]\r\n int catanhf (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosf' [-Wincompatible-library-redeclaration]\r\n int ccosf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshf' [-Wincompatible-library-redeclaration]\r\n int ccoshf (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpf' [-Wincompatible-library-redeclaration]\r\n int cexpf (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagf' [-Wincompatible-library-redeclaration]\r\n int cimagf (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogf' [-Wincompatible-library-redeclaration]\r\n int clogf (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjf' [-Wincompatible-library-redeclaration]\r\n int conjf (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowf' [-Wincompatible-library-redeclaration]\r\n int cpowf (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowf' is a builtin with type '_Complex float (_Complex float, _Complex float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojf' [-Wincompatible-library-redeclaration]\r\n int cprojf (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'crealf' [-Wincompatible-library-redeclaration]\r\n int crealf (void);\r\n ^\r\n _configtest.c:17:5: note: 'crealf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinf' [-Wincompatible-library-redeclaration]\r\n int csinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhf' [-Wincompatible-library-redeclaration]\r\n int csinhf (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtf' [-Wincompatible-library-redeclaration]\r\n int csqrtf (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanf' [-Wincompatible-library-redeclaration]\r\n int ctanf (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhf' [-Wincompatible-library-redeclaration]\r\n int ctanhf (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhf' is a builtin with type '_Complex float (_Complex float)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsl' [-Wincompatible-library-redeclaration]\r\n int cabsl (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosl' [-Wincompatible-library-redeclaration]\r\n int cacosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshl' [-Wincompatible-library-redeclaration]\r\n int cacoshl (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargl' [-Wincompatible-library-redeclaration]\r\n int cargl (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinl' [-Wincompatible-library-redeclaration]\r\n int casinl (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhl' [-Wincompatible-library-redeclaration]\r\n int casinhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanl' [-Wincompatible-library-redeclaration]\r\n int catanl (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhl' [-Wincompatible-library-redeclaration]\r\n int catanhl (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosl' [-Wincompatible-library-redeclaration]\r\n int ccosl (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshl' [-Wincompatible-library-redeclaration]\r\n int ccoshl (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpl' [-Wincompatible-library-redeclaration]\r\n int cexpl (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagl' [-Wincompatible-library-redeclaration]\r\n int cimagl (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogl' [-Wincompatible-library-redeclaration]\r\n int clogl (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjl' [-Wincompatible-library-redeclaration]\r\n int conjl (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowl' [-Wincompatible-library-redeclaration]\r\n int cpowl (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowl' is a builtin with type '_Complex long double (_Complex long double, _Complex long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojl' [-Wincompatible-library-redeclaration]\r\n int cprojl (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creall' [-Wincompatible-library-redeclaration]\r\n int creall (void);\r\n ^\r\n _configtest.c:17:5: note: 'creall' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinl' [-Wincompatible-library-redeclaration]\r\n int csinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhl' [-Wincompatible-library-redeclaration]\r\n int csinhl (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtl' [-Wincompatible-library-redeclaration]\r\n int csqrtl (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanl' [-Wincompatible-library-redeclaration]\r\n int ctanl (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhl' [-Wincompatible-library-redeclaration]\r\n int ctanhl (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:2:12: warning: unused function 'static_func' [-Wunused-function]\r\n static int static_func (char * restrict a)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:3:19: warning: unused function 'static_func' [-Wunused-function]\r\n static inline int static_func (void)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h\r\n #define SIZEOF_PY_INTPTR_T 8\r\n #define SIZEOF_OFF_T 8\r\n #define SIZEOF_PY_LONG_LONG 8\r\n #define MATHLIB\r\n #define HAVE_SIN 1\r\n #define HAVE_COS 1\r\n #define HAVE_TAN 1\r\n #define HAVE_SINH 1\r\n #define HAVE_COSH 1\r\n #define HAVE_TANH 1\r\n #define HAVE_FABS 1\r\n #define HAVE_FLOOR 1\r\n #define HAVE_CEIL 1\r\n #define HAVE_SQRT 1\r\n #define HAVE_LOG10 1\r\n #define HAVE_LOG 1\r\n #define HAVE_EXP 1\r\n #define HAVE_ASIN 1\r\n #define HAVE_ACOS 1\r\n #define HAVE_ATAN 1\r\n #define HAVE_FMOD 1\r\n #define HAVE_MODF 1\r\n #define HAVE_FREXP 1\r\n #define HAVE_LDEXP 1\r\n #define HAVE_RINT 1\r\n #define HAVE_TRUNC 1\r\n #define HAVE_EXP2 1\r\n #define HAVE_LOG2 1\r\n #define HAVE_ATAN2 1\r\n #define HAVE_POW 1\r\n #define HAVE_NEXTAFTER 1\r\n #define HAVE_STRTOLL 1\r\n #define HAVE_STRTOULL 1\r\n #define HAVE_CBRT 1\r\n #define HAVE_STRTOLD_L 1\r\n #define HAVE_BACKTRACE 1\r\n #define HAVE_MADVISE 1\r\n #define HAVE_XMMINTRIN_H 1\r\n #define HAVE_EMMINTRIN_H 1\r\n #define HAVE_XLOCALE_H 1\r\n #define HAVE_DLFCN_H 1\r\n #define HAVE_SYS_MMAN_H 1\r\n #define HAVE___BUILTIN_ISNAN 1\r\n #define HAVE___BUILTIN_ISINF 1\r\n #define HAVE___BUILTIN_ISFINITE 1\r\n #define HAVE___BUILTIN_BSWAP32 1\r\n #define HAVE___BUILTIN_BSWAP64 1\r\n #define HAVE___BUILTIN_EXPECT 1\r\n #define HAVE___BUILTIN_MUL_OVERFLOW 1\r\n #define HAVE___BUILTIN_CPU_SUPPORTS 1\r\n #define HAVE__M_FROM_INT64 1\r\n #define HAVE__MM_LOAD_PS 1\r\n #define HAVE__MM_PREFETCH 1\r\n #define HAVE__MM_LOAD_PD 1\r\n #define HAVE___BUILTIN_PREFETCH 1\r\n #define HAVE_LINK_AVX 1\r\n #define HAVE_LINK_AVX2 1\r\n #define HAVE_XGETBV 1\r\n #define HAVE_ATTRIBUTE_NONNULL 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX2 1\r\n #define HAVE___THREAD 1\r\n #define HAVE_SINF 1\r\n #define HAVE_COSF 1\r\n #define HAVE_TANF 1\r\n #define HAVE_SINHF 1\r\n #define HAVE_COSHF 1\r\n #define HAVE_TANHF 1\r\n #define HAVE_FABSF 1\r\n #define HAVE_FLOORF 1\r\n #define HAVE_CEILF 1\r\n #define HAVE_RINTF 1\r\n #define HAVE_TRUNCF 1\r\n #define HAVE_SQRTF 1\r\n #define HAVE_LOG10F 1\r\n #define HAVE_LOGF 1\r\n #define HAVE_LOG1PF 1\r\n #define HAVE_EXPF 1\r\n #define HAVE_EXPM1F 1\r\n #define HAVE_ASINF 1\r\n #define HAVE_ACOSF 1\r\n #define HAVE_ATANF 1\r\n #define HAVE_ASINHF 1\r\n #define HAVE_ACOSHF 1\r\n #define HAVE_ATANHF 1\r\n #define HAVE_HYPOTF 1\r\n #define HAVE_ATAN2F 1\r\n #define HAVE_POWF 1\r\n #define HAVE_FMODF 1\r\n #define HAVE_MODFF 1\r\n #define HAVE_FREXPF 1\r\n #define HAVE_LDEXPF 1\r\n #define HAVE_EXP2F 1\r\n #define HAVE_LOG2F 1\r\n #define HAVE_COPYSIGNF 1\r\n #define HAVE_NEXTAFTERF 1\r\n #define HAVE_CBRTF 1\r\n #define HAVE_SINL 1\r\n #define HAVE_COSL 1\r\n #define HAVE_TANL 1\r\n #define HAVE_SINHL 1\r\n #define HAVE_COSHL 1\r\n #define HAVE_TANHL 1\r\n #define HAVE_FABSL 1\r\n #define HAVE_FLOORL 1\r\n #define HAVE_CEILL 1\r\n #define HAVE_RINTL 1\r\n #define HAVE_TRUNCL 1\r\n #define HAVE_SQRTL 1\r\n #define HAVE_LOG10L 1\r\n #define HAVE_LOGL 1\r\n #define HAVE_LOG1PL 1\r\n #define HAVE_EXPL 1\r\n #define HAVE_EXPM1L 1\r\n #define HAVE_ASINL 1\r\n #define HAVE_ACOSL 1\r\n #define HAVE_ATANL 1\r\n #define HAVE_ASINHL 1\r\n #define HAVE_ACOSHL 1\r\n #define HAVE_ATANHL 1\r\n #define HAVE_HYPOTL 1\r\n #define HAVE_ATAN2L 1\r\n #define HAVE_POWL 1\r\n #define HAVE_FMODL 1\r\n #define HAVE_MODFL 1\r\n #define HAVE_FREXPL 1\r\n #define HAVE_LDEXPL 1\r\n #define HAVE_EXP2L 1\r\n #define HAVE_LOG2L 1\r\n #define HAVE_COPYSIGNL 1\r\n #define HAVE_NEXTAFTERL 1\r\n #define HAVE_CBRTL 1\r\n #define HAVE_DECL_SIGNBIT\r\n #define HAVE_COMPLEX_H 1\r\n #define HAVE_CABS 1\r\n #define HAVE_CACOS 1\r\n #define HAVE_CACOSH 1\r\n #define HAVE_CARG 1\r\n #define HAVE_CASIN 1\r\n #define HAVE_CASINH 1\r\n #define HAVE_CATAN 1\r\n #define HAVE_CATANH 1\r\n #define HAVE_CCOS 1\r\n #define HAVE_CCOSH 1\r\n #define HAVE_CEXP 1\r\n #define HAVE_CIMAG 1\r\n #define HAVE_CLOG 1\r\n #define HAVE_CONJ 1\r\n #define HAVE_CPOW 1\r\n #define HAVE_CPROJ 1\r\n #define HAVE_CREAL 1\r\n #define HAVE_CSIN 1\r\n #define HAVE_CSINH 1\r\n #define HAVE_CSQRT 1\r\n #define HAVE_CTAN 1\r\n #define HAVE_CTANH 1\r\n #define HAVE_CABSF 1\r\n #define HAVE_CACOSF 1\r\n #define HAVE_CACOSHF 1\r\n #define HAVE_CARGF 1\r\n #define HAVE_CASINF 1\r\n #define HAVE_CASINHF 1\r\n #define HAVE_CATANF 1\r\n #define HAVE_CATANHF 1\r\n #define HAVE_CCOSF 1\r\n #define HAVE_CCOSHF 1\r\n #define HAVE_CEXPF 1\r\n #define HAVE_CIMAGF 1\r\n #define HAVE_CLOGF 1\r\n #define HAVE_CONJF 1\r\n #define HAVE_CPOWF 1\r\n #define HAVE_CPROJF 1\r\n #define HAVE_CREALF 1\r\n #define HAVE_CSINF 1\r\n #define HAVE_CSINHF 1\r\n #define HAVE_CSQRTF 1\r\n #define HAVE_CTANF 1\r\n #define HAVE_CTANHF 1\r\n #define HAVE_CABSL 1\r\n #define HAVE_CACOSL 1\r\n #define HAVE_CACOSHL 1\r\n #define HAVE_CARGL 1\r\n #define HAVE_CASINL 1\r\n #define HAVE_CASINHL 1\r\n #define HAVE_CATANL 1\r\n #define HAVE_CATANHL 1\r\n #define HAVE_CCOSL 1\r\n #define HAVE_CCOSHL 1\r\n #define HAVE_CEXPL 1\r\n #define HAVE_CIMAGL 1\r\n #define HAVE_CLOGL 1\r\n #define HAVE_CONJL 1\r\n #define HAVE_CPOWL 1\r\n #define HAVE_CPROJL 1\r\n #define HAVE_CREALL 1\r\n #define HAVE_CSINL 1\r\n #define HAVE_CSINHL 1\r\n #define HAVE_CSQRTL 1\r\n #define HAVE_CTANL 1\r\n #define HAVE_CTANHL 1\r\n #define NPY_RESTRICT restrict\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define HAVE_LDOUBLE_INTEL_EXTENDED_16_BYTES_LE 1\r\n #define NPY_PY3K 1\r\n #ifndef __cplusplus\r\n \/* #undef inline *\/\r\n #endif\r\n \r\n #ifndef _NPY_NPY_CONFIG_H_\r\n #error config.h should never be included directly, include npy_config.h instead\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n Generating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'exp' [-Wincompatible-library-redeclaration]\r\n int exp (void);\r\n ^\r\n _configtest.c:1:5: note: 'exp' is a builtin with type 'double (double)'\r\n 1 warning generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n #define NPY_SIZEOF_SHORT SIZEOF_SHORT\r\n #define NPY_SIZEOF_INT SIZEOF_INT\r\n #define NPY_SIZEOF_LONG SIZEOF_LONG\r\n #define NPY_SIZEOF_FLOAT 4\r\n #define NPY_SIZEOF_COMPLEX_FLOAT 8\r\n #define NPY_SIZEOF_DOUBLE 8\r\n #define NPY_SIZEOF_COMPLEX_DOUBLE 16\r\n #define NPY_SIZEOF_LONGDOUBLE 16\r\n #define NPY_SIZEOF_COMPLEX_LONGDOUBLE 32\r\n #define NPY_SIZEOF_PY_INTPTR_T 8\r\n #define NPY_SIZEOF_OFF_T 8\r\n #define NPY_SIZEOF_PY_LONG_LONG 8\r\n #define NPY_SIZEOF_LONGLONG 8\r\n #define NPY_NO_SMP 0\r\n #define NPY_HAVE_DECL_ISNAN\r\n #define NPY_HAVE_DECL_ISINF\r\n #define NPY_HAVE_DECL_ISFINITE\r\n #define NPY_HAVE_DECL_SIGNBIT\r\n #define NPY_USE_C99_COMPLEX 1\r\n #define NPY_HAVE_COMPLEX_DOUBLE 1\r\n #define NPY_HAVE_COMPLEX_FLOAT 1\r\n #define NPY_HAVE_COMPLEX_LONG_DOUBLE 1\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define NPY_USE_C99_FORMATS 1\r\n #define NPY_VISIBILITY_HIDDEN __attribute__((visibility(\"hidden\")))\r\n #define NPY_ABI_VERSION 0x01000009\r\n #define NPY_API_VERSION 0x0000000D\r\n \r\n #ifndef __STDC_FORMAT_MACROS\r\n #define __STDC_FORMAT_MACROS 1\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h']\r\n building extension \"numpy.core._multiarray_tests\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n building extension \"numpy.core._multiarray_umath\" sources\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_ufunc_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common' to include_dirs.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_internal.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h']\r\n building extension \"numpy.core._umath_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n building extension \"numpy.core._rational_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n building extension \"numpy.core._struct_ufunc_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n building extension \"numpy.core._operand_flag_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n building extension \"numpy.fft.fftpack_lite\" sources\r\n building extension \"numpy.linalg.lapack_lite\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n building extension \"numpy.linalg._umath_linalg\" sources\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n building extension \"numpy.random.mtrand\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/random\r\n building data_files sources\r\n build_src: building npy-pkg config files\r\n running build_py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/conftest.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_globals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/dual.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_distributor_init.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/ctypeslib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/matlib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_pytesttester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/py3k.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/_inspect.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/fromnumeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_add_newdocs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_methods.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_internal.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_string_helpers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/multiarray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/records.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup_common.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_aliased_types.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/memmap.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/overrides.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/getlimits.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype_ctypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/defchararray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/machar.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/einsumfunc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath_tests.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numerictypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_type_aliases.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/cversions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/arrayprint.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/code_generators\/generate_numpy_api.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/unixccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/numpy_distribution.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/conv_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/cpuinfo.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvc9compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/npy_pkg_config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/compat.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/misc_util.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/log.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/line_endings.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/lib2def.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/pathccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/system_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/exec_command.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/from_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/mingw32ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/extension.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/intelccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/distutils\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config_compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_ext.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_headers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_py.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_src.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/sdist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_scripts.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/bdist_rpm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/autodist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/egg_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/develop.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_data.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/gnu.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/compaq.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/intel.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/none.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/nag.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/ibm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/sun.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/lahey.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/g95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/mips.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/hpux.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/environment.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pathf95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/absoft.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/vast.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/misc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/internals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/creation.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/constants.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/ufuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/broadcasting.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/basics.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/subclassing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/indexing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/byteswapping.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/structured_arrays.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/glossary.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/common_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/crackfortran.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cb_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py2e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/func2subr.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/diagnose.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/capi_maps.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f90mod_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py_testing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/use_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/auxfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__main__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/helper.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/fftpack.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_iotools.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/mixins.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/nanfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/recfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/histograms.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/scimath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/user_array.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/format.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/twodim_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/financial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/index_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/npyio.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/stride_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arrayterator.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraysetops.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraypad.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/type_check.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_datasource.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/ufunclike.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/linalg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/extras.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/testutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/bench.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/timer_comparison.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/mrecords.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/defmatrix.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/laguerre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/_polybase.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polyutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite_e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/chebyshev.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/legendre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/print_coercion_tables.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/parameterized.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n running build_clib\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_clib\r\n building 'npymath' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n ar: adding 4 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n building 'npysort' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/quicksort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/mergesort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/heapsort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/selection.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/binsearch.c\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n 22 warnings generated.\r\n ar: adding 5 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n running build_ext\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_ext\r\n building 'numpy.core._dummy' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/dummymodule.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/dummymodule.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_dummy.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_tests.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_umath' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/core\/src\/multiarray\/alloc.c\r\n clang: numpy\/core\/src\/multiarray\/calculation.cclang: numpy\/core\/src\/multiarray\/array_assign_scalar.c\r\n clang: numpy\/core\/src\/multiarray\/convert.c\r\n \r\n clang: numpy\/core\/src\/multiarray\/ctors.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busday.c\r\n clang: numpy\/core\/src\/multiarray\/dragon4.cclang: numpy\/core\/src\/multiarray\/flagsobject.c\r\n \r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/arrayobject.c\r\n clang: numpy\/core\/src\/multiarray\/array_assign_array.c\r\n clang: numpy\/core\/src\/multiarray\/convert_datatype.c\r\n clang: numpy\/core\/src\/multiarray\/getset.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busdaycal.c\r\n clang: numpy\/core\/src\/multiarray\/buffer.c\r\n clang: numpy\/core\/src\/multiarray\/compiled_base.c\r\n clang: numpy\/core\/src\/multiarray\/hashdescr.c\r\n clang: numpy\/core\/src\/multiarray\/descriptor.c\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/conversion_utils.c\r\n clang: numpy\/core\/src\/multiarray\/item_selection.c\r\n clang: numpy\/core\/src\/multiarray\/dtype_transfer.c\r\n clang: numpy\/core\/src\/multiarray\/mapping.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n 3 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/datetime.c\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:477:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ptr = PyUnicode_AS_UNICODE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/common.c\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 6 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/nditer_pywrap.c\r\n 9 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/sequence.c\r\n clang: numpy\/core\/src\/multiarray\/shape.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n clang: numpy\/core\/src\/multiarray\/methods.c\r\n clang: numpy\/core\/src\/multiarray\/iterators.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_strings.c\r\n clang: numpy\/core\/src\/multiarray\/number.c\r\n clang: numpy\/core\/src\/multiarray\/scalarapi.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n numpy\/core\/src\/multiarray\/scalarapi.c:74:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:135:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1849:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buffer = PyUnicode_AS_DATA(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 5 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/typeinfo.c\r\n clang: numpy\/core\/src\/multiarray\/refcount.c\r\n clang: numpy\/core\/src\/multiarray\/usertypes.c\r\n clang: numpy\/core\/src\/multiarray\/multiarraymodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n clang: numpy\/core\/src\/multiarray\/vdot.c\r\n clang: numpy\/core\/src\/umath\/umathmodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n clang: numpy\/core\/src\/umath\/reduction.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n clang: numpy\/core\/src\/multiarray\/nditer_api.c\r\n 14 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/strfuncs.c\r\n numpy\/core\/src\/umath\/loops.c.src:655:18: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n result = PyEval_CallObject(tocall, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:178:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_ReprFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:195:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_StrFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 2 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/temp_elide.c\r\n clang: numpy\/core\/src\/umath\/cpuid.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n clang: numpy\/core\/src\/umath\/ufunc_object.c\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'byte_long' [-Wunused-function]\r\n byte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ubyte_long' [-Wunused-function]\r\n ubyte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'short_long' [-Wunused-function]\r\n short_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ushort_long' [-Wunused-function]\r\n ushort_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'int_long' [-Wunused-function]\r\n int_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'uint_long' [-Wunused-function]\r\n uint_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'long_long' [-Wunused-function]\r\n long_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulong_long' [-Wunused-function]\r\n ulong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longlong_long' [-Wunused-function]\r\n longlong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulonglong_long' [-Wunused-function]\r\n ulonglong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'half_long' [-Wunused-function]\r\n half_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'float_long' [-Wunused-function]\r\n float_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'double_long' [-Wunused-function]\r\n double_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longdouble_long' [-Wunused-function]\r\n longdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cfloat_long' [-Wunused-function]\r\n cfloat_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cdouble_long' [-Wunused-function]\r\n cdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'clongdouble_long' [-Wunused-function]\r\n clongdouble_long(PyObject *obj)\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/nditer_constr.c\r\n numpy\/core\/src\/umath\/ufunc_object.c:657:19: warning: comparison of integers of different signs: 'int' and 'size_t' (aka 'unsigned long') [-Wsign-compare]\r\n for (i = 0; i < len; i++) {\r\n ~ ^ ~~~\r\n clang: numpy\/core\/src\/umath\/override.c\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n clang: numpy\/core\/src\/common\/array_assign.c\r\n clang: numpy\/core\/src\/common\/ufunc_override.c\r\n clang: numpy\/core\/src\/common\/npy_longdouble.c\r\n clang: numpy\/core\/src\/common\/numpyos.c\r\n clang: numpy\/core\/src\/common\/ucsnarrow.c\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/umath\/extobj.c\r\n numpy\/core\/src\/common\/ucsnarrow.c:139:34: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n ret = (PyUnicodeObject *)PyUnicode_FromUnicode((Py_UNICODE*)buf,\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/common\/python_xerbla.c\r\n clang: numpy\/core\/src\/common\/cblasfuncs.c\r\n clang: \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c\r\n In file included from \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c:26:\r\n In file included from numpy\/core\/include\/numpy\/arrayobject.h:4:\r\n In file included from numpy\/core\/include\/numpy\/ndarrayobject.h:21:\r\n build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h:1463:1: warning: unused function '_import_array' [-Wunused-function]\r\n _import_array(void)\r\n ^\r\n 1 warning generated.\r\n 17 warnings generated.\r\n clang: numpy\/core\/src\/umath\/ufunc_type_resolution.c\r\n 4 warnings generated.\r\n 4 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/alloc.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arrayobject.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_scalar.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_array.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/buffer.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/calculation.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/compiled_base.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/common.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert_datatype.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/conversion_utils.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/ctors.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_strings.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busday.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busdaycal.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/descriptor.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dragon4.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dtype_transfer.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/flagsobject.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/getset.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/hashdescr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/item_selection.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/iterators.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/mapping.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/methods.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/multiarraymodule.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_api.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_constr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_pywrap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/number.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/refcount.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/sequence.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/shape.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalarapi.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/strfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/temp_elide.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/typeinfo.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/usertypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/vdot.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/umathmodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/reduction.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_object.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/extobj.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/cpuid.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_type_resolution.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/halffloat.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/array_assign.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/npy_longdouble.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ucsnarrow.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ufunc_override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/numpyos.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/cblasfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/python_xerbla.o build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -lnpysort -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_umath.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.core._umath_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_umath_tests.cpython-39-darwin.so\r\n building 'numpy.core._rational_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_rational_tests.cpython-39-darwin.so\r\n building 'numpy.core._struct_ufunc_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_struct_ufunc_tests.cpython-39-darwin.so\r\n building 'numpy.core._operand_flag_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_operand_flag_tests.cpython-39-darwin.so\r\n building 'numpy.fft.fftpack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n compile options: '-Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/fft\/fftpack_litemodule.c\r\n clang: numpy\/fft\/fftpack.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_lite.cpython-39-darwin.so\r\n building 'numpy.linalg.lapack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/linalg\/lapack_litemodule.c\r\n clang: numpy\/linalg\/lapack_lite\/python_xerbla.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.linalg._umath_linalg' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n numpy\/linalg\/umath_linalg.c.src:735:32: warning: unknown warning group '-Wmaybe-uninitialized', ignored [-Wunknown-warning-option]\r\n #pragma GCC diagnostic ignored \"-Wmaybe-uninitialized\"\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:541:1: warning: unused function 'dump_ufunc_object' [-Wunused-function]\r\n dump_ufunc_object(PyUFuncObject* ufunc)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:566:1: warning: unused function 'dump_linearize_data' [-Wunused-function]\r\n dump_linearize_data(const char* name, const LINEARIZE_DATA_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_FLOAT_matrix' [-Wunused-function]\r\n dump_FLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_DOUBLE_matrix' [-Wunused-function]\r\n dump_DOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CFLOAT_matrix' [-Wunused-function]\r\n dump_CFLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CDOUBLE_matrix' [-Wunused-function]\r\n dump_CDOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_FLOAT_matrix' [-Wunused-function]\r\n zero_FLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_DOUBLE_matrix' [-Wunused-function]\r\n zero_DOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CFLOAT_matrix' [-Wunused-function]\r\n zero_CFLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CDOUBLE_matrix' [-Wunused-function]\r\n zero_CDOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:1862:1: warning: unused function 'dump_geev_params' [-Wunused-function]\r\n dump_geev_params(const char *name, GEEV_PARAMS_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2132:1: warning: unused function 'init_cgeev' [-Wunused-function]\r\n init_cgeev(GEEV_PARAMS_t* params,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2213:1: warning: unused function 'process_cgeev_results' [-Wunused-function]\r\n process_cgeev_results(GEEV_PARAMS_t *NPY_UNUSED(params))\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2376:1: warning: unused function 'dump_gesdd_params' [-Wunused-function]\r\n dump_gesdd_params(const char *name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2864:1: warning: unused function 'dump_gelsd_params' [-Wunused-function]\r\n dump_gelsd_params(const char *name,\r\n ^\r\n 16 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/_umath_linalg.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.random.mtrand' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\r\n compile options: '-D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/random\/mtrand\/mtrand.c\r\n clang: numpy\/random\/mtrand\/initarray.cclang: numpy\/random\/mtrand\/randomkit.c\r\n \r\n clang: numpy\/random\/mtrand\/distributions.c\r\n numpy\/random\/mtrand\/mtrand.c:40400:34: error: no member named 'tp_print' in 'struct _typeobject'\r\n __pyx_type_6mtrand_RandomState.tp_print = 0;\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 12 warnings and 1 error generated.\r\n error: Command \"clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c numpy\/random\/mtrand\/mtrand.c -o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o -MMD -MF build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o.d\" failed with exit status 1\nDamn sorry","embeddings":[-0.372736305,-0.0714854077,-0.0845708922,0.2079993039,0.2768215835,0.0631752834,-0.2505984604,0.0786435232,-0.2070377618,-0.0340085514,0.0305880085,0.3349942863,-0.1231444255,0.0895890892,0.0807427391,0.0722528622,0.189591229,0.2151816189,-0.3760947287,0.0339735448,0.0351168886,0.1739135981,-0.1176334247,-0.0769590884,-0.1111136004,-0.2832199335,0.0382908024,0.1094798371,-0.532479167,-0.3132808805,0.291446507,0.0252360553,-0.0163556207,0.5489337444,-0.0001225818,0.1088191941,0.1684155613,0.018390229,-0.3556295037,-0.4525223374,-0.2284404039,-0.4977667034,0.1029686555,0.0328003541,0.2697275281,-0.2269828916,-0.1955371499,-0.3135961592,0.7321167588,0.186695531,0.1548049748,0.3139568865,0.1280289441,0.0492491126,0.1173379719,0.3006310165,-0.3177905381,0.3576257825,0.2570489943,-0.1687269062,0.2200257182,0.1589555889,-0.2279057503,0.2082943469,0.5441326499,-0.1704113036,-0.0771712363,-0.4264783263,-0.2593541741,0.2732714415,0.5033100843,-0.2703568041,-0.6291657686,-0.0928478241,0.0548009649,-0.2381634116,0.3260372877,0.223864004,-0.2062843293,0.0802830532,-0.4288931787,-0.280636698,-0.1864999384,0.2670033574,-0.0238301586,0.164580822,-0.1459159702,0.1176900715,0.2985162139,-0.0034400814,0.2251620293,0.0791203454,-0.0705096424,0.0032631624,-0.2801685035,0.0655268356,-0.2036794722,0.2356178463,0.1992911696,0.1688992977,-0.1797399819,-0.2005181015,-0.2243610919,0.0616602823,0.066762954,0.2027508169,-0.1514637321,0.1150361076,0.3352227509,0.2663400173,0.4560080171,0.0843001902,0.0707917213,-0.1902450472,-0.209478572,-0.1111337766,0.1524448395,-0.3028428555,-0.4012164772,0.1194737256,0.0135395043,0.1056581885,0.0688990727,0.1510053128,-0.1464695483,0.1697680354,0.2481278181,0.3707648218,-0.232209444,0.0304407589,-0.0021012723,-0.0312458072,-0.2525970638,0.0168092307,0.1831142902,-0.1020761356,0.3657963574,0.2121508121,0.1536038369,0.1180837601,0.2853684127,0.0168460198,-0.191518262,0.4379527569,-0.2559596002,0.2257401049,0.0813604146,0.3052143753,0.0385886207,-0.0087099858,0.1200957373,-0.0894460753,-0.1464142799,0.0493404903,-0.2953626513,0.051196415,-0.16042234,-0.0320122093,0.0422822423,-0.0837064162,0.015641503,-0.2536138892,0.0998989642,-0.1036708131,0.3004768193,0.414516747,-0.51960361,-0.038707234,-0.1943658739,-0.1891062409,-0.0495771132,0.1207961589,-0.1165530235,0.1363425553,-0.164071247,-0.4210309684,0.2341786772,-0.3646256328,-0.367303431,0.1608417332,-0.1646376401,-0.086746715,0.0553342775,0.0455297008,-0.0180610958,0.0614581928,-0.0477216803,0.326702714,-0.1066503674,-0.2653565109,0.0084401397,-0.5503122211,0.0675886497,0.0629741848,-0.0602851324,0.0150617491,-0.0644270927,0.139841333,0.0511045381,0.0582192577,-0.0014922149,0.6988541484,0.416780889,0.0951713175,-0.2119343877,-0.2936756313,-0.3661433756,0.1583058536,0.0255879425,0.0643135011,-0.3344601095,-0.0186114982,-0.1848731637,0.128934145,-0.0381260812,-0.0527399741,-0.0203619599,0.0027848945,0.2437096834,-0.2807203233,-0.233386904,0.5044182539,0.1574977785,0.2694315314,-0.4862544239,0.0366727933,-0.3219184875,-0.1439168453,0.1915305704,0.2339345664,0.0553197712,-0.2622070611,-0.1443571001,0.3139041066,-0.1654947251,-0.1707349569,-0.0076072491,0.1440196335,0.2182890773,-0.2099311203,0.2413927317,-0.0720721483,-0.1406363845,0.0983242542,-0.0743637085,-0.1697815061,0.0582351796,0.3382672966,-0.0074652079,0.3404535055,0.3141412735,0.0584230796,-0.1450206935,-0.4074391723,0.396992296,-0.0027272592,0.4638343155,-0.1445932984,0.0697035044,0.053041961,-0.1363226324,0.0916765481,-0.0233560391,0.0049676136,-0.072430402,0.1785400212,0.185349673,0.0348233506,0.5852743983,0.0619082898,-0.1347470284,0.1598725617,-0.0658946782,-0.0729199722,0.0373055674,0.1743800044,-0.0199271459,-0.0330145024,0.0282345489,-0.0624522939,-0.0412191078,-0.207671389,0.1253645271,0.2953568995,-0.3419783413,0.1114570275,-0.5203249454,-0.0051726159,-0.0906571373,-0.0719850361,-0.4134817719,-0.2235808372,-0.1058893874,0.1500547081,0.2082414627,0.1049688607,0.1166992337,0.0297720432,0.001696527,-0.209155798,-0.1628497243,-0.1249261945,-0.0730912238,-0.0193224195,0.1679847986,0.1203907058,0.4714294374,-0.0997886732,-0.1613499224,-0.4218867421,-0.261598736,0.107161954,-0.398339957,0.742423296,0.469727993,0.202538535,0.0647644475,0.1159555018,0.1875865012,-0.1887231469,-0.1085140929,-0.0309973564,-0.213460058,-0.2505242229,-0.1225516498,-0.3344918489,-0.32494241,-0.2550111711,0.1577945501,-0.0034383023,-0.1047476232,0.3142209351,0.2670063674,0.0404987186,-0.165805921,0.0468914136,-0.0268509146,-0.2246979922,0.177915588,-0.201831609,-0.2568696737,0.3019668162,0.2744387388,0.1829818487,0.2670661211,-0.7614741325,-0.2670167983,-0.3597413003,0.2398702502,-0.1860815585,0.2179733366,0.1911988556,0.0769287571,0.0779739991,-0.2079752088,-0.4608823061,-0.0097571854,0.0752782673,0.4969078898,0.0363830738,0.3437930942,-0.0585201345,0.6186058521,0.397529155,0.0103019681,0.1783744097,0.2334723771,0.5895819068,0.0778009444,-0.4482364953,0.2397813201,-0.0084547177,0.1363245547,-0.1942522824,0.2197874635,0.2414545864,-0.277361989,0.2657617033,0.0042451476,0.0055124569,-0.2210909873,-0.13537018,0.3050734699,0.2228150517,0.0064602434,0.0430852436,-0.1302471906,0.0280230325,0.1745655835,0.010982167,0.3000519872,-0.1954792738,-0.1373809576,-0.183105126,0.2909622788,-0.0428895615,0.4082678556,-0.2521476746,0.4757547975,0.0345090851,0.0108251274,0.475733906,-0.3083002269,-0.1645100862,0.1024655625,0.1262150258,-0.6377895474,-0.1250775605,-0.1298962086,0.1647143066,0.248098284,0.2099779695,-0.4318394959,-0.1485227644,0.2467826158,0.5641607046,0.0371086225,-0.1298632622,-0.3288927078,-0.3882852793,-0.2767420411,-0.1371272653,-0.058102645,-0.1284619868,0.2001442164,-0.0588850789,0.1972738355,0.1628406644,0.1069483757,-0.0801087394,0.2060672939,-0.1776001751,0.2210780978,0.3728102148,-0.1271787584,0.1659846008,0.737244606,-0.2061559558,-0.3858459294,-0.0367117412,0.1224328429,0.1555299163,0.3913010359,-0.2864465714,-0.1263913363,0.2399162054,0.2055450827,0.1049313173,0.14691028,0.289960444,0.1443811804,-0.195566237,-0.4653308094,0.1165174395,0.4374177158,0.1289001107,0.4145717323,0.0137373563,0.0171636641,-0.0216146,0.2067595273,0.8317614794,-0.0005798346,0.0876606479,0.3389310539,-0.0641187727,0.5443438888,-0.4247684479,0.04840637,-0.0802192986,-0.143647477,-0.1364471316,-0.3701676726,0.3992536366,0.1148540974,-0.2142028362,0.3585865796,-0.1218098328,0.29375422,0.0482256673,0.1957940161,-0.0632860139,-0.2802743912,-0.1753268093,0.1845365763,-0.233880952,0.1757169068,-0.1705885977,0.083735086,-0.3188177347,-0.0867874846,-0.5056916475,0.0665702522,-0.443862617,0.2258304805,-0.0344874859,-0.2310214341,0.3094379306,0.1276277453,0.3448578715,0.4515912831,-0.3112368286,0.1110853925,-0.2714110315,-0.2467177361,0.3035934865,0.0439759605,0.0467805564,-0.0107217925,-0.1727659255,-0.0625354499,-0.2353877276,0.0169170164,0.0544819459,-0.0357779078,-0.1322258413,-0.1305123866,-0.1653528064,-0.1025140062,0.2076852918,-0.2839880288,0.0506421104,0.07781367,-0.0196289588,0.4598361552,-0.1750662029,-0.2117223293,-0.1011548117,0.212752223,-0.1656555086,-0.2982017696,0.234419629,0.2739279866,-0.3369333148,-0.1571041346,-0.094751589,-0.1557620764,-0.4031636715,0.1162526757,0.1045675799,-0.1815831959,-0.1145593971,0.2249464691,0.2387459278,-0.2340987474,0.0011506572,-0.4424839914,-0.2236991078,0.2046369612,-0.1303168684,0.0251450893,0.1048556641,0.0480054915,-0.0482010245,0.1132196635,-0.1702207923,-0.1554876715,0.0242763478,-0.0713960975,0.3412455618,-0.0761460811,0.0959786624,-0.2396223247,-0.03813833,0.1586219817,-0.2529083788,-0.076279223,-0.0355480611,0.137926355,0.0933924913,-0.3160541952,0.212218523,-0.4301118255,-0.1320781857,-0.1007387489,0.183366701,0.2039175779,0.0625451133,-0.2080432922,-0.4835402668,0.0814061463,0.0820645839,0.2181892693,0.0904808417,0.0420457087,0.2254475355,0.0915967301,0.1900683045,-0.09184926,-0.3334983885,0.0691129938,0.0046931463,0.2152244151,0.1156246215,-0.3056086898,0.2381537408,0.2715603411,0.4740514159,0.5606653094,-0.0505268052,0.2009599954,0.3780045509,0.0619235337,-0.3095085323,-0.1179374307,0.0978347436,0.211131677,-0.0669629127,0.294595927,0.2904630005,-0.1662479341,0.1414123923,0.2137935907,0.2448209524,0.0442248173,0.2401469797,0.1740344614,0.1456380934,-0.0191844087,0.2478772849,0.2596318722,0.1486105621,0.2685900927,-0.0940031186,-0.0129717235,0.1056639478,0.2145727575,-0.1893776506,-0.3154191971,-0.0778527632,-0.14081873,0.3885463476,0.0168743245,-0.0760412812,0.2765650153,-0.3946425617,-0.3519563675,-0.1730092168,0.2614917159,0.0503016636,0.2280521244,0.3215574324,0.0809561014,0.1780232787,0.0048405635,0.0397591293,-0.4191242456,0.2413220108,0.1764912158,-0.1236993223,-0.770821631,0.2125341445,0.1664997637,-0.1692295074,-0.0538866408,0.3296135664,0.1637779623,-0.1454275995,-0.2812110484,0.5068164468,0.5802744627,0.3409990072,-0.0469977558,0.1705626249,-0.2142198682,0.0777670294,-0.0155832991,0.2700774372,-0.1369200051,0.154964745,0.236733079,0.0225825254,-0.1113553271,-0.1749951243,0.05340923,0.4135027826,-0.1955527812,0.1946346015,-0.1500745118,-0.0271252468,-0.0973940268,-0.0908994824,-0.4811028838,0.2690776885,0.5110985637,0.0316910855,-0.0370604061,-0.1930583566,0.0272696707,-0.0880471319,0.3524516523,0.495470494,0.191963926,-0.4115162492,-0.0745142847,-0.6475214958,-0.2043262273,-0.3776461184,-0.2580498755,0.1304243505,-0.0521419086,0.0602565371,0.3123220801,0.2321897298,0.2287330478,0.1469683051,-0.0183709953,-0.3618494868,-0.1180011779,0.0987984166,0.1109670922,0.0805131122,-0.2171806246,0.0052382955,0.2312532216,-0.2043973804,-0.2440494448,-0.0672755912,0.0961156338,0.0278290417,0.3226872683,0.0853667632,0.3677543998,0.2596828043,-0.2449216247,-0.4981413782,-0.2963164449,-0.1827402264,0.3557329476,-0.046792753,0.2464546263,-0.5367340446,0.0496459119,-0.5474626422,0.097678706,0.0588760599,-0.0952013209,-0.2168268859,0.2665975988,-0.215844512,0.3078436255,0.0800647959,0.1987073421,-0.0550648682,0.2909039557,-0.4305544794,-0.322242856,0.3843740821,0.168199867,-0.3109259009,0.3755890727,0.0602345243,0.2313626111,-0.4108432531,-0.5497607589,0.1153397188,0.0295620337,-0.0591647848,-0.2287070155,0.1518161595,-0.1175127327,-0.0596970692,0.009724074,0.5295980573,0.0082971165,-0.254521817,0.3482492864,0.1984189898]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1696","title":"Unable to install datasets","comments":"Damn sorry","body":"** Edit **\r\nI believe there's a bug with the package when you're installing it with Python 3.9. I recommend sticking with previous versions. Thanks, @thomwolf for the insight! \r\n\r\n**Short description**\r\n\r\nI followed the instructions for installing datasets (https:\/\/huggingface.co\/docs\/datasets\/installation.html). However, while I tried to download datasets using `pip install datasets` I got a massive error message after getting stuck at \"Installing build dependencies...\" \r\n\r\nI was wondering if this problem can be fixed by creating a virtual environment, but it didn't help. Can anyone offer some advice on how to fix this issue? \r\n\r\nHere's an error message: \r\n\r\n`(env) Gas-MacBook-Pro:Downloads destiny$ pip install datasets\r\nCollecting datasets\r\n Using cached datasets-1.2.0-py3-none-any.whl (159 kB)\r\nCollecting numpy>=1.17\r\n Using cached numpy-1.19.5-cp39-cp39-macosx_10_9_x86_64.whl (15.6 MB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-2.0.0.tar.gz (58.9 MB)\r\n....\r\n\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceilf' [-Wincompatible-library-redeclaration]\r\n int ceilf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceilf' is a builtin with type 'float (float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintf' [-Wincompatible-library-redeclaration]\r\n int rintf (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintf' is a builtin with type 'float (float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncf' [-Wincompatible-library-redeclaration]\r\n int truncf (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncf' is a builtin with type 'float (float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtf' [-Wincompatible-library-redeclaration]\r\n int sqrtf (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtf' is a builtin with type 'float (float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10f' [-Wincompatible-library-redeclaration]\r\n int log10f (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10f' is a builtin with type 'float (float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logf' [-Wincompatible-library-redeclaration]\r\n int logf (void);\r\n ^\r\n _configtest.c:14:5: note: 'logf' is a builtin with type 'float (float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pf' [-Wincompatible-library-redeclaration]\r\n int log1pf (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pf' is a builtin with type 'float (float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expf' [-Wincompatible-library-redeclaration]\r\n int expf (void);\r\n ^\r\n _configtest.c:16:5: note: 'expf' is a builtin with type 'float (float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1f' [-Wincompatible-library-redeclaration]\r\n int expm1f (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1f' is a builtin with type 'float (float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinf' [-Wincompatible-library-redeclaration]\r\n int asinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinf' is a builtin with type 'float (float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosf' [-Wincompatible-library-redeclaration]\r\n int acosf (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosf' is a builtin with type 'float (float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanf' [-Wincompatible-library-redeclaration]\r\n int atanf (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanf' is a builtin with type 'float (float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhf' [-Wincompatible-library-redeclaration]\r\n int asinhf (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhf' is a builtin with type 'float (float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshf' [-Wincompatible-library-redeclaration]\r\n int acoshf (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshf' is a builtin with type 'float (float)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhf' [-Wincompatible-library-redeclaration]\r\n int atanhf (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhf' is a builtin with type 'float (float)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotf' [-Wincompatible-library-redeclaration]\r\n int hypotf (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotf' is a builtin with type 'float (float, float)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2f' [-Wincompatible-library-redeclaration]\r\n int atan2f (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2f' is a builtin with type 'float (float, float)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powf' [-Wincompatible-library-redeclaration]\r\n int powf (void);\r\n ^\r\n _configtest.c:26:5: note: 'powf' is a builtin with type 'float (float, float)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodf' [-Wincompatible-library-redeclaration]\r\n int fmodf (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodf' is a builtin with type 'float (float, float)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modff' [-Wincompatible-library-redeclaration]\r\n int modff (void);\r\n ^\r\n _configtest.c:28:5: note: 'modff' is a builtin with type 'float (float, float *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpf' [-Wincompatible-library-redeclaration]\r\n int frexpf (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpf' is a builtin with type 'float (float, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpf' [-Wincompatible-library-redeclaration]\r\n int ldexpf (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpf' is a builtin with type 'float (float, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2f' [-Wincompatible-library-redeclaration]\r\n int exp2f (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2f' is a builtin with type 'float (float)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2f' [-Wincompatible-library-redeclaration]\r\n int log2f (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2f' is a builtin with type 'float (float)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignf' [-Wincompatible-library-redeclaration]\r\n int copysignf (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignf' is a builtin with type 'float (float, float)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterf' [-Wincompatible-library-redeclaration]\r\n int nextafterf (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterf' is a builtin with type 'float (float, float)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtf' [-Wincompatible-library-redeclaration]\r\n int cbrtf (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtf' is a builtin with type 'float (float)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'sinl' [-Wincompatible-library-redeclaration]\r\n int sinl (void);\r\n ^\r\n _configtest.c:1:5: note: 'sinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cosl' [-Wincompatible-library-redeclaration]\r\n int cosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'tanl' [-Wincompatible-library-redeclaration]\r\n int tanl (void);\r\n ^\r\n _configtest.c:3:5: note: 'tanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'sinhl' [-Wincompatible-library-redeclaration]\r\n int sinhl (void);\r\n ^\r\n _configtest.c:4:5: note: 'sinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'coshl' [-Wincompatible-library-redeclaration]\r\n int coshl (void);\r\n ^\r\n _configtest.c:5:5: note: 'coshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'tanhl' [-Wincompatible-library-redeclaration]\r\n int tanhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'tanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'fabsl' [-Wincompatible-library-redeclaration]\r\n int fabsl (void);\r\n ^\r\n _configtest.c:7:5: note: 'fabsl' is a builtin with type 'long double (long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'floorl' [-Wincompatible-library-redeclaration]\r\n int floorl (void);\r\n ^\r\n _configtest.c:8:5: note: 'floorl' is a builtin with type 'long double (long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceill' [-Wincompatible-library-redeclaration]\r\n int ceill (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceill' is a builtin with type 'long double (long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintl' [-Wincompatible-library-redeclaration]\r\n int rintl (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintl' is a builtin with type 'long double (long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncl' [-Wincompatible-library-redeclaration]\r\n int truncl (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncl' is a builtin with type 'long double (long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtl' [-Wincompatible-library-redeclaration]\r\n int sqrtl (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtl' is a builtin with type 'long double (long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10l' [-Wincompatible-library-redeclaration]\r\n int log10l (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10l' is a builtin with type 'long double (long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logl' [-Wincompatible-library-redeclaration]\r\n int logl (void);\r\n ^\r\n _configtest.c:14:5: note: 'logl' is a builtin with type 'long double (long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pl' [-Wincompatible-library-redeclaration]\r\n int log1pl (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pl' is a builtin with type 'long double (long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expl' [-Wincompatible-library-redeclaration]\r\n int expl (void);\r\n ^\r\n _configtest.c:16:5: note: 'expl' is a builtin with type 'long double (long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1l' [-Wincompatible-library-redeclaration]\r\n int expm1l (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1l' is a builtin with type 'long double (long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinl' [-Wincompatible-library-redeclaration]\r\n int asinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosl' [-Wincompatible-library-redeclaration]\r\n int acosl (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanl' [-Wincompatible-library-redeclaration]\r\n int atanl (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhl' [-Wincompatible-library-redeclaration]\r\n int asinhl (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshl' [-Wincompatible-library-redeclaration]\r\n int acoshl (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhl' [-Wincompatible-library-redeclaration]\r\n int atanhl (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotl' [-Wincompatible-library-redeclaration]\r\n int hypotl (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2l' [-Wincompatible-library-redeclaration]\r\n int atan2l (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2l' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powl' [-Wincompatible-library-redeclaration]\r\n int powl (void);\r\n ^\r\n _configtest.c:26:5: note: 'powl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodl' [-Wincompatible-library-redeclaration]\r\n int fmodl (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modfl' [-Wincompatible-library-redeclaration]\r\n int modfl (void);\r\n ^\r\n _configtest.c:28:5: note: 'modfl' is a builtin with type 'long double (long double, long double *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpl' [-Wincompatible-library-redeclaration]\r\n int frexpl (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpl' is a builtin with type 'long double (long double, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpl' [-Wincompatible-library-redeclaration]\r\n int ldexpl (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpl' is a builtin with type 'long double (long double, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2l' [-Wincompatible-library-redeclaration]\r\n int exp2l (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2l' [-Wincompatible-library-redeclaration]\r\n int log2l (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignl' [-Wincompatible-library-redeclaration]\r\n int copysignl (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterl' [-Wincompatible-library-redeclaration]\r\n int nextafterl (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtl' [-Wincompatible-library-redeclaration]\r\n int cbrtl (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtl' is a builtin with type 'long double (long double)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:8:12: error: use of undeclared identifier 'HAVE_DECL_SIGNBIT'\r\n (void) HAVE_DECL_SIGNBIT;\r\n ^\r\n 1 error generated.\r\n failure.\r\n removing: _configtest.c _configtest.o\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabs' [-Wincompatible-library-redeclaration]\r\n int cabs (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabs' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacos' [-Wincompatible-library-redeclaration]\r\n int cacos (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacosh' [-Wincompatible-library-redeclaration]\r\n int cacosh (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'carg' [-Wincompatible-library-redeclaration]\r\n int carg (void);\r\n ^\r\n _configtest.c:4:5: note: 'carg' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casin' [-Wincompatible-library-redeclaration]\r\n int casin (void);\r\n ^\r\n _configtest.c:5:5: note: 'casin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinh' [-Wincompatible-library-redeclaration]\r\n int casinh (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catan' [-Wincompatible-library-redeclaration]\r\n int catan (void);\r\n ^\r\n _configtest.c:7:5: note: 'catan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanh' [-Wincompatible-library-redeclaration]\r\n int catanh (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccos' [-Wincompatible-library-redeclaration]\r\n int ccos (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccosh' [-Wincompatible-library-redeclaration]\r\n int ccosh (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexp' [-Wincompatible-library-redeclaration]\r\n int cexp (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexp' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimag' [-Wincompatible-library-redeclaration]\r\n int cimag (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimag' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clog' [-Wincompatible-library-redeclaration]\r\n int clog (void);\r\n ^\r\n _configtest.c:13:5: note: 'clog' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conj' [-Wincompatible-library-redeclaration]\r\n int conj (void);\r\n ^\r\n _configtest.c:14:5: note: 'conj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpow' [-Wincompatible-library-redeclaration]\r\n int cpow (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpow' is a builtin with type '_Complex double (_Complex double, _Complex double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cproj' [-Wincompatible-library-redeclaration]\r\n int cproj (void);\r\n ^\r\n _configtest.c:16:5: note: 'cproj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creal' [-Wincompatible-library-redeclaration]\r\n int creal (void);\r\n ^\r\n _configtest.c:17:5: note: 'creal' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csin' [-Wincompatible-library-redeclaration]\r\n int csin (void);\r\n ^\r\n _configtest.c:18:5: note: 'csin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinh' [-Wincompatible-library-redeclaration]\r\n int csinh (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrt' [-Wincompatible-library-redeclaration]\r\n int csqrt (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrt' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctan' [-Wincompatible-library-redeclaration]\r\n int ctan (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanh' [-Wincompatible-library-redeclaration]\r\n int ctanh (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanh' is a builtin with type '_Complex double (_Complex double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsf' [-Wincompatible-library-redeclaration]\r\n int cabsf (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosf' [-Wincompatible-library-redeclaration]\r\n int cacosf (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshf' [-Wincompatible-library-redeclaration]\r\n int cacoshf (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargf' [-Wincompatible-library-redeclaration]\r\n int cargf (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinf' [-Wincompatible-library-redeclaration]\r\n int casinf (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhf' [-Wincompatible-library-redeclaration]\r\n int casinhf (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanf' [-Wincompatible-library-redeclaration]\r\n int catanf (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhf' [-Wincompatible-library-redeclaration]\r\n int catanhf (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosf' [-Wincompatible-library-redeclaration]\r\n int ccosf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshf' [-Wincompatible-library-redeclaration]\r\n int ccoshf (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpf' [-Wincompatible-library-redeclaration]\r\n int cexpf (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagf' [-Wincompatible-library-redeclaration]\r\n int cimagf (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogf' [-Wincompatible-library-redeclaration]\r\n int clogf (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjf' [-Wincompatible-library-redeclaration]\r\n int conjf (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowf' [-Wincompatible-library-redeclaration]\r\n int cpowf (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowf' is a builtin with type '_Complex float (_Complex float, _Complex float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojf' [-Wincompatible-library-redeclaration]\r\n int cprojf (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'crealf' [-Wincompatible-library-redeclaration]\r\n int crealf (void);\r\n ^\r\n _configtest.c:17:5: note: 'crealf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinf' [-Wincompatible-library-redeclaration]\r\n int csinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhf' [-Wincompatible-library-redeclaration]\r\n int csinhf (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtf' [-Wincompatible-library-redeclaration]\r\n int csqrtf (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanf' [-Wincompatible-library-redeclaration]\r\n int ctanf (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhf' [-Wincompatible-library-redeclaration]\r\n int ctanhf (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhf' is a builtin with type '_Complex float (_Complex float)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsl' [-Wincompatible-library-redeclaration]\r\n int cabsl (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosl' [-Wincompatible-library-redeclaration]\r\n int cacosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshl' [-Wincompatible-library-redeclaration]\r\n int cacoshl (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargl' [-Wincompatible-library-redeclaration]\r\n int cargl (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinl' [-Wincompatible-library-redeclaration]\r\n int casinl (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhl' [-Wincompatible-library-redeclaration]\r\n int casinhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanl' [-Wincompatible-library-redeclaration]\r\n int catanl (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhl' [-Wincompatible-library-redeclaration]\r\n int catanhl (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosl' [-Wincompatible-library-redeclaration]\r\n int ccosl (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshl' [-Wincompatible-library-redeclaration]\r\n int ccoshl (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpl' [-Wincompatible-library-redeclaration]\r\n int cexpl (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagl' [-Wincompatible-library-redeclaration]\r\n int cimagl (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogl' [-Wincompatible-library-redeclaration]\r\n int clogl (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjl' [-Wincompatible-library-redeclaration]\r\n int conjl (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowl' [-Wincompatible-library-redeclaration]\r\n int cpowl (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowl' is a builtin with type '_Complex long double (_Complex long double, _Complex long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojl' [-Wincompatible-library-redeclaration]\r\n int cprojl (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creall' [-Wincompatible-library-redeclaration]\r\n int creall (void);\r\n ^\r\n _configtest.c:17:5: note: 'creall' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinl' [-Wincompatible-library-redeclaration]\r\n int csinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhl' [-Wincompatible-library-redeclaration]\r\n int csinhl (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtl' [-Wincompatible-library-redeclaration]\r\n int csqrtl (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanl' [-Wincompatible-library-redeclaration]\r\n int ctanl (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhl' [-Wincompatible-library-redeclaration]\r\n int ctanhl (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:2:12: warning: unused function 'static_func' [-Wunused-function]\r\n static int static_func (char * restrict a)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:3:19: warning: unused function 'static_func' [-Wunused-function]\r\n static inline int static_func (void)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h\r\n #define SIZEOF_PY_INTPTR_T 8\r\n #define SIZEOF_OFF_T 8\r\n #define SIZEOF_PY_LONG_LONG 8\r\n #define MATHLIB\r\n #define HAVE_SIN 1\r\n #define HAVE_COS 1\r\n #define HAVE_TAN 1\r\n #define HAVE_SINH 1\r\n #define HAVE_COSH 1\r\n #define HAVE_TANH 1\r\n #define HAVE_FABS 1\r\n #define HAVE_FLOOR 1\r\n #define HAVE_CEIL 1\r\n #define HAVE_SQRT 1\r\n #define HAVE_LOG10 1\r\n #define HAVE_LOG 1\r\n #define HAVE_EXP 1\r\n #define HAVE_ASIN 1\r\n #define HAVE_ACOS 1\r\n #define HAVE_ATAN 1\r\n #define HAVE_FMOD 1\r\n #define HAVE_MODF 1\r\n #define HAVE_FREXP 1\r\n #define HAVE_LDEXP 1\r\n #define HAVE_RINT 1\r\n #define HAVE_TRUNC 1\r\n #define HAVE_EXP2 1\r\n #define HAVE_LOG2 1\r\n #define HAVE_ATAN2 1\r\n #define HAVE_POW 1\r\n #define HAVE_NEXTAFTER 1\r\n #define HAVE_STRTOLL 1\r\n #define HAVE_STRTOULL 1\r\n #define HAVE_CBRT 1\r\n #define HAVE_STRTOLD_L 1\r\n #define HAVE_BACKTRACE 1\r\n #define HAVE_MADVISE 1\r\n #define HAVE_XMMINTRIN_H 1\r\n #define HAVE_EMMINTRIN_H 1\r\n #define HAVE_XLOCALE_H 1\r\n #define HAVE_DLFCN_H 1\r\n #define HAVE_SYS_MMAN_H 1\r\n #define HAVE___BUILTIN_ISNAN 1\r\n #define HAVE___BUILTIN_ISINF 1\r\n #define HAVE___BUILTIN_ISFINITE 1\r\n #define HAVE___BUILTIN_BSWAP32 1\r\n #define HAVE___BUILTIN_BSWAP64 1\r\n #define HAVE___BUILTIN_EXPECT 1\r\n #define HAVE___BUILTIN_MUL_OVERFLOW 1\r\n #define HAVE___BUILTIN_CPU_SUPPORTS 1\r\n #define HAVE__M_FROM_INT64 1\r\n #define HAVE__MM_LOAD_PS 1\r\n #define HAVE__MM_PREFETCH 1\r\n #define HAVE__MM_LOAD_PD 1\r\n #define HAVE___BUILTIN_PREFETCH 1\r\n #define HAVE_LINK_AVX 1\r\n #define HAVE_LINK_AVX2 1\r\n #define HAVE_XGETBV 1\r\n #define HAVE_ATTRIBUTE_NONNULL 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX2 1\r\n #define HAVE___THREAD 1\r\n #define HAVE_SINF 1\r\n #define HAVE_COSF 1\r\n #define HAVE_TANF 1\r\n #define HAVE_SINHF 1\r\n #define HAVE_COSHF 1\r\n #define HAVE_TANHF 1\r\n #define HAVE_FABSF 1\r\n #define HAVE_FLOORF 1\r\n #define HAVE_CEILF 1\r\n #define HAVE_RINTF 1\r\n #define HAVE_TRUNCF 1\r\n #define HAVE_SQRTF 1\r\n #define HAVE_LOG10F 1\r\n #define HAVE_LOGF 1\r\n #define HAVE_LOG1PF 1\r\n #define HAVE_EXPF 1\r\n #define HAVE_EXPM1F 1\r\n #define HAVE_ASINF 1\r\n #define HAVE_ACOSF 1\r\n #define HAVE_ATANF 1\r\n #define HAVE_ASINHF 1\r\n #define HAVE_ACOSHF 1\r\n #define HAVE_ATANHF 1\r\n #define HAVE_HYPOTF 1\r\n #define HAVE_ATAN2F 1\r\n #define HAVE_POWF 1\r\n #define HAVE_FMODF 1\r\n #define HAVE_MODFF 1\r\n #define HAVE_FREXPF 1\r\n #define HAVE_LDEXPF 1\r\n #define HAVE_EXP2F 1\r\n #define HAVE_LOG2F 1\r\n #define HAVE_COPYSIGNF 1\r\n #define HAVE_NEXTAFTERF 1\r\n #define HAVE_CBRTF 1\r\n #define HAVE_SINL 1\r\n #define HAVE_COSL 1\r\n #define HAVE_TANL 1\r\n #define HAVE_SINHL 1\r\n #define HAVE_COSHL 1\r\n #define HAVE_TANHL 1\r\n #define HAVE_FABSL 1\r\n #define HAVE_FLOORL 1\r\n #define HAVE_CEILL 1\r\n #define HAVE_RINTL 1\r\n #define HAVE_TRUNCL 1\r\n #define HAVE_SQRTL 1\r\n #define HAVE_LOG10L 1\r\n #define HAVE_LOGL 1\r\n #define HAVE_LOG1PL 1\r\n #define HAVE_EXPL 1\r\n #define HAVE_EXPM1L 1\r\n #define HAVE_ASINL 1\r\n #define HAVE_ACOSL 1\r\n #define HAVE_ATANL 1\r\n #define HAVE_ASINHL 1\r\n #define HAVE_ACOSHL 1\r\n #define HAVE_ATANHL 1\r\n #define HAVE_HYPOTL 1\r\n #define HAVE_ATAN2L 1\r\n #define HAVE_POWL 1\r\n #define HAVE_FMODL 1\r\n #define HAVE_MODFL 1\r\n #define HAVE_FREXPL 1\r\n #define HAVE_LDEXPL 1\r\n #define HAVE_EXP2L 1\r\n #define HAVE_LOG2L 1\r\n #define HAVE_COPYSIGNL 1\r\n #define HAVE_NEXTAFTERL 1\r\n #define HAVE_CBRTL 1\r\n #define HAVE_DECL_SIGNBIT\r\n #define HAVE_COMPLEX_H 1\r\n #define HAVE_CABS 1\r\n #define HAVE_CACOS 1\r\n #define HAVE_CACOSH 1\r\n #define HAVE_CARG 1\r\n #define HAVE_CASIN 1\r\n #define HAVE_CASINH 1\r\n #define HAVE_CATAN 1\r\n #define HAVE_CATANH 1\r\n #define HAVE_CCOS 1\r\n #define HAVE_CCOSH 1\r\n #define HAVE_CEXP 1\r\n #define HAVE_CIMAG 1\r\n #define HAVE_CLOG 1\r\n #define HAVE_CONJ 1\r\n #define HAVE_CPOW 1\r\n #define HAVE_CPROJ 1\r\n #define HAVE_CREAL 1\r\n #define HAVE_CSIN 1\r\n #define HAVE_CSINH 1\r\n #define HAVE_CSQRT 1\r\n #define HAVE_CTAN 1\r\n #define HAVE_CTANH 1\r\n #define HAVE_CABSF 1\r\n #define HAVE_CACOSF 1\r\n #define HAVE_CACOSHF 1\r\n #define HAVE_CARGF 1\r\n #define HAVE_CASINF 1\r\n #define HAVE_CASINHF 1\r\n #define HAVE_CATANF 1\r\n #define HAVE_CATANHF 1\r\n #define HAVE_CCOSF 1\r\n #define HAVE_CCOSHF 1\r\n #define HAVE_CEXPF 1\r\n #define HAVE_CIMAGF 1\r\n #define HAVE_CLOGF 1\r\n #define HAVE_CONJF 1\r\n #define HAVE_CPOWF 1\r\n #define HAVE_CPROJF 1\r\n #define HAVE_CREALF 1\r\n #define HAVE_CSINF 1\r\n #define HAVE_CSINHF 1\r\n #define HAVE_CSQRTF 1\r\n #define HAVE_CTANF 1\r\n #define HAVE_CTANHF 1\r\n #define HAVE_CABSL 1\r\n #define HAVE_CACOSL 1\r\n #define HAVE_CACOSHL 1\r\n #define HAVE_CARGL 1\r\n #define HAVE_CASINL 1\r\n #define HAVE_CASINHL 1\r\n #define HAVE_CATANL 1\r\n #define HAVE_CATANHL 1\r\n #define HAVE_CCOSL 1\r\n #define HAVE_CCOSHL 1\r\n #define HAVE_CEXPL 1\r\n #define HAVE_CIMAGL 1\r\n #define HAVE_CLOGL 1\r\n #define HAVE_CONJL 1\r\n #define HAVE_CPOWL 1\r\n #define HAVE_CPROJL 1\r\n #define HAVE_CREALL 1\r\n #define HAVE_CSINL 1\r\n #define HAVE_CSINHL 1\r\n #define HAVE_CSQRTL 1\r\n #define HAVE_CTANL 1\r\n #define HAVE_CTANHL 1\r\n #define NPY_RESTRICT restrict\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define HAVE_LDOUBLE_INTEL_EXTENDED_16_BYTES_LE 1\r\n #define NPY_PY3K 1\r\n #ifndef __cplusplus\r\n \/* #undef inline *\/\r\n #endif\r\n \r\n #ifndef _NPY_NPY_CONFIG_H_\r\n #error config.h should never be included directly, include npy_config.h instead\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n Generating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'exp' [-Wincompatible-library-redeclaration]\r\n int exp (void);\r\n ^\r\n _configtest.c:1:5: note: 'exp' is a builtin with type 'double (double)'\r\n 1 warning generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n #define NPY_SIZEOF_SHORT SIZEOF_SHORT\r\n #define NPY_SIZEOF_INT SIZEOF_INT\r\n #define NPY_SIZEOF_LONG SIZEOF_LONG\r\n #define NPY_SIZEOF_FLOAT 4\r\n #define NPY_SIZEOF_COMPLEX_FLOAT 8\r\n #define NPY_SIZEOF_DOUBLE 8\r\n #define NPY_SIZEOF_COMPLEX_DOUBLE 16\r\n #define NPY_SIZEOF_LONGDOUBLE 16\r\n #define NPY_SIZEOF_COMPLEX_LONGDOUBLE 32\r\n #define NPY_SIZEOF_PY_INTPTR_T 8\r\n #define NPY_SIZEOF_OFF_T 8\r\n #define NPY_SIZEOF_PY_LONG_LONG 8\r\n #define NPY_SIZEOF_LONGLONG 8\r\n #define NPY_NO_SMP 0\r\n #define NPY_HAVE_DECL_ISNAN\r\n #define NPY_HAVE_DECL_ISINF\r\n #define NPY_HAVE_DECL_ISFINITE\r\n #define NPY_HAVE_DECL_SIGNBIT\r\n #define NPY_USE_C99_COMPLEX 1\r\n #define NPY_HAVE_COMPLEX_DOUBLE 1\r\n #define NPY_HAVE_COMPLEX_FLOAT 1\r\n #define NPY_HAVE_COMPLEX_LONG_DOUBLE 1\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define NPY_USE_C99_FORMATS 1\r\n #define NPY_VISIBILITY_HIDDEN __attribute__((visibility(\"hidden\")))\r\n #define NPY_ABI_VERSION 0x01000009\r\n #define NPY_API_VERSION 0x0000000D\r\n \r\n #ifndef __STDC_FORMAT_MACROS\r\n #define __STDC_FORMAT_MACROS 1\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h']\r\n building extension \"numpy.core._multiarray_tests\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n building extension \"numpy.core._multiarray_umath\" sources\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_ufunc_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common' to include_dirs.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_internal.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h']\r\n building extension \"numpy.core._umath_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n building extension \"numpy.core._rational_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n building extension \"numpy.core._struct_ufunc_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n building extension \"numpy.core._operand_flag_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n building extension \"numpy.fft.fftpack_lite\" sources\r\n building extension \"numpy.linalg.lapack_lite\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n building extension \"numpy.linalg._umath_linalg\" sources\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n building extension \"numpy.random.mtrand\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/random\r\n building data_files sources\r\n build_src: building npy-pkg config files\r\n running build_py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/conftest.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_globals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/dual.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_distributor_init.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/ctypeslib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/matlib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_pytesttester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/py3k.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/_inspect.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/fromnumeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_add_newdocs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_methods.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_internal.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_string_helpers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/multiarray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/records.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup_common.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_aliased_types.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/memmap.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/overrides.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/getlimits.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype_ctypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/defchararray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/machar.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/einsumfunc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath_tests.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numerictypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_type_aliases.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/cversions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/arrayprint.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/code_generators\/generate_numpy_api.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/unixccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/numpy_distribution.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/conv_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/cpuinfo.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvc9compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/npy_pkg_config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/compat.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/misc_util.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/log.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/line_endings.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/lib2def.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/pathccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/system_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/exec_command.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/from_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/mingw32ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/extension.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/intelccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/distutils\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config_compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_ext.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_headers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_py.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_src.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/sdist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_scripts.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/bdist_rpm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/autodist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/egg_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/develop.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_data.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/gnu.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/compaq.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/intel.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/none.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/nag.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/ibm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/sun.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/lahey.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/g95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/mips.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/hpux.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/environment.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pathf95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/absoft.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/vast.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/misc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/internals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/creation.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/constants.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/ufuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/broadcasting.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/basics.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/subclassing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/indexing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/byteswapping.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/structured_arrays.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/glossary.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/common_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/crackfortran.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cb_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py2e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/func2subr.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/diagnose.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/capi_maps.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f90mod_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py_testing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/use_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/auxfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__main__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/helper.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/fftpack.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_iotools.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/mixins.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/nanfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/recfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/histograms.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/scimath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/user_array.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/format.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/twodim_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/financial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/index_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/npyio.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/stride_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arrayterator.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraysetops.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraypad.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/type_check.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_datasource.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/ufunclike.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/linalg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/extras.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/testutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/bench.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/timer_comparison.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/mrecords.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/defmatrix.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/laguerre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/_polybase.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polyutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite_e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/chebyshev.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/legendre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/print_coercion_tables.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/parameterized.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n running build_clib\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_clib\r\n building 'npymath' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n ar: adding 4 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n building 'npysort' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/quicksort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/mergesort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/heapsort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/selection.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/binsearch.c\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n 22 warnings generated.\r\n ar: adding 5 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n running build_ext\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_ext\r\n building 'numpy.core._dummy' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/dummymodule.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/dummymodule.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_dummy.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_tests.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_umath' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/core\/src\/multiarray\/alloc.c\r\n clang: numpy\/core\/src\/multiarray\/calculation.cclang: numpy\/core\/src\/multiarray\/array_assign_scalar.c\r\n clang: numpy\/core\/src\/multiarray\/convert.c\r\n \r\n clang: numpy\/core\/src\/multiarray\/ctors.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busday.c\r\n clang: numpy\/core\/src\/multiarray\/dragon4.cclang: numpy\/core\/src\/multiarray\/flagsobject.c\r\n \r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/arrayobject.c\r\n clang: numpy\/core\/src\/multiarray\/array_assign_array.c\r\n clang: numpy\/core\/src\/multiarray\/convert_datatype.c\r\n clang: numpy\/core\/src\/multiarray\/getset.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busdaycal.c\r\n clang: numpy\/core\/src\/multiarray\/buffer.c\r\n clang: numpy\/core\/src\/multiarray\/compiled_base.c\r\n clang: numpy\/core\/src\/multiarray\/hashdescr.c\r\n clang: numpy\/core\/src\/multiarray\/descriptor.c\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/conversion_utils.c\r\n clang: numpy\/core\/src\/multiarray\/item_selection.c\r\n clang: numpy\/core\/src\/multiarray\/dtype_transfer.c\r\n clang: numpy\/core\/src\/multiarray\/mapping.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n 3 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/datetime.c\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:477:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ptr = PyUnicode_AS_UNICODE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/common.c\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 6 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/nditer_pywrap.c\r\n 9 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/sequence.c\r\n clang: numpy\/core\/src\/multiarray\/shape.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n clang: numpy\/core\/src\/multiarray\/methods.c\r\n clang: numpy\/core\/src\/multiarray\/iterators.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_strings.c\r\n clang: numpy\/core\/src\/multiarray\/number.c\r\n clang: numpy\/core\/src\/multiarray\/scalarapi.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n numpy\/core\/src\/multiarray\/scalarapi.c:74:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:135:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1849:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buffer = PyUnicode_AS_DATA(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 5 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/typeinfo.c\r\n clang: numpy\/core\/src\/multiarray\/refcount.c\r\n clang: numpy\/core\/src\/multiarray\/usertypes.c\r\n clang: numpy\/core\/src\/multiarray\/multiarraymodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n clang: numpy\/core\/src\/multiarray\/vdot.c\r\n clang: numpy\/core\/src\/umath\/umathmodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n clang: numpy\/core\/src\/umath\/reduction.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n clang: numpy\/core\/src\/multiarray\/nditer_api.c\r\n 14 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/strfuncs.c\r\n numpy\/core\/src\/umath\/loops.c.src:655:18: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n result = PyEval_CallObject(tocall, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:178:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_ReprFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:195:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_StrFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 2 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/temp_elide.c\r\n clang: numpy\/core\/src\/umath\/cpuid.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n clang: numpy\/core\/src\/umath\/ufunc_object.c\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'byte_long' [-Wunused-function]\r\n byte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ubyte_long' [-Wunused-function]\r\n ubyte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'short_long' [-Wunused-function]\r\n short_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ushort_long' [-Wunused-function]\r\n ushort_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'int_long' [-Wunused-function]\r\n int_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'uint_long' [-Wunused-function]\r\n uint_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'long_long' [-Wunused-function]\r\n long_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulong_long' [-Wunused-function]\r\n ulong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longlong_long' [-Wunused-function]\r\n longlong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulonglong_long' [-Wunused-function]\r\n ulonglong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'half_long' [-Wunused-function]\r\n half_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'float_long' [-Wunused-function]\r\n float_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'double_long' [-Wunused-function]\r\n double_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longdouble_long' [-Wunused-function]\r\n longdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cfloat_long' [-Wunused-function]\r\n cfloat_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cdouble_long' [-Wunused-function]\r\n cdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'clongdouble_long' [-Wunused-function]\r\n clongdouble_long(PyObject *obj)\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/nditer_constr.c\r\n numpy\/core\/src\/umath\/ufunc_object.c:657:19: warning: comparison of integers of different signs: 'int' and 'size_t' (aka 'unsigned long') [-Wsign-compare]\r\n for (i = 0; i < len; i++) {\r\n ~ ^ ~~~\r\n clang: numpy\/core\/src\/umath\/override.c\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n clang: numpy\/core\/src\/common\/array_assign.c\r\n clang: numpy\/core\/src\/common\/ufunc_override.c\r\n clang: numpy\/core\/src\/common\/npy_longdouble.c\r\n clang: numpy\/core\/src\/common\/numpyos.c\r\n clang: numpy\/core\/src\/common\/ucsnarrow.c\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/umath\/extobj.c\r\n numpy\/core\/src\/common\/ucsnarrow.c:139:34: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n ret = (PyUnicodeObject *)PyUnicode_FromUnicode((Py_UNICODE*)buf,\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/common\/python_xerbla.c\r\n clang: numpy\/core\/src\/common\/cblasfuncs.c\r\n clang: \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c\r\n In file included from \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c:26:\r\n In file included from numpy\/core\/include\/numpy\/arrayobject.h:4:\r\n In file included from numpy\/core\/include\/numpy\/ndarrayobject.h:21:\r\n build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h:1463:1: warning: unused function '_import_array' [-Wunused-function]\r\n _import_array(void)\r\n ^\r\n 1 warning generated.\r\n 17 warnings generated.\r\n clang: numpy\/core\/src\/umath\/ufunc_type_resolution.c\r\n 4 warnings generated.\r\n 4 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/alloc.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arrayobject.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_scalar.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_array.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/buffer.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/calculation.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/compiled_base.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/common.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert_datatype.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/conversion_utils.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/ctors.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_strings.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busday.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busdaycal.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/descriptor.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dragon4.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dtype_transfer.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/flagsobject.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/getset.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/hashdescr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/item_selection.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/iterators.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/mapping.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/methods.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/multiarraymodule.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_api.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_constr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_pywrap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/number.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/refcount.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/sequence.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/shape.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalarapi.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/strfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/temp_elide.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/typeinfo.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/usertypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/vdot.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/umathmodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/reduction.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_object.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/extobj.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/cpuid.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_type_resolution.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/halffloat.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/array_assign.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/npy_longdouble.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ucsnarrow.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ufunc_override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/numpyos.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/cblasfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/python_xerbla.o build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -lnpysort -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_umath.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.core._umath_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_umath_tests.cpython-39-darwin.so\r\n building 'numpy.core._rational_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_rational_tests.cpython-39-darwin.so\r\n building 'numpy.core._struct_ufunc_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_struct_ufunc_tests.cpython-39-darwin.so\r\n building 'numpy.core._operand_flag_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_operand_flag_tests.cpython-39-darwin.so\r\n building 'numpy.fft.fftpack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n compile options: '-Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/fft\/fftpack_litemodule.c\r\n clang: numpy\/fft\/fftpack.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_lite.cpython-39-darwin.so\r\n building 'numpy.linalg.lapack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/linalg\/lapack_litemodule.c\r\n clang: numpy\/linalg\/lapack_lite\/python_xerbla.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.linalg._umath_linalg' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n numpy\/linalg\/umath_linalg.c.src:735:32: warning: unknown warning group '-Wmaybe-uninitialized', ignored [-Wunknown-warning-option]\r\n #pragma GCC diagnostic ignored \"-Wmaybe-uninitialized\"\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:541:1: warning: unused function 'dump_ufunc_object' [-Wunused-function]\r\n dump_ufunc_object(PyUFuncObject* ufunc)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:566:1: warning: unused function 'dump_linearize_data' [-Wunused-function]\r\n dump_linearize_data(const char* name, const LINEARIZE_DATA_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_FLOAT_matrix' [-Wunused-function]\r\n dump_FLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_DOUBLE_matrix' [-Wunused-function]\r\n dump_DOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CFLOAT_matrix' [-Wunused-function]\r\n dump_CFLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CDOUBLE_matrix' [-Wunused-function]\r\n dump_CDOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_FLOAT_matrix' [-Wunused-function]\r\n zero_FLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_DOUBLE_matrix' [-Wunused-function]\r\n zero_DOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CFLOAT_matrix' [-Wunused-function]\r\n zero_CFLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CDOUBLE_matrix' [-Wunused-function]\r\n zero_CDOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:1862:1: warning: unused function 'dump_geev_params' [-Wunused-function]\r\n dump_geev_params(const char *name, GEEV_PARAMS_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2132:1: warning: unused function 'init_cgeev' [-Wunused-function]\r\n init_cgeev(GEEV_PARAMS_t* params,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2213:1: warning: unused function 'process_cgeev_results' [-Wunused-function]\r\n process_cgeev_results(GEEV_PARAMS_t *NPY_UNUSED(params))\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2376:1: warning: unused function 'dump_gesdd_params' [-Wunused-function]\r\n dump_gesdd_params(const char *name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2864:1: warning: unused function 'dump_gelsd_params' [-Wunused-function]\r\n dump_gelsd_params(const char *name,\r\n ^\r\n 16 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/_umath_linalg.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.random.mtrand' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\r\n compile options: '-D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/random\/mtrand\/mtrand.c\r\n clang: numpy\/random\/mtrand\/initarray.cclang: numpy\/random\/mtrand\/randomkit.c\r\n \r\n clang: numpy\/random\/mtrand\/distributions.c\r\n numpy\/random\/mtrand\/mtrand.c:40400:34: error: no member named 'tp_print' in 'struct _typeobject'\r\n __pyx_type_6mtrand_RandomState.tp_print = 0;\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 12 warnings and 1 error generated.\r\n error: Command \"clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c numpy\/random\/mtrand\/mtrand.c -o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o -MMD -MF build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o.d\" failed with exit status 1","comment_length":2,"text":"Unable to install datasets\n** Edit **\r\nI believe there's a bug with the package when you're installing it with Python 3.9. I recommend sticking with previous versions. Thanks, @thomwolf for the insight! \r\n\r\n**Short description**\r\n\r\nI followed the instructions for installing datasets (https:\/\/huggingface.co\/docs\/datasets\/installation.html). However, while I tried to download datasets using `pip install datasets` I got a massive error message after getting stuck at \"Installing build dependencies...\" \r\n\r\nI was wondering if this problem can be fixed by creating a virtual environment, but it didn't help. Can anyone offer some advice on how to fix this issue? \r\n\r\nHere's an error message: \r\n\r\n`(env) Gas-MacBook-Pro:Downloads destiny$ pip install datasets\r\nCollecting datasets\r\n Using cached datasets-1.2.0-py3-none-any.whl (159 kB)\r\nCollecting numpy>=1.17\r\n Using cached numpy-1.19.5-cp39-cp39-macosx_10_9_x86_64.whl (15.6 MB)\r\nCollecting pyarrow>=0.17.1\r\n Using cached pyarrow-2.0.0.tar.gz (58.9 MB)\r\n....\r\n\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceilf' [-Wincompatible-library-redeclaration]\r\n int ceilf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceilf' is a builtin with type 'float (float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintf' [-Wincompatible-library-redeclaration]\r\n int rintf (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintf' is a builtin with type 'float (float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncf' [-Wincompatible-library-redeclaration]\r\n int truncf (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncf' is a builtin with type 'float (float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtf' [-Wincompatible-library-redeclaration]\r\n int sqrtf (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtf' is a builtin with type 'float (float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10f' [-Wincompatible-library-redeclaration]\r\n int log10f (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10f' is a builtin with type 'float (float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logf' [-Wincompatible-library-redeclaration]\r\n int logf (void);\r\n ^\r\n _configtest.c:14:5: note: 'logf' is a builtin with type 'float (float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pf' [-Wincompatible-library-redeclaration]\r\n int log1pf (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pf' is a builtin with type 'float (float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expf' [-Wincompatible-library-redeclaration]\r\n int expf (void);\r\n ^\r\n _configtest.c:16:5: note: 'expf' is a builtin with type 'float (float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1f' [-Wincompatible-library-redeclaration]\r\n int expm1f (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1f' is a builtin with type 'float (float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinf' [-Wincompatible-library-redeclaration]\r\n int asinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinf' is a builtin with type 'float (float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosf' [-Wincompatible-library-redeclaration]\r\n int acosf (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosf' is a builtin with type 'float (float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanf' [-Wincompatible-library-redeclaration]\r\n int atanf (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanf' is a builtin with type 'float (float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhf' [-Wincompatible-library-redeclaration]\r\n int asinhf (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhf' is a builtin with type 'float (float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshf' [-Wincompatible-library-redeclaration]\r\n int acoshf (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshf' is a builtin with type 'float (float)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhf' [-Wincompatible-library-redeclaration]\r\n int atanhf (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhf' is a builtin with type 'float (float)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotf' [-Wincompatible-library-redeclaration]\r\n int hypotf (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotf' is a builtin with type 'float (float, float)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2f' [-Wincompatible-library-redeclaration]\r\n int atan2f (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2f' is a builtin with type 'float (float, float)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powf' [-Wincompatible-library-redeclaration]\r\n int powf (void);\r\n ^\r\n _configtest.c:26:5: note: 'powf' is a builtin with type 'float (float, float)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodf' [-Wincompatible-library-redeclaration]\r\n int fmodf (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodf' is a builtin with type 'float (float, float)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modff' [-Wincompatible-library-redeclaration]\r\n int modff (void);\r\n ^\r\n _configtest.c:28:5: note: 'modff' is a builtin with type 'float (float, float *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpf' [-Wincompatible-library-redeclaration]\r\n int frexpf (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpf' is a builtin with type 'float (float, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpf' [-Wincompatible-library-redeclaration]\r\n int ldexpf (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpf' is a builtin with type 'float (float, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2f' [-Wincompatible-library-redeclaration]\r\n int exp2f (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2f' is a builtin with type 'float (float)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2f' [-Wincompatible-library-redeclaration]\r\n int log2f (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2f' is a builtin with type 'float (float)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignf' [-Wincompatible-library-redeclaration]\r\n int copysignf (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignf' is a builtin with type 'float (float, float)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterf' [-Wincompatible-library-redeclaration]\r\n int nextafterf (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterf' is a builtin with type 'float (float, float)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtf' [-Wincompatible-library-redeclaration]\r\n int cbrtf (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtf' is a builtin with type 'float (float)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'sinl' [-Wincompatible-library-redeclaration]\r\n int sinl (void);\r\n ^\r\n _configtest.c:1:5: note: 'sinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cosl' [-Wincompatible-library-redeclaration]\r\n int cosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'tanl' [-Wincompatible-library-redeclaration]\r\n int tanl (void);\r\n ^\r\n _configtest.c:3:5: note: 'tanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'sinhl' [-Wincompatible-library-redeclaration]\r\n int sinhl (void);\r\n ^\r\n _configtest.c:4:5: note: 'sinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'coshl' [-Wincompatible-library-redeclaration]\r\n int coshl (void);\r\n ^\r\n _configtest.c:5:5: note: 'coshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'tanhl' [-Wincompatible-library-redeclaration]\r\n int tanhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'tanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'fabsl' [-Wincompatible-library-redeclaration]\r\n int fabsl (void);\r\n ^\r\n _configtest.c:7:5: note: 'fabsl' is a builtin with type 'long double (long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'floorl' [-Wincompatible-library-redeclaration]\r\n int floorl (void);\r\n ^\r\n _configtest.c:8:5: note: 'floorl' is a builtin with type 'long double (long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ceill' [-Wincompatible-library-redeclaration]\r\n int ceill (void);\r\n ^\r\n _configtest.c:9:5: note: 'ceill' is a builtin with type 'long double (long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'rintl' [-Wincompatible-library-redeclaration]\r\n int rintl (void);\r\n ^\r\n _configtest.c:10:5: note: 'rintl' is a builtin with type 'long double (long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'truncl' [-Wincompatible-library-redeclaration]\r\n int truncl (void);\r\n ^\r\n _configtest.c:11:5: note: 'truncl' is a builtin with type 'long double (long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'sqrtl' [-Wincompatible-library-redeclaration]\r\n int sqrtl (void);\r\n ^\r\n _configtest.c:12:5: note: 'sqrtl' is a builtin with type 'long double (long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'log10l' [-Wincompatible-library-redeclaration]\r\n int log10l (void);\r\n ^\r\n _configtest.c:13:5: note: 'log10l' is a builtin with type 'long double (long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'logl' [-Wincompatible-library-redeclaration]\r\n int logl (void);\r\n ^\r\n _configtest.c:14:5: note: 'logl' is a builtin with type 'long double (long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'log1pl' [-Wincompatible-library-redeclaration]\r\n int log1pl (void);\r\n ^\r\n _configtest.c:15:5: note: 'log1pl' is a builtin with type 'long double (long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'expl' [-Wincompatible-library-redeclaration]\r\n int expl (void);\r\n ^\r\n _configtest.c:16:5: note: 'expl' is a builtin with type 'long double (long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'expm1l' [-Wincompatible-library-redeclaration]\r\n int expm1l (void);\r\n ^\r\n _configtest.c:17:5: note: 'expm1l' is a builtin with type 'long double (long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'asinl' [-Wincompatible-library-redeclaration]\r\n int asinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'asinl' is a builtin with type 'long double (long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'acosl' [-Wincompatible-library-redeclaration]\r\n int acosl (void);\r\n ^\r\n _configtest.c:19:5: note: 'acosl' is a builtin with type 'long double (long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'atanl' [-Wincompatible-library-redeclaration]\r\n int atanl (void);\r\n ^\r\n _configtest.c:20:5: note: 'atanl' is a builtin with type 'long double (long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'asinhl' [-Wincompatible-library-redeclaration]\r\n int asinhl (void);\r\n ^\r\n _configtest.c:21:5: note: 'asinhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'acoshl' [-Wincompatible-library-redeclaration]\r\n int acoshl (void);\r\n ^\r\n _configtest.c:22:5: note: 'acoshl' is a builtin with type 'long double (long double)'\r\n _configtest.c:23:5: warning: incompatible redeclaration of library function 'atanhl' [-Wincompatible-library-redeclaration]\r\n int atanhl (void);\r\n ^\r\n _configtest.c:23:5: note: 'atanhl' is a builtin with type 'long double (long double)'\r\n _configtest.c:24:5: warning: incompatible redeclaration of library function 'hypotl' [-Wincompatible-library-redeclaration]\r\n int hypotl (void);\r\n ^\r\n _configtest.c:24:5: note: 'hypotl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:25:5: warning: incompatible redeclaration of library function 'atan2l' [-Wincompatible-library-redeclaration]\r\n int atan2l (void);\r\n ^\r\n _configtest.c:25:5: note: 'atan2l' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:26:5: warning: incompatible redeclaration of library function 'powl' [-Wincompatible-library-redeclaration]\r\n int powl (void);\r\n ^\r\n _configtest.c:26:5: note: 'powl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:27:5: warning: incompatible redeclaration of library function 'fmodl' [-Wincompatible-library-redeclaration]\r\n int fmodl (void);\r\n ^\r\n _configtest.c:27:5: note: 'fmodl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:28:5: warning: incompatible redeclaration of library function 'modfl' [-Wincompatible-library-redeclaration]\r\n int modfl (void);\r\n ^\r\n _configtest.c:28:5: note: 'modfl' is a builtin with type 'long double (long double, long double *)'\r\n _configtest.c:29:5: warning: incompatible redeclaration of library function 'frexpl' [-Wincompatible-library-redeclaration]\r\n int frexpl (void);\r\n ^\r\n _configtest.c:29:5: note: 'frexpl' is a builtin with type 'long double (long double, int *)'\r\n _configtest.c:30:5: warning: incompatible redeclaration of library function 'ldexpl' [-Wincompatible-library-redeclaration]\r\n int ldexpl (void);\r\n ^\r\n _configtest.c:30:5: note: 'ldexpl' is a builtin with type 'long double (long double, int)'\r\n _configtest.c:31:5: warning: incompatible redeclaration of library function 'exp2l' [-Wincompatible-library-redeclaration]\r\n int exp2l (void);\r\n ^\r\n _configtest.c:31:5: note: 'exp2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:32:5: warning: incompatible redeclaration of library function 'log2l' [-Wincompatible-library-redeclaration]\r\n int log2l (void);\r\n ^\r\n _configtest.c:32:5: note: 'log2l' is a builtin with type 'long double (long double)'\r\n _configtest.c:33:5: warning: incompatible redeclaration of library function 'copysignl' [-Wincompatible-library-redeclaration]\r\n int copysignl (void);\r\n ^\r\n _configtest.c:33:5: note: 'copysignl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:34:5: warning: incompatible redeclaration of library function 'nextafterl' [-Wincompatible-library-redeclaration]\r\n int nextafterl (void);\r\n ^\r\n _configtest.c:34:5: note: 'nextafterl' is a builtin with type 'long double (long double, long double)'\r\n _configtest.c:35:5: warning: incompatible redeclaration of library function 'cbrtl' [-Wincompatible-library-redeclaration]\r\n int cbrtl (void);\r\n ^\r\n _configtest.c:35:5: note: 'cbrtl' is a builtin with type 'long double (long double)'\r\n 35 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:8:12: error: use of undeclared identifier 'HAVE_DECL_SIGNBIT'\r\n (void) HAVE_DECL_SIGNBIT;\r\n ^\r\n 1 error generated.\r\n failure.\r\n removing: _configtest.c _configtest.o\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabs' [-Wincompatible-library-redeclaration]\r\n int cabs (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabs' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacos' [-Wincompatible-library-redeclaration]\r\n int cacos (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacosh' [-Wincompatible-library-redeclaration]\r\n int cacosh (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'carg' [-Wincompatible-library-redeclaration]\r\n int carg (void);\r\n ^\r\n _configtest.c:4:5: note: 'carg' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casin' [-Wincompatible-library-redeclaration]\r\n int casin (void);\r\n ^\r\n _configtest.c:5:5: note: 'casin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinh' [-Wincompatible-library-redeclaration]\r\n int casinh (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catan' [-Wincompatible-library-redeclaration]\r\n int catan (void);\r\n ^\r\n _configtest.c:7:5: note: 'catan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanh' [-Wincompatible-library-redeclaration]\r\n int catanh (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccos' [-Wincompatible-library-redeclaration]\r\n int ccos (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccos' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccosh' [-Wincompatible-library-redeclaration]\r\n int ccosh (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccosh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexp' [-Wincompatible-library-redeclaration]\r\n int cexp (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexp' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimag' [-Wincompatible-library-redeclaration]\r\n int cimag (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimag' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clog' [-Wincompatible-library-redeclaration]\r\n int clog (void);\r\n ^\r\n _configtest.c:13:5: note: 'clog' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conj' [-Wincompatible-library-redeclaration]\r\n int conj (void);\r\n ^\r\n _configtest.c:14:5: note: 'conj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpow' [-Wincompatible-library-redeclaration]\r\n int cpow (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpow' is a builtin with type '_Complex double (_Complex double, _Complex double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cproj' [-Wincompatible-library-redeclaration]\r\n int cproj (void);\r\n ^\r\n _configtest.c:16:5: note: 'cproj' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creal' [-Wincompatible-library-redeclaration]\r\n int creal (void);\r\n ^\r\n _configtest.c:17:5: note: 'creal' is a builtin with type 'double (_Complex double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csin' [-Wincompatible-library-redeclaration]\r\n int csin (void);\r\n ^\r\n _configtest.c:18:5: note: 'csin' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinh' [-Wincompatible-library-redeclaration]\r\n int csinh (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinh' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrt' [-Wincompatible-library-redeclaration]\r\n int csqrt (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrt' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctan' [-Wincompatible-library-redeclaration]\r\n int ctan (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctan' is a builtin with type '_Complex double (_Complex double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanh' [-Wincompatible-library-redeclaration]\r\n int ctanh (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanh' is a builtin with type '_Complex double (_Complex double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsf' [-Wincompatible-library-redeclaration]\r\n int cabsf (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosf' [-Wincompatible-library-redeclaration]\r\n int cacosf (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshf' [-Wincompatible-library-redeclaration]\r\n int cacoshf (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargf' [-Wincompatible-library-redeclaration]\r\n int cargf (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinf' [-Wincompatible-library-redeclaration]\r\n int casinf (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhf' [-Wincompatible-library-redeclaration]\r\n int casinhf (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanf' [-Wincompatible-library-redeclaration]\r\n int catanf (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhf' [-Wincompatible-library-redeclaration]\r\n int catanhf (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosf' [-Wincompatible-library-redeclaration]\r\n int ccosf (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshf' [-Wincompatible-library-redeclaration]\r\n int ccoshf (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpf' [-Wincompatible-library-redeclaration]\r\n int cexpf (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagf' [-Wincompatible-library-redeclaration]\r\n int cimagf (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogf' [-Wincompatible-library-redeclaration]\r\n int clogf (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjf' [-Wincompatible-library-redeclaration]\r\n int conjf (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowf' [-Wincompatible-library-redeclaration]\r\n int cpowf (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowf' is a builtin with type '_Complex float (_Complex float, _Complex float)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojf' [-Wincompatible-library-redeclaration]\r\n int cprojf (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'crealf' [-Wincompatible-library-redeclaration]\r\n int crealf (void);\r\n ^\r\n _configtest.c:17:5: note: 'crealf' is a builtin with type 'float (_Complex float)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinf' [-Wincompatible-library-redeclaration]\r\n int csinf (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhf' [-Wincompatible-library-redeclaration]\r\n int csinhf (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtf' [-Wincompatible-library-redeclaration]\r\n int csqrtf (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanf' [-Wincompatible-library-redeclaration]\r\n int ctanf (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanf' is a builtin with type '_Complex float (_Complex float)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhf' [-Wincompatible-library-redeclaration]\r\n int ctanhf (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhf' is a builtin with type '_Complex float (_Complex float)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'cabsl' [-Wincompatible-library-redeclaration]\r\n int cabsl (void);\r\n ^\r\n _configtest.c:1:5: note: 'cabsl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:2:5: warning: incompatible redeclaration of library function 'cacosl' [-Wincompatible-library-redeclaration]\r\n int cacosl (void);\r\n ^\r\n _configtest.c:2:5: note: 'cacosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:3:5: warning: incompatible redeclaration of library function 'cacoshl' [-Wincompatible-library-redeclaration]\r\n int cacoshl (void);\r\n ^\r\n _configtest.c:3:5: note: 'cacoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:4:5: warning: incompatible redeclaration of library function 'cargl' [-Wincompatible-library-redeclaration]\r\n int cargl (void);\r\n ^\r\n _configtest.c:4:5: note: 'cargl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:5:5: warning: incompatible redeclaration of library function 'casinl' [-Wincompatible-library-redeclaration]\r\n int casinl (void);\r\n ^\r\n _configtest.c:5:5: note: 'casinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:6:5: warning: incompatible redeclaration of library function 'casinhl' [-Wincompatible-library-redeclaration]\r\n int casinhl (void);\r\n ^\r\n _configtest.c:6:5: note: 'casinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:7:5: warning: incompatible redeclaration of library function 'catanl' [-Wincompatible-library-redeclaration]\r\n int catanl (void);\r\n ^\r\n _configtest.c:7:5: note: 'catanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:8:5: warning: incompatible redeclaration of library function 'catanhl' [-Wincompatible-library-redeclaration]\r\n int catanhl (void);\r\n ^\r\n _configtest.c:8:5: note: 'catanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:9:5: warning: incompatible redeclaration of library function 'ccosl' [-Wincompatible-library-redeclaration]\r\n int ccosl (void);\r\n ^\r\n _configtest.c:9:5: note: 'ccosl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:10:5: warning: incompatible redeclaration of library function 'ccoshl' [-Wincompatible-library-redeclaration]\r\n int ccoshl (void);\r\n ^\r\n _configtest.c:10:5: note: 'ccoshl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:11:5: warning: incompatible redeclaration of library function 'cexpl' [-Wincompatible-library-redeclaration]\r\n int cexpl (void);\r\n ^\r\n _configtest.c:11:5: note: 'cexpl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:12:5: warning: incompatible redeclaration of library function 'cimagl' [-Wincompatible-library-redeclaration]\r\n int cimagl (void);\r\n ^\r\n _configtest.c:12:5: note: 'cimagl' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:13:5: warning: incompatible redeclaration of library function 'clogl' [-Wincompatible-library-redeclaration]\r\n int clogl (void);\r\n ^\r\n _configtest.c:13:5: note: 'clogl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:14:5: warning: incompatible redeclaration of library function 'conjl' [-Wincompatible-library-redeclaration]\r\n int conjl (void);\r\n ^\r\n _configtest.c:14:5: note: 'conjl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:15:5: warning: incompatible redeclaration of library function 'cpowl' [-Wincompatible-library-redeclaration]\r\n int cpowl (void);\r\n ^\r\n _configtest.c:15:5: note: 'cpowl' is a builtin with type '_Complex long double (_Complex long double, _Complex long double)'\r\n _configtest.c:16:5: warning: incompatible redeclaration of library function 'cprojl' [-Wincompatible-library-redeclaration]\r\n int cprojl (void);\r\n ^\r\n _configtest.c:16:5: note: 'cprojl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:17:5: warning: incompatible redeclaration of library function 'creall' [-Wincompatible-library-redeclaration]\r\n int creall (void);\r\n ^\r\n _configtest.c:17:5: note: 'creall' is a builtin with type 'long double (_Complex long double)'\r\n _configtest.c:18:5: warning: incompatible redeclaration of library function 'csinl' [-Wincompatible-library-redeclaration]\r\n int csinl (void);\r\n ^\r\n _configtest.c:18:5: note: 'csinl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:19:5: warning: incompatible redeclaration of library function 'csinhl' [-Wincompatible-library-redeclaration]\r\n int csinhl (void);\r\n ^\r\n _configtest.c:19:5: note: 'csinhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:20:5: warning: incompatible redeclaration of library function 'csqrtl' [-Wincompatible-library-redeclaration]\r\n int csqrtl (void);\r\n ^\r\n _configtest.c:20:5: note: 'csqrtl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:21:5: warning: incompatible redeclaration of library function 'ctanl' [-Wincompatible-library-redeclaration]\r\n int ctanl (void);\r\n ^\r\n _configtest.c:21:5: note: 'ctanl' is a builtin with type '_Complex long double (_Complex long double)'\r\n _configtest.c:22:5: warning: incompatible redeclaration of library function 'ctanhl' [-Wincompatible-library-redeclaration]\r\n int ctanhl (void);\r\n ^\r\n _configtest.c:22:5: note: 'ctanhl' is a builtin with type '_Complex long double (_Complex long double)'\r\n 22 warnings generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:2:12: warning: unused function 'static_func' [-Wunused-function]\r\n static int static_func (char * restrict a)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:3:19: warning: unused function 'static_func' [-Wunused-function]\r\n static inline int static_func (void)\r\n ^\r\n 1 warning generated.\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h\r\n #define SIZEOF_PY_INTPTR_T 8\r\n #define SIZEOF_OFF_T 8\r\n #define SIZEOF_PY_LONG_LONG 8\r\n #define MATHLIB\r\n #define HAVE_SIN 1\r\n #define HAVE_COS 1\r\n #define HAVE_TAN 1\r\n #define HAVE_SINH 1\r\n #define HAVE_COSH 1\r\n #define HAVE_TANH 1\r\n #define HAVE_FABS 1\r\n #define HAVE_FLOOR 1\r\n #define HAVE_CEIL 1\r\n #define HAVE_SQRT 1\r\n #define HAVE_LOG10 1\r\n #define HAVE_LOG 1\r\n #define HAVE_EXP 1\r\n #define HAVE_ASIN 1\r\n #define HAVE_ACOS 1\r\n #define HAVE_ATAN 1\r\n #define HAVE_FMOD 1\r\n #define HAVE_MODF 1\r\n #define HAVE_FREXP 1\r\n #define HAVE_LDEXP 1\r\n #define HAVE_RINT 1\r\n #define HAVE_TRUNC 1\r\n #define HAVE_EXP2 1\r\n #define HAVE_LOG2 1\r\n #define HAVE_ATAN2 1\r\n #define HAVE_POW 1\r\n #define HAVE_NEXTAFTER 1\r\n #define HAVE_STRTOLL 1\r\n #define HAVE_STRTOULL 1\r\n #define HAVE_CBRT 1\r\n #define HAVE_STRTOLD_L 1\r\n #define HAVE_BACKTRACE 1\r\n #define HAVE_MADVISE 1\r\n #define HAVE_XMMINTRIN_H 1\r\n #define HAVE_EMMINTRIN_H 1\r\n #define HAVE_XLOCALE_H 1\r\n #define HAVE_DLFCN_H 1\r\n #define HAVE_SYS_MMAN_H 1\r\n #define HAVE___BUILTIN_ISNAN 1\r\n #define HAVE___BUILTIN_ISINF 1\r\n #define HAVE___BUILTIN_ISFINITE 1\r\n #define HAVE___BUILTIN_BSWAP32 1\r\n #define HAVE___BUILTIN_BSWAP64 1\r\n #define HAVE___BUILTIN_EXPECT 1\r\n #define HAVE___BUILTIN_MUL_OVERFLOW 1\r\n #define HAVE___BUILTIN_CPU_SUPPORTS 1\r\n #define HAVE__M_FROM_INT64 1\r\n #define HAVE__MM_LOAD_PS 1\r\n #define HAVE__MM_PREFETCH 1\r\n #define HAVE__MM_LOAD_PD 1\r\n #define HAVE___BUILTIN_PREFETCH 1\r\n #define HAVE_LINK_AVX 1\r\n #define HAVE_LINK_AVX2 1\r\n #define HAVE_XGETBV 1\r\n #define HAVE_ATTRIBUTE_NONNULL 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX 1\r\n #define HAVE_ATTRIBUTE_TARGET_AVX2 1\r\n #define HAVE___THREAD 1\r\n #define HAVE_SINF 1\r\n #define HAVE_COSF 1\r\n #define HAVE_TANF 1\r\n #define HAVE_SINHF 1\r\n #define HAVE_COSHF 1\r\n #define HAVE_TANHF 1\r\n #define HAVE_FABSF 1\r\n #define HAVE_FLOORF 1\r\n #define HAVE_CEILF 1\r\n #define HAVE_RINTF 1\r\n #define HAVE_TRUNCF 1\r\n #define HAVE_SQRTF 1\r\n #define HAVE_LOG10F 1\r\n #define HAVE_LOGF 1\r\n #define HAVE_LOG1PF 1\r\n #define HAVE_EXPF 1\r\n #define HAVE_EXPM1F 1\r\n #define HAVE_ASINF 1\r\n #define HAVE_ACOSF 1\r\n #define HAVE_ATANF 1\r\n #define HAVE_ASINHF 1\r\n #define HAVE_ACOSHF 1\r\n #define HAVE_ATANHF 1\r\n #define HAVE_HYPOTF 1\r\n #define HAVE_ATAN2F 1\r\n #define HAVE_POWF 1\r\n #define HAVE_FMODF 1\r\n #define HAVE_MODFF 1\r\n #define HAVE_FREXPF 1\r\n #define HAVE_LDEXPF 1\r\n #define HAVE_EXP2F 1\r\n #define HAVE_LOG2F 1\r\n #define HAVE_COPYSIGNF 1\r\n #define HAVE_NEXTAFTERF 1\r\n #define HAVE_CBRTF 1\r\n #define HAVE_SINL 1\r\n #define HAVE_COSL 1\r\n #define HAVE_TANL 1\r\n #define HAVE_SINHL 1\r\n #define HAVE_COSHL 1\r\n #define HAVE_TANHL 1\r\n #define HAVE_FABSL 1\r\n #define HAVE_FLOORL 1\r\n #define HAVE_CEILL 1\r\n #define HAVE_RINTL 1\r\n #define HAVE_TRUNCL 1\r\n #define HAVE_SQRTL 1\r\n #define HAVE_LOG10L 1\r\n #define HAVE_LOGL 1\r\n #define HAVE_LOG1PL 1\r\n #define HAVE_EXPL 1\r\n #define HAVE_EXPM1L 1\r\n #define HAVE_ASINL 1\r\n #define HAVE_ACOSL 1\r\n #define HAVE_ATANL 1\r\n #define HAVE_ASINHL 1\r\n #define HAVE_ACOSHL 1\r\n #define HAVE_ATANHL 1\r\n #define HAVE_HYPOTL 1\r\n #define HAVE_ATAN2L 1\r\n #define HAVE_POWL 1\r\n #define HAVE_FMODL 1\r\n #define HAVE_MODFL 1\r\n #define HAVE_FREXPL 1\r\n #define HAVE_LDEXPL 1\r\n #define HAVE_EXP2L 1\r\n #define HAVE_LOG2L 1\r\n #define HAVE_COPYSIGNL 1\r\n #define HAVE_NEXTAFTERL 1\r\n #define HAVE_CBRTL 1\r\n #define HAVE_DECL_SIGNBIT\r\n #define HAVE_COMPLEX_H 1\r\n #define HAVE_CABS 1\r\n #define HAVE_CACOS 1\r\n #define HAVE_CACOSH 1\r\n #define HAVE_CARG 1\r\n #define HAVE_CASIN 1\r\n #define HAVE_CASINH 1\r\n #define HAVE_CATAN 1\r\n #define HAVE_CATANH 1\r\n #define HAVE_CCOS 1\r\n #define HAVE_CCOSH 1\r\n #define HAVE_CEXP 1\r\n #define HAVE_CIMAG 1\r\n #define HAVE_CLOG 1\r\n #define HAVE_CONJ 1\r\n #define HAVE_CPOW 1\r\n #define HAVE_CPROJ 1\r\n #define HAVE_CREAL 1\r\n #define HAVE_CSIN 1\r\n #define HAVE_CSINH 1\r\n #define HAVE_CSQRT 1\r\n #define HAVE_CTAN 1\r\n #define HAVE_CTANH 1\r\n #define HAVE_CABSF 1\r\n #define HAVE_CACOSF 1\r\n #define HAVE_CACOSHF 1\r\n #define HAVE_CARGF 1\r\n #define HAVE_CASINF 1\r\n #define HAVE_CASINHF 1\r\n #define HAVE_CATANF 1\r\n #define HAVE_CATANHF 1\r\n #define HAVE_CCOSF 1\r\n #define HAVE_CCOSHF 1\r\n #define HAVE_CEXPF 1\r\n #define HAVE_CIMAGF 1\r\n #define HAVE_CLOGF 1\r\n #define HAVE_CONJF 1\r\n #define HAVE_CPOWF 1\r\n #define HAVE_CPROJF 1\r\n #define HAVE_CREALF 1\r\n #define HAVE_CSINF 1\r\n #define HAVE_CSINHF 1\r\n #define HAVE_CSQRTF 1\r\n #define HAVE_CTANF 1\r\n #define HAVE_CTANHF 1\r\n #define HAVE_CABSL 1\r\n #define HAVE_CACOSL 1\r\n #define HAVE_CACOSHL 1\r\n #define HAVE_CARGL 1\r\n #define HAVE_CASINL 1\r\n #define HAVE_CASINHL 1\r\n #define HAVE_CATANL 1\r\n #define HAVE_CATANHL 1\r\n #define HAVE_CCOSL 1\r\n #define HAVE_CCOSHL 1\r\n #define HAVE_CEXPL 1\r\n #define HAVE_CIMAGL 1\r\n #define HAVE_CLOGL 1\r\n #define HAVE_CONJL 1\r\n #define HAVE_CPOWL 1\r\n #define HAVE_CPROJL 1\r\n #define HAVE_CREALL 1\r\n #define HAVE_CSINL 1\r\n #define HAVE_CSINHL 1\r\n #define HAVE_CSQRTL 1\r\n #define HAVE_CTANL 1\r\n #define HAVE_CTANHL 1\r\n #define NPY_RESTRICT restrict\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define HAVE_LDOUBLE_INTEL_EXTENDED_16_BYTES_LE 1\r\n #define NPY_PY3K 1\r\n #ifndef __cplusplus\r\n \/* #undef inline *\/\r\n #endif\r\n \r\n #ifndef _NPY_NPY_CONFIG_H_\r\n #error config.h should never be included directly, include npy_config.h instead\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n Generating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n _configtest.c:1:5: warning: incompatible redeclaration of library function 'exp' [-Wincompatible-library-redeclaration]\r\n int exp (void);\r\n ^\r\n _configtest.c:1:5: note: 'exp' is a builtin with type 'double (double)'\r\n 1 warning generated.\r\n clang _configtest.o -o _configtest\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d _configtest\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -c'\r\n clang: _configtest.c\r\n success!\r\n removing: _configtest.c _configtest.o _configtest.o.d\r\n File: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h\r\n #define NPY_SIZEOF_SHORT SIZEOF_SHORT\r\n #define NPY_SIZEOF_INT SIZEOF_INT\r\n #define NPY_SIZEOF_LONG SIZEOF_LONG\r\n #define NPY_SIZEOF_FLOAT 4\r\n #define NPY_SIZEOF_COMPLEX_FLOAT 8\r\n #define NPY_SIZEOF_DOUBLE 8\r\n #define NPY_SIZEOF_COMPLEX_DOUBLE 16\r\n #define NPY_SIZEOF_LONGDOUBLE 16\r\n #define NPY_SIZEOF_COMPLEX_LONGDOUBLE 32\r\n #define NPY_SIZEOF_PY_INTPTR_T 8\r\n #define NPY_SIZEOF_OFF_T 8\r\n #define NPY_SIZEOF_PY_LONG_LONG 8\r\n #define NPY_SIZEOF_LONGLONG 8\r\n #define NPY_NO_SMP 0\r\n #define NPY_HAVE_DECL_ISNAN\r\n #define NPY_HAVE_DECL_ISINF\r\n #define NPY_HAVE_DECL_ISFINITE\r\n #define NPY_HAVE_DECL_SIGNBIT\r\n #define NPY_USE_C99_COMPLEX 1\r\n #define NPY_HAVE_COMPLEX_DOUBLE 1\r\n #define NPY_HAVE_COMPLEX_FLOAT 1\r\n #define NPY_HAVE_COMPLEX_LONG_DOUBLE 1\r\n #define NPY_RELAXED_STRIDES_CHECKING 1\r\n #define NPY_USE_C99_FORMATS 1\r\n #define NPY_VISIBILITY_HIDDEN __attribute__((visibility(\"hidden\")))\r\n #define NPY_ABI_VERSION 0x01000009\r\n #define NPY_API_VERSION 0x0000000D\r\n \r\n #ifndef __STDC_FORMAT_MACROS\r\n #define __STDC_FORMAT_MACROS 1\r\n #endif\r\n \r\n EOF\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h']\r\n building extension \"numpy.core._multiarray_tests\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n building extension \"numpy.core._multiarray_umath\" sources\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h' to sources.\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_numpy_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h' to sources.\r\n executing numpy\/core\/code_generators\/generate_ufunc_api.py\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath' to include_dirs.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h\r\n adding 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common' to include_dirs.\r\n numpy.core - nothing done with h_files = ['build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/funcs.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/simd.inc', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_internal.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/templ_common.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/config.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/_numpyconfig.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h', 'build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__ufunc_api.h']\r\n building extension \"numpy.core._umath_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n building extension \"numpy.core._rational_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n building extension \"numpy.core._struct_ufunc_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n building extension \"numpy.core._operand_flag_tests\" sources\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n building extension \"numpy.fft.fftpack_lite\" sources\r\n building extension \"numpy.linalg.lapack_lite\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n building extension \"numpy.linalg._umath_linalg\" sources\r\n adding 'numpy\/linalg\/lapack_lite\/python_xerbla.c' to sources.\r\n conv_template:> build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n building extension \"numpy.random.mtrand\" sources\r\n creating build\/src.macosx-10.15-x86_64-3.9\/numpy\/random\r\n building data_files sources\r\n build_src: building npy-pkg config files\r\n running build_py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/conftest.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_globals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/dual.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_distributor_init.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/ctypeslib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/matlib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying numpy\/_pytesttester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/py3k.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n copying numpy\/compat\/_inspect.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/compat\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/fromnumeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_add_newdocs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_methods.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_internal.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_string_helpers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/multiarray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/records.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup_common.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_aliased_types.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/memmap.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/overrides.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/getlimits.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_dtype_ctypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/defchararray.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/machar.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numeric.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/einsumfunc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/umath_tests.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/numerictypes.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/_type_aliases.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/cversions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/arrayprint.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n copying numpy\/core\/code_generators\/generate_numpy_api.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/unixccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/numpy_distribution.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/conv_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/cpuinfo.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvc9compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/npy_pkg_config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/compat.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/misc_util.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/log.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/line_endings.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/lib2def.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/pathccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/system_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/exec_command.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/from_template.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/mingw32ccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/extension.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/msvccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/intelccompiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying numpy\/distutils\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n copying build\/src.macosx-10.15-x86_64-3.9\/numpy\/distutils\/__config__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config_compiler.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_ext.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/config.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_headers.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_py.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_src.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/sdist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_scripts.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/bdist_rpm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/build_clib.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/autodist.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/egg_info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/develop.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n copying numpy\/distutils\/command\/install_data.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/command\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/gnu.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/compaq.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/intel.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/none.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/nag.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/ibm.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/sun.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/lahey.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/g95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/mips.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/hpux.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/environment.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/pathf95.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/absoft.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n copying numpy\/distutils\/fcompiler\/vast.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/distutils\/fcompiler\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/misc.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/internals.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/creation.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/constants.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/ufuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/broadcasting.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/basics.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/subclassing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/indexing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/byteswapping.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/structured_arrays.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n copying numpy\/doc\/glossary.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/doc\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/common_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/crackfortran.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/cb_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py2e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/func2subr.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__version__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/diagnose.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/capi_maps.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f90mod_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/f2py_testing.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/use_rules.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/auxfuncs.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n copying numpy\/f2py\/__main__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/f2py\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/helper.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/fftpack.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n copying numpy\/fft\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_iotools.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/mixins.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/nanfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/recfunctions.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/histograms.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/scimath.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/user_array.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/format.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/twodim_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/financial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/index_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/npyio.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/shape_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/stride_tricks.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arrayterator.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/function_base.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraysetops.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/arraypad.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/type_check.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/_datasource.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n copying numpy\/lib\/ufunclike.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/lib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/linalg.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n copying numpy\/linalg\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/extras.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/version.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/testutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/core.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/bench.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/timer_comparison.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n copying numpy\/ma\/mrecords.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/ma\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n copying numpy\/matrixlib\/defmatrix.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/matrixlib\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/laguerre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/_polybase.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polyutils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite_e.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/chebyshev.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/polynomial.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/legendre.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n copying numpy\/polynomial\/hermite.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/polynomial\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n copying numpy\/random\/info.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/setup.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/print_coercion_tables.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n copying numpy\/testing\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\r\n creating build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/nosetester.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/__init__.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/noseclasses.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/utils.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/parameterized.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n copying numpy\/testing\/_private\/decorators.py -> build\/lib.macosx-10.15-x86_64-3.9\/numpy\/testing\/_private\r\n running build_clib\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_clib\r\n building 'npymath' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n ar: adding 4 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpymath.a\r\n building 'npysort' library\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\r\n compile options: '-Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/quicksort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/mergesort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/heapsort.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/selection.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npysort\/binsearch.c\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/npysort\/selection.c.src:328:9: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp k;\r\n ^~~~~~~~~~~\r\n numpy\/core\/src\/npysort\/selection.c.src:326:14: note: silence by adding parentheses to mark code as explicitly dead\r\n else if (0 && kth == num - 1) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n 22 warnings generated.\r\n ar: adding 5 object files to build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n ranlib:@ build\/temp.macosx-10.15-x86_64-3.9\/libnpysort.a\r\n running build_ext\r\n customize UnixCCompiler\r\n customize UnixCCompiler using build_ext\r\n building 'numpy.core._dummy' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/core\/src\/dummymodule.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/dummymodule.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_dummy.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.c\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/_multiarray_tests.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_tests.cpython-39-darwin.so\r\n building 'numpy.core._multiarray_umath' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/core\/src\/multiarray\/alloc.c\r\n clang: numpy\/core\/src\/multiarray\/calculation.cclang: numpy\/core\/src\/multiarray\/array_assign_scalar.c\r\n clang: numpy\/core\/src\/multiarray\/convert.c\r\n \r\n clang: numpy\/core\/src\/multiarray\/ctors.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busday.c\r\n clang: numpy\/core\/src\/multiarray\/dragon4.cclang: numpy\/core\/src\/multiarray\/flagsobject.c\r\n \r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/ctors.c:2261:36: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (!(PyUString_Check(name) && PyUString_GET_SIZE(name) == 0)) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/arrayobject.c\r\n clang: numpy\/core\/src\/multiarray\/array_assign_array.c\r\n clang: numpy\/core\/src\/multiarray\/convert_datatype.c\r\n clang: numpy\/core\/src\/multiarray\/getset.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_busdaycal.c\r\n clang: numpy\/core\/src\/multiarray\/buffer.c\r\n clang: numpy\/core\/src\/multiarray\/compiled_base.c\r\n clang: numpy\/core\/src\/multiarray\/hashdescr.c\r\n clang: numpy\/core\/src\/multiarray\/descriptor.c\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:453:13: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n if (PyUString_GET_SIZE(name) == 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/descriptor.c:460:48: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n else if (PyUString_Check(title) && PyUString_GET_SIZE(title) > 0) {\r\n ^\r\n numpy\/core\/include\/numpy\/npy_3kcompat.h:110:28: note: expanded from macro 'PyUString_GET_SIZE'\r\n #define PyUString_GET_SIZE PyUnicode_GET_SIZE\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/conversion_utils.c\r\n clang: numpy\/core\/src\/multiarray\/item_selection.c\r\n clang: numpy\/core\/src\/multiarray\/dtype_transfer.c\r\n clang: numpy\/core\/src\/multiarray\/mapping.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.c\r\n 3 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/datetime.c\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:477:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ptr = PyUnicode_AS_UNICODE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/arraytypes.c.src:482:15: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n datalen = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/common.c\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:187:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:239:28: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n itemsize = PyUnicode_GET_DATA_SIZE(temp);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/common.c:282:24: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n int itemsize = PyUnicode_GET_DATA_SIZE(obj);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 6 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/nditer_pywrap.c\r\n 9 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/sequence.c\r\n clang: numpy\/core\/src\/multiarray\/shape.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.c\r\n clang: numpy\/core\/src\/multiarray\/methods.c\r\n clang: numpy\/core\/src\/multiarray\/iterators.c\r\n clang: numpy\/core\/src\/multiarray\/datetime_strings.c\r\n clang: numpy\/core\/src\/multiarray\/number.c\r\n clang: numpy\/core\/src\/multiarray\/scalarapi.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.c\r\n numpy\/core\/src\/multiarray\/scalarapi.c:74:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:135:28: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n return (void *)PyUnicode_AS_DATA(scalar);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalarapi.c:568:29: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n descr->elsize = PyUnicode_GET_DATA_SIZE(sc);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:475:17: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n ip = dptr = PyUnicode_AS_UNICODE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:476:11: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n len = PyUnicode_GET_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:481:11: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n new = PyUnicode_FromUnicode(ip, len);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1849:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buffer = PyUnicode_AS_DATA(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:283:21: note: expanded from macro 'PyUnicode_AS_DATA'\r\n ((const char *)(PyUnicode_AS_UNICODE(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:279:7: note: expanded from macro 'PyUnicode_AS_UNICODE'\r\n PyUnicode_AsUnicode(_PyObject_CAST(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/scalartypes.c.src:1850:18: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n buflen = PyUnicode_GET_DATA_SIZE(self);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:268:6: note: expanded from macro 'PyUnicode_GET_DATA_SIZE'\r\n (PyUnicode_GET_SIZE(op) * Py_UNICODE_SIZE)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 5 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/typeinfo.c\r\n clang: numpy\/core\/src\/multiarray\/refcount.c\r\n clang: numpy\/core\/src\/multiarray\/usertypes.c\r\n clang: numpy\/core\/src\/multiarray\/multiarraymodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.c\r\n clang: numpy\/core\/src\/multiarray\/vdot.c\r\n clang: numpy\/core\/src\/umath\/umathmodule.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.c\r\n clang: numpy\/core\/src\/umath\/reduction.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.c\r\n clang: numpy\/core\/src\/multiarray\/nditer_api.c\r\n 14 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/strfuncs.c\r\n numpy\/core\/src\/umath\/loops.c.src:655:18: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n result = PyEval_CallObject(tocall, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:178:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_ReprFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/core\/src\/multiarray\/strfuncs.c:195:13: warning: 'PyEval_CallObjectWithKeywords' is deprecated [-Wdeprecated-declarations]\r\n s = PyEval_CallObject(PyArray_StrFunction, arglist);\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:24:5: note: expanded from macro 'PyEval_CallObject'\r\n PyEval_CallObjectWithKeywords(callable, arg, (PyObject *)NULL)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/ceval.h:17:1: note: 'PyEval_CallObjectWithKeywords' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.9) PyAPI_FUNC(PyObject *) PyEval_CallObjectWithKeywords(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 2 warnings generated.\r\n clang: numpy\/core\/src\/multiarray\/temp_elide.c\r\n clang: numpy\/core\/src\/umath\/cpuid.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.c\r\n clang: numpy\/core\/src\/umath\/ufunc_object.c\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'byte_long' [-Wunused-function]\r\n byte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ubyte_long' [-Wunused-function]\r\n ubyte_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'short_long' [-Wunused-function]\r\n short_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ushort_long' [-Wunused-function]\r\n ushort_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'int_long' [-Wunused-function]\r\n int_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'uint_long' [-Wunused-function]\r\n uint_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'long_long' [-Wunused-function]\r\n long_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulong_long' [-Wunused-function]\r\n ulong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longlong_long' [-Wunused-function]\r\n longlong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'ulonglong_long' [-Wunused-function]\r\n ulonglong_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'half_long' [-Wunused-function]\r\n half_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'float_long' [-Wunused-function]\r\n float_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'double_long' [-Wunused-function]\r\n double_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'longdouble_long' [-Wunused-function]\r\n longdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cfloat_long' [-Wunused-function]\r\n cfloat_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'cdouble_long' [-Wunused-function]\r\n cdouble_long(PyObject *obj)\r\n ^\r\n numpy\/core\/src\/umath\/scalarmath.c.src:1449:1: warning: unused function 'clongdouble_long' [-Wunused-function]\r\n clongdouble_long(PyObject *obj)\r\n ^\r\n clang: numpy\/core\/src\/multiarray\/nditer_constr.c\r\n numpy\/core\/src\/umath\/ufunc_object.c:657:19: warning: comparison of integers of different signs: 'int' and 'size_t' (aka 'unsigned long') [-Wsign-compare]\r\n for (i = 0; i < len; i++) {\r\n ~ ^ ~~~\r\n clang: numpy\/core\/src\/umath\/override.c\r\n clang: numpy\/core\/src\/npymath\/npy_math.c\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.c\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n numpy\/core\/src\/umath\/loops.c.src:2527:22: warning: code will never be executed [-Wunreachable-code]\r\n npy_intp n = dimensions[0];\r\n ^~~~~~~~~~\r\n numpy\/core\/src\/umath\/loops.c.src:2526:29: note: silence by adding parentheses to mark code as explicitly dead\r\n if (IS_BINARY_REDUCE && 0) {\r\n ^\r\n \/* DISABLES CODE *\/ ( )\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.c\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:48:33: warning: unused variable 'tiny' [-Wunused-const-variable]\r\n static const volatile npy_float tiny = 3.9443045e-31f;\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:25: warning: unused variable 'c_halff' [-Wunused-const-variable]\r\n static const npy_cfloat c_halff = {0.5F, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:25: warning: unused variable 'c_if' [-Wunused-const-variable]\r\n static const npy_cfloat c_if = {0.0, 1.0F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:25: warning: unused variable 'c_ihalff' [-Wunused-const-variable]\r\n static const npy_cfloat c_ihalff = {0.0, 0.5F};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddf' [-Wunused-function]\r\n caddf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubf' [-Wunused-function]\r\n csubf(npy_cfloat a, npy_cfloat b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegf' [-Wunused-function]\r\n cnegf(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulif' [-Wunused-function]\r\n cmulif(npy_cfloat a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:26: warning: unused variable 'c_half' [-Wunused-const-variable]\r\n static const npy_cdouble c_half = {0.5, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:26: warning: unused variable 'c_i' [-Wunused-const-variable]\r\n static const npy_cdouble c_i = {0.0, 1.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:26: warning: unused variable 'c_ihalf' [-Wunused-const-variable]\r\n static const npy_cdouble c_ihalf = {0.0, 0.5};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'cadd' [-Wunused-function]\r\n cadd(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csub' [-Wunused-function]\r\n csub(npy_cdouble a, npy_cdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cneg' [-Wunused-function]\r\n cneg(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmuli' [-Wunused-function]\r\n cmuli(npy_cdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:67:30: warning: unused variable 'c_halfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_halfl = {0.5L, 0.0};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:68:30: warning: unused variable 'c_il' [-Wunused-const-variable]\r\n static const npy_clongdouble c_il = {0.0, 1.0L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:69:30: warning: unused variable 'c_ihalfl' [-Wunused-const-variable]\r\n static const npy_clongdouble c_ihalfl = {0.0, 0.5L};\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:79:1: warning: unused function 'caddl' [-Wunused-function]\r\n caddl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:87:1: warning: unused function 'csubl' [-Wunused-function]\r\n csubl(npy_clongdouble a, npy_clongdouble b)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:137:1: warning: unused function 'cnegl' [-Wunused-function]\r\n cnegl(npy_clongdouble a)\r\n ^\r\n numpy\/core\/src\/npymath\/npy_math_complex.c.src:144:1: warning: unused function 'cmulil' [-Wunused-function]\r\n cmulil(npy_clongdouble a)\r\n ^\r\n 22 warnings generated.\r\n clang: numpy\/core\/src\/common\/mem_overlap.c\r\n clang: numpy\/core\/src\/npymath\/halffloat.c\r\n clang: numpy\/core\/src\/common\/array_assign.c\r\n clang: numpy\/core\/src\/common\/ufunc_override.c\r\n clang: numpy\/core\/src\/common\/npy_longdouble.c\r\n clang: numpy\/core\/src\/common\/numpyos.c\r\n clang: numpy\/core\/src\/common\/ucsnarrow.c\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/umath\/extobj.c\r\n numpy\/core\/src\/common\/ucsnarrow.c:139:34: warning: 'PyUnicode_FromUnicode' is deprecated [-Wdeprecated-declarations]\r\n ret = (PyUnicodeObject *)PyUnicode_FromUnicode((Py_UNICODE*)buf,\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:551:1: note: 'PyUnicode_FromUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(PyObject*) PyUnicode_FromUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 1 warning generated.\r\n clang: numpy\/core\/src\/common\/python_xerbla.c\r\n clang: numpy\/core\/src\/common\/cblasfuncs.c\r\n clang: \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c\r\n In file included from \/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.c:26:\r\n In file included from numpy\/core\/include\/numpy\/arrayobject.h:4:\r\n In file included from numpy\/core\/include\/numpy\/ndarrayobject.h:21:\r\n build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy\/__multiarray_api.h:1463:1: warning: unused function '_import_array' [-Wunused-function]\r\n _import_array(void)\r\n ^\r\n 1 warning generated.\r\n 17 warnings generated.\r\n clang: numpy\/core\/src\/umath\/ufunc_type_resolution.c\r\n 4 warnings generated.\r\n 4 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/alloc.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arrayobject.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/arraytypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_scalar.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/array_assign_array.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/buffer.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/calculation.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/compiled_base.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/common.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/convert_datatype.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/conversion_utils.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/ctors.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_strings.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busday.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/datetime_busdaycal.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/descriptor.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dragon4.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/dtype_transfer.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/einsum.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/flagsobject.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/getset.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/hashdescr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/item_selection.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/iterators.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/lowlevel_strided_loops.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/mapping.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/methods.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/multiarraymodule.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_templ.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_api.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_constr.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/nditer_pywrap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/number.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/refcount.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/sequence.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/shape.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalarapi.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/scalartypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/strfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/temp_elide.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/typeinfo.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/usertypes.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/multiarray\/vdot.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/umathmodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/reduction.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/loops.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/matmul.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_object.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/extobj.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/cpuid.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/scalarmath.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/ufunc_type_resolution.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/ieee754.o build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/npy_math_complex.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath\/halffloat.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/array_assign.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/mem_overlap.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/npy_longdouble.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ucsnarrow.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/ufunc_override.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/numpyos.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/cblasfuncs.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common\/python_xerbla.o build\/temp.macosx-10.15-x86_64-3.9\/private\/var\/folders\/fz\/0j719tys48x7jlnjnwc69smr0000gn\/T\/pip-install-ufzck51l\/numpy_b0e8a3953a1d4b46801f12bcea55536e\/numpy\/_build_utils\/src\/apple_sgemv_fix.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -lnpysort -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_multiarray_umath.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.core._umath_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_umath_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_umath_tests.cpython-39-darwin.so\r\n building 'numpy.core._rational_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_rational_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_rational_tests.cpython-39-darwin.so\r\n building 'numpy.core._struct_ufunc_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_struct_ufunc_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_struct_ufunc_tests.cpython-39-darwin.so\r\n building 'numpy.core._operand_flag_tests' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n compile options: '-DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/umath\/_operand_flag_tests.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/core\/_operand_flag_tests.cpython-39-darwin.so\r\n building 'numpy.fft.fftpack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\r\n compile options: '-Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/fft\/fftpack_litemodule.c\r\n clang: numpy\/fft\/fftpack.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/fft\/fftpack_lite.cpython-39-darwin.so\r\n building 'numpy.linalg.lapack_lite' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: numpy\/linalg\/lapack_litemodule.c\r\n clang: numpy\/linalg\/lapack_lite\/python_xerbla.c\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_litemodule.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.linalg._umath_linalg' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\r\n compile options: '-DNO_ATLAS_INFO=3 -DHAVE_CBLAS -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n extra options: '-msse3 -I\/System\/Library\/Frameworks\/vecLib.framework\/Headers'\r\n clang: build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.c\r\n numpy\/linalg\/umath_linalg.c.src:735:32: warning: unknown warning group '-Wmaybe-uninitialized', ignored [-Wunknown-warning-option]\r\n #pragma GCC diagnostic ignored \"-Wmaybe-uninitialized\"\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:541:1: warning: unused function 'dump_ufunc_object' [-Wunused-function]\r\n dump_ufunc_object(PyUFuncObject* ufunc)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:566:1: warning: unused function 'dump_linearize_data' [-Wunused-function]\r\n dump_linearize_data(const char* name, const LINEARIZE_DATA_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_FLOAT_matrix' [-Wunused-function]\r\n dump_FLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_DOUBLE_matrix' [-Wunused-function]\r\n dump_DOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CFLOAT_matrix' [-Wunused-function]\r\n dump_CFLOAT_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:602:1: warning: unused function 'dump_CDOUBLE_matrix' [-Wunused-function]\r\n dump_CDOUBLE_matrix(const char* name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_FLOAT_matrix' [-Wunused-function]\r\n zero_FLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_DOUBLE_matrix' [-Wunused-function]\r\n zero_DOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CFLOAT_matrix' [-Wunused-function]\r\n zero_CFLOAT_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:865:1: warning: unused function 'zero_CDOUBLE_matrix' [-Wunused-function]\r\n zero_CDOUBLE_matrix(void *dst_in, const LINEARIZE_DATA_t* data)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:1862:1: warning: unused function 'dump_geev_params' [-Wunused-function]\r\n dump_geev_params(const char *name, GEEV_PARAMS_t* params)\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2132:1: warning: unused function 'init_cgeev' [-Wunused-function]\r\n init_cgeev(GEEV_PARAMS_t* params,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2213:1: warning: unused function 'process_cgeev_results' [-Wunused-function]\r\n process_cgeev_results(GEEV_PARAMS_t *NPY_UNUSED(params))\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2376:1: warning: unused function 'dump_gesdd_params' [-Wunused-function]\r\n dump_gesdd_params(const char *name,\r\n ^\r\n numpy\/linalg\/umath_linalg.c.src:2864:1: warning: unused function 'dump_gelsd_params' [-Wunused-function]\r\n dump_gelsd_params(const char *name,\r\n ^\r\n 16 warnings generated.\r\n clang -bundle -undefined dynamic_lookup -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk build\/temp.macosx-10.15-x86_64-3.9\/build\/src.macosx-10.15-x86_64-3.9\/numpy\/linalg\/umath_linalg.o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/linalg\/lapack_lite\/python_xerbla.o -L\/usr\/local\/lib -L\/usr\/local\/opt\/openssl@1.1\/lib -L\/usr\/local\/opt\/sqlite\/lib -Lbuild\/temp.macosx-10.15-x86_64-3.9 -lnpymath -o build\/lib.macosx-10.15-x86_64-3.9\/numpy\/linalg\/_umath_linalg.cpython-39-darwin.so -Wl,-framework -Wl,Accelerate\r\n building 'numpy.random.mtrand' extension\r\n compiling C sources\r\n C compiler: clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers\r\n \r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\r\n creating build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\r\n compile options: '-D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c'\r\n clang: numpy\/random\/mtrand\/mtrand.c\r\n clang: numpy\/random\/mtrand\/initarray.cclang: numpy\/random\/mtrand\/randomkit.c\r\n \r\n clang: numpy\/random\/mtrand\/distributions.c\r\n numpy\/random\/mtrand\/mtrand.c:40400:34: error: no member named 'tp_print' in 'struct _typeobject'\r\n __pyx_type_6mtrand_RandomState.tp_print = 0;\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:22: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42673:52: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**name) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:26: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:261:7: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op) : \\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: 'PyUnicode_AsUnicode' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:262:14: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n ((void)PyUnicode_AsUnicode(_PyObject_CAST(op)),\\\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:580:1: note: 'PyUnicode_AsUnicode' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3) PyAPI_FUNC(Py_UNICODE *) PyUnicode_AsUnicode(\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n numpy\/random\/mtrand\/mtrand.c:42689:59: warning: '_PyUnicode_get_wstr_length' is deprecated [-Wdeprecated-declarations]\r\n (PyUnicode_GET_SIZE(**argname) != PyUnicode_GET_SIZE(key)) ? 1 :\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:264:8: note: expanded from macro 'PyUnicode_GET_SIZE'\r\n PyUnicode_WSTR_LENGTH(op)))\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:451:35: note: expanded from macro 'PyUnicode_WSTR_LENGTH'\r\n #define PyUnicode_WSTR_LENGTH(op) _PyUnicode_get_wstr_length((PyObject*)op)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/cpython\/unicodeobject.h:445:1: note: '_PyUnicode_get_wstr_length' has been explicitly marked deprecated here\r\n Py_DEPRECATED(3.3)\r\n ^\r\n \/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9\/pyport.h:508:54: note: expanded from macro 'Py_DEPRECATED'\r\n #define Py_DEPRECATED(VERSION_UNUSED) __attribute__((__deprecated__))\r\n ^\r\n 12 warnings and 1 error generated.\r\n error: Command \"clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot \/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/usr\/include -I\/Library\/Developer\/CommandLineTools\/SDKs\/MacOSX10.15.sdk\/System\/Library\/Frameworks\/Tk.framework\/Versions\/8.5\/Headers -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Inumpy\/core\/include -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/include\/numpy -Inumpy\/core\/src\/common -Inumpy\/core\/src -Inumpy\/core -Inumpy\/core\/src\/npymath -Inumpy\/core\/src\/multiarray -Inumpy\/core\/src\/umath -Inumpy\/core\/src\/npysort -I\/usr\/local\/include -I\/usr\/local\/opt\/openssl@1.1\/include -I\/usr\/local\/opt\/sqlite\/include -I\/Users\/destiny\/Downloads\/env\/include -I\/usr\/local\/Cellar\/python@3.9\/3.9.0_1\/Frameworks\/Python.framework\/Versions\/3.9\/include\/python3.9 -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/common -Ibuild\/src.macosx-10.15-x86_64-3.9\/numpy\/core\/src\/npymath -c numpy\/random\/mtrand\/mtrand.c -o build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o -MMD -MF build\/temp.macosx-10.15-x86_64-3.9\/numpy\/random\/mtrand\/mtrand.o.d\" failed with exit status 1\nDamn sorry","embeddings":[-0.372736305,-0.0714854077,-0.0845708922,0.2079993039,0.2768215835,0.0631752834,-0.2505984604,0.0786435232,-0.2070377618,-0.0340085514,0.0305880085,0.3349942863,-0.1231444255,0.0895890892,0.0807427391,0.0722528622,0.189591229,0.2151816189,-0.3760947287,0.0339735448,0.0351168886,0.1739135981,-0.1176334247,-0.0769590884,-0.1111136004,-0.2832199335,0.0382908024,0.1094798371,-0.532479167,-0.3132808805,0.291446507,0.0252360553,-0.0163556207,0.5489337444,-0.0001225818,0.1088191941,0.1684155613,0.018390229,-0.3556295037,-0.4525223374,-0.2284404039,-0.4977667034,0.1029686555,0.0328003541,0.2697275281,-0.2269828916,-0.1955371499,-0.3135961592,0.7321167588,0.186695531,0.1548049748,0.3139568865,0.1280289441,0.0492491126,0.1173379719,0.3006310165,-0.3177905381,0.3576257825,0.2570489943,-0.1687269062,0.2200257182,0.1589555889,-0.2279057503,0.2082943469,0.5441326499,-0.1704113036,-0.0771712363,-0.4264783263,-0.2593541741,0.2732714415,0.5033100843,-0.2703568041,-0.6291657686,-0.0928478241,0.0548009649,-0.2381634116,0.3260372877,0.223864004,-0.2062843293,0.0802830532,-0.4288931787,-0.280636698,-0.1864999384,0.2670033574,-0.0238301586,0.164580822,-0.1459159702,0.1176900715,0.2985162139,-0.0034400814,0.2251620293,0.0791203454,-0.0705096424,0.0032631624,-0.2801685035,0.0655268356,-0.2036794722,0.2356178463,0.1992911696,0.1688992977,-0.1797399819,-0.2005181015,-0.2243610919,0.0616602823,0.066762954,0.2027508169,-0.1514637321,0.1150361076,0.3352227509,0.2663400173,0.4560080171,0.0843001902,0.0707917213,-0.1902450472,-0.209478572,-0.1111337766,0.1524448395,-0.3028428555,-0.4012164772,0.1194737256,0.0135395043,0.1056581885,0.0688990727,0.1510053128,-0.1464695483,0.1697680354,0.2481278181,0.3707648218,-0.232209444,0.0304407589,-0.0021012723,-0.0312458072,-0.2525970638,0.0168092307,0.1831142902,-0.1020761356,0.3657963574,0.2121508121,0.1536038369,0.1180837601,0.2853684127,0.0168460198,-0.191518262,0.4379527569,-0.2559596002,0.2257401049,0.0813604146,0.3052143753,0.0385886207,-0.0087099858,0.1200957373,-0.0894460753,-0.1464142799,0.0493404903,-0.2953626513,0.051196415,-0.16042234,-0.0320122093,0.0422822423,-0.0837064162,0.015641503,-0.2536138892,0.0998989642,-0.1036708131,0.3004768193,0.414516747,-0.51960361,-0.038707234,-0.1943658739,-0.1891062409,-0.0495771132,0.1207961589,-0.1165530235,0.1363425553,-0.164071247,-0.4210309684,0.2341786772,-0.3646256328,-0.367303431,0.1608417332,-0.1646376401,-0.086746715,0.0553342775,0.0455297008,-0.0180610958,0.0614581928,-0.0477216803,0.326702714,-0.1066503674,-0.2653565109,0.0084401397,-0.5503122211,0.0675886497,0.0629741848,-0.0602851324,0.0150617491,-0.0644270927,0.139841333,0.0511045381,0.0582192577,-0.0014922149,0.6988541484,0.416780889,0.0951713175,-0.2119343877,-0.2936756313,-0.3661433756,0.1583058536,0.0255879425,0.0643135011,-0.3344601095,-0.0186114982,-0.1848731637,0.128934145,-0.0381260812,-0.0527399741,-0.0203619599,0.0027848945,0.2437096834,-0.2807203233,-0.233386904,0.5044182539,0.1574977785,0.2694315314,-0.4862544239,0.0366727933,-0.3219184875,-0.1439168453,0.1915305704,0.2339345664,0.0553197712,-0.2622070611,-0.1443571001,0.3139041066,-0.1654947251,-0.1707349569,-0.0076072491,0.1440196335,0.2182890773,-0.2099311203,0.2413927317,-0.0720721483,-0.1406363845,0.0983242542,-0.0743637085,-0.1697815061,0.0582351796,0.3382672966,-0.0074652079,0.3404535055,0.3141412735,0.0584230796,-0.1450206935,-0.4074391723,0.396992296,-0.0027272592,0.4638343155,-0.1445932984,0.0697035044,0.053041961,-0.1363226324,0.0916765481,-0.0233560391,0.0049676136,-0.072430402,0.1785400212,0.185349673,0.0348233506,0.5852743983,0.0619082898,-0.1347470284,0.1598725617,-0.0658946782,-0.0729199722,0.0373055674,0.1743800044,-0.0199271459,-0.0330145024,0.0282345489,-0.0624522939,-0.0412191078,-0.207671389,0.1253645271,0.2953568995,-0.3419783413,0.1114570275,-0.5203249454,-0.0051726159,-0.0906571373,-0.0719850361,-0.4134817719,-0.2235808372,-0.1058893874,0.1500547081,0.2082414627,0.1049688607,0.1166992337,0.0297720432,0.001696527,-0.209155798,-0.1628497243,-0.1249261945,-0.0730912238,-0.0193224195,0.1679847986,0.1203907058,0.4714294374,-0.0997886732,-0.1613499224,-0.4218867421,-0.261598736,0.107161954,-0.398339957,0.742423296,0.469727993,0.202538535,0.0647644475,0.1159555018,0.1875865012,-0.1887231469,-0.1085140929,-0.0309973564,-0.213460058,-0.2505242229,-0.1225516498,-0.3344918489,-0.32494241,-0.2550111711,0.1577945501,-0.0034383023,-0.1047476232,0.3142209351,0.2670063674,0.0404987186,-0.165805921,0.0468914136,-0.0268509146,-0.2246979922,0.177915588,-0.201831609,-0.2568696737,0.3019668162,0.2744387388,0.1829818487,0.2670661211,-0.7614741325,-0.2670167983,-0.3597413003,0.2398702502,-0.1860815585,0.2179733366,0.1911988556,0.0769287571,0.0779739991,-0.2079752088,-0.4608823061,-0.0097571854,0.0752782673,0.4969078898,0.0363830738,0.3437930942,-0.0585201345,0.6186058521,0.397529155,0.0103019681,0.1783744097,0.2334723771,0.5895819068,0.0778009444,-0.4482364953,0.2397813201,-0.0084547177,0.1363245547,-0.1942522824,0.2197874635,0.2414545864,-0.277361989,0.2657617033,0.0042451476,0.0055124569,-0.2210909873,-0.13537018,0.3050734699,0.2228150517,0.0064602434,0.0430852436,-0.1302471906,0.0280230325,0.1745655835,0.010982167,0.3000519872,-0.1954792738,-0.1373809576,-0.183105126,0.2909622788,-0.0428895615,0.4082678556,-0.2521476746,0.4757547975,0.0345090851,0.0108251274,0.475733906,-0.3083002269,-0.1645100862,0.1024655625,0.1262150258,-0.6377895474,-0.1250775605,-0.1298962086,0.1647143066,0.248098284,0.2099779695,-0.4318394959,-0.1485227644,0.2467826158,0.5641607046,0.0371086225,-0.1298632622,-0.3288927078,-0.3882852793,-0.2767420411,-0.1371272653,-0.058102645,-0.1284619868,0.2001442164,-0.0588850789,0.1972738355,0.1628406644,0.1069483757,-0.0801087394,0.2060672939,-0.1776001751,0.2210780978,0.3728102148,-0.1271787584,0.1659846008,0.737244606,-0.2061559558,-0.3858459294,-0.0367117412,0.1224328429,0.1555299163,0.3913010359,-0.2864465714,-0.1263913363,0.2399162054,0.2055450827,0.1049313173,0.14691028,0.289960444,0.1443811804,-0.195566237,-0.4653308094,0.1165174395,0.4374177158,0.1289001107,0.4145717323,0.0137373563,0.0171636641,-0.0216146,0.2067595273,0.8317614794,-0.0005798346,0.0876606479,0.3389310539,-0.0641187727,0.5443438888,-0.4247684479,0.04840637,-0.0802192986,-0.143647477,-0.1364471316,-0.3701676726,0.3992536366,0.1148540974,-0.2142028362,0.3585865796,-0.1218098328,0.29375422,0.0482256673,0.1957940161,-0.0632860139,-0.2802743912,-0.1753268093,0.1845365763,-0.233880952,0.1757169068,-0.1705885977,0.083735086,-0.3188177347,-0.0867874846,-0.5056916475,0.0665702522,-0.443862617,0.2258304805,-0.0344874859,-0.2310214341,0.3094379306,0.1276277453,0.3448578715,0.4515912831,-0.3112368286,0.1110853925,-0.2714110315,-0.2467177361,0.3035934865,0.0439759605,0.0467805564,-0.0107217925,-0.1727659255,-0.0625354499,-0.2353877276,0.0169170164,0.0544819459,-0.0357779078,-0.1322258413,-0.1305123866,-0.1653528064,-0.1025140062,0.2076852918,-0.2839880288,0.0506421104,0.07781367,-0.0196289588,0.4598361552,-0.1750662029,-0.2117223293,-0.1011548117,0.212752223,-0.1656555086,-0.2982017696,0.234419629,0.2739279866,-0.3369333148,-0.1571041346,-0.094751589,-0.1557620764,-0.4031636715,0.1162526757,0.1045675799,-0.1815831959,-0.1145593971,0.2249464691,0.2387459278,-0.2340987474,0.0011506572,-0.4424839914,-0.2236991078,0.2046369612,-0.1303168684,0.0251450893,0.1048556641,0.0480054915,-0.0482010245,0.1132196635,-0.1702207923,-0.1554876715,0.0242763478,-0.0713960975,0.3412455618,-0.0761460811,0.0959786624,-0.2396223247,-0.03813833,0.1586219817,-0.2529083788,-0.076279223,-0.0355480611,0.137926355,0.0933924913,-0.3160541952,0.212218523,-0.4301118255,-0.1320781857,-0.1007387489,0.183366701,0.2039175779,0.0625451133,-0.2080432922,-0.4835402668,0.0814061463,0.0820645839,0.2181892693,0.0904808417,0.0420457087,0.2254475355,0.0915967301,0.1900683045,-0.09184926,-0.3334983885,0.0691129938,0.0046931463,0.2152244151,0.1156246215,-0.3056086898,0.2381537408,0.2715603411,0.4740514159,0.5606653094,-0.0505268052,0.2009599954,0.3780045509,0.0619235337,-0.3095085323,-0.1179374307,0.0978347436,0.211131677,-0.0669629127,0.294595927,0.2904630005,-0.1662479341,0.1414123923,0.2137935907,0.2448209524,0.0442248173,0.2401469797,0.1740344614,0.1456380934,-0.0191844087,0.2478772849,0.2596318722,0.1486105621,0.2685900927,-0.0940031186,-0.0129717235,0.1056639478,0.2145727575,-0.1893776506,-0.3154191971,-0.0778527632,-0.14081873,0.3885463476,0.0168743245,-0.0760412812,0.2765650153,-0.3946425617,-0.3519563675,-0.1730092168,0.2614917159,0.0503016636,0.2280521244,0.3215574324,0.0809561014,0.1780232787,0.0048405635,0.0397591293,-0.4191242456,0.2413220108,0.1764912158,-0.1236993223,-0.770821631,0.2125341445,0.1664997637,-0.1692295074,-0.0538866408,0.3296135664,0.1637779623,-0.1454275995,-0.2812110484,0.5068164468,0.5802744627,0.3409990072,-0.0469977558,0.1705626249,-0.2142198682,0.0777670294,-0.0155832991,0.2700774372,-0.1369200051,0.154964745,0.236733079,0.0225825254,-0.1113553271,-0.1749951243,0.05340923,0.4135027826,-0.1955527812,0.1946346015,-0.1500745118,-0.0271252468,-0.0973940268,-0.0908994824,-0.4811028838,0.2690776885,0.5110985637,0.0316910855,-0.0370604061,-0.1930583566,0.0272696707,-0.0880471319,0.3524516523,0.495470494,0.191963926,-0.4115162492,-0.0745142847,-0.6475214958,-0.2043262273,-0.3776461184,-0.2580498755,0.1304243505,-0.0521419086,0.0602565371,0.3123220801,0.2321897298,0.2287330478,0.1469683051,-0.0183709953,-0.3618494868,-0.1180011779,0.0987984166,0.1109670922,0.0805131122,-0.2171806246,0.0052382955,0.2312532216,-0.2043973804,-0.2440494448,-0.0672755912,0.0961156338,0.0278290417,0.3226872683,0.0853667632,0.3677543998,0.2596828043,-0.2449216247,-0.4981413782,-0.2963164449,-0.1827402264,0.3557329476,-0.046792753,0.2464546263,-0.5367340446,0.0496459119,-0.5474626422,0.097678706,0.0588760599,-0.0952013209,-0.2168268859,0.2665975988,-0.215844512,0.3078436255,0.0800647959,0.1987073421,-0.0550648682,0.2909039557,-0.4305544794,-0.322242856,0.3843740821,0.168199867,-0.3109259009,0.3755890727,0.0602345243,0.2313626111,-0.4108432531,-0.5497607589,0.1153397188,0.0295620337,-0.0591647848,-0.2287070155,0.1518161595,-0.1175127327,-0.0596970692,0.009724074,0.5295980573,0.0082971165,-0.254521817,0.3482492864,0.1984189898]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1687","title":"Question: Shouldn't .info be a part of DatasetDict?","comments":"We could do something. There is a part of `.info` which is split specific (cache files, split instructions) but maybe if could be made to work.","body":"Currently, only `Dataset` contains the .info or .features, but as many datasets contains standard splits (train, test) and thus the underlying information is the same (or at least should be) across the datasets. \r\n\r\nFor instance:\r\n```\r\n>>> ds = datasets.load_dataset(\"conll2002\", \"es\")\r\n>>> ds.info\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nAttributeError: 'DatasetDict' object has no attribute 'info'\r\n```\r\n\r\nI could imagine that this wouldn't work for datasets dicts which hold entirely different datasets (multimodal datasets), but it seems odd that splits of the same dataset is treated the same as what is essentially different datasets. \r\n\r\nIntuitively it would also make sense that if a dataset is supplied via. the load_dataset that is have a common .info which covers the entire dataset.\r\n\r\nIt is entirely possible that I am missing another perspective","comment_length":26,"text":"Question: Shouldn't .info be a part of DatasetDict?\nCurrently, only `Dataset` contains the .info or .features, but as many datasets contains standard splits (train, test) and thus the underlying information is the same (or at least should be) across the datasets. \r\n\r\nFor instance:\r\n```\r\n>>> ds = datasets.load_dataset(\"conll2002\", \"es\")\r\n>>> ds.info\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nAttributeError: 'DatasetDict' object has no attribute 'info'\r\n```\r\n\r\nI could imagine that this wouldn't work for datasets dicts which hold entirely different datasets (multimodal datasets), but it seems odd that splits of the same dataset is treated the same as what is essentially different datasets. \r\n\r\nIntuitively it would also make sense that if a dataset is supplied via. the load_dataset that is have a common .info which covers the entire dataset.\r\n\r\nIt is entirely possible that I am missing another perspective\nWe could do something. There is a part of `.info` which is split specific (cache files, split instructions) but maybe if could be made to work.","embeddings":[0.0918117985,-0.1329854578,-0.0601033382,0.317915678,0.1159813926,0.2852307856,0.4842665493,0.104040511,0.1170904636,-0.0210736264,0.1198941171,-0.0122084459,0.1128129661,0.5723268986,-0.0662022829,-0.1641823649,-0.0705984831,0.148715958,0.4008666873,-0.217286855,-0.0350506678,-0.0796024948,-0.28389889,0.0697829798,-0.5201468468,0.1602342576,-0.1704638451,0.1148243397,0.1587544233,-0.2171791047,0.3359515667,0.079117544,0.2234056294,0.0951015726,-0.0001184233,0.0710313171,0.4172881246,-0.1005561054,-0.585830152,-0.0511120446,-0.4921867847,-0.2075820267,0.2555397153,-0.3876200318,0.1147451177,-0.0305139609,0.0198400728,-0.3870246708,-0.1367856413,0.0425017625,0.0913463607,0.2089169919,0.0781250373,0.1249756664,0.4283175468,0.440869838,-0.0037694895,-0.0738444403,0.0077280076,0.3080052435,0.1272393465,0.34235695,-0.0232125279,0.133755222,0.3417018354,0.1327822953,0.1155046076,-0.0591512546,0.1476278752,0.2283668965,0.7229958773,-0.1421965659,-0.5014352202,-0.4906575084,0.0095369993,-0.0605854467,0.0451939628,0.2498720139,0.1005141586,0.2560992241,-0.0096330466,-0.2655453086,0.0091192648,0.0590938404,-0.3169502318,0.2660659254,-0.2420030385,0.3724735379,-0.1499092728,0.2580593824,0.3194072247,-0.4223934412,-0.1593245566,-0.3487668335,0.0991014019,-0.1485087723,-0.3174663186,0.0585657805,0.2237863541,0.0928342566,-0.1094228253,0.1233303994,-0.3389116228,0.1974087358,0.6273680329,-0.1340590715,0.4555550814,0.3131785989,0.4928690195,-0.2206749767,-0.1782532334,-0.0730944946,0.0329697728,0.1471543014,0.0598963723,-0.0845930278,0.3374170959,-0.1648215353,-0.3835932612,-0.1304867864,-0.1527523547,-0.2434693724,0.1898399442,-0.0050661671,0.1400835514,0.2932530344,-0.1279888898,0.0353852585,0.2271337062,-0.4226893783,-0.1293718517,-0.2358970046,-0.0385828502,0.3673235774,-0.197487548,-0.1680685133,0.3483381271,-0.1482053995,-0.0717323422,-0.0511191003,0.0808620229,-0.1957073957,0.0654319674,0.3250560462,-0.082434684,0.39034307,0.1030830741,-0.1345704347,-0.2661689222,0.1504443288,-0.6408148408,-0.4634259939,0.2243560255,0.0530499816,-0.1358696967,0.2209243029,-0.1107752249,0.2632291615,0.3134664595,-0.0908262879,0.1531967074,-0.1356644779,-0.296440959,-0.1926665306,0.283311367,0.2104682177,-0.3970859647,-0.2136283517,-0.1504446268,-0.1436151117,0.0299902745,0.1212649345,-0.3859351575,0.2841060758,-0.2248269469,-0.3534811139,0.3347937465,-0.1946114004,-0.1144688651,0.4579925835,0.1861239821,0.0381088071,0.1050548404,0.2736823261,-0.0585425459,0.0713181198,0.1039349735,0.2583704889,-0.2353020757,-0.2740975618,0.0002957224,-0.1138076633,0.310577184,0.2131095827,-0.1159447134,-0.0384645127,-0.2344484329,0.2109272778,0.3655130863,-0.0272810124,0.0955586061,0.0140591124,0.1028714553,0.4357927144,0.2865960598,-0.2845999002,-0.4904100597,0.1311647147,0.1655048877,-0.395472616,0.1204107329,-0.4200781584,0.0800234452,-0.1782492846,-0.1572500467,-0.3746443391,-0.0022305641,-0.1508160681,-0.1624032706,-0.1777560413,-0.4235555232,0.4239542782,-0.2499213666,0.1152750477,-0.2289767861,0.2394696921,0.0387869477,0.2098432183,-0.1989639848,0.2727729082,-0.0657186434,-0.0491792671,0.110736534,0.3047856688,0.5942277312,-0.286133945,0.1041956469,0.0203605127,-0.0166928843,-0.1124620214,0.1144106612,-0.0618982203,0.0706393644,-0.2937588394,-0.1092709526,0.1607099175,-0.1580697894,0.3453448117,0.0543774143,0.174341619,0.169173792,-0.0406468436,-0.2175850719,-0.3095995486,-0.2781988382,-0.2834964395,0.3399203718,0.3674307466,-0.1479644775,0.1553955823,0.2227449864,0.0576762483,0.1480972469,-0.1831474453,-0.4656175077,0.0250392407,-0.0894895196,0.1794130355,0.3915216029,0.1990486979,0.154834792,-0.0784718394,0.3370197713,-0.1943040937,-0.072354205,0.4055977464,0.1690877527,-0.1283873171,-0.2659930587,0.0911554322,0.1691453457,-0.1398743987,-0.0753898025,-0.1289141476,-0.2589673996,0.1873482168,-0.2755749822,-0.1545038074,-0.2051540315,-0.1453302205,-0.3799349368,-0.4896480143,-0.0038610965,-0.0541994981,-0.2509652078,0.0167104527,-0.3541566133,0.2501591742,-0.026182007,-0.1255615354,0.0276781227,-0.3448781669,-0.0640548021,0.0100141894,0.0690846294,0.4820319712,0.1558512449,-0.2240110338,-0.0550546758,-0.4028355181,-0.1728874147,-0.0414768569,-0.1749634892,0.0103770448,0.1427564919,0.0835348517,0.0280568656,-0.1347143203,0.030028129,0.052287925,-0.0131286634,-0.1029175743,0.1530731171,-0.2047076225,-0.1125838533,-0.1419579238,-0.5771692395,-0.2368137538,0.2238031477,-0.1635049433,0.0578630194,0.1183964461,-0.2589979768,0.0879827216,0.0384262949,0.2131520063,-0.1650425792,-0.1554406285,0.1244345233,-0.0911379084,-0.0739802867,0.0501793288,-0.1604059786,0.0683867186,0.1043668166,-0.470696032,0.2623697519,-0.3920534849,0.0128376316,0.0992572457,0.2085049897,0.2222805172,-0.0709316358,0.1464599669,-0.0629284382,-0.2990629673,-0.0041911704,-0.2204363644,-0.0384828113,0.0579742566,-0.0815701187,-0.1445349753,0.5173141956,0.1857023984,0.0303034056,-0.1386299878,0.1564418077,0.4514404237,-0.2710128129,-0.2426128983,0.1004729941,0.0746991113,-0.236952126,0.2484903187,-0.0011771657,0.2600431144,-0.1933954358,-0.2747224867,0.0274014641,-0.378411293,0.1079768091,-0.3394284546,0.1228848621,-0.061063312,0.0279221088,0.0699402392,0.0313592963,-0.0473089665,0.68699193,0.2125435174,-0.0544528142,-0.7601422668,-0.5233815312,0.3629417121,0.132928282,0.0649044588,0.3725583255,-0.0514236465,-0.1317767352,-0.1335675865,0.2957074344,0.308726728,0.0008376676,-0.4387602806,0.0320483707,-0.2063318193,-0.4428910911,-0.1868585646,-0.0106602181,0.409609437,-0.0056822966,0.198918134,-0.2873961329,-0.1382307559,0.4776052833,0.4554176331,-0.1339907199,-0.3033165932,-0.2194998264,0.1771643907,0.1213681921,-0.478410691,0.0207735058,-0.2603870034,-0.1603539884,0.0404707231,-0.0451318361,0.0173410624,0.1864453852,-0.0271525234,0.1437589824,-0.1432802826,0.1059502289,0.3473182619,0.2760670781,0.2497085035,0.6773898602,-0.1211696863,-0.6218709946,0.0552464984,-0.2108050883,0.6216723323,0.1541042775,-0.0865342245,-0.0053953221,0.0068762638,-0.2875773609,-0.3153980672,-0.0931713879,0.0803465322,-0.068841964,-0.5808680058,-0.7927826643,0.3102198243,0.1676645279,-0.0158811472,0.4327215254,0.3155716062,-0.1672296375,0.2504054904,-0.4388625324,0.6755127311,0.1444548517,0.1506430656,-0.2772524953,0.2773195803,0.5739865899,0.0357267931,0.0863612965,-0.3134250641,-0.479962945,-0.0779711604,-0.2438887209,0.148746103,0.1541532576,-0.0418953598,0.2184707075,-0.1195214465,0.0541671477,-0.1822516173,0.3473967314,0.0076188101,0.0396562293,0.0260313749,-0.0304024462,0.0507251844,0.2001179606,0.2681599855,0.0471880436,-0.3281910121,-0.0386859551,-0.229134053,-0.3129148781,0.1777766943,0.4434984028,0.0024858881,-0.1892047375,0.1715146899,0.1304222196,0.669478476,0.0187586918,-0.2068848759,0.0130617237,-0.1004848331,0.0997364372,0.007352002,0.0586407296,0.2492208481,0.2794010937,-0.1724518389,0.0720134154,-0.1735914946,-0.2988300622,-0.3996561468,0.106174238,0.3400160968,-0.4496112764,0.0272467323,0.0921721235,0.0765690953,-0.2017685175,0.0139200501,0.1829655319,0.0334848762,0.2177861482,-0.069895573,-0.3221408725,-0.1218642145,-0.078536503,0.3185659945,0.006492462,0.2477065921,-0.2155152559,-0.2510914505,-0.0643264204,0.5433021784,0.1061232388,0.2311720848,-0.0636541098,0.1837387532,0.1524506509,-0.2490389645,-0.0781513676,0.0095972624,-0.082040593,-0.233327359,-0.2443508506,-0.2849214971,0.534414351,-0.1239704341,0.3136835396,0.0972733349,0.114989467,-0.0264210142,0.256306082,-0.1571938545,-0.1378162652,0.3991807103,0.2493344545,0.1371568888,0.2059926391,0.1675762385,0.0426624641,-0.0096541038,-0.3132180274,0.2585717738,-0.0766199604,-0.0210978165,0.2216659784,0.1844183505,0.1518867165,-0.0572775714,-0.2669841349,-0.0206332672,-0.2713765204,0.4246802032,0.3646689951,0.1537020057,0.6921367645,0.2261843681,0.2610784471,-0.2724517882,0.425239712,-0.0351863727,0.3260367513,0.061127536,0.1071466357,-0.2512070835,0.1255657375,-0.2718522549,0.0091490131,-0.2118488997,0.0561064295,-0.1930880547,-0.0750973746,0.4025211334,0.0802955329,0.1000443771,0.7510889769,-0.2151395231,0.1651091576,0.1358086616,0.0396690965,0.0208546724,0.1299548298,0.3349671364,0.0463065393,0.0596183054,0.2890665233,-0.0499845371,0.1184663847,0.1100523546,-0.0038776849,0.4358593225,0.0654246584,0.3864558935,0.3260371685,0.1769787371,0.0664719716,0.0723322481,0.3067263365,0.2118372023,0.5174105167,-0.0414484851,0.0469897725,0.3193527162,-0.1149276718,0.4228002131,-0.0783750042,-0.1103232354,0.1462045908,-0.0399976783,-0.3343277574,0.0088816006,0.2850526571,-0.4110283256,-0.156128034,0.1367542893,-0.4379039705,-0.1316363961,0.0749813691,0.0354778767,-0.014055416,0.1506729573,-0.2756997645,-0.0160298869,-0.1391406804,0.2499789745,-0.1785297692,-0.2056026757,-0.2083389461,0.3819350302,0.1837308258,0.1191876233,-0.0517097302,-0.2124293298,-0.0864067227,0.1472593993,0.5784415007,0.0440549329,-0.0509758033,0.3064542711,0.2258925438,-0.1777759194,-0.0701105297,-0.2706612051,-0.0872593522,0.2107874751,-0.26080966,-0.12770392,0.2905161679,0.1410027146,0.0792058483,-0.0067763044,0.369265765,0.337859571,-0.5210638046,0.3581643701,-0.0435105562,-0.2012472749,0.2808779776,0.0114161354,-0.211636737,-0.3510106206,0.1508915126,0.066733025,0.0462453663,0.0443047807,0.0183870103,-0.0416028425,0.0827523544,0.4115047753,0.1575247496,-0.1438445747,0.0859446377,-0.4509420097,0.0690573007,-0.2467597425,-0.2733412981,0.1212219298,0.1675492078,0.1472270936,0.2236828357,-0.0884711295,0.0046353494,0.0026609353,-0.4381082654,-0.1862560511,-0.4560613632,-0.2429239899,0.118826963,-0.0072454168,-0.356009692,0.1922553033,0.1653686166,-0.1472022831,0.0131182224,-0.1230940968,0.5838446617,0.6622903943,0.1086524203,-0.0602959432,0.5134742856,0.1388151497,0.008031439,-0.2333011478,-0.3123189211,-0.1814205796,0.4386747181,0.0361131765,-0.1086723134,-0.4148912132,-0.2217723727,-0.1412605643,0.5059242845,-0.1026513577,-0.0880628824,-0.0933946297,0.2447440624,-0.1111092642,0.2183565795,0.1561423838,0.0836762041,-0.0692711771,-0.2627574205,-0.0073285485,-0.0385815464,0.3029502034,-0.039808251,-0.1337749064,0.1064071208,0.2277564704,0.1762737632,0.0262918044,-0.4592275321,-0.2288872451,0.6860064864,-0.0836353451,-0.021555081,-0.2875185907,0.1600333154,-0.1428746879,-0.2343402654,0.2004330009,-0.104952082,-0.3920252323,0.0454398654,-0.1185890213]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1687","title":"Question: Shouldn't .info be a part of DatasetDict?","comments":"Yes this was kinda the idea I was going for. DatasetDict.info would be the shared info amongs the datasets (maybe even some info on how they differ). ","body":"Currently, only `Dataset` contains the .info or .features, but as many datasets contains standard splits (train, test) and thus the underlying information is the same (or at least should be) across the datasets. \r\n\r\nFor instance:\r\n```\r\n>>> ds = datasets.load_dataset(\"conll2002\", \"es\")\r\n>>> ds.info\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nAttributeError: 'DatasetDict' object has no attribute 'info'\r\n```\r\n\r\nI could imagine that this wouldn't work for datasets dicts which hold entirely different datasets (multimodal datasets), but it seems odd that splits of the same dataset is treated the same as what is essentially different datasets. \r\n\r\nIntuitively it would also make sense that if a dataset is supplied via. the load_dataset that is have a common .info which covers the entire dataset.\r\n\r\nIt is entirely possible that I am missing another perspective","comment_length":27,"text":"Question: Shouldn't .info be a part of DatasetDict?\nCurrently, only `Dataset` contains the .info or .features, but as many datasets contains standard splits (train, test) and thus the underlying information is the same (or at least should be) across the datasets. \r\n\r\nFor instance:\r\n```\r\n>>> ds = datasets.load_dataset(\"conll2002\", \"es\")\r\n>>> ds.info\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nAttributeError: 'DatasetDict' object has no attribute 'info'\r\n```\r\n\r\nI could imagine that this wouldn't work for datasets dicts which hold entirely different datasets (multimodal datasets), but it seems odd that splits of the same dataset is treated the same as what is essentially different datasets. \r\n\r\nIntuitively it would also make sense that if a dataset is supplied via. the load_dataset that is have a common .info which covers the entire dataset.\r\n\r\nIt is entirely possible that I am missing another perspective\nYes this was kinda the idea I was going for. DatasetDict.info would be the shared info amongs the datasets (maybe even some info on how they differ). ","embeddings":[0.109234564,-0.1863408387,-0.0522325188,0.3248639107,0.1043859795,0.2396507561,0.4616636634,0.0180867035,0.076928921,0.0197503231,0.1676876396,0.0010697683,0.0968394801,0.5461275578,-0.0978985056,-0.103799358,-0.0594279096,0.138886869,0.3905678689,-0.251594305,-0.000719186,-0.0910816714,-0.2596049607,0.098815605,-0.5232893825,0.1915978938,-0.1881441921,0.0919921249,0.1326063573,-0.2878218591,0.3615059257,0.1595394015,0.1973745078,0.1136342287,-0.0001183565,0.0224601906,0.4010705352,-0.0624151528,-0.5887030363,-0.052405633,-0.4848047197,-0.2119043916,0.3201882839,-0.3865294755,0.1211567298,-0.0696784258,0.0017127135,-0.4104648829,-0.2561742067,0.0234128647,0.0837033615,0.209484905,0.0782388151,0.1111850813,0.3784434199,0.4335740209,0.0096172029,0.0070683011,0.0320767947,0.2960720956,0.1255498976,0.3308481574,-0.0533949137,0.1116912067,0.4069631398,0.1460915059,0.1254811585,-0.0638061538,0.1105677485,0.2188150883,0.7944613695,-0.1607745588,-0.4864640832,-0.4499758482,0.0171879325,-0.053866718,0.0481885485,0.2049067616,0.1058291867,0.2054299861,-0.0294155907,-0.2366993278,-0.0012191251,0.087443091,-0.3106896877,0.2355629057,-0.2511017621,0.3670661151,-0.1846856624,0.2112950236,0.2617731094,-0.4690439701,-0.1036319882,-0.3024092913,0.1225377768,-0.1447336674,-0.2677628696,0.0410825089,0.1964635104,0.0816415474,-0.0882414952,0.0786856711,-0.3459379375,0.1982536167,0.6331498027,-0.1691520363,0.4581494331,0.3049709201,0.4819129407,-0.21129556,-0.1644596606,-0.0888506919,0.0035896231,0.1620987803,0.1140550375,-0.0938202739,0.3855833113,-0.1798552275,-0.3533093631,-0.0674244687,-0.1632728875,-0.2639004886,0.2029367834,-0.0227611251,0.1098847762,0.274785459,-0.088450022,-0.0320448205,0.2030628473,-0.3814787567,-0.1435827911,-0.2690339983,-0.069170326,0.3542395532,-0.2017106414,-0.1512514502,0.3689916134,-0.1425230205,0.0116628725,-0.0521384068,0.0536363795,-0.156157136,0.0642834306,0.3008127511,-0.1564820558,0.4282224178,0.0956930518,-0.1618210971,-0.2783508003,0.1801878661,-0.6619879603,-0.4488373399,0.2047552168,0.0613590963,-0.1576842517,0.1795571893,-0.1415863335,0.2406547815,0.2539420128,-0.0723043233,0.1906903386,-0.1591925472,-0.2682466805,-0.2610308826,0.2825265825,0.2337578684,-0.3897713423,-0.2297365963,-0.0804793686,-0.147293821,-0.0020092546,0.0697090328,-0.4143724144,0.2756814063,-0.2169331759,-0.3133245707,0.3193613887,-0.2567077577,-0.1279162019,0.4607948661,0.2108805031,0.0802655518,0.1223868281,0.3123600483,-0.1063220799,0.0454800278,0.1513047814,0.2537457645,-0.1846103221,-0.2593037784,-0.0427560024,-0.0478005558,0.3199851811,0.2106951475,-0.1783976108,-0.0691636875,-0.2170659751,0.1388737559,0.4230898917,-0.0462925956,0.0533452481,0.0540667772,0.1988194734,0.4479038119,0.2968827188,-0.2312646508,-0.4689643681,0.1559650749,0.1894424558,-0.3594921231,0.1258040667,-0.4240337312,0.0922378972,-0.1444715112,-0.1887962669,-0.3890085518,-0.0082081845,-0.1356377304,-0.223087579,-0.1789464504,-0.4328570962,0.4074249864,-0.287383914,0.1064642519,-0.250410974,0.2724049985,0.0433237664,0.2123250663,-0.1825290322,0.2804706097,-0.0611763746,-0.0184207931,0.1108569875,0.322879225,0.5905709267,-0.2544253767,0.09543778,0.0868837982,-0.0070470329,-0.1313748211,0.1871860772,-0.0639396608,0.0843911394,-0.2608762681,-0.1119738966,0.2004221231,-0.165943563,0.2659816444,0.0471609011,0.1501613706,0.1141627356,0.0130631216,-0.2590413094,-0.2280737907,-0.2409589142,-0.2200492322,0.3644494116,0.3960295618,-0.1986416429,0.1761419326,0.2211500108,0.0366481841,0.0990409032,-0.1873274148,-0.4361338615,0.0404323302,-0.1034636348,0.1863884777,0.4261831939,0.1874251962,0.1363627911,-0.0564795174,0.2948385179,-0.1621707231,-0.056969583,0.3631010056,0.1800805628,-0.1710114777,-0.2477538586,0.1055670902,0.1776001304,-0.2268302739,-0.1198953986,-0.1188481525,-0.2635375559,0.1931549758,-0.2276278585,-0.1255797148,-0.2238827199,-0.1674411893,-0.408067286,-0.470836997,-0.041249413,-0.0914924666,-0.2265574783,0.0324298739,-0.3502626419,0.2202373892,-0.027290795,-0.1370193511,0.0849741399,-0.3411065638,-0.0684848353,-0.0016225099,0.0199214891,0.4772357941,0.1395870894,-0.1437512487,0.0297061838,-0.3783156574,-0.2370514572,-0.0676569119,-0.2039967477,0.0427970327,0.1739371717,0.1056649163,0.0046056714,-0.1211947277,0.0363043435,0.0951905772,-0.0085574389,-0.0960078314,0.1449934542,-0.2609967589,-0.078554742,-0.1492268741,-0.5878081918,-0.2085361481,0.2513512969,-0.1442806274,0.0605821349,0.1840535253,-0.1893706024,0.0794620961,0.0890742838,0.2794890404,-0.1544049382,-0.222745508,0.1043919176,-0.0826161057,-0.083708398,0.0354142375,-0.1492855251,0.116129972,0.0625252649,-0.5031933784,0.1963873208,-0.3549006581,0.0471700951,0.0823211148,0.2206005603,0.2216579169,-0.0433164574,0.1332761645,-0.0704778135,-0.3033605218,-0.0373763293,-0.2083110362,-0.0586094595,0.0915668532,-0.1116663441,-0.1912150234,0.4794411957,0.1923932135,0.0343192369,-0.1583175808,0.1479472369,0.4536730051,-0.2255889773,-0.2565922141,0.0839340463,0.1156938896,-0.2311653495,0.2009001523,-0.0077882814,0.3088763654,-0.1947282553,-0.3063077927,0.0388078801,-0.3404156268,0.1031360477,-0.3827613592,0.1496530622,-0.0974634439,-0.0163760316,0.0242986996,0.0544932149,-0.062305402,0.6804859638,0.1500389874,-0.0791757032,-0.7875328064,-0.5043078065,0.3543514013,0.1103458628,0.033800967,0.3633687496,-0.047958117,-0.0858700201,-0.1417991966,0.2948877215,0.3043975234,-0.0153172072,-0.418092221,0.0420073904,-0.1848290712,-0.4713216126,-0.200055778,-0.0243441518,0.3679104447,-0.0322527587,0.1814920306,-0.2694959939,-0.1731759012,0.4519659877,0.4222109914,-0.1693472415,-0.3046568334,-0.2661786377,0.1701023132,0.1248489022,-0.5041344166,-0.0035559062,-0.2079113871,-0.1949191988,0.1103091016,-0.016112648,0.0882917419,0.2064442039,-0.0423499234,0.1636591405,-0.1593437493,0.123211816,0.3756864965,0.2779029608,0.2583453357,0.6850102544,-0.0984587818,-0.6121306419,0.0078619709,-0.249033615,0.6869461536,0.1279260516,-0.0900214389,-0.0067450632,-0.0160082113,-0.269726932,-0.3494412005,-0.1217173114,0.111990124,-0.0605169274,-0.5644884109,-0.7933292389,0.3231039047,0.1790104955,-0.0328593254,0.456007123,0.355999738,-0.201115936,0.235147655,-0.3832592964,0.691309154,0.17690292,0.1861806512,-0.2361904383,0.2525369823,0.5729828477,0.0631193817,0.0276864674,-0.3370052576,-0.4901946783,-0.0737897456,-0.2707250416,0.1726966798,0.130748868,-0.0557373986,0.21151115,-0.1061846763,0.029245371,-0.1781507879,0.3440638781,0.0306046139,0.0739399865,0.0062905652,-0.0323983692,0.0484623015,0.1883675903,0.2416067719,0.0242225323,-0.3567381799,-0.0699104369,-0.2014046907,-0.32207039,0.1840605587,0.4327131212,0.1349834353,-0.1679757535,0.1662174463,0.1572209895,0.6592188478,0.0222421046,-0.2307641953,-0.0212972797,-0.1278990805,0.1114328504,-0.0143335378,0.0161695052,0.2778161466,0.2751942575,-0.1926818192,0.104561083,-0.1761861742,-0.2330978662,-0.3902464211,0.07419312,0.3489204049,-0.4822297394,0.0250340104,0.1101201475,0.0408796296,-0.178389892,0.0120050982,0.2466860265,-0.005420838,0.2095369101,-0.0163705815,-0.2943812311,-0.0740962029,-0.0691502616,0.3758611679,-0.0444387719,0.2839975059,-0.2256935984,-0.2468979955,-0.0345865712,0.5593050122,0.1610975116,0.2006607801,-0.0909960121,0.2291177213,0.1600698531,-0.2370390892,-0.0087889144,-0.008673775,-0.0612079017,-0.1965877861,-0.2428296953,-0.2928566337,0.5185007453,-0.1055701748,0.3687820137,0.0104603227,0.0972035155,0.0215418693,0.1943461448,-0.1704063714,-0.1012378931,0.3857163191,0.2459945977,0.207506761,0.1977914423,0.1395378262,0.0345862135,0.001112855,-0.3333174288,0.2167121321,-0.0601824895,-0.0714589059,0.2272607088,0.1871923655,0.129270643,-0.0412131511,-0.2895896733,0.0127084292,-0.2425256371,0.3665853739,0.3701936901,0.1453304142,0.7033871412,0.2130199671,0.2670083642,-0.2300622314,0.4268025756,-0.005306175,0.3281851113,0.1175382957,0.0664870664,-0.3007629514,0.1175704375,-0.2519350648,0.0608937889,-0.256125927,0.0925450176,-0.1749143153,-0.0935115442,0.4097734392,0.0751038268,0.1259775907,0.7074720263,-0.2628042102,0.15441975,0.0697276518,0.0529177412,0.0444896668,0.0754190683,0.3436790705,0.093657352,0.0155060124,0.2582239509,-0.0460689552,0.1242726818,0.0824613422,-0.0076017152,0.4043335319,0.0479064137,0.354824245,0.221626997,0.1696322411,0.0462902114,0.1006653681,0.3653800189,0.1909578592,0.5154171586,-0.0293354075,0.0451469384,0.3739592433,-0.1580110192,0.382643491,-0.0960672796,-0.0135467779,0.1442211121,-0.0574904196,-0.3808069229,0.0007179086,0.1923139393,-0.3736928403,-0.1198177487,0.1813124865,-0.4378457367,-0.1517007649,0.0372377485,0.0433833636,-0.0408421829,0.1608287543,-0.2679539919,-0.0194446351,-0.1617603004,0.2487408072,-0.2149016708,-0.1819607764,-0.2487837076,0.3825968206,0.1977815628,0.1454022378,-0.0649392903,-0.2216399759,-0.0543532334,0.1535514146,0.5740056038,0.0758354813,0.0051942272,0.3431329429,0.2081570923,-0.1150468215,-0.0558258742,-0.250798285,-0.0516077131,0.1842878163,-0.2408695966,-0.1199262738,0.2851547599,0.1447785199,0.0790030882,0.0148640927,0.3991677165,0.3199594617,-0.5323261619,0.3043732941,-0.0565003976,-0.2260834873,0.2637318671,0.0439351276,-0.2527672946,-0.3791944981,0.2322812676,0.1450575143,0.0595515519,0.0132898539,0.0195109881,-0.0589776076,0.0886097923,0.4871283174,0.1757132113,-0.208682403,0.0501776412,-0.4910962582,0.0962912291,-0.1997430176,-0.3158703148,0.1360269487,0.1735565811,0.1628629267,0.210858658,-0.0823882222,0.0359608829,-0.0585147999,-0.442592591,-0.203019619,-0.5107286572,-0.2458575368,0.0823237449,-0.0229788199,-0.2609199882,0.1512150019,0.1273559481,-0.1151454374,-0.0306611285,-0.1179502234,0.5505143404,0.7200504541,0.1110587344,-0.0958185047,0.5151230693,0.1796166003,-0.064357169,-0.1798269004,-0.3064324856,-0.1049068123,0.472091943,-0.0040374189,-0.0576276556,-0.3662098646,-0.2138501257,-0.1761783212,0.462944448,-0.0740015134,-0.1350767314,-0.0880296603,0.2622854114,-0.0707028285,0.1925247759,0.0988910198,0.0881729126,-0.0489777327,-0.1841164231,0.0227208212,-0.0173229259,0.3316030204,-0.0744464993,-0.0924083889,0.0756232738,0.2343603224,0.1036890596,0.0704876408,-0.5209988356,-0.1857074052,0.7407820821,-0.098055236,-0.0945293754,-0.2407279909,0.1190873906,-0.1634677798,-0.2099820822,0.2338991314,-0.1494737715,-0.3507017493,0.0544901974,-0.1084604561]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1686","title":"Dataset Error: DaNE contains empty samples at the end","comments":"Thanks for reporting, I opened a PR to fix that","body":"The dataset DaNE, contains empty samples at the end. It is naturally easy to remove using a filter but should probably not be there, to begin with as it can cause errors.\r\n\r\n```python\r\n>>> import datasets\r\n[...]\r\n>>> dataset = datasets.load_dataset(\"dane\")\r\n[...]\r\n>>> dataset[\"test\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n>>> dataset[\"train\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n```\r\n\r\nBest,\r\nKenneth","comment_length":10,"text":"Dataset Error: DaNE contains empty samples at the end\nThe dataset DaNE, contains empty samples at the end. It is naturally easy to remove using a filter but should probably not be there, to begin with as it can cause errors.\r\n\r\n```python\r\n>>> import datasets\r\n[...]\r\n>>> dataset = datasets.load_dataset(\"dane\")\r\n[...]\r\n>>> dataset[\"test\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n>>> dataset[\"train\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n```\r\n\r\nBest,\r\nKenneth\nThanks for reporting, I opened a PR to fix that","embeddings":[0.0047835,-0.2403898537,-0.2218231708,0.0121033434,0.2349358797,0.1396797299,0.3500452936,0.2725845873,0.2178974301,0.2462106794,0.1730049998,0.2148654759,-0.1180567443,0.0999728292,-0.1143105701,-0.2607348561,0.0481527299,0.1995811164,-0.2149447948,-0.2789696455,-0.12232811,0.0250022188,-0.2561839521,-0.0289916713,-0.2914612293,-0.0998405814,-0.0971537232,-0.0776925385,-0.1865427494,-0.2878666818,0.1821992099,-0.0646890178,-0.0565813482,0.4812009931,-0.0000981325,0.0520607419,0.2300475389,0.013646801,-0.2183106095,-0.1275827289,-0.2707432806,-0.0882126912,0.0181330871,-0.3365775943,-0.0597366691,-0.0881359354,-0.1734833717,-0.1927319914,0.3999645114,0.3933990896,0.3475658298,-0.027789887,-0.0682239607,-0.2029356658,0.0823580027,-0.0273386165,0.0160632525,0.0282933619,0.1392069161,-0.1185554788,0.1856298149,0.3940171897,-0.1089104265,0.0233575795,0.0072888792,0.0720625445,0.4298939407,-0.4152368903,0.1382918358,0.1620172709,0.4731559455,-0.1426445097,-0.192247808,-0.1013650075,0.0646559,-0.2950894833,0.0091218557,0.1583848894,-0.1961812526,0.1582535952,-0.1982882172,0.1433401406,-0.0380259976,0.1466831714,-0.2527247667,0.3247109652,0.0829712227,-0.0312618725,-0.0317802913,-0.1319001019,-0.0139622856,-0.1529698521,-0.091470845,0.123409912,-0.1767315716,-0.1820046902,0.1194465533,-0.3166241944,0.0377648361,0.1620371938,0.0684630126,-0.1212072372,0.1164520457,0.2122350186,0.3267807066,0.1368256062,0.0242593512,0.206952706,-0.0740302876,-0.0487566628,-0.1185042411,0.0628318489,-0.0338665731,-0.3912796676,0.1759446412,0.0750385523,0.1084893867,-0.0728363171,-0.3022869825,0.2066791803,-0.024565991,0.1258664578,0.0035599764,0.3290969431,-0.0437329859,0.1228177175,-0.1003170609,0.1424917579,-0.0824723765,0.0509787202,-0.3363825381,0.1188664883,-0.2112102658,-0.5104455352,0.1090329662,0.1662182957,0.2951400578,0.1420786232,-0.1548848748,-0.0109105082,0.3115044236,-0.3557296097,0.1705346704,0.3001939356,0.1166939214,0.1585432589,0.1480792165,-0.2643091679,-0.0803127736,0.1741277874,-0.274545759,-0.1432556659,-0.2465383708,0.4661228359,0.0423090272,-0.1393921375,-0.3265492618,0.1380064785,0.0630980581,-0.1019717753,-0.0211525299,-0.0761060119,-0.0854233876,-0.1376732737,0.285274148,0.2476820648,-0.2472970039,-0.1127839386,0.044312425,-0.1165418029,0.2401260734,0.2916325629,-0.0737272277,0.1984496862,-0.197626546,0.1159071326,0.3468621373,-0.2681879699,-0.2044838816,0.1080804691,-0.0212782342,-0.0594710074,-0.1420601159,-0.076414898,0.0131461695,0.1946516484,0.1796733886,0.2343626022,-0.0465882085,-0.0177180897,-0.2471743524,-0.0847195685,0.2942718863,0.1914488524,0.1427886039,0.0277405847,-0.1754400581,0.128274709,0.2297190428,-0.0438825674,-0.0091347052,0.4321836531,0.4600350857,0.0109708561,-0.0966791809,-0.0511314869,-0.0742258281,0.0413917005,0.1286696941,0.2854814529,0.0122192772,-0.1294894814,-0.2764928043,-0.1619101465,-0.2304739207,-0.1355530769,0.3535484076,-0.0445195511,-0.0859319419,0.1230479404,-0.2056271285,0.1109890789,-0.1798298806,0.1318146139,-0.1031758934,0.3387281597,-0.2518329024,-0.0023516021,-0.0435927883,0.1376651078,0.1118887514,0.056201864,-0.0165225621,0.1821834594,0.1526858956,-0.0873604268,0.1373738796,-0.1631267071,0.0495133474,-0.2184158415,-0.1555243433,0.3890599906,0.1427763402,-0.0272259675,-0.2053173035,0.0669293106,0.1496278197,0.2045339346,0.10755983,0.0162611455,0.1132226735,-0.2187414169,0.1720788032,-0.1833790839,0.1725168228,-0.1377925873,0.1262438595,-0.0168160554,-0.1199564189,-0.114815101,0.2722395957,0.1316289455,0.217995435,0.1578475386,-0.1575408429,0.0378864035,0.0702383593,0.2949165404,0.239563942,0.2868806124,0.0837227032,0.0290413704,0.0989120528,-0.0350016467,0.1151226014,0.0313579477,-0.0286040809,0.2699772716,0.0538867675,0.0568375029,-0.2583304048,-0.1918704808,0.021856701,0.3623172343,-0.1548786163,-0.0067171738,-0.2401380241,-0.1583110243,0.0296415146,0.1109195799,-0.0863590762,-0.3013877571,0.2298639417,-0.0257336106,-0.2985094786,0.2507893145,-0.0469851755,0.0732867792,0.2300786972,-0.0003457732,-0.1248533428,-0.243791461,-0.2222489715,0.2859991193,0.0418851711,0.3473754525,0.2189565301,-0.2635963857,-0.1986075342,0.0423633642,-0.3450841606,0.0556564778,-0.102708973,0.3029245734,0.1472289264,0.3699688911,-0.3311337233,-0.2749993205,0.2255137563,-0.1373625845,-0.2359548062,0.0741085559,0.1343904436,0.0035222515,-0.2461089492,-0.4167608321,-0.31868577,-0.3201167583,-0.1698961109,0.0469901823,-0.0084044393,0.1322675496,0.3167234659,0.0521147139,-0.0223643333,0.1049882844,-0.1745298356,-0.3384003043,0.3079465628,-0.1925238371,-0.3766245246,0.1279185414,-0.1847530603,0.306612134,-0.254062742,-0.4626396,-0.1359268576,-0.0298976116,-0.0540175475,-0.0890751779,0.0387769192,0.1308973432,0.0678489432,-0.2722516358,-0.1435526013,-0.0076316292,0.0690955967,-0.0854022428,0.2879109383,-0.3592693508,0.3144870698,-0.0002528063,0.2587943673,0.242315203,-0.147080183,0.1402680427,-0.1283783764,0.1531906724,-0.1803449094,-0.4023916721,0.0913544521,0.0497151017,-0.0440250598,0.1417559981,-0.0282214358,-0.1757350564,-0.2262155414,0.1929253638,-0.2730834484,-0.1948021203,0.0201747436,0.0151661411,0.0430835038,-0.0608068816,0.077146709,-0.2662429214,-0.0004526147,-0.092808798,0.1869510263,-0.0280966945,-0.0561328605,-0.5485401154,0.2478294075,-0.2951685786,0.100545086,-0.1098497957,0.070332624,-0.1313753575,0.0816552937,-0.0624933988,0.0440219827,0.5133491755,-0.2394293249,-0.0786851719,0.0392570086,0.1794705242,-0.1437889338,0.1132250875,-0.3266349733,-0.1082744971,0.2915794849,0.0095197195,-0.2043699473,-0.046947822,0.3059433699,0.0892328992,-0.1894133389,-0.0858808383,-0.3344301283,-0.3493281901,-0.2633576393,-0.0007080277,0.2335875779,0.1501875222,-0.1447902918,-0.1981935352,-0.0966329575,-0.036843162,0.3144248128,0.127920866,0.3973326683,0.0410998054,0.1794861257,-0.1315343529,0.2101925761,0.3389058709,0.4087629616,-0.0568696521,-0.0976154804,0.1304517984,-0.2605876327,0.2948584259,0.0800035819,0.0388831124,0.1459211111,-0.3462323248,0.2745558321,-0.0908023342,0.3677177131,0.4316271842,0.1831064373,-0.2209733576,-0.266283989,0.2885432839,-0.018535722,0.0534063689,0.1570128649,0.1279863417,-0.3031210303,0.4539655447,-0.000090308,0.8482797742,0.2338560969,0.0953580141,0.2726929486,0.0893241987,0.2085202336,-0.0704030171,0.1366905272,-0.1685815305,-0.112199448,-0.0238081664,-0.0332794562,0.0855829716,0.3287417591,-0.311324507,0.0973818824,-0.1732923537,0.1516099274,-0.1335392147,0.0897617936,-0.0003263683,0.1290322393,-0.1407247335,0.3338052928,-0.0387004986,0.1216558293,-0.0947829708,-0.0934407264,-0.1199554205,-0.1500315368,-0.0232508499,-0.0091101741,0.1293339133,0.206850931,-0.0860598236,-0.3148771524,0.0286952835,0.2329126447,0.2909002602,0.243552357,0.0292561296,0.3258609772,0.301759243,0.0451742634,-0.0190018341,0.1354712397,0.2342634797,-0.1793263704,-0.3946053982,0.2025227994,-0.0998145789,-0.2074992657,-0.0870115981,-0.1002438962,0.0387293659,-0.4835878015,-0.197048977,0.1472851038,0.0120593105,-0.287594676,0.3185055554,-0.0215437766,-0.2567573786,0.147907123,0.0393999964,-0.1561187506,-0.0637795106,0.1460044831,0.0951265022,-0.1999863386,0.6061750054,-0.1075160652,-0.1518532634,-0.4008369446,0.0688672885,0.2677804232,-0.3163879812,0.2189088166,-0.0548601411,0.0077233268,0.062240947,0.2127064317,0.0940153673,0.0704409033,0.0955473483,-0.2247422189,-0.4444610178,0.2742581666,-0.0529290289,0.3093001842,-0.0958635584,0.1811676919,0.0260560624,-0.014212979,-0.5368689299,0.0945702046,-0.4050330222,0.1345218718,-0.1684113443,-0.2167639285,-0.1456323117,-0.1639456749,0.3472725749,0.036672771,-0.0651461259,-0.4538067281,-0.0199015439,0.035038013,-0.0185412932,-0.1189372167,0.1891777366,0.0076570436,-0.3035958707,-0.0127255842,0.0243536234,0.0795274898,0.080250375,0.2908804119,0.188194558,0.1708170623,-0.1665554047,-0.2071214616,0.0984644219,0.0631441623,0.0563477837,0.0034731431,0.0193798114,-0.0138189038,0.1760663688,0.0360364802,-0.0571162216,0.0308247358,0.2885244787,-0.3951023817,0.1735807061,0.042241253,0.1654189676,0.1357655823,-0.4708945155,0.1168869883,0.344905138,0.4896976948,-0.4784144759,0.0553887561,0.1259483397,0.2000157982,0.2356615067,0.0114221135,-0.0590032265,0.0051942677,-0.0602955259,0.3224766552,0.101981923,-0.1183427125,0.1651211679,0.6090922356,-0.1725051403,-0.031126976,-0.0928597823,0.0774212852,0.0199709628,0.5830510855,-0.004162339,0.5754637718,-0.0340071432,0.0469603464,-0.0318826288,-0.1659197956,-0.0541259423,0.1469039619,-0.1344867945,0.1831884384,0.1293453872,0.1463635862,-0.1289156079,-0.3962649703,-0.2750422955,0.0502915718,-0.3425110281,-0.0458284803,-0.435800761,-0.0862511918,-0.2428584695,-0.0162065402,0.077118732,-0.2512909174,0.0255448353,0.0292144418,-0.1950488687,-0.1913188696,-0.1978898644,0.1151302606,0.1773782372,-0.0726195946,0.0984161869,0.4463357925,-0.1606779695,0.068149142,0.34711954,0.366761148,0.184931308,-0.0674304962,0.0616979189,0.0482148677,-0.2647487819,-0.0619952269,0.2996995449,0.2890268266,0.0464400649,0.2956807017,0.3151909113,-0.3501691818,0.0766294375,0.1603610665,0.1473896354,-0.2927682996,0.1704170853,0.0011580101,-0.0200375002,-0.3928390741,-0.2031603158,-0.3765420318,-0.0609580092,0.4621415436,0.0733076707,0.127336815,-0.0972785056,0.1736003608,-0.1397905499,0.2555043995,0.210987404,0.2537482977,-0.3867299557,-0.2867272496,-0.5975883603,0.0897155181,-0.2448736429,-0.1250065267,-0.0841022432,-0.0061890315,0.0108968681,0.1621705592,-0.0555657297,0.1872788072,-0.0163139068,0.1272355765,-0.5106663704,-0.1320037693,-0.2302888632,0.0923608392,0.0824379548,-0.3296445906,0.0360213816,-0.1105762348,0.2333438993,0.119439967,-0.0284429267,-0.0081256414,0.3200005591,0.152040571,0.0456842855,0.447648257,-0.0517545156,0.0579842851,-0.0447141603,-0.4001832604,-0.1197107285,0.1868249327,0.4128521979,0.4482226968,-0.0364909656,0.1490909308,0.0196347032,0.0350223295,0.3146042824,-0.3895193636,-0.4056879282,0.1191590354,0.2422042787,0.1679243892,0.0857926458,0.1925950795,-0.0515476502,-0.0481506288,-0.2457739413,-0.4889572561,0.481797874,-0.3192516565,-0.2032634318,-0.290134728,0.0802739114,0.3635623157,-0.0708549023,-0.6132722497,-0.0189178418,0.2579334378,-0.0174567997,-0.2655264437,0.2674396932,-0.0597252809,-0.0127943214,0.0020227807,0.1840706915,0.0970245153,-0.0951152742,0.0696482956,-0.3605858982]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1686","title":"Dataset Error: DaNE contains empty samples at the end","comments":"One the PR is merged the fix will be available in the next release of `datasets`.\r\n\r\nIf you don't want to wait the next release you can still load the script from the master branch with\r\n\r\n```python\r\nload_dataset(\"dane\", script_version=\"master\")\r\n```","body":"The dataset DaNE, contains empty samples at the end. It is naturally easy to remove using a filter but should probably not be there, to begin with as it can cause errors.\r\n\r\n```python\r\n>>> import datasets\r\n[...]\r\n>>> dataset = datasets.load_dataset(\"dane\")\r\n[...]\r\n>>> dataset[\"test\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n>>> dataset[\"train\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n```\r\n\r\nBest,\r\nKenneth","comment_length":40,"text":"Dataset Error: DaNE contains empty samples at the end\nThe dataset DaNE, contains empty samples at the end. It is naturally easy to remove using a filter but should probably not be there, to begin with as it can cause errors.\r\n\r\n```python\r\n>>> import datasets\r\n[...]\r\n>>> dataset = datasets.load_dataset(\"dane\")\r\n[...]\r\n>>> dataset[\"test\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n>>> dataset[\"train\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n```\r\n\r\nBest,\r\nKenneth\nOne the PR is merged the fix will be available in the next release of `datasets`.\r\n\r\nIf you don't want to wait the next release you can still load the script from the master branch with\r\n\r\n```python\r\nload_dataset(\"dane\", script_version=\"master\")\r\n```","embeddings":[-0.1232687905,-0.1505773515,-0.2077827305,-0.0634229407,0.2767205834,0.1347071677,0.3861126602,0.3464493752,0.2240535468,0.2550488412,0.1427296549,0.2462115139,-0.1021410227,0.1205538809,-0.0594546609,-0.189043507,0.0712748617,0.2300376892,-0.2116483599,-0.2163271606,-0.1598127633,0.0545276366,-0.2814107239,-0.0294996072,-0.2529697418,-0.0086915977,-0.0732685253,0.000903021,-0.181566745,-0.339361459,0.1875905842,0.0553387143,-0.0221887324,0.472243011,-0.0000967098,0.0357762761,0.2746002972,0.0090655759,-0.2073474526,-0.1868923903,-0.2046446502,-0.1032754928,0.016476728,-0.2946446836,-0.0948273465,-0.119078815,-0.1217465252,-0.2512132227,0.2761532068,0.3788674176,0.3510321975,-0.0203607064,-0.1133409366,-0.233607173,0.0938120931,0.0186805166,0.0067269937,0.0080389073,0.1170281321,-0.158007279,0.1715448797,0.3733156025,-0.1356579363,-0.0398959704,0.1303146333,0.0744022503,0.3552424908,-0.396034658,0.1225465909,0.1647951901,0.476218909,-0.2359470278,-0.2052012682,-0.095294483,0.0928829834,-0.3406413198,0.0278559774,0.0826479718,-0.1892317683,0.1356574893,-0.1832286417,0.1252121031,-0.0296160169,0.1406727582,-0.22200495,0.2924418449,0.1040622592,0.0105551416,0.0352573544,-0.1616911292,-0.0072539877,-0.1755838096,-0.0992462188,0.1262255907,-0.1759640276,-0.178135559,0.1104308814,-0.3320736587,0.0922228247,0.2341314256,0.0549229272,-0.0728612691,0.0724771842,0.2050118595,0.3890073001,0.1582598388,0.1183600649,0.1860062629,0.0431001447,0.0392940342,-0.0792645365,0.0688089505,-0.0300588105,-0.3939440846,0.2165605873,0.0729623586,0.1235312074,-0.0328079984,-0.3163638711,0.1423445195,0.0572096668,0.105277501,-0.0099749016,0.3338144422,-0.073283188,0.1919693947,-0.0296997838,0.1066347957,-0.1325368583,0.0875127912,-0.3342106938,0.1476406306,-0.1924995184,-0.4847509265,0.1393510848,0.1399340928,0.259888798,0.1122575328,-0.0875671804,0.0437997393,0.2365788519,-0.2299486399,0.1596099138,0.2711679637,0.2072999328,0.1032985076,0.1049730405,-0.3105363846,-0.0968494788,0.1971830428,-0.2414344996,-0.1270575225,-0.2391119301,0.4493574202,-0.0192511845,-0.1367019564,-0.3514400125,0.1592812836,0.0225653257,-0.1922840178,-0.0744035766,-0.1213260442,-0.0971533135,-0.1482298672,0.3134791553,0.2801155448,-0.3103978038,-0.1445904672,0.0392432958,-0.1823048592,0.1620915234,0.2586766481,-0.1053339988,0.1341364384,-0.2297564,0.1065836996,0.3605087101,-0.3388747275,-0.1952200383,0.1234817058,-0.0189936142,-0.0715915188,-0.1197774783,-0.0189193822,0.0368326977,0.1500175595,0.1854106933,0.2362961918,-0.0505880937,-0.0243352931,-0.1854705513,-0.0553223044,0.2880092561,0.1449862272,0.09852916,0.035712488,-0.142558381,0.1447718889,0.2507312894,-0.0788625032,0.0050781532,0.4619210958,0.4397904277,0.0230416376,-0.0082344152,-0.0932403952,-0.1799519658,0.0302312542,0.1275368333,0.2458103448,-0.0125913695,-0.1820110828,-0.2893087268,-0.1439445317,-0.2131991386,-0.1518132538,0.3282752931,0.0149561418,-0.0991956368,0.0585254207,-0.2611064017,0.1758178025,-0.1464342922,0.1212603897,-0.2127426416,0.3394411504,-0.2513133585,0.0080370512,-0.0149493618,0.129990235,0.117881909,0.0251401346,-0.0241430234,0.259624362,0.1666731536,-0.0331909545,0.1301199645,-0.1600823551,0.0945522934,-0.1683371812,-0.1287045479,0.3237847686,0.1293809712,-0.0326088183,-0.2479771376,0.1031958833,0.106739305,0.1904259771,0.1538407505,0.0503829904,0.1857273877,-0.2024238408,0.1165678501,-0.2406700701,0.1367219537,-0.1152270287,0.1116154268,-0.0330059007,-0.1394121051,-0.0811002851,0.3397686183,0.0697710812,0.2045668215,0.1477720141,-0.1715923399,0.0466526486,0.0244666263,0.3108148575,0.2799334824,0.310554564,0.0669835135,0.0863323882,0.0967035815,-0.0259545837,0.1524069607,0.0576315708,-0.0488425605,0.267711997,0.0449269712,0.0368419029,-0.2552139163,-0.2440847903,0.0002030683,0.3363626897,-0.1759390831,-0.0267413799,-0.3138723075,-0.1280578077,0.0148754241,0.1995943338,-0.0780418813,-0.3177387714,0.2236334085,0.0047461521,-0.3061591089,0.2955164909,-0.0162079223,0.0966826603,0.1927609295,-0.0293254815,-0.1610547006,-0.2629687488,-0.234137401,0.2838199437,0.0835726261,0.2928678393,0.2347479761,-0.2635474801,-0.1672602892,-0.043928232,-0.4346902072,0.049269069,-0.1462500989,0.2544342577,0.2226761281,0.3057717383,-0.2177507579,-0.2630028725,0.2211736292,-0.2390861958,-0.2774768174,0.0902240425,0.1239492893,-0.0732543916,-0.2961527407,-0.5132988691,-0.2596700191,-0.3618588448,-0.1057882681,0.1025441512,-0.0004766732,0.0969894305,0.3217784166,0.0193617698,-0.0249998663,0.0289331451,-0.1692986488,-0.2669264674,0.3266661763,-0.2702935934,-0.3725543916,0.1384802908,-0.1809381992,0.2575161457,-0.2692220509,-0.4536553323,-0.1316346526,-0.0194885172,-0.0114393728,-0.0574130863,0.0379805006,0.1411522925,0.134551093,-0.2727219164,-0.1700938195,-0.0769099072,0.0196130797,-0.1201296896,0.2499282509,-0.2742264867,0.3223972023,0.0168702379,0.2349817008,0.2456393689,-0.175684765,0.073930636,-0.1861131936,0.2673369348,-0.1300089657,-0.3965242803,0.0441881865,0.0597977117,0.0294842925,0.1088710353,-0.0278626941,-0.1776435375,-0.2493134737,0.186298728,-0.1851068884,-0.1896832436,0.0200834088,0.0145738609,-0.0138831865,-0.0301512796,0.1286458969,-0.3162184954,0.0096851774,-0.0390825532,0.1935116798,-0.045454517,-0.0527265966,-0.5001696944,0.2549444735,-0.345723629,0.1618809253,-0.0631518066,0.0598971508,-0.0876944289,0.044563923,-0.104246527,0.001842959,0.5315636992,-0.3192275167,-0.0705184639,0.0981983095,0.2009631842,-0.1757179648,0.0994821191,-0.3373843431,-0.1074302793,0.2964458764,0.0823939219,-0.2276563644,-0.1090291589,0.3578347862,0.0679270253,-0.1750819534,-0.1112864837,-0.3121805191,-0.353567034,-0.351639986,-0.0649890527,0.2422629744,0.1737551838,-0.2100568861,-0.211878106,-0.0522688329,-0.0621638559,0.312102288,0.1146017537,0.4049684703,0.0548356324,0.1624941379,-0.1140965819,0.1351618022,0.3803835511,0.4657444358,-0.049191542,-0.2040527612,0.1282796711,-0.1826682985,0.3136865199,0.0569626838,-0.0169248991,0.0337806344,-0.3138438761,0.2500200868,-0.1137787029,0.3845151961,0.4363157451,0.1238589585,-0.2079488784,-0.3572323322,0.284261167,0.052402176,0.0299960356,0.1809986234,0.0503299534,-0.3033679426,0.4107021689,-0.0137907388,0.8179671764,0.2246252447,0.094548136,0.2602621913,0.028795762,0.2276393771,-0.1138888225,0.09623117,-0.2584106922,-0.0580951199,-0.0227292683,-0.0341435522,0.0731226727,0.3025133312,-0.3570846617,0.1366188824,-0.1738310158,0.1611589938,-0.1448457092,0.2015021443,-0.0203194246,0.0394432731,-0.1272226572,0.3340436518,-0.0134764789,0.1012718976,-0.0598256737,-0.1300127357,-0.0954313204,-0.1965573728,-0.0335769206,0.0927530378,0.1977063566,0.2040991634,-0.1111574173,-0.2627123594,0.0110584833,0.28479895,0.2877002656,0.2532496452,0.0456558652,0.3643516302,0.2367525995,0.0686033964,0.0154254958,0.1436347812,0.2437866926,-0.157482326,-0.3793878257,0.1241312325,-0.0492140651,-0.238768205,-0.0866311416,-0.0765679032,0.0707118884,-0.469509691,-0.1692463905,0.1563895643,-0.021885965,-0.3271130621,0.3178136945,0.0043692309,-0.2989166081,0.1441240758,0.0638882294,-0.1407249421,-0.0377772748,0.2160801291,0.1302144527,-0.2545407712,0.5490442514,-0.0538430288,-0.1450417787,-0.389069289,0.0515379608,0.3508819044,-0.3129739463,0.2727482915,-0.014698226,0.0150267975,0.134629339,0.1828005463,-0.0136690792,0.0424894765,0.0678109899,-0.2157609165,-0.4011380076,0.2692801356,-0.039416682,0.3781545758,0.0293614101,0.1548749059,-0.037504904,-0.0215579998,-0.5434963107,0.0568274856,-0.3389387429,0.0999237895,-0.1396113932,-0.1531351805,-0.1597623527,-0.1850288361,0.3447807133,0.0766849145,-0.0454825088,-0.4464409649,-0.02671979,0.0371738784,-0.0955550522,-0.0747711957,0.1484639645,0.0051724473,-0.3181422651,-0.0577873252,0.0885732919,0.0849686787,0.0727113634,0.3011941612,0.1814303845,0.1627578586,-0.086352244,-0.1732787341,0.1043921262,0.1138423756,-0.0053466344,0.0460276194,-0.0707834885,-0.0294746589,0.2215913385,0.0282799397,0.0247881133,0.1064960212,0.3007612824,-0.4033624828,0.1334044039,0.0323999859,0.2279305756,0.211489141,-0.4869907796,0.1296071857,0.2989042401,0.453907758,-0.4949323237,0.0021950048,0.1348945051,0.2098198831,0.214147687,-0.0227024425,-0.020530276,0.0191045851,-0.0242959708,0.2847110927,0.1492114365,-0.18376863,0.2255083621,0.5618138909,-0.1565525234,-0.0521000586,-0.0835756212,0.119702585,0.0532143004,0.6409173608,-0.0213495009,0.5188952684,-0.0208637547,0.05305494,0.0179419592,-0.1945541054,0.0241334196,0.1158734709,-0.1578848064,0.1754607409,0.179054901,0.1270076185,-0.1554494202,-0.3245255947,-0.2416407168,0.0661976263,-0.3337956071,-0.0492084846,-0.4465231597,-0.0548964143,-0.2335655242,-0.0768387541,0.1006761342,-0.2483944297,0.0378010832,0.027534917,-0.1604023874,-0.1588710845,-0.1390854269,0.0429744944,0.188163802,-0.0569791123,0.1025161594,0.4437999129,-0.1285183132,0.0798109621,0.2866291404,0.439479351,0.2118879557,-0.0260807537,0.0607133061,0.019935783,-0.2901849151,-0.0570248701,0.3167178929,0.2316608131,0.0652734786,0.285494864,0.3337288499,-0.3265680969,0.0426011272,0.1756025106,0.1506907642,-0.2586565018,0.2594536543,0.0497828759,-0.058079686,-0.4016959369,-0.1670692414,-0.3539691865,-0.1325335205,0.5236925483,0.1074893326,0.1265448332,-0.2104561776,0.185171932,-0.0609450974,0.2532323301,0.1499818861,0.15763475,-0.3179418445,-0.2323916256,-0.6231598854,0.0902824774,-0.165164426,-0.1386429518,-0.0792434588,0.0411353707,-0.0158721283,0.1360889822,-0.0043476559,0.0932962522,0.0431356691,0.0918102115,-0.4863611758,-0.1456211358,-0.1995879114,0.0900729969,0.1656545848,-0.3219249547,-0.0019834945,-0.1730484068,0.2241112292,0.0928028077,-0.0431549996,0.0421920978,0.2434081733,0.189623341,0.0440151133,0.4735169411,-0.1255483925,0.0708761439,-0.0993169174,-0.3905445039,-0.1866025925,0.2100106925,0.3252286613,0.3909318149,-0.0734648332,0.0594277941,-0.0229493268,0.0177038275,0.2530505061,-0.3288391232,-0.3871287405,0.1173000634,0.2303164899,0.1771806777,0.0450130105,0.2078751624,-0.0694295466,-0.0191220976,-0.2118053436,-0.5480454564,0.5345568657,-0.2942785323,-0.2483920455,-0.2666711509,0.075276792,0.2875222564,-0.0574494377,-0.6307256222,0.0487267636,0.225745365,0.0503624901,-0.2204595655,0.2630021274,-0.1538781673,0.0320738144,-0.0350458734,0.1697673798,0.1194176897,-0.0898677781,0.0852507427,-0.4025442898]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1686","title":"Dataset Error: DaNE contains empty samples at the end","comments":"If you have other questions feel free to reopen :) ","body":"The dataset DaNE, contains empty samples at the end. It is naturally easy to remove using a filter but should probably not be there, to begin with as it can cause errors.\r\n\r\n```python\r\n>>> import datasets\r\n[...]\r\n>>> dataset = datasets.load_dataset(\"dane\")\r\n[...]\r\n>>> dataset[\"test\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n>>> dataset[\"train\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n```\r\n\r\nBest,\r\nKenneth","comment_length":10,"text":"Dataset Error: DaNE contains empty samples at the end\nThe dataset DaNE, contains empty samples at the end. It is naturally easy to remove using a filter but should probably not be there, to begin with as it can cause errors.\r\n\r\n```python\r\n>>> import datasets\r\n[...]\r\n>>> dataset = datasets.load_dataset(\"dane\")\r\n[...]\r\n>>> dataset[\"test\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n>>> dataset[\"train\"][-1]\r\n{'dep_ids': [], 'dep_labels': [], 'lemmas': [], 'morph_tags': [], 'ner_tags': [], 'pos_tags': [], 'sent_id': '', 'text': '', 'tok_ids': [], 'tokens': []}\r\n```\r\n\r\nBest,\r\nKenneth\nIf you have other questions feel free to reopen :) ","embeddings":[-0.0087827081,-0.2528685927,-0.2576634884,0.0576860681,0.1794044226,0.1376597583,0.3596493006,0.2523090243,0.2156166434,0.2398332208,0.1235913783,0.2309537828,-0.1224744245,0.1184972078,-0.1270742565,-0.2640180886,0.0517120622,0.1811549366,-0.2238840312,-0.2953351438,-0.118509382,0.0087724617,-0.2490935624,-0.0385834575,-0.2775420249,-0.0959450081,-0.1200972572,-0.0129094552,-0.1990781575,-0.2694941163,0.188322708,0.0015739419,-0.0609245785,0.4917179048,-0.0000975056,0.0017022546,0.1622211039,0.0289291069,-0.1781413555,-0.1023186594,-0.2686603367,-0.0648631677,-0.0053314324,-0.3946037889,-0.0605807491,-0.103768453,-0.1588978171,-0.1989334524,0.4108237922,0.4010691643,0.3534663022,-0.0369922407,-0.1012689844,-0.1676898003,0.1319390237,-0.030133212,0.0374472626,0.0501694679,0.1013665497,-0.1148374006,0.1707385033,0.4196330011,-0.1159297451,0.0311602671,0.0510260388,0.0615610816,0.3688761592,-0.4099339843,0.1001946181,0.1819634438,0.4772837758,-0.1559392661,-0.1937507689,-0.0891528577,0.0080051869,-0.3031517565,-0.0022984985,0.0896999091,-0.1564149559,0.1526387036,-0.2052850574,0.1050892174,-0.0559000038,0.1347101927,-0.2636934221,0.3467005193,0.1006384119,-0.0175151769,-0.0205102656,-0.107763201,-0.0259352401,-0.182934314,-0.0089031886,0.1415729821,-0.156035617,-0.1763229668,0.111014016,-0.3220408559,0.0348874182,0.1592565924,0.0751841664,-0.1134902015,0.1059953272,0.21932742,0.3133505285,0.1490701586,0.017227415,0.2010683566,-0.0560698472,-0.0406455286,-0.1685321033,0.0778167397,-0.0533454083,-0.3706967533,0.2049519867,0.0763685629,0.1059745997,-0.0951959044,-0.2958321571,0.1806061715,-0.0530536883,0.0939997584,0.0362456255,0.3219203055,-0.0485755391,0.1078944802,-0.1208767369,0.1286029816,-0.0968275517,0.0788571909,-0.3413962126,0.1563995332,-0.189183712,-0.4768800735,0.0848740861,0.1796709448,0.2841749191,0.123243615,-0.1260395497,0.0168166496,0.2998917997,-0.3544846475,0.200521633,0.2558670938,0.0697595179,0.1479008049,0.1470105797,-0.2539596856,-0.1154209003,0.1417219788,-0.2468060106,-0.122458674,-0.1931499392,0.4696199298,0.0595177487,-0.1585585624,-0.3030904233,0.1590217948,0.0401956849,-0.1254237443,-0.0066904272,-0.0998815596,-0.0449821576,-0.18425183,0.2927543521,0.2386265397,-0.2530131936,-0.1020051613,0.0545462668,-0.1255522668,0.2371043265,0.289305836,-0.0711132735,0.1974641085,-0.1728929728,0.1437257379,0.3753038943,-0.2874373496,-0.1717739105,0.0850092769,-0.0275696944,-0.0287467837,-0.1171760708,-0.0410310142,-0.0163081754,0.1722483337,0.1418990791,0.2506377995,-0.0348552614,-0.0410547554,-0.2224924713,-0.0917345434,0.299679935,0.1869345456,0.126592502,0.0414937027,-0.1621229202,0.1079267859,0.2392853796,-0.0969982892,-0.0490484051,0.4101432264,0.4733175635,-0.0049493643,-0.0934816301,-0.0711248145,-0.0870098323,0.038273111,0.1196504906,0.2970600128,0.0151462685,-0.1416698992,-0.2871637344,-0.1827297509,-0.2078363597,-0.1048666313,0.3724482059,-0.0161990933,-0.0858469233,0.0948889703,-0.2158031464,0.1218946129,-0.1483168602,0.1267935634,-0.1084107235,0.313228786,-0.2740926743,0.0230492949,-0.0353167802,0.1532702446,0.132852152,0.0184629858,0.0279998723,0.2017269731,0.1320418864,-0.0360711589,0.1593159735,-0.1273363382,0.0647495985,-0.246217832,-0.1140120104,0.351655215,0.1303078979,-0.0129029807,-0.1886965036,0.100664407,0.1392619461,0.1955137849,0.1406377852,-0.0254850667,0.0817802548,-0.2165912539,0.1425827742,-0.1665834934,0.1881197244,-0.147393465,0.111858502,-0.0138847437,-0.1394920647,-0.0806574076,0.2845634818,0.1198689267,0.1990845948,0.1066892743,-0.1692792922,0.0505695976,0.0733407736,0.2940853834,0.2179565281,0.3106139898,0.151266098,0.0259634424,0.0198358577,-0.0003564683,0.1143894568,0.0595137626,-0.009459395,0.2395635992,0.0250093155,0.0597684979,-0.2623164952,-0.2347584218,-0.0482555255,0.325361222,-0.1502872556,0.0406979136,-0.2231764197,-0.1631902009,0.01019709,0.0840463489,-0.0832122862,-0.2975433469,0.22469531,-0.0553328954,-0.318711549,0.2492684722,-0.0490496233,0.0805435926,0.1914149672,-0.0229898095,-0.1252213717,-0.2333340943,-0.2079240233,0.3087169528,0.0446011983,0.3896153569,0.234034881,-0.2760981321,-0.1594332755,0.0684595257,-0.3692917824,0.0527573042,-0.138315171,0.2823215127,0.1188490093,0.3785992265,-0.3570213914,-0.2987491488,0.222328186,-0.1452458799,-0.2071302235,-0.0020421962,0.1549544483,-0.0080676191,-0.2175498605,-0.4752557576,-0.3175473213,-0.2985446453,-0.1227937713,0.0936222821,0.0306719169,0.1239099205,0.32659477,0.0782102048,-0.0123684872,0.1709992141,-0.1721755415,-0.3479922116,0.3115019798,-0.1935701817,-0.3593387902,0.0677169338,-0.1752615571,0.2893318832,-0.2256548405,-0.4472939372,-0.1465665847,-0.0239398181,-0.0559677631,-0.0756405741,0.0594547465,0.1131987497,0.1109491959,-0.2847737372,-0.1200758442,0.017602386,0.078210108,-0.0671762377,0.2682172656,-0.3446980417,0.2564875782,-0.0336771086,0.2492434382,0.2093647122,-0.1382615268,0.1677184701,-0.133941099,0.1469819397,-0.1307879686,-0.4095375538,0.0755980387,0.0104824556,-0.0461549163,0.1671387404,-0.0284304172,-0.1860110313,-0.2574830949,0.2011730373,-0.3035578728,-0.1724817008,0.0321546867,0.0207936093,0.0328282937,-0.0657130107,0.0483215339,-0.2690802813,0.0460009053,-0.0638225675,0.1801231802,-0.0397088043,-0.0601799935,-0.5665232539,0.2141902894,-0.3019243777,0.1203776523,-0.0930203497,-0.0038671612,-0.1174967289,0.0491410755,-0.0701529309,0.0004092084,0.4788865745,-0.2301075608,-0.0946066231,0.0356620066,0.2123615295,-0.1236984134,0.1448026448,-0.3667244613,-0.1372152269,0.3035237789,0.0281829163,-0.2356809676,-0.043769937,0.3208115697,0.1242141053,-0.168105945,-0.0766018406,-0.3198749721,-0.3513250351,-0.2911549509,0.0021395846,0.1964918524,0.181097284,-0.1606839597,-0.1845306605,-0.0641729981,-0.0032272134,0.3613939285,0.1569884568,0.3749766946,0.0123617994,0.106891267,-0.0869820565,0.2289482057,0.3257833123,0.4271508157,-0.0777539238,-0.0832742006,0.0742323324,-0.2451578677,0.3269693851,0.0732990503,0.0633613914,0.1179466695,-0.3316694498,0.2391514927,-0.1006749645,0.3783435822,0.4543425739,0.1790101081,-0.2172308564,-0.2906177044,0.285582602,-0.0150095848,-0.0008469737,0.1325833648,0.128169477,-0.3329293132,0.4500831962,0.0244302284,0.889890492,0.220629707,0.1113051623,0.2622612417,0.048582647,0.201104179,-0.0773344412,0.1124376506,-0.167997852,-0.122720927,-0.0483323224,-0.040488448,0.1187710613,0.359477371,-0.3046232164,0.1015611887,-0.1773266941,0.1907563359,-0.1729645133,0.102091752,0.0216703843,0.116254501,-0.1771526486,0.3393171728,-0.0001824149,0.0877291188,-0.1067018434,-0.0856093913,-0.1729564816,-0.1730306596,0.0256419145,-0.0195340049,0.1074651331,0.1967837214,-0.1298638731,-0.3096008599,-0.0051908349,0.2595408261,0.3156982362,0.1838159859,0.0416116081,0.2915273309,0.3210722804,0.0623942874,-0.01376931,0.0963217542,0.2553298771,-0.1677721888,-0.3701055944,0.1993849725,-0.0815807655,-0.2089570016,-0.1142246127,-0.0831497237,0.0306818038,-0.4577130675,-0.1546398252,0.2002801299,-0.0142162945,-0.2908188999,0.3242892921,-0.02571024,-0.244985655,0.1968368739,0.0521595776,-0.172662288,-0.0866321623,0.1433287561,0.1402669698,-0.194291696,0.5821241736,-0.1109672263,-0.1741411984,-0.4146704376,0.0611985438,0.2791421413,-0.2848740518,0.1960008293,-0.0702042058,-0.0287029743,0.040207535,0.2035458833,0.1208562329,0.0724744648,0.0649223626,-0.2137127817,-0.4518822432,0.3301442862,-0.0426453464,0.3445533514,-0.1134966835,0.1701533347,0.0020740391,0.0099245412,-0.5574768782,0.0807306692,-0.382907778,0.1232377291,-0.1394735724,-0.2097100466,-0.1324911118,-0.1601646692,0.3690021932,-0.0138543453,-0.0825749561,-0.4717026055,-0.0203014016,0.0303624645,-0.0196897984,-0.1406430751,0.1768588722,0.0068233092,-0.2591423988,-0.0225785896,0.0254020188,0.0738901868,0.1075764149,0.295928061,0.2089688629,0.1749445945,-0.1700543016,-0.2150179446,0.1107613966,0.0396129861,0.0560182184,-0.0070181503,-0.0346032381,-0.0661379471,0.220232293,0.0415780358,0.0018536649,0.0333853401,0.2651948631,-0.3506431878,0.149618417,0.0364343077,0.2110625505,0.1954626888,-0.4245160818,0.1244408935,0.2930161357,0.4895514846,-0.468247354,0.0615027845,0.1548897326,0.1271440387,0.2500427961,0.0388114937,-0.0682592019,0.0438774712,-0.0648984611,0.2865451872,0.0858797207,-0.1008341983,0.1428443193,0.5819150209,-0.1847387254,-0.0429146402,-0.054413043,0.1227537096,0.0590762384,0.553710103,-0.0107750092,0.5505725145,-0.04445526,0.0436747223,0.0109899584,-0.1032541394,-0.036962565,0.1635693312,-0.1095460281,0.1937700659,0.1498686075,0.1516484022,-0.1271873862,-0.4009828866,-0.2688106,0.0930766761,-0.2920280397,-0.100798279,-0.4344131351,-0.1309577823,-0.2581528425,-0.0565526932,0.0614262819,-0.2726224065,0.0385194868,0.0062200082,-0.1610476971,-0.1825039089,-0.1819554567,0.0613022484,0.1656577885,-0.0697031617,0.1065820083,0.4792703688,-0.1324564815,0.0883134604,0.3319242597,0.3408558667,0.1487443596,-0.1005847678,0.0587183982,0.0104224021,-0.2727804184,-0.0709117427,0.2689335644,0.2741424441,0.0162612461,0.332382232,0.3282858133,-0.3602384925,0.0857774392,0.1606110483,0.1352502704,-0.2674843669,0.1580325067,0.0087353913,-0.0336123705,-0.3335416019,-0.2084522545,-0.3918228447,-0.0752740353,0.4282594621,0.059537068,0.1389325112,-0.0717158765,0.1756020635,-0.1139597371,0.2647707462,0.1949827224,0.2815316021,-0.3879762292,-0.2716321051,-0.6381655335,0.0820890814,-0.2083566636,-0.1406586766,-0.0875541791,0.0072205341,0.0505715795,0.1724829376,-0.0298361946,0.1632436961,0.0053462996,0.1094117984,-0.5134953856,-0.1266054809,-0.2119727731,0.1242992654,0.0756983235,-0.2973659933,0.0396903157,-0.1272921264,0.2465234995,0.1043683141,-0.0511381179,0.0163112618,0.2909140885,0.1819716543,-0.0094297752,0.4871368706,-0.0971395746,0.0571655594,-0.0052275886,-0.3524596393,-0.1157449856,0.1712235361,0.3778580129,0.4640810192,-0.0465628728,0.1493870318,0.0536738448,0.064154543,0.3176574707,-0.3937202096,-0.3784159124,0.1026911661,0.2740381658,0.144763276,0.0548746064,0.2194031924,-0.035124965,-0.034621764,-0.2298497856,-0.485512048,0.5131320953,-0.2854625881,-0.1921599954,-0.2857821882,0.082039699,0.3764765561,-0.0321725942,-0.6157333255,-0.0097124372,0.2536058426,-0.0427849069,-0.2621743381,0.2792547941,-0.0362273045,0.0018955504,-0.0111546805,0.1586034298,0.062508598,-0.0544215664,0.0110630821,-0.3584385514]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1683","title":"`ArrowInvalid` occurs while running `Dataset.map()` function for DPRContext","comments":"Looks like the mapping function returns a dictionary with a 768-dim array in the `embeddings` field. Since the map is batched, we actually expect the `embeddings` field to be an array of shape (batch_size, 768) to have one embedding per example in the batch.\r\n\r\nTo fix that can you try to remove one of the `[0]` ? In my opinion you only need one of them, not two.","body":"It seems to fail the final batch ):\r\n\r\nsteps to reproduce:\r\n```\r\nfrom datasets import load_dataset\r\nfrom elasticsearch import Elasticsearch\r\nimport torch\r\nfrom transformers import file_utils, set_seed\r\nfrom transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast\r\nMAX_SEQ_LENGTH = 256\r\nctx_encoder = DPRContextEncoder.from_pretrained(\"facebook\/dpr-ctx_encoder-single-nq-base\", cache_dir=\"..\/datasets\/\")\r\nctx_tokenizer = DPRContextEncoderTokenizerFast.from_pretrained(\r\n \"facebook\/dpr-ctx_encoder-single-nq-base\", \r\n cache_dir=\"..datasets\/\"\r\n)\r\n\r\ndataset = load_dataset('text', \r\n data_files='data\/raw\/ARC_Corpus.txt',\r\n cache_dir='..\/datasets')\r\n\r\ntorch.set_grad_enabled(False)\r\nds_with_embeddings = dataset.map(\r\n lambda example: {\r\n 'embeddings': ctx_encoder(\r\n **ctx_tokenizer(\r\n example[\"text\"], \r\n padding='max_length', \r\n truncation=True, \r\n max_length=MAX_SEQ_LENGTH,\r\n return_tensors=\"pt\"\r\n )\r\n )[0][0].numpy(),\r\n },\r\n batched=True,\r\n load_from_cache_file=False,\r\n batch_size=1000\r\n)\r\n```\r\nARC Corpus can be obtained from [here](https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/arc\/ARC-V1-Feb2018.zip)\r\n\r\nAnd then the error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-13-67d139bb2ed3> in <module>\r\n 14 batched=True,\r\n 15 load_from_cache_file=False,\r\n---> 16 batch_size=1000\r\n 17 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in <dictcomp>(.0)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1257 fn_kwargs=fn_kwargs,\r\n 1258 new_fingerprint=new_fingerprint,\r\n-> 1259 update_data=update_data,\r\n 1260 )\r\n 1261 else:\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 155 }\r\n 156 # apply actual function\r\n--> 157 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 158 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 159 # re-apply format to the output\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1526 if update_data:\r\n 1527 batch = cast_to_python_objects(batch)\r\n-> 1528 writer.write_batch(batch)\r\n 1529 if update_data:\r\n 1530 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 276 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 277 typed_sequence_examples[col] = typed_sequence\r\n--> 278 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 279 self.write_table(pa_table)\r\n 280 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_arrays()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.validate()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Column 1 named text expected length 768 but got length 1000\r\n```","comment_length":68,"text":"`ArrowInvalid` occurs while running `Dataset.map()` function for DPRContext\nIt seems to fail the final batch ):\r\n\r\nsteps to reproduce:\r\n```\r\nfrom datasets import load_dataset\r\nfrom elasticsearch import Elasticsearch\r\nimport torch\r\nfrom transformers import file_utils, set_seed\r\nfrom transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast\r\nMAX_SEQ_LENGTH = 256\r\nctx_encoder = DPRContextEncoder.from_pretrained(\"facebook\/dpr-ctx_encoder-single-nq-base\", cache_dir=\"..\/datasets\/\")\r\nctx_tokenizer = DPRContextEncoderTokenizerFast.from_pretrained(\r\n \"facebook\/dpr-ctx_encoder-single-nq-base\", \r\n cache_dir=\"..datasets\/\"\r\n)\r\n\r\ndataset = load_dataset('text', \r\n data_files='data\/raw\/ARC_Corpus.txt',\r\n cache_dir='..\/datasets')\r\n\r\ntorch.set_grad_enabled(False)\r\nds_with_embeddings = dataset.map(\r\n lambda example: {\r\n 'embeddings': ctx_encoder(\r\n **ctx_tokenizer(\r\n example[\"text\"], \r\n padding='max_length', \r\n truncation=True, \r\n max_length=MAX_SEQ_LENGTH,\r\n return_tensors=\"pt\"\r\n )\r\n )[0][0].numpy(),\r\n },\r\n batched=True,\r\n load_from_cache_file=False,\r\n batch_size=1000\r\n)\r\n```\r\nARC Corpus can be obtained from [here](https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/arc\/ARC-V1-Feb2018.zip)\r\n\r\nAnd then the error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-13-67d139bb2ed3> in <module>\r\n 14 batched=True,\r\n 15 load_from_cache_file=False,\r\n---> 16 batch_size=1000\r\n 17 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in <dictcomp>(.0)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1257 fn_kwargs=fn_kwargs,\r\n 1258 new_fingerprint=new_fingerprint,\r\n-> 1259 update_data=update_data,\r\n 1260 )\r\n 1261 else:\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 155 }\r\n 156 # apply actual function\r\n--> 157 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 158 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 159 # re-apply format to the output\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1526 if update_data:\r\n 1527 batch = cast_to_python_objects(batch)\r\n-> 1528 writer.write_batch(batch)\r\n 1529 if update_data:\r\n 1530 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 276 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 277 typed_sequence_examples[col] = typed_sequence\r\n--> 278 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 279 self.write_table(pa_table)\r\n 280 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_arrays()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.validate()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Column 1 named text expected length 768 but got length 1000\r\n```\nLooks like the mapping function returns a dictionary with a 768-dim array in the `embeddings` field. Since the map is batched, we actually expect the `embeddings` field to be an array of shape (batch_size, 768) to have one embedding per example in the batch.\r\n\r\nTo fix that can you try to remove one of the `[0]` ? In my opinion you only need one of them, not two.","embeddings":[-0.5113657713,-0.340737462,-0.1262720525,0.0608496368,0.0776377022,0.1417664886,-0.018712217,0.2698065937,-0.2622149587,0.0491960905,0.1108063161,0.5883788466,-0.0307827368,-0.1317675114,-0.4038944542,-0.0905743986,-0.0616570264,0.1219060645,0.2424558401,-0.089935638,-0.3668372035,0.0883089751,-0.2699972093,-0.1003229246,-0.1718577743,-0.2321972847,0.2313919067,-0.0377138183,0.003301251,-0.6815324426,0.6883983016,-0.2978188097,0.2429684848,0.1627676785,-0.0001183986,0.2131496072,0.4500014782,0.0111980177,-0.0065675713,0.080258362,-0.2403251082,0.194777444,0.0855849758,-0.1638190746,0.0858565792,-0.0353686213,-0.0441782735,-0.427708596,0.3808458447,0.195334971,0.1014310941,0.2174219787,0.0514748059,-0.0151883606,-0.0687452629,0.1960485876,-0.1333302259,-0.2255924344,0.2797706425,0.1810365021,0.1458378285,0.5812528729,-0.0482616462,0.2692404985,0.1576283872,-0.1972575933,0.0579736903,-0.3331651092,0.1915374994,-0.1121736318,0.2094191313,-0.2229378819,-0.2514075637,0.1366557777,-0.1958025694,-0.3983700573,0.1623699516,-0.1368175,0.051956825,0.036535617,-0.1455416232,0.139839381,-0.2271478921,0.24815014,-0.1595502347,0.158727631,-0.0363634005,0.3123621643,-0.0161653459,-0.1319190264,0.1796032041,0.1299351603,-0.1233132333,0.0910461843,-0.2346746773,-0.1503433883,-0.2761262655,-0.3692839444,0.0213410761,-0.1542137265,-0.1089864597,0.1875861883,0.540751636,0.1412303448,0.0859375447,0.1132425293,0.242615208,0.8317361474,-0.0535715595,-0.6531259418,0.0678203776,0.0701442957,-0.0338044018,-0.3524084985,0.256862551,0.1501419544,0.1474722475,0.0349667855,-0.279946506,0.1407098919,-0.438165009,0.0077407854,-0.2134210616,0.3998953104,0.2957156003,-0.1497726738,0.2224410474,0.1331149936,0.1421479881,-0.1057698429,-0.0059415619,-0.0893860981,-0.1892879009,-0.1479352415,0.1425799429,-0.1967147738,0.2172679454,-0.2264876217,-0.0113379322,-0.0850389674,0.3367851079,0.0655955449,0.4326565266,0.2180086821,0.0231516995,0.2297055721,-0.0104063805,0.1910956949,-0.1135064363,-0.0793193206,-0.1578666717,-0.3309911489,-0.3654819429,0.1561720073,0.3952356279,-0.028295733,-0.0704108402,0.0058605992,0.5582671165,-0.0122786565,0.3374227583,-0.2378217578,0.1193173453,0.0060363831,-0.0863270462,0.184869349,-0.3499477208,-0.0525056981,0.0317243487,0.280678153,0.0482468344,0.0491322652,-0.3638934791,0.0915571451,-0.1223412901,0.5995523334,0.4406459033,-0.7461100221,-0.3960719705,0.5548933744,0.0000231852,-0.2747846246,-0.0202935971,-0.0612332262,0.4194323719,-0.1286513954,0.1339125484,0.0815142617,-0.0028018788,0.0602680556,-0.0769046396,0.0175310653,0.2058888823,0.130353868,0.0866857395,-0.0168716852,0.2668988109,-0.0338122919,-0.0874593481,0.0137310801,0.3313800395,0.4629959762,-0.1180306226,-0.1043538675,0.2504345775,0.0164172519,-0.3472789526,0.1601376832,-0.1680821031,-0.0168550368,-0.343441695,-0.2692348957,-0.1579743922,0.1876064539,-0.3379364312,-0.2994133234,0.0545108058,0.0237101484,-0.023135744,-0.1506970078,-0.2348758578,-0.3109050393,-0.0031263814,0.0129285213,-0.371036768,-0.081058912,-0.1028760746,-0.1666499376,-0.2229582369,-0.0823807046,0.1944281906,-0.1868701428,-0.2059344798,0.3810172379,-0.3576883078,0.1575224847,-0.2036373615,0.0353445709,0.0965872109,-0.7572340965,-0.1935379505,-0.0853928477,0.0226561949,-0.1436367929,0.2973734736,0.0332045108,0.4531671405,0.32563591,-0.2301132083,0.0423419885,0.1064608321,-0.0633365661,-0.1695415229,-0.2364452034,-0.1860896647,-0.1981634051,0.1607283205,-0.0286631752,-0.1010631099,0.033602152,0.1005503759,0.032848306,0.051831156,0.0886938199,-0.3016294837,0.1139409393,0.2114739269,0.1100961193,0.1420832723,0.0793418661,0.0846390128,0.0585502237,-0.0404893346,0.0700683594,0.3686750531,0.2179716974,-0.0281671304,0.257704705,0.0106333932,0.315379262,-0.256785363,-0.225231126,0.2095829397,0.0708614513,-0.2471293062,0.1702718586,-0.1032790393,-0.1220331639,0.0348452218,-0.1799590886,0.2601111531,-0.0945540741,-0.0504049994,0.0102419816,-0.061194364,0.4547729194,-0.0581415668,-0.0782623664,0.3865163326,-0.1864851266,-0.2859807014,-0.1884989291,-0.2462268472,-0.1040349305,0.0507002547,0.0092101879,0.2503574491,0.1913103312,0.0721683577,-0.3318425417,-0.2322167158,0.3374351859,-0.0510996468,-0.3957129419,0.1137748212,-0.2494402528,-0.382923454,-0.4515089691,0.1395195872,0.1534736603,-0.2197473049,0.1152105033,-0.1087608859,-0.22306256,0.1489222348,-0.5410323143,-0.2287052721,-0.2145576924,-0.0828818455,-0.3405638635,-0.0386743769,0.182748735,0.1774793565,0.0308759715,-0.0209289491,-0.0506052971,-0.1819638908,0.199599281,0.2329193056,-0.3783580363,0.0198900681,-0.0408562608,-0.1620057225,0.3158453703,0.5299435854,-0.1742092669,0.4087524414,-0.0935854092,0.0415871069,-0.3694301844,0.1419584006,0.5250942707,0.2435423434,-0.0890006721,-0.0797964483,0.1214501634,0.0565610826,-0.1646126956,0.2331374139,-0.1432997137,0.5282084942,0.0230465196,0.7464090586,0.1865098327,-0.6531713605,0.3472135961,-0.0085074985,-0.0273385886,0.0829499215,-0.1887917072,-0.094296217,-0.1831742972,0.080202572,-0.1202504113,-0.1443750709,-0.5767668486,0.1931263804,0.0903160498,-0.4708704948,0.0725259855,0.0098931044,-0.2897835374,0.4527910948,0.0263338014,0.1306546777,-0.0530324057,0.0912237614,0.1545127034,-0.2886983156,0.0147817573,-0.0437882356,-0.0654516518,-0.2173924893,-0.367125541,0.4868406951,0.2577666938,0.3479820788,-0.11464414,0.1626833528,0.0557988808,0.1853664666,0.4528127611,-0.4465717971,0.1091212779,0.0074966587,-0.1906968951,-0.2724784315,0.094267413,-0.0605693534,0.5769729614,0.4145486057,0.4333523214,-0.1463117748,-0.0121417092,0.0991358161,-0.2778380215,0.0614904016,-0.0091800224,0.0020793576,-0.1659399271,-0.2891857326,0.074616313,0.5696228743,0.0475608967,0.3033523858,-0.3070929348,-0.5790467858,-0.0392232463,-0.226955846,0.0006710388,0.2106073201,-0.0364595689,0.1529902518,-0.0096756527,0.3127455115,0.5661595464,0.1893276125,-0.0359560363,-0.1788259149,0.1438657343,-0.0896839425,0.0997157767,0.1143462583,-0.1353585869,-0.1474956274,0.0493534021,-0.0527137257,-0.0971566215,-0.0111675179,0.281041801,0.133730337,-0.3588031828,0.0294272844,0.2876640558,-0.1065136045,-0.022147404,0.5833806396,0.1208186522,-0.3441368341,0.5862824917,0.4361251891,0.6721884608,0.2287106067,-0.1058533788,-0.0988228917,0.2796190381,-0.0167952534,-0.0684407726,0.161393106,-0.2341504097,0.1617098749,-0.0667752624,-0.1077655181,-0.0533003174,0.265442431,-0.2107126415,-0.1009784192,0.1353304088,-0.2278293371,-0.0918503553,0.0910272747,-0.0353779607,-0.2407876402,0.2644459009,0.078892611,0.1308480203,0.1920498908,-0.111443907,-0.1472589523,0.0668359473,-0.3950742483,-0.233878687,0.0655947849,-0.4276450276,0.6017538905,0.1960137933,-0.1862182468,0.243661359,0.2843470275,0.2751020789,-0.2213450968,-0.0998323485,0.2911045253,0.4625076354,-0.2195213139,0.0896592662,-0.1184225976,0.1319423914,0.0171379577,-0.0814791471,0.2218269855,-0.1132743582,-0.383653909,0.1449813247,0.0136611583,-0.1129018962,-0.2023650259,0.0042820321,-0.0683048591,-0.3333327472,-0.1713193506,0.022209296,-0.1941104382,-0.1613932848,0.1137868688,0.0862243846,-0.3374771774,0.251973927,0.6712554097,-0.1556506604,-0.3268623352,0.4741314352,-0.0137802921,-0.1899153739,-0.2256557345,0.0754358768,0.5069118142,-0.5794172883,0.0827752575,-0.3235072792,0.0102473591,0.1571317315,0.5728722811,0.0356439389,0.3189882636,-0.1937944591,-0.0653920025,-0.2630267143,0.0230289586,-0.0755285397,0.4225475192,-0.1140358523,0.1403530538,-0.0659269243,0.0310643725,-0.212642163,-0.0343555883,-0.2249841988,0.0371422395,-0.1169101447,0.1496392787,0.0520565398,0.1060676053,-0.0331649072,0.4258903265,-0.1706050932,-0.0995732993,0.0934457034,0.1573872566,0.1690530479,-0.1038612425,-0.2477400601,-0.1066066623,-0.2969983816,-0.0554750152,0.1336551458,0.0964181572,-0.0245325211,0.1679580063,-0.0560057722,0.0031046285,0.0417178869,-0.2056293339,0.1120127961,0.1913034171,-0.0987461731,0.242951259,0.0712005496,-0.0102206515,-0.1499932855,-0.0407378189,0.0804265961,0.0391499847,0.2788161039,-0.4909853637,-0.2165833116,0.1341007054,0.1585979611,0.0885760412,-0.5477801561,-0.2336433083,-0.0200244058,0.2141940743,-0.0538470782,0.0383823253,-0.118472591,0.1434016228,0.0451569483,-0.028205758,0.1623191535,-0.0965995044,0.1533261985,0.2675468326,0.3809916675,-0.4468233883,0.0574765243,0.2488967478,-0.1121353135,0.0673063025,0.1593638957,0.3147460222,0.5915929079,0.3026495278,-0.1068798006,0.6875780821,0.2359990478,0.1713031679,0.0590547957,-0.0296096448,0.0559683926,-0.1344697475,0.1046848819,0.2520949841,0.2909655571,0.1931342483,-0.0982567817,-0.2824084163,0.0062947846,0.022751499,-0.1574299335,-0.1467564404,-0.3718587458,0.0348342098,0.0237774923,-0.0485217609,-0.1388556212,0.3094592094,0.2812197506,0.0936241448,-0.3840429485,-0.1180859134,0.1315281987,-0.0105849952,0.1634087116,-0.2039889991,0.2102388591,0.0912892297,-0.162929967,0.050826408,0.5249247551,0.387380898,0.2320964187,-0.3360838592,0.0496417992,0.0634767562,-0.0855809227,0.0324373469,0.2687185109,0.003969796,-0.0099392906,0.0981375873,0.0593787096,-0.0846748799,-0.0261173509,0.1643423289,0.4865007699,0.3900396526,-0.0844932497,-0.1798851043,-0.1179356873,0.0608870834,0.1091435999,-0.4464127123,0.1287889183,0.4633158147,-0.1877123564,0.0811693221,0.012980517,0.0467701368,-0.1753967851,0.4479142427,0.4729489684,-0.1310104281,-0.2427333295,-0.1254306138,-0.115588747,0.4575336576,-0.1092391312,-0.0273592565,0.3644365966,0.3641228378,0.1350461543,0.0312272143,0.2069369107,0.0161172766,-0.2688085139,0.4063840508,-0.0002255975,-0.0250777677,-0.0781739727,0.1112399697,0.1547817141,-0.2695218325,0.0635711923,0.2607866228,0.0267178547,-0.3894838095,-0.2669806778,0.1352346241,-0.259506166,0.6943307519,0.0299417879,0.0702526122,-0.2511700988,0.0753903538,-0.2062064111,-0.3413657844,0.0098226294,-0.0084072109,0.2645082474,0.274427861,-0.0939271152,-0.0926951319,-0.3260191679,-0.2867220044,-0.0668520108,-0.2104846388,-0.277595669,0.3050505519,-0.2955505252,0.1799634099,-0.1510111988,0.1711075306,0.2621454,0.2405257225,-0.3641981483,-0.588109076,0.5087835789,-0.3397665024,-0.3331579566,0.3033267856,0.0289909914,0.6112289429,0.1801222861,-0.537248075,0.0097092548,0.1904549301,-0.0897613615,-0.4189183414,-0.217498377,0.1601080447,0.0964678675,-0.099575378,0.3238032758,0.1050207615,0.1133849099,-0.2556917667,-0.3263609707]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1683","title":"`ArrowInvalid` occurs while running `Dataset.map()` function for DPRContext","comments":"It makes sense :D\r\n\r\nIt seems to work! Thanks a lot :))\r\n\r\nClosing the issue","body":"It seems to fail the final batch ):\r\n\r\nsteps to reproduce:\r\n```\r\nfrom datasets import load_dataset\r\nfrom elasticsearch import Elasticsearch\r\nimport torch\r\nfrom transformers import file_utils, set_seed\r\nfrom transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast\r\nMAX_SEQ_LENGTH = 256\r\nctx_encoder = DPRContextEncoder.from_pretrained(\"facebook\/dpr-ctx_encoder-single-nq-base\", cache_dir=\"..\/datasets\/\")\r\nctx_tokenizer = DPRContextEncoderTokenizerFast.from_pretrained(\r\n \"facebook\/dpr-ctx_encoder-single-nq-base\", \r\n cache_dir=\"..datasets\/\"\r\n)\r\n\r\ndataset = load_dataset('text', \r\n data_files='data\/raw\/ARC_Corpus.txt',\r\n cache_dir='..\/datasets')\r\n\r\ntorch.set_grad_enabled(False)\r\nds_with_embeddings = dataset.map(\r\n lambda example: {\r\n 'embeddings': ctx_encoder(\r\n **ctx_tokenizer(\r\n example[\"text\"], \r\n padding='max_length', \r\n truncation=True, \r\n max_length=MAX_SEQ_LENGTH,\r\n return_tensors=\"pt\"\r\n )\r\n )[0][0].numpy(),\r\n },\r\n batched=True,\r\n load_from_cache_file=False,\r\n batch_size=1000\r\n)\r\n```\r\nARC Corpus can be obtained from [here](https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/arc\/ARC-V1-Feb2018.zip)\r\n\r\nAnd then the error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-13-67d139bb2ed3> in <module>\r\n 14 batched=True,\r\n 15 load_from_cache_file=False,\r\n---> 16 batch_size=1000\r\n 17 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in <dictcomp>(.0)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1257 fn_kwargs=fn_kwargs,\r\n 1258 new_fingerprint=new_fingerprint,\r\n-> 1259 update_data=update_data,\r\n 1260 )\r\n 1261 else:\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 155 }\r\n 156 # apply actual function\r\n--> 157 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 158 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 159 # re-apply format to the output\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1526 if update_data:\r\n 1527 batch = cast_to_python_objects(batch)\r\n-> 1528 writer.write_batch(batch)\r\n 1529 if update_data:\r\n 1530 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 276 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 277 typed_sequence_examples[col] = typed_sequence\r\n--> 278 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 279 self.write_table(pa_table)\r\n 280 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_arrays()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.validate()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Column 1 named text expected length 768 but got length 1000\r\n```","comment_length":15,"text":"`ArrowInvalid` occurs while running `Dataset.map()` function for DPRContext\nIt seems to fail the final batch ):\r\n\r\nsteps to reproduce:\r\n```\r\nfrom datasets import load_dataset\r\nfrom elasticsearch import Elasticsearch\r\nimport torch\r\nfrom transformers import file_utils, set_seed\r\nfrom transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast\r\nMAX_SEQ_LENGTH = 256\r\nctx_encoder = DPRContextEncoder.from_pretrained(\"facebook\/dpr-ctx_encoder-single-nq-base\", cache_dir=\"..\/datasets\/\")\r\nctx_tokenizer = DPRContextEncoderTokenizerFast.from_pretrained(\r\n \"facebook\/dpr-ctx_encoder-single-nq-base\", \r\n cache_dir=\"..datasets\/\"\r\n)\r\n\r\ndataset = load_dataset('text', \r\n data_files='data\/raw\/ARC_Corpus.txt',\r\n cache_dir='..\/datasets')\r\n\r\ntorch.set_grad_enabled(False)\r\nds_with_embeddings = dataset.map(\r\n lambda example: {\r\n 'embeddings': ctx_encoder(\r\n **ctx_tokenizer(\r\n example[\"text\"], \r\n padding='max_length', \r\n truncation=True, \r\n max_length=MAX_SEQ_LENGTH,\r\n return_tensors=\"pt\"\r\n )\r\n )[0][0].numpy(),\r\n },\r\n batched=True,\r\n load_from_cache_file=False,\r\n batch_size=1000\r\n)\r\n```\r\nARC Corpus can be obtained from [here](https:\/\/ai2-datasets.s3-us-west-2.amazonaws.com\/arc\/ARC-V1-Feb2018.zip)\r\n\r\nAnd then the error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-13-67d139bb2ed3> in <module>\r\n 14 batched=True,\r\n 15 load_from_cache_file=False,\r\n---> 16 batch_size=1000\r\n 17 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in map(self, function, with_indices, input_columns, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/dataset_dict.py in <dictcomp>(.0)\r\n 301 num_proc=num_proc,\r\n 302 )\r\n--> 303 for k, dataset in self.items()\r\n 304 }\r\n 305 )\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1257 fn_kwargs=fn_kwargs,\r\n 1258 new_fingerprint=new_fingerprint,\r\n-> 1259 update_data=update_data,\r\n 1260 )\r\n 1261 else:\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 155 }\r\n 156 # apply actual function\r\n--> 157 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 158 datasets: List[\"Dataset\"] = list(out.values()) if isinstance(out, dict) else [out]\r\n 159 # re-apply format to the output\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1526 if update_data:\r\n 1527 batch = cast_to_python_objects(batch)\r\n-> 1528 writer.write_batch(batch)\r\n 1529 if update_data:\r\n 1530 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 276 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 277 typed_sequence_examples[col] = typed_sequence\r\n--> 278 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 279 self.write_table(pa_table)\r\n 280 \r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_arrays()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.validate()\r\n\r\n~\/.cache\/pypoetry\/virtualenvs\/masters-utTTC0p8-py3.7\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Column 1 named text expected length 768 but got length 1000\r\n```\nIt makes sense :D\r\n\r\nIt seems to work! Thanks a lot :))\r\n\r\nClosing the issue","embeddings":[-0.5113657713,-0.340737462,-0.1262720525,0.0608496368,0.0776377022,0.1417664886,-0.018712217,0.2698065937,-0.2622149587,0.0491960905,0.1108063161,0.5883788466,-0.0307827368,-0.1317675114,-0.4038944542,-0.0905743986,-0.0616570264,0.1219060645,0.2424558401,-0.089935638,-0.3668372035,0.0883089751,-0.2699972093,-0.1003229246,-0.1718577743,-0.2321972847,0.2313919067,-0.0377138183,0.003301251,-0.6815324426,0.6883983016,-0.2978188097,0.2429684848,0.1627676785,-0.0001183986,0.2131496072,0.4500014782,0.0111980177,-0.0065675713,0.080258362,-0.2403251082,0.194777444,0.0855849758,-0.1638190746,0.0858565792,-0.0353686213,-0.0441782735,-0.427708596,0.3808458447,0.195334971,0.1014310941,0.2174219787,0.0514748059,-0.0151883606,-0.0687452629,0.1960485876,-0.1333302259,-0.2255924344,0.2797706425,0.1810365021,0.1458378285,0.5812528729,-0.0482616462,0.2692404985,0.1576283872,-0.1972575933,0.0579736903,-0.3331651092,0.1915374994,-0.1121736318,0.2094191313,-0.2229378819,-0.2514075637,0.1366557777,-0.1958025694,-0.3983700573,0.1623699516,-0.1368175,0.051956825,0.036535617,-0.1455416232,0.139839381,-0.2271478921,0.24815014,-0.1595502347,0.158727631,-0.0363634005,0.3123621643,-0.0161653459,-0.1319190264,0.1796032041,0.1299351603,-0.1233132333,0.0910461843,-0.2346746773,-0.1503433883,-0.2761262655,-0.3692839444,0.0213410761,-0.1542137265,-0.1089864597,0.1875861883,0.540751636,0.1412303448,0.0859375447,0.1132425293,0.242615208,0.8317361474,-0.0535715595,-0.6531259418,0.0678203776,0.0701442957,-0.0338044018,-0.3524084985,0.256862551,0.1501419544,0.1474722475,0.0349667855,-0.279946506,0.1407098919,-0.438165009,0.0077407854,-0.2134210616,0.3998953104,0.2957156003,-0.1497726738,0.2224410474,0.1331149936,0.1421479881,-0.1057698429,-0.0059415619,-0.0893860981,-0.1892879009,-0.1479352415,0.1425799429,-0.1967147738,0.2172679454,-0.2264876217,-0.0113379322,-0.0850389674,0.3367851079,0.0655955449,0.4326565266,0.2180086821,0.0231516995,0.2297055721,-0.0104063805,0.1910956949,-0.1135064363,-0.0793193206,-0.1578666717,-0.3309911489,-0.3654819429,0.1561720073,0.3952356279,-0.028295733,-0.0704108402,0.0058605992,0.5582671165,-0.0122786565,0.3374227583,-0.2378217578,0.1193173453,0.0060363831,-0.0863270462,0.184869349,-0.3499477208,-0.0525056981,0.0317243487,0.280678153,0.0482468344,0.0491322652,-0.3638934791,0.0915571451,-0.1223412901,0.5995523334,0.4406459033,-0.7461100221,-0.3960719705,0.5548933744,0.0000231852,-0.2747846246,-0.0202935971,-0.0612332262,0.4194323719,-0.1286513954,0.1339125484,0.0815142617,-0.0028018788,0.0602680556,-0.0769046396,0.0175310653,0.2058888823,0.130353868,0.0866857395,-0.0168716852,0.2668988109,-0.0338122919,-0.0874593481,0.0137310801,0.3313800395,0.4629959762,-0.1180306226,-0.1043538675,0.2504345775,0.0164172519,-0.3472789526,0.1601376832,-0.1680821031,-0.0168550368,-0.343441695,-0.2692348957,-0.1579743922,0.1876064539,-0.3379364312,-0.2994133234,0.0545108058,0.0237101484,-0.023135744,-0.1506970078,-0.2348758578,-0.3109050393,-0.0031263814,0.0129285213,-0.371036768,-0.081058912,-0.1028760746,-0.1666499376,-0.2229582369,-0.0823807046,0.1944281906,-0.1868701428,-0.2059344798,0.3810172379,-0.3576883078,0.1575224847,-0.2036373615,0.0353445709,0.0965872109,-0.7572340965,-0.1935379505,-0.0853928477,0.0226561949,-0.1436367929,0.2973734736,0.0332045108,0.4531671405,0.32563591,-0.2301132083,0.0423419885,0.1064608321,-0.0633365661,-0.1695415229,-0.2364452034,-0.1860896647,-0.1981634051,0.1607283205,-0.0286631752,-0.1010631099,0.033602152,0.1005503759,0.032848306,0.051831156,0.0886938199,-0.3016294837,0.1139409393,0.2114739269,0.1100961193,0.1420832723,0.0793418661,0.0846390128,0.0585502237,-0.0404893346,0.0700683594,0.3686750531,0.2179716974,-0.0281671304,0.257704705,0.0106333932,0.315379262,-0.256785363,-0.225231126,0.2095829397,0.0708614513,-0.2471293062,0.1702718586,-0.1032790393,-0.1220331639,0.0348452218,-0.1799590886,0.2601111531,-0.0945540741,-0.0504049994,0.0102419816,-0.061194364,0.4547729194,-0.0581415668,-0.0782623664,0.3865163326,-0.1864851266,-0.2859807014,-0.1884989291,-0.2462268472,-0.1040349305,0.0507002547,0.0092101879,0.2503574491,0.1913103312,0.0721683577,-0.3318425417,-0.2322167158,0.3374351859,-0.0510996468,-0.3957129419,0.1137748212,-0.2494402528,-0.382923454,-0.4515089691,0.1395195872,0.1534736603,-0.2197473049,0.1152105033,-0.1087608859,-0.22306256,0.1489222348,-0.5410323143,-0.2287052721,-0.2145576924,-0.0828818455,-0.3405638635,-0.0386743769,0.182748735,0.1774793565,0.0308759715,-0.0209289491,-0.0506052971,-0.1819638908,0.199599281,0.2329193056,-0.3783580363,0.0198900681,-0.0408562608,-0.1620057225,0.3158453703,0.5299435854,-0.1742092669,0.4087524414,-0.0935854092,0.0415871069,-0.3694301844,0.1419584006,0.5250942707,0.2435423434,-0.0890006721,-0.0797964483,0.1214501634,0.0565610826,-0.1646126956,0.2331374139,-0.1432997137,0.5282084942,0.0230465196,0.7464090586,0.1865098327,-0.6531713605,0.3472135961,-0.0085074985,-0.0273385886,0.0829499215,-0.1887917072,-0.094296217,-0.1831742972,0.080202572,-0.1202504113,-0.1443750709,-0.5767668486,0.1931263804,0.0903160498,-0.4708704948,0.0725259855,0.0098931044,-0.2897835374,0.4527910948,0.0263338014,0.1306546777,-0.0530324057,0.0912237614,0.1545127034,-0.2886983156,0.0147817573,-0.0437882356,-0.0654516518,-0.2173924893,-0.367125541,0.4868406951,0.2577666938,0.3479820788,-0.11464414,0.1626833528,0.0557988808,0.1853664666,0.4528127611,-0.4465717971,0.1091212779,0.0074966587,-0.1906968951,-0.2724784315,0.094267413,-0.0605693534,0.5769729614,0.4145486057,0.4333523214,-0.1463117748,-0.0121417092,0.0991358161,-0.2778380215,0.0614904016,-0.0091800224,0.0020793576,-0.1659399271,-0.2891857326,0.074616313,0.5696228743,0.0475608967,0.3033523858,-0.3070929348,-0.5790467858,-0.0392232463,-0.226955846,0.0006710388,0.2106073201,-0.0364595689,0.1529902518,-0.0096756527,0.3127455115,0.5661595464,0.1893276125,-0.0359560363,-0.1788259149,0.1438657343,-0.0896839425,0.0997157767,0.1143462583,-0.1353585869,-0.1474956274,0.0493534021,-0.0527137257,-0.0971566215,-0.0111675179,0.281041801,0.133730337,-0.3588031828,0.0294272844,0.2876640558,-0.1065136045,-0.022147404,0.5833806396,0.1208186522,-0.3441368341,0.5862824917,0.4361251891,0.6721884608,0.2287106067,-0.1058533788,-0.0988228917,0.2796190381,-0.0167952534,-0.0684407726,0.161393106,-0.2341504097,0.1617098749,-0.0667752624,-0.1077655181,-0.0533003174,0.265442431,-0.2107126415,-0.1009784192,0.1353304088,-0.2278293371,-0.0918503553,0.0910272747,-0.0353779607,-0.2407876402,0.2644459009,0.078892611,0.1308480203,0.1920498908,-0.111443907,-0.1472589523,0.0668359473,-0.3950742483,-0.233878687,0.0655947849,-0.4276450276,0.6017538905,0.1960137933,-0.1862182468,0.243661359,0.2843470275,0.2751020789,-0.2213450968,-0.0998323485,0.2911045253,0.4625076354,-0.2195213139,0.0896592662,-0.1184225976,0.1319423914,0.0171379577,-0.0814791471,0.2218269855,-0.1132743582,-0.383653909,0.1449813247,0.0136611583,-0.1129018962,-0.2023650259,0.0042820321,-0.0683048591,-0.3333327472,-0.1713193506,0.022209296,-0.1941104382,-0.1613932848,0.1137868688,0.0862243846,-0.3374771774,0.251973927,0.6712554097,-0.1556506604,-0.3268623352,0.4741314352,-0.0137802921,-0.1899153739,-0.2256557345,0.0754358768,0.5069118142,-0.5794172883,0.0827752575,-0.3235072792,0.0102473591,0.1571317315,0.5728722811,0.0356439389,0.3189882636,-0.1937944591,-0.0653920025,-0.2630267143,0.0230289586,-0.0755285397,0.4225475192,-0.1140358523,0.1403530538,-0.0659269243,0.0310643725,-0.212642163,-0.0343555883,-0.2249841988,0.0371422395,-0.1169101447,0.1496392787,0.0520565398,0.1060676053,-0.0331649072,0.4258903265,-0.1706050932,-0.0995732993,0.0934457034,0.1573872566,0.1690530479,-0.1038612425,-0.2477400601,-0.1066066623,-0.2969983816,-0.0554750152,0.1336551458,0.0964181572,-0.0245325211,0.1679580063,-0.0560057722,0.0031046285,0.0417178869,-0.2056293339,0.1120127961,0.1913034171,-0.0987461731,0.242951259,0.0712005496,-0.0102206515,-0.1499932855,-0.0407378189,0.0804265961,0.0391499847,0.2788161039,-0.4909853637,-0.2165833116,0.1341007054,0.1585979611,0.0885760412,-0.5477801561,-0.2336433083,-0.0200244058,0.2141940743,-0.0538470782,0.0383823253,-0.118472591,0.1434016228,0.0451569483,-0.028205758,0.1623191535,-0.0965995044,0.1533261985,0.2675468326,0.3809916675,-0.4468233883,0.0574765243,0.2488967478,-0.1121353135,0.0673063025,0.1593638957,0.3147460222,0.5915929079,0.3026495278,-0.1068798006,0.6875780821,0.2359990478,0.1713031679,0.0590547957,-0.0296096448,0.0559683926,-0.1344697475,0.1046848819,0.2520949841,0.2909655571,0.1931342483,-0.0982567817,-0.2824084163,0.0062947846,0.022751499,-0.1574299335,-0.1467564404,-0.3718587458,0.0348342098,0.0237774923,-0.0485217609,-0.1388556212,0.3094592094,0.2812197506,0.0936241448,-0.3840429485,-0.1180859134,0.1315281987,-0.0105849952,0.1634087116,-0.2039889991,0.2102388591,0.0912892297,-0.162929967,0.050826408,0.5249247551,0.387380898,0.2320964187,-0.3360838592,0.0496417992,0.0634767562,-0.0855809227,0.0324373469,0.2687185109,0.003969796,-0.0099392906,0.0981375873,0.0593787096,-0.0846748799,-0.0261173509,0.1643423289,0.4865007699,0.3900396526,-0.0844932497,-0.1798851043,-0.1179356873,0.0608870834,0.1091435999,-0.4464127123,0.1287889183,0.4633158147,-0.1877123564,0.0811693221,0.012980517,0.0467701368,-0.1753967851,0.4479142427,0.4729489684,-0.1310104281,-0.2427333295,-0.1254306138,-0.115588747,0.4575336576,-0.1092391312,-0.0273592565,0.3644365966,0.3641228378,0.1350461543,0.0312272143,0.2069369107,0.0161172766,-0.2688085139,0.4063840508,-0.0002255975,-0.0250777677,-0.0781739727,0.1112399697,0.1547817141,-0.2695218325,0.0635711923,0.2607866228,0.0267178547,-0.3894838095,-0.2669806778,0.1352346241,-0.259506166,0.6943307519,0.0299417879,0.0702526122,-0.2511700988,0.0753903538,-0.2062064111,-0.3413657844,0.0098226294,-0.0084072109,0.2645082474,0.274427861,-0.0939271152,-0.0926951319,-0.3260191679,-0.2867220044,-0.0668520108,-0.2104846388,-0.277595669,0.3050505519,-0.2955505252,0.1799634099,-0.1510111988,0.1711075306,0.2621454,0.2405257225,-0.3641981483,-0.588109076,0.5087835789,-0.3397665024,-0.3331579566,0.3033267856,0.0289909914,0.6112289429,0.1801222861,-0.537248075,0.0097092548,0.1904549301,-0.0897613615,-0.4189183414,-0.217498377,0.1601080447,0.0964678675,-0.099575378,0.3238032758,0.1050207615,0.1133849099,-0.2556917667,-0.3263609707]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1681","title":"Dataset \"dane\" missing","comments":"Hi @KennethEnevoldsen ,\r\nI think the issue might be that this dataset was added during the community sprint and has not been released yet. It will be available with the v2 of datasets.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of datasets using pip:\r\npip install git+https:\/\/github.com\/huggingface\/datasets.git@master","body":"the `dane` dataset appear to be missing in the latest version (1.1.3).\r\n\r\n```python\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.3'\r\n>>> \"dane\" in datasets.list_datasets()\r\nTrue\r\n```\r\n\r\nAs we can see it should be present, but doesn't seem to be findable when using `load_dataset`.\r\n\r\n```python\r\n>>> datasets.load_dataset(\"dane\")\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 588, in load_dataset\r\n module_path, hash = prepare_module(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 280, in prepare_module\r\n raise FileNotFoundError(\r\nFileNotFoundError: Couldn't find file locally at dane\/dane.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n```\r\n\r\nThis issue might be relevant to @ophelielacroix from the Alexandra Institut whom created the data.","comment_length":56,"text":"Dataset \"dane\" missing\nthe `dane` dataset appear to be missing in the latest version (1.1.3).\r\n\r\n```python\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.3'\r\n>>> \"dane\" in datasets.list_datasets()\r\nTrue\r\n```\r\n\r\nAs we can see it should be present, but doesn't seem to be findable when using `load_dataset`.\r\n\r\n```python\r\n>>> datasets.load_dataset(\"dane\")\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 588, in load_dataset\r\n module_path, hash = prepare_module(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 280, in prepare_module\r\n raise FileNotFoundError(\r\nFileNotFoundError: Couldn't find file locally at dane\/dane.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n```\r\n\r\nThis issue might be relevant to @ophelielacroix from the Alexandra Institut whom created the data.\nHi @KennethEnevoldsen ,\r\nI think the issue might be that this dataset was added during the community sprint and has not been released yet. It will be available with the v2 of datasets.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of datasets using pip:\r\npip install git+https:\/\/github.com\/huggingface\/datasets.git@master","embeddings":[-0.0066535207,-0.0612563305,-0.0992998853,0.1284519881,0.2658341527,0.2934038043,0.4646624029,0.1487023532,0.2840229869,0.0827282742,0.1303288788,0.0408590771,-0.0920234174,-0.1245878786,0.2971785367,-0.2793469429,0.2202173769,0.0512977093,0.1837167591,-0.1475976259,-0.2369974256,0.254214108,-0.2888587415,-0.090038985,-0.0924229249,0.0823839828,-0.101888001,0.139782533,-0.0649531037,-0.6014867425,0.642777741,0.0417708345,0.3791090846,0.4024262726,-0.0001110658,0.0540831015,0.2432838678,0.0235112887,-0.4869872332,-0.3152087033,-0.6912311316,-0.4789561927,0.1658877283,-0.1466356218,-0.0642058998,0.1301175803,0.0734554678,-0.4600993395,0.0440699533,0.2021569163,0.2513066232,0.0974107385,0.1838599741,-0.2588645518,0.0307783745,0.2624579072,-0.1520386636,0.1597358584,0.1850463301,0.198756665,0.3487948477,0.307885021,0.148994565,0.0193946417,0.2877072096,0.1553142667,0.2871349752,-0.1641148031,-0.0070116892,0.0631371737,0.6540952325,-0.140773505,-0.7040755749,-0.3533023298,0.1485740989,-0.3803811371,0.1837790608,-0.1397375762,-0.1242031157,0.1525912434,-0.2222140729,-0.2455667406,0.0295432061,0.2610390186,-0.2916905284,0.3245614767,-0.1398500353,-0.021707261,-0.2066091746,-0.0592098124,-0.0247959569,-0.1159208044,0.1139637828,0.2741007805,-0.2491634041,-0.0674469098,0.1185003966,-0.2061502486,-0.0327990986,0.1498194188,0.0173948165,-0.171906203,-0.0063734083,0.1006967723,0.353479445,0.2497101575,0.1984507889,-0.0159756802,0.257517904,0.3605662584,0.1864643693,-0.1657250971,-0.2060384154,-0.2400937676,0.2269962579,0.0962073132,0.2703241408,-0.0575917736,-0.2305234224,-0.0208626539,0.3652468622,-0.1499254704,-0.0439337641,0.1124223247,-0.0977388695,0.3643301725,-0.0803220347,0.1735870242,-0.1319596618,0.031080192,-0.248398155,-0.0731434748,-0.0761916637,-0.3178832531,0.2205925286,-0.1794002652,0.4029653072,-0.0104438094,-0.2701882422,0.1550481021,-0.0678124875,0.0290778093,-0.1138985455,0.2672907114,-0.0218402129,0.1548041701,0.0522527695,-0.4291962683,-0.1739230156,-0.0561629161,-0.2590444982,-0.1839904487,-0.5109384656,0.2261655629,-0.1876500249,-0.3296757936,-0.4136569798,0.1709020734,-0.0866976231,-0.2888192534,-0.0393894129,-0.1422567815,-0.3087976575,-0.1636295021,0.3612360954,0.4254740477,-0.3754051328,-0.2220775485,-0.2986675799,-0.1379340887,0.0524830744,0.0378002822,-0.1871097982,0.1209383085,-0.2908596396,-0.1904940009,0.5775368214,-0.5136830807,-0.2738463879,0.0612903386,0.1417578012,0.1254551262,0.0390238091,-0.1962939799,0.0934103876,-0.0833633393,0.0603376888,0.4086598456,0.0167406127,-0.1915555745,0.0878219157,-0.0913955644,-0.0085834293,0.2126212865,0.1517920941,0.0569025017,0.1812868565,0.076675944,0.1600443721,0.0915104002,0.0494675525,0.4845329821,0.1955732405,0.3065835238,0.038520325,-0.0549055301,-0.5542677641,0.4240736961,0.3048123419,-0.1309112906,-0.1590324044,0.033567857,-0.2935138941,-0.0039210613,-0.2610232234,-0.1701722294,0.0945450142,0.2176745832,0.1462025344,0.2157948613,-0.3023059666,0.1623969078,0.1202042028,0.2067893296,-0.2911315262,0.1884755641,-0.0311723724,0.044327125,0.022034185,0.2709725797,0.1022873223,-0.0330045931,-0.0559417345,0.1777423471,0.1901222318,0.2124840021,0.2953181863,-0.1923425347,0.2499668002,-0.0279774927,0.129624784,0.0328726359,0.0824973509,0.0115447761,-0.2562762201,0.0155946324,0.1473733485,0.1286751181,0.0872903168,0.253621012,0.2938048542,-0.0222274382,0.3696301281,-0.3660769165,0.0822673589,-0.1461671442,0.4610933959,-0.1435316205,-0.0517950356,0.1415269822,0.343142122,-0.0648482218,0.0488349125,0.2475116551,-0.2129967809,0.0466700569,0.1423934847,0.5272797942,0.6025570035,0.2355483919,0.3136324883,-0.1858260036,0.0877967328,-0.0984764025,0.1036738679,0.1213465407,0.0915519297,-0.0236821659,0.1333335042,0.0904912278,-0.1985484511,-0.1832408756,0.0555040538,0.1326709241,-0.0741935521,0.063425824,-0.368658036,-0.0351405106,-0.125416398,0.1148936823,-0.2466208041,-0.4466947913,-0.0464512855,-0.0794595033,0.0844679847,0.0710717216,-0.1870524734,0.0809297487,-0.101423122,-0.084213011,-0.2959759533,-0.1488332897,-0.2996692359,0.0720545575,0.0946173742,-0.1181530878,0.1933088899,-0.3682515919,0.1760058254,-0.4141762555,-0.4692050815,-0.0031071687,-0.0848735571,0.3349460959,0.4064882696,0.026953673,-0.2145118117,-0.1444413215,0.2156667858,-0.2421434224,-0.1893730909,0.0909964368,-0.0489478149,-0.0040461635,0.0551481545,-0.2715997994,-0.5440988541,-0.2131244242,0.0858798698,0.2569324374,-0.0612959377,0.1413228214,0.0455826446,0.1604282856,-0.0533168055,0.3135327399,-0.205416441,-0.4513518512,0.3560549021,-0.2946029305,-0.2575352192,0.2806711495,-0.1606235802,0.4164973199,-0.2432595491,-0.5850550532,0.3026738763,-0.1646489948,0.2575046718,-0.0604185946,0.18009606,0.3521588147,-0.0163948424,-0.0678970367,-0.2659996748,-0.3340036273,-0.2347367257,-0.0102192378,0.4603375196,-0.096155718,0.3872702122,0.0390981138,0.5496211052,0.1955940872,-0.0551208816,0.1557836682,-0.0753046945,0.3423407078,0.0128896516,-0.3261141777,0.1782972962,0.0754669681,0.207522288,0.0982226506,0.2704727054,-0.2166920155,-0.3083969951,-0.0837837309,-0.1627886891,-0.1371242255,-0.2482927144,-0.1863251477,0.1393463314,0.1061242893,0.166903615,-0.4805337489,-0.1498396844,-0.0822561085,0.4817940593,0.2816618979,0.2097400427,-0.1798550934,0.1377669275,-0.6376325488,0.1711969525,-0.0944816247,0.4824328423,-0.0260465741,0.0418074951,-0.0371493325,0.2225906551,0.4343495965,-0.4581441879,-0.0952571407,0.1160130426,-0.1704671681,-0.3259852827,0.0058239554,0.0822885931,-0.1273268163,0.1212462038,0.3126390874,-0.0887572169,-0.2010982633,0.3699717522,0.1298400611,-0.2001170516,-0.2146290839,-0.3223248124,-0.2684320807,-0.1973860115,-0.3638213277,-0.1186069623,0.0679431036,0.1455649585,0.0368243828,-0.2880838215,0.1053815261,0.09487167,0.1589982808,0.2666432559,0.2433153689,0.1687088311,0.4035201073,0.1185985133,0.1826077998,0.4878797531,-0.1087843627,-0.4713645577,0.1359987408,0.0008555739,0.4082768857,0.0806282163,-0.1543272436,-0.0314881392,0.0084957909,0.0452308953,-0.3203417063,-0.0046980618,0.3944860101,0.1607013494,-0.1158392504,-0.3708450496,0.3312446475,0.1204178706,0.0984683037,0.3976605535,0.110556446,-0.3682602048,0.1173178703,-0.043426618,0.5109540224,0.0518225171,0.2117459625,0.1790180057,-0.1995453387,0.4608455598,0.2042568177,0.0624886379,-0.0796047375,-0.0577878281,-0.1564309001,-0.1497076303,0.2499609143,0.1446952373,-0.0942074358,0.3691319525,0.0745987222,0.2523350716,0.0144616514,0.2626576126,-0.1812769771,-0.1476722807,-0.2805743814,0.173419863,-0.0306048114,0.457046479,-0.1766513139,0.0182180759,-0.0262842234,-0.0635707304,-0.2283398956,-0.1456351876,0.0864704773,0.3676978052,0.0269398987,-0.6471763849,0.2223007828,0.3653705716,0.4824870825,0.0132814208,-0.3053214252,0.3802862465,0.0809771717,0.1492304057,0.049927,0.0952791497,0.2767326534,-0.0868488699,-0.2505894005,0.1121464819,-0.2018728554,-0.3773930073,-0.0706979111,-0.0171871856,-0.1237414479,-0.0389026888,-0.0129357902,-0.0261108726,0.1824720502,-0.3737734854,0.1526242942,0.2982405126,0.0043716417,-0.0554953367,-0.195722267,-0.1303299814,-0.1379314661,0.4166831672,-0.0922932774,-0.4069572687,0.5085174441,0.1203408539,-0.178344205,-0.1431147903,0.1727579087,0.0648818016,-0.1887298077,0.0020651075,-0.1986137927,0.2678204477,0.0857286602,-0.017884044,0.1039521918,-0.4878866971,0.0685084239,-0.5261338949,-0.3209863305,0.2049984336,-0.2184117287,0.1394477189,0.0297863111,-0.2115044147,0.0463660024,-0.2618275285,-0.3148195744,0.0966334119,-0.1712365597,-0.1676990539,-0.1800212413,0.0274010263,0.1157131046,-0.1935607046,0.2423854917,-0.3069908023,0.0083602322,-0.1954171658,-0.1288330704,0.1148744226,-0.1486091465,-0.1268499047,0.118657954,-0.0555263013,-0.2969623506,-0.0920072719,0.1906292588,0.1511767358,-0.0156381205,0.3036956787,0.0122054005,0.0471307561,0.0090311533,-0.097926721,0.1498332918,0.2871651053,0.0977827907,0.1878941208,0.0580555163,0.0973010585,0.049825523,0.0251335744,-0.4133030176,0.1737002432,0.5069929957,-0.4505013227,0.1844443679,0.1284921169,0.1411400437,0.2169173062,-0.278776139,0.3034454882,0.1904480457,0.2245934904,-0.4017912447,-0.2237738371,0.1278615594,0.3568142653,0.0566926971,0.0396723822,0.1326604337,-0.1031146571,-0.0563396886,0.2900175452,0.2105329186,-0.2017617524,0.3308552802,0.5534643531,0.0951264575,0.0724660978,0.0177593492,0.2613463104,-0.0322929882,0.6142643094,-0.070329845,0.3135185242,-0.2665214837,0.1789724678,0.0311904736,-0.4226337075,0.1440199614,-0.1150465086,-0.1819406003,0.0142988833,-0.0033606219,0.360152036,-0.4505881071,-0.0028726708,-0.2931550443,-0.0364465006,-0.2229077369,0.0485663079,-0.16775994,-0.0463837087,-0.040631678,-0.0540507697,-0.0421168059,-0.3212351501,0.1275526732,0.0758551806,-0.0538994484,-0.4687039852,0.0144989565,0.1112592667,0.1754005998,-0.1473204941,0.280513823,0.2702254951,-0.2324672043,0.1633627415,0.5229644775,0.3031785786,-0.0058326549,0.2003104389,-0.1027538106,0.0387290493,-0.2582164705,-0.1050583571,0.3067097366,0.0261483341,0.2868366539,0.1429959387,0.1997896582,-0.0207554214,-0.0108358292,0.1403168291,0.0292250179,-0.4682282209,0.5080540776,-0.4775103927,0.0210620575,-0.0722803473,0.268632412,-0.3505006433,-0.0773351565,0.5598424673,0.2219349295,0.0016675442,-0.1637536138,0.0724010691,0.0291877035,0.1771844923,0.2654812932,-0.0842675939,-0.1829138845,-0.2766012847,-0.6952949762,-0.0747822672,0.029075278,-0.3284610808,0.1751838475,0.0519132502,-0.0468813814,0.0537993908,0.2827133238,0.1716702878,0.0251605213,-0.0062626116,-0.4210507274,-0.523738265,-0.0046749692,-0.0143313631,-0.1315938681,-0.202611655,0.1678825915,-0.178652972,-0.0221266411,-0.1578670144,0.0730144083,0.152875036,0.1502260864,0.2609064579,-0.0170134697,0.2680147588,0.0157072451,-0.0942323357,-0.2700975835,-0.3498823047,-0.111992307,0.3450046778,-0.3022577167,0.4650198519,-0.2893925607,-0.0877455771,-0.0973017141,0.1885309517,0.1759335846,-0.1363103986,-0.3141354322,0.324311167,-0.1108165383,0.2402154356,0.5251003504,0.1949782968,-0.1034947261,-0.030580828,-0.1608141959,-0.2133240849,0.3224272728,-0.346948117,-0.1973442137,-0.2657952309,-0.0163369756,0.2487490475,-0.1727483869,-0.5405055285,0.1857834309,0.4451215267,0.1455664486,-0.1368567646,0.3280616999,0.0225098394,0.0800469592,0.0172947869,0.4166647196,-0.0715255439,-0.1197772026,-0.0335528664,0.0291472208]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1681","title":"Dataset \"dane\" missing","comments":"The `dane` dataset was added recently, that's why it wasn't available yet. We did an intermediate release today just before the v2.0.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `dane` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"dane\")\r\n```","body":"the `dane` dataset appear to be missing in the latest version (1.1.3).\r\n\r\n```python\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.3'\r\n>>> \"dane\" in datasets.list_datasets()\r\nTrue\r\n```\r\n\r\nAs we can see it should be present, but doesn't seem to be findable when using `load_dataset`.\r\n\r\n```python\r\n>>> datasets.load_dataset(\"dane\")\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 588, in load_dataset\r\n module_path, hash = prepare_module(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 280, in prepare_module\r\n raise FileNotFoundError(\r\nFileNotFoundError: Couldn't find file locally at dane\/dane.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n```\r\n\r\nThis issue might be relevant to @ophelielacroix from the Alexandra Institut whom created the data.","comment_length":52,"text":"Dataset \"dane\" missing\nthe `dane` dataset appear to be missing in the latest version (1.1.3).\r\n\r\n```python\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.3'\r\n>>> \"dane\" in datasets.list_datasets()\r\nTrue\r\n```\r\n\r\nAs we can see it should be present, but doesn't seem to be findable when using `load_dataset`.\r\n\r\n```python\r\n>>> datasets.load_dataset(\"dane\")\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 588, in load_dataset\r\n module_path, hash = prepare_module(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 280, in prepare_module\r\n raise FileNotFoundError(\r\nFileNotFoundError: Couldn't find file locally at dane\/dane.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n```\r\n\r\nThis issue might be relevant to @ophelielacroix from the Alexandra Institut whom created the data.\nThe `dane` dataset was added recently, that's why it wasn't available yet. We did an intermediate release today just before the v2.0.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `dane` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"dane\")\r\n```","embeddings":[-0.0066535207,-0.0612563305,-0.0992998853,0.1284519881,0.2658341527,0.2934038043,0.4646624029,0.1487023532,0.2840229869,0.0827282742,0.1303288788,0.0408590771,-0.0920234174,-0.1245878786,0.2971785367,-0.2793469429,0.2202173769,0.0512977093,0.1837167591,-0.1475976259,-0.2369974256,0.254214108,-0.2888587415,-0.090038985,-0.0924229249,0.0823839828,-0.101888001,0.139782533,-0.0649531037,-0.6014867425,0.642777741,0.0417708345,0.3791090846,0.4024262726,-0.0001110658,0.0540831015,0.2432838678,0.0235112887,-0.4869872332,-0.3152087033,-0.6912311316,-0.4789561927,0.1658877283,-0.1466356218,-0.0642058998,0.1301175803,0.0734554678,-0.4600993395,0.0440699533,0.2021569163,0.2513066232,0.0974107385,0.1838599741,-0.2588645518,0.0307783745,0.2624579072,-0.1520386636,0.1597358584,0.1850463301,0.198756665,0.3487948477,0.307885021,0.148994565,0.0193946417,0.2877072096,0.1553142667,0.2871349752,-0.1641148031,-0.0070116892,0.0631371737,0.6540952325,-0.140773505,-0.7040755749,-0.3533023298,0.1485740989,-0.3803811371,0.1837790608,-0.1397375762,-0.1242031157,0.1525912434,-0.2222140729,-0.2455667406,0.0295432061,0.2610390186,-0.2916905284,0.3245614767,-0.1398500353,-0.021707261,-0.2066091746,-0.0592098124,-0.0247959569,-0.1159208044,0.1139637828,0.2741007805,-0.2491634041,-0.0674469098,0.1185003966,-0.2061502486,-0.0327990986,0.1498194188,0.0173948165,-0.171906203,-0.0063734083,0.1006967723,0.353479445,0.2497101575,0.1984507889,-0.0159756802,0.257517904,0.3605662584,0.1864643693,-0.1657250971,-0.2060384154,-0.2400937676,0.2269962579,0.0962073132,0.2703241408,-0.0575917736,-0.2305234224,-0.0208626539,0.3652468622,-0.1499254704,-0.0439337641,0.1124223247,-0.0977388695,0.3643301725,-0.0803220347,0.1735870242,-0.1319596618,0.031080192,-0.248398155,-0.0731434748,-0.0761916637,-0.3178832531,0.2205925286,-0.1794002652,0.4029653072,-0.0104438094,-0.2701882422,0.1550481021,-0.0678124875,0.0290778093,-0.1138985455,0.2672907114,-0.0218402129,0.1548041701,0.0522527695,-0.4291962683,-0.1739230156,-0.0561629161,-0.2590444982,-0.1839904487,-0.5109384656,0.2261655629,-0.1876500249,-0.3296757936,-0.4136569798,0.1709020734,-0.0866976231,-0.2888192534,-0.0393894129,-0.1422567815,-0.3087976575,-0.1636295021,0.3612360954,0.4254740477,-0.3754051328,-0.2220775485,-0.2986675799,-0.1379340887,0.0524830744,0.0378002822,-0.1871097982,0.1209383085,-0.2908596396,-0.1904940009,0.5775368214,-0.5136830807,-0.2738463879,0.0612903386,0.1417578012,0.1254551262,0.0390238091,-0.1962939799,0.0934103876,-0.0833633393,0.0603376888,0.4086598456,0.0167406127,-0.1915555745,0.0878219157,-0.0913955644,-0.0085834293,0.2126212865,0.1517920941,0.0569025017,0.1812868565,0.076675944,0.1600443721,0.0915104002,0.0494675525,0.4845329821,0.1955732405,0.3065835238,0.038520325,-0.0549055301,-0.5542677641,0.4240736961,0.3048123419,-0.1309112906,-0.1590324044,0.033567857,-0.2935138941,-0.0039210613,-0.2610232234,-0.1701722294,0.0945450142,0.2176745832,0.1462025344,0.2157948613,-0.3023059666,0.1623969078,0.1202042028,0.2067893296,-0.2911315262,0.1884755641,-0.0311723724,0.044327125,0.022034185,0.2709725797,0.1022873223,-0.0330045931,-0.0559417345,0.1777423471,0.1901222318,0.2124840021,0.2953181863,-0.1923425347,0.2499668002,-0.0279774927,0.129624784,0.0328726359,0.0824973509,0.0115447761,-0.2562762201,0.0155946324,0.1473733485,0.1286751181,0.0872903168,0.253621012,0.2938048542,-0.0222274382,0.3696301281,-0.3660769165,0.0822673589,-0.1461671442,0.4610933959,-0.1435316205,-0.0517950356,0.1415269822,0.343142122,-0.0648482218,0.0488349125,0.2475116551,-0.2129967809,0.0466700569,0.1423934847,0.5272797942,0.6025570035,0.2355483919,0.3136324883,-0.1858260036,0.0877967328,-0.0984764025,0.1036738679,0.1213465407,0.0915519297,-0.0236821659,0.1333335042,0.0904912278,-0.1985484511,-0.1832408756,0.0555040538,0.1326709241,-0.0741935521,0.063425824,-0.368658036,-0.0351405106,-0.125416398,0.1148936823,-0.2466208041,-0.4466947913,-0.0464512855,-0.0794595033,0.0844679847,0.0710717216,-0.1870524734,0.0809297487,-0.101423122,-0.084213011,-0.2959759533,-0.1488332897,-0.2996692359,0.0720545575,0.0946173742,-0.1181530878,0.1933088899,-0.3682515919,0.1760058254,-0.4141762555,-0.4692050815,-0.0031071687,-0.0848735571,0.3349460959,0.4064882696,0.026953673,-0.2145118117,-0.1444413215,0.2156667858,-0.2421434224,-0.1893730909,0.0909964368,-0.0489478149,-0.0040461635,0.0551481545,-0.2715997994,-0.5440988541,-0.2131244242,0.0858798698,0.2569324374,-0.0612959377,0.1413228214,0.0455826446,0.1604282856,-0.0533168055,0.3135327399,-0.205416441,-0.4513518512,0.3560549021,-0.2946029305,-0.2575352192,0.2806711495,-0.1606235802,0.4164973199,-0.2432595491,-0.5850550532,0.3026738763,-0.1646489948,0.2575046718,-0.0604185946,0.18009606,0.3521588147,-0.0163948424,-0.0678970367,-0.2659996748,-0.3340036273,-0.2347367257,-0.0102192378,0.4603375196,-0.096155718,0.3872702122,0.0390981138,0.5496211052,0.1955940872,-0.0551208816,0.1557836682,-0.0753046945,0.3423407078,0.0128896516,-0.3261141777,0.1782972962,0.0754669681,0.207522288,0.0982226506,0.2704727054,-0.2166920155,-0.3083969951,-0.0837837309,-0.1627886891,-0.1371242255,-0.2482927144,-0.1863251477,0.1393463314,0.1061242893,0.166903615,-0.4805337489,-0.1498396844,-0.0822561085,0.4817940593,0.2816618979,0.2097400427,-0.1798550934,0.1377669275,-0.6376325488,0.1711969525,-0.0944816247,0.4824328423,-0.0260465741,0.0418074951,-0.0371493325,0.2225906551,0.4343495965,-0.4581441879,-0.0952571407,0.1160130426,-0.1704671681,-0.3259852827,0.0058239554,0.0822885931,-0.1273268163,0.1212462038,0.3126390874,-0.0887572169,-0.2010982633,0.3699717522,0.1298400611,-0.2001170516,-0.2146290839,-0.3223248124,-0.2684320807,-0.1973860115,-0.3638213277,-0.1186069623,0.0679431036,0.1455649585,0.0368243828,-0.2880838215,0.1053815261,0.09487167,0.1589982808,0.2666432559,0.2433153689,0.1687088311,0.4035201073,0.1185985133,0.1826077998,0.4878797531,-0.1087843627,-0.4713645577,0.1359987408,0.0008555739,0.4082768857,0.0806282163,-0.1543272436,-0.0314881392,0.0084957909,0.0452308953,-0.3203417063,-0.0046980618,0.3944860101,0.1607013494,-0.1158392504,-0.3708450496,0.3312446475,0.1204178706,0.0984683037,0.3976605535,0.110556446,-0.3682602048,0.1173178703,-0.043426618,0.5109540224,0.0518225171,0.2117459625,0.1790180057,-0.1995453387,0.4608455598,0.2042568177,0.0624886379,-0.0796047375,-0.0577878281,-0.1564309001,-0.1497076303,0.2499609143,0.1446952373,-0.0942074358,0.3691319525,0.0745987222,0.2523350716,0.0144616514,0.2626576126,-0.1812769771,-0.1476722807,-0.2805743814,0.173419863,-0.0306048114,0.457046479,-0.1766513139,0.0182180759,-0.0262842234,-0.0635707304,-0.2283398956,-0.1456351876,0.0864704773,0.3676978052,0.0269398987,-0.6471763849,0.2223007828,0.3653705716,0.4824870825,0.0132814208,-0.3053214252,0.3802862465,0.0809771717,0.1492304057,0.049927,0.0952791497,0.2767326534,-0.0868488699,-0.2505894005,0.1121464819,-0.2018728554,-0.3773930073,-0.0706979111,-0.0171871856,-0.1237414479,-0.0389026888,-0.0129357902,-0.0261108726,0.1824720502,-0.3737734854,0.1526242942,0.2982405126,0.0043716417,-0.0554953367,-0.195722267,-0.1303299814,-0.1379314661,0.4166831672,-0.0922932774,-0.4069572687,0.5085174441,0.1203408539,-0.178344205,-0.1431147903,0.1727579087,0.0648818016,-0.1887298077,0.0020651075,-0.1986137927,0.2678204477,0.0857286602,-0.017884044,0.1039521918,-0.4878866971,0.0685084239,-0.5261338949,-0.3209863305,0.2049984336,-0.2184117287,0.1394477189,0.0297863111,-0.2115044147,0.0463660024,-0.2618275285,-0.3148195744,0.0966334119,-0.1712365597,-0.1676990539,-0.1800212413,0.0274010263,0.1157131046,-0.1935607046,0.2423854917,-0.3069908023,0.0083602322,-0.1954171658,-0.1288330704,0.1148744226,-0.1486091465,-0.1268499047,0.118657954,-0.0555263013,-0.2969623506,-0.0920072719,0.1906292588,0.1511767358,-0.0156381205,0.3036956787,0.0122054005,0.0471307561,0.0090311533,-0.097926721,0.1498332918,0.2871651053,0.0977827907,0.1878941208,0.0580555163,0.0973010585,0.049825523,0.0251335744,-0.4133030176,0.1737002432,0.5069929957,-0.4505013227,0.1844443679,0.1284921169,0.1411400437,0.2169173062,-0.278776139,0.3034454882,0.1904480457,0.2245934904,-0.4017912447,-0.2237738371,0.1278615594,0.3568142653,0.0566926971,0.0396723822,0.1326604337,-0.1031146571,-0.0563396886,0.2900175452,0.2105329186,-0.2017617524,0.3308552802,0.5534643531,0.0951264575,0.0724660978,0.0177593492,0.2613463104,-0.0322929882,0.6142643094,-0.070329845,0.3135185242,-0.2665214837,0.1789724678,0.0311904736,-0.4226337075,0.1440199614,-0.1150465086,-0.1819406003,0.0142988833,-0.0033606219,0.360152036,-0.4505881071,-0.0028726708,-0.2931550443,-0.0364465006,-0.2229077369,0.0485663079,-0.16775994,-0.0463837087,-0.040631678,-0.0540507697,-0.0421168059,-0.3212351501,0.1275526732,0.0758551806,-0.0538994484,-0.4687039852,0.0144989565,0.1112592667,0.1754005998,-0.1473204941,0.280513823,0.2702254951,-0.2324672043,0.1633627415,0.5229644775,0.3031785786,-0.0058326549,0.2003104389,-0.1027538106,0.0387290493,-0.2582164705,-0.1050583571,0.3067097366,0.0261483341,0.2868366539,0.1429959387,0.1997896582,-0.0207554214,-0.0108358292,0.1403168291,0.0292250179,-0.4682282209,0.5080540776,-0.4775103927,0.0210620575,-0.0722803473,0.268632412,-0.3505006433,-0.0773351565,0.5598424673,0.2219349295,0.0016675442,-0.1637536138,0.0724010691,0.0291877035,0.1771844923,0.2654812932,-0.0842675939,-0.1829138845,-0.2766012847,-0.6952949762,-0.0747822672,0.029075278,-0.3284610808,0.1751838475,0.0519132502,-0.0468813814,0.0537993908,0.2827133238,0.1716702878,0.0251605213,-0.0062626116,-0.4210507274,-0.523738265,-0.0046749692,-0.0143313631,-0.1315938681,-0.202611655,0.1678825915,-0.178652972,-0.0221266411,-0.1578670144,0.0730144083,0.152875036,0.1502260864,0.2609064579,-0.0170134697,0.2680147588,0.0157072451,-0.0942323357,-0.2700975835,-0.3498823047,-0.111992307,0.3450046778,-0.3022577167,0.4650198519,-0.2893925607,-0.0877455771,-0.0973017141,0.1885309517,0.1759335846,-0.1363103986,-0.3141354322,0.324311167,-0.1108165383,0.2402154356,0.5251003504,0.1949782968,-0.1034947261,-0.030580828,-0.1608141959,-0.2133240849,0.3224272728,-0.346948117,-0.1973442137,-0.2657952309,-0.0163369756,0.2487490475,-0.1727483869,-0.5405055285,0.1857834309,0.4451215267,0.1455664486,-0.1368567646,0.3280616999,0.0225098394,0.0800469592,0.0172947869,0.4166647196,-0.0715255439,-0.1197772026,-0.0335528664,0.0291472208]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1681","title":"Dataset \"dane\" missing","comments":"Thanks. Solved the problem.","body":"the `dane` dataset appear to be missing in the latest version (1.1.3).\r\n\r\n```python\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.3'\r\n>>> \"dane\" in datasets.list_datasets()\r\nTrue\r\n```\r\n\r\nAs we can see it should be present, but doesn't seem to be findable when using `load_dataset`.\r\n\r\n```python\r\n>>> datasets.load_dataset(\"dane\")\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 588, in load_dataset\r\n module_path, hash = prepare_module(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 280, in prepare_module\r\n raise FileNotFoundError(\r\nFileNotFoundError: Couldn't find file locally at dane\/dane.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n```\r\n\r\nThis issue might be relevant to @ophelielacroix from the Alexandra Institut whom created the data.","comment_length":4,"text":"Dataset \"dane\" missing\nthe `dane` dataset appear to be missing in the latest version (1.1.3).\r\n\r\n```python\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.3'\r\n>>> \"dane\" in datasets.list_datasets()\r\nTrue\r\n```\r\n\r\nAs we can see it should be present, but doesn't seem to be findable when using `load_dataset`.\r\n\r\n```python\r\n>>> datasets.load_dataset(\"dane\")\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 588, in load_dataset\r\n module_path, hash = prepare_module(\r\n File \"\/home\/kenneth\/.Envs\/EDP\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 280, in prepare_module\r\n raise FileNotFoundError(\r\nFileNotFoundError: Couldn't find file locally at dane\/dane.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dane\/dane.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dane\/dane.py\r\n```\r\n\r\nThis issue might be relevant to @ophelielacroix from the Alexandra Institut whom created the data.\nThanks. Solved the problem.","embeddings":[-0.0066535207,-0.0612563305,-0.0992998853,0.1284519881,0.2658341527,0.2934038043,0.4646624029,0.1487023532,0.2840229869,0.0827282742,0.1303288788,0.0408590771,-0.0920234174,-0.1245878786,0.2971785367,-0.2793469429,0.2202173769,0.0512977093,0.1837167591,-0.1475976259,-0.2369974256,0.254214108,-0.2888587415,-0.090038985,-0.0924229249,0.0823839828,-0.101888001,0.139782533,-0.0649531037,-0.6014867425,0.642777741,0.0417708345,0.3791090846,0.4024262726,-0.0001110658,0.0540831015,0.2432838678,0.0235112887,-0.4869872332,-0.3152087033,-0.6912311316,-0.4789561927,0.1658877283,-0.1466356218,-0.0642058998,0.1301175803,0.0734554678,-0.4600993395,0.0440699533,0.2021569163,0.2513066232,0.0974107385,0.1838599741,-0.2588645518,0.0307783745,0.2624579072,-0.1520386636,0.1597358584,0.1850463301,0.198756665,0.3487948477,0.307885021,0.148994565,0.0193946417,0.2877072096,0.1553142667,0.2871349752,-0.1641148031,-0.0070116892,0.0631371737,0.6540952325,-0.140773505,-0.7040755749,-0.3533023298,0.1485740989,-0.3803811371,0.1837790608,-0.1397375762,-0.1242031157,0.1525912434,-0.2222140729,-0.2455667406,0.0295432061,0.2610390186,-0.2916905284,0.3245614767,-0.1398500353,-0.021707261,-0.2066091746,-0.0592098124,-0.0247959569,-0.1159208044,0.1139637828,0.2741007805,-0.2491634041,-0.0674469098,0.1185003966,-0.2061502486,-0.0327990986,0.1498194188,0.0173948165,-0.171906203,-0.0063734083,0.1006967723,0.353479445,0.2497101575,0.1984507889,-0.0159756802,0.257517904,0.3605662584,0.1864643693,-0.1657250971,-0.2060384154,-0.2400937676,0.2269962579,0.0962073132,0.2703241408,-0.0575917736,-0.2305234224,-0.0208626539,0.3652468622,-0.1499254704,-0.0439337641,0.1124223247,-0.0977388695,0.3643301725,-0.0803220347,0.1735870242,-0.1319596618,0.031080192,-0.248398155,-0.0731434748,-0.0761916637,-0.3178832531,0.2205925286,-0.1794002652,0.4029653072,-0.0104438094,-0.2701882422,0.1550481021,-0.0678124875,0.0290778093,-0.1138985455,0.2672907114,-0.0218402129,0.1548041701,0.0522527695,-0.4291962683,-0.1739230156,-0.0561629161,-0.2590444982,-0.1839904487,-0.5109384656,0.2261655629,-0.1876500249,-0.3296757936,-0.4136569798,0.1709020734,-0.0866976231,-0.2888192534,-0.0393894129,-0.1422567815,-0.3087976575,-0.1636295021,0.3612360954,0.4254740477,-0.3754051328,-0.2220775485,-0.2986675799,-0.1379340887,0.0524830744,0.0378002822,-0.1871097982,0.1209383085,-0.2908596396,-0.1904940009,0.5775368214,-0.5136830807,-0.2738463879,0.0612903386,0.1417578012,0.1254551262,0.0390238091,-0.1962939799,0.0934103876,-0.0833633393,0.0603376888,0.4086598456,0.0167406127,-0.1915555745,0.0878219157,-0.0913955644,-0.0085834293,0.2126212865,0.1517920941,0.0569025017,0.1812868565,0.076675944,0.1600443721,0.0915104002,0.0494675525,0.4845329821,0.1955732405,0.3065835238,0.038520325,-0.0549055301,-0.5542677641,0.4240736961,0.3048123419,-0.1309112906,-0.1590324044,0.033567857,-0.2935138941,-0.0039210613,-0.2610232234,-0.1701722294,0.0945450142,0.2176745832,0.1462025344,0.2157948613,-0.3023059666,0.1623969078,0.1202042028,0.2067893296,-0.2911315262,0.1884755641,-0.0311723724,0.044327125,0.022034185,0.2709725797,0.1022873223,-0.0330045931,-0.0559417345,0.1777423471,0.1901222318,0.2124840021,0.2953181863,-0.1923425347,0.2499668002,-0.0279774927,0.129624784,0.0328726359,0.0824973509,0.0115447761,-0.2562762201,0.0155946324,0.1473733485,0.1286751181,0.0872903168,0.253621012,0.2938048542,-0.0222274382,0.3696301281,-0.3660769165,0.0822673589,-0.1461671442,0.4610933959,-0.1435316205,-0.0517950356,0.1415269822,0.343142122,-0.0648482218,0.0488349125,0.2475116551,-0.2129967809,0.0466700569,0.1423934847,0.5272797942,0.6025570035,0.2355483919,0.3136324883,-0.1858260036,0.0877967328,-0.0984764025,0.1036738679,0.1213465407,0.0915519297,-0.0236821659,0.1333335042,0.0904912278,-0.1985484511,-0.1832408756,0.0555040538,0.1326709241,-0.0741935521,0.063425824,-0.368658036,-0.0351405106,-0.125416398,0.1148936823,-0.2466208041,-0.4466947913,-0.0464512855,-0.0794595033,0.0844679847,0.0710717216,-0.1870524734,0.0809297487,-0.101423122,-0.084213011,-0.2959759533,-0.1488332897,-0.2996692359,0.0720545575,0.0946173742,-0.1181530878,0.1933088899,-0.3682515919,0.1760058254,-0.4141762555,-0.4692050815,-0.0031071687,-0.0848735571,0.3349460959,0.4064882696,0.026953673,-0.2145118117,-0.1444413215,0.2156667858,-0.2421434224,-0.1893730909,0.0909964368,-0.0489478149,-0.0040461635,0.0551481545,-0.2715997994,-0.5440988541,-0.2131244242,0.0858798698,0.2569324374,-0.0612959377,0.1413228214,0.0455826446,0.1604282856,-0.0533168055,0.3135327399,-0.205416441,-0.4513518512,0.3560549021,-0.2946029305,-0.2575352192,0.2806711495,-0.1606235802,0.4164973199,-0.2432595491,-0.5850550532,0.3026738763,-0.1646489948,0.2575046718,-0.0604185946,0.18009606,0.3521588147,-0.0163948424,-0.0678970367,-0.2659996748,-0.3340036273,-0.2347367257,-0.0102192378,0.4603375196,-0.096155718,0.3872702122,0.0390981138,0.5496211052,0.1955940872,-0.0551208816,0.1557836682,-0.0753046945,0.3423407078,0.0128896516,-0.3261141777,0.1782972962,0.0754669681,0.207522288,0.0982226506,0.2704727054,-0.2166920155,-0.3083969951,-0.0837837309,-0.1627886891,-0.1371242255,-0.2482927144,-0.1863251477,0.1393463314,0.1061242893,0.166903615,-0.4805337489,-0.1498396844,-0.0822561085,0.4817940593,0.2816618979,0.2097400427,-0.1798550934,0.1377669275,-0.6376325488,0.1711969525,-0.0944816247,0.4824328423,-0.0260465741,0.0418074951,-0.0371493325,0.2225906551,0.4343495965,-0.4581441879,-0.0952571407,0.1160130426,-0.1704671681,-0.3259852827,0.0058239554,0.0822885931,-0.1273268163,0.1212462038,0.3126390874,-0.0887572169,-0.2010982633,0.3699717522,0.1298400611,-0.2001170516,-0.2146290839,-0.3223248124,-0.2684320807,-0.1973860115,-0.3638213277,-0.1186069623,0.0679431036,0.1455649585,0.0368243828,-0.2880838215,0.1053815261,0.09487167,0.1589982808,0.2666432559,0.2433153689,0.1687088311,0.4035201073,0.1185985133,0.1826077998,0.4878797531,-0.1087843627,-0.4713645577,0.1359987408,0.0008555739,0.4082768857,0.0806282163,-0.1543272436,-0.0314881392,0.0084957909,0.0452308953,-0.3203417063,-0.0046980618,0.3944860101,0.1607013494,-0.1158392504,-0.3708450496,0.3312446475,0.1204178706,0.0984683037,0.3976605535,0.110556446,-0.3682602048,0.1173178703,-0.043426618,0.5109540224,0.0518225171,0.2117459625,0.1790180057,-0.1995453387,0.4608455598,0.2042568177,0.0624886379,-0.0796047375,-0.0577878281,-0.1564309001,-0.1497076303,0.2499609143,0.1446952373,-0.0942074358,0.3691319525,0.0745987222,0.2523350716,0.0144616514,0.2626576126,-0.1812769771,-0.1476722807,-0.2805743814,0.173419863,-0.0306048114,0.457046479,-0.1766513139,0.0182180759,-0.0262842234,-0.0635707304,-0.2283398956,-0.1456351876,0.0864704773,0.3676978052,0.0269398987,-0.6471763849,0.2223007828,0.3653705716,0.4824870825,0.0132814208,-0.3053214252,0.3802862465,0.0809771717,0.1492304057,0.049927,0.0952791497,0.2767326534,-0.0868488699,-0.2505894005,0.1121464819,-0.2018728554,-0.3773930073,-0.0706979111,-0.0171871856,-0.1237414479,-0.0389026888,-0.0129357902,-0.0261108726,0.1824720502,-0.3737734854,0.1526242942,0.2982405126,0.0043716417,-0.0554953367,-0.195722267,-0.1303299814,-0.1379314661,0.4166831672,-0.0922932774,-0.4069572687,0.5085174441,0.1203408539,-0.178344205,-0.1431147903,0.1727579087,0.0648818016,-0.1887298077,0.0020651075,-0.1986137927,0.2678204477,0.0857286602,-0.017884044,0.1039521918,-0.4878866971,0.0685084239,-0.5261338949,-0.3209863305,0.2049984336,-0.2184117287,0.1394477189,0.0297863111,-0.2115044147,0.0463660024,-0.2618275285,-0.3148195744,0.0966334119,-0.1712365597,-0.1676990539,-0.1800212413,0.0274010263,0.1157131046,-0.1935607046,0.2423854917,-0.3069908023,0.0083602322,-0.1954171658,-0.1288330704,0.1148744226,-0.1486091465,-0.1268499047,0.118657954,-0.0555263013,-0.2969623506,-0.0920072719,0.1906292588,0.1511767358,-0.0156381205,0.3036956787,0.0122054005,0.0471307561,0.0090311533,-0.097926721,0.1498332918,0.2871651053,0.0977827907,0.1878941208,0.0580555163,0.0973010585,0.049825523,0.0251335744,-0.4133030176,0.1737002432,0.5069929957,-0.4505013227,0.1844443679,0.1284921169,0.1411400437,0.2169173062,-0.278776139,0.3034454882,0.1904480457,0.2245934904,-0.4017912447,-0.2237738371,0.1278615594,0.3568142653,0.0566926971,0.0396723822,0.1326604337,-0.1031146571,-0.0563396886,0.2900175452,0.2105329186,-0.2017617524,0.3308552802,0.5534643531,0.0951264575,0.0724660978,0.0177593492,0.2613463104,-0.0322929882,0.6142643094,-0.070329845,0.3135185242,-0.2665214837,0.1789724678,0.0311904736,-0.4226337075,0.1440199614,-0.1150465086,-0.1819406003,0.0142988833,-0.0033606219,0.360152036,-0.4505881071,-0.0028726708,-0.2931550443,-0.0364465006,-0.2229077369,0.0485663079,-0.16775994,-0.0463837087,-0.040631678,-0.0540507697,-0.0421168059,-0.3212351501,0.1275526732,0.0758551806,-0.0538994484,-0.4687039852,0.0144989565,0.1112592667,0.1754005998,-0.1473204941,0.280513823,0.2702254951,-0.2324672043,0.1633627415,0.5229644775,0.3031785786,-0.0058326549,0.2003104389,-0.1027538106,0.0387290493,-0.2582164705,-0.1050583571,0.3067097366,0.0261483341,0.2868366539,0.1429959387,0.1997896582,-0.0207554214,-0.0108358292,0.1403168291,0.0292250179,-0.4682282209,0.5080540776,-0.4775103927,0.0210620575,-0.0722803473,0.268632412,-0.3505006433,-0.0773351565,0.5598424673,0.2219349295,0.0016675442,-0.1637536138,0.0724010691,0.0291877035,0.1771844923,0.2654812932,-0.0842675939,-0.1829138845,-0.2766012847,-0.6952949762,-0.0747822672,0.029075278,-0.3284610808,0.1751838475,0.0519132502,-0.0468813814,0.0537993908,0.2827133238,0.1716702878,0.0251605213,-0.0062626116,-0.4210507274,-0.523738265,-0.0046749692,-0.0143313631,-0.1315938681,-0.202611655,0.1678825915,-0.178652972,-0.0221266411,-0.1578670144,0.0730144083,0.152875036,0.1502260864,0.2609064579,-0.0170134697,0.2680147588,0.0157072451,-0.0942323357,-0.2700975835,-0.3498823047,-0.111992307,0.3450046778,-0.3022577167,0.4650198519,-0.2893925607,-0.0877455771,-0.0973017141,0.1885309517,0.1759335846,-0.1363103986,-0.3141354322,0.324311167,-0.1108165383,0.2402154356,0.5251003504,0.1949782968,-0.1034947261,-0.030580828,-0.1608141959,-0.2133240849,0.3224272728,-0.346948117,-0.1973442137,-0.2657952309,-0.0163369756,0.2487490475,-0.1727483869,-0.5405055285,0.1857834309,0.4451215267,0.1455664486,-0.1368567646,0.3280616999,0.0225098394,0.0800469592,0.0172947869,0.4166647196,-0.0715255439,-0.1197772026,-0.0335528664,0.0291472208]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1679","title":"Can't import cc100 dataset","comments":"cc100 was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `cc100` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nlang = \"en\"\r\ndataset = load_dataset(\"cc100\", lang=lang, split=\"train\")\r\n```","body":"There is some issue to import cc100 dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"cc100\")\r\n```\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/cc100\/cc100.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/cc100\/cc100.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 280 raise FileNotFoundError(\r\n 281 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 282 combined_path, github_file_path, file_path\r\n 283 )\r\n 284 )\r\n\r\nFileNotFoundError: Couldn't find file locally at cc100\/cc100.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/cc100\/cc100.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/cc100\/cc100.py","comment_length":45,"text":"Can't import cc100 dataset\nThere is some issue to import cc100 dataset.\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"cc100\")\r\n```\r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/cc100\/cc100.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/cc100\/cc100.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 280 raise FileNotFoundError(\r\n 281 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 282 combined_path, github_file_path, file_path\r\n 283 )\r\n 284 )\r\n\r\nFileNotFoundError: Couldn't find file locally at cc100\/cc100.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/cc100\/cc100.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/cc100\/cc100.py\ncc100 was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `cc100` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nlang = \"en\"\r\ndataset = load_dataset(\"cc100\", lang=lang, split=\"train\")\r\n```","embeddings":[-0.2749530673,-0.3488157392,-0.1628864706,0.2345296443,0.376396805,0.1457457542,0.0886662602,0.0246397574,0.0158823952,0.2067761719,-0.1255749017,0.058414761,0.0227050297,0.3194364309,0.0983807594,-0.005402416,0.1296453923,0.1373708844,-0.3452748954,0.0146202436,-0.0807719454,0.2495070249,-0.085307695,-0.1574202925,-0.2753781974,-0.1641575098,-0.0514012128,0.1796361804,-0.3012573421,-0.4615064263,0.6025828719,0.1790027618,0.2557296753,0.4899493456,-0.0001045807,0.0461228415,0.3770821393,0.0239945892,-0.4418845177,-0.3079560697,-0.2751539052,-0.2639507353,0.1835988909,-0.0436674058,-0.1347506642,0.201985091,-0.1778894812,-0.2138003409,0.3010198772,0.0500014164,0.3266026676,0.3059599996,0.3633733094,-0.3747950196,0.0341681354,0.0937169716,0.0094008027,0.3512433767,0.2063264847,0.1514515132,0.5028682351,0.1674754173,-0.0955764651,-0.0340652615,0.2659811676,-0.1589617431,-0.1439435929,-0.366864562,0.1385574937,0.1350733191,0.5673771501,-0.2247601897,-0.3478521109,0.0083858827,0.1170568168,-0.5926403999,0.1050112918,0.110301204,0.1089515463,0.2087957114,-0.3087430894,-0.1390147954,-0.1013323516,0.2098259628,-0.1649814993,-0.0933538303,-0.320030123,-0.0166480467,0.1543783695,-0.1073016152,-0.1980895847,0.0044630994,-0.031594675,0.181926474,-0.2933140397,0.0186268166,0.0426845625,0.4454786181,0.1327247769,0.1679537594,0.0847056061,-0.0797265843,0.044417832,0.244242847,0.0643805116,0.1364514083,-0.0355199352,0.0517161451,0.2195052952,0.2635697722,-0.0347134359,-0.2701790929,-0.1346898377,-0.2716605067,-0.0493012518,0.0229159426,0.350949496,-0.2061715126,-0.2172701508,0.0527681075,0.2327941358,0.0468728989,0.136940673,0.5143303275,-0.003946661,0.1669722646,-0.0405057818,0.1405848414,-0.01133803,-0.0045871357,-0.2370131612,0.0864198953,-0.1366158724,-0.1119011864,0.2177563012,-0.5572939515,0.2780295312,0.1383111924,0.0646387264,0.0165715162,-0.1541118622,0.0250986814,-0.0274689011,0.2688960433,0.0131191416,0.1907918453,0.141219601,-0.0905668959,-0.1207038984,-0.1980187744,-0.4211934209,-0.1977618337,-0.2723000348,0.2675927579,-0.1505774111,-0.1318505704,-0.2584421933,-0.1919153333,-0.0768684968,-0.1178111583,-0.0770635381,-0.0027255749,-0.0963254198,-0.1600158215,0.2762827277,0.52097404,-0.3317390084,-0.0715729371,-0.181727618,-0.1196101457,0.0628939942,0.0969097689,-0.2397205383,-0.2497815043,-0.14684771,-0.3261528313,-0.1567421705,-0.5559664965,-0.4181702435,0.1911296546,-0.0677199587,0.0445034578,0.0309950747,-0.1653467119,-0.0856773779,-0.0136372391,-0.204861328,0.0122111384,-0.049923528,-0.0878971294,-0.0892958939,-0.2032905519,0.2163340598,-0.0259719323,-0.1203064099,-0.0841754079,0.2792485952,-0.178420037,0.0360985212,-0.1472934633,-0.0455059558,0.2831318378,0.28106758,0.1500968039,-0.0131318895,-0.096602425,-0.5498054028,0.1840482354,0.0591754019,0.2070122808,-0.2800960839,-0.1901307702,-0.4212655723,0.0353376679,-0.1500855237,-0.1384312958,0.1622018814,-0.0023704628,0.2669632435,0.0558734536,-0.1457965225,0.3959629834,-0.0193073023,0.2013513297,-0.2098923773,0.1347904652,-0.1404971182,0.1456405371,0.1808162928,0.209728241,0.1623943299,-0.1691637337,-0.0460646078,0.3943288028,-0.1744179726,0.530454576,0.0294824895,-0.0860908329,0.3313724697,-0.4779743254,-0.0278088786,-0.1411225796,0.277826339,0.0638917759,-0.0410025679,0.1373913735,-0.1340174973,0.2316181362,0.1290092617,0.0426906608,0.2338794917,0.1133559048,-0.0380368419,-0.1478625238,0.2096565962,0.0880584717,0.612246573,0.0224679932,-0.125888899,-0.047208406,0.1204791516,0.0573649704,-0.0698103905,0.0109027401,-0.3110825121,0.2613328695,0.2233102471,0.2125795633,0.1484139413,0.2743114829,-0.1790673435,-0.057594832,0.0863918066,-0.112077035,0.2854093909,0.0789457187,0.1874593347,-0.0491946973,0.130128026,0.0304458383,-0.2594585121,-0.1271762699,-0.032301113,0.146824643,-0.2639596164,0.090302065,-0.4293751419,-0.1814528555,-0.0166764092,-0.1330062002,-0.0983510911,-0.1650461107,-0.3125822842,0.122342065,0.0383008756,0.0193284359,-0.1376667917,0.0725121945,-0.1019553989,-0.1644335389,-0.1568349302,-0.3101901412,0.025382679,0.1606173515,0.3345532417,-0.0981037766,0.4525795281,-0.1812592149,0.1574148089,0.0079964912,-0.4218347967,0.1091921479,-0.2060811967,0.2028568983,0.2215285897,0.1929978579,-0.2877087295,-0.156739518,0.3035696149,-0.0503761843,-0.0871437639,0.0492058061,-0.1616818309,-0.3335810006,-0.0298849978,-0.2988216281,-0.5281748176,-0.3252052963,0.1262627095,0.2590164244,-0.0417268686,0.4280340672,0.1962852925,0.0906430706,0.0179838836,0.1230181158,-0.0773329362,-0.6638807654,0.4207583368,-0.3592381477,-0.4082060754,0.1353978068,0.0066997116,0.4727200568,-0.0671639889,-0.2308905274,0.0575049445,-0.2374916375,0.0838246197,0.2534953654,0.273425281,0.1750728786,-0.0735047385,-0.1103817448,-0.0898962766,-0.1655896604,-0.2545473576,-0.1683807671,0.1834418029,-0.1109104827,0.2447137386,-0.1323466897,0.6540396214,0.2099393904,-0.1649854034,0.3992759585,-0.1709168255,0.5142349005,-0.0351865031,-0.4970191121,-0.0146232592,-0.0221135225,0.0487140343,0.0795000866,0.0645030588,-0.0016462541,-0.4643183053,-0.2883170545,-0.3414288163,-0.1591687053,-0.3202172816,0.1914167702,0.298622638,0.0857731849,-0.1307843029,-0.0182891693,-0.2483211011,0.0578744709,0.322111398,0.0105764605,0.0752973333,-0.4150483608,-0.3769316375,-0.504602313,0.3550254107,0.0271565225,0.0787502751,0.0839341432,0.0212864764,-0.0299006011,-0.0707841739,0.2992476225,0.0286083315,-0.0222569723,0.1022903398,-0.2124718428,-0.2089952081,0.061171025,-0.0229470544,0.2134125233,-0.0891142264,-0.1713017225,-0.2848718166,-0.0710542873,0.1226265356,0.2774398327,-0.1200054809,-0.046256572,-0.261556536,-0.1937488765,-0.4382615089,-0.185262084,-0.0885528997,0.2786754966,-0.0242833812,-0.1072542742,-0.127537027,-0.1385217309,0.3748293817,0.321849972,0.3681971431,0.1434695721,0.1886389703,0.1020502225,0.1352093071,0.4293233454,0.7669625282,0.0246013608,-0.2726376653,-0.0431875363,0.0048076599,0.2650564909,0.0473108999,-0.1300270408,0.0622840859,0.2921971977,0.1550611556,-0.3183419406,0.0320762619,0.1724954695,0.1257959157,-0.3858863115,-0.1640508175,0.4455737174,0.0851221904,0.1408640742,0.1478751749,0.3699252903,-0.0830996186,-0.0491790101,-0.0902429968,0.7487875819,0.0010086116,-0.0992411152,0.1067605689,-0.0872874781,0.6782954335,0.0400448851,-0.0908959955,-0.1455075294,0.0014086592,0.0021553768,-0.1016371548,0.2758297324,-0.0598340742,-0.2138642371,0.2948035598,-0.2063918114,0.3195100427,-0.1048730537,0.2965541482,-0.1936094612,-0.1699323058,-0.0740502626,0.2467357367,0.0452323966,0.4164851904,-0.201131314,-0.0169361662,-0.1460291892,-0.3335603774,-0.0945572779,-0.0162595864,-0.4031131268,0.1428786814,-0.0505346656,-0.4562132359,0.3574999571,0.2802865803,0.2976773977,0.1905578524,-0.3624043763,0.1698271483,-0.0540937483,-0.2864755094,-0.1399847269,0.2435581237,0.2589796185,-0.1527511179,-0.1368127167,0.2127375752,0.067193836,-0.0705490112,0.2041629702,0.0739911944,0.0268295724,-0.1725116223,-0.2176495939,-0.1118209437,0.036377307,-0.1863762289,0.1992070824,0.0600971244,-0.0420961305,0.0213780794,0.2636808753,-0.1200767383,-0.2585512996,0.337177515,-0.0661532953,-0.1846172959,0.3840769231,0.3715595007,-0.2279283404,-0.238503471,-0.1110123247,0.117292814,-0.4654299915,0.119449079,0.3156570196,0.2769638598,-0.0174405761,0.2775643766,0.1805330962,-0.1019524261,0.1772877425,-0.380125761,-0.2860305905,0.1585612446,-0.056727197,0.1450148374,0.2007385045,0.3208969235,0.0306510013,0.0196826831,-0.4130367339,0.0822302625,-0.1529958397,-0.0324869044,0.6231765151,0.1154276803,0.194410488,-0.2506250143,0.2638086677,0.1378501803,-0.1559448689,-0.333789736,-0.1039380953,0.091930218,-0.0452092029,-0.2206749469,0.1881062239,0.0847531557,0.0613887161,-0.0452282727,0.0839515179,0.1012864932,0.1009422913,0.0681770965,-0.1082367823,-0.0102526862,-0.2600283325,0.1204632223,0.1696056724,0.2260053009,0.0941865072,0.0407167114,0.0963998139,-0.0871348903,0.1342705488,-0.1065194681,-0.1059433967,-0.0376034752,0.4198008478,-0.2765180171,0.0969426706,0.3451443315,0.2989571989,0.2348744571,-0.1580512971,0.282315433,0.2273000181,0.2511664033,-0.307489872,-0.0096709793,-0.0989003703,0.1118579358,0.0866718143,0.1117515191,0.3189949691,-0.1143728346,0.1151209846,0.1248405352,0.4290117025,-0.0957763642,0.4838536978,0.3432486057,0.2017801106,-0.0089379363,0.0586411022,0.2886769474,0.0158026107,0.2157494575,-0.4133892357,0.0226052962,-0.1441311389,0.1580992639,0.157125026,-0.2874079049,0.2587372959,0.1885423809,-0.0733521357,-0.0322902389,-0.0436370932,0.7672813535,-0.0632773936,0.0335436575,-0.2181836367,0.3860799968,-0.0850509703,-0.0513550155,-0.1774566174,-0.231209144,-0.1051059291,0.0002876093,-0.0691204444,-0.339807868,0.2412613779,0.1286629289,-0.1383032799,-0.5828998685,0.0625800341,0.2402125895,0.1811665446,-0.1473495811,0.3023958802,0.3403142095,-0.2132931352,0.0783959925,0.3581850231,0.6212481856,0.1875988692,0.2288161963,0.1830417067,0.0733367056,-0.1954087168,0.0972382054,0.0000464228,-0.1444268823,0.2164273113,0.0926919952,0.2745135725,-0.2333946377,-0.3043327332,-0.0082233567,0.1176785976,-0.3661317527,-0.0558013991,-0.251891464,0.0503883362,-0.2957037389,0.1296907812,-0.4416199625,0.2058071047,0.3274718225,-0.0615106635,0.0302690994,-0.2500506639,0.1146479025,-0.0776505545,0.5049266219,0.2232315987,0.2362313867,-0.2586010993,-0.3236328661,-0.6043619514,-0.017657917,-0.0000915033,-0.0327479057,-0.1260682642,-0.0306668431,-0.0529073142,0.1829476506,0.2557895482,-0.1284562349,0.0552978963,-0.1421186626,-0.0976200104,-0.0224556942,-0.2848815322,-0.128166154,-0.0005137995,-0.1004381776,0.0310888439,-0.2098394632,0.0556075014,-0.3399514854,0.2074807733,-0.2634787858,-0.1694066226,0.3223171532,0.0923560113,0.5297517776,-0.112827681,0.0303640328,-0.1091181263,-0.2837510705,-0.0753787458,0.3584705591,0.1657080948,0.2871477604,-0.241602093,-0.4416572154,-0.2381151468,0.1624214798,0.1115581468,0.0039232257,-0.1397883147,-0.0290725157,-0.1245803908,0.0117553892,0.2453525513,0.1738391221,-0.0560710877,0.0446739122,-0.1264391541,-0.5482059121,0.3843716681,-0.2695604265,-0.0471228734,0.0564557426,0.1962691247,0.3459501863,-0.2228515297,-0.3835343719,0.3413447738,0.2327558696,0.1056846827,-0.1048971564,0.417034626,-0.050641425,-0.1036453545,-0.0186631307,0.3969450295,0.150920704,-0.2328805476,-0.0551846959,-0.1623315513]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1675","title":"Add the 800GB Pile dataset?","comments":"The pile dataset would be very nice.\r\nBenchmarks show that pile trained models achieve better results than most of actually trained models","body":"## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n","comment_length":22,"text":"Add the 800GB Pile dataset?\n## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n\nThe pile dataset would be very nice.\r\nBenchmarks show that pile trained models achieve better results than most of actually trained models","embeddings":[-0.3252147138,0.213067472,-0.1506743729,0.1371282637,0.000942951,0.2578674257,0.1434531361,0.2466031909,-0.0307177734,-0.0639290512,-0.157744959,0.1487944871,-0.5934290886,0.1841434836,0.2048877925,-0.0223660879,-0.0074280403,-0.122964114,-0.0401776135,-0.074812524,-0.0162710417,-0.0986746922,0.06882447,-0.3840371966,-0.219153434,-0.0709641501,-0.12500377,0.0149499346,-0.3641270101,-0.2253900915,-0.2300112844,0.0985262915,0.3934330046,0.5549378991,-0.0001021537,-0.128280431,0.1410385668,-0.1368499845,-0.3820027113,-0.2080664486,-0.071841605,-0.3678537309,-0.3328624666,-0.0382189862,-0.1916804314,-0.1729317158,0.1150261238,-0.3564876914,0.0243452359,0.3020795584,0.2485696375,0.1346296668,0.006665613,-0.1158389524,0.3296718895,0.0767238587,-0.1908534616,0.3099156022,0.3554446995,0.1383180171,-0.0885919705,0.4444621205,0.4074594975,-0.2330576032,0.2445586175,0.029440945,-0.0288664382,-0.1090516001,-0.2375527471,0.3558644652,0.277790904,-0.1494378,-0.5054074526,-0.4087862968,-0.0100187752,-0.393754065,-0.3037469089,0.3672268987,0.0821367726,0.3314937353,0.052197542,-0.2601241171,-0.0611016341,0.177352652,0.0485189781,0.3551804423,-0.022327356,-0.1445569247,0.1392926127,0.0402459465,-0.1318549514,0.0138096297,-0.0106844706,-0.0227933358,-0.2930926681,-0.2389979511,0.4097944796,0.344365716,0.3930755854,-0.2281851768,-0.0614296682,0.0060236086,-0.0072429222,-0.0502688065,0.4607635736,0.0007306704,-0.2595378458,-0.2574682534,0.2438705862,0.0729763731,0.0823986679,0.1097905561,-0.0664121136,-0.0415099002,-0.1491446793,-0.1901452988,-0.0808310732,0.1290602237,0.0055192425,-0.008858148,0.2112541497,-0.0898280293,0.0002598176,0.0070809335,-0.0278747659,0.256572336,0.0403307118,0.2176799625,0.1866182685,-0.5832268596,-0.2047815919,0.1769856662,-0.4584914148,-0.075730674,0.0986139551,-0.1288845092,0.054991927,-0.1199430674,0.1810481995,0.2751230299,0.152120918,0.0321043879,0.0874717012,-0.058149457,-0.0981405601,-0.2579050362,0.0763818622,0.0706216171,-0.2920556366,0.0801397339,-0.0535616837,-0.2715631723,-0.3757432997,0.2224727124,-0.1484063417,0.0717100874,-0.4174520969,0.7332292199,-0.0663980097,0.2488409132,0.0603188016,-0.0827520713,0.0672256276,0.0345236063,0.0980906785,0.1966949105,-0.5776969194,-0.077310659,-0.6373913288,0.0914145783,-0.0582604185,0.2447685301,-0.1526014954,0.1631354094,-0.0524855554,0.0299164336,0.6078365445,0.1548920274,-0.1823148429,0.1506227106,-0.0588281527,-0.0163766947,-0.0104910471,0.1764247864,0.0489749797,-0.0654991567,0.2746493816,0.3799539804,-0.164845407,-0.1337458193,-0.3065443337,-0.4034540951,0.0495001823,0.4762968123,0.1166485697,-0.2422853708,0.1984212548,0.0635692403,0.1890216619,0.0624795705,0.0929858014,-0.1868360788,0.2389918417,0.1390586197,-0.0402011089,-0.2588374913,-0.3763410747,0.0564037524,0.0325516686,0.3952606022,0.0391411148,-0.1904402375,0.1332283914,0.059420865,0.1557436585,-0.0427025557,0.2095918208,0.0314803645,-0.0233042259,-0.1712443084,-0.2686154544,-0.1257418096,-0.1535976231,-0.1023858041,-0.2796764672,-0.0285101179,-0.1850658655,-0.1535534859,0.0187313762,0.3044503927,-0.5161637068,-0.1331953108,0.204218626,0.0854381248,-0.1772891581,-0.208417654,0.2370718122,0.6692658663,0.2030736208,-0.1166468337,0.2004453242,-0.146037519,-0.083937794,0.0267567299,-0.1868268549,0.3955789208,0.3413773775,-0.0392296575,0.0915246308,-0.1056938171,0.1067870334,0.0587483086,-0.0062876018,-0.2438683361,0.2649517655,0.0563978739,0.2367112637,-0.0421917401,-0.3611216843,0.1217309311,0.3609569967,0.1336941123,0.2278655469,0.2681151032,-0.1685228944,-0.2820704579,0.0066008158,-0.2736427486,0.2471921146,0.4055266976,0.2380433083,0.1153727099,0.2009988129,-0.1359192431,-0.0178490188,0.0349395536,-0.0803052038,0.1921199709,0.4058767557,0.0943996683,-0.2614095509,-0.2708651423,-0.1733201146,0.1677818596,0.1304055452,-0.0607841462,0.1282447129,-0.0524662398,0.0956667885,0.0754774213,-0.1807923168,-0.1974586695,0.1726610214,-0.1768560261,-0.2444763035,0.0030569064,-0.1337911636,0.5722124577,-0.0788939297,0.181721434,0.0383824445,-0.2997011542,-0.0945089385,0.2508624792,0.3035569191,-0.2810708284,0.3139433861,0.0578413457,0.1643300354,-0.0403545909,-0.5640271902,0.1866859794,-0.088725917,0.3829104006,0.251629144,-0.0220005922,0.2324692905,-0.0640577227,-0.1304810047,-0.1618163884,-0.0281373542,-0.2053381801,0.0447272472,0.0356374569,-0.2012237906,-0.3236881495,-0.3910394907,-0.4839153886,0.2440131456,0.0915533453,-0.0329911001,-0.049640961,0.152875483,0.3634531498,-0.0162587017,0.0674708635,-0.0638787299,-0.308283478,0.3143343329,-0.0352327377,-0.2157191336,-0.1052859426,-0.2631336749,-0.0849087983,0.0710202977,-0.580185473,0.1611856669,-0.0692842007,0.1904709488,0.0425881222,0.0420679748,0.1455496103,-0.0173330046,-0.2072859257,-0.0365572758,-0.008288648,0.0741132274,0.3788351119,0.3502980769,-0.2368447632,0.296453774,0.2226049453,0.4703178108,-0.0883284956,-0.3011774421,0.3410257101,0.3796491623,0.3692949414,-0.2438226938,-0.0545627624,0.2698502243,0.0895341635,-0.0264575407,0.6250420809,0.2273140699,-0.3611325324,0.0328242071,0.083496511,0.0900719538,-0.090858914,0.3243923783,-0.193723157,-0.0023732486,-0.0843033791,-0.5924363136,-0.1767882556,-0.1897264719,0.0006864526,0.2817796469,0.0461571962,0.0339418761,-0.1756245345,0.3312980235,-0.2871048152,0.1208339259,-0.1268336177,-0.0563569069,-0.1131872237,-0.2872135043,-0.1494746506,0.054323934,0.3943405747,-0.3943569362,-0.2768113017,-0.0082405061,0.3708256483,-0.0801422074,-0.0053257924,-0.4047894478,-0.254878372,0.5810701251,0.5500808954,-0.2733107209,0.101757817,0.1855352223,0.2776249349,0.1024071723,-0.2307991236,-0.1536742449,-0.1661076993,-0.0927811041,0.1778385788,0.2747174501,-0.0061795176,0.1294382364,0.1172046065,0.0633925796,0.0669287294,0.25828439,-0.0058503817,-0.0609230101,0.2784731686,-0.0916512534,-0.1798344851,0.4046490788,0.1912713051,0.3315361738,-0.1893270463,-0.2400706559,0.1354093403,-0.4178825617,0.3177898526,0.183900252,0.1554373056,0.1418527812,-0.2126809657,0.1653679013,-0.0878929272,0.4779783785,-0.041074194,0.0481498651,-0.5286179185,-0.3487732112,0.3135917485,-0.0111735398,0.097677052,0.2518340647,-0.0679865405,-0.4288959205,0.3860784769,-0.0323410146,0.9558492899,-0.0225028694,0.3515888155,-0.2127989084,0.1410267055,0.4181644619,-0.3968192041,0.0073517226,0.0929678828,-0.1612190455,0.0693853348,-0.0192997679,0.0155805033,0.5426028371,-0.1792238206,0.016451234,-0.1553492695,-0.1700279862,0.0339674205,0.5154805183,0.0222005975,-0.1957754344,-0.271030128,0.2747103274,0.0273143705,-0.1845241785,-0.2289824784,-0.2321474105,-0.0969978645,0.1343789399,-0.4439438581,-0.02583066,-0.090696156,0.020812925,0.0142245134,-0.2122339457,0.4201295376,-0.0418774709,0.2150019407,0.2508339882,-0.1258823723,-0.0159896202,-0.0829438344,-0.0248233527,0.3419751823,-0.1127628312,0.4251574278,-0.0343185738,-0.0647145286,0.0974693671,0.130223453,-0.4447056651,-0.1778984964,-0.0072652381,0.1249743477,0.0954266191,-0.2544194162,0.3434922695,-0.1642208695,-0.1362981349,0.217880398,0.1743402779,-0.088428244,0.0305740144,0.0895982459,-0.1760764569,-0.2223832905,-0.0123737464,0.0476531312,-0.0151196113,0.0274803285,0.4605931342,-0.198206827,-0.2967045605,-0.1264927089,0.2688869238,-0.1782015115,0.1364583224,0.1329890192,-0.1882604361,-0.0775606483,-0.100669764,0.4525688589,-0.0238422938,0.0272304565,-0.074436076,-0.1755801886,0.1121716797,-0.0670982972,0.2054094225,-0.2702157795,0.0308277439,0.0449826568,0.1887643486,-0.4568798542,-0.0358540416,-0.0951191112,0.1719471514,-0.0421656556,-0.0899104849,0.0482700244,-0.0313367173,0.1201030686,-0.1586467773,-0.2573413253,-0.3211759627,-0.1014236137,0.0550560057,0.1992428005,-0.0053389939,0.2748622894,-0.0263861511,-0.2926286459,-0.1035335734,0.1493404806,-0.2090720087,0.0748578608,-0.0019893094,0.3594571352,-0.0098029524,0.2656903863,0.0337566659,0.1791223139,0.0440017544,0.2843107879,0.104788743,0.3512016237,-0.0611014627,-0.1160922498,0.3266392648,0.0138862878,0.3696794808,0.0959681422,0.1060837433,-0.0046598529,0.4014838934,0.2480711639,0.0003906432,-0.121202901,0.0127563654,0.1862211376,0.3145943582,-0.4905485809,0.0201823283,0.0709840208,0.089693673,0.2564951777,0.1013110951,0.2407882959,0.3983295858,-0.2662469447,0.0535799265,0.1283688098,-0.0832779557,0.2789346278,0.0439869724,0.1234266311,0.1276889592,0.2497949898,0.1447680295,0.1770759523,-0.1079279333,-0.112192899,0.2233372778,0.1968613565,-0.1436490566,0.154433161,-0.1914402246,0.0226359405,0.4870806336,0.2763770223,0.2709682286,-0.0751851872,0.1060113013,0.1277764142,-0.0181067996,0.0097329095,-0.0434012078,0.005022665,-0.0997806787,0.2557622194,-0.1935979724,-0.2548975945,0.0115684876,-0.1140682846,-0.2391190976,-0.1452971101,0.1114658341,0.0533444621,-0.3808151186,0.1327935755,0.1021698117,0.0770187005,-0.2044328004,0.1731252819,0.1261539608,-0.0261200648,0.128987968,0.1516894698,-0.0273051802,0.1524228901,-0.0788352564,0.1962604672,-0.0650071651,-0.0121111395,-0.4677234292,0.1735625714,0.1755782515,0.1563398242,0.1874565333,0.198434487,-0.3066502213,0.2008961588,-0.0270945542,-0.0888833553,-0.2535094023,0.3709411323,0.0992707983,-0.2311715335,-0.2780436277,0.2098616809,-0.5753176212,-0.1409056634,0.7585268021,-0.3803829253,-0.0125929294,-0.2522753179,0.1364706755,-0.1514411718,0.4402095377,0.6380358934,0.1436520666,-0.0411110297,-0.0766839609,-0.4288348258,0.0336953886,-0.1114848256,0.3441734612,-0.1771566123,0.4110723138,0.2709236443,0.2339823842,-0.2049842328,0.1825208068,-0.4394817352,-0.1053749919,-0.4030176699,0.0419513844,-0.1118452027,0.4399303496,0.0765123889,-0.015151998,-0.0362367369,0.0228044987,0.1420082748,-0.4613074064,-0.0078921681,0.1073617488,0.2951264679,0.272369653,-0.0705059841,0.1057790369,-0.1023393273,-0.2758906782,-0.5270483494,-0.1035613269,-0.2224801481,0.0660461187,-0.0118253864,0.2248227894,-0.3254526854,-0.0818865746,0.0326576009,0.0385359712,-0.0797877088,-0.1976593733,-0.230678916,-0.0725272372,-0.001549151,-0.0150148459,-0.0211381223,0.1985312253,-0.0059173005,-0.256364584,-0.0262428969,-0.1197190136,0.0835959241,-0.188711971,-0.3257734179,-0.1102397665,-0.0160246585,0.0708808973,0.0067614368,-0.5428434014,0.0760669187,0.3034561276,0.1014957726,-0.0819886476,0.4366681576,0.1028063372,-0.0226395186,-0.1713512838,0.0976246744,-0.1331772506,-0.1530857384,-0.2508779466,-0.2792596817]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1675","title":"Add the 800GB Pile dataset?","comments":"The pile can very easily be added and adapted using this [tfds implementation](https:\/\/github.com\/EleutherAI\/The-Pile\/blob\/master\/the_pile\/tfds_pile.py) from the repo. \r\n\r\nHowever, the question is whether you'd be ok with 800GB+ cached in your local disk, since the tfds implementation was designed to offload the storage to Google Cloud Storage.","body":"## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n","comment_length":45,"text":"Add the 800GB Pile dataset?\n## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n\nThe pile can very easily be added and adapted using this [tfds implementation](https:\/\/github.com\/EleutherAI\/The-Pile\/blob\/master\/the_pile\/tfds_pile.py) from the repo. \r\n\r\nHowever, the question is whether you'd be ok with 800GB+ cached in your local disk, since the tfds implementation was designed to offload the storage to Google Cloud Storage.","embeddings":[-0.3054965138,0.0699859783,-0.1604184061,0.161225304,0.0583351254,0.2056465596,0.1664933413,0.1528969705,0.0987018198,0.0379916802,-0.2099467963,-0.0474001765,-0.5426978469,0.1175630987,0.1921430081,0.0196068771,0.0165603738,-0.1210943982,-0.1276661754,-0.0019259665,0.0934902132,0.0706935674,0.1867747158,-0.4521124065,-0.1669261009,-0.0579509772,-0.2515397668,0.0387953669,-0.0755197853,-0.101735726,-0.2362293303,0.0868579373,0.3087849021,0.6071380377,-0.0001004333,-0.0695983022,0.0087373825,-0.2291680127,-0.3281329572,-0.0924974158,0.0223346297,-0.2448978126,-0.3246685266,-0.0821539685,-0.2599293888,-0.1494977027,0.0683995262,-0.4652225375,0.0738678202,0.2932696044,0.2845729589,0.0713733211,-0.0794752464,0.0155267641,0.3248164058,-0.0263188742,-0.14824,0.3343794048,0.3295799196,0.1583121121,-0.1424901485,0.4226370454,0.2748495936,0.0003275149,0.2505562901,0.0873639584,-0.0578725003,-0.0570576638,-0.0305908471,0.2280007303,0.3887753189,-0.2194996923,-0.4567513764,-0.3718363643,-0.1433018595,-0.4234705269,-0.0850543976,0.2680353224,0.0284770839,0.3589910865,-0.0444991626,-0.3819787204,-0.0959534347,0.0799565613,-0.0777403563,0.3094981611,-0.0480185784,-0.2049657553,0.0719822496,-0.0821328014,-0.1087798402,0.0608082339,-0.0129865585,0.0472399443,-0.2522495687,-0.2563916445,0.2787882388,0.2331831306,0.3749370873,-0.1088677272,0.0821086094,0.0843164846,0.0355407856,-0.0263236314,0.3399219215,0.2041399628,-0.3900613189,-0.2537352741,0.2743602097,-0.0891911983,-0.0249115061,0.0381722935,-0.0104355151,-0.0446640924,0.0524151959,-0.1799826473,-0.2425102741,0.0263723508,0.1079890579,0.0000515683,0.3417654634,-0.0609996729,0.0119816698,0.1120091081,0.0219103228,0.2458597869,0.1014774516,0.0701586008,0.1030491367,-0.3814472258,-0.2073855996,0.1403629035,-0.3521869779,0.0542536937,0.1413021386,-0.0207382943,0.0582091212,-0.3302528262,0.1605791003,0.3598946035,0.2365097255,-0.0915248245,0.0723782778,0.0284404065,-0.0561790876,-0.2589203119,-0.0186113194,0.1120792776,-0.3431424499,0.067394644,-0.0311440565,-0.3437911868,-0.3613585532,0.2279112339,-0.1845790893,-0.0468106456,-0.3756109476,0.6113613248,0.1705449969,0.1302203387,-0.0316049978,0.0682153106,-0.0785706192,-0.0239683427,0.0126636978,0.2618785799,-0.5343835354,-0.0026729736,-0.4895244837,0.1290100217,-0.2406866252,0.3616557121,-0.0843996853,0.0237889271,-0.1299888492,0.0761487484,0.3966332376,-0.0157754458,-0.2992013991,0.2768241465,-0.1049594283,-0.1416923106,0.072183609,0.1846334785,0.1140413657,-0.0324722305,0.2382099628,0.4528157115,-0.0604710542,-0.0016306908,-0.4045578837,-0.5051792264,0.0401144437,0.4231114984,0.1265842915,-0.1892950237,0.237639606,0.0895424336,0.0545406006,-0.0421656445,0.1965840012,0.0215575211,0.288571775,0.1429424286,-0.0291415788,-0.3155581951,-0.587836504,0.1303302944,-0.079941541,0.2485728264,-0.017826356,-0.2165259868,0.1982438266,-0.114903979,0.2045803517,0.0319182612,0.2385396063,0.0654960349,0.1839899719,-0.159246996,-0.1978841871,-0.0996308327,0.0096465666,-0.2245691717,-0.1309184134,0.0693737417,-0.1822370142,-0.1854407191,-0.0076618758,0.0389042795,-0.5003730059,-0.2084504962,0.0860297084,0.2411580533,-0.2672135234,0.0174496602,0.2214572132,0.6760438085,0.2526566386,-0.0342112072,0.208221823,-0.0302790608,-0.0804125145,0.0326918438,-0.268369168,0.2687629759,0.2835396826,-0.1074249223,0.1430109143,-0.159606263,0.239254266,0.0550443046,0.0441909283,-0.328019321,0.2263125479,0.0250183959,0.2887698114,0.0275505055,-0.3536763787,0.234021768,0.3536356688,0.1528300792,0.298683852,0.33782956,-0.0547360517,-0.3734267354,0.1202701256,-0.2180446982,0.3336068392,0.3641212881,0.2784278691,0.0397910289,0.2182824016,-0.1479755193,0.0195082091,-0.1615649462,-0.02755519,0.2779072821,0.4574326575,0.0318875201,-0.2473473847,-0.1694093347,-0.0303081218,0.1659447104,0.0831669345,-0.0231984463,-0.0686668009,-0.1499892771,0.1826805323,0.1798735708,-0.2760950923,-0.0844960362,0.2508695424,-0.00810239,-0.3004891574,-0.0554528795,-0.0380510651,0.5070264339,0.0164479129,-0.0195976291,-0.0083081992,-0.3167681992,-0.1373906583,0.2386581302,0.3709865808,-0.3009616137,0.3809199035,0.0201062653,0.0053937854,-0.1210008115,-0.3262602985,0.2334696203,-0.0303064343,0.3070221245,0.0583283417,0.1241592914,0.151471898,0.0820878819,-0.1439037919,-0.2824913263,-0.0357832462,-0.2636392117,0.118278563,0.0336226746,-0.3632322252,-0.2356481105,-0.5376784801,-0.4607844949,0.2093794197,0.1517200172,0.0935026258,-0.0114702918,0.1578441113,0.2768280804,0.253629297,0.0150033208,-0.251732856,-0.3874271214,0.3524133861,-0.2271275371,-0.2334832102,-0.0054174131,-0.1251117736,-0.050418742,0.1901734918,-0.609280467,-0.1003181413,-0.1245458797,0.2317742705,-0.0553525016,0.0214531515,0.2684258819,-0.0660313815,-0.2008191943,-0.0331001244,0.0089810323,0.2427134067,0.3633560836,0.2199008912,-0.116218552,0.2175286114,0.2947781086,0.4891134202,-0.2678934932,-0.251932323,0.4020811617,0.3352089822,0.2886543572,-0.2938062549,0.1223429963,0.292454958,0.0916426852,-0.2098829448,0.6140494347,0.2262099236,-0.4091822803,-0.0378631093,0.0658067241,0.0875459686,-0.070081003,0.3082033992,-0.1408266425,0.1322071403,-0.0281330831,-0.5833309293,-0.2649160624,-0.2499229163,-0.0137892989,0.1876520962,0.0687693283,0.1264136583,-0.0854870081,0.343885839,-0.3455129564,0.1323498338,0.0159987062,-0.1227381751,-0.2380903065,-0.1587122977,-0.1047981009,-0.1711214632,0.3415547311,-0.3185347617,-0.1155377552,0.032757055,0.2325633913,-0.0489835963,-0.0136708291,-0.2610040009,-0.224687025,0.4029815197,0.5296343565,-0.3832054436,0.0340124816,-0.0429399051,0.3307134211,-0.1017483622,-0.2309643179,-0.1190294996,-0.2533489466,-0.1862151325,0.1779835969,0.2870961726,-0.0126780961,0.1965269595,0.1609135717,0.1286884248,-0.0208329018,0.1567957252,-0.0048182486,0.1233115345,0.1387028545,-0.1354106814,-0.2249894291,0.2443581074,0.1303787977,0.3796402216,-0.129796654,0.1473888308,0.1988919675,-0.3397561014,-0.0037406923,0.2533743382,0.0739241391,0.1150248796,-0.2285284102,0.2516635656,-0.2066577524,0.3521853685,-0.1946512908,0.0684046596,-0.2621572614,-0.2607871592,0.3508726954,0.0376652926,0.0656041056,0.1809139252,-0.1079268903,-0.578687191,0.4100573361,0.1836815476,0.9564607739,-0.1332676262,0.3486765921,-0.1356176883,0.1329642087,0.3100843728,-0.4754393995,0.0331932865,0.1005341634,-0.1023322567,0.0606770553,0.092418991,-0.007769323,0.4886503816,-0.2206032127,0.1095141992,-0.1505853981,-0.0877460912,0.0057056113,0.5700282454,-0.003950261,-0.2818449438,-0.2881192863,0.3334217668,0.0195513479,-0.1422273666,-0.2530405819,-0.1252621114,-0.0284563936,0.2278012484,-0.4362996817,-0.1414025426,-0.1020910814,0.1376575679,-0.1483317465,-0.1609148532,0.1276311427,-0.0775422603,0.2671681345,0.2540808618,-0.2023924291,-0.034013629,-0.0697055981,0.0114882924,0.1946485192,-0.0168610997,0.4102245867,-0.1721378118,-0.0313171186,0.0191245489,0.0555946864,-0.4859297276,-0.1235734522,-0.1589237899,-0.0269786306,0.1921118647,-0.1967698634,0.2476175874,-0.2998570502,0.0078700585,0.2287457287,0.1789417714,-0.0579697154,0.0536750145,-0.0165708549,-0.1226278022,-0.2901830375,0.0806116536,0.1469686031,-0.0448513962,0.0697116926,0.4147088826,-0.1879293472,-0.3427499235,-0.2588159442,-0.0683767721,-0.0841010734,0.2128602266,0.1063573062,-0.2842147648,-0.1736703962,-0.1278042644,0.4878836572,0.1443256438,-0.0150567563,0.0125421556,-0.3584056795,0.0962196514,0.0439115725,0.2152088881,-0.2082722485,-0.0210100114,0.0988937989,0.1707865447,-0.4913286269,0.0373089947,-0.1311993301,0.1306352615,-0.0491909906,-0.1189427078,0.1715483516,-0.014053273,0.1218107045,-0.0293353777,-0.2440227121,-0.3242494762,-0.0493636802,0.0477148741,0.2057030648,-0.0865186453,0.2269387841,-0.1222289652,-0.1621656865,-0.0707706437,0.2802378237,-0.0175998062,-0.0244199876,-0.0470491499,0.2736540437,-0.0562888756,0.2223466039,0.0667048842,0.2154333442,0.0602245182,0.1073815078,0.2274560183,0.3288936317,-0.0791207999,-0.1017635018,0.2705051601,0.0372098535,0.2520081699,0.0831186101,0.0202686768,-0.0796207413,0.3269741833,0.2955205441,-0.1121633947,-0.1510842443,-0.0740488619,0.2572688758,0.3584209979,-0.3802066445,-0.0175964069,-0.0326066799,-0.0205728188,0.3032337427,0.2015851438,0.3765598536,0.2834504247,-0.1388183236,0.2128321081,0.2518252134,-0.0828197151,0.0960178375,-0.1193466932,0.0716375485,0.0877459347,0.1453257948,0.0948277339,0.2414845973,-0.0838297009,-0.096191816,0.1077276543,0.16844666,-0.0799635649,-0.0267886631,-0.2819987237,0.0509507768,0.3446570635,0.3080427051,0.4171383977,-0.1255759746,0.1118625253,0.0843612924,-0.1455593258,0.0150852716,0.0471889228,-0.041543372,-0.1534627527,0.3584344983,-0.1329726726,-0.2213218063,0.197309196,-0.1227588207,-0.2807620764,-0.1509593129,0.1184963658,0.1186973304,-0.1819481999,0.1790840626,0.0902823657,0.1049885154,-0.1272965819,0.3646610975,0.2069532871,-0.072718963,0.0699027479,0.0895766914,-0.0457633212,0.1938197315,-0.1762323976,0.2385967076,-0.0766305774,0.0749088153,-0.4782021642,0.1429201066,0.0959983692,0.0413343944,0.2420551181,0.1821168214,-0.3164461553,0.403277427,-0.0659610406,-0.0439251326,-0.1647762656,0.1812407672,0.187090084,-0.1419082433,-0.1560183764,0.1450746655,-0.6401208639,-0.200298354,0.6081376076,-0.4078819156,0.1453166902,-0.1463488489,0.1537319273,-0.0997353867,0.6128956079,0.4654806554,0.1116485298,-0.0085954489,-0.0731741115,-0.4749073088,0.1111096218,-0.1730326116,0.3495074213,-0.2958161831,0.5248234868,0.1997130811,0.1749622524,-0.1954124868,0.2283548862,-0.4133356512,0.012483866,-0.4292391241,0.1578932405,-0.2818533182,0.4780580401,0.1679557413,-0.1854649186,0.0158092808,0.0171661675,0.1792792231,-0.3698543608,0.0624510758,0.0449266434,0.2174945325,0.2862019241,-0.0621053465,0.1056593508,-0.0976624861,-0.2774540782,-0.4645417035,-0.0273561291,-0.2737820745,-0.0667864159,0.0582765527,0.2316478789,-0.2983383238,-0.1406377852,0.00256123,0.1681706756,-0.0644554794,-0.0705550462,-0.0256327987,-0.0901363418,0.0226222631,-0.0277267583,0.0479574241,0.3123017251,-0.0394278541,-0.167139262,-0.2133657485,-0.1223945692,0.0956287533,-0.1032185405,-0.2685498893,0.002295698,-0.0024153397,0.2139842063,-0.1677069515,-0.5281786323,0.1257605404,0.2320607752,0.1098962277,-0.0873398855,0.3371965587,0.082501471,-0.119318001,-0.1614409834,-0.0165591054,-0.0636978447,-0.0904239938,-0.3649248183,-0.2698542178]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1675","title":"Add the 800GB Pile dataset?","comments":"With the dataset streaming feature (see #2375) it will be more convenient to play with such big datasets :)\r\nI'm currently adding C4 (see #2511 ) but I can probably start working on this afterwards","body":"## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n","comment_length":35,"text":"Add the 800GB Pile dataset?\n## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n\nWith the dataset streaming feature (see #2375) it will be more convenient to play with such big datasets :)\r\nI'm currently adding C4 (see #2511 ) but I can probably start working on this afterwards","embeddings":[-0.4098331034,0.1263111085,-0.1624413729,0.1556787789,0.0171111673,0.1715144813,0.0879780725,0.2739042044,0.011103591,0.019485062,-0.1574598253,0.1408568323,-0.5226863027,0.192629233,0.1664162129,-0.0006090707,-0.0227244943,-0.0548107065,-0.1018029079,-0.0283942707,-0.042558752,-0.0505990051,0.0230189022,-0.4305403531,-0.2151899636,-0.1103021279,-0.1426184922,-0.0110516138,-0.3041887879,-0.2418607175,-0.2696644962,0.1471849829,0.4014956653,0.5093926191,-0.0000992309,-0.1877329946,0.1837464273,-0.1823872179,-0.3778296709,-0.1405142546,-0.0579623953,-0.3062309921,-0.2828385234,-0.0362581983,-0.2138492912,-0.1490485072,0.1165734082,-0.4075603187,0.0837146118,0.3522391617,0.2853194773,0.1292305589,0.028258143,-0.1187565252,0.3448964953,-0.0372173861,-0.1411593109,0.2165897042,0.3609104753,0.1932851821,-0.1349958479,0.42317608,0.3995490372,-0.1363987327,0.2003374249,-0.0137786493,-0.0137635078,-0.1184780449,-0.0939949825,0.3289185464,0.3396664262,-0.1999338567,-0.416947186,-0.3320512474,0.0004093489,-0.4861880541,-0.1864102781,0.2802343667,0.0130024422,0.2654020786,0.0118499016,-0.2891337276,-0.1023034602,0.2159856856,0.0264109615,0.2852253318,-0.0916785821,-0.1814775318,0.1847205013,-0.0452431478,-0.1685019284,0.0620114021,-0.0701507702,0.0169383176,-0.3413780332,-0.2053023726,0.3815687597,0.3479739726,0.4029490054,-0.2013194859,-0.0007641301,0.0900324583,0.0076760538,-0.0363318399,0.4183678627,0.0228529926,-0.2576278746,-0.2429987937,0.3162668347,0.0615197308,0.0845505148,0.0780184045,-0.0812592208,-0.109158121,-0.1546163857,-0.1037386954,-0.131176576,0.1341690719,0.024996819,-0.0110356547,0.2473470122,-0.0406346321,-0.0132628605,0.1060569137,-0.0132587207,0.2237422913,0.0462209955,0.179082036,0.1645918041,-0.5682691932,-0.2182086557,0.1096915454,-0.4494864643,-0.0582360178,0.1746323854,-0.0966311768,0.0979664624,-0.093815811,0.2153852135,0.2556799948,0.1541841477,-0.032111045,0.0867116377,-0.0207196753,-0.0739556402,-0.219107002,0.0245267358,-0.0203439575,-0.2771773338,0.0620876178,0.008937764,-0.2364451438,-0.3791476786,0.2482879758,-0.1052084714,0.0647910833,-0.3975297511,0.6843155026,-0.0518117882,0.2303264588,0.0298150163,-0.0720965415,0.1268395782,0.0094611244,0.0543078631,0.1860060841,-0.5823349953,-0.0688423142,-0.6220449209,0.1332652271,-0.048302792,0.23295708,-0.1152959913,0.0331350043,-0.1026658043,-0.0032313957,0.5506247878,0.0903272852,-0.1896681786,0.1784516424,-0.0943489075,-0.0666787103,0.0360305645,0.166351527,0.1789021492,-0.0276673492,0.2125173658,0.3831298649,-0.1318815202,-0.0458412506,-0.2773420811,-0.3492940664,0.0065038488,0.4455781579,0.0955940038,-0.3235457242,0.2116486728,0.0790048391,0.2011428326,-0.0387859792,0.1662488878,-0.0941343755,0.3228883147,0.0941027626,-0.0326916911,-0.2544544935,-0.4213958085,0.0129303262,0.0664446577,0.3924232721,0.0327239223,-0.2007573843,0.1176968887,-0.0324536301,0.0745065138,-0.0610411279,0.24777475,0.0449669287,0.0116757331,-0.1699575335,-0.2375018448,-0.0224223435,-0.0895086527,-0.1390724927,-0.2041968554,-0.0277161859,-0.211376071,-0.1865868419,0.0991932601,0.2034651339,-0.3931650221,-0.1438767463,0.1542894244,0.1564688385,-0.0995218977,-0.1483729929,0.1739114374,0.589936018,0.2877263427,-0.1312032491,0.2223951221,-0.0966124907,-0.0549186617,0.0523668528,-0.1657312959,0.3457504213,0.2843394279,-0.0630708039,0.1501615494,-0.1299120188,0.2165811807,0.0404009894,0.0111626014,-0.3215346336,0.2626162469,0.017928483,0.2361111343,-0.0284166448,-0.378952831,0.1693694144,0.380448252,0.1784754843,0.2226568162,0.2837972939,-0.2051712424,-0.285794735,0.041163601,-0.2445848286,0.1751973927,0.4254031181,0.2662835717,0.0939711854,0.1909112632,-0.1473923177,0.0472895131,-0.0415800102,-0.0142360535,0.2097346634,0.3988668919,-0.0117707476,-0.3059681058,-0.2796691954,-0.0838693455,0.1702593267,0.125722155,-0.1005367264,-0.0214955118,-0.1116395518,0.1045796573,0.0312195905,-0.2022458166,-0.1765328199,0.1887108237,-0.0844262689,-0.2668888867,0.0207093898,-0.1601111591,0.4609529376,0.0269854292,0.1990112662,0.0462487005,-0.2685699463,-0.1023698896,0.2859627306,0.2906585038,-0.256511718,0.3059352338,0.0507842563,0.1975107938,-0.1086905003,-0.5585720539,0.2651889622,-0.0712607875,0.3164297044,0.2004011571,0.0077922037,0.1900359094,-0.0581640974,-0.1472811252,-0.2070458978,-0.0193798356,-0.153292194,-0.0415223315,0.030775249,-0.2517190576,-0.3398703933,-0.4018904567,-0.5428557396,0.2682358325,0.0710560605,0.0720275193,-0.004570663,0.1488371491,0.3448063731,0.0270824581,0.0566587001,-0.1289211065,-0.3212639987,0.3212068081,-0.149311319,-0.240281716,-0.0765934363,-0.2067047805,-0.0238753948,0.0814590231,-0.501999557,0.1597308815,-0.1152676418,0.2250425965,0.0298296846,0.0394681692,0.1675286591,-0.0220976397,-0.2421857864,-0.0792160332,0.0473500453,0.1114284396,0.3575184047,0.330948025,-0.244711563,0.2910519838,0.2060319632,0.3839906454,-0.0426986925,-0.2878244817,0.406321615,0.3795241117,0.3640135825,-0.2093789876,-0.0007664657,0.2635859549,0.0807814151,-0.0966672227,0.6487154961,0.1601150632,-0.4226080179,0.035568554,0.0923501551,0.0280518942,-0.1093086451,0.3098239899,-0.1567347199,0.0372957699,-0.1065694988,-0.5659021735,-0.2080825269,-0.2272190899,-0.0668059811,0.2432921678,0.0154429302,0.0377539843,-0.1647177786,0.2961852252,-0.3002596498,0.1578607857,-0.0841639638,-0.0066452078,-0.1013781503,-0.2576545775,-0.1832234859,-0.0373369269,0.430185616,-0.3462508917,-0.2166752219,0.050768327,0.3493843079,-0.1120416895,0.0604224689,-0.4069740772,-0.2059410065,0.6046828032,0.4675154984,-0.3034148216,0.0278696734,0.1316444725,0.2617976665,0.1110644117,-0.2128383368,-0.175514847,-0.1478177607,-0.1704755425,0.1762155294,0.2725251913,-0.0027763657,0.0747247711,0.0604748875,0.1321131736,-0.0399940312,0.2389431447,0.0417566113,0.0186076444,0.226168856,-0.0878184363,-0.1445903331,0.3352924883,0.2573812902,0.3048608005,-0.1247100756,-0.1524936259,0.1539131254,-0.3727530241,0.2853825986,0.1742954701,0.1367963701,0.1844531149,-0.2605435848,0.2182884067,-0.1256576329,0.4825087786,-0.0610661805,0.0999267474,-0.4868613482,-0.2704685032,0.3152832985,-0.0656304434,0.0766545683,0.2514153123,-0.0321724415,-0.4504584372,0.2987638414,-0.0284120832,0.9278072715,-0.0053632641,0.2693897188,-0.2089679688,0.0972138047,0.3174709082,-0.4739414454,0.0579383112,0.0496929251,-0.1479236633,0.0744377151,0.0002514924,0.0623320676,0.4972245395,-0.2921587825,0.0621023588,-0.1469142437,-0.1745944619,-0.0159543715,0.5426499248,-0.0461146384,-0.2445471585,-0.303629756,0.3128995299,-0.0087838909,-0.178237319,-0.2275765389,-0.2171167284,-0.0276978407,0.1086652055,-0.3882672191,0.0299073793,-0.1745706499,0.017036451,-0.0396176279,-0.1559927464,0.3619067669,-0.0859262943,0.210545972,0.2843986452,-0.1356962025,0.0590244047,-0.0542575456,-0.0351910032,0.238589406,-0.0666113794,0.3549990058,-0.0717319697,-0.1068443656,0.1439717114,0.1412447691,-0.4076227844,-0.0638784841,-0.1237678677,0.0953618437,0.180132851,-0.1763366908,0.2990343273,-0.230769664,-0.1559379101,0.2434671372,0.1524496526,-0.1237744093,0.0562221073,0.1281016916,-0.1516967416,-0.2540131807,0.0688450336,0.0104803182,-0.0695487484,0.0541339368,0.5405293107,-0.2351927757,-0.3617608845,-0.1462940276,0.1553830057,-0.1333180815,0.156272307,0.1209397316,-0.2216409296,-0.043845173,-0.004107778,0.3809800744,0.014676434,0.0183272995,-0.0393645205,-0.2908758223,0.0985571146,-0.0675146505,0.2288955897,-0.2520004511,0.0393728465,0.042782139,0.242646113,-0.4967220724,-0.0217153598,-0.1245098934,0.1922948807,0.0089229997,-0.1470744163,0.102487497,0.0160690546,0.1619407535,-0.1062357426,-0.2696929276,-0.349036634,-0.112516202,0.0441824645,0.1769063473,-0.0322488695,0.2614036798,-0.0853855163,-0.2715211213,-0.1099483669,0.2870598435,-0.2080384195,0.0509494729,0.043771714,0.3296026289,-0.126469627,0.3109326065,-0.0449662544,0.2604689598,0.1011301503,0.1784499139,0.1243858933,0.3225096762,-0.1056765765,-0.0566259213,0.3306291699,0.0424305797,0.291728884,0.0795196667,0.0630710348,-0.0454287902,0.3690237105,0.2322791964,-0.0418736823,-0.0801910087,-0.0635387897,0.2289940119,0.3526460528,-0.4441505969,0.0261414107,-0.0003549254,0.0660296082,0.3163895011,0.0754539371,0.2048867643,0.3034871817,-0.2193251401,0.1052556038,0.1245333552,-0.0666771159,0.2558591068,0.0101925181,0.0778098926,0.0834266394,0.2537027001,0.1096563786,0.2113311887,0.0206476189,-0.1559851915,0.2174354196,0.2494317442,-0.1217414737,0.1984139085,-0.2705112994,0.0142900981,0.4408418536,0.2304972112,0.334853977,-0.0280814283,0.1581738293,0.1911634356,-0.0242969729,-0.0747416541,-0.0105192363,-0.0324999988,-0.1061252579,0.2486440092,-0.2432185411,-0.2727807462,0.0305127688,-0.1603690237,-0.1590132713,-0.0842064396,0.124872379,0.096042946,-0.3879925609,0.1526907533,0.0539046451,0.0714255422,-0.1914815903,0.281496048,0.1722510606,-0.0372691154,0.0974465534,0.0637184232,-0.0154829854,0.1945394576,-0.1336380541,0.1822951436,-0.0614411831,-0.0422142968,-0.4361591041,0.1846748292,0.2109833509,0.12667422,0.1863869131,0.2063037604,-0.3460839689,0.1753544509,-0.0321185887,-0.1129937023,-0.1564688087,0.2638242841,0.0976850539,-0.1863017082,-0.2644141316,0.1608151793,-0.6029196978,-0.1301615685,0.7552982569,-0.4130332172,0.0005653648,-0.2642828524,0.1572831273,-0.1625027508,0.5178126097,0.4893517196,0.1626828462,-0.0336117335,-0.1965469271,-0.4162887335,0.1185081303,-0.1516164094,0.3029322922,-0.205378592,0.4095591307,0.2892982066,0.2066662908,-0.106643796,0.2366363406,-0.3802986741,-0.0713696927,-0.4423670471,0.0858564824,-0.1462791413,0.4189071059,0.0946603268,-0.1232274771,0.0203778222,0.0184335131,0.1869545132,-0.4501388669,0.0716281235,0.0896570832,0.2031956762,0.2926829755,-0.046856828,0.1830469817,-0.1631487757,-0.2863172293,-0.4637686908,-0.1018195674,-0.2286796719,-0.0029147982,0.0460253507,0.2699045539,-0.2357537895,-0.0778399035,0.0077008149,0.093764551,-0.0757311061,-0.1177697703,-0.1592990011,-0.048414439,-0.0241812468,-0.0201824345,-0.0309325028,0.2383468896,-0.0153583167,-0.2196117938,-0.041362077,-0.2070821822,0.1371308714,-0.1975238323,-0.3689701855,-0.1258336753,0.0563211851,0.0922310278,-0.0358198658,-0.5341690779,0.1531904489,0.2678014934,0.1322012097,-0.1233044043,0.4166258276,0.0848037601,-0.0559648909,-0.167246148,0.1795664877,-0.0280024763,-0.1260197908,-0.3196700215,-0.3176398873]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1675","title":"Add the 800GB Pile dataset?","comments":"Hi folks! Just wanted to follow up on this -- would be really nice to get the Pile on HF Datasets... unclear if it would be easy to also add partitions of the Pile subject to the original 22 datasets used, but that would be nice too!","body":"## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n","comment_length":47,"text":"Add the 800GB Pile dataset?\n## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n\nHi folks! Just wanted to follow up on this -- would be really nice to get the Pile on HF Datasets... unclear if it would be easy to also add partitions of the Pile subject to the original 22 datasets used, but that would be nice too!","embeddings":[-0.2462544739,0.1785694659,-0.1075271368,0.2075333744,-0.0628189445,0.2498984635,0.1501436085,0.1807032526,0.1075051203,-0.0014590707,-0.3454721272,0.0975166634,-0.4746151567,0.2544946969,0.1928560734,-0.0652398095,-0.032202214,-0.2071850151,-0.1370310485,-0.1494104564,-0.06065806,-0.0598101765,0.0282126125,-0.3862361908,-0.1723033935,0.004146033,-0.3139385283,0.0742300451,-0.3200513422,-0.1177404895,-0.1273551136,0.1806448847,0.432210654,0.5957205296,-0.0001031671,-0.1962912828,0.024265008,-0.1736913919,-0.3333541155,-0.1857631207,-0.0035195744,-0.3440428674,-0.4653924406,0.0684908777,-0.1177530363,-0.2050124109,0.12677145,-0.3534972966,-0.0913880542,0.1521177739,0.2241453081,0.149783507,0.0541178472,-0.1607087702,0.4485835135,0.1492411792,-0.130315721,0.273645699,0.3025313914,0.041326385,-0.0410155579,0.3837419152,0.4280636907,-0.1955244541,0.2748353779,-0.0034523201,-0.0380527973,-0.1331808567,-0.1561854482,0.3036518991,0.3889554739,-0.1051894799,-0.487639159,-0.3463985026,-0.107987918,-0.315996021,-0.2576558292,0.412088871,0.0678864419,0.2484545857,0.1746350676,-0.3019080758,-0.1186246648,0.2390800864,-0.0445908681,0.3197979033,-0.0408892706,-0.1356092095,0.1942106783,-0.0039509782,-0.0605598278,-0.0825138167,-0.1060341001,0.1061292589,-0.261744678,-0.1970933378,0.33981511,0.360209614,0.4732288718,-0.343878895,-0.1671775728,0.0080113756,-0.1477238685,-0.0648703426,0.5987275839,-0.0423334874,-0.3046991527,-0.2190734744,0.2182767987,0.0477403998,0.1607235819,0.0447870195,-0.0993976444,-0.0348435156,-0.2174886316,-0.2602546215,-0.1236837283,-0.0555649623,0.0519261882,-0.117215611,0.1994657964,-0.1382020265,-0.0506327599,0.1108685061,0.1273375005,0.2053894997,-0.0192008335,0.2568579912,0.121724695,-0.4795180261,-0.1827813089,0.1190001294,-0.4479356706,-0.1070014089,0.101488933,-0.1534747928,-0.0588073991,-0.0950969383,0.2930878401,0.370467931,0.1002270803,0.0000388057,0.1805009544,0.0745559484,-0.0098785674,-0.1554697156,0.0954120681,0.0298353042,-0.3180525601,0.1465946734,-0.05322586,-0.2526592016,-0.4137487411,0.1962387115,-0.1080147251,-0.001286297,-0.5398012996,0.6560422778,-0.0744825155,0.3185945153,0.0379105546,-0.0938696861,0.1458636522,-0.0064918697,0.0280066803,0.1340823323,-0.6778559089,-0.0360390209,-0.666544497,0.0231845882,-0.0199850611,0.2567601204,-0.1694158167,0.1129339933,-0.2431821823,-0.0524932332,0.5646255612,0.2066770047,-0.1768901497,0.2358557731,-0.0061378046,0.0615539253,-0.0677525848,0.1928419024,0.0997923687,-0.0896512419,0.2306838632,0.3897214234,-0.1044965535,-0.153836742,-0.2931567729,-0.2979383171,0.0302868839,0.477201134,0.1861846894,-0.2131325454,0.3239218593,0.0539867505,0.2610428631,-0.0270515569,0.1740340292,-0.1687909812,0.4095945954,0.2165106237,0.0469269343,-0.1935575008,-0.3710539043,0.0883245543,-0.0564857982,0.4205211699,0.1146350726,-0.1526171118,0.1054009423,-0.0260515492,0.1093332246,-0.0851702541,0.1698955446,0.0053043626,-0.0936794803,-0.3116691113,-0.2342717797,0.0772787854,-0.0754422694,-0.0829077065,-0.1737835258,0.1861634254,-0.2260220349,-0.1404219568,-0.0134574817,0.2356663048,-0.4520278573,-0.1736225188,0.1470247209,0.0358707085,-0.1718818694,-0.1805039495,0.3125281334,0.593639791,0.2700743377,-0.1895742565,0.3841978312,-0.2826440334,-0.0948998332,-0.037112467,-0.0977637842,0.3466219306,0.141826272,-0.0143075706,0.0712347478,-0.103793636,0.1114261001,0.1357508451,0.0565716289,-0.2372091562,0.2847626507,0.0584240519,0.1625307053,-0.0734737441,-0.3168832362,0.2308573872,0.2888997793,0.188281104,0.1037332714,0.2862325013,-0.2110560387,-0.2141566426,0.0827909186,-0.3201047778,0.29331249,0.4313607812,0.3334782422,0.1507135481,0.1923593432,-0.0678071603,-0.0338193439,0.0887047648,-0.0418704785,0.1549538523,0.4150850773,0.0493681133,-0.2622881234,-0.2563100457,-0.0996499062,0.1201703921,0.0827069953,-0.0965997875,0.0342443101,-0.1537794769,0.0391568877,-0.0015208692,-0.2632212043,-0.2255914509,0.1800459027,-0.1216264814,-0.2089032829,0.0168169048,-0.1368888617,0.4582442343,-0.148108691,0.2264863104,0.0739084706,-0.2982954085,-0.0681068748,0.2394589633,0.3514464796,-0.1933685243,0.383833766,0.0832781866,0.0436142683,-0.1888152659,-0.5572310686,0.2148355544,-0.131484881,0.3213920295,0.1718160957,-0.0137503073,0.1581781954,-0.0338955112,-0.1430461556,-0.0654320568,-0.0057403627,-0.2261835486,-0.0729996413,0.0239962153,-0.2119414508,-0.2607219219,-0.4658943713,-0.4638248384,0.2482122332,0.0824124664,0.0538888127,-0.0242492966,0.0566325188,0.2967192531,0.0132699283,-0.1570475996,-0.1987161487,-0.3436394632,0.340357244,-0.1229745075,-0.2107488066,0.0278549567,-0.239579916,-0.1833103448,0.1086484417,-0.594632566,0.1073942259,-0.0048695211,0.2758696973,0.0069912579,0.1516615152,0.1241982505,-0.1518832296,-0.1421293616,-0.081577599,-0.0377443545,0.1081052721,0.5077723861,0.3683866858,-0.1968954504,0.2154929489,0.2673070133,0.5058031678,-0.0379618742,-0.1533984244,0.4270523787,0.5014266372,0.392534107,-0.238448143,-0.0938844606,0.3584389985,0.0730214864,-0.0942887217,0.5424857736,0.2252040207,-0.3714762032,0.0863374025,0.0207918324,0.1657278091,-0.0736840665,0.2951665819,-0.1628871113,0.0000465995,-0.0031870066,-0.6648690701,-0.2963987887,-0.2286062092,0.0514846891,0.3544838428,0.1379663646,0.0954989567,-0.2339651287,0.3559135199,-0.2122379541,0.1766771078,-0.1295895278,-0.0875875279,-0.0914662406,-0.1989506632,-0.0549692065,0.1005001888,0.5109840035,-0.3662701249,-0.3167026937,0.1177468598,0.3514556885,-0.0848128498,-0.0594587438,-0.4126622975,-0.2113266736,0.5140792727,0.502702415,-0.3331848979,0.0236152392,0.1986258924,0.259747982,0.0325455405,-0.3211179972,-0.0426874608,-0.204508841,-0.10946244,0.1658123732,0.2859381437,-0.0510991588,0.0039805938,0.1611012518,0.1152085662,0.0895007923,0.2027987093,-0.1108847037,-0.0331287794,0.2975577116,-0.0294309035,-0.1174037009,0.4459313452,0.3497481048,0.4114852548,-0.2457731068,-0.1615236998,0.2627890408,-0.4169073999,0.3343498707,0.228466332,0.2370265871,0.0411271565,-0.2406501472,0.1091442853,-0.1777591258,0.3633443415,-0.0939075202,0.0275256224,-0.5469814539,-0.237522155,0.3390755057,-0.0768830478,0.0732175708,0.2748345435,0.1847644746,-0.4231977165,0.5059276819,-0.1159733906,1.0700716972,-0.080769375,0.3042270839,-0.217544049,-0.0211432781,0.2948868573,-0.4703840017,-0.0475963578,-0.0133465063,-0.2273612171,0.0808859169,-0.0867928565,0.0344966911,0.5482620597,-0.0845296904,0.0840836689,-0.1508916169,-0.170410648,0.0115116686,0.4651859403,-0.0262098461,-0.2943552434,-0.2766018212,0.2728030384,0.0622450635,-0.0274280626,-0.3297658861,-0.2247918248,-0.0615264438,0.1578472108,-0.4502339661,-0.0080671124,-0.078588821,-0.0440625623,0.0156752467,-0.1983697563,0.4680590332,-0.139866665,0.3715138733,0.1521102339,-0.0564447083,0.0242805704,-0.0350761898,-0.0258218423,0.3160747886,-0.0757735372,0.4437025189,-0.0432162732,-0.1435642093,0.0124272965,0.0758765489,-0.4124035835,-0.2277436107,-0.0692635775,0.1264894009,0.114988707,-0.1996549666,0.3787492514,-0.0596692264,-0.1180609092,0.2050382495,0.2202064097,-0.1078031212,0.084893994,0.0587841682,-0.1283051968,-0.2801240087,0.0071991021,0.0045524668,-0.1225725859,0.0323063135,0.5429108739,-0.2912156284,-0.2394305915,-0.1506330818,0.2842727005,-0.1853858978,0.0086245798,0.1156838536,-0.1242171675,-0.1738389283,-0.0819338262,0.4243218899,-0.0707866699,-0.0540440157,-0.1348333955,-0.1471021622,0.136967361,-0.2005058527,0.2099662274,-0.2749506235,0.1014334187,0.0236300807,0.2915304899,-0.4439957738,0.1228215471,0.0282780323,0.1778004616,-0.0181761924,-0.0624047481,0.0216264389,-0.110428147,0.1079770103,-0.122729823,-0.2240716666,-0.2821012139,-0.1286136061,0.0871806592,0.2167997807,-0.0982912555,0.2894611955,-0.1814132631,-0.2590300441,-0.0796576738,0.2726959288,-0.103851676,0.0400890037,0.0235557854,0.5215920806,-0.0635343194,0.2113538384,0.1882427931,0.35488832,0.1207743958,0.1915183067,0.0807865486,0.2534691393,-0.0892502815,-0.1287544817,0.3614580333,0.0671223253,0.3614724576,0.0984225422,0.0668548495,0.0416596048,0.3551243544,0.273886323,-0.082525745,-0.0077844015,0.0128559778,0.2061582208,0.270334512,-0.4512035847,0.0081876526,0.1076396629,0.0954559743,0.1759246588,0.1175028235,0.3312568963,0.4307403862,-0.2234125733,0.1161267012,0.0938006938,-0.0999250412,0.2012495399,0.0098465625,0.0189983454,0.110868901,0.1790065318,0.0791346431,0.1057787091,-0.1001973003,-0.1811123639,0.1659988016,0.2869562209,-0.1682220846,0.2207459211,-0.3757106662,-0.0267041028,0.515219152,0.4073400497,0.2285590023,-0.0850902647,0.1573411226,0.1210770234,-0.0336671509,0.0843485966,-0.0181367323,-0.0292824749,-0.0162932333,0.4008811712,-0.2416741699,-0.3885984123,-0.0067313216,-0.1727814972,-0.3297215402,-0.073975727,0.0763520151,0.0797685161,-0.3823194802,0.1355735213,0.0760656148,-0.0144498646,-0.1072490737,0.2297531366,0.0830257609,-0.0512251556,0.2122186571,0.0801176354,-0.0374674164,0.0851842389,-0.1067230031,0.1989655942,-0.0075915833,-0.0174301136,-0.3655427992,0.1198865324,0.0601712354,0.1083435118,0.2009350806,0.1586287171,-0.2674557269,0.050679937,0.0067178216,-0.0756036192,-0.2069849372,0.4589696527,0.0882485807,-0.2043780982,-0.1409869194,0.1264381558,-0.4311354756,-0.2203212082,0.704680264,-0.3813634217,0.03679033,-0.1178375855,0.1220143065,-0.1921441108,0.3321495056,0.6509487629,0.038350597,-0.1129873022,-0.0989616811,-0.5288535953,0.0839564279,-0.1306719631,0.3303055763,-0.2055213898,0.4203825295,0.4353750646,0.129807204,-0.0557136759,0.0865161866,-0.2963830233,-0.0059642107,-0.4423348308,0.0430950075,-0.097234875,0.4771197438,0.0220445748,-0.0307972487,-0.04293615,0.1832573116,0.0724949986,-0.3979876339,0.0213230699,0.1321704388,0.0940716267,0.182655558,-0.0758395121,0.2021499723,-0.0717784688,-0.2632469535,-0.4402863979,-0.262280345,-0.1788100004,0.0394663736,-0.1695528924,0.2703260183,-0.3259003162,-0.2024612278,0.0125678908,0.0474358127,-0.1753879786,-0.0761072412,-0.276617974,0.0672173724,-0.0704903156,-0.0284488872,0.0527558103,0.221197769,0.1244694293,-0.1655367464,-0.0721092001,-0.0642676651,0.1269426793,-0.0891807526,-0.393109262,-0.1120236889,0.0302626602,0.0496613011,-0.0979816988,-0.5905724168,0.10672158,0.3350083232,0.0582951717,0.0444456227,0.3272448778,0.1226269528,0.0199758951,-0.191902101,0.0963458866,-0.0984059721,-0.1776347756,-0.2727271616,-0.2248992771]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1675","title":"Add the 800GB Pile dataset?","comments":"Hi folks, thanks to some awesome work by @lhoestq and @albertvillanova you can now stream the Pile as follows:\r\n\r\n```python\r\n# Install master branch of `datasets`\r\npip install git+https:\/\/github.com\/huggingface\/datasets.git#egg=datasets[streaming]\r\npip install zstandard\r\n\r\nfrom datasets import load_dataset\r\n\r\ndset = load_dataset(\"json\", data_files=\"https:\/\/the-eye.eu\/public\/AI\/pile\/train\/00.jsonl.zst\", streaming=True, split=\"train\")\r\nnext(iter(dset))\r\n# {'meta': {'pile_set_name': 'Pile-CC'},\r\n# 'text': 'It is done, and submitted. You can play \u201cSurvival of the Tastiest\u201d on Android, and on the web ... '}\r\n```\r\n\r\nNext step is to add the Pile as a \"canonical\" dataset that can be streamed without specifying the file names explicitly :)","body":"## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n","comment_length":92,"text":"Add the 800GB Pile dataset?\n## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n\nHi folks, thanks to some awesome work by @lhoestq and @albertvillanova you can now stream the Pile as follows:\r\n\r\n```python\r\n# Install master branch of `datasets`\r\npip install git+https:\/\/github.com\/huggingface\/datasets.git#egg=datasets[streaming]\r\npip install zstandard\r\n\r\nfrom datasets import load_dataset\r\n\r\ndset = load_dataset(\"json\", data_files=\"https:\/\/the-eye.eu\/public\/AI\/pile\/train\/00.jsonl.zst\", streaming=True, split=\"train\")\r\nnext(iter(dset))\r\n# {'meta': {'pile_set_name': 'Pile-CC'},\r\n# 'text': 'It is done, and submitted. You can play \u201cSurvival of the Tastiest\u201d on Android, and on the web ... '}\r\n```\r\n\r\nNext step is to add the Pile as a \"canonical\" dataset that can be streamed without specifying the file names explicitly :)","embeddings":[-0.3223569691,-0.0436929688,-0.1469397694,0.0411979742,0.2053308487,0.0567985885,0.1199193895,0.2633144259,-0.0495686866,0.1010038331,-0.2534806132,0.1415169835,-0.453261137,0.1955078095,0.157208994,0.0147248944,-0.0914569497,-0.0514797457,0.0360941589,-0.0275311302,0.1478369832,-0.0608970076,0.132908076,-0.3931410313,0.1203901395,0.0574042201,-0.1081212461,0.0364227742,-0.3161754906,-0.2588157058,-0.2680271268,0.0863038376,0.3348543942,0.5441961884,-0.0001011025,-0.1211931184,0.2058726847,-0.3016029894,-0.3517262042,-0.291303277,0.0903971344,-0.2420322001,-0.3066551089,-0.130350709,-0.1966381073,-0.2078080326,0.1834151745,-0.3442660272,0.28452003,0.341537267,0.2460384667,0.1315052509,-0.0304514002,0.0425570235,0.3156402409,-0.0613367409,-0.1409148872,0.2879666388,0.4384595454,0.131225422,-0.0695037246,0.390666008,0.2101586908,-0.0690988302,0.0879188105,-0.102880843,0.0603302978,-0.2037926316,-0.090234071,0.4779412746,0.2258951813,-0.3107229173,-0.4418420494,-0.2665180266,-0.0485248864,-0.5264925957,-0.2674874663,0.4536276162,-0.1578510851,0.2263097614,0.0106866965,-0.1838753074,-0.1374783814,0.2025263608,0.0351605341,0.2958590686,-0.050167013,-0.1749762297,0.0153176319,-0.0711371228,-0.202226758,0.0046313517,-0.0453153811,0.0850836933,-0.4025754035,-0.248159647,0.3415242136,0.1628439724,0.331109643,-0.1066936255,0.2155127078,0.1476169676,0.0707817525,-0.0287207086,0.450353533,0.0320552066,-0.2049900591,-0.3580890894,0.2392516732,0.2649254799,0.0162714999,0.0166314021,-0.0751740858,0.0401686393,-0.1365391314,-0.0610720403,-0.0502994284,-0.0035038099,-0.0086774798,-0.0772495791,0.0323535725,0.0265965629,-0.0499776192,0.1562220752,-0.0496002212,0.3794095814,-0.0218507051,0.2079280466,0.0882696956,-0.436363548,-0.1796023548,0.0789486542,-0.3663024902,-0.1088271961,0.1638758481,-0.1643036902,0.1702489853,-0.2545806766,0.3023360968,0.2581272721,0.1764388829,-0.0198114831,0.0513743199,0.0377178267,0.1977613866,-0.2589660585,0.0989611,-0.1191890612,-0.3118190169,0.0759711415,0.030774245,-0.2165573686,-0.2595750988,0.2630078197,-0.2195437998,0.1194420084,-0.339209348,0.5862973332,0.0106500732,0.1276053786,-0.0754742846,-0.0523760803,0.0848475993,-0.0357882008,0.0291658957,0.1398172826,-0.5509948134,-0.131236732,-0.5367796421,0.1494332552,0.0903180093,0.1384186596,-0.0874186307,-0.0488986634,-0.0450920947,0.045368772,0.5971069932,-0.0393128358,-0.3197206855,0.1944663227,-0.2178276479,0.082122609,0.1501761079,0.1490412802,0.3288455009,-0.0766837522,0.1707689464,0.4196424782,-0.082190901,0.0360239856,-0.2637160718,-0.4479620457,-0.0619452186,0.4485101998,-0.0011915879,-0.2839028239,0.2976228297,0.1281043142,0.3932348788,-0.0307257567,0.1780168116,-0.0209076554,0.3714098632,0.2155564427,-0.1040035263,-0.2432037741,-0.3883219659,-0.0004523165,0.1256648004,0.3175265491,-0.0576177426,-0.176865831,0.0995964855,-0.0897155628,-0.0054206564,-0.0083347308,0.2842760682,0.215275079,-0.0720103234,-0.205806613,-0.2596538067,-0.0198443774,-0.2136233747,-0.1910217255,-0.2605654895,0.0733034611,-0.2180713862,-0.2537564933,0.15475218,0.0511855371,-0.3584303856,-0.0731421113,0.0983453393,0.1717747003,-0.1573737264,0.0545240566,0.2587210834,0.6038361788,0.2455324531,-0.2508397102,0.1503962874,0.1796720177,0.0597122386,0.0767359883,-0.328433156,0.2855479419,0.2483708709,0.0522318855,0.2417669296,-0.123791337,0.3073280752,0.0731959119,-0.2831407785,-0.1669170409,0.3048636317,0.0451128893,0.0771817043,-0.1612460613,-0.4005177617,0.265421629,0.4768961966,0.1106094122,0.1505624652,0.3148940504,-0.1522031128,-0.283857882,0.0053650695,-0.2078521997,0.1680807769,0.3904982805,0.3401862085,0.0501147099,0.1904364824,-0.1668256521,0.0935665965,-0.0266901311,-0.1783637851,0.3081232309,0.4929266572,-0.0784886405,-0.3999209702,-0.244451493,-0.1103419811,0.0061995997,0.1426928788,-0.0133943185,0.0114843175,-0.1119273305,-0.0182371214,-0.1914657056,-0.1208035871,-0.2223506719,0.1170553789,0.1016682386,-0.262424171,0.0269336216,-0.1841677874,0.2350763977,0.0711468831,0.069557935,0.0398696773,-0.3471658528,-0.0365973935,0.2696183026,0.1716452986,-0.1558896452,0.2812961638,-0.0741847456,0.0166828353,-0.0781815127,-0.4989754558,0.266802907,-0.0905107483,0.2021902055,0.1259572804,0.2397974133,0.1978953332,-0.0552802645,-0.0750290602,-0.3651872277,-0.0118545061,-0.0598126352,0.034238115,0.0973975956,-0.2828164399,-0.3978762031,-0.4410095215,-0.5356467366,0.2150579691,0.1740148664,0.1467034072,0.0685802847,0.1819939911,0.2245611697,0.1612925828,0.1206091717,-0.1552394032,-0.4950091541,0.3499086499,-0.1595913619,-0.3062624931,-0.0049661868,-0.014390111,-0.0345366821,0.1636470258,-0.4432415664,-0.0470626205,-0.0167907476,0.0818646252,0.0182889774,0.104462944,0.118752338,0.0270555075,-0.2058415711,-0.097542949,0.1790212393,0.0209461302,0.4673111439,0.3781637847,-0.09892492,0.4266079664,0.2625722587,0.4059220254,-0.0381824225,-0.1138415709,0.4830045402,0.346957922,0.3760917485,-0.1459474415,0.0600003265,0.2370053083,0.002286854,-0.0762184486,0.5582517385,0.1973243654,-0.4028145075,-0.0382022858,0.2850158215,-0.0976750329,-0.0763764828,0.2500544488,-0.0523850545,-0.0720954537,-0.0710293502,-0.4156255126,-0.1283507645,-0.241866827,-0.0514256544,0.2994118631,-0.0427013859,0.1738257557,-0.1776525825,0.2880126238,-0.3454490006,0.0637273192,-0.0849911049,-0.0478322729,-0.0885403976,-0.3148698211,-0.0483548008,-0.1767076105,0.3204275966,-0.1517396122,0.0316219553,0.045060467,0.3639653325,-0.0953784063,0.0259391163,-0.5017448068,-0.2487821281,0.5851411819,0.3097583652,-0.4286117256,0.1080940515,-0.0699658394,0.284966737,0.089711085,-0.0237537306,-0.28659603,-0.3304406106,-0.2712455094,0.1759819835,0.2770758569,0.0879162252,-0.0682353377,0.1106632128,0.1559013277,0.0852566212,0.2094977796,0.037067607,0.0897322446,0.2009016424,-0.1001159549,-0.1530419886,0.2777677774,0.1319482774,0.4925675392,0.0307235811,-0.2574204803,0.1374424696,-0.2497544885,0.0639543608,0.3131730258,0.2076773942,0.005387052,-0.1623273939,0.3597307205,-0.0301491115,0.5377065539,-0.0259130914,-0.0057319878,-0.3113618791,-0.3881619573,0.2694009542,-0.1336192638,0.1377859116,0.2873788774,-0.1377398223,-0.4819787443,0.350813061,0.1015332416,1.157471776,0.0225036591,0.3382509351,-0.0684627518,-0.0187517069,0.1488356441,-0.5433734655,0.1053291559,-0.0305756871,-0.1621502042,0.0120473709,0.0488125086,-0.0241122358,0.5330657959,-0.3105174899,0.2377952337,-0.0263203513,-0.0531155691,0.0488952994,0.4132896662,-0.1325357407,-0.2765012085,-0.4921767116,0.3025732934,-0.0954798162,-0.1069525927,-0.1177936941,-0.3090376854,-0.0039090374,0.0507025644,-0.3224793673,0.1603117734,-0.1182793081,0.0092114015,-0.2497262955,-0.2439662516,0.0596254691,-0.0124506578,0.0123518538,0.2281520218,-0.0170158278,0.1412080377,0.1210275814,-0.009968034,0.2040768564,0.0225620717,0.185852915,-0.1174793318,-0.2016074061,0.2129385769,0.2023432851,-0.3256718516,-0.1188547164,-0.1152733937,0.1606315225,0.3437890708,-0.1396983713,0.1534810364,-0.306320101,-0.1026401073,0.2062493861,0.0449412391,-0.253957361,-0.0041741608,0.0636998937,-0.0873417109,-0.245027408,0.181099385,-0.0150198918,0.0991083682,0.1201462224,0.5236203671,-0.2086773664,-0.3327823579,-0.1803805679,-0.0764839277,-0.011857979,0.0165105835,0.0068746586,-0.2667149603,-0.0784184188,0.1140018478,0.2360067368,0.1239779368,-0.0908996984,-0.137626633,-0.2385466099,0.0986465737,0.1860430688,0.2600179017,-0.1946840882,-0.0757202283,0.0515561551,0.0670828223,-0.4871875346,-0.0137961544,-0.0547762811,0.1994472444,-0.0557226911,-0.151614517,0.2296517789,-0.03277722,0.2004781067,-0.1023901924,-0.3184906244,-0.3430956602,-0.0756846443,0.0706375986,0.245582357,-0.079261981,0.0490383208,-0.0899716541,-0.2021515965,-0.0873928145,0.2722050548,-0.0695240051,0.1285642982,-0.0134236384,0.2994030416,0.0349527523,0.2877044678,0.0054352591,0.1629658192,0.1627283692,0.1801445335,0.0608673431,0.3414309621,-0.0923992023,-0.1042084992,0.1743165404,0.1463618279,0.3186542392,0.1799396873,0.0390047245,-0.048608236,0.2051264644,0.3666827977,0.177592665,-0.1128845885,-0.1774240583,0.2098622918,0.3532809019,-0.4156755507,-0.0538531281,-0.038775228,0.0090608317,0.3116801679,0.0625318661,0.0261146687,0.2102063447,-0.2683260441,0.1273122728,0.2595116198,-0.0007237605,0.2829007506,0.0719393864,0.040477775,0.0894797817,-0.0437463,0.067051053,0.2175388932,-0.021305237,-0.1862262189,0.2245434374,0.0799612775,-0.2385454923,0.0303309206,-0.1407351345,-0.1264450699,0.4498063326,0.098227255,0.547855258,0.0159651618,0.117114976,0.1456205398,-0.0917285234,-0.1369493157,0.0603200756,0.1622690558,-0.1530496627,0.1642970741,-0.2015284002,-0.2752503455,0.1849632412,0.0402365513,-0.0874898657,-0.1899811625,0.1981969476,0.1518744677,-0.3833711445,-0.1464333832,0.1075043604,0.1457350701,-0.2104996145,0.2389494032,0.4500501752,0.0200156439,-0.0056789862,-0.0901041254,0.0518383011,0.1532124579,-0.1451708078,0.2344295681,0.0289015137,-0.0984183103,-0.3744084239,0.0603867397,0.1292493641,0.0513804443,0.2384419143,0.2543418407,-0.3856631219,0.2379293442,0.0397523977,-0.2122367471,-0.0748326257,0.304633379,0.0831705481,-0.0551889092,-0.2444656938,-0.0099749109,-0.7704511881,-0.0414515845,0.4132132232,-0.2534708977,0.0785747021,-0.2475295365,0.1518910378,-0.1163163558,0.4799453914,0.3843393326,0.026327271,0.0351226963,-0.1932334453,-0.5093250275,0.0783378482,-0.1550762802,0.162762329,-0.2799509764,0.4626987278,0.2352755517,0.248616457,-0.2323393226,0.2263725698,-0.3380233347,0.1256545484,-0.3879914582,-0.01444891,-0.0728383586,0.5229436755,0.1007208973,-0.2218314111,-0.047730051,-0.0742279291,0.2376425713,-0.2841776013,0.0521264486,-0.0273147468,0.0885236561,0.2240050584,-0.2203160226,0.0539591797,-0.2076574564,-0.4573860168,-0.3317125738,-0.0282235984,-0.3018070459,0.0516992807,-0.1037742943,0.4044714868,-0.1579030454,-0.0006497208,-0.0794937611,0.1015086919,-0.0591059327,-0.2413428426,-0.080272764,-0.1176308095,0.0569489263,0.1257924289,-0.1227526814,0.2547261119,-0.0747400075,-0.1236220002,-0.1223706529,-0.3211065233,0.3737632334,-0.2486292869,-0.3403843045,-0.0965086892,0.040513359,0.0772689357,0.0772471353,-0.5529177785,0.3258179724,0.2028731257,0.067170307,-0.1410716027,0.3578510284,0.0323955417,-0.0576107986,-0.1391894668,0.1053263098,0.1142416745,-0.1572898179,-0.251586169,-0.3276928663]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1675","title":"Add the 800GB Pile dataset?","comments":"> Hi folks! Just wanted to follow up on this -- would be really nice to get the Pile on HF Datasets... unclear if it would be easy to also add partitions of the Pile subject to the original 22 datasets used, but that would be nice too!\r\n\r\nHi @siddk thanks to a tip from @richarddwang it seems we can access some of the partitions that EleutherAI created for the Pile [here](https:\/\/the-eye.eu\/public\/AI\/pile_preliminary_components\/). What's missing are links to the preprocessed versions of pre-existing datasets like DeepMind Mathematics and OpenSubtitles, but worst case we do the processing ourselves and host these components on the Hub.\r\n\r\nMy current idea is that we could provide 23 configs: one for each of the 22 datasets and an `all` config that links to the train \/ dev \/ test splits that EleutherAI released [here](https:\/\/the-eye.eu\/public\/AI\/pile\/), e.g.\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n# Load a single component\r\nyoutube_subtitles = load_dataset(\"the_pile\", \"youtube_subtitles\")\r\n# Load the train \/ dev \/ test splits of the whole corpus\r\ndset = load_dataset(\"the_pile\", \"all\")\r\n```\r\n\r\nIdeally we'd like everything to be compatible with the streaming API and there's ongoing work by @albertvillanova to make this happen for the various compression algorithms.\r\n\r\ncc @lhoestq ","body":"## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n","comment_length":199,"text":"Add the 800GB Pile dataset?\n## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n\n> Hi folks! Just wanted to follow up on this -- would be really nice to get the Pile on HF Datasets... unclear if it would be easy to also add partitions of the Pile subject to the original 22 datasets used, but that would be nice too!\r\n\r\nHi @siddk thanks to a tip from @richarddwang it seems we can access some of the partitions that EleutherAI created for the Pile [here](https:\/\/the-eye.eu\/public\/AI\/pile_preliminary_components\/). What's missing are links to the preprocessed versions of pre-existing datasets like DeepMind Mathematics and OpenSubtitles, but worst case we do the processing ourselves and host these components on the Hub.\r\n\r\nMy current idea is that we could provide 23 configs: one for each of the 22 datasets and an `all` config that links to the train \/ dev \/ test splits that EleutherAI released [here](https:\/\/the-eye.eu\/public\/AI\/pile\/), e.g.\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n# Load a single component\r\nyoutube_subtitles = load_dataset(\"the_pile\", \"youtube_subtitles\")\r\n# Load the train \/ dev \/ test splits of the whole corpus\r\ndset = load_dataset(\"the_pile\", \"all\")\r\n```\r\n\r\nIdeally we'd like everything to be compatible with the streaming API and there's ongoing work by @albertvillanova to make this happen for the various compression algorithms.\r\n\r\ncc @lhoestq ","embeddings":[-0.2282571793,0.1881410331,-0.0650017411,0.1254124492,0.0126015767,0.2586630285,0.3119372427,0.3322311938,0.1376132816,0.0271330234,-0.4386693537,0.0903872401,-0.4766185284,0.3780531883,0.2840977013,-0.1374839693,-0.0587182418,-0.250131458,-0.0747230574,-0.1796042472,0.0673381463,-0.1525621563,0.0313818492,-0.3628998101,-0.3110481203,0.0765463263,-0.3312490582,0.0986899137,-0.1548926532,-0.1428652108,-0.0868680477,0.2483041584,0.4204017222,0.643060267,-0.0001067324,-0.1299245059,-0.0083429599,-0.2832959592,-0.2589152455,-0.1079280898,0.0129947169,-0.2474836558,-0.3683611453,-0.0455847569,-0.139142707,-0.1270540804,0.1864703298,-0.3560496271,0.0931270495,0.1874836534,0.1514540911,0.1044343412,-0.0270650294,-0.0475407988,0.4256104231,0.1131251678,-0.0817006156,0.3710246086,0.3601985276,-0.0385705717,-0.197845459,0.4168069661,0.4534835219,-0.2206376791,0.1778477281,0.0370863788,-0.0558865964,-0.3194620311,-0.2085234523,0.3520064056,0.1860892922,-0.2281012535,-0.5114623904,-0.390768528,-0.0690834001,-0.274343878,-0.2978233397,0.455956459,-0.0434006527,0.2798133194,0.1603559405,-0.3045364618,-0.2127908766,0.2923324108,0.1032138839,0.3265909553,0.0732522011,-0.0842014253,0.1150962561,0.042542316,-0.0186907537,-0.0760523155,0.0499152653,0.1292003691,-0.2761754096,-0.2716826499,0.3116219938,0.2378759086,0.3485050499,-0.2937188745,-0.0690670311,0.1573792249,-0.110458605,-0.0658649728,0.5716243386,0.0531895086,-0.376991421,-0.1046238616,0.2488781959,0.1903036386,0.0661697388,0.1693439037,-0.1435924768,-0.0210490357,-0.2816229165,-0.1246438771,-0.0998213366,0.0529648215,0.042696856,-0.1287230849,0.0037251222,-0.1155982316,0.0569202937,0.1292072982,-0.0793142021,0.3255789876,-0.0271163043,0.3327146173,0.0585907288,-0.3411454558,-0.159672007,0.2507674992,-0.439252466,-0.0159345046,0.2574296594,-0.2227126807,-0.0154155409,-0.056561213,0.3749923408,0.2365031987,0.1827092171,0.0334611349,0.0821232721,0.0265833028,0.2038087994,-0.2123887986,0.0997455791,-0.00956822,-0.3655012548,0.0555023327,-0.0903909132,-0.3846963942,-0.2294014692,0.1550245732,-0.1646524221,0.0437303372,-0.5862874985,0.5789603591,-0.117109172,0.3604901135,-0.0000940233,-0.0841739178,0.1785383672,0.0156737659,0.0960438475,0.3012169302,-0.6603952646,-0.0925348178,-0.5667022467,0.1003710032,-0.0186371841,0.2854993045,-0.1411906928,0.0862474367,-0.1919643134,-0.0177196395,0.5501359105,0.0586882383,-0.179905206,0.2326256931,0.0663630664,0.1149931625,0.0443727784,0.250171721,0.2186097801,-0.1673351079,0.2972685695,0.4012947977,-0.1819023192,-0.0803402662,-0.2413843125,-0.3239442706,-0.1327838004,0.5722842216,0.3206012845,-0.2686688006,0.2041142732,0.1243723929,0.339437753,-0.0848251358,0.2756646574,-0.1774342358,0.3395742178,0.2443036437,0.0192866866,-0.1585206538,-0.3723514676,0.1865617484,-0.0838787407,0.4694602489,0.0942402855,-0.2287684828,0.0601031594,-0.1053943858,0.0143343499,-0.0598728247,0.1379510462,-0.0303854793,-0.0853952691,-0.2968528569,-0.3704718947,0.0920602456,-0.1619082838,-0.0688595474,-0.2895305157,0.303963244,-0.282283932,-0.1348469555,-0.0668507516,0.2432847768,-0.3899496794,-0.1626827419,0.1621795893,0.0692065209,-0.1194823012,-0.1231318489,0.2450623661,0.6768254638,0.3665265143,-0.348058641,0.4001242816,-0.1731018722,-0.1120726466,-0.053560704,-0.0680566132,0.3849924207,0.0789143592,0.1181934997,0.1623069346,-0.0590924546,0.0239694398,0.1912855506,-0.0903871804,-0.2839310765,0.1618857533,0.0931975842,0.1746960431,-0.1144882888,-0.2631929517,0.2389910966,0.2992624044,0.1440161467,0.0741515607,0.3139259815,-0.2024215609,-0.2755900025,0.0226800255,-0.258484602,0.1707413495,0.3357011974,0.3421949744,0.2455014586,0.0908137485,-0.0897762924,0.0264359191,0.1777727157,-0.0633540899,0.0865309611,0.289778471,0.0180611983,-0.1132533476,-0.3876161277,0.0010842518,0.0349820293,-0.090488337,-0.0753129497,0.0485066436,-0.1141976118,-0.0166435763,-0.1363122165,-0.2334698886,-0.2300331742,0.1964631826,-0.107794404,-0.3001351357,0.1516125053,-0.1213786006,0.2916135192,-0.1180161163,0.1591349095,0.090658702,-0.3925730884,-0.030238606,0.1686822176,0.4147643447,-0.174088791,0.1542633325,-0.0453838557,-0.1625928581,-0.1450780332,-0.5141839385,0.289390713,-0.2778079808,0.3317552209,0.1051537171,0.0289108213,0.1600422412,-0.1231144071,-0.0683486313,0.0452122577,0.0994686708,-0.2118147463,-0.0526102446,0.1033179238,-0.0715308785,-0.3749701381,-0.4768829346,-0.4420588911,0.2274209708,0.1049331948,0.1688858122,-0.0329905301,0.1517014056,0.2041810155,0.285754323,-0.0660426021,-0.1892328709,-0.3686844409,0.3979457617,-0.1961515546,-0.1966491938,0.0384902507,-0.2632104754,-0.1312866211,0.1356367916,-0.5101802349,0.0295195114,0.0120013049,0.1627230793,0.0133303553,0.196385622,0.1155848727,-0.0585150309,-0.0557785742,-0.0347956792,-0.0180799197,0.150504902,0.4623231888,0.2625662088,-0.1689417064,0.3510466218,0.2542749643,0.727611959,0.003083626,-0.0559934862,0.4064157009,0.4397130609,0.3380452096,-0.1789106727,0.0039283652,0.3281744421,0.0187001005,-0.1475957334,0.517652154,0.1488894075,-0.4262508452,0.0026980259,0.1242028549,0.0983084291,-0.20254381,0.3485305607,-0.2985153794,0.0200031884,-0.0312003586,-0.5389335752,-0.221586436,-0.1953792274,0.088497512,0.3120163977,0.1095740944,0.10188438,-0.0763872713,0.2693141103,-0.1887980998,0.1580265313,-0.0437003449,-0.1328924149,-0.0466227196,-0.2509896755,0.0107633006,0.1253485084,0.4681356549,-0.3941885233,-0.1235310361,0.1212869808,0.3260887861,-0.127685234,-0.0355436429,-0.5294525027,-0.4004548192,0.5900310874,0.3432855904,-0.3308141232,-0.0361624844,0.2174320966,0.1503800303,0.0279115923,-0.1188354567,-0.0395783223,-0.1805312335,-0.169441089,0.1693010479,0.370588541,-0.0405523144,-0.0470398441,0.1441902667,0.0930871442,0.0253579039,0.2950045764,-0.0540947095,-0.0159559064,0.3271071911,0.0599807575,-0.2474498898,0.3872924149,0.3461895883,0.4749763906,-0.1169205904,-0.0905513093,0.3398268223,-0.2036012858,0.2306207418,0.3679645658,0.2484651059,-0.0886403546,-0.2434787154,0.1212006137,-0.1880738139,0.5451241732,-0.0940475911,-0.0260816831,-0.5294877291,-0.3803672194,0.3644817472,-0.0992156044,0.0354740024,0.1674586684,0.2813960612,-0.4829722941,0.6364092827,-0.0168574154,1.1025232077,-0.0652918965,0.3537803888,-0.1294853091,0.0039829887,0.2202463001,-0.5112812519,-0.0253644735,-0.0458270423,-0.2003690898,0.0251344256,-0.0575812347,-0.159829393,0.5191497803,-0.1203580201,0.2486139983,-0.1480068564,-0.2651326954,0.020204002,0.4700296819,-0.1056299731,-0.385422498,-0.3597135246,0.2270652205,0.0098067252,0.0193854701,-0.233650133,-0.2644320726,-0.1241327301,0.1691194773,-0.4793208838,0.0896330774,-0.1576261222,0.0248133708,-0.2018629164,-0.2102385163,0.4235244095,-0.0873093382,0.35237059,0.1215920597,-0.0807893276,0.1512176394,0.0451418199,-0.0527395979,0.3716473281,-0.0390892252,0.5940967202,-0.008322116,-0.2317297459,0.0769675002,0.14276959,-0.3230777383,-0.1906630546,-0.1975677907,0.255591929,0.1390405595,-0.1689262986,0.3088465929,-0.1813957095,-0.0680170655,0.1618088931,0.1335716099,-0.2017662823,0.0590178333,0.0189702343,-0.1300776452,-0.2557782531,0.0183555093,0.0582696572,-0.1565388441,0.1021979153,0.5984769464,-0.1940762103,-0.1768921614,-0.0995306373,0.280919075,-0.2014945,-0.0990256965,-0.0217411052,-0.149495095,-0.2289838344,-0.1485708207,0.4502548575,0.1329182386,-0.0309694912,-0.1021904722,-0.2635548115,0.2107949853,-0.023498185,0.3028500676,-0.364964515,0.065363273,0.0227702707,0.2464584559,-0.401044935,0.1114065647,0.0063210777,0.1552125812,-0.1150989532,-0.0658409297,0.1508653015,-0.0921790153,0.0876242742,-0.1946215779,-0.2820923328,-0.2190267295,-0.2060976923,0.1083389148,0.2033401877,-0.125287354,0.1312322468,-0.0882523656,-0.2762526274,-0.1511943936,0.3431833684,-0.0181443933,0.1012833938,-0.0046780198,0.5204725862,-0.0797795355,0.1138587296,0.0979010388,0.3665081561,0.2086997628,0.2944311798,0.1023333147,0.2033561021,-0.133848384,-0.1399561167,0.3338398635,0.1092657968,0.3891882002,0.1512448639,0.0868345201,-0.0023411785,0.2002908438,0.3097852468,-0.0276385862,-0.0176400226,-0.0481520221,0.2271760404,0.227232933,-0.3550765812,0.1756113917,-0.0071314294,0.0143562471,0.2544665933,0.227611199,0.2620123923,0.5055462122,-0.2609362304,0.1652662158,-0.0090645077,-0.0192007963,0.3018051386,-0.0261375904,-0.0782491788,0.1375224143,-0.0037861085,0.07977584,0.1533231586,-0.1607873142,-0.2781437039,0.1463104635,0.0951159596,-0.2183500677,0.2714690566,-0.272634536,-0.2061923444,0.4248819649,0.2532886565,0.2978368402,-0.2112036347,0.1144422665,0.0731411576,-0.0343912058,0.0772482976,-0.0096379714,0.0482389368,-0.0574829616,0.2725512385,-0.1869561523,-0.5308090448,0.0211125929,-0.1288434714,-0.378148973,-0.1739601195,0.2517412305,0.0448923744,-0.1875323802,0.0295009464,0.132082954,0.1364844143,-0.0829810649,0.2108623236,0.1295992136,-0.0900362283,0.0825520828,-0.017804211,-0.0014581142,0.0515290499,-0.2987748384,0.3172309399,0.1981982291,-0.0147839142,-0.2029738724,-0.0378414877,0.0955373868,0.1478348076,0.1982405484,0.1261266023,-0.2556395829,-0.1641161293,-0.0558718443,-0.0261695981,-0.114230983,0.6312190294,0.1253979355,-0.1434576064,-0.1313256621,0.0720063001,-0.6766440272,-0.2525229156,0.6505020857,-0.3081345558,0.064097777,-0.1045376435,0.0819040984,-0.2183979154,0.4375668466,0.5349731445,-0.080837287,-0.1756248921,-0.0339159742,-0.4111699462,0.0426710583,-0.065084517,0.1337044239,-0.2017540336,0.4288891256,0.3276130557,0.1650901586,-0.1294502616,0.0138971899,-0.276652813,0.1141718403,-0.3966600597,0.0286500771,-0.1805902272,0.529540062,0.0643056408,-0.0955123603,-0.1499082446,0.1289159209,-0.0185426157,-0.2123022974,-0.0372615345,0.192766577,-0.0630088076,0.2179369628,-0.1568393856,0.3154942393,-0.0734699741,-0.2787453532,-0.5120803118,-0.3022730649,-0.1770476401,0.0238355361,-0.2438126802,0.3752899468,-0.2595465481,-0.2580469847,-0.0995239466,0.0577655397,-0.1248800233,-0.2029437274,-0.1791109443,-0.0107746674,-0.0259504467,0.1387142092,0.174814105,0.2442113012,0.0098783346,-0.0478954241,-0.1346514076,-0.2120028436,0.3039341569,-0.0588455237,-0.3372973502,-0.1390295923,0.106036,0.0217880625,0.051223062,-0.6042736173,0.1963683367,0.2461274564,0.0073622167,-0.124506034,0.28544572,0.05073791,0.0038656292,-0.1812218577,-0.0281392038,-0.1132021695,-0.1597997993,-0.3635296524,-0.3330015242]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1675","title":"Add the 800GB Pile dataset?","comments":"Ah I just saw that @lhoestq is already thinking about the specifying of one or more subsets in [this PR](https:\/\/github.com\/huggingface\/datasets\/pull\/2817#issuecomment-901874049) :)","body":"## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n","comment_length":21,"text":"Add the 800GB Pile dataset?\n## Adding a Dataset\r\n- **Name:** The Pile\r\n- **Description:** The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. See [here](https:\/\/twitter.com\/nabla_theta\/status\/1345130408170541056?s=20) for the Twitter announcement\r\n- **Paper:** https:\/\/pile.eleuther.ai\/paper.pdf\r\n- **Data:** https:\/\/pile.eleuther.ai\/\r\n- **Motivation:** Enables hardcore (GPT-3 scale!) language modelling\r\n\r\n## Remarks\r\nGiven the extreme size of this dataset, I'm not sure how feasible this will be to include in `datasets` \ud83e\udd2f . I'm also unsure how many `datasets` users are pretraining LMs, so the usage of this dataset may not warrant the effort to integrate it.\r\n\nAh I just saw that @lhoestq is already thinking about the specifying of one or more subsets in [this PR](https:\/\/github.com\/huggingface\/datasets\/pull\/2817#issuecomment-901874049) :)","embeddings":[-0.3140250444,0.1560877711,-0.157755807,0.1820491105,-0.027711926,0.2084934264,0.1576407254,0.1906449646,0.0126804197,-0.0463636816,-0.1654876918,0.1902200729,-0.4964599013,0.2447127551,0.2264724523,-0.036290586,-0.0239854418,-0.117916815,-0.0887088031,-0.0244249031,-0.0195851661,-0.0417656042,0.0144501971,-0.4073126018,-0.167057544,-0.0601609461,-0.1501535177,0.0104317749,-0.3633566797,-0.200304538,-0.237642318,0.1232232302,0.3964087367,0.502411902,-0.0001000399,-0.1629320234,0.1292916089,-0.1930175722,-0.362067163,-0.2543430924,-0.1096574739,-0.3248431087,-0.3422295749,-0.0353268646,-0.1831485331,-0.1879443377,0.1316821724,-0.3572055995,0.0566397607,0.2555432618,0.2685053051,0.2546108663,0.0170974322,-0.1162618324,0.3399556577,0.0631387904,-0.169928968,0.257361263,0.3639290929,0.1733178347,-0.0872363523,0.4222867191,0.4526141882,-0.2000396252,0.258446604,0.0176956858,-0.0658392385,-0.1251063049,-0.1947441995,0.3474349678,0.3824141026,-0.1905806363,-0.5288470387,-0.4103041887,-0.0564834289,-0.4034163356,-0.2425067425,0.344522655,0.0575302169,0.332824558,0.0233515445,-0.2500684559,-0.0692026913,0.1912065297,0.0271331575,0.3145010471,-0.1006790847,-0.1568935961,0.120230563,0.0075310203,-0.1577138305,0.0121188499,-0.0387093984,-0.0001409127,-0.3294295967,-0.2154621929,0.3981156051,0.3535758257,0.4854921699,-0.1341152042,0.0018772121,0.0193458535,-0.0771902874,-0.0032304337,0.5196982622,0.024914287,-0.2072120011,-0.2758996189,0.2686921358,0.0638092756,0.1761342883,0.086870119,-0.047157757,-0.0809495747,-0.1676562726,-0.1773598045,-0.0507859848,0.127234295,-0.0632344708,-0.0446244888,0.2358297855,-0.1035479829,-0.0050903591,0.055820968,0.0438573286,0.1919285655,0.0104029188,0.2573389709,0.1459525377,-0.4894524515,-0.2063297629,0.183729887,-0.4527623951,-0.0691873208,0.1031783223,-0.1375151873,0.0783142745,-0.1020732895,0.2345260382,0.2292526215,0.1346583813,0.0306340363,0.1371716857,-0.0269555617,-0.0739035681,-0.2033777982,0.0568886176,0.0850887001,-0.3138953447,0.1317081004,0.0163719989,-0.2196156681,-0.3382868171,0.2284154743,-0.2176370472,0.0264339987,-0.4154157341,0.7231199145,-0.0329928249,0.2563389242,0.0751016885,-0.0533589609,0.0356976353,0.0082695978,0.0736957863,0.1863314807,-0.5746176243,-0.1319470406,-0.5834824443,0.0483263321,-0.1022877768,0.2690888643,-0.1963113695,0.1377868205,-0.1480196118,0.0586056672,0.5879366994,0.1688159108,-0.1505534798,0.1420860738,-0.1099500656,0.009615723,-0.0103814499,0.1500778496,0.1274719089,-0.0512013063,0.2615861595,0.3900918663,-0.1389925778,-0.1138648242,-0.2623089552,-0.377990514,0.0675957352,0.3910391033,0.1070266664,-0.2736819386,0.2186495215,0.0501531251,0.1668393761,-0.0123430695,0.1976042986,-0.1248912737,0.2841337025,0.1168326139,-0.0177331921,-0.3275248408,-0.4190087616,0.0483474024,0.0419962704,0.3821239471,0.0728235766,-0.2030946314,0.0589445308,0.005783753,0.1332860589,-0.1204906926,0.2177647799,0.0485019125,-0.0301405974,-0.2270911634,-0.3296844959,-0.0630915835,-0.0892224908,-0.0877662897,-0.2958340943,0.0060742958,-0.1867309809,-0.1411352009,0.0270735174,0.2707060277,-0.4790112972,-0.1836998314,0.2493359745,0.0859609172,-0.130210802,-0.1862048656,0.2459146976,0.6053516865,0.2642050683,-0.1209844127,0.2040129453,-0.2006687373,-0.0422985032,0.0218253527,-0.1840438098,0.4000040591,0.2446723282,-0.0618102476,0.1006869897,-0.1020419598,0.1329485625,0.0554240122,-0.0784556791,-0.2822150588,0.2581147552,0.1021332741,0.2248966396,0.0236439798,-0.3749103248,0.1169881374,0.2765392661,0.1346707046,0.2371223718,0.2638746798,-0.1729671657,-0.2526938915,-0.0043546734,-0.2683978379,0.2763849199,0.4303976893,0.271143496,0.0645138994,0.184170261,-0.1474291235,0.0317661613,0.0813246891,-0.0447636172,0.1689247936,0.402156055,0.0353724696,-0.3077982068,-0.2223669291,-0.0800784528,0.0917734131,0.0741328597,-0.0449300483,0.0448769554,-0.0952308401,0.0547971576,0.0300011206,-0.2347225398,-0.1994048208,0.1858643442,-0.1503640413,-0.3074035347,0.0467671491,-0.1291742474,0.5056379437,-0.0863483921,0.1858696043,0.0908082128,-0.2864463329,-0.0711079612,0.2793172598,0.2715525329,-0.2613441646,0.4053997695,0.0030316042,0.0979189053,-0.2049828768,-0.5722048879,0.2662444711,-0.1563777179,0.3840333819,0.2316687405,0.0047286176,0.1451892257,-0.0371908583,-0.1277272105,-0.1264147758,-0.0322177187,-0.2840772569,-0.0060539264,0.0704042837,-0.2615520358,-0.3562017679,-0.3958789706,-0.5152435303,0.3340537548,0.0680235848,0.0318070501,-0.000996761,0.0940476283,0.3164206147,-0.0580389835,-0.012133399,-0.1471477747,-0.2930302024,0.3443108499,-0.1152916849,-0.2608080208,-0.0380178615,-0.2216433883,-0.0507546738,0.104109779,-0.4849395454,0.0580915697,-0.0900533721,0.2685629725,0.0396258198,0.1388692707,0.1441310793,-0.1144477725,-0.221587196,-0.1144572422,-0.0163236819,0.1449478269,0.3994907737,0.3285677731,-0.1540107131,0.2489542365,0.2208058834,0.4202838242,-0.0282783546,-0.2335667461,0.4258720875,0.3500135839,0.4011774063,-0.1807069331,-0.0191543009,0.2930106819,0.0757228956,-0.0897006541,0.6187443137,0.2041490078,-0.2972596884,0.0342047587,0.0647785664,0.0405765921,-0.1080315188,0.3248300552,-0.1217569858,0.0495291688,-0.0646893829,-0.6107584834,-0.2354198992,-0.1349544525,-0.0441820696,0.3229153454,0.0953680947,0.0575048178,-0.2516891062,0.3593747318,-0.3118148744,0.1145659611,-0.088508375,0.0156534091,-0.1086449549,-0.287297219,-0.1381918341,0.0626946166,0.4251104295,-0.3361886144,-0.2940456271,0.0159608312,0.3112529218,-0.1158224046,0.0431156196,-0.3933472633,-0.1939311177,0.5661793947,0.4673762619,-0.3769964576,0.0618888922,0.1933876574,0.2966935635,0.149527669,-0.2640249431,-0.1340187043,-0.2297497988,-0.18490237,0.1448799223,0.3940480351,0.0007442465,0.1037832126,0.139274016,0.1136061326,0.0214907583,0.2478018999,-0.0123321488,-0.001199484,0.2568435967,-0.1122081652,-0.152970776,0.3542766869,0.2100278139,0.398237139,-0.1596426666,-0.2744925618,0.1917804331,-0.3813211024,0.303258121,0.2107843012,0.1493547559,0.1431998909,-0.216464445,0.1274141669,-0.1360334903,0.4089274108,-0.0726562217,0.0683230832,-0.4745434821,-0.3134112954,0.3272920251,-0.0328480341,0.0708321929,0.2631412446,-0.0244733747,-0.4681372046,0.356305629,-0.0487756804,0.972084105,-0.0315379351,0.3297254443,-0.1556223482,0.1254192889,0.406547755,-0.3534582257,0.0052825054,0.0605981685,-0.1813676804,0.0550211966,-0.0546496995,0.0490667522,0.5103421807,-0.1810548455,0.0517501384,-0.1288799644,-0.1312520653,0.0034578745,0.48246032,-0.0369956382,-0.2597953677,-0.3652406633,0.2879696488,0.0420823954,-0.1093499959,-0.2554651797,-0.2379550636,-0.1509443223,0.083564885,-0.4433173537,-0.0807977095,-0.1639738679,0.0077205137,0.0364821106,-0.2103143781,0.4322669208,-0.0128375851,0.2519204915,0.1670984328,-0.1127913222,0.0176694654,-0.111227423,-0.018626865,0.2736057043,-0.0821469575,0.3878295124,-0.0404396206,-0.0624104515,0.09686248,0.1057486832,-0.3731237352,-0.1521080732,-0.0252921171,0.074453488,0.1005660072,-0.2579145432,0.289891839,-0.1230167076,-0.1712987572,0.2289017141,0.2383226752,-0.0725237355,0.0059795072,0.0627453849,-0.1664611101,-0.2630989254,0.0295568276,-0.0158199277,-0.0853959024,0.0257694665,0.5174471736,-0.2413362861,-0.3063099086,-0.1214732453,0.2176185399,-0.1551899165,0.0993228331,0.1783810705,-0.1808441132,-0.1649969518,0.0031081256,0.3988018334,-0.0037677865,0.0122639062,-0.0902003497,-0.1694751829,0.1409019679,-0.0748112872,0.2321824878,-0.2394777983,0.0583787337,0.0653490871,0.1996252835,-0.4790588915,-0.0172106773,-0.1031065509,0.1940823644,0.0257966854,-0.1157992557,0.0800538734,-0.0498215407,0.1445262581,-0.1368616968,-0.2744659483,-0.3361320198,-0.1224716976,0.0602241233,0.2010567784,-0.0505763367,0.2600130141,-0.0751723275,-0.2805403769,-0.117541492,0.2650120258,-0.1448343545,0.0390710272,-0.006452715,0.31725806,-0.0623279437,0.2591876686,0.024976844,0.2109283358,0.1000114679,0.2151434869,0.1083607152,0.3262650669,-0.0670704022,-0.0964009762,0.315030992,0.092123799,0.3351786435,-0.0102924863,0.0919969454,-0.0295383818,0.3819145858,0.279818058,-0.0355544761,-0.1299985051,0.0084709004,0.158623755,0.3344058394,-0.4310395122,0.0342153721,0.0924785584,0.0528948233,0.2981903255,0.1584796011,0.2950378358,0.4187948704,-0.2616935968,0.036277514,0.1286183596,-0.0724784732,0.2660198808,0.001015306,0.0877121985,0.1428154409,0.2457883507,0.1364241838,0.2143191695,-0.0979570299,-0.1044535413,0.2325045764,0.2193846554,-0.0818717703,0.1846458316,-0.3057016432,0.0541619062,0.4232201278,0.3285285234,0.2546018064,-0.0541002229,0.1648483276,0.1153752729,0.0459735729,0.038816262,0.0204470828,-0.0355903506,-0.1078863442,0.2713014185,-0.2485500872,-0.2481558174,-0.0301026218,-0.1608668715,-0.2640420198,-0.1002404839,0.0802676454,0.075196296,-0.4136738777,0.126817897,0.0742893517,0.0162865017,-0.2125294507,0.2674420476,0.0987028182,-0.0569280684,0.143157959,0.1705259979,0.0042503756,0.132899031,-0.0127465101,0.1644603312,-0.1137919053,0.0115691246,-0.4354976118,0.144881621,0.1139830053,0.1660974622,0.2558999062,0.2096755207,-0.3054894209,0.1138866171,0.0108472276,-0.1492231637,-0.2690040171,0.3347718418,0.0259919893,-0.1921249628,-0.2373899966,0.1815179288,-0.5247017741,-0.1559541225,0.7176670432,-0.3689734638,0.0299224537,-0.229393363,0.146516338,-0.1644469202,0.4301392436,0.5728159547,0.1341493875,-0.0498819649,-0.097431615,-0.4856863022,0.0876668245,-0.1041567177,0.247343421,-0.146143049,0.3865130246,0.3208931386,0.2434921712,-0.1550246775,0.232075721,-0.365811646,-0.0333920941,-0.410318166,0.0164919551,-0.0933658704,0.4397864044,0.1092086583,-0.0113494117,-0.0103890495,0.0238902364,0.1371145397,-0.4912763536,0.0674182773,0.0880611539,0.1417625397,0.2719899714,-0.0375982337,0.2052155733,-0.1154116541,-0.2631016672,-0.4812112451,-0.12634857,-0.224243775,0.1023272797,-0.0559035167,0.2449292541,-0.312764436,-0.1222844124,0.0259012245,0.0865536109,-0.1114763692,-0.109012112,-0.154275015,-0.0126512693,-0.0579320379,-0.03762215,0.015035118,0.2603275478,0.0042966329,-0.1916681826,-0.0162360109,-0.1827394515,0.1051878929,-0.2134497166,-0.3895477057,-0.1133389249,0.0187837426,0.0992126316,-0.0601748563,-0.6072430611,0.1306893528,0.3180980384,0.1082284525,-0.0802904889,0.3524813652,0.090518713,-0.0578104518,-0.1608340591,0.1676839739,-0.0985128209,-0.1351516098,-0.2600592375,-0.2852146029]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1674","title":"dutch_social can't be loaded","comments":"exactly the same issue in some other datasets.\r\nDid you find any solution??\r\n","body":"Hi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```","comment_length":13,"text":"dutch_social can't be loaded\nHi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```\nexactly the same issue in some other datasets.\r\nDid you find any solution??\r\n","embeddings":[-0.1440130919,-0.1277865618,-0.1560761034,0.2933968306,0.2167993635,-0.1348885447,-0.0120579246,0.0937409475,0.3958035707,-0.0862966329,-0.2941347361,-0.0078357765,0.0196760818,-0.0421471745,0.1373223066,0.1136013195,0.2246757895,0.0936486647,0.144613862,-0.1379558891,-0.2554816604,0.4084303081,0.0364790969,0.1195254549,0.0535758808,0.2432906479,-0.0837842226,0.3191888928,-0.0633874536,-0.2948622704,0.3469977081,0.1354592592,0.1217035055,0.5947429538,-0.0001220059,0.0153691275,0.4793719053,-0.0934169069,-0.6241431832,-0.4102129638,0.0205617975,-0.2827773392,0.0646422952,-0.0720438063,-0.0622491986,-0.2690680027,-0.0105327405,-0.474494487,0.2986169159,0.4165929258,0.1114755422,-0.0108238775,0.0761229396,-0.0746558085,0.0295876656,0.3834807873,0.0695490241,0.5175902247,0.2559808195,0.0191765875,0.4757993519,0.139318645,-0.1636501998,-0.1841302067,0.1682499945,-0.1772353053,-0.0807946846,-0.2221838981,0.1476258039,-0.074964039,0.4582820535,-0.0886856616,-0.264942199,-0.2424675226,0.0461236127,0.0547623821,0.276529938,0.1222155318,-0.3778547943,0.2942798138,-0.1758255213,-0.2210489362,-0.1348641068,0.542472899,0.0668623,0.0545753092,-0.1998028606,0.0972559825,0.2403409779,-0.1409109533,-0.0025449127,-0.1723716855,0.1879131049,0.2741535306,0.0476628169,0.275170058,0.109135069,0.1229047924,0.0124698281,0.2088605762,-0.0083227707,0.0611036904,-0.2934058607,0.2661069036,0.1451622844,-0.063431032,0.1614208817,-0.0578602739,0.1387722641,0.2737375498,-0.2378363311,-0.0220349152,-0.5131173134,-0.2261775881,-0.4047652185,-0.1627119333,0.0526988059,-0.1887488216,-0.2392709255,-0.0537533425,0.2383565456,0.1124858856,0.1500710696,0.4265648723,-0.0568392836,0.1309242398,0.0157592501,0.3348211646,-0.0972028598,0.1625242531,0.0043877871,0.0115900803,-0.2395053059,-0.1033063531,0.3564893603,-0.1404215097,0.3913652897,-0.0133134797,0.0193966776,-0.2262601852,-0.0651753247,-0.1723169386,-0.0208002795,0.4110073447,-0.0504028127,0.1848825663,0.2301233411,-0.3826444447,-0.2856127322,0.0220117606,-0.4266529381,-0.2579602599,-0.1370801181,0.0865042135,-0.3576443791,-0.0483067632,-0.2518779337,0.0720809549,-0.2459039837,-0.3608632684,-0.1423634887,-0.0626245961,0.0569238551,-0.1635662615,0.4047258198,0.6451685429,-0.259563148,-0.1338703334,-0.2519385815,-0.0979309827,0.1166802645,0.1549301893,-0.2774693072,0.1003124788,-0.4217948914,-0.1295697242,0.5071360469,-0.3881822526,-0.1034957767,0.043269366,0.0442071557,0.4348086715,0.0677174553,-0.0021925054,-0.2236290574,0.0936073661,0.1177948564,0.365460515,0.2958675623,-0.3151164651,-0.0436034799,-0.08808887,0.2152402699,0.4358336627,-0.0789023265,-0.065279454,0.3209886551,-0.0179740302,0.0676686615,-0.0443313085,-0.1122181565,0.3995007277,-0.1002356708,0.6040776372,-0.1201559752,0.0579323918,-0.1615790576,0.1620278955,0.3629038632,-0.0949904695,-0.2131394148,0.0145916874,-0.5914059281,0.1086386964,-0.2440940589,-0.0796029046,0.0442409627,0.3385514021,0.2097835839,0.4188049734,-0.1303366423,0.5874919891,-0.1820048988,-0.0004448143,-0.136841476,0.3376156092,-0.077340886,0.0069129346,-0.025883032,0.4640442133,0.19163315,-0.0753638968,0.1306604892,0.1775824875,0.2757100463,0.1431743801,0.3152182102,-0.1459710449,0.013373767,-0.1851287335,0.0264447127,0.2751175463,0.1668300778,-0.0880569145,-0.2502496839,0.1557591856,-0.0062821158,0.2842067778,0.1452289671,0.0032311126,0.2915836871,-0.0235694479,-0.0459581092,-0.1600633413,0.3323235512,-0.1702431589,0.5120025873,-0.006221151,-0.2748087049,0.1538019031,0.0134518212,-0.1130990013,0.0663182139,0.2667791247,-0.1694904119,0.1820977181,0.1649412066,0.2306359261,0.5306184292,0.1501768082,0.0584150627,-0.1042482406,0.1984350532,-0.226388678,0.196112588,0.0779637173,0.3595768809,0.1046478078,0.1316241622,0.1227424741,-0.2572664917,-0.5361442566,-0.149571985,0.0522079468,-0.2917066514,0.3217304945,-0.185494706,-0.4013975859,-0.3152742684,-0.1355844587,-0.2492753416,-0.3631968498,-0.1829154044,0.1399168819,0.0379369855,-0.0491964184,-0.1167751774,0.1715294868,-0.2130713314,-0.3847946823,-0.3287187815,-0.0870077685,-0.049715437,-0.0937781259,0.4112957716,0.100119248,0.1564811319,-0.4932972789,0.1215285957,0.094456993,-0.4239718616,-0.1459390521,-0.0658269599,0.1950836778,-0.0813921466,0.3485629559,-0.2351174653,-0.3012833893,0.3619758785,-0.019668784,-0.0797758698,0.1490667313,0.2029022425,-0.042098213,-0.1645146012,-0.0364278592,-0.8044582009,-0.3288797438,0.1570771933,0.2175286114,0.0154114179,0.4426355064,0.3873107731,0.0452336743,-0.0348101072,0.1303268522,-0.1764192879,-0.3531501889,0.6498783231,-0.3994504213,-0.2847266793,0.1182230935,0.2317685336,0.2142133862,-0.3526259363,-0.4864182472,0.0804351568,0.0773071796,0.2877722383,-0.0448323004,0.1416716278,-0.0867542922,-0.0689283088,-0.0460393466,-0.1349615753,-0.0135841006,-0.116391331,-0.3259768188,0.2562913597,0.0006416994,0.1128346696,-0.2048086971,0.5850800276,0.1846069694,0.1297905594,0.3108214438,-0.0963969231,0.686024785,-0.1007470489,-0.4953562319,-0.0185463745,-0.0426108614,0.0646766201,0.1125227883,0.302135855,-0.1705505848,-0.4108614326,0.052665893,-0.6017850637,-0.0814373419,0.1122563258,-0.1788872331,0.0257286951,0.0925906599,-0.1898592263,-0.3393605649,-0.1695629656,0.0055447565,0.5575093627,-0.0501216576,0.3014118075,-0.5939533114,-0.2490032464,-0.1569108814,0.1927903295,-0.1274876893,0.2557533979,-0.3264309466,0.0863464475,0.1326007843,-0.0135395527,0.4668221474,-0.0073006665,-0.0018592202,0.2152351737,0.21335949,-0.6372420788,0.1366122514,0.1852865368,0.1388081014,0.4634910524,0.3116728663,-0.1862471849,-0.0533104017,-0.0528364554,0.3992613852,-0.099411957,-0.1412382126,-0.4389512837,-0.5290256739,-0.3330169022,-0.3466005623,0.041629184,0.3000886738,-0.1127192676,0.1479955167,-0.0392461792,0.08828035,0.1973963529,0.215036571,0.3390181661,0.2029387206,-0.1517691016,0.2156927139,0.207391575,0.186458379,0.8277205825,-0.0280642305,-0.472987473,-0.0676505715,-0.1684446484,0.2539019883,0.2572754323,-0.0525502823,-0.264387399,0.0862524435,-0.0616413429,-0.1060521901,0.2844638824,-0.0042873053,0.1814712882,-0.1484197974,-0.3832840323,0.4944527149,-0.0522756279,-0.0187165849,0.1487674266,0.1603042036,-0.1453430653,0.2155505419,-0.2960323989,1.1821036339,0.0836563557,0.1619860977,0.1922802627,-0.1424458325,0.3413336873,-0.2491705269,-0.0668278858,-0.0165302716,0.1841173321,-0.0074367477,-0.150554657,0.2806473672,0.1858281046,0.0204978194,0.0036976642,-0.0962960869,0.0299856495,0.0989257619,0.4379950464,0.1277776212,-0.0180496313,-0.5319270492,0.1279685497,-0.0070529534,0.6483965516,-0.0589642003,-0.1501560062,-0.6219805479,0.0417346805,-0.1697573066,0.0559509322,-0.1763844192,0.3046876788,-0.0242759045,-0.4557029903,-0.131096974,0.3823457658,0.4158375561,0.0628483519,-0.2100280225,0.1942519248,-0.0006204766,-0.2378745824,0.1408261955,-0.2639215887,0.0989748389,-0.1657351702,-0.3407813311,0.2149266452,-0.0537951812,-0.3472219706,0.1409178823,0.0401334427,0.2564260364,0.0465484746,-0.4316743016,0.0652011633,0.2491727322,-0.1211000159,0.0465644635,0.0329136997,0.0470805205,-0.20226264,-0.223509267,-0.0786677599,-0.0912784413,0.4528014958,-0.2509536445,-0.1783564538,0.3761715591,0.1313108206,-0.2986024618,-0.094971858,0.1794088036,-0.0717810094,-0.4908860326,-0.3483213782,0.0258632731,-0.0363163538,-0.3221884966,0.1668507606,0.2664426863,-0.2686613202,0.1378796399,-0.7992363572,-0.009050753,0.2148845643,0.0030076047,0.0240182113,0.149524793,-0.1888760328,0.1773919314,-0.1492820978,-0.2086732835,0.1037804931,-0.2042625546,-0.0708857626,0.3645169139,0.0797675326,0.1445762366,-0.2587434351,0.0331773981,-0.3906826675,0.0711602569,-0.1381139606,-0.0232484508,0.167788595,-0.030742228,-0.0736366659,-0.1844884306,-0.423468709,0.0495093353,-0.0976031423,-0.1504056454,-0.0286711175,0.0516630039,-0.1411792636,-0.0938978791,0.3486450016,-0.0014615451,0.2108521163,0.0660802647,0.5716895461,0.0365405045,0.0253206547,0.1066599116,0.0104909716,-0.052856829,0.1187435761,-0.1242643297,0.1727949828,0.3727681637,-0.3196749091,0.3249733746,0.120964922,-0.0035803635,0.5077812672,-0.441368252,0.1462892592,0.3029463887,0.0475829877,-0.2808189392,-0.1859784126,0.4009407163,0.0555047877,-0.0616027452,0.2878829837,0.2537675798,-0.1890611053,0.389048636,0.1293126196,0.4627902508,0.4501369894,0.2029798329,0.0939915404,0.0795385763,0.0393343084,-0.0704107285,0.320171982,-0.0320692174,0.4540109932,-0.261349231,0.1091690212,0.0425857119,0.1719930172,-0.1905469745,-0.2267156392,0.1692724973,-0.1652827412,0.0399112813,0.0455428585,-0.1602568477,0.4140933454,-0.4061875045,-0.1118158549,0.0197634734,0.2420539558,0.0456830412,-0.0484396964,-0.2243011445,-0.2722094655,-0.1357747316,-0.184592098,0.1357543319,-0.5232682228,0.2045866102,0.1481912583,-0.1573607177,-0.2957580388,-0.0227050185,0.268961817,0.0716758817,-0.1148370355,0.326603651,0.2135397494,0.0652032346,0.1118482724,0.2881093621,0.4642577171,0.2599305511,-0.1109033898,-0.0053320415,-0.0054772645,-0.1516873986,-0.2587798536,0.2607496977,0.0704799891,0.0574535392,0.3448998928,0.166607663,-0.0934453532,-0.1094970554,0.2727870345,0.0146258213,-0.1755820364,0.1559967399,-0.4784351885,-0.0105871502,-0.3096095324,0.042605143,-0.3361676335,-0.0115367193,0.2119975388,0.0244800057,0.1457011402,-0.2048273087,0.0240990482,0.0394072495,0.4086048901,0.3885657787,0.2942712009,-0.2616147101,-0.5872488022,-0.8084726334,0.1373805106,-0.2018896788,-0.4872536361,0.0315862224,-0.1072384417,0.0769952685,0.2530308664,0.0159672648,0.3099045753,-0.0801284462,-0.1084850878,-0.2610019445,-0.2317061871,-0.2546134293,0.0900814682,-0.1903463602,-0.1421096772,0.1901785582,-0.5949850082,0.0008110312,0.163948983,-0.1057656333,-0.2082068324,-0.0478758253,0.5655726194,-0.1149526983,0.2684367895,0.0479270406,-0.2305282652,-0.1125653535,-0.252161473,0.1632338315,0.6077491045,0.0070399325,0.2551538348,-0.4812214077,0.0169900656,-0.0885087028,-0.026375182,0.2812675536,-0.1549213678,0.1070497036,-0.0785577297,-0.1062413678,0.2599004805,0.2936939001,0.3048795164,-0.2413358092,0.176688835,-0.1652900726,-0.1124801785,0.3234544694,-0.3052375019,-0.261243552,0.1704805791,0.1355917305,0.0898253173,0.1752025783,-0.2406037897,0.1892071515,0.3313728273,0.1525108516,-0.1949869394,0.2746852636,-0.3622052372,-0.1652604491,-0.0452168286,0.133011803,0.0323434584,0.0108250966,0.0206802953,-0.0255035013]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1674","title":"dutch_social can't be loaded","comments":"Hi @koenvandenberge and @alighofrani95!\r\nThe datasets you're experiencing issues with were most likely added recently to the `datasets` library, meaning they have not been released yet. They will be released with the v2 of the library.\r\nMeanwhile, you can still load the datasets using one of the techniques described in this issue: #1641 \r\nLet me know if this helps!","body":"Hi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```","comment_length":59,"text":"dutch_social can't be loaded\nHi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```\nHi @koenvandenberge and @alighofrani95!\r\nThe datasets you're experiencing issues with were most likely added recently to the `datasets` library, meaning they have not been released yet. They will be released with the v2 of the library.\r\nMeanwhile, you can still load the datasets using one of the techniques described in this issue: #1641 \r\nLet me know if this helps!","embeddings":[-0.1440130919,-0.1277865618,-0.1560761034,0.2933968306,0.2167993635,-0.1348885447,-0.0120579246,0.0937409475,0.3958035707,-0.0862966329,-0.2941347361,-0.0078357765,0.0196760818,-0.0421471745,0.1373223066,0.1136013195,0.2246757895,0.0936486647,0.144613862,-0.1379558891,-0.2554816604,0.4084303081,0.0364790969,0.1195254549,0.0535758808,0.2432906479,-0.0837842226,0.3191888928,-0.0633874536,-0.2948622704,0.3469977081,0.1354592592,0.1217035055,0.5947429538,-0.0001220059,0.0153691275,0.4793719053,-0.0934169069,-0.6241431832,-0.4102129638,0.0205617975,-0.2827773392,0.0646422952,-0.0720438063,-0.0622491986,-0.2690680027,-0.0105327405,-0.474494487,0.2986169159,0.4165929258,0.1114755422,-0.0108238775,0.0761229396,-0.0746558085,0.0295876656,0.3834807873,0.0695490241,0.5175902247,0.2559808195,0.0191765875,0.4757993519,0.139318645,-0.1636501998,-0.1841302067,0.1682499945,-0.1772353053,-0.0807946846,-0.2221838981,0.1476258039,-0.074964039,0.4582820535,-0.0886856616,-0.264942199,-0.2424675226,0.0461236127,0.0547623821,0.276529938,0.1222155318,-0.3778547943,0.2942798138,-0.1758255213,-0.2210489362,-0.1348641068,0.542472899,0.0668623,0.0545753092,-0.1998028606,0.0972559825,0.2403409779,-0.1409109533,-0.0025449127,-0.1723716855,0.1879131049,0.2741535306,0.0476628169,0.275170058,0.109135069,0.1229047924,0.0124698281,0.2088605762,-0.0083227707,0.0611036904,-0.2934058607,0.2661069036,0.1451622844,-0.063431032,0.1614208817,-0.0578602739,0.1387722641,0.2737375498,-0.2378363311,-0.0220349152,-0.5131173134,-0.2261775881,-0.4047652185,-0.1627119333,0.0526988059,-0.1887488216,-0.2392709255,-0.0537533425,0.2383565456,0.1124858856,0.1500710696,0.4265648723,-0.0568392836,0.1309242398,0.0157592501,0.3348211646,-0.0972028598,0.1625242531,0.0043877871,0.0115900803,-0.2395053059,-0.1033063531,0.3564893603,-0.1404215097,0.3913652897,-0.0133134797,0.0193966776,-0.2262601852,-0.0651753247,-0.1723169386,-0.0208002795,0.4110073447,-0.0504028127,0.1848825663,0.2301233411,-0.3826444447,-0.2856127322,0.0220117606,-0.4266529381,-0.2579602599,-0.1370801181,0.0865042135,-0.3576443791,-0.0483067632,-0.2518779337,0.0720809549,-0.2459039837,-0.3608632684,-0.1423634887,-0.0626245961,0.0569238551,-0.1635662615,0.4047258198,0.6451685429,-0.259563148,-0.1338703334,-0.2519385815,-0.0979309827,0.1166802645,0.1549301893,-0.2774693072,0.1003124788,-0.4217948914,-0.1295697242,0.5071360469,-0.3881822526,-0.1034957767,0.043269366,0.0442071557,0.4348086715,0.0677174553,-0.0021925054,-0.2236290574,0.0936073661,0.1177948564,0.365460515,0.2958675623,-0.3151164651,-0.0436034799,-0.08808887,0.2152402699,0.4358336627,-0.0789023265,-0.065279454,0.3209886551,-0.0179740302,0.0676686615,-0.0443313085,-0.1122181565,0.3995007277,-0.1002356708,0.6040776372,-0.1201559752,0.0579323918,-0.1615790576,0.1620278955,0.3629038632,-0.0949904695,-0.2131394148,0.0145916874,-0.5914059281,0.1086386964,-0.2440940589,-0.0796029046,0.0442409627,0.3385514021,0.2097835839,0.4188049734,-0.1303366423,0.5874919891,-0.1820048988,-0.0004448143,-0.136841476,0.3376156092,-0.077340886,0.0069129346,-0.025883032,0.4640442133,0.19163315,-0.0753638968,0.1306604892,0.1775824875,0.2757100463,0.1431743801,0.3152182102,-0.1459710449,0.013373767,-0.1851287335,0.0264447127,0.2751175463,0.1668300778,-0.0880569145,-0.2502496839,0.1557591856,-0.0062821158,0.2842067778,0.1452289671,0.0032311126,0.2915836871,-0.0235694479,-0.0459581092,-0.1600633413,0.3323235512,-0.1702431589,0.5120025873,-0.006221151,-0.2748087049,0.1538019031,0.0134518212,-0.1130990013,0.0663182139,0.2667791247,-0.1694904119,0.1820977181,0.1649412066,0.2306359261,0.5306184292,0.1501768082,0.0584150627,-0.1042482406,0.1984350532,-0.226388678,0.196112588,0.0779637173,0.3595768809,0.1046478078,0.1316241622,0.1227424741,-0.2572664917,-0.5361442566,-0.149571985,0.0522079468,-0.2917066514,0.3217304945,-0.185494706,-0.4013975859,-0.3152742684,-0.1355844587,-0.2492753416,-0.3631968498,-0.1829154044,0.1399168819,0.0379369855,-0.0491964184,-0.1167751774,0.1715294868,-0.2130713314,-0.3847946823,-0.3287187815,-0.0870077685,-0.049715437,-0.0937781259,0.4112957716,0.100119248,0.1564811319,-0.4932972789,0.1215285957,0.094456993,-0.4239718616,-0.1459390521,-0.0658269599,0.1950836778,-0.0813921466,0.3485629559,-0.2351174653,-0.3012833893,0.3619758785,-0.019668784,-0.0797758698,0.1490667313,0.2029022425,-0.042098213,-0.1645146012,-0.0364278592,-0.8044582009,-0.3288797438,0.1570771933,0.2175286114,0.0154114179,0.4426355064,0.3873107731,0.0452336743,-0.0348101072,0.1303268522,-0.1764192879,-0.3531501889,0.6498783231,-0.3994504213,-0.2847266793,0.1182230935,0.2317685336,0.2142133862,-0.3526259363,-0.4864182472,0.0804351568,0.0773071796,0.2877722383,-0.0448323004,0.1416716278,-0.0867542922,-0.0689283088,-0.0460393466,-0.1349615753,-0.0135841006,-0.116391331,-0.3259768188,0.2562913597,0.0006416994,0.1128346696,-0.2048086971,0.5850800276,0.1846069694,0.1297905594,0.3108214438,-0.0963969231,0.686024785,-0.1007470489,-0.4953562319,-0.0185463745,-0.0426108614,0.0646766201,0.1125227883,0.302135855,-0.1705505848,-0.4108614326,0.052665893,-0.6017850637,-0.0814373419,0.1122563258,-0.1788872331,0.0257286951,0.0925906599,-0.1898592263,-0.3393605649,-0.1695629656,0.0055447565,0.5575093627,-0.0501216576,0.3014118075,-0.5939533114,-0.2490032464,-0.1569108814,0.1927903295,-0.1274876893,0.2557533979,-0.3264309466,0.0863464475,0.1326007843,-0.0135395527,0.4668221474,-0.0073006665,-0.0018592202,0.2152351737,0.21335949,-0.6372420788,0.1366122514,0.1852865368,0.1388081014,0.4634910524,0.3116728663,-0.1862471849,-0.0533104017,-0.0528364554,0.3992613852,-0.099411957,-0.1412382126,-0.4389512837,-0.5290256739,-0.3330169022,-0.3466005623,0.041629184,0.3000886738,-0.1127192676,0.1479955167,-0.0392461792,0.08828035,0.1973963529,0.215036571,0.3390181661,0.2029387206,-0.1517691016,0.2156927139,0.207391575,0.186458379,0.8277205825,-0.0280642305,-0.472987473,-0.0676505715,-0.1684446484,0.2539019883,0.2572754323,-0.0525502823,-0.264387399,0.0862524435,-0.0616413429,-0.1060521901,0.2844638824,-0.0042873053,0.1814712882,-0.1484197974,-0.3832840323,0.4944527149,-0.0522756279,-0.0187165849,0.1487674266,0.1603042036,-0.1453430653,0.2155505419,-0.2960323989,1.1821036339,0.0836563557,0.1619860977,0.1922802627,-0.1424458325,0.3413336873,-0.2491705269,-0.0668278858,-0.0165302716,0.1841173321,-0.0074367477,-0.150554657,0.2806473672,0.1858281046,0.0204978194,0.0036976642,-0.0962960869,0.0299856495,0.0989257619,0.4379950464,0.1277776212,-0.0180496313,-0.5319270492,0.1279685497,-0.0070529534,0.6483965516,-0.0589642003,-0.1501560062,-0.6219805479,0.0417346805,-0.1697573066,0.0559509322,-0.1763844192,0.3046876788,-0.0242759045,-0.4557029903,-0.131096974,0.3823457658,0.4158375561,0.0628483519,-0.2100280225,0.1942519248,-0.0006204766,-0.2378745824,0.1408261955,-0.2639215887,0.0989748389,-0.1657351702,-0.3407813311,0.2149266452,-0.0537951812,-0.3472219706,0.1409178823,0.0401334427,0.2564260364,0.0465484746,-0.4316743016,0.0652011633,0.2491727322,-0.1211000159,0.0465644635,0.0329136997,0.0470805205,-0.20226264,-0.223509267,-0.0786677599,-0.0912784413,0.4528014958,-0.2509536445,-0.1783564538,0.3761715591,0.1313108206,-0.2986024618,-0.094971858,0.1794088036,-0.0717810094,-0.4908860326,-0.3483213782,0.0258632731,-0.0363163538,-0.3221884966,0.1668507606,0.2664426863,-0.2686613202,0.1378796399,-0.7992363572,-0.009050753,0.2148845643,0.0030076047,0.0240182113,0.149524793,-0.1888760328,0.1773919314,-0.1492820978,-0.2086732835,0.1037804931,-0.2042625546,-0.0708857626,0.3645169139,0.0797675326,0.1445762366,-0.2587434351,0.0331773981,-0.3906826675,0.0711602569,-0.1381139606,-0.0232484508,0.167788595,-0.030742228,-0.0736366659,-0.1844884306,-0.423468709,0.0495093353,-0.0976031423,-0.1504056454,-0.0286711175,0.0516630039,-0.1411792636,-0.0938978791,0.3486450016,-0.0014615451,0.2108521163,0.0660802647,0.5716895461,0.0365405045,0.0253206547,0.1066599116,0.0104909716,-0.052856829,0.1187435761,-0.1242643297,0.1727949828,0.3727681637,-0.3196749091,0.3249733746,0.120964922,-0.0035803635,0.5077812672,-0.441368252,0.1462892592,0.3029463887,0.0475829877,-0.2808189392,-0.1859784126,0.4009407163,0.0555047877,-0.0616027452,0.2878829837,0.2537675798,-0.1890611053,0.389048636,0.1293126196,0.4627902508,0.4501369894,0.2029798329,0.0939915404,0.0795385763,0.0393343084,-0.0704107285,0.320171982,-0.0320692174,0.4540109932,-0.261349231,0.1091690212,0.0425857119,0.1719930172,-0.1905469745,-0.2267156392,0.1692724973,-0.1652827412,0.0399112813,0.0455428585,-0.1602568477,0.4140933454,-0.4061875045,-0.1118158549,0.0197634734,0.2420539558,0.0456830412,-0.0484396964,-0.2243011445,-0.2722094655,-0.1357747316,-0.184592098,0.1357543319,-0.5232682228,0.2045866102,0.1481912583,-0.1573607177,-0.2957580388,-0.0227050185,0.268961817,0.0716758817,-0.1148370355,0.326603651,0.2135397494,0.0652032346,0.1118482724,0.2881093621,0.4642577171,0.2599305511,-0.1109033898,-0.0053320415,-0.0054772645,-0.1516873986,-0.2587798536,0.2607496977,0.0704799891,0.0574535392,0.3448998928,0.166607663,-0.0934453532,-0.1094970554,0.2727870345,0.0146258213,-0.1755820364,0.1559967399,-0.4784351885,-0.0105871502,-0.3096095324,0.042605143,-0.3361676335,-0.0115367193,0.2119975388,0.0244800057,0.1457011402,-0.2048273087,0.0240990482,0.0394072495,0.4086048901,0.3885657787,0.2942712009,-0.2616147101,-0.5872488022,-0.8084726334,0.1373805106,-0.2018896788,-0.4872536361,0.0315862224,-0.1072384417,0.0769952685,0.2530308664,0.0159672648,0.3099045753,-0.0801284462,-0.1084850878,-0.2610019445,-0.2317061871,-0.2546134293,0.0900814682,-0.1903463602,-0.1421096772,0.1901785582,-0.5949850082,0.0008110312,0.163948983,-0.1057656333,-0.2082068324,-0.0478758253,0.5655726194,-0.1149526983,0.2684367895,0.0479270406,-0.2305282652,-0.1125653535,-0.252161473,0.1632338315,0.6077491045,0.0070399325,0.2551538348,-0.4812214077,0.0169900656,-0.0885087028,-0.026375182,0.2812675536,-0.1549213678,0.1070497036,-0.0785577297,-0.1062413678,0.2599004805,0.2936939001,0.3048795164,-0.2413358092,0.176688835,-0.1652900726,-0.1124801785,0.3234544694,-0.3052375019,-0.261243552,0.1704805791,0.1355917305,0.0898253173,0.1752025783,-0.2406037897,0.1892071515,0.3313728273,0.1525108516,-0.1949869394,0.2746852636,-0.3622052372,-0.1652604491,-0.0452168286,0.133011803,0.0323434584,0.0108250966,0.0206802953,-0.0255035013]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1674","title":"dutch_social can't be loaded","comments":"Maybe we should do a small release on Monday in the meantime @lhoestq ?","body":"Hi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```","comment_length":14,"text":"dutch_social can't be loaded\nHi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```\nMaybe we should do a small release on Monday in the meantime @lhoestq ?","embeddings":[-0.1440130919,-0.1277865618,-0.1560761034,0.2933968306,0.2167993635,-0.1348885447,-0.0120579246,0.0937409475,0.3958035707,-0.0862966329,-0.2941347361,-0.0078357765,0.0196760818,-0.0421471745,0.1373223066,0.1136013195,0.2246757895,0.0936486647,0.144613862,-0.1379558891,-0.2554816604,0.4084303081,0.0364790969,0.1195254549,0.0535758808,0.2432906479,-0.0837842226,0.3191888928,-0.0633874536,-0.2948622704,0.3469977081,0.1354592592,0.1217035055,0.5947429538,-0.0001220059,0.0153691275,0.4793719053,-0.0934169069,-0.6241431832,-0.4102129638,0.0205617975,-0.2827773392,0.0646422952,-0.0720438063,-0.0622491986,-0.2690680027,-0.0105327405,-0.474494487,0.2986169159,0.4165929258,0.1114755422,-0.0108238775,0.0761229396,-0.0746558085,0.0295876656,0.3834807873,0.0695490241,0.5175902247,0.2559808195,0.0191765875,0.4757993519,0.139318645,-0.1636501998,-0.1841302067,0.1682499945,-0.1772353053,-0.0807946846,-0.2221838981,0.1476258039,-0.074964039,0.4582820535,-0.0886856616,-0.264942199,-0.2424675226,0.0461236127,0.0547623821,0.276529938,0.1222155318,-0.3778547943,0.2942798138,-0.1758255213,-0.2210489362,-0.1348641068,0.542472899,0.0668623,0.0545753092,-0.1998028606,0.0972559825,0.2403409779,-0.1409109533,-0.0025449127,-0.1723716855,0.1879131049,0.2741535306,0.0476628169,0.275170058,0.109135069,0.1229047924,0.0124698281,0.2088605762,-0.0083227707,0.0611036904,-0.2934058607,0.2661069036,0.1451622844,-0.063431032,0.1614208817,-0.0578602739,0.1387722641,0.2737375498,-0.2378363311,-0.0220349152,-0.5131173134,-0.2261775881,-0.4047652185,-0.1627119333,0.0526988059,-0.1887488216,-0.2392709255,-0.0537533425,0.2383565456,0.1124858856,0.1500710696,0.4265648723,-0.0568392836,0.1309242398,0.0157592501,0.3348211646,-0.0972028598,0.1625242531,0.0043877871,0.0115900803,-0.2395053059,-0.1033063531,0.3564893603,-0.1404215097,0.3913652897,-0.0133134797,0.0193966776,-0.2262601852,-0.0651753247,-0.1723169386,-0.0208002795,0.4110073447,-0.0504028127,0.1848825663,0.2301233411,-0.3826444447,-0.2856127322,0.0220117606,-0.4266529381,-0.2579602599,-0.1370801181,0.0865042135,-0.3576443791,-0.0483067632,-0.2518779337,0.0720809549,-0.2459039837,-0.3608632684,-0.1423634887,-0.0626245961,0.0569238551,-0.1635662615,0.4047258198,0.6451685429,-0.259563148,-0.1338703334,-0.2519385815,-0.0979309827,0.1166802645,0.1549301893,-0.2774693072,0.1003124788,-0.4217948914,-0.1295697242,0.5071360469,-0.3881822526,-0.1034957767,0.043269366,0.0442071557,0.4348086715,0.0677174553,-0.0021925054,-0.2236290574,0.0936073661,0.1177948564,0.365460515,0.2958675623,-0.3151164651,-0.0436034799,-0.08808887,0.2152402699,0.4358336627,-0.0789023265,-0.065279454,0.3209886551,-0.0179740302,0.0676686615,-0.0443313085,-0.1122181565,0.3995007277,-0.1002356708,0.6040776372,-0.1201559752,0.0579323918,-0.1615790576,0.1620278955,0.3629038632,-0.0949904695,-0.2131394148,0.0145916874,-0.5914059281,0.1086386964,-0.2440940589,-0.0796029046,0.0442409627,0.3385514021,0.2097835839,0.4188049734,-0.1303366423,0.5874919891,-0.1820048988,-0.0004448143,-0.136841476,0.3376156092,-0.077340886,0.0069129346,-0.025883032,0.4640442133,0.19163315,-0.0753638968,0.1306604892,0.1775824875,0.2757100463,0.1431743801,0.3152182102,-0.1459710449,0.013373767,-0.1851287335,0.0264447127,0.2751175463,0.1668300778,-0.0880569145,-0.2502496839,0.1557591856,-0.0062821158,0.2842067778,0.1452289671,0.0032311126,0.2915836871,-0.0235694479,-0.0459581092,-0.1600633413,0.3323235512,-0.1702431589,0.5120025873,-0.006221151,-0.2748087049,0.1538019031,0.0134518212,-0.1130990013,0.0663182139,0.2667791247,-0.1694904119,0.1820977181,0.1649412066,0.2306359261,0.5306184292,0.1501768082,0.0584150627,-0.1042482406,0.1984350532,-0.226388678,0.196112588,0.0779637173,0.3595768809,0.1046478078,0.1316241622,0.1227424741,-0.2572664917,-0.5361442566,-0.149571985,0.0522079468,-0.2917066514,0.3217304945,-0.185494706,-0.4013975859,-0.3152742684,-0.1355844587,-0.2492753416,-0.3631968498,-0.1829154044,0.1399168819,0.0379369855,-0.0491964184,-0.1167751774,0.1715294868,-0.2130713314,-0.3847946823,-0.3287187815,-0.0870077685,-0.049715437,-0.0937781259,0.4112957716,0.100119248,0.1564811319,-0.4932972789,0.1215285957,0.094456993,-0.4239718616,-0.1459390521,-0.0658269599,0.1950836778,-0.0813921466,0.3485629559,-0.2351174653,-0.3012833893,0.3619758785,-0.019668784,-0.0797758698,0.1490667313,0.2029022425,-0.042098213,-0.1645146012,-0.0364278592,-0.8044582009,-0.3288797438,0.1570771933,0.2175286114,0.0154114179,0.4426355064,0.3873107731,0.0452336743,-0.0348101072,0.1303268522,-0.1764192879,-0.3531501889,0.6498783231,-0.3994504213,-0.2847266793,0.1182230935,0.2317685336,0.2142133862,-0.3526259363,-0.4864182472,0.0804351568,0.0773071796,0.2877722383,-0.0448323004,0.1416716278,-0.0867542922,-0.0689283088,-0.0460393466,-0.1349615753,-0.0135841006,-0.116391331,-0.3259768188,0.2562913597,0.0006416994,0.1128346696,-0.2048086971,0.5850800276,0.1846069694,0.1297905594,0.3108214438,-0.0963969231,0.686024785,-0.1007470489,-0.4953562319,-0.0185463745,-0.0426108614,0.0646766201,0.1125227883,0.302135855,-0.1705505848,-0.4108614326,0.052665893,-0.6017850637,-0.0814373419,0.1122563258,-0.1788872331,0.0257286951,0.0925906599,-0.1898592263,-0.3393605649,-0.1695629656,0.0055447565,0.5575093627,-0.0501216576,0.3014118075,-0.5939533114,-0.2490032464,-0.1569108814,0.1927903295,-0.1274876893,0.2557533979,-0.3264309466,0.0863464475,0.1326007843,-0.0135395527,0.4668221474,-0.0073006665,-0.0018592202,0.2152351737,0.21335949,-0.6372420788,0.1366122514,0.1852865368,0.1388081014,0.4634910524,0.3116728663,-0.1862471849,-0.0533104017,-0.0528364554,0.3992613852,-0.099411957,-0.1412382126,-0.4389512837,-0.5290256739,-0.3330169022,-0.3466005623,0.041629184,0.3000886738,-0.1127192676,0.1479955167,-0.0392461792,0.08828035,0.1973963529,0.215036571,0.3390181661,0.2029387206,-0.1517691016,0.2156927139,0.207391575,0.186458379,0.8277205825,-0.0280642305,-0.472987473,-0.0676505715,-0.1684446484,0.2539019883,0.2572754323,-0.0525502823,-0.264387399,0.0862524435,-0.0616413429,-0.1060521901,0.2844638824,-0.0042873053,0.1814712882,-0.1484197974,-0.3832840323,0.4944527149,-0.0522756279,-0.0187165849,0.1487674266,0.1603042036,-0.1453430653,0.2155505419,-0.2960323989,1.1821036339,0.0836563557,0.1619860977,0.1922802627,-0.1424458325,0.3413336873,-0.2491705269,-0.0668278858,-0.0165302716,0.1841173321,-0.0074367477,-0.150554657,0.2806473672,0.1858281046,0.0204978194,0.0036976642,-0.0962960869,0.0299856495,0.0989257619,0.4379950464,0.1277776212,-0.0180496313,-0.5319270492,0.1279685497,-0.0070529534,0.6483965516,-0.0589642003,-0.1501560062,-0.6219805479,0.0417346805,-0.1697573066,0.0559509322,-0.1763844192,0.3046876788,-0.0242759045,-0.4557029903,-0.131096974,0.3823457658,0.4158375561,0.0628483519,-0.2100280225,0.1942519248,-0.0006204766,-0.2378745824,0.1408261955,-0.2639215887,0.0989748389,-0.1657351702,-0.3407813311,0.2149266452,-0.0537951812,-0.3472219706,0.1409178823,0.0401334427,0.2564260364,0.0465484746,-0.4316743016,0.0652011633,0.2491727322,-0.1211000159,0.0465644635,0.0329136997,0.0470805205,-0.20226264,-0.223509267,-0.0786677599,-0.0912784413,0.4528014958,-0.2509536445,-0.1783564538,0.3761715591,0.1313108206,-0.2986024618,-0.094971858,0.1794088036,-0.0717810094,-0.4908860326,-0.3483213782,0.0258632731,-0.0363163538,-0.3221884966,0.1668507606,0.2664426863,-0.2686613202,0.1378796399,-0.7992363572,-0.009050753,0.2148845643,0.0030076047,0.0240182113,0.149524793,-0.1888760328,0.1773919314,-0.1492820978,-0.2086732835,0.1037804931,-0.2042625546,-0.0708857626,0.3645169139,0.0797675326,0.1445762366,-0.2587434351,0.0331773981,-0.3906826675,0.0711602569,-0.1381139606,-0.0232484508,0.167788595,-0.030742228,-0.0736366659,-0.1844884306,-0.423468709,0.0495093353,-0.0976031423,-0.1504056454,-0.0286711175,0.0516630039,-0.1411792636,-0.0938978791,0.3486450016,-0.0014615451,0.2108521163,0.0660802647,0.5716895461,0.0365405045,0.0253206547,0.1066599116,0.0104909716,-0.052856829,0.1187435761,-0.1242643297,0.1727949828,0.3727681637,-0.3196749091,0.3249733746,0.120964922,-0.0035803635,0.5077812672,-0.441368252,0.1462892592,0.3029463887,0.0475829877,-0.2808189392,-0.1859784126,0.4009407163,0.0555047877,-0.0616027452,0.2878829837,0.2537675798,-0.1890611053,0.389048636,0.1293126196,0.4627902508,0.4501369894,0.2029798329,0.0939915404,0.0795385763,0.0393343084,-0.0704107285,0.320171982,-0.0320692174,0.4540109932,-0.261349231,0.1091690212,0.0425857119,0.1719930172,-0.1905469745,-0.2267156392,0.1692724973,-0.1652827412,0.0399112813,0.0455428585,-0.1602568477,0.4140933454,-0.4061875045,-0.1118158549,0.0197634734,0.2420539558,0.0456830412,-0.0484396964,-0.2243011445,-0.2722094655,-0.1357747316,-0.184592098,0.1357543319,-0.5232682228,0.2045866102,0.1481912583,-0.1573607177,-0.2957580388,-0.0227050185,0.268961817,0.0716758817,-0.1148370355,0.326603651,0.2135397494,0.0652032346,0.1118482724,0.2881093621,0.4642577171,0.2599305511,-0.1109033898,-0.0053320415,-0.0054772645,-0.1516873986,-0.2587798536,0.2607496977,0.0704799891,0.0574535392,0.3448998928,0.166607663,-0.0934453532,-0.1094970554,0.2727870345,0.0146258213,-0.1755820364,0.1559967399,-0.4784351885,-0.0105871502,-0.3096095324,0.042605143,-0.3361676335,-0.0115367193,0.2119975388,0.0244800057,0.1457011402,-0.2048273087,0.0240990482,0.0394072495,0.4086048901,0.3885657787,0.2942712009,-0.2616147101,-0.5872488022,-0.8084726334,0.1373805106,-0.2018896788,-0.4872536361,0.0315862224,-0.1072384417,0.0769952685,0.2530308664,0.0159672648,0.3099045753,-0.0801284462,-0.1084850878,-0.2610019445,-0.2317061871,-0.2546134293,0.0900814682,-0.1903463602,-0.1421096772,0.1901785582,-0.5949850082,0.0008110312,0.163948983,-0.1057656333,-0.2082068324,-0.0478758253,0.5655726194,-0.1149526983,0.2684367895,0.0479270406,-0.2305282652,-0.1125653535,-0.252161473,0.1632338315,0.6077491045,0.0070399325,0.2551538348,-0.4812214077,0.0169900656,-0.0885087028,-0.026375182,0.2812675536,-0.1549213678,0.1070497036,-0.0785577297,-0.1062413678,0.2599004805,0.2936939001,0.3048795164,-0.2413358092,0.176688835,-0.1652900726,-0.1124801785,0.3234544694,-0.3052375019,-0.261243552,0.1704805791,0.1355917305,0.0898253173,0.1752025783,-0.2406037897,0.1892071515,0.3313728273,0.1525108516,-0.1949869394,0.2746852636,-0.3622052372,-0.1652604491,-0.0452168286,0.133011803,0.0323434584,0.0108250966,0.0206802953,-0.0255035013]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1674","title":"dutch_social can't be loaded","comments":"Yes sure !","body":"Hi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```","comment_length":3,"text":"dutch_social can't be loaded\nHi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```\nYes sure !","embeddings":[-0.1440130919,-0.1277865618,-0.1560761034,0.2933968306,0.2167993635,-0.1348885447,-0.0120579246,0.0937409475,0.3958035707,-0.0862966329,-0.2941347361,-0.0078357765,0.0196760818,-0.0421471745,0.1373223066,0.1136013195,0.2246757895,0.0936486647,0.144613862,-0.1379558891,-0.2554816604,0.4084303081,0.0364790969,0.1195254549,0.0535758808,0.2432906479,-0.0837842226,0.3191888928,-0.0633874536,-0.2948622704,0.3469977081,0.1354592592,0.1217035055,0.5947429538,-0.0001220059,0.0153691275,0.4793719053,-0.0934169069,-0.6241431832,-0.4102129638,0.0205617975,-0.2827773392,0.0646422952,-0.0720438063,-0.0622491986,-0.2690680027,-0.0105327405,-0.474494487,0.2986169159,0.4165929258,0.1114755422,-0.0108238775,0.0761229396,-0.0746558085,0.0295876656,0.3834807873,0.0695490241,0.5175902247,0.2559808195,0.0191765875,0.4757993519,0.139318645,-0.1636501998,-0.1841302067,0.1682499945,-0.1772353053,-0.0807946846,-0.2221838981,0.1476258039,-0.074964039,0.4582820535,-0.0886856616,-0.264942199,-0.2424675226,0.0461236127,0.0547623821,0.276529938,0.1222155318,-0.3778547943,0.2942798138,-0.1758255213,-0.2210489362,-0.1348641068,0.542472899,0.0668623,0.0545753092,-0.1998028606,0.0972559825,0.2403409779,-0.1409109533,-0.0025449127,-0.1723716855,0.1879131049,0.2741535306,0.0476628169,0.275170058,0.109135069,0.1229047924,0.0124698281,0.2088605762,-0.0083227707,0.0611036904,-0.2934058607,0.2661069036,0.1451622844,-0.063431032,0.1614208817,-0.0578602739,0.1387722641,0.2737375498,-0.2378363311,-0.0220349152,-0.5131173134,-0.2261775881,-0.4047652185,-0.1627119333,0.0526988059,-0.1887488216,-0.2392709255,-0.0537533425,0.2383565456,0.1124858856,0.1500710696,0.4265648723,-0.0568392836,0.1309242398,0.0157592501,0.3348211646,-0.0972028598,0.1625242531,0.0043877871,0.0115900803,-0.2395053059,-0.1033063531,0.3564893603,-0.1404215097,0.3913652897,-0.0133134797,0.0193966776,-0.2262601852,-0.0651753247,-0.1723169386,-0.0208002795,0.4110073447,-0.0504028127,0.1848825663,0.2301233411,-0.3826444447,-0.2856127322,0.0220117606,-0.4266529381,-0.2579602599,-0.1370801181,0.0865042135,-0.3576443791,-0.0483067632,-0.2518779337,0.0720809549,-0.2459039837,-0.3608632684,-0.1423634887,-0.0626245961,0.0569238551,-0.1635662615,0.4047258198,0.6451685429,-0.259563148,-0.1338703334,-0.2519385815,-0.0979309827,0.1166802645,0.1549301893,-0.2774693072,0.1003124788,-0.4217948914,-0.1295697242,0.5071360469,-0.3881822526,-0.1034957767,0.043269366,0.0442071557,0.4348086715,0.0677174553,-0.0021925054,-0.2236290574,0.0936073661,0.1177948564,0.365460515,0.2958675623,-0.3151164651,-0.0436034799,-0.08808887,0.2152402699,0.4358336627,-0.0789023265,-0.065279454,0.3209886551,-0.0179740302,0.0676686615,-0.0443313085,-0.1122181565,0.3995007277,-0.1002356708,0.6040776372,-0.1201559752,0.0579323918,-0.1615790576,0.1620278955,0.3629038632,-0.0949904695,-0.2131394148,0.0145916874,-0.5914059281,0.1086386964,-0.2440940589,-0.0796029046,0.0442409627,0.3385514021,0.2097835839,0.4188049734,-0.1303366423,0.5874919891,-0.1820048988,-0.0004448143,-0.136841476,0.3376156092,-0.077340886,0.0069129346,-0.025883032,0.4640442133,0.19163315,-0.0753638968,0.1306604892,0.1775824875,0.2757100463,0.1431743801,0.3152182102,-0.1459710449,0.013373767,-0.1851287335,0.0264447127,0.2751175463,0.1668300778,-0.0880569145,-0.2502496839,0.1557591856,-0.0062821158,0.2842067778,0.1452289671,0.0032311126,0.2915836871,-0.0235694479,-0.0459581092,-0.1600633413,0.3323235512,-0.1702431589,0.5120025873,-0.006221151,-0.2748087049,0.1538019031,0.0134518212,-0.1130990013,0.0663182139,0.2667791247,-0.1694904119,0.1820977181,0.1649412066,0.2306359261,0.5306184292,0.1501768082,0.0584150627,-0.1042482406,0.1984350532,-0.226388678,0.196112588,0.0779637173,0.3595768809,0.1046478078,0.1316241622,0.1227424741,-0.2572664917,-0.5361442566,-0.149571985,0.0522079468,-0.2917066514,0.3217304945,-0.185494706,-0.4013975859,-0.3152742684,-0.1355844587,-0.2492753416,-0.3631968498,-0.1829154044,0.1399168819,0.0379369855,-0.0491964184,-0.1167751774,0.1715294868,-0.2130713314,-0.3847946823,-0.3287187815,-0.0870077685,-0.049715437,-0.0937781259,0.4112957716,0.100119248,0.1564811319,-0.4932972789,0.1215285957,0.094456993,-0.4239718616,-0.1459390521,-0.0658269599,0.1950836778,-0.0813921466,0.3485629559,-0.2351174653,-0.3012833893,0.3619758785,-0.019668784,-0.0797758698,0.1490667313,0.2029022425,-0.042098213,-0.1645146012,-0.0364278592,-0.8044582009,-0.3288797438,0.1570771933,0.2175286114,0.0154114179,0.4426355064,0.3873107731,0.0452336743,-0.0348101072,0.1303268522,-0.1764192879,-0.3531501889,0.6498783231,-0.3994504213,-0.2847266793,0.1182230935,0.2317685336,0.2142133862,-0.3526259363,-0.4864182472,0.0804351568,0.0773071796,0.2877722383,-0.0448323004,0.1416716278,-0.0867542922,-0.0689283088,-0.0460393466,-0.1349615753,-0.0135841006,-0.116391331,-0.3259768188,0.2562913597,0.0006416994,0.1128346696,-0.2048086971,0.5850800276,0.1846069694,0.1297905594,0.3108214438,-0.0963969231,0.686024785,-0.1007470489,-0.4953562319,-0.0185463745,-0.0426108614,0.0646766201,0.1125227883,0.302135855,-0.1705505848,-0.4108614326,0.052665893,-0.6017850637,-0.0814373419,0.1122563258,-0.1788872331,0.0257286951,0.0925906599,-0.1898592263,-0.3393605649,-0.1695629656,0.0055447565,0.5575093627,-0.0501216576,0.3014118075,-0.5939533114,-0.2490032464,-0.1569108814,0.1927903295,-0.1274876893,0.2557533979,-0.3264309466,0.0863464475,0.1326007843,-0.0135395527,0.4668221474,-0.0073006665,-0.0018592202,0.2152351737,0.21335949,-0.6372420788,0.1366122514,0.1852865368,0.1388081014,0.4634910524,0.3116728663,-0.1862471849,-0.0533104017,-0.0528364554,0.3992613852,-0.099411957,-0.1412382126,-0.4389512837,-0.5290256739,-0.3330169022,-0.3466005623,0.041629184,0.3000886738,-0.1127192676,0.1479955167,-0.0392461792,0.08828035,0.1973963529,0.215036571,0.3390181661,0.2029387206,-0.1517691016,0.2156927139,0.207391575,0.186458379,0.8277205825,-0.0280642305,-0.472987473,-0.0676505715,-0.1684446484,0.2539019883,0.2572754323,-0.0525502823,-0.264387399,0.0862524435,-0.0616413429,-0.1060521901,0.2844638824,-0.0042873053,0.1814712882,-0.1484197974,-0.3832840323,0.4944527149,-0.0522756279,-0.0187165849,0.1487674266,0.1603042036,-0.1453430653,0.2155505419,-0.2960323989,1.1821036339,0.0836563557,0.1619860977,0.1922802627,-0.1424458325,0.3413336873,-0.2491705269,-0.0668278858,-0.0165302716,0.1841173321,-0.0074367477,-0.150554657,0.2806473672,0.1858281046,0.0204978194,0.0036976642,-0.0962960869,0.0299856495,0.0989257619,0.4379950464,0.1277776212,-0.0180496313,-0.5319270492,0.1279685497,-0.0070529534,0.6483965516,-0.0589642003,-0.1501560062,-0.6219805479,0.0417346805,-0.1697573066,0.0559509322,-0.1763844192,0.3046876788,-0.0242759045,-0.4557029903,-0.131096974,0.3823457658,0.4158375561,0.0628483519,-0.2100280225,0.1942519248,-0.0006204766,-0.2378745824,0.1408261955,-0.2639215887,0.0989748389,-0.1657351702,-0.3407813311,0.2149266452,-0.0537951812,-0.3472219706,0.1409178823,0.0401334427,0.2564260364,0.0465484746,-0.4316743016,0.0652011633,0.2491727322,-0.1211000159,0.0465644635,0.0329136997,0.0470805205,-0.20226264,-0.223509267,-0.0786677599,-0.0912784413,0.4528014958,-0.2509536445,-0.1783564538,0.3761715591,0.1313108206,-0.2986024618,-0.094971858,0.1794088036,-0.0717810094,-0.4908860326,-0.3483213782,0.0258632731,-0.0363163538,-0.3221884966,0.1668507606,0.2664426863,-0.2686613202,0.1378796399,-0.7992363572,-0.009050753,0.2148845643,0.0030076047,0.0240182113,0.149524793,-0.1888760328,0.1773919314,-0.1492820978,-0.2086732835,0.1037804931,-0.2042625546,-0.0708857626,0.3645169139,0.0797675326,0.1445762366,-0.2587434351,0.0331773981,-0.3906826675,0.0711602569,-0.1381139606,-0.0232484508,0.167788595,-0.030742228,-0.0736366659,-0.1844884306,-0.423468709,0.0495093353,-0.0976031423,-0.1504056454,-0.0286711175,0.0516630039,-0.1411792636,-0.0938978791,0.3486450016,-0.0014615451,0.2108521163,0.0660802647,0.5716895461,0.0365405045,0.0253206547,0.1066599116,0.0104909716,-0.052856829,0.1187435761,-0.1242643297,0.1727949828,0.3727681637,-0.3196749091,0.3249733746,0.120964922,-0.0035803635,0.5077812672,-0.441368252,0.1462892592,0.3029463887,0.0475829877,-0.2808189392,-0.1859784126,0.4009407163,0.0555047877,-0.0616027452,0.2878829837,0.2537675798,-0.1890611053,0.389048636,0.1293126196,0.4627902508,0.4501369894,0.2029798329,0.0939915404,0.0795385763,0.0393343084,-0.0704107285,0.320171982,-0.0320692174,0.4540109932,-0.261349231,0.1091690212,0.0425857119,0.1719930172,-0.1905469745,-0.2267156392,0.1692724973,-0.1652827412,0.0399112813,0.0455428585,-0.1602568477,0.4140933454,-0.4061875045,-0.1118158549,0.0197634734,0.2420539558,0.0456830412,-0.0484396964,-0.2243011445,-0.2722094655,-0.1357747316,-0.184592098,0.1357543319,-0.5232682228,0.2045866102,0.1481912583,-0.1573607177,-0.2957580388,-0.0227050185,0.268961817,0.0716758817,-0.1148370355,0.326603651,0.2135397494,0.0652032346,0.1118482724,0.2881093621,0.4642577171,0.2599305511,-0.1109033898,-0.0053320415,-0.0054772645,-0.1516873986,-0.2587798536,0.2607496977,0.0704799891,0.0574535392,0.3448998928,0.166607663,-0.0934453532,-0.1094970554,0.2727870345,0.0146258213,-0.1755820364,0.1559967399,-0.4784351885,-0.0105871502,-0.3096095324,0.042605143,-0.3361676335,-0.0115367193,0.2119975388,0.0244800057,0.1457011402,-0.2048273087,0.0240990482,0.0394072495,0.4086048901,0.3885657787,0.2942712009,-0.2616147101,-0.5872488022,-0.8084726334,0.1373805106,-0.2018896788,-0.4872536361,0.0315862224,-0.1072384417,0.0769952685,0.2530308664,0.0159672648,0.3099045753,-0.0801284462,-0.1084850878,-0.2610019445,-0.2317061871,-0.2546134293,0.0900814682,-0.1903463602,-0.1421096772,0.1901785582,-0.5949850082,0.0008110312,0.163948983,-0.1057656333,-0.2082068324,-0.0478758253,0.5655726194,-0.1149526983,0.2684367895,0.0479270406,-0.2305282652,-0.1125653535,-0.252161473,0.1632338315,0.6077491045,0.0070399325,0.2551538348,-0.4812214077,0.0169900656,-0.0885087028,-0.026375182,0.2812675536,-0.1549213678,0.1070497036,-0.0785577297,-0.1062413678,0.2599004805,0.2936939001,0.3048795164,-0.2413358092,0.176688835,-0.1652900726,-0.1124801785,0.3234544694,-0.3052375019,-0.261243552,0.1704805791,0.1355917305,0.0898253173,0.1752025783,-0.2406037897,0.1892071515,0.3313728273,0.1525108516,-0.1949869394,0.2746852636,-0.3622052372,-0.1652604491,-0.0452168286,0.133011803,0.0323434584,0.0108250966,0.0206802953,-0.0255035013]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1674","title":"dutch_social can't be loaded","comments":"I just did the release :)\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `dutch_social` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"dutch_social\")\r\n```","body":"Hi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```","comment_length":36,"text":"dutch_social can't be loaded\nHi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```\nI just did the release :)\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `dutch_social` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"dutch_social\")\r\n```","embeddings":[-0.1440130919,-0.1277865618,-0.1560761034,0.2933968306,0.2167993635,-0.1348885447,-0.0120579246,0.0937409475,0.3958035707,-0.0862966329,-0.2941347361,-0.0078357765,0.0196760818,-0.0421471745,0.1373223066,0.1136013195,0.2246757895,0.0936486647,0.144613862,-0.1379558891,-0.2554816604,0.4084303081,0.0364790969,0.1195254549,0.0535758808,0.2432906479,-0.0837842226,0.3191888928,-0.0633874536,-0.2948622704,0.3469977081,0.1354592592,0.1217035055,0.5947429538,-0.0001220059,0.0153691275,0.4793719053,-0.0934169069,-0.6241431832,-0.4102129638,0.0205617975,-0.2827773392,0.0646422952,-0.0720438063,-0.0622491986,-0.2690680027,-0.0105327405,-0.474494487,0.2986169159,0.4165929258,0.1114755422,-0.0108238775,0.0761229396,-0.0746558085,0.0295876656,0.3834807873,0.0695490241,0.5175902247,0.2559808195,0.0191765875,0.4757993519,0.139318645,-0.1636501998,-0.1841302067,0.1682499945,-0.1772353053,-0.0807946846,-0.2221838981,0.1476258039,-0.074964039,0.4582820535,-0.0886856616,-0.264942199,-0.2424675226,0.0461236127,0.0547623821,0.276529938,0.1222155318,-0.3778547943,0.2942798138,-0.1758255213,-0.2210489362,-0.1348641068,0.542472899,0.0668623,0.0545753092,-0.1998028606,0.0972559825,0.2403409779,-0.1409109533,-0.0025449127,-0.1723716855,0.1879131049,0.2741535306,0.0476628169,0.275170058,0.109135069,0.1229047924,0.0124698281,0.2088605762,-0.0083227707,0.0611036904,-0.2934058607,0.2661069036,0.1451622844,-0.063431032,0.1614208817,-0.0578602739,0.1387722641,0.2737375498,-0.2378363311,-0.0220349152,-0.5131173134,-0.2261775881,-0.4047652185,-0.1627119333,0.0526988059,-0.1887488216,-0.2392709255,-0.0537533425,0.2383565456,0.1124858856,0.1500710696,0.4265648723,-0.0568392836,0.1309242398,0.0157592501,0.3348211646,-0.0972028598,0.1625242531,0.0043877871,0.0115900803,-0.2395053059,-0.1033063531,0.3564893603,-0.1404215097,0.3913652897,-0.0133134797,0.0193966776,-0.2262601852,-0.0651753247,-0.1723169386,-0.0208002795,0.4110073447,-0.0504028127,0.1848825663,0.2301233411,-0.3826444447,-0.2856127322,0.0220117606,-0.4266529381,-0.2579602599,-0.1370801181,0.0865042135,-0.3576443791,-0.0483067632,-0.2518779337,0.0720809549,-0.2459039837,-0.3608632684,-0.1423634887,-0.0626245961,0.0569238551,-0.1635662615,0.4047258198,0.6451685429,-0.259563148,-0.1338703334,-0.2519385815,-0.0979309827,0.1166802645,0.1549301893,-0.2774693072,0.1003124788,-0.4217948914,-0.1295697242,0.5071360469,-0.3881822526,-0.1034957767,0.043269366,0.0442071557,0.4348086715,0.0677174553,-0.0021925054,-0.2236290574,0.0936073661,0.1177948564,0.365460515,0.2958675623,-0.3151164651,-0.0436034799,-0.08808887,0.2152402699,0.4358336627,-0.0789023265,-0.065279454,0.3209886551,-0.0179740302,0.0676686615,-0.0443313085,-0.1122181565,0.3995007277,-0.1002356708,0.6040776372,-0.1201559752,0.0579323918,-0.1615790576,0.1620278955,0.3629038632,-0.0949904695,-0.2131394148,0.0145916874,-0.5914059281,0.1086386964,-0.2440940589,-0.0796029046,0.0442409627,0.3385514021,0.2097835839,0.4188049734,-0.1303366423,0.5874919891,-0.1820048988,-0.0004448143,-0.136841476,0.3376156092,-0.077340886,0.0069129346,-0.025883032,0.4640442133,0.19163315,-0.0753638968,0.1306604892,0.1775824875,0.2757100463,0.1431743801,0.3152182102,-0.1459710449,0.013373767,-0.1851287335,0.0264447127,0.2751175463,0.1668300778,-0.0880569145,-0.2502496839,0.1557591856,-0.0062821158,0.2842067778,0.1452289671,0.0032311126,0.2915836871,-0.0235694479,-0.0459581092,-0.1600633413,0.3323235512,-0.1702431589,0.5120025873,-0.006221151,-0.2748087049,0.1538019031,0.0134518212,-0.1130990013,0.0663182139,0.2667791247,-0.1694904119,0.1820977181,0.1649412066,0.2306359261,0.5306184292,0.1501768082,0.0584150627,-0.1042482406,0.1984350532,-0.226388678,0.196112588,0.0779637173,0.3595768809,0.1046478078,0.1316241622,0.1227424741,-0.2572664917,-0.5361442566,-0.149571985,0.0522079468,-0.2917066514,0.3217304945,-0.185494706,-0.4013975859,-0.3152742684,-0.1355844587,-0.2492753416,-0.3631968498,-0.1829154044,0.1399168819,0.0379369855,-0.0491964184,-0.1167751774,0.1715294868,-0.2130713314,-0.3847946823,-0.3287187815,-0.0870077685,-0.049715437,-0.0937781259,0.4112957716,0.100119248,0.1564811319,-0.4932972789,0.1215285957,0.094456993,-0.4239718616,-0.1459390521,-0.0658269599,0.1950836778,-0.0813921466,0.3485629559,-0.2351174653,-0.3012833893,0.3619758785,-0.019668784,-0.0797758698,0.1490667313,0.2029022425,-0.042098213,-0.1645146012,-0.0364278592,-0.8044582009,-0.3288797438,0.1570771933,0.2175286114,0.0154114179,0.4426355064,0.3873107731,0.0452336743,-0.0348101072,0.1303268522,-0.1764192879,-0.3531501889,0.6498783231,-0.3994504213,-0.2847266793,0.1182230935,0.2317685336,0.2142133862,-0.3526259363,-0.4864182472,0.0804351568,0.0773071796,0.2877722383,-0.0448323004,0.1416716278,-0.0867542922,-0.0689283088,-0.0460393466,-0.1349615753,-0.0135841006,-0.116391331,-0.3259768188,0.2562913597,0.0006416994,0.1128346696,-0.2048086971,0.5850800276,0.1846069694,0.1297905594,0.3108214438,-0.0963969231,0.686024785,-0.1007470489,-0.4953562319,-0.0185463745,-0.0426108614,0.0646766201,0.1125227883,0.302135855,-0.1705505848,-0.4108614326,0.052665893,-0.6017850637,-0.0814373419,0.1122563258,-0.1788872331,0.0257286951,0.0925906599,-0.1898592263,-0.3393605649,-0.1695629656,0.0055447565,0.5575093627,-0.0501216576,0.3014118075,-0.5939533114,-0.2490032464,-0.1569108814,0.1927903295,-0.1274876893,0.2557533979,-0.3264309466,0.0863464475,0.1326007843,-0.0135395527,0.4668221474,-0.0073006665,-0.0018592202,0.2152351737,0.21335949,-0.6372420788,0.1366122514,0.1852865368,0.1388081014,0.4634910524,0.3116728663,-0.1862471849,-0.0533104017,-0.0528364554,0.3992613852,-0.099411957,-0.1412382126,-0.4389512837,-0.5290256739,-0.3330169022,-0.3466005623,0.041629184,0.3000886738,-0.1127192676,0.1479955167,-0.0392461792,0.08828035,0.1973963529,0.215036571,0.3390181661,0.2029387206,-0.1517691016,0.2156927139,0.207391575,0.186458379,0.8277205825,-0.0280642305,-0.472987473,-0.0676505715,-0.1684446484,0.2539019883,0.2572754323,-0.0525502823,-0.264387399,0.0862524435,-0.0616413429,-0.1060521901,0.2844638824,-0.0042873053,0.1814712882,-0.1484197974,-0.3832840323,0.4944527149,-0.0522756279,-0.0187165849,0.1487674266,0.1603042036,-0.1453430653,0.2155505419,-0.2960323989,1.1821036339,0.0836563557,0.1619860977,0.1922802627,-0.1424458325,0.3413336873,-0.2491705269,-0.0668278858,-0.0165302716,0.1841173321,-0.0074367477,-0.150554657,0.2806473672,0.1858281046,0.0204978194,0.0036976642,-0.0962960869,0.0299856495,0.0989257619,0.4379950464,0.1277776212,-0.0180496313,-0.5319270492,0.1279685497,-0.0070529534,0.6483965516,-0.0589642003,-0.1501560062,-0.6219805479,0.0417346805,-0.1697573066,0.0559509322,-0.1763844192,0.3046876788,-0.0242759045,-0.4557029903,-0.131096974,0.3823457658,0.4158375561,0.0628483519,-0.2100280225,0.1942519248,-0.0006204766,-0.2378745824,0.1408261955,-0.2639215887,0.0989748389,-0.1657351702,-0.3407813311,0.2149266452,-0.0537951812,-0.3472219706,0.1409178823,0.0401334427,0.2564260364,0.0465484746,-0.4316743016,0.0652011633,0.2491727322,-0.1211000159,0.0465644635,0.0329136997,0.0470805205,-0.20226264,-0.223509267,-0.0786677599,-0.0912784413,0.4528014958,-0.2509536445,-0.1783564538,0.3761715591,0.1313108206,-0.2986024618,-0.094971858,0.1794088036,-0.0717810094,-0.4908860326,-0.3483213782,0.0258632731,-0.0363163538,-0.3221884966,0.1668507606,0.2664426863,-0.2686613202,0.1378796399,-0.7992363572,-0.009050753,0.2148845643,0.0030076047,0.0240182113,0.149524793,-0.1888760328,0.1773919314,-0.1492820978,-0.2086732835,0.1037804931,-0.2042625546,-0.0708857626,0.3645169139,0.0797675326,0.1445762366,-0.2587434351,0.0331773981,-0.3906826675,0.0711602569,-0.1381139606,-0.0232484508,0.167788595,-0.030742228,-0.0736366659,-0.1844884306,-0.423468709,0.0495093353,-0.0976031423,-0.1504056454,-0.0286711175,0.0516630039,-0.1411792636,-0.0938978791,0.3486450016,-0.0014615451,0.2108521163,0.0660802647,0.5716895461,0.0365405045,0.0253206547,0.1066599116,0.0104909716,-0.052856829,0.1187435761,-0.1242643297,0.1727949828,0.3727681637,-0.3196749091,0.3249733746,0.120964922,-0.0035803635,0.5077812672,-0.441368252,0.1462892592,0.3029463887,0.0475829877,-0.2808189392,-0.1859784126,0.4009407163,0.0555047877,-0.0616027452,0.2878829837,0.2537675798,-0.1890611053,0.389048636,0.1293126196,0.4627902508,0.4501369894,0.2029798329,0.0939915404,0.0795385763,0.0393343084,-0.0704107285,0.320171982,-0.0320692174,0.4540109932,-0.261349231,0.1091690212,0.0425857119,0.1719930172,-0.1905469745,-0.2267156392,0.1692724973,-0.1652827412,0.0399112813,0.0455428585,-0.1602568477,0.4140933454,-0.4061875045,-0.1118158549,0.0197634734,0.2420539558,0.0456830412,-0.0484396964,-0.2243011445,-0.2722094655,-0.1357747316,-0.184592098,0.1357543319,-0.5232682228,0.2045866102,0.1481912583,-0.1573607177,-0.2957580388,-0.0227050185,0.268961817,0.0716758817,-0.1148370355,0.326603651,0.2135397494,0.0652032346,0.1118482724,0.2881093621,0.4642577171,0.2599305511,-0.1109033898,-0.0053320415,-0.0054772645,-0.1516873986,-0.2587798536,0.2607496977,0.0704799891,0.0574535392,0.3448998928,0.166607663,-0.0934453532,-0.1094970554,0.2727870345,0.0146258213,-0.1755820364,0.1559967399,-0.4784351885,-0.0105871502,-0.3096095324,0.042605143,-0.3361676335,-0.0115367193,0.2119975388,0.0244800057,0.1457011402,-0.2048273087,0.0240990482,0.0394072495,0.4086048901,0.3885657787,0.2942712009,-0.2616147101,-0.5872488022,-0.8084726334,0.1373805106,-0.2018896788,-0.4872536361,0.0315862224,-0.1072384417,0.0769952685,0.2530308664,0.0159672648,0.3099045753,-0.0801284462,-0.1084850878,-0.2610019445,-0.2317061871,-0.2546134293,0.0900814682,-0.1903463602,-0.1421096772,0.1901785582,-0.5949850082,0.0008110312,0.163948983,-0.1057656333,-0.2082068324,-0.0478758253,0.5655726194,-0.1149526983,0.2684367895,0.0479270406,-0.2305282652,-0.1125653535,-0.252161473,0.1632338315,0.6077491045,0.0070399325,0.2551538348,-0.4812214077,0.0169900656,-0.0885087028,-0.026375182,0.2812675536,-0.1549213678,0.1070497036,-0.0785577297,-0.1062413678,0.2599004805,0.2936939001,0.3048795164,-0.2413358092,0.176688835,-0.1652900726,-0.1124801785,0.3234544694,-0.3052375019,-0.261243552,0.1704805791,0.1355917305,0.0898253173,0.1752025783,-0.2406037897,0.1892071515,0.3313728273,0.1525108516,-0.1949869394,0.2746852636,-0.3622052372,-0.1652604491,-0.0452168286,0.133011803,0.0323434584,0.0108250966,0.0206802953,-0.0255035013]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1674","title":"dutch_social can't be loaded","comments":"@lhoestq could you also shed light on the Hindi Wikipedia Dataset for issue number #1673. Will this also be available in the new release that you committed recently?","body":"Hi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```","comment_length":28,"text":"dutch_social can't be loaded\nHi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```\n@lhoestq could you also shed light on the Hindi Wikipedia Dataset for issue number #1673. Will this also be available in the new release that you committed recently?","embeddings":[-0.1440130919,-0.1277865618,-0.1560761034,0.2933968306,0.2167993635,-0.1348885447,-0.0120579246,0.0937409475,0.3958035707,-0.0862966329,-0.2941347361,-0.0078357765,0.0196760818,-0.0421471745,0.1373223066,0.1136013195,0.2246757895,0.0936486647,0.144613862,-0.1379558891,-0.2554816604,0.4084303081,0.0364790969,0.1195254549,0.0535758808,0.2432906479,-0.0837842226,0.3191888928,-0.0633874536,-0.2948622704,0.3469977081,0.1354592592,0.1217035055,0.5947429538,-0.0001220059,0.0153691275,0.4793719053,-0.0934169069,-0.6241431832,-0.4102129638,0.0205617975,-0.2827773392,0.0646422952,-0.0720438063,-0.0622491986,-0.2690680027,-0.0105327405,-0.474494487,0.2986169159,0.4165929258,0.1114755422,-0.0108238775,0.0761229396,-0.0746558085,0.0295876656,0.3834807873,0.0695490241,0.5175902247,0.2559808195,0.0191765875,0.4757993519,0.139318645,-0.1636501998,-0.1841302067,0.1682499945,-0.1772353053,-0.0807946846,-0.2221838981,0.1476258039,-0.074964039,0.4582820535,-0.0886856616,-0.264942199,-0.2424675226,0.0461236127,0.0547623821,0.276529938,0.1222155318,-0.3778547943,0.2942798138,-0.1758255213,-0.2210489362,-0.1348641068,0.542472899,0.0668623,0.0545753092,-0.1998028606,0.0972559825,0.2403409779,-0.1409109533,-0.0025449127,-0.1723716855,0.1879131049,0.2741535306,0.0476628169,0.275170058,0.109135069,0.1229047924,0.0124698281,0.2088605762,-0.0083227707,0.0611036904,-0.2934058607,0.2661069036,0.1451622844,-0.063431032,0.1614208817,-0.0578602739,0.1387722641,0.2737375498,-0.2378363311,-0.0220349152,-0.5131173134,-0.2261775881,-0.4047652185,-0.1627119333,0.0526988059,-0.1887488216,-0.2392709255,-0.0537533425,0.2383565456,0.1124858856,0.1500710696,0.4265648723,-0.0568392836,0.1309242398,0.0157592501,0.3348211646,-0.0972028598,0.1625242531,0.0043877871,0.0115900803,-0.2395053059,-0.1033063531,0.3564893603,-0.1404215097,0.3913652897,-0.0133134797,0.0193966776,-0.2262601852,-0.0651753247,-0.1723169386,-0.0208002795,0.4110073447,-0.0504028127,0.1848825663,0.2301233411,-0.3826444447,-0.2856127322,0.0220117606,-0.4266529381,-0.2579602599,-0.1370801181,0.0865042135,-0.3576443791,-0.0483067632,-0.2518779337,0.0720809549,-0.2459039837,-0.3608632684,-0.1423634887,-0.0626245961,0.0569238551,-0.1635662615,0.4047258198,0.6451685429,-0.259563148,-0.1338703334,-0.2519385815,-0.0979309827,0.1166802645,0.1549301893,-0.2774693072,0.1003124788,-0.4217948914,-0.1295697242,0.5071360469,-0.3881822526,-0.1034957767,0.043269366,0.0442071557,0.4348086715,0.0677174553,-0.0021925054,-0.2236290574,0.0936073661,0.1177948564,0.365460515,0.2958675623,-0.3151164651,-0.0436034799,-0.08808887,0.2152402699,0.4358336627,-0.0789023265,-0.065279454,0.3209886551,-0.0179740302,0.0676686615,-0.0443313085,-0.1122181565,0.3995007277,-0.1002356708,0.6040776372,-0.1201559752,0.0579323918,-0.1615790576,0.1620278955,0.3629038632,-0.0949904695,-0.2131394148,0.0145916874,-0.5914059281,0.1086386964,-0.2440940589,-0.0796029046,0.0442409627,0.3385514021,0.2097835839,0.4188049734,-0.1303366423,0.5874919891,-0.1820048988,-0.0004448143,-0.136841476,0.3376156092,-0.077340886,0.0069129346,-0.025883032,0.4640442133,0.19163315,-0.0753638968,0.1306604892,0.1775824875,0.2757100463,0.1431743801,0.3152182102,-0.1459710449,0.013373767,-0.1851287335,0.0264447127,0.2751175463,0.1668300778,-0.0880569145,-0.2502496839,0.1557591856,-0.0062821158,0.2842067778,0.1452289671,0.0032311126,0.2915836871,-0.0235694479,-0.0459581092,-0.1600633413,0.3323235512,-0.1702431589,0.5120025873,-0.006221151,-0.2748087049,0.1538019031,0.0134518212,-0.1130990013,0.0663182139,0.2667791247,-0.1694904119,0.1820977181,0.1649412066,0.2306359261,0.5306184292,0.1501768082,0.0584150627,-0.1042482406,0.1984350532,-0.226388678,0.196112588,0.0779637173,0.3595768809,0.1046478078,0.1316241622,0.1227424741,-0.2572664917,-0.5361442566,-0.149571985,0.0522079468,-0.2917066514,0.3217304945,-0.185494706,-0.4013975859,-0.3152742684,-0.1355844587,-0.2492753416,-0.3631968498,-0.1829154044,0.1399168819,0.0379369855,-0.0491964184,-0.1167751774,0.1715294868,-0.2130713314,-0.3847946823,-0.3287187815,-0.0870077685,-0.049715437,-0.0937781259,0.4112957716,0.100119248,0.1564811319,-0.4932972789,0.1215285957,0.094456993,-0.4239718616,-0.1459390521,-0.0658269599,0.1950836778,-0.0813921466,0.3485629559,-0.2351174653,-0.3012833893,0.3619758785,-0.019668784,-0.0797758698,0.1490667313,0.2029022425,-0.042098213,-0.1645146012,-0.0364278592,-0.8044582009,-0.3288797438,0.1570771933,0.2175286114,0.0154114179,0.4426355064,0.3873107731,0.0452336743,-0.0348101072,0.1303268522,-0.1764192879,-0.3531501889,0.6498783231,-0.3994504213,-0.2847266793,0.1182230935,0.2317685336,0.2142133862,-0.3526259363,-0.4864182472,0.0804351568,0.0773071796,0.2877722383,-0.0448323004,0.1416716278,-0.0867542922,-0.0689283088,-0.0460393466,-0.1349615753,-0.0135841006,-0.116391331,-0.3259768188,0.2562913597,0.0006416994,0.1128346696,-0.2048086971,0.5850800276,0.1846069694,0.1297905594,0.3108214438,-0.0963969231,0.686024785,-0.1007470489,-0.4953562319,-0.0185463745,-0.0426108614,0.0646766201,0.1125227883,0.302135855,-0.1705505848,-0.4108614326,0.052665893,-0.6017850637,-0.0814373419,0.1122563258,-0.1788872331,0.0257286951,0.0925906599,-0.1898592263,-0.3393605649,-0.1695629656,0.0055447565,0.5575093627,-0.0501216576,0.3014118075,-0.5939533114,-0.2490032464,-0.1569108814,0.1927903295,-0.1274876893,0.2557533979,-0.3264309466,0.0863464475,0.1326007843,-0.0135395527,0.4668221474,-0.0073006665,-0.0018592202,0.2152351737,0.21335949,-0.6372420788,0.1366122514,0.1852865368,0.1388081014,0.4634910524,0.3116728663,-0.1862471849,-0.0533104017,-0.0528364554,0.3992613852,-0.099411957,-0.1412382126,-0.4389512837,-0.5290256739,-0.3330169022,-0.3466005623,0.041629184,0.3000886738,-0.1127192676,0.1479955167,-0.0392461792,0.08828035,0.1973963529,0.215036571,0.3390181661,0.2029387206,-0.1517691016,0.2156927139,0.207391575,0.186458379,0.8277205825,-0.0280642305,-0.472987473,-0.0676505715,-0.1684446484,0.2539019883,0.2572754323,-0.0525502823,-0.264387399,0.0862524435,-0.0616413429,-0.1060521901,0.2844638824,-0.0042873053,0.1814712882,-0.1484197974,-0.3832840323,0.4944527149,-0.0522756279,-0.0187165849,0.1487674266,0.1603042036,-0.1453430653,0.2155505419,-0.2960323989,1.1821036339,0.0836563557,0.1619860977,0.1922802627,-0.1424458325,0.3413336873,-0.2491705269,-0.0668278858,-0.0165302716,0.1841173321,-0.0074367477,-0.150554657,0.2806473672,0.1858281046,0.0204978194,0.0036976642,-0.0962960869,0.0299856495,0.0989257619,0.4379950464,0.1277776212,-0.0180496313,-0.5319270492,0.1279685497,-0.0070529534,0.6483965516,-0.0589642003,-0.1501560062,-0.6219805479,0.0417346805,-0.1697573066,0.0559509322,-0.1763844192,0.3046876788,-0.0242759045,-0.4557029903,-0.131096974,0.3823457658,0.4158375561,0.0628483519,-0.2100280225,0.1942519248,-0.0006204766,-0.2378745824,0.1408261955,-0.2639215887,0.0989748389,-0.1657351702,-0.3407813311,0.2149266452,-0.0537951812,-0.3472219706,0.1409178823,0.0401334427,0.2564260364,0.0465484746,-0.4316743016,0.0652011633,0.2491727322,-0.1211000159,0.0465644635,0.0329136997,0.0470805205,-0.20226264,-0.223509267,-0.0786677599,-0.0912784413,0.4528014958,-0.2509536445,-0.1783564538,0.3761715591,0.1313108206,-0.2986024618,-0.094971858,0.1794088036,-0.0717810094,-0.4908860326,-0.3483213782,0.0258632731,-0.0363163538,-0.3221884966,0.1668507606,0.2664426863,-0.2686613202,0.1378796399,-0.7992363572,-0.009050753,0.2148845643,0.0030076047,0.0240182113,0.149524793,-0.1888760328,0.1773919314,-0.1492820978,-0.2086732835,0.1037804931,-0.2042625546,-0.0708857626,0.3645169139,0.0797675326,0.1445762366,-0.2587434351,0.0331773981,-0.3906826675,0.0711602569,-0.1381139606,-0.0232484508,0.167788595,-0.030742228,-0.0736366659,-0.1844884306,-0.423468709,0.0495093353,-0.0976031423,-0.1504056454,-0.0286711175,0.0516630039,-0.1411792636,-0.0938978791,0.3486450016,-0.0014615451,0.2108521163,0.0660802647,0.5716895461,0.0365405045,0.0253206547,0.1066599116,0.0104909716,-0.052856829,0.1187435761,-0.1242643297,0.1727949828,0.3727681637,-0.3196749091,0.3249733746,0.120964922,-0.0035803635,0.5077812672,-0.441368252,0.1462892592,0.3029463887,0.0475829877,-0.2808189392,-0.1859784126,0.4009407163,0.0555047877,-0.0616027452,0.2878829837,0.2537675798,-0.1890611053,0.389048636,0.1293126196,0.4627902508,0.4501369894,0.2029798329,0.0939915404,0.0795385763,0.0393343084,-0.0704107285,0.320171982,-0.0320692174,0.4540109932,-0.261349231,0.1091690212,0.0425857119,0.1719930172,-0.1905469745,-0.2267156392,0.1692724973,-0.1652827412,0.0399112813,0.0455428585,-0.1602568477,0.4140933454,-0.4061875045,-0.1118158549,0.0197634734,0.2420539558,0.0456830412,-0.0484396964,-0.2243011445,-0.2722094655,-0.1357747316,-0.184592098,0.1357543319,-0.5232682228,0.2045866102,0.1481912583,-0.1573607177,-0.2957580388,-0.0227050185,0.268961817,0.0716758817,-0.1148370355,0.326603651,0.2135397494,0.0652032346,0.1118482724,0.2881093621,0.4642577171,0.2599305511,-0.1109033898,-0.0053320415,-0.0054772645,-0.1516873986,-0.2587798536,0.2607496977,0.0704799891,0.0574535392,0.3448998928,0.166607663,-0.0934453532,-0.1094970554,0.2727870345,0.0146258213,-0.1755820364,0.1559967399,-0.4784351885,-0.0105871502,-0.3096095324,0.042605143,-0.3361676335,-0.0115367193,0.2119975388,0.0244800057,0.1457011402,-0.2048273087,0.0240990482,0.0394072495,0.4086048901,0.3885657787,0.2942712009,-0.2616147101,-0.5872488022,-0.8084726334,0.1373805106,-0.2018896788,-0.4872536361,0.0315862224,-0.1072384417,0.0769952685,0.2530308664,0.0159672648,0.3099045753,-0.0801284462,-0.1084850878,-0.2610019445,-0.2317061871,-0.2546134293,0.0900814682,-0.1903463602,-0.1421096772,0.1901785582,-0.5949850082,0.0008110312,0.163948983,-0.1057656333,-0.2082068324,-0.0478758253,0.5655726194,-0.1149526983,0.2684367895,0.0479270406,-0.2305282652,-0.1125653535,-0.252161473,0.1632338315,0.6077491045,0.0070399325,0.2551538348,-0.4812214077,0.0169900656,-0.0885087028,-0.026375182,0.2812675536,-0.1549213678,0.1070497036,-0.0785577297,-0.1062413678,0.2599004805,0.2936939001,0.3048795164,-0.2413358092,0.176688835,-0.1652900726,-0.1124801785,0.3234544694,-0.3052375019,-0.261243552,0.1704805791,0.1355917305,0.0898253173,0.1752025783,-0.2406037897,0.1892071515,0.3313728273,0.1525108516,-0.1949869394,0.2746852636,-0.3622052372,-0.1652604491,-0.0452168286,0.133011803,0.0323434584,0.0108250966,0.0206802953,-0.0255035013]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1674","title":"dutch_social can't be loaded","comments":"The issue is different for this one, let me give more details in the issue","body":"Hi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```","comment_length":15,"text":"dutch_social can't be loaded\nHi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```\nThe issue is different for this one, let me give more details in the issue","embeddings":[-0.1440130919,-0.1277865618,-0.1560761034,0.2933968306,0.2167993635,-0.1348885447,-0.0120579246,0.0937409475,0.3958035707,-0.0862966329,-0.2941347361,-0.0078357765,0.0196760818,-0.0421471745,0.1373223066,0.1136013195,0.2246757895,0.0936486647,0.144613862,-0.1379558891,-0.2554816604,0.4084303081,0.0364790969,0.1195254549,0.0535758808,0.2432906479,-0.0837842226,0.3191888928,-0.0633874536,-0.2948622704,0.3469977081,0.1354592592,0.1217035055,0.5947429538,-0.0001220059,0.0153691275,0.4793719053,-0.0934169069,-0.6241431832,-0.4102129638,0.0205617975,-0.2827773392,0.0646422952,-0.0720438063,-0.0622491986,-0.2690680027,-0.0105327405,-0.474494487,0.2986169159,0.4165929258,0.1114755422,-0.0108238775,0.0761229396,-0.0746558085,0.0295876656,0.3834807873,0.0695490241,0.5175902247,0.2559808195,0.0191765875,0.4757993519,0.139318645,-0.1636501998,-0.1841302067,0.1682499945,-0.1772353053,-0.0807946846,-0.2221838981,0.1476258039,-0.074964039,0.4582820535,-0.0886856616,-0.264942199,-0.2424675226,0.0461236127,0.0547623821,0.276529938,0.1222155318,-0.3778547943,0.2942798138,-0.1758255213,-0.2210489362,-0.1348641068,0.542472899,0.0668623,0.0545753092,-0.1998028606,0.0972559825,0.2403409779,-0.1409109533,-0.0025449127,-0.1723716855,0.1879131049,0.2741535306,0.0476628169,0.275170058,0.109135069,0.1229047924,0.0124698281,0.2088605762,-0.0083227707,0.0611036904,-0.2934058607,0.2661069036,0.1451622844,-0.063431032,0.1614208817,-0.0578602739,0.1387722641,0.2737375498,-0.2378363311,-0.0220349152,-0.5131173134,-0.2261775881,-0.4047652185,-0.1627119333,0.0526988059,-0.1887488216,-0.2392709255,-0.0537533425,0.2383565456,0.1124858856,0.1500710696,0.4265648723,-0.0568392836,0.1309242398,0.0157592501,0.3348211646,-0.0972028598,0.1625242531,0.0043877871,0.0115900803,-0.2395053059,-0.1033063531,0.3564893603,-0.1404215097,0.3913652897,-0.0133134797,0.0193966776,-0.2262601852,-0.0651753247,-0.1723169386,-0.0208002795,0.4110073447,-0.0504028127,0.1848825663,0.2301233411,-0.3826444447,-0.2856127322,0.0220117606,-0.4266529381,-0.2579602599,-0.1370801181,0.0865042135,-0.3576443791,-0.0483067632,-0.2518779337,0.0720809549,-0.2459039837,-0.3608632684,-0.1423634887,-0.0626245961,0.0569238551,-0.1635662615,0.4047258198,0.6451685429,-0.259563148,-0.1338703334,-0.2519385815,-0.0979309827,0.1166802645,0.1549301893,-0.2774693072,0.1003124788,-0.4217948914,-0.1295697242,0.5071360469,-0.3881822526,-0.1034957767,0.043269366,0.0442071557,0.4348086715,0.0677174553,-0.0021925054,-0.2236290574,0.0936073661,0.1177948564,0.365460515,0.2958675623,-0.3151164651,-0.0436034799,-0.08808887,0.2152402699,0.4358336627,-0.0789023265,-0.065279454,0.3209886551,-0.0179740302,0.0676686615,-0.0443313085,-0.1122181565,0.3995007277,-0.1002356708,0.6040776372,-0.1201559752,0.0579323918,-0.1615790576,0.1620278955,0.3629038632,-0.0949904695,-0.2131394148,0.0145916874,-0.5914059281,0.1086386964,-0.2440940589,-0.0796029046,0.0442409627,0.3385514021,0.2097835839,0.4188049734,-0.1303366423,0.5874919891,-0.1820048988,-0.0004448143,-0.136841476,0.3376156092,-0.077340886,0.0069129346,-0.025883032,0.4640442133,0.19163315,-0.0753638968,0.1306604892,0.1775824875,0.2757100463,0.1431743801,0.3152182102,-0.1459710449,0.013373767,-0.1851287335,0.0264447127,0.2751175463,0.1668300778,-0.0880569145,-0.2502496839,0.1557591856,-0.0062821158,0.2842067778,0.1452289671,0.0032311126,0.2915836871,-0.0235694479,-0.0459581092,-0.1600633413,0.3323235512,-0.1702431589,0.5120025873,-0.006221151,-0.2748087049,0.1538019031,0.0134518212,-0.1130990013,0.0663182139,0.2667791247,-0.1694904119,0.1820977181,0.1649412066,0.2306359261,0.5306184292,0.1501768082,0.0584150627,-0.1042482406,0.1984350532,-0.226388678,0.196112588,0.0779637173,0.3595768809,0.1046478078,0.1316241622,0.1227424741,-0.2572664917,-0.5361442566,-0.149571985,0.0522079468,-0.2917066514,0.3217304945,-0.185494706,-0.4013975859,-0.3152742684,-0.1355844587,-0.2492753416,-0.3631968498,-0.1829154044,0.1399168819,0.0379369855,-0.0491964184,-0.1167751774,0.1715294868,-0.2130713314,-0.3847946823,-0.3287187815,-0.0870077685,-0.049715437,-0.0937781259,0.4112957716,0.100119248,0.1564811319,-0.4932972789,0.1215285957,0.094456993,-0.4239718616,-0.1459390521,-0.0658269599,0.1950836778,-0.0813921466,0.3485629559,-0.2351174653,-0.3012833893,0.3619758785,-0.019668784,-0.0797758698,0.1490667313,0.2029022425,-0.042098213,-0.1645146012,-0.0364278592,-0.8044582009,-0.3288797438,0.1570771933,0.2175286114,0.0154114179,0.4426355064,0.3873107731,0.0452336743,-0.0348101072,0.1303268522,-0.1764192879,-0.3531501889,0.6498783231,-0.3994504213,-0.2847266793,0.1182230935,0.2317685336,0.2142133862,-0.3526259363,-0.4864182472,0.0804351568,0.0773071796,0.2877722383,-0.0448323004,0.1416716278,-0.0867542922,-0.0689283088,-0.0460393466,-0.1349615753,-0.0135841006,-0.116391331,-0.3259768188,0.2562913597,0.0006416994,0.1128346696,-0.2048086971,0.5850800276,0.1846069694,0.1297905594,0.3108214438,-0.0963969231,0.686024785,-0.1007470489,-0.4953562319,-0.0185463745,-0.0426108614,0.0646766201,0.1125227883,0.302135855,-0.1705505848,-0.4108614326,0.052665893,-0.6017850637,-0.0814373419,0.1122563258,-0.1788872331,0.0257286951,0.0925906599,-0.1898592263,-0.3393605649,-0.1695629656,0.0055447565,0.5575093627,-0.0501216576,0.3014118075,-0.5939533114,-0.2490032464,-0.1569108814,0.1927903295,-0.1274876893,0.2557533979,-0.3264309466,0.0863464475,0.1326007843,-0.0135395527,0.4668221474,-0.0073006665,-0.0018592202,0.2152351737,0.21335949,-0.6372420788,0.1366122514,0.1852865368,0.1388081014,0.4634910524,0.3116728663,-0.1862471849,-0.0533104017,-0.0528364554,0.3992613852,-0.099411957,-0.1412382126,-0.4389512837,-0.5290256739,-0.3330169022,-0.3466005623,0.041629184,0.3000886738,-0.1127192676,0.1479955167,-0.0392461792,0.08828035,0.1973963529,0.215036571,0.3390181661,0.2029387206,-0.1517691016,0.2156927139,0.207391575,0.186458379,0.8277205825,-0.0280642305,-0.472987473,-0.0676505715,-0.1684446484,0.2539019883,0.2572754323,-0.0525502823,-0.264387399,0.0862524435,-0.0616413429,-0.1060521901,0.2844638824,-0.0042873053,0.1814712882,-0.1484197974,-0.3832840323,0.4944527149,-0.0522756279,-0.0187165849,0.1487674266,0.1603042036,-0.1453430653,0.2155505419,-0.2960323989,1.1821036339,0.0836563557,0.1619860977,0.1922802627,-0.1424458325,0.3413336873,-0.2491705269,-0.0668278858,-0.0165302716,0.1841173321,-0.0074367477,-0.150554657,0.2806473672,0.1858281046,0.0204978194,0.0036976642,-0.0962960869,0.0299856495,0.0989257619,0.4379950464,0.1277776212,-0.0180496313,-0.5319270492,0.1279685497,-0.0070529534,0.6483965516,-0.0589642003,-0.1501560062,-0.6219805479,0.0417346805,-0.1697573066,0.0559509322,-0.1763844192,0.3046876788,-0.0242759045,-0.4557029903,-0.131096974,0.3823457658,0.4158375561,0.0628483519,-0.2100280225,0.1942519248,-0.0006204766,-0.2378745824,0.1408261955,-0.2639215887,0.0989748389,-0.1657351702,-0.3407813311,0.2149266452,-0.0537951812,-0.3472219706,0.1409178823,0.0401334427,0.2564260364,0.0465484746,-0.4316743016,0.0652011633,0.2491727322,-0.1211000159,0.0465644635,0.0329136997,0.0470805205,-0.20226264,-0.223509267,-0.0786677599,-0.0912784413,0.4528014958,-0.2509536445,-0.1783564538,0.3761715591,0.1313108206,-0.2986024618,-0.094971858,0.1794088036,-0.0717810094,-0.4908860326,-0.3483213782,0.0258632731,-0.0363163538,-0.3221884966,0.1668507606,0.2664426863,-0.2686613202,0.1378796399,-0.7992363572,-0.009050753,0.2148845643,0.0030076047,0.0240182113,0.149524793,-0.1888760328,0.1773919314,-0.1492820978,-0.2086732835,0.1037804931,-0.2042625546,-0.0708857626,0.3645169139,0.0797675326,0.1445762366,-0.2587434351,0.0331773981,-0.3906826675,0.0711602569,-0.1381139606,-0.0232484508,0.167788595,-0.030742228,-0.0736366659,-0.1844884306,-0.423468709,0.0495093353,-0.0976031423,-0.1504056454,-0.0286711175,0.0516630039,-0.1411792636,-0.0938978791,0.3486450016,-0.0014615451,0.2108521163,0.0660802647,0.5716895461,0.0365405045,0.0253206547,0.1066599116,0.0104909716,-0.052856829,0.1187435761,-0.1242643297,0.1727949828,0.3727681637,-0.3196749091,0.3249733746,0.120964922,-0.0035803635,0.5077812672,-0.441368252,0.1462892592,0.3029463887,0.0475829877,-0.2808189392,-0.1859784126,0.4009407163,0.0555047877,-0.0616027452,0.2878829837,0.2537675798,-0.1890611053,0.389048636,0.1293126196,0.4627902508,0.4501369894,0.2029798329,0.0939915404,0.0795385763,0.0393343084,-0.0704107285,0.320171982,-0.0320692174,0.4540109932,-0.261349231,0.1091690212,0.0425857119,0.1719930172,-0.1905469745,-0.2267156392,0.1692724973,-0.1652827412,0.0399112813,0.0455428585,-0.1602568477,0.4140933454,-0.4061875045,-0.1118158549,0.0197634734,0.2420539558,0.0456830412,-0.0484396964,-0.2243011445,-0.2722094655,-0.1357747316,-0.184592098,0.1357543319,-0.5232682228,0.2045866102,0.1481912583,-0.1573607177,-0.2957580388,-0.0227050185,0.268961817,0.0716758817,-0.1148370355,0.326603651,0.2135397494,0.0652032346,0.1118482724,0.2881093621,0.4642577171,0.2599305511,-0.1109033898,-0.0053320415,-0.0054772645,-0.1516873986,-0.2587798536,0.2607496977,0.0704799891,0.0574535392,0.3448998928,0.166607663,-0.0934453532,-0.1094970554,0.2727870345,0.0146258213,-0.1755820364,0.1559967399,-0.4784351885,-0.0105871502,-0.3096095324,0.042605143,-0.3361676335,-0.0115367193,0.2119975388,0.0244800057,0.1457011402,-0.2048273087,0.0240990482,0.0394072495,0.4086048901,0.3885657787,0.2942712009,-0.2616147101,-0.5872488022,-0.8084726334,0.1373805106,-0.2018896788,-0.4872536361,0.0315862224,-0.1072384417,0.0769952685,0.2530308664,0.0159672648,0.3099045753,-0.0801284462,-0.1084850878,-0.2610019445,-0.2317061871,-0.2546134293,0.0900814682,-0.1903463602,-0.1421096772,0.1901785582,-0.5949850082,0.0008110312,0.163948983,-0.1057656333,-0.2082068324,-0.0478758253,0.5655726194,-0.1149526983,0.2684367895,0.0479270406,-0.2305282652,-0.1125653535,-0.252161473,0.1632338315,0.6077491045,0.0070399325,0.2551538348,-0.4812214077,0.0169900656,-0.0885087028,-0.026375182,0.2812675536,-0.1549213678,0.1070497036,-0.0785577297,-0.1062413678,0.2599004805,0.2936939001,0.3048795164,-0.2413358092,0.176688835,-0.1652900726,-0.1124801785,0.3234544694,-0.3052375019,-0.261243552,0.1704805791,0.1355917305,0.0898253173,0.1752025783,-0.2406037897,0.1892071515,0.3313728273,0.1525108516,-0.1949869394,0.2746852636,-0.3622052372,-0.1652604491,-0.0452168286,0.133011803,0.0323434584,0.0108250966,0.0206802953,-0.0255035013]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1674","title":"dutch_social can't be loaded","comments":"Okay. Could you comment on the #1673 thread? Actually @thomwolf had commented that if i use datasets library from source, it would allow me to download the Hindi Wikipedia Dataset but even the version 1.1.3 gave me the same issue. The details are there in the issue #1673 thread.","body":"Hi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```","comment_length":49,"text":"dutch_social can't be loaded\nHi all,\r\n\r\nI'm trying to import the `dutch_social` dataset described [here](https:\/\/huggingface.co\/datasets\/dutch_social).\r\n\r\nHowever, the code that should load the data doesn't seem to be working, in particular because the corresponding files can't be found at the provided links.\r\n\r\n```\r\n(base) Koens-MacBook-Pro:~ koenvandenberge$ python\r\nPython 3.7.4 (default, Aug 13 2019, 15:17:50) \r\n[Clang 4.0.1 (tags\/RELEASE_401\/final)] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from datasets import load_dataset\r\ndataset = load_dataset(\r\n 'dutch_social')\r\n>>> dataset = load_dataset(\r\n... 'dutch_social')\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 2, in <module>\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/koenvandenberge\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at dutch_social\/dutch_social.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/dutch_social\/dutch_social.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/dutch_social\/dutch_social.py\r\n```\nOkay. Could you comment on the #1673 thread? Actually @thomwolf had commented that if i use datasets library from source, it would allow me to download the Hindi Wikipedia Dataset but even the version 1.1.3 gave me the same issue. The details are there in the issue #1673 thread.","embeddings":[-0.1440130919,-0.1277865618,-0.1560761034,0.2933968306,0.2167993635,-0.1348885447,-0.0120579246,0.0937409475,0.3958035707,-0.0862966329,-0.2941347361,-0.0078357765,0.0196760818,-0.0421471745,0.1373223066,0.1136013195,0.2246757895,0.0936486647,0.144613862,-0.1379558891,-0.2554816604,0.4084303081,0.0364790969,0.1195254549,0.0535758808,0.2432906479,-0.0837842226,0.3191888928,-0.0633874536,-0.2948622704,0.3469977081,0.1354592592,0.1217035055,0.5947429538,-0.0001220059,0.0153691275,0.4793719053,-0.0934169069,-0.6241431832,-0.4102129638,0.0205617975,-0.2827773392,0.0646422952,-0.0720438063,-0.0622491986,-0.2690680027,-0.0105327405,-0.474494487,0.2986169159,0.4165929258,0.1114755422,-0.0108238775,0.0761229396,-0.0746558085,0.0295876656,0.3834807873,0.0695490241,0.5175902247,0.2559808195,0.0191765875,0.4757993519,0.139318645,-0.1636501998,-0.1841302067,0.1682499945,-0.1772353053,-0.0807946846,-0.2221838981,0.1476258039,-0.074964039,0.4582820535,-0.0886856616,-0.264942199,-0.2424675226,0.0461236127,0.0547623821,0.276529938,0.1222155318,-0.3778547943,0.2942798138,-0.1758255213,-0.2210489362,-0.1348641068,0.542472899,0.0668623,0.0545753092,-0.1998028606,0.0972559825,0.2403409779,-0.1409109533,-0.0025449127,-0.1723716855,0.1879131049,0.2741535306,0.0476628169,0.275170058,0.109135069,0.1229047924,0.0124698281,0.2088605762,-0.0083227707,0.0611036904,-0.2934058607,0.2661069036,0.1451622844,-0.063431032,0.1614208817,-0.0578602739,0.1387722641,0.2737375498,-0.2378363311,-0.0220349152,-0.5131173134,-0.2261775881,-0.4047652185,-0.1627119333,0.0526988059,-0.1887488216,-0.2392709255,-0.0537533425,0.2383565456,0.1124858856,0.1500710696,0.4265648723,-0.0568392836,0.1309242398,0.0157592501,0.3348211646,-0.0972028598,0.1625242531,0.0043877871,0.0115900803,-0.2395053059,-0.1033063531,0.3564893603,-0.1404215097,0.3913652897,-0.0133134797,0.0193966776,-0.2262601852,-0.0651753247,-0.1723169386,-0.0208002795,0.4110073447,-0.0504028127,0.1848825663,0.2301233411,-0.3826444447,-0.2856127322,0.0220117606,-0.4266529381,-0.2579602599,-0.1370801181,0.0865042135,-0.3576443791,-0.0483067632,-0.2518779337,0.0720809549,-0.2459039837,-0.3608632684,-0.1423634887,-0.0626245961,0.0569238551,-0.1635662615,0.4047258198,0.6451685429,-0.259563148,-0.1338703334,-0.2519385815,-0.0979309827,0.1166802645,0.1549301893,-0.2774693072,0.1003124788,-0.4217948914,-0.1295697242,0.5071360469,-0.3881822526,-0.1034957767,0.043269366,0.0442071557,0.4348086715,0.0677174553,-0.0021925054,-0.2236290574,0.0936073661,0.1177948564,0.365460515,0.2958675623,-0.3151164651,-0.0436034799,-0.08808887,0.2152402699,0.4358336627,-0.0789023265,-0.065279454,0.3209886551,-0.0179740302,0.0676686615,-0.0443313085,-0.1122181565,0.3995007277,-0.1002356708,0.6040776372,-0.1201559752,0.0579323918,-0.1615790576,0.1620278955,0.3629038632,-0.0949904695,-0.2131394148,0.0145916874,-0.5914059281,0.1086386964,-0.2440940589,-0.0796029046,0.0442409627,0.3385514021,0.2097835839,0.4188049734,-0.1303366423,0.5874919891,-0.1820048988,-0.0004448143,-0.136841476,0.3376156092,-0.077340886,0.0069129346,-0.025883032,0.4640442133,0.19163315,-0.0753638968,0.1306604892,0.1775824875,0.2757100463,0.1431743801,0.3152182102,-0.1459710449,0.013373767,-0.1851287335,0.0264447127,0.2751175463,0.1668300778,-0.0880569145,-0.2502496839,0.1557591856,-0.0062821158,0.2842067778,0.1452289671,0.0032311126,0.2915836871,-0.0235694479,-0.0459581092,-0.1600633413,0.3323235512,-0.1702431589,0.5120025873,-0.006221151,-0.2748087049,0.1538019031,0.0134518212,-0.1130990013,0.0663182139,0.2667791247,-0.1694904119,0.1820977181,0.1649412066,0.2306359261,0.5306184292,0.1501768082,0.0584150627,-0.1042482406,0.1984350532,-0.226388678,0.196112588,0.0779637173,0.3595768809,0.1046478078,0.1316241622,0.1227424741,-0.2572664917,-0.5361442566,-0.149571985,0.0522079468,-0.2917066514,0.3217304945,-0.185494706,-0.4013975859,-0.3152742684,-0.1355844587,-0.2492753416,-0.3631968498,-0.1829154044,0.1399168819,0.0379369855,-0.0491964184,-0.1167751774,0.1715294868,-0.2130713314,-0.3847946823,-0.3287187815,-0.0870077685,-0.049715437,-0.0937781259,0.4112957716,0.100119248,0.1564811319,-0.4932972789,0.1215285957,0.094456993,-0.4239718616,-0.1459390521,-0.0658269599,0.1950836778,-0.0813921466,0.3485629559,-0.2351174653,-0.3012833893,0.3619758785,-0.019668784,-0.0797758698,0.1490667313,0.2029022425,-0.042098213,-0.1645146012,-0.0364278592,-0.8044582009,-0.3288797438,0.1570771933,0.2175286114,0.0154114179,0.4426355064,0.3873107731,0.0452336743,-0.0348101072,0.1303268522,-0.1764192879,-0.3531501889,0.6498783231,-0.3994504213,-0.2847266793,0.1182230935,0.2317685336,0.2142133862,-0.3526259363,-0.4864182472,0.0804351568,0.0773071796,0.2877722383,-0.0448323004,0.1416716278,-0.0867542922,-0.0689283088,-0.0460393466,-0.1349615753,-0.0135841006,-0.116391331,-0.3259768188,0.2562913597,0.0006416994,0.1128346696,-0.2048086971,0.5850800276,0.1846069694,0.1297905594,0.3108214438,-0.0963969231,0.686024785,-0.1007470489,-0.4953562319,-0.0185463745,-0.0426108614,0.0646766201,0.1125227883,0.302135855,-0.1705505848,-0.4108614326,0.052665893,-0.6017850637,-0.0814373419,0.1122563258,-0.1788872331,0.0257286951,0.0925906599,-0.1898592263,-0.3393605649,-0.1695629656,0.0055447565,0.5575093627,-0.0501216576,0.3014118075,-0.5939533114,-0.2490032464,-0.1569108814,0.1927903295,-0.1274876893,0.2557533979,-0.3264309466,0.0863464475,0.1326007843,-0.0135395527,0.4668221474,-0.0073006665,-0.0018592202,0.2152351737,0.21335949,-0.6372420788,0.1366122514,0.1852865368,0.1388081014,0.4634910524,0.3116728663,-0.1862471849,-0.0533104017,-0.0528364554,0.3992613852,-0.099411957,-0.1412382126,-0.4389512837,-0.5290256739,-0.3330169022,-0.3466005623,0.041629184,0.3000886738,-0.1127192676,0.1479955167,-0.0392461792,0.08828035,0.1973963529,0.215036571,0.3390181661,0.2029387206,-0.1517691016,0.2156927139,0.207391575,0.186458379,0.8277205825,-0.0280642305,-0.472987473,-0.0676505715,-0.1684446484,0.2539019883,0.2572754323,-0.0525502823,-0.264387399,0.0862524435,-0.0616413429,-0.1060521901,0.2844638824,-0.0042873053,0.1814712882,-0.1484197974,-0.3832840323,0.4944527149,-0.0522756279,-0.0187165849,0.1487674266,0.1603042036,-0.1453430653,0.2155505419,-0.2960323989,1.1821036339,0.0836563557,0.1619860977,0.1922802627,-0.1424458325,0.3413336873,-0.2491705269,-0.0668278858,-0.0165302716,0.1841173321,-0.0074367477,-0.150554657,0.2806473672,0.1858281046,0.0204978194,0.0036976642,-0.0962960869,0.0299856495,0.0989257619,0.4379950464,0.1277776212,-0.0180496313,-0.5319270492,0.1279685497,-0.0070529534,0.6483965516,-0.0589642003,-0.1501560062,-0.6219805479,0.0417346805,-0.1697573066,0.0559509322,-0.1763844192,0.3046876788,-0.0242759045,-0.4557029903,-0.131096974,0.3823457658,0.4158375561,0.0628483519,-0.2100280225,0.1942519248,-0.0006204766,-0.2378745824,0.1408261955,-0.2639215887,0.0989748389,-0.1657351702,-0.3407813311,0.2149266452,-0.0537951812,-0.3472219706,0.1409178823,0.0401334427,0.2564260364,0.0465484746,-0.4316743016,0.0652011633,0.2491727322,-0.1211000159,0.0465644635,0.0329136997,0.0470805205,-0.20226264,-0.223509267,-0.0786677599,-0.0912784413,0.4528014958,-0.2509536445,-0.1783564538,0.3761715591,0.1313108206,-0.2986024618,-0.094971858,0.1794088036,-0.0717810094,-0.4908860326,-0.3483213782,0.0258632731,-0.0363163538,-0.3221884966,0.1668507606,0.2664426863,-0.2686613202,0.1378796399,-0.7992363572,-0.009050753,0.2148845643,0.0030076047,0.0240182113,0.149524793,-0.1888760328,0.1773919314,-0.1492820978,-0.2086732835,0.1037804931,-0.2042625546,-0.0708857626,0.3645169139,0.0797675326,0.1445762366,-0.2587434351,0.0331773981,-0.3906826675,0.0711602569,-0.1381139606,-0.0232484508,0.167788595,-0.030742228,-0.0736366659,-0.1844884306,-0.423468709,0.0495093353,-0.0976031423,-0.1504056454,-0.0286711175,0.0516630039,-0.1411792636,-0.0938978791,0.3486450016,-0.0014615451,0.2108521163,0.0660802647,0.5716895461,0.0365405045,0.0253206547,0.1066599116,0.0104909716,-0.052856829,0.1187435761,-0.1242643297,0.1727949828,0.3727681637,-0.3196749091,0.3249733746,0.120964922,-0.0035803635,0.5077812672,-0.441368252,0.1462892592,0.3029463887,0.0475829877,-0.2808189392,-0.1859784126,0.4009407163,0.0555047877,-0.0616027452,0.2878829837,0.2537675798,-0.1890611053,0.389048636,0.1293126196,0.4627902508,0.4501369894,0.2029798329,0.0939915404,0.0795385763,0.0393343084,-0.0704107285,0.320171982,-0.0320692174,0.4540109932,-0.261349231,0.1091690212,0.0425857119,0.1719930172,-0.1905469745,-0.2267156392,0.1692724973,-0.1652827412,0.0399112813,0.0455428585,-0.1602568477,0.4140933454,-0.4061875045,-0.1118158549,0.0197634734,0.2420539558,0.0456830412,-0.0484396964,-0.2243011445,-0.2722094655,-0.1357747316,-0.184592098,0.1357543319,-0.5232682228,0.2045866102,0.1481912583,-0.1573607177,-0.2957580388,-0.0227050185,0.268961817,0.0716758817,-0.1148370355,0.326603651,0.2135397494,0.0652032346,0.1118482724,0.2881093621,0.4642577171,0.2599305511,-0.1109033898,-0.0053320415,-0.0054772645,-0.1516873986,-0.2587798536,0.2607496977,0.0704799891,0.0574535392,0.3448998928,0.166607663,-0.0934453532,-0.1094970554,0.2727870345,0.0146258213,-0.1755820364,0.1559967399,-0.4784351885,-0.0105871502,-0.3096095324,0.042605143,-0.3361676335,-0.0115367193,0.2119975388,0.0244800057,0.1457011402,-0.2048273087,0.0240990482,0.0394072495,0.4086048901,0.3885657787,0.2942712009,-0.2616147101,-0.5872488022,-0.8084726334,0.1373805106,-0.2018896788,-0.4872536361,0.0315862224,-0.1072384417,0.0769952685,0.2530308664,0.0159672648,0.3099045753,-0.0801284462,-0.1084850878,-0.2610019445,-0.2317061871,-0.2546134293,0.0900814682,-0.1903463602,-0.1421096772,0.1901785582,-0.5949850082,0.0008110312,0.163948983,-0.1057656333,-0.2082068324,-0.0478758253,0.5655726194,-0.1149526983,0.2684367895,0.0479270406,-0.2305282652,-0.1125653535,-0.252161473,0.1632338315,0.6077491045,0.0070399325,0.2551538348,-0.4812214077,0.0169900656,-0.0885087028,-0.026375182,0.2812675536,-0.1549213678,0.1070497036,-0.0785577297,-0.1062413678,0.2599004805,0.2936939001,0.3048795164,-0.2413358092,0.176688835,-0.1652900726,-0.1124801785,0.3234544694,-0.3052375019,-0.261243552,0.1704805791,0.1355917305,0.0898253173,0.1752025783,-0.2406037897,0.1892071515,0.3313728273,0.1525108516,-0.1949869394,0.2746852636,-0.3622052372,-0.1652604491,-0.0452168286,0.133011803,0.0323434584,0.0108250966,0.0206802953,-0.0255035013]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1673","title":"Unable to Download Hindi Wikipedia Dataset","comments":"Currently this dataset is only available when the library is installed from source since it was added after the last release.\r\n\r\nWe pin the dataset version with the library version so that people can have a reproducible dataset and processing when pinning the library.\r\n\r\nWe'll see if we can provide access to newer datasets with a warning that they are newer than your library version, that would help in cases like yours.","body":"I used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n","comment_length":72,"text":"Unable to Download Hindi Wikipedia Dataset\nI used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n\nCurrently this dataset is only available when the library is installed from source since it was added after the last release.\r\n\r\nWe pin the dataset version with the library version so that people can have a reproducible dataset and processing when pinning the library.\r\n\r\nWe'll see if we can provide access to newer datasets with a warning that they are newer than your library version, that would help in cases like yours.","embeddings":[-0.1846334487,0.0373013057,-0.0611194484,0.2292183638,0.053209424,0.1153836846,-0.0216608755,0.3467042446,0.2222079933,0.0246456787,0.398733139,0.074465245,0.0466891788,0.2172787488,0.1481214762,-0.3831970096,0.0711239949,-0.0032809081,0.0407369472,0.1277780086,0.0758774877,0.2772662044,-0.2721773088,-0.0085311243,-0.5351586938,-0.0147672929,0.2069775462,-0.1029816121,-0.3329057097,-0.3878406584,0.453076005,0.0746859983,0.3272311687,0.2894625962,-0.000117929,0.0840014741,0.5217901468,-0.1599767059,-0.3209836483,-0.4747384191,-0.1340344846,-0.3125717938,0.1033498496,-0.2150286734,-0.1281321496,0.2258340567,0.3669331372,-0.2974428236,0.1387929171,0.1242169738,0.1640716642,-0.1203647479,0.386944592,-0.1464688033,0.203081727,0.2014706582,-0.0552636422,0.2983807623,0.0545501336,0.1054257676,-0.0732383355,0.1809621155,0.0445848443,-0.1273255348,0.3103481233,-0.3043245375,-0.2426697612,-0.4269345701,0.4759139121,0.3865130842,1.1295317411,-0.3715818822,-0.1677476466,0.1515140682,-0.0448209159,0.2206408381,0.4101853073,0.4490496218,-0.2433068007,-0.0120517062,0.1708031297,-0.3701305687,-0.157833159,0.3645862639,0.0316441916,0.2004282922,0.0091386847,0.1854778826,-0.0307361204,-0.2145123035,-0.02501937,-0.0500654131,0.120045878,0.3106352389,-0.2569937706,0.1385988295,-0.127401948,0.1306022853,0.0987364575,-0.1533349901,-0.0414458886,-0.0029117034,-0.0950385034,-0.0041448404,0.3619229496,-0.0806551874,0.0078141429,0.2096252292,0.256455183,0.2672003806,-0.0139932446,-0.0607570074,0.0611205436,-0.4209552705,-0.5915756226,-0.0259812288,-0.0162632968,-0.2765037715,-0.1287475973,0.0044562402,-0.3457309008,-0.2075077295,0.0737193525,0.0334329158,0.0313108116,0.1312571019,0.0619303472,0.0491787419,-0.1501957923,-0.3245725036,-0.050696224,0.2413212657,-0.300229311,0.0361862145,0.2658056617,-0.2571329176,0.3741752803,0.0476975664,-0.1977427155,0.0010906622,-0.0669957474,-0.2411717176,-0.3081691265,0.197338298,0.1855812818,0.4436905682,-0.2013188899,-0.1000121832,-0.038498003,0.1480777115,-0.2725658119,-0.012922259,0.0705824345,0.074526526,-0.2666237652,-0.1187896803,0.0480042249,0.2205324024,-0.0582599565,-0.4137549102,0.009203448,-0.0192557257,-0.1150910109,-0.3171354234,0.3101770878,0.4562660754,-0.7055321336,0.0829726681,-0.1259951144,-0.0040928465,0.0826830119,-0.0513191372,-0.2339766324,0.5790066719,-0.2916668653,0.0442174561,0.5752339363,-0.2936263978,-0.5257562399,-0.0133434022,-0.02399645,0.057880789,0.0556753725,0.2580733299,0.1837553829,0.006373134,-0.1392722875,0.3432026803,0.097191155,-0.1311730295,-0.2822996378,-0.0902168378,0.3322113454,0.153289929,0.2401341796,-0.1168657541,0.1273595244,0.4729164839,0.3813784719,0.0354138464,0.1606386602,0.3658227324,-0.2541016042,0.09451738,0.2299745679,-0.088264592,-0.0512497053,0.2550355494,-0.0206331778,-0.071366027,-0.1211463735,0.0199509189,-0.403144598,0.0117723644,-0.2332554311,0.1814034879,0.0689975396,0.0132382959,0.1103869081,0.3523840308,-0.0559859313,0.1727205366,-0.2753531933,0.030645363,-0.3404828608,0.4449898899,-0.1094449088,-0.0813350677,0.0301134735,0.1406916529,0.1717761159,-0.0056818076,-0.151619792,0.0053042262,0.1093395278,0.0264145229,0.2505449057,-0.0275604129,0.1490451247,-0.5346199274,0.2565496862,0.5513415933,0.0377882309,-0.0155817978,0.0395439006,0.0551243499,-0.2308329344,0.0603769757,-0.069590807,0.1959153116,0.3569629788,0.0926788896,0.0055166283,0.0128296521,0.2818467319,0.4182784557,0.2217084467,-0.0167972948,-0.0961955115,0.0408422239,0.57889992,0.0564463139,0.0157234799,0.1267824769,-0.2564577162,-0.0707025006,0.1073689535,0.1339578629,0.038056232,0.1063374206,0.0162664559,-0.207527101,0.0751836598,-0.1791910976,0.2018443793,0.1663945168,0.2951563597,0.1557773203,-0.1634331346,0.0725022703,-0.1826339811,-0.0771216527,0.0450495668,0.3206934333,-0.1707876623,-0.050545305,-0.3483173251,-0.518684566,-0.3468508124,-0.306591779,-0.486923337,-0.2183832675,-0.0752952248,-0.1249911115,0.0173766036,0.0453165025,-0.0387022905,-0.1222059354,0.0449226983,-0.3332616687,-0.0671157539,-0.3668973446,-0.3348782063,-0.0264077224,0.5341210365,0.0615052544,0.2562438548,-0.2601419389,-0.0517231934,-0.6239216328,-0.229999125,0.0948204845,-0.046230305,-0.1074339077,0.073105216,0.6945303679,-0.1953697503,0.0020850198,0.0359014384,0.2028860152,-0.0650187805,-0.1063526422,-0.0644028634,0.0037223068,0.1378206462,-0.5049888492,-0.3756238818,-0.2445554584,-0.0648974106,0.1017139927,0.056053523,-0.0199140925,0.0931154191,0.0310917161,0.267934233,0.151895389,-0.0283386596,-0.1419942528,0.347107619,-0.1432011425,-0.5207159519,0.5949020982,0.0072855256,0.1874605864,0.0847608,-0.3583533168,0.3445145488,0.0796519369,0.2515495121,0.1710092872,-0.0006898724,0.2284472585,-0.2805490494,0.2448115349,0.0850279555,-0.1836838275,0.1268242598,-0.3386278152,0.4829137921,0.3686639667,0.2770755887,-0.0187711753,0.6054495573,0.3167237937,0.0584708042,0.3383820057,0.1432934701,0.0349115692,-0.2171928585,-0.2373215407,0.1322418898,-0.0000805427,0.0751705319,0.1479733139,0.0855295807,-0.1966651976,-0.2690395713,-0.3766618371,-0.2506541312,-0.3482201993,-0.0689137504,-0.2725534439,0.3399359286,0.1153769046,0.0445707105,0.0012760905,-0.2917280197,0.1944931448,0.4144558609,-0.0743595436,0.1778516769,0.0403146446,-0.4357806742,-0.2594370544,0.0352744237,0.0874824226,0.3519076109,-0.1406172067,0.2277020067,0.2488122135,0.0583522543,0.5387908816,-0.2299266756,0.1355606467,0.1591628194,0.0505382046,-0.2174153179,-0.163083151,-0.006016246,0.3007087708,0.2635967731,-0.1130733788,-0.4702248871,-0.1034749523,0.3776741028,0.3612948954,-0.00210738,-0.0958080143,-0.2353926301,-0.2012509406,-0.277040869,-0.2497873008,-0.1831016839,0.4537198246,0.1865486354,0.1727632284,0.0504775569,-0.0033689544,-0.1720174104,-0.0968926176,0.3721061051,0.2365717888,-0.0514718145,0.3459805548,0.2768327594,-0.2307365984,0.3108764887,0.2015568763,-0.1960392594,0.064004533,-0.0363841057,-0.0918103606,0.0926545858,-0.1090955287,0.0549148321,0.0343550108,-0.3928489089,0.0695433617,-0.140405938,0.1414228976,0.1021879837,-0.5075061321,-0.4818611741,0.5160732865,0.2066749632,0.1307382733,0.3050281107,0.1219189242,-0.2265904546,0.1187318712,0.0871231183,0.8414729238,-0.0588590167,0.1046173945,-0.0235368907,0.0664840266,0.4696100354,-0.1648102105,0.0213939045,-0.4580952227,0.0641556233,-0.1256150752,0.0395131484,0.0341214314,-0.1010839343,-0.2395250201,0.4299021959,0.1962433904,0.3532970548,0.1065807268,0.3824318349,0.0614488907,-0.1783649027,-0.200961262,0.0373246334,-0.1750662923,0.4360246062,-0.2624443173,-0.1721881032,-0.1308653057,-0.0097462554,-0.4735486805,0.111117065,-0.2853594124,0.3070903122,-0.1701979041,-0.4766334891,0.4796658754,0.374638021,0.3110159039,0.296115309,-0.3534682989,0.2413543612,-0.4690703154,-0.4751164019,0.0149061047,0.2051029056,0.3296316862,-0.1623258293,-0.3806777596,0.2140043527,-0.0429228619,-0.01550439,-0.0898473859,-0.0840382203,0.2972671688,0.0985151231,-0.353441447,-0.0484921001,-0.0322286673,-0.1068561599,0.1144239604,0.0624976493,0.1496981978,-0.0541652739,0.1990107596,-0.1092151254,0.0480472632,0.3589994609,0.0623320453,-0.1001624614,0.4084860981,0.3789877295,-0.4220362604,-0.1429107934,-0.0566173606,-0.3723242581,-0.256806165,-0.0918068811,0.0396639965,0.2331899256,-0.1711028069,0.2240111828,0.062479727,-0.2768573165,-0.0716160312,-0.4444007277,-0.2308716178,0.2884555757,-0.2411358654,0.0103185363,0.0273759067,-0.28836447,0.4284653366,-0.2648025155,-0.191591084,-0.0075471904,-0.1405308247,-0.0364875942,0.1321454346,0.1201081574,-0.1327278763,-0.0646248087,0.0431292281,-0.1425334066,-0.1169387251,-0.0816131383,0.2025775164,0.1139765233,-0.002746701,0.0553283468,0.0104171541,-0.5299056768,-0.1002087072,0.0279056747,-0.1141846627,0.0384867936,-0.0241202991,0.1766662747,-0.0177953094,0.0356094986,-0.4783553183,0.3255082965,0.032209456,0.2005625367,0.0932023376,0.3240570128,0.2393976897,-0.066435717,-0.4538112581,0.155031547,-0.0625684485,-0.0609817207,0.2337652147,-0.337271452,0.3416976929,0.2021016926,0.1301508099,0.0836523026,-0.1786543876,-0.0231602974,0.3648687303,0.0775286108,-0.461861074,0.0256229099,0.6571726203,0.1592990011,-0.0764518604,0.0052086739,0.3300807774,-0.2423384786,-0.0210619718,0.2571818233,0.2027352452,-0.0896665156,0.257897377,0.3942117095,0.3652914166,0.2093596607,0.0093701901,0.114569515,-0.1352539659,0.5476802588,-0.3880693913,0.1480910182,0.1981539726,-0.1265384108,-0.0837161541,-0.3219208419,0.1154315993,0.2119636685,-0.2054352462,-0.069517754,-0.0338217244,0.13775222,0.3002785146,-0.0574842393,-0.2226241827,0.0185337029,-0.236528933,0.2147398293,-0.4541546404,-0.3500849009,-0.053001754,0.148167789,0.1076802909,-0.1282634735,0.0066438536,0.222690478,-0.1914744675,-0.1691995114,0.7468152642,-0.1285256445,0.0256769434,-0.1639954597,0.0634751692,-0.2326100618,-0.0989580825,0.0470707305,0.23978962,0.1900180131,0.247612536,-0.2872261107,-0.0197458249,0.0469244607,-0.0128357215,0.1226417273,0.1487042904,0.0920745283,0.0069978833,0.2984030545,0.0865888223,-0.0603494421,-0.224329859,0.3149134517,0.2164528221,-0.0790289193,-0.040113885,-0.1905999184,0.0408140123,-0.2115936726,-0.0317706987,-0.5151205063,-0.1172866449,0.3442645371,0.1591210961,-0.0020007223,-0.3801005781,0.0205824599,-0.2063930184,0.3930348456,0.3170326948,0.1858787686,-0.3537291586,-0.3557287157,-0.5848137736,0.1577866971,-0.1198754907,-0.1324210465,0.0998257846,0.0973670185,-0.1109461561,0.1328789294,0.3184336722,0.0585601814,0.0231853351,0.3147509694,-0.1297197044,-0.1978247464,-0.1940149069,-0.0542207398,-0.3312115967,-0.0908293352,0.1407482028,-0.4636225104,-0.007864072,-0.0642924383,-0.1705764383,0.21816881,-0.2165547311,0.1140213907,-0.1979753822,0.4842721522,0.0799906999,-0.0016849003,-0.2617835999,0.1757388711,-0.0425585993,0.4966691732,-0.0828532428,0.3954638839,-0.1983480155,0.3092242181,-0.1897019744,0.383469522,-0.163951084,0.3447006047,-0.2370954901,-0.0953472406,-0.0656844005,0.3868778348,0.0679276437,0.1910696626,-0.18565166,0.0544342287,-0.362799108,-0.1660848856,0.2119629234,-0.4191954732,-0.184048593,-0.0631768182,0.0822247565,0.1713710278,0.1093432307,-0.2676993012,0.1202481762,0.1447652876,0.0203519929,-0.0533106402,0.0885918438,-0.131630078,0.1653034687,0.0196170565,0.0060189161,-0.1020064801,-0.2280569375,-0.1537579894,-0.3175723553]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1673","title":"Unable to Download Hindi Wikipedia Dataset","comments":"So for now, should i try and install the library from source and then try out the same piece of code? Will it work then, considering both the versions will match then?","body":"I used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n","comment_length":32,"text":"Unable to Download Hindi Wikipedia Dataset\nI used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n\nSo for now, should i try and install the library from source and then try out the same piece of code? Will it work then, considering both the versions will match then?","embeddings":[-0.1116185784,0.1364629418,-0.0870575085,0.2553951144,0.0008571927,0.0589530841,-0.0200850423,0.3475781381,0.1966427863,-0.031186102,0.4182320535,0.0535380915,0.0738414675,0.2070575953,0.1762293726,-0.3135467768,0.1087693498,0.1200553849,0.0042044101,0.0592677221,0.0683595464,0.2611467242,-0.2298688442,-0.0048639877,-0.4776678979,0.0012445815,0.2221390009,-0.0502518117,-0.2584412098,-0.3565151095,0.4109671116,0.0497172885,0.2961830199,0.3686352372,-0.0001199416,0.1368225962,0.4905959666,-0.1850735098,-0.2386745512,-0.4604792297,-0.0924596488,-0.3030233681,0.0296326261,-0.2857756913,-0.1497669965,0.1927249879,0.3124535382,-0.2782360315,0.1295937598,0.1415741444,0.170101583,-0.0960211977,0.3933209181,-0.0845570192,0.2663611174,0.1897556335,0.0195259675,0.3279036582,-0.0306914207,0.1860106587,-0.0879555792,0.2282851338,0.1306127608,-0.1248108596,0.2921603918,-0.2631483376,-0.2344123572,-0.3487629294,0.4009373188,0.2763460875,1.0225684643,-0.3250437081,-0.2268597484,0.1726696491,-0.0803990588,0.1519275308,0.4632243216,0.3294928074,-0.200940758,-0.0213521793,0.1633627117,-0.4233598709,-0.1424345672,0.3965321183,-0.0582766086,0.1761892885,-0.0224964004,0.1639204472,-0.0848843008,-0.2016354501,0.0008347647,-0.1162046045,0.1215322092,0.2226236612,-0.1990119219,0.1172073185,-0.1028320417,0.154689759,0.0547818169,-0.1945848018,0.0123956772,0.0211747456,-0.114752695,-0.0663920864,0.3393965662,-0.0875509679,-0.0246436354,0.2418071628,0.2941322327,0.2672395706,-0.0053095212,-0.0251796693,0.135896638,-0.4400852919,-0.5205779672,-0.0086791916,-0.010131767,-0.3164439797,-0.1114299819,0.0814468339,-0.2926316261,-0.2897930741,0.0461605452,0.0422914028,-0.0381762385,0.1850257218,0.0670786351,0.0251767281,-0.183930099,-0.2609624267,-0.0484742597,0.3083530366,-0.3088774681,0.0374192894,0.2097858936,-0.2037178576,0.4332460761,0.0159764159,-0.1486077309,0.0410353504,-0.0718801171,-0.2721830606,-0.2053116113,0.1476372182,0.1482718438,0.4279994667,-0.2541537285,-0.0450664088,-0.0453885384,0.16307652,-0.1861615479,0.0238913614,0.1941388696,0.0877622217,-0.3134622872,-0.0832369104,0.0934656262,0.1890525818,0.0000830017,-0.5406222939,-0.0220427122,-0.0361358151,-0.0432399698,-0.2813766003,0.3444929719,0.4325047433,-0.7107768059,0.131054014,-0.1130259186,0.0302892216,0.0470137857,-0.0309073608,-0.2615732849,0.6532417536,-0.3513746262,0.0233924296,0.6675368547,-0.3990435004,-0.4707461596,-0.0478582866,-0.0367799699,0.0960422829,0.0560680963,0.2257205248,0.166121304,-0.0386955701,-0.0958093107,0.2801710963,0.0846248269,-0.1513461024,-0.3334873617,-0.089302972,0.3929633796,0.1288661063,0.2982840538,-0.1237699762,0.0904327407,0.444034487,0.3604435027,-0.0058935569,0.0943313688,0.3124651909,-0.2774246037,0.0735908374,0.1572695673,-0.1025939211,-0.0308641419,0.1844977736,-0.0134440791,-0.1441207379,-0.0776043087,-0.0127051789,-0.4661594331,-0.0546104833,-0.2460792959,0.1904679686,0.0802894384,0.0531753562,0.1517300755,0.2807612121,-0.0819265917,0.1642741114,-0.2459920198,0.0311402213,-0.4025273025,0.5011726618,-0.1499479413,-0.0836176574,0.07487005,0.1212408692,0.184238866,-0.0470681638,-0.1886332482,0.0152834347,0.0330634303,0.0542501546,0.1806891263,-0.133483544,0.2101243138,-0.4950474501,0.2544282079,0.581487596,0.0968516245,-0.0262989048,0.0826093629,-0.0026080934,-0.2105977833,0.0477733761,0.0073690796,0.0598729588,0.3191693425,0.1081823781,-0.0205736812,-0.0164077785,0.2179352939,0.4814141393,0.2567302585,-0.0057961638,-0.1379146874,0.0355149694,0.6405736804,0.0470729433,-0.0323285684,0.0782957971,-0.1548196524,-0.0652377531,0.1192677319,0.1161776558,0.120758757,0.1098808125,-0.0378070809,-0.2099820524,0.0756020471,-0.1881776303,0.1930868626,0.2020437121,0.3380408883,0.2567362189,-0.1854952574,0.0773996189,-0.1787091643,-0.0802071989,0.0539321303,0.2617740631,-0.1984218061,-0.0188062768,-0.3552405238,-0.5050302148,-0.4395568073,-0.33516711,-0.4656643867,-0.1712058336,-0.0870599672,-0.1600824296,-0.0653273538,0.1061923206,-0.0200817306,-0.1783478409,-0.0367398299,-0.3817269802,-0.0138923321,-0.3240439296,-0.3724229336,-0.0291352049,0.5464525819,0.0917799473,0.2754091024,-0.3189601004,-0.0864539817,-0.6308422089,-0.2829709649,0.1234155968,-0.0257624649,-0.1141891479,0.1440612227,0.6267017722,-0.1173535436,-0.0635198429,0.0890231133,0.1732285023,-0.0416595303,-0.2011984438,-0.033210665,0.0631767586,0.1438006759,-0.5816696882,-0.3861139119,-0.2711100578,-0.1043009833,0.1433054358,0.0773236081,-0.1253488362,0.0453780778,0.0498803593,0.3385449648,0.1959766448,-0.0305155218,-0.0556683466,0.3645651937,-0.1890694946,-0.5188432336,0.5080653429,-0.0794422776,0.2510534823,0.1673784256,-0.3368943632,0.2751417458,0.0962032452,0.2145345211,0.179372564,0.1414342374,0.2075333595,-0.2547532618,0.2343722582,0.068531923,-0.1730420589,0.2150754035,-0.35712713,0.4659358859,0.4194562733,0.2468277365,-0.1778564304,0.6441901326,0.2383691072,-0.0121935122,0.3310514092,0.1444189847,0.1175789535,-0.191721186,-0.1829325855,0.0724566877,0.0049510603,0.0267769713,0.127186805,0.0087249493,-0.1607027799,-0.3184440434,-0.3897809684,-0.316933006,-0.4087817073,-0.0390388519,-0.2150545418,0.3358775973,0.1023711935,0.1053743288,-0.0104409093,-0.2369319946,0.232788831,0.3670194447,-0.0745427087,0.1514535546,0.0156741291,-0.4938832521,-0.2678689361,0.1095770001,0.1559029371,0.3675425947,-0.0375842638,0.1003234088,0.1462160945,-0.1329313368,0.5034664273,-0.2104183137,0.0751681104,0.1487030685,0.0313598737,-0.1780384779,-0.1782197505,-0.0204721075,0.3454664052,0.286937207,-0.1696615964,-0.5061261058,-0.1280063391,0.4084997773,0.4383280873,-0.0520065911,-0.0538254194,-0.2738624215,-0.2603932619,-0.2510298789,-0.2795884311,-0.2178498656,0.4130958915,0.2153591961,0.1571729034,0.0203839764,-0.0494049489,-0.1336654425,-0.0247951299,0.4649862051,0.1659041494,0.0092912894,0.3347705901,0.2640925944,-0.1937426329,0.3560106754,0.1706126779,-0.0900506228,0.0901788846,-0.0634825453,-0.0927116498,0.1813076735,-0.1247133985,0.0746713728,-0.0042833341,-0.4284429252,0.109696798,-0.1176905781,0.0552225076,0.1345377266,-0.4963887632,-0.5103209615,0.4413460493,0.2300394475,0.0885900706,0.285692215,0.108013913,-0.298810631,0.0704561099,0.0863484517,0.8700020909,0.0319475234,0.1575450152,0.0469298884,0.0209545996,0.494130671,-0.1662573963,0.0198419914,-0.4533278048,0.1052881852,-0.08929573,0.0262282044,0.0791772604,-0.1005076244,-0.2168672383,0.5290193558,0.2651736736,0.3622131646,0.0197631121,0.4233469069,0.0628685206,-0.1593638808,-0.2850795984,0.0130707165,-0.1367735267,0.4633344412,-0.3115536571,-0.1587217599,-0.2044164985,-0.1001026928,-0.3631284833,0.054539863,-0.2672945857,0.2633298337,-0.2275199592,-0.5686739087,0.4709669948,0.3436931074,0.3785256445,0.335740447,-0.2580091655,0.2218454033,-0.3946754038,-0.3821427226,0.0196000747,0.2535357177,0.2988277674,-0.2233186066,-0.3286412954,0.2567256093,-0.079725191,0.0227971822,-0.1368798018,-0.1016602367,0.2737199962,0.0324057713,-0.3632273674,0.0122348899,-0.0522090979,-0.078893058,0.1035783738,0.002720207,0.1011657938,-0.0409109071,0.1235653907,-0.0433087908,-0.0022660834,0.2945349813,0.1677813232,-0.1583174616,0.3771737218,0.265979141,-0.3943152428,-0.1178563684,-0.1152650714,-0.3469323516,-0.3014823794,-0.0758019462,0.0294291396,0.1562173218,-0.2124088407,0.2315443903,0.1200786307,-0.1702415347,0.0003157854,-0.3563271761,-0.2629660666,0.3253136873,-0.0238719564,0.0774775371,0.0476338156,-0.2397531122,0.31454584,-0.2913704813,-0.1907004118,-0.04249008,-0.1248415411,0.0411162637,0.0679339021,0.1169426367,-0.1829056889,-0.0374281146,0.043380551,-0.0657028109,-0.0960236117,-0.0634440556,0.2495626807,0.1393845677,-0.0221888553,0.0468064584,0.0602462366,-0.458016336,-0.0198058467,0.061682485,-0.1221447065,-0.038698405,-0.018839607,0.1796775013,-0.0458944105,-0.0174047761,-0.5574914217,0.3837867677,0.0576646291,0.2761916518,0.1330737919,0.3843609393,0.2052498013,-0.0646792725,-0.4044151604,0.0942887664,-0.1726572067,0.0191469099,0.1824294329,-0.3665362597,0.352892369,0.2252864242,0.1696192473,0.2171684206,-0.1902458221,-0.0315770321,0.3662296832,0.0816733614,-0.5075830817,0.0873115286,0.5958000422,-0.0074478774,-0.0206947364,0.0827271938,0.3130984604,-0.2065874189,-0.0258113258,0.2825277746,0.14409329,-0.018781852,0.260368526,0.3330653906,0.337862283,0.2519086599,-0.013018121,0.086908713,-0.1973246634,0.5674139261,-0.4624263942,0.207490027,0.1939550191,-0.0950284451,-0.1086816043,-0.3034704328,0.1681504846,0.2585104406,-0.1355085671,-0.0362329371,0.0352825224,0.051706735,0.2671379447,-0.1691539437,-0.1282800287,0.0409756899,-0.1965313703,0.2740449607,-0.4266390204,-0.3769052029,-0.08376351,0.1964244992,0.0770260245,-0.0948444977,0.0063408175,0.28209427,-0.1801148206,-0.1813605428,0.7708126307,-0.2583170831,0.0691535547,-0.136505723,0.1843926758,-0.1348950118,-0.1944842637,0.0975573659,0.2608771324,0.1480577588,0.2769491076,-0.2119622231,-0.0158364549,0.095279634,0.0245059133,0.1141698658,0.1354493201,0.1343690455,0.055015821,0.316403836,0.1058052853,-0.0945035145,-0.193210274,0.3063552082,0.1115049943,-0.1694963127,0.0868595466,-0.144230321,0.0057424507,-0.201288715,-0.0225918815,-0.5413014889,-0.1206553057,0.2794725001,0.1686347127,0.0083403783,-0.3549897373,0.0123478845,-0.2008835971,0.3789103925,0.2389475554,0.2652489841,-0.3975165784,-0.3675655723,-0.665909946,0.1524261385,-0.1094152331,-0.1661465466,0.0317767821,0.0933844596,-0.0817312375,0.1665581167,0.4052265584,0.0637402907,0.0003764066,0.3149206936,-0.1184576303,-0.1961195022,-0.2516298592,-0.0719270632,-0.3258891404,-0.1477111727,0.0725457072,-0.4315220118,-0.0068006581,-0.0493637882,-0.1983939856,0.2371880859,-0.2616608739,0.1034406647,-0.1572091281,0.495467484,0.0302815884,-0.0985058099,-0.1353366524,0.1375449151,-0.0723123774,0.5550069213,-0.0299398582,0.4324308038,-0.1629996896,0.3704671562,-0.248319298,0.3966754079,-0.104514122,0.3519795239,-0.1611779332,-0.0905516371,-0.0499827676,0.2990833521,0.1529003829,0.3317386508,-0.1846023798,0.1033961624,-0.447414875,-0.2045026124,0.2256164253,-0.4589468241,-0.1408321559,-0.107719332,0.1248961538,0.1781666428,0.0744521096,-0.3265519142,0.1241816357,0.124581717,0.0142088551,-0.0345542394,0.0618678667,-0.1609525383,0.0698502511,-0.0277726464,-0.0126877725,-0.1279167384,-0.2843229175,-0.1038047522,-0.326433152]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1673","title":"Unable to Download Hindi Wikipedia Dataset","comments":"Yes","body":"I used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n","comment_length":1,"text":"Unable to Download Hindi Wikipedia Dataset\nI used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n\nYes","embeddings":[-0.1605094373,-0.0589019731,-0.0731838197,0.2721837163,0.0850575864,0.1388684958,-0.0598694906,0.3268081844,0.2434153706,0.0265026987,0.3429498971,0.0398125499,0.0904918313,0.2291966975,0.1593858749,-0.3564942479,0.0861499533,0.0321488865,0.0813349634,0.1238899752,0.044392366,0.2952028811,-0.3024786711,-0.0217906702,-0.5379131436,0.027862899,0.1875984073,-0.0768048763,-0.2954098582,-0.3628420234,0.3753118813,0.028087439,0.271067977,0.3725534081,-0.0001193605,0.1147044674,0.5256556273,-0.1985906214,-0.2826633751,-0.3983022869,-0.1666505784,-0.3197062016,0.073206231,-0.2472494692,-0.1435828358,0.1750800461,0.3648650944,-0.2410789728,0.1311997026,0.2046667188,0.1513615847,-0.1262546033,0.394685328,-0.1381669044,0.1539474428,0.2273595929,-0.0256191362,0.3042186797,-0.0333038792,0.1054554582,-0.068603605,0.1929351389,0.0848935246,-0.1442332417,0.2640812993,-0.2755292058,-0.1651940495,-0.4362581968,0.4817917347,0.3000311852,1.0910084248,-0.3456637263,-0.1421257854,0.2106039673,-0.039889615,0.2130126804,0.4133118987,0.4615986049,-0.3004294932,0.0160235055,0.1329021305,-0.3872614205,-0.1459991336,0.3871781826,0.0161216445,0.1660468429,-0.0041151051,0.1377120167,-0.0622211583,-0.1825374961,-0.0384258814,-0.0747040734,0.1315123886,0.3147633076,-0.320635587,0.1226087809,-0.146208033,0.165122807,0.0533686765,-0.1872973293,0.0146528799,-0.0050718132,-0.0669585615,-0.0177325923,0.3500912189,-0.0895412937,-0.0027685016,0.2228836268,0.2675311565,0.2771115601,0.0213915613,-0.0628242716,0.1056637466,-0.4156900048,-0.545568943,-0.0630755126,-0.0153888082,-0.2428346276,-0.1343560517,0.0423583463,-0.2883388996,-0.2154269069,0.0730632544,0.073552303,0.0380227827,0.1008385867,0.0314704925,0.0471871309,-0.1405236274,-0.3007537723,-0.0628437698,0.2707746625,-0.3178549111,0.0353584401,0.2519268394,-0.1611049324,0.4106163383,0.0315900519,-0.1820542961,-0.0103667686,-0.101630725,-0.2721654177,-0.2873457074,0.18253842,0.1821099073,0.4131452739,-0.1726799309,-0.0755077526,-0.0042283968,0.1528403312,-0.2636603415,0.005348356,0.1163316593,0.0834820494,-0.240772441,-0.1017809808,0.0897008181,0.2115001827,-0.0283658877,-0.4113608599,0.0092541957,-0.0006716783,-0.1397946179,-0.296159327,0.3243661821,0.4458423257,-0.6702910066,0.058018025,-0.0843310803,-0.0144877331,0.1221236661,-0.0089642433,-0.2122647464,0.6341120005,-0.3332958221,0.0769342333,0.6100831628,-0.2970023453,-0.4984974265,0.0057410533,-0.0487234592,0.0190654974,0.0859883726,0.2429790348,0.184711501,0.022435151,-0.0668176934,0.3275715411,0.0829798579,-0.1105106845,-0.2874236405,-0.0442936607,0.359883517,0.1515224278,0.2221645713,-0.1088690534,0.1149634868,0.4083543122,0.388651222,0.0295695905,0.1573453248,0.3802975714,-0.2826012373,0.0672298521,0.2630061209,-0.0529416166,-0.0304865651,0.2086316943,-0.0376441106,-0.0500971004,-0.1485052258,0.0266477596,-0.4856742322,0.0030756677,-0.2630516589,0.1602405161,0.0851066113,0.0129328957,0.0991667062,0.339410156,-0.0630374253,0.1467555016,-0.2361355722,0.0354604051,-0.3051142395,0.4568080902,-0.1394803077,-0.10046193,0.0196837373,0.140835166,0.180949226,0.0216747876,-0.1554729939,0.0108136712,0.1426948011,-0.0130629186,0.2968997359,-0.1324211657,0.1337527484,-0.5891224146,0.2304272652,0.6115739942,0.064025037,-0.045290526,-0.0053810407,-0.0050290413,-0.2087323815,0.0683087558,-0.0842311233,0.1821877658,0.3720775545,0.0535424799,0.0421014503,0.0151511477,0.2406859845,0.4382483661,0.1884684861,-0.0141391475,-0.1273211688,0.0129033215,0.5432699919,0.0407056808,0.0224693995,0.1229108945,-0.2417102158,-0.0932946727,0.1059861556,0.0797303021,0.1043850183,0.1059884131,0.0047396906,-0.2038253993,0.1125255078,-0.1904201061,0.2157669663,0.1743095368,0.2927399874,0.2298192978,-0.1701170504,0.0074671153,-0.2090406418,-0.0950380862,0.0483490638,0.3153557479,-0.2132389396,-0.085035637,-0.30358845,-0.4926317334,-0.3385273814,-0.2720499337,-0.4327945709,-0.2279996723,-0.0536544546,-0.1586786211,0.0025110976,0.0612145141,-0.0282514058,-0.1696145982,0.0964076295,-0.3396134377,-0.0974454805,-0.4050168693,-0.2985163629,-0.0441204011,0.5184716582,0.0992789418,0.2387679517,-0.3066699207,-0.0604786016,-0.6261211634,-0.1583919674,0.1038863957,-0.0176697094,-0.1010785624,0.0776790902,0.7180455923,-0.1672855765,0.0475734994,0.0865858644,0.2649796903,-0.1098124757,-0.0619459935,-0.0309847035,0.0130844023,0.1669309586,-0.5004774928,-0.4018400311,-0.2508095801,-0.1229101494,0.099883616,0.0594453029,-0.0717921704,0.1009274349,0.0329655111,0.2432644367,0.1539010108,-0.0353285521,-0.0930482298,0.3889521658,-0.1604363471,-0.520616591,0.5911819339,-0.0003299813,0.2096369117,0.0718200505,-0.3233703673,0.3878663182,0.065479137,0.2459874749,0.1294103861,0.0882707983,0.1859444678,-0.2868738472,0.2297470123,0.0959097296,-0.1337715536,0.1621992737,-0.3626516461,0.4980724752,0.3622515202,0.24644427,-0.0765812173,0.6048179865,0.3418103456,0.0895532891,0.3380911052,0.0924113914,0.0151144927,-0.1843877733,-0.2653961778,0.1310109794,0.0158264879,0.0645560026,0.175936982,0.0584587157,-0.1805885434,-0.2670194805,-0.405521065,-0.28889817,-0.3485139012,-0.0725284517,-0.2756978273,0.3652024567,0.0767382756,0.0120223295,0.0314706378,-0.2870505154,0.209088102,0.4102123678,-0.124655515,0.1648157835,0.104323402,-0.4456487894,-0.2801963389,0.0112817092,0.0761576369,0.3569319546,-0.1600037962,0.2255741656,0.220486626,0.0538239554,0.610521853,-0.2355066538,0.1243463606,0.1458099186,0.0473249666,-0.2075496614,-0.1417548507,-0.0470546335,0.3314688206,0.2772872448,-0.0808869526,-0.4756385684,-0.0971618369,0.3879501522,0.3497825861,-0.0356194526,-0.1325241774,-0.2621866763,-0.1984988451,-0.2926791012,-0.2464211434,-0.1558120847,0.4800246954,0.2240490019,0.171518445,0.0370813012,0.0240213759,-0.1540606171,-0.0758261681,0.3812219799,0.2119681537,-0.0815967768,0.322134912,0.2707491517,-0.2722895145,0.3187689185,0.2224117517,-0.2428434938,0.0772033855,-0.0727828071,-0.0580751598,0.0302102156,-0.1047419161,0.0409114212,0.0042237192,-0.3539040685,0.1132711545,-0.1279489994,0.1167868301,0.1498719901,-0.5149570704,-0.4158942699,0.4565430582,0.1544733942,0.1537587047,0.2667201757,0.1273205578,-0.2488879859,0.0883766413,0.0588093624,0.8252080083,0.0091054477,0.1090072468,0.0067121056,0.0732324645,0.3703391254,-0.1616588682,0.0372080803,-0.449844867,0.0723428205,-0.1119557545,0.0251955483,0.0027147869,-0.0445876867,-0.2878738344,0.4787789881,0.1817352772,0.3474054039,0.1091555953,0.3650166392,0.1241484955,-0.1863378882,-0.1897559613,0.0289191846,-0.1667471379,0.4385502636,-0.2515373826,-0.1634045839,-0.1669480354,0.0105923396,-0.3938354254,0.1037142277,-0.3050714433,0.32458812,-0.2027768195,-0.5057308674,0.4895987809,0.4194814265,0.2902536392,0.297259599,-0.3529237211,0.2120724618,-0.431961596,-0.4760139585,-0.016307313,0.2057943046,0.3349253237,-0.1602193117,-0.4165842235,0.2256052196,-0.0263743903,0.0490834527,-0.0841180906,-0.0811103731,0.2431376278,0.0481107198,-0.397492677,-0.0884245113,-0.081346646,-0.1112220883,0.1078829393,0.0491446927,0.0949701369,-0.0275430065,0.2022016346,-0.0962586254,0.0664816201,0.3319656849,0.0539384447,-0.138992548,0.4182008803,0.3320670426,-0.4354907572,-0.168938309,-0.0620141923,-0.4026759863,-0.2683436275,-0.0261089653,0.072581552,0.2527644932,-0.2051793635,0.1648885757,0.0648480281,-0.2528846562,-0.0287054796,-0.4495550692,-0.2546976209,0.244433552,-0.1910889298,0.0287143625,0.0285416562,-0.2896346748,0.4298817813,-0.2682135403,-0.1925943792,-0.0165124517,-0.1737764478,-0.0314546525,0.1541928053,0.0888139829,-0.1691808254,-0.0777919069,0.0553867221,-0.1659153998,-0.0588208698,-0.0841589272,0.2421483099,0.1150474623,0.0202118829,0.1051489264,0.0146904243,-0.4531366527,-0.0808843374,0.0380328894,-0.1343271583,-0.0040126643,-0.0169828217,0.2123029381,-0.0058259922,-0.0128563605,-0.4818258286,0.3460794985,0.0610232539,0.2028129995,0.082620725,0.3021574318,0.2422981113,-0.0907605737,-0.4522017837,0.1685255766,-0.0488993302,-0.0524872541,0.2265900671,-0.3653360009,0.340266645,0.1443543285,0.1173543483,0.0745395645,-0.2104662955,0.0039854874,0.3420725167,0.0822453722,-0.5024277568,0.0302734319,0.7149345279,0.0944142491,-0.0405831225,0.0125121213,0.2565906048,-0.2539511323,-0.0621398352,0.2852384448,0.1532427967,-0.0215492304,0.181558162,0.3633982837,0.3452807367,0.2424183637,0.0405558757,0.0934704319,-0.1735609621,0.5893962979,-0.3875778913,0.1492934823,0.1892826408,-0.1223274618,-0.1034696102,-0.2939020395,0.1187872589,0.1451515704,-0.2234536409,-0.0563212149,-0.0678069368,0.0658459067,0.3206883371,-0.064446032,-0.1964605749,0.0659352019,-0.2450640202,0.2609987259,-0.4734373689,-0.3543959856,-0.0448825024,0.1742208898,0.1315744519,-0.1167959422,0.0053067212,0.2405185401,-0.208242476,-0.078331545,0.7926011682,-0.1295003444,0.0495412089,-0.2018618286,0.0168298036,-0.2061580867,-0.0869030505,-0.0022853347,0.2512445748,0.1913696975,0.286267221,-0.314800024,-0.0369853675,0.0900882259,-0.0438745581,0.1217202768,0.1788881421,0.0825875551,0.0276678093,0.2691628039,0.0833424255,-0.080766812,-0.2714516819,0.3711175919,0.2146866173,-0.0477384515,-0.0640802756,-0.1572515666,-0.0002863591,-0.2405088544,-0.0408919118,-0.5615615249,-0.1216875836,0.3374466598,0.1423927397,-0.0000504295,-0.3678945005,0.0196368955,-0.2126250863,0.4451298416,0.2890412509,0.2002770305,-0.4392872453,-0.3620164096,-0.5997402668,0.1559374481,-0.0828626603,-0.1907678992,0.0861904621,0.0905093029,-0.1249707639,0.1612413228,0.3472520411,0.0788456276,-0.0078203287,0.3881794512,-0.1292658299,-0.1948549747,-0.2178632319,-0.005208577,-0.3052527905,-0.0919377133,0.1604423821,-0.4431411326,0.0117266821,-0.0460536331,-0.1504233927,0.1968577653,-0.2122927457,0.0744058266,-0.2147575468,0.5154703259,0.0306754,0.0362926722,-0.1485247463,0.1850177497,-0.0411020331,0.514929235,-0.0426730998,0.3657058477,-0.1815209985,0.2725125551,-0.1720046252,0.385101229,-0.137690112,0.3313766122,-0.1540521234,-0.0616534278,-0.0109499674,0.3789134622,0.0885166675,0.1892623603,-0.2270369083,0.0563183986,-0.3690291643,-0.2070324272,0.1984246075,-0.400462985,-0.2269666642,-0.082641691,0.0924226195,0.1307742894,0.0853291675,-0.2597124875,0.1172086224,0.1462719142,0.0106670093,-0.0683091879,0.1265809983,-0.1666316241,0.1546225697,0.0336109959,-0.0115993386,-0.0716538951,-0.1951651275,-0.2224350274,-0.3237542808]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1673","title":"Unable to Download Hindi Wikipedia Dataset","comments":"Hey, so i tried installing the library from source using the commands : **git clone https:\/\/github.com\/huggingface\/datasets**, **cd datasets** and then **pip3 install -e .**. But i still am facing the same error that file is not found. Please advise.\r\n\r\nThe Datasets library version now is 1.1.3 by installing from source as compared to the earlier 1.0.3 that i had loaded using pip command but I am still getting same error\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103479005-69f3b080-4df0-11eb-83ae-58d7bb56a90e.png)\r\n","body":"I used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n","comment_length":71,"text":"Unable to Download Hindi Wikipedia Dataset\nI used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n\nHey, so i tried installing the library from source using the commands : **git clone https:\/\/github.com\/huggingface\/datasets**, **cd datasets** and then **pip3 install -e .**. But i still am facing the same error that file is not found. Please advise.\r\n\r\nThe Datasets library version now is 1.1.3 by installing from source as compared to the earlier 1.0.3 that i had loaded using pip command but I am still getting same error\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103479005-69f3b080-4df0-11eb-83ae-58d7bb56a90e.png)\r\n","embeddings":[-0.1465503126,0.0689324141,-0.0685979575,0.2491659224,0.0940222442,0.0981624722,-0.0108038336,0.300265044,0.2056314349,0.0105530787,0.3993399441,0.1249028668,0.094412744,0.1712741852,0.2068688273,-0.3449133039,0.0655948073,0.1091009825,-0.0037668047,0.1099171191,0.1411299109,0.3182806969,-0.2545852959,-0.0223163571,-0.4677815437,0.0209264811,0.1805852801,-0.1294429004,-0.3263757825,-0.3986920714,0.4391380847,0.0544427261,0.2593984604,0.3542151749,-0.0001193817,0.098571986,0.5011398792,-0.1628396213,-0.3172684014,-0.4880221188,-0.1099457964,-0.3709641099,0.0777489766,-0.2636030614,-0.1019325778,0.1779790819,0.3313590288,-0.1715381294,0.0750310197,0.166800946,0.1584782153,-0.0895957723,0.4187116921,-0.1109170541,0.2155592144,0.2446872592,-0.0607995801,0.2880851626,0.0487542935,0.0981846377,-0.0256014336,0.1852132529,0.050882712,-0.0827116445,0.3174597919,-0.2934673429,-0.2573735714,-0.3573926985,0.4213846028,0.3379175663,1.02250278,-0.398583889,-0.273886472,0.1128012687,-0.005135268,0.1822863072,0.4232428968,0.4133462906,-0.2505421638,-0.015567231,0.0776664764,-0.3905896842,-0.1346762478,0.376213938,0.0835084394,0.107283175,-0.0483075678,0.1569276154,-0.0117750047,-0.193001911,-0.0286203381,-0.0906697512,0.121988818,0.3296678364,-0.2907230556,0.1777351499,-0.1184544936,0.156657815,0.0986154079,-0.180471912,-0.019992372,0.0240814723,-0.1072402447,0.0255629774,0.3462226093,-0.0297972932,-0.0316005982,0.2042375654,0.2528335452,0.3454883099,0.0378434807,-0.057752721,0.066608347,-0.3518583179,-0.5248504281,-0.0550721474,-0.0665054545,-0.2746732533,-0.1836228073,0.0162873939,-0.2484945506,-0.1921250671,0.1000580043,0.0523053966,0.0006563097,0.1718006432,0.0245504193,0.0292796567,-0.1951865256,-0.2448223531,-0.0415181071,0.3407041728,-0.3416930437,0.0294138417,0.2804209888,-0.268083632,0.451384604,0.0620260201,-0.1783179641,0.0249203332,-0.0775749758,-0.2494254261,-0.2391320914,0.2149781138,0.1716963947,0.4294681251,-0.1778288484,-0.0899565071,-0.0230659153,0.0799571648,-0.2160615623,0.0227570962,0.0705771744,0.0649512336,-0.3262450099,-0.1010823548,-0.0728589147,0.1652237177,-0.0362290777,-0.4435551763,0.0267364979,-0.0415836424,-0.0482419953,-0.2789181173,0.3309089839,0.4413673878,-0.670966804,0.052452635,-0.1534172446,0.000172739,0.1014323533,0.0084355427,-0.2476768047,0.5937273502,-0.3557330072,0.0363896377,0.5849054456,-0.3862292469,-0.5270808339,0.0094767921,-0.0043359026,0.0828576759,0.0765917078,0.2425976694,0.1698553264,0.0387084521,-0.190120399,0.313441664,0.0714813843,-0.0468932912,-0.2660342157,-0.1286066324,0.3043842912,0.1595204026,0.2136811316,-0.1706256419,0.165395841,0.4260933399,0.3458833396,-0.0076898569,0.1791165918,0.4467867315,-0.1761472821,0.140531674,0.2131312788,-0.1383676827,-0.0526880063,0.1961433589,0.0010356202,-0.0422132239,-0.2385879159,-0.0071198586,-0.4226621687,-0.0625063404,-0.2837450802,0.1142023727,0.0646001622,0.0346823633,0.132375136,0.2899646759,-0.0708186701,0.2151221633,-0.233903721,0.0451227576,-0.420712471,0.4064812064,-0.1321009248,-0.1083956137,0.054941576,0.1968874335,0.1279750168,-0.0218865201,-0.1831482351,0.0394128859,0.0695022941,0.0251194462,0.2400258034,-0.120562315,0.1684936434,-0.5526188016,0.2112795562,0.4790015519,0.0803966075,0.0479688346,-0.0337376334,-0.0516845994,-0.2463632226,0.0547255278,-0.0189790037,0.1343645751,0.3730354905,0.1178276092,-0.0066963318,-0.0023127843,0.2520800233,0.4505518377,0.2667425871,0.0331759229,-0.052323088,0.0564053319,0.6210753322,0.0723135322,0.0047869459,0.1327172667,-0.2396251708,-0.1066733897,0.0785985589,0.1280668676,0.1346791685,0.1067078412,0.0131321168,-0.2335936278,0.1373036653,-0.1548273414,0.1748354882,0.1962943375,0.2368424684,0.1871284395,-0.1233921275,0.0761108994,-0.1863161474,-0.0816784725,0.0659447089,0.3190715015,-0.1899593771,-0.0299229771,-0.3684742749,-0.5022326708,-0.3432816863,-0.3228256702,-0.4344170988,-0.2218503952,-0.0949574262,-0.0881347433,0.0537266769,0.0500501208,0.0405996032,-0.2014796585,0.0236134008,-0.340752244,-0.0830789432,-0.325429827,-0.3207295835,-0.050599739,0.5199856758,0.1368112415,0.3006576598,-0.3158456385,-0.0924753398,-0.6595547199,-0.2461890876,0.1009241417,-0.0530718677,-0.0160337575,0.1856158823,0.6253229976,-0.2144840062,0.0281605292,0.0658021718,0.1511944532,-0.081744507,-0.0750561059,-0.0750973895,-0.0248522907,0.0920032933,-0.5493107438,-0.369910121,-0.272647649,-0.0164931472,0.0344290137,0.058830712,-0.0658104271,0.0304345563,0.1151822954,0.1985549629,0.2137798816,-0.012564918,-0.1578126401,0.4093869328,-0.1883184463,-0.5284603834,0.6092122793,-0.0119419405,0.2732456028,0.0532611683,-0.313065052,0.2904496789,0.0179675743,0.2536846399,0.1646978408,0.1323915124,0.1832594424,-0.2988812625,0.2472976446,0.0479275919,-0.1605923772,0.1197418198,-0.3451824784,0.4536977112,0.3488155603,0.2859294415,-0.0269231107,0.6235115528,0.3517344296,0.0441815183,0.4012577236,0.0901484564,0.1008643657,-0.1746268421,-0.2704689205,0.1210416704,0.0026568915,0.1452383846,0.1305583268,0.0497689955,-0.1058067828,-0.2321633101,-0.3345561028,-0.3322811425,-0.3249893188,-0.1172049791,-0.1760758311,0.3473748267,0.1111967638,0.0025076906,-0.0000054874,-0.3106088936,0.1543326676,0.4284178615,-0.0713842884,0.2148959041,0.0458995253,-0.4885782897,-0.2591528594,-0.0091248155,0.1251290739,0.3944340348,-0.1315015703,0.2689720094,0.2365816832,0.0512200035,0.5420583487,-0.2351160198,0.0345107466,0.0849743858,0.0403554365,-0.2170709372,-0.1607278585,-0.0016796716,0.32254076,0.2367368937,-0.0638760105,-0.5603260994,-0.1590347886,0.418708235,0.4167237878,-0.0038904652,-0.1183226183,-0.2904560566,-0.2515203953,-0.2778495252,-0.2819132805,-0.1444506049,0.4368573725,0.1976367086,0.1351343691,0.0809952989,0.0641106293,-0.1684580147,-0.1230759248,0.3975958526,0.2212846726,-0.0324387513,0.295231998,0.2529827356,-0.1995541602,0.3428439796,0.1516623199,-0.1987978965,0.0646717101,-0.0657847971,-0.05218786,0.0521096401,-0.133967787,0.0974387303,0.0062319427,-0.327023536,0.1063331962,-0.136215508,0.120766528,0.1430390179,-0.4724414051,-0.5154546499,0.4462742805,0.2011181116,0.173393324,0.2453350425,0.1415679008,-0.2159323692,0.0663926005,0.0042481045,0.7669913769,-0.0608967021,0.1652669758,0.0332882367,0.1004792526,0.5030773878,-0.1388473213,0.024729982,-0.4496907294,0.0175354518,-0.1247142479,0.0173496399,0.0424387231,-0.0899445266,-0.2532258332,0.5594903827,0.283392489,0.3509947658,0.0959280059,0.3910237551,0.0198713783,-0.2344890833,-0.2341431081,0.0261879489,-0.1998327672,0.4416565299,-0.2183288783,-0.1659997255,-0.1723139137,-0.0182355158,-0.429384917,0.1123374179,-0.2618550062,0.3072473705,-0.1522675604,-0.4519041479,0.5024827719,0.3713585436,0.2734575272,0.3167138398,-0.3428615034,0.1986531615,-0.4735795856,-0.420625627,0.0418328121,0.2026548535,0.3022682369,-0.1343312114,-0.3665067852,0.1866331846,-0.089621596,0.0395742096,-0.0258383285,-0.0852846578,0.2640121579,0.0369029753,-0.4308215976,-0.1102356911,-0.057109043,-0.1375552118,0.092674315,0.0807585716,0.0554212891,0.0378444642,0.1084174141,-0.1010378003,0.0652038455,0.3398969769,-0.0431594364,-0.1945652515,0.3939539194,0.3470808566,-0.4483514726,-0.147780031,-0.0957914069,-0.3243696988,-0.3189645112,-0.0977712348,0.0892468244,0.2573280036,-0.1305138916,0.166051209,0.0075252014,-0.3035092056,-0.085876897,-0.4674795568,-0.2315335721,0.2842712402,-0.1237466335,0.0332615525,0.0810022056,-0.3440442681,0.3855434656,-0.3068344295,-0.1745396703,0.0019873977,-0.1550554335,0.0048715551,0.1338542402,0.117433548,-0.1343636513,-0.0779162496,0.0519600175,-0.106362246,-0.0498794541,-0.0814205781,0.2275131941,0.1191380918,0.001336985,0.0868032128,0.0504445583,-0.4908800721,-0.0974357128,-0.0086311297,-0.0168312564,0.0106222872,-0.0420549437,0.2127013803,-0.0749426484,-0.0033190595,-0.4467348158,0.3419929147,0.0156423505,0.185180366,0.0935416371,0.3374811709,0.2530322373,-0.0360997133,-0.4261891246,0.1366702467,-0.082328558,-0.0347029306,0.2462344915,-0.366951406,0.3672495484,0.2243917882,0.1517209858,0.116137743,-0.2107327878,0.0282621514,0.3478685021,0.083705768,-0.4758676291,0.0165562741,0.604745388,0.1022493094,-0.053295847,0.0489847288,0.3269353807,-0.2605017722,-0.0398252867,0.2566744387,0.2291958183,-0.0207243487,0.2210004777,0.3879458606,0.3946740627,0.201382935,0.0284474529,0.0693977997,-0.2000321299,0.6186951995,-0.4213269651,0.1975308061,0.1657887548,-0.0326081775,-0.1267840266,-0.2862227559,0.1191459373,0.1686941534,-0.092341952,-0.0536616072,-0.0389852338,0.1953288615,0.2621173859,-0.0370301008,-0.1986600906,0.0807429105,-0.2414989769,0.2650493383,-0.4243359566,-0.3099016547,-0.0961157009,0.1578333974,0.0980414748,-0.0903996825,0.0047102524,0.284439683,-0.1545059234,-0.2538588047,0.7121483684,-0.0825034454,0.0564970151,-0.1628472507,0.0946196467,-0.1678534299,-0.1083511263,-0.0534195602,0.2686801553,0.217285037,0.2691023052,-0.2216748893,-0.000525924,0.0462639667,-0.0440052673,0.1139596999,0.1794165075,0.0153985796,0.0634766892,0.2147220522,0.0782423094,-0.063840203,-0.2578328848,0.3082065582,0.1774465144,-0.0848162025,0.033944793,-0.2267199159,0.0619380064,-0.2382612079,0.0397553705,-0.5503433347,-0.0726515427,0.420170486,0.1671354175,-0.0077170813,-0.4254102409,0.0166796558,-0.1950621754,0.3528371453,0.3148078322,0.186162293,-0.3998849392,-0.3406915963,-0.6655068994,0.1275738776,-0.1036615744,-0.1927263737,0.1470295489,0.0725466982,-0.1872205585,0.0950243622,0.3276728988,0.0989128426,0.0804901049,0.3717657328,-0.1100379154,-0.236791566,-0.2163107246,-0.0296308361,-0.2719547451,-0.1440063715,0.1214045212,-0.4027549922,-0.0222945698,-0.1011069342,-0.0847585425,0.2049462795,-0.2255083919,0.0470510684,-0.1775285304,0.4820343554,0.0612374023,-0.0259463135,-0.1860626191,0.1464214027,-0.0075715557,0.5809044838,-0.1022229567,0.410345912,-0.1863963604,0.2700352967,-0.2632559538,0.3378363848,-0.1088456735,0.2931085825,-0.203467384,-0.0697926357,-0.0784943849,0.4220157862,0.0747844279,0.1694676131,-0.2019442618,0.0627131313,-0.381829679,-0.2224589586,0.1896549761,-0.4549949169,-0.1909749955,-0.0554454066,0.1062466055,0.1403681636,0.0127047729,-0.3124849796,0.1340280026,0.156761989,0.0722616464,-0.0718391687,0.1322742552,-0.1541549861,0.200661689,0.0356682204,0.0074676578,-0.0969307199,-0.2021696568,-0.1120137647,-0.2864823341]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1673","title":"Unable to Download Hindi Wikipedia Dataset","comments":"Looks like the wikipedia dump for hindi at the date of 05\/05\/2020 is not available anymore.\r\nYou can try to load a more recent version of wikipedia\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nd = load_dataset(\"wikipedia\", language=\"hi\", date=\"20210101\", split=\"train\", beam_runner=\"DirectRunner\")\r\n```","body":"I used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n","comment_length":40,"text":"Unable to Download Hindi Wikipedia Dataset\nI used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n\nLooks like the wikipedia dump for hindi at the date of 05\/05\/2020 is not available anymore.\r\nYou can try to load a more recent version of wikipedia\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nd = load_dataset(\"wikipedia\", language=\"hi\", date=\"20210101\", split=\"train\", beam_runner=\"DirectRunner\")\r\n```","embeddings":[-0.1142456904,0.0112914881,-0.0827291012,0.188191995,0.0503296554,0.203110978,-0.0217098705,0.4027991295,0.2084719241,-0.0058662095,0.2971007526,-0.01063844,0.1007415429,0.1128164455,0.1189810112,-0.3750500679,0.0604316592,-0.0714059174,0.1233816445,0.0742849037,0.0124701401,0.2530111372,-0.2966147363,-0.000915119,-0.5076264143,0.0811902434,0.0929644704,-0.1746855974,-0.3081796169,-0.3290758431,0.3946946263,0.032021571,0.3380321264,0.267832309,-0.0001185627,0.0496360399,0.5291934013,-0.1734580994,-0.3872661889,-0.3783107996,-0.1402136832,-0.3079809248,0.0476080514,-0.2449628711,-0.0817109272,0.0995807946,0.3573267162,-0.2165108323,0.119702585,0.2080284059,0.1690301895,-0.0981415585,0.3833535612,-0.2002480477,0.1789742857,0.2349620014,-0.0561930016,0.3583011329,-0.0010121538,-0.0040538209,-0.0369790271,0.1388836801,0.046540305,-0.1358784586,0.2164182961,-0.2033793777,-0.1063366607,-0.380720377,0.4467658699,0.2975744009,1.0478886366,-0.3198606968,-0.2823239565,0.214803502,0.0255117435,0.1601150185,0.3726471961,0.3713812828,-0.2544006407,0.0224460494,0.2328926027,-0.4139357805,-0.0935643092,0.4551313818,0.0004587615,0.1820247024,0.0393327549,0.1147852093,-0.0187585372,-0.265148133,-0.0293263476,-0.1063046306,0.1576577574,0.3844853044,-0.2153396159,0.0957845375,-0.1552342474,0.150580287,0.0482715629,-0.2602938116,-0.0801319703,-0.0422489978,-0.0851458535,-0.0182959754,0.4143214822,-0.1023116261,0.0436332449,0.2053697705,0.2462697923,0.3137840331,0.0165848266,-0.0224873368,0.1246273592,-0.4093409777,-0.4507754147,-0.0420527533,0.0805299729,-0.2413734794,-0.1452439725,0.0494721644,-0.3812088668,-0.2045060843,-0.0651382655,0.0024246085,-0.0156679265,0.1061700061,0.1126058847,0.0522613674,-0.1497132927,-0.3151382804,-0.0803760365,0.1926774681,-0.3331976235,0.1103744507,0.2177271396,-0.2642215788,0.2979155779,0.1510023922,-0.2088230401,-0.0396617167,-0.1556182653,-0.2326006144,-0.3740469515,0.1542014033,0.1772901863,0.4825568795,-0.1454354376,-0.1282797903,-0.0095931599,0.1143854484,-0.3579782844,0.0362945534,0.0355905481,0.084700264,-0.2616971731,-0.0836329684,0.0690908581,0.177265659,-0.0340588652,-0.4436794519,0.0225847475,-0.0526538417,-0.1188783944,-0.2595073581,0.3008177876,0.4370456338,-0.5658004284,0.0131420568,-0.208628282,0.0295671951,0.0542645529,0.0161206499,-0.2735312283,0.66962111,-0.2844646275,-0.0045776512,0.5242993832,-0.266395241,-0.4838491976,-0.014880049,-0.0151251247,0.09028925,-0.0813735351,0.2021463066,0.2086894065,-0.0237698592,0.0005578518,0.3112415969,0.0934147984,-0.194594577,-0.2605804801,-0.0950673744,0.4047927856,0.1130468547,0.2699636519,-0.0457578339,0.1129858568,0.6010671258,0.4021862149,0.1348844469,0.1459055841,0.3607157469,-0.240617767,0.1704009771,0.2450812459,-0.1539526582,-0.0987970456,0.2499047071,0.0002758898,0.0224659145,-0.1439981014,-0.0125864912,-0.44327721,0.0313378796,-0.2176449746,0.1527222246,0.0891195312,-0.0159838405,0.0716809407,0.430372715,0.0207281336,0.0172469784,-0.1806950718,0.016178282,-0.3506973088,0.4398330748,-0.1203364357,-0.0601956099,-0.0120089427,0.1021561399,0.2002846897,0.0723434389,-0.1172060594,-0.0264040567,0.2186687887,0.055788178,0.2651782334,-0.1387015134,0.1483439654,-0.5472911,0.2296742052,0.5981794596,0.0541749336,-0.0634030178,0.0449294746,0.0571675524,-0.1294469535,0.0145988837,-0.1710965186,0.2312998176,0.3597036898,0.0177478809,0.033297196,-0.0226083715,0.2005066723,0.388009429,0.160476163,0.0257180482,-0.0517596751,-0.0501822196,0.5676726699,0.013026678,0.0404762663,0.1214449555,-0.2276733816,-0.1060064733,0.07224489,-0.0035087985,0.0372728705,0.0625212267,0.0145789078,-0.2079574913,0.13756603,-0.1434830576,0.2105375975,0.1933050752,0.2895304263,0.1703340262,-0.1188584939,0.0961954817,-0.1260169297,-0.001003818,0.032303568,0.3559486866,-0.142358005,-0.0748434812,-0.314160496,-0.3821249306,-0.346678853,-0.3569432497,-0.4646942019,-0.320218116,-0.0665283948,-0.1526101232,0.0118520819,0.0595371015,-0.0201086067,-0.1933003515,0.0828795731,-0.3502337337,-0.0672850534,-0.3931134939,-0.2638324201,-0.0219011996,0.4575640559,0.1035848036,0.1910099834,-0.2944647372,-0.1393680423,-0.6602706909,-0.2131776959,0.0333165675,-0.0553252883,-0.0693884715,0.0611698329,0.7040323019,-0.1448130459,-0.0219321381,0.0930676758,0.125358507,-0.0922289193,-0.0308833253,-0.0188181549,0.0705837309,0.1434679627,-0.4446650743,-0.3900807798,-0.1972253025,-0.0793293864,0.046562694,0.0004414085,-0.0096682468,0.10412018,-0.0156586692,0.2523651421,0.0786527023,-0.0663231984,-0.0295413435,0.4097995758,-0.1150566563,-0.4930728376,0.5355672836,-0.0323044658,0.0980366543,0.0833553746,-0.3445494771,0.3909450173,0.1141969487,0.3440346718,0.178769514,-0.0629390851,0.1142304763,-0.3059741557,0.2277783304,0.1264754087,-0.0927611738,0.0243690051,-0.2515067458,0.5262636542,0.3555581272,0.2398677319,-0.0959963873,0.6890134811,0.309278667,0.09135212,0.3108575046,0.1147760823,0.0427565351,-0.1477428824,-0.3294082582,0.0390819907,-0.0017178936,0.0603982992,0.1246396005,-0.0083485739,-0.1762297302,-0.2674736083,-0.3386886418,-0.3087988794,-0.4182576835,-0.031201113,-0.1827286631,0.3806190789,0.1019648612,0.0398650318,0.0076799993,-0.3526425958,0.2515993714,0.3737899959,-0.0607439131,0.1500243545,0.1255884171,-0.382512331,-0.2390328795,0.0255235266,0.0756099746,0.3893580735,-0.1512797773,0.266002357,0.1895328909,0.1310961396,0.4795293808,-0.1971653104,0.1819534302,0.1743209213,0.0958776996,-0.2066469789,-0.1298273951,-0.0494711772,0.3432275951,0.3920090795,-0.0364791639,-0.4435792863,-0.0967472941,0.4116117656,0.3721291423,-0.1031901091,-0.1285009086,-0.1971745044,-0.2075055689,-0.2654876113,-0.3012401164,-0.1196202338,0.4736841917,0.1639019698,0.205946371,0.0096731465,0.1201705188,-0.0808968619,-0.109549053,0.3244810104,0.2624776363,-0.0497949831,0.3203046322,0.0954141319,-0.1609316021,0.3310253024,0.1415163577,-0.3441977203,0.0146330353,-0.0268190224,-0.0948610678,0.0826931,-0.1075916961,0.0648720413,0.0690131858,-0.3516659141,0.0110548194,-0.2043764591,0.2243968397,0.1059361398,-0.4895793498,-0.4795033038,0.4767798483,0.1361737698,0.1465826482,0.222384274,0.1183655262,-0.2488027811,0.0812116116,0.0417956077,0.8113127947,-0.0041978266,0.1299043745,-0.0159675423,0.2420138121,0.4224827588,-0.0433779545,-0.0316000804,-0.4124185443,0.1048569381,-0.1195014641,0.0927783102,0.0567311198,-0.0683569312,-0.252859354,0.4712554514,0.0911273807,0.4095675349,0.1020707712,0.3847072124,0.072476469,-0.0978337526,-0.2327296883,0.0697219148,-0.1691249311,0.4914325178,-0.2189213037,-0.1809260547,-0.1554926336,0.0135536268,-0.4811661243,0.0763368011,-0.2148295343,0.3751368523,-0.1967709661,-0.4316458404,0.42507568,0.4711770713,0.2186172754,0.3196096122,-0.3090032637,0.2268408686,-0.4928474724,-0.5243959427,-0.08519537,0.199942857,0.333986789,-0.0711626783,-0.4607299864,0.29904598,-0.0264660791,-0.0045425408,-0.0245852843,-0.0175213944,0.1757826209,-0.0199280027,-0.3917600214,-0.1023278981,-0.0440567993,-0.1440081,0.1156576201,0.0669276789,0.1960700005,-0.0841203555,0.1445452869,-0.0760212243,0.1062943861,0.3669896126,0.0686006472,-0.1909133494,0.4533345997,0.3784365356,-0.3925841153,-0.1547713131,-0.0374611542,-0.5030235648,-0.2975404859,-0.0450540446,0.0101436172,0.383577764,-0.2043282241,0.1496369988,0.0886124969,-0.2962018549,-0.0348610096,-0.4667868614,-0.171869874,0.3102963269,-0.0983341187,-0.0165487751,0.0724731088,-0.3425881565,0.392580539,-0.2337954044,-0.1992261708,0.0446427539,-0.1082525253,0.012196701,0.1057272255,0.0677056387,-0.1116083488,-0.1628446728,0.0814592838,-0.1640484035,-0.013022718,-0.0976503715,0.1480904222,0.1212373823,-0.0113537712,0.0907410085,0.0184453018,-0.5164158344,-0.0096518788,-0.0228640288,-0.1621379107,0.0907080397,0.0803442374,0.2449538261,0.1194228604,0.0543588921,-0.3819981217,0.3313151598,0.0914922431,0.269565016,0.0802761614,0.3349442482,0.310198158,-0.0144962855,-0.5009613037,0.1729118973,-0.1319601089,-0.0840191245,0.2616025805,-0.3245726228,0.3326554,0.230360508,0.0892667174,0.1715657711,-0.2180767357,-0.0087181684,0.3220073581,0.0964973196,-0.4737289846,0.0498639978,0.6553481817,0.1684619039,-0.1458322406,-0.0410756208,0.1880706698,-0.268345654,-0.0888959467,0.2554204464,0.068015255,-0.0005088054,0.2828114033,0.3966422081,0.4411618114,0.2044318467,0.0622822754,0.1076688394,-0.1690013409,0.4482859671,-0.366330266,0.0783493221,0.1258277595,-0.0819659308,-0.0718963146,-0.3179202974,0.1306102425,0.2340452969,-0.2829898,-0.0061121057,-0.0568183027,0.126840353,0.2818172574,-0.004787324,-0.2626942396,0.0689860433,-0.2399255037,0.232270211,-0.3234376311,-0.3213647008,-0.0518087074,0.124185577,0.1495206654,-0.2248169929,0.0528924167,0.2691709399,-0.2421857715,-0.1750237048,0.6996942759,-0.1228057817,0.0791605935,-0.2152735144,-0.0682317689,-0.170502767,-0.1181153208,-0.0143594891,0.235020563,0.2145269662,0.2371811122,-0.3281451464,0.0307349022,0.0474661887,-0.0158790741,0.1379213333,0.1811313033,0.1174186394,0.0486818403,0.2684905827,0.1020906046,-0.0721880123,-0.3152783215,0.3781619668,0.3369340897,-0.0896062031,0.007954753,-0.1436550617,0.0161772296,-0.2486831844,0.0341167375,-0.4928399324,-0.1555254012,0.3062030971,0.1962838024,-0.0103812171,-0.3409282863,0.0249345992,-0.2520850599,0.3639448583,0.4127925634,0.1538990736,-0.4125403464,-0.3143726587,-0.6434620023,0.1342368573,-0.1958826929,-0.1014530957,0.1521614492,0.1058575138,-0.1459162235,0.1916059256,0.3052421212,-0.0401923098,0.0191674661,0.3578122258,-0.1499290168,-0.2280533314,-0.1995406747,-0.037548501,-0.2679667771,-0.0898636207,0.1356250197,-0.5401660204,-0.0031084481,-0.1343589574,-0.21719414,0.2494380772,-0.1276286542,0.1384731382,-0.2062591016,0.4088146985,0.0422141701,-0.0548482202,-0.248868838,0.0978807658,0.0775737464,0.498691082,-0.1390182823,0.3550518155,-0.1974675953,0.2864797711,-0.1888196915,0.4310458601,-0.193979159,0.1719306856,-0.2592929602,-0.1134958789,-0.0173824467,0.3172316551,0.0900116339,0.2336299568,-0.2447543442,0.1031289771,-0.2843617499,-0.1764483899,0.1598710567,-0.4107302129,-0.2433446497,-0.0845405459,0.0889583752,0.0971836001,0.1263245195,-0.3352665901,0.1043180376,0.232870549,0.0663595721,-0.0430896208,0.1159902364,-0.1386879534,0.1604825407,0.0646861866,0.0074926019,-0.0901330188,-0.2122354656,-0.152677238,-0.3400040567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1673","title":"Unable to Download Hindi Wikipedia Dataset","comments":"Okay, thank you so much","body":"I used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n","comment_length":5,"text":"Unable to Download Hindi Wikipedia Dataset\nI used the Dataset Library in Python to load the wikipedia dataset with the Hindi Config 20200501.hi along with something called beam_runner='DirectRunner' and it keeps giving me the error that the file is not found. I have attached the screenshot of the error and the code both. Please help me to understand how to resolve this issue.\r\n\r\n![Code](https:\/\/user-images.githubusercontent.com\/30871963\/103437466-1f3a3300-4c4e-11eb-9d54-fc9601abfeec.png)\r\n\r\n![Error](https:\/\/user-images.githubusercontent.com\/30871963\/103437407-7ee40e80-4c4d-11eb-8151-a86eb664e6be.png)\r\n\nOkay, thank you so much","embeddings":[-0.1544660628,-0.0721434727,-0.0782444105,0.2858543396,0.0982340798,0.1330393404,-0.0724583641,0.3185287416,0.2396122366,0.0294536892,0.3412092328,0.0395786688,0.086340867,0.2323967963,0.1463043094,-0.3630146384,0.0804661736,0.0282439087,0.0815504417,0.1279117763,0.0420934558,0.2924534082,-0.3064701855,-0.0259541199,-0.5272388458,0.021078337,0.1810135543,-0.0831703916,-0.3026345968,-0.3641305566,0.3693988323,0.0267378986,0.2806425691,0.3802900016,-0.0001191297,0.096371524,0.5213109255,-0.1985916644,-0.2784032226,-0.3928823173,-0.1708170474,-0.3152721524,0.0753380731,-0.2581853569,-0.1426196098,0.1710348576,0.358614713,-0.2376520634,0.1351703703,0.2139098197,0.1550106406,-0.1278423667,0.3964315653,-0.1360064447,0.1529307365,0.2170000672,-0.0235807002,0.306530416,-0.0221805032,0.1045761555,-0.0570672713,0.1927705407,0.079027079,-0.1408817768,0.2720823884,-0.2797789872,-0.1543446183,-0.4466505051,0.4750225544,0.2934005558,1.1102946997,-0.3459926248,-0.1341136247,0.2147438526,-0.0399844684,0.2101479024,0.4064826369,0.4715131521,-0.2968216836,0.0193644091,0.1271566898,-0.372848779,-0.1539758593,0.3899965286,0.0201689396,0.1592190713,-0.006147909,0.1367648691,-0.0632511452,-0.203677088,-0.0362618193,-0.0792663991,0.1378755867,0.3165574074,-0.319770515,0.1130008996,-0.1496423334,0.1606165022,0.0482048355,-0.1750207692,0.0073336493,-0.0020144675,-0.0587585606,-0.0153953899,0.3481206894,-0.080156818,0.0007042593,0.2280527949,0.2620669305,0.2659781277,0.0179692805,-0.0697338134,0.1051103845,-0.4174843729,-0.5391157269,-0.0592969023,-0.0076484382,-0.2503898442,-0.1234888509,0.0393375754,-0.2853660583,-0.2141133398,0.0742918625,0.0778508708,0.0393389575,0.1033064574,0.0315554291,0.0453953184,-0.1332724839,-0.3060471117,-0.0593290925,0.2680808008,-0.3243452013,0.0303311329,0.2490432858,-0.1545991004,0.4089983106,0.027964212,-0.1793727577,-0.01339598,-0.0996172801,-0.2807816863,-0.2931293845,0.1710564494,0.1852927953,0.4125385582,-0.1714462042,-0.0686739236,0.0020378511,0.1581555456,-0.261714071,0.0053702774,0.114339143,0.0889700204,-0.2383293658,-0.0964265615,0.0840188339,0.2037483305,-0.0320790485,-0.3954046071,0.0085450429,-0.0023272652,-0.1440215111,-0.301432997,0.3227068484,0.4405107498,-0.6614277959,0.0618948042,-0.0755787119,-0.0233578458,0.1245468706,-0.0061097625,-0.2114979029,0.6333741546,-0.3329261839,0.0803981796,0.6194733381,-0.2973737419,-0.5007474422,0.0095737893,-0.0553726181,0.0010042792,0.087760061,0.2446612716,0.1954912841,0.0271862708,-0.0764117837,0.3369124234,0.0817677379,-0.100069195,-0.2989677489,-0.0442967229,0.3664987683,0.1510113925,0.2168452144,-0.0962253064,0.1240594238,0.4006758332,0.3860539198,0.0269697867,0.161554724,0.3824412227,-0.2670179605,0.0780897439,0.2692109644,-0.0580925532,-0.031063918,0.2051740587,-0.0426302217,-0.0577145591,-0.146538347,0.0329763405,-0.4887000322,-0.0016235977,-0.2607757747,0.171416223,0.0884326547,0.0079613337,0.0991738364,0.3326156735,-0.0576219894,0.1450489014,-0.2374307066,0.0349875912,-0.3115936816,0.4610894322,-0.1430303305,-0.1084198579,0.0254769083,0.13862212,0.1934860051,0.0242618676,-0.1540277749,0.0072123455,0.1385707557,-0.0073538893,0.3043992221,-0.1231292337,0.1278697699,-0.6036294699,0.2295841575,0.5985167623,0.065348506,-0.0474319085,0.0045026354,-0.0075645675,-0.2017413229,0.0680267438,-0.095225729,0.1842343509,0.3622424304,0.0442763902,0.0466329046,0.0213652756,0.2476721555,0.4392839372,0.1812026948,-0.0121779749,-0.1218423247,0.0205138437,0.5322778821,0.0401079915,0.0245826673,0.1226280108,-0.2385936379,-0.0881685466,0.0980908722,0.0632226542,0.1057419553,0.105239518,-0.0011534694,-0.2018024921,0.1303739101,-0.1912591159,0.2202856243,0.1736283004,0.292899549,0.2245385051,-0.1742150486,0.0038601225,-0.1947142333,-0.0952856392,0.0534788035,0.3299643099,-0.2140192986,-0.0967301875,-0.3055926561,-0.5007157922,-0.3475244343,-0.264854908,-0.429040134,-0.2261990905,-0.0580335632,-0.1543405056,0.0117289452,0.0637181103,-0.0180052575,-0.1597044915,0.0954117626,-0.3176746964,-0.0970718935,-0.4169743955,-0.288769424,-0.0414447822,0.5214242935,0.1051607504,0.2443814129,-0.3002617359,-0.0540624261,-0.6274703741,-0.1528620124,0.0925307721,-0.0100058699,-0.1005221605,0.0747316554,0.7218778133,-0.1563831866,0.044910606,0.081985265,0.2664487064,-0.1069424897,-0.0569588318,-0.0269218367,0.0112580275,0.1692335159,-0.4935762882,-0.4119775295,-0.2474576533,-0.1224936098,0.0906109139,0.0580290444,-0.0799992904,0.095782958,0.0328031182,0.2309182286,0.163202256,-0.0341523625,-0.0913365111,0.3812626898,-0.1620968729,-0.5197646618,0.5947902203,0.0025378282,0.2048813701,0.062333025,-0.3140587509,0.3865538836,0.0551575273,0.2401403785,0.1303135455,0.0880254135,0.179777056,-0.2849123478,0.2292980403,0.08946006,-0.1342566162,0.1538980007,-0.356256634,0.506131053,0.3717095256,0.244882077,-0.0816908032,0.6173710227,0.3413203657,0.0802210271,0.3299131691,0.0929101035,-0.004047622,-0.1697464138,-0.280405432,0.1379494816,0.0203912202,0.069739379,0.1738162339,0.0503287464,-0.1785613894,-0.2657451928,-0.411215961,-0.304885596,-0.3493551314,-0.0749760717,-0.2692434788,0.3695923388,0.0727806687,-0.0019479359,0.0307633001,-0.2862039804,0.2022672296,0.4112346768,-0.1254363358,0.1613654494,0.1038396582,-0.4470138848,-0.2737796903,0.0176014379,0.0664800033,0.3522173762,-0.1632555127,0.2347335368,0.2213146389,0.0631033555,0.6174063087,-0.2410918474,0.1231301129,0.1409139037,0.0463269688,-0.2100592256,-0.1399165541,-0.0533392392,0.3263232708,0.2737722099,-0.0964152515,-0.4731025696,-0.092761524,0.3770270646,0.3455289602,-0.0427816771,-0.1404837072,-0.2554599643,-0.1840831786,-0.2903753221,-0.2442982942,-0.1607371271,0.4859986007,0.2192781866,0.1705099791,0.0355055407,0.0236004833,-0.1540169865,-0.0877860039,0.3820419312,0.2176900953,-0.0780159608,0.321941793,0.2717167735,-0.2625775337,0.3052928448,0.2179190665,-0.2362350821,0.0752056912,-0.0806717798,-0.0548533835,0.0343100242,-0.1003134176,0.057555221,0.0007072043,-0.3401350379,0.1082236543,-0.1407517642,0.1250985265,0.1463369727,-0.508708775,-0.4146636128,0.4452124238,0.1523190439,0.1477224678,0.2689031661,0.1154369861,-0.2491787821,0.0868198797,0.062269859,0.8538019657,0.0174916033,0.0933099315,-0.0014657398,0.0703958645,0.3658012748,-0.1613146812,0.0340354517,-0.4584129155,0.0727859214,-0.1066282913,0.020563595,-0.0034456518,-0.0442338921,-0.2948286235,0.4744451344,0.1720268875,0.3448114395,0.1034467071,0.3655560315,0.1243489161,-0.1789473444,-0.1919023544,0.0303007979,-0.1685818285,0.4330063164,-0.2511150539,-0.1698122025,-0.1603194773,0.0096165193,-0.4012070596,0.1005691737,-0.3151165843,0.3195386231,-0.2101977319,-0.5005089045,0.5095863938,0.4216287136,0.2967089713,0.3003125489,-0.3543883562,0.2088653594,-0.4134968817,-0.4805890322,-0.0284546111,0.2009240687,0.3245066404,-0.1654487997,-0.4284565449,0.2324867398,-0.0165745746,0.0600194372,-0.0953146815,-0.0886579901,0.2420407832,0.0340101682,-0.4005761147,-0.0878786892,-0.0855741352,-0.1068992615,0.1101084575,0.0537295379,0.087402083,-0.0102595063,0.2032286823,-0.093603462,0.0720717311,0.3388929069,0.0555737242,-0.1401110739,0.4156932235,0.3276913464,-0.4365460575,-0.1717865616,-0.0493329242,-0.3964349627,-0.281804353,-0.0305678714,0.07727395,0.2643360794,-0.1970328242,0.1667823344,0.0596887432,-0.2420454323,-0.0289623942,-0.4527751803,-0.2567241788,0.2492405027,-0.1957126409,0.0254627615,0.0181463286,-0.2907323539,0.4297772646,-0.2620403469,-0.1986252964,-0.0056474102,-0.1812485009,-0.019536335,0.1642434001,0.0805090889,-0.1818983257,-0.0837814361,0.0516699962,-0.154701218,-0.063612178,-0.0876018628,0.2460533381,0.1148716658,0.0230890438,0.1047420353,0.0073687918,-0.4586683512,-0.0697614774,0.0475307107,-0.1456328779,-0.0026589215,-0.0070402417,0.2137502134,-0.003045796,-0.0091692982,-0.4876939356,0.347692281,0.0606977157,0.1982127875,0.0842109621,0.3014647961,0.2383134663,-0.09511666,-0.4642399848,0.158900246,-0.0491828509,-0.0507599227,0.2213474214,-0.3610607982,0.3399216831,0.1522394121,0.1247406974,0.0678348243,-0.2161260396,0.0011982692,0.3491067886,0.0880112126,-0.4908066392,0.0356411897,0.7196968198,0.0941659063,-0.0426123515,0.0082958769,0.2468200028,-0.2452961802,-0.0559202731,0.279343009,0.1568006128,-0.0161189474,0.183208406,0.355576694,0.3492599726,0.230570063,0.0449083671,0.1013950333,-0.1727814227,0.5815096498,-0.3934648335,0.1534000039,0.1812427193,-0.1221508682,-0.091534242,-0.2921939492,0.1161322892,0.1483723968,-0.2219302505,-0.0617361739,-0.0822316408,0.0707783848,0.3258110881,-0.0752228349,-0.1835063994,0.0838745907,-0.2397321463,0.2679246366,-0.4864017963,-0.3571428061,-0.0450620055,0.1659560502,0.1331091523,-0.1108922213,0.0036961678,0.2428188175,-0.2171884626,-0.0850231797,0.7850094438,-0.1197789684,0.0560128316,-0.2051125318,0.010395119,-0.2043113261,-0.0836280286,-0.0064414949,0.2666910887,0.1907544881,0.2884558439,-0.3206646442,-0.0330195688,0.1004974842,-0.0482489429,0.1334809363,0.1762001961,0.0831411257,0.0233719163,0.2674332261,0.0865976512,-0.0833109021,-0.2680189013,0.3720452785,0.2156118751,-0.0466302969,-0.0770528093,-0.1392226368,0.0006486932,-0.2475322187,-0.0428945124,-0.5624539256,-0.1179527193,0.3318536878,0.1428346336,-0.003033774,-0.3703491986,0.0201280937,-0.2060913593,0.4508662224,0.2982859612,0.2062217593,-0.4542810619,-0.3683733642,-0.6015108824,0.1596116573,-0.0803872272,-0.1807602644,0.0870203674,0.0868055448,-0.1267277449,0.1653955281,0.3478473723,0.0804981962,-0.0185929462,0.4016457498,-0.1411094964,-0.1890689284,-0.2136560231,-0.0092338575,-0.2982210815,-0.0845095664,0.1623221338,-0.4422294199,0.0130587472,-0.0488164276,-0.1438509375,0.1918028146,-0.2053667754,0.0765018314,-0.2165559977,0.5170443058,0.0338731557,0.0493363999,-0.1397150606,0.1761347055,-0.0382342078,0.5141885281,-0.0324444138,0.360409826,-0.1803201437,0.2668999434,-0.1827866882,0.3837479651,-0.1474260241,0.3254166543,-0.1499433815,-0.0680286735,-0.0013500791,0.3765622079,0.0846283287,0.1798509955,-0.2285605222,0.0651243031,-0.3756961226,-0.2088284194,0.2017837167,-0.3975315392,-0.2280014157,-0.0762504041,0.0938087106,0.1275345981,0.082052052,-0.2573546469,0.1174415872,0.1554253697,0.0105864173,-0.063894935,0.1308356076,-0.1639447212,0.1602512151,0.0318803042,-0.021741638,-0.0629349947,-0.1923876256,-0.2258144021,-0.3185673356]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1672","title":"load_dataset hang on file_lock","comments":"Can you try to upgrade to a more recent version of datasets?","body":"I am trying to load the squad dataset. Fails on Windows 10 but succeeds in Colab.\r\nTransformers: 3.3.1\r\nDatasets: 1.0.2\r\nWindows 10 (also tested in WSL)\r\n\r\n```\r\ndatasets.logging.set_verbosity_debug()\r\ndatasets.\r\ntrain_dataset = load_dataset('squad', split='train')\r\nvalid_dataset = load_dataset('squad', split='validation')\r\n\r\ntrain_dataset.features\r\n```\r\n\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py not found in cache or force_download set to True, downloading to C:\\Users\\simpl\\.cache\\huggingface\\datasets\\tmpzj_o_6u7\r\nDownloading:\r\n5.24k\/? [00:00<00:00, 134kB\/s]\r\nstoring https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py in cache at C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\ncreating metadata file for C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\n\r\nChecking C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py to C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.json\r\nNo config specified, defaulting to first: squad\/plain_text\r\n```\r\n\r\nInterrupting the jupyter kernel we are in a file lock.\r\n\r\nIn Google Colab the download is ok. In contrast to a local run in colab dataset_infos.json is downloaded\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/dataset_infos.json not found in cache or force_download set to True, downloading to \/root\/.cache\/huggingface\/datasets\/tmptl9ha_ad\r\n\r\nDownloading:\r\n2.19k\/? [00:00<00:00, 26.2kB\/s]\r\n```","comment_length":12,"text":"load_dataset hang on file_lock\nI am trying to load the squad dataset. Fails on Windows 10 but succeeds in Colab.\r\nTransformers: 3.3.1\r\nDatasets: 1.0.2\r\nWindows 10 (also tested in WSL)\r\n\r\n```\r\ndatasets.logging.set_verbosity_debug()\r\ndatasets.\r\ntrain_dataset = load_dataset('squad', split='train')\r\nvalid_dataset = load_dataset('squad', split='validation')\r\n\r\ntrain_dataset.features\r\n```\r\n\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py not found in cache or force_download set to True, downloading to C:\\Users\\simpl\\.cache\\huggingface\\datasets\\tmpzj_o_6u7\r\nDownloading:\r\n5.24k\/? [00:00<00:00, 134kB\/s]\r\nstoring https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py in cache at C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\ncreating metadata file for C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\n\r\nChecking C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py to C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.json\r\nNo config specified, defaulting to first: squad\/plain_text\r\n```\r\n\r\nInterrupting the jupyter kernel we are in a file lock.\r\n\r\nIn Google Colab the download is ok. In contrast to a local run in colab dataset_infos.json is downloaded\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/dataset_infos.json not found in cache or force_download set to True, downloading to \/root\/.cache\/huggingface\/datasets\/tmptl9ha_ad\r\n\r\nDownloading:\r\n2.19k\/? [00:00<00:00, 26.2kB\/s]\r\n```\nCan you try to upgrade to a more recent version of datasets?","embeddings":[-0.3021154404,-0.0215215832,-0.0879950076,0.178936407,0.5179888606,0.1499255449,0.5325869918,-0.0388461575,0.0199962519,0.0043181255,-0.2535487115,0.25555709,-0.0103855347,-0.1183214337,-0.0975556076,0.0359974504,0.0950897262,0.1051173806,0.0844943151,0.1125974804,-0.0238301829,0.4240731001,-0.3112899065,-0.2331582755,-0.7260748744,-0.0725926012,0.0726541728,0.2130880803,-0.268955946,-0.2011618763,0.5048645139,0.1790501773,0.2133804262,0.5419107676,-0.0001225752,0.0556077734,0.2359667122,0.0357728191,-0.2530398667,-0.2616369724,-0.2748567462,-0.2957787216,0.1831113249,0.012881794,-0.1750219017,0.7024013996,0.1763481945,-0.4491305947,0.4564883709,0.2201069146,0.1482032835,0.3765569031,-0.0643603578,-0.2676066756,0.1338297576,-0.1435051858,-0.2925020754,0.4845799506,0.4809325337,-0.1317549795,0.1103980094,0.1550246179,-0.0124723623,-0.168261677,0.0610582121,0.0491241217,-0.2083521485,-0.4002087712,0.4759041369,0.1590743959,0.462074101,-0.1445714831,-0.2766668797,-0.0671958849,0.3508067429,0.1159304976,0.5585261583,0.2401416451,-0.0911733434,0.154171586,-0.0766754821,0.1578611284,0.0243432038,-0.0620982833,0.0891702473,0.2551713884,-0.0876449049,0.0554294996,0.1462422758,0.0595081635,0.0929599479,-0.0263972823,-0.0479124971,0.1122642159,-0.5934708714,0.1357288659,-0.330829829,0.3397850692,-0.1689827591,0.0296289809,0.0081195422,0.1290463209,0.3702765703,0.1749778092,-0.0086369375,0.0606242754,0.1405549794,0.0206459891,0.3935605884,0.128599301,-0.3604629338,0.1481845677,-0.3828315437,-0.3367047906,0.1523135751,0.0648473948,0.1792359948,-0.2723613679,-0.2743980885,-0.0824539289,0.0945712551,0.0038924548,0.3614347279,0.5836243629,-0.0673661828,-0.0278212335,-0.1538965106,-0.0384557545,-0.3419782817,0.1740929484,-0.1228892058,0.0187594723,-0.1587666869,0.1141898558,0.4133546352,-0.4438305497,0.3186836243,0.0627911836,0.1318190396,-0.0032936137,0.0621202029,-0.2714750171,-0.2003184855,0.3339361548,0.0880344287,0.2697606981,-0.0520738587,-0.4359939694,-0.1121630594,-0.0120501053,-0.1778403372,-0.0692958608,0.0639475808,0.1069511026,-0.287111789,0.012819564,-0.1811254025,-0.0161171686,0.1178004816,-0.061824739,0.1045139655,-0.2027835846,-0.2577009499,-0.1545438617,0.1867063791,0.6341595054,-0.4503531754,0.0353727788,0.3156071305,-0.3282568157,-0.1789106429,0.2429857254,-0.1054776087,-0.193110764,-0.4120923877,0.0113430647,0.2249480039,-0.557004571,-0.7323595285,0.2544721067,-0.2879525423,-0.108532533,0.3105348647,0.24202995,0.3244645894,-0.1146094203,0.1471154541,0.0610773563,-0.0618501641,0.0426971018,-0.1446962357,-0.1810022891,0.0861845315,0.2156251222,0.0729595497,-0.0619786195,0.1285819709,0.6232088804,0.1470398307,0.1933713555,-0.0859030411,0.3554547429,0.3095434904,-0.1419392377,-0.0124201002,-0.2539836466,-0.7716570497,0.4025574923,-0.064985469,0.1990300715,-0.0898044258,-0.2064938247,-0.1742604226,0.0108299283,-0.3812264502,-0.1427406967,0.0050982153,-0.0223159324,0.1340504885,-0.0619087219,0.0002701475,0.6825737357,-0.0598089062,0.1077914089,-0.2840501368,0.1238967851,-0.2325538248,-0.2640245557,0.0058739875,-0.0486497916,0.1146543249,-0.2998555303,-0.1581552476,0.1780064106,-0.017896438,0.1784740686,0.0875405744,-0.0402759425,0.2334705144,-0.1706898659,-0.0394272134,-0.0271221623,0.10674119,-0.2455772758,-0.1253591627,0.2161942422,-0.3088027835,-0.0180273224,-0.0053029177,-0.1447438002,0.1779081523,0.1409148723,-0.0248716529,0.0228372086,0.2542496026,0.3635209501,0.5597595572,0.2122932971,0.2021749467,0.1637285352,0.6630067229,0.0458530672,-0.1473339051,-0.0978469625,-0.0810114443,0.0671536699,0.1066267192,0.0425916016,0.5582988858,0.1253881305,0.0176518466,0.0071026068,-0.0917788669,-0.3237576485,0.1106104702,0.0651510507,0.0299530141,0.2512910366,0.1034510806,0.0331908837,-0.1568048894,-0.284660995,0.2054218203,0.1000614911,-0.1784664094,0.0483886711,-0.1830164194,0.1056167856,-0.1569919437,0.1695332527,-0.3918553889,-0.1113442555,-0.2496089041,0.4572722614,0.4709657729,0.1210713387,-0.2346424907,0.0072332034,0.146064505,-0.3017837107,-0.1085824668,-0.0973616913,-0.0893343315,-0.0566942245,0.6666872501,-0.149059698,0.2231175005,-0.0336790383,-0.0768713951,-0.003031556,-0.2530536354,0.1237075552,-0.1360322833,0.4206443429,0.0739684999,0.1953096539,-0.2648068368,-0.0657883957,0.2203661799,-0.1555208713,-0.1244605482,-0.184746936,-0.0504730865,-0.1968497336,-0.0472335778,-0.1776357591,-0.4348340333,-0.3862979114,0.020922523,-0.2410447896,-0.1923210323,0.340686053,0.0218808055,0.3659332991,0.1288529783,0.0627439097,-0.1347733289,-0.621917069,0.2681832612,-0.0112997582,-0.3586463332,0.1750245243,-0.0209160596,-0.1368217021,-0.0861629397,-0.3688417971,-0.200804159,-0.2929191589,0.3318698704,-0.1156508848,-0.0950155482,0.3846439719,-0.1934524626,0.0707099363,-0.0797456577,-0.2999759614,0.0513800345,-0.037532907,0.1813334972,-0.0543623082,0.07557524,-0.0318417996,0.6378086805,0.346427381,-0.0828953087,0.2503506839,0.0616514124,0.3179033399,-0.105567269,-0.3772533536,0.1237836331,0.0217826236,0.0070413435,0.2305400968,0.2159917951,0.0657451674,-0.2100141644,-0.1933851242,-0.0691947937,-0.2989993393,0.122077316,0.2476724833,-0.1481981725,0.1170200258,0.2134620547,0.01936372,-0.1991245002,-0.0055058487,0.7495666146,-0.2478500307,0.0212370474,-0.1894887239,-0.080604881,-0.4696344435,0.007457444,0.1202898324,0.3606142998,-0.1436661333,0.0714179575,0.2087249011,-0.22599262,0.876454711,0.1221743003,0.2058797777,0.10523206,-0.1461683065,-0.0911150798,-0.1707974672,-0.09223786,0.2615541816,0.4532833993,0.6301118135,-0.0385815278,-0.3543898165,-0.143497318,-0.0167233869,-0.1722128838,-0.2208207399,-0.2302740365,-0.1717423201,-0.2692455053,0.0475822985,0.0881757885,0.0841906443,-0.0633177459,-0.0908300132,-0.0351934992,-0.2652930617,0.0685051605,0.1933160424,-0.0476473048,0.206092,0.4185990989,0.0047986852,0.2482732981,-0.1124578789,0.5424154401,-0.1702809334,-0.2747665644,0.2352747768,-0.1403506249,-0.0199131742,0.4306934178,-0.1972316951,-0.0197949801,0.0696696267,0.1121012866,-0.0169117786,0.2231727988,0.0980751067,0.184104979,-0.2574020028,-0.0331873037,0.2595709562,-0.0530035906,-0.0126302792,0.1406695247,0.0316335894,-0.1650970131,0.1152468547,-0.0691314116,0.8369383812,-0.4818124771,0.1879780889,-0.0125688929,0.0311591271,0.6355465651,-0.157927826,0.0729460046,-0.262182802,-0.5761338472,0.1016746834,-0.1309420019,0.1884804666,0.0120954765,-0.2257237881,0.0585116893,0.2267003357,0.4708392918,-0.0111767035,0.1981204748,-0.3919526041,-0.2028503716,-0.5293018222,0.114245452,-0.0139455823,0.3403401971,-0.2284068614,0.0226034913,-0.134627074,-0.0921075866,-0.2757972479,0.1896314174,-0.5999928713,0.1270852536,-0.1051605642,-0.5103439093,0.2686319947,0.3383281231,0.2421117127,0.1845908463,-0.1392537802,0.2685423493,0.0077807885,-0.1580024213,-0.0752274916,0.0376364328,0.3033438325,-0.0706577972,-0.19111076,-0.1814781427,-0.0465255082,-0.1682810634,0.0252479129,0.0902535096,0.0156260598,-0.1082209945,-0.2895772755,-0.2553245127,-0.1348431408,-0.1357499808,0.0663055032,-0.0476754643,0.0069899359,0.1341112107,0.1573957354,-0.2546168566,-0.125838533,0.7812122703,-0.4224651754,-0.1673939377,0.531730473,0.1821883172,-0.2767602503,-0.2578667998,0.1053774357,-0.3079369068,-0.3172455728,-0.0627216101,0.1154197976,-0.1104095429,0.2417919189,0.1300779879,0.4141307771,-0.3900484145,0.221310541,-0.6019890904,-0.4822528064,-0.1206789911,-0.0000178164,0.2279850394,0.1503324658,0.1250676364,0.1534176916,0.0684849247,-0.205185324,0.1389004886,-0.3487364948,-0.0442082807,0.3459945023,-0.2889596522,0.2424169779,0.0132993087,0.035027463,0.339216888,-0.1399859488,-0.2143962681,-0.1179978773,0.1341082901,0.0462917201,-0.0544811264,-0.1868456155,-0.2604002953,0.0475720651,-0.029738022,0.3557330072,0.2684982717,0.1377952099,-0.0292272996,-0.1497192085,0.3529652655,-0.1717207432,0.2208089679,0.0136517519,-0.0164809842,0.3168202341,0.11235746,-0.0279756337,0.1515489519,-0.3096603155,-0.1124946326,0.1072988808,0.1982225627,0.0474058539,-0.346159339,0.0111010699,0.2825109065,-0.0127815511,0.2522990704,0.0485424921,0.0202336162,0.2640166283,0.1004365608,-0.1800275743,0.1281296313,0.2956616282,-0.0242180172,0.0614341758,0.1758879274,-0.0279410463,0.1119092852,0.1224343553,0.1633555442,0.8019188046,0.4035438001,-0.0315273553,0.1679771692,0.2547037005,0.2746591568,0.1705827564,0.1046107337,-0.0673757568,0.2202755362,-0.3067367077,0.3227696717,-0.563652277,0.3911701739,0.2394527644,-0.3281103373,-0.0788825825,0.2526055276,-0.0484685116,0.0107441712,-0.1404701918,0.683973968,-0.475199759,-0.1997077316,-0.0535995588,0.123673752,-0.2316673845,-0.1022131667,0.1347427815,0.0309787728,0.0777465403,0.0872583836,0.0274678357,-0.0436853096,0.1842945516,0.3241153359,0.0563391522,-0.0849167332,-0.0122196628,0.1307576597,-0.1130595431,-0.0915634781,0.32018888,0.0766087398,0.1584496498,0.1065644696,0.0596486405,0.3337251544,0.2250444144,-0.0984768048,0.0011566288,0.0703722686,0.0187049918,-0.1584299654,0.2194297463,0.0479044728,-0.1294260025,0.0755432397,0.0610052906,-0.1300656945,-0.0836252198,0.0809474513,-0.2034854442,-0.0816971585,0.1932331026,-0.0650670156,-0.101938121,-0.052704975,0.127596423,-0.2227206379,0.1510452777,0.0266778711,-0.1832033098,0.232638225,-0.0965286195,0.0490203388,0.0260616615,0.4137379527,0.1740300208,0.2699415982,-0.285795629,0.0306574088,-0.7201026082,0.1451295316,0.30068928,-0.0133312605,-0.224881053,0.0997479036,-0.0822949335,0.0449661016,0.241230607,-0.1539658606,0.4191064835,0.1802007854,-0.3893116415,-0.0677264854,0.0581257902,-0.2105434984,-0.0633471534,-0.3447371721,0.120507963,0.0495378003,-0.0942728445,-0.3394928873,-0.1467994153,-0.1389024556,-0.3585758805,0.2288987637,-0.1144960672,0.5724791884,-0.0151055036,-0.0887346342,-0.2234006077,0.0460456982,0.0766761675,-0.0917804837,0.2714507878,0.2899795473,-0.3411710858,-0.0053973435,-0.2218616605,0.2548294663,-0.0433399267,0.0390743315,-0.3349056542,-0.0858316198,-0.0717895627,0.2222231477,0.112218067,0.4876586795,-0.0883960426,0.1779204458,-0.4131251574,-0.2066215575,0.3055531383,-0.5895540118,-0.2191735357,-0.0081723118,0.1451423913,-0.0080041615,-0.1097811013,-0.3922094405,0.2485863417,0.0465252511,-0.1773832738,-0.2967149913,0.3516765237,-0.0151687097,0.0579132438,-0.16020298,0.1675810218,-0.0894695744,-0.0663724914,-0.1200742126,-0.2072401345]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1672","title":"load_dataset hang on file_lock","comments":"Thank, upgrading to 1.1.3 resolved the issue.","body":"I am trying to load the squad dataset. Fails on Windows 10 but succeeds in Colab.\r\nTransformers: 3.3.1\r\nDatasets: 1.0.2\r\nWindows 10 (also tested in WSL)\r\n\r\n```\r\ndatasets.logging.set_verbosity_debug()\r\ndatasets.\r\ntrain_dataset = load_dataset('squad', split='train')\r\nvalid_dataset = load_dataset('squad', split='validation')\r\n\r\ntrain_dataset.features\r\n```\r\n\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py not found in cache or force_download set to True, downloading to C:\\Users\\simpl\\.cache\\huggingface\\datasets\\tmpzj_o_6u7\r\nDownloading:\r\n5.24k\/? [00:00<00:00, 134kB\/s]\r\nstoring https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py in cache at C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\ncreating metadata file for C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\n\r\nChecking C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py to C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.json\r\nNo config specified, defaulting to first: squad\/plain_text\r\n```\r\n\r\nInterrupting the jupyter kernel we are in a file lock.\r\n\r\nIn Google Colab the download is ok. In contrast to a local run in colab dataset_infos.json is downloaded\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/dataset_infos.json not found in cache or force_download set to True, downloading to \/root\/.cache\/huggingface\/datasets\/tmptl9ha_ad\r\n\r\nDownloading:\r\n2.19k\/? [00:00<00:00, 26.2kB\/s]\r\n```","comment_length":7,"text":"load_dataset hang on file_lock\nI am trying to load the squad dataset. Fails on Windows 10 but succeeds in Colab.\r\nTransformers: 3.3.1\r\nDatasets: 1.0.2\r\nWindows 10 (also tested in WSL)\r\n\r\n```\r\ndatasets.logging.set_verbosity_debug()\r\ndatasets.\r\ntrain_dataset = load_dataset('squad', split='train')\r\nvalid_dataset = load_dataset('squad', split='validation')\r\n\r\ntrain_dataset.features\r\n```\r\n\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py not found in cache or force_download set to True, downloading to C:\\Users\\simpl\\.cache\\huggingface\\datasets\\tmpzj_o_6u7\r\nDownloading:\r\n5.24k\/? [00:00<00:00, 134kB\/s]\r\nstoring https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py in cache at C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\ncreating metadata file for C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\n\r\nChecking C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py to C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.json\r\nNo config specified, defaulting to first: squad\/plain_text\r\n```\r\n\r\nInterrupting the jupyter kernel we are in a file lock.\r\n\r\nIn Google Colab the download is ok. In contrast to a local run in colab dataset_infos.json is downloaded\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/dataset_infos.json not found in cache or force_download set to True, downloading to \/root\/.cache\/huggingface\/datasets\/tmptl9ha_ad\r\n\r\nDownloading:\r\n2.19k\/? [00:00<00:00, 26.2kB\/s]\r\n```\nThank, upgrading to 1.1.3 resolved the issue.","embeddings":[-0.3021154404,-0.0215215832,-0.0879950076,0.178936407,0.5179888606,0.1499255449,0.5325869918,-0.0388461575,0.0199962519,0.0043181255,-0.2535487115,0.25555709,-0.0103855347,-0.1183214337,-0.0975556076,0.0359974504,0.0950897262,0.1051173806,0.0844943151,0.1125974804,-0.0238301829,0.4240731001,-0.3112899065,-0.2331582755,-0.7260748744,-0.0725926012,0.0726541728,0.2130880803,-0.268955946,-0.2011618763,0.5048645139,0.1790501773,0.2133804262,0.5419107676,-0.0001225752,0.0556077734,0.2359667122,0.0357728191,-0.2530398667,-0.2616369724,-0.2748567462,-0.2957787216,0.1831113249,0.012881794,-0.1750219017,0.7024013996,0.1763481945,-0.4491305947,0.4564883709,0.2201069146,0.1482032835,0.3765569031,-0.0643603578,-0.2676066756,0.1338297576,-0.1435051858,-0.2925020754,0.4845799506,0.4809325337,-0.1317549795,0.1103980094,0.1550246179,-0.0124723623,-0.168261677,0.0610582121,0.0491241217,-0.2083521485,-0.4002087712,0.4759041369,0.1590743959,0.462074101,-0.1445714831,-0.2766668797,-0.0671958849,0.3508067429,0.1159304976,0.5585261583,0.2401416451,-0.0911733434,0.154171586,-0.0766754821,0.1578611284,0.0243432038,-0.0620982833,0.0891702473,0.2551713884,-0.0876449049,0.0554294996,0.1462422758,0.0595081635,0.0929599479,-0.0263972823,-0.0479124971,0.1122642159,-0.5934708714,0.1357288659,-0.330829829,0.3397850692,-0.1689827591,0.0296289809,0.0081195422,0.1290463209,0.3702765703,0.1749778092,-0.0086369375,0.0606242754,0.1405549794,0.0206459891,0.3935605884,0.128599301,-0.3604629338,0.1481845677,-0.3828315437,-0.3367047906,0.1523135751,0.0648473948,0.1792359948,-0.2723613679,-0.2743980885,-0.0824539289,0.0945712551,0.0038924548,0.3614347279,0.5836243629,-0.0673661828,-0.0278212335,-0.1538965106,-0.0384557545,-0.3419782817,0.1740929484,-0.1228892058,0.0187594723,-0.1587666869,0.1141898558,0.4133546352,-0.4438305497,0.3186836243,0.0627911836,0.1318190396,-0.0032936137,0.0621202029,-0.2714750171,-0.2003184855,0.3339361548,0.0880344287,0.2697606981,-0.0520738587,-0.4359939694,-0.1121630594,-0.0120501053,-0.1778403372,-0.0692958608,0.0639475808,0.1069511026,-0.287111789,0.012819564,-0.1811254025,-0.0161171686,0.1178004816,-0.061824739,0.1045139655,-0.2027835846,-0.2577009499,-0.1545438617,0.1867063791,0.6341595054,-0.4503531754,0.0353727788,0.3156071305,-0.3282568157,-0.1789106429,0.2429857254,-0.1054776087,-0.193110764,-0.4120923877,0.0113430647,0.2249480039,-0.557004571,-0.7323595285,0.2544721067,-0.2879525423,-0.108532533,0.3105348647,0.24202995,0.3244645894,-0.1146094203,0.1471154541,0.0610773563,-0.0618501641,0.0426971018,-0.1446962357,-0.1810022891,0.0861845315,0.2156251222,0.0729595497,-0.0619786195,0.1285819709,0.6232088804,0.1470398307,0.1933713555,-0.0859030411,0.3554547429,0.3095434904,-0.1419392377,-0.0124201002,-0.2539836466,-0.7716570497,0.4025574923,-0.064985469,0.1990300715,-0.0898044258,-0.2064938247,-0.1742604226,0.0108299283,-0.3812264502,-0.1427406967,0.0050982153,-0.0223159324,0.1340504885,-0.0619087219,0.0002701475,0.6825737357,-0.0598089062,0.1077914089,-0.2840501368,0.1238967851,-0.2325538248,-0.2640245557,0.0058739875,-0.0486497916,0.1146543249,-0.2998555303,-0.1581552476,0.1780064106,-0.017896438,0.1784740686,0.0875405744,-0.0402759425,0.2334705144,-0.1706898659,-0.0394272134,-0.0271221623,0.10674119,-0.2455772758,-0.1253591627,0.2161942422,-0.3088027835,-0.0180273224,-0.0053029177,-0.1447438002,0.1779081523,0.1409148723,-0.0248716529,0.0228372086,0.2542496026,0.3635209501,0.5597595572,0.2122932971,0.2021749467,0.1637285352,0.6630067229,0.0458530672,-0.1473339051,-0.0978469625,-0.0810114443,0.0671536699,0.1066267192,0.0425916016,0.5582988858,0.1253881305,0.0176518466,0.0071026068,-0.0917788669,-0.3237576485,0.1106104702,0.0651510507,0.0299530141,0.2512910366,0.1034510806,0.0331908837,-0.1568048894,-0.284660995,0.2054218203,0.1000614911,-0.1784664094,0.0483886711,-0.1830164194,0.1056167856,-0.1569919437,0.1695332527,-0.3918553889,-0.1113442555,-0.2496089041,0.4572722614,0.4709657729,0.1210713387,-0.2346424907,0.0072332034,0.146064505,-0.3017837107,-0.1085824668,-0.0973616913,-0.0893343315,-0.0566942245,0.6666872501,-0.149059698,0.2231175005,-0.0336790383,-0.0768713951,-0.003031556,-0.2530536354,0.1237075552,-0.1360322833,0.4206443429,0.0739684999,0.1953096539,-0.2648068368,-0.0657883957,0.2203661799,-0.1555208713,-0.1244605482,-0.184746936,-0.0504730865,-0.1968497336,-0.0472335778,-0.1776357591,-0.4348340333,-0.3862979114,0.020922523,-0.2410447896,-0.1923210323,0.340686053,0.0218808055,0.3659332991,0.1288529783,0.0627439097,-0.1347733289,-0.621917069,0.2681832612,-0.0112997582,-0.3586463332,0.1750245243,-0.0209160596,-0.1368217021,-0.0861629397,-0.3688417971,-0.200804159,-0.2929191589,0.3318698704,-0.1156508848,-0.0950155482,0.3846439719,-0.1934524626,0.0707099363,-0.0797456577,-0.2999759614,0.0513800345,-0.037532907,0.1813334972,-0.0543623082,0.07557524,-0.0318417996,0.6378086805,0.346427381,-0.0828953087,0.2503506839,0.0616514124,0.3179033399,-0.105567269,-0.3772533536,0.1237836331,0.0217826236,0.0070413435,0.2305400968,0.2159917951,0.0657451674,-0.2100141644,-0.1933851242,-0.0691947937,-0.2989993393,0.122077316,0.2476724833,-0.1481981725,0.1170200258,0.2134620547,0.01936372,-0.1991245002,-0.0055058487,0.7495666146,-0.2478500307,0.0212370474,-0.1894887239,-0.080604881,-0.4696344435,0.007457444,0.1202898324,0.3606142998,-0.1436661333,0.0714179575,0.2087249011,-0.22599262,0.876454711,0.1221743003,0.2058797777,0.10523206,-0.1461683065,-0.0911150798,-0.1707974672,-0.09223786,0.2615541816,0.4532833993,0.6301118135,-0.0385815278,-0.3543898165,-0.143497318,-0.0167233869,-0.1722128838,-0.2208207399,-0.2302740365,-0.1717423201,-0.2692455053,0.0475822985,0.0881757885,0.0841906443,-0.0633177459,-0.0908300132,-0.0351934992,-0.2652930617,0.0685051605,0.1933160424,-0.0476473048,0.206092,0.4185990989,0.0047986852,0.2482732981,-0.1124578789,0.5424154401,-0.1702809334,-0.2747665644,0.2352747768,-0.1403506249,-0.0199131742,0.4306934178,-0.1972316951,-0.0197949801,0.0696696267,0.1121012866,-0.0169117786,0.2231727988,0.0980751067,0.184104979,-0.2574020028,-0.0331873037,0.2595709562,-0.0530035906,-0.0126302792,0.1406695247,0.0316335894,-0.1650970131,0.1152468547,-0.0691314116,0.8369383812,-0.4818124771,0.1879780889,-0.0125688929,0.0311591271,0.6355465651,-0.157927826,0.0729460046,-0.262182802,-0.5761338472,0.1016746834,-0.1309420019,0.1884804666,0.0120954765,-0.2257237881,0.0585116893,0.2267003357,0.4708392918,-0.0111767035,0.1981204748,-0.3919526041,-0.2028503716,-0.5293018222,0.114245452,-0.0139455823,0.3403401971,-0.2284068614,0.0226034913,-0.134627074,-0.0921075866,-0.2757972479,0.1896314174,-0.5999928713,0.1270852536,-0.1051605642,-0.5103439093,0.2686319947,0.3383281231,0.2421117127,0.1845908463,-0.1392537802,0.2685423493,0.0077807885,-0.1580024213,-0.0752274916,0.0376364328,0.3033438325,-0.0706577972,-0.19111076,-0.1814781427,-0.0465255082,-0.1682810634,0.0252479129,0.0902535096,0.0156260598,-0.1082209945,-0.2895772755,-0.2553245127,-0.1348431408,-0.1357499808,0.0663055032,-0.0476754643,0.0069899359,0.1341112107,0.1573957354,-0.2546168566,-0.125838533,0.7812122703,-0.4224651754,-0.1673939377,0.531730473,0.1821883172,-0.2767602503,-0.2578667998,0.1053774357,-0.3079369068,-0.3172455728,-0.0627216101,0.1154197976,-0.1104095429,0.2417919189,0.1300779879,0.4141307771,-0.3900484145,0.221310541,-0.6019890904,-0.4822528064,-0.1206789911,-0.0000178164,0.2279850394,0.1503324658,0.1250676364,0.1534176916,0.0684849247,-0.205185324,0.1389004886,-0.3487364948,-0.0442082807,0.3459945023,-0.2889596522,0.2424169779,0.0132993087,0.035027463,0.339216888,-0.1399859488,-0.2143962681,-0.1179978773,0.1341082901,0.0462917201,-0.0544811264,-0.1868456155,-0.2604002953,0.0475720651,-0.029738022,0.3557330072,0.2684982717,0.1377952099,-0.0292272996,-0.1497192085,0.3529652655,-0.1717207432,0.2208089679,0.0136517519,-0.0164809842,0.3168202341,0.11235746,-0.0279756337,0.1515489519,-0.3096603155,-0.1124946326,0.1072988808,0.1982225627,0.0474058539,-0.346159339,0.0111010699,0.2825109065,-0.0127815511,0.2522990704,0.0485424921,0.0202336162,0.2640166283,0.1004365608,-0.1800275743,0.1281296313,0.2956616282,-0.0242180172,0.0614341758,0.1758879274,-0.0279410463,0.1119092852,0.1224343553,0.1633555442,0.8019188046,0.4035438001,-0.0315273553,0.1679771692,0.2547037005,0.2746591568,0.1705827564,0.1046107337,-0.0673757568,0.2202755362,-0.3067367077,0.3227696717,-0.563652277,0.3911701739,0.2394527644,-0.3281103373,-0.0788825825,0.2526055276,-0.0484685116,0.0107441712,-0.1404701918,0.683973968,-0.475199759,-0.1997077316,-0.0535995588,0.123673752,-0.2316673845,-0.1022131667,0.1347427815,0.0309787728,0.0777465403,0.0872583836,0.0274678357,-0.0436853096,0.1842945516,0.3241153359,0.0563391522,-0.0849167332,-0.0122196628,0.1307576597,-0.1130595431,-0.0915634781,0.32018888,0.0766087398,0.1584496498,0.1065644696,0.0596486405,0.3337251544,0.2250444144,-0.0984768048,0.0011566288,0.0703722686,0.0187049918,-0.1584299654,0.2194297463,0.0479044728,-0.1294260025,0.0755432397,0.0610052906,-0.1300656945,-0.0836252198,0.0809474513,-0.2034854442,-0.0816971585,0.1932331026,-0.0650670156,-0.101938121,-0.052704975,0.127596423,-0.2227206379,0.1510452777,0.0266778711,-0.1832033098,0.232638225,-0.0965286195,0.0490203388,0.0260616615,0.4137379527,0.1740300208,0.2699415982,-0.285795629,0.0306574088,-0.7201026082,0.1451295316,0.30068928,-0.0133312605,-0.224881053,0.0997479036,-0.0822949335,0.0449661016,0.241230607,-0.1539658606,0.4191064835,0.1802007854,-0.3893116415,-0.0677264854,0.0581257902,-0.2105434984,-0.0633471534,-0.3447371721,0.120507963,0.0495378003,-0.0942728445,-0.3394928873,-0.1467994153,-0.1389024556,-0.3585758805,0.2288987637,-0.1144960672,0.5724791884,-0.0151055036,-0.0887346342,-0.2234006077,0.0460456982,0.0766761675,-0.0917804837,0.2714507878,0.2899795473,-0.3411710858,-0.0053973435,-0.2218616605,0.2548294663,-0.0433399267,0.0390743315,-0.3349056542,-0.0858316198,-0.0717895627,0.2222231477,0.112218067,0.4876586795,-0.0883960426,0.1779204458,-0.4131251574,-0.2066215575,0.3055531383,-0.5895540118,-0.2191735357,-0.0081723118,0.1451423913,-0.0080041615,-0.1097811013,-0.3922094405,0.2485863417,0.0465252511,-0.1773832738,-0.2967149913,0.3516765237,-0.0151687097,0.0579132438,-0.16020298,0.1675810218,-0.0894695744,-0.0663724914,-0.1200742126,-0.2072401345]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1672","title":"load_dataset hang on file_lock","comments":"Having the same issue with `datasets 1.1.3` of `1.5.0` (both tracebacks look the same) and `kilt_wikipedia`, Ubuntu 20.04\r\n\r\n```py\r\nIn [1]: from datasets import load_dataset \r\n\r\nIn [2]: wikipedia = load_dataset('kilt_wikipedia')['full'] \r\nDownloading: 7.37kB [00:00, 2.74MB\/s] \r\nDownloading: 3.33kB [00:00, 1.44MB\/s] \r\n^C---------------------------------------------------------------------------\r\nOSError Traceback (most recent call last)\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/utils\/filelock.py in _acquire(self)\r\n 380 try:\r\n--> 381 fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\n 382 except (IOError, OSError):\r\n\r\nOSError: [Errno 37] No locks available\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nKeyboardInterrupt Traceback (most recent call last)\r\n<ipython-input-2-f412d3d46ec9> in <module>\r\n----> 1 wikipedia = load_dataset('kilt_wikipedia')['full']\r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, sav\r\ne_infos, script_version, **config_kwargs)\r\n 601 hash=hash,\r\n 602 features=features,\r\n--> 603 **config_kwargs,\r\n 604 )\r\n 605 \r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/builder.py in __init__(self, *args, **kwargs)\r\n 841 def __init__(self, *args, **kwargs):\r\n 842 self._writer_batch_size = kwargs.pop(\"writer_batch_size\", self._writer_batch_size)\r\n--> 843 super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n 844 \r\n 845 @abc.abstractmethod\r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)\r\n 174 os.makedirs(self._cache_dir_root, exist_ok=True)\r\n 175 lock_path = os.path.join(self._cache_dir_root, self._cache_dir.replace(os.sep, \"_\") + \".lock\")\r\n--> 176 with FileLock(lock_path):\r\n 177 if os.path.exists(self._cache_dir): # check if data exist\r\n 178 if len(os.listdir(self._cache_dir)) > 0:\r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/utils\/filelock.py in __enter__(self)\r\n 312 \r\n 313 def __enter__(self):\r\n--> 314 self.acquire()\r\n 315 return self\r\n 316 \r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/utils\/filelock.py in acquire(self, timeout, poll_intervall)\r\n 261 if not self.is_locked:\r\n 262 logger().debug(\"Attempting to acquire lock %s on %s\", lock_id, lock_filename)\r\n--> 263 self._acquire()\r\n 264 \r\n 265 if self.is_locked:\r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/utils\/filelock.py in _acquire(self)\r\n 379 \r\n 380 try:\r\n--> 381 fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\n 382 except (IOError, OSError):\r\n 383 os.close(fd)\r\n\r\nKeyboardInterrupt: \r\n\r\n```","body":"I am trying to load the squad dataset. Fails on Windows 10 but succeeds in Colab.\r\nTransformers: 3.3.1\r\nDatasets: 1.0.2\r\nWindows 10 (also tested in WSL)\r\n\r\n```\r\ndatasets.logging.set_verbosity_debug()\r\ndatasets.\r\ntrain_dataset = load_dataset('squad', split='train')\r\nvalid_dataset = load_dataset('squad', split='validation')\r\n\r\ntrain_dataset.features\r\n```\r\n\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py not found in cache or force_download set to True, downloading to C:\\Users\\simpl\\.cache\\huggingface\\datasets\\tmpzj_o_6u7\r\nDownloading:\r\n5.24k\/? [00:00<00:00, 134kB\/s]\r\nstoring https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py in cache at C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\ncreating metadata file for C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\n\r\nChecking C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py to C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.json\r\nNo config specified, defaulting to first: squad\/plain_text\r\n```\r\n\r\nInterrupting the jupyter kernel we are in a file lock.\r\n\r\nIn Google Colab the download is ok. In contrast to a local run in colab dataset_infos.json is downloaded\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/dataset_infos.json not found in cache or force_download set to True, downloading to \/root\/.cache\/huggingface\/datasets\/tmptl9ha_ad\r\n\r\nDownloading:\r\n2.19k\/? [00:00<00:00, 26.2kB\/s]\r\n```","comment_length":234,"text":"load_dataset hang on file_lock\nI am trying to load the squad dataset. Fails on Windows 10 but succeeds in Colab.\r\nTransformers: 3.3.1\r\nDatasets: 1.0.2\r\nWindows 10 (also tested in WSL)\r\n\r\n```\r\ndatasets.logging.set_verbosity_debug()\r\ndatasets.\r\ntrain_dataset = load_dataset('squad', split='train')\r\nvalid_dataset = load_dataset('squad', split='validation')\r\n\r\ntrain_dataset.features\r\n```\r\n\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py not found in cache or force_download set to True, downloading to C:\\Users\\simpl\\.cache\\huggingface\\datasets\\tmpzj_o_6u7\r\nDownloading:\r\n5.24k\/? [00:00<00:00, 134kB\/s]\r\nstoring https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py in cache at C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\ncreating metadata file for C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py\r\n\r\nChecking C:\\Users\\simpl\\.cache\\huggingface\\datasets\\f6877c8d2e01e8fcb60dc101be28b54a7522feac756deb9ac5c39c6d8ebef1ce.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py to C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/squad.py at C:\\Users\\simpl\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\squad\\1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41\\squad.json\r\nNo config specified, defaulting to first: squad\/plain_text\r\n```\r\n\r\nInterrupting the jupyter kernel we are in a file lock.\r\n\r\nIn Google Colab the download is ok. In contrast to a local run in colab dataset_infos.json is downloaded\r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/squad\/dataset_infos.json not found in cache or force_download set to True, downloading to \/root\/.cache\/huggingface\/datasets\/tmptl9ha_ad\r\n\r\nDownloading:\r\n2.19k\/? [00:00<00:00, 26.2kB\/s]\r\n```\nHaving the same issue with `datasets 1.1.3` of `1.5.0` (both tracebacks look the same) and `kilt_wikipedia`, Ubuntu 20.04\r\n\r\n```py\r\nIn [1]: from datasets import load_dataset \r\n\r\nIn [2]: wikipedia = load_dataset('kilt_wikipedia')['full'] \r\nDownloading: 7.37kB [00:00, 2.74MB\/s] \r\nDownloading: 3.33kB [00:00, 1.44MB\/s] \r\n^C---------------------------------------------------------------------------\r\nOSError Traceback (most recent call last)\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/utils\/filelock.py in _acquire(self)\r\n 380 try:\r\n--> 381 fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\n 382 except (IOError, OSError):\r\n\r\nOSError: [Errno 37] No locks available\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nKeyboardInterrupt Traceback (most recent call last)\r\n<ipython-input-2-f412d3d46ec9> in <module>\r\n----> 1 wikipedia = load_dataset('kilt_wikipedia')['full']\r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, sav\r\ne_infos, script_version, **config_kwargs)\r\n 601 hash=hash,\r\n 602 features=features,\r\n--> 603 **config_kwargs,\r\n 604 )\r\n 605 \r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/builder.py in __init__(self, *args, **kwargs)\r\n 841 def __init__(self, *args, **kwargs):\r\n 842 self._writer_batch_size = kwargs.pop(\"writer_batch_size\", self._writer_batch_size)\r\n--> 843 super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n 844 \r\n 845 @abc.abstractmethod\r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)\r\n 174 os.makedirs(self._cache_dir_root, exist_ok=True)\r\n 175 lock_path = os.path.join(self._cache_dir_root, self._cache_dir.replace(os.sep, \"_\") + \".lock\")\r\n--> 176 with FileLock(lock_path):\r\n 177 if os.path.exists(self._cache_dir): # check if data exist\r\n 178 if len(os.listdir(self._cache_dir)) > 0:\r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/utils\/filelock.py in __enter__(self)\r\n 312 \r\n 313 def __enter__(self):\r\n--> 314 self.acquire()\r\n 315 return self\r\n 316 \r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/utils\/filelock.py in acquire(self, timeout, poll_intervall)\r\n 261 if not self.is_locked:\r\n 262 logger().debug(\"Attempting to acquire lock %s on %s\", lock_id, lock_filename)\r\n--> 263 self._acquire()\r\n 264 \r\n 265 if self.is_locked:\r\n\r\n~\/anaconda3\/envs\/transformers2\/lib\/python3.7\/site-packages\/datasets\/utils\/filelock.py in _acquire(self)\r\n 379 \r\n 380 try:\r\n--> 381 fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB)\r\n 382 except (IOError, OSError):\r\n 383 os.close(fd)\r\n\r\nKeyboardInterrupt: \r\n\r\n```","embeddings":[-0.3021154404,-0.0215215832,-0.0879950076,0.178936407,0.5179888606,0.1499255449,0.5325869918,-0.0388461575,0.0199962519,0.0043181255,-0.2535487115,0.25555709,-0.0103855347,-0.1183214337,-0.0975556076,0.0359974504,0.0950897262,0.1051173806,0.0844943151,0.1125974804,-0.0238301829,0.4240731001,-0.3112899065,-0.2331582755,-0.7260748744,-0.0725926012,0.0726541728,0.2130880803,-0.268955946,-0.2011618763,0.5048645139,0.1790501773,0.2133804262,0.5419107676,-0.0001225752,0.0556077734,0.2359667122,0.0357728191,-0.2530398667,-0.2616369724,-0.2748567462,-0.2957787216,0.1831113249,0.012881794,-0.1750219017,0.7024013996,0.1763481945,-0.4491305947,0.4564883709,0.2201069146,0.1482032835,0.3765569031,-0.0643603578,-0.2676066756,0.1338297576,-0.1435051858,-0.2925020754,0.4845799506,0.4809325337,-0.1317549795,0.1103980094,0.1550246179,-0.0124723623,-0.168261677,0.0610582121,0.0491241217,-0.2083521485,-0.4002087712,0.4759041369,0.1590743959,0.462074101,-0.1445714831,-0.2766668797,-0.0671958849,0.3508067429,0.1159304976,0.5585261583,0.2401416451,-0.0911733434,0.154171586,-0.0766754821,0.1578611284,0.0243432038,-0.0620982833,0.0891702473,0.2551713884,-0.0876449049,0.0554294996,0.1462422758,0.0595081635,0.0929599479,-0.0263972823,-0.0479124971,0.1122642159,-0.5934708714,0.1357288659,-0.330829829,0.3397850692,-0.1689827591,0.0296289809,0.0081195422,0.1290463209,0.3702765703,0.1749778092,-0.0086369375,0.0606242754,0.1405549794,0.0206459891,0.3935605884,0.128599301,-0.3604629338,0.1481845677,-0.3828315437,-0.3367047906,0.1523135751,0.0648473948,0.1792359948,-0.2723613679,-0.2743980885,-0.0824539289,0.0945712551,0.0038924548,0.3614347279,0.5836243629,-0.0673661828,-0.0278212335,-0.1538965106,-0.0384557545,-0.3419782817,0.1740929484,-0.1228892058,0.0187594723,-0.1587666869,0.1141898558,0.4133546352,-0.4438305497,0.3186836243,0.0627911836,0.1318190396,-0.0032936137,0.0621202029,-0.2714750171,-0.2003184855,0.3339361548,0.0880344287,0.2697606981,-0.0520738587,-0.4359939694,-0.1121630594,-0.0120501053,-0.1778403372,-0.0692958608,0.0639475808,0.1069511026,-0.287111789,0.012819564,-0.1811254025,-0.0161171686,0.1178004816,-0.061824739,0.1045139655,-0.2027835846,-0.2577009499,-0.1545438617,0.1867063791,0.6341595054,-0.4503531754,0.0353727788,0.3156071305,-0.3282568157,-0.1789106429,0.2429857254,-0.1054776087,-0.193110764,-0.4120923877,0.0113430647,0.2249480039,-0.557004571,-0.7323595285,0.2544721067,-0.2879525423,-0.108532533,0.3105348647,0.24202995,0.3244645894,-0.1146094203,0.1471154541,0.0610773563,-0.0618501641,0.0426971018,-0.1446962357,-0.1810022891,0.0861845315,0.2156251222,0.0729595497,-0.0619786195,0.1285819709,0.6232088804,0.1470398307,0.1933713555,-0.0859030411,0.3554547429,0.3095434904,-0.1419392377,-0.0124201002,-0.2539836466,-0.7716570497,0.4025574923,-0.064985469,0.1990300715,-0.0898044258,-0.2064938247,-0.1742604226,0.0108299283,-0.3812264502,-0.1427406967,0.0050982153,-0.0223159324,0.1340504885,-0.0619087219,0.0002701475,0.6825737357,-0.0598089062,0.1077914089,-0.2840501368,0.1238967851,-0.2325538248,-0.2640245557,0.0058739875,-0.0486497916,0.1146543249,-0.2998555303,-0.1581552476,0.1780064106,-0.017896438,0.1784740686,0.0875405744,-0.0402759425,0.2334705144,-0.1706898659,-0.0394272134,-0.0271221623,0.10674119,-0.2455772758,-0.1253591627,0.2161942422,-0.3088027835,-0.0180273224,-0.0053029177,-0.1447438002,0.1779081523,0.1409148723,-0.0248716529,0.0228372086,0.2542496026,0.3635209501,0.5597595572,0.2122932971,0.2021749467,0.1637285352,0.6630067229,0.0458530672,-0.1473339051,-0.0978469625,-0.0810114443,0.0671536699,0.1066267192,0.0425916016,0.5582988858,0.1253881305,0.0176518466,0.0071026068,-0.0917788669,-0.3237576485,0.1106104702,0.0651510507,0.0299530141,0.2512910366,0.1034510806,0.0331908837,-0.1568048894,-0.284660995,0.2054218203,0.1000614911,-0.1784664094,0.0483886711,-0.1830164194,0.1056167856,-0.1569919437,0.1695332527,-0.3918553889,-0.1113442555,-0.2496089041,0.4572722614,0.4709657729,0.1210713387,-0.2346424907,0.0072332034,0.146064505,-0.3017837107,-0.1085824668,-0.0973616913,-0.0893343315,-0.0566942245,0.6666872501,-0.149059698,0.2231175005,-0.0336790383,-0.0768713951,-0.003031556,-0.2530536354,0.1237075552,-0.1360322833,0.4206443429,0.0739684999,0.1953096539,-0.2648068368,-0.0657883957,0.2203661799,-0.1555208713,-0.1244605482,-0.184746936,-0.0504730865,-0.1968497336,-0.0472335778,-0.1776357591,-0.4348340333,-0.3862979114,0.020922523,-0.2410447896,-0.1923210323,0.340686053,0.0218808055,0.3659332991,0.1288529783,0.0627439097,-0.1347733289,-0.621917069,0.2681832612,-0.0112997582,-0.3586463332,0.1750245243,-0.0209160596,-0.1368217021,-0.0861629397,-0.3688417971,-0.200804159,-0.2929191589,0.3318698704,-0.1156508848,-0.0950155482,0.3846439719,-0.1934524626,0.0707099363,-0.0797456577,-0.2999759614,0.0513800345,-0.037532907,0.1813334972,-0.0543623082,0.07557524,-0.0318417996,0.6378086805,0.346427381,-0.0828953087,0.2503506839,0.0616514124,0.3179033399,-0.105567269,-0.3772533536,0.1237836331,0.0217826236,0.0070413435,0.2305400968,0.2159917951,0.0657451674,-0.2100141644,-0.1933851242,-0.0691947937,-0.2989993393,0.122077316,0.2476724833,-0.1481981725,0.1170200258,0.2134620547,0.01936372,-0.1991245002,-0.0055058487,0.7495666146,-0.2478500307,0.0212370474,-0.1894887239,-0.080604881,-0.4696344435,0.007457444,0.1202898324,0.3606142998,-0.1436661333,0.0714179575,0.2087249011,-0.22599262,0.876454711,0.1221743003,0.2058797777,0.10523206,-0.1461683065,-0.0911150798,-0.1707974672,-0.09223786,0.2615541816,0.4532833993,0.6301118135,-0.0385815278,-0.3543898165,-0.143497318,-0.0167233869,-0.1722128838,-0.2208207399,-0.2302740365,-0.1717423201,-0.2692455053,0.0475822985,0.0881757885,0.0841906443,-0.0633177459,-0.0908300132,-0.0351934992,-0.2652930617,0.0685051605,0.1933160424,-0.0476473048,0.206092,0.4185990989,0.0047986852,0.2482732981,-0.1124578789,0.5424154401,-0.1702809334,-0.2747665644,0.2352747768,-0.1403506249,-0.0199131742,0.4306934178,-0.1972316951,-0.0197949801,0.0696696267,0.1121012866,-0.0169117786,0.2231727988,0.0980751067,0.184104979,-0.2574020028,-0.0331873037,0.2595709562,-0.0530035906,-0.0126302792,0.1406695247,0.0316335894,-0.1650970131,0.1152468547,-0.0691314116,0.8369383812,-0.4818124771,0.1879780889,-0.0125688929,0.0311591271,0.6355465651,-0.157927826,0.0729460046,-0.262182802,-0.5761338472,0.1016746834,-0.1309420019,0.1884804666,0.0120954765,-0.2257237881,0.0585116893,0.2267003357,0.4708392918,-0.0111767035,0.1981204748,-0.3919526041,-0.2028503716,-0.5293018222,0.114245452,-0.0139455823,0.3403401971,-0.2284068614,0.0226034913,-0.134627074,-0.0921075866,-0.2757972479,0.1896314174,-0.5999928713,0.1270852536,-0.1051605642,-0.5103439093,0.2686319947,0.3383281231,0.2421117127,0.1845908463,-0.1392537802,0.2685423493,0.0077807885,-0.1580024213,-0.0752274916,0.0376364328,0.3033438325,-0.0706577972,-0.19111076,-0.1814781427,-0.0465255082,-0.1682810634,0.0252479129,0.0902535096,0.0156260598,-0.1082209945,-0.2895772755,-0.2553245127,-0.1348431408,-0.1357499808,0.0663055032,-0.0476754643,0.0069899359,0.1341112107,0.1573957354,-0.2546168566,-0.125838533,0.7812122703,-0.4224651754,-0.1673939377,0.531730473,0.1821883172,-0.2767602503,-0.2578667998,0.1053774357,-0.3079369068,-0.3172455728,-0.0627216101,0.1154197976,-0.1104095429,0.2417919189,0.1300779879,0.4141307771,-0.3900484145,0.221310541,-0.6019890904,-0.4822528064,-0.1206789911,-0.0000178164,0.2279850394,0.1503324658,0.1250676364,0.1534176916,0.0684849247,-0.205185324,0.1389004886,-0.3487364948,-0.0442082807,0.3459945023,-0.2889596522,0.2424169779,0.0132993087,0.035027463,0.339216888,-0.1399859488,-0.2143962681,-0.1179978773,0.1341082901,0.0462917201,-0.0544811264,-0.1868456155,-0.2604002953,0.0475720651,-0.029738022,0.3557330072,0.2684982717,0.1377952099,-0.0292272996,-0.1497192085,0.3529652655,-0.1717207432,0.2208089679,0.0136517519,-0.0164809842,0.3168202341,0.11235746,-0.0279756337,0.1515489519,-0.3096603155,-0.1124946326,0.1072988808,0.1982225627,0.0474058539,-0.346159339,0.0111010699,0.2825109065,-0.0127815511,0.2522990704,0.0485424921,0.0202336162,0.2640166283,0.1004365608,-0.1800275743,0.1281296313,0.2956616282,-0.0242180172,0.0614341758,0.1758879274,-0.0279410463,0.1119092852,0.1224343553,0.1633555442,0.8019188046,0.4035438001,-0.0315273553,0.1679771692,0.2547037005,0.2746591568,0.1705827564,0.1046107337,-0.0673757568,0.2202755362,-0.3067367077,0.3227696717,-0.563652277,0.3911701739,0.2394527644,-0.3281103373,-0.0788825825,0.2526055276,-0.0484685116,0.0107441712,-0.1404701918,0.683973968,-0.475199759,-0.1997077316,-0.0535995588,0.123673752,-0.2316673845,-0.1022131667,0.1347427815,0.0309787728,0.0777465403,0.0872583836,0.0274678357,-0.0436853096,0.1842945516,0.3241153359,0.0563391522,-0.0849167332,-0.0122196628,0.1307576597,-0.1130595431,-0.0915634781,0.32018888,0.0766087398,0.1584496498,0.1065644696,0.0596486405,0.3337251544,0.2250444144,-0.0984768048,0.0011566288,0.0703722686,0.0187049918,-0.1584299654,0.2194297463,0.0479044728,-0.1294260025,0.0755432397,0.0610052906,-0.1300656945,-0.0836252198,0.0809474513,-0.2034854442,-0.0816971585,0.1932331026,-0.0650670156,-0.101938121,-0.052704975,0.127596423,-0.2227206379,0.1510452777,0.0266778711,-0.1832033098,0.232638225,-0.0965286195,0.0490203388,0.0260616615,0.4137379527,0.1740300208,0.2699415982,-0.285795629,0.0306574088,-0.7201026082,0.1451295316,0.30068928,-0.0133312605,-0.224881053,0.0997479036,-0.0822949335,0.0449661016,0.241230607,-0.1539658606,0.4191064835,0.1802007854,-0.3893116415,-0.0677264854,0.0581257902,-0.2105434984,-0.0633471534,-0.3447371721,0.120507963,0.0495378003,-0.0942728445,-0.3394928873,-0.1467994153,-0.1389024556,-0.3585758805,0.2288987637,-0.1144960672,0.5724791884,-0.0151055036,-0.0887346342,-0.2234006077,0.0460456982,0.0766761675,-0.0917804837,0.2714507878,0.2899795473,-0.3411710858,-0.0053973435,-0.2218616605,0.2548294663,-0.0433399267,0.0390743315,-0.3349056542,-0.0858316198,-0.0717895627,0.2222231477,0.112218067,0.4876586795,-0.0883960426,0.1779204458,-0.4131251574,-0.2066215575,0.3055531383,-0.5895540118,-0.2191735357,-0.0081723118,0.1451423913,-0.0080041615,-0.1097811013,-0.3922094405,0.2485863417,0.0465252511,-0.1773832738,-0.2967149913,0.3516765237,-0.0151687097,0.0579132438,-0.16020298,0.1675810218,-0.0894695744,-0.0663724914,-0.1200742126,-0.2072401345]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1671","title":"connection issue ","comments":"Also, mayjor issue for me is the format issue, even if I go through changing the whole code to use load_from_disk, then if I do \r\n\r\nd = datasets.load_from_disk(\"imdb\")\r\nd = d[\"train\"][:10] => the format of this is no more in datasets format\r\nthis is different from you call load_datasets(\"train[10]\")\r\n\r\ncould you tell me how I can make the two datastes the same format @lhoestq \r\n\r\n","body":"Hi\r\nI am getting this connection issue, resulting in large failure on cloud, @lhoestq I appreciate your help on this.\r\n\r\nIf I want to keep the codes the same, so not using save_to_disk, load_from_disk, but save the datastes in the way load_dataset reads from and copy the files in the same folder the datasets library reads from, could you assist me how this can be done, thanks\r\n\r\nI tried to do read the data, save it to a path and then set HF_HOME, which does not work and this is still not reading from the old set path, could you assist me how to save the datasets in a path, and let dataset library read from this path to avoid connection issue. thanks\r\n\r\n```\r\nimdb = datasets.load_dataset(\"imdb\")\r\nimdb.save_to_disk(\"\/idiap\/temp\/rkarimi\/hf_datasets\/imdb\")\r\n>>> os.environ[\"HF_HOME\"]=\"\/idiap\/temp\/rkarimi\/hf_datasets\/\"\r\n>>> imdb = datasets.load_dataset(\"imdb\")\r\nReusing dataset imdb (\/idiap\/temp\/rkarimi\/cache_home_2\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3)\r\n```\r\n\r\nI tried afterwards to set HF_HOME in bash, this makes it read from it, but it cannot let dataset library load from the saved path and still downloading data. could you tell me how to fix this issue @lhoestq thanks \r\n\r\nAlso this is on cloud, so I save them in a path, copy it to \"another machine\" to load the data\r\n\r\n### Error stack\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/finetune_t5_trainer.py\", line 344, in <module>\r\n main()\r\n File \".\/finetune_t5_trainer.py\", line 232, in main\r\n for task in data_args.eval_tasks} if training_args.do_test else None\r\n File \".\/finetune_t5_trainer.py\", line 232, in <dictcomp>\r\n for task in data_args.eval_tasks} if training_args.do_test else None\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 136, in get_dataset\r\n split = self.get_sampled_split(split, n_obs)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 64, in get_sampled_split\r\n dataset = self.load_dataset(split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 454, in load_dataset\r\n split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return http_head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 403, in http_head\r\n url, proxies=proxies, headers=headers, cookies=cookies, allow_redirects=allow_redirects, timeout=timeout\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 542, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 655, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7ff6d6c60a20>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n```\r\n","comment_length":64,"text":"connection issue \nHi\r\nI am getting this connection issue, resulting in large failure on cloud, @lhoestq I appreciate your help on this.\r\n\r\nIf I want to keep the codes the same, so not using save_to_disk, load_from_disk, but save the datastes in the way load_dataset reads from and copy the files in the same folder the datasets library reads from, could you assist me how this can be done, thanks\r\n\r\nI tried to do read the data, save it to a path and then set HF_HOME, which does not work and this is still not reading from the old set path, could you assist me how to save the datasets in a path, and let dataset library read from this path to avoid connection issue. thanks\r\n\r\n```\r\nimdb = datasets.load_dataset(\"imdb\")\r\nimdb.save_to_disk(\"\/idiap\/temp\/rkarimi\/hf_datasets\/imdb\")\r\n>>> os.environ[\"HF_HOME\"]=\"\/idiap\/temp\/rkarimi\/hf_datasets\/\"\r\n>>> imdb = datasets.load_dataset(\"imdb\")\r\nReusing dataset imdb (\/idiap\/temp\/rkarimi\/cache_home_2\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3)\r\n```\r\n\r\nI tried afterwards to set HF_HOME in bash, this makes it read from it, but it cannot let dataset library load from the saved path and still downloading data. could you tell me how to fix this issue @lhoestq thanks \r\n\r\nAlso this is on cloud, so I save them in a path, copy it to \"another machine\" to load the data\r\n\r\n### Error stack\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/finetune_t5_trainer.py\", line 344, in <module>\r\n main()\r\n File \".\/finetune_t5_trainer.py\", line 232, in main\r\n for task in data_args.eval_tasks} if training_args.do_test else None\r\n File \".\/finetune_t5_trainer.py\", line 232, in <dictcomp>\r\n for task in data_args.eval_tasks} if training_args.do_test else None\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 136, in get_dataset\r\n split = self.get_sampled_split(split, n_obs)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 64, in get_sampled_split\r\n dataset = self.load_dataset(split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 454, in load_dataset\r\n split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return http_head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 403, in http_head\r\n url, proxies=proxies, headers=headers, cookies=cookies, allow_redirects=allow_redirects, timeout=timeout\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 542, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 655, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7ff6d6c60a20>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n```\r\n\nAlso, mayjor issue for me is the format issue, even if I go through changing the whole code to use load_from_disk, then if I do \r\n\r\nd = datasets.load_from_disk(\"imdb\")\r\nd = d[\"train\"][:10] => the format of this is no more in datasets format\r\nthis is different from you call load_datasets(\"train[10]\")\r\n\r\ncould you tell me how I can make the two datastes the same format @lhoestq \r\n\r\n","embeddings":[-0.4165683091,0.2338398546,0.0117213819,0.3550438881,0.4361407161,-0.1840167344,0.1718588769,0.203017801,-0.1961055249,0.0112288008,-0.2930115461,0.1609504372,0.1448516548,0.2955971062,0.0645913035,-0.0235872623,-0.1781959832,0.0235761348,-0.4069025815,-0.1952103525,0.0163711477,0.1511496902,0.1948605627,0.1142529994,-0.1687250137,-0.2025160342,0.0332008079,0.3280472457,0.0736503303,-0.1991920322,0.2894149423,0.2229556888,0.1790002882,0.4932547808,-0.0001225407,-0.0549697652,0.1337938607,-0.1734607965,-0.4072666168,-0.4631280303,-0.0064323978,0.1484323442,0.2245416194,-0.3105448782,-0.0386022925,0.1304718256,-0.020976603,-0.4665897191,0.4220196903,0.3426231742,0.1340467781,-0.0510152169,0.1519358307,0.1134421676,0.2371073812,0.1746752858,0.0939037874,0.2888202071,0.2243531346,0.0739746988,0.1331334859,-0.2380449176,-0.2881765068,-0.2595244944,0.3810807765,0.0185023304,0.0558203273,-0.1920615882,0.1628101617,0.3489706218,0.6868793964,-0.2939348221,-0.3240501285,0.1296465993,-0.1057170257,-0.1006093845,0.2473123968,0.1119014323,-0.1696666926,0.2362319827,-0.1059877202,-0.2554987967,-0.5657853484,0.3649443686,0.23361893,-0.1368011981,-0.0151887229,0.0324518047,0.0685580447,-0.1694662422,0.572822988,-0.2556990087,0.1473989487,-0.095241107,-0.2188861817,0.0466725528,-0.3066892028,0.0719366148,0.2427719384,0.0897350758,0.1054017991,0.1554961056,-0.2294098139,0.1013626903,0.181036979,-0.0353980511,-0.0266640913,0.1002140269,0.5406512618,0.3721492589,-0.1249257624,-0.129481554,0.0568670519,-0.0367878564,0.1227776483,0.0311886594,0.2147913277,-0.3598895669,-0.1131691262,0.0711495131,0.0917075723,-0.1393203884,0.1415489912,0.7215830088,-0.1206659973,0.1398363858,0.0471956655,0.1483438015,-0.1907476187,0.0251497291,-0.0653437078,-0.0253225043,0.1732768714,0.1365409195,0.1921574473,-0.1690635383,0.2207565159,0.1226691455,0.267064184,-0.2381130755,0.0871987194,-0.3716704845,0.1613682657,0.2053639442,0.1392564476,0.2550626695,-0.0422037803,-0.0452656262,-0.0251929462,0.0961957425,-0.3489897847,-0.365759939,-0.2603074908,0.0855558664,-0.0185977984,-0.1519390792,-0.5155616403,-0.4576686323,-0.0135056907,-0.275706321,-0.1152858436,-0.1120437831,0.036160294,-0.4071178436,0.4432790875,0.4497457743,-0.3629339337,0.0974959061,-0.0111646624,0.0004489943,0.0442566611,0.335114181,-0.1948211938,0.09084443,-0.3700456619,-0.1470950842,0.4817884862,-0.6308919191,-0.5266734958,0.4691863954,-0.0156159634,-0.1747933924,0.1873798221,0.1840229332,0.2436390519,-0.0247960705,0.0157400817,0.2771080434,0.1599723846,-0.0193772558,-0.2106706798,-0.3089704812,0.0360030867,0.1118314564,0.1359898448,0.0014546949,0.4625453353,-0.0260967836,0.3786990643,-0.1464327723,0.1524419188,0.3833598197,0.4820323884,0.2073187679,-0.1340682507,0.2457385808,-0.4730626345,-0.0478930026,0.1075474471,-0.4402324557,-0.2094289511,-0.1622176468,-0.0057171565,-0.1584044248,-0.2368806303,0.1601702571,-0.0140387705,0.0610916875,-0.1786529869,-0.1449938715,-0.1914950013,0.3710275292,-0.1508071274,0.0346166715,-0.1332603097,0.6925702095,-0.3196692765,-0.0803304911,0.0977722034,-0.1760411114,0.2887410522,-0.2172831893,-0.2229076326,0.3530237675,-0.2944036424,0.5860389471,-0.1928031445,0.3995559812,0.1996216178,-0.1799443513,0.3573048711,-0.0611427613,0.2429034859,0.0051078419,-0.0665533468,0.2225022167,-0.3444164693,0.3657674789,0.4451170564,-0.1658440083,0.1474837661,-0.0515021794,0.1195932254,-0.010836835,0.2701353729,-0.0191503912,0.2505918145,-0.1071067154,-0.2997033298,0.338786006,0.1614518166,-0.0755201876,0.2802351117,0.1416828334,0.3327289224,-0.1510933936,0.0215278435,0.2114137709,0.350869596,0.1427525729,-0.1889165491,0.2852966785,0.3808850348,-0.290107131,0.3794764578,0.0049454826,0.2021841407,0.4133448005,0.0140298577,-0.0563507676,-0.1360384822,-0.2094858736,-0.0645500645,0.0968624502,-0.5591400266,0.0748847723,-0.330696851,-0.2309482694,-0.3336356878,0.1958571523,-0.0456189476,-0.189805612,-0.3087564707,0.5337558389,0.1978805512,-0.0155970724,-0.0680829212,0.107169278,-0.2461314499,-0.2705364823,-0.2247102708,-0.0855608284,-0.1869094521,-0.0415779278,0.2394335717,0.067908287,0.1281425506,-0.147133559,-0.3475012183,-0.5005226135,-0.025833169,-0.012016207,0.3362388313,0.3599482775,0.1549184173,0.5422621369,0.0101410924,-0.2701672316,0.0528144576,0.0025792038,-0.0369844846,-0.0801247582,-0.0444890447,-0.062526755,0.1591738462,-0.376952976,-0.4660072625,-0.272729218,0.0742559433,-0.1292780489,0.4784614146,0.0407723263,-0.0397594422,0.0824822262,0.0171046741,0.2734913826,-0.104222171,-0.6493803859,0.1882528812,-0.2475812584,-0.2464491576,0.0708665103,0.1661071628,0.3669281304,-0.103027828,-0.7127384543,-0.3805267513,0.1247557551,0.4643068016,-0.2653199136,0.1684736907,0.2085427046,-0.0138028627,0.0622594543,-0.0191278141,-0.1240472794,0.1722651124,0.2472811341,0.2826841772,0.0826563165,0.4393643439,-0.3292027116,0.4117842615,0.2907215655,0.052050855,0.6116373539,-0.0687991828,0.3417039216,-0.1881387979,-0.2586171627,0.1163145974,0.0173529331,-0.2553443015,-0.0844282359,-0.1897151619,0.2259676307,-0.2243671864,-0.4729626775,-0.3650858998,-0.4024696052,-0.0116716987,-0.0944223031,0.0846013203,-0.0072426349,0.1034730524,0.0596796088,-0.1606900543,0.262101233,0.5211011767,-0.1559973806,0.1142594293,-0.1473679245,-0.1331163496,-0.3509144783,0.2465466559,-0.0877445564,0.388772577,-0.3745872676,0.3151901662,0.0184102505,-0.2634575665,0.7598444819,-0.340820998,0.2458364516,-0.0185975432,0.0526477993,-0.2140598595,0.056219928,0.1495471746,-0.0919722021,-0.0185823888,0.3741905689,-0.1134554818,-0.1500326544,0.0334630609,0.2197373211,-0.1522755474,-0.0592103712,0.0857426822,-0.3991432488,-0.2136281729,-0.2902208269,-0.3151868284,0.0152739137,-0.1055769473,-0.1324443966,0.0001828343,0.0047289184,0.0593171529,-0.2191532105,0.3563083112,-0.1530951411,0.2984580994,0.0896889791,0.0829967707,0.7660005093,0.5066099763,-0.3261796236,0.2330475748,0.0288674477,-0.0180783123,0.3309481442,0.4470452964,-0.0877508968,-0.0187325664,0.2051284611,-0.0134756332,0.0337531976,-0.2055154145,-0.0325732902,0.0433270559,-0.1204738468,-0.49457407,0.3498412669,0.0218594503,-0.102430217,-0.0798329711,0.154010877,-0.228349179,0.4239747524,0.1683757752,1.1728327274,-0.0085933125,0.407662183,-0.0464611202,-0.323951304,0.0329189822,-0.3433544338,-0.0580771677,-0.5660092235,-0.2452655882,-0.0923602581,-0.2051523775,-0.1008117497,0.0198146906,-0.2067545056,0.6651971936,-0.4083108902,-0.1387770027,-0.1261160672,0.3855769336,-0.202336669,-0.0526325181,-0.2361828089,0.0526998192,-0.2278695256,0.1461645067,-0.0807892606,-0.0475775041,-0.0285189245,-0.2451552302,-0.1060827076,0.1113935858,-0.6149661541,0.2729372978,-0.2194919884,-0.3267889321,0.1695637107,0.1855261624,0.2594456673,0.2520202696,-0.0014896638,0.1197755784,0.0358878709,0.0964157134,0.0441535637,-0.0308905803,0.1161925718,-0.0914627761,-0.3058595359,-0.0451437905,-0.0193823818,-0.0452691652,0.0230773706,-0.3219892979,0.2040060759,-0.3108045161,-0.2169777751,-0.1227657199,-0.1670120806,-0.2081405818,-0.0219599325,0.0332980007,-0.0036613685,0.3311715126,0.2592055202,-0.0984126404,-0.0400426947,0.7084525824,0.0247398913,0.0371671617,0.6152187586,0.0180473328,-0.1918168515,-0.1420805007,0.3525808454,0.011256054,-0.1141099259,0.0025599184,-0.2095712721,0.0917290673,-0.018081475,0.2597585022,0.0363621376,0.3591594994,-0.196846813,-0.4136096835,-0.3947238028,0.0499111116,0.0878014639,0.1732582003,0.2596768737,-0.2510609925,-0.1806850135,-0.1892799288,-0.196068719,0.0868523046,-0.2406547368,-0.0528017692,0.3912680447,-0.235229224,0.2191277295,-0.2918088138,0.0669735968,-0.219236955,-0.1210115105,-0.0586634129,-0.2007757425,0.2231466025,-0.1096741036,-0.0928883478,-0.2505300939,-0.5493827462,-0.0648497269,0.1458461583,0.1204134971,0.1669927984,0.0496790595,-0.1041597202,-0.2347135842,-0.2688831985,-0.4103073776,0.1611325145,0.1240783483,0.1731611639,-0.1562412977,0.2550908029,-0.0834015831,-0.2990101278,-0.2021295726,-0.0415378921,-0.01654776,0.1194501892,0.2923726439,-0.1619505435,-0.3103433847,0.13136971,0.1914162338,0.5632075071,-0.1509615481,-0.0557343587,0.1753114313,0.0461311974,-0.0049435506,-0.1973165125,0.0394362621,-0.2136090696,-0.0814311653,0.2099070251,0.0970344692,-0.161249727,-0.2320539951,-0.0296471752,0.5801815987,-0.2229622304,0.0660822317,0.1348181516,-0.0585788265,-0.0523232184,-0.0407470986,0.2509841919,0.1600105315,0.3742899597,-0.3978114426,0.1851177961,0.261072576,0.241316244,0.0902866796,-0.4189111292,-0.0002381842,-0.1706841737,-0.1442528963,-0.006130225,-0.047476843,0.2074502558,0.1213902757,-0.4080002904,0.1326811463,0.2399294674,-0.0919525549,-0.1811965108,-0.0814219117,-0.23401241,-0.0908504203,-0.0121558513,-0.2507009506,-0.0693347156,0.15660429,0.4984638393,0.1658786088,-0.2415312976,0.2404808104,0.0544873774,-0.0064413268,-0.1461607218,0.5727660656,0.225406751,0.0644956455,-0.1021256149,0.2984086871,0.6130160093,0.3451414406,0.2410974801,0.3834908903,0.2394800335,-0.0181098022,0.1865581721,-0.4949088097,0.3360934854,0.3135307729,0.4099771082,0.0559390336,0.0829235613,0.2300323099,-0.0025673273,0.1394113749,-0.0711572245,0.0936671048,0.2314650565,-0.2152895033,-0.37132442,0.005285278,-0.3311599195,-0.0767544061,0.5433692336,-0.111650236,0.2025739998,0.026869148,0.0055434955,-0.0697944537,0.5532213449,0.2662383616,0.1484848708,-0.4372366965,0.1165821776,-0.4312411845,-0.1177116185,0.0813818276,0.1280947328,-0.1398498416,0.1143865362,-0.0723445192,-0.1362173706,0.4376218319,0.1402816474,0.0691804886,0.3971472979,-0.1524437815,-0.0732319877,0.1713889986,0.1265297681,-0.0854833275,-0.3033074439,0.3095442951,-0.0572910272,-0.0626063868,0.0117075676,0.0055476134,-0.0062817321,-0.2815601826,0.3604546785,-0.0331917666,0.3741059303,-0.0612893701,-0.0597654991,0.1181571707,-0.2616134882,-0.1063210666,0.0853293389,0.3383645713,0.1440545768,0.0114563936,-0.1457680315,-0.5411687493,0.0251784977,-0.1509300768,0.0171873961,0.0581273809,-0.1721749604,0.0490011089,0.0729078576,0.120352909,0.1281689405,-0.0488925725,0.2641179562,-0.4909194112,-0.3520485461,0.6163319945,0.0661152378,0.0947724581,-0.2561660111,0.5273269415,-0.0475769043,0.126377359,-0.4738439322,0.1005249768,0.2954449654,-0.2473184168,-0.1833804548,0.2309462875,-0.1437371373,0.344193995,-0.0171284154,0.0471277647,0.0349635407,-0.2077980936,-0.2120913714,-0.1606736034]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1671","title":"connection issue ","comments":"> `\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7ff6d6c60a20>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))`\r\n\r\nDo you have an internet connection on the machine ? Is there a proxy that might block requests to aws ?\r\n\r\n> I tried to do read the data, save it to a path and then set HF_HOME, which does not work and this is still not reading from the old set path, could you assist me how to save the datasets in a path, and let dataset library read from this path to avoid connection issue. thanks\r\n\r\nHF_HOME is used to specify the directory for the cache files of this library.\r\nYou can use save_to_disk and load_from_disk without changing the HF_HOME:\r\n```python\r\nimdb = datasets.load_dataset(\"imdb\")\r\nimdb.save_to_disk(\"\/idiap\/temp\/rkarimi\/hf_datasets\/imdb\")\r\nimdb = datasets.load_from_disk(\"\/idiap\/temp\/rkarimi\/hf_datasets\/imdb\")\r\n```\r\n\r\n> could you tell me how I can make the two datastes the same format\r\n\r\nIndeed they returns different things:\r\n- `load_dataset` returns a `Dataset` object if the split is specified, or a `DatasetDict` if no split is given. Therefore `load_datasets(\"imdb\", split=\"train[10]\")` returns a `Dataset` object containing 10 elements.\r\n- doing `d[\"train\"][:10]` on a DatasetDict \"d\" gets the train split `d[\"train\"]` as a `Dataset` object and then gets the first 10 elements as a dictionary","body":"Hi\r\nI am getting this connection issue, resulting in large failure on cloud, @lhoestq I appreciate your help on this.\r\n\r\nIf I want to keep the codes the same, so not using save_to_disk, load_from_disk, but save the datastes in the way load_dataset reads from and copy the files in the same folder the datasets library reads from, could you assist me how this can be done, thanks\r\n\r\nI tried to do read the data, save it to a path and then set HF_HOME, which does not work and this is still not reading from the old set path, could you assist me how to save the datasets in a path, and let dataset library read from this path to avoid connection issue. thanks\r\n\r\n```\r\nimdb = datasets.load_dataset(\"imdb\")\r\nimdb.save_to_disk(\"\/idiap\/temp\/rkarimi\/hf_datasets\/imdb\")\r\n>>> os.environ[\"HF_HOME\"]=\"\/idiap\/temp\/rkarimi\/hf_datasets\/\"\r\n>>> imdb = datasets.load_dataset(\"imdb\")\r\nReusing dataset imdb (\/idiap\/temp\/rkarimi\/cache_home_2\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3)\r\n```\r\n\r\nI tried afterwards to set HF_HOME in bash, this makes it read from it, but it cannot let dataset library load from the saved path and still downloading data. could you tell me how to fix this issue @lhoestq thanks \r\n\r\nAlso this is on cloud, so I save them in a path, copy it to \"another machine\" to load the data\r\n\r\n### Error stack\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/finetune_t5_trainer.py\", line 344, in <module>\r\n main()\r\n File \".\/finetune_t5_trainer.py\", line 232, in main\r\n for task in data_args.eval_tasks} if training_args.do_test else None\r\n File \".\/finetune_t5_trainer.py\", line 232, in <dictcomp>\r\n for task in data_args.eval_tasks} if training_args.do_test else None\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 136, in get_dataset\r\n split = self.get_sampled_split(split, n_obs)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 64, in get_sampled_split\r\n dataset = self.load_dataset(split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 454, in load_dataset\r\n split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return http_head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 403, in http_head\r\n url, proxies=proxies, headers=headers, cookies=cookies, allow_redirects=allow_redirects, timeout=timeout\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 542, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 655, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7ff6d6c60a20>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n```\r\n","comment_length":210,"text":"connection issue \nHi\r\nI am getting this connection issue, resulting in large failure on cloud, @lhoestq I appreciate your help on this.\r\n\r\nIf I want to keep the codes the same, so not using save_to_disk, load_from_disk, but save the datastes in the way load_dataset reads from and copy the files in the same folder the datasets library reads from, could you assist me how this can be done, thanks\r\n\r\nI tried to do read the data, save it to a path and then set HF_HOME, which does not work and this is still not reading from the old set path, could you assist me how to save the datasets in a path, and let dataset library read from this path to avoid connection issue. thanks\r\n\r\n```\r\nimdb = datasets.load_dataset(\"imdb\")\r\nimdb.save_to_disk(\"\/idiap\/temp\/rkarimi\/hf_datasets\/imdb\")\r\n>>> os.environ[\"HF_HOME\"]=\"\/idiap\/temp\/rkarimi\/hf_datasets\/\"\r\n>>> imdb = datasets.load_dataset(\"imdb\")\r\nReusing dataset imdb (\/idiap\/temp\/rkarimi\/cache_home_2\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3)\r\n```\r\n\r\nI tried afterwards to set HF_HOME in bash, this makes it read from it, but it cannot let dataset library load from the saved path and still downloading data. could you tell me how to fix this issue @lhoestq thanks \r\n\r\nAlso this is on cloud, so I save them in a path, copy it to \"another machine\" to load the data\r\n\r\n### Error stack\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/finetune_t5_trainer.py\", line 344, in <module>\r\n main()\r\n File \".\/finetune_t5_trainer.py\", line 232, in main\r\n for task in data_args.eval_tasks} if training_args.do_test else None\r\n File \".\/finetune_t5_trainer.py\", line 232, in <dictcomp>\r\n for task in data_args.eval_tasks} if training_args.do_test else None\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 136, in get_dataset\r\n split = self.get_sampled_split(split, n_obs)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 64, in get_sampled_split\r\n dataset = self.load_dataset(split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 454, in load_dataset\r\n split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return http_head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 403, in http_head\r\n url, proxies=proxies, headers=headers, cookies=cookies, allow_redirects=allow_redirects, timeout=timeout\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 542, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 655, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7ff6d6c60a20>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n```\r\n\n> `\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7ff6d6c60a20>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))`\r\n\r\nDo you have an internet connection on the machine ? Is there a proxy that might block requests to aws ?\r\n\r\n> I tried to do read the data, save it to a path and then set HF_HOME, which does not work and this is still not reading from the old set path, could you assist me how to save the datasets in a path, and let dataset library read from this path to avoid connection issue. thanks\r\n\r\nHF_HOME is used to specify the directory for the cache files of this library.\r\nYou can use save_to_disk and load_from_disk without changing the HF_HOME:\r\n```python\r\nimdb = datasets.load_dataset(\"imdb\")\r\nimdb.save_to_disk(\"\/idiap\/temp\/rkarimi\/hf_datasets\/imdb\")\r\nimdb = datasets.load_from_disk(\"\/idiap\/temp\/rkarimi\/hf_datasets\/imdb\")\r\n```\r\n\r\n> could you tell me how I can make the two datastes the same format\r\n\r\nIndeed they returns different things:\r\n- `load_dataset` returns a `Dataset` object if the split is specified, or a `DatasetDict` if no split is given. Therefore `load_datasets(\"imdb\", split=\"train[10]\")` returns a `Dataset` object containing 10 elements.\r\n- doing `d[\"train\"][:10]` on a DatasetDict \"d\" gets the train split `d[\"train\"]` as a `Dataset` object and then gets the first 10 elements as a dictionary","embeddings":[-0.4165683091,0.2338398546,0.0117213819,0.3550438881,0.4361407161,-0.1840167344,0.1718588769,0.203017801,-0.1961055249,0.0112288008,-0.2930115461,0.1609504372,0.1448516548,0.2955971062,0.0645913035,-0.0235872623,-0.1781959832,0.0235761348,-0.4069025815,-0.1952103525,0.0163711477,0.1511496902,0.1948605627,0.1142529994,-0.1687250137,-0.2025160342,0.0332008079,0.3280472457,0.0736503303,-0.1991920322,0.2894149423,0.2229556888,0.1790002882,0.4932547808,-0.0001225407,-0.0549697652,0.1337938607,-0.1734607965,-0.4072666168,-0.4631280303,-0.0064323978,0.1484323442,0.2245416194,-0.3105448782,-0.0386022925,0.1304718256,-0.020976603,-0.4665897191,0.4220196903,0.3426231742,0.1340467781,-0.0510152169,0.1519358307,0.1134421676,0.2371073812,0.1746752858,0.0939037874,0.2888202071,0.2243531346,0.0739746988,0.1331334859,-0.2380449176,-0.2881765068,-0.2595244944,0.3810807765,0.0185023304,0.0558203273,-0.1920615882,0.1628101617,0.3489706218,0.6868793964,-0.2939348221,-0.3240501285,0.1296465993,-0.1057170257,-0.1006093845,0.2473123968,0.1119014323,-0.1696666926,0.2362319827,-0.1059877202,-0.2554987967,-0.5657853484,0.3649443686,0.23361893,-0.1368011981,-0.0151887229,0.0324518047,0.0685580447,-0.1694662422,0.572822988,-0.2556990087,0.1473989487,-0.095241107,-0.2188861817,0.0466725528,-0.3066892028,0.0719366148,0.2427719384,0.0897350758,0.1054017991,0.1554961056,-0.2294098139,0.1013626903,0.181036979,-0.0353980511,-0.0266640913,0.1002140269,0.5406512618,0.3721492589,-0.1249257624,-0.129481554,0.0568670519,-0.0367878564,0.1227776483,0.0311886594,0.2147913277,-0.3598895669,-0.1131691262,0.0711495131,0.0917075723,-0.1393203884,0.1415489912,0.7215830088,-0.1206659973,0.1398363858,0.0471956655,0.1483438015,-0.1907476187,0.0251497291,-0.0653437078,-0.0253225043,0.1732768714,0.1365409195,0.1921574473,-0.1690635383,0.2207565159,0.1226691455,0.267064184,-0.2381130755,0.0871987194,-0.3716704845,0.1613682657,0.2053639442,0.1392564476,0.2550626695,-0.0422037803,-0.0452656262,-0.0251929462,0.0961957425,-0.3489897847,-0.365759939,-0.2603074908,0.0855558664,-0.0185977984,-0.1519390792,-0.5155616403,-0.4576686323,-0.0135056907,-0.275706321,-0.1152858436,-0.1120437831,0.036160294,-0.4071178436,0.4432790875,0.4497457743,-0.3629339337,0.0974959061,-0.0111646624,0.0004489943,0.0442566611,0.335114181,-0.1948211938,0.09084443,-0.3700456619,-0.1470950842,0.4817884862,-0.6308919191,-0.5266734958,0.4691863954,-0.0156159634,-0.1747933924,0.1873798221,0.1840229332,0.2436390519,-0.0247960705,0.0157400817,0.2771080434,0.1599723846,-0.0193772558,-0.2106706798,-0.3089704812,0.0360030867,0.1118314564,0.1359898448,0.0014546949,0.4625453353,-0.0260967836,0.3786990643,-0.1464327723,0.1524419188,0.3833598197,0.4820323884,0.2073187679,-0.1340682507,0.2457385808,-0.4730626345,-0.0478930026,0.1075474471,-0.4402324557,-0.2094289511,-0.1622176468,-0.0057171565,-0.1584044248,-0.2368806303,0.1601702571,-0.0140387705,0.0610916875,-0.1786529869,-0.1449938715,-0.1914950013,0.3710275292,-0.1508071274,0.0346166715,-0.1332603097,0.6925702095,-0.3196692765,-0.0803304911,0.0977722034,-0.1760411114,0.2887410522,-0.2172831893,-0.2229076326,0.3530237675,-0.2944036424,0.5860389471,-0.1928031445,0.3995559812,0.1996216178,-0.1799443513,0.3573048711,-0.0611427613,0.2429034859,0.0051078419,-0.0665533468,0.2225022167,-0.3444164693,0.3657674789,0.4451170564,-0.1658440083,0.1474837661,-0.0515021794,0.1195932254,-0.010836835,0.2701353729,-0.0191503912,0.2505918145,-0.1071067154,-0.2997033298,0.338786006,0.1614518166,-0.0755201876,0.2802351117,0.1416828334,0.3327289224,-0.1510933936,0.0215278435,0.2114137709,0.350869596,0.1427525729,-0.1889165491,0.2852966785,0.3808850348,-0.290107131,0.3794764578,0.0049454826,0.2021841407,0.4133448005,0.0140298577,-0.0563507676,-0.1360384822,-0.2094858736,-0.0645500645,0.0968624502,-0.5591400266,0.0748847723,-0.330696851,-0.2309482694,-0.3336356878,0.1958571523,-0.0456189476,-0.189805612,-0.3087564707,0.5337558389,0.1978805512,-0.0155970724,-0.0680829212,0.107169278,-0.2461314499,-0.2705364823,-0.2247102708,-0.0855608284,-0.1869094521,-0.0415779278,0.2394335717,0.067908287,0.1281425506,-0.147133559,-0.3475012183,-0.5005226135,-0.025833169,-0.012016207,0.3362388313,0.3599482775,0.1549184173,0.5422621369,0.0101410924,-0.2701672316,0.0528144576,0.0025792038,-0.0369844846,-0.0801247582,-0.0444890447,-0.062526755,0.1591738462,-0.376952976,-0.4660072625,-0.272729218,0.0742559433,-0.1292780489,0.4784614146,0.0407723263,-0.0397594422,0.0824822262,0.0171046741,0.2734913826,-0.104222171,-0.6493803859,0.1882528812,-0.2475812584,-0.2464491576,0.0708665103,0.1661071628,0.3669281304,-0.103027828,-0.7127384543,-0.3805267513,0.1247557551,0.4643068016,-0.2653199136,0.1684736907,0.2085427046,-0.0138028627,0.0622594543,-0.0191278141,-0.1240472794,0.1722651124,0.2472811341,0.2826841772,0.0826563165,0.4393643439,-0.3292027116,0.4117842615,0.2907215655,0.052050855,0.6116373539,-0.0687991828,0.3417039216,-0.1881387979,-0.2586171627,0.1163145974,0.0173529331,-0.2553443015,-0.0844282359,-0.1897151619,0.2259676307,-0.2243671864,-0.4729626775,-0.3650858998,-0.4024696052,-0.0116716987,-0.0944223031,0.0846013203,-0.0072426349,0.1034730524,0.0596796088,-0.1606900543,0.262101233,0.5211011767,-0.1559973806,0.1142594293,-0.1473679245,-0.1331163496,-0.3509144783,0.2465466559,-0.0877445564,0.388772577,-0.3745872676,0.3151901662,0.0184102505,-0.2634575665,0.7598444819,-0.340820998,0.2458364516,-0.0185975432,0.0526477993,-0.2140598595,0.056219928,0.1495471746,-0.0919722021,-0.0185823888,0.3741905689,-0.1134554818,-0.1500326544,0.0334630609,0.2197373211,-0.1522755474,-0.0592103712,0.0857426822,-0.3991432488,-0.2136281729,-0.2902208269,-0.3151868284,0.0152739137,-0.1055769473,-0.1324443966,0.0001828343,0.0047289184,0.0593171529,-0.2191532105,0.3563083112,-0.1530951411,0.2984580994,0.0896889791,0.0829967707,0.7660005093,0.5066099763,-0.3261796236,0.2330475748,0.0288674477,-0.0180783123,0.3309481442,0.4470452964,-0.0877508968,-0.0187325664,0.2051284611,-0.0134756332,0.0337531976,-0.2055154145,-0.0325732902,0.0433270559,-0.1204738468,-0.49457407,0.3498412669,0.0218594503,-0.102430217,-0.0798329711,0.154010877,-0.228349179,0.4239747524,0.1683757752,1.1728327274,-0.0085933125,0.407662183,-0.0464611202,-0.323951304,0.0329189822,-0.3433544338,-0.0580771677,-0.5660092235,-0.2452655882,-0.0923602581,-0.2051523775,-0.1008117497,0.0198146906,-0.2067545056,0.6651971936,-0.4083108902,-0.1387770027,-0.1261160672,0.3855769336,-0.202336669,-0.0526325181,-0.2361828089,0.0526998192,-0.2278695256,0.1461645067,-0.0807892606,-0.0475775041,-0.0285189245,-0.2451552302,-0.1060827076,0.1113935858,-0.6149661541,0.2729372978,-0.2194919884,-0.3267889321,0.1695637107,0.1855261624,0.2594456673,0.2520202696,-0.0014896638,0.1197755784,0.0358878709,0.0964157134,0.0441535637,-0.0308905803,0.1161925718,-0.0914627761,-0.3058595359,-0.0451437905,-0.0193823818,-0.0452691652,0.0230773706,-0.3219892979,0.2040060759,-0.3108045161,-0.2169777751,-0.1227657199,-0.1670120806,-0.2081405818,-0.0219599325,0.0332980007,-0.0036613685,0.3311715126,0.2592055202,-0.0984126404,-0.0400426947,0.7084525824,0.0247398913,0.0371671617,0.6152187586,0.0180473328,-0.1918168515,-0.1420805007,0.3525808454,0.011256054,-0.1141099259,0.0025599184,-0.2095712721,0.0917290673,-0.018081475,0.2597585022,0.0363621376,0.3591594994,-0.196846813,-0.4136096835,-0.3947238028,0.0499111116,0.0878014639,0.1732582003,0.2596768737,-0.2510609925,-0.1806850135,-0.1892799288,-0.196068719,0.0868523046,-0.2406547368,-0.0528017692,0.3912680447,-0.235229224,0.2191277295,-0.2918088138,0.0669735968,-0.219236955,-0.1210115105,-0.0586634129,-0.2007757425,0.2231466025,-0.1096741036,-0.0928883478,-0.2505300939,-0.5493827462,-0.0648497269,0.1458461583,0.1204134971,0.1669927984,0.0496790595,-0.1041597202,-0.2347135842,-0.2688831985,-0.4103073776,0.1611325145,0.1240783483,0.1731611639,-0.1562412977,0.2550908029,-0.0834015831,-0.2990101278,-0.2021295726,-0.0415378921,-0.01654776,0.1194501892,0.2923726439,-0.1619505435,-0.3103433847,0.13136971,0.1914162338,0.5632075071,-0.1509615481,-0.0557343587,0.1753114313,0.0461311974,-0.0049435506,-0.1973165125,0.0394362621,-0.2136090696,-0.0814311653,0.2099070251,0.0970344692,-0.161249727,-0.2320539951,-0.0296471752,0.5801815987,-0.2229622304,0.0660822317,0.1348181516,-0.0585788265,-0.0523232184,-0.0407470986,0.2509841919,0.1600105315,0.3742899597,-0.3978114426,0.1851177961,0.261072576,0.241316244,0.0902866796,-0.4189111292,-0.0002381842,-0.1706841737,-0.1442528963,-0.006130225,-0.047476843,0.2074502558,0.1213902757,-0.4080002904,0.1326811463,0.2399294674,-0.0919525549,-0.1811965108,-0.0814219117,-0.23401241,-0.0908504203,-0.0121558513,-0.2507009506,-0.0693347156,0.15660429,0.4984638393,0.1658786088,-0.2415312976,0.2404808104,0.0544873774,-0.0064413268,-0.1461607218,0.5727660656,0.225406751,0.0644956455,-0.1021256149,0.2984086871,0.6130160093,0.3451414406,0.2410974801,0.3834908903,0.2394800335,-0.0181098022,0.1865581721,-0.4949088097,0.3360934854,0.3135307729,0.4099771082,0.0559390336,0.0829235613,0.2300323099,-0.0025673273,0.1394113749,-0.0711572245,0.0936671048,0.2314650565,-0.2152895033,-0.37132442,0.005285278,-0.3311599195,-0.0767544061,0.5433692336,-0.111650236,0.2025739998,0.026869148,0.0055434955,-0.0697944537,0.5532213449,0.2662383616,0.1484848708,-0.4372366965,0.1165821776,-0.4312411845,-0.1177116185,0.0813818276,0.1280947328,-0.1398498416,0.1143865362,-0.0723445192,-0.1362173706,0.4376218319,0.1402816474,0.0691804886,0.3971472979,-0.1524437815,-0.0732319877,0.1713889986,0.1265297681,-0.0854833275,-0.3033074439,0.3095442951,-0.0572910272,-0.0626063868,0.0117075676,0.0055476134,-0.0062817321,-0.2815601826,0.3604546785,-0.0331917666,0.3741059303,-0.0612893701,-0.0597654991,0.1181571707,-0.2616134882,-0.1063210666,0.0853293389,0.3383645713,0.1440545768,0.0114563936,-0.1457680315,-0.5411687493,0.0251784977,-0.1509300768,0.0171873961,0.0581273809,-0.1721749604,0.0490011089,0.0729078576,0.120352909,0.1281689405,-0.0488925725,0.2641179562,-0.4909194112,-0.3520485461,0.6163319945,0.0661152378,0.0947724581,-0.2561660111,0.5273269415,-0.0475769043,0.126377359,-0.4738439322,0.1005249768,0.2954449654,-0.2473184168,-0.1833804548,0.2309462875,-0.1437371373,0.344193995,-0.0171284154,0.0471277647,0.0349635407,-0.2077980936,-0.2120913714,-0.1606736034]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1670","title":"wiki_dpr pre-processing performance","comments":"Hi ! And thanks for the tips :) \r\n\r\nIndeed currently `wiki_dpr` takes some time to be processed.\r\nMultiprocessing for dataset generation is definitely going to speed up things.\r\n\r\nRegarding the index note that for the default configurations, the index is downloaded instead of being built, which avoid spending time on constructing the index. However in other cases it would be awesome to make the construction faster.\r\n\r\nAny contribution that can help things faster are welcome. In particular in you have some code that can build a wiki_dpr IVF PQ index in a sharded GPU setup and would like to share it, we can add it to an `examples` folder. In particular since faiss is becoming the library of reference for dataset indexing for tasks like Open Domain Question Answering.\r\n\r\n","body":"I've been working with wiki_dpr and noticed that the dataset processing is seriously impaired in performance [1]. It takes about 12h to process the entire dataset. Most of this time is simply loading and processing the data, but the actual indexing is also quite slow (3h).\r\n\r\nI won't repeat the concerns around multiprocessing as they are addressed in other issues (#786), but this is the first obvious thing to do. Using cython to speed up the text manipulation may be also help. Loading and processing a dataset of this size in under 15 minutes does not seem unreasonable on a modern multi-core machine. I have hit such targets myself on similar tasks. Would love to see this improve.\r\n\r\nThe other issue is that it takes 3h to construct the FAISS index. If only we could use GPUs with HNSW, but we can't. My sharded GPU indexing code can build an IVF + PQ index in 10 minutes on 20 million vectors. Still, 3h seems slow even for the CPU.\r\n\r\nIt looks like HF is adding only 1000 vectors at a time by default [2], whereas the faiss benchmarks adds 1 million vectors at a time (effectively) [3]. It's possible the runtime could be reduced with a larger batch. Also, it looks like project dependencies ultimately use OpenBLAS, but this is known to have issues when combined with OpenMP, which HNSW does [3]. A workaround is to set the environment variable `OMP_WAIT_POLICY=PASSIVE` via `os.environ` or similar.\r\n\r\nReferences:\r\n[1] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py\r\n[2] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/search.py\r\n[3] https:\/\/github.com\/facebookresearch\/faiss\/blob\/master\/benchs\/bench_hnsw.py\r\n[4] https:\/\/github.com\/facebookresearch\/faiss\/issues\/422","comment_length":129,"text":"wiki_dpr pre-processing performance\nI've been working with wiki_dpr and noticed that the dataset processing is seriously impaired in performance [1]. It takes about 12h to process the entire dataset. Most of this time is simply loading and processing the data, but the actual indexing is also quite slow (3h).\r\n\r\nI won't repeat the concerns around multiprocessing as they are addressed in other issues (#786), but this is the first obvious thing to do. Using cython to speed up the text manipulation may be also help. Loading and processing a dataset of this size in under 15 minutes does not seem unreasonable on a modern multi-core machine. I have hit such targets myself on similar tasks. Would love to see this improve.\r\n\r\nThe other issue is that it takes 3h to construct the FAISS index. If only we could use GPUs with HNSW, but we can't. My sharded GPU indexing code can build an IVF + PQ index in 10 minutes on 20 million vectors. Still, 3h seems slow even for the CPU.\r\n\r\nIt looks like HF is adding only 1000 vectors at a time by default [2], whereas the faiss benchmarks adds 1 million vectors at a time (effectively) [3]. It's possible the runtime could be reduced with a larger batch. Also, it looks like project dependencies ultimately use OpenBLAS, but this is known to have issues when combined with OpenMP, which HNSW does [3]. A workaround is to set the environment variable `OMP_WAIT_POLICY=PASSIVE` via `os.environ` or similar.\r\n\r\nReferences:\r\n[1] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py\r\n[2] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/search.py\r\n[3] https:\/\/github.com\/facebookresearch\/faiss\/blob\/master\/benchs\/bench_hnsw.py\r\n[4] https:\/\/github.com\/facebookresearch\/faiss\/issues\/422\nHi ! And thanks for the tips :) \r\n\r\nIndeed currently `wiki_dpr` takes some time to be processed.\r\nMultiprocessing for dataset generation is definitely going to speed up things.\r\n\r\nRegarding the index note that for the default configurations, the index is downloaded instead of being built, which avoid spending time on constructing the index. However in other cases it would be awesome to make the construction faster.\r\n\r\nAny contribution that can help things faster are welcome. In particular in you have some code that can build a wiki_dpr IVF PQ index in a sharded GPU setup and would like to share it, we can add it to an `examples` folder. In particular since faiss is becoming the library of reference for dataset indexing for tasks like Open Domain Question Answering.\r\n\r\n","embeddings":[-0.2196649015,-0.1832672656,-0.1137228534,0.088132672,-0.1135564297,-0.0815373436,0.0223083887,0.3311191797,0.1895469576,0.0709527209,0.0205211919,-0.1025920883,0.3246052861,0.1452307552,-0.3447281122,-0.2720117271,0.2055481523,0.0610853657,0.1536909789,0.0085712625,-0.2336437255,-0.0478305891,-0.4042540789,-0.0988772362,-0.1302234828,-0.2336065769,-0.0790712684,-0.1159494072,-0.1927477121,-0.4908298254,0.2295529395,0.4961724281,0.1521090567,0.1528985053,-0.000108029,-0.1149050891,0.202179715,0.2288627923,0.1197232381,0.6329683065,0.0736345351,-0.1388030648,0.1686449498,-0.2438187152,0.0160673484,-0.3819811642,-0.1399827302,0.1675919443,0.2192769051,-0.1553871781,0.1405241936,0.0523834117,-0.1207046583,0.1294268966,-0.051441852,-0.1134570464,-0.1975087225,-0.2734824419,0.1838988364,-0.1971022189,-0.1978050768,0.5801024437,-0.2147762626,0.074397549,0.2620584369,0.1684101373,0.2045363486,-0.2048115581,0.1987678111,0.5227435231,0.2674139738,0.1252056807,-0.2163727283,-0.2284141034,-0.0392408296,-0.3450886905,0.2761880159,-0.2448882461,-0.178367734,0.0715071261,0.0493536294,0.0162273142,0.0047916044,-0.0439381488,0.0087107196,0.4345324636,0.1707940251,-0.0183573328,0.2114876658,-0.1735899448,-0.1470002234,-0.1979638934,-0.0012505206,0.2027682811,-0.3930261433,-0.0495891608,0.1050152108,-0.1647939384,0.1057785749,-0.172263056,-0.3507420719,0.4661415815,0.1852914244,-0.0895521417,0.1366778612,-0.0003281521,-0.3072820306,-0.1281069666,0.3641516566,0.0541132689,-0.1369956583,0.2023543417,0.2030726075,-0.110679388,-0.1670619398,-0.3106616735,-0.2953485847,-0.1063729525,0.2886683941,0.1519557238,0.1150512993,-0.2484764755,-0.0148870749,0.5208369493,-0.1647288203,0.5626708269,0.1504448801,-0.1578185707,-0.3267688453,0.2533043027,-0.0347385146,-0.1732836664,-0.1848648787,0.2981399,0.0767066926,0.0865779296,-0.0797210634,0.0952973589,0.1707717776,-0.0256922953,-0.0198636167,-0.0553642437,0.2363325804,0.2633557916,-0.1075721607,0.1206818521,-0.2376253307,-0.0400305279,-0.2965534925,-0.1210606694,-0.0640706867,-0.3717763424,-0.1162206233,0.2038687766,0.1534664482,0.0098027978,-0.0241119564,0.3252814412,0.0509022176,0.007238368,-0.0024153227,0.0057540494,-0.1513900757,-0.1015185714,0.4997426867,0.2514150143,-0.0847324133,0.1747264266,0.025348749,0.1252768636,0.1669304371,0.4507260323,-0.1114014834,0.1618746668,0.0395002887,0.2550393641,0.0195264257,-0.3203860819,-0.2272886336,0.378285557,0.163028419,0.0185769629,0.1553283632,0.01525026,0.2176691741,0.0091753835,-0.065287143,0.4854988456,0.1750310212,0.1409177184,-0.4725474119,-0.4832615852,-0.0594813675,0.3461815417,0.1197740883,-0.226567924,-0.2047182173,-0.1846186817,0.3050087392,-0.3265382349,0.098593615,0.1026033312,0.1075408608,0.2157734931,0.2896379232,-0.0459009036,-0.1346040815,0.2787211239,-0.3080227077,0.3445780575,0.277059257,-0.289414227,0.1209242195,0.030660117,-0.1056921631,-0.1341904551,0.1402307153,-0.1129245162,0.0192971174,0.0391548127,-0.1006057411,-0.0919625461,-0.2119908631,-0.0181858987,0.0214301459,-0.0155423917,-0.3207993209,0.0573524646,0.0892536417,0.2041374445,0.0866511241,0.1444229335,-0.0217300933,0.0549624152,0.0868354291,-0.0066821561,0.520585835,0.2361232042,0.061384175,-0.1906127334,-0.0632872432,-0.0609438457,0.0114322649,-0.2304022908,-0.2520048618,0.5083364844,0.0967105404,0.672422111,0.0362220407,-0.1513773799,0.1983903944,-0.092809923,0.3236225545,0.1621481329,0.096767351,0.0861774608,-0.15423356,0.1096738502,-0.2470085621,0.3957998157,0.0847710297,-0.1390085518,-0.1460458934,0.3455000222,0.1476966441,-0.4026895761,0.1203723028,-0.3250585198,-0.1167353988,0.234021306,0.0778978541,-0.0932746679,-0.0612068288,-0.1905838549,0.2561009526,0.0626242831,-0.1506821215,0.1598194391,0.0484115444,0.1712654829,-0.0974179357,-0.2836385965,-0.1396322846,0.0632073507,-0.0014035341,0.0198498406,-0.0967805237,-0.0367061719,0.4275031984,0.0264048632,-0.4450348914,-0.3590585291,0.2395977229,-0.1972917616,-0.1552642286,0.2693766952,-0.0099309999,0.0478624813,-0.105079107,-0.2813078761,-0.453338474,-0.3079473972,-0.2773779333,0.0130231027,0.1892403513,0.6149399281,0.2304699421,0.2074116468,-0.232235074,-0.0473324023,-0.4137030542,0.1432667375,-0.0947487503,0.0748724937,0.0620125197,0.1043494195,-0.2676288784,-0.2927021086,0.0963432938,-0.1845993698,0.073781468,-0.1074340865,-0.2597098947,-0.002387204,0.0752095431,-0.3678357601,-0.0981878564,-0.2722530067,0.1236127838,-0.1237497404,0.1488938332,-0.2418092638,0.0643003285,0.1365865469,-0.1227324083,0.0969901755,-0.1594769806,-0.0740229934,0.3361534476,0.0502667464,-0.2920596302,-0.0997658446,-0.1421093345,-0.0260732789,0.0583991557,-0.4986907244,0.1020059213,-0.4466369152,0.4375715554,0.1235352606,0.2664425075,0.3579042554,0.0905895829,-0.2238862216,0.0039343988,-0.0138119599,0.0772024393,-0.2698499262,-0.0257808845,-0.0899664909,0.027619997,-0.2029920518,1.1685949564,0.1803689152,0.060284365,0.2578519285,0.2749099433,0.1021383107,-0.1524999142,0.0782125816,0.1474120915,-0.1481910199,0.2420825809,0.2913126349,0.0530166514,-0.3329035342,-0.0764566436,0.0347350985,-0.1852193028,-0.3656598926,0.2082986087,0.4082408845,0.3489249051,-0.000184287,0.1110990196,-0.2183866501,-0.2367400527,0.2086192667,0.2917195559,0.0299605411,-0.24308002,-0.0582612641,0.0630525872,-0.7865275741,0.2313333005,0.1303321272,0.1013951078,-0.042363897,0.2195341289,0.3353596032,-0.3801006675,0.3456801772,-0.4950584173,-0.0331721231,0.3131635785,-0.2899520695,-0.0476486646,0.0994896218,-0.2029057145,-0.5299870968,0.373526454,0.2760537565,-0.4261974692,-0.3881303966,0.1481103301,-0.0957378075,-0.1505559236,-0.3641572595,-0.1820687056,-0.3787018061,-0.1126397923,0.245660916,-0.075872831,0.1786431521,-0.1438387334,0.0683246106,-0.22472471,0.3002116084,-0.0248447545,0.3001645505,0.0009156851,-0.023906745,0.3277934194,0.0894171894,-0.1370222121,0.2020126134,-0.0360448807,0.0479540117,0.3339440525,-0.2293162942,0.0398639068,0.0343011692,0.4406486452,0.093110092,0.3388126791,-0.0096827606,0.4912431836,-0.4976097941,0.0112885786,0.3263988197,0.010954014,-0.3418415487,-0.28491804,0.5926836729,0.0981411189,-0.2826972008,0.2403193861,0.6242871881,-0.0457813181,0.7988659739,0.4379025698,0.9731782079,-0.3236016929,0.182946831,0.0516982228,-0.1141508371,0.1772158295,-0.5803914666,0.0132795926,-0.3441277742,-0.0608092099,0.1457227916,0.0660938248,0.0998156369,0.079050459,-0.2112663835,0.0565325059,0.2663472891,-0.3910273314,-0.1033741832,0.6442167163,0.2642791271,-0.5839915872,-0.0688980371,0.0940648839,0.1049236283,-0.0070528071,-0.0335591212,-0.1499771476,0.0058964575,-0.0157538969,-0.4263886511,-0.078843385,0.2132332772,0.0951080024,-0.1442391872,-0.1015795991,0.3289506733,0.1994958818,0.0446215197,0.0828214213,-0.3486490846,0.4796303511,-0.2325749546,-0.0525605753,0.0644851252,-0.2691119015,0.0368863046,-0.1021122932,-0.3996189535,-0.3449296057,-0.1726768613,-0.1318245232,-0.3958916962,-0.1175536141,-0.1286820471,0.0277437363,0.17468445,0.1159787104,-0.3212335408,-0.1571043283,0.1689956784,-0.0084781172,-0.114650175,0.5580577254,0.1456538737,-0.1181377843,-0.1447271407,0.2413382381,0.2218540758,-0.1261492223,0.1711150706,0.3306896687,-0.3781175315,-0.253942579,-0.2330624908,0.1613133848,0.0913039371,-0.0462992638,0.1414882094,-0.4187159836,0.1323078573,0.185975045,-0.0451858267,-0.389975518,-0.1491930485,-0.0778695345,-0.1537671685,0.1333177537,-0.1149328128,0.1478138864,-0.1276235729,-0.0718946084,-0.0769596919,0.5021989346,-0.3233356774,-0.0379970931,-0.3658128977,0.1154035255,-0.2046507299,-0.249543488,0.1450018585,0.1177471727,0.0801378489,0.0474302173,-0.1114959121,-0.2268679738,-0.1165582389,0.1802139133,-0.0278919339,-0.2174326032,0.2300642282,0.0541063175,-0.4773789048,-0.5050476789,0.1447993964,-0.0598669462,-0.1000759378,0.0180108994,0.0252684914,0.097836785,-0.1535309702,0.2102260292,0.0410500988,0.3504740596,0.0708817318,-0.0282138586,0.1094341874,-0.2575258315,0.2309644222,0.1266764253,0.2536876202,0.031752184,0.4446632266,0.0696464553,-0.3849517405,0.0196001325,0.4330316186,0.4753473103,0.1411868632,-0.0167593211,0.1932483464,0.179166615,-0.213744536,-0.1301713288,0.3691519797,-0.1887833625,0.1737031341,0.0434081927,0.4589833319,0.1746049076,-0.1596093029,0.053206794,-0.0747288391,-0.3618654311,-0.1350294054,0.0654059872,0.2945966125,-0.0154725444,0.1813447773,0.148032099,0.2206320018,0.4996941388,0.127193287,0.4412364364,-0.0816404596,0.5757459998,-0.0370632485,-0.1976613104,0.334217906,0.2281460762,0.038185019,0.0533296727,-0.0587760545,-0.1067265496,0.1567356139,-0.3506704271,-0.1394798011,0.2525772452,-0.1414334923,-0.3310838044,0.1789602339,-0.0030862256,0.1212244406,0.2332031876,0.0462858304,-0.0102013163,0.2023239136,0.1436454356,-0.1063007712,-0.3107666075,0.0503134616,-0.0770487264,0.6449183226,0.0399300978,-0.2119450271,0.1370384842,-0.1598820388,-0.0736915544,0.206590116,0.2697362602,0.1512991935,-0.6558002234,0.0792783648,0.0481703654,0.0943072066,0.0159029663,0.2695036232,0.0123863807,0.3510616124,0.2135608047,0.0070445086,-0.2483609021,-0.7480022907,-0.1176292971,0.2216346264,0.0305377729,-0.038206771,0.1356329173,-0.1258384734,0.0279582627,0.1113023013,-0.2532184124,0.3649376035,0.2434102148,-0.0819707066,0.0315249078,-0.0477426313,0.0999827757,-0.0567487031,0.8836591244,0.1941982508,0.1804577708,-0.2134257704,-0.1510017365,-0.5924735069,-0.0190747529,-0.3248521984,-0.0911901742,0.5681339502,0.3624492586,0.0040307147,0.1747616976,0.2194930762,-0.0107988231,0.1119870394,0.4059708416,-0.2775332332,0.2101975083,-0.2691062391,-0.1409318,-0.0756237209,-0.3053850234,0.4251206517,0.2309118956,-0.0066821855,-0.0091214916,0.1979620308,0.1358774751,0.1123121455,0.2765578032,0.2952306867,0.1052660793,-0.1817988157,-0.10903164,-0.2036578655,0.0550288782,-0.0732898489,0.0469975881,-0.0669147521,0.0866826251,-0.1295159757,-0.3051650226,-0.2181547731,0.1757832021,-0.0741405487,0.136987552,-0.2045065761,-0.1023224443,0.2649249136,0.3677134812,0.0878140926,0.2704728842,-0.0025316062,-0.0173629429,-0.4523029625,-0.3206740618,0.3238148391,-0.37162745,-0.1630413979,-0.0864442065,0.0670392066,0.1396896094,0.2970749438,-0.5577389002,-0.0970949382,0.1017499715,0.1856281161,-0.4030950367,0.0158737898,-0.0454704575,-0.2367836535,-0.2156967074,-0.0411097705,0.0656712651,-0.1367792338,-0.2975480855,-0.2228102088]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1670","title":"wiki_dpr pre-processing performance","comments":"I'd be happy to contribute something when I get the time, probably adding multiprocessing and \/ or cython support to wiki_dpr. I've written cythonized apache beam code before as well.\r\n\r\nFor sharded index building, I used the FAISS example code for indexing 1 billion vectors as a start. I'm sure you're aware that the documentation isn't great, but the source code is fairly easy to follow.","body":"I've been working with wiki_dpr and noticed that the dataset processing is seriously impaired in performance [1]. It takes about 12h to process the entire dataset. Most of this time is simply loading and processing the data, but the actual indexing is also quite slow (3h).\r\n\r\nI won't repeat the concerns around multiprocessing as they are addressed in other issues (#786), but this is the first obvious thing to do. Using cython to speed up the text manipulation may be also help. Loading and processing a dataset of this size in under 15 minutes does not seem unreasonable on a modern multi-core machine. I have hit such targets myself on similar tasks. Would love to see this improve.\r\n\r\nThe other issue is that it takes 3h to construct the FAISS index. If only we could use GPUs with HNSW, but we can't. My sharded GPU indexing code can build an IVF + PQ index in 10 minutes on 20 million vectors. Still, 3h seems slow even for the CPU.\r\n\r\nIt looks like HF is adding only 1000 vectors at a time by default [2], whereas the faiss benchmarks adds 1 million vectors at a time (effectively) [3]. It's possible the runtime could be reduced with a larger batch. Also, it looks like project dependencies ultimately use OpenBLAS, but this is known to have issues when combined with OpenMP, which HNSW does [3]. A workaround is to set the environment variable `OMP_WAIT_POLICY=PASSIVE` via `os.environ` or similar.\r\n\r\nReferences:\r\n[1] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py\r\n[2] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/search.py\r\n[3] https:\/\/github.com\/facebookresearch\/faiss\/blob\/master\/benchs\/bench_hnsw.py\r\n[4] https:\/\/github.com\/facebookresearch\/faiss\/issues\/422","comment_length":66,"text":"wiki_dpr pre-processing performance\nI've been working with wiki_dpr and noticed that the dataset processing is seriously impaired in performance [1]. It takes about 12h to process the entire dataset. Most of this time is simply loading and processing the data, but the actual indexing is also quite slow (3h).\r\n\r\nI won't repeat the concerns around multiprocessing as they are addressed in other issues (#786), but this is the first obvious thing to do. Using cython to speed up the text manipulation may be also help. Loading and processing a dataset of this size in under 15 minutes does not seem unreasonable on a modern multi-core machine. I have hit such targets myself on similar tasks. Would love to see this improve.\r\n\r\nThe other issue is that it takes 3h to construct the FAISS index. If only we could use GPUs with HNSW, but we can't. My sharded GPU indexing code can build an IVF + PQ index in 10 minutes on 20 million vectors. Still, 3h seems slow even for the CPU.\r\n\r\nIt looks like HF is adding only 1000 vectors at a time by default [2], whereas the faiss benchmarks adds 1 million vectors at a time (effectively) [3]. It's possible the runtime could be reduced with a larger batch. Also, it looks like project dependencies ultimately use OpenBLAS, but this is known to have issues when combined with OpenMP, which HNSW does [3]. A workaround is to set the environment variable `OMP_WAIT_POLICY=PASSIVE` via `os.environ` or similar.\r\n\r\nReferences:\r\n[1] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py\r\n[2] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/search.py\r\n[3] https:\/\/github.com\/facebookresearch\/faiss\/blob\/master\/benchs\/bench_hnsw.py\r\n[4] https:\/\/github.com\/facebookresearch\/faiss\/issues\/422\nI'd be happy to contribute something when I get the time, probably adding multiprocessing and \/ or cython support to wiki_dpr. I've written cythonized apache beam code before as well.\r\n\r\nFor sharded index building, I used the FAISS example code for indexing 1 billion vectors as a start. I'm sure you're aware that the documentation isn't great, but the source code is fairly easy to follow.","embeddings":[-0.2264903635,-0.1561907828,-0.1201761216,0.0668545812,-0.1370591819,-0.0890740529,0.0256168693,0.334238261,0.1915272623,0.0689785779,0.0395436101,-0.0971363634,0.3024311662,0.1471910179,-0.3419129848,-0.2786286175,0.2023194581,0.0613186844,0.1484082639,0.0297786724,-0.2428166419,-0.0547438562,-0.4051635563,-0.1212011278,-0.119506523,-0.2362726927,-0.0799640715,-0.1265879571,-0.1769073606,-0.5026086569,0.2207076252,0.5021685362,0.1483295709,0.1311967075,-0.0001081542,-0.1195450798,0.1936886162,0.2271693647,0.1224437505,0.6476590633,0.0664061159,-0.1502612829,0.1650483757,-0.2387850881,0.0332406126,-0.3853898346,-0.122053422,0.1576739997,0.2102280259,-0.1659050584,0.1324095279,0.0579918399,-0.1257512271,0.1390791982,-0.0618019663,-0.0908403471,-0.2100697458,-0.2791371644,0.1850635409,-0.2045671642,-0.2020124942,0.5718423724,-0.2305089682,0.0829863176,0.2626013756,0.1567495167,0.2092082649,-0.2050276399,0.1888362169,0.533824861,0.2657999694,0.1225182116,-0.2275393754,-0.23706384,-0.027647024,-0.3344913125,0.279099375,-0.2398142517,-0.1816463619,0.087790437,0.05900703,0.0167671703,-0.0018296944,-0.0482963026,0.010347249,0.4794566333,0.1762813926,-0.0241318401,0.2075946778,-0.1637800485,-0.1426846534,-0.1857211739,0.0056334799,0.2005604655,-0.391604811,-0.0591739677,0.087165013,-0.177181229,0.1087731794,-0.1624542773,-0.3400701284,0.4601314366,0.1689214855,-0.0850517601,0.1326358318,-0.0110870143,-0.3062150776,-0.1471126676,0.3762043118,0.0522954911,-0.117648311,0.2342726737,0.2008639723,-0.1009126306,-0.1655819714,-0.3135617375,-0.2978515327,-0.0925514102,0.2849405706,0.1605560184,0.1177737638,-0.256334722,-0.0207051486,0.5065271854,-0.1621720046,0.5769702196,0.1474592388,-0.1734802723,-0.304023087,0.2322988808,-0.0271006376,-0.1695578396,-0.1649038643,0.2978241444,0.0679754764,0.104746148,-0.0775455981,0.0752315074,0.1736094505,-0.0095366463,-0.0072202981,-0.0319361798,0.2566510141,0.2684512734,-0.1274978071,0.0969416574,-0.2341869175,-0.0310036857,-0.3091947734,-0.1136879176,-0.0527657308,-0.3701983988,-0.1326064467,0.1966398209,0.14800933,-0.0075231032,-0.0231632758,0.3451903164,0.0442126617,-0.0044656643,0.0304510053,0.0039309552,-0.1693896055,-0.1058666036,0.5083746314,0.2476486564,-0.0840832293,0.1602045149,0.0268207882,0.1279104501,0.160217762,0.4359187782,-0.1009852514,0.1620683223,0.0551536605,0.2438957542,0.0188094378,-0.3310874999,-0.236484468,0.3677527905,0.1674934626,0.0178088266,0.1697303653,0.0338237472,0.2224253863,-0.0007783477,-0.0635619313,0.5007228851,0.1689953208,0.1448026597,-0.4762882888,-0.5098897219,-0.0542118438,0.3578077257,0.1260609478,-0.2346327454,-0.1879566759,-0.1784204245,0.3020689785,-0.3102788627,0.1019865498,0.0898889154,0.1041042805,0.194513768,0.2758499086,-0.0493191928,-0.1281998754,0.2644064724,-0.3105755746,0.3434853554,0.2918092906,-0.2907407284,0.1353847682,0.0444958396,-0.0907858908,-0.1451968849,0.1301653683,-0.101052098,0.0234394185,0.0444510952,-0.1051182523,-0.0882676542,-0.2204465419,-0.0174610801,0.0375105478,-0.0215444453,-0.2987062037,0.0531780496,0.0902995393,0.2185144275,0.0656749681,0.1421156824,-0.0148042403,0.0567219518,0.0849608406,-0.0036726377,0.5223878026,0.2626947761,0.0666323453,-0.1789102703,-0.0565042235,-0.0545864329,-0.0102756908,-0.2095191479,-0.2721387446,0.5150545835,0.0993194133,0.6662998199,0.0335601419,-0.1411223412,0.1910708398,-0.0800541118,0.2929942906,0.1612861753,0.0908899903,0.08388751,-0.1554394364,0.1086618304,-0.2457912862,0.3985975385,0.0887197182,-0.1418863386,-0.1238221675,0.3410217166,0.1518612355,-0.431748718,0.1378454566,-0.3139727414,-0.1318142861,0.239617303,0.0760304257,-0.0933852643,-0.06432987,-0.1795387566,0.2557094097,0.0687241405,-0.1604065001,0.1599912345,0.048968602,0.1897314638,-0.1004807502,-0.3037284315,-0.150357917,0.0547477938,0.0158069301,0.0114368899,-0.0945765972,-0.0329721645,0.4362289608,0.0277712569,-0.4427422285,-0.3615911007,0.2546587288,-0.2031912953,-0.1740405858,0.2699273229,-0.0333381668,0.0738854781,-0.1160988063,-0.3093850017,-0.4688055217,-0.3147058487,-0.2602574229,0.0121996589,0.1835529804,0.6119245291,0.2436839044,0.2118457407,-0.2368309051,-0.0472815,-0.4198971093,0.1470544189,-0.0797184184,0.0652085841,0.0760584027,0.0987823457,-0.2667643726,-0.3064781725,0.0766442046,-0.200657323,0.0864774957,-0.1277013719,-0.254714191,0.0155785289,0.0817716196,-0.352517575,-0.0994156301,-0.2811506093,0.131622389,-0.1253048927,0.1471805871,-0.2470509261,0.0699926689,0.148690328,-0.1375655681,0.0945608765,-0.1460377723,-0.079986386,0.3313512802,0.0447334126,-0.2690945864,-0.1204663813,-0.1354925185,-0.0198030267,0.0591920093,-0.5038510561,0.0904157087,-0.431381464,0.4464229047,0.108937338,0.2679252326,0.3700772226,0.107183136,-0.2281227708,0.0174601376,-0.0057210559,0.0528921708,-0.2638611495,-0.049014058,-0.0867943317,0.0085858125,-0.1958406866,1.1640775204,0.1669705361,0.066325888,0.2571604252,0.2840445042,0.1098176464,-0.1424652785,0.1016769186,0.1536926776,-0.1440338492,0.2449607402,0.2913011611,0.0464445278,-0.3404867351,-0.0658448115,0.0091318237,-0.1790581346,-0.3628080189,0.1982976496,0.4147167802,0.3583087027,-0.0132025778,0.1047783345,-0.2162606567,-0.2308458686,0.2271834016,0.293974936,0.0361808166,-0.2375804633,-0.045625288,0.0703794211,-0.7937565446,0.2271479517,0.1352057457,0.1062423289,-0.044931367,0.2049987167,0.3181872666,-0.3824547827,0.3311475217,-0.5110362172,-0.034223929,0.3105385005,-0.2685112357,-0.026773151,0.1010705829,-0.1987965852,-0.5399523973,0.3662314713,0.2730425894,-0.404184401,-0.3946310282,0.1638050824,-0.105685845,-0.1411682963,-0.349920094,-0.1689430922,-0.3743134141,-0.1158240065,0.2541350126,-0.0730223879,0.1617851555,-0.1162712052,0.0757676587,-0.2403760552,0.3053177595,-0.0359443128,0.2997192144,-0.0203903578,-0.0206663683,0.3212929666,0.0841277987,-0.1593244523,0.2003445774,-0.0690931901,0.0427263193,0.3371325135,-0.2086455524,0.0539321713,0.041836191,0.4395431876,0.0943160057,0.3287370205,0.0064216875,0.4958718717,-0.4995758832,0.0186160393,0.316724211,0.005829467,-0.3524342477,-0.3060558438,0.6003261209,0.1124924272,-0.2819769084,0.2517493367,0.6379445791,-0.0435167365,0.8059248924,0.4664424658,0.9706417918,-0.3135972321,0.2017132938,0.0578306504,-0.1052981168,0.186831668,-0.5955249071,0.0106223654,-0.3307252824,-0.0507868752,0.1385859847,0.071747154,0.107087411,0.0696205869,-0.2108366191,0.0408564247,0.2892589271,-0.3912501931,-0.0888436735,0.651011467,0.2506239414,-0.5994696021,-0.062719509,0.0864966437,0.1103718579,-0.0299356598,-0.0331214145,-0.1477195323,0.0312305354,-0.0010646326,-0.4338219166,-0.0924841762,0.2150058448,0.0937632173,-0.1425295919,-0.0979944542,0.3362613022,0.2008272111,0.0377694778,0.0891830027,-0.3503617346,0.4732806683,-0.228855744,-0.0520562753,0.0807799324,-0.2716549933,0.0637093112,-0.097920306,-0.3850010335,-0.3434790671,-0.171693638,-0.1527827084,-0.3854253292,-0.1245309114,-0.130957678,0.0431425981,0.1877592653,0.1281505227,-0.327159375,-0.1681949198,0.1640872061,-0.0141303027,-0.1119824648,0.5528222919,0.1349529624,-0.1208633929,-0.1516037583,0.2280513346,0.22843422,-0.1255816221,0.1546091139,0.3147464991,-0.3647041917,-0.2627891302,-0.224992469,0.1548174471,0.1012103558,-0.0364480913,0.1330310106,-0.4334462583,0.1299832314,0.1956923902,-0.0455755554,-0.3782905042,-0.1649549305,-0.0718958825,-0.1449116617,0.1371908188,-0.1202352643,0.1591122597,-0.1099165529,-0.0799399763,-0.0874128565,0.4848991632,-0.3216123581,-0.0750118718,-0.3507187963,0.0784443468,-0.2023091465,-0.2345542759,0.1431981027,0.1367759109,0.0717569292,0.049555175,-0.1006561667,-0.2216841877,-0.1346908659,0.1836602986,-0.0282211099,-0.21051009,0.2355115116,0.05850292,-0.4899175167,-0.5152711868,0.1423525959,-0.0661608204,-0.0958241299,0.0199262202,0.0216853097,0.1306713372,-0.1422111392,0.2207502276,0.0553043671,0.3395251632,0.0574595556,-0.0085228831,0.1168655455,-0.2495571673,0.2408250719,0.1467067748,0.2563509643,0.0192094445,0.4395847321,0.0862904266,-0.4105634987,0.0197775979,0.4301698208,0.4558754265,0.1562863886,-0.0293102879,0.1915709674,0.1753004789,-0.2196451873,-0.1398758739,0.3598823845,-0.1931513101,0.1824486852,0.0544721745,0.4774359763,0.1810075194,-0.1471675038,0.0437454768,-0.0729643553,-0.3672818244,-0.1262561083,0.0754416957,0.3041472435,-0.0196465962,0.1780897975,0.152285248,0.2421380877,0.4696971476,0.1346750408,0.4433130324,-0.0614603981,0.5766792297,-0.0347284488,-0.1724632978,0.3463389277,0.2108273208,0.0394405313,0.0757887289,-0.056696143,-0.117715396,0.1503635049,-0.3285142183,-0.1473157108,0.2560689449,-0.1253513247,-0.3405516148,0.1767050922,-0.0147075504,0.1190310344,0.2388461977,0.0501365811,-0.015960088,0.2008631825,0.1352664381,-0.0882199854,-0.3003579974,0.0668340847,-0.0721721947,0.649107933,0.0258754678,-0.2342167944,0.1350666881,-0.1706244498,-0.0879401192,0.2100565284,0.2735480368,0.1326279491,-0.6634915471,0.0866360664,0.0214948785,0.1066218093,0.0001624249,0.2667883337,0.0129998103,0.3445979357,0.1935042441,0.0065663229,-0.2392437309,-0.7350765467,-0.1268932223,0.197494626,0.0194032714,-0.0277862325,0.1513046771,-0.1215536445,0.0380776301,0.1309849918,-0.2416858375,0.3493694961,0.2266610712,-0.073034279,0.0265201703,-0.0334448777,0.0996011645,-0.055462759,0.880910635,0.1900393814,0.1785145849,-0.1936183572,-0.1442979127,-0.5910930037,-0.0252808183,-0.2965450287,-0.1120122299,0.5680955052,0.357169807,0.0106257014,0.1882096678,0.2214713991,-0.0089690685,0.1164626032,0.3898643851,-0.2631464005,0.2222468257,-0.2608741224,-0.1377269775,-0.0771225765,-0.3192447126,0.415912509,0.2360437214,-0.0172014926,-0.0248354375,0.1768144518,0.1454500705,0.1142577529,0.2799149454,0.2969346344,0.0768237785,-0.1839601547,-0.1251140386,-0.2327936292,0.0658806413,-0.0831589103,0.0293316115,-0.0746744722,0.0847039744,-0.1430925876,-0.3185128868,-0.217784211,0.1827890426,-0.0721334442,0.1368626952,-0.198474288,-0.1087496355,0.2646808922,0.3724131882,0.085557282,0.2582963407,0.0161590148,-0.0237191338,-0.4486636519,-0.3001714051,0.3009140193,-0.3872106075,-0.1610067934,-0.1063421145,0.0541980676,0.1399506629,0.2979110479,-0.548587501,-0.114374578,0.099874422,0.2049383223,-0.3972114325,0.014841456,-0.0455899723,-0.2327032536,-0.226829201,-0.0411338657,0.0479602925,-0.1373187304,-0.3126225471,-0.2336394489]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1670","title":"wiki_dpr pre-processing performance","comments":"Nice thanks ! That would be awesome to make its construction faster :) ","body":"I've been working with wiki_dpr and noticed that the dataset processing is seriously impaired in performance [1]. It takes about 12h to process the entire dataset. Most of this time is simply loading and processing the data, but the actual indexing is also quite slow (3h).\r\n\r\nI won't repeat the concerns around multiprocessing as they are addressed in other issues (#786), but this is the first obvious thing to do. Using cython to speed up the text manipulation may be also help. Loading and processing a dataset of this size in under 15 minutes does not seem unreasonable on a modern multi-core machine. I have hit such targets myself on similar tasks. Would love to see this improve.\r\n\r\nThe other issue is that it takes 3h to construct the FAISS index. If only we could use GPUs with HNSW, but we can't. My sharded GPU indexing code can build an IVF + PQ index in 10 minutes on 20 million vectors. Still, 3h seems slow even for the CPU.\r\n\r\nIt looks like HF is adding only 1000 vectors at a time by default [2], whereas the faiss benchmarks adds 1 million vectors at a time (effectively) [3]. It's possible the runtime could be reduced with a larger batch. Also, it looks like project dependencies ultimately use OpenBLAS, but this is known to have issues when combined with OpenMP, which HNSW does [3]. A workaround is to set the environment variable `OMP_WAIT_POLICY=PASSIVE` via `os.environ` or similar.\r\n\r\nReferences:\r\n[1] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py\r\n[2] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/search.py\r\n[3] https:\/\/github.com\/facebookresearch\/faiss\/blob\/master\/benchs\/bench_hnsw.py\r\n[4] https:\/\/github.com\/facebookresearch\/faiss\/issues\/422","comment_length":13,"text":"wiki_dpr pre-processing performance\nI've been working with wiki_dpr and noticed that the dataset processing is seriously impaired in performance [1]. It takes about 12h to process the entire dataset. Most of this time is simply loading and processing the data, but the actual indexing is also quite slow (3h).\r\n\r\nI won't repeat the concerns around multiprocessing as they are addressed in other issues (#786), but this is the first obvious thing to do. Using cython to speed up the text manipulation may be also help. Loading and processing a dataset of this size in under 15 minutes does not seem unreasonable on a modern multi-core machine. I have hit such targets myself on similar tasks. Would love to see this improve.\r\n\r\nThe other issue is that it takes 3h to construct the FAISS index. If only we could use GPUs with HNSW, but we can't. My sharded GPU indexing code can build an IVF + PQ index in 10 minutes on 20 million vectors. Still, 3h seems slow even for the CPU.\r\n\r\nIt looks like HF is adding only 1000 vectors at a time by default [2], whereas the faiss benchmarks adds 1 million vectors at a time (effectively) [3]. It's possible the runtime could be reduced with a larger batch. Also, it looks like project dependencies ultimately use OpenBLAS, but this is known to have issues when combined with OpenMP, which HNSW does [3]. A workaround is to set the environment variable `OMP_WAIT_POLICY=PASSIVE` via `os.environ` or similar.\r\n\r\nReferences:\r\n[1] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py\r\n[2] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/search.py\r\n[3] https:\/\/github.com\/facebookresearch\/faiss\/blob\/master\/benchs\/bench_hnsw.py\r\n[4] https:\/\/github.com\/facebookresearch\/faiss\/issues\/422\nNice thanks ! That would be awesome to make its construction faster :) ","embeddings":[-0.2318795472,-0.1633458883,-0.1273872852,0.0676816031,-0.1257117242,-0.0887089297,0.0157347564,0.3411875069,0.1872670949,0.0617169999,0.0255135056,-0.0995670557,0.2901466787,0.1591228396,-0.3300855756,-0.2848236263,0.2114918828,0.0640387312,0.1474420577,0.019559741,-0.2316578329,-0.044841677,-0.4052048326,-0.114139244,-0.122649014,-0.2355428189,-0.0896494016,-0.1171084642,-0.2057312727,-0.5016263723,0.2147744894,0.5130922198,0.1529366076,0.1363784969,-0.0001084032,-0.1110991463,0.2000096142,0.231851235,0.1145225465,0.6541631818,0.0636401102,-0.1518353373,0.1586824358,-0.2378550023,0.0315564834,-0.3885970712,-0.1275275052,0.1579280943,0.2077937573,-0.1720529348,0.13291201,0.0550700985,-0.1112723127,0.1340446472,-0.0559436455,-0.0834023878,-0.2229518443,-0.2891471982,0.1877396852,-0.2127058208,-0.2015827745,0.58578825,-0.2353854924,0.0716937482,0.2801270187,0.1579944044,0.2157767415,-0.1971387714,0.1781517118,0.5282712579,0.2601043582,0.111258097,-0.2368027717,-0.2400523573,-0.0429853201,-0.342202872,0.2677705288,-0.2373063117,-0.1902873069,0.0848125964,0.0484610721,0.008471814,-0.0004502972,-0.0514163375,0.0092774471,0.4775907099,0.1702142358,-0.0268835854,0.2124189138,-0.1617868692,-0.1474876255,-0.1860709786,0.0085436525,0.1914845556,-0.3870609701,-0.0780896321,0.0927761197,-0.1582889259,0.1184421778,-0.1542232931,-0.3467411697,0.4518052936,0.1752861291,-0.0933610424,0.1297524422,-0.005095412,-0.2984670699,-0.1445990056,0.3730093241,0.0554404929,-0.1084550321,0.2364843935,0.2187488079,-0.0950943157,-0.1949125975,-0.309728384,-0.2924931347,-0.0990269929,0.2865154147,0.1592615396,0.1280232519,-0.257745415,-0.0406779498,0.5231061578,-0.1696135849,0.5679964423,0.1358149499,-0.1637239605,-0.3046422005,0.2395743579,-0.039300561,-0.1675253063,-0.1614587009,0.2968652546,0.0766699314,0.0984948575,-0.081587404,0.0839075446,0.1874306947,-0.0131710833,0.0050523099,-0.0285720825,0.2554186881,0.2728478611,-0.11553213,0.0958262607,-0.2132821083,-0.0429379568,-0.3091226816,-0.1123294681,-0.0497058406,-0.3710502684,-0.1375023723,0.1933423877,0.1350494474,-0.0138746751,-0.0367150679,0.3477531672,0.0384723581,0.0093777403,0.0331538282,0.0124304621,-0.1650294513,-0.0969632417,0.5097306371,0.2461217493,-0.0840916261,0.1548502594,0.0348874517,0.1156728491,0.1688044816,0.4428002238,-0.0970725045,0.1533998102,0.0535579696,0.2547484934,0.0231049657,-0.3281795382,-0.2199200839,0.3778908253,0.1665830016,0.0112062609,0.1714670956,0.0255196504,0.2197301835,0.0005339556,-0.0690525249,0.4742009342,0.1622311026,0.1454572529,-0.4817976952,-0.5109411478,-0.0448342562,0.3609303534,0.1384111792,-0.2336106002,-0.1913503855,-0.1772473603,0.3152966499,-0.3165588379,0.1103295013,0.0892324671,0.1024338529,0.2066496313,0.2870973945,-0.0534526445,-0.1386283934,0.2572834492,-0.3068487048,0.3497416079,0.2761106193,-0.2887627482,0.1311763972,0.0376550965,-0.0937155783,-0.1481046677,0.1308268607,-0.1168851927,0.0357082896,0.0547654293,-0.1201132014,-0.0924624577,-0.2198781818,-0.0225060657,0.0375098437,-0.0174355917,-0.3033360839,0.064588882,0.0784540623,0.2213786542,0.0579347089,0.1385455132,-0.0071469541,0.0511829853,0.0749275386,-0.0173947271,0.518862009,0.2479435056,0.0659699589,-0.1907912195,-0.0508986786,-0.0605635531,-0.0032296288,-0.2214653343,-0.2592741847,0.5167284012,0.0858862251,0.668425262,0.0349187106,-0.1513179839,0.1868754178,-0.0781709924,0.27561903,0.1594088078,0.1061212942,0.0823388025,-0.1651682407,0.0797619075,-0.2548782527,0.3942470551,0.0943051055,-0.1368441433,-0.1250788271,0.3314998746,0.1502057612,-0.4074688256,0.1359499842,-0.3245105743,-0.1146673337,0.2377145588,0.0712408423,-0.082198441,-0.0482552573,-0.1773051322,0.262324363,0.068676874,-0.1564032882,0.1667716801,0.0314658768,0.1884938926,-0.1071376652,-0.3112437129,-0.1569243819,0.0542974956,0.0107906414,0.0028774273,-0.065827094,-0.0440344401,0.4319081604,0.0250311941,-0.4408817589,-0.3584357202,0.2601487041,-0.2100221813,-0.1636828482,0.2841120362,-0.0329459719,0.0762379467,-0.1264891624,-0.3078449368,-0.4645067155,-0.3170798123,-0.248818785,0.0132162822,0.192607671,0.6036788225,0.2483331561,0.2200794816,-0.237392664,-0.0516549647,-0.4258534014,0.1425076574,-0.0970106274,0.0585978702,0.0739234835,0.1106786355,-0.2637264729,-0.3140029311,0.0851816162,-0.1886493713,0.0632204041,-0.1345437616,-0.2544957697,0.0121663669,0.0669726878,-0.3527224064,-0.0975862667,-0.2743484676,0.1429994553,-0.1092786342,0.1338851005,-0.2444852889,0.078387022,0.1428751647,-0.1425639242,0.096160762,-0.1478597373,-0.0851158202,0.3233517706,0.0418121777,-0.2701201737,-0.1257798374,-0.1346145719,-0.0249628332,0.0466806889,-0.5134690404,0.1002744809,-0.4299108386,0.4347678125,0.1152651757,0.2719830573,0.3662225008,0.1111784577,-0.2252222002,0.0244157426,-0.0120108156,0.0287362933,-0.2713853121,-0.0510435216,-0.0881780535,0.0216038693,-0.1973343641,1.1873936653,0.1820472181,0.0868717581,0.2510290444,0.2780432701,0.1054804549,-0.1531313509,0.0748835728,0.1587738991,-0.1548447162,0.2487384975,0.2862161696,0.0485797748,-0.354518503,-0.0821707696,0.0353741199,-0.1858396679,-0.3550292552,0.190037936,0.4087063372,0.3410896659,-0.0128190676,0.0941332132,-0.2041866332,-0.2268597484,0.2347063869,0.3078115582,0.0216735136,-0.235052675,-0.0572070703,0.0807366818,-0.8000960946,0.2205490321,0.1284079701,0.0774195567,-0.0418416262,0.2058659494,0.3268979788,-0.3908889592,0.3424009085,-0.4953644574,-0.0376280174,0.3090656996,-0.2671429217,-0.0311206114,0.1148487926,-0.1918589771,-0.5422781706,0.3624515235,0.2930097282,-0.4010971487,-0.3888965845,0.1606419981,-0.1194810122,-0.1378680021,-0.350006938,-0.1862537712,-0.379275918,-0.1343246698,0.2573320866,-0.0487571619,0.1741050035,-0.1222213283,0.0709896013,-0.2246670723,0.3019061685,-0.0343511105,0.2882450223,-0.0383480191,-0.014738081,0.3213673234,0.0841618031,-0.1546815932,0.1910128295,-0.0538536794,0.0346633941,0.331417948,-0.2054711133,0.0571645498,0.0417812057,0.4502899051,0.0958505124,0.3248740733,0.0181003939,0.5038387179,-0.5029913783,0.0108438209,0.3101133704,-0.0016629448,-0.3724009395,-0.3109590411,0.6045933962,0.1139979959,-0.2738061249,0.2547681332,0.6593227983,-0.0315924175,0.7872525454,0.4603203833,0.9856671691,-0.3197063804,0.2168227732,0.0471273363,-0.1069233343,0.1985638291,-0.6014535427,0.0240456276,-0.3375656605,-0.0594403334,0.1335201263,0.0692974031,0.1002259925,0.069216162,-0.2128610313,0.0377188139,0.2949275672,-0.3882154524,-0.0900771543,0.6486504674,0.2490663677,-0.5918107629,-0.0641975924,0.0892680362,0.1236478984,-0.0096676704,-0.032952182,-0.1488458514,0.0230558831,-0.0090645868,-0.4399823844,-0.0937442854,0.2081181854,0.0960430652,-0.1504496634,-0.1060330644,0.3365857303,0.2083316445,0.0465735868,0.085882768,-0.358081013,0.4610836208,-0.2289597243,-0.0559002161,0.079835549,-0.2707687616,0.0605372041,-0.1074079573,-0.3731227815,-0.3459534049,-0.1685772687,-0.160092622,-0.4023573697,-0.1099916026,-0.1279063672,0.041284699,0.1937730908,0.1276643127,-0.306012094,-0.1592543125,0.1621572971,-0.0062498208,-0.109861739,0.5462223291,0.1473308951,-0.1128806025,-0.1486033648,0.2269417644,0.2331135571,-0.1390191466,0.1622928381,0.3187988102,-0.3738892674,-0.2679356933,-0.2200105041,0.1700532883,0.0937741175,-0.0383657143,0.1376073658,-0.4239279032,0.1290637106,0.2114301026,-0.03544515,-0.3637802601,-0.1406973898,-0.0633297861,-0.1500644684,0.1435389072,-0.1309657544,0.1705197394,-0.106049858,-0.0653440878,-0.0825865045,0.4688975513,-0.3209694326,-0.0614215024,-0.3466937244,0.0727935433,-0.188404575,-0.2265416831,0.1384356469,0.1393159777,0.0746078119,0.0553300492,-0.1177409664,-0.2225760072,-0.1293331832,0.1832370907,-0.0115592936,-0.2057181746,0.2456824034,0.0795391649,-0.4934978187,-0.4994137883,0.1606570184,-0.0577320047,-0.0910572857,0.0247230884,0.0368502885,0.1109039858,-0.1460771114,0.2085186392,0.0710162297,0.3521419764,0.0705292374,-0.0107005611,0.1145542637,-0.2416501492,0.2522212863,0.158439815,0.2457416058,0.0323527157,0.4416692555,0.0669105425,-0.4044258296,0.0169186629,0.4389098883,0.4633316994,0.148161456,-0.020635521,0.1948250532,0.1707723141,-0.2112469524,-0.1419454664,0.3636691868,-0.1944027394,0.1724708378,0.0558043122,0.4619053304,0.1797498465,-0.1496186554,0.0515748076,-0.0670887828,-0.3777537644,-0.1270960271,0.0673969314,0.2989116609,-0.0239993669,0.1773394793,0.1540950835,0.2271533161,0.4735501409,0.1278076619,0.4462398291,-0.0716930553,0.5769329667,-0.0390099846,-0.1757934242,0.3473793268,0.2176821381,0.0300713815,0.0722852051,-0.0644396022,-0.1147946492,0.1626926959,-0.3407691121,-0.1404659301,0.2563560307,-0.1320921183,-0.338011235,0.1731590629,-0.0210183281,0.1225554645,0.242792204,0.0406568423,-0.0254542381,0.2076282799,0.1349643618,-0.0983606204,-0.2980774343,0.0608019754,-0.0721971542,0.6442844272,0.0236689355,-0.2409149259,0.1434140801,-0.1719368398,-0.0985332206,0.2129693925,0.2823128402,0.1345399022,-0.6549844742,0.0867626444,0.0207808949,0.107198976,0.0107509727,0.2637124062,0.0298817828,0.3383969069,0.2010136247,0.0050290274,-0.2323578298,-0.7312870622,-0.1145754084,0.2129694521,0.0197297744,-0.0279247835,0.1377134323,-0.1125159785,0.0250829831,0.1360195279,-0.2605473697,0.3457467258,0.2331478447,-0.0711879507,0.026733866,-0.0427551866,0.0981942713,-0.0581381246,0.8748050928,0.2115671039,0.1841970235,-0.1979470998,-0.1487395912,-0.5790196657,-0.0241489969,-0.3100542724,-0.0870132819,0.5665012598,0.3728150725,0.0230553765,0.2013586313,0.2094534039,-0.0177904852,0.1123705506,0.386555016,-0.2594490051,0.2265853882,-0.2453865111,-0.1210877225,-0.0708780065,-0.3086999357,0.4005036354,0.2349149734,-0.017745167,-0.0296837874,0.1898634285,0.1325110495,0.0975260362,0.2627242506,0.2973817289,0.0880223364,-0.1842723936,-0.1333335787,-0.2278998792,0.0624450929,-0.0910314769,0.0342569798,-0.0607914329,0.0845985636,-0.1470314264,-0.3235642612,-0.2145537436,0.1878663003,-0.0682612658,0.1329569966,-0.1905940771,-0.1105648354,0.2698450685,0.3655023277,0.0790009946,0.2604525685,0.0080608418,-0.0145262731,-0.4633788168,-0.3155335188,0.3208703995,-0.3779631555,-0.1658343822,-0.1081847623,0.0537885092,0.1408564895,0.300988704,-0.5592666864,-0.103454873,0.0956058577,0.2044902295,-0.3980048001,0.0314418972,-0.0376871713,-0.231986776,-0.2307401001,-0.0467479043,0.0481606163,-0.1388861239,-0.3054636419,-0.2494749576]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1669","title":"wiki_dpr dataset pre-processesing performance","comments":"Sorry, double posted.","body":"I've been working with wiki_dpr and noticed that the dataset processing is seriously impaired in performance [1]. It takes about 12h to process the entire dataset. Most of this time is simply loading and processing the data, but the actual indexing is also quite slow (3h).\r\n\r\nI won't repeat the concerns around multiprocessing as they are addressed in other issues (#786), but this is the first obvious thing to do. Using cython to speed up the text manipulation may be also help. Loading and processing a dataset of this size in under 15 minutes does not seem unreasonable on a modern multi-core machine. I have hit such targets myself on similar tasks. Would love to see this improve.\r\n\r\nThe other issue is that it takes 3h to construct the FAISS index. If only we could use GPUs with HNSW, but we can't. My sharded GPU indexing code can build an IVF + PQ index in 10 minutes on 20 million vectors. Still, 3h seems slow even for the CPU.\r\n\r\nIt looks like HF is adding only 1000 vectors at a time by default [2], whereas the faiss benchmarks adds 1 million vectors at a time (effectively) [3]. It's possible the runtime could be reduced with a larger batch. Also, it looks like project dependencies ultimately use OpenBLAS, but this is known to have issues when combined with OpenMP, which HNSW does [3]. A workaround is to set the environment variable `OMP_WAIT_POLICY=PASSIVE` via `os.environ` or similar.\r\n\r\nReferences:\r\n[1] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py\r\n[2] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/search.py\r\n[3] https:\/\/github.com\/facebookresearch\/faiss\/blob\/master\/benchs\/bench_hnsw.py\r\n[4] https:\/\/github.com\/facebookresearch\/faiss\/issues\/422","comment_length":3,"text":"wiki_dpr dataset pre-processesing performance\nI've been working with wiki_dpr and noticed that the dataset processing is seriously impaired in performance [1]. It takes about 12h to process the entire dataset. Most of this time is simply loading and processing the data, but the actual indexing is also quite slow (3h).\r\n\r\nI won't repeat the concerns around multiprocessing as they are addressed in other issues (#786), but this is the first obvious thing to do. Using cython to speed up the text manipulation may be also help. Loading and processing a dataset of this size in under 15 minutes does not seem unreasonable on a modern multi-core machine. I have hit such targets myself on similar tasks. Would love to see this improve.\r\n\r\nThe other issue is that it takes 3h to construct the FAISS index. If only we could use GPUs with HNSW, but we can't. My sharded GPU indexing code can build an IVF + PQ index in 10 minutes on 20 million vectors. Still, 3h seems slow even for the CPU.\r\n\r\nIt looks like HF is adding only 1000 vectors at a time by default [2], whereas the faiss benchmarks adds 1 million vectors at a time (effectively) [3]. It's possible the runtime could be reduced with a larger batch. Also, it looks like project dependencies ultimately use OpenBLAS, but this is known to have issues when combined with OpenMP, which HNSW does [3]. A workaround is to set the environment variable `OMP_WAIT_POLICY=PASSIVE` via `os.environ` or similar.\r\n\r\nReferences:\r\n[1] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py\r\n[2] https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/search.py\r\n[3] https:\/\/github.com\/facebookresearch\/faiss\/blob\/master\/benchs\/bench_hnsw.py\r\n[4] https:\/\/github.com\/facebookresearch\/faiss\/issues\/422\nSorry, double posted.","embeddings":[-0.2976846099,-0.122228615,-0.1067941487,0.1046242341,-0.0974260941,-0.0534375869,0.0623664409,0.329993695,0.2043515593,0.0627962276,-0.0362101942,-0.0759141147,0.3217522204,0.1443755031,-0.329991281,-0.2232453823,0.1921799481,0.0447993577,0.1362492442,0.0089026475,-0.2738982439,-0.0583764166,-0.3977983892,-0.1027324647,-0.1429358125,-0.2541145682,-0.0842311606,-0.0911638737,-0.1966576278,-0.5001938343,0.2629151642,0.4861430526,0.1652962565,0.1834714711,-0.0001054152,-0.1001169086,0.1930698603,0.1926413625,0.0480890125,0.6098158956,0.0376013219,-0.1446354389,0.1473102272,-0.2647178471,0.012251921,-0.3938210011,-0.0985934064,0.1016964316,0.1992801875,-0.1435369551,0.1727650911,0.0814296529,-0.1328088641,0.096579805,-0.0477576517,-0.120473139,-0.2167093903,-0.2427264154,0.1850718856,-0.1586667895,-0.2104788125,0.5752785206,-0.2118805498,0.102905646,0.287722677,0.1624941528,0.1683534086,-0.1930993199,0.220011279,0.534470439,0.3223836422,0.0620260797,-0.235484913,-0.2457633466,-0.0465837978,-0.3243643641,0.2690390646,-0.2020264715,-0.1782935709,0.0964055061,0.0425666906,0.0330756642,0.0112167047,-0.0702812672,-0.007210678,0.4143025279,0.1632209271,-0.0044304561,0.1625287533,-0.1745149791,-0.0824148208,-0.2490358651,-0.0147156948,0.1770819426,-0.4193206131,-0.0398419499,0.1095605642,-0.1348030269,0.1472896338,-0.1609082967,-0.3319703341,0.4500738978,0.1630414873,-0.0756092817,0.1569504142,-0.0168816037,-0.3011651039,-0.1090048403,0.3752598763,0.017726684,-0.1548790932,0.1711194664,0.1973835975,-0.1147617772,-0.1229803339,-0.3387014568,-0.2557950318,-0.1036427021,0.2353302389,0.1299839765,0.1344509274,-0.256744802,-0.031048838,0.545059979,-0.1647080928,0.5105432272,0.0966034904,-0.1278076768,-0.323789984,0.2488512844,-0.0711700842,-0.1843919307,-0.1672294289,0.3012452126,0.096030347,0.0622227602,-0.0511641055,0.0579931326,0.1516148597,0.0214871932,-0.0008305582,-0.0594323799,0.2263106555,0.2509995103,-0.1030343175,0.1034574434,-0.2432741523,-0.0524805039,-0.3027390838,-0.1318610609,-0.03770658,-0.3637017012,-0.0789661929,0.2259154916,0.153771773,-0.0095955245,-0.0386791974,0.3573207259,0.0636771768,-0.0005329757,-0.0027698134,-0.0109065585,-0.1507496983,-0.1132374778,0.4777676761,0.3021310568,-0.0963883698,0.1477075517,0.0146033606,0.1196590066,0.1188180298,0.464194119,-0.1150398031,0.14634794,0.0402115099,0.2481357753,0.0273248814,-0.304608047,-0.2665833831,0.3552547693,0.1579463333,0.0052716085,0.1489366442,0.0712228194,0.25510934,0.0345554017,-0.0680296943,0.4842619002,0.1524915099,0.130175218,-0.435459882,-0.4753507376,-0.0334945507,0.3714721799,0.1413951069,-0.2113063633,-0.1654024124,-0.1668291986,0.3127175868,-0.3167105019,0.1136797294,0.1072490364,0.1147617996,0.1989511102,0.2834100723,-0.0874506906,-0.1996229738,0.2902365029,-0.2615647018,0.3387904763,0.2491782457,-0.2969032824,0.1201612875,0.0612251237,-0.122620374,-0.1599895209,0.1563497335,-0.074274838,0.0213259626,0.0181642119,-0.1195273027,-0.0558518842,-0.236286357,-0.0168638304,-0.0327991433,0.0077429144,-0.2935707569,0.0623862259,0.0980656594,0.1793346256,0.0511564501,0.120849967,-0.0044876765,0.0846828371,0.0893447772,0.0004128199,0.5253211856,0.239787668,0.0636632144,-0.19897677,-0.0448684879,-0.0791466832,0.0191594679,-0.2078748792,-0.2949854732,0.4744634926,0.0535268299,0.639801085,0.0521392263,-0.1695007086,0.2496026903,-0.0907273293,0.3009428978,0.1477142423,0.0492625125,0.0726563334,-0.1201697662,0.1294329762,-0.296716094,0.3991198838,0.134882316,-0.1264396459,-0.1145361885,0.3231076896,0.1218427196,-0.3963513076,0.1333666444,-0.279892534,-0.0742253587,0.2621766031,0.0965581164,-0.0938909948,-0.0835145488,-0.213496089,0.2458111048,0.0594613366,-0.137094453,0.1648263931,0.0653128847,0.1471574605,-0.1903486699,-0.2941790223,-0.11321491,0.0761379674,-0.0161962043,0.0117167067,-0.1046688631,-0.0277570132,0.3916325867,0.0615943894,-0.437181294,-0.3558013439,0.2353981286,-0.1715197116,-0.1559889466,0.2822730541,-0.0554340295,0.0628556609,-0.0633971468,-0.3186098337,-0.4480033219,-0.3094058633,-0.2336214334,0.0440039076,0.1775153428,0.6337993145,0.2542839348,0.1765007228,-0.2109744996,-0.1072526723,-0.3896984756,0.1459154934,-0.1024214402,0.1071886942,0.0790902227,0.1280239373,-0.2580246031,-0.2735026479,0.1032147482,-0.2019751072,0.0492336228,-0.1289891452,-0.2297789305,-0.0238723159,0.0750538185,-0.4068042338,-0.1270494759,-0.3025960624,0.130313918,-0.1128370911,0.132321611,-0.2320055515,0.0666955486,0.1342340112,-0.0795182288,0.1049049348,-0.1775390506,-0.0486542061,0.3422835171,0.0246740896,-0.2993176281,-0.1019132137,-0.1464487463,-0.0267161801,0.0754812509,-0.5214396715,0.1073006615,-0.4605299532,0.4406935871,0.1462900043,0.2562380135,0.3611198068,0.1144773439,-0.2418409139,-0.028342776,-0.0074090646,0.0752480552,-0.2573398948,-0.0447609052,-0.0906380117,0.0503440909,-0.2395912111,1.1339564323,0.1891254783,0.0338071845,0.2934496701,0.2454099208,0.1794987172,-0.1947926879,0.0681175217,0.1677199304,-0.137259379,0.2168342769,0.2975349128,0.059100695,-0.3269144595,-0.0443464927,0.0449896716,-0.1534859091,-0.3492105305,0.1776080877,0.3953426182,0.3450183272,0.0592684522,0.1262679696,-0.2123252004,-0.2321273983,0.1792701781,0.2656973004,0.0164759234,-0.2449573427,-0.0578423031,0.1206894964,-0.8101043701,0.2261852622,0.1092268601,0.1086989194,-0.0489868596,0.1958331019,0.3047571778,-0.3645422459,0.3748657703,-0.4867278039,-0.0223069713,0.2978295684,-0.2447893173,-0.1038804501,0.12525855,-0.2135812789,-0.5068948865,0.3344529867,0.2769773006,-0.4533751905,-0.3698814511,0.1629190892,-0.0549654178,-0.1317180246,-0.3748954237,-0.1765306145,-0.3780054748,-0.1153636277,0.2514256537,-0.0873970762,0.1720983535,-0.1337898821,0.0358246267,-0.2623511255,0.2735429704,-0.0069037797,0.3046072125,0.0209567025,0.0163489617,0.3673815727,0.0791151971,-0.0779408142,0.2594228685,0.0409225971,0.0328048654,0.303296268,-0.2176844776,0.0534449667,-0.0048671807,0.4127340615,0.0625834838,0.3030747771,-0.0089853704,0.4678243399,-0.4732510746,0.0134106204,0.298879981,0.0575696751,-0.3464903831,-0.2968801558,0.5739645362,0.0968784839,-0.2765947878,0.2757393122,0.572203517,-0.0631936714,0.7707854509,0.4574108422,0.9429730773,-0.333697021,0.1652428061,0.0859334469,-0.1097120643,0.1944678575,-0.5661563873,0.0193617623,-0.3551995158,-0.087693423,0.1354873329,0.0509273969,0.1005476117,0.1055908427,-0.194656238,0.0615506247,0.2874813974,-0.3494615853,-0.1280440837,0.665809691,0.1729646772,-0.5658151507,-0.0844349787,0.1249911934,0.1187894642,-0.0280022472,-0.0297368523,-0.1307890713,0.0050624087,-0.0016357695,-0.3985829949,-0.0908581689,0.1637729704,0.1294056475,-0.1353537589,-0.0828441009,0.3283472061,0.2061927021,0.062794745,0.0567427166,-0.3397096992,0.4979533851,-0.2290307879,-0.0229700375,0.1016626582,-0.2771745026,0.0599106587,-0.0731600448,-0.3809022307,-0.3557908237,-0.1788033843,-0.1548418701,-0.3439089954,-0.113287136,-0.119530946,-0.0133219203,0.1681835949,0.1202941388,-0.3320031762,-0.185490787,0.1826682985,-0.0091558984,-0.0585378334,0.5724567771,0.1440650523,-0.1502670795,-0.1820012927,0.2258184105,0.226647839,-0.0777247623,0.1830065995,0.318462491,-0.3850703537,-0.2688331902,-0.1929434836,0.1863595545,0.0740043968,-0.0011355872,0.1304306686,-0.3938735425,0.1175080091,0.1844231933,-0.0592183769,-0.3597862124,-0.1919208616,-0.0902539492,-0.1423663646,0.1451710761,-0.1071953848,0.1852142513,-0.1258393824,-0.0865794122,-0.0949737504,0.46288082,-0.3512984216,-0.0637625307,-0.3861006796,0.1190607175,-0.1809659749,-0.224087894,0.1660089791,0.097483851,0.0899728686,0.0341694839,-0.1423296481,-0.2536726296,-0.0995618328,0.1635569036,-0.0292390808,-0.2137809843,0.1806176007,0.0187171996,-0.5077250004,-0.5286760926,0.1538659334,-0.0402507745,-0.1219654083,-0.0120273102,0.0180595014,0.0612126254,-0.1259870082,0.2321688831,0.0379535742,0.3111078143,0.0477277488,-0.0065337745,0.1023024172,-0.2351322472,0.2159159333,0.1425229013,0.2501585484,0.0089295665,0.420866251,0.0437395237,-0.3991687596,0.0718968436,0.4821221232,0.495102793,0.1559592336,-0.0484922044,0.1892528981,0.2103253007,-0.238516286,-0.1419201195,0.3773721159,-0.1846658438,0.2097035348,0.0612115599,0.4228207767,0.1977474242,-0.1567343473,0.0405817926,-0.0248098094,-0.3437446654,-0.1171779037,0.0575121939,0.2765234411,0.000982926,0.2095994949,0.1603706181,0.2390824258,0.5420674682,0.1391662508,0.4469085038,-0.0583580136,0.5305747986,-0.0257439837,-0.2098160237,0.3437813222,0.1991034597,0.0135696419,0.0506672636,-0.0445410125,-0.1020090505,0.1227667481,-0.3168584108,-0.1529571265,0.2340817004,-0.1515928954,-0.3366781473,0.1648574173,-0.0178142823,0.149652794,0.2424030006,0.0485199206,-0.0395805761,0.1717758775,0.1030312777,-0.0953134894,-0.3470870256,0.1057704464,-0.0689334571,0.5805917978,0.0336862914,-0.205776915,0.1399796903,-0.1491163224,-0.0548835844,0.247816354,0.25419873,0.1394347996,-0.6033108234,0.0793581381,0.0119883241,0.0894395411,-0.0339520052,0.2869886756,-0.0519130342,0.2915502787,0.2786697149,0.0312563814,-0.2566646039,-0.7414156795,-0.081006825,0.1912506968,-0.009917344,-0.0572246276,0.109960027,-0.1264655292,0.0263365638,0.127186656,-0.2384957075,0.3384816349,0.285992831,-0.1356897056,0.0507645346,-0.0765869021,0.1138283685,-0.1017821059,0.9140748978,0.2056731433,0.1896191984,-0.1857462674,-0.1361594647,-0.6289613247,-0.0274248365,-0.3157174587,-0.1245245636,0.5484994054,0.3674613833,0.0524273701,0.177240625,0.1689513624,0.0400310978,0.0932650566,0.3857700229,-0.2792920768,0.202565223,-0.2978861332,-0.1105804741,-0.0784514099,-0.3529699147,0.4510814548,0.1768638194,0.0070843277,-0.0134673482,0.2138262093,0.1227809489,0.0763865188,0.2768260241,0.2763860524,0.1189723164,-0.1907364726,-0.0736437365,-0.2219382524,0.0264439806,-0.0790330619,0.076230295,-0.0682541728,0.0982754901,-0.1232704744,-0.3180605769,-0.1976687759,0.1924562305,-0.0448966548,0.1668352932,-0.1656690091,-0.0665471479,0.2271424234,0.3578555882,0.0714293346,0.2557866871,0.01782603,-0.0276381485,-0.3982237279,-0.3530806005,0.3242488503,-0.3849734366,-0.1858806759,-0.0938888118,0.0504415669,0.1830487251,0.2395279408,-0.5299116969,-0.049471084,0.1246989444,0.1816522926,-0.4058226049,0.0020056672,-0.0518036447,-0.2203704715,-0.2374937981,-0.051779706,0.032949239,-0.1410067081,-0.3000696301,-0.2273116112]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1662","title":"Arrow file is too large when saving vector data","comments":"Hi !\r\nThe arrow file size is due to the embeddings. Indeed if they're stored as float32 then the total size of the embeddings is\r\n\r\n20 000 000 vectors * 768 dimensions * 4 bytes per dimension ~= 60GB\r\n\r\nIf you want to reduce the size you can consider using quantization for example, or maybe using dimension reduction techniques.\r\n","body":"I computed the sentence embedding of each sentence of bookcorpus data using bert base and saved them to disk. I used 20M sentences and the obtained arrow file is about 59GB while the original text file is only about 1.3GB. Are there any ways to reduce the size of the arrow file?","comment_length":59,"text":"Arrow file is too large when saving vector data\nI computed the sentence embedding of each sentence of bookcorpus data using bert base and saved them to disk. I used 20M sentences and the obtained arrow file is about 59GB while the original text file is only about 1.3GB. Are there any ways to reduce the size of the arrow file?\nHi !\r\nThe arrow file size is due to the embeddings. Indeed if they're stored as float32 then the total size of the embeddings is\r\n\r\n20 000 000 vectors * 768 dimensions * 4 bytes per dimension ~= 60GB\r\n\r\nIf you want to reduce the size you can consider using quantization for example, or maybe using dimension reduction techniques.\r\n","embeddings":[0.1137216091,-0.3316557705,-0.0589662418,0.4534536302,0.1343118995,-0.116157122,-0.1786454469,0.4608655572,-0.4080996215,0.3328985274,0.1982362866,-0.0884741992,-0.1221771985,-0.1876727194,-0.1028836071,-0.1468020082,0.1206523329,0.4042373896,0.0463363416,-0.0668317229,0.2086952031,0.2305034697,0.0889814273,0.0583834723,-0.3092935979,-0.2625985146,-0.1549984366,0.0315525122,-0.3289449811,-0.1841223538,-0.1933373958,-0.1647046059,0.3871693313,0.4107970595,-0.000129293,-0.1663684547,-0.0067866617,-0.080024071,-0.1275971979,0.3673682809,0.0773602054,-0.3972461224,0.01280786,-0.262186408,0.2411106825,-0.3113117516,-0.1806944609,-0.2490857095,0.5573136806,0.0042652586,0.0689282045,-0.2304158509,0.0921434909,0.347211659,0.2535005212,-0.0635205135,-0.0713259727,0.1529286802,0.471544981,0.2860766947,-0.1823883057,0.363199383,0.110103935,-0.208829388,0.3755971491,-0.2341856956,0.6091362238,-0.0399567485,0.2837225199,0.3255839348,0.6376829743,-0.1370803267,0.0512443036,0.0781831965,0.3942013085,-0.0722277611,0.0660198182,0.2971102297,-0.0139591889,-0.0113798045,-0.0376826189,-0.6596640944,-0.3486739397,0.0186890159,0.1498577446,-0.2296135724,-0.1608579159,-0.1307564974,0.3024173975,-0.3091194928,0.0180879068,0.0056584124,-0.1944341063,-0.0326657034,-0.2165947109,-0.2044761628,-0.6254200935,-0.0533948205,-0.0757654458,-0.1932907104,0.6419781446,-0.1503009349,0.0239941832,0.0240580477,0.1890589446,0.4034417272,-0.1562243998,0.2484546602,0.0352215841,0.0753098577,-0.0337957032,-0.1493660361,-0.4900302887,-0.1498427391,0.2052189112,-0.3869850338,-0.280672282,-0.0732591748,0.2934067249,0.1845793724,-0.4840515852,0.2669833899,0.0975444838,0.1559239924,0.1078798324,0.2910142243,-0.0011189451,0.0923160613,-0.065930061,0.0094854841,0.0277631208,-0.0981323868,-0.142045781,0.0060584713,-0.0782943964,-0.0065724929,0.1483317316,-0.3339111209,-0.0072672791,-0.3509913385,0.0287624039,-0.3294256628,0.2498751283,-0.0001296597,-0.1208896637,0.1071232408,-0.272860378,0.1755198091,-0.1996482164,0.2861877084,-0.3546923101,-0.1992547512,-0.4079980254,0.0020834215,-0.070258759,0.2342661321,-0.0766559169,0.0782729685,0.2531573474,-0.1458364427,0.0639629588,-0.1947598904,0.019837359,-0.1918919384,0.203684628,0.0059584631,-0.5826382637,0.0113032684,-0.213216126,0.2590471208,0.2356853485,0.6174716949,-0.0363893546,0.0585219972,0.1287312806,0.4063420296,0.1356898844,0.0480834953,-0.5367757678,0.217118755,0.2386541963,-0.1011692807,0.1859214008,0.2929548621,0.0892949253,0.0071523362,-0.2879393399,0.1310657859,0.1717647612,0.1636817306,-0.2564821541,-0.441157192,-0.2164209038,-0.0965436473,-0.3174477816,-0.6281605363,-0.0682073161,0.025035454,0.272800386,-0.2577063739,0.2252342701,0.492847234,0.2753368914,-0.000695367,0.2238850147,0.1669015735,-0.1212274358,-0.1580065787,-0.2481136918,-0.4369781613,-0.2673237324,-0.198346287,-0.012754973,-0.1155380681,0.0847476721,0.2097585201,-0.0684809238,-0.4338859022,-0.2588705719,-0.2044482231,-0.0028295205,-0.1649833918,-0.1741110981,0.0226179603,-0.0813772157,-0.0307607241,-0.038439855,-0.2457405776,0.247255221,0.1347584128,-0.1484475136,0.1164523661,-0.2407489568,0.1598006934,-0.167587474,0.4313790202,0.1129634604,0.0386708044,-0.020548841,-0.64373523,0.7238839269,0.1546060443,0.1889218539,0.1275749654,-0.2429232001,-0.349193722,0.0800835863,0.2482022792,0.093988277,-0.22779046,-0.1839208454,0.067819491,-0.1057577431,-0.1484805644,0.1197732911,0.2603482902,0.34256953,0.4101367593,0.1127658933,-0.1909984201,0.343357563,0.0198469441,0.1349350959,0.5249637961,-0.6002112627,-0.3416878581,0.0518216789,-0.1772211194,-0.0939612314,0.126302734,0.0625123084,-0.125511691,0.4896761477,-0.0078556342,0.3011786342,0.3242177963,0.4192395508,0.1767469794,-0.0108457673,0.0837734193,-0.1029767692,0.062662974,-0.2192377597,0.0466163233,-0.2451952845,0.2424612343,-0.2749293149,-0.2344910651,-0.1921862066,-0.0173116568,-0.1019467115,0.0515865199,0.2451206446,-0.1880639195,-0.1591486186,-0.211703971,0.1817723513,0.4403665066,0.2060803175,0.0918668434,0.1535120904,-0.0351310559,-0.1502384543,-0.0109650763,0.0438867137,-0.014695243,0.1125260592,0.1260084361,0.1070844159,-0.1678998917,-0.1243059039,-0.0153395627,-0.0361870378,-0.1662033349,-0.1303281933,0.0919687748,-0.786903739,-0.0398700088,-0.2734321952,-0.0855514109,-0.1463751346,0.1608956605,-0.2135525942,-0.0653852299,0.1312352121,-0.1547039151,-0.0369184352,-0.2339600474,0.2178565115,-0.0222987738,0.1768089086,-0.0442196913,0.3882011771,0.1545167118,0.0012794459,0.0640717447,0.2359921187,-0.3928713799,0.3850260675,0.0214982945,-0.1801554859,0.0799703673,0.0457315855,0.1141679585,0.0209659413,-0.3639511764,0.2030226737,-0.2756900787,-0.278595984,-0.3433789909,0.2026297897,0.0911812708,0.054971572,-0.1060471535,0.1742748171,0.1105966941,0.2940917611,0.4214516878,0.4746117592,-0.1692388207,-0.2786592543,0.0714132786,0.3714794517,0.6862313747,0.1074137762,0.1681974679,0.3925536573,0.1168598086,-0.0548362769,0.2649499476,-0.0390816852,0.0188127756,0.0521350093,0.3258905113,0.4453130364,0.2013317198,0.4860762954,-0.5145969391,-0.1084597558,-0.3140788972,0.5424493551,-0.3409871757,0.3015163541,-0.281758517,-0.3549107909,-0.0133195883,-0.2670260072,-0.0733144954,-0.0885907486,0.1736109108,0.161736533,0.1939890832,-0.1758621335,-0.4150085747,0.2750554979,0.1352818012,-0.2880963385,0.1549356133,0.1337933987,0.2831843793,0.0482739992,-0.0358571447,0.1166321263,-0.2747629583,-0.1158392951,0.2354984283,-0.3694680631,0.3124309778,-0.0899734572,-0.0067961793,0.0879629999,0.0653086081,-0.1585932523,0.0240448546,0.0099530648,0.4706496596,-0.1509796679,-0.0376149043,-0.1293762922,0.0860333219,-0.46331352,-0.0107386028,0.1371354014,0.1705424041,0.3386723399,-0.3994383812,-0.0127625791,-0.4001532793,0.1063597128,0.0559597239,0.5959634781,-0.3194734156,-0.0572467074,0.1228641942,0.2665352225,0.2798195183,0.1751795262,0.0332852229,-0.2518422008,-0.0498186089,-0.2080980837,0.4521415234,0.3885692656,-0.1870096475,0.0093597621,0.1944096684,0.126067251,0.1910994202,-0.0013707286,0.0459874682,0.2445211262,-0.1782577336,0.1882913411,-0.064453274,0.1291636229,0.0304573402,0.0923735872,0.5049011111,0.0511386283,0.4228149652,0.225539133,1.2819310427,-0.1344436556,0.2138414979,-0.190648675,0.0315619148,0.1171220839,-0.0890502483,0.2319933474,-0.0786614269,0.1264596581,0.0736227483,-0.0634129271,0.0136233959,0.031883765,-0.3941905797,0.1372601092,-0.3030980229,-0.3074610233,-0.3690083623,0.1829670519,-0.0208761636,-0.2820615768,-0.1807436496,-0.0662737861,-0.047713168,-0.1506426185,0.0146177197,0.0692235008,-0.2681740522,-0.0985623747,-0.0979989097,-0.1027334556,-0.2832162976,0.1115963459,-0.2369615287,-0.3864511549,0.3440918624,0.4148671925,-0.1306281239,0.2602443695,-0.1367514729,0.1721029133,-0.0593322776,-0.2699760497,-0.0435662493,-0.1252530366,-0.1257939488,-0.0844017491,0.0121079125,0.0373379216,0.1637973338,-0.2950190902,0.3602509797,-0.1748200357,0.7085203528,0.0546970665,-0.0832191408,-0.1273793727,-0.5596513748,0.0113310413,0.0413069203,0.3711184859,0.1836883575,0.2382619381,0.3910109103,-0.3079733253,0.0975241885,0.0628666431,0.3214305043,-0.0193516295,0.7334330082,-0.1669426858,-0.2465697527,-0.2655655742,0.3216527104,0.0326206796,-0.4623304307,0.1976676285,0.0064532855,-0.3059625924,-0.1030947715,0.235552147,0.2818929553,0.0115014818,-0.3024431467,0.0060576745,-0.38651371,0.4225132167,-0.1388490945,0.2432640493,-0.036615856,-0.4012190998,-0.0374315642,0.3353601396,-0.1049782708,0.3595975339,-0.1137710065,0.0796488747,-0.1602434516,-0.0459150746,0.0498032942,-0.300337553,-0.0874512494,0.2291772813,0.1241424456,-0.0989670232,-0.1797979325,0.1781725287,-0.0540412404,-0.059090469,0.0503462926,-0.4137822092,-0.0927604213,-0.2805047333,-0.1733571887,-0.2545133233,0.3050201237,-0.261765033,-0.128130272,-0.1699758619,-0.1282246411,0.1567588001,0.1440479159,0.2888166308,0.0827175304,-0.045481354,0.236332342,-0.3123509586,-0.2136643678,0.2970193028,0.4324501157,0.2068550587,0.1090134829,0.3513637185,-0.1013693884,-0.1990477741,-0.162529543,0.137292996,-0.062797077,-0.0920757055,0.2627412677,0.0169004016,-0.1033727825,-0.2043920904,-0.121281065,-0.4736107886,-0.4672915936,0.0142504796,0.0597827584,0.273442328,-0.2132897079,-0.0273086391,-0.2114281803,0.1240290031,0.0010524523,0.3375549912,0.0598064736,-0.1298225522,0.0006453398,-0.0257925987,0.5211464763,0.2667185664,-0.3445760906,-0.0237662029,0.4986928403,0.3806135356,0.0009741188,0.109378159,-0.1961462647,0.22603333,0.2355386913,0.1819458157,0.1937902868,0.5880278945,0.3157110512,-0.0874496475,-0.1110390946,0.0368193872,-0.1495421827,0.0311907511,0.0440904275,0.0746323988,-0.0530978963,0.0082553411,-0.3844767511,0.0205221474,0.0415400527,0.3682130575,0.0305648483,-0.093818903,0.1930747181,0.0438018292,0.2861435413,-0.2255739868,0.0276724324,-0.1405593157,0.3002783954,-0.1159047782,0.2977313101,0.1508740485,0.0979457349,0.145440504,0.1547387391,0.2353406698,0.0387759805,-0.0379785672,-0.2686649859,0.4310177565,0.0131004872,-0.081076704,-0.0771013275,-0.0239915438,-0.1285038888,0.0003872976,0.3382590115,0.0446582101,-0.2085217535,-0.0749502704,-0.3417833447,-0.3779195845,-0.0687136352,-0.2507584393,0.4585436881,0.1665566713,-0.127736032,-0.0884364173,-0.0201416593,-0.0298000909,-0.3091538548,0.4782453477,0.3961673379,-0.0792749301,-0.2440525293,-0.2150600553,-0.2994366884,0.0474866666,-0.0821966603,0.1778809726,0.2118955255,0.2039851248,0.0466624871,0.1977080703,0.3431822062,0.0931052268,-0.3775729239,-0.045984894,-0.2014560103,0.0397860631,0.5905766487,0.0623002648,0.1177547202,-0.2467526644,0.2514085174,0.2052885443,-0.1412032098,0.1503601372,-0.13232252,-0.2257355154,0.2298247218,0.5471312404,-0.0382751189,-0.2566404641,-0.0233353507,-0.179323405,0.0776947737,0.1479667574,-0.0277354009,0.2186078727,0.3663982749,0.2814730406,-0.470096916,0.3040874898,-0.0057607256,-0.0943539217,-0.101290673,0.0121159945,-0.1702946126,-0.0032212862,0.1802985668,0.3396620154,-0.0955097005,0.0863409713,-0.2226566374,-0.3381663561,-0.1471143663,-0.0893623903,0.3378700912,0.3517481983,0.0487759598,0.1546375453,0.3332777321,0.8088229895,0.146721378,-0.8357337117,0.1606571376,0.1101628765,0.0341617763,-0.2272118181,0.3260708451,-0.0811272934,0.1249610037,-0.1021847501,0.6838572025,-0.0438783206,0.0102338074,-0.3169232905,-0.4951130748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1662","title":"Arrow file is too large when saving vector data","comments":"Thanks for your reply @lhoestq.\r\nI want to save original embedding for these sentences for subsequent calculations. So does arrow have a way to save in a compressed format to reduce the size of the file?","body":"I computed the sentence embedding of each sentence of bookcorpus data using bert base and saved them to disk. I used 20M sentences and the obtained arrow file is about 59GB while the original text file is only about 1.3GB. Are there any ways to reduce the size of the arrow file?","comment_length":36,"text":"Arrow file is too large when saving vector data\nI computed the sentence embedding of each sentence of bookcorpus data using bert base and saved them to disk. I used 20M sentences and the obtained arrow file is about 59GB while the original text file is only about 1.3GB. Are there any ways to reduce the size of the arrow file?\nThanks for your reply @lhoestq.\r\nI want to save original embedding for these sentences for subsequent calculations. So does arrow have a way to save in a compressed format to reduce the size of the file?","embeddings":[0.067263931,-0.2825369835,-0.0641458035,0.402374655,0.0808355585,-0.053776551,-0.2692085207,0.4793400168,-0.5607516766,0.3315005004,0.1089846343,0.1340743154,-0.1180463359,-0.2116098851,0.0485850386,-0.0961459577,0.0595268831,0.4329222441,0.0893179998,0.0133465035,0.2631244659,0.3559788764,0.0939318687,0.0067862286,-0.3308284283,-0.1569064856,-0.1461255252,0.0108346129,-0.1975939274,-0.2436818033,-0.111749433,-0.1628140807,0.3664756417,0.2997963428,-0.0001282434,-0.155340448,0.0799876302,-0.1324873716,-0.1543596387,0.350638628,0.0955116302,-0.5090038776,0.0873981565,-0.3085873425,0.1737430841,-0.402608633,-0.1041352078,-0.2626860142,0.6798275709,-0.0405709147,0.0581341572,-0.187691167,0.0452428348,0.2361843884,0.1947540641,-0.0782135054,-0.1130747423,0.1938004643,0.3574740291,0.3629294634,-0.1065315455,0.2919212282,0.1914874613,-0.2800036669,0.2227589339,-0.2258037478,0.4473464489,0.0007933857,0.225431934,0.3514633477,0.5140762925,-0.1759474128,-0.0603622533,0.0637928993,0.4757856131,-0.2109968364,0.0174834989,0.286449641,-0.0265951119,0.0579873584,0.0427703559,-0.6632637978,-0.3350250721,-0.0129562095,0.1566061527,-0.2324803472,-0.243476972,-0.2268908918,0.3395007849,-0.2392470539,0.029350739,-0.0688615665,-0.1880541891,-0.0309015494,-0.1763373017,-0.2596206665,-0.77453053,-0.1786165833,-0.0796045065,-0.2344761342,0.6635527015,-0.0720998943,0.1642229855,-0.00721636,0.1766322851,0.3590497673,0.0140034854,0.2010472417,0.0712022483,0.0396765769,0.0171178654,-0.1618068665,-0.5301873684,-0.0396627933,0.1944205314,-0.3796878755,-0.328019321,-0.0312103201,0.3187498152,0.1300549507,-0.335705936,0.228883937,0.0295665171,0.0492363684,0.1042785197,0.3126418293,-0.0208850373,0.0403178036,0.0544417799,0.0019847983,0.0394293331,-0.1343961507,-0.0781032443,0.0259398296,-0.0275111292,-0.2190253884,0.2604770958,-0.3533632457,-0.0077571361,-0.2964896262,0.0408075601,-0.2585963011,0.2016054541,-0.0302536003,0.0025413448,0.0752286091,-0.3014023304,0.1610336155,-0.1245599538,0.3366652131,-0.4297147691,-0.1650011241,-0.3299642205,-0.0179977175,-0.1047722399,0.1742520183,0.0003865594,0.1255943328,0.2257331014,-0.1661272943,0.0993145332,-0.1296684444,0.0494081452,-0.2312420905,0.2432983816,-0.0923740864,-0.6731775403,-0.0857579261,-0.2777484357,0.1920726895,0.1400735378,0.6400219202,-0.0744174495,-0.0322014168,0.0783581287,0.3584970236,0.3260307312,0.0667536333,-0.4791609347,0.1774730831,0.2246381789,-0.0786180794,0.1037978977,0.2265954316,0.1492664814,-0.0604442693,-0.2285675853,0.0938980207,0.1577383876,0.0399236791,-0.1941211522,-0.3479957879,-0.1874046624,-0.1616005301,-0.3704382479,-0.556866169,-0.0573177338,0.0993908495,0.2825308442,-0.2271603346,0.225230366,0.4774558246,0.1413949728,-0.0760712996,0.3271079659,0.1416174024,-0.0021055548,-0.1257630587,-0.3621923327,-0.5235086083,-0.1756133139,-0.190572843,0.0293155946,-0.121326156,0.0939254388,0.1542140841,-0.0850109085,-0.4176620245,-0.1665883511,-0.1945740879,-0.1212180406,-0.1485397071,-0.1625646353,0.0435752422,-0.0093543902,-0.0929440409,0.035087578,-0.2618329525,0.1685757786,0.049347043,-0.2110937685,0.0334855728,-0.2219132781,0.1685773581,-0.1188398153,0.4782136381,0.076688461,-0.1507988572,0.0236453805,-0.5871305466,0.5806174874,0.1016194671,0.1639539897,0.1385890245,-0.4002934992,-0.2357888073,0.1655118316,0.0865475535,0.1155845299,-0.2047315091,-0.1576061398,0.0299940892,-0.1912243962,-0.2066265792,0.0113874953,0.2525856197,0.3043104708,0.2930154502,0.1326592267,-0.2342439294,0.3456522822,0.0127602592,0.1384345591,0.5068286061,-0.590945363,-0.3025445342,-0.0011266111,-0.1128183678,-0.0976115465,0.1313309968,0.0698328018,-0.1697295159,0.5162374377,-0.0031064451,0.3938286304,0.270829916,0.5881363153,0.2482156903,0.0130917123,0.1224994883,-0.1683842391,0.2059627771,-0.123657085,-0.1146202683,-0.1343303025,0.1422505081,-0.3604365885,-0.1974012852,-0.1886174828,0.0823270231,-0.0327368937,0.0508299693,0.2275571376,-0.2626132071,-0.0950481668,-0.1746382415,0.1507802308,0.3636558354,0.1913493723,-0.0163355395,0.2036825567,-0.0591201037,-0.1144365445,0.0136475526,-0.064595215,-0.1314064264,0.0420030691,0.0426589921,0.1085365787,-0.1422220021,-0.1445227712,0.084627755,0.0082741491,-0.2499080151,-0.1291313916,0.0014617112,-0.7448284626,0.1219720989,-0.2711542547,-0.1041727141,-0.2187893093,0.0720855296,-0.1498136371,-0.0517308488,0.1437551379,-0.1157066822,-0.0283922385,-0.1755820364,0.291978091,0.0340435579,0.1302584112,-0.0711229295,0.3743300736,0.2452189028,-0.0035659089,-0.0189086702,0.2167963386,-0.3478161395,0.4853478968,-0.0247741044,-0.2298325151,0.0000074246,0.0339117311,0.0755403712,0.1155659705,-0.2431470305,0.2363770157,-0.1075066254,-0.2725660801,-0.3093037307,0.178979218,0.1244450286,-0.058327686,-0.1150035635,0.0773663744,0.1102449819,0.3835187256,0.4883324504,0.5294553041,-0.1333200783,-0.24173069,0.0422669388,0.3603091836,0.6346899867,0.120903194,0.1444972456,0.2817271948,0.1929074973,0.0391993187,0.4585183859,-0.1066624895,0.0548391752,0.162365973,0.2848105431,0.4691654742,0.1828640997,0.4499879777,-0.5186035633,-0.0367425159,-0.1853714734,0.4856633246,-0.3974145949,0.3809382021,-0.1953455061,-0.2994411588,0.08462964,-0.2304918915,0.007099086,-0.0964522213,0.260533452,0.0451736599,0.2662026286,-0.1417333186,-0.4890057147,0.1582086235,0.1310581118,-0.2440658957,0.1971672475,0.0474171974,0.1548453569,-0.0201170854,-0.0414036028,0.1315921247,-0.3944899738,-0.0863266364,0.1839589924,-0.4335387945,0.2960476279,-0.0694704726,-0.0164886825,0.0242046043,0.1298882365,-0.1280749738,-0.0804572329,-0.0356766842,0.4960733354,-0.2424960434,0.0180646442,-0.0489427671,0.1010118201,-0.4381713271,-0.0898477361,0.102928184,0.0489280932,0.3025795519,-0.4152060151,-0.0937522128,-0.289640367,0.0332685523,0.0367536247,0.6221728325,-0.3188767433,-0.0418372676,0.0982410759,0.122875236,0.2350888997,0.171498999,0.1339370012,-0.2176385671,0.0213347077,-0.1492290497,0.4313943982,0.2597925663,-0.2377799898,-0.0589340776,0.1089125872,0.0741329268,0.1757491827,-0.1656113267,0.0866549015,0.1817600727,-0.1917173117,0.2376742512,-0.0064083892,0.226253435,0.1244768426,0.1428155154,0.325280875,0.0712398738,0.4024209082,0.3104081452,1.0934176445,-0.1287821829,0.2878456712,-0.1855242252,-0.0314821899,0.1677292138,-0.0775820762,0.2207026482,-0.0507251211,0.1350910813,0.048908107,-0.0110023702,0.03768453,0.0935291573,-0.4008229077,0.1416651607,-0.2495075911,-0.3455013931,-0.2802385688,0.264498055,-0.0119484263,-0.30631724,-0.1614439934,-0.0556845143,0.0325534418,-0.1607539952,0.0973842591,-0.0465735793,-0.2276340872,-0.0740784854,-0.0539508797,0.037938118,-0.2111909688,0.105247125,-0.2001098096,-0.5033581853,0.3286201358,0.3860142827,-0.1774036586,0.1504752785,-0.0988138616,0.180958569,0.0230001044,-0.2624498904,-0.0634004027,-0.1264546663,-0.1668697149,-0.1170619354,0.1283912212,0.0339487344,0.0910818204,-0.2994737923,0.5842024088,-0.2078561485,0.5878828168,0.0970592052,-0.1030642241,-0.1162516549,-0.5892644525,-0.0096269595,0.0382418446,0.3565510809,0.1828390658,0.1842283159,0.4391725063,-0.1612726152,0.1305127591,0.0740120187,0.3857982159,-0.0120790387,0.7654311657,-0.192737326,-0.164211005,-0.2633431554,0.2644882798,-0.0846717134,-0.472369343,0.3465470374,-0.0600116961,-0.2479175329,-0.0745704323,0.2044215798,0.23022452,0.0835557953,-0.4369381368,0.0162837729,-0.4646851718,0.3239001036,-0.1200581044,0.2729882896,-0.0041064946,-0.3141970038,-0.0186182223,0.3333195746,-0.1124104932,0.3276797533,0.0084069772,0.0749627575,-0.1830726862,0.0497311167,0.1053538546,-0.2597889304,-0.0891984105,0.3804856539,0.1472004205,-0.1042716354,-0.1940243989,0.185444355,-0.1057780907,-0.0335897617,0.0290962141,-0.3154132962,-0.1044459268,-0.2183835804,-0.1982858628,-0.2835426033,0.2915591896,-0.314152658,-0.0030917155,-0.207375437,-0.1813281626,0.1156785563,0.1284759492,0.3049497306,0.037167035,0.0612643361,0.1911238134,-0.3139713705,-0.3012051582,0.1760341525,0.4389144778,0.0999903679,0.11993251,0.2517718971,-0.2074714154,-0.2415664643,-0.1788192391,0.188997969,-0.0664439127,-0.0680996776,0.2118321061,-0.0175427776,-0.1619874835,-0.154440999,-0.2356278002,-0.4070380032,-0.5107870698,-0.0505880713,0.091530785,0.2492302358,-0.2055696398,-0.0225741751,-0.3179937303,0.1100414544,0.1095681861,0.3319498003,0.2289375812,-0.0958464518,0.0498471931,-0.0874892995,0.512218833,-0.0179323163,-0.3362059295,-0.0426701233,0.618886292,0.3766671419,0.0590201542,0.1007626504,-0.1343935728,0.1774658859,0.214203462,0.2379975468,0.2395053357,0.6591510177,0.3992859423,-0.0830742717,-0.1581904888,0.1943127066,-0.1205557659,-0.0382822752,0.0432950407,0.0945108309,0.0109623075,0.034207698,-0.3181349039,0.0844226554,0.0497327037,0.4067024589,0.1238243356,0.0285054445,0.1804178953,0.0569752567,0.3310377598,-0.2036266476,0.0831326395,-0.1041354537,0.279720962,-0.1739206165,0.3154942393,0.1687639058,-0.0076066311,0.1997717321,0.1897363812,0.2928083837,0.0257325191,0.1126205325,-0.2340721488,0.5015303493,-0.1101212502,-0.0384547412,-0.0451747477,-0.0203724317,-0.0814571828,-0.0159712657,0.1981264502,0.0288142934,-0.1625747532,0.0618569516,-0.2549828589,-0.3699814081,-0.1174420938,-0.1436953694,0.4433273673,0.1981670558,-0.1535850316,-0.1786174476,0.0041289981,-0.0225224886,-0.2683810592,0.5162475109,0.2965544164,-0.205163151,-0.0771106035,-0.2799364924,-0.205603689,0.02792711,-0.1277355403,0.1398620754,0.2190639824,0.2399235964,0.0545045882,0.2773197889,0.3435633779,0.0985492542,-0.3065315783,-0.0582888871,-0.1980984658,0.0425648279,0.6677463651,0.1645656973,0.1425884515,-0.2501245439,0.1892293543,0.2020077854,-0.1647945791,0.1294285506,-0.1772539914,-0.1251276881,0.1388525665,0.5862326026,-0.1196151152,-0.360925436,-0.1206138134,-0.1635197997,0.1863893121,0.2066577375,-0.0159587245,0.1790756583,0.4032120109,0.2369689345,-0.398357451,0.2444373518,0.0063421237,-0.1069057062,-0.1695092469,-0.0891232789,-0.0445543267,-0.0449707806,0.1237526909,0.2627573311,0.0513751544,0.113261655,-0.2934110761,-0.3352183104,-0.2135386765,-0.1042336673,0.2608759701,0.3320406973,0.1046812236,0.2116081417,0.2618841529,0.7541187406,0.1268854588,-0.7142351866,0.2270789444,0.1679456085,0.0354758389,-0.1394239664,0.2423329055,-0.0989552736,0.1849943101,-0.1064541414,0.6509243846,-0.0033633984,-0.0971023142,-0.4538102746,-0.4735901058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1662","title":"Arrow file is too large when saving vector data","comments":"Arrow doesn't have compression since it is designed to have no serialization overhead","body":"I computed the sentence embedding of each sentence of bookcorpus data using bert base and saved them to disk. I used 20M sentences and the obtained arrow file is about 59GB while the original text file is only about 1.3GB. Are there any ways to reduce the size of the arrow file?","comment_length":13,"text":"Arrow file is too large when saving vector data\nI computed the sentence embedding of each sentence of bookcorpus data using bert base and saved them to disk. I used 20M sentences and the obtained arrow file is about 59GB while the original text file is only about 1.3GB. Are there any ways to reduce the size of the arrow file?\nArrow doesn't have compression since it is designed to have no serialization overhead","embeddings":[0.0507503264,-0.1890882105,-0.0481205396,0.4096188843,0.1655380428,-0.1161523238,-0.1979257911,0.4760343134,-0.5176417828,0.3480738401,0.1787733436,0.036895752,-0.0722984001,-0.2448065132,-0.1110658869,-0.1173491478,0.1286960989,0.4373621643,0.0612979345,-0.07270284,0.2508214116,0.2598432004,0.0781849027,0.0291273501,-0.3022218347,-0.2502729595,-0.0998189896,0.0724829212,-0.3011452258,-0.2625126243,-0.1283480376,-0.1378051341,0.4053733647,0.3970190287,-0.0001282424,-0.16256392,0.0895902887,-0.0594073981,-0.1381272972,0.4011262059,0.1029510275,-0.4261998236,0.0292587429,-0.2789821029,0.2387601733,-0.2446385026,-0.1166483313,-0.1664191335,0.526529789,0.0165715422,0.0585713573,-0.2652911246,0.1759495288,0.314307332,0.2205957323,-0.0668862462,-0.1156277955,0.169132933,0.4665040076,0.3631112874,-0.2120838463,0.3825870752,0.1959662884,-0.1987203211,0.2826405168,-0.2619544268,0.5988512635,-0.0072824201,0.2758460939,0.3210677207,0.6042670012,-0.1604376733,0.0022472576,0.1364461631,0.4048812985,-0.1223638728,0.0436504632,0.2967832685,0.0007919627,0.0305735599,0.0428958535,-0.6791768074,-0.3727332354,-0.0391066447,0.1542566121,-0.1809260398,-0.138619408,-0.1515310109,0.3044012189,-0.2638312876,0.0057943575,0.0414171033,-0.2033098191,-0.0345911533,-0.2742111683,-0.1971734762,-0.6642643213,-0.180932343,-0.076967597,-0.224204734,0.6161234975,-0.1189014763,0.0792511329,-0.0018123015,0.1857958436,0.3607159853,-0.1320570409,0.2629351318,0.0967257097,0.0627169088,-0.0658576563,-0.1772944629,-0.5473857522,-0.0522804074,0.2111370116,-0.3625531495,-0.2314305305,-0.0631910712,0.2877734601,0.1227807924,-0.4077186286,0.2274799496,0.0933762267,0.1170782894,0.0775150955,0.2975284457,-0.0711270347,0.0274639186,0.0082984939,-0.0344498679,0.0231469646,-0.1661094725,-0.1152518615,-0.0009006085,-0.113107346,-0.0745722204,0.0951939598,-0.3681463599,-0.0617448017,-0.2640174925,0.0579709858,-0.3185176551,0.1638256758,-0.0389143825,-0.0550109111,0.1336952597,-0.3181382716,0.2252838463,-0.1759065241,0.3231898844,-0.3950983584,-0.2102268785,-0.4000717998,-0.0085916817,-0.0432525091,0.1995307058,-0.0100092795,0.0897501484,0.2363730669,-0.1866234392,0.0764531568,-0.122878246,0.043755047,-0.2060817778,0.1769923717,0.0002048277,-0.5913956165,0.0206888691,-0.2263724655,0.2290249616,0.2718520761,0.6582818031,-0.011123199,0.0398029462,0.1661523581,0.3295146227,0.2715435624,-0.0775721148,-0.5677331686,0.2573771477,0.1854646057,-0.1532877386,0.1948246807,0.2738877535,0.0458326228,-0.0740796328,-0.2793557942,0.1502473652,0.1047165841,0.1436359435,-0.184927687,-0.4282307923,-0.2048595846,-0.08162947,-0.3423552215,-0.6096333265,-0.0234281588,0.0381434038,0.2595015764,-0.1989536732,0.2322840542,0.5008648634,0.2704702914,-0.0680022016,0.2858623862,0.1859237999,-0.0035750305,-0.170546934,-0.2777453661,-0.4026816189,-0.2564080358,-0.1937722713,-0.013911834,-0.1038093641,0.1279554367,0.3090303242,-0.0833140537,-0.415664345,-0.2412571311,-0.1775528193,0.0063225972,-0.156959936,-0.176566869,0.0353177227,-0.0314238258,-0.1440633833,-0.0652855709,-0.2538245022,0.2781003118,0.1048421785,-0.123852089,0.0606218614,-0.2358367592,0.1837593764,-0.1627874076,0.4289818108,0.1266688406,-0.014523345,-0.0098014371,-0.5674405694,0.6338784099,0.2279627025,0.2302377373,0.1270756572,-0.2987273932,-0.3776149154,0.1645563543,0.1192288473,0.1711425036,-0.2152514011,-0.1111674756,0.0900983289,-0.1383969188,-0.1143791527,0.0824730545,0.2274533212,0.3786367774,0.4382952452,0.0540299341,-0.1855411828,0.4365794361,0.0096390164,0.1055398211,0.5411300659,-0.5602242351,-0.3091391921,0.0371651761,-0.1885406077,-0.1661314368,0.1323100328,0.0788812935,-0.1810807437,0.4968096614,-0.0163468011,0.286501199,0.3092641234,0.4892221093,0.1690260768,0.0823395252,0.1356552243,-0.1037537903,0.0130194398,-0.1327822506,0.0086663328,-0.1274090558,0.2101376057,-0.2826264501,-0.1781879067,-0.2139097601,-0.000316147,-0.0137713412,0.0813742876,0.2349838465,-0.1763648689,-0.0955405682,-0.1751874089,0.1633109152,0.3974402547,0.2533738911,-0.0106993355,0.2196204513,-0.0343784131,-0.1630561948,-0.0101575321,0.0394023806,-0.086733222,0.0662226528,0.1505589485,0.1859569997,-0.0867656991,-0.1688192785,0.0107819354,-0.0130925598,-0.2625660002,-0.1719710976,-0.0327368788,-0.8235833645,0.0386458635,-0.2697506547,-0.1164409146,-0.1738630533,0.1901939213,-0.2358833402,-0.1505157053,0.1694946885,-0.0956837535,-0.0340741277,-0.2430463582,0.2441932857,0.0293397177,0.1342104971,-0.022095222,0.2907994092,0.1299889833,0.0782613978,0.0570188835,0.248000592,-0.3324787617,0.4345035851,0.0006668789,-0.256079942,0.0394594483,0.0595602095,0.1081749573,0.0553325526,-0.3088936806,0.3090564609,-0.255630672,-0.294121176,-0.3409182429,0.2198177874,0.1440718174,0.0519992523,-0.1078686863,0.1429902911,0.0750048012,0.3306616843,0.4145927429,0.4684438407,-0.1529823095,-0.3367233276,0.0221690275,0.4113886654,0.646091342,0.1295186281,0.1901556104,0.3390545845,0.0841626301,-0.0177234225,0.3273495436,-0.0011622963,0.071725592,0.0609242916,0.3329191506,0.4339378178,0.1684678197,0.4799724817,-0.547095716,-0.067200765,-0.2880034447,0.5035071969,-0.3974097669,0.362754494,-0.2314971983,-0.3303448558,-0.0106226327,-0.2617217004,-0.0120286224,-0.1350444257,0.1795544177,0.1289420873,0.1481875479,-0.2065446973,-0.4695161581,0.2139192075,0.1611497849,-0.2541449666,0.1671234667,0.0817026123,0.2239688486,-0.0433683842,-0.0606019124,0.1685401499,-0.2610401213,-0.0933518708,0.198401764,-0.4394463003,0.2965718806,-0.1228943542,-0.0041612205,0.0814093798,0.0163912456,-0.1339261532,-0.0023834386,-0.0524343476,0.4400238395,-0.215056479,-0.0116168419,-0.1521436125,0.0855752528,-0.4503020942,-0.0044086678,0.1019541323,0.1755741388,0.3336870074,-0.3974992633,-0.0933346897,-0.3852759004,0.0821812898,0.0864809155,0.5707866549,-0.3757106662,-0.0128902504,0.0560714118,0.1500386447,0.2598499358,0.1578243226,0.1400973201,-0.309347719,-0.0309037305,-0.1919126809,0.3767790794,0.3295960426,-0.1719195843,-0.0071551572,0.1474732608,0.0339590944,0.2138871253,-0.0686460659,0.0292640477,0.187453866,-0.2609530687,0.2355771959,-0.1287193894,0.1452952176,0.0070167263,0.1465732604,0.4493563473,0.141200006,0.3773002625,0.1937418878,1.1788514853,-0.1270531416,0.2193982303,-0.1883725822,-0.0078594135,0.1291565895,-0.0895995125,0.2481951863,-0.1077605262,0.1205272228,0.0789498687,-0.0188815035,-0.0304000806,0.0569621436,-0.3748676181,0.1275316179,-0.2308950871,-0.2811104953,-0.3568494022,0.2820073962,0.0221187007,-0.3117346168,-0.1965523958,-0.0731102228,-0.0167824998,-0.1666530818,0.0748491138,-0.0209850594,-0.2613743544,-0.1517742276,-0.0827333182,-0.0238964595,-0.2259732336,0.1030336022,-0.120352082,-0.418338716,0.3453137279,0.3520294428,-0.0738298222,0.2160603553,-0.0845438018,0.1696390212,-0.0429370068,-0.2299388349,-0.0424261764,-0.0892668515,-0.1649933308,-0.1250689775,0.0521127358,0.0584607013,0.1609445661,-0.295016706,0.5063447952,-0.1816982925,0.6465438604,0.09123943,-0.0740310699,-0.1664719731,-0.5908510685,0.0154771823,0.0319925807,0.3636799157,0.1264856607,0.1891272962,0.4088579416,-0.2687449157,0.1189008728,0.0579097606,0.3435485363,0.0067312638,0.7002112865,-0.18974711,-0.1908385903,-0.2504741251,0.2762667537,0.0408988446,-0.4976556301,0.2203977704,-0.0779492781,-0.3504694402,-0.0512309186,0.27572909,0.1996073425,0.0855066478,-0.3620643914,0.0325320177,-0.4356412888,0.360871315,-0.1325119883,0.178705439,0.0172832385,-0.3698696792,-0.0366573706,0.3729791641,-0.1106708944,0.3083665073,-0.088040635,0.095480971,-0.1674074233,-0.0715765581,0.0627679974,-0.2792836428,-0.0987624377,0.2936201692,0.1565817147,-0.0949352831,-0.2163523734,0.184978202,-0.0726151988,-0.0267056506,0.0369366854,-0.4130551815,-0.1585206836,-0.2537832558,-0.1916853786,-0.289200902,0.3083759844,-0.2461355925,-0.1496095955,-0.152912423,-0.1866732836,0.131913349,0.1641736627,0.2739393115,0.0982140452,-0.0085038105,0.2684277892,-0.2657713592,-0.2786466479,0.2612834573,0.4838646054,0.2478047311,0.1102971956,0.2955198288,-0.1140328795,-0.2087934762,-0.1579713672,0.1052324325,-0.0812026635,-0.1091110259,0.2207968831,0.0067117088,-0.1522182524,-0.2571261823,-0.2142486721,-0.4619248807,-0.4495341182,0.0026219473,0.0700542852,0.2550339103,-0.1918656677,0.0280368421,-0.2222548574,0.2091040164,-0.0107655739,0.3839108348,0.0869313106,-0.1118685454,0.0469535068,-0.01055534,0.4986552,0.2798518538,-0.3051370084,-0.0402658246,0.6297525764,0.3693649471,0.0286890808,0.1231182292,-0.2344676703,0.1594972312,0.2328419983,0.151596874,0.2458892763,0.548817873,0.3126243651,-0.1325707585,-0.0902569667,-0.0247076675,-0.1300999224,0.0346370004,-0.01865815,0.1239150017,0.0235382058,0.0246532187,-0.3665665388,0.1525657177,0.1151593551,0.3465064466,0.1174302474,-0.0770815313,0.1658726633,0.0917935148,0.3202593327,-0.2011279911,0.0528881475,-0.1256152093,0.2939895391,-0.1100596935,0.2815486789,0.141452536,0.0670417547,0.1533951759,0.1706710011,0.2629418075,-0.0394961089,-0.0060418369,-0.2295688838,0.4855218828,0.0242939088,-0.1285311282,-0.0815535337,-0.0177085642,-0.0170439743,0.025501864,0.2159310579,0.0682787448,-0.0893943533,-0.0377070829,-0.3476885557,-0.3928695619,-0.0646102503,-0.1617678553,0.434271574,0.1767680049,-0.152278319,-0.1089238748,-0.0649212673,-0.0246207751,-0.2665039599,0.4759874046,0.3449437022,-0.0826516077,-0.1698708683,-0.2397712618,-0.2476901561,0.0822790042,-0.115754053,0.1484807283,0.2369828224,0.2136703879,0.0496859327,0.2275607586,0.329105556,0.0857308358,-0.3935407996,-0.029657064,-0.1754957736,0.0332392119,0.5905086994,0.0425914377,0.0936291143,-0.3016751409,0.2567046285,0.2504894733,-0.1454404742,0.1128226742,-0.2696452737,-0.241854772,0.1737627536,0.5679293275,-0.0559421889,-0.3346838057,-0.0955080539,-0.1680918634,0.0773420855,0.2241217792,-0.1075552478,0.190219,0.3954317272,0.2511384487,-0.4164696932,0.361214906,-0.0142293489,-0.0625579879,-0.1254799217,0.0377196595,-0.1990657002,0.0539327413,0.07548967,0.3355110884,-0.0638324991,0.0761748031,-0.2123362124,-0.3827673793,-0.1612919867,-0.1045696437,0.3224183321,0.3041210473,0.0128134461,0.0972217992,0.2628237009,0.8297924399,0.13784118,-0.8076389432,0.1921735257,0.0814647451,0.0375299044,-0.2672510445,0.3228561282,-0.1470949203,0.1519895941,-0.0857507735,0.7052807808,-0.0630456582,-0.0266039018,-0.3932935596,-0.4349771738]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1662","title":"Arrow file is too large when saving vector data","comments":"I see. Thank you.","body":"I computed the sentence embedding of each sentence of bookcorpus data using bert base and saved them to disk. I used 20M sentences and the obtained arrow file is about 59GB while the original text file is only about 1.3GB. Are there any ways to reduce the size of the arrow file?","comment_length":4,"text":"Arrow file is too large when saving vector data\nI computed the sentence embedding of each sentence of bookcorpus data using bert base and saved them to disk. I used 20M sentences and the obtained arrow file is about 59GB while the original text file is only about 1.3GB. Are there any ways to reduce the size of the arrow file?\nI see. Thank you.","embeddings":[0.0401681177,-0.2502167821,-0.0606315695,0.4922181666,0.1117284447,-0.1174077541,-0.1834441423,0.432300359,-0.5243883729,0.348590374,0.1785313487,-0.0209325589,-0.0856497511,-0.2135419399,-0.1167742312,-0.1533538252,0.1172391847,0.4240454733,0.0469511077,-0.0726861656,0.2308914363,0.2567802668,0.0635340288,0.0628622845,-0.3285108209,-0.275255084,-0.1115148962,0.0961957499,-0.2844336629,-0.2232511342,-0.1700700223,-0.1266344488,0.4344142675,0.4224643409,-0.0001286193,-0.1795559525,0.0395773202,-0.0844501629,-0.1227384433,0.4055206478,0.0336733907,-0.4020327032,0.0187182315,-0.3380127251,0.2295054197,-0.1917050034,-0.1575640291,-0.1507896781,0.5684316158,0.0564378053,0.0649439767,-0.2654737234,0.1426814198,0.3326765597,0.2908854485,-0.0981659293,-0.0735566989,0.1320353895,0.4601346552,0.3115095496,-0.1797334105,0.3663775325,0.1735219955,-0.210288465,0.3022196293,-0.2312025428,0.600092113,-0.0385223962,0.2519407272,0.3201055825,0.6274766326,-0.1581799239,0.030799754,0.1190490425,0.3379938006,-0.0965935588,0.0550756231,0.2763803899,-0.0186816957,0.0347851254,0.0503512621,-0.6824251413,-0.4179260135,0.0033212092,0.1648013145,-0.1794048101,-0.1642932892,-0.1617091596,0.3105744421,-0.2711760104,0.0022674687,0.0393699482,-0.1972249448,-0.0399636962,-0.2539339066,-0.2047715783,-0.6492016912,-0.101767607,-0.0728560612,-0.2742416859,0.6638318896,-0.1338996887,0.0388555415,0.0183084346,0.1861018986,0.3444779813,-0.1739724427,0.2610003352,0.1364305913,0.0836249664,-0.1120015606,-0.1662757695,-0.5035558343,-0.1357659847,0.2288920432,-0.3346647024,-0.2377629578,-0.0849989057,0.2673284411,0.14608109,-0.4199253023,0.2093949318,0.1226423681,0.1503825933,0.064206861,0.2517827451,-0.0512024239,0.0547161773,0.0011381212,-0.0504450873,0.0269948691,-0.1135569066,-0.155779019,-0.0167514663,-0.1002417579,-0.0612435266,0.1280459017,-0.3992787302,-0.0386271365,-0.3119467497,-0.0124516701,-0.3539374173,0.1587024033,-0.0529073142,-0.0660966113,0.106406793,-0.3241746724,0.2241577357,-0.1534020454,0.3436032832,-0.3936292529,-0.2413055748,-0.3603664935,-0.0142078921,-0.0708622336,0.1943975538,-0.0465516858,0.1097716391,0.2278118283,-0.1848271638,0.0685653985,-0.1283160299,0.027652707,-0.2286006808,0.1719055921,0.0822104588,-0.585508883,0.0495146289,-0.226515159,0.1975293756,0.2678924203,0.6588305831,-0.0314263962,0.0741421133,0.1063889861,0.3289429843,0.3129524291,-0.0441116728,-0.5733448267,0.2363132387,0.2131867111,-0.1698504835,0.201679185,0.2972208261,0.0809140205,-0.0582793579,-0.3040889204,0.1478198618,0.1280446798,0.1651148945,-0.241630435,-0.4002753496,-0.1937331408,-0.0586507544,-0.2884289026,-0.5854561329,0.0311416648,0.039749667,0.2916388214,-0.2184117436,0.2416806072,0.5469619036,0.3443410099,-0.0172786843,0.2508615553,0.1268852204,-0.0612170994,-0.1582131833,-0.2605774105,-0.4318853617,-0.218930915,-0.1431433409,-0.0602716729,-0.1350262612,0.077918306,0.2748738825,-0.0864611119,-0.3960631192,-0.289070189,-0.2268977165,0.0076694745,-0.0946824625,-0.2342787236,0.0090760207,0.0074250884,-0.04668447,-0.101155363,-0.2463829219,0.2778315842,0.1151733622,-0.1157192662,0.066434063,-0.2269319296,0.1660059243,-0.1596755981,0.462877661,0.1538438499,-0.0168100595,-0.0322945938,-0.6204246283,0.6692790389,0.120661065,0.1904894263,0.1371716708,-0.2479414344,-0.3891344965,0.120631136,0.1334734708,0.1441586167,-0.2442245781,-0.1417155862,0.063943103,-0.086252965,-0.1361387968,0.1042678133,0.3097651005,0.3902880847,0.3872959316,0.0454629362,-0.186847657,0.4395573437,0.015808275,0.1544037759,0.5543648005,-0.5477025509,-0.3331469893,-0.0185143631,-0.1916823238,-0.1034362316,0.1407268196,0.0729665905,-0.1279083341,0.517770648,-0.0266544949,0.3052236438,0.344401598,0.4658587277,0.181632787,0.0818277374,0.0918701142,-0.1218573749,0.0032617305,-0.1254239082,0.0765673816,-0.1801028252,0.1720205992,-0.2441990823,-0.2108225673,-0.2391768247,-0.0138324797,-0.0479215123,0.0630955026,0.2553275526,-0.1563220918,-0.1236587986,-0.1632957906,0.230626896,0.4489123523,0.2504657507,0.0108743031,0.1958824098,-0.0417763963,-0.1730007231,-0.0023989452,0.0152288936,-0.0023609188,0.0655906945,0.098827675,0.2117336243,-0.1687851548,-0.1518230885,0.0088939788,0.0141337635,-0.1680267155,-0.1450348794,0.0491596088,-0.7630653381,0.0441092663,-0.2764770389,-0.0825790465,-0.1903212219,0.1538662761,-0.242993623,-0.1233120933,0.2258817106,-0.1502488106,-0.0820894018,-0.285369277,0.222840026,0.0259570386,0.1522601396,-0.0797647387,0.3396663964,0.1145102158,0.0517652594,0.0781786516,0.2324468493,-0.3610811532,0.4152543545,0.007003326,-0.2377122939,0.050790377,0.0705837309,0.1087314636,0.0329805203,-0.3378190398,0.2583277524,-0.2557246387,-0.2365308404,-0.3306213617,0.2413688153,0.122035861,0.0458613038,-0.0821188465,0.1498086601,0.0750838965,0.3138791621,0.4652401209,0.4440535307,-0.1327842325,-0.3041436076,0.0312642679,0.4554575682,0.646759212,0.0978459716,0.2178622633,0.3232795,0.0737975761,-0.0183729641,0.2618082762,-0.0294100661,0.1021652892,0.0362139121,0.3458236456,0.4168952703,0.1593423635,0.4333873093,-0.6014704108,-0.1576856226,-0.2879237533,0.4981635809,-0.3650660217,0.3872124553,-0.2095000893,-0.4004956484,-0.0282099396,-0.2771818638,-0.0258960053,-0.0534408726,0.177983582,0.1160623431,0.2024091333,-0.190310359,-0.43911165,0.2450433224,0.1579977572,-0.3007817864,0.1206685081,0.1303693801,0.2283747345,-0.0073374701,-0.020128075,0.1223431155,-0.2180293947,-0.1236213073,0.1609844565,-0.3719014227,0.3145336211,-0.1329615265,-0.0917178914,0.1535023004,0.0764228776,-0.1236389056,0.0039265458,0.0118986657,0.4527677894,-0.1906811148,-0.0262283236,-0.1256581098,0.1165075153,-0.4562981725,0.062576063,0.114491038,0.1511439383,0.3452139795,-0.3747163117,-0.0374740697,-0.3903982937,0.0626906827,0.0952233151,0.5715113282,-0.3790847361,-0.0236535463,0.084771812,0.1295925081,0.260528177,0.1309926808,0.0377252251,-0.2597981691,-0.0211575199,-0.1988270134,0.4318984151,0.3716220558,-0.1794538945,0.0165136736,0.1905692369,0.0582225583,0.2170664966,-0.1066569239,0.0299279839,0.2599633038,-0.197500512,0.1873573065,-0.1205553189,0.1466885954,-0.0159913301,0.0897313133,0.4402635694,0.0966242254,0.403575778,0.1546315104,1.2652608156,-0.0998996645,0.2459024489,-0.2130903304,0.0404121615,0.0984010398,-0.1375276446,0.2302640378,-0.1352615803,0.096162498,0.0631494895,-0.062917605,-0.0446256772,0.0734868795,-0.3492338359,0.1608658731,-0.2465384752,-0.2415938079,-0.3692933619,0.2377439886,0.0881816968,-0.2524377108,-0.2255166918,-0.0897941813,0.0086944671,-0.1156172901,0.0044343406,-0.0077430937,-0.2654407918,-0.1154301614,-0.0990106165,-0.0644185543,-0.2917867005,0.1252423972,-0.1963357478,-0.4238348603,0.386985451,0.4624296725,-0.0486160927,0.2235883623,-0.0993737131,0.165273875,-0.0300081652,-0.2651404738,-0.0724051893,-0.0509691238,-0.1578860432,-0.1643757671,-0.0044378075,0.0658525676,0.1339127421,-0.3458079398,0.4053433239,-0.1757678837,0.6359392405,0.0514977649,-0.0941911563,-0.1268847585,-0.61334759,-0.0099055562,0.0298989173,0.412249893,0.1501037776,0.2313750833,0.4212037325,-0.2835696638,0.1003473252,0.0666420832,0.3366231918,0.0394402519,0.6982759237,-0.1524215937,-0.2044593245,-0.2787381113,0.2851925194,-0.023322843,-0.4781748056,0.2006470859,-0.030151099,-0.3285941482,-0.0789929554,0.2878503799,0.2553307414,0.0816761851,-0.34760952,-0.0468964763,-0.4687311351,0.3861242831,-0.1280089915,0.1672700942,0.0141889136,-0.3654478788,-0.0450613685,0.3550875187,-0.1161690801,0.3924208283,-0.095574066,0.0827976018,-0.0622972511,-0.0838774741,0.0686334148,-0.2529419959,-0.0917365253,0.3060911,0.1195064485,-0.0861565769,-0.1950665563,0.1949021071,-0.0490286425,-0.0330398828,0.0273656603,-0.436013937,-0.1277302802,-0.2227757871,-0.1667703986,-0.2422769368,0.3457096517,-0.235631749,-0.1824890226,-0.1996704638,-0.1950463206,0.1690808088,0.1441928148,0.230447486,0.0544794947,-0.0151511217,0.2499816865,-0.2832714617,-0.2594648302,0.2461014241,0.4713946581,0.2636876702,0.0529167093,0.3233456016,-0.0963284746,-0.2167811394,-0.1488299668,0.0746619403,-0.0626435131,-0.0943479016,0.2579266131,0.0066459547,-0.1898040771,-0.247162655,-0.1938010007,-0.4886606038,-0.3924585879,0.0292536039,0.0667307004,0.2952417731,-0.1803411394,0.0142749976,-0.2083939612,0.1683333516,-0.0248628575,0.3416135609,0.0800142139,-0.0930528864,0.049111709,0.0127929309,0.4713212252,0.3575207889,-0.2989110351,-0.0723857284,0.5432665944,0.3444216251,0.0609396808,0.0861343816,-0.1868287176,0.2233648747,0.1998151541,0.1254771501,0.1808994114,0.5768271685,0.348693043,-0.077460885,-0.1155926809,0.0149464142,-0.1622900963,0.0452485383,0.0013015671,0.0450223461,0.0267802309,0.0214781091,-0.3917791545,0.1150666922,0.0902354568,0.3548915982,0.0814954638,-0.0529371835,0.2113244385,0.0953202397,0.2596791089,-0.2216045856,0.0748322532,-0.1788131595,0.2607171535,-0.1030409634,0.315246433,0.1532089859,0.0856604949,0.1789261401,0.1649800986,0.2569923997,-0.0423140936,-0.0185160805,-0.2795057297,0.5011304617,0.0627136156,-0.0950939655,-0.0760617182,-0.0381542183,-0.0247397609,0.0540781505,0.2384482175,0.0373048484,-0.188086167,-0.036318507,-0.3648893535,-0.4148682058,-0.089489907,-0.1868912429,0.4207116365,0.1567149162,-0.1543002576,-0.0834370628,-0.0269110687,-0.0361566059,-0.2679473758,0.4784450233,0.3761763275,-0.0690760761,-0.222693041,-0.26335603,-0.3243176937,0.0979259089,-0.0710040107,0.2090918273,0.2155864388,0.1644325256,0.0445775166,0.2213982344,0.4204846919,0.1185588464,-0.430980444,0.0166761372,-0.2237553895,0.0226161517,0.5614058375,0.0574660413,0.1108510345,-0.2704607844,0.2426544279,0.291117847,-0.1399082989,0.0940671638,-0.210179016,-0.2247830629,0.180493176,0.556265533,-0.1041135937,-0.2705646157,-0.0832489356,-0.1750449687,0.1156646013,0.1535115689,-0.0436564125,0.1956659704,0.4387933314,0.267495811,-0.4105098844,0.3038395345,-0.0560668036,-0.0950059593,-0.0999838859,0.0273132641,-0.1709038317,0.0241212007,0.1385346651,0.3278045058,-0.0291186813,0.0789938122,-0.252196461,-0.391112715,-0.1869300008,-0.0919420347,0.3112695813,0.3301733732,0.0310856402,0.0605020002,0.3079842925,0.8441234827,0.1711013168,-0.795668304,0.1858357042,0.1393157095,0.0641254559,-0.2501521707,0.3305053115,-0.1243715882,0.1354574412,-0.0699529573,0.6754998565,-0.0663308203,-0.0549734831,-0.3742741346,-0.476138413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1647","title":"NarrativeQA fails to load with `load_dataset`","comments":"Hi @eric-mitchell,\r\nI think the issue might be that this dataset was added during the community sprint and has not been released yet. It will be available with the v2 of `datasets`.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of `datasets` using pip:\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","body":"When loading the NarrativeQA dataset with `load_dataset('narrativeqa')` as given in the documentation [here](https:\/\/huggingface.co\/datasets\/narrativeqa), I receive a cascade of exceptions, ending with\r\n\r\n FileNotFoundError: Couldn't find file locally at narrativeqa\/narrativeqa.py, or remotely at \r\n https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/narrativeqa\/narrativeqa.py or \r\n https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/narrativeqa\/narrativeqa.py\r\n\r\nWorkaround: manually copy the `narrativeqa.py` builder into my local directory with \r\n\r\n curl https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/narrativeqa\/narrativeqa.py -o narrativeqa.py\r\n\r\nand load the dataset as `load_dataset('narrativeqa.py')` everything works fine. I'm on datasets v1.1.3 using Python 3.6.10.","comment_length":55,"text":"NarrativeQA fails to load with `load_dataset`\nWhen loading the NarrativeQA dataset with `load_dataset('narrativeqa')` as given in the documentation [here](https:\/\/huggingface.co\/datasets\/narrativeqa), I receive a cascade of exceptions, ending with\r\n\r\n FileNotFoundError: Couldn't find file locally at narrativeqa\/narrativeqa.py, or remotely at \r\n https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/narrativeqa\/narrativeqa.py or \r\n https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/narrativeqa\/narrativeqa.py\r\n\r\nWorkaround: manually copy the `narrativeqa.py` builder into my local directory with \r\n\r\n curl https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/narrativeqa\/narrativeqa.py -o narrativeqa.py\r\n\r\nand load the dataset as `load_dataset('narrativeqa.py')` everything works fine. I'm on datasets v1.1.3 using Python 3.6.10.\nHi @eric-mitchell,\r\nI think the issue might be that this dataset was added during the community sprint and has not been released yet. It will be available with the v2 of `datasets`.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of `datasets` using pip:\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","embeddings":[-0.2821941078,0.1054382026,0.0326368958,0.2439089417,0.190005973,0.1799186468,0.1333401948,0.0412888601,-0.1444729418,0.0441531092,-0.014136171,-0.0756954849,-0.0687526986,0.3853098452,0.0838763937,-0.1709612459,0.0442917012,0.0163581055,0.0471326895,0.0996936783,-0.2275328636,0.6004674435,-0.3060184121,0.0204679333,-0.3162917495,-0.0214897413,-0.1606631875,0.3630981743,-0.1869076639,-0.3998726904,0.6071943045,-0.0712309107,0.2232357115,0.5504921079,-0.00011022,0.0608154461,0.4106917381,-0.1400963366,-0.2848300636,-0.545355022,0.2669924498,-0.2112297714,0.2129349858,0.1392676383,-0.1671685725,-0.3874749839,0.2078933418,-0.0572460517,0.321726203,0.0858349651,0.1936791837,0.3349677324,0.2189273536,-0.1860184222,-0.1006725505,0.4217249751,-0.2878890634,0.5078809261,0.3055493236,-0.4217514992,0.1246747151,0.1989541054,-0.0314400643,-0.005937275,0.3751299381,-0.0012227655,0.0501069874,-0.2876365185,0.1067650765,0.3626172245,0.6540010571,-0.3735683262,-0.5450147986,-0.183354333,0.1010366306,-0.084696874,0.2412417382,0.2746928334,-0.2394714504,0.1392699778,-0.1500250399,-0.116059199,-0.1332204044,0.2248403281,-0.1042077318,-0.1651473194,-0.2865085304,0.0356239192,0.0348373204,-0.2221505195,-0.1596003324,0.1280231774,-0.2268224061,0.2955132127,-0.2854912579,0.1019492745,0.0155473538,0.225426659,0.1659238935,-0.2129214704,0.028150741,-0.0490586534,-0.0764428303,0.1599115729,0.2491157055,0.1045944691,0.3280425072,0.0166725665,0.0025238984,0.4532367289,-0.0647722185,-0.2219322473,-0.0902573466,-0.1459997892,-0.4528244436,-0.2933596671,0.4718654156,-0.30325827,-0.3878256679,-0.1083442271,-0.187420994,0.0374947675,0.0771065801,0.4981051683,0.0121543976,0.1539297551,0.0155747412,0.3006159663,-0.2579495609,0.1018595099,-0.2480658889,0.0056528724,-0.0319857895,0.1371661723,0.2997520268,-0.3352165818,0.3940825164,-0.0288877748,-0.0616462827,0.1528032422,-0.0062926491,0.0956953615,-0.2000017166,0.2249585986,-0.1231740341,0.0305550043,0.2513726354,-0.0039111096,-0.0630689114,-0.060995277,-0.1491312832,-0.2974407375,-0.0398866013,0.1490037739,-0.1069887877,-0.2064623535,-0.0754367337,0.1805878431,0.0128380544,-0.127744019,-0.088431038,0.0661217421,-0.0049029607,-0.0903493464,0.2024713159,0.6703977585,-0.2659667432,-0.3223662674,0.0665737167,-0.2062622905,-0.002002259,0.205178231,-0.2598896027,0.0763844475,-0.2033912838,-0.1399738789,0.0594103932,-0.4468630552,-0.2169282287,0.2421035022,-0.0552593172,0.409137398,0.0252109431,-0.0529107004,-0.0884344503,-0.1937901825,0.1179099977,0.2477221787,0.041410502,-0.180747062,-0.0769385174,-0.1732111126,0.1643526256,0.2965083718,-0.2560911775,0.0753234625,0.1443372518,-0.1822345108,0.1934351921,-0.0239516459,-0.0791850686,0.2923217118,0.2021273822,0.1244123653,0.0860616416,0.1987719983,-0.5315409899,0.2751492262,-0.2863561213,-0.0349830911,-0.1951927692,-0.0762917846,-0.1019935831,0.0727367327,-0.3476805091,-0.3338475525,0.0303023178,0.2231727242,0.051065281,0.1007084996,-0.3703853488,0.3647746444,-0.2149709016,0.1742464453,-0.5457684398,0.0198504608,0.059093833,-0.109520033,0.122496672,0.2437279373,0.0530007482,-0.2464805096,0.0897159576,0.4100808501,-0.2697848976,0.4130957425,0.0533893108,0.0024124011,0.1878376901,-0.3031235933,0.1404956579,-0.0539183393,0.0428540744,0.0091652777,-0.0933393985,0.2232495397,-0.1476233006,0.3639097214,0.0349284671,0.1885386556,0.2245209068,0.0839198828,-0.2038252205,-0.2347534746,0.434181571,-0.1314287931,0.3960675895,-0.0928729326,-0.3104823828,0.0197391361,0.1395178586,0.1080144867,-0.0624740012,0.0629142523,-0.3400149047,0.1228983179,0.087337561,-0.149281919,0.4034076035,0.1285380274,-0.0878837779,0.1276677847,-0.0151850637,-0.3860321939,0.1712200493,0.0358905382,-0.288698107,0.0302725509,0.0978229865,0.0187041406,-0.361423403,0.1283961535,-0.0121853743,0.0879093707,-0.2980452776,0.2551460266,-0.0776269361,-0.3964503109,-0.1199690253,-0.4994731545,-0.2983438969,-0.225864917,-0.1354241669,0.113485992,-0.0269375853,0.2017488629,0.3073233962,0.0679067969,-0.2341177016,-0.1906663924,-0.072332561,-0.2339591384,-0.1884654313,0.0955380574,0.296767801,-0.1169552803,0.2403816581,-0.3985351026,0.0019971929,-0.2876364291,-0.2706886232,0.1007398292,-0.2863255739,0.4284830689,0.0129717411,0.2201460004,-0.0836322457,0.0230730586,0.439858377,-0.0221727323,-0.1318078339,0.1779700667,-0.186746195,-0.1351217628,-0.1410122067,-0.1960711926,-0.30781129,-0.4054515958,0.3266366422,0.052409023,-0.1396427751,0.637440145,-0.0049181371,0.1109821573,-0.4007156491,0.0691167936,-0.0332454443,-0.5922862291,0.238506645,-0.3008897901,-0.2814723253,0.3023081422,0.1887840033,0.1527570635,0.2165230811,-0.4129439294,-0.1828762442,0.0208987184,0.0596235693,0.1238240898,0.1198371798,0.268009901,-0.1689155102,0.0583419874,0.0049574403,-0.1630619317,0.0573053621,-0.0954310149,0.2994501889,-0.1553699225,0.4328054786,-0.0993204266,0.5152155757,0.4010253847,0.0893512145,0.4676293433,-0.0574609526,0.5533753633,-0.1913194358,-0.3868480623,0.1434226632,0.0086409105,0.2190764397,0.2548012435,0.1087997556,0.2240044624,-0.2882213593,-0.23322393,-0.1790243983,-0.1051175818,-0.0885645747,-0.2226638347,0.5598247051,0.0422958732,0.1839492619,0.117294848,-0.171197325,0.0792702362,0.4047596455,0.3063110411,0.1532229185,0.0855194554,-0.0205499306,-0.5750069618,0.2002411038,-0.1005980447,0.2336289436,-0.1493739933,-0.0945477486,0.0758243278,-0.0533303134,0.5612584949,0.1697458178,-0.2268660516,0.0816910639,-0.0993022397,-0.7485927343,-0.0011172717,0.0711871088,0.5246649981,-0.2983523905,0.4550998211,-0.3250734508,-0.1037764475,-0.0049161837,0.280115217,-0.0287978854,-0.2050734609,-0.4159323573,-0.1526643783,-0.3211813569,-0.139282912,-0.2078924328,0.2848507464,0.0500270687,0.1092008427,-0.0233779363,-0.0825082958,0.1994103342,0.190078795,0.3466509581,-0.1118840873,0.3190791607,0.3728345335,0.1810114831,0.187776491,0.7771446705,0.15043284,-0.6387341619,-0.1201392412,0.1999837458,0.1508354396,0.2416085452,-0.0357491672,0.0066250395,0.1359489262,-0.1526327431,-0.3339592516,-0.0301000904,0.2143161446,-0.0788479969,-0.2795948386,-0.1993094236,0.5535469651,-0.1464820951,0.2264552414,0.1538517773,0.1967406124,-0.0732810795,0.0016761695,-0.1573602855,0.6390523911,-0.040313188,0.1748256832,0.3679147661,0.1531437039,0.2759170234,-0.0173136387,0.0256162789,-0.3230863512,0.2338577956,-0.058319252,-0.155377239,0.2595311999,0.0256213732,-0.1812216192,0.3464086652,-0.0382865556,0.2396208197,0.0347959734,0.1220935509,-0.2781035006,-0.277084738,-0.4702249467,0.1231399104,0.0421370044,0.3807178736,-0.2301452309,-0.0641089827,0.0069545615,-0.0681097358,-0.154933542,0.104905203,-0.4039308429,-0.0393745378,0.4102818072,-0.4239174426,0.1326032877,0.3218355179,0.296606034,-0.1543131322,-0.4461571872,0.2538772821,-0.3292990327,-0.4046239853,0.0833637863,-0.0610091127,0.1630056947,-0.0150951473,0.0507323071,0.1275271326,-0.1174856126,-0.0183891673,-0.1843884736,-0.0077184057,-0.067583032,-0.1112013683,-0.2188766003,-0.1859840006,0.2508607805,-0.1036204919,0.1100484207,0.2406917661,-0.0587901212,-0.1924421489,-0.2930653095,-0.1108549088,-0.1738733351,0.2027392387,-0.3240361214,0.0001264135,0.5568037033,0.3961308599,0.0718973801,-0.0772814378,-0.1610660553,0.3245172501,-0.4137250185,0.1049698815,0.5443444848,0.1243741289,0.1150998026,0.5972671509,0.0085318713,-0.488622576,-0.0376533195,-0.3821846545,-0.4004566371,0.3767909706,-0.0392235927,-0.0035063336,0.1043881103,0.3638806939,0.1484958529,-0.1086734012,-0.2812004387,0.0144610321,0.0260914676,0.0665959194,0.3481182754,0.0831234828,0.2603158951,0.1296880692,0.0548093431,0.0585671701,-0.1510332674,-0.1564381868,0.0636930987,0.109740667,0.0839065909,-0.1260282546,0.030983679,-0.0964156762,-0.1620913297,-0.2794083059,0.2913079262,0.1293537021,-0.0418578982,0.0481650531,-0.0309824124,0.2061075568,-0.1004877985,0.1286318302,-0.0037093274,0.1389215738,0.0646109655,-0.1588986218,-0.2791421115,-0.0699502304,0.1689421833,0.1885785908,0.0188460574,0.064279072,0.2225042135,-0.2409181595,0.0808429718,0.0880453587,0.1194436848,0.3508829176,-0.2469259053,0.2167237401,0.1472097188,0.1262485534,-0.2034908831,-0.0014663305,-0.1817890108,0.147103399,-0.0147106452,0.0656108856,0.1921749562,-0.1525169164,0.1016340628,0.1648362279,0.2551507354,0.0291612428,0.2111836672,-0.0184635129,0.0826422796,0.0121679883,0.1049643755,0.100631848,0.2894346714,0.2616158426,-0.3571737707,-0.035940405,-0.1055735275,0.2749790847,-0.0500702597,-0.4363563359,-0.0092428513,-0.0109622171,0.2346219569,0.0550125651,-0.2336055636,0.6237167716,-0.3811263144,0.2054310888,-0.2912210524,0.0795696452,-0.1503755301,-0.1913666576,-0.0197431147,-0.1661552191,0.1980034113,-0.0751412064,-0.1810738295,-0.2111661136,0.0607774742,-0.0983832628,-0.1694709212,-0.4221380055,0.1217107773,0.3971255124,0.0550698452,-0.0399877392,0.5224871635,0.291990608,-0.0746912882,0.2625925541,0.4273853302,0.5418922305,0.1653062254,-0.0831643417,0.2354131192,-0.3207223713,-0.1076363921,0.0870465711,0.3370732963,-0.1440215558,0.147631377,0.2105644643,0.1527604759,-0.081590496,-0.079956159,-0.0137942843,0.432841152,-0.3931150436,0.0917451084,-0.5244226456,0.0656390712,-0.2222937793,0.1447220892,-0.5138548017,0.1262402087,0.1807167828,0.1617513895,-0.1603774279,-0.1651047766,0.0718926042,-0.0402664877,0.4889549613,0.5259319544,0.2498774976,-0.166514501,-0.319968164,-0.7123633623,0.0642145947,-0.068987377,0.0215055011,0.0160256643,0.1856743246,0.1465887874,0.1642286181,-0.1459349543,0.2376634777,0.0286108255,-0.2953656316,-0.2187983096,-0.2316807806,-0.0019517813,-0.0532662384,-0.0607996657,0.0391262285,0.0842231363,-0.1931580901,-0.0525390729,-0.0814345479,0.0231455453,0.026958216,-0.3341568112,0.2552153468,0.0374402665,-0.0859819353,0.0870682597,-0.2809714079,-0.4722728133,-0.3770329058,-0.0823964402,0.2043832988,-0.1128700823,0.4890654087,-0.3747421801,-0.1170655042,-0.2143312395,0.1286319643,0.057052739,-0.0012537003,-0.1321219504,0.2001547515,0.0097506912,-0.0023391431,0.0858018026,0.4958351254,0.0586853512,0.0648389906,-0.2739212215,-0.2833625972,0.4915392995,-0.3666218519,-0.0029920216,0.2970085442,0.0630753487,0.0621237494,-0.2294496745,-0.7798358798,0.3636758924,0.2381771356,0.1048714668,-0.0375920273,0.0724144578,0.0000913051,0.0771133155,-0.1479493082,0.2218929082,0.193659842,0.0095020067,0.2759428322,0.043186266]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1647","title":"NarrativeQA fails to load with `load_dataset`","comments":"@bhavitvyamalik Great, thanks for this! Confirmed that the problem is resolved on master at [cbbda53](https:\/\/github.com\/huggingface\/datasets\/commit\/cbbda53ac1520b01f0f67ed6017003936c41ec59).","body":"When loading the NarrativeQA dataset with `load_dataset('narrativeqa')` as given in the documentation [here](https:\/\/huggingface.co\/datasets\/narrativeqa), I receive a cascade of exceptions, ending with\r\n\r\n FileNotFoundError: Couldn't find file locally at narrativeqa\/narrativeqa.py, or remotely at \r\n https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/narrativeqa\/narrativeqa.py or \r\n https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/narrativeqa\/narrativeqa.py\r\n\r\nWorkaround: manually copy the `narrativeqa.py` builder into my local directory with \r\n\r\n curl https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/narrativeqa\/narrativeqa.py -o narrativeqa.py\r\n\r\nand load the dataset as `load_dataset('narrativeqa.py')` everything works fine. I'm on datasets v1.1.3 using Python 3.6.10.","comment_length":15,"text":"NarrativeQA fails to load with `load_dataset`\nWhen loading the NarrativeQA dataset with `load_dataset('narrativeqa')` as given in the documentation [here](https:\/\/huggingface.co\/datasets\/narrativeqa), I receive a cascade of exceptions, ending with\r\n\r\n FileNotFoundError: Couldn't find file locally at narrativeqa\/narrativeqa.py, or remotely at \r\n https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/narrativeqa\/narrativeqa.py or \r\n https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/narrativeqa\/narrativeqa.py\r\n\r\nWorkaround: manually copy the `narrativeqa.py` builder into my local directory with \r\n\r\n curl https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/narrativeqa\/narrativeqa.py -o narrativeqa.py\r\n\r\nand load the dataset as `load_dataset('narrativeqa.py')` everything works fine. I'm on datasets v1.1.3 using Python 3.6.10.\n@bhavitvyamalik Great, thanks for this! Confirmed that the problem is resolved on master at [cbbda53](https:\/\/github.com\/huggingface\/datasets\/commit\/cbbda53ac1520b01f0f67ed6017003936c41ec59).","embeddings":[-0.2146174461,0.1621950269,0.0469725914,0.3080255389,0.2021450698,0.169084534,0.2221396714,0.0322584324,-0.103125982,0.0728139579,0.0291966219,-0.1251862049,-0.0299643874,0.3841863275,0.0823540092,-0.1388388425,0.0632621124,0.0123042315,-0.0613596551,0.1301435977,-0.2554582953,0.5650151968,-0.3087537587,0.0355321765,-0.3673131764,0.0776510313,-0.2068326622,0.4470644593,-0.1237637252,-0.3570647538,0.6157504916,-0.0831176266,0.1994945705,0.6125927567,-0.000114849,0.1384543031,0.408637464,-0.1238553002,-0.3053527474,-0.5542643666,0.1990433633,-0.1439836174,0.2387083769,0.117693983,-0.148766309,-0.3430629969,0.2187850326,-0.0382941999,0.2938016653,0.0769625679,0.1457052976,0.3507170677,0.177833572,-0.2063507587,-0.1319084764,0.3832830191,-0.2404083163,0.5567546487,0.2944138646,-0.4671268761,0.027822556,0.1364254802,-0.0403410308,-0.0717985332,0.353333205,-0.0810704827,0.0842030197,-0.2240511477,0.1320184618,0.2990549505,0.6650859714,-0.4016878009,-0.5489897728,-0.2329103202,0.0682106763,0.0124167837,0.3016398847,0.2311384678,-0.2660416067,0.1436997354,-0.1927509457,-0.141125083,-0.1281029284,0.2137604654,-0.1912476569,-0.2185758352,-0.2847173512,0.0178695172,0.0121312123,-0.1582885385,-0.1042521074,0.0220639072,-0.2405802459,0.2440605611,-0.3618841171,0.1149763241,0.057331454,0.2311739177,0.137490198,-0.1825457215,0.0808030292,-0.0564137511,-0.1437974125,0.2165841162,0.2996948957,0.1469930261,0.3425021172,-0.0270491019,0.0094865393,0.4379321337,-0.1447564512,-0.2090357095,-0.0926615,-0.2464589179,-0.4077750146,-0.2653168142,0.4451967478,-0.3163428307,-0.3708595932,-0.064520739,-0.2783451378,0.0840341672,0.1538729817,0.4943803549,0.037317358,0.1750158817,0.0117192874,0.2877277136,-0.2577525675,0.1054750606,-0.2827109098,-0.0197969954,-0.0996643677,0.1394269764,0.2734495103,-0.3940111399,0.4211376309,0.0197580308,-0.0570916049,0.0674900189,-0.0350548849,0.1052402854,-0.1928251684,0.2470906228,-0.1226534471,0.0505480431,0.2573309243,0.0499989949,-0.0173703041,-0.0448864326,-0.1615893096,-0.3788504601,-0.0054585892,0.1200404614,-0.1170029491,-0.1644010395,-0.1368739605,0.1259985715,0.0317245387,-0.1399798393,-0.0896632746,0.0444329381,-0.0163216311,-0.0951922908,0.2221472412,0.735368073,-0.2197562605,-0.2599994242,0.0744204223,-0.1951038688,-0.012916266,0.2640272379,-0.2990819216,0.1117490903,-0.3075962961,-0.1000628993,0.0635027513,-0.426094234,-0.2178532928,0.3083328009,-0.0923927873,0.3939522207,0.079688035,-0.1241802499,-0.1295693666,-0.2833921611,0.1085198447,0.2146522701,0.0405110195,-0.1524080783,-0.0743230432,-0.1279241294,0.1148213968,0.3057292998,-0.2773502469,0.0831520036,0.2071806043,-0.1830666661,0.1444313824,-0.0727255642,-0.1193615645,0.3105682135,0.1420010477,0.0390350185,0.0583795421,0.2525664866,-0.5856948495,0.3229648769,-0.1995595694,-0.0202266127,-0.2081411034,-0.0651887208,-0.0880818889,0.0970029831,-0.3952653706,-0.2827707529,0.0033317667,0.1922079027,-0.0563780032,0.0967646316,-0.3770219088,0.3486073613,-0.2858105004,0.2077604085,-0.5515011549,0.0119297551,0.1011612713,-0.1136537641,0.0857456848,0.2023074478,0.1008401811,-0.2398427576,0.0222162511,0.436226517,-0.2940654457,0.4611859918,0.0441422388,-0.0169496071,0.2168501467,-0.2394043207,0.1074519455,-0.0062757218,0.0134399179,-0.0078331949,-0.1315651983,0.1931906343,-0.1191278547,0.4174310565,0.0629791543,0.1257967651,0.1674371809,0.0682838112,-0.1594029814,-0.2018905431,0.481660068,-0.0612480454,0.4410198629,0.0328838341,-0.2958055735,0.008220844,0.109601967,0.0405688137,-0.0726576075,0.0754446164,-0.3002235889,0.0911761746,0.1303593069,-0.078536205,0.4333189428,0.0693844855,-0.0300202034,0.1547783166,-0.0282457247,-0.4389360845,0.2356312722,0.0272422638,-0.2732403576,0.1139834151,0.0955523998,0.0208956543,-0.3669918776,0.0474730805,0.0337353051,0.0470863655,-0.3688087463,0.2499431223,-0.1245177239,-0.401827693,-0.1399184018,-0.4225893617,-0.2988222539,-0.2281948924,-0.1726802737,0.2035043687,-0.1204859167,0.1782665849,0.2477364838,0.0940613151,-0.2449440807,-0.2789854705,-0.0715118647,-0.2198172063,-0.1797338426,0.0273520388,0.3871209919,-0.1646442562,0.1834332198,-0.3827611506,-0.0687506497,-0.277572155,-0.2188063115,0.1348913908,-0.1599506736,0.4039121866,0.0292279292,0.2063154578,-0.0765430704,0.0076923654,0.4492745697,0.0775068849,-0.159770444,0.2132053375,-0.1320456266,-0.0816050619,-0.1577841192,-0.1657156646,-0.2101219445,-0.426560998,0.2500154674,-0.0910461321,-0.092654258,0.6166639328,0.004651051,0.08927349,-0.385927856,0.070687227,0.0219043382,-0.6743842363,0.2216651291,-0.2568488717,-0.2252064049,0.2474649847,0.1921599805,0.1570555121,0.2915989757,-0.4048011303,-0.1714767665,0.062208835,0.0561975241,0.1158516333,0.1912910491,0.2059281915,-0.13010782,0.0826881826,0.0281859916,-0.1721683741,0.1100228727,-0.0622509308,0.2831320465,-0.1431958973,0.4539726377,-0.098724775,0.5330193043,0.3615877032,0.0840133876,0.5270130038,0.0332350582,0.5299533606,-0.2488363981,-0.3644041717,0.1072978303,0.0029711905,0.1466450691,0.235255152,0.1280418187,0.2274150699,-0.283578217,-0.3866213858,-0.2305119485,-0.1620416641,-0.052501034,-0.2758104801,0.5465677381,0.0055164169,0.1790507287,0.1705774665,-0.1023233831,0.0717407912,0.4783906043,0.3385443687,0.1209379211,0.1416810006,-0.0282678436,-0.5779919028,0.1698120385,-0.1130734384,0.3064842522,-0.1265113056,-0.0579124615,0.0574075803,-0.0623510145,0.6362998486,0.1348244548,-0.1851315498,0.0258076359,-0.186287269,-0.6747944951,-0.0075316955,0.1693586707,0.5964778066,-0.2711527348,0.4412065148,-0.3415164053,-0.1504614651,-0.0846861303,0.2501829267,0.011067017,-0.2838842273,-0.3590967059,-0.1527989358,-0.2953270078,-0.1433203071,-0.1690993458,0.2608083785,0.0259243008,0.0974371061,-0.0716464669,-0.0736684129,0.1769043505,0.1483183503,0.3201086223,-0.1983536482,0.3618120551,0.3499523699,0.2348385155,0.2460372001,0.7621443868,0.1836948395,-0.6073385477,-0.0746026933,0.2376005799,0.187359944,0.3218595386,-0.0312619694,-0.0110278148,0.11505799,-0.1895249933,-0.2616353631,-0.0623114035,0.1828150749,-0.090976052,-0.2991221547,-0.226908803,0.5725125074,-0.1815988719,0.2102489024,0.1859620512,0.1739569306,-0.1209988073,0.0738930851,-0.1936712414,0.6427900195,-0.0052243648,0.235856235,0.3127553463,0.1965877861,0.3089160025,-0.0041255774,0.0445766374,-0.3982970417,0.2030352205,-0.0756714121,-0.1754728854,0.2166680694,0.1020744592,-0.0711062476,0.4338657856,-0.0329049639,0.282284677,0.0951789618,0.0698153153,-0.3119854927,-0.2750599086,-0.4901397228,0.0664377511,0.1039369926,0.382106632,-0.2585290074,-0.0889499336,0.0673580319,-0.0326871723,-0.149503991,0.1522733867,-0.4771089554,-0.0611595549,0.3813790083,-0.433074683,0.1843552887,0.3204255104,0.2854408622,-0.1109440401,-0.4035055637,0.2071638405,-0.2614501417,-0.2983446419,0.122514464,-0.0834669918,0.1966561675,0.0306285676,0.0580132939,0.0717354193,-0.1479816437,-0.0626849234,-0.1410788,0.0039058055,-0.1419255137,-0.0912808478,-0.3038577139,-0.2230893672,0.2321185321,-0.1176489294,0.0783443302,0.2039361745,-0.0537172183,-0.185981676,-0.3743906319,-0.1384998113,-0.155934006,0.287420541,-0.4009866714,0.0391904935,0.5536509752,0.3473566771,0.1075940281,-0.0485558137,-0.1260890067,0.3787122965,-0.3931472301,0.1417203546,0.5003685355,0.0832104832,0.1458871216,0.5787090659,0.0265984796,-0.4708382785,-0.0205160175,-0.4227881134,-0.374252975,0.3393880129,-0.014589279,0.0151414555,0.1070604622,0.3381356895,0.1141017154,-0.2151004821,-0.2370010167,0.0396194309,0.0000489417,0.0543142855,0.356471628,0.1638313085,0.2628135085,0.0966783985,0.0229096487,0.0894809812,-0.1203771085,-0.1286022663,0.0804093555,0.1275303364,0.1376614422,-0.0657070875,0.0254489966,-0.1241058484,-0.1004903466,-0.2596068978,0.2955764234,0.1360803843,-0.03385479,0.001025568,-0.0531112142,0.2549782097,-0.2354426533,0.2052734047,-0.0013956552,0.1115899459,0.045069579,-0.1806114614,-0.3205581307,-0.0623727106,0.1998338997,0.1362001747,-0.0737061054,0.1318746507,0.2352555096,-0.2472324669,0.0349434204,0.066586636,0.1725869626,0.3456260562,-0.2550762296,0.1907642037,0.167418927,0.1002467051,-0.1097235084,-0.0239301622,-0.1150398627,0.1351688951,-0.0069649201,0.0809305906,0.2123229504,-0.2478269786,0.0750644132,0.1526243985,0.2110058218,0.1399419904,0.1826868057,0.0981250703,0.1237965152,0.083442159,0.1007994115,0.0721094683,0.3199776411,0.3281069398,-0.3053922653,-0.023678286,-0.1173122674,0.2757751942,-0.0800385475,-0.4861638248,0.0719867125,0.0340557694,0.2367089838,0.0175772682,-0.2896146476,0.6188409328,-0.4086520672,0.1708027571,-0.2145146132,0.0755852461,-0.2199035585,-0.1818612665,-0.1147664785,-0.1545165181,0.212845102,-0.0028064274,-0.2620030344,-0.1782354563,0.0483921468,-0.0540687814,-0.1229379326,-0.3825924397,0.0483867861,0.3731901348,0.0301401131,-0.0544938259,0.5565445423,0.3213451505,-0.1277495474,0.2489935309,0.396607548,0.5509128571,0.1886550188,-0.1313251704,0.1760171652,-0.2741215229,-0.0742812604,0.0741349012,0.2771350145,-0.2073177248,0.1415545791,0.2127040476,0.1096521318,-0.0533019081,-0.0182497203,0.028615417,0.3480343521,-0.4390378296,0.1017834917,-0.4681577384,0.0401121303,-0.1791360676,0.1357258856,-0.4765306115,0.0706511438,0.2811730504,0.1465136707,-0.1551318169,-0.1196208596,0.0539175868,0.0072260834,0.542480588,0.552113235,0.211193189,-0.2412104458,-0.3233866394,-0.7340269685,0.1115999818,-0.038914144,0.0910504162,0.0555803962,0.1410765797,0.1251658052,0.1048565209,-0.1580174416,0.3051593304,0.0879080743,-0.194863528,-0.220454514,-0.25251019,-0.053110905,-0.1248834208,-0.1028793827,-0.0417206623,0.1304092556,-0.2274564952,-0.068707101,-0.0489047468,0.0816658586,0.0335287675,-0.3293924332,0.2416538745,0.0205731168,-0.0748860016,0.0924659669,-0.3012628257,-0.3769254386,-0.3255306184,-0.1103333533,0.1994029284,-0.1206556782,0.476130873,-0.3579111397,-0.0742779151,-0.2165854126,0.1418120265,0.0997277275,-0.0331582427,-0.1861753017,0.1787978709,-0.0187120792,-0.0330479704,0.1263455153,0.5544756651,0.0174622089,0.1222795472,-0.3360943496,-0.3248312771,0.5100073814,-0.4035036564,0.0199378151,0.2556400299,0.109710373,0.063278392,-0.1745129228,-0.7413820028,0.3158416748,0.1933721155,0.0889077634,-0.0852426738,0.0519512333,0.0333435796,0.1355819404,-0.1307328343,0.1727855802,0.2235262394,-0.04869131,0.2874524593,0.0256264675]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1647","title":"NarrativeQA fails to load with `load_dataset`","comments":"Update: HuggingFace did an intermediate release yesterday just before the v2.0.\r\n\r\nTo load it you can just update `datasets`\r\n\r\n`pip install --upgrade datasets`","body":"When loading the NarrativeQA dataset with `load_dataset('narrativeqa')` as given in the documentation [here](https:\/\/huggingface.co\/datasets\/narrativeqa), I receive a cascade of exceptions, ending with\r\n\r\n FileNotFoundError: Couldn't find file locally at narrativeqa\/narrativeqa.py, or remotely at \r\n https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/narrativeqa\/narrativeqa.py or \r\n https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/narrativeqa\/narrativeqa.py\r\n\r\nWorkaround: manually copy the `narrativeqa.py` builder into my local directory with \r\n\r\n curl https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/narrativeqa\/narrativeqa.py -o narrativeqa.py\r\n\r\nand load the dataset as `load_dataset('narrativeqa.py')` everything works fine. I'm on datasets v1.1.3 using Python 3.6.10.","comment_length":23,"text":"NarrativeQA fails to load with `load_dataset`\nWhen loading the NarrativeQA dataset with `load_dataset('narrativeqa')` as given in the documentation [here](https:\/\/huggingface.co\/datasets\/narrativeqa), I receive a cascade of exceptions, ending with\r\n\r\n FileNotFoundError: Couldn't find file locally at narrativeqa\/narrativeqa.py, or remotely at \r\n https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/narrativeqa\/narrativeqa.py or \r\n https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/narrativeqa\/narrativeqa.py\r\n\r\nWorkaround: manually copy the `narrativeqa.py` builder into my local directory with \r\n\r\n curl https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/narrativeqa\/narrativeqa.py -o narrativeqa.py\r\n\r\nand load the dataset as `load_dataset('narrativeqa.py')` everything works fine. I'm on datasets v1.1.3 using Python 3.6.10.\nUpdate: HuggingFace did an intermediate release yesterday just before the v2.0.\r\n\r\nTo load it you can just update `datasets`\r\n\r\n`pip install --upgrade datasets`","embeddings":[-0.2199551761,0.0480084457,0.0608192459,0.312422812,0.1940941066,0.1517978162,0.1400994509,0.0211620592,-0.1197655275,0.0751328319,0.0060817758,-0.1055012122,-0.0284701269,0.328783989,0.1128825396,-0.1749626398,0.0848928317,0.0000286327,-0.0132892365,0.1011296585,-0.2477790415,0.5799874067,-0.3045037091,0.0495427139,-0.3111803234,0.0400928035,-0.1827201545,0.4143431485,-0.134498775,-0.3618732095,0.5955119133,-0.0901765451,0.1628862619,0.5966639519,-0.0001159591,0.0845015123,0.3932443559,-0.1201084405,-0.2960406244,-0.5694472194,0.2616091669,-0.1980767101,0.2148328125,0.1369700134,-0.1391425133,-0.3809946775,0.2064242959,-0.015644094,0.3636644483,0.0311052669,0.1502898186,0.3551585078,0.1913443357,-0.1960780472,-0.1232333258,0.4534870386,-0.2778141499,0.5499457717,0.3091154397,-0.4308938682,0.0927215219,0.1659913957,-0.0754026026,-0.0536760539,0.39754951,-0.024482822,0.0316641182,-0.2200985551,0.1294304579,0.3337748051,0.5657204986,-0.3766566217,-0.5654232502,-0.2516533732,0.0828977525,-0.0521907918,0.3052110076,0.2050047815,-0.2722261846,0.1463726908,-0.2035996914,-0.1753699481,-0.1267252564,0.1935858279,-0.1112736091,-0.2434892654,-0.2949478626,0.0435590893,0.0719450936,-0.1950484067,-0.1617371589,0.0534055047,-0.238451764,0.2782220542,-0.2544952929,0.1104443818,0.0095407376,0.2557695806,0.1553553045,-0.2342424244,0.0359687246,-0.0622371808,-0.0856734887,0.1938934624,0.2793362737,0.1482362747,0.3574495912,-0.0253055226,-0.0095609967,0.4405895472,-0.0695300847,-0.208296597,-0.0765138119,-0.1682606637,-0.4591735303,-0.3255860507,0.4318531752,-0.3057953417,-0.4052734077,-0.0722541139,-0.2397596389,0.0862227082,0.1374139935,0.4865350425,0.0081579303,0.1582750976,0.0058746026,0.2865971923,-0.2645154595,0.0871506706,-0.2418975085,-0.0022812472,-0.0693875924,0.1570147574,0.2398714125,-0.3612253666,0.3871788979,-0.017591022,-0.0449736305,0.0816111639,-0.0984736979,0.1081791595,-0.2023358643,0.2186914086,-0.1624392867,0.0317286365,0.2772150636,-0.0326874964,-0.0554952323,-0.0622329935,-0.137747556,-0.3022783101,-0.0104777804,0.1069138348,-0.0946805701,-0.1699189395,-0.1125650629,0.1576721817,-0.0130356383,-0.1991377473,-0.0738989785,0.0706882998,0.0106178634,-0.073409833,0.2454290986,0.7580022812,-0.2056699097,-0.312558651,0.071137473,-0.1893505454,0.0001482788,0.2615221441,-0.2317508012,0.0680976883,-0.2531420588,-0.1607623249,0.0935508609,-0.4468418658,-0.1955764443,0.2799386978,-0.0556045361,0.4567216039,0.0581239238,-0.0801238641,-0.1273096502,-0.2414393574,0.1230968684,0.1893331259,0.024409499,-0.206517607,-0.0782089755,-0.1571550071,0.1286076456,0.2847905159,-0.2935614586,0.077220507,0.145658657,-0.1646689028,0.1315743774,-0.0165994577,-0.1360207945,0.2849665582,0.1409733444,0.1329290271,0.0182084814,0.2053834796,-0.5579439998,0.2882403731,-0.2570711672,-0.0820594132,-0.1659240127,-0.0840731114,-0.0773310065,0.0983150154,-0.3395567238,-0.3217682838,-0.012640723,0.2045415044,0.0194325894,0.1464295685,-0.3868191838,0.3574392796,-0.2472709864,0.2067177594,-0.5592110753,-0.0302885082,0.0747719556,-0.1133629307,0.0375892036,0.26366961,0.0972482264,-0.2482202649,0.0932326689,0.459361881,-0.2942497432,0.4402934313,0.0197736491,-0.0383595861,0.230896011,-0.2425242513,0.0496424139,-0.0545340925,-0.009617541,0.0093187159,-0.1185193583,0.1728489101,-0.1332919747,0.4073511362,0.0459632091,0.181058377,0.1797826737,0.0661977828,-0.181844309,-0.2293711901,0.4912464023,-0.1180183291,0.4162700772,-0.0020716565,-0.3156051636,-0.017534731,0.1378815174,0.0692394525,-0.0539511181,0.079073526,-0.3096742332,0.1229100451,0.1191897839,-0.140136838,0.3699494004,0.0957033783,-0.0847904533,0.1523884386,-0.0509501025,-0.3968065381,0.2081427425,0.0215498786,-0.3026605844,0.048985716,0.0666019022,0.0369160734,-0.3345046937,0.0931314453,-0.0384076945,0.0460374355,-0.3374765217,0.2550656497,-0.1003413945,-0.4144649804,-0.1569963396,-0.4590072334,-0.3220498264,-0.2167787254,-0.1421779394,0.19984442,-0.0557115711,0.1832412779,0.3134799898,0.0883676261,-0.2705671489,-0.2283976227,-0.1064613387,-0.1779920757,-0.2053411752,0.0222982261,0.3676596284,-0.1473156363,0.2259963453,-0.4360030293,-0.071560286,-0.2555087805,-0.2175281793,0.1356736869,-0.2151544541,0.431124419,0.0441277251,0.1951391399,-0.1098990589,-0.0055347406,0.5068662167,0.0420540832,-0.1580023468,0.2077817768,-0.1208435148,-0.1133539528,-0.1543784589,-0.1071836203,-0.2348463386,-0.4226069748,0.3171221316,-0.0450610556,-0.116146937,0.6448138356,0.0387850404,0.1422766447,-0.4684151113,0.0847676247,0.0091073811,-0.5962807536,0.2181991339,-0.2305508703,-0.2328947932,0.2591850162,0.2447499633,0.0769027323,0.2677282095,-0.3939247429,-0.2405212373,0.0683698058,0.0462512821,0.1119177192,0.1560761034,0.2483342141,-0.1363087296,0.0755938292,0.0432069562,-0.1572255045,0.0981369168,-0.1003512517,0.3137177229,-0.1538291126,0.4582180083,-0.1005096957,0.5343384147,0.4117761254,0.0967059731,0.4817893803,-0.0264523011,0.581859529,-0.2475986183,-0.4075126052,0.0939726904,0.0179189742,0.1671089679,0.2258089781,0.140760988,0.2846951485,-0.3021670282,-0.309294492,-0.1881640553,-0.1235716045,-0.0852918923,-0.2234611064,0.5553427935,0.0112731475,0.2170986682,0.1922613829,-0.1264065951,0.087604925,0.4325501621,0.3353114426,0.152274549,0.0859318972,-0.0275147967,-0.5519149303,0.1919887811,-0.0515101254,0.2728056908,-0.1816072464,-0.0822812617,0.065009512,-0.0405395105,0.6027062535,0.2003560811,-0.2387973964,0.0110560469,-0.1418334693,-0.7813404799,0.0276849009,0.1436466426,0.5761937499,-0.2153860927,0.4796074033,-0.3531573415,-0.1438581049,-0.0496489219,0.3080472052,0.0000997901,-0.22613056,-0.4201829433,-0.1625661254,-0.2803269923,-0.1416435689,-0.1996248513,0.2476374507,0.093294248,0.0994479209,-0.0437136069,-0.0913077295,0.1925930828,0.1824090481,0.3472195268,-0.2103386819,0.2896926701,0.3869659901,0.2320165187,0.2081537992,0.7227814198,0.1646444201,-0.5916885734,-0.1534425765,0.2696811259,0.1832048893,0.2895334363,-0.0287019983,0.0045344103,0.1195639521,-0.1513383538,-0.3312912583,-0.0692972839,0.2510878444,-0.0724763423,-0.2780544162,-0.1971238106,0.5810284019,-0.1712813973,0.2124407589,0.1725043058,0.2566058636,-0.0607082956,0.0138224466,-0.1972826272,0.6852072477,-0.0016984127,0.2276739776,0.3490580916,0.1740031689,0.3668690026,0.0413299836,0.0558345914,-0.3259003162,0.2142915875,-0.0621550158,-0.1550914794,0.2504835427,0.0143205281,-0.1438208073,0.4169284999,-0.0492007062,0.2519986629,0.0512324013,0.0585064813,-0.3035952449,-0.2866882384,-0.49859038,0.06551525,0.0861330107,0.3852570057,-0.2290011346,-0.055278495,0.043406941,-0.0650592893,-0.1822628379,0.0976705849,-0.4539618492,-0.0852340311,0.4453539848,-0.4227737784,0.132439211,0.3079871237,0.2770318985,-0.1352959722,-0.4182508588,0.1905224919,-0.3107000589,-0.3352680206,0.1088909134,-0.0673647746,0.1250195354,0.018082859,0.0961855724,0.0950738043,-0.1389359385,-0.0259650927,-0.1720430702,0.0189174488,-0.1168286726,-0.091314666,-0.2645078003,-0.2000600547,0.2647669017,-0.113291055,0.073735252,0.1806695014,-0.0412886888,-0.2020790577,-0.3155838847,-0.113793008,-0.1568174213,0.2447517961,-0.4322889149,0.029142512,0.5220119953,0.3119788766,0.0903040618,-0.0378106162,-0.1419821829,0.326726675,-0.4389435649,0.0869346112,0.4958404005,0.0871375874,0.0812375769,0.6029357314,0.0888436213,-0.5001256466,-0.0295987539,-0.3501308858,-0.3667313457,0.3960722387,0.0544702262,-0.0350407064,0.1130215377,0.3675639331,0.1275762767,-0.1201763526,-0.2314170897,-0.0022265255,0.0070114215,0.0237147678,0.3608201742,0.1345325112,0.2697116137,0.0971635655,0.0258939993,0.0691878423,-0.129836157,-0.1108961478,0.0662152544,0.1297897547,0.0817599222,-0.1085813195,0.0188027006,-0.0833506435,-0.0660318807,-0.2945555449,0.2419247627,0.1824625134,-0.0405431353,-0.0263476111,-0.0977531075,0.3069406748,-0.1608067006,0.149248302,0.0016987051,0.174300015,0.066671893,-0.2106344253,-0.2734610438,-0.0773224086,0.1642080098,0.1716503799,0.0075364131,0.0281192288,0.204660058,-0.1968458146,0.0656945184,0.1100027785,0.1370832622,0.3960878253,-0.2315135747,0.2100993842,0.1842642128,0.0929268077,-0.1274147928,-0.0038554207,-0.1139460281,0.1378071755,-0.0260082781,0.1068404317,0.1758494526,-0.2236815542,0.1428069472,0.132709071,0.2636273503,0.1022378653,0.1659463495,0.00606019,0.1174133345,0.0537047684,0.0871424973,0.0680989772,0.3419531286,0.2420545816,-0.3453662097,-0.0766909346,-0.1559669971,0.3327166736,-0.0734205842,-0.4263690412,-0.003476229,0.0283628237,0.3028205335,0.0520639271,-0.309976697,0.6268877983,-0.4704827666,0.1806904972,-0.2335541397,0.1000594422,-0.1739255339,-0.2023798376,-0.0350860283,-0.1601095051,0.1972403824,-0.0691827759,-0.1901581138,-0.2009134442,0.0884609818,-0.0684195459,-0.1163364872,-0.3947867453,0.0498186834,0.3899997473,0.0864816606,-0.0608530305,0.5724349618,0.3500392735,-0.0964195579,0.2373733222,0.4372108281,0.5460038781,0.2139693201,-0.0959323496,0.1794986129,-0.341055274,-0.0520879515,0.0628493056,0.3029104173,-0.1970634013,0.1611323655,0.218563512,0.1102535725,-0.0631715357,-0.0313356407,-0.0528936908,0.3784719408,-0.4556914866,0.1444167793,-0.5341811776,0.0613518022,-0.2188896537,0.1397595108,-0.5079593658,0.1194732711,0.1734573096,0.1381901354,-0.1787901372,-0.1093848348,0.0446788631,0.0417435542,0.4834540784,0.5565122366,0.2354832441,-0.1938709021,-0.3390642107,-0.7109742761,0.1004295647,-0.0644232258,0.0836052224,0.0371627733,0.1825920343,0.1044417098,0.1517058164,-0.1548493356,0.2474969029,0.0749198273,-0.2547399104,-0.1842651665,-0.2168312371,-0.0328449309,-0.1198426038,-0.0634594709,0.0122602666,0.135617137,-0.2194178104,-0.0862190351,-0.0113357389,0.1002514064,-0.0181537624,-0.3268493116,0.2521870434,0.0448701605,-0.11176797,0.1069106832,-0.3588003218,-0.3740367591,-0.3377349377,-0.0780079737,0.1916562319,-0.1386501491,0.4262621701,-0.3807392418,-0.1261508465,-0.2059417665,0.1716354042,0.094605051,-0.0150345284,-0.1638607532,0.1481854916,0.0023829474,-0.0677164942,0.118384257,0.5463752151,0.0304560345,0.1110923439,-0.3167284131,-0.2582370937,0.5261232853,-0.397346884,0.0353844166,0.3021899462,0.0952848494,0.1364236176,-0.2403280884,-0.7155963182,0.2655899823,0.2215318531,0.1262856424,-0.0346115828,0.0403748564,-0.0620652735,0.069711931,-0.1273891032,0.2238001227,0.2203967273,-0.0228032283,0.3593526185,0.0462817512]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1644","title":"HoVeR dataset fails to load","comments":"Hover was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `hover` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"hover\")\r\n```","body":"Hi! I'm getting an error when trying to load **HoVeR** dataset. Another one (**SQuAD**) does work for me. I'm using the latest (1.1.3) version of the library.\r\n\r\nSteps to reproduce the error:\r\n\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"hover\")\r\nTraceback (most recent call last):\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/hover\/hover.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/hover\/hover.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at hover\/hover.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/hover\/hover.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/hover\/hover.py\r\n```","comment_length":40,"text":"HoVeR dataset fails to load\nHi! I'm getting an error when trying to load **HoVeR** dataset. Another one (**SQuAD**) does work for me. I'm using the latest (1.1.3) version of the library.\r\n\r\nSteps to reproduce the error:\r\n\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"hover\")\r\nTraceback (most recent call last):\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/hover\/hover.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/hover\/hover.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/Users\/urikz\/anaconda\/envs\/mentionmemory\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at hover\/hover.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/hover\/hover.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/hover\/hover.py\r\n```\nHover was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `hover` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"hover\")\r\n```","embeddings":[-0.221469745,0.0588316135,0.0173894744,0.2946339846,0.2871959507,0.1036297306,0.2778086066,0.2113136351,0.055386655,0.0441018417,-0.1638761312,-0.0258904286,0.0066259452,-0.1750875711,0.3238515556,-0.1725585759,-0.1127176136,-0.0135882031,-0.2122376859,0.1584800482,-0.0319676921,0.2331239581,-0.1512842625,0.2855141759,-0.034938287,0.188546896,-0.0476546511,0.1102088764,-0.0013451164,-0.5571446419,0.6485419273,0.1964526772,0.2534165084,0.6113681793,-0.0001130312,0.1879943311,0.427631259,0.0710939392,-0.4378682077,-0.4516066909,0.1970393956,-0.0665755048,0.401270926,0.088631399,-0.1916379929,-0.1743660718,-0.0482023507,-0.356459558,0.1199604869,0.1515654474,0.2370973378,0.5524005294,-0.0181829743,-0.2322163582,0.2086297274,-0.0932899341,-0.1618075073,0.4859958291,0.277926445,0.0000181067,0.0415287912,0.1254346818,-0.1783881336,0.1609940231,0.4485349655,-0.0253838282,-0.1275161803,-0.0778705627,0.1246726066,0.2353181243,0.5451827049,-0.2534771264,-0.3022951782,-0.173381567,0.0765558109,0.1363622248,0.2944325209,-0.2126211375,-0.0716271028,0.1803832352,-0.1028136313,-0.1887427866,-0.0607598387,0.2107258141,0.1004542783,-0.158973828,-0.2308696061,0.2097773254,0.1637531519,-0.05227549,0.0160596594,-0.0153824734,0.0535579473,0.2114631385,-0.396317631,0.0593051873,0.0366261862,0.1347655654,0.2433893085,0.3089216948,0.3861995041,-0.0397093259,-0.2499678433,0.1359069943,0.4187658131,0.0782540217,-0.0268121343,-0.3456539214,0.439147234,0.2833787799,0.1257042885,-0.0463790931,-0.0763063729,-0.0516994074,-0.1337563246,-0.1697103679,0.2196262628,-0.2770833075,-0.2052009851,0.1551857591,-0.1102024615,-0.0619326904,0.1705599576,0.3384833932,-0.1790276617,0.2826485932,0.2799422443,0.2793718576,0.0097070495,-0.1391048133,-0.2306847572,0.3470246792,-0.1135188118,0.1684673876,0.4636680186,-0.3708824217,0.2146412134,0.0527159609,0.2620734274,0.0118495738,-0.064015612,-0.1308800727,-0.2860842049,0.262001574,0.0375623852,0.3730460107,0.3032679558,-0.3780884445,-0.0785978958,0.1303363591,-0.1046970189,-0.2736743093,-0.1131843925,0.140702635,-0.5425629616,-0.1690017283,-0.7952097058,-0.1158957928,-0.0374030024,-0.1901670247,-0.1757319272,-0.2521235943,0.1131115183,-0.1844933033,0.2569434643,0.5731605291,-0.4683733881,-0.1417627186,-0.3210857809,-0.2450504303,-0.1895079464,0.0499389432,-0.2290991247,0.0145849064,-0.3759831786,0.2342487425,0.201632306,-0.5678400397,-0.3043695986,0.1881448478,-0.2737540305,0.1391416043,-0.1096155122,-0.0431161709,0.0342822187,-0.1298496127,-0.2751350701,0.3710016608,0.0552587844,-0.1252245307,-0.152625978,-0.0876119956,0.0242702495,0.481423825,-0.0050684409,-0.0346196033,0.1112635955,0.4048300385,0.1238478646,0.0082403999,-0.002646581,0.2363563478,0.1557277441,0.0660671145,0.0265987385,-0.2205318511,-0.5138401985,0.279047966,0.2804960012,0.0934022292,-0.0520986617,0.1098806933,-0.3560436666,-0.0623568818,-0.4687760472,-0.257620424,0.042077519,0.2640281022,-0.1119287014,-0.0139558166,-0.2272306979,0.1317779124,-0.1492676586,0.2208481133,-0.2036609352,0.351572752,-0.1274198592,-0.1394954026,0.1126090959,0.1909775883,0.0127466023,-0.2036845237,-0.1120589375,0.5353549719,-0.19174546,-0.1702150106,0.1466422528,-0.0516479574,0.1855684668,-0.3324547708,0.019998746,0.0225046352,0.0527466536,0.0093191722,-0.1755173802,-0.0428609997,-0.0302115995,0.422205627,0.1818193346,0.1183508858,0.3142014444,0.0689993054,-0.0676742047,-0.1829338819,0.3034820855,0.062292099,0.4900861979,-0.0120340073,-0.197725758,0.0360332131,-0.0401958451,-0.007314716,0.1719882339,0.2215848565,-0.2188159376,-0.0418814793,-0.076336585,0.1008972302,0.5481566787,0.09765248,-0.0673388541,-0.1769091785,0.0582956001,-0.1067657098,0.1759354621,-0.1082828864,-0.0781492665,0.0818073526,0.1608298421,-0.0328737162,-0.2620770931,-0.4119637311,-0.0328710265,0.1852684617,-0.2394344509,0.1062231287,-0.3376254737,-0.1877980381,-0.0378028154,0.0568638258,-0.4906938374,-0.3384881616,-0.2438513488,0.2719503939,0.3034607172,0.0520159602,-0.0791255087,0.2371322662,0.0541921705,0.0087981354,-0.1003618538,0.0227770153,-0.2968294621,0.0321643203,0.1251295209,-0.1656052917,0.1294487268,0.0444601886,0.1059454307,-0.3842816651,-0.1393769234,0.2893945277,-0.0943957046,0.4804418087,0.2228505611,0.3556439877,0.0432693847,0.0097817592,0.5540043116,-0.0970824137,-0.0247857142,0.1292891353,0.061432723,0.0392577611,-0.1232213378,-0.0799865648,-0.4014402926,-0.3176139891,-0.003122418,-0.0119200535,-0.1667927057,0.4903976619,-0.0261619668,0.2140633762,-0.1069947779,0.1902738214,-0.2419166118,-0.4892966151,0.200207904,-0.2261955738,-0.2358217239,-0.0773115382,0.2203531563,0.3298978508,-0.1658444703,-0.5135371089,-0.1725959629,-0.0367091224,0.1087123677,-0.0219955295,0.1146963835,0.0110178953,-0.0514286198,-0.0525038578,-0.1449035555,-0.1751174331,-0.1460386962,-0.1188498512,0.2608271539,0.0687822402,0.0731951147,0.1224943027,0.6985940337,-0.1189664379,-0.2699800432,0.5415757895,0.0427996106,0.492747426,-0.0116330767,-0.5201570988,0.0955592021,-0.0993093848,0.3007572293,-0.0471367761,0.0544559136,0.1366566867,-0.2283730507,0.1179624647,-0.146733582,-0.1641769111,-0.0845774561,0.1379181743,0.0436215997,0.1699550301,0.0198067762,0.0150240473,-0.2037185133,0.1452025175,0.5837768912,0.1453821063,0.106602788,-0.1139515936,-0.1601517648,-0.3059076369,0.2898315787,-0.1667885035,0.3880302012,-0.0833021328,0.0869269669,0.0775844678,0.1479599178,0.4434062839,-0.3287041783,0.1585523784,0.1336524487,-0.0857756585,-0.6703960299,-0.0936662108,0.0994088575,0.3416185379,-0.0096508805,0.0618187711,-0.2640918493,-0.3057759702,0.4000406563,0.3831837773,-0.0875029117,-0.3635737896,-0.2049144059,-0.2763392925,-0.2349275798,-0.3325669765,-0.1548427939,0.2314252108,0.0645101592,0.0727364048,0.0704078525,-0.2061458379,0.2263251245,0.0480815135,0.3262000978,0.2228341401,0.1129967868,0.2624368072,0.0505169928,0.0808393359,0.276473403,-0.0935819969,-0.0407480337,0.4285108447,0.0415433384,0.1869734526,0.2134996206,-0.2970018387,0.2622372508,0.071264632,0.0613886863,-0.0912041962,0.1217717007,0.2187208086,-0.180023849,0.0382180624,-0.6317686439,0.5229841471,0.2532395124,0.0820076466,0.1012394503,0.051872123,-0.0263960492,0.1271338463,-0.0124298865,0.8495185971,-0.2115578651,0.2392898947,0.1863348186,0.3600391448,0.5700742006,-0.1596258432,0.0261815768,-0.3569960892,-0.2408921272,-0.0472788401,-0.3012674749,0.1071692109,0.096842505,-0.0497713611,0.2951699495,-0.2144957036,0.1786037087,0.0760802254,0.2879984677,-0.2931635082,-0.3046196401,-0.259069562,0.1664928496,-0.3988608122,0.4500918686,-0.1243478507,0.1604433507,-0.0989370868,-0.0784366652,-0.4062737226,0.1623120904,-0.022279989,0.2918296456,-0.1571384668,-0.4060109556,0.2369616628,0.1566004455,0.0825853124,0.213132292,-0.1353421658,0.2415904254,-0.2805398703,-0.144226715,0.000985063,-0.0398392975,0.1898031235,-0.0533527546,-0.1139114201,0.0885142982,-0.0280687287,-0.2107796222,0.1150694117,0.1050563306,0.0150268096,-0.2030857801,-0.2619379461,-0.1495789438,0.1153659001,-0.2147351652,0.0846839398,0.3800062239,-0.0407465436,-0.0872093216,0.0637581274,-0.1531886905,0.0290357377,0.5284175873,-0.0379123129,-0.2446852773,0.4941783249,0.4384137392,-0.1883351803,-0.1068596467,0.1098428965,0.2774757445,-0.5570066571,0.0944477618,-0.0762709975,0.1007295176,-0.0816820934,0.2877082527,0.215319261,-0.0438448079,-0.0691148266,-0.6689061522,-0.1225342155,0.2263715118,-0.2136110663,-0.0091676135,-0.1261732131,0.0823611245,0.0398913808,-0.3140805364,-0.2577999532,0.1932243705,-0.338650912,-0.157736823,0.3997710943,0.0153687298,0.2695322037,-0.1794127971,0.1243730634,0.0134058436,-0.1539351344,-0.1391739547,-0.1338413507,0.1500630677,0.0249882024,-0.082199432,-0.0661346987,-0.1386979818,0.0426825322,-0.0968800709,0.1971869469,0.1137086004,0.1165089384,-0.2642098367,-0.2311448008,0.0192840435,-0.0896336287,0.16840294,-0.000908838,0.3030080199,0.2946841121,-0.1176850796,0.077614814,-0.0906430706,0.017833421,0.1215566471,-0.0687292293,0.1031277776,0.1715617329,-0.3984612226,-0.0187457334,0.4810518622,0.3851553798,0.4579800963,-0.1927492172,0.178535074,-0.0602040216,0.1391738057,-0.1916448474,-0.1017361432,0.1052829549,0.0532130077,-0.0369514525,0.2517877519,-0.0764853805,0.0979356468,0.134403646,0.1943236887,0.6998489499,0.0628598258,0.1014312208,0.0436776951,-0.0285174232,-0.0237177927,0.2893110216,0.0461046733,0.1912942082,0.2021325827,-0.3849471807,0.1216018498,-0.1005506888,0.0210117456,0.0795859769,-0.459258765,0.2204454839,0.2550876737,0.3501327336,-0.1038687378,-0.1861909181,0.6454380751,-0.3548586071,0.0363705568,-0.0615884364,-0.0515921526,-0.1458561271,0.1176288873,-0.2180452645,-0.2793150544,-0.2996639311,-0.050261043,-0.1636514962,-0.2826920152,0.142309621,0.078260906,-0.0187922586,-0.5962900519,-0.208734408,0.3028778136,-0.2005786002,-0.025480682,0.2560763955,0.4521099925,-0.0031772675,0.184930712,0.3444438577,0.4056625366,0.1872998327,-0.169851914,0.0210561454,-0.2616790235,-0.0209682938,0.0674625486,-0.0407870971,0.0780815706,-0.0311201643,0.3041681647,0.168119669,-0.1418996006,0.1346779168,0.072129637,0.2286484092,-0.4309054613,0.2187297195,-0.3335321546,0.1112832204,-0.2731332183,0.0653153881,-0.5620749593,0.1158099473,0.4431761801,0.2090492696,-0.2461424768,-0.3371422589,0.0658825859,0.174587965,0.292445749,0.5197380185,0.0768842623,-0.3513478041,-0.2063227892,-0.9277210832,0.1879741251,0.0629231483,0.0292701796,0.0561481901,0.2212741524,0.1257389784,0.2855866253,0.5260552764,-0.1164274365,0.2458110154,-0.1491671205,-0.1590224355,-0.3091545105,-0.0292106438,0.0668421909,-0.1883630604,-0.0389772765,-0.0656963289,-0.091453366,0.0146777825,-0.3132792413,-0.2859941125,0.0990199298,-0.2107786536,0.4215120673,-0.2497771084,0.3580860794,0.1815736741,-0.2679289281,-0.2454044968,-0.418530494,0.0515818708,0.3490275443,-0.1713770926,0.3201974928,-0.1454135627,-0.1534007788,-0.5415129066,0.0478423424,0.0129160965,0.0887644961,-0.38911587,0.0631579086,-0.1504519284,0.1437314004,0.0170258991,0.0090682134,-0.1232908592,0.3748603761,-0.2689987123,-0.4195800126,0.674582839,-0.1699885428,-0.0913293883,-0.0927453861,0.1541092843,-0.4061711431,-0.1239946857,-0.3438657522,-0.0074313032,0.2635872662,0.2550987601,-0.3452776968,-0.1030685753,-0.2710345685,0.0684489012,-0.0651410446,0.2863773108,0.0968328938,-0.1559881419,0.0386836641,0.0821691826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1643","title":"Dataset social_bias_frames 404","comments":"I see, master is already fixed in https:\/\/github.com\/huggingface\/datasets\/commit\/9e058f098a0919efd03a136b9b9c3dec5076f626","body":"```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"social_bias_frames\")\r\n...\r\nDownloading and preparing dataset social_bias_frames\/default\r\n...\r\n~\/.pyenv\/versions\/3.7.6\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 484 )\r\n 485 elif response is not None and response.status_code == 404:\r\n--> 486 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 487 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 488 \r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/homes.cs.washington.edu\/~msap\/social-bias-frames\/SocialBiasFrames_v2.tgz\r\n```\r\n[Here](https:\/\/homes.cs.washington.edu\/~msap\/social-bias-frames\/) we find button `Download data` with the correct URL for the data: https:\/\/homes.cs.washington.edu\/~msap\/social-bias-frames\/SBIC.v2.tgz","comment_length":8,"text":"Dataset social_bias_frames 404\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"social_bias_frames\")\r\n...\r\nDownloading and preparing dataset social_bias_frames\/default\r\n...\r\n~\/.pyenv\/versions\/3.7.6\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 484 )\r\n 485 elif response is not None and response.status_code == 404:\r\n--> 486 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n 487 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 488 \r\n\r\nFileNotFoundError: Couldn't find file at https:\/\/homes.cs.washington.edu\/~msap\/social-bias-frames\/SocialBiasFrames_v2.tgz\r\n```\r\n[Here](https:\/\/homes.cs.washington.edu\/~msap\/social-bias-frames\/) we find button `Download data` with the correct URL for the data: https:\/\/homes.cs.washington.edu\/~msap\/social-bias-frames\/SBIC.v2.tgz\nI see, master is already fixed in https:\/\/github.com\/huggingface\/datasets\/commit\/9e058f098a0919efd03a136b9b9c3dec5076f626","embeddings":[-0.2098474801,0.1111924574,-0.053344503,0.2385647893,0.2090139836,0.2027729303,0.1050811186,0.3743502498,0.0982944071,0.0731079876,-0.0326477699,-0.0548801012,0.0335738584,0.0678444132,0.0022525573,0.0796444714,-0.0243726298,0.2905724049,-0.1208122745,-0.0473792553,-0.2406603247,0.2053297907,0.0453498103,0.2186264545,-0.2140847147,0.1705987006,-0.2395168245,0.2992097735,-0.2762687206,-0.356826216,0.2826701105,0.0345984921,-0.0641023219,0.9442012906,-0.0001120225,0.0753628537,0.4498872161,0.0282068215,-0.3442862928,-0.3886438012,-0.0447048023,-0.2145807743,0.0109283272,-0.1543274522,0.0279417168,-0.2381560951,0.1522223651,-0.0094564483,0.0127149876,0.2685523331,0.1904910952,0.2019750774,0.0501174554,-0.2279178798,0.279917866,-0.0372411385,0.0557786301,0.4019738436,-0.265060544,-0.0484823883,0.0578802936,0.2305274755,-0.0566751994,0.0048077069,0.10462749,-0.0195915177,0.3661519885,-0.3660225868,0.0167344753,0.2486035079,0.5347079039,-0.1409321576,-0.2660863996,-0.1473281085,-0.0535562634,-0.0095342221,0.1917833537,0.137161687,-0.1302834898,0.1674599051,-0.0781814009,-0.2985760868,0.0276964549,0.2571815848,-0.1344664395,-0.0669357777,-0.0570943281,0.0094113313,0.1080967635,0.1610836536,0.1646664739,-0.0206699166,0.0899631456,-0.0866397545,-0.3476577103,-0.0269126799,0.2916951776,0.0422945283,-0.1356101632,0.1983241886,-0.0659756511,-0.0475719795,0.0619228259,0.0797864422,0.1923474371,0.2181216031,-0.352396667,0.0627278611,0.3862148821,0.5282068253,-0.0134463226,0.1261308044,-0.1003616005,-0.4384925961,-0.1573803127,0.295748651,0.4241423011,-0.050136473,-0.2582071722,0.0162240844,-0.3396864533,0.0090155303,-0.0547303148,0.377489686,-0.16898413,0.4107787907,0.0144685302,-0.0604572147,-0.1940338314,-0.0494644418,-0.2822501957,0.0456456803,-0.2959097326,-0.0313950926,0.2800116539,-0.2669191658,0.3399345577,-0.0624204203,-0.00097144,-0.0411756709,0.0457775518,-0.1431257129,-0.0152161513,0.3294964135,0.0758070797,0.402235359,0.1823309064,0.0166756585,-0.0114719812,-0.0594818741,-0.360802561,-0.5093393326,0.1325879246,0.2506570816,-0.0136338864,0.0596523173,0.0138469134,0.1379792392,0.0724257082,-0.0839443207,0.0299600586,-0.1338659227,-0.2054453939,-0.2116020918,0.2489109486,0.6369197965,-0.5422288179,0.0209275912,0.0193885043,-0.1649934649,-0.0641714111,0.0966819003,-0.0876015723,0.2664028704,-0.5535143018,-0.0137661286,0.0920415223,-0.3822830021,-0.7280783057,0.0400738157,-0.3880797923,-0.070903942,0.2703590989,0.0125725986,0.2009661198,-0.3005084693,0.0112655303,0.197763741,0.1628630608,-0.1564534307,-0.1623266041,-0.3487687111,0.0012981336,0.0842259526,0.1741548926,-0.0590998344,0.1029692665,0.1173920408,0.1571041495,-0.0511767641,0.1202722713,0.0434849449,0.1744966507,0.2552499175,0.1108184159,-0.245760873,-0.1902864873,0.0971273929,-0.1238118261,0.20662947,-0.0759854987,-0.2639983296,-0.5531086922,-0.2392774224,-0.2680138052,-0.0838796124,0.1968520433,0.3375245631,-0.0846537277,0.1434592009,-0.1244340315,0.1702635586,-0.1858063191,0.2499278486,0.0951921418,0.3299043775,-0.339045465,-0.0091309715,0.1977005452,0.1693873107,0.3545876145,-0.1239779517,-0.0344968215,0.4405347407,-0.1668229699,0.3275703192,0.1681506783,0.0251322724,0.279561758,-0.375259012,-0.0334265046,0.5464473367,0.1408960819,0.1158302501,0.1147421822,0.0381866023,0.0255398098,0.1369411498,0.1179926097,-0.0161888264,0.0865341127,-0.0874529332,0.0624739602,-0.2558326125,0.1102042198,0.2350814342,0.1033314317,0.029883774,0.0794714838,0.2092346698,0.5072583556,-0.2483462393,0.0452072062,0.0952371359,-0.2486372143,-0.1584298909,-0.0317662917,0.2283471078,0.280216068,0.1279201508,0.2134591043,0.1634104997,0.0358686112,-0.416556567,0.3729899228,0.2356551439,0.0156986136,0.4263197482,0.1669783443,0.0003017726,-0.2919352353,-0.1101355031,0.0678243712,0.1711080968,-0.4833144546,-0.2249368876,-0.4628627896,-0.4728516936,0.0195844918,-0.0154681867,0.1251662374,-0.3728902638,0.0245442241,0.1369907409,-0.2168293595,0.1992839575,-0.2581530213,0.3217778504,0.1486123502,-0.060177967,-0.1912156641,-0.0799507573,-0.2047985643,0.0735364184,0.0437261499,-0.2282186896,0.0005648846,-0.4256437719,-0.0435288437,-0.3909265697,-0.3479516506,0.2178191394,-0.0415736735,0.3232354522,0.2756815851,0.2247872502,0.0706380755,0.040776737,0.2471167147,-0.2329016328,-0.1083757728,0.0048705065,-0.0607215837,0.2648850977,-0.1456330866,-0.4023025036,-0.3844248652,-0.5067508221,-0.2048302442,0.1040663049,0.2067419887,-0.138846606,0.092314966,0.1261505783,-0.0711597949,0.2243455201,-0.3653639853,-0.4046927691,0.4314951897,-0.327855885,-0.3112736046,0.030419942,-0.0609949045,-0.0330893919,-0.2245313078,-0.6929161549,-0.5804825425,0.0027419552,0.0602081865,0.1250902116,0.2182761729,0.2117684633,0.0665425807,-0.2961374819,-0.06036634,-0.1922163367,-0.0109697357,0.2412168086,0.0730148032,0.2032002062,0.2869644165,0.3205823004,1.0216187239,0.1667382121,0.1277287304,0.4178949594,-0.0099670719,0.5008906126,-0.0620308258,-0.3529949486,0.186988771,-0.1541184187,0.2188275903,0.1832528114,0.0776065737,-0.0023542121,-0.2552806735,-0.1439381093,-0.4999407828,-0.1860260963,-0.2159650028,-0.0800111517,-0.0754182413,0.2099187374,0.0935301259,-0.055735752,-0.0791440308,0.4894644916,0.5222302675,-0.0056233727,0.2281974554,0.0420654789,-0.0536742397,-0.331299752,0.1460684836,-0.1344989836,0.15336968,0.0767949447,-0.0066826013,0.0665214807,-0.1268408448,0.7887833118,-0.0818759128,0.1751987487,-0.024247596,-0.0297758915,-0.241424948,-0.117555663,-0.0005868002,0.0390302427,0.1496471614,0.159306854,-0.4819425642,-0.1992821693,0.1578953564,-0.0611289516,-0.1904979348,-0.043536365,-0.2497501671,-0.1658351272,-0.2533925772,-0.0159786027,-0.1207225621,0.2289427817,-0.052921854,-0.0219551027,-0.0219693147,-0.1139530391,-0.0549201332,-0.0739728212,0.2412612885,0.1037407368,-0.1478657871,0.2495818436,0.271956116,0.1041885242,0.5311152935,-0.1478559673,-0.4457691014,0.3920868337,-0.0329484753,0.1293473989,0.4671718776,-0.129233703,-0.2533002496,0.2704401314,0.1427875608,-0.3337450325,0.2694020867,0.3513500094,-0.152217865,0.0920545608,-0.5682854056,0.1011168882,0.0428420641,-0.1151355878,0.0557985194,0.2945498228,-0.1974063069,0.2343465686,-0.1150298044,0.9693986177,0.2121682912,0.2511304915,0.3040933609,-0.0756709948,0.6251871586,-0.2648700476,0.1886938363,-0.0212146062,-0.192134887,-0.164372161,-0.0367619693,-0.1330448836,-0.0670188591,-0.0710915476,0.629566431,0.0382665396,0.6191273928,0.2376244366,0.2181519419,0.0842577964,0.0371331163,-0.1240339205,0.1235710979,0.0869940072,0.3769938052,-0.1547830403,-0.1451921761,-0.2227137536,-0.002257762,-0.1026351005,0.0241144747,-0.5471087694,0.0100001376,-0.4304994643,-0.3737229109,-0.148955822,-0.0299298204,0.1399924904,0.072395429,-0.1944545954,0.1610311866,0.2001040876,0.1434689313,0.2175752223,-0.136646688,0.3786768615,-0.0965508521,-0.2449299991,0.2489450425,0.0803773776,-0.202172786,0.148855716,-0.2205262631,0.2397594452,-0.1724004298,-0.169494465,-0.2050329149,0.1000964642,-0.3536850512,0.1334026605,-0.1545274407,-0.0183829349,-0.3068176806,-0.094250828,-0.2561788261,-0.1719969213,0.4904083014,-0.212503314,-0.0849834979,0.5271114111,0.0801063403,-0.1972787082,-0.1352923959,-0.1579718292,-0.0943057537,-0.6694433093,0.0909164697,-0.2347221524,-0.1317743659,0.0747305453,0.1490826011,0.3879945874,0.1240790337,-0.0968662277,-0.5632417798,-0.076647155,0.0742406473,-0.1292859167,0.1073165014,0.0680827722,-0.1170225069,0.2376324385,-0.185667485,-0.3654306233,0.059109509,-0.2046435177,-0.0410509892,-0.0032711215,0.0408883505,0.0148241883,0.0399135202,0.1238470525,-0.0419313237,-0.3148235083,-0.1778647006,-0.0791726783,0.1504039615,0.0136460671,-0.0786526799,0.0053795641,-0.0451803245,0.0070417738,-0.0137825292,0.0805253461,-0.1401961893,0.0881935135,0.0309154596,0.0143246241,-0.2370708585,-0.2071877867,0.1762292236,-0.0045640743,0.2455942929,0.0635654926,0.2978509367,-0.1763697416,-0.0650826842,0.1044569016,0.1533678323,0.1145586371,0.1920506954,0.434371382,-0.2884493172,-0.113235265,0.3283295035,0.4680194259,0.1478162408,-0.2039265633,-0.2075850517,0.3011679351,0.1965369731,-0.2452543974,0.1817252934,0.481446147,0.0820581242,0.143911019,-0.06402383,0.3400917947,-0.1360027343,-0.0654033571,0.300129056,0.50192523,0.0328584164,0.2212764621,0.280140698,-0.0897448808,0.3480209112,0.4836561084,0.1234064773,0.1221599728,0.5380609035,0.0913976803,0.0177102052,0.0468856879,0.1876242757,-0.0240686648,-0.1022261158,0.3132666051,0.2068858296,-0.1923198253,-0.1565623432,-0.1085583791,0.3954344392,-0.1905355603,-0.0402568951,-0.1978207082,0.1671878397,-0.1273419708,-0.1217039227,-0.3736745417,-0.2436671555,0.1499159634,0.0995829329,0.0763271078,-0.2748232782,-0.0615305267,0.038208887,-0.2058518678,-0.2610404491,-0.2454831898,-0.0306615215,0.245479852,-0.1283017546,0.2768103778,0.346391201,-0.087487556,0.0224509407,0.6105306149,0.3521999419,0.110335581,-0.4123037755,-0.1189723387,0.3080906868,-0.0995051861,-0.0870715678,0.1559164077,0.1816922128,-0.1471176893,0.1930500269,0.2055704743,-0.3172944188,-0.0486883335,0.1823025346,0.1613593996,-0.4219179749,0.2056095004,-0.0976724699,-0.0608371235,0.0360572226,0.013177962,-0.5479457974,-0.0275617354,0.4733474553,-0.2056146115,0.1207630858,-0.3906956017,0.108569026,0.0653605536,0.3230491877,0.3463680148,0.0527830496,-0.3446643949,0.0624774508,-0.7606028318,0.375875324,0.2229195535,-0.0561665855,-0.0379028022,0.240829736,-0.0579963215,0.0942816809,0.2952294946,0.2034649849,0.2838564813,0.0945355967,-0.213350296,0.003470774,-0.3917798996,0.0766047686,-0.0628876612,-0.2554247379,0.0678384379,-0.5057089925,0.0567470901,0.1362043619,-0.2449789643,0.2034899443,-0.2312484831,0.5664471388,0.0223718286,-0.0055097621,-0.1833533347,-0.1146762744,-0.2839719951,0.0409587137,-0.3648301363,-0.0556930192,-0.1445595175,0.2403933406,-0.2561407089,0.0355748571,-0.3186377585,0.2782375216,0.2585745156,-0.2200102359,-0.2147115022,0.0988973007,-0.1935963333,-0.163733229,0.2075300962,0.0311086122,-0.1277403533,0.0409926921,-0.1782700866,-0.4862091541,0.3834860921,-0.5889070034,0.01994933,-0.262265116,0.4592326581,0.0149317095,0.3003078401,0.0191766806,0.1807073206,0.256326139,0.3521980345,-0.3601977527,0.363006115,-0.3978638947,0.094347991,-0.0146915503,0.041662544,0.2383701801,-0.3019502759,-0.063198328,-0.1551961601]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1641","title":"muchocine dataset cannot be dowloaded","comments":"I have encountered the same error with `v1.0.1` and `v1.0.2` on both Windows and Linux environments. However, cloning the repo and using the path to the dataset's root directory worked for me. Even after having the dataset cached - passing the path is the only way (for now) to load the dataset.\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"squad\") # Works\r\ndataset = load_dataset(\"code_search_net\", \"python\") # Error\r\ndataset = load_dataset(\"covid_qa_deepset\") # Error\r\n\r\npath = \"\/huggingface\/datasets\/datasets\/{}\/\"\r\ndataset = load_dataset(path.format(\"code_search_net\"), \"python\") # Works\r\ndataset = load_dataset(path.format(\"covid_qa_deepset\")) # Works\r\n```\r\n\r\n","body":"```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```","comment_length":88,"text":"muchocine dataset cannot be dowloaded\n```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```\nI have encountered the same error with `v1.0.1` and `v1.0.2` on both Windows and Linux environments. However, cloning the repo and using the path to the dataset's root directory worked for me. Even after having the dataset cached - passing the path is the only way (for now) to load the dataset.\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"squad\") # Works\r\ndataset = load_dataset(\"code_search_net\", \"python\") # Error\r\ndataset = load_dataset(\"covid_qa_deepset\") # Error\r\n\r\npath = \"\/huggingface\/datasets\/datasets\/{}\/\"\r\ndataset = load_dataset(path.format(\"code_search_net\"), \"python\") # Works\r\ndataset = load_dataset(path.format(\"covid_qa_deepset\")) # Works\r\n```\r\n\r\n","embeddings":[-0.3657995164,-0.1491084993,-0.0535814576,0.3329214454,0.4308770597,0.1230888665,0.359962672,0.3173546791,0.3176317215,0.0625982285,-0.2380269468,0.0012010086,-0.088910602,0.0610442795,0.0386018343,0.0196577273,-0.09693712,0.0420849808,-0.0913112238,0.0183463041,-0.2103825361,0.0953873545,-0.0354201607,-0.1008452177,-0.1107043624,-0.1373711079,-0.0456292927,0.3429242671,-0.4638158977,-0.2475145757,0.3933928609,-0.0597475134,0.234281376,0.4329293668,-0.0001103552,0.1410666555,0.3332580626,-0.1318264008,-0.4365814328,-0.5427050591,-0.2340316921,-0.2482806295,0.1217606962,-0.1392974555,-0.0860487893,-0.2626752853,0.0757315606,-0.1615948975,0.3259954154,0.4596894383,0.2739565372,0.380763948,0.1729074568,-0.2033828199,0.4229698777,-0.0247138757,-0.1236245483,0.4684586823,0.0693040863,-0.024407642,0.0798502043,0.0708083808,-0.188991338,0.1367580295,0.1010280177,-0.1283799112,-0.0986570567,-0.3918550313,0.1986587644,0.2060496807,0.2593174577,-0.1449202597,-0.4517106712,-0.0636464879,0.0942918658,-0.4788383842,0.0918764025,0.2396257669,0.1592647433,0.1096555591,0.0373005494,-0.249765873,-0.1381266117,0.3669583201,-0.1173515543,0.0664340779,-0.2460633069,-0.1064698994,0.0680627376,-0.1647121012,-0.1510896981,0.1155014038,-0.2062194943,0.102867052,-0.2802644968,-0.056664519,-0.0772117227,-0.0013284244,0.0971117243,0.2565974891,0.0405130051,-0.0053874771,0.0977683812,0.0714682937,0.0552441515,0.1962952316,-0.0038632948,0.1495791376,0.0931143016,0.3699893057,-0.0203259401,-0.0598866008,-0.4647661746,-0.250232935,-0.2243047506,-0.2275567502,0.3014701009,-0.2856255472,-0.3109283149,-0.0650905594,-0.0692585632,-0.0173054188,0.2589639723,0.591776371,-0.2199613154,0.1064362228,0.0758352801,0.3028856218,-0.0017310816,0.2429312021,-0.2484750748,0.4639939368,-0.1429483891,-0.0564340316,0.251064837,-0.4133730233,0.4145965874,-0.1181406304,0.0109985182,-0.1353993267,-0.0377918668,-0.1631853282,0.1514472365,0.224788323,-0.0626443401,0.0758576468,0.2453983724,0.0033513503,-0.1945626736,-0.2132005095,-0.1482447833,-0.4352903068,-0.1609933972,0.2583007216,-0.0089885276,-0.1736529022,-0.1358133107,-0.366612941,0.1069539264,-0.0593503825,-0.0898369849,-0.2057666928,-0.1339452416,-0.2123783231,0.4348729551,0.6249145865,0.0577635914,-0.1927994341,-0.1481889635,0.1223766208,0.1011742428,0.1879488081,-0.256937474,0.1325680017,-0.3040416241,-0.2483960986,0.5892069936,-0.478657186,-0.3480915129,0.4911933541,-0.2445544153,0.0593723133,0.1439642012,-0.1001187265,0.2385188937,0.0485155918,0.5149019957,0.1525217742,0.0060279611,-0.0757718012,-0.2067523301,-0.3548433185,-0.1694717556,0.2668253481,0.2473154366,0.1203408763,0.2451995462,0.3110216558,0.0581675619,-0.0235748794,-0.0413322337,0.251663208,0.2244619727,0.0511546023,-0.2082608342,-0.060526751,-0.373261869,0.2185180634,-0.0019837443,0.0943156704,-0.1907487065,0.1080496237,-0.5253341198,-0.041457817,-0.2475919873,-0.1638884544,0.163851127,0.2125468105,0.256457001,0.1356508881,-0.2132635266,0.3509989381,-0.1266469955,0.0677978545,-0.2135928869,0.0486418754,-0.169808507,-0.1537032425,0.1298379898,0.1843456626,0.2591026723,-0.2886403799,0.0465948954,0.2431535423,-0.0469079241,-0.2477146685,0.040527422,0.2363331914,0.2393021435,-0.0464564115,0.1544012427,-0.0636134222,0.1630853117,0.0505272746,0.1560126543,0.3476131558,-0.1160445437,0.2653279901,0.2960418165,-0.1287916303,0.3048322797,0.0371037014,0.1048189625,-0.2303738743,0.2754022777,0.1103917286,0.4532395899,0.0878859088,-0.244653523,-0.0362577438,0.1542772949,0.1417571306,-0.100424476,0.088071011,-0.1078079715,0.0503145568,-0.1643220484,0.1259839535,0.4394570589,0.3152256012,-0.003013884,-0.1307429522,0.1457129568,-0.1901173443,0.22852467,0.0828294829,0.1319265068,0.0903955325,0.2431523353,0.0604094751,-0.2865849137,-0.4136480987,0.0604158528,0.2175486982,-0.1233721375,0.0543118194,-0.3415805101,-0.4350299239,-0.1832307875,-0.0812561661,-0.1189615279,-0.2224486619,-0.1093107909,0.3850286901,0.0820262358,0.1743492037,-0.3616304398,0.1311659068,0.0034895537,-0.463675499,-0.0208469816,-0.0498597361,0.0238630641,0.0916467384,0.3101963997,-0.3293982744,0.2989204228,-0.2584398687,-0.3787842393,-0.3437911272,-0.0524499752,0.1639781594,0.0416437387,0.2919322252,0.0584816337,0.2683333457,-0.0478225909,-0.1814966053,0.2785207331,-0.0333806127,-0.101678744,-0.0568195209,-0.0152252857,0.004719927,0.1111131161,-0.3484595418,-0.3846930861,-0.5702346563,0.001228385,0.0241096672,0.0976283625,0.1633743942,0.0823181942,0.1395433843,-0.2771349549,0.2637175024,0.0028176897,-0.6356604695,0.3077563643,-0.3354674876,-0.4740182459,0.0636546239,0.0719056129,0.1483119577,0.0598436333,-0.3255211115,-0.1184874848,-0.2935323417,0.0969748423,0.0113175409,0.1488799453,0.1564039141,-0.0627485886,-0.1523783356,-0.0394820422,-0.0744168758,0.0284218621,0.0755534172,0.1952203214,-0.1556702703,0.3054393232,-0.1872801781,0.8105006814,-0.0446907543,-0.2567236423,0.6232312918,-0.3263625205,0.543845892,-0.1952953637,-0.5162585378,-0.088937521,0.0012898825,0.2168815732,0.047435265,-0.1229545251,-0.0691595674,-0.388232708,-0.1752273887,-0.3440136611,-0.1467586905,0.0254923478,-0.0559352078,-0.0069198064,-0.0019944157,-0.1349003911,0.0762924775,-0.0448449962,0.1067680791,0.4677458704,0.1058969349,0.1338784844,0.2699663043,-0.125907734,-0.315382272,0.229288891,0.0990082175,0.1840803921,-0.0263515394,-0.1844694316,-0.0765056908,-0.1125784293,0.5402199626,0.0232715495,0.1116978228,0.3033772409,-0.3210595846,-0.4074937105,-0.0094456505,-0.1867676228,0.0728378072,0.4294902384,0.1813544333,-0.3219102323,-0.1675598919,0.3808234334,0.276889354,-0.1648255736,-0.0369857587,-0.3196110725,-0.2903357744,-0.3800643384,-0.0481718555,0.1076278836,0.4150267541,0.0340872742,0.1011691615,-0.2938578725,-0.0260187201,-0.0172521118,0.0750777125,0.3295866251,-0.0051776441,0.2987500429,0.0001269509,0.4790292084,0.5336917639,0.608314693,-0.0790778473,-0.168248862,0.1742368191,0.2388187945,0.056849625,0.0911385417,-0.1643012762,-0.0760752335,0.1003498286,0.1237166375,0.0757639632,0.4928692579,0.1997979432,-0.101721786,-0.3060792685,-0.2848143578,0.3799953163,0.0059596449,0.1839969456,0.3617464006,0.1735025197,-0.1793793142,0.1578198373,-0.2865616083,0.8279553056,-0.088301301,0.1863437593,0.2034919858,-0.1707055271,0.4270977676,0.0044306559,0.2608374953,-0.3066089451,-0.1917940527,0.1068552136,-0.2196325362,-0.0452402234,0.037217807,-0.0017503271,0.2464989573,-0.171620518,-0.2535544038,-0.1752812415,0.4142414927,-0.0600915626,-0.1273560226,-0.2972863317,0.1233005896,0.0636470839,0.47649014,-0.2641223967,-0.022815749,-0.1227031499,-0.2786701322,-0.2843520641,0.2614377439,-0.0923176631,0.319498539,-0.3142413497,-0.199461177,-0.0117253065,-0.1410720944,0.0119980527,-0.0121824266,-0.1136445478,0.09910357,0.025367748,-0.4313670993,-0.012650975,0.10936331,0.4606412053,-0.1748532355,-0.0569606051,-0.0439526103,-0.1209330708,-0.0764138401,0.0666377321,0.0606505424,-0.1704290211,-0.2852598131,-0.3929140866,0.0057140384,-0.0192151498,-0.2860834301,0.1495039016,0.0541306585,-0.1489330828,-0.141005531,0.0631333962,-0.369474411,-0.261447221,0.5848149657,-0.2290470451,-0.1672657281,0.4755726159,0.1389335841,-0.2453746349,-0.2389470786,-0.057756912,0.2142287195,-0.3769004345,0.1933730543,0.1458589584,0.210410431,0.0994283855,0.4053385556,0.2238170207,-0.0655532181,0.163526088,-0.5226733685,-0.2048822045,0.2629172504,-0.0754835159,0.1675189137,0.0834503621,0.0323742144,0.1139416397,-0.0411145985,-0.3581664562,0.0279709678,-0.2944269776,0.0520219095,0.1817686409,-0.0211625379,0.357322067,0.2104436606,0.2412478775,-0.095286347,-0.1011711806,-0.2811157405,-0.1054296345,0.1371492594,-0.0351853035,-0.0646686256,-0.1630118936,-0.1591403484,-0.1330506504,-0.0642687753,0.2809072137,0.1647329032,0.0117628146,-0.284776479,0.0981535316,-0.0421203226,-0.1719227284,0.2252429426,0.1075998917,0.0509256944,0.2489593774,0.2051205486,0.0316969715,-0.1099842787,0.1387232691,-0.2905350626,0.0681333989,-0.0240929034,0.2176180631,-0.2760128379,-0.2741425633,0.1538185179,0.274525255,0.4074634016,0.0081729321,0.0470387936,0.2161379755,0.2668988705,-0.3773310483,0.0189239737,0.0285646226,-0.1256906539,0.2468962669,0.255073756,-0.106856972,0.0481030643,0.1360794455,0.2231007814,0.4210959971,-0.0778731555,0.2160880417,-0.1783435792,0.0489661582,-0.1044896618,0.3343425989,0.2829759419,0.4328795373,0.2902895808,-0.1596602052,0.3407626152,-0.1692937315,0.1661964655,-0.0429286435,-0.4089371264,0.3431918025,0.155533269,0.1181677505,-0.0545903109,-0.2855466604,-0.0084881354,-0.0694933012,0.2013454139,-0.1192132086,0.2931375206,-0.1820884347,-0.1090704128,0.0934147164,-0.4112871587,0.079889521,0.0461072847,0.1834768653,-0.1655807644,0.1462471783,0.2682876885,-0.1675834507,-0.0777960047,-0.0180541705,0.3332146406,0.1199546829,-0.2109153569,0.2081679106,0.2205871046,-0.1167477816,-0.1216178611,0.355735302,0.5909681916,0.3230317831,0.1966940165,0.0063568894,0.1522777081,-0.1297103465,-0.1076303869,0.1393267363,-0.0713555291,0.2352316678,0.175805375,0.2756284177,-0.2915883958,0.0513826981,-0.0899239033,-0.0341186747,-0.0582948178,0.2172850668,-0.0469961315,-0.1263225377,-0.3412036002,0.2155513465,-0.5843129158,0.4456870556,0.2231096029,0.010193604,0.1192841381,0.0343007967,0.1173476353,0.1104774997,0.4590658844,0.3333301544,0.2629753649,-0.3781104088,-0.0899777859,-0.5046923757,0.2131059915,-0.1343518198,0.0054264665,0.0841841772,0.236955747,-0.0553477108,0.0025013217,0.1508084685,0.2183922082,0.0902341604,-0.0175418463,-0.2637939453,-0.0161259435,0.0706616193,0.0177553222,0.1297650337,-0.3305588663,0.0024301747,-0.4772558808,0.095216684,-0.1454280168,-0.2248672396,-0.1889298856,-0.342071265,0.3520760536,-0.1208494604,0.159047097,-0.1975364834,-0.1927533597,-0.2993255258,-0.4522846639,-0.1979387552,0.3398811817,0.3392877579,0.2523281574,-0.2662300169,-0.1787528545,-0.4373828769,0.2831882834,-0.0156293456,-0.2114800364,0.1319566667,-0.0878060684,0.0294063874,-0.0275121499,0.0575888157,0.1131119058,0.0304558743,0.0242230408,-0.3524520993,-0.321449101,0.5321993232,-0.1855504811,-0.0950867981,-0.0717512891,0.2673282027,0.0405889079,-0.0020066877,-0.5231081247,0.5159048438,0.4153162539,0.0681953207,-0.1566002816,0.1899167001,-0.2691119313,0.0465724282,-0.0812959895,0.1766751111,-0.006289144,-0.2810450196,-0.1985638887,-0.0266595725]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1641","title":"muchocine dataset cannot be dowloaded","comments":"Hi @mrm8488 and @amoux!\r\n The datasets you are trying to load have been added to the library during the community sprint for v2 last month. They will be available with the v2 release!\r\nFor now, there are still a couple of solutions to load the datasets:\r\n1. As suggested by @amoux, you can clone the git repo and pass the local path to the script\r\n2. You can also install the latest (master) version of `datasets` using pip: `pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","body":"```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```","comment_length":81,"text":"muchocine dataset cannot be dowloaded\n```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```\nHi @mrm8488 and @amoux!\r\n The datasets you are trying to load have been added to the library during the community sprint for v2 last month. They will be available with the v2 release!\r\nFor now, there are still a couple of solutions to load the datasets:\r\n1. As suggested by @amoux, you can clone the git repo and pass the local path to the script\r\n2. You can also install the latest (master) version of `datasets` using pip: `pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","embeddings":[-0.3657995164,-0.1491084993,-0.0535814576,0.3329214454,0.4308770597,0.1230888665,0.359962672,0.3173546791,0.3176317215,0.0625982285,-0.2380269468,0.0012010086,-0.088910602,0.0610442795,0.0386018343,0.0196577273,-0.09693712,0.0420849808,-0.0913112238,0.0183463041,-0.2103825361,0.0953873545,-0.0354201607,-0.1008452177,-0.1107043624,-0.1373711079,-0.0456292927,0.3429242671,-0.4638158977,-0.2475145757,0.3933928609,-0.0597475134,0.234281376,0.4329293668,-0.0001103552,0.1410666555,0.3332580626,-0.1318264008,-0.4365814328,-0.5427050591,-0.2340316921,-0.2482806295,0.1217606962,-0.1392974555,-0.0860487893,-0.2626752853,0.0757315606,-0.1615948975,0.3259954154,0.4596894383,0.2739565372,0.380763948,0.1729074568,-0.2033828199,0.4229698777,-0.0247138757,-0.1236245483,0.4684586823,0.0693040863,-0.024407642,0.0798502043,0.0708083808,-0.188991338,0.1367580295,0.1010280177,-0.1283799112,-0.0986570567,-0.3918550313,0.1986587644,0.2060496807,0.2593174577,-0.1449202597,-0.4517106712,-0.0636464879,0.0942918658,-0.4788383842,0.0918764025,0.2396257669,0.1592647433,0.1096555591,0.0373005494,-0.249765873,-0.1381266117,0.3669583201,-0.1173515543,0.0664340779,-0.2460633069,-0.1064698994,0.0680627376,-0.1647121012,-0.1510896981,0.1155014038,-0.2062194943,0.102867052,-0.2802644968,-0.056664519,-0.0772117227,-0.0013284244,0.0971117243,0.2565974891,0.0405130051,-0.0053874771,0.0977683812,0.0714682937,0.0552441515,0.1962952316,-0.0038632948,0.1495791376,0.0931143016,0.3699893057,-0.0203259401,-0.0598866008,-0.4647661746,-0.250232935,-0.2243047506,-0.2275567502,0.3014701009,-0.2856255472,-0.3109283149,-0.0650905594,-0.0692585632,-0.0173054188,0.2589639723,0.591776371,-0.2199613154,0.1064362228,0.0758352801,0.3028856218,-0.0017310816,0.2429312021,-0.2484750748,0.4639939368,-0.1429483891,-0.0564340316,0.251064837,-0.4133730233,0.4145965874,-0.1181406304,0.0109985182,-0.1353993267,-0.0377918668,-0.1631853282,0.1514472365,0.224788323,-0.0626443401,0.0758576468,0.2453983724,0.0033513503,-0.1945626736,-0.2132005095,-0.1482447833,-0.4352903068,-0.1609933972,0.2583007216,-0.0089885276,-0.1736529022,-0.1358133107,-0.366612941,0.1069539264,-0.0593503825,-0.0898369849,-0.2057666928,-0.1339452416,-0.2123783231,0.4348729551,0.6249145865,0.0577635914,-0.1927994341,-0.1481889635,0.1223766208,0.1011742428,0.1879488081,-0.256937474,0.1325680017,-0.3040416241,-0.2483960986,0.5892069936,-0.478657186,-0.3480915129,0.4911933541,-0.2445544153,0.0593723133,0.1439642012,-0.1001187265,0.2385188937,0.0485155918,0.5149019957,0.1525217742,0.0060279611,-0.0757718012,-0.2067523301,-0.3548433185,-0.1694717556,0.2668253481,0.2473154366,0.1203408763,0.2451995462,0.3110216558,0.0581675619,-0.0235748794,-0.0413322337,0.251663208,0.2244619727,0.0511546023,-0.2082608342,-0.060526751,-0.373261869,0.2185180634,-0.0019837443,0.0943156704,-0.1907487065,0.1080496237,-0.5253341198,-0.041457817,-0.2475919873,-0.1638884544,0.163851127,0.2125468105,0.256457001,0.1356508881,-0.2132635266,0.3509989381,-0.1266469955,0.0677978545,-0.2135928869,0.0486418754,-0.169808507,-0.1537032425,0.1298379898,0.1843456626,0.2591026723,-0.2886403799,0.0465948954,0.2431535423,-0.0469079241,-0.2477146685,0.040527422,0.2363331914,0.2393021435,-0.0464564115,0.1544012427,-0.0636134222,0.1630853117,0.0505272746,0.1560126543,0.3476131558,-0.1160445437,0.2653279901,0.2960418165,-0.1287916303,0.3048322797,0.0371037014,0.1048189625,-0.2303738743,0.2754022777,0.1103917286,0.4532395899,0.0878859088,-0.244653523,-0.0362577438,0.1542772949,0.1417571306,-0.100424476,0.088071011,-0.1078079715,0.0503145568,-0.1643220484,0.1259839535,0.4394570589,0.3152256012,-0.003013884,-0.1307429522,0.1457129568,-0.1901173443,0.22852467,0.0828294829,0.1319265068,0.0903955325,0.2431523353,0.0604094751,-0.2865849137,-0.4136480987,0.0604158528,0.2175486982,-0.1233721375,0.0543118194,-0.3415805101,-0.4350299239,-0.1832307875,-0.0812561661,-0.1189615279,-0.2224486619,-0.1093107909,0.3850286901,0.0820262358,0.1743492037,-0.3616304398,0.1311659068,0.0034895537,-0.463675499,-0.0208469816,-0.0498597361,0.0238630641,0.0916467384,0.3101963997,-0.3293982744,0.2989204228,-0.2584398687,-0.3787842393,-0.3437911272,-0.0524499752,0.1639781594,0.0416437387,0.2919322252,0.0584816337,0.2683333457,-0.0478225909,-0.1814966053,0.2785207331,-0.0333806127,-0.101678744,-0.0568195209,-0.0152252857,0.004719927,0.1111131161,-0.3484595418,-0.3846930861,-0.5702346563,0.001228385,0.0241096672,0.0976283625,0.1633743942,0.0823181942,0.1395433843,-0.2771349549,0.2637175024,0.0028176897,-0.6356604695,0.3077563643,-0.3354674876,-0.4740182459,0.0636546239,0.0719056129,0.1483119577,0.0598436333,-0.3255211115,-0.1184874848,-0.2935323417,0.0969748423,0.0113175409,0.1488799453,0.1564039141,-0.0627485886,-0.1523783356,-0.0394820422,-0.0744168758,0.0284218621,0.0755534172,0.1952203214,-0.1556702703,0.3054393232,-0.1872801781,0.8105006814,-0.0446907543,-0.2567236423,0.6232312918,-0.3263625205,0.543845892,-0.1952953637,-0.5162585378,-0.088937521,0.0012898825,0.2168815732,0.047435265,-0.1229545251,-0.0691595674,-0.388232708,-0.1752273887,-0.3440136611,-0.1467586905,0.0254923478,-0.0559352078,-0.0069198064,-0.0019944157,-0.1349003911,0.0762924775,-0.0448449962,0.1067680791,0.4677458704,0.1058969349,0.1338784844,0.2699663043,-0.125907734,-0.315382272,0.229288891,0.0990082175,0.1840803921,-0.0263515394,-0.1844694316,-0.0765056908,-0.1125784293,0.5402199626,0.0232715495,0.1116978228,0.3033772409,-0.3210595846,-0.4074937105,-0.0094456505,-0.1867676228,0.0728378072,0.4294902384,0.1813544333,-0.3219102323,-0.1675598919,0.3808234334,0.276889354,-0.1648255736,-0.0369857587,-0.3196110725,-0.2903357744,-0.3800643384,-0.0481718555,0.1076278836,0.4150267541,0.0340872742,0.1011691615,-0.2938578725,-0.0260187201,-0.0172521118,0.0750777125,0.3295866251,-0.0051776441,0.2987500429,0.0001269509,0.4790292084,0.5336917639,0.608314693,-0.0790778473,-0.168248862,0.1742368191,0.2388187945,0.056849625,0.0911385417,-0.1643012762,-0.0760752335,0.1003498286,0.1237166375,0.0757639632,0.4928692579,0.1997979432,-0.101721786,-0.3060792685,-0.2848143578,0.3799953163,0.0059596449,0.1839969456,0.3617464006,0.1735025197,-0.1793793142,0.1578198373,-0.2865616083,0.8279553056,-0.088301301,0.1863437593,0.2034919858,-0.1707055271,0.4270977676,0.0044306559,0.2608374953,-0.3066089451,-0.1917940527,0.1068552136,-0.2196325362,-0.0452402234,0.037217807,-0.0017503271,0.2464989573,-0.171620518,-0.2535544038,-0.1752812415,0.4142414927,-0.0600915626,-0.1273560226,-0.2972863317,0.1233005896,0.0636470839,0.47649014,-0.2641223967,-0.022815749,-0.1227031499,-0.2786701322,-0.2843520641,0.2614377439,-0.0923176631,0.319498539,-0.3142413497,-0.199461177,-0.0117253065,-0.1410720944,0.0119980527,-0.0121824266,-0.1136445478,0.09910357,0.025367748,-0.4313670993,-0.012650975,0.10936331,0.4606412053,-0.1748532355,-0.0569606051,-0.0439526103,-0.1209330708,-0.0764138401,0.0666377321,0.0606505424,-0.1704290211,-0.2852598131,-0.3929140866,0.0057140384,-0.0192151498,-0.2860834301,0.1495039016,0.0541306585,-0.1489330828,-0.141005531,0.0631333962,-0.369474411,-0.261447221,0.5848149657,-0.2290470451,-0.1672657281,0.4755726159,0.1389335841,-0.2453746349,-0.2389470786,-0.057756912,0.2142287195,-0.3769004345,0.1933730543,0.1458589584,0.210410431,0.0994283855,0.4053385556,0.2238170207,-0.0655532181,0.163526088,-0.5226733685,-0.2048822045,0.2629172504,-0.0754835159,0.1675189137,0.0834503621,0.0323742144,0.1139416397,-0.0411145985,-0.3581664562,0.0279709678,-0.2944269776,0.0520219095,0.1817686409,-0.0211625379,0.357322067,0.2104436606,0.2412478775,-0.095286347,-0.1011711806,-0.2811157405,-0.1054296345,0.1371492594,-0.0351853035,-0.0646686256,-0.1630118936,-0.1591403484,-0.1330506504,-0.0642687753,0.2809072137,0.1647329032,0.0117628146,-0.284776479,0.0981535316,-0.0421203226,-0.1719227284,0.2252429426,0.1075998917,0.0509256944,0.2489593774,0.2051205486,0.0316969715,-0.1099842787,0.1387232691,-0.2905350626,0.0681333989,-0.0240929034,0.2176180631,-0.2760128379,-0.2741425633,0.1538185179,0.274525255,0.4074634016,0.0081729321,0.0470387936,0.2161379755,0.2668988705,-0.3773310483,0.0189239737,0.0285646226,-0.1256906539,0.2468962669,0.255073756,-0.106856972,0.0481030643,0.1360794455,0.2231007814,0.4210959971,-0.0778731555,0.2160880417,-0.1783435792,0.0489661582,-0.1044896618,0.3343425989,0.2829759419,0.4328795373,0.2902895808,-0.1596602052,0.3407626152,-0.1692937315,0.1661964655,-0.0429286435,-0.4089371264,0.3431918025,0.155533269,0.1181677505,-0.0545903109,-0.2855466604,-0.0084881354,-0.0694933012,0.2013454139,-0.1192132086,0.2931375206,-0.1820884347,-0.1090704128,0.0934147164,-0.4112871587,0.079889521,0.0461072847,0.1834768653,-0.1655807644,0.1462471783,0.2682876885,-0.1675834507,-0.0777960047,-0.0180541705,0.3332146406,0.1199546829,-0.2109153569,0.2081679106,0.2205871046,-0.1167477816,-0.1216178611,0.355735302,0.5909681916,0.3230317831,0.1966940165,0.0063568894,0.1522777081,-0.1297103465,-0.1076303869,0.1393267363,-0.0713555291,0.2352316678,0.175805375,0.2756284177,-0.2915883958,0.0513826981,-0.0899239033,-0.0341186747,-0.0582948178,0.2172850668,-0.0469961315,-0.1263225377,-0.3412036002,0.2155513465,-0.5843129158,0.4456870556,0.2231096029,0.010193604,0.1192841381,0.0343007967,0.1173476353,0.1104774997,0.4590658844,0.3333301544,0.2629753649,-0.3781104088,-0.0899777859,-0.5046923757,0.2131059915,-0.1343518198,0.0054264665,0.0841841772,0.236955747,-0.0553477108,0.0025013217,0.1508084685,0.2183922082,0.0902341604,-0.0175418463,-0.2637939453,-0.0161259435,0.0706616193,0.0177553222,0.1297650337,-0.3305588663,0.0024301747,-0.4772558808,0.095216684,-0.1454280168,-0.2248672396,-0.1889298856,-0.342071265,0.3520760536,-0.1208494604,0.159047097,-0.1975364834,-0.1927533597,-0.2993255258,-0.4522846639,-0.1979387552,0.3398811817,0.3392877579,0.2523281574,-0.2662300169,-0.1787528545,-0.4373828769,0.2831882834,-0.0156293456,-0.2114800364,0.1319566667,-0.0878060684,0.0294063874,-0.0275121499,0.0575888157,0.1131119058,0.0304558743,0.0242230408,-0.3524520993,-0.321449101,0.5321993232,-0.1855504811,-0.0950867981,-0.0717512891,0.2673282027,0.0405889079,-0.0020066877,-0.5231081247,0.5159048438,0.4153162539,0.0681953207,-0.1566002816,0.1899167001,-0.2691119313,0.0465724282,-0.0812959895,0.1766751111,-0.006289144,-0.2810450196,-0.1985638887,-0.0266595725]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1641","title":"muchocine dataset cannot be dowloaded","comments":"If you don't want to clone entire `datasets` repo, just download the `muchocine` directory and pass the local path to the directory. Cheers!","body":"```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```","comment_length":23,"text":"muchocine dataset cannot be dowloaded\n```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```\nIf you don't want to clone entire `datasets` repo, just download the `muchocine` directory and pass the local path to the directory. Cheers!","embeddings":[-0.3657995164,-0.1491084993,-0.0535814576,0.3329214454,0.4308770597,0.1230888665,0.359962672,0.3173546791,0.3176317215,0.0625982285,-0.2380269468,0.0012010086,-0.088910602,0.0610442795,0.0386018343,0.0196577273,-0.09693712,0.0420849808,-0.0913112238,0.0183463041,-0.2103825361,0.0953873545,-0.0354201607,-0.1008452177,-0.1107043624,-0.1373711079,-0.0456292927,0.3429242671,-0.4638158977,-0.2475145757,0.3933928609,-0.0597475134,0.234281376,0.4329293668,-0.0001103552,0.1410666555,0.3332580626,-0.1318264008,-0.4365814328,-0.5427050591,-0.2340316921,-0.2482806295,0.1217606962,-0.1392974555,-0.0860487893,-0.2626752853,0.0757315606,-0.1615948975,0.3259954154,0.4596894383,0.2739565372,0.380763948,0.1729074568,-0.2033828199,0.4229698777,-0.0247138757,-0.1236245483,0.4684586823,0.0693040863,-0.024407642,0.0798502043,0.0708083808,-0.188991338,0.1367580295,0.1010280177,-0.1283799112,-0.0986570567,-0.3918550313,0.1986587644,0.2060496807,0.2593174577,-0.1449202597,-0.4517106712,-0.0636464879,0.0942918658,-0.4788383842,0.0918764025,0.2396257669,0.1592647433,0.1096555591,0.0373005494,-0.249765873,-0.1381266117,0.3669583201,-0.1173515543,0.0664340779,-0.2460633069,-0.1064698994,0.0680627376,-0.1647121012,-0.1510896981,0.1155014038,-0.2062194943,0.102867052,-0.2802644968,-0.056664519,-0.0772117227,-0.0013284244,0.0971117243,0.2565974891,0.0405130051,-0.0053874771,0.0977683812,0.0714682937,0.0552441515,0.1962952316,-0.0038632948,0.1495791376,0.0931143016,0.3699893057,-0.0203259401,-0.0598866008,-0.4647661746,-0.250232935,-0.2243047506,-0.2275567502,0.3014701009,-0.2856255472,-0.3109283149,-0.0650905594,-0.0692585632,-0.0173054188,0.2589639723,0.591776371,-0.2199613154,0.1064362228,0.0758352801,0.3028856218,-0.0017310816,0.2429312021,-0.2484750748,0.4639939368,-0.1429483891,-0.0564340316,0.251064837,-0.4133730233,0.4145965874,-0.1181406304,0.0109985182,-0.1353993267,-0.0377918668,-0.1631853282,0.1514472365,0.224788323,-0.0626443401,0.0758576468,0.2453983724,0.0033513503,-0.1945626736,-0.2132005095,-0.1482447833,-0.4352903068,-0.1609933972,0.2583007216,-0.0089885276,-0.1736529022,-0.1358133107,-0.366612941,0.1069539264,-0.0593503825,-0.0898369849,-0.2057666928,-0.1339452416,-0.2123783231,0.4348729551,0.6249145865,0.0577635914,-0.1927994341,-0.1481889635,0.1223766208,0.1011742428,0.1879488081,-0.256937474,0.1325680017,-0.3040416241,-0.2483960986,0.5892069936,-0.478657186,-0.3480915129,0.4911933541,-0.2445544153,0.0593723133,0.1439642012,-0.1001187265,0.2385188937,0.0485155918,0.5149019957,0.1525217742,0.0060279611,-0.0757718012,-0.2067523301,-0.3548433185,-0.1694717556,0.2668253481,0.2473154366,0.1203408763,0.2451995462,0.3110216558,0.0581675619,-0.0235748794,-0.0413322337,0.251663208,0.2244619727,0.0511546023,-0.2082608342,-0.060526751,-0.373261869,0.2185180634,-0.0019837443,0.0943156704,-0.1907487065,0.1080496237,-0.5253341198,-0.041457817,-0.2475919873,-0.1638884544,0.163851127,0.2125468105,0.256457001,0.1356508881,-0.2132635266,0.3509989381,-0.1266469955,0.0677978545,-0.2135928869,0.0486418754,-0.169808507,-0.1537032425,0.1298379898,0.1843456626,0.2591026723,-0.2886403799,0.0465948954,0.2431535423,-0.0469079241,-0.2477146685,0.040527422,0.2363331914,0.2393021435,-0.0464564115,0.1544012427,-0.0636134222,0.1630853117,0.0505272746,0.1560126543,0.3476131558,-0.1160445437,0.2653279901,0.2960418165,-0.1287916303,0.3048322797,0.0371037014,0.1048189625,-0.2303738743,0.2754022777,0.1103917286,0.4532395899,0.0878859088,-0.244653523,-0.0362577438,0.1542772949,0.1417571306,-0.100424476,0.088071011,-0.1078079715,0.0503145568,-0.1643220484,0.1259839535,0.4394570589,0.3152256012,-0.003013884,-0.1307429522,0.1457129568,-0.1901173443,0.22852467,0.0828294829,0.1319265068,0.0903955325,0.2431523353,0.0604094751,-0.2865849137,-0.4136480987,0.0604158528,0.2175486982,-0.1233721375,0.0543118194,-0.3415805101,-0.4350299239,-0.1832307875,-0.0812561661,-0.1189615279,-0.2224486619,-0.1093107909,0.3850286901,0.0820262358,0.1743492037,-0.3616304398,0.1311659068,0.0034895537,-0.463675499,-0.0208469816,-0.0498597361,0.0238630641,0.0916467384,0.3101963997,-0.3293982744,0.2989204228,-0.2584398687,-0.3787842393,-0.3437911272,-0.0524499752,0.1639781594,0.0416437387,0.2919322252,0.0584816337,0.2683333457,-0.0478225909,-0.1814966053,0.2785207331,-0.0333806127,-0.101678744,-0.0568195209,-0.0152252857,0.004719927,0.1111131161,-0.3484595418,-0.3846930861,-0.5702346563,0.001228385,0.0241096672,0.0976283625,0.1633743942,0.0823181942,0.1395433843,-0.2771349549,0.2637175024,0.0028176897,-0.6356604695,0.3077563643,-0.3354674876,-0.4740182459,0.0636546239,0.0719056129,0.1483119577,0.0598436333,-0.3255211115,-0.1184874848,-0.2935323417,0.0969748423,0.0113175409,0.1488799453,0.1564039141,-0.0627485886,-0.1523783356,-0.0394820422,-0.0744168758,0.0284218621,0.0755534172,0.1952203214,-0.1556702703,0.3054393232,-0.1872801781,0.8105006814,-0.0446907543,-0.2567236423,0.6232312918,-0.3263625205,0.543845892,-0.1952953637,-0.5162585378,-0.088937521,0.0012898825,0.2168815732,0.047435265,-0.1229545251,-0.0691595674,-0.388232708,-0.1752273887,-0.3440136611,-0.1467586905,0.0254923478,-0.0559352078,-0.0069198064,-0.0019944157,-0.1349003911,0.0762924775,-0.0448449962,0.1067680791,0.4677458704,0.1058969349,0.1338784844,0.2699663043,-0.125907734,-0.315382272,0.229288891,0.0990082175,0.1840803921,-0.0263515394,-0.1844694316,-0.0765056908,-0.1125784293,0.5402199626,0.0232715495,0.1116978228,0.3033772409,-0.3210595846,-0.4074937105,-0.0094456505,-0.1867676228,0.0728378072,0.4294902384,0.1813544333,-0.3219102323,-0.1675598919,0.3808234334,0.276889354,-0.1648255736,-0.0369857587,-0.3196110725,-0.2903357744,-0.3800643384,-0.0481718555,0.1076278836,0.4150267541,0.0340872742,0.1011691615,-0.2938578725,-0.0260187201,-0.0172521118,0.0750777125,0.3295866251,-0.0051776441,0.2987500429,0.0001269509,0.4790292084,0.5336917639,0.608314693,-0.0790778473,-0.168248862,0.1742368191,0.2388187945,0.056849625,0.0911385417,-0.1643012762,-0.0760752335,0.1003498286,0.1237166375,0.0757639632,0.4928692579,0.1997979432,-0.101721786,-0.3060792685,-0.2848143578,0.3799953163,0.0059596449,0.1839969456,0.3617464006,0.1735025197,-0.1793793142,0.1578198373,-0.2865616083,0.8279553056,-0.088301301,0.1863437593,0.2034919858,-0.1707055271,0.4270977676,0.0044306559,0.2608374953,-0.3066089451,-0.1917940527,0.1068552136,-0.2196325362,-0.0452402234,0.037217807,-0.0017503271,0.2464989573,-0.171620518,-0.2535544038,-0.1752812415,0.4142414927,-0.0600915626,-0.1273560226,-0.2972863317,0.1233005896,0.0636470839,0.47649014,-0.2641223967,-0.022815749,-0.1227031499,-0.2786701322,-0.2843520641,0.2614377439,-0.0923176631,0.319498539,-0.3142413497,-0.199461177,-0.0117253065,-0.1410720944,0.0119980527,-0.0121824266,-0.1136445478,0.09910357,0.025367748,-0.4313670993,-0.012650975,0.10936331,0.4606412053,-0.1748532355,-0.0569606051,-0.0439526103,-0.1209330708,-0.0764138401,0.0666377321,0.0606505424,-0.1704290211,-0.2852598131,-0.3929140866,0.0057140384,-0.0192151498,-0.2860834301,0.1495039016,0.0541306585,-0.1489330828,-0.141005531,0.0631333962,-0.369474411,-0.261447221,0.5848149657,-0.2290470451,-0.1672657281,0.4755726159,0.1389335841,-0.2453746349,-0.2389470786,-0.057756912,0.2142287195,-0.3769004345,0.1933730543,0.1458589584,0.210410431,0.0994283855,0.4053385556,0.2238170207,-0.0655532181,0.163526088,-0.5226733685,-0.2048822045,0.2629172504,-0.0754835159,0.1675189137,0.0834503621,0.0323742144,0.1139416397,-0.0411145985,-0.3581664562,0.0279709678,-0.2944269776,0.0520219095,0.1817686409,-0.0211625379,0.357322067,0.2104436606,0.2412478775,-0.095286347,-0.1011711806,-0.2811157405,-0.1054296345,0.1371492594,-0.0351853035,-0.0646686256,-0.1630118936,-0.1591403484,-0.1330506504,-0.0642687753,0.2809072137,0.1647329032,0.0117628146,-0.284776479,0.0981535316,-0.0421203226,-0.1719227284,0.2252429426,0.1075998917,0.0509256944,0.2489593774,0.2051205486,0.0316969715,-0.1099842787,0.1387232691,-0.2905350626,0.0681333989,-0.0240929034,0.2176180631,-0.2760128379,-0.2741425633,0.1538185179,0.274525255,0.4074634016,0.0081729321,0.0470387936,0.2161379755,0.2668988705,-0.3773310483,0.0189239737,0.0285646226,-0.1256906539,0.2468962669,0.255073756,-0.106856972,0.0481030643,0.1360794455,0.2231007814,0.4210959971,-0.0778731555,0.2160880417,-0.1783435792,0.0489661582,-0.1044896618,0.3343425989,0.2829759419,0.4328795373,0.2902895808,-0.1596602052,0.3407626152,-0.1692937315,0.1661964655,-0.0429286435,-0.4089371264,0.3431918025,0.155533269,0.1181677505,-0.0545903109,-0.2855466604,-0.0084881354,-0.0694933012,0.2013454139,-0.1192132086,0.2931375206,-0.1820884347,-0.1090704128,0.0934147164,-0.4112871587,0.079889521,0.0461072847,0.1834768653,-0.1655807644,0.1462471783,0.2682876885,-0.1675834507,-0.0777960047,-0.0180541705,0.3332146406,0.1199546829,-0.2109153569,0.2081679106,0.2205871046,-0.1167477816,-0.1216178611,0.355735302,0.5909681916,0.3230317831,0.1966940165,0.0063568894,0.1522777081,-0.1297103465,-0.1076303869,0.1393267363,-0.0713555291,0.2352316678,0.175805375,0.2756284177,-0.2915883958,0.0513826981,-0.0899239033,-0.0341186747,-0.0582948178,0.2172850668,-0.0469961315,-0.1263225377,-0.3412036002,0.2155513465,-0.5843129158,0.4456870556,0.2231096029,0.010193604,0.1192841381,0.0343007967,0.1173476353,0.1104774997,0.4590658844,0.3333301544,0.2629753649,-0.3781104088,-0.0899777859,-0.5046923757,0.2131059915,-0.1343518198,0.0054264665,0.0841841772,0.236955747,-0.0553477108,0.0025013217,0.1508084685,0.2183922082,0.0902341604,-0.0175418463,-0.2637939453,-0.0161259435,0.0706616193,0.0177553222,0.1297650337,-0.3305588663,0.0024301747,-0.4772558808,0.095216684,-0.1454280168,-0.2248672396,-0.1889298856,-0.342071265,0.3520760536,-0.1208494604,0.159047097,-0.1975364834,-0.1927533597,-0.2993255258,-0.4522846639,-0.1979387552,0.3398811817,0.3392877579,0.2523281574,-0.2662300169,-0.1787528545,-0.4373828769,0.2831882834,-0.0156293456,-0.2114800364,0.1319566667,-0.0878060684,0.0294063874,-0.0275121499,0.0575888157,0.1131119058,0.0304558743,0.0242230408,-0.3524520993,-0.321449101,0.5321993232,-0.1855504811,-0.0950867981,-0.0717512891,0.2673282027,0.0405889079,-0.0020066877,-0.5231081247,0.5159048438,0.4153162539,0.0681953207,-0.1566002816,0.1899167001,-0.2691119313,0.0465724282,-0.0812959895,0.1766751111,-0.006289144,-0.2810450196,-0.1985638887,-0.0266595725]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1641","title":"muchocine dataset cannot be dowloaded","comments":"Muchocine was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `muchocine` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"muchocine\", split=\"train\")\r\n```","body":"```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```","comment_length":41,"text":"muchocine dataset cannot be dowloaded\n```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```\nMuchocine was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `muchocine` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"muchocine\", split=\"train\")\r\n```","embeddings":[-0.3657995164,-0.1491084993,-0.0535814576,0.3329214454,0.4308770597,0.1230888665,0.359962672,0.3173546791,0.3176317215,0.0625982285,-0.2380269468,0.0012010086,-0.088910602,0.0610442795,0.0386018343,0.0196577273,-0.09693712,0.0420849808,-0.0913112238,0.0183463041,-0.2103825361,0.0953873545,-0.0354201607,-0.1008452177,-0.1107043624,-0.1373711079,-0.0456292927,0.3429242671,-0.4638158977,-0.2475145757,0.3933928609,-0.0597475134,0.234281376,0.4329293668,-0.0001103552,0.1410666555,0.3332580626,-0.1318264008,-0.4365814328,-0.5427050591,-0.2340316921,-0.2482806295,0.1217606962,-0.1392974555,-0.0860487893,-0.2626752853,0.0757315606,-0.1615948975,0.3259954154,0.4596894383,0.2739565372,0.380763948,0.1729074568,-0.2033828199,0.4229698777,-0.0247138757,-0.1236245483,0.4684586823,0.0693040863,-0.024407642,0.0798502043,0.0708083808,-0.188991338,0.1367580295,0.1010280177,-0.1283799112,-0.0986570567,-0.3918550313,0.1986587644,0.2060496807,0.2593174577,-0.1449202597,-0.4517106712,-0.0636464879,0.0942918658,-0.4788383842,0.0918764025,0.2396257669,0.1592647433,0.1096555591,0.0373005494,-0.249765873,-0.1381266117,0.3669583201,-0.1173515543,0.0664340779,-0.2460633069,-0.1064698994,0.0680627376,-0.1647121012,-0.1510896981,0.1155014038,-0.2062194943,0.102867052,-0.2802644968,-0.056664519,-0.0772117227,-0.0013284244,0.0971117243,0.2565974891,0.0405130051,-0.0053874771,0.0977683812,0.0714682937,0.0552441515,0.1962952316,-0.0038632948,0.1495791376,0.0931143016,0.3699893057,-0.0203259401,-0.0598866008,-0.4647661746,-0.250232935,-0.2243047506,-0.2275567502,0.3014701009,-0.2856255472,-0.3109283149,-0.0650905594,-0.0692585632,-0.0173054188,0.2589639723,0.591776371,-0.2199613154,0.1064362228,0.0758352801,0.3028856218,-0.0017310816,0.2429312021,-0.2484750748,0.4639939368,-0.1429483891,-0.0564340316,0.251064837,-0.4133730233,0.4145965874,-0.1181406304,0.0109985182,-0.1353993267,-0.0377918668,-0.1631853282,0.1514472365,0.224788323,-0.0626443401,0.0758576468,0.2453983724,0.0033513503,-0.1945626736,-0.2132005095,-0.1482447833,-0.4352903068,-0.1609933972,0.2583007216,-0.0089885276,-0.1736529022,-0.1358133107,-0.366612941,0.1069539264,-0.0593503825,-0.0898369849,-0.2057666928,-0.1339452416,-0.2123783231,0.4348729551,0.6249145865,0.0577635914,-0.1927994341,-0.1481889635,0.1223766208,0.1011742428,0.1879488081,-0.256937474,0.1325680017,-0.3040416241,-0.2483960986,0.5892069936,-0.478657186,-0.3480915129,0.4911933541,-0.2445544153,0.0593723133,0.1439642012,-0.1001187265,0.2385188937,0.0485155918,0.5149019957,0.1525217742,0.0060279611,-0.0757718012,-0.2067523301,-0.3548433185,-0.1694717556,0.2668253481,0.2473154366,0.1203408763,0.2451995462,0.3110216558,0.0581675619,-0.0235748794,-0.0413322337,0.251663208,0.2244619727,0.0511546023,-0.2082608342,-0.060526751,-0.373261869,0.2185180634,-0.0019837443,0.0943156704,-0.1907487065,0.1080496237,-0.5253341198,-0.041457817,-0.2475919873,-0.1638884544,0.163851127,0.2125468105,0.256457001,0.1356508881,-0.2132635266,0.3509989381,-0.1266469955,0.0677978545,-0.2135928869,0.0486418754,-0.169808507,-0.1537032425,0.1298379898,0.1843456626,0.2591026723,-0.2886403799,0.0465948954,0.2431535423,-0.0469079241,-0.2477146685,0.040527422,0.2363331914,0.2393021435,-0.0464564115,0.1544012427,-0.0636134222,0.1630853117,0.0505272746,0.1560126543,0.3476131558,-0.1160445437,0.2653279901,0.2960418165,-0.1287916303,0.3048322797,0.0371037014,0.1048189625,-0.2303738743,0.2754022777,0.1103917286,0.4532395899,0.0878859088,-0.244653523,-0.0362577438,0.1542772949,0.1417571306,-0.100424476,0.088071011,-0.1078079715,0.0503145568,-0.1643220484,0.1259839535,0.4394570589,0.3152256012,-0.003013884,-0.1307429522,0.1457129568,-0.1901173443,0.22852467,0.0828294829,0.1319265068,0.0903955325,0.2431523353,0.0604094751,-0.2865849137,-0.4136480987,0.0604158528,0.2175486982,-0.1233721375,0.0543118194,-0.3415805101,-0.4350299239,-0.1832307875,-0.0812561661,-0.1189615279,-0.2224486619,-0.1093107909,0.3850286901,0.0820262358,0.1743492037,-0.3616304398,0.1311659068,0.0034895537,-0.463675499,-0.0208469816,-0.0498597361,0.0238630641,0.0916467384,0.3101963997,-0.3293982744,0.2989204228,-0.2584398687,-0.3787842393,-0.3437911272,-0.0524499752,0.1639781594,0.0416437387,0.2919322252,0.0584816337,0.2683333457,-0.0478225909,-0.1814966053,0.2785207331,-0.0333806127,-0.101678744,-0.0568195209,-0.0152252857,0.004719927,0.1111131161,-0.3484595418,-0.3846930861,-0.5702346563,0.001228385,0.0241096672,0.0976283625,0.1633743942,0.0823181942,0.1395433843,-0.2771349549,0.2637175024,0.0028176897,-0.6356604695,0.3077563643,-0.3354674876,-0.4740182459,0.0636546239,0.0719056129,0.1483119577,0.0598436333,-0.3255211115,-0.1184874848,-0.2935323417,0.0969748423,0.0113175409,0.1488799453,0.1564039141,-0.0627485886,-0.1523783356,-0.0394820422,-0.0744168758,0.0284218621,0.0755534172,0.1952203214,-0.1556702703,0.3054393232,-0.1872801781,0.8105006814,-0.0446907543,-0.2567236423,0.6232312918,-0.3263625205,0.543845892,-0.1952953637,-0.5162585378,-0.088937521,0.0012898825,0.2168815732,0.047435265,-0.1229545251,-0.0691595674,-0.388232708,-0.1752273887,-0.3440136611,-0.1467586905,0.0254923478,-0.0559352078,-0.0069198064,-0.0019944157,-0.1349003911,0.0762924775,-0.0448449962,0.1067680791,0.4677458704,0.1058969349,0.1338784844,0.2699663043,-0.125907734,-0.315382272,0.229288891,0.0990082175,0.1840803921,-0.0263515394,-0.1844694316,-0.0765056908,-0.1125784293,0.5402199626,0.0232715495,0.1116978228,0.3033772409,-0.3210595846,-0.4074937105,-0.0094456505,-0.1867676228,0.0728378072,0.4294902384,0.1813544333,-0.3219102323,-0.1675598919,0.3808234334,0.276889354,-0.1648255736,-0.0369857587,-0.3196110725,-0.2903357744,-0.3800643384,-0.0481718555,0.1076278836,0.4150267541,0.0340872742,0.1011691615,-0.2938578725,-0.0260187201,-0.0172521118,0.0750777125,0.3295866251,-0.0051776441,0.2987500429,0.0001269509,0.4790292084,0.5336917639,0.608314693,-0.0790778473,-0.168248862,0.1742368191,0.2388187945,0.056849625,0.0911385417,-0.1643012762,-0.0760752335,0.1003498286,0.1237166375,0.0757639632,0.4928692579,0.1997979432,-0.101721786,-0.3060792685,-0.2848143578,0.3799953163,0.0059596449,0.1839969456,0.3617464006,0.1735025197,-0.1793793142,0.1578198373,-0.2865616083,0.8279553056,-0.088301301,0.1863437593,0.2034919858,-0.1707055271,0.4270977676,0.0044306559,0.2608374953,-0.3066089451,-0.1917940527,0.1068552136,-0.2196325362,-0.0452402234,0.037217807,-0.0017503271,0.2464989573,-0.171620518,-0.2535544038,-0.1752812415,0.4142414927,-0.0600915626,-0.1273560226,-0.2972863317,0.1233005896,0.0636470839,0.47649014,-0.2641223967,-0.022815749,-0.1227031499,-0.2786701322,-0.2843520641,0.2614377439,-0.0923176631,0.319498539,-0.3142413497,-0.199461177,-0.0117253065,-0.1410720944,0.0119980527,-0.0121824266,-0.1136445478,0.09910357,0.025367748,-0.4313670993,-0.012650975,0.10936331,0.4606412053,-0.1748532355,-0.0569606051,-0.0439526103,-0.1209330708,-0.0764138401,0.0666377321,0.0606505424,-0.1704290211,-0.2852598131,-0.3929140866,0.0057140384,-0.0192151498,-0.2860834301,0.1495039016,0.0541306585,-0.1489330828,-0.141005531,0.0631333962,-0.369474411,-0.261447221,0.5848149657,-0.2290470451,-0.1672657281,0.4755726159,0.1389335841,-0.2453746349,-0.2389470786,-0.057756912,0.2142287195,-0.3769004345,0.1933730543,0.1458589584,0.210410431,0.0994283855,0.4053385556,0.2238170207,-0.0655532181,0.163526088,-0.5226733685,-0.2048822045,0.2629172504,-0.0754835159,0.1675189137,0.0834503621,0.0323742144,0.1139416397,-0.0411145985,-0.3581664562,0.0279709678,-0.2944269776,0.0520219095,0.1817686409,-0.0211625379,0.357322067,0.2104436606,0.2412478775,-0.095286347,-0.1011711806,-0.2811157405,-0.1054296345,0.1371492594,-0.0351853035,-0.0646686256,-0.1630118936,-0.1591403484,-0.1330506504,-0.0642687753,0.2809072137,0.1647329032,0.0117628146,-0.284776479,0.0981535316,-0.0421203226,-0.1719227284,0.2252429426,0.1075998917,0.0509256944,0.2489593774,0.2051205486,0.0316969715,-0.1099842787,0.1387232691,-0.2905350626,0.0681333989,-0.0240929034,0.2176180631,-0.2760128379,-0.2741425633,0.1538185179,0.274525255,0.4074634016,0.0081729321,0.0470387936,0.2161379755,0.2668988705,-0.3773310483,0.0189239737,0.0285646226,-0.1256906539,0.2468962669,0.255073756,-0.106856972,0.0481030643,0.1360794455,0.2231007814,0.4210959971,-0.0778731555,0.2160880417,-0.1783435792,0.0489661582,-0.1044896618,0.3343425989,0.2829759419,0.4328795373,0.2902895808,-0.1596602052,0.3407626152,-0.1692937315,0.1661964655,-0.0429286435,-0.4089371264,0.3431918025,0.155533269,0.1181677505,-0.0545903109,-0.2855466604,-0.0084881354,-0.0694933012,0.2013454139,-0.1192132086,0.2931375206,-0.1820884347,-0.1090704128,0.0934147164,-0.4112871587,0.079889521,0.0461072847,0.1834768653,-0.1655807644,0.1462471783,0.2682876885,-0.1675834507,-0.0777960047,-0.0180541705,0.3332146406,0.1199546829,-0.2109153569,0.2081679106,0.2205871046,-0.1167477816,-0.1216178611,0.355735302,0.5909681916,0.3230317831,0.1966940165,0.0063568894,0.1522777081,-0.1297103465,-0.1076303869,0.1393267363,-0.0713555291,0.2352316678,0.175805375,0.2756284177,-0.2915883958,0.0513826981,-0.0899239033,-0.0341186747,-0.0582948178,0.2172850668,-0.0469961315,-0.1263225377,-0.3412036002,0.2155513465,-0.5843129158,0.4456870556,0.2231096029,0.010193604,0.1192841381,0.0343007967,0.1173476353,0.1104774997,0.4590658844,0.3333301544,0.2629753649,-0.3781104088,-0.0899777859,-0.5046923757,0.2131059915,-0.1343518198,0.0054264665,0.0841841772,0.236955747,-0.0553477108,0.0025013217,0.1508084685,0.2183922082,0.0902341604,-0.0175418463,-0.2637939453,-0.0161259435,0.0706616193,0.0177553222,0.1297650337,-0.3305588663,0.0024301747,-0.4772558808,0.095216684,-0.1454280168,-0.2248672396,-0.1889298856,-0.342071265,0.3520760536,-0.1208494604,0.159047097,-0.1975364834,-0.1927533597,-0.2993255258,-0.4522846639,-0.1979387552,0.3398811817,0.3392877579,0.2523281574,-0.2662300169,-0.1787528545,-0.4373828769,0.2831882834,-0.0156293456,-0.2114800364,0.1319566667,-0.0878060684,0.0294063874,-0.0275121499,0.0575888157,0.1131119058,0.0304558743,0.0242230408,-0.3524520993,-0.321449101,0.5321993232,-0.1855504811,-0.0950867981,-0.0717512891,0.2673282027,0.0405889079,-0.0020066877,-0.5231081247,0.5159048438,0.4153162539,0.0681953207,-0.1566002816,0.1899167001,-0.2691119313,0.0465724282,-0.0812959895,0.1766751111,-0.006289144,-0.2810450196,-0.1985638887,-0.0266595725]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1641","title":"muchocine dataset cannot be dowloaded","comments":"Thanks @lhoestq ","body":"```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```","comment_length":2,"text":"muchocine dataset cannot be dowloaded\n```python\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n\r\n7 frames\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 281 raise FileNotFoundError(\r\n 282 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 283 combined_path, github_file_path, file_path\r\n 284 )\r\n 285 )\r\n\r\nFileNotFoundError: Couldn't find file locally at muchocine\/muchocine.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/muchocine\/muchocine.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/muchocine\/muchocine.py\r\n```\nThanks @lhoestq ","embeddings":[-0.3657995164,-0.1491084993,-0.0535814576,0.3329214454,0.4308770597,0.1230888665,0.359962672,0.3173546791,0.3176317215,0.0625982285,-0.2380269468,0.0012010086,-0.088910602,0.0610442795,0.0386018343,0.0196577273,-0.09693712,0.0420849808,-0.0913112238,0.0183463041,-0.2103825361,0.0953873545,-0.0354201607,-0.1008452177,-0.1107043624,-0.1373711079,-0.0456292927,0.3429242671,-0.4638158977,-0.2475145757,0.3933928609,-0.0597475134,0.234281376,0.4329293668,-0.0001103552,0.1410666555,0.3332580626,-0.1318264008,-0.4365814328,-0.5427050591,-0.2340316921,-0.2482806295,0.1217606962,-0.1392974555,-0.0860487893,-0.2626752853,0.0757315606,-0.1615948975,0.3259954154,0.4596894383,0.2739565372,0.380763948,0.1729074568,-0.2033828199,0.4229698777,-0.0247138757,-0.1236245483,0.4684586823,0.0693040863,-0.024407642,0.0798502043,0.0708083808,-0.188991338,0.1367580295,0.1010280177,-0.1283799112,-0.0986570567,-0.3918550313,0.1986587644,0.2060496807,0.2593174577,-0.1449202597,-0.4517106712,-0.0636464879,0.0942918658,-0.4788383842,0.0918764025,0.2396257669,0.1592647433,0.1096555591,0.0373005494,-0.249765873,-0.1381266117,0.3669583201,-0.1173515543,0.0664340779,-0.2460633069,-0.1064698994,0.0680627376,-0.1647121012,-0.1510896981,0.1155014038,-0.2062194943,0.102867052,-0.2802644968,-0.056664519,-0.0772117227,-0.0013284244,0.0971117243,0.2565974891,0.0405130051,-0.0053874771,0.0977683812,0.0714682937,0.0552441515,0.1962952316,-0.0038632948,0.1495791376,0.0931143016,0.3699893057,-0.0203259401,-0.0598866008,-0.4647661746,-0.250232935,-0.2243047506,-0.2275567502,0.3014701009,-0.2856255472,-0.3109283149,-0.0650905594,-0.0692585632,-0.0173054188,0.2589639723,0.591776371,-0.2199613154,0.1064362228,0.0758352801,0.3028856218,-0.0017310816,0.2429312021,-0.2484750748,0.4639939368,-0.1429483891,-0.0564340316,0.251064837,-0.4133730233,0.4145965874,-0.1181406304,0.0109985182,-0.1353993267,-0.0377918668,-0.1631853282,0.1514472365,0.224788323,-0.0626443401,0.0758576468,0.2453983724,0.0033513503,-0.1945626736,-0.2132005095,-0.1482447833,-0.4352903068,-0.1609933972,0.2583007216,-0.0089885276,-0.1736529022,-0.1358133107,-0.366612941,0.1069539264,-0.0593503825,-0.0898369849,-0.2057666928,-0.1339452416,-0.2123783231,0.4348729551,0.6249145865,0.0577635914,-0.1927994341,-0.1481889635,0.1223766208,0.1011742428,0.1879488081,-0.256937474,0.1325680017,-0.3040416241,-0.2483960986,0.5892069936,-0.478657186,-0.3480915129,0.4911933541,-0.2445544153,0.0593723133,0.1439642012,-0.1001187265,0.2385188937,0.0485155918,0.5149019957,0.1525217742,0.0060279611,-0.0757718012,-0.2067523301,-0.3548433185,-0.1694717556,0.2668253481,0.2473154366,0.1203408763,0.2451995462,0.3110216558,0.0581675619,-0.0235748794,-0.0413322337,0.251663208,0.2244619727,0.0511546023,-0.2082608342,-0.060526751,-0.373261869,0.2185180634,-0.0019837443,0.0943156704,-0.1907487065,0.1080496237,-0.5253341198,-0.041457817,-0.2475919873,-0.1638884544,0.163851127,0.2125468105,0.256457001,0.1356508881,-0.2132635266,0.3509989381,-0.1266469955,0.0677978545,-0.2135928869,0.0486418754,-0.169808507,-0.1537032425,0.1298379898,0.1843456626,0.2591026723,-0.2886403799,0.0465948954,0.2431535423,-0.0469079241,-0.2477146685,0.040527422,0.2363331914,0.2393021435,-0.0464564115,0.1544012427,-0.0636134222,0.1630853117,0.0505272746,0.1560126543,0.3476131558,-0.1160445437,0.2653279901,0.2960418165,-0.1287916303,0.3048322797,0.0371037014,0.1048189625,-0.2303738743,0.2754022777,0.1103917286,0.4532395899,0.0878859088,-0.244653523,-0.0362577438,0.1542772949,0.1417571306,-0.100424476,0.088071011,-0.1078079715,0.0503145568,-0.1643220484,0.1259839535,0.4394570589,0.3152256012,-0.003013884,-0.1307429522,0.1457129568,-0.1901173443,0.22852467,0.0828294829,0.1319265068,0.0903955325,0.2431523353,0.0604094751,-0.2865849137,-0.4136480987,0.0604158528,0.2175486982,-0.1233721375,0.0543118194,-0.3415805101,-0.4350299239,-0.1832307875,-0.0812561661,-0.1189615279,-0.2224486619,-0.1093107909,0.3850286901,0.0820262358,0.1743492037,-0.3616304398,0.1311659068,0.0034895537,-0.463675499,-0.0208469816,-0.0498597361,0.0238630641,0.0916467384,0.3101963997,-0.3293982744,0.2989204228,-0.2584398687,-0.3787842393,-0.3437911272,-0.0524499752,0.1639781594,0.0416437387,0.2919322252,0.0584816337,0.2683333457,-0.0478225909,-0.1814966053,0.2785207331,-0.0333806127,-0.101678744,-0.0568195209,-0.0152252857,0.004719927,0.1111131161,-0.3484595418,-0.3846930861,-0.5702346563,0.001228385,0.0241096672,0.0976283625,0.1633743942,0.0823181942,0.1395433843,-0.2771349549,0.2637175024,0.0028176897,-0.6356604695,0.3077563643,-0.3354674876,-0.4740182459,0.0636546239,0.0719056129,0.1483119577,0.0598436333,-0.3255211115,-0.1184874848,-0.2935323417,0.0969748423,0.0113175409,0.1488799453,0.1564039141,-0.0627485886,-0.1523783356,-0.0394820422,-0.0744168758,0.0284218621,0.0755534172,0.1952203214,-0.1556702703,0.3054393232,-0.1872801781,0.8105006814,-0.0446907543,-0.2567236423,0.6232312918,-0.3263625205,0.543845892,-0.1952953637,-0.5162585378,-0.088937521,0.0012898825,0.2168815732,0.047435265,-0.1229545251,-0.0691595674,-0.388232708,-0.1752273887,-0.3440136611,-0.1467586905,0.0254923478,-0.0559352078,-0.0069198064,-0.0019944157,-0.1349003911,0.0762924775,-0.0448449962,0.1067680791,0.4677458704,0.1058969349,0.1338784844,0.2699663043,-0.125907734,-0.315382272,0.229288891,0.0990082175,0.1840803921,-0.0263515394,-0.1844694316,-0.0765056908,-0.1125784293,0.5402199626,0.0232715495,0.1116978228,0.3033772409,-0.3210595846,-0.4074937105,-0.0094456505,-0.1867676228,0.0728378072,0.4294902384,0.1813544333,-0.3219102323,-0.1675598919,0.3808234334,0.276889354,-0.1648255736,-0.0369857587,-0.3196110725,-0.2903357744,-0.3800643384,-0.0481718555,0.1076278836,0.4150267541,0.0340872742,0.1011691615,-0.2938578725,-0.0260187201,-0.0172521118,0.0750777125,0.3295866251,-0.0051776441,0.2987500429,0.0001269509,0.4790292084,0.5336917639,0.608314693,-0.0790778473,-0.168248862,0.1742368191,0.2388187945,0.056849625,0.0911385417,-0.1643012762,-0.0760752335,0.1003498286,0.1237166375,0.0757639632,0.4928692579,0.1997979432,-0.101721786,-0.3060792685,-0.2848143578,0.3799953163,0.0059596449,0.1839969456,0.3617464006,0.1735025197,-0.1793793142,0.1578198373,-0.2865616083,0.8279553056,-0.088301301,0.1863437593,0.2034919858,-0.1707055271,0.4270977676,0.0044306559,0.2608374953,-0.3066089451,-0.1917940527,0.1068552136,-0.2196325362,-0.0452402234,0.037217807,-0.0017503271,0.2464989573,-0.171620518,-0.2535544038,-0.1752812415,0.4142414927,-0.0600915626,-0.1273560226,-0.2972863317,0.1233005896,0.0636470839,0.47649014,-0.2641223967,-0.022815749,-0.1227031499,-0.2786701322,-0.2843520641,0.2614377439,-0.0923176631,0.319498539,-0.3142413497,-0.199461177,-0.0117253065,-0.1410720944,0.0119980527,-0.0121824266,-0.1136445478,0.09910357,0.025367748,-0.4313670993,-0.012650975,0.10936331,0.4606412053,-0.1748532355,-0.0569606051,-0.0439526103,-0.1209330708,-0.0764138401,0.0666377321,0.0606505424,-0.1704290211,-0.2852598131,-0.3929140866,0.0057140384,-0.0192151498,-0.2860834301,0.1495039016,0.0541306585,-0.1489330828,-0.141005531,0.0631333962,-0.369474411,-0.261447221,0.5848149657,-0.2290470451,-0.1672657281,0.4755726159,0.1389335841,-0.2453746349,-0.2389470786,-0.057756912,0.2142287195,-0.3769004345,0.1933730543,0.1458589584,0.210410431,0.0994283855,0.4053385556,0.2238170207,-0.0655532181,0.163526088,-0.5226733685,-0.2048822045,0.2629172504,-0.0754835159,0.1675189137,0.0834503621,0.0323742144,0.1139416397,-0.0411145985,-0.3581664562,0.0279709678,-0.2944269776,0.0520219095,0.1817686409,-0.0211625379,0.357322067,0.2104436606,0.2412478775,-0.095286347,-0.1011711806,-0.2811157405,-0.1054296345,0.1371492594,-0.0351853035,-0.0646686256,-0.1630118936,-0.1591403484,-0.1330506504,-0.0642687753,0.2809072137,0.1647329032,0.0117628146,-0.284776479,0.0981535316,-0.0421203226,-0.1719227284,0.2252429426,0.1075998917,0.0509256944,0.2489593774,0.2051205486,0.0316969715,-0.1099842787,0.1387232691,-0.2905350626,0.0681333989,-0.0240929034,0.2176180631,-0.2760128379,-0.2741425633,0.1538185179,0.274525255,0.4074634016,0.0081729321,0.0470387936,0.2161379755,0.2668988705,-0.3773310483,0.0189239737,0.0285646226,-0.1256906539,0.2468962669,0.255073756,-0.106856972,0.0481030643,0.1360794455,0.2231007814,0.4210959971,-0.0778731555,0.2160880417,-0.1783435792,0.0489661582,-0.1044896618,0.3343425989,0.2829759419,0.4328795373,0.2902895808,-0.1596602052,0.3407626152,-0.1692937315,0.1661964655,-0.0429286435,-0.4089371264,0.3431918025,0.155533269,0.1181677505,-0.0545903109,-0.2855466604,-0.0084881354,-0.0694933012,0.2013454139,-0.1192132086,0.2931375206,-0.1820884347,-0.1090704128,0.0934147164,-0.4112871587,0.079889521,0.0461072847,0.1834768653,-0.1655807644,0.1462471783,0.2682876885,-0.1675834507,-0.0777960047,-0.0180541705,0.3332146406,0.1199546829,-0.2109153569,0.2081679106,0.2205871046,-0.1167477816,-0.1216178611,0.355735302,0.5909681916,0.3230317831,0.1966940165,0.0063568894,0.1522777081,-0.1297103465,-0.1076303869,0.1393267363,-0.0713555291,0.2352316678,0.175805375,0.2756284177,-0.2915883958,0.0513826981,-0.0899239033,-0.0341186747,-0.0582948178,0.2172850668,-0.0469961315,-0.1263225377,-0.3412036002,0.2155513465,-0.5843129158,0.4456870556,0.2231096029,0.010193604,0.1192841381,0.0343007967,0.1173476353,0.1104774997,0.4590658844,0.3333301544,0.2629753649,-0.3781104088,-0.0899777859,-0.5046923757,0.2131059915,-0.1343518198,0.0054264665,0.0841841772,0.236955747,-0.0553477108,0.0025013217,0.1508084685,0.2183922082,0.0902341604,-0.0175418463,-0.2637939453,-0.0161259435,0.0706616193,0.0177553222,0.1297650337,-0.3305588663,0.0024301747,-0.4772558808,0.095216684,-0.1454280168,-0.2248672396,-0.1889298856,-0.342071265,0.3520760536,-0.1208494604,0.159047097,-0.1975364834,-0.1927533597,-0.2993255258,-0.4522846639,-0.1979387552,0.3398811817,0.3392877579,0.2523281574,-0.2662300169,-0.1787528545,-0.4373828769,0.2831882834,-0.0156293456,-0.2114800364,0.1319566667,-0.0878060684,0.0294063874,-0.0275121499,0.0575888157,0.1131119058,0.0304558743,0.0242230408,-0.3524520993,-0.321449101,0.5321993232,-0.1855504811,-0.0950867981,-0.0717512891,0.2673282027,0.0405889079,-0.0020066877,-0.5231081247,0.5159048438,0.4153162539,0.0681953207,-0.1566002816,0.1899167001,-0.2691119313,0.0465724282,-0.0812959895,0.1766751111,-0.006289144,-0.2810450196,-0.1985638887,-0.0266595725]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1639","title":"bug with sst2 in glue ","comments":"Maybe you can use nltk's treebank detokenizer ?\r\n```python\r\nfrom nltk.tokenize.treebank import TreebankWordDetokenizer\r\n\r\nTreebankWordDetokenizer().detokenize(\"it 's a charming and often affecting journey . \".split())\r\n# \"it's a charming and often affecting journey.\"\r\n```","body":"Hi\r\nI am getting very low accuracy on SST2 I investigate this and observe that for this dataset sentences are tokenized, while this is correct for the other datasets in GLUE, please see below.\r\nIs there any alternatives I could get untokenized sentences? I am unfortunately under time pressure to report some results on this dataset. thank you for your help. @lhoestq \r\n \r\n```\r\n>>> a = datasets.load_dataset('glue', 'sst2', split=\"validation\", script_version=\"master\")\r\nReusing dataset glue (\/julia\/datasets\/glue\/sst2\/1.0.0\/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4)\r\n>>> a[:10]\r\n{'idx': [0, 1, 2, 3, 4, 5, 6, 7, 8, 9], 'label': [1, 0, 1, 1, 0, 1, 0, 0, 1, 0], 'sentence': [\"it 's a charming and often affecting journey . \", 'unflinchingly bleak and desperate ', 'allows us to hope that nolan is poised to embark a major career as a commercial yet inventive filmmaker . ', \"the acting , costumes , music , cinematography and sound are all astounding given the production 's austere locales . \", \"it 's slow -- very , very slow . \", 'although laced with humor and a few fanciful touches , the film is a refreshingly serious look at young women . ', 'a sometimes tedious film . ', \"or doing last year 's taxes with your ex-wife . \", \"you do n't have to know about music to appreciate the film 's easygoing blend of comedy and romance . \", \"in exactly 89 minutes , most of which passed as slowly as if i 'd been sitting naked on an igloo , formula 51 sank from quirky to jerky to utter turkey . \"]}\r\n\r\n```","comment_length":32,"text":"bug with sst2 in glue \nHi\r\nI am getting very low accuracy on SST2 I investigate this and observe that for this dataset sentences are tokenized, while this is correct for the other datasets in GLUE, please see below.\r\nIs there any alternatives I could get untokenized sentences? I am unfortunately under time pressure to report some results on this dataset. thank you for your help. @lhoestq \r\n \r\n```\r\n>>> a = datasets.load_dataset('glue', 'sst2', split=\"validation\", script_version=\"master\")\r\nReusing dataset glue (\/julia\/datasets\/glue\/sst2\/1.0.0\/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4)\r\n>>> a[:10]\r\n{'idx': [0, 1, 2, 3, 4, 5, 6, 7, 8, 9], 'label': [1, 0, 1, 1, 0, 1, 0, 0, 1, 0], 'sentence': [\"it 's a charming and often affecting journey . \", 'unflinchingly bleak and desperate ', 'allows us to hope that nolan is poised to embark a major career as a commercial yet inventive filmmaker . ', \"the acting , costumes , music , cinematography and sound are all astounding given the production 's austere locales . \", \"it 's slow -- very , very slow . \", 'although laced with humor and a few fanciful touches , the film is a refreshingly serious look at young women . ', 'a sometimes tedious film . ', \"or doing last year 's taxes with your ex-wife . \", \"you do n't have to know about music to appreciate the film 's easygoing blend of comedy and romance . \", \"in exactly 89 minutes , most of which passed as slowly as if i 'd been sitting naked on an igloo , formula 51 sank from quirky to jerky to utter turkey . \"]}\r\n\r\n```\nMaybe you can use nltk's treebank detokenizer ?\r\n```python\r\nfrom nltk.tokenize.treebank import TreebankWordDetokenizer\r\n\r\nTreebankWordDetokenizer().detokenize(\"it 's a charming and often affecting journey . \".split())\r\n# \"it's a charming and often affecting journey.\"\r\n```","embeddings":[0.1141937822,-0.191114828,0.0535292141,0.1571243256,0.1562406421,-0.3605676293,0.1134419814,0.4331327975,-0.0932659656,0.0230266862,-0.1225385964,0.1391696036,-0.0974969193,0.0586754978,-0.16858913,-0.1118808985,0.2775104642,0.0810054019,-0.3122582436,-0.2152646184,0.1465487629,0.3253771961,-0.2217946947,0.1962158382,-0.3749184012,0.0005402695,-0.0025017459,-0.1653343439,-0.147021547,-0.1796660125,-0.361616075,0.1227448359,0.142877385,0.2076182663,-0.0001284339,-0.2215033621,0.2694629729,-0.2194067836,0.0018538549,0.2560395002,0.0565842278,-0.5237965584,0.0025379027,-0.3194332123,-0.3669329584,0.1096748784,-0.0938306451,-0.2007995695,0.5336482525,-0.1431433856,0.076762639,0.2225267589,-0.0619566552,0.1098598987,0.1986176372,0.0025218893,-0.0819646716,0.0560987182,0.0948231965,-0.0120578278,0.1472665966,0.2358227223,-0.124583222,0.0914944485,0.0435767211,-0.0318556316,-0.3331217766,-0.3354297876,0.1253537089,0.3483810127,0.295286864,-0.0750595182,-0.5162137747,-0.2952619195,-0.0524900481,0.2702248991,-0.1150405854,0.058658164,0.1640500277,0.4188830256,-0.1041115001,-0.2002945542,-0.0514230505,-0.0067006447,0.1387176812,0.3399919868,0.0612359568,0.1504879743,-0.1737190634,-0.2166426033,-0.4400324523,0.1399834752,-0.5158091187,-0.0077790557,-0.464527607,-0.1171266735,-0.3323483467,0.0888856053,-0.094950065,0.2479356527,-0.2073272169,0.1058496088,0.3114411533,-0.2439594269,0.155563429,0.2846984863,0.2777003348,0.3676304817,0.1604797393,-0.4486633539,-0.1351236403,-0.0723751634,-0.1148506701,0.0376185514,-0.104485102,0.0774470791,-0.078829743,-0.2549705803,-0.375463903,0.0849398673,-0.525841713,0.2092369944,-0.3027280867,-0.0012374806,-0.1326924413,0.1751765609,0.0157193076,0.2238351852,-0.0487246774,-0.5905393958,-0.0669179112,0.0601003617,-0.0542647317,0.0192328487,-0.1725622267,0.1348020285,0.0291421395,-0.0153351994,-0.2141216993,-0.0353840515,-0.24141936,-0.3270127177,0.1687343866,-0.0320347399,-0.2492668033,0.3267421722,-0.1242565215,-0.3437110186,0.1826789826,-0.0952266902,-0.2481368035,0.0048111398,-0.3740372956,-0.072741501,-0.7066127658,-0.0345190354,0.0738087818,0.3483802676,-0.0170580912,0.3711566925,0.056876231,-0.1334564388,-0.0393762775,-0.0836961716,0.3642608821,0.5312820673,-0.7039783597,-0.1444991976,0.2193377018,0.037804082,0.4260115027,0.5988708138,0.3259574771,0.3910284042,-0.1165050268,0.6638131738,0.0409003161,-0.1775779426,-0.4084926248,0.3672143817,-0.0454975516,0.0683856681,0.1460836977,-0.0276188366,0.6220766902,0.0999988765,0.13522695,0.1166748703,0.2677462995,-0.0580955893,-0.3883869946,-0.1797920465,0.2969615161,-0.0649921969,-0.0748581663,-0.1737874299,-0.0679684281,-0.0006405268,0.1131375358,-0.0443205461,0.1762990952,0.1299658269,0.2252661288,0.3682638407,0.0285546631,0.0050624097,0.0588335507,-0.0166372955,-0.106705755,0.2543140352,-0.1692504883,-0.0433502607,0.0708888695,-0.1402979493,0.2227377146,-0.0166765843,0.0620166548,-0.0265861265,-0.097810708,0.1673720181,-0.0492919907,0.1474598199,0.1114816144,0.1355089992,0.1175305471,0.4481175244,0.2730052471,-0.5152079463,0.0349517204,0.2944254577,0.2193368226,-0.3180552423,-0.1498562545,0.4196158051,-0.2714055479,0.170766592,-0.1787165105,0.2072255611,-0.1719323099,-0.3247873783,0.1816176623,0.2580245137,0.0087032681,0.1195492223,0.1745596975,0.2186061442,0.2944153249,0.0248743445,-0.2141067535,0.0806526095,-0.0175328758,-0.0525173694,-0.3154951334,-0.2962321639,0.2288537174,-0.2663196623,0.0239960402,0.0651253909,-0.2167958617,0.2262640297,0.3028301895,-0.0247993115,0.3780799806,-0.0241712723,-0.4174866676,0.0829204246,0.1892524064,0.2949672043,0.0403658897,0.1943738759,-0.1516488045,-0.0976008922,0.3454198539,-0.0168022793,0.2071207911,0.1002804637,-0.2213463336,0.1448145807,0.1498964429,0.1462793648,-0.0215860773,0.4351988137,-0.2174084038,0.2660402954,-0.5064103603,-0.2188868523,-0.5575281978,-0.5232078433,0.0320059061,0.270314455,-0.0204196051,-0.0448600911,0.0462543927,-0.3525875211,0.0211024024,0.2890893519,0.1296738833,0.2662609518,-0.0789649412,0.0987234786,-0.0341484174,0.1235249415,-0.3036597669,-0.1553719491,0.0000800888,0.0990958661,0.1361237764,-0.207561478,-0.0551052243,-0.4055820405,-0.2890824974,0.1528496295,-0.2902615368,0.3968253732,0.3595591784,0.0558852814,-0.2514448166,0.0870464295,0.2250032425,0.0028077357,-0.0131191052,0.1350421757,-0.1564460397,-0.2393328846,-0.1441880018,-0.0708739907,-0.1651630253,0.1086726934,0.3374187648,-0.3074238598,-0.0873105377,-0.1631095707,-0.425998956,-0.3015058041,-0.2955172956,-0.1434105188,0.0432185121,-0.3151704371,0.4570654631,0.243942067,-0.2985363007,-0.2668834925,0.0724310428,-0.369666487,-0.3661431074,-0.0754311308,-0.1193569973,-0.4101173878,0.0520401448,-0.160611555,-0.1784392148,0.1711137891,-0.0700922534,0.0546809845,-0.0876444876,-0.4567007422,0.62574929,0.2299714088,-0.004340386,-0.0116407303,0.5590554476,0.5441452861,-0.0169792324,0.1982869655,-0.1012941003,0.1787251234,-0.1645834297,0.1216292232,0.0102444943,-0.1149952784,0.4203284383,0.104905501,0.2937069535,0.3908886313,0.055480279,0.0491352193,0.1942737848,-0.0904615745,-0.4015384316,0.0305713955,-0.1464715153,-0.2181120068,0.0794878006,0.0151801668,0.1855860353,0.0239886492,-0.3088745475,0.3613342643,-0.1971510202,0.5832237601,0.1977724582,-0.6095137596,0.0065384391,-0.4076495171,0.4030979574,0.0859934464,0.520755589,-0.0730758011,0.0114242295,0.2866304219,0.0292703733,0.6724801064,-0.2632599473,0.1373614222,-0.0254670121,0.0512801707,-0.5708633065,-0.4011025429,-0.3983349204,-0.0258871615,0.655374825,0.3363977373,-0.3256053329,-0.2960943282,0.1046884954,0.0701974407,-0.196932748,-0.246164605,0.1725447625,0.278608501,0.0833283961,-0.1072144583,0.3308826983,-0.0268225428,0.2947429419,-0.0791192651,-0.1034616083,-0.280448705,0.2461508512,0.0850126818,0.5140622854,-0.2361581773,0.1135241538,0.096375905,0.2140444517,-0.2740834653,0.0490073524,-0.0269697644,-0.193842262,-0.0021207044,-0.3352290094,0.3915080428,0.1036204547,-0.1898939908,0.3069439828,-0.496329844,0.2118797749,-0.157628566,0.0535491258,0.384251833,0.2778591514,-0.0135934195,0.2110220939,-0.1121271253,0.0732405037,-0.2107790112,-0.7628881335,0.4255465865,-0.3396132886,0.2636586428,0.1486216187,0.9446775317,0.1546033174,-0.1362092346,-0.1566053778,-0.1592583656,0.6799587011,-0.3395628333,-0.0568055436,-0.3854139745,-0.0903909206,0.0365038775,0.2030341178,-0.1233872175,0.3410128951,-0.3238176405,0.4729294479,-0.1593893915,0.2125936002,0.0230199583,0.2544322908,0.4401401281,0.0512923636,0.1633998454,-0.003491038,-0.1585545689,0.0276129376,0.043096371,-0.0826073065,0.1365879923,-0.0041123689,-0.5293961167,-0.0904470459,-0.2544338405,-0.1894884109,0.4944370389,0.0769486204,0.1126224101,0.1134762317,0.4654404223,0.4525073171,-0.0271510929,0.00619906,0.1584913731,0.0939728245,0.4269345403,0.2403831035,0.0354300812,-0.1022597924,-0.2098260224,-0.3369402587,0.1585659534,-0.2532201111,-0.2527690828,-0.3797810972,0.4875214994,-0.149960041,-0.0046301652,0.072149463,0.2281367332,-0.0877102688,-0.0392666534,0.0319559053,-0.0908121318,0.3694455624,0.1919644624,-0.3186237216,0.0501737744,0.3725883663,-0.2465496063,0.12455789,0.5588754416,-0.0406712033,-0.1990904808,-0.2013646513,0.1970104426,-0.0559379086,-0.3591531515,0.1621854901,-0.1399756074,0.0351485014,-0.1826298386,0.0436186865,0.2351416647,-0.1816456765,-0.4590078294,-0.1890336722,0.085193716,0.0789415836,0.0679363534,0.0708927363,-0.1127550006,0.1678717136,-0.0820318982,-0.0385819413,-0.1471495926,0.1248494759,0.0054170727,0.0471552722,0.2646822631,-0.0535983369,-0.3119609058,-0.0513138138,0.0377619006,0.1042291,0.1823804677,-0.0492523313,-0.0915895849,0.1792733669,-0.0362687372,0.0124706868,0.1246877089,-0.0439785831,-0.051332999,-0.1533373594,0.0440196246,0.2600450814,0.0642997548,0.5977873802,0.6043590307,-0.0524598025,-0.2097425759,0.1772461236,0.0624981411,0.1499682218,0.0009378643,0.2768410742,0.4794325829,-0.0167957488,-0.0201252978,0.1143737733,0.1681886613,0.4496702552,0.0179574192,-0.1475858092,-0.110593386,-0.0708448589,0.0144898137,0.0213268623,-0.0437376574,-0.5781181455,0.3267469108,0.0831645951,-0.145402506,0.0391671211,0.1399260908,0.1559184194,-0.1391393542,0.1385854632,0.1983237267,0.0575756542,0.0945804343,-0.0920750871,0.4965296388,0.0697272494,-0.1750065386,0.0354388058,0.0901099518,-0.2029357255,-0.0117359012,-0.4933754206,-0.0545051247,0.1416080147,0.4102309644,0.0503334031,0.6172514558,0.4189927876,0.1054476798,0.2707446814,0.025804257,0.257129848,0.0158989783,0.0999811068,0.0105539896,0.4147476554,0.2474181503,-0.6086637378,-0.2462768257,0.2609453797,-0.1339474767,-0.4249436259,-0.1618660539,0.0342887193,-0.17129682,0.0046701445,-0.211333409,0.0599800721,-0.0153554128,0.0578866601,0.1201384962,-0.1129411757,-0.4338991046,-0.0775609016,0.4074472785,-0.1977061927,0.5040383935,-0.2116229832,0.1686112881,0.1107261404,0.7201171517,0.1898798943,0.2110995054,-0.1318566203,-0.1851348728,-0.1237183511,0.0508316234,0.0986404046,0.3515788019,-0.1974090189,-0.248429507,0.1329896301,-0.0446643978,0.0137184402,0.0979646519,-0.2688438296,0.0428907797,-0.2306030393,0.2195553929,0.0032144624,-0.1488931328,0.0130949318,-0.1742384732,-0.245241493,-0.1885525137,0.0455966108,-0.3163469732,0.0331861489,-0.0043870509,0.0160191804,-0.2716775537,0.3423826098,0.1782590598,0.0973865837,0.0179106276,-0.130710572,-0.0308426414,-0.0516334698,0.0578917637,-0.107844986,-0.0937212259,0.0682279393,0.2241958529,0.0398801565,0.2100558877,0.6183136106,-0.5300003886,-0.0648628771,-0.5419040322,0.4250277877,0.2149814069,0.3155127168,0.1228665859,-0.1175931245,0.1638846844,0.0078269998,-0.15112634,0.2757664323,-0.0338716917,-0.0472641326,-0.0918110833,0.4175808132,0.4229001701,0.1106428504,0.0629245192,0.1894167662,-0.0759948194,-0.376068294,-0.501830399,0.1399919093,-0.0538614467,0.2418343425,-0.0273274444,0.1390213221,0.0733145922,-0.1430762708,0.0869128257,0.4620881379,-0.3423890471,0.1094644368,-0.0858717859,-0.0586789735,-0.3165747821,-0.1729416102,0.0850435868,0.0671132058,-0.2730297446,-0.0989847034,0.5375218391,-0.406519562,-0.0524047427,-0.2308149785,0.2738251686,0.4895904362,-0.1418654621,-0.3272663057,-0.4253805876,0.5455459952,0.2990843356,-0.0738516375,0.3723992109,0.0187542103,0.0109987585,0.0532985777,-0.0145342238,0.1540331095,-0.1106906757,-0.0508363843,-0.2397189438]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1639","title":"bug with sst2 in glue ","comments":"I am looking for alternative file URL here instead of adding extra processing code: https:\/\/github.com\/huggingface\/datasets\/blob\/171f2bba9dd8b92006b13cf076a5bf31d67d3e69\/datasets\/glue\/glue.py#L174","body":"Hi\r\nI am getting very low accuracy on SST2 I investigate this and observe that for this dataset sentences are tokenized, while this is correct for the other datasets in GLUE, please see below.\r\nIs there any alternatives I could get untokenized sentences? I am unfortunately under time pressure to report some results on this dataset. thank you for your help. @lhoestq \r\n \r\n```\r\n>>> a = datasets.load_dataset('glue', 'sst2', split=\"validation\", script_version=\"master\")\r\nReusing dataset glue (\/julia\/datasets\/glue\/sst2\/1.0.0\/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4)\r\n>>> a[:10]\r\n{'idx': [0, 1, 2, 3, 4, 5, 6, 7, 8, 9], 'label': [1, 0, 1, 1, 0, 1, 0, 0, 1, 0], 'sentence': [\"it 's a charming and often affecting journey . \", 'unflinchingly bleak and desperate ', 'allows us to hope that nolan is poised to embark a major career as a commercial yet inventive filmmaker . ', \"the acting , costumes , music , cinematography and sound are all astounding given the production 's austere locales . \", \"it 's slow -- very , very slow . \", 'although laced with humor and a few fanciful touches , the film is a refreshingly serious look at young women . ', 'a sometimes tedious film . ', \"or doing last year 's taxes with your ex-wife . \", \"you do n't have to know about music to appreciate the film 's easygoing blend of comedy and romance . \", \"in exactly 89 minutes , most of which passed as slowly as if i 'd been sitting naked on an igloo , formula 51 sank from quirky to jerky to utter turkey . \"]}\r\n\r\n```","comment_length":15,"text":"bug with sst2 in glue \nHi\r\nI am getting very low accuracy on SST2 I investigate this and observe that for this dataset sentences are tokenized, while this is correct for the other datasets in GLUE, please see below.\r\nIs there any alternatives I could get untokenized sentences? I am unfortunately under time pressure to report some results on this dataset. thank you for your help. @lhoestq \r\n \r\n```\r\n>>> a = datasets.load_dataset('glue', 'sst2', split=\"validation\", script_version=\"master\")\r\nReusing dataset glue (\/julia\/datasets\/glue\/sst2\/1.0.0\/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4)\r\n>>> a[:10]\r\n{'idx': [0, 1, 2, 3, 4, 5, 6, 7, 8, 9], 'label': [1, 0, 1, 1, 0, 1, 0, 0, 1, 0], 'sentence': [\"it 's a charming and often affecting journey . \", 'unflinchingly bleak and desperate ', 'allows us to hope that nolan is poised to embark a major career as a commercial yet inventive filmmaker . ', \"the acting , costumes , music , cinematography and sound are all astounding given the production 's austere locales . \", \"it 's slow -- very , very slow . \", 'although laced with humor and a few fanciful touches , the film is a refreshingly serious look at young women . ', 'a sometimes tedious film . ', \"or doing last year 's taxes with your ex-wife . \", \"you do n't have to know about music to appreciate the film 's easygoing blend of comedy and romance . \", \"in exactly 89 minutes , most of which passed as slowly as if i 'd been sitting naked on an igloo , formula 51 sank from quirky to jerky to utter turkey . \"]}\r\n\r\n```\nI am looking for alternative file URL here instead of adding extra processing code: https:\/\/github.com\/huggingface\/datasets\/blob\/171f2bba9dd8b92006b13cf076a5bf31d67d3e69\/datasets\/glue\/glue.py#L174","embeddings":[0.0745585188,-0.1421597451,0.0459705293,0.1520761102,0.174217999,-0.3497954607,0.1239791811,0.4414982796,-0.0777454078,0.042793002,-0.1272771657,0.0841267407,-0.1058053002,0.1174532697,-0.1189543828,-0.0650619343,0.2794892192,0.0854180381,-0.2919892669,-0.1956674308,0.1490904689,0.3207139373,-0.2032112777,0.2129850239,-0.3381327987,0.0296920799,0.0325951204,-0.1514992416,-0.1419923156,-0.180888325,-0.3845718801,0.1444615275,0.1710386574,0.2168252319,-0.0001281228,-0.1787853688,0.2562443018,-0.2393691689,0.057118807,0.2527914643,0.018208988,-0.5082676411,-0.0631703734,-0.3029416203,-0.3309911788,0.0838461667,-0.0600790344,-0.2228627056,0.5005704165,-0.1284409612,0.0735829622,0.1987253726,-0.1111456901,0.0749078467,0.2200063467,-0.0057845651,-0.0736334845,0.0745038539,0.1413089037,0.0020450223,0.1571166068,0.2682349086,-0.130225867,0.0629358366,0.0115578277,-0.0619002134,-0.3378021419,-0.2963066995,0.1673830301,0.3620509505,0.3472058475,-0.0826249048,-0.541495204,-0.2534787655,-0.0288481768,0.2996871471,-0.0991654098,0.0660687983,0.1693732142,0.4488847852,-0.0876655355,-0.2370864302,-0.0684333146,-0.0525636673,0.1725391448,0.3133468628,0.0454012789,0.130567193,-0.1665721536,-0.2188275009,-0.4552412629,0.152153194,-0.5214688182,-0.0590890869,-0.4590846896,-0.1299513727,-0.3927189112,0.0470860191,-0.0484406911,0.2012428194,-0.1717160493,0.13263385,0.3389309943,-0.2486032993,0.1147647277,0.2529625297,0.2547122836,0.3145361245,0.1837242693,-0.3959207237,-0.1219858304,-0.0871436521,-0.1205767989,-0.0264458582,-0.1362355649,0.0567488708,-0.1064728722,-0.2545301616,-0.387507081,0.0558294803,-0.509673357,0.2154132426,-0.2997604609,0.0689242557,-0.1426026821,0.1608933061,0.0399775542,0.1592312455,-0.0576289818,-0.6441297531,-0.0639934912,0.0799270645,-0.0836720094,0.070674412,-0.1725990325,0.1365127861,0.037578661,0.0006292518,-0.2282691747,-0.0240339153,-0.2165538222,-0.291898936,0.1699992865,-0.0456942692,-0.2129768431,0.3382775784,-0.1610167474,-0.343121469,0.190484345,-0.0913120061,-0.2817988992,-0.0491657294,-0.3789661825,-0.0685515106,-0.6872428656,-0.0541146174,0.0517443381,0.3418417871,-0.0545005016,0.3128675818,0.051270809,-0.1338974535,-0.0421197899,-0.0858567432,0.410367161,0.5998943448,-0.7683393955,-0.1543697566,0.2466992289,0.0499928258,0.4027200043,0.5945767164,0.3157579005,0.3988308609,-0.1388723254,0.653686583,0.0469471291,-0.2293391675,-0.425393343,0.3818599284,-0.0297449082,0.0337096415,0.2003352344,-0.0215147827,0.6545792222,0.1253730655,0.1215310916,0.0623579919,0.2530224919,-0.0708083063,-0.3732642829,-0.2148115337,0.2607968152,-0.0442267805,-0.1063527241,-0.189644888,-0.0677399039,-0.0212944727,0.1065382436,-0.0734855235,0.1825780272,0.1393817067,0.1886266172,0.3983444571,0.0521911979,0.0067722201,0.0300913844,-0.0233629271,-0.0955456346,0.1813118011,-0.1720803678,-0.0324676745,0.0661886111,-0.186360687,0.2101015449,-0.0362818018,0.0577339455,-0.0077825561,-0.1124014333,0.1344249249,-0.0700101554,0.1543043703,0.0546942092,0.1232320368,0.109074153,0.4224333465,0.2634762824,-0.5047082901,0.0588282123,0.2525965273,0.1997754425,-0.3339441717,-0.1483002454,0.4213383198,-0.2713469565,0.2161646485,-0.0709468052,0.2263563126,-0.1845433861,-0.3455144465,0.1679254621,0.2203757167,-0.0039858548,0.1773364991,0.1183517501,0.1743477434,0.2753558755,0.0114889555,-0.1884243339,0.0631098598,0.0448137522,-0.0833020136,-0.3350508511,-0.271776706,0.1933527738,-0.2434646934,0.0680718496,0.0611235127,-0.2262559533,0.216725409,0.3562335968,-0.0367188454,0.3578252494,-0.0042302059,-0.4323099852,0.0910971835,0.1740958691,0.3240964413,0.0096430611,0.2090380937,-0.1724333465,-0.1125440076,0.3725584745,-0.0330926031,0.1964356303,0.076853618,-0.2449943572,0.2020053267,0.1366926581,0.1378288269,-0.0142513895,0.4006648064,-0.2186388969,0.2517690063,-0.5362452269,-0.2559605837,-0.575268805,-0.4609850645,0.0074297236,0.3293522298,0.0011627615,0.0368644856,0.0714094937,-0.388866961,-0.000480765,0.2885791957,0.1008934081,0.3087309003,-0.0650193542,0.1032405645,-0.0150299519,0.1434754133,-0.2794549465,-0.1621658355,0.0366956145,0.0794391334,0.1174676046,-0.2465489805,-0.0508102551,-0.4670237303,-0.2551806271,0.1549929231,-0.2163087875,0.4059416354,0.3499673903,0.0839058608,-0.2182663977,0.1059995666,0.2052865326,0.0020178554,0.012316335,0.0673857629,-0.1622826904,-0.2396754175,-0.1231449619,-0.1074826419,-0.1655832529,0.0853161663,0.2936834991,-0.3694170713,-0.0902197137,-0.2226535231,-0.3832156062,-0.2802067697,-0.2846973538,-0.1446311176,0.0462168604,-0.359788388,0.4777298272,0.2219733149,-0.305857718,-0.2205470055,0.0583810657,-0.3366681635,-0.3445280194,-0.1096875891,-0.1245107427,-0.3707384169,-0.0057431948,-0.1713929474,-0.2006957829,0.2087658346,-0.0704255551,0.0521358103,-0.1239510998,-0.4854849875,0.6383935809,0.2014977783,-0.0095859896,-0.0016563386,0.5295645595,0.5737708807,0.0658330098,0.1802886277,-0.0984318033,0.1484482288,-0.1683570296,0.1294168532,0.0778244436,-0.1267541647,0.3822645545,0.1119606942,0.3343070447,0.3808689415,0.054794848,0.1188559607,0.2067398578,-0.0416861922,-0.4273850024,0.0604476742,-0.134518072,-0.1697849631,0.1159317046,0.0176229943,0.1561875045,0.0044255215,-0.2575780451,0.423966527,-0.1089444533,0.6160780191,0.2164475322,-0.6053961515,0.0175656341,-0.3483980894,0.4273315072,0.1124111414,0.5075610876,-0.1169241965,0.0240345057,0.257777065,-0.0092395805,0.7340770364,-0.3072482347,0.1153621376,-0.0316919498,0.0508111604,-0.5489913225,-0.3965366781,-0.3833631277,-0.0441944636,0.6220102906,0.38655442,-0.3147189021,-0.3090786636,0.0973459035,0.0343532152,-0.2295440137,-0.226224944,0.1903225332,0.2566261888,0.060188029,-0.1067835391,0.309340477,-0.0607509874,0.3010923564,-0.1651995033,-0.0684712678,-0.2529852986,0.2650109828,0.0873100758,0.5050092936,-0.2309360653,0.1478457302,0.1361488104,0.2218758911,-0.1797953546,0.0261122938,-0.0525787845,-0.2611317635,0.0446112044,-0.3197729588,0.3771215677,0.1262725145,-0.175247848,0.3117197454,-0.4282632172,0.1855051517,-0.1974612176,0.0809398368,0.4267759323,0.2725757658,-0.0366484858,0.1700849682,-0.131605953,0.0822238848,-0.2032157034,-0.7338915467,0.4210218787,-0.3402369916,0.2357174605,0.1311230659,0.9469517469,0.1278921068,-0.1374029666,-0.1934461594,-0.1525736749,0.6441377997,-0.3562262952,-0.0304254871,-0.3888594508,-0.1142040193,0.0278939586,0.1569900364,-0.0823465362,0.3520277739,-0.333571285,0.4684766233,-0.1305635422,0.2177507877,0.0314206108,0.2866870165,0.4293083847,0.0279990546,0.1872750074,-0.0150948679,-0.1520209908,-0.0095794238,0.0567854717,-0.1066754907,0.1708406061,0.0038383394,-0.5374071002,-0.1086087599,-0.2934653163,-0.2303611487,0.4758560061,0.0884269774,0.1581940353,0.0834505782,0.4324457049,0.4546601474,-0.0270784758,0.0384245031,0.120095998,0.082983613,0.4012922049,0.2306598127,0.0121553233,-0.0953677371,-0.1974153519,-0.3199685812,0.2188015282,-0.3044162393,-0.2192700207,-0.409432292,0.510081768,-0.2094505429,-0.0166579317,0.0747929588,0.2353459299,-0.1047526449,-0.0440190807,0.0447952598,-0.1008649766,0.3739907742,0.209572196,-0.3199503422,0.0513214469,0.3427315056,-0.2861298919,0.1079257056,0.5481642485,-0.0708785802,-0.1995642036,-0.1974071115,0.2181756794,-0.0762734339,-0.3623617887,0.1718565971,-0.1532324553,0.0445444211,-0.1948426217,0.0003652898,0.2345201522,-0.1928902417,-0.4613094032,-0.2070197761,0.1286639869,0.1100764647,0.0697770938,0.0669837594,-0.108362563,0.1621014774,-0.1433106214,-0.0471334979,-0.1552604139,0.1899020523,0.0024320513,0.0548200682,0.2747702003,-0.0299594309,-0.3083523512,-0.0683917031,0.022541007,0.101236321,0.1495862901,-0.0483529083,-0.1127404273,0.1808944494,-0.0526941009,0.0313757956,0.1248609871,-0.0452271998,-0.054568477,-0.1377556324,0.0301712099,0.2743552923,0.0400803834,0.5759860873,0.5702604651,-0.0799893215,-0.236589402,0.1911097616,0.0989343151,0.2294850498,0.0231660716,0.2979408801,0.4484456182,-0.0284092221,0.0209743865,0.1467289925,0.227995187,0.4814066887,0.0386313498,-0.1284182817,-0.1062688157,0.0070917909,-0.0023648073,0.0057242247,-0.0813803822,-0.5254739523,0.3409698904,0.0757299811,-0.1325325221,0.0460400581,0.1315566599,0.0972111672,-0.1394571811,0.1022302657,0.2212896347,0.0730976239,0.1850413531,-0.0878993347,0.4996817112,0.0819106996,-0.1560335159,0.0169126336,0.0723280683,-0.1931060106,-0.0008757007,-0.4522145689,-0.0277084056,0.1565909684,0.4299027324,0.00783653,0.6429364681,0.4613454342,0.0601639971,0.2086228877,0.0290443804,0.2467204928,0.0139253149,0.1379866153,-0.0010694013,0.4090626836,0.2206773311,-0.5823296905,-0.2110348493,0.219209373,-0.1023736596,-0.4218310118,-0.1671013385,-0.0017164857,-0.1707947999,0.0472070426,-0.1818715036,0.0602118857,-0.008427063,0.0785045847,0.1160311326,-0.0688953176,-0.450640142,-0.0662756935,0.3940516412,-0.1867667586,0.5020816326,-0.192545563,0.20047611,0.1575916857,0.7569479346,0.1931516975,0.1784943789,-0.1290636659,-0.1858231723,-0.1339013278,0.0831062198,0.1162535027,0.3591114581,-0.2091014534,-0.3103852272,0.097144559,-0.0366843976,0.0286213271,0.0866496414,-0.2670866847,0.0310391281,-0.2565017045,0.1823730171,0.0261645336,-0.1244462579,-0.0019599637,-0.1575094312,-0.249512434,-0.2015335709,0.0923577175,-0.3580392599,0.0330192745,0.0082934005,0.0183555149,-0.2366770953,0.3184246719,0.1800001562,0.1012202054,0.0007681306,-0.115864411,-0.021633571,-0.0368281491,0.0871987864,-0.029143082,-0.1245075241,0.0705245212,0.2240510732,0.0044589275,0.2844087481,0.558818996,-0.4886174202,-0.0806684867,-0.5145741701,0.436722368,0.2527570724,0.354859978,0.1189614013,-0.1582122445,0.1974625587,0.0122648953,-0.1493365765,0.3073656559,-0.0562026203,-0.0320240594,-0.1418784857,0.4778134823,0.3932236731,0.0907470509,0.0236927439,0.1769576967,-0.0606026687,-0.4032883048,-0.4964973032,0.1318101287,-0.0502242856,0.2067283541,-0.0436508916,0.0945211127,0.0296184029,-0.1499549448,0.0504882559,0.4450700283,-0.3452331424,0.0767427757,-0.072020404,-0.1048899367,-0.3214932084,-0.1773848534,0.0487442948,0.0493692309,-0.2866925299,-0.0542983823,0.5470992923,-0.3944108784,-0.0523121059,-0.2148030102,0.2991603911,0.4605790079,-0.0626913235,-0.2892771661,-0.4080733657,0.542745471,0.3417378366,-0.0617529415,0.3759545684,0.0178890806,0.0287658758,0.0590533689,-0.0123589328,0.1946405619,-0.1627248377,-0.0458943062,-0.2679139674]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1639","title":"bug with sst2 in glue ","comments":"I don't know if there exists a detokenized version somewhere. Even the version on kaggle is tokenized","body":"Hi\r\nI am getting very low accuracy on SST2 I investigate this and observe that for this dataset sentences are tokenized, while this is correct for the other datasets in GLUE, please see below.\r\nIs there any alternatives I could get untokenized sentences? I am unfortunately under time pressure to report some results on this dataset. thank you for your help. @lhoestq \r\n \r\n```\r\n>>> a = datasets.load_dataset('glue', 'sst2', split=\"validation\", script_version=\"master\")\r\nReusing dataset glue (\/julia\/datasets\/glue\/sst2\/1.0.0\/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4)\r\n>>> a[:10]\r\n{'idx': [0, 1, 2, 3, 4, 5, 6, 7, 8, 9], 'label': [1, 0, 1, 1, 0, 1, 0, 0, 1, 0], 'sentence': [\"it 's a charming and often affecting journey . \", 'unflinchingly bleak and desperate ', 'allows us to hope that nolan is poised to embark a major career as a commercial yet inventive filmmaker . ', \"the acting , costumes , music , cinematography and sound are all astounding given the production 's austere locales . \", \"it 's slow -- very , very slow . \", 'although laced with humor and a few fanciful touches , the film is a refreshingly serious look at young women . ', 'a sometimes tedious film . ', \"or doing last year 's taxes with your ex-wife . \", \"you do n't have to know about music to appreciate the film 's easygoing blend of comedy and romance . \", \"in exactly 89 minutes , most of which passed as slowly as if i 'd been sitting naked on an igloo , formula 51 sank from quirky to jerky to utter turkey . \"]}\r\n\r\n```","comment_length":17,"text":"bug with sst2 in glue \nHi\r\nI am getting very low accuracy on SST2 I investigate this and observe that for this dataset sentences are tokenized, while this is correct for the other datasets in GLUE, please see below.\r\nIs there any alternatives I could get untokenized sentences? I am unfortunately under time pressure to report some results on this dataset. thank you for your help. @lhoestq \r\n \r\n```\r\n>>> a = datasets.load_dataset('glue', 'sst2', split=\"validation\", script_version=\"master\")\r\nReusing dataset glue (\/julia\/datasets\/glue\/sst2\/1.0.0\/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4)\r\n>>> a[:10]\r\n{'idx': [0, 1, 2, 3, 4, 5, 6, 7, 8, 9], 'label': [1, 0, 1, 1, 0, 1, 0, 0, 1, 0], 'sentence': [\"it 's a charming and often affecting journey . \", 'unflinchingly bleak and desperate ', 'allows us to hope that nolan is poised to embark a major career as a commercial yet inventive filmmaker . ', \"the acting , costumes , music , cinematography and sound are all astounding given the production 's austere locales . \", \"it 's slow -- very , very slow . \", 'although laced with humor and a few fanciful touches , the film is a refreshingly serious look at young women . ', 'a sometimes tedious film . ', \"or doing last year 's taxes with your ex-wife . \", \"you do n't have to know about music to appreciate the film 's easygoing blend of comedy and romance . \", \"in exactly 89 minutes , most of which passed as slowly as if i 'd been sitting naked on an igloo , formula 51 sank from quirky to jerky to utter turkey . \"]}\r\n\r\n```\nI don't know if there exists a detokenized version somewhere. Even the version on kaggle is tokenized","embeddings":[0.1053772271,-0.1916711181,0.0535599329,0.0827101097,0.1684059799,-0.3281565011,0.1854297668,0.4321678579,-0.0919407979,0.0309742317,-0.0943492949,0.1462736279,-0.1057705283,0.138015911,-0.1469752043,-0.0554611683,0.3053624332,0.0720337108,-0.2480497807,-0.1800416261,0.1838430911,0.3005387485,-0.2079133391,0.1841128916,-0.358858794,0.0333110765,0.0494763404,-0.2354058623,-0.1768406481,-0.2107705027,-0.3141484857,0.1596002728,0.193716079,0.2096603662,-0.0001275378,-0.2242546529,0.2872391939,-0.2060788572,0.0091374535,0.281778127,0.0366854705,-0.5301072598,-0.0000769759,-0.2940465212,-0.3653716445,0.1080832183,-0.0872452185,-0.235310331,0.5179437399,-0.1901039183,0.0725806132,0.2578645349,-0.101898618,0.0968488604,0.2016626596,-0.0233195685,-0.1244882122,0.1061551198,0.08858715,0.0119531509,0.1987287104,0.3125153482,-0.0967756957,0.0525541343,0.0309715979,-0.0561690629,-0.3840512931,-0.3257555664,0.1482466757,0.3792476654,0.3166408837,-0.0897192881,-0.5367329121,-0.2675744593,-0.041843567,0.2583253682,-0.1118208319,0.0156590901,0.1824868917,0.4119680226,-0.064618364,-0.2049472332,-0.0397313088,-0.0683754832,0.1081719175,0.3712735176,0.0445146784,0.1453737766,-0.1497977674,-0.2369308472,-0.447389096,0.1808105707,-0.4920337498,-0.0555570088,-0.4540756941,-0.1429215968,-0.38015908,0.0319992192,-0.0619642287,0.2108943164,-0.182887435,0.1540302038,0.3370283246,-0.2426043302,0.203800261,0.3650320768,0.318313688,0.3222047091,0.1798518449,-0.4099115431,-0.1268843263,-0.069605425,-0.1418879032,0.0408858433,-0.1073176712,0.0882590786,-0.0875706896,-0.2446332276,-0.4116359055,0.064416036,-0.5661545396,0.223309651,-0.3138051331,0.0089959437,-0.1119709462,0.1845390052,-0.0280314293,0.2074243277,-0.0727613941,-0.6290407777,-0.0687368512,-0.0062103234,-0.063937746,0.0341500528,-0.1824687868,0.0688116848,0.0377292112,-0.0475334823,-0.1774316877,-0.0282412209,-0.2062680423,-0.2804603279,0.1551904082,-0.0316387527,-0.2349665314,0.3630231321,-0.0941834971,-0.3709600568,0.1779542863,-0.0937622339,-0.2195090353,0.0055769766,-0.3337346613,-0.0651310906,-0.6949193478,-0.0588202439,0.0399165191,0.3861969411,-0.0074766479,0.3310188651,0.0459608547,-0.1881844252,0.0179588348,-0.0791794211,0.3777060807,0.5376538038,-0.762662828,-0.1581082791,0.1939786673,0.0355225392,0.4186126292,0.5911073089,0.3239365816,0.3647040725,-0.1242012754,0.6969619989,0.0042047054,-0.208952114,-0.4275865555,0.3860729933,-0.0376090258,0.090931505,0.1503320485,-0.0861776471,0.6578813791,0.0823523477,0.103084825,0.0325747095,0.2396647483,-0.1090309694,-0.3792620301,-0.1749333441,0.3303101659,-0.0758963674,-0.097798191,-0.1917981803,-0.0626145527,-0.0028709034,0.128158316,0.021715872,0.2007224709,0.0928254351,0.2245758623,0.3747392297,0.0294538494,0.0179387014,0.0467925519,-0.0380257629,-0.0738161653,0.2513394654,-0.1488924623,-0.0840390846,0.0735711753,-0.1498223394,0.210893631,-0.066608876,0.0683752224,-0.0589382276,-0.1446881741,0.1409900934,-0.0863773301,0.1108239144,0.1083044484,0.1027767733,0.1052518189,0.3852163553,0.221386373,-0.5122485757,0.0479111709,0.2678563297,0.214848876,-0.3405073583,-0.1625143439,0.3817720711,-0.2230755389,0.15265508,-0.1253413111,0.2494989187,-0.1301655173,-0.3278444409,0.1481739581,0.2530640662,-0.0347637199,0.1585804969,0.1489637941,0.2186182886,0.2874464989,-0.0190626495,-0.2203672677,0.1157481819,0.051628083,-0.0309161991,-0.3549937308,-0.3772973716,0.1641123891,-0.3063129187,-0.0073033432,0.0423836485,-0.2061954141,0.2887103558,0.300026834,-0.0405451059,0.3853990734,-0.0378463641,-0.4604098201,0.057044223,0.2201635838,0.2649001181,-0.0403721631,0.1799645871,-0.1132788137,-0.1288619488,0.3557143807,-0.034090627,0.171519354,0.0668156967,-0.2446060181,0.1565224379,0.134263441,0.1531965882,-0.0122299809,0.4428392947,-0.1954577863,0.2174531668,-0.4741294384,-0.2130085677,-0.587084353,-0.455011785,0.0164153036,0.2930877805,-0.0381961986,-0.0149812931,0.0942262709,-0.3488217592,0.0119788041,0.3152786195,0.1148794293,0.2802421451,-0.1082459837,0.0956882536,-0.0202123653,0.097623378,-0.3434951603,-0.1455566436,-0.0335753523,0.0611251183,0.1160847321,-0.2009541839,-0.0082516456,-0.4096842408,-0.3374337852,0.1583040059,-0.2983544469,0.4141184986,0.3482485414,0.063403815,-0.2890954018,0.1293140501,0.2046280503,-0.0259066932,-0.0365047194,0.0962841585,-0.159734115,-0.2362699062,-0.1160138696,-0.0897850245,-0.1597997397,0.1439817995,0.3131771982,-0.3221341968,-0.1053250507,-0.2012376636,-0.3947537541,-0.3288001418,-0.2766459584,-0.2073532045,0.0142798573,-0.3016605377,0.4418472052,0.2269349098,-0.3200713992,-0.2176027,0.0909380838,-0.3419412374,-0.3723222911,-0.0544716083,-0.0980043411,-0.3579964638,0.0681325346,-0.1384385675,-0.2349672019,0.2289199829,-0.0151854921,0.044205185,-0.0976490602,-0.5201345086,0.6407052875,0.2494034916,0.0104425531,-0.0207645502,0.50195086,0.5810190439,0.0396907292,0.2045617402,-0.105859749,0.1814359576,-0.1387355477,0.1529069096,-0.0077936118,-0.0941909626,0.4422138333,0.0720098615,0.2870484889,0.3915069699,0.0355925113,0.0571254306,0.2165171951,-0.0703724474,-0.3903559148,0.0493725352,-0.1600362659,-0.2360286564,0.0569487587,0.0182953477,0.2200924456,0.0204203892,-0.2840134501,0.4270584583,-0.1736555398,0.577036202,0.2225105166,-0.6082280874,-0.0222303402,-0.4122837782,0.3947543502,0.0779662728,0.5645366311,-0.0995319635,-0.0090480838,0.3066927791,0.0025272043,0.6872395873,-0.278049916,0.1542811096,-0.0125413816,0.0632714853,-0.5838255882,-0.4179600775,-0.3970817626,-0.076409705,0.6016299725,0.3341150284,-0.3360376656,-0.2747053802,0.0150328493,0.0707114711,-0.2207321078,-0.2018721849,0.1271374375,0.2625389099,0.0893819556,-0.105067268,0.3133582175,-0.0805003569,0.2673579156,-0.0805507675,-0.092421554,-0.2919735014,0.2052495629,0.1271133572,0.5027503371,-0.2434448451,0.100867033,0.1818566024,0.2136202753,-0.1835634559,0.0245420206,-0.0190701094,-0.2138529718,0.0146806957,-0.2935297191,0.3276799023,0.1193412021,-0.1721226871,0.265717566,-0.436075151,0.1697473228,-0.1880754828,0.0773708224,0.4160399139,0.257317394,-0.0128923776,0.1774767339,-0.1614689827,0.0643766671,-0.2097385824,-0.7001760602,0.4232565165,-0.3076153696,0.3103661835,0.198983416,0.9896322489,0.1034158543,-0.128158927,-0.1873940825,-0.1472871602,0.6541970968,-0.3806745112,-0.0115183806,-0.3637718558,-0.1058204174,0.0314450935,0.2072007805,-0.1031240225,0.3262476027,-0.3529521525,0.4330963492,-0.2215720117,0.2752585113,0.0378009826,0.2362699062,0.4365597963,0.0669674501,0.2220270634,0.0047120899,-0.1494491845,0.0237568822,0.035163898,-0.076216422,0.153446883,0.0169060118,-0.5261893272,-0.0714147314,-0.2255996019,-0.1951127946,0.4405925274,0.0815305263,0.0614727661,0.1002305523,0.4885941148,0.4361701906,-0.0634725392,0.0374194421,0.2201316804,0.0860925168,0.4140884578,0.1954234838,0.0417411439,-0.1117576584,-0.2063276917,-0.3461257815,0.204515934,-0.2436403334,-0.2923876345,-0.3536171913,0.4286381304,-0.184463948,0.0681486279,0.0913905054,0.2552635372,-0.1097519994,-0.0369252935,0.0021962279,-0.0902769864,0.3685131371,0.1963047832,-0.2931803167,0.0417205282,0.4069791436,-0.232502833,0.1254948527,0.5497884154,-0.1017915457,-0.2252684832,-0.2067078501,0.1969598979,-0.0603260361,-0.3254715204,0.1602658331,-0.1623904854,0.0118255047,-0.1835958958,0.0668832734,0.2456146181,-0.2072423995,-0.4763456881,-0.1797680855,0.1073337197,0.0839170516,0.0476500615,0.0586452298,-0.1267219335,0.1897962689,-0.1093968377,-0.0093376497,-0.158412993,0.120235078,0.021454433,0.0271577798,0.2310236394,-0.0305370428,-0.2976668477,-0.0744141862,0.0330074765,0.1023901254,0.1723152399,-0.0497026704,-0.1122442931,0.1763163954,-0.0285951477,0.0156684574,0.092694208,-0.016871566,-0.046250578,-0.1839940697,0.079182975,0.3121452332,0.0969770849,0.5788119435,0.5832945108,-0.033908356,-0.222071588,0.1607327759,0.0873374045,0.1841948032,0.0529938079,0.2401016653,0.5068276525,-0.0162094608,0.0280093066,0.1571312547,0.2266868204,0.4644916058,0.0047274544,-0.1873105615,-0.1078376323,-0.0579006039,0.0208873227,0.0725469515,-0.0402345955,-0.6182408333,0.3836396635,0.0868523791,-0.1440355927,0.0142529979,0.1585737765,0.1346081942,-0.1789777279,0.1089547798,0.2710599005,0.1026725397,0.1148479953,-0.1038566977,0.4935482442,0.0700599402,-0.1272036135,-0.0059617008,0.0814884454,-0.1780944914,0.0028196378,-0.4517574906,0.0024042593,0.1429587454,0.4093230665,0.0426885076,0.673463881,0.4398098588,0.0856782421,0.2413147092,-0.0056111896,0.2390040606,0.0125074573,0.0981615856,0.0005225074,0.3858040571,0.2122180611,-0.6277556419,-0.2471204698,0.1745300591,-0.1434087753,-0.4272122979,-0.1796654165,0.0361138582,-0.1521292329,0.0125042209,-0.1833848804,0.0802486241,0.0216519888,0.0369823426,0.0786463395,-0.0942126736,-0.4781661928,-0.1087018549,0.4069794416,-0.1498499662,0.4832811952,-0.1814756095,0.1454333365,0.1216650978,0.676325202,0.1954018772,0.186401546,-0.1615892947,-0.1933141798,-0.1266186982,0.0618981831,0.0646427795,0.3771719337,-0.1495801508,-0.2651422918,0.1260564774,-0.0423136503,0.0203797445,0.0936334878,-0.2254330814,0.0614366569,-0.1877754331,0.1608242691,-0.0641598329,-0.1141403466,0.0275460109,-0.1480630487,-0.2405239493,-0.151717931,0.0845284387,-0.2936899364,0.0058082659,-0.0316544026,0.0293737389,-0.2334011942,0.3396129012,0.154254809,0.0370926149,0.0378440432,-0.0779611692,-0.0145464893,0.0138079282,0.0658976585,-0.0732400045,-0.1268460155,0.0944130495,0.2512851357,0.0523795746,0.2203788608,0.5503807664,-0.4273731411,-0.0579757877,-0.4943718612,0.4398273528,0.2314485908,0.3380961716,0.1491912305,-0.1296543479,0.1890538037,-0.0348265171,-0.1559411138,0.2760892808,-0.0666887835,0.0218982808,-0.1562603563,0.5143120885,0.4250681102,0.1174677238,0.0720208809,0.1500927657,-0.086658068,-0.3402732015,-0.5251982808,0.1291481405,-0.0712801516,0.1653262675,-0.0385844186,0.1468282193,0.0343764797,-0.0710097253,0.0820452645,0.4577865005,-0.3480732143,0.0715712011,-0.0907173157,-0.14454633,-0.265804857,-0.1429156214,0.0634149313,0.0624262318,-0.2742138207,-0.0804677159,0.5503883362,-0.4103033543,-0.0559439547,-0.2348195016,0.2618363798,0.5103293657,-0.1085254923,-0.3054231703,-0.3657253385,0.5240724087,0.3138027489,-0.0678080693,0.347936213,0.0191900302,0.0052172206,0.0415776744,0.014119477,0.142597571,-0.0807207674,-0.0625645891,-0.2172840536]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1636","title":"winogrande cannot be dowloaded ","comments":"I have same issue for other datasets (`myanmar_news` in my case).\r\n\r\nA version of `datasets` runs correctly on my local machine (**without GPU**) which looking for the dataset at \r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/myanmar_news\/myanmar_news.py\r\n```\r\n\r\nMeanwhile, other version runs on Colab (**with GPU**) failed to download the dataset. It try to find the dataset at `1.1.3` instead of `master` . If I disable GPU on my Colab, the code can load the dataset without any problem.\r\n\r\nMaybe there is some version missmatch with the GPU and CPU version of code for these datasets?","body":"Hi,\r\nI am getting this error when trying to run the codes on the cloud. Thank you for any suggestion and help on this @lhoestq \r\n\r\n```\r\n File \".\/finetune_trainer.py\", line 318, in <module>\r\n main()\r\n File \".\/finetune_trainer.py\", line 148, in main\r\n for task in data_args.tasks]\r\n File \".\/finetune_trainer.py\", line 148, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 65, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 466, in load_dataset\r\n return datasets.load_dataset('winogrande', 'winogrande_l', split=split)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/winogrande\/winogrande.py\r\nyo\/0 I1224 14:17:46.419031 31226 main shadow.py:122 > Traceback (most recent call last):\r\n File \"\/usr\/lib\/python3.6\/runpy.py\", line 193, in _run_module_as_main\r\n \"__main__\", mod_spec)\r\n File \"\/usr\/lib\/python3.6\/runpy.py\", line 85, in _run_code\r\n exec(code, run_globals)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/distributed\/launch.py\", line 260, in <module>\r\n main()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/distributed\/launch.py\", line 256, in main\r\n cmd=cmd)\r\n```","comment_length":90,"text":"winogrande cannot be dowloaded \nHi,\r\nI am getting this error when trying to run the codes on the cloud. Thank you for any suggestion and help on this @lhoestq \r\n\r\n```\r\n File \".\/finetune_trainer.py\", line 318, in <module>\r\n main()\r\n File \".\/finetune_trainer.py\", line 148, in main\r\n for task in data_args.tasks]\r\n File \".\/finetune_trainer.py\", line 148, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 65, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 466, in load_dataset\r\n return datasets.load_dataset('winogrande', 'winogrande_l', split=split)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/winogrande\/winogrande.py\r\nyo\/0 I1224 14:17:46.419031 31226 main shadow.py:122 > Traceback (most recent call last):\r\n File \"\/usr\/lib\/python3.6\/runpy.py\", line 193, in _run_module_as_main\r\n \"__main__\", mod_spec)\r\n File \"\/usr\/lib\/python3.6\/runpy.py\", line 85, in _run_code\r\n exec(code, run_globals)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/distributed\/launch.py\", line 260, in <module>\r\n main()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/distributed\/launch.py\", line 256, in main\r\n cmd=cmd)\r\n```\nI have same issue for other datasets (`myanmar_news` in my case).\r\n\r\nA version of `datasets` runs correctly on my local machine (**without GPU**) which looking for the dataset at \r\n```\r\nhttps:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/myanmar_news\/myanmar_news.py\r\n```\r\n\r\nMeanwhile, other version runs on Colab (**with GPU**) failed to download the dataset. It try to find the dataset at `1.1.3` instead of `master` . If I disable GPU on my Colab, the code can load the dataset without any problem.\r\n\r\nMaybe there is some version missmatch with the GPU and CPU version of code for these datasets?","embeddings":[-0.368696928,0.1048511565,-0.0660381317,0.1727429926,0.2931744158,0.0646404102,0.6664779782,0.0742833689,0.32062307,0.0376309268,-0.1209185869,0.0811130106,-0.0274400972,0.3085554242,0.0915508866,-0.1352519244,-0.1551791877,0.0680280253,-0.3883652091,-0.1035755351,-0.3014270067,0.1449964643,-0.3381091654,0.1980768144,-0.1758435071,-0.3244913816,-0.0872360617,0.2303459346,-0.4127731025,-0.1883580089,0.3303743899,-0.1787249595,0.2364586294,0.1999069154,-0.0001119528,0.0268156249,0.2326923311,-0.0847722366,-0.1867612749,-0.1960292011,-0.0253983475,0.3051585853,0.2187183499,-0.184618175,-0.0229243319,0.1664848775,-0.1624972969,0.0773649812,0.3128141165,0.5773444772,0.1898547262,0.3975589275,0.2045686692,0.0331234634,-0.4235072732,-0.3424312174,0.0212687682,0.6150701642,0.1908356696,-0.4338017404,0.2548056841,0.0014486369,0.0207589474,0.1185219586,0.0675943941,-0.1765601337,0.2509324253,-0.419539243,-0.1311502308,-0.0256385114,0.5311209559,0.0215759054,-0.1296834201,0.2047387362,0.1120898128,-0.2739987075,0.284229666,0.1477161348,-0.1760560274,0.1649956256,-0.3080173731,-0.0580678545,-0.3582025766,0.3218635321,0.1481005102,0.1265269965,0.0242350139,0.0446994193,0.2552797198,0.0882350504,0.1398637444,0.0488179848,0.0544819944,0.0351311862,-0.4112992287,-0.0904320851,-0.2612808347,0.2552453876,-0.3865052164,0.1501704752,0.1091204733,-0.1255415678,0.4431173503,0.1739905626,0.1946264058,0.3714409769,-0.2129581422,0.2366685122,0.0951329246,0.0697603747,0.1420715451,0.082355991,-0.6819787025,-0.5539950728,-0.1986275911,0.1179601252,0.1390322596,-0.2845466435,-0.2745520771,-0.0962962508,0.3059588969,0.0497980826,0.3026941717,0.2752325535,-0.0745164827,-0.2629697025,0.277310878,0.2931385636,-0.1387893111,-0.0831434205,-0.0040009543,0.2816756666,-0.1500656903,-0.1428893209,0.2244484723,0.1805662811,0.3548044562,-0.0919252411,0.3388037384,-0.2103509307,0.0394848287,-0.2153775543,-0.2308950871,0.0822293311,-0.1234867275,-0.1133310646,0.3215297163,0.0159397442,-0.0242304523,0.1006711274,-0.3189349174,-0.5324395895,-0.148290664,0.1527325511,0.0759669244,-0.1154205725,-0.6399577856,-0.1232644543,0.0951323062,-0.046619948,0.0554771423,-0.4480449557,-0.0278314501,-0.097169362,0.2478317022,0.4105527699,-0.3679771721,-0.1377865523,0.0975870043,-0.1613768488,-0.0650710016,0.1746538728,-0.3203060925,0.4273764491,-0.4233532846,-0.1221769452,0.4045239985,-0.6057487726,-0.3403845727,0.3660116792,-0.1654238105,-0.1204435304,-0.054276295,-0.1046243161,0.4172097147,-0.1382526308,0.3981581926,0.4014757574,0.0502186231,-0.1129810363,-0.103532888,-0.2852769196,-0.0202987473,0.3621865809,0.2654350102,0.1293463856,0.0755940005,-0.0660735965,0.060531009,0.0940140858,-0.2598439157,0.3106851578,-0.0612095036,0.1996018887,-0.0799068734,0.2906145453,-0.1260471642,0.1627796441,-0.0035591011,-0.0161473919,0.042138584,0.1411819458,-0.1098405421,-0.1448151171,-0.3546047807,-0.108858645,0.0833577812,-0.0719759688,0.3594585955,-0.1155302525,-0.1014227644,0.2762686014,-0.2571201921,-0.0231026448,-0.0483524762,0.1581780165,-0.2223320305,-0.2964906693,0.1315133125,-0.0264314245,0.2973249257,-0.1454710066,-0.3102459013,0.3024873137,-0.2499793172,0.1710686535,-0.0369232446,0.2473673075,0.1887207776,-0.1569518745,0.1100813076,0.3358827531,0.1203901619,-0.0491737686,0.5043795109,0.1778320223,-0.2820903063,0.3455861211,0.0722008571,0.1410001516,-0.0977449194,0.0539246649,-0.0214269403,0.1489199549,0.326510191,0.069836095,0.3296034038,-0.2740300596,-0.081678547,-0.0127635589,0.1321129948,-0.0582861416,0.1922532618,0.0241464414,-0.058877904,0.1225888729,-0.138240248,-0.3534490764,0.6288217902,-0.0779441819,-0.1627549827,0.2168244272,0.0880993605,-0.3022451401,0.4201553464,0.1474983692,0.2916156054,0.1335139573,-0.1542398781,-0.2613516152,-0.2166075557,-0.251067549,0.1798330992,0.0996108577,-0.2337482125,0.1668332368,-0.2003485709,0.0094736833,-0.2737947404,-0.2217485309,0.0794080794,0.1521431506,0.1278292239,0.3142968118,0.1968609989,0.1366688311,-0.0264539234,0.3919760585,0.0437615626,-0.1345431209,-0.1614870876,-0.2623594999,-0.1333007216,0.0577557385,0.1983505785,-0.4033934176,0.3068819344,-0.1677616835,-0.1851496696,0.0981961116,0.1332039982,0.0097874077,0.110046871,0.6873757243,-0.0991813987,0.3886087835,-0.1293687224,0.0391853452,0.2474743873,-0.1137061566,-0.0018731067,-0.2590671182,0.1835538596,-0.0399589501,0.0559465811,-0.4450828731,-0.6970521212,-0.2890126407,-0.0726742893,-0.2989792824,0.28790766,-0.134353295,0.1210230514,0.0026091747,0.2019261122,0.0667655617,-0.1324052066,-0.4622768164,0.2771128118,-0.3519118428,-0.3357568979,0.1141091138,0.395082891,0.2223625481,-0.4351142049,-0.2844531834,-0.3437049687,0.0159265492,0.0304134078,-0.1782898754,-0.1346721351,0.3508553803,-0.1107115448,-0.03814926,-0.0219648555,0.024923183,0.0567572452,-0.0207183138,0.2247359157,0.1840579212,0.3671400547,-0.1862891912,0.6083549261,0.0779509991,-0.1369999945,0.1451140046,-0.1870760173,-0.0541699491,-0.0692810565,-0.4518500268,0.4742793739,-0.1131831333,-0.0614738353,0.2452813834,-0.1743471771,0.2207771987,-0.3289113641,-0.1733208299,-0.4900580049,-0.0917509943,-0.1131490022,0.0208363403,0.002460879,0.0226041563,0.1967402399,0.1265710294,-0.0200672112,0.2225946039,0.5301418304,0.0282060541,0.2667176127,0.591996789,-0.1696763337,-0.0755533278,0.275557816,-0.117418386,0.3571249843,-0.4477061331,0.1704387963,0.0834037215,0.1157481149,0.6483278871,-0.1920786947,0.2554333806,0.0080347247,-0.1107422933,0.0455936119,-0.3255800903,-0.2942346632,-0.0918782949,0.4250554442,0.2422790229,-0.0626886114,0.1446256489,0.11392387,0.1978151947,-0.3709716201,-0.0854698569,-0.5120494366,-0.3674280345,-0.3352569938,0.3530168831,-0.0670926943,0.3027692735,-0.1915315688,0.1227203533,-0.0513175838,0.1204482242,-0.002483753,-0.0087142428,0.1735261679,-0.063360557,0.0260047521,0.0835358948,0.2604655027,0.271879971,0.3621813953,0.2598895133,-0.1369645596,0.139536202,-0.0141690718,0.0902519003,0.1537485421,-0.2411435694,-0.320723474,0.1457295418,0.1455246806,-0.1394383013,0.087427929,0.2518631518,-0.0247282498,-0.2617937922,-0.2276068032,0.114930518,-0.1276950687,-0.0375007913,0.3017038107,-0.1389856786,-0.151504606,-0.08415512,0.1552571654,0.9107421041,-0.2133566737,-0.1285695285,0.3481628299,-0.2790898085,0.0201086234,-0.1385186166,0.3538869023,-0.2007005364,-0.0924270526,0.0408170633,-0.1368802339,0.2002818137,0.1912966222,-0.2707364559,0.4681877196,-0.0015254169,0.1334586143,0.1498990506,0.040440809,0.2626670003,-0.0780851692,-0.1359547526,0.084515065,-0.1086390391,0.3842449784,-0.5040246248,-0.2170932591,0.2379847765,-0.4305113852,-0.1765408218,0.451271832,-0.5486161709,0.325914681,-0.3944927752,-0.2378854156,-0.0847825557,0.20163396,0.1886088848,0.0921094641,-0.0877842084,0.165805459,0.1182443351,-0.2723978162,-0.15852198,0.0094331466,-0.0041274941,-0.2431104034,-0.4131644368,0.0712870881,0.0180768184,-0.2828376293,-0.2171703726,0.0145923756,0.1336223036,-0.0166084711,-0.1656593382,-0.216467917,-0.0426400602,-0.2078529,0.102247335,-0.2015220225,-0.1709569395,0.2988410294,0.0344409011,-0.2984281778,-0.0502208322,0.5843225121,-0.1991005689,-0.0769307688,0.5926619768,0.3001182377,-0.1717824638,-0.3273946345,-0.1680516452,-0.0253276732,-0.3014725447,0.1512160301,0.2889396846,0.2211104482,-0.0231381971,0.3339020908,0.0578122921,0.0522143431,0.0001322072,-0.506719172,-0.1778177023,0.0708361045,0.1334398538,-0.0263048187,0.1574452072,-0.0285426602,0.1735300124,-0.1489845514,-0.2994318306,-0.0567264706,-0.2236979455,-0.0013177,0.2572506666,-0.0540261902,0.2678607702,0.3288381398,0.1292611808,-0.1097879559,0.1241292432,-0.2014394999,-0.0468051285,0.1505444348,0.0206485409,-0.1661668867,-0.2040661871,-0.3078506291,-0.1011749506,0.1202470511,0.1411782801,-0.0410150662,0.0667842925,-0.0007048233,-0.4208702147,0.1353564858,-0.0623972714,-0.0482190922,-0.1271520853,0.3525931239,-0.14984034,0.0040281876,0.2936516106,-0.035684675,-0.2039978355,-0.2254557014,0.0852013007,0.1038977206,0.1968708187,-0.3131240904,0.1954004616,-0.2847557962,0.0798463449,0.3708586693,-0.2202080637,-0.1123923808,0.1443208456,0.1717065424,-0.261895895,-0.0288327895,-0.2897729576,0.0112013826,-0.0316609703,0.3273953497,-0.0662682652,-0.1877393275,-0.1899750233,-0.0890927911,0.3179430366,0.1313030869,-0.0474189781,0.4649277031,0.0435289294,-0.0652489886,-0.1027757749,0.1915580779,0.1136752889,0.1237245873,-0.7052069306,-0.0146799432,-0.1433242708,0.1492024511,-0.2080516964,-0.2169017494,-0.0615348481,0.0089414967,0.195224911,0.0333750732,-0.242160216,-0.2797775269,-0.2829131484,-0.1272009909,-0.4734354615,0.2137410343,-0.173505798,0.0427742936,-0.2581561506,-0.0992375687,0.0819458961,0.2739166617,0.1573874205,-0.1379285157,0.211448282,0.1874432564,-0.1046094596,-0.1669309139,-0.0586813651,0.3726677001,-0.1559056789,-0.001970361,0.4480982423,0.3400500417,0.2911922336,0.1144417971,0.552649796,0.4909820855,0.4593272507,-0.085553959,0.2298236191,-0.0359859094,-0.0366633721,0.0817338079,0.3253512681,0.233419925,0.2773746848,0.1695165038,0.156104058,-0.0981179178,0.4486888945,-0.1526051164,-0.1355650872,0.0243609715,0.2683689296,0.225122273,0.1164354384,-0.3357058764,0.0403325632,-0.6587376595,0.1144017205,0.2917356789,0.255114913,0.0153100286,-0.0458030142,0.0590034239,-0.1430252194,0.5999675393,0.5157787204,0.4286439121,-0.5772804022,0.0536262318,-0.3934624791,0.1048484892,0.0279974565,-0.2297892272,-0.006612326,-0.0525765866,-0.1068507582,0.1597436368,0.3771780729,-0.1941118389,0.0713170022,0.3440300524,-0.1549400538,0.1846998036,0.0904052407,0.4121747911,0.2145600021,-0.4363020658,0.1314442009,-0.1117016971,0.0171968024,-0.0213399921,0.1246045604,-0.2903750539,-0.3366801143,0.5827984214,0.0390419066,0.5013961792,0.0446205847,0.127710551,-0.0943086296,-0.021208765,-0.1863379031,0.3635920584,0.2075771838,-0.1896802783,-0.1403339356,-0.057942789,-0.5768238306,0.1074187458,-0.0909880102,-0.2019696832,-0.1147235706,-0.2546962202,0.0155286239,0.0891181231,0.1202306896,0.057911884,-0.0667156056,0.1066170335,-0.4519064128,-0.1203515977,0.5304166079,-0.4497497976,-0.1057529971,-0.1027746648,0.3755018413,0.1887564361,0.1216079667,-0.4843398929,0.3328354359,0.1578608602,0.1881427914,-0.068051368,0.2138949931,-0.2201521993,0.2617906034,-0.0151782334,0.0882182568,-0.2545225918,-0.0324254818,-0.205749616,-0.1825577319]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1636","title":"winogrande cannot be dowloaded ","comments":"It looks like they're two different issues\r\n\r\n----------\r\n\r\nFirst for `myanmar_news`: \r\n\r\nIt must come from the way you installed `datasets`.\r\nIf you install `datasets` from source, then the `myanmar_news` script will be loaded from `master`.\r\nHowever if you install from `pip` it will get it using the version of the lib (here `1.1.3`) and `myanmar_news` is not available in `1.1.3`.\r\n\r\nThe difference between your GPU and CPU executions must be the environment, one seems to have installed `datasets` from source and not the other.\r\n\r\n----------\r\n\r\nThen for `winogrande`:\r\n\r\nThe errors says that the url https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/winogrande\/winogrande.py is not reachable.\r\nHowever it works fine on my side.\r\n\r\nDoes your machine have an internet connection ? Are connections to github blocked by some sort of proxy ?\r\nCan you also try again in case github had issues when you tried the first time ?\r\n","body":"Hi,\r\nI am getting this error when trying to run the codes on the cloud. Thank you for any suggestion and help on this @lhoestq \r\n\r\n```\r\n File \".\/finetune_trainer.py\", line 318, in <module>\r\n main()\r\n File \".\/finetune_trainer.py\", line 148, in main\r\n for task in data_args.tasks]\r\n File \".\/finetune_trainer.py\", line 148, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 65, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 466, in load_dataset\r\n return datasets.load_dataset('winogrande', 'winogrande_l', split=split)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/winogrande\/winogrande.py\r\nyo\/0 I1224 14:17:46.419031 31226 main shadow.py:122 > Traceback (most recent call last):\r\n File \"\/usr\/lib\/python3.6\/runpy.py\", line 193, in _run_module_as_main\r\n \"__main__\", mod_spec)\r\n File \"\/usr\/lib\/python3.6\/runpy.py\", line 85, in _run_code\r\n exec(code, run_globals)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/distributed\/launch.py\", line 260, in <module>\r\n main()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/distributed\/launch.py\", line 256, in main\r\n cmd=cmd)\r\n```","comment_length":141,"text":"winogrande cannot be dowloaded \nHi,\r\nI am getting this error when trying to run the codes on the cloud. Thank you for any suggestion and help on this @lhoestq \r\n\r\n```\r\n File \".\/finetune_trainer.py\", line 318, in <module>\r\n main()\r\n File \".\/finetune_trainer.py\", line 148, in main\r\n for task in data_args.tasks]\r\n File \".\/finetune_trainer.py\", line 148, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 65, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 466, in load_dataset\r\n return datasets.load_dataset('winogrande', 'winogrande_l', split=split)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/winogrande\/winogrande.py\r\nyo\/0 I1224 14:17:46.419031 31226 main shadow.py:122 > Traceback (most recent call last):\r\n File \"\/usr\/lib\/python3.6\/runpy.py\", line 193, in _run_module_as_main\r\n \"__main__\", mod_spec)\r\n File \"\/usr\/lib\/python3.6\/runpy.py\", line 85, in _run_code\r\n exec(code, run_globals)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/distributed\/launch.py\", line 260, in <module>\r\n main()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/distributed\/launch.py\", line 256, in main\r\n cmd=cmd)\r\n```\nIt looks like they're two different issues\r\n\r\n----------\r\n\r\nFirst for `myanmar_news`: \r\n\r\nIt must come from the way you installed `datasets`.\r\nIf you install `datasets` from source, then the `myanmar_news` script will be loaded from `master`.\r\nHowever if you install from `pip` it will get it using the version of the lib (here `1.1.3`) and `myanmar_news` is not available in `1.1.3`.\r\n\r\nThe difference between your GPU and CPU executions must be the environment, one seems to have installed `datasets` from source and not the other.\r\n\r\n----------\r\n\r\nThen for `winogrande`:\r\n\r\nThe errors says that the url https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/winogrande\/winogrande.py is not reachable.\r\nHowever it works fine on my side.\r\n\r\nDoes your machine have an internet connection ? Are connections to github blocked by some sort of proxy ?\r\nCan you also try again in case github had issues when you tried the first time ?\r\n","embeddings":[-0.368696928,0.1048511565,-0.0660381317,0.1727429926,0.2931744158,0.0646404102,0.6664779782,0.0742833689,0.32062307,0.0376309268,-0.1209185869,0.0811130106,-0.0274400972,0.3085554242,0.0915508866,-0.1352519244,-0.1551791877,0.0680280253,-0.3883652091,-0.1035755351,-0.3014270067,0.1449964643,-0.3381091654,0.1980768144,-0.1758435071,-0.3244913816,-0.0872360617,0.2303459346,-0.4127731025,-0.1883580089,0.3303743899,-0.1787249595,0.2364586294,0.1999069154,-0.0001119528,0.0268156249,0.2326923311,-0.0847722366,-0.1867612749,-0.1960292011,-0.0253983475,0.3051585853,0.2187183499,-0.184618175,-0.0229243319,0.1664848775,-0.1624972969,0.0773649812,0.3128141165,0.5773444772,0.1898547262,0.3975589275,0.2045686692,0.0331234634,-0.4235072732,-0.3424312174,0.0212687682,0.6150701642,0.1908356696,-0.4338017404,0.2548056841,0.0014486369,0.0207589474,0.1185219586,0.0675943941,-0.1765601337,0.2509324253,-0.419539243,-0.1311502308,-0.0256385114,0.5311209559,0.0215759054,-0.1296834201,0.2047387362,0.1120898128,-0.2739987075,0.284229666,0.1477161348,-0.1760560274,0.1649956256,-0.3080173731,-0.0580678545,-0.3582025766,0.3218635321,0.1481005102,0.1265269965,0.0242350139,0.0446994193,0.2552797198,0.0882350504,0.1398637444,0.0488179848,0.0544819944,0.0351311862,-0.4112992287,-0.0904320851,-0.2612808347,0.2552453876,-0.3865052164,0.1501704752,0.1091204733,-0.1255415678,0.4431173503,0.1739905626,0.1946264058,0.3714409769,-0.2129581422,0.2366685122,0.0951329246,0.0697603747,0.1420715451,0.082355991,-0.6819787025,-0.5539950728,-0.1986275911,0.1179601252,0.1390322596,-0.2845466435,-0.2745520771,-0.0962962508,0.3059588969,0.0497980826,0.3026941717,0.2752325535,-0.0745164827,-0.2629697025,0.277310878,0.2931385636,-0.1387893111,-0.0831434205,-0.0040009543,0.2816756666,-0.1500656903,-0.1428893209,0.2244484723,0.1805662811,0.3548044562,-0.0919252411,0.3388037384,-0.2103509307,0.0394848287,-0.2153775543,-0.2308950871,0.0822293311,-0.1234867275,-0.1133310646,0.3215297163,0.0159397442,-0.0242304523,0.1006711274,-0.3189349174,-0.5324395895,-0.148290664,0.1527325511,0.0759669244,-0.1154205725,-0.6399577856,-0.1232644543,0.0951323062,-0.046619948,0.0554771423,-0.4480449557,-0.0278314501,-0.097169362,0.2478317022,0.4105527699,-0.3679771721,-0.1377865523,0.0975870043,-0.1613768488,-0.0650710016,0.1746538728,-0.3203060925,0.4273764491,-0.4233532846,-0.1221769452,0.4045239985,-0.6057487726,-0.3403845727,0.3660116792,-0.1654238105,-0.1204435304,-0.054276295,-0.1046243161,0.4172097147,-0.1382526308,0.3981581926,0.4014757574,0.0502186231,-0.1129810363,-0.103532888,-0.2852769196,-0.0202987473,0.3621865809,0.2654350102,0.1293463856,0.0755940005,-0.0660735965,0.060531009,0.0940140858,-0.2598439157,0.3106851578,-0.0612095036,0.1996018887,-0.0799068734,0.2906145453,-0.1260471642,0.1627796441,-0.0035591011,-0.0161473919,0.042138584,0.1411819458,-0.1098405421,-0.1448151171,-0.3546047807,-0.108858645,0.0833577812,-0.0719759688,0.3594585955,-0.1155302525,-0.1014227644,0.2762686014,-0.2571201921,-0.0231026448,-0.0483524762,0.1581780165,-0.2223320305,-0.2964906693,0.1315133125,-0.0264314245,0.2973249257,-0.1454710066,-0.3102459013,0.3024873137,-0.2499793172,0.1710686535,-0.0369232446,0.2473673075,0.1887207776,-0.1569518745,0.1100813076,0.3358827531,0.1203901619,-0.0491737686,0.5043795109,0.1778320223,-0.2820903063,0.3455861211,0.0722008571,0.1410001516,-0.0977449194,0.0539246649,-0.0214269403,0.1489199549,0.326510191,0.069836095,0.3296034038,-0.2740300596,-0.081678547,-0.0127635589,0.1321129948,-0.0582861416,0.1922532618,0.0241464414,-0.058877904,0.1225888729,-0.138240248,-0.3534490764,0.6288217902,-0.0779441819,-0.1627549827,0.2168244272,0.0880993605,-0.3022451401,0.4201553464,0.1474983692,0.2916156054,0.1335139573,-0.1542398781,-0.2613516152,-0.2166075557,-0.251067549,0.1798330992,0.0996108577,-0.2337482125,0.1668332368,-0.2003485709,0.0094736833,-0.2737947404,-0.2217485309,0.0794080794,0.1521431506,0.1278292239,0.3142968118,0.1968609989,0.1366688311,-0.0264539234,0.3919760585,0.0437615626,-0.1345431209,-0.1614870876,-0.2623594999,-0.1333007216,0.0577557385,0.1983505785,-0.4033934176,0.3068819344,-0.1677616835,-0.1851496696,0.0981961116,0.1332039982,0.0097874077,0.110046871,0.6873757243,-0.0991813987,0.3886087835,-0.1293687224,0.0391853452,0.2474743873,-0.1137061566,-0.0018731067,-0.2590671182,0.1835538596,-0.0399589501,0.0559465811,-0.4450828731,-0.6970521212,-0.2890126407,-0.0726742893,-0.2989792824,0.28790766,-0.134353295,0.1210230514,0.0026091747,0.2019261122,0.0667655617,-0.1324052066,-0.4622768164,0.2771128118,-0.3519118428,-0.3357568979,0.1141091138,0.395082891,0.2223625481,-0.4351142049,-0.2844531834,-0.3437049687,0.0159265492,0.0304134078,-0.1782898754,-0.1346721351,0.3508553803,-0.1107115448,-0.03814926,-0.0219648555,0.024923183,0.0567572452,-0.0207183138,0.2247359157,0.1840579212,0.3671400547,-0.1862891912,0.6083549261,0.0779509991,-0.1369999945,0.1451140046,-0.1870760173,-0.0541699491,-0.0692810565,-0.4518500268,0.4742793739,-0.1131831333,-0.0614738353,0.2452813834,-0.1743471771,0.2207771987,-0.3289113641,-0.1733208299,-0.4900580049,-0.0917509943,-0.1131490022,0.0208363403,0.002460879,0.0226041563,0.1967402399,0.1265710294,-0.0200672112,0.2225946039,0.5301418304,0.0282060541,0.2667176127,0.591996789,-0.1696763337,-0.0755533278,0.275557816,-0.117418386,0.3571249843,-0.4477061331,0.1704387963,0.0834037215,0.1157481149,0.6483278871,-0.1920786947,0.2554333806,0.0080347247,-0.1107422933,0.0455936119,-0.3255800903,-0.2942346632,-0.0918782949,0.4250554442,0.2422790229,-0.0626886114,0.1446256489,0.11392387,0.1978151947,-0.3709716201,-0.0854698569,-0.5120494366,-0.3674280345,-0.3352569938,0.3530168831,-0.0670926943,0.3027692735,-0.1915315688,0.1227203533,-0.0513175838,0.1204482242,-0.002483753,-0.0087142428,0.1735261679,-0.063360557,0.0260047521,0.0835358948,0.2604655027,0.271879971,0.3621813953,0.2598895133,-0.1369645596,0.139536202,-0.0141690718,0.0902519003,0.1537485421,-0.2411435694,-0.320723474,0.1457295418,0.1455246806,-0.1394383013,0.087427929,0.2518631518,-0.0247282498,-0.2617937922,-0.2276068032,0.114930518,-0.1276950687,-0.0375007913,0.3017038107,-0.1389856786,-0.151504606,-0.08415512,0.1552571654,0.9107421041,-0.2133566737,-0.1285695285,0.3481628299,-0.2790898085,0.0201086234,-0.1385186166,0.3538869023,-0.2007005364,-0.0924270526,0.0408170633,-0.1368802339,0.2002818137,0.1912966222,-0.2707364559,0.4681877196,-0.0015254169,0.1334586143,0.1498990506,0.040440809,0.2626670003,-0.0780851692,-0.1359547526,0.084515065,-0.1086390391,0.3842449784,-0.5040246248,-0.2170932591,0.2379847765,-0.4305113852,-0.1765408218,0.451271832,-0.5486161709,0.325914681,-0.3944927752,-0.2378854156,-0.0847825557,0.20163396,0.1886088848,0.0921094641,-0.0877842084,0.165805459,0.1182443351,-0.2723978162,-0.15852198,0.0094331466,-0.0041274941,-0.2431104034,-0.4131644368,0.0712870881,0.0180768184,-0.2828376293,-0.2171703726,0.0145923756,0.1336223036,-0.0166084711,-0.1656593382,-0.216467917,-0.0426400602,-0.2078529,0.102247335,-0.2015220225,-0.1709569395,0.2988410294,0.0344409011,-0.2984281778,-0.0502208322,0.5843225121,-0.1991005689,-0.0769307688,0.5926619768,0.3001182377,-0.1717824638,-0.3273946345,-0.1680516452,-0.0253276732,-0.3014725447,0.1512160301,0.2889396846,0.2211104482,-0.0231381971,0.3339020908,0.0578122921,0.0522143431,0.0001322072,-0.506719172,-0.1778177023,0.0708361045,0.1334398538,-0.0263048187,0.1574452072,-0.0285426602,0.1735300124,-0.1489845514,-0.2994318306,-0.0567264706,-0.2236979455,-0.0013177,0.2572506666,-0.0540261902,0.2678607702,0.3288381398,0.1292611808,-0.1097879559,0.1241292432,-0.2014394999,-0.0468051285,0.1505444348,0.0206485409,-0.1661668867,-0.2040661871,-0.3078506291,-0.1011749506,0.1202470511,0.1411782801,-0.0410150662,0.0667842925,-0.0007048233,-0.4208702147,0.1353564858,-0.0623972714,-0.0482190922,-0.1271520853,0.3525931239,-0.14984034,0.0040281876,0.2936516106,-0.035684675,-0.2039978355,-0.2254557014,0.0852013007,0.1038977206,0.1968708187,-0.3131240904,0.1954004616,-0.2847557962,0.0798463449,0.3708586693,-0.2202080637,-0.1123923808,0.1443208456,0.1717065424,-0.261895895,-0.0288327895,-0.2897729576,0.0112013826,-0.0316609703,0.3273953497,-0.0662682652,-0.1877393275,-0.1899750233,-0.0890927911,0.3179430366,0.1313030869,-0.0474189781,0.4649277031,0.0435289294,-0.0652489886,-0.1027757749,0.1915580779,0.1136752889,0.1237245873,-0.7052069306,-0.0146799432,-0.1433242708,0.1492024511,-0.2080516964,-0.2169017494,-0.0615348481,0.0089414967,0.195224911,0.0333750732,-0.242160216,-0.2797775269,-0.2829131484,-0.1272009909,-0.4734354615,0.2137410343,-0.173505798,0.0427742936,-0.2581561506,-0.0992375687,0.0819458961,0.2739166617,0.1573874205,-0.1379285157,0.211448282,0.1874432564,-0.1046094596,-0.1669309139,-0.0586813651,0.3726677001,-0.1559056789,-0.001970361,0.4480982423,0.3400500417,0.2911922336,0.1144417971,0.552649796,0.4909820855,0.4593272507,-0.085553959,0.2298236191,-0.0359859094,-0.0366633721,0.0817338079,0.3253512681,0.233419925,0.2773746848,0.1695165038,0.156104058,-0.0981179178,0.4486888945,-0.1526051164,-0.1355650872,0.0243609715,0.2683689296,0.225122273,0.1164354384,-0.3357058764,0.0403325632,-0.6587376595,0.1144017205,0.2917356789,0.255114913,0.0153100286,-0.0458030142,0.0590034239,-0.1430252194,0.5999675393,0.5157787204,0.4286439121,-0.5772804022,0.0536262318,-0.3934624791,0.1048484892,0.0279974565,-0.2297892272,-0.006612326,-0.0525765866,-0.1068507582,0.1597436368,0.3771780729,-0.1941118389,0.0713170022,0.3440300524,-0.1549400538,0.1846998036,0.0904052407,0.4121747911,0.2145600021,-0.4363020658,0.1314442009,-0.1117016971,0.0171968024,-0.0213399921,0.1246045604,-0.2903750539,-0.3366801143,0.5827984214,0.0390419066,0.5013961792,0.0446205847,0.127710551,-0.0943086296,-0.021208765,-0.1863379031,0.3635920584,0.2075771838,-0.1896802783,-0.1403339356,-0.057942789,-0.5768238306,0.1074187458,-0.0909880102,-0.2019696832,-0.1147235706,-0.2546962202,0.0155286239,0.0891181231,0.1202306896,0.057911884,-0.0667156056,0.1066170335,-0.4519064128,-0.1203515977,0.5304166079,-0.4497497976,-0.1057529971,-0.1027746648,0.3755018413,0.1887564361,0.1216079667,-0.4843398929,0.3328354359,0.1578608602,0.1881427914,-0.068051368,0.2138949931,-0.2201521993,0.2617906034,-0.0151782334,0.0882182568,-0.2545225918,-0.0324254818,-0.205749616,-0.1825577319]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1634","title":"Inspecting datasets per category","comments":"That's interesting, can you tell me what you think would be useful to access to inspect a dataset?\r\n\r\nYou can filter them in the hub with the search by the way: https:\/\/huggingface.co\/datasets have you seen it?","body":"Hi\r\nIs there a way I could get all NLI datasets\/all QA datasets to get some understanding of available datasets per category? this is hard for me to inspect the datasets one by one in the webpage, thanks for the suggestions @lhoestq ","comment_length":36,"text":"Inspecting datasets per category\nHi\r\nIs there a way I could get all NLI datasets\/all QA datasets to get some understanding of available datasets per category? this is hard for me to inspect the datasets one by one in the webpage, thanks for the suggestions @lhoestq \nThat's interesting, can you tell me what you think would be useful to access to inspect a dataset?\r\n\r\nYou can filter them in the hub with the search by the way: https:\/\/huggingface.co\/datasets have you seen it?","embeddings":[-0.2455319911,-0.1529337466,-0.2294568419,0.3999654353,0.0459513701,0.2248052955,0.0731056929,0.4707452953,-0.0032283345,-0.1497346461,-0.5485332608,-0.1118844971,-0.1011159196,0.3258757293,0.0880427063,0.1338709444,-0.0260168537,0.0386467539,-0.0946159959,0.0354207307,-0.0761443749,-0.1488868296,-0.2528313398,0.239694044,0.0124907149,-0.0431006625,-0.3931308985,-0.1143400446,-0.2368001491,-0.3348473608,0.1663315743,0.395935297,0.3783767521,0.0771230236,-0.000107972,-0.0072599016,0.0652775094,0.046406921,-0.1809976399,0.1047304571,-0.6043121815,-0.2638504207,0.1473354548,-0.1237352937,0.0415516011,-0.0603904389,0.4771701992,-0.2552815676,-0.2283760458,-0.0516712368,0.1873449236,0.2058766037,-0.4735043347,0.078137584,-0.1382701844,0.2351609468,-0.198034063,-0.0474467836,0.5839135647,-0.0409656614,-0.177018851,0.3836557269,0.1003370807,0.0608108565,0.1391247511,-0.0921693072,-0.0813433826,-0.4509172738,0.0876605883,0.5685508251,0.5253902078,-0.0804104581,-0.4150145352,-0.2234537899,-0.184940502,0.1121361852,-0.4191016257,0.850507617,-0.0754681379,0.1847095788,-0.1760089695,-0.0661096498,-0.2954183519,0.0319227092,-0.1430185437,0.2825137973,-0.0627331883,-0.1289217621,0.2206001282,-0.0508234985,0.3368180692,-0.0682560429,-0.3944413364,0.1516827345,-0.6407146454,-0.2385390848,0.3341816962,0.6969444156,0.1839789003,0.175811708,0.2753145993,0.2173420489,-0.230701074,-0.1931372434,0.461650908,0.0655522794,0.2504457533,0.0461897664,0.303167522,-0.0581082106,-0.0202748384,0.0779221654,-0.1189201623,0.170808807,-0.2324444503,-0.0415991992,0.1922338754,-0.1545951813,-0.1994924843,-0.3903557658,-0.1095198095,-0.4240267277,0.0493930168,0.3936800659,0.1353563666,-0.0321975946,-0.3744388223,0.2982469797,-0.1149066612,-0.5472370386,-0.1787673533,0.035287071,0.0234185643,-0.0174971763,0.1506324261,-0.1031354144,-0.0708575845,-0.1618688554,0.0054196925,0.0382021777,-0.029030351,-0.0673473254,0.0550171286,0.0873191357,0.312037617,-0.1972889751,0.0786146894,0.2358674854,-0.4207357764,-0.1280501336,-0.0361292511,-0.4050108194,-0.0450908951,0.1007256955,-0.3277039528,-0.2192318887,-0.2970399559,0.611412406,-0.4679879248,0.2038843781,0.0732191131,0.314232111,0.283782959,-0.1606057286,-0.1994332671,0.4373907149,-0.1983653009,-0.0861080885,-0.0749574229,-0.247760281,-0.104261741,0.1070460677,-0.1139393598,0.2101263404,-0.3110682964,0.0409238376,0.8389814496,-0.2249023318,-0.3771826029,0.1156929955,0.3103725612,-0.1505962312,0.1415734738,0.3947982192,-0.0614907369,-0.010832048,-0.2910554707,0.2192099243,-0.0396326594,-0.0137523348,0.0243359152,-0.315728128,0.1325859278,0.1902198046,0.0364495404,0.2422332019,0.2564565241,-0.1025398523,0.4131573439,-0.114271149,-0.1415675431,-0.0168976523,0.1025638655,-0.1194425002,0.0969595686,-0.300057292,-0.3621153831,0.1579020172,-0.0509960502,0.1799060702,0.0329711325,-0.2369339913,0.0066283979,0.0663223639,-0.1448626816,-0.0071046934,0.2586657405,0.178842634,0.0613704473,0.1345415562,-0.2937826514,0.2766747475,0.0193650555,0.0965427384,-0.1030542329,0.3132304847,0.1412937194,0.3934921324,0.2139265984,0.1543788016,-0.2115135789,-0.0644509569,0.0090086861,-0.3272518814,-0.2054938525,0.1604975015,0.8943871856,0.6694521904,0.2533872724,-0.5422837138,0.2767968178,-0.1760485619,-0.025960017,0.0198597591,-0.2345683128,0.3427531421,-0.2071764767,0.4622580707,-0.0255446602,-0.0318100527,0.0727981925,0.036438819,-0.1948776096,0.0417967588,0.4725963175,0.2244651914,0.0703530535,0.1599472314,-0.2708268166,0.280457288,0.3801903129,-0.1596799642,0.1027686074,0.0950903222,-0.145575583,-0.0148924505,0.2546900809,0.0104940627,0.0758028328,0.1760607362,0.0171193983,-0.1715416163,0.0746806264,-0.0338230766,-0.0092220474,0.1070100516,-0.0957994685,-0.2957440615,0.397996366,-0.0150660882,-0.2960931957,-0.1552005261,-0.1319240034,0.1509987265,-0.046348989,0.290843606,-0.0595920347,-0.5412447453,0.3073117435,-0.29486233,-0.1921954006,-0.2989366949,0.1141650677,-0.2823266089,0.130208984,-0.2304908782,-0.4811024368,0.6867519021,-0.0976978093,0.2228114754,-0.0934261531,-0.1628004313,-0.1883012056,0.1311374158,0.440828681,0.216014877,0.3410702646,-0.143998161,0.5487811565,-0.1458551288,-0.3076657951,0.1404660493,0.0481414795,0.1991781592,-0.2155126631,0.1595045626,-0.0369566306,0.0478237718,0.0600176752,0.2438788712,-0.089751035,0.2800734043,-0.2317492962,0.0928029045,0.23717013,-0.1616496146,-0.4624883831,-0.1963436902,0.1247851551,-0.3123081028,0.3161909282,-0.2666243911,-0.3154160082,-0.1647447795,-0.1834567785,-0.1416295916,-0.0881386176,-0.2221992016,0.0457377024,-0.0412664115,-0.1307436228,-0.1385354847,-0.2107693255,0.3759195209,0.0217984933,-0.399569273,-0.1029036418,-0.1500617713,0.0271741692,0.1704886258,-0.1327169836,0.0620967411,-0.1057753041,0.0755798146,0.1335565746,0.0940785483,-0.1578190923,-0.1179904714,0.2207147777,0.0705570206,-0.1607826948,0.2513234615,0.3187564015,0.203243196,0.1024870798,0.2634471655,-0.1615879089,0.1351609081,0.319488287,-0.4736805558,-0.0829159766,0.1024240404,0.0007219925,0.3073074222,0.2434829623,-0.4967825115,-0.1959158629,0.2715581059,-0.5491095781,-0.2005575299,0.0628585145,-0.0746424645,0.1725081652,0.072630845,-0.1223684549,-0.0668064058,-0.4072676003,-0.0691316128,0.3100281656,-0.0219103228,-0.17774795,-0.3566366434,-0.3992902339,-0.093053259,0.2959269285,-0.1511599272,0.1808461547,-0.0303153433,0.1074213088,0.3834602237,0.1661670208,0.3568627536,-0.4037394822,-0.0006812536,0.0243587084,-0.0173624549,-0.3910657167,0.1619275659,-0.2636987865,-0.0248567127,0.2297753394,-0.0767989978,-0.0128447628,-0.2526217699,-0.0165061001,-0.0320344903,-0.281953603,-0.2206416577,-0.0981287658,0.1151982918,-0.0012347116,0.039134331,-0.3254961967,-0.0503391549,-0.2086155117,0.0370726101,-0.244877845,-0.152414605,0.2816011608,0.5065797567,-0.1005822346,0.2178728729,0.0389209539,0.2029356211,0.0159156397,0.0536582619,0.4275124967,0.2068452239,-0.4748890102,0.0239341483,0.0655188039,0.2111892998,0.1162533686,0.0514622554,0.0544395633,-0.3059339225,-0.0435880609,-0.4261354208,0.1663311273,0.1503156424,-0.0694124699,-0.5391888022,-0.1627792865,0.7504374385,0.1072373092,-0.2833897769,0.2404269874,0.3694290817,-0.1128291711,0.0911417082,0.0249155872,0.859405458,-0.0010103866,-0.139820084,-0.0666873232,-0.1861011982,0.2593496144,-0.4081709981,0.132809177,-0.236267224,-0.1062207744,-0.081392467,-0.1731308848,-0.0003388301,0.3037554324,-0.0273835398,0.0514978841,0.0020788691,-0.1182083711,-0.0313686319,-0.0631342381,-0.4598028064,0.0767731518,-0.3565748632,0.1627537608,-0.1747123152,0.4422894716,-0.1572094113,-0.1597495228,-0.1271939129,0.2501044869,-0.7062593699,-0.0980553627,0.0848825201,-0.1956662983,-0.0745778605,-0.0288046896,0.1047401577,0.2502745688,0.3318220675,-0.0971478373,-0.269290477,0.1816768199,-0.4351150692,0.1336654574,0.0469590724,0.0543607511,-0.0616497658,0.0081185214,-0.4391889572,0.4193753302,0.4069055319,-0.0423470922,-0.355089277,0.1938813478,-0.0727896541,-0.127609089,0.1314764172,0.2027264833,0.184370473,0.2467302978,0.1358047873,0.3787496984,0.0870233029,0.333812207,0.1907683462,-0.1340194792,-0.0871977434,-0.1878649145,0.3211949766,0.0499797724,-0.0058583166,-0.0291577168,0.0536780506,-0.2007260919,0.3166264594,0.0587172173,-0.1811845899,-0.1178200245,-0.2657906115,0.1536209732,0.1292905658,0.0801983848,0.0935249627,-0.0665163547,0.0345748588,-0.451985091,0.1431039274,0.4041517079,-0.2118909359,0.3051808178,-0.3709364235,-0.3015109897,-0.0248065088,0.2729508877,-0.2766903937,0.0267839283,-0.181380406,-0.0523765162,-0.1771395206,0.1926278174,-0.0447737724,-0.0737126097,-0.0212035347,-0.1468823701,-0.1794738024,-0.2239365578,0.2025604397,0.1314595044,0.0777185261,0.1646578759,0.008061667,0.0091514764,0.2661632299,-0.1670531332,-0.0388434529,-0.3671861887,0.2220809609,0.2739847898,0.4554353654,-0.1185231805,0.0535644218,0.1602439135,-0.0340114422,0.3793414831,0.4118108749,0.0656532794,0.0429078415,-0.0282267984,0.0663387552,0.2493350804,0.2505869567,0.0101996232,-0.1839417517,0.0641720444,0.5837347507,0.2311163545,-0.0074668033,0.1920495927,-0.0305192135,0.1468071043,0.0628003851,0.2199812531,-0.150026381,-0.1281102151,-0.0851435214,0.2809866965,0.0331245922,0.2353118807,0.2495196313,0.2642408311,0.5634089708,-0.1710542291,0.4700336158,-0.088851124,0.3717053235,0.1326763779,-0.0357450694,-0.0936154202,0.2024019212,0.0258683413,0.0911720023,0.2170695066,-0.1402004212,0.1877423227,-0.1060445234,0.137800172,0.3250533044,-0.2639152706,-0.179434821,0.2496218085,0.015910076,0.24676162,-0.1017866507,-0.034046866,0.156511575,-0.0489472114,-0.2703662515,0.1094438434,0.2058236897,-0.1287345439,-0.0462116338,-0.3164763451,0.0939147845,0.0347564593,0.0394147746,-0.5336431861,-0.1789910644,-0.0088282516,-0.0143991672,-0.3907335103,0.3065728843,0.4233435094,0.2012784481,-0.0695150271,-0.0741470009,0.0912312418,-0.0197193101,0.337695092,0.5650493503,-0.0577122159,-0.1012709737,-0.1964596659,0.1104789749,-0.3864605427,-0.1585022211,-0.1047345549,0.068528831,-0.0597408302,0.0311170891,0.1922676563,0.1358437985,0.0110430019,0.2848908305,-0.1230020225,0.2313934863,-0.2062752843,0.1371389031,-0.1495082676,0.0089321807,-0.071321018,-0.3597351611,-0.4178462625,-0.0991976932,0.2232005149,-0.1679434329,0.0097352788,-0.1797056198,0.0455248207,0.0283364877,0.2231453806,0.5108691454,0.2698989511,0.1474293619,-0.4483953714,-0.2105752975,-0.0687660724,-0.1219296753,-0.0573349558,0.0807898268,0.2413982749,-0.0606409758,0.1788728833,0.1604474336,0.141503334,0.1274715662,0.3011879325,-0.7209573984,-0.1776541471,-0.1319405735,0.420244962,-0.2823081613,-0.1069947034,0.1888155639,-0.2594947517,-0.1056063101,0.1073795632,0.2328464389,0.1289750338,0.0555762723,0.1899854988,0.0424443409,0.2643534839,0.1386158019,0.1542827636,-0.5443912745,-0.2363073677,-0.0446845256,0.0662066266,0.2053744346,0.086939998,-0.186775282,-0.2433930933,0.3166288733,-0.0002630944,0.1859564483,0.0702337027,-0.0096768709,-0.1132529974,0.1296984702,-0.127762869,0.2176863551,0.1216072217,0.1292817742,-0.0450114757,0.2169523984,-0.0759634525,0.1270125657,-0.336224407,-0.170635134,0.1558655649,0.0282379948,-0.1027618572,0.1368471533,-0.259578377,0.0393415019,0.5676419139,-0.1922877133,0.3778695166,0.4671713114,0.4792613089,-0.1509285867,-0.1778731942,-0.0198874697,0.1768817455,-0.344584018,-0.0993165076,-0.0676412135]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1634","title":"Inspecting datasets per category","comments":"Hi @thomwolf \r\nthank you, I was not aware of this, I was looking into the data viewer linked into readme page. \r\n\r\nThis is exactly what I was looking for, but this does not work currently, please see the attached \r\nI am selecting to see all nli datasets in english and it retrieves none. thanks\r\n\r\n![5tarDHn9CP6ngeM](https:\/\/user-images.githubusercontent.com\/53898419\/103107612-1509aa80-4638-11eb-85b5-0c995a189969.png)\r\n\r\n\r\n\r\n","body":"Hi\r\nIs there a way I could get all NLI datasets\/all QA datasets to get some understanding of available datasets per category? this is hard for me to inspect the datasets one by one in the webpage, thanks for the suggestions @lhoestq ","comment_length":55,"text":"Inspecting datasets per category\nHi\r\nIs there a way I could get all NLI datasets\/all QA datasets to get some understanding of available datasets per category? this is hard for me to inspect the datasets one by one in the webpage, thanks for the suggestions @lhoestq \nHi @thomwolf \r\nthank you, I was not aware of this, I was looking into the data viewer linked into readme page. \r\n\r\nThis is exactly what I was looking for, but this does not work currently, please see the attached \r\nI am selecting to see all nli datasets in english and it retrieves none. thanks\r\n\r\n![5tarDHn9CP6ngeM](https:\/\/user-images.githubusercontent.com\/53898419\/103107612-1509aa80-4638-11eb-85b5-0c995a189969.png)\r\n\r\n\r\n\r\n","embeddings":[-0.3020195365,0.0937979147,-0.1714936346,0.4588537514,0.0671875924,0.2000089139,0.1474443525,0.49897331,-0.1237441152,-0.0721698329,-0.6115688682,-0.0818746164,-0.0379927382,0.2360660881,0.131478712,0.0705120787,0.065998964,0.0717614144,-0.2098444998,-0.0730781108,-0.059058927,-0.0043414631,-0.2953975797,0.2639480531,0.1482008398,-0.0464194864,-0.4075677693,-0.291372925,-0.2740218639,-0.3585765958,0.164096415,0.2567878962,0.4316548407,0.0354354978,-0.000108613,-0.0164748654,0.1057887301,0.0324192271,-0.0531503595,0.0628628582,-0.4935716391,-0.3013846874,0.2357198596,-0.1952267736,0.0598739646,-0.2024279088,0.5626852512,-0.3242330253,-0.1774676293,0.1138124242,0.1750924736,-0.0224909093,-0.3953582644,0.1345527172,-0.1644510925,0.1767985076,-0.1878391653,-0.130546689,0.4642978609,-0.0359445103,-0.2221885026,0.3846780062,0.1228795126,0.0805994868,-0.0227045789,-0.0394640118,0.0611899346,-0.4473687112,0.1216736063,0.6197764277,0.6888504028,-0.0595749803,-0.4407017827,-0.0740549341,-0.1628539264,0.1237755194,-0.3919146657,0.9289369583,-0.1156804115,0.1182618216,-0.1310912371,-0.056632258,-0.3187266588,0.0347070284,-0.294816345,0.3741955459,-0.0890563503,-0.0939147621,0.3317596912,-0.1761251241,0.3616896272,0.0232480131,-0.4417388439,0.1799140722,-0.7394194007,-0.1312709004,0.2512081861,0.54157722,0.2040117234,-0.0172321368,0.2887922823,0.2197982669,-0.2134324163,-0.1720743775,0.4058315158,0.1329377294,0.3010050356,0.0407841131,0.1693354994,-0.1552112699,-0.0337218195,0.0742785707,-0.2333092541,0.0204621889,-0.3318940103,-0.0512507595,0.0436876714,-0.1987778246,-0.2195052356,-0.3356381655,-0.1315139681,-0.4507161677,-0.0831772387,0.3694804907,0.1268960685,-0.0060849669,-0.3100359738,0.3781314194,-0.2511692941,-0.6250892282,-0.0985437706,0.0786972567,-0.0395732038,-0.1394738555,0.1178268567,-0.132292375,-0.0797973722,-0.1720879674,0.0024927517,0.0442238301,-0.0058483463,-0.0924382731,0.0156469904,-0.037989933,0.4044882953,-0.1612757593,0.1049123332,-0.0051472778,-0.3377614319,0.0264738481,0.087710388,-0.2825858891,-0.1999628544,0.1168851554,-0.2446359694,-0.2116881758,-0.1490552574,0.6655653715,-0.3820025921,0.1058220938,0.0460244417,0.1420821995,0.2311930209,-0.1116814986,-0.1651346236,0.4212861657,-0.4193429053,-0.0928275809,-0.1446888745,-0.3321814835,-0.035393361,0.1578119099,-0.0548250973,0.0212567579,-0.3922074735,0.0473742113,0.8782119155,-0.138991788,-0.2263593376,0.2655929029,0.3861093223,-0.2313533574,0.2574380636,0.2039435059,0.0125827817,0.0419330858,-0.3922658563,0.1225627586,-0.0020044029,0.0424103327,0.0268573072,-0.4239691496,0.3247992992,0.179589808,0.0849992931,0.1655385047,0.3775900602,-0.1253252923,0.587793529,-0.1668925136,-0.0504017398,0.0300414674,-0.0740754157,-0.0916089043,0.1358509511,-0.2105307132,-0.3396768868,0.119444333,-0.082640782,0.320548296,0.0602466837,-0.1397473067,0.1110084727,-0.0298251733,-0.2207383364,-0.0601396561,0.2412538528,0.2704812586,0.0336975642,0.215193823,-0.2410713285,0.2090810388,0.0245016143,-0.0129749887,-0.0265538804,0.2569160461,0.0907474309,0.3026961684,0.2979464233,0.126738295,-0.1749521941,-0.0169488527,-0.0762933418,-0.3363853395,-0.1274318248,0.1786785573,0.7207256556,0.5630325675,0.2103105187,-0.4626798034,0.3300400674,-0.0529621504,-0.0632888079,0.0695943236,-0.3170131147,0.3304572403,-0.0975014716,0.4795435667,0.0679147243,-0.0239531547,0.219502598,0.0179052707,-0.1856084466,-0.1168108657,0.48088485,0.2276680171,0.0520079918,0.1034202427,-0.3596718013,0.3000658154,0.4525820911,-0.1030228212,0.0546753742,0.0290706661,-0.0871543735,0.0506488159,0.2534509003,0.0919288173,0.0513923354,0.2142790705,0.1353226453,-0.1168498173,0.1431589127,-0.0556243286,0.0240946878,0.0016845705,-0.0039028097,-0.1994135529,0.3179620504,-0.1397604346,-0.3216177821,-0.1213301495,-0.0041835429,0.1333161741,-0.068133086,0.1726482958,-0.2653362751,-0.7509955168,0.0605957881,-0.2273226529,-0.1888329834,-0.2517419755,-0.0104628066,-0.4002918303,0.1387111992,-0.1137303039,-0.5434058309,0.509698689,-0.1051074937,0.3512748182,-0.070585914,-0.1407811642,-0.3351511359,0.1513425857,0.4208589792,0.1502161175,0.1985864341,-0.2169165909,0.4200828671,-0.0584665202,-0.3545227051,0.2315334976,-0.0147799132,0.209756434,-0.3728648424,0.0447465964,-0.1688536257,0.0584379993,0.0314166956,0.2289773673,-0.123162955,0.2000934482,-0.1428049058,0.0288113058,0.1368291974,-0.3826674223,-0.5265018344,-0.1817942262,0.0331892557,-0.2954965532,0.438386023,-0.2491925061,-0.3484262526,-0.1118429676,-0.107559368,-0.0286989696,-0.0434639677,-0.0734534934,0.1195354685,-0.129232794,-0.2260616124,0.0163660757,-0.1137905791,0.3740551472,-0.0836806074,-0.442250371,0.0434216335,-0.1405605525,-0.0144093074,0.1422806531,-0.0484289639,0.0285579599,-0.0508260615,0.0869196057,0.0409819409,0.0735117123,-0.0676587075,-0.1629754603,0.1693504751,0.0401130021,-0.043169789,0.2874175608,0.1787169725,0.208769545,0.1954711676,0.2579610348,-0.1945156455,0.1424736977,0.3222069442,-0.3774318993,0.0478432477,0.0303767808,0.0558932573,0.3996391594,0.1450021267,-0.628862381,-0.2084129751,0.3221397102,-0.5552141666,-0.2265350074,0.088812612,-0.149536103,0.2075648457,0.0207619518,-0.0775999874,0.0837036222,-0.3250208795,-0.1575468332,0.3913148046,0.0509020239,-0.0621320307,-0.3364440799,-0.3859550953,-0.026969308,0.286414355,-0.1135942414,0.2579450607,-0.0970210433,0.1719461679,0.3771682978,0.0909999534,0.3423734009,-0.2114246786,0.0279691219,0.0932235941,0.0083120959,-0.2816431522,0.0696272776,-0.278834343,-0.0912206545,0.277518332,-0.3083480597,-0.0473563969,-0.3179281056,-0.0532584973,0.0690794513,-0.3432067037,-0.1513072401,-0.0900534019,0.114066191,0.030963853,-0.0589107126,-0.3538697362,-0.0032025399,-0.2659939229,0.0684546754,-0.1913894415,-0.1656214893,0.2111181468,0.377758503,-0.0121983998,0.2206458002,0.021257868,0.1479032189,0.0771995187,-0.0063233813,0.4515425265,0.2630317807,-0.5572960973,0.0850619674,-0.0800327584,0.2294076979,0.033795312,0.022400979,0.0002941984,-0.141196698,-0.1852438152,-0.1836123616,0.1143568084,0.1947583407,-0.0497117452,-0.4863708019,-0.3311288953,0.6556469202,0.1367168427,-0.221098274,0.0318397172,0.3164769709,-0.1561466753,0.0797094107,0.0147972656,0.8090863824,0.0801175684,-0.1332387328,-0.045145411,-0.3093684018,0.1627748758,-0.3230440617,0.2105727792,-0.3043628931,-0.0359013751,-0.0438176356,-0.1073023081,0.0576686896,0.3989127278,-0.0960458741,0.0985174701,0.1258398741,-0.3313917816,0.0753259733,0.0210373476,-0.3838549256,0.1116892844,-0.2615568936,0.1308001876,-0.1490581036,0.2083064318,-0.0944458842,-0.1825577617,-0.0428501293,0.3301802278,-0.6485316157,0.0021135267,0.0714440495,-0.1575013697,-0.1743376851,-0.1107510924,0.2586371303,0.3972183764,0.1567978561,-0.0609822869,-0.2657901645,0.1670799255,-0.2500766516,0.1785642803,0.1506242752,0.1089887172,-0.0166761391,-0.0209060684,-0.4588471651,0.354034692,0.3280014396,0.0789715722,-0.4013212919,0.2116367966,-0.0427926183,-0.0684071258,0.2023304701,0.2025087178,0.174160406,0.2887026668,0.1374204755,0.3309772015,0.0645346791,0.3572795093,0.2028447688,-0.1039934829,-0.092031233,-0.19272089,0.3683835268,0.0001067464,0.0997368991,0.0861869454,-0.0519411825,-0.1495075524,0.2414253801,-0.0497906879,-0.1681580991,-0.231042698,-0.2521412969,0.3320322633,0.1683422327,0.0790033117,-0.0618122667,-0.0134208007,-0.0308932811,-0.5010133982,0.1910608262,0.3252383769,-0.3416942358,0.2430303246,-0.1877229214,-0.3507114947,-0.0251149144,0.2102278322,-0.2612726092,-0.0132477423,-0.0355978794,0.0339167491,-0.0463921055,0.2421016097,-0.1471871585,0.0151670044,-0.024263205,-0.1183610335,-0.0831626654,-0.2031707168,0.229935959,0.1283489168,0.1762433499,0.0689413026,-0.0022135004,-0.1151167303,0.1910297275,-0.0516595729,-0.0038870147,-0.379530251,0.0463465638,0.4171484411,0.2537681758,-0.0822300091,0.070400238,0.0926540643,0.0384930037,0.3643124998,0.4042586386,0.1657267064,0.09381219,0.0037083558,0.075603798,0.2010842115,0.1566458195,0.0732639655,-0.1140944064,-0.0839857757,0.6470817327,0.0230679847,0.0876667425,0.1875973046,-0.0351429023,0.1391638517,0.2540566325,0.2066786587,-0.1567818224,-0.1206426546,-0.1098901033,0.159055233,-0.0080295205,0.268357873,0.2346677035,0.0748463944,0.649841845,-0.1366043836,0.3954975009,-0.122255981,0.3186435401,0.3236806691,-0.09209802,-0.1056098491,0.2292933464,-0.0402623266,0.0379138365,0.3127177358,-0.2224684358,0.0657831877,-0.0051798359,0.1318513751,0.2921599746,-0.2281017601,-0.1554634571,0.2287596315,-0.0486883223,0.1844877452,-0.030736221,-0.103726618,0.2893851101,-0.0508962274,-0.2401756942,0.1599112004,0.1840174794,-0.0789205208,-0.0847471356,-0.3151169419,0.03318353,0.1819830537,-0.0495014451,-0.4605704248,-0.2441245764,0.1062988192,-0.0710188225,-0.3761577308,0.2659576237,0.3715748787,0.1353331655,-0.1623423249,0.0586528778,0.1455444992,0.0116410609,0.2370201051,0.6227419972,0.0092102261,0.0220941491,-0.0794003606,-0.0717060715,-0.2960782647,-0.0912010744,-0.0708020627,0.07689742,-0.012136912,0.0488520972,0.2988015413,0.0873346031,0.0341220759,0.3583627045,-0.0538328402,0.0921837687,-0.2237171233,0.0200411268,-0.0467168875,0.0596249513,-0.0720158741,-0.4267560244,-0.5183417797,-0.0902432799,0.1778504699,-0.080242753,0.0134646632,-0.2390890419,0.0593329333,0.1258343756,0.1538510174,0.4542290568,0.3827670217,0.0820902437,-0.4627159536,-0.2488354594,-0.1256547272,-0.0287638921,-0.0238947775,0.0432420224,0.2596847415,-0.0028939366,0.2010757029,0.1798426807,0.2567205131,0.0987332761,0.4813728929,-0.6798579693,-0.2280742824,0.0302763097,0.4915425181,-0.3918305933,-0.0666948035,0.1848249137,-0.3721673489,-0.0672245249,0.0714228973,0.2307497561,0.1590185761,0.1329328269,0.1774269789,0.136746332,0.3408711851,0.1042286381,0.1905265003,-0.5206269026,-0.1819166243,-0.172936216,0.2167003453,0.2120067328,0.0511450544,-0.1508770883,-0.222330749,0.2895967364,-0.0461106934,0.0446382277,0.0916934907,0.07095965,-0.1301771998,0.1377518624,-0.200435251,0.1721227765,0.108612664,0.1958357543,-0.078768529,0.0830467343,-0.0166957285,-0.1073185131,-0.4275692105,-0.1512579769,0.1423338503,0.0000478543,-0.0957090035,0.183661893,-0.2501519322,0.0647431761,0.5259506106,-0.2010783702,0.3949595988,0.3639182448,0.4660130739,-0.1426101625,-0.2556731999,-0.0267155748,0.3524829447,-0.3383699059,-0.1640617549,-0.1033092663]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1634","title":"Inspecting datasets per category","comments":"I see 4 results for NLI in English but indeed some are not tagged yet and missing (GLUE), we will focus on that in January (cc @yjernite): https:\/\/huggingface.co\/datasets?filter=task_ids:natural-language-inference,languages:en","body":"Hi\r\nIs there a way I could get all NLI datasets\/all QA datasets to get some understanding of available datasets per category? this is hard for me to inspect the datasets one by one in the webpage, thanks for the suggestions @lhoestq ","comment_length":28,"text":"Inspecting datasets per category\nHi\r\nIs there a way I could get all NLI datasets\/all QA datasets to get some understanding of available datasets per category? this is hard for me to inspect the datasets one by one in the webpage, thanks for the suggestions @lhoestq \nI see 4 results for NLI in English but indeed some are not tagged yet and missing (GLUE), we will focus on that in January (cc @yjernite): https:\/\/huggingface.co\/datasets?filter=task_ids:natural-language-inference,languages:en","embeddings":[-0.1127985418,0.0156888273,-0.2197402567,0.4125655591,-0.0079149175,0.026991684,0.0843824744,0.4735390544,-0.0621107072,-0.1604093611,-0.5733580589,-0.1227079108,0.0521350764,0.3426798284,0.1644427776,0.0418968275,0.100888744,-0.0833265334,-0.2709667981,-0.1761794686,-0.0469946191,0.178943187,-0.2701209188,0.3549456298,0.0659168661,-0.0939937755,-0.4046341181,-0.3289308548,-0.2377311885,-0.2400817424,0.1890156269,0.2508487105,0.4176180065,-0.0365298539,-0.0001102242,-0.0850438923,0.0039951331,-0.0407811664,-0.0730079189,0.0611323453,-0.4303300381,-0.3196310103,0.1286224574,-0.1766109169,-0.0106974281,-0.1142883599,0.4511528611,-0.3239992261,-0.1636002958,0.0433886871,0.1818676591,0.0093082245,-0.4813211262,0.146073699,-0.1470403373,0.1292658448,-0.0827775225,-0.0604550317,0.5571928024,-0.2251417935,-0.1827837378,0.4445842206,0.1540631205,-0.0277672932,-0.1510073245,0.0132741947,-0.0464177206,-0.5090123415,0.0752446428,0.6124581099,0.496126622,-0.0473681651,-0.5170997977,-0.1746857762,-0.1840706319,0.0752101094,-0.4121730328,0.8895875216,-0.1419305205,0.2088404447,0.0084068812,-0.0851003453,-0.2214701623,0.1187308356,-0.1582270563,0.4387899637,-0.0738752633,-0.117742613,0.2705903947,-0.2010828406,0.062749058,0.0608076528,-0.3862090111,0.1860885322,-0.69509691,-0.1772442013,0.3227952421,0.5855038762,0.2710989714,0.0167680308,0.1947613806,0.1557922065,-0.1826265454,-0.161041379,0.4567234814,0.1354044229,0.2885792255,0.0037540917,0.1845114082,-0.1134006605,-0.0974414125,0.0830178782,-0.1895136833,0.1046670973,-0.2143543214,0.0113174664,0.0283199977,-0.2254145145,-0.2909221947,-0.3071687222,-0.1820036918,-0.4849576652,-0.1087528318,0.339627713,0.1163031757,0.0348418914,-0.2204786688,0.4193568528,-0.2193921953,-0.6845995784,-0.0882566422,0.0998279229,-0.0802271515,-0.0805768818,0.0993035138,-0.0153533192,-0.057511773,-0.2019354403,-0.0699019358,0.1084723249,-0.0063275839,-0.189262554,0.1160751283,-0.0741073564,0.2871043682,-0.109026581,0.0747684762,0.0665938407,-0.4490146935,-0.1377215832,0.1082337275,-0.236370638,-0.1890600473,0.0896724612,-0.2355096787,-0.2154624313,-0.2315912545,0.8258110881,-0.3214225769,0.2188818455,0.1348439902,0.1498854905,0.2046007663,-0.1159629449,-0.1072126552,0.3968374729,-0.3240875602,-0.2145777345,-0.0769570172,-0.2934623361,0.0621811412,0.3232209086,-0.0426837839,0.1701610684,-0.3040454388,0.1332400739,0.8199773431,-0.1311862916,-0.2302034944,0.1967497021,0.3711154163,-0.1904722005,0.2296781242,0.1832924336,0.0072219879,0.0040391102,-0.2037837356,0.3367217481,0.0340469331,-0.0324148946,0.0162279028,-0.4255031645,0.3276370466,0.1472854465,0.0769338682,0.1007180735,0.2246593535,-0.2018427551,0.385843724,-0.1028623506,-0.0476812311,-0.0138876671,0.0274659079,-0.1208754182,0.0670092627,-0.272803992,-0.3843850195,0.1236685589,-0.082649067,0.4163319468,-0.073247388,-0.1229216382,0.1007919684,-0.0096191,-0.1153100729,-0.0703625381,0.2301851064,0.3147184849,0.118111603,0.2442222238,-0.3133840561,0.2815248668,-0.0372196063,0.0223637763,-0.1091490313,0.2844394445,0.087567918,0.1831196249,0.2248633206,0.2734616995,-0.2931275666,-0.0311005656,-0.0402636416,-0.4010964632,-0.2481615692,0.1165764853,0.8087966442,0.5512917042,0.174339518,-0.5052917004,0.2770757377,-0.133137241,-0.0335992761,0.0797795355,-0.0714485496,0.3099631071,-0.0887234509,0.5227497816,0.0742839053,-0.0779778659,0.1878208667,0.1012386903,-0.2409105301,-0.0713264123,0.53509444,0.0795138851,0.0601890832,0.161670208,-0.43609339,0.2146095634,0.532804966,-0.1646728814,0.149516046,0.101268664,-0.0416984819,-0.0310222078,0.2356841117,0.0367941149,0.1024970114,0.2288898677,0.1723770499,-0.2392042577,0.1234094128,-0.0517314635,0.073348321,-0.0261238907,-0.0974049121,-0.2387784123,0.3937008083,-0.0026428597,-0.3572147191,-0.066080302,-0.1198119298,0.1046438217,0.0034508468,0.2326081246,-0.2002726346,-0.8213554621,0.1187007576,-0.2470667958,-0.2600858808,-0.1919220239,-0.0024958323,-0.569118619,0.1178916544,-0.1062088087,-0.4092338979,0.4692082703,-0.2063668817,0.1681923121,-0.1286978126,-0.2273899615,-0.3320856094,0.1356230974,0.3242715597,0.1927656978,0.2986070216,-0.3433547616,0.2622004747,0.0103463726,-0.3482649326,0.186990872,-0.0681907609,0.2572000921,-0.2491106987,0.102082409,-0.0885111466,0.0174125433,0.1467038542,0.2707752287,-0.1872179955,0.1284037828,-0.2476276308,0.0773552582,0.1484889835,-0.4460300207,-0.5780335665,-0.1683107913,0.0925660804,-0.357511431,0.3129068017,-0.149653405,-0.4064166248,-0.1207008511,-0.2696771622,0.0610913672,-0.0928570107,-0.0928908065,0.0211014841,0.0906309932,-0.111561358,-0.1502605677,-0.1687968969,0.4150243998,-0.0158124343,-0.3585648537,-0.022690298,-0.1867684275,0.1630617082,0.1342711896,0.0555054843,0.1085530892,-0.0928499475,0.0801583007,0.0759951547,-0.0502451286,0.0357467681,-0.0165054202,0.2407966256,0.0827601105,0.0190385152,0.23616229,0.3050002456,0.2867231369,0.0190422051,0.3718002141,-0.0424514525,0.1373587102,0.1845540255,-0.4306498468,0.0667117238,0.1064558476,0.0897859931,0.3318333924,0.1840807199,-0.5415427685,-0.2435428798,0.2247100472,-0.6125219464,-0.2315607518,0.1145230159,-0.0413795896,0.2235575616,0.0544813089,-0.1867135614,0.0731294006,-0.2060892731,-0.0863276199,0.3274203837,0.0444091782,-0.038226407,-0.4139119089,-0.3529693484,-0.106470041,0.249264583,-0.0923417434,0.2507375181,-0.1812244505,0.1780225188,0.3566058278,0.1714970171,0.2985116541,-0.40016523,-0.0698320195,0.0733716413,-0.027570596,-0.2477744669,0.1407765448,-0.2982812822,-0.2307197452,0.5297487378,-0.1279580742,-0.2148176879,-0.4016868174,-0.0655421317,0.1682411283,-0.1684613228,-0.1904632598,-0.1303549707,0.0730409771,0.029120259,0.1132497638,-0.2631711662,-0.0910049602,-0.1226071194,0.0415868238,-0.1461172104,-0.1333321035,0.232002303,0.3934748769,-0.0892106965,0.1989943981,0.0685409531,0.0750705674,0.0807505399,-0.1342319101,0.4367637336,0.0533443913,-0.5336012244,0.0033448581,0.0795833915,0.4183849394,0.1011236832,0.0538832657,0.1381784827,-0.2014583498,-0.1979631186,-0.2396720648,0.1622288078,0.3017291725,0.0723012015,-0.4462114871,-0.2070462555,0.5728653073,0.1923396289,-0.3388906717,0.0443107747,0.3722043633,-0.1955408305,0.1116704494,-0.0236386824,0.831828475,-0.0113868583,-0.0249485914,-0.0516159795,-0.2059978396,0.332496196,-0.1751130968,0.0926594958,-0.1611166745,0.0311760474,0.0351069942,-0.0833073035,-0.0892659947,0.3349159956,-0.0461584106,0.3119551539,0.0801177397,-0.2309140116,0.0895702764,-0.0007283898,-0.3086006641,-0.0069984235,-0.3852083981,0.1337172538,-0.1953079104,0.3649826944,-0.1054603681,-0.1772147864,0.0195661616,0.2956980765,-0.6136066914,0.0287193004,-0.0179091897,-0.2332647592,-0.0462430343,-0.0340144224,0.2420563847,0.3691946566,0.3097740114,-0.0812722072,-0.1127198935,0.0926257148,-0.3355250657,0.2084056884,0.1703900993,0.0901057348,0.0688824803,0.0377710164,-0.4751331508,0.2571143806,0.3688825369,0.2071536332,-0.4394441843,0.1714011729,-0.0394590385,-0.0476527661,0.0917825326,0.3183694184,0.1799232662,0.3305414617,0.1340893656,0.4675774276,0.1035875753,0.285561204,0.1866770685,-0.1476492882,-0.1027946547,-0.2752368152,0.3838764131,0.0956323519,0.0610088818,0.0361117087,-0.001712613,-0.1880677938,0.1357816011,0.0026864894,-0.2526122928,-0.226101011,-0.1686774641,0.1145653129,0.029958846,0.103838414,0.0409370884,-0.1136746779,0.0526609011,-0.4389400482,0.1644084901,0.4572065771,-0.351732105,0.1087554917,-0.1565157771,-0.3679513633,-0.033091031,0.2793291807,-0.2604285181,-0.011719862,-0.0793886334,0.035030555,0.0180344805,0.1826602072,0.0352749191,0.0661131442,-0.0231497362,-0.125864923,-0.1180765852,-0.231439814,0.1804789156,0.1233825758,0.1091677621,0.1509302109,0.1224827021,-0.20742172,0.3152784705,-0.223726958,-0.0464904457,-0.4435856342,0.136268571,0.5214185119,0.2045168579,-0.0018961651,0.0614307709,0.1071740165,-0.0241139531,0.219435364,0.3966358602,0.088531211,0.1594089717,-0.0084563307,-0.0201986749,0.1451419145,0.1739730239,0.0265043899,-0.2240360826,0.0596449822,0.493840754,0.0146694798,0.0791525692,0.1161772832,0.0149493106,0.0795855001,0.2458634228,0.2282721251,-0.2086607516,0.0324506573,-0.0744107515,0.2234495282,0.0267187338,0.3163642287,0.2787234783,0.1507699639,0.6196379662,-0.2598477006,0.2862073481,-0.1257545948,0.3093803227,0.1347479671,0.1441496462,-0.0486202836,0.1830772907,-0.0801406875,0.0158884861,0.1508247852,-0.1272360384,0.120108515,-0.0241700653,0.1604949981,0.2226890773,-0.1539549381,-0.1497023851,0.3707998097,0.0718796104,0.182846278,-0.0070506567,0.1533213258,0.2372010201,-0.132270664,-0.1776412278,0.1726247221,0.1134885922,-0.1595599055,-0.0777849779,-0.2316888422,-0.0062043252,0.1058845073,-0.1173260659,-0.5116460323,-0.1307653189,0.1200508699,-0.0045922059,-0.4869020879,0.0960843712,0.3425357044,0.172542423,-0.1181792915,0.2598576546,0.0871040896,-0.0676201507,0.1424230933,0.6825017929,0.027133489,-0.0993227884,-0.1503872573,0.0349730738,-0.3102734387,-0.0708337128,-0.1033069193,0.149336949,0.0227612182,0.0881430954,0.2851001918,0.0849182457,0.0288160872,0.2295914292,-0.0570198447,0.0951020569,-0.2746156454,0.1256340444,-0.2114600688,0.1131284833,-0.2031775117,-0.3252632916,-0.5201652646,0.0268219505,0.1028184146,-0.0754069611,0.0176996961,-0.2109995186,0.0402792506,0.0440324955,0.1029941365,0.5331733823,0.404358387,0.0349197015,-0.4275890887,-0.2034166306,-0.1857515723,-0.1308433563,0.0431984104,-0.0820477083,0.2520276904,0.0909275115,0.1533996314,0.0318215899,0.2669098973,0.1237296388,0.3093392551,-0.7733200192,-0.2053343356,-0.1485769898,0.5449175835,-0.3012162149,-0.0835859627,0.0205549467,-0.394986093,-0.1135889143,0.0780164748,0.1775304377,0.1522607803,0.2333381921,0.14840433,0.2053357214,0.3941547573,0.0528875515,0.0722308382,-0.3629334271,-0.2571231723,-0.1393941194,0.2454861999,0.1717245877,0.0829621553,-0.2216353416,-0.1020846665,0.2900405228,-0.0459350497,0.0188567974,0.0039924332,0.0209292527,-0.0718698502,0.2503847778,-0.3015879989,0.1815918088,0.253226459,0.1314146817,-0.0951132029,0.1234531477,0.0051023909,-0.0372708775,-0.4289672375,-0.1710698307,0.1141090319,0.0715207979,0.0237286873,0.1730563045,-0.330994159,0.0133583751,0.5041663051,-0.1386148483,0.3007952869,0.3942507505,0.5251535773,-0.0820360631,-0.1610569656,-0.066708073,0.2772645056,-0.3969744444,-0.0334269106,-0.1631158292]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1633","title":"social_i_qa wrong format of labels","comments":"@lhoestq, should I raise a PR for this? Just a minor change while reading labels text file","body":"Hi,\r\nthere is extra \"\\n\" in labels of social_i_qa datasets, no big deal, but I was wondering if you could remove it to make it consistent.\r\nso label is 'label': '1\\n', not '1'\r\nthanks\r\n\r\n```\r\n>>> import datasets \r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\r\n... 'social_i_qa')\r\ncahce dir \/julia\/cache\/datasets\r\nDownloading: 4.72kB [00:00, 3.52MB\/s] \r\ncahce dir \/julia\/cache\/datasets\r\nDownloading: 2.19kB [00:00, 1.81MB\/s] \r\nUsing custom data configuration default\r\nReusing dataset social_i_qa (\/julia\/datasets\/social_i_qa\/default\/0.1.0\/4a4190cc2d2482d43416c2167c0c5dccdd769d4482e84893614bd069e5c3ba06)\r\n>>> dataset['train'][0]\r\n{'answerA': 'like attending', 'answerB': 'like staying home', 'answerC': 'a good friend to have', 'context': 'Cameron decided to have a barbecue and gathered her friends together.', 'label': '1\\n', 'question': 'How would Others feel as a result?'}\r\n\r\n```\r\n\r\n","comment_length":17,"text":"social_i_qa wrong format of labels\nHi,\r\nthere is extra \"\\n\" in labels of social_i_qa datasets, no big deal, but I was wondering if you could remove it to make it consistent.\r\nso label is 'label': '1\\n', not '1'\r\nthanks\r\n\r\n```\r\n>>> import datasets \r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\r\n... 'social_i_qa')\r\ncahce dir \/julia\/cache\/datasets\r\nDownloading: 4.72kB [00:00, 3.52MB\/s] \r\ncahce dir \/julia\/cache\/datasets\r\nDownloading: 2.19kB [00:00, 1.81MB\/s] \r\nUsing custom data configuration default\r\nReusing dataset social_i_qa (\/julia\/datasets\/social_i_qa\/default\/0.1.0\/4a4190cc2d2482d43416c2167c0c5dccdd769d4482e84893614bd069e5c3ba06)\r\n>>> dataset['train'][0]\r\n{'answerA': 'like attending', 'answerB': 'like staying home', 'answerC': 'a good friend to have', 'context': 'Cameron decided to have a barbecue and gathered her friends together.', 'label': '1\\n', 'question': 'How would Others feel as a result?'}\r\n\r\n```\r\n\r\n\n@lhoestq, should I raise a PR for this? Just a minor change while reading labels text file","embeddings":[0.0132785533,-0.2219308168,-0.0782271028,0.3724220097,0.1299581081,-0.1916471422,0.0608600564,0.2249795496,-0.2229051143,0.3125813603,-0.16639103,-0.3445963562,-0.078576982,0.5373507738,-0.1730184257,0.0900040045,0.1220888272,0.2645392716,0.6475253105,-0.138331458,-0.236980021,0.105072692,-0.0336987302,0.3101118803,-0.2785474956,-0.2114457488,0.0093865162,-0.3806235194,-0.1008315459,-0.4061317444,0.1285590231,0.3338832855,0.1116476506,0.3532100618,-0.0001260965,-0.0991124064,0.0040047453,-0.1466123015,0.0817428753,0.1455848068,-0.2214523405,-0.1744339764,-0.1749262065,0.2110028565,-0.1103216335,-0.5590037107,-0.1015454158,-0.1866581887,-0.0710556731,0.2300129682,0.0624605045,-0.2600317001,-0.2793854177,0.094369702,0.2292268723,0.2006987482,0.2353515625,0.1537181437,0.0999924541,0.2369868606,0.624463737,0.7521879077,-0.3069121242,-0.1839202493,0.1455566585,0.0703373626,0.4446062446,-0.339754045,0.0911542103,0.2607860267,0.8203286529,-0.1031029671,-0.2267248631,-0.1876580417,0.1667058021,-0.267074883,-0.2224024534,0.0834130719,-0.0434053317,0.2503166199,-0.0584254488,-0.2887458801,-0.0803501084,0.0705720782,0.1526282281,0.4675337076,-0.0119621744,0.1469963938,-0.2605960369,-0.0835636333,-0.0143391658,0.0301113464,-0.3768481314,-0.1946709603,-0.1289010793,-0.0656705648,-0.2519749403,-0.0269439276,-0.0413859449,0.1595534682,0.0440837257,-0.1755918711,0.0466385409,-0.1568239182,0.3857022524,-0.0102505358,0.3896238506,0.4299487174,0.1011596173,-0.2825559974,-0.1060048789,-0.1774226576,0.1876734346,-0.2209671289,0.1700438857,-0.014759968,0.2454372197,-0.3865423799,-0.239273414,0.0199281685,-0.5819667578,0.0038895183,-0.1383946538,0.0742783472,-0.118504025,0.4221025407,-0.0419078022,0.12136399,-0.232444331,-0.3014339209,-0.1508969069,-0.2516450882,0.1877764612,-0.1084897369,-0.0364330821,0.0238047671,-0.1202055737,-0.016203776,0.1037535593,-0.003325715,-0.1993272305,-0.2320940048,0.3165769279,0.0554180779,-0.1698958129,0.2157836258,-0.1535643488,-0.3747541308,-0.2571625412,0.1514257491,-0.3373299837,-0.2044187635,-0.0937221646,0.1288665533,0.2472561747,-0.2678976953,-0.1742645502,0.3157111406,0.3574001491,-0.2569462061,0.3162155747,-0.3247472346,0.2047551125,0.0265886262,-0.0607221201,0.3076834977,-0.5260057449,-0.2504105568,0.3707752526,0.004788144,0.5158595443,0.5865712166,0.103598401,0.2158721685,-0.0601794347,0.0508755036,0.3392345607,-0.6111678481,-0.2679282725,0.267490983,0.1576338857,0.0104533965,0.2217943519,0.183943525,0.3711941838,0.0081192078,0.1132846028,0.0204729512,-0.0163176339,-0.14601928,-0.3352743089,-0.0568639822,0.2459256351,0.1221683025,-0.2678988278,0.0576463602,-0.2807745934,-0.5657582283,0.2322619557,-0.2211879343,0.0653799027,0.2120517045,0.7162328362,0.4489491582,0.0694976747,0.0874293521,-0.2416492254,-0.0466044508,-0.0882573426,0.1843825877,0.1429970115,-0.4669782221,-0.1074726433,-0.1339941025,-0.1955721527,0.003102463,-0.0129684294,0.0612174757,-0.1843079478,-0.0037840984,0.0523441434,0.4653782547,0.1108090356,0.1275124699,-0.0339819677,0.0352421179,-0.0814232454,-0.0387319438,0.0021282395,0.4082733095,0.3573410213,0.1827614009,0.0059242947,0.3441536427,-0.1253070086,-0.100705564,0.1207252368,0.1969972849,0.1039842293,-0.3600364923,0.1242761239,0.1653449088,-0.2149889767,0.0778153017,-0.1491081715,0.202808097,0.1893986315,0.0931470096,-0.4089239836,-0.1642623097,0.2763912082,0.0495437011,-0.1534664184,-0.0670580119,0.2308488935,-0.4607717991,0.1187176481,-0.0620222576,-0.5775554776,0.3442960382,0.4050579369,-0.0120826093,0.0264269877,0.1860125959,-0.094049044,-0.1570965648,-0.0063945749,0.2693271339,0.1008819193,0.0970500186,-0.1272014976,0.0594439469,0.3202559948,-0.1740100086,0.283994168,-0.0439146087,-0.1747615784,0.0919813737,0.2362814993,0.0894765109,0.1576460153,-0.1308176219,-0.3797545135,0.266120404,-0.4184514582,0.1302920878,-0.0984707698,-0.4785610735,-0.1266468316,-0.6135700941,-0.4269012213,-0.2420767248,0.1106072664,-0.2983385623,0.1699324995,0.276706934,0.4350067675,0.3092648685,-0.2435679138,-0.2629216313,-0.0271620601,-0.1783802956,-0.3288605213,-0.1130735725,-0.1880236715,0.2747863829,-0.0625045672,-0.3044040799,-0.0050127776,0.0677356496,-0.3274458647,0.1188004836,-0.079474479,0.0203505736,0.2615310252,0.024734132,-0.1166634262,0.2391006052,-0.0881174058,-0.1330470294,0.2544910908,0.1026047692,-0.4490543902,-0.0183136594,-0.2510599494,-0.3221260309,-0.6444743872,0.2027774155,0.1054435968,-0.0162830967,0.0016732258,0.191882655,-0.1805253774,-0.1840477586,0.0804873854,-0.2243872583,-0.1326185912,-0.3191981018,0.1793168038,0.020745514,0.1199438795,-0.0446727015,0.1645192951,-0.2015758455,-0.1402556747,-0.6035840511,-0.0757532716,-0.0402427986,0.1338956654,-0.2781124413,0.0321618766,0.1169930771,0.0527104549,0.0304119308,0.0637016669,-0.1129297167,0.2649347782,0.1484680027,-0.0023496815,-0.1855294406,-0.0081923464,-0.0780180022,0.2050890923,0.3046781123,0.0093889795,-0.1823556125,-0.0574351437,0.4810902774,0.1528856903,-0.0101815434,0.1953223944,0.0905661732,0.0288440809,0.2890706062,-0.0138944872,0.2910845578,0.1026208997,0.1299511641,-0.1687362939,0.0180363432,0.0537560843,-0.0116699208,0.3652868271,0.0965513662,-0.075143531,-0.284778893,-0.3224139512,-0.1540867686,0.1093940064,0.0982109159,0.0685299113,-0.7366599441,0.231404379,-0.2745263577,0.4588467181,0.1567831039,0.1876427829,-0.0823956653,0.5042703748,0.0336672887,0.2031309009,0.7495101094,-0.4833982587,-0.1992151439,-0.0576261915,0.5623437762,-0.4447718859,-0.190942362,-0.185264647,-0.0466107801,0.38808465,0.5814822912,-0.1556289941,-0.2454755455,0.1311356723,0.0130167194,-0.3311719894,-0.1396134645,-0.2438621223,-0.2152126729,-0.0462429821,0.2171622664,0.2212100774,0.1851467937,-0.0994262621,0.0331832506,0.2082673907,0.158976987,0.3634381592,-0.0392948054,0.5451880097,0.0777360797,-0.0865150392,0.1271380782,-0.0555551425,0.4668223858,0.4707154334,-0.3658136725,-0.8872762322,-0.2774435282,-0.5907796025,0.350810051,0.2769891024,0.2057451159,0.0682351142,0.112934418,0.372730881,0.0148341004,0.0234886315,0.1559741944,0.0804019123,-0.5741952658,-0.2932038903,0.1184714288,0.4049559236,-0.1286848933,-0.2283144295,0.3316228092,-0.485101074,0.2146522105,0.0287613682,1.1100389957,0.1928315759,0.0075463187,0.0507378392,-0.5905679464,0.5279996395,-0.2389278561,0.0801340416,-0.4131192267,-0.2406007349,-0.0036555461,0.0536892265,-0.0153252231,0.4041559398,-0.4088608027,0.4386075437,-0.3602978885,0.1145674437,-0.022909632,-0.0635534301,0.4366986454,0.2198300958,-0.0525720716,-0.0685366094,-0.5413904786,-0.2055844367,-0.146025151,0.1214243174,-0.1592824757,-0.1959633976,-0.270105809,-0.2113203406,-0.1561495811,-0.4304659069,0.3617223501,-0.2764850855,0.1328816563,0.4572089314,0.6024870276,0.1135130078,-0.1463374048,0.1434184611,0.0936287269,0.3183327913,0.128956601,0.1119609848,-0.2895762622,-0.0769177899,-0.4365361333,-0.1468356103,0.1348627508,-0.2034775317,-0.568425715,-0.341550827,0.1622770876,-0.2131261081,0.0719181225,0.043410439,0.271058321,-0.0427333862,-0.0344418287,0.0456839502,-0.0594333522,0.1785791665,0.0671230331,-0.3771670759,-0.1731173396,-0.0144643402,0.2468800992,-0.0816213638,0.2710582614,-0.0192070436,-0.1909543425,-0.2025274336,0.2634912431,0.1464578956,-0.5441134572,0.0609948114,-0.1638534963,0.0794604197,-0.0308083091,0.2564077079,0.0040271087,-0.0315976106,0.2597407997,-0.3851787746,0.2269774675,0.1086504087,-0.0326007716,0.1026049703,0.2620751262,0.2181413174,-0.0853399038,0.1288945079,-0.2033515871,0.1636164635,0.088328667,0.2151657045,0.0252732597,0.0324871764,0.1106406301,-0.0860736445,-0.0772574916,0.0030452779,0.0700701624,-0.0435471646,-0.0792809352,0.2268335223,-0.0363675207,-0.1224818304,0.049041979,0.117039144,-0.0786334574,0.0142125925,-0.078560017,0.1819281727,0.3069089651,0.616104722,-0.1188578308,0.1722613722,-0.3997147679,0.0934919715,0.1612648666,0.2738203406,0.2434922755,0.2088905871,-0.0797688216,-0.1504358053,-0.2347466052,0.2914884984,0.477630645,0.441883862,-0.1535138786,-0.4760377109,0.2765363753,0.2433062345,0.0944893807,0.2376839817,-0.3268941939,-0.2415629029,0.3689542413,0.0949162692,-0.0711161643,-0.0249736067,0.225707382,0.036576692,-0.2706282437,0.2700268924,0.3887725174,-0.1036335304,0.0677157715,0.1532880068,0.2603302598,-0.1335884631,0.1977947354,0.0420339704,-0.1860108078,-0.0569822043,0.36569947,0.1309813857,-0.3041418791,0.3884842098,0.3095372319,0.0262841713,0.5513893366,0.0968694538,-0.0718506277,-0.3541102409,-0.0751095265,0.2516711652,-0.229266867,0.1444121599,0.1754047424,0.0736990348,-0.18050991,-0.5274277925,0.195385024,-0.1064903885,-0.1218877733,-0.1460493803,-0.114431344,0.1414320022,-0.0376041681,-0.0730083212,-0.2854776382,-0.0415714234,-0.1823082715,-0.0289493222,-0.3202752471,-0.3320940435,-0.2481919825,-0.1464116126,0.5078229308,-0.2194050252,0.0949153453,0.2847426534,-0.0256363098,0.1872621179,0.1846110374,0.3080628514,0.0011358861,-0.0831549615,0.0417360738,-0.0124912448,-0.0269360077,0.3436032832,-0.0796079487,0.3835317194,-0.1874523759,0.0750447661,0.0862250999,-0.0140716713,0.282384783,0.3320313394,0.5783448219,-0.1813524663,0.2441785038,-0.0180410277,-0.237215057,-0.164609015,-0.0471900627,-0.243540898,0.0256754626,0.3854943514,0.1520514339,0.1211454868,-0.0349456482,-0.0011294144,0.0641206279,0.3271696866,0.5245546103,0.0989328846,-0.0985149518,-0.0012728596,-0.3565157652,0.3220432401,-0.2169582099,0.0042105843,0.0871713534,-0.0041145501,0.0526883341,-0.0861484781,0.0071786083,0.1673326045,-0.1168887615,-0.0456161685,-0.3623970449,-0.0929426625,-0.2473563999,0.2926359475,0.2325423956,-0.1625220329,0.1749672294,-0.056589365,-0.0589161478,0.012455212,-0.0841823667,0.0819656625,0.2315058261,0.2678607404,0.2439595312,0.1575015634,0.0843430087,-0.4018859267,0.1683827192,-0.0069988454,-0.4111431241,0.4559905827,0.0454118326,0.4668777883,-0.2287034392,0.186150521,-0.0069975932,0.0059602708,0.1837860942,0.3315390646,-0.2472690642,0.2213975638,0.3472178876,-0.2395342737,-0.0176324006,0.1237426549,0.1633205265,0.0596294925,-0.3737597764,-0.0125350114,0.3990119994,0.1499468684,-0.0689650178,-0.0541441627,0.4218946099,0.3916333318,-0.1186876744,-0.4830706716,-0.0004910202,0.5178112984,-0.2028841823,-0.3391341865,0.2773770392,-0.0100760786,0.0231584497,0.0059452923,-0.1608887315,0.3616690636,-0.2252886444,0.1217500344,-0.2094773948]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1633","title":"social_i_qa wrong format of labels","comments":"Sure feel free to open a PR thanks !","body":"Hi,\r\nthere is extra \"\\n\" in labels of social_i_qa datasets, no big deal, but I was wondering if you could remove it to make it consistent.\r\nso label is 'label': '1\\n', not '1'\r\nthanks\r\n\r\n```\r\n>>> import datasets \r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\r\n... 'social_i_qa')\r\ncahce dir \/julia\/cache\/datasets\r\nDownloading: 4.72kB [00:00, 3.52MB\/s] \r\ncahce dir \/julia\/cache\/datasets\r\nDownloading: 2.19kB [00:00, 1.81MB\/s] \r\nUsing custom data configuration default\r\nReusing dataset social_i_qa (\/julia\/datasets\/social_i_qa\/default\/0.1.0\/4a4190cc2d2482d43416c2167c0c5dccdd769d4482e84893614bd069e5c3ba06)\r\n>>> dataset['train'][0]\r\n{'answerA': 'like attending', 'answerB': 'like staying home', 'answerC': 'a good friend to have', 'context': 'Cameron decided to have a barbecue and gathered her friends together.', 'label': '1\\n', 'question': 'How would Others feel as a result?'}\r\n\r\n```\r\n\r\n","comment_length":9,"text":"social_i_qa wrong format of labels\nHi,\r\nthere is extra \"\\n\" in labels of social_i_qa datasets, no big deal, but I was wondering if you could remove it to make it consistent.\r\nso label is 'label': '1\\n', not '1'\r\nthanks\r\n\r\n```\r\n>>> import datasets \r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\r\n... 'social_i_qa')\r\ncahce dir \/julia\/cache\/datasets\r\nDownloading: 4.72kB [00:00, 3.52MB\/s] \r\ncahce dir \/julia\/cache\/datasets\r\nDownloading: 2.19kB [00:00, 1.81MB\/s] \r\nUsing custom data configuration default\r\nReusing dataset social_i_qa (\/julia\/datasets\/social_i_qa\/default\/0.1.0\/4a4190cc2d2482d43416c2167c0c5dccdd769d4482e84893614bd069e5c3ba06)\r\n>>> dataset['train'][0]\r\n{'answerA': 'like attending', 'answerB': 'like staying home', 'answerC': 'a good friend to have', 'context': 'Cameron decided to have a barbecue and gathered her friends together.', 'label': '1\\n', 'question': 'How would Others feel as a result?'}\r\n\r\n```\r\n\r\n\nSure feel free to open a PR thanks !","embeddings":[0.0492675304,-0.2175826877,-0.1146280468,0.3597561419,0.0827393383,-0.1826645136,0.0871730521,0.3020369411,-0.1670477092,0.2547192574,-0.2077218443,-0.2840921879,-0.1246122867,0.5941567421,-0.1946665943,0.1424549967,0.1104016677,0.2856855989,0.4830946922,-0.1500134319,-0.2150883228,0.1565583944,0.0123902587,0.2791870236,-0.305372715,-0.1653001457,-0.0955901295,-0.3455423117,-0.1604853719,-0.331695497,0.1878923178,0.3406995237,0.0470067002,0.4007871151,-0.0001234023,-0.1523970664,0.0031859821,-0.1023092717,0.0160015803,0.0187387671,-0.2862668335,-0.2020233274,-0.0972680822,0.222902447,-0.1030854359,-0.5315406919,-0.1165262088,-0.2239706963,-0.0206856057,0.197236836,0.0966224521,-0.2455823272,-0.3549914062,-0.0081001157,0.1493125856,0.2337798029,0.1948210001,0.1643984169,0.0994026512,0.1099623889,0.7370941639,0.7524778843,-0.3016228378,-0.2200663686,0.1416431218,0.0015039262,0.3840678632,-0.4318003356,0.0401714072,0.3197075427,0.8595770001,-0.0675518811,-0.1793605387,-0.133588016,0.1398671269,-0.1493010372,-0.2797861695,0.0350596979,-0.0186483711,0.2289393991,-0.0241802838,-0.2905867398,-0.0740971714,0.1610723287,0.1447565109,0.4783553481,0.0263266694,0.1136759073,-0.3074854016,-0.1269353032,-0.054561533,0.0331953466,-0.3618002832,-0.1357371956,-0.1701066494,-0.0904569328,-0.195728302,0.0018529841,0.0158508644,0.1256081164,0.0572507791,-0.16076608,0.0272233263,-0.1757714897,0.4196959138,0.0221645907,0.3865008056,0.2871238291,0.0461797677,-0.3137209415,-0.156686157,-0.1754077077,0.1210731789,-0.2085587233,0.1048446223,-0.0243082717,0.3228280246,-0.3610476851,-0.1793159395,0.0087279202,-0.5811860561,0.056506563,-0.1170795336,0.1577971876,-0.1790129095,0.3697304428,-0.0539915711,0.1601100266,-0.2399178296,-0.2951287329,-0.1926085353,-0.2407290637,0.0951222852,-0.1706256568,0.0102103231,-0.0234704763,-0.1224563867,0.0313778296,0.1445281655,-0.043277353,-0.2112134546,-0.2344214469,0.3149112761,0.046189528,-0.1469930559,0.3031124771,-0.0837646872,-0.3455062807,-0.2188545614,0.1297385097,-0.3649470508,-0.2317498177,-0.1919657588,0.1740821004,0.2092262954,-0.2640437782,-0.1622005552,0.3280293047,0.23984541,-0.122354053,0.2219696045,-0.3194843531,0.1829773337,-0.0347140878,0.035813123,0.2626775801,-0.4858740866,-0.2344525754,0.2351628691,0.0283233933,0.5021169782,0.6369107366,0.1147554144,0.2140600383,-0.0553338341,0.1085227504,0.3854341805,-0.5723042488,-0.2084782124,0.2587687969,0.1542992443,0.0711279884,0.0906824023,0.1683838218,0.414193064,0.0300964303,0.0776780322,0.1801206619,-0.0378817357,-0.1099722385,-0.3294333518,-0.0980934128,0.2831754088,0.1345637888,-0.2457045615,0.0815357864,-0.2745985389,-0.5596561432,0.2566483319,-0.2206740677,0.0184369814,0.1762643605,0.7755270004,0.4390074015,0.0823443234,0.1032791063,-0.3953659236,-0.0303536374,-0.0948869139,0.2099940926,0.0843557194,-0.4223873913,-0.1347944885,-0.1243720651,-0.1844382435,0.0104303351,0.0495498665,0.0010783013,-0.2413250357,-0.0488821194,0.0157972332,0.517778039,0.1951727271,0.1292706281,-0.0376305245,0.0876956135,-0.1091357991,-0.0192680657,0.0413900949,0.4722794294,0.4014358521,0.1888507456,0.0323939063,0.4012379646,-0.1812178046,-0.0780621767,0.0789962038,0.1964983195,0.0583518073,-0.4091486931,0.0678516701,0.2528767884,-0.211604774,0.0334812403,-0.099763386,0.17983073,0.1561919749,0.0953878537,-0.4194253981,-0.0718524754,0.266789943,0.0315450653,-0.1895132661,-0.1170196086,0.2359025776,-0.522295475,0.0798055977,-0.0602366775,-0.5524230003,0.3089454472,0.3664543033,0.0541738234,0.0270602368,0.1164237857,-0.1554220766,-0.1644230783,-0.0304602589,0.2548182011,0.0490171947,0.1128314883,-0.0930855945,0.0159091037,0.3078115582,-0.1659633666,0.2943143547,-0.105185762,-0.0720913559,0.0804309472,0.2894303501,0.0573670231,0.1701980978,-0.1093950197,-0.4221563935,0.2582370341,-0.4182316065,0.1076513603,-0.1867103875,-0.5037405491,-0.0595207922,-0.6948615909,-0.4230277538,-0.2947793007,0.138970077,-0.2494706064,0.1251666099,0.3119556904,0.4005903602,0.2774753273,-0.2863791585,-0.1974100024,-0.0472028032,-0.2003667653,-0.2647204399,-0.0674997047,-0.2079251558,0.2966348231,-0.0688879862,-0.2936260104,0.0168265942,0.0904686898,-0.3641434312,0.1066294238,-0.0727154985,0.0296140015,0.2317038924,0.1346534938,-0.1741945297,0.1904316396,-0.0100960005,-0.1401831657,0.2369793952,0.1520740986,-0.4376102686,0.0450570807,-0.2669122815,-0.3110018075,-0.6636099815,0.0975837559,0.0393084027,-0.1424746811,0.0208666883,0.2844303846,-0.1426301897,-0.2162969857,0.0383509211,-0.1997513771,-0.1600333452,-0.2897958457,0.2265136093,-0.0132308397,0.0865872875,-0.0595655814,0.1560061127,-0.1151093319,-0.1606010199,-0.5146896243,-0.032458052,-0.0361670218,0.0990679935,-0.2181613594,-0.0381256118,0.1183151752,0.0463773757,-0.012150296,-0.012815428,-0.0607888475,0.2177023441,0.2263573706,-0.0112417908,-0.210613668,0.0526356623,-0.0653119013,0.2096883953,0.3150632977,0.0293807276,-0.1163096502,-0.0231194086,0.4625680447,0.1432843953,-0.1250663102,0.1314727068,0.1274516732,0.0427371487,0.3241705596,-0.0157151725,0.2316337377,0.075024426,0.1733077914,-0.2372737229,0.0474821739,0.07134597,-0.1032160074,0.323079139,0.1511119008,-0.1391716599,-0.219004184,-0.2911585271,-0.0664831251,0.1136574298,0.0878168344,0.0572460778,-0.7075658441,0.2029919773,-0.2292716205,0.457125932,0.0761138722,0.1839648336,-0.1191651598,0.4944078326,0.0110479351,0.2394255549,0.7984461784,-0.4692696035,-0.2774220705,-0.0275517628,0.5898339748,-0.5009109974,-0.1685484499,-0.1483477503,-0.0219929311,0.3954065144,0.5819657445,-0.1212751642,-0.2927293777,0.0714922547,0.0133368634,-0.2968727946,-0.1390559971,-0.2110069692,-0.1730104089,-0.0795374736,0.1535517722,0.2592167556,0.1927838176,-0.0386887304,-0.0255755894,0.1915324479,0.1735647023,0.384005934,-0.0094179306,0.5969157219,0.1028055772,-0.0208487585,0.0739944205,0.0187106356,0.4176024795,0.5068894029,-0.3736259937,-0.8385084867,-0.2634973824,-0.5094601512,0.358494401,0.245862484,0.2262129039,0.0737932548,0.0791352168,0.3238598108,-0.0021326011,0.0422303677,0.1783323735,0.0145489704,-0.5428771973,-0.2960020304,0.062749967,0.3621221185,-0.1319404542,-0.1133226678,0.379150331,-0.430819124,0.2610625327,-0.0462963954,1.1743556261,0.1828978956,-0.0266877841,0.1069278345,-0.5944348574,0.5097916722,-0.1815715134,0.0601803027,-0.4020953178,-0.2302296907,0.0262540951,0.0312317908,-0.109014079,0.4008765221,-0.4283944368,0.423917979,-0.4446191788,0.2541781366,-0.0468180664,-0.0365727097,0.438487649,0.2025996298,-0.054126367,-0.0097287828,-0.5424133539,-0.1681402773,-0.2136386931,0.0887273923,-0.1500101984,-0.1728537977,-0.1941417754,-0.1019403487,-0.2035408467,-0.3995812237,0.3055725098,-0.3100845814,0.0902428627,0.428298533,0.5867320895,0.0686218441,-0.1519067287,0.1764768511,0.1503113508,0.3493874371,0.0502194315,0.0734410137,-0.2112915218,-0.0997003466,-0.4160128534,-0.1802494973,0.1898326725,-0.1684298813,-0.5501136184,-0.3138190508,0.0908476189,-0.2141149789,0.0399908461,0.1171493083,0.3292885125,-0.0812217966,0.0096825287,0.0855085403,-0.0563021414,0.2044134736,0.1017864347,-0.3957956731,-0.1540033221,-0.0030557055,0.2875192761,-0.1080458239,0.297110945,0.0741568804,-0.2029008567,-0.2510919273,0.1808166504,0.2512027919,-0.559864819,0.0774165019,-0.2254369706,0.0591033734,-0.1087489277,0.2665822804,-0.0057874965,0.0113249803,0.2599097788,-0.3497242332,0.2594105303,0.2272412777,-0.0572631545,0.1351504773,0.1914437264,0.1798921227,-0.0370056033,0.1609745771,-0.2522886693,0.148413524,0.1358340383,0.2104094476,0.0060510291,0.0504164509,0.1601724029,-0.1208601519,-0.0360853262,-0.0589548163,0.0367273502,-0.0976446718,-0.038998995,0.2057211399,0.0011582319,-0.1003906876,0.0658493042,0.0559698045,0.0048924726,0.008905231,-0.0555447191,0.1647893786,0.3081882894,0.5039407015,-0.0488125943,0.1316040009,-0.3656660318,0.0739696547,0.255259037,0.2199977636,0.2436835617,0.1244073212,-0.1082055867,-0.2069259286,-0.201015383,0.2829546332,0.4556480646,0.5049985051,-0.1738657951,-0.4445304275,0.2419833839,0.2094389945,0.145646289,0.1555333287,-0.2993703783,-0.1875873506,0.3917114139,0.1564093828,-0.0685902759,0.0185108036,0.1613234133,0.0136810634,-0.240531072,0.22280249,0.3437613845,-0.0376867466,0.0766181648,0.148094371,0.1928062737,-0.0783459619,0.2567575574,0.0305081438,-0.2233735025,-0.0980505496,0.3291513026,0.0619321764,-0.360542804,0.3138459325,0.2692574263,0.0362600423,0.6332076788,0.0358523987,0.0303602759,-0.2294175476,-0.1422742456,0.2493295074,-0.1707580686,0.1328101456,0.1534189433,0.0718001798,-0.1517773718,-0.6201999784,0.2181252986,-0.0600129552,-0.0863846168,-0.1243586019,-0.1976813078,0.1226294488,-0.1099826694,-0.1550379694,-0.2757197917,-0.0980877355,-0.2079045773,0.0103621688,-0.3322546184,-0.3119708002,-0.3043994009,0.0008077544,0.4630056918,-0.2119540572,0.129340753,0.2429623008,-0.0332319327,0.1143078655,0.2529857159,0.3321436942,0.0115827862,-0.122516416,0.0830710605,0.0583351217,-0.0891162381,0.3502973318,-0.0906197131,0.3809944689,-0.2181500047,0.042554792,0.1218439117,-0.0522025228,0.265529871,0.3343238235,0.6018249989,-0.1089960858,0.1572355181,-0.0316123702,-0.2037417591,-0.1436356604,-0.0863300711,-0.336127758,0.0044468096,0.4384143949,0.1850224584,0.1105297431,-0.0524606816,0.023422474,0.0553178713,0.2760488093,0.5201635361,0.2002545893,-0.1342838407,-0.121195063,-0.3335158825,0.3160597384,-0.2906930149,0.0581070669,0.072195448,0.0314369462,0.1385425627,-0.1177740172,0.1014432237,0.1045867801,-0.0863208994,-0.095484592,-0.3825448751,-0.0397346616,-0.3146309555,0.3228771091,0.228508994,-0.1164077371,0.1152494624,-0.0691637099,-0.0268834922,0.0601613,-0.1489181072,0.0806708708,0.1669828147,0.2448673397,0.2325315475,0.1157566831,0.1112831756,-0.3771116138,0.0747936219,-0.1061992496,-0.3484564126,0.439209193,0.0841140226,0.5032382011,-0.187432006,0.211608544,-0.0122201201,-0.0329964086,0.1632774174,0.2964802682,-0.2247528136,0.1982660592,0.3651622832,-0.2293666601,-0.0611922853,0.1720227897,0.0962494537,0.0816758648,-0.3966222405,-0.1229754463,0.4166907072,0.1597227603,-0.1423600912,-0.0685318261,0.4567958117,0.3548553288,-0.0360653885,-0.4897309542,-0.0113639673,0.4418072402,-0.1694203317,-0.2959726751,0.2665332258,0.1306941956,0.0852689147,0.0015672614,-0.2236550897,0.4432822764,-0.1901572943,0.063675493,-0.2016966343]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1630","title":"Adding UKP Argument Aspect Similarity Corpus","comments":"Adding a link to the guide on adding a dataset if someone want to give it a try: https:\/\/github.com\/huggingface\/datasets#add-a-new-dataset-to-the-hub\r\n\r\nwe should add this guide to the issue template @lhoestq ","body":"Hi, this would be great to have this dataset included.\r\n\r\n## Adding a Dataset\r\n- **Name:** UKP Argument Aspect Similarity Corpus\r\n- **Description:** The UKP Argument Aspect Similarity Corpus (UKP ASPECT) includes 3,595 sentence pairs over 28 controversial topics. Each sentence pair was annotated via crowdsourcing as either \u201chigh similarity\u201d, \u201csome similarity\u201d, \u201cno similarity\u201d or \u201cnot related\u201d with respect to the topic.\r\n- **Paper:** https:\/\/www.aclweb.org\/anthology\/P19-1054\/\r\n- **Data:** https:\/\/tudatalib.ulb.tu-darmstadt.de\/handle\/tudatalib\/1998\r\n- **Motivation:** this is one of the datasets currently used frequently in recent adapter papers like https:\/\/arxiv.org\/pdf\/2005.00247.pdf \r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nThank you","comment_length":29,"text":"Adding UKP Argument Aspect Similarity Corpus\nHi, this would be great to have this dataset included.\r\n\r\n## Adding a Dataset\r\n- **Name:** UKP Argument Aspect Similarity Corpus\r\n- **Description:** The UKP Argument Aspect Similarity Corpus (UKP ASPECT) includes 3,595 sentence pairs over 28 controversial topics. Each sentence pair was annotated via crowdsourcing as either \u201chigh similarity\u201d, \u201csome similarity\u201d, \u201cno similarity\u201d or \u201cnot related\u201d with respect to the topic.\r\n- **Paper:** https:\/\/www.aclweb.org\/anthology\/P19-1054\/\r\n- **Data:** https:\/\/tudatalib.ulb.tu-darmstadt.de\/handle\/tudatalib\/1998\r\n- **Motivation:** this is one of the datasets currently used frequently in recent adapter papers like https:\/\/arxiv.org\/pdf\/2005.00247.pdf \r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nThank you\nAdding a link to the guide on adding a dataset if someone want to give it a try: https:\/\/github.com\/huggingface\/datasets#add-a-new-dataset-to-the-hub\r\n\r\nwe should add this guide to the issue template @lhoestq ","embeddings":[-0.3177589476,-0.2620921433,-0.0742940754,-0.1360726207,-0.1152319089,0.1707084626,0.3080384135,0.203022629,-0.2397362143,0.0552949868,-0.2352134734,0.4752635658,-0.0952458978,-0.1193298921,0.301674366,-0.4320202768,-0.0477607138,0.1427704841,0.0484669209,-0.1363328099,0.0759682804,0.1880860329,-0.1696922183,-0.2604604363,-0.26458776,-0.0204983279,-0.1806991994,0.1152987406,0.0702264085,-0.3448495269,0.0981593132,0.1290148497,-0.1533488035,0.0382176004,-0.0001020611,-0.2599645257,0.3763659894,-0.0408452936,-0.0781309307,-0.0949171036,-0.1256621331,0.2132274508,-0.1509781778,-0.0368380174,-0.2771830857,0.0343147777,-0.0654125214,0.1563650817,0.3473089933,0.234363243,0.2541932762,0.143404454,0.1492129564,-0.2305727601,0.0840076879,0.3737621307,-0.1466037333,0.0393851176,0.4819653332,0.3865528703,-0.1464485079,0.3024272323,-0.0252901055,-0.0557186455,0.035598468,-0.1498378515,-0.1170112938,-0.0455061756,0.0358292311,0.2339223176,0.5025715828,-0.0695720688,-0.2430479676,-0.4196177125,0.1549947262,0.2311789393,-0.0697871447,0.0160776209,-0.1710774153,0.255038619,0.1053161398,-0.0129054748,-0.4324845672,0.0423391461,-0.0131294001,0.328864187,-0.1389487833,-0.1893429905,0.0843321234,-0.1206893176,-0.131541267,-0.1259637773,0.2005767226,0.0272814799,0.1782055199,-0.2032244802,0.2273506224,0.0372804105,0.3695320189,-0.3269723654,0.1083113775,0.0024058116,-0.0184228383,0.1728197038,0.175915882,0.0522628836,0.0287914742,-0.1338439733,-0.0270637106,0.0693990886,-0.0369082205,0.2978389561,0.039893344,-0.0069525167,-0.3751886487,-0.0728975758,-0.0188814737,-0.5017318726,-0.3217943609,-0.1345870644,0.2176480889,-0.0170383509,0.0088497866,0.1097394004,0.1124256328,0.1390443444,-0.0787663013,0.1546752453,-0.0838784724,-0.2307536751,-0.1956539899,0.1075070053,-0.2300258875,-0.0905876905,0.4115203917,-0.1559683532,0.2945146859,0.0293781217,0.2517844141,0.1479529738,-0.1327795535,0.2535916865,0.1956558377,-0.0287257694,-0.1515611708,-0.2560281456,0.1043768451,-0.0417254828,-0.3461176455,0.0345863402,0.0430074222,-0.1798159033,-0.2725968361,0.2764825225,-0.558665514,-0.4412409663,-0.0476597436,0.5198410153,-0.1433010399,-0.1662922502,0.0136279296,0.2717223465,-0.5072974563,-0.0581072234,-0.1543831676,0.4345178902,-0.1015696079,-0.4384256601,-0.0022473237,-0.0258314516,0.0073461914,0.1587067097,-0.1857937723,0.0062166159,-0.1964743286,0.1717140228,0.2090325654,-0.1932772398,-0.1742429733,0.2781268954,-0.2839746475,0.1765062064,0.2557897568,-0.0770390034,0.2923836112,-0.1722425073,0.4016663134,0.4029042423,-0.0046590436,0.015304774,-0.2578356266,-0.1196821705,-0.2878045738,0.1495059431,-0.1504549384,0.1132013947,0.2190835625,0.0312549248,0.4210237265,-0.1299681664,-0.0844264105,0.3525913358,0.2251746953,-0.3548794389,0.0616183951,-0.4081225395,0.0213565994,-0.1696057618,-0.2275652885,0.3027378917,0.0940209255,-0.1567711234,-0.2218085676,-0.0960050374,0.0356404632,-0.2785292268,0.2104332,-0.1456791162,0.0542766117,0.0303602964,-0.0065948116,0.2362563908,-0.0733030885,-0.088628307,-0.2577840984,0.1787191629,-0.0821367949,-0.1371123046,0.4384849966,0.5372505188,-0.1583573669,0.0528508313,0.0140156904,-0.2217489481,0.0223165248,0.3042853773,0.0700785294,0.2915846705,0.3826595843,-0.436252743,0.2204376906,-0.2790266871,-0.1133593693,0.0008918793,-0.4453908503,0.4637394547,0.0930498242,-0.1239488646,-0.0855586305,0.0722268149,0.014726162,-0.1199679002,-0.192541182,-0.356274128,0.5151850581,0.1732232422,0.1354529709,0.1301218122,-0.0516964048,0.2454971522,0.0071555288,-0.025641907,0.115761742,0.3170927167,-0.382230252,-0.2852944732,0.2455774397,-0.277792424,0.31807217,0.5747028589,0.1599914283,-0.0199756995,-0.1691667438,-0.2188318372,0.0733575895,0.0193080604,0.0886423662,0.1600494683,0.4310274422,0.1673225164,-0.5958175659,0.0126943914,-0.1276692748,-0.0557531975,-0.1462134421,-0.1251993179,-0.1335916817,-0.1093116626,-0.162848711,0.2715628445,0.1880187988,-0.2080157548,0.1844824404,0.0354403667,-0.0365183651,0.3215628862,-0.0671000108,0.3225325644,-0.2832520008,-0.0594839305,-0.0153339291,-0.2736524642,-0.1572203934,0.3251060545,0.0364069194,0.1241931394,0.3659246564,-0.1462062746,0.221101895,-0.211540252,-0.7749061584,0.2080138773,-0.1193342358,0.3887394965,0.3073036373,-0.3305588365,-0.1670753509,-0.180423364,0.0694085434,-0.0857750103,-0.2479082644,0.0334359519,-0.5105165243,-0.2646910548,-0.0085194213,-0.2055128962,0.0208577923,-0.4457691014,0.4158594012,-0.1666579843,0.0813387632,0.4588474035,0.0284741167,0.1281661093,-0.3406171799,0.4250284731,-0.2878658473,-0.1325178146,0.1804171652,-0.3023087382,-0.3168964088,-0.3453714848,-0.1991608888,0.0902965069,-0.1152852774,-0.242961809,-0.4526661336,-0.0250471421,0.1881505996,0.0353634581,0.0056605861,0.116214931,0.0669867694,-0.2145439833,-0.1214645728,-0.091514729,0.0681779012,0.2476431727,0.1903648674,-0.1128423885,-0.1599102318,0.2046280205,0.1031314358,0.2823708355,-0.1871026754,0.1593344361,0.1396972537,-0.0036140417,-0.0806222931,-0.235210225,0.4480724633,0.0809463486,-0.0335974991,0.5477821231,0.197495386,-0.3934002221,-0.0771774426,-0.3385390937,-0.1824477464,-0.1524591446,0.1273847371,0.0976106375,0.2027918696,-0.0670545474,-0.4088734686,-0.4249365628,-0.2789643109,0.2690777481,0.298563689,0.2101811916,0.0120691778,0.0173096582,0.3822860122,-0.3703687191,0.2534401715,0.1584357023,-0.1759842485,-0.4221259952,-0.1351892352,0.0388307795,0.1601971239,0.5583922863,-0.1225711182,-0.1864280403,-0.00815967,0.2202051133,-0.2568174899,0.2670732737,-0.2381974012,-0.1835955828,0.2431992888,0.0195286274,-0.2546620667,0.0663005635,0.458640188,0.0561383925,0.1179872379,-0.5310832858,0.0742273033,-0.056055773,-0.3493674695,0.0451166257,0.2192827016,-0.019862989,0.1994464695,0.0001767818,0.0362888314,0.1676735878,0.3099418283,0.1876984388,-0.0515293553,0.1045764461,0.2032250315,-0.0672683418,0.0847293586,0.0891208127,0.310158819,-0.1134526655,-0.193609491,-0.0442448854,-0.1725541055,0.2790647447,0.2448085397,0.3807350099,0.0686099753,-0.1615297496,-0.1260262281,-0.2227454782,-0.1062129512,-0.1079663336,0.1105871052,-0.06545908,-0.0359881595,0.6073560119,0.1287794262,-0.17757909,-0.0960507095,0.2519664466,-0.2140782624,0.520673573,0.1697616428,0.7450040579,0.1591945589,0.2997280657,0.20001553,0.013876535,0.3534680009,0.152964592,0.0292008501,0.0341241807,-0.0971202999,-0.0013617951,-0.1323670745,0.1203544736,0.1660012603,-0.3217508495,-0.1649725437,-0.0383109152,-0.2365002334,0.2211611569,0.2023329288,0.0448926315,-0.3651108444,-0.4240857661,0.2469709367,-0.1343730539,0.1706786007,0.0380583555,-0.2022247016,-0.0429292507,-0.031007776,-0.5856846571,0.055348929,-0.0752944797,-0.0777065381,0.2622393966,-0.1873436421,0.5644802451,0.1807140261,0.3590945005,0.0280325338,-0.2775943279,0.2399737835,0.0437060222,-0.0789766908,0.1165658608,-0.0578531362,0.5153161883,-0.2004475147,-0.1924440116,-0.0860133022,0.1779169142,-0.1607517004,-0.268941313,-0.068569921,-0.2105764747,-0.0977662727,-0.2652698457,0.205113396,0.0032294455,-0.3100395799,0.181535989,0.2395736575,-0.2013947666,0.2902668417,0.240561083,-0.1960428059,-0.0010216059,0.3272322416,0.1324819028,-0.5037202239,0.2505846024,0.3955571651,-0.1932807267,-0.2894613445,-0.3408525884,-0.1844338477,-0.3201626539,-0.0685250387,0.2198492438,-0.2964031994,0.0162885115,0.2609229684,0.336414367,0.1512771398,0.0417309925,-0.0908873528,-0.2304932326,0.1080990136,-0.1421355903,0.1311363876,-0.0238789562,0.1346590966,0.0977774784,0.148961693,-0.4810189307,0.052830901,-0.017637033,-0.1355933249,-0.0027507762,-0.3389395177,-0.0822087973,-0.1554580778,0.2025390714,-0.0650407225,-0.1779796332,-0.3447924256,-0.2345919609,0.0986318812,-0.080866456,0.0406872109,0.0841215327,0.0809598938,0.0970656201,0.0801882744,0.3145186603,0.4301785529,0.0809029415,0.0007431479,-0.0957488045,-0.2233079523,0.3939963281,0.0598546527,0.0491617285,-0.0076550022,0.1516756266,0.0278847404,-0.0662032813,0.0025705688,0.1824302226,0.4090398252,0.2144497186,0.1624259353,0.0095827291,0.1689972579,-0.1434502751,0.4485791624,-0.1726193577,-0.2681815624,-0.0173277836,-0.1098328382,-0.1975366324,0.3261049092,-0.4386687279,-0.1557978541,0.0030846717,0.0961803123,0.1537089944,0.1624153703,0.4235972464,0.109748818,0.1121354997,0.1038233489,0.348664254,-0.2083105594,-0.0960369781,-0.037378639,0.251318723,-0.1241373569,-0.1307868958,0.3298060596,0.3460961282,-0.2174161673,-0.1143987924,0.1837361008,-0.0592028238,0.2494989187,0.455999732,0.1129975915,0.2678765655,0.2334602326,0.1402200758,0.3024090528,0.1312415749,0.4737325311,0.3824749887,-0.1872193962,-0.0622899123,0.2784100175,0.0341414064,-0.2236101776,-0.311796844,-0.1966885626,-0.3885117769,-0.3081871867,-0.0887898207,0.0542945676,0.2801856995,-0.0539484732,0.1133825853,-0.2112012208,0.0237370972,0.1791402102,0.2579673231,-0.1199946478,0.2507947683,0.1295731068,0.0075703892,0.3768984377,0.4593679905,0.4203956723,-0.1359322071,0.0292555466,0.0508754626,-0.4598286152,-0.0823909342,-0.144973889,0.2034954131,0.1862889528,-0.0845372602,0.3650686741,0.2170902044,-0.2189106047,0.1119980514,-0.0321333483,-0.3523701131,0.0021477102,-0.0276169106,-0.0794108361,0.0214131996,-0.1064477637,0.2010622174,-0.2225444913,0.1151829064,0.284294486,-0.2529411018,-0.0244949516,0.1907769144,0.1436193287,-0.0637636408,0.0968323722,0.6004698873,0.0889655203,-0.0785521418,-0.0010809513,-0.3615006506,0.1615792215,0.1743486077,-0.0473132692,0.0898526087,-0.0591105185,0.3126476109,0.1648336649,-0.1453869045,-0.0824154094,-0.280449748,-0.0088792192,-0.4129318595,0.2215529233,0.3215803504,0.0825803056,0.0092245406,0.155118987,-0.1830786467,0.2540987432,0.1466398388,-0.4447722733,0.0790033042,-0.0705621094,-0.2315805107,-0.1591315567,0.1449802667,0.3741682172,-0.2030189186,-0.1047645807,-0.0694706887,-0.4452858567,-0.0914897695,0.1806924343,-0.0444985144,0.2864478827,0.0250813998,-0.3154509068,0.4350380898,0.0454146117,0.0664860383,-0.0794602558,-0.4980970025,0.4053112268,-0.0995938405,0.1716542542,-0.0001339865,0.2879062593,0.154643923,0.0581428893,0.0973362848,-0.5477496386,0.4831718802,-0.0398250297,-0.1598912179,0.0188673772,0.1734883189,-0.1272482872,-0.0517871231,-0.7925340533,-0.1058034152,0.2655445337,-0.0639549345,0.065486677,0.2744801641,0.2028084844,-0.2423504293,-0.1427162737,0.2173300833,0.0030860673,0.0426199585,-0.1558924764,-0.1839764416]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1630","title":"Adding UKP Argument Aspect Similarity Corpus","comments":"thanks @thomwolf , this is added now. The template is correct, sorry my mistake not to include it. ","body":"Hi, this would be great to have this dataset included.\r\n\r\n## Adding a Dataset\r\n- **Name:** UKP Argument Aspect Similarity Corpus\r\n- **Description:** The UKP Argument Aspect Similarity Corpus (UKP ASPECT) includes 3,595 sentence pairs over 28 controversial topics. Each sentence pair was annotated via crowdsourcing as either \u201chigh similarity\u201d, \u201csome similarity\u201d, \u201cno similarity\u201d or \u201cnot related\u201d with respect to the topic.\r\n- **Paper:** https:\/\/www.aclweb.org\/anthology\/P19-1054\/\r\n- **Data:** https:\/\/tudatalib.ulb.tu-darmstadt.de\/handle\/tudatalib\/1998\r\n- **Motivation:** this is one of the datasets currently used frequently in recent adapter papers like https:\/\/arxiv.org\/pdf\/2005.00247.pdf \r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nThank you","comment_length":18,"text":"Adding UKP Argument Aspect Similarity Corpus\nHi, this would be great to have this dataset included.\r\n\r\n## Adding a Dataset\r\n- **Name:** UKP Argument Aspect Similarity Corpus\r\n- **Description:** The UKP Argument Aspect Similarity Corpus (UKP ASPECT) includes 3,595 sentence pairs over 28 controversial topics. Each sentence pair was annotated via crowdsourcing as either \u201chigh similarity\u201d, \u201csome similarity\u201d, \u201cno similarity\u201d or \u201cnot related\u201d with respect to the topic.\r\n- **Paper:** https:\/\/www.aclweb.org\/anthology\/P19-1054\/\r\n- **Data:** https:\/\/tudatalib.ulb.tu-darmstadt.de\/handle\/tudatalib\/1998\r\n- **Motivation:** this is one of the datasets currently used frequently in recent adapter papers like https:\/\/arxiv.org\/pdf\/2005.00247.pdf \r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md).\r\n\r\nThank you\nthanks @thomwolf , this is added now. The template is correct, sorry my mistake not to include it. ","embeddings":[-0.2550905347,-0.2173558474,-0.0847493783,-0.2169910371,-0.0947980359,0.1451424211,0.367269069,0.2097543478,-0.2897710204,0.0634276718,-0.166504547,0.4542146325,-0.02984377,-0.1806160808,0.2641555369,-0.4156152606,-0.0277210958,0.2105742842,0.049068287,-0.1246545091,0.0629920959,0.2326800078,-0.1302823275,-0.2014721781,-0.2289610505,-0.0196438804,-0.1739866883,0.147197783,0.0351302736,-0.2967180908,0.1124121696,0.0916619152,-0.1418098062,0.0146097736,-0.0000990571,-0.2117740959,0.3862992227,-0.0606792867,-0.0464645252,-0.0332469381,-0.1312147081,0.1633898914,-0.0849606618,-0.0597518794,-0.2971861362,-0.1249030158,-0.0628761426,0.197969079,0.3083144724,0.2179932445,0.2830431759,0.0412785523,0.1532015353,-0.2231764495,0.0542273745,0.372484535,-0.1918721348,-0.0373891182,0.5217646956,0.3527919054,-0.0901430324,0.3414720893,-0.0033210132,-0.0511504635,0.0003618104,-0.1810986549,-0.0129327364,-0.0153201716,0.0367778353,0.1964602768,0.5437291861,-0.0190001447,-0.16203475,-0.3095144033,0.1646744311,0.2289224118,-0.0552968979,0.0089115892,-0.1429760158,0.2687635422,0.15167059,-0.0105946213,-0.4237985611,0.0962055102,-0.0234361459,0.2819739878,-0.106898509,-0.2380457371,0.0380231366,-0.1212884784,-0.0602225624,-0.0872784778,0.1236409917,-0.0316625461,0.1669944823,-0.1671116948,0.2100124955,-0.0729077309,0.3489116728,-0.2743950486,0.1291020364,-0.0405261442,0.0396197177,0.2048200816,0.1561380178,-0.0048238346,-0.0407022797,-0.0937014818,-0.0361111201,0.1035664827,0.0218055174,0.2918518782,0.0199267827,-0.1053200364,-0.3434861898,-0.0291532483,0.0300549306,-0.453024894,-0.2599156499,-0.138784036,0.1700324714,0.0228823945,-0.1020828262,0.0902135223,0.1458735764,0.1284774244,-0.0262469985,0.1271075457,-0.0804239884,-0.2611059844,-0.2323881686,0.1441008896,-0.2995675206,-0.1328026503,0.3934412897,-0.1138909385,0.3081302941,0.0744938627,0.2730951309,0.1036646888,-0.0693196133,0.2856170535,0.2269999236,0.0151602514,-0.0742057189,-0.1895555556,0.1480439603,-0.0515319332,-0.2750719786,0.1307035685,-0.0398677699,-0.1713178456,-0.2809180319,0.2983584404,-0.4681397676,-0.3936270177,-0.0139237782,0.4882874787,-0.084912844,-0.193204537,-0.0053284918,0.2781386375,-0.454872489,-0.0543081351,-0.1208577752,0.4197540581,-0.1798436493,-0.3813486099,-0.0347594246,-0.0119616352,0.1090534851,0.2143625766,-0.2166351825,0.0478458963,-0.2560178339,0.1200586483,0.2329182476,-0.1001328602,-0.1621562541,0.301857084,-0.2582181394,0.1179015785,0.2696000338,-0.1470792145,0.3066031039,-0.1011559218,0.3036890924,0.3539858162,0.0040287641,0.0786729977,-0.2685270011,-0.0639335662,-0.2409072667,0.0782698691,-0.1772473156,0.0890614465,0.1935023814,0.0182819832,0.4311459064,-0.1252668053,-0.0860974044,0.4274209142,0.2498184294,-0.407520622,0.1211673245,-0.3079815805,0.0360293724,-0.2448688149,-0.2205982506,0.3174855411,0.1179848164,-0.0811023712,-0.2976610661,-0.1536769867,-0.0206593387,-0.3938336372,0.2214022428,-0.1008376777,-0.0391285159,0.0101959473,0.064515911,0.1718185097,-0.1350694299,-0.1406488866,-0.2126213908,0.1825343221,-0.1382214129,-0.1204563007,0.4487544,0.4845866263,-0.1282463521,0.0805834457,-0.0232475214,-0.2206363082,0.119718574,0.2751038671,0.0227779765,0.2776458561,0.3360043466,-0.5113830566,0.1997106075,-0.2488138527,-0.1310189366,-0.0479061902,-0.4224360883,0.4277729094,0.1416531652,-0.1345336586,-0.0911373869,0.0684768483,0.0927173272,-0.1663668901,-0.1857132316,-0.3488000333,0.4322529435,0.2175365984,0.1175949126,0.1099050641,-0.0720189214,0.2330865562,0.0871167853,0.0113648688,0.1306214333,0.2664994299,-0.3690463006,-0.3079594076,0.2083461136,-0.2946323156,0.274823159,0.5568578243,0.1525157094,-0.0265704282,-0.1436187327,-0.2622219324,0.1515434235,0.0207322557,0.0483719707,0.2199467272,0.4446839392,0.0752179101,-0.5359095931,0.058967486,-0.0868408009,-0.0623485073,-0.1661193818,-0.1392113268,-0.1426318288,-0.1033686027,-0.1718831956,0.2552261651,0.2509300113,-0.2508099675,0.1935921013,-0.0319356211,-0.0621485524,0.4448681474,-0.0399218835,0.2686149776,-0.2233770043,-0.0599934272,-0.0297684092,-0.2206225693,-0.205453679,0.3607836366,-0.031765271,0.0920393169,0.3115673065,-0.1617511064,0.2578003109,-0.2807171941,-0.7797292471,0.2488978803,-0.0763469785,0.3654778898,0.3032671809,-0.2814694047,-0.1847647876,-0.1216378212,0.1028083041,-0.0673624799,-0.2881733179,0.1080121771,-0.5512277484,-0.2237079293,-0.0371234976,-0.2512822151,0.0282511935,-0.445605129,0.2751121521,-0.1886967719,0.0891922414,0.4735910594,0.0987272784,0.0626444966,-0.3358846009,0.3852186203,-0.2408429831,-0.1512866616,0.2256619036,-0.3280216753,-0.3354620934,-0.2680983245,-0.1632638723,0.0857123509,-0.1484076232,-0.2060193717,-0.385687232,0.0287413765,0.1281868964,0.012660386,0.0350101404,0.1469385922,0.0528247729,-0.2421970367,-0.1627655774,-0.0361217782,0.055664517,0.122708343,0.2691582143,-0.1286115199,-0.1232236177,0.1824790239,0.1725967079,0.277220279,-0.1674591005,0.1601112634,0.1729978323,0.0230118036,-0.108789973,-0.2326396853,0.4022704959,0.1066690311,-0.0373140797,0.5552401543,0.190486148,-0.448520273,-0.0446929298,-0.2731250525,-0.2130150646,-0.1309690922,0.1554552317,0.1348466873,0.2547484636,-0.0432786942,-0.3884185553,-0.3537809849,-0.2758013904,0.2993523777,0.3802369833,0.2101033628,0.0178511553,0.0381271765,0.4500536323,-0.3387806118,0.2127801925,0.0977317318,-0.1708395332,-0.3681964576,-0.1383675188,0.0188468024,0.1710695326,0.5239738226,-0.1302961707,-0.1656358093,0.039746888,0.2249462903,-0.182634294,0.1840699017,-0.2657267153,-0.151166141,0.3482765853,-0.0605064072,-0.2417074442,0.0755711719,0.3531863689,-0.0065203458,0.0821735486,-0.5406859517,0.0463679545,-0.0891818032,-0.3288062513,0.0031811388,0.2573785484,-0.0292349365,0.1765136421,-0.0873762146,0.0234758351,0.1436362267,0.2640921772,0.1392315626,-0.0146644469,0.0631770864,0.1693930924,-0.1259153038,0.1562687606,0.0881801546,0.2713107169,-0.196443975,-0.1842610538,0.0467689149,-0.247399658,0.3068912029,0.2026403546,0.3275253475,0.0152703281,-0.2273040712,-0.1682826579,-0.1677910686,-0.1398306638,-0.0610210709,0.1291090399,-0.0080484366,-0.1244053021,0.507653296,0.1256543249,-0.1580706537,-0.0865477845,0.2473796606,-0.222559154,0.4721938372,0.1624655426,0.7175643444,0.2077745944,0.2980909944,0.22703816,0.0330809765,0.3603802323,0.1263258904,0.0329847224,-0.0445428863,-0.124085702,0.011043326,-0.1848861277,0.1366097182,0.1808668226,-0.2793701291,-0.1187781394,-0.0766880438,-0.2133894563,0.2605491579,0.1746497601,0.1605074108,-0.3124157488,-0.3768992126,0.2605460286,-0.1239089072,0.0777714178,0.0038375692,-0.1890075505,-0.0612027943,-0.066235885,-0.5578238964,0.0630300641,-0.1044206992,-0.0495284125,0.2819229066,-0.2459743619,0.608746171,0.1159588769,0.2833862603,0.061880067,-0.2882776558,0.2918896973,0.1036826074,-0.0951202437,0.024364261,0.0019365547,0.529743135,-0.2225661278,-0.193957746,-0.0920378417,0.113076292,-0.1554878801,-0.303040415,-0.1045726165,-0.1748305708,-0.1785483062,-0.2851916552,0.1601521224,0.0173597001,-0.3412758112,0.2080582976,0.2593668401,-0.1455940753,0.2571590245,0.2816723287,-0.1585880369,0.0449749865,0.2441085428,0.1089026108,-0.4613351524,0.2977910638,0.4687363207,-0.2047411054,-0.3268458545,-0.3245159686,-0.1668380648,-0.2777350545,-0.0971876681,0.1831482649,-0.2413516492,0.0488648899,0.2636703551,0.3334893584,0.1942250431,0.0276737902,-0.1031775773,-0.1991484463,0.0799935311,-0.2333468795,0.1095084399,0.0106014265,0.1481612474,0.0906601027,0.1369414628,-0.509827137,0.0013321673,0.0321966745,-0.1800955683,-0.019320542,-0.3365171254,-0.0585694723,-0.1174954623,0.2156559974,-0.0507016182,-0.1558035165,-0.3780556023,-0.1990741789,0.0961516947,-0.0790365413,0.0357509442,0.1184432134,0.0634322613,0.0410715677,0.1009868532,0.2178911269,0.3117172122,0.0949667543,0.0490227081,-0.1454351097,-0.281373024,0.3790169954,-0.002232885,0.0324597545,0.0060734618,0.0834418237,0.0164492298,-0.0347172543,-0.0203404296,0.1815967709,0.4051027596,0.1678255945,0.2308509052,0.0203474909,0.1641140729,-0.1045196727,0.3807808757,-0.1927611828,-0.3170745373,-0.0597017035,-0.1339392066,-0.2030964047,0.353492856,-0.4849466383,-0.1797751337,-0.0094665913,0.1785134971,0.1269884855,0.121398218,0.4179604948,0.0471645668,0.0809601992,0.0976384506,0.3344087005,-0.2412098795,-0.1142839342,-0.0048665935,0.227893427,-0.1108392999,-0.1188251227,0.3489711583,0.2673996389,-0.1727413982,-0.0244318713,0.1487492919,-0.0130890543,0.2514117658,0.3952703476,0.0530291237,0.2509377301,0.2094529271,0.1447479129,0.24467507,0.1645479947,0.4418738484,0.5188907981,-0.0920517221,-0.047869686,0.3086419404,-0.012114441,-0.2144568861,-0.3662442267,-0.155716911,-0.3725658953,-0.3333562315,-0.1056345701,0.1067121401,0.2752205729,-0.0830851123,0.1189955994,-0.2257290184,0.0540933684,0.2005104274,0.2216286808,-0.1693561077,0.2730742991,0.1126309559,-0.0113197481,0.3106688261,0.4625347555,0.40745309,-0.13555035,0.0733168572,-0.0245073624,-0.4439842999,-0.1523750126,-0.1230384037,0.1964580417,0.1910107136,-0.1391980946,0.3820296526,0.221654132,-0.2278256565,0.1216119155,0.0571626984,-0.3473909795,0.0289457142,0.0110182455,-0.0861840546,0.0219264701,-0.0924543589,0.2301411927,-0.222368896,0.0538558252,0.3591844738,-0.1849939525,-0.0838169903,0.1405888498,0.1608646363,-0.0925200731,0.0957496464,0.5441216826,0.085520871,-0.0714193508,-0.057641279,-0.3061354458,0.2435710579,0.1159412637,-0.0556581356,0.0707347915,-0.0005549128,0.3331015408,0.1093010455,-0.0713380724,0.0372496024,-0.2677500844,0.0354900807,-0.4806912839,0.1615231931,0.2575652599,0.098780252,-0.0019961367,0.1756018996,-0.1971773058,0.2820112705,0.1982287914,-0.4632117748,0.0152611323,-0.0114713004,-0.1330865622,-0.1617699116,0.2044117153,0.3616984785,-0.2048712224,-0.1179869026,-0.0483480766,-0.4750306606,-0.1003059074,0.2863422334,-0.0007374223,0.2644278705,0.051846683,-0.3618206084,0.3571457267,0.0211622715,0.0354206152,-0.1531350613,-0.5680055618,0.4435119331,-0.0379807092,0.1529134512,0.0000779166,0.2405637801,0.0878383294,-0.0247105714,0.0571204834,-0.6042075753,0.4317805767,-0.0480581224,-0.1693988591,0.0191914923,0.1876175106,-0.1691665351,0.0348274149,-0.7300879359,-0.1169720665,0.2655218244,-0.0394160971,0.092452623,0.2321476936,0.2379316241,-0.1249122247,-0.1324487031,0.2693978548,0.0733239725,0.029936051,-0.2465262115,-0.2017636746]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1627","title":"`Dataset.map` disable progress bar","comments":"Progress bar can be disabled like this:\r\n```python\r\nfrom datasets.utils.logging import set_verbosity_error\r\nset_verbosity_error()\r\n```\r\n\r\nThere is this line in `Dataset.map`:\r\n```python\r\nnot_verbose = bool(logger.getEffectiveLevel() > WARNING)\r\n```\r\n\r\nSo any logging level higher than `WARNING` turns off the progress bar.","body":"I can't find anything to turn off the `tqdm` progress bars while running a preprocessing function using `Dataset.map`. I want to do akin to `disable_tqdm=True` in the case of `transformers`. Is there something like that?","comment_length":39,"text":"`Dataset.map` disable progress bar\nI can't find anything to turn off the `tqdm` progress bars while running a preprocessing function using `Dataset.map`. I want to do akin to `disable_tqdm=True` in the case of `transformers`. Is there something like that?\nProgress bar can be disabled like this:\r\n```python\r\nfrom datasets.utils.logging import set_verbosity_error\r\nset_verbosity_error()\r\n```\r\n\r\nThere is this line in `Dataset.map`:\r\n```python\r\nnot_verbose = bool(logger.getEffectiveLevel() > WARNING)\r\n```\r\n\r\nSo any logging level higher than `WARNING` turns off the progress bar.","embeddings":[-0.345834583,-0.3181484342,-0.0507016964,-0.1878176481,0.3254140317,-0.0002219211,0.2311783284,0.1533992589,-0.3758892715,0.2149729431,0.1782131493,0.6198928952,-0.2077849209,0.0813507065,-0.1250561923,0.0541477911,-0.0937978029,0.2880425155,-0.4447112978,0.1668906212,-0.1205287501,-0.0268635228,-0.4199817181,0.1133120134,-0.098049596,-0.0874668136,0.4577662945,-0.2100177258,0.1111780331,-0.5028156042,0.1255543977,0.5249206424,-0.1022467762,0.2963654995,-0.0001341136,0.024378268,0.3334862888,0.1161456704,0.0576809719,0.2084132433,-0.3663836718,-0.3946353495,0.3381592035,-0.1342763007,0.0732083544,-0.172539711,-0.2612757683,-0.5232918262,0.4219367802,-0.0776959732,-0.0365383662,0.503811419,-0.467764616,0.5086688399,-0.0745022893,0.0810276791,-0.3309802115,-0.0999305546,0.5109071136,-0.0141434502,-0.2387829423,0.3522944152,-0.2959550023,0.1542491317,0.2584664524,-0.0826860368,0.605276823,-0.5217351317,0.2537859976,0.1346681714,0.5032689571,-0.5142654777,-0.103938736,-0.1459460407,-0.0485781766,0.0544987172,0.1487227976,-0.3223780096,-0.3646895587,0.1625011861,-0.9887856245,-0.3662439585,-0.0918549299,-0.0601705424,-0.1202919558,0.0436481424,-0.0014366576,0.0460398644,0.1100329161,-0.0151390852,0.2032525688,0.0090326387,-0.0887708217,0.2580164075,-0.255309701,-0.3512195051,0.1488700509,-0.0420721844,0.0288523752,0.367156446,-0.1467290372,0.2798498869,0.0295170005,0.1189122275,0.1812521815,0.0157953911,0.4012874961,-0.0109180687,0.3395793438,-0.4546201527,0.050877247,0.3440840244,0.4707216322,-0.5852707028,0.575140059,0.3569725454,0.2022000998,-0.109085694,0.1236839592,0.0169166792,0.2879556119,-0.0798742622,0.2036167681,0.0256176069,0.267922312,-0.1138722375,-0.2338978052,0.0312828198,-0.1787957549,-0.1378449798,0.1107528955,-0.2916526496,-0.2107667178,0.1103599146,0.0275119357,0.2005032748,-0.0148581173,0.0525610186,0.0456209332,0.0795346051,0.452668488,-0.0510114431,0.5489702225,0.622191608,-0.2569610476,0.1888652593,0.0575466119,0.2622515261,-0.4132283032,0.479164958,0.1666655838,-0.2782751918,-0.0373225622,-0.1018301398,-0.1409644932,0.178478837,-0.4876033962,0.2415321469,0.2238913774,0.0873818621,0.38781479,-0.1732172519,-0.3146266937,-0.0629522726,0.2620423138,0.4065777957,-0.3015166223,-0.2856298387,-0.1173471361,-0.241612196,0.0504980199,-0.1054368392,-0.2011386901,0.2815577686,-0.0167808644,-0.0703995004,0.3614420295,-0.2343790382,-0.3531353176,0.4110332727,-0.0478603989,-0.071903035,-0.0759115368,0.2448171228,0.0474888794,-0.0725209787,-0.1860623509,0.0920645967,0.0307610128,0.0773212016,0.1497756541,-0.1875065714,0.2341778874,-0.0015334329,0.1241452917,0.2763523459,0.1399574131,-0.420553863,0.1893345863,0.2317565829,0.1503296643,0.355152905,-0.1651833951,-0.4015514851,-0.1150018945,-0.2628898919,-0.2299821228,0.0786335915,-0.0617304593,-0.1473404765,-0.1688000858,-0.2442699522,-0.0655416101,0.0673422515,-0.1299708635,0.2982673049,-0.1344028115,-0.5122274756,0.0367587544,-0.1499243528,-0.3869235814,0.2167724669,0.1945154071,0.1012107059,0.1696923226,0.0502697118,0.459985137,-0.1136488244,-0.0803473443,0.1388212293,0.04870243,0.0829548314,-0.0290944334,0.3649711013,0.1031373292,0.252364248,0.1710756421,0.3974097371,0.037687175,-0.224794358,-0.1037115678,0.2083734721,-0.4188669622,0.1820928901,0.1995804161,0.1599497646,-0.1794397086,0.234629944,-0.1804733127,0.1407560408,0.1444905102,-0.0888667107,-0.4731482565,-0.3031069338,0.068484731,0.3422614038,0.3420707881,-0.0974302292,0.2473237365,-0.0083739804,0.3244614601,0.1424059421,0.2136680335,-0.1983314902,-0.3455728889,0.2194939256,-0.045962207,0.1714564711,0.4166837037,0.0125212241,0.3845880926,0.2453301251,-0.1530783921,0.0026161363,-0.0621065646,0.3627313673,0.0730372146,0.0545571297,-0.0962882191,-0.1259753853,-0.226972416,-0.2155729979,-0.1228719503,0.0408077799,-0.1454324424,0.1795247644,-0.3290635943,-0.0077629453,0.0007107481,-0.1432679594,-0.480457902,-0.1354398131,0.2090540528,0.0530926064,-0.0528867468,0.5098966956,0.1242874414,0.1399916708,0.0301752854,-0.2715519071,-0.2854590416,-0.2685713172,0.3451518714,-0.1722872555,-0.0865654498,-0.205834344,0.1100154966,-0.1852139533,0.1396550983,-0.2164420784,0.0435309522,0.1771795452,0.0013085809,0.0881358832,-0.0593979508,0.1097072065,0.0103697209,0.5180321932,-0.0942613408,-0.0591810234,0.0995153561,-0.1263444573,-0.0315817818,0.0843847543,-0.1217757612,0.3325320482,-0.1045642644,-0.099418819,-0.083496742,-0.3843751848,0.0502827391,0.1206876636,-0.1646551937,-0.0517021455,-0.1406535804,-0.1926172823,0.1343061328,-0.4898988605,0.3091023266,-0.1470202953,0.1182011515,0.0801663697,0.167426616,0.1182433292,0.263187319,-0.3549214602,0.0283561442,-0.1727309078,-0.0023675128,-0.2131805718,0.0833469704,0.3554988205,0.1952704191,0.0604897067,-0.0558646135,-0.1040972248,-0.0567384958,0.0631672069,0.0698353723,0.0344594829,-0.0473815687,-0.2366431057,0.6141837239,0.0919991732,-0.0919125304,-0.1033864692,0.0090510743,0.1637837589,0.1960847825,-0.0705555379,-0.5262030959,0.0143890986,-0.045922,-0.0322818607,0.1214521304,-0.0631750673,0.0450207219,-0.0504896827,0.1169421971,-0.2403010279,-0.1419678181,-0.4802061915,0.4614352286,-0.0157534778,0.2269478738,-0.2995601594,0.1065014377,0.1062132418,0.2454842329,0.3297358751,-0.2437499166,-0.0746303499,-0.0446578152,-0.3279978633,0.2383661568,0.0933354199,0.0316674486,-0.0943403095,-0.0265827179,0.1586090326,0.2483362705,0.3697468042,-0.4077409208,0.259309262,0.1666104794,-0.208727017,-0.2014466077,-0.0463936627,0.1423126459,0.4433344007,0.2257812321,0.162608698,0.148593992,-0.2219480276,-0.0578015633,0.0410224609,-0.3618226945,-0.3129102588,-0.5531741381,-0.0424817353,-0.0629407167,0.1748670936,0.0701643974,-0.3610954881,-0.1632426381,-0.2949319184,-0.0610833503,-0.32101053,0.0023079275,0.4309772849,0.0183988288,0.0223182198,-0.1053875312,0.219559893,-0.1172479689,-0.1520588398,-0.0362153016,0.2892939448,-0.1550923884,0.0741342157,-0.2477943152,0.3164279163,-0.0828590468,-0.2551887333,0.2867896557,0.008752441,0.2370634526,-0.2747247517,-0.0644562021,0.3154395521,0.100409165,-0.3306770325,0.0442068502,0.275470674,0.1331278086,-0.122319594,0.1277172863,-0.1690709442,-0.1833424866,-0.3525828719,0.2918332219,0.7135696411,-0.4319210052,-0.1446543038,-0.3559749722,0.0088928128,0.1857150197,-0.0889951885,0.2825490534,-0.1481867731,-0.3673080802,-0.2275435627,-0.302615881,0.440259248,0.3619307876,-0.1523835063,0.0919800699,-0.1903012395,0.0956289768,-0.1654461026,-0.2392514348,0.2474027872,-0.2798160911,-0.3376834691,-0.0935482234,0.0362191051,0.158659339,-0.0739717409,0.1744252294,0.2933245301,0.1652284861,-0.2226187438,-0.2081990987,0.1240016297,0.1661001444,0.2709668875,-0.3187802732,0.3715474308,0.1175296679,0.5764491558,0.1615431458,-0.1230596974,0.0815664828,0.2311410755,0.2143230587,-0.144782424,0.2635621727,0.1395494193,-0.0501189791,-0.1251632869,0.3880839348,0.0825444609,-0.3765778542,0.0979466215,0.0449484624,0.3170093894,-0.0457231477,0.5854789019,0.2599982917,0.4253346622,-0.0329947881,-0.0644232854,0.2772458792,-0.0199049208,0.653132081,-0.2949050367,-0.4707323313,0.0545517579,-0.0316644758,0.0615848266,-0.129902333,0.1589777619,0.038032826,-0.1793585867,-0.2140564919,0.0968190283,-0.0529563427,-0.2472262233,0.2621440589,0.2397036999,0.1260241717,-0.1100012362,0.1642518193,-0.3169594109,0.096654363,-0.5039532185,-0.1002561674,-0.5480384231,-0.0317325443,-0.1239491031,0.2156188935,0.0691370517,0.293961823,0.1357737929,-0.0552852415,-0.0605779327,-0.29185763,-0.1900506467,0.102883853,0.2943024337,0.2551986575,-0.0562939048,-0.1330565661,-0.2916058898,-0.12271896,-0.2887803912,-0.0360205285,-0.1181157753,0.2696971297,-0.1220968738,-0.3067593873,-0.0056852014,-0.0613391958,-0.0833405256,-0.1033747867,0.2095609158,0.1420368254,0.2168642282,0.4606212974,0.0934376642,-0.0980474874,-0.4693944752,-0.1366827935,0.2118953466,0.1991723031,0.0767967254,0.28376472,0.0422669388,0.1052961424,-0.1642039567,0.1334051788,0.4515683949,0.106548585,0.1396982968,-0.0659108162,0.1160324216,-0.1108912006,0.1632678658,0.0049915216,-0.2694737017,-0.4546253383,-0.0257953554,-0.0599919185,-0.1077878624,-0.389177978,0.3527149856,0.0281374045,0.1355990171,0.3533260524,-0.2131982148,0.4176568985,0.3382109404,0.1407421827,0.0804940239,-0.220490545,-0.0312426835,0.3701893985,0.1710315198,0.0887414962,0.566994369,0.6067858338,0.053177733,0.1392555982,0.1154805645,0.3828787506,0.1580106616,0.1414523423,-0.3168924451,-0.1274823099,-0.096369043,-0.0097292978,-0.2600937486,-0.0668183193,0.1962119043,-0.2595645785,0.072573036,-0.1358903348,-0.1593066603,0.3143870831,0.0949296951,-0.321303606,0.4296512008,0.1611703336,0.1368116289,0.0168906637,0.0410612486,0.15977557,0.4636191428,-0.0748362839,0.3082242012,0.0199698415,0.2071753591,0.0732519776,0.105549641,-0.0212373324,-0.4432514906,-0.4240786731,0.4105767608,0.1780603975,0.199593991,-0.0591985732,0.1031317189,-0.340703994,-0.1046201959,-0.1414486021,0.0907053277,-0.2635655403,0.3098531365,-0.2103798538,0.1371961683,-0.0053020599,-0.1227390096,0.137365669,-0.1533679515,-0.0779598653,0.3819878399,-0.2443204373,0.6768480539,0.319635272,-0.203360498,0.0486039743,0.1490587741,-0.1084140763,-0.1907948554,0.5342121124,-0.1282657236,-0.0296314955,-0.0715151131,-0.0445488468,-0.2661742568,0.5012587309,0.1911076307,0.4887460172,-0.1981479824,-0.1337868124,-0.6350818276,-0.0376051404,0.4233620465,-0.138216868,0.2652036548,-0.0654605925,0.0015621692,0.5450614691,0.535159409,-0.286691308,-0.1760884076,-0.2413429171,0.0049282112,0.109022066,-0.287825495,0.149126932,-0.258589685,-0.198503986,0.276910156,0.2439939678,-0.1533011943,0.1096713319,-0.2206153125,0.2532630861,0.0940567404,0.2637478411,0.0382155329,0.0974496603,0.07152652,-0.0023837779,0.1134411842,-0.0338568762,-0.0548510365,0.1085759029,-0.0150144286,0.0029875713,-0.3970349133,0.3604340553,-0.2424626052,-0.0326247886,0.1972207725,0.199883461,-0.0487743244,-0.0447782911,0.1410310566,0.1802861392,-0.0711050481,0.0462826863,0.2530254722,0.581379056,-0.170636937,0.0019506036,0.1594063044,-0.30259794,-0.4452957511,-0.2526145875,0.3208004236,-0.0966458395,-0.3266314864,-0.5128279924,-0.131495595,0.1436121464,-0.1941707879,-0.2234551162,0.2881712615,-0.4045099914,0.1213745996,-0.0012250087,0.3339444101,0.0121283932,0.3669067919,-0.4686109126,-0.0919876993]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1624","title":"Cannot download ade_corpus_v2","comments":"Hi @him1411, the dataset you are trying to load has been added during the community sprint and has not been released yet. It will be available with the v2 of `datasets`.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of `datasets` using pip:\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","body":"I tried this to get the dataset following this url : https:\/\/huggingface.co\/datasets\/ade_corpus_v2\r\n\r\nbut received this error : \r\n\r\n`Traceback (most recent call last):\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/ade_corpus_v2\/ade_corpus_v2.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/ade_corpus_v2\/ade_corpus_v2.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at ade_corpus_v2\/ade_corpus_v2.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/ade_corpus_v2\/ade_corpus_v2.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/ade_corpus_v2\/ade_corpus_v2.py`\r\n\r\n\r\n","comment_length":54,"text":"Cannot download ade_corpus_v2\nI tried this to get the dataset following this url : https:\/\/huggingface.co\/datasets\/ade_corpus_v2\r\n\r\nbut received this error : \r\n\r\n`Traceback (most recent call last):\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/ade_corpus_v2\/ade_corpus_v2.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/ade_corpus_v2\/ade_corpus_v2.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at ade_corpus_v2\/ade_corpus_v2.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/ade_corpus_v2\/ade_corpus_v2.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/ade_corpus_v2\/ade_corpus_v2.py`\r\n\r\n\r\n\nHi @him1411, the dataset you are trying to load has been added during the community sprint and has not been released yet. It will be available with the v2 of `datasets`.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of `datasets` using pip:\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","embeddings":[-0.0399101153,-0.1982669234,-0.113394618,0.2741905153,0.2603805959,0.2967451215,-0.0410111733,0.3299029171,0.0405992754,-0.1682073772,-0.2708159387,-0.2759156823,0.0790571347,-0.2060552537,-0.0267115999,-0.323489368,-0.1370057613,0.0029619159,-0.0265697669,-0.0274784267,-0.1152140573,0.3363958001,-0.2073079646,-0.0367349312,0.2109981328,0.0620413721,-0.1915316135,0.0425680354,-0.2808462083,-0.3533932865,0.53807652,-0.1105325148,0.2315054089,0.4590536654,-0.0001128967,0.159000352,0.2790457606,-0.0251458976,-0.365804255,-0.5255252123,-0.0857978389,0.0248197503,-0.0926860869,-0.3844043911,-0.1996401548,-0.3515524268,0.1897516251,-0.121204555,0.1625540704,0.300662756,0.2591804862,-0.0057054446,0.2623307705,-0.4079148769,-0.0430632941,0.0092382785,-0.0960070938,0.1479903311,0.075496003,0.2773635089,0.5707489848,0.2245631963,-0.2179099917,0.1162523031,-0.0670321286,0.100350976,0.1089014634,-0.3214739263,0.3586633801,-0.0397363305,0.6245003939,-0.1722650677,-0.5032495856,0.0513363592,0.0833838657,-0.1121204123,0.1236503422,0.0861939713,-0.2576685548,0.2251090109,-0.271217525,-0.2157211602,-0.2600765824,0.0914617851,0.0068647084,-0.0785092935,-0.1740344763,-0.019697912,-0.0309119318,0.080880858,-0.3524926901,0.0483626872,-0.1184762642,0.3002521396,0.0673744977,-0.0911049545,-0.1955265105,-0.0883468688,0.1906267405,0.0668392628,0.2124758512,-0.2406171709,-0.0119449897,-0.0469680093,-0.0392432958,0.2032940686,0.1763388067,-0.0667514876,0.2336608768,0.4772131443,0.1515923887,-0.2167411596,-0.1939044446,-0.2442658097,-0.2289280444,0.1582281142,0.2803454697,-0.447412163,-0.2276506722,-0.0553199723,-0.0535495766,0.197368741,0.0759611502,0.2929617763,-0.2096658796,-0.0176630169,-0.0392605364,0.2517108321,-0.0915941373,-0.0751967505,-0.1726430207,0.0985154733,-0.1267765015,-0.071235925,0.4334601164,-0.172145322,0.5568087697,-0.01730459,-0.165155232,0.0689437911,-0.0026068252,-0.2857436836,-0.1562635452,0.4075728059,0.2570572495,0.4013083279,0.0511066392,-0.1140500754,-0.0957691148,-0.2148584425,-0.1625625491,-0.2975975871,0.1640312821,0.2230684161,-0.0380365215,-0.0300294422,-0.2216090858,-0.06868653,0.1481539458,-0.0852437839,-0.0049362229,0.1350704432,-0.572783649,-0.2564504445,0.2576949894,0.3892453015,-0.1078387275,0.0925006121,-0.1284802705,-0.0780445039,-0.0828075334,0.1136744544,-0.1936371028,0.1439759731,-0.3953800201,0.0960064754,0.3923988938,-0.9619325995,-0.5316289663,0.3074181378,-0.0790592358,-0.0476625413,0.2594870925,-0.193416208,0.3689076602,-0.3970071971,0.0991809666,0.1445134729,0.0960305706,-0.1170032397,0.0447865799,-0.045579128,-0.124108091,0.054028295,-0.0139340973,0.190086633,0.1639372259,0.0328419097,0.3211568594,0.1087778062,0.181635797,0.3467017114,0.1078858748,-0.1950686425,0.1039923355,-0.1141309366,-0.3088447154,0.3227958381,-0.0806961358,0.0740596354,-0.2461928576,-0.1048443615,-0.4744685292,-0.1219155565,-0.1964444965,-0.4550785422,0.0906680822,0.0183877759,0.5066975355,0.1317300051,-0.0604760759,0.3836167157,-0.0854322091,0.142460525,-0.1503279358,0.0147880698,0.0733482689,0.053287223,-0.0068445369,-0.0800044015,0.1633701622,0.0168759283,-0.0752949789,0.1261767894,-0.014680774,0.2781453431,0.1799918562,-0.0110479174,0.2677811384,-0.3115859032,0.0727299973,-0.0469257645,0.2421076,0.0392748155,0.0623304248,-0.1299505532,0.1001603454,0.1143912748,0.3588016033,0.1484081596,0.3363650739,-0.1169097647,0.141220659,-0.1831562817,0.2509795129,0.1245312616,0.114884451,-0.154750973,-0.0881509483,0.274064064,0.1330340058,0.1747714728,-0.0855381563,0.0776381493,-0.1877564043,0.1635003835,-0.0824789479,0.45902583,0.4656788111,0.2134691328,0.3276383281,0.1585112959,0.0019474645,-0.1873863041,0.324993223,0.0360771008,0.1105055809,0.1557202041,0.1118210033,0.146860078,-0.0375960097,-0.1002386212,0.2401796877,0.2203896046,-0.131401822,-0.1094390303,-0.1590308398,-0.4430620968,-0.2277200371,-0.1068865955,0.0623177551,0.0080609368,-0.0241542794,0.424263984,0.1728570908,-0.0849471092,-0.3658924997,-0.0858860016,-0.0495841317,-0.1270220429,-0.0928509459,0.2467514724,-0.2537060678,0.1124774516,0.14385432,0.1789725572,0.3550393283,-0.2721133232,0.0480044447,-0.3824070096,-0.080881089,0.2435293049,0.0341095068,0.3125128746,0.1089713648,0.2086796612,-0.1823421866,-0.0530172512,0.2419634461,-0.087568678,-0.4155235291,-0.0122696823,-0.0029781247,-0.0665701255,-0.1537419707,-0.3599541783,-0.2867087424,-0.251372695,-0.2509933114,0.0242556222,0.1879399717,0.4431188405,-0.2806210816,0.0617797747,0.2239062786,0.2829473913,-0.2250178158,-0.2212397009,0.361212194,-0.5924842954,-0.3680436611,0.1045943275,0.1452607214,0.4514105618,0.2817083299,-0.4592278302,-0.2388827056,-0.0816094652,-0.2589484155,0.0334601477,-0.1563779563,0.3027187884,-0.2187416255,-0.0749977157,0.0001980195,0.0595103949,0.2751667202,-0.0508786924,0.1851986796,-0.0113794683,0.3512956798,0.1126058176,0.4582701921,0.6416583657,0.0769231766,0.5418627262,-0.1050295159,0.1122363955,-0.0519141667,-0.5397056937,-0.1024657413,0.1521128267,0.2426543236,0.1649681181,0.2039411515,-0.0557572916,-0.3971022069,-0.3225966394,-0.4174732864,-0.1801713854,-0.2219049335,0.0995136425,0.2833354175,0.062963523,0.2435507029,0.0485479571,-0.068222411,0.1176408529,0.3647827506,-0.0434271023,0.2248773724,-0.0899403691,-0.0428513028,-0.6552762389,0.3240651786,0.1203537956,0.3302398622,-0.1474135816,0.2084894925,0.0919598937,0.1300566047,0.8983021379,-0.344905287,0.2532890141,-0.1114863977,-0.0709566101,-0.3883152008,0.0821616426,0.1213578507,0.6417463422,0.0756441131,0.3298549652,-0.2659645975,0.0007725815,0.1921421587,-0.0217406135,-0.1426725686,-0.1155107617,-0.0976328328,-0.6872632504,-0.1449578404,-0.2012650073,0.0322593451,0.2486923784,0.1411321163,-0.1197467297,-0.0315406919,0.0247778129,0.133472532,0.0047180867,0.2767627537,0.1753459424,0.3898344636,0.5165261626,0.0436670929,0.1491284966,0.5392688513,-0.1278941929,-0.2325411886,0.0555956513,0.0158540122,-0.0460565239,0.3378149271,-0.0406878442,-0.0027506745,-0.0463297516,-0.1784972847,-0.0693198815,-0.1893952638,0.4800873995,0.0287377294,-0.4528841376,-0.4792282581,0.137744233,0.0232265797,-0.0972000286,0.1893782467,-0.0432939418,-0.2816011608,0.0984732732,-0.1655670106,0.831946075,0.1218220815,-0.1014555022,0.0863723457,-0.017924739,0.618704617,0.0380186848,0.173671335,0.0325756371,-0.2418027669,-0.1273514479,-0.0844213888,0.1253200769,0.0615587719,-0.2623636723,0.4733439088,-0.0283274986,0.1172506735,0.0141176414,0.2744521499,-0.334546268,-0.1542872787,-0.5114957094,0.1424471438,-0.0512752198,0.5752465725,-0.1724167466,-0.1829849184,-0.1918567568,-0.313367039,-0.3538986444,0.1942897141,-0.0838299245,0.2777224183,-0.0915049836,-0.1545034349,0.0208190605,0.283585757,0.7378914952,0.0675429925,-0.4196716845,0.2290886194,-0.0960143879,-0.2218797654,0.2674856186,-0.3062758446,0.2200663388,-0.300724864,-0.1965103149,0.1097697094,0.1154362336,-0.0586118177,0.2039769292,-0.3506623507,-0.3065339029,-0.0893342122,-0.2451552153,0.1280300766,0.0809906945,-0.1501711011,0.0821089521,-0.1429533213,-0.0594753884,-0.022308955,-0.0092505869,-0.2508853972,0.0321429111,0.5760474801,0.0949670672,-0.2015913129,0.5825818777,0.0990263969,-0.1363515407,-0.2064478397,-0.4992934167,-0.2328624129,-0.5102233291,0.0513311848,-0.0161336772,0.4604524672,0.0225792322,0.0589982085,0.2213426083,-0.2192216069,0.0122765889,-0.5906417966,-0.143237114,0.201896891,-0.0184506681,-0.0465257242,-0.1227690727,0.1095513701,0.2810791433,-0.2713392675,-0.288939774,0.1452508569,-0.1061071903,-0.230622068,0.0081703616,-0.0111998729,0.3746191859,0.0014030535,0.1206454709,0.1764123142,-0.2742566764,-0.1349224448,-0.0515173115,0.1365890503,-0.2135878652,-0.193538934,-0.0490220264,-0.3488192856,-0.2868466079,0.1396255791,0.0483084098,0.3250525296,0.1891445071,-0.2940942645,0.118036218,0.0976238623,-0.0445429422,0.4540589154,-0.139983952,0.2237018198,0.242046833,0.2468589842,-0.1591075808,-0.1558391452,-0.0036875461,-0.1054708883,-0.1114056781,-0.2081737965,0.2927430272,-0.122975871,0.2513773441,0.3190283477,0.194258377,0.0602114648,-0.2426787466,0.2417406887,0.1946669668,0.2053769231,-0.3600097299,0.0526831038,0.1588477045,0.065615885,-0.1489723027,0.1919577271,0.1248755306,-0.063121371,0.2120789587,0.2209670842,0.7554401755,-0.0236869995,0.0659239814,0.0336201303,0.0927207768,0.0200682897,-0.0304563958,0.1907741725,0.3175688982,0.2540689111,-0.3543515205,0.1352820545,-0.0578900799,0.1832671762,0.0791436136,-0.3790208101,0.0091440203,0.3755663335,0.0854998678,0.281034559,0.0554608516,0.3250522614,-0.0968430638,0.0595964901,-0.5663063526,0.3975794911,0.0382076986,-0.0101011032,-0.2789667547,-0.3378447592,-0.0248788819,0.0041617863,0.0006079846,-0.2313052714,-0.0328009836,-0.0848590806,-0.3778826296,-0.3994798958,-0.11124181,0.1067070588,0.145710662,-0.1347804666,0.4259522259,0.2621905804,-0.1036841944,0.2320249826,0.6412649155,0.1760902852,0.1427491903,0.0704455525,0.0519339629,0.1193026006,-0.1221634671,-0.0769705102,0.2211875916,0.1523688585,-0.0002039674,0.2625030279,0.1621084362,-0.0736407638,0.0611341447,0.0105877537,0.1776607484,-0.0668970123,0.343339622,-0.2465384156,0.2775495052,-0.1333295852,0.1362704933,-0.6502066255,-0.1536979377,0.3432265222,-0.4141463339,0.1078461558,-0.0682408512,0.0560545176,0.0745089948,0.0869652927,0.4335426986,0.3255806565,-0.197047174,-0.3015869856,-0.5478926897,0.1837693304,-0.0247350764,0.138073504,0.0065424037,0.067382887,0.0349252969,0.0804399401,0.2238636017,0.1808240712,-0.1317716688,0.1575792283,-0.2936446369,-0.1027537659,-0.1600828469,0.0632451549,-0.0879277065,-0.2994442284,0.1270012707,-0.0999277011,0.0527354702,-0.1937147677,0.2205796093,0.0570682622,-0.2798716426,0.3468836844,0.1318029314,0.1966952085,-0.1534296721,-0.1781551391,-0.23232086,-0.1524001062,-0.1193514988,0.3400589824,0.0874051079,-0.0301337112,-0.233188048,0.0259148609,-0.0119094057,0.1198291332,0.0347451642,-0.008333711,-0.3551753759,0.1296233833,-0.4348855913,0.0018161762,-0.0895785615,0.0752892494,-0.0213381611,-0.0050704181,-0.1325352341,-0.3424767256,0.7437351942,-0.4379320741,0.0565478019,0.1573339999,0.3597735167,-0.1136438027,-0.0894941613,-0.4221367836,0.0342017077,0.3630839288,0.0660159141,-0.254144907,0.2254090756,-0.1688335538,-0.0743738264,-0.0662858039,0.3791212142,-0.1086538285,-0.3819715083,0.0567828156,-0.0417264625]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1624","title":"Cannot download ade_corpus_v2","comments":"`ade_corpus_v2` was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `ade_corpus_v2` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"ade_corpus_v2\", \"Ade_corpos_v2_drug_ade_relation\")\r\n```\r\n\r\n(looks like there is a typo in the configuration name, we'll fix it for the v2.0 release of `datasets` soon)","body":"I tried this to get the dataset following this url : https:\/\/huggingface.co\/datasets\/ade_corpus_v2\r\n\r\nbut received this error : \r\n\r\n`Traceback (most recent call last):\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/ade_corpus_v2\/ade_corpus_v2.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/ade_corpus_v2\/ade_corpus_v2.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at ade_corpus_v2\/ade_corpus_v2.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/ade_corpus_v2\/ade_corpus_v2.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/ade_corpus_v2\/ade_corpus_v2.py`\r\n\r\n\r\n","comment_length":61,"text":"Cannot download ade_corpus_v2\nI tried this to get the dataset following this url : https:\/\/huggingface.co\/datasets\/ade_corpus_v2\r\n\r\nbut received this error : \r\n\r\n`Traceback (most recent call last):\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/ade_corpus_v2\/ade_corpus_v2.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 278, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 486, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/ade_corpus_v2\/ade_corpus_v2.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/opt\/anaconda3\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 282, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at ade_corpus_v2\/ade_corpus_v2.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/ade_corpus_v2\/ade_corpus_v2.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/ade_corpus_v2\/ade_corpus_v2.py`\r\n\r\n\r\n\n`ade_corpus_v2` was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `ade_corpus_v2` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"ade_corpus_v2\", \"Ade_corpos_v2_drug_ade_relation\")\r\n```\r\n\r\n(looks like there is a typo in the configuration name, we'll fix it for the v2.0 release of `datasets` soon)","embeddings":[-0.0399101153,-0.1982669234,-0.113394618,0.2741905153,0.2603805959,0.2967451215,-0.0410111733,0.3299029171,0.0405992754,-0.1682073772,-0.2708159387,-0.2759156823,0.0790571347,-0.2060552537,-0.0267115999,-0.323489368,-0.1370057613,0.0029619159,-0.0265697669,-0.0274784267,-0.1152140573,0.3363958001,-0.2073079646,-0.0367349312,0.2109981328,0.0620413721,-0.1915316135,0.0425680354,-0.2808462083,-0.3533932865,0.53807652,-0.1105325148,0.2315054089,0.4590536654,-0.0001128967,0.159000352,0.2790457606,-0.0251458976,-0.365804255,-0.5255252123,-0.0857978389,0.0248197503,-0.0926860869,-0.3844043911,-0.1996401548,-0.3515524268,0.1897516251,-0.121204555,0.1625540704,0.300662756,0.2591804862,-0.0057054446,0.2623307705,-0.4079148769,-0.0430632941,0.0092382785,-0.0960070938,0.1479903311,0.075496003,0.2773635089,0.5707489848,0.2245631963,-0.2179099917,0.1162523031,-0.0670321286,0.100350976,0.1089014634,-0.3214739263,0.3586633801,-0.0397363305,0.6245003939,-0.1722650677,-0.5032495856,0.0513363592,0.0833838657,-0.1121204123,0.1236503422,0.0861939713,-0.2576685548,0.2251090109,-0.271217525,-0.2157211602,-0.2600765824,0.0914617851,0.0068647084,-0.0785092935,-0.1740344763,-0.019697912,-0.0309119318,0.080880858,-0.3524926901,0.0483626872,-0.1184762642,0.3002521396,0.0673744977,-0.0911049545,-0.1955265105,-0.0883468688,0.1906267405,0.0668392628,0.2124758512,-0.2406171709,-0.0119449897,-0.0469680093,-0.0392432958,0.2032940686,0.1763388067,-0.0667514876,0.2336608768,0.4772131443,0.1515923887,-0.2167411596,-0.1939044446,-0.2442658097,-0.2289280444,0.1582281142,0.2803454697,-0.447412163,-0.2276506722,-0.0553199723,-0.0535495766,0.197368741,0.0759611502,0.2929617763,-0.2096658796,-0.0176630169,-0.0392605364,0.2517108321,-0.0915941373,-0.0751967505,-0.1726430207,0.0985154733,-0.1267765015,-0.071235925,0.4334601164,-0.172145322,0.5568087697,-0.01730459,-0.165155232,0.0689437911,-0.0026068252,-0.2857436836,-0.1562635452,0.4075728059,0.2570572495,0.4013083279,0.0511066392,-0.1140500754,-0.0957691148,-0.2148584425,-0.1625625491,-0.2975975871,0.1640312821,0.2230684161,-0.0380365215,-0.0300294422,-0.2216090858,-0.06868653,0.1481539458,-0.0852437839,-0.0049362229,0.1350704432,-0.572783649,-0.2564504445,0.2576949894,0.3892453015,-0.1078387275,0.0925006121,-0.1284802705,-0.0780445039,-0.0828075334,0.1136744544,-0.1936371028,0.1439759731,-0.3953800201,0.0960064754,0.3923988938,-0.9619325995,-0.5316289663,0.3074181378,-0.0790592358,-0.0476625413,0.2594870925,-0.193416208,0.3689076602,-0.3970071971,0.0991809666,0.1445134729,0.0960305706,-0.1170032397,0.0447865799,-0.045579128,-0.124108091,0.054028295,-0.0139340973,0.190086633,0.1639372259,0.0328419097,0.3211568594,0.1087778062,0.181635797,0.3467017114,0.1078858748,-0.1950686425,0.1039923355,-0.1141309366,-0.3088447154,0.3227958381,-0.0806961358,0.0740596354,-0.2461928576,-0.1048443615,-0.4744685292,-0.1219155565,-0.1964444965,-0.4550785422,0.0906680822,0.0183877759,0.5066975355,0.1317300051,-0.0604760759,0.3836167157,-0.0854322091,0.142460525,-0.1503279358,0.0147880698,0.0733482689,0.053287223,-0.0068445369,-0.0800044015,0.1633701622,0.0168759283,-0.0752949789,0.1261767894,-0.014680774,0.2781453431,0.1799918562,-0.0110479174,0.2677811384,-0.3115859032,0.0727299973,-0.0469257645,0.2421076,0.0392748155,0.0623304248,-0.1299505532,0.1001603454,0.1143912748,0.3588016033,0.1484081596,0.3363650739,-0.1169097647,0.141220659,-0.1831562817,0.2509795129,0.1245312616,0.114884451,-0.154750973,-0.0881509483,0.274064064,0.1330340058,0.1747714728,-0.0855381563,0.0776381493,-0.1877564043,0.1635003835,-0.0824789479,0.45902583,0.4656788111,0.2134691328,0.3276383281,0.1585112959,0.0019474645,-0.1873863041,0.324993223,0.0360771008,0.1105055809,0.1557202041,0.1118210033,0.146860078,-0.0375960097,-0.1002386212,0.2401796877,0.2203896046,-0.131401822,-0.1094390303,-0.1590308398,-0.4430620968,-0.2277200371,-0.1068865955,0.0623177551,0.0080609368,-0.0241542794,0.424263984,0.1728570908,-0.0849471092,-0.3658924997,-0.0858860016,-0.0495841317,-0.1270220429,-0.0928509459,0.2467514724,-0.2537060678,0.1124774516,0.14385432,0.1789725572,0.3550393283,-0.2721133232,0.0480044447,-0.3824070096,-0.080881089,0.2435293049,0.0341095068,0.3125128746,0.1089713648,0.2086796612,-0.1823421866,-0.0530172512,0.2419634461,-0.087568678,-0.4155235291,-0.0122696823,-0.0029781247,-0.0665701255,-0.1537419707,-0.3599541783,-0.2867087424,-0.251372695,-0.2509933114,0.0242556222,0.1879399717,0.4431188405,-0.2806210816,0.0617797747,0.2239062786,0.2829473913,-0.2250178158,-0.2212397009,0.361212194,-0.5924842954,-0.3680436611,0.1045943275,0.1452607214,0.4514105618,0.2817083299,-0.4592278302,-0.2388827056,-0.0816094652,-0.2589484155,0.0334601477,-0.1563779563,0.3027187884,-0.2187416255,-0.0749977157,0.0001980195,0.0595103949,0.2751667202,-0.0508786924,0.1851986796,-0.0113794683,0.3512956798,0.1126058176,0.4582701921,0.6416583657,0.0769231766,0.5418627262,-0.1050295159,0.1122363955,-0.0519141667,-0.5397056937,-0.1024657413,0.1521128267,0.2426543236,0.1649681181,0.2039411515,-0.0557572916,-0.3971022069,-0.3225966394,-0.4174732864,-0.1801713854,-0.2219049335,0.0995136425,0.2833354175,0.062963523,0.2435507029,0.0485479571,-0.068222411,0.1176408529,0.3647827506,-0.0434271023,0.2248773724,-0.0899403691,-0.0428513028,-0.6552762389,0.3240651786,0.1203537956,0.3302398622,-0.1474135816,0.2084894925,0.0919598937,0.1300566047,0.8983021379,-0.344905287,0.2532890141,-0.1114863977,-0.0709566101,-0.3883152008,0.0821616426,0.1213578507,0.6417463422,0.0756441131,0.3298549652,-0.2659645975,0.0007725815,0.1921421587,-0.0217406135,-0.1426725686,-0.1155107617,-0.0976328328,-0.6872632504,-0.1449578404,-0.2012650073,0.0322593451,0.2486923784,0.1411321163,-0.1197467297,-0.0315406919,0.0247778129,0.133472532,0.0047180867,0.2767627537,0.1753459424,0.3898344636,0.5165261626,0.0436670929,0.1491284966,0.5392688513,-0.1278941929,-0.2325411886,0.0555956513,0.0158540122,-0.0460565239,0.3378149271,-0.0406878442,-0.0027506745,-0.0463297516,-0.1784972847,-0.0693198815,-0.1893952638,0.4800873995,0.0287377294,-0.4528841376,-0.4792282581,0.137744233,0.0232265797,-0.0972000286,0.1893782467,-0.0432939418,-0.2816011608,0.0984732732,-0.1655670106,0.831946075,0.1218220815,-0.1014555022,0.0863723457,-0.017924739,0.618704617,0.0380186848,0.173671335,0.0325756371,-0.2418027669,-0.1273514479,-0.0844213888,0.1253200769,0.0615587719,-0.2623636723,0.4733439088,-0.0283274986,0.1172506735,0.0141176414,0.2744521499,-0.334546268,-0.1542872787,-0.5114957094,0.1424471438,-0.0512752198,0.5752465725,-0.1724167466,-0.1829849184,-0.1918567568,-0.313367039,-0.3538986444,0.1942897141,-0.0838299245,0.2777224183,-0.0915049836,-0.1545034349,0.0208190605,0.283585757,0.7378914952,0.0675429925,-0.4196716845,0.2290886194,-0.0960143879,-0.2218797654,0.2674856186,-0.3062758446,0.2200663388,-0.300724864,-0.1965103149,0.1097697094,0.1154362336,-0.0586118177,0.2039769292,-0.3506623507,-0.3065339029,-0.0893342122,-0.2451552153,0.1280300766,0.0809906945,-0.1501711011,0.0821089521,-0.1429533213,-0.0594753884,-0.022308955,-0.0092505869,-0.2508853972,0.0321429111,0.5760474801,0.0949670672,-0.2015913129,0.5825818777,0.0990263969,-0.1363515407,-0.2064478397,-0.4992934167,-0.2328624129,-0.5102233291,0.0513311848,-0.0161336772,0.4604524672,0.0225792322,0.0589982085,0.2213426083,-0.2192216069,0.0122765889,-0.5906417966,-0.143237114,0.201896891,-0.0184506681,-0.0465257242,-0.1227690727,0.1095513701,0.2810791433,-0.2713392675,-0.288939774,0.1452508569,-0.1061071903,-0.230622068,0.0081703616,-0.0111998729,0.3746191859,0.0014030535,0.1206454709,0.1764123142,-0.2742566764,-0.1349224448,-0.0515173115,0.1365890503,-0.2135878652,-0.193538934,-0.0490220264,-0.3488192856,-0.2868466079,0.1396255791,0.0483084098,0.3250525296,0.1891445071,-0.2940942645,0.118036218,0.0976238623,-0.0445429422,0.4540589154,-0.139983952,0.2237018198,0.242046833,0.2468589842,-0.1591075808,-0.1558391452,-0.0036875461,-0.1054708883,-0.1114056781,-0.2081737965,0.2927430272,-0.122975871,0.2513773441,0.3190283477,0.194258377,0.0602114648,-0.2426787466,0.2417406887,0.1946669668,0.2053769231,-0.3600097299,0.0526831038,0.1588477045,0.065615885,-0.1489723027,0.1919577271,0.1248755306,-0.063121371,0.2120789587,0.2209670842,0.7554401755,-0.0236869995,0.0659239814,0.0336201303,0.0927207768,0.0200682897,-0.0304563958,0.1907741725,0.3175688982,0.2540689111,-0.3543515205,0.1352820545,-0.0578900799,0.1832671762,0.0791436136,-0.3790208101,0.0091440203,0.3755663335,0.0854998678,0.281034559,0.0554608516,0.3250522614,-0.0968430638,0.0595964901,-0.5663063526,0.3975794911,0.0382076986,-0.0101011032,-0.2789667547,-0.3378447592,-0.0248788819,0.0041617863,0.0006079846,-0.2313052714,-0.0328009836,-0.0848590806,-0.3778826296,-0.3994798958,-0.11124181,0.1067070588,0.145710662,-0.1347804666,0.4259522259,0.2621905804,-0.1036841944,0.2320249826,0.6412649155,0.1760902852,0.1427491903,0.0704455525,0.0519339629,0.1193026006,-0.1221634671,-0.0769705102,0.2211875916,0.1523688585,-0.0002039674,0.2625030279,0.1621084362,-0.0736407638,0.0611341447,0.0105877537,0.1776607484,-0.0668970123,0.343339622,-0.2465384156,0.2775495052,-0.1333295852,0.1362704933,-0.6502066255,-0.1536979377,0.3432265222,-0.4141463339,0.1078461558,-0.0682408512,0.0560545176,0.0745089948,0.0869652927,0.4335426986,0.3255806565,-0.197047174,-0.3015869856,-0.5478926897,0.1837693304,-0.0247350764,0.138073504,0.0065424037,0.067382887,0.0349252969,0.0804399401,0.2238636017,0.1808240712,-0.1317716688,0.1575792283,-0.2936446369,-0.1027537659,-0.1600828469,0.0632451549,-0.0879277065,-0.2994442284,0.1270012707,-0.0999277011,0.0527354702,-0.1937147677,0.2205796093,0.0570682622,-0.2798716426,0.3468836844,0.1318029314,0.1966952085,-0.1534296721,-0.1781551391,-0.23232086,-0.1524001062,-0.1193514988,0.3400589824,0.0874051079,-0.0301337112,-0.233188048,0.0259148609,-0.0119094057,0.1198291332,0.0347451642,-0.008333711,-0.3551753759,0.1296233833,-0.4348855913,0.0018161762,-0.0895785615,0.0752892494,-0.0213381611,-0.0050704181,-0.1325352341,-0.3424767256,0.7437351942,-0.4379320741,0.0565478019,0.1573339999,0.3597735167,-0.1136438027,-0.0894941613,-0.4221367836,0.0342017077,0.3630839288,0.0660159141,-0.254144907,0.2254090756,-0.1688335538,-0.0743738264,-0.0662858039,0.3791212142,-0.1086538285,-0.3819715083,0.0567828156,-0.0417264625]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1622","title":"Can't call shape on the output of select()","comments":"Indeed that's a typo, do you want to open a PR to fix it?","body":"I get the error `TypeError: tuple expected at most 1 argument, got 2` when calling `shape` on the output of `select()`.\r\nIt's line 531 in shape in arrow_dataset.py that causes the problem:\r\n``return tuple(self._indices.num_rows, self._data.num_columns)``\r\nThis makes sense, since `tuple(num1, num2)` is not a valid call.\r\n \r\nFull code to reproduce:\r\n\r\n```python\r\ndataset = load_dataset(\"cnn_dailymail\", \"3.0.0\")\r\ntrain_set = dataset[\"train\"]\r\nt = train_set.select(range(10))\r\nprint(t.shape)","comment_length":14,"text":"Can't call shape on the output of select()\nI get the error `TypeError: tuple expected at most 1 argument, got 2` when calling `shape` on the output of `select()`.\r\nIt's line 531 in shape in arrow_dataset.py that causes the problem:\r\n``return tuple(self._indices.num_rows, self._data.num_columns)``\r\nThis makes sense, since `tuple(num1, num2)` is not a valid call.\r\n \r\nFull code to reproduce:\r\n\r\n```python\r\ndataset = load_dataset(\"cnn_dailymail\", \"3.0.0\")\r\ntrain_set = dataset[\"train\"]\r\nt = train_set.select(range(10))\r\nprint(t.shape)\nIndeed that's a typo, do you want to open a PR to fix it?","embeddings":[-0.0470377393,-0.4228085577,-0.05458517,0.4962621927,0.4015932381,-0.1582879573,0.2413294613,0.1844179034,0.1243640929,0.2816537917,0.4273205101,0.5230286717,-0.2628481984,0.2571104765,0.0753411502,-0.3647647798,-0.1854680032,-0.0963822603,-0.3527422845,-0.0875790417,-0.3882706761,0.195346579,-0.2464379817,0.3379766047,-0.3010381162,-0.1039655879,0.2737055719,-0.1765854657,-0.2550579607,-0.1675288826,0.4781428576,-0.5915694833,-0.2588697374,0.5355428457,-0.0001188768,-0.2104697526,0.2604023516,0.0585208125,-0.0436491147,-0.4457792342,-0.3621533513,-0.266669631,0.3599996269,-0.4325096905,-0.0586748794,0.1775175482,-0.4364939928,0.1112539321,0.3602995872,0.3592143059,0.1865002513,0.086321339,0.035094101,-0.2154311538,0.0698668957,0.4926719069,-0.1458601952,0.206092611,-0.0200955346,-0.1013900563,0.4696869254,0.3399440646,-0.1314168721,-0.1154825389,-0.2768054008,0.3161586225,0.3987176716,-0.1668971777,-0.0530957542,0.0092599774,0.3950342834,-0.1525374055,-0.0543019399,-0.008967963,0.1482392699,-0.0774619505,0.0809276551,0.1816561967,-0.1800653487,-0.1858369112,-0.0684749857,0.1424336135,-0.2900226712,0.4068798721,0.099041827,-0.1054547802,-0.1119286343,0.2486342341,0.1469237804,-0.0573163703,0.2072727382,0.124449037,0.1611034274,0.05262357,-0.0081547257,-0.0314563401,-0.0411109552,-0.4254201949,-0.2153451294,0.2024807632,0.100707829,0.0449661426,-0.0691680685,0.2488932908,0.4562994242,0.1436239183,-0.1817147881,0.5431018472,-0.2839054763,-0.1197859198,0.1606404334,-0.0349111296,-0.0524605587,-0.3946494162,0.1851395518,0.1531423777,0.4736814499,-0.066520147,-0.0715572312,-0.176655978,0.0497312583,-0.0276216418,0.096202299,0.0958063677,0.0275640879,0.0023270112,0.2205130011,0.2326973528,-0.0719956607,-0.0343656838,-0.2960288823,0.2070693374,-0.2140553147,-0.1953027397,-0.0464093015,0.177453652,-0.1027235165,-0.0197712295,0.3589083254,-0.2652836144,0.0083826166,-0.2436528802,0.1588269621,0.1548443139,-0.2379318625,-0.0903308168,0.6245591044,-0.0391789004,-0.1223990917,0.3134678006,-0.4360381663,-0.4864941239,-0.0218582675,0.1067902148,0.3197747469,0.0522656292,-0.1391057223,0.1030034572,0.4132131338,0.0231100805,0.1258755624,-0.4864802361,0.1184678003,-0.2565171123,0.2478175014,0.1292173415,-0.4465169609,-0.0700905249,-0.0579654016,0.2399684489,0.5588713884,0.16332075,-0.3197975457,-0.0417979546,0.0698817819,0.2086014003,0.7656643987,0.0460004434,-0.1716541648,-0.2349739373,-0.1394957453,-0.2143571675,-0.2995622456,0.1019404531,-0.3265525401,0.2007898539,0.2591881156,0.1284368485,-0.0652106702,-0.1127852798,-0.080797635,-0.0037580824,0.3521812558,0.0758939981,0.0579943992,-0.0355261974,-0.2347131521,-0.3663568497,0.0393079929,0.1217567548,-0.2626023889,-0.3624141514,0.6732469201,-0.0079495516,-0.1479475498,-0.214624837,0.0105187213,0.0521400161,0.0099255778,0.0279053878,0.0101479003,0.0369738303,-0.2270056307,0.2777214646,0.0180340577,0.4766046703,0.0573056862,-0.2790991366,-0.0437551551,-0.1206625998,-0.1592674255,-0.1953876764,-0.2885147631,0.0410323925,-0.2372155041,0.2732338607,-0.0943456218,-0.3818604052,-0.0660734475,0.0008187569,0.2613957822,0.266834408,-0.1375806779,0.1451402605,0.1919214875,-0.3736019433,-0.4903031886,-0.0623981133,-0.26165995,-0.1530424654,0.0326998718,-0.0882536843,0.1038381234,-0.2574428618,0.0410155579,0.4187330306,-0.2218528092,0.363012284,-0.2917590737,0.3342907131,-0.227323845,0.1742348075,-0.0595010966,-0.2280034572,-0.1375850439,-0.2673730552,-0.1406325996,-0.2142278999,-0.3067266643,0.068119444,0.4543756843,0.1517698616,0.2748110294,0.2409110665,0.1608039141,0.1882739216,-0.1404475719,0.0437897854,0.4074276984,0.094189167,-0.1331113577,-0.1683916301,-0.1207028329,-0.1065248623,0.1485988349,0.2123779058,-0.2214858234,-0.0754372552,0.0026963982,-0.1212880164,0.0165537335,0.1698452383,0.0216117688,-0.028874388,-0.0738584623,0.150796786,-0.186494112,-0.291891098,-0.081881918,-0.3803558052,0.1646812409,-0.0638524368,-0.1389657408,-0.2848677039,0.0235501397,-0.0294120796,-0.1128010824,0.2831394374,0.3751420677,-0.1691699475,0.0063945921,0.090026848,-0.1733530164,0.0231487006,-0.0228854027,0.2784872949,0.2102824599,0.2229383588,0.0430557728,0.0588748567,-0.1544149965,-0.0590619855,-0.1342177838,0.01582589,0.5002955198,0.1574064195,-0.1284950525,-0.104463324,0.375408113,0.2326102555,0.1548407376,0.3631447256,0.0243192147,-0.1787174642,-0.0445611402,-0.1969809234,-0.176798299,-0.1733492911,0.3666982055,0.4090999961,0.2740924358,0.0612849034,0.3769086301,0.0869456008,0.0161019154,-0.1786843389,-0.2038548738,-0.018064132,0.1669633836,0.0343210921,-0.1533752531,-0.0564232804,0.1536925435,0.248720333,-0.1362985969,-0.2743715942,-0.1736924648,-0.2008845806,0.040541511,-0.2940834165,0.0820135847,0.2473533899,0.2332096845,-0.1112003997,0.071670644,-0.1438536048,0.3329837024,0.7839489579,0.014374584,0.1257480383,0.4443817139,0.0883427709,0.2688307166,0.0673190132,-0.2459847182,0.0519969463,0.174667567,-0.1827870011,-0.1125850156,-0.3226052523,-0.067009218,-0.257450819,-0.0802701339,-0.2390282899,-0.1113915071,0.0213735048,-0.2107905149,0.3718149364,-0.1758585125,-0.1814594716,0.2716073096,-0.0619729199,-0.0245222021,-0.1469324827,0.1201453879,-0.3471196294,0.1353096664,0.0956441015,-0.2385640293,0.4574941695,-0.0466000289,-0.3622934222,-0.3926819861,-0.0834703147,0.3379406035,-0.1005646661,-0.0518903695,-0.0928199068,-0.1288390905,0.0053977631,0.1908105463,0.3414446115,0.0860869363,-0.1452296078,0.3320047855,-0.3009617329,-0.3423114717,-0.1490440369,-0.2258183062,0.2665384412,-0.2603875101,0.0987412184,-0.3667381108,0.144107461,0.3722532988,0.202408582,-0.3718597293,-0.0206134822,-0.0197923556,-0.1894572377,0.215703249,0.1023068279,0.0800010711,0.0914677233,-0.0534814149,-0.1425516903,-0.2604981065,-0.2714175582,-0.1782323569,-0.1146483049,0.3787490129,-0.3430660069,-0.1463321149,0.1056833938,-0.0990838706,0.0780547559,0.4944097698,-0.1834997386,0.1288505942,0.0658819303,-0.0681412891,0.4537705481,0.2096381336,-0.0904765949,0.0653213337,-0.1159728765,0.233355388,0.0277854484,-0.1321607232,0.5920399427,0.0790633708,-0.0563223585,-0.645644486,0.4978291392,-0.0503813401,0.0823023319,0.6249057055,0.2053252161,0.0579818003,0.303319037,0.1999613047,1.1740193367,0.0621390045,-0.057393115,0.3479493558,0.347242415,0.1811217666,0.277217567,0.0079740779,-0.1794171929,-0.3934550583,-0.0758555457,-0.3500045538,-0.0607277416,0.2292089462,-0.2567815483,-0.0966894776,-0.5463039875,-0.3924412131,-0.0158449933,-0.2748791873,0.4156419933,0.0331304371,-0.1788587868,0.1109423414,-0.1653793007,0.0259915404,-0.1099004298,0.1078779399,0.0596344881,-0.4403730333,-0.2078887671,0.0338182487,-0.4202952087,0.2071086764,0.1784718633,-0.6904373765,0.2330351472,0.3497958779,-0.1364519298,0.1547100395,0.0815872699,-0.1553228348,0.451144278,0.0849727839,-0.1755827665,-0.0025747083,0.2627048492,-0.0600029677,-0.1698242277,0.4156903028,0.1319287568,-0.1810973883,-0.122091271,0.1428507715,-0.0317375846,-0.5531700253,-0.7149989009,-0.1789058447,0.0562764443,-0.2111257464,0.0881917849,0.2862835228,-0.1538762301,0.3407685161,0.3803727627,-0.1462604702,0.0856358707,-0.0664050654,0.3169367909,0.1656140834,0.3174603879,0.3740447164,-0.0328008793,-0.1693746746,-0.0567469262,0.1774099767,-0.4131915271,0.1781737506,-0.0102268495,0.1516098529,0.073400788,0.327486217,0.2364445776,0.0285587087,-0.2062256336,-0.1395036876,-0.0853864402,0.4555494487,0.2616821826,0.1119833291,-0.2282098085,0.0007809647,-0.0325835831,0.3278935552,-0.2387200594,-0.0329764262,-0.2055392265,0.1574757397,-0.003163975,0.0049088728,0.1189223155,0.1919005215,0.1092501357,0.1136285886,0.1550551802,-0.1717953384,-0.0347039066,0.1507946849,0.0500523709,0.0161755942,-0.0109661371,-0.037614204,-0.25240013,0.0426616371,-0.186899066,0.0114912111,-0.1802593768,0.0909843817,0.1449192017,0.0389402248,-0.0319085792,0.2023855746,-0.3208241463,-0.0968116224,0.2228619009,-0.0150543805,0.2975263596,-0.1032742262,-0.3742576838,-0.2031520903,-0.0017775485,0.0130211608,0.1270575374,-0.3851843476,-0.0132964076,-0.4733631611,0.1699326783,0.1951006949,-0.027522685,0.1172356606,0.1307801306,0.1512405276,-0.3993054628,-0.2393215448,-0.3210450709,0.2072241306,-0.0055682431,0.2578728497,-0.5235138535,0.3827012181,-0.4200551808,0.1686742306,-0.0352344401,-0.1851469427,0.3063251674,0.3574311137,-0.1304576248,-0.1007420868,0.0183395296,0.2261874825,0.027742777,0.0257515516,-0.0631419942,0.1923684627,0.1819628328,-0.2738062441,0.2572651803,-0.1151979938,-0.3272055387,0.1112862378,-0.2255074233,0.0606316477,-0.0279000029,0.3182542026,-0.1943730265,-0.2074221969,-0.2497271299,-0.1812240928,-0.0127505232,0.2768714428,-0.1519773751,-0.077072598,-0.0513809212,-0.3217807412,-0.0382355861,-0.1144886613,0.5726268888,-0.010419528,0.1198346317,-0.0428387448,-0.2650414109,0.0560755506,0.0045285071,-0.3024837971,0.0073340544,0.4698960185,0.007029762,-0.1764830798,0.0723152608,0.3144408464,0.4511796236,-0.308050245,-0.0601624958,-0.0082941651,-0.3057603836,0.1013157591,-0.0262377821,0.3494472206,0.015881341,0.1785336435,0.1708808541,-0.1412305534,0.1281048805,0.0031581153,0.4791848063,-0.5300072432,0.6371521354,-0.2300274521,-0.3389557898,-0.3594483137,-0.0373611152,-0.2228579968,-0.125490129,0.3143988848,0.274418205,0.0743374974,0.2216509283,0.037483111,0.0039911945,0.4801988006,-0.1606903821,0.2985351682,-0.2730500102,-0.1079555228,0.0584310852,0.2213099301,-0.089204751,0.0666987225,0.1274843216,-0.0383662023,0.0088963173,0.1928820312,0.1847096384,0.1636067927,0.0680056363,0.5173586607,-0.3658384681,-0.1786599159,0.000730927,0.087186642,0.1694237888,-0.11183507,-0.0815074593,0.0804727376,0.0240682047,-0.2291939557,-0.1171690822,0.1030247584,0.8488240838,0.4012433589,0.2638866901,0.2260148674,0.1123797297,0.0031815011,0.3141271472,-0.4309876561,0.011654608,0.5943624973,0.0966387391,0.2790845633,0.011201038,0.0578476414,-0.369628191,-0.0305520929,-0.1509279162,-0.3673889637,-0.1616608799,-0.1469755322,0.0742135495,-0.2159608006,0.2327011228,0.0244939905,0.1480062157,-0.0258398894,0.0259540733,-0.2149510235,0.715084374,-0.2513640821,-0.3628267348,0.0827531964,0.1390394419,-0.0658968762,-0.0272014551,-0.1854968816,-0.2418183833,0.2443928719,0.0095135113,-0.0690847784,0.0245333798,-0.0135955149,0.0363637023,-0.0897580385,1.0810700655,-0.015928885,-0.0569855459,-0.2624447048,-0.2546088994]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1622","title":"Can't call shape on the output of select()","comments":"Yes, created a PR","body":"I get the error `TypeError: tuple expected at most 1 argument, got 2` when calling `shape` on the output of `select()`.\r\nIt's line 531 in shape in arrow_dataset.py that causes the problem:\r\n``return tuple(self._indices.num_rows, self._data.num_columns)``\r\nThis makes sense, since `tuple(num1, num2)` is not a valid call.\r\n \r\nFull code to reproduce:\r\n\r\n```python\r\ndataset = load_dataset(\"cnn_dailymail\", \"3.0.0\")\r\ntrain_set = dataset[\"train\"]\r\nt = train_set.select(range(10))\r\nprint(t.shape)","comment_length":4,"text":"Can't call shape on the output of select()\nI get the error `TypeError: tuple expected at most 1 argument, got 2` when calling `shape` on the output of `select()`.\r\nIt's line 531 in shape in arrow_dataset.py that causes the problem:\r\n``return tuple(self._indices.num_rows, self._data.num_columns)``\r\nThis makes sense, since `tuple(num1, num2)` is not a valid call.\r\n \r\nFull code to reproduce:\r\n\r\n```python\r\ndataset = load_dataset(\"cnn_dailymail\", \"3.0.0\")\r\ntrain_set = dataset[\"train\"]\r\nt = train_set.select(range(10))\r\nprint(t.shape)\nYes, created a PR","embeddings":[-0.0810442343,-0.448440522,-0.0203575641,0.5366585255,0.4365376234,-0.2255450785,0.1969325244,0.1591523737,0.1245262846,0.2971191704,0.4932901859,0.4957280457,-0.2938623428,0.2642907202,0.0866098851,-0.388372153,-0.1834037006,-0.0725857094,-0.2607069016,-0.0407379158,-0.4226823747,0.1807110757,-0.2416140735,0.3705961704,-0.3218380809,-0.0949034244,0.2719432712,-0.1892885864,-0.2582831979,-0.1613218933,0.4867624044,-0.575411737,-0.2284760177,0.5059751272,-0.0001227193,-0.1656772643,0.2563094795,0.0664867982,-0.0176184736,-0.4644877315,-0.3376430571,-0.263628751,0.3109247088,-0.4003588855,-0.0697052628,0.0958191827,-0.4252294004,0.1407826841,0.3522883058,0.3772268891,0.154947266,0.1494374275,0.0323755853,-0.2121151239,0.047344476,0.5477138758,-0.1462652087,0.1569454372,-0.0408781841,-0.096605815,0.5169168115,0.2898950577,-0.1091238409,-0.0972314551,-0.3117538393,0.3430264592,0.4587482214,-0.1749192923,-0.048476845,-0.016478559,0.3277834356,-0.1233518422,-0.0651844367,-0.0134960786,0.1555029303,-0.1151269078,0.089263171,0.2067295015,-0.2247527242,-0.2363147289,-0.0977918282,0.2039568424,-0.3255081773,0.407130003,0.0976327062,-0.1123478413,-0.0550756492,0.2773999274,0.2207739055,-0.0660730377,0.2118868381,0.1216159835,0.1629530042,0.008867084,0.0603171363,-0.0029032291,-0.0002297509,-0.37706393,-0.2150975764,0.2203426361,0.0464344919,0.0381844603,-0.0961980224,0.2687527835,0.4953636825,0.127137363,-0.1981516033,0.622933507,-0.2506791353,-0.146554634,0.1602905393,-0.0002590035,-0.0413001589,-0.3569540977,0.2039907575,0.1432541907,0.4940039217,-0.0112765888,-0.0710590854,-0.1401797533,0.0010109978,-0.0207535438,0.079763189,0.0895230547,0.046158012,0.0265104584,0.2161242217,0.2439101636,-0.0881297141,0.0047445907,-0.2481705099,0.2396360487,-0.223168686,-0.2340228111,-0.0295937918,0.1969685853,-0.1196092367,-0.0280852038,0.372985214,-0.2715993226,0.0616924055,-0.2778755426,0.1959121227,0.185273245,-0.2213430256,-0.1699109226,0.6077177525,-0.1147207916,-0.1113921925,0.3293159008,-0.3465791047,-0.4822809994,0.0008470003,0.0685442761,0.3426304162,0.1077087224,-0.1228462607,0.1240555644,0.469376117,0.024612518,0.1352027953,-0.5629082322,0.1095781997,-0.2079113722,0.2168492973,0.176891759,-0.4521920979,-0.0872612372,0.0019671721,0.2058628947,0.5516195893,0.1854635775,-0.3273363411,-0.0435867235,0.0542554818,0.1600991189,0.7668824196,0.0973923653,-0.1317916214,-0.2546821535,-0.1502264887,-0.2387654036,-0.2278369814,0.1061070934,-0.3716361821,0.1996467859,0.2436064929,0.0832172483,-0.1223746985,-0.1185391396,-0.0501355007,0.0358989313,0.3412739635,0.0599598438,0.064134486,-0.0096181454,-0.2666327059,-0.3474270403,0.0028790848,0.0815622881,-0.2434237003,-0.4241832495,0.6810166836,-0.0307851266,-0.185056448,-0.2334622294,0.1024297476,0.0540552475,-0.0213977937,0.0781796649,0.0456882715,0.0239646416,-0.2189668715,0.3267194629,0.0285368636,0.4593761861,0.0078644557,-0.2110620439,-0.0157846455,-0.1551178843,-0.1490161419,-0.1665537059,-0.2627483606,0.043499358,-0.249544546,0.2721625566,-0.1043839678,-0.4157685041,-0.0899160355,0.0385525934,0.244740203,0.2617596388,-0.0723417997,0.10261181,0.2384411246,-0.3926970065,-0.4720010757,-0.0801082999,-0.214778617,-0.0980708078,0.0410595685,-0.1334885806,0.0951865539,-0.246344313,-0.021632025,0.4602743983,-0.1699912101,0.3820121884,-0.2695949674,0.3370124698,-0.1975646764,0.1833867729,-0.0439667031,-0.2178823948,-0.1283906847,-0.1868088245,-0.1920999736,-0.2080623955,-0.3216575384,0.0944255367,0.4264374077,0.1369016469,0.2713209689,0.2505295277,0.2146810591,0.1599149704,-0.158043474,0.0534154139,0.4421800375,0.0902534947,-0.1352275908,-0.1598554999,-0.1478198618,-0.1179928109,0.1066171452,0.2031083405,-0.2962730527,-0.0891605616,0.0208927114,-0.1603391171,0.0114792436,0.1894539446,0.0533808693,-0.0114493081,-0.0814308822,0.1859892309,-0.1506511271,-0.3086385727,-0.0696374997,-0.3318116665,0.2145080268,-0.0567340441,-0.1569761932,-0.2405668199,0.0736221448,-0.0310366247,-0.0893721282,0.2650822997,0.3892347813,-0.1885790378,-0.0059396573,0.0819514841,-0.1823845804,0.0037549362,-0.0479230508,0.2662864625,0.2609318197,0.2505930066,0.036589995,0.0217128452,-0.1796160042,-0.0397490039,-0.1610120088,0.0301881488,0.5293599367,0.1090990454,-0.1003758237,-0.0732717365,0.3680845201,0.237463966,0.1340377182,0.3665767312,-0.0437354036,-0.2147997618,-0.0693129376,-0.1980808526,-0.1574750394,-0.1181488186,0.4395238161,0.4891446531,0.2650692165,-0.0084205298,0.3994103074,0.0505997464,0.0302904584,-0.280451715,-0.2244341075,0.0255690906,0.0991249233,0.047760196,-0.1255227625,-0.0561413579,0.1314303726,0.2199629247,-0.1132858023,-0.2488484383,-0.2497840524,-0.1813891977,0.1497071534,-0.2670969665,0.1080466062,0.2908415496,0.2504702806,-0.1026651859,0.1138133109,-0.187885195,0.3628815413,0.7914664745,0.0212379657,0.1433212459,0.4485126734,0.0532733835,0.2772982419,0.0911845341,-0.2908481359,0.0434397496,0.1497602612,-0.1791304946,-0.0892049968,-0.2832585275,-0.0301335398,-0.2511665225,-0.0807303265,-0.2437735051,-0.0907446966,0.0254927706,-0.2094248086,0.4344292581,-0.1547563672,-0.1505004466,0.2331961095,-0.0771869048,-0.0034030771,-0.1718850136,0.1365785599,-0.3716364801,0.1593056619,0.0212079305,-0.2419027984,0.4667959809,-0.0624097846,-0.378903091,-0.3248585165,-0.1193891838,0.3631056547,-0.0798089877,-0.0780945942,-0.1007343829,-0.1271291375,0.0219345745,0.1466200054,0.3869648278,0.0638781041,-0.1221100837,0.3411920667,-0.3207015395,-0.299186945,-0.1154122502,-0.2274173796,0.2656183839,-0.2612388134,0.1275576502,-0.4099822044,0.1777211428,0.373696059,0.1972324401,-0.360456109,0.0188383628,-0.055197455,-0.2555289567,0.2238495946,0.198745057,0.0403482094,0.0669729039,-0.0535465144,-0.1440561414,-0.2700187862,-0.3227533996,-0.161488831,-0.106112957,0.341711849,-0.3877434433,-0.1504971236,0.0896974355,-0.1542555988,0.0985996425,0.4740499556,-0.181906268,0.1413002461,0.0939123183,-0.0853722766,0.446697861,0.1941199303,-0.1052447557,0.0521551184,-0.1218516454,0.2602003217,0.0170154907,-0.162791118,0.5975410342,0.0780544803,-0.0597615726,-0.6645178795,0.4541007876,-0.0255364012,0.0872894824,0.5409530401,0.1936110705,0.0774113685,0.2591387033,0.2110262662,1.1670898199,0.0389210284,-0.0693527088,0.3744927049,0.3166927099,0.2141484171,0.2353299409,0.0112127606,-0.1814055145,-0.3702332079,-0.1123933047,-0.3745963275,-0.042781204,0.2077685893,-0.2791653574,-0.1145881638,-0.5167747736,-0.4453575313,0.033203546,-0.248930037,0.4109396935,0.0352981314,-0.2459571064,0.0905078873,-0.1627560705,0.0320015289,-0.1215785891,0.1346378773,0.068604894,-0.4393228292,-0.2190674096,0.0345423631,-0.3867521286,0.1812264323,0.2189218998,-0.7066283226,0.1858794987,0.3812211156,-0.0986056775,0.126771003,0.0413699858,-0.1500877142,0.4577012062,0.0900539383,-0.1657815874,-0.041638948,0.258256793,-0.0382844135,-0.1779128164,0.4321312904,0.1285254359,-0.177892372,-0.1539597064,0.1312244684,-0.036881417,-0.535025835,-0.6554537416,-0.160224542,-0.0141166467,-0.2137791663,0.0495577343,0.2439855933,-0.1374420673,0.3192067742,0.3910159469,-0.1397670656,0.0561190173,-0.0884322301,0.3146240115,0.1873202175,0.2213303149,0.3963392377,-0.0141464882,-0.1666247696,-0.0555865206,0.1595844179,-0.4215203822,0.2075127959,0.0660622641,0.11285346,0.1200139001,0.3032578528,0.2074245661,0.0246706866,-0.1545776725,-0.1010726616,-0.1060640588,0.3889962435,0.251029551,0.0908254161,-0.1812780201,0.0562700741,-0.0464377329,0.318091929,-0.2020421326,-0.0986995101,-0.2439707518,0.1756654084,0.011228025,-0.0249750763,0.0996169597,0.1679449081,0.0731725246,0.1398207992,0.1478020102,-0.1303932071,-0.0265725218,0.1788624972,0.084704496,-0.0072225225,-0.0074182367,-0.0152501808,-0.2600252926,0.0222829562,-0.1341565847,0.0098835249,-0.2415359616,-0.0062135039,0.167486921,0.0786823034,-0.0048481156,0.1694024652,-0.3719232678,-0.0925126299,0.2327801436,-0.0419246778,0.2999942005,-0.0940979496,-0.3820677996,-0.1765323877,0.0502057336,0.006432944,0.1369892061,-0.3717018664,-0.0393385924,-0.539639771,0.1999913007,0.2091818303,0.0061789639,0.1427614689,0.113549307,0.1006976441,-0.3950443864,-0.2466530055,-0.3561444581,0.1964244395,-0.0066468064,0.3080424964,-0.5246571898,0.3882297873,-0.446523577,0.1767617464,-0.0332074724,-0.2684451938,0.2678388059,0.349150449,-0.144858554,-0.0527342372,0.0372153781,0.2449204773,-0.0165143106,-0.000293428,-0.0627287403,0.2056648135,0.1864418238,-0.2409685701,0.2604153156,-0.0929950625,-0.3145061135,0.0827515423,-0.2199824154,0.0402063429,0.013163208,0.3143455684,-0.2151416242,-0.151289776,-0.3199323118,-0.142523095,0.0116979564,0.2442023307,-0.1183800772,-0.0830043554,-0.0491945557,-0.280297339,-0.0327553079,-0.1090858281,0.5771696568,-0.0107622687,0.1151219532,-0.0644634366,-0.2668882906,-0.0246576313,-0.0388485007,-0.3241917491,0.0232625753,0.4801786244,-0.0164076127,-0.1893214434,0.0426808633,0.3323248327,0.4641762078,-0.3293123841,-0.0684827641,-0.031375654,-0.288239032,0.0298448298,0.0088382969,0.349794805,0.0500692017,0.150397107,0.1348635256,-0.1106357425,0.1542778164,-0.0102848876,0.4663149118,-0.5364181399,0.7102099657,-0.229513377,-0.3411076963,-0.3358285725,0.0049455557,-0.1696669161,-0.087707974,0.2728877366,0.2685566545,0.0893405229,0.220100984,0.0127971508,0.0014876536,0.4662843645,-0.1748277396,0.2501419783,-0.2883366942,-0.0481789596,0.0794204697,0.2155248225,-0.1334902793,0.037704818,0.0737105981,-0.0705049038,-0.0045323726,0.2002752572,0.1126687452,0.2215344608,0.0961978287,0.4994592667,-0.3505905867,-0.2091571093,0.0270421356,0.0723361447,0.1551727355,-0.1319336593,-0.1007075533,0.1205014586,0.0004223504,-0.2571084201,-0.1147305593,0.0836865976,0.9033868313,0.4040676057,0.3495993912,0.2750531137,0.1125499308,-0.0115058944,0.3220203817,-0.397166729,-0.0166366268,0.5977190733,0.0810239464,0.2494258434,0.0182429068,0.0431041308,-0.3890536427,-0.0568206459,-0.1133599654,-0.3312774301,-0.1369463801,-0.1534977257,0.0492529459,-0.2235886753,0.2117848992,-0.0260111261,0.1885564923,-0.0552672558,0.0295143463,-0.1540117115,0.6913304329,-0.2283856869,-0.337223053,0.0977494046,0.0985652506,-0.0863200203,-0.0680218861,-0.1605506986,-0.2323010862,0.2586404681,-0.0328074545,-0.0957883298,0.0044714403,-0.0590825714,0.0155145871,-0.1321324408,1.0939145088,-0.0075476496,-0.0904488042,-0.1956271231,-0.2209260613]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1618","title":"Can't filter language:EN on https:\/\/huggingface.co\/datasets","comments":"cc'ing @mapmeld ","body":"When visiting https:\/\/huggingface.co\/datasets, I don't see an obvious way to filter only English datasets. This is unexpected for me, am I missing something? I'd expect English to be selectable in the language widget. This problem reproduced on Mozilla Firefox and MS Edge:\r\n\r\n![screenshot](https:\/\/user-images.githubusercontent.com\/4547987\/102792244-892e1f00-43a8-11eb-9e89-4826ca201a87.png)\r\n","comment_length":2,"text":"Can't filter language:EN on https:\/\/huggingface.co\/datasets\nWhen visiting https:\/\/huggingface.co\/datasets, I don't see an obvious way to filter only English datasets. This is unexpected for me, am I missing something? I'd expect English to be selectable in the language widget. This problem reproduced on Mozilla Firefox and MS Edge:\r\n\r\n![screenshot](https:\/\/user-images.githubusercontent.com\/4547987\/102792244-892e1f00-43a8-11eb-9e89-4826ca201a87.png)\r\n\ncc'ing @mapmeld ","embeddings":[0.0712649003,-0.203909725,-0.0886661336,-0.1041690558,0.3083346784,-0.1375826448,0.0199502725,0.2050477117,0.1892900616,0.1316726059,-0.3945536315,-0.033305712,0.1600336134,0.3280676603,0.0115702031,-0.0026568682,0.0062019341,-0.2479204684,-0.0557594374,-0.3372556269,-0.0063091144,0.2191529423,-0.0113107096,0.1485607624,0.0023008443,0.1764297634,0.2464780211,0.1226496026,-0.312805146,0.1580780149,0.4409834146,0.0622682646,-0.2356739491,0.0706171989,-0.0001155677,0.1962729543,0.2634896338,-0.2374809831,-0.3352334499,-0.4733934999,0.0596489049,0.0180573482,0.1146264747,0.1466006041,-0.1841839105,-0.1078458279,0.0484989993,-0.5110805631,0.1605988443,0.2958293855,0.1093299985,0.2792661488,-0.239305988,0.202344805,0.1713742465,0.2830444574,-0.1865931749,-0.1926135421,0.5360184908,0.0953860581,-0.0591686517,0.3080681264,0.2046958059,-0.1543609351,-0.0100669153,-0.0748888329,-0.0931045339,-0.3115885258,0.4341978431,0.4047341049,0.3687337041,0.1538343877,-0.3722068965,-0.0459247567,-0.2387470752,0.258292377,0.2565599382,0.3066347241,-0.4852936268,0.3200512826,-0.1850732863,-0.6047545671,0.1087156758,0.4450512528,-0.1151872054,0.8385538459,0.011042783,0.1085000336,0.3898685873,-0.1659885645,0.079401724,-0.1831424236,0.1976145208,0.261328876,0.002512235,0.2438256592,0.0016086893,0.556172967,0.2018950135,-0.0784955099,-0.5124129653,0.48327595,-0.10160546,-0.0314961821,0.3086353242,-0.0942561179,0.3645786047,-0.0565553233,0.4635268748,-0.033798296,0.0163634215,0.0319390111,0.1787707657,-0.0126495371,-0.3762994707,-0.2955794632,0.0698452443,-0.415441364,-0.3156930804,-0.089760989,0.027304858,-0.4920036495,0.3816134334,0.208004728,0.2462244779,-0.0082501322,-0.0835864022,0.3539087772,-0.3154687285,-0.3446996212,-0.0641522408,0.0836094841,-0.4090605676,-0.2925333679,0.1264155954,-0.4408020377,0.1894478351,0.0851460621,0.0243218821,-0.1159978881,-0.2009438127,0.1619760394,0.1632102877,0.1534820646,-0.1358368397,0.2921904624,0.4921875,-0.6711062789,-0.0357447527,-0.0422422513,-0.339440316,0.0618739948,-0.421248585,0.0394224674,-0.1475532651,-0.2016592473,0.2313350141,0.6244534254,-0.0572226644,0.0203815158,0.1073547527,0.3173674643,-0.3983130157,0.2762416303,0.1082000583,0.1727572083,-0.4608765543,-0.3962003589,-0.0496789999,-0.2595998943,0.1555467546,0.1666608304,-0.1157277375,-0.5279542208,-0.0661848262,0.1150735319,0.2018103302,-0.0606534481,-0.3337161541,0.0604692958,0.0656847432,0.0518145971,-0.1414961964,-0.1272581667,-0.1632457525,-0.0003639027,0.0861877874,0.1503097862,-0.0540120527,-0.1687471569,-0.0669957697,-0.0620272048,0.6305915117,0.1910134703,0.1610699296,-0.1329347342,0.1740878522,-0.0096104098,0.3087939322,-0.2276963443,0.0003904953,0.1484153122,0.1585028023,0.1505011171,0.1637895405,-0.1347787976,-0.2974293828,-0.0764401108,0.0667651445,0.4012443125,-0.110983111,-0.1153440773,-0.1670814008,0.0385651737,-0.2971960902,-0.2686656713,0.068510294,0.1331034154,-0.2421048433,0.279022783,-0.1503769904,0.0037618354,-0.0408320501,0.0402018689,-0.1638589948,0.0810517222,-0.0219202824,0.0792077407,-0.2360189408,0.1624454409,0.054734312,-0.1304504275,-0.1553989202,-0.0929353014,0.3744748831,0.0154421264,0.490000844,0.1132253408,0.3263052404,-0.2689576447,-0.0508305393,0.259190321,0.0754899234,0.1214104071,-0.4280153513,0.2211889178,0.187098667,0.34717381,0.1697783768,0.1524695754,0.5072755814,0.0195484329,-0.3341378868,-0.488673687,0.4389429688,-0.0283279959,0.1989305913,0.1239012107,-0.506401062,0.1154176965,0.3294615448,0.2849546969,0.4273667336,0.377458334,0.0363669023,-0.2009702921,0.1783004403,-0.0026062743,-0.1111702174,0.1506242603,0.3659340441,0.0481176674,0.1113682836,-0.2801690996,-0.0019762702,0.0750682428,-0.0121618854,-0.1603061557,0.2949730456,-0.3922141194,-0.8334483504,0.4709528387,-0.0834686831,-0.1360102296,0.0379438475,0.3020277321,-0.6944234967,-0.4544697404,0.109899424,-0.039500311,-0.5918177962,-0.3072559536,0.3294537365,-0.152127102,-0.0496314578,0.0690049753,0.0387725793,0.0894497484,-0.2400896251,0.0454905219,-0.3391710222,-0.2388224751,-0.2403541803,0.0097292298,0.3720571101,0.2128883153,0.198884666,-0.1167152971,0.1409712732,-0.1453161836,-0.7914140821,0.1717515439,0.0132543426,0.4704160988,-0.0930741429,0.1926689744,-0.3161944747,0.1308477819,0.4974772334,-0.0545007847,-0.2133378386,0.1762443781,0.1208089888,0.0761057138,-0.0363448709,-0.239923805,-0.4338032901,-0.0016850309,-0.3799986243,-0.1961051524,-0.0796064064,0.0066597769,-0.1504183859,-0.1135632172,-0.4480303824,-0.0562914163,-0.419439286,-0.1335778087,0.2822260559,0.0476596057,-0.0731578767,0.1890612096,0.2412533015,-0.0512308888,0.0269752517,-0.2688790262,-0.0765325651,-0.1560602188,0.1199483946,0.1128923744,0.0838463157,0.2787014246,-0.0938100293,0.1023631766,-0.2217251807,-0.3002575636,0.0581806749,-0.1103226617,0.3345350325,-0.1413022876,0.3879980445,0.3063770533,0.3826068044,-0.0162334554,0.3615572453,0.7596259117,0.3340232074,0.4011025429,-0.1775020808,-0.3090931475,0.1661457717,-0.0449451506,0.0853889659,0.1878568679,0.2542714179,-0.2463531643,-0.2662661374,-0.1878265291,-0.2441607714,-0.1634013802,0.0282877013,0.1910261959,-0.0127448756,0.2129616886,0.1248001233,0.2771047056,-0.1216549128,-0.0715501234,0.451603502,0.0112356236,0.2100963891,-0.3773725927,-0.3749032617,-0.2237638235,0.1521587819,0.07607463,0.454819262,-0.4694688022,0.0913868919,0.4561005533,0.1276338696,0.4923042357,-0.2783865929,-0.1922507733,-0.2048921883,-0.3398918211,-0.3169366419,0.136370182,-0.0246188305,-0.1060378402,0.3287743628,0.1526444554,-0.3777374327,-0.3466691673,-0.2282950133,0.3110028803,-0.091993548,0.0635270104,-0.15032354,-0.2430593371,-0.3747086227,0.2916134596,0.0111535918,-0.0125994347,0.1495083869,0.1398875117,0.1269243062,0.2980006933,0.0141032496,0.1026981622,-0.2368700653,0.1163388193,-0.4076665938,0.2278395444,0.2769062519,-0.2171473801,0.6303996444,0.0644657612,-0.4821489751,-0.1143561751,-0.1218727082,0.232590124,0.2908240259,0.2083908468,0.1868188083,0.4718681574,-0.1171726063,0.2331198454,-0.034993194,0.1160567179,0.354291141,-0.4070611298,-0.1926321089,0.2785891891,-0.2177587301,0.0837334916,0.1061876193,0.5938468575,-0.1112322286,0.0016801639,0.0356728174,0.8598227501,0.0114701129,-0.0422550291,-0.0632631928,-0.1351647973,0.1898655295,0.0051230062,-0.1498517394,0.0696624219,-0.0685011521,-0.1525757611,0.1334790736,0.1347663999,-0.0340468399,-0.0614738427,0.2230993509,-0.2211748958,-0.3094959855,0.4750598073,0.2038014382,-0.2874107361,-0.0428198539,-0.1119717285,0.0500384904,-0.0147251915,0.4676228762,0.1972468048,-0.0219572261,-0.0670800507,-0.3027565777,-0.4004336894,0.0365603231,-0.1964637637,-0.0187329426,-0.050683856,-0.4862181544,0.3315670788,0.0023576503,0.3614227474,0.2044858634,-0.2563692629,0.3900311589,-0.0011372836,0.0186914951,-0.1889792234,-0.0003651979,0.3222512603,-0.0926529765,-0.4408873916,-0.1057565436,0.0562024564,-0.1629535556,-0.2878610492,0.0097810244,-0.0238246098,0.3092347682,-0.3322202265,0.0132238511,0.1192681193,-0.3350393176,0.0213719048,0.1368556023,0.210725382,-0.3415746987,0.4144303501,-0.1305995882,-0.2066421956,0.2852000296,0.0647995248,0.1260679364,0.3355803192,0.4387979805,-0.3554387987,-0.1083728895,-0.0532358699,0.1310398281,-0.132474497,-0.3146373928,0.1057740003,-0.120007135,-0.1701256186,-0.0308883768,-0.2209148705,-0.0169366319,-0.2928876281,-0.4045007527,0.2457154691,0.1522338539,-0.1934017688,-0.1423903555,-0.0104840193,-0.196358338,0.2197694182,0.0458058715,-0.1691336334,-0.0889849588,0.0910511091,0.0169191808,0.1376324296,0.296662569,0.0636887774,-0.2114676684,-0.0372050181,-0.2118415385,-0.0520336255,-0.1175607517,-0.0983525515,0.0242402013,0.2368707359,-0.2630985081,0.3164887428,0.0838194788,0.2777095139,-0.1751616895,0.1103627235,0.2720095515,0.0717302561,0.0422530323,-0.1457779408,-0.1232271045,0.3492973745,-0.0251401793,0.0181821976,0.0012384978,0.1706898212,-0.0504945815,0.3915721774,0.0299105961,-0.1185261533,-0.0604476631,-0.0089885304,0.3398258388,-0.3278770745,-0.0873255879,0.2653989792,0.0519300289,-0.0317687094,0.1901384294,0.0018098281,0.2365339845,0.3689267933,0.0932446644,-0.3170773983,-0.283860296,0.3992066085,0.0300887171,0.0849805102,0.0444642156,0.0696303248,0.1965900809,0.5751878023,-0.2572990954,-0.0610420555,-0.2396674156,0.3775023222,0.1689632237,0.0144324219,-0.0286073461,0.2725023627,-0.059081953,-0.1945975572,0.2466426194,0.0532256439,0.1483435929,0.0785846412,0.2926907539,-0.0862934068,-0.1950780898,-0.339104414,0.0055466737,-0.1505809128,0.0282158293,0.135961026,0.0614703447,0.1552891433,-0.2563390732,-0.1886484027,0.4413140118,-0.0399709903,-0.2145157158,0.0158637036,-0.2117143273,-0.0312073547,0.2065152824,-0.0624163486,-0.2431289405,0.0497769117,-0.1008213013,-0.055511646,-0.2424364239,0.0972600728,-0.221978128,-0.2190913558,0.127073586,0.3672893941,-0.015232618,0.0385338478,0.0334603041,0.0744959339,-0.1084097847,0.1096876711,0.0543491207,-0.1116965264,-0.2205373049,-0.0845597014,-0.1444923878,0.3815061152,0.2566306591,0.0975165069,0.2900044024,0.0829823539,-0.0382903218,0.2038792223,0.120549649,0.0187793002,-0.1002659798,-0.2331330627,-0.2339171469,0.5030563474,-0.3746004105,-0.3565005064,-0.2633960247,0.3633213639,0.082061559,-0.1590732187,0.0050514019,-0.2541117072,0.0376752205,0.1089991331,0.1569944024,0.3069289625,0.2274844199,0.1588534117,-0.3174373806,-0.4950583577,-0.088460058,-0.1966401637,0.0395246223,0.0046345145,-0.098652564,0.3217895031,0.2754671276,0.3676724732,0.080314599,-0.0894379392,0.1323614568,-0.1470140219,0.0420907401,0.255567342,0.2716406584,-0.0091772703,-0.0107892537,0.1540186405,-0.072639741,-0.0700674653,-0.187438786,0.3717726469,-0.2337078303,-0.3814877868,0.2093182951,0.1271326691,0.5057985783,-0.1662695557,0.1854801774,-0.0086603612,-0.2380880713,-0.3505830169,0.5087690949,0.1734792292,0.2276091576,-0.055408705,-0.1669967771,0.0535807833,0.3670722246,-0.145240739,0.2868474424,0.0754509643,0.0530662388,0.2478293628,-0.2227004766,-0.085566543,0.1000543311,0.1163361073,-0.3260931671,0.0109937051,-0.1401001066,0.3009521961,-0.1136833876,-0.3456959426,-0.156965524,-0.1674185991,0.3713665307,-0.30137977,-0.1691807806,0.0259526577,-0.0576219633,0.0109374691,0.0615954064,0.1418251991,-0.0753642693,-0.1327927858,-0.3029811978,0.223247081,0.3998853862,-0.0319559537,-0.0014904335,-0.102285251]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1618","title":"Can't filter language:EN on https:\/\/huggingface.co\/datasets","comments":"Full language list is now deployed to https:\/\/huggingface.co\/datasets ! Recommend close","body":"When visiting https:\/\/huggingface.co\/datasets, I don't see an obvious way to filter only English datasets. This is unexpected for me, am I missing something? I'd expect English to be selectable in the language widget. This problem reproduced on Mozilla Firefox and MS Edge:\r\n\r\n![screenshot](https:\/\/user-images.githubusercontent.com\/4547987\/102792244-892e1f00-43a8-11eb-9e89-4826ca201a87.png)\r\n","comment_length":11,"text":"Can't filter language:EN on https:\/\/huggingface.co\/datasets\nWhen visiting https:\/\/huggingface.co\/datasets, I don't see an obvious way to filter only English datasets. This is unexpected for me, am I missing something? I'd expect English to be selectable in the language widget. This problem reproduced on Mozilla Firefox and MS Edge:\r\n\r\n![screenshot](https:\/\/user-images.githubusercontent.com\/4547987\/102792244-892e1f00-43a8-11eb-9e89-4826ca201a87.png)\r\n\nFull language list is now deployed to https:\/\/huggingface.co\/datasets ! Recommend close","embeddings":[0.1120311394,-0.2155438811,-0.1076241881,-0.1051538885,0.3186013699,-0.097181417,0.0044593154,0.2149730772,0.2302737683,0.1157945022,-0.3949679434,-0.0211918671,0.1348247528,0.3168137968,0.0361360237,-0.0104697971,-0.012229695,-0.2563101351,-0.0395958126,-0.3618176579,-0.01815819,0.2271939069,-0.0093411673,0.1427730322,0.0398445837,0.1924277395,0.2289363891,0.1136408299,-0.304995805,0.1055653393,0.4417517185,0.0045837965,-0.22409527,0.0421747863,-0.0001144573,0.1883409023,0.268116504,-0.2337490767,-0.3595429063,-0.5061581135,0.0589805283,-0.01650385,0.1243720651,0.1144398674,-0.1792945862,-0.1101235822,0.0440315679,-0.5277249217,0.1793083251,0.3008905649,0.1203853786,0.2735803425,-0.2166124135,0.1867288798,0.1646858156,0.2502915561,-0.1821792871,-0.1745153368,0.5653476715,0.1209592745,-0.0642866418,0.2877334952,0.2086322606,-0.1767961979,-0.027709607,-0.0740753338,-0.0838215798,-0.3479908705,0.4390347004,0.4125545323,0.3513755202,0.1383458972,-0.3747716248,-0.0430838354,-0.2192296386,0.2165109664,0.2373345792,0.3184575737,-0.4999711215,0.3170444667,-0.1696970165,-0.6163952351,0.1039317474,0.4199258983,-0.118719466,0.8482972383,-0.0052762916,0.0905176476,0.3776390851,-0.1412192434,0.0747989416,-0.1940534413,0.1897844076,0.271895498,-0.0119173871,0.2583353221,0.0379228592,0.5404688716,0.2091648728,-0.0585809238,-0.498776257,0.4788305163,-0.1025350839,-0.0354847461,0.291893959,-0.0639307797,0.3769117296,-0.0721852779,0.4319868982,-0.0308555253,0.0672808439,0.0393010825,0.1540822685,-0.0023991733,-0.3998157084,-0.3008528054,0.0537242815,-0.4041450024,-0.3408955634,-0.0671953782,0.0100269457,-0.4760325849,0.3784677684,0.2316599786,0.2582052648,-0.02121282,-0.0762191415,0.322614193,-0.2820214033,-0.31064412,-0.0700309873,0.0649016276,-0.429577738,-0.2937310636,0.1091246456,-0.46873492,0.2114049196,0.0892856941,0.0363471247,-0.0873654485,-0.2137511224,0.167890355,0.1746302098,0.1263638139,-0.1101218238,0.2924096286,0.5167384744,-0.6423665881,-0.0323215835,-0.0353688598,-0.3255827129,0.0658938214,-0.4093610346,0.0467116609,-0.1560136378,-0.1943836212,0.2345942557,0.6182880402,-0.0351016968,0.0723646581,0.1242142916,0.3179014027,-0.3916143179,0.2450409085,0.1235217452,0.1831074804,-0.4521826804,-0.4047275782,-0.0314627253,-0.2393603027,0.1583274156,0.1819259822,-0.1347611994,-0.5159894824,-0.0964190662,0.0826371536,0.1765784174,-0.0156956557,-0.3378778398,0.0896264091,0.0241585411,0.0503077284,-0.1384800375,-0.1312785894,-0.1268920749,0.0056333658,0.0972217694,0.13719666,-0.0579507947,-0.1927022934,-0.0836405978,-0.0653851405,0.6196441054,0.1692211926,0.1618299782,-0.1331748515,0.1471350342,0.0242928863,0.3361760974,-0.2165207267,0.0175903421,0.178276062,0.1666291356,0.1928970665,0.1758861393,-0.1339346021,-0.3218634427,-0.059037298,0.0420511104,0.4022675753,-0.120913133,-0.0885327458,-0.1788202822,0.0485464595,-0.2895299196,-0.2746122479,0.081682235,0.1406804323,-0.2365437746,0.3107298315,-0.1547267288,-0.0120535176,-0.0663098544,0.0402437858,-0.2031192631,0.0900440738,-0.0218844693,0.0819720849,-0.2212992162,0.1633861214,0.0516433604,-0.1742553115,-0.1556434333,-0.1032257155,0.3930206895,-0.0038711659,0.4741039276,0.1074234396,0.338729471,-0.3001138866,-0.0432531238,0.2748066187,0.0708666444,0.1312387586,-0.4056914151,0.2212802321,0.1336282641,0.3869348466,0.1573727131,0.19164823,0.5241050124,0.0275678337,-0.3398288786,-0.4691794813,0.4551407397,-0.0117823342,0.211574778,0.1090026051,-0.4846510291,0.1153237671,0.3246816993,0.2996408641,0.3962058127,0.3725768924,0.0452950895,-0.176377207,0.1757803857,-0.0318450034,-0.1087485477,0.1612419933,0.350482136,0.0430148728,0.1358725727,-0.2899954915,0.0256336965,0.0833778903,0.0027306816,-0.1383337975,0.2985559106,-0.3724452853,-0.7993243933,0.4308419228,-0.0624503866,-0.1318611354,0.0596467331,0.2711990774,-0.6991379857,-0.436563462,0.1000925228,-0.0938142017,-0.5903603435,-0.293751061,0.3032997549,-0.2026658505,-0.0530889332,0.0850410387,0.0196970887,0.0702312887,-0.2596761882,0.0421805456,-0.3292359412,-0.2338102609,-0.2469685078,0.0242743716,0.3838761151,0.2071104348,0.2090962976,-0.1018813998,0.1041693315,-0.1403813809,-0.8133555651,0.1855370849,0.0245240964,0.4382332861,-0.0739241838,0.1892478317,-0.3409240544,0.1052319333,0.4869645238,-0.0514197275,-0.2153250277,0.1790964603,0.1294008791,0.0737120584,-0.0329983607,-0.2660450935,-0.4398469031,-0.0092964033,-0.3581671119,-0.1730986387,-0.0436964259,-0.0260819178,-0.1956897676,-0.0844984949,-0.4665608406,-0.0636109784,-0.3857705593,-0.090283826,0.2481485605,0.0533263758,-0.0977281556,0.2143172026,0.2227563709,-0.0137598664,0.0052484642,-0.2989774942,-0.0602851138,-0.1738796532,0.1506641656,0.0825978816,0.0695962086,0.2936536074,-0.0982539877,0.0905273929,-0.2186816782,-0.3055318892,0.0274623502,-0.1469065398,0.3695085943,-0.1333203614,0.382061094,0.3290556371,0.3918676376,0.029314097,0.3648798168,0.728402555,0.3427478075,0.3821211159,-0.1820872873,-0.3398932815,0.1409284323,-0.045234222,0.0894172117,0.1858822405,0.2586691678,-0.2272899151,-0.2510143518,-0.157592684,-0.2209074795,-0.155969575,0.042930685,0.1672978699,0.0394123271,0.231786713,0.1218779758,0.2904919684,-0.131463781,-0.0712638572,0.4473520219,0.0302475132,0.1951948404,-0.3649285436,-0.3943358958,-0.2149038166,0.1239273548,0.084650889,0.468010664,-0.4666343927,0.1197471693,0.4367208481,0.1478333473,0.4935410023,-0.2566252649,-0.1947069466,-0.1690028459,-0.3233855665,-0.313498646,0.1579969376,-0.0374303497,-0.1090682745,0.3835737407,0.1526128948,-0.3815174699,-0.3396859765,-0.2420625836,0.302746594,-0.0895149782,0.0680898353,-0.1567274928,-0.2535893023,-0.3728054166,0.3014578819,-0.0007712842,-0.0118069677,0.1261574477,0.1582237929,0.1476901472,0.2811851501,-0.0271504093,0.0775865838,-0.2310688496,0.1361652911,-0.377350688,0.1974871457,0.2720196545,-0.2064410597,0.6487461329,0.0506878942,-0.4973319769,-0.0810481533,-0.1218702942,0.225520283,0.2946873307,0.2121329308,0.2020648569,0.4631165862,-0.1225869507,0.2496923506,-0.0441319272,0.1234133914,0.3303916752,-0.3896813989,-0.2096736878,0.258423239,-0.2308732271,0.0849537775,0.104556419,0.5696938038,-0.1266635209,0.016422281,0.0689750016,0.8592752814,0.0118932044,-0.0512237214,-0.0676823929,-0.1429949403,0.2415514141,0.0671175718,-0.1559198797,0.0427581929,-0.0668419003,-0.1374647468,0.1114680544,0.1385151446,-0.0426659919,-0.0817625895,0.2111934721,-0.185306251,-0.29040429,0.4768542051,0.1970906407,-0.290766716,-0.0852174312,-0.1486865282,0.0547248162,-0.013887438,0.4871543944,0.1739480644,-0.0454733893,-0.0792745948,-0.3181619048,-0.433619231,0.0350196473,-0.2215066105,-0.0042509111,-0.0666092336,-0.5058730245,0.3729265332,0.0054541039,0.3769128025,0.2376338392,-0.2666000128,0.374650687,-0.0049311584,-0.0010145231,-0.1615674198,0.0087547302,0.2896641791,-0.0985525772,-0.4565821588,-0.1162096411,0.0455793627,-0.1398112029,-0.3063097894,0.0421901383,0.0112403613,0.3315787315,-0.3271600604,0.0309188794,0.1063601598,-0.3374302387,0.0333221033,0.1397780329,0.1629459262,-0.3492730856,0.4133201838,-0.1051888689,-0.2226077765,0.2896215022,0.0923855752,0.089498587,0.3390019536,0.4615411758,-0.3525945842,-0.1221927255,-0.0345773548,0.0904573649,-0.1401395798,-0.3077332079,0.1120960861,-0.0751901492,-0.1932171881,-0.0330829918,-0.2008970082,-0.0147514576,-0.2953464389,-0.4184730351,0.2229775488,0.1710959673,-0.1959963143,-0.1764539629,-0.0415519178,-0.151652202,0.2014570087,0.0388323069,-0.1790861487,-0.0899479836,0.0686907619,0.0277999137,0.1296998411,0.2859016061,0.0732748508,-0.1853261143,-0.0315016173,-0.1689818799,-0.0603178553,-0.1314495355,-0.0840753615,0.0182945151,0.2378659546,-0.2608808577,0.2782637179,0.0760700703,0.2821528912,-0.1561598927,0.1029899195,0.2580300272,0.0828597099,0.0387868024,-0.132813707,-0.1150907427,0.3535358906,-0.0172529053,-0.0280713253,-0.0053205327,0.1891827732,-0.020043727,0.3975832164,0.0476846807,-0.1357571334,-0.0758917406,-0.032479126,0.3543391228,-0.3002408743,-0.1062132195,0.2705930769,0.0619662292,-0.019343216,0.1940082461,-0.007890949,0.2368567288,0.390094161,0.0954182595,-0.32243523,-0.2728776634,0.3895712495,0.0622533783,0.0941218808,0.0629193634,0.1000322849,0.1647623479,0.5825904012,-0.2589819133,-0.0307009965,-0.2876173258,0.3642105162,0.1908920705,0.0501729436,-0.0113997301,0.2687364817,-0.0586696342,-0.1908094585,0.2208967358,0.0824774802,0.1683085412,0.0589620918,0.2954146266,-0.0772754028,-0.2363804728,-0.3361929953,0.0195006542,-0.1749787778,0.0347999968,0.1233355775,0.0844358429,0.1429627091,-0.2464597821,-0.1666059643,0.4248023331,-0.0551519692,-0.1892596185,0.0360207371,-0.2122129351,-0.0143470904,0.190067932,-0.0330294035,-0.2618407905,0.0358783565,-0.110321708,-0.0783036947,-0.2803925276,0.051015716,-0.1862318218,-0.2330653518,0.1339383274,0.3796361983,0.0080277687,0.037812259,-0.014792392,0.0669371113,-0.1465551108,0.1049454883,0.0832323506,-0.1107471511,-0.2035045922,-0.0691311508,-0.1622146517,0.4190335274,0.2568343282,0.1044526547,0.2764020264,0.0870141014,-0.0405758694,0.2325242609,0.1293253303,0.0112956567,-0.0840232298,-0.1968409866,-0.2180151194,0.4770917594,-0.3992695808,-0.3500246406,-0.2802976966,0.3517712951,0.057285551,-0.1621023864,0.0070662224,-0.2458392829,0.0520906337,0.1046433747,0.1212040335,0.2964487672,0.248434484,0.1509830803,-0.3503214121,-0.4745776057,-0.0843739212,-0.1956329197,0.0469326973,0.015086201,-0.0773086175,0.3101496398,0.2662801147,0.3402161896,0.0772052035,-0.1144307703,0.1488952637,-0.1540661007,0.0498516783,0.2406947911,0.2896236777,-0.0295773353,0.0111492202,0.1668169796,-0.059530165,-0.0689276829,-0.200897947,0.3713825941,-0.2420729697,-0.3273513913,0.194400087,0.1543398201,0.4877352417,-0.1634872854,0.1494901031,-0.0449660383,-0.2719577253,-0.3588313758,0.5270806551,0.1826506555,0.2184613794,-0.0684086904,-0.2510745227,0.051751256,0.3741071522,-0.1371735483,0.2624453902,0.0261291265,0.0355887227,0.2356258631,-0.2076958567,-0.0890196115,0.1177430302,0.1378079802,-0.3427234888,0.0170849171,-0.1289710104,0.2869729698,-0.1058070436,-0.3378205001,-0.1504335701,-0.1341374815,0.3393765092,-0.324749887,-0.1525639892,0.0043834918,-0.019981904,0.0325764827,0.0641651675,0.122162886,-0.0856664553,-0.139870435,-0.3027666211,0.2001475692,0.3930492103,-0.0498169437,0.0373952314,-0.1194087416]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1618","title":"Can't filter language:EN on https:\/\/huggingface.co\/datasets","comments":"Cool @mapmeld ! My 2 cents (for a next iteration), it would be cool to have a small search widget in the filter dropdown as you have a ton of languages now here! Closing this in the meantime.","body":"When visiting https:\/\/huggingface.co\/datasets, I don't see an obvious way to filter only English datasets. This is unexpected for me, am I missing something? I'd expect English to be selectable in the language widget. This problem reproduced on Mozilla Firefox and MS Edge:\r\n\r\n![screenshot](https:\/\/user-images.githubusercontent.com\/4547987\/102792244-892e1f00-43a8-11eb-9e89-4826ca201a87.png)\r\n","comment_length":38,"text":"Can't filter language:EN on https:\/\/huggingface.co\/datasets\nWhen visiting https:\/\/huggingface.co\/datasets, I don't see an obvious way to filter only English datasets. This is unexpected for me, am I missing something? I'd expect English to be selectable in the language widget. This problem reproduced on Mozilla Firefox and MS Edge:\r\n\r\n![screenshot](https:\/\/user-images.githubusercontent.com\/4547987\/102792244-892e1f00-43a8-11eb-9e89-4826ca201a87.png)\r\n\nCool @mapmeld ! My 2 cents (for a next iteration), it would be cool to have a small search widget in the filter dropdown as you have a ton of languages now here! Closing this in the meantime.","embeddings":[0.0697203875,-0.1904691607,-0.0992743149,-0.1423217207,0.3043785393,-0.1967637241,-0.0242899098,0.2436978966,0.2299478501,0.0734315589,-0.3851575255,-0.0264967307,0.1065776572,0.3587076068,0.0037252805,-0.0293827187,0.0082300613,-0.242115587,-0.0193542466,-0.3352008164,-0.0219079703,0.2305799723,0.0144517198,0.1968016028,0.0344665274,0.1441680491,0.2395025939,0.1189103574,-0.336981833,0.161104992,0.4430655837,0.0839436278,-0.2019207776,0.0722684786,-0.0001132097,0.1843021512,0.2821063697,-0.2375727147,-0.3122716844,-0.4901600182,0.0833079889,0.0298773609,0.0978154019,0.1604104936,-0.1911302358,-0.0888534859,0.0164860357,-0.5040478706,0.1358761489,0.2542887032,0.1379289478,0.2267963141,-0.230095014,0.182113722,0.1706558913,0.2834809422,-0.1793498844,-0.2151714116,0.614452064,0.0103617357,-0.024525905,0.3743491471,0.189716965,-0.2065367103,-0.0197540335,-0.0695601776,-0.0864234343,-0.3363251388,0.4389978051,0.4179053307,0.3462484777,0.1578739285,-0.369800508,-0.034940578,-0.2311831117,0.2896886468,0.2406505644,0.2826818526,-0.4962031245,0.292322427,-0.1496741176,-0.622476697,0.1252483577,0.4565009475,-0.1259596497,0.8481764793,0.0109492494,0.0648580492,0.3868456185,-0.2022313327,0.0874971673,-0.1736609042,0.1670818478,0.2750412822,0.0258528944,0.2047601789,0.0499382615,0.5662867427,0.1952952743,-0.068339631,-0.5048609376,0.4946347475,-0.0847920105,-0.0357238911,0.2893125713,-0.1108626351,0.3451492786,-0.0452296995,0.4507942498,-0.0567076541,-0.0216337461,-0.030431971,0.1588544399,-0.0109640537,-0.4000426233,-0.3252842724,0.1004615277,-0.4018624723,-0.2509342134,-0.0828103423,0.0253898688,-0.5085737705,0.372757256,0.2321229726,0.2330646813,0.0200557187,-0.0536398739,0.3188359141,-0.2947303355,-0.3379003704,-0.0659277886,0.0888256207,-0.3805269599,-0.2679069042,0.0916711017,-0.4386591911,0.1977417767,0.0660373792,0.042881541,-0.0982045755,-0.1369184852,0.1586972028,0.1299182922,0.2043980062,-0.1609536558,0.259979248,0.503883481,-0.7266700268,-0.0265471209,-0.0072138193,-0.3502404988,0.039247971,-0.4442353845,0.0769849792,-0.1553951055,-0.1846591979,0.2390430719,0.6273361444,-0.0862734243,-0.0393217131,0.0991693884,0.2599027455,-0.3728788495,0.2773992717,0.1285035014,0.1500095427,-0.5414364338,-0.3975726664,-0.1406737268,-0.204844594,0.1602687091,0.1611751318,-0.1280632168,-0.4564213455,-0.0561366789,0.1231921762,0.2289880216,-0.0248813666,-0.325927943,0.025333602,0.0851873457,0.0772140548,-0.1757469028,-0.1210657507,-0.1756032258,-0.0099734394,0.0850169137,0.1311845332,-0.1131326109,-0.181431964,-0.1108245179,-0.023086803,0.6042264104,0.2240638286,0.2061372846,-0.1574541479,0.1637067944,0.0172632169,0.2774575651,-0.2118106484,-0.0415724181,0.1440002471,0.2356201708,0.1567976475,0.1479810029,-0.1542041749,-0.2422254682,-0.0708123371,0.0586463436,0.3345796466,-0.0909061357,-0.0999172032,-0.1584774703,0.055048652,-0.2936167717,-0.2847729623,0.1034499854,0.0707144216,-0.2660553753,0.2952750027,-0.1530410647,0.003563449,-0.0228449423,0.012876993,-0.1686631292,0.1423533708,-0.0322412215,0.0786100999,-0.2693579197,0.1288791001,0.0960844383,-0.090056181,-0.1398057938,-0.1038971171,0.3578149378,-0.0511831976,0.5464028716,0.0946927369,0.3221946359,-0.2679735124,-0.0512743555,0.2520977557,0.0933540463,0.0747131258,-0.4248587489,0.2352568805,0.2121915817,0.3593796194,0.1562715471,0.1533129811,0.477858156,0.0400025323,-0.3378351927,-0.4407450259,0.4604885578,-0.0364161246,0.1299155354,0.0974904224,-0.4817985594,0.1562607884,0.3481703401,0.2584572732,0.4431127608,0.4171910882,0.0544425547,-0.159989208,0.1677643061,0.0575118698,-0.0727854818,0.1639977396,0.3393564224,0.045149561,0.1474932879,-0.2835951447,0.0294036139,0.0779328346,-0.0032880101,-0.1536536217,0.2630381882,-0.388395071,-0.7758849859,0.3943957984,-0.1288506389,-0.1343992651,0.0731743425,0.2974516749,-0.6716725826,-0.4274066389,0.1434940249,-0.0696998909,-0.6090425253,-0.3300592899,0.3609874547,-0.1772851497,-0.0729263127,0.087815918,0.0043422952,0.1457873434,-0.2441672832,0.0668262616,-0.3475687504,-0.2859016657,-0.2279474437,0.0368862636,0.3765652776,0.2082701027,0.2132933736,-0.0707803145,0.0950175598,-0.1185419708,-0.8488957882,0.1651747227,0.043694146,0.4854832292,-0.0589543432,0.1918137968,-0.336092025,0.0960922316,0.4884258807,-0.034146715,-0.2212391794,0.1713183373,0.1258920282,0.1197915524,-0.0225212928,-0.2072440684,-0.4753740728,-0.0032552248,-0.3889010847,-0.1890711635,-0.0813273713,0.0611331128,-0.1392064393,-0.1334471852,-0.4243969321,-0.0804809481,-0.4438278675,-0.1025070325,0.2734499276,-0.0141879087,-0.0718323663,0.1732351482,0.2118355781,-0.0628603697,-0.0226121489,-0.252859503,-0.0648088232,-0.1417482942,0.0499465913,0.1261299998,0.0976682827,0.2344747782,-0.0770998448,0.0769518837,-0.2106241882,-0.24988392,-0.0035374023,-0.1133511737,0.3706026673,-0.0966546908,0.3994728327,0.2778926194,0.4070092738,-0.0409842841,0.3978936076,0.729614675,0.3698035181,0.3786163628,-0.1808141023,-0.3395781815,0.1518745571,-0.0530514792,0.0674843043,0.1528007835,0.2572992444,-0.2912737429,-0.254226625,-0.1762201935,-0.2075944543,-0.1326984167,0.0750071928,0.2263453901,-0.0904252529,0.2002207786,0.0971447006,0.2921693027,-0.0977365077,-0.0951216593,0.4100566804,-0.0252291858,0.2025512606,-0.3146877885,-0.4060757756,-0.2273303866,0.1209266335,0.0852564797,0.3663113117,-0.4623540044,0.0997481346,0.4209723771,0.1423618346,0.5113817453,-0.3275405765,-0.2169968933,-0.1932693124,-0.36658746,-0.2917199135,0.133318916,-0.0415318199,-0.0890664086,0.325494051,0.1138990074,-0.297760427,-0.3260324299,-0.2396512032,0.2806367278,-0.0741269365,0.0695190653,-0.1420497149,-0.2998773456,-0.3416671455,0.2777191401,0.0276308469,0.0101212738,0.1436524391,0.1166003123,0.1328862309,0.2618616819,0.0203748755,0.1074721292,-0.1997938156,0.1203533337,-0.3877083957,0.1785957813,0.2634789348,-0.2361579239,0.5840196609,0.0544875413,-0.4106300473,-0.1029427946,-0.1549547017,0.2495317459,0.3277280927,0.2176396698,0.1701346487,0.4617418945,-0.0736793578,0.2461997569,-0.0704003125,0.0876337588,0.3339954913,-0.417286247,-0.2025863528,0.2991057038,-0.1963202059,0.0929352418,0.1767478585,0.6078708768,-0.1159728244,0.0043917857,0.0315837152,0.9024523497,0.0455393307,-0.0817784816,-0.1098654196,-0.1339853555,0.1832498163,-0.0273210369,-0.1489991993,0.0625030175,-0.1196947172,-0.1612454206,0.1652187258,0.1311919987,-0.0519263856,-0.1083442569,0.2160116136,-0.2077485323,-0.368796587,0.490970403,0.1842208803,-0.2601138949,-0.0338322781,-0.0949097723,0.0862209052,-0.0158458445,0.4588500559,0.1462963372,-0.0267404914,-0.0477765501,-0.2950968742,-0.4784201682,0.020577129,-0.2033832669,-0.0197258722,-0.0415498801,-0.4611433744,0.3367361724,-0.0387237333,0.3308035433,0.2357173413,-0.2417199314,0.4382307529,0.0296571814,0.0616570041,-0.1589969397,-0.0421616584,0.3224777877,-0.0999506861,-0.4679793715,-0.0917439908,0.0648094192,-0.1604277194,-0.3316780031,0.0301953312,-0.0140682869,0.3370954692,-0.3215791583,0.0008817861,0.1160256267,-0.3203574717,0.0510471314,0.1539743692,0.1602160037,-0.3130500615,0.4173514545,-0.0630557984,-0.2024502754,0.2692019939,0.088876754,0.1634570509,0.3131775856,0.4789273441,-0.374109,-0.1280457675,-0.0092916479,0.1501131952,-0.0383824743,-0.3002012074,0.0630461648,-0.1466517448,-0.1306985915,-0.0374566875,-0.1890237778,-0.0550261289,-0.2686322927,-0.4194504917,0.2642682195,0.156321615,-0.1615405679,-0.1581414193,-0.0588086322,-0.2340725511,0.2292050719,0.0600800551,-0.1987729669,-0.0952123031,0.0708448216,0.0004403669,0.0809509009,0.3135180473,0.0671560243,-0.231526345,-0.0141637158,-0.1911023408,-0.0547632463,-0.1556666493,-0.0770100728,0.0129507659,0.2241217643,-0.2560294271,0.3213593662,0.0987443104,0.2911232412,-0.1497278661,0.060009066,0.237149924,0.1311628819,0.021934174,-0.1233183518,-0.1723652482,0.3582632542,-0.0262937117,0.0191973373,0.0292989332,0.1904426813,-0.0600689836,0.4209103286,0.0199787784,-0.0522016212,-0.067400977,-0.10507337,0.3562832773,-0.2561872303,-0.0784981847,0.2487531602,0.0899206549,-0.0238220394,0.2156779021,-0.0067197098,0.2575020492,0.3767143786,0.1271656454,-0.3480820358,-0.2708782852,0.4393673241,0.0696188807,0.0912882909,0.0518657789,-0.0209962148,0.1900276542,0.5934480429,-0.2173599452,0.000598211,-0.2812071145,0.3986510634,0.1168006659,0.041448988,0.003843965,0.2109670639,-0.0958435237,-0.1941691786,0.2776306272,0.0010256139,0.1392679513,0.0157836434,0.2865011394,-0.04347625,-0.1826279759,-0.2778844535,0.0630052015,-0.1719837785,0.0254733693,0.1456583142,0.0085596433,0.1686031073,-0.2946828008,-0.1567503512,0.4429513216,-0.0348585956,-0.1660729945,-0.0161463376,-0.2395260483,-0.0327039361,0.1964886039,-0.0132301236,-0.2229390591,0.0400850475,-0.0554275215,-0.0373079404,-0.260814935,0.0578331277,-0.2298514098,-0.2040998638,0.0964579284,0.3305189908,0.0145139433,0.0386466086,0.0385616906,0.0743961707,-0.1065115407,0.1705517173,0.0026629199,-0.0675765052,-0.2130514979,-0.1034101695,-0.1013573259,0.3337801695,0.2396072,0.0420402847,0.2500343919,0.1211283803,-0.0448550396,0.214299202,0.14403373,-0.0385878198,-0.1112138554,-0.2471660823,-0.1967730522,0.5320686698,-0.3661890328,-0.402320385,-0.2896669209,0.3555041552,0.0956301242,-0.1350448579,-0.0017207933,-0.2280672491,0.055331897,0.1424283385,0.126975581,0.3029918373,0.197389394,0.1620062739,-0.2989750504,-0.4763529897,-0.1380309463,-0.2268099189,0.0338528827,0.0332463048,-0.0797083154,0.3156283498,0.2846124768,0.3745504618,0.085373871,-0.1024450436,0.0912109315,-0.1856529117,0.0226052068,0.2534652054,0.2765002847,-0.0069928472,0.0258967746,0.133643955,-0.1313361675,-0.0321137309,-0.1581094563,0.3250004053,-0.2064154744,-0.3413321078,0.1750477999,0.0974927545,0.4911982119,-0.1743947268,0.2077256143,0.025287943,-0.2233003527,-0.3599919677,0.4798054993,0.1770716161,0.2094188035,-0.06207636,-0.242306143,0.0154048577,0.365665555,-0.1404817551,0.2785497308,0.0435755961,0.0257527679,0.264467895,-0.2237758487,-0.1201341823,0.1116731092,0.0457647741,-0.2715981603,-0.0138406819,-0.1557691991,0.3085579276,-0.1301255375,-0.3394054174,-0.1737615317,-0.1860187203,0.3233477771,-0.2438561916,-0.1015706658,-0.0065151802,-0.0547778793,0.0343064815,0.0915483013,0.1175946742,-0.0001389792,-0.11577674,-0.3208641112,0.1700104773,0.4163022935,-0.0359049141,-0.0218567811,-0.0723691732]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"Hi @SapirWeissbuch,\r\nWhen you are saying it freezes, at that time it is unzipping the file from the zip file it downloaded. Since it's a very heavy file it'll take some time. It was taking ~11GB after unzipping when it started reading examples for me. Hope that helps!\r\n![Screenshot 2020-12-21 at 23 40 52](https:\/\/user-images.githubusercontent.com\/19718818\/102808355-3b380c00-43e6-11eb-81ab-c31019ae6322.png)\r\n","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":54,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nHi @SapirWeissbuch,\r\nWhen you are saying it freezes, at that time it is unzipping the file from the zip file it downloaded. Since it's a very heavy file it'll take some time. It was taking ~11GB after unzipping when it started reading examples for me. Hope that helps!\r\n![Screenshot 2020-12-21 at 23 40 52](https:\/\/user-images.githubusercontent.com\/19718818\/102808355-3b380c00-43e6-11eb-81ab-c31019ae6322.png)\r\n","embeddings":[-0.0419084467,0.172738567,-0.0584622957,0.448946923,0.0830925331,0.2269361615,0.2345205098,0.0880642906,0.1740850508,-0.000837622,-0.2182780355,0.2595328987,0.0018136572,0.0818539709,-0.0989711285,-0.0069779027,-0.0974632129,-0.1411154121,0.209971875,0.1532503814,-0.1402238458,0.5456249118,-0.2876877189,-0.3265475035,-0.1779135317,-0.056436412,0.0590684079,0.2066374421,-0.0397573076,-0.2740011215,0.5494995117,0.2960214317,0.0283417329,0.5103057623,-0.0001163505,-0.2048004419,0.3131906688,-0.0038417126,-0.4972884059,-0.2235812694,-0.3429577351,-0.0255894195,0.057130184,0.0733717233,0.0538625382,0.4552014768,-0.1131191179,-0.6822659969,0.0399417654,0.0720884874,0.1493647844,-0.0646007583,-0.199761346,-0.1022181883,0.1219624579,-0.1928976029,-0.2828271389,0.4720593989,0.2885800898,0.0657839403,0.2345456928,0.2593853474,0.0046880711,0.3377595544,0.0988686234,0.0581915826,-0.325982064,-0.3113313019,0.4224762917,0.2790928483,0.8703471422,-0.2963435948,-0.175425455,-0.0423670523,0.0715548322,-0.1038353592,0.2577138245,0.088854,-0.4103254676,-0.0335079283,-0.1232032478,-0.1644373834,-0.0424618758,0.0381719917,-0.1927553713,-0.2027969658,-0.1956015676,-0.0244732946,0.121129334,0.0154822152,0.3752960861,-0.266582936,-0.2203110307,0.1416900307,-0.6825615764,0.0889082029,-0.1432446986,0.1520679891,0.1661089212,0.1773971468,0.2389192879,0.250829041,-0.2416718155,-0.0300459806,0.2960170209,0.4848723114,-0.0671037883,-0.0906197503,0.3128862679,0.2891587019,-0.116367057,0.0646981373,-0.11278864,-0.0540562049,-0.0088184932,-0.1395270377,0.1751488596,-0.3751081228,-0.3067250848,0.0658895671,0.0155868111,0.1677224189,0.0349553004,0.3543732464,-0.3763412237,0.2156981528,-0.1120009422,-0.045582965,-0.1661397219,-0.041979067,-0.086148724,-0.1404378414,0.0543522239,0.1692944914,0.411878109,-0.1329503953,0.0746368617,-0.2131661922,0.172381416,-0.0418972,0.0963757411,-0.456872046,-0.0322256424,0.19248043,0.0370418951,0.4138689041,-0.2077854425,-0.1593095809,-0.1645265073,0.2517725527,-0.2718759179,-0.1436472535,-0.007459823,0.0797771662,-0.4005433023,-0.0722330213,-0.168295905,0.123660475,0.1961122453,-0.0432695486,-0.0236979295,0.0501558706,-0.33690238,-0.2600096762,0.245839864,0.6521677971,-0.4494019747,0.3058450818,-0.3420861661,0.1299992949,0.0804680139,0.2561940253,-0.0010044767,0.0337726139,-0.4061591327,0.0149713336,0.3212281764,-0.3201260567,-0.7326216698,0.3826341033,-0.0256310049,0.251427412,0.0303474572,0.2256506681,0.1226950809,-0.0749238282,0.1661731154,0.4375931621,-0.1711920202,-0.031693168,-0.4060000777,-0.27152583,0.1005339772,0.1699432135,-0.1659566909,0.0679287687,-0.036284972,0.1187874749,0.1556513757,0.3864189982,0.0588685088,0.3071848452,0.2669268847,0.00021677,0.0709671304,-0.067124255,-0.8073760271,0.4343363047,-0.0653368309,-0.1522076726,-0.19800587,-0.0142491395,-0.4346853793,-0.1675317585,-0.2741286755,-0.2488716394,-0.0109109906,0.0987044871,0.3192574084,0.2353910059,0.1014146581,0.648337841,0.3068910837,-0.0480939932,-0.4378825426,0.0892025158,0.0519032031,-0.1164015979,-0.0646075159,-0.0341096781,0.0176020544,-0.2008464038,-0.21785447,0.2084423602,0.1733133793,0.1311374009,-0.0172833204,0.0190534666,0.2716843486,0.0193918105,0.286211133,0.0508091412,0.3185540438,-0.2622424662,-0.1959272921,-0.0148142194,-0.4580947161,0.2177153826,-0.1105609089,-0.0237456206,0.2106559426,0.0273159984,0.2022928298,-0.0464965068,0.296946913,0.1381292492,0.3851886988,-0.1820930392,0.0159382094,0.3154764771,0.2811936736,0.1673019528,-0.0332994573,0.1528603882,-0.1099497899,-0.0577406734,0.1584849656,0.4288981557,0.6317715049,0.0146721471,0.1970623732,-0.0221591387,0.2154348344,-0.2755388916,0.0512860119,0.082552731,-0.1301777363,0.1544675082,0.1960978359,-0.0384207219,0.1795618236,-0.1035709754,0.1096571609,0.1532823741,-0.0308224317,0.2952177227,-0.2102186233,-0.191859901,0.0659354851,-0.0963198915,-0.0047730547,-0.4441174865,-0.1619956195,0.3785782158,0.2477114648,-0.0651289448,-0.0693964809,0.2523093522,0.1083593071,-0.5406515002,-0.3403006196,-0.2089052647,-0.2180715799,0.0151600912,0.271680057,0.0632411689,0.2063985914,-0.0864794552,-0.0475741588,-0.181358844,-0.1395511925,-0.0201594122,-0.0543301664,0.4290530384,0.0674380362,0.3619537354,-0.1399614513,0.2467644066,0.0067677926,0.0304791257,-0.0865390971,0.0126491021,0.0340264253,0.1970240325,-0.1144118533,-0.3764761388,-0.3245883286,-0.4204921424,0.2019141614,-0.0184990447,0.003930144,0.3973693252,0.0829567239,0.0703602433,0.2658996582,-0.1668619066,-0.3439247608,-0.3762113452,0.6043846011,-0.1868122518,-0.1460451782,0.0393626504,-0.0498542413,0.3061037064,0.1134551391,-0.5427657366,0.0421469919,-0.5677765608,0.5096668005,-0.1733064055,-0.1721161604,0.2683787644,-0.221430406,0.0925935358,-0.0034911465,-0.1249541789,0.0168372486,0.2290721536,0.2733463645,0.1916391999,0.1165311038,0.1638864726,0.5300636292,0.1157094315,0.1098057777,0.3600562811,0.2078381777,0.4153865576,-0.2022200376,-0.0981784761,-0.024988126,-0.1080678627,-0.1985759437,0.0585559569,0.0758882686,0.113443248,-0.1827463359,0.2020843625,0.1277654022,-0.1485204399,0.0409211256,-0.4448772967,-0.0398676395,0.175667122,0.1237269118,0.1660163999,-0.4054832757,0.1143273786,0.1749941111,-0.1603663415,-0.0097187385,0.3262452185,0.2560048103,-0.5234526396,0.2162799686,0.0181004722,0.2151677012,-0.29408288,0.1594325006,0.1931457072,-0.033380881,0.6104550362,-0.073506549,0.1198480502,0.0844236463,-0.0722127184,-0.2800079286,0.0625383258,-0.0327555463,0.2215899825,0.094677873,0.084027335,-0.3307253122,0.2568326294,-0.0967839956,0.3848520219,-0.128627032,-0.2640788257,-0.2131868154,-0.3541562855,-0.1759675294,-0.1663416475,0.0562220439,0.1060224101,-0.0501417965,-0.1664782465,-0.0232503247,-0.1409790516,-0.0645930395,0.2014039606,0.1898768097,0.0323210284,0.5507431626,0.1635599732,-0.5343498588,0.0028341657,0.5873194337,0.0728155375,-0.2510009408,0.0864331648,-0.0227485262,-0.3373125494,0.3260510564,-0.2240002602,-0.0134175085,-0.1696369052,0.0024737238,0.0549915098,0.174631387,-0.0746801198,-0.0657136887,-0.151986286,-0.4081593752,0.4954634905,0.0350018293,-0.1528702229,0.5628780723,-0.3258577287,-0.1096098348,0.1997948587,-0.17324543,0.8595941663,-0.1777355671,0.0855187997,0.1173197404,-0.2510174513,0.4928015471,-0.2547518909,0.0500189997,-0.2874680161,-0.3076962531,0.0784626156,-0.1739573926,0.1624069959,-0.062266387,-0.346033603,0.2888248563,-0.3253986239,0.0183193944,-0.0400674306,0.115846239,-0.2250914872,-0.1719713211,-0.2207202166,0.1739456058,-0.2580852509,0.3437282741,-0.177933991,0.0057819127,-0.325363189,0.1621319801,-0.377212584,0.1598799229,-0.4309802055,0.0780469626,-0.1338422,-0.4891131222,0.0282179061,0.2355607897,0.230025351,0.091631867,-0.294498682,0.4285390377,0.0786144733,-0.1168581098,0.1433288902,-0.0975631177,-0.1673152447,-0.1013596505,0.0093222363,0.0414463282,-0.1491098404,-0.0571762361,0.0773034245,0.1830576807,0.1978448629,0.077308856,-0.3150310814,-0.1432825476,0.1029699743,-0.0597883202,0.0848004296,0.3302456737,0.0688858852,0.2345088571,-0.080310747,-0.1395228505,-0.1402884126,0.5465254188,-0.0952441469,-0.0586933605,0.4514365196,0.1515827477,-0.2001711577,-0.0938321352,-0.0367093794,-0.6449838281,-0.4405821562,0.2830787599,-0.1435875595,0.0039357189,-0.1014511734,0.215230912,0.0586709157,-0.2457254827,0.0370211788,-0.541117847,-0.1821194291,-0.3006326258,-0.2627495527,0.2188415825,-0.1836178601,-0.3439795673,0.0516597331,0.2093585581,-0.1738290489,0.1009985507,-0.2741859555,-0.1593555808,0.1626309603,0.104561612,0.3247634172,0.0082544973,-0.0520348214,0.2829304934,-0.0652701706,-0.191411227,-0.0854738131,0.1704089493,-0.0347761363,-0.2156842202,-0.0541878529,-0.2251030505,-0.0897261724,0.0570358261,0.3465260267,0.27925542,-0.0226034503,-0.132548064,-0.0539561473,0.2009240091,-0.4248509407,0.5406425595,-0.1742587984,0.3315263987,0.2469300777,0.0721240044,0.0538678393,0.0042933286,-0.2793897092,-0.1636240631,0.0419360325,-0.2720910013,0.2115260661,-0.2299984396,0.1230467558,0.0393713266,-0.1588976532,0.4445647597,-0.1811893582,-0.4000502527,0.1709839255,0.1026598588,0.0542881303,0.0240070485,-0.2044904828,0.1132172942,-0.1070365235,0.2897759974,0.1330543309,-0.0436142869,0.1880614012,0.2820530534,0.6532971263,-0.0395050123,0.1079608053,0.025754571,-0.0673298687,0.1677751839,0.0750852302,0.0890249759,-0.1121623665,0.4364631772,-0.15692994,0.1402921528,-0.0204460025,0.2242277265,0.011509845,-0.4680805504,0.0595380925,0.2120230794,-0.2010051906,0.132539168,0.0797109604,0.1168012246,-0.2172458917,0.0740720332,-0.1964280456,-0.1980274022,0.0284438729,-0.008488344,0.4307757914,-0.1018736437,-0.0949766561,0.165265426,-0.0193188936,-0.2035589814,0.4076844156,0.2153974473,-0.2317448258,-0.3878201246,0.0999843404,0.1282314807,0.0356578045,-0.0697673336,0.4195338786,-0.0426922366,-0.0523490235,0.0882742256,0.1833008528,0.1563060731,0.3036764562,-0.2038546205,-0.0398274846,0.0033325318,-0.0403041467,0.0935934931,0.4269134104,0.0705635175,-0.0670418665,0.2274570912,0.0696365237,-0.1309253424,0.0751293153,0.238056913,0.1169582084,-0.3764619231,0.2378476262,-0.1862945706,0.119346045,-0.1318241954,0.012940282,-0.2357870489,-0.0373216756,-0.0426656678,-0.1476791799,-0.0136869652,0.0601971298,0.0524177402,-0.0117672607,0.6775050759,0.2256251723,0.2454310358,-0.3152849972,-0.1166141331,-0.6835512519,0.1176371053,0.024574168,-0.0928709209,-0.1354145706,0.3266229331,-0.200543493,0.3659254909,-0.1487881839,0.3139801025,0.2909752131,0.1791518927,-0.49015975,-0.3483687639,-0.0771437883,-0.1987233013,-0.0226224959,-0.5086995363,-0.0547381863,0.0389294848,-0.0785040259,-0.041259665,-0.4627976716,0.1432774961,0.3093457818,0.4588360786,0.0028567654,0.2817789614,0.2478705645,-0.2167214304,-0.1851252913,0.0283570383,0.1950254887,-0.1751901656,0.1846858859,-0.0769165233,-0.1916371882,0.0418164916,-0.28418383,0.5085655451,0.1008726582,0.4262541533,-0.0078399964,-0.137731567,0.0082964813,0.3841107488,0.0598078407,0.3055108488,-0.1552421302,0.2351806313,-0.3621495366,-0.1250896603,0.3949168026,-0.0471910909,-0.3295769393,0.2524252236,0.2086874545,-0.1783497483,-0.3466308713,-0.4967410564,0.2499596328,0.2593311071,-0.2156809121,-0.0225337334,0.2097008675,-0.1952105761,0.1252015382,-0.1547338814,0.4324015677,0.1071854904,0.0767815039,-0.0055144327,-0.0240292195]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"Hi @bhavitvyamalik \r\nThanks for the reply!\r\nActually I let it run for 30 minutes before I killed the process. In this time, 30GB were extracted (much more than 11GB), I checked the size of the destination directory.\r\n\r\nWhat version of Datasets are you using?\r\n","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":44,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nHi @bhavitvyamalik \r\nThanks for the reply!\r\nActually I let it run for 30 minutes before I killed the process. In this time, 30GB were extracted (much more than 11GB), I checked the size of the destination directory.\r\n\r\nWhat version of Datasets are you using?\r\n","embeddings":[-0.072755143,0.1626463085,-0.0521008894,0.4159293771,0.0719526932,0.1767252088,0.1985662133,0.0667587072,0.1333532035,-0.0039923522,-0.2109427154,0.2015078813,-0.0472181328,0.0925029591,-0.1151683033,0.0005983744,-0.1870140433,-0.0596627817,0.1579323113,0.1352991462,-0.1677170992,0.439342469,-0.2720969319,-0.3291651607,-0.2195433378,0.0065108202,0.0340620987,0.1153706238,-0.0439226031,-0.2515452802,0.5903285146,0.1817964613,0.2022955716,0.5596300364,-0.00011643,-0.2690275013,0.3349490762,-0.0813676193,-0.5700404644,-0.2561136186,-0.4124467671,0.0133202085,0.0350640342,0.0512670875,0.0637827441,0.4205184281,-0.1521245837,-0.64205724,0.0961365029,0.184820652,0.2039233148,-0.1425318271,-0.1532195061,-0.105651319,0.1025357097,-0.0433299318,-0.2304594815,0.4613931775,0.1876121461,0.0937481448,0.2468427271,0.3056149483,-0.0021749551,0.3733741343,0.1243720353,0.0224541258,-0.2359964401,-0.3585821986,0.4380189478,0.2638860345,0.8583325744,-0.2614651024,-0.2454858124,-0.1408300549,0.0629529282,-0.1338534206,0.2592305541,0.2633804381,-0.3231261969,-0.0073956484,-0.1510874331,-0.1000832841,-0.0636823475,-0.0040893252,-0.2076932192,-0.2849427164,-0.1990710497,0.0020443622,0.0280593559,-0.00062262,0.3906316459,-0.1917218268,-0.1724092066,0.3107558787,-0.5941225886,0.1101470888,-0.0551612638,0.1852333695,0.2100424469,0.173067838,0.160839662,0.08328861,-0.3457117677,-0.0585267134,0.2925242186,0.5779297948,-0.0692556798,-0.0797537342,0.1991256326,0.2562425733,-0.0700737685,0.0698362142,-0.0884934738,-0.0246249195,0.0572537109,-0.1104035899,0.2593982816,-0.3535263538,-0.3831744194,0.0682486594,0.0211931132,0.2233143002,-0.0483936593,0.2390223891,-0.2393819839,0.2680348158,-0.0848907307,0.060208302,-0.1209563091,-0.1047899053,-0.1099950448,-0.0689461753,0.0120426118,0.1089615747,0.4129388332,-0.1468844563,0.0874883607,-0.1852960438,0.1007984355,-0.0752597973,0.0441004112,-0.4850211143,-0.0442705639,0.285319984,0.1247140467,0.3971720636,-0.2388545573,-0.138188079,-0.1328911185,0.3041514158,-0.3151180744,-0.2827414274,-0.0892390534,0.1152573451,-0.4591413438,-0.0980610847,-0.3268722594,0.1572990417,0.2685288191,0.1059336066,-0.0093497122,-0.0360065624,-0.3958993554,-0.2581654191,0.2020711601,0.5899381042,-0.4222159386,0.2609266043,-0.3011381328,0.1883919388,0.1586330533,0.2318543792,0.0258223563,0.1076005399,-0.4511807561,0.0197257139,0.2745524943,-0.3494895101,-0.7386118174,0.3036574721,0.021286672,0.2213606387,0.0366427563,0.0763042793,0.1528085172,-0.0280256998,0.16202043,0.4609229863,-0.1844620705,-0.0038157732,-0.385767132,-0.3302130401,0.0818016604,0.1815673113,-0.0305189472,0.0676134229,0.0258463323,0.0914006531,0.2147233635,0.3795999885,0.1587375998,0.2579649389,0.3154444396,-0.0450445525,0.0312176943,-0.0865430459,-0.7296493053,0.4170327187,-0.1016145498,-0.1146012545,-0.1999589801,-0.0041276789,-0.5134187937,-0.0837907493,-0.2601849139,-0.2858148217,-0.0128261345,0.0724398345,0.3398637474,0.1317500025,0.0976457074,0.6660851836,0.2577014863,-0.12178348,-0.4322471023,0.0490944795,0.0339934193,-0.102925472,-0.0327064544,0.010017273,0.0939169005,-0.1624750495,-0.2578213811,0.1050464064,0.1943413317,0.0221010577,0.0371774621,0.0491048954,0.2568796873,0.0424103886,0.2400223911,0.0523945093,0.2376604378,-0.1321962029,-0.1441477388,-0.0882766098,-0.3817676008,0.2373189926,-0.2045960426,0.0101193106,0.2657052577,-0.0246239956,0.2844458818,-0.0999696851,0.2773302495,0.1908300519,0.2834511101,-0.1209324971,0.1273938566,0.2658622563,0.1088988706,0.1941094548,-0.0323126167,0.2162722945,-0.1747235805,-0.1451085657,0.1201923266,0.4297387004,0.6090689898,0.0016629801,0.270540148,-0.0489554182,0.3418198228,-0.1985937655,0.0410673879,0.0859616771,-0.1470722109,0.187978819,0.2925009727,0.0380579941,0.197683081,-0.0623431243,0.1066617742,0.2068859041,-0.0509885959,0.1892718971,-0.1957104355,-0.1634073853,-0.0388464779,-0.1722293645,0.0784078911,-0.3618616462,-0.1269526333,0.3402609229,0.2416095138,-0.0549681224,-0.1024231687,0.1973516792,0.1123743206,-0.478464663,-0.2655420601,-0.1195397303,-0.2441501468,0.044734966,0.1455067992,-0.0701164529,0.2637217641,-0.0968733057,-0.034589503,-0.213289842,-0.1277239025,-0.0152927311,0.0476173051,0.3991951942,0.1074263006,0.2642818093,-0.0586503781,0.2124877572,-0.0502059944,-0.0096272482,-0.0813302696,0.0515185595,-0.0399384424,0.2434220463,-0.1355899274,-0.326969862,-0.3629831374,-0.3707721531,0.1907312572,0.1076776683,0.0279355422,0.4600449502,0.0399975292,-0.0275215711,0.3603964746,-0.1745766401,-0.4181405008,-0.3557340503,0.6125190854,-0.1771862209,-0.1476271003,0.1279090047,-0.0705701038,0.3535722792,0.1714600772,-0.5831279159,0.0055774353,-0.4850946069,0.5077188015,-0.1956724972,-0.1334520131,0.1899163276,-0.3022510409,0.1054920405,0.076634936,-0.1188502386,-0.0458650179,0.2866661251,0.3172294199,0.1094697714,0.1482848525,0.1944676489,0.4785459638,0.215719983,0.1037300304,0.3804444671,0.179491818,0.3754736483,-0.3020776808,-0.08651568,0.0627321377,-0.1200739145,-0.2153888047,0.1527853012,0.1057171971,0.2084663361,-0.1314737797,0.0989515483,0.0226343852,-0.1160672531,0.0014186777,-0.4102527499,0.050651148,0.2017982304,0.0600185208,0.1254478693,-0.401425302,0.0162632503,0.0974212885,-0.1297282726,0.1409917474,0.369335413,0.3272194266,-0.5318821669,0.2155187428,-0.0294578094,0.2194763124,-0.196953088,0.1962719709,0.2081558704,0.1019167826,0.5212138891,-0.0897870436,0.2266061753,-0.01907208,0.0000495937,-0.3613800704,-0.0660124943,0.0565532558,0.2414767742,0.0898291767,0.0517472439,-0.2726706266,0.3449422419,-0.0023381829,0.3766367733,-0.2011573464,-0.2737296224,-0.2575483024,-0.4044120014,-0.2176550776,-0.1345187575,0.1277777702,0.2258788943,-0.0346930251,-0.1814047545,-0.0072682798,-0.1588960588,-0.1298326552,0.0914265513,0.2349980325,0.1482093632,0.5017930269,0.1631287038,-0.470064491,0.0298714936,0.6188191772,0.0043137502,-0.2765968442,0.0663196892,-0.0915620103,-0.2987781763,0.2302347124,-0.2355352044,-0.0846044049,-0.1964434683,-0.0188092235,-0.0477936901,0.17243734,-0.1896478385,-0.1684861034,-0.1341042668,-0.4031732976,0.4983971417,-0.0103195701,-0.132339403,0.4759447575,-0.315389514,-0.1550495923,0.1900364459,-0.1825044602,0.8272014856,-0.1437197775,-0.0429147743,0.1082746387,-0.305735141,0.4609217644,-0.1710369438,-0.0367403589,-0.2875880599,-0.2151854336,0.1003731266,-0.1226035506,-0.0063100811,-0.0448034145,-0.2065382749,0.3282616138,-0.3333375454,0.0914622694,0.0447994173,0.1773787886,-0.1810258627,-0.1805230677,-0.1682491004,0.1849960983,-0.2625401318,0.3540967107,-0.2125169188,0.069457598,-0.255556345,0.0848933831,-0.440630883,0.1657100022,-0.4408501685,0.1346239746,-0.1146128029,-0.5683187842,-0.0345532373,0.2617666423,0.2147224993,0.0392890722,-0.3168460131,0.3903346062,0.0329713076,-0.0617152676,0.1376775801,-0.077906087,-0.2076827437,-0.0793446749,-0.0017444626,0.0331547782,-0.141885668,-0.0404806621,0.0148302913,0.1529010981,0.2462220043,0.0414788984,-0.2445770353,-0.225831151,0.2302803248,-0.0697725937,0.0885313153,0.2931302786,0.0569550991,0.2398427725,-0.1555963457,-0.2222203314,-0.1356586963,0.4944388568,-0.0258462951,-0.0093853502,0.4466128647,0.1005354673,-0.1745759994,-0.0451196022,-0.088202104,-0.6157084703,-0.4520606995,0.2825577855,-0.1205759346,0.1409267187,-0.0820494667,0.243770957,0.1357149333,-0.2271658927,0.0787746087,-0.5944200754,-0.221812591,-0.2898914218,-0.3759373724,0.1205818281,-0.1268992275,-0.3020056188,0.1421276331,0.2786503732,-0.1813616455,0.0741475001,-0.1590571702,-0.0898460075,0.2024529576,0.0792619884,0.2377504259,0.0176319629,-0.0414703041,0.2652091384,-0.0846115947,-0.1973303854,-0.1915541142,0.1515131295,0.0887515321,-0.1477024257,-0.0745694935,-0.311770469,-0.1129619628,-0.0278399792,0.3011051118,0.361538589,0.0165561344,-0.0479446463,-0.0378436148,0.1085291952,-0.4083346128,0.5186579823,-0.1535273194,0.3087811768,0.1834799498,0.1362285316,0.0959779546,0.0195671488,-0.3008526862,-0.237510398,-0.0483946577,-0.1842811406,0.0666152313,-0.2637198865,0.2281456888,0.0626926497,-0.142483905,0.3330272436,-0.2362486571,-0.3347247541,0.0777773783,0.113819465,-0.0388804227,0.0164992716,-0.2113914639,0.1722695231,-0.1770913005,0.2823286057,0.2327589691,-0.1868277937,0.2290091366,0.2101494521,0.6139941216,-0.0540427789,0.0400846936,0.0927275494,-0.1185182557,0.1473406702,0.1073774844,0.0372937135,-0.0559446625,0.5519288778,-0.2281755954,0.1509530395,0.0169233289,0.0981225669,0.0009083137,-0.4996128976,-0.0433160253,0.1512283683,-0.1827793717,0.1465857774,0.0932254195,0.1557900757,-0.1304771453,0.1328514516,-0.2237262279,-0.2250801921,0.0161184054,0.0076746019,0.3835389614,-0.1032451391,-0.0754264444,0.1986467242,-0.0239694435,-0.213213861,0.3885928988,0.1910746843,-0.3590888083,-0.4073275328,0.1236299351,0.2064030766,-0.0704110041,-0.1513513625,0.4083332419,-0.0228311624,-0.0186206847,0.1225939393,0.199271962,0.25999856,0.2954671383,-0.2308009118,0.0117439143,0.0455792025,0.0164119359,-0.02973965,0.4536008537,0.0333793126,-0.0418842882,0.1907137632,0.0988925472,-0.1316246539,0.0989831313,0.302183181,0.1275603324,-0.2749022841,0.2052880824,-0.1854934394,0.1603644043,-0.1391321272,0.1036335975,-0.1344947219,0.0123619111,-0.018133672,-0.1250477731,0.0103912465,0.0776279494,0.0477039069,-0.1266246587,0.5700390935,0.2417907417,0.2312133759,-0.3278712928,-0.1018165946,-0.6946564317,0.1600179672,-0.0046687517,-0.1457926184,-0.1520231515,0.3000464737,-0.2754313052,0.2337021828,-0.1188789085,0.3639770746,0.2919640541,0.2082565576,-0.4241272211,-0.3640205264,-0.0461475737,-0.1400150955,0.0453945436,-0.5439508557,-0.0457185544,0.0318986923,-0.0767632574,-0.0202268157,-0.4368483424,0.2137728482,0.3540940881,0.5459756851,0.0308637563,0.2489111871,0.189694196,-0.095907554,-0.2870941758,0.0376062319,0.0817401782,-0.0968897715,0.1043605506,-0.0295551773,-0.201837942,0.1148269102,-0.2297531366,0.4332323074,0.1188292652,0.3623374104,-0.1103126258,-0.0917160586,-0.0786883533,0.4053358734,0.0381542221,0.2004565746,-0.2115886658,0.2952994406,-0.2809065878,-0.2454272211,0.3921199441,-0.0651435331,-0.3290114701,0.2062491924,0.2611353993,-0.1149535477,-0.3742113113,-0.5145344138,0.2866063416,0.3664518297,-0.1729808301,-0.0789417326,0.2095696032,-0.2188344002,0.1235141307,-0.1210314184,0.4332136214,0.0994341224,0.039684169,0.0183965843,-0.0409614444]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"I'm using datasets version: 1.1.3. I think you should drop `cache_dir` and use only\r\n`dataset = datasets.load_dataset(\"trivia_qa\", \"rc\")`\r\n\r\nTried that on colab and it's working there too\r\n![image](https:\/\/user-images.githubusercontent.com\/19718818\/102814269-4db74300-43f0-11eb-8f26-ecfcf4632002.png)\r\n","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":28,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nI'm using datasets version: 1.1.3. I think you should drop `cache_dir` and use only\r\n`dataset = datasets.load_dataset(\"trivia_qa\", \"rc\")`\r\n\r\nTried that on colab and it's working there too\r\n![image](https:\/\/user-images.githubusercontent.com\/19718818\/102814269-4db74300-43f0-11eb-8f26-ecfcf4632002.png)\r\n","embeddings":[-0.0324455127,0.2749699056,-0.0218249429,0.3239808977,0.1333954036,0.1968196183,0.2799459994,0.0583024845,0.0948531032,-0.059717834,-0.2450279444,0.2710150778,-0.0008884075,0.1060801074,-0.1267104149,0.0433528572,-0.114684917,-0.0377161428,0.2132201493,0.1566438973,-0.1349406689,0.4377724528,-0.2820023,-0.3223251998,-0.2458768636,0.0691145957,0.0668900087,0.1003829762,0.0343146771,-0.2480177581,0.67411834,0.3076746166,0.149097681,0.5383524895,-0.0001176828,-0.1858210415,0.3122101426,-0.0681562498,-0.625824213,-0.2935509086,-0.4505560398,0.0806286782,0.0823827982,0.0722627118,-0.0465550199,0.4582152367,-0.0952111408,-0.6741397977,0.0192471445,0.1569029987,0.1905343533,-0.1172520518,-0.2548100948,-0.0953706503,0.1114436388,-0.0803015232,-0.27206406,0.5007556677,0.2071802467,0.0817496479,0.2020538896,0.3206925094,-0.0337509513,0.39436692,0.1626695246,0.0672909617,-0.2473103404,-0.307952553,0.4526152015,0.2340932935,0.9158679843,-0.3237604201,-0.2919473052,-0.1021097675,0.0687192604,-0.1473658085,0.2955060601,0.1430480182,-0.2731516361,-0.0047070812,-0.1451757252,-0.1480061263,-0.0216660183,-0.0137277143,-0.1773999482,-0.2071225792,-0.1470057219,-0.004651607,0.041708041,0.0171908811,0.510862112,-0.2119361758,-0.1754100323,0.248605594,-0.623901844,0.1116386428,-0.0553300567,0.2207639962,0.1355824471,0.1886218488,0.1547271311,0.1280135959,-0.3434439898,-0.0555103943,0.2494443655,0.5355216861,-0.0092455531,-0.0484632105,0.2394342124,0.2919661403,-0.0908755213,0.0781818926,-0.074564226,-0.0653682649,0.0330923907,-0.0501569174,0.2361414284,-0.3740792274,-0.3454495072,-0.0093842046,-0.0154698594,0.2531176805,0.0263857655,0.2701820433,-0.2840004861,0.2326665819,-0.0908644348,0.0841919929,-0.1722839177,-0.0141151398,-0.0864259601,-0.1260414422,0.0389911756,0.1875640303,0.4296388626,-0.1550474912,0.1072456837,-0.1846923828,0.1514116973,-0.0643315837,0.0191757977,-0.4474502802,-0.1141708121,0.2350038588,0.0871080831,0.4529133737,-0.1869063079,-0.1543507725,-0.0992894471,0.2812709212,-0.3120135665,-0.2740344405,-0.0483691655,0.0841600969,-0.4793226719,-0.1381507814,-0.3207099736,0.1803421527,0.2674741447,0.0174977332,-0.0052148467,-0.0017862897,-0.4459944665,-0.2575539649,0.2131515294,0.6501175165,-0.4160538614,0.192986995,-0.3035592139,0.186270237,0.1453388035,0.1869918108,-0.0152987568,0.0502583906,-0.3998034596,-0.0738630965,0.2610940635,-0.3984422088,-0.7728948593,0.2378232479,0.0084293252,0.188441962,0.0210103337,0.1339071393,0.0947505683,-0.0753481686,0.1275687516,0.455789268,-0.1201919094,0.0158131346,-0.3521460891,-0.3146685362,0.0798838809,0.2370212227,-0.0563240834,0.1390461177,-0.0251109377,0.0512885526,0.1590341032,0.3394579291,0.1158736944,0.2126523256,0.3286348581,-0.0262288544,0.0397684798,-0.0002226138,-0.8115261197,0.4374570847,-0.0522571355,-0.1002551541,-0.2450171113,-0.040324863,-0.467741698,-0.1434410214,-0.2980220318,-0.2695590854,-0.0153903868,0.1407362819,0.3109505177,0.1854406297,0.087342687,0.6520835161,0.307155937,-0.1022906676,-0.3502542973,0.0763698146,0.0567293838,-0.0485360622,-0.027627591,0.0433520973,0.1094926298,-0.1354460567,-0.2275478542,0.1687919796,0.1888411194,0.0848039016,0.0694317594,0.0469981581,0.2705690265,0.0723999366,0.179262653,0.0418834873,0.248191461,-0.1535512954,-0.1856684387,-0.0135972165,-0.3767162859,0.2876800001,-0.2158856541,0.0418824516,0.2819383144,0.0126301255,0.2349714637,-0.0830989033,0.2622601688,0.2052707374,0.3273557127,-0.1094087288,0.0862701535,0.3412249684,0.1120816469,0.2141619921,-0.0404642113,0.1691351235,-0.1454675496,-0.139097482,0.1596773714,0.4336571097,0.6067399979,-0.0081361635,0.2251293361,-0.0391937681,0.3126021028,-0.2367356867,0.0741251335,0.0563662015,-0.1426462978,0.1152948439,0.3039670885,-0.0187689774,0.2213625312,-0.0872809589,0.1258670688,0.1848024279,-0.0884196013,0.2492524534,-0.2403630763,-0.1852182299,-0.0158719402,-0.2151351273,0.0807840154,-0.3866450489,-0.1256521344,0.3427274227,0.2241571248,0.0063512516,-0.106487155,0.2327479124,0.080698289,-0.6001936197,-0.316981703,-0.2200912535,-0.2548022866,0.0200473219,0.1552534848,-0.0421639681,0.2343332171,-0.1184985191,-0.0251719654,-0.2593436241,-0.1221707091,-0.0266620032,-0.0059835552,0.3980239928,0.0872639567,0.206764549,-0.1229046881,0.1910785586,-0.0037700424,0.0013773274,-0.0179857742,-0.0032086673,-0.0109097874,0.2183505148,-0.1473802626,-0.3007587194,-0.3201138079,-0.3525672257,0.147317782,0.0582233183,0.04454704,0.5137864947,0.0451366603,-0.015032338,0.3702687621,-0.203423664,-0.3940540552,-0.374451369,0.5996718407,-0.184115991,-0.1274546832,0.1135366485,-0.0431589969,0.3891008198,0.1707431227,-0.5868318677,-0.0655888245,-0.4273253679,0.5127469301,-0.1010312513,-0.1719828397,0.1673006564,-0.2365752459,0.1099241376,0.0389212146,-0.1907941401,-0.0395424888,0.253788799,0.303168118,0.154148221,0.1072200984,0.1426324397,0.506226182,0.1756164283,0.0887815803,0.3587370217,0.1716665179,0.4075106084,-0.2761300504,-0.1543662697,0.0304713529,-0.0825059116,-0.2064793855,0.1651031375,0.1080181673,0.1601503044,-0.1984343231,0.1768435985,0.0752851292,-0.1809390038,-0.0501997024,-0.4193104804,0.06138945,0.1943118572,0.1231007949,0.1041473299,-0.4013205171,0.0772134513,0.1993872076,-0.1166353002,0.098635681,0.3743680716,0.2576519251,-0.5474616885,0.2236439884,-0.0078034978,0.2726119161,-0.1983117461,0.2351412177,0.1919603199,0.0969454199,0.5006756783,-0.0991206393,0.2050292492,0.0334523395,-0.0060104434,-0.3170508146,-0.0772121325,0.0845157206,0.3015796542,0.0356599763,0.0596462265,-0.261874795,0.2602558136,-0.0659120455,0.2567504942,-0.1806650758,-0.2428630292,-0.2566303611,-0.4131553471,-0.182514742,-0.1818362474,0.0476025045,0.1722752452,-0.0373496041,-0.1239414662,0.010648015,-0.161299631,-0.1508211344,0.1064650714,0.2024849504,0.2022666633,0.4871498048,0.1861310899,-0.5531713963,0.0806216896,0.6059457064,-0.0138776405,-0.2692315876,0.0502422117,-0.0712924972,-0.3364036679,0.2518009245,-0.2729458511,-0.0500291139,-0.2049168199,-0.1072276905,0.0157931242,0.1768847704,-0.2031434476,-0.1740594953,-0.1204413921,-0.4807562232,0.5313028693,0.013543047,-0.1668869108,0.4396960437,-0.3095037341,-0.1804394275,0.1788161099,-0.1127643958,0.7681950927,-0.1555906683,-0.0050704316,0.1729530096,-0.2942380607,0.508491993,-0.1210597828,0.0329205059,-0.264914006,-0.227643773,0.0819310397,-0.1351887882,0.0032549053,-0.0637610182,-0.2479319721,0.3533605337,-0.3189290166,0.0806481391,0.0407207087,0.0440123715,-0.171577543,-0.2244061679,-0.2230705768,0.1778987348,-0.3040218055,0.365304023,-0.1601791978,0.0671746731,-0.3220827281,0.1346186399,-0.4186404943,0.169611603,-0.4425396919,0.0899516717,-0.1080944836,-0.5266512036,0.0117193386,0.2386697084,0.2837730348,0.0198051203,-0.3421069682,0.3862946332,0.0622824915,0.0004882537,0.1050111204,-0.1384168565,-0.1765523255,-0.0272213649,0.0477185659,-0.0033274279,-0.1261083335,-0.0894590914,-0.0017527336,0.1335174292,0.1867548376,0.1096469909,-0.1947399229,-0.1613244414,0.231558457,-0.0728066564,0.0829450339,0.2448138148,0.1242415681,0.1920788288,-0.1916362941,-0.2255367637,-0.0904904678,0.5441234112,-0.0244290885,0.0056946822,0.4305721521,0.092589356,-0.1747611463,-0.0247634668,-0.1550227553,-0.6290492415,-0.4330122471,0.2905843556,-0.1344904602,0.1045579761,-0.0686680973,0.2606515884,0.0709484145,-0.1818849742,0.1006852761,-0.5233536363,-0.2221821547,-0.3009106219,-0.3789571524,0.1374900788,-0.1544614583,-0.2642764747,0.1347445995,0.2034654021,-0.1666638106,0.0633800626,-0.1732999831,-0.1007649377,0.1682905108,0.0854131654,0.2637555301,0.0089568691,-0.0644388273,0.1917527616,-0.1090219021,-0.1644964367,-0.1586015522,0.1599348634,0.0910769776,-0.1871575266,-0.0783970952,-0.2666660845,-0.0998355225,-0.0338013358,0.3180635273,0.3632739186,-0.0149722006,-0.05271364,-0.0648542121,0.1390859038,-0.3818118274,0.4824162126,-0.138551861,0.272847712,0.2167392671,0.1214706302,0.055833932,0.0440627262,-0.1804901809,-0.2143296599,-0.0466567725,-0.1535296291,0.1018019766,-0.3093038499,0.2173660845,0.1267255843,-0.1526685655,0.3397646248,-0.2066389918,-0.3721071184,0.1103957668,0.0945676789,-0.040858794,-0.0062043117,-0.198647812,0.1783259213,-0.1379200071,0.2900523841,0.242014423,-0.1941052228,0.2613470554,0.2286372334,0.5862898827,-0.0528428033,0.0804848596,0.0580657013,-0.1086462066,0.1888636798,0.0946381986,0.0340854935,-0.1129748002,0.5088777542,-0.1981893629,0.1434391439,0.0143694868,0.0490407012,0.0007850596,-0.4658516943,0.0166368578,0.1260615438,-0.210554257,0.0891597271,0.0873117372,0.14191401,-0.1962425262,0.1303602308,-0.1648844332,-0.2436531484,0.0011645299,-0.0606833883,0.3894938231,-0.0845030099,-0.0755919144,0.2057424039,-0.0703736544,-0.1915276349,0.3536187708,0.1826485395,-0.3285893202,-0.4096682668,0.0885547325,0.2275131494,0.0071074124,-0.0833047479,0.4019204974,-0.0832633451,-0.0604124181,0.1581520736,0.1826946288,0.1861813813,0.291275233,-0.2104233205,-0.1039792374,0.0379631408,0.0533444695,0.0217702445,0.4283697009,0.0490645729,-0.0768105909,0.1765136719,0.0655078068,-0.0873609036,0.1394458711,0.2569184005,0.1283602566,-0.2736841738,0.2357609719,-0.1489082128,0.1756699234,-0.088376388,0.1440530568,-0.1109415367,-0.0232593287,-0.0274233036,-0.1055704877,0.0255751349,0.0853485838,0.0447417535,-0.0342579372,0.6097824574,0.2330071479,0.2112393975,-0.3550788462,-0.0700534284,-0.6480929255,0.1435248107,0.0637284815,-0.0543989502,-0.158087641,0.2777583897,-0.2702069581,0.2383376956,-0.1530757695,0.2816689909,0.3561630845,0.1254971325,-0.4057527781,-0.3556218743,-0.0862514973,-0.1886681914,0.0047452119,-0.5090947151,-0.0442722142,0.0472228788,-0.0740122721,-0.0124101844,-0.4688417912,0.2420438975,0.277911216,0.5243609548,0.0045865662,0.2276543528,0.2296289206,-0.1136583462,-0.3318481445,-0.0115297381,0.082251057,-0.1448966414,0.1083540097,-0.0515407249,-0.1801005602,0.1096671894,-0.2946832478,0.5148976445,0.1734207124,0.3646084368,-0.0654628724,-0.0829370841,-0.1057925001,0.3901943266,-0.0210062955,0.1711703092,-0.1948108971,0.326755017,-0.3708812296,-0.24455823,0.4485140741,-0.1018966734,-0.3124238253,0.1936094165,0.3034094572,-0.1146598533,-0.366076231,-0.5041533709,0.2702483535,0.3064178526,-0.153346613,-0.0954082087,0.2173030525,-0.1908441782,0.1835220754,-0.1543015242,0.4260324538,0.0941904187,0.0712041706,-0.0044627115,-0.0562521517]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"Train, Validation, and Test splits contain 138384, 18669, and 17210 samples respectively. It takes some time to read the samples. Even in your colab notebook it was reading the samples before you killed the process. Let me know if it works now!","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":42,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nTrain, Validation, and Test splits contain 138384, 18669, and 17210 samples respectively. It takes some time to read the samples. Even in your colab notebook it was reading the samples before you killed the process. Let me know if it works now!","embeddings":[-0.0954272225,0.2150017768,-0.0282702781,0.3146703243,0.0880519301,0.1548192501,0.2518567741,0.0587237068,0.1317174435,0.0135483062,-0.2907213271,0.294244945,-0.063027285,0.1593239903,-0.1379980147,0.0480366312,-0.1314404756,-0.0800191611,0.2093431652,0.1527232826,-0.1440872252,0.4468840361,-0.3239098787,-0.327598691,-0.2357051075,-0.0038085489,0.0367073528,0.0954877064,-0.0514701568,-0.2540023625,0.6315826774,0.2810488939,0.1258323491,0.5134747028,-0.000117358,-0.1675780565,0.3026215732,-0.0567497648,-0.5882295966,-0.2423142344,-0.4385442436,0.0669240281,0.0713514611,0.0721423477,-0.0678216144,0.468113631,-0.0136968521,-0.5675046444,0.1036921069,0.1793596745,0.180880487,-0.0697262362,-0.1773590595,-0.0844901055,0.1710763574,-0.0346744396,-0.2726801932,0.5722362399,0.2749502957,0.0593251102,0.1192120835,0.3345854282,-0.013206346,0.3771262467,0.0707204193,0.0227241423,-0.2844106555,-0.3562527001,0.397049129,0.2216862887,0.8158916831,-0.2651734948,-0.1943282187,-0.1256033629,0.0726431385,-0.1700063646,0.3319043517,0.1924577057,-0.3744276762,0.0152338324,-0.1312677264,-0.0496278256,-0.0233993158,-0.0152375614,-0.1556309611,-0.1486995816,-0.1854819208,-0.0202173553,0.1402362436,0.0790977105,0.5242319703,-0.1135060042,-0.1165755615,0.2436380982,-0.6248546839,0.0691105053,-0.1014846414,0.1613452584,0.1315224171,0.1775840819,0.2185835987,0.0994748026,-0.2863441706,-0.0031777616,0.2560694814,0.4834381342,-0.0459775254,-0.0941640139,0.203132093,0.3059603572,-0.1675646305,0.0793081895,-0.1169579551,-0.0772604048,-0.0377766639,-0.0337748528,0.2018983662,-0.3512710929,-0.3899106085,0.0125035541,-0.0930636674,0.2361389697,0.0449043214,0.2959346473,-0.34680444,0.148555547,-0.0739104077,-0.038953416,-0.2628845274,-0.0872745365,-0.0986863896,-0.0980872884,-0.0486942939,0.1933555901,0.4289930165,-0.1102925837,0.1828455329,-0.2092844397,0.1033675447,-0.1057472751,0.0984404087,-0.5431810617,-0.0812800452,0.3238000274,0.1397364885,0.3871344328,-0.2562746406,-0.20229581,-0.0877487361,0.3268471062,-0.3405276239,-0.2281249613,0.0071293581,0.0892666951,-0.4724737406,-0.1162059009,-0.3092406988,0.1435221285,0.2485707402,-0.0102978908,-0.0947474614,-0.0128461253,-0.4346059263,-0.2804276347,0.2533690631,0.5699442625,-0.4258940816,0.230786249,-0.334975183,0.224328354,0.193610996,0.1867098957,-0.0140071502,0.1114510074,-0.4291025698,-0.0067005889,0.291716218,-0.2781499624,-0.7794819474,0.2389283627,0.0014892045,0.2033474594,0.0229181163,0.1304776371,0.1507204622,-0.043587327,0.1604697257,0.4117088914,-0.1476098746,0.0172978677,-0.3770851791,-0.3359038532,0.0895883515,0.2272990346,0.0135471635,0.0313052759,-0.0042720078,0.0734458417,0.2714964151,0.3763743341,0.0417548977,0.207447201,0.3754951358,-0.0400265343,0.0129543627,0.005521643,-0.6925061941,0.4320077598,-0.0775319412,-0.087106064,-0.2032730877,0.0011354347,-0.49914819,-0.113199681,-0.3160092235,-0.2926142514,-0.0279086549,0.1228660718,0.2305107713,0.2348573804,0.1045904905,0.6910942197,0.1962856203,-0.0885030776,-0.393810153,0.0615453273,0.0296903159,-0.1168797836,-0.0716760606,0.0280773155,0.0582787357,-0.161901325,-0.2185112536,0.1395226568,0.2593931854,0.0189049225,0.0082036965,0.0197802801,0.3181707263,0.0067134425,0.1502089798,0.0071472749,0.2437195778,-0.152905494,-0.1781631857,0.0367755331,-0.473328203,0.1957154125,-0.2445640862,0.0079243025,0.2596302629,0.0045232824,0.2119437605,0.0076272758,0.3155985475,0.1694469899,0.3344169855,-0.109226048,0.0524763614,0.2959811091,0.1642036289,0.220667541,-0.0794057846,0.1246937364,-0.151683569,-0.1268700212,0.2336960435,0.4344563484,0.6323289275,-0.035062246,0.2059133053,0.0076473556,0.2383215129,-0.2561874986,0.0199435093,0.0333449021,-0.0229340717,0.1256630719,0.2581071854,-0.0045014932,0.1622480154,-0.0430843793,0.1665069461,0.1929653585,-0.0923789516,0.2687755227,-0.1619881988,-0.1913446933,-0.0973138362,-0.1816344708,0.1139667332,-0.3931698799,-0.1319476366,0.3868918419,0.2352635413,0.0197710879,-0.0125679215,0.2920116186,0.0494219176,-0.471886009,-0.2409401685,-0.178058207,-0.2163993865,0.0519582927,0.1794566512,0.0317140371,0.2411491871,-0.1188347191,-0.1186477244,-0.1004496217,-0.1711639315,0.0169870779,-0.0109774359,0.4662133753,0.0267574321,0.2249595821,-0.1297753751,0.0577979274,-0.0152402101,-0.0280416422,-0.0870885849,-0.0635520816,-0.0419487357,0.2378916889,-0.128480196,-0.3079329431,-0.3237249851,-0.3039703369,0.1589350402,0.0231039766,0.01735349,0.4270168543,0.0706182793,0.0454905815,0.3967775106,-0.2290724069,-0.349501282,-0.378993541,0.5690661073,-0.2083933949,-0.1274505258,0.1044181138,-0.0623076111,0.4087072611,0.1532644331,-0.6130222082,-0.0492451154,-0.4274726808,0.4506584108,-0.0927331895,-0.2406839877,0.1335648596,-0.2650046647,0.1155035496,0.0462870076,-0.1447690129,-0.0226460733,0.2541779876,0.3913129568,0.0966663212,0.0932568237,0.162793681,0.578135252,0.1583214402,0.0928006098,0.3064733446,0.1476077139,0.3610422909,-0.2567945123,-0.1881389767,0.0833066329,-0.1589289606,-0.2403326929,0.1650352329,0.0436960757,0.0667289719,-0.1750543714,0.1710814834,0.0225268696,-0.2322509587,0.0235162154,-0.366807878,0.0244963057,0.1655193269,0.173923254,0.141658172,-0.4322201908,0.0605734512,0.1856381297,-0.1419761181,0.046496354,0.4162323177,0.22572954,-0.5468367934,0.1498769373,0.0075770039,0.2178840935,-0.1962669045,0.2093794048,0.2442544848,0.0909829959,0.5848752856,-0.053497497,0.206171304,0.0429372489,-0.0160159618,-0.2721914649,-0.0848733485,0.0593085662,0.223954916,0.1089034379,0.0138874371,-0.2128853351,0.3208968341,0.0627664402,0.2301818281,-0.1963506788,-0.1968142092,-0.2777685225,-0.4279959202,-0.1541115046,-0.0918154418,0.0949116871,0.1662311107,-0.1125854328,-0.1619186997,0.0057878084,-0.188952595,-0.0994115621,0.1571962237,0.1876955479,0.2310693115,0.5053078532,0.2218860835,-0.4908116162,0.0266726408,0.6061955094,-0.0717741847,-0.331628561,0.1051932201,-0.1066438183,-0.2893044651,0.3677919209,-0.2560786307,-0.0485876538,-0.1713088304,-0.1133215353,-0.0192626063,0.3453820646,-0.1500332355,-0.123181425,-0.1232584193,-0.4716365933,0.47737813,0.0228648055,-0.1714971215,0.4526947141,-0.3457268775,-0.1525050998,0.2804819942,-0.1153851524,0.8671264648,-0.2140952349,-0.0482101925,0.1657728702,-0.2647894621,0.4800903201,-0.1820510179,0.1149591804,-0.2979268432,-0.354640305,0.1082071736,-0.1670413166,0.0255148821,-0.136976406,-0.2145306617,0.2820848823,-0.227153033,0.0815107226,0.0031540752,0.0136150345,-0.1160615012,-0.1112677902,-0.1584165841,0.1759820431,-0.2606734931,0.376732707,-0.1981402189,0.0269297343,-0.2867681384,0.1745902002,-0.444704324,0.1477521807,-0.5393671393,0.1069383249,-0.1090276167,-0.5295168757,-0.0282899924,0.2990350425,0.2836942077,0.112820819,-0.2986032665,0.4241594374,0.1085070372,-0.0465257168,0.0907010734,-0.1420481801,-0.1320208758,-0.0631338507,0.0012275053,0.0357645378,-0.1529458761,-0.1545387059,0.1024789661,0.1676947773,0.2249674797,0.0401971824,-0.2084401399,-0.1563213915,0.2295824736,-0.0791280419,0.0847896039,0.2503472269,0.1096874177,0.2126041651,-0.1298063099,-0.243284598,-0.1145165265,0.4787815511,-0.0118386028,0.0389651507,0.4395754635,0.0964130759,-0.2019702196,-0.0327584669,-0.1077834591,-0.6246580482,-0.5375119448,0.3230868876,-0.1863794476,0.1012113318,-0.0705594048,0.3247739375,0.1237588972,-0.2367018312,0.093149431,-0.5435065031,-0.19857876,-0.3119516969,-0.3757375181,0.1384144276,-0.0823714882,-0.2047027051,0.0634241924,0.2262847573,-0.1809338331,0.0895229951,-0.2237007469,-0.0998614281,0.1161865816,0.0179475043,0.2611815631,0.0079672597,-0.0674410611,0.2899776101,-0.0709101632,-0.1712058932,-0.1838661879,0.1547561437,0.0427690074,-0.1486199945,-0.1340640485,-0.2484936267,-0.1180677041,-0.010785358,0.3065955341,0.2694148421,0.0331191085,-0.1046931222,0.0180803388,0.1706997305,-0.3825434446,0.5315548778,-0.1494026035,0.3177739382,0.2272865772,0.1073211953,0.0518265665,0.0529540181,-0.1267414838,-0.2516174018,-0.0183454286,-0.2249572128,0.1314525008,-0.2701285183,0.1803934574,0.0017732905,-0.1325886101,0.3472148478,-0.1771712005,-0.3890730739,0.183413893,0.0779343247,-0.0400381088,0.0436174385,-0.182576254,0.2390052974,-0.1587893218,0.3141076267,0.2402286977,-0.0527438,0.2177194357,0.2396764606,0.5507900119,0.0252178144,0.046455577,0.0385227613,-0.130009532,0.1951766014,0.0742680281,-0.0186083335,-0.1376002878,0.419469595,-0.2050711215,0.1979962736,-0.02792494,0.1023582295,0.0021869992,-0.4949430227,0.0312654935,0.1538413614,-0.3032110333,0.1422282308,0.1170836389,0.2208842486,-0.2133762687,0.0772616416,-0.2105186731,-0.1401477605,-0.0516908169,-0.0028005252,0.4223904312,-0.116657041,-0.1491424143,0.2172988206,-0.0990600139,-0.2019015849,0.3717299402,0.1664247215,-0.3519618213,-0.3245631456,0.0373971239,0.2248092592,0.0322136208,-0.119534269,0.4193250239,-0.1714903265,-0.0099398177,0.0891676396,0.1964031011,0.1845820844,0.2700957954,-0.2681439221,-0.1380283386,0.0516293757,0.0328449681,0.0191889424,0.4536051154,0.185045734,-0.0622499399,0.2061809748,0.0547684804,-0.1205268577,0.1513219923,0.3119697273,0.1272037178,-0.2443739325,0.1705385745,-0.1646249294,0.1715108603,-0.153963387,0.1098515168,-0.137508437,0.0280085262,-0.06858854,-0.1961844862,0.0537496768,0.1352626383,0.0498126447,-0.0500344709,0.6777929068,0.2417130917,0.27138111,-0.3485797346,-0.1659467071,-0.6396212578,0.148750782,0.0483723953,-0.017407015,-0.1623742282,0.2578816116,-0.2073294669,0.2430495322,-0.2053161711,0.3471247852,0.3449111581,0.195066005,-0.4843505621,-0.3318236172,-0.0824968442,-0.2425240874,0.0108208256,-0.4769305885,-0.0164229479,-0.0085531333,-0.0897246003,0.0408441946,-0.5267087221,0.1907693893,0.3206741512,0.5116618276,0.0020494156,0.313945055,0.2300007492,-0.1094815135,-0.2432365566,-0.0129269771,0.133626923,-0.2035776824,0.1809323281,0.0233198255,-0.1101616621,0.2038064301,-0.2465336025,0.4927625358,0.1781431288,0.3462760746,-0.1061306745,-0.1308370978,-0.1282340884,0.4184765816,0.0044470681,0.1910317987,-0.2103832811,0.2931016982,-0.3047272563,-0.1421368271,0.4516492188,-0.1008560136,-0.2481240332,0.1200752407,0.2046428472,-0.123000592,-0.2994627655,-0.5540912747,0.2221765518,0.2787110507,-0.172343865,-0.0971547291,0.2578881681,-0.1244222224,0.1691584885,-0.1640617698,0.4610251188,0.0983029678,-0.0039858893,-0.0489947982,-0.0819673687]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"Hi, it works on colab but it still doesn't work on my computer, same problem as before - overly large and long extraction process.\r\nI have to use a custom 'cache_dir' because I don't have any space left in my home directory where it is defaulted, maybe this could be the issue?","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":52,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nHi, it works on colab but it still doesn't work on my computer, same problem as before - overly large and long extraction process.\r\nI have to use a custom 'cache_dir' because I don't have any space left in my home directory where it is defaulted, maybe this could be the issue?","embeddings":[-0.1124394983,0.2039625049,-0.0727101192,0.4566599131,0.1053852066,0.1380977482,0.2412602454,0.0399661735,0.1459222138,0.0728044286,-0.2571244538,0.1797399968,-0.0175125841,0.0835775211,-0.1670748293,0.0181838013,-0.2205189466,-0.0765432492,0.262817055,0.0581117049,-0.1664710045,0.4603987336,-0.2236673981,-0.2790875435,-0.2222800702,-0.0290817581,0.0343550332,0.0978260487,0.0028531079,-0.1748681068,0.6022046804,0.2330117822,0.1949602365,0.6118329167,-0.0001210657,-0.2649900019,0.2710088491,-0.0766073689,-0.5360262394,-0.1091710255,-0.3738023341,0.061139483,0.1372760236,-0.0116485972,0.0458871648,0.4654710591,-0.1074189246,-0.6454579234,0.104428418,0.1348217726,0.1602918208,-0.148973316,-0.3102101982,-0.089912802,0.0827999935,0.0267981123,-0.2206467539,0.4412756562,0.2664740384,0.0577256456,0.2510488033,0.2638855577,-0.0377928242,0.3813094497,0.1449037641,0.1518656909,-0.2558335364,-0.4166747332,0.4611064196,0.2921358943,0.7200582027,-0.133162111,-0.2939246595,-0.1821670234,0.0125792157,-0.2129628509,0.3156765997,0.2628422976,-0.3349860609,0.0770834088,-0.11117284,-0.1183197349,-0.1295943558,-0.0991460308,-0.2954618633,-0.2909482121,-0.1909498721,-0.0468303934,0.1284815967,0.0477972813,0.4253875613,-0.2764372826,-0.0477600619,0.1911709905,-0.5783612728,0.1121393219,0.0455096029,0.245426178,0.0799599513,0.2400475591,0.2162092477,0.1265865862,-0.283814162,-0.1221358627,0.2870098948,0.5951668024,-0.0842282921,-0.1412150711,0.2576093972,0.2235172093,-0.145800665,0.0932792574,-0.1343962103,0.0148859564,0.019345643,-0.1112164333,0.1985664368,-0.3478776813,-0.3684642017,0.0084945308,-0.0260760728,0.2039990574,-0.0027995019,0.2565117478,-0.2200749069,0.270526737,-0.2517562509,0.0077803265,-0.1173054054,-0.0626892522,-0.0948332027,-0.0068682325,-0.0089578992,0.1656866521,0.4055348635,-0.1726150811,-0.039167393,-0.2994236648,0.0978527963,-0.0339586474,-0.000242449,-0.5300260186,0.0057029636,0.2260296345,0.1368425339,0.4079750776,-0.1774229258,-0.1665692478,-0.2452620715,0.2282978892,-0.2902781367,-0.3212238848,0.0549395941,0.0653731301,-0.4587558508,-0.0608705282,-0.3371566534,0.0665167496,0.2378190458,0.1256387234,-0.0948304906,-0.0256254077,-0.3592674136,-0.2082239985,0.1178154424,0.5704817176,-0.4226481318,0.350644201,-0.3547403812,0.2162886262,0.2554175556,0.2924048305,0.0121941697,0.1714260131,-0.3983627856,0.0105082095,0.2924824357,-0.386695981,-0.8326298594,0.3234568834,-0.0628117323,0.2581585646,0.1022337005,0.0824794322,0.1330956668,-0.0432955921,0.1874770522,0.5307177305,-0.1809251904,-0.0270353146,-0.3147310615,-0.4350322187,0.130693242,0.1489083171,-0.0150489034,0.1042971686,0.0000300174,0.091236718,0.1730369031,0.2804522812,0.0889832824,0.3110314012,0.3881545365,-0.037390966,-0.0309918877,0.0304189995,-0.5355272889,0.432238847,-0.0747448355,-0.0441823192,-0.2751703858,-0.0550684109,-0.3800143301,-0.0545448512,-0.3257084787,-0.2278085053,-0.0647570565,0.1403909773,0.4081245959,0.083685495,0.0989000276,0.6743146777,0.3070165515,-0.1923600733,-0.4451985061,-0.007311421,0.0292270388,-0.0255627371,-0.0624024756,-0.1033241823,0.0079640774,-0.2448357195,-0.1993350238,0.0276979785,0.2593225837,-0.0258349907,0.0311506726,0.0664681941,0.2851371467,0.0224811938,0.3190886378,0.0880980417,0.2208914012,-0.1117255092,-0.2005067468,-0.0054202103,-0.3680719733,0.3170793056,-0.1383224577,-0.0302100535,0.3640990555,-0.0329875723,0.3663946092,-0.1630755514,0.2493323386,0.2290354222,0.3185803294,-0.1538075209,0.037300881,0.3991546035,0.1025302187,0.2353081554,-0.0605215542,0.2296677381,-0.0669093058,-0.1127091646,0.1428695619,0.4477927983,0.6377922297,-0.0017876205,0.2312705666,-0.0116687426,0.4445112944,-0.2531865835,0.1265708655,0.1394788176,-0.1144221053,0.1315434724,0.3815363944,0.1249604672,0.1835181713,-0.1551882774,0.0522776879,0.1829489022,0.0505098701,0.3114051223,-0.2933896482,-0.1412698328,0.0102272574,-0.1402571499,0.1915502697,-0.3343356848,-0.0997165963,0.2889537513,0.1851547062,-0.0220518783,-0.0802043006,0.2349568307,0.1068214029,-0.6577318311,-0.1860103756,-0.12593548,-0.1675988287,-0.0205535926,0.16597794,-0.0659078807,0.1569801271,-0.1125473902,-0.0452650748,-0.2311018854,-0.1205702573,0.1382140517,0.1124632284,0.3400293589,0.1230643764,0.1951078773,-0.0293242093,0.0919511989,-0.0122229662,-0.0432704799,-0.0923534036,0.0983579084,-0.0017712895,0.2034452111,-0.1427102983,-0.218968913,-0.3988292217,-0.3529396951,0.1966072619,0.1087290347,0.1134831756,0.5008627176,-0.0045686108,-0.0660050139,0.3374961913,-0.1798048168,-0.3153896034,-0.2892059684,0.6500527859,-0.1588740349,-0.1606778353,0.0356204323,-0.1235070154,0.3259694576,0.311514765,-0.596796453,0.0108579295,-0.5298214555,0.5150713921,-0.1468387395,-0.150575608,0.2410836965,-0.2570917308,0.0925663486,0.1877712309,-0.1138511747,-0.0390502475,0.361372143,0.324747324,0.1278100461,0.0414549448,0.1006369293,0.4964287281,0.181912452,0.1965619922,0.2644200623,0.2189560533,0.3408922255,-0.310798347,-0.1178690642,0.0881003141,-0.1094503775,-0.3138961494,0.2442173511,0.0095251556,0.1493338048,-0.1414439976,0.1356516927,0.0246780142,-0.2404597849,-0.0037691572,-0.3402994871,0.0382092707,0.1952247918,0.1561392546,0.0683294758,-0.3506827652,0.0516655929,0.0779483691,-0.1360313296,0.0936768055,0.362006247,0.1988880038,-0.6867262721,0.2976361811,-0.0673015714,0.1260100007,-0.1006123796,0.13430731,0.1694939286,0.0397765376,0.4954964817,-0.0662682056,0.1842703223,0.0231706258,-0.1181606501,-0.3150578439,-0.053742677,0.0341007374,0.2516176403,0.1035482138,-0.0376144685,-0.2355343848,0.2097641081,-0.0143410927,0.396070987,-0.2495847195,-0.2628289461,-0.1983995736,-0.5200641155,-0.1323180795,-0.1656607091,0.1134948358,0.2910748422,-0.0096833333,-0.080279991,-0.064998582,-0.0512035005,0.0603083819,0.1282627285,0.2356875539,0.0832986236,0.4498069882,0.054903429,-0.3317554593,0.0727713928,0.5874845982,-0.0691995546,-0.2378368378,0.1447430253,-0.1471159458,-0.2464408278,0.3908081949,-0.1837968826,-0.0847089589,-0.2200707495,0.0337483548,-0.0675563589,0.1908866167,-0.2788873017,-0.2726649344,-0.0432045385,-0.4134301543,0.4249438047,-0.0096348859,-0.1675673425,0.4580017626,-0.2474299222,-0.1639781892,0.2043065429,-0.1585938781,0.9138550758,-0.0805505142,-0.0034696206,0.1507047862,-0.2079460919,0.4514724314,-0.1793815643,0.0711289346,-0.2143279761,-0.267608881,0.0974445939,-0.0965109468,-0.0214511566,0.0284408871,-0.1200101227,0.3581600785,-0.3298619688,0.0152348168,-0.050055705,0.207313776,-0.1306747496,-0.2421005964,-0.2531361878,0.1241845414,-0.1924165487,0.3656939864,-0.1789283454,0.0241738856,-0.3195342124,0.0672561601,-0.3876726627,0.1608821601,-0.4400922358,0.1835901439,-0.0828127488,-0.6697801948,0.0536903962,0.2453933358,0.2026811093,0.0219541695,-0.3091062307,0.3031800687,-0.0374058187,-0.0393263474,0.152852729,-0.2124556005,-0.1889004558,-0.0733266771,0.0999685675,-0.1232620627,-0.1935406923,-0.0720458701,0.1344902366,0.219412908,0.2226474434,0.1160175502,-0.3141058683,-0.1978795826,0.1567371041,-0.0357589349,0.0456863195,0.2854549289,0.0556826554,0.3891106248,-0.3272897601,-0.2102783322,-0.129797101,0.4838750958,-0.0143465204,-0.0976509079,0.4774668515,0.0779028758,-0.2222819775,-0.053730119,-0.0966772288,-0.6926589012,-0.3858878613,0.2867058218,-0.2331440449,0.0293819699,0.0122663938,0.1816850454,0.0957051292,-0.2289653271,0.0206919145,-0.5060122013,-0.1740646511,-0.2611438632,-0.3473089039,0.0542530902,-0.140859291,-0.236777395,0.0571804158,0.2233096808,-0.1493703723,0.0729501992,-0.1299003363,-0.1688790917,0.1508990526,0.0259112362,0.3095939457,-0.0542562716,-0.0970711783,0.2362462133,-0.0349564105,-0.15880391,-0.200898841,0.1948210895,-0.0205996055,-0.0862790495,0.0120639596,-0.2491587251,-0.1334414333,-0.1085189655,0.3305954337,0.3707275391,0.055456873,-0.2095954418,-0.1150323525,0.1533996463,-0.4814027548,0.5189342499,-0.1384075284,0.4532432854,0.2359662354,0.1354412436,0.1091645211,0.0201759748,-0.1902893633,-0.1820201278,-0.0362317972,-0.2306596935,0.1343621761,-0.2551439404,0.1454495341,0.1495040506,-0.2014383525,0.4630461335,-0.1462032497,-0.37997967,0.0736227483,0.0506367758,-0.0483804792,-0.0101916511,-0.1379198134,0.1518627554,-0.2151199281,0.3758493662,0.2288944721,-0.0797499493,0.1907574832,0.2228185534,0.6447782516,-0.0168135818,-0.0011569094,0.0510929041,-0.0798631981,0.1751161963,0.1628041267,0.1315268427,0.0321823806,0.6310331821,-0.2361091226,0.112343803,-0.0444177501,0.1189578027,-0.0537397563,-0.5293367505,-0.1235545874,0.1374858469,-0.1704308093,0.1545784473,0.0553770438,0.1203292534,-0.2863411009,0.1980026215,-0.2747959197,-0.1305593997,0.0702808276,0.0432925671,0.4179134071,-0.0326130055,-0.0213607624,0.2362586409,-0.0460084267,-0.1254775673,0.3595543504,0.1674815416,-0.2902520597,-0.30452618,0.1098576486,0.0775432587,-0.0075031309,-0.1655341238,0.3842706084,0.0379329287,-0.0154903624,0.0174712073,0.187150225,0.1906708628,0.3469963372,-0.3113017678,-0.1345649213,-0.0530818701,-0.0076328875,-0.1113528088,0.3449161649,0.0573172309,0.0014296599,0.2149898261,0.0319243297,-0.0873318911,0.0654171556,0.3005395532,0.1960020363,-0.2324970514,0.2070972621,-0.0842723921,0.1450963169,-0.177627027,0.1186173558,-0.1595536172,0.055943422,0.0064926003,-0.1260638088,0.0329538547,0.1499284357,0.0170865562,-0.1517686546,0.5939472318,0.2363781035,0.2501508296,-0.2938681245,-0.0702615902,-0.6929665208,0.1366198212,-0.0587029271,-0.1349783242,-0.1556126922,0.4354976416,-0.2148852944,0.2857705653,-0.1770946532,0.2804149091,0.2239731252,0.1362756938,-0.531121552,-0.284810096,0.0016272641,-0.1559209526,0.0432397984,-0.6099458933,-0.0377910286,-0.000166414,-0.136899963,-0.000647529,-0.4042120278,0.1251641363,0.38561818,0.5973390341,0.1163575128,0.1736588627,0.2634854317,-0.1178943291,-0.2479631156,0.1287066191,0.1317954063,-0.108243078,0.1779836565,-0.0249573868,-0.2282369137,0.1317043453,-0.2528842688,0.523578763,0.2184141129,0.2672612369,-0.1370813996,-0.0560126714,-0.1602931619,0.4422332346,0.1186151952,0.2526869476,-0.1493698657,0.2445573509,-0.3856603801,-0.2008840889,0.4115626514,-0.0085625742,-0.4528909028,0.1688714921,0.2824739516,-0.1177220345,-0.2932255566,-0.425193876,0.2317073196,0.2437926382,-0.1594468355,-0.1062924191,0.181069985,-0.2410930097,0.0200015418,-0.1364493519,0.3371227086,0.1063127145,-0.0014481164,-0.147181794,0.0167581495]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"I tried running this again - More details of the problem:\r\nCode:\r\n```\r\ndatasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir=\"\/path\/to\/cache\")\r\n```\r\n\r\nThe output:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to path\/to\/cache\/trivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.67G\/2.67G [03:38<00:00, 12.2MB\/s]\r\n\r\n```\r\nThe process continues (no progress bar is visible).\r\nI tried `du -sh .` in `path\/to\/cache`, and the size keeps increasing, reached 35G before I killed the process.\r\n\r\nGoogle Colab with custom `cache_dir` has same issue.\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing#scrollTo=2G2O0AeNIXan","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":81,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nI tried running this again - More details of the problem:\r\nCode:\r\n```\r\ndatasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir=\"\/path\/to\/cache\")\r\n```\r\n\r\nThe output:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to path\/to\/cache\/trivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.67G\/2.67G [03:38<00:00, 12.2MB\/s]\r\n\r\n```\r\nThe process continues (no progress bar is visible).\r\nI tried `du -sh .` in `path\/to\/cache`, and the size keeps increasing, reached 35G before I killed the process.\r\n\r\nGoogle Colab with custom `cache_dir` has same issue.\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing#scrollTo=2G2O0AeNIXan","embeddings":[-0.0191132333,0.2346220762,-0.0358678177,0.3387272656,0.1412011832,0.1956146508,0.2685050666,0.0284198001,0.1188834086,-0.0826994851,-0.2390524,0.2821522653,0.0216848385,0.0726570189,-0.0835466832,0.0234450512,-0.1319268644,-0.077235058,0.1958685517,0.1156656966,-0.1322735101,0.4592790306,-0.274974972,-0.317925185,-0.2377310246,0.0502318218,0.0789388195,0.1147859767,0.0367471986,-0.2417893857,0.6353093982,0.2854878306,0.1081413254,0.5510926843,-0.0001176656,-0.1716504246,0.3202268481,-0.0672563165,-0.6127192974,-0.2736315131,-0.4003522098,0.0468358658,0.0461337902,0.0724672452,0.0060809413,0.4613676965,-0.1097799614,-0.7131503224,0.0690336972,0.1353898197,0.1855592728,-0.1125454381,-0.2207427323,-0.0838005543,0.0829930454,-0.0740477368,-0.2689909935,0.4667278826,0.1987545192,0.0861581787,0.2101964206,0.260833323,0.0145356264,0.384326309,0.1283073723,0.0625407398,-0.2769593894,-0.3085980713,0.4427739084,0.2299291641,0.8865922093,-0.3260815144,-0.3149341345,-0.1055868566,0.0569770783,-0.1421446353,0.2633228004,0.1837041527,-0.2723793387,-0.0140484553,-0.1401576847,-0.0825646371,-0.0307545867,-0.0464234352,-0.2129700482,-0.241199702,-0.1517520696,0.0083343768,0.0067779361,0.0274224039,0.4106859267,-0.1950242072,-0.1808835566,0.2906986475,-0.6349016428,0.1571294069,-0.0411919281,0.1883413643,0.1735039949,0.1483285576,0.2045854181,0.1414367259,-0.3397989273,-0.0378533378,0.2626331449,0.531612277,-0.0408639833,-0.0834982246,0.2299902588,0.2655797899,-0.1089846715,0.0714503825,-0.0711207166,-0.047081247,0.0569608398,-0.0621099919,0.241721943,-0.3457325101,-0.4039240777,0.0383145735,-0.0017094251,0.2466624677,0.0015263157,0.2440172732,-0.275067836,0.2244942486,-0.0682820752,0.074724786,-0.1628013402,-0.0360300541,-0.0955939367,-0.1428514123,0.0032888087,0.1493494064,0.4283840954,-0.1551055759,0.1066986248,-0.156065464,0.1478950381,-0.1040507704,0.0231474452,-0.4320104122,-0.107814461,0.2410247624,0.0746769011,0.4414884746,-0.2072996646,-0.1470959932,-0.0880824551,0.2640244067,-0.293947041,-0.2577490509,-0.0445576347,0.0855401382,-0.4759041071,-0.1131555066,-0.294577986,0.1743015945,0.303640455,0.0474472009,-0.0126435179,-0.0173471384,-0.4301814437,-0.2669342756,0.2036341727,0.6221860051,-0.4207316041,0.1967634559,-0.3082355261,0.194241032,0.1443872899,0.179546684,0.0068101212,0.0972253382,-0.4354678392,-0.0398846604,0.2702302337,-0.3908962309,-0.7491314411,0.2808628082,-0.0086194472,0.2237785906,0.0303814951,0.1010896787,0.1038411781,-0.0577310324,0.1362883002,0.4570526183,-0.1250881404,-0.0197001416,-0.3529679775,-0.2770241201,0.0781979263,0.2064096481,-0.0646447986,0.1350308359,0.0128663061,0.0768369287,0.2177930772,0.3736059666,0.1321482807,0.1880967021,0.2888700962,-0.0331768543,0.0478501543,0.0071891216,-0.8005840778,0.4807349145,-0.0945226029,-0.1170418188,-0.259799242,-0.0258311182,-0.5113695264,-0.0753076971,-0.2906000018,-0.283392936,-0.0104413424,0.1654554307,0.2727052271,0.1715573221,0.0851592645,0.6717213988,0.2998747826,-0.1202849969,-0.3922044337,0.069436118,0.0767221674,-0.0590650626,-0.0433744974,0.0797144324,0.1047338471,-0.1470641196,-0.203613326,0.1653554142,0.1932532787,0.075868085,0.0367464796,0.0325158127,0.253762424,0.0919819102,0.1649685204,0.0511354506,0.2706748545,-0.1777030528,-0.1767133027,-0.0247356147,-0.3629616499,0.2979874611,-0.2117595077,0.0285086688,0.2780199051,0.0143422978,0.2297800183,-0.0664697737,0.282533139,0.1815014631,0.3434978724,-0.1430892646,0.0949134603,0.3221319616,0.1214525625,0.1996028721,-0.0438179523,0.1478548944,-0.1202018857,-0.1116029993,0.1134591326,0.4167696834,0.6022350192,-0.0363141373,0.2517687678,-0.0315059014,0.2953365445,-0.2402058989,0.0384855978,0.058603812,-0.1457096338,0.1686426252,0.3002696037,-0.0121428287,0.2025443316,-0.0837552994,0.1454544961,0.1655578762,-0.1184707209,0.2592589557,-0.2133530974,-0.1156928539,-0.0370280258,-0.203744024,0.0560552962,-0.4093008339,-0.1734046042,0.3552066982,0.246805802,-0.0204446856,-0.107994929,0.2236550301,0.0996357426,-0.5879987478,-0.299007237,-0.1832721829,-0.2583683133,0.0222613122,0.1391981989,-0.0598681718,0.2214749008,-0.0970888883,-0.0285948794,-0.2190076858,-0.1292646676,-0.0258197282,-0.0072490936,0.4299786985,0.0988122672,0.203066349,-0.0951739252,0.2066628933,0.0094749322,-0.0003330787,-0.0543339662,0.0226713307,0.0008212873,0.1979881078,-0.15129219,-0.309789598,-0.3121215999,-0.35847193,0.1870757043,0.0355167873,0.0285188444,0.4856404662,0.0236524437,-0.012974075,0.3645753562,-0.2017696798,-0.4081791639,-0.3453952968,0.5892272592,-0.2119027674,-0.1583936363,0.1197955683,-0.0461748429,0.4122824967,0.1915090084,-0.583807528,0.0116565665,-0.4313183129,0.5346263647,-0.1392471045,-0.1815200895,0.1536774337,-0.262444973,0.1151670367,0.0053054686,-0.1958691627,0.002066188,0.2577970922,0.2908815145,0.1448958218,0.1264071465,0.1457356662,0.4955901206,0.1865119487,0.0836846232,0.3631515801,0.1688725948,0.4154903293,-0.2770463228,-0.1153901145,0.0173894502,-0.0924721062,-0.2025767267,0.1484269798,0.1045753807,0.1818237305,-0.1652718186,0.1738299131,0.0539359599,-0.1709293574,-0.0176242273,-0.4410480857,0.0775508881,0.1798340529,0.1241006926,0.1254856884,-0.3712650537,0.0546165258,0.1717485189,-0.0691058859,0.0827201456,0.3950489759,0.2830405831,-0.5196454525,0.1963971555,-0.0285703354,0.3078733683,-0.2408765852,0.2567875683,0.1941807121,0.0760756359,0.5441389084,-0.094689779,0.2514659166,0.058081653,-0.0221393127,-0.3381403983,-0.0738257468,0.0900607705,0.2846440375,0.0447066948,0.017563194,-0.2622878253,0.3053124845,-0.0892377645,0.2814872861,-0.1907856166,-0.2799037099,-0.2527105808,-0.3981196582,-0.1556324065,-0.1745398492,0.0683006197,0.1983937621,-0.062065769,-0.1559293568,0.0003540722,-0.170691967,-0.1382349133,0.0894411579,0.2080431879,0.1888033748,0.5020067096,0.194750607,-0.5206371546,0.0263912342,0.5982953906,-0.0079599712,-0.2585740089,0.0666796565,-0.1071565151,-0.3194825649,0.2468573302,-0.2781352401,-0.0596500002,-0.1639051586,-0.1132393852,0.0210730694,0.1549765915,-0.1706638783,-0.1680389047,-0.128439948,-0.4932682812,0.5392371416,-0.0051737749,-0.184492588,0.4501324594,-0.3309911191,-0.199438706,0.203665331,-0.1315498948,0.7409493327,-0.1532824934,-0.0219075773,0.1534796208,-0.3121378124,0.4592402577,-0.1370946765,0.0242901221,-0.2828537822,-0.2036294043,0.0698862821,-0.1276931316,-0.0080811959,-0.0273979455,-0.2072892338,0.3313939869,-0.3650595248,0.1006971598,0.0516707562,0.0472790115,-0.1720716953,-0.2131943852,-0.2071440071,0.17771855,-0.2885351181,0.3594101071,-0.1883131266,0.0432253554,-0.3233253062,0.1506308615,-0.4407791793,0.1631144732,-0.4221429527,0.0890822634,-0.1084937155,-0.5437339544,0.0140564255,0.2419080436,0.2625370026,0.0221083872,-0.3237560391,0.3971302509,0.0593231171,-0.0015365478,0.1392301321,-0.1147689447,-0.1784151345,-0.0305855945,0.0294051301,-0.000315078,-0.1538187861,-0.0790109187,0.0043022768,0.1569542587,0.1965064704,0.1055629402,-0.1669839025,-0.1564686745,0.2185391635,-0.070940271,0.0791637301,0.2919027507,0.0854126886,0.2118907124,-0.1748491824,-0.2381022722,-0.0882193744,0.5300829411,-0.0337807126,-0.0233992115,0.4653880298,0.0922389254,-0.162016049,-0.0205393732,-0.1402368248,-0.6055622101,-0.4811120927,0.2822018862,-0.1345674247,0.1384404004,-0.0690787211,0.2516883612,0.0785120651,-0.1928959787,0.0841267258,-0.5883567929,-0.2071657926,-0.3393982649,-0.3894107342,0.127051264,-0.176082626,-0.2725810111,0.1475329697,0.203455314,-0.1656276435,0.0422469229,-0.1583331078,-0.1188724786,0.1814659685,0.1070237979,0.2363226712,0.0029268339,-0.0544721074,0.2204471231,-0.1146871075,-0.1680714786,-0.1807156056,0.1661593169,0.0735968649,-0.1433023661,-0.0776597261,-0.2586244643,-0.1208892316,-0.0386880301,0.304530561,0.3473717272,-0.0146406284,-0.0699677244,-0.0420328192,0.2171614468,-0.3868172467,0.4742593765,-0.1237716079,0.2924684584,0.2344244272,0.1151464805,0.0340176858,0.0401512682,-0.1915518641,-0.2325370908,-0.0525234863,-0.1627646834,0.1260344684,-0.3381893635,0.2135657817,0.0655585229,-0.1083589494,0.3362697065,-0.2090881765,-0.379950285,0.0725133643,0.0941547528,-0.0089040734,0.0128663098,-0.1946444213,0.1920184493,-0.1543806642,0.2885892093,0.2812696397,-0.2293297946,0.2159969211,0.2215421498,0.6129818559,-0.0681276992,0.0681177005,0.088719815,-0.1129189432,0.2292172909,0.0956120118,0.0488049313,-0.1103403419,0.5271968842,-0.1658565849,0.1481882781,0.0386571884,0.0795997158,0.0036874532,-0.4858110845,0.0137196537,0.1520477086,-0.1910654455,0.1370664686,0.0971600041,0.1184920892,-0.1806696355,0.113299042,-0.1919310838,-0.2406312972,0.0084101046,-0.046036616,0.3963513076,-0.10627646,-0.0845840648,0.1787839681,-0.0532914586,-0.1944562048,0.3645761907,0.1588547826,-0.3332526684,-0.4112211168,0.0916550085,0.224017024,-0.0072518624,-0.1169890985,0.4023095965,-0.0411143973,-0.0485871881,0.172427237,0.1727314293,0.2019569725,0.3180104792,-0.2198442817,-0.0466550551,0.0393261611,0.0616999418,0.0069271848,0.4818116724,0.0255242735,-0.0796539262,0.2072611898,0.0694755614,-0.0898694322,0.1094669476,0.3004081249,0.1705380678,-0.2708091438,0.2643252611,-0.1876896322,0.1683550477,-0.1009247527,0.1464059204,-0.1244278401,-0.0345959105,-0.0075278976,-0.1026270092,0.0213439167,0.0891810954,0.0441833958,-0.0608277544,0.6165288091,0.2023578733,0.2472774088,-0.3340443075,-0.0762168169,-0.6680170298,0.1565926671,0.0737018064,-0.0707147866,-0.1641814858,0.2768978477,-0.2316913456,0.217199415,-0.1182686538,0.2808210254,0.3394160569,0.1648774147,-0.4000347257,-0.4039432108,-0.0748031139,-0.1645415574,-0.0130372765,-0.4930550754,-0.046119839,0.0432861708,-0.0618439689,-0.0203051399,-0.4816482961,0.2203567177,0.3078497648,0.5403585434,0.0276348908,0.242197454,0.2108855546,-0.1230483428,-0.3688016832,-0.0013881596,0.099699311,-0.1126253083,0.0906551406,-0.0464227162,-0.1530633569,0.1046940461,-0.2722202241,0.4707448184,0.1713112146,0.3378372192,-0.0568043366,-0.0787906051,-0.1075412259,0.3662706316,-0.0116847381,0.1998780817,-0.1851875186,0.3140194714,-0.309692204,-0.2390624434,0.4232536256,-0.1126867533,-0.3413986564,0.2263753414,0.2746394873,-0.1444095969,-0.3413013518,-0.5097907782,0.2802210748,0.3395372927,-0.215257585,-0.071280174,0.1830790192,-0.2083290219,0.1669014543,-0.1470669657,0.4213539362,0.0891154557,0.0543557964,0.0091716275,-0.062296506]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"1) You can clear the huggingface folder in your `.cache` directory to use default directory for datasets. Speed of extraction and loading of samples depends a lot on your machine's configurations too.\r\n\r\n2) I tried on colab `dataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")`. After memory usage reached around 42GB (starting from 32GB used already), the dataset was loaded in the memory. Even Your colab notebook shows \r\n![image](https:\/\/user-images.githubusercontent.com\/19718818\/102852229-c7c4e780-4443-11eb-91d6-bf21024358a3.png)\r\nwhich means it's loaded now.","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":73,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\n1) You can clear the huggingface folder in your `.cache` directory to use default directory for datasets. Speed of extraction and loading of samples depends a lot on your machine's configurations too.\r\n\r\n2) I tried on colab `dataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")`. After memory usage reached around 42GB (starting from 32GB used already), the dataset was loaded in the memory. Even Your colab notebook shows \r\n![image](https:\/\/user-images.githubusercontent.com\/19718818\/102852229-c7c4e780-4443-11eb-91d6-bf21024358a3.png)\r\nwhich means it's loaded now.","embeddings":[-0.0378204137,0.1969360858,-0.0213918537,0.3770355284,0.0714803264,0.1753475517,0.2409699857,0.0573944151,0.127870515,-0.0225050412,-0.2598694861,0.2617431879,-0.0099751623,0.1539277285,-0.0962107927,0.0181324072,-0.1629698277,-0.0272244308,0.1827304214,0.1249765903,-0.1456334442,0.4732732773,-0.3000766337,-0.3055313528,-0.2189928889,0.0209773611,0.0766963586,0.104297854,-0.0106252572,-0.2406463474,0.6824563146,0.2984650433,0.1582715958,0.5067254305,-0.000116904,-0.1950888783,0.2863745391,-0.0882358924,-0.5817727447,-0.2903612852,-0.4384790659,0.0601245612,0.0849269405,0.0675119832,-0.0535828397,0.4907841086,-0.0729141086,-0.6632893085,0.0589521229,0.1665002257,0.1875519454,-0.0474640839,-0.2352565974,-0.1112776622,0.1040510833,-0.0874529034,-0.2666222751,0.4951730669,0.2078435719,0.1018992737,0.2289200723,0.3050989807,-0.0203275681,0.3870891333,0.1716640741,0.0868210569,-0.2917646468,-0.3670197725,0.4509593248,0.2974137664,0.8650926352,-0.2715186775,-0.2730097473,-0.0695945993,0.0907623842,-0.1313936114,0.307498157,0.1710153371,-0.2946821749,-0.0152117237,-0.1419036239,-0.1120753363,-0.0633661151,-0.0124609992,-0.1693012416,-0.2273414433,-0.2012782097,-0.0119249281,0.0618782379,0.0034097272,0.4565645158,-0.2102435231,-0.1713304818,0.266010046,-0.6071332097,0.1296246797,-0.1088141128,0.2369006276,0.1592016965,0.1369526237,0.1727169454,0.0987171531,-0.3527310193,-0.0351227634,0.2121552825,0.5477255583,-0.0109770065,-0.1223005727,0.2560369372,0.3002017438,-0.1371740848,0.0925453752,-0.0757906809,-0.0648038387,-0.0123179266,-0.081881091,0.2244823426,-0.3794716001,-0.3778353333,-0.0424906127,-0.0378495008,0.2366098315,0.0364250094,0.2636562288,-0.2741851509,0.1741034091,-0.0838901401,0.0420697518,-0.1603948772,-0.0573036633,-0.0917438492,-0.1155974492,0.0423483029,0.1906615347,0.4011397064,-0.2065199465,0.1264312863,-0.1874910593,0.1777374744,-0.0363082439,0.0199493393,-0.4155361652,-0.0973082781,0.247795105,0.1213410646,0.4248805344,-0.2058926374,-0.1840034872,-0.1319413632,0.23598589,-0.3165967166,-0.2726558745,-0.0210115407,0.0798255131,-0.4653993547,-0.1361622065,-0.3558708727,0.1462778002,0.234274298,0.0311608352,-0.0463752747,-0.0202396959,-0.4285145998,-0.2600781322,0.2213883847,0.6070357561,-0.398234427,0.2053857297,-0.2950755358,0.2050085664,0.1462543756,0.2117609382,-0.0300156977,0.0446040221,-0.4265945852,-0.0395600498,0.2726883888,-0.4085456133,-0.7721036673,0.202782169,0.033973407,0.2119552642,0.0570143573,0.12341737,0.1194112375,-0.0648046583,0.1202156469,0.4723507166,-0.1216243804,0.0043084258,-0.3479326069,-0.3304363787,0.0952135995,0.1849820465,-0.057031922,0.1299722642,-0.0222084355,0.0487578362,0.170876801,0.3199008107,0.0690201372,0.1946125329,0.3535335958,-0.0481612496,0.0248623658,0.0106180422,-0.7797961831,0.4660256207,-0.083256565,-0.0681072921,-0.2493628114,-0.0339063071,-0.4600460827,-0.1263608634,-0.3008000851,-0.3012436926,-0.0207808614,0.1813722551,0.3229856789,0.1746608317,0.0593355633,0.7220692635,0.2963217497,-0.0923906565,-0.3953096271,0.0675364062,0.0664163604,-0.0736004114,-0.0174641833,0.0514175482,0.0939883366,-0.1732202768,-0.2242990881,0.1480604112,0.1868628263,0.0319901146,0.0709332675,0.0308086295,0.2856915593,0.0174240693,0.1968009472,0.0230444875,0.2305075079,-0.1218319982,-0.1542066336,0.0039639054,-0.4010272026,0.2766900957,-0.2161407024,0.0326379389,0.2775684595,0.0468800366,0.2360722125,-0.0684322044,0.2617351115,0.2082517892,0.3502916694,-0.1108467206,0.0505608879,0.3242493272,0.1505035907,0.2410300076,-0.0496642962,0.1662349999,-0.1415006369,-0.1254716367,0.1596951485,0.412473768,0.5925504565,-0.0173770096,0.2083378136,-0.0246682446,0.2961340845,-0.2416315228,0.0521872267,0.0349838287,-0.1199700162,0.0931526795,0.3018989563,0.01543862,0.1826259345,-0.083641395,0.084378846,0.1672872901,-0.0572505742,0.2478214055,-0.2375026941,-0.2205661237,-0.0380853713,-0.19256109,0.0287783667,-0.341162622,-0.1547432393,0.3447343409,0.2560048103,-0.0188446585,-0.0518012792,0.2708902061,0.0623524971,-0.5690119863,-0.278600961,-0.1567381173,-0.2715739012,0.0196659621,0.1388840377,-0.0198112577,0.2030585706,-0.1676170975,-0.0171669684,-0.2220652252,-0.1390314549,0.0090787802,0.0060735,0.4389398396,0.1061840579,0.229515478,-0.1326919794,0.1424076855,0.0056625861,-0.0073766881,-0.0655236989,-0.0197632238,-0.0305116735,0.2000202984,-0.1168845668,-0.3033442795,-0.3647994697,-0.335850507,0.1801334172,0.0498668514,0.0223348122,0.4906466901,0.0020801595,0.0206448827,0.330388248,-0.148971647,-0.3603318036,-0.4174559712,0.6014065742,-0.2016437352,-0.1822627932,0.1378254443,-0.0615997352,0.40658921,0.1509123296,-0.586907208,-0.0809577405,-0.4267997444,0.5569659472,-0.0758911967,-0.1767404228,0.1463646293,-0.2511541545,0.1215645447,0.0279136822,-0.2074700445,-0.0166781601,0.2717612386,0.3219144344,0.1806330234,0.0880651399,0.1262785047,0.5181487799,0.2077556401,0.0797469616,0.3638018966,0.1753033996,0.3964679837,-0.2793215811,-0.1735205352,0.0575630032,-0.0832296237,-0.203456074,0.1787530929,0.0977607146,0.18786484,-0.1921940595,0.1710539609,0.0384585783,-0.1988487393,-0.046292711,-0.3616319001,0.0609244853,0.1831429005,0.1158303842,0.1133392155,-0.3928630352,0.0840607211,0.2526940405,-0.1262196749,0.1214868873,0.3857606649,0.230997473,-0.553461194,0.2372852713,0.0018783974,0.2450123429,-0.2240564972,0.2852639854,0.1927921325,0.1100050136,0.5230179429,-0.0884748623,0.1987654418,0.0231842529,-0.0740380064,-0.2951200306,-0.0542874336,0.0760575533,0.2667066157,0.0670180991,0.045284234,-0.2487856001,0.2691775858,-0.0243187118,0.2936882973,-0.1839977801,-0.2175713032,-0.2482480258,-0.4296460748,-0.1410507113,-0.1887928396,0.0589421652,0.2183938622,-0.0335481651,-0.1168378219,-0.0244840924,-0.1741449237,-0.1122997478,0.1407258809,0.2288200557,0.2203713208,0.4981544912,0.2033435553,-0.524617672,0.0551405177,0.5850197077,-0.0583945997,-0.3004552424,0.0111046014,-0.0649569407,-0.302300781,0.2721959651,-0.2975708842,-0.0676441342,-0.1775530726,-0.1155495197,0.0611535758,0.2059529275,-0.1886909753,-0.1520280391,-0.114458099,-0.4708744287,0.5124666095,0.0411495604,-0.1552053243,0.3966341317,-0.2734747231,-0.1805183887,0.1921937764,-0.1469857693,0.775680542,-0.1921826154,-0.0300763808,0.1549893916,-0.2656217813,0.5389317274,-0.0968945697,0.0451293066,-0.2695481479,-0.3054897189,0.0805153698,-0.1332689971,0.0123725487,-0.0849461257,-0.2445513457,0.351154685,-0.279024899,0.0592627823,0.0335564464,0.0595695823,-0.1348479837,-0.2305525243,-0.2572596967,0.1751279086,-0.3139318824,0.416554451,-0.1755643785,0.0114753274,-0.2852796912,0.164512977,-0.4088957906,0.1740792245,-0.4878209233,0.0743987486,-0.0582339838,-0.5423918962,-0.0075046024,0.2715565264,0.3043860495,0.0374094397,-0.3208135366,0.336948514,0.0693715289,-0.0517882854,0.104011789,-0.115268223,-0.1634016484,-0.0540847294,0.0470896289,-0.0023301835,-0.127288416,-0.0996445566,0.0815312713,0.1404269636,0.205503732,0.1185604781,-0.1997354776,-0.1690602601,0.2651328146,-0.0764438957,0.0808740631,0.2749656439,0.0785579085,0.2011075914,-0.1354622394,-0.243091166,-0.0981511846,0.4980009496,-0.0553408973,-0.00109521,0.4211676717,0.1036757901,-0.1822806895,-0.0191634689,-0.115427278,-0.652387619,-0.446203351,0.3132854998,-0.1218560711,0.1035806984,-0.0763662979,0.3590603769,0.1304398477,-0.2463641018,0.0967590138,-0.562264204,-0.2025083303,-0.2914939523,-0.3396280706,0.1370552331,-0.1141140908,-0.2731301785,0.1120392233,0.1898609251,-0.164766714,0.0618520342,-0.1895029545,-0.1244169772,0.1951440424,0.1091445163,0.2884381115,0.0401099063,-0.0651831925,0.2081223577,-0.1011092439,-0.1718211472,-0.1848806441,0.1607465744,0.0726068318,-0.2027557939,-0.0686552301,-0.2751318514,-0.1262378246,-0.0444717593,0.3397128284,0.3327182829,-0.0025919306,-0.0679652765,-0.0563223809,0.1373612732,-0.4157472849,0.5315402746,-0.1125098541,0.301159054,0.2221142352,0.1018565297,-0.0056066248,0.0165555812,-0.1387830228,-0.2388537973,-0.0481826253,-0.1930406094,0.0811532885,-0.297442615,0.2074742913,0.0975277349,-0.1519770324,0.309433639,-0.1741313636,-0.3294197023,0.1066070199,0.0786715895,-0.0416874774,-0.000476144,-0.1906666011,0.2172415107,-0.1319389492,0.3239268959,0.2600619197,-0.1550657451,0.2694016695,0.1862222701,0.5831526518,0.0079974234,0.0914920419,0.0158877335,-0.1148679629,0.2185667455,0.101017803,0.0639428496,-0.1130049005,0.4651258588,-0.2042327225,0.1500175297,0.0146561628,0.0891997516,0.0744852126,-0.4607025385,0.0579125769,0.1220576316,-0.1654013246,0.0956625491,0.0996743068,0.1712952554,-0.2308000922,0.0981799141,-0.1879240274,-0.1944777071,0.0011100378,-0.0574755073,0.4275349677,-0.0939213037,-0.0768165365,0.2017904669,-0.0709806234,-0.2179564983,0.3591507375,0.1810360551,-0.3327528834,-0.3817484379,0.0677004457,0.217260778,-0.0335039534,-0.0596167706,0.4418260753,-0.1036954299,-0.0293520316,0.1637021452,0.2213601023,0.1865794957,0.2992683649,-0.2348520607,-0.1026837528,0.000645592,0.0496760122,0.0035828855,0.4059077799,0.0496884063,-0.0829558,0.2037239373,0.0742981061,-0.0955299884,0.1310674995,0.2435009629,0.154842481,-0.2750955224,0.2033429444,-0.1607804149,0.1884043366,-0.1200278848,0.1289809495,-0.1276045442,0.0546827428,-0.0252689943,-0.1080456898,0.0213514082,0.1120825484,0.0471017919,-0.0440152362,0.6185708046,0.2233749479,0.2531940937,-0.3250675797,-0.0921133012,-0.6767205596,0.1313294321,0.100149475,-0.088599667,-0.140391022,0.2652161121,-0.2414724678,0.2453544885,-0.1460153013,0.2898319066,0.343424499,0.1123667732,-0.4177535176,-0.3659463823,-0.0639835596,-0.2085926533,0.0235484131,-0.494494319,-0.0506686345,-0.0019278878,-0.0757290646,-0.004220855,-0.4750475883,0.1863630861,0.2682702243,0.5125700235,0.0185879357,0.2594862878,0.2298403531,-0.1068245098,-0.2795262933,-0.0157521088,0.0940179601,-0.1190589219,0.1042476967,-0.0274683964,-0.1487519443,0.1516392976,-0.2929864824,0.5097359419,0.1991877705,0.3153412044,-0.1011395752,-0.0605889671,-0.1269253194,0.3645149767,0.0174745657,0.1931539029,-0.2062769979,0.2854265869,-0.3366298676,-0.198577404,0.4555933177,-0.1429150105,-0.2909006476,0.1750326753,0.2778065801,-0.1199425161,-0.3622121513,-0.5101880431,0.254742831,0.3145081103,-0.1905055344,-0.0828981176,0.2124556452,-0.1858970225,0.1646559834,-0.1686094403,0.4358358383,0.0952632427,0.0214769822,-0.0326855965,-0.0608769245]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"Facing the same issue.\r\nI am able to download datasets without `cache_dir`, however, when I specify the `cache_dir`, the process hangs indefinitely after partial download. \r\nTried for `data = load_dataset(\"cnn_dailymail\", \"3.0.0\")`","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":31,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nFacing the same issue.\r\nI am able to download datasets without `cache_dir`, however, when I specify the `cache_dir`, the process hangs indefinitely after partial download. \r\nTried for `data = load_dataset(\"cnn_dailymail\", \"3.0.0\")`","embeddings":[0.0042780945,0.2504116297,-0.0458030961,0.3203591704,0.174059093,0.2203368247,0.2745952606,0.0188670121,0.1395895034,-0.0835989192,-0.2411101013,0.2049974203,-0.0157355629,0.0229584035,-0.1130747646,0.0706264004,-0.0831784979,-0.012375203,0.214881748,0.1076202914,-0.1432205886,0.4405174851,-0.2540305853,-0.3808916807,-0.2399228662,0.0780592263,0.0524286777,0.1447728127,0.0600571446,-0.2211668193,0.6519641876,0.2406462729,0.0253143925,0.502795577,-0.0001188678,-0.1076321304,0.350839287,-0.0716238022,-0.7364218235,-0.3258126974,-0.3748835921,0.0870155171,0.109717764,0.0362841338,0.0253028739,0.408542335,-0.0688275471,-0.6611969471,0.0648548454,0.1356102377,0.1606152356,-0.1279445589,-0.3012706637,-0.0829707533,0.0818593428,-0.1440077573,-0.2933000326,0.4838713408,0.1657514572,0.0694014207,0.1291466802,0.2706472576,0.0797717944,0.4111890197,0.0540815964,0.0604026616,-0.3164115548,-0.359775126,0.4886941016,0.2121761292,0.8598021269,-0.2700321376,-0.3039605021,-0.1749399304,0.0525521673,-0.1236264482,0.290151149,0.1497558802,-0.3275518417,-0.0069505651,-0.1564629227,-0.0704139322,-0.0106756752,0.0054445537,-0.2916892171,-0.2443006337,-0.1538294554,-0.0073771616,-0.0015822977,0.0826448426,0.4729681313,-0.2602844834,-0.0807422027,0.2202804536,-0.5595542789,0.2158766687,0.000559893,0.1892478913,0.1013273299,0.1822555214,0.1477422714,0.2065310478,-0.3083066344,-0.087026976,0.2530344725,0.4764888585,-0.0430285707,-0.0697925165,0.283046782,0.2299452722,-0.0114323553,0.0710697472,-0.1208941266,0.0165324863,0.0011087955,-0.0767835081,0.1764149815,-0.2972975075,-0.3234760165,0.0074924449,0.0409277193,0.3652978241,0.0059952508,0.2722024024,-0.242434144,0.2520879805,-0.1457788795,0.013629158,-0.1482284367,0.0827375427,-0.0643170774,-0.1548316777,0.0538986698,0.0969780013,0.4505031407,-0.1375050992,0.1490903497,-0.1958867311,0.2819207311,-0.0487325191,0.0776979998,-0.4769000411,-0.0573143624,0.3119814396,0.1081625447,0.5107375979,-0.1896294802,-0.0414041132,-0.0710497648,0.2285241187,-0.1788782328,-0.263349086,0.0674281269,0.0627503768,-0.4993799925,-0.0282171965,-0.2152943164,0.1369773746,0.2798576057,0.0653764382,0.0029331285,-0.0338038728,-0.392280668,-0.2411888838,0.2537250519,0.6359077692,-0.3161544502,0.2139904648,-0.2323968112,0.1923343241,0.0928999335,0.2078724802,-0.0724156275,-0.0071817767,-0.4010176957,-0.0495647304,0.2674063146,-0.3931474686,-0.7098671794,0.3313200176,0.0407892279,0.2903725207,0.0077557415,0.0894190222,0.0465986431,-0.0027359091,0.1709143072,0.500800252,-0.0923290178,0.0558370352,-0.3301295936,-0.2972788215,0.0386329032,0.2795317769,-0.0485461131,0.1419136673,-0.1011564583,0.1120294109,0.2013295293,0.3345919847,0.1605563164,0.231125623,0.2642463148,-0.036222633,0.0119830258,-0.060543254,-0.8455967903,0.4726104736,0.0385342017,-0.1076812297,-0.297590822,0.0029712343,-0.422159344,-0.1490445584,-0.2707059979,-0.2411905378,-0.0184294619,0.1848862022,0.3131714165,0.1435331255,0.0976728126,0.6807877421,0.3323359191,-0.0728507414,-0.35404706,0.0430621877,0.1104802936,-0.0161763746,-0.0810256675,0.0414480492,0.0675494671,-0.1951831579,-0.2127973586,0.1842820942,0.1314559579,0.0822755545,0.0119864093,0.0230521932,0.2591760755,0.1992135793,0.1444220096,0.0551135689,0.2881848514,-0.1542217284,-0.1795338243,0.0172716696,-0.310582757,0.3332156539,-0.1253064424,0.0291084964,0.3228703141,-0.007037492,0.1806444973,-0.0623966157,0.3442015052,0.1316409558,0.4150342643,-0.1498939544,0.0717741624,0.2736897171,0.1089301705,0.1691252887,-0.052160766,0.169222787,-0.1906673163,-0.0682865903,0.0960837007,0.4361806512,0.5844067335,-0.0283588301,0.3492489159,-0.1001049206,0.2558563948,-0.2869028747,0.1058244258,0.1019388735,-0.1843168885,0.2019687444,0.2924491167,-0.0843551606,0.1706735343,-0.0542840846,0.221521154,0.0816672668,-0.0835195929,0.2589274347,-0.199834764,-0.1345647722,0.0375201106,-0.2513227463,0.1366075277,-0.4426152706,-0.1470975578,0.3263929188,0.2022882551,-0.026781654,-0.1754199564,0.2357199043,0.0906463861,-0.6479483247,-0.3592618704,-0.1837377101,-0.1817581654,0.0170165244,0.1409485936,-0.0831934884,0.1492794305,-0.0941857696,-0.1126950458,-0.2637678385,-0.0968256369,-0.0063732327,-0.0246003587,0.4230521023,0.1201583222,0.2072814405,-0.0478807241,0.2065472007,0.0156726111,0.0309749153,0.0067942669,-0.0120712314,0.0788661838,0.2587018013,-0.1997960657,-0.2379235774,-0.3741104901,-0.3802778125,0.2016575485,-0.0252621844,0.0028096561,0.4357881844,-0.0093410453,-0.068602711,0.3549776673,-0.2090492696,-0.3691793978,-0.3391586244,0.6691077352,-0.2349963933,-0.1425837725,0.0567806512,-0.0644834638,0.2819382548,0.2827350199,-0.5972637534,-0.0077203391,-0.5064714551,0.5216751099,-0.1362667531,-0.2705992162,0.1317769438,-0.2185649872,0.0977876112,0.0000025939,-0.1688382477,-0.0128682489,0.2656877637,0.2363733947,0.1391280144,0.0743553862,0.1826934665,0.571531117,0.1059387997,0.0792105943,0.2879718542,0.22063905,0.4312893152,-0.299883455,-0.0697754622,0.0152765382,-0.1323608756,-0.1477147341,0.1845574975,0.1168322042,0.0568664595,-0.1868288219,0.179595232,-0.0052232328,-0.1844436675,0.034964554,-0.4463770986,0.0876688883,0.1846206635,0.1202608272,0.0857125819,-0.3432097733,0.0198365021,0.1667864025,-0.0695132762,0.0387733392,0.4060423374,0.3434112072,-0.5011444092,0.1534785479,-0.0241144001,0.3109670877,-0.1532038301,0.1822483391,0.1199699715,0.0131131802,0.6287562251,-0.0832201391,0.2388140261,0.0373101607,-0.0618585907,-0.40424034,-0.0380957425,0.0854319185,0.3196202815,0.1477718204,0.0252171475,-0.2612608969,0.2810675204,-0.1347262561,0.2684253156,-0.1581935138,-0.2248172313,-0.2795633674,-0.3588253558,-0.1442420483,-0.1729663312,0.0967129692,0.0821413845,-0.0519323684,-0.1567452848,-0.0261831991,-0.1752167195,-0.1594083607,0.0351321511,0.1524333209,0.1793830991,0.5609030724,0.2166100889,-0.4731970429,0.1231369898,0.5334889293,-0.1100142002,-0.2519619465,0.126699701,-0.0722753257,-0.3700700998,0.3975606859,-0.2603934109,-0.0688087493,-0.1713931859,-0.0747619346,0.0283565931,0.2050452381,-0.1576775759,-0.1865011603,-0.1050529853,-0.4660967588,0.497192204,-0.0504607186,-0.2050777823,0.4447928071,-0.3642542362,-0.2078467458,0.1940716952,-0.2189795524,0.6659548879,-0.0548519194,0.0585260466,0.0749524385,-0.3400513232,0.471611619,-0.1628007591,0.0473022312,-0.1543657482,-0.3128213882,0.0367220603,-0.1653552353,0.025001036,-0.0089768758,-0.0966470391,0.3350319266,-0.390489161,0.154302761,0.0578945577,0.0460313559,-0.2451273948,-0.2718791366,-0.2216625214,0.1411841959,-0.1814484,0.2911992073,-0.100908421,0.034097679,-0.3563494682,0.1824293286,-0.3708055913,0.1940934509,-0.4626452625,0.0550779551,-0.1170081347,-0.5315765142,0.0289932974,0.1455118507,0.2777690291,-0.0075456328,-0.3407298028,0.4241767228,0.0544475056,0.0108051961,0.1159152687,-0.133043766,-0.2331975549,0.0014964623,0.1132599637,-0.0380919091,-0.2732329667,-0.1269539595,0.0510478169,0.129347235,0.1726423055,0.1227828041,-0.2189679593,-0.1275241375,0.2238183469,-0.1205579862,0.0730810389,0.1893675029,0.1084733084,0.2053567171,-0.2464368492,-0.2010922581,-0.0815198123,0.5874224901,-0.0695189387,-0.0102478312,0.466127187,-0.0028551838,-0.2535869777,-0.0019551299,-0.1758167595,-0.5710939765,-0.4229907691,0.2562327683,-0.1278041303,0.0750891641,-0.1116855592,0.1165079996,0.0880360603,-0.1511273831,0.0247770827,-0.5702730417,-0.1585041732,-0.3132454157,-0.2943415046,0.1866932064,-0.1497496665,-0.2294154167,0.1861995757,0.151118055,-0.1596882045,0.0921521559,-0.2148970366,-0.0692122579,0.1858190745,0.0222222693,0.3231983483,0.0430676937,-0.0717356727,0.2392369211,-0.1080570295,-0.1421725154,-0.1173829958,0.1702847481,0.0648732036,-0.1915672123,-0.091481775,-0.259032011,-0.0760588422,0.0086048217,0.3184524179,0.3038001657,-0.0199087951,-0.1090331376,-0.0729848295,0.2976176441,-0.401753664,0.5114492774,-0.2153646648,0.2653167248,0.234475866,0.1604985297,-0.0283346083,0.0851914212,-0.1821054369,-0.2493754923,-0.0763730109,-0.1426789314,0.1569521874,-0.2790204883,0.2373750508,0.1195063144,-0.0738399923,0.4100891352,-0.2310433537,-0.4143728316,0.1516560465,0.0842765421,-0.002360503,-0.0003500717,-0.2928047776,0.08541549,-0.0513891503,0.247413978,0.2849433422,-0.2616274059,0.2267559916,0.201951012,0.6356399059,-0.0344095938,-0.0013816202,0.0709957331,-0.136079073,0.1964983791,0.1004635766,0.070625104,-0.0739177316,0.6703199744,-0.1315604597,0.1332929432,0.0280555263,0.0371905789,-0.0613713525,-0.4371674061,0.0159355644,0.2454336882,-0.2146484107,0.1507847011,-0.0005333283,0.0983297601,-0.2034166306,0.07963074,-0.1947490871,-0.1547715068,0.0915198624,-0.0996360704,0.390889436,-0.118751429,-0.0895854235,0.18494986,-0.0739461109,-0.1946815848,0.452024579,0.1953646243,-0.2572306097,-0.3443217576,0.0304246955,0.1361244768,0.0041117687,-0.1579862982,0.357560128,0.0026053947,-0.0937052071,0.0748277009,0.1263249218,0.1466230154,0.2916063368,-0.2349766046,-0.0911232755,0.1312728226,0.0511716194,-0.0246798582,0.4464437068,0.1023718417,-0.0395441428,0.241896525,0.053006947,-0.1436316222,0.1022617221,0.2840794325,0.1139982641,-0.3254925013,0.2658189535,-0.1117463559,0.1561945528,-0.0179239213,0.0991589576,-0.1910628825,-0.0357492529,-0.0382312201,-0.1588800997,0.056013383,0.163610056,0.0332066864,-0.0116036115,0.642475009,0.1006132364,0.2702061236,-0.35883075,-0.139992848,-0.5773681998,0.0902809873,0.0509208217,-0.0261379201,-0.1565818191,0.2753821015,-0.195538342,0.1995488554,-0.2208450288,0.262683928,0.3423651457,0.1516291499,-0.4110071957,-0.3417984545,-0.0746529102,-0.1671658456,0.01595027,-0.5533842444,-0.0790738836,0.0727833658,-0.0447733775,0.0705998167,-0.5323393941,0.2670734823,0.3064929247,0.484817028,0.0064849858,0.2496199757,0.1884728074,-0.1957107186,-0.2607659101,-0.0291103013,0.1365922242,-0.1425150484,0.1535532922,-0.0711966455,-0.1974887401,0.1009490192,-0.2389451116,0.4061661065,0.1774387509,0.3550842404,-0.0169354212,-0.0751556382,-0.1135328412,0.3664140105,0.0257317368,0.2572016418,-0.1933922768,0.2369487733,-0.3427282572,-0.2245134562,0.4319783151,-0.1386662573,-0.3572745025,0.1980278492,0.3005154133,-0.1102453023,-0.2278943509,-0.4596554935,0.2537398636,0.2981156111,-0.1519326568,-0.1614125222,0.130281955,-0.191617161,0.1208890378,-0.2207437307,0.4610864818,0.1243968233,-0.0114325304,-0.0419233926,-0.0763099268]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"Hi @ashutoshml,\r\nI tried this and it worked for me:\r\n`data = load_dataset(\"cnn_dailymail\", \"3.0.0\", cache_dir=\".\/dummy\")`\r\n\r\nI'm using datasets==1.8.0. It took around 3-4 mins for dataset to unpack and start loading examples.","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":31,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nHi @ashutoshml,\r\nI tried this and it worked for me:\r\n`data = load_dataset(\"cnn_dailymail\", \"3.0.0\", cache_dir=\".\/dummy\")`\r\n\r\nI'm using datasets==1.8.0. It took around 3-4 mins for dataset to unpack and start loading examples.","embeddings":[-0.0161708575,0.2297847122,-0.0322397873,0.3434302211,0.1108229533,0.2220182866,0.3055529296,0.0372812152,0.10845723,-0.0921418965,-0.2221646756,0.2408726811,-0.0385525823,0.1162008494,-0.1133522615,0.0309860595,-0.1359189302,-0.0631945282,0.2353850603,0.1180652082,-0.1302544326,0.4372630119,-0.319630444,-0.3554959893,-0.2387880683,0.0237787813,0.0636806637,0.0957122222,0.0047329515,-0.2345732152,0.6385351419,0.2791397274,0.1181462184,0.5157641172,-0.0001169941,-0.215261206,0.3333494067,-0.0684954599,-0.6077932715,-0.2789219916,-0.3930041492,0.0457532331,0.0335161351,0.0610926859,-0.0509885475,0.4967774451,-0.1270866692,-0.68250072,0.0614344291,0.1803266108,0.1823161244,-0.0502500981,-0.171939522,-0.0725494921,0.0638539717,-0.0905586258,-0.268150121,0.4946600199,0.1551376581,0.0975516513,0.2529839873,0.331369698,0.0248196889,0.4136205614,0.1185360849,0.0480060801,-0.2777655423,-0.3346879184,0.4164537489,0.2303195447,0.8650121689,-0.2543489635,-0.3269252479,-0.0918510035,0.0841074139,-0.1494254023,0.2977922261,0.1900098473,-0.3102540672,-0.016460225,-0.1508080512,-0.0905476511,-0.0704411194,0.007946115,-0.2361735106,-0.2268900275,-0.1544542313,0.0180781335,0.019396238,-0.0038164356,0.456076026,-0.2130610645,-0.1223457232,0.3066767156,-0.65750283,0.1063818857,-0.0774628073,0.217916131,0.1605318338,0.1575927585,0.1439424455,0.1163089126,-0.3457915783,-0.0678958967,0.2134515345,0.5641543865,-0.0443470739,-0.0618443526,0.2127415836,0.2808054388,-0.0879347697,0.0965339541,-0.0884197056,-0.0386992842,0.0230213664,-0.056780152,0.2750973701,-0.354660511,-0.3671809733,0.0226325095,-0.0311452299,0.2224421203,0.0147286085,0.2650328875,-0.3251701891,0.2021937072,-0.0550872535,0.0201080665,-0.1727339774,-0.0391677469,-0.0853923783,-0.110216707,0.0527203865,0.1796500236,0.4412157536,-0.1889049858,0.0968768895,-0.205994308,0.1617501974,-0.0732570812,0.0391553342,-0.4348950386,-0.1178796068,0.225325644,0.0690706298,0.4374957085,-0.225779146,-0.1089572459,-0.1188760772,0.2386876792,-0.2894028127,-0.2850323915,-0.0544328168,0.0862685591,-0.4735613763,-0.1438089013,-0.2476587296,0.1591524482,0.23349154,0.0608700961,-0.0520559698,0.0331652872,-0.4323675931,-0.2124741673,0.2255028188,0.6246807575,-0.4017665386,0.2280920893,-0.2959651947,0.2312543094,0.1291996092,0.1825995743,-0.0365856327,0.0507646762,-0.4023651481,-0.066472508,0.2884465754,-0.3448900878,-0.7637765408,0.2402676046,-0.0072369417,0.2049037218,0.0244494379,0.1251177788,0.0759882927,-0.0541138202,0.1253979355,0.514295578,-0.1276306808,0.0129121635,-0.3770835102,-0.3079400361,0.1181101054,0.2464576662,-0.0579772703,0.1241133362,-0.039152056,0.0778740719,0.1895922422,0.3612080812,0.1256436855,0.2161608934,0.3133903444,-0.0379740372,0.056541618,0.0117612993,-0.8077424169,0.4555340707,-0.027855536,-0.0420433581,-0.2352332771,-0.0102761416,-0.5106800795,-0.1290507764,-0.2872656882,-0.2603490055,-0.0208017975,0.1453473717,0.340796113,0.1737203747,0.0690539032,0.6878806949,0.3122401834,-0.1150005162,-0.4065687358,0.0771846324,0.0412027799,-0.0303805787,-0.0158531889,0.0304849856,0.1075851917,-0.1671617627,-0.2501544356,0.1136160642,0.1644426733,0.0490662605,0.0501779951,0.013094632,0.2476552874,0.0753425658,0.2009117603,0.0724595711,0.2859911323,-0.1650985926,-0.2105876356,0.002755848,-0.3375557065,0.2851551771,-0.2091512531,0.0070916349,0.2657751441,0.0062512872,0.2497428805,-0.0618720166,0.294426322,0.1491237879,0.3727253079,-0.1142761484,0.0861810446,0.3334768713,0.1487441063,0.1839720011,-0.0698969737,0.1951130182,-0.1243645325,-0.1425950825,0.1091438308,0.4017100036,0.5841244459,-0.0246835425,0.2626039684,-0.0332573317,0.2991403937,-0.2309496254,0.0736985207,0.0413556881,-0.1135909557,0.1331713796,0.2872654498,-0.015570784,0.1911846995,-0.1201843098,0.1731726527,0.2000922263,-0.0830413252,0.2018013746,-0.1967657208,-0.1593468338,-0.0198100135,-0.2326493114,0.0836899877,-0.3867376149,-0.1762474328,0.3189530075,0.2630099356,0.0070880526,-0.1054377481,0.2337634414,0.1556800455,-0.5326211452,-0.3184814155,-0.1940667778,-0.2282195836,0.0227809176,0.1533803791,-0.0598246381,0.2215584666,-0.1137655601,-0.0823396742,-0.2419231832,-0.152606979,-0.0322496705,0.0191484652,0.4340485036,0.0852257311,0.2672777176,-0.0650587976,0.2055649906,0.0122110434,0.0226376634,-0.0593248978,0.0130007574,-0.0156129673,0.217313841,-0.0895089135,-0.3338562548,-0.3593615294,-0.3331201077,0.1772400141,0.0283102542,0.0318437628,0.4842742383,0.0203257129,-0.0261697695,0.3580590189,-0.1635661125,-0.3964845836,-0.3687490225,0.6486810446,-0.2061018348,-0.151373595,0.1052231714,-0.0556890629,0.3705451787,0.1249329001,-0.585927546,-0.0531539619,-0.4565603733,0.4831457138,-0.1305647343,-0.2152395695,0.1574800164,-0.2882137597,0.1117240041,0.0590288602,-0.1880036443,-0.0513585582,0.251760602,0.28122437,0.1736371666,0.164906323,0.1558852047,0.4896213114,0.2271452844,0.0765749663,0.3385466337,0.1747327149,0.3446930051,-0.26328969,-0.1565910876,0.038693402,-0.0647285581,-0.1784442365,0.1594429612,0.0945797563,0.1281877458,-0.1914579868,0.1282638907,0.0179177728,-0.1522355527,-0.0236677118,-0.4125364721,0.0673777536,0.2339377552,0.1071430519,0.0859969929,-0.4340619743,0.0650222301,0.1726836562,-0.1214197502,0.0862312913,0.4207447767,0.2988087237,-0.5537292361,0.2220349461,-0.0466593057,0.2986119986,-0.1947982609,0.2443314046,0.2043375522,0.0654062554,0.5053154826,-0.0991744027,0.2398491055,0.0395721272,-0.0184289142,-0.3319123089,-0.0879440382,0.0857420191,0.2489284873,0.0268401094,0.0984614268,-0.2554751635,0.2832726836,0.0093980217,0.2727679312,-0.1783520579,-0.2880887985,-0.254629612,-0.3815025389,-0.2035190165,-0.1745370775,-0.0006424907,0.2138524503,-0.0135615813,-0.1656665355,0.0001876962,-0.1296426207,-0.1350315362,0.0774730816,0.1817211658,0.2171587646,0.5387612581,0.1875288785,-0.5295698643,0.0161454417,0.6190474629,-0.0002219721,-0.2811990678,0.0352763906,-0.0614152253,-0.331024915,0.2810474038,-0.2723965645,-0.0786228776,-0.148221612,-0.0570588633,0.0151261557,0.167252332,-0.1680762917,-0.1296951324,-0.1499880701,-0.4540210068,0.5328175426,-0.0014154706,-0.1512827277,0.4374180436,-0.337777406,-0.1537493467,0.1955570877,-0.1587496996,0.7808175683,-0.1520689279,-0.0213917922,0.1761063486,-0.2561190724,0.4869804084,-0.1592504829,0.0299089309,-0.2471296936,-0.2300305963,0.0666506365,-0.123602137,0.0069490271,-0.0075517711,-0.1988683492,0.3442739248,-0.3252334893,0.1087375879,0.0285194237,0.099735558,-0.173204571,-0.2128087133,-0.2275398672,0.1706978828,-0.2817318439,0.3777637184,-0.1735588014,0.0559374467,-0.3191872835,0.141797781,-0.4367451072,0.1777481139,-0.4636301398,0.0767847151,-0.1152112633,-0.4964627624,-0.0009552438,0.2023342699,0.2688336372,0.0195267648,-0.3288549483,0.4005285203,0.0492472127,-0.0533989035,0.1303325742,-0.1444353163,-0.165199995,-0.0500952415,0.0372057073,0.0066389688,-0.1685594171,-0.0726506189,-0.0072567128,0.1448432058,0.1887490749,0.1230240315,-0.1996091008,-0.1614856124,0.2216001302,-0.0751433969,0.0868100151,0.2447747886,0.0833371133,0.2246736437,-0.1345857382,-0.2351902276,-0.0945636854,0.5425859094,-0.0418153666,-0.0157618634,0.4603332579,0.1284795403,-0.2134692371,-0.0289991442,-0.1296995878,-0.6000588536,-0.4761018157,0.288210392,-0.0929749385,0.1675271094,-0.0723382011,0.2882305086,0.0734288022,-0.2019598484,0.1117247641,-0.5971550941,-0.1664825082,-0.3297134638,-0.3448544443,0.1408458948,-0.1521685421,-0.2515463829,0.1187152714,0.2488016486,-0.1690047085,0.0734422505,-0.1707433313,-0.1087722033,0.1898916811,0.079047747,0.2814862132,0.0398601927,-0.0455877557,0.2012580931,-0.0795357451,-0.1799441874,-0.1656100452,0.1628962457,0.0664254427,-0.1888504028,-0.054277584,-0.2952574492,-0.1321925372,-0.000810941,0.2849336863,0.3268962204,-0.0081289299,-0.0827440247,-0.0454863533,0.126787886,-0.3804497421,0.493010968,-0.1351150721,0.3272512853,0.2206288129,0.1308666468,0.0543920957,0.0594297163,-0.1953427494,-0.1989383996,-0.0607008785,-0.1933353692,0.1201185361,-0.3254982531,0.1937719584,0.0853085369,-0.1041855291,0.3221391737,-0.2265310138,-0.3556316197,0.0959326252,0.0901494473,-0.0654294863,-0.0049341065,-0.1978555918,0.1507649273,-0.1145720109,0.2622528374,0.2281636745,-0.1705436409,0.2230267674,0.2137537897,0.6045224071,-0.0438825823,0.0442142561,0.0157341715,-0.1103893146,0.1768483073,0.1116627976,0.055496756,-0.0949215889,0.4914776385,-0.1949403882,0.117182143,0.0167324506,0.1446588635,-0.0071431375,-0.4608992934,0.0759697706,0.1625014395,-0.219024539,0.1113846824,0.0828363001,0.1521740258,-0.1638479084,0.1271162182,-0.2203707695,-0.167126745,0.036259979,-0.0248119086,0.4233195186,-0.0978647992,-0.1179638728,0.2062977999,-0.0396516547,-0.2228570282,0.3689396083,0.2101279795,-0.3132458329,-0.398396343,0.0771701857,0.2319147587,0.0192524064,-0.1468176097,0.4307014048,-0.074193731,-0.0568372458,0.1226646528,0.1887847781,0.1796087027,0.2943630219,-0.2133621275,-0.0261738282,0.0397639051,0.0103664128,0.0414894596,0.462603569,0.0618491657,-0.0755709931,0.188543573,0.0601768196,-0.101666078,0.081050612,0.2556931078,0.1381815821,-0.2983431518,0.2092870027,-0.1891839802,0.2116075903,-0.1481431872,0.112528123,-0.1796244085,-0.0210958123,-0.0209635049,-0.1094904542,0.0349578634,0.088953279,0.0428268798,-0.0672253743,0.6162350774,0.1745605916,0.2748784125,-0.3566158116,-0.0978997573,-0.6595902443,0.1062238142,0.0629369691,-0.0669710562,-0.1566729993,0.2737516165,-0.2397699207,0.2131851465,-0.1346968859,0.3115025461,0.3181229532,0.164714992,-0.4332875609,-0.3533026874,-0.0477129444,-0.1604931056,0.0171137601,-0.4949307442,-0.0348050818,0.0197237153,-0.0829360485,-0.0402799398,-0.4827023447,0.2306899726,0.3190772235,0.5184900165,0.001926869,0.2857753038,0.2384925187,-0.1143461242,-0.2785441577,0.0299347248,0.1139574647,-0.1097860262,0.1091727316,-0.0705160722,-0.1354418248,0.056626156,-0.2698991299,0.4596205354,0.1217449009,0.2896387875,-0.0564037599,-0.0923894048,-0.1420256048,0.4115003943,-0.0225745253,0.1825933456,-0.1708189398,0.3060689867,-0.3263866901,-0.226297155,0.448361069,-0.0971601531,-0.3082883358,0.2209600955,0.2452325374,-0.1234275177,-0.3260552883,-0.5284139514,0.3083971739,0.3418225646,-0.1708075404,-0.0704443753,0.2244172841,-0.1444666386,0.1790290624,-0.1673061997,0.4373211563,0.0691850409,0.0663147271,-0.0562921055,-0.0439465195]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1615","title":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`","comments":"Ok. I waited for 20-30 mins, and it still is stuck.\r\nI am using datasets==1.8.0.\r\n\r\nIs there anyway to check what is happening? like a` --verbose` flag?\r\n\r\n![Screenshot 2021-06-25 at 6 37 43 PM](https:\/\/user-images.githubusercontent.com\/2375919\/123429653-cdfb7280-d5e4-11eb-9fa7-ff295800cc86.png)\r\n","body":"Hello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n","comment_length":34,"text":"Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`\nHello,\r\nI'm having issue downloading TriviaQA dataset with `load_dataset`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \".\/datasets\")\r\n```\r\n\r\n## The output:\r\n1. Download begins:\r\n```\r\nDownloading and preparing dataset trivia_qa\/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to \/cs\/labs\/gabis\/sapirweissbuch\/tr\r\nivia_qa\/rc\/1.1.0\/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 17%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2589 | 446M\/2.67G [00:37<04:45, 7.77MB\/s]\r\n```\r\n2. 100% is reached\r\n3. It got stuck here for about an hour, and added additional 30G of data to \".\/datasets\" directory. I killed the process eventually.\r\n\r\nA similar issue can be observed in Google Colab:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing\r\n\r\n## Expected behaviour:\r\nThe dataset \"TriviaQA\" should be successfully downloaded.\r\n\nOk. I waited for 20-30 mins, and it still is stuck.\r\nI am using datasets==1.8.0.\r\n\r\nIs there anyway to check what is happening? like a` --verbose` flag?\r\n\r\n![Screenshot 2021-06-25 at 6 37 43 PM](https:\/\/user-images.githubusercontent.com\/2375919\/123429653-cdfb7280-d5e4-11eb-9fa7-ff295800cc86.png)\r\n","embeddings":[-0.0052458039,0.1308883131,-0.0646162108,0.3441570997,0.0718581229,0.210291639,0.17244111,0.1310920417,0.1391692013,-0.0710647553,-0.1699336171,0.231231302,-0.0192254949,0.0066203936,-0.1691107601,0.2085393071,-0.1640118062,-0.0868696421,0.2699991167,0.0834092572,-0.038244091,0.4004627764,-0.2224357128,-0.2988501787,-0.1695676297,-0.0254348591,0.0705305934,0.0405973941,0.0147904633,-0.3218882978,0.6169593334,0.3028764427,0.009877827,0.521084547,-0.0001201023,-0.1916544735,0.4338795245,0.0159098525,-0.7354334593,-0.2126133889,-0.411706686,0.0140906833,0.0877757818,0.0253632683,0.0474618189,0.4686435461,-0.0080202678,-0.680444181,0.0924841166,0.0909667015,0.1437713504,-0.0238936823,-0.2291480601,-0.0375164934,0.0284584425,0.0096831163,-0.2325122207,0.4369439483,0.2176711112,0.0913062915,0.2389549613,0.252281189,-0.0492004231,0.3928609788,-0.0071388446,0.0374720767,-0.1860148758,-0.3192746043,0.4707184732,0.3660711348,0.8030222058,-0.1876770109,-0.3021990061,-0.1832547039,0.1051823273,-0.0451491773,0.413316071,0.1583818048,-0.336879909,-0.0022909432,-0.1966359764,-0.0813699439,0.0054737031,-0.0495399833,-0.2693914771,-0.261174351,-0.1570450515,-0.0208111741,0.0786024854,0.0757898912,0.331874609,-0.2209260613,-0.1263227463,0.3372460604,-0.6835750341,0.1361633986,0.072355479,0.2225887775,0.1222133413,0.1436494887,0.2464012951,0.1499664932,-0.3367521167,-0.0158412661,0.2920416594,0.4770641625,0.0216687676,-0.1471895725,0.2946625352,0.2768577933,-0.0165046696,0.1851952672,-0.1194929481,-0.1138872355,0.1638812125,-0.1195146367,0.3472663462,-0.420897752,-0.4582388997,0.0693760514,-0.1012671068,0.2368401289,0.0996522456,0.2810336053,-0.3059815466,0.1489553154,-0.0852394775,-0.0406089798,-0.176011309,0.0588091128,-0.0030846773,-0.1624728441,0.0150796827,0.1091506332,0.4115763903,-0.2335549593,-0.0193372164,-0.1226827353,0.1655345261,-0.1065096632,0.040022023,-0.4161472917,-0.0327929854,0.3291257918,0.0912063345,0.4472205341,-0.1924474239,-0.1869480759,-0.1487813592,0.1830757111,-0.1687164009,-0.1786273271,-0.0225004498,0.0721458644,-0.5353237391,-0.1041093096,-0.2382825613,0.1389842033,0.1975210309,0.093503885,0.0546241254,-0.0165530462,-0.3146316707,-0.2055378258,0.2628719211,0.7534264922,-0.4006408155,0.2732146978,-0.3736734986,0.197200492,0.0929380357,0.2091330439,0.0395619236,0.1090814918,-0.4702808559,-0.0468017869,0.2117950022,-0.3955647647,-0.5958545804,0.3525334001,-0.0282600336,0.2908311486,0.0996687785,0.0646447912,-0.0339887924,-0.0744108632,0.1817499697,0.3799595833,-0.0818123296,-0.030382704,-0.3953439593,-0.2925280333,0.0339508913,0.2108902037,-0.095669277,-0.0097001679,-0.0184888877,0.0432756208,0.1957745552,0.384322077,0.038434092,0.1697260439,0.3557586074,0.0017097529,-0.0794348642,-0.0463985428,-0.8227305412,0.4676510394,0.0079893582,-0.1329741031,-0.286468029,-0.1050078943,-0.5263269544,0.0180345941,-0.3432105482,-0.1726970226,-0.0191411693,0.2503947318,0.3605091274,0.255169481,0.0629131943,0.6656355262,0.2315622419,-0.1129195243,-0.3988396227,0.0818557218,0.176239863,-0.0596291497,0.0037791485,0.0058503971,0.0960794389,-0.2225623876,-0.1633274108,0.1464251727,0.1269235909,0.1480647177,0.0653381199,0.0636171326,0.2154255956,0.1169756204,0.0703701675,0.0574338362,0.2074352652,-0.1087787524,-0.2665083706,-0.0270735361,-0.3863425255,0.315282613,-0.1466709822,-0.034661144,0.2687528729,-0.0348880775,0.1267798692,-0.0633670911,0.4413505495,0.2584002614,0.3805348575,-0.053692665,0.040112108,0.2626843154,0.2711616158,0.118067421,-0.012823306,0.1971620172,-0.005204225,-0.0706369504,0.177028209,0.3518933654,0.4828983545,-0.0416873321,0.2759874761,-0.0799083486,0.1331025064,-0.1560465991,0.0334436744,0.1082303226,-0.1268747151,0.1741719693,0.309795171,-0.0264489166,0.0924227014,-0.1649825573,0.1287864298,0.1788528413,-0.1485517621,0.198604241,-0.2390971035,-0.0132886302,-0.0422030911,-0.2888913751,0.1038679853,-0.493181169,-0.1106578112,0.2955605984,0.283478111,0.0019748365,-0.0721168146,0.2240708172,0.0732660294,-0.5757681131,-0.4541899264,-0.1719468981,-0.2614944279,-0.0203753747,0.19481574,-0.0408428162,0.250739783,-0.0865891799,-0.10126701,-0.200471431,-0.2505178452,0.0091868388,0.0438208655,0.481872797,0.0933851972,0.289922148,0.007487243,0.2269599438,-0.0196238793,0.0244032089,-0.0774511546,-0.0027823437,0.0303157717,0.2225931287,-0.1063723639,-0.2903220952,-0.3319783509,-0.4047456682,0.2436876744,-0.0214492511,0.0139278583,0.4712770879,0.0456215478,0.0825299621,0.2933071554,-0.143687427,-0.3843383491,-0.3096329272,0.6168893576,-0.1876232028,-0.0910336152,0.0997093245,-0.0327564068,0.2840221226,0.1509723514,-0.6188516021,0.0646927357,-0.6073617935,0.5238555074,-0.1325420737,-0.2632633746,0.213538304,-0.1811937243,0.1117006019,0.0977270678,-0.1261819601,-0.0444909185,0.2625546753,0.3519196212,0.0681914166,0.1534047872,0.183074683,0.494686693,0.1345431656,0.022358112,0.2706026435,0.1346401423,0.4930537939,-0.2187382281,-0.1133547127,-0.0532007627,-0.1161283106,-0.2943062782,0.1397247016,0.0782917589,0.1330399066,-0.1606243849,0.2421140671,0.042788282,-0.2041894346,-0.023750959,-0.5263646841,0.0355342366,0.1687311083,0.1239354759,0.124277845,-0.33073318,-0.0145646557,0.2276058644,-0.1181540564,-0.0128104053,0.3142594099,0.3351431191,-0.4688970149,0.2364880443,-0.0593995973,0.3144456744,-0.3311344683,0.1470077038,0.2424875498,-0.0506540686,0.6517907381,-0.0769302696,0.3268947005,0.0154485935,0.0725687444,-0.3916998804,-0.1152863353,0.0467255451,0.2104929984,0.1444844604,0.0007258535,-0.3484277129,0.347165972,-0.0237742793,0.1957208514,-0.1668449193,-0.2916873693,-0.3935706317,-0.4005663097,-0.2025601268,-0.0671448782,0.0672191232,0.0531281382,-0.0583435036,-0.2078257054,-0.0089258896,-0.1617198288,-0.072663188,0.1248307675,0.0469690338,0.2352912426,0.5087922812,0.2496747524,-0.4768317342,-0.0204205438,0.534622848,0.0526247993,-0.2157335579,0.0872021168,-0.0667998269,-0.2258852124,0.3088751137,-0.266990602,-0.0345387459,-0.1484565288,0.0098614106,-0.0521905124,0.1935735643,-0.0570386462,-0.036355745,-0.2147771716,-0.4271722436,0.6526007652,0.067254059,-0.2232859731,0.4791546464,-0.343891263,-0.1222775728,0.1828718185,-0.214143604,0.8169724941,-0.1977908015,-0.1324986964,0.1182645336,-0.3915683329,0.5272780061,-0.2375798821,-0.0293316189,-0.0916936472,-0.1518983245,0.0798478201,-0.1431253105,0.0844403729,-0.0487608165,-0.2028613091,0.287643373,-0.4040241241,0.0848370492,0.0553879254,0.0853973776,-0.2655195892,-0.2070168108,-0.39656353,0.1311640888,-0.2325733304,0.481934309,-0.1615771204,0.0359020159,-0.2869976461,0.2073231786,-0.5213598013,0.0980341285,-0.3350685239,0.0059338594,-0.1224386692,-0.5067163706,-0.0475260504,0.2077906132,0.1990365684,0.1179095283,-0.3257424235,0.489261657,0.0707578361,-0.0793528631,0.1297968179,-0.1865677387,-0.1506747454,0.0341408215,0.0379216112,0.0501757972,-0.1818911135,0.0431218036,-0.0111010894,0.132974878,0.1897034049,0.1139445379,-0.1433441192,-0.1560094357,0.1866177768,-0.0310062263,0.0623791516,0.1810165644,0.061332982,0.2440014482,-0.1498767138,-0.2352419496,-0.1162512526,0.5939406157,-0.228521198,0.0415870063,0.4236773551,0.1007682681,-0.1361696422,-0.0026853941,-0.1311409622,-0.5875869989,-0.4679343998,0.2783744335,-0.0468705446,0.0869078785,-0.0268221591,0.1812011153,0.136582002,-0.2843505144,0.05238593,-0.5536264181,-0.1574616283,-0.3288672864,-0.3142514229,0.1687850952,-0.1865963638,-0.2416708618,0.1827360988,0.1942668557,-0.1549937427,-0.0939509496,-0.38877666,-0.1536055058,0.2354278415,0.0443239138,0.2058865726,-0.0806987733,-0.0613724887,0.2169472277,-0.0262628719,-0.1529832929,-0.1057881117,0.180307135,-0.0461629629,0.0125949681,-0.0243286919,-0.2785234153,0.0204307772,0.0062243212,0.2195919752,0.2657715678,-0.0172807425,-0.171486944,-0.1798634082,0.2139878422,-0.3819269836,0.453910768,-0.1530853212,0.2628614604,0.2321184725,0.1231578663,-0.0785785019,0.0676737055,-0.1806049794,-0.2075751275,-0.0240934808,-0.167075783,0.0803436488,-0.2063756138,0.2021932304,-0.0706200451,-0.0909149274,0.4165438712,-0.1716995835,-0.4082164168,0.046172373,0.0658009648,-0.0336195938,0.0806748569,-0.2799673378,0.1415288001,-0.0916928425,0.2812607288,0.1598957479,-0.1819912046,0.3058791757,0.2701832056,0.6407557726,-0.124585107,-0.0139955096,0.1008381248,-0.1573378146,0.1835170537,0.176716432,-0.0525839515,-0.0714948848,0.4804334342,-0.1396228075,0.1702184379,0.0308630969,0.2109649628,-0.024375286,-0.3796406686,0.0421280637,0.322737515,-0.1420819908,0.2416046858,0.1274545938,0.0618408993,-0.1851192117,0.143048197,-0.2687399983,-0.1166612655,0.1116792485,0.0122858919,0.3113307953,-0.0601462275,-0.1574994177,0.1922563016,-0.002223314,-0.2239465117,0.2905063927,0.2160765529,-0.2472400665,-0.3749713004,0.0771255195,0.1127564535,0.1163311377,-0.1847728491,0.3728837669,-0.019561328,-0.0630163997,0.113641113,0.2130895406,0.3010067344,0.3170288503,-0.2346718609,0.0336291045,0.0587065034,0.0358819515,-0.0091549028,0.5393313766,-0.0781274214,-0.0539051034,0.1863538176,0.07834059,-0.1479100436,0.2179923803,0.1138939932,0.111622788,-0.3968809247,0.2204314768,-0.0829083994,0.1531684846,-0.0872341245,0.1031033769,-0.1507106274,-0.0423172042,-0.0470859036,-0.1635811329,-0.0052166223,0.0656392574,0.0323734358,-0.053288687,0.5978507996,0.2575446963,0.2881396711,-0.3217953742,-0.0932294503,-0.7962046266,0.008507343,0.0062572984,-0.1606842279,-0.0932786837,0.289486289,-0.2070467919,0.2126640081,-0.0544079132,0.2795097828,0.243060708,0.1278769821,-0.3115137219,-0.3779143691,-0.0557407849,-0.2249976546,-0.0575228781,-0.5158137083,-0.0359895453,0.0204638932,-0.1113478094,-0.0172113571,-0.4779676497,0.219913885,0.3332752883,0.4278501868,0.0355083384,0.3024101853,0.2180191725,-0.1101591662,-0.3326716125,0.0565736666,0.1805068254,-0.0433879606,-0.0057403389,-0.1517131925,-0.1478455067,0.1530557275,-0.2317729741,0.4001648128,0.2216174006,0.3796208203,-0.020435052,-0.2293595821,-0.1397007108,0.3396216929,0.0430633575,0.2814917266,-0.1215864271,0.3195244074,-0.3485707343,-0.2198424488,0.3764307201,-0.1246417239,-0.297565937,0.2786642611,0.3085492849,-0.107005626,-0.264531821,-0.4146088064,0.2370793372,0.3796964884,-0.180357337,-0.0862861276,0.177196458,-0.1125818416,0.1785973907,-0.1542252153,0.4451508224,0.0631568581,0.1238295808,0.0102812191,-0.0419790745]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1611","title":"shuffle with torch generator ","comments":"Is there a way one can convert the two generator? not sure overall what alternatives I could have to shuffle the datasets with a torch generator, thanks ","body":"Hi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq ","comment_length":27,"text":"shuffle with torch generator \nHi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq \nIs there a way one can convert the two generator? not sure overall what alternatives I could have to shuffle the datasets with a torch generator, thanks ","embeddings":[-0.4141784906,-0.2276341915,-0.0309860893,0.4363026023,0.2791832387,0.043277476,0.0350645818,0.0862027258,-0.0705147088,0.5191633105,-0.2391806394,0.4969671667,-0.330867976,-0.3823140264,-0.2315443754,-0.3067396283,0.2168588638,-0.0674591884,-0.3701868951,0.0901022404,-0.1378070265,-0.2120033205,-0.1668649018,0.0467480533,-0.1056925207,0.0289946701,0.0033921432,-0.0965117216,-0.3189224899,0.134345144,-0.193623811,0.3798199892,0.1345464438,0.5104945898,-0.0001205206,-0.0924573243,0.2343459725,-0.0886435509,-0.1473664045,-0.3304494917,0.2770422995,0.3916114271,-0.497562021,0.0803651139,-0.1436500996,-0.134876579,0.1823467016,-0.4246356487,0.025369674,0.0088243959,0.146384567,-0.1882594973,0.2212270498,-0.1267284453,0.4813113511,0.4027077854,0.0825443491,-0.0884001628,0.5316643119,0.2585778236,-0.222730726,0.0401541106,-0.1297207475,0.0570802614,0.3467617631,-0.377928108,-0.5593923926,-0.1208927184,0.2527587712,0.3894735873,0.2606964707,-0.0788615718,-0.0548112579,-0.0975281075,-0.0268234983,0.1572411656,0.0690156966,0.1864593923,-0.3888832033,-0.0558707416,-0.3143303394,-0.2865081429,0.059359964,-0.1246102825,0.1501453221,0.5047636628,0.0944000855,0.1592345089,0.429700762,-0.2380346209,0.7994137406,-0.1298983991,-0.0251097046,0.0777098238,-0.2114275247,-0.0427819937,-0.2100488842,0.481218636,0.4092452824,-0.1242537498,0.0871970356,0.1063172743,0.0312605128,0.1096846312,-0.0036943706,-0.1530969739,-0.0645572767,0.4121882021,0.1983651668,-0.1908563226,-0.0955185816,0.0059050391,-0.2408456802,-0.6150102615,-0.2558302581,0.2141474485,-0.3566376269,-0.2235177308,-0.0417441837,-0.4753188193,0.0988784209,-0.2133395076,0.1529147327,0.2406569719,0.0333356075,0.2746611834,-0.010903758,0.1564359665,0.1632135063,-0.4084785581,0.0813561752,-0.0123460209,-0.2426604182,0.1039214507,-0.0549301319,-0.3516777754,0.1047479734,0.0525300242,0.1870106012,0.5445389152,0.0593385696,-0.2262383848,0.0869415328,-0.3754952252,0.2271413803,-0.0004926358,-0.1230563819,0.4934233725,-0.0944036543,0.387047708,-0.1626808643,-0.1309407353,0.1281172335,0.0729740635,-0.125826627,0.0293600895,-0.1989664733,0.0780624002,0.1250072271,-0.4125432372,-0.0224302672,-0.1792347878,-0.1869376898,-0.2180637121,-0.0891267136,0.00144904,-0.0504480861,0.0199258886,-0.0462862141,-0.0942552388,0.3332948387,0.3242069781,-0.1496121436,0.0513040163,-0.2401927561,-0.5925034881,0.3951981068,-0.0906561837,-0.2823819518,-0.1787954867,-0.0289292913,-0.1212859526,0.3703610897,0.1865266711,0.1186726913,-0.2664369941,0.5015241504,0.3267721534,-0.2621350288,0.1053579301,-0.2406796068,-0.1173427925,0.5927738547,0.37823686,0.2078772634,0.0389382839,-0.0008431331,-0.0676842183,0.0872129425,-0.1589268297,0.0897461548,0.0231049769,-0.2916110754,0.1719158292,-0.0735472143,-0.0347446315,0.2200122029,-0.0979226232,0.0656043962,-0.0383214243,0.217228815,-0.1234984323,0.2181232274,-0.3128986657,-0.0517180413,-0.0244372617,0.0106973164,0.1245209947,0.3151629865,-0.3106861413,0.0900882557,-0.1076287404,-0.1986149848,-0.2256064266,-0.6309060454,0.0666367114,-0.1373061985,-0.0613078028,0.1682926416,0.1063309014,-0.0261792112,0.0117537547,-0.0639230907,0.0541573465,-0.1843316704,0.0759909078,0.1398366094,-0.0266890153,-0.0728359818,-0.0398983844,0.3686387837,0.0236871112,0.1820199341,-0.0388193764,0.0369778313,0.3304706514,-0.4579139054,0.1511119604,0.0395483151,-0.1713802665,0.0006632492,0.1531132609,0.0429793149,0.4353911877,0.4335139692,0.1709699929,0.2457942516,0.1246993169,-0.5354149938,-0.032986138,0.3636902869,0.072829254,-0.1253970712,0.1292566806,-0.2272519022,0.0214630552,0.001182365,-0.1308523864,0.4655730426,0.2820750475,0.245953396,-0.0134935612,0.3454642892,-0.0944776908,0.0651640743,-0.2085628062,-0.0908692777,0.1638874859,0.1917981058,-0.2428211719,-0.05599989,-0.3100757003,0.0884887576,0.3063648045,-0.0932318121,-0.0870061219,-0.2038206458,-0.1799857467,-0.4440018237,-0.0827579796,-0.093373701,-0.1509070396,0.0896420032,0.6144540906,0.069410421,0.0637290999,0.5797442794,0.0669962987,-0.0739754364,0.0863265544,-0.1775232404,-0.3066989481,-0.1291801035,0.1254413575,0.4116496742,-0.1509331018,0.2816527486,-0.3073143065,-0.1129855588,-0.4432057142,0.3672915697,0.0439493507,0.0977441669,-0.2580509186,0.2157942504,-0.172171101,0.1490639597,-0.249466151,0.2054423392,-0.4076107442,0.066296339,0.1641972959,-0.3200441897,0.0756505132,0.1044751704,-0.0764465332,-0.3199960887,-0.3747588396,-0.1709561348,-0.0763250887,0.1783693284,-0.3522421122,0.2795055509,-0.3748940825,-0.035457965,-0.1900365055,-0.1652492285,-0.8801335692,0.5325649977,-0.1929665953,-0.2212184072,-0.2686172426,-0.0485665351,0.3006251156,0.7538975477,0.2688275874,0.3970760107,-0.0204230472,0.1205137298,-0.1282154918,0.3746427,0.1638052315,0.0094232224,0.0517136157,0.0818366483,0.3039545715,-0.1065535918,-0.3358479738,0.6419594288,0.4148103595,0.1426938921,-0.0218239632,0.8628772497,-0.2597814202,-0.3316825628,-0.0852450505,0.0241199769,0.1927410811,-0.1950429976,-0.1512049139,0.5222127438,0.1137923971,-0.2648955286,0.0093717612,-0.2193320692,0.1323524266,-0.2966076434,0.0547251888,0.0778380707,-0.1388813555,0.3595013916,-0.1910057515,-0.0436083339,-0.1585786492,0.1617463231,-0.3032835722,-0.2201804072,0.0962135941,0.0943289846,0.3186903,0.0489130951,-0.2407928258,-0.1534567624,-0.3318790793,0.320088923,0.1681140512,0.0793098509,0.0895788446,0.0772297233,0.0609393455,0.0380619764,0.5132349133,-0.3216819167,-0.0405177996,0.0383862406,0.1502503902,-0.2556785941,0.0548255518,-0.0441342033,-0.0220030304,-0.1036843732,-0.1374330819,-0.1989562511,-0.0475106239,-0.261544317,0.3407230675,-0.0951274782,-0.1046417579,-0.2002945691,-0.0085903527,-0.2690851092,-0.1879465878,-0.0314430445,-0.0943555906,-0.436457485,0.0207448415,-0.1020894051,0.2858409286,0.3050039709,0.0867652893,0.182099402,-0.2733401358,-0.1042109877,0.2747334838,0.0136288479,0.4649119973,-0.0637806281,-0.0677898303,0.0011765476,0.1711737812,-0.1105522737,0.0236908942,0.065423578,0.0274464004,0.197584793,-0.1957109869,0.0513343811,-0.1650893092,-0.1151413992,-0.1891136914,0.0799325109,-0.4065696001,-0.8603869677,-0.0436613932,0.5672543645,-0.0909672678,0.5497948527,-0.5639747977,-0.1765224487,0.1159188673,0.1881151497,0.9552465677,-0.2742542922,0.031555064,-0.1030604541,0.0200800914,0.3085851669,-0.2204601318,0.2421175689,-0.4099314213,-0.1911153942,-0.008422072,-0.1529792249,0.0274688229,0.3680694401,-0.1963088363,0.483823061,0.04898718,-0.1099390015,-0.0645064041,0.1021210551,-0.5011495948,-0.1054381207,0.0654442832,0.0676890761,-0.2397897393,0.3317898214,-0.0004017874,-0.013044999,-0.2659433484,0.0709797591,-0.0599929467,-0.0080447942,0.0041974899,-0.1237485334,-0.0390883461,-0.203925699,0.3492183089,0.2434874922,0.2807121575,0.0620373338,0.0845598504,0.0678337142,0.1642637551,0.0732620656,-0.0731094629,0.0714492798,-0.0255446807,0.0269059613,-0.0850747749,0.2939510942,0.2857492268,-0.422847569,-0.6244634986,-0.2806549966,0.5599281788,-0.1390550733,0.1827697009,0.1398071051,0.0464449935,-0.0374701321,0.0336063579,0.1038947478,-0.0967346132,0.2264552116,-0.1312228292,0.1111344919,0.097185567,0.1996379495,0.0059636254,0.0504396483,-0.0069548571,0.1060209125,-0.1473900676,-0.0759517401,0.0564645827,0.2092258632,-0.1446640193,0.019859178,-0.1947276592,-0.0836814567,-0.0051459442,0.1247399673,-0.0820257813,0.1713254303,0.0929279849,-0.1992405355,-0.4890235662,-0.0062794052,0.0420524999,0.0166612994,0.1151491925,-0.1389395297,-0.2639862895,-0.0036174376,-0.2014114112,0.0498905815,0.3590525985,0.3101041019,-0.1562046111,-0.0584852397,-0.6464868188,-0.0773523003,0.0138306068,0.0583698973,-0.0764324069,-0.0466555506,0.0781551301,0.1411399692,0.1059897915,0.0242953766,-0.0589825436,-0.3023313582,0.0000440315,-0.1626368463,0.2617577314,0.5220800638,-0.0685373098,-0.0004951603,0.3048141003,0.0445070118,-0.2492462248,0.29278934,-0.0068648886,-0.1153539345,-0.0106467381,0.1660510451,0.3458784819,-0.2346951067,-0.1977744699,0.1273856014,0.4180144668,-0.0411198772,0.207416594,0.0483072437,0.3283166885,0.3353272676,0.0286517795,0.4495698214,-0.0541584119,-0.0828978792,0.4580724835,0.0879213512,-0.0568051562,-0.0126572773,-0.3755943775,0.0665516779,0.0292026699,0.0846221298,-0.2136851996,-0.3802140951,-0.2528567314,-0.1441884786,0.2995219231,-0.1484571993,0.039308928,0.3494199216,0.2734491229,-0.129717797,0.2299493849,-0.1882284731,-0.017375553,0.3861589432,0.1005264297,0.8498154283,0.2162032127,-0.0834516659,0.4184253812,-0.1986812949,0.0932203159,0.5848050117,0.1946249455,0.0003466784,0.3048976958,0.306298852,-0.0073704855,0.0857690498,0.2058199793,0.2283825576,-0.0321625508,-0.0414981283,-0.267683953,-0.3867055476,-0.2157637775,0.2375889719,-0.029111078,0.1414607167,0.4236831367,0.1156170666,0.0444019474,-0.2688699365,0.4230864346,0.3364492655,0.0513530858,-0.1241722479,-0.2339475006,0.1084054112,-0.0956396759,0.6494945288,0.1055609584,0.2005015314,0.3770092428,-0.2455316782,-0.01501182,-0.3588199019,-0.0848660767,-0.0581516623,0.2315097153,0.1129073128,-0.1762229949,0.2547445893,0.0632023513,-0.0935541913,-0.1136514619,0.0774555877,-0.0238624122,0.0948963016,0.0694527924,0.0061010113,-0.0824831799,-0.0016696027,-0.0006492323,-0.3499497175,-0.0622027442,-0.2825419307,-0.091355212,0.0080647431,-0.1870278865,0.0182276852,0.2150028348,0.1014336422,0.1864675432,-0.0531871058,-0.0050464203,-0.5153775215,-0.4280236661,0.2614643872,0.1175278351,0.2586258948,-0.5599026084,-0.0521414615,0.2529878914,0.2502704561,-0.1713238657,0.3089397848,0.0850189105,0.8079195619,-0.0663371235,-0.1776545793,-0.0787273645,0.011566163,0.0503055602,-0.3968661129,0.0147081036,0.0894795284,0.0464646257,-0.0948257819,0.1943320781,-0.0527126603,-0.1016651616,-0.0094184875,0.1006199941,0.4464061856,0.0209837556,0.2246930897,0.0280237384,0.2385798097,-0.2178929597,0.0807863921,0.231175229,0.0845921338,0.1365789175,-0.1305394918,-0.369074285,0.3513293564,-0.1636218876,0.3346005678,-0.5669949651,0.1291413605,-0.0576452278,0.2332994044,-0.0486172959,0.0324713401,0.1134440079,-0.1191108823,-0.4796673656,0.2342282832,0.3663766086,-0.1289347112,-0.3619128764,-0.5091979504,-0.1815521717,0.3375093639,-0.0819151178,-0.268293947,-0.0149021102,0.0216666516,0.0662198737,0.6247010231,-0.0031188878,0.0953419805,0.0813566446,-0.1854307353,0.1069899276,0.0714304969,-0.0352780856,-0.1530918926,-0.1130390093]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1611","title":"shuffle with torch generator ","comments":"@lhoestq let me please expalin in more details, maybe you could help me suggesting an alternative to solve the issue for now, I have multiple large datasets using huggingface library, then I need to define a distributed sampler on top of it, for this I need to shard the datasets and give each shard to each core, but before sharding I need to shuffle the dataset, if you are familiar with distributed sampler in pytorch, this needs to be done based on seed+epoch generator to make it consistent across the cores they do it through defining a torch generator, I was wondering if you could tell me how I can shuffle the data for now, I am unfortunately blocked by this and have a limited time left, and I greatly appreciate your help on this. thanks ","body":"Hi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq ","comment_length":136,"text":"shuffle with torch generator \nHi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq \n@lhoestq let me please expalin in more details, maybe you could help me suggesting an alternative to solve the issue for now, I have multiple large datasets using huggingface library, then I need to define a distributed sampler on top of it, for this I need to shard the datasets and give each shard to each core, but before sharding I need to shuffle the dataset, if you are familiar with distributed sampler in pytorch, this needs to be done based on seed+epoch generator to make it consistent across the cores they do it through defining a torch generator, I was wondering if you could tell me how I can shuffle the data for now, I am unfortunately blocked by this and have a limited time left, and I greatly appreciate your help on this. thanks ","embeddings":[-0.345851928,-0.4519608021,-0.0053851502,0.3233452737,0.2008843571,-0.1706957966,0.1138170585,0.0025900884,0.0121341171,0.5316617489,-0.1594806314,0.4323351681,-0.3134092093,-0.2165889293,-0.2019849569,-0.3771593869,0.107542105,-0.0871212408,-0.2455119044,0.0093543027,-0.0153144468,-0.2353334427,-0.186999172,-0.0573527105,-0.3854927421,-0.059340138,-0.0702758729,0.0212741178,-0.2236671746,0.0718621314,-0.2819514275,0.4691458344,0.2383426726,0.4779495001,-0.0001151688,-0.1170631275,0.2073725462,-0.0082180416,-0.1322351694,-0.2457087636,0.5509521365,0.43428424,-0.3845115006,-0.0679859817,-0.1916062832,-0.068793714,0.1744905263,-0.2716235518,0.0208925493,-0.1335559636,0.1496651471,0.035266649,0.104959622,-0.0843408182,0.4105360508,0.3694914281,0.1240980029,-0.0748827383,0.6229237318,0.134675324,-0.4861242175,0.0889942795,-0.081676051,0.2442923784,0.3888377547,-0.3940316141,-0.7607227564,-0.2191768438,0.1398545355,0.3866349757,0.139064461,-0.2253740281,-0.1394059658,-0.1977352947,-0.1394049227,-0.0148732699,0.1322558522,0.2197691649,-0.3946428895,-0.0954117328,-0.2467139363,-0.1916704029,-0.0495113395,-0.0948894471,0.2863081694,0.5201362371,-0.0334931873,0.1497374475,0.3512330055,-0.0548648499,0.4070957303,-0.1119518876,0.0795881748,0.14323771,-0.3439023793,0.0141528668,-0.1334604025,0.5297748446,0.4111104012,0.0513974167,0.0837284029,0.2188688368,0.2293460518,0.11625579,0.0340936594,-0.0582391247,-0.1435583234,0.3440670967,0.3032718003,-0.0598213226,-0.0261203386,0.0129351998,-0.1129943207,-0.4922784269,-0.1642145663,0.2016288638,-0.3297792077,-0.0942341685,0.1179650873,-0.4185841382,0.2435651273,-0.2392206043,0.2254231721,0.3411636353,-0.0781826451,0.1544607878,-0.0698721781,0.0171650685,-0.0401932411,-0.3330687582,-0.0707998276,-0.0575136319,-0.2132867575,0.1547156274,0.1180803329,-0.3479045928,0.1632682681,0.0476759225,0.4263076782,0.3616333902,-0.0101900781,-0.0751890317,0.1006940231,-0.2142037153,0.0173566006,-0.0203871354,-0.0840773955,0.5804338455,-0.0719616115,0.3543978333,-0.239458859,-0.3152093887,0.1301570386,0.0954240412,-0.289603889,0.0723858327,-0.1350549906,-0.0199878812,0.1318057775,-0.2996225655,0.0490128584,-0.0807092786,-0.2355407774,-0.1869644821,0.0455323011,-0.0539968386,0.2070120573,-0.182871446,-0.0343383439,-0.0598994046,0.3174870908,0.2542971373,-0.1248109415,0.0288087521,-0.1999742389,-0.3596103191,0.2687471509,-0.0409890451,-0.3759897649,-0.2268588394,-0.0936549827,-0.0587956831,0.3629342318,0.252592653,0.3305433393,-0.1851219684,0.282153964,0.3671049774,-0.1621299088,0.1097866446,-0.3579723239,0.0516373254,0.3286982477,0.601562202,0.0406840183,-0.0156480968,0.0674102604,-0.1858720332,0.1879710704,-0.0775896162,0.2044013888,0.144687742,-0.1698748022,0.2384008616,0.0738100484,-0.1249532402,-0.0200469866,0.0654329211,0.0921319872,0.0891097561,0.1101812422,-0.1243180409,0.342443049,-0.2894906402,-0.163928017,-0.1577590108,-0.0244304482,0.075328894,0.3603484333,-0.2597909272,-0.0181979295,-0.1288227886,-0.251880765,-0.1574526131,-0.6665961146,0.1126433536,-0.1576205939,-0.1328803152,0.028896952,0.1954203993,-0.1411717385,0.0061728014,-0.1035900414,0.1642970741,-0.1483186483,-0.1238739416,-0.0707352608,0.0913665891,0.0446030684,-0.0432604663,0.2310305536,-0.0208403952,0.0873576179,-0.0920630544,0.206313774,0.3520578742,-0.5542064309,0.0431295298,-0.0027142284,-0.3050481677,0.0146015929,0.2018087655,-0.0797998533,0.4644012451,0.7137275338,0.0023212756,0.2510870099,0.0740201473,-0.6002058983,-0.0779328868,0.1259373873,-0.0623260587,-0.0496206135,0.040826302,-0.1125167236,0.0960838646,0.0660910159,-0.1044833809,0.363651067,0.2306364179,0.1789245754,-0.0428005271,0.3528853059,-0.1432031095,-0.1455960423,-0.1294303238,-0.1495319903,0.2132592797,0.2005718201,-0.1272635758,-0.0802867636,-0.3866424859,0.2035643607,0.319191426,-0.0650558248,-0.0954838321,-0.0337795913,-0.2052563429,-0.4084914029,-0.0787668377,-0.3317169547,0.060965199,0.131063208,0.6428599358,0.0693952292,0.2565937042,0.6699770689,0.3664533198,-0.0731634647,0.0235493518,-0.1242146567,-0.262521714,-0.031487368,0.092764616,0.2420937419,-0.0858208761,0.3420304954,-0.1620462835,-0.1218222007,-0.3247768581,0.2326404154,0.0363441743,-0.0060370862,-0.2195035964,0.1252288818,-0.1376129985,0.0825688913,-0.1996157467,0.2960529029,-0.5989969373,-0.0129783917,0.308318764,-0.1864526719,0.0051311762,0.0418456271,-0.045492772,-0.3525609076,-0.445735544,0.0362416878,-0.1430865973,0.1574224085,-0.171244204,0.1446669251,-0.4235257506,-0.1788658053,-0.0983342752,-0.1551620364,-0.943046093,0.35212484,-0.2260384262,-0.2373831719,-0.3247357607,-0.1194115058,0.1464328021,0.6899400949,0.1432884187,0.166545406,-0.078536123,0.1501694173,-0.0889793336,0.4013924897,0.2770030499,-0.0338909067,0.1004374474,0.0633189678,0.1788790822,-0.0051381639,-0.2224503309,0.4549425244,0.3356425762,0.0940947086,-0.0164455641,0.9399464726,-0.2725069225,-0.3625983596,0.0532111079,0.1038637608,0.138283819,-0.0798450634,-0.2272517979,0.603448391,0.0597777329,-0.3100404739,0.0303451773,-0.2174509466,0.1743360013,-0.2652322948,-0.0129558388,0.0935492814,-0.1643845588,0.3257824183,-0.017114222,0.1035677418,-0.2892692983,0.0285932682,-0.2621687651,-0.3024075031,0.0492133088,0.1850677431,0.3610880077,0.0330644399,-0.2837132812,-0.1188779697,-0.3910366297,0.2486174256,0.2084727883,0.0118035302,0.0517613068,0.0321484357,0.1518404186,0.1040157825,0.5262333155,-0.0668495148,0.0267591644,-0.0700907931,-0.1551864594,-0.2295379937,0.0387763493,0.0071792058,0.0054755402,-0.1355658621,-0.0498802364,-0.3229092956,-0.092971243,-0.1937994361,0.2434325814,-0.178122893,-0.2946726084,-0.3549219966,-0.0209314302,-0.3090896308,-0.0975425988,0.0878840834,-0.0623149686,-0.3908461928,0.2824152112,0.0873687491,0.3011774421,0.0681506395,0.0609089918,0.2478549033,-0.2877689898,-0.0115335248,0.4761492014,0.0819202289,0.3673387766,0.077598758,-0.2444377244,-0.0897748321,0.1344442517,-0.0287955329,0.0469174422,0.0651500821,0.112304166,0.2911318243,-0.0937491506,0.1156600788,-0.36795187,-0.0949881002,-0.2512465119,0.074259229,-0.3805732429,-0.7667902112,-0.0251205601,0.4275448918,-0.0719509199,0.6165530086,-0.4705485404,-0.0007681603,0.2282354832,0.0578255504,0.9757539034,-0.2650170624,0.1749456227,-0.0817330554,0.2150145173,0.2316199839,-0.1121115088,0.3865836859,-0.4202061296,-0.3344765604,0.0655876696,-0.1213555932,-0.0136234723,0.3198573291,-0.1518234313,0.3902928829,0.1975786239,-0.2236306518,-0.188392207,0.2172449231,-0.4041140676,-0.2321407199,0.223590225,0.0526877418,-0.3155839741,0.4675931036,-0.0234459005,0.0612374395,-0.0555636883,0.0253859889,-0.0596048087,-0.1049631462,-0.0072984272,-0.1076095924,0.1160128936,-0.2225757092,0.2116524875,0.1921595782,0.1540047079,0.0478674434,0.1716740876,0.1495461166,0.1490918845,-0.062526606,-0.0437728278,0.1130809113,0.0944623947,-0.0324051566,-0.2577424049,0.1130004227,0.2747409046,-0.5833414197,-0.6232048869,-0.1929500997,0.4472710788,-0.0037890915,0.2241951674,0.056864474,-0.0005588444,-0.0667751878,0.0481132716,0.2962720692,-0.2568406761,0.2525820136,-0.1374906898,0.1244170889,-0.0520478897,0.3734005988,-0.1756168753,0.1279884428,0.047285784,0.2227570117,-0.1582924575,-0.0809741244,-0.1138635129,0.1484412253,-0.2834390104,0.0025406803,-0.1708136201,-0.072135821,-0.0848249644,0.3120902181,0.0033092019,0.1405466646,0.0977611989,-0.2794311643,-0.4836830497,-0.0000736054,0.1074065343,0.0881596059,0.1918315887,0.0200110804,-0.1853941828,0.2085170299,-0.2603884339,0.020750111,0.2845562398,0.2545334697,-0.0673075691,-0.0597851537,-0.379696697,-0.0725220516,-0.0394894145,0.0963218659,-0.0515610427,-0.0860726088,0.0954792649,0.1055665836,0.1000622362,0.016032353,-0.2372811288,-0.3166929483,0.0201249421,-0.147283718,0.2543066144,0.5206654668,-0.0236139186,0.1284666061,0.4752301276,-0.0080879796,-0.1643273681,0.3528121114,-0.0160429832,-0.1884899139,0.0344028249,-0.0217603501,0.2198740244,-0.1818753481,-0.1424254477,0.0736819357,0.4221869409,-0.0828216821,0.1516513824,-0.0835954994,0.2910152078,0.1936585605,0.3331764638,0.3068487942,-0.0069001662,-0.1370609403,0.5387890339,0.1367722899,-0.0360357054,0.0125110811,-0.2074843198,0.2590881288,0.2254396081,0.2310744077,-0.2801939547,-0.3183999956,-0.3892592192,-0.0399671718,0.3736358881,-0.1069794148,-0.0237322673,0.0826502815,0.2065705657,-0.1519041359,0.354624331,-0.0035373289,0.0229499079,0.1869133562,0.1136195809,0.8747284412,0.1322331131,-0.1067267284,0.2627243698,-0.4496672451,0.1241895109,0.2922258973,0.152738452,-0.015194281,0.0863655284,0.4797901213,-0.0959049463,-0.0657423288,0.1199486405,0.2468841672,-0.0611760579,0.0521727018,-0.1522693038,-0.2944077849,-0.0605370104,0.142175898,0.0676258802,0.1978852302,0.6124059558,0.0396468788,-0.1269851178,-0.3056103885,0.1819963604,0.3696608245,0.1412160993,-0.0132030332,-0.1103711277,0.1692485511,-0.144636631,0.6867967844,0.0824758634,0.1232614294,0.2310813665,-0.3189182878,0.1211749539,-0.5023207664,-0.0131494245,-0.0883576348,0.3387950361,0.1338323951,-0.220248729,0.2695659995,0.0262322798,-0.1303270608,-0.049400609,0.2014977932,0.0908792168,0.1367034316,0.1747650653,0.0276177041,-0.1055993959,0.0116057331,0.0025131823,-0.4752290547,0.0908925086,-0.1595612317,-0.1860006303,-0.0748368502,-0.2754307687,0.0322028063,0.2151003331,0.2799641192,0.3108185828,-0.0046652323,-0.0596654043,-0.4469118714,-0.4370977879,0.3125188053,-0.0649749041,0.1066955104,-0.3467547297,-0.0941027105,0.1747890413,0.1493840218,-0.2726558745,0.2127731591,-0.0749748349,0.720463872,-0.2018407583,-0.0875979438,-0.2993690073,0.0995970666,0.1795666963,-0.3586974442,0.0004146701,0.1934260279,0.0342141055,0.0003761825,0.1864843965,-0.1078130081,-0.2186216265,-0.0687923282,0.0262937248,0.3666085005,0.0433549732,0.1332029253,-0.0128260711,0.231478557,-0.3193357289,-0.0119051244,0.1782919914,0.1290525794,0.1647719592,-0.2314359248,-0.5010430813,0.435654372,-0.2152931988,0.114751339,-0.4243612289,0.1917989105,0.0247903336,0.2560216188,0.0127110099,0.2756125033,0.0281508788,-0.0167041309,-0.397084713,0.1323711574,0.5565843582,-0.3701780736,-0.2819609642,-0.5279883146,-0.0292996205,0.1957954168,-0.1203954145,-0.4141183197,0.0774533376,-0.0736492947,0.1590770334,0.4474384189,0.0008261672,0.0332030132,0.035031192,-0.1430095434,0.1949490011,0.0375350416,-0.0205244441,-0.0015497194,-0.0931260064]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1611","title":"shuffle with torch generator ","comments":"@lhoestq Is there a way I could shuffle the datasets from this library with a custom defined shuffle function? thanks for your help on this. ","body":"Hi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq ","comment_length":25,"text":"shuffle with torch generator \nHi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq \n@lhoestq Is there a way I could shuffle the datasets from this library with a custom defined shuffle function? thanks for your help on this. ","embeddings":[-0.3035784662,-0.1541758925,-0.0428936779,0.3373784721,0.1244964898,-0.0405245647,0.1495991498,0.0698686391,-0.0988659188,0.4983230233,-0.1478007436,0.5367980599,-0.3674835861,-0.3628564477,-0.1735696793,-0.2025685161,0.1165713891,-0.0829895064,-0.229901731,0.0604962781,-0.1584002376,-0.2253881246,-0.2725287974,0.0495431572,-0.114835389,0.0130128255,0.0354387164,-0.0336142704,-0.2428324223,0.1144076362,-0.0953740999,0.5520756841,0.1718186736,0.4014338851,-0.0001167589,-0.2259283215,0.2070662528,-0.1696791649,-0.1488472521,-0.251193434,0.4779121876,0.4454227686,-0.3977607787,0.0311327539,-0.1117493659,-0.1513141394,0.1459348649,-0.5109872222,-0.069283247,0.0557293184,0.1596289426,-0.2149686664,0.1080581322,-0.0395502113,0.6648828983,0.466170162,0.0026009066,-0.0728148744,0.6146603823,0.0956551582,-0.2529292405,-0.034004312,-0.0446134284,0.179815501,0.4139145315,-0.326271981,-0.7432286143,-0.1033257619,0.1801627874,0.4821199477,0.3446732461,-0.3318504691,-0.0814986676,-0.1926306039,0.0067854966,0.0644697472,0.1991120726,0.1914750934,-0.2767150998,-0.1970661134,-0.224372223,-0.1730788648,0.0746031702,-0.058240477,0.3356627822,0.6138373613,0.0105088428,0.1713776588,0.4257211685,-0.1807716191,0.7319177985,-0.0452859439,-0.0550397895,0.311068207,-0.2374525368,-0.0331255384,-0.1561368704,0.3568582535,0.432932049,-0.0203721374,0.1030462384,0.23465617,0.0394732431,0.1086723953,0.1352599114,-0.1630287319,-0.1442575157,0.3600514829,0.3103067875,-0.1091387272,-0.0970998481,-0.0809088945,-0.1866468936,-0.3205394149,-0.3309747577,0.1076548174,-0.4172235131,-0.096133098,0.0702721253,-0.4154550731,0.1184190735,-0.2412600219,0.1686436087,0.2605020404,-0.0206941124,0.1192469522,-0.0961938351,-0.0250684973,0.0824843571,-0.2402795553,0.0107150944,-0.0049069598,-0.3077452481,0.0775055364,0.108058162,-0.3181445301,0.1922660917,-0.0376173742,0.3117220402,0.3771187961,0.0379488878,-0.1347333342,0.1117306501,-0.2570710778,0.1197576299,0.0064496049,-0.1585110277,0.4980469942,-0.2637287974,0.3558024764,-0.2256187946,-0.1845002174,0.0244657416,0.0772549734,-0.2556601167,0.0563771836,-0.1895473748,0.0211307611,0.0347858071,-0.4529851079,0.0363938808,-0.1487643719,-0.187309742,-0.2074573785,-0.1261648238,-0.1203788519,-0.0316562206,-0.038546443,-0.1719409376,0.0178977661,0.316730082,0.1443868726,-0.1533988416,0.1072942615,-0.3012363613,-0.4549753666,0.4599805772,0.0481922701,-0.306215018,-0.2992991507,0.0671848208,-0.0260197781,0.2625827491,0.2745354176,0.2286716551,-0.2559615076,0.2982192934,0.3607936203,-0.1393992156,0.0964509249,-0.1335204095,-0.081289351,0.4252729118,0.5648540258,0.1300880015,-0.053502515,0.1655643731,0.0240847562,0.123947233,-0.0519501455,0.08254949,0.009626586,-0.2304044664,0.3558782339,0.036301218,-0.1382434666,0.1211545542,-0.0442458093,0.1756918579,-0.086056754,0.2649950385,-0.1033013165,0.2996015549,-0.3137801886,-0.0838229954,-0.1393998265,0.0080720494,0.243737936,0.2525041103,-0.2178008407,-0.0947144777,-0.0906122178,-0.3220292926,-0.2203038335,-0.670090735,0.0559604578,-0.0062423358,-0.1308740377,0.1150240749,0.1816488653,-0.1719783396,0.0419495888,-0.1544873565,0.1902092099,-0.0864718333,-0.1496556699,0.0263892952,0.0747575834,-0.061107371,-0.0418951586,0.2515585423,0.0435165651,0.1604134291,-0.0506227203,0.0799464583,0.4821775854,-0.5742489696,0.0972163454,-0.0065116836,-0.3149254024,0.0320318304,0.2417598218,0.0035514485,0.4444433451,0.5479028225,0.1385601163,0.2927290499,-0.0133888163,-0.4744552374,-0.2436847389,0.3090004921,0.108270444,-0.0405207686,0.1083001122,-0.1526260525,0.1104533151,-0.0019484891,-0.09884011,0.4094269574,0.2262386084,0.0735641122,-0.0030621758,0.361911118,-0.1118232235,-0.1624463201,-0.1876970232,-0.2690888047,0.1019109786,0.1199697405,-0.1798645258,-0.1780310422,-0.3670764863,0.0940080062,0.3032658696,-0.0573142692,-0.0924269632,-0.1232123449,-0.1683565527,-0.2929193377,-0.1665654927,-0.2037306279,-0.0161374379,0.0990225598,0.5631673932,0.0536621362,0.1482138783,0.6211823225,0.2635795474,-0.0518165417,-0.1397404373,-0.0813622847,-0.2025576979,-0.1179957017,0.1520366669,0.2128129303,-0.0477299653,0.3596486151,-0.1595647037,-0.0359830782,-0.3572947383,0.1892105639,0.0093765985,0.0859096125,-0.2573554814,0.3506405056,-0.3062865436,0.1170390397,-0.2053722143,0.1324309111,-0.6321336031,0.1780333221,0.1444718242,-0.1617561877,0.1821977049,-0.0648906156,-0.0845097378,-0.2052391022,-0.4204808772,-0.012672158,0.0059512216,0.1796849668,-0.3157528341,0.1628570408,-0.2896281183,-0.1375176609,-0.213557139,-0.2477021068,-0.9783408642,0.4515042007,-0.2405683696,-0.1577394456,-0.1128492281,-0.0633360669,0.2025256455,0.7796785831,0.1309731007,0.3953173161,0.0678880885,0.0575833619,-0.1564130783,0.4933382273,0.1891448647,-0.1289324015,0.1232082397,0.0086534992,0.0994754061,-0.0338101797,-0.208881706,0.6188428402,0.4272312224,0.1347936392,-0.0998386592,0.7730320096,-0.4162921309,-0.3284847438,0.0762144625,-0.0208371095,0.3319323659,-0.1206751391,-0.0334844775,0.5191320181,0.2483401746,-0.2873217762,0.0842626095,-0.1239076704,0.0359829515,-0.2930499017,0.063214466,0.1309969574,-0.1277483553,0.3286906481,-0.1418532431,-0.0763857067,-0.2606885731,0.149863556,-0.268756032,-0.2925872505,0.0051216274,0.0267195869,0.4023666978,-0.0203187242,-0.1828022301,-0.2186228335,-0.2036701739,0.1599788964,0.1347583234,0.0530838445,0.0940866247,0.1123060435,0.1004628539,0.1352196485,0.3906803429,-0.2722322047,-0.0659665018,0.0562625006,0.0407925025,-0.2542857528,0.0097185923,0.0409327261,-0.058035925,-0.1821449995,-0.2019442171,-0.1584849358,0.047980722,-0.231315285,0.2062178105,-0.1143954843,-0.261292249,-0.2440740466,-0.1170788035,-0.3092528582,-0.1473234594,-0.0473459661,-0.1787184477,-0.2922314703,0.101461269,0.0012467179,0.1970525235,0.1632234752,-0.0167886447,0.2373551428,-0.1206472218,-0.1505354643,0.3263992369,-0.0997742042,0.3666453063,-0.0150788883,-0.0371746309,0.0433073379,0.1564738303,-0.1848741174,-0.006305329,0.0173953138,0.0224342365,0.2908084393,-0.3455904722,0.007311183,-0.1409434676,-0.0249327105,-0.1763724089,-0.1474721283,-0.3159902394,-1.0068508387,0.0434527248,0.4618212283,-0.0273278058,0.7483633757,-0.61966151,-0.1764824092,0.1454708874,0.2752572,0.8909563422,-0.260507077,-0.01411492,-0.1914790273,0.0850102082,0.3317534626,-0.3643650711,0.151826635,-0.3686317801,-0.1354744136,0.0021885016,-0.0910619795,0.0369184501,0.2798548937,-0.1181955263,0.4254440963,0.1225406155,-0.0409282632,-0.1367304474,0.1231365353,-0.4476029873,-0.2032689899,0.0812265798,0.0647803545,-0.1984678507,0.323350966,0.027360389,0.0293345768,-0.0888989121,0.0090712365,-0.2012014538,-0.1212560013,0.0371648148,-0.2789334059,-0.0391377807,-0.0707089379,0.3410106301,0.2221719176,0.2066981643,0.2115163952,0.0631891936,0.2110839337,0.1262263507,0.0196329467,-0.1309486032,0.2468429953,-0.0190643184,-0.037335217,-0.076184839,0.3246568739,0.295907408,-0.5829527378,-0.6144095659,-0.1921777427,0.5890190005,0.0232761484,0.2478560507,-0.0332033262,0.0249769781,-0.004700406,0.0635930374,0.2283790857,-0.2038342208,0.2793621719,-0.244149372,0.1313410401,-0.0981325135,0.272023052,0.0156825874,0.0714938864,-0.0281190239,0.1495898962,-0.0055015925,-0.0735473633,-0.0477740131,0.1613066047,-0.0796335116,-0.0481188819,-0.1457571238,-0.1371886581,0.0558167659,0.2236286849,-0.1674131155,0.2410680354,0.0528243631,-0.2295515239,-0.3880849779,-0.0995869264,0.0591049753,0.0893568397,0.1823503822,-0.1040366665,-0.1720720977,-0.0243583415,-0.2216076255,0.0458721034,0.3218085766,0.2535408735,-0.1767077744,0.2660332322,-0.5071166754,0.0088510606,-0.024115691,0.2084841877,-0.0804460123,-0.0580192395,0.0701223835,0.1233227775,0.1140223891,0.1510863006,-0.2069822103,-0.3900058568,-0.1493141204,-0.1796876192,0.1897665709,0.3922872543,-0.0636844337,0.1426587552,0.4087819755,0.0329274349,-0.1814809889,0.3096798956,-0.1086526513,-0.1174209192,0.0616277196,0.0041311756,0.1979899853,-0.1851326972,-0.2097730339,0.0540555343,0.2748886347,-0.0556197762,0.2277856469,-0.0168296546,0.3121976554,0.3089565039,0.1517380625,0.5174764991,-0.0071756113,-0.1955630034,0.5407363772,0.0763467699,-0.017942166,0.0213461332,-0.3296198249,0.3085441589,0.0836119503,0.163182199,-0.1892545372,-0.3940080404,-0.3025415838,-0.1744622588,0.367448926,-0.1503718346,0.0802067891,0.3846774101,0.1008530557,-0.1109745279,0.2033129036,-0.0933495685,-0.0091554765,0.5316179991,0.1259895861,0.7887239456,0.2126030028,-0.1383998543,0.2269643992,-0.3732977509,0.1743650436,0.5442454219,0.2306514978,0.100784488,0.1758381426,0.3211457431,0.0741185322,0.1789502949,0.1295501143,0.0976886898,0.0289660599,0.1193645597,-0.1495663822,-0.3216155767,-0.1720866114,0.1685184985,-0.0052108564,0.1426666826,0.4082621932,0.0039067739,0.0199951679,-0.3250437677,0.4161902666,0.1790066212,0.1097269282,-0.0725976527,-0.2148453593,0.0006911955,-0.0770651922,0.721503377,-0.1062089354,0.0172639769,0.3384530246,-0.2357688099,-0.0183827523,-0.4908436239,-0.0298218522,-0.126352042,0.3056735098,0.1938974261,-0.1186314896,0.269404918,0.0555923544,-0.1257508695,-0.0872212574,0.1030608788,0.1090540886,0.0505762473,0.0969417319,0.0275772233,-0.0019593118,-0.0081817722,-0.0185193364,-0.2531485558,-0.1131973565,-0.2802993059,-0.0125279743,-0.0606909171,-0.1967709959,0.0197615102,0.1342768669,0.1909321398,0.1829273105,-0.0272442512,0.1152950153,-0.5936115384,-0.4396766126,0.2917504013,0.0376804508,0.1616150588,-0.28668648,-0.0221072715,0.2116844058,0.1467067003,-0.2590263188,0.244322896,0.0968485624,0.7042540312,-0.1514727175,-0.2618570328,-0.3209061921,0.1366693676,0.0931891352,-0.3202073872,-0.0915446877,0.0136399129,-0.0026882412,-0.0724902377,0.1163150147,-0.1075307503,-0.0412446447,0.1209606007,-0.0739696324,0.316601783,0.0193904042,0.271443665,-0.1365826875,0.309722811,-0.2087182999,-0.0733660683,0.1528330445,0.0813795924,0.1304224133,-0.0680079609,-0.4727525711,0.3995134532,-0.1341679096,0.4174215198,-0.4970863163,0.1263099462,-0.0073547154,0.235527426,-0.0987217575,0.1173049957,0.0108465096,-0.1050689295,-0.3378245234,0.2594768405,0.4067914784,-0.2430846393,-0.4009336829,-0.476026386,-0.0760572627,0.2441125363,-0.1144973114,-0.4248805046,0.1785087138,0.0727160498,0.2316798866,0.5036876202,-0.0080118924,-0.0069168177,0.1653009504,-0.0975839719,0.0311260726,0.1335442513,0.0665061921,0.0065384856,-0.1202148274]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1611","title":"shuffle with torch generator ","comments":"Right now the shuffle method only accepts the `seed` (optional int) or `generator` (optional `np.random.Generator`) parameters.\r\n\r\nHere is a suggestion to shuffle the data using your own shuffle method using `select`.\r\n`select` can be used to re-order the dataset samples or simply pick a few ones if you want.\r\nIt's what is used under the hood when you call `dataset.shuffle`.\r\n\r\nTo use `select` you must have the list of re-ordered indices of your samples.\r\n\r\nLet's say you have a `shuffle` methods that you want to use. Then you can first build your shuffled list of indices:\r\n```python\r\nshuffled_indices = shuffle(range(len(dataset)))\r\n```\r\n\r\nThen you can shuffle your dataset using the shuffled indices with \r\n```python\r\nshuffled_dataset = dataset.select(shuffled_indices)\r\n```\r\n\r\nHope that helps","body":"Hi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq ","comment_length":120,"text":"shuffle with torch generator \nHi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq \nRight now the shuffle method only accepts the `seed` (optional int) or `generator` (optional `np.random.Generator`) parameters.\r\n\r\nHere is a suggestion to shuffle the data using your own shuffle method using `select`.\r\n`select` can be used to re-order the dataset samples or simply pick a few ones if you want.\r\nIt's what is used under the hood when you call `dataset.shuffle`.\r\n\r\nTo use `select` you must have the list of re-ordered indices of your samples.\r\n\r\nLet's say you have a `shuffle` methods that you want to use. Then you can first build your shuffled list of indices:\r\n```python\r\nshuffled_indices = shuffle(range(len(dataset)))\r\n```\r\n\r\nThen you can shuffle your dataset using the shuffled indices with \r\n```python\r\nshuffled_dataset = dataset.select(shuffled_indices)\r\n```\r\n\r\nHope that helps","embeddings":[-0.2320701778,-0.1594077945,-0.0132597713,0.249286592,0.1917155385,0.016128635,0.086092934,0.1256412417,-0.0738925636,0.5062565804,-0.0565603077,0.6657477021,-0.3163603842,-0.3614477515,-0.1684427559,-0.2553729117,0.0437459536,-0.1630821377,-0.2135148495,0.032560911,-0.2393629104,-0.2892496884,-0.2646623254,-0.005998516,-0.119662039,0.0604129359,-0.0957779735,0.0697435588,-0.2313453108,-0.034571778,-0.1720061898,0.5989302397,0.1985050887,0.4059646726,-0.0001133756,-0.1537085474,0.2525965571,-0.1660947502,-0.2985817492,-0.2896046937,0.3812659681,0.308853507,-0.4335247278,0.0859259665,-0.2295082361,-0.3008211851,0.1807852536,-0.3355771005,-0.018414218,0.0614014305,0.1672649235,0.0012047766,0.2648460567,-0.0670845434,0.519961834,0.3149352074,0.093722187,-0.0753147975,0.5529654026,0.1821962595,-0.1812990904,-0.0414466672,-0.0807733834,0.176660493,0.3287064135,-0.2766603827,-0.5377693772,-0.1196920723,0.1335413605,0.4067071974,0.2531914413,-0.1997767985,-0.2101837546,-0.3013878465,-0.0292875133,-0.0357886665,0.1097449958,0.2107415795,-0.3999198973,-0.1496413797,-0.2572130561,-0.0833966285,0.1118390039,-0.1043697894,0.2718878984,0.6229640841,0.0889731348,0.140069589,0.4378804564,-0.1311396956,0.81343472,-0.0863590613,-0.0097718071,0.2210773677,-0.2670795321,0.0049428302,-0.0320762694,0.3473730683,0.377025038,-0.0491665564,0.1465638578,0.1273645312,0.042808637,0.0636321902,0.1232161373,-0.1424030364,-0.1858207136,0.3928990662,0.2217474431,-0.1759120822,0.1167247891,-0.0850033313,-0.1812994629,-0.5007681847,-0.1929596961,0.2007648051,-0.4478093982,0.0526513308,0.0036981783,-0.4550374448,0.0964477584,-0.2533772886,0.2330251038,0.2721402943,-0.068879813,0.1646943837,-0.028656682,0.0458342843,0.0906960294,-0.2706475258,-0.0352261774,-0.1291973442,-0.1892329454,0.0456854962,0.0670174509,-0.2744113207,0.1627708077,0.083790347,0.2174981833,0.4782420099,-0.0006481794,-0.0075603076,0.1372971684,-0.1494775414,0.2428740412,0.0184350647,-0.1708400697,0.3659123778,-0.1772079319,0.4577911794,-0.295635134,-0.2572662532,0.0998008177,0.1069459841,-0.1186270118,0.0358435139,-0.2069556117,0.0063397097,0.0167940035,-0.3003845215,-0.0010501485,-0.218586117,-0.0986286253,-0.238034457,0.05922959,-0.056848567,0.0386081189,0.0080923988,-0.1043094099,0.1638102382,0.3211078644,0.1710610986,-0.1124110669,0.1007732973,-0.2374942154,-0.521967411,0.3435090184,0.047422938,-0.2410604656,-0.2898803651,0.0686896443,-0.047338672,0.2372686714,0.238971889,0.0942459255,-0.2064083368,0.4309085608,0.4226894677,-0.2441977412,0.1156652942,-0.2087157071,0.0345320404,0.3860126436,0.5716748238,0.0606420599,-0.0480869561,-0.0235483851,-0.1555312872,0.1788862944,-0.0860240087,-0.0381914563,0.0053408854,-0.232880339,0.2788631618,-0.0127549833,-0.0245501604,0.0993406177,0.0880826935,0.1925429702,-0.0531476997,0.1646049172,-0.0536928214,0.2090454847,-0.318884939,-0.1554498076,-0.1669819504,0.0440988317,0.1306563318,0.3118703961,-0.2316394448,-0.0347861759,-0.0988282785,-0.1887603551,-0.144592911,-0.6924346685,0.0650368854,-0.1370727122,-0.2032377869,0.1183274239,0.108606644,0.0162565354,0.0743982419,-0.0976934433,0.2368776798,-0.0716199726,-0.2531092763,-0.0759782195,-0.0043030628,-0.0251436271,-0.0400895551,0.1345903277,0.1320821196,0.1572430879,-0.0739689916,0.0745791867,0.5301996469,-0.716584444,0.0497625172,0.0402512364,-0.1974244267,-0.0040386356,0.2728999555,-0.0175354853,0.4752105176,0.5115851164,0.2074907571,0.1819639653,-0.0543831028,-0.5250878334,-0.11700131,0.2466132939,0.0588052645,-0.0901359543,-0.0052354415,-0.2030215859,0.1712268889,-0.0845850706,-0.0675661564,0.3254946768,0.2940306664,0.0934721157,-0.042188257,0.3154894412,-0.1702660918,-0.0833766237,-0.1547855437,-0.3174166679,0.1140315831,0.190901354,-0.1367492974,-0.0972053558,-0.1878259629,0.1252653897,0.324699074,-0.0931765214,-0.001537644,-0.1604825407,-0.0584675893,-0.2479928583,-0.2115260661,-0.191160053,-0.0813388899,0.1489317566,0.5911383629,-0.0183113813,0.1634088308,0.6033519506,0.093340449,0.1049074903,-0.0179258101,-0.135517329,-0.2551341355,-0.0939079821,0.1050561219,0.1692867279,-0.1076337695,0.3447979987,-0.1817201525,-0.183020696,-0.1967984587,0.2760646641,-0.0145381819,-0.0120235104,-0.113359943,0.1915734708,-0.2135903388,0.130221799,-0.2264458537,0.2502763867,-0.5216255784,0.1627506167,0.2632404566,-0.0664846972,0.0604910627,0.0069141998,-0.1209042743,-0.2186222076,-0.4105161428,-0.0653211474,0.0126492381,0.173568204,-0.1661871225,0.2061451226,-0.3581282496,0.0035157301,-0.2108099163,-0.1508387625,-0.8749733567,0.518968761,-0.2251122594,-0.1516726315,-0.2170840651,-0.0929623693,0.0123957107,0.7811891437,0.320733577,0.3386201859,-0.0524125248,0.1517129689,-0.1009292752,0.5109634995,0.2728753686,-0.050905861,0.1122866794,0.0777172372,0.0903858095,-0.0663493872,-0.0504682288,0.5741662383,0.4247804284,0.22136347,-0.0017520887,0.8899260759,-0.2905632854,-0.26383847,0.042600777,-0.0307404492,0.2523863316,-0.1464753896,-0.1405351311,0.3846739233,0.1928543895,-0.2621259391,-0.0209299885,-0.1710183918,0.1717054248,-0.2726278305,0.1028138548,0.0703629926,-0.1773781478,0.3632208109,-0.1383733898,0.0071825571,-0.2943245471,0.2401369661,-0.2781677246,-0.2899659872,-0.0448491573,-0.0141976522,0.4808888733,-0.0303472057,-0.1537941843,-0.1666100472,-0.3638842404,0.1629349142,0.109147191,0.0897328928,0.2293405086,0.0162330493,0.0828845203,0.0360428728,0.5378273726,-0.238909781,-0.0615054592,0.0643534064,0.0022729572,-0.3917084038,-0.0274627823,-0.0353392735,-0.0840165764,-0.1636383981,-0.2483158112,-0.2289425433,0.0092206933,-0.2370860428,0.2372020483,-0.2000842094,-0.2044199705,-0.2620078325,-0.1691819131,-0.2880616188,-0.1470840722,0.0261420086,-0.1124192849,-0.3721059859,0.2072004974,-0.0124046868,0.1687181741,0.1065880507,0.0220672097,0.2278780937,-0.1550828069,0.0151363304,0.4244745672,-0.0999181047,0.445358336,0.0162491966,-0.0349131711,0.0400872193,0.0022082615,-0.1164841279,-0.0166078825,0.035370402,0.0350533575,0.1604710966,-0.3022455573,0.0573443882,-0.1874417365,-0.1031927392,-0.098775357,-0.2239027321,-0.2644951344,-0.9420631528,0.0981853232,0.3958291113,-0.0640433133,0.6643548608,-0.495169878,-0.1747865528,0.2509578466,0.1397776753,0.7741404176,-0.2292695194,0.0014747363,-0.0686025247,0.1083281636,0.2215510607,-0.0978512391,0.2858168483,-0.3349285126,-0.2567651272,0.0439297296,-0.1366047561,-0.0236119889,0.3318227828,-0.1152201742,0.4559642076,-0.0042942767,-0.245009914,-0.1390374005,0.0753836483,-0.4081730545,-0.1559440643,0.1624041498,0.042834118,-0.1644506305,0.3086212277,0.1053434908,-0.0197795071,-0.1058141142,-0.0229315348,-0.2260188013,-0.1232331172,-0.0799682811,-0.2085879296,0.013517973,-0.1567155123,0.2688783109,0.2193946689,0.3042417169,0.1892430931,0.0748777241,0.1946423054,0.2330042124,-0.0012368895,-0.1648614556,0.1622194499,0.1420402378,0.0310534425,-0.1079802066,0.2484639585,0.3312575519,-0.5866292119,-0.7023997903,-0.2392367125,0.50469172,0.0051348982,0.1765240431,-0.065566726,0.0413446352,-0.0742791966,0.0900724977,0.2556134164,-0.2051658332,0.2700253427,-0.3084162772,0.201078251,-0.065916039,0.2871934772,-0.0891789421,0.1666192263,0.0605418757,0.1801241636,-0.0421833433,-0.0471887589,-0.0409849286,0.2056340426,-0.0253041293,0.0547933392,-0.1302044094,0.0602421872,0.077040188,0.2232198417,-0.1271429807,0.0943940058,0.0939896703,-0.1584349573,-0.3777198493,-0.0785425678,0.2063245475,0.0897401944,0.0748428479,0.0442228839,-0.229246825,0.0879579708,-0.2615520954,-0.0207175408,0.3389473855,0.2913643122,-0.1677930653,0.0946381241,-0.5044462085,-0.0965234488,0.0486081913,0.0726100057,-0.0808961838,-0.1088952571,0.067602694,0.0700225905,0.232036218,0.0154324491,-0.2066195309,-0.4514906704,-0.1086149141,-0.1996985823,0.2130438685,0.4048780799,-0.0517927632,0.2019563615,0.4848182797,0.1200300604,-0.1224202514,0.2897352278,-0.1992783993,-0.1089760065,0.1263922751,0.0384809971,0.1871341467,-0.1477587372,-0.3249956071,0.1066503525,0.2479911894,-0.0242060833,0.2589034438,-0.1300383806,0.2051178813,0.1977376193,0.1431545317,0.5490160584,-0.0241697505,-0.1290996373,0.5621985197,0.1282063574,-0.0485049933,-0.0113104424,-0.3999853134,0.359602958,0.2187380791,0.1192381904,-0.2792426944,-0.4171812832,-0.5137891173,-0.0829114467,0.3224644065,-0.1983130872,0.0232497454,0.4262234569,0.1984590292,-0.1827002615,0.1873044223,-0.2013271749,-0.0598200932,0.422106117,0.1413745582,0.8368740678,0.2010379285,-0.2540157735,0.1464507729,-0.3265659213,0.1080767512,0.3612487614,0.1531527489,0.0878704563,0.2605609894,0.3248278499,-0.0665344596,0.1227415279,0.1017878503,0.1108718142,0.0082595963,-0.0168305617,-0.2495732009,-0.3581893444,-0.1869255602,0.1896223128,0.0208462104,0.0653165653,0.3232699633,0.0491845831,-0.0217556171,-0.4332725406,0.2967405617,0.1902730763,0.2016113847,-0.0842584372,-0.2452662885,0.0960096568,-0.1634313464,0.5964711308,-0.0654608756,0.106040597,0.3574059606,-0.4151381254,0.0308307279,-0.494976908,-0.0706249699,-0.0862888694,0.2017359287,0.048964683,-0.1722868532,0.204906404,0.082596086,-0.1169800609,-0.1994989365,0.1994546652,0.118688263,0.0434386544,0.2000209391,0.1226272434,-0.1002101377,0.1600586623,-0.0721282959,-0.2333839089,0.0152561944,-0.278139472,0.049072016,-0.0462956764,-0.1497135907,0.05244454,0.0912027135,0.2286255062,0.1965642422,-0.0882154331,0.0501561835,-0.4392889142,-0.3919751644,0.0951574817,0.0139532406,0.1518120468,-0.1636558771,-0.0021698033,0.2206371427,0.1337607652,-0.3114302456,0.2564142048,0.0888693482,0.8290222883,-0.1680541486,-0.264500469,-0.3452558219,0.0891216695,0.0270202216,-0.3515399098,-0.0868410617,0.0194348544,0.0622427426,-0.1024460718,0.2027429342,0.0275940113,0.1038210243,0.0102281449,0.0441325381,0.2663683891,0.0345681943,0.2282507867,-0.0968144983,0.3079462647,-0.3173826337,-0.0665174499,0.2037760615,0.0388911478,0.1681126952,-0.2812403142,-0.5187476277,0.4463419318,-0.1663926989,0.2029132247,-0.4137054384,0.1552191228,0.0025386501,0.2402257472,-0.0740205646,0.1094953194,-0.0402411185,-0.0066903871,-0.38485232,0.1179637164,0.4230724573,-0.2074630558,-0.3417066932,-0.5690091848,-0.0971395299,0.2035880089,0.0993620902,-0.2195867151,0.161626175,0.0346583873,0.0751518011,0.3864593506,0.0540911891,0.0597443208,0.094203651,-0.1130542904,0.1797307879,0.1795636714,-0.0864981711,-0.0678467527,-0.0125811193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1611","title":"shuffle with torch generator ","comments":"thank you @lhoestq thank you very much for responding to my question, this greatly helped me and remove the blocking for continuing my work, thanks. ","body":"Hi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq ","comment_length":25,"text":"shuffle with torch generator \nHi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq \nthank you @lhoestq thank you very much for responding to my question, this greatly helped me and remove the blocking for continuing my work, thanks. ","embeddings":[-0.4117562175,-0.2818109393,-0.0233389847,0.3429308534,0.288166523,-0.0083279759,0.1853650659,0.0202477071,-0.0382129364,0.5751110911,-0.0707360506,0.5064848661,-0.2460879833,-0.398517102,-0.2954906225,-0.2201804221,0.1450132132,-0.0676352233,-0.2230791301,0.0393518955,-0.1021092534,-0.307774514,-0.2854660749,0.0284374729,-0.2981184423,-0.0404999778,-0.0155293923,-0.0678452924,-0.2791276872,0.0653344765,-0.1847226322,0.5040544271,0.1741338819,0.423199743,-0.0001157135,-0.1107380018,0.3379265368,-0.0809172317,-0.1778486073,-0.2761378884,0.2808982134,0.4443747997,-0.4619713426,0.0814752057,-0.0761001855,-0.1191534251,0.262019664,-0.3990255296,-0.0382396542,-0.0323174298,0.1677490622,-0.1495522857,0.1914846897,-0.1773432046,0.5097533464,0.4011304975,0.0406591371,-0.1047625244,0.5944638252,0.1509049684,-0.31341362,-0.0186422076,-0.0470052995,0.2459580898,0.2551367879,-0.4061440527,-0.5669309497,-0.1412596405,0.2915235162,0.432637006,0.3464850485,-0.1009705514,-0.0876359418,-0.1382441372,-0.0516583472,0.070657067,0.1695021391,0.2097512186,-0.3878733218,-0.076031357,-0.2963160574,-0.2245026827,0.0772745758,-0.1248117983,0.3338493109,0.5769516826,0.0548146665,0.1275304705,0.3952463865,-0.1300410032,0.7564070821,-0.0276903547,-0.1322856694,0.180185467,-0.2901794314,0.0592062138,-0.1038615331,0.4589678049,0.4592785239,-0.1179590523,0.1412024349,0.1927354932,0.1631337553,0.2064422816,0.0835959464,-0.2326903194,-0.1683253199,0.371948719,0.3275441825,-0.1216964871,-0.1029051691,-0.0803461745,-0.2736363411,-0.4016214013,-0.2905214429,0.1997108907,-0.3854078948,-0.1315917969,-0.0026497534,-0.3606237471,0.1022426561,-0.2754894793,0.1183914021,0.3188547492,0.039439328,0.2343288511,-0.1622237861,0.0793385431,0.1468222588,-0.3387516737,-0.0117167896,-0.0898968801,-0.2668285966,0.0389011987,0.0953970551,-0.2885343432,0.1794987768,0.0288267527,0.2084522396,0.4494912028,0.0679666474,-0.212199375,0.0274584927,-0.2296057492,0.1874508709,0.0382102132,-0.1417286247,0.4477780163,-0.0680315793,0.4003006518,-0.2112270147,-0.1952126771,0.1401190013,0.0867423341,-0.2420710325,0.0497155078,-0.1160952896,-0.0214937422,0.0174704771,-0.4353491068,0.038603358,-0.1387663931,-0.1997874826,-0.2105035484,-0.0770201683,-0.0960155502,-0.0629139543,-0.0114829261,-0.1344392449,0.0013911501,0.3399463296,0.2091975957,-0.1234331205,0.1512252539,-0.2698124051,-0.5102877617,0.373535037,0.0481293723,-0.3487653434,-0.1753028482,-0.0551074669,-0.0634961575,0.3548519015,0.1620312482,0.1916329265,-0.2087263018,0.4538601339,0.3311853409,-0.2098890245,0.0973577127,-0.2796488106,-0.0218266938,0.3807961345,0.5322903395,0.1090922579,0.001875726,0.0981852338,-0.0412237868,0.1082879752,-0.0693678558,0.1326419711,0.0782246366,-0.0831430778,0.2248944342,-0.0522771664,-0.060416732,0.1672779322,0.0103391204,0.1992212683,-0.0871164724,0.2522365749,-0.0813509673,0.3240052164,-0.2543725073,-0.1569236666,-0.1398229003,0.0288605876,0.2397935092,0.2405674309,-0.2476200908,0.0059128054,-0.085572347,-0.1806353629,-0.2193390727,-0.6085528135,0.0424942859,-0.1866376996,-0.1833738983,0.135542661,0.1437725723,-0.1615741104,-0.0219171401,-0.1176350564,0.1262602657,-0.1486366987,-0.0632254034,0.1355209351,0.0105550587,-0.0843917057,-0.0737552345,0.224462837,-0.0117331166,0.1923964918,-0.0517516769,0.1302827001,0.3118939698,-0.5202637315,0.0663340762,0.0422188267,-0.2777667642,0.0717525184,0.1890022457,0.0568766631,0.4894283414,0.6646026373,0.1860252768,0.2167235464,0.0068792459,-0.5235305429,-0.1244735867,0.2731057703,0.1018129513,-0.052761551,0.056969583,-0.1429014802,0.0878971592,-0.023370415,-0.1077101454,0.4354138672,0.2545705736,0.1574991345,-0.1007210687,0.3602264225,-0.1870089024,-0.1009567901,-0.1962851137,-0.1689577252,0.2180515379,0.206701532,-0.2285235226,-0.13107647,-0.3742406964,0.1812791079,0.3783466518,-0.0451701507,-0.0854163244,-0.1065683961,-0.0906846076,-0.3173723221,-0.0933452621,-0.1963348091,-0.0564508401,0.1738361418,0.6990808845,0.0711857975,0.2079226673,0.610693872,0.1983982176,-0.0311340168,0.0571435057,-0.1184085235,-0.2773786783,-0.1420054138,0.1446620673,0.2860934734,-0.1460135132,0.3171652257,-0.2155869156,-0.0514003448,-0.2703024447,0.298989892,0.0336650275,0.0519145764,-0.2218716145,0.2022777796,-0.1200759858,0.0867903307,-0.1907147616,0.2468302697,-0.608558476,0.1329551488,0.2232781351,-0.1191163808,0.0708387345,-0.0028893054,-0.0119696045,-0.1814792752,-0.4910068512,0.0055967756,-0.116973944,0.1098065823,-0.2618025839,0.248354882,-0.3301255107,-0.1022090167,-0.1899798214,-0.222785309,-0.9283091426,0.4673413038,-0.1702859551,-0.2366674691,-0.168489635,-0.0929057896,0.1874645948,0.7728921771,0.1805966049,0.4282252789,-0.0203868728,0.1144338176,-0.1508905441,0.4392540157,0.2280040383,-0.1003654599,0.0300874747,0.0201426167,0.2003813237,-0.0565354563,-0.257371217,0.660173893,0.3345237672,0.1454714239,-0.03929675,0.7990567088,-0.3135177791,-0.2797806263,0.0481291488,0.0203874707,0.2880475223,-0.0831072628,-0.1258369386,0.5681530237,0.1507304609,-0.3594087362,0.041585125,-0.2139391154,0.0977132767,-0.326113224,0.0696129352,0.0967370942,-0.1345916688,0.3717278838,-0.1410247386,0.0119782556,-0.1898114979,0.1996208876,-0.2064849734,-0.2596210539,0.0001750296,0.0459646098,0.3754582107,-0.107609868,-0.2370132357,-0.2379765958,-0.3076353073,0.2023234367,0.112306796,0.2105600089,-0.0607610606,0.1083498597,0.1141337603,0.1026585251,0.5240514278,-0.3630339503,-0.0054853433,0.0366804153,0.1418018788,-0.1902390569,-0.015198864,-0.0458702184,-0.0310828723,-0.1114495769,-0.1698953509,-0.1657045484,0.0426526479,-0.3309707344,0.2020040751,-0.1166919768,-0.2460793853,-0.3332740366,0.0175152868,-0.3408659995,-0.1500421464,-0.0059507447,-0.1389127821,-0.3660569787,0.1460535377,-0.0065317592,0.2338988781,0.1313283294,0.0485682338,0.1726669371,-0.2197837085,-0.0706781,0.4024194479,-0.0574021377,0.4202818573,-0.0177923068,-0.036364466,0.0685212016,0.1908304989,-0.2722372711,0.0684208795,0.1038242206,0.0827696919,0.2520070374,-0.2654286921,0.0069518606,-0.2293833792,-0.0952079594,-0.1460232884,-0.0636298358,-0.3268696368,-0.8247862458,-0.0776376873,0.4725459218,-0.0429716073,0.6893948317,-0.6805840731,-0.0998140946,0.1598953307,0.1242811456,0.9412471056,-0.1447330117,-0.0118783796,-0.2176925838,0.0676959902,0.1971279979,-0.2554460764,0.2600313723,-0.3068582416,-0.2737262249,0.0232750159,-0.1342376769,0.0583815575,0.3374805748,-0.1816402227,0.3982696533,0.0990778431,-0.0826403052,-0.0641167387,0.0807637274,-0.4968972802,-0.10822694,0.1658361107,0.0543700792,-0.1641697884,0.3615782559,-0.0182314683,0.0040453249,-0.1734045744,-0.0340255275,-0.216410622,-0.0998029783,0.0204524826,-0.2153551579,0.0194683038,-0.1478786021,0.3500259519,0.1577974856,0.2589779198,0.1333605945,0.0187996496,0.1726667285,0.183564797,-0.0758863762,-0.1434714943,0.1428443044,-0.0079611465,-0.0302392747,-0.1500080079,0.3065977395,0.2275142819,-0.4429503679,-0.6769730449,-0.2381721437,0.515284121,-0.0193677004,0.1992014945,-0.0631208345,-0.0515103303,-0.0745770037,0.081195958,0.2024207115,-0.1846558303,0.1817303002,-0.1301301867,0.1931194812,-0.0219942331,0.3089700043,-0.1147696376,0.1241697297,-0.0144804381,0.1725402623,-0.0930835605,-0.1138999984,-0.014830322,0.2143169343,-0.0955195352,-0.1233250797,-0.1381346732,-0.1326003075,-0.0305135306,0.1899707317,-0.1120757759,0.0922216922,0.0561104529,-0.3106084466,-0.4455008507,-0.0199979879,-0.0204239488,0.010037289,0.0621727072,-0.0606129095,-0.1569641232,0.1017797515,-0.2399470806,0.0108937221,0.2947953045,0.2491133809,-0.1631932557,0.0049435617,-0.4621196389,-0.0994754136,0.0336514935,0.143073231,-0.0800847411,-0.0845405459,0.1580217481,0.1471370757,0.174471125,0.1282200962,-0.1408059746,-0.3532658517,-0.0574856065,-0.1220607907,0.1553482711,0.377789557,-0.0772094131,0.0837781727,0.367572695,0.1479174495,-0.1945850998,0.2345262319,-0.1320666373,-0.1055456847,0.0293109082,-0.0479655042,0.2639685273,-0.1812372357,-0.2999149561,0.0523467064,0.2912943661,-0.0139581691,0.2302351743,-0.1000665575,0.3668320179,0.2731445134,0.1395557076,0.4675050676,-0.0470695458,-0.1518773437,0.5485619307,0.1394267678,-0.0277040582,0.0231959131,-0.293907553,0.2056673914,0.1130916104,0.0779518858,-0.2080104351,-0.4566201866,-0.3766264319,-0.0330740884,0.316485852,-0.202466175,-0.0190888252,0.3381612897,0.2045635581,-0.0729380623,0.2119436115,-0.0825426504,-0.010702841,0.4644418955,0.1448298991,0.770401299,0.1706359386,-0.1047091857,0.273893714,-0.3623510301,0.1926516742,0.4783174396,0.1935777962,-0.0316708237,0.1713208705,0.3620586097,-0.0268517807,0.189533785,0.0881519243,0.0456706546,-0.0529977977,0.0693024471,-0.1655278802,-0.3278186321,-0.1903777421,0.1844021827,0.037951421,0.2915658653,0.4617573023,0.0100461487,0.0244599525,-0.3408915699,0.4167200923,0.2853493392,0.0605904274,-0.0948177353,-0.1319025457,0.00258886,-0.1960088313,0.6983327866,0.0147394715,0.1742140055,0.3289167583,-0.2885518074,0.0193946939,-0.5623590946,-0.1129851565,-0.0801945403,0.3762345314,0.1828697473,-0.1873295754,0.2603107095,0.0653274208,-0.1364607215,-0.0657808483,0.1286490262,0.005838545,0.1601885557,0.0603146143,0.0711839795,-0.0930918083,0.0907164216,-0.1030724347,-0.332295239,-0.0431362018,-0.2789057493,-0.0934678167,-0.0896466151,-0.2784148753,0.0246645305,0.0955812633,0.1655137539,0.2832508683,-0.0924110115,0.0494510196,-0.5450074673,-0.4882777631,0.3877263069,-0.0940085873,0.1768872887,-0.3144972622,-0.0330036655,0.2500275671,0.1767668575,-0.2163219899,0.2864139378,0.0643693134,0.8334419131,-0.1497571915,-0.2087248415,-0.1839602739,0.1089250892,0.0128711462,-0.3402650654,0.0944580734,0.0786103532,0.067849569,-0.0774593353,0.1345703751,-0.1287464947,-0.080243066,0.0074744504,0.005931423,0.4130083323,-0.0254197977,0.2427446991,-0.1136677861,0.2786675692,-0.1852512211,-0.0410143547,0.2943844199,0.0766365156,0.1211097836,-0.1242446825,-0.4374044538,0.3772437274,-0.2000864446,0.3277047276,-0.4550852478,0.1901217103,-0.0159993172,0.2746884823,-0.0587309897,0.0019138717,0.0185848959,-0.1545495391,-0.3842070997,0.1506045312,0.4040511847,-0.2496137619,-0.454316467,-0.5537909865,-0.0687911287,0.2581370175,-0.0609534755,-0.3576658964,0.1179709435,0.0138706854,0.2096756995,0.4233917892,0.0027504689,0.1004507914,0.0986904576,-0.1208767295,0.09532848,0.0466701984,0.0047049122,-0.0340922922,-0.0419218801]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1611","title":"shuffle with torch generator ","comments":"@lhoestq could you confirm the method proposed does not bring the whole data into memory? thanks ","body":"Hi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq ","comment_length":16,"text":"shuffle with torch generator \nHi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq \n@lhoestq could you confirm the method proposed does not bring the whole data into memory? thanks ","embeddings":[-0.3857867122,-0.2282416075,-0.0162493158,0.4048088491,0.2082107365,-0.0341905281,0.0570118502,0.0808268562,-0.0665858388,0.5318525434,-0.0752742067,0.4812488556,-0.2632014751,-0.4862922728,-0.1832594126,-0.0692368224,0.209336713,-0.0377673917,-0.3435776234,0.031989038,-0.1403766125,-0.3306186497,-0.2982028127,-0.0842134878,-0.2607178688,0.0126891034,-0.1362400502,0.0697403848,-0.2460333109,0.0968679339,-0.2931047976,0.4951768517,0.0974509344,0.4980046749,-0.0001194008,-0.1465671659,0.314034164,-0.1183268875,-0.1504206061,-0.1998485625,0.2539368272,0.3593318164,-0.4748353958,0.1151401922,-0.0948923677,-0.0983739272,0.2776004374,-0.4339160323,-0.0195627324,-0.0745983571,0.1194336936,-0.2219698876,0.1998452246,-0.0841578543,0.5198373199,0.36665079,0.1169424802,-0.0849211365,0.5057342052,0.1422061175,-0.31400159,0.0430175178,-0.0097538056,0.171885401,0.2746033072,-0.3088372052,-0.6670969725,-0.0731379911,0.2729103863,0.3703113198,0.3169616461,-0.1528983712,-0.2055038363,-0.1536009908,-0.1124182269,0.0022435456,0.1612336189,0.2617281973,-0.4892468154,-0.0784293562,-0.342598021,-0.2791954279,0.1053367928,-0.167631641,0.1774778217,0.6292580962,-0.0122196078,0.110183619,0.4721578956,-0.1184208542,0.763874352,-0.0575809851,-0.0921976939,0.2165194154,-0.3028174043,-0.032636717,-0.059119381,0.4093033075,0.5086134076,-0.1756494045,0.176848799,0.2054343224,0.1596412659,0.1293462515,0.1709764302,-0.2012197524,-0.2537612021,0.366658628,0.2740383744,-0.2214652002,-0.0844080746,-0.135183692,-0.2312447131,-0.3415017426,-0.1987459511,0.0932107642,-0.45950073,-0.1035667658,-0.0030539483,-0.372389704,0.1141156703,-0.2568842173,0.0909757912,0.3584319353,-0.0127113601,0.2207596749,-0.1634804159,0.0045393351,0.1502262056,-0.3092858493,-0.0220165849,-0.0666015521,-0.2956474423,0.0870512947,0.0762138814,-0.2042885721,0.1803091317,-0.0373950787,0.182115227,0.4441623688,0.0517636761,-0.2536513805,0.1751945019,-0.2355471104,0.1352522671,0.0250756424,-0.175233826,0.5327965021,-0.1040987819,0.4893251061,-0.2315270305,-0.224889949,0.2488964349,0.060175404,-0.2154031396,0.1239278838,-0.1270207465,0.0521255881,0.1170879379,-0.4041868448,0.0113249468,-0.1173706725,-0.1828221679,-0.2459083498,0.0161361508,-0.0269769933,-0.0341455936,0.0278567336,-0.0910011679,0.0044788048,0.4119673669,0.2299843282,-0.0978752226,0.1580456346,-0.2728039622,-0.5346218944,0.3825376928,0.061661128,-0.3585659564,-0.1421957165,-0.0271634534,-0.0516204424,0.3775514662,0.2457604706,0.1548477858,-0.1419527531,0.5249986649,0.3873007596,-0.2463421226,0.1804000288,-0.2839829326,-0.0025058826,0.3841550052,0.5206304193,0.0664393306,0.000791532,0.1284059733,0.0195430145,0.203168273,-0.058235459,0.1055608019,0.0748063102,-0.2058852762,0.1975598335,-0.0436069407,-0.0503901616,0.1585489064,-0.0291767381,0.2243088186,-0.1018485948,0.2184841335,-0.0466072559,0.3369511664,-0.2232076228,-0.096533902,-0.1030953303,-0.0129000274,0.2410169393,0.244253695,-0.2526803315,-0.0090333084,-0.0129416501,-0.2078794092,-0.170349136,-0.5943424702,0.0505106226,-0.0962300971,-0.1664659083,0.1006870121,0.1108159721,-0.1965514123,-0.0120118763,-0.0871617645,0.1609720439,-0.0896545574,-0.0794939026,0.0786018074,-0.0006750373,-0.0545690656,-0.0681432858,0.3016478121,-0.0086859483,0.2611523271,-0.0938592032,0.1107312888,0.2868093848,-0.4786905944,0.0866324455,0.118223682,-0.3833280802,-0.0188733023,0.2372222841,0.0608041175,0.4270816445,0.5575503111,0.256670028,0.2012055069,0.076348342,-0.5414386392,-0.17388767,0.3183691502,0.0981515497,-0.029399626,0.0670163035,-0.2116579711,-0.0411356986,0.024119949,-0.1402640492,0.48291412,0.2563227415,0.2310438305,-0.0845211223,0.4668060839,-0.1006577983,-0.0706994757,-0.1270671636,-0.1596906483,0.1911114454,0.2077067792,-0.2050695717,-0.1365255564,-0.234330222,0.2272626013,0.3151206374,-0.0151874106,-0.1364256293,-0.0912374333,-0.114206858,-0.3149419725,-0.0350303315,-0.1822603941,-0.0354245789,0.1197428405,0.6539214849,0.0629940927,0.0963203907,0.5549211502,0.2750637233,-0.0267802607,0.0436420105,-0.0518874042,-0.2414872497,-0.0461279936,0.1141184568,0.2237266302,-0.1593601406,0.3454262018,-0.1068453491,-0.0952300206,-0.2359415293,0.3460628986,0.0228703897,0.0347066633,-0.2167380303,0.2048020512,-0.1543881893,0.1254279017,-0.1257212609,0.1758277863,-0.5927733779,0.0427868776,0.2018331289,-0.0510552637,0.0954355076,0.0087260567,0.0055768657,-0.1682067662,-0.4804471135,-0.0460185185,-0.0284464341,0.1921399683,-0.3154762089,0.243944481,-0.2847138345,-0.1365345567,-0.2100237459,-0.2071885318,-0.8603690863,0.4527459145,-0.1606403142,-0.1866165698,-0.2010495514,-0.0664193332,0.1636928618,0.8908515573,0.1798513085,0.3920753896,-0.066160813,0.0514746867,-0.1675226986,0.4999174178,0.2530493736,-0.1604395062,0.0146283293,0.041597154,0.1934222728,-0.123470813,-0.2549111545,0.6271773577,0.3988313377,0.1586617529,0.0381274335,0.8380124569,-0.3068429232,-0.1997573376,-0.013116762,0.042593509,0.3154979348,-0.1849274933,-0.106050618,0.5697467327,0.1633919477,-0.4058308303,0.0360882729,-0.1433459371,0.088269718,-0.3547773361,0.1121891066,0.1103249267,-0.1092029661,0.3583007753,-0.1524185985,0.0267834514,-0.1555154771,0.2084495276,-0.2398774475,-0.2918745875,-0.0470709614,-0.0607747175,0.4366775751,-0.1337849945,-0.3084259331,-0.2118674517,-0.4083958864,0.2108348012,0.144338131,0.1832794696,0.0769476965,-0.0280637275,0.0944011435,0.0602171384,0.5309451818,-0.2350263,-0.0730379075,-0.0475254618,0.0290766936,-0.2154396176,0.0757037401,-0.057270091,-0.0354391709,-0.1516382247,-0.2124570161,-0.2400843054,0.0213416368,-0.3269408941,0.1960454583,-0.1080246419,-0.2059813142,-0.3332966864,-0.0137518412,-0.3194843233,-0.1466920525,0.0508034378,-0.1080151498,-0.3333901167,0.1372601837,0.0277323741,0.2616928816,0.094177939,0.0252877902,0.213062942,-0.2242095023,-0.0468299165,0.3251302242,-0.0547463149,0.4139700532,0.0516986586,-0.0104444092,0.1350213885,0.1824507266,-0.2453268915,0.1804997921,0.1243954077,0.0725270435,0.186986044,-0.3097443879,0.0393320099,-0.3164407313,-0.0102104992,-0.1883475035,-0.1315409392,-0.2916305363,-0.8355184793,-0.064751327,0.4750263691,-0.0714280084,0.7121176124,-0.5950436592,-0.1893879324,0.205436483,0.2334749252,0.9532245994,-0.2147202045,0.1107863858,-0.2075082064,0.1003838629,0.2247795314,-0.2290544361,0.3168805242,-0.2985694706,-0.2406842262,0.0319079123,-0.1628288925,-0.0188270174,0.3456649482,-0.1245016232,0.3787989914,0.0945547596,-0.085395515,-0.1201913953,0.0439822972,-0.5103490949,-0.1726794988,0.2018914968,0.0375602283,-0.1433513314,0.3429370224,-0.0027487322,0.0653845891,-0.195876956,-0.0033345264,-0.1820876002,-0.0864257962,-0.0061657387,-0.2308732569,-0.0923180431,-0.1324801892,0.2446677685,0.1509974003,0.2180853933,0.1461435258,0.1269291639,0.1244303957,0.1544623673,0.0436406583,-0.0863195434,0.1542251408,-0.0138645954,-0.0195204448,-0.0341742262,0.3160554171,0.1929140836,-0.4901720285,-0.7082418203,-0.1624746323,0.5463738441,-0.0350535251,0.1603608727,0.0658219978,-0.0294624139,-0.0667241961,0.0447956324,0.3199017048,-0.2208365202,0.2455599755,-0.1981829554,0.1636130214,-0.0153016523,0.350476414,-0.0838297233,0.0898247883,-0.0239886306,0.064966999,-0.1133532897,-0.0825203061,0.0120757986,0.3228746951,-0.0203041844,-0.0716733411,-0.2228623331,-0.1686680764,-0.1006511897,0.1929335147,-0.085514769,0.116653271,0.0942629278,-0.2360298932,-0.4817040563,0.0820524022,-0.0004006698,0.0293428767,0.0344310887,-0.119694382,-0.1943013668,0.0941484421,-0.2016472816,-0.0253233947,0.3116519153,0.2752121687,-0.1117419004,0.06824819,-0.4960809946,-0.031426169,-0.0068720174,0.197293207,-0.0367345773,-0.0618252009,0.2443395406,0.1506058127,0.2171289027,0.0612501726,-0.1379719079,-0.4391351044,-0.0773075595,-0.1876847893,0.2167333812,0.3900285661,-0.0632094368,0.0774064064,0.3624557853,0.1067811772,-0.1825013608,0.2662628293,-0.0892119929,-0.079182364,0.001112328,0.0310549252,0.2785971463,-0.2274038345,-0.2887029946,0.1435763389,0.3019222915,-0.0708108321,0.2406218648,-0.0909319147,0.2875429392,0.2140903026,0.1873806566,0.497702837,-0.0202794801,-0.1313207,0.4899718463,0.1112990528,0.0164811816,-0.0264144931,-0.3831151724,0.0968847051,0.0691775829,0.1185045391,-0.2059750259,-0.3913081586,-0.446385771,-0.0980848297,0.3503064513,-0.2127090991,-0.084658429,0.2895328999,0.1823283732,-0.0620737858,0.2297113687,-0.0335863307,-0.0436902717,0.5628730655,0.2256509066,0.8604160547,0.1539479792,-0.0754235908,0.1859445572,-0.4044262469,0.1407749057,0.5778158307,0.17845653,0.0551765338,0.0977434888,0.2856556177,-0.0216079038,0.1266252249,0.1554320455,-0.0014043318,-0.0719268247,0.115689978,-0.1180154905,-0.3915811479,-0.1777413636,0.2243792713,0.0020682141,0.1624231786,0.4545370936,-0.0230247173,0.0201804861,-0.2767600119,0.4297183156,0.2775364816,0.0739217326,-0.21070005,-0.1627358794,-0.0457329899,-0.1128472835,0.6732831001,0.0102164792,0.0859267414,0.3581016958,-0.192353934,0.0435049683,-0.5329005122,-0.1255059987,-0.1774291545,0.3884708881,0.196757108,-0.1939303428,0.1862166524,0.0270525031,-0.0982532427,-0.0543634519,0.1873421669,0.0025461949,0.0650909841,0.0881177559,0.0764523223,-0.1814599037,0.1462593377,-0.0796540454,-0.3200246096,-0.0405041687,-0.317478776,-0.2077348232,-0.0626734644,-0.2483086288,0.0030036238,0.1489840299,0.1660983115,0.2623519599,-0.0357371084,-0.0119936597,-0.6107524037,-0.4758829474,0.2888514996,0.002634512,0.2379199117,-0.3334869146,0.1120851859,0.2663121223,0.2129977196,-0.3366816044,0.2122365087,0.0567593798,0.7562640309,-0.0995334312,-0.2308259457,-0.2375010401,0.1835942864,-0.0024824722,-0.4147565365,0.0883702785,0.008879073,0.0334366001,-0.020860929,0.1814442575,-0.1497179568,-0.0138838682,0.0449698642,-0.0042351591,0.3337017298,-0.001201582,0.1898442954,-0.0937556028,0.3001071215,-0.2552218735,-0.1624148041,0.228230685,0.0586303547,0.1370463967,-0.1148883849,-0.3927904367,0.3071728945,-0.1936856955,0.3434869945,-0.3691320121,0.1301064938,-0.0194497667,0.2719525397,0.0371953547,0.0644819364,0.0066810716,-0.230805546,-0.4162609577,0.1733791232,0.3045825958,-0.18138583,-0.3846291304,-0.6210350394,-0.0147980088,0.2028828263,-0.0576893091,-0.4087681174,0.0544934161,0.0544613414,0.2451592684,0.3816484213,0.0486064851,0.0662929714,0.0239229016,-0.1650191396,0.2102393806,-0.0094542196,-0.0084677208,-0.0699319988,-0.065320462]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1611","title":"shuffle with torch generator ","comments":"Yes the dataset is not loaded into memory","body":"Hi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq ","comment_length":8,"text":"shuffle with torch generator \nHi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq \nYes the dataset is not loaded into memory","embeddings":[-0.4172564149,-0.2576541305,-0.0073525999,0.4541786611,0.2750394046,0.0776224062,0.1353669614,-0.0089471824,0.0177605525,0.5187441111,-0.0689717457,0.452678144,-0.3000228703,-0.4433021545,-0.1881413311,-0.1356245428,0.1738031507,-0.0080666002,-0.3012467623,0.0126533266,-0.1754168868,-0.26055637,-0.3218832612,-0.0195965879,-0.2852362096,0.003637942,-0.095186308,0.041683957,-0.2667783201,0.0640376508,-0.2018730938,0.5141158104,0.2154843807,0.4320467412,-0.0001192661,-0.1136964783,0.3135232925,-0.1518145353,-0.2011432499,-0.2675199211,0.2165659368,0.3686363101,-0.4315787256,0.0947271809,-0.1187008321,-0.1437410563,0.2882807851,-0.453918159,-0.0208016317,-0.0443204232,0.1375888586,-0.1794014573,0.1709111333,-0.1291816384,0.5070601106,0.3871383965,0.0790661573,-0.1003779396,0.4632022679,0.162472114,-0.2687987983,-0.0537219606,-0.0138264997,0.2142766118,0.3232809603,-0.3323692679,-0.5895082355,-0.0905475467,0.2511155605,0.4039445519,0.3741146624,-0.1131486669,-0.1184150204,-0.1280744225,-0.0512158684,-0.0077771223,0.2349291891,0.2542429268,-0.388184607,-0.0622315742,-0.3345398307,-0.3378303945,0.0948628411,-0.1163159013,0.2202768475,0.5894504189,-0.0162086412,0.1504888535,0.3903769851,-0.0895751864,0.7511168122,-0.0665607303,-0.1625945866,0.21521613,-0.2788399458,0.058413811,-0.1372701824,0.4218731523,0.4573015273,-0.1608717293,0.1525581628,0.1771359295,0.1148109138,0.1450178176,0.1879624873,-0.223184526,-0.2192224562,0.3973278403,0.2741510272,-0.212725848,-0.1176295057,-0.1372213662,-0.3294937015,-0.323487252,-0.1633132249,0.1382562667,-0.3763881922,-0.1267920434,-0.0366699658,-0.3873645365,0.1286102682,-0.2604784071,0.0878584757,0.4022954106,-0.0226660129,0.1568280756,-0.1313514858,0.1161480993,0.121035412,-0.3480960727,-0.0365916006,-0.1322951615,-0.2802034914,0.0361868963,0.0943706632,-0.2529618144,0.2332258075,-0.0416237041,0.1727854311,0.4228042662,0.0580074824,-0.2110720426,0.1135472059,-0.1646860987,0.2231448293,0.0395862088,-0.1371367425,0.5068792105,-0.1222118437,0.5125057697,-0.2489311844,-0.1513232887,0.2081185877,0.0531801358,-0.2951083481,0.0772988498,-0.1722038537,-0.0078702215,0.104617916,-0.4314145446,-0.0013342852,-0.1631149203,-0.2166919857,-0.2541194558,-0.0228160862,0.0427333936,-0.0695155039,-0.0413107537,-0.1004256234,-0.0501646884,0.361086905,0.1661679447,-0.0800995082,0.1352221966,-0.3510179818,-0.5362803936,0.424234271,0.0911071077,-0.330681175,-0.2192699462,-0.0341309868,-0.0443285853,0.2918801606,0.2465581596,0.1523038298,-0.1270693094,0.4730614126,0.3895826638,-0.1796916574,0.1198036745,-0.1932836771,-0.0393821783,0.3962083757,0.4849229455,0.1491127461,0.0459242761,0.1072564498,0.0198977552,0.1271045506,-0.000964325,0.0852108896,0.1456622481,-0.195342958,0.2648466527,-0.0072809421,-0.043233186,0.1127988696,-0.0111801177,0.2179225385,-0.1530665606,0.2776351869,-0.0147876516,0.3134383261,-0.1966531575,-0.1914327294,-0.0871848688,-0.0034652178,0.2629280686,0.2148148268,-0.252697885,-0.0586534031,0.0060223457,-0.2438748181,-0.1776144654,-0.593398571,0.0574669987,-0.1160838455,-0.1856956035,0.1330730915,0.1388080567,-0.1709258407,-0.0291883815,-0.1173846275,0.1701013595,-0.0733082518,-0.0819162428,0.046974048,-0.047145918,-0.0727824643,-0.1120413542,0.2687520981,0.0220983066,0.1941319555,-0.0847183466,0.1551468074,0.3070307672,-0.4834379256,0.0479591116,0.0929590687,-0.3256712556,0.0540596433,0.2536333501,0.0370441228,0.402979672,0.5861380696,0.1654393673,0.2613726556,0.0839670226,-0.4801838696,-0.2429644912,0.3104994595,0.1122323126,-0.0176166259,0.0483176149,-0.3052173853,0.048692856,-0.003248573,-0.1563994139,0.5221976638,0.2573235333,0.1609569192,-0.1242510378,0.4241300821,-0.1579315662,-0.0844814479,-0.181147173,-0.1038622782,0.1797628552,0.1539476365,-0.2377531379,-0.2140821815,-0.2357824594,0.2061816156,0.3395680189,-0.0809368789,-0.132942766,-0.0830390155,-0.0489908569,-0.2663186491,-0.0504123606,-0.210602507,-0.0366888605,0.0969640464,0.6521186233,0.1429716796,0.1414892823,0.5446645021,0.202637732,0.0078050205,0.0180150457,-0.0818114132,-0.2625120878,-0.1385827065,0.1199160367,0.2001273036,-0.1715459228,0.3412088156,-0.1686442941,-0.0393398181,-0.2356674075,0.3507167101,0.0536683314,0.0858690515,-0.1954689026,0.1845224649,-0.0792537555,0.0953188464,-0.0841185898,0.2353701144,-0.617030561,0.0580501854,0.1935273558,-0.060814701,0.1131364405,0.0389918461,-0.0217002816,-0.1848857254,-0.4413002729,-0.0921057686,0.0094671911,0.0906393826,-0.2733371854,0.2702860534,-0.3127991855,-0.1518042386,-0.1919543594,-0.2601256371,-0.8989681602,0.4933840334,-0.1639874876,-0.2278005481,-0.1197731942,-0.0777771771,0.2266803533,0.8367548585,0.1030227318,0.5156798959,0.0099568265,0.0744335949,-0.1481958926,0.4359954298,0.2790837288,-0.1030105576,0.0612501688,-0.0041277432,0.185875237,-0.0796302706,-0.2546643913,0.6798548698,0.3548481762,0.1641103923,-0.1280556172,0.8535884619,-0.3665952384,-0.2430506349,0.0770197064,-0.0183583759,0.380810827,-0.1264204979,-0.1208880022,0.521894753,0.1749370545,-0.3622096777,0.0534838662,-0.1152756885,0.0713691041,-0.3536589146,0.1040801778,0.0915587097,-0.0873621851,0.3006140888,-0.1783081293,0.0031706241,-0.1649320275,0.1865438074,-0.2562504113,-0.2888714969,-0.0315827392,0.0256936476,0.4224099219,-0.1086862683,-0.2643993199,-0.2481362969,-0.3114062846,0.1468345821,0.0789487734,0.1764778346,0.0044002724,0.0588959306,0.1306385845,0.1162516549,0.5068795085,-0.3161603212,-0.0416194312,0.0309419129,0.0869178027,-0.2146820426,0.0100355335,-0.0505239367,-0.0263474043,-0.1317317784,-0.2067909092,-0.1458236426,0.0112127885,-0.2981491983,0.1864436418,-0.2082519978,-0.3046280146,-0.4174587429,0.0356002562,-0.3386410773,-0.1908564717,-0.0015541976,-0.1236179844,-0.3270394802,0.1095510349,-0.0662461296,0.235781312,0.1348797679,0.0473581776,0.2282689661,-0.1919890791,-0.0819855779,0.4000643194,-0.0553886294,0.4316190779,0.0753356367,-0.0031253456,0.0529211909,0.2218060046,-0.2250640541,0.1489886642,0.0119941691,0.018229302,0.1480341405,-0.2990829647,-0.0274937674,-0.3127638996,-0.0080620172,-0.1077896282,-0.0951907337,-0.3252776563,-0.874861002,-0.0281357747,0.4811258912,-0.0472454838,0.7214950919,-0.6548552513,-0.1485260427,0.1800335497,0.1849250942,0.8657466769,-0.322825402,0.0527380668,-0.1469038874,0.1306144148,0.2256391048,-0.2790879309,0.2732538879,-0.2907871008,-0.2251387686,0.0081772134,-0.1328106076,0.0363632254,0.3856485486,-0.1042183787,0.3783327043,0.0793752,-0.0466748178,-0.0846769959,0.0514418967,-0.446795404,-0.1238093525,0.1740867943,0.0354871452,-0.15434663,0.4210706651,-0.0274827033,0.0874101892,-0.1390148252,-0.0388088077,-0.190545097,-0.08887247,-0.0207098201,-0.220550403,-0.0904445201,-0.1284923255,0.2760154605,0.2277702689,0.2679238319,0.1522907317,0.0785872862,0.1871243417,0.1540585756,-0.0897450969,-0.1385737062,0.2342576832,-0.0396103971,-0.0153998146,-0.1068155617,0.385404706,0.1906506717,-0.4774028063,-0.6623187661,-0.1466487348,0.5143477917,-0.0488489941,0.1290789098,-0.0170109291,0.0285299625,-0.0348691307,0.0468598306,0.2688119113,-0.1619102508,0.1532989442,-0.0957584307,0.0981480405,-0.0502895564,0.3655324578,-0.0371738821,0.0713994727,0.0464504026,0.10790319,-0.0709504038,-0.100297533,0.0023291435,0.3129950166,-0.04915151,-0.12902385,-0.0950135142,-0.1110265926,-0.0752718523,0.1934663057,-0.0982480571,0.074654609,0.1184294149,-0.2689265013,-0.5081481934,-0.008781503,-0.0621132441,0.036382366,0.1471021026,-0.1928832084,-0.1312040687,0.1693210751,-0.1911110878,-0.036146041,0.2934356332,0.2451007962,-0.0961462632,0.0997860655,-0.4704484344,-0.086979717,0.0106464056,0.0810489804,-0.0275142975,-0.0559317507,0.2362267822,0.1514450908,0.1700442731,0.1411237568,-0.2151871026,-0.4494339824,-0.0785080045,-0.167414993,0.2260550857,0.3869441152,-0.0889179483,0.0908590406,0.39167732,0.1159122214,-0.2110989988,0.3005178273,-0.1524670124,-0.0919817388,0.0155773303,-0.0564423241,0.2979203463,-0.1829430014,-0.3130237758,0.051431004,0.2878363729,-0.0609176345,0.2400103807,-0.1417956501,0.3698171079,0.2088542879,0.1158744097,0.4728652537,-0.0395756885,-0.1239001453,0.459094584,0.101862587,-0.0513953269,0.0402031019,-0.2938166559,0.1959022731,0.0449455343,0.1200676039,-0.2403406799,-0.4203652143,-0.3791614473,-0.055692479,0.3557011485,-0.1911322325,-0.0368803144,0.3124746084,0.1791945994,-0.0231008139,0.213212952,0.0154688181,-0.0399669297,0.5714643002,0.1879053265,0.804721415,0.171346277,-0.0645318106,0.2010854781,-0.4597210884,0.1304383427,0.5441171527,0.1955968887,0.0286760703,0.1236555353,0.3123939335,-0.0113598118,0.1274809092,0.0722063631,0.0012584975,-0.0999426544,0.076247327,-0.1542072445,-0.3255072236,-0.1429716051,0.2212847769,0.0260194838,0.155599609,0.4209599495,0.0172581561,0.0210216157,-0.3052648902,0.4754823148,0.3088971972,-0.0312408954,-0.1539303064,-0.1148339137,-0.0703257844,-0.1357994974,0.7224518061,0.0266271047,0.1380037069,0.3618769646,-0.195149526,0.0243431702,-0.4962057769,-0.1478854269,-0.1315387189,0.3894463778,0.1904330254,-0.1434013098,0.2386009842,0.047999423,-0.0684619471,-0.0584193841,0.1440997571,0.0487344041,0.1067546457,0.0810362846,-0.0232138149,-0.0956474617,0.0792068243,-0.0900278315,-0.3392567039,-0.0081157787,-0.2431724221,-0.1532803923,-0.0668408722,-0.2661015689,0.0004239056,0.079190515,0.1911295205,0.3354842365,-0.0574785732,0.017516356,-0.5783042908,-0.5666031837,0.3918582201,-0.0980672091,0.2230306864,-0.3148997426,0.0362562239,0.2233536392,0.1836007237,-0.2710875571,0.3008387983,0.0619994216,0.7649537921,-0.1491718739,-0.2832429111,-0.2479470372,0.1978462934,-0.0094157467,-0.3456152678,0.0473972969,-0.0056105671,0.008892809,-0.062325798,0.1459884495,-0.1727380008,-0.0311553981,0.0478592739,-0.0815186799,0.4574809372,-0.0131746307,0.2769522667,-0.0912522674,0.2331746072,-0.1630414426,-0.0727345049,0.2737430632,0.0515466332,0.0599919409,-0.0843287781,-0.4109747112,0.3249227107,-0.2246989757,0.349756211,-0.4349318743,0.1519673318,-0.0518246405,0.3190074563,0.0332220718,-0.016493557,0.0266288053,-0.2032495886,-0.3578418195,0.1876041591,0.370582372,-0.2307288349,-0.3869144917,-0.6003525853,-0.068201229,0.2432967722,-0.0714902058,-0.4167401195,0.1659068465,0.1098588631,0.2194311321,0.4185340703,0.0149032045,0.0553715266,0.074083671,-0.1586630493,0.2096779644,-0.0111005697,0.0205009282,-0.0895950794,-0.0694929659]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1611","title":"shuffle with torch generator ","comments":"great. thanks a lot.","body":"Hi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq ","comment_length":4,"text":"shuffle with torch generator \nHi\r\nI need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help \r\n\r\n@lhoestq \ngreat. thanks a lot.","embeddings":[-0.4197355509,-0.3070660233,-0.0212705228,0.3629998565,0.2520819604,-0.003981764,0.1494320184,0.0312668756,-0.0511774197,0.5634387136,-0.0856267512,0.5248087049,-0.266100049,-0.380915761,-0.2814808488,-0.2384521067,0.1838306636,-0.0642709956,-0.2371503711,0.030540945,-0.1117504239,-0.3193590939,-0.2920676172,0.0236403625,-0.2711904347,-0.0057208831,-0.0739056543,-0.0782106146,-0.2642969489,0.0803858638,-0.1954066157,0.5167433619,0.2295534015,0.4324381948,-0.0001177257,-0.1295404732,0.3284322917,-0.0967333168,-0.1528692245,-0.2359811366,0.2653710246,0.4253270328,-0.4704970419,0.075608179,-0.0780158043,-0.1182838753,0.2700307965,-0.4204277396,-0.0432951674,-0.0421132743,0.1442633867,-0.154702425,0.1818864793,-0.129000023,0.5224452019,0.4289605916,0.0664919466,-0.1066406369,0.5992360711,0.1449380517,-0.2970390916,-0.034734074,-0.0465201326,0.2256380618,0.2866126597,-0.384519875,-0.6026039124,-0.1381328106,0.2458613664,0.4439589083,0.3392515779,-0.1258958876,-0.0714056119,-0.1451542675,-0.0615012236,0.0517364517,0.1619475633,0.2075221539,-0.3595267236,-0.1039451063,-0.3077360392,-0.2286465913,0.0542927571,-0.1261933893,0.3305749893,0.5788705945,0.0450259,0.1383505613,0.4010510743,-0.1266091019,0.7531282306,-0.0456650034,-0.1257606298,0.2054907382,-0.282171011,0.0216344409,-0.1524496228,0.4349313676,0.4673796594,-0.1056509838,0.1151001602,0.1772566736,0.1261613816,0.2099637091,0.1112674028,-0.2187282145,-0.1568425298,0.3814126253,0.3154910207,-0.1637527794,-0.1103892699,-0.0865468308,-0.2904098034,-0.3661591113,-0.2439682484,0.2003751695,-0.3990047276,-0.1177270636,0.0264556743,-0.3834649026,0.1068516523,-0.3020831943,0.0974896848,0.307493031,0.0093325479,0.2069602013,-0.1848154515,0.0981769487,0.1526456922,-0.3571290672,-0.008488318,-0.0953390673,-0.2527647614,0.0392060913,0.0828915834,-0.2813788652,0.1723152995,0.0163646564,0.1942354441,0.4501247108,0.0608263202,-0.185039714,0.0046986691,-0.2672036588,0.1911841184,0.0280286204,-0.1456118673,0.4797125459,-0.0873901248,0.4146392941,-0.2289162129,-0.2036859393,0.136439234,0.0752304196,-0.2472047359,0.0276677553,-0.1545600593,0.0057584345,0.0278609879,-0.4062950015,0.0466707647,-0.1435768753,-0.1840311438,-0.2005269676,-0.0881888792,-0.0991066024,-0.0201792102,-0.0115152942,-0.1041229889,-0.0234266631,0.3517248034,0.1976461112,-0.1052308902,0.1239225864,-0.2614882886,-0.5104750991,0.3952814043,0.069689326,-0.3050644696,-0.1755915284,-0.020200599,-0.0846522748,0.3434268832,0.1995253861,0.1951178163,-0.202645272,0.402875036,0.3555676043,-0.2097111046,0.0928196535,-0.2656313181,0.0070950794,0.3947251737,0.5484529138,0.1123577356,-0.0001295548,0.0963770822,-0.0400959514,0.1211276427,-0.0430287495,0.1129154861,0.0630560815,-0.1035681814,0.2557709515,-0.0332453139,-0.0566230305,0.1901441067,-0.0130154854,0.1959370375,-0.1175603569,0.2923246324,-0.0687357187,0.3405936658,-0.2617041767,-0.15106076,-0.1114368439,0.0073435195,0.2355025411,0.2526477575,-0.2357226312,-0.0298456922,-0.1002676487,-0.1733666658,-0.2130662203,-0.6108283997,0.0620438308,-0.1512788981,-0.1538875252,0.147184357,0.1229750291,-0.1597095579,-0.0370754674,-0.1427028477,0.1498075277,-0.1404726207,-0.0819080248,0.1242695972,0.0443410017,-0.077000536,-0.0880899876,0.2272853702,-0.0090157697,0.1742033958,-0.0793914348,0.1377439648,0.3418453038,-0.5362430811,0.0671041682,0.0085408762,-0.2792390883,0.0381972492,0.2064945102,0.066002652,0.5116209388,0.6407732368,0.1745316386,0.204980284,0.0035285773,-0.5204299092,-0.1307123452,0.2500272989,0.0597138479,-0.0535972342,0.0575685576,-0.1665382385,0.0943816453,-0.0368408337,-0.1147362962,0.462102592,0.2305439562,0.1652555019,-0.0763757154,0.3656658828,-0.1679429263,-0.0901842192,-0.2244118303,-0.1679785997,0.1832863241,0.1692134887,-0.2225479335,-0.131765008,-0.3737792373,0.1679139733,0.4090622663,-0.0402210839,-0.0660147071,-0.0861207396,-0.1031983942,-0.3454341888,-0.0992413461,-0.1918388307,-0.0425412841,0.1558878869,0.663386941,0.0984335169,0.205568403,0.616019249,0.2280511111,-0.013978933,0.057768736,-0.1016871557,-0.274633944,-0.1491436064,0.1298324913,0.2785968184,-0.1499190331,0.3096302748,-0.2348050475,-0.0208435021,-0.2732505798,0.3347325027,0.0161079913,0.0476161651,-0.2531341016,0.1988132447,-0.1413429677,0.0944512933,-0.1916799098,0.2425931096,-0.6137949228,0.1236052513,0.2375344485,-0.1163332388,0.0605261624,0.0345206,0.0164676625,-0.1889287978,-0.4479548633,-0.0159244798,-0.0899565965,0.1062138453,-0.267652154,0.2399479151,-0.3426269591,-0.1120214239,-0.1891501546,-0.2095315903,-0.961660862,0.4455385804,-0.1672071964,-0.19705832,-0.1807746589,-0.0793027207,0.1851182133,0.7851221561,0.1649507582,0.4142147005,-0.0038027121,0.0974696279,-0.1468311101,0.4828581512,0.1704725623,-0.1005701125,0.0668402016,0.0547160544,0.2082857639,-0.0615531318,-0.2441043556,0.6736752987,0.3530527651,0.1209079847,-0.0642888844,0.7953141928,-0.2940876782,-0.2954678237,0.0199210383,0.0043484438,0.2979815304,-0.0477251299,-0.1194577366,0.5893548727,0.1705934405,-0.3371338248,0.0303796064,-0.2132965326,0.0744192898,-0.3423092067,0.0387023985,0.1213111132,-0.1111398488,0.3693023026,-0.1770630628,0.020687202,-0.214894712,0.1823052019,-0.2145400941,-0.2726568282,0.0249449573,0.057732828,0.3653135002,-0.0949217603,-0.2399185151,-0.2603058517,-0.2724301517,0.195758462,0.0858864263,0.1976623535,-0.0027946932,0.1025118306,0.1031434536,0.1188765094,0.5143902302,-0.3377234638,-0.0128019154,0.0133263897,0.1259002835,-0.1830615252,-0.0294640306,-0.064090468,-0.0274418853,-0.1250414103,-0.1676830947,-0.1506131142,0.0544140153,-0.3024461269,0.1949875057,-0.1317970008,-0.2670108974,-0.3139562309,0.0292556416,-0.302390039,-0.1519881934,-0.0506527089,-0.147661671,-0.3674740493,0.1529659033,-0.029641794,0.2204916477,0.1304358691,0.0576662086,0.1770299077,-0.2458881438,-0.0841316432,0.4033033252,-0.0666820258,0.4059658349,-0.0310312398,-0.068176575,0.0536922701,0.1603971273,-0.2530567944,0.0838613212,0.039036002,0.0878179818,0.2342342436,-0.279724896,0.0144835599,-0.2438729107,-0.0831610337,-0.1451356113,-0.0621040724,-0.3343507349,-0.8782353997,-0.0298608001,0.4620232284,-0.0370983705,0.6940491796,-0.699451685,-0.1106030121,0.1744893789,0.1490394324,0.9318234921,-0.1476341039,-0.041496899,-0.2223467231,0.0894169211,0.2219632864,-0.2644958794,0.2394836247,-0.3370785713,-0.2438518107,0.0081070401,-0.128838852,0.0434722751,0.3482920229,-0.1881698519,0.4099946022,0.132515654,-0.1002013162,-0.0802490115,0.0712564141,-0.4675809741,-0.0826677233,0.199695617,0.040171802,-0.1799409837,0.3627780676,0.0000685917,0.0167336762,-0.1263702363,-0.015519754,-0.203967467,-0.1113609523,0.0273404084,-0.2002467364,0.0590927973,-0.1437794566,0.321171999,0.1856298894,0.2769383788,0.1164317504,0.0410367772,0.1507927179,0.2219494581,-0.0734380335,-0.1472444683,0.1574625522,-0.0255627427,-0.0218991432,-0.1466015279,0.3181573153,0.2507700324,-0.445129782,-0.6563791633,-0.2501164675,0.5176440477,0.009860185,0.207080394,-0.0528004244,-0.0395616628,-0.0488239266,0.0579019897,0.250934273,-0.2032769322,0.2124416381,-0.1465697885,0.1912404448,-0.044636637,0.2651584148,-0.1085812524,0.0969727635,-0.0488834679,0.1530075073,-0.0649077222,-0.0813804939,0.0282103177,0.2128171623,-0.0656934604,-0.1267228425,-0.1551629901,-0.090963386,-0.0150578739,0.1940467805,-0.0997666195,0.1083663702,0.0774615854,-0.2912533283,-0.4233079851,-0.0453211218,-0.0123326313,0.0281603429,0.0692533627,-0.0848819092,-0.1453825384,0.1197234318,-0.2124211639,0.0209694598,0.3275422752,0.2408879399,-0.1552738547,0.0688285753,-0.5024100542,-0.079311423,0.0107036354,0.1252432168,-0.0749498978,-0.0557715446,0.1537416577,0.1509513706,0.1740913093,0.1404607892,-0.1564353853,-0.3381943703,-0.0744335651,-0.1283567697,0.1231037602,0.3715675473,-0.0858027339,0.1302604079,0.3891245425,0.0926166549,-0.2167634219,0.2585119903,-0.1042531356,-0.1060233414,0.0358165652,-0.0291795488,0.2716376483,-0.2185893357,-0.2742494345,0.0803918764,0.2716799378,-0.0556507371,0.2390550375,-0.1102045104,0.3766094446,0.2702060044,0.1285731643,0.4839735031,-0.0639681667,-0.1528812051,0.5370222926,0.119154945,-0.0177627653,0.011496597,-0.3022294641,0.261248976,0.10633564,0.1253052056,-0.2175364643,-0.4405028522,-0.4188861847,-0.052490063,0.3262076974,-0.18933478,-0.0069998433,0.3430101275,0.2044764757,-0.0679215491,0.204364717,-0.0651205257,-0.009094731,0.4823642671,0.1631817669,0.7924056053,0.1771524251,-0.1363172531,0.2628966868,-0.352701664,0.1929054409,0.480938822,0.1574998051,-0.0088533368,0.1770243645,0.3692762852,-0.0464614667,0.1698323041,0.1017849743,0.040010795,-0.031335365,0.073882632,-0.1621073633,-0.3382014334,-0.1932371855,0.1933868378,0.0340749249,0.2588867247,0.4588282108,0.0197645612,0.0391288474,-0.3481440544,0.4250051975,0.3162654936,0.0736589283,-0.0901714116,-0.1503045708,0.0103920549,-0.2070543468,0.698733449,0.0127836429,0.1618237644,0.3391192853,-0.2906762064,0.036768999,-0.5638827085,-0.0996599495,-0.0602338463,0.3483062983,0.1705809236,-0.2240268141,0.2494500577,0.0466321148,-0.1044528335,-0.0417220928,0.1299710423,0.0164251179,0.1465009004,0.0753576607,0.0651675686,-0.088432461,0.0800279081,-0.0973285586,-0.3252268136,-0.0508030951,-0.2830638289,-0.0865568519,-0.0877807513,-0.2704348564,0.0099375648,0.1143937111,0.1765779257,0.3051988184,-0.08596389,0.0428845733,-0.567115128,-0.4892757535,0.3687989712,-0.0727415755,0.1936689168,-0.3219328523,-0.0370962918,0.2352254242,0.2038365006,-0.2201818675,0.2973442078,0.0533442721,0.8139891624,-0.1658262759,-0.2305332422,-0.2215096056,0.1412493587,0.013297799,-0.3174909949,0.0889745355,0.0633765981,0.0388632603,-0.0589310862,0.145030573,-0.1249011979,-0.0946803614,-0.0112928553,-0.0175971761,0.3787477911,-0.0019923402,0.2658336461,-0.0702659711,0.2923523188,-0.2100884467,-0.0481999628,0.2818689942,0.0683599785,0.1209639981,-0.1309904158,-0.4302137196,0.3588554561,-0.1651386172,0.3096039295,-0.4629254341,0.161924541,0.0004584421,0.2471159995,-0.0212661233,0.0157194622,0.01762175,-0.1447013617,-0.3781061172,0.1858365536,0.3723663986,-0.2351487726,-0.4557902217,-0.5890541673,-0.0868969709,0.2575583458,-0.0346079543,-0.383536607,0.113779135,0.0475450009,0.204261452,0.4237786233,0.0011210282,0.100408718,0.0995463729,-0.1305478215,0.0958576724,0.0548480488,0.0053823371,-0.0157013014,-0.0410898104]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1610","title":"shuffle does not accept seed ","comments":"Hi, did you check the doc on `shuffle`?\r\nhttps:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?datasets.Dataset.shuffle#datasets.Dataset.shuffle","body":"Hi\r\nI need to shuffle the dataset, but this needs to be based on epoch+seed to be consistent across the cores, when I pass seed to shuffle, this does not accept seed, could you assist me with this? thanks @lhoestq\r\n ","comment_length":9,"text":"shuffle does not accept seed \nHi\r\nI need to shuffle the dataset, but this needs to be based on epoch+seed to be consistent across the cores, when I pass seed to shuffle, this does not accept seed, could you assist me with this? thanks @lhoestq\r\n \nHi, did you check the doc on `shuffle`?\r\nhttps:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?datasets.Dataset.shuffle#datasets.Dataset.shuffle","embeddings":[-0.2270244658,-0.3764101863,-0.0347145386,0.0943192467,0.3178338706,-0.0177694447,0.0859302431,-0.0452611633,-0.0717076659,0.4275326133,0.07351356,0.2512517869,-0.2236372083,0.5158159137,-0.2870840132,0.0741080493,0.16077438,0.0093610883,0.0747007653,-0.064170517,-0.2869283259,-0.156941995,-0.2876102924,-0.3281210959,-0.2399264425,0.0930236727,-0.0817826763,0.1465920508,-0.3244695663,0.0034953998,-0.0870955512,0.5332155228,0.1753993928,0.5836760998,-0.0001161893,-0.1176713556,0.0537752286,-0.1494767219,-0.0911681727,0.1039269567,-0.339792192,0.6304675341,-0.3235476613,-0.0956925452,-0.4316250682,-0.0046027638,0.0314693823,-0.0276742131,0.1604911536,-0.2300323695,0.138571009,0.0853514895,0.1943444014,-0.1003290638,0.3118543923,0.3102669716,0.0372060053,0.2198551148,0.0236507598,0.2999255359,-0.0057597538,0.0188830942,0.1028316841,0.2216716111,0.4829649627,-0.2831609249,-0.0619048737,-0.1659423262,-0.0794781074,0.2915311754,0.6015408635,-0.0583482645,-0.2350591868,-0.0776054189,-0.1141482368,-0.6467604041,-0.0209242105,0.1110306382,-0.0934297368,0.0325319543,-0.0719309673,-0.0646490231,0.0207141787,-0.0758901089,0.2913765311,0.5887156725,-0.0165456682,0.0836897567,0.5269889235,-0.1077489257,0.0430145003,0.1221886054,-0.2218786031,0.1637289226,-0.2919984758,-0.2278884053,-0.2662092745,0.7055479884,0.3622874916,0.1038907543,-0.1570546329,0.2060819566,-0.0271294303,0.2249286026,0.1571903676,0.1134995967,0.1775589138,-0.0966342613,0.0902489871,0.1410679966,-0.000145155,0.0524043702,-0.0407670215,-0.0318055153,-0.0772356391,0.0057111508,-0.2298544496,-0.2930741906,0.0319502689,-0.5817496181,-0.031546656,-0.3947674334,0.4023291171,0.3781959414,-0.1151031107,0.1313766539,-0.0595691092,0.1259260178,0.0091188625,-0.3755142689,-0.3232355118,-0.1133233085,-0.2807452083,0.225225091,0.0116768004,-0.3272070289,0.3453485966,0.3433463573,0.1351230145,0.1227802262,-0.2631951272,0.0617210157,-0.1014179662,-0.0258127581,-0.0246376097,0.0667796507,0.1053773612,0.0885520503,0.0979260653,0.2774964273,-0.4386060238,-0.1242101789,0.2671051621,0.1075255871,-0.104808487,-0.2052968889,-0.6282554269,0.1451850832,-0.015884839,-0.077090919,-0.0621270984,0.0018710046,-0.2864667177,-0.177394703,0.3764268458,-0.1183036342,-0.0113375131,-0.1621814519,0.1931909323,-0.0514833927,0.1639101505,0.6112811565,-0.0210086629,-0.1909288466,-0.0078680096,-0.0006425783,0.0970515758,0.1385854036,-0.2902570963,-0.3103143275,-0.0468755029,-0.4180058837,0.0449452922,0.1815032214,0.1104879454,-0.278262347,0.132797718,0.3897395134,-0.1999438256,0.1246545315,-0.2816767097,-0.0888316035,0.0549316853,0.4264356196,-0.1624675244,0.0481621884,0.036051847,-0.2870487869,0.3077259064,-0.0842451006,-0.0053018103,0.025502732,0.2694328129,0.3269290924,0.1680730134,-0.0561404936,0.0010099692,-0.0257957373,0.320104152,-0.0115530947,0.4818516374,-0.1168672144,-0.229482308,-0.2664842904,-0.181918323,-0.0011727004,-0.0064593549,0.0404428691,0.1324110031,-0.118868649,-0.1984938681,-0.1135683656,-0.2762009203,0.0578474477,-0.5004374981,0.0688507631,-0.1163398698,-0.3386281133,0.06052012,0.0628021955,0.2084770203,0.0091140801,-0.0389623083,0.4128287435,0.1446229368,-0.2592532039,-0.0984322727,0.2539079785,0.0530858561,0.0195054933,-0.0004132227,-0.4485636055,0.2694408596,-0.2347464412,0.1787746102,0.2207800448,-0.3151164353,-0.2249840945,-0.0424867086,-0.2713719308,0.1822666526,0.0879910067,-0.165549472,0.2098504752,0.3585236073,-0.0200881753,-0.0097992113,0.0676245764,-0.4856426418,-0.14947097,0.240477249,0.0213830099,-0.0720862746,-0.1118322983,-0.1505843699,0.0641885996,0.0276651848,0.007142277,0.193111062,0.253731221,0.0201879852,0.0204906557,0.080447793,-0.1983421594,-0.2080911845,-0.2214631885,0.2855843902,-0.0763717741,-0.0695587024,-0.1448711902,-0.3748391569,-0.0765010566,0.178778246,0.3144549429,-0.2389584929,-0.2472753972,-0.325997144,-0.0508077852,-0.057484556,-0.3587117493,-0.3645914197,0.0336525887,0.0716465265,0.7233041525,0.1904637665,0.0767840371,0.3986012042,0.3073830903,-0.0746853054,0.2320514917,-0.3147574067,-0.3007510602,0.0871328712,0.1018441617,0.1810417771,0.1761606187,0.307376653,-0.1510211825,-0.2707589269,0.0231163166,-0.2049140334,-0.0938397497,-0.1768683791,0.2378317863,-0.2554350495,0.3177197874,-0.2649537921,-0.2518545091,0.3849167228,-0.426133126,-0.180822745,-0.025373213,0.0004500409,-0.1857567877,-0.1347625107,-0.0278676581,-0.0359490737,-0.2764140964,0.2758582234,-0.0604480878,-0.1504130065,0.2112683952,0.0622021034,-0.3640274405,-0.231360361,-0.2156235427,-0.4414393902,-0.8115921617,0.2658408582,-0.1714396626,-0.1759794503,0.0475743227,-0.027079571,-0.0197971426,0.5089735985,0.1174869835,0.102772899,-0.1537671834,-0.0598781295,-0.1125238687,0.2270331979,0.5488132238,-0.2175704837,0.1878561229,0.0694919527,-0.4736878872,0.1309925169,0.3545611501,0.4553125203,0.2023475915,0.6820059419,0.0435426123,0.8500704169,-0.4729242921,-0.0909291655,0.3273846209,0.1331480891,0.2055802047,-0.2246967703,-0.0422127768,0.454071939,0.3447186351,-0.2404646873,0.1776558459,-0.0365534201,0.0319156125,-0.2894175351,-0.040006429,-0.2981526256,-0.07788001,-0.0720067397,0.3880673349,-0.0472228937,-0.0719470754,0.1240249202,0.0330882743,-0.0487786233,-0.0523966812,0.4692560136,0.2600713372,0.0858076215,-0.4809788465,-0.0198217779,-0.3292879164,0.0618244186,-0.0531065427,0.0597123057,-0.0946550146,0.0423378274,0.356790483,0.0577638671,0.5726420879,-0.0670684129,0.2023458481,-0.1834189594,-0.1766930968,-0.2006648332,-0.146684289,-0.1638046801,-0.4664914608,-0.1116445214,0.2392878979,-0.3602339923,-0.0928704292,0.0474609099,-0.2146106958,-0.0480218455,-0.3192707896,-0.2203256488,-0.2238973379,-0.2937514186,0.2858097255,0.1526587605,0.0536992475,-0.1762460917,0.288957566,0.0832172558,0.0491196513,0.1584063917,0.1869951785,0.2496871948,0.2257428765,-0.0272987075,0.5357376933,-0.2613735795,0.3733626008,0.0238520931,0.183850944,-0.1990629286,-0.0252552945,-0.0805901587,0.1963320524,0.0023373866,0.2353169769,0.0922188684,0.2591042519,0.0732418299,-0.3619025648,-0.1329853386,0.1683292389,0.0018058742,-0.510114491,-0.7767041922,0.1530824602,0.3185724616,-0.11471688,0.4655044675,-0.025435226,-0.0361088216,0.329774648,0.019791225,0.7866495252,-0.2761667073,-0.0838119313,0.4054812491,0.2307415456,0.0236884952,0.295725137,0.4031051099,-0.415068835,-0.5347543955,-0.0373035446,-0.0529488064,0.0931367204,0.4020694494,-0.1363642663,0.3450940251,0.0268075075,-0.0504757501,-0.1316199601,-0.000983378,-0.1253965348,0.1095830873,-0.2135537267,0.1441949308,-0.2713958919,0.2783390582,-0.0551924855,-0.0661337525,-0.0537681952,-0.1698430032,-0.1974130571,-0.0751348138,-0.1201173887,-0.2786743939,-0.1620788276,0.0678898022,-0.0184619352,0.1348995119,0.4895372689,0.1435419619,-0.0163704623,0.2112687975,0.0688908026,-0.1089422032,-0.2695263624,0.0802741349,0.1413731873,-0.0909629911,-0.2697945237,-0.2727810144,0.2585124373,-0.7629036307,-0.5136640668,-0.2042930126,0.0638476089,-0.029672496,-0.0455519706,0.1344175041,0.0955742374,-0.0198702738,0.0835156888,0.1212682873,-0.0097375726,0.2349770963,0.1764284372,-0.0745973811,-0.2346412539,0.4504235685,-0.138905704,0.2147396058,0.1396713257,0.0597854629,-0.0464177206,-0.0714687556,0.068506971,0.2950207591,-0.0701668933,-0.0258644242,0.1548923403,-0.0656013489,0.1829087883,0.2000296563,0.0866750032,0.0267215576,0.2052281201,-0.4999912679,-0.2299271226,0.2843564749,-0.2115427256,0.0513736121,0.249633193,-0.021429915,-0.0435978286,0.0905324966,-0.2631253004,0.0788745955,0.175004825,0.2438214123,0.2040696293,-0.1113540754,0.1109043509,-0.065405637,0.1028018892,0.0023243234,-0.1651374102,-0.1198477671,-0.0137525909,0.0889318287,0.1737233549,0.1622076184,-0.1194945499,-0.4061360359,-0.1948967278,-0.1028355286,0.1423201412,0.4355356991,0.061666362,0.1749191433,1.0447347164,0.3954499066,-0.0503818057,0.2874034345,-0.1300049424,-0.2139330357,0.0294342358,-0.1671153158,0.0138625521,-0.0931023732,-0.3125652671,-0.1311164647,0.1889927238,-0.0830208063,0.0892362595,-0.0553422607,0.0761592314,0.3265877962,0.3987318575,0.3013869524,0.2824039161,-0.1422480196,0.437099725,0.1988721341,0.0914352015,0.1177674755,0.278595686,0.4097176492,0.0371400826,0.1517917514,-0.2478079796,-0.1026738882,-0.3526520729,0.0710508823,0.3615995049,-0.4970799983,0.1596379876,0.0683633536,-0.1370611489,-0.1605355889,0.3122987151,0.026089659,0.0397712104,-0.0136185661,0.1464554518,0.345926851,0.1769397408,-0.0540728681,0.1060349718,-0.4014311433,-0.2482421845,0.6673712134,-0.0102960002,0.0948918387,0.1413756013,0.4323264956,0.0061391625,-0.1415293664,0.008223542,0.1333757341,-0.21292831,0.112737678,-0.2970485091,-0.241170004,-0.1083396077,-0.0901389867,-0.210779801,0.0570228472,0.5410450697,-0.043190226,-0.1806109101,-0.363999486,-0.303987354,0.391821444,0.0895706043,0.0930958465,0.0872961804,-0.1694203168,-0.3139250576,0.8918427229,0.2435021698,0.0858810022,0.2433633506,0.0584663637,0.1616476029,-0.3251116872,-0.2513420582,0.085042946,0.1629290283,0.3028137088,-0.0886207446,0.0022017164,0.1729615629,-0.0911633596,0.0102840122,-0.116444096,0.4658021927,-0.2209759504,-0.0021471311,0.0630773157,0.18876037,0.1679881513,0.0536472015,-0.2533687949,0.1768382043,-0.5851315856,-0.0773613006,-0.0601318777,0.0402972661,0.0224634893,0.1051836535,0.1662865728,0.2038718909,-0.3807618916,0.0285576228,-0.3001153171,-0.562564671,0.0181057137,0.1876414567,0.1887489557,-0.1379501373,0.0149770072,0.254432559,0.0273440611,0.0954004973,-0.0867264867,0.0113995522,0.3835161924,-0.1654761285,0.0993808359,-0.10658326,0.0172311347,0.1863534302,-0.0806246325,0.2554670572,0.2830182612,-0.0343926176,-0.0360285901,0.3049871624,-0.2444048226,-0.364435941,-0.177275002,0.1247338429,0.446044147,-0.1615052074,0.1612081975,-0.0301489793,-0.0932284072,-0.569683075,0.0014497569,-0.1189164668,0.3761825264,-0.0181238204,-0.2275947183,-0.3866176605,0.4435469508,-0.2874620855,0.0973358005,-0.4451755583,0.3856310844,0.1680960953,0.0923693851,0.0771509111,0.0387629084,0.0452558435,-0.0640121251,-0.12644656,-0.0576052591,0.6830077171,-0.0980130807,-0.0177716464,-0.585709691,0.030918492,0.0552420542,-0.2197270393,-0.0703961775,0.1641987562,0.0008173359,0.2290299684,0.0857625008,0.3644435406,-0.001277984,-0.1020380706,-0.0343418457,0.2222901583,0.2071916163,-0.1870925874,0.079635039,-0.2017952949]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1610","title":"shuffle does not accept seed ","comments":"Hi Thomas\r\nthanks for reponse, yes, I did checked it, but this does not work for me please see \r\n\r\n```\r\n(internship) rkarimi@italix17:\/idiap\/user\/rkarimi\/dev$ python \r\nPython 3.7.9 (default, Aug 31 2020, 12:42:55) \r\n[GCC 7.3.0] :: Anaconda, Inc. on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import datasets \r\n2020-12-20 01:48:50.766004: W tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:60] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudart.so.11.0: cannot open shared object file: No such file or directory\r\n2020-12-20 01:48:50.766029: I tensorflow\/stream_executor\/cuda\/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\n>>> data = datasets.load_dataset(\"scitail\", \"snli_format\")\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nReusing dataset scitail (\/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/scitail\/snli_format\/1.1.0\/fd8ccdfc3134ce86eb4ef10ba7f21ee2a125c946e26bb1dd3625fe74f48d3b90)\r\n>>> data.shuffle(seed=2)\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nTypeError: shuffle() got an unexpected keyword argument 'seed'\r\n\r\n```\r\n\r\ndatasets version\r\n`datasets 1.1.2 <pip>\r\n`\r\n","body":"Hi\r\nI need to shuffle the dataset, but this needs to be based on epoch+seed to be consistent across the cores, when I pass seed to shuffle, this does not accept seed, could you assist me with this? thanks @lhoestq\r\n ","comment_length":134,"text":"shuffle does not accept seed \nHi\r\nI need to shuffle the dataset, but this needs to be based on epoch+seed to be consistent across the cores, when I pass seed to shuffle, this does not accept seed, could you assist me with this? thanks @lhoestq\r\n \nHi Thomas\r\nthanks for reponse, yes, I did checked it, but this does not work for me please see \r\n\r\n```\r\n(internship) rkarimi@italix17:\/idiap\/user\/rkarimi\/dev$ python \r\nPython 3.7.9 (default, Aug 31 2020, 12:42:55) \r\n[GCC 7.3.0] :: Anaconda, Inc. on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import datasets \r\n2020-12-20 01:48:50.766004: W tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:60] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudart.so.11.0: cannot open shared object file: No such file or directory\r\n2020-12-20 01:48:50.766029: I tensorflow\/stream_executor\/cuda\/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\n>>> data = datasets.load_dataset(\"scitail\", \"snli_format\")\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nReusing dataset scitail (\/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/scitail\/snli_format\/1.1.0\/fd8ccdfc3134ce86eb4ef10ba7f21ee2a125c946e26bb1dd3625fe74f48d3b90)\r\n>>> data.shuffle(seed=2)\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nTypeError: shuffle() got an unexpected keyword argument 'seed'\r\n\r\n```\r\n\r\ndatasets version\r\n`datasets 1.1.2 <pip>\r\n`\r\n","embeddings":[-0.3611408174,-0.2253305167,-0.0523433983,0.1002608612,0.2563570142,-0.1110154986,0.1722029597,0.0852967724,-0.0057133436,0.3612514436,0.144108519,0.3828558326,-0.2792423368,0.4236404002,-0.317094028,0.1824320108,0.1758906841,0.0246543158,0.0966961905,-0.2187662274,-0.2138340175,-0.1767721623,-0.3849220574,-0.392226547,-0.0966786593,0.0997234434,-0.0631705746,-0.0549211204,-0.2333814502,0.0829104856,0.0291087646,0.4695254266,0.2254402041,0.543418169,-0.0001184209,0.031107679,0.1028720811,-0.1222021952,-0.1623677164,-0.2037837803,-0.4092410505,0.7001438737,-0.2987084985,-0.1451266557,-0.4335159957,0.0217807833,-0.0585794561,-0.3505462706,0.2419661433,0.0192649774,0.0743444338,-0.0877100602,0.1874300539,0.045032803,0.3148187399,0.1693404168,-0.0267277192,0.1806753725,0.1170011461,0.141771391,0.1314733326,-0.0136718452,0.0271484889,0.2115646452,0.4446219504,-0.2895041108,0.0597491972,-0.3515516818,-0.022754604,0.2394785285,0.6474811435,-0.1044927686,-0.3328814507,0.027288055,-0.1217833981,-0.7533959746,-0.1434610784,0.2448544502,-0.2657306492,-0.0243245978,0.0012280716,0.0514400005,-0.1524673551,0.06739714,0.1754340529,0.6841983795,0.0115244631,0.1180583239,0.4326361716,-0.0639571697,0.3201989532,0.032385245,-0.057937406,0.1822149307,-0.3543589115,-0.189820081,-0.3696711063,0.1263028532,0.1526526362,0.1987504214,0.0989194363,0.2543503344,-0.0179729294,0.2990306616,-0.0273408722,0.1894387454,0.1938605756,0.1103519574,-0.0203140359,-0.0033378631,0.0162443109,0.0666703433,-0.1197569966,-0.1333681792,-0.2709949315,0.065532282,-0.1789503247,-0.1597332656,-0.1167848408,-0.4980922341,-0.3183217049,-0.383685261,0.4276821613,0.3747380078,-0.1726719737,0.1342258751,0.0090161609,0.2096813023,0.0860612616,-0.2013346702,-0.267860502,0.0493237115,-0.2658182085,0.2053087056,0.0714615434,-0.2550654411,0.2359534204,0.2998144329,0.180451721,-0.0260684006,-0.0820793882,0.0033828055,-0.0434186012,0.0275081135,-0.0385255069,0.0666897446,0.1341362596,-0.0235903002,0.0407774411,0.4544708729,-0.5137068629,-0.0959360525,0.1449231654,0.080949977,-0.031147819,-0.1751510203,-0.5291618109,0.0693709999,0.0980649665,0.0767359361,-0.0567197688,-0.1430936456,-0.2210722417,-0.2535941601,0.4845531285,-0.2019007206,-0.1611191332,-0.0025037813,0.1482612342,0.2340067476,0.3288626075,0.4299624562,-0.0824116766,-0.2007815838,0.0891031995,-0.0327073224,0.2788090706,-0.111896567,-0.2866583467,-0.371278882,0.0071844622,-0.2699796557,-0.0005191008,0.2127477378,0.0968917906,-0.3004891276,0.2491336912,0.4734845161,-0.0651877597,0.1062624678,-0.2993537784,-0.0737838373,0.0894062892,0.5021672845,-0.1099241674,0.0022002284,0.0911847427,-0.1518818885,0.4024082124,-0.0993448198,-0.0192164183,0.0537746586,0.4163772762,0.1382247359,0.0518619418,0.050076507,0.0253946539,0.0496197604,0.4589367807,0.0498924553,0.3490516245,0.0063804165,-0.2450549603,-0.2303178012,-0.2283510715,0.0235861186,-0.0238635521,0.0090243444,0.1661869437,-0.0394502543,-0.1085365787,0.0519927889,-0.5093948841,0.0178662818,-0.3958610892,0.1027719229,-0.0856780186,-0.3175879717,0.0648345202,0.0358296186,0.1814566553,0.0059541785,-0.0573778749,0.3696957231,0.1335140914,-0.2687822282,-0.3451565504,0.2437442392,0.0348926447,0.0060798237,0.0194668155,-0.363627255,0.1999331564,-0.2298474908,0.2633122802,0.3140081465,-0.2565154135,-0.114737764,0.0411611907,-0.2615285218,0.2098861635,0.0645635426,-0.1613117605,0.1803571582,0.3249889314,0.0004540306,0.0708319843,-0.0369634517,-0.3684694767,-0.0874771997,0.2278015167,0.0122382622,-0.1194051877,-0.1054250896,-0.1213169619,0.106050767,0.1194289252,0.2121357769,0.2475169301,0.1334799826,0.0652236938,-0.0148238074,0.1242563277,-0.3287729323,-0.1415364444,-0.1527764797,0.3554347157,-0.1534022093,0.1107402891,-0.0885211974,-0.1478292793,-0.2959602773,0.2944551706,0.4305765033,-0.1968322843,-0.1009759232,-0.3533166647,-0.0418738499,-0.1448265016,-0.3894093931,-0.0849636048,0.0953142941,-0.0746296123,0.7804245353,0.2613789737,0.1790030599,0.4629921913,0.2231993973,-0.0111896936,0.008618732,-0.1971976906,-0.2141981274,0.1154349595,0.094418481,0.309376359,0.1184532344,0.2321745753,-0.0511785261,-0.2368722111,0.1078239232,-0.2652058005,-0.1191261634,-0.0676183403,0.1537304074,-0.5521233082,0.2170722187,-0.2828325033,-0.4466594756,0.3089562058,-0.1998510212,-0.1039057001,0.0575454906,0.0287181959,-0.2272478938,-0.2291841209,-0.1263979375,-0.1747358739,-0.3220356107,0.1404100657,0.1458332539,-0.0682677403,0.438867718,0.278580755,-0.382552892,0.0910977125,-0.2483300716,-0.3010384142,-0.5564925075,0.3111891747,-0.1494143158,-0.1545503885,-0.0209704097,-0.1006934345,0.2367001176,0.7145299315,0.0014578539,0.2914760411,-0.0760957226,-0.148301065,-0.0383760892,0.0375671387,0.5418398976,-0.2448954731,0.2498478144,0.1140450686,-0.1856481433,0.1597103328,0.2866020799,0.4009343386,0.2772111595,0.6139820218,0.1146798804,0.8872416019,-0.323028028,-0.3530137241,0.2038476914,0.0597041175,0.0461004451,-0.1533634067,-0.004342542,0.4281462431,0.4057964981,-0.2340331972,0.0565506779,-0.1704082638,-0.2394576222,-0.369557023,0.0714306682,-0.3473468125,-0.1121447235,0.0824679509,0.1314641386,-0.0287999474,-0.1044372916,0.1211365983,0.0069418154,-0.1989909261,-0.1880779713,0.4354414642,0.0926452056,0.0561244637,-0.3997306228,-0.1434668899,-0.3735197484,0.0937168747,-0.1710083485,0.2613600791,-0.0232575275,0.0718953907,0.3596392274,0.0759317651,0.4168021679,-0.0979146883,0.1123276129,-0.1094888672,-0.0538831279,-0.2615788877,-0.1849402636,-0.1586206257,-0.4113390744,-0.0098012639,0.1744081378,-0.2656899393,0.021301901,-0.1195965856,-0.2482982427,-0.0566112697,-0.1592416167,-0.193607986,-0.2221111804,-0.3232959807,0.0619873628,0.1619882733,0.1035289913,-0.1610773057,0.0532945879,0.057074789,-0.0148694878,0.2875737548,0.2513359487,0.2465616614,0.2588960528,0.1504959762,0.1836815774,-0.3022022545,0.5362427831,-0.1622434109,0.2666838765,-0.2363552153,-0.0117449686,-0.1168003082,0.1234583333,-0.063419424,0.0496844389,0.0545561239,0.0064278771,-0.0274295472,-0.1465011239,-0.0108464779,0.1814292222,0.0382841304,-0.5410442352,-1.0802172422,0.1903682947,0.3028416038,-0.1298493147,0.7586442828,-0.3035853505,-0.0043945094,0.4236359,0.0830999091,0.7637459636,-0.2922925949,-0.203382954,0.4367133975,0.2170633227,0.1630804837,0.0495186821,0.546277225,-0.406116426,-0.4143505394,-0.0269855354,-0.1834913939,0.0943835378,0.4430631101,-0.1552726775,0.2226473689,0.0186110511,-0.1358989179,-0.0702841878,0.0810505748,-0.2048526108,0.0815507695,0.0058770622,0.1108387336,-0.2439632714,0.2806298733,0.0505448915,-0.0571790673,-0.367010802,-0.0800879821,-0.213044703,0.1272124946,-0.1728191078,-0.1039242595,-0.281616956,-0.1353356838,0.0687894002,0.2096081376,0.5077693462,0.1432932317,0.0618176945,0.2626107931,0.0739550069,-0.1254739761,-0.0339673683,-0.040966589,0.204384312,-0.0336338766,-0.2631893456,-0.0979691669,0.3640576899,-0.7248569727,-0.4052668512,-0.1746648848,0.2550791502,0.1793935001,-0.1126201823,0.2795415819,-0.1132468283,0.1159828976,0.0910145044,-0.0799330845,0.0348872617,0.2517882884,-0.0340326987,-0.0980082378,-0.1022260711,0.3270203769,0.1678155363,0.2097698152,0.3061744869,0.0538999513,-0.0496066399,-0.0596302189,-0.1551058292,0.1178596988,-0.0458976403,-0.0972229317,0.0574440025,-0.1233138591,0.3824639022,0.234532997,-0.0655943751,0.2924669087,0.2172478884,-0.5897246003,-0.0509739555,0.1335219592,-0.3107289374,0.1417053938,0.1097967103,0.0991513282,-0.1965574175,-0.050078854,-0.2414623648,-0.0285362508,0.2438211888,0.153517589,0.1518406719,0.030171277,0.0799925253,-0.09929239,0.0824380443,0.1317389011,-0.1655619889,-0.0888633728,0.0432293639,0.0811352134,0.2323162854,0.1018951386,-0.1587355733,-0.5502097607,-0.2331139743,-0.1039926708,0.0548733771,0.4458000064,0.062444143,0.153755039,0.8165636063,0.4594482481,-0.0468276851,0.2826242149,-0.3036560416,-0.1551722884,0.0453600176,-0.0843893215,0.0835873187,-0.028856298,-0.5530840755,-0.24562186,-0.1104248837,-0.0208631102,0.2164385915,-0.111887373,0.1159099713,0.2026201338,0.2151464224,0.297758013,0.1507697552,-0.2387458533,0.385727644,0.1774684489,0.0762745738,0.0942931771,0.0872187316,0.4449030459,0.0488536283,0.2637331486,-0.2752606869,-0.3965362012,-0.3330903351,0.0899172127,0.4578497112,-0.4908415675,0.3078712821,0.1843553782,-0.0155400848,-0.1607227474,0.0816479772,0.0181556493,0.0323112607,0.1464743167,0.0237537008,0.284340173,0.1885379255,-0.1568910778,-0.0847822651,-0.3486177027,-0.291893661,0.7141577601,-0.1767001897,0.2091710418,0.29779163,0.2677813768,0.1328713596,0.0064644394,0.0997213647,0.0216236189,-0.1695972383,0.2545367777,-0.4078198373,-0.1922158599,-0.136377722,-0.1428141296,-0.1083559245,0.0413887799,0.3476500213,0.0592189319,-0.2778981626,-0.0720920935,-0.3138539493,0.4026613832,0.0394624099,0.0047962461,0.1703034341,-0.2112160176,-0.2643462718,0.6609690785,0.0015209117,0.1338514388,0.4175248742,0.1288035363,0.2596217096,-0.0248103943,-0.2193275392,0.2074463665,0.1799837202,0.506182909,0.0677464977,-0.0868817344,0.1744322032,-0.0787593797,0.0454309136,0.0781336501,0.5217003226,-0.267285794,0.0767546296,0.0434467942,0.1062491685,0.1414196789,0.0462317243,-0.330640316,0.1009448841,-0.4752773643,-0.0678268522,0.0255446117,0.0667860955,-0.0065075294,-0.034880396,0.034626238,0.094032675,-0.370652169,-0.0032605727,-0.3976333439,-0.4175258279,0.0983742625,-0.0486312173,0.0690122023,-0.2239977568,0.2311510146,0.2991912365,-0.0187734701,-0.0545410439,0.0081284894,-0.0669468045,0.3860735595,-0.270565629,0.0497124195,-0.370413661,0.0301734768,0.1646143198,-0.1682939976,0.1987192035,0.1453211159,-0.0244066678,-0.0999523103,0.2395856827,-0.0390974022,-0.2174385637,0.0375294276,-0.0112730768,0.3982421756,-0.1108964682,0.102721259,-0.1268592924,-0.1312946081,-0.5119926333,0.085713923,-0.0813269317,0.4255468845,-0.0233871005,-0.2648216188,-0.3979429007,0.4278897345,-0.2156163901,-0.3002790809,-0.2899352312,0.3450368047,0.1262332797,0.3591560125,-0.0169633701,0.0757392421,-0.0092611006,-0.0694122165,-0.1976409405,-0.1824417859,0.6684938669,-0.3057699203,-0.0798934028,-0.4659472704,0.0939518735,-0.0872713029,0.0380332768,-0.0911187306,0.199557662,0.0209581666,0.198491767,0.1439945102,0.222377345,0.1808533818,0.0005039272,-0.0991008803,0.114485465,0.1380208433,-0.3079781234,-0.0292004719,-0.2287420481]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1610","title":"shuffle does not accept seed ","comments":"Thanks for reporting ! \r\n\r\nIndeed it looks like an issue with `suffle` on `DatasetDict`. We're going to fix that.\r\nIn the meantime you can shuffle each split (train, validation, test) separately:\r\n```python\r\nshuffled_train_dataset = data[\"train\"].shuffle(seed=42)\r\n```\r\n","body":"Hi\r\nI need to shuffle the dataset, but this needs to be based on epoch+seed to be consistent across the cores, when I pass seed to shuffle, this does not accept seed, could you assist me with this? thanks @lhoestq\r\n ","comment_length":36,"text":"shuffle does not accept seed \nHi\r\nI need to shuffle the dataset, but this needs to be based on epoch+seed to be consistent across the cores, when I pass seed to shuffle, this does not accept seed, could you assist me with this? thanks @lhoestq\r\n \nThanks for reporting ! \r\n\r\nIndeed it looks like an issue with `suffle` on `DatasetDict`. We're going to fix that.\r\nIn the meantime you can shuffle each split (train, validation, test) separately:\r\n```python\r\nshuffled_train_dataset = data[\"train\"].shuffle(seed=42)\r\n```\r\n","embeddings":[-0.1347804815,-0.2200724334,-0.0721425191,0.0236785691,0.2914151549,0.0666157305,0.1913048029,0.1325144321,-0.1194050908,0.3138949275,0.134742856,0.3167942166,-0.3220253885,0.3731395304,-0.2299949974,0.0033440839,0.1122753695,-0.0262981746,0.1395083368,-0.0770237893,-0.2660176158,-0.1038577631,-0.4424047172,-0.3410610557,-0.2854548097,0.0956010371,-0.1741223484,0.1295313388,-0.2702615559,-0.0985725001,-0.0670805499,0.3628925979,-0.009106623,0.5489265919,-0.0001074182,-0.0270830337,-0.0419341661,-0.1488695443,-0.1034156308,0.0203134511,-0.3263738453,0.5454821587,-0.4160846472,-0.1584909707,-0.4424951673,-0.037389271,-0.1593659818,-0.069139488,0.3702379167,0.110146746,0.2067461908,0.0331087261,0.0294005517,-0.0273064841,0.2497342974,0.0464598499,0.0437573493,0.0516144149,0.0109907668,0.0839249492,-0.0199910924,-0.0127108265,0.2320576608,0.2207108587,0.2102143914,-0.2231942415,0.0342869908,-0.253064394,-0.0665040687,0.1607753336,0.2795592844,-0.1354098469,-0.3697174489,-0.1066208705,-0.0143135777,-0.6579858661,-0.0505812541,0.3136821985,-0.0247074813,0.0323528349,-0.0712482333,0.1344084442,-0.0298695043,-0.0848142356,0.0765891373,0.699716568,0.0725433081,0.0611053966,0.423551172,0.0133899152,0.1375877261,0.0112705929,-0.0726978183,0.1279779822,-0.3236722052,-0.2280355841,-0.2975527644,0.2766510844,0.1149020791,0.1078327671,-0.0889401585,0.2085072398,0.0864677727,0.1739783436,0.1651337892,0.2276163697,0.0355324335,0.3068234026,-0.0154891489,0.0969886109,-0.1259662658,0.1423597634,0.0188584886,-0.2953653634,-0.0575607941,0.1596961915,-0.2405930758,-0.1067030132,-0.2917694747,-0.3292226195,-0.2995627224,-0.2655608356,0.2093190402,0.3032734394,-0.0809738189,0.0591247566,-0.0797601491,0.1136507764,-0.0517724566,-0.2500488758,-0.43158108,0.0078702839,-0.3319699466,0.1614773124,0.0961486921,-0.2158621848,0.3259159029,0.3171091974,0.0427013934,-0.1045705527,-0.0938387662,-0.0080860965,0.0616393201,0.0213751812,0.0275105126,-0.0125426231,0.146946013,-0.0121893724,0.10842143,0.2525421679,-0.3640213013,-0.2342913747,0.4208706021,0.2410050035,-0.1537894607,-0.0020534673,-0.3313791156,0.0717510432,0.1327168792,0.1020409539,-0.1301130503,-0.1859665215,-0.2507869601,-0.1024178416,0.4589520097,-0.1903232634,-0.0930913761,-0.0459462591,0.1666830033,0.0458181128,0.2529376745,0.5030997396,-0.0493666828,-0.0394233875,-0.0561052486,0.0591851585,0.1244835407,0.0668547079,-0.2905207872,-0.2616107166,-0.1367256492,-0.1420549899,-0.096235238,0.0116049051,0.1021196693,-0.2010557652,0.3246784508,0.3828471303,-0.2970031798,0.1307126433,-0.4009949863,0.0544542857,0.1685847044,0.4805800915,0.0144611215,-0.0717338324,-0.1493746787,-0.1211038083,0.4762872159,0.0101417443,-0.029344473,0.0018905971,0.2415610105,0.1914938241,0.1795387268,-0.0784265473,0.2410700917,0.1030149981,0.3657622933,0.0524096079,0.3203825057,-0.1304528713,-0.3920786679,-0.2696313262,-0.3058220148,-0.1071614623,0.1618437171,-0.170145154,0.2359348238,-0.1054814756,-0.0968539566,-0.135586217,-0.2334574312,0.0112656048,-0.3087714911,0.0612226799,-0.135071665,-0.292391181,0.0529397465,0.1395294517,0.102069214,0.0419107638,-0.1187908277,0.409160018,0.3247660398,-0.4278866053,-0.3233824372,0.0829123929,-0.0699155778,0.1035030857,-0.1071636528,-0.0716485009,0.1234953627,-0.2072857916,0.1270310432,0.4635509551,-0.3934854865,-0.1097532064,0.0302342214,-0.1540511549,0.2201206684,0.058988221,-0.2735131383,0.0595620312,0.333858788,-0.2367123961,0.1785191596,-0.0354736522,-0.2713895142,-0.1414781213,0.3098217845,-0.0125956545,-0.1193342432,-0.2523328364,-0.1314130127,0.1176525056,-0.0989534184,0.291110158,0.2756367326,0.2380652279,-0.0021677529,0.0973027647,-0.023164466,-0.2529582083,-0.2076808959,-0.125868991,0.2291233391,0.0015448589,0.0062876199,-0.0961295441,-0.1727616638,-0.0996162519,0.2327681184,0.3319702148,-0.2466630042,-0.1438867301,-0.3990576267,0.1052869111,-0.3117464185,-0.274704963,-0.1441822797,-0.1001659483,0.180894807,0.4816630781,-0.0424368717,0.2302291691,0.2112178504,0.0923637897,0.021320127,0.2123742253,-0.0577039495,-0.2570916414,0.0187981017,0.1791387647,0.2132530212,0.1310414821,0.3282648325,-0.09263096,-0.2903418541,0.1300692111,-0.151954487,-0.1913879663,-0.181830138,0.1780179441,-0.330005765,0.1186374649,-0.1865497977,-0.3050548732,0.3582506776,-0.2353467494,-0.1327146739,0.0911906138,0.1785441935,-0.175164476,-0.1757030934,-0.2797611654,-0.1637222916,-0.2218892872,0.1413851082,-0.126843065,-0.126042828,0.1422865391,0.0657179356,-0.4844720066,-0.0367817581,-0.3342334628,-0.2686417401,-0.4810955822,0.1757566631,-0.1975926012,-0.1893901229,0.0030248165,-0.0851678997,0.0662170053,0.4775412977,0.0966284424,0.2267088592,-0.1116108,-0.1041982174,-0.1159769818,0.0072134668,0.4805997908,-0.1754934341,0.1418778747,0.0142328534,-0.2586209774,0.2727104425,0.2904882729,0.3311959207,-0.0986543447,0.4549436271,0.1734048277,0.9153655171,-0.234760657,-0.0158350207,0.074536182,0.0564704128,-0.0724040046,-0.1511095315,-0.0849986374,0.3989588022,0.2620506287,-0.3004077375,0.1801551282,-0.064949438,-0.0543625131,-0.1536415517,0.0852355212,-0.1919069886,-0.1828248352,0.0762912482,0.0657035708,0.0463273562,-0.0170308538,0.2455512881,0.1083478257,-0.0097480258,-0.1477893591,0.1956631243,0.1891005933,-0.0157258734,-0.441832006,-0.263048768,-0.1869140267,0.0048825708,-0.034709923,0.1843194067,0.0014082303,-0.1263059974,0.236039564,0.0822398141,0.5527573824,-0.0894096196,0.2011180818,0.0401285514,-0.0949510038,-0.2554173768,-0.4019107521,-0.2634103894,-0.3782823384,-0.0155929644,0.2363538742,-0.2493738383,0.002353505,0.0031834585,-0.0522731878,-0.0939383879,-0.2544519603,-0.2734539509,-0.1572751105,-0.2095811963,0.3014733493,0.245157823,0.0709425956,-0.3887239695,0.2107310593,0.0932001919,0.0424989648,0.1839707047,0.2184982002,0.302426964,0.0780557543,0.2418813258,0.2548441589,-0.2479174435,0.3776123524,0.0538217202,0.0363324843,-0.1269844025,-0.0503967032,-0.1830578744,0.0775400475,0.039892029,0.1088717803,0.003435547,0.1038621366,0.1826262325,-0.3009642363,0.1088776812,0.2284180969,-0.0530878566,-0.4583853781,-0.9026972651,-0.0421940275,0.2998148799,-0.0727342367,0.4875212014,-0.127234295,-0.0758399293,0.5407730937,-0.033857327,0.5933590531,-0.2362514138,-0.0324080586,0.363093853,0.2730620205,-0.0930330828,0.1895392686,0.5196412206,-0.3974256814,-0.5951387286,0.0554496907,-0.0291510765,0.01991928,0.4732430577,-0.1833691448,0.3339941204,0.12418174,-0.0870800391,-0.209052369,0.0343970358,-0.0963270664,0.1541951895,-0.1947224289,0.2323649973,-0.0871942639,0.2187417448,0.0684457719,-0.0216793977,-0.2913120091,-0.2065157145,-0.1912405789,0.0320260972,-0.0968425348,-0.0009727283,-0.2653569877,-0.0049357871,-0.0882458985,0.0100512719,0.3758583367,0.2907517552,0.0881950334,0.2946348786,0.2940891385,-0.0657177493,-0.0618017986,-0.0307536237,0.2619324327,-0.0523355044,-0.2637942433,-0.2150695175,0.1631853133,-0.6124819517,-0.4894150496,-0.1502243429,0.0032472343,-0.1396195441,-0.1096203998,0.1575693488,0.1193602309,-0.1767693311,0.1998736262,0.0432941876,-0.1546570957,0.4796597958,0.0211599544,-0.1229075119,-0.1563744694,0.2903960347,0.028807763,0.2739140987,0.3979180455,0.0208079014,0.0969830006,-0.2394681871,0.0971226692,0.2445236742,-0.037060488,0.1178652048,0.0201305635,0.0256518181,0.2794482708,0.168609038,0.0359277874,0.1518794894,0.1496250778,-0.359816283,-0.1710107923,0.1275084019,0.0513410084,0.2123155445,0.0412966385,0.3212027252,-0.2519981563,0.1561409831,-0.3886310458,-0.0051879189,0.0066387397,0.2691411972,-0.0974374563,-0.1446208954,-0.0377121344,-0.0845918208,0.228706494,0.0854053348,-0.1447926462,-0.2572974861,0.0149831157,0.0022737312,0.1695061028,0.1784314066,-0.0713163316,-0.2094126791,-0.3208102584,-0.1243991405,0.1283066422,0.3556047678,0.0702229664,0.3206741512,0.8912523985,0.4156518281,0.1616137028,0.2012948245,-0.2163653523,-0.2438079119,0.0597872734,-0.057931494,0.0938092396,0.0217382386,-0.3837366104,-0.2344218343,-0.0374069288,0.0267625991,0.1878928989,-0.2423620075,0.1023165137,0.1466929764,0.5108510256,0.472543329,0.1019658893,-0.0770142153,0.4637640119,0.3311007023,-0.0307503678,0.1320494562,0.2979548872,0.3410888314,0.1816857606,0.1704537421,-0.349471122,-0.1290711761,-0.4043290913,0.1619679332,0.2966355979,-0.4832224548,0.2506630719,0.373759985,-0.0283314418,-0.1797098815,0.0534491204,-0.2817390263,0.056112729,0.0431374684,0.1456333548,0.4172185361,0.1747671962,-0.1647617072,0.0128790466,-0.3011213541,-0.3469181657,0.5019988418,-0.155496031,0.1566528976,0.1851504594,0.381993711,-0.2040082514,-0.2225101143,-0.0756685734,0.0416096672,-0.1850988567,0.1485219747,-0.2696807086,-0.1130308211,-0.1673610359,-0.0102916313,-0.1045152768,0.1405149847,0.3648936749,-0.0483915657,-0.2564620078,-0.2033610195,-0.366635412,0.3606003821,0.1531335562,-0.0168159045,-0.0943886787,-0.0686606988,-0.3091454804,0.8062518239,-0.0120763117,0.1151587293,0.2985970974,-0.1529327631,0.1008982286,-0.1636233777,-0.2112581879,0.04956007,0.2045854032,0.3445229232,0.0160698518,0.2288042605,0.2510470748,-0.165821299,0.0390937254,0.0174412485,0.4443405271,-0.2231723219,0.2361667603,0.2337810844,0.0554710068,0.1201831996,-0.0006375272,-0.2975916266,0.1300688833,-0.5377172828,0.0589534193,0.0563983843,-0.0056589958,0.0871114805,0.0090976423,0.3264142275,0.0765326545,-0.3135475516,-0.0910588726,-0.2488686442,-0.5535826683,0.0051725544,-0.152915597,0.0167994443,-0.1696963906,0.1382557452,0.1882765442,0.0613843016,0.0348909311,-0.0258202385,-0.0900949761,0.5307045579,-0.3389757276,0.171340853,-0.1916727573,-0.03225936,0.2219179124,-0.1237616837,0.1312693208,0.2837960422,0.0880220979,0.1741428971,0.0920335725,0.051997453,-0.124101907,-0.0937301144,0.0602673329,0.3613370657,-0.1068921611,0.2886893749,-0.2155925333,0.0198509488,-0.6188889146,-0.0410012379,0.0147861568,0.4648216367,0.0527441651,-0.2882517576,-0.3264468312,0.4086205363,-0.1227190942,-0.1337548941,-0.2514201701,0.3094482124,0.3073427379,0.2038547844,0.0057368767,0.2329697311,-0.037258096,-0.0484105945,-0.0549767129,-0.1712824851,0.6249601245,-0.1791591644,-0.1430870444,-0.532391727,0.0688989684,-0.0192841403,-0.0696745738,-0.0515775494,0.0550968423,0.001222825,0.1214959249,-0.0043943678,0.2406889945,0.0904304683,-0.0547527149,-0.009614775,0.1765805185,0.1083490029,-0.3192111254,0.1146696582,-0.2614700496]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1609","title":"Not able to use 'jigsaw_toxicity_pred' dataset","comments":"Hi @jassimran,\r\nThe `jigsaw_toxicity_pred` dataset has not been released yet, it will be available with version 2 of `datasets`, coming soon.\r\nYou can still access it by installing the master (unreleased) version of datasets directly :\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`\r\nPlease let me know if this helps","body":" When trying to use jigsaw_toxicity_pred dataset, like this in a [colab](https:\/\/colab.research.google.com\/drive\/1LwO2A5M2X5dvhkAFYE4D2CUT3WUdWnkn?usp=sharing):\r\n```\r\nfrom datasets import list_datasets, list_metrics, load_dataset, load_metric\r\n\r\nds = load_dataset(\"jigsaw_toxicity_pred\")\r\n```\r\n \r\nI see below error:\r\n\r\n> FileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 280 raise FileNotFoundError(\r\n 281 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 282 combined_path, github_file_path, file_path\r\n 283 )\r\n 284 )\r\n\r\nFileNotFoundError: Couldn't find file locally at jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py","comment_length":46,"text":"Not able to use 'jigsaw_toxicity_pred' dataset\n When trying to use jigsaw_toxicity_pred dataset, like this in a [colab](https:\/\/colab.research.google.com\/drive\/1LwO2A5M2X5dvhkAFYE4D2CUT3WUdWnkn?usp=sharing):\r\n```\r\nfrom datasets import list_datasets, list_metrics, load_dataset, load_metric\r\n\r\nds = load_dataset(\"jigsaw_toxicity_pred\")\r\n```\r\n \r\nI see below error:\r\n\r\n> FileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 280 raise FileNotFoundError(\r\n 281 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 282 combined_path, github_file_path, file_path\r\n 283 )\r\n 284 )\r\n\r\nFileNotFoundError: Couldn't find file locally at jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\nHi @jassimran,\r\nThe `jigsaw_toxicity_pred` dataset has not been released yet, it will be available with version 2 of `datasets`, coming soon.\r\nYou can still access it by installing the master (unreleased) version of datasets directly :\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`\r\nPlease let me know if this helps","embeddings":[-0.1765043288,-0.07363718,-0.0608391762,0.2410436869,0.3765500188,0.2648481727,0.1960246563,0.0077844937,-0.1372358799,0.1830359697,-0.2765671909,0.3002050519,-0.0954562724,0.0872111022,-0.1608526856,0.1532540321,0.0479022898,0.2332412452,-0.1677392572,0.2624707222,-0.1808016002,0.1098075956,-0.2976607084,0.1541490108,-0.4638727307,-0.1733397245,0.0785701647,0.4033997655,-0.1453522146,-0.3795202076,0.5730249286,-0.0468719266,0.1658125073,0.7571409941,-0.0001131657,0.3079266548,0.403950721,-0.07764256,-0.3698799014,-0.5823313594,-0.5913619995,-0.1218554154,0.1567143202,-0.1497648954,-0.3388918042,-0.0401686765,0.1544191539,-0.3331000805,0.3245493472,0.3865208328,0.2066322863,0.3872398436,0.1222632974,-0.3384933174,-0.0258551668,0.2666294873,-0.1972867548,0.6044800878,0.3443886936,-0.043817237,0.2820273042,0.0967959613,-0.0518094562,0.1000471711,0.0944557488,0.0319421031,-0.14020437,-0.3953693509,0.3337980807,0.4867743552,0.2882891893,-0.3954355717,-0.2368726581,-0.1654328406,-0.0092283944,-0.2622863352,-0.0164637957,0.0009507873,0.0711009502,0.1785696,-0.02291191,0.1373260766,-0.1632054299,0.2048930526,0.2333356142,0.3115946352,-0.2750920057,0.0986382738,0.1140073687,-0.4757558703,0.1821375787,0.053063944,-0.055582568,0.0987114236,-0.3599222004,0.1672172993,-0.1756782979,0.2006390989,0.2447078824,-0.4047555923,0.0635226369,-0.154194504,-0.0503103919,0.0883153006,0.1189346239,0.0482748933,0.144855082,0.4108527601,0.2634599507,0.5451515913,-0.2810816169,-0.1524787694,-0.1094568223,-0.2906466126,-0.0365780257,-0.2321513444,0.468872577,-0.3515856564,-0.4451787174,-0.3080203533,-0.1883712411,0.1652941406,0.3322097659,0.4939045906,0.1570606083,-0.1572115421,-0.0750433728,0.3428778648,-0.0558868386,0.1293441951,-0.2999328971,0.2289604694,-0.2291098982,0.0647563487,0.1287426651,-0.4026706815,0.4692715704,-0.1230796278,0.1983292699,-0.0562339425,-0.1611506492,-0.1195021868,-0.1380105764,0.248663187,0.0260710046,0.0224840399,0.0836298689,-0.0628136694,-0.0675119013,0.0591872223,-0.3778591454,-0.4255342782,0.0368446857,0.1483312547,-0.4374786913,-0.1637428552,0.0553643554,-0.1182392314,0.1565728635,-0.1337582022,-0.1645391881,-0.0758990571,-0.4998377264,-0.2826627493,0.1935234666,0.4830069244,-0.2996079922,0.042469319,-0.3080744743,-0.0867858678,0.1806526482,0.1393012106,-0.3239135444,0.0703095943,-0.2412335426,-0.0015097894,0.3545083404,-0.5732682943,-0.2533680201,0.1079819202,0.0063852407,0.0841249228,0.1747066528,-0.3414479792,-0.2044880688,-0.1492318958,0.1607513875,0.0345812254,-0.0251700934,-0.1039276347,0.0745938495,-0.300321579,-0.0428007878,0.2549883425,-0.0014975984,0.178963989,0.337190181,-0.0854072347,0.1417766362,-0.1695163995,-0.0791049078,0.1477522999,0.004492451,0.0200523064,-0.0354111455,-0.0347735882,-0.2954501212,0.2822983265,-0.1452663541,-0.0098495539,-0.3215234876,-0.061694432,-0.5691983104,0.0971442834,-0.2326286733,-0.0552996621,0.1148718297,0.0202192422,0.0741456002,0.2099332213,-0.0664540157,0.303404808,-0.1201207116,0.0736225247,-0.3704358041,0.1775779873,-0.1285678893,-0.3357838392,0.1636983454,0.2939181924,0.2214094847,-0.1230493411,-0.0515593998,0.2560582161,0.007353568,0.0841370597,0.2337305546,0.2524306178,0.1431983262,-0.1125041172,0.2018995136,-0.0870922133,0.3635498285,-0.0120474827,0.1510169804,0.401984036,-0.2108748257,0.2795546353,0.0922547653,-0.0128316488,0.2691123188,-0.0919803232,0.0392131172,-0.3684935868,0.1300258487,-0.0170000475,0.1427501142,0.118689537,-0.3156521916,0.0353442915,0.1817990243,0.0676587746,0.026217714,0.005977266,-0.1093193814,0.0335148275,0.2066028118,0.076037623,0.5343503952,0.2244418263,0.0346319638,0.073210299,0.032673452,-0.3638688922,0.0351182409,0.1861788481,0.4019531012,0.1055076569,0.179616183,-0.1125291809,0.0497633927,-0.0733935162,0.0133888135,0.3247944713,-0.2611782849,0.4062215984,-0.6884021759,-0.849063158,0.0187445264,-0.1318324208,-0.0035038979,-0.1434836835,-0.2425691336,-0.0528350025,0.0637289807,0.2590059638,0.0430169031,0.0015332315,-0.1681162566,-0.2226403505,-0.0675634071,-0.2243787944,-0.2410758287,0.1178557202,0.2741014063,0.0301817339,0.3007182777,-0.2623437047,-0.0789944902,-0.1063176543,-0.0491989404,0.1852882355,0.0297127925,0.5172261,0.2643103302,0.0573489554,-0.232159853,-0.3317301571,0.3848645389,-0.230599001,0.0403684676,-0.0265582688,-0.4671980143,-0.0780436993,0.0726623535,0.0134151634,-0.3836221397,-0.3464674652,-0.2247263789,0.1520386636,-0.0058445763,0.4043499231,0.1393468082,0.3607227504,-0.0079633249,0.1871718466,-0.1175756231,-0.7591186166,0.3459428549,-0.1522512734,-0.1781123281,0.2786597908,-0.1888881624,0.3604364097,0.0063017262,-0.2204868048,-0.3749679625,-0.0712205693,0.0595394559,0.2680722177,0.1431832612,-0.0766409934,-0.2826042175,0.0295414124,-0.2402922511,-0.3323746324,0.1549118608,-0.0324948691,0.4485782683,0.2557002902,0.2735060155,0.0265375283,0.5264333487,0.2063240707,-0.310410291,0.560850203,-0.2335514724,0.5399410129,-0.2854922712,-0.3986561596,-0.0295694638,0.2363219857,0.1110223532,0.0497976281,0.0686873496,0.0622299872,-0.3711649477,-0.2727765441,-0.4604558349,-0.22083731,-0.1248241067,0.0627700239,0.0486788563,0.1545646787,-0.1930720061,-0.0020582129,-0.2547899485,-0.1212504581,0.5862550735,0.0754125938,0.0770373046,0.4145714343,-0.0855188742,-0.403793782,0.4634940922,-0.0040488704,0.194873333,-0.1834444702,0.0782377869,0.1105068475,0.058107283,0.3156017661,-0.1711797863,0.2286424637,0.2236150652,-0.3791867495,-0.1125552505,-0.0982640684,-0.038982287,0.2180468291,0.3534886539,0.0354411341,-0.1983342618,-0.1873654425,0.4676233828,0.1350232065,-0.0776861683,-0.2284842432,-0.3260590434,-0.3493535817,-0.3408049047,-0.283942312,0.0043522064,0.2995140254,0.2653265893,0.0243891422,-0.1407870203,-0.1731911153,0.0951183289,0.3628186285,0.2854585946,0.0919266492,0.208934024,0.2666968107,-0.0291621573,0.4811285734,0.7712754607,-0.1457217485,-0.6279449463,-0.1264186949,0.2441602945,0.1976390034,0.2365232855,-0.1961033344,-0.0069632921,0.0578687005,-0.1744995564,0.0216790345,0.3970034122,0.186040923,-0.0640664175,-0.5427675843,-0.3402173519,0.2356946915,0.1827408671,-0.0238550305,0.190159753,0.1988546848,-0.2527967691,0.2567949593,0.0346994735,0.7595465183,-0.1744944304,-0.0428739153,0.1884584427,0.1158021986,0.5992568135,0.1882604957,-0.0707108155,-0.2075582594,-0.2978479266,0.1058532,-0.1279861331,0.3986432254,0.0456813648,-0.1773797274,0.5035871267,-0.1521178484,0.1690407097,-0.0002238388,0.2144718021,0.2041130066,-0.1077796593,-0.2365299463,0.1485313028,-0.0898998529,0.1902943403,-0.2229714245,-0.1885854006,-0.2068694681,-0.0418122523,-0.3231393695,-0.013327688,-0.2607429624,0.0772032067,-0.0433270223,-0.4872079194,0.1385264397,0.2872037888,0.2798929214,0.1584499031,-0.236288026,0.2044305801,0.0388747714,-0.3264046609,-0.3186455965,0.0816777721,0.3960490823,-0.0962193981,-0.199808076,-0.0842521638,0.1123977676,-0.210931778,0.0993198678,-0.0172653906,-0.3980234563,-0.2877750099,-0.2272575647,-0.1440537572,0.2280877531,-0.1238338873,0.1092354283,0.3238578737,-0.1020536944,-0.0421350151,0.0472226143,-0.3129553795,-0.1615106612,0.2259002626,-0.0680491328,0.1246451885,0.2166646272,0.1262734234,-0.1239750311,-0.1283407658,0.015923949,0.110541746,0.0809169412,0.1783045381,0.3882919252,0.1421783566,0.04564346,0.1837655604,0.2708903551,-0.2175178528,0.0005880228,-0.6009100676,-0.3271937072,0.0152038941,-0.1389222592,0.1773627847,-0.1620715261,0.3467495441,0.0747586936,-0.0876115859,-0.2790808082,0.1054969802,-0.2524275482,-0.0330542363,0.3787266314,0.0321678743,0.1882181913,0.0604914092,0.1364929974,-0.0075695026,-0.1669653654,-0.2190154642,-0.0968082175,0.1552595794,0.0360927582,-0.324672997,-0.1140930653,-0.2800056934,-0.1091171578,-0.1485867053,0.2010942996,0.2201131284,0.0697049573,-0.0142667945,0.2874352336,0.0023168079,0.182918787,0.2478019446,0.022718478,0.1067924201,0.2279693484,0.0498295277,0.079078123,-0.0503811948,0.4414775372,-0.2765816152,-0.1662036628,-0.0622410774,0.266651839,-0.0797385871,0.0001466705,0.1708437353,0.2072013468,0.5527578592,-0.0051489123,0.2276041508,0.2421133071,0.0600854233,-0.268250078,0.0043666405,0.2005743682,0.2052980214,0.0675548762,0.3758973777,0.1295620948,0.2704569101,0.0920095593,-0.0519322865,0.4057397842,0.0340241082,0.2595987916,0.0937975422,-0.0268633459,0.0012690169,0.2760650516,0.1775625646,0.3242695034,0.2217540592,-0.3799788952,0.4237817526,-0.17972745,0.24527134,0.1152561754,-0.4413736761,0.2160060108,0.1960758865,-0.1192201898,-0.1913729608,-0.018545581,0.2208849788,-0.2242123038,0.1731158644,-0.1550617814,0.1390928775,-0.1601520181,-0.1405973136,0.0376081169,-0.2851132751,0.0589855872,-0.1458445489,-0.0228641834,-0.1104943901,-0.1318608224,-0.1499732882,-0.158703208,-0.0365961529,0.2429273725,0.304975003,-0.0171866007,0.0441878997,0.3076301813,0.0726316795,0.0201910324,0.2118661851,0.360176146,0.6329972148,0.1458148658,0.0199772157,-0.0067221224,-0.1254881769,-0.2086032331,-0.0578629896,0.403976202,0.0339972638,0.1046539545,0.3079672754,0.1693718284,-0.169427827,0.0600415133,0.0038525024,0.0138388844,-0.1192600429,0.0365674943,-0.0875124335,0.0739951581,-0.0295899156,0.1677485406,-0.3000261486,0.3487210572,0.1864134669,-0.291365236,0.2167501152,-0.0426208079,0.0767088905,0.1015828401,0.4885877967,0.1608777046,0.3075758219,-0.3083042502,-0.0974685699,-0.6733787656,0.0455199033,-0.0631597936,-0.1657150388,0.0463938452,-0.1422862113,-0.2273872048,0.1067602038,0.3253476024,0.0365948454,0.2370185554,-0.0189558994,-0.179011479,-0.4035907388,-0.0597984195,-0.4032366574,0.0397742689,-0.0347359814,-0.0172816925,-0.2817715406,-0.0508392863,-0.1897322536,-0.183355093,-0.0974668637,-0.1760172993,0.2545200288,0.1525194198,0.4518975019,-0.0744162723,0.2492251247,-0.3362928927,-0.4395931661,-0.3038597703,0.2104604393,0.1770783067,0.0404433794,-0.3705193102,-0.1775516123,-0.2045305371,-0.1834138185,-0.1058307365,-0.1154196486,-0.1602137983,0.1122977138,-0.1001882777,0.0281389523,-0.0156753343,0.2281958908,-0.1216267422,0.1054669097,-0.0954061449,-0.1676407009,0.7022178769,0.051288873,-0.061452616,-0.1471153051,-0.0944868326,-0.0570841096,-0.1484263837,-0.4888674617,0.2672984302,0.3948301971,0.041690439,-0.108330667,0.2988126576,-0.1936224699,0.0061627477,0.036324203,0.0121704927,0.0570545085,-0.1334373951,0.027306553,-0.0796211734]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1609","title":"Not able to use 'jigsaw_toxicity_pred' dataset","comments":"Thanks.That works for now.","body":" When trying to use jigsaw_toxicity_pred dataset, like this in a [colab](https:\/\/colab.research.google.com\/drive\/1LwO2A5M2X5dvhkAFYE4D2CUT3WUdWnkn?usp=sharing):\r\n```\r\nfrom datasets import list_datasets, list_metrics, load_dataset, load_metric\r\n\r\nds = load_dataset(\"jigsaw_toxicity_pred\")\r\n```\r\n \r\nI see below error:\r\n\r\n> FileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 280 raise FileNotFoundError(\r\n 281 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 282 combined_path, github_file_path, file_path\r\n 283 )\r\n 284 )\r\n\r\nFileNotFoundError: Couldn't find file locally at jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py","comment_length":4,"text":"Not able to use 'jigsaw_toxicity_pred' dataset\n When trying to use jigsaw_toxicity_pred dataset, like this in a [colab](https:\/\/colab.research.google.com\/drive\/1LwO2A5M2X5dvhkAFYE4D2CUT3WUdWnkn?usp=sharing):\r\n```\r\nfrom datasets import list_datasets, list_metrics, load_dataset, load_metric\r\n\r\nds = load_dataset(\"jigsaw_toxicity_pred\")\r\n```\r\n \r\nI see below error:\r\n\r\n> FileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nFileNotFoundError Traceback (most recent call last)\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 280 raise FileNotFoundError(\r\n 281 \"Couldn't find file locally at {}, or remotely at {} or {}\".format(\r\n--> 282 combined_path, github_file_path, file_path\r\n 283 )\r\n 284 )\r\n\r\nFileNotFoundError: Couldn't find file locally at jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/jigsaw_toxicity_pred\/jigsaw_toxicity_pred.py\nThanks.That works for now.","embeddings":[-0.1765043288,-0.07363718,-0.0608391762,0.2410436869,0.3765500188,0.2648481727,0.1960246563,0.0077844937,-0.1372358799,0.1830359697,-0.2765671909,0.3002050519,-0.0954562724,0.0872111022,-0.1608526856,0.1532540321,0.0479022898,0.2332412452,-0.1677392572,0.2624707222,-0.1808016002,0.1098075956,-0.2976607084,0.1541490108,-0.4638727307,-0.1733397245,0.0785701647,0.4033997655,-0.1453522146,-0.3795202076,0.5730249286,-0.0468719266,0.1658125073,0.7571409941,-0.0001131657,0.3079266548,0.403950721,-0.07764256,-0.3698799014,-0.5823313594,-0.5913619995,-0.1218554154,0.1567143202,-0.1497648954,-0.3388918042,-0.0401686765,0.1544191539,-0.3331000805,0.3245493472,0.3865208328,0.2066322863,0.3872398436,0.1222632974,-0.3384933174,-0.0258551668,0.2666294873,-0.1972867548,0.6044800878,0.3443886936,-0.043817237,0.2820273042,0.0967959613,-0.0518094562,0.1000471711,0.0944557488,0.0319421031,-0.14020437,-0.3953693509,0.3337980807,0.4867743552,0.2882891893,-0.3954355717,-0.2368726581,-0.1654328406,-0.0092283944,-0.2622863352,-0.0164637957,0.0009507873,0.0711009502,0.1785696,-0.02291191,0.1373260766,-0.1632054299,0.2048930526,0.2333356142,0.3115946352,-0.2750920057,0.0986382738,0.1140073687,-0.4757558703,0.1821375787,0.053063944,-0.055582568,0.0987114236,-0.3599222004,0.1672172993,-0.1756782979,0.2006390989,0.2447078824,-0.4047555923,0.0635226369,-0.154194504,-0.0503103919,0.0883153006,0.1189346239,0.0482748933,0.144855082,0.4108527601,0.2634599507,0.5451515913,-0.2810816169,-0.1524787694,-0.1094568223,-0.2906466126,-0.0365780257,-0.2321513444,0.468872577,-0.3515856564,-0.4451787174,-0.3080203533,-0.1883712411,0.1652941406,0.3322097659,0.4939045906,0.1570606083,-0.1572115421,-0.0750433728,0.3428778648,-0.0558868386,0.1293441951,-0.2999328971,0.2289604694,-0.2291098982,0.0647563487,0.1287426651,-0.4026706815,0.4692715704,-0.1230796278,0.1983292699,-0.0562339425,-0.1611506492,-0.1195021868,-0.1380105764,0.248663187,0.0260710046,0.0224840399,0.0836298689,-0.0628136694,-0.0675119013,0.0591872223,-0.3778591454,-0.4255342782,0.0368446857,0.1483312547,-0.4374786913,-0.1637428552,0.0553643554,-0.1182392314,0.1565728635,-0.1337582022,-0.1645391881,-0.0758990571,-0.4998377264,-0.2826627493,0.1935234666,0.4830069244,-0.2996079922,0.042469319,-0.3080744743,-0.0867858678,0.1806526482,0.1393012106,-0.3239135444,0.0703095943,-0.2412335426,-0.0015097894,0.3545083404,-0.5732682943,-0.2533680201,0.1079819202,0.0063852407,0.0841249228,0.1747066528,-0.3414479792,-0.2044880688,-0.1492318958,0.1607513875,0.0345812254,-0.0251700934,-0.1039276347,0.0745938495,-0.300321579,-0.0428007878,0.2549883425,-0.0014975984,0.178963989,0.337190181,-0.0854072347,0.1417766362,-0.1695163995,-0.0791049078,0.1477522999,0.004492451,0.0200523064,-0.0354111455,-0.0347735882,-0.2954501212,0.2822983265,-0.1452663541,-0.0098495539,-0.3215234876,-0.061694432,-0.5691983104,0.0971442834,-0.2326286733,-0.0552996621,0.1148718297,0.0202192422,0.0741456002,0.2099332213,-0.0664540157,0.303404808,-0.1201207116,0.0736225247,-0.3704358041,0.1775779873,-0.1285678893,-0.3357838392,0.1636983454,0.2939181924,0.2214094847,-0.1230493411,-0.0515593998,0.2560582161,0.007353568,0.0841370597,0.2337305546,0.2524306178,0.1431983262,-0.1125041172,0.2018995136,-0.0870922133,0.3635498285,-0.0120474827,0.1510169804,0.401984036,-0.2108748257,0.2795546353,0.0922547653,-0.0128316488,0.2691123188,-0.0919803232,0.0392131172,-0.3684935868,0.1300258487,-0.0170000475,0.1427501142,0.118689537,-0.3156521916,0.0353442915,0.1817990243,0.0676587746,0.026217714,0.005977266,-0.1093193814,0.0335148275,0.2066028118,0.076037623,0.5343503952,0.2244418263,0.0346319638,0.073210299,0.032673452,-0.3638688922,0.0351182409,0.1861788481,0.4019531012,0.1055076569,0.179616183,-0.1125291809,0.0497633927,-0.0733935162,0.0133888135,0.3247944713,-0.2611782849,0.4062215984,-0.6884021759,-0.849063158,0.0187445264,-0.1318324208,-0.0035038979,-0.1434836835,-0.2425691336,-0.0528350025,0.0637289807,0.2590059638,0.0430169031,0.0015332315,-0.1681162566,-0.2226403505,-0.0675634071,-0.2243787944,-0.2410758287,0.1178557202,0.2741014063,0.0301817339,0.3007182777,-0.2623437047,-0.0789944902,-0.1063176543,-0.0491989404,0.1852882355,0.0297127925,0.5172261,0.2643103302,0.0573489554,-0.232159853,-0.3317301571,0.3848645389,-0.230599001,0.0403684676,-0.0265582688,-0.4671980143,-0.0780436993,0.0726623535,0.0134151634,-0.3836221397,-0.3464674652,-0.2247263789,0.1520386636,-0.0058445763,0.4043499231,0.1393468082,0.3607227504,-0.0079633249,0.1871718466,-0.1175756231,-0.7591186166,0.3459428549,-0.1522512734,-0.1781123281,0.2786597908,-0.1888881624,0.3604364097,0.0063017262,-0.2204868048,-0.3749679625,-0.0712205693,0.0595394559,0.2680722177,0.1431832612,-0.0766409934,-0.2826042175,0.0295414124,-0.2402922511,-0.3323746324,0.1549118608,-0.0324948691,0.4485782683,0.2557002902,0.2735060155,0.0265375283,0.5264333487,0.2063240707,-0.310410291,0.560850203,-0.2335514724,0.5399410129,-0.2854922712,-0.3986561596,-0.0295694638,0.2363219857,0.1110223532,0.0497976281,0.0686873496,0.0622299872,-0.3711649477,-0.2727765441,-0.4604558349,-0.22083731,-0.1248241067,0.0627700239,0.0486788563,0.1545646787,-0.1930720061,-0.0020582129,-0.2547899485,-0.1212504581,0.5862550735,0.0754125938,0.0770373046,0.4145714343,-0.0855188742,-0.403793782,0.4634940922,-0.0040488704,0.194873333,-0.1834444702,0.0782377869,0.1105068475,0.058107283,0.3156017661,-0.1711797863,0.2286424637,0.2236150652,-0.3791867495,-0.1125552505,-0.0982640684,-0.038982287,0.2180468291,0.3534886539,0.0354411341,-0.1983342618,-0.1873654425,0.4676233828,0.1350232065,-0.0776861683,-0.2284842432,-0.3260590434,-0.3493535817,-0.3408049047,-0.283942312,0.0043522064,0.2995140254,0.2653265893,0.0243891422,-0.1407870203,-0.1731911153,0.0951183289,0.3628186285,0.2854585946,0.0919266492,0.208934024,0.2666968107,-0.0291621573,0.4811285734,0.7712754607,-0.1457217485,-0.6279449463,-0.1264186949,0.2441602945,0.1976390034,0.2365232855,-0.1961033344,-0.0069632921,0.0578687005,-0.1744995564,0.0216790345,0.3970034122,0.186040923,-0.0640664175,-0.5427675843,-0.3402173519,0.2356946915,0.1827408671,-0.0238550305,0.190159753,0.1988546848,-0.2527967691,0.2567949593,0.0346994735,0.7595465183,-0.1744944304,-0.0428739153,0.1884584427,0.1158021986,0.5992568135,0.1882604957,-0.0707108155,-0.2075582594,-0.2978479266,0.1058532,-0.1279861331,0.3986432254,0.0456813648,-0.1773797274,0.5035871267,-0.1521178484,0.1690407097,-0.0002238388,0.2144718021,0.2041130066,-0.1077796593,-0.2365299463,0.1485313028,-0.0898998529,0.1902943403,-0.2229714245,-0.1885854006,-0.2068694681,-0.0418122523,-0.3231393695,-0.013327688,-0.2607429624,0.0772032067,-0.0433270223,-0.4872079194,0.1385264397,0.2872037888,0.2798929214,0.1584499031,-0.236288026,0.2044305801,0.0388747714,-0.3264046609,-0.3186455965,0.0816777721,0.3960490823,-0.0962193981,-0.199808076,-0.0842521638,0.1123977676,-0.210931778,0.0993198678,-0.0172653906,-0.3980234563,-0.2877750099,-0.2272575647,-0.1440537572,0.2280877531,-0.1238338873,0.1092354283,0.3238578737,-0.1020536944,-0.0421350151,0.0472226143,-0.3129553795,-0.1615106612,0.2259002626,-0.0680491328,0.1246451885,0.2166646272,0.1262734234,-0.1239750311,-0.1283407658,0.015923949,0.110541746,0.0809169412,0.1783045381,0.3882919252,0.1421783566,0.04564346,0.1837655604,0.2708903551,-0.2175178528,0.0005880228,-0.6009100676,-0.3271937072,0.0152038941,-0.1389222592,0.1773627847,-0.1620715261,0.3467495441,0.0747586936,-0.0876115859,-0.2790808082,0.1054969802,-0.2524275482,-0.0330542363,0.3787266314,0.0321678743,0.1882181913,0.0604914092,0.1364929974,-0.0075695026,-0.1669653654,-0.2190154642,-0.0968082175,0.1552595794,0.0360927582,-0.324672997,-0.1140930653,-0.2800056934,-0.1091171578,-0.1485867053,0.2010942996,0.2201131284,0.0697049573,-0.0142667945,0.2874352336,0.0023168079,0.182918787,0.2478019446,0.022718478,0.1067924201,0.2279693484,0.0498295277,0.079078123,-0.0503811948,0.4414775372,-0.2765816152,-0.1662036628,-0.0622410774,0.266651839,-0.0797385871,0.0001466705,0.1708437353,0.2072013468,0.5527578592,-0.0051489123,0.2276041508,0.2421133071,0.0600854233,-0.268250078,0.0043666405,0.2005743682,0.2052980214,0.0675548762,0.3758973777,0.1295620948,0.2704569101,0.0920095593,-0.0519322865,0.4057397842,0.0340241082,0.2595987916,0.0937975422,-0.0268633459,0.0012690169,0.2760650516,0.1775625646,0.3242695034,0.2217540592,-0.3799788952,0.4237817526,-0.17972745,0.24527134,0.1152561754,-0.4413736761,0.2160060108,0.1960758865,-0.1192201898,-0.1913729608,-0.018545581,0.2208849788,-0.2242123038,0.1731158644,-0.1550617814,0.1390928775,-0.1601520181,-0.1405973136,0.0376081169,-0.2851132751,0.0589855872,-0.1458445489,-0.0228641834,-0.1104943901,-0.1318608224,-0.1499732882,-0.158703208,-0.0365961529,0.2429273725,0.304975003,-0.0171866007,0.0441878997,0.3076301813,0.0726316795,0.0201910324,0.2118661851,0.360176146,0.6329972148,0.1458148658,0.0199772157,-0.0067221224,-0.1254881769,-0.2086032331,-0.0578629896,0.403976202,0.0339972638,0.1046539545,0.3079672754,0.1693718284,-0.169427827,0.0600415133,0.0038525024,0.0138388844,-0.1192600429,0.0365674943,-0.0875124335,0.0739951581,-0.0295899156,0.1677485406,-0.3000261486,0.3487210572,0.1864134669,-0.291365236,0.2167501152,-0.0426208079,0.0767088905,0.1015828401,0.4885877967,0.1608777046,0.3075758219,-0.3083042502,-0.0974685699,-0.6733787656,0.0455199033,-0.0631597936,-0.1657150388,0.0463938452,-0.1422862113,-0.2273872048,0.1067602038,0.3253476024,0.0365948454,0.2370185554,-0.0189558994,-0.179011479,-0.4035907388,-0.0597984195,-0.4032366574,0.0397742689,-0.0347359814,-0.0172816925,-0.2817715406,-0.0508392863,-0.1897322536,-0.183355093,-0.0974668637,-0.1760172993,0.2545200288,0.1525194198,0.4518975019,-0.0744162723,0.2492251247,-0.3362928927,-0.4395931661,-0.3038597703,0.2104604393,0.1770783067,0.0404433794,-0.3705193102,-0.1775516123,-0.2045305371,-0.1834138185,-0.1058307365,-0.1154196486,-0.1602137983,0.1122977138,-0.1001882777,0.0281389523,-0.0156753343,0.2281958908,-0.1216267422,0.1054669097,-0.0954061449,-0.1676407009,0.7022178769,0.051288873,-0.061452616,-0.1471153051,-0.0944868326,-0.0570841096,-0.1484263837,-0.4888674617,0.2672984302,0.3948301971,0.041690439,-0.108330667,0.2988126576,-0.1936224699,0.0061627477,0.036324203,0.0121704927,0.0570545085,-0.1334373951,0.027306553,-0.0796211734]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1600","title":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comments":"Hi @david-waterworth!\r\n\r\nAs indicated in the error message, `load_dataset(\"csv\")` returns a `DatasetDict` object, which is mapping of `str` to `Dataset` objects. I believe in this case the behavior is to return a `train` split with all the data.\r\n`train_test_split` is a method of the `Dataset` object, so you will need to do something like this:\r\n```python\r\ndataset_dict = load_dataset(`'csv', data_files='data.txt')\r\ndataset = dataset_dict['split name, eg train']\r\ndataset.train_test_split(test_size=0.1)\r\n```\r\n\r\nPlease let me know if this helps. \ud83d\ude42 ","body":"The following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comment_length":76,"text":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\nThe following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\nHi @david-waterworth!\r\n\r\nAs indicated in the error message, `load_dataset(\"csv\")` returns a `DatasetDict` object, which is mapping of `str` to `Dataset` objects. I believe in this case the behavior is to return a `train` split with all the data.\r\n`train_test_split` is a method of the `Dataset` object, so you will need to do something like this:\r\n```python\r\ndataset_dict = load_dataset(`'csv', data_files='data.txt')\r\ndataset = dataset_dict['split name, eg train']\r\ndataset.train_test_split(test_size=0.1)\r\n```\r\n\r\nPlease let me know if this helps. \ud83d\ude42 ","embeddings":[-0.0721847042,-0.1323110312,-0.0861678347,0.2685681283,0.3532631397,0.1787524074,0.427296102,0.3118065894,0.3319097757,0.1663205922,0.2769959569,0.1833032519,-0.2980105579,0.307574451,-0.033894226,-0.2959113717,0.0081043774,0.1070964187,0.1237977371,-0.0300345644,-0.1895722151,0.1840578765,-0.2195200324,0.0997344851,-0.3673037589,0.0206725802,-0.2920014262,0.1307963878,-0.0135357408,-0.2515274882,0.4045374691,-0.1375559121,0.1239313036,0.4964110553,-0.0001121337,-0.0254925154,0.0358922109,-0.0644961447,-0.2525277436,-0.4415965676,-0.2628276348,-0.2395177335,0.1976013929,-0.3958483636,-0.1714026779,-0.0224701688,-0.136529088,-0.4225856364,0.3209575713,0.5206329823,0.2207773924,-0.00219544,0.0334863551,-0.1584517956,0.1464577466,0.1532027125,0.0414273925,0.1552959681,-0.1339991391,-0.0168723613,0.295756489,0.0661466122,-0.0692673475,0.3163212836,0.2319248468,0.2917082906,-0.2003952712,-0.2420010567,-0.115467906,0.1132740378,0.5742337704,-0.3167564869,-0.3579134941,-0.0600361638,0.1263863295,-0.5130979419,-0.0208156444,0.2185101807,-0.0335691981,0.0874094442,-0.1007624567,-0.2014103234,-0.1574539095,0.0674097985,-0.4065105021,-0.0739222988,-0.1161908805,0.1030807048,0.1283176243,0.0969465002,-0.0590991862,-0.0893520191,0.0394241959,0.1225550622,-0.2608906627,-0.1216239333,-0.1725812703,-0.3684391379,-0.0792468116,0.139998138,0.1815395951,-0.1973561049,0.0081711588,0.1937788427,0.3508922756,0.2967844009,0.0391871184,0.5758778453,0.0592384674,-0.1832799762,-0.338832736,-0.2924592793,-0.058088135,-0.2673820555,0.2529995441,0.1948087811,0.444372952,-0.2247351259,-0.3625485301,0.2208096534,-0.3055536747,-0.0277326219,-0.0909044445,0.3527173698,-0.0477382019,0.122298196,0.1689243615,0.0412398092,-0.1388924122,-0.2089310288,-0.2548887134,0.3821985722,-0.1558446139,-0.0423373803,0.109929584,-0.1273092777,0.0165456794,-0.2022984177,0.1279786378,-0.2566141784,-0.0368140228,-0.2804859281,0.0832359418,0.1974511147,-0.0184952468,0.2927273214,0.3714340031,-0.4367980063,-0.1055992544,0.2160589248,-0.500698328,-0.4635074735,-0.040211942,0.2280384004,-0.2061602324,-0.087254256,-0.3410853446,0.0228510052,0.1365407407,0.031679187,-0.0430204719,-0.3725330234,-0.187078774,-0.289393127,0.339697957,0.3746131957,-0.5248161554,-0.0159844737,0.0027712686,-0.200794816,0.1853204817,0.0941373035,-0.0391452014,0.4101447165,-0.1283570379,-0.0481366552,0.3652197421,-0.2434015423,-0.2943015099,0.2787620425,-0.0637711659,-0.033641208,0.0105867274,0.0398907028,0.200393483,0.1239671633,0.0806268975,0.2507042885,-0.1206177399,-0.0842709243,-0.0462476388,0.1164425239,0.5629889369,0.29239133,0.1966350079,0.1795151383,-0.0754685923,-0.0262020715,0.1159027442,-0.0579619519,0.0302031673,0.0934660584,0.3688443303,0.116268605,-0.0223210435,-0.4006477892,-0.1254470348,0.0952937305,0.4420140684,0.0403981693,-0.0990110189,-0.0385203436,-0.2707170546,0.0100218048,-0.3413084745,-0.2407452017,0.1200156957,-0.16371575,0.0094290171,-0.013194927,-0.3408820629,0.3551131189,-0.223589018,0.1766072363,-0.1674160957,0.3753546476,-0.1724620312,-0.0743992627,-0.1578885913,0.1958103627,0.0619020648,-0.0317661911,-0.1056192294,0.3751665354,0.2643580735,-0.2826120555,0.0902113691,-0.0859476104,0.1390044838,-0.2407877445,-0.1898607612,0.2528318167,0.2463705391,-0.1360628307,-0.030519506,0.408488065,-0.1322629601,0.3659788072,0.0389474072,0.3024173379,0.0064868806,-0.0319371074,-0.0359364115,-0.0593069829,0.1365471184,-0.4171840549,0.1285039783,-0.2234471589,-0.2419031262,-0.2094892561,0.3478752077,-0.069305405,-0.0220943298,0.0216235146,-0.2207664847,0.2940608263,-0.1583036184,0.4630827606,0.5181429982,0.1379354894,-0.1991990209,0.1414236426,0.0021805759,0.0898431987,0.2335760593,0.0540669002,0.1229796708,0.0310472734,-0.2389370948,0.0470307432,-0.053731937,-0.5640873909,-0.071783714,0.4171223938,-0.4143362641,0.0641179085,-0.1269441247,0.0092918146,-0.296294719,-0.3549583852,-0.0284615308,-0.3789285719,-0.1835002601,-0.1405200064,-0.0239546187,-0.1128611416,-0.4261830747,-0.0856428817,0.0932454914,-0.2486566603,0.2697195113,-0.3531538844,-0.079539068,0.1293804348,0.0736549795,0.1292053759,0.1668207049,-0.1089604497,0.0651179254,0.040707428,0.0916746855,-0.0229641031,-0.2543025017,0.135082975,0.2855051756,0.2967996001,-0.1143374816,-0.262509048,0.3381903172,-0.0877907351,-0.0605214052,0.243462652,0.275780797,-0.2561478317,-0.0971169397,-0.6580390334,-0.4108822644,-0.2691025138,0.0630399659,0.0739004984,0.1592146605,0.1093687788,0.2830112576,0.1262419373,0.2075158209,0.0419592038,-0.0460016616,-0.2564496398,0.3345264792,-0.2586972117,-0.3425795138,0.1432675272,0.0259500034,0.0694944337,-0.0448084921,-0.3738967478,-0.0027365733,-0.1055493429,0.1427330375,-0.1437899917,0.0626069754,0.2344849259,0.1276384741,-0.1057046577,-0.0078382008,-0.1594961286,0.0380888022,0.0540230423,0.1010552198,-0.1472778916,0.2825528979,-0.0749214366,0.5764611363,0.0969001651,-0.1595266163,0.0693654194,-0.2750759423,0.0639482737,-0.0680162758,-0.6389513612,-0.0329338908,-0.2159185708,-0.2822295725,0.1167115048,-0.0733789504,0.2000922412,-0.3924251497,-0.0912912637,-0.2240173966,-0.2448515594,0.0542313084,-0.2960803807,0.1903056502,-0.248251304,0.0775618851,-0.1729650348,0.0909667537,-0.0160467979,0.3166525662,0.1189932972,-0.0575218052,-0.5402713418,-0.3924870193,-0.2511126697,0.2855437994,-0.0690759867,0.2957165837,-0.019741267,-0.0272285566,-0.0459334813,0.1055565327,0.674249053,0.0671969056,-0.1452751905,0.0725868568,-0.1232419908,-0.1783969402,-0.0942352861,-0.2604153156,0.449637264,-0.1186182871,-0.0161083937,0.0296085868,-0.01995451,0.299239248,0.1922976226,-0.2455410063,-0.0722681209,-0.1721977592,-0.1288796514,-0.1834926009,0.0396891981,-0.0242904592,0.2716973722,-0.069162719,0.0858444497,-0.1916061193,-0.1549893767,0.5035322905,0.0667256266,0.2780402601,-0.2251093984,0.1154664233,0.0561304428,0.1723070741,0.2256921828,0.5942599177,-0.0520958304,-0.4422063828,-0.0910880938,-0.2161379755,0.6388921142,-0.0520616584,-0.2286679447,-0.007967432,-0.1412947327,0.0992834494,-0.2276455015,0.0082798544,0.3840450346,0.0977487937,-0.4098173678,-0.6983443499,0.2479275316,0.0678379983,0.1709145755,0.2534902692,0.0964543447,-0.2580840588,0.4556393325,-0.0939265341,0.8387123346,0.1053680256,0.0990953967,0.0341435783,0.0411580577,0.4353800118,-0.0145551777,0.1437955797,-0.2871617079,-0.5176244974,-0.2060649991,-0.1592775136,0.3610612452,0.1694132388,-0.197415337,0.4240471721,-0.137970835,0.1385344118,-0.1215623766,0.1294140816,0.1554100811,0.0039540585,-0.2922899127,0.1733409017,-0.0998881906,0.2435767651,0.1092956588,0.0772559047,-0.3266753256,-0.184703514,-0.2942938805,-0.0714986026,-0.0502400063,0.1856258661,0.0318278484,-0.4340490699,0.2708715796,0.2985517979,0.2258134931,0.1510853767,-0.0733152181,0.1236286461,-0.0843476057,0.0912762582,-0.0423797853,0.0655172914,0.5445194244,0.0768014267,-0.2678830922,0.3966128826,-0.0029904689,-0.1281079501,0.1212416291,0.2293196768,0.0765843093,-0.6178494096,-0.3464152217,0.0288612954,0.0228959937,-0.1572952121,0.163443014,0.1338638216,-0.4194944799,0.3092005551,0.0632803813,-0.2606849372,-0.0963098779,0.2163581103,0.1629343778,0.2465141565,0.6398243308,0.092577666,-0.0098999618,-0.2739763856,0.2244112045,0.4415142238,-0.2741951048,0.2020126134,-0.022194529,0.3782289028,-0.0185004901,0.2545011342,0.0148525368,0.1679700613,0.0134746404,-0.5874105096,-0.1785158068,0.0299021266,0.0678364262,0.2657144368,-0.0682324022,0.3186632991,0.044655472,0.1009659916,-0.3291759193,0.129448846,-0.1768713593,0.1042376906,0.1987215579,0.2297051251,0.1792407334,-0.2211219519,0.1787106097,-0.0331516825,0.0672197565,-0.2704760134,-0.1381520629,0.1154435799,0.0779267624,0.1170513704,0.0263843723,-0.0415176786,-0.0412724316,-0.1786631942,-0.0748812109,0.3920443654,-0.124678649,0.6344898343,0.139708519,0.084443517,-0.2582957745,0.1275230795,-0.0869969428,0.3201051056,0.3684004545,0.125464499,0.0744451955,0.0658873841,-0.354642421,0.0452725291,-0.2939605713,-0.1350603402,0.3098613322,-0.4562231302,0.200041756,0.0263524409,0.4301758409,0.4376136363,-0.2431810051,0.4377040267,0.2264554054,0.2433790565,-0.2659005225,0.1741059572,0.3031703532,-0.1066059992,-0.0247050934,0.2842214108,-0.1608769298,0.0681732222,-0.4807461202,0.0944099352,0.4180377722,-0.1298874319,0.3879742622,0.3791291416,-0.0596056879,0.0380722806,0.0702856258,0.1264149249,-0.0031306872,0.6698891521,-0.0753341466,0.2561290562,0.2050168514,0.0044148616,0.1049541086,-0.2050678134,0.2074905187,0.051977247,-0.1348827183,-0.0753559023,-0.0836353302,0.4223738313,-0.4901798368,-0.0410122797,0.0220178831,0.026942743,-0.0455720164,0.1274721324,-0.0465647541,-0.0742499083,-0.2240782082,-0.135683015,0.1360314339,-0.3794770241,0.1740129739,-0.0083090188,-0.051539354,-0.2948414683,0.1352845132,0.1303865314,0.1709826142,-0.3162543476,-0.0131663857,0.1555113792,0.0070483279,0.3802857697,-0.001292633,0.3857502639,0.2601336539,-0.1090431586,-0.0563489497,-0.1175308749,-0.2950937152,0.2205930799,0.1814935356,-0.1840782315,0.1059237421,0.3545348346,0.2340081781,-0.0747618303,0.169818908,0.1993200183,0.1871189028,-0.4202073514,0.6271944046,0.0535096079,-0.255823046,-0.2602987587,-0.1061137542,-0.3823805749,-0.3083065152,0.1675444543,0.1097627655,0.1084963232,-0.0165266097,0.0893463343,0.0704536885,0.2227946669,0.279964149,0.4098479152,-0.488460958,-0.0084739318,-0.4700657725,0.0326474421,-0.1541023999,-0.1359292418,-0.1391151994,0.0506295487,-0.0652023852,0.1427138448,0.0248275436,0.1623287499,-0.0151281189,0.2275580913,-0.1682758927,-0.4612595439,-0.325255096,0.16801171,0.0045839762,-0.0336374789,-0.1374645233,-0.0759002566,0.0395974554,-0.0020673173,-0.0511302873,0.0990502909,-0.0323346741,0.2202933431,0.0572283231,0.7479214072,0.0703615323,0.1436002702,0.0226143524,-0.1677639484,-0.2627967,0.3687815964,0.1317407638,0.3843569458,0.025984453,-0.2073770463,-0.2479707003,0.2719694376,0.0841691047,-0.093893975,-0.249550879,0.1534935385,0.0458176583,0.0400983989,0.2332106084,0.3029092848,-0.2424895018,0.2095244229,-0.1211486831,-0.3196155429,0.3894041181,-0.3805125654,-0.1750677526,-0.12260364,0.1846761107,0.1710370928,0.0157212019,-0.5626366138,-0.098763749,0.4462661743,-0.1658881605,-0.2387492359,0.2059510201,-0.0196027067,0.1822865605,0.0629559606,0.4821690023,-0.0643890724,-0.0762837455,0.1273597479,-0.2176991552]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1600","title":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comments":"Thanks, that's working - the same issue also tripped me up with training. \r\n\r\nI also agree https:\/\/github.com\/huggingface\/datasets\/issues\/767 would be a useful addition. ","body":"The following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comment_length":22,"text":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\nThe following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\nThanks, that's working - the same issue also tripped me up with training. \r\n\r\nI also agree https:\/\/github.com\/huggingface\/datasets\/issues\/767 would be a useful addition. ","embeddings":[-0.1846861988,-0.1984477043,-0.0545882955,0.2467512041,0.3953751922,0.1030242667,0.4267278612,0.3831037879,0.3261997998,0.1956540048,0.1507687718,0.0640658215,-0.2686586082,0.4507163763,0.0584371015,-0.3567760587,0.0626722947,0.1936277598,0.06078168,0.0350720882,-0.1242044792,0.1890627295,-0.2912185192,-0.0519424826,-0.473429352,0.0503040664,-0.2243108004,0.0437319949,-0.0271128807,-0.2051904351,0.384031713,-0.1047580689,0.0608503558,0.5816102028,-0.0001103235,0.060649313,0.1080931276,-0.106884189,-0.2396588326,-0.2753520906,-0.1038397253,-0.2271676064,0.1324585676,-0.2241396159,-0.3628971875,0.0539404266,-0.1217899323,-0.2161211818,0.4450163841,0.4613480866,0.2605397105,0.2811121047,0.0929356888,-0.2848074436,0.1258871704,0.1839955896,-0.1198847815,0.1295715719,-0.0749584511,-0.0903258845,0.2761040628,0.2351316512,0.0938194171,0.2621213198,0.2497683316,0.1438686401,-0.3698174357,-0.1744376868,-0.1129837185,0.2359529883,0.3810487986,-0.2765884697,-0.5040438771,-0.2638499141,0.153602168,-0.5238580108,0.0427222475,0.1760730147,0.0776904523,0.1835557669,-0.241923213,-0.1039561778,-0.0991920009,0.0538911782,-0.283732444,-0.0754010156,-0.2382143587,0.0181718338,0.2245601565,0.0913871676,-0.2342689335,0.0713992715,0.0532810763,0.0892882496,-0.2784926295,-0.1460763663,-0.2125331461,-0.1639090329,0.110178858,0.2204519808,0.0217942372,-0.1837146878,-0.0039266036,0.1535110474,0.2553347647,0.3511749208,0.1957256645,0.4229312241,0.223684594,0.0411520191,-0.2873191237,-0.2798427939,0.0130214998,-0.3616330922,0.0008081726,0.1058954522,0.3911060393,-0.1769423336,-0.3389345109,0.0766282454,-0.2040091008,-0.0077176387,0.0730956048,0.3337322772,-0.0573752075,-0.0588300377,0.1081146076,0.1537029743,-0.3166412115,-0.3853555322,-0.326241374,0.288695842,-0.1831528842,0.0292907339,0.2001987845,-0.2009829581,0.2049362659,-0.200357452,0.2115934342,-0.1807635278,-0.090694949,-0.1443425715,-0.0365465432,0.3677008748,0.0280801263,0.1323035806,0.2988390326,-0.2814490497,-0.0717439801,-0.0172024891,-0.31913203,-0.3248497248,0.0714159012,0.1977607012,-0.2869434953,-0.1072334051,-0.3161544502,0.0980216712,0.159886837,0.0231092069,-0.0298051164,-0.2496530712,-0.2829373181,-0.1383283585,0.446244508,0.4211852551,-0.3492752016,-0.1563429385,0.1918721348,-0.2150983512,0.0671740621,0.2506552637,0.0235544313,0.3546364009,-0.0481557287,0.1400209367,0.1740016043,-0.1990397573,-0.2208687961,0.2080297023,-0.026805561,-0.0019288772,-0.0054318677,-0.154068172,0.0715041384,0.0562889911,0.1036762521,0.293361485,-0.2000457197,-0.0187269431,-0.0270957574,-0.0861532465,0.303768158,0.2829179168,0.1473238766,0.0183842424,-0.1201270446,0.026026804,0.1397976577,-0.1667535007,0.0329543799,0.1168683842,0.3362462819,0.1048713997,-0.025149459,-0.3535371721,-0.1979879141,0.1710967571,0.4560193717,0.062550813,-0.0488089956,-0.0994438156,-0.2215573341,-0.0261200927,-0.2876888812,-0.4046596587,0.0574959479,-0.2550922036,0.1605911255,0.1340693384,-0.3540807068,0.3058643341,-0.2963332534,0.3191065192,-0.2286272198,0.3604705632,-0.1177028343,-0.2119923979,-0.0572769567,0.2541782558,-0.0181351062,-0.1646514833,-0.0841823667,0.3766163588,0.1199678108,-0.1657221466,0.0825534016,-0.109981589,0.2615176439,-0.1722089946,-0.2904167473,0.0505828597,0.0591344684,-0.1039418578,-0.1310251802,0.4590165317,-0.0876019076,0.2237435579,0.0420509167,0.2742888033,0.0848186612,-0.092228964,-0.0446876772,-0.1057677716,0.2735306621,-0.5087496042,0.1001402587,-0.2306523025,-0.2755360305,-0.0906738117,0.4155738354,-0.0166744106,0.0677953362,-0.0181375351,-0.1673465222,0.3644117415,-0.0292870849,0.4095970988,0.4842481315,0.1815354824,-0.2047944516,0.1650668532,-0.0136250993,-0.0591130964,0.1353191137,0.0115996152,0.1306747794,-0.0279480349,-0.2468860447,-0.042242635,-0.2045158297,-0.3201777041,-0.0408037901,0.2895835638,-0.4424277246,0.0485311113,-0.243034929,-0.0320067368,-0.2458418906,-0.263430804,-0.155266434,-0.2756096721,-0.0537177585,-0.0249828901,-0.0737604871,0.018274622,-0.3404821754,0.0271311104,0.1059866548,-0.0963923782,0.1737786233,-0.222713992,-0.0749430284,0.100150153,0.2192967832,0.0530372486,0.2077374756,-0.1844676137,0.0754775405,-0.0088191824,-0.1656457633,0.0247168038,-0.3662691712,0.279427141,0.2628706694,0.2170624435,-0.1536860764,-0.1248969287,0.4182597995,-0.2302126437,-0.0924310014,0.0869576409,0.3433737457,-0.2754804492,-0.3146719038,-0.5279149413,-0.2542979717,-0.2518594265,0.2830945551,-0.0502455086,-0.0180444401,0.3214957118,0.2813249826,0.2104564756,0.0030243883,0.0112941489,-0.052640032,-0.3032206893,0.2851372063,-0.391002208,-0.4183826447,0.0890901312,0.1104732752,0.1094342098,-0.0517444722,-0.3512325883,-0.0866654813,-0.1428495198,0.1426784396,0.0041108085,0.1272309273,0.2416519523,-0.1699460745,-0.0425243713,-0.0882617161,-0.2323176414,0.1702661365,-0.0606296957,0.1275812238,-0.2310269326,0.2783776522,0.0382479951,0.6902598739,0.1376738399,-0.0335256271,0.1117489412,-0.311475575,0.1596425772,-0.1432040036,-0.547619462,0.1341440976,-0.059121836,-0.1061896011,0.1755037606,0.0977407768,0.2639809847,-0.2763020396,-0.1354800761,-0.1459861398,-0.4110564291,-0.0224735234,-0.0475868806,0.1665591151,-0.2509230673,0.1649682373,0.0239431709,0.1250636876,-0.036216177,0.2875620425,0.0130773811,-0.0930172428,-0.4692695737,-0.352461338,-0.440759927,0.2361113429,-0.1203943118,0.2364783883,-0.0943083093,-0.1046336889,-0.0983239114,0.1375365853,0.7563708425,-0.0219588019,-0.117466256,-0.0697756857,-0.2654159069,-0.2875236273,-0.1143938303,-0.1660134196,0.281614244,0.0405383594,0.1399068236,-0.0395577587,-0.1005631685,0.3904910684,0.1394097954,-0.0860415474,-0.0941690952,-0.1784464121,-0.1805272996,-0.2042564303,0.0865899771,0.1529073119,0.1488432586,0.0072235223,0.1084178388,-0.1991005093,-0.2399874777,0.4517407119,0.0708558559,0.1966763586,-0.1215073317,0.2498881072,0.2474304587,0.0659671277,0.1725645512,0.6067808867,-0.1809215248,-0.5803419352,-0.0645788386,-0.2436720133,0.5723367929,0.1711210757,-0.120610632,0.0035680383,-0.035753075,0.1895620674,-0.2105436921,0.0060102153,0.3294207156,0.1648205966,-0.2364555895,-0.5567426085,0.2764131427,0.0994316265,0.3125402033,0.3097650409,0.3089547753,-0.2004816979,0.2576605976,-0.2299964577,0.7025114894,0.1212065071,0.0765970349,0.1732084006,0.0672574788,0.4859927595,-0.0007577693,0.1452780962,-0.2311039567,-0.4378169477,-0.099671647,-0.0695250407,0.4050200582,-0.0732656568,-0.2276267111,0.3836452067,-0.0714426413,0.1355580986,-0.2266634703,0.0504279993,0.0069392156,-0.1747953743,-0.4288173318,0.1857798398,0.0523424968,0.3444141448,0.1153296381,-0.018928498,-0.2602659166,-0.2005392015,-0.2727979422,-0.0671606809,-0.1454299986,0.1588488519,0.3116418719,-0.3745336235,0.3346984982,0.2419031411,0.3966185451,0.0982908607,-0.193219468,0.180407092,-0.2654661834,-0.0271420833,-0.0957375094,-0.0074779508,0.529916048,-0.0249193702,-0.2789433897,0.0683570206,-0.0169676803,-0.1393125951,0.1542785168,0.1149234623,-0.1036340073,-0.5509044528,-0.2589760721,-0.0202846937,0.125142619,-0.240995869,0.1862434149,0.141300261,-0.3067742586,0.1170662865,0.1744794101,-0.2401868105,-0.1974875182,0.3025266528,-0.0414185338,0.142427206,0.5993548036,0.2066729516,-0.0137996515,-0.3160971105,0.1751202047,0.5526399016,-0.3977743983,0.1992246956,-0.0278255846,0.2351665646,0.0384805202,0.324749738,0.0003817113,-0.0852931067,0.0933169946,-0.4125089645,-0.1559256166,0.1037855595,0.0040022144,0.2791255116,0.0261521172,0.3407078683,0.1152539402,0.0288767442,-0.3290396333,0.1591720283,-0.1257724315,0.0665454119,0.2596556246,0.0864379406,0.3596298695,-0.2830477655,0.152535677,0.0681307688,-0.0249203984,-0.2598589957,-0.1910291016,0.0989779979,-0.0232913997,0.1772610247,0.161465168,0.0732407868,-0.1714469194,-0.270218581,0.070489794,0.288820982,-0.0642657354,0.4903995097,0.1225862727,0.1161269695,-0.0744379088,0.0421854369,-0.0498997644,0.2619653642,0.2407581359,-0.0472239107,-0.0636247769,0.1204841509,-0.200685516,0.0221741237,-0.2818071842,-0.1100149378,0.2184884846,-0.4265004098,0.1330872178,0.1526703537,0.4794739187,0.3238247037,-0.2561483383,0.4194527566,0.3094594777,0.2326135188,-0.2462056875,0.2292548269,0.3853669167,0.0372099467,0.0800132006,0.3398469687,-0.0205736496,0.1317349672,-0.3172639012,0.2038926184,0.2978074551,-0.0991519392,0.4517715871,0.3082892299,-0.1605488062,0.079776518,0.1402213573,0.0540033467,0.1489086151,0.5775363445,-0.2341184765,0.2427436113,0.0030869602,0.0952907205,0.1125719324,-0.2341766804,0.0753906444,-0.0279167425,-0.123877719,-0.0927224383,-0.2193617225,0.7663367391,-0.6305839419,-0.168550387,-0.091338791,0.1773313731,-0.0440921299,0.1119597331,0.0201199707,0.0009325509,-0.1366668195,-0.1476505846,0.1166810989,-0.2911379933,0.2370639294,-0.0175177101,-0.0739107952,-0.3109386563,0.0810782909,0.0583282262,0.2695934474,-0.2086713612,0.0049610157,0.0519668572,-0.0117209116,0.2858563066,0.234013319,0.4605006278,0.1309825033,0.0718592629,-0.1021860614,-0.1684698761,-0.1850577891,0.1636552513,0.2709507346,-0.0707170665,0.1986960918,0.320268482,0.2055373788,-0.1299634576,0.1336038858,0.0303097107,0.1736452878,-0.4789002836,0.6228631735,-0.1364354938,-0.132383585,-0.3485612869,-0.1829468906,-0.4489800036,-0.1026140377,0.1511127204,-0.0268433653,0.0801587254,-0.0825148001,0.0809317753,0.0874390528,0.3327470124,0.2745979726,0.2439937443,-0.3794067204,-0.0223782584,-0.4814208746,-0.1162690297,-0.2013105154,0.0326568969,-0.1707203239,0.0711300448,-0.0923437849,0.1257176101,0.0916626528,-0.1133765951,0.1285658777,-0.0238352735,-0.2652895153,-0.3344790041,-0.3479903638,-0.0067151352,0.1544408649,0.0164716672,-0.0083656097,-0.1240032464,-0.0236130916,0.1138249859,0.0847188458,-0.0404525623,-0.1717931926,0.1929931045,-0.0098008113,0.6767649651,0.0883392841,0.1176790521,-0.1806137115,-0.2579259872,-0.2844210267,0.3459977508,0.0008281137,0.4123235643,-0.1656765044,-0.3102829754,-0.2536058724,0.3791850507,0.1961625814,-0.0965305194,-0.107935518,0.2606863081,0.0420684814,0.0455177277,0.3739990294,0.3746081591,-0.4039810598,0.1107295305,-0.1688305289,-0.3724434972,0.5001776218,-0.4634429812,-0.1900966167,-0.0587620325,0.1768218428,0.4209754169,-0.2111109197,-0.7313174009,-0.0243655667,0.2395711988,-0.0905218646,-0.3277503252,0.3913854659,0.0184688549,0.1046299636,0.1292184889,0.565600276,0.0339646675,-0.2450905144,0.3566367328,-0.1714496911]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1600","title":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comments":"Closing this now","body":"The following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comment_length":3,"text":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\nThe following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\nClosing this now","embeddings":[-0.2165911794,-0.1156142652,-0.1547282785,0.2151317745,0.3571775258,0.1413639635,0.5365412235,0.3127042651,0.3181775808,0.1944720745,0.2975084484,0.0776169747,-0.2691287994,0.3204883635,-0.038517978,-0.2649321258,0.0942774042,0.1988207698,0.1877980977,-0.0531631336,-0.2478244603,0.1089841872,-0.3069024682,-0.0011863795,-0.3518226445,0.0564454198,-0.2351336479,0.0053524063,-0.0120735448,-0.2356038541,0.3126182556,-0.1432137489,0.0425492935,0.53373909,-0.0001090274,0.0181358065,0.0995279402,-0.0520238243,-0.2296444923,-0.163515538,-0.2892935574,-0.2679252923,0.1138843447,-0.4231679738,-0.1366590261,-0.033350192,-0.0937222093,-0.3226298392,0.3383784592,0.4891922176,0.2533312738,0.0895991996,0.0977732837,-0.197966978,0.1345553398,0.0374647453,0.0080510499,0.1918721199,-0.2124295384,-0.0636410862,0.3371697068,0.1788046509,-0.0764160231,0.2346238196,0.1490535438,0.1634343266,-0.1840076894,-0.1770638227,-0.1066140831,0.1118588597,0.4941656888,-0.2869966924,-0.4327162802,-0.0352169983,0.1153446659,-0.6305786371,-0.0150108859,0.2010841221,0.1294112056,0.1747963578,-0.1192245707,-0.1438951492,-0.189056024,0.0728510395,-0.2733741999,0.0854621306,-0.1297172606,0.0956041589,0.0638080016,0.0505722165,0.0299795065,-0.0783824027,-0.0003061116,0.0460150875,-0.2582685649,-0.1595558524,-0.224438712,-0.2551240027,-0.0071703941,0.1954799592,0.0834679306,-0.2581485808,0.1441468298,0.2644067705,0.2381157875,0.2939353883,0.0818283632,0.4858795404,0.1240001097,-0.1211706027,-0.4056529105,-0.2653479874,-0.0177646764,-0.4540929794,0.2339570075,0.3106167912,0.4484662116,-0.1542458087,-0.4364591837,0.0921983346,-0.2854396105,-0.0553880557,-0.0608465448,0.3634897172,-0.0914010033,0.0905076936,0.066812709,0.0424913615,-0.2713073194,-0.4141191244,-0.2956366241,0.4248864949,-0.0910388082,-0.1010410935,0.035382811,-0.1067000702,0.1549993008,-0.2643965781,0.1464879364,-0.2786694169,-0.0329398662,-0.3335040808,-0.1193906814,0.2968860269,-0.0017660502,0.1050235704,0.2884653211,-0.3055106103,-0.0257490724,0.1157168448,-0.4319779873,-0.345002085,0.0268859491,0.2645799816,-0.1423332244,-0.1463484019,-0.2597032785,-0.1157563254,0.1658685207,-0.0365705043,0.0283689518,-0.293009907,-0.1593287587,-0.3196315765,0.3288811743,0.4261953235,-0.4610563815,0.0136687793,-0.0004151607,-0.1202633828,0.1518007666,0.2024272233,-0.0645125583,0.4418674707,-0.0669878125,0.0290897563,0.3928308189,-0.2728186548,-0.2350717634,0.2091980726,-0.0401921645,-0.1506180018,0.0266689621,-0.083354488,0.0861813053,0.0038654662,0.169004485,0.1759024411,-0.1612701863,-0.0864271447,-0.039993193,-0.0090903817,0.4947561622,0.2123094201,0.1852391809,0.0133841326,-0.0851732269,0.0326037668,0.1271774024,-0.0897482634,-0.1222278699,0.0833534449,0.3813501894,0.0089769084,-0.0743109584,-0.3862054348,-0.0954936817,0.1279567778,0.4396855533,-0.1411217451,0.0103900265,-0.1170453206,-0.2942935526,0.0934688747,-0.3223962784,-0.242617175,0.1269471198,-0.2005779594,0.0141041875,0.0695964098,-0.3414818645,0.268877387,-0.3259544671,0.2087711841,-0.0256987792,0.3815901577,-0.1455621868,-0.1301416457,-0.1395665109,0.1818435341,0.0828669816,-0.0846908614,-0.1352252215,0.5056771636,0.2155066878,-0.1756013781,0.0466925241,-0.1187109202,0.1168624759,-0.1688813269,-0.2603200674,0.1460049003,0.1726823896,-0.0407361649,-0.013184201,0.3930956125,-0.0136945006,0.2308031172,0.0665716752,0.1515289843,-0.0061800596,-0.0826549232,-0.0230962448,-0.0489674583,0.1830333769,-0.3848026693,0.0442114957,-0.2535817623,-0.1567917913,-0.2104139924,0.5644355416,-0.0018281379,0.105323121,-0.0669735745,-0.1270076632,0.3720360398,-0.0044259126,0.4165984094,0.5667843223,0.1680355817,-0.2547520101,0.1733216792,-0.0222381819,-0.0088334372,0.2370300442,0.0902127698,0.2322402745,0.0691882595,-0.284433037,-0.0060901153,-0.1311072409,-0.5641340017,-0.0599371046,0.3495235145,-0.3726667166,0.1666741222,-0.1952664554,0.0161237735,-0.2712702453,-0.246906817,0.0143208597,-0.3337243795,-0.1270708591,-0.1824924946,-0.0806117952,-0.087285161,-0.4195177257,-0.1725065112,0.1035882533,-0.1436437517,0.2560679615,-0.2111967355,-0.037509948,0.1302948594,0.1189977303,0.0760556012,0.1828053743,-0.1434537768,0.070686467,0.0975748077,0.0508804321,-0.0417591073,-0.2369385958,0.171375975,0.2784315646,0.3062544167,-0.0461075492,-0.2266159505,0.354929179,0.0085549671,-0.0210317019,0.0305725485,0.386033684,-0.2867516577,-0.1829062849,-0.6356946826,-0.3298898339,-0.2980462611,0.0733502582,0.0582421087,0.1174155474,0.2191250771,0.3627149463,0.1856225431,0.2546183765,-0.0323586836,0.0772550255,-0.1355672628,0.2608924806,-0.2951762974,-0.2984301448,0.2334867716,0.0620910823,0.1757426411,-0.0988403112,-0.281455338,-0.0046965429,-0.074887,-0.0039030411,-0.1338421404,0.0841653123,0.2322718501,-0.0377798937,-0.1172686517,-0.0265078396,-0.1011085957,0.0061259051,-0.0543272682,0.0968607143,-0.2461405993,0.3613415956,-0.1464352906,0.5832125545,0.096812807,-0.1301891804,0.0600384958,-0.2970152497,0.0223419424,-0.1021500155,-0.5819538832,0.043777138,-0.0424773842,-0.2993447483,0.1438539922,-0.0801239386,0.2276609391,-0.3226840794,-0.2137098312,-0.2583842874,-0.2935443521,-0.0377027467,-0.1549055576,0.1861222535,-0.3185546994,0.1475453377,-0.0807061791,0.2439084798,-0.0488179885,0.3401106894,0.0355689637,-0.1899778694,-0.4260028005,-0.4987967908,-0.1465920061,0.3066388369,-0.1029705703,0.3194201291,-0.0198442973,0.0527501367,-0.1185950562,0.0624469183,0.7235286236,-0.114799805,-0.1402062923,0.0619871952,-0.0181684252,-0.1503733695,-0.1230168641,-0.1693427563,0.3475127816,0.0019467194,-0.0577291735,0.0319135077,0.0163247008,0.4474137425,0.1859904081,-0.1865291595,-0.1161393151,-0.2203061581,-0.109708719,-0.2236854583,0.0162650775,0.029435562,0.176536411,-0.050415948,0.0282655545,-0.2751162946,-0.300966084,0.4796034694,0.0856256187,0.2076580077,-0.1781844497,0.117130138,-0.0092694387,0.077001147,0.2147418559,0.5269619226,-0.0997855961,-0.4530473351,-0.0890103132,-0.1683455408,0.6559542418,0.0544735678,-0.2130184323,0.0260835048,-0.1006519496,0.0876087919,-0.0240925625,-0.0545567125,0.3582224548,0.1941832304,-0.320639044,-0.6474256516,0.2927321792,0.1043300182,0.2346655726,0.2579133511,0.0694305599,-0.2256586999,0.3464525044,-0.1637236178,0.7412594557,0.1849204004,0.0755435526,0.0746622235,0.1013701484,0.3852808177,-0.1551359147,0.1317245215,-0.2539622784,-0.4530781806,-0.1685425192,-0.116039291,0.5304476023,0.0925190225,-0.2881455719,0.4626777768,-0.2120892256,0.1902077347,-0.259200573,0.0591863096,0.0608854592,-0.019769812,-0.3515643775,0.1760729551,0.0256895609,0.2534195781,-0.0419180989,0.0528494567,-0.2269741297,-0.1792686135,-0.2312848419,-0.0455552004,-0.1676415056,0.202506274,0.0447157621,-0.3418423235,0.2854591012,0.2689253986,0.2888093591,0.1962767541,-0.0882456526,0.1963537931,-0.059526857,-0.0096929586,-0.1476804018,0.0122023774,0.4989795685,-0.0255187824,-0.343739152,0.3685621917,-0.0733161345,-0.13860403,0.2333033234,0.2125757188,0.1514417976,-0.6578701138,-0.4184582829,0.0036578006,-0.0318122804,-0.2067884058,0.1865403354,0.0564888418,-0.2605488002,0.2692466676,0.0800908655,-0.2042624503,-0.1532877535,0.2130814493,0.1203946993,0.2510177195,0.6617763042,0.0363804102,0.0087632025,-0.3664483726,0.1927245408,0.3925562501,-0.3502240777,0.1454834789,-0.0583933815,0.3458991945,0.1220119074,0.2794335485,0.0692506656,0.1159897521,-0.0018097056,-0.4988094866,-0.1236371771,0.0465842783,0.1180531085,0.269464016,-0.0423240773,0.3419875801,0.1194563508,-0.0115779918,-0.3360959291,0.1061675996,-0.1934532672,0.0005650156,0.1845773906,0.250087887,0.2099890411,-0.1973270476,0.2170882821,0.0621271282,-0.0025013199,-0.2863835692,-0.1583697349,0.1054932028,0.0405070111,0.1535470337,0.1466705799,-0.0161215309,-0.1002528742,-0.1373338103,-0.1136863679,0.141477555,-0.0646040365,0.5911881328,0.142031312,0.2084574699,-0.2193463147,0.0237852167,-0.1651867777,0.383505255,0.2950944304,0.030664403,0.1013301536,0.0557842739,-0.4075573385,-0.1082286686,-0.3492741287,-0.0797759816,0.2286151797,-0.4208650291,0.182153523,0.0250073262,0.3542855382,0.4450029731,-0.3114097416,0.4460113645,0.2845194936,0.2701590359,-0.2517151833,0.2345972508,0.3516770303,-0.0038509457,0.0568124354,0.3100918829,-0.0778241605,0.1231834516,-0.3356384635,0.1348239928,0.3014735579,-0.0893196613,0.4813706875,0.4545948505,-0.0768297389,0.0823231488,0.0147258881,0.065214999,0.1745178849,0.6956907511,-0.1533078253,0.2525840104,0.2250845879,-0.0029491663,-0.0243793502,-0.2266538739,0.1784846187,-0.0199479293,-0.1523839086,-0.0288140103,-0.0412768982,0.4014104009,-0.4244107902,-0.0429851376,-0.0125266016,0.1330595613,-0.0182098653,0.1443836242,-0.0584068075,0.0210327003,-0.1469732821,-0.2018046677,0.1118900478,-0.2449975163,0.0894883573,-0.0109168598,0.0037192453,-0.1792200208,0.269859314,0.1038400754,0.216663301,-0.3752128184,-0.00864131,0.1712271422,0.0273964629,0.2192055583,0.088655673,0.4643575549,0.2432167977,0.0227781087,-0.1735148281,-0.1253430545,-0.2138860971,0.2321545035,0.0691760704,-0.1305679083,0.2395526469,0.3109047711,0.2799113393,-0.1638040394,0.2190139443,0.1651636064,0.078693077,-0.5224715471,0.586877048,0.0275851302,-0.2259419858,-0.3782295287,-0.0934485048,-0.4507813752,-0.2449449301,0.1233231276,0.0670284703,0.1184033602,0.0011271564,0.0876048505,0.0215856452,0.3194788396,0.2368379086,0.4214295447,-0.4459475875,0.0069971727,-0.3962905407,-0.0559561737,-0.2600261271,-0.087601997,-0.0659512579,0.0526603386,-0.2088647038,0.1042289138,0.0982794091,0.0651775077,0.0645567402,0.1420287639,-0.2647394538,-0.4244701564,-0.2892549038,0.0840255022,0.1236817539,-0.1019685715,-0.0578659028,-0.2058138698,0.0622332841,0.0216079913,-0.0832080543,0.1487599909,-0.0493154339,0.3157121241,-0.0450369157,0.7179903984,0.0238307957,0.1227547079,-0.0465452671,-0.1962791383,-0.2451653183,0.3873879015,0.2226451933,0.3785984516,-0.0414750278,-0.2001888305,-0.1518626809,0.359952122,0.0342201777,-0.1487209499,-0.0703517497,0.1318968534,0.1085756943,0.053175237,0.2291684747,0.2418136746,-0.3520118594,0.1373569965,-0.1714948565,-0.3081892133,0.4470797479,-0.3700246513,-0.2186496556,-0.0812620744,0.2401710749,0.261038363,0.0419585891,-0.5829893351,-0.0866912454,0.4555713236,-0.0499029495,-0.3509840369,0.3089895844,0.0051721185,0.1206335053,0.1789471805,0.4244954884,-0.1273816675,-0.162857607,0.1818776131,-0.2482377291]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1600","title":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comments":"> ```python\r\n> dataset_dict = load_dataset(`'csv', data_files='data.txt')\r\n> dataset = dataset_dict['split name, eg train']\r\n> dataset.train_test_split(test_size=0.1)\r\n> ```\r\n\r\nI am getting error like\r\nKeyError: 'split name, eg train'\r\nCould you please tell me how to solve this?","body":"The following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comment_length":37,"text":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\nThe following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\n> ```python\r\n> dataset_dict = load_dataset(`'csv', data_files='data.txt')\r\n> dataset = dataset_dict['split name, eg train']\r\n> dataset.train_test_split(test_size=0.1)\r\n> ```\r\n\r\nI am getting error like\r\nKeyError: 'split name, eg train'\r\nCould you please tell me how to solve this?","embeddings":[-0.1205269769,-0.2044595033,-0.1178737432,0.3427510262,0.2748533487,0.2128635794,0.417621702,0.2791910172,0.3849134147,0.1766941845,0.2599661648,0.2442348003,-0.3558097184,0.4015853405,-0.1800408959,-0.3566969633,0.067672275,0.2050315142,0.2042681277,-0.0092861699,-0.2286288738,0.1024326384,-0.3420259058,0.0578026101,-0.3522693217,-0.0670242831,-0.3373372853,0.0235142987,-0.0220198687,-0.3306672275,0.3673844337,-0.1288262606,0.0630214587,0.4926139414,-0.0001131172,0.0144109484,0.0952198729,-0.0939696506,-0.3082743287,-0.2538853586,-0.4034256935,-0.202546224,0.1784944981,-0.5090008974,-0.102444768,-0.0878219306,-0.1878918558,-0.3301984966,0.4390422702,0.4318709075,0.2212600708,-0.0175378602,0.1199533343,-0.1634230316,0.0466196314,0.1382140666,0.0279129241,0.1376452893,-0.1803663224,-0.0400673151,0.3375974894,0.0766585469,-0.0616211817,0.3274958134,0.2132835984,0.1494387984,-0.1793500632,-0.3218154609,-0.0956685841,0.1868734211,0.5323076844,-0.2772233486,-0.4459287822,-0.1072070524,0.1357379705,-0.4340752065,-0.0066019399,0.0991728157,0.103386566,0.0961878225,0.0116508557,-0.2029306293,-0.1861786246,0.0810089484,-0.355663836,0.019007111,-0.1421079338,0.1209091917,0.0314170085,-0.017612258,-0.0384897254,-0.1144160181,0.0046987622,0.0224142,-0.2375529706,-0.2133391798,-0.2970763743,-0.2553686202,-0.019230146,0.0633097738,0.0608688407,-0.2469666153,0.1673274487,0.2162225097,0.3056078255,0.2904831469,0.2199985683,0.5732904673,0.0975123495,-0.09575665,-0.3294288218,-0.2312819809,-0.1072048396,-0.28599298,0.2528921068,0.276391089,0.4462272823,-0.0980492383,-0.3808518648,0.0926919058,-0.2468934953,-0.0697228089,-0.011388368,0.3455750048,-0.0857682526,0.1073075607,-0.0358099639,0.0881662741,-0.1349001974,-0.2533355951,-0.2457509339,0.4074082971,-0.0367090926,-0.0669158176,0.0501543581,-0.0713475645,0.183567673,-0.2005105764,0.0712040439,-0.2747604549,-0.089260377,-0.3202910721,-0.1533060521,0.2038680315,0.0455179252,0.1852240264,0.2966268659,-0.3811807632,-0.101030305,0.102444075,-0.4939915836,-0.2768363059,-0.0385558307,0.2153871953,-0.1205421016,-0.1385895908,-0.3266150355,-0.0319312476,0.1544391662,-0.0379985981,-0.0203255415,-0.2732340991,-0.144125089,-0.3367162049,0.2728694975,0.442786485,-0.4894809723,-0.1112404764,-0.0402708277,-0.1182367355,0.1037686318,0.0835876912,-0.0242834799,0.4687138796,-0.1057143733,0.0184721779,0.3932589293,-0.3447049856,-0.3340929747,0.2828659117,-0.0146962339,-0.0924784318,0.1201412156,-0.0028033424,0.2520149648,0.0232443549,0.0854704529,0.2322946787,-0.1012347639,-0.0957083106,0.0600435659,-0.059937641,0.4805958569,0.2306594998,0.1884990036,0.0880542323,-0.0794905797,0.0212223791,0.0340896733,-0.0553408042,-0.0318096951,0.1547145545,0.3268216252,0.0971263945,0.0214630086,-0.3522357047,-0.2091632932,0.1463080943,0.4027765393,-0.1277881414,-0.1587322205,-0.0898962542,-0.3355984688,0.0228168685,-0.4687238038,-0.2829129398,0.1041628271,-0.2288238406,0.0968894362,0.0673707202,-0.3220150769,0.3304789662,-0.1484997272,0.2350497693,-0.0644603446,0.3642419279,-0.2054459155,-0.0867250413,-0.1168430969,0.1835987419,0.1139998958,-0.0121547868,-0.1291647702,0.412443608,0.2695295215,-0.265586704,0.1130643785,-0.150993228,0.1742063761,-0.2123697698,-0.1195155904,0.1874687821,0.2570874989,-0.075355716,0.0458243713,0.4258392155,-0.0667375922,0.3461900353,-0.0190855768,0.2911891043,0.0023090953,-0.1240794659,0.037286561,-0.0840156674,0.1021381319,-0.4604661763,0.2323112488,-0.2532217801,-0.0918207392,-0.2488137037,0.4830524623,0.0224960204,0.1657025814,-0.0495697223,-0.2577195168,0.3827397823,-0.1473414004,0.4524562955,0.488855511,0.1363449991,-0.3093397021,0.0952892229,0.0685566887,0.1056234762,0.194423005,0.0804859027,0.2588699758,0.085517548,-0.3087680936,0.0358704738,-0.027666593,-0.5531837344,-0.0048054676,0.3986990452,-0.3466000855,0.0871452391,-0.141471684,-0.0603525266,-0.1292672753,-0.2361915112,0.0036394603,-0.3927210569,-0.1627770215,-0.1968063116,0.0466178469,-0.0247741044,-0.2414561212,-0.2356132716,0.1663375497,-0.2057171762,0.2398035675,-0.3715010285,-0.0335458443,0.1028845385,0.0890767947,0.0590357147,0.1915899068,-0.1959491819,0.0938040093,0.1138881072,0.0716952235,0.0083815558,-0.2420751154,0.1781362742,0.2741517723,0.3571283817,-0.0458537564,-0.3544014692,0.3528320789,-0.0043780128,-0.0839072242,0.1576317996,0.260635078,-0.1917724758,-0.0259466432,-0.5728279948,-0.4099226296,-0.2271383256,0.0313870013,0.0239049792,0.1485964954,0.1456320733,0.3056466281,0.1615990549,0.2220598161,0.107079722,0.0387223028,-0.1193662509,0.3027025461,-0.2321112007,-0.2960785627,0.1044882387,-0.0693777278,0.1490538567,-0.1352611929,-0.3425974548,-0.08992102,-0.1535870433,0.1381921172,-0.1096958742,0.0668885857,0.1864421219,0.1027308181,-0.0582427457,-0.0311725661,-0.0990062281,0.0555013791,-0.012225708,0.0967250541,-0.1267628968,0.3843730092,-0.1644410491,0.5157351494,0.1411608458,-0.1086243689,0.1097046956,-0.3059410751,0.0143279238,-0.0158294626,-0.6153257489,-0.0605167411,-0.0618869923,-0.2104400694,0.1313346922,-0.1136223376,0.1917755753,-0.3463659585,-0.2849587202,-0.2858389914,-0.2135683149,-0.12596412,-0.329318136,0.140188098,-0.3368887901,0.1928606033,-0.1289025396,0.1866608411,-0.0011483905,0.3112514913,0.0087834727,-0.1630607098,-0.4543210864,-0.4370806813,-0.1529827714,0.265203923,-0.0970758572,0.3282500207,-0.0340490714,0.0941461995,-0.0764287338,0.0943195596,0.7349424362,-0.0714670792,-0.1001705006,-0.0332464278,-0.0521326587,-0.1900089681,-0.1048291102,-0.2299948335,0.5030241609,-0.0876661837,0.0027429417,-0.0310242679,0.0344090238,0.3113953471,0.239801228,-0.2419746071,-0.1511629969,-0.1678676754,-0.0941835418,-0.3122289181,-0.0014196824,-0.0375216305,0.2183442116,0.0448481813,0.0967204571,-0.3013869822,-0.246302247,0.4358212054,0.1434432566,0.2560602725,-0.2165164053,0.1422136277,0.0582229868,0.0989068896,0.2481687516,0.4858223796,-0.1307382435,-0.5411925316,-0.1599549502,-0.2037206739,0.6125230789,-0.0984083042,-0.1674028635,0.1258666068,-0.152134642,0.0942442343,-0.1171960011,-0.1505892128,0.4400135875,0.1157613844,-0.4736856818,-0.6331071258,0.2424561679,0.1834883541,0.1777931601,0.2150507718,0.193329528,-0.188003391,0.4287166297,-0.1920146644,0.66092664,0.1043300182,0.0495335646,0.0827895477,0.0691626295,0.4578764737,0.0339507312,0.0691423416,-0.2622512281,-0.3694926798,-0.1801963598,-0.0797598287,0.4680801928,0.0982966498,-0.3375288844,0.5543084741,-0.3565013707,0.1723984331,-0.2492762357,0.1135993972,0.0654879212,0.002702534,-0.2002151906,0.1248351783,-0.0273621883,0.3215284944,0.0665844157,0.0451522432,-0.2416020781,-0.1770335883,-0.3047278821,-0.0593625046,-0.1201234981,0.2717935145,0.0178721119,-0.4535498917,0.3685100973,0.4197378159,0.2819157243,0.2263425738,-0.1074008346,0.2145715952,-0.1238343269,-0.0308244079,-0.1158101708,0.0609073192,0.4535270035,0.0621369183,-0.3640810251,0.3706395328,-0.0544335246,-0.0718912706,0.1885049194,0.1930850595,0.0826645195,-0.6170632839,-0.3145345449,-0.0730235204,0.0198251493,-0.202337116,0.1379705817,0.0599233881,-0.3034594357,0.3259062767,0.0855778009,-0.2349764109,-0.0773345903,0.1898122579,0.192624867,0.2795056403,0.7549842,-0.0061500296,-0.0099874428,-0.3380096257,0.2171352804,0.3891827464,-0.2369085401,0.1518325508,0.0105898147,0.3085423112,0.0520969257,0.1962128431,0.1046637222,0.2140082866,0.048670046,-0.5679895878,-0.2412710786,0.085939236,0.0973166749,0.1689918786,-0.0799370855,0.169477582,0.018362591,0.1771763414,-0.2893379927,0.1424961388,-0.2494203001,0.045074638,0.2369282842,0.2075513154,0.1613523513,-0.1639759243,0.2104177624,-0.0610006899,0.0906000212,-0.2415872961,-0.171802029,0.1104260832,0.0230544601,0.1284041703,-0.0357600786,-0.1128595397,-0.0447493717,-0.123884432,-0.0921593457,0.2925587595,-0.0106437607,0.6938596368,0.153323561,0.1420180351,-0.2456716597,0.1975135505,-0.2029734254,0.266824156,0.2879685462,0.1317586303,0.1380711049,0.0665026158,-0.4070839286,-0.0571364798,-0.3679028451,-0.1692455262,0.2420723289,-0.4453582168,0.2781388462,0.1007578969,0.337528497,0.4312368631,-0.2580162287,0.4187628031,0.2723287642,0.2146221995,-0.2926924229,0.1353315413,0.3706128299,-0.0750511587,0.0406157784,0.2696177363,-0.1462730169,0.1138869002,-0.4549008012,0.1108095571,0.2901775241,-0.0216060169,0.405675441,0.5363781452,-0.0205577109,-0.0157294776,0.0540847145,0.1852179915,0.0616678223,0.7057472467,-0.2238819897,0.251622349,0.1816998273,0.0022837538,0.2152891606,-0.14604038,0.2892875373,0.0238012224,-0.1140925959,-0.2394763529,-0.0767548308,0.338811785,-0.437171936,0.0026954322,0.0326131769,0.065116711,-0.0625541508,0.1184768155,-0.0029002768,-0.0547837168,-0.1409326345,-0.2161055952,0.1037255824,-0.4050590098,0.0682860911,0.0501564778,-0.091871962,-0.308848381,0.3248203993,0.2112972587,0.2408234477,-0.3704355061,-0.0629427955,0.2284309417,0.0591781363,0.321791172,0.1235974655,0.4195916355,0.2080782354,-0.1655699611,-0.1063493863,-0.0696405694,-0.3451307118,0.2758285999,0.0558696054,-0.2152123004,0.2649638057,0.3252098262,0.218501702,-0.1124547347,0.0673572645,0.2014528215,0.0604508296,-0.3707568049,0.570656538,-0.0448758118,-0.2099321336,-0.2597061396,-0.0337093808,-0.4932359755,-0.3157055974,0.1626577675,0.083075501,0.0874734744,0.0040430538,0.0666708946,-0.0347679853,0.311960429,0.3699966967,0.447342366,-0.4583413601,0.0587364733,-0.4064313471,-0.0625117421,-0.1706451178,-0.267593354,-0.0750989988,0.0279810727,-0.21851255,0.1184733734,0.0458286218,0.1959801167,0.0537378639,0.2616840899,-0.2815915644,-0.3883526325,-0.1944139004,0.1575339735,0.0974523649,-0.0869667605,-0.0311086271,-0.0532725491,0.0260924194,-0.0358333029,-0.0239420291,0.1411376297,-0.1114035249,0.179639712,-0.0342300013,0.8244967461,0.0411320142,0.139193058,0.1016960591,-0.1549030989,-0.2476409078,0.3596130908,0.2093880326,0.3200776577,-0.0601792671,-0.1672213376,-0.1139270887,0.3152984679,-0.022855904,-0.125820905,-0.2007557899,0.1408032328,0.1141410694,0.0304080658,0.1940094084,0.1550713032,-0.2861592472,0.1971878111,-0.0772078037,-0.2748168111,0.3750349283,-0.3308017552,-0.1739963442,-0.0950516388,0.1784998775,0.2217277437,0.0613464601,-0.4198961854,-0.0326580852,0.5135567784,-0.127369985,-0.2425843924,0.2391482145,0.0569447689,0.1867638081,0.1496245563,0.412231952,-0.179799512,-0.1278124601,0.1793389022,-0.1812321544]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1600","title":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comments":"dataset = load_dataset('csv', data_files=['files\/datasets\/dataset.csv'])\r\ndataset = dataset['train']\r\ndataset = dataset.train_test_split(test_size=0.1)","body":"The following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'","comment_length":10,"text":"AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\nThe following code fails with \"'DatasetDict' object has no attribute 'train_test_split'\" - am I doing something wrong?\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='data.txt')\r\ndataset = dataset.train_test_split(test_size=0.1)\r\n```\r\n\r\n> AttributeError: 'DatasetDict' object has no attribute 'train_test_split'\ndataset = load_dataset('csv', data_files=['files\/datasets\/dataset.csv'])\r\ndataset = dataset['train']\r\ndataset = dataset.train_test_split(test_size=0.1)","embeddings":[-0.168819949,-0.0895657241,-0.1656550467,0.2573613822,0.3516077399,0.1596391946,0.4991727471,0.3440155685,0.3641399741,0.1974352151,0.3051607311,0.1201410666,-0.2989811897,0.3131076097,-0.0644738525,-0.2879690826,0.0944450349,0.1789191514,0.1777369529,-0.0448618345,-0.2227533311,0.0891705453,-0.2522353828,-0.0573025681,-0.3801820576,0.0993146747,-0.2507269681,0.0183739439,-0.0053784563,-0.2344795465,0.3619081676,-0.1088499576,0.0490436517,0.5366619229,-0.0001086665,0.0727502033,0.1202280596,-0.0763781443,-0.2129806727,-0.2104907781,-0.2536337972,-0.285449177,0.1103726849,-0.4407963157,-0.1876960397,-0.0188312884,-0.1063611433,-0.3393587768,0.3114046454,0.4762917459,0.2473164648,0.1197987497,0.0659805387,-0.2427176386,0.1712759882,0.1113829687,0.0273008961,0.1780010462,-0.2250906825,-0.0259752665,0.3210640848,0.1848746091,-0.0560377873,0.2440651357,0.1898109168,0.1676557362,-0.2004002035,-0.1804250181,-0.1261216849,0.1463595182,0.4885717928,-0.2605952024,-0.4726575613,-0.0384933092,0.1402429342,-0.5402607918,0.0229078718,0.1370058656,0.1438129544,0.1158675775,-0.1002938822,-0.1335130483,-0.1582290083,0.0894621536,-0.3131714165,0.0464290231,-0.1156654432,0.0719725266,0.0483563393,0.0378157124,0.0597522631,-0.0551366359,0.0054276069,0.0491910465,-0.2332330793,-0.1202067807,-0.244792223,-0.2842235565,-0.0207614508,0.1565630436,0.1195522994,-0.2945025563,0.0667671114,0.2293013036,0.2508392334,0.2581825554,0.0659290552,0.5071336627,0.0988596156,-0.0914861187,-0.3934693038,-0.2994187772,-0.0122880721,-0.4554179907,0.2599800229,0.2914750576,0.4554568529,-0.1712064892,-0.3794609904,0.0475901552,-0.2830673456,-0.0803878382,0.0080843382,0.3440116346,-0.0730253533,0.1176920682,0.0474533066,0.1071521193,-0.2226705998,-0.3976819217,-0.3104906082,0.4239194095,-0.0405552275,-0.0585984811,0.1145501435,-0.0799061209,0.2132354975,-0.236697033,0.1418080777,-0.2447731197,-0.0430568494,-0.3072315454,-0.0946555734,0.3031710684,0.0191136021,0.1456106305,0.3015467823,-0.3156770766,-0.0584817417,0.1207301542,-0.4160334468,-0.3725086749,0.0739330873,0.2541280389,-0.1318846196,-0.1620066613,-0.29088521,-0.0698176399,0.2092530727,-0.0402286127,-0.0017364639,-0.3285949826,-0.1299798191,-0.3370349705,0.3269532025,0.390373081,-0.4283120334,0.0044121034,-0.0150055122,-0.1524241418,0.1445885599,0.1761439592,-0.0720389485,0.4559329748,-0.0796152428,0.0221958794,0.3807896972,-0.3545580208,-0.2053935826,0.2088555247,0.0421170816,-0.1104533225,0.1073822305,-0.0410548411,0.0823836848,0.0067538186,0.1500345618,0.2034300119,-0.1165105179,-0.1001956984,-0.0015229592,0.0375634693,0.5044831634,0.2675765455,0.1665123254,0.0698160753,-0.0968032554,0.0237547252,0.097398892,-0.1025888026,-0.1135391816,0.1295613199,0.2837895155,0.0444203168,-0.0763650537,-0.3487949371,-0.0994067639,0.1074484363,0.4071891904,-0.1156628877,-0.002422296,-0.0985592827,-0.2511886954,0.0237705857,-0.2945945859,-0.2689823806,0.1368894875,-0.1609607786,0.0419246964,0.0131160133,-0.3447531462,0.2833332121,-0.2583819926,0.2114238441,-0.0707356557,0.3754177988,-0.1742719263,-0.1601150036,-0.0800437331,0.1617924422,0.1292862147,-0.0362550318,-0.1228793934,0.4906884134,0.2007080466,-0.1766177714,0.0294972826,-0.0699162334,0.1476002485,-0.1843979955,-0.227977559,0.1840877235,0.2027374208,-0.1034414396,0.0010619638,0.4035955071,-0.1009403691,0.271948725,0.0634897649,0.1719463766,0.0384318754,-0.04552022,-0.048934523,-0.0337489285,0.1923883706,-0.3728203475,0.1006772071,-0.2757467926,-0.1720796824,-0.1762153059,0.4889076948,-0.0141296368,0.0433668457,-0.1584323645,-0.1451762319,0.3581564724,-0.0656810403,0.4384078681,0.5434161425,0.1883122325,-0.2372898608,0.1206557155,-0.0332639478,-0.0015461469,0.2120706737,0.0736538842,0.2344748378,0.0401762724,-0.2763022482,0.0112905018,-0.1381621361,-0.553427279,-0.0732117593,0.3723132908,-0.4140865803,0.1054813564,-0.1986807436,-0.0271586124,-0.2552155852,-0.2677759528,-0.0329104289,-0.3597173095,-0.148433581,-0.1531469375,-0.0818554461,-0.110881649,-0.3579110205,-0.2296980917,0.062295448,-0.1413431764,0.2354056984,-0.2632505894,-0.0043682451,0.1379502416,0.1119893715,0.104106836,0.1517979503,-0.1348432451,0.078154929,0.0871082246,0.0391448364,-0.0355797447,-0.2476497442,0.1708665639,0.2690754831,0.2707629502,0.016823655,-0.271261245,0.36338076,-0.0193695333,-0.0200490989,0.0650246888,0.3267942667,-0.2929120064,-0.156831488,-0.6081111431,-0.3394301236,-0.2986710966,0.0212261081,0.0966049209,0.1187644973,0.2269054353,0.3368711174,0.2051041722,0.2455575615,0.0351011492,0.0566224083,-0.1955898702,0.3052218258,-0.3106729984,-0.3222133517,0.1538508385,0.0299641062,0.1528657079,-0.033043284,-0.3032085299,-0.0197187308,-0.1425570846,0.0076173353,-0.1455681175,0.0826331601,0.2302883863,0.0088860169,-0.112516962,-0.0765442029,-0.153372094,0.0593599416,-0.0175255854,0.0927364081,-0.2482847273,0.2780461311,-0.1550465524,0.5371037722,0.0937599838,-0.1058342382,0.1459674835,-0.3088464737,0.0757639334,-0.0632644966,-0.5643721223,0.0336019397,-0.0478203818,-0.30448699,0.1093090847,-0.0693638548,0.2176673263,-0.3312976658,-0.2337933481,-0.2767468393,-0.2501311302,-0.0655181259,-0.1635406017,0.174129039,-0.3099191785,0.1690339744,-0.1200230792,0.260163933,0.0005322673,0.3420070112,0.0566733591,-0.1568801552,-0.5062533617,-0.4710015357,-0.1576303095,0.2538839877,-0.0837317258,0.2915797234,-0.0167058073,0.0405231193,-0.1477217376,0.0282261986,0.7611704469,-0.1178448051,-0.1256505698,0.0833425373,-0.0168453231,-0.2172798663,-0.1400920153,-0.2243396789,0.3834550679,-0.0311221145,-0.0480314679,0.0290372688,0.0217702165,0.3870057464,0.2060580254,-0.1887333989,-0.0896786377,-0.181367591,-0.1398535818,-0.208843261,-0.0330697075,0.0074658021,0.1847509742,-0.0233503096,0.0872295424,-0.3147545457,-0.2543343902,0.4516355097,0.0761826858,0.2854202986,-0.2226526141,0.1195965186,-0.0078860056,0.0645478591,0.2240463644,0.5583817959,-0.091721341,-0.4606264532,-0.1279206425,-0.2117185146,0.6439974904,0.028433986,-0.1661851257,-0.0421620235,-0.1095042154,0.028179111,-0.0324227959,-0.0695855767,0.3697930276,0.1168479994,-0.3224905729,-0.6984113455,0.3057525754,0.1458669454,0.2552419305,0.2518700361,0.0959991589,-0.2404976487,0.4288192391,-0.1548580229,0.7264741063,0.1791732907,0.0580282174,0.1287309378,0.1081115007,0.3967598975,-0.0653943643,0.1376120001,-0.2413049638,-0.4725702703,-0.1473034769,-0.1150015593,0.4794711471,0.1320770383,-0.2815611064,0.4910962582,-0.2486865371,0.1916026622,-0.3053161204,0.0730569661,0.049116157,-0.0344338119,-0.3382943571,0.1842460781,0.0213761274,0.2319994867,-0.0180254281,0.042443607,-0.2744146883,-0.183791548,-0.236561805,-0.0350775234,-0.182508409,0.2074574232,0.0706694573,-0.356223166,0.2360308319,0.3269349039,0.306539923,0.1521999687,-0.1256082654,0.2026683092,-0.0746597946,0.0055209491,-0.1425758749,0.0459774882,0.5135546327,-0.0398419201,-0.3316503167,0.2558654547,-0.0886059925,-0.1759694368,0.2042644173,0.1838205308,0.0873777419,-0.6025166512,-0.3732924163,0.0797222257,-0.0353354774,-0.2248951644,0.1874084324,0.0696175694,-0.2910529375,0.2453750968,0.0999751538,-0.1887649298,-0.1370452642,0.1982908845,0.1347460896,0.2708365917,0.6741462946,0.0200035553,0.0011707756,-0.3490204513,0.1410797536,0.4322199523,-0.321978271,0.1786877662,-0.1031469554,0.3110319674,0.0959434956,0.2559028864,0.0638471544,0.1389840692,0.0223944001,-0.5178411603,-0.1152924597,0.0549450405,0.134317711,0.2719337344,-0.0907014757,0.2887223661,0.0862716064,0.0210504681,-0.3396008313,0.1180333197,-0.158177793,0.0615521856,0.1575445235,0.2425912321,0.2182817012,-0.1722273529,0.233536154,-0.0175760332,0.041208677,-0.2908467352,-0.1720313877,0.0954012275,0.0733480677,0.1195537746,0.1099965647,-0.0420814231,-0.1052544266,-0.1844243407,-0.1322168261,0.2120015472,-0.069417432,0.6132817864,0.1790233701,0.2729517221,-0.2195266336,0.0718034655,-0.1647596508,0.3104909956,0.2944766581,0.0732924044,0.1142209321,0.0576492697,-0.3414894342,-0.0934145898,-0.3648389578,-0.0157237947,0.1690131724,-0.429779321,0.2277672738,0.056447763,0.3770379424,0.4141546488,-0.2568547428,0.4750721157,0.2273683995,0.2634678781,-0.2536590099,0.2126416117,0.3488264382,0.0062457812,0.0305259302,0.3018937111,-0.0666267797,0.1198053658,-0.3919455707,0.0923535004,0.2720296383,-0.0566477887,0.4515748322,0.4090518355,-0.1121446937,0.0581531376,-0.0248704441,0.0757708624,0.1189167574,0.6997756362,-0.1471264213,0.2850115299,0.2353593558,-0.0726913065,0.026645029,-0.1609303951,0.1711770594,-0.0140155517,-0.0685094148,-0.0506316796,-0.0290280711,0.3723839819,-0.4240604937,-0.0366046354,0.0107944217,0.0859630182,0.0174333751,0.1197142154,-0.0467850119,0.0003125382,-0.1687886268,-0.1891402602,0.131118089,-0.2676811516,0.1033350304,0.0473796353,-0.0168627854,-0.2472664565,0.2927853167,0.0932651609,0.181112349,-0.3328249454,-0.0045866054,0.2346045524,0.008305449,0.3388751149,0.1500240564,0.4661251903,0.2234011889,0.01139485,-0.1630111337,-0.1544198096,-0.2657023072,0.2025712132,0.0683809817,-0.1555551738,0.2297577709,0.3257814944,0.2775682807,-0.1522037089,0.1673384458,0.1561962515,0.0599574,-0.4817698002,0.6579479575,-0.0081980415,-0.2306609005,-0.3388347924,-0.0889268517,-0.4525027275,-0.2989162505,0.1037303954,0.0518531203,0.1155767664,0.0408537872,0.09004163,0.0258315839,0.3186717927,0.270629704,0.4159637392,-0.4118186235,0.0297546349,-0.4427039325,-0.0654199719,-0.2368257195,-0.0545959882,-0.1016374826,0.0701078624,-0.1959482133,0.0956335291,0.135280028,0.0461024716,0.0988546386,0.125721857,-0.2588565648,-0.4474729598,-0.2150309533,0.047473006,0.1021521688,-0.0943430141,-0.0699177757,-0.2015238553,0.087519981,0.0054536089,-0.0750793889,0.1296038628,-0.0366264023,0.2989094257,-0.0080806809,0.7428685427,0.0078888778,0.0614368618,-0.0699976012,-0.1894464791,-0.2999547422,0.3851858377,0.1553598046,0.4146446586,-0.0710941926,-0.1760358214,-0.1673569232,0.3979777396,-0.017419219,-0.1127231717,-0.1355445236,0.1739277244,0.0904098302,0.0304981582,0.2065194845,0.2269862592,-0.295018971,0.2064044327,-0.1786989272,-0.2924422622,0.437289387,-0.4076159894,-0.186259523,-0.0808788091,0.2252394259,0.2994106114,0.0831965059,-0.5973834991,-0.085950762,0.4264029562,-0.0901948363,-0.3252311349,0.3191953003,0.0261365566,0.125688374,0.1545255184,0.3867102265,-0.0998460427,-0.191755116,0.1619733125,-0.2207740098]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1594","title":"connection error ","comments":"This happen quite often when they are too many concurrent requests to github.\r\n\r\ni can understand it\u2019s a bit cumbersome to handle on the user side. Maybe we should try a few times in the lib (eg with timeout) before failing, what do you think @lhoestq ?","body":"Hi\r\nI am hitting to this error, thanks \r\n\r\n```\r\n> Traceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 379, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 208, in main\r\n if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO\r\n File \"finetune_t5_trainer.py\", line 207, in <dictcomp>\r\n for task in data_args.eval_tasks}\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 66, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/boolq\/boolq.py\r\nel\/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED\r\n```","comment_length":47,"text":"connection error \nHi\r\nI am hitting to this error, thanks \r\n\r\n```\r\n> Traceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 379, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 208, in main\r\n if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO\r\n File \"finetune_t5_trainer.py\", line 207, in <dictcomp>\r\n for task in data_args.eval_tasks}\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 66, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/boolq\/boolq.py\r\nel\/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED\r\n```\nThis happen quite often when they are too many concurrent requests to github.\r\n\r\ni can understand it\u2019s a bit cumbersome to handle on the user side. Maybe we should try a few times in the lib (eg with timeout) before failing, what do you think @lhoestq ?","embeddings":[-0.3293083012,-0.1345852762,-0.2016925067,0.3231747746,0.5001865029,-0.1886005998,0.1602738351,0.2997075617,-0.3055787683,0.1561734229,0.012656942,-0.0056440663,0.2251010835,0.1642649472,-0.0455740355,-0.0660552904,-0.2141919881,0.0516924225,-0.1668691933,0.1017805934,-0.0759902298,0.0589256845,-0.135654822,0.1066663116,-0.0014342469,-0.0925262198,0.0482695475,0.0914029479,-0.1770411283,-0.3981225789,0.3847735524,-0.0475312918,-0.0878206268,0.2440319657,-0.0001026009,0.2239751518,0.1840706468,0.0613851659,-0.2191357613,-0.165497303,-0.1825285554,-0.1764973998,0.3466186523,-0.2045548558,-0.323461026,0.156537205,0.0194423646,0.0245129764,0.6095243096,0.5391385555,0.3651290834,0.1239325255,0.2508780658,-0.1263591349,0.0121593978,-0.4649489522,-0.0548876636,0.2500353456,0.2293919772,-0.1382567436,0.0677758455,0.0419034436,0.0411453992,-0.0182585083,0.0719028115,-0.1040173098,0.1631603092,-0.3608938754,0.2469529659,0.1152732894,0.5638700128,0.0516682491,-0.1345327944,0.2436355948,0.0145922545,-0.2167304009,0.2574052811,0.0059991763,-0.2366169691,0.1825777292,0.1746409535,-0.1708101779,-0.5097798705,0.229768768,-0.093413502,0.2104021162,-0.0483462103,0.0594062284,0.0371313393,-0.2034060955,0.3239794672,0.2822365165,0.0639957413,0.095603019,-0.4873182774,-0.0258345623,-0.0517093278,-0.1743869781,-0.0484157056,0.1437137574,0.2108507156,0.0826344863,0.1433905512,0.0551506393,0.0877403691,0.1378038824,-0.1206299141,0.2517746687,0.2280283868,0.3546999991,-0.1444411874,-0.0109692346,-0.1517023295,-0.524538219,0.0227191225,0.3155859113,0.3285774887,-0.0800363272,-0.2814172208,0.1608621478,-0.073479332,0.0374105051,-0.0147579722,0.5862597227,-0.4222016037,0.1364524961,-0.0385238677,0.0206818804,-0.1698478311,0.0816259533,-0.2103554457,0.128688857,-0.1109872386,-0.2046239078,0.1267408729,-0.0979281366,0.1036819816,-0.1254515946,0.0796652734,-0.1540387422,0.2098276913,-0.3071586192,0.0403931066,0.2144149095,-0.0093317134,-0.0397677943,0.256238997,0.0868284404,0.1426610798,-0.0583050475,-0.3986396194,-0.4699076712,-0.0246357843,0.3337431848,0.1888823807,-0.1018986925,0.242503643,-0.239442572,0.1015270948,-0.1859617233,-0.1586317867,-0.3791605234,-0.1898315698,-0.2785962224,0.3494767845,0.2411410213,-0.2269233316,-0.0555184782,-0.0063581206,-0.2299630344,-0.0841925368,0.1094194576,-0.1654981077,0.1853645891,-0.2126497775,-0.0556343347,0.6239518523,-0.6403186321,-0.5884609222,0.451880306,-0.2488524467,-0.2172074914,0.0026223192,0.0128350044,0.31572631,0.1871753633,0.2534630299,0.2531638145,-0.0963229612,-0.0373811014,-0.1279286742,-0.2783178389,0.210431397,0.2529270649,0.338386029,-0.0542290099,0.1496915519,-0.2224408686,0.2122184187,0.1233668029,0.055118911,0.2695868015,0.2858492136,0.1266420037,-0.0151618924,0.2609181404,-0.2238556445,0.1011382937,-0.0823274553,0.1316879541,-0.4020649791,0.2130479068,-0.3046886325,-0.070966728,-0.4243715107,-0.1110660285,0.3290539384,0.0198725909,0.1009851322,0.0605011731,-0.0466322526,0.2943363786,-0.151177302,0.1532868445,-0.0333023034,0.2780037224,-0.3268547356,-0.0552035421,0.0937715843,0.0030752169,0.3442077935,-0.1848721802,-0.1886288971,0.2643786371,-0.159370333,0.1300932914,0.2049147338,0.0657288283,0.11827223,-0.448369503,-0.0238029417,0.0437438563,0.1285386086,-0.0440290347,0.4191423953,0.3036993146,0.029638866,0.3536557257,0.1969596446,0.1627426744,0.30114609,0.0441382863,-0.1183543652,0.1699621975,0.1193344668,0.1104473621,0.2826167047,-0.2836562991,-0.0844766945,0.0332247652,0.164361313,-0.0335939415,0.0535354093,0.0687814355,0.2974520922,-0.0559393056,-0.1830288321,0.1009347588,0.4425370097,0.0588133782,-0.1450980455,0.2595957816,0.0428838842,-0.1643870771,0.3133176267,0.0137783522,0.231696859,0.3095141351,0.0246753413,-0.0696165934,-0.1814823747,-0.2767195404,-0.1199273095,0.334752351,-0.1648125499,0.2214969695,-0.0891913995,-0.1461733282,0.0719188824,-0.1294409186,-0.1530590802,-0.2924027145,-0.0076095946,-0.0767282844,0.1505491138,0.2648813426,-0.1862804443,0.2516584396,0.3692764044,-0.1475120485,0.0229872093,-0.3701071739,-0.1531626284,0.2031988502,-0.0022343313,-0.3685530126,0.2077259868,-0.1924225092,-0.218137607,-0.1249535531,-0.1477749348,0.023245059,0.0012742316,0.0432976708,0.256049335,0.3906775117,0.1269900203,-0.0520312451,0.4064903259,-0.0884209946,-0.0913150236,0.0280723516,-0.0236177891,0.0279100221,0.0800108612,-0.6284663677,-0.5869358182,-0.294239372,0.1182904765,0.1431990862,0.1980281472,0.24892205,0.1683931798,0.1541638076,0.0365752652,0.318130374,-0.2569623291,-0.5267910361,0.1774822176,-0.1706408262,-0.4868075848,0.2212446034,0.1338598132,0.5196148753,-0.0392800197,-0.4721224606,-0.4179493487,0.0700986385,0.1188787967,-0.316524297,-0.1002306342,0.2164641619,-0.0834572911,-0.1960401684,0.0654206201,-0.0279408097,0.178632319,-0.000116027,0.2409600466,-0.1367491931,0.4993528426,-0.0066893827,0.4120452404,-0.0015904063,0.0416391492,0.173962459,-0.2517149448,0.1825367659,0.0320855603,-0.3973609209,0.1971779168,-0.0506810173,-0.1291242838,0.0730498731,-0.1645355672,0.1909097433,-0.5621136427,0.0427492596,-0.3912924826,-0.2510988116,-0.1200514883,-0.266015619,0.2810431421,0.0346744284,0.0859910771,-0.0982130542,0.0313237309,0.0420694649,0.3615584671,-0.1767034531,0.0078708092,0.0109080775,-0.0729205012,-0.352217108,0.2808557749,-0.0276010856,0.5090871453,-0.4446038604,0.0221328009,0.086178422,-0.0871059448,0.6551527381,-0.2658198178,0.1087110192,0.1216820404,-0.0544465892,-0.0370873436,-0.1218068078,-0.11756441,0.0180059187,0.1102109849,0.0187622011,-0.2171086669,0.0363610685,0.1376867294,-0.2447354645,-0.3150658011,0.0254317112,-0.2848104239,-0.3694692552,-0.0877855867,0.020495709,-0.0204049256,0.2766889632,0.0878966674,-0.1161073819,0.0979298651,-0.113330394,0.0173429102,0.0017108119,-0.0560506918,-0.1787100434,0.0288764928,0.1128022969,0.1006798819,0.2440427393,0.4235987365,-0.0448745713,-0.430036068,0.107609652,0.0284702014,0.2794895172,0.1612032652,-0.1627589315,0.2013744712,0.1291829497,0.3139091134,0.1362462342,0.102741234,0.5461699963,0.0042751282,-0.2236464322,-0.3965302408,-0.0348817892,-0.0644250065,0.0113105457,-0.0274784882,-0.2196493,-0.0800055563,0.0886778384,-0.1124985889,0.8294611573,0.0559734516,0.1079068482,0.3164263368,-0.1620282382,0.3252575099,-0.3066837192,0.273648113,-0.3532904387,-0.2786344588,-0.0291732792,-0.149663344,0.1126587912,-0.1508237273,-0.2704812586,0.3974288404,-0.1038571522,-0.015174468,-0.0373830833,0.0277132653,-0.056335479,0.1403032243,-0.3439783156,0.2541297972,-0.2757758796,0.3824891746,-0.2450356781,-0.2291565984,0.0792995021,-0.3686657548,-0.0293783993,0.255264461,-0.5707988143,0.3553519547,0.1545434296,-0.258582592,-0.0773392543,0.0948887467,0.0552250035,0.0117143029,-0.0373347886,0.253847003,-0.1581549197,0.0301682055,0.031538289,0.2192122787,0.1645492911,-0.2323606014,-0.2736302614,0.4664444923,-0.0604994744,-0.192182079,0.1606217474,0.0782424733,0.1615356952,-0.244144246,0.026363764,-0.2938996553,-0.2802439928,-0.1792211235,0.2058230788,-0.024694819,-0.2133593708,0.2009063959,0.229608044,-0.266056776,-0.0023667321,0.5637319684,0.0339706391,-0.1038133577,0.5721572042,0.0662537441,-0.0789105073,-0.3221276104,0.0294782743,-0.3145597577,-0.1750094444,-0.0270837154,0.0321228728,0.1296553314,0.0526438318,0.1740814894,0.0958074257,0.3109664321,-0.0165245272,-0.4803707898,-0.3675088584,-0.0251885485,-0.0307892747,0.1159087345,0.0154813239,0.3045248985,0.1293400526,-0.1455638111,-0.4137910903,0.038231656,-0.4060228467,-0.0250341333,0.163233608,-0.4319891632,0.2081795931,0.2303438932,0.2680647075,-0.1451085359,-0.303848058,-0.2964724898,0.0754910186,0.0826493502,0.0192133542,-0.1086061224,-0.1849854439,-0.2091994435,-0.0872026086,0.2196312249,0.0172618683,-0.052918084,0.0027895425,0.0397939347,-0.2459587455,0.0118430201,-0.0590337887,0.1182148233,0.0088244537,0.3148765564,0.0764694884,0.093558155,-0.2288632393,-0.2149313837,-0.049458921,0.0181147903,0.1174892932,-0.1726869643,0.4006973803,-0.2829346657,-0.0837286264,-0.1904572248,0.2987390757,0.4727976918,-0.247380197,0.0802390203,-0.0273023061,0.319771558,-0.1681524664,-0.0463615805,-0.1242021769,-0.0514262319,-0.0253467876,-0.0628121272,-0.2026316971,-0.1088778377,0.0428660214,0.1006502435,0.3122604787,-0.1423399001,0.0757191032,0.2222028077,-0.1228759959,-0.0651845485,0.1654010862,-0.1106474549,0.157167688,0.2767724395,-0.2756948173,0.1414103806,-0.2967186868,-0.0063293558,-0.0764336661,-0.30286327,-0.1283053905,0.0337629728,-0.225846678,-0.2506494224,-0.1360699683,0.3380437195,-0.0385781787,-0.1033947766,-0.3024477661,0.120288752,-0.2053774893,-0.2039037198,0.0223923493,-0.1146414801,-0.1198232919,-0.0627378151,0.0718694031,0.001259504,0.0875681713,0.1861035377,0.0112271942,-0.0528696217,-0.0873607695,0.2029916495,-0.0180562679,-0.1929122955,0.288275212,0.2056521177,0.0747296587,-0.2149980813,0.2968590856,0.5699777007,0.3594922721,-0.1515306085,0.1315045655,0.1248291582,-0.0330185369,-0.0403567292,-0.0303487796,0.2941537499,0.0763506144,0.3400413394,0.2877489626,-0.1894121766,0.3914562464,-0.0686069652,0.1110327691,-0.0636682585,0.1955822557,0.1219102293,-0.0993115231,-0.4362105429,0.2374273688,-0.687033534,0.017641779,0.1313972324,0.0855285898,0.2730656862,-0.2660717666,0.1238822043,-0.2409586012,0.6968002915,0.2872368693,0.2240161002,-0.3663333058,-0.0454765372,-0.419279635,0.3314238787,0.0393611975,0.0967454389,-0.1452009976,0.1707276851,-0.13172777,0.192718491,0.3683382571,-0.0617673397,0.0475332029,0.0752871931,-0.3002041876,-0.0893396437,0.1133302972,0.1212982759,0.0776123852,-0.187150389,0.1649665833,-0.4068341553,0.2258765996,0.0177901667,-0.2796664834,0.0114736119,-0.0562083535,0.4100193679,0.0544079691,0.4602577984,-0.0172595866,0.0323372036,-0.063857846,-0.3920346797,0.1263432801,0.1834335625,0.21989429,-0.0330676883,0.0796693563,0.0409015231,-0.3526593745,0.1972060949,-0.0334698819,-0.051827807,-0.0524145886,-0.0014031201,0.3012179434,0.087695457,-0.0379099473,-0.1161682904,-0.0168769825,-0.0071442807,-0.5099958181,-0.3519430459,0.4959525168,0.1254892349,-0.0674904734,-0.2154363692,0.3892975152,-0.0426002704,0.2377911806,-0.3732434809,0.3449625969,0.3041445315,-0.0472727157,-0.3802525699,0.3203733563,-0.1481854171,0.3062384129,0.002188995,-0.0563650243,-0.105674617,-0.0378368646,-0.1125168428,-0.109581545]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1594","title":"connection error ","comments":"Yes currently there's no retry afaik. We should add retries","body":"Hi\r\nI am hitting to this error, thanks \r\n\r\n```\r\n> Traceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 379, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 208, in main\r\n if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO\r\n File \"finetune_t5_trainer.py\", line 207, in <dictcomp>\r\n for task in data_args.eval_tasks}\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 66, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/boolq\/boolq.py\r\nel\/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED\r\n```","comment_length":10,"text":"connection error \nHi\r\nI am hitting to this error, thanks \r\n\r\n```\r\n> Traceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 379, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 208, in main\r\n if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO\r\n File \"finetune_t5_trainer.py\", line 207, in <dictcomp>\r\n for task in data_args.eval_tasks}\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 66, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/boolq\/boolq.py\r\nel\/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED\r\n```\nYes currently there's no retry afaik. We should add retries","embeddings":[-0.3293083012,-0.1345852762,-0.2016925067,0.3231747746,0.5001865029,-0.1886005998,0.1602738351,0.2997075617,-0.3055787683,0.1561734229,0.012656942,-0.0056440663,0.2251010835,0.1642649472,-0.0455740355,-0.0660552904,-0.2141919881,0.0516924225,-0.1668691933,0.1017805934,-0.0759902298,0.0589256845,-0.135654822,0.1066663116,-0.0014342469,-0.0925262198,0.0482695475,0.0914029479,-0.1770411283,-0.3981225789,0.3847735524,-0.0475312918,-0.0878206268,0.2440319657,-0.0001026009,0.2239751518,0.1840706468,0.0613851659,-0.2191357613,-0.165497303,-0.1825285554,-0.1764973998,0.3466186523,-0.2045548558,-0.323461026,0.156537205,0.0194423646,0.0245129764,0.6095243096,0.5391385555,0.3651290834,0.1239325255,0.2508780658,-0.1263591349,0.0121593978,-0.4649489522,-0.0548876636,0.2500353456,0.2293919772,-0.1382567436,0.0677758455,0.0419034436,0.0411453992,-0.0182585083,0.0719028115,-0.1040173098,0.1631603092,-0.3608938754,0.2469529659,0.1152732894,0.5638700128,0.0516682491,-0.1345327944,0.2436355948,0.0145922545,-0.2167304009,0.2574052811,0.0059991763,-0.2366169691,0.1825777292,0.1746409535,-0.1708101779,-0.5097798705,0.229768768,-0.093413502,0.2104021162,-0.0483462103,0.0594062284,0.0371313393,-0.2034060955,0.3239794672,0.2822365165,0.0639957413,0.095603019,-0.4873182774,-0.0258345623,-0.0517093278,-0.1743869781,-0.0484157056,0.1437137574,0.2108507156,0.0826344863,0.1433905512,0.0551506393,0.0877403691,0.1378038824,-0.1206299141,0.2517746687,0.2280283868,0.3546999991,-0.1444411874,-0.0109692346,-0.1517023295,-0.524538219,0.0227191225,0.3155859113,0.3285774887,-0.0800363272,-0.2814172208,0.1608621478,-0.073479332,0.0374105051,-0.0147579722,0.5862597227,-0.4222016037,0.1364524961,-0.0385238677,0.0206818804,-0.1698478311,0.0816259533,-0.2103554457,0.128688857,-0.1109872386,-0.2046239078,0.1267408729,-0.0979281366,0.1036819816,-0.1254515946,0.0796652734,-0.1540387422,0.2098276913,-0.3071586192,0.0403931066,0.2144149095,-0.0093317134,-0.0397677943,0.256238997,0.0868284404,0.1426610798,-0.0583050475,-0.3986396194,-0.4699076712,-0.0246357843,0.3337431848,0.1888823807,-0.1018986925,0.242503643,-0.239442572,0.1015270948,-0.1859617233,-0.1586317867,-0.3791605234,-0.1898315698,-0.2785962224,0.3494767845,0.2411410213,-0.2269233316,-0.0555184782,-0.0063581206,-0.2299630344,-0.0841925368,0.1094194576,-0.1654981077,0.1853645891,-0.2126497775,-0.0556343347,0.6239518523,-0.6403186321,-0.5884609222,0.451880306,-0.2488524467,-0.2172074914,0.0026223192,0.0128350044,0.31572631,0.1871753633,0.2534630299,0.2531638145,-0.0963229612,-0.0373811014,-0.1279286742,-0.2783178389,0.210431397,0.2529270649,0.338386029,-0.0542290099,0.1496915519,-0.2224408686,0.2122184187,0.1233668029,0.055118911,0.2695868015,0.2858492136,0.1266420037,-0.0151618924,0.2609181404,-0.2238556445,0.1011382937,-0.0823274553,0.1316879541,-0.4020649791,0.2130479068,-0.3046886325,-0.070966728,-0.4243715107,-0.1110660285,0.3290539384,0.0198725909,0.1009851322,0.0605011731,-0.0466322526,0.2943363786,-0.151177302,0.1532868445,-0.0333023034,0.2780037224,-0.3268547356,-0.0552035421,0.0937715843,0.0030752169,0.3442077935,-0.1848721802,-0.1886288971,0.2643786371,-0.159370333,0.1300932914,0.2049147338,0.0657288283,0.11827223,-0.448369503,-0.0238029417,0.0437438563,0.1285386086,-0.0440290347,0.4191423953,0.3036993146,0.029638866,0.3536557257,0.1969596446,0.1627426744,0.30114609,0.0441382863,-0.1183543652,0.1699621975,0.1193344668,0.1104473621,0.2826167047,-0.2836562991,-0.0844766945,0.0332247652,0.164361313,-0.0335939415,0.0535354093,0.0687814355,0.2974520922,-0.0559393056,-0.1830288321,0.1009347588,0.4425370097,0.0588133782,-0.1450980455,0.2595957816,0.0428838842,-0.1643870771,0.3133176267,0.0137783522,0.231696859,0.3095141351,0.0246753413,-0.0696165934,-0.1814823747,-0.2767195404,-0.1199273095,0.334752351,-0.1648125499,0.2214969695,-0.0891913995,-0.1461733282,0.0719188824,-0.1294409186,-0.1530590802,-0.2924027145,-0.0076095946,-0.0767282844,0.1505491138,0.2648813426,-0.1862804443,0.2516584396,0.3692764044,-0.1475120485,0.0229872093,-0.3701071739,-0.1531626284,0.2031988502,-0.0022343313,-0.3685530126,0.2077259868,-0.1924225092,-0.218137607,-0.1249535531,-0.1477749348,0.023245059,0.0012742316,0.0432976708,0.256049335,0.3906775117,0.1269900203,-0.0520312451,0.4064903259,-0.0884209946,-0.0913150236,0.0280723516,-0.0236177891,0.0279100221,0.0800108612,-0.6284663677,-0.5869358182,-0.294239372,0.1182904765,0.1431990862,0.1980281472,0.24892205,0.1683931798,0.1541638076,0.0365752652,0.318130374,-0.2569623291,-0.5267910361,0.1774822176,-0.1706408262,-0.4868075848,0.2212446034,0.1338598132,0.5196148753,-0.0392800197,-0.4721224606,-0.4179493487,0.0700986385,0.1188787967,-0.316524297,-0.1002306342,0.2164641619,-0.0834572911,-0.1960401684,0.0654206201,-0.0279408097,0.178632319,-0.000116027,0.2409600466,-0.1367491931,0.4993528426,-0.0066893827,0.4120452404,-0.0015904063,0.0416391492,0.173962459,-0.2517149448,0.1825367659,0.0320855603,-0.3973609209,0.1971779168,-0.0506810173,-0.1291242838,0.0730498731,-0.1645355672,0.1909097433,-0.5621136427,0.0427492596,-0.3912924826,-0.2510988116,-0.1200514883,-0.266015619,0.2810431421,0.0346744284,0.0859910771,-0.0982130542,0.0313237309,0.0420694649,0.3615584671,-0.1767034531,0.0078708092,0.0109080775,-0.0729205012,-0.352217108,0.2808557749,-0.0276010856,0.5090871453,-0.4446038604,0.0221328009,0.086178422,-0.0871059448,0.6551527381,-0.2658198178,0.1087110192,0.1216820404,-0.0544465892,-0.0370873436,-0.1218068078,-0.11756441,0.0180059187,0.1102109849,0.0187622011,-0.2171086669,0.0363610685,0.1376867294,-0.2447354645,-0.3150658011,0.0254317112,-0.2848104239,-0.3694692552,-0.0877855867,0.020495709,-0.0204049256,0.2766889632,0.0878966674,-0.1161073819,0.0979298651,-0.113330394,0.0173429102,0.0017108119,-0.0560506918,-0.1787100434,0.0288764928,0.1128022969,0.1006798819,0.2440427393,0.4235987365,-0.0448745713,-0.430036068,0.107609652,0.0284702014,0.2794895172,0.1612032652,-0.1627589315,0.2013744712,0.1291829497,0.3139091134,0.1362462342,0.102741234,0.5461699963,0.0042751282,-0.2236464322,-0.3965302408,-0.0348817892,-0.0644250065,0.0113105457,-0.0274784882,-0.2196493,-0.0800055563,0.0886778384,-0.1124985889,0.8294611573,0.0559734516,0.1079068482,0.3164263368,-0.1620282382,0.3252575099,-0.3066837192,0.273648113,-0.3532904387,-0.2786344588,-0.0291732792,-0.149663344,0.1126587912,-0.1508237273,-0.2704812586,0.3974288404,-0.1038571522,-0.015174468,-0.0373830833,0.0277132653,-0.056335479,0.1403032243,-0.3439783156,0.2541297972,-0.2757758796,0.3824891746,-0.2450356781,-0.2291565984,0.0792995021,-0.3686657548,-0.0293783993,0.255264461,-0.5707988143,0.3553519547,0.1545434296,-0.258582592,-0.0773392543,0.0948887467,0.0552250035,0.0117143029,-0.0373347886,0.253847003,-0.1581549197,0.0301682055,0.031538289,0.2192122787,0.1645492911,-0.2323606014,-0.2736302614,0.4664444923,-0.0604994744,-0.192182079,0.1606217474,0.0782424733,0.1615356952,-0.244144246,0.026363764,-0.2938996553,-0.2802439928,-0.1792211235,0.2058230788,-0.024694819,-0.2133593708,0.2009063959,0.229608044,-0.266056776,-0.0023667321,0.5637319684,0.0339706391,-0.1038133577,0.5721572042,0.0662537441,-0.0789105073,-0.3221276104,0.0294782743,-0.3145597577,-0.1750094444,-0.0270837154,0.0321228728,0.1296553314,0.0526438318,0.1740814894,0.0958074257,0.3109664321,-0.0165245272,-0.4803707898,-0.3675088584,-0.0251885485,-0.0307892747,0.1159087345,0.0154813239,0.3045248985,0.1293400526,-0.1455638111,-0.4137910903,0.038231656,-0.4060228467,-0.0250341333,0.163233608,-0.4319891632,0.2081795931,0.2303438932,0.2680647075,-0.1451085359,-0.303848058,-0.2964724898,0.0754910186,0.0826493502,0.0192133542,-0.1086061224,-0.1849854439,-0.2091994435,-0.0872026086,0.2196312249,0.0172618683,-0.052918084,0.0027895425,0.0397939347,-0.2459587455,0.0118430201,-0.0590337887,0.1182148233,0.0088244537,0.3148765564,0.0764694884,0.093558155,-0.2288632393,-0.2149313837,-0.049458921,0.0181147903,0.1174892932,-0.1726869643,0.4006973803,-0.2829346657,-0.0837286264,-0.1904572248,0.2987390757,0.4727976918,-0.247380197,0.0802390203,-0.0273023061,0.319771558,-0.1681524664,-0.0463615805,-0.1242021769,-0.0514262319,-0.0253467876,-0.0628121272,-0.2026316971,-0.1088778377,0.0428660214,0.1006502435,0.3122604787,-0.1423399001,0.0757191032,0.2222028077,-0.1228759959,-0.0651845485,0.1654010862,-0.1106474549,0.157167688,0.2767724395,-0.2756948173,0.1414103806,-0.2967186868,-0.0063293558,-0.0764336661,-0.30286327,-0.1283053905,0.0337629728,-0.225846678,-0.2506494224,-0.1360699683,0.3380437195,-0.0385781787,-0.1033947766,-0.3024477661,0.120288752,-0.2053774893,-0.2039037198,0.0223923493,-0.1146414801,-0.1198232919,-0.0627378151,0.0718694031,0.001259504,0.0875681713,0.1861035377,0.0112271942,-0.0528696217,-0.0873607695,0.2029916495,-0.0180562679,-0.1929122955,0.288275212,0.2056521177,0.0747296587,-0.2149980813,0.2968590856,0.5699777007,0.3594922721,-0.1515306085,0.1315045655,0.1248291582,-0.0330185369,-0.0403567292,-0.0303487796,0.2941537499,0.0763506144,0.3400413394,0.2877489626,-0.1894121766,0.3914562464,-0.0686069652,0.1110327691,-0.0636682585,0.1955822557,0.1219102293,-0.0993115231,-0.4362105429,0.2374273688,-0.687033534,0.017641779,0.1313972324,0.0855285898,0.2730656862,-0.2660717666,0.1238822043,-0.2409586012,0.6968002915,0.2872368693,0.2240161002,-0.3663333058,-0.0454765372,-0.419279635,0.3314238787,0.0393611975,0.0967454389,-0.1452009976,0.1707276851,-0.13172777,0.192718491,0.3683382571,-0.0617673397,0.0475332029,0.0752871931,-0.3002041876,-0.0893396437,0.1133302972,0.1212982759,0.0776123852,-0.187150389,0.1649665833,-0.4068341553,0.2258765996,0.0177901667,-0.2796664834,0.0114736119,-0.0562083535,0.4100193679,0.0544079691,0.4602577984,-0.0172595866,0.0323372036,-0.063857846,-0.3920346797,0.1263432801,0.1834335625,0.21989429,-0.0330676883,0.0796693563,0.0409015231,-0.3526593745,0.1972060949,-0.0334698819,-0.051827807,-0.0524145886,-0.0014031201,0.3012179434,0.087695457,-0.0379099473,-0.1161682904,-0.0168769825,-0.0071442807,-0.5099958181,-0.3519430459,0.4959525168,0.1254892349,-0.0674904734,-0.2154363692,0.3892975152,-0.0426002704,0.2377911806,-0.3732434809,0.3449625969,0.3041445315,-0.0472727157,-0.3802525699,0.3203733563,-0.1481854171,0.3062384129,0.002188995,-0.0563650243,-0.105674617,-0.0378368646,-0.1125168428,-0.109581545]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1594","title":"connection error ","comments":"Retries were added in #1603 :) \r\nIt will be available in the next release","body":"Hi\r\nI am hitting to this error, thanks \r\n\r\n```\r\n> Traceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 379, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 208, in main\r\n if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO\r\n File \"finetune_t5_trainer.py\", line 207, in <dictcomp>\r\n for task in data_args.eval_tasks}\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 66, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/boolq\/boolq.py\r\nel\/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED\r\n```","comment_length":14,"text":"connection error \nHi\r\nI am hitting to this error, thanks \r\n\r\n```\r\n> Traceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 379, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 208, in main\r\n if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO\r\n File \"finetune_t5_trainer.py\", line 207, in <dictcomp>\r\n for task in data_args.eval_tasks}\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 66, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/boolq\/boolq.py\r\nel\/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED\r\n```\nRetries were added in #1603 :) \r\nIt will be available in the next release","embeddings":[-0.3293083012,-0.1345852762,-0.2016925067,0.3231747746,0.5001865029,-0.1886005998,0.1602738351,0.2997075617,-0.3055787683,0.1561734229,0.012656942,-0.0056440663,0.2251010835,0.1642649472,-0.0455740355,-0.0660552904,-0.2141919881,0.0516924225,-0.1668691933,0.1017805934,-0.0759902298,0.0589256845,-0.135654822,0.1066663116,-0.0014342469,-0.0925262198,0.0482695475,0.0914029479,-0.1770411283,-0.3981225789,0.3847735524,-0.0475312918,-0.0878206268,0.2440319657,-0.0001026009,0.2239751518,0.1840706468,0.0613851659,-0.2191357613,-0.165497303,-0.1825285554,-0.1764973998,0.3466186523,-0.2045548558,-0.323461026,0.156537205,0.0194423646,0.0245129764,0.6095243096,0.5391385555,0.3651290834,0.1239325255,0.2508780658,-0.1263591349,0.0121593978,-0.4649489522,-0.0548876636,0.2500353456,0.2293919772,-0.1382567436,0.0677758455,0.0419034436,0.0411453992,-0.0182585083,0.0719028115,-0.1040173098,0.1631603092,-0.3608938754,0.2469529659,0.1152732894,0.5638700128,0.0516682491,-0.1345327944,0.2436355948,0.0145922545,-0.2167304009,0.2574052811,0.0059991763,-0.2366169691,0.1825777292,0.1746409535,-0.1708101779,-0.5097798705,0.229768768,-0.093413502,0.2104021162,-0.0483462103,0.0594062284,0.0371313393,-0.2034060955,0.3239794672,0.2822365165,0.0639957413,0.095603019,-0.4873182774,-0.0258345623,-0.0517093278,-0.1743869781,-0.0484157056,0.1437137574,0.2108507156,0.0826344863,0.1433905512,0.0551506393,0.0877403691,0.1378038824,-0.1206299141,0.2517746687,0.2280283868,0.3546999991,-0.1444411874,-0.0109692346,-0.1517023295,-0.524538219,0.0227191225,0.3155859113,0.3285774887,-0.0800363272,-0.2814172208,0.1608621478,-0.073479332,0.0374105051,-0.0147579722,0.5862597227,-0.4222016037,0.1364524961,-0.0385238677,0.0206818804,-0.1698478311,0.0816259533,-0.2103554457,0.128688857,-0.1109872386,-0.2046239078,0.1267408729,-0.0979281366,0.1036819816,-0.1254515946,0.0796652734,-0.1540387422,0.2098276913,-0.3071586192,0.0403931066,0.2144149095,-0.0093317134,-0.0397677943,0.256238997,0.0868284404,0.1426610798,-0.0583050475,-0.3986396194,-0.4699076712,-0.0246357843,0.3337431848,0.1888823807,-0.1018986925,0.242503643,-0.239442572,0.1015270948,-0.1859617233,-0.1586317867,-0.3791605234,-0.1898315698,-0.2785962224,0.3494767845,0.2411410213,-0.2269233316,-0.0555184782,-0.0063581206,-0.2299630344,-0.0841925368,0.1094194576,-0.1654981077,0.1853645891,-0.2126497775,-0.0556343347,0.6239518523,-0.6403186321,-0.5884609222,0.451880306,-0.2488524467,-0.2172074914,0.0026223192,0.0128350044,0.31572631,0.1871753633,0.2534630299,0.2531638145,-0.0963229612,-0.0373811014,-0.1279286742,-0.2783178389,0.210431397,0.2529270649,0.338386029,-0.0542290099,0.1496915519,-0.2224408686,0.2122184187,0.1233668029,0.055118911,0.2695868015,0.2858492136,0.1266420037,-0.0151618924,0.2609181404,-0.2238556445,0.1011382937,-0.0823274553,0.1316879541,-0.4020649791,0.2130479068,-0.3046886325,-0.070966728,-0.4243715107,-0.1110660285,0.3290539384,0.0198725909,0.1009851322,0.0605011731,-0.0466322526,0.2943363786,-0.151177302,0.1532868445,-0.0333023034,0.2780037224,-0.3268547356,-0.0552035421,0.0937715843,0.0030752169,0.3442077935,-0.1848721802,-0.1886288971,0.2643786371,-0.159370333,0.1300932914,0.2049147338,0.0657288283,0.11827223,-0.448369503,-0.0238029417,0.0437438563,0.1285386086,-0.0440290347,0.4191423953,0.3036993146,0.029638866,0.3536557257,0.1969596446,0.1627426744,0.30114609,0.0441382863,-0.1183543652,0.1699621975,0.1193344668,0.1104473621,0.2826167047,-0.2836562991,-0.0844766945,0.0332247652,0.164361313,-0.0335939415,0.0535354093,0.0687814355,0.2974520922,-0.0559393056,-0.1830288321,0.1009347588,0.4425370097,0.0588133782,-0.1450980455,0.2595957816,0.0428838842,-0.1643870771,0.3133176267,0.0137783522,0.231696859,0.3095141351,0.0246753413,-0.0696165934,-0.1814823747,-0.2767195404,-0.1199273095,0.334752351,-0.1648125499,0.2214969695,-0.0891913995,-0.1461733282,0.0719188824,-0.1294409186,-0.1530590802,-0.2924027145,-0.0076095946,-0.0767282844,0.1505491138,0.2648813426,-0.1862804443,0.2516584396,0.3692764044,-0.1475120485,0.0229872093,-0.3701071739,-0.1531626284,0.2031988502,-0.0022343313,-0.3685530126,0.2077259868,-0.1924225092,-0.218137607,-0.1249535531,-0.1477749348,0.023245059,0.0012742316,0.0432976708,0.256049335,0.3906775117,0.1269900203,-0.0520312451,0.4064903259,-0.0884209946,-0.0913150236,0.0280723516,-0.0236177891,0.0279100221,0.0800108612,-0.6284663677,-0.5869358182,-0.294239372,0.1182904765,0.1431990862,0.1980281472,0.24892205,0.1683931798,0.1541638076,0.0365752652,0.318130374,-0.2569623291,-0.5267910361,0.1774822176,-0.1706408262,-0.4868075848,0.2212446034,0.1338598132,0.5196148753,-0.0392800197,-0.4721224606,-0.4179493487,0.0700986385,0.1188787967,-0.316524297,-0.1002306342,0.2164641619,-0.0834572911,-0.1960401684,0.0654206201,-0.0279408097,0.178632319,-0.000116027,0.2409600466,-0.1367491931,0.4993528426,-0.0066893827,0.4120452404,-0.0015904063,0.0416391492,0.173962459,-0.2517149448,0.1825367659,0.0320855603,-0.3973609209,0.1971779168,-0.0506810173,-0.1291242838,0.0730498731,-0.1645355672,0.1909097433,-0.5621136427,0.0427492596,-0.3912924826,-0.2510988116,-0.1200514883,-0.266015619,0.2810431421,0.0346744284,0.0859910771,-0.0982130542,0.0313237309,0.0420694649,0.3615584671,-0.1767034531,0.0078708092,0.0109080775,-0.0729205012,-0.352217108,0.2808557749,-0.0276010856,0.5090871453,-0.4446038604,0.0221328009,0.086178422,-0.0871059448,0.6551527381,-0.2658198178,0.1087110192,0.1216820404,-0.0544465892,-0.0370873436,-0.1218068078,-0.11756441,0.0180059187,0.1102109849,0.0187622011,-0.2171086669,0.0363610685,0.1376867294,-0.2447354645,-0.3150658011,0.0254317112,-0.2848104239,-0.3694692552,-0.0877855867,0.020495709,-0.0204049256,0.2766889632,0.0878966674,-0.1161073819,0.0979298651,-0.113330394,0.0173429102,0.0017108119,-0.0560506918,-0.1787100434,0.0288764928,0.1128022969,0.1006798819,0.2440427393,0.4235987365,-0.0448745713,-0.430036068,0.107609652,0.0284702014,0.2794895172,0.1612032652,-0.1627589315,0.2013744712,0.1291829497,0.3139091134,0.1362462342,0.102741234,0.5461699963,0.0042751282,-0.2236464322,-0.3965302408,-0.0348817892,-0.0644250065,0.0113105457,-0.0274784882,-0.2196493,-0.0800055563,0.0886778384,-0.1124985889,0.8294611573,0.0559734516,0.1079068482,0.3164263368,-0.1620282382,0.3252575099,-0.3066837192,0.273648113,-0.3532904387,-0.2786344588,-0.0291732792,-0.149663344,0.1126587912,-0.1508237273,-0.2704812586,0.3974288404,-0.1038571522,-0.015174468,-0.0373830833,0.0277132653,-0.056335479,0.1403032243,-0.3439783156,0.2541297972,-0.2757758796,0.3824891746,-0.2450356781,-0.2291565984,0.0792995021,-0.3686657548,-0.0293783993,0.255264461,-0.5707988143,0.3553519547,0.1545434296,-0.258582592,-0.0773392543,0.0948887467,0.0552250035,0.0117143029,-0.0373347886,0.253847003,-0.1581549197,0.0301682055,0.031538289,0.2192122787,0.1645492911,-0.2323606014,-0.2736302614,0.4664444923,-0.0604994744,-0.192182079,0.1606217474,0.0782424733,0.1615356952,-0.244144246,0.026363764,-0.2938996553,-0.2802439928,-0.1792211235,0.2058230788,-0.024694819,-0.2133593708,0.2009063959,0.229608044,-0.266056776,-0.0023667321,0.5637319684,0.0339706391,-0.1038133577,0.5721572042,0.0662537441,-0.0789105073,-0.3221276104,0.0294782743,-0.3145597577,-0.1750094444,-0.0270837154,0.0321228728,0.1296553314,0.0526438318,0.1740814894,0.0958074257,0.3109664321,-0.0165245272,-0.4803707898,-0.3675088584,-0.0251885485,-0.0307892747,0.1159087345,0.0154813239,0.3045248985,0.1293400526,-0.1455638111,-0.4137910903,0.038231656,-0.4060228467,-0.0250341333,0.163233608,-0.4319891632,0.2081795931,0.2303438932,0.2680647075,-0.1451085359,-0.303848058,-0.2964724898,0.0754910186,0.0826493502,0.0192133542,-0.1086061224,-0.1849854439,-0.2091994435,-0.0872026086,0.2196312249,0.0172618683,-0.052918084,0.0027895425,0.0397939347,-0.2459587455,0.0118430201,-0.0590337887,0.1182148233,0.0088244537,0.3148765564,0.0764694884,0.093558155,-0.2288632393,-0.2149313837,-0.049458921,0.0181147903,0.1174892932,-0.1726869643,0.4006973803,-0.2829346657,-0.0837286264,-0.1904572248,0.2987390757,0.4727976918,-0.247380197,0.0802390203,-0.0273023061,0.319771558,-0.1681524664,-0.0463615805,-0.1242021769,-0.0514262319,-0.0253467876,-0.0628121272,-0.2026316971,-0.1088778377,0.0428660214,0.1006502435,0.3122604787,-0.1423399001,0.0757191032,0.2222028077,-0.1228759959,-0.0651845485,0.1654010862,-0.1106474549,0.157167688,0.2767724395,-0.2756948173,0.1414103806,-0.2967186868,-0.0063293558,-0.0764336661,-0.30286327,-0.1283053905,0.0337629728,-0.225846678,-0.2506494224,-0.1360699683,0.3380437195,-0.0385781787,-0.1033947766,-0.3024477661,0.120288752,-0.2053774893,-0.2039037198,0.0223923493,-0.1146414801,-0.1198232919,-0.0627378151,0.0718694031,0.001259504,0.0875681713,0.1861035377,0.0112271942,-0.0528696217,-0.0873607695,0.2029916495,-0.0180562679,-0.1929122955,0.288275212,0.2056521177,0.0747296587,-0.2149980813,0.2968590856,0.5699777007,0.3594922721,-0.1515306085,0.1315045655,0.1248291582,-0.0330185369,-0.0403567292,-0.0303487796,0.2941537499,0.0763506144,0.3400413394,0.2877489626,-0.1894121766,0.3914562464,-0.0686069652,0.1110327691,-0.0636682585,0.1955822557,0.1219102293,-0.0993115231,-0.4362105429,0.2374273688,-0.687033534,0.017641779,0.1313972324,0.0855285898,0.2730656862,-0.2660717666,0.1238822043,-0.2409586012,0.6968002915,0.2872368693,0.2240161002,-0.3663333058,-0.0454765372,-0.419279635,0.3314238787,0.0393611975,0.0967454389,-0.1452009976,0.1707276851,-0.13172777,0.192718491,0.3683382571,-0.0617673397,0.0475332029,0.0752871931,-0.3002041876,-0.0893396437,0.1133302972,0.1212982759,0.0776123852,-0.187150389,0.1649665833,-0.4068341553,0.2258765996,0.0177901667,-0.2796664834,0.0114736119,-0.0562083535,0.4100193679,0.0544079691,0.4602577984,-0.0172595866,0.0323372036,-0.063857846,-0.3920346797,0.1263432801,0.1834335625,0.21989429,-0.0330676883,0.0796693563,0.0409015231,-0.3526593745,0.1972060949,-0.0334698819,-0.051827807,-0.0524145886,-0.0014031201,0.3012179434,0.087695457,-0.0379099473,-0.1161682904,-0.0168769825,-0.0071442807,-0.5099958181,-0.3519430459,0.4959525168,0.1254892349,-0.0674904734,-0.2154363692,0.3892975152,-0.0426002704,0.2377911806,-0.3732434809,0.3449625969,0.3041445315,-0.0472727157,-0.3802525699,0.3203733563,-0.1481854171,0.3062384129,0.002188995,-0.0563650243,-0.105674617,-0.0378368646,-0.1125168428,-0.109581545]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1594","title":"connection error ","comments":"Hi @lhoestq thank you for the modification, I will use`script_version=\"master\"` for now :), to my experience, also setting timeout to a larger number like 3*60 which I normally use helps a lot on this.\r\n","body":"Hi\r\nI am hitting to this error, thanks \r\n\r\n```\r\n> Traceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 379, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 208, in main\r\n if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO\r\n File \"finetune_t5_trainer.py\", line 207, in <dictcomp>\r\n for task in data_args.eval_tasks}\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 66, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/boolq\/boolq.py\r\nel\/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED\r\n```","comment_length":34,"text":"connection error \nHi\r\nI am hitting to this error, thanks \r\n\r\n```\r\n> Traceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 379, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 208, in main\r\n if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO\r\n File \"finetune_t5_trainer.py\", line 207, in <dictcomp>\r\n for task in data_args.eval_tasks}\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 66, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split, script_version=\"master\")\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 267, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 487, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/master\/datasets\/boolq\/boolq.py\r\nel\/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED\r\n```\nHi @lhoestq thank you for the modification, I will use`script_version=\"master\"` for now :), to my experience, also setting timeout to a larger number like 3*60 which I normally use helps a lot on this.\r\n","embeddings":[-0.3293083012,-0.1345852762,-0.2016925067,0.3231747746,0.5001865029,-0.1886005998,0.1602738351,0.2997075617,-0.3055787683,0.1561734229,0.012656942,-0.0056440663,0.2251010835,0.1642649472,-0.0455740355,-0.0660552904,-0.2141919881,0.0516924225,-0.1668691933,0.1017805934,-0.0759902298,0.0589256845,-0.135654822,0.1066663116,-0.0014342469,-0.0925262198,0.0482695475,0.0914029479,-0.1770411283,-0.3981225789,0.3847735524,-0.0475312918,-0.0878206268,0.2440319657,-0.0001026009,0.2239751518,0.1840706468,0.0613851659,-0.2191357613,-0.165497303,-0.1825285554,-0.1764973998,0.3466186523,-0.2045548558,-0.323461026,0.156537205,0.0194423646,0.0245129764,0.6095243096,0.5391385555,0.3651290834,0.1239325255,0.2508780658,-0.1263591349,0.0121593978,-0.4649489522,-0.0548876636,0.2500353456,0.2293919772,-0.1382567436,0.0677758455,0.0419034436,0.0411453992,-0.0182585083,0.0719028115,-0.1040173098,0.1631603092,-0.3608938754,0.2469529659,0.1152732894,0.5638700128,0.0516682491,-0.1345327944,0.2436355948,0.0145922545,-0.2167304009,0.2574052811,0.0059991763,-0.2366169691,0.1825777292,0.1746409535,-0.1708101779,-0.5097798705,0.229768768,-0.093413502,0.2104021162,-0.0483462103,0.0594062284,0.0371313393,-0.2034060955,0.3239794672,0.2822365165,0.0639957413,0.095603019,-0.4873182774,-0.0258345623,-0.0517093278,-0.1743869781,-0.0484157056,0.1437137574,0.2108507156,0.0826344863,0.1433905512,0.0551506393,0.0877403691,0.1378038824,-0.1206299141,0.2517746687,0.2280283868,0.3546999991,-0.1444411874,-0.0109692346,-0.1517023295,-0.524538219,0.0227191225,0.3155859113,0.3285774887,-0.0800363272,-0.2814172208,0.1608621478,-0.073479332,0.0374105051,-0.0147579722,0.5862597227,-0.4222016037,0.1364524961,-0.0385238677,0.0206818804,-0.1698478311,0.0816259533,-0.2103554457,0.128688857,-0.1109872386,-0.2046239078,0.1267408729,-0.0979281366,0.1036819816,-0.1254515946,0.0796652734,-0.1540387422,0.2098276913,-0.3071586192,0.0403931066,0.2144149095,-0.0093317134,-0.0397677943,0.256238997,0.0868284404,0.1426610798,-0.0583050475,-0.3986396194,-0.4699076712,-0.0246357843,0.3337431848,0.1888823807,-0.1018986925,0.242503643,-0.239442572,0.1015270948,-0.1859617233,-0.1586317867,-0.3791605234,-0.1898315698,-0.2785962224,0.3494767845,0.2411410213,-0.2269233316,-0.0555184782,-0.0063581206,-0.2299630344,-0.0841925368,0.1094194576,-0.1654981077,0.1853645891,-0.2126497775,-0.0556343347,0.6239518523,-0.6403186321,-0.5884609222,0.451880306,-0.2488524467,-0.2172074914,0.0026223192,0.0128350044,0.31572631,0.1871753633,0.2534630299,0.2531638145,-0.0963229612,-0.0373811014,-0.1279286742,-0.2783178389,0.210431397,0.2529270649,0.338386029,-0.0542290099,0.1496915519,-0.2224408686,0.2122184187,0.1233668029,0.055118911,0.2695868015,0.2858492136,0.1266420037,-0.0151618924,0.2609181404,-0.2238556445,0.1011382937,-0.0823274553,0.1316879541,-0.4020649791,0.2130479068,-0.3046886325,-0.070966728,-0.4243715107,-0.1110660285,0.3290539384,0.0198725909,0.1009851322,0.0605011731,-0.0466322526,0.2943363786,-0.151177302,0.1532868445,-0.0333023034,0.2780037224,-0.3268547356,-0.0552035421,0.0937715843,0.0030752169,0.3442077935,-0.1848721802,-0.1886288971,0.2643786371,-0.159370333,0.1300932914,0.2049147338,0.0657288283,0.11827223,-0.448369503,-0.0238029417,0.0437438563,0.1285386086,-0.0440290347,0.4191423953,0.3036993146,0.029638866,0.3536557257,0.1969596446,0.1627426744,0.30114609,0.0441382863,-0.1183543652,0.1699621975,0.1193344668,0.1104473621,0.2826167047,-0.2836562991,-0.0844766945,0.0332247652,0.164361313,-0.0335939415,0.0535354093,0.0687814355,0.2974520922,-0.0559393056,-0.1830288321,0.1009347588,0.4425370097,0.0588133782,-0.1450980455,0.2595957816,0.0428838842,-0.1643870771,0.3133176267,0.0137783522,0.231696859,0.3095141351,0.0246753413,-0.0696165934,-0.1814823747,-0.2767195404,-0.1199273095,0.334752351,-0.1648125499,0.2214969695,-0.0891913995,-0.1461733282,0.0719188824,-0.1294409186,-0.1530590802,-0.2924027145,-0.0076095946,-0.0767282844,0.1505491138,0.2648813426,-0.1862804443,0.2516584396,0.3692764044,-0.1475120485,0.0229872093,-0.3701071739,-0.1531626284,0.2031988502,-0.0022343313,-0.3685530126,0.2077259868,-0.1924225092,-0.218137607,-0.1249535531,-0.1477749348,0.023245059,0.0012742316,0.0432976708,0.256049335,0.3906775117,0.1269900203,-0.0520312451,0.4064903259,-0.0884209946,-0.0913150236,0.0280723516,-0.0236177891,0.0279100221,0.0800108612,-0.6284663677,-0.5869358182,-0.294239372,0.1182904765,0.1431990862,0.1980281472,0.24892205,0.1683931798,0.1541638076,0.0365752652,0.318130374,-0.2569623291,-0.5267910361,0.1774822176,-0.1706408262,-0.4868075848,0.2212446034,0.1338598132,0.5196148753,-0.0392800197,-0.4721224606,-0.4179493487,0.0700986385,0.1188787967,-0.316524297,-0.1002306342,0.2164641619,-0.0834572911,-0.1960401684,0.0654206201,-0.0279408097,0.178632319,-0.000116027,0.2409600466,-0.1367491931,0.4993528426,-0.0066893827,0.4120452404,-0.0015904063,0.0416391492,0.173962459,-0.2517149448,0.1825367659,0.0320855603,-0.3973609209,0.1971779168,-0.0506810173,-0.1291242838,0.0730498731,-0.1645355672,0.1909097433,-0.5621136427,0.0427492596,-0.3912924826,-0.2510988116,-0.1200514883,-0.266015619,0.2810431421,0.0346744284,0.0859910771,-0.0982130542,0.0313237309,0.0420694649,0.3615584671,-0.1767034531,0.0078708092,0.0109080775,-0.0729205012,-0.352217108,0.2808557749,-0.0276010856,0.5090871453,-0.4446038604,0.0221328009,0.086178422,-0.0871059448,0.6551527381,-0.2658198178,0.1087110192,0.1216820404,-0.0544465892,-0.0370873436,-0.1218068078,-0.11756441,0.0180059187,0.1102109849,0.0187622011,-0.2171086669,0.0363610685,0.1376867294,-0.2447354645,-0.3150658011,0.0254317112,-0.2848104239,-0.3694692552,-0.0877855867,0.020495709,-0.0204049256,0.2766889632,0.0878966674,-0.1161073819,0.0979298651,-0.113330394,0.0173429102,0.0017108119,-0.0560506918,-0.1787100434,0.0288764928,0.1128022969,0.1006798819,0.2440427393,0.4235987365,-0.0448745713,-0.430036068,0.107609652,0.0284702014,0.2794895172,0.1612032652,-0.1627589315,0.2013744712,0.1291829497,0.3139091134,0.1362462342,0.102741234,0.5461699963,0.0042751282,-0.2236464322,-0.3965302408,-0.0348817892,-0.0644250065,0.0113105457,-0.0274784882,-0.2196493,-0.0800055563,0.0886778384,-0.1124985889,0.8294611573,0.0559734516,0.1079068482,0.3164263368,-0.1620282382,0.3252575099,-0.3066837192,0.273648113,-0.3532904387,-0.2786344588,-0.0291732792,-0.149663344,0.1126587912,-0.1508237273,-0.2704812586,0.3974288404,-0.1038571522,-0.015174468,-0.0373830833,0.0277132653,-0.056335479,0.1403032243,-0.3439783156,0.2541297972,-0.2757758796,0.3824891746,-0.2450356781,-0.2291565984,0.0792995021,-0.3686657548,-0.0293783993,0.255264461,-0.5707988143,0.3553519547,0.1545434296,-0.258582592,-0.0773392543,0.0948887467,0.0552250035,0.0117143029,-0.0373347886,0.253847003,-0.1581549197,0.0301682055,0.031538289,0.2192122787,0.1645492911,-0.2323606014,-0.2736302614,0.4664444923,-0.0604994744,-0.192182079,0.1606217474,0.0782424733,0.1615356952,-0.244144246,0.026363764,-0.2938996553,-0.2802439928,-0.1792211235,0.2058230788,-0.024694819,-0.2133593708,0.2009063959,0.229608044,-0.266056776,-0.0023667321,0.5637319684,0.0339706391,-0.1038133577,0.5721572042,0.0662537441,-0.0789105073,-0.3221276104,0.0294782743,-0.3145597577,-0.1750094444,-0.0270837154,0.0321228728,0.1296553314,0.0526438318,0.1740814894,0.0958074257,0.3109664321,-0.0165245272,-0.4803707898,-0.3675088584,-0.0251885485,-0.0307892747,0.1159087345,0.0154813239,0.3045248985,0.1293400526,-0.1455638111,-0.4137910903,0.038231656,-0.4060228467,-0.0250341333,0.163233608,-0.4319891632,0.2081795931,0.2303438932,0.2680647075,-0.1451085359,-0.303848058,-0.2964724898,0.0754910186,0.0826493502,0.0192133542,-0.1086061224,-0.1849854439,-0.2091994435,-0.0872026086,0.2196312249,0.0172618683,-0.052918084,0.0027895425,0.0397939347,-0.2459587455,0.0118430201,-0.0590337887,0.1182148233,0.0088244537,0.3148765564,0.0764694884,0.093558155,-0.2288632393,-0.2149313837,-0.049458921,0.0181147903,0.1174892932,-0.1726869643,0.4006973803,-0.2829346657,-0.0837286264,-0.1904572248,0.2987390757,0.4727976918,-0.247380197,0.0802390203,-0.0273023061,0.319771558,-0.1681524664,-0.0463615805,-0.1242021769,-0.0514262319,-0.0253467876,-0.0628121272,-0.2026316971,-0.1088778377,0.0428660214,0.1006502435,0.3122604787,-0.1423399001,0.0757191032,0.2222028077,-0.1228759959,-0.0651845485,0.1654010862,-0.1106474549,0.157167688,0.2767724395,-0.2756948173,0.1414103806,-0.2967186868,-0.0063293558,-0.0764336661,-0.30286327,-0.1283053905,0.0337629728,-0.225846678,-0.2506494224,-0.1360699683,0.3380437195,-0.0385781787,-0.1033947766,-0.3024477661,0.120288752,-0.2053774893,-0.2039037198,0.0223923493,-0.1146414801,-0.1198232919,-0.0627378151,0.0718694031,0.001259504,0.0875681713,0.1861035377,0.0112271942,-0.0528696217,-0.0873607695,0.2029916495,-0.0180562679,-0.1929122955,0.288275212,0.2056521177,0.0747296587,-0.2149980813,0.2968590856,0.5699777007,0.3594922721,-0.1515306085,0.1315045655,0.1248291582,-0.0330185369,-0.0403567292,-0.0303487796,0.2941537499,0.0763506144,0.3400413394,0.2877489626,-0.1894121766,0.3914562464,-0.0686069652,0.1110327691,-0.0636682585,0.1955822557,0.1219102293,-0.0993115231,-0.4362105429,0.2374273688,-0.687033534,0.017641779,0.1313972324,0.0855285898,0.2730656862,-0.2660717666,0.1238822043,-0.2409586012,0.6968002915,0.2872368693,0.2240161002,-0.3663333058,-0.0454765372,-0.419279635,0.3314238787,0.0393611975,0.0967454389,-0.1452009976,0.1707276851,-0.13172777,0.192718491,0.3683382571,-0.0617673397,0.0475332029,0.0752871931,-0.3002041876,-0.0893396437,0.1133302972,0.1212982759,0.0776123852,-0.187150389,0.1649665833,-0.4068341553,0.2258765996,0.0177901667,-0.2796664834,0.0114736119,-0.0562083535,0.4100193679,0.0544079691,0.4602577984,-0.0172595866,0.0323372036,-0.063857846,-0.3920346797,0.1263432801,0.1834335625,0.21989429,-0.0330676883,0.0796693563,0.0409015231,-0.3526593745,0.1972060949,-0.0334698819,-0.051827807,-0.0524145886,-0.0014031201,0.3012179434,0.087695457,-0.0379099473,-0.1161682904,-0.0168769825,-0.0071442807,-0.5099958181,-0.3519430459,0.4959525168,0.1254892349,-0.0674904734,-0.2154363692,0.3892975152,-0.0426002704,0.2377911806,-0.3732434809,0.3449625969,0.3041445315,-0.0472727157,-0.3802525699,0.3203733563,-0.1481854171,0.3062384129,0.002188995,-0.0563650243,-0.105674617,-0.0378368646,-0.1125168428,-0.109581545]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1593","title":"Access to key in DatasetDict map","comments":"Indeed that would be cool\r\n\r\nAlso FYI right now the easiest way to do this is\r\n```python\r\ndataset_dict[\"train\"] = dataset_dict[\"train\"].map(my_transform_for_the_train_set)\r\ndataset_dict[\"test\"] = dataset_dict[\"test\"].map(my_transform_for_the_test_set)\r\n```","body":"It is possible that we want to do different things in the `map` function (and possibly other functions too) of a `DatasetDict`, depending on the key. I understand that `DatasetDict.map` is a really thin wrapper of `Dataset.map`, so it is easy to directly implement this functionality in the client code. Still, it'd be nice if there can be a flag, similar to `with_indices`, that allows the callable to know the key inside `DatasetDict`.","comment_length":24,"text":"Access to key in DatasetDict map\nIt is possible that we want to do different things in the `map` function (and possibly other functions too) of a `DatasetDict`, depending on the key. I understand that `DatasetDict.map` is a really thin wrapper of `Dataset.map`, so it is easy to directly implement this functionality in the client code. Still, it'd be nice if there can be a flag, similar to `with_indices`, that allows the callable to know the key inside `DatasetDict`.\nIndeed that would be cool\r\n\r\nAlso FYI right now the easiest way to do this is\r\n```python\r\ndataset_dict[\"train\"] = dataset_dict[\"train\"].map(my_transform_for_the_train_set)\r\ndataset_dict[\"test\"] = dataset_dict[\"test\"].map(my_transform_for_the_test_set)\r\n```","embeddings":[0.0123616559,-0.0980538875,-0.1504415572,0.0315879881,0.1485255063,0.0519785024,0.2157814354,0.1856265515,0.2782222033,0.1137610897,0.0187772661,0.6971397996,-0.194628492,0.3624713123,0.0075963112,0.1261140257,-0.082267195,0.1726022512,0.259760946,0.0200805254,-0.5007085204,-0.2413183302,-0.2991085351,0.1724794954,0.1813092083,-0.058190126,-0.0412486903,-0.1254830509,0.0863479972,-0.4971494079,0.3150211871,0.3500140011,-0.0713424459,0.0319679268,-0.0001151014,0.165341422,0.1797243357,-0.0615737364,-0.2317584604,-0.2660992742,-0.4753479064,-0.2223521471,-0.0665304959,-0.648195982,0.0736432299,-0.2658869922,0.0899404213,-0.3869361579,0.1715967357,0.1491442025,0.1471835971,0.1689081043,-0.0654482991,0.2037469,-0.0100793177,0.0594722889,-0.0677720681,-0.1251585186,0.0909942687,-0.1609252095,0.1930097193,0.0897944868,-0.0347207636,-0.1856620163,0.4800027609,0.1962268353,-0.2003143132,-0.2015486062,-0.035012912,0.1073151156,0.257601738,-0.464533478,-0.4607774317,-0.349897176,0.0422466584,0.0605236515,0.1202891022,-0.1899610758,0.1220527738,0.2078469992,-0.0567756705,-0.0148770856,-0.1210499257,0.069634445,-0.1026101559,0.3232737184,0.0501525626,0.3766308129,0.1448031515,0.1034575552,-0.050728824,-0.1709432006,0.0642506629,0.4460495412,-0.0070257471,-0.1497470886,0.2523940206,-0.1559841633,0.2075985819,-0.3478935063,0.1729162782,0.1256625801,-0.4662780166,0.3151428699,0.382566154,0.1909714043,0.3423776031,0.4960179925,0.4341569245,-0.2414720207,-0.1015854925,0.0000163147,0.0529405624,-0.0984004512,0.3259883821,0.449416995,0.4494536519,0.1942183673,-0.0887192115,-0.142272234,-0.1836024076,-0.2230855525,0.2525042892,0.1441774219,0.1953484714,0.1295474321,-0.0843947977,-0.1063770875,0.1183463037,0.0830751881,-0.1132612005,-0.1972928047,0.1457876861,-0.0385118201,0.0284653343,-0.0229258575,0.0479106344,-0.1424472928,0.1269001067,0.1206097007,0.2216482311,-0.1108534709,0.4448646903,0.0949500576,-0.3591142893,-0.02338765,0.1648881137,-0.2535071373,-0.5074768066,0.0631086677,-0.2506268322,-0.3553037345,0.5641079545,0.1127244756,0.1348879039,0.0448333025,-0.1656806767,0.4262408614,0.0385414772,-0.1489331275,0.2924417555,-0.1452793479,-0.6354404688,-0.4432385862,-0.1224961877,0.0715213642,-0.4225706458,-0.2193415612,0.193688795,-0.2955155671,-0.2337773442,-0.1400133073,-0.4305417538,0.1845209599,-0.1133115068,0.0842298418,0.5983136892,-0.5417661667,-0.7054085732,-0.0194750242,-0.0734470561,-0.0909807459,0.1061515138,0.1983310878,0.3027149737,-0.1075788587,0.3345128894,0.2482724488,0.0704048499,0.0739826337,0.3476862907,-0.0775430724,0.0440010428,-0.1386542767,-0.2733567357,0.012396723,-0.0209333003,-0.1570063531,0.1605323255,-0.0444436334,0.1319596767,-0.0316115208,0.1468590945,-0.014786616,0.0897356346,-0.0011813468,-0.1574274004,0.4363519549,-0.0688039362,-0.1161404923,-0.3282233477,-0.5135014653,0.0457432643,0.3376213014,-0.1675526947,0.0446512923,0.1876629889,-0.0961411968,0.0183243807,-0.0490369909,-0.3668048382,0.0269409791,-0.0686387569,-0.1688911319,0.0345465615,-0.1081574336,0.2327875048,0.2283022553,0.0057086465,0.1308414042,0.1189313233,0.1284094006,-0.1436486691,-0.0352761261,0.3456194699,-0.0780207291,-0.1358102262,0.4973068535,0.1956263185,0.0363544971,0.0547037758,0.2495283335,0.1709625572,-0.1297852248,-0.142611891,0.5926755071,0.0285720006,0.5155330896,-0.0098347757,0.0482867248,0.0550444834,0.057674747,-0.4270036817,-0.3303960562,-0.4731330574,-0.1918272525,0.2141078115,0.0192024149,0.1139845699,0.1446620822,0.6021998525,0.0569199324,0.1399388462,-0.2470755428,-0.1606605053,0.1173473224,-0.0280296579,0.0974755585,0.229315713,0.1352882832,-0.0288372599,0.0935410336,-0.1350356638,0.1313690245,0.1300598681,0.3321389854,-0.1751943827,-0.285826683,0.0560939834,-0.0637323633,-0.0282833148,-0.3308681846,-0.0558249801,-0.0974879339,-0.1905638427,0.241342634,-0.0426226109,-0.0978617445,0.0608884171,-0.2186288387,0.0325925685,-0.3761589825,-0.0131312925,0.0083226291,-0.2312106341,0.1049044132,-0.1630046368,0.1837065965,0.1399992853,-0.3912048936,0.0317251533,-0.4720563293,0.036917083,0.0322630927,0.0279489849,-0.3296690881,0.3956876993,0.2419899553,0.0166119263,-0.1163709685,-0.2459565401,-0.005321688,-0.1730002463,-0.2134752423,0.3398188353,0.1414757967,-0.175737828,0.1314511448,0.1284939945,0.0330686271,-0.2387274206,-0.1916492283,0.055913426,-0.0331564508,0.0056467326,-0.0744236708,-0.2525494993,-0.3577523232,0.5741587877,-0.1613992602,0.2467688471,-0.0077684666,-0.0785122663,0.1846992224,0.1403214037,0.1610547751,-0.2782910168,0.2995181084,0.1700734049,-0.1171790063,-0.1032219529,-0.111432977,-0.2328278869,0.2313754708,0.1477987915,-0.1864003241,-0.1199128479,-0.0949169323,0.3826450408,-0.1426097453,0.0625523701,0.3247343898,0.3371560574,0.0288054533,-0.1656000465,-0.2728478014,-0.1100341082,0.3427269757,-0.3026394248,0.4693041444,0.3372187018,-0.0239047296,0.4323866665,-0.0281220581,-0.0266241487,0.086345084,-0.0989653617,0.0983515903,0.0146527821,-0.3411612511,-0.2429898828,0.1172859967,0.0156982988,0.0947756246,-0.1910479218,-0.191097796,-0.0792386234,-0.5995090604,-0.0115666874,-0.1916851997,0.2917588651,-0.1449211687,0.056501653,0.0385610722,0.252251029,-0.3942568004,0.1860198826,0.0070533748,-0.2436265647,0.3206495643,-0.2016779631,-0.1346752048,-0.4841751158,-0.2913571894,0.228464514,0.1066894904,0.1827000678,0.1247128397,-0.1861442477,0.0424533598,0.0351591036,0.6711293459,-0.090238072,-0.0433978252,0.1302250624,-0.1705904007,-0.472612381,0.1933234632,0.0955845267,0.6799287796,-0.4885380268,0.4142966866,-0.2821058333,0.1681352258,-0.2788057625,0.1424884349,-0.2792942524,-0.2370332927,-0.1003668159,0.0329027362,-0.3489783704,0.1554313153,0.0556014664,-0.205723688,-0.0410665199,0.0037355931,-0.1011340022,-0.0587351732,-0.0257110074,0.2758450806,0.1817137301,0.0462844782,0.148060292,0.3917503059,-0.0673678964,0.001990509,0.4635414481,0.0900651664,-0.6185475588,-0.2677064538,0.2093935609,0.3605566621,0.0001386261,0.0508040115,0.3652799428,0.2122522742,-0.3647414446,-0.2920520306,-0.0912646577,0.1370174438,0.1409702003,-0.6226903796,-0.8021765947,0.5334590673,0.1424503475,-0.4570407867,0.5476131439,0.1369370967,-0.5080655217,0.6939821243,0.0317472816,0.818079114,0.0909691676,0.2459729165,-0.1142297164,-0.2517312467,0.3870451152,-0.0805384368,-0.0942546427,-0.1135750934,-0.2276619822,-0.147387743,-0.0596042983,-0.1450408846,0.2415314913,-0.4093046486,0.1480799317,-0.0883655176,0.1170580611,-0.1651399434,0.0385311097,0.0123297134,-0.2198592871,-0.1952332109,0.0750327781,0.1890084147,0.1283808649,0.1366223097,-0.0801974609,-0.2220903635,-0.1737529188,0.0364484042,-0.3822324276,-0.519430995,0.4279226959,-0.249238044,0.21816881,0.0673161149,0.2284749597,0.4209683836,0.2316121459,-0.0853805542,0.1967369169,0.5073007345,0.2895573974,-0.0951865017,-0.1096264943,0.3067339957,0.238043353,-0.3285782039,0.3536573946,0.0079675196,-0.3847073019,-0.4506922364,0.4155710042,-0.0181305557,-0.2002573907,-0.0622832067,-0.1920297891,-0.1150195152,-0.0430936888,0.0938725099,-0.0051013767,0.0501457043,0.3668041527,-0.0498032346,-0.2165305763,0.1042746156,0.509755969,0.2294653803,0.3759138286,0.3757973015,-0.4711036086,-0.0614993274,-0.1008530036,-0.0579059981,0.2963011861,0.1827339679,0.0935970917,0.1716170907,0.0842455253,-0.3170439899,0.310810864,-0.0857541859,0.2972805798,-0.3940306902,-0.2748644352,-0.7087803483,0.3164017498,0.2487419695,0.2865431309,0.1933692247,-0.0142206606,-0.2886118591,-0.0271037724,-0.2733268738,-0.3420169055,0.0181578882,-0.0102351066,0.3442188203,0.4479816854,0.1761229187,0.0933119655,0.0755361542,0.0744336173,0.0688588843,-0.2235023975,-0.1801404804,0.1493206471,0.0245124996,0.2308112085,-0.3096633852,-0.4354640543,-0.0078956522,-0.2798788846,0.1082426384,0.4523063302,0.1347955912,0.2953732908,0.2547713816,0.0669461265,-0.0244195461,0.29912287,0.0422925763,0.0645180196,0.1117193326,0.362396121,-0.3781584203,0.0310489126,-0.0688535571,0.1380793899,-0.0055247415,-0.1330182552,0.0332085267,0.3281262517,0.1403977573,0.2353489101,0.0820999667,0.276720047,-0.1882340461,0.1560410559,0.1065798849,0.1200828403,-0.177686736,-0.2091618776,0.215586409,-0.1703969091,0.1577081084,0.187066108,-0.0103386808,0.1187836453,-0.1394832283,-0.221241191,0.1681769788,-0.2791554928,0.1245211363,0.4301840961,0.0467428602,-0.1407841593,0.2656863928,0.3378235698,0.4944858551,0.2655309141,-0.3144285083,0.1665557027,0.0808337256,-0.1713611186,0.4370556772,0.0694448277,0.1463284492,0.1456014067,-0.0595269054,0.1259955764,0.2828309536,0.0892844126,-0.126315847,0.097394146,-0.0245555509,0.1320469528,-0.0004669731,-0.2421218306,-0.2539230585,-0.0854687244,0.1972647309,-0.2781001329,0.1593440026,-0.1881109476,0.1652384698,-0.0784458593,0.1206083223,0.2307350338,0.0284567866,0.0906465948,0.3625095487,-0.0431696437,-0.4325211644,-0.0575030968,0.3947230577,0.4380664825,-0.0266026873,0.0425503291,-0.0863508284,-0.2363494784,0.219547987,-0.0105046388,-0.2339505404,-0.3468417227,0.1312466711,-0.2367573231,-0.0248994026,0.2180727273,0.1072021201,-0.0457139835,-0.10694056,0.058142256,0.243554458,0.0644963682,0.0964427963,0.1433320642,-0.5415202379,0.229074806,0.0276765376,0.0710890517,-0.2556727827,0.1122746691,0.2078499645,0.2689954042,-0.0512942225,0.0678157955,0.0928567126,0.1738771498,0.0270611849,0.0502625778,-0.0984318182,0.2000484467,-0.3773201704,-0.0089052226,0.0027763098,-0.2553081512,0.0847193301,-0.0904905349,0.1383438259,-0.0266533755,0.2075954378,-0.385506779,-0.101936236,0.0522675477,-0.0027397822,-0.5162889361,-0.0027051433,0.0192057732,0.1361534446,-0.2607427835,0.0488843769,0.0608717352,0.0216409434,-0.4746630788,0.2268773913,0.042819012,0.3522044718,0.5298022032,0.0421164222,0.3925606608,-0.1250881702,0.3032928705,-0.0094712116,0.1285093725,-0.1753814965,0.2215262353,-0.2331988066,-0.0719531551,-0.2032157034,-0.3315406144,-0.0679726079,-0.0388560742,-0.0194204655,-0.3444187343,-0.0503284261,0.1241626516,-0.3433810771,-0.162735939,-0.3086554408,0.170132339,0.0229629241,-0.0601679273,0.1718798727,-0.2669443786,0.5446509719,-0.3649508357,-0.1904860437,0.401833266,0.0341416299,-0.3851489425,0.1420315206,-0.3323297501,-0.0212924164,0.4848424792,-0.1330303848,-0.187727049,-0.0205653869,0.204070136,0.486679554,-0.3850891292,0.4077715874,-0.0966907963,-0.0449476168,0.0255766045,-0.1758658439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1592","title":"Using datasets.Metric with Trainer()","comments":"We are indeed working on the integration with `Trainer` :)","body":"## Using datasets.Metric with Trainer()\r\n\r\nHi team, I was quite surprised in the [Metric documentation](https:\/\/huggingface.co\/docs\/datasets\/using_metrics.html) I don't see how it can be used with `Trainer()`. That would be the most intuitive use case instead of having to iterate the batches and add predictions and references to the metric, then compute the metric manually. Ideally, any pre-built metrics can be added to `compute_metrics` argument of `Trainer()` and they will be calculated at an interval specified by `TrainingArguments.evaluation_strategy`. \r\n\r\nIs this option available but just not mentioned in the documentation or it's not possible at the moment? I notice in the [Transformer | Training and fine-tuning](https:\/\/huggingface.co\/transformers\/training.html) tutorial, you are using custom scripts to calculate the accuracy, P\/R\/F, which are already in the pre-built metrics.","comment_length":10,"text":"Using datasets.Metric with Trainer()\n## Using datasets.Metric with Trainer()\r\n\r\nHi team, I was quite surprised in the [Metric documentation](https:\/\/huggingface.co\/docs\/datasets\/using_metrics.html) I don't see how it can be used with `Trainer()`. That would be the most intuitive use case instead of having to iterate the batches and add predictions and references to the metric, then compute the metric manually. Ideally, any pre-built metrics can be added to `compute_metrics` argument of `Trainer()` and they will be calculated at an interval specified by `TrainingArguments.evaluation_strategy`. \r\n\r\nIs this option available but just not mentioned in the documentation or it's not possible at the moment? I notice in the [Transformer | Training and fine-tuning](https:\/\/huggingface.co\/transformers\/training.html) tutorial, you are using custom scripts to calculate the accuracy, P\/R\/F, which are already in the pre-built metrics.\nWe are indeed working on the integration with `Trainer` :)","embeddings":[-0.2474181652,-0.3850362897,0.0257479325,0.2080757916,0.3879331648,-0.1194488704,0.30666399,-0.2372968793,-0.0428988934,0.342396915,-0.3008914292,0.0187155362,-0.1781838685,0.2172097564,0.0734231994,-0.3760861754,-0.1171281785,0.2103784829,-0.4320205748,0.2859491408,-0.0255832598,0.1156030074,-0.1282717139,-0.0701691061,-0.4342900217,0.0711919367,0.0177618675,-0.3503544033,-0.2118661702,-0.3951251209,0.5193301439,0.2450224459,0.1800102741,0.6852629781,-0.0001189353,0.1074586734,0.1646833569,-0.132955566,-0.0809742808,-0.3278051615,0.0896859244,-0.0989331007,0.0744482502,-0.2389069945,-0.2973430455,-0.1060979143,0.2252854854,0.1257541478,0.4554151893,0.4069104493,0.0824791417,0.6397571564,-0.0572001673,-0.2873821855,-0.3974006176,0.0900289789,-0.1208119541,0.215618968,-0.1471707076,0.1280863881,-0.3028177023,0.0718122199,0.4839321971,0.0053508948,0.7120957971,0.1612315625,0.0015737112,-0.155317381,-0.3926597834,0.1309252232,0.193563953,-0.1052640155,-0.4079487324,-0.2382798493,-0.018263137,-0.3920497298,0.000775871,-0.4115589857,-0.014657123,-0.0075306762,-0.6212767363,0.0678091273,-0.3307400942,-0.0808574259,-0.1358795762,0.3300500512,-0.2015007287,0.0654123053,0.4275785387,0.1592972279,0.0709772557,-0.0942509249,0.1236419082,0.3919994235,-0.1527172625,-0.426292181,-0.0082795722,0.2485240847,0.0186698455,0.0426451601,-0.0092896363,-0.0642451122,-0.0321502797,0.1371351331,0.1439744383,0.6142418981,0.1426247507,0.2963454127,0.1829791516,0.4542326629,0.1667043865,-0.1172694191,0.0802247152,-0.3124661148,0.1349399686,0.0484147221,-0.3514042199,-0.0216245316,-0.0412236415,-0.2391423732,0.2914250791,-0.3801036477,0.2948004305,0.1478450149,-0.2263271958,0.0713993087,-0.0289954059,0.0434850976,-0.0493525416,-0.0510356985,-0.0811661705,0.1979245394,-0.37776649,0.2127141953,0.0988299921,-0.2628761828,0.2297246158,-0.1811588705,0.5006491542,-0.0296049602,-0.1601079404,-0.1320691556,0.1584343016,0.1678454727,-0.045761168,-0.2741928995,-0.051947128,-0.1012646928,-0.2154012471,-0.0806124136,-0.1527254283,-0.3111388981,0.1104299799,0.0686637908,-0.4799492061,-0.0747028664,-0.4332622588,0.463463366,0.0383136012,-0.0925498828,-0.1033725664,-0.1296517849,-0.4964143932,0.0283729956,0.3167053759,-0.0086944951,0.0586105175,-0.5211880207,0.2385725379,-0.2391110957,-0.354388088,0.1642661691,-0.343750149,0.6294376254,-0.127108857,-0.2060199529,0.4423575699,-0.5158532858,0.0200254153,-0.1627130806,0.0853708759,0.1416789144,-0.0504966304,-0.0326637514,0.3765693605,-0.1079683676,-0.0343376249,0.4080233276,-0.0194844548,-0.2552847862,0.0388965569,-0.1046069115,0.0093344273,0.1177665666,0.5218383074,0.1408085972,-0.1105374917,0.3205921948,-0.0832674652,-0.2468085289,0.0808039829,0.1201564968,-0.14042826,-0.0820569098,0.1239421368,-0.1839583516,-0.2167746276,0.1625044495,-0.2273501307,0.1667478085,0.3292058408,-0.1657514721,-0.1713364124,0.0260874256,-0.1458047628,-0.1515581757,0.0158232786,-0.0682863295,0.2552261055,-0.1351118088,-0.3420325816,0.2075924277,-0.1991106719,0.2746566236,-0.3068739176,0.1280629188,0.0000729242,0.169503361,-0.1228381023,0.2922434807,0.1110473946,-0.2874439955,0.075268887,0.3535060883,-0.1520002782,0.0635323301,0.0323790163,0.3906010389,0.4177722037,-0.0997805595,-0.2440576106,-0.040229477,0.0803201646,-0.174799785,-0.2873756289,0.5528020859,0.1747088134,0.1873034984,0.1093343496,-0.1203084439,-0.1469258666,0.0711559653,-0.4561134875,-0.0101409405,0.0976289287,-0.4425823987,0.0434975103,-0.0644262508,0.0107878214,-0.2935643792,0.2968790233,-0.191521436,-0.1206448898,0.3105800748,0.0461693928,-0.0497037172,-0.3816053867,-0.1967928559,0.4754009247,0.0334714875,0.2440705746,0.1943615526,-0.2052806616,-0.088724874,0.1129222512,0.3002788126,-0.3368463814,-0.0100208754,0.0280211288,0.0092442352,-0.1769950986,-0.166245833,-0.0559981465,0.0438800603,-0.2950829268,0.2084054202,0.0241600722,0.1000913829,-0.3222527504,-0.1761387289,-0.2168163061,-0.2026246935,0.0684797093,0.1627049148,0.1790479124,0.2171555907,0.2991215587,0.4885181785,-0.0703182891,0.115984045,0.0414453037,0.0175753105,0.0377183259,0.0655799061,-0.1054354459,-0.452842772,0.2034274787,-0.2322074473,0.1123660132,-0.143109113,-0.4587962627,0.1486978233,-0.0435031354,0.5108934641,0.2665725052,-0.1593791246,-0.1225007996,0.1150519252,0.256739378,-0.2228027731,0.0162314977,-0.2743196487,-0.2379311919,-0.0066830646,-0.1635070443,0.0229980536,-0.1073651835,-0.1071498469,0.0694869533,0.3258016407,-0.050033249,0.1532923132,-0.099301286,0.256316185,-0.1824662238,-0.0664686337,-0.121891059,-0.472296387,0.6279153824,-0.3637926877,-0.3172843456,-0.0497692078,-0.2774666548,0.1644995064,-0.1088888645,-0.3189424276,-0.5487595797,0.046850726,0.0111929923,-0.0768262967,0.4687302411,0.2049355507,-0.1982397884,0.0909614637,-0.1020432115,-0.2695018947,0.1338992864,-0.04957968,0.1479867399,-0.0341416746,0.2101357728,0.1038700119,0.7977967858,0.117837511,-0.3861602247,0.1088779196,-0.0882880464,0.5125672221,-0.1759129316,-0.0761897638,0.1731626242,0.0162539873,0.0578149557,0.3063440621,0.0831324607,0.2554545105,-0.2631804347,-0.0299163964,0.0245583951,-0.1272460073,0.1280688345,-0.0425177924,0.2454518974,0.0342255868,0.2574927509,-0.1258933097,-0.0606662408,0.0948393196,0.1774844229,0.2354320735,0.1962610483,-0.3426889479,-0.1223653257,-0.5203502774,0.1703490615,-0.0661531538,-0.045874197,0.0411795639,0.1293184012,-0.0124778384,0.0744824708,0.4363785684,-0.0314238183,-0.2003368139,-0.1205889806,-0.4391428232,-0.0926182866,0.0696210712,-0.0986395776,-0.1916569322,-0.2187715471,0.7211732864,-0.3712581098,-0.1550571024,0.3753459156,0.038998846,-0.339677006,-0.1098971143,-0.0499222651,-0.2238604426,0.1400763243,0.2173035592,0.1871692091,0.2120724022,0.100343734,0.3417859674,-0.0453215018,0.262947768,0.2529036403,0.0849978551,0.1397034675,0.0069238627,0.0230456032,0.5692472458,-0.1186617166,-0.0967542753,0.0615866855,-0.1215934008,-0.4094952941,0.1356686205,0.1670743674,0.4793884456,0.2322282344,0.0693337321,-0.0126632899,0.0701041445,0.2605131865,-0.2346936762,0.2445284277,0.3314377069,0.2349517941,-0.4556227028,-0.2539581358,0.4741811752,0.276049614,0.1293674111,0.2798882127,0.1850745827,-0.3786095381,0.0825382993,0.1590105891,0.7304536104,0.0699055344,0.274399966,0.3215593398,-0.0960041806,0.3534730375,-0.3638696074,0.0533687137,-0.038948331,-0.1597730815,-0.1810319573,-0.0441066734,0.2800954878,-0.0369021706,-0.1634791046,0.5254698396,-0.001144495,0.3161703944,0.0714374855,-0.0053016664,0.1980736405,-0.3482282758,-0.0614770614,0.0878688768,0.0305717979,0.1969439983,-0.1972517967,0.0183981955,0.2361974567,-0.1693622023,0.0898393318,-0.4161946177,-0.1481168419,-0.1641549021,0.440610528,-0.0447270386,0.1428455412,0.3011421263,0.500325799,-0.0531021506,-0.1641024798,0.3018398881,-0.5876498818,0.2311695069,-0.0568313487,0.0314591303,0.2614786923,-0.0475093089,0.2523694038,0.1411481351,-0.0403486826,-0.1902470142,-0.0251571946,-0.0374179184,-0.2744523883,-0.2127095312,0.0778785497,-0.0557180569,0.0230705403,-0.1875377595,0.0632608086,0.0964934677,-0.4356296659,0.2317132801,0.1103945598,-0.2406062633,-0.1028976664,0.4297071993,-0.0251223426,-0.3107573986,0.0404439345,-0.1815176308,-0.1786714792,0.1237982363,-0.0251143426,0.3360742927,-0.3282156885,0.1146954447,0.1105986759,0.0064051743,-0.1873124987,0.5700149536,0.2056779861,-0.1686207056,-0.1493895799,-0.2142711133,-0.4429731965,-0.0053691831,0.2647723854,0.1888806075,-0.0764142498,0.4597077668,0.1361926943,0.2230239958,-0.257591337,-0.0367928781,-0.0757719278,0.20413661,0.2133338302,-0.1671492308,0.2670021355,0.1046893448,-0.1050826684,0.1421863139,-0.4355385602,-0.1020752713,-0.3743812144,0.1707039326,0.0471868552,-0.0596053265,-0.0470826179,0.2318725139,-0.0904182494,-0.2807016373,0.0795145333,0.3877389729,0.0321131945,0.1465234309,-0.0715363175,0.1243628785,0.2473858744,-0.1152001545,0.665112555,0.1809127033,0.1469387412,0.139908731,-0.4225590825,-0.046342019,-0.0031370439,0.217508629,0.4922202229,-0.146825999,0.213011086,-0.0444167964,0.0257466547,0.0281290803,0.5642774105,0.0810528398,-0.1343717575,0.0400781631,0.0051854369,-0.0003875529,-0.137219727,0.0365948007,0.153337732,0.1171102673,0.0845763907,0.2582672238,0.0975541472,0.2102125138,0.2089064717,-0.0333383493,0.1285621524,-0.2454733104,0.286310792,0.1349142492,-0.0417224057,0.138935104,0.717155695,0.1280668676,0.3870888352,0.1013723537,-0.1950636059,0.5523722768,-0.4398640692,0.0834312513,0.0417185202,-0.0726478547,-0.301633954,0.0268867351,0.3445600569,-0.0531549305,-0.1667992473,0.4838912487,-0.2710378766,-0.273765415,-0.2815027833,0.1785681993,-0.0133887753,-0.232047528,0.0334949084,-0.2236304283,-0.120307304,-0.0470623188,0.0243664514,-0.2001582086,0.1597925127,-0.1961752027,-0.0111924298,0.1304205209,-0.2128395736,0.1206471398,0.221767962,-0.1276659369,0.2330983728,0.0106193833,0.1573678404,0.0201493222,0.5378271341,0.0974688232,-0.158879891,0.0727203265,0.1027351171,-0.3154653311,-0.0125628151,-0.0369883329,0.4414526522,0.0136744631,-0.5041055083,0.3826314211,0.0459667481,-0.230868265,0.5877290964,-0.3877696991,0.1373041123,-0.6298614144,0.4512999654,0.0388439,-0.0191086009,-0.0143254679,-0.0883770734,-0.1741521657,0.1414028257,0.088854298,-0.0653965473,0.0376834199,-0.2667644918,0.0236312021,0.1610345989,0.374659121,0.2754173875,0.0171311963,-0.3490069509,-0.1206556037,-0.5873494148,-0.1536122859,0.2258851528,0.3459435701,-0.0852070004,-0.0613135435,0.018442858,0.2417157888,0.3303927779,-0.2441107184,0.1439419836,-0.0913535357,-0.1031144485,-0.0229552574,-0.3671602607,-0.0017460164,0.1433882862,-0.0283084679,0.0573914722,-0.2008614689,-0.0444523506,0.1146985218,0.2084314227,0.124752894,0.2515338957,0.2952916026,0.077409938,0.3986713588,0.0873477682,0.3844888806,0.0444195345,0.056172736,-0.1759569645,-0.0580493994,-0.1091292128,0.2443763912,-0.4411914349,-0.1447463334,-0.2042586356,-0.0989384279,0.2672798932,-0.2314569503,-0.2200585157,-0.0222923215,-0.4364183247,-0.1029474661,0.3537239134,0.6288295388,-0.0263052825,0.0219238307,-0.1602925807,0.0556085333,0.5448101759,-0.5092915297,-0.2690834403,-0.2043043375,-0.1899021417,0.4362276793,-0.4492767155,-0.6769992709,0.113772653,0.2011606246,-0.0570605509,-0.1274568588,0.3429963887,-0.2157069445,0.0853474662,-0.1112411246,0.5011616349,0.0438352823,-0.038071394,-0.0908693075,-0.1998866349]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1591","title":"IWSLT-17 Link Broken","comments":"Sorry, this is a duplicate of #1287. Not sure why it didn't come up when I searched `iwslt` in the issues list.","body":"```\r\nFileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n```","comment_length":22,"text":"IWSLT-17 Link Broken\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n```\nSorry, this is a duplicate of #1287. Not sure why it didn't come up when I searched `iwslt` in the issues list.","embeddings":[0.0006844614,-0.4732919931,-0.0473029353,-0.0252460092,-0.072372213,-0.1377999783,0.4260438979,0.2196717709,0.0861584172,-0.0510080159,0.1014911458,0.01117827,0.234124288,0.1399892271,0.068584688,0.3315440118,0.1696381569,-0.0591772646,-0.2145912647,0.0649079904,-0.2930266559,0.4922843874,-0.1860420853,-0.0230990555,-0.4022633433,0.1943532526,-0.0546754338,0.3688046932,0.0660205707,-0.31805709,0.0928245783,-0.2664248943,-0.1155584902,0.0801602453,-0.0001027269,0.0725963414,0.3312967718,0.0182316564,-0.039724309,0.0038765541,-0.4141123891,-0.360758245,-0.3401444852,-0.2679913342,0.083196342,-0.1541734785,0.0127533665,0.4328042269,0.4594027698,0.2857708037,0.2857938707,0.0177020784,0.3464299738,-0.2142918855,-0.033776205,-0.1964155138,-0.1482028365,0.0175156519,-0.0572566092,-0.0490614623,0.0739294812,0.4202610552,-0.0014970835,-0.0906919241,0.0832032412,-0.1663054526,0.6395931244,-0.2843967378,0.3459231853,0.0326452889,0.3625123501,0.0956142023,-0.024690453,0.239191696,-0.0066234395,0.1453074068,0.4166468084,-0.0576201826,-0.108288981,0.1479462981,0.0997009128,-0.1229253262,0.0443260074,0.1056374758,0.0061146948,0.709802866,-0.0680117458,-0.0042495984,0.2167512625,-0.2968645692,-0.0850264803,0.3304434121,-0.2750071287,0.0375195071,0.0305321943,-0.1025233194,0.1772165745,0.0931896865,-0.156649217,-0.1070362926,-0.1556742787,0.0075665819,-0.0228822865,-0.0098473709,-0.1456732303,0.0217749588,0.0329333246,-0.1880739331,0.3502298892,0.2299733311,-0.0562805794,0.0538175255,0.0640300959,-0.2839123607,-0.2731635571,0.0800906494,0.0603784025,0.014891156,-0.3269964159,-0.0366399214,-0.0750796422,0.1032091603,0.0396595038,0.1052277684,0.1486850977,0.1083868593,0.0737453252,-0.1670612395,-0.145693019,-0.1165530309,-0.0839784294,0.0587396324,-0.0097535206,0.0754460767,0.1167550012,0.2683935165,0.1553176641,-0.0396597013,-0.3994235992,-0.3470971584,-0.3023123443,0.2189502269,0.0932866335,0.2179436833,-0.1534749269,0.1959988624,0.1381669492,-0.2337029725,0.091553919,0.1750638038,-0.2198525071,-0.3070573211,-0.3303204775,0.3187909424,-0.0890317112,-0.062058486,0.592000246,0.1136392131,-0.2392677814,-0.2184087783,-0.0538383052,0.2799759209,-0.5356158614,-0.0395276621,0.1645094007,0.2866088748,-0.4959416389,-0.2172528803,-0.157035768,-0.045078706,0.0142990323,0.2759904265,-0.098097913,0.2701850832,-0.3180533946,0.3719313741,0.211175859,-0.1238324866,-0.3806768358,0.2075423598,-0.2282713056,-0.10474433,-0.1508671492,0.1886084825,-0.0305274352,-0.2770000696,-0.2378554344,0.023651721,0.1344772577,0.1217098087,-0.4252086282,-0.4194178581,0.1552657634,-0.1698845029,0.0007901921,-0.2586677969,0.1822284013,0.424702704,0.1566957235,-0.0342610441,-0.0651527345,0.179021433,0.476087451,0.1283302903,0.3319918215,0.0866797343,-0.0711061656,0.1365967393,0.2252643555,0.1361791939,0.1692947149,0.062154945,-0.2857600451,-0.0129033951,-0.2341265231,-0.2812821865,0.299508661,0.1546766013,-0.1314084977,0.0246637166,-0.1131095663,-0.0352985673,-0.3365405202,-0.0846760347,0.0395616442,0.3213646114,-0.3308702111,-0.1463459283,-0.0502646044,0.2350209206,0.3047770262,-0.0263919551,-0.1382350773,0.1381839514,-0.2247166932,0.104634434,0.0036255436,0.3050383031,0.1095009297,-0.6218362451,-0.1884242296,0.5265060067,0.010966639,-0.0426728055,0.1724851727,-0.3564175367,0.1630452871,0.1086338684,-0.0718908906,-0.0120034106,-0.0416823737,-0.3572949767,0.0865869522,-0.1812565029,0.1133780628,-0.1891230643,-0.3709328473,-0.10582757,0.2016046196,0.2312986255,0.3363108337,0.0031455364,0.2152808756,0.1793357432,0.0183640625,-0.015806498,-0.1573385894,0.1023871675,0.220795989,0.0757260323,-0.040155828,-0.0022344652,-0.0290193856,-0.1879840642,0.3390098214,0.0486641303,-0.0258562863,0.3497512341,0.1588130295,0.1261734515,-0.2462342978,0.1571489275,0.0975352302,0.0104038026,-0.2005771846,-0.0918560997,-0.3729525805,-0.6177487969,-0.2585009933,0.1413275152,-0.1600207984,-0.2828134596,0.2161189318,0.0731555,-0.1082731038,0.3150978088,0.2780673206,0.4048838615,0.0364424214,0.5076473951,-0.2202726007,-0.0938652083,-0.3835863173,0.1270001978,0.4048323333,-0.052804023,0.1483353674,-0.1700387895,-0.0775309503,-0.6530392766,-0.5673181415,0.048985064,0.0108172055,0.2896001935,0.0804569051,0.3780201972,-0.1454787254,-0.3218047619,0.3558204174,-0.0894986987,0.1128537282,-0.279014051,-0.081680648,0.0054657566,-0.2410790175,-0.0255379099,-0.1625329852,-0.4413508773,0.1885261834,-0.3920689821,0.0677083731,0.1095629111,0.0386846252,-0.0286509693,0.0545981824,-0.0528109483,-0.0466461331,0.258187741,0.2488073856,-0.1049061567,-0.3156597018,0.2583633363,0.2335182428,0.1036950126,-0.5245766044,-0.5797256231,-0.1467759013,0.0488854237,0.0952193588,-0.1783047616,0.0995092615,0.0492230803,0.0556763858,-0.24199754,-0.1582617462,-0.0889321789,0.065396443,-0.0556306615,0.4127615094,0.1255225241,0.113470301,0.1550671607,0.1015897989,0.1160734743,0.0212453716,0.1573708504,-0.0480968505,-0.0487801097,0.0341104269,-0.3927622736,0.2581166625,-0.0589644164,0.0655069351,0.409242481,0.0354209282,-0.0891521424,-0.0291421022,-0.0985524505,-0.1568245143,-0.3099571764,-0.077726841,0.5121833682,0.0831044465,0.1202711239,-0.0283231437,-0.0700543821,-0.2511938214,0.3918587267,0.3374727368,-0.2671720088,-0.0110454224,0.4114719331,-0.3707741797,-0.1234085485,0.0122503666,0.1930211037,0.1237206236,-0.2686843872,-0.2343182862,0.0099318586,-0.2200825065,0.5752206445,-0.0396777205,0.2547216713,0.1274146736,0.341920346,0.2140260339,-0.0659979656,0.0977605581,0.1244032905,0.7113208771,0.3554584384,-0.0087933298,0.0870206803,0.2661788464,0.1590253711,-0.113955006,-0.3886222839,-0.0166429002,-0.3141437471,-0.3908333182,-0.3156846166,0.0792690814,-0.0788986906,-0.0519410037,0.0826048255,0.2171731889,-0.3817941844,-0.0929287821,0.1053541228,-0.0473163836,0.2801974714,0.0733596087,0.0461424589,0.1459236145,-0.5549019575,-0.0152227571,0.3242753148,-0.0138773555,0.0529942848,-0.0627305955,-0.2190223634,0.3730353117,0.0389321484,0.0311073437,-0.0219494998,0.1446044445,-0.0320432149,-0.1585561782,0.3358007371,-0.0466830209,-0.0937906802,0.244624272,0.4014318883,-0.2614635527,0.0688641369,0.3759871423,0.1437222213,-0.10759224,-0.0036890656,-0.1040746719,0.8769706488,0.5410659909,-0.1670460701,0.1796724498,0.180556193,0.4220653176,-0.1025893539,0.1661435664,-0.059216287,0.1265650988,0.0549818985,0.2576271296,-0.1995970607,0.1188718826,-0.1496516466,0.0614204891,0.3874852359,-0.0402340032,0.1643796563,0.2193497866,0.0151952654,-0.0196010247,-0.1709966511,0.2933668792,0.0980625227,0.3090145886,-0.1907305717,-0.1337441355,0.2583162785,-0.097166121,-0.4179595709,0.2403559238,-0.3239867389,0.5455456972,0.1046036482,0.2023061067,0.1196498722,-0.0137771228,0.2258885503,0.100983642,-0.3179384768,0.3280110955,0.2294306457,-0.1505611092,-0.3034003377,0.1024809852,0.0019912012,-0.3366168737,-0.4029811919,-0.1634420604,-0.0276879091,0.1027876213,-0.0655167699,-0.0362562016,0.2170410901,0.1709929109,0.0847850591,0.2113866955,-0.1790823489,-0.2219301909,0.2504846454,-0.272734046,-0.1662693918,-0.0086997943,-0.2595964968,-0.0586968027,0.0802591667,0.3711563349,-0.4774598777,0.0836863965,0.5087407827,0.3086289167,-0.2049729377,-0.2534293234,0.053838484,-0.508384645,-0.5010902286,-0.0978129953,0.0397705734,-0.0769959688,0.0426667705,0.7937260866,0.0808824599,0.1489170045,-0.198944822,-0.2461434752,0.086452499,0.2057931274,0.2014051676,0.1129096597,-0.4223166406,0.2567222714,0.1314021349,-0.3457979262,-0.4207066596,0.1900727451,-0.1318689138,-0.1235417724,0.0241343901,-0.1946233958,0.2517677248,0.226744771,0.1914000213,0.257119596,0.0673249662,-0.3255093396,-0.2029978335,0.0958786234,0.1198637336,-0.3343102634,-0.0767678171,-0.150572747,0.1334246099,0.3163568377,-0.0953289866,-0.1780115515,0.0482162014,-0.1081425324,-0.0137935607,0.0125727374,0.1259266138,0.1934914291,0.3097809553,0.6810576916,-0.0056447037,0.0859884396,0.047094848,0.0795342028,0.1490433067,0.0681175813,0.0055858772,-0.0693299472,0.0836990029,-0.2331851721,-0.0160219315,0.1874295175,-0.085494794,0.2330173999,-0.3215359747,-0.1663146466,0.2321325839,0.2929686904,-0.1608548462,0.0125245852,0.0311915334,-0.126288414,0.0330006368,0.0258744042,0.0692848563,-0.2237233818,0.3405435979,0.1852992922,0.2580711246,-0.4347940385,-0.185751304,-0.0469825789,-0.059480343,0.0724557489,-0.3675893843,-0.270092845,0.2561877072,0.1028906703,-0.1038204953,-0.0006864072,-0.0012103049,0.3537601233,0.1350405216,-0.0775277913,0.1472008228,-0.2014288157,0.0481883809,-0.1448144764,-0.0977461115,-0.0273426957,-0.1606376022,-0.151738435,-0.2279828787,0.4944896102,-0.1046966612,0.0031334748,-0.252553165,0.0089335823,-0.3407759666,0.0217199773,-0.0995987058,0.1278247088,-0.1337421685,0.1325188279,-0.1817132682,0.0401826501,-0.372585535,-0.0080487942,0.3462360501,-0.1691352874,0.4039767087,0.2556940317,-0.0111859553,0.0878763199,0.4890322387,0.3612070978,0.2412551641,0.0626913086,0.084457241,-0.0879357755,-0.2101274133,-0.0667084157,0.2262799889,0.0728764534,0.1894913912,0.1027543694,0.1991948634,-0.2840892375,-0.1034412012,-0.0921614841,0.0008311747,0.1463597566,-0.1153679714,-0.4012374878,0.1403783262,-0.2646595836,0.0403613038,-0.4237565398,0.1678449959,0.2200509906,0.0891449377,0.1705370098,0.0166035332,0.1612883061,-0.2782751024,0.442760855,0.2120123506,-0.1959426552,-0.1066589728,-0.4702953696,-0.3975894153,0.1634074748,-0.1076740101,0.1429194361,-0.1334154904,0.0323043652,-0.2335060239,-0.091564253,0.0825941935,-0.136957705,-0.1162934154,0.0918511748,-0.1264710724,-0.1245790347,0.1015842557,0.0530874096,-0.0182603858,-0.1163272858,0.0951023251,0.0089897616,0.1616520584,0.1846157759,-0.4957481325,-0.3013752103,-0.0680713356,-0.067057699,0.1549978554,0.1991527528,-0.2349580973,-0.38188833,-0.293097496,-0.3963512778,-0.0302287117,0.2401717901,0.0433713421,0.192104876,0.0443056412,0.4203975201,-0.2631629109,-0.1371117383,-0.1592000127,0.2826224267,-0.2275381088,0.136279285,-0.1149270013,0.121718049,0.2010377795,0.2671245933,-0.1260782927,-0.137155965,-0.2957424521,-0.4097743034,0.4783676863,-0.4301977158,-0.1116491333,-0.0150298439,0.2523158193,-0.0244043786,0.2548389137,-0.3287775517,-0.116100952,0.2193017602,0.2975826561,-0.0058920374,0.0242541991,0.0652033985,-0.1835387498,0.0413877927,-0.2219201773,0.412830025,-0.0829784423,0.3143583238,0.1022474617]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1591","title":"IWSLT-17 Link Broken","comments":"Closing this since its a duplicate","body":"```\r\nFileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n```","comment_length":6,"text":"IWSLT-17 Link Broken\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n```\nClosing this since its a duplicate","embeddings":[0.0831616968,-0.5686072707,-0.0968643874,0.0713846236,-0.1165643483,-0.1431536227,0.2291705012,0.2218179405,0.0600371584,-0.020788068,-0.0552899428,0.013893595,0.318267405,0.1358832121,-0.092038393,0.3631219566,0.0809323415,0.1052210778,-0.274597168,-0.0537774898,-0.2358401269,0.4682433903,-0.1429242194,0.1117994487,-0.3499359488,0.1402864307,-0.0763485506,0.4525981247,-0.0706484765,-0.1207786575,0.0061448836,-0.1302866638,-0.1855068058,0.1564768553,-0.0001021994,0.0922966897,0.2846952081,0.0005464505,-0.0397692285,0.0426784493,-0.2905516922,-0.2663292885,-0.3602590263,-0.3216896057,0.0409675874,-0.2774332166,0.0417888202,0.338765651,0.494383961,0.2500095069,0.2696195245,-0.0216090735,0.3248042166,-0.1653766483,-0.0248939842,-0.2747733295,-0.0524993017,0.0491496101,-0.0208766516,0.0229235385,0.1124512702,0.4092994034,-0.1133442298,-0.0469918959,0.0596248433,-0.222063601,0.6550581455,-0.2645055354,0.3368583322,-0.0047617112,0.3498798311,-0.046324756,0.0413104855,0.3761866987,-0.0395712741,0.2063697726,0.2212893665,-0.0891298354,-0.0808519199,0.2244429588,-0.0284345318,-0.0811310634,-0.0343026333,0.1090448946,0.086173892,0.6233494282,-0.0593286529,0.0357746482,0.2963439822,-0.1840645969,-0.0373902246,0.1650033593,-0.266479522,0.029688051,0.2354712039,-0.1793920249,0.1223831698,-0.0037010431,-0.1731685251,0.0214696731,-0.1568003744,0.0926480293,-0.0452649184,0.0264067836,-0.2177922279,0.0488871336,0.1257401258,-0.1463859081,0.3011942208,0.1998756379,-0.1514950842,0.0480695665,0.0614691339,-0.3649395406,-0.2310220152,0.1377445161,0.1012885123,0.0308493674,-0.2135957628,-0.174905777,-0.0059589422,0.0939757675,-0.0065597957,0.228217423,0.1351718307,0.0425274335,-0.0104241706,-0.2052965611,-0.1959784478,-0.1430972815,-0.1678319275,0.1924172491,0.0863867626,-0.0439598486,-0.0065093185,0.2545860112,0.1732105166,-0.0070542884,-0.3897497952,-0.3435227573,-0.165314123,0.1516070515,0.2259953767,0.2059527636,-0.070077762,0.1747810692,0.1966535151,-0.2174871266,0.0576472282,0.1684357524,-0.2287733108,-0.3422927558,-0.4209682345,0.364087075,-0.0050806212,-0.032373175,0.5156806111,-0.0220998116,-0.2657917738,-0.1465589553,0.0206465926,0.3768517375,-0.4398686886,-0.0640688762,0.1177519113,0.3027053773,-0.3267705142,-0.1339076012,-0.1139408648,-0.0371476375,0.076168485,0.4699297547,-0.1663464159,0.3143123686,-0.3344708383,0.3740330338,0.3728163242,-0.1688221246,-0.3415885568,0.3076369464,-0.1568152457,-0.1157844141,-0.0409531035,0.2506449223,0.0031218135,-0.2764055133,-0.2066980153,0.001798369,0.219884485,0.1364579946,-0.4613340497,-0.4543723762,0.1538166404,-0.2023334801,-0.0126419999,-0.360901624,0.1518281847,0.2744230926,0.2581032813,-0.0859739333,-0.1493846625,0.1878559142,0.4139761627,0.1272265017,0.2639655173,0.1411165446,0.0308665074,0.0487971865,0.1286526173,0.07547196,0.1312799156,-0.0736789033,-0.3246115148,-0.0228433609,-0.3789179921,-0.2867704332,0.3185165823,0.1420701742,-0.0635734051,0.0441864356,-0.1164839193,-0.0999765173,-0.3733233809,-0.1365934908,0.0242694095,0.2532382011,-0.4090290964,0.0740580931,-0.0902486145,0.2467478812,0.3461641967,-0.0937997401,-0.0975790918,0.1700029671,-0.2260325402,0.2691893876,0.0576086752,0.141540423,0.0865352228,-0.5924528241,-0.0999835134,0.4683946371,0.0210832208,-0.1006890535,0.0902093574,-0.348077327,0.2123730928,-0.0635624751,-0.0460788012,-0.2318045795,-0.0908840969,-0.3009371161,0.193783164,-0.2026719302,0.0262137651,-0.12633802,-0.4059934616,0.1324629933,0.3044635057,0.1458447874,0.3614330292,-0.0750303715,0.2356820107,0.1981844008,0.0174185447,-0.037479084,-0.2257994264,0.0598973744,0.2024338096,0.1166694388,0.0277387574,0.009894195,0.0539934896,-0.2053185701,0.3360818028,0.0876697898,-0.0424149334,0.3322218955,0.0754371136,0.0575721711,-0.3303678632,0.1785151511,0.1278708279,-0.0139354216,-0.1590652764,-0.1029870883,-0.3878301382,-0.5936242938,-0.2456642389,0.2025418431,-0.0545147881,-0.2372082919,0.3086202443,0.1690987051,-0.0916460529,0.3008931875,0.3340730965,0.2995609641,-0.0212452058,0.3505654633,-0.2780423462,-0.0214976594,-0.267148912,0.1793767661,0.4665489495,-0.029245181,0.1136392951,-0.1783156395,-0.2011390924,-0.752204299,-0.5163903832,-0.0072126566,0.0149687771,0.1707967967,0.1117455438,0.4072034359,-0.1123942956,-0.3575185537,0.3140276372,-0.0059534903,-0.0043792501,-0.3099623322,-0.0813262016,-0.0183645282,-0.2500096858,-0.1193755493,-0.1743379831,-0.3706058562,0.2102298588,-0.3708045185,0.0713409707,0.1772649139,0.0876884386,-0.1151644066,0.1466215104,-0.017135635,-0.1028354764,0.230227828,0.2748604715,-0.0834925994,-0.3293525875,0.1878342628,0.2470542341,0.1983070523,-0.6520739198,-0.4865690172,-0.2667278051,0.040965572,0.0388285108,-0.11196886,0.0513638705,0.0389262438,-0.0513654314,-0.2543958724,-0.0634825975,-0.0336693041,0.0665671974,-0.1507166773,0.3128427565,0.1067347676,0.1976371855,0.1102048531,0.073576346,0.1716753989,0.0349147804,0.1240604222,-0.0991062,-0.0164577682,0.011540262,-0.3740276694,0.2392348647,-0.0166662168,0.001735204,0.436003536,0.0011935598,-0.0289094243,0.0057189772,-0.2705666721,-0.1982909888,-0.2628411055,-0.0521498248,0.5193050504,-0.1110357344,0.0556781963,-0.1012390181,-0.0015995727,-0.1593154967,0.4202523828,0.3708048761,-0.4092085063,0.0447789431,0.4111919105,-0.2707718015,-0.0644954816,0.0988038555,0.2239130884,0.003082715,-0.2134789377,-0.1558633596,0.1541232169,-0.1610955149,0.7585225105,-0.0389431044,0.3208904862,0.1230188608,0.2718275487,0.2305573672,0.0542253479,0.2061479688,0.1448653042,0.562187016,0.3099980056,-0.0196944848,0.0971736684,0.3264108002,0.0630993545,-0.0672442839,-0.3433494866,-0.1086802632,-0.4698187411,-0.5020904541,-0.3492280245,0.186240226,-0.1147777885,-0.0567495115,-0.0262678452,0.1924752295,-0.4037152827,-0.0402305871,0.1861366928,-0.0963740647,0.290584296,-0.0736789852,0.0086626718,0.0947704986,-0.600821197,0.0399194509,0.3729530275,0.0501877442,-0.045901794,-0.049256742,-0.1823404878,0.2974664271,0.0599193126,0.0156688597,0.0565445498,0.1243474633,0.1701496392,-0.1073746681,0.3296920359,-0.009176976,-0.255836606,0.239184171,0.3976652622,-0.2288036197,0.0554700531,0.2924399674,0.104001753,-0.0533754304,-0.135449633,-0.0865137056,0.9239500761,0.4039599299,-0.0879211798,0.1360990256,0.0571726225,0.3250687718,-0.1673019081,0.0233501177,-0.0407405235,0.1561028212,0.035459742,0.2771095932,-0.2004754096,0.1922136098,-0.1422634125,0.0777810663,0.4323305786,0.0320542827,0.1780149639,0.3054023981,0.027436642,0.1152802035,-0.1805297881,0.2694378197,0.0630186796,0.2970751822,-0.1738760024,-0.2601338327,0.2822410762,-0.0949978083,-0.2343672514,0.2006904483,-0.4120526016,0.6487286091,-0.1319775283,0.1568270624,-0.0133425174,-0.0437755026,0.3030739129,0.0411281176,-0.2076859027,0.3271149397,0.2293645442,-0.1224403381,-0.2963023484,0.0512540974,-0.0923784152,-0.3777457774,-0.4222824574,0.0106383087,0.0484350622,0.0501606688,0.0060766558,-0.0367044099,0.2529656589,0.1965109259,-0.0656022951,0.2037361413,-0.3845929503,-0.19370161,0.2164293677,-0.3658157289,-0.2183419615,-0.1053692102,-0.2583471835,-0.2529479861,0.0709113553,0.3226187527,-0.2966625392,0.1166301668,0.4994286895,0.2290470898,-0.1556761712,-0.2609316409,0.0270839948,-0.5316746831,-0.4793562591,-0.0083187772,0.0745421275,-0.1114722416,0.0502201021,0.8319503069,0.1867344677,0.2823089957,-0.1856089234,-0.1821729392,0.0956995264,0.1608040631,0.2060508132,0.1743947566,-0.1765375584,0.2800470591,0.1834611446,-0.3471688926,-0.4221918583,0.2426226288,-0.1282037795,-0.2381153405,-0.0323537737,-0.1730272472,0.1891024262,0.1605658233,0.2417709231,0.2806186378,0.0259012934,-0.3505977988,-0.0810698196,0.0855247453,0.0579464175,-0.2438059598,-0.0828482658,-0.1742305011,0.1049230322,0.3559297025,-0.1319721192,-0.1443390697,0.1581627578,-0.0379949138,0.0086918185,-0.1188556179,-0.0614716411,0.0629177466,0.2807857096,0.5709138513,-0.0325559787,0.0065180915,-0.0738030002,0.0328642167,0.2689120173,-0.0943007693,0.1302497089,-0.1055922732,0.0436333232,-0.1390289664,-0.0741520673,0.1080576554,-0.0620591901,0.1913295984,-0.3745895028,-0.2634865046,0.1098070592,0.2776018381,-0.2236057222,0.0375326388,-0.0347842947,-0.3300499916,0.0958203375,0.0828517303,-0.0204195194,-0.1092336103,0.354085356,0.2324723899,0.2803201973,-0.4758968949,-0.238209486,-0.1816982627,-0.1316816658,-0.0154634072,-0.3503430784,-0.1771941036,0.3194108605,0.0781034008,-0.1703231633,0.0188190956,-0.0197055954,0.3411801457,0.1485902071,0.044083301,0.1610946506,-0.0599152446,0.0003502345,-0.0241924468,0.0293868072,-0.1098009124,-0.0195796955,-0.1973915547,-0.1539185494,0.619312048,-0.1055206582,0.0252174344,-0.2616587877,0.00696665,-0.403270334,-0.0074385726,-0.0815744251,0.0685799196,-0.1953364015,0.14531371,-0.1830940545,0.1016634777,-0.2391079515,-0.005184765,0.4331629276,-0.278980732,0.4016618431,0.4132735431,-0.0282858498,0.0438968018,0.3856124282,0.3969130218,0.1132662967,0.1107795238,0.1566824317,-0.0954248235,-0.2278515995,-0.0503253974,0.1480786204,-0.0246370062,0.1073471233,0.1450694352,0.2636876106,-0.2828329504,-0.1049006954,-0.090897955,0.020933846,0.0220387522,-0.1580241323,-0.2956181765,0.084152557,-0.2866209149,0.0994707868,-0.4870062172,0.0996255204,0.1636638343,0.0525531098,0.2180423886,-0.0397622287,0.1467602551,-0.1625911146,0.4775078595,0.2726364136,-0.1844510436,-0.1412128061,-0.4386581779,-0.4556953907,0.1453968883,-0.0879260376,0.0670265406,-0.1083729938,-0.0057791988,-0.0784822479,-0.1159134805,0.0541461222,0.0634011626,-0.1502088308,0.0462342836,-0.0497652292,-0.0989074931,0.0101955077,0.0693768412,0.0542018786,-0.2284081876,0.1736979932,-0.0407651924,0.1971675605,0.2268520743,-0.3216748834,-0.1897658259,-0.0956411958,-0.1758786142,0.1044713408,0.1514083147,-0.3203735948,-0.2788774967,-0.0516733453,-0.2463127375,-0.0294289477,0.3377509713,0.1665875614,0.1081342846,0.0734200403,0.4041680098,-0.1356108487,-0.0791942701,-0.0962046608,0.1723338515,-0.2020142972,0.0716027096,-0.0295140091,0.1594515443,0.1339401752,0.2817716002,-0.1828497201,-0.1493097395,-0.2904231548,-0.4855021536,0.5116254687,-0.2880659997,-0.0736498609,0.0456325971,0.2489917576,0.0118769072,0.3490402102,-0.2351570129,0.0681398734,0.2542476058,0.280811131,-0.0141268391,0.1741475761,0.0088763591,-0.1793542206,-0.0369106345,-0.1932938844,0.4421988428,-0.1399347782,0.3511251211,-0.026191406]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1590","title":"Add helper to resolve namespace collision","comments":"Do you have an example?","body":"Many projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. ","comment_length":5,"text":"Add helper to resolve namespace collision\nMany projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. \nDo you have an example?","embeddings":[-0.2538418174,0.0961689427,-0.1048552692,0.3527351618,0.1007491201,-0.0802501142,0.1382944435,0.2031300366,0.2435297817,0.2675812542,0.0080044167,0.0642518774,-0.2843755484,0.3053634465,0.0437697843,-0.1240668967,0.0446554124,0.136503309,-0.1512791514,0.2778183222,-0.2029750943,0.1177269742,-0.2499900311,0.1129506677,-0.1931457371,0.2057630718,0.2455874234,-0.0963441432,0.2151639462,-0.2511823475,0.6898103952,0.2073137164,0.3034313023,0.5816520452,-0.0001147484,-0.0382627472,0.3849980533,-0.0327037163,-0.1130325422,0.0168179758,-0.2535123229,-0.2632159591,0.1146168336,-0.076939337,-0.2012659907,-0.1324408948,-0.1562743038,-0.4227061272,0.107440643,0.0672538951,0.2215976715,0.2550992668,0.1861998588,0.0390672088,0.4427598417,0.1391973644,-0.2214101404,0.1489810795,0.3609550595,-0.2571060956,0.2527554333,0.2875652015,-0.3082804084,-0.1428987086,0.4232206941,0.0421376638,-0.2081580609,-0.0925102159,-0.1983264536,0.237156719,0.4816095233,-0.2598926127,-0.0672907233,-0.5967507362,-0.1074745581,-0.1123104021,0.3792419732,-0.3263862431,-0.1080235392,0.0183419492,-0.113366127,-0.283708781,-0.1945106536,0.2124413252,0.2003504336,0.077375181,-0.2997067571,0.1280593872,0.3740561306,-0.0547262877,-0.0288329646,-0.0401463397,0.060977146,-0.0784668401,-0.4546447396,-0.2980474532,-0.2049228251,0.1942734271,0.0665678829,0.0862723514,-0.1126920581,0.2162448764,-0.1452969164,0.1723137051,0.0969699398,-0.1027059779,0.1917689443,-0.0659987628,0.6449652314,0.2755848169,0.0048661307,0.042575784,-0.0409067236,-0.2722920477,-0.0030715729,-0.1023664251,0.2628497779,-0.1352831572,-0.0593192242,0.02344938,0.1603905261,0.0403919071,0.3161291778,-0.0133334463,-0.1267312467,0.0397000574,0.1072195396,-0.0825680345,0.0088224253,-0.4559513628,0.0206844229,-0.1146828532,-0.1312288642,0.0368769951,-0.056184724,0.2121676505,-0.1602214873,-0.1597829759,0.4318936467,0.0960522592,0.1113494858,0.234685272,-0.0772275999,0.1026117206,-0.1817710847,0.0135282828,-0.060621582,-0.2066824436,-0.239440456,0.0388296433,-0.4515498877,-0.0914450064,-0.0558065698,0.1212720871,0.1953946054,0.0926626846,-0.3738143146,0.0759549811,-0.1693217605,-0.4652647674,0.105242528,0.1391468495,-0.3210276961,-0.3292497694,0.1632020921,0.0451056361,-0.1833807081,-0.0543675087,-0.3837215006,0.04266591,-0.1691864133,-0.0557868853,0.1400238425,-0.1044549122,-0.3842353225,-0.1512086242,0.3036570847,-0.3763741553,-0.4241859913,-0.061348822,0.0057179714,0.0574397445,0.0883519277,0.5008837581,-0.1767054051,-0.2199774683,-0.1210777313,0.4968911707,-0.2484409213,-0.0903464332,0.1726094931,-0.0776643753,0.2433721274,0.2886810005,0.0796866193,-0.0278001726,0.0857152641,-0.0358646102,-0.0934001654,-0.2335895449,0.1872473657,0.2613114119,0.2896867096,0.272923857,0.0619632117,-0.5637215972,-0.6234832406,0.0877535641,-0.0268706344,-0.0894969851,0.0343999639,-0.1456347108,0.1271457076,-0.0467213318,-0.0432528183,0.1458862573,0.0109718302,-0.0470808782,0.3973810971,-0.3103847206,-0.2432839572,0.4124102294,-0.2882077694,0.3399765193,-0.4274469316,0.2822182178,0.2417871356,-0.045593407,0.0521373264,0.4664885402,-0.0599731132,-0.0098295668,-0.0049075792,0.4178019464,0.0526211374,0.2185953259,0.0686928779,0.3343612552,0.0211415812,-0.1276829541,-0.1912592649,0.0610071681,-0.1501164883,0.1168429181,0.008776865,0.2090116143,-0.244951278,0.2986317277,-0.0624939613,0.0102264015,-0.1723419279,-0.1886092722,-0.1446910352,-0.0196682438,0.2162183821,-0.1508025974,0.5300851464,0.0923176929,-0.1171914786,-0.0683377311,0.1988733858,0.0686829537,0.2940938771,0.3263250887,-0.2623599172,0.3358141482,-0.0984892547,0.2808036804,0.206392616,0.2032473534,-0.1857610494,0.2121060044,-0.027122857,-0.0467674471,0.1147648096,-0.0198614653,0.014685954,-0.1993021667,0.0781110078,0.0217220057,-0.0207986608,0.1387491971,-0.4275842011,0.0856281444,-0.2543712854,-0.0667687505,-0.3877125978,-0.5908028483,-0.2239125967,-0.1092461869,-0.5629714131,-0.164021343,0.1072710678,0.2337984741,-0.353776902,0.2698173523,0.2720312774,0.4828246534,-0.0018154554,0.0551374964,0.0841678157,-0.2862923145,-0.3080463409,0.0652754232,0.2411570698,-0.1108855754,0.4057464004,-0.2041593492,0.1402556896,-0.3862427175,-0.6766159534,0.0721303523,0.0908243731,0.1528758109,0.3252756,0.0997726545,0.3070839345,-0.2105762661,0.0443002172,-0.2851426005,-0.006242652,-0.0568206124,0.1022390723,0.1083042398,-0.051497817,0.1293146014,-0.0713769719,-0.5083194971,0.1659556329,-0.3258289099,0.0507177226,0.1405841261,-0.0457964316,0.1721437275,-0.5115203857,0.0921980888,0.0305316634,-0.5022735,0.024268087,-0.061016243,0.0178527478,-0.1287599951,0.0796119422,0.1421637237,-0.0172335971,-0.2008799911,-0.3147956431,-0.091942735,0.2054114938,0.3539818525,0.1961414367,0.128298834,0.0468301736,0.0232187044,-0.3142698109,-0.1382821053,0.074768737,0.0069934903,0.1724202782,0.4037826657,-0.0330142714,-0.3397261202,0.1108384654,0.0660185516,-0.1625700295,0.2853940129,0.1654502749,0.5372005105,0.2059573084,-0.1825682372,-0.0313363522,0.0404894128,0.055192966,0.2559786439,0.2254238576,0.2610011697,-0.3073948026,-0.4325138927,-0.0620122813,-0.3766532838,-0.1653377116,-0.0112851458,0.1544512361,-0.301330179,-0.6369467378,-0.142523393,-0.1239088923,0.3116272986,0.3184878528,0.2294988483,-0.0741070658,-0.7395923734,-0.005351393,0.0494383015,0.2673211396,0.0296204984,-0.0325730406,-0.1134854332,0.0090959864,-0.3395829201,0.3071244955,0.6736313701,-0.3017759025,-0.3071037531,-0.5296486616,0.0962029025,-0.1865166128,0.0752460733,0.1557705998,0.0653562397,-0.255454123,0.5097441077,-0.1124906689,-0.1085587963,0.0988620371,0.289855063,0.07323277,-0.1310596019,-0.0288930703,0.1607823819,-0.3899274766,0.0137002338,-0.0402098224,-0.1847279668,0.1465595365,-0.2294798493,0.0804490969,-0.0362670943,0.2168859988,-0.0301695839,0.2258195579,-0.1210257411,0.1758917272,0.2178521454,0.3441887498,0.239540264,0.2552680373,-0.3525523543,-0.7641575933,0.1778807342,0.2231353819,0.8594169021,0.1457107961,0.0137525201,0.3202854693,0.01260777,0.2635712624,-0.4891685545,-0.0116715143,0.1362303048,0.3724038303,-0.4171175361,-0.2620909512,0.4302103221,0.4017979503,0.0442449003,0.307314992,0.6909219623,-0.3515101969,-0.2056997418,-0.100140132,0.9675010443,0.2342445403,0.1087684333,0.0133813154,-0.141324982,1.0758259296,-0.1875181347,0.0104643982,-0.1215908378,-0.2269923687,-0.0349279381,-0.0509167686,0.0453685224,-0.1325257123,-0.1474636793,0.4256058037,-0.0167740546,0.4041464925,-0.1572455317,-0.0021185023,0.3457830548,-0.401499331,-0.3410648704,0.0303107332,-0.1625098586,0.0340353586,-0.1504639536,0.0565039888,0.1405218244,0.0250331834,-0.2415799648,-0.0997374281,0.2815174162,-0.2573367059,0.553876698,0.0631766766,0.210131526,0.098263666,0.6500083804,0.2453757077,-0.0782010928,0.0449781679,-0.305676043,0.2232055068,-0.2460215092,-0.1691181362,0.0145349093,0.0560590662,-0.0333339125,-0.0642208606,0.0880725086,-0.515781939,-0.2315044403,-0.036852967,-0.17259866,-0.1401658505,-0.0024619941,-0.2608411908,0.1435942054,-0.3010339141,0.1221664697,-0.1107269526,-0.0699000135,-0.0066904626,0.3066425323,-0.2258799523,-0.0970937237,0.1114780828,0.166870147,-0.1768072397,0.1894590855,0.21188429,-0.078574352,-0.11906261,-0.1267661452,0.424736768,-0.0011417181,0.2611353397,-0.1119388714,-0.1861346364,0.0932771489,-0.0432221666,0.0342101306,-0.1750774533,0.0275063198,-0.3880865276,-0.2356522381,0.4783287942,-0.1559481025,0.3966647089,0.2742111087,-0.2130383849,-0.1889297664,0.1342142075,-0.2216285616,-0.0693426654,-0.1607335657,0.0200870298,0.4670378268,0.258382678,-0.0689764321,-0.107364215,-0.0673812777,0.2223566175,0.1297643185,-0.081042096,-0.2038494945,0.1759560406,-0.0974894315,-0.052285213,0.2200405151,-0.178564027,0.1169208139,-0.2760953903,0.0183229651,-0.0496766567,-0.0549373701,0.0647497103,-0.0021813842,-0.1598231047,0.1033108607,0.6095228195,-0.0416728407,0.1216465384,-0.0048978389,-0.0087418044,-0.0542296246,-0.1067968607,0.3450841606,0.1909642816,0.4489977062,-0.2865984142,0.1781409979,0.0736023188,-0.1990318149,0.3774769902,0.1955315471,-0.0003022672,-0.0613838658,-0.0660529211,0.2279129028,0.0837559626,-0.2515330017,-0.3117325902,0.3738553524,-0.2350184321,0.1068815067,0.2489638031,0.1967437863,0.2975732088,0.2924079001,0.0283576213,0.3168817163,0.1892175227,0.3971841633,0.2350289524,-0.1507320553,-0.0646856427,0.3269256055,0.1080195382,0.1305189133,0.3392693996,-0.0596348979,0.4132078886,-0.0557436869,-0.0689234063,0.3169462979,-0.2499602586,0.2420344651,0.1486692131,-0.0749455541,0.0002004722,0.0143597992,0.1854296029,-0.1661150604,0.1431762129,0.1067466512,0.2738803923,-0.208683446,0.0204398092,-0.0245606191,-0.3529637456,-0.2033442259,-0.0749036297,0.0887106583,-0.1941024363,0.494707197,-0.1139589623,0.3081118464,-0.0448622517,0.051756952,0.3246577978,0.0916967243,-0.0580018498,0.1214546785,0.0258018132,-0.0823126212,0.3426403105,0.2101829797,0.3627081811,0.1299014837,-0.2497846186,-0.0747203454,-0.0661180615,-0.0174080972,-0.2725378573,0.0989694744,-0.1969267279,-0.1200926378,-0.0531753525,0.1615745276,-0.0532515235,-0.0430777185,-0.340511024,0.0020411627,-0.7071827054,0.0156794358,-0.2116163671,0.1047016084,-0.271953553,0.1201132685,-0.2376570255,-0.2066096514,0.5702577233,-0.0159309581,-0.1505480707,-0.0900887847,0.0588706471,0.0560878366,0.7925460339,0.4971894622,0.0527797006,0.0227187239,-0.4446482658,-0.6077390313,0.3915790617,-0.1931376904,0.1890028715,0.0305475388,-0.5825127959,0.224363029,0.1521835029,0.3870790303,0.0400181301,0.1929533929,-0.5062060356,-0.4849377275,-0.2349159569,0.192718178,-0.3917532265,0.1395439357,-0.1101194471,0.2122626901,0.3934630752,-0.078282997,-0.2256264836,0.3774135113,0.0176219661,0.1123294756,0.3234955966,0.1748977453,0.4409450591,-0.143911764,-0.305682987,0.0190054402,-0.142989561,-0.2796932161,0.0868803114,-0.3088429868,0.0259062182,-0.4057562947,-0.1393394172,-0.4208799303,-0.1245991439,0.073517926,0.2940758467,-0.4099245071,-0.1489328444,-0.1545719653,0.0847480744,0.1030103117,0.0573721491,0.0545935109,0.1392380595,-0.1471322179,0.0348082744,0.2748749256,0.0439790711,-0.2906852365,-0.2531274259,-0.0505416915,-0.1064447984,-0.1510363072,-0.4887059629,-0.1812971532,0.2067199051,0.0575402454,0.2754487693,0.2701907158,-0.0706040785,0.1844311506,0.0323125832,0.2122756392,0.3713397384,-0.0974227414,0.3297072351,-0.0362022743]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1590","title":"Add helper to resolve namespace collision","comments":"I was thinking about using something like [importlib](https:\/\/docs.python.org\/3\/library\/importlib.html#importing-a-source-file-directly) to over-ride the collision. \r\n\r\n**Reason requested**: I use the [following template](https:\/\/github.com\/jramapuram\/ml_base\/) repo where I house all my datasets as a submodule.","body":"Many projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. ","comment_length":29,"text":"Add helper to resolve namespace collision\nMany projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. \nI was thinking about using something like [importlib](https:\/\/docs.python.org\/3\/library\/importlib.html#importing-a-source-file-directly) to over-ride the collision. \r\n\r\n**Reason requested**: I use the [following template](https:\/\/github.com\/jramapuram\/ml_base\/) repo where I house all my datasets as a submodule.","embeddings":[-0.156945169,0.085805431,-0.0459537581,0.2750368118,0.1614482999,-0.0471068732,0.1879473031,0.2351552397,0.1399725974,0.0676217303,-0.1060155556,0.0726316646,-0.379468441,0.365108043,0.1478723288,-0.0800853819,0.1018497273,0.0871386304,-0.1015021279,0.2270358056,-0.2017204762,0.1321813911,-0.1773248911,0.0692506805,-0.1361082792,0.2808799446,0.2181895971,0.0411920734,0.0783126503,-0.3763766289,0.6635086536,0.2178044617,0.3690737188,0.6192420721,-0.0001174822,0.1117214039,0.3730975389,-0.0607055351,-0.1470485926,-0.1178849488,-0.0896531269,-0.398647964,0.157278046,-0.0462405048,-0.1469103098,-0.1982926875,-0.250744313,-0.3956302106,0.2224406898,0.1046413779,0.1570375562,0.3138197362,0.3482822776,0.1955201179,0.4042695761,0.0580365136,-0.103069976,0.2518261075,0.3102469146,-0.2774194777,0.1863786131,0.2999771833,-0.4233086109,-0.1609926224,0.5737731457,0.0806408972,-0.133542031,-0.1649817675,-0.3388842046,0.2239961177,0.3205664456,-0.3809448779,-0.2733329535,-0.6625205278,-0.0953679755,-0.3431422114,0.3581660986,-0.2587684691,-0.0489988141,0.1162643433,-0.0102240816,-0.3572150171,-0.2172235548,0.1401239187,0.0621522479,0.265432477,-0.118245177,0.0382897072,0.4718239605,-0.0413214862,0.1391025037,-0.1224656329,0.1292885244,0.0249687508,-0.3898670971,-0.2220895588,-0.0659959465,0.1947240829,-0.0027468349,0.0927888975,-0.1214878112,0.1434396654,-0.2469825,0.089137055,0.1438131034,-0.0759093091,0.0228997152,-0.0074788737,0.717302084,0.278201133,0.0356224775,0.1428511143,-0.0309190508,-0.279659003,-0.1277398765,-0.0557538271,0.3128400445,-0.083722353,-0.1144183874,0.1310545057,-0.0295079686,-0.01024203,0.3119973242,-0.0674568936,-0.1261175126,0.2743070722,0.2009346187,-0.014787293,-0.0914753675,-0.3621776998,0.0768510923,-0.0200122334,-0.1118421853,-0.1172181293,-0.0236462485,0.1487340629,-0.1232327819,-0.1212419197,0.4266929924,0.0588852316,0.1227620766,0.2128577232,-0.0646680519,0.130825609,-0.1384952068,0.0264550764,0.0455058515,-0.2981049418,-0.2503246367,0.0837171525,-0.2816407681,-0.1105155125,0.0168741271,0.1137429625,0.0872917548,0.0983204395,-0.3846315742,0.066032514,-0.1101949885,-0.3452208936,0.045744326,0.0482117534,-0.174320519,-0.2597379982,0.2053050697,0.1476268619,-0.3792133629,-0.0783925578,-0.3599071801,0.0991626084,-0.1537526548,-0.0157151036,0.0686499849,-0.1013021022,-0.3450179994,-0.1573765278,0.1345909834,-0.4595775902,-0.4601137936,-0.0651671961,-0.005425611,0.1478833109,0.1906592399,0.5217299461,-0.3122331798,-0.2124907225,0.0728029087,0.5595843196,-0.2102899998,-0.0309661087,0.1056246608,-0.1280076951,0.1873516887,0.175775826,0.0800944716,-0.1775848716,-0.0847342461,0.0296681896,0.0435900353,-0.2440781742,0.0962440521,0.2909335792,0.3882429004,0.2304106057,-0.048839692,-0.4740625024,-0.5224717259,0.0260031708,0.0659537911,0.1904855072,-0.1272553951,-0.1919789612,0.1316481978,0.0562363416,-0.0914121121,0.1116883308,-0.023463849,-0.1359507591,0.3357867599,-0.211732924,-0.2838371396,0.3958899379,-0.1835471839,0.4851572216,-0.473492682,0.3816845417,0.1658197939,-0.0012573909,0.0393790901,0.5233588815,-0.0207728297,-0.0889599398,0.0956602693,0.3063693643,0.0091867987,0.2391800135,-0.0697517172,0.3499189913,0.1179017425,-0.0932851136,-0.1358383149,0.0945345461,-0.0933089778,0.1238581985,0.1265600771,0.4233871996,-0.2325821668,0.277510941,0.0566695891,0.0093641998,-0.0860435218,-0.1140576079,-0.1919590831,-0.0608415939,0.2643036842,-0.0066962368,0.5013024211,0.1276096553,-0.0960039794,-0.0228632148,0.1501436234,0.0308485869,0.1486248672,0.1569752544,-0.3729528487,0.3065495491,-0.0669645965,0.3952649832,0.2192809135,0.1973306686,-0.2224758267,0.3172088563,-0.0031693033,-0.0333979614,0.2044757456,0.1071964651,-0.0041245315,-0.1295863539,0.1589461863,-0.0098833358,0.011780642,0.1770435423,-0.3526941538,0.1459255069,-0.398136586,0.059504699,-0.3560387194,-0.5332534909,-0.3027786314,-0.3312542737,-0.6207664609,-0.2648798525,0.0466144644,0.3321264982,-0.3618955314,0.1726870686,0.215398252,0.3085461855,0.014952302,-0.2485610843,0.2147975713,-0.2350459099,-0.2394423932,0.0423847996,0.4659856558,-0.1626095027,0.2937952876,-0.1141601205,-0.0810242593,-0.2225541025,-0.6898431778,0.0872551426,-0.0175350346,0.227743268,0.3318365514,0.1090879366,0.3108024001,-0.1535188705,-0.0013428192,-0.2860572934,-0.0153256878,-0.0740682483,0.1047858819,0.2111226767,-0.0123377629,0.1076954082,-0.2220643461,-0.4446206689,0.1081407964,-0.1349276304,-0.0310644787,0.2012324035,-0.0743323788,0.1763153821,-0.2543161511,-0.0639796034,0.1600046009,-0.3523983657,0.1451933831,-0.0190420412,-0.0227381904,-0.098684676,-0.0657074824,0.0990856513,-0.0748306289,-0.2034688443,-0.4110997617,-0.1161350384,0.2301383615,0.3544764817,0.2806236148,0.0717646703,0.152112022,0.1068637371,-0.2129294872,-0.0888511017,0.0074360888,0.0716136098,0.0171761755,0.4399723411,0.0482326634,-0.2288437635,0.2650896907,0.1216901988,-0.1875744164,0.3132732511,0.1984449625,0.5431454182,0.1646975875,-0.247422725,-0.1733193249,-0.0199507885,-0.0514642708,0.210297972,0.1896425188,0.3960986435,-0.3102763295,-0.2964318395,-0.056904044,-0.4034613669,-0.024837127,0.1067196652,0.0819286555,-0.1414052397,-0.5966582298,-0.1719109714,-0.2243926674,0.2201675177,0.400853008,0.1940349042,0.0167091582,-0.6984470487,-0.0778701156,-0.2181310207,0.2828424573,0.0598677061,-0.1170995906,0.1309958547,-0.0083762035,-0.2977267504,0.091097638,0.5523031354,-0.2209061086,-0.3059738874,-0.4303902686,0.076212883,-0.1707992852,-0.0047949166,0.1300356835,0.0926482752,0.0008676559,0.4298421741,-0.1858716905,-0.065000698,0.0752333552,0.512958467,0.0454751067,0.0409818329,-0.1126858518,-0.0259399973,-0.4455135763,-0.0904305726,0.1137190834,0.0034029849,0.1350713968,-0.0784126744,0.0346181057,-0.1736755818,0.3371174037,-0.0142901037,0.2323843539,-0.109348163,0.1724034846,0.1334079057,0.3156104982,0.3287723362,0.4573601186,-0.4361449778,-0.7642879486,-0.0867230818,0.194463864,0.7389248013,0.2048791051,-0.1225195304,0.3060646951,0.04740135,0.0907476246,-0.490866065,-0.016903637,-0.0246188678,0.4625957012,-0.3601637483,-0.4996226728,0.4395856261,0.4005669355,0.1093072817,0.1701033413,0.6332274079,-0.374368161,-0.0511947796,-0.1848616153,0.9142554998,0.1587291211,0.2362566739,0.0915777832,0.0089893239,1.0984016657,-0.2501588166,0.1661792099,-0.1380213499,-0.1625464261,-0.0547795594,-0.059036117,0.0768417343,-0.2288736552,0.0413752161,0.4591515064,0.0121202823,0.3413606584,-0.2457136512,0.0985910594,0.3570766747,-0.3516695201,-0.6398153305,-0.0089944648,-0.0716337264,0.0040827817,-0.070828408,0.0169754978,-0.0489109941,-0.0213700924,-0.2332990766,-0.0892185494,0.1305248588,-0.2368203551,0.3903694153,-0.0991284847,0.1750008315,0.173531428,0.6006442904,0.2176595926,-0.0340732262,-0.0452482067,-0.3088108897,0.192715019,-0.0972944275,-0.1421903372,0.2157854885,0.0186987221,0.0276928339,-0.0319658332,0.1370048225,-0.5409186482,-0.0854680985,-0.0652296767,0.0633702055,-0.1768949926,-0.1592888236,-0.1832463145,0.1586174071,-0.254870683,0.0855118036,-0.0772722438,-0.2614862919,-0.0641113669,0.122467503,-0.1924477518,-0.1129999906,0.1861415058,0.1958611459,-0.0697996989,0.3000404239,0.3983023465,-0.0896990076,0.0027384625,-0.0739803314,0.3805935383,0.0014836602,0.2747151554,-0.0524310768,-0.1981116086,0.0061163721,-0.0309930462,0.0871137232,-0.1627963781,0.0946489424,-0.4005634785,-0.2875578105,0.398167491,-0.12937437,0.2608496249,0.3441787064,0.0217090677,-0.2792416513,0.0087309256,-0.2015838623,-0.0921011344,-0.0532379672,-0.065086171,0.3431401253,0.1689762622,-0.0307396092,-0.1561994255,-0.0663721189,0.216906324,0.1314416826,-0.0688231811,-0.2563326359,0.167384252,-0.0377693474,-0.2141346484,0.2650285363,-0.2362389863,0.0974013507,-0.2687414587,0.0230754111,-0.0017983164,0.0086684627,0.0040368983,-0.0032617357,-0.1228569001,0.0132449483,0.5187285542,-0.0651389286,0.1203402653,0.2021673769,0.1430522203,-0.0020971415,0.0071394616,0.2849828005,0.1805705875,0.3072295785,-0.2070939839,0.2245764881,-0.0459452197,-0.2484918088,0.475443393,0.3130010366,0.2250478417,-0.1340699047,-0.030816514,0.2520117164,0.051124014,-0.4314862788,-0.4235868752,0.4019429386,-0.2347677052,0.1546393186,0.3005873263,0.2775352299,0.2926278412,0.0977240652,0.2149344385,0.2707152367,0.3168258369,0.4239744246,0.3118394315,-0.0848891065,-0.0783633143,0.2291675657,0.0989611521,0.13558577,0.4182701707,-0.2440088838,0.3526082337,-0.180142805,-0.1247176677,0.1538202316,-0.3897278011,0.188908875,0.1089881733,0.0265675727,0.0866498575,0.0747835413,0.2063043863,-0.2038428336,0.1101153269,0.1600789279,0.3905013204,-0.1938611269,-0.1050224975,0.1225996315,-0.2557954788,-0.1593945026,0.0323051922,0.0722959489,-0.2947138548,0.2978249788,-0.0655343905,0.297848165,-0.1992426664,-0.0010342131,0.1497366875,0.1076391414,-0.0917860791,0.0447564535,0.111897327,-0.022515595,0.2294863015,0.02457509,0.3366083503,0.1335032284,-0.2941164374,-0.0256585646,-0.1241124496,0.0392422453,-0.3147287667,-0.0958593115,-0.2602224946,0.1392739862,-0.0068141571,0.110634625,-0.0508338809,-0.1520015746,-0.2369687855,-0.0735137984,-0.6723147035,0.1920060664,-0.066082947,0.0562682152,-0.329603821,0.1361298561,-0.2355738878,-0.158604756,0.7526233196,0.0516193584,-0.0734772608,-0.0743599012,0.0441377461,0.0399626195,0.5922054648,0.4784473181,-0.0028789761,-0.0771372169,-0.2362503111,-0.7712089419,0.3004339337,-0.2262996733,-0.0103526926,-0.0312571749,-0.5115121007,0.2582508624,0.229048267,0.1962855607,0.0668907017,0.1850357354,-0.4785355031,-0.4799033701,-0.3376756608,-0.0363710187,-0.3446259797,0.091180034,-0.0946876928,0.0666482449,0.294754982,-0.0861897841,-0.2589507103,0.2926538885,0.0432743579,0.1321386099,0.3314138055,0.2384781837,0.4119388461,-0.100386411,-0.3274398446,-0.0184658598,0.0076052407,-0.4604479074,0.0969589129,-0.3055747747,0.1370216757,-0.471842587,-0.1260854751,-0.5198672414,0.03025591,0.1490972787,0.0618255325,-0.4667416811,0.0181366839,-0.233824715,0.0921585262,0.0501262471,0.2110858858,-0.0013973444,0.2109672129,-0.1268211156,-0.0309888087,0.2041135579,-0.2729181349,-0.1864511669,-0.2045772523,-0.0682547688,-0.0898671895,-0.1786532849,-0.5367215276,-0.1130992919,0.1399765462,-0.0038185213,0.1610143632,0.262250185,0.0091805467,0.15462403,0.0039020951,0.0273160394,0.3570766449,-0.1450828016,0.3185430169,0.0174268242]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1590","title":"Add helper to resolve namespace collision","comments":"Alternatively huggingface could consider some submodule type structure like:\r\n\r\n`import huggingface.datasets`\r\n`import huggingface.transformers`\r\n\r\n`datasets` is a very common module in ML and should be an end-user decision and not scope all of python \u00af\\_(\u30c4)_\/\u00af \r\n","body":"Many projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. ","comment_length":34,"text":"Add helper to resolve namespace collision\nMany projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. \nAlternatively huggingface could consider some submodule type structure like:\r\n\r\n`import huggingface.datasets`\r\n`import huggingface.transformers`\r\n\r\n`datasets` is a very common module in ML and should be an end-user decision and not scope all of python \u00af\\_(\u30c4)_\/\u00af \r\n","embeddings":[-0.0202236325,-0.2652241588,-0.0927194953,0.1554595232,0.3644935787,-0.1625663787,0.2206045687,0.1671950966,0.0891767815,0.2554261982,-0.2716357112,0.1697729826,-0.2384689003,0.2573799193,0.0592853911,-0.2396575511,0.0753815696,0.197005704,-0.3191300631,0.0870092362,-0.2288310677,0.1971452683,-0.1359622777,0.0589765124,-0.2770675421,0.0928416848,0.2053655386,-0.0865936726,0.0594990812,-0.3789724708,0.4213656783,0.1152880341,0.0546915084,0.5897495151,-0.0001011381,0.1318942457,0.337695092,-0.0408486165,-0.0935960487,-0.0902457908,0.0646301135,-0.2467926443,0.2434570044,-0.0631534308,-0.2757797539,-0.1726580709,-0.0201392658,-0.1908741295,0.4177950025,0.0541873388,0.346364826,0.5181908011,0.233932972,0.0629720017,0.2500138879,0.1026623249,-0.1660955846,0.115399763,0.191190809,-0.1256443113,0.0923160687,0.4077009261,-0.2057329267,-0.1077073067,0.4733327329,0.1147729009,-0.1144275665,-0.3207325935,-0.1828877181,0.2236241102,0.1393297613,-0.4643296003,-0.3387849629,-0.4886345267,-0.1247690469,-0.1406293362,0.3330661952,-0.1330355108,-0.04382772,0.1750166118,-0.0336099826,-0.0629154816,-0.1290286928,0.088739343,0.0112887789,0.3285007775,-0.2876746058,0.0572053492,0.2873640954,-0.1634868681,-0.1059212163,0.0114281531,0.1962403804,-0.0631782189,-0.23037377,-0.2078415751,0.0271324832,0.2500008941,0.0400675647,0.0164244473,-0.1680965573,0.2096388489,0.1057278141,0.1243800223,0.0933449343,-0.0037212193,-0.1013072059,-0.1184755191,0.5116112232,0.103482753,-0.0130224088,0.0168297067,0.017249098,-0.224902004,-0.1072269008,-0.1891673058,0.3918105066,-0.0000874776,-0.1438305229,-0.098392278,0.084255442,0.1389573067,0.2741027176,0.3232211471,-0.0582580082,0.1231165156,0.1442242414,0.050936278,-0.1939392984,-0.1619465649,-0.0936441422,-0.0890825838,-0.2535675168,-0.1062750742,0.021815002,0.1464740187,0.1363381892,-0.082087256,0.3387714028,0.0901669711,0.122593388,0.2884990573,0.0387651324,0.2551595569,-0.0981007591,0.0819577947,0.0615825318,-0.1290495694,-0.2916733623,-0.255592376,-0.2272693068,-0.0853667483,-0.0852701813,0.2785585821,0.0571323857,0.1005708426,-0.0861962512,0.0224285461,-0.0613554418,-0.2259920388,0.0058093392,0.0954289213,-0.137312457,-0.2678747177,0.1736903787,0.1147441566,0.0881103128,-0.221149683,-0.1986528337,-0.1133023798,-0.0881039575,0.1867656857,-0.0536876321,-0.1450605243,-0.2509948611,-0.1305901259,0.0392693505,-0.3478722572,-0.4469206929,0.003473382,-0.0512945242,0.0921060294,0.2216733098,0.2875139415,-0.2845281363,-0.1318760365,0.1497124135,0.334582448,-0.1078119054,0.1419370323,-0.0702395961,-0.1103658304,0.0854812488,0.0956269726,-0.0715063512,-0.0109138954,-0.1561855972,0.0679795146,0.0651495233,-0.1832594573,0.012015162,0.3451623619,0.4077172577,0.2324431092,0.0379570313,-0.3982560039,-0.5612103343,0.032327652,-0.2237790078,0.1317364872,-0.102172792,-0.381174773,-0.0273611676,-0.0074822535,-0.101243414,-0.0579985864,0.2481876612,-0.033184994,0.2494141012,-0.1152242944,-0.1502001137,0.2772931755,0.026908299,0.293605715,-0.5201870203,0.2467380464,-0.0133098299,-0.0575284772,0.1687585562,0.4031574428,-0.0246705879,-0.1086483672,0.0968585834,0.3235653341,0.0206217505,0.2406048775,-0.1278152019,0.3849263489,0.0608320311,-0.1427503079,-0.0769751817,0.2066701651,-0.0725471973,0.176276207,0.0158786476,0.4241325557,-0.1905089021,0.1394430697,0.1381799281,0.0197811555,0.1021703109,-0.1923259348,-0.313311398,-0.2394379079,0.0874391869,-0.0588207208,0.521186769,0.27946648,-0.2153863162,-0.1565445662,0.3494957089,0.1497510225,0.2577457428,0.1061117202,-0.4559352398,0.2515173554,-0.0359121971,0.2682374418,0.1603031605,0.3987917006,-0.159025833,0.2306382805,-0.0639001951,-0.1053712294,0.1615660042,0.2154632956,-0.0500359945,-0.0688724667,0.0756266043,-0.0285415687,-0.2224746048,0.0728974417,-0.3403604329,-0.0364852473,-0.3240096569,0.1397470832,-0.2611646354,-0.3896709979,-0.4315490425,-0.1077602357,-0.5112162828,-0.1372930259,0.0964360014,0.1642341763,-0.3801087737,0.3237053454,0.2867611051,0.2110137194,-0.0020325056,-0.1493339837,-0.0526552461,-0.1076910049,-0.2154489309,0.1281558126,0.2769778967,-0.030529527,0.391313374,-0.1259329319,0.0550007448,-0.1950947493,-0.6819356084,0.0630114228,-0.0705563426,0.1962787956,0.3726926744,0.0421325117,0.3110900819,-0.3523676395,0.1790613234,-0.2390560806,-0.1668489277,-0.0133356946,-0.0222005192,-0.0204774402,-0.140968889,-0.0746177286,-0.1765633076,-0.5822033286,0.3250768185,-0.1797237992,-0.0627534762,0.3977240026,-0.0689526424,0.1572034359,-0.3618356884,0.2050260305,0.0232381057,-0.2114356011,0.2111343443,-0.1115906388,-0.1141852885,-0.1968887597,-0.1522673517,0.2833880186,-0.1439309716,-0.1671087444,-0.3801899552,-0.2101474404,0.0034125147,0.327509284,0.2715934515,0.1351822764,0.2047846019,-0.1520042866,-0.2211496085,-0.1301624477,-0.0582694225,-0.0515990667,0.1095560119,0.1124327257,0.1236451566,-0.2242598236,0.278770715,0.2801963091,-0.2398181558,0.252494812,0.0529353768,0.4887318909,-0.0350748785,-0.3629176319,-0.0896551386,-0.0427206457,0.0346949995,0.2865269482,0.0250556599,0.2448738664,-0.1581250876,-0.1891269833,-0.0824503079,-0.4748337567,-0.0026112495,0.1852330714,0.1192943379,-0.0933772326,-0.3426036239,-0.1773270518,-0.0428630188,0.2780269086,0.2471112311,-0.093053326,-0.0193603225,-0.5284514427,-0.0593466461,-0.2361170053,0.2037996352,-0.0679994226,-0.0176141001,0.0776781589,-0.1631179005,-0.1769984066,0.2165578753,0.6753464937,-0.0499911606,-0.2833184004,-0.2240954489,-0.0680461675,-0.3274958432,0.1146276891,-0.0210918374,0.1283043772,0.0931736827,0.4130401909,-0.3220464587,-0.2418492287,0.0810483098,0.2850701213,0.0428887159,0.002952751,-0.3901442289,-0.0289212726,-0.3902353942,-0.1144888699,0.1473126262,-0.2407577485,0.14699395,-0.1408347487,-0.041487202,-0.2923750281,0.2965859175,0.2387551665,0.3894861042,-0.0658739805,0.1300346553,0.2103615254,0.4148670137,0.2032137811,0.4545965791,-0.4082917571,-0.6085441113,0.0584281459,0.15215078,0.4647033811,0.2070516348,0.0766121149,0.3528668284,0.0812071264,0.2780640721,-0.4330651462,0.1400018036,0.2168093771,0.334644258,-0.201672554,-0.2722825408,0.4416021407,0.090062283,0.0070571285,0.1701144725,0.6453772783,-0.3645779788,0.1618507057,-0.0969085842,0.9009621143,0.0809037536,0.2035181522,0.1620769203,-0.029331319,0.9161871672,-0.1057747677,0.182910502,-0.1810831279,0.0233982261,0.0717688426,-0.0271749217,0.059339121,-0.1545616835,-0.2843590081,0.308226794,0.0685515925,0.2132790387,-0.2964932919,0.2089805305,0.1228097305,-0.4371929765,-0.5631967187,0.1779294163,0.0335508473,0.2441462725,-0.0881635845,-0.1733076572,0.0018395197,-0.098458603,0.0139704291,-0.1171323508,-0.0323285051,0.0407738313,0.4415958226,0.0378803611,0.2005828023,0.2026889026,0.5129510164,0.2344173044,-0.263061136,0.126295954,-0.2502359152,0.1097511947,-0.2437956482,-0.0091928924,0.2004680634,-0.0229763128,-0.0851151869,-0.1354885399,0.124202691,-0.2236679196,-0.2185149789,0.0265588313,0.0029758487,-0.2992792726,0.0492848232,-0.1533480138,0.1045874283,-0.2569082677,0.2286133766,-0.1673405766,-0.1289894134,0.0257409196,0.2801822424,-0.2060462236,-0.107290335,0.1382480264,0.1594892442,-0.0734928399,0.2619618773,0.1753440946,-0.1808309406,-0.2089405805,-0.1262280643,0.3283751905,-0.2221455872,0.3311892748,0.0556369834,-0.292907387,0.0798499733,0.0831328705,0.1865786761,-0.1984474361,-0.1031372175,-0.0254355334,-0.4370602071,0.4916541278,-0.0650669932,0.3749508858,0.1244235262,0.3562999666,-0.1330062151,0.0781966299,-0.4297306538,-0.1096433476,-0.3257800937,0.0809883997,0.2810426056,-0.177066341,0.1059286743,0.0081501603,0.1589883268,0.1806968153,-0.176199168,-0.2516176701,-0.2799301744,0.0567980856,-0.1202951968,0.0320821069,0.2483464628,0.0161120165,0.0384872332,-0.2850759625,0.1489430666,0.1143903136,0.1309931427,0.0645134076,0.0306733828,-0.1078486964,0.09680029,0.0926954746,-0.0895815715,-0.0007836206,0.159846738,-0.0319489688,-0.2284413874,-0.0313014276,0.2267944813,0.1584984064,0.3250283599,-0.1643187553,0.0868043974,0.0367913097,-0.3148100972,0.3681905568,0.2952308357,0.187450245,-0.1608654857,0.0955852494,0.2768040299,0.2737602592,-0.4097084403,-0.1916365623,0.2237161249,-0.2172228694,0.3088392019,0.3486171663,0.1564535052,0.4141766429,0.326905936,0.1284358054,0.3097162247,0.1166208014,0.3002035618,0.0921415389,-0.0174397603,-0.0431226976,0.3794749379,0.1942272782,0.33039096,0.4142693579,-0.2103564888,0.3748688996,-0.4779445529,-0.0294906795,0.2914586365,-0.1085098758,-0.0213150252,-0.0218520314,0.0324620493,0.2077528536,-0.0438360348,0.2561402321,-0.2644205689,0.0026093933,-0.0429600514,0.2710442543,-0.2268692106,-0.2666248381,0.2371620536,-0.1558845639,-0.0860588551,-0.1094587818,0.1541882157,-0.1426190585,0.2459130883,-0.0812180564,0.1960862875,-0.291028142,-0.0773115382,0.2626054883,0.293114543,-0.1907213479,0.0867767259,0.1322947592,-0.1513995379,-0.0454268195,0.1123116016,0.3918342888,0.1082319766,-0.2031149566,-0.0421204641,-0.0660861284,-0.0447556973,-0.3440140784,-0.0971739888,-0.0816090405,0.3048807085,0.0610508434,0.2392605096,-0.2412866056,-0.3084822595,-0.2968113422,0.0002041004,-0.4568963051,0.1837311983,-0.2699301243,0.0193448439,-0.3872521818,0.0133888647,-0.5692485571,-0.1178877205,0.4266919494,-0.0584457107,-0.008873472,-0.2350172997,0.1313991547,0.1579085439,0.6225574017,0.2383096516,0.2647868395,-0.091632396,-0.246731326,-0.6086645126,0.313717097,-0.1753590554,0.0879136845,0.0039587975,-0.2102925181,0.1343161762,0.0930103138,0.0652892143,-0.1212890297,0.0376515426,-0.4913036823,-0.4235429168,-0.1050194725,-0.2243243903,-0.2587658465,0.3489259183,-0.1268956214,0.0430244841,0.2042160332,0.0871865153,-0.3307678998,0.3949809968,0.0761258975,0.1387530118,0.2521378398,0.1064618081,0.3198624551,-0.2111983597,-0.3475871086,-0.0720267519,-0.2325626016,-0.3384207487,-0.0668736175,-0.0558733456,0.2629168332,-0.3570220172,-0.3147628009,-0.3939545155,0.047957141,0.2586736679,0.001433684,-0.2825321853,0.1819225848,-0.0618685335,0.1458425075,0.0404771008,0.3666268885,0.0287965536,0.115091674,-0.1892728657,-0.2648586035,0.4914456308,-0.2128135562,-0.2221877575,-0.0923758075,0.1810762286,-0.0729140788,-0.2240480036,-0.6383808851,-0.0246796887,0.0754350722,0.0113524785,-0.0957603827,0.2912734449,0.0606011748,-0.0935273394,-0.0304584987,0.0751857013,0.2853314877,-0.13556467,0.3221633434,-0.1054508686]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1590","title":"Add helper to resolve namespace collision","comments":"That's a interesting option indeed. We'll think about it.","body":"Many projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. ","comment_length":9,"text":"Add helper to resolve namespace collision\nMany projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. \nThat's a interesting option indeed. We'll think about it.","embeddings":[-0.333307296,-0.0606815554,-0.1414774507,0.1775356382,0.296657145,-0.2553654909,0.2027596533,0.3330878913,0.1583759487,0.4334588945,-0.2013991773,0.1073923707,-0.1952550858,0.3127579391,-0.0596751682,-0.0131464833,0.0125782546,0.2596198916,-0.2767188847,0.1369097084,-0.2634368241,0.1955543756,0.0868326351,0.0119076297,-0.1752835214,0.0200133622,0.202593118,0.1165703908,0.0287667867,-0.3427736163,0.3376656771,0.2818188667,0.0009694082,0.4470309317,-0.0000963333,0.0757819116,0.2724318802,0.0310884956,-0.06162972,0.1533776671,-0.1032938212,-0.1571265757,0.0647667199,-0.2174678594,-0.3371703923,-0.1311044395,-0.0275202431,-0.3242069483,0.2175551802,-0.1206497103,0.3979304731,0.3877052367,-0.0547075644,-0.0234938599,0.3167296946,0.0453084148,-0.248452425,0.0248797927,0.2681359053,-0.1039410532,-0.0684755892,0.4525310099,-0.1548681706,-0.0115815839,0.3274198174,0.0816635787,-0.2268363535,0.011442191,-0.1157343984,0.2353540659,0.1948137879,-0.3045111597,-0.2073897272,-0.3909028471,-0.1736368984,-0.2232000381,0.3208946884,-0.1932113022,-0.117592819,0.2257913053,-0.0904175267,-0.192390725,-0.2032581866,0.0769213438,0.1911443323,0.1222716942,-0.2283041775,0.0101811942,0.2673758268,-0.0551152714,-0.122013472,0.0923819914,0.0590345375,-0.367832005,-0.2345868945,-0.2298954278,0.0517702997,0.1737319529,0.2540224195,0.0652254149,0.0100802854,0.314270705,0.0841415599,0.1480083466,0.0666729137,-0.0999607518,-0.0079377554,-0.2983982265,0.6054139137,0.2149369866,-0.0465402529,0.0280576274,0.1141888574,-0.1979860067,-0.0695745498,-0.1555554122,0.3000428975,-0.0941214636,-0.0925313681,-0.0740909278,0.2610558867,0.2631855309,0.1889272332,0.2856072485,-0.0196399726,0.0613721535,0.1225191206,0.0117630549,-0.1332187206,-0.3549470901,-0.122941792,-0.1513354331,-0.1137515977,-0.0135091478,0.0628323629,0.1073825955,0.097193554,-0.1323292851,0.2690577507,0.1273274422,0.1895700544,0.3062681854,0.1926297098,0.1599747688,-0.0834835246,-0.1412400305,-0.073420085,-0.1403066367,-0.1725201458,-0.1769116074,-0.165719524,-0.0564307198,-0.1896015257,0.3019244075,0.1814424694,0.0908604413,-0.0468594283,0.1786413044,-0.168395102,-0.347379297,-0.0218040384,0.2809993625,-0.3013656139,-0.1715974063,0.1279107481,0.089594312,-0.032764297,-0.0996069312,-0.049016133,-0.1074845493,-0.0903386176,0.1784599125,-0.042396985,-0.2677704394,-0.248689115,0.0147513589,0.0557963923,-0.2430986613,-0.3025396168,-0.0196337905,-0.1484732926,-0.0924589932,0.2653501034,0.3865014613,-0.1153534055,-0.1740977764,0.0265773647,0.3150395155,-0.1612882614,0.2428852171,-0.1515519172,-0.2352839857,0.0520283394,0.1850319505,0.0128649389,-0.0507159233,0.1278318018,-0.013348137,0.0157123376,-0.3812432587,0.1598541141,0.2632340789,0.2368267179,0.0907846615,-0.0371300057,-0.3642901778,-0.6232244372,-0.0254432857,-0.0827104673,0.006116882,0.1265015304,-0.4249512255,-0.026979506,-0.0313404761,-0.003818606,0.029945327,0.3004103601,0.1754710972,0.0527469441,-0.278400749,-0.2285149992,0.2843088508,-0.1754951626,0.1712910384,-0.3724256754,0.1086039245,0.0209976975,0.036255382,0.2177303433,0.4133128822,-0.0669860244,-0.0386518352,0.1480653584,0.3652946353,-0.07374385,0.2067081332,0.1551920921,0.3724741638,0.0995223448,-0.0549602695,-0.0395490229,-0.0391394049,-0.0434592292,0.18170771,-0.0161170345,0.376588583,-0.093348287,0.0696438327,0.1228674129,-0.0027425925,0.1634846032,-0.3034064174,-0.268925935,-0.1945384592,0.1027919874,-0.0511493199,0.3917599022,0.2089340389,-0.3882794082,0.0168349519,0.3002970219,0.2633997202,0.3474872708,0.185023129,-0.1761864126,0.1676325947,0.0106448513,0.2298925519,0.1048761383,0.4484815598,-0.0449456275,0.1367322952,-0.0329614319,-0.1282529086,0.0727512836,0.0390520804,0.0137084266,-0.0432737134,0.1587101519,-0.0845085829,-0.4501670599,0.0919886678,-0.3051521182,-0.1634059995,-0.3020127118,-0.0191347077,-0.2026892155,-0.3608114123,-0.1311849058,0.0304527692,-0.4165005386,-0.1175317913,0.3669065833,0.280711025,-0.5986717939,0.3736685812,0.1856944859,0.4932311773,-0.1376316547,0.0980879813,-0.074211441,-0.2706514001,-0.0640834868,0.1934779882,0.180268541,-0.0183864441,0.4568412006,-0.1273225695,0.1533268094,-0.3703358173,-0.6064677238,0.0894348249,0.0195191298,0.1440528482,0.2594346404,0.1393136233,0.2021205574,-0.3281260133,0.0329373442,-0.3339532912,-0.211526975,-0.0277394652,-0.0166426916,-0.0477615111,-0.2526448965,-0.116301164,0.0074345777,-0.5913141966,0.2939190865,-0.2362995595,0.0466195419,0.2157261968,-0.0407978706,0.0813178867,-0.4590120614,0.2263478041,-0.1973293275,-0.4127545953,0.0572145171,-0.2190865427,-0.1954445541,-0.2230259925,-0.014571989,0.2234169245,-0.0295960903,-0.2719365656,-0.3736749589,-0.0365193225,0.1959301382,0.3376681805,0.0000063875,0.1604583412,0.1264894456,-0.2860069573,-0.2656854093,-0.0720099956,0.1886471361,-0.1372971088,-0.0720968172,0.0591898821,-0.0591771193,-0.1188002825,0.1287934929,0.0130108027,-0.1580894142,0.4367743731,0.034222357,0.6551917791,-0.0385365523,-0.1920598596,0.1061223447,0.0970548466,0.0270564668,0.4082346261,0.086765781,0.0654512718,-0.1742072552,-0.0921192542,-0.125732556,-0.4641284347,-0.073496528,0.3061481416,0.030937491,0.0265894588,-0.4601980746,-0.1231579855,-0.0292792078,0.3053368032,0.219337672,0.0154745383,-0.092461817,-0.5202347636,0.1809443682,-0.2354489267,0.23567155,0.0364424549,-0.1015507355,-0.1704501957,-0.2158120424,-0.2505109608,0.1614661664,0.6110161543,-0.0891980231,-0.1973164827,-0.3674799502,-0.0013952475,-0.129200235,0.268769145,0.009201522,-0.1414214671,-0.0075133163,0.4450765252,-0.3248300254,-0.1765654087,0.0908282474,0.2006067932,0.0376323983,0.0878013074,-0.1453626454,-0.1393622309,-0.4348025918,-0.0208223853,0.0282157101,-0.189934954,0.2025924325,-0.3461400867,-0.0354302078,-0.1563955843,0.0745557025,0.2588439286,0.2893253267,-0.0501379408,0.1223113686,0.1274967492,0.450963527,0.24933137,0.3772580326,-0.3201477528,-0.4753778875,0.1778146923,-0.0470593907,0.4743435681,0.2156001627,0.034156289,0.2161576748,-0.0185944289,0.3497695923,-0.3186557293,0.1515969634,0.0757260248,0.3175280392,-0.0986533314,-0.0241907947,0.3491724133,0.1370459944,-0.139048785,0.2540732026,0.4911814332,-0.3232056201,0.0032343701,-0.0299674049,0.8855907321,0.1056121439,0.2487405837,0.1936124563,-0.1301231682,0.7416334748,-0.1253148615,-0.0057813255,-0.2042514384,-0.2316677421,0.0227827784,0.122446917,-0.0359402411,-0.1028271839,-0.1277359575,0.1617044806,0.1966115385,0.2098567486,-0.3110846877,0.1352752596,0.1470895708,-0.4514621794,-0.4612251818,0.2248733491,-0.1161238998,0.0211705957,-0.1824131608,-0.1019477919,0.124123551,-0.072244592,-0.019357888,-0.0483580492,0.0004145927,-0.0297621451,0.2275688946,0.0535414852,-0.0901788399,0.0702582374,0.5872550607,0.1994176358,-0.2770316899,0.1689130515,-0.2158319652,0.2534058392,-0.2243984491,-0.1491058171,0.1442285925,-0.1082656458,0.0214667711,-0.2319771349,0.1533290148,-0.3724336922,-0.1074941978,-0.0835541561,-0.1208783239,-0.2034196556,0.2934625149,-0.0886076167,0.0162199344,-0.359207809,0.2496371567,-0.2317857444,-0.0014839346,-0.0571932606,0.5067799687,-0.2434356511,-0.1012933925,0.1239106432,0.0796422884,-0.0826620981,0.1734454036,0.1539868116,-0.1471027583,-0.3617188334,-0.1794024706,0.34062621,-0.1203764454,0.3272584081,0.0687457398,-0.4154073298,0.0429234952,0.2170188427,0.1591478139,-0.1347607374,-0.0347159281,-0.0699220896,-0.3016020954,0.4503432512,-0.2702894211,0.4737462103,0.254745692,0.0541086867,-0.1629879624,0.0234031845,-0.4845525026,-0.1527335942,-0.303902328,0.0911612958,0.300135076,-0.0241648238,0.0992228836,-0.0303565022,0.1799135506,0.1419876665,-0.1217101738,-0.3179635704,-0.232382223,0.0683443844,-0.1001814976,0.1548478156,0.1733206362,0.0577317849,-0.0607292727,-0.2522354722,0.2309528887,0.0776543096,0.001853809,0.0294039678,0.0922851637,-0.0230657216,0.1448910832,0.2576679587,0.0151514728,0.0999400988,-0.208949253,-0.0915964022,-0.2376726568,-0.1425906271,0.4495269358,0.2095391601,0.3572067916,0.0453723483,0.0566358082,0.1708986163,-0.2864087224,0.4544510543,0.3953225911,-0.0079190414,-0.0786270052,-0.0443958342,0.1653842479,0.3480187058,-0.3592630327,-0.2240286171,0.2418738902,-0.2962069511,0.3216108382,0.2612406015,0.1631905735,0.2102433592,0.2711853683,0.0929905474,0.1685416996,-0.1193306521,0.2019393444,0.029899342,-0.1064256057,-0.1050612032,0.3100690246,0.1835039258,0.3039821982,0.3560727239,-0.1540013403,0.4656894803,-0.1631238461,0.0865291879,0.1899957061,-0.0961969644,0.1543879509,-0.010904843,0.0776908994,0.1670867652,0.1116076931,0.2875139117,-0.1223222464,-0.0762883201,-0.1443474144,0.3021418452,-0.2169876695,-0.1225275621,-0.1053380817,-0.1828467995,-0.067014806,-0.0610575229,0.0619340725,0.0217973273,0.4482190311,-0.0210211314,0.2921095788,-0.2495602071,0.0262455847,0.1570696384,0.2387662679,-0.2262055278,0.1011970416,0.2060342133,-0.1575079262,0.2370788753,0.3511713743,0.285367012,0.0231998861,-0.1266076416,-0.0413745642,-0.0943210274,-0.0178203415,-0.5750265121,-0.0273738205,0.0038964308,0.069557488,0.2204520553,0.2723742127,-0.2449919879,-0.0317997336,-0.2435388863,-0.1067462564,-0.2998780608,-0.0084183142,0.0174773633,-0.0818328187,-0.3114442825,-0.0309391953,-0.3184853494,-0.2450156063,0.3068661988,-0.2121073157,0.0368838198,-0.1902482063,0.1602801681,0.1622106433,0.7105545402,0.1880345792,-0.0317467898,-0.0500356443,-0.2814380825,-0.558380127,0.4863836169,-0.1233729795,0.069958128,-0.0457054377,-0.2542594671,0.1808734387,0.16514045,0.1295319647,-0.0287137963,0.0479160473,-0.5361844301,-0.6508665681,-0.0382979251,0.1014942899,-0.4157022834,0.31544137,-0.1538863927,0.1669650823,0.2104787678,0.1281074733,-0.1026253998,0.3133718967,-0.0284486637,0.1136712283,0.1749945879,0.0963793322,0.2970145047,-0.3960744441,-0.2823928297,-0.0043029538,-0.3137663901,-0.3562425971,-0.0081997691,-0.0123135354,0.1174056754,-0.2758523524,-0.2045809925,-0.2278908491,-0.0091855247,0.1416176558,0.2925693393,-0.2401115745,0.1742570251,-0.0232698023,0.1487606913,-0.0618378818,0.197418198,0.0688479468,0.0721602142,-0.1814211607,-0.4055908024,0.4400929511,0.0274842754,-0.2121569365,-0.0254108664,0.1884889007,0.0754702464,-0.2529133558,-0.4663053155,0.0657795295,0.0391491093,0.1063443497,0.0582093522,0.2895167172,-0.1638449132,-0.0274682865,-0.0430258214,0.2346630543,0.3723336458,-0.2161186635,0.1957447678,-0.1520841569]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1590","title":"Add helper to resolve namespace collision","comments":"It also wasn't initially obvious to me that the samples which contain `import datasets` were in fact importing a huggingface library (in fact all the huggingface imports are very generic - transformers, tokenizers, datasets...)","body":"Many projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. ","comment_length":34,"text":"Add helper to resolve namespace collision\nMany projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict. \nIt also wasn't initially obvious to me that the samples which contain `import datasets` were in fact importing a huggingface library (in fact all the huggingface imports are very generic - transformers, tokenizers, datasets...)","embeddings":[-0.1716234088,-0.054331854,-0.1060904935,0.1661169678,0.3200764358,-0.2631010115,0.2233148962,0.2214922905,0.1152144894,0.3468360901,-0.2177070975,0.0769941732,-0.2270479798,0.2291831821,0.0510888733,-0.096799396,0.1414095759,0.1882733256,-0.3230443299,0.0956738368,-0.2504667938,0.367636472,-0.1152618155,0.0489351638,-0.2631996274,0.1357953101,0.1781104803,0.0242974274,-0.1031204388,-0.3912395537,0.4888105392,0.2592075467,0.0982602835,0.5347704887,-0.0001008511,0.0321501456,0.3329060674,-0.0276504848,0.0007692393,-0.047589425,0.0364767201,-0.1708824188,0.2306132913,-0.0239406116,-0.1686529517,-0.2056214064,-0.1198736578,-0.2970339954,0.2763868868,0.0841663331,0.3598596156,0.4617402256,0.1313858181,-0.0768034458,0.3627577722,0.0963326097,-0.2227722704,0.057724949,0.3163791895,-0.1041614413,0.0468180887,0.3070700169,-0.1991894394,-0.0762047842,0.4549307525,0.0259640291,-0.3586925566,-0.3287334144,-0.1528097689,0.1908170134,0.2731131315,-0.2819491923,-0.1536628753,-0.4264019728,-0.1069417149,-0.0959383398,0.3473276496,-0.3036998808,-0.0107707921,0.1086404622,-0.1284351051,-0.1054658592,-0.2067374438,0.0488111526,0.1202847734,0.182044819,-0.2933821082,0.0484388955,0.3704205453,-0.1188665405,-0.1089816317,0.0120951794,0.0670721531,-0.1090289876,-0.3166467845,-0.1150623709,0.0921506062,0.237753734,0.1597691774,0.1767258346,-0.2038028091,0.2182209194,0.0010576987,0.0623374693,-0.000859837,0.0542277135,-0.0342781059,-0.198603794,0.5527970791,0.2132434249,0.0499669015,0.0854636878,-0.0799074247,-0.2244858742,-0.1833886206,-0.2480139881,0.2999303341,-0.129007861,-0.0830086395,-0.0633308589,0.2499461174,0.1599763185,0.167051509,0.2875520289,-0.1070812419,0.0924921855,0.1625744551,-0.0009797555,-0.1402504891,-0.2982375622,-0.1280475706,-0.0877862871,-0.2294624746,-0.1101402715,0.0930651426,0.001677885,0.1425856054,-0.0730218738,0.3274430335,0.1099218428,0.1255504787,0.387673676,0.1581718028,0.2085342705,-0.1047998071,0.1603885442,0.0640211031,-0.1535737664,-0.2043193579,-0.1815986931,-0.3285942674,-0.1244390756,-0.2427878231,0.2783993185,-0.0005783685,0.0742766187,-0.1776985228,0.0787390396,-0.2354187071,-0.3015482426,-0.0025794331,0.089835383,-0.1285718232,-0.1933573037,0.2595249116,0.1862837672,-0.0397231281,-0.15157184,-0.2312023938,-0.2090185732,-0.0258405879,0.1009816453,-0.0768338814,-0.0912891179,-0.4290794134,0.0888910294,0.0363239795,-0.341076076,-0.3444636464,-0.0266798232,-0.122072041,0.0966459885,0.1180794463,0.2478375137,-0.0789285526,-0.0897482783,0.0506904311,0.3743130863,-0.0806017816,0.1113695726,-0.0837374702,-0.1473541558,0.0619588085,0.1496371627,0.0322673842,-0.0506884307,0.0818423778,0.0967354923,0.0244633369,-0.2688745856,0.051268708,0.3549134433,0.2928172648,0.1757862419,0.0566872731,-0.2884209454,-0.5094377995,0.0490486473,-0.1337141693,0.1777652055,0.0579022653,-0.2778173685,-0.1193551496,-0.0210711509,-0.1724146456,-0.132930249,0.2425019145,0.0880356133,0.2063600868,-0.1213769615,-0.1532645673,0.3962278962,-0.0709776133,0.2890957594,-0.5446171761,0.136016503,-0.0133889224,-0.0801835358,0.2109015733,0.3362109363,-0.0737625286,-0.1314787269,0.012851377,0.4666207135,0.0388001427,0.2169883698,-0.0610328466,0.3250129819,0.0382029973,-0.0809074268,-0.1348622888,-0.0163693149,-0.1158625707,0.1263958216,0.0752323493,0.323456794,-0.1517747045,0.1348322332,0.1206180081,0.0085516339,0.0611655228,-0.1266064495,-0.2937129438,-0.2063525468,0.1784841269,-0.0399721339,0.5010519624,0.2095853388,-0.1630304307,-0.0904012397,0.3006502092,0.1256808788,0.1690942347,0.1850461513,-0.3328076303,0.1235525161,0.0621022321,0.3123052716,0.1769826561,0.3749838471,-0.1040768772,0.1340055466,-0.0802904889,-0.194326207,0.1071156189,0.0810026154,-0.0310398173,-0.0732616782,0.0953732356,-0.0209314171,-0.4374803007,0.004406306,-0.3678159714,-0.0123512559,-0.3567407131,0.0487865619,-0.3243841231,-0.4119811952,-0.2333052158,-0.0788863972,-0.4529622197,-0.2149646431,0.2001776993,0.1844459027,-0.3822817504,0.3398858905,0.2563820481,0.3452131152,-0.1577738971,0.0183630418,-0.096030958,-0.0979340076,-0.2653641701,0.1727079153,0.2459769547,0.0696047544,0.4489150047,-0.1880033165,0.2625869811,-0.2494304478,-0.7585305572,0.0928101167,-0.01130387,0.2021751255,0.3293008804,0.0585579537,0.2586821616,-0.3670679033,0.0069151097,-0.3343408108,-0.1845338941,0.0125687122,-0.1436916888,-0.162294507,-0.1993564963,-0.1186818257,-0.0368571021,-0.5511009693,0.2229225039,-0.203947112,-0.1150979176,0.3167098463,-0.0235494208,0.2907909155,-0.4422486126,0.1673290431,-0.0310630128,-0.40988639,0.2175301015,-0.2191780061,-0.2563723624,-0.0489579104,0.0027548892,0.2813417912,-0.1618327051,-0.238123253,-0.2436077297,-0.1488197148,0.1200854182,0.3816463351,0.1798692048,0.2096943408,0.1286325753,-0.168921262,-0.2520650923,-0.1157061458,0.0651172996,-0.202627629,0.0197580419,0.0236343928,-0.0603592545,-0.2330673635,0.204626888,0.1164685562,-0.1896945834,0.3827630877,0.080948256,0.6727414727,-0.1006569639,-0.2928420603,0.0680517405,-0.0194776971,0.1335468888,0.3965210021,0.0475236624,0.2117363811,-0.2093107104,-0.0853887275,-0.0438354909,-0.3954039216,-0.1390277743,0.1724425703,0.0150353927,-0.0690044016,-0.4618407488,-0.1063507199,-0.0794302151,0.3416058719,0.2545205057,-0.0045329034,0.0160307717,-0.6660512686,0.085122861,-0.2873129249,0.2038649172,-0.0358028077,-0.0691054836,-0.0643516257,-0.1157974005,-0.1800161898,0.2391929626,0.6643900871,-0.0601360835,-0.3686954081,-0.2361992449,-0.0007079775,-0.306897819,0.0636752918,-0.0611853525,-0.1025839373,0.1228528768,0.4552626014,-0.2801752388,-0.1845309138,0.1217532083,0.292319864,0.1544237435,0.0998882353,-0.2492325753,-0.1559447944,-0.4418144226,-0.1102470607,0.0781989396,-0.138884902,0.0660432354,-0.2089808732,-0.0043862974,-0.1665541381,0.2741232514,0.2290883809,0.3621069193,-0.0192536134,0.1066801101,0.2744940519,0.3847666681,0.1978495717,0.3844273984,-0.4340292811,-0.6006050706,0.2385069132,0.0832810476,0.4434199631,0.2322692722,-0.0080619669,0.2868749201,-0.027306458,0.3315168619,-0.3384514749,0.2471526712,0.1939098984,0.2923811078,-0.1095429659,-0.1528539658,0.3241997659,0.2008042634,-0.0320927873,0.3843852282,0.6873418093,-0.2957417369,0.073272422,-0.1591493636,0.8891833425,0.1530972421,0.1167262197,0.2754866481,-0.0570452921,0.9147965908,-0.1653279215,-0.0125802215,-0.1258677095,-0.113602668,0.0887031257,0.0125505952,0.0921703279,-0.2225005329,-0.2066338658,0.1720073074,0.1576691568,0.2362180054,-0.2536675632,0.2309844196,0.2133428156,-0.4571326673,-0.2572706938,0.211906895,-0.1044952571,0.2102664113,-0.1652732044,-0.0287872478,0.040350683,-0.1484420598,-0.0739227384,-0.0127679547,0.053322129,0.0021357972,0.4662960172,0.1825069338,0.1402405351,0.3002677262,0.662923336,0.1969093084,-0.3239462078,0.1812441349,-0.2685647607,0.1098470241,-0.1346364468,-0.0888000503,0.2029492259,-0.063436918,-0.0337746069,-0.128730908,0.0737041235,-0.2807866335,-0.189218834,-0.0564695895,-0.0368611105,-0.2680021226,0.0537569895,-0.0889052823,0.165779382,-0.2806427777,0.2254053503,-0.2069068551,-0.0022463498,0.0168894827,0.3366632164,-0.2985759377,-0.1466330737,0.1754275262,0.0110222716,-0.277926445,0.1927544624,0.2056207657,-0.1907159537,-0.1894997507,-0.114094913,0.4079243243,-0.2680709362,0.2316064686,0.0067962334,-0.3190090358,0.1115989238,0.1799429953,0.1409953237,-0.261115551,0.0276526213,-0.0929687247,-0.3116619289,0.4535996318,-0.2223198116,0.4007025361,0.1895668209,0.0251828898,-0.1024947017,0.086461477,-0.4182649851,-0.0864680484,-0.2789540887,0.0182401445,0.3211550117,-0.0419289656,0.0900927112,-0.0578940511,0.1070772409,0.1915733218,-0.1554874182,-0.2392191291,-0.3062343299,0.0637635365,-0.0972100794,-0.0153373638,0.2928591073,0.0228652358,-0.0140390629,-0.254178226,0.1784348935,0.0603085831,-0.0205013007,0.0031806584,0.1131246537,-0.0742533132,0.1106413305,0.1956300735,0.0138042783,0.1485376805,0.0268243924,-0.107699886,-0.204392463,-0.1640634686,0.2915218472,0.3105290532,0.3763986528,-0.0743134618,0.0403737538,0.0356446207,-0.1852066815,0.3360448778,0.3035385013,0.076306425,-0.1580760628,-0.009023793,0.2509440184,0.2672972977,-0.2801643312,-0.314005971,0.1554099023,-0.1322545856,0.1599821895,0.1995127052,0.2533139586,0.2555080354,0.3761181235,0.1835836172,0.371293366,0.1086527109,0.385723412,0.0714859962,-0.1753883809,-0.0757980421,0.3299860656,0.0942314267,0.2346596718,0.4205245078,-0.2518236935,0.4789220393,-0.3529281616,0.0927990377,0.2049230337,-0.2354155332,0.1524898708,-0.0175274573,0.0026283988,0.1835927963,0.0110215051,0.3121464252,-0.239510715,-0.0141938748,-0.106003955,0.3352072537,-0.1968841255,-0.1233881712,0.0529490039,-0.2466178089,-0.1814153045,0.0332600698,0.0691587627,-0.0349599272,0.3456926942,-0.0826918706,0.1044543013,-0.339196682,-0.048967652,0.0682276487,0.1889819354,-0.1521361768,0.1890357733,0.1043093055,-0.1827581078,0.2095462531,0.2704191208,0.4066824913,0.1755354255,-0.0343752205,-0.0935190618,-0.1054311991,-0.0592624657,-0.3675794601,0.0433288515,-0.058506798,0.1019843519,0.0710639358,0.2339399457,-0.2674208283,-0.2310080379,-0.2215320319,0.0321755186,-0.4046780467,0.1092642099,-0.2300724834,0.0373568423,-0.3563767374,0.0359813869,-0.4958161116,-0.0656956956,0.3878440559,-0.0793619901,-0.0656025931,-0.3321765363,0.1231692955,0.1516615748,0.6866532564,0.2619680464,0.119071655,-0.1069376543,-0.5023650527,-0.7254452705,0.3800845742,-0.0637717247,0.1848857403,-0.0030615816,-0.3144658804,0.1168701947,0.1347701102,0.1403687894,-0.0114432462,0.0391351506,-0.5646096468,-0.5288473368,0.0057916231,-0.0229796488,-0.3092227876,0.2121955156,-0.1331285238,0.1813654006,0.120687373,0.0685025454,-0.2805921137,0.2687823176,0.0800034627,0.1308311671,0.2709142268,0.1827312708,0.3108758032,-0.235372439,-0.2521847785,-0.1005694643,-0.2863892615,-0.273715198,0.0415699594,-0.0826508328,0.2577743232,-0.3124459088,-0.3013809025,-0.3488104343,0.0015416591,0.2928626835,0.2274935246,-0.4342717528,0.1474634558,-0.054169327,0.2132262439,0.0837321058,0.2255559415,0.0721460953,0.1237296015,-0.2248812169,-0.2834953666,0.4166070521,-0.1089392304,-0.1908558756,-0.1489652991,0.1179046631,-0.0791965723,-0.3435354233,-0.5919113755,0.0233605579,0.0869260281,0.129053399,-0.0486894287,0.3910192549,-0.0909191594,0.0499208048,-0.0145056611,0.2419301271,0.4145051539,-0.1119269431,0.3631415665,-0.0466982014]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1585","title":"FileNotFoundError for `amazon_polarity`","comments":"Hi @phtephanx , the `amazon_polarity` dataset has not been released yet. It will be available in the coming soon v2of `datasets` :) \r\n\r\nYou can still access it now if you want, but you will need to install datasets via the master branch:\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","body":"Version: `datasets==v1.1.3`\r\n\r\n### Reproduction\r\n```python\r\nfrom datasets import load_dataset\r\ndata = load_dataset(\"amazon_polarity\")\r\n```\r\ncrashes with\r\n```bash\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/amazon_polarity\/amazon_polarity.py\r\n```\r\nand \r\n```bash\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/amazon_polarity\/amazon_polarity.py\r\n```\r\nand\r\n```bash\r\nFileNotFoundError: Couldn't find file locally at amazon_polarity\/amazon_polarity.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/amazon_polarity\/amazon_polarity.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/amazon_polarity\/amazon_polarity.py\r\n```","comment_length":45,"text":"FileNotFoundError for `amazon_polarity`\nVersion: `datasets==v1.1.3`\r\n\r\n### Reproduction\r\n```python\r\nfrom datasets import load_dataset\r\ndata = load_dataset(\"amazon_polarity\")\r\n```\r\ncrashes with\r\n```bash\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/amazon_polarity\/amazon_polarity.py\r\n```\r\nand \r\n```bash\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/amazon_polarity\/amazon_polarity.py\r\n```\r\nand\r\n```bash\r\nFileNotFoundError: Couldn't find file locally at amazon_polarity\/amazon_polarity.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/amazon_polarity\/amazon_polarity.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/amazon_polarity\/amazon_polarity.py\r\n```\nHi @phtephanx , the `amazon_polarity` dataset has not been released yet. It will be available in the coming soon v2of `datasets` :) \r\n\r\nYou can still access it now if you want, but you will need to install datasets via the master branch:\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","embeddings":[-0.3479880095,-0.4079693556,-0.1594295353,0.1468008608,0.3409609795,0.1245037466,0.1368293166,0.0708829165,-0.0147829289,0.2100149393,-0.0304772444,-0.0856967419,-0.2752964795,0.0079954294,-0.083372511,-0.1797067821,0.0836886317,0.1474182904,-0.1062064916,0.144836992,-0.1857416183,0.3676777184,-0.2011100203,-0.0363301374,-0.1535672098,0.0321417078,-0.1244175658,0.2561988235,-0.3571479619,-0.1719460189,0.6051847339,-0.1438151598,0.116646491,0.4674164355,-0.0001094239,0.121839717,0.2970665395,0.1170058623,-0.2840379775,-0.6816424727,0.1893191934,-0.1137111932,0.1002502143,-0.0300899651,-0.1724277139,-0.4273685515,0.3060815334,0.06626679,0.4094986022,0.0599564649,0.2611253858,0.4023949802,0.4682300389,-0.1297983825,0.4718745947,0.3050636947,-0.1289897412,0.1152159944,0.0195885096,-0.0436794125,0.2055599838,0.3719202876,0.0113057951,-0.0789840817,0.2118726075,-0.0042474139,0.1750297099,-0.1100075245,0.0232831724,0.2148030698,0.2996555865,-0.3209213614,-0.5227025151,-0.0029204008,0.011715604,-0.3397398591,0.1643906385,0.1410713345,-0.0013033786,-0.1067062169,0.0194250494,-0.2168499976,-0.3099757731,0.394634515,-0.074210383,0.5269249678,-0.3228904307,0.0313689671,0.1816498935,-0.4286199808,-0.3890161514,0.1138272062,-0.1071820781,0.1441517919,-0.2899423242,-0.075633727,-0.1392741054,0.0748889819,0.1399799585,0.2753765583,-0.074021332,-0.0055774814,-0.0318000428,0.170882374,0.077100262,0.034131851,0.1519277245,0.3041283786,0.1158187687,0.3378837705,0.1673315763,-0.0622062683,-0.0027534203,-0.3101986945,-0.0584132709,-0.1008433327,0.4976038337,-0.1935555488,-0.4300061166,-0.0161387399,-0.2263162285,0.1012486294,0.1638518721,0.4274002314,0.0445426889,-0.028920725,0.1809092015,0.1616847813,-0.2305804193,0.2029237449,-0.2381457984,0.2560457885,-0.1458518952,-0.0051119602,0.1202393174,-0.1230172813,0.4229930043,-0.1286453754,-0.0903867185,0.0102752652,0.1358660012,0.0473022275,0.0937853083,0.342635721,-0.1503624171,-0.075939998,0.1274428368,0.0928332508,-0.1470044255,-0.1799183488,0.0203040726,-0.0959174559,-0.2991919518,0.207498163,-0.2199320644,-0.1230285391,0.0789890736,-0.024935998,-0.1744746268,-0.3944456577,0.0188542735,-0.0501719341,-0.153336212,-0.1192642674,0.2974745631,0.6852934361,-0.2487232983,-0.2089466453,-0.3055853546,-0.2532973588,0.1533036679,0.0427145138,-0.0661377087,-0.131085366,-0.3379172683,-0.2395652235,0.1273248941,-0.34221071,-0.219497785,0.1426003575,-0.39196527,0.1090528369,0.2609383166,0.1072728559,-0.4844020605,-0.0023038345,0.072755076,-0.2347019166,0.1089569405,0.0022045132,-0.0262717772,-0.3088288307,-0.0118465014,0.1687119305,-0.010352876,-0.1097878367,0.239464134,0.1723675132,0.1868828386,-0.015090609,-0.1329808533,0.4036785662,0.430098176,0.1812001914,-0.043866355,-0.4173109531,-0.5634769797,0.0696375445,-0.1297064722,-0.0143243913,-0.1950017214,-0.1236344129,-0.1820705384,-0.1635348201,-0.0530653857,-0.1849985868,0.1603649855,0.1631918401,0.3020811379,0.0290639997,-0.1731580943,0.1303827018,-0.2044846565,0.2718700469,-0.5783797503,0.0628575236,-0.1096294373,-0.0123150293,0.148976624,0.1735637784,0.0894442797,-0.2606093287,0.2154633403,0.4597685933,-0.2463010401,0.4114013314,0.1603762209,0.174153924,0.2682320476,-0.0960373655,-0.0961564556,-0.0080652973,0.0273449682,0.1278957278,-0.0698075444,0.385148108,-0.1351651549,0.1483459175,0.1445954889,0.0045578103,0.1259079725,0.2072737366,-0.2886961699,-0.3195874393,0.3829951286,-0.0956356525,0.2078200877,-0.2757167518,-0.3405966461,-0.0480875447,0.1322781593,-0.0102751087,0.164276734,-0.099009566,-0.2142263502,0.1930595189,0.1184575707,0.4621918201,0.4799415469,0.3050818443,-0.0855378881,0.0466704704,-0.1272161007,-0.160594821,0.4043017924,0.2406167239,0.0038172002,0.0504072271,-0.1336296946,-0.0189378634,-0.2433249801,-0.3935017288,-0.2533627152,0.1306104809,-0.2505778074,-0.0321471058,-0.4261734784,-0.2258952856,-0.3069345355,0.1278160065,-0.1757679582,-0.1779222339,-0.1088575646,0.2513082922,-0.1686687618,0.0835060179,-0.0559361838,0.0933208317,-0.0435148589,-0.3251518011,-0.2229113281,-0.1641744226,-0.0613541678,0.1425081193,0.2137884796,-0.1246764585,0.4876343012,-0.2830376029,0.0688867271,-0.3899007142,-0.4924218059,0.0189632233,-0.2072217911,0.0949290171,0.4582244754,0.1032193527,-0.2475157976,-0.0739440322,0.1684570312,-0.1575738043,-0.2392317504,-0.0062979213,-0.0697318763,-0.1733554006,-0.0253236219,-0.402853936,-0.1508603245,-0.337118268,0.2300270051,0.148672834,-0.0006279053,0.1806941628,0.1901656538,0.3676950336,-0.3491116762,0.2490089387,0.0143668177,-0.2989014089,0.2959597111,-0.2435079366,-0.2643984556,0.2790344954,0.1617137939,0.0167356525,0.0499413833,-0.2836008668,-0.06305857,-0.12184567,-0.1174788252,0.0111594694,0.3160513639,0.3380795121,0.0302935038,-0.100764282,0.0467141345,-0.1240636781,-0.1301065385,-0.2254658639,0.3550337851,-0.2552389801,0.076704964,-0.2182062417,0.6340251565,0.0623792075,-0.1936589777,0.0910418853,-0.1483147144,0.4900240004,-0.1918846816,-0.2894430459,-0.2767001987,0.257278949,0.3088666201,0.1075014174,0.0433204509,0.2568730116,-0.3664234281,-0.0917699039,-0.3213224113,-0.2131362408,-0.3383034766,0.1778076291,0.1570366025,-0.0186243597,-0.0073519871,0.0493267812,-0.1662915796,-0.0711467192,0.4121952951,-0.1035990417,0.0997330323,-0.0668363869,-0.3051210344,-0.1988105178,0.4756160975,0.030413365,0.0509602912,-0.0937960669,-0.1918788552,0.0009239058,-0.1081313789,0.4818147123,0.0903185979,-0.0801216736,0.0228532832,-0.1577991396,-0.4005216658,-0.0274782088,-0.2612534463,0.4249591827,0.0342727117,0.5901094079,-0.3759629726,-0.2228934318,0.2796900272,0.1097374037,-0.0025948372,-0.0770180225,-0.2817257941,0.0272907689,-0.0402558967,0.0364875756,-0.0845616981,0.2202616185,-0.175041154,0.1786637157,0.0618274026,-0.1259283721,0.0754038095,0.0294994563,0.3922421634,-0.2385631204,0.0622898489,0.250308454,-0.0664571673,0.2658132315,0.6059344411,0.2176028043,-0.277572602,0.1582262963,0.1005288363,-0.1520284265,0.3247642815,-0.0392042324,0.0130165871,0.3081917465,0.1735008508,-0.1358238012,-0.3210608959,0.393696934,0.1134105027,0.0918206051,-0.0360333771,0.2441853583,0.0196916889,0.1410986781,0.321239084,0.5112134218,-0.1586908102,0.2731305659,0.0461124592,0.8052716255,0.0167238321,0.262868315,0.1304103881,-0.0347270891,0.9862422347,-0.1025052294,0.1490567774,-0.3020689487,0.19349204,0.1081602499,-0.0407787114,0.4197332859,-0.2651719451,-0.1427745521,0.3299347758,0.2675995827,0.0835018232,-0.2166481763,-0.0107500525,-0.4383032024,0.0878242627,-0.1916678399,0.1452476382,0.079135403,0.4863111377,-0.2161916047,0.0072140102,-0.1626043618,-0.3805135787,-0.3852407634,-0.0912893713,-0.1410883218,0.2592483759,0.1741279662,-0.0907114446,-0.0190740954,0.3718284667,0.2497218549,0.0609979481,-0.396470964,0.0922153965,-0.2150311619,-0.3787040114,-0.1657681614,0.1121566743,0.2095563412,-0.0374903232,-0.2007800043,-0.035363812,0.047026068,-0.0080671879,-0.3303602934,-0.0268963184,-0.2138495892,0.0041921628,-0.1896992773,-0.0082467822,-0.2543064654,-0.1578424573,0.177546382,0.1121898964,-0.2231851369,-0.2031597644,0.1021467298,-0.0587706603,-0.1308757514,0.5322408676,-0.1828396171,0.0836191028,0.4300751388,0.2786898613,-0.0971618369,-0.3899719417,0.1866092533,0.2124700397,-0.4469741881,0.0227388889,0.1510967016,0.191763103,0.1680829078,0.1403881758,0.2693332732,-0.0948630869,0.0875577256,-0.3882385194,-0.0769423172,0.3205413818,0.0332594588,0.1899958849,0.1736549288,-0.2284808457,0.0975497812,-0.2836911678,-0.3788015842,0.0786603093,-0.2254058123,-0.0939114913,0.0192459505,-0.0073985299,0.5711520314,0.1018613279,0.2320117056,0.2181948572,-0.3243469298,-0.2544843554,-0.1084256098,0.0945680514,-0.0346432664,-0.0125215165,-0.0249573048,-0.0222223569,0.0687841326,0.1524954438,0.1734759808,0.3341299295,0.1582510918,0.1446084231,-0.1194521263,0.1690199822,0.0117783919,0.0195734166,-0.071988903,0.1289560497,0.1779900938,0.0775710493,0.0067379321,-0.1191647723,0.0373140424,-0.0810109153,-0.188047722,0.1136232316,0.3063569665,-0.0758865699,-0.2214574814,-0.1626741737,0.4580843151,0.3845559359,-0.2183555067,0.1248593032,0.2854144871,0.2142765373,-0.2931698263,0.0103997122,0.0190629307,0.1760827601,0.1890486032,0.1412040442,-0.2063532621,-0.0167129748,0.2385961562,0.3029824197,0.3455760479,-0.3921842575,0.2276841551,0.1078593209,0.0631872565,0.0041161785,0.0418097265,0.2804657817,0.2554586232,0.1395144761,-0.253901273,0.2305685729,-0.417719245,0.2897433043,0.0536154471,-0.298011899,0.1886996329,0.1838290244,-0.1125131473,0.0073551307,-0.199933365,0.7434198856,-0.3424648046,0.293061316,-0.2554724514,0.0252120234,-0.1385477185,-0.1526787579,-0.2909658253,-0.2425522804,-0.0798388347,-0.2007221878,0.1263447553,-0.2806013227,-0.0877748132,0.0587936938,-0.0349412449,-0.3793413043,0.0086649647,0.0944180638,0.2072147131,-0.1528529078,0.0437530316,0.4355440438,-0.0391812064,0.0630143881,0.3242606223,0.431612283,0.0738174245,-0.0038447289,0.154587999,-0.0941189975,-0.2815183997,0.101015456,0.1761722863,-0.0095712906,0.0294628218,0.3469564915,0.2939645052,-0.2292863876,0.3707342148,0.1491947919,0.2623641789,-0.1603526473,0.2588389814,-0.567165792,-0.0026885467,-0.2175392061,0.1489082575,-0.7439574599,-0.0749540403,-0.0734461546,0.1557801813,-0.0134731187,-0.0751893222,0.1184415221,0.1225307584,0.3655386269,0.4004965127,0.1081300154,-0.0999230593,-0.0020383843,-0.3938156962,0.0723533854,-0.0834049657,0.1173952743,-0.2401403189,0.2559217513,-0.1104337052,0.0697369426,0.2139820904,-0.3945614994,0.0260401014,0.1659129858,-0.1430796087,-0.0883520022,0.2128640711,-0.1136935949,0.1055822894,-0.1475958079,-0.086824514,-0.1874763072,0.0144141093,-0.1302590221,0.0870231614,-0.1773097366,-0.3022311628,0.2704393864,-0.0357645936,0.0042448388,-0.1206798777,-0.2427915186,-0.4267435968,-0.1690506041,0.0312255584,0.0688577071,0.1093186364,0.6689342856,-0.2943127453,-0.6157664061,-0.2046460062,0.2288289219,-0.1594868302,-0.1637366712,-0.0575525612,0.1553613544,-0.0737826228,-0.0551081486,-0.1251988113,0.2218609899,0.0011763651,-0.0023418465,-0.2762530744,-0.2309858799,0.5606697202,-0.298710674,0.0917576849,0.1425621957,0.3285603821,0.5192729235,-0.1770681888,-0.4749755561,0.5117938519,0.1641480476,0.2177815437,0.1199834347,0.2224524915,-0.1886867732,0.0596453883,0.0232484341,0.198811695,0.0254378468,-0.0106476005,0.1978369355,-0.1795605123]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1581","title":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'","comments":"Thanks for reporting !\r\nYou can override the directory in which cache file are stored using for example\r\n```\r\nENV HF_HOME=\"\/root\/cache\/hf_cache_home\"\r\n```\r\n\r\nThis way both `transformers` and `datasets` will use this directory instead of the default `.cache`","body":"I am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n","comment_length":37,"text":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'\nI am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n\nThanks for reporting !\r\nYou can override the directory in which cache file are stored using for example\r\n```\r\nENV HF_HOME=\"\/root\/cache\/hf_cache_home\"\r\n```\r\n\r\nThis way both `transformers` and `datasets` will use this directory instead of the default `.cache`","embeddings":[-0.2601940632,0.1537577212,-0.087916173,0.1361187547,0.0533822998,0.0593192391,0.656006515,0.1960314065,0.161277622,-0.0136698019,-0.3051161468,-0.168390587,-0.0324441306,-0.5109063387,-0.0259655211,-0.2055929005,0.0551733933,-0.0676345825,-0.3101544976,-0.1253348142,-0.0913535804,-0.0168972444,-0.0075819497,0.1020792797,-0.3513800204,0.0235478356,0.333840996,-0.0979835615,0.0049274717,-0.1136808172,0.3865947425,0.1387621909,0.1959901601,0.8908079863,-0.0001138685,0.364698112,0.1204629838,0.0754132047,0.0866398886,-0.2086414546,-0.1387084424,-0.2477509379,-0.1201189086,-0.0817605555,0.0790471286,0.0090726782,0.0140740611,-0.1214024574,0.6044265628,0.3587148786,0.2027944475,0.4903935492,0.0908189341,0.123684071,0.2537278831,-0.0316032395,-0.2483440191,-0.0738279372,-0.1243278161,-0.2798888981,-0.0047014416,0.2760694027,-0.0540417656,0.1426951587,0.2170650065,-0.1940569878,0.1225119233,-0.3570998013,-0.0412276238,0.0453185365,0.5151111484,-0.065642722,-0.5028495193,-0.1864654869,-0.1189291701,0.0584088117,0.302632153,0.3046842217,-0.0361773148,0.2790159285,-0.2144559771,-0.1514513791,-0.4730093479,0.095054619,-0.2394710332,0.3513744175,0.0620445088,0.1947925389,-0.0911417976,-0.0715904012,0.6296569109,-0.0060653305,0.1474893838,0.2979938984,-0.3239389062,0.2316237688,0.0331782922,-0.5658233166,-0.3128330112,0.0865982324,-0.2314409018,-0.0924056843,-0.0704433471,0.3829046488,-0.2021469325,0.4120383263,-0.3363805413,0.7340712547,0.3856590092,-0.3047202826,0.0332622379,-0.0432148613,-0.0374102555,-0.0881388262,-0.1260698587,-0.0319224447,0.14577277,0.0939131603,-0.098433651,-0.034681011,0.086896874,-0.0100169806,0.1854218394,0.3815259933,0.1036055908,0.0991566405,0.3901361525,0.1196964458,-0.0152244633,0.0574717671,-0.2500608265,0.072026372,-0.0946181267,-0.0641494319,0.1987671703,0.0866814032,0.0298909303,-0.1037014425,0.2303165644,-0.1406204551,0.4395029545,-0.2077358812,0.1142411977,0.7253692746,0.1236809492,0.040582139,0.3252632618,-0.1411764622,0.0385678783,0.1115270332,-0.3354057968,-0.1929328144,0.0483054183,0.1385359615,-0.133828178,0.2708508074,-0.2460403144,-0.1343725771,0.3796409369,-0.0594122037,-0.0387408696,-0.2568915188,0.0189277288,-0.1345490813,-0.0994921178,0.2476488799,-0.1799386889,-0.1785385013,-0.0476372056,0.0107344426,-0.0975554064,0.4373539984,-0.1490976363,0.3612090945,-0.012258118,0.1215736493,0.0213417336,-0.5040411353,-0.4507939517,-0.0658406541,-0.0317620635,0.0552451983,0.0770501271,0.0080055622,0.3263442218,-0.2391067594,-0.2724581957,0.3224855959,0.2141270787,0.0887082145,0.22979559,-0.1883960515,-0.0167393964,0.0299266316,0.5327509046,0.2355925739,0.0445311628,0.4013054967,-0.1830300242,-0.342936188,0.1383527666,0.4372718632,0.3134897947,0.0098209782,0.0195712894,-0.0359986238,-0.3422741592,0.3768794537,-0.1361479759,0.1983713806,-0.0884066597,-0.001343661,-0.1373440027,0.3042966723,-0.0974787921,-0.0624615848,0.1574099958,0.0914994776,-0.0297624823,-0.0523535758,0.2131533921,0.4122335017,0.0303778984,0.1924372464,-0.013811538,0.4103766978,-0.2020863146,-0.0439700596,-0.0946270898,-0.0712999925,0.1517301798,-0.1395329535,-0.205512628,0.1627992392,-0.0285273809,0.1827243418,-0.5095719695,0.427361697,0.2366541922,-0.0472982004,0.0571585074,0.1110836491,-0.002362475,-0.0089515029,-0.0438567549,0.0467418507,0.0703111067,0.0665785149,-0.0663850605,0.1159160361,0.2043628991,0.0928915516,0.0580297783,-0.0351178497,0.0987340659,0.1005920395,-0.0032767237,-0.1262062788,0.3256442547,-0.2848223448,0.2149302065,0.1594939679,0.0941943005,0.0378742032,-0.0761301368,0.1714250296,0.3254772723,0.3470335901,0.2231839895,0.0314969905,0.0246453322,-0.0549993627,-0.0549255013,-0.0206400882,-0.1116272211,0.1299436986,0.1952213347,0.0732089356,0.0617158674,0.0350541733,-0.1838503778,-0.2796123028,0.0911719501,0.3946083188,-0.3553246856,0.1200247556,-0.039686121,0.0555186421,0.1301530749,0.0189663973,-0.4570333958,-0.2215031236,-0.2029660195,0.4210834503,-0.1194587797,0.1644146591,-0.0810745582,0.0509676039,0.2963262796,-0.1255412996,-0.259331286,0.104335174,0.0625970736,0.0425717719,0.1925008893,-0.5198504925,0.1900203079,0.1750378907,0.1842620224,-0.0083790794,-0.5007292032,0.0480755009,-0.1537050158,0.1523056179,-0.1455800682,0.2838713229,0.0492252409,-0.2909571528,-0.0210785959,-0.1372650564,-0.0680153519,-0.0453203991,-0.0619840845,-0.2530847788,-0.1954331398,-0.069822371,-0.2933194339,-0.4212827682,0.1225009188,0.371108681,0.2529848516,0.4407685995,0.1377109438,0.0921132565,0.2690692544,0.0994600207,-0.0383629762,-0.107421346,0.2451769412,-0.3104769289,-0.2563972175,0.0066135097,0.0111053875,0.2837287486,0.337452203,-0.3564890623,-0.1421988308,-0.312304467,-0.0041876035,-0.0811958015,0.1192051098,0.045397982,-0.0257927254,0.1307452321,-0.1638926119,0.0347888954,-0.0795593485,0.1042242572,-0.0728223324,0.0976093784,0.0997843817,0.2115115076,0.6869980693,-0.1291385889,0.1413394809,0.018308932,0.2151246369,0.4513024092,0.0434456803,-0.2967938781,0.0668857172,0.0837084129,-0.206221208,-0.1333539337,0.0458995514,-0.1497395933,0.0206685234,0.1178896278,-0.0471880026,-0.2259911448,0.1134366393,-0.0898844674,0.3340997994,0.187919274,0.0507023931,-0.4434088469,-0.1077162847,-0.0562812537,0.3804533184,0.0898376033,0.1508090645,-0.3076767027,-0.0904264003,-0.2749199867,0.0611348636,-0.0811378509,0.5575761199,-0.1852121502,0.3113373816,-0.0447095297,0.1809080094,0.4595344961,-0.2760755718,0.005393195,-0.1066814512,-0.281932354,-0.3433832824,-0.2390835285,-0.0859023929,-0.06393902,0.1136865318,0.1906843483,-0.0791468695,-0.0084775966,-0.1428744942,-0.2193469405,-0.1370996833,-0.2479740083,-0.1115662828,-0.2691839337,-0.4986168742,-0.1251070797,0.3895351589,-0.1092313603,-0.1225882396,0.1031751484,-0.2260031849,0.0671980307,0.1904061288,0.1915994138,0.0266411565,-0.009566904,-0.1999968439,0.2065167427,-0.5022637248,0.062692441,0.19732593,-0.1263116598,-0.3567327857,-0.016329661,0.1559303403,-0.1357991099,0.0637567714,-0.174356252,-0.0275263917,0.0254739188,-0.0504413955,-0.071854718,0.3104915917,-0.055890426,0.195660457,-0.1012471616,-0.4688993096,0.19685857,0.2237318605,-0.0757478699,0.1689159125,0.0943683311,-0.3989681005,0.4668301046,0.1583802253,0.5565386415,-0.359916836,0.0974431783,0.3196151853,0.067224808,0.6953794956,-0.6555452943,0.3161601126,-0.2079330236,-0.3527629972,-0.1493595541,-0.3817493916,0.181630671,0.0691904053,-0.0273417775,0.0941415504,-0.0852651075,0.3506408036,0.1797598153,0.2020633519,0.0884777084,-0.4448313415,0.1528819054,0.2294978499,-0.2720116377,0.3177681863,-0.1780319661,-0.0277537853,-0.2918524146,0.0254852902,-0.2747633159,0.1161500663,-0.2991321683,0.5190944672,0.1113141179,0.0464832298,0.2879340053,0.2445803732,0.2088550776,0.2620199323,-0.2269392312,0.0637965649,-0.130311653,-0.0931614488,-0.0392174013,0.0066591357,0.335981667,0.0596252605,-0.0585409142,-0.0665251166,-0.1696169376,-0.3609890938,0.0174033288,0.2294077426,0.2024174631,0.006458255,-0.0510129295,-0.0119281607,-0.3825114667,0.0726773962,0.1492640525,0.0109344935,-0.2534069121,0.2334099412,-0.045485761,-0.2722888291,0.0967666358,0.3703703582,0.1194974408,-0.1971344352,0.559756279,0.2043311149,-0.2898311615,-0.0391659811,-0.2040651292,0.0931312293,-0.4541436136,-0.0446451493,-0.4564749002,-0.4141295254,-0.0188431554,-0.4614747167,0.3459784389,-0.0074619078,-0.1736587584,-0.4079886973,-0.3153894246,-0.0964628011,-0.3273630738,-0.0001220054,-0.0069627934,-0.0182451922,-0.0710390657,0.1626549214,-0.259177804,-0.2372059524,-0.3444597125,0.006200335,0.3342401981,-0.2212559134,0.0436337963,-0.1810938865,0.0256804116,0.1466897726,-0.387232244,-0.1846114099,-0.2641800344,0.1107682213,0.2623438537,-0.3276488781,-0.0680264756,-0.3471541703,-0.0914151073,-0.1358476132,0.1620068699,0.26431036,-0.1182678416,0.1854373962,-0.1474619806,0.2718785405,-0.019502759,0.1773811579,-0.0268762168,0.0644539818,0.0384416021,0.3823345602,0.1326795816,0.0603220277,-0.6228336692,0.1132022813,0.0375844911,0.1517114937,0.3931129277,-0.3460859954,0.3375230134,-0.0362393744,0.0605528355,-0.152901575,-0.3420254588,-0.0691744462,0.1160460636,0.1642261893,0.0276592579,-0.096303314,0.0548701957,0.032719519,0.0987614095,0.2690493762,0.2619537115,-0.0406337827,-0.0124462042,0.3720317185,0.6341760755,-0.2661919594,0.2063339502,0.2476853728,-0.2334974557,0.1992752403,0.2232558876,-0.002031469,0.0001358897,0.1990372539,0.1524522156,-0.0540718548,0.0205246545,-0.1834712923,0.336091429,-0.2124366164,-0.1275319606,-0.0399635695,0.1181048453,0.2308101356,-0.0397053212,-0.3249217868,0.0261873621,-0.0104645165,-0.2011923939,0.1950682998,-0.1402630657,0.0443249755,0.0428522676,0.203814134,0.0132805081,-0.2877651155,0.1271378845,-0.4779744148,-0.3058706224,0.3640507758,-0.1340831816,0.0985853449,0.0616869591,0.2978540659,-0.2202402949,-0.196892485,0.0838604197,0.1072050706,-0.1116496697,-0.082573697,0.0934753716,0.5431833267,0.2893403471,0.1599655449,0.2631548941,0.0980713665,-0.0594070368,-0.0906574205,0.3076898456,-0.2290304899,0.1799558848,0.1197366491,0.0925129056,-0.2126800418,-0.2806037664,0.093236357,0.0901547596,-0.0003153438,0.2890575826,-0.2614881396,-0.2587867677,0.0748261362,-0.0476370901,-0.1973496825,0.1740469337,0.8346324563,-0.1961341202,0.2542154193,-0.0428735539,0.0120079843,-0.1633709967,0.385330081,0.3202114701,0.0207920764,-0.2639081478,0.2996233106,-0.2990526259,0.120006606,-0.1916530728,-0.0139936889,0.1818581522,0.1701564491,-0.0613947324,-0.0301435366,0.1060565785,-0.4536414444,-0.0028722938,0.0240874756,-0.2195233554,-0.1286708415,0.1216328964,0.0295880251,0.0657106042,-0.3943275809,0.0754953325,0.0724963695,0.0972357243,-0.3166572452,0.0166931972,0.2151388079,0.2526507676,0.2539834082,0.2079452425,0.3316455483,0.0703696534,0.0958230048,-0.3994336128,-0.349573344,-0.3445076644,0.0160421301,-0.1569499522,0.5965117812,-0.2879594266,-0.5061549544,-0.3620113134,0.0960142016,0.105117783,-0.2808544934,-0.2017903328,0.1834330857,-0.4025213122,0.3427912593,0.2535629272,-0.0980608538,0.244820863,0.3333668411,-0.3015104234,-0.3771386743,0.5819663405,-0.4455092549,-0.4940264225,0.2248333991,0.0352744348,-0.1100796983,-0.0267488882,-0.4583678842,0.2484734058,0.1376974732,-0.3479514122,-0.3160350621,0.0086388001,0.1702643633,0.2086642385,-0.0559566729,0.1456699669,-0.0756859928,-0.3460895717,0.118467927,-0.0319126099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1581","title":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'","comments":"Great, thanks. I didn't see documentation about than ENV variable, looks like an obvious solution. ","body":"I am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n","comment_length":15,"text":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'\nI am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n\nGreat, thanks. I didn't see documentation about than ENV variable, looks like an obvious solution. ","embeddings":[-0.2601940632,0.1537577212,-0.087916173,0.1361187547,0.0533822998,0.0593192391,0.656006515,0.1960314065,0.161277622,-0.0136698019,-0.3051161468,-0.168390587,-0.0324441306,-0.5109063387,-0.0259655211,-0.2055929005,0.0551733933,-0.0676345825,-0.3101544976,-0.1253348142,-0.0913535804,-0.0168972444,-0.0075819497,0.1020792797,-0.3513800204,0.0235478356,0.333840996,-0.0979835615,0.0049274717,-0.1136808172,0.3865947425,0.1387621909,0.1959901601,0.8908079863,-0.0001138685,0.364698112,0.1204629838,0.0754132047,0.0866398886,-0.2086414546,-0.1387084424,-0.2477509379,-0.1201189086,-0.0817605555,0.0790471286,0.0090726782,0.0140740611,-0.1214024574,0.6044265628,0.3587148786,0.2027944475,0.4903935492,0.0908189341,0.123684071,0.2537278831,-0.0316032395,-0.2483440191,-0.0738279372,-0.1243278161,-0.2798888981,-0.0047014416,0.2760694027,-0.0540417656,0.1426951587,0.2170650065,-0.1940569878,0.1225119233,-0.3570998013,-0.0412276238,0.0453185365,0.5151111484,-0.065642722,-0.5028495193,-0.1864654869,-0.1189291701,0.0584088117,0.302632153,0.3046842217,-0.0361773148,0.2790159285,-0.2144559771,-0.1514513791,-0.4730093479,0.095054619,-0.2394710332,0.3513744175,0.0620445088,0.1947925389,-0.0911417976,-0.0715904012,0.6296569109,-0.0060653305,0.1474893838,0.2979938984,-0.3239389062,0.2316237688,0.0331782922,-0.5658233166,-0.3128330112,0.0865982324,-0.2314409018,-0.0924056843,-0.0704433471,0.3829046488,-0.2021469325,0.4120383263,-0.3363805413,0.7340712547,0.3856590092,-0.3047202826,0.0332622379,-0.0432148613,-0.0374102555,-0.0881388262,-0.1260698587,-0.0319224447,0.14577277,0.0939131603,-0.098433651,-0.034681011,0.086896874,-0.0100169806,0.1854218394,0.3815259933,0.1036055908,0.0991566405,0.3901361525,0.1196964458,-0.0152244633,0.0574717671,-0.2500608265,0.072026372,-0.0946181267,-0.0641494319,0.1987671703,0.0866814032,0.0298909303,-0.1037014425,0.2303165644,-0.1406204551,0.4395029545,-0.2077358812,0.1142411977,0.7253692746,0.1236809492,0.040582139,0.3252632618,-0.1411764622,0.0385678783,0.1115270332,-0.3354057968,-0.1929328144,0.0483054183,0.1385359615,-0.133828178,0.2708508074,-0.2460403144,-0.1343725771,0.3796409369,-0.0594122037,-0.0387408696,-0.2568915188,0.0189277288,-0.1345490813,-0.0994921178,0.2476488799,-0.1799386889,-0.1785385013,-0.0476372056,0.0107344426,-0.0975554064,0.4373539984,-0.1490976363,0.3612090945,-0.012258118,0.1215736493,0.0213417336,-0.5040411353,-0.4507939517,-0.0658406541,-0.0317620635,0.0552451983,0.0770501271,0.0080055622,0.3263442218,-0.2391067594,-0.2724581957,0.3224855959,0.2141270787,0.0887082145,0.22979559,-0.1883960515,-0.0167393964,0.0299266316,0.5327509046,0.2355925739,0.0445311628,0.4013054967,-0.1830300242,-0.342936188,0.1383527666,0.4372718632,0.3134897947,0.0098209782,0.0195712894,-0.0359986238,-0.3422741592,0.3768794537,-0.1361479759,0.1983713806,-0.0884066597,-0.001343661,-0.1373440027,0.3042966723,-0.0974787921,-0.0624615848,0.1574099958,0.0914994776,-0.0297624823,-0.0523535758,0.2131533921,0.4122335017,0.0303778984,0.1924372464,-0.013811538,0.4103766978,-0.2020863146,-0.0439700596,-0.0946270898,-0.0712999925,0.1517301798,-0.1395329535,-0.205512628,0.1627992392,-0.0285273809,0.1827243418,-0.5095719695,0.427361697,0.2366541922,-0.0472982004,0.0571585074,0.1110836491,-0.002362475,-0.0089515029,-0.0438567549,0.0467418507,0.0703111067,0.0665785149,-0.0663850605,0.1159160361,0.2043628991,0.0928915516,0.0580297783,-0.0351178497,0.0987340659,0.1005920395,-0.0032767237,-0.1262062788,0.3256442547,-0.2848223448,0.2149302065,0.1594939679,0.0941943005,0.0378742032,-0.0761301368,0.1714250296,0.3254772723,0.3470335901,0.2231839895,0.0314969905,0.0246453322,-0.0549993627,-0.0549255013,-0.0206400882,-0.1116272211,0.1299436986,0.1952213347,0.0732089356,0.0617158674,0.0350541733,-0.1838503778,-0.2796123028,0.0911719501,0.3946083188,-0.3553246856,0.1200247556,-0.039686121,0.0555186421,0.1301530749,0.0189663973,-0.4570333958,-0.2215031236,-0.2029660195,0.4210834503,-0.1194587797,0.1644146591,-0.0810745582,0.0509676039,0.2963262796,-0.1255412996,-0.259331286,0.104335174,0.0625970736,0.0425717719,0.1925008893,-0.5198504925,0.1900203079,0.1750378907,0.1842620224,-0.0083790794,-0.5007292032,0.0480755009,-0.1537050158,0.1523056179,-0.1455800682,0.2838713229,0.0492252409,-0.2909571528,-0.0210785959,-0.1372650564,-0.0680153519,-0.0453203991,-0.0619840845,-0.2530847788,-0.1954331398,-0.069822371,-0.2933194339,-0.4212827682,0.1225009188,0.371108681,0.2529848516,0.4407685995,0.1377109438,0.0921132565,0.2690692544,0.0994600207,-0.0383629762,-0.107421346,0.2451769412,-0.3104769289,-0.2563972175,0.0066135097,0.0111053875,0.2837287486,0.337452203,-0.3564890623,-0.1421988308,-0.312304467,-0.0041876035,-0.0811958015,0.1192051098,0.045397982,-0.0257927254,0.1307452321,-0.1638926119,0.0347888954,-0.0795593485,0.1042242572,-0.0728223324,0.0976093784,0.0997843817,0.2115115076,0.6869980693,-0.1291385889,0.1413394809,0.018308932,0.2151246369,0.4513024092,0.0434456803,-0.2967938781,0.0668857172,0.0837084129,-0.206221208,-0.1333539337,0.0458995514,-0.1497395933,0.0206685234,0.1178896278,-0.0471880026,-0.2259911448,0.1134366393,-0.0898844674,0.3340997994,0.187919274,0.0507023931,-0.4434088469,-0.1077162847,-0.0562812537,0.3804533184,0.0898376033,0.1508090645,-0.3076767027,-0.0904264003,-0.2749199867,0.0611348636,-0.0811378509,0.5575761199,-0.1852121502,0.3113373816,-0.0447095297,0.1809080094,0.4595344961,-0.2760755718,0.005393195,-0.1066814512,-0.281932354,-0.3433832824,-0.2390835285,-0.0859023929,-0.06393902,0.1136865318,0.1906843483,-0.0791468695,-0.0084775966,-0.1428744942,-0.2193469405,-0.1370996833,-0.2479740083,-0.1115662828,-0.2691839337,-0.4986168742,-0.1251070797,0.3895351589,-0.1092313603,-0.1225882396,0.1031751484,-0.2260031849,0.0671980307,0.1904061288,0.1915994138,0.0266411565,-0.009566904,-0.1999968439,0.2065167427,-0.5022637248,0.062692441,0.19732593,-0.1263116598,-0.3567327857,-0.016329661,0.1559303403,-0.1357991099,0.0637567714,-0.174356252,-0.0275263917,0.0254739188,-0.0504413955,-0.071854718,0.3104915917,-0.055890426,0.195660457,-0.1012471616,-0.4688993096,0.19685857,0.2237318605,-0.0757478699,0.1689159125,0.0943683311,-0.3989681005,0.4668301046,0.1583802253,0.5565386415,-0.359916836,0.0974431783,0.3196151853,0.067224808,0.6953794956,-0.6555452943,0.3161601126,-0.2079330236,-0.3527629972,-0.1493595541,-0.3817493916,0.181630671,0.0691904053,-0.0273417775,0.0941415504,-0.0852651075,0.3506408036,0.1797598153,0.2020633519,0.0884777084,-0.4448313415,0.1528819054,0.2294978499,-0.2720116377,0.3177681863,-0.1780319661,-0.0277537853,-0.2918524146,0.0254852902,-0.2747633159,0.1161500663,-0.2991321683,0.5190944672,0.1113141179,0.0464832298,0.2879340053,0.2445803732,0.2088550776,0.2620199323,-0.2269392312,0.0637965649,-0.130311653,-0.0931614488,-0.0392174013,0.0066591357,0.335981667,0.0596252605,-0.0585409142,-0.0665251166,-0.1696169376,-0.3609890938,0.0174033288,0.2294077426,0.2024174631,0.006458255,-0.0510129295,-0.0119281607,-0.3825114667,0.0726773962,0.1492640525,0.0109344935,-0.2534069121,0.2334099412,-0.045485761,-0.2722888291,0.0967666358,0.3703703582,0.1194974408,-0.1971344352,0.559756279,0.2043311149,-0.2898311615,-0.0391659811,-0.2040651292,0.0931312293,-0.4541436136,-0.0446451493,-0.4564749002,-0.4141295254,-0.0188431554,-0.4614747167,0.3459784389,-0.0074619078,-0.1736587584,-0.4079886973,-0.3153894246,-0.0964628011,-0.3273630738,-0.0001220054,-0.0069627934,-0.0182451922,-0.0710390657,0.1626549214,-0.259177804,-0.2372059524,-0.3444597125,0.006200335,0.3342401981,-0.2212559134,0.0436337963,-0.1810938865,0.0256804116,0.1466897726,-0.387232244,-0.1846114099,-0.2641800344,0.1107682213,0.2623438537,-0.3276488781,-0.0680264756,-0.3471541703,-0.0914151073,-0.1358476132,0.1620068699,0.26431036,-0.1182678416,0.1854373962,-0.1474619806,0.2718785405,-0.019502759,0.1773811579,-0.0268762168,0.0644539818,0.0384416021,0.3823345602,0.1326795816,0.0603220277,-0.6228336692,0.1132022813,0.0375844911,0.1517114937,0.3931129277,-0.3460859954,0.3375230134,-0.0362393744,0.0605528355,-0.152901575,-0.3420254588,-0.0691744462,0.1160460636,0.1642261893,0.0276592579,-0.096303314,0.0548701957,0.032719519,0.0987614095,0.2690493762,0.2619537115,-0.0406337827,-0.0124462042,0.3720317185,0.6341760755,-0.2661919594,0.2063339502,0.2476853728,-0.2334974557,0.1992752403,0.2232558876,-0.002031469,0.0001358897,0.1990372539,0.1524522156,-0.0540718548,0.0205246545,-0.1834712923,0.336091429,-0.2124366164,-0.1275319606,-0.0399635695,0.1181048453,0.2308101356,-0.0397053212,-0.3249217868,0.0261873621,-0.0104645165,-0.2011923939,0.1950682998,-0.1402630657,0.0443249755,0.0428522676,0.203814134,0.0132805081,-0.2877651155,0.1271378845,-0.4779744148,-0.3058706224,0.3640507758,-0.1340831816,0.0985853449,0.0616869591,0.2978540659,-0.2202402949,-0.196892485,0.0838604197,0.1072050706,-0.1116496697,-0.082573697,0.0934753716,0.5431833267,0.2893403471,0.1599655449,0.2631548941,0.0980713665,-0.0594070368,-0.0906574205,0.3076898456,-0.2290304899,0.1799558848,0.1197366491,0.0925129056,-0.2126800418,-0.2806037664,0.093236357,0.0901547596,-0.0003153438,0.2890575826,-0.2614881396,-0.2587867677,0.0748261362,-0.0476370901,-0.1973496825,0.1740469337,0.8346324563,-0.1961341202,0.2542154193,-0.0428735539,0.0120079843,-0.1633709967,0.385330081,0.3202114701,0.0207920764,-0.2639081478,0.2996233106,-0.2990526259,0.120006606,-0.1916530728,-0.0139936889,0.1818581522,0.1701564491,-0.0613947324,-0.0301435366,0.1060565785,-0.4536414444,-0.0028722938,0.0240874756,-0.2195233554,-0.1286708415,0.1216328964,0.0295880251,0.0657106042,-0.3943275809,0.0754953325,0.0724963695,0.0972357243,-0.3166572452,0.0166931972,0.2151388079,0.2526507676,0.2539834082,0.2079452425,0.3316455483,0.0703696534,0.0958230048,-0.3994336128,-0.349573344,-0.3445076644,0.0160421301,-0.1569499522,0.5965117812,-0.2879594266,-0.5061549544,-0.3620113134,0.0960142016,0.105117783,-0.2808544934,-0.2017903328,0.1834330857,-0.4025213122,0.3427912593,0.2535629272,-0.0980608538,0.244820863,0.3333668411,-0.3015104234,-0.3771386743,0.5819663405,-0.4455092549,-0.4940264225,0.2248333991,0.0352744348,-0.1100796983,-0.0267488882,-0.4583678842,0.2484734058,0.1376974732,-0.3479514122,-0.3160350621,0.0086388001,0.1702643633,0.2086642385,-0.0559566729,0.1456699669,-0.0756859928,-0.3460895717,0.118467927,-0.0319126099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1581","title":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'","comments":"> Thanks for reporting !\r\n> You can override the directory in which cache file are stored using for example\r\n> \r\n> ```\r\n> ENV HF_HOME=\"\/root\/cache\/hf_cache_home\"\r\n> ```\r\n> \r\n> This way both `transformers` and `datasets` will use this directory instead of the default `.cache`\r\n\r\ncan we disable caching directly?","body":"I am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n","comment_length":50,"text":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'\nI am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n\n> Thanks for reporting !\r\n> You can override the directory in which cache file are stored using for example\r\n> \r\n> ```\r\n> ENV HF_HOME=\"\/root\/cache\/hf_cache_home\"\r\n> ```\r\n> \r\n> This way both `transformers` and `datasets` will use this directory instead of the default `.cache`\r\n\r\ncan we disable caching directly?","embeddings":[-0.2601940632,0.1537577212,-0.087916173,0.1361187547,0.0533822998,0.0593192391,0.656006515,0.1960314065,0.161277622,-0.0136698019,-0.3051161468,-0.168390587,-0.0324441306,-0.5109063387,-0.0259655211,-0.2055929005,0.0551733933,-0.0676345825,-0.3101544976,-0.1253348142,-0.0913535804,-0.0168972444,-0.0075819497,0.1020792797,-0.3513800204,0.0235478356,0.333840996,-0.0979835615,0.0049274717,-0.1136808172,0.3865947425,0.1387621909,0.1959901601,0.8908079863,-0.0001138685,0.364698112,0.1204629838,0.0754132047,0.0866398886,-0.2086414546,-0.1387084424,-0.2477509379,-0.1201189086,-0.0817605555,0.0790471286,0.0090726782,0.0140740611,-0.1214024574,0.6044265628,0.3587148786,0.2027944475,0.4903935492,0.0908189341,0.123684071,0.2537278831,-0.0316032395,-0.2483440191,-0.0738279372,-0.1243278161,-0.2798888981,-0.0047014416,0.2760694027,-0.0540417656,0.1426951587,0.2170650065,-0.1940569878,0.1225119233,-0.3570998013,-0.0412276238,0.0453185365,0.5151111484,-0.065642722,-0.5028495193,-0.1864654869,-0.1189291701,0.0584088117,0.302632153,0.3046842217,-0.0361773148,0.2790159285,-0.2144559771,-0.1514513791,-0.4730093479,0.095054619,-0.2394710332,0.3513744175,0.0620445088,0.1947925389,-0.0911417976,-0.0715904012,0.6296569109,-0.0060653305,0.1474893838,0.2979938984,-0.3239389062,0.2316237688,0.0331782922,-0.5658233166,-0.3128330112,0.0865982324,-0.2314409018,-0.0924056843,-0.0704433471,0.3829046488,-0.2021469325,0.4120383263,-0.3363805413,0.7340712547,0.3856590092,-0.3047202826,0.0332622379,-0.0432148613,-0.0374102555,-0.0881388262,-0.1260698587,-0.0319224447,0.14577277,0.0939131603,-0.098433651,-0.034681011,0.086896874,-0.0100169806,0.1854218394,0.3815259933,0.1036055908,0.0991566405,0.3901361525,0.1196964458,-0.0152244633,0.0574717671,-0.2500608265,0.072026372,-0.0946181267,-0.0641494319,0.1987671703,0.0866814032,0.0298909303,-0.1037014425,0.2303165644,-0.1406204551,0.4395029545,-0.2077358812,0.1142411977,0.7253692746,0.1236809492,0.040582139,0.3252632618,-0.1411764622,0.0385678783,0.1115270332,-0.3354057968,-0.1929328144,0.0483054183,0.1385359615,-0.133828178,0.2708508074,-0.2460403144,-0.1343725771,0.3796409369,-0.0594122037,-0.0387408696,-0.2568915188,0.0189277288,-0.1345490813,-0.0994921178,0.2476488799,-0.1799386889,-0.1785385013,-0.0476372056,0.0107344426,-0.0975554064,0.4373539984,-0.1490976363,0.3612090945,-0.012258118,0.1215736493,0.0213417336,-0.5040411353,-0.4507939517,-0.0658406541,-0.0317620635,0.0552451983,0.0770501271,0.0080055622,0.3263442218,-0.2391067594,-0.2724581957,0.3224855959,0.2141270787,0.0887082145,0.22979559,-0.1883960515,-0.0167393964,0.0299266316,0.5327509046,0.2355925739,0.0445311628,0.4013054967,-0.1830300242,-0.342936188,0.1383527666,0.4372718632,0.3134897947,0.0098209782,0.0195712894,-0.0359986238,-0.3422741592,0.3768794537,-0.1361479759,0.1983713806,-0.0884066597,-0.001343661,-0.1373440027,0.3042966723,-0.0974787921,-0.0624615848,0.1574099958,0.0914994776,-0.0297624823,-0.0523535758,0.2131533921,0.4122335017,0.0303778984,0.1924372464,-0.013811538,0.4103766978,-0.2020863146,-0.0439700596,-0.0946270898,-0.0712999925,0.1517301798,-0.1395329535,-0.205512628,0.1627992392,-0.0285273809,0.1827243418,-0.5095719695,0.427361697,0.2366541922,-0.0472982004,0.0571585074,0.1110836491,-0.002362475,-0.0089515029,-0.0438567549,0.0467418507,0.0703111067,0.0665785149,-0.0663850605,0.1159160361,0.2043628991,0.0928915516,0.0580297783,-0.0351178497,0.0987340659,0.1005920395,-0.0032767237,-0.1262062788,0.3256442547,-0.2848223448,0.2149302065,0.1594939679,0.0941943005,0.0378742032,-0.0761301368,0.1714250296,0.3254772723,0.3470335901,0.2231839895,0.0314969905,0.0246453322,-0.0549993627,-0.0549255013,-0.0206400882,-0.1116272211,0.1299436986,0.1952213347,0.0732089356,0.0617158674,0.0350541733,-0.1838503778,-0.2796123028,0.0911719501,0.3946083188,-0.3553246856,0.1200247556,-0.039686121,0.0555186421,0.1301530749,0.0189663973,-0.4570333958,-0.2215031236,-0.2029660195,0.4210834503,-0.1194587797,0.1644146591,-0.0810745582,0.0509676039,0.2963262796,-0.1255412996,-0.259331286,0.104335174,0.0625970736,0.0425717719,0.1925008893,-0.5198504925,0.1900203079,0.1750378907,0.1842620224,-0.0083790794,-0.5007292032,0.0480755009,-0.1537050158,0.1523056179,-0.1455800682,0.2838713229,0.0492252409,-0.2909571528,-0.0210785959,-0.1372650564,-0.0680153519,-0.0453203991,-0.0619840845,-0.2530847788,-0.1954331398,-0.069822371,-0.2933194339,-0.4212827682,0.1225009188,0.371108681,0.2529848516,0.4407685995,0.1377109438,0.0921132565,0.2690692544,0.0994600207,-0.0383629762,-0.107421346,0.2451769412,-0.3104769289,-0.2563972175,0.0066135097,0.0111053875,0.2837287486,0.337452203,-0.3564890623,-0.1421988308,-0.312304467,-0.0041876035,-0.0811958015,0.1192051098,0.045397982,-0.0257927254,0.1307452321,-0.1638926119,0.0347888954,-0.0795593485,0.1042242572,-0.0728223324,0.0976093784,0.0997843817,0.2115115076,0.6869980693,-0.1291385889,0.1413394809,0.018308932,0.2151246369,0.4513024092,0.0434456803,-0.2967938781,0.0668857172,0.0837084129,-0.206221208,-0.1333539337,0.0458995514,-0.1497395933,0.0206685234,0.1178896278,-0.0471880026,-0.2259911448,0.1134366393,-0.0898844674,0.3340997994,0.187919274,0.0507023931,-0.4434088469,-0.1077162847,-0.0562812537,0.3804533184,0.0898376033,0.1508090645,-0.3076767027,-0.0904264003,-0.2749199867,0.0611348636,-0.0811378509,0.5575761199,-0.1852121502,0.3113373816,-0.0447095297,0.1809080094,0.4595344961,-0.2760755718,0.005393195,-0.1066814512,-0.281932354,-0.3433832824,-0.2390835285,-0.0859023929,-0.06393902,0.1136865318,0.1906843483,-0.0791468695,-0.0084775966,-0.1428744942,-0.2193469405,-0.1370996833,-0.2479740083,-0.1115662828,-0.2691839337,-0.4986168742,-0.1251070797,0.3895351589,-0.1092313603,-0.1225882396,0.1031751484,-0.2260031849,0.0671980307,0.1904061288,0.1915994138,0.0266411565,-0.009566904,-0.1999968439,0.2065167427,-0.5022637248,0.062692441,0.19732593,-0.1263116598,-0.3567327857,-0.016329661,0.1559303403,-0.1357991099,0.0637567714,-0.174356252,-0.0275263917,0.0254739188,-0.0504413955,-0.071854718,0.3104915917,-0.055890426,0.195660457,-0.1012471616,-0.4688993096,0.19685857,0.2237318605,-0.0757478699,0.1689159125,0.0943683311,-0.3989681005,0.4668301046,0.1583802253,0.5565386415,-0.359916836,0.0974431783,0.3196151853,0.067224808,0.6953794956,-0.6555452943,0.3161601126,-0.2079330236,-0.3527629972,-0.1493595541,-0.3817493916,0.181630671,0.0691904053,-0.0273417775,0.0941415504,-0.0852651075,0.3506408036,0.1797598153,0.2020633519,0.0884777084,-0.4448313415,0.1528819054,0.2294978499,-0.2720116377,0.3177681863,-0.1780319661,-0.0277537853,-0.2918524146,0.0254852902,-0.2747633159,0.1161500663,-0.2991321683,0.5190944672,0.1113141179,0.0464832298,0.2879340053,0.2445803732,0.2088550776,0.2620199323,-0.2269392312,0.0637965649,-0.130311653,-0.0931614488,-0.0392174013,0.0066591357,0.335981667,0.0596252605,-0.0585409142,-0.0665251166,-0.1696169376,-0.3609890938,0.0174033288,0.2294077426,0.2024174631,0.006458255,-0.0510129295,-0.0119281607,-0.3825114667,0.0726773962,0.1492640525,0.0109344935,-0.2534069121,0.2334099412,-0.045485761,-0.2722888291,0.0967666358,0.3703703582,0.1194974408,-0.1971344352,0.559756279,0.2043311149,-0.2898311615,-0.0391659811,-0.2040651292,0.0931312293,-0.4541436136,-0.0446451493,-0.4564749002,-0.4141295254,-0.0188431554,-0.4614747167,0.3459784389,-0.0074619078,-0.1736587584,-0.4079886973,-0.3153894246,-0.0964628011,-0.3273630738,-0.0001220054,-0.0069627934,-0.0182451922,-0.0710390657,0.1626549214,-0.259177804,-0.2372059524,-0.3444597125,0.006200335,0.3342401981,-0.2212559134,0.0436337963,-0.1810938865,0.0256804116,0.1466897726,-0.387232244,-0.1846114099,-0.2641800344,0.1107682213,0.2623438537,-0.3276488781,-0.0680264756,-0.3471541703,-0.0914151073,-0.1358476132,0.1620068699,0.26431036,-0.1182678416,0.1854373962,-0.1474619806,0.2718785405,-0.019502759,0.1773811579,-0.0268762168,0.0644539818,0.0384416021,0.3823345602,0.1326795816,0.0603220277,-0.6228336692,0.1132022813,0.0375844911,0.1517114937,0.3931129277,-0.3460859954,0.3375230134,-0.0362393744,0.0605528355,-0.152901575,-0.3420254588,-0.0691744462,0.1160460636,0.1642261893,0.0276592579,-0.096303314,0.0548701957,0.032719519,0.0987614095,0.2690493762,0.2619537115,-0.0406337827,-0.0124462042,0.3720317185,0.6341760755,-0.2661919594,0.2063339502,0.2476853728,-0.2334974557,0.1992752403,0.2232558876,-0.002031469,0.0001358897,0.1990372539,0.1524522156,-0.0540718548,0.0205246545,-0.1834712923,0.336091429,-0.2124366164,-0.1275319606,-0.0399635695,0.1181048453,0.2308101356,-0.0397053212,-0.3249217868,0.0261873621,-0.0104645165,-0.2011923939,0.1950682998,-0.1402630657,0.0443249755,0.0428522676,0.203814134,0.0132805081,-0.2877651155,0.1271378845,-0.4779744148,-0.3058706224,0.3640507758,-0.1340831816,0.0985853449,0.0616869591,0.2978540659,-0.2202402949,-0.196892485,0.0838604197,0.1072050706,-0.1116496697,-0.082573697,0.0934753716,0.5431833267,0.2893403471,0.1599655449,0.2631548941,0.0980713665,-0.0594070368,-0.0906574205,0.3076898456,-0.2290304899,0.1799558848,0.1197366491,0.0925129056,-0.2126800418,-0.2806037664,0.093236357,0.0901547596,-0.0003153438,0.2890575826,-0.2614881396,-0.2587867677,0.0748261362,-0.0476370901,-0.1973496825,0.1740469337,0.8346324563,-0.1961341202,0.2542154193,-0.0428735539,0.0120079843,-0.1633709967,0.385330081,0.3202114701,0.0207920764,-0.2639081478,0.2996233106,-0.2990526259,0.120006606,-0.1916530728,-0.0139936889,0.1818581522,0.1701564491,-0.0613947324,-0.0301435366,0.1060565785,-0.4536414444,-0.0028722938,0.0240874756,-0.2195233554,-0.1286708415,0.1216328964,0.0295880251,0.0657106042,-0.3943275809,0.0754953325,0.0724963695,0.0972357243,-0.3166572452,0.0166931972,0.2151388079,0.2526507676,0.2539834082,0.2079452425,0.3316455483,0.0703696534,0.0958230048,-0.3994336128,-0.349573344,-0.3445076644,0.0160421301,-0.1569499522,0.5965117812,-0.2879594266,-0.5061549544,-0.3620113134,0.0960142016,0.105117783,-0.2808544934,-0.2017903328,0.1834330857,-0.4025213122,0.3427912593,0.2535629272,-0.0980608538,0.244820863,0.3333668411,-0.3015104234,-0.3771386743,0.5819663405,-0.4455092549,-0.4940264225,0.2248333991,0.0352744348,-0.1100796983,-0.0267488882,-0.4583678842,0.2484734058,0.1376974732,-0.3479514122,-0.3160350621,0.0086388001,0.1702643633,0.2086642385,-0.0559566729,0.1456699669,-0.0756859928,-0.3460895717,0.118467927,-0.0319126099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1581","title":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'","comments":"Hi ! Unfortunately no since we need this directory to load datasets.\r\nWhen you load a dataset, it downloads the raw data files in the cache directory inside <cache_dir>\/downloads. Then it builds the dataset and saves it as arrow data inside <cache_dir>\/<dataset_name>.\r\n\r\nHowever you can specify the directory of your choice, and it can be a temporary directory if you want to clean everything up at one point.","body":"I am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n","comment_length":68,"text":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'\nI am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n\nHi ! Unfortunately no since we need this directory to load datasets.\r\nWhen you load a dataset, it downloads the raw data files in the cache directory inside <cache_dir>\/downloads. Then it builds the dataset and saves it as arrow data inside <cache_dir>\/<dataset_name>.\r\n\r\nHowever you can specify the directory of your choice, and it can be a temporary directory if you want to clean everything up at one point.","embeddings":[-0.2601940632,0.1537577212,-0.087916173,0.1361187547,0.0533822998,0.0593192391,0.656006515,0.1960314065,0.161277622,-0.0136698019,-0.3051161468,-0.168390587,-0.0324441306,-0.5109063387,-0.0259655211,-0.2055929005,0.0551733933,-0.0676345825,-0.3101544976,-0.1253348142,-0.0913535804,-0.0168972444,-0.0075819497,0.1020792797,-0.3513800204,0.0235478356,0.333840996,-0.0979835615,0.0049274717,-0.1136808172,0.3865947425,0.1387621909,0.1959901601,0.8908079863,-0.0001138685,0.364698112,0.1204629838,0.0754132047,0.0866398886,-0.2086414546,-0.1387084424,-0.2477509379,-0.1201189086,-0.0817605555,0.0790471286,0.0090726782,0.0140740611,-0.1214024574,0.6044265628,0.3587148786,0.2027944475,0.4903935492,0.0908189341,0.123684071,0.2537278831,-0.0316032395,-0.2483440191,-0.0738279372,-0.1243278161,-0.2798888981,-0.0047014416,0.2760694027,-0.0540417656,0.1426951587,0.2170650065,-0.1940569878,0.1225119233,-0.3570998013,-0.0412276238,0.0453185365,0.5151111484,-0.065642722,-0.5028495193,-0.1864654869,-0.1189291701,0.0584088117,0.302632153,0.3046842217,-0.0361773148,0.2790159285,-0.2144559771,-0.1514513791,-0.4730093479,0.095054619,-0.2394710332,0.3513744175,0.0620445088,0.1947925389,-0.0911417976,-0.0715904012,0.6296569109,-0.0060653305,0.1474893838,0.2979938984,-0.3239389062,0.2316237688,0.0331782922,-0.5658233166,-0.3128330112,0.0865982324,-0.2314409018,-0.0924056843,-0.0704433471,0.3829046488,-0.2021469325,0.4120383263,-0.3363805413,0.7340712547,0.3856590092,-0.3047202826,0.0332622379,-0.0432148613,-0.0374102555,-0.0881388262,-0.1260698587,-0.0319224447,0.14577277,0.0939131603,-0.098433651,-0.034681011,0.086896874,-0.0100169806,0.1854218394,0.3815259933,0.1036055908,0.0991566405,0.3901361525,0.1196964458,-0.0152244633,0.0574717671,-0.2500608265,0.072026372,-0.0946181267,-0.0641494319,0.1987671703,0.0866814032,0.0298909303,-0.1037014425,0.2303165644,-0.1406204551,0.4395029545,-0.2077358812,0.1142411977,0.7253692746,0.1236809492,0.040582139,0.3252632618,-0.1411764622,0.0385678783,0.1115270332,-0.3354057968,-0.1929328144,0.0483054183,0.1385359615,-0.133828178,0.2708508074,-0.2460403144,-0.1343725771,0.3796409369,-0.0594122037,-0.0387408696,-0.2568915188,0.0189277288,-0.1345490813,-0.0994921178,0.2476488799,-0.1799386889,-0.1785385013,-0.0476372056,0.0107344426,-0.0975554064,0.4373539984,-0.1490976363,0.3612090945,-0.012258118,0.1215736493,0.0213417336,-0.5040411353,-0.4507939517,-0.0658406541,-0.0317620635,0.0552451983,0.0770501271,0.0080055622,0.3263442218,-0.2391067594,-0.2724581957,0.3224855959,0.2141270787,0.0887082145,0.22979559,-0.1883960515,-0.0167393964,0.0299266316,0.5327509046,0.2355925739,0.0445311628,0.4013054967,-0.1830300242,-0.342936188,0.1383527666,0.4372718632,0.3134897947,0.0098209782,0.0195712894,-0.0359986238,-0.3422741592,0.3768794537,-0.1361479759,0.1983713806,-0.0884066597,-0.001343661,-0.1373440027,0.3042966723,-0.0974787921,-0.0624615848,0.1574099958,0.0914994776,-0.0297624823,-0.0523535758,0.2131533921,0.4122335017,0.0303778984,0.1924372464,-0.013811538,0.4103766978,-0.2020863146,-0.0439700596,-0.0946270898,-0.0712999925,0.1517301798,-0.1395329535,-0.205512628,0.1627992392,-0.0285273809,0.1827243418,-0.5095719695,0.427361697,0.2366541922,-0.0472982004,0.0571585074,0.1110836491,-0.002362475,-0.0089515029,-0.0438567549,0.0467418507,0.0703111067,0.0665785149,-0.0663850605,0.1159160361,0.2043628991,0.0928915516,0.0580297783,-0.0351178497,0.0987340659,0.1005920395,-0.0032767237,-0.1262062788,0.3256442547,-0.2848223448,0.2149302065,0.1594939679,0.0941943005,0.0378742032,-0.0761301368,0.1714250296,0.3254772723,0.3470335901,0.2231839895,0.0314969905,0.0246453322,-0.0549993627,-0.0549255013,-0.0206400882,-0.1116272211,0.1299436986,0.1952213347,0.0732089356,0.0617158674,0.0350541733,-0.1838503778,-0.2796123028,0.0911719501,0.3946083188,-0.3553246856,0.1200247556,-0.039686121,0.0555186421,0.1301530749,0.0189663973,-0.4570333958,-0.2215031236,-0.2029660195,0.4210834503,-0.1194587797,0.1644146591,-0.0810745582,0.0509676039,0.2963262796,-0.1255412996,-0.259331286,0.104335174,0.0625970736,0.0425717719,0.1925008893,-0.5198504925,0.1900203079,0.1750378907,0.1842620224,-0.0083790794,-0.5007292032,0.0480755009,-0.1537050158,0.1523056179,-0.1455800682,0.2838713229,0.0492252409,-0.2909571528,-0.0210785959,-0.1372650564,-0.0680153519,-0.0453203991,-0.0619840845,-0.2530847788,-0.1954331398,-0.069822371,-0.2933194339,-0.4212827682,0.1225009188,0.371108681,0.2529848516,0.4407685995,0.1377109438,0.0921132565,0.2690692544,0.0994600207,-0.0383629762,-0.107421346,0.2451769412,-0.3104769289,-0.2563972175,0.0066135097,0.0111053875,0.2837287486,0.337452203,-0.3564890623,-0.1421988308,-0.312304467,-0.0041876035,-0.0811958015,0.1192051098,0.045397982,-0.0257927254,0.1307452321,-0.1638926119,0.0347888954,-0.0795593485,0.1042242572,-0.0728223324,0.0976093784,0.0997843817,0.2115115076,0.6869980693,-0.1291385889,0.1413394809,0.018308932,0.2151246369,0.4513024092,0.0434456803,-0.2967938781,0.0668857172,0.0837084129,-0.206221208,-0.1333539337,0.0458995514,-0.1497395933,0.0206685234,0.1178896278,-0.0471880026,-0.2259911448,0.1134366393,-0.0898844674,0.3340997994,0.187919274,0.0507023931,-0.4434088469,-0.1077162847,-0.0562812537,0.3804533184,0.0898376033,0.1508090645,-0.3076767027,-0.0904264003,-0.2749199867,0.0611348636,-0.0811378509,0.5575761199,-0.1852121502,0.3113373816,-0.0447095297,0.1809080094,0.4595344961,-0.2760755718,0.005393195,-0.1066814512,-0.281932354,-0.3433832824,-0.2390835285,-0.0859023929,-0.06393902,0.1136865318,0.1906843483,-0.0791468695,-0.0084775966,-0.1428744942,-0.2193469405,-0.1370996833,-0.2479740083,-0.1115662828,-0.2691839337,-0.4986168742,-0.1251070797,0.3895351589,-0.1092313603,-0.1225882396,0.1031751484,-0.2260031849,0.0671980307,0.1904061288,0.1915994138,0.0266411565,-0.009566904,-0.1999968439,0.2065167427,-0.5022637248,0.062692441,0.19732593,-0.1263116598,-0.3567327857,-0.016329661,0.1559303403,-0.1357991099,0.0637567714,-0.174356252,-0.0275263917,0.0254739188,-0.0504413955,-0.071854718,0.3104915917,-0.055890426,0.195660457,-0.1012471616,-0.4688993096,0.19685857,0.2237318605,-0.0757478699,0.1689159125,0.0943683311,-0.3989681005,0.4668301046,0.1583802253,0.5565386415,-0.359916836,0.0974431783,0.3196151853,0.067224808,0.6953794956,-0.6555452943,0.3161601126,-0.2079330236,-0.3527629972,-0.1493595541,-0.3817493916,0.181630671,0.0691904053,-0.0273417775,0.0941415504,-0.0852651075,0.3506408036,0.1797598153,0.2020633519,0.0884777084,-0.4448313415,0.1528819054,0.2294978499,-0.2720116377,0.3177681863,-0.1780319661,-0.0277537853,-0.2918524146,0.0254852902,-0.2747633159,0.1161500663,-0.2991321683,0.5190944672,0.1113141179,0.0464832298,0.2879340053,0.2445803732,0.2088550776,0.2620199323,-0.2269392312,0.0637965649,-0.130311653,-0.0931614488,-0.0392174013,0.0066591357,0.335981667,0.0596252605,-0.0585409142,-0.0665251166,-0.1696169376,-0.3609890938,0.0174033288,0.2294077426,0.2024174631,0.006458255,-0.0510129295,-0.0119281607,-0.3825114667,0.0726773962,0.1492640525,0.0109344935,-0.2534069121,0.2334099412,-0.045485761,-0.2722888291,0.0967666358,0.3703703582,0.1194974408,-0.1971344352,0.559756279,0.2043311149,-0.2898311615,-0.0391659811,-0.2040651292,0.0931312293,-0.4541436136,-0.0446451493,-0.4564749002,-0.4141295254,-0.0188431554,-0.4614747167,0.3459784389,-0.0074619078,-0.1736587584,-0.4079886973,-0.3153894246,-0.0964628011,-0.3273630738,-0.0001220054,-0.0069627934,-0.0182451922,-0.0710390657,0.1626549214,-0.259177804,-0.2372059524,-0.3444597125,0.006200335,0.3342401981,-0.2212559134,0.0436337963,-0.1810938865,0.0256804116,0.1466897726,-0.387232244,-0.1846114099,-0.2641800344,0.1107682213,0.2623438537,-0.3276488781,-0.0680264756,-0.3471541703,-0.0914151073,-0.1358476132,0.1620068699,0.26431036,-0.1182678416,0.1854373962,-0.1474619806,0.2718785405,-0.019502759,0.1773811579,-0.0268762168,0.0644539818,0.0384416021,0.3823345602,0.1326795816,0.0603220277,-0.6228336692,0.1132022813,0.0375844911,0.1517114937,0.3931129277,-0.3460859954,0.3375230134,-0.0362393744,0.0605528355,-0.152901575,-0.3420254588,-0.0691744462,0.1160460636,0.1642261893,0.0276592579,-0.096303314,0.0548701957,0.032719519,0.0987614095,0.2690493762,0.2619537115,-0.0406337827,-0.0124462042,0.3720317185,0.6341760755,-0.2661919594,0.2063339502,0.2476853728,-0.2334974557,0.1992752403,0.2232558876,-0.002031469,0.0001358897,0.1990372539,0.1524522156,-0.0540718548,0.0205246545,-0.1834712923,0.336091429,-0.2124366164,-0.1275319606,-0.0399635695,0.1181048453,0.2308101356,-0.0397053212,-0.3249217868,0.0261873621,-0.0104645165,-0.2011923939,0.1950682998,-0.1402630657,0.0443249755,0.0428522676,0.203814134,0.0132805081,-0.2877651155,0.1271378845,-0.4779744148,-0.3058706224,0.3640507758,-0.1340831816,0.0985853449,0.0616869591,0.2978540659,-0.2202402949,-0.196892485,0.0838604197,0.1072050706,-0.1116496697,-0.082573697,0.0934753716,0.5431833267,0.2893403471,0.1599655449,0.2631548941,0.0980713665,-0.0594070368,-0.0906574205,0.3076898456,-0.2290304899,0.1799558848,0.1197366491,0.0925129056,-0.2126800418,-0.2806037664,0.093236357,0.0901547596,-0.0003153438,0.2890575826,-0.2614881396,-0.2587867677,0.0748261362,-0.0476370901,-0.1973496825,0.1740469337,0.8346324563,-0.1961341202,0.2542154193,-0.0428735539,0.0120079843,-0.1633709967,0.385330081,0.3202114701,0.0207920764,-0.2639081478,0.2996233106,-0.2990526259,0.120006606,-0.1916530728,-0.0139936889,0.1818581522,0.1701564491,-0.0613947324,-0.0301435366,0.1060565785,-0.4536414444,-0.0028722938,0.0240874756,-0.2195233554,-0.1286708415,0.1216328964,0.0295880251,0.0657106042,-0.3943275809,0.0754953325,0.0724963695,0.0972357243,-0.3166572452,0.0166931972,0.2151388079,0.2526507676,0.2539834082,0.2079452425,0.3316455483,0.0703696534,0.0958230048,-0.3994336128,-0.349573344,-0.3445076644,0.0160421301,-0.1569499522,0.5965117812,-0.2879594266,-0.5061549544,-0.3620113134,0.0960142016,0.105117783,-0.2808544934,-0.2017903328,0.1834330857,-0.4025213122,0.3427912593,0.2535629272,-0.0980608538,0.244820863,0.3333668411,-0.3015104234,-0.3771386743,0.5819663405,-0.4455092549,-0.4940264225,0.2248333991,0.0352744348,-0.1100796983,-0.0267488882,-0.4583678842,0.2484734058,0.1376974732,-0.3479514122,-0.3160350621,0.0086388001,0.1702643633,0.2086642385,-0.0559566729,0.1456699669,-0.0756859928,-0.3460895717,0.118467927,-0.0319126099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1581","title":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'","comments":"I'm closing this to keep issues a bit cleaner","body":"I am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n","comment_length":9,"text":"Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'\nI am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `\/.cache`:\r\n\r\n```\r\n$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)\/data:\/root\/data -v $(pwd):\/root -v $(pwd)\/models\/:\/root\/models -v $(pwd)\/saved_models\/:\/root\/saved_models -e \"HOST_HOSTNAME=$(hostname)\" hf-error:latest \/bin\/bash\r\n\r\n________ _______________ \r\n___ __\/__________________________________ ____\/__ \/________ __\r\n__ \/ _ _ \\_ __ \\_ ___\/ __ \\_ ___\/_ \/_ __ \/_ __ \\_ | \/| \/ \/\r\n_ \/ \/ __\/ \/ \/ \/(__ )\/ \/_\/ \/ \/ _ __\/ _ \/ \/ \/_\/ \/_ |\/ |\/ \/ \r\n\/_\/ \\___\/\/_\/ \/_\/\/____\/ \\____\/\/_\/ \/_\/ \/_\/ \\____\/____\/|__\/\r\n\r\n\r\nYou are running this container as user with ID 1000 and group 1000,\r\nwhich should map to the ID and group for your user on the Docker host. Great!\r\n\r\ntf-docker \/root > python\r\nPython 3.6.9 (default, Oct 8 2020, 12:12:24) \r\n[GCC 8.4.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import transformers\r\n2020-12-15 23:53:21.165827: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/__init__.py\", line 22, in <module>\r\n from .integrations import ( # isort:skip\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/integrations.py\", line 5, in <module>\r\n from .trainer_utils import EvaluationStrategy\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/trainer_utils.py\", line 25, in <module>\r\n from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/file_utils.py\", line 88, in <module>\r\n import datasets # noqa: F401\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/__init__.py\", line 26, in <module>\r\n from .arrow_dataset import Dataset, concatenate_datasets\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py\", line 40, in <module>\r\n from .arrow_reader import ArrowReader\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_reader.py\", line 31, in <module>\r\n from .utils import cached_path, logging\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/download_manager.py\", line 25, in <module>\r\n from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 118, in <module>\r\n os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 210, in makedirs\r\n makedirs(head, mode, exist_ok)\r\n File \"\/usr\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nPermissionError: [Errno 13] Permission denied: '\/.cache'\r\n```\r\nI've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile.\r\n\r\n```\r\nFROM tensorflow\/tensorflow:latest-gpu-jupyter\r\nWORKDIR \/root\r\n\r\nEXPOSE 80\r\nEXPOSE 8888\r\nEXPOSE 6006\r\n\r\nENV SHELL \/bin\/bash\r\nENV PATH=\"\/root\/.local\/bin:${PATH}\"\r\n\r\nENV CUDA_CACHE_PATH=\"\/root\/cache\/cuda\"\r\nENV CUDA_CACHE_MAXSIZE=\"4294967296\"\r\n\r\nENV TFHUB_CACHE_DIR=\"\/root\/cache\/tfhub\"\r\n\r\nRUN pip install --upgrade pip\r\n\r\nRUN apt update -y && apt upgrade -y\r\n\r\nRUN pip install transformers\r\n\r\n#Installing datasets will throw the error, try commenting and rebuilding\r\nRUN pip install datasets\r\n\r\n#Another workaround is creating the directory and give permissions explicitly\r\n#RUN mkdir \/.cache\r\n#RUN chmod 777 \/.cache\r\n```\r\n\nI'm closing this to keep issues a bit cleaner","embeddings":[-0.2601940632,0.1537577212,-0.087916173,0.1361187547,0.0533822998,0.0593192391,0.656006515,0.1960314065,0.161277622,-0.0136698019,-0.3051161468,-0.168390587,-0.0324441306,-0.5109063387,-0.0259655211,-0.2055929005,0.0551733933,-0.0676345825,-0.3101544976,-0.1253348142,-0.0913535804,-0.0168972444,-0.0075819497,0.1020792797,-0.3513800204,0.0235478356,0.333840996,-0.0979835615,0.0049274717,-0.1136808172,0.3865947425,0.1387621909,0.1959901601,0.8908079863,-0.0001138685,0.364698112,0.1204629838,0.0754132047,0.0866398886,-0.2086414546,-0.1387084424,-0.2477509379,-0.1201189086,-0.0817605555,0.0790471286,0.0090726782,0.0140740611,-0.1214024574,0.6044265628,0.3587148786,0.2027944475,0.4903935492,0.0908189341,0.123684071,0.2537278831,-0.0316032395,-0.2483440191,-0.0738279372,-0.1243278161,-0.2798888981,-0.0047014416,0.2760694027,-0.0540417656,0.1426951587,0.2170650065,-0.1940569878,0.1225119233,-0.3570998013,-0.0412276238,0.0453185365,0.5151111484,-0.065642722,-0.5028495193,-0.1864654869,-0.1189291701,0.0584088117,0.302632153,0.3046842217,-0.0361773148,0.2790159285,-0.2144559771,-0.1514513791,-0.4730093479,0.095054619,-0.2394710332,0.3513744175,0.0620445088,0.1947925389,-0.0911417976,-0.0715904012,0.6296569109,-0.0060653305,0.1474893838,0.2979938984,-0.3239389062,0.2316237688,0.0331782922,-0.5658233166,-0.3128330112,0.0865982324,-0.2314409018,-0.0924056843,-0.0704433471,0.3829046488,-0.2021469325,0.4120383263,-0.3363805413,0.7340712547,0.3856590092,-0.3047202826,0.0332622379,-0.0432148613,-0.0374102555,-0.0881388262,-0.1260698587,-0.0319224447,0.14577277,0.0939131603,-0.098433651,-0.034681011,0.086896874,-0.0100169806,0.1854218394,0.3815259933,0.1036055908,0.0991566405,0.3901361525,0.1196964458,-0.0152244633,0.0574717671,-0.2500608265,0.072026372,-0.0946181267,-0.0641494319,0.1987671703,0.0866814032,0.0298909303,-0.1037014425,0.2303165644,-0.1406204551,0.4395029545,-0.2077358812,0.1142411977,0.7253692746,0.1236809492,0.040582139,0.3252632618,-0.1411764622,0.0385678783,0.1115270332,-0.3354057968,-0.1929328144,0.0483054183,0.1385359615,-0.133828178,0.2708508074,-0.2460403144,-0.1343725771,0.3796409369,-0.0594122037,-0.0387408696,-0.2568915188,0.0189277288,-0.1345490813,-0.0994921178,0.2476488799,-0.1799386889,-0.1785385013,-0.0476372056,0.0107344426,-0.0975554064,0.4373539984,-0.1490976363,0.3612090945,-0.012258118,0.1215736493,0.0213417336,-0.5040411353,-0.4507939517,-0.0658406541,-0.0317620635,0.0552451983,0.0770501271,0.0080055622,0.3263442218,-0.2391067594,-0.2724581957,0.3224855959,0.2141270787,0.0887082145,0.22979559,-0.1883960515,-0.0167393964,0.0299266316,0.5327509046,0.2355925739,0.0445311628,0.4013054967,-0.1830300242,-0.342936188,0.1383527666,0.4372718632,0.3134897947,0.0098209782,0.0195712894,-0.0359986238,-0.3422741592,0.3768794537,-0.1361479759,0.1983713806,-0.0884066597,-0.001343661,-0.1373440027,0.3042966723,-0.0974787921,-0.0624615848,0.1574099958,0.0914994776,-0.0297624823,-0.0523535758,0.2131533921,0.4122335017,0.0303778984,0.1924372464,-0.013811538,0.4103766978,-0.2020863146,-0.0439700596,-0.0946270898,-0.0712999925,0.1517301798,-0.1395329535,-0.205512628,0.1627992392,-0.0285273809,0.1827243418,-0.5095719695,0.427361697,0.2366541922,-0.0472982004,0.0571585074,0.1110836491,-0.002362475,-0.0089515029,-0.0438567549,0.0467418507,0.0703111067,0.0665785149,-0.0663850605,0.1159160361,0.2043628991,0.0928915516,0.0580297783,-0.0351178497,0.0987340659,0.1005920395,-0.0032767237,-0.1262062788,0.3256442547,-0.2848223448,0.2149302065,0.1594939679,0.0941943005,0.0378742032,-0.0761301368,0.1714250296,0.3254772723,0.3470335901,0.2231839895,0.0314969905,0.0246453322,-0.0549993627,-0.0549255013,-0.0206400882,-0.1116272211,0.1299436986,0.1952213347,0.0732089356,0.0617158674,0.0350541733,-0.1838503778,-0.2796123028,0.0911719501,0.3946083188,-0.3553246856,0.1200247556,-0.039686121,0.0555186421,0.1301530749,0.0189663973,-0.4570333958,-0.2215031236,-0.2029660195,0.4210834503,-0.1194587797,0.1644146591,-0.0810745582,0.0509676039,0.2963262796,-0.1255412996,-0.259331286,0.104335174,0.0625970736,0.0425717719,0.1925008893,-0.5198504925,0.1900203079,0.1750378907,0.1842620224,-0.0083790794,-0.5007292032,0.0480755009,-0.1537050158,0.1523056179,-0.1455800682,0.2838713229,0.0492252409,-0.2909571528,-0.0210785959,-0.1372650564,-0.0680153519,-0.0453203991,-0.0619840845,-0.2530847788,-0.1954331398,-0.069822371,-0.2933194339,-0.4212827682,0.1225009188,0.371108681,0.2529848516,0.4407685995,0.1377109438,0.0921132565,0.2690692544,0.0994600207,-0.0383629762,-0.107421346,0.2451769412,-0.3104769289,-0.2563972175,0.0066135097,0.0111053875,0.2837287486,0.337452203,-0.3564890623,-0.1421988308,-0.312304467,-0.0041876035,-0.0811958015,0.1192051098,0.045397982,-0.0257927254,0.1307452321,-0.1638926119,0.0347888954,-0.0795593485,0.1042242572,-0.0728223324,0.0976093784,0.0997843817,0.2115115076,0.6869980693,-0.1291385889,0.1413394809,0.018308932,0.2151246369,0.4513024092,0.0434456803,-0.2967938781,0.0668857172,0.0837084129,-0.206221208,-0.1333539337,0.0458995514,-0.1497395933,0.0206685234,0.1178896278,-0.0471880026,-0.2259911448,0.1134366393,-0.0898844674,0.3340997994,0.187919274,0.0507023931,-0.4434088469,-0.1077162847,-0.0562812537,0.3804533184,0.0898376033,0.1508090645,-0.3076767027,-0.0904264003,-0.2749199867,0.0611348636,-0.0811378509,0.5575761199,-0.1852121502,0.3113373816,-0.0447095297,0.1809080094,0.4595344961,-0.2760755718,0.005393195,-0.1066814512,-0.281932354,-0.3433832824,-0.2390835285,-0.0859023929,-0.06393902,0.1136865318,0.1906843483,-0.0791468695,-0.0084775966,-0.1428744942,-0.2193469405,-0.1370996833,-0.2479740083,-0.1115662828,-0.2691839337,-0.4986168742,-0.1251070797,0.3895351589,-0.1092313603,-0.1225882396,0.1031751484,-0.2260031849,0.0671980307,0.1904061288,0.1915994138,0.0266411565,-0.009566904,-0.1999968439,0.2065167427,-0.5022637248,0.062692441,0.19732593,-0.1263116598,-0.3567327857,-0.016329661,0.1559303403,-0.1357991099,0.0637567714,-0.174356252,-0.0275263917,0.0254739188,-0.0504413955,-0.071854718,0.3104915917,-0.055890426,0.195660457,-0.1012471616,-0.4688993096,0.19685857,0.2237318605,-0.0757478699,0.1689159125,0.0943683311,-0.3989681005,0.4668301046,0.1583802253,0.5565386415,-0.359916836,0.0974431783,0.3196151853,0.067224808,0.6953794956,-0.6555452943,0.3161601126,-0.2079330236,-0.3527629972,-0.1493595541,-0.3817493916,0.181630671,0.0691904053,-0.0273417775,0.0941415504,-0.0852651075,0.3506408036,0.1797598153,0.2020633519,0.0884777084,-0.4448313415,0.1528819054,0.2294978499,-0.2720116377,0.3177681863,-0.1780319661,-0.0277537853,-0.2918524146,0.0254852902,-0.2747633159,0.1161500663,-0.2991321683,0.5190944672,0.1113141179,0.0464832298,0.2879340053,0.2445803732,0.2088550776,0.2620199323,-0.2269392312,0.0637965649,-0.130311653,-0.0931614488,-0.0392174013,0.0066591357,0.335981667,0.0596252605,-0.0585409142,-0.0665251166,-0.1696169376,-0.3609890938,0.0174033288,0.2294077426,0.2024174631,0.006458255,-0.0510129295,-0.0119281607,-0.3825114667,0.0726773962,0.1492640525,0.0109344935,-0.2534069121,0.2334099412,-0.045485761,-0.2722888291,0.0967666358,0.3703703582,0.1194974408,-0.1971344352,0.559756279,0.2043311149,-0.2898311615,-0.0391659811,-0.2040651292,0.0931312293,-0.4541436136,-0.0446451493,-0.4564749002,-0.4141295254,-0.0188431554,-0.4614747167,0.3459784389,-0.0074619078,-0.1736587584,-0.4079886973,-0.3153894246,-0.0964628011,-0.3273630738,-0.0001220054,-0.0069627934,-0.0182451922,-0.0710390657,0.1626549214,-0.259177804,-0.2372059524,-0.3444597125,0.006200335,0.3342401981,-0.2212559134,0.0436337963,-0.1810938865,0.0256804116,0.1466897726,-0.387232244,-0.1846114099,-0.2641800344,0.1107682213,0.2623438537,-0.3276488781,-0.0680264756,-0.3471541703,-0.0914151073,-0.1358476132,0.1620068699,0.26431036,-0.1182678416,0.1854373962,-0.1474619806,0.2718785405,-0.019502759,0.1773811579,-0.0268762168,0.0644539818,0.0384416021,0.3823345602,0.1326795816,0.0603220277,-0.6228336692,0.1132022813,0.0375844911,0.1517114937,0.3931129277,-0.3460859954,0.3375230134,-0.0362393744,0.0605528355,-0.152901575,-0.3420254588,-0.0691744462,0.1160460636,0.1642261893,0.0276592579,-0.096303314,0.0548701957,0.032719519,0.0987614095,0.2690493762,0.2619537115,-0.0406337827,-0.0124462042,0.3720317185,0.6341760755,-0.2661919594,0.2063339502,0.2476853728,-0.2334974557,0.1992752403,0.2232558876,-0.002031469,0.0001358897,0.1990372539,0.1524522156,-0.0540718548,0.0205246545,-0.1834712923,0.336091429,-0.2124366164,-0.1275319606,-0.0399635695,0.1181048453,0.2308101356,-0.0397053212,-0.3249217868,0.0261873621,-0.0104645165,-0.2011923939,0.1950682998,-0.1402630657,0.0443249755,0.0428522676,0.203814134,0.0132805081,-0.2877651155,0.1271378845,-0.4779744148,-0.3058706224,0.3640507758,-0.1340831816,0.0985853449,0.0616869591,0.2978540659,-0.2202402949,-0.196892485,0.0838604197,0.1072050706,-0.1116496697,-0.082573697,0.0934753716,0.5431833267,0.2893403471,0.1599655449,0.2631548941,0.0980713665,-0.0594070368,-0.0906574205,0.3076898456,-0.2290304899,0.1799558848,0.1197366491,0.0925129056,-0.2126800418,-0.2806037664,0.093236357,0.0901547596,-0.0003153438,0.2890575826,-0.2614881396,-0.2587867677,0.0748261362,-0.0476370901,-0.1973496825,0.1740469337,0.8346324563,-0.1961341202,0.2542154193,-0.0428735539,0.0120079843,-0.1633709967,0.385330081,0.3202114701,0.0207920764,-0.2639081478,0.2996233106,-0.2990526259,0.120006606,-0.1916530728,-0.0139936889,0.1818581522,0.1701564491,-0.0613947324,-0.0301435366,0.1060565785,-0.4536414444,-0.0028722938,0.0240874756,-0.2195233554,-0.1286708415,0.1216328964,0.0295880251,0.0657106042,-0.3943275809,0.0754953325,0.0724963695,0.0972357243,-0.3166572452,0.0166931972,0.2151388079,0.2526507676,0.2539834082,0.2079452425,0.3316455483,0.0703696534,0.0958230048,-0.3994336128,-0.349573344,-0.3445076644,0.0160421301,-0.1569499522,0.5965117812,-0.2879594266,-0.5061549544,-0.3620113134,0.0960142016,0.105117783,-0.2808544934,-0.2017903328,0.1834330857,-0.4025213122,0.3427912593,0.2535629272,-0.0980608538,0.244820863,0.3333668411,-0.3015104234,-0.3771386743,0.5819663405,-0.4455092549,-0.4940264225,0.2248333991,0.0352744348,-0.1100796983,-0.0267488882,-0.4583678842,0.2484734058,0.1376974732,-0.3479514122,-0.3160350621,0.0086388001,0.1702643633,0.2086642385,-0.0559566729,0.1456699669,-0.0756859928,-0.3460895717,0.118467927,-0.0319126099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1541","title":"connection issue while downloading data","comments":"could you tell me how I can avoid download, by pre-downloading the data first, put them in a folder so the code does not try to redownload? could you tell me the path to put the downloaded data, and how to do it? thanks\r\n@lhoestq ","body":"Hi\r\nI am running my codes on google cloud, and I am getting this error resulting in the failure of the codes when trying to download the data, could you assist me to solve this? also as a temporary solution, could you tell me how I can increase the number of retries and timeout to at least let the models run for now. thanks \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 361, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 269, in main\r\n add_prefix=False if training_args.train_adapters else True)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 306, in load_dataset\r\n return datasets.load_dataset('glue', 'cola', split=split)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return http_head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 403, in http_head\r\n url, proxies=proxies, headers=headers, cookies=cookies, allow_redirects=allow_redirects, timeout=timeout\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 542, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 655, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7f47db511e80>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n```","comment_length":45,"text":"connection issue while downloading data\nHi\r\nI am running my codes on google cloud, and I am getting this error resulting in the failure of the codes when trying to download the data, could you assist me to solve this? also as a temporary solution, could you tell me how I can increase the number of retries and timeout to at least let the models run for now. thanks \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 361, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 269, in main\r\n add_prefix=False if training_args.train_adapters else True)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 306, in load_dataset\r\n return datasets.load_dataset('glue', 'cola', split=split)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return http_head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 403, in http_head\r\n url, proxies=proxies, headers=headers, cookies=cookies, allow_redirects=allow_redirects, timeout=timeout\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 542, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 655, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7f47db511e80>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n```\ncould you tell me how I can avoid download, by pre-downloading the data first, put them in a folder so the code does not try to redownload? could you tell me the path to put the downloaded data, and how to do it? thanks\r\n@lhoestq ","embeddings":[-0.2348409742,-0.0264184847,-0.1684961021,0.3393738568,0.4391576052,-0.1407120377,0.1563979387,0.3509828448,-0.2138073742,0.0685087293,-0.0208971277,-0.1096859127,0.0704352781,0.536990881,-0.1515039504,-0.2388876677,-0.1877327263,0.0280617326,-0.2314403802,0.0420351811,-0.1295740306,0.2590333223,-0.0621024258,0.1730599999,0.0358291566,-0.1131441817,0.1226589531,0.1459693611,-0.2801611722,-0.3838599026,0.1542091817,0.1745480001,-0.0517693125,0.2062220573,-0.0001088944,0.1076903939,0.307946831,-0.1070469543,-0.2203330994,-0.1559585482,0.0400469601,0.0956171006,0.1934548467,-0.0438331701,-0.1365971714,0.434663713,0.0030512204,0.0237636659,0.4515209794,0.4471628666,0.2572606802,0.1820829362,0.3130951524,-0.1857011616,0.0406049751,-0.4782786965,0.0231450051,0.3593066633,0.262865752,-0.2473660111,0.2174865901,0.121585317,0.0678002685,-0.1103276387,0.0881747901,-0.0993106067,0.0909342691,-0.3702736497,0.0434348211,0.1779972017,0.6764319539,0.0420952141,-0.1345541775,0.2067839056,-0.0563103072,-0.202206403,0.3282392919,-0.0293237772,-0.1838054359,0.1603696644,-0.0458533056,-0.0682923198,-0.3555369377,0.2799238265,-0.0877680257,0.0889927521,0.0597958229,0.1290601194,-0.0441392027,-0.0850264803,0.1178375557,0.1198882908,-0.0155313239,-0.0139372749,-0.4044622779,-0.0382161178,-0.0493910164,-0.1649002284,0.0893062726,0.1916986555,0.386172086,-0.1656158417,0.1851050854,0.195149079,0.3030425608,0.0251457244,-0.1535613388,0.2802621722,0.485332787,0.358232379,0.1091378555,-0.1168528646,-0.0793488696,-0.4436267912,-0.1070655286,0.1931964457,0.1481758207,-0.1713293791,-0.2580232322,0.1838609725,-0.1554706842,0.2243509144,0.1914636046,0.4051600397,-0.2933759391,-0.1206312254,0.205328241,0.0556207299,-0.1346567273,-0.2475837469,-0.177614972,0.1270141453,0.0032318893,-0.1961533576,0.112988241,-0.1806070954,0.3181100488,-0.135443598,0.1654633135,-0.1547960788,-0.0770795271,-0.3587570786,-0.4265533388,0.342864722,-0.0515845008,-0.1006182507,0.1217820197,-0.0271780584,0.1396220624,-0.1921745092,-0.2594857812,-0.4317109585,-0.2148144096,0.3007487655,0.0068481578,-0.0460898913,-0.0009702805,-0.1341415346,0.1635116786,-0.39072451,0.0341531858,-0.3428096473,0.016795313,-0.3032230735,0.2672646046,0.5000346899,-0.4822481871,0.00113251,-0.0824278891,-0.2504619658,0.1914057732,0.2989630699,-0.24371849,0.1051212549,-0.1535787433,0.0678506047,0.3929648101,-0.3507179916,-0.4748890698,0.3805271983,-0.1470362097,-0.2910526693,-0.0379408561,0.064364858,0.2132600844,0.0781288892,0.1432295442,0.3018902838,-0.0660439059,-0.2896078825,-0.2765758634,-0.2182384431,0.0853394121,0.4488564432,0.3002042472,0.0365872607,0.303275764,-0.3245083392,-0.0462646373,0.0777715594,0.0712761208,0.3233808875,0.3449003398,0.1176356673,-0.0203552675,0.3761102557,-0.549890995,0.173010692,-0.4278171957,0.2051826715,-0.3752169609,0.1459923834,-0.1822524518,-0.0354656875,-0.1066942215,-0.0441699438,0.2470875382,-0.311899066,0.0967386141,0.1461656541,0.1997203827,0.4978630841,-0.2475582361,0.1445532143,-0.0677045658,0.5404032469,-0.3357788622,-0.2070756704,0.3209988177,-0.287119031,0.2573584616,-0.0747268349,-0.2176730037,0.3465878367,-0.2453757226,0.3602873087,0.0862292126,0.2653048635,0.1509155035,-0.1417225152,-0.0060187192,0.2202348262,0.3463457227,-0.1375163347,0.2603257,0.2672841549,0.0849163458,0.1005876735,0.1340834051,0.1851221174,0.290214479,0.055423528,-0.0036834383,0.1190352961,0.2799320817,0.1297676265,0.1541454643,-0.2395042926,-0.3208917081,0.0304419827,0.187446937,-0.1364736408,0.0050887545,0.095593743,0.3231284916,-0.2867569923,-0.1466925889,-0.0483226217,0.3849936426,-0.0679269657,-0.2181104124,0.3391186297,0.1424878389,-0.3514521122,0.3687647581,-0.0148211624,0.1530308723,0.4343566298,0.0197932944,-0.1724939644,-0.2720381916,-0.5410823822,0.172705397,0.5017488003,-0.3676473796,0.028473923,-0.0609808527,-0.1288914084,0.1670016497,0.1663617939,-0.2979949117,-0.302639693,0.0902758762,0.1386961788,-0.1017941013,0.0400378779,0.0291711539,0.2307890654,0.306073457,-0.0174340829,-0.1619219929,-0.293936342,-0.1299986094,0.0691568255,0.0803927258,-0.2212884873,0.1464639306,-0.0845270455,-0.0798363164,-0.1683973521,-0.1975148916,0.0711754262,0.0564199798,0.3333255649,0.3781257868,0.4529694021,-0.0027468328,-0.0465435982,0.2666623294,-0.201058805,-0.1093119979,0.0565123335,-0.0399197973,-0.0880911946,-0.1407164335,-0.4213065803,-0.6614133716,-0.2270002961,-0.018772373,-0.1047685444,0.202976346,0.2386614531,0.168742612,0.0918304995,0.2343424708,-0.0431979001,-0.3158206642,-0.4506627917,0.258895874,-0.1679927409,-0.2692593634,0.1085924357,0.3187679946,0.4288701117,-0.0719965622,-0.4718382359,-0.2920773327,-0.0196224786,0.131521821,-0.2031797171,-0.1841822714,0.328887105,-0.1671674997,-0.1352888495,0.0198434796,0.1486324817,0.3061666489,0.3233553171,0.1663650423,-0.0350661464,0.6165557504,-0.1657438129,0.3962852657,0.199892655,0.0060246023,0.2777493,-0.0252536535,-0.0822132155,-0.0938235968,-0.0131384525,0.3243084848,-0.1069097072,-0.2881341279,0.1485267729,-0.4488637149,0.0011849821,-0.4261848927,-0.0540868677,-0.4263953567,-0.1368150115,-0.0879227221,0.0861332342,0.242628172,-0.0162403937,-0.0412735716,0.0788463801,0.023074612,0.0587047637,0.3784459531,-0.0227943771,0.0318992026,0.2137587368,-0.2503308654,-0.3188165128,0.3502259552,0.0155304112,0.3136876523,-0.4380701184,0.2704896033,0.0670155808,-0.1444276571,0.5130234361,-0.4947693944,0.06634859,-0.0227241535,0.0036184967,0.2418506891,-0.285808742,-0.1581924707,-0.2280465662,0.0677400753,0.0179840717,-0.4608237743,0.0354513489,0.0661661029,-0.218942076,-0.2123014778,-0.0013238468,-0.060536705,-0.4850711524,-0.1683726758,0.0531065054,-0.0144173885,0.2999936044,-0.0625381768,-0.0462606139,0.2140882909,0.0120987156,0.0838258266,0.0084540462,0.0792363063,-0.2319022119,0.228739053,0.1301504523,-0.2059690356,0.4853291214,0.4809721112,0.2270241976,-0.16932863,0.1286838651,0.0329689682,0.1585355699,0.2000292391,-0.2111003399,0.2564331889,0.0856139883,0.36133641,0.1694755852,-0.1247670278,0.2746269405,-0.1899612546,-0.3355725408,-0.3043577671,0.0813347697,-0.0354939513,-0.0266097412,-0.0049335645,-0.1673104763,-0.1212773696,0.135221526,0.0712553859,0.9138278365,0.0209953655,0.0808267519,0.3139433861,-0.3383994699,0.1514257938,-0.3617417216,0.26392892,-0.3707275987,-0.2160765827,-0.0849531665,-0.0661517456,0.1525574923,0.0710653216,-0.3401987851,0.4297800064,-0.4042372406,0.1665910184,-0.1783964187,0.0652062669,0.0074960049,0.07184349,-0.2217148244,0.1884853393,-0.3539247513,0.3225484192,-0.3001503646,-0.2629715502,0.0268249363,-0.3401606977,-0.1810239106,0.1295182705,-0.5444908142,0.2351482064,-0.0711172298,-0.2762077749,0.3558166027,0.0676489621,0.1636480838,0.2084148973,-0.1242882609,0.3041799366,-0.0570820943,-0.0589084737,-0.096169509,0.2220032364,-0.0016728493,-0.2251731455,-0.3161198795,0.2273073941,-0.0993589237,-0.2042077333,-0.0531466566,-0.2324140817,-0.5021007061,-0.2361730188,0.0360459425,-0.1957389861,-0.3002686799,-0.1727752239,0.1554190964,0.1541719288,-0.1244439632,0.0952770934,0.2707772851,-0.1908460259,-0.1712900251,0.5571674109,0.1096606478,0.0587171726,0.6863898635,0.2083504498,-0.2349707782,-0.2707695961,-0.1917540431,0.0602437891,-0.4200970829,0.2383085191,-0.0531689897,0.1985812485,0.0607349649,0.4470109642,0.1206930503,0.2824609578,0.220891133,-0.5104773641,-0.2640198469,0.1446656436,-0.2137591541,0.258762449,-0.0870598257,0.265463382,0.0221863799,-0.0770348459,-0.3624051213,0.0866945982,-0.6442704201,-0.0238353256,0.2712138891,-0.5243382454,0.1725560874,0.0610062815,0.1990177333,-0.0153145837,-0.3990594745,-0.1956991553,0.0424407423,0.1484948993,0.0466089882,-0.1798482835,-0.0302988235,-0.186284095,-0.2644752264,0.3944396973,0.2726374865,-0.0735798106,0.172169894,-0.0297063962,-0.2637216151,0.0076816757,-0.070131883,0.0800346583,0.1399908066,0.1601366401,0.001673621,0.0794283599,-0.0097057475,-0.1773547381,-0.0999605581,0.1669428945,-0.0649405792,-0.0240601096,0.4716748893,-0.4384282827,-0.0445466749,-0.1281055659,0.2768379152,0.5457190871,-0.2084845155,0.0477337018,0.1734747142,0.2687778771,-0.1477644891,-0.1217350662,-0.2008672655,-0.0067314543,-0.0093575632,0.0009221805,0.1110901162,0.0385773964,-0.0286063161,0.0887344927,0.4709984958,0.0415340103,-0.0143642509,0.2882629037,0.0084490245,-0.0822439045,0.0564002693,0.1108721718,0.1142779812,0.0868126824,-0.3503553569,-0.2556738853,-0.1168416888,0.2562548518,-0.114198491,-0.4329715967,-0.4023192823,0.1917792559,-0.3171721399,-0.0475568436,-0.1428562403,0.2396066785,0.0299904849,-0.1882735789,-0.2870416045,0.307928741,-0.145800218,-0.1753404886,-0.2540184855,-0.2020064592,-0.343349874,0.1994732022,0.0569151305,0.1301012784,0.162605077,0.1688883007,-0.0059497016,-0.1788111031,-0.0957144424,0.2316546887,0.0032797323,-0.2984701097,0.2564791739,0.2168835551,0.0045155692,-0.0090244263,0.3461565375,0.6126069427,0.1562518477,-0.0602722093,0.1735377014,0.0777150393,-0.0152061107,0.1278726161,-0.0035640784,0.6596262455,0.128115505,0.2581997514,0.1150671393,-0.0949970931,0.2833233178,-0.0089173671,-0.0268701874,-0.1481945664,0.0960942581,0.0650024489,0.0760132596,-0.281614095,-0.0964308009,-0.5803738832,-0.0902787372,0.1286623925,0.1684319824,0.1892680675,-0.0685638338,0.0735563338,-0.2700653374,0.7045841217,0.3630319536,0.3230106831,-0.351567328,-0.0615594909,-0.2411384284,0.0012109979,-0.1764067858,0.2934834957,0.0082162879,0.2739601731,-0.0044442308,-0.056747701,0.3255103827,0.026548598,-0.3016248941,0.2652546465,-0.1580959409,0.0128404293,-0.1093129218,-0.103370443,0.0730004236,-0.064684324,0.120557338,-0.1764977127,0.1828400195,-0.2580232322,0.0219348073,-0.0369349495,-0.2141056657,0.4725675881,0.1146926656,0.4608373046,0.0695562586,-0.0373116173,-0.1260103285,-0.3512321711,0.0559628494,0.1321918815,0.2736310363,0.1281115115,0.2889132798,-0.042218063,-0.30183056,0.2375255674,-0.0987885967,0.0442330949,-0.0262362789,-0.0956498086,0.1066227853,-0.0608495921,0.031822186,0.0039755586,0.067622371,-0.1332004815,-0.5958692431,-0.4013921618,0.5534636378,0.1308204383,-0.0247525014,-0.03565979,0.2427159697,0.1207464859,0.2269695103,-0.5182474852,0.229388684,0.1252570897,-0.2398459911,-0.4368201494,0.2778034508,0.1732063442,0.0596353784,0.0441511795,-0.0060801459,-0.0349971466,-0.0930217728,-0.1105709374,-0.1281407624]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1541","title":"connection issue while downloading data","comments":"Does your instance have an internet connection ?\r\n\r\nIf you don't have an internet connection you'll need to have the dataset on the instance disk.\r\nTo do so first download the dataset on another machine using `load_dataset` and then you can save it in a folder using `my_dataset.save_to_disk(\"path\/to\/folder\")`. Once the folder is copied on your instance you can reload the dataset with `datasets.load_from_disk(\"path\/to\/folder\")`","body":"Hi\r\nI am running my codes on google cloud, and I am getting this error resulting in the failure of the codes when trying to download the data, could you assist me to solve this? also as a temporary solution, could you tell me how I can increase the number of retries and timeout to at least let the models run for now. thanks \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 361, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 269, in main\r\n add_prefix=False if training_args.train_adapters else True)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 306, in load_dataset\r\n return datasets.load_dataset('glue', 'cola', split=split)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return http_head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 403, in http_head\r\n url, proxies=proxies, headers=headers, cookies=cookies, allow_redirects=allow_redirects, timeout=timeout\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 542, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 655, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7f47db511e80>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n```","comment_length":63,"text":"connection issue while downloading data\nHi\r\nI am running my codes on google cloud, and I am getting this error resulting in the failure of the codes when trying to download the data, could you assist me to solve this? also as a temporary solution, could you tell me how I can increase the number of retries and timeout to at least let the models run for now. thanks \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 361, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 269, in main\r\n add_prefix=False if training_args.train_adapters else True)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 70, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/workdir\/seq2seq\/data\/tasks.py\", line 306, in load_dataset\r\n return datasets.load_dataset('glue', 'cola', split=split)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 589, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py\", line 263, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return http_head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py\", line 403, in http_head\r\n url, proxies=proxies, headers=headers, cookies=cookies, allow_redirects=allow_redirects, timeout=timeout\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 542, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/sessions.py\", line 655, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/requests\/adapters.py\", line 504, in send\r\n raise ConnectTimeout(e, request=request)\r\nrequests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/glue\/glue.py (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7f47db511e80>, 'Connection to s3.amazonaws.com timed out. (connect timeout=10)'))\r\n```\nDoes your instance have an internet connection ?\r\n\r\nIf you don't have an internet connection you'll need to have the dataset on the instance disk.\r\nTo do so first download the dataset on another machine using `load_dataset` and then you can save it in a folder using `my_dataset.save_to_disk(\"path\/to\/folder\")`. Once the folder is copied on your instance you can reload the dataset with `datasets.load_from_disk(\"path\/to\/folder\")`","embeddings":[-0.2348409742,-0.0264184847,-0.1684961021,0.3393738568,0.4391576052,-0.1407120377,0.1563979387,0.3509828448,-0.2138073742,0.0685087293,-0.0208971277,-0.1096859127,0.0704352781,0.536990881,-0.1515039504,-0.2388876677,-0.1877327263,0.0280617326,-0.2314403802,0.0420351811,-0.1295740306,0.2590333223,-0.0621024258,0.1730599999,0.0358291566,-0.1131441817,0.1226589531,0.1459693611,-0.2801611722,-0.3838599026,0.1542091817,0.1745480001,-0.0517693125,0.2062220573,-0.0001088944,0.1076903939,0.307946831,-0.1070469543,-0.2203330994,-0.1559585482,0.0400469601,0.0956171006,0.1934548467,-0.0438331701,-0.1365971714,0.434663713,0.0030512204,0.0237636659,0.4515209794,0.4471628666,0.2572606802,0.1820829362,0.3130951524,-0.1857011616,0.0406049751,-0.4782786965,0.0231450051,0.3593066633,0.262865752,-0.2473660111,0.2174865901,0.121585317,0.0678002685,-0.1103276387,0.0881747901,-0.0993106067,0.0909342691,-0.3702736497,0.0434348211,0.1779972017,0.6764319539,0.0420952141,-0.1345541775,0.2067839056,-0.0563103072,-0.202206403,0.3282392919,-0.0293237772,-0.1838054359,0.1603696644,-0.0458533056,-0.0682923198,-0.3555369377,0.2799238265,-0.0877680257,0.0889927521,0.0597958229,0.1290601194,-0.0441392027,-0.0850264803,0.1178375557,0.1198882908,-0.0155313239,-0.0139372749,-0.4044622779,-0.0382161178,-0.0493910164,-0.1649002284,0.0893062726,0.1916986555,0.386172086,-0.1656158417,0.1851050854,0.195149079,0.3030425608,0.0251457244,-0.1535613388,0.2802621722,0.485332787,0.358232379,0.1091378555,-0.1168528646,-0.0793488696,-0.4436267912,-0.1070655286,0.1931964457,0.1481758207,-0.1713293791,-0.2580232322,0.1838609725,-0.1554706842,0.2243509144,0.1914636046,0.4051600397,-0.2933759391,-0.1206312254,0.205328241,0.0556207299,-0.1346567273,-0.2475837469,-0.177614972,0.1270141453,0.0032318893,-0.1961533576,0.112988241,-0.1806070954,0.3181100488,-0.135443598,0.1654633135,-0.1547960788,-0.0770795271,-0.3587570786,-0.4265533388,0.342864722,-0.0515845008,-0.1006182507,0.1217820197,-0.0271780584,0.1396220624,-0.1921745092,-0.2594857812,-0.4317109585,-0.2148144096,0.3007487655,0.0068481578,-0.0460898913,-0.0009702805,-0.1341415346,0.1635116786,-0.39072451,0.0341531858,-0.3428096473,0.016795313,-0.3032230735,0.2672646046,0.5000346899,-0.4822481871,0.00113251,-0.0824278891,-0.2504619658,0.1914057732,0.2989630699,-0.24371849,0.1051212549,-0.1535787433,0.0678506047,0.3929648101,-0.3507179916,-0.4748890698,0.3805271983,-0.1470362097,-0.2910526693,-0.0379408561,0.064364858,0.2132600844,0.0781288892,0.1432295442,0.3018902838,-0.0660439059,-0.2896078825,-0.2765758634,-0.2182384431,0.0853394121,0.4488564432,0.3002042472,0.0365872607,0.303275764,-0.3245083392,-0.0462646373,0.0777715594,0.0712761208,0.3233808875,0.3449003398,0.1176356673,-0.0203552675,0.3761102557,-0.549890995,0.173010692,-0.4278171957,0.2051826715,-0.3752169609,0.1459923834,-0.1822524518,-0.0354656875,-0.1066942215,-0.0441699438,0.2470875382,-0.311899066,0.0967386141,0.1461656541,0.1997203827,0.4978630841,-0.2475582361,0.1445532143,-0.0677045658,0.5404032469,-0.3357788622,-0.2070756704,0.3209988177,-0.287119031,0.2573584616,-0.0747268349,-0.2176730037,0.3465878367,-0.2453757226,0.3602873087,0.0862292126,0.2653048635,0.1509155035,-0.1417225152,-0.0060187192,0.2202348262,0.3463457227,-0.1375163347,0.2603257,0.2672841549,0.0849163458,0.1005876735,0.1340834051,0.1851221174,0.290214479,0.055423528,-0.0036834383,0.1190352961,0.2799320817,0.1297676265,0.1541454643,-0.2395042926,-0.3208917081,0.0304419827,0.187446937,-0.1364736408,0.0050887545,0.095593743,0.3231284916,-0.2867569923,-0.1466925889,-0.0483226217,0.3849936426,-0.0679269657,-0.2181104124,0.3391186297,0.1424878389,-0.3514521122,0.3687647581,-0.0148211624,0.1530308723,0.4343566298,0.0197932944,-0.1724939644,-0.2720381916,-0.5410823822,0.172705397,0.5017488003,-0.3676473796,0.028473923,-0.0609808527,-0.1288914084,0.1670016497,0.1663617939,-0.2979949117,-0.302639693,0.0902758762,0.1386961788,-0.1017941013,0.0400378779,0.0291711539,0.2307890654,0.306073457,-0.0174340829,-0.1619219929,-0.293936342,-0.1299986094,0.0691568255,0.0803927258,-0.2212884873,0.1464639306,-0.0845270455,-0.0798363164,-0.1683973521,-0.1975148916,0.0711754262,0.0564199798,0.3333255649,0.3781257868,0.4529694021,-0.0027468328,-0.0465435982,0.2666623294,-0.201058805,-0.1093119979,0.0565123335,-0.0399197973,-0.0880911946,-0.1407164335,-0.4213065803,-0.6614133716,-0.2270002961,-0.018772373,-0.1047685444,0.202976346,0.2386614531,0.168742612,0.0918304995,0.2343424708,-0.0431979001,-0.3158206642,-0.4506627917,0.258895874,-0.1679927409,-0.2692593634,0.1085924357,0.3187679946,0.4288701117,-0.0719965622,-0.4718382359,-0.2920773327,-0.0196224786,0.131521821,-0.2031797171,-0.1841822714,0.328887105,-0.1671674997,-0.1352888495,0.0198434796,0.1486324817,0.3061666489,0.3233553171,0.1663650423,-0.0350661464,0.6165557504,-0.1657438129,0.3962852657,0.199892655,0.0060246023,0.2777493,-0.0252536535,-0.0822132155,-0.0938235968,-0.0131384525,0.3243084848,-0.1069097072,-0.2881341279,0.1485267729,-0.4488637149,0.0011849821,-0.4261848927,-0.0540868677,-0.4263953567,-0.1368150115,-0.0879227221,0.0861332342,0.242628172,-0.0162403937,-0.0412735716,0.0788463801,0.023074612,0.0587047637,0.3784459531,-0.0227943771,0.0318992026,0.2137587368,-0.2503308654,-0.3188165128,0.3502259552,0.0155304112,0.3136876523,-0.4380701184,0.2704896033,0.0670155808,-0.1444276571,0.5130234361,-0.4947693944,0.06634859,-0.0227241535,0.0036184967,0.2418506891,-0.285808742,-0.1581924707,-0.2280465662,0.0677400753,0.0179840717,-0.4608237743,0.0354513489,0.0661661029,-0.218942076,-0.2123014778,-0.0013238468,-0.060536705,-0.4850711524,-0.1683726758,0.0531065054,-0.0144173885,0.2999936044,-0.0625381768,-0.0462606139,0.2140882909,0.0120987156,0.0838258266,0.0084540462,0.0792363063,-0.2319022119,0.228739053,0.1301504523,-0.2059690356,0.4853291214,0.4809721112,0.2270241976,-0.16932863,0.1286838651,0.0329689682,0.1585355699,0.2000292391,-0.2111003399,0.2564331889,0.0856139883,0.36133641,0.1694755852,-0.1247670278,0.2746269405,-0.1899612546,-0.3355725408,-0.3043577671,0.0813347697,-0.0354939513,-0.0266097412,-0.0049335645,-0.1673104763,-0.1212773696,0.135221526,0.0712553859,0.9138278365,0.0209953655,0.0808267519,0.3139433861,-0.3383994699,0.1514257938,-0.3617417216,0.26392892,-0.3707275987,-0.2160765827,-0.0849531665,-0.0661517456,0.1525574923,0.0710653216,-0.3401987851,0.4297800064,-0.4042372406,0.1665910184,-0.1783964187,0.0652062669,0.0074960049,0.07184349,-0.2217148244,0.1884853393,-0.3539247513,0.3225484192,-0.3001503646,-0.2629715502,0.0268249363,-0.3401606977,-0.1810239106,0.1295182705,-0.5444908142,0.2351482064,-0.0711172298,-0.2762077749,0.3558166027,0.0676489621,0.1636480838,0.2084148973,-0.1242882609,0.3041799366,-0.0570820943,-0.0589084737,-0.096169509,0.2220032364,-0.0016728493,-0.2251731455,-0.3161198795,0.2273073941,-0.0993589237,-0.2042077333,-0.0531466566,-0.2324140817,-0.5021007061,-0.2361730188,0.0360459425,-0.1957389861,-0.3002686799,-0.1727752239,0.1554190964,0.1541719288,-0.1244439632,0.0952770934,0.2707772851,-0.1908460259,-0.1712900251,0.5571674109,0.1096606478,0.0587171726,0.6863898635,0.2083504498,-0.2349707782,-0.2707695961,-0.1917540431,0.0602437891,-0.4200970829,0.2383085191,-0.0531689897,0.1985812485,0.0607349649,0.4470109642,0.1206930503,0.2824609578,0.220891133,-0.5104773641,-0.2640198469,0.1446656436,-0.2137591541,0.258762449,-0.0870598257,0.265463382,0.0221863799,-0.0770348459,-0.3624051213,0.0866945982,-0.6442704201,-0.0238353256,0.2712138891,-0.5243382454,0.1725560874,0.0610062815,0.1990177333,-0.0153145837,-0.3990594745,-0.1956991553,0.0424407423,0.1484948993,0.0466089882,-0.1798482835,-0.0302988235,-0.186284095,-0.2644752264,0.3944396973,0.2726374865,-0.0735798106,0.172169894,-0.0297063962,-0.2637216151,0.0076816757,-0.070131883,0.0800346583,0.1399908066,0.1601366401,0.001673621,0.0794283599,-0.0097057475,-0.1773547381,-0.0999605581,0.1669428945,-0.0649405792,-0.0240601096,0.4716748893,-0.4384282827,-0.0445466749,-0.1281055659,0.2768379152,0.5457190871,-0.2084845155,0.0477337018,0.1734747142,0.2687778771,-0.1477644891,-0.1217350662,-0.2008672655,-0.0067314543,-0.0093575632,0.0009221805,0.1110901162,0.0385773964,-0.0286063161,0.0887344927,0.4709984958,0.0415340103,-0.0143642509,0.2882629037,0.0084490245,-0.0822439045,0.0564002693,0.1108721718,0.1142779812,0.0868126824,-0.3503553569,-0.2556738853,-0.1168416888,0.2562548518,-0.114198491,-0.4329715967,-0.4023192823,0.1917792559,-0.3171721399,-0.0475568436,-0.1428562403,0.2396066785,0.0299904849,-0.1882735789,-0.2870416045,0.307928741,-0.145800218,-0.1753404886,-0.2540184855,-0.2020064592,-0.343349874,0.1994732022,0.0569151305,0.1301012784,0.162605077,0.1688883007,-0.0059497016,-0.1788111031,-0.0957144424,0.2316546887,0.0032797323,-0.2984701097,0.2564791739,0.2168835551,0.0045155692,-0.0090244263,0.3461565375,0.6126069427,0.1562518477,-0.0602722093,0.1735377014,0.0777150393,-0.0152061107,0.1278726161,-0.0035640784,0.6596262455,0.128115505,0.2581997514,0.1150671393,-0.0949970931,0.2833233178,-0.0089173671,-0.0268701874,-0.1481945664,0.0960942581,0.0650024489,0.0760132596,-0.281614095,-0.0964308009,-0.5803738832,-0.0902787372,0.1286623925,0.1684319824,0.1892680675,-0.0685638338,0.0735563338,-0.2700653374,0.7045841217,0.3630319536,0.3230106831,-0.351567328,-0.0615594909,-0.2411384284,0.0012109979,-0.1764067858,0.2934834957,0.0082162879,0.2739601731,-0.0044442308,-0.056747701,0.3255103827,0.026548598,-0.3016248941,0.2652546465,-0.1580959409,0.0128404293,-0.1093129218,-0.103370443,0.0730004236,-0.064684324,0.120557338,-0.1764977127,0.1828400195,-0.2580232322,0.0219348073,-0.0369349495,-0.2141056657,0.4725675881,0.1146926656,0.4608373046,0.0695562586,-0.0373116173,-0.1260103285,-0.3512321711,0.0559628494,0.1321918815,0.2736310363,0.1281115115,0.2889132798,-0.042218063,-0.30183056,0.2375255674,-0.0987885967,0.0442330949,-0.0262362789,-0.0956498086,0.1066227853,-0.0608495921,0.031822186,0.0039755586,0.067622371,-0.1332004815,-0.5958692431,-0.4013921618,0.5534636378,0.1308204383,-0.0247525014,-0.03565979,0.2427159697,0.1207464859,0.2269695103,-0.5182474852,0.229388684,0.1252570897,-0.2398459911,-0.4368201494,0.2778034508,0.1732063442,0.0596353784,0.0441511795,-0.0060801459,-0.0349971466,-0.0930217728,-0.1105709374,-0.1281407624]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1514","title":"how to get all the options of a property in datasets ","comments":"In a dataset, labels correspond to the `ClassLabel` feature that has the `names` property that returns string represenation of the integer classes (or `num_classes` to get the number of different classes).","body":"Hi\r\ncould you tell me how I can get all unique options of a property of dataset?\r\nfor instance in case of boolq, if the user wants to know which unique labels it has, is there a way to access unique labels without getting all training data lables and then forming a set i mean? thanks","comment_length":31,"text":"how to get all the options of a property in datasets \nHi\r\ncould you tell me how I can get all unique options of a property of dataset?\r\nfor instance in case of boolq, if the user wants to know which unique labels it has, is there a way to access unique labels without getting all training data lables and then forming a set i mean? thanks\nIn a dataset, labels correspond to the `ClassLabel` feature that has the `names` property that returns string represenation of the integer classes (or `num_classes` to get the number of different classes).","embeddings":[-0.3508508503,-0.5276566148,-0.0328901485,0.3197903633,-0.0088524511,0.1327080727,0.0669215247,0.0627553165,-0.1925359815,0.2703049183,-0.2775320113,0.2990584671,-0.2552965283,0.1220931783,-0.0826280192,0.2548773885,-0.1904511899,0.1400496662,-0.09147764,-0.3434121013,-0.4865431786,-0.2970269322,-0.458103627,0.0147145968,-0.1832594424,-0.4297350049,-0.1620201468,-0.3819602132,-0.2474734187,-0.4416976869,-0.1803476363,0.3577092886,-0.0052560666,-0.1465243846,-0.0001120292,-0.1863492727,-0.0877825618,0.1195253432,-0.2179146856,-0.1170703992,-0.208410114,0.0798479095,0.0001605893,-0.2902169824,-0.2380115986,-0.1694646478,0.1302454025,-0.3067198396,-0.2818312943,0.0357822031,0.1677719653,-0.1580674499,-0.1943156421,0.0313468911,0.2919986844,0.3407337368,-0.1350935102,-0.1522563994,0.5592805743,0.2760846317,0.1553929299,0.1595231146,0.1639227718,-0.065399915,-0.1125929207,0.0601310804,-0.0172084011,-0.0513155013,0.1700330079,0.4939925671,0.8889746666,-0.0780139416,-0.555224359,-0.222472623,0.3864808679,-0.2008572221,-0.072684668,0.4160019159,-0.021555854,0.0239764266,-0.5343723893,-0.1047019809,-0.2602104545,0.0041006017,0.0951181576,0.4499118328,-0.0597374104,0.0480396077,0.0435346849,-0.1944862306,0.3572511375,0.0081248814,-0.0555034131,0.2131772637,-0.180676043,-0.1613455266,-0.0790354833,0.3115596771,0.4610341489,0.2292971611,-0.0018088394,-0.2025976181,-0.2045291215,0.0713876858,0.1316974461,0.1638012081,0.2778801322,0.4300645292,-0.090524666,-0.2591495514,-0.5156908035,-0.0436422862,-0.1627069116,0.5808580518,0.0987301767,0.0906743333,0.1428923756,-0.0987817422,0.0261776708,-0.1224273294,-0.5973362327,-0.1279473752,0.3690081239,0.0653480664,-0.1663045436,-0.2808000743,-0.1604654193,-0.3047666848,0.0698339045,-0.5663423538,-0.24556835,0.0211142674,-0.0447985902,0.0290551689,0.1502798051,0.0854732171,0.2885451913,0.2838986814,-0.0270618312,-0.0989094749,0.1929654628,-0.1295313537,0.073515676,-0.2468218803,-0.0755354539,-0.1078964993,-0.0959062129,-0.0326659568,-0.4610232115,-0.0920462757,0.0755554736,-0.3649215102,0.0379907489,0.1456158459,-0.1221426204,-0.1351405084,-0.234095186,0.5512105227,-0.1914868504,0.2397634685,0.131430462,-0.098591961,-0.2369734794,-0.0555083305,-0.1372011453,0.1909402013,-0.4381872118,-0.1115924791,0.1313178986,-0.1416656375,-0.1695653647,-0.0368135758,-0.3678408563,0.2495279312,-0.198859483,-0.0420435667,0.2656773329,0.0845837221,-0.4063229859,-0.0453337878,0.3962123394,-0.3897976279,0.2347951531,0.3541610837,0.5342296362,0.155602932,0.0012972433,-0.0104314303,-0.1367512047,-0.4786080718,0.2135133296,-0.2491564751,0.0555046983,-0.127150178,-0.2984972596,0.0940545499,0.2261770964,0.0374370739,0.1945082843,0.0008146421,-0.2040582448,0.3154881597,0.1634713411,0.3308371305,-0.140871644,-0.5227249265,-0.4604741931,0.2434815466,-0.1847603023,-0.1982658803,0.2715202868,-0.3839793801,0.2311086953,0.1584577709,0.0020248869,0.0699903071,0.1261916459,0.1834160835,0.0030262142,-0.2639645636,-0.3475838602,0.1739324182,-0.1795786023,-0.0859248266,-0.0596419498,-0.2804452181,0.3700778484,0.1967259049,-0.0314778648,0.0921080634,0.1555160284,0.1380991042,-0.0922473669,-0.0289695244,0.0102938851,-0.0202162024,0.1972585469,0.2296967357,0.1364670694,-0.2346062958,-0.1057953537,-0.0488548726,0.0834585354,0.0277628601,-0.2555168569,0.2668005228,-0.1034759358,0.2608881593,-0.0779846907,0.1833365858,0.0540484488,-0.0457128584,-0.0940100104,-0.1271288991,0.1622812301,-0.1134361476,0.4942336679,0.3270045817,-0.4405329525,0.1776648015,0.2505425811,-0.1502178609,0.0749085695,-0.2717399001,0.1221662238,0.121959269,0.4038164914,0.1231876388,0.2708517015,0.2157386839,0.0594006851,0.1966955811,-0.1114786863,0.1793223321,0.2144340128,0.0014691475,-0.2242041379,-0.3802837431,-0.0141311381,0.100709416,-0.032080587,-0.3933856487,0.2156148404,0.0209182147,-0.0542869009,-0.0450572558,-0.411819607,-0.0780554116,0.1107792109,-0.0728577822,-0.0198096782,-0.1078429073,0.1396297365,-0.089996092,0.0984444246,0.1316862255,-0.491967082,0.4692613482,-0.0703241378,0.0017363622,0.372818768,-0.1602380574,-0.317358017,0.0002689914,-0.3346790671,0.1726289988,0.4815166295,0.2825987041,0.2387952656,-0.4937898517,-0.0780083463,0.1100169718,0.1204605103,-0.0002427356,0.0564519651,-0.1279259473,-0.3248132467,0.3949852586,0.1154254153,0.2966908514,-0.0409834608,-0.1411793232,-0.1113710031,-0.2185892016,0.0414262228,-0.4756484032,-0.509113431,-0.0658925474,0.1078628227,-0.2145976722,0.2966996729,-0.0095512038,-0.4133852422,0.1338648647,0.0358329788,0.4845369756,-0.1738816351,-0.4474420249,0.0916540176,0.0210226625,0.0016238815,-0.2051277757,-0.200714916,0.0482299402,0.2721456885,-0.1739082634,-0.2373665273,-0.2317432761,0.0977828205,-0.1227781028,0.0688818768,0.1421620697,0.115131326,0.0753470436,-0.0545673594,-0.3315775394,0.1224296242,-0.2906486988,0.3100847304,0.3690904677,-0.0175023321,0.0651027709,0.2933643758,0.5332478881,-0.03682952,-0.0299321152,-0.4662999213,0.1133903638,0.4116409719,-0.454364121,-0.1225723252,0.3822774887,-0.1244691461,0.1978470683,0.2040620297,0.266358912,0.0306801796,0.1350489259,-0.2073510289,-0.2362264544,-0.3709763885,-0.2445937246,0.1783780754,-0.1887317002,0.0496794619,-0.1382540762,0.0604829304,-0.0667935833,-0.1624006629,-0.1108362451,0.0495775864,-0.6374655962,-0.2879964411,0.568338871,0.1488939971,0.1635270566,0.2337237,-0.1037099287,0.3242292106,0.4476892948,0.367551446,0.612274766,-0.1161097884,-0.0738219395,0.0576057844,-0.3960897923,-0.0004855786,-0.0506570041,-0.3637347221,-0.0429009236,-0.1238821745,0.4561243951,-0.057260219,-0.1372626275,-0.1734610498,0.3180663288,-0.2492275238,-0.3283992708,0.2676925659,-0.0034218025,0.0775881484,0.1097712517,-0.1344329566,0.0195838623,0.1759527326,-0.1185571924,-0.2799161971,0.034668047,0.4038734138,0.0199873056,-0.0403884239,-0.0033356722,0.3267429769,0.0396660082,0.0282007977,0.0530344211,0.5687279105,-0.1818538457,-0.4036194682,-0.3023500144,-0.0770729259,0.1083744913,-0.0863679275,-0.043961551,0.2125876546,-0.1236129403,-0.048100058,-0.1968957782,0.0352679491,0.0430281423,-0.0405988991,-0.2414752841,-0.3577510715,0.6491355896,0.074869208,-0.4582131803,0.2182478607,0.3473543525,-0.2755720019,-0.1457881927,0.1741806716,0.7046633959,0.0962943211,-0.2293825299,-0.259524554,0.0824263021,0.5855372548,0.2945800424,-0.0808784515,-0.0581787676,-0.336227566,-0.0368251763,-0.0157255046,0.0476300567,0.3330006301,0.0333092138,0.2259677052,0.0565386079,0.2718309462,0.1054319441,0.0779374093,-0.1414922029,0.0247663744,0.2523499429,0.1107065082,-0.1785781235,0.2311081588,0.1454965174,0.1079127118,0.2482429594,0.2883224189,-0.4448489547,-0.0917815417,-0.2850385308,0.2031097263,0.4126000702,0.0605051927,0.1747208685,0.217066884,0.4050023258,0.4423659444,-0.1711285561,0.0804444999,0.2109609395,0.4124330282,-0.2929895222,0.190917179,0.304148972,0.1589387804,-0.2294656634,0.2323849499,0.2367485315,0.3315257132,-0.491717279,0.026816342,0.1548868716,-0.3165367246,0.2781927288,-0.2163077891,0.0856903046,-0.0395413302,0.0612801872,0.1692212224,-0.1062486842,0.522518158,0.1892319024,-0.152371645,-0.1467066705,-0.181987375,0.2093459368,0.107500203,0.3382543027,-0.1614833325,-0.0882725716,-0.1336886287,0.4419907033,-0.3051992655,0.0378131941,0.0915211663,-0.0191944521,0.3036922812,-0.0070886174,0.0418284647,0.2091380209,0.2866424918,0.1424587518,-0.1397879273,-0.2146412432,0.1631245762,-0.0729961917,0.4103816748,0.0091073904,-0.1632106751,-0.2773331106,0.1723169237,-0.2245134562,0.0188675765,-0.275585562,0.1324807703,0.090389967,0.3165181279,-0.3393513262,0.1298603565,0.1120263413,-0.1530304849,0.1716082096,-0.1140109599,-0.3153392375,0.096324116,0.240869388,0.3602523506,-0.2150143236,-0.3067852259,0.03835335,-0.0774378181,-0.0664263144,0.1932676733,0.22584185,0.5054675937,0.1278170049,-0.2530193627,-0.4976937771,0.3159550726,0.0121904407,-0.0291647743,0.1767419726,0.154979229,0.0083202012,-0.2764666975,-0.6644695401,-0.020938931,-0.1368759274,-0.3500558138,-0.0383888073,0.3078997135,0.5481111407,0.3904308081,-0.1895967722,0.0405789874,0.0470066741,0.0424424335,0.4098187983,0.1500736922,-0.1682856083,-0.0705759823,-0.0809309334,0.0962925106,0.1378865987,0.427256912,-0.1457203776,0.2353934348,-0.0750229657,-0.4264290035,0.2358671576,-0.1205943078,0.0669380426,0.361348182,-0.0428197086,-0.2482873499,0.2149004787,0.6761207581,0.1334044337,0.1710664779,-0.2072856128,-0.03495593,0.026625758,0.022543408,0.6205440164,-0.1656975746,0.2913399041,0.1755541414,-0.2380995154,-0.07148505,-0.2300848812,0.1034753397,-0.3074240983,0.0097692935,0.1138041168,0.2060405016,0.1070099846,-0.1604612619,-0.3988549709,-0.0769754797,0.0729233995,0.0950030386,0.0571960174,-0.3662110269,0.0002713618,-0.0797158554,0.1783390343,-0.526930511,0.0930289775,0.331849277,0.0669465885,-0.206877768,0.0980242863,0.0045663919,0.0911098123,0.2443209141,0.1618795395,-0.1217998341,-0.0071256165,0.0803609043,0.1737852842,-0.1217867881,-0.2072756439,-0.0474165641,-0.1494160891,-0.0112560885,-0.2534238994,0.2478480339,-0.0191509034,0.0451480225,0.3839637935,-0.2259897888,0.2885332704,-0.2560207844,0.3926220834,0.3788807392,-0.1923001409,0.1074584275,0.0872374326,-0.1191890761,-0.415992558,0.5320670605,0.295224905,0.2267304808,0.1194133461,0.0369841494,-0.2051904798,-0.0487762019,0.4255405664,-0.0299868025,0.0174363907,0.4553385079,-0.2576415241,-0.1752889752,0.0695095956,-0.2328025103,0.4561752677,-0.0199789517,-0.1047270447,-0.1117710769,-0.0148594258,-0.1067886502,0.090809986,0.1928249449,-0.3207894564,-0.448330313,0.2457567006,0.0998100191,-0.004557143,-0.2362709045,0.2528703213,-0.2717032135,-0.0297287088,-0.13529405,0.3959843814,0.110851422,-0.046965342,-0.2839601934,0.1314224303,0.1565364599,0.1620585918,0.4353887439,0.1063679457,0.0358092748,-0.1216440648,0.0820017979,0.1073435023,0.1328491122,-0.0450523011,-0.171400547,0.194701761,-0.0374763012,-0.1566369385,-0.2194830775,-0.0624624304,0.0988234356,-0.2983466685,0.0126312431,0.3846224546,-0.0683057085,-0.1487444937,0.142832607,0.159328416,-0.190164566,0.5893236995,-0.066215992,-0.0594213717,0.1112929508,0.3183358908,0.2559436262,0.0200316049,-0.1685690433,-0.066998072,0.8531072736,-0.5415777564,-0.1140703186,0.1884545237,0.3835060894,0.1347115338,-0.0845647603,0.1515162289,-0.0027909456,-0.1205839366,0.3433541059,0.0145807164]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1514","title":"how to get all the options of a property in datasets ","comments":"I think the `features` attribute of the dataset object is what you are looking for:\r\n```\r\n>>> dataset.features\r\n{'sentence1': Value(dtype='string', id=None),\r\n 'sentence2': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=2, names=['not_equivalent', 'equivalent'], names_file=None, id=None),\r\n 'idx': Value(dtype='int32', id=None)\r\n}\r\n>>> dataset.features[\"label\"].names\r\n['not_equivalent', 'equivalent']\r\n```\r\n\r\nFor reference: https:\/\/huggingface.co\/docs\/datasets\/exploring.html","body":"Hi\r\ncould you tell me how I can get all unique options of a property of dataset?\r\nfor instance in case of boolq, if the user wants to know which unique labels it has, is there a way to access unique labels without getting all training data lables and then forming a set i mean? thanks","comment_length":42,"text":"how to get all the options of a property in datasets \nHi\r\ncould you tell me how I can get all unique options of a property of dataset?\r\nfor instance in case of boolq, if the user wants to know which unique labels it has, is there a way to access unique labels without getting all training data lables and then forming a set i mean? thanks\nI think the `features` attribute of the dataset object is what you are looking for:\r\n```\r\n>>> dataset.features\r\n{'sentence1': Value(dtype='string', id=None),\r\n 'sentence2': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=2, names=['not_equivalent', 'equivalent'], names_file=None, id=None),\r\n 'idx': Value(dtype='int32', id=None)\r\n}\r\n>>> dataset.features[\"label\"].names\r\n['not_equivalent', 'equivalent']\r\n```\r\n\r\nFor reference: https:\/\/huggingface.co\/docs\/datasets\/exploring.html","embeddings":[-0.2287427336,-0.7055018544,0.0021907324,0.3662371039,0.0509751029,0.1418879181,0.0267284531,0.0608512163,-0.1547129899,0.2690688372,-0.4087378383,0.2562119663,-0.2511259019,0.4104970396,-0.0019404892,0.1533517241,-0.1557465792,0.1335611194,-0.1536472142,-0.2861343324,-0.4006635249,-0.1075738594,-0.4153287113,0.0467985049,-0.1314859986,-0.3693220019,-0.2243348658,-0.2647702992,-0.3092392087,-0.4690890908,-0.1003361195,0.4063610137,-0.0191017445,-0.0985037237,-0.0001168394,-0.0966578275,-0.08514373,0.0530237928,-0.195506528,-0.1642183661,-0.0560531989,0.0330166593,0.0604154281,-0.3104787171,-0.2448531687,-0.2659602463,0.063094683,-0.2815676033,-0.0964833125,0.0834135562,0.1124279872,0.122878924,-0.1639308035,0.0523825288,0.2354943752,0.3953170478,-0.1452857256,-0.1424498409,0.3674568236,0.2163593024,0.2054813504,0.1736009419,0.1753453612,0.0197265018,0.0648147613,0.1097645015,-0.1048912853,-0.1086171567,0.1151849255,0.6177591681,0.7410152555,-0.0802450255,-0.612547338,-0.3371491432,0.265163064,-0.2187315077,-0.0421202779,0.3485732079,-0.0364700854,0.0196287557,-0.6166024804,-0.1469857246,-0.2099895477,0.0341836624,0.0878430605,0.3314309418,-0.2299049497,0.107028015,0.1994532347,-0.2394700199,0.1162179559,-0.0532050356,-0.160334751,0.2908092439,-0.2628741264,-0.2033747286,0.0598603152,0.3661664128,0.6192520261,0.2419130355,0.0272459202,-0.2319880575,-0.2885713279,0.0393390879,0.2170250863,0.2005427778,0.2487373948,0.3153836429,0.0167148542,-0.0292649139,-0.3435151577,-0.0755878538,-0.081313394,0.4947544038,-0.0487851687,0.0495078489,0.3242346644,-0.1381803751,0.0299292915,-0.1983109564,-0.5271268487,-0.095401071,0.3363982737,0.0749922022,-0.135993138,-0.2295092493,-0.0980432183,-0.1476190835,-0.0068864855,-0.4394291937,-0.2158149183,0.0091785034,0.0450276509,0.0627499819,0.1190925166,-0.1027028337,0.3910476565,0.2135322392,-0.0078681912,-0.1423297822,0.0254099611,-0.0637543425,0.0469572209,-0.1330059916,-0.0796604156,-0.0924228877,-0.0602715649,-0.1388602257,-0.460755825,-0.1905695349,0.1057441831,-0.4433948994,-0.028986495,0.097121276,-0.3224673867,-0.1415637434,-0.4366815984,0.6044265628,-0.2416905761,0.350016892,0.1379347593,0.017984312,-0.3908011019,-0.046833761,-0.0240165386,0.3638699055,-0.3369419873,-0.2405359,0.179744035,-0.2132232487,-0.3713692725,0.0807779655,-0.3572637141,0.2071002722,-0.2416037768,-0.0352623798,0.2858682573,0.0241249148,-0.331443131,0.0052651698,0.3683824539,-0.2202334404,0.1870829165,0.2353427708,0.5210514069,0.1221040264,-0.0985236019,0.1144925132,-0.0286738817,-0.3989966214,0.2246572524,-0.2133475691,0.0700233206,-0.145232439,-0.3617038429,0.1528536528,0.0965439603,-0.1897106767,0.2468205839,0.0048565585,-0.2093896568,0.3276514113,0.2211042941,0.3890689313,-0.1036694273,-0.5267808437,-0.6104329824,0.3131175041,0.0081161158,-0.1598109752,0.2112653106,-0.4473211765,0.1214880645,0.1537570804,-0.0376690403,-0.0609103665,0.0337237902,0.1989069581,-0.000506168,-0.1429089159,-0.4254854023,0.272126466,-0.1248877794,0.0505873822,-0.1900824159,-0.1255626231,0.4010555446,0.2112405896,0.0548178665,0.0028917007,0.2383458763,-0.010320913,-0.0924656242,0.06456732,-0.033551883,0.1128497869,0.2294691205,0.2720689774,0.2494476438,-0.2392003387,-0.1886919737,-0.0828718841,0.1910863072,0.018258119,-0.3324782252,0.3907878101,-0.0463843308,0.3015041649,-0.0743616149,0.1733385921,0.0144688906,-0.030882081,-0.1836419255,-0.1196030751,0.1472086459,-0.0567021742,0.3591713607,0.3756254017,-0.4823968112,0.1143965274,0.3670990467,-0.1303100586,-0.1336413771,-0.2044655383,0.0917301923,0.0992022231,0.3935653269,0.0503240153,0.2244663686,0.18861489,0.0470325425,0.2774070799,0.0206078831,0.2673248649,0.1599066257,0.0343190916,-0.1635104567,-0.3330229223,0.1298903525,0.067790553,-0.1989113688,-0.2625770271,0.1614474952,0.0883533433,-0.1697712094,0.0232223105,-0.3889833391,-0.0831419975,0.0753609017,-0.0631954893,-0.0673457459,-0.0844076052,0.0233297255,0.0526544079,0.1011319757,0.0764356479,-0.3535858989,0.392444104,-0.080462046,0.0092259431,0.2691920996,-0.0380385555,-0.2051478773,-0.0644438863,-0.3056202531,0.2183968872,0.4517791271,0.1415783763,0.1047298834,-0.5595074296,-0.0770265386,0.0939003527,0.0703040957,0.2038977295,0.1465806514,-0.0607118271,-0.2809715867,0.3665855527,0.1948534995,0.272393167,-0.1599522233,-0.1031919196,-0.0639202073,-0.1811526269,0.0288371444,-0.4219209254,-0.4215579629,-0.0707494467,0.3276548386,-0.1999600977,0.2525439262,-0.0054208282,-0.2850001454,0.0456598215,-0.1100159883,0.3793177009,-0.1815898418,-0.5349197388,0.1260641515,0.0119428365,-0.0821237266,-0.1585401595,-0.1444892585,-0.1269701123,0.1906870306,-0.2452409863,-0.3869085312,-0.3368161917,0.2491073757,-0.0860304832,0.1027117223,0.2500685155,0.0352764949,0.1486334801,-0.0380931124,-0.5036994219,0.1496802419,-0.0694487318,0.2438976616,0.3529318571,0.0748306811,0.0213789437,0.3612077236,0.5248460174,0.010234491,0.0907141715,-0.5891011357,0.2321334332,0.3434927762,-0.5600323677,-0.0695580468,0.3056469858,-0.0159543399,0.1589618474,0.2492464632,0.3992693722,-0.0480947755,0.120454289,-0.172225967,-0.2478890866,-0.4388039708,-0.1080582216,0.1012217328,-0.1443817317,0.0406376086,-0.0792639479,0.0588060804,-0.037686538,0.0627976507,-0.0023788128,0.1430988014,-0.6517040133,-0.2346108854,0.3950178921,0.1208132729,0.1292471588,0.1279474497,-0.0468617566,0.3369092643,0.4993626773,0.3162549138,0.7801806331,0.0260330662,-0.1128731668,-0.0357378535,-0.5820832253,-0.0472151376,-0.0763568953,-0.2561796904,0.0139712496,-0.132951349,0.597763896,-0.1236235201,-0.3026717007,-0.1031474546,0.3476691544,-0.3011568487,-0.3065579832,0.1594865918,-0.0749160349,-0.0214034449,0.0576465726,-0.1002711356,0.0472491421,0.1667348593,-0.138536647,-0.2284671813,0.0289722737,0.3732237518,0.063187778,-0.0361330807,0.0028208136,0.3145392835,0.1896236986,-0.0453872159,-0.0175956152,0.6283410788,-0.0717599019,-0.4086485803,-0.5035921335,0.0082438178,0.1751670688,-0.0565965213,-0.0774242729,0.1841045022,-0.0997278541,0.0023480055,-0.2769601345,0.1276538521,0.2268241644,0.0325884111,-0.1749291271,-0.2764024138,0.7093744874,0.0163552742,-0.3961479962,0.1268813461,0.6459160447,-0.2218303829,-0.2099890858,-0.0077350526,0.6521463394,0.0345272385,-0.1381374747,-0.1319208443,0.016617123,0.6354857683,0.3060136139,-0.0797815099,-0.0553421676,-0.309368372,-0.0541975349,-0.0453262776,0.0968269184,0.1455923915,0.0276085585,0.3248861134,0.1116391867,0.28083691,-0.0057665268,0.1019061282,-0.1137156785,-0.000339987,0.1069571301,0.0792500451,-0.2261991352,0.4505364001,0.1749609262,0.0701608285,0.3037329912,0.1329969913,-0.505233705,-0.228966549,-0.3203359544,0.0170860421,0.4922897518,-0.0518641435,0.1833898127,0.3056118488,0.3739116192,0.3710641861,-0.2079503238,0.004276786,0.0710262433,0.3601531386,-0.377346307,0.1411982626,0.2795593441,0.1284252405,-0.2771090865,0.1511865109,0.2461718619,0.3247496188,-0.3241710365,-0.0185615867,-0.0576823317,-0.3309652209,0.171170637,-0.1988740116,0.1035678312,-0.0958092883,0.0093035456,0.2089435458,-0.2372797728,0.3808008432,0.2046243399,-0.1594769955,-0.0842825398,-0.0232442357,0.12445461,0.1620335579,0.3429060578,-0.0653317794,-0.0637663081,-0.0858425125,0.3398194313,-0.2361554801,0.0771326646,0.1924656928,0.0288874097,0.2318576574,-0.1331951916,0.1773483455,0.2025745958,0.2019983977,0.2058050036,-0.2115893066,-0.2841511071,0.1425227672,-0.050017219,0.3632566333,-0.0028419262,-0.0485502705,-0.1789270788,0.1734474003,-0.1874321848,0.1480184942,-0.2888017297,0.0832714811,0.2191857249,0.262532711,-0.2179463506,0.0098096514,0.0409732796,-0.1540768147,0.0811912343,-0.0582363643,-0.4584405124,0.1241144091,0.1723208427,0.2276168168,-0.0747313946,-0.2618873715,0.0364910848,-0.1576020122,0.0799195915,0.3282824457,0.2694257498,0.5273607373,0.1394509673,-0.3102009594,-0.3670921922,0.2909505665,0.0024575258,-0.061373204,0.2268462628,0.1074510589,-0.0988561958,-0.2987816334,-0.5390706062,-0.031558957,-0.0642853528,-0.3648092747,-0.0317706317,0.3734407127,0.3671704829,0.3179034293,-0.0042953673,0.1758875698,0.0637014881,0.2036920786,0.3741742969,0.0896876976,-0.2198189944,-0.102780804,0.0574634671,0.100705348,0.1698739678,0.3659593463,-0.190816015,0.2398902625,-0.1170777157,-0.4271604121,0.1381697953,0.0387119576,0.1131403521,0.2040216625,-0.1371590048,-0.2020518333,0.301949501,0.6552519202,0.1903934032,0.1541807503,-0.2928784192,-0.1105856746,-0.0413479432,0.0894722119,0.7109920979,-0.1782188118,0.1906860471,0.0730054453,-0.1023827493,-0.1321014613,-0.1981784403,0.3249554336,-0.2759926617,-0.0703855082,0.0229763072,0.3993243277,0.1270033568,-0.3190347254,-0.1811656654,-0.0869245902,-0.0180620681,0.1969266236,-0.0030602058,-0.5139037371,0.1683699191,-0.1095313951,0.1652678996,-0.5792856216,0.0690940022,0.2280593812,0.1499494314,-0.1392500848,0.069512479,0.10275314,0.1299785972,0.2193460315,0.2406876683,-0.1010680348,-0.0733277723,0.0974663422,0.2790150642,-0.1512628794,-0.1866816133,-0.0302194059,-0.1159710437,-0.1603491753,-0.1726065725,0.2221860439,-0.0752003863,0.0711882785,0.3703368306,-0.3475040793,0.2174993008,-0.4432836175,0.3771455884,0.1857397854,-0.1287379712,-0.0365791023,0.0839677453,-0.0648310855,-0.3235014081,0.4672772884,0.3401453793,0.2324958742,0.08562316,0.0133122392,-0.2151787579,0.0051042978,0.5135836005,0.0474880263,0.0052232957,0.4486808479,-0.3901946247,-0.3382209539,0.0740337223,-0.2447770983,0.4777833223,0.0145509103,-0.1510381103,-0.1116698459,-0.0144168716,-0.0516223796,0.1202390641,0.1177463979,-0.2833405435,-0.4744014144,0.0319928266,0.0141881695,0.0148386108,-0.1975267529,0.2254750878,-0.2813904583,-0.0948786438,-0.1291196048,0.4669210017,-0.0685881376,-0.1200574562,-0.3512737155,0.1876307577,0.1087181345,0.225184992,0.4164706767,0.0754590407,0.0582730398,-0.1651848257,0.0757233948,0.0470713377,0.1906221658,-0.0528146885,-0.1015421897,0.1430753022,0.0815638006,-0.0329462662,-0.3104358315,-0.0263723619,0.2088253349,-0.2509791553,-0.0839005113,0.4362268746,0.0926924199,-0.2283170372,0.0312082302,0.0780953169,-0.1871642023,0.7877019048,-0.0610989667,-0.0189592913,0.1026379764,0.2547629774,0.5230182409,-0.1736035347,-0.3759199381,-0.0495071895,0.7507701516,-0.4973528981,-0.0682793409,0.2854039967,0.2297147661,0.0863565654,-0.0432122983,0.2874252796,0.0999551639,-0.0676929653,0.3426644504,-0.0046434547]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1478","title":"Inconsistent argument names.","comments":"Also for the `Accuracy` metric the `accuracy_score` method should have its args in the opposite order so `accuracy_score(predictions, references,,,)`.","body":"Just find it a wee bit odd that in the transformers library `predictions` are those made by the model:\r\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/trainer_utils.py#L51-L61\r\n\r\nWhile in many datasets metrics they are the ground truth labels:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/c3f53792a744ede18d748a1133b6597fdd2d8d18\/metrics\/accuracy\/accuracy.py#L31-L40\r\n\r\nDo you think predictions & references should be swapped? I'd be willing to do some refactoring here if you agree.","comment_length":19,"text":"Inconsistent argument names.\nJust find it a wee bit odd that in the transformers library `predictions` are those made by the model:\r\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/trainer_utils.py#L51-L61\r\n\r\nWhile in many datasets metrics they are the ground truth labels:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/c3f53792a744ede18d748a1133b6597fdd2d8d18\/metrics\/accuracy\/accuracy.py#L31-L40\r\n\r\nDo you think predictions & references should be swapped? I'd be willing to do some refactoring here if you agree.\nAlso for the `Accuracy` metric the `accuracy_score` method should have its args in the opposite order so `accuracy_score(predictions, references,,,)`.","embeddings":[0.2731813788,-0.3669873476,0.0105247274,0.2615756392,0.4064151645,-0.2207668722,0.2435519993,-0.032928925,-0.1960491836,0.0497912131,-0.1712593883,0.110400103,0.0239894912,0.1010815725,-0.013367855,-0.3983656764,0.0884952918,-0.1208069101,-0.2717051208,-0.2669148743,-0.0417980738,0.3241396844,-0.1340851486,0.1431115568,-0.3410859108,0.1306581944,0.0284888018,-0.2280272394,-0.165961355,-0.3751151264,0.2316489071,0.1122537553,-0.2013053298,0.5761030912,-0.0001090249,-0.0986292958,0.4503882229,0.0429984331,0.0979532823,-0.1389461309,0.2040537149,0.0040316018,0.3800654411,-0.2709213793,-0.1811490357,0.0455643758,-0.1698185354,0.3196667433,0.5733880997,0.2023383826,0.2764056921,0.4994625151,0.0267404187,-0.1820530742,-0.0167505071,0.3684441447,-0.1435261369,0.2194007188,-0.0225852784,0.1702657491,-0.2521291077,0.2552764118,0.2196521759,-0.079987891,0.4930861592,0.0928412154,0.0441884212,0.0080637373,-0.2042267025,0.0753111169,0.098084718,-0.2048594058,-0.5751951337,-0.1471262276,-0.1671092808,-0.2236678004,0.0386000089,-0.0628711656,-0.0837079734,-0.0743604824,-0.3173147142,0.0666565821,-0.2484065741,-0.1054518223,-0.0565710105,0.3778792024,-0.0511569865,0.0255960952,0.2035983503,-0.0259750765,0.0071266093,-0.0220366679,0.0458586961,0.2064843774,-0.2466256917,-0.2352699488,0.0978803411,0.3341541886,0.0830831751,-0.0604761653,-0.050915353,0.133210063,-0.0257559661,-0.0544035584,0.2670994103,0.4839111567,0.1295625716,0.1448741704,0.1113019958,0.1941492856,0.0352858938,0.0035420808,0.3475992978,-0.0608533807,0.1952458024,-0.0376509018,-0.2405876368,-0.4538148642,0.0780746415,-0.0467881598,0.1735589057,-0.2127188295,0.3083499372,0.2447806597,-0.4023400843,0.1304409951,0.1004997417,0.1457679123,-0.1560167223,0.0451313332,-0.2794976532,-0.0354639553,-0.5579994321,0.2265981734,-0.0161454044,0.1510367841,0.4049934447,0.0445037112,0.4226751924,-0.116089344,-0.2303492129,0.0658566579,0.1480440795,0.0695276335,-0.3464034498,-0.0202824045,0.0759003237,-0.1423481405,-0.1665621996,-0.1238345876,-0.1915263236,-0.2033853084,0.0824014917,0.2798464596,-0.3662946522,-0.0239016935,-0.1684896648,0.1240849942,0.032805346,-0.005280144,0.0329590328,-0.1358790845,-0.0155637451,-0.1864932179,0.279056251,0.2494159192,0.1906177998,-0.3905467391,0.1428603381,-0.1148656607,-0.1832009256,0.3622815013,-0.2361149937,-0.022618385,-0.0717781708,-0.0623340681,0.1299925148,-0.3559307456,-0.1099891067,0.0228483547,0.2844336927,0.1188532338,-0.0488430411,-0.0386589468,-0.0623236038,-0.24523516,0.0481164716,-0.133120209,0.1778406799,0.0154769756,-0.5007194281,-0.2730275989,-0.1432960182,0.097972095,0.1249622032,0.1286139488,-0.2446158379,0.1608840972,0.3392624557,-0.0987039059,-0.3262064457,0.1028044671,0.230823487,-0.1984651685,0.0530900508,0.0304738618,-0.4228440225,-0.0428095423,-0.2905333042,0.4193040133,0.2434006184,-0.2294383794,-0.1947897822,0.0939320251,-0.021168163,-0.1322898269,0.2385233492,-0.0693557039,0.1061500758,-0.0104011251,-0.1873704791,0.0234177262,-0.2420466542,0.2245829403,-0.2961933911,0.1095297933,-0.1609722972,-0.1399498731,0.0054599759,0.2729823589,0.1944747567,-0.1046131551,-0.1219412461,0.4222339094,0.02305682,0.215087831,-0.147157684,0.5913020968,0.3448300064,-0.1929062158,-0.0175549462,0.1209854111,-0.028342478,-0.2437526584,-0.3149511814,0.4001962543,-0.085733138,0.1407819837,0.3331146538,-0.1448747814,-0.0783898383,0.0735543594,-0.3948216438,-0.0057423394,0.2205551118,-0.161967352,0.2159049809,0.1072520688,-0.251216948,-0.1485432684,0.3663340807,-0.0666086376,-0.1350828707,0.2656883597,-0.0148499086,-0.0443814471,-0.1563386768,0.1039333865,0.4005812109,0.2021436393,-0.0817384645,0.1937269717,-0.1493364125,-0.1747569293,0.1467334032,0.3263317347,-0.3253344595,-0.0154467486,0.0735313147,0.1794380099,-0.3161493242,-0.056428954,-0.3818185329,0.0026232249,-0.1877410263,0.1746571958,0.0561089478,0.0912809893,-0.4408505857,-0.4194336236,-0.5085701346,-0.37309587,0.0834481493,-0.0125853643,-0.1168553606,0.2910577655,0.0102692302,0.2725937665,-0.2012775242,0.3011476696,-0.231815815,-0.1809814721,-0.1449623257,0.0649188533,-0.0005280174,-0.1864445359,0.3517945111,-0.1220342591,-0.0329939984,0.18036744,-0.7035816312,0.1564162225,-0.059343271,0.4357468486,0.2694195211,-0.0466163754,0.0246974695,-0.2172188014,0.472784698,-0.3604517877,0.1338212788,-0.0346326418,0.0381666161,-0.3561862409,-0.1619342715,-0.2841198146,-0.023173647,-0.221237272,0.1652252823,0.0890472084,0.0157862101,0.6461840868,-0.0748782903,-0.0009122782,-0.1254460216,0.2407364547,-0.1744231582,-0.3264143765,0.3574675024,-0.1375958025,-0.3368236721,-0.2210949808,0.0389712229,0.2903655171,-0.2625148296,-0.2988057137,-0.528434813,-0.1652683169,-0.2844391167,0.1087446958,0.2394404262,0.0225937013,0.0692526102,-0.0820013359,-0.082464464,-0.538801074,0.2279700786,0.0763646513,0.3385048807,-0.1704762876,0.0478972048,-0.0700189695,0.6143711209,0.2681386471,-0.3585923016,-0.0221968349,0.1695961803,0.0751660764,-0.1357620806,-0.3061009943,0.087646246,0.0679087117,-0.1022746265,0.3983539641,0.1747196317,0.0820224658,-0.2248396128,0.0019411991,-0.0715102181,-0.3880083561,0.0747024193,0.0903993845,0.0090203369,0.0017708013,0.3643043041,-0.2551608384,-0.0260043573,0.3872633278,0.1652656943,0.0230906401,0.0596461296,-0.3090392649,-0.2635150254,-0.3135920763,0.0789282471,-0.1081813723,0.0101663722,0.0114543382,-0.2042497545,0.0426032171,0.0314128064,0.3680113256,0.1540136486,-0.1849602014,-0.0439781025,-0.2024407834,-0.1204517931,-0.136641562,-0.0866008699,-0.2303578258,0.1441722512,0.5868595839,-0.4578207135,-0.263900131,0.1538532078,-0.0666936934,-0.2070074677,-0.1865273118,-0.1079442725,-0.0729937404,0.1400522441,-0.0854692981,-0.0651711673,-0.1613924056,-0.1461491436,0.1129131094,0.2235856354,0.1055683717,0.237654373,0.4936466217,0.2925416529,-0.277631402,0.0513539538,0.213357687,0.1717784256,-0.0803305358,0.2920035422,-0.2146088034,-0.122795105,-0.1981621385,-0.1105415598,0.214896962,0.2506068349,0.1510549635,0.0156921595,-0.006319141,0.2994708717,-0.1539943218,0.0385589488,0.4508613348,0.039573621,-0.1558486521,-0.1706232131,0.6738684177,-0.060156472,-0.2831735909,0.1638411134,0.4449319243,-0.2272998691,0.5194252729,0.2175828218,1.0889228582,0.1069160402,0.0494518317,0.31581375,-0.2730346024,0.4263893664,0.0115113314,0.0828661323,-0.3455776572,-0.1167231947,0.0976561457,-0.0800307989,-0.0665786564,0.0281959325,-0.2774503529,0.331680119,-0.0596056953,0.2070012093,-0.153654471,0.0992837176,0.2525378764,-0.1773385257,-0.3854656518,0.2241417766,-0.0717903003,0.2167675793,-0.1630468816,-0.1830930412,-0.0060108448,-0.2771737576,-0.1023511887,0.0304844938,-0.0059706657,-0.1198209375,0.6014568806,0.0198700819,0.1522496343,0.1620300114,0.4937833846,0.1199937165,-0.2232742459,0.2148794234,-0.3081080019,0.2238098234,-0.0079232054,0.1512021422,0.1854408234,-0.0896100327,-0.1417721361,-0.0066043371,0.0774523914,-0.1209677234,-0.336794436,-0.2036021054,0.0267844144,-0.3191963136,0.1342593431,-0.0585564747,-0.1731138676,-0.0645457953,0.2134260088,0.0138396993,-0.1995619386,0.464274466,0.3106339276,-0.0782203451,-0.193254441,0.2929206192,-0.0322320797,-0.2332820594,0.0905383304,0.0629579276,-0.0986728594,-0.1577107906,-0.442443192,0.0355701596,-0.4661077559,0.1386987269,-0.1761010736,-0.3279391229,-0.0027026327,0.4734790027,0.5512046218,0.0011475972,-0.0902011618,-0.0343905278,-0.3842211366,0.2968620956,0.2703802586,0.2962860465,-0.2652155459,0.2663923204,0.0294160936,0.3402479589,-0.4273926914,-0.2028061002,-0.2410658896,0.0717972741,0.0532618873,-0.3644260466,0.1708068699,-0.0033058114,0.1121911034,-0.0753499344,-0.2686131597,-0.179839462,-0.2429631352,0.0790381283,0.0719847232,-0.0633974969,-0.084333688,0.2383141965,0.123514086,-0.2001894712,0.0944143757,0.2596897781,0.1465177983,0.5000568032,-0.0304101072,0.2394499183,0.060353633,-0.0846126452,0.4427678883,0.0116532817,0.3352669179,-0.0743968189,-0.2876386046,-0.2415056676,0.1656606644,0.3212962449,0.0125831245,0.194329977,0.2605950534,-0.00973057,-0.1582778394,0.2121722251,0.3628947735,0.3294598162,-0.0170120616,-0.1882125139,-0.1788672954,0.2839373648,-0.182759881,-0.1095077991,0.0024615137,0.2359223962,0.0959534049,0.2922531068,0.023467293,-0.0353057198,0.2685286403,0.1122431234,0.5112106204,-0.3401263356,0.2363712341,0.152733624,-0.0366102159,0.0856956169,0.4524398148,-0.0876653641,0.3393719196,-0.0827498436,-0.2079927027,0.2195245922,-0.4933938682,-0.0465504751,0.2815237045,0.3500543237,-0.2611863911,0.194048509,0.1739081293,0.147628963,-0.0357834958,0.4031772614,-0.0133874789,-0.4535732865,-0.4093821049,-0.0276817326,-0.0757883638,-0.2337372005,-0.1527151018,-0.2250807732,-0.1819275022,-0.12741445,-0.0998891369,-0.1553858817,0.5217834115,-0.0358606614,0.0308542512,-0.064052552,-0.3481558859,0.1288260221,0.4875507355,-0.0767111778,0.2586067915,0.4656591117,-0.0471522473,0.0578512624,0.3424310386,0.2485369146,0.1830042005,-0.0542474911,0.054702051,0.0416977964,-0.0485284552,0.0732840598,0.3361122012,0.0001353821,-0.008693656,0.1509361416,0.2178554833,-0.2653509676,0.315387547,-0.5025444627,0.1285457164,-0.2653689981,0.7372075319,-0.0105512263,-0.2024535835,-0.1395481527,-0.1946981996,-0.3662175536,0.0185859278,0.055879619,0.0265462697,0.0848879367,-0.1573365033,0.1181368455,0.1454439759,0.5106711984,0.3234314919,0.225402832,-0.391756624,0.0454440042,-0.3647670746,0.0143846637,0.444429785,0.3885421157,-0.1138791293,-0.0852999538,0.0599838682,0.0212448481,0.1439101398,-0.3385232985,-0.0759206042,-0.1304984242,-0.3831557035,0.0957409963,-0.3658342063,-0.2712197006,0.246621877,0.1024474204,-0.0437813364,-0.0287931673,0.0867957622,-0.0375148356,0.1586443782,0.1548332721,0.2553787529,0.101979278,0.172439009,0.3268676102,0.0141367614,-0.1190444827,0.0525727123,-0.1921419352,-0.0984352529,-0.0220793374,-0.3004070818,0.3137197793,0.0419788733,-0.1632678807,-0.3625546396,0.0762948841,0.1909289062,-0.3590405285,-0.34429878,0.1590275168,-0.0391314588,-0.2292040139,0.268091023,0.5187247992,0.0958219767,0.2948406637,-0.235216409,-0.1957510561,0.663007319,-0.3643350899,0.0121414866,-0.0396223217,0.2570455074,-0.0394446664,-0.2340297699,-0.9655610323,-0.0292498861,0.0918550938,-0.2649493515,-0.219156146,0.2893411815,-0.0328868218,-0.1080773771,-0.0495715514,0.5406030416,0.1458814442,-0.1375697404,0.0589522533,-0.0361502059]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1478","title":"Inconsistent argument names.","comments":"Thanks for pointing this out ! \ud83d\udd75\ud83c\udffb \r\nPredictions and references should indeed be swapped in the docstring.\r\nHowever, the call to `accuracy_score` should not be changed, it [signature](https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.accuracy_score.html#sklearn.metrics.accuracy_score) being:\r\n```\r\nsklearn.metrics.accuracy_score(y_true, y_pred, *, normalize=True, sample_weight=None)\r\n```\r\n\r\nFeel free to open a PR if you want to fix this :)","body":"Just find it a wee bit odd that in the transformers library `predictions` are those made by the model:\r\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/trainer_utils.py#L51-L61\r\n\r\nWhile in many datasets metrics they are the ground truth labels:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/c3f53792a744ede18d748a1133b6597fdd2d8d18\/metrics\/accuracy\/accuracy.py#L31-L40\r\n\r\nDo you think predictions & references should be swapped? I'd be willing to do some refactoring here if you agree.","comment_length":49,"text":"Inconsistent argument names.\nJust find it a wee bit odd that in the transformers library `predictions` are those made by the model:\r\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/master\/src\/transformers\/trainer_utils.py#L51-L61\r\n\r\nWhile in many datasets metrics they are the ground truth labels:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/c3f53792a744ede18d748a1133b6597fdd2d8d18\/metrics\/accuracy\/accuracy.py#L31-L40\r\n\r\nDo you think predictions & references should be swapped? I'd be willing to do some refactoring here if you agree.\nThanks for pointing this out ! \ud83d\udd75\ud83c\udffb \r\nPredictions and references should indeed be swapped in the docstring.\r\nHowever, the call to `accuracy_score` should not be changed, it [signature](https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.metrics.accuracy_score.html#sklearn.metrics.accuracy_score) being:\r\n```\r\nsklearn.metrics.accuracy_score(y_true, y_pred, *, normalize=True, sample_weight=None)\r\n```\r\n\r\nFeel free to open a PR if you want to fix this :)","embeddings":[0.256931603,-0.439550519,-0.0103000738,0.1632704586,0.4465913773,-0.2115409821,0.1434418708,-0.0334956869,0.0414287522,0.0814790502,-0.0405135155,0.1803880632,-0.0548593551,0.1349070221,0.0187775791,-0.4100299776,0.1643256396,-0.1507949233,-0.2882173657,-0.2812667489,-0.0857049823,0.316467911,-0.1572450399,0.100552015,-0.4445046186,0.0573824234,0.0731789321,-0.2022777051,-0.1855972856,-0.3112768531,0.2928377688,0.0348808169,-0.1435786635,0.5684481263,-0.0001043543,-0.0801726133,0.5083734989,-0.0048199133,0.0420004018,-0.1177941635,0.1701010466,0.0062208744,0.2916088998,-0.3434048891,-0.193949163,0.2135120183,-0.1863261312,0.1779732704,0.4698330164,0.2248388529,0.3030235767,0.440297097,0.064599596,-0.1933786869,-0.0585102588,0.2556340694,-0.0898000821,0.1844703853,-0.1508141309,0.1044708788,-0.2257791758,0.3445389271,0.1872504205,-0.1968569905,0.568767786,0.1029587388,0.079935573,-0.1289537251,-0.0912590697,0.0444573052,0.0709270686,-0.2305466086,-0.4193950593,-0.1462265998,-0.1154459268,-0.2439677417,0.0441442579,-0.003800323,-0.1105913967,-0.1083739847,-0.3040143251,0.004342108,-0.2384107411,-0.0871796086,-0.1200482175,0.4932813644,-0.044805076,0.0596698895,0.2030075341,-0.0181639157,0.0755194277,-0.0440237559,-0.0175675377,0.1420896649,-0.2674137056,-0.2524872422,-0.0743403137,0.2494470328,0.0487836823,-0.1247599721,0.018226834,0.117621921,-0.0264747404,-0.0782871321,0.2774725556,0.5153080821,0.0807188824,0.2178994119,0.0439514443,0.2514868379,0.0921024159,-0.0154798897,0.2969269753,-0.1088856757,0.1562731415,-0.0452339128,-0.1819580346,-0.2827305198,0.0063982559,0.0453011133,0.0737396255,-0.0733467042,0.2071209252,0.1893324256,-0.3304433525,0.2138773203,0.2055527419,0.1191916168,-0.0907178596,0.0011589869,-0.3113301098,0.0857705697,-0.5817627311,0.2180779576,0.0201008264,0.2139526904,0.3889667094,0.0482910089,0.3449585438,-0.17496714,-0.2742282748,0.0641731992,0.1296645105,0.0910407528,-0.371928215,0.0573801994,0.0904335752,-0.0964154452,-0.1146554649,-0.002281853,-0.2376356274,-0.1269274205,0.1172142848,0.3066257536,-0.3674982786,-0.0712011233,-0.1644512117,0.1566270888,0.0240525641,0.1329238862,0.0971584097,-0.1436061561,-0.071237579,-0.1228671968,0.3704476058,0.1328169107,0.2067985237,-0.316463232,0.1525969803,-0.0030898147,-0.2281825244,0.2791340053,-0.2480566353,0.0794885308,-0.0573107414,-0.0417285301,0.0855614096,-0.4337061048,-0.1464656889,-0.023914637,0.1614613533,0.017851403,-0.1277337372,-0.0667634234,-0.0342098922,-0.1476313025,-0.0077677043,-0.033803694,0.1741591841,0.0677702725,-0.6005351543,-0.3513386846,-0.1266983151,-0.0123309819,0.2001012862,0.1121135131,-0.2058453709,0.3587814867,0.3848200142,-0.060632389,-0.2027910352,0.2178660184,0.3127883673,-0.1990872622,0.0022054054,0.0562240593,-0.4301520288,-0.0479111075,-0.2720617652,0.4051241875,0.1613992751,-0.2514888644,-0.2848739326,0.1218406782,-0.0194588695,-0.2196539789,0.3248638809,-0.0722801909,0.2226168513,-0.011103617,-0.0927924961,-0.0237485468,-0.2221786082,0.1867496073,-0.3372329175,0.2868709564,-0.2119273245,-0.0809812248,0.0009326966,0.2739035487,0.2689195871,-0.0639837906,-0.0738747865,0.4159818292,-0.1009960026,0.1458866596,-0.1311028749,0.3805533051,0.3150107861,-0.2862721682,-0.0108403536,0.0776697546,-0.0136617944,-0.1669187099,-0.2959807217,0.2807634771,-0.2125167102,0.1141026914,0.1887623817,-0.0884110928,-0.1432165056,0.043011412,-0.290368408,-0.078977339,0.0140888663,-0.2077496052,0.10352391,0.0311889369,-0.1021187678,-0.1167837977,0.4021645486,-0.1101825386,-0.1710701734,0.1190947667,-0.1837718785,-0.0675297305,-0.1294608265,0.1501441598,0.3384487033,0.1897508353,-0.1371939331,0.1450781077,-0.1647858769,-0.1246198788,0.15455392,0.3616389632,-0.2243347466,0.0703088567,0.1447581798,0.1423943937,-0.3193593323,0.0456405729,-0.2093979418,0.0649728402,-0.1356942207,0.1865713,0.0880624801,0.2206481546,-0.4930603802,-0.3447504044,-0.3960666955,-0.4616573751,0.0575411543,0.0512602292,-0.0770540014,0.270311296,0.0260869879,0.1901890337,-0.0972842425,0.1863812506,-0.1831385791,-0.2294822335,-0.1294647306,0.1167707443,-0.0222530197,-0.1021947637,0.2585438192,-0.0462663807,-0.0047253119,0.089859955,-0.6740859747,0.0580762029,-0.1147524938,0.3619232178,0.2027481794,-0.0060335714,-0.077190198,-0.3147086501,0.4627213478,-0.3848543167,0.0855311975,0.0152146835,0.1265775859,-0.2489892691,-0.1569399685,-0.282684654,0.0228785723,-0.1770999432,0.0959240124,0.0997364223,-0.1777557433,0.6479942203,0.0956636742,-0.0052459999,-0.2033057064,0.1182725951,-0.2391469032,-0.2335713953,0.356430769,-0.0545614548,-0.4124352932,-0.2016922235,-0.0423714966,0.3321449161,-0.2471004725,-0.3998714387,-0.4859279096,-0.136632219,-0.1886876822,0.0765183568,0.188956067,0.041314967,0.0390404686,-0.0667268336,-0.1327357292,-0.4137021601,0.2914661467,0.0208934862,0.2645951807,-0.1987661272,0.0895240083,-0.0390896089,0.7449858189,0.2926265299,-0.3071352839,-0.0790222287,0.2613869905,0.1781193465,-0.0992988944,-0.2288461924,0.0927711129,0.0014606455,-0.0256205313,0.3370515108,0.0153883481,0.1114616692,-0.1759306788,0.0077854451,-0.0346546136,-0.4441078603,0.0643239543,0.1354659498,-0.0013579817,0.0173564255,0.3422452211,-0.2973372638,-0.1130083501,0.4060263336,0.0941159353,-0.0518892705,0.0938118026,-0.3671156466,-0.3119596541,-0.3852484226,-0.0452662073,-0.1475196034,0.0864070803,0.0665900931,-0.1999535412,0.1263783872,0.0075904569,0.3109396994,0.1741897166,-0.0744166896,-0.0606325567,-0.1424402744,-0.1088782176,-0.1683872938,-0.1783252954,-0.1928320974,0.0903369039,0.4623688459,-0.4577736259,-0.2010156214,0.2234459668,-0.012054096,-0.2121215165,-0.1751923412,-0.0460777991,-0.0623125583,0.0899677947,-0.0738728493,0.1063747555,-0.000462584,-0.2603627741,0.1670318395,0.1651109457,0.0645127818,0.1582478136,0.3301131129,0.3451453447,-0.1230527759,0.0834306628,0.1829057634,0.2164638489,-0.0981518552,0.361376822,-0.1825079769,-0.1572299004,-0.137950778,-0.0892177671,0.0655862987,0.2963197231,0.1103030741,0.069054015,0.0225031488,0.2366209626,-0.1196970344,0.0871279985,0.359415859,0.0525657088,-0.1456014365,-0.0958312452,0.7624723315,-0.0418551862,-0.281208992,0.0825654492,0.3479245305,-0.3535614312,0.4915427268,0.3038066328,0.98012954,0.1556286812,0.1015792489,0.2742386162,-0.2376195639,0.5145319104,-0.1806176603,0.1069838405,-0.3739768863,-0.0533205122,0.1637927592,-0.0009505122,-0.1805530638,0.0130477687,-0.2989775538,0.2325779349,-0.0967232361,0.2205069661,-0.1810116619,0.1618667841,0.3003582954,-0.1440810412,-0.26390481,0.2612044215,-0.0236060191,0.1913763881,-0.113020055,-0.2304831147,-0.1182729304,-0.2138363421,-0.0463454276,0.0372554697,-0.0597796999,-0.080876708,0.4814066887,-0.0448383242,0.1836674809,0.1811635643,0.5718559623,0.1594437212,-0.2871662974,0.2436348498,-0.2239193469,0.1114661247,-0.0298064705,0.1648176163,0.1402430534,-0.0821546316,-0.2135183364,0.1373218447,-0.007769953,-0.0537045486,-0.4129833877,-0.1863075793,0.0327015296,-0.3877943456,0.0754231066,0.00815548,-0.0997478142,-0.0771082491,0.2453721315,-0.0471380018,-0.1394182146,0.270265907,0.4160871208,-0.0851860791,-0.1407625526,0.3222423792,0.0007677799,-0.2211221009,0.1694956422,-0.056291841,-0.1175068989,-0.1953238994,-0.4273698032,0.0388566628,-0.5343936682,0.0756664202,-0.031113755,-0.2663197219,0.1231165454,0.5384747386,0.5222584009,-0.0018293555,-0.0101204747,-0.030452298,-0.3189579546,0.3316340148,0.0909774452,0.2659347355,-0.2274134159,0.3254045844,-0.0325057209,0.3130789697,-0.4708201587,-0.1727944314,-0.1393198371,0.1091416627,0.1204730868,-0.4166150987,0.0991415083,0.0321758427,0.1179071441,-0.0295724273,-0.2310783863,-0.2092258036,-0.1939097643,0.0637802333,0.1857426465,-0.0329041891,-0.1109312996,0.2121653259,0.1622433811,-0.1922196448,0.1191680655,0.2229339331,0.2008503377,0.433079809,-0.0359401181,0.3352210522,0.0477251969,-0.1254093498,0.3776908219,-0.0110147614,0.2012267262,-0.0752759874,-0.0959924981,-0.1424960345,0.0520465784,0.2900220156,-0.1264625192,0.2532617748,0.230369553,-0.1209504828,-0.1893377751,0.234085843,0.322961241,0.2558214962,-0.1402049512,-0.2384914756,-0.1198649928,0.3315635026,-0.1774673164,-0.089236483,0.0248329956,0.1517831385,0.1211133972,0.2600542009,0.0938527957,-0.2307929993,0.3086584806,0.2216135859,0.4670119584,-0.3117228746,0.3354870677,0.2046530843,-0.0056985314,0.0224104133,0.4360669553,-0.0458877832,0.278954953,0.0891938061,-0.1828472465,0.1972232312,-0.4349935651,-0.0839253068,0.2375442088,0.3870135546,-0.1375892162,0.1356675774,0.0856374949,0.1301252395,-0.1282192916,0.3575029075,-0.0262750629,-0.4033465385,-0.4696719944,-0.0326711461,-0.1958782226,-0.2381101549,-0.1704584807,-0.1990292966,-0.1440724283,-0.1020294055,-0.0963798389,-0.1623236537,0.4438832998,-0.0886595324,-0.1972816736,0.0583592057,-0.2875333726,0.267683953,0.4577464163,-0.0942046419,0.3053969741,0.3139282465,-0.0677824542,0.0710598007,0.1999145001,0.1426868588,0.2681037188,-0.0580841042,0.0186995007,0.1312802285,-0.0954223424,0.123498708,0.3089662492,0.1333374232,-0.1635093093,0.2376047224,0.2542573214,-0.3229692578,0.2580934465,-0.3386286795,0.2281223089,-0.2395282686,0.4624901116,0.0390589684,-0.1340334415,-0.1528572738,-0.1531211287,-0.3504591286,0.043332912,-0.0394208319,0.0340955667,0.1457487643,-0.2540512681,0.1440104544,0.1248264313,0.4362035096,0.2979634106,-0.0121709164,-0.3062780499,0.0834584832,-0.3498923182,0.0501785204,0.3667325079,0.3661016822,-0.166788578,-0.1049022973,0.0701217875,0.04937879,0.0706794932,-0.1864989996,-0.0569111444,-0.1365015656,-0.3562349379,0.0285187736,-0.3498349786,-0.2375133187,0.330996722,0.0615641288,-0.0713767856,-0.0111634824,0.1290488392,-0.0126869641,0.1184190288,0.0774327889,0.2908157706,0.0786843151,0.158237502,0.3548883498,0.00733341,-0.1421247572,-0.0491093211,-0.1930250674,-0.1361330897,0.0086778337,-0.2023632973,0.3826273382,0.0849796906,-0.2019885033,-0.2672128081,0.0327144228,0.2549814582,-0.3714531362,-0.3458369672,0.1677333117,-0.0547167659,-0.1213578358,0.322085768,0.478711307,0.0301892236,0.2532159686,-0.1803782582,-0.3619759381,0.5924825668,-0.4610414207,-0.0664849877,0.0346095376,0.1938368827,-0.0236744843,-0.2490109652,-1.0206305981,0.0651262328,0.0875698179,-0.169146657,-0.1690046191,0.1944060177,-0.0718152821,-0.0802268982,-0.0244347434,0.5383815169,0.1742252558,-0.1455207765,0.0278859399,-0.0938529223]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1452","title":"SNLI dataset contains labels with value -1","comments":"I believe the `-1` label is used for missing\/NULL data as per HuggingFace Dataset conventions. If I recall correctly SNLI has some entries with no (gold) labels in the dataset.","body":"```\r\nimport datasets\r\nnli_data = datasets.load_dataset(\"snli\")\r\ntrain_data = nli_data['train']\r\ntrain_labels = train_data['label']\r\nlabel_set = set(train_labels)\r\nprint(label_set)\r\n```\r\n\r\n**Output:**\r\n`{0, 1, 2, -1}`","comment_length":30,"text":"SNLI dataset contains labels with value -1\n```\r\nimport datasets\r\nnli_data = datasets.load_dataset(\"snli\")\r\ntrain_data = nli_data['train']\r\ntrain_labels = train_data['label']\r\nlabel_set = set(train_labels)\r\nprint(label_set)\r\n```\r\n\r\n**Output:**\r\n`{0, 1, 2, -1}`\nI believe the `-1` label is used for missing\/NULL data as per HuggingFace Dataset conventions. If I recall correctly SNLI has some entries with no (gold) labels in the dataset.","embeddings":[0.2351298779,-0.490634948,-0.1780145317,0.3363342583,0.2337820381,0.0211006664,0.3015899956,0.2003286481,0.0807732418,0.2864296734,-0.2351152003,0.3536292017,-0.2061203569,0.2780169249,0.0497128963,0.0089706723,0.2335757613,0.3703630269,0.0620434359,-0.4029215276,-0.3136185408,0.2042293549,-0.3197530508,0.1720517725,-0.4605438709,0.2223438323,-0.0642018318,-0.0900376439,-0.0218581893,-0.4933305979,0.16226013,-0.0321304351,-0.0364017226,0.0682409629,-0.000104066,-0.2530908883,0.0445113033,-0.027761437,-0.4121533334,-0.2150870562,-0.1446629316,-0.1115380302,-0.0711202994,-0.2451922894,-0.2267181277,-0.1674472392,0.1702845544,-0.0743094236,-0.0091543021,0.2677800059,0.2388068587,0.2588718534,-0.2726499438,0.0377832241,0.3642984629,-0.2887417376,0.1562072635,-0.0347504877,0.0894331336,0.0672424585,0.4728746712,0.5562069416,-0.137286514,-0.0182487294,0.0932062343,0.2765094638,0.0019010649,-0.3113327324,0.194612965,0.3709362745,0.1806552559,-0.2979120314,-0.2675397396,-0.0996704921,0.0889209583,-0.400216639,-0.0638480633,0.2426854521,0.1563144922,0.2327328622,0.0338694304,0.1923460215,-0.2313928604,0.1665087938,-0.0847164243,0.7823176384,-0.1968450695,0.1007970646,0.1636362374,0.0839669183,-0.4554844499,-0.0083741294,0.1143384129,0.4739844203,-0.3202089965,-0.0803492367,-0.0371239595,0.0814103708,0.0021936407,0.1008560136,-0.0091487048,0.0005345253,-0.0671000481,0.2667031884,-0.0283777863,-0.003456393,0.3293767273,0.1094885617,0.1501355916,-0.1802954674,0.2476408631,-0.0512554944,-0.2885903418,-0.1748425812,0.4545010328,-0.0278989784,0.1671315432,-0.1204602346,-0.4925311208,0.1128047779,-0.2256963253,0.0788223892,0.0458180904,0.1962345541,-0.0339996964,0.1010887474,-0.0013508926,0.0699644536,-0.0285988636,-0.2663111985,-0.2864689529,0.1551371664,-0.0046351412,-0.255432874,0.0295977704,-0.2740984261,0.2123831958,0.0571984798,-0.2523680329,-0.1367340088,0.1693802029,-0.0879595131,0.3424265981,0.3663026989,-0.224009648,0.3934674263,0.0625335276,-0.2190621793,-0.1465875208,-0.0780828521,-0.0950909108,-0.1828194708,-0.1656190008,0.2921012044,-0.1998002529,-0.0757763386,0.1041484401,0.0385775976,-0.0655758008,0.1044519171,0.2772592306,-0.1649906337,-0.0727408752,-0.0392621532,0.258243829,0.1227393299,-0.3179697096,-0.2274758369,0.1154534742,-0.384016186,0.2018665224,0.2267106622,0.1190630421,-0.1606129408,-0.1722411662,0.2534030676,0.3473176956,-0.4435730278,-0.2762626708,-0.0471902117,-0.161497131,-0.3439723551,0.116528824,0.4465350807,-0.1905210316,0.2841313779,0.2110200822,-0.0879750624,0.1887966096,-0.0729888156,-0.2380004674,-0.0798481405,0.3521511555,0.1141519696,0.003062258,-0.0117154438,-0.2967057824,-0.0361137688,0.3701810241,-0.0732097775,0.116582036,0.1806010157,0.3337184191,0.1263370663,-0.0010266864,-0.2666113675,-0.1928448379,0.2538926005,0.0759872124,0.4210845232,0.0298257601,-0.1766707599,-0.3001585901,-0.0803937614,-0.3594002128,-0.1270229965,0.2832456529,0.2410756797,0.2093353719,0.0433067717,-0.2880734205,0.1890072227,-0.1809074283,0.0934363306,-0.3956659436,0.1654046476,0.0461279452,0.0933330059,-0.1129261702,0.4331927001,0.2286773175,0.1468732506,0.2520022988,0.2583920062,-0.2196794748,-0.1835345328,0.2187638432,0.276529938,0.0116540277,-0.4192703366,-0.0756351277,0.3144541681,0.2531535625,0.0568554029,-0.1340450197,0.4621799886,-0.0223456342,0.0066585857,-0.1698778123,0.1676912904,-0.0149603896,-0.0020620213,-0.2303067148,-0.4399383664,0.0580249019,-0.391689837,0.2669937611,0.1903838515,-0.6285805106,0.0602436475,0.3843151331,-0.3520504832,0.117489174,-0.0791908652,-0.2204506546,0.117962271,0.1759145558,0.1723509878,0.0378966071,0.3321821988,-0.0567628257,0.0690156668,-0.323064208,-0.1077095866,0.0968855023,0.0975665525,-0.0577555373,-0.054227829,0.2304429859,0.0467052869,-0.2609626651,-0.2921113372,-0.2107824236,0.2805331349,-0.3731403649,0.1426354945,-0.0735427737,-0.2771177292,-0.3835356832,-0.1542982608,-0.1928972304,-0.2834320962,0.0728442371,-0.2848047018,0.0335918926,-0.0416482314,0.056503389,0.1506259441,0.0116777811,-0.0249490235,-0.3300718069,-0.3428486884,-0.2137782276,0.1240366548,-0.3182232976,0.3774763942,0.3344863355,-0.3859687448,-0.0977402106,-0.0646956787,-0.2464907765,0.0238352492,-0.2468114942,-0.0295265559,0.3473882079,0.1475397497,-0.1745041013,0.2958403528,0.1671305448,-0.0695272759,-0.1272739172,0.005984548,-0.1463558078,-0.0574937053,-0.3506551385,-0.2518244982,-0.4022501111,-0.0249263868,0.1182148606,-0.0469041616,0.1613894701,0.4060355425,0.0036259489,0.3636782467,0.1491140574,0.2072501779,-0.2078464627,-0.2352870703,0.2357077748,0.0460595004,-0.2393185943,0.0106626861,0.094345361,-0.0006302742,-0.2364697605,-0.2919704616,-0.1215140894,-0.0245993957,0.0083640981,0.1204760596,0.1074016318,0.0959753916,0.0974492133,-0.0904043391,-0.230564937,0.0119007919,0.2191186845,-0.2883580327,0.0082270857,-0.081275098,0.1632949114,0.0013201949,0.0399134755,0.3995077312,-0.1092988849,0.0766125172,-0.2272634357,0.2812047303,0.0771833584,-0.0067515741,0.0989886969,0.4109141827,-0.0732335672,0.2068363279,-0.0151301771,0.214256987,-0.0641547292,0.1985780746,-0.0919158831,-0.0029180991,-0.0905835479,0.0777685195,0.1036107466,-0.1578301936,-0.0772637948,-0.1821531802,-0.0799096376,-0.1854282767,0.0974824354,-0.1388551295,0.0557350479,-0.5141428113,-0.082998842,-0.3476426303,0.0720675513,0.1631190628,0.158831194,0.1256497502,-0.0945825502,0.3085624874,0.1368498504,0.5765370727,-0.3166086078,0.1981485486,0.212803781,0.2803311646,-0.4869130254,-0.113967441,-0.6099460125,-0.1725613475,0.3706698716,0.3140948415,-0.0789950639,-0.4624698162,0.3771751523,0.1205480322,-0.1256605536,-0.1421991885,-0.3774523139,-0.2581156194,-0.0622942336,0.0054920632,0.1518622637,0.1125731096,-0.0510511249,0.0586017706,0.0383023359,0.1489058435,0.4121738672,0.22285752,0.2234487683,0.1637652516,-0.1800130755,0.2618964911,0.1510832608,-0.2314300984,0.4501062036,-0.097887978,-0.2910020649,0.0848188847,-0.2944493294,0.3888799548,0.0898749158,0.0088282833,0.0789896175,-0.1120534241,0.236193195,-0.2655686438,0.1670556664,0.3529888391,0.2594587207,-0.4387022555,-0.3621265292,0.6647841334,0.0958042294,-0.1892320663,-0.0729789436,0.2487386316,-0.3200927079,0.1794999689,0.2729803324,0.9641368389,-0.1005128995,0.2129392177,0.3113630712,-0.3147005439,0.6079534292,0.1732765734,-0.0148405023,-0.3615508974,0.0299907234,-0.0505760126,0.0933048651,-0.1934852749,0.2853369117,-0.0241359491,0.3201922476,-0.185781002,0.1043140367,0.0621879995,-0.1088867188,0.1585689485,0.0066224425,-0.4599723518,0.2189105898,-0.0848495662,0.2482399642,-0.1908284426,0.0670366511,-0.2170333713,0.0169962682,0.0059342515,-0.2288414836,0.1517658532,-0.0113800522,0.231186077,-0.1229619607,-0.3427266181,0.4207301438,0.6655433178,0.080902949,0.0118340123,-0.0321959145,0.0452999473,0.1970064342,0.1671718806,0.1227251962,0.4367147982,0.0465491265,-0.346909672,0.1530185938,0.1269425601,0.1641944498,-0.2171158493,0.1008448079,-0.3158276379,0.0655721724,-0.1465639919,0.0275816415,0.0687630624,-0.0932060853,0.1963314265,0.0360523202,-0.2817460597,-0.1251504421,-0.0924172848,-0.434812367,-0.1411595047,0.2630099654,0.0323561132,0.2662035525,0.2066871375,0.000849844,-0.0056230417,-0.2599628866,-0.0178740229,0.3614188731,-0.4468573928,0.0775829703,0.184569791,-0.0173151847,0.005907509,0.3781409562,0.0668684542,-0.0067959181,0.0433615558,-0.2623486817,-0.3767210841,-0.046002984,0.0937625393,0.3868141472,0.1167083085,0.1600836217,-0.0736195073,-0.3393360078,-0.4597358406,-0.0735426545,-0.4624356329,0.2527546883,0.2372574359,0.0936302841,0.0148801031,0.0273297075,0.266951412,-0.0802696794,-0.0791352764,-0.3020402491,-0.2825827003,0.0617700741,-0.0219961368,-0.1211457998,-0.011522756,0.1759750843,-0.0505349636,-0.0988349691,0.1627531797,0.325003624,0.0983197466,0.8718969822,-0.2352777123,0.2100133747,-0.0423877612,-0.2241737247,0.091741249,-0.1011335403,0.1130430326,0.1086645722,-0.0958105847,0.0072402619,-0.0508465096,0.148025319,-0.2378524542,-0.0637744218,-0.0144621646,-0.3602891862,0.2820595801,0.1690960974,0.3595198095,0.1125513092,-0.2995539606,-0.1943112761,0.1972056925,0.367429018,-0.2160392553,-0.0263514854,0.1794310957,0.1220029742,0.2011300474,0.3365058303,0.0186138246,-0.0376086347,-0.0143329371,-0.0522533283,0.0955803916,-0.2456497103,0.056246601,0.264077425,-0.0146668842,-0.0296910163,0.2569749653,0.0768447444,0.1075203642,0.5465226173,0.0581043698,0.5124669671,-0.06780646,0.1534053385,-0.2671483159,0.0794395432,-0.0952102914,0.215826124,-0.3798171282,0.2653621435,-0.0469482876,0.2389761209,-0.2679867744,-0.0305678509,0.0572863743,0.2600264549,-0.2000262588,-0.2792088687,-0.5016416907,0.0313145593,-0.1620900929,0.0636745915,0.0177423749,-0.234311074,0.1490168273,0.1415742785,0.0570540465,-0.4010354578,-0.3689585328,0.36513412,0.3934513032,-0.2435117364,-0.0449131355,0.2358343899,-0.0070815594,0.2254355401,0.0187666155,0.3661857545,0.0601861365,0.1167529598,0.0989403948,-0.2857356966,-0.2569532096,-0.0397981293,-0.053873565,0.1812048703,-0.0081215166,0.2478680462,0.2175568938,-0.2464453429,0.2887548506,-0.0403528586,0.359058857,-0.3402103484,0.1050338,-0.1236345544,-0.2808960378,-0.3866416514,-0.0705356747,-0.4358713031,0.1042103469,0.0963247791,0.2498170435,0.2405853868,-0.1760198176,0.1389347762,0.1296518892,0.3520661592,0.1182737499,0.1525215507,-0.086435169,-0.1423437744,-0.5975673199,-0.066822201,0.1150040999,-0.143396765,0.0619998351,-0.0243961029,-0.2969183922,-0.2536074221,0.0132277757,0.1678119004,0.0532695539,-0.0115240291,-0.1281196028,-0.1661501229,-0.1828193516,-0.108283639,0.1383168697,-0.0779633299,-0.1140060946,-0.2393924445,0.1913024783,-0.1337547749,0.2242268771,-0.2953458726,0.0906505436,0.1201316267,-0.03050665,0.3859786689,-0.0153775094,0.083463043,-0.087238282,0.052930221,-0.3612129092,0.1725237519,-0.0918775573,0.4230141044,0.0075020557,-0.0877741426,0.0177342631,0.1135893464,0.2244577855,-0.0923166946,-0.2933936119,0.3331451416,-0.1683173329,0.0176842343,0.1092719957,0.219321236,0.213530004,-0.0095569342,-0.2414336503,-0.3486387432,0.3660676181,-0.3710889518,-0.2608675063,-0.0415771976,0.2282085568,0.3876202404,-0.1055226773,-0.6232703328,0.1003511399,0.356475234,-0.0327017643,-0.5908319354,0.2276556194,0.1633147001,0.1289622635,0.0048582838,-0.0563322939,0.2881031632,-0.1017462686,0.290903151,-0.4343141913]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1452","title":"SNLI dataset contains labels with value -1","comments":"Ah, you're right. The dataset has some pairs with missing labels. Thanks for reminding me.","body":"```\r\nimport datasets\r\nnli_data = datasets.load_dataset(\"snli\")\r\ntrain_data = nli_data['train']\r\ntrain_labels = train_data['label']\r\nlabel_set = set(train_labels)\r\nprint(label_set)\r\n```\r\n\r\n**Output:**\r\n`{0, 1, 2, -1}`","comment_length":15,"text":"SNLI dataset contains labels with value -1\n```\r\nimport datasets\r\nnli_data = datasets.load_dataset(\"snli\")\r\ntrain_data = nli_data['train']\r\ntrain_labels = train_data['label']\r\nlabel_set = set(train_labels)\r\nprint(label_set)\r\n```\r\n\r\n**Output:**\r\n`{0, 1, 2, -1}`\nAh, you're right. The dataset has some pairs with missing labels. Thanks for reminding me.","embeddings":[0.1118536741,-0.2780376375,-0.1819718182,0.1631645411,0.2689473331,0.0285462718,0.5036272407,0.2536185682,0.0604351163,0.2276120782,-0.1744589061,0.4532493949,-0.210382387,0.2768723965,-0.0508687831,0.0627899021,0.2159684598,0.3498456776,-0.0242513493,-0.4335756004,-0.2908861041,0.142149955,-0.3404208124,0.1497154683,-0.3666659594,0.1060911715,-0.0292138495,-0.0403701775,0.0229313765,-0.4227949977,0.1341409981,0.0434117131,0.0409138575,0.1052899733,-0.0001008777,-0.2558995187,0.0653119013,-0.0549712218,-0.3258263469,-0.0199053809,-0.2499684989,-0.1186199486,0.0023276156,-0.3762792349,-0.2135950476,-0.2262270302,0.1018225998,-0.2142590731,0.0586634129,0.3627851009,0.3158906996,0.15790084,-0.2155623138,0.0024745879,0.4544513226,-0.2641556859,0.1115611941,0.0459361114,0.1250978708,-0.0291645937,0.4434398115,0.4544566572,-0.1860801131,-0.1342370063,0.0469911024,0.1790825129,0.0347375609,-0.4117902815,0.1620332152,0.3017691076,0.3887634277,-0.2639425695,-0.302341491,-0.0734768361,0.1300907731,-0.2908623815,-0.0451200865,0.1392708272,0.08795771,0.1763615906,-0.0263210069,0.1870730519,-0.1172016338,0.1945297122,-0.0672754496,0.6152953506,-0.0560118295,0.1212967038,-0.0269814637,0.0957662538,-0.1853598058,0.0174402893,0.0846525505,0.4305848777,-0.3594730496,-0.1304369569,-0.0068986462,-0.1462401003,-0.1168877706,0.2244693339,-0.0339934155,-0.0151705984,0.0769722164,0.2177132964,0.0424441621,-0.0703018606,0.2687713504,0.4172580242,0.0470611304,-0.19404082,0.1028210446,0.0042303796,-0.2441621721,-0.3367707133,0.340361923,-0.008628225,0.0897424817,-0.1094033942,-0.3745248616,0.1177113876,-0.3621172011,0.0231859013,0.0122725228,0.2220629305,-0.1283331364,0.0967051759,0.0591875017,0.1713957042,-0.0693793073,-0.2387370914,-0.3129944205,0.1732561588,-0.0757716075,-0.2522183359,0.0876326114,-0.1049505398,0.3333207369,0.0357388146,-0.0978717133,-0.1024185717,0.1471026391,-0.079775013,0.2700354457,0.1682723612,-0.0486292019,0.4202074707,0.0123415645,-0.2450473607,-0.0535009056,-0.0629094914,-0.1146162003,-0.018771477,-0.1340129524,0.366250366,-0.1971644759,-0.1101498306,0.1922436506,-0.1310351193,0.084574081,-0.0899315178,0.1767875999,-0.2608612478,0.0718951374,-0.1219825149,0.2854155302,0.0125711653,-0.3457044363,-0.0725089982,0.2392299473,-0.1646461636,0.3382320702,0.339040339,0.0200880468,0.0111369202,-0.1895346642,0.1582575589,0.4480031431,-0.2867081761,-0.1963155866,0.0087498277,-0.0136905219,-0.3934727013,0.1116809621,0.4097698927,0.0208251644,0.138966009,0.2405127585,-0.0829332545,0.0617567822,-0.0589002967,-0.2983292937,-0.0328534842,0.3897428811,0.1370458454,0.1605696827,0.0718540624,-0.1543074548,-0.0297723543,0.3749241829,-0.1329959631,0.1035239547,0.3646071851,0.2937957644,0.0940033197,0.0096858768,-0.1585902572,-0.1178036183,0.234958142,-0.0907361805,0.3485963345,0.1171335801,-0.0694336742,-0.3225132227,-0.1373307556,-0.2918079793,-0.109129943,0.3371516466,0.1363331676,0.0403683297,0.1040598005,-0.1811056584,0.0598758198,-0.183663398,-0.0280997902,-0.3217225373,0.3370443881,-0.04671707,0.0504470915,-0.0166396145,0.398273319,0.1611090153,0.1465103477,0.1317023039,0.2127883434,-0.0392171703,-0.2444071025,0.0878971368,0.1537600309,-0.0672802255,-0.491674304,-0.1090056151,0.2931222022,0.3065393865,-0.0393610075,-0.1812731773,0.3904075623,-0.0518389009,0.0518175028,-0.1955671459,0.1358830333,0.0696080402,-0.1048024297,-0.1713804752,-0.2462313324,0.009794197,-0.3247595429,0.2665183842,0.1878144294,-0.4635910392,0.0227727499,0.4532364905,-0.1767725796,0.069463253,-0.1261389107,-0.1291095465,0.0945222527,0.0371354558,0.2148673832,0.1471679509,0.3857699633,-0.0607934482,0.0064881872,-0.2905076146,-0.1038385928,0.1719707996,0.1223573834,-0.0235670991,-0.029720569,0.1485604495,0.0945165157,-0.1279315501,-0.3228694797,-0.1710501313,0.2836472392,-0.3586219251,0.0445262268,-0.0937035158,-0.272796452,-0.3074238002,-0.0526748076,-0.1450656205,-0.3768035769,0.1448462903,-0.2679312229,0.0589643233,0.0580109544,-0.0319974758,0.0395493805,0.0537365712,-0.0339079462,-0.1718099266,-0.4197889566,-0.330914259,0.2308084816,-0.1267460734,0.293194592,0.2365594208,-0.3838067353,-0.2064955831,-0.1219534129,-0.2666961253,-0.0873747319,-0.1195058078,0.0401754938,0.2885735035,0.181355387,-0.1752677858,0.1303755343,0.142224595,-0.0287497435,-0.1489929557,0.0204918105,-0.1133561954,-0.2219990045,-0.2576594949,-0.4762954414,-0.2694855332,-0.1307218075,-0.0201449506,0.0376781486,0.1003485098,0.2586572766,0.0139030134,0.1723561287,0.1337086409,0.1764549762,-0.2077111602,-0.1488771737,0.2580248415,-0.0248956271,-0.3455135226,-0.0087045832,0.0358207114,0.1769574434,-0.1822732985,-0.2311969548,0.0559090748,-0.0600130297,-0.0319903791,-0.0126013923,0.1555624455,0.0959043577,0.2128648758,-0.1529211402,-0.2857674658,-0.06215325,0.2661264241,-0.2060146928,0.1070382446,-0.1710791886,0.2117890567,-0.0926895589,0.0687291324,0.4884346426,-0.2214795947,0.1204956099,-0.3573571742,0.1877124608,0.0524740368,-0.1540985405,0.2141720206,0.2431252301,-0.1410363913,0.2924849093,-0.0452337228,-0.0149453441,-0.2022334784,0.2277644873,-0.2570956051,-0.1121347547,-0.0439009964,0.077723071,0.0478486456,-0.0562553108,-0.0066201556,-0.1942148656,-0.1253768355,-0.1539738327,-0.0516392849,-0.1641162038,0.1019854769,-0.6157753468,-0.1331816912,-0.337094605,-0.0067198551,0.128797695,0.3186765015,0.1370190084,-0.0484169088,0.1838338822,0.1871806532,0.53604877,-0.4153202176,0.0978542939,0.227091521,0.4576844871,-0.4525152743,-0.1272295117,-0.6044983864,-0.0763088316,0.4541174471,0.1446553916,-0.0078381496,-0.3718630373,0.2816555798,0.2030185312,-0.1784158051,-0.1690223515,-0.3429640234,-0.2450873256,-0.1208986342,-0.1261227578,0.1934665143,0.1034550667,-0.0692914948,-0.0848251879,0.0481572859,-0.0086913593,0.3802253604,0.2562136948,0.2714602351,0.0841505677,-0.0305391829,-0.0119460598,0.2066926956,-0.135387823,0.4792304635,-0.0089818081,-0.157875061,0.1499581039,-0.4739341736,0.4508332312,-0.0433704294,-0.0479226932,-0.0485182889,-0.2707800567,0.2366050631,-0.0340845883,0.1342850477,0.3640725613,0.175876528,-0.4084053338,-0.337128222,0.4565478265,0.0549930669,-0.1540043652,0.034854617,0.0469578095,-0.2804580033,0.2358067036,0.3273253441,0.9063718915,0.0007837245,0.1231529266,0.3364728093,-0.2744974494,0.6885092854,0.2298232913,0.0742127374,-0.3973150849,-0.0229825657,-0.0168596879,0.0473411232,-0.2256451696,0.2902073562,-0.094923757,0.4155393839,-0.1947015077,-0.0730092376,0.1185048223,-0.0672015026,0.4083646536,0.0843995512,-0.335192889,0.238118723,-0.0637855902,0.1856127232,-0.1690883338,-0.0605456308,-0.2150921822,-0.0806534812,-0.0445048474,-0.0561183393,0.0690634623,0.037315663,0.1476336271,0.088920109,-0.1210791692,0.4016357362,0.6263202429,0.1479914486,-0.0350456797,-0.0202401951,0.1046754047,0.2282796353,0.225617677,0.1912357658,0.3700456917,-0.0072713541,-0.2491001487,0.054462675,0.0236032736,0.1254260987,-0.2921708822,0.0956685841,-0.2048648894,-0.1443951577,-0.090412043,0.1094155982,-0.0645864084,-0.2053784281,0.2465061992,0.0030322932,-0.1927841306,0.1943092197,-0.0878915414,-0.3196943998,-0.0912416279,0.2692928016,0.0545456968,0.1364841014,0.3814725876,0.0249946825,-0.0380914882,-0.316624403,0.0439099632,0.2432015836,-0.5036354661,0.0205990896,0.0040114955,0.0839461237,0.1525068283,0.308416307,0.1293535829,0.0853622258,0.1321308613,-0.3326323032,-0.3325412869,-0.0103168739,0.035903275,0.3189948201,0.0556394495,0.241280064,-0.0106051667,-0.2725116313,-0.495752871,-0.1186678261,-0.5839552879,0.2587632239,0.0860810429,-0.0056848908,-0.040470168,-0.0325991511,0.3335573673,-0.1813172698,-0.0341323353,-0.3547062278,-0.180824548,0.0251951609,-0.0195422973,-0.0087054186,0.0031828058,0.1268298626,-0.1882519126,-0.06436795,-0.0866366997,0.2892719507,0.0460953899,0.7262243629,-0.1603816748,0.12108583,-0.1881635785,-0.2265730351,-0.0524255484,-0.1894924045,0.0947076902,0.142619729,0.1234540418,-0.059150029,-0.0671540201,0.0521568805,-0.3606184125,-0.0546072237,0.0892574787,-0.4208975732,0.2675087452,0.206945911,0.3510307372,0.1475009471,-0.229669854,-0.141553238,0.2146027088,0.4330559075,-0.4210015535,-0.0867048949,0.1990365237,0.15756993,0.142960459,0.3227541447,-0.0193873756,-0.2346490026,0.095640339,-0.062227048,0.0465174839,-0.3910162747,0.0741765425,0.2747507989,-0.0270378534,-0.0535501242,0.1003342792,0.0184096955,0.0815019011,0.6174554229,-0.0041080094,0.5148960948,0.0431082472,0.0536883362,-0.2062141448,0.0563800968,-0.0069791735,0.2371980697,-0.3062588871,0.181102246,-0.0525687598,0.1173020229,-0.1255834997,-0.0598856509,0.1619001478,0.1122158989,-0.1767848581,-0.1399238259,-0.5250555277,-0.0684196725,-0.2403157204,-0.2016872466,0.0308872573,-0.1433385313,0.1405465305,0.2282343209,0.0738630369,-0.3885368705,-0.3406557143,0.2546333969,0.2827319503,-0.2683543563,0.0681291521,0.184567824,-0.117741026,0.1976631135,0.0212832708,0.4366680682,0.2140670866,0.1502760798,-0.0331765823,-0.1218941808,-0.2761456966,-0.0452231467,0.0216358043,0.1430811286,-0.110946022,0.330550909,0.2662889659,-0.2441501468,0.2916612625,-0.033834897,0.1981048137,-0.1975321472,0.2196088731,-0.0722686499,-0.247362107,-0.317083627,-0.0829906836,-0.456032604,-0.0038557621,0.0857834071,0.1851464063,0.2101688832,-0.1925017685,0.1610216051,0.0609055646,0.4192157686,0.1005906612,0.1539625973,-0.1062553152,-0.0732606649,-0.3974675536,0.0089585222,0.0021361727,-0.1896007955,0.1017434746,0.0753091723,-0.2065397799,-0.2281098962,0.1403315514,0.2636688352,-0.0185287334,0.134640649,-0.3385530412,-0.1680946648,-0.2514616847,0.0045157066,0.1108765379,-0.0740717649,-0.0734854564,-0.3135527968,0.2651238441,-0.1156844422,0.1227104589,-0.0880045369,0.0885892138,0.0854889974,0.089568615,0.4334324896,-0.1032164469,0.0500132404,-0.0275047943,-0.1733646095,-0.2060021758,0.2303587347,0.1327759922,0.4021478295,-0.0755728409,-0.0240012445,-0.065914385,0.0404378735,0.2750587165,0.0078834333,-0.312038213,0.3482681513,0.0035180717,0.0903298706,0.0997659117,0.2146539986,0.0612340383,-0.0031070574,-0.2937858403,-0.4520533681,0.5011490583,-0.3566154838,-0.2193627954,-0.1289221793,0.3157896996,0.3374814987,0.0022223666,-0.4355823994,0.0485707857,0.4443815351,-0.0137845296,-0.4638033509,0.2154532373,0.0795976892,0.1536393762,0.0835807472,-0.1729295254,0.1779200435,-0.0587782487,0.2044994086,-0.4551286399]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1444","title":"FileNotFound remotly, can't load a dataset","comments":"This dataset will be available in version-2 of the library. If you want to use this dataset now, install datasets from `master` branch rather.\r\n\r\nCommand to install datasets from `master` branch:\r\n`!pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","body":"```py\r\n!pip install datasets\r\nimport datasets as ds\r\n\r\ncorpus = ds.load_dataset('large_spanish_corpus')\r\n```\r\ngives the error\r\n\r\n> FileNotFoundError: Couldn't find file locally at large_spanish_corpus\/large_spanish_corpus.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/large_spanish_corpus\/large_spanish_corpus.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/large_spanish_corpus\/large_spanish_corpus.py\r\n\r\nnot just `large_spanish_corpus`, `zest` too, but `squad` is available. \r\n\r\nthis was using colab and localy ","comment_length":34,"text":"FileNotFound remotly, can't load a dataset\n```py\r\n!pip install datasets\r\nimport datasets as ds\r\n\r\ncorpus = ds.load_dataset('large_spanish_corpus')\r\n```\r\ngives the error\r\n\r\n> FileNotFoundError: Couldn't find file locally at large_spanish_corpus\/large_spanish_corpus.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/large_spanish_corpus\/large_spanish_corpus.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/large_spanish_corpus\/large_spanish_corpus.py\r\n\r\nnot just `large_spanish_corpus`, `zest` too, but `squad` is available. \r\n\r\nthis was using colab and localy \nThis dataset will be available in version-2 of the library. If you want to use this dataset now, install datasets from `master` branch rather.\r\n\r\nCommand to install datasets from `master` branch:\r\n`!pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`","embeddings":[-0.3270967305,-0.3888709247,-0.0211854372,0.3421646357,0.3611239195,0.053929776,-0.0934886783,0.212861225,-0.0617659502,0.2608962655,-0.3553176522,0.1730859429,0.0300486814,-0.1594371796,0.0316852406,0.0206086189,-0.0119563369,0.0681066439,0.0414810441,-0.0984659716,-0.1051172316,0.337292552,-0.2512364686,-0.050593365,-0.1271476299,0.0490329787,-0.1710327119,0.1196063608,-0.3393884599,-0.1975401193,0.3731856644,-0.0539984293,0.3303165436,0.7346161604,-0.0001053096,0.0331307463,0.3021110892,-0.165072754,-0.1891661733,-0.5733721852,0.2214238346,-0.5341050625,-0.0140152387,-0.1214286312,-0.3000083864,-0.1858647615,0.2292091548,-0.0646855161,0.4581916332,0.138992846,0.2760677934,0.1265496016,0.1662908792,-0.3696480691,0.2097802907,0.1956458986,-0.1781926453,0.2711293101,0.5124987364,0.0008453406,0.3474407494,0.2254715711,-0.1975982338,-0.0633731633,0.0224851724,-0.0391207598,-0.1415897161,-0.397159338,0.2629119158,0.2328353524,0.3422170281,-0.3914035559,-0.2908690572,-0.1270831078,0.0680985674,-0.0180698503,-0.0308835004,0.195157066,-0.0137036536,0.1463201791,0.1868037134,-0.1229998842,-0.047105141,0.3313314319,-0.0886648819,0.2586908042,-0.3090187013,0.0099390605,0.1905119717,-0.178924948,-0.2507298589,0.0302959979,-0.0183176827,0.4132662117,-0.0670698136,0.0363565609,0.1135698333,0.0308769792,0.3723825514,-0.1460781097,-0.0953407064,0.0370077156,-0.1044247374,0.2302518487,0.2692061365,0.045740705,0.1354756653,0.1478246897,0.1838281304,0.2538388968,-0.0949297026,-0.1023403108,-0.2020095885,-0.2282737941,-0.3279549778,-0.1099150777,0.4180825353,-0.368394345,-0.1520365477,-0.1110243574,-0.0217322372,0.035854809,0.2241002768,0.6009091139,-0.1057141572,0.0214474481,-0.2264181376,0.1999215186,-0.1387862265,-0.0065258481,-0.2208006829,0.1473014504,-0.2257985026,-0.0462417379,0.2515354156,-0.5578830838,0.393538475,-0.0472707152,0.1960410029,-0.0681858808,-0.1338357925,-0.301745981,0.0997979343,0.229762733,0.101319164,0.0338174775,0.2995767891,-0.2491473556,-0.2520236969,0.0262057353,-0.2204843909,-0.1265044808,-0.3007340133,0.1876771152,-0.3505807221,-0.0759404302,-0.3589547575,0.2076333761,-0.0608204342,-0.4315320253,-0.000035826,-0.0217535459,-0.2501894236,-0.1095410511,0.425739795,0.7670357227,-0.2865989804,-0.1230983064,0.0340257213,-0.126667127,0.101105772,0.126527518,-0.1167581826,-0.0218859594,-0.265375644,0.1027911976,0.2992968559,-0.615336597,-0.3468686044,0.2056128085,-0.0500283204,0.1586273313,-0.0413157791,-0.1025643721,-0.0431875326,0.050346978,0.0029872649,0.0049375533,0.0312955864,0.0295400694,-0.0962504521,-0.1457029581,-0.1982443333,0.3119064271,0.1383856982,0.0675422177,0.3202099502,0.0876349062,0.0316906422,0.0958209634,-0.0984545723,0.2816782892,0.3037556708,0.1534382701,0.0837896764,-0.2247743607,-0.3634960353,0.2498729378,0.0457058363,0.2284103036,-0.2949457169,-0.188501969,-0.1586822271,-0.0201391019,-0.2509643435,-0.1564370543,0.1302569211,0.0791035146,0.1900039166,0.1363538951,-0.2076636851,0.1021811813,-0.1985594481,0.0608884916,-0.3810701668,0.0536024906,-0.1443722248,-0.0908582732,0.0266818665,0.3134473264,-0.0061015519,-0.2961701751,0.1275138259,0.2339981198,-0.0410703793,0.0831171274,0.0651024356,-0.1028961241,0.1492845863,-0.1187017262,0.1305786818,0.0326571129,0.1909237206,-0.0021715867,-0.1348509341,0.138798058,0.0411254317,0.0962456986,0.1917950064,-0.2008479834,0.3672035336,0.112716496,-0.2047113925,-0.1698101759,0.3535322249,0.0969964117,0.1955730617,0.0430092216,-0.3212142587,-0.1043168306,0.3177223802,0.1372816265,0.1213165373,0.2048687339,-0.2800172865,0.186897859,0.003449304,0.0588296801,0.4702047706,0.3391547799,-0.0764484555,0.0293206442,0.1730826646,-0.0994678661,0.0943906382,0.1055547893,0.1211164296,0.2762360275,0.0952636003,0.111055091,-0.3392531872,-0.1590596288,-0.2432276756,0.2127058059,-0.0921130925,0.0312307067,-0.1918958724,-0.5405955911,-0.2883024216,-0.1157423779,-0.3506484628,0.1294652373,-0.0584437922,-0.0211829785,0.2439412177,0.1399305463,0.0428650416,0.1373578906,-0.1882007718,-0.4878160954,-0.0153020732,-0.2103477567,-0.126777187,0.0949553549,0.373341918,-0.0983640179,0.3098777235,-0.2427178919,0.0662054196,-0.1999077499,-0.318315357,0.120759435,-0.196837157,0.2978352308,0.0523591004,0.2083314359,-0.3186392784,-0.0556888133,0.3241010904,-0.2407051027,-0.1146929413,-0.0510720573,-0.2177354693,-0.2655118704,-0.0859508514,-0.4589136243,-0.5289609432,-0.447973907,0.1381776482,0.3176063001,-0.0419272594,0.4903953373,0.2172504216,0.4897038043,-0.1627669334,0.2261385322,-0.1228660271,-0.4091244936,0.2950487733,-0.2791319191,-0.3444777429,0.1522465646,0.0878732055,0.2448349744,-0.3050942421,-0.4055993855,-0.2700473964,-0.0187662374,-0.0090881865,-0.0477787703,0.328910172,0.2568496764,-0.2854099572,-0.0629229397,-0.0179242268,-0.1917268783,-0.1979177296,0.0373480804,0.3480443954,-0.2784844339,0.1747761518,-0.0510753207,0.5100226402,0.1080506817,-0.0539683104,0.4590839148,-0.2125363797,0.5354276896,-0.284610182,-0.6068034768,0.1230493113,0.1881648451,0.1007851958,0.455753684,0.2662453949,-0.0571786948,-0.4042778313,-0.253518641,-0.1881132871,-0.0671611428,-0.0563629046,0.3478078544,-0.0905106217,0.0250823442,-0.0433408767,-0.0595117845,-0.2525693476,-0.1394330263,0.3875393867,-0.1102691293,0.1785722077,-0.1752706021,-0.0578829534,-0.1995628625,0.3943420649,0.0865487233,-0.1424325407,-0.1837637275,-0.0137153687,0.097867243,-0.1053808108,0.5849302411,0.2099873126,-0.1671550274,0.0606411733,-0.1672187001,-0.3760756254,-0.0778248459,0.0220037047,0.2661767304,0.3142610788,0.5334757566,-0.3083975017,-0.1032577828,0.25345999,0.3023595512,0.0048125433,-0.1182588264,-0.5544285178,-0.4240008891,-0.4105862379,-0.0481040403,0.0090005966,0.1661432385,0.2040583193,-0.0146053433,0.0687740073,0.0684202239,0.2253904939,0.299115479,0.2291793078,0.3138447404,0.065086633,0.2294562906,0.248440966,0.1432071626,0.7460249066,-0.1061469316,-0.4913912416,0.0051837605,0.0553579181,0.1756643951,0.2071933299,-0.0321507789,0.0447965711,0.0930327848,-0.1065680012,-0.0349840112,0.0573667809,0.2532098889,0.1384339929,-0.3063986003,-0.3154132068,0.3437559605,0.1638239175,0.2035823464,0.2294928432,0.1355138421,0.0448230393,0.2078261971,-0.1876066923,0.907562077,-0.2314331234,0.0208020248,0.0834301859,0.2398640811,0.6653989553,-0.2646847963,0.0139116989,-0.2222963274,-0.2613443136,0.0821412727,-0.0289324839,0.332262814,-0.0175855029,-0.3768773973,0.3186191618,0.1252050698,0.3583048582,-0.16409868,0.1369104683,-0.142059207,-0.2050007731,-0.4637773037,0.2265389562,-0.1797637343,0.5575463176,-0.1739387959,0.0781848654,-0.2595531344,-0.2640449405,-0.2268749923,0.1334368736,-0.3323608637,0.1006983146,0.1577433348,-0.3055326045,0.0968877375,0.5088331103,0.2002818733,0.0978932008,-0.3811895847,0.1110065579,0.1457075775,-0.4354255795,-0.0231535938,0.0302781444,0.2411985993,-0.1258546263,-0.1988929659,-0.0128455125,0.1473236233,-0.055887498,-0.0240393989,0.1696569324,-0.0047227722,-0.0821274966,-0.5319322348,0.0294233058,0.1550682038,-0.0894571394,0.1784019917,0.1721092463,-0.159109965,0.01337595,0.2607854903,-0.2644824684,-0.0693470612,0.498560816,-0.2105444372,-0.2688063383,0.3399003148,0.427875042,-0.1696823835,-0.2209756672,0.1368850321,-0.0918848962,-0.3659348786,-0.0913234577,0.299831748,0.0477569513,0.1249719709,0.4114444256,0.1034818813,-0.3905794621,0.0265023317,-0.3580159247,-0.2339127213,0.1635785103,0.0341691636,0.0446513332,0.3250163198,0.0347864591,0.3023769259,-0.0056316881,-0.369738102,0.1797963977,-0.3807907104,0.0691766813,0.1513292044,0.1309876889,0.2864605486,-0.0649409145,0.1327305138,0.1592631191,-0.1887765378,-0.3201285601,0.0252338499,0.1318091154,0.0088003604,0.0407903232,-0.0902465135,-0.1275175363,-0.1669250578,-0.1647980362,0.0729558244,0.2761351168,0.0528861433,-0.0219793469,-0.1367124766,0.0181748681,0.1757992953,0.112044625,0.0819193944,0.0469869003,0.2217239141,-0.0142504917,-0.042215623,-0.0695015714,-0.0026878694,0.0092809964,0.1468575299,0.0915894136,0.201220572,-0.1149231344,0.2597365975,0.2934654355,0.1343341172,0.1988358945,-0.2045553774,0.3173618019,0.2910456955,0.1649031639,-0.4927283823,0.0109484764,0.3847287893,-0.0523795225,0.1156073287,0.3433591425,-0.0403706655,0.2120542377,0.220249936,0.1148749366,0.6210146546,-0.0351018719,0.2023153305,0.2659614384,0.0513977073,0.0443716235,0.0082494467,0.1925211847,0.1516172886,0.20732373,-0.4689477682,0.2712054551,-0.2565205097,0.1767780632,0.1840639114,-0.4482612908,0.2845285535,0.2756141722,0.1606398523,-0.1030310243,-0.184969008,0.6661188006,-0.2973182201,0.0942683965,-0.2109468132,0.1413542181,-0.0934720412,-0.0145489331,0.0580372326,-0.2697334588,0.0645316392,-0.0911812037,0.0593159162,-0.2050245255,-0.1774323136,0.0446235947,-0.1793869734,-0.1205453798,0.0697872341,-0.0349678174,0.0536233522,-0.0868523642,0.3013354838,-0.0068913461,0.1842117161,0.0858196467,0.5155832767,0.5716382265,0.3280612528,0.2361672968,0.1018439978,-0.1570679694,-0.0988454223,-0.1237168312,0.2292480171,-0.0049438421,0.0981467217,0.2715250552,0.2406728715,-0.191452086,0.0523064844,0.0405184925,-0.0540999137,-0.0939719006,0.1424347907,-0.3754615486,-0.040780928,-0.3616989851,0.1225066409,-0.574762702,0.2454855293,0.2230425328,0.0232154448,0.0817761123,-0.281914115,0.1061741859,0.0963528976,0.2875329554,0.3922448754,0.4183029234,-0.1742403507,-0.0814740211,-0.7985011339,0.1864479333,-0.0288854912,0.0994036719,-0.1754780859,0.0637720451,0.1289838254,0.2351397723,-0.1391017586,-0.2754873931,0.1332513392,-0.1303336918,-0.2426570803,-0.1969063282,-0.1147121489,0.086795561,0.0514457412,-0.1422484368,-0.1107990667,-0.3931455612,0.0234793425,-0.194093734,-0.1383961141,-0.3310091794,-0.3074293733,0.2228418887,-0.1930968016,0.3065578341,-0.1752336323,0.0472529605,-0.2333415598,-0.3475571871,-0.0999896154,0.3762139678,0.2235003263,0.1463289112,-0.3078841567,-0.2681342959,-0.202549085,0.0366116427,0.0824194402,0.0228216778,-0.1170497388,-0.1455352753,-0.0485010669,-0.017513413,-0.1628396511,0.2385034114,-0.1185432822,-0.0291283187,-0.2439675778,-0.1928910017,0.6545346379,-0.2642819285,0.0807748064,-0.1202555671,0.0465950705,0.1699085087,-0.4187501073,-0.5220379233,0.3417106271,0.2361607701,0.1816197485,-0.2510513663,0.2188769579,-0.291934669,-0.0108342888,-0.0916387588,0.0521893278,0.0451445282,0.0432635099,0.2105447799,-0.0679319575]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1444","title":"FileNotFound remotly, can't load a dataset","comments":"Closing this, thanks @VasudevGupta7 ","body":"```py\r\n!pip install datasets\r\nimport datasets as ds\r\n\r\ncorpus = ds.load_dataset('large_spanish_corpus')\r\n```\r\ngives the error\r\n\r\n> FileNotFoundError: Couldn't find file locally at large_spanish_corpus\/large_spanish_corpus.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/large_spanish_corpus\/large_spanish_corpus.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/large_spanish_corpus\/large_spanish_corpus.py\r\n\r\nnot just `large_spanish_corpus`, `zest` too, but `squad` is available. \r\n\r\nthis was using colab and localy ","comment_length":4,"text":"FileNotFound remotly, can't load a dataset\n```py\r\n!pip install datasets\r\nimport datasets as ds\r\n\r\ncorpus = ds.load_dataset('large_spanish_corpus')\r\n```\r\ngives the error\r\n\r\n> FileNotFoundError: Couldn't find file locally at large_spanish_corpus\/large_spanish_corpus.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/large_spanish_corpus\/large_spanish_corpus.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/large_spanish_corpus\/large_spanish_corpus.py\r\n\r\nnot just `large_spanish_corpus`, `zest` too, but `squad` is available. \r\n\r\nthis was using colab and localy \nClosing this, thanks @VasudevGupta7 ","embeddings":[-0.318844974,-0.3623417616,-0.0187154878,0.3630237281,0.3586478829,-0.0406982675,-0.0311189108,0.1687899828,0.0016797043,0.2309849113,-0.3423127532,0.1984071434,0.079322122,-0.2001295984,0.0361942425,0.0450788848,0.0581420474,0.0609196387,0.1336325854,-0.1898285449,-0.2286264598,0.2249712348,-0.2781754434,-0.060955748,-0.1133666113,0.1477235705,-0.2026691288,0.1306924671,-0.4269887805,-0.1987339109,0.4409176707,-0.1318822801,0.2792572081,0.7935501933,-0.0001119634,0.1272509843,0.3094501197,-0.0960893184,-0.2836238742,-0.5593124032,0.0830454975,-0.516407609,0.057928998,-0.0445662849,-0.1804553121,-0.297049731,0.2368040979,-0.0564736091,0.4217143953,0.1436933726,0.1947663277,0.0987897664,0.1234080642,-0.3420368731,0.1211903468,0.2655274868,-0.2022892982,0.3334844708,0.5162982345,-0.1877292097,0.4161331356,0.1824971586,-0.4205686152,-0.029600814,-0.0682301223,-0.0317780599,-0.0980928242,-0.4560987949,0.2474964857,0.2424454689,0.446729362,-0.3158263564,-0.3863340318,-0.0750260428,0.0814233795,0.0097904345,-0.0001135968,0.2344245166,-0.0801856965,0.1933421046,0.1069619507,-0.1412335485,-0.0420395397,0.404332608,-0.1961879134,0.3434244394,-0.2501561642,0.040604528,0.0815431774,-0.1162961796,-0.0951064304,-0.0920645148,-0.0225540642,0.3995039761,-0.1042857468,0.035189759,0.0940692574,0.1090650782,0.2983530164,-0.2015966475,-0.0513805635,-0.1063900962,-0.1460077018,0.2570865154,0.2468955815,0.0630103797,0.1124749556,0.255949676,0.1772999763,0.1814268231,-0.0923162252,-0.0464251079,-0.1950502396,-0.2668486238,-0.3082130253,-0.1235987246,0.5392895937,-0.3685724139,-0.278549403,-0.018559996,-0.1544689387,0.0568310581,0.1439264566,0.5655876398,-0.1516357064,0.1474625766,-0.3086224198,0.2302378714,-0.1212392822,0.0085487394,-0.1927378476,0.1663312465,-0.1916408688,-0.020915607,0.2365944684,-0.423504144,0.3702076972,0.0119346669,0.0343449079,-0.1759414524,-0.1278447807,-0.3771776855,0.1120416224,0.3615404367,-0.0125500839,0.054547511,0.3059005737,-0.1605529636,-0.1137443334,0.1615891904,-0.2779767811,-0.0708428994,-0.2451229244,0.1576605588,-0.2633892298,-0.1346046925,-0.3717961609,0.1198330447,-0.125720799,-0.4005056322,0.10249345,-0.0675686821,-0.3011534512,-0.120644711,0.4150713086,0.8518055677,-0.3355196714,-0.093294248,-0.0604912341,-0.1148239449,0.1904436052,0.1587476283,-0.1758161783,0.1265174299,-0.3173572719,0.0852932408,0.3570447862,-0.6555161476,-0.3502781987,0.3940269649,-0.1724870652,0.1743158102,-0.0031939433,-0.1204576716,-0.105079256,0.0691681951,0.1237685904,-0.0654860586,-0.0663059279,-0.0503761917,-0.142225191,-0.1863250732,-0.1296483576,0.3998391628,0.2517876029,0.0284954812,0.3122981191,0.0064349538,0.056459602,0.1409973055,-0.2018603832,0.2789867222,0.3711987734,0.0615504757,0.0130872354,-0.1971012503,-0.4304905832,0.3101559579,0.0188057944,0.1489988416,-0.3316302598,-0.1355324686,-0.1616557688,0.0700269043,-0.3332709074,-0.1619988829,0.0675913244,0.0231912974,0.1319550425,0.1637164801,-0.2757384181,0.1554542482,-0.2386261225,0.1178620458,-0.2853386402,0.0133797722,-0.0643008947,0.0012760872,-0.1768088341,0.2192420214,0.080203332,-0.3044885695,0.1513867825,0.2341832072,0.0449803136,0.0103129139,0.1267170608,-0.1656969637,0.1939175874,-0.122781679,0.1364119351,0.0978603214,0.2889787257,-0.0309188105,-0.115271233,0.0863264427,0.0478846207,0.1270620823,0.1546972692,-0.2131677121,0.2714660466,-0.0019443034,-0.1627203077,-0.085153833,0.4590184093,0.1962234229,0.1736622304,0.137475118,-0.287707448,-0.1821365952,0.1407131106,0.0629014894,0.1138783246,0.183448568,-0.2541672289,0.2530087829,-0.0514143817,0.1553941816,0.5468078256,0.2619109154,-0.1404587477,0.0524425507,0.1771872938,-0.1258466095,0.10322804,0.1554161906,0.0822554082,0.4120635688,0.1051189676,0.1485672891,-0.2611756325,-0.1467762738,-0.1277178079,0.1874726862,-0.1077103019,0.0245930124,-0.2794661522,-0.5051726699,-0.3058557808,-0.0146225905,-0.2405150235,0.0940984413,-0.0536517166,-0.0190877654,0.237966001,0.1414179057,-0.0791394413,0.0709032565,-0.2862451971,-0.6054201722,0.1051735207,-0.1415938139,-0.0785738155,0.0245204698,0.4939727187,-0.1373359859,0.2790651619,-0.2592563927,-0.0737913251,-0.183931604,-0.1656147689,0.093360886,-0.1004273146,0.3379811049,-0.0167553294,0.2939297259,-0.3947290182,0.0188096389,0.3196304739,-0.1041871458,-0.0589189418,-0.0309407245,-0.1934084743,-0.2796659768,-0.0250590947,-0.5509610772,-0.6147885323,-0.360651046,0.0698504373,0.17899625,-0.0543606766,0.4562773108,0.2112663537,0.4162027538,-0.1866580844,0.1510326415,-0.0113200368,-0.415324688,0.2779581249,-0.1183177978,-0.25416556,0.1727492213,0.1229851395,0.1345045418,-0.2037549615,-0.4326478243,-0.248353377,-0.0015366314,-0.0084353378,-0.0158456843,0.2570870221,0.2936086655,-0.3017704487,-0.0146989375,0.0673354268,-0.0560011193,-0.2546598911,0.0902753398,0.3254760206,-0.2873962522,0.2540574968,-0.1594658345,0.5002567768,0.0868263543,0.0357575938,0.4872640073,-0.1455090642,0.4775304794,-0.3247684538,-0.6194015145,0.0318724401,0.232213676,-0.0674543381,0.3852866292,0.2430349439,-0.0869529545,-0.453145057,-0.2870657742,-0.2236241102,-0.0078404117,0.0051518269,0.3224627674,-0.0569308847,0.0329461023,-0.0538649224,0.0784727857,-0.2104668468,-0.1713650227,0.4470370114,0.0395404547,0.1404958665,-0.1172326878,-0.065638572,-0.0503196083,0.3196733594,0.1239723265,-0.031794101,-0.2450338155,0.0032575652,0.0897798464,-0.083825469,0.6060542464,0.125977397,-0.2148017734,0.0619142763,-0.1242173538,-0.3558597863,-0.1397267729,0.1241109446,0.33854568,0.2068275958,0.5484955311,-0.2395518422,0.0007114509,0.3126532137,0.3900602758,0.0251356196,-0.2338775694,-0.4893727005,-0.4368323386,-0.3035600185,-0.0894210562,0.0003592575,0.0473273285,0.2134342492,-0.0106994277,0.0144244283,0.1609066874,0.1945794821,0.2030138373,0.155582279,0.1789845973,0.1334356666,0.1948364228,0.3523103595,0.2230377346,0.7691653967,-0.1178849488,-0.5521108508,-0.0847059116,0.1034232229,0.2201524526,0.2720788419,-0.0638969094,0.000765383,0.1117356792,-0.1247651801,-0.0085505797,-0.019835202,0.2024696916,0.1292486489,-0.4431892037,-0.3632142246,0.433834672,0.1714629382,0.1282337755,0.2642171383,0.0986125916,0.0996947214,0.2125950456,-0.3209170103,0.9649803042,-0.1976135969,0.0362592861,0.0771165788,0.3622811139,0.6574353576,-0.3590168953,0.0445196517,-0.2230015695,-0.2257985473,0.0043686656,-0.0580592155,0.3891059458,-0.0474551655,-0.2929098308,0.4159952998,0.0153065901,0.3994216025,-0.0599418618,0.0025063592,-0.1263850927,-0.0684088841,-0.3582015336,0.1462122649,-0.1549631804,0.6251106858,-0.2535955906,0.077225551,-0.3234472275,-0.2228563726,-0.2617683113,0.1159332022,-0.4040341079,0.1770790517,0.1299504191,-0.3495531678,0.0338614993,0.5772147775,0.1931933314,0.0745151564,-0.3255049586,0.0933134705,0.1815036386,-0.4349440634,-0.000763265,-0.0813703388,0.2118201107,-0.0388680845,-0.2481323034,-0.0548109598,0.0382058732,-0.0521591567,-0.021917779,0.1444230974,0.106216073,-0.1031226739,-0.6530137658,-0.08413212,0.1828003079,-0.0499580987,0.1217646897,0.1779325753,-0.0956843421,0.0431085005,0.166189611,-0.3415124714,-0.1132987887,0.5113555789,-0.1856525689,-0.2996102273,0.36169222,0.3614443243,-0.1663857102,-0.2654777765,0.1330931485,-0.0957556963,-0.3206220567,-0.094085373,0.2909229994,0.1932491064,0.109358795,0.3781818449,0.1973588765,-0.4038244188,0.0249055419,-0.4093193114,-0.1236805916,0.1327896863,0.1057254747,-0.0049707382,0.3236237168,-0.0337340832,0.3437898457,-0.1685631275,-0.2829583585,0.1128235459,-0.3283310235,0.0372480191,0.1804626435,0.2854779363,0.3007024825,-0.0607787408,0.1297686547,0.0727349967,-0.0878174677,-0.281730026,0.0913285688,0.1639157087,0.1169719025,0.0177246872,-0.0559778698,-0.2437392473,0.0019007773,-0.0929029435,0.0203969926,0.3053594828,0.0740510672,-0.0741511881,-0.0874260813,0.044532463,0.1253588796,0.2892073095,0.0188889615,0.1319877803,0.1818603426,-0.0731024146,0.1626294106,-0.0380570926,-0.0687274858,-0.038027823,0.0540184937,0.1009544656,0.1793214828,-0.2335824966,0.3291619122,0.2407063395,0.1341376454,0.3426393867,-0.285951525,0.3464767039,0.3896774352,0.1448640674,-0.4216090143,0.0075634248,0.5237147808,0.0645895302,0.0583187714,0.3181272149,-0.0544047654,0.112127997,0.2146092653,0.1580101699,0.5828193426,-0.0742316619,0.2061963528,0.3621203005,0.1632302105,0.1010898277,-0.0977707878,0.1651702672,0.1123132631,0.3069590628,-0.3984757364,0.2375390828,-0.1787928343,0.2657759786,0.1091773808,-0.5371880531,0.343121171,0.2038057297,0.046059452,-0.1184164435,-0.1416187882,0.4960412383,-0.3275724351,0.1069446057,-0.2543179095,0.1637641639,-0.1826739609,-0.0228513405,0.0249206182,-0.2967809439,0.1943347603,-0.0360241644,0.0471088178,-0.1881922036,-0.19042629,0.1351135075,-0.3087254167,-0.0898034722,0.1096820459,0.0639084801,0.027124865,-0.1606835872,0.2069001347,-0.0040159156,0.142120868,-0.0576762073,0.5198544264,0.5531443954,0.5145381689,0.1421940476,0.1212123334,-0.1646075845,-0.111194402,-0.1009603366,0.2181242257,-0.1094547212,0.1719533652,0.2905450165,0.2046411932,-0.1378827393,0.2495927811,0.0381854102,0.0015574556,-0.1165596843,0.1595306993,-0.3742489517,-0.0799477845,-0.3764404655,0.2038151324,-0.5713078976,0.2765871882,0.339463979,0.1170363799,0.0929874927,-0.1818301827,0.0704221949,0.034915451,0.2177680284,0.5072876811,0.4780167043,-0.3832007051,-0.0000326404,-0.7233870625,0.1582027525,-0.1702185571,0.0586374924,-0.217623204,-0.02439107,0.0711091384,0.2534762323,-0.2483322322,-0.1785003394,0.1768391281,0.0696047843,-0.1952718198,-0.2398768812,-0.1178399548,0.1501638591,-0.00287463,-0.2730621696,0.0009602974,-0.4926660657,-0.033117082,-0.1619898826,-0.1544716507,-0.2251114398,-0.292810142,0.204383567,-0.1892799586,0.3745802641,-0.061655391,0.0555200353,-0.2636980712,-0.3919501603,-0.1114182025,0.3940665126,0.1919886768,0.0916118547,-0.3440899253,-0.1759020239,-0.1788997799,-0.039456889,-0.0636948496,0.0664487928,-0.0992949978,-0.2773171663,-0.0121658267,-0.0423315503,-0.1503696591,0.1343099773,-0.1156885326,-0.0959422141,-0.2692499161,-0.2346588522,0.5547620058,-0.2860987186,0.1067291424,-0.1147617251,0.0544556156,0.0170521513,-0.2258203924,-0.4898255169,0.2988792062,0.3092295825,0.1333564818,-0.2277311534,0.1967218071,-0.1528733671,0.0018915375,-0.0348544121,0.0822892785,-0.0721235424,0.0014291957,0.1483232379,-0.0604110472]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1422","title":"Can't map dataset (loaded from csv)","comments":"Please could you post the whole script? I can't reproduce your issue. After updating the feature names\/labels to match with the data, everything works fine for me. Try to update datasets\/transformers to the newest version.","body":"Hello! I am trying to load single csv file with two columns: ('label': str, 'text' str), where is label is str of two possible classes.\r\n\r\nBelow steps are similar with [this notebook](https:\/\/colab.research.google.com\/drive\/1-JIJlao4dI-Ilww_NnTc0rxtp-ymgDgM?usp=sharing), where bert model and tokenizer are used to classify lmdb loaded dataset. Only one difference it is the dataset loaded from .csv file.\r\nHere is how I load it:\r\n\r\n```python\r\ndata_path = 'data.csv'\r\ndata = pd.read_csv(data_path)\r\n\r\n# process class name to indices\r\nclasses = ['neg', 'pos']\r\nclass_to_idx = { cl: i for i, cl in enumerate(classes) }\r\n\r\n# now data is like {'label': int, 'text' str}\r\ndata['label'] = data['label'].apply(lambda x: class_to_idx[x])\r\n\r\n# load dataset and map it with defined `tokenize` function\r\nfeatures = Features({\r\n target: ClassLabel(num_classes=2, names=['neg', 'pos'], names_file=None, id=None),\r\n feature: Value(dtype='string', id=None),\r\n})\r\ndataset = Dataset.from_pandas(data, features=features)\r\ndataset.map(tokenize, batched=True, batch_size=len(dataset))\r\n```\r\n\r\nIt ruins on the last line with following error:\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<ipython-input-112-32b6275ce418> in <module>()\r\n 9 })\r\n 10 dataset = Dataset.from_pandas(data, features=features)\r\n---> 11 dataset.map(tokenizer, batched=True, batch_size=len(dataset))\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1237 test_inputs = self[:2] if batched else self[0]\r\n 1238 test_indices = [0, 1] if batched else 0\r\n-> 1239 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1240 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1241 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1208 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1209 processed_inputs = (\r\n-> 1210 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1211 )\r\n 1212 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/tokenization_utils_base.py in __call__(self, text, text_pair, add_special_tokens, padding, truncation, max_length, stride, is_split_into_words, pad_to_multiple_of, return_tensors, return_token_type_ids, return_attention_mask, return_overflowing_tokens, return_special_tokens_mask, return_offsets_mapping, return_length, verbose, **kwargs)\r\n 2281 )\r\n 2282 ), (\r\n-> 2283 \"text input must of type `str` (single example), `List[str]` (batch or single pretokenized example) \"\r\n 2284 \"or `List[List[str]]` (batch of pretokenized examples).\"\r\n 2285 )\r\n\r\nAssertionError: text input must of type `str` (single example), `List[str]` (batch or single pretokenized example) or `List[List[str]]` (batch of pretokenized examples).\r\n```\r\n\r\nwhich I think is not expected. I also tried the same steps using `Dataset.from_csv` which resulted in the same error.\r\n\r\nFor reproducing this, I used [this dataset from kaggle](https:\/\/www.kaggle.com\/team-ai\/spam-text-message-classification)","comment_length":35,"text":"Can't map dataset (loaded from csv)\nHello! I am trying to load single csv file with two columns: ('label': str, 'text' str), where is label is str of two possible classes.\r\n\r\nBelow steps are similar with [this notebook](https:\/\/colab.research.google.com\/drive\/1-JIJlao4dI-Ilww_NnTc0rxtp-ymgDgM?usp=sharing), where bert model and tokenizer are used to classify lmdb loaded dataset. Only one difference it is the dataset loaded from .csv file.\r\nHere is how I load it:\r\n\r\n```python\r\ndata_path = 'data.csv'\r\ndata = pd.read_csv(data_path)\r\n\r\n# process class name to indices\r\nclasses = ['neg', 'pos']\r\nclass_to_idx = { cl: i for i, cl in enumerate(classes) }\r\n\r\n# now data is like {'label': int, 'text' str}\r\ndata['label'] = data['label'].apply(lambda x: class_to_idx[x])\r\n\r\n# load dataset and map it with defined `tokenize` function\r\nfeatures = Features({\r\n target: ClassLabel(num_classes=2, names=['neg', 'pos'], names_file=None, id=None),\r\n feature: Value(dtype='string', id=None),\r\n})\r\ndataset = Dataset.from_pandas(data, features=features)\r\ndataset.map(tokenize, batched=True, batch_size=len(dataset))\r\n```\r\n\r\nIt ruins on the last line with following error:\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<ipython-input-112-32b6275ce418> in <module>()\r\n 9 })\r\n 10 dataset = Dataset.from_pandas(data, features=features)\r\n---> 11 dataset.map(tokenizer, batched=True, batch_size=len(dataset))\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1237 test_inputs = self[:2] if batched else self[0]\r\n 1238 test_indices = [0, 1] if batched else 0\r\n-> 1239 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1240 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1241 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1208 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1209 processed_inputs = (\r\n-> 1210 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1211 )\r\n 1212 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/tokenization_utils_base.py in __call__(self, text, text_pair, add_special_tokens, padding, truncation, max_length, stride, is_split_into_words, pad_to_multiple_of, return_tensors, return_token_type_ids, return_attention_mask, return_overflowing_tokens, return_special_tokens_mask, return_offsets_mapping, return_length, verbose, **kwargs)\r\n 2281 )\r\n 2282 ), (\r\n-> 2283 \"text input must of type `str` (single example), `List[str]` (batch or single pretokenized example) \"\r\n 2284 \"or `List[List[str]]` (batch of pretokenized examples).\"\r\n 2285 )\r\n\r\nAssertionError: text input must of type `str` (single example), `List[str]` (batch or single pretokenized example) or `List[List[str]]` (batch of pretokenized examples).\r\n```\r\n\r\nwhich I think is not expected. I also tried the same steps using `Dataset.from_csv` which resulted in the same error.\r\n\r\nFor reproducing this, I used [this dataset from kaggle](https:\/\/www.kaggle.com\/team-ai\/spam-text-message-classification)\nPlease could you post the whole script? I can't reproduce your issue. After updating the feature names\/labels to match with the data, everything works fine for me. Try to update datasets\/transformers to the newest version.","embeddings":[-0.0363199227,-0.0493307747,0.0591480546,0.1960300207,0.3175801933,0.1296344399,0.6318867803,0.3784266114,0.2701599002,-0.0147236995,-0.2211088985,0.26376459,-0.0631853342,-0.0150062218,-0.0317861624,-0.2046317905,0.1337798685,0.2118782401,-0.1829666644,-0.0350301787,-0.2907462418,0.0249158051,-0.1068111807,0.2036802769,0.0197142474,-0.2090110034,-0.0746262372,0.0459839106,-0.2114800513,-0.2632806599,0.2159728855,-0.0493636914,0.5007882714,0.435937047,-0.0001207136,-0.0859811977,-0.1823789477,-0.1009574682,-0.0814799741,-0.3854593635,-0.1374601573,0.0300164334,0.1163092926,-0.2970100343,-0.0744085535,0.0227589831,-0.0245813169,-0.1638604254,0.3432323635,0.3329411149,0.1156951711,-0.0070006349,0.1165167913,-0.1090124175,0.3139908314,-0.0830129161,-0.0401910394,0.167835325,0.315456003,-0.4825088978,-0.0649220943,0.1928553283,-0.3011556566,-0.0419434011,0.4646375477,0.2137138247,0.158956483,-0.2898328602,0.1365338266,0.1331964433,0.4657473564,-0.1151477545,-0.219532147,0.0547608286,0.0344038568,-0.2702280879,0.02838636,0.0456837155,-0.1139841303,0.1086886153,-0.1024555862,-0.1770985872,-0.0279027931,0.2061924785,-0.4087243676,0.2057975978,-0.0557475276,0.2325825244,0.0402006321,-0.2316696346,-0.155206576,-0.0399327539,0.2145067751,0.5941803455,-0.1974852681,-0.1333898157,0.0279572308,-0.3895782828,-0.1100072786,-0.2322983891,-0.1188440323,-0.1543679386,-0.1276846826,0.2993394732,0.2697301507,0.5647977591,0.0272395723,0.5422596931,0.2139974535,0.1010005549,-0.2506912947,0.0377508141,-0.1541895866,-0.2827709019,0.2870100737,0.0581633784,-0.0641857162,-0.0956934169,-0.329421699,0.2402531505,-0.415389061,0.1328236461,0.0743058175,0.6524133086,0.2075130641,-0.0942461863,0.269875139,0.2843061686,0.1186549515,0.0051627969,-0.0669963583,0.3031971455,-0.3530656993,-0.1006992534,0.1698996574,0.0514027327,0.0464357473,0.3085812926,-0.0923380554,-0.0356982835,-0.032873325,-0.1898667365,0.1616642773,0.1182357222,0.0726162568,0.3539725244,0.4711993337,-0.3452658951,-0.1083572954,0.2828636467,-0.4599839449,-0.0054730126,-0.1010404974,0.0946433693,0.1262521148,0.2099694759,-0.4163970351,0.1459005773,0.3574437201,-0.3219867051,-0.19480443,-0.3051259816,-0.0836994946,-0.2269810587,0.2432650626,0.2348656654,-0.5438092947,0.1403555721,0.0783428252,0.3336324692,0.2918049991,0.3815695941,-0.2207279652,0.2234846205,-0.2278954387,0.3127995133,0.2383131385,-0.3298654556,-0.3597221076,0.312369585,-0.0331654698,-0.0880806893,0.0080464864,-0.0217954461,0.4882386625,-0.0339082405,0.1162630394,0.0372320972,-0.152098611,0.0862343982,-0.1676580757,-0.0399070941,0.4897908568,-0.1700118482,0.2294394821,0.1159251705,0.1349700987,0.2924374938,0.0103068613,-0.2915747464,0.1749004275,-0.0661578178,0.1333984882,-0.0705089644,0.0366534069,-0.2851987481,-0.4596106112,0.0767471641,0.0754321218,0.1340249181,-0.1838854849,-0.1106934026,-0.2779102921,-0.0232510995,-0.0741265714,-0.1999337673,0.0292622577,-0.0154434657,-0.256698966,-0.055601567,-0.0836829692,-0.2426069677,-0.0368225984,0.1588586122,-0.0782571733,0.0076647373,-0.1949110031,0.0822894722,-0.0772336274,0.04394456,0.1950021982,-0.0023535059,-0.1986827552,0.0559201054,0.0518036969,-0.0809343755,-0.0811867788,-0.0375422128,0.2180506885,-0.1274510622,-0.1690516621,-0.0750652924,0.2907112837,-0.1380863637,-0.252577126,0.3186502755,0.0168296322,0.3122312129,0.1154069602,0.1145328283,0.2599707246,0.0418411791,0.0865600631,-0.3737351894,-0.1926257759,0.2935028672,0.2589741647,0.154698506,-0.0777677372,-0.0609958172,0.142658025,-0.1335907578,0.0308212563,0.0094727771,-0.199971512,-0.0852358714,0.2894076705,-0.0857502222,0.1652871817,0.0406221375,0.0625380576,-0.2115490884,-0.2276137322,0.0822504163,0.0605193712,-0.2268910855,-0.0607029349,0.0529958569,0.0683715418,0.0882520303,-0.0817593038,-0.0190588739,0.2863254249,0.5847138166,-0.390991956,-0.0683166012,-0.214716047,0.1170497239,-0.1058593616,-0.496304363,-0.1311950833,-0.1060906202,-0.2085026801,0.2625114024,0.0616770796,0.2166935802,0.0551218204,0.0939407945,0.3615973592,0.0393193439,0.1899323314,-0.174316138,-0.2117988765,0.0167031586,0.2753624618,0.0007117973,0.1711568981,-0.0011612893,-0.2233868986,-0.3630214334,-0.4320555627,0.0203379411,-0.0752345324,0.3070160747,0.207945168,0.1353277117,-0.3077080846,-0.3446017802,0.1428049356,0.0589852557,-0.1098983586,0.0713910237,0.0034607612,-0.3001106679,0.1107848436,-0.556061089,-0.166413784,-0.0522097237,-0.1271596849,-0.2033301145,0.2183383703,-0.3293950558,0.0247675478,0.0507252067,0.1594140083,0.0182025936,-0.1527701765,-0.1177452058,0.2939305305,-0.1057843566,-0.2996169925,0.0283035357,-0.1588464677,0.2352195829,-0.0126873646,-0.1893120706,-0.1807990968,-0.0377765708,0.0976352319,-0.0214477852,0.071006991,0.0548175536,0.2307506055,-0.0060728281,0.0111752106,-0.1993457675,0.077370353,-0.0068606897,0.3096449971,-0.1190208122,0.5365262032,-0.001956112,0.9932539463,0.0855753198,-0.1577251405,0.4048094749,-0.0419143699,0.1154531464,-0.1302984357,-0.7429162264,-0.0135826888,-0.189058274,-0.1387497187,0.3288029432,-0.1925850511,-0.3625994623,-0.0441093445,0.1770317703,-0.2216781229,-0.1894226223,0.2895946205,-0.2774533927,0.0420526303,0.157803759,-0.0477606952,-0.3704247773,-0.213369891,-0.1930531114,0.261454016,0.0047652978,-0.0229739621,0.0135734361,-0.2664872706,-0.5215991735,0.2433543801,0.1242299303,0.4188009501,-0.2098043263,0.1426520944,0.0610149801,0.1687998325,0.6888326406,-0.0195550825,0.1125068963,0.1391798258,-0.1512920707,-0.0791232362,0.0676735938,-0.1994972527,0.0883932114,0.346932143,0.5298926234,-0.0986488611,0.0164254028,0.0190519542,0.0271240026,-0.2292007655,-0.1593633741,-0.1253409684,0.1321070492,-0.3919849992,0.1279266626,0.2497394979,0.2374415994,0.0471080169,-0.2645955384,-0.0324906334,-0.1234860867,0.4109721184,0.4043199122,-0.0464400724,0.2494567484,0.0538211241,-0.1417577565,0.3292892277,0.408164531,0.3600576818,-0.4821265936,-0.6383057833,0.0391044766,0.0401061065,0.3005616069,0.0823301002,-0.1973975748,0.1902044117,-0.2747591436,-0.0287708528,-0.1576127112,0.4213906229,0.3624390066,0.2754831612,-0.4533807039,-0.4862160385,0.1569014639,-0.0935743228,-0.3375833333,0.035652183,0.1259561479,-0.4417129159,0.4667490423,-0.0186617859,0.8538818359,-0.4546049833,0.0962253436,0.1599946618,0.2531588972,0.7502367496,-0.0199419931,-0.0957798287,-0.2133440971,0.0240359288,-0.1087242588,0.1119210124,0.1089091524,0.253631562,-0.222307831,0.4156003296,0.0304187555,0.4982374907,0.1266560256,0.0278673694,0.4766612053,-0.3088310659,-0.3115204573,0.0545618087,-0.0697770715,0.0356489271,-0.0045600785,-0.0352154225,-0.1696330607,-0.2989766002,-0.2679096162,0.0762702227,-0.1809964329,0.3267801404,-0.3523544669,-0.3634991944,0.4128607213,0.4410256445,0.0437680222,0.1301425397,-0.1129651517,0.0003238506,0.0872352421,0.0596228056,-0.3403808475,-0.0629584566,0.3871629238,0.0844535306,-0.1124230325,0.4013395011,-0.0449005365,-0.1486762315,0.0666267574,0.174853459,0.5714659095,-0.5062837005,-0.3370099664,-0.2767016888,0.0419976003,-0.3917419612,0.0281824637,-0.1948964596,-0.1239202023,0.0166914072,0.0216483809,-0.4055575728,0.1933333427,0.6294291615,-0.0493522808,0.0708502084,0.7852793932,0.0728306323,-0.1149651036,-0.1665186286,0.0860115662,0.5840911269,-0.14866817,0.233825773,0.0933881551,-0.0641303733,0.0181239638,0.4313964546,0.5497334599,-0.0057319538,0.1652353555,-0.4803129435,-0.4752922058,0.0116617782,-0.0688422099,0.1330046952,0.4147024155,0.158065632,-0.1791446656,0.5363548994,-0.1994547248,0.0471610166,-0.179473117,0.2883890867,0.2534252107,0.1340298951,-0.2279547751,-0.0515116639,0.1157054976,-0.125923261,0.0553603284,-0.1513466686,-0.180310443,0.162634477,0.0242551975,-0.0538024716,-0.1285189986,-0.3979252875,0.093452543,-0.4114852846,0.2004027963,0.2641579807,0.0309233088,0.0558289401,0.0943148509,-0.2631490231,-0.1857144237,0.038532149,-0.1030581146,-0.1144090891,0.0192986745,0.1908440739,0.0535025112,-0.034686137,-0.025311904,-0.0308934189,-0.0405474678,-0.0744228438,0.3435593843,-0.112395674,0.2276649028,0.2359961867,0.1910891086,0.0069904225,-0.1660908014,0.0699272603,0.3582183421,0.043475911,-0.4803177118,0.0895875469,0.1368270516,-0.0299050808,0.0883752331,0.1709184498,-0.0528236739,0.0629411787,-0.1101972163,-0.2367053479,0.3873440623,-0.0596081205,0.1175315008,0.5157994032,-0.1992423832,0.2741678655,0.4069995284,0.2518590093,0.1012815014,0.5119723678,-0.1945046633,0.16651842,-0.2078119218,0.3426595926,0.0851847082,-0.3729570508,0.1948928088,0.2040794939,-0.1129731014,0.1360031515,0.0149118369,0.6207628846,-0.4058395922,-0.2460162491,0.0260691065,0.362469703,-0.3382800817,-0.2492062151,-0.0257639401,-0.2772208452,0.1468814015,-0.0313286819,-0.3017627895,-0.0870868564,0.1407350451,0.0974299833,-0.0951563418,-0.388295114,-0.3287885785,0.109848842,-0.0754377097,-0.1581839174,0.428452611,-0.149591133,0.0571290404,0.0845344439,0.1054724678,0.4288723171,0.6453410387,0.0887052193,-0.1818567961,0.1211570278,-0.1629321724,-0.0693932772,0.0466387048,0.0910195038,0.0710614994,0.2339627594,0.0203048363,-0.0890840515,0.2069789171,0.1943567842,0.2529907823,0.1798135489,0.0852779597,-0.1781769693,-0.1139146835,-0.1873798072,0.0658743605,-0.3288325369,0.2687665224,0.0924542397,-0.1591729522,0.2595170438,0.0588992722,0.0379386842,0.2263917625,0.1651692241,0.2591685355,-0.092806749,-0.6077032685,0.1426296532,-0.8178735971,0.3019883335,-0.2084612399,-0.0313831493,-0.2205667347,0.0342906564,0.2456472665,-0.127258271,0.1502779424,0.023984829,0.0380285643,0.3548089564,-0.0126671987,-0.0467083417,-0.2404095232,0.1079238057,-0.1733584404,-0.3466698527,0.2856391668,0.1229375228,-0.1231837273,-0.195590511,0.178684473,0.0083468761,0.0940715745,0.4659883976,0.1278936714,0.7292595506,-0.0321749821,0.2721481025,0.0740848854,-0.1821622849,-0.1105566993,0.0749446675,0.202351585,0.2000992149,-0.195001781,-0.1445555091,-0.2320061326,-0.0755844414,-0.0144242607,-0.2899771929,-0.6900325418,0.1493444592,-0.0430814438,-0.0434927978,-0.018656997,0.2625669539,-0.0919618383,0.1879252493,-0.0971285477,-0.4130038917,0.3670980036,-0.554874301,-0.4784332514,-0.024199456,-0.1244087294,0.1122574136,-0.0611881986,-0.6009425521,-0.0534130372,0.2493426055,-0.0713379085,-0.4005272985,-0.0427324288,-0.0534496233,0.0576560237,-0.1844071895,0.4424070716,0.1898754984,-0.1703061908,-0.0935344249,-0.4564055502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1422","title":"Can't map dataset (loaded from csv)","comments":"Actually, the problem was how `tokenize` function was defined. This was completely my side mistake, so there are really no needs in this issue anymore","body":"Hello! I am trying to load single csv file with two columns: ('label': str, 'text' str), where is label is str of two possible classes.\r\n\r\nBelow steps are similar with [this notebook](https:\/\/colab.research.google.com\/drive\/1-JIJlao4dI-Ilww_NnTc0rxtp-ymgDgM?usp=sharing), where bert model and tokenizer are used to classify lmdb loaded dataset. Only one difference it is the dataset loaded from .csv file.\r\nHere is how I load it:\r\n\r\n```python\r\ndata_path = 'data.csv'\r\ndata = pd.read_csv(data_path)\r\n\r\n# process class name to indices\r\nclasses = ['neg', 'pos']\r\nclass_to_idx = { cl: i for i, cl in enumerate(classes) }\r\n\r\n# now data is like {'label': int, 'text' str}\r\ndata['label'] = data['label'].apply(lambda x: class_to_idx[x])\r\n\r\n# load dataset and map it with defined `tokenize` function\r\nfeatures = Features({\r\n target: ClassLabel(num_classes=2, names=['neg', 'pos'], names_file=None, id=None),\r\n feature: Value(dtype='string', id=None),\r\n})\r\ndataset = Dataset.from_pandas(data, features=features)\r\ndataset.map(tokenize, batched=True, batch_size=len(dataset))\r\n```\r\n\r\nIt ruins on the last line with following error:\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<ipython-input-112-32b6275ce418> in <module>()\r\n 9 })\r\n 10 dataset = Dataset.from_pandas(data, features=features)\r\n---> 11 dataset.map(tokenizer, batched=True, batch_size=len(dataset))\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1237 test_inputs = self[:2] if batched else self[0]\r\n 1238 test_indices = [0, 1] if batched else 0\r\n-> 1239 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1240 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1241 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1208 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1209 processed_inputs = (\r\n-> 1210 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1211 )\r\n 1212 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/tokenization_utils_base.py in __call__(self, text, text_pair, add_special_tokens, padding, truncation, max_length, stride, is_split_into_words, pad_to_multiple_of, return_tensors, return_token_type_ids, return_attention_mask, return_overflowing_tokens, return_special_tokens_mask, return_offsets_mapping, return_length, verbose, **kwargs)\r\n 2281 )\r\n 2282 ), (\r\n-> 2283 \"text input must of type `str` (single example), `List[str]` (batch or single pretokenized example) \"\r\n 2284 \"or `List[List[str]]` (batch of pretokenized examples).\"\r\n 2285 )\r\n\r\nAssertionError: text input must of type `str` (single example), `List[str]` (batch or single pretokenized example) or `List[List[str]]` (batch of pretokenized examples).\r\n```\r\n\r\nwhich I think is not expected. I also tried the same steps using `Dataset.from_csv` which resulted in the same error.\r\n\r\nFor reproducing this, I used [this dataset from kaggle](https:\/\/www.kaggle.com\/team-ai\/spam-text-message-classification)","comment_length":25,"text":"Can't map dataset (loaded from csv)\nHello! I am trying to load single csv file with two columns: ('label': str, 'text' str), where is label is str of two possible classes.\r\n\r\nBelow steps are similar with [this notebook](https:\/\/colab.research.google.com\/drive\/1-JIJlao4dI-Ilww_NnTc0rxtp-ymgDgM?usp=sharing), where bert model and tokenizer are used to classify lmdb loaded dataset. Only one difference it is the dataset loaded from .csv file.\r\nHere is how I load it:\r\n\r\n```python\r\ndata_path = 'data.csv'\r\ndata = pd.read_csv(data_path)\r\n\r\n# process class name to indices\r\nclasses = ['neg', 'pos']\r\nclass_to_idx = { cl: i for i, cl in enumerate(classes) }\r\n\r\n# now data is like {'label': int, 'text' str}\r\ndata['label'] = data['label'].apply(lambda x: class_to_idx[x])\r\n\r\n# load dataset and map it with defined `tokenize` function\r\nfeatures = Features({\r\n target: ClassLabel(num_classes=2, names=['neg', 'pos'], names_file=None, id=None),\r\n feature: Value(dtype='string', id=None),\r\n})\r\ndataset = Dataset.from_pandas(data, features=features)\r\ndataset.map(tokenize, batched=True, batch_size=len(dataset))\r\n```\r\n\r\nIt ruins on the last line with following error:\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<ipython-input-112-32b6275ce418> in <module>()\r\n 9 })\r\n 10 dataset = Dataset.from_pandas(data, features=features)\r\n---> 11 dataset.map(tokenizer, batched=True, batch_size=len(dataset))\r\n\r\n2 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1237 test_inputs = self[:2] if batched else self[0]\r\n 1238 test_indices = [0, 1] if batched else 0\r\n-> 1239 update_data = does_function_return_dict(test_inputs, test_indices)\r\n 1240 logger.info(\"Testing finished, running the mapping function on the dataset\")\r\n 1241 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in does_function_return_dict(inputs, indices)\r\n 1208 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]\r\n 1209 processed_inputs = (\r\n-> 1210 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n 1211 )\r\n 1212 does_return_dict = isinstance(processed_inputs, Mapping)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/tokenization_utils_base.py in __call__(self, text, text_pair, add_special_tokens, padding, truncation, max_length, stride, is_split_into_words, pad_to_multiple_of, return_tensors, return_token_type_ids, return_attention_mask, return_overflowing_tokens, return_special_tokens_mask, return_offsets_mapping, return_length, verbose, **kwargs)\r\n 2281 )\r\n 2282 ), (\r\n-> 2283 \"text input must of type `str` (single example), `List[str]` (batch or single pretokenized example) \"\r\n 2284 \"or `List[List[str]]` (batch of pretokenized examples).\"\r\n 2285 )\r\n\r\nAssertionError: text input must of type `str` (single example), `List[str]` (batch or single pretokenized example) or `List[List[str]]` (batch of pretokenized examples).\r\n```\r\n\r\nwhich I think is not expected. I also tried the same steps using `Dataset.from_csv` which resulted in the same error.\r\n\r\nFor reproducing this, I used [this dataset from kaggle](https:\/\/www.kaggle.com\/team-ai\/spam-text-message-classification)\nActually, the problem was how `tokenize` function was defined. This was completely my side mistake, so there are really no needs in this issue anymore","embeddings":[-0.0363199227,-0.0493307747,0.0591480546,0.1960300207,0.3175801933,0.1296344399,0.6318867803,0.3784266114,0.2701599002,-0.0147236995,-0.2211088985,0.26376459,-0.0631853342,-0.0150062218,-0.0317861624,-0.2046317905,0.1337798685,0.2118782401,-0.1829666644,-0.0350301787,-0.2907462418,0.0249158051,-0.1068111807,0.2036802769,0.0197142474,-0.2090110034,-0.0746262372,0.0459839106,-0.2114800513,-0.2632806599,0.2159728855,-0.0493636914,0.5007882714,0.435937047,-0.0001207136,-0.0859811977,-0.1823789477,-0.1009574682,-0.0814799741,-0.3854593635,-0.1374601573,0.0300164334,0.1163092926,-0.2970100343,-0.0744085535,0.0227589831,-0.0245813169,-0.1638604254,0.3432323635,0.3329411149,0.1156951711,-0.0070006349,0.1165167913,-0.1090124175,0.3139908314,-0.0830129161,-0.0401910394,0.167835325,0.315456003,-0.4825088978,-0.0649220943,0.1928553283,-0.3011556566,-0.0419434011,0.4646375477,0.2137138247,0.158956483,-0.2898328602,0.1365338266,0.1331964433,0.4657473564,-0.1151477545,-0.219532147,0.0547608286,0.0344038568,-0.2702280879,0.02838636,0.0456837155,-0.1139841303,0.1086886153,-0.1024555862,-0.1770985872,-0.0279027931,0.2061924785,-0.4087243676,0.2057975978,-0.0557475276,0.2325825244,0.0402006321,-0.2316696346,-0.155206576,-0.0399327539,0.2145067751,0.5941803455,-0.1974852681,-0.1333898157,0.0279572308,-0.3895782828,-0.1100072786,-0.2322983891,-0.1188440323,-0.1543679386,-0.1276846826,0.2993394732,0.2697301507,0.5647977591,0.0272395723,0.5422596931,0.2139974535,0.1010005549,-0.2506912947,0.0377508141,-0.1541895866,-0.2827709019,0.2870100737,0.0581633784,-0.0641857162,-0.0956934169,-0.329421699,0.2402531505,-0.415389061,0.1328236461,0.0743058175,0.6524133086,0.2075130641,-0.0942461863,0.269875139,0.2843061686,0.1186549515,0.0051627969,-0.0669963583,0.3031971455,-0.3530656993,-0.1006992534,0.1698996574,0.0514027327,0.0464357473,0.3085812926,-0.0923380554,-0.0356982835,-0.032873325,-0.1898667365,0.1616642773,0.1182357222,0.0726162568,0.3539725244,0.4711993337,-0.3452658951,-0.1083572954,0.2828636467,-0.4599839449,-0.0054730126,-0.1010404974,0.0946433693,0.1262521148,0.2099694759,-0.4163970351,0.1459005773,0.3574437201,-0.3219867051,-0.19480443,-0.3051259816,-0.0836994946,-0.2269810587,0.2432650626,0.2348656654,-0.5438092947,0.1403555721,0.0783428252,0.3336324692,0.2918049991,0.3815695941,-0.2207279652,0.2234846205,-0.2278954387,0.3127995133,0.2383131385,-0.3298654556,-0.3597221076,0.312369585,-0.0331654698,-0.0880806893,0.0080464864,-0.0217954461,0.4882386625,-0.0339082405,0.1162630394,0.0372320972,-0.152098611,0.0862343982,-0.1676580757,-0.0399070941,0.4897908568,-0.1700118482,0.2294394821,0.1159251705,0.1349700987,0.2924374938,0.0103068613,-0.2915747464,0.1749004275,-0.0661578178,0.1333984882,-0.0705089644,0.0366534069,-0.2851987481,-0.4596106112,0.0767471641,0.0754321218,0.1340249181,-0.1838854849,-0.1106934026,-0.2779102921,-0.0232510995,-0.0741265714,-0.1999337673,0.0292622577,-0.0154434657,-0.256698966,-0.055601567,-0.0836829692,-0.2426069677,-0.0368225984,0.1588586122,-0.0782571733,0.0076647373,-0.1949110031,0.0822894722,-0.0772336274,0.04394456,0.1950021982,-0.0023535059,-0.1986827552,0.0559201054,0.0518036969,-0.0809343755,-0.0811867788,-0.0375422128,0.2180506885,-0.1274510622,-0.1690516621,-0.0750652924,0.2907112837,-0.1380863637,-0.252577126,0.3186502755,0.0168296322,0.3122312129,0.1154069602,0.1145328283,0.2599707246,0.0418411791,0.0865600631,-0.3737351894,-0.1926257759,0.2935028672,0.2589741647,0.154698506,-0.0777677372,-0.0609958172,0.142658025,-0.1335907578,0.0308212563,0.0094727771,-0.199971512,-0.0852358714,0.2894076705,-0.0857502222,0.1652871817,0.0406221375,0.0625380576,-0.2115490884,-0.2276137322,0.0822504163,0.0605193712,-0.2268910855,-0.0607029349,0.0529958569,0.0683715418,0.0882520303,-0.0817593038,-0.0190588739,0.2863254249,0.5847138166,-0.390991956,-0.0683166012,-0.214716047,0.1170497239,-0.1058593616,-0.496304363,-0.1311950833,-0.1060906202,-0.2085026801,0.2625114024,0.0616770796,0.2166935802,0.0551218204,0.0939407945,0.3615973592,0.0393193439,0.1899323314,-0.174316138,-0.2117988765,0.0167031586,0.2753624618,0.0007117973,0.1711568981,-0.0011612893,-0.2233868986,-0.3630214334,-0.4320555627,0.0203379411,-0.0752345324,0.3070160747,0.207945168,0.1353277117,-0.3077080846,-0.3446017802,0.1428049356,0.0589852557,-0.1098983586,0.0713910237,0.0034607612,-0.3001106679,0.1107848436,-0.556061089,-0.166413784,-0.0522097237,-0.1271596849,-0.2033301145,0.2183383703,-0.3293950558,0.0247675478,0.0507252067,0.1594140083,0.0182025936,-0.1527701765,-0.1177452058,0.2939305305,-0.1057843566,-0.2996169925,0.0283035357,-0.1588464677,0.2352195829,-0.0126873646,-0.1893120706,-0.1807990968,-0.0377765708,0.0976352319,-0.0214477852,0.071006991,0.0548175536,0.2307506055,-0.0060728281,0.0111752106,-0.1993457675,0.077370353,-0.0068606897,0.3096449971,-0.1190208122,0.5365262032,-0.001956112,0.9932539463,0.0855753198,-0.1577251405,0.4048094749,-0.0419143699,0.1154531464,-0.1302984357,-0.7429162264,-0.0135826888,-0.189058274,-0.1387497187,0.3288029432,-0.1925850511,-0.3625994623,-0.0441093445,0.1770317703,-0.2216781229,-0.1894226223,0.2895946205,-0.2774533927,0.0420526303,0.157803759,-0.0477606952,-0.3704247773,-0.213369891,-0.1930531114,0.261454016,0.0047652978,-0.0229739621,0.0135734361,-0.2664872706,-0.5215991735,0.2433543801,0.1242299303,0.4188009501,-0.2098043263,0.1426520944,0.0610149801,0.1687998325,0.6888326406,-0.0195550825,0.1125068963,0.1391798258,-0.1512920707,-0.0791232362,0.0676735938,-0.1994972527,0.0883932114,0.346932143,0.5298926234,-0.0986488611,0.0164254028,0.0190519542,0.0271240026,-0.2292007655,-0.1593633741,-0.1253409684,0.1321070492,-0.3919849992,0.1279266626,0.2497394979,0.2374415994,0.0471080169,-0.2645955384,-0.0324906334,-0.1234860867,0.4109721184,0.4043199122,-0.0464400724,0.2494567484,0.0538211241,-0.1417577565,0.3292892277,0.408164531,0.3600576818,-0.4821265936,-0.6383057833,0.0391044766,0.0401061065,0.3005616069,0.0823301002,-0.1973975748,0.1902044117,-0.2747591436,-0.0287708528,-0.1576127112,0.4213906229,0.3624390066,0.2754831612,-0.4533807039,-0.4862160385,0.1569014639,-0.0935743228,-0.3375833333,0.035652183,0.1259561479,-0.4417129159,0.4667490423,-0.0186617859,0.8538818359,-0.4546049833,0.0962253436,0.1599946618,0.2531588972,0.7502367496,-0.0199419931,-0.0957798287,-0.2133440971,0.0240359288,-0.1087242588,0.1119210124,0.1089091524,0.253631562,-0.222307831,0.4156003296,0.0304187555,0.4982374907,0.1266560256,0.0278673694,0.4766612053,-0.3088310659,-0.3115204573,0.0545618087,-0.0697770715,0.0356489271,-0.0045600785,-0.0352154225,-0.1696330607,-0.2989766002,-0.2679096162,0.0762702227,-0.1809964329,0.3267801404,-0.3523544669,-0.3634991944,0.4128607213,0.4410256445,0.0437680222,0.1301425397,-0.1129651517,0.0003238506,0.0872352421,0.0596228056,-0.3403808475,-0.0629584566,0.3871629238,0.0844535306,-0.1124230325,0.4013395011,-0.0449005365,-0.1486762315,0.0666267574,0.174853459,0.5714659095,-0.5062837005,-0.3370099664,-0.2767016888,0.0419976003,-0.3917419612,0.0281824637,-0.1948964596,-0.1239202023,0.0166914072,0.0216483809,-0.4055575728,0.1933333427,0.6294291615,-0.0493522808,0.0708502084,0.7852793932,0.0728306323,-0.1149651036,-0.1665186286,0.0860115662,0.5840911269,-0.14866817,0.233825773,0.0933881551,-0.0641303733,0.0181239638,0.4313964546,0.5497334599,-0.0057319538,0.1652353555,-0.4803129435,-0.4752922058,0.0116617782,-0.0688422099,0.1330046952,0.4147024155,0.158065632,-0.1791446656,0.5363548994,-0.1994547248,0.0471610166,-0.179473117,0.2883890867,0.2534252107,0.1340298951,-0.2279547751,-0.0515116639,0.1157054976,-0.125923261,0.0553603284,-0.1513466686,-0.180310443,0.162634477,0.0242551975,-0.0538024716,-0.1285189986,-0.3979252875,0.093452543,-0.4114852846,0.2004027963,0.2641579807,0.0309233088,0.0558289401,0.0943148509,-0.2631490231,-0.1857144237,0.038532149,-0.1030581146,-0.1144090891,0.0192986745,0.1908440739,0.0535025112,-0.034686137,-0.025311904,-0.0308934189,-0.0405474678,-0.0744228438,0.3435593843,-0.112395674,0.2276649028,0.2359961867,0.1910891086,0.0069904225,-0.1660908014,0.0699272603,0.3582183421,0.043475911,-0.4803177118,0.0895875469,0.1368270516,-0.0299050808,0.0883752331,0.1709184498,-0.0528236739,0.0629411787,-0.1101972163,-0.2367053479,0.3873440623,-0.0596081205,0.1175315008,0.5157994032,-0.1992423832,0.2741678655,0.4069995284,0.2518590093,0.1012815014,0.5119723678,-0.1945046633,0.16651842,-0.2078119218,0.3426595926,0.0851847082,-0.3729570508,0.1948928088,0.2040794939,-0.1129731014,0.1360031515,0.0149118369,0.6207628846,-0.4058395922,-0.2460162491,0.0260691065,0.362469703,-0.3382800817,-0.2492062151,-0.0257639401,-0.2772208452,0.1468814015,-0.0313286819,-0.3017627895,-0.0870868564,0.1407350451,0.0974299833,-0.0951563418,-0.388295114,-0.3287885785,0.109848842,-0.0754377097,-0.1581839174,0.428452611,-0.149591133,0.0571290404,0.0845344439,0.1054724678,0.4288723171,0.6453410387,0.0887052193,-0.1818567961,0.1211570278,-0.1629321724,-0.0693932772,0.0466387048,0.0910195038,0.0710614994,0.2339627594,0.0203048363,-0.0890840515,0.2069789171,0.1943567842,0.2529907823,0.1798135489,0.0852779597,-0.1781769693,-0.1139146835,-0.1873798072,0.0658743605,-0.3288325369,0.2687665224,0.0924542397,-0.1591729522,0.2595170438,0.0588992722,0.0379386842,0.2263917625,0.1651692241,0.2591685355,-0.092806749,-0.6077032685,0.1426296532,-0.8178735971,0.3019883335,-0.2084612399,-0.0313831493,-0.2205667347,0.0342906564,0.2456472665,-0.127258271,0.1502779424,0.023984829,0.0380285643,0.3548089564,-0.0126671987,-0.0467083417,-0.2404095232,0.1079238057,-0.1733584404,-0.3466698527,0.2856391668,0.1229375228,-0.1231837273,-0.195590511,0.178684473,0.0083468761,0.0940715745,0.4659883976,0.1278936714,0.7292595506,-0.0321749821,0.2721481025,0.0740848854,-0.1821622849,-0.1105566993,0.0749446675,0.202351585,0.2000992149,-0.195001781,-0.1445555091,-0.2320061326,-0.0755844414,-0.0144242607,-0.2899771929,-0.6900325418,0.1493444592,-0.0430814438,-0.0434927978,-0.018656997,0.2625669539,-0.0919618383,0.1879252493,-0.0971285477,-0.4130038917,0.3670980036,-0.554874301,-0.4784332514,-0.024199456,-0.1244087294,0.1122574136,-0.0611881986,-0.6009425521,-0.0534130372,0.2493426055,-0.0713379085,-0.4005272985,-0.0427324288,-0.0534496233,0.0576560237,-0.1844071895,0.4424070716,0.1898754984,-0.1703061908,-0.0935344249,-0.4564055502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1324","title":"\u2753 Sharing ElasticSearch indexed dataset ","comments":"Hello @pietrolesci , I am not sure to understand what you are trying to do here.\r\n\r\nIf you're looking for ways to save a dataset on disk, you can you the `save_to_disk` method:\r\n```python\r\n>>> import datasets\r\n>>> loaded_dataset = datasets.load(\"dataset_name\")\r\n>>> loaded_dataset.save_to_disk(\"\/path\/on\/your\/disk\")\r\n```\r\n\r\nThe saved dataset can later be retrieved using:\r\n```python\r\n>>> loaded_dataset = datasets.Dataset.load_from_disk(\"\/path\/on\/your\/disk\")\r\n```\r\n\r\nAlso, I'd recommend posting your question directly in the issue section of the [elasticsearch repo](https:\/\/github.com\/elastic\/elasticsearch)","body":"Hi there,\r\n\r\nFirst of all, thank you very much for this amazing library. Datasets have become my preferred data structure for basically everything I am currently doing.\r\n\r\n**Question:** I'm working with a dataset and I have an elasticsearch container running at localhost:9200. I added an elasticsearch index and I was wondering\r\n\r\n- how can I know where it has been saved?\r\n\r\n- how can I share the indexed dataset with others?\r\n\r\nI tried to dig into the docs, but could not find anything about that.\r\n\r\nThank you very much for your help.\r\n\r\nBest,\r\nPietro\r\n\r\nEdit: apologies for the wrong label","comment_length":73,"text":"\u2753 Sharing ElasticSearch indexed dataset \nHi there,\r\n\r\nFirst of all, thank you very much for this amazing library. Datasets have become my preferred data structure for basically everything I am currently doing.\r\n\r\n**Question:** I'm working with a dataset and I have an elasticsearch container running at localhost:9200. I added an elasticsearch index and I was wondering\r\n\r\n- how can I know where it has been saved?\r\n\r\n- how can I share the indexed dataset with others?\r\n\r\nI tried to dig into the docs, but could not find anything about that.\r\n\r\nThank you very much for your help.\r\n\r\nBest,\r\nPietro\r\n\r\nEdit: apologies for the wrong label\nHello @pietrolesci , I am not sure to understand what you are trying to do here.\r\n\r\nIf you're looking for ways to save a dataset on disk, you can you the `save_to_disk` method:\r\n```python\r\n>>> import datasets\r\n>>> loaded_dataset = datasets.load(\"dataset_name\")\r\n>>> loaded_dataset.save_to_disk(\"\/path\/on\/your\/disk\")\r\n```\r\n\r\nThe saved dataset can later be retrieved using:\r\n```python\r\n>>> loaded_dataset = datasets.Dataset.load_from_disk(\"\/path\/on\/your\/disk\")\r\n```\r\n\r\nAlso, I'd recommend posting your question directly in the issue section of the [elasticsearch repo](https:\/\/github.com\/elastic\/elasticsearch)","embeddings":[-0.1429361105,-0.0638304427,-0.1649869233,0.115781866,-0.2390129417,0.181620121,0.2375487387,0.0645310879,0.1375180483,0.2632859945,-0.2542998195,0.1102019325,-0.0403488576,-0.1499138623,0.1162049919,0.2783807516,0.1407323331,-0.0522752032,0.1609809846,-0.1157200262,-0.0960236788,-0.119999066,0.0747434869,-0.0017549342,-0.0378025025,-0.1324211657,-0.4750579894,-0.1286243647,0.0660043433,-0.2219629884,0.5905970931,0.228190884,0.1744388193,0.3492886424,-0.00011476,0.0371787362,0.3044726849,-0.1059916615,-0.0600125231,-0.3337877989,-0.581566155,-0.215335533,0.3577898443,-0.5136823654,-0.0529217497,-0.2876566947,0.215880692,-0.6586778164,-0.0800799802,-0.0880682096,0.1715499014,0.1409371942,-0.1088337451,-0.0197396595,0.1892026067,0.1232945025,-0.0214467831,-0.0853259489,0.0034694511,0.0862663165,0.3268643916,0.1831186116,-0.1671927869,0.0605216324,0.1701910049,0.1723883599,-0.5219464302,-0.2311159968,-0.071456939,-0.0918339714,0.8283691406,-0.2656334937,-0.2670591772,0.0779282004,-0.0186363161,-0.1364419609,0.0233307462,0.1480533183,0.1877560019,0.1393417418,-0.2092095912,-0.4387285113,-0.4489777386,0.1656543314,-0.0322407037,0.5990275145,-0.1216457784,0.1211185828,-0.1504866928,0.0379112922,0.3560683131,-0.0478119478,0.1332057416,0.1248650327,0.2179226428,-0.1643845588,-0.152307421,-0.3114429414,-0.0014166407,0.1400372088,-0.1719859093,-0.2091900259,0.2432661802,0.3475384414,0.3548472822,0.2206200361,0.1119060665,0.3693484664,0.2211711854,-0.3995965719,-0.1336420625,-0.3954393864,-0.002178137,0.2996359468,0.4795609117,-0.384178102,0.1183941662,-0.1556719542,-0.0499291383,-0.1355164498,-0.2635773122,-0.2346781343,0.1576255262,0.2230813056,-0.0426803678,-0.4526020885,0.0704217851,0.053011056,0.0606575049,0.0951057225,-0.1951247007,-0.1827514917,0.1014507264,0.3022234738,0.1282158643,-0.431622237,0.1091167778,-0.2733577192,0.2553037405,-0.0140679283,0.429731071,0.1854222566,0.2484382838,0.3468284905,0.3137482703,0.1941066235,0.0212832112,-0.1582435519,-0.4444412589,-0.0081882961,-0.2976227403,-0.4269990623,-0.2789047062,0.0796810836,-0.14677082,0.1225128844,-0.2636526823,0.1629920751,0.0278550591,0.1676827371,0.0675702393,0.3594226241,-0.3890078664,-0.3475756347,0.4555885196,-0.1132136956,-0.1483667642,-0.1401285827,-0.4394765794,0.3204064667,-0.1318344772,-0.0364204086,-0.2487839162,0.1561017185,0.0096513871,0.3899398744,0.3576794565,-0.2793350816,-0.4653582871,0.0308838598,-0.0632729381,-0.4568527043,-0.090642795,0.2196332067,0.5229578614,-0.1691964865,-0.0543145537,0.5200255513,0.130307585,-0.1105512753,-0.2245105654,-0.105370082,-0.100602217,-0.0674336106,-0.2246749103,-0.0784225389,0.4058630764,0.083795689,-0.1331441849,0.3465709686,0.2486115694,0.2592419088,0.6578143239,0.3961037695,0.0544923209,-0.000192329,-0.1552846283,-0.0264302511,0.1335355639,-0.325779438,-0.0197524186,-0.4639935195,-0.1371489465,0.2006537616,-0.0217944086,0.0898777023,0.1105008945,0.0291883945,0.0281217005,0.3715507388,-0.2713485658,-0.1345423907,0.0544304997,-0.1252729148,-0.1900705695,0.4088031948,-0.0426318161,0.055150643,-0.1099947318,-0.0405121073,0.2740743756,-0.0997938141,-0.1949898452,0.4314921796,-0.4871744215,0.4711259902,0.3968756199,0.4983282089,0.0694006085,-0.6583535075,0.5281495452,-0.0021941918,-0.1115332916,0.1143557876,-0.3864212334,0.2820076048,-0.3749865294,0.1623958796,-0.1657324582,-0.0703583956,-0.1311714053,0.009526629,-0.3378591239,0.0164274033,-0.1932197362,-0.2723450959,-0.1785287708,0.1191431507,-0.1280243248,0.3031448424,0.0438921526,-0.0066568959,0.213763088,0.1184215173,0.044545982,-0.2419227511,-0.0614567101,-0.0233433712,0.3800972402,0.1158345491,0.3460180461,-0.1429543495,0.2170419097,0.1167357787,0.2452485263,0.1727616191,0.0757868513,-0.3122872114,0.1125931963,0.0014003376,0.3942690194,-0.1093649566,-0.2946109772,-0.0856119469,0.068510212,0.4260074794,0.0198859684,-0.0122362422,-0.0640035197,-0.1772958189,-0.2049411535,-0.2950291038,-0.0858226418,0.3247539401,0.1016401574,-0.1656903923,-0.4062299728,0.3377700746,0.0262661129,-0.1283356398,-0.102193132,-0.0353158005,0.0038677722,-0.0308065824,0.1300443113,0.3277303576,0.4591417909,0.3788124025,0.3379550874,-0.0890286043,0.0154948421,-0.1455993354,0.1245422736,0.083466284,-0.0886008218,0.1689879596,-0.7019357681,-0.2014822364,-0.0496284701,0.0304674767,0.0253383685,-0.1356609464,0.1490632892,0.1728926003,0.3836007714,-0.359167397,-0.4527559876,-0.0289978329,0.1434897929,-0.138441056,0.1597138494,-0.2611454427,-0.2167722583,0.0661258474,0.2811373174,-0.2025665492,0.0066708107,-0.6452115774,0.1128072739,-0.4477564692,-0.0002353656,0.2146393508,0.0456143841,0.0255773813,0.0032441698,-0.3282503486,0.3077257872,0.1301650107,-0.0031391599,0.0891865417,0.2140133083,0.3354142904,0.0677713454,0.0188867711,0.0027593495,-0.2675281465,-0.0721446946,0.428008765,0.1089506745,0.2220333368,-0.0767017007,0.0200661104,0.713788569,0.1559154987,-0.0385167301,0.1857172102,0.1338333189,0.2979231477,-0.0266185235,-0.3098938167,-0.2448050678,0.2814114094,-0.3961248994,0.0849685594,0.1135468036,-0.1524419188,-0.1195737422,-0.56090343,-0.3739663363,0.0048772967,0.011040966,0.2319194376,0.3605761528,-0.2060416192,-0.4837602079,-0.0038880042,-0.3766238093,-0.3294091821,0.3415917158,0.1157604828,0.1399447322,-0.1882131994,-0.2318633795,-0.1870823801,0.1716303825,-0.1504940689,-0.0476321056,-0.0803429857,0.1951390058,0.4556745291,0.2745219171,0.1774972528,-0.1910548657,-0.3307288587,-0.2974391282,-0.0213380065,-0.2842198908,0.0587756149,0.0432242528,0.3016094267,-0.3704270422,0.3083726764,0.0989039838,-0.0571459271,-0.0637233928,-0.0265495274,-0.4934938252,-0.0928100944,-0.264077872,-0.356697768,0.2192667127,0.1003524363,-0.0361859351,0.0916430876,0.0977054462,0.1210551634,-0.1716818064,0.0780136287,0.0592731424,-0.0973030552,0.2781715989,0.1711909026,0.1679454446,0.7190199494,0.149673298,0.2815488577,0.312074393,-0.0907720774,0.4290629327,-0.0787078962,0.0842247531,0.4779815078,-0.065704748,0.0993307903,-0.1357584298,-0.0922005549,-0.1045402288,-0.2849583626,-0.2307431102,0.0904515535,0.0187982377,-0.2011897415,-0.4521866739,0.4708428085,-0.1084275171,-0.1266343892,0.4583641887,0.3381693661,-0.1887979656,0.3753177226,0.1698606014,0.9062938094,-0.1186842918,0.0228271689,-0.2764090896,0.1808335781,0.3432733715,-0.4557620585,0.0925634652,-0.0950530842,-0.4550405145,-0.0503502972,-0.1456655264,-0.0431459658,-0.1382171363,-0.3969094753,-0.2094334513,-0.2419341356,-0.1358187646,-0.2971745729,0.2549379468,0.0892326534,-0.2151841223,0.0032019261,0.0994381756,-0.0560693257,0.3162005842,0.1561443508,-0.0820658058,-0.0567035079,0.1120547876,-0.2932804227,-0.288851738,0.3653303683,0.3622822762,-0.1373373568,-0.1008210555,0.0408040844,0.1724918634,0.135817796,0.0985700488,-0.1121763587,-0.2138378322,-0.2195040584,-0.0197745897,0.1174197569,-0.1770546734,0.1229785532,0.1347723752,-0.3482441008,0.11546655,0.3369024098,-0.1293959171,-0.2612890899,0.1749848723,0.0769294575,-0.2425855696,0.0109512592,-0.0012125166,-0.1433361918,0.0580404848,0.0989300534,0.2997498214,-0.1633694023,0.338821739,0.2223599404,-0.1853231788,0.0945519879,0.5310108066,0.2706599832,-0.1918641627,0.2153995484,-0.2635295987,-0.1084835753,-0.0033488169,0.6778512597,0.3922754824,0.4330102503,-0.2178932428,-0.3760693669,0.6171231866,-0.0813298002,-0.1064816266,0.1463917494,0.0585174821,-0.5026091337,-0.1749236286,0.0992423818,0.2952950895,0.2668551505,0.355117172,-0.2838893831,-0.6877222657,0.0752503276,-0.0786379874,-0.2646252811,-0.0151813412,-0.2381038666,-0.1330007464,0.3729129136,0.3742672503,0.2937640846,-0.3062369227,0.0346297361,-0.1212885529,0.063552171,-0.0984661728,-0.0472080596,0.1676334739,0.0018581169,-0.0642295927,-0.2943470478,-0.6276305914,0.0924844667,-0.0185315255,-0.2699891627,0.2725459635,0.2530996501,-0.0331856683,0.1095570996,-0.1364646554,0.0254691355,0.1434615552,0.3007857203,0.2595262825,-0.1352567673,0.2295941561,0.1930019557,-0.1594734788,-0.2823141515,0.0256345607,0.0973566398,0.1198134795,0.0809880942,0.1485732049,0.1498058438,0.2380693853,0.3956745267,-0.0017428051,-0.199137643,0.0691332221,0.0027165075,0.1522466987,0.1693266034,0.0266512223,0.1461408883,0.2341642231,-0.3636999428,0.2730685174,0.1167218015,0.1973476559,-0.3014603257,0.028656682,0.5244446397,0.0212020781,0.0693769529,-0.2014276236,-0.0417225435,0.0697335005,0.156946525,0.5201879144,0.4039348364,0.2948313355,0.1010181308,0.4887929559,0.3322434723,-0.2144626081,0.4194937348,-0.0409087576,0.1117508486,-0.1217951104,0.3228593171,0.1213344485,-0.0352262184,0.350155592,-0.1719780117,-0.1148760244,0.0394033454,0.0711812451,0.0735252276,-0.0129791712,0.1209477037,-0.0950400978,0.0616097003,-0.2645799518,0.1729436219,-0.3572736681,-0.2551344335,0.0647488981,0.1035549194,0.0677632019,0.2595201433,0.198686108,0.0589179024,-0.0807159618,-0.1421077847,0.0400519893,0.2084175944,-0.1009969041,0.0228561666,-0.0579562709,-0.1007582098,0.0253019314,0.626326561,0.0562480725,-0.3144046962,0.0496787913,-0.2114910036,-0.2944129705,-0.4331623018,-0.1437261552,0.1286887228,0.1327508241,0.2015917152,0.2349300385,0.272030592,-0.0070940596,-0.3087901473,-0.0508810729,-0.1066234335,0.0405167453,-0.0261629112,-0.3537275493,-0.1799957305,0.5780165792,0.1312900484,-0.0121270958,0.0221168306,0.0324428976,-0.0481945053,0.1600104123,0.3391343057,-0.1215308234,-0.0368460119,-0.2133181989,-0.2488237768,0.1263718605,-0.2509039342,-0.0169714093,0.2048119307,0.4461942315,-0.0847688839,0.0112572117,-0.0300212596,-0.0601800084,-0.0930337012,0.1181418002,0.2180832773,-0.3541246355,-0.1416830868,0.2230700105,-0.0088978643,0.1475783288,0.1561979502,-0.232343033,0.0083543099,-0.4974645972,0.0556207635,0.0654125512,0.2187635154,0.1594608128,-0.3096054792,0.1000325158,0.0664421022,0.0703707114,0.1690435857,-0.0246858392,-0.0373543911,0.2794217765,-0.0495902747,-0.1328596622,-0.5643982887,-0.3422437906,-0.3078683019,-0.1439306289,-0.2064444125,-0.2464855909,0.1733019352,-0.431735009,-0.2348001897,0.097313337,0.4112352431,0.0503337048,-0.0008298988,0.053816665,-0.1150373146,-0.067503348,0.1778490543,-0.3819772899,0.0685225129,-0.0919138417,0.1249930784,0.2355664521,0.164540723,-0.1812952161,0.0762334019,0.6318387389,-0.2289964706,0.1625603586,0.0231881011,0.3164775372,0.228188917,-0.1420792192,0.0780965388,0.0015668669,0.1062558144,-0.0677429512,-0.0993951038]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1324","title":"\u2753 Sharing ElasticSearch indexed dataset ","comments":"Hi @SBrandeis,\n\nThanks a lot for picking up my request. \n\nMaybe I can clarify my use-case with a bit of context. Say I have the IMDb dataset. I create an ES index on it. Now I can save and reload the dataset from disk normally. Once I reload the dataset, it is easy to retrieve the ES index on my machine. I was wondering: is there a way I can share the (now) indexed version of the IMDb dataset with my colleagues without requiring them to re-index it?\n\nThanks a lot in advance for your consideration.\n\nBest,\n\nPietro","body":"Hi there,\r\n\r\nFirst of all, thank you very much for this amazing library. Datasets have become my preferred data structure for basically everything I am currently doing.\r\n\r\n**Question:** I'm working with a dataset and I have an elasticsearch container running at localhost:9200. I added an elasticsearch index and I was wondering\r\n\r\n- how can I know where it has been saved?\r\n\r\n- how can I share the indexed dataset with others?\r\n\r\nI tried to dig into the docs, but could not find anything about that.\r\n\r\nThank you very much for your help.\r\n\r\nBest,\r\nPietro\r\n\r\nEdit: apologies for the wrong label","comment_length":98,"text":"\u2753 Sharing ElasticSearch indexed dataset \nHi there,\r\n\r\nFirst of all, thank you very much for this amazing library. Datasets have become my preferred data structure for basically everything I am currently doing.\r\n\r\n**Question:** I'm working with a dataset and I have an elasticsearch container running at localhost:9200. I added an elasticsearch index and I was wondering\r\n\r\n- how can I know where it has been saved?\r\n\r\n- how can I share the indexed dataset with others?\r\n\r\nI tried to dig into the docs, but could not find anything about that.\r\n\r\nThank you very much for your help.\r\n\r\nBest,\r\nPietro\r\n\r\nEdit: apologies for the wrong label\nHi @SBrandeis,\n\nThanks a lot for picking up my request. \n\nMaybe I can clarify my use-case with a bit of context. Say I have the IMDb dataset. I create an ES index on it. Now I can save and reload the dataset from disk normally. Once I reload the dataset, it is easy to retrieve the ES index on my machine. I was wondering: is there a way I can share the (now) indexed version of the IMDb dataset with my colleagues without requiring them to re-index it?\n\nThanks a lot in advance for your consideration.\n\nBest,\n\nPietro","embeddings":[-0.1988601536,0.1247475296,-0.119677715,0.0339192636,-0.3991583884,0.1631872207,0.2689709365,0.0641190484,0.0988282338,0.2021004707,-0.228173539,0.1584118009,0.0233792495,-0.0414418392,-0.0148405479,0.3394413888,0.0893101469,0.0631404296,0.11034666,-0.1242943257,0.0559722446,-0.2349125296,0.0433592908,-0.0645673648,-0.0652778074,-0.061832089,-0.4033026993,-0.1812634468,0.0436657816,-0.2119496018,0.5912082791,0.301091224,0.2538352609,0.2416616231,-0.0001154477,0.059577629,0.276428014,-0.0930100605,-0.1243460774,-0.3499150574,-0.6613681316,-0.1312082708,0.2852573693,-0.4554820657,-0.0966447666,-0.3421029449,0.2402936816,-0.7295071483,-0.0450430326,-0.2401281297,0.1476607323,0.0412187912,-0.1752968878,-0.0152987074,0.1940662861,0.113657333,0.0129568912,-0.1093859896,0.1031234711,0.1652024686,0.205419451,0.1775828004,-0.0940437019,0.0174312349,0.1710152775,0.0823680609,-0.524279058,-0.1088333279,-0.0281775277,0.0563650094,0.9450519681,-0.1793226451,-0.419952482,0.0407501198,0.0339218043,-0.0398226306,-0.0109042032,-0.0083920714,0.2608441114,0.1329615861,-0.2057773024,-0.4527464509,-0.3319205642,0.0893896967,0.0239438452,0.5131623149,-0.054500781,0.0551777445,-0.166105777,-0.0404138342,0.4516308308,-0.0163610075,0.0697255582,0.1077305153,0.2860757709,-0.1387107968,-0.1411644071,-0.290079236,-0.0777222589,0.1675433367,-0.1779172421,-0.1583892852,0.229165405,0.2637395859,0.4209385812,0.2178203613,0.1772520393,0.3229680657,0.1465878785,-0.352927804,-0.1640387475,-0.3661541939,-0.0132499412,0.3466744125,0.4428035915,-0.3139209151,0.0692416802,-0.1900915205,-0.074898243,-0.2175788134,-0.2848870456,-0.1972461045,0.1211084202,0.0965382755,-0.0309478715,-0.372071743,0.0214540251,-0.0606534444,-0.0201938078,-0.0292197559,-0.1957124025,-0.2964936495,0.0171738043,0.3570494652,0.1961147338,-0.4609553516,0.0394107066,-0.244851768,0.223120898,0.0160479229,0.3973346949,0.2916308343,0.2900517583,0.1884032786,0.257761538,0.2254975289,-0.0337665081,-0.2146922797,-0.4397146106,0.0227637663,-0.2895736992,-0.4273771048,-0.2026579678,0.0500161946,-0.1111699566,0.1029620916,-0.1753638536,0.2210798413,-0.0205312632,0.0416545309,0.0754981786,0.3739998341,-0.3952652216,-0.2210346758,0.414886713,0.0035593063,-0.2440735847,-0.2064417303,-0.5053339601,0.3462714255,-0.2634888887,0.0357699208,-0.280257076,0.0588382967,0.0382040925,0.3766159713,0.3078480065,-0.2098412663,-0.6120325327,0.0018124729,-0.0938809216,-0.3875820637,-0.004830536,0.2987863123,0.5886089206,-0.1760359704,-0.0411486402,0.4830717742,0.0590430647,-0.0885216147,-0.1775696427,-0.1968394965,-0.0563885197,-0.1380819231,-0.0941395834,-0.0182277765,0.3711380363,0.1541195214,-0.1286540627,0.3143368363,0.2648582757,0.1926955879,0.486101687,0.427141726,0.0616721921,0.0193928163,-0.258890003,0.0219211411,0.0880433843,-0.3332216144,0.1101880223,-0.5423186421,-0.132173717,0.1099899933,0.0572945066,0.0492972583,0.0710489675,0.0199016817,-0.0125617124,0.313713342,-0.2198410183,-0.1813590527,0.1798437834,-0.1857519895,-0.3203570545,0.3934904933,-0.0204347074,0.0663026199,-0.1220833287,0.0164788682,0.191543445,-0.1148840338,-0.1754519641,0.3342700303,-0.519467473,0.4946865737,0.3578430116,0.500210762,0.0858272761,-0.6622859836,0.5527696609,-0.0361749306,-0.1229287684,0.0886336714,-0.3847316802,0.2940207422,-0.4438501298,0.1080007255,-0.1244286001,-0.1022992879,-0.1508060396,-0.0758102834,-0.3517902195,-0.0726294816,-0.2458980531,-0.3034408092,-0.20773004,0.1918961555,-0.0563424863,0.3178856075,-0.0210104138,0.1341428459,0.1603891701,0.0969604477,-0.038572602,-0.2863000333,-0.0457849167,-0.1007416099,0.2109859884,0.1735736579,0.3986305296,-0.1023572609,0.3224239051,0.0732471496,0.1670353711,0.0395460837,0.1103518903,-0.2544102073,0.107952252,0.04026898,0.3734403551,-0.0160953645,-0.1876020432,-0.128962934,0.0510351732,0.3918848336,-0.0940178856,-0.0854631141,-0.0150619606,-0.1260554045,-0.1579370052,-0.2407260388,-0.0270373318,0.4399486482,0.0583895408,-0.0824674368,-0.4287155569,0.3352515697,-0.0003864482,-0.1406266838,-0.0541082919,-0.1411371976,0.0395008922,-0.0201627724,0.1621989608,0.3113367558,0.4181275666,0.3823670447,0.3871317804,-0.1832802743,-0.0685769245,-0.1460180283,0.1803927422,0.1241778135,-0.136332795,0.054550156,-0.6782448292,-0.2409548163,-0.1003120914,-0.0597910695,0.0098857796,-0.2054486573,0.1269374043,0.184761256,0.4511450231,-0.4822726846,-0.2808845639,0.0073809791,0.2037153989,-0.2172985673,0.0744234696,-0.1621911675,-0.1661675572,-0.0636928603,0.3251665831,-0.248437956,-0.0824173763,-0.6632475257,0.0341141559,-0.4512145519,0.0057522156,0.0873015672,-0.0184708405,-0.0174825881,0.1385416985,-0.3052280545,0.4069462419,0.1929593682,0.0120888986,0.0669228211,0.089193657,0.3754009008,0.0703555495,-0.0106692379,0.032443732,-0.2659999728,-0.0522957966,0.4143597186,0.1227937266,0.1801627576,-0.0679911599,0.0717385486,0.7623034716,0.1164719164,-0.1460179687,0.1448879689,0.1330832243,0.3890991807,-0.0884660035,-0.233131513,-0.2329468131,0.3048665226,-0.3906289637,0.0088465977,0.079992637,-0.2279769033,-0.1015568376,-0.5722395778,-0.2830640376,-0.0507570691,0.0366561152,0.1230342761,0.3523516953,0.0107647311,-0.422739327,-0.0648380965,-0.4604937732,-0.3073159456,0.229919821,0.2506479919,0.1428655982,-0.1529070288,-0.1323896199,-0.0686790571,0.1539710015,-0.041075293,0.0654674023,-0.0034509203,0.1661800444,0.4228128791,0.3001916409,0.1424068958,-0.2715088725,-0.2307691723,-0.2508732975,0.0491177849,-0.2428726405,0.0660156384,0.0827064812,0.3585073352,-0.3965243399,0.2971867323,0.1547397971,0.0060447874,-0.0213040393,-0.0852674246,-0.4650022686,-0.0787568614,-0.2405430824,-0.2119168192,0.2239191085,0.1603489369,-0.0894413143,-0.0315386541,0.0900234208,0.1248378083,-0.2611791193,0.133764267,-0.0053243418,-0.0403085873,0.2585460842,0.3238072395,0.2644466758,0.7711284757,0.1207401752,0.3748743534,0.353918165,-0.0917024985,0.4947232902,-0.1578239202,0.0938214436,0.3423300683,0.0029231685,0.1041040048,-0.1784408689,-0.0110084396,-0.1501689553,-0.3299858272,-0.2239181995,0.0409728102,-0.0595214516,-0.2411114573,-0.4408234954,0.4113924801,-0.0485622808,-0.1598932147,0.4277384579,0.3819778264,-0.1581193209,0.3384886384,0.2151459605,0.7895462513,-0.153599456,0.0697657242,-0.3745653331,0.2819425762,0.4274340272,-0.3671251237,0.15229927,-0.1027528942,-0.5429016352,-0.0907826796,-0.0935062841,-0.0923119783,-0.0650366545,-0.268784076,-0.1713573039,-0.1788930893,-0.1259432584,-0.2673239112,0.3235788643,0.1354744285,-0.1565599442,0.1838376224,0.0625622049,0.0260708947,0.2199367136,0.1365500093,-0.0409249738,-0.0357426479,0.1524196416,-0.3662888408,-0.2778866291,0.3187919557,0.4105169177,-0.1524470747,-0.1143142506,0.140115872,0.0055305162,0.1207776517,0.1060691178,-0.0635837168,-0.1557033807,-0.1551452428,-0.048988983,0.0908610001,-0.1937014461,0.2439700812,0.2436434478,-0.4006450772,0.0392426625,0.3353351355,-0.1123581827,-0.2934342325,0.105327934,0.0497858003,-0.2370606214,0.1523443609,0.0401251987,-0.0984494463,0.1003220826,0.0716194585,0.1845344305,-0.0219404045,0.2324468344,0.35365206,-0.2250643969,0.1306175739,0.7560180426,0.301176697,-0.1902536899,0.2914906442,-0.2839081883,-0.0618213564,0.03446104,0.5927833319,0.4039462507,0.378546536,-0.2801948488,-0.4006794691,0.6040418148,0.0050997688,-0.140352577,0.1100044996,0.1292929351,-0.4848043025,-0.0552801527,0.1767627746,0.3700094223,0.1789283454,0.3861180842,-0.3293561637,-0.6366773844,0.0367073528,-0.0532693788,-0.2445711046,0.0393371321,-0.1165656298,-0.0479989089,0.307490766,0.3928630948,0.2351212502,-0.361112237,-0.0121808052,-0.115854986,0.0772446245,-0.0817568004,-0.0255222172,0.1632356197,0.0478632562,-0.0336269625,-0.3351222873,-0.6004692912,0.0131640434,-0.0295976028,-0.2250585407,0.3415967226,0.2611825466,-0.0255071837,0.0792087913,-0.1520790458,0.0596672222,0.2393163294,0.3122388124,0.3275751472,-0.1785263121,0.2807263732,0.1692123115,-0.0819033682,-0.2514819205,-0.0153171699,0.1791565716,0.1710134149,0.0182963312,0.0997864157,0.1434958726,0.3075676262,0.3659248352,0.0741650239,-0.1473126113,-0.0506463386,0.0160377193,0.1272450686,0.1177117452,0.0958656073,0.1720528454,0.214163065,-0.3437947929,0.2365575284,0.2027544975,0.1931516677,-0.2582730055,-0.0245818906,0.3759483099,0.0024668633,0.1191950068,-0.286424458,-0.0343849137,0.0204687826,0.1288377047,0.5339488983,0.4390485585,0.287253499,0.1947346032,0.4252044857,0.326595962,-0.0952542946,0.2823249996,-0.129513815,0.0875483975,-0.0259349067,0.4635297358,0.1435147077,0.0034317132,0.2549614906,0.0141069861,-0.1103914902,0.0870040879,0.0234131869,0.0990672782,-0.1082645282,0.1965470761,-0.1438850909,-0.0461317264,-0.2504726648,0.1138726473,-0.2143408209,-0.2665426731,0.0252654161,0.1398301572,-0.0576944686,0.1907279193,0.1885733306,0.0953207761,-0.0905579329,-0.1448542774,0.0238224212,0.1769090295,0.0349898152,0.105783008,-0.1384929419,-0.2238197923,-0.0477365069,0.6414462924,0.0690552071,-0.2220045924,0.0221631005,-0.2016858011,-0.221901685,-0.5017582774,-0.135553658,0.0744630694,0.1801434457,0.247764349,0.2241043895,0.2597483397,-0.0983477831,-0.3767017722,0.0423207171,-0.0012834473,0.0804956779,-0.0712692142,-0.2423620373,-0.1729291677,0.5879146457,0.0399129875,-0.0257998221,-0.0133588016,0.0198875181,-0.0317281224,0.2002620846,0.3906554878,-0.1158633903,0.0656955093,-0.1754956841,-0.3162516057,0.2365531623,-0.2546003759,0.1089096814,0.1835517585,0.4819937348,-0.0048102797,0.0258817505,-0.0044518602,-0.0784697756,-0.0661040395,0.1084615141,0.2437612563,-0.4186978638,-0.1691356003,0.0946247503,-0.0129651465,0.1304560006,0.0540064052,-0.1932363063,-0.0159187764,-0.479165554,-0.0254825391,0.0818016827,0.2168731987,0.2079459131,-0.2851437628,0.0684062093,0.0941850394,0.0374050438,0.0386183821,-0.0879768133,-0.1031288952,0.3056048751,0.0441691615,-0.1872268766,-0.6146065593,-0.2730353177,-0.3219205439,-0.2103547305,-0.1793844551,-0.1831953973,-0.0646429658,-0.3975069821,-0.276853919,0.1195989177,0.2697784007,0.1655711532,0.0162621085,0.1240483299,-0.0733270198,-0.048156444,0.1637559533,-0.3931723535,0.1680311412,-0.1358804107,0.1666194499,0.2945106626,0.0460224003,-0.0858491361,0.2235157043,0.6327011585,-0.2470054477,0.2508889437,-0.1537608355,0.2452571541,0.262622565,-0.2423036546,0.1168691814,0.0135355629,0.1160094365,-0.0772315487,-0.0991740078]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1324","title":"\u2753 Sharing ElasticSearch indexed dataset ","comments":"Thanks for the clarification.\r\n\r\nI am not familiar with ElasticSearch, but if I understand well you're trying to migrate your data along with the ES index.\r\nMy advice would be to check out ES documentation, for instance, this might help you: https:\/\/www.elastic.co\/guide\/en\/cloud\/current\/ec-migrate-data.html\r\n\r\nLet me know if it helps","body":"Hi there,\r\n\r\nFirst of all, thank you very much for this amazing library. Datasets have become my preferred data structure for basically everything I am currently doing.\r\n\r\n**Question:** I'm working with a dataset and I have an elasticsearch container running at localhost:9200. I added an elasticsearch index and I was wondering\r\n\r\n- how can I know where it has been saved?\r\n\r\n- how can I share the indexed dataset with others?\r\n\r\nI tried to dig into the docs, but could not find anything about that.\r\n\r\nThank you very much for your help.\r\n\r\nBest,\r\nPietro\r\n\r\nEdit: apologies for the wrong label","comment_length":48,"text":"\u2753 Sharing ElasticSearch indexed dataset \nHi there,\r\n\r\nFirst of all, thank you very much for this amazing library. Datasets have become my preferred data structure for basically everything I am currently doing.\r\n\r\n**Question:** I'm working with a dataset and I have an elasticsearch container running at localhost:9200. I added an elasticsearch index and I was wondering\r\n\r\n- how can I know where it has been saved?\r\n\r\n- how can I share the indexed dataset with others?\r\n\r\nI tried to dig into the docs, but could not find anything about that.\r\n\r\nThank you very much for your help.\r\n\r\nBest,\r\nPietro\r\n\r\nEdit: apologies for the wrong label\nThanks for the clarification.\r\n\r\nI am not familiar with ElasticSearch, but if I understand well you're trying to migrate your data along with the ES index.\r\nMy advice would be to check out ES documentation, for instance, this might help you: https:\/\/www.elastic.co\/guide\/en\/cloud\/current\/ec-migrate-data.html\r\n\r\nLet me know if it helps","embeddings":[-0.1457756609,-0.045346763,-0.1310747564,0.0862376168,-0.3029541075,0.1440629959,0.168963179,0.0394937918,0.0905099213,0.167836979,-0.2247992307,0.1182387918,0.0383169986,-0.0483014807,0.0813873261,0.2858945429,0.2005696893,-0.0846778378,0.1571799964,-0.1761487871,-0.0924894065,-0.1813904643,0.0924829692,0.0360959023,-0.0428278074,-0.0231426042,-0.4308625162,-0.1304083616,0.0299442615,-0.1795429438,0.469704628,0.1927620769,0.275839597,0.3492526412,-0.0001151063,0.0133273397,0.2313718647,-0.1313443929,-0.0700168684,-0.3784717023,-0.7095398903,-0.1439011991,0.3159016967,-0.4703968465,-0.0717966259,-0.3100250661,0.1885525137,-0.7246231437,-0.0082336534,0.0296246633,0.1764687002,0.0073751709,-0.0613887832,-0.0677640513,0.2456240207,0.0809686035,-0.0013593545,-0.1211186722,0.0571740121,0.1488161236,0.2379551381,0.158067733,-0.0959754065,-0.0015726967,0.173545748,0.1030583009,-0.4810886979,-0.1708822548,-0.0100431833,-0.0526020043,0.7975552082,-0.1531261057,-0.2209452242,0.1287153512,-0.0147749521,-0.0849731192,0.0635086447,0.0574337319,0.2169608623,0.1254877448,-0.236580953,-0.4179345965,-0.4528800547,0.0952967554,-0.1337144673,0.6397081614,-0.1087928787,0.0641462654,-0.1296021342,0.0166655034,0.5582042336,0.0055199042,0.1231711358,0.1779037416,0.2077921331,-0.1689664871,-0.1721054316,-0.3097271025,-0.0356601365,0.1470437199,-0.1803037226,-0.1733400375,0.2875542343,0.3189651668,0.3117184341,0.2659611106,0.1532383412,0.2794008553,0.0862537697,-0.4532053471,-0.0932770297,-0.3655295372,-0.0941490158,0.2714438438,0.2927916944,-0.3570107222,0.0569864549,-0.2542733252,-0.0886999667,-0.1678356379,-0.1851098984,-0.2368090302,0.1837714016,0.1748078614,0.0455106199,-0.4377539754,0.0412700698,-0.009010422,0.0657930449,-0.0398498476,-0.2080461383,-0.1812649071,-0.0013803255,0.2258405834,0.0353704728,-0.3625747561,0.0773466825,-0.2569189966,0.2154798508,-0.0168701336,0.4023520648,0.1661077589,0.315841943,0.2647255063,0.2203661352,0.1481296122,0.0576660708,-0.1004638523,-0.4170539379,-0.0847444758,-0.2948832214,-0.4469672441,-0.3006444275,0.0753790811,-0.121927321,0.023230629,-0.1892858297,0.320091933,-0.04891764,0.0919416696,0.0651453212,0.429769218,-0.3482821286,-0.2637747824,0.4185822308,-0.0818905458,-0.2158023268,-0.1309787184,-0.4038163722,0.3640606999,-0.0214126687,0.0175250582,-0.2373889089,0.0326789208,0.0393478908,0.4336847365,0.2714959085,-0.1292854548,-0.4007872939,0.0380572714,-0.1345510632,-0.5735340714,0.0136672873,0.1823256463,0.5256351829,-0.29685691,-0.071812667,0.5366872549,0.0097737322,-0.0651436225,-0.2112269253,-0.0803556964,-0.0552819371,-0.1397622228,-0.226126045,-0.1280879676,0.399525106,0.1132630184,-0.2694829702,0.3338492811,0.271224618,0.235200122,0.5771109462,0.3885474503,0.0084548704,0.1116839498,-0.1991685778,-0.0603856705,0.0679599941,-0.3469825983,0.0607089736,-0.4277478456,-0.2286389172,0.1216201857,-0.0791684166,0.0404194891,0.1097247973,0.0403867438,-0.0493519753,0.3112525642,-0.2652309537,-0.2971449494,0.0204566047,-0.1574488729,-0.2572397292,0.4892600477,0.0032491321,0.0899046957,-0.1495745331,-0.0277398415,0.2485863119,-0.1627816409,-0.3001868427,0.3699037731,-0.5102575421,0.4789199233,0.4329916835,0.5023630857,0.0516019203,-0.7361007929,0.6079818606,-0.0727884397,-0.1887968481,0.0689285696,-0.4780113697,0.1741653681,-0.5054827332,0.1937195808,-0.1432692409,-0.0824300125,-0.1361819357,-0.0746347681,-0.2095844895,0.0672513098,-0.1215950623,-0.230892688,-0.1964156926,0.2698990107,-0.0475250818,0.3268631399,0.0483636893,0.1000007242,0.1638967544,0.093992956,-0.066106841,-0.3816815019,0.0605910309,-0.0777859837,0.336887002,0.1360581219,0.4463493526,-0.1025030836,0.2812378109,0.0636640787,0.2965095639,0.0862451717,0.0999866426,-0.3449760675,0.0289299879,0.0542095788,0.3567815125,-0.0990595073,-0.2287160009,-0.006872871,0.0799518079,0.4823937714,0.0181276258,-0.0609977134,-0.0829006955,-0.1063559651,-0.1488537788,-0.2547786534,0.0012727914,0.3403769433,0.0274672173,-0.1734492779,-0.4013466239,0.3015756309,0.0434722342,-0.081888251,-0.1868399382,-0.173001498,-0.0350471511,-0.0165757686,0.2086957693,0.348998487,0.3991431594,0.4420820773,0.3904934525,-0.1034120172,-0.1730882674,-0.1159180552,0.1397185773,0.0154737933,-0.2044383436,0.1714712828,-0.6772081256,-0.1900226325,-0.0780407786,-0.039935112,-0.0449576154,-0.0345212407,0.0807467028,0.1899700165,0.5178708434,-0.4853645265,-0.4031346738,-0.0117790578,0.104527235,-0.2234055102,0.1328673512,-0.2687136233,-0.1638313532,-0.0073705218,0.2812463939,-0.2817280293,-0.0349922329,-0.6651854515,0.0611876808,-0.4426783323,0.0827700421,0.2310274094,0.0671742782,0.0836090073,0.0312215593,-0.2561567426,0.383764267,0.2040835172,0.0096310116,0.0569279604,0.1633331627,0.41004318,0.0706089437,-0.019218307,0.0783297271,-0.189073801,-0.0556713231,0.4821393788,0.164823547,0.161321491,-0.0689434782,0.0577240959,0.7453504205,0.1732260734,-0.115282096,0.1567568779,0.2003351897,0.197094366,-0.0396700427,-0.2997998595,-0.0960258842,0.3181846142,-0.3749881387,0.1859306544,0.0552676842,-0.1796358824,-0.1480715126,-0.5336785913,-0.2598907351,0.0502286777,0.091606319,0.1981266886,0.3708256483,-0.1932546794,-0.5065416098,0.007285763,-0.4207778573,-0.3672535717,0.2285424024,0.1188808978,0.1398949176,-0.2085955292,-0.164408356,-0.1218948886,0.2319167107,-0.1496206671,-0.1080234051,-0.1033150479,0.1664573699,0.4637421668,0.3541485667,0.04552586,-0.2242466807,-0.3911664486,-0.3795759678,0.0313668624,-0.1973666698,0.0807813704,0.0721732453,0.3277914822,-0.2468873113,0.1882665157,0.0223068409,-0.043685995,0.0979662016,-0.082546562,-0.3813145161,-0.0678571016,-0.1511036158,-0.2786829472,0.227304548,0.2290167809,0.0225729682,-0.0241477117,0.0265186224,0.1267311722,-0.1894010752,0.1069185957,0.0395644493,-0.0802999809,0.1730447561,0.2275231183,0.1138883829,0.8261784315,0.2076793462,0.3811038136,0.3793133497,-0.1029414684,0.3582115769,-0.1079950184,0.114134036,0.4698265195,-0.0042554364,0.1704633236,-0.0857633278,-0.0289557762,-0.145104602,-0.1974016726,-0.1476940066,0.0058576735,0.0794493333,-0.1790492237,-0.3474200964,0.435664475,-0.1267926246,-0.1528051794,0.3582469523,0.2564387321,-0.2605204284,0.2792612016,0.1858705729,0.9290233254,-0.0519198216,-0.0397878885,-0.3440650105,0.258150667,0.3788025677,-0.5314864516,-0.0194415636,-0.0580636077,-0.5285541415,-0.0637198985,-0.0934446752,-0.0235432107,-0.0735021234,-0.3854603171,-0.1883977801,-0.163809821,-0.0478344299,-0.1594842374,0.2797237039,0.1993723214,-0.1660588831,0.0671453327,0.0898785368,-0.146774441,0.2420863807,0.1903401315,-0.1246158034,-0.0721091181,0.1775363833,-0.3136197031,-0.2241411209,0.2794348598,0.3773968816,-0.2146373242,-0.1604981124,0.1585399657,0.0062597152,0.2064766586,0.0457429849,-0.0918211564,-0.1513352394,-0.1130282581,-0.0509921126,0.1414051056,-0.239877224,0.1609662324,0.1948281378,-0.324144572,0.2213176936,0.3517825902,-0.1224717796,-0.3239691257,0.1079271063,0.0976787806,-0.2376699597,0.0775839016,0.09700308,-0.1314314008,0.0492523983,0.0934397876,0.3093383014,-0.1214598939,0.2844890356,0.2661388516,-0.1769104898,0.0404785573,0.5570268035,0.2246734351,-0.2920640111,0.207924366,-0.2699432373,-0.0589593127,0.0366437398,0.6599254012,0.515642643,0.3183205724,-0.316064924,-0.3338963091,0.5548591018,-0.0476834141,-0.1209005937,0.1960240304,0.070760116,-0.4755544066,-0.0214261077,0.1016719714,0.2840730548,0.2609853148,0.3835826516,-0.3919514418,-0.6248450875,0.1104677171,-0.0119401589,-0.2682052851,-0.0095654177,-0.1776363403,-0.0691406727,0.4399552941,0.3236010373,0.1586662084,-0.3802181184,0.0243614782,-0.1664460003,0.0887474343,-0.0843420625,0.0068047666,0.1514452547,0.0892770216,-0.0574725904,-0.4319677055,-0.6510105133,0.0220575631,-0.0257921964,-0.3172126114,0.1984330267,0.2094599754,-0.0439779609,0.084516786,-0.2157479674,0.0832351372,0.1431358457,0.3438169956,0.3215337098,-0.2229130417,0.2338875681,0.1753680557,-0.1830292791,-0.2176288962,0.0401376858,0.1843344867,0.1131009981,0.0319416858,0.1458634436,0.1764743328,0.2464808971,0.1897221506,0.0596828163,-0.1466869563,-0.0125309937,0.0494462401,0.1684558541,0.1212230101,0.0886721537,0.0933919027,0.2917613983,-0.3603587151,0.165633291,0.1971262246,0.2134271115,-0.0954940543,0.0658894032,0.4099615216,0.1071264148,0.0901237875,-0.3227034807,-0.0585172139,0.0760971159,0.166794911,0.554987967,0.4529893696,0.2450195998,0.1695380509,0.5373660922,0.2956964076,-0.1101278812,0.3896268606,-0.135958001,-0.0248856824,-0.1249081492,0.3491644263,0.1191426069,0.0707084835,0.3210742176,-0.0247135069,-0.2048633844,0.0466394536,0.1512058526,0.1134105995,-0.030605495,0.0082231378,-0.1360280514,0.0073532113,-0.2749246359,0.099161692,-0.3227596581,-0.2442674935,0.0788005292,0.1789403558,-0.0337308608,0.2129299939,0.2452941239,0.0286623277,-0.0275922455,0.0212372206,0.0430193059,0.1435919255,0.0164045338,0.0253444742,-0.1698683947,-0.1757557392,0.0151037388,0.6578924656,0.0672674701,-0.3307161927,0.0559033938,-0.1781045198,-0.2360547185,-0.4900830984,-0.1800144613,0.1053707004,0.1492767781,0.1928827316,0.1296411753,0.2398085743,0.0768740177,-0.4116147161,0.0720799342,-0.0667614937,0.0655934289,-0.1078821421,-0.3489001989,-0.17940633,0.4989196956,0.0881647095,-0.0655881613,0.078154996,0.0292589217,0.0218474958,0.1311138123,0.3386644125,-0.0239716545,-0.1400454193,-0.3222853541,-0.2095175683,0.211889714,-0.1874551326,0.088702783,0.2236839831,0.4133697152,-0.0065184892,-0.0283104889,0.0297666173,-0.0905525163,-0.0517752208,0.1508619934,0.184623614,-0.3065720201,-0.0720717981,0.2092580199,-0.0009828982,0.3000794053,0.2042940408,-0.122089088,-0.0033513301,-0.5728544593,0.0765561685,0.0602841228,0.1637622714,0.0912501588,-0.3142159581,0.1307994127,0.0859457329,0.2388208658,0.2297323048,-0.0563217811,-0.0285836272,0.3390567899,0.0303539112,-0.1852874756,-0.4962013364,-0.4122146666,-0.2655918598,-0.1981622279,-0.1734670401,-0.2189140767,0.0473858118,-0.4242605567,-0.2663052678,0.1756478399,0.3380908072,0.0397033915,0.0451988019,0.0734681338,-0.1010468826,-0.0787755325,0.1480128616,-0.3056065142,0.097636044,-0.0687249899,0.1028552726,0.2726200521,0.1440494806,-0.0303102992,0.0846571848,0.5821902156,-0.2577877343,0.1556805223,-0.0446349159,0.2648150623,0.2533211708,-0.2322271764,0.085444741,0.0414164625,0.1609145552,-0.0746956766,-0.1033958718]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1299","title":"can't load \"german_legal_entity_recognition\" dataset","comments":"Please if you could tell me more about the error? \r\n\r\n1. Please check the directory you've been working on\r\n2. Check for any typos","body":"FileNotFoundError: Couldn't find file locally at german_legal_entity_recognition\/german_legal_entity_recognition.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py\r\n","comment_length":24,"text":"can't load \"german_legal_entity_recognition\" dataset\nFileNotFoundError: Couldn't find file locally at german_legal_entity_recognition\/german_legal_entity_recognition.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py\r\n\nPlease if you could tell me more about the error? \r\n\r\n1. Please check the directory you've been working on\r\n2. Check for any typos","embeddings":[-0.2461109459,-0.2954854667,-0.040122088,0.4629744589,0.134741649,0.0751197487,0.1829098016,0.2415550202,0.2743229866,0.084339641,-0.1621886194,-0.2467778772,-0.0410388298,0.2271589786,0.1816771328,-0.3239541352,-0.022687709,0.043106351,0.1395513117,-0.2421332002,-0.1837922037,0.3086111248,-0.1309448034,0.2035576105,-0.2785853744,0.1524153352,-0.1371282935,0.4542097747,-0.3045515418,-0.2061940581,0.243312344,-0.2026706636,0.3079298735,0.4991836846,-0.0001100421,0.0578923151,0.2836090922,-0.0098897405,-0.2552669048,-0.4077288806,-0.000673596,-0.2990614176,-0.1864303499,-0.1148855016,-0.2259489447,-0.2796554267,0.1619674116,0.1354253888,0.2634932995,0.1191348284,0.2091219276,0.2212058753,0.2193627506,-0.3371452987,-0.1168966964,0.2205763161,-0.0645754561,0.5331406593,0.1345074773,-0.2835020423,0.2549849451,0.227091521,-0.0557186753,-0.0208476763,0.3751378357,-0.0840708092,0.3238506317,-0.3081605136,0.2847310305,0.234388113,0.580249548,-0.0297926161,-0.4290724099,-0.0293356739,0.2880429029,0.0057340362,0.2407151312,0.3006296158,-0.0669369847,0.1814675182,-0.1373806745,-0.4443282783,-0.3213987947,0.2860621214,-0.2091830373,0.0494364537,-0.2200017869,-0.0653783605,0.2275342345,-0.2309976071,-0.4672821164,0.02005703,0.0080825547,0.5115160942,-0.1440299898,0.026818607,-0.0979861096,0.3844959438,0.1937473714,0.1407542825,-0.0220368207,-0.081481725,-0.1746383458,0.4402425885,0.0594136417,0.3439055383,0.1706577986,0.2345796674,0.1332697272,0.1937059164,-0.1103216931,0.0529637858,-0.4952788651,-0.3095989525,-0.35193488,-0.2179227024,0.3497495055,-0.3086481988,-0.2129980773,0.0519950278,-0.188149184,-0.0218673814,0.34618783,0.5144165754,-0.064978011,-0.2256052792,-0.1134259552,0.173306644,-0.1949893981,-0.0240159798,-0.2273389846,0.2044352293,-0.2663430274,0.1077500731,0.1484897137,-0.4823882282,0.4201607704,-0.2502062023,0.0515125133,0.0119270589,-0.1076055914,-0.07985273,-0.1945853382,0.2724869847,-0.2817009687,-0.1495464295,0.2713314295,-0.2432877421,-0.1272886097,-0.0838614032,-0.2914420664,-0.3538348675,-0.3816328645,0.1404944211,-0.1598204821,-0.1504817158,-0.1611048132,0.1011340767,-0.1878173947,-0.3983779848,0.003076813,0.0209349561,-0.1749327779,-0.1412834227,0.3550225496,0.660187006,-0.1571733057,-0.1841322929,0.0235803425,-0.2447856367,0.1131006628,0.7056491971,-0.23356691,0.0617510192,-0.4061587155,0.2985373139,0.2582632899,-0.5748142004,-0.242413938,0.3591725528,-0.2599228621,0.055071637,-0.0022604682,-0.0761059597,-0.2731065154,-0.1170467883,-0.0694173649,0.1008342132,0.0887652189,-0.2138914913,-0.1731622517,-0.1359585822,0.1755654216,0.3268742859,0.1123142168,0.102979809,0.2103891224,-0.0414946526,-0.0276233666,-0.0393747576,-0.1953698993,0.3031049669,0.2408164889,0.3216609657,0.0204128716,-0.1257183254,-0.3142140508,0.3031443954,-0.3899177015,0.2161759436,-0.2050553113,-0.1065893546,-0.2555696368,-0.0077495999,-0.3642155826,-0.1737731546,0.0880405232,0.0162563585,0.2212073952,0.1731007546,-0.0889178962,0.1511776149,-0.0848262981,0.1308608204,-0.3367919922,0.2155641615,-0.2358353138,0.1930098683,-0.0413667262,0.1421066225,0.1962637007,-0.2131726146,-0.0544220507,0.3109876215,-0.2180805802,0.2646730244,0.4681869745,0.3286012113,0.2611548603,-0.26838094,0.3497458696,-0.0667929128,0.1807490736,0.1038902849,-0.1683859229,0.083073236,0.1250580996,0.0417994745,0.0687506497,-0.0625735596,0.0969749838,-0.0657422021,-0.0689738393,-0.0130336583,0.4850594103,0.0714358389,0.3537172675,-0.0129299164,-0.375918895,-0.2800389528,0.2990945876,-0.1732295603,0.1164980382,0.2613541484,-0.2369155884,0.0857469961,0.0461429916,0.0166967437,0.3218719661,0.1582996845,-0.1274445355,0.0714788437,-0.0126155028,-0.1146031097,0.0880259499,-0.0003688995,0.0651855022,0.240393728,-0.107860893,0.0802324042,-0.4135636985,-0.3385950923,-0.0826898366,0.2061855793,-0.2287418395,0.0982242972,-0.2093147635,-0.478164494,-0.2285084724,0.0613498837,-0.5594624281,0.0963179693,0.066031158,-0.033255972,0.0308852997,0.1123059765,0.039493341,0.2607886195,-0.184775278,-0.2399189472,-0.2281203717,-0.1024246737,-0.2071873993,0.0067287451,0.5578662753,0.196608901,0.430419445,-0.2647414207,0.0509952605,-0.3066093028,-0.4186324477,0.0406331159,-0.1041649207,0.2540677786,0.0432715751,0.2298954725,-0.3527813554,-0.1797481328,0.5061764717,-0.1521884501,-0.2286915183,0.0058346274,-0.0436440408,-0.2822234631,0.0267798062,-0.1938976645,-0.4402248859,-0.3264966607,0.4318302572,0.0639394745,0.0518777966,0.3492003083,-0.0634190589,0.3247578442,-0.1961455494,0.1281203479,-0.1535876542,-0.4176025689,0.2414850891,-0.3585348427,-0.3183611631,0.0801431462,0.1686861813,0.2564223111,-0.0917546004,-0.173003152,-0.2125993967,0.005323017,-0.0363134146,-0.0836634859,0.1677423567,0.1912681609,-0.1686710268,-0.0399982147,0.1529936194,-0.0419996828,0.0001621202,0.1548144817,0.198603943,-0.0571862459,0.2133805752,-0.1873622388,0.670863688,0.1992785782,0.0671118051,0.4373435378,-0.0646627024,0.2162336409,-0.1551831663,-0.4231314957,-0.0120128877,0.1890667379,-0.0339833163,0.3280712664,0.2138596773,0.0712214708,-0.3305560052,-0.5055467486,-0.382048279,-0.0762637332,-0.0656919852,0.0381204039,0.3100767136,-0.0659812018,-0.158364892,-0.0825098678,-0.0293632373,0.3220584989,0.6643012166,0.165501982,0.1561715156,-0.3311134279,-0.2678553462,-0.2221661657,0.4375950098,0.0761239976,0.0173630547,-0.449706614,0.1185971871,0.1560296416,-0.1163388491,0.5943135023,-0.0935931727,0.1053122208,-0.1405805945,-0.1460764557,0.0205050074,0.0110441055,0.2064360082,0.2797380984,0.0606897399,0.3156520426,-0.072266154,-0.1791745722,0.1312684417,0.1703105867,-0.1624177247,-0.3358234167,-0.3524512947,-0.1666998416,-0.4119659066,-0.1022674739,0.0257699788,0.1470804662,0.1410583109,0.185291931,0.0059481137,0.100339666,0.1731918752,0.1170120612,0.2713026404,-0.0170245804,-0.0409970656,0.5033988953,0.3873663843,0.3642572165,0.8100678921,0.0326409638,-0.6243072152,-0.0792534649,0.1536627412,0.1142801344,0.3046535254,0.0781558678,-0.1562288254,0.3815321922,0.1248529628,-0.3096081614,-0.1216216907,0.230099082,0.0362181626,-0.2854736447,-0.1932158172,0.4634526968,-0.0314112008,0.1584514976,0.1678153425,0.3962773085,0.0294215586,0.1493307203,-0.2323090285,1.117784977,-0.2501133382,-0.0336063839,0.0783227608,0.079431735,0.3679837584,-0.2429523319,-0.0302174911,-0.2846556008,-0.0506035909,-0.0966738164,-0.0900375769,0.3480014205,0.1589499265,0.0190859139,0.3904221654,-0.0050295321,0.0213309824,-0.1037063897,0.5435551405,-0.1875599772,0.0483768992,-0.38116166,0.1587076038,0.1709783971,0.9175265431,-0.240570277,-0.2204122692,-0.1932104081,-0.1632933915,-0.2511785626,0.1232749745,-0.171674937,0.191244185,0.0557752512,-0.1219136715,0.208379373,0.3882264197,0.1826113462,-0.0764015988,-0.522999227,0.0537897274,0.0085395575,-0.4368182421,-0.1741906404,0.0277117621,0.3646520376,-0.1686219126,-0.2650587559,0.008052174,-0.0954893157,-0.0469613299,-0.15455392,-0.2065621018,-0.0931533054,0.1195601076,-0.220408693,-0.1327029914,-0.0482055806,0.0758237466,0.107942231,0.2924996614,-0.120882839,-0.3240250349,0.1940034628,-0.1589924842,0.0031829062,0.5526271462,-0.2241658866,-0.2374397814,0.5092985034,0.0583982989,-0.0459430367,-0.2427573204,-0.085586898,0.2943687439,-0.6377717257,-0.1279045492,0.4180642664,0.1766909808,0.1304137111,0.3004823625,0.379340142,-0.4835888743,0.0352339372,-0.4407584667,-0.0572316758,0.2066504806,-0.0869095027,0.0205387715,0.1090745553,0.1655276865,0.2694160342,-0.0816535875,-0.3106639981,0.1539812982,-0.2633136213,-0.1432812512,0.2972684801,0.0894583017,0.3182830513,0.0721745715,0.158940196,-0.0666907206,-0.2371606678,-0.1986877173,0.0156590138,0.1709641665,0.01937126,-0.096448794,-0.0131185222,-0.2603687048,-0.0547049604,0.1063845232,-0.0237938706,0.1455268413,0.0162385236,0.0693527833,0.0125743626,0.0340332463,0.0533114932,0.2057837546,0.0302573331,0.0266188029,0.0853226632,-0.134053722,-0.0473757274,0.0088502439,-0.1550019979,-0.0149423825,-0.0044979891,-0.0131079573,0.2423468828,-0.2492470592,0.1270388663,0.0687045008,0.1330344826,0.1983101368,-0.3698763251,0.248224929,0.5175201297,0.2364893407,-0.4058769345,-0.1334460527,0.132076472,-0.2481372654,-0.1152819693,0.1054717973,0.239909485,0.2900750339,-0.0831424966,0.0457253531,0.2982501686,-0.1296290457,0.1374139339,0.2040880769,-0.149570629,0.0734469295,0.0053792875,0.3130514324,0.201922372,0.2272089571,-0.2314146161,-0.110761255,-0.5035055876,0.5509961843,0.1265520006,-0.446438998,0.0432436094,0.2720758319,0.1229607314,0.0511240587,-0.2502061725,0.5838778019,-0.2880882025,0.0189929754,-0.2728719413,0.3538155556,-0.1155236214,0.0331161432,-0.2132678032,-0.2617594004,-0.0792261288,-0.1460847408,0.0441380627,-0.1830733865,-0.1822914183,0.1761893779,-0.1251681894,-0.1299707294,-0.1938256323,0.3341676891,0.1599289179,-0.1620254964,0.2216958851,0.1497132182,-0.0616784655,0.2485283017,0.3011996746,0.554566741,0.1641010642,-0.0174190942,0.0392229669,-0.0806010067,-0.0786468759,0.0641745403,0.3036924899,0.0605673641,0.1531082243,0.3303853869,0.129868418,-0.1067740321,0.084173061,-0.2326614559,0.1154810414,-0.1030599624,0.3758984506,-0.4738535881,0.0834749788,-0.4334076345,-0.012331672,-0.5144891143,-0.010893207,0.1059370935,0.1479999423,0.0318016745,-0.0529207103,0.0680522323,0.0072843153,0.4513627291,0.7064887285,0.2970288396,-0.3031168282,-0.1937813908,-0.6371819973,0.154656902,-0.0132896462,0.0063500162,0.0004408012,-0.0258289799,0.0355728045,-0.0980792865,-0.0328977741,-0.1459824592,0.1888629645,0.1436649561,-0.0407395549,0.1399946362,-0.1291047633,0.1612912416,0.0439980179,-0.1437498033,0.1138733625,-0.1729668081,-0.0409156904,-0.2266374826,0.3342975974,-0.4176327288,-0.3125066161,0.2645232081,0.0058600549,0.469393909,0.0195388272,-0.1019147635,-0.0850109458,-0.1184910163,-0.1584532857,0.3720178008,0.150270015,0.2475247979,-0.1619675159,-0.2581013441,-0.2843648195,0.0392013304,-0.1302444041,0.1434408724,-0.2634316385,-0.1647253633,0.0208955836,-0.1051236466,0.1348500401,0.0568217225,-0.0135655254,-0.1285772473,-0.2913085818,0.0122211138,0.8393113613,-0.3831662834,0.0767079666,0.1538659483,0.2517720461,0.1479532719,-0.1441248357,-0.6370150447,0.2956531048,0.3046897948,0.282369554,-0.064346984,0.2194420248,-0.1137199774,-0.092481792,0.0013684818,-0.0549191087,-0.0685572401,-0.1624987274,0.116110377,-0.0317010172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1299","title":"can't load \"german_legal_entity_recognition\" dataset","comments":"> Please if you could tell me more about the error?\r\n> \r\n> 1. Please check the directory you've been working on\r\n> 2. Check for any typos\r\n\r\nError happens during the execution of this line:\r\ndataset = load_dataset(\"german_legal_entity_recognition\")\r\n\r\nAlso, when I try to open mentioned links via Opera I have errors \"404: Not Found\" and \"This XML file does not appear to have any style information associated with it. The document tree is shown below.\" respectively.","body":"FileNotFoundError: Couldn't find file locally at german_legal_entity_recognition\/german_legal_entity_recognition.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py\r\n","comment_length":77,"text":"can't load \"german_legal_entity_recognition\" dataset\nFileNotFoundError: Couldn't find file locally at german_legal_entity_recognition\/german_legal_entity_recognition.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py\r\n\n> Please if you could tell me more about the error?\r\n> \r\n> 1. Please check the directory you've been working on\r\n> 2. Check for any typos\r\n\r\nError happens during the execution of this line:\r\ndataset = load_dataset(\"german_legal_entity_recognition\")\r\n\r\nAlso, when I try to open mentioned links via Opera I have errors \"404: Not Found\" and \"This XML file does not appear to have any style information associated with it. The document tree is shown below.\" respectively.","embeddings":[-0.1284616888,-0.0970662758,-0.0301719233,0.52423352,0.2877425551,0.138086766,0.1956101209,0.2292719632,0.1878093183,0.1808798313,-0.3449552059,-0.1354647428,0.0454051569,0.238665849,0.1107380167,-0.3510169089,-0.1016481742,0.0041541751,0.1530084312,-0.1696580201,-0.1890911311,0.3157928884,-0.1670924276,0.2493717968,-0.12845397,0.2280391455,-0.1583690494,0.5008804202,-0.267236799,-0.3464731872,0.1775410622,-0.1727831811,0.1381816566,0.5028778911,-0.0001160817,0.1684517264,0.3584832847,0.0463054292,-0.3028357327,-0.4721885622,-0.0444821008,0.011026741,-0.1089340225,-0.1386529654,-0.1597964764,-0.2255044729,0.0787687302,0.0283097997,0.1838006675,0.3876415789,0.1323372275,0.2076338083,0.2036084682,-0.239934504,0.112439163,0.069475919,-0.0881248787,0.3908723891,0.1431781501,0.0015836535,0.3734642267,0.2545952499,-0.214456439,0.0117633203,0.3610550165,-0.0947234556,0.2621243894,-0.4173006713,0.3253712952,0.3081182539,0.8236352801,-0.0829197094,-0.4649599791,-0.0205962099,0.1881633401,-0.0181164816,0.3648922443,0.2760870755,-0.0488722064,0.1934266239,-0.1533073336,-0.3488994837,-0.289683938,0.2990824878,-0.1846633554,0.0604278706,-0.1255752444,-0.0613287054,0.3649992645,-0.1658501327,-0.3809635341,-0.1153136492,0.0675977021,0.5025035739,-0.0852106288,0.1173983365,-0.0282755103,0.3216753006,0.1948667914,0.1033321321,0.0905859619,0.0120888771,-0.2064005136,0.4004949331,0.0734971538,0.2989094853,0.030573098,0.006263718,0.2821767926,0.1956744492,-0.0332635939,0.1224434227,-0.4561715126,-0.3617556095,-0.4565536082,-0.1500506103,0.4977904856,-0.4203483164,-0.3818337321,0.1402867287,-0.2901342511,-0.1732571274,0.335816592,0.6531698108,-0.2122738212,-0.1192939207,-0.1761870831,0.1861924231,-0.3083966374,-0.0409370661,-0.1233855113,0.1253173202,-0.245634079,0.2390463352,0.144863382,-0.3951359689,0.2847658098,-0.2692220807,0.0242194589,-0.1080353558,-0.1189616099,-0.2255274355,-0.249380976,0.3539412618,-0.1349590868,-0.1933986098,0.2049238831,-0.3587966561,-0.1289900541,0.0766790211,-0.327272743,-0.4347092807,-0.2450088412,0.1028737351,-0.3333546519,-0.1754341424,-0.1302225441,0.0702588037,-0.1435457915,-0.3391142488,-0.0810110644,0.0006705662,-0.2801930308,-0.1255639642,0.3089601099,0.6876109838,-0.0440485701,-0.1589181423,0.1731986552,-0.2280242145,0.1038119122,0.6274731159,-0.191130802,0.1403751969,-0.319752574,0.1875180453,0.4359926283,-0.6256730556,-0.2637548149,0.2893291414,-0.1598046869,0.0991236046,-0.065715462,-0.1781258434,-0.159752056,0.074345462,-0.0614383258,0.0533809662,0.1809496582,-0.0987008363,-0.1542083323,-0.1613453031,-0.0043838653,0.3532508314,0.1352611035,0.0834900662,0.2695771456,-0.286185056,0.0958966538,-0.1564672887,-0.1349140555,0.4070647955,0.1309380829,0.5612075329,0.1535349339,-0.1372198015,-0.1892559379,0.2605710626,-0.2483993024,0.250526607,-0.1847560704,-0.1427255273,-0.3887182772,-0.0137939975,-0.4441559017,-0.1686789691,-0.0096325669,0.0590864867,0.1441983134,0.2651419342,-0.1881757528,0.1842205971,0.0520941168,0.2091677934,-0.4159216881,0.3306803405,-0.2947822213,0.1462970525,-0.1381468624,0.2533885837,0.1048906371,-0.201987505,-0.1726142317,0.3779194057,0.0250423625,0.1596824527,0.3278082013,0.471762985,0.2786517739,-0.4465465248,0.24975577,0.0216959529,0.2192718834,0.1120481789,-0.2551671863,-0.0079791378,0.1175501049,0.1259956658,0.2405890971,-0.0571400039,0.0639624596,-0.0355957896,-0.1666634381,0.0263212621,0.5831589103,-0.0020470028,0.4861436486,-0.0022115733,-0.4058570862,-0.2668928504,0.2205027193,-0.1580255777,0.1402618587,0.2830582261,-0.2300815582,0.0876422971,-0.0906660259,0.1340228319,0.2167288959,0.1192997843,-0.1595410854,0.2266142219,0.0307773091,-0.185721457,0.0778872073,0.1208221912,0.1685411632,0.1628404707,-0.0735765547,0.1814069301,-0.3839359283,-0.4361386597,-0.0667828843,0.0940121189,-0.2802446485,0.2051865309,-0.1672077477,-0.3744663894,-0.1607932895,0.0722543374,-0.485195905,0.0286906119,0.0245605502,0.1296552122,-0.037622828,0.0515218414,-0.1298662275,0.2387040704,-0.1413996965,-0.2614262402,-0.0420084484,-0.0587313473,-0.2014176548,-0.0789254531,0.4222655296,0.1445723325,0.17905505,-0.3623304963,0.1587977111,-0.2472373098,-0.3430608213,0.1865047961,0.0980219468,0.1405782253,-0.1274766922,0.102063328,-0.2902251184,-0.108414501,0.5380041599,0.0940812379,-0.1029296443,0.0803398043,0.008339758,-0.2383141369,0.0861875042,-0.2965488434,-0.5001032948,-0.3831173778,0.2841564715,-0.1005473286,0.2432362735,0.4525517523,-0.0493821613,0.3326973319,-0.0926731676,0.2460079044,-0.1583628803,-0.4093944132,0.3129261136,-0.4540614784,-0.3857223094,0.0652327985,0.1502108574,0.2502662838,-0.0103789754,-0.3031935394,-0.0848824158,-0.0673022941,-0.0191881657,-0.1619923711,-0.018712664,0.2730546892,-0.027946746,-0.0371688157,0.0137552675,-0.0009872422,-0.0250708554,0.1494068205,0.2231205553,0.0141198635,0.2424313426,-0.1294840574,0.6173750758,0.1517210007,0.1059064865,0.5841574669,-0.170213908,0.2498150766,-0.0573802926,-0.352294594,-0.0659816042,-0.0265934356,-0.134411931,0.3559343219,0.319192946,-0.0453878306,-0.4974301457,-0.2689446807,-0.2581544518,-0.0719563887,-0.0121803498,0.0068369079,0.2516325414,-0.0952537656,0.0993749648,0.0060829371,0.0294295419,0.2890433073,0.6507748961,0.1626077145,0.0567377321,-0.210628137,-0.2528006732,-0.3029285073,0.3592164516,0.0446701869,-0.0879454315,-0.4004314542,0.0914236158,0.1656910032,-0.149118349,0.6489861608,-0.1175707504,0.0423878096,0.0016189333,0.0530691594,-0.1806643158,-0.0166646577,0.1738822162,0.2548440397,0.0215677619,0.3470984697,-0.0611725226,-0.1713438183,0.0787485316,0.161940217,-0.1692982614,-0.2994202375,-0.3161098361,-0.2759823501,-0.4376220405,-0.1102142632,-0.0771074966,0.1999564171,0.032008417,0.050813213,0.1499728709,0.1454230547,0.1553143412,0.1777428091,0.2041412592,-0.0463148691,-0.0611939691,0.5820173621,0.3182827532,0.4499576986,0.9425209165,0.0126380222,-0.6109728217,-0.0634975731,0.1455767453,-0.0340406261,0.2525072694,0.1000837311,-0.3182469308,0.350146085,0.022013288,-0.3123271465,0.0172680691,0.2669469714,0.1146985218,-0.2009669244,-0.3476184905,0.5151138306,-0.049219951,-0.0451108143,0.2192759514,0.4564252496,0.0591217726,0.1847100854,-0.1947250217,1.1007053852,-0.2717573941,-0.0526265725,0.2020338923,-0.016612621,0.07605315,-0.2747394443,0.0693777055,-0.1266998053,-0.0832686648,-0.1203595772,-0.2141693681,0.2875407636,0.1524730623,-0.0292159952,0.3771092296,-0.002062192,-0.1013656706,0.0334239267,0.2994051278,-0.1524931341,-0.0252194069,-0.2375162244,0.1009337977,0.0430852398,1.0740029812,-0.1547463387,-0.1891571581,-0.2770196199,-0.1850480586,-0.2150519341,0.3156625032,-0.2351907939,0.1516550183,-0.201139614,-0.1391521096,0.0346118771,0.5869801641,0.0920319259,-0.1826558411,-0.4452542961,0.1263446659,0.0421454944,-0.3262606263,-0.094537735,-0.0742372274,0.4136698246,-0.1770841926,-0.3557676077,0.0859502554,-0.0714371502,-0.1525382996,0.0953514352,-0.3138718605,-0.2005218416,0.1050699353,-0.1540896595,-0.117580317,-0.0927768722,0.0422036983,0.0258594863,0.1876173615,-0.1187161803,-0.1854834855,0.1494340748,-0.2381450683,0.0512902439,0.588802278,-0.2491498441,-0.2321004421,0.4923295975,0.1366745234,-0.1005723774,-0.1319222152,-0.1162578389,0.1764931828,-0.3826814592,-0.0913547352,0.022586694,0.1735225171,-0.1091829985,0.085293375,0.2728910744,-0.3769960403,-0.0233340599,-0.5408440828,-0.1972014457,0.3759463131,0.234879896,0.0364417024,-0.0130627537,0.0862952918,0.3445598483,-0.3093902767,-0.233953312,0.1378255486,-0.2675618529,-0.1870756745,0.4149504602,0.0234971792,0.2746323943,-0.1204062328,0.0425672606,-0.2816507816,-0.29673177,-0.0935780182,-0.0209885743,0.1799411774,-0.0201436095,-0.1855713725,-0.0260274969,-0.2796741128,0.0380660146,0.0263078474,0.0986684263,0.2541894317,-0.0806547105,-0.1497017145,-0.0389992483,0.1606203318,-0.0285040662,0.1658855379,-0.0250785556,0.2170055211,0.1362279207,-0.1153361276,-0.0526594967,-0.0562997684,-0.2351646572,-0.1351748556,0.1196875423,-0.1064519882,0.3019508421,-0.3200264871,0.2494128495,0.2363189608,0.1513902545,0.2151640207,-0.3591763973,0.2953815162,0.3524010181,0.1845543534,-0.2272461504,-0.1679018438,0.0308032874,-0.1785126626,-0.0863494575,0.0024558555,0.2729935944,0.3457306027,-0.1994317919,0.0774596483,0.4343309104,0.145699203,0.1985399872,0.2813809812,-0.2331503034,0.1715175658,-0.002649066,0.1979810745,0.2629511058,0.4025625288,-0.3038870692,0.1397800744,-0.0790857077,0.6005098224,0.1578238755,-0.4051303267,-0.0515115038,0.073706992,0.2421593964,0.0153266955,-0.1426185071,0.2035051435,-0.2709720135,-0.0616923869,-0.3238830268,0.355550766,0.0263442025,-0.0678499267,-0.1990638524,-0.3584107757,-0.0612693317,-0.082978569,-0.0345828719,-0.1905530393,-0.1526883394,0.0129223289,-0.1009118259,-0.1997891366,-0.2719909549,0.2633304298,0.2288675308,-0.1135562807,0.3110853434,0.2184485942,-0.1059438065,0.1514738053,0.3076580167,0.5152637362,0.2543149889,-0.1152786687,0.0155461971,-0.2091235071,-0.0948872194,0.1704777926,0.1511585116,0.1068275645,0.0460992046,0.2802452445,0.0151049094,-0.0751340687,0.1925445497,-0.2286165655,0.2965637743,-0.1104427204,0.2212409228,-0.3549813628,0.0751564503,-0.3543505073,-0.0653172284,-0.5037228465,0.1704410762,0.2489495575,0.0723910183,0.0488217138,0.0041601779,0.0330383517,0.1352712661,0.376422435,0.6063412428,0.5643181801,-0.2106999904,-0.0940877944,-0.6297900081,0.2138065845,-0.1053326577,0.0101098903,0.1011180282,-0.0779324174,-0.1457156986,0.0350367501,0.0920496881,-0.1372896135,0.3103255033,0.1286796778,-0.1351249963,0.0758847296,-0.1106577963,0.2553890944,-0.0463654995,-0.2306019217,0.1476125866,-0.0851738304,-0.1164185852,-0.1416279227,0.2353745699,-0.4128286242,-0.5597071648,0.4009501338,-0.0996348709,0.575242579,-0.0311707892,-0.0989595428,-0.0761373714,-0.2527616322,0.0294083841,0.3914283216,0.1576301903,0.1113566607,-0.1932994425,-0.2572252452,-0.2031803876,0.2207051963,-0.1014235094,0.0736487955,-0.2639419138,-0.2942397594,0.0518652946,-0.0410699286,0.0597422719,0.0186676197,-0.0547693893,-0.0834889561,-0.1638526469,-0.0529189743,0.9457668066,-0.2903810441,0.1950922459,-0.0955955684,0.2677007914,0.2320239097,0.0235326104,-0.5080148578,0.2224430442,0.2420266122,0.1784065217,-0.2893014848,0.0993578136,-0.0520632155,-0.1598246247,-0.0468340889,0.0172137562,-0.1551774889,-0.0938545167,0.0368307978,-0.1078802273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1299","title":"can't load \"german_legal_entity_recognition\" dataset","comments":"Hello @nataly-obr, the `german_legal_entity_recognition` dataset has not yet been released (it is part of the coming soon v2 release).\r\n\r\nYou can still access it now if you want, but you will need to install `datasets` via the master branch:\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`\r\n\r\nPlease let me know if it solves the issue :) ","body":"FileNotFoundError: Couldn't find file locally at german_legal_entity_recognition\/german_legal_entity_recognition.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py\r\n","comment_length":52,"text":"can't load \"german_legal_entity_recognition\" dataset\nFileNotFoundError: Couldn't find file locally at german_legal_entity_recognition\/german_legal_entity_recognition.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/german_legal_entity_recognition\/german_legal_entity_recognition.py\r\n\nHello @nataly-obr, the `german_legal_entity_recognition` dataset has not yet been released (it is part of the coming soon v2 release).\r\n\r\nYou can still access it now if you want, but you will need to install `datasets` via the master branch:\r\n`pip install git+https:\/\/github.com\/huggingface\/datasets.git@master`\r\n\r\nPlease let me know if it solves the issue :) ","embeddings":[-0.287135601,-0.2293046713,-0.0558743365,0.3209625185,0.1926872134,0.1035037786,0.1227492914,0.239558354,0.2547089756,-0.0033335367,-0.1283259392,-0.1231847852,-0.0620686971,0.3180630505,0.1877844334,-0.3089747429,0.0005011986,0.015175079,0.0721771196,-0.129286781,-0.1097700596,0.3460334539,-0.0871673822,0.1794844419,-0.2606520057,0.1178169921,-0.0188958384,0.2385766953,-0.4208890796,-0.3372544944,0.2726183534,-0.0110955201,0.2925595641,0.371425271,-0.0001043722,0.0022043132,0.3241093755,-0.0111567322,-0.2839604914,-0.3169110417,-0.0197815206,-0.3420846462,-0.1262253225,-0.0518512838,-0.3168023825,-0.2389598191,0.1567593962,0.1090816408,0.2710792124,0.1588666737,0.2833471596,0.1991866529,0.270620048,-0.360001564,-0.0701991469,0.0936058387,-0.1158882603,0.4267020822,0.1928993613,-0.2007417381,0.2586779892,0.2370307893,0.116937682,0.0021750489,0.3039171398,-0.024705613,0.3294273913,-0.2710103095,0.2744709849,0.2596640587,0.4311144054,-0.123146072,-0.4543117583,-0.101801075,0.2980352342,-0.0944480523,0.2075906694,0.2430980504,-0.0850081593,0.2237718999,-0.1252354681,-0.3121505082,-0.2083556652,0.2568840683,-0.2121196538,0.10334263,-0.2279229611,-0.063815482,0.305175215,-0.2178628445,-0.4415865242,0.1773997098,-0.0384146236,0.5027360916,-0.0967360586,0.1253807396,0.0440346785,0.348957032,0.281717509,0.1656108946,-0.0340988412,0.0228382517,-0.0802756473,0.4368696809,0.0577675477,0.3527481556,0.0837927386,0.2032633126,0.1320182085,0.280441016,0.0159925073,0.0769498572,-0.4701928198,-0.170616135,-0.3189519644,-0.1559458673,0.2127619684,-0.2779736519,-0.2171692252,0.0097234305,-0.0927772149,0.0084325718,0.3076982796,0.4883945286,-0.0327314734,-0.2342407852,-0.0934844688,0.1717046797,-0.1615845859,0.0802362263,-0.2343033999,0.0346785411,-0.2837478518,-0.0286612548,0.2533976436,-0.4500764012,0.4466090202,-0.1697286069,0.0600390323,0.1076340079,-0.0652886629,0.0629625693,-0.1449688226,0.2517574131,-0.1662921458,-0.1733265817,0.1900547892,-0.3055607378,-0.1252636909,-0.1787098497,-0.2166444659,-0.3605930507,-0.37529248,0.2259646505,-0.2043678761,-0.187016964,-0.1884661317,0.1423495114,-0.1045726389,-0.4150547385,-0.0911060497,-0.015621705,-0.0996251032,-0.0927589387,0.3819972575,0.4307042062,-0.1753313094,-0.2079688311,-0.0456437804,-0.2844352722,0.1065340787,0.4751934111,-0.2280706912,-0.0386946984,-0.396861434,0.2115337998,0.1432289928,-0.453361243,-0.3421435356,0.2101932317,-0.1873482615,0.0749401674,0.0014123755,-0.1085762084,-0.2418076843,-0.0224312004,-0.1778215766,0.1832603663,0.1322078854,-0.1387587935,-0.0946837962,-0.1494649351,0.0830519348,0.1617780328,0.0679191053,0.1183405146,0.1734056771,0.0127754696,0.0263376106,0.0217664279,-0.1108661219,0.2096807808,0.224819839,0.3467616439,0.0571014285,-0.1895009875,-0.2203349471,0.307705611,-0.4281440973,0.2901538312,-0.3179882467,-0.2025410235,-0.3226408362,0.0154769868,-0.3389259875,-0.2649016976,0.1795160919,0.1342784911,0.2389125377,0.1872345358,-0.0360569693,0.0982850417,0.1055332795,0.1050726026,-0.402998805,0.1303628236,-0.2962359488,0.1427173615,0.0135174822,0.1573440433,0.083754845,-0.1933580935,-0.0413600132,0.291136831,-0.2500871718,0.3001385927,0.288407445,0.3958922923,0.2157092541,-0.2101268321,0.2361978441,-0.1107795089,0.1215829477,0.0962872505,-0.2546933889,0.1321818978,0.0794701874,0.0011697293,0.051285997,0.043703761,0.1992012411,0.0104948403,-0.0914536193,-0.1404322237,0.2713919282,-0.0491027795,0.3081584871,-0.1044256687,-0.2544832826,-0.1351487786,0.387834996,-0.020589333,0.1426635832,0.2112210989,-0.2736790478,0.0250289291,0.1215223297,0.0918626711,0.2886164188,0.2752539814,-0.0311059374,0.0427798927,-0.0110768965,-0.0889173299,0.0887093395,-0.0508761816,-0.0482559986,0.1743098646,-0.0122377193,0.0207016468,-0.4914883375,-0.1776544154,-0.0759711117,0.2441278845,-0.1945655346,0.0935321674,-0.2653297782,-0.410626471,-0.1424117684,-0.0438923091,-0.5049240589,0.0443081111,0.1313380003,0.0712512136,0.0490951054,0.129050374,0.0700280741,0.1916438043,-0.1662600338,-0.134924069,-0.310583204,-0.0739901289,-0.2353853434,0.1117409691,0.3545113504,0.2079512775,0.4717747271,-0.1894608289,0.1324892938,-0.3272043169,-0.600632906,0.0623441152,-0.2352684885,0.3087221384,0.1546491235,0.1451974809,-0.3562120497,-0.1046805456,0.4863615036,-0.1898590624,-0.2475607246,-0.0297015887,-0.1127689481,-0.3060136139,-0.041085545,-0.1921435744,-0.4251380563,-0.384382993,0.548238337,0.1713629663,-0.0013478778,0.3274297416,-0.088318333,0.2806099355,-0.1597571969,0.1602456421,-0.2694223523,-0.3390863538,0.2167554349,-0.4446284473,-0.3358726501,0.0180750545,0.1143524498,0.3700516224,-0.0964458957,-0.1132557914,-0.2231542468,-0.0799965337,-0.0062224856,-0.0623166449,0.1297807693,0.2979497612,-0.0672560036,-0.0971203595,0.0130105522,-0.0781469196,-0.0373245329,0.0681313425,0.2425166667,-0.0593306758,0.1630978733,0.0247505642,0.6636180878,0.219872117,0.0549555272,0.3495468199,-0.1756163687,0.2357822061,-0.1428213269,-0.4567848444,-0.019317124,0.1067793369,0.1134361103,0.3309997916,0.2169079781,0.1099371389,-0.2286787778,-0.2032390386,-0.2612337768,-0.0527685992,-0.1159261763,0.0216245968,0.3059684336,0.0276983511,-0.0514889993,-0.1710950732,-0.1610162109,0.2887646258,0.4294677675,0.0548807494,0.2797007561,-0.3239167035,-0.20926781,-0.2900282741,0.4059990942,0.0259999819,-0.0143817831,-0.3674530387,0.0639025122,0.2022509277,-0.0671821758,0.5411589146,-0.0151449339,0.1318528205,-0.1551022381,-0.1166047379,-0.0568027459,0.0321312174,0.0681479797,0.1473077983,0.0709411502,0.2918481827,-0.1973031312,-0.2109269202,0.1356880665,0.0536990836,-0.1384829134,-0.2008591294,-0.3638655543,-0.1265242547,-0.4643529654,-0.0363947749,0.0795841813,0.1581552029,-0.0118486276,0.0888735875,-0.0055110841,0.0582237169,0.1181041971,0.2581231296,0.2853506207,0.061050076,-0.0329073332,0.4962373376,0.2412015349,0.2928660512,0.903095603,0.0374446474,-0.5974399447,-0.0672150329,0.1330420822,0.0157623217,0.2243538499,-0.0302627385,-0.1523289531,0.3946187496,0.1091537923,-0.3597810566,0.0396902412,0.3105780184,0.0445691869,-0.1691173166,-0.1938370466,0.4223771691,-0.0063675665,0.0590270795,0.1012548357,0.4643971026,-0.100590758,0.222247079,-0.244068861,0.9018837214,-0.1857047677,-0.0163326282,0.0905331895,0.0189460609,0.295566082,-0.2095983624,0.0179672185,-0.1722640991,-0.0136424126,-0.06711483,-0.0457918346,0.2570757866,0.1821769774,-0.0685173944,0.2682501376,0.0485039949,0.0206053406,-0.0939894468,0.5839898586,-0.2126607299,-0.0314419046,-0.279080838,0.2627642155,0.066398792,0.8657299876,-0.1349084973,-0.2178676873,-0.167290926,-0.1910547167,-0.2989360392,0.0837852359,-0.1352346689,0.256293118,-0.0097763352,-0.1183076128,0.151492998,0.321002394,0.1314985752,0.0574066676,-0.5936903954,0.142104283,0.1098857373,-0.4285562932,-0.1501795799,0.0605323985,0.3860791624,-0.1960684359,-0.2655944824,-0.0198910702,0.0448563658,-0.0204233564,-0.24806045,-0.1849880964,-0.051631324,0.0376829095,-0.0638323426,-0.0995162651,-0.0327224359,-0.0414391868,0.1771371663,0.2839221656,-0.2054964751,-0.2806945145,0.2618798614,-0.1227718443,-0.0083155185,0.5535974503,-0.2329237163,-0.235673666,0.4879062772,0.1241211519,-0.0839796811,-0.2452081144,-0.1700535268,0.2742600739,-0.7010433674,-0.1427640468,0.4160333574,0.0046225283,0.2194000334,0.2663269341,0.2969088554,-0.5169305205,0.0058192816,-0.4191864431,-0.1297298521,0.1292425841,-0.1462572068,0.0867163092,0.1373927891,0.2032428682,0.1536358893,0.0081006959,-0.3894360065,0.1659612805,-0.2964887917,-0.1010633484,0.2427747548,0.0080073848,0.2464223206,0.0020827043,0.2410346866,0.049094934,-0.3328877389,-0.2498598695,-0.0370018966,0.1327316612,0.0028105178,-0.1149384901,-0.0287480298,-0.1482413858,-0.1678066254,0.0718372092,0.136245057,0.1104840189,0.0671192706,0.1133614853,0.0027967973,-0.0251711477,0.2837280035,0.0395055227,0.012746837,-0.0555565581,0.086874187,-0.0207448639,-0.090487659,0.012700608,-0.0977362543,0.0426986478,0.1011068895,0.0205859635,0.3517746925,-0.2655716538,0.1165345684,0.0739995688,0.185393393,0.1270457655,-0.2866701186,0.2289478481,0.4426570237,0.2985438704,-0.47711128,-0.1298857927,-0.0006360873,-0.1966404468,-0.0282347724,-0.013260711,0.3188249171,0.283752799,-0.076736182,0.1066010669,0.3333195448,-0.2685655355,0.1106194332,0.1932328343,-0.2465140671,-0.0085457573,0.01131998,0.2905161381,0.1318642497,0.3088471591,-0.2265245616,-0.0315658376,-0.4782330096,0.5385349393,0.2666293085,-0.3581576943,-0.090971753,0.0968580991,0.071271956,0.1364169419,-0.1544249505,0.7230104208,-0.3374552429,0.0211152993,-0.3394336104,0.2355688661,-0.0821216255,-0.0698981956,-0.1313723475,-0.2217419297,-0.0670109242,-0.1869946867,0.0176493917,-0.2153773308,-0.1389422715,0.0830589309,-0.0692196563,-0.3543566763,-0.2979280353,0.2269875109,0.2037164718,-0.131247431,0.2813130617,0.1106147468,-0.0874739885,0.2627348304,0.3017261326,0.5873758197,0.0351856053,0.0740019679,0.0617890432,-0.2088027298,-0.0921293646,-0.109564431,0.3769292831,0.2304144949,0.1554087549,0.2603459954,0.1857828945,-0.1731237918,-0.0711420923,-0.2792942226,0.0958354846,-0.0564152114,0.4539793432,-0.452164799,0.062757656,-0.3875875473,0.0286676344,-0.522289753,-0.0055026873,0.1947346479,0.1312919855,0.0624886528,-0.2410089225,0.1104958653,0.0426301621,0.4625753164,0.570676446,0.2936342657,-0.1461770236,-0.0949215665,-0.6033201814,0.1542592198,0.0418437533,0.0344360806,-0.0132869855,-0.0234472863,0.0551848821,-0.0365119278,-0.0200043693,-0.2295777947,0.1678578854,0.0440307409,-0.014884253,0.149932757,-0.1426046491,0.1320004016,0.0895937905,-0.0728232265,0.0851186067,-0.1527349651,0.0326881222,-0.2568894923,0.3353639543,-0.3484842181,-0.2517906129,0.2978453338,0.1351178735,0.3951489627,-0.0444623679,-0.034926299,-0.2083920091,-0.1573612839,-0.2170385271,0.3865812719,0.0789019987,0.2669255435,-0.1007362679,-0.4330461025,-0.2646440864,0.0012700515,-0.0981018171,-0.0115452521,-0.3078569472,-0.0789810568,-0.0546493195,0.0014905818,0.1194022447,0.1274342984,0.0348049887,-0.1514230222,-0.1979456544,-0.1850954741,0.8110972643,-0.3621260524,0.0602045059,0.1660539657,0.1678345799,0.148713842,-0.3176252842,-0.6711137295,0.2956936955,0.2326421887,0.1906607151,-0.068932794,0.2591880262,-0.1669406593,-0.0353300199,-0.0263389312,0.1093059182,0.0340025388,-0.1503763199,0.1687027067,-0.0347632207]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1290","title":"imdb dataset cannot be downloaded","comments":"Hi @rabeehk , I am unable to reproduce your problem locally.\r\nCan you try emptying the cache (removing the content of `\/idiap\/temp\/rkarimi\/cache_home_1\/datasets`) and retry ?","body":"hi\r\nplease find error below getting imdb train spli:\r\nthanks\r\n\r\n`\r\ndatasets.load_dataset>>> datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\n\r\nerrors\r\n\r\n\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown size, total: 207.28 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=7486451, num_examples=5628, dataset_name='imdb')}]\r\n\r\n\r\n```","comment_length":25,"text":"imdb dataset cannot be downloaded\nhi\r\nplease find error below getting imdb train spli:\r\nthanks\r\n\r\n`\r\ndatasets.load_dataset>>> datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\n\r\nerrors\r\n\r\n\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown size, total: 207.28 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=7486451, num_examples=5628, dataset_name='imdb')}]\r\n\r\n\r\n```\nHi @rabeehk , I am unable to reproduce your problem locally.\r\nCan you try emptying the cache (removing the content of `\/idiap\/temp\/rkarimi\/cache_home_1\/datasets`) and retry ?","embeddings":[-0.4860101342,0.0299127121,-0.1837317646,0.2636173666,0.3486504555,0.3077562451,0.3157056868,0.4439661801,0.2162813097,0.0212673135,-0.0990766734,-0.0724211112,-0.0315420739,0.1580998898,-0.0292219277,-0.1209031865,-0.1767851561,0.0127698407,-0.1587827206,0.0179681294,-0.1283203512,0.238350302,-0.1618407369,-0.0335202925,-0.4301397502,-0.3121137321,0.139201194,0.0827983245,-0.3193941116,-0.4591375291,0.392133534,-0.0082366345,0.2423053384,0.2826945782,-0.0001035062,-0.036567647,0.5334970951,-0.1074083596,-0.3835625947,-0.2877946198,-0.327385217,0.0555791669,-0.002764019,-0.2751792669,0.0047641462,-0.0253852848,0.0113665806,-0.3959704638,0.037413571,0.4941405356,0.2971445024,-0.1297646463,0.2451955676,-0.2180922627,0.2423908561,0.0137036452,0.0529719256,0.2488452196,0.140880987,0.2723523378,0.2486406863,0.1801100224,-0.3034388125,-0.0220966898,0.2865758538,-0.0972823352,0.1653281152,-0.4706068635,0.1722737402,0.2280539721,0.7382758856,-0.2388616651,-0.3810375333,0.1447506249,0.3771938384,-0.1702625602,0.100001283,0.2182770073,-0.1827754229,0.101738967,-0.2199486643,-0.0662822351,-0.1369595826,0.1590718776,0.0886127204,-0.1605072618,0.0495887101,-0.0400308669,0.0987498984,-0.1344026029,0.2517719865,0.0384320244,-0.0021123588,0.236170128,-0.260240078,-0.1002123132,-0.2237588018,0.0192964952,0.0649492294,0.2488008887,-0.1534933597,-0.1331486851,0.016815085,0.1392866224,0.182145372,0.1841432154,-0.058251448,0.3144685328,0.1950910091,0.4538915157,-0.1490833759,-0.1017040089,-0.2541640997,-0.1346600354,-0.0824786797,0.143608585,0.0311763398,-0.1240337119,-0.4782600105,0.0448355265,-0.2012896836,0.1291317046,0.2305186391,0.3140892088,-0.1347789168,0.1046995968,0.2271656245,0.252812922,0.1159728169,-0.1042450368,-0.246147573,0.3043960631,-0.0936458483,-0.0565775447,0.3721653223,-0.1996798813,0.3377612829,0.0696253031,0.0036384759,-0.2105993181,0.2095436156,-0.3007852435,0.0287867971,0.3584664762,0.2556399703,0.2001560777,-0.0840682834,0.1235398501,-0.0588877834,0.0461486988,-0.2743842006,-0.319922477,-0.0881592631,0.3507601321,0.1262279451,-0.1656855792,-0.3235688508,-0.1998392344,0.09139698,-0.2130101174,0.0047630412,-0.1729651242,-0.025870176,-0.133760795,0.3516489267,0.4913713336,-0.2694042921,0.0458030626,-0.4400700331,-0.1443544328,0.0440970734,0.3456625044,-0.1699273437,0.3116580248,-0.32995224,0.1360130161,0.2640047073,-0.6436082721,-0.5091253519,0.325425595,-0.0387009382,-0.11843279,0.2331247181,0.2476876527,0.3442249298,0.0431956872,0.1554581672,0.3124530315,-0.153066501,-0.0956452191,-0.1574309617,-0.2806973457,-0.0521216802,0.1857506186,0.3627323508,0.1378489584,0.114368625,0.2967046201,0.2887882888,0.1775026917,0.1041831821,0.3223493993,0.2164294571,0.1731072664,-0.1532247961,-0.1303550899,-0.2971837521,0.080735229,0.2190829068,-0.0312121324,-0.2718612254,-0.2509310544,-0.3407040238,-0.2066206634,-0.1483213454,0.0180350691,0.2382639647,-0.0485067032,-0.0493651181,-0.0633319318,-0.2919754088,0.4999574423,-0.2259841561,0.0901971832,-0.292301327,0.1588745862,-0.2442293465,-0.1025532484,0.124649033,-0.0284797642,0.1251573414,-0.0823221952,-0.0637026131,0.3320347369,0.0375415012,0.0506807417,0.0673423484,-0.0951315463,0.1241538003,-0.3400791883,0.1844294965,-0.2295352072,0.0710089952,-0.0017823211,-0.0044872374,-0.0572311543,-0.2801706791,0.0628418326,0.2860819995,0.0964290351,0.298142463,-0.1603615433,0.0415811948,-0.4865525067,0.0784966424,0.0981659293,0.2395064682,-0.0675733835,-0.2537468672,0.0527260862,0.3675588369,0.1426952928,0.0361450277,0.1218582988,-0.0939497575,0.0451115519,-0.1279809475,0.2976512313,0.1854215115,0.1711168885,0.0062498222,-0.1330317408,0.228055194,-0.2165435851,0.2388888299,-0.0532691255,0.0778323933,0.3355535567,-0.092541486,0.0923875868,-0.1577505767,-0.3601544201,0.085387975,0.2455344051,-0.2284465432,-0.2190803885,-0.3301956058,-0.2610275745,-0.0584802814,-0.1618005931,0.0648657903,-0.173955366,-0.1136664376,0.3175649643,0.252427876,0.0781725198,-0.287586838,-0.0884352028,0.0608519055,-0.2129155397,-0.1243227869,-0.1231795996,-0.0458455533,0.2266675681,0.3349539936,-0.021135442,0.2578156292,-0.4402718544,-0.0080837542,-0.2770047784,0.0826189965,0.0398202352,0.0108164065,0.095336467,0.124990724,0.4773280323,0.1202947125,-0.3079587817,0.0788394511,-0.0808625892,-0.1750845313,-0.0007010662,-0.0925777853,-0.0714502633,0.0025721136,-0.4792709947,-0.4157936573,-0.316506058,-0.0713702664,0.0571880192,0.0436627604,0.0977312848,0.2333915681,0.2032420188,0.3070598245,0.1822692305,-0.0370333716,-0.3868349493,0.302156955,-0.2015400529,-0.2068994343,0.2725228369,-0.0110130245,0.321393162,0.0716894716,-0.6101097465,0.1016692817,-0.0240966566,0.3465483487,-0.098723419,-0.109618932,0.0603595115,-0.065568395,-0.1339580268,-0.0619546995,-0.0375211053,-0.0004061266,-0.0901715308,0.1974919438,-0.2317510247,0.5183964968,-0.0380557179,0.3936097324,0.327231586,-0.1218411848,0.4882543683,-0.0589160919,0.2292048335,-0.2034323961,-0.4554421902,0.0337610655,-0.0116053149,-0.1714819819,-0.1340015084,-0.139235422,-0.2006169409,-0.0880587623,0.0487595052,-0.3500264287,-0.0818259716,-0.0632227436,-0.1893380731,0.1761308014,-0.0381598659,0.1182602495,-0.0236347299,-0.2639210224,0.1144766435,0.3832407296,0.0313140601,0.1572652757,0.0041668452,-0.2751965225,-0.1234828457,0.1956590563,0.0952752233,0.2581464946,-0.2238431424,0.2099879086,0.1383150518,0.1370850354,0.6306633949,-0.3738168776,0.1606445312,0.1012409031,0.1247325316,-0.085095644,-0.0353820398,-0.1879751831,0.0969590023,0.4624360204,0.313814491,-0.2233282328,-0.0177924689,0.446682781,0.1704700738,0.0012421121,-0.1044087335,-0.0297605917,-0.2990216613,-0.3173526824,-0.0554821603,0.0616683662,0.1669998318,-0.0458416566,-0.1391413212,-0.0763726979,0.058955241,-0.0352792367,0.0062580821,0.4235682786,0.1951737702,0.2749030888,0.1604067087,-0.1231250763,0.7554829121,0.5446822643,-0.3073401451,0.0237262882,0.0381548181,-0.108909443,-0.0058781877,0.1639737785,-0.1571988016,-0.3096763492,0.1150087416,-0.0502805226,0.1761645973,0.0670876428,0.4062772095,-0.0849895701,-0.2734709382,-0.6353532076,0.2006477416,0.2205908746,0.075873144,0.1138357893,0.001441384,-0.2162543237,0.1600407809,0.1431051046,0.6717502475,-0.0306913126,-0.1283620447,-0.0737282336,0.2314275205,0.1600268483,-0.3901946247,0.2138733417,-0.248378545,-0.2130271792,-0.0515333414,-0.1116768494,-0.1148253009,0.3104119003,-0.3014479876,0.2266770899,-0.31661731,-0.1173048168,-0.1931751966,0.5951215625,-0.4062134624,0.0443438888,0.0933396295,0.2500638068,-0.0803388208,0.2330947965,0.0176276974,-0.005986582,-0.1392312646,-0.0871677548,-0.5714972615,0.0955903009,-0.3902885318,0.4118309021,-0.2546887398,-0.1696923226,0.0545233265,0.0963137001,0.1940549612,0.140099436,-0.0896286741,0.1585943848,0.3080503345,-0.319887042,0.1317542046,0.1286303103,0.5172037482,-0.1429077536,-0.1903925538,0.1281249076,0.0822041407,-0.0944063291,0.2196153998,-0.113035731,0.1230945066,-0.2164566219,-0.3051655293,-0.1710217595,-0.0309378468,-0.1239987388,0.1908607483,-0.1052556783,0.0750182271,0.1135814488,0.2069985718,-0.2609417737,-0.0699764043,0.5473296046,0.0353881046,-0.0399466455,0.4528162181,0.1376165003,-0.2352596819,-0.2972407341,0.0628512949,-0.0171452072,-0.2241660506,0.1007652357,-0.0658697411,0.2046489269,-0.033007741,-0.0393612199,0.2021432966,0.07974758,-0.1549257338,-0.5466320515,-0.3308158815,0.2572823763,-0.0040129968,0.1021018103,-0.226144895,-0.004490938,0.1607132107,-0.2122690231,-0.3538897038,0.0809048787,-0.2972688079,0.0604099818,0.3226789534,-0.0839445516,0.1275472492,-0.0507345349,0.260458082,-0.0283886697,-0.0389637761,-0.2842151523,0.0097504398,0.0598692596,-0.0420603193,-0.092815429,-0.1324003637,-0.3429490328,-0.339350462,0.0148780914,-0.0271453448,0.1792007238,0.1183891669,0.137585327,-0.1262245774,-0.0520076193,-0.1079084054,0.1508829892,-0.0076595279,0.0836496726,0.0269258469,0.173334837,0.16698125,-0.0562452041,-0.1608996987,-0.0270904601,-0.1440099031,0.0731736124,0.314510107,-0.2280226648,0.1771394461,0.2457535714,0.0589395389,0.4200526178,-0.1652677208,-0.1035108417,-0.0913868919,0.3593567312,-0.2875733376,0.0594157763,0.1363564134,0.0103493119,-0.0321375206,0.0257978179,0.092798546,-0.0330738015,0.1518532932,0.128941223,0.5847817659,-0.0939864144,0.2283103317,0.1213584766,0.0341578908,-0.0537566096,0.0660596639,0.2479736507,0.1333173215,0.6074611545,-0.3323966265,0.021696534,-0.0137874614,0.0380411521,0.0430706888,-0.3360207677,0.0479255281,0.0449727252,-0.0241372604,-0.1163440794,-0.0760938078,0.1558948457,-0.0984017849,-0.0509514883,-0.2309980541,0.0863991827,-0.2810724378,0.1381596178,-0.2467105836,-0.1639960855,-0.2269250453,0.0585490614,-0.0784547403,0.0134896468,0.0897028968,0.3838812411,-0.2677740157,-0.338809371,0.1131344512,0.1015294641,0.2099556029,-0.3549320698,0.2294833362,0.1947603375,0.0789522305,0.171492815,0.2270049751,0.4746427536,0.4007495344,-0.0522254892,0.0508304685,0.0979081392,-0.2286951989,-0.0948867872,0.1375479102,0.3165980577,0.199182108,0.1111147478,0.2761748433,-0.2056592256,0.0442580581,0.0271602068,0.3460959494,-0.0298096035,0.1888675988,-0.1929424405,-0.1620156914,-0.3676898777,0.0509318449,-0.3131984174,-0.1232887357,0.3973481059,-0.2198224068,-0.0307696238,-0.1220248342,0.114908129,-0.1704159677,0.3895864487,0.407718569,0.022323817,-0.4315789938,-0.0374502204,-0.5283246636,-0.0234989859,-0.1055567712,0.0213449188,0.0183603335,0.3473741412,-0.3096303642,0.0190831069,0.3375191391,0.1437054873,0.0124574797,0.3879181743,-0.1668470502,-0.0650721863,-0.1635156274,-0.0045003239,-0.0164063927,-0.2656364739,0.0661236122,-0.5669974685,0.1465960294,-0.0391405337,-0.374306947,0.2084971517,0.364037931,0.4081876874,-0.1317987591,0.686552465,-0.1391124427,-0.1931458563,-0.2702291906,-0.2662594914,-0.2637103796,0.2612058818,0.2799525261,0.2722635865,-0.0940014422,-0.1878320426,-0.2483573407,0.3727858961,0.0062568532,-0.2396901846,-0.3305553198,-0.20875296,-0.1693937331,0.0842131227,0.1114842966,0.0353930667,0.0012898342,0.0618680194,-0.338870883,-0.4206544757,0.4592054188,-0.0953678638,0.0193864685,-0.1915466785,0.2458002567,0.2035212517,-0.0267859176,-0.4084336758,0.2136357129,0.3001485467,-0.1244565621,-0.1870332062,0.1222754568,0.0408127569,0.2679310143,-0.0382403135,-0.0211121086,0.0068067512,-0.189632833,-0.2512019277,-0.1015275717]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1290","title":"imdb dataset cannot be downloaded","comments":"Hi,\r\nthanks, I did remove the cache and still the same error here\r\n\r\n```\r\n>>> a = datasets.load_dataset(\"imdb\", split=\"train\")\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown size, total: 207.28 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=4902716, num_examples=3680, dataset_name='imdb')}]\r\n```\r\n\r\ndatasets version\r\n```\r\ndatasets 1.1.2 <pip>\r\ntensorflow-datasets 4.1.0 <pip>\r\n\r\n```","body":"hi\r\nplease find error below getting imdb train spli:\r\nthanks\r\n\r\n`\r\ndatasets.load_dataset>>> datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\n\r\nerrors\r\n\r\n\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown size, total: 207.28 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=7486451, num_examples=5628, dataset_name='imdb')}]\r\n\r\n\r\n```","comment_length":115,"text":"imdb dataset cannot be downloaded\nhi\r\nplease find error below getting imdb train spli:\r\nthanks\r\n\r\n`\r\ndatasets.load_dataset>>> datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\n\r\nerrors\r\n\r\n\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown size, total: 207.28 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=7486451, num_examples=5628, dataset_name='imdb')}]\r\n\r\n\r\n```\nHi,\r\nthanks, I did remove the cache and still the same error here\r\n\r\n```\r\n>>> a = datasets.load_dataset(\"imdb\", split=\"train\")\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown size, total: 207.28 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=4902716, num_examples=3680, dataset_name='imdb')}]\r\n```\r\n\r\ndatasets version\r\n```\r\ndatasets 1.1.2 <pip>\r\ntensorflow-datasets 4.1.0 <pip>\r\n\r\n```","embeddings":[-0.4860101342,0.0299127121,-0.1837317646,0.2636173666,0.3486504555,0.3077562451,0.3157056868,0.4439661801,0.2162813097,0.0212673135,-0.0990766734,-0.0724211112,-0.0315420739,0.1580998898,-0.0292219277,-0.1209031865,-0.1767851561,0.0127698407,-0.1587827206,0.0179681294,-0.1283203512,0.238350302,-0.1618407369,-0.0335202925,-0.4301397502,-0.3121137321,0.139201194,0.0827983245,-0.3193941116,-0.4591375291,0.392133534,-0.0082366345,0.2423053384,0.2826945782,-0.0001035062,-0.036567647,0.5334970951,-0.1074083596,-0.3835625947,-0.2877946198,-0.327385217,0.0555791669,-0.002764019,-0.2751792669,0.0047641462,-0.0253852848,0.0113665806,-0.3959704638,0.037413571,0.4941405356,0.2971445024,-0.1297646463,0.2451955676,-0.2180922627,0.2423908561,0.0137036452,0.0529719256,0.2488452196,0.140880987,0.2723523378,0.2486406863,0.1801100224,-0.3034388125,-0.0220966898,0.2865758538,-0.0972823352,0.1653281152,-0.4706068635,0.1722737402,0.2280539721,0.7382758856,-0.2388616651,-0.3810375333,0.1447506249,0.3771938384,-0.1702625602,0.100001283,0.2182770073,-0.1827754229,0.101738967,-0.2199486643,-0.0662822351,-0.1369595826,0.1590718776,0.0886127204,-0.1605072618,0.0495887101,-0.0400308669,0.0987498984,-0.1344026029,0.2517719865,0.0384320244,-0.0021123588,0.236170128,-0.260240078,-0.1002123132,-0.2237588018,0.0192964952,0.0649492294,0.2488008887,-0.1534933597,-0.1331486851,0.016815085,0.1392866224,0.182145372,0.1841432154,-0.058251448,0.3144685328,0.1950910091,0.4538915157,-0.1490833759,-0.1017040089,-0.2541640997,-0.1346600354,-0.0824786797,0.143608585,0.0311763398,-0.1240337119,-0.4782600105,0.0448355265,-0.2012896836,0.1291317046,0.2305186391,0.3140892088,-0.1347789168,0.1046995968,0.2271656245,0.252812922,0.1159728169,-0.1042450368,-0.246147573,0.3043960631,-0.0936458483,-0.0565775447,0.3721653223,-0.1996798813,0.3377612829,0.0696253031,0.0036384759,-0.2105993181,0.2095436156,-0.3007852435,0.0287867971,0.3584664762,0.2556399703,0.2001560777,-0.0840682834,0.1235398501,-0.0588877834,0.0461486988,-0.2743842006,-0.319922477,-0.0881592631,0.3507601321,0.1262279451,-0.1656855792,-0.3235688508,-0.1998392344,0.09139698,-0.2130101174,0.0047630412,-0.1729651242,-0.025870176,-0.133760795,0.3516489267,0.4913713336,-0.2694042921,0.0458030626,-0.4400700331,-0.1443544328,0.0440970734,0.3456625044,-0.1699273437,0.3116580248,-0.32995224,0.1360130161,0.2640047073,-0.6436082721,-0.5091253519,0.325425595,-0.0387009382,-0.11843279,0.2331247181,0.2476876527,0.3442249298,0.0431956872,0.1554581672,0.3124530315,-0.153066501,-0.0956452191,-0.1574309617,-0.2806973457,-0.0521216802,0.1857506186,0.3627323508,0.1378489584,0.114368625,0.2967046201,0.2887882888,0.1775026917,0.1041831821,0.3223493993,0.2164294571,0.1731072664,-0.1532247961,-0.1303550899,-0.2971837521,0.080735229,0.2190829068,-0.0312121324,-0.2718612254,-0.2509310544,-0.3407040238,-0.2066206634,-0.1483213454,0.0180350691,0.2382639647,-0.0485067032,-0.0493651181,-0.0633319318,-0.2919754088,0.4999574423,-0.2259841561,0.0901971832,-0.292301327,0.1588745862,-0.2442293465,-0.1025532484,0.124649033,-0.0284797642,0.1251573414,-0.0823221952,-0.0637026131,0.3320347369,0.0375415012,0.0506807417,0.0673423484,-0.0951315463,0.1241538003,-0.3400791883,0.1844294965,-0.2295352072,0.0710089952,-0.0017823211,-0.0044872374,-0.0572311543,-0.2801706791,0.0628418326,0.2860819995,0.0964290351,0.298142463,-0.1603615433,0.0415811948,-0.4865525067,0.0784966424,0.0981659293,0.2395064682,-0.0675733835,-0.2537468672,0.0527260862,0.3675588369,0.1426952928,0.0361450277,0.1218582988,-0.0939497575,0.0451115519,-0.1279809475,0.2976512313,0.1854215115,0.1711168885,0.0062498222,-0.1330317408,0.228055194,-0.2165435851,0.2388888299,-0.0532691255,0.0778323933,0.3355535567,-0.092541486,0.0923875868,-0.1577505767,-0.3601544201,0.085387975,0.2455344051,-0.2284465432,-0.2190803885,-0.3301956058,-0.2610275745,-0.0584802814,-0.1618005931,0.0648657903,-0.173955366,-0.1136664376,0.3175649643,0.252427876,0.0781725198,-0.287586838,-0.0884352028,0.0608519055,-0.2129155397,-0.1243227869,-0.1231795996,-0.0458455533,0.2266675681,0.3349539936,-0.021135442,0.2578156292,-0.4402718544,-0.0080837542,-0.2770047784,0.0826189965,0.0398202352,0.0108164065,0.095336467,0.124990724,0.4773280323,0.1202947125,-0.3079587817,0.0788394511,-0.0808625892,-0.1750845313,-0.0007010662,-0.0925777853,-0.0714502633,0.0025721136,-0.4792709947,-0.4157936573,-0.316506058,-0.0713702664,0.0571880192,0.0436627604,0.0977312848,0.2333915681,0.2032420188,0.3070598245,0.1822692305,-0.0370333716,-0.3868349493,0.302156955,-0.2015400529,-0.2068994343,0.2725228369,-0.0110130245,0.321393162,0.0716894716,-0.6101097465,0.1016692817,-0.0240966566,0.3465483487,-0.098723419,-0.109618932,0.0603595115,-0.065568395,-0.1339580268,-0.0619546995,-0.0375211053,-0.0004061266,-0.0901715308,0.1974919438,-0.2317510247,0.5183964968,-0.0380557179,0.3936097324,0.327231586,-0.1218411848,0.4882543683,-0.0589160919,0.2292048335,-0.2034323961,-0.4554421902,0.0337610655,-0.0116053149,-0.1714819819,-0.1340015084,-0.139235422,-0.2006169409,-0.0880587623,0.0487595052,-0.3500264287,-0.0818259716,-0.0632227436,-0.1893380731,0.1761308014,-0.0381598659,0.1182602495,-0.0236347299,-0.2639210224,0.1144766435,0.3832407296,0.0313140601,0.1572652757,0.0041668452,-0.2751965225,-0.1234828457,0.1956590563,0.0952752233,0.2581464946,-0.2238431424,0.2099879086,0.1383150518,0.1370850354,0.6306633949,-0.3738168776,0.1606445312,0.1012409031,0.1247325316,-0.085095644,-0.0353820398,-0.1879751831,0.0969590023,0.4624360204,0.313814491,-0.2233282328,-0.0177924689,0.446682781,0.1704700738,0.0012421121,-0.1044087335,-0.0297605917,-0.2990216613,-0.3173526824,-0.0554821603,0.0616683662,0.1669998318,-0.0458416566,-0.1391413212,-0.0763726979,0.058955241,-0.0352792367,0.0062580821,0.4235682786,0.1951737702,0.2749030888,0.1604067087,-0.1231250763,0.7554829121,0.5446822643,-0.3073401451,0.0237262882,0.0381548181,-0.108909443,-0.0058781877,0.1639737785,-0.1571988016,-0.3096763492,0.1150087416,-0.0502805226,0.1761645973,0.0670876428,0.4062772095,-0.0849895701,-0.2734709382,-0.6353532076,0.2006477416,0.2205908746,0.075873144,0.1138357893,0.001441384,-0.2162543237,0.1600407809,0.1431051046,0.6717502475,-0.0306913126,-0.1283620447,-0.0737282336,0.2314275205,0.1600268483,-0.3901946247,0.2138733417,-0.248378545,-0.2130271792,-0.0515333414,-0.1116768494,-0.1148253009,0.3104119003,-0.3014479876,0.2266770899,-0.31661731,-0.1173048168,-0.1931751966,0.5951215625,-0.4062134624,0.0443438888,0.0933396295,0.2500638068,-0.0803388208,0.2330947965,0.0176276974,-0.005986582,-0.1392312646,-0.0871677548,-0.5714972615,0.0955903009,-0.3902885318,0.4118309021,-0.2546887398,-0.1696923226,0.0545233265,0.0963137001,0.1940549612,0.140099436,-0.0896286741,0.1585943848,0.3080503345,-0.319887042,0.1317542046,0.1286303103,0.5172037482,-0.1429077536,-0.1903925538,0.1281249076,0.0822041407,-0.0944063291,0.2196153998,-0.113035731,0.1230945066,-0.2164566219,-0.3051655293,-0.1710217595,-0.0309378468,-0.1239987388,0.1908607483,-0.1052556783,0.0750182271,0.1135814488,0.2069985718,-0.2609417737,-0.0699764043,0.5473296046,0.0353881046,-0.0399466455,0.4528162181,0.1376165003,-0.2352596819,-0.2972407341,0.0628512949,-0.0171452072,-0.2241660506,0.1007652357,-0.0658697411,0.2046489269,-0.033007741,-0.0393612199,0.2021432966,0.07974758,-0.1549257338,-0.5466320515,-0.3308158815,0.2572823763,-0.0040129968,0.1021018103,-0.226144895,-0.004490938,0.1607132107,-0.2122690231,-0.3538897038,0.0809048787,-0.2972688079,0.0604099818,0.3226789534,-0.0839445516,0.1275472492,-0.0507345349,0.260458082,-0.0283886697,-0.0389637761,-0.2842151523,0.0097504398,0.0598692596,-0.0420603193,-0.092815429,-0.1324003637,-0.3429490328,-0.339350462,0.0148780914,-0.0271453448,0.1792007238,0.1183891669,0.137585327,-0.1262245774,-0.0520076193,-0.1079084054,0.1508829892,-0.0076595279,0.0836496726,0.0269258469,0.173334837,0.16698125,-0.0562452041,-0.1608996987,-0.0270904601,-0.1440099031,0.0731736124,0.314510107,-0.2280226648,0.1771394461,0.2457535714,0.0589395389,0.4200526178,-0.1652677208,-0.1035108417,-0.0913868919,0.3593567312,-0.2875733376,0.0594157763,0.1363564134,0.0103493119,-0.0321375206,0.0257978179,0.092798546,-0.0330738015,0.1518532932,0.128941223,0.5847817659,-0.0939864144,0.2283103317,0.1213584766,0.0341578908,-0.0537566096,0.0660596639,0.2479736507,0.1333173215,0.6074611545,-0.3323966265,0.021696534,-0.0137874614,0.0380411521,0.0430706888,-0.3360207677,0.0479255281,0.0449727252,-0.0241372604,-0.1163440794,-0.0760938078,0.1558948457,-0.0984017849,-0.0509514883,-0.2309980541,0.0863991827,-0.2810724378,0.1381596178,-0.2467105836,-0.1639960855,-0.2269250453,0.0585490614,-0.0784547403,0.0134896468,0.0897028968,0.3838812411,-0.2677740157,-0.338809371,0.1131344512,0.1015294641,0.2099556029,-0.3549320698,0.2294833362,0.1947603375,0.0789522305,0.171492815,0.2270049751,0.4746427536,0.4007495344,-0.0522254892,0.0508304685,0.0979081392,-0.2286951989,-0.0948867872,0.1375479102,0.3165980577,0.199182108,0.1111147478,0.2761748433,-0.2056592256,0.0442580581,0.0271602068,0.3460959494,-0.0298096035,0.1888675988,-0.1929424405,-0.1620156914,-0.3676898777,0.0509318449,-0.3131984174,-0.1232887357,0.3973481059,-0.2198224068,-0.0307696238,-0.1220248342,0.114908129,-0.1704159677,0.3895864487,0.407718569,0.022323817,-0.4315789938,-0.0374502204,-0.5283246636,-0.0234989859,-0.1055567712,0.0213449188,0.0183603335,0.3473741412,-0.3096303642,0.0190831069,0.3375191391,0.1437054873,0.0124574797,0.3879181743,-0.1668470502,-0.0650721863,-0.1635156274,-0.0045003239,-0.0164063927,-0.2656364739,0.0661236122,-0.5669974685,0.1465960294,-0.0391405337,-0.374306947,0.2084971517,0.364037931,0.4081876874,-0.1317987591,0.686552465,-0.1391124427,-0.1931458563,-0.2702291906,-0.2662594914,-0.2637103796,0.2612058818,0.2799525261,0.2722635865,-0.0940014422,-0.1878320426,-0.2483573407,0.3727858961,0.0062568532,-0.2396901846,-0.3305553198,-0.20875296,-0.1693937331,0.0842131227,0.1114842966,0.0353930667,0.0012898342,0.0618680194,-0.338870883,-0.4206544757,0.4592054188,-0.0953678638,0.0193864685,-0.1915466785,0.2458002567,0.2035212517,-0.0267859176,-0.4084336758,0.2136357129,0.3001485467,-0.1244565621,-0.1870332062,0.1222754568,0.0408127569,0.2679310143,-0.0382403135,-0.0211121086,0.0068067512,-0.189632833,-0.2512019277,-0.1015275717]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1290","title":"imdb dataset cannot be downloaded","comments":"resolved with moving to version 1.1.3","body":"hi\r\nplease find error below getting imdb train spli:\r\nthanks\r\n\r\n`\r\ndatasets.load_dataset>>> datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\n\r\nerrors\r\n\r\n\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown size, total: 207.28 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=7486451, num_examples=5628, dataset_name='imdb')}]\r\n\r\n\r\n```","comment_length":6,"text":"imdb dataset cannot be downloaded\nhi\r\nplease find error below getting imdb train spli:\r\nthanks\r\n\r\n`\r\ndatasets.load_dataset>>> datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\n\r\nerrors\r\n\r\n\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown size, total: 207.28 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/imdb\/plain_text\/1.0.0\/90099cb476936b753383ba2ae6ab2eae419b2e87f71cd5189cb9c8e5814d12a3...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=7486451, num_examples=5628, dataset_name='imdb')}]\r\n\r\n\r\n```\nresolved with moving to version 1.1.3","embeddings":[-0.4860101342,0.0299127121,-0.1837317646,0.2636173666,0.3486504555,0.3077562451,0.3157056868,0.4439661801,0.2162813097,0.0212673135,-0.0990766734,-0.0724211112,-0.0315420739,0.1580998898,-0.0292219277,-0.1209031865,-0.1767851561,0.0127698407,-0.1587827206,0.0179681294,-0.1283203512,0.238350302,-0.1618407369,-0.0335202925,-0.4301397502,-0.3121137321,0.139201194,0.0827983245,-0.3193941116,-0.4591375291,0.392133534,-0.0082366345,0.2423053384,0.2826945782,-0.0001035062,-0.036567647,0.5334970951,-0.1074083596,-0.3835625947,-0.2877946198,-0.327385217,0.0555791669,-0.002764019,-0.2751792669,0.0047641462,-0.0253852848,0.0113665806,-0.3959704638,0.037413571,0.4941405356,0.2971445024,-0.1297646463,0.2451955676,-0.2180922627,0.2423908561,0.0137036452,0.0529719256,0.2488452196,0.140880987,0.2723523378,0.2486406863,0.1801100224,-0.3034388125,-0.0220966898,0.2865758538,-0.0972823352,0.1653281152,-0.4706068635,0.1722737402,0.2280539721,0.7382758856,-0.2388616651,-0.3810375333,0.1447506249,0.3771938384,-0.1702625602,0.100001283,0.2182770073,-0.1827754229,0.101738967,-0.2199486643,-0.0662822351,-0.1369595826,0.1590718776,0.0886127204,-0.1605072618,0.0495887101,-0.0400308669,0.0987498984,-0.1344026029,0.2517719865,0.0384320244,-0.0021123588,0.236170128,-0.260240078,-0.1002123132,-0.2237588018,0.0192964952,0.0649492294,0.2488008887,-0.1534933597,-0.1331486851,0.016815085,0.1392866224,0.182145372,0.1841432154,-0.058251448,0.3144685328,0.1950910091,0.4538915157,-0.1490833759,-0.1017040089,-0.2541640997,-0.1346600354,-0.0824786797,0.143608585,0.0311763398,-0.1240337119,-0.4782600105,0.0448355265,-0.2012896836,0.1291317046,0.2305186391,0.3140892088,-0.1347789168,0.1046995968,0.2271656245,0.252812922,0.1159728169,-0.1042450368,-0.246147573,0.3043960631,-0.0936458483,-0.0565775447,0.3721653223,-0.1996798813,0.3377612829,0.0696253031,0.0036384759,-0.2105993181,0.2095436156,-0.3007852435,0.0287867971,0.3584664762,0.2556399703,0.2001560777,-0.0840682834,0.1235398501,-0.0588877834,0.0461486988,-0.2743842006,-0.319922477,-0.0881592631,0.3507601321,0.1262279451,-0.1656855792,-0.3235688508,-0.1998392344,0.09139698,-0.2130101174,0.0047630412,-0.1729651242,-0.025870176,-0.133760795,0.3516489267,0.4913713336,-0.2694042921,0.0458030626,-0.4400700331,-0.1443544328,0.0440970734,0.3456625044,-0.1699273437,0.3116580248,-0.32995224,0.1360130161,0.2640047073,-0.6436082721,-0.5091253519,0.325425595,-0.0387009382,-0.11843279,0.2331247181,0.2476876527,0.3442249298,0.0431956872,0.1554581672,0.3124530315,-0.153066501,-0.0956452191,-0.1574309617,-0.2806973457,-0.0521216802,0.1857506186,0.3627323508,0.1378489584,0.114368625,0.2967046201,0.2887882888,0.1775026917,0.1041831821,0.3223493993,0.2164294571,0.1731072664,-0.1532247961,-0.1303550899,-0.2971837521,0.080735229,0.2190829068,-0.0312121324,-0.2718612254,-0.2509310544,-0.3407040238,-0.2066206634,-0.1483213454,0.0180350691,0.2382639647,-0.0485067032,-0.0493651181,-0.0633319318,-0.2919754088,0.4999574423,-0.2259841561,0.0901971832,-0.292301327,0.1588745862,-0.2442293465,-0.1025532484,0.124649033,-0.0284797642,0.1251573414,-0.0823221952,-0.0637026131,0.3320347369,0.0375415012,0.0506807417,0.0673423484,-0.0951315463,0.1241538003,-0.3400791883,0.1844294965,-0.2295352072,0.0710089952,-0.0017823211,-0.0044872374,-0.0572311543,-0.2801706791,0.0628418326,0.2860819995,0.0964290351,0.298142463,-0.1603615433,0.0415811948,-0.4865525067,0.0784966424,0.0981659293,0.2395064682,-0.0675733835,-0.2537468672,0.0527260862,0.3675588369,0.1426952928,0.0361450277,0.1218582988,-0.0939497575,0.0451115519,-0.1279809475,0.2976512313,0.1854215115,0.1711168885,0.0062498222,-0.1330317408,0.228055194,-0.2165435851,0.2388888299,-0.0532691255,0.0778323933,0.3355535567,-0.092541486,0.0923875868,-0.1577505767,-0.3601544201,0.085387975,0.2455344051,-0.2284465432,-0.2190803885,-0.3301956058,-0.2610275745,-0.0584802814,-0.1618005931,0.0648657903,-0.173955366,-0.1136664376,0.3175649643,0.252427876,0.0781725198,-0.287586838,-0.0884352028,0.0608519055,-0.2129155397,-0.1243227869,-0.1231795996,-0.0458455533,0.2266675681,0.3349539936,-0.021135442,0.2578156292,-0.4402718544,-0.0080837542,-0.2770047784,0.0826189965,0.0398202352,0.0108164065,0.095336467,0.124990724,0.4773280323,0.1202947125,-0.3079587817,0.0788394511,-0.0808625892,-0.1750845313,-0.0007010662,-0.0925777853,-0.0714502633,0.0025721136,-0.4792709947,-0.4157936573,-0.316506058,-0.0713702664,0.0571880192,0.0436627604,0.0977312848,0.2333915681,0.2032420188,0.3070598245,0.1822692305,-0.0370333716,-0.3868349493,0.302156955,-0.2015400529,-0.2068994343,0.2725228369,-0.0110130245,0.321393162,0.0716894716,-0.6101097465,0.1016692817,-0.0240966566,0.3465483487,-0.098723419,-0.109618932,0.0603595115,-0.065568395,-0.1339580268,-0.0619546995,-0.0375211053,-0.0004061266,-0.0901715308,0.1974919438,-0.2317510247,0.5183964968,-0.0380557179,0.3936097324,0.327231586,-0.1218411848,0.4882543683,-0.0589160919,0.2292048335,-0.2034323961,-0.4554421902,0.0337610655,-0.0116053149,-0.1714819819,-0.1340015084,-0.139235422,-0.2006169409,-0.0880587623,0.0487595052,-0.3500264287,-0.0818259716,-0.0632227436,-0.1893380731,0.1761308014,-0.0381598659,0.1182602495,-0.0236347299,-0.2639210224,0.1144766435,0.3832407296,0.0313140601,0.1572652757,0.0041668452,-0.2751965225,-0.1234828457,0.1956590563,0.0952752233,0.2581464946,-0.2238431424,0.2099879086,0.1383150518,0.1370850354,0.6306633949,-0.3738168776,0.1606445312,0.1012409031,0.1247325316,-0.085095644,-0.0353820398,-0.1879751831,0.0969590023,0.4624360204,0.313814491,-0.2233282328,-0.0177924689,0.446682781,0.1704700738,0.0012421121,-0.1044087335,-0.0297605917,-0.2990216613,-0.3173526824,-0.0554821603,0.0616683662,0.1669998318,-0.0458416566,-0.1391413212,-0.0763726979,0.058955241,-0.0352792367,0.0062580821,0.4235682786,0.1951737702,0.2749030888,0.1604067087,-0.1231250763,0.7554829121,0.5446822643,-0.3073401451,0.0237262882,0.0381548181,-0.108909443,-0.0058781877,0.1639737785,-0.1571988016,-0.3096763492,0.1150087416,-0.0502805226,0.1761645973,0.0670876428,0.4062772095,-0.0849895701,-0.2734709382,-0.6353532076,0.2006477416,0.2205908746,0.075873144,0.1138357893,0.001441384,-0.2162543237,0.1600407809,0.1431051046,0.6717502475,-0.0306913126,-0.1283620447,-0.0737282336,0.2314275205,0.1600268483,-0.3901946247,0.2138733417,-0.248378545,-0.2130271792,-0.0515333414,-0.1116768494,-0.1148253009,0.3104119003,-0.3014479876,0.2266770899,-0.31661731,-0.1173048168,-0.1931751966,0.5951215625,-0.4062134624,0.0443438888,0.0933396295,0.2500638068,-0.0803388208,0.2330947965,0.0176276974,-0.005986582,-0.1392312646,-0.0871677548,-0.5714972615,0.0955903009,-0.3902885318,0.4118309021,-0.2546887398,-0.1696923226,0.0545233265,0.0963137001,0.1940549612,0.140099436,-0.0896286741,0.1585943848,0.3080503345,-0.319887042,0.1317542046,0.1286303103,0.5172037482,-0.1429077536,-0.1903925538,0.1281249076,0.0822041407,-0.0944063291,0.2196153998,-0.113035731,0.1230945066,-0.2164566219,-0.3051655293,-0.1710217595,-0.0309378468,-0.1239987388,0.1908607483,-0.1052556783,0.0750182271,0.1135814488,0.2069985718,-0.2609417737,-0.0699764043,0.5473296046,0.0353881046,-0.0399466455,0.4528162181,0.1376165003,-0.2352596819,-0.2972407341,0.0628512949,-0.0171452072,-0.2241660506,0.1007652357,-0.0658697411,0.2046489269,-0.033007741,-0.0393612199,0.2021432966,0.07974758,-0.1549257338,-0.5466320515,-0.3308158815,0.2572823763,-0.0040129968,0.1021018103,-0.226144895,-0.004490938,0.1607132107,-0.2122690231,-0.3538897038,0.0809048787,-0.2972688079,0.0604099818,0.3226789534,-0.0839445516,0.1275472492,-0.0507345349,0.260458082,-0.0283886697,-0.0389637761,-0.2842151523,0.0097504398,0.0598692596,-0.0420603193,-0.092815429,-0.1324003637,-0.3429490328,-0.339350462,0.0148780914,-0.0271453448,0.1792007238,0.1183891669,0.137585327,-0.1262245774,-0.0520076193,-0.1079084054,0.1508829892,-0.0076595279,0.0836496726,0.0269258469,0.173334837,0.16698125,-0.0562452041,-0.1608996987,-0.0270904601,-0.1440099031,0.0731736124,0.314510107,-0.2280226648,0.1771394461,0.2457535714,0.0589395389,0.4200526178,-0.1652677208,-0.1035108417,-0.0913868919,0.3593567312,-0.2875733376,0.0594157763,0.1363564134,0.0103493119,-0.0321375206,0.0257978179,0.092798546,-0.0330738015,0.1518532932,0.128941223,0.5847817659,-0.0939864144,0.2283103317,0.1213584766,0.0341578908,-0.0537566096,0.0660596639,0.2479736507,0.1333173215,0.6074611545,-0.3323966265,0.021696534,-0.0137874614,0.0380411521,0.0430706888,-0.3360207677,0.0479255281,0.0449727252,-0.0241372604,-0.1163440794,-0.0760938078,0.1558948457,-0.0984017849,-0.0509514883,-0.2309980541,0.0863991827,-0.2810724378,0.1381596178,-0.2467105836,-0.1639960855,-0.2269250453,0.0585490614,-0.0784547403,0.0134896468,0.0897028968,0.3838812411,-0.2677740157,-0.338809371,0.1131344512,0.1015294641,0.2099556029,-0.3549320698,0.2294833362,0.1947603375,0.0789522305,0.171492815,0.2270049751,0.4746427536,0.4007495344,-0.0522254892,0.0508304685,0.0979081392,-0.2286951989,-0.0948867872,0.1375479102,0.3165980577,0.199182108,0.1111147478,0.2761748433,-0.2056592256,0.0442580581,0.0271602068,0.3460959494,-0.0298096035,0.1888675988,-0.1929424405,-0.1620156914,-0.3676898777,0.0509318449,-0.3131984174,-0.1232887357,0.3973481059,-0.2198224068,-0.0307696238,-0.1220248342,0.114908129,-0.1704159677,0.3895864487,0.407718569,0.022323817,-0.4315789938,-0.0374502204,-0.5283246636,-0.0234989859,-0.1055567712,0.0213449188,0.0183603335,0.3473741412,-0.3096303642,0.0190831069,0.3375191391,0.1437054873,0.0124574797,0.3879181743,-0.1668470502,-0.0650721863,-0.1635156274,-0.0045003239,-0.0164063927,-0.2656364739,0.0661236122,-0.5669974685,0.1465960294,-0.0391405337,-0.374306947,0.2084971517,0.364037931,0.4081876874,-0.1317987591,0.686552465,-0.1391124427,-0.1931458563,-0.2702291906,-0.2662594914,-0.2637103796,0.2612058818,0.2799525261,0.2722635865,-0.0940014422,-0.1878320426,-0.2483573407,0.3727858961,0.0062568532,-0.2396901846,-0.3305553198,-0.20875296,-0.1693937331,0.0842131227,0.1114842966,0.0353930667,0.0012898342,0.0618680194,-0.338870883,-0.4206544757,0.4592054188,-0.0953678638,0.0193864685,-0.1915466785,0.2458002567,0.2035212517,-0.0267859176,-0.4084336758,0.2136357129,0.3001485467,-0.1244565621,-0.1870332062,0.1222754568,0.0408127569,0.2679310143,-0.0382403135,-0.0211121086,0.0068067512,-0.189632833,-0.2512019277,-0.1015275717]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1287","title":"'iwslt2017-ro-nl', cannot be downloaded ","comments":"the same issue with datasets.load_dataset(\"iwslt2017\", 'iwslt2017-en-nl', split=split), ..... ","body":"Hi\r\nI am trying \r\n`>>> datasets.load_dataset(\"iwslt2017\", 'iwslt2017-ro-nl', split=\"train\")`\r\n\r\ngetting this error thank you for your help\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset iwsl_t217\/iwslt2017-ro-nl (download: 314.07 MiB, generated: 39.92 MiB, post-processed: Unknown size, total: 354.00 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/iwsl_t217\/iwslt2017-ro-nl\/1.0.0\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/iwslt2017\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd\/iwslt2017.py\", line 118, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(MULTI_URL)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n\r\n```","comment_length":8,"text":"'iwslt2017-ro-nl', cannot be downloaded \nHi\r\nI am trying \r\n`>>> datasets.load_dataset(\"iwslt2017\", 'iwslt2017-ro-nl', split=\"train\")`\r\n\r\ngetting this error thank you for your help\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset iwsl_t217\/iwslt2017-ro-nl (download: 314.07 MiB, generated: 39.92 MiB, post-processed: Unknown size, total: 354.00 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/iwsl_t217\/iwslt2017-ro-nl\/1.0.0\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/iwslt2017\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd\/iwslt2017.py\", line 118, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(MULTI_URL)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n\r\n```\nthe same issue with datasets.load_dataset(\"iwslt2017\", 'iwslt2017-en-nl', split=split), ..... ","embeddings":[-0.3419763446,-0.30060938,-0.1523725539,0.2706135213,0.2139090747,0.2157970816,0.2569400668,0.3650550246,0.1976396739,-0.0201298464,-0.0654523149,-0.1176212877,0.1440995634,-0.0291719362,-0.1518213302,-0.2699722052,-0.0663298965,0.1131053045,-0.1293237358,0.0422331616,-0.3337647617,0.2209876627,-0.1799138039,0.0285914224,-0.3148556948,-0.2158248723,-0.1557890624,0.0668197274,-0.376522094,-0.3967740536,0.4967109859,0.0440180227,0.3563047349,0.3447470367,-0.0001073461,0.0428802036,0.3648747802,-0.1117397919,-0.2904242277,-0.1810606569,-0.5819841027,-0.3666986525,-0.1755940318,-0.3403346837,0.0163967311,-0.4022048116,-0.1394961774,-0.1585931033,0.3906446397,0.388822943,0.2995847166,0.1232836023,0.2581681013,-0.2987955511,-0.0361348465,-0.283228755,-0.0201401301,0.2171174586,-0.0970925391,-0.020589469,0.2999322712,0.20792225,-0.1322476715,0.0845834091,0.0945169777,-0.137185201,0.2849365473,-0.4758993089,0.306342274,0.1419929713,0.6968966126,-0.1230078042,-0.2451383919,0.1330755949,0.0477134213,-0.1949062496,0.1256547123,0.3085877597,-0.1431127042,0.1953275204,-0.1524961442,-0.0715682507,-0.1325484961,0.2132592052,-0.1820759028,0.3298105896,-0.0642741919,0.1544097811,0.3366713822,-0.0729813129,0.2190582752,-0.0332605802,-0.1243837699,0.3861820102,-0.338873446,0.002449838,-0.2990521193,0.0185885374,-0.1256584078,0.0470617972,-0.2219414711,-0.2800481319,0.1643054038,0.1297833622,0.0439549685,0.2667294443,-0.0249498244,0.2557565868,-0.029544726,0.3510740101,0.0903158039,-0.128841728,-0.183530882,-0.3094523549,-0.1967954487,0.0998697355,0.063150391,-0.1231369227,-0.488530159,-0.1932276338,-0.0544629619,-0.0670500621,0.0719540119,0.4097508788,-0.0886905044,0.3924994767,0.163899228,0.2355045825,0.0368230082,-0.2544769943,-0.258556813,0.3158781826,-0.1857732832,-0.0204413049,0.2933999598,0.0800527781,0.101256527,0.0159880724,-0.2408084869,-0.1507217437,0.1572513729,-0.2279166579,0.0483161844,0.3992284238,0.1482194215,0.3353996873,0.0706556588,0.0414648466,-0.1198901832,0.1366238296,-0.2916611433,-0.2356133014,-0.0833062753,0.2989494205,-0.0838587806,-0.1165730953,0.0086517511,-0.0890580639,0.217648685,-0.0663128868,-0.1008108631,-0.0856411159,-0.2119235545,-0.3296175599,0.2840553522,0.3511959612,-0.3999688625,-0.1349375695,-0.2301706076,-0.2683898211,0.3191880584,0.2966541648,-0.1974266917,0.3559424579,-0.2710108459,0.1405541301,0.4296115339,-0.3251674473,-0.5949853063,0.3352683187,0.0314203426,-0.1941059083,0.2613495588,0.1419058591,0.0988269597,-0.0779613182,-0.0219251625,0.3322188854,-0.0165831987,-0.097443454,-0.1484275609,-0.4670132697,0.2685246468,0.2582086325,0.2376422286,0.1317265779,0.0829866901,0.2184657454,0.3316060901,0.0024225796,0.2145042121,0.1683985293,0.2144212276,-0.0120791206,-0.0896465406,-0.016023051,-0.291590035,0.2323364913,0.0352539532,-0.1178119481,-0.1290831864,0.0036250551,-0.4926204383,-0.1394286007,-0.4215623736,-0.0569159538,0.190910399,0.1853034496,0.0793977603,-0.1578582674,-0.077596359,0.5676696897,-0.3102144301,-0.0721075162,-0.2714800537,0.2507627308,-0.1448059827,-0.0085442523,0.0757114217,0.0322704986,0.318339318,-0.035067495,-0.1745737493,0.2032273263,-0.1742767543,0.1249080747,0.0113041038,-0.0183872506,0.0161141548,-0.4125339091,0.1873710155,-0.023024777,0.1227757186,-0.0164506584,0.043144878,0.0571190715,0.0457890294,0.137702629,0.0710880011,0.0011138548,0.3079363406,-0.3163119256,0.0914543346,-0.1924466044,0.1263335347,-0.0542972386,0.0092262914,-0.0536241569,0.1528143138,-0.0957075283,0.4249424636,0.0955039039,0.0690912232,0.0036228031,0.0028498142,0.1753265113,-0.2760487497,0.5185399055,0.3655532598,0.0751129761,0.0784046054,-0.1150381565,0.1288037151,-0.1084229127,0.2247704417,0.0476504453,-0.0041810251,0.2197386175,-0.0676042587,-0.0982570946,-0.0923172608,-0.0493154153,0.2656356096,0.3520200253,-0.1040189266,0.0060471906,-0.2463860065,-0.2953222394,-0.27870363,-0.1008472964,-0.0477192849,-0.1535633057,-0.2378375381,0.1771686673,0.2375490665,0.1192688122,-0.2360265255,0.1147160009,0.025813099,-0.0294521693,-0.1604108065,0.0218824688,-0.1970059723,0.2072577775,0.3102039993,0.0527941659,0.5148772597,-0.2965624928,-0.0328365453,-0.2607624233,0.0909985304,-0.0588743016,-0.0106225405,0.0126678767,0.1085725948,0.5615016222,0.1883694828,-0.2239586264,0.3116089106,-0.0666209757,-0.0488946363,-0.1512012929,-0.1495619714,-0.0486634113,-0.0859781131,-0.4663751423,-0.6394459605,-0.4285477102,-0.1054912657,-0.1299048066,0.0185630955,0.10352052,0.1360276043,-0.0806234851,0.1401029527,-0.1422324181,-0.1530722529,-0.1259031445,0.4883402884,-0.1384137869,-0.2780931592,0.2463521957,0.0517431833,0.3558724821,0.1222597435,-0.4612587392,0.2047154158,-0.1789457053,0.2339013368,-0.123334527,0.1875346601,0.1291775852,-0.12157882,-0.0701697245,-0.0898954198,-0.0077213193,-0.0767805576,0.0147693707,0.5862768888,0.0112557458,0.3250548244,-0.0636339188,0.5220999122,0.3522051573,-0.1157827452,0.5259754062,0.0313490592,0.0135138659,-0.0820038617,-0.633777976,0.0230399352,-0.1061978936,0.0868648589,0.2244164795,-0.1441634148,-0.2223575413,-0.1677296609,-0.0043811467,-0.148821041,-0.0474478975,0.0343091525,0.3830757439,0.4424400032,0.0338994041,-0.048039794,-0.2091472298,-0.1317834407,0.0974779278,0.2961211801,-0.1347294599,0.2479324192,-0.0165558532,-0.1964795738,-0.2588474154,0.1430327594,0.236054197,0.3298193514,-0.2033838779,0.1699988842,0.1397537142,0.0282663312,0.5790627003,-0.1956948638,0.1182057783,0.2313864678,0.2583263218,-0.0488119945,-0.1092301533,-0.1493160427,0.2338519841,0.3252852857,0.2348416597,-0.2585186064,-0.1194622889,0.2451254874,0.3395388126,-0.297731787,-0.3928254843,-0.318682313,-0.4542882144,-0.4696657658,-0.1589414328,0.0924784914,0.3104878962,-0.2088991851,0.065757513,-0.0496007204,-0.1515216529,0.0941471905,0.1550450921,0.3905794919,0.2684715092,0.1671625674,0.2203499526,0.1206488833,0.0937275141,0.5806069374,-0.0820148289,-0.1815625131,0.1683234572,-0.0675854161,-0.0086019738,0.1252853274,-0.1580293328,0.0009671754,0.0902222097,0.0446865596,0.1281659454,0.1415161639,0.3960202336,-0.1422026455,-0.569853425,-0.5394011736,0.4596917629,0.062678799,0.0551528931,0.2357880026,-0.0408788286,-0.1155170724,-0.1661672741,-0.1388940066,0.8912362456,-0.0221582334,-0.1548423767,0.1937688291,0.1800160259,0.3420315385,-0.2002798915,0.1621747464,-0.2236508131,0.0076990472,-0.0533528291,-0.1299173534,0.0873081163,0.2866203487,-0.2430064529,0.3370768726,-0.1529806703,0.0282465853,0.153639257,0.4686248302,-0.2369810492,-0.1234860495,-0.2134869695,0.2599333823,-0.0697904751,0.2494766861,-0.2205187231,-0.1804306507,0.0077849352,-0.1079011783,-0.2287700772,0.1813034266,-0.2994191647,0.59278965,-0.1998260468,0.00426909,0.2811477482,0.1610227078,0.3320013881,0.0170541592,-0.2426313609,0.2426300943,-0.0703173429,-0.2951053083,-0.2690463662,0.1414592713,0.1052307561,-0.1932798773,-0.3641048968,0.1524308175,-0.0982600376,-0.1537448168,0.1027298346,-0.0649383813,0.4251005054,-0.222037971,-0.1376624554,-0.1038458347,0.0651887953,-0.1359352618,0.1879301816,-0.0017168126,-0.1478768438,0.0615535192,-0.1109071523,-0.2153674513,-0.0525974594,0.4336761236,0.0625075027,0.0277480204,0.6179629564,0.0915395692,-0.1221234575,-0.3233804703,-0.0919291601,-0.0856461897,-0.4771237969,0.0789747238,0.113899909,0.0712306201,-0.185955584,0.4050688446,0.2536170483,0.0127548082,0.0016877064,-0.654889822,-0.3016872704,-0.0393077284,-0.021214975,0.159457311,-0.1859031767,0.0273677949,0.1794208437,-0.1460598707,-0.3506726027,0.0116716139,-0.220564127,0.0751525685,0.4576438963,-0.1315953881,0.1161369309,0.1173945889,0.2125164866,0.076565057,-0.1298950762,-0.3067094088,-0.1219613627,0.1114019677,-0.0835447311,-0.1904917657,-0.0903211758,-0.3495313227,-0.1241815314,0.1055315956,-0.1069640517,0.0205262285,0.0696955919,0.0964210704,0.1702435464,-0.022158891,-0.2776302397,0.2832111418,0.0862473249,0.1938600838,0.0527240932,0.290671438,0.1994037032,0.0168558266,-0.1140226573,-0.171184659,0.2523238659,-0.1742706001,0.0158356875,-0.3120677173,0.2289433181,0.182236135,0.0388923697,0.5452494621,-0.1949207187,0.1702246964,0.1145869568,0.2919920385,-0.3564121425,0.0131854489,0.1363899857,-0.012141061,-0.1282097399,0.0647349134,0.19133389,-0.1645380408,0.1727178991,0.0361656807,0.4443298876,-0.1137610525,0.0333962366,0.3205941021,0.1102307662,0.0073839156,0.1503521353,0.0869146287,0.1049035192,0.3699611127,-0.3519082963,0.334184587,-0.0292459577,0.1960637122,-0.0807112232,-0.463588804,-0.0132800359,-0.0271061286,-0.0280076005,-0.064129591,-0.0836857557,0.2307623029,-0.0578371771,-0.2075787187,-0.3942587376,0.2716932595,-0.1383715421,-0.0168737527,-0.2087452561,-0.1575703323,-0.3138845265,0.0200664438,0.0944762826,-0.2163674533,0.276887536,0.1739004105,-0.0674182326,-0.2864430547,-0.0354458913,0.3620596528,0.0075486349,-0.3814489245,0.2536882162,0.2413000017,-0.1637511253,0.1019428447,0.2105297595,0.3905052245,0.3419500291,0.0518311784,0.1003485695,0.0410616472,-0.1755670011,-0.052547358,0.2474575043,0.1228797287,0.2133717984,0.3228830397,0.2559886575,-0.2055215538,-0.2537412941,-0.1055223793,0.1573281139,-0.0536228269,0.3031086028,-0.3327677548,-0.0193307158,-0.3722386658,0.1489567608,-0.5131070018,0.0265103281,0.3228518367,-0.1077064797,0.1639059335,0.0533327833,0.0815824345,-0.3227351904,0.5315322876,0.3320508003,0.194754377,-0.4337192476,-0.3740612268,-0.7417867184,-0.060878139,-0.206008628,0.02456595,-0.1929228157,0.1862428188,-0.1321678311,0.037540324,0.1825464368,0.2190940231,0.236361295,0.2899665534,-0.2522948384,-0.2005007118,0.0371855609,0.2285678089,0.0390388146,-0.4399996102,0.1195399463,-0.3286365867,0.1335948259,-0.0845858455,-0.283585757,0.1401758939,0.0447453745,0.3270394504,0.1665534228,0.6542968154,-0.0836465061,-0.1344577968,-0.3480279744,-0.1684510857,-0.1131030992,0.1500907987,0.2106296569,0.3543981016,-0.111497499,0.3759313524,-0.3795743585,0.0954524279,-0.0690874904,-0.1211291328,-0.2700442374,-0.1315339059,-0.1551492959,0.1941407472,0.2851369977,0.1061910316,-0.0925377384,-0.0111997407,-0.1500885785,-0.2279428691,0.5045260191,-0.3369373679,-0.3883259296,-0.0437305719,0.1152147502,0.0646315664,0.2162813395,-0.2958243787,0.2166941762,0.3361885548,-0.0963055119,-0.1000958234,0.2369075567,0.0455233082,0.1424679309,0.0629596338,-0.0166317429,0.0973836705,-0.2265188247,-0.0544902533,-0.2047988921]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1287","title":"'iwslt2017-ro-nl', cannot be downloaded ","comments":"even with setting master like the following command, still remains \r\n\r\ndatasets.load_dataset(\"iwslt2017\", 'iwslt2017-en-nl', split=\"train\", script_version=\"master\")\r\n","body":"Hi\r\nI am trying \r\n`>>> datasets.load_dataset(\"iwslt2017\", 'iwslt2017-ro-nl', split=\"train\")`\r\n\r\ngetting this error thank you for your help\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset iwsl_t217\/iwslt2017-ro-nl (download: 314.07 MiB, generated: 39.92 MiB, post-processed: Unknown size, total: 354.00 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/iwsl_t217\/iwslt2017-ro-nl\/1.0.0\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/iwslt2017\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd\/iwslt2017.py\", line 118, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(MULTI_URL)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n\r\n```","comment_length":14,"text":"'iwslt2017-ro-nl', cannot be downloaded \nHi\r\nI am trying \r\n`>>> datasets.load_dataset(\"iwslt2017\", 'iwslt2017-ro-nl', split=\"train\")`\r\n\r\ngetting this error thank you for your help\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset iwsl_t217\/iwslt2017-ro-nl (download: 314.07 MiB, generated: 39.92 MiB, post-processed: Unknown size, total: 354.00 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/iwsl_t217\/iwslt2017-ro-nl\/1.0.0\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/iwslt2017\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd\/iwslt2017.py\", line 118, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(MULTI_URL)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n\r\n```\neven with setting master like the following command, still remains \r\n\r\ndatasets.load_dataset(\"iwslt2017\", 'iwslt2017-en-nl', split=\"train\", script_version=\"master\")\r\n","embeddings":[-0.3419763446,-0.30060938,-0.1523725539,0.2706135213,0.2139090747,0.2157970816,0.2569400668,0.3650550246,0.1976396739,-0.0201298464,-0.0654523149,-0.1176212877,0.1440995634,-0.0291719362,-0.1518213302,-0.2699722052,-0.0663298965,0.1131053045,-0.1293237358,0.0422331616,-0.3337647617,0.2209876627,-0.1799138039,0.0285914224,-0.3148556948,-0.2158248723,-0.1557890624,0.0668197274,-0.376522094,-0.3967740536,0.4967109859,0.0440180227,0.3563047349,0.3447470367,-0.0001073461,0.0428802036,0.3648747802,-0.1117397919,-0.2904242277,-0.1810606569,-0.5819841027,-0.3666986525,-0.1755940318,-0.3403346837,0.0163967311,-0.4022048116,-0.1394961774,-0.1585931033,0.3906446397,0.388822943,0.2995847166,0.1232836023,0.2581681013,-0.2987955511,-0.0361348465,-0.283228755,-0.0201401301,0.2171174586,-0.0970925391,-0.020589469,0.2999322712,0.20792225,-0.1322476715,0.0845834091,0.0945169777,-0.137185201,0.2849365473,-0.4758993089,0.306342274,0.1419929713,0.6968966126,-0.1230078042,-0.2451383919,0.1330755949,0.0477134213,-0.1949062496,0.1256547123,0.3085877597,-0.1431127042,0.1953275204,-0.1524961442,-0.0715682507,-0.1325484961,0.2132592052,-0.1820759028,0.3298105896,-0.0642741919,0.1544097811,0.3366713822,-0.0729813129,0.2190582752,-0.0332605802,-0.1243837699,0.3861820102,-0.338873446,0.002449838,-0.2990521193,0.0185885374,-0.1256584078,0.0470617972,-0.2219414711,-0.2800481319,0.1643054038,0.1297833622,0.0439549685,0.2667294443,-0.0249498244,0.2557565868,-0.029544726,0.3510740101,0.0903158039,-0.128841728,-0.183530882,-0.3094523549,-0.1967954487,0.0998697355,0.063150391,-0.1231369227,-0.488530159,-0.1932276338,-0.0544629619,-0.0670500621,0.0719540119,0.4097508788,-0.0886905044,0.3924994767,0.163899228,0.2355045825,0.0368230082,-0.2544769943,-0.258556813,0.3158781826,-0.1857732832,-0.0204413049,0.2933999598,0.0800527781,0.101256527,0.0159880724,-0.2408084869,-0.1507217437,0.1572513729,-0.2279166579,0.0483161844,0.3992284238,0.1482194215,0.3353996873,0.0706556588,0.0414648466,-0.1198901832,0.1366238296,-0.2916611433,-0.2356133014,-0.0833062753,0.2989494205,-0.0838587806,-0.1165730953,0.0086517511,-0.0890580639,0.217648685,-0.0663128868,-0.1008108631,-0.0856411159,-0.2119235545,-0.3296175599,0.2840553522,0.3511959612,-0.3999688625,-0.1349375695,-0.2301706076,-0.2683898211,0.3191880584,0.2966541648,-0.1974266917,0.3559424579,-0.2710108459,0.1405541301,0.4296115339,-0.3251674473,-0.5949853063,0.3352683187,0.0314203426,-0.1941059083,0.2613495588,0.1419058591,0.0988269597,-0.0779613182,-0.0219251625,0.3322188854,-0.0165831987,-0.097443454,-0.1484275609,-0.4670132697,0.2685246468,0.2582086325,0.2376422286,0.1317265779,0.0829866901,0.2184657454,0.3316060901,0.0024225796,0.2145042121,0.1683985293,0.2144212276,-0.0120791206,-0.0896465406,-0.016023051,-0.291590035,0.2323364913,0.0352539532,-0.1178119481,-0.1290831864,0.0036250551,-0.4926204383,-0.1394286007,-0.4215623736,-0.0569159538,0.190910399,0.1853034496,0.0793977603,-0.1578582674,-0.077596359,0.5676696897,-0.3102144301,-0.0721075162,-0.2714800537,0.2507627308,-0.1448059827,-0.0085442523,0.0757114217,0.0322704986,0.318339318,-0.035067495,-0.1745737493,0.2032273263,-0.1742767543,0.1249080747,0.0113041038,-0.0183872506,0.0161141548,-0.4125339091,0.1873710155,-0.023024777,0.1227757186,-0.0164506584,0.043144878,0.0571190715,0.0457890294,0.137702629,0.0710880011,0.0011138548,0.3079363406,-0.3163119256,0.0914543346,-0.1924466044,0.1263335347,-0.0542972386,0.0092262914,-0.0536241569,0.1528143138,-0.0957075283,0.4249424636,0.0955039039,0.0690912232,0.0036228031,0.0028498142,0.1753265113,-0.2760487497,0.5185399055,0.3655532598,0.0751129761,0.0784046054,-0.1150381565,0.1288037151,-0.1084229127,0.2247704417,0.0476504453,-0.0041810251,0.2197386175,-0.0676042587,-0.0982570946,-0.0923172608,-0.0493154153,0.2656356096,0.3520200253,-0.1040189266,0.0060471906,-0.2463860065,-0.2953222394,-0.27870363,-0.1008472964,-0.0477192849,-0.1535633057,-0.2378375381,0.1771686673,0.2375490665,0.1192688122,-0.2360265255,0.1147160009,0.025813099,-0.0294521693,-0.1604108065,0.0218824688,-0.1970059723,0.2072577775,0.3102039993,0.0527941659,0.5148772597,-0.2965624928,-0.0328365453,-0.2607624233,0.0909985304,-0.0588743016,-0.0106225405,0.0126678767,0.1085725948,0.5615016222,0.1883694828,-0.2239586264,0.3116089106,-0.0666209757,-0.0488946363,-0.1512012929,-0.1495619714,-0.0486634113,-0.0859781131,-0.4663751423,-0.6394459605,-0.4285477102,-0.1054912657,-0.1299048066,0.0185630955,0.10352052,0.1360276043,-0.0806234851,0.1401029527,-0.1422324181,-0.1530722529,-0.1259031445,0.4883402884,-0.1384137869,-0.2780931592,0.2463521957,0.0517431833,0.3558724821,0.1222597435,-0.4612587392,0.2047154158,-0.1789457053,0.2339013368,-0.123334527,0.1875346601,0.1291775852,-0.12157882,-0.0701697245,-0.0898954198,-0.0077213193,-0.0767805576,0.0147693707,0.5862768888,0.0112557458,0.3250548244,-0.0636339188,0.5220999122,0.3522051573,-0.1157827452,0.5259754062,0.0313490592,0.0135138659,-0.0820038617,-0.633777976,0.0230399352,-0.1061978936,0.0868648589,0.2244164795,-0.1441634148,-0.2223575413,-0.1677296609,-0.0043811467,-0.148821041,-0.0474478975,0.0343091525,0.3830757439,0.4424400032,0.0338994041,-0.048039794,-0.2091472298,-0.1317834407,0.0974779278,0.2961211801,-0.1347294599,0.2479324192,-0.0165558532,-0.1964795738,-0.2588474154,0.1430327594,0.236054197,0.3298193514,-0.2033838779,0.1699988842,0.1397537142,0.0282663312,0.5790627003,-0.1956948638,0.1182057783,0.2313864678,0.2583263218,-0.0488119945,-0.1092301533,-0.1493160427,0.2338519841,0.3252852857,0.2348416597,-0.2585186064,-0.1194622889,0.2451254874,0.3395388126,-0.297731787,-0.3928254843,-0.318682313,-0.4542882144,-0.4696657658,-0.1589414328,0.0924784914,0.3104878962,-0.2088991851,0.065757513,-0.0496007204,-0.1515216529,0.0941471905,0.1550450921,0.3905794919,0.2684715092,0.1671625674,0.2203499526,0.1206488833,0.0937275141,0.5806069374,-0.0820148289,-0.1815625131,0.1683234572,-0.0675854161,-0.0086019738,0.1252853274,-0.1580293328,0.0009671754,0.0902222097,0.0446865596,0.1281659454,0.1415161639,0.3960202336,-0.1422026455,-0.569853425,-0.5394011736,0.4596917629,0.062678799,0.0551528931,0.2357880026,-0.0408788286,-0.1155170724,-0.1661672741,-0.1388940066,0.8912362456,-0.0221582334,-0.1548423767,0.1937688291,0.1800160259,0.3420315385,-0.2002798915,0.1621747464,-0.2236508131,0.0076990472,-0.0533528291,-0.1299173534,0.0873081163,0.2866203487,-0.2430064529,0.3370768726,-0.1529806703,0.0282465853,0.153639257,0.4686248302,-0.2369810492,-0.1234860495,-0.2134869695,0.2599333823,-0.0697904751,0.2494766861,-0.2205187231,-0.1804306507,0.0077849352,-0.1079011783,-0.2287700772,0.1813034266,-0.2994191647,0.59278965,-0.1998260468,0.00426909,0.2811477482,0.1610227078,0.3320013881,0.0170541592,-0.2426313609,0.2426300943,-0.0703173429,-0.2951053083,-0.2690463662,0.1414592713,0.1052307561,-0.1932798773,-0.3641048968,0.1524308175,-0.0982600376,-0.1537448168,0.1027298346,-0.0649383813,0.4251005054,-0.222037971,-0.1376624554,-0.1038458347,0.0651887953,-0.1359352618,0.1879301816,-0.0017168126,-0.1478768438,0.0615535192,-0.1109071523,-0.2153674513,-0.0525974594,0.4336761236,0.0625075027,0.0277480204,0.6179629564,0.0915395692,-0.1221234575,-0.3233804703,-0.0919291601,-0.0856461897,-0.4771237969,0.0789747238,0.113899909,0.0712306201,-0.185955584,0.4050688446,0.2536170483,0.0127548082,0.0016877064,-0.654889822,-0.3016872704,-0.0393077284,-0.021214975,0.159457311,-0.1859031767,0.0273677949,0.1794208437,-0.1460598707,-0.3506726027,0.0116716139,-0.220564127,0.0751525685,0.4576438963,-0.1315953881,0.1161369309,0.1173945889,0.2125164866,0.076565057,-0.1298950762,-0.3067094088,-0.1219613627,0.1114019677,-0.0835447311,-0.1904917657,-0.0903211758,-0.3495313227,-0.1241815314,0.1055315956,-0.1069640517,0.0205262285,0.0696955919,0.0964210704,0.1702435464,-0.022158891,-0.2776302397,0.2832111418,0.0862473249,0.1938600838,0.0527240932,0.290671438,0.1994037032,0.0168558266,-0.1140226573,-0.171184659,0.2523238659,-0.1742706001,0.0158356875,-0.3120677173,0.2289433181,0.182236135,0.0388923697,0.5452494621,-0.1949207187,0.1702246964,0.1145869568,0.2919920385,-0.3564121425,0.0131854489,0.1363899857,-0.012141061,-0.1282097399,0.0647349134,0.19133389,-0.1645380408,0.1727178991,0.0361656807,0.4443298876,-0.1137610525,0.0333962366,0.3205941021,0.1102307662,0.0073839156,0.1503521353,0.0869146287,0.1049035192,0.3699611127,-0.3519082963,0.334184587,-0.0292459577,0.1960637122,-0.0807112232,-0.463588804,-0.0132800359,-0.0271061286,-0.0280076005,-0.064129591,-0.0836857557,0.2307623029,-0.0578371771,-0.2075787187,-0.3942587376,0.2716932595,-0.1383715421,-0.0168737527,-0.2087452561,-0.1575703323,-0.3138845265,0.0200664438,0.0944762826,-0.2163674533,0.276887536,0.1739004105,-0.0674182326,-0.2864430547,-0.0354458913,0.3620596528,0.0075486349,-0.3814489245,0.2536882162,0.2413000017,-0.1637511253,0.1019428447,0.2105297595,0.3905052245,0.3419500291,0.0518311784,0.1003485695,0.0410616472,-0.1755670011,-0.052547358,0.2474575043,0.1228797287,0.2133717984,0.3228830397,0.2559886575,-0.2055215538,-0.2537412941,-0.1055223793,0.1573281139,-0.0536228269,0.3031086028,-0.3327677548,-0.0193307158,-0.3722386658,0.1489567608,-0.5131070018,0.0265103281,0.3228518367,-0.1077064797,0.1639059335,0.0533327833,0.0815824345,-0.3227351904,0.5315322876,0.3320508003,0.194754377,-0.4337192476,-0.3740612268,-0.7417867184,-0.060878139,-0.206008628,0.02456595,-0.1929228157,0.1862428188,-0.1321678311,0.037540324,0.1825464368,0.2190940231,0.236361295,0.2899665534,-0.2522948384,-0.2005007118,0.0371855609,0.2285678089,0.0390388146,-0.4399996102,0.1195399463,-0.3286365867,0.1335948259,-0.0845858455,-0.283585757,0.1401758939,0.0447453745,0.3270394504,0.1665534228,0.6542968154,-0.0836465061,-0.1344577968,-0.3480279744,-0.1684510857,-0.1131030992,0.1500907987,0.2106296569,0.3543981016,-0.111497499,0.3759313524,-0.3795743585,0.0954524279,-0.0690874904,-0.1211291328,-0.2700442374,-0.1315339059,-0.1551492959,0.1941407472,0.2851369977,0.1061910316,-0.0925377384,-0.0111997407,-0.1500885785,-0.2279428691,0.5045260191,-0.3369373679,-0.3883259296,-0.0437305719,0.1152147502,0.0646315664,0.2162813395,-0.2958243787,0.2166941762,0.3361885548,-0.0963055119,-0.1000958234,0.2369075567,0.0455233082,0.1424679309,0.0629596338,-0.0166317429,0.0973836705,-0.2265188247,-0.0544902533,-0.2047988921]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1287","title":"'iwslt2017-ro-nl', cannot be downloaded ","comments":"Looks like the data has been moved from its original location to google drive\r\n\r\nNew url: https:\/\/drive.google.com\/u\/0\/uc?id=12ycYSzLIG253AFN35Y6qoyf9wtkOjakp&export=download","body":"Hi\r\nI am trying \r\n`>>> datasets.load_dataset(\"iwslt2017\", 'iwslt2017-ro-nl', split=\"train\")`\r\n\r\ngetting this error thank you for your help\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset iwsl_t217\/iwslt2017-ro-nl (download: 314.07 MiB, generated: 39.92 MiB, post-processed: Unknown size, total: 354.00 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/iwsl_t217\/iwslt2017-ro-nl\/1.0.0\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/iwslt2017\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd\/iwslt2017.py\", line 118, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(MULTI_URL)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n\r\n```","comment_length":17,"text":"'iwslt2017-ro-nl', cannot be downloaded \nHi\r\nI am trying \r\n`>>> datasets.load_dataset(\"iwslt2017\", 'iwslt2017-ro-nl', split=\"train\")`\r\n\r\ngetting this error thank you for your help\r\n```\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nDownloading and preparing dataset iwsl_t217\/iwslt2017-ro-nl (download: 314.07 MiB, generated: 39.92 MiB, post-processed: Unknown size, total: 354.00 MiB) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/iwsl_t217\/iwslt2017-ro-nl\/1.0.0\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/iwslt2017\/cca6935a0851a8ceac1202a62c958738bdfa23c57a51bc52ac1c5ebd2aa172cd\/iwslt2017.py\", line 118, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(MULTI_URL)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/wit3.fbk.eu\/archive\/2017-01-trnmted\/\/texts\/DeEnItNlRo\/DeEnItNlRo\/DeEnItNlRo-DeEnItNlRo.tgz\r\n\r\n```\nLooks like the data has been moved from its original location to google drive\r\n\r\nNew url: https:\/\/drive.google.com\/u\/0\/uc?id=12ycYSzLIG253AFN35Y6qoyf9wtkOjakp&export=download","embeddings":[-0.3419763446,-0.30060938,-0.1523725539,0.2706135213,0.2139090747,0.2157970816,0.2569400668,0.3650550246,0.1976396739,-0.0201298464,-0.0654523149,-0.1176212877,0.1440995634,-0.0291719362,-0.1518213302,-0.2699722052,-0.0663298965,0.1131053045,-0.1293237358,0.0422331616,-0.3337647617,0.2209876627,-0.1799138039,0.0285914224,-0.3148556948,-0.2158248723,-0.1557890624,0.0668197274,-0.376522094,-0.3967740536,0.4967109859,0.0440180227,0.3563047349,0.3447470367,-0.0001073461,0.0428802036,0.3648747802,-0.1117397919,-0.2904242277,-0.1810606569,-0.5819841027,-0.3666986525,-0.1755940318,-0.3403346837,0.0163967311,-0.4022048116,-0.1394961774,-0.1585931033,0.3906446397,0.388822943,0.2995847166,0.1232836023,0.2581681013,-0.2987955511,-0.0361348465,-0.283228755,-0.0201401301,0.2171174586,-0.0970925391,-0.020589469,0.2999322712,0.20792225,-0.1322476715,0.0845834091,0.0945169777,-0.137185201,0.2849365473,-0.4758993089,0.306342274,0.1419929713,0.6968966126,-0.1230078042,-0.2451383919,0.1330755949,0.0477134213,-0.1949062496,0.1256547123,0.3085877597,-0.1431127042,0.1953275204,-0.1524961442,-0.0715682507,-0.1325484961,0.2132592052,-0.1820759028,0.3298105896,-0.0642741919,0.1544097811,0.3366713822,-0.0729813129,0.2190582752,-0.0332605802,-0.1243837699,0.3861820102,-0.338873446,0.002449838,-0.2990521193,0.0185885374,-0.1256584078,0.0470617972,-0.2219414711,-0.2800481319,0.1643054038,0.1297833622,0.0439549685,0.2667294443,-0.0249498244,0.2557565868,-0.029544726,0.3510740101,0.0903158039,-0.128841728,-0.183530882,-0.3094523549,-0.1967954487,0.0998697355,0.063150391,-0.1231369227,-0.488530159,-0.1932276338,-0.0544629619,-0.0670500621,0.0719540119,0.4097508788,-0.0886905044,0.3924994767,0.163899228,0.2355045825,0.0368230082,-0.2544769943,-0.258556813,0.3158781826,-0.1857732832,-0.0204413049,0.2933999598,0.0800527781,0.101256527,0.0159880724,-0.2408084869,-0.1507217437,0.1572513729,-0.2279166579,0.0483161844,0.3992284238,0.1482194215,0.3353996873,0.0706556588,0.0414648466,-0.1198901832,0.1366238296,-0.2916611433,-0.2356133014,-0.0833062753,0.2989494205,-0.0838587806,-0.1165730953,0.0086517511,-0.0890580639,0.217648685,-0.0663128868,-0.1008108631,-0.0856411159,-0.2119235545,-0.3296175599,0.2840553522,0.3511959612,-0.3999688625,-0.1349375695,-0.2301706076,-0.2683898211,0.3191880584,0.2966541648,-0.1974266917,0.3559424579,-0.2710108459,0.1405541301,0.4296115339,-0.3251674473,-0.5949853063,0.3352683187,0.0314203426,-0.1941059083,0.2613495588,0.1419058591,0.0988269597,-0.0779613182,-0.0219251625,0.3322188854,-0.0165831987,-0.097443454,-0.1484275609,-0.4670132697,0.2685246468,0.2582086325,0.2376422286,0.1317265779,0.0829866901,0.2184657454,0.3316060901,0.0024225796,0.2145042121,0.1683985293,0.2144212276,-0.0120791206,-0.0896465406,-0.016023051,-0.291590035,0.2323364913,0.0352539532,-0.1178119481,-0.1290831864,0.0036250551,-0.4926204383,-0.1394286007,-0.4215623736,-0.0569159538,0.190910399,0.1853034496,0.0793977603,-0.1578582674,-0.077596359,0.5676696897,-0.3102144301,-0.0721075162,-0.2714800537,0.2507627308,-0.1448059827,-0.0085442523,0.0757114217,0.0322704986,0.318339318,-0.035067495,-0.1745737493,0.2032273263,-0.1742767543,0.1249080747,0.0113041038,-0.0183872506,0.0161141548,-0.4125339091,0.1873710155,-0.023024777,0.1227757186,-0.0164506584,0.043144878,0.0571190715,0.0457890294,0.137702629,0.0710880011,0.0011138548,0.3079363406,-0.3163119256,0.0914543346,-0.1924466044,0.1263335347,-0.0542972386,0.0092262914,-0.0536241569,0.1528143138,-0.0957075283,0.4249424636,0.0955039039,0.0690912232,0.0036228031,0.0028498142,0.1753265113,-0.2760487497,0.5185399055,0.3655532598,0.0751129761,0.0784046054,-0.1150381565,0.1288037151,-0.1084229127,0.2247704417,0.0476504453,-0.0041810251,0.2197386175,-0.0676042587,-0.0982570946,-0.0923172608,-0.0493154153,0.2656356096,0.3520200253,-0.1040189266,0.0060471906,-0.2463860065,-0.2953222394,-0.27870363,-0.1008472964,-0.0477192849,-0.1535633057,-0.2378375381,0.1771686673,0.2375490665,0.1192688122,-0.2360265255,0.1147160009,0.025813099,-0.0294521693,-0.1604108065,0.0218824688,-0.1970059723,0.2072577775,0.3102039993,0.0527941659,0.5148772597,-0.2965624928,-0.0328365453,-0.2607624233,0.0909985304,-0.0588743016,-0.0106225405,0.0126678767,0.1085725948,0.5615016222,0.1883694828,-0.2239586264,0.3116089106,-0.0666209757,-0.0488946363,-0.1512012929,-0.1495619714,-0.0486634113,-0.0859781131,-0.4663751423,-0.6394459605,-0.4285477102,-0.1054912657,-0.1299048066,0.0185630955,0.10352052,0.1360276043,-0.0806234851,0.1401029527,-0.1422324181,-0.1530722529,-0.1259031445,0.4883402884,-0.1384137869,-0.2780931592,0.2463521957,0.0517431833,0.3558724821,0.1222597435,-0.4612587392,0.2047154158,-0.1789457053,0.2339013368,-0.123334527,0.1875346601,0.1291775852,-0.12157882,-0.0701697245,-0.0898954198,-0.0077213193,-0.0767805576,0.0147693707,0.5862768888,0.0112557458,0.3250548244,-0.0636339188,0.5220999122,0.3522051573,-0.1157827452,0.5259754062,0.0313490592,0.0135138659,-0.0820038617,-0.633777976,0.0230399352,-0.1061978936,0.0868648589,0.2244164795,-0.1441634148,-0.2223575413,-0.1677296609,-0.0043811467,-0.148821041,-0.0474478975,0.0343091525,0.3830757439,0.4424400032,0.0338994041,-0.048039794,-0.2091472298,-0.1317834407,0.0974779278,0.2961211801,-0.1347294599,0.2479324192,-0.0165558532,-0.1964795738,-0.2588474154,0.1430327594,0.236054197,0.3298193514,-0.2033838779,0.1699988842,0.1397537142,0.0282663312,0.5790627003,-0.1956948638,0.1182057783,0.2313864678,0.2583263218,-0.0488119945,-0.1092301533,-0.1493160427,0.2338519841,0.3252852857,0.2348416597,-0.2585186064,-0.1194622889,0.2451254874,0.3395388126,-0.297731787,-0.3928254843,-0.318682313,-0.4542882144,-0.4696657658,-0.1589414328,0.0924784914,0.3104878962,-0.2088991851,0.065757513,-0.0496007204,-0.1515216529,0.0941471905,0.1550450921,0.3905794919,0.2684715092,0.1671625674,0.2203499526,0.1206488833,0.0937275141,0.5806069374,-0.0820148289,-0.1815625131,0.1683234572,-0.0675854161,-0.0086019738,0.1252853274,-0.1580293328,0.0009671754,0.0902222097,0.0446865596,0.1281659454,0.1415161639,0.3960202336,-0.1422026455,-0.569853425,-0.5394011736,0.4596917629,0.062678799,0.0551528931,0.2357880026,-0.0408788286,-0.1155170724,-0.1661672741,-0.1388940066,0.8912362456,-0.0221582334,-0.1548423767,0.1937688291,0.1800160259,0.3420315385,-0.2002798915,0.1621747464,-0.2236508131,0.0076990472,-0.0533528291,-0.1299173534,0.0873081163,0.2866203487,-0.2430064529,0.3370768726,-0.1529806703,0.0282465853,0.153639257,0.4686248302,-0.2369810492,-0.1234860495,-0.2134869695,0.2599333823,-0.0697904751,0.2494766861,-0.2205187231,-0.1804306507,0.0077849352,-0.1079011783,-0.2287700772,0.1813034266,-0.2994191647,0.59278965,-0.1998260468,0.00426909,0.2811477482,0.1610227078,0.3320013881,0.0170541592,-0.2426313609,0.2426300943,-0.0703173429,-0.2951053083,-0.2690463662,0.1414592713,0.1052307561,-0.1932798773,-0.3641048968,0.1524308175,-0.0982600376,-0.1537448168,0.1027298346,-0.0649383813,0.4251005054,-0.222037971,-0.1376624554,-0.1038458347,0.0651887953,-0.1359352618,0.1879301816,-0.0017168126,-0.1478768438,0.0615535192,-0.1109071523,-0.2153674513,-0.0525974594,0.4336761236,0.0625075027,0.0277480204,0.6179629564,0.0915395692,-0.1221234575,-0.3233804703,-0.0919291601,-0.0856461897,-0.4771237969,0.0789747238,0.113899909,0.0712306201,-0.185955584,0.4050688446,0.2536170483,0.0127548082,0.0016877064,-0.654889822,-0.3016872704,-0.0393077284,-0.021214975,0.159457311,-0.1859031767,0.0273677949,0.1794208437,-0.1460598707,-0.3506726027,0.0116716139,-0.220564127,0.0751525685,0.4576438963,-0.1315953881,0.1161369309,0.1173945889,0.2125164866,0.076565057,-0.1298950762,-0.3067094088,-0.1219613627,0.1114019677,-0.0835447311,-0.1904917657,-0.0903211758,-0.3495313227,-0.1241815314,0.1055315956,-0.1069640517,0.0205262285,0.0696955919,0.0964210704,0.1702435464,-0.022158891,-0.2776302397,0.2832111418,0.0862473249,0.1938600838,0.0527240932,0.290671438,0.1994037032,0.0168558266,-0.1140226573,-0.171184659,0.2523238659,-0.1742706001,0.0158356875,-0.3120677173,0.2289433181,0.182236135,0.0388923697,0.5452494621,-0.1949207187,0.1702246964,0.1145869568,0.2919920385,-0.3564121425,0.0131854489,0.1363899857,-0.012141061,-0.1282097399,0.0647349134,0.19133389,-0.1645380408,0.1727178991,0.0361656807,0.4443298876,-0.1137610525,0.0333962366,0.3205941021,0.1102307662,0.0073839156,0.1503521353,0.0869146287,0.1049035192,0.3699611127,-0.3519082963,0.334184587,-0.0292459577,0.1960637122,-0.0807112232,-0.463588804,-0.0132800359,-0.0271061286,-0.0280076005,-0.064129591,-0.0836857557,0.2307623029,-0.0578371771,-0.2075787187,-0.3942587376,0.2716932595,-0.1383715421,-0.0168737527,-0.2087452561,-0.1575703323,-0.3138845265,0.0200664438,0.0944762826,-0.2163674533,0.276887536,0.1739004105,-0.0674182326,-0.2864430547,-0.0354458913,0.3620596528,0.0075486349,-0.3814489245,0.2536882162,0.2413000017,-0.1637511253,0.1019428447,0.2105297595,0.3905052245,0.3419500291,0.0518311784,0.1003485695,0.0410616472,-0.1755670011,-0.052547358,0.2474575043,0.1228797287,0.2133717984,0.3228830397,0.2559886575,-0.2055215538,-0.2537412941,-0.1055223793,0.1573281139,-0.0536228269,0.3031086028,-0.3327677548,-0.0193307158,-0.3722386658,0.1489567608,-0.5131070018,0.0265103281,0.3228518367,-0.1077064797,0.1639059335,0.0533327833,0.0815824345,-0.3227351904,0.5315322876,0.3320508003,0.194754377,-0.4337192476,-0.3740612268,-0.7417867184,-0.060878139,-0.206008628,0.02456595,-0.1929228157,0.1862428188,-0.1321678311,0.037540324,0.1825464368,0.2190940231,0.236361295,0.2899665534,-0.2522948384,-0.2005007118,0.0371855609,0.2285678089,0.0390388146,-0.4399996102,0.1195399463,-0.3286365867,0.1335948259,-0.0845858455,-0.283585757,0.1401758939,0.0447453745,0.3270394504,0.1665534228,0.6542968154,-0.0836465061,-0.1344577968,-0.3480279744,-0.1684510857,-0.1131030992,0.1500907987,0.2106296569,0.3543981016,-0.111497499,0.3759313524,-0.3795743585,0.0954524279,-0.0690874904,-0.1211291328,-0.2700442374,-0.1315339059,-0.1551492959,0.1941407472,0.2851369977,0.1061910316,-0.0925377384,-0.0111997407,-0.1500885785,-0.2279428691,0.5045260191,-0.3369373679,-0.3883259296,-0.0437305719,0.1152147502,0.0646315664,0.2162813395,-0.2958243787,0.2166941762,0.3361885548,-0.0963055119,-0.1000958234,0.2369075567,0.0455233082,0.1424679309,0.0629596338,-0.0166317429,0.0973836705,-0.2265188247,-0.0544902533,-0.2047988921]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1286","title":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted","comments":"I remember also getting the same issue for several other translation datasets like all the iwslt2017 group, this is blokcing me and I really need to fix it and I was wondering if you have an idea on this. @lhoestq thanks,. ","body":"Hi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n","comment_length":41,"text":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted\nHi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n\nI remember also getting the same issue for several other translation datasets like all the iwslt2017 group, this is blokcing me and I really need to fix it and I was wondering if you have an idea on this. @lhoestq thanks,. ","embeddings":[-0.0197352897,-0.7191827893,0.0039322353,0.349368006,0.484241426,0.0076734996,0.2424444556,0.1216879562,-0.2513463497,0.3279392719,-0.0821483955,-0.1225553527,-0.1213336438,0.3986013532,0.024262283,-0.2556790411,-0.0995945483,0.0782158896,-0.3232808411,0.1850418597,-0.0975957587,0.2666340768,-0.3790644109,0.2954924107,-0.077346921,0.0884474367,0.1073483229,0.0845447779,-0.0165824965,-0.3517342508,0.0590415671,-0.0919298232,-0.3089157045,0.3623276055,-0.0001231714,-0.1029601544,0.2742366493,0.044334691,-0.0298559833,-0.1064275429,0.264336437,0.5919190049,-0.1285633147,0.155628711,-0.1199958175,0.0390061848,0.131960839,-0.02621934,0.2547330558,0.2040431052,0.1139890254,0.3572780192,0.4076769948,0.0077055944,-0.0644371659,-0.047120966,-0.2066649497,0.1545071602,0.0657213703,-0.1996357739,0.1090153903,-0.0004496211,0.1906781793,-0.0334959961,-0.2198651433,-0.3123973608,-0.0370121896,-0.457115531,0.0150993029,0.4477550387,-0.0995786712,-0.2989667356,-0.1321324855,0.2411860675,0.0361542404,-0.7871572971,0.0136458101,0.1679371297,-0.0502248146,0.2885320783,-0.137509048,0.1263263226,-0.2918356359,0.0164490547,-0.0907559618,0.296466738,-0.1985564381,0.2684500217,0.3229812086,0.1940049678,-0.3849190474,0.1892252117,0.0081450799,0.1069084778,-0.3058028519,-0.2175878286,0.1347346306,0.0532117598,0.291143328,0.1250610203,-0.0492140464,0.0266518034,-0.0314312987,0.1974624097,0.2514950335,-0.0322238952,0.0891965702,0.1261572838,0.157536611,0.5037890673,0.053735029,0.2031769902,0.0966271609,-0.2018554062,0.0816629007,-0.0702104494,0.5441356897,-0.1696140319,-0.5968497992,0.4957826138,-0.4942743778,-0.0085916603,0.2324092388,0.4670405984,-0.1604339182,0.0753958225,0.251483798,-0.1050434783,0.033637356,0.03451325,-0.0803707466,0.2207057327,0.1228926927,0.2221151143,-0.1487662196,-0.4933043122,0.1599145234,-0.1897882819,0.2903176546,-0.6953832507,-0.2288500816,-0.366612047,-0.0604084767,0.1137853563,-0.1019294262,-0.1594460607,0.2681770623,0.0848985463,-0.013057163,0.0232961178,0.2334747612,-0.3981198967,0.1256912202,0.1727491468,-0.1056274921,0.1610930264,0.0997461826,-0.0547138155,-0.037544772,0.2110037208,0.0817849934,-0.023355253,-0.5620622039,0.0634886622,0.2465737909,0.3047662675,0.0822414607,-0.0770296752,0.1512786895,-0.0502387509,0.054323215,0.4209915996,-0.0538993366,0.4664910734,-0.2711001635,0.1770830005,0.277297765,-0.3652850688,-0.0003139204,0.1826036125,-0.2375145853,0.1117426082,-0.2652013898,-0.4663338959,-0.1654590964,-0.1183409244,0.1349778473,-0.0789419264,-0.2356961071,-0.0736637264,-0.3157056272,-0.0500996858,0.2459145337,0.1108266786,0.0505057685,-0.2129725069,0.1019253507,-0.0896832198,0.2275961339,0.0058478625,-0.2316600084,0.11156746,0.4407880902,-0.0369520411,-0.0522972122,-0.1141529381,-0.116719678,0.1146957576,-0.0951117575,0.2058560997,0.2640617192,-0.0140006486,-0.222789079,0.3215104342,-0.2788105011,-0.2408593893,0.1297049522,-0.0126401894,-0.1174058467,0.0466670133,0.0669539571,-0.0052326033,-0.5102657676,0.1701666266,0.1292891949,0.047954496,-0.4098981023,-0.0345081277,-0.0219770689,0.1678383499,0.0930051282,-0.2305410653,-0.0614778958,0.4744261503,-0.0609407164,0.1962468177,-0.0178826489,-0.2281215787,-0.019167576,-0.6979162097,-0.3184904456,-0.0092402529,0.1823446751,-0.1330951899,0.4334768355,0.2950298786,0.1894212216,-0.0313181132,0.1306529492,-0.1452960372,-0.1054134294,-0.0241774917,-0.1970705837,-0.0412810817,0.4775482118,0.0791641027,-0.083314769,-0.0907501876,-0.5611173511,-0.1841323078,0.3183330894,-0.2006358951,0.0832459405,0.2185398936,0.2986201346,-0.1066382155,-0.174405396,-0.2389483154,0.4070246518,-0.0666783303,-0.0943469256,0.3965403438,-0.1829442829,0.1772942096,0.30615291,0.1469939947,0.2547303438,0.3355357349,0.0759050399,0.0548323095,-0.3406339884,0.1482571512,-0.1149554327,0.2089336812,-0.2045271844,0.2236363739,-0.2222569287,0.1700025946,-0.2625818849,-0.2926515341,-0.1523296684,-0.0743206516,0.020759508,0.3121842742,0.1147071868,0.2348940372,0.3013081253,0.2710024714,0.1796744317,0.24606888,0.0581639707,0.1148115098,-0.2809874415,-0.0713953227,0.409381628,-0.3983355463,0.437820226,0.0031497572,-0.3976996541,-0.1845485568,-0.3869085014,0.2667299807,-0.3574424982,0.4122446179,0.23498936,0.0989026725,0.2883796394,-0.0820783675,0.2549666166,-0.1147719994,-0.2736467421,0.1317070425,0.0489208996,-0.0504887998,-0.1470930874,-0.1372155547,-0.076008074,-0.1587286741,0.4361136556,0.0166942198,0.2049041837,0.3152825236,0.0969091505,0.2270750254,-0.0253577381,0.1294412464,-0.3533500433,-0.2036488354,-0.0627004877,0.0351996385,-0.2237658501,-0.0332907401,0.0955420956,0.3138401806,-0.115697816,-0.3986230791,-0.2539067566,0.2589060962,-0.4011817276,0.0286345966,-0.0587614216,0.2250745445,-0.0227604751,0.0023586736,0.1034687683,-0.1657687128,0.2917427123,0.0014571624,0.4208348393,-0.0582709499,0.4073520303,0.1907383204,0.2507788241,0.2218789309,-0.1074185222,0.2332936674,-0.0760882199,0.2300221622,0.0626940429,-0.2186158597,0.1225989982,-0.1815152466,-0.0678680167,0.2153835744,-0.2848355174,0.1016196385,-0.2425624877,0.3786380589,-0.3925980031,-0.2305805087,-0.0728006586,0.084850952,0.2980444729,-0.218303442,-0.1612262875,0.3346336484,0.1729964465,0.1649609953,0.1857609898,-0.0097199064,-0.2947641015,0.1768292785,-0.4341126978,-0.1974519044,0.2037676871,0.1779196262,0.3842518628,-0.1591307223,-0.1912715137,0.0965435058,-0.3532340825,0.2157043666,-0.1090002358,0.1710741967,-0.0191141888,-0.2454843968,0.0143333711,-0.1819860488,-0.2506864071,-0.1394036263,-0.0508332662,0.2705188692,-0.4435863197,-0.0580348633,0.9266176224,-0.0307043791,-0.2027936578,-0.0545093827,-0.3822521865,-0.3325395584,-0.2178062052,0.3268850744,0.0803599432,0.1595795751,-0.2594973445,0.054746896,0.1613204628,-0.0796455517,0.1856283695,0.0180702563,-0.1477721334,-0.2234720886,-0.0046396349,-0.0809065476,-0.2214631587,0.3577394187,0.066345498,0.1223880351,0.0403788388,-0.1450326741,0.2732554376,0.2892638147,0.4261297882,-0.2185800523,0.2583729923,-0.1215013042,0.2301147729,-0.2938078642,0.114960365,0.5577464104,0.1897827685,-0.3376097083,0.3320045173,0.1806747466,-0.3360965848,0.121775113,0.473772347,0.2906243801,-0.034681879,0.0033714068,0.0544294566,1.076790452,-0.1063169241,-0.1410206258,0.3265558183,-0.2886441648,-0.0629893243,0.1711464971,0.074469611,-0.3738822639,-0.2014196664,-0.1141918153,-0.099731572,0.182915628,-0.5049661398,-0.46861884,0.270686388,0.2902297676,0.047393661,0.1362045556,0.0351531766,0.0513243116,0.0296752714,-0.1473932117,0.075758256,-0.1771012098,0.4955222607,-0.3521041572,0.006089679,0.1425145864,-0.2339378148,-0.5489826798,0.1974165887,-0.7597703934,-0.2172420472,0.1435348392,-0.0956566334,0.2164642364,-0.0425305665,0.1211306229,0.2369874418,-0.0984941423,-0.0710498318,-0.0660440177,-0.2263087183,-0.278244406,0.080314897,0.1488425285,-0.1012301594,-0.3922503591,0.2363767028,-0.0378921032,-0.1245239452,0.2563019693,-0.1621961296,0.0216446221,-0.1495987326,-0.2880109847,0.1166077033,-0.4050696194,-0.3859924376,0.0441254005,-0.0842707083,-0.6793951392,-0.0104836412,0.2277966589,-0.2703295052,0.0709382817,0.3031373918,-0.4074920118,0.4476473927,0.2836990356,0.3070386052,-0.0456531011,-0.2508759201,-0.2475806177,-0.096215196,-0.239861846,0.1254095435,-0.1179654077,0.0117817651,-0.0179658718,0.3521892726,0.2009031028,0.130558461,-0.0421137735,-0.1326913387,-0.4322311878,0.1133480817,0.0895270556,0.3084596992,0.0070495745,0.8161113858,0.0050012395,0.1381090879,-0.2168498635,0.1537139416,-0.5475091934,0.1356402934,0.5144497156,-0.3967116475,0.2740752399,0.1211526692,0.1307330132,0.3177930415,-0.2503612041,-0.1576968282,-0.028146917,0.1464556307,0.1394556761,-0.2126519084,0.0227943938,0.170848608,0.0856941938,0.169199571,0.0142538045,-0.1359180212,0.0720585734,0.0227145944,0.5171648264,0.066288501,-0.133510977,-0.3067292571,-0.0594982728,0.2850827277,0.1547500938,-0.0466246828,-0.004913874,-0.3229970634,-0.0157679971,0.0993029922,0.2712794542,-0.0949808732,0.1953766793,-0.2645510137,-0.3805359304,-0.2151184231,0.5645757914,0.3280044198,-0.3266963065,0.086527355,-0.0304161627,0.018389836,-0.2668769658,0.1713368893,-0.1892541796,-0.1514424533,-0.0306949466,0.5003805757,-0.1047546417,0.2028654367,0.0371597931,0.213701725,-0.00022627,-0.1407934874,-0.0896461084,0.0310132988,-0.2003608495,0.1695739627,0.1094613299,-0.1566507071,0.3708167672,-0.323977381,-0.0258680321,0.3565204144,-0.5007627606,0.3787528872,0.1755802333,-0.2727207839,0.0502875932,0.2554655969,0.0517259911,0.1064760238,0.0574634559,0.5036066771,0.0453846492,-0.3039710224,-0.0631596595,0.4623815715,-0.1607019305,-0.2788872123,0.1073022485,-0.0637273565,-0.3115138113,0.134660393,0.1420977265,-0.02707389,0.0633941963,0.0979784504,-0.3087844551,-0.1646403223,-0.401157558,0.2207766622,0.2365179211,-0.2379700541,0.5151308179,0.2662872076,-0.0371031389,-0.2489867359,0.7079617977,0.3704000711,-0.044673983,-0.1711800694,0.1376731247,-0.2246428877,0.1120398939,0.2599118352,-0.0763739496,0.0478594899,-0.0139040696,0.1707173288,0.0092956806,-0.0898856148,0.1108519435,0.1339693218,-0.3126707971,-0.2695256174,0.1654578,0.014774763,-0.0065957443,-0.4888792634,0.381357491,-0.2600221634,0.6083681583,0.3633648157,-0.0167721026,0.1414662302,0.2023964226,0.0427674167,0.2453739345,0.3704001904,0.2853342891,-0.1474468559,-0.3849023879,-0.3310810626,-0.3185833395,0.2492609769,-0.0115360944,0.4620011747,0.2291846871,0.134952575,-0.2289453298,0.0339672528,0.4562621713,0.2448747754,0.0648012906,0.1269744784,-0.2102719843,0.0880673677,-0.1616372764,-0.2616662979,0.0603457205,-0.1920471936,0.1918265373,-0.2449938953,-0.0385891534,-0.0182405785,0.0576571263,-0.4193947613,-0.0711653382,0.1217793003,-0.0717778057,-0.0353250094,-0.1474374086,-0.3151301742,0.029259732,-0.0187704731,-0.0378767923,0.2944749594,0.0775456652,-0.0054252218,0.1476282179,-0.4997337461,-0.2743061185,0.0258237924,0.4406977594,-0.2658599317,-0.2488072366,0.1811849922,0.0307605229,-0.2400042415,0.132065475,0.3002210855,0.0648869798,0.091827713,-0.559091866,-0.1621075869,0.6869646907,0.3171429932,0.0053969757,-0.1448841244,0.1080972478,0.214095071,-0.2759179175,-0.8308287263,0.0773528218,0.1110792533,0.0343419835,-0.3358500004,0.4130390882,-0.396550715,0.0305419471,0.146956265,0.2626532614,0.2095080614,0.0917727649,0.3162976801,-0.1423849612]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1286","title":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted","comments":"maybe there is an empty line or something inside these datasets? could you tell me why this is happening? thanks ","body":"Hi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n","comment_length":20,"text":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted\nHi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n\nmaybe there is an empty line or something inside these datasets? could you tell me why this is happening? thanks ","embeddings":[-0.0197352897,-0.7191827893,0.0039322353,0.349368006,0.484241426,0.0076734996,0.2424444556,0.1216879562,-0.2513463497,0.3279392719,-0.0821483955,-0.1225553527,-0.1213336438,0.3986013532,0.024262283,-0.2556790411,-0.0995945483,0.0782158896,-0.3232808411,0.1850418597,-0.0975957587,0.2666340768,-0.3790644109,0.2954924107,-0.077346921,0.0884474367,0.1073483229,0.0845447779,-0.0165824965,-0.3517342508,0.0590415671,-0.0919298232,-0.3089157045,0.3623276055,-0.0001231714,-0.1029601544,0.2742366493,0.044334691,-0.0298559833,-0.1064275429,0.264336437,0.5919190049,-0.1285633147,0.155628711,-0.1199958175,0.0390061848,0.131960839,-0.02621934,0.2547330558,0.2040431052,0.1139890254,0.3572780192,0.4076769948,0.0077055944,-0.0644371659,-0.047120966,-0.2066649497,0.1545071602,0.0657213703,-0.1996357739,0.1090153903,-0.0004496211,0.1906781793,-0.0334959961,-0.2198651433,-0.3123973608,-0.0370121896,-0.457115531,0.0150993029,0.4477550387,-0.0995786712,-0.2989667356,-0.1321324855,0.2411860675,0.0361542404,-0.7871572971,0.0136458101,0.1679371297,-0.0502248146,0.2885320783,-0.137509048,0.1263263226,-0.2918356359,0.0164490547,-0.0907559618,0.296466738,-0.1985564381,0.2684500217,0.3229812086,0.1940049678,-0.3849190474,0.1892252117,0.0081450799,0.1069084778,-0.3058028519,-0.2175878286,0.1347346306,0.0532117598,0.291143328,0.1250610203,-0.0492140464,0.0266518034,-0.0314312987,0.1974624097,0.2514950335,-0.0322238952,0.0891965702,0.1261572838,0.157536611,0.5037890673,0.053735029,0.2031769902,0.0966271609,-0.2018554062,0.0816629007,-0.0702104494,0.5441356897,-0.1696140319,-0.5968497992,0.4957826138,-0.4942743778,-0.0085916603,0.2324092388,0.4670405984,-0.1604339182,0.0753958225,0.251483798,-0.1050434783,0.033637356,0.03451325,-0.0803707466,0.2207057327,0.1228926927,0.2221151143,-0.1487662196,-0.4933043122,0.1599145234,-0.1897882819,0.2903176546,-0.6953832507,-0.2288500816,-0.366612047,-0.0604084767,0.1137853563,-0.1019294262,-0.1594460607,0.2681770623,0.0848985463,-0.013057163,0.0232961178,0.2334747612,-0.3981198967,0.1256912202,0.1727491468,-0.1056274921,0.1610930264,0.0997461826,-0.0547138155,-0.037544772,0.2110037208,0.0817849934,-0.023355253,-0.5620622039,0.0634886622,0.2465737909,0.3047662675,0.0822414607,-0.0770296752,0.1512786895,-0.0502387509,0.054323215,0.4209915996,-0.0538993366,0.4664910734,-0.2711001635,0.1770830005,0.277297765,-0.3652850688,-0.0003139204,0.1826036125,-0.2375145853,0.1117426082,-0.2652013898,-0.4663338959,-0.1654590964,-0.1183409244,0.1349778473,-0.0789419264,-0.2356961071,-0.0736637264,-0.3157056272,-0.0500996858,0.2459145337,0.1108266786,0.0505057685,-0.2129725069,0.1019253507,-0.0896832198,0.2275961339,0.0058478625,-0.2316600084,0.11156746,0.4407880902,-0.0369520411,-0.0522972122,-0.1141529381,-0.116719678,0.1146957576,-0.0951117575,0.2058560997,0.2640617192,-0.0140006486,-0.222789079,0.3215104342,-0.2788105011,-0.2408593893,0.1297049522,-0.0126401894,-0.1174058467,0.0466670133,0.0669539571,-0.0052326033,-0.5102657676,0.1701666266,0.1292891949,0.047954496,-0.4098981023,-0.0345081277,-0.0219770689,0.1678383499,0.0930051282,-0.2305410653,-0.0614778958,0.4744261503,-0.0609407164,0.1962468177,-0.0178826489,-0.2281215787,-0.019167576,-0.6979162097,-0.3184904456,-0.0092402529,0.1823446751,-0.1330951899,0.4334768355,0.2950298786,0.1894212216,-0.0313181132,0.1306529492,-0.1452960372,-0.1054134294,-0.0241774917,-0.1970705837,-0.0412810817,0.4775482118,0.0791641027,-0.083314769,-0.0907501876,-0.5611173511,-0.1841323078,0.3183330894,-0.2006358951,0.0832459405,0.2185398936,0.2986201346,-0.1066382155,-0.174405396,-0.2389483154,0.4070246518,-0.0666783303,-0.0943469256,0.3965403438,-0.1829442829,0.1772942096,0.30615291,0.1469939947,0.2547303438,0.3355357349,0.0759050399,0.0548323095,-0.3406339884,0.1482571512,-0.1149554327,0.2089336812,-0.2045271844,0.2236363739,-0.2222569287,0.1700025946,-0.2625818849,-0.2926515341,-0.1523296684,-0.0743206516,0.020759508,0.3121842742,0.1147071868,0.2348940372,0.3013081253,0.2710024714,0.1796744317,0.24606888,0.0581639707,0.1148115098,-0.2809874415,-0.0713953227,0.409381628,-0.3983355463,0.437820226,0.0031497572,-0.3976996541,-0.1845485568,-0.3869085014,0.2667299807,-0.3574424982,0.4122446179,0.23498936,0.0989026725,0.2883796394,-0.0820783675,0.2549666166,-0.1147719994,-0.2736467421,0.1317070425,0.0489208996,-0.0504887998,-0.1470930874,-0.1372155547,-0.076008074,-0.1587286741,0.4361136556,0.0166942198,0.2049041837,0.3152825236,0.0969091505,0.2270750254,-0.0253577381,0.1294412464,-0.3533500433,-0.2036488354,-0.0627004877,0.0351996385,-0.2237658501,-0.0332907401,0.0955420956,0.3138401806,-0.115697816,-0.3986230791,-0.2539067566,0.2589060962,-0.4011817276,0.0286345966,-0.0587614216,0.2250745445,-0.0227604751,0.0023586736,0.1034687683,-0.1657687128,0.2917427123,0.0014571624,0.4208348393,-0.0582709499,0.4073520303,0.1907383204,0.2507788241,0.2218789309,-0.1074185222,0.2332936674,-0.0760882199,0.2300221622,0.0626940429,-0.2186158597,0.1225989982,-0.1815152466,-0.0678680167,0.2153835744,-0.2848355174,0.1016196385,-0.2425624877,0.3786380589,-0.3925980031,-0.2305805087,-0.0728006586,0.084850952,0.2980444729,-0.218303442,-0.1612262875,0.3346336484,0.1729964465,0.1649609953,0.1857609898,-0.0097199064,-0.2947641015,0.1768292785,-0.4341126978,-0.1974519044,0.2037676871,0.1779196262,0.3842518628,-0.1591307223,-0.1912715137,0.0965435058,-0.3532340825,0.2157043666,-0.1090002358,0.1710741967,-0.0191141888,-0.2454843968,0.0143333711,-0.1819860488,-0.2506864071,-0.1394036263,-0.0508332662,0.2705188692,-0.4435863197,-0.0580348633,0.9266176224,-0.0307043791,-0.2027936578,-0.0545093827,-0.3822521865,-0.3325395584,-0.2178062052,0.3268850744,0.0803599432,0.1595795751,-0.2594973445,0.054746896,0.1613204628,-0.0796455517,0.1856283695,0.0180702563,-0.1477721334,-0.2234720886,-0.0046396349,-0.0809065476,-0.2214631587,0.3577394187,0.066345498,0.1223880351,0.0403788388,-0.1450326741,0.2732554376,0.2892638147,0.4261297882,-0.2185800523,0.2583729923,-0.1215013042,0.2301147729,-0.2938078642,0.114960365,0.5577464104,0.1897827685,-0.3376097083,0.3320045173,0.1806747466,-0.3360965848,0.121775113,0.473772347,0.2906243801,-0.034681879,0.0033714068,0.0544294566,1.076790452,-0.1063169241,-0.1410206258,0.3265558183,-0.2886441648,-0.0629893243,0.1711464971,0.074469611,-0.3738822639,-0.2014196664,-0.1141918153,-0.099731572,0.182915628,-0.5049661398,-0.46861884,0.270686388,0.2902297676,0.047393661,0.1362045556,0.0351531766,0.0513243116,0.0296752714,-0.1473932117,0.075758256,-0.1771012098,0.4955222607,-0.3521041572,0.006089679,0.1425145864,-0.2339378148,-0.5489826798,0.1974165887,-0.7597703934,-0.2172420472,0.1435348392,-0.0956566334,0.2164642364,-0.0425305665,0.1211306229,0.2369874418,-0.0984941423,-0.0710498318,-0.0660440177,-0.2263087183,-0.278244406,0.080314897,0.1488425285,-0.1012301594,-0.3922503591,0.2363767028,-0.0378921032,-0.1245239452,0.2563019693,-0.1621961296,0.0216446221,-0.1495987326,-0.2880109847,0.1166077033,-0.4050696194,-0.3859924376,0.0441254005,-0.0842707083,-0.6793951392,-0.0104836412,0.2277966589,-0.2703295052,0.0709382817,0.3031373918,-0.4074920118,0.4476473927,0.2836990356,0.3070386052,-0.0456531011,-0.2508759201,-0.2475806177,-0.096215196,-0.239861846,0.1254095435,-0.1179654077,0.0117817651,-0.0179658718,0.3521892726,0.2009031028,0.130558461,-0.0421137735,-0.1326913387,-0.4322311878,0.1133480817,0.0895270556,0.3084596992,0.0070495745,0.8161113858,0.0050012395,0.1381090879,-0.2168498635,0.1537139416,-0.5475091934,0.1356402934,0.5144497156,-0.3967116475,0.2740752399,0.1211526692,0.1307330132,0.3177930415,-0.2503612041,-0.1576968282,-0.028146917,0.1464556307,0.1394556761,-0.2126519084,0.0227943938,0.170848608,0.0856941938,0.169199571,0.0142538045,-0.1359180212,0.0720585734,0.0227145944,0.5171648264,0.066288501,-0.133510977,-0.3067292571,-0.0594982728,0.2850827277,0.1547500938,-0.0466246828,-0.004913874,-0.3229970634,-0.0157679971,0.0993029922,0.2712794542,-0.0949808732,0.1953766793,-0.2645510137,-0.3805359304,-0.2151184231,0.5645757914,0.3280044198,-0.3266963065,0.086527355,-0.0304161627,0.018389836,-0.2668769658,0.1713368893,-0.1892541796,-0.1514424533,-0.0306949466,0.5003805757,-0.1047546417,0.2028654367,0.0371597931,0.213701725,-0.00022627,-0.1407934874,-0.0896461084,0.0310132988,-0.2003608495,0.1695739627,0.1094613299,-0.1566507071,0.3708167672,-0.323977381,-0.0258680321,0.3565204144,-0.5007627606,0.3787528872,0.1755802333,-0.2727207839,0.0502875932,0.2554655969,0.0517259911,0.1064760238,0.0574634559,0.5036066771,0.0453846492,-0.3039710224,-0.0631596595,0.4623815715,-0.1607019305,-0.2788872123,0.1073022485,-0.0637273565,-0.3115138113,0.134660393,0.1420977265,-0.02707389,0.0633941963,0.0979784504,-0.3087844551,-0.1646403223,-0.401157558,0.2207766622,0.2365179211,-0.2379700541,0.5151308179,0.2662872076,-0.0371031389,-0.2489867359,0.7079617977,0.3704000711,-0.044673983,-0.1711800694,0.1376731247,-0.2246428877,0.1120398939,0.2599118352,-0.0763739496,0.0478594899,-0.0139040696,0.1707173288,0.0092956806,-0.0898856148,0.1108519435,0.1339693218,-0.3126707971,-0.2695256174,0.1654578,0.014774763,-0.0065957443,-0.4888792634,0.381357491,-0.2600221634,0.6083681583,0.3633648157,-0.0167721026,0.1414662302,0.2023964226,0.0427674167,0.2453739345,0.3704001904,0.2853342891,-0.1474468559,-0.3849023879,-0.3310810626,-0.3185833395,0.2492609769,-0.0115360944,0.4620011747,0.2291846871,0.134952575,-0.2289453298,0.0339672528,0.4562621713,0.2448747754,0.0648012906,0.1269744784,-0.2102719843,0.0880673677,-0.1616372764,-0.2616662979,0.0603457205,-0.1920471936,0.1918265373,-0.2449938953,-0.0385891534,-0.0182405785,0.0576571263,-0.4193947613,-0.0711653382,0.1217793003,-0.0717778057,-0.0353250094,-0.1474374086,-0.3151301742,0.029259732,-0.0187704731,-0.0378767923,0.2944749594,0.0775456652,-0.0054252218,0.1476282179,-0.4997337461,-0.2743061185,0.0258237924,0.4406977594,-0.2658599317,-0.2488072366,0.1811849922,0.0307605229,-0.2400042415,0.132065475,0.3002210855,0.0648869798,0.091827713,-0.559091866,-0.1621075869,0.6869646907,0.3171429932,0.0053969757,-0.1448841244,0.1080972478,0.214095071,-0.2759179175,-0.8308287263,0.0773528218,0.1110792533,0.0343419835,-0.3358500004,0.4130390882,-0.396550715,0.0305419471,0.146956265,0.2626532614,0.2095080614,0.0917727649,0.3162976801,-0.1423849612]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1286","title":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted","comments":"I just checked and the wmt16 en-ro doesn't have empty lines\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nd = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nlen(d) # 610320\r\nlen(d.filter(lambda x: len(x[\"translation\"][\"en\"].strip()) > 0)) # 610320\r\nlen(d.filter(lambda x: len(x[\"translation\"][\"ro\"].strip()) > 0)) # 610320\r\n# also tested for split=\"validation\" and \"test\"\r\n```\r\n\r\nCan you open an issue on the `transformers` repo ? also cc @sgugger ","body":"Hi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n","comment_length":59,"text":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted\nHi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n\nI just checked and the wmt16 en-ro doesn't have empty lines\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nd = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nlen(d) # 610320\r\nlen(d.filter(lambda x: len(x[\"translation\"][\"en\"].strip()) > 0)) # 610320\r\nlen(d.filter(lambda x: len(x[\"translation\"][\"ro\"].strip()) > 0)) # 610320\r\n# also tested for split=\"validation\" and \"test\"\r\n```\r\n\r\nCan you open an issue on the `transformers` repo ? also cc @sgugger ","embeddings":[-0.0197352897,-0.7191827893,0.0039322353,0.349368006,0.484241426,0.0076734996,0.2424444556,0.1216879562,-0.2513463497,0.3279392719,-0.0821483955,-0.1225553527,-0.1213336438,0.3986013532,0.024262283,-0.2556790411,-0.0995945483,0.0782158896,-0.3232808411,0.1850418597,-0.0975957587,0.2666340768,-0.3790644109,0.2954924107,-0.077346921,0.0884474367,0.1073483229,0.0845447779,-0.0165824965,-0.3517342508,0.0590415671,-0.0919298232,-0.3089157045,0.3623276055,-0.0001231714,-0.1029601544,0.2742366493,0.044334691,-0.0298559833,-0.1064275429,0.264336437,0.5919190049,-0.1285633147,0.155628711,-0.1199958175,0.0390061848,0.131960839,-0.02621934,0.2547330558,0.2040431052,0.1139890254,0.3572780192,0.4076769948,0.0077055944,-0.0644371659,-0.047120966,-0.2066649497,0.1545071602,0.0657213703,-0.1996357739,0.1090153903,-0.0004496211,0.1906781793,-0.0334959961,-0.2198651433,-0.3123973608,-0.0370121896,-0.457115531,0.0150993029,0.4477550387,-0.0995786712,-0.2989667356,-0.1321324855,0.2411860675,0.0361542404,-0.7871572971,0.0136458101,0.1679371297,-0.0502248146,0.2885320783,-0.137509048,0.1263263226,-0.2918356359,0.0164490547,-0.0907559618,0.296466738,-0.1985564381,0.2684500217,0.3229812086,0.1940049678,-0.3849190474,0.1892252117,0.0081450799,0.1069084778,-0.3058028519,-0.2175878286,0.1347346306,0.0532117598,0.291143328,0.1250610203,-0.0492140464,0.0266518034,-0.0314312987,0.1974624097,0.2514950335,-0.0322238952,0.0891965702,0.1261572838,0.157536611,0.5037890673,0.053735029,0.2031769902,0.0966271609,-0.2018554062,0.0816629007,-0.0702104494,0.5441356897,-0.1696140319,-0.5968497992,0.4957826138,-0.4942743778,-0.0085916603,0.2324092388,0.4670405984,-0.1604339182,0.0753958225,0.251483798,-0.1050434783,0.033637356,0.03451325,-0.0803707466,0.2207057327,0.1228926927,0.2221151143,-0.1487662196,-0.4933043122,0.1599145234,-0.1897882819,0.2903176546,-0.6953832507,-0.2288500816,-0.366612047,-0.0604084767,0.1137853563,-0.1019294262,-0.1594460607,0.2681770623,0.0848985463,-0.013057163,0.0232961178,0.2334747612,-0.3981198967,0.1256912202,0.1727491468,-0.1056274921,0.1610930264,0.0997461826,-0.0547138155,-0.037544772,0.2110037208,0.0817849934,-0.023355253,-0.5620622039,0.0634886622,0.2465737909,0.3047662675,0.0822414607,-0.0770296752,0.1512786895,-0.0502387509,0.054323215,0.4209915996,-0.0538993366,0.4664910734,-0.2711001635,0.1770830005,0.277297765,-0.3652850688,-0.0003139204,0.1826036125,-0.2375145853,0.1117426082,-0.2652013898,-0.4663338959,-0.1654590964,-0.1183409244,0.1349778473,-0.0789419264,-0.2356961071,-0.0736637264,-0.3157056272,-0.0500996858,0.2459145337,0.1108266786,0.0505057685,-0.2129725069,0.1019253507,-0.0896832198,0.2275961339,0.0058478625,-0.2316600084,0.11156746,0.4407880902,-0.0369520411,-0.0522972122,-0.1141529381,-0.116719678,0.1146957576,-0.0951117575,0.2058560997,0.2640617192,-0.0140006486,-0.222789079,0.3215104342,-0.2788105011,-0.2408593893,0.1297049522,-0.0126401894,-0.1174058467,0.0466670133,0.0669539571,-0.0052326033,-0.5102657676,0.1701666266,0.1292891949,0.047954496,-0.4098981023,-0.0345081277,-0.0219770689,0.1678383499,0.0930051282,-0.2305410653,-0.0614778958,0.4744261503,-0.0609407164,0.1962468177,-0.0178826489,-0.2281215787,-0.019167576,-0.6979162097,-0.3184904456,-0.0092402529,0.1823446751,-0.1330951899,0.4334768355,0.2950298786,0.1894212216,-0.0313181132,0.1306529492,-0.1452960372,-0.1054134294,-0.0241774917,-0.1970705837,-0.0412810817,0.4775482118,0.0791641027,-0.083314769,-0.0907501876,-0.5611173511,-0.1841323078,0.3183330894,-0.2006358951,0.0832459405,0.2185398936,0.2986201346,-0.1066382155,-0.174405396,-0.2389483154,0.4070246518,-0.0666783303,-0.0943469256,0.3965403438,-0.1829442829,0.1772942096,0.30615291,0.1469939947,0.2547303438,0.3355357349,0.0759050399,0.0548323095,-0.3406339884,0.1482571512,-0.1149554327,0.2089336812,-0.2045271844,0.2236363739,-0.2222569287,0.1700025946,-0.2625818849,-0.2926515341,-0.1523296684,-0.0743206516,0.020759508,0.3121842742,0.1147071868,0.2348940372,0.3013081253,0.2710024714,0.1796744317,0.24606888,0.0581639707,0.1148115098,-0.2809874415,-0.0713953227,0.409381628,-0.3983355463,0.437820226,0.0031497572,-0.3976996541,-0.1845485568,-0.3869085014,0.2667299807,-0.3574424982,0.4122446179,0.23498936,0.0989026725,0.2883796394,-0.0820783675,0.2549666166,-0.1147719994,-0.2736467421,0.1317070425,0.0489208996,-0.0504887998,-0.1470930874,-0.1372155547,-0.076008074,-0.1587286741,0.4361136556,0.0166942198,0.2049041837,0.3152825236,0.0969091505,0.2270750254,-0.0253577381,0.1294412464,-0.3533500433,-0.2036488354,-0.0627004877,0.0351996385,-0.2237658501,-0.0332907401,0.0955420956,0.3138401806,-0.115697816,-0.3986230791,-0.2539067566,0.2589060962,-0.4011817276,0.0286345966,-0.0587614216,0.2250745445,-0.0227604751,0.0023586736,0.1034687683,-0.1657687128,0.2917427123,0.0014571624,0.4208348393,-0.0582709499,0.4073520303,0.1907383204,0.2507788241,0.2218789309,-0.1074185222,0.2332936674,-0.0760882199,0.2300221622,0.0626940429,-0.2186158597,0.1225989982,-0.1815152466,-0.0678680167,0.2153835744,-0.2848355174,0.1016196385,-0.2425624877,0.3786380589,-0.3925980031,-0.2305805087,-0.0728006586,0.084850952,0.2980444729,-0.218303442,-0.1612262875,0.3346336484,0.1729964465,0.1649609953,0.1857609898,-0.0097199064,-0.2947641015,0.1768292785,-0.4341126978,-0.1974519044,0.2037676871,0.1779196262,0.3842518628,-0.1591307223,-0.1912715137,0.0965435058,-0.3532340825,0.2157043666,-0.1090002358,0.1710741967,-0.0191141888,-0.2454843968,0.0143333711,-0.1819860488,-0.2506864071,-0.1394036263,-0.0508332662,0.2705188692,-0.4435863197,-0.0580348633,0.9266176224,-0.0307043791,-0.2027936578,-0.0545093827,-0.3822521865,-0.3325395584,-0.2178062052,0.3268850744,0.0803599432,0.1595795751,-0.2594973445,0.054746896,0.1613204628,-0.0796455517,0.1856283695,0.0180702563,-0.1477721334,-0.2234720886,-0.0046396349,-0.0809065476,-0.2214631587,0.3577394187,0.066345498,0.1223880351,0.0403788388,-0.1450326741,0.2732554376,0.2892638147,0.4261297882,-0.2185800523,0.2583729923,-0.1215013042,0.2301147729,-0.2938078642,0.114960365,0.5577464104,0.1897827685,-0.3376097083,0.3320045173,0.1806747466,-0.3360965848,0.121775113,0.473772347,0.2906243801,-0.034681879,0.0033714068,0.0544294566,1.076790452,-0.1063169241,-0.1410206258,0.3265558183,-0.2886441648,-0.0629893243,0.1711464971,0.074469611,-0.3738822639,-0.2014196664,-0.1141918153,-0.099731572,0.182915628,-0.5049661398,-0.46861884,0.270686388,0.2902297676,0.047393661,0.1362045556,0.0351531766,0.0513243116,0.0296752714,-0.1473932117,0.075758256,-0.1771012098,0.4955222607,-0.3521041572,0.006089679,0.1425145864,-0.2339378148,-0.5489826798,0.1974165887,-0.7597703934,-0.2172420472,0.1435348392,-0.0956566334,0.2164642364,-0.0425305665,0.1211306229,0.2369874418,-0.0984941423,-0.0710498318,-0.0660440177,-0.2263087183,-0.278244406,0.080314897,0.1488425285,-0.1012301594,-0.3922503591,0.2363767028,-0.0378921032,-0.1245239452,0.2563019693,-0.1621961296,0.0216446221,-0.1495987326,-0.2880109847,0.1166077033,-0.4050696194,-0.3859924376,0.0441254005,-0.0842707083,-0.6793951392,-0.0104836412,0.2277966589,-0.2703295052,0.0709382817,0.3031373918,-0.4074920118,0.4476473927,0.2836990356,0.3070386052,-0.0456531011,-0.2508759201,-0.2475806177,-0.096215196,-0.239861846,0.1254095435,-0.1179654077,0.0117817651,-0.0179658718,0.3521892726,0.2009031028,0.130558461,-0.0421137735,-0.1326913387,-0.4322311878,0.1133480817,0.0895270556,0.3084596992,0.0070495745,0.8161113858,0.0050012395,0.1381090879,-0.2168498635,0.1537139416,-0.5475091934,0.1356402934,0.5144497156,-0.3967116475,0.2740752399,0.1211526692,0.1307330132,0.3177930415,-0.2503612041,-0.1576968282,-0.028146917,0.1464556307,0.1394556761,-0.2126519084,0.0227943938,0.170848608,0.0856941938,0.169199571,0.0142538045,-0.1359180212,0.0720585734,0.0227145944,0.5171648264,0.066288501,-0.133510977,-0.3067292571,-0.0594982728,0.2850827277,0.1547500938,-0.0466246828,-0.004913874,-0.3229970634,-0.0157679971,0.0993029922,0.2712794542,-0.0949808732,0.1953766793,-0.2645510137,-0.3805359304,-0.2151184231,0.5645757914,0.3280044198,-0.3266963065,0.086527355,-0.0304161627,0.018389836,-0.2668769658,0.1713368893,-0.1892541796,-0.1514424533,-0.0306949466,0.5003805757,-0.1047546417,0.2028654367,0.0371597931,0.213701725,-0.00022627,-0.1407934874,-0.0896461084,0.0310132988,-0.2003608495,0.1695739627,0.1094613299,-0.1566507071,0.3708167672,-0.323977381,-0.0258680321,0.3565204144,-0.5007627606,0.3787528872,0.1755802333,-0.2727207839,0.0502875932,0.2554655969,0.0517259911,0.1064760238,0.0574634559,0.5036066771,0.0453846492,-0.3039710224,-0.0631596595,0.4623815715,-0.1607019305,-0.2788872123,0.1073022485,-0.0637273565,-0.3115138113,0.134660393,0.1420977265,-0.02707389,0.0633941963,0.0979784504,-0.3087844551,-0.1646403223,-0.401157558,0.2207766622,0.2365179211,-0.2379700541,0.5151308179,0.2662872076,-0.0371031389,-0.2489867359,0.7079617977,0.3704000711,-0.044673983,-0.1711800694,0.1376731247,-0.2246428877,0.1120398939,0.2599118352,-0.0763739496,0.0478594899,-0.0139040696,0.1707173288,0.0092956806,-0.0898856148,0.1108519435,0.1339693218,-0.3126707971,-0.2695256174,0.1654578,0.014774763,-0.0065957443,-0.4888792634,0.381357491,-0.2600221634,0.6083681583,0.3633648157,-0.0167721026,0.1414662302,0.2023964226,0.0427674167,0.2453739345,0.3704001904,0.2853342891,-0.1474468559,-0.3849023879,-0.3310810626,-0.3185833395,0.2492609769,-0.0115360944,0.4620011747,0.2291846871,0.134952575,-0.2289453298,0.0339672528,0.4562621713,0.2448747754,0.0648012906,0.1269744784,-0.2102719843,0.0880673677,-0.1616372764,-0.2616662979,0.0603457205,-0.1920471936,0.1918265373,-0.2449938953,-0.0385891534,-0.0182405785,0.0576571263,-0.4193947613,-0.0711653382,0.1217793003,-0.0717778057,-0.0353250094,-0.1474374086,-0.3151301742,0.029259732,-0.0187704731,-0.0378767923,0.2944749594,0.0775456652,-0.0054252218,0.1476282179,-0.4997337461,-0.2743061185,0.0258237924,0.4406977594,-0.2658599317,-0.2488072366,0.1811849922,0.0307605229,-0.2400042415,0.132065475,0.3002210855,0.0648869798,0.091827713,-0.559091866,-0.1621075869,0.6869646907,0.3171429932,0.0053969757,-0.1448841244,0.1080972478,0.214095071,-0.2759179175,-0.8308287263,0.0773528218,0.1110792533,0.0343419835,-0.3358500004,0.4130390882,-0.396550715,0.0305419471,0.146956265,0.2626532614,0.2095080614,0.0917727649,0.3162976801,-0.1423849612]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1286","title":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted","comments":"Hi @lhoestq \r\nI am not really sure which part is causing this, to me this is more related to dataset library as this is happening for some of the datassets below please find the information to reprodcue the bug, this is really blocking me and I appreciate your help\r\n\r\n\r\n## Environment info\r\n- `transformers` version: 3.5.1\r\n- Platform: GPU\r\n- Python version: 3.7 \r\n- PyTorch version (GPU?): 1.0.4\r\n- Tensorflow version (GPU?): - \r\n- Using GPU in script?: - \r\n- Using distributed or parallel set-up in script?: - \r\n\r\n### Who can help\r\n tokenizers: @mfuntowicz\r\n Trainer: @sgugger\r\n TextGeneration: @TevenLeScao \r\n nlp datasets: [different repo](https:\/\/github.com\/huggingface\/nlp)\r\n rust tokenizers: [different repo](https:\/\/github.com\/huggingface\/tokenizers)\r\n examples\/seq2seq: @patil-suraj\r\n\r\n## Information\r\nHi\r\nI am testing seq2seq model with T5 on different datasets and this is always getting the following bug, this is really blocking me as this fails for many datasets. could you have a look please? thanks \r\n\r\n```\r\n[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n\r\n```\r\n\r\nTo reproduce the error please run on 1 GPU:\r\n```\r\ngit clone git@github.com:rabeehk\/debug-seq2seq.git\r\npython setup.py develop \r\ncd seq2seq \r\npython finetune_t5_trainer.py temp.json\r\n\r\n```\r\n\r\nFull output of the program:\r\n\r\n```\r\n(internship) rkarimi@vgnh008:\/idiap\/user\/rkarimi\/dev\/debug-seq2seq\/seq2seq$ python finetune_t5_trainer.py temp.json \r\n2020-12-12 15:38:16.234542: W tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:60] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudart.so.11.0: cannot open shared object file: No such file or directory\r\n2020-12-12 15:38:16.234598: I tensorflow\/stream_executor\/cuda\/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\n12\/12\/2020 15:38:32 - WARNING - __main__ - Process rank: -1, device: cuda:0, n_gpu: 1, distributed training: False, 16-bits training: False\r\n12\/12\/2020 15:38:32 - INFO - __main__ - Training\/evaluation parameters Seq2SeqTrainingArguments(output_dir='outputs\/test', overwrite_output_dir=True, do_train=True, do_eval=True, do_predict=False, evaluate_during_training=False, evaluation_strategy=<EvaluationStrategy.NO: 'no'>, prediction_loss_only=False, per_device_train_batch_size=64, per_device_eval_batch_size=64, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=1, eval_accumulation_steps=None, learning_rate=0.01, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=2, max_steps=-1, warmup_steps=500, logging_dir='runs\/Dec12_15-38-32_vgnh008', logging_first_step=True, logging_steps=200, save_steps=200, save_total_limit=1, no_cuda=False, seed=42, fp16=False, fp16_opt_level='O1', local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=200, dataloader_num_workers=0, past_index=-1, run_name='outputs\/test', disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=None, label_smoothing=0.1, sortish_sampler=False, predict_with_generate=True, adafactor=False, encoder_layerdrop=None, decoder_layerdrop=None, dropout=None, attention_dropout=None, lr_scheduler='linear', fixed_length_emb=None, encoder_projection=None, encoder_pooling=None, projection_length=None, only_projection_bottleneck=False, concat_projection_token=False, gcs_bucket='ruse-xcloud-bucket', temperature=10, train_adapters=True, do_finetune=True, parametric_task_embedding=False, eval_output_dir='outputs\/finetune-adapter\/test-n-1-lr-1e-02-e-20')\r\nSome weights of T5ForConditionalGeneration were not initialized from the model checkpoint at t5-small and are newly initialized: ['encoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.0.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.0.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.0.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.0.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.1.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.1.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.1.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.1.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.2.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.2.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.2.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.2.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.3.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.3.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.3.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.3.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.4.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.4.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.4.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.4.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.5.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.5.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.5.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.5.layer.1.adapter_controller.post_layer_norm.bias', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.0.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.0.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.0.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.0.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.1.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.1.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.1.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.1.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.2.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.2.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.2.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.2.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.3.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.3.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.3.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.3.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.4.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.4.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.4.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.4.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.5.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.5.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.5.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.5.layer.2.adapter_controller.post_layer_norm.bias']\r\nYou should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140079090376272 acquired on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140079090376272 released on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\nUsing custom data configuration default\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140082549312272 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140082549312272 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140082549365648 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nReusing dataset boolq (\/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534)\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140082549365648 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nLoading cached processed dataset at \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534\/cache-6810ece2a440c3be.arrow\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549560848 acquired on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549560848 released on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\nUsing custom data configuration default\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549560848 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549560848 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549365200 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nReusing dataset boolq (\/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534)\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549365200 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nLoading cached processed dataset at \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534\/cache-9a2822394a3a4e34.arrow\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.metrics.metrics - selected metric <function build_compute_metrics_fn.<locals>.classification_metrics at 0x7f66b464cc20> for task boolq\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - ***** Running training *****\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Num examples = 10\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Num Epochs = 2\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Instantaneous batch size per device = 64\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Total train batch size (w. parallel, distributed & accumulation) = 64\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Gradient Accumulation steps = 1\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Total optimization steps = 2\r\n{'loss': 529.79443359375, 'learning_rate': 2e-05, 'epoch': 1.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 2.37it\/s]12\/12\/2020 15:38:46 - INFO - seq2seq.trainers.trainer - \r\n\r\nTraining completed. Do not forget to share your model on huggingface.co\/models =)\r\n\r\n\r\n{'epoch': 2.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 2.43it\/s]\r\n12\/12\/2020 15:38:46 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/test\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079084929680 acquired on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079084929680 released on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\nUsing custom data configuration default\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079084929360 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079084929360 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079085355216 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nReusing dataset boolq (\/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534)\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079085355216 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nLoading cached processed dataset at \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534\/cache-164dd1d57e9fa69a.arrow\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.metrics.metrics - selected metric <function build_compute_metrics_fn.<locals>.classification_metrics at 0x7f66b40c67a0> for task boolq\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - ***** Running training *****\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Num examples = 1\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Num Epochs = 2\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Instantaneous batch size per device = 64\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Total train batch size (w. parallel, distributed & accumulation) = 64\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Gradient Accumulation steps = 1\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Total optimization steps = 2\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Continuing training from checkpoint, will skip to saved global_step\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Continuing training from epoch 2\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Continuing training from global step 2\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Will skip the first 0 steps in the first epoch\r\n 0%| | 0\/2 [00:00<?, ?it\/s]12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - \r\n\r\nTraining completed. Do not forget to share your model on huggingface.co\/models =)\r\n\r\n\r\n{'epoch': 2.0} \r\n 0%| | 0\/2 [00:00<?, ?it\/s]\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/finetune-adapter\/test-n-1-lr-1e-02-e-20\/boolq\r\n12\/12\/2020 15:39:07 - INFO - seq2seq.utils.utils - using task specific params for boolq: {'max_length': 3}\r\n12\/12\/2020 15:39:07 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/12\/2020 15:39:07 - INFO - seq2seq.trainers.trainer - Num examples = 3269\r\n12\/12\/2020 15:39:07 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 52\/52 [00:12<00:00, 4.86it\/s][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n```\r\n\r\n\r\n\r\n","body":"Hi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n","comment_length":1524,"text":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted\nHi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n\nHi @lhoestq \r\nI am not really sure which part is causing this, to me this is more related to dataset library as this is happening for some of the datassets below please find the information to reprodcue the bug, this is really blocking me and I appreciate your help\r\n\r\n\r\n## Environment info\r\n- `transformers` version: 3.5.1\r\n- Platform: GPU\r\n- Python version: 3.7 \r\n- PyTorch version (GPU?): 1.0.4\r\n- Tensorflow version (GPU?): - \r\n- Using GPU in script?: - \r\n- Using distributed or parallel set-up in script?: - \r\n\r\n### Who can help\r\n tokenizers: @mfuntowicz\r\n Trainer: @sgugger\r\n TextGeneration: @TevenLeScao \r\n nlp datasets: [different repo](https:\/\/github.com\/huggingface\/nlp)\r\n rust tokenizers: [different repo](https:\/\/github.com\/huggingface\/tokenizers)\r\n examples\/seq2seq: @patil-suraj\r\n\r\n## Information\r\nHi\r\nI am testing seq2seq model with T5 on different datasets and this is always getting the following bug, this is really blocking me as this fails for many datasets. could you have a look please? thanks \r\n\r\n```\r\n[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n\r\n```\r\n\r\nTo reproduce the error please run on 1 GPU:\r\n```\r\ngit clone git@github.com:rabeehk\/debug-seq2seq.git\r\npython setup.py develop \r\ncd seq2seq \r\npython finetune_t5_trainer.py temp.json\r\n\r\n```\r\n\r\nFull output of the program:\r\n\r\n```\r\n(internship) rkarimi@vgnh008:\/idiap\/user\/rkarimi\/dev\/debug-seq2seq\/seq2seq$ python finetune_t5_trainer.py temp.json \r\n2020-12-12 15:38:16.234542: W tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:60] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudart.so.11.0: cannot open shared object file: No such file or directory\r\n2020-12-12 15:38:16.234598: I tensorflow\/stream_executor\/cuda\/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\n12\/12\/2020 15:38:32 - WARNING - __main__ - Process rank: -1, device: cuda:0, n_gpu: 1, distributed training: False, 16-bits training: False\r\n12\/12\/2020 15:38:32 - INFO - __main__ - Training\/evaluation parameters Seq2SeqTrainingArguments(output_dir='outputs\/test', overwrite_output_dir=True, do_train=True, do_eval=True, do_predict=False, evaluate_during_training=False, evaluation_strategy=<EvaluationStrategy.NO: 'no'>, prediction_loss_only=False, per_device_train_batch_size=64, per_device_eval_batch_size=64, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=1, eval_accumulation_steps=None, learning_rate=0.01, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=2, max_steps=-1, warmup_steps=500, logging_dir='runs\/Dec12_15-38-32_vgnh008', logging_first_step=True, logging_steps=200, save_steps=200, save_total_limit=1, no_cuda=False, seed=42, fp16=False, fp16_opt_level='O1', local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=200, dataloader_num_workers=0, past_index=-1, run_name='outputs\/test', disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=None, label_smoothing=0.1, sortish_sampler=False, predict_with_generate=True, adafactor=False, encoder_layerdrop=None, decoder_layerdrop=None, dropout=None, attention_dropout=None, lr_scheduler='linear', fixed_length_emb=None, encoder_projection=None, encoder_pooling=None, projection_length=None, only_projection_bottleneck=False, concat_projection_token=False, gcs_bucket='ruse-xcloud-bucket', temperature=10, train_adapters=True, do_finetune=True, parametric_task_embedding=False, eval_output_dir='outputs\/finetune-adapter\/test-n-1-lr-1e-02-e-20')\r\nSome weights of T5ForConditionalGeneration were not initialized from the model checkpoint at t5-small and are newly initialized: ['encoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.0.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.0.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.0.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.0.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.0.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.0.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.1.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.1.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.1.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.1.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.1.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.1.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.2.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.2.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.2.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.2.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.2.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.2.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.3.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.3.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.3.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.3.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.3.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.3.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.4.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.4.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.4.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.4.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.4.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.4.layer.1.adapter_controller.post_layer_norm.bias', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.5.layer.0.adapter_controller.post_layer_norm.weight', 'encoder.block.5.layer.0.adapter_controller.post_layer_norm.bias', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'encoder.block.5.layer.1.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'encoder.block.5.layer.1.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'encoder.block.5.layer.1.adapter_controller.post_layer_norm.weight', 'encoder.block.5.layer.1.adapter_controller.post_layer_norm.bias', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.0.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.0.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.0.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.0.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.0.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.0.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.0.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.0.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.1.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.1.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.1.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.1.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.1.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.1.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.1.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.1.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.2.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.2.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.2.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.2.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.2.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.2.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.2.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.2.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.3.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.3.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.3.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.3.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.3.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.3.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.3.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.3.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.4.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.4.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.4.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.4.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.4.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.4.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.4.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.4.layer.2.adapter_controller.post_layer_norm.bias', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.5.layer.0.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.5.layer.0.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.5.layer.0.adapter_controller.post_layer_norm.weight', 'decoder.block.5.layer.0.adapter_controller.post_layer_norm.bias', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.weight', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.weight_generator.0.bias', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.weight', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.weight_generator.1.bias', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.weight', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.bias_generator.0.bias', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.weight', 'decoder.block.5.layer.2.adapter_controller.meta_up_sampler.bias_generator.1.bias', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.weight', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.weight_generator.0.bias', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.weight', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.weight_generator.1.bias', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.weight', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.bias_generator.0.bias', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.weight', 'decoder.block.5.layer.2.adapter_controller.meta_down_sampler.bias_generator.1.bias', 'decoder.block.5.layer.2.adapter_controller.post_layer_norm.weight', 'decoder.block.5.layer.2.adapter_controller.post_layer_norm.bias']\r\nYou should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140079090376272 acquired on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140079090376272 released on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\nUsing custom data configuration default\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140082549312272 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140082549312272 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140082549365648 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nReusing dataset boolq (\/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534)\r\n12\/12\/2020 15:38:44 - INFO - filelock - Lock 140082549365648 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nLoading cached processed dataset at \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534\/cache-6810ece2a440c3be.arrow\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549560848 acquired on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549560848 released on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\nUsing custom data configuration default\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549560848 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549560848 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549365200 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nReusing dataset boolq (\/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534)\r\n12\/12\/2020 15:38:45 - INFO - filelock - Lock 140082549365200 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nLoading cached processed dataset at \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534\/cache-9a2822394a3a4e34.arrow\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.metrics.metrics - selected metric <function build_compute_metrics_fn.<locals>.classification_metrics at 0x7f66b464cc20> for task boolq\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - ***** Running training *****\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Num examples = 10\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Num Epochs = 2\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Instantaneous batch size per device = 64\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Total train batch size (w. parallel, distributed & accumulation) = 64\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Gradient Accumulation steps = 1\r\n12\/12\/2020 15:38:45 - INFO - seq2seq.trainers.trainer - Total optimization steps = 2\r\n{'loss': 529.79443359375, 'learning_rate': 2e-05, 'epoch': 1.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 2.37it\/s]12\/12\/2020 15:38:46 - INFO - seq2seq.trainers.trainer - \r\n\r\nTraining completed. Do not forget to share your model on huggingface.co\/models =)\r\n\r\n\r\n{'epoch': 2.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 2.43it\/s]\r\n12\/12\/2020 15:38:46 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/test\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079084929680 acquired on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079084929680 released on \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/4c7b1146606607c193d1ef601d8d0c134521b2ac59f61ee98c09119be925ee16.7ad892de9d7f1b4f9dfc598ef31e4a398a7224176bc9a3110e0e2075ff943e8f.py.lock\r\nUsing custom data configuration default\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079084929360 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079084929360 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079085355216 acquired on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nReusing dataset boolq (\/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534)\r\n12\/12\/2020 15:38:59 - INFO - filelock - Lock 140079085355216 released on \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/_idiap_temp_rkarimi_cache_home_1_datasets_boolq_default_0.1.0_1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534.lock\r\nLoading cached processed dataset at \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/default\/0.1.0\/1fcfdc6f36dc89a2245ffbbd5248ab33890594b50396731ebc78411bdd2ca534\/cache-164dd1d57e9fa69a.arrow\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.metrics.metrics - selected metric <function build_compute_metrics_fn.<locals>.classification_metrics at 0x7f66b40c67a0> for task boolq\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - ***** Running training *****\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Num examples = 1\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Num Epochs = 2\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Instantaneous batch size per device = 64\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Total train batch size (w. parallel, distributed & accumulation) = 64\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Gradient Accumulation steps = 1\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Total optimization steps = 2\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Continuing training from checkpoint, will skip to saved global_step\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Continuing training from epoch 2\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Continuing training from global step 2\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Will skip the first 0 steps in the first epoch\r\n 0%| | 0\/2 [00:00<?, ?it\/s]12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - \r\n\r\nTraining completed. Do not forget to share your model on huggingface.co\/models =)\r\n\r\n\r\n{'epoch': 2.0} \r\n 0%| | 0\/2 [00:00<?, ?it\/s]\r\n12\/12\/2020 15:38:59 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/finetune-adapter\/test-n-1-lr-1e-02-e-20\/boolq\r\n12\/12\/2020 15:39:07 - INFO - seq2seq.utils.utils - using task specific params for boolq: {'max_length': 3}\r\n12\/12\/2020 15:39:07 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/12\/2020 15:39:07 - INFO - seq2seq.trainers.trainer - Num examples = 3269\r\n12\/12\/2020 15:39:07 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 52\/52 [00:12<00:00, 4.86it\/s][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n```\r\n\r\n\r\n\r\n","embeddings":[-0.0197352897,-0.7191827893,0.0039322353,0.349368006,0.484241426,0.0076734996,0.2424444556,0.1216879562,-0.2513463497,0.3279392719,-0.0821483955,-0.1225553527,-0.1213336438,0.3986013532,0.024262283,-0.2556790411,-0.0995945483,0.0782158896,-0.3232808411,0.1850418597,-0.0975957587,0.2666340768,-0.3790644109,0.2954924107,-0.077346921,0.0884474367,0.1073483229,0.0845447779,-0.0165824965,-0.3517342508,0.0590415671,-0.0919298232,-0.3089157045,0.3623276055,-0.0001231714,-0.1029601544,0.2742366493,0.044334691,-0.0298559833,-0.1064275429,0.264336437,0.5919190049,-0.1285633147,0.155628711,-0.1199958175,0.0390061848,0.131960839,-0.02621934,0.2547330558,0.2040431052,0.1139890254,0.3572780192,0.4076769948,0.0077055944,-0.0644371659,-0.047120966,-0.2066649497,0.1545071602,0.0657213703,-0.1996357739,0.1090153903,-0.0004496211,0.1906781793,-0.0334959961,-0.2198651433,-0.3123973608,-0.0370121896,-0.457115531,0.0150993029,0.4477550387,-0.0995786712,-0.2989667356,-0.1321324855,0.2411860675,0.0361542404,-0.7871572971,0.0136458101,0.1679371297,-0.0502248146,0.2885320783,-0.137509048,0.1263263226,-0.2918356359,0.0164490547,-0.0907559618,0.296466738,-0.1985564381,0.2684500217,0.3229812086,0.1940049678,-0.3849190474,0.1892252117,0.0081450799,0.1069084778,-0.3058028519,-0.2175878286,0.1347346306,0.0532117598,0.291143328,0.1250610203,-0.0492140464,0.0266518034,-0.0314312987,0.1974624097,0.2514950335,-0.0322238952,0.0891965702,0.1261572838,0.157536611,0.5037890673,0.053735029,0.2031769902,0.0966271609,-0.2018554062,0.0816629007,-0.0702104494,0.5441356897,-0.1696140319,-0.5968497992,0.4957826138,-0.4942743778,-0.0085916603,0.2324092388,0.4670405984,-0.1604339182,0.0753958225,0.251483798,-0.1050434783,0.033637356,0.03451325,-0.0803707466,0.2207057327,0.1228926927,0.2221151143,-0.1487662196,-0.4933043122,0.1599145234,-0.1897882819,0.2903176546,-0.6953832507,-0.2288500816,-0.366612047,-0.0604084767,0.1137853563,-0.1019294262,-0.1594460607,0.2681770623,0.0848985463,-0.013057163,0.0232961178,0.2334747612,-0.3981198967,0.1256912202,0.1727491468,-0.1056274921,0.1610930264,0.0997461826,-0.0547138155,-0.037544772,0.2110037208,0.0817849934,-0.023355253,-0.5620622039,0.0634886622,0.2465737909,0.3047662675,0.0822414607,-0.0770296752,0.1512786895,-0.0502387509,0.054323215,0.4209915996,-0.0538993366,0.4664910734,-0.2711001635,0.1770830005,0.277297765,-0.3652850688,-0.0003139204,0.1826036125,-0.2375145853,0.1117426082,-0.2652013898,-0.4663338959,-0.1654590964,-0.1183409244,0.1349778473,-0.0789419264,-0.2356961071,-0.0736637264,-0.3157056272,-0.0500996858,0.2459145337,0.1108266786,0.0505057685,-0.2129725069,0.1019253507,-0.0896832198,0.2275961339,0.0058478625,-0.2316600084,0.11156746,0.4407880902,-0.0369520411,-0.0522972122,-0.1141529381,-0.116719678,0.1146957576,-0.0951117575,0.2058560997,0.2640617192,-0.0140006486,-0.222789079,0.3215104342,-0.2788105011,-0.2408593893,0.1297049522,-0.0126401894,-0.1174058467,0.0466670133,0.0669539571,-0.0052326033,-0.5102657676,0.1701666266,0.1292891949,0.047954496,-0.4098981023,-0.0345081277,-0.0219770689,0.1678383499,0.0930051282,-0.2305410653,-0.0614778958,0.4744261503,-0.0609407164,0.1962468177,-0.0178826489,-0.2281215787,-0.019167576,-0.6979162097,-0.3184904456,-0.0092402529,0.1823446751,-0.1330951899,0.4334768355,0.2950298786,0.1894212216,-0.0313181132,0.1306529492,-0.1452960372,-0.1054134294,-0.0241774917,-0.1970705837,-0.0412810817,0.4775482118,0.0791641027,-0.083314769,-0.0907501876,-0.5611173511,-0.1841323078,0.3183330894,-0.2006358951,0.0832459405,0.2185398936,0.2986201346,-0.1066382155,-0.174405396,-0.2389483154,0.4070246518,-0.0666783303,-0.0943469256,0.3965403438,-0.1829442829,0.1772942096,0.30615291,0.1469939947,0.2547303438,0.3355357349,0.0759050399,0.0548323095,-0.3406339884,0.1482571512,-0.1149554327,0.2089336812,-0.2045271844,0.2236363739,-0.2222569287,0.1700025946,-0.2625818849,-0.2926515341,-0.1523296684,-0.0743206516,0.020759508,0.3121842742,0.1147071868,0.2348940372,0.3013081253,0.2710024714,0.1796744317,0.24606888,0.0581639707,0.1148115098,-0.2809874415,-0.0713953227,0.409381628,-0.3983355463,0.437820226,0.0031497572,-0.3976996541,-0.1845485568,-0.3869085014,0.2667299807,-0.3574424982,0.4122446179,0.23498936,0.0989026725,0.2883796394,-0.0820783675,0.2549666166,-0.1147719994,-0.2736467421,0.1317070425,0.0489208996,-0.0504887998,-0.1470930874,-0.1372155547,-0.076008074,-0.1587286741,0.4361136556,0.0166942198,0.2049041837,0.3152825236,0.0969091505,0.2270750254,-0.0253577381,0.1294412464,-0.3533500433,-0.2036488354,-0.0627004877,0.0351996385,-0.2237658501,-0.0332907401,0.0955420956,0.3138401806,-0.115697816,-0.3986230791,-0.2539067566,0.2589060962,-0.4011817276,0.0286345966,-0.0587614216,0.2250745445,-0.0227604751,0.0023586736,0.1034687683,-0.1657687128,0.2917427123,0.0014571624,0.4208348393,-0.0582709499,0.4073520303,0.1907383204,0.2507788241,0.2218789309,-0.1074185222,0.2332936674,-0.0760882199,0.2300221622,0.0626940429,-0.2186158597,0.1225989982,-0.1815152466,-0.0678680167,0.2153835744,-0.2848355174,0.1016196385,-0.2425624877,0.3786380589,-0.3925980031,-0.2305805087,-0.0728006586,0.084850952,0.2980444729,-0.218303442,-0.1612262875,0.3346336484,0.1729964465,0.1649609953,0.1857609898,-0.0097199064,-0.2947641015,0.1768292785,-0.4341126978,-0.1974519044,0.2037676871,0.1779196262,0.3842518628,-0.1591307223,-0.1912715137,0.0965435058,-0.3532340825,0.2157043666,-0.1090002358,0.1710741967,-0.0191141888,-0.2454843968,0.0143333711,-0.1819860488,-0.2506864071,-0.1394036263,-0.0508332662,0.2705188692,-0.4435863197,-0.0580348633,0.9266176224,-0.0307043791,-0.2027936578,-0.0545093827,-0.3822521865,-0.3325395584,-0.2178062052,0.3268850744,0.0803599432,0.1595795751,-0.2594973445,0.054746896,0.1613204628,-0.0796455517,0.1856283695,0.0180702563,-0.1477721334,-0.2234720886,-0.0046396349,-0.0809065476,-0.2214631587,0.3577394187,0.066345498,0.1223880351,0.0403788388,-0.1450326741,0.2732554376,0.2892638147,0.4261297882,-0.2185800523,0.2583729923,-0.1215013042,0.2301147729,-0.2938078642,0.114960365,0.5577464104,0.1897827685,-0.3376097083,0.3320045173,0.1806747466,-0.3360965848,0.121775113,0.473772347,0.2906243801,-0.034681879,0.0033714068,0.0544294566,1.076790452,-0.1063169241,-0.1410206258,0.3265558183,-0.2886441648,-0.0629893243,0.1711464971,0.074469611,-0.3738822639,-0.2014196664,-0.1141918153,-0.099731572,0.182915628,-0.5049661398,-0.46861884,0.270686388,0.2902297676,0.047393661,0.1362045556,0.0351531766,0.0513243116,0.0296752714,-0.1473932117,0.075758256,-0.1771012098,0.4955222607,-0.3521041572,0.006089679,0.1425145864,-0.2339378148,-0.5489826798,0.1974165887,-0.7597703934,-0.2172420472,0.1435348392,-0.0956566334,0.2164642364,-0.0425305665,0.1211306229,0.2369874418,-0.0984941423,-0.0710498318,-0.0660440177,-0.2263087183,-0.278244406,0.080314897,0.1488425285,-0.1012301594,-0.3922503591,0.2363767028,-0.0378921032,-0.1245239452,0.2563019693,-0.1621961296,0.0216446221,-0.1495987326,-0.2880109847,0.1166077033,-0.4050696194,-0.3859924376,0.0441254005,-0.0842707083,-0.6793951392,-0.0104836412,0.2277966589,-0.2703295052,0.0709382817,0.3031373918,-0.4074920118,0.4476473927,0.2836990356,0.3070386052,-0.0456531011,-0.2508759201,-0.2475806177,-0.096215196,-0.239861846,0.1254095435,-0.1179654077,0.0117817651,-0.0179658718,0.3521892726,0.2009031028,0.130558461,-0.0421137735,-0.1326913387,-0.4322311878,0.1133480817,0.0895270556,0.3084596992,0.0070495745,0.8161113858,0.0050012395,0.1381090879,-0.2168498635,0.1537139416,-0.5475091934,0.1356402934,0.5144497156,-0.3967116475,0.2740752399,0.1211526692,0.1307330132,0.3177930415,-0.2503612041,-0.1576968282,-0.028146917,0.1464556307,0.1394556761,-0.2126519084,0.0227943938,0.170848608,0.0856941938,0.169199571,0.0142538045,-0.1359180212,0.0720585734,0.0227145944,0.5171648264,0.066288501,-0.133510977,-0.3067292571,-0.0594982728,0.2850827277,0.1547500938,-0.0466246828,-0.004913874,-0.3229970634,-0.0157679971,0.0993029922,0.2712794542,-0.0949808732,0.1953766793,-0.2645510137,-0.3805359304,-0.2151184231,0.5645757914,0.3280044198,-0.3266963065,0.086527355,-0.0304161627,0.018389836,-0.2668769658,0.1713368893,-0.1892541796,-0.1514424533,-0.0306949466,0.5003805757,-0.1047546417,0.2028654367,0.0371597931,0.213701725,-0.00022627,-0.1407934874,-0.0896461084,0.0310132988,-0.2003608495,0.1695739627,0.1094613299,-0.1566507071,0.3708167672,-0.323977381,-0.0258680321,0.3565204144,-0.5007627606,0.3787528872,0.1755802333,-0.2727207839,0.0502875932,0.2554655969,0.0517259911,0.1064760238,0.0574634559,0.5036066771,0.0453846492,-0.3039710224,-0.0631596595,0.4623815715,-0.1607019305,-0.2788872123,0.1073022485,-0.0637273565,-0.3115138113,0.134660393,0.1420977265,-0.02707389,0.0633941963,0.0979784504,-0.3087844551,-0.1646403223,-0.401157558,0.2207766622,0.2365179211,-0.2379700541,0.5151308179,0.2662872076,-0.0371031389,-0.2489867359,0.7079617977,0.3704000711,-0.044673983,-0.1711800694,0.1376731247,-0.2246428877,0.1120398939,0.2599118352,-0.0763739496,0.0478594899,-0.0139040696,0.1707173288,0.0092956806,-0.0898856148,0.1108519435,0.1339693218,-0.3126707971,-0.2695256174,0.1654578,0.014774763,-0.0065957443,-0.4888792634,0.381357491,-0.2600221634,0.6083681583,0.3633648157,-0.0167721026,0.1414662302,0.2023964226,0.0427674167,0.2453739345,0.3704001904,0.2853342891,-0.1474468559,-0.3849023879,-0.3310810626,-0.3185833395,0.2492609769,-0.0115360944,0.4620011747,0.2291846871,0.134952575,-0.2289453298,0.0339672528,0.4562621713,0.2448747754,0.0648012906,0.1269744784,-0.2102719843,0.0880673677,-0.1616372764,-0.2616662979,0.0603457205,-0.1920471936,0.1918265373,-0.2449938953,-0.0385891534,-0.0182405785,0.0576571263,-0.4193947613,-0.0711653382,0.1217793003,-0.0717778057,-0.0353250094,-0.1474374086,-0.3151301742,0.029259732,-0.0187704731,-0.0378767923,0.2944749594,0.0775456652,-0.0054252218,0.1476282179,-0.4997337461,-0.2743061185,0.0258237924,0.4406977594,-0.2658599317,-0.2488072366,0.1811849922,0.0307605229,-0.2400042415,0.132065475,0.3002210855,0.0648869798,0.091827713,-0.559091866,-0.1621075869,0.6869646907,0.3171429932,0.0053969757,-0.1448841244,0.1080972478,0.214095071,-0.2759179175,-0.8308287263,0.0773528218,0.1110792533,0.0343419835,-0.3358500004,0.4130390882,-0.396550715,0.0305419471,0.146956265,0.2626532614,0.2095080614,0.0917727649,0.3162976801,-0.1423849612]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1286","title":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted","comments":"solved see https:\/\/github.com\/huggingface\/transformers\/issues\/9079?_pjax=%23js-repo-pjax-container ","body":"Hi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n","comment_length":3,"text":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted\nHi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n\nsolved see https:\/\/github.com\/huggingface\/transformers\/issues\/9079?_pjax=%23js-repo-pjax-container ","embeddings":[-0.0197352897,-0.7191827893,0.0039322353,0.349368006,0.484241426,0.0076734996,0.2424444556,0.1216879562,-0.2513463497,0.3279392719,-0.0821483955,-0.1225553527,-0.1213336438,0.3986013532,0.024262283,-0.2556790411,-0.0995945483,0.0782158896,-0.3232808411,0.1850418597,-0.0975957587,0.2666340768,-0.3790644109,0.2954924107,-0.077346921,0.0884474367,0.1073483229,0.0845447779,-0.0165824965,-0.3517342508,0.0590415671,-0.0919298232,-0.3089157045,0.3623276055,-0.0001231714,-0.1029601544,0.2742366493,0.044334691,-0.0298559833,-0.1064275429,0.264336437,0.5919190049,-0.1285633147,0.155628711,-0.1199958175,0.0390061848,0.131960839,-0.02621934,0.2547330558,0.2040431052,0.1139890254,0.3572780192,0.4076769948,0.0077055944,-0.0644371659,-0.047120966,-0.2066649497,0.1545071602,0.0657213703,-0.1996357739,0.1090153903,-0.0004496211,0.1906781793,-0.0334959961,-0.2198651433,-0.3123973608,-0.0370121896,-0.457115531,0.0150993029,0.4477550387,-0.0995786712,-0.2989667356,-0.1321324855,0.2411860675,0.0361542404,-0.7871572971,0.0136458101,0.1679371297,-0.0502248146,0.2885320783,-0.137509048,0.1263263226,-0.2918356359,0.0164490547,-0.0907559618,0.296466738,-0.1985564381,0.2684500217,0.3229812086,0.1940049678,-0.3849190474,0.1892252117,0.0081450799,0.1069084778,-0.3058028519,-0.2175878286,0.1347346306,0.0532117598,0.291143328,0.1250610203,-0.0492140464,0.0266518034,-0.0314312987,0.1974624097,0.2514950335,-0.0322238952,0.0891965702,0.1261572838,0.157536611,0.5037890673,0.053735029,0.2031769902,0.0966271609,-0.2018554062,0.0816629007,-0.0702104494,0.5441356897,-0.1696140319,-0.5968497992,0.4957826138,-0.4942743778,-0.0085916603,0.2324092388,0.4670405984,-0.1604339182,0.0753958225,0.251483798,-0.1050434783,0.033637356,0.03451325,-0.0803707466,0.2207057327,0.1228926927,0.2221151143,-0.1487662196,-0.4933043122,0.1599145234,-0.1897882819,0.2903176546,-0.6953832507,-0.2288500816,-0.366612047,-0.0604084767,0.1137853563,-0.1019294262,-0.1594460607,0.2681770623,0.0848985463,-0.013057163,0.0232961178,0.2334747612,-0.3981198967,0.1256912202,0.1727491468,-0.1056274921,0.1610930264,0.0997461826,-0.0547138155,-0.037544772,0.2110037208,0.0817849934,-0.023355253,-0.5620622039,0.0634886622,0.2465737909,0.3047662675,0.0822414607,-0.0770296752,0.1512786895,-0.0502387509,0.054323215,0.4209915996,-0.0538993366,0.4664910734,-0.2711001635,0.1770830005,0.277297765,-0.3652850688,-0.0003139204,0.1826036125,-0.2375145853,0.1117426082,-0.2652013898,-0.4663338959,-0.1654590964,-0.1183409244,0.1349778473,-0.0789419264,-0.2356961071,-0.0736637264,-0.3157056272,-0.0500996858,0.2459145337,0.1108266786,0.0505057685,-0.2129725069,0.1019253507,-0.0896832198,0.2275961339,0.0058478625,-0.2316600084,0.11156746,0.4407880902,-0.0369520411,-0.0522972122,-0.1141529381,-0.116719678,0.1146957576,-0.0951117575,0.2058560997,0.2640617192,-0.0140006486,-0.222789079,0.3215104342,-0.2788105011,-0.2408593893,0.1297049522,-0.0126401894,-0.1174058467,0.0466670133,0.0669539571,-0.0052326033,-0.5102657676,0.1701666266,0.1292891949,0.047954496,-0.4098981023,-0.0345081277,-0.0219770689,0.1678383499,0.0930051282,-0.2305410653,-0.0614778958,0.4744261503,-0.0609407164,0.1962468177,-0.0178826489,-0.2281215787,-0.019167576,-0.6979162097,-0.3184904456,-0.0092402529,0.1823446751,-0.1330951899,0.4334768355,0.2950298786,0.1894212216,-0.0313181132,0.1306529492,-0.1452960372,-0.1054134294,-0.0241774917,-0.1970705837,-0.0412810817,0.4775482118,0.0791641027,-0.083314769,-0.0907501876,-0.5611173511,-0.1841323078,0.3183330894,-0.2006358951,0.0832459405,0.2185398936,0.2986201346,-0.1066382155,-0.174405396,-0.2389483154,0.4070246518,-0.0666783303,-0.0943469256,0.3965403438,-0.1829442829,0.1772942096,0.30615291,0.1469939947,0.2547303438,0.3355357349,0.0759050399,0.0548323095,-0.3406339884,0.1482571512,-0.1149554327,0.2089336812,-0.2045271844,0.2236363739,-0.2222569287,0.1700025946,-0.2625818849,-0.2926515341,-0.1523296684,-0.0743206516,0.020759508,0.3121842742,0.1147071868,0.2348940372,0.3013081253,0.2710024714,0.1796744317,0.24606888,0.0581639707,0.1148115098,-0.2809874415,-0.0713953227,0.409381628,-0.3983355463,0.437820226,0.0031497572,-0.3976996541,-0.1845485568,-0.3869085014,0.2667299807,-0.3574424982,0.4122446179,0.23498936,0.0989026725,0.2883796394,-0.0820783675,0.2549666166,-0.1147719994,-0.2736467421,0.1317070425,0.0489208996,-0.0504887998,-0.1470930874,-0.1372155547,-0.076008074,-0.1587286741,0.4361136556,0.0166942198,0.2049041837,0.3152825236,0.0969091505,0.2270750254,-0.0253577381,0.1294412464,-0.3533500433,-0.2036488354,-0.0627004877,0.0351996385,-0.2237658501,-0.0332907401,0.0955420956,0.3138401806,-0.115697816,-0.3986230791,-0.2539067566,0.2589060962,-0.4011817276,0.0286345966,-0.0587614216,0.2250745445,-0.0227604751,0.0023586736,0.1034687683,-0.1657687128,0.2917427123,0.0014571624,0.4208348393,-0.0582709499,0.4073520303,0.1907383204,0.2507788241,0.2218789309,-0.1074185222,0.2332936674,-0.0760882199,0.2300221622,0.0626940429,-0.2186158597,0.1225989982,-0.1815152466,-0.0678680167,0.2153835744,-0.2848355174,0.1016196385,-0.2425624877,0.3786380589,-0.3925980031,-0.2305805087,-0.0728006586,0.084850952,0.2980444729,-0.218303442,-0.1612262875,0.3346336484,0.1729964465,0.1649609953,0.1857609898,-0.0097199064,-0.2947641015,0.1768292785,-0.4341126978,-0.1974519044,0.2037676871,0.1779196262,0.3842518628,-0.1591307223,-0.1912715137,0.0965435058,-0.3532340825,0.2157043666,-0.1090002358,0.1710741967,-0.0191141888,-0.2454843968,0.0143333711,-0.1819860488,-0.2506864071,-0.1394036263,-0.0508332662,0.2705188692,-0.4435863197,-0.0580348633,0.9266176224,-0.0307043791,-0.2027936578,-0.0545093827,-0.3822521865,-0.3325395584,-0.2178062052,0.3268850744,0.0803599432,0.1595795751,-0.2594973445,0.054746896,0.1613204628,-0.0796455517,0.1856283695,0.0180702563,-0.1477721334,-0.2234720886,-0.0046396349,-0.0809065476,-0.2214631587,0.3577394187,0.066345498,0.1223880351,0.0403788388,-0.1450326741,0.2732554376,0.2892638147,0.4261297882,-0.2185800523,0.2583729923,-0.1215013042,0.2301147729,-0.2938078642,0.114960365,0.5577464104,0.1897827685,-0.3376097083,0.3320045173,0.1806747466,-0.3360965848,0.121775113,0.473772347,0.2906243801,-0.034681879,0.0033714068,0.0544294566,1.076790452,-0.1063169241,-0.1410206258,0.3265558183,-0.2886441648,-0.0629893243,0.1711464971,0.074469611,-0.3738822639,-0.2014196664,-0.1141918153,-0.099731572,0.182915628,-0.5049661398,-0.46861884,0.270686388,0.2902297676,0.047393661,0.1362045556,0.0351531766,0.0513243116,0.0296752714,-0.1473932117,0.075758256,-0.1771012098,0.4955222607,-0.3521041572,0.006089679,0.1425145864,-0.2339378148,-0.5489826798,0.1974165887,-0.7597703934,-0.2172420472,0.1435348392,-0.0956566334,0.2164642364,-0.0425305665,0.1211306229,0.2369874418,-0.0984941423,-0.0710498318,-0.0660440177,-0.2263087183,-0.278244406,0.080314897,0.1488425285,-0.1012301594,-0.3922503591,0.2363767028,-0.0378921032,-0.1245239452,0.2563019693,-0.1621961296,0.0216446221,-0.1495987326,-0.2880109847,0.1166077033,-0.4050696194,-0.3859924376,0.0441254005,-0.0842707083,-0.6793951392,-0.0104836412,0.2277966589,-0.2703295052,0.0709382817,0.3031373918,-0.4074920118,0.4476473927,0.2836990356,0.3070386052,-0.0456531011,-0.2508759201,-0.2475806177,-0.096215196,-0.239861846,0.1254095435,-0.1179654077,0.0117817651,-0.0179658718,0.3521892726,0.2009031028,0.130558461,-0.0421137735,-0.1326913387,-0.4322311878,0.1133480817,0.0895270556,0.3084596992,0.0070495745,0.8161113858,0.0050012395,0.1381090879,-0.2168498635,0.1537139416,-0.5475091934,0.1356402934,0.5144497156,-0.3967116475,0.2740752399,0.1211526692,0.1307330132,0.3177930415,-0.2503612041,-0.1576968282,-0.028146917,0.1464556307,0.1394556761,-0.2126519084,0.0227943938,0.170848608,0.0856941938,0.169199571,0.0142538045,-0.1359180212,0.0720585734,0.0227145944,0.5171648264,0.066288501,-0.133510977,-0.3067292571,-0.0594982728,0.2850827277,0.1547500938,-0.0466246828,-0.004913874,-0.3229970634,-0.0157679971,0.0993029922,0.2712794542,-0.0949808732,0.1953766793,-0.2645510137,-0.3805359304,-0.2151184231,0.5645757914,0.3280044198,-0.3266963065,0.086527355,-0.0304161627,0.018389836,-0.2668769658,0.1713368893,-0.1892541796,-0.1514424533,-0.0306949466,0.5003805757,-0.1047546417,0.2028654367,0.0371597931,0.213701725,-0.00022627,-0.1407934874,-0.0896461084,0.0310132988,-0.2003608495,0.1695739627,0.1094613299,-0.1566507071,0.3708167672,-0.323977381,-0.0258680321,0.3565204144,-0.5007627606,0.3787528872,0.1755802333,-0.2727207839,0.0502875932,0.2554655969,0.0517259911,0.1064760238,0.0574634559,0.5036066771,0.0453846492,-0.3039710224,-0.0631596595,0.4623815715,-0.1607019305,-0.2788872123,0.1073022485,-0.0637273565,-0.3115138113,0.134660393,0.1420977265,-0.02707389,0.0633941963,0.0979784504,-0.3087844551,-0.1646403223,-0.401157558,0.2207766622,0.2365179211,-0.2379700541,0.5151308179,0.2662872076,-0.0371031389,-0.2489867359,0.7079617977,0.3704000711,-0.044673983,-0.1711800694,0.1376731247,-0.2246428877,0.1120398939,0.2599118352,-0.0763739496,0.0478594899,-0.0139040696,0.1707173288,0.0092956806,-0.0898856148,0.1108519435,0.1339693218,-0.3126707971,-0.2695256174,0.1654578,0.014774763,-0.0065957443,-0.4888792634,0.381357491,-0.2600221634,0.6083681583,0.3633648157,-0.0167721026,0.1414662302,0.2023964226,0.0427674167,0.2453739345,0.3704001904,0.2853342891,-0.1474468559,-0.3849023879,-0.3310810626,-0.3185833395,0.2492609769,-0.0115360944,0.4620011747,0.2291846871,0.134952575,-0.2289453298,0.0339672528,0.4562621713,0.2448747754,0.0648012906,0.1269744784,-0.2102719843,0.0880673677,-0.1616372764,-0.2616662979,0.0603457205,-0.1920471936,0.1918265373,-0.2449938953,-0.0385891534,-0.0182405785,0.0576571263,-0.4193947613,-0.0711653382,0.1217793003,-0.0717778057,-0.0353250094,-0.1474374086,-0.3151301742,0.029259732,-0.0187704731,-0.0378767923,0.2944749594,0.0775456652,-0.0054252218,0.1476282179,-0.4997337461,-0.2743061185,0.0258237924,0.4406977594,-0.2658599317,-0.2488072366,0.1811849922,0.0307605229,-0.2400042415,0.132065475,0.3002210855,0.0648869798,0.091827713,-0.559091866,-0.1621075869,0.6869646907,0.3171429932,0.0053969757,-0.1448841244,0.1080972478,0.214095071,-0.2759179175,-0.8308287263,0.0773528218,0.1110792533,0.0343419835,-0.3358500004,0.4130390882,-0.396550715,0.0305419471,0.146956265,0.2626532614,0.2095080614,0.0917727649,0.3162976801,-0.1423849612]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1286","title":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted","comments":"Hii please follow me","body":"Hi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n","comment_length":4,"text":"[libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted\nHi\r\nI am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help\r\n\r\n{'epoch': 20.0} \r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 20\/20 [00:16<00:00, 1.22it\/s]\r\n12\/08\/2020 10:41:19 - INFO - seq2seq.trainers.trainer - Saving model checkpoint to outputs\/experiment\/joint\/finetune\/lr-2e-5\r\n12\/08\/2020 10:41:24 - INFO - __main__ - {'wmt16-en-ro': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1998), 'qnli': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 5462), 'scitail': Dataset(features: {'src_texts': Value(dtype='string', id=None), 'task': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 1303)}\r\n12\/08\/2020 10:41:24 - INFO - __main__ - *** Evaluate ***\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.utils.utils - using task specific params for wmt16-en-ro: {'max_length': 300, 'num_beams': 4}\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - ***** Running Evaluation *****\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Num examples = 1998\r\n12\/08\/2020 10:41:24 - INFO - seq2seq.trainers.trainer - Batch size = 64\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32\/32 [00:37<00:00, 1.19s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nAborted\r\n\nHii please follow me","embeddings":[-0.0197352897,-0.7191827893,0.0039322353,0.349368006,0.484241426,0.0076734996,0.2424444556,0.1216879562,-0.2513463497,0.3279392719,-0.0821483955,-0.1225553527,-0.1213336438,0.3986013532,0.024262283,-0.2556790411,-0.0995945483,0.0782158896,-0.3232808411,0.1850418597,-0.0975957587,0.2666340768,-0.3790644109,0.2954924107,-0.077346921,0.0884474367,0.1073483229,0.0845447779,-0.0165824965,-0.3517342508,0.0590415671,-0.0919298232,-0.3089157045,0.3623276055,-0.0001231714,-0.1029601544,0.2742366493,0.044334691,-0.0298559833,-0.1064275429,0.264336437,0.5919190049,-0.1285633147,0.155628711,-0.1199958175,0.0390061848,0.131960839,-0.02621934,0.2547330558,0.2040431052,0.1139890254,0.3572780192,0.4076769948,0.0077055944,-0.0644371659,-0.047120966,-0.2066649497,0.1545071602,0.0657213703,-0.1996357739,0.1090153903,-0.0004496211,0.1906781793,-0.0334959961,-0.2198651433,-0.3123973608,-0.0370121896,-0.457115531,0.0150993029,0.4477550387,-0.0995786712,-0.2989667356,-0.1321324855,0.2411860675,0.0361542404,-0.7871572971,0.0136458101,0.1679371297,-0.0502248146,0.2885320783,-0.137509048,0.1263263226,-0.2918356359,0.0164490547,-0.0907559618,0.296466738,-0.1985564381,0.2684500217,0.3229812086,0.1940049678,-0.3849190474,0.1892252117,0.0081450799,0.1069084778,-0.3058028519,-0.2175878286,0.1347346306,0.0532117598,0.291143328,0.1250610203,-0.0492140464,0.0266518034,-0.0314312987,0.1974624097,0.2514950335,-0.0322238952,0.0891965702,0.1261572838,0.157536611,0.5037890673,0.053735029,0.2031769902,0.0966271609,-0.2018554062,0.0816629007,-0.0702104494,0.5441356897,-0.1696140319,-0.5968497992,0.4957826138,-0.4942743778,-0.0085916603,0.2324092388,0.4670405984,-0.1604339182,0.0753958225,0.251483798,-0.1050434783,0.033637356,0.03451325,-0.0803707466,0.2207057327,0.1228926927,0.2221151143,-0.1487662196,-0.4933043122,0.1599145234,-0.1897882819,0.2903176546,-0.6953832507,-0.2288500816,-0.366612047,-0.0604084767,0.1137853563,-0.1019294262,-0.1594460607,0.2681770623,0.0848985463,-0.013057163,0.0232961178,0.2334747612,-0.3981198967,0.1256912202,0.1727491468,-0.1056274921,0.1610930264,0.0997461826,-0.0547138155,-0.037544772,0.2110037208,0.0817849934,-0.023355253,-0.5620622039,0.0634886622,0.2465737909,0.3047662675,0.0822414607,-0.0770296752,0.1512786895,-0.0502387509,0.054323215,0.4209915996,-0.0538993366,0.4664910734,-0.2711001635,0.1770830005,0.277297765,-0.3652850688,-0.0003139204,0.1826036125,-0.2375145853,0.1117426082,-0.2652013898,-0.4663338959,-0.1654590964,-0.1183409244,0.1349778473,-0.0789419264,-0.2356961071,-0.0736637264,-0.3157056272,-0.0500996858,0.2459145337,0.1108266786,0.0505057685,-0.2129725069,0.1019253507,-0.0896832198,0.2275961339,0.0058478625,-0.2316600084,0.11156746,0.4407880902,-0.0369520411,-0.0522972122,-0.1141529381,-0.116719678,0.1146957576,-0.0951117575,0.2058560997,0.2640617192,-0.0140006486,-0.222789079,0.3215104342,-0.2788105011,-0.2408593893,0.1297049522,-0.0126401894,-0.1174058467,0.0466670133,0.0669539571,-0.0052326033,-0.5102657676,0.1701666266,0.1292891949,0.047954496,-0.4098981023,-0.0345081277,-0.0219770689,0.1678383499,0.0930051282,-0.2305410653,-0.0614778958,0.4744261503,-0.0609407164,0.1962468177,-0.0178826489,-0.2281215787,-0.019167576,-0.6979162097,-0.3184904456,-0.0092402529,0.1823446751,-0.1330951899,0.4334768355,0.2950298786,0.1894212216,-0.0313181132,0.1306529492,-0.1452960372,-0.1054134294,-0.0241774917,-0.1970705837,-0.0412810817,0.4775482118,0.0791641027,-0.083314769,-0.0907501876,-0.5611173511,-0.1841323078,0.3183330894,-0.2006358951,0.0832459405,0.2185398936,0.2986201346,-0.1066382155,-0.174405396,-0.2389483154,0.4070246518,-0.0666783303,-0.0943469256,0.3965403438,-0.1829442829,0.1772942096,0.30615291,0.1469939947,0.2547303438,0.3355357349,0.0759050399,0.0548323095,-0.3406339884,0.1482571512,-0.1149554327,0.2089336812,-0.2045271844,0.2236363739,-0.2222569287,0.1700025946,-0.2625818849,-0.2926515341,-0.1523296684,-0.0743206516,0.020759508,0.3121842742,0.1147071868,0.2348940372,0.3013081253,0.2710024714,0.1796744317,0.24606888,0.0581639707,0.1148115098,-0.2809874415,-0.0713953227,0.409381628,-0.3983355463,0.437820226,0.0031497572,-0.3976996541,-0.1845485568,-0.3869085014,0.2667299807,-0.3574424982,0.4122446179,0.23498936,0.0989026725,0.2883796394,-0.0820783675,0.2549666166,-0.1147719994,-0.2736467421,0.1317070425,0.0489208996,-0.0504887998,-0.1470930874,-0.1372155547,-0.076008074,-0.1587286741,0.4361136556,0.0166942198,0.2049041837,0.3152825236,0.0969091505,0.2270750254,-0.0253577381,0.1294412464,-0.3533500433,-0.2036488354,-0.0627004877,0.0351996385,-0.2237658501,-0.0332907401,0.0955420956,0.3138401806,-0.115697816,-0.3986230791,-0.2539067566,0.2589060962,-0.4011817276,0.0286345966,-0.0587614216,0.2250745445,-0.0227604751,0.0023586736,0.1034687683,-0.1657687128,0.2917427123,0.0014571624,0.4208348393,-0.0582709499,0.4073520303,0.1907383204,0.2507788241,0.2218789309,-0.1074185222,0.2332936674,-0.0760882199,0.2300221622,0.0626940429,-0.2186158597,0.1225989982,-0.1815152466,-0.0678680167,0.2153835744,-0.2848355174,0.1016196385,-0.2425624877,0.3786380589,-0.3925980031,-0.2305805087,-0.0728006586,0.084850952,0.2980444729,-0.218303442,-0.1612262875,0.3346336484,0.1729964465,0.1649609953,0.1857609898,-0.0097199064,-0.2947641015,0.1768292785,-0.4341126978,-0.1974519044,0.2037676871,0.1779196262,0.3842518628,-0.1591307223,-0.1912715137,0.0965435058,-0.3532340825,0.2157043666,-0.1090002358,0.1710741967,-0.0191141888,-0.2454843968,0.0143333711,-0.1819860488,-0.2506864071,-0.1394036263,-0.0508332662,0.2705188692,-0.4435863197,-0.0580348633,0.9266176224,-0.0307043791,-0.2027936578,-0.0545093827,-0.3822521865,-0.3325395584,-0.2178062052,0.3268850744,0.0803599432,0.1595795751,-0.2594973445,0.054746896,0.1613204628,-0.0796455517,0.1856283695,0.0180702563,-0.1477721334,-0.2234720886,-0.0046396349,-0.0809065476,-0.2214631587,0.3577394187,0.066345498,0.1223880351,0.0403788388,-0.1450326741,0.2732554376,0.2892638147,0.4261297882,-0.2185800523,0.2583729923,-0.1215013042,0.2301147729,-0.2938078642,0.114960365,0.5577464104,0.1897827685,-0.3376097083,0.3320045173,0.1806747466,-0.3360965848,0.121775113,0.473772347,0.2906243801,-0.034681879,0.0033714068,0.0544294566,1.076790452,-0.1063169241,-0.1410206258,0.3265558183,-0.2886441648,-0.0629893243,0.1711464971,0.074469611,-0.3738822639,-0.2014196664,-0.1141918153,-0.099731572,0.182915628,-0.5049661398,-0.46861884,0.270686388,0.2902297676,0.047393661,0.1362045556,0.0351531766,0.0513243116,0.0296752714,-0.1473932117,0.075758256,-0.1771012098,0.4955222607,-0.3521041572,0.006089679,0.1425145864,-0.2339378148,-0.5489826798,0.1974165887,-0.7597703934,-0.2172420472,0.1435348392,-0.0956566334,0.2164642364,-0.0425305665,0.1211306229,0.2369874418,-0.0984941423,-0.0710498318,-0.0660440177,-0.2263087183,-0.278244406,0.080314897,0.1488425285,-0.1012301594,-0.3922503591,0.2363767028,-0.0378921032,-0.1245239452,0.2563019693,-0.1621961296,0.0216446221,-0.1495987326,-0.2880109847,0.1166077033,-0.4050696194,-0.3859924376,0.0441254005,-0.0842707083,-0.6793951392,-0.0104836412,0.2277966589,-0.2703295052,0.0709382817,0.3031373918,-0.4074920118,0.4476473927,0.2836990356,0.3070386052,-0.0456531011,-0.2508759201,-0.2475806177,-0.096215196,-0.239861846,0.1254095435,-0.1179654077,0.0117817651,-0.0179658718,0.3521892726,0.2009031028,0.130558461,-0.0421137735,-0.1326913387,-0.4322311878,0.1133480817,0.0895270556,0.3084596992,0.0070495745,0.8161113858,0.0050012395,0.1381090879,-0.2168498635,0.1537139416,-0.5475091934,0.1356402934,0.5144497156,-0.3967116475,0.2740752399,0.1211526692,0.1307330132,0.3177930415,-0.2503612041,-0.1576968282,-0.028146917,0.1464556307,0.1394556761,-0.2126519084,0.0227943938,0.170848608,0.0856941938,0.169199571,0.0142538045,-0.1359180212,0.0720585734,0.0227145944,0.5171648264,0.066288501,-0.133510977,-0.3067292571,-0.0594982728,0.2850827277,0.1547500938,-0.0466246828,-0.004913874,-0.3229970634,-0.0157679971,0.0993029922,0.2712794542,-0.0949808732,0.1953766793,-0.2645510137,-0.3805359304,-0.2151184231,0.5645757914,0.3280044198,-0.3266963065,0.086527355,-0.0304161627,0.018389836,-0.2668769658,0.1713368893,-0.1892541796,-0.1514424533,-0.0306949466,0.5003805757,-0.1047546417,0.2028654367,0.0371597931,0.213701725,-0.00022627,-0.1407934874,-0.0896461084,0.0310132988,-0.2003608495,0.1695739627,0.1094613299,-0.1566507071,0.3708167672,-0.323977381,-0.0258680321,0.3565204144,-0.5007627606,0.3787528872,0.1755802333,-0.2727207839,0.0502875932,0.2554655969,0.0517259911,0.1064760238,0.0574634559,0.5036066771,0.0453846492,-0.3039710224,-0.0631596595,0.4623815715,-0.1607019305,-0.2788872123,0.1073022485,-0.0637273565,-0.3115138113,0.134660393,0.1420977265,-0.02707389,0.0633941963,0.0979784504,-0.3087844551,-0.1646403223,-0.401157558,0.2207766622,0.2365179211,-0.2379700541,0.5151308179,0.2662872076,-0.0371031389,-0.2489867359,0.7079617977,0.3704000711,-0.044673983,-0.1711800694,0.1376731247,-0.2246428877,0.1120398939,0.2599118352,-0.0763739496,0.0478594899,-0.0139040696,0.1707173288,0.0092956806,-0.0898856148,0.1108519435,0.1339693218,-0.3126707971,-0.2695256174,0.1654578,0.014774763,-0.0065957443,-0.4888792634,0.381357491,-0.2600221634,0.6083681583,0.3633648157,-0.0167721026,0.1414662302,0.2023964226,0.0427674167,0.2453739345,0.3704001904,0.2853342891,-0.1474468559,-0.3849023879,-0.3310810626,-0.3185833395,0.2492609769,-0.0115360944,0.4620011747,0.2291846871,0.134952575,-0.2289453298,0.0339672528,0.4562621713,0.2448747754,0.0648012906,0.1269744784,-0.2102719843,0.0880673677,-0.1616372764,-0.2616662979,0.0603457205,-0.1920471936,0.1918265373,-0.2449938953,-0.0385891534,-0.0182405785,0.0576571263,-0.4193947613,-0.0711653382,0.1217793003,-0.0717778057,-0.0353250094,-0.1474374086,-0.3151301742,0.029259732,-0.0187704731,-0.0378767923,0.2944749594,0.0775456652,-0.0054252218,0.1476282179,-0.4997337461,-0.2743061185,0.0258237924,0.4406977594,-0.2658599317,-0.2488072366,0.1811849922,0.0307605229,-0.2400042415,0.132065475,0.3002210855,0.0648869798,0.091827713,-0.559091866,-0.1621075869,0.6869646907,0.3171429932,0.0053969757,-0.1448841244,0.1080972478,0.214095071,-0.2759179175,-0.8308287263,0.0773528218,0.1110792533,0.0343419835,-0.3358500004,0.4130390882,-0.396550715,0.0305419471,0.146956265,0.2626532614,0.2095080614,0.0917727649,0.3162976801,-0.1423849612]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1285","title":"boolq does not work ","comments":"here is the minimal code to reproduce\r\n\r\n`datasets>>> datasets.load_dataset(\"boolq\", \"train\")\r\n\r\nthe errors\r\n\r\n```\r\n`cahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nUsing custom data configuration train\r\nDownloading and preparing dataset boolq\/train (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/train\/0.1.0\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\n\r\n\r\n\r\n```","body":"Hi\r\nI am getting this error when trying to load boolq, thanks for your help\r\n\r\nts_boolq_default_0.1.0_2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11.lock\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 274, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 147, in main\r\n for task in data_args.tasks]\r\n File \"finetune_t5_trainer.py\", line 147, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 58, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 54, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\ntensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists\r\n\r\n","comment_length":115,"text":"boolq does not work \nHi\r\nI am getting this error when trying to load boolq, thanks for your help\r\n\r\nts_boolq_default_0.1.0_2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11.lock\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 274, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 147, in main\r\n for task in data_args.tasks]\r\n File \"finetune_t5_trainer.py\", line 147, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 58, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 54, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\ntensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists\r\n\r\n\nhere is the minimal code to reproduce\r\n\r\n`datasets>>> datasets.load_dataset(\"boolq\", \"train\")\r\n\r\nthe errors\r\n\r\n```\r\n`cahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\nUsing custom data configuration train\r\nDownloading and preparing dataset boolq\/train (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/boolq\/train\/0.1.0\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home_1\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\n\r\n\r\n\r\n```","embeddings":[-0.2191431969,-0.205223009,-0.1021828577,0.056804508,0.0433856584,0.0777705982,0.3575468659,0.2529199123,0.2545483708,-0.0290935356,-0.137118265,0.3180469871,-0.2989818454,0.435959518,-0.0375271626,-0.0104442453,-0.2001055479,0.0617069267,0.0194576196,0.1253176033,-0.3767187595,0.4527184665,-0.3950281441,0.1717271507,0.039792046,-0.1584234387,0.0690807477,0.2183280289,-0.3045058548,-0.367679894,0.1875551343,-0.0160979461,0.1602543443,0.1953685731,-0.000100287,0.0124218855,0.3434242904,0.0293650366,-0.1709919572,-0.3191271424,-0.1750642061,-0.3502523601,0.2896553278,-0.0837975815,-0.0092108911,0.0844620094,-0.040360827,0.1825928986,0.3829545379,0.3910404146,0.3393365741,0.0961863548,0.1874352694,-0.2147234976,0.2771983445,-0.5643305779,0.0346739218,0.1041202247,0.2713558376,-0.1368377209,0.066627875,0.1479776949,-0.2208089978,-0.2162980139,-0.1051368415,-0.2081234902,-0.0779546276,-0.3245044649,0.1794871092,0.0790971294,0.6055165529,-0.0848833472,-0.0521295182,0.1929166764,0.2385609597,-0.1072742119,0.1638332754,0.0137318,0.0599494018,0.0544823632,0.2340850532,-0.3852249384,-0.0247666705,0.0413612984,-0.0779064223,0.2063310295,-0.0665245429,0.0390985645,0.2114246339,-0.2090814412,0.1663435102,0.508505702,-0.109657988,-0.0038654527,-0.6627800465,0.024767397,0.2779552937,0.107672438,-0.0527212545,0.0032308488,-0.265878886,0.0674191788,0.1694182903,0.1585515887,-0.1650961787,0.3916872144,0.2289465219,0.2797166705,0.0727470815,0.2224866748,-0.2273650616,0.2074246705,-0.2910070717,-0.492138654,-0.0160816219,0.1348418295,0.3629411459,-0.0850622207,-0.2391541153,-0.1036297455,-0.1959228367,-0.0248908307,0.3913449347,0.7891902924,0.1488980204,-0.3131633401,0.1699921936,0.0756845623,-0.2671214938,-0.1868343651,-0.3410372734,0.2802948356,0.0750981793,-0.1892602742,0.1544688791,-0.3493214548,0.1605904698,0.1638125777,-0.0757626295,-0.1015292183,0.2639566064,-0.1807930171,-0.2796195447,0.2034140676,-0.1660426855,0.0412304625,0.222171858,0.0821633413,-0.063733995,0.2537777126,-0.3442661166,-0.4547116458,-0.1617778689,0.3301042616,0.1894203126,-0.27367872,0.1826002151,0.2261167169,0.1871689409,-0.1125649959,-0.1956937462,-0.3212908208,-0.2962062657,-0.1822833419,0.2448646575,0.3433954418,-0.4542224705,-0.0933891684,-0.2166294158,-0.2461134046,-0.054044459,0.1466912478,-0.2197429091,0.398642987,-0.1403478235,0.0188561901,0.1012149081,-0.2950314879,-0.4969398975,0.2633554637,-0.2121250033,-0.1468689889,-0.2692807019,0.0484414436,0.1941547692,0.1539271921,0.0051376866,0.0565571487,0.2050610185,-0.2150689512,-0.1294436157,-0.0946026072,0.2298936993,0.3611728549,0.1907626092,-0.0996231213,-0.0038339908,0.0660987124,0.0545602627,0.2124175131,-0.2959643602,0.2606880069,0.2629840374,-0.1679650694,0.0497377403,0.0490054712,-0.3281366825,0.1711824983,-0.1773874164,0.1284162402,0.0088677797,0.2201396525,-0.3857020736,0.124791272,-0.2456311733,0.0568337888,0.2308734208,-0.2148270011,-0.0974920318,0.0471501648,-0.1539621353,0.2841616571,-0.1739644557,0.055631686,-0.071144715,-0.0564676598,-0.3174751997,-0.1609033346,0.2278515995,0.1587961763,0.1742808223,0.0216859486,0.0050521316,0.3554823399,0.0803116187,-0.0317420475,-0.2111491263,-0.4885710478,-0.0345933922,-0.1139063314,0.0384387746,0.102731511,0.1070292294,0.1205408573,0.4304286242,0.2363624573,-0.1677540839,-0.0515252911,-0.0633639246,0.1917677373,0.1616193652,0.0279592704,-0.0829842314,-0.0570030771,0.4093349576,0.0435367897,-0.0650580376,-0.0590648204,-0.0195259042,0.014713522,0.4898835123,0.1344736814,0.0625171438,-0.1088181809,0.2517293394,0.1401916444,0.0302508697,0.1328377724,0.277749151,0.1064233333,-0.4235241711,-0.0366860069,-0.0264514554,-0.2671252787,0.0989147648,-0.3150690496,-0.1128067076,0.1915855706,-0.0246155076,-0.162846759,-0.1511110663,-0.2533967495,0.1131062135,0.344355613,-0.1373842359,-0.0577376187,-0.1854242533,-0.2737466395,0.3402540386,-0.0681999996,0.0371059217,-0.316401571,0.2217026651,-0.0770749301,0.138763383,0.3388938606,-0.0285247955,0.1094074845,0.3005981147,-0.2708330452,-0.0180830099,-0.2160334438,-0.1804675907,0.1515997946,0.111972183,-0.2137085795,0.0305303,-0.1087111309,-0.0169140063,-0.004822446,-0.3049965501,-0.1138391271,-0.3007988334,0.3057751358,0.2515809238,0.0596834607,-0.3211912513,-0.0032180632,0.1792205125,-0.2130450308,-0.0801892728,0.1580676436,0.1449140161,-0.0205494296,-0.204799369,-0.4926880896,-0.5303481817,-0.4888476729,-0.0292072166,-0.0471295975,0.1707663089,0.3728390336,-0.0082199024,0.3335877657,0.1069243327,0.1427049488,-0.0587089285,-0.2668074667,0.2721175849,0.0249675456,-0.3096513748,0.1576294154,0.1298577338,0.3517134786,-0.2375917286,-0.1885915101,0.1306701601,0.0096215066,0.0643238276,-0.2209012657,-0.1836156547,0.0529978499,-0.0396089479,-0.0411025397,0.0911995918,-0.2069228888,0.2017320096,0.0643187687,0.1448442787,0.0772933438,0.2657338977,-0.0761375427,0.5265434384,0.0392807871,0.1855731457,0.2640739977,0.130462423,0.134970814,-0.0506642237,-0.5398096442,-0.0671861321,0.4233142436,-0.1721684039,-0.0361371003,-0.3168293238,0.1760747284,-0.2813004851,-0.0863358006,-0.1637949795,-0.1836038679,-0.1436770856,-0.2176184952,0.003678957,-0.0227848347,0.3067978621,0.1663006693,-0.0702654868,0.0677278712,0.3873063624,-0.1091511622,-0.0157409012,-0.2296463996,-0.2658676207,-0.2586579919,0.3157219291,-0.2792370319,0.3146839142,-0.1532648057,-0.0428509377,0.1665743738,-0.0807461962,0.4986982942,0.0501148775,-0.0191343743,0.1252700835,-0.2378567606,-0.0747508928,-0.1307845712,-0.1986892968,0.1724558771,-0.1019118577,0.233688727,-0.2286940366,0.1395965666,-0.1511369348,-0.1400825977,-0.2261397243,-0.4447718859,-0.4023564756,-0.2775964439,-0.3293209672,0.0714741349,-0.0625686571,0.3856148124,-0.0554290116,0.0615473762,-0.0310035963,-0.3100096881,0.092790179,0.0753413886,0.3129445314,-0.1278225631,0.0731345192,-0.1084556058,0.0209055487,0.0822391137,0.2575199902,-0.1618361473,-0.1628741324,0.0495073274,0.0791191757,0.1855982542,0.0263856649,-0.402969569,0.2586004734,0.0187967625,0.0338427462,0.3420662284,-0.1090370268,0.1140786931,0.0074263974,0.1009620354,-0.2298306525,-0.0616125651,-0.058305379,0.0528999977,0.1004934162,-0.0807929933,-0.0299653281,0.2240544558,-0.1453294307,0.506583333,-0.4068631828,-0.0270505585,0.3145805299,0.1090787724,0.4495962858,-0.0112947337,0.3527529538,-0.508730948,-0.3376002908,0.1896848232,-0.0802515373,-0.020310564,-0.4732180238,-0.5903365016,0.2061620504,-0.3476190567,0.2962293029,-0.0182416569,0.1656645536,-0.1126884371,0.1834341139,-0.2206294388,0.2286557853,-0.1790461391,0.247985974,-0.2823167741,-0.0033894579,0.1613980383,-0.1582825631,-0.1249284893,0.285826087,-0.1620238274,0.2359033674,0.203013733,-0.1350350231,0.1564134955,0.4539047778,-0.1938336343,0.283121109,-0.0773664489,0.0888468996,0.0780039355,-0.1140508503,-0.1461127847,0.2270624936,0.4083356261,-0.0326155946,-0.1881405115,0.6344335675,0.0801858529,0.0685991719,-0.2457523346,0.1895558685,-0.0615431778,-0.2341832668,-0.2373738438,-0.3077602386,-0.0010710877,-0.1609265208,0.1954807043,-0.0249681585,-0.1262477785,0.0540262945,-0.08676783,-0.4619655311,-0.0013346053,0.3493010104,0.0624585226,0.234109655,0.7029871345,0.3712255657,-0.0397965647,-0.3196722865,-0.0188007914,-0.0711913705,0.0093136588,0.2136898339,0.1183323935,0.0440297686,0.3165587485,0.2889674604,-0.0132552655,0.0372627228,0.0755368769,-0.380961895,-0.3957485855,0.2548047304,-0.2124716341,0.0100132292,-0.1032276899,0.1452225745,0.0724800304,0.1129650176,-0.4067446589,0.019754393,-0.5416104198,-0.1763852984,-0.0903903693,-0.1376826316,0.0024854471,0.2650108039,0.2780073285,0.1707057208,-0.2160606831,-0.3226590753,-0.0614078231,0.0523257367,0.1314618289,0.1478700936,-0.2413240224,-0.3203828633,-0.0740061328,0.3170636594,-0.1286108643,-0.0382293165,0.0708732903,0.3496511579,0.0035032725,0.0252895448,-0.0936604515,0.2850466967,-0.1372584254,-0.0343026966,0.0765368044,0.121716775,0.1666172147,-0.0452565625,-0.3469081521,0.0563661084,0.1295736283,0.0026262011,-0.030915482,-0.1531497389,0.0734745041,-0.1278023869,0.1320957392,0.5306506157,-0.225074023,-0.2179568708,0.0909496248,0.3123808801,-0.2310650349,0.2480469495,-0.2824618816,0.0132380547,-0.0537811294,0.1957368106,-0.2150955349,-0.1293255091,-0.0300089587,0.1327185929,0.136807695,0.1238769218,0.0107436711,0.3166515827,-0.2526705861,-0.1326469332,0.252402097,0.183785215,-0.0579355136,0.4016678333,-0.379914999,0.0427691974,-0.1585952193,0.3562553227,0.1159767285,-0.3820240498,-0.0618344136,0.3805193007,0.0734974816,-0.4506339729,-0.3727971613,0.4307152331,0.1382676959,-0.0312939622,0.091887705,0.2125879079,-0.4361151755,-0.2103605419,-0.1892086416,-0.0458007418,-0.1491173953,-0.0098613715,0.0262135956,0.0587635562,0.1239584982,0.2194059044,-0.006783633,-0.3287954628,0.2241207361,0.244951278,-0.3615393043,-0.2153161913,0.2987223864,0.0637276694,0.2853659689,0.3019364178,-0.1928016543,0.281584233,0.4341529012,-0.3539808095,0.0686111972,-0.0249796696,-0.0522812642,0.0304837395,0.3618318439,0.3311666548,0.4536408186,0.3099969327,0.1705660522,-0.1611032784,0.1405387074,0.1924778372,0.0576292202,-0.0661225244,0.5304341912,0.0419044942,0.0930425674,-0.1055197865,0.2225124091,-0.4005809426,-0.1301390082,0.1982863992,0.2625137269,0.1857812256,-0.2342473865,0.1494355351,-0.2071273029,0.3105294108,0.3286698759,0.5006896257,-0.4303507209,0.1608488262,-0.4329150021,0.2768636048,0.1256214976,-0.1307919025,-0.1068229303,0.1194247156,-0.111103788,-0.0143294847,0.1969831884,-0.0060373223,0.1079833359,-0.1549903154,-0.2892766893,0.1259484738,0.0802692324,0.0460333191,0.0018901629,-0.0472125784,0.0844692513,-0.4870571196,0.1282548308,-0.3443618417,-0.3882578909,0.2019646019,-0.043356806,0.2446198463,-0.0467390828,0.1558375806,-0.0514056198,0.0405545458,-0.0031453269,-0.1967606246,-0.14371939,-0.0071743135,0.2906820774,-0.0417024493,-0.1355441064,0.2434600145,-0.2620758116,0.2542633116,-0.1207381412,-0.0912974998,-0.2947988808,-0.0520454496,0.1674501002,-0.0951958746,0.1116770878,0.229479596,0.0491243228,0.2940317988,-0.5586979389,-0.1086154431,0.5152429938,0.093815662,-0.0625304431,0.0418021679,0.1600454897,0.0820968747,0.0147416862,-0.2646854818,0.1551657021,0.2506594062,-0.0373107567,0.0082782172,0.3977273405,0.173977077,0.1924737543,-0.1579850465,-0.1946321726,0.2570955455,0.1329276264,-0.0124456529,0.063294217]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1285","title":"boolq does not work ","comments":"This has been fixed by #881 \r\nthis fix will be available in the next release soon.\r\n\r\nIf you don't want to wait for the release you can actually load the latest version of boolq by specifying `script_version=\"master\"` in `load_dataset`","body":"Hi\r\nI am getting this error when trying to load boolq, thanks for your help\r\n\r\nts_boolq_default_0.1.0_2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11.lock\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 274, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 147, in main\r\n for task in data_args.tasks]\r\n File \"finetune_t5_trainer.py\", line 147, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 58, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 54, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\ntensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists\r\n\r\n","comment_length":39,"text":"boolq does not work \nHi\r\nI am getting this error when trying to load boolq, thanks for your help\r\n\r\nts_boolq_default_0.1.0_2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11.lock\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 274, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 147, in main\r\n for task in data_args.tasks]\r\n File \"finetune_t5_trainer.py\", line 147, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 58, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 54, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\ntensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists\r\n\r\n\nThis has been fixed by #881 \r\nthis fix will be available in the next release soon.\r\n\r\nIf you don't want to wait for the release you can actually load the latest version of boolq by specifying `script_version=\"master\"` in `load_dataset`","embeddings":[-0.2191431969,-0.205223009,-0.1021828577,0.056804508,0.0433856584,0.0777705982,0.3575468659,0.2529199123,0.2545483708,-0.0290935356,-0.137118265,0.3180469871,-0.2989818454,0.435959518,-0.0375271626,-0.0104442453,-0.2001055479,0.0617069267,0.0194576196,0.1253176033,-0.3767187595,0.4527184665,-0.3950281441,0.1717271507,0.039792046,-0.1584234387,0.0690807477,0.2183280289,-0.3045058548,-0.367679894,0.1875551343,-0.0160979461,0.1602543443,0.1953685731,-0.000100287,0.0124218855,0.3434242904,0.0293650366,-0.1709919572,-0.3191271424,-0.1750642061,-0.3502523601,0.2896553278,-0.0837975815,-0.0092108911,0.0844620094,-0.040360827,0.1825928986,0.3829545379,0.3910404146,0.3393365741,0.0961863548,0.1874352694,-0.2147234976,0.2771983445,-0.5643305779,0.0346739218,0.1041202247,0.2713558376,-0.1368377209,0.066627875,0.1479776949,-0.2208089978,-0.2162980139,-0.1051368415,-0.2081234902,-0.0779546276,-0.3245044649,0.1794871092,0.0790971294,0.6055165529,-0.0848833472,-0.0521295182,0.1929166764,0.2385609597,-0.1072742119,0.1638332754,0.0137318,0.0599494018,0.0544823632,0.2340850532,-0.3852249384,-0.0247666705,0.0413612984,-0.0779064223,0.2063310295,-0.0665245429,0.0390985645,0.2114246339,-0.2090814412,0.1663435102,0.508505702,-0.109657988,-0.0038654527,-0.6627800465,0.024767397,0.2779552937,0.107672438,-0.0527212545,0.0032308488,-0.265878886,0.0674191788,0.1694182903,0.1585515887,-0.1650961787,0.3916872144,0.2289465219,0.2797166705,0.0727470815,0.2224866748,-0.2273650616,0.2074246705,-0.2910070717,-0.492138654,-0.0160816219,0.1348418295,0.3629411459,-0.0850622207,-0.2391541153,-0.1036297455,-0.1959228367,-0.0248908307,0.3913449347,0.7891902924,0.1488980204,-0.3131633401,0.1699921936,0.0756845623,-0.2671214938,-0.1868343651,-0.3410372734,0.2802948356,0.0750981793,-0.1892602742,0.1544688791,-0.3493214548,0.1605904698,0.1638125777,-0.0757626295,-0.1015292183,0.2639566064,-0.1807930171,-0.2796195447,0.2034140676,-0.1660426855,0.0412304625,0.222171858,0.0821633413,-0.063733995,0.2537777126,-0.3442661166,-0.4547116458,-0.1617778689,0.3301042616,0.1894203126,-0.27367872,0.1826002151,0.2261167169,0.1871689409,-0.1125649959,-0.1956937462,-0.3212908208,-0.2962062657,-0.1822833419,0.2448646575,0.3433954418,-0.4542224705,-0.0933891684,-0.2166294158,-0.2461134046,-0.054044459,0.1466912478,-0.2197429091,0.398642987,-0.1403478235,0.0188561901,0.1012149081,-0.2950314879,-0.4969398975,0.2633554637,-0.2121250033,-0.1468689889,-0.2692807019,0.0484414436,0.1941547692,0.1539271921,0.0051376866,0.0565571487,0.2050610185,-0.2150689512,-0.1294436157,-0.0946026072,0.2298936993,0.3611728549,0.1907626092,-0.0996231213,-0.0038339908,0.0660987124,0.0545602627,0.2124175131,-0.2959643602,0.2606880069,0.2629840374,-0.1679650694,0.0497377403,0.0490054712,-0.3281366825,0.1711824983,-0.1773874164,0.1284162402,0.0088677797,0.2201396525,-0.3857020736,0.124791272,-0.2456311733,0.0568337888,0.2308734208,-0.2148270011,-0.0974920318,0.0471501648,-0.1539621353,0.2841616571,-0.1739644557,0.055631686,-0.071144715,-0.0564676598,-0.3174751997,-0.1609033346,0.2278515995,0.1587961763,0.1742808223,0.0216859486,0.0050521316,0.3554823399,0.0803116187,-0.0317420475,-0.2111491263,-0.4885710478,-0.0345933922,-0.1139063314,0.0384387746,0.102731511,0.1070292294,0.1205408573,0.4304286242,0.2363624573,-0.1677540839,-0.0515252911,-0.0633639246,0.1917677373,0.1616193652,0.0279592704,-0.0829842314,-0.0570030771,0.4093349576,0.0435367897,-0.0650580376,-0.0590648204,-0.0195259042,0.014713522,0.4898835123,0.1344736814,0.0625171438,-0.1088181809,0.2517293394,0.1401916444,0.0302508697,0.1328377724,0.277749151,0.1064233333,-0.4235241711,-0.0366860069,-0.0264514554,-0.2671252787,0.0989147648,-0.3150690496,-0.1128067076,0.1915855706,-0.0246155076,-0.162846759,-0.1511110663,-0.2533967495,0.1131062135,0.344355613,-0.1373842359,-0.0577376187,-0.1854242533,-0.2737466395,0.3402540386,-0.0681999996,0.0371059217,-0.316401571,0.2217026651,-0.0770749301,0.138763383,0.3388938606,-0.0285247955,0.1094074845,0.3005981147,-0.2708330452,-0.0180830099,-0.2160334438,-0.1804675907,0.1515997946,0.111972183,-0.2137085795,0.0305303,-0.1087111309,-0.0169140063,-0.004822446,-0.3049965501,-0.1138391271,-0.3007988334,0.3057751358,0.2515809238,0.0596834607,-0.3211912513,-0.0032180632,0.1792205125,-0.2130450308,-0.0801892728,0.1580676436,0.1449140161,-0.0205494296,-0.204799369,-0.4926880896,-0.5303481817,-0.4888476729,-0.0292072166,-0.0471295975,0.1707663089,0.3728390336,-0.0082199024,0.3335877657,0.1069243327,0.1427049488,-0.0587089285,-0.2668074667,0.2721175849,0.0249675456,-0.3096513748,0.1576294154,0.1298577338,0.3517134786,-0.2375917286,-0.1885915101,0.1306701601,0.0096215066,0.0643238276,-0.2209012657,-0.1836156547,0.0529978499,-0.0396089479,-0.0411025397,0.0911995918,-0.2069228888,0.2017320096,0.0643187687,0.1448442787,0.0772933438,0.2657338977,-0.0761375427,0.5265434384,0.0392807871,0.1855731457,0.2640739977,0.130462423,0.134970814,-0.0506642237,-0.5398096442,-0.0671861321,0.4233142436,-0.1721684039,-0.0361371003,-0.3168293238,0.1760747284,-0.2813004851,-0.0863358006,-0.1637949795,-0.1836038679,-0.1436770856,-0.2176184952,0.003678957,-0.0227848347,0.3067978621,0.1663006693,-0.0702654868,0.0677278712,0.3873063624,-0.1091511622,-0.0157409012,-0.2296463996,-0.2658676207,-0.2586579919,0.3157219291,-0.2792370319,0.3146839142,-0.1532648057,-0.0428509377,0.1665743738,-0.0807461962,0.4986982942,0.0501148775,-0.0191343743,0.1252700835,-0.2378567606,-0.0747508928,-0.1307845712,-0.1986892968,0.1724558771,-0.1019118577,0.233688727,-0.2286940366,0.1395965666,-0.1511369348,-0.1400825977,-0.2261397243,-0.4447718859,-0.4023564756,-0.2775964439,-0.3293209672,0.0714741349,-0.0625686571,0.3856148124,-0.0554290116,0.0615473762,-0.0310035963,-0.3100096881,0.092790179,0.0753413886,0.3129445314,-0.1278225631,0.0731345192,-0.1084556058,0.0209055487,0.0822391137,0.2575199902,-0.1618361473,-0.1628741324,0.0495073274,0.0791191757,0.1855982542,0.0263856649,-0.402969569,0.2586004734,0.0187967625,0.0338427462,0.3420662284,-0.1090370268,0.1140786931,0.0074263974,0.1009620354,-0.2298306525,-0.0616125651,-0.058305379,0.0528999977,0.1004934162,-0.0807929933,-0.0299653281,0.2240544558,-0.1453294307,0.506583333,-0.4068631828,-0.0270505585,0.3145805299,0.1090787724,0.4495962858,-0.0112947337,0.3527529538,-0.508730948,-0.3376002908,0.1896848232,-0.0802515373,-0.020310564,-0.4732180238,-0.5903365016,0.2061620504,-0.3476190567,0.2962293029,-0.0182416569,0.1656645536,-0.1126884371,0.1834341139,-0.2206294388,0.2286557853,-0.1790461391,0.247985974,-0.2823167741,-0.0033894579,0.1613980383,-0.1582825631,-0.1249284893,0.285826087,-0.1620238274,0.2359033674,0.203013733,-0.1350350231,0.1564134955,0.4539047778,-0.1938336343,0.283121109,-0.0773664489,0.0888468996,0.0780039355,-0.1140508503,-0.1461127847,0.2270624936,0.4083356261,-0.0326155946,-0.1881405115,0.6344335675,0.0801858529,0.0685991719,-0.2457523346,0.1895558685,-0.0615431778,-0.2341832668,-0.2373738438,-0.3077602386,-0.0010710877,-0.1609265208,0.1954807043,-0.0249681585,-0.1262477785,0.0540262945,-0.08676783,-0.4619655311,-0.0013346053,0.3493010104,0.0624585226,0.234109655,0.7029871345,0.3712255657,-0.0397965647,-0.3196722865,-0.0188007914,-0.0711913705,0.0093136588,0.2136898339,0.1183323935,0.0440297686,0.3165587485,0.2889674604,-0.0132552655,0.0372627228,0.0755368769,-0.380961895,-0.3957485855,0.2548047304,-0.2124716341,0.0100132292,-0.1032276899,0.1452225745,0.0724800304,0.1129650176,-0.4067446589,0.019754393,-0.5416104198,-0.1763852984,-0.0903903693,-0.1376826316,0.0024854471,0.2650108039,0.2780073285,0.1707057208,-0.2160606831,-0.3226590753,-0.0614078231,0.0523257367,0.1314618289,0.1478700936,-0.2413240224,-0.3203828633,-0.0740061328,0.3170636594,-0.1286108643,-0.0382293165,0.0708732903,0.3496511579,0.0035032725,0.0252895448,-0.0936604515,0.2850466967,-0.1372584254,-0.0343026966,0.0765368044,0.121716775,0.1666172147,-0.0452565625,-0.3469081521,0.0563661084,0.1295736283,0.0026262011,-0.030915482,-0.1531497389,0.0734745041,-0.1278023869,0.1320957392,0.5306506157,-0.225074023,-0.2179568708,0.0909496248,0.3123808801,-0.2310650349,0.2480469495,-0.2824618816,0.0132380547,-0.0537811294,0.1957368106,-0.2150955349,-0.1293255091,-0.0300089587,0.1327185929,0.136807695,0.1238769218,0.0107436711,0.3166515827,-0.2526705861,-0.1326469332,0.252402097,0.183785215,-0.0579355136,0.4016678333,-0.379914999,0.0427691974,-0.1585952193,0.3562553227,0.1159767285,-0.3820240498,-0.0618344136,0.3805193007,0.0734974816,-0.4506339729,-0.3727971613,0.4307152331,0.1382676959,-0.0312939622,0.091887705,0.2125879079,-0.4361151755,-0.2103605419,-0.1892086416,-0.0458007418,-0.1491173953,-0.0098613715,0.0262135956,0.0587635562,0.1239584982,0.2194059044,-0.006783633,-0.3287954628,0.2241207361,0.244951278,-0.3615393043,-0.2153161913,0.2987223864,0.0637276694,0.2853659689,0.3019364178,-0.1928016543,0.281584233,0.4341529012,-0.3539808095,0.0686111972,-0.0249796696,-0.0522812642,0.0304837395,0.3618318439,0.3311666548,0.4536408186,0.3099969327,0.1705660522,-0.1611032784,0.1405387074,0.1924778372,0.0576292202,-0.0661225244,0.5304341912,0.0419044942,0.0930425674,-0.1055197865,0.2225124091,-0.4005809426,-0.1301390082,0.1982863992,0.2625137269,0.1857812256,-0.2342473865,0.1494355351,-0.2071273029,0.3105294108,0.3286698759,0.5006896257,-0.4303507209,0.1608488262,-0.4329150021,0.2768636048,0.1256214976,-0.1307919025,-0.1068229303,0.1194247156,-0.111103788,-0.0143294847,0.1969831884,-0.0060373223,0.1079833359,-0.1549903154,-0.2892766893,0.1259484738,0.0802692324,0.0460333191,0.0018901629,-0.0472125784,0.0844692513,-0.4870571196,0.1282548308,-0.3443618417,-0.3882578909,0.2019646019,-0.043356806,0.2446198463,-0.0467390828,0.1558375806,-0.0514056198,0.0405545458,-0.0031453269,-0.1967606246,-0.14371939,-0.0071743135,0.2906820774,-0.0417024493,-0.1355441064,0.2434600145,-0.2620758116,0.2542633116,-0.1207381412,-0.0912974998,-0.2947988808,-0.0520454496,0.1674501002,-0.0951958746,0.1116770878,0.229479596,0.0491243228,0.2940317988,-0.5586979389,-0.1086154431,0.5152429938,0.093815662,-0.0625304431,0.0418021679,0.1600454897,0.0820968747,0.0147416862,-0.2646854818,0.1551657021,0.2506594062,-0.0373107567,0.0082782172,0.3977273405,0.173977077,0.1924737543,-0.1579850465,-0.1946321726,0.2570955455,0.1329276264,-0.0124456529,0.063294217]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1285","title":"boolq does not work ","comments":"thank you this solved this issue, for now seems to work, thanks ","body":"Hi\r\nI am getting this error when trying to load boolq, thanks for your help\r\n\r\nts_boolq_default_0.1.0_2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11.lock\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 274, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 147, in main\r\n for task in data_args.tasks]\r\n File \"finetune_t5_trainer.py\", line 147, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 58, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 54, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\ntensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists\r\n\r\n","comment_length":12,"text":"boolq does not work \nHi\r\nI am getting this error when trying to load boolq, thanks for your help\r\n\r\nts_boolq_default_0.1.0_2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11.lock\r\nTraceback (most recent call last):\r\n File \"finetune_t5_trainer.py\", line 274, in <module>\r\n main()\r\n File \"finetune_t5_trainer.py\", line 147, in main\r\n for task in data_args.tasks]\r\n File \"finetune_t5_trainer.py\", line 147, in <listcomp>\r\n for task in data_args.tasks]\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 58, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/ruse\/seq2seq\/tasks\/tasks.py\", line 54, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\ntensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists\r\n\r\n\nthank you this solved this issue, for now seems to work, thanks ","embeddings":[-0.2191431969,-0.205223009,-0.1021828577,0.056804508,0.0433856584,0.0777705982,0.3575468659,0.2529199123,0.2545483708,-0.0290935356,-0.137118265,0.3180469871,-0.2989818454,0.435959518,-0.0375271626,-0.0104442453,-0.2001055479,0.0617069267,0.0194576196,0.1253176033,-0.3767187595,0.4527184665,-0.3950281441,0.1717271507,0.039792046,-0.1584234387,0.0690807477,0.2183280289,-0.3045058548,-0.367679894,0.1875551343,-0.0160979461,0.1602543443,0.1953685731,-0.000100287,0.0124218855,0.3434242904,0.0293650366,-0.1709919572,-0.3191271424,-0.1750642061,-0.3502523601,0.2896553278,-0.0837975815,-0.0092108911,0.0844620094,-0.040360827,0.1825928986,0.3829545379,0.3910404146,0.3393365741,0.0961863548,0.1874352694,-0.2147234976,0.2771983445,-0.5643305779,0.0346739218,0.1041202247,0.2713558376,-0.1368377209,0.066627875,0.1479776949,-0.2208089978,-0.2162980139,-0.1051368415,-0.2081234902,-0.0779546276,-0.3245044649,0.1794871092,0.0790971294,0.6055165529,-0.0848833472,-0.0521295182,0.1929166764,0.2385609597,-0.1072742119,0.1638332754,0.0137318,0.0599494018,0.0544823632,0.2340850532,-0.3852249384,-0.0247666705,0.0413612984,-0.0779064223,0.2063310295,-0.0665245429,0.0390985645,0.2114246339,-0.2090814412,0.1663435102,0.508505702,-0.109657988,-0.0038654527,-0.6627800465,0.024767397,0.2779552937,0.107672438,-0.0527212545,0.0032308488,-0.265878886,0.0674191788,0.1694182903,0.1585515887,-0.1650961787,0.3916872144,0.2289465219,0.2797166705,0.0727470815,0.2224866748,-0.2273650616,0.2074246705,-0.2910070717,-0.492138654,-0.0160816219,0.1348418295,0.3629411459,-0.0850622207,-0.2391541153,-0.1036297455,-0.1959228367,-0.0248908307,0.3913449347,0.7891902924,0.1488980204,-0.3131633401,0.1699921936,0.0756845623,-0.2671214938,-0.1868343651,-0.3410372734,0.2802948356,0.0750981793,-0.1892602742,0.1544688791,-0.3493214548,0.1605904698,0.1638125777,-0.0757626295,-0.1015292183,0.2639566064,-0.1807930171,-0.2796195447,0.2034140676,-0.1660426855,0.0412304625,0.222171858,0.0821633413,-0.063733995,0.2537777126,-0.3442661166,-0.4547116458,-0.1617778689,0.3301042616,0.1894203126,-0.27367872,0.1826002151,0.2261167169,0.1871689409,-0.1125649959,-0.1956937462,-0.3212908208,-0.2962062657,-0.1822833419,0.2448646575,0.3433954418,-0.4542224705,-0.0933891684,-0.2166294158,-0.2461134046,-0.054044459,0.1466912478,-0.2197429091,0.398642987,-0.1403478235,0.0188561901,0.1012149081,-0.2950314879,-0.4969398975,0.2633554637,-0.2121250033,-0.1468689889,-0.2692807019,0.0484414436,0.1941547692,0.1539271921,0.0051376866,0.0565571487,0.2050610185,-0.2150689512,-0.1294436157,-0.0946026072,0.2298936993,0.3611728549,0.1907626092,-0.0996231213,-0.0038339908,0.0660987124,0.0545602627,0.2124175131,-0.2959643602,0.2606880069,0.2629840374,-0.1679650694,0.0497377403,0.0490054712,-0.3281366825,0.1711824983,-0.1773874164,0.1284162402,0.0088677797,0.2201396525,-0.3857020736,0.124791272,-0.2456311733,0.0568337888,0.2308734208,-0.2148270011,-0.0974920318,0.0471501648,-0.1539621353,0.2841616571,-0.1739644557,0.055631686,-0.071144715,-0.0564676598,-0.3174751997,-0.1609033346,0.2278515995,0.1587961763,0.1742808223,0.0216859486,0.0050521316,0.3554823399,0.0803116187,-0.0317420475,-0.2111491263,-0.4885710478,-0.0345933922,-0.1139063314,0.0384387746,0.102731511,0.1070292294,0.1205408573,0.4304286242,0.2363624573,-0.1677540839,-0.0515252911,-0.0633639246,0.1917677373,0.1616193652,0.0279592704,-0.0829842314,-0.0570030771,0.4093349576,0.0435367897,-0.0650580376,-0.0590648204,-0.0195259042,0.014713522,0.4898835123,0.1344736814,0.0625171438,-0.1088181809,0.2517293394,0.1401916444,0.0302508697,0.1328377724,0.277749151,0.1064233333,-0.4235241711,-0.0366860069,-0.0264514554,-0.2671252787,0.0989147648,-0.3150690496,-0.1128067076,0.1915855706,-0.0246155076,-0.162846759,-0.1511110663,-0.2533967495,0.1131062135,0.344355613,-0.1373842359,-0.0577376187,-0.1854242533,-0.2737466395,0.3402540386,-0.0681999996,0.0371059217,-0.316401571,0.2217026651,-0.0770749301,0.138763383,0.3388938606,-0.0285247955,0.1094074845,0.3005981147,-0.2708330452,-0.0180830099,-0.2160334438,-0.1804675907,0.1515997946,0.111972183,-0.2137085795,0.0305303,-0.1087111309,-0.0169140063,-0.004822446,-0.3049965501,-0.1138391271,-0.3007988334,0.3057751358,0.2515809238,0.0596834607,-0.3211912513,-0.0032180632,0.1792205125,-0.2130450308,-0.0801892728,0.1580676436,0.1449140161,-0.0205494296,-0.204799369,-0.4926880896,-0.5303481817,-0.4888476729,-0.0292072166,-0.0471295975,0.1707663089,0.3728390336,-0.0082199024,0.3335877657,0.1069243327,0.1427049488,-0.0587089285,-0.2668074667,0.2721175849,0.0249675456,-0.3096513748,0.1576294154,0.1298577338,0.3517134786,-0.2375917286,-0.1885915101,0.1306701601,0.0096215066,0.0643238276,-0.2209012657,-0.1836156547,0.0529978499,-0.0396089479,-0.0411025397,0.0911995918,-0.2069228888,0.2017320096,0.0643187687,0.1448442787,0.0772933438,0.2657338977,-0.0761375427,0.5265434384,0.0392807871,0.1855731457,0.2640739977,0.130462423,0.134970814,-0.0506642237,-0.5398096442,-0.0671861321,0.4233142436,-0.1721684039,-0.0361371003,-0.3168293238,0.1760747284,-0.2813004851,-0.0863358006,-0.1637949795,-0.1836038679,-0.1436770856,-0.2176184952,0.003678957,-0.0227848347,0.3067978621,0.1663006693,-0.0702654868,0.0677278712,0.3873063624,-0.1091511622,-0.0157409012,-0.2296463996,-0.2658676207,-0.2586579919,0.3157219291,-0.2792370319,0.3146839142,-0.1532648057,-0.0428509377,0.1665743738,-0.0807461962,0.4986982942,0.0501148775,-0.0191343743,0.1252700835,-0.2378567606,-0.0747508928,-0.1307845712,-0.1986892968,0.1724558771,-0.1019118577,0.233688727,-0.2286940366,0.1395965666,-0.1511369348,-0.1400825977,-0.2261397243,-0.4447718859,-0.4023564756,-0.2775964439,-0.3293209672,0.0714741349,-0.0625686571,0.3856148124,-0.0554290116,0.0615473762,-0.0310035963,-0.3100096881,0.092790179,0.0753413886,0.3129445314,-0.1278225631,0.0731345192,-0.1084556058,0.0209055487,0.0822391137,0.2575199902,-0.1618361473,-0.1628741324,0.0495073274,0.0791191757,0.1855982542,0.0263856649,-0.402969569,0.2586004734,0.0187967625,0.0338427462,0.3420662284,-0.1090370268,0.1140786931,0.0074263974,0.1009620354,-0.2298306525,-0.0616125651,-0.058305379,0.0528999977,0.1004934162,-0.0807929933,-0.0299653281,0.2240544558,-0.1453294307,0.506583333,-0.4068631828,-0.0270505585,0.3145805299,0.1090787724,0.4495962858,-0.0112947337,0.3527529538,-0.508730948,-0.3376002908,0.1896848232,-0.0802515373,-0.020310564,-0.4732180238,-0.5903365016,0.2061620504,-0.3476190567,0.2962293029,-0.0182416569,0.1656645536,-0.1126884371,0.1834341139,-0.2206294388,0.2286557853,-0.1790461391,0.247985974,-0.2823167741,-0.0033894579,0.1613980383,-0.1582825631,-0.1249284893,0.285826087,-0.1620238274,0.2359033674,0.203013733,-0.1350350231,0.1564134955,0.4539047778,-0.1938336343,0.283121109,-0.0773664489,0.0888468996,0.0780039355,-0.1140508503,-0.1461127847,0.2270624936,0.4083356261,-0.0326155946,-0.1881405115,0.6344335675,0.0801858529,0.0685991719,-0.2457523346,0.1895558685,-0.0615431778,-0.2341832668,-0.2373738438,-0.3077602386,-0.0010710877,-0.1609265208,0.1954807043,-0.0249681585,-0.1262477785,0.0540262945,-0.08676783,-0.4619655311,-0.0013346053,0.3493010104,0.0624585226,0.234109655,0.7029871345,0.3712255657,-0.0397965647,-0.3196722865,-0.0188007914,-0.0711913705,0.0093136588,0.2136898339,0.1183323935,0.0440297686,0.3165587485,0.2889674604,-0.0132552655,0.0372627228,0.0755368769,-0.380961895,-0.3957485855,0.2548047304,-0.2124716341,0.0100132292,-0.1032276899,0.1452225745,0.0724800304,0.1129650176,-0.4067446589,0.019754393,-0.5416104198,-0.1763852984,-0.0903903693,-0.1376826316,0.0024854471,0.2650108039,0.2780073285,0.1707057208,-0.2160606831,-0.3226590753,-0.0614078231,0.0523257367,0.1314618289,0.1478700936,-0.2413240224,-0.3203828633,-0.0740061328,0.3170636594,-0.1286108643,-0.0382293165,0.0708732903,0.3496511579,0.0035032725,0.0252895448,-0.0936604515,0.2850466967,-0.1372584254,-0.0343026966,0.0765368044,0.121716775,0.1666172147,-0.0452565625,-0.3469081521,0.0563661084,0.1295736283,0.0026262011,-0.030915482,-0.1531497389,0.0734745041,-0.1278023869,0.1320957392,0.5306506157,-0.225074023,-0.2179568708,0.0909496248,0.3123808801,-0.2310650349,0.2480469495,-0.2824618816,0.0132380547,-0.0537811294,0.1957368106,-0.2150955349,-0.1293255091,-0.0300089587,0.1327185929,0.136807695,0.1238769218,0.0107436711,0.3166515827,-0.2526705861,-0.1326469332,0.252402097,0.183785215,-0.0579355136,0.4016678333,-0.379914999,0.0427691974,-0.1585952193,0.3562553227,0.1159767285,-0.3820240498,-0.0618344136,0.3805193007,0.0734974816,-0.4506339729,-0.3727971613,0.4307152331,0.1382676959,-0.0312939622,0.091887705,0.2125879079,-0.4361151755,-0.2103605419,-0.1892086416,-0.0458007418,-0.1491173953,-0.0098613715,0.0262135956,0.0587635562,0.1239584982,0.2194059044,-0.006783633,-0.3287954628,0.2241207361,0.244951278,-0.3615393043,-0.2153161913,0.2987223864,0.0637276694,0.2853659689,0.3019364178,-0.1928016543,0.281584233,0.4341529012,-0.3539808095,0.0686111972,-0.0249796696,-0.0522812642,0.0304837395,0.3618318439,0.3311666548,0.4536408186,0.3099969327,0.1705660522,-0.1611032784,0.1405387074,0.1924778372,0.0576292202,-0.0661225244,0.5304341912,0.0419044942,0.0930425674,-0.1055197865,0.2225124091,-0.4005809426,-0.1301390082,0.1982863992,0.2625137269,0.1857812256,-0.2342473865,0.1494355351,-0.2071273029,0.3105294108,0.3286698759,0.5006896257,-0.4303507209,0.1608488262,-0.4329150021,0.2768636048,0.1256214976,-0.1307919025,-0.1068229303,0.1194247156,-0.111103788,-0.0143294847,0.1969831884,-0.0060373223,0.1079833359,-0.1549903154,-0.2892766893,0.1259484738,0.0802692324,0.0460333191,0.0018901629,-0.0472125784,0.0844692513,-0.4870571196,0.1282548308,-0.3443618417,-0.3882578909,0.2019646019,-0.043356806,0.2446198463,-0.0467390828,0.1558375806,-0.0514056198,0.0405545458,-0.0031453269,-0.1967606246,-0.14371939,-0.0071743135,0.2906820774,-0.0417024493,-0.1355441064,0.2434600145,-0.2620758116,0.2542633116,-0.1207381412,-0.0912974998,-0.2947988808,-0.0520454496,0.1674501002,-0.0951958746,0.1116770878,0.229479596,0.0491243228,0.2940317988,-0.5586979389,-0.1086154431,0.5152429938,0.093815662,-0.0625304431,0.0418021679,0.1600454897,0.0820968747,0.0147416862,-0.2646854818,0.1551657021,0.2506594062,-0.0373107567,0.0082782172,0.3977273405,0.173977077,0.1924737543,-0.1579850465,-0.1946321726,0.2570955455,0.1329276264,-0.0124456529,0.063294217]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1167","title":"\u2753 On-the-fly tokenization with datasets, tokenizers, and torch Datasets and Dataloaders","comments":"We're working on adding on-the-fly transforms in datasets.\r\nCurrently the only on-the-fly functions that can be applied are in `set_format` in which we transform the data in either numpy\/torch\/tf tensors or pandas.\r\nFor example\r\n```python\r\ndataset.set_format(\"torch\")\r\n```\r\napplies `torch.Tensor` to the dataset entries on-the-fly.\r\n\r\nWe plan to extend this to user-defined formatting transforms.\r\nFor example\r\n```python\r\ndataset.set_format(transform=tokenize)\r\n```\r\n\r\nWhat do you think ?","body":"Hi there,\r\n\r\nI have a question regarding \"on-the-fly\" tokenization. This question was elicited by reading the \"How to train a new language model from scratch using Transformers and Tokenizers\" [here](https:\/\/huggingface.co\/blog\/how-to-train). Towards the end there is this sentence: \"If your dataset is very large, you can opt to load and tokenize examples on the fly, rather than as a preprocessing step\". I've tried coming up with a solution that would combine both `datasets` and `tokenizers`, but did not manage to find a good pattern.\r\n\r\nI guess the solution would entail wrapping a dataset into a Pytorch dataset.\r\n\r\nAs a concrete example from the [docs](https:\/\/huggingface.co\/transformers\/custom_datasets.html)\r\n\r\n```python\r\nimport torch\r\n\r\nclass SquadDataset(torch.utils.data.Dataset):\r\n def __init__(self, encodings):\r\n # instead of doing this beforehand, I'd like to do tokenization on the fly\r\n self.encodings = encodings \r\n\r\n def __getitem__(self, idx):\r\n return {key: torch.tensor(val[idx]) for key, val in self.encodings.items()}\r\n\r\n def __len__(self):\r\n return len(self.encodings.input_ids)\r\n\r\ntrain_dataset = SquadDataset(train_encodings)\r\n```\r\n\r\nHow would one implement this with \"on-the-fly\" tokenization exploiting the vectorized capabilities of tokenizers?\r\n\r\n\r\n----\r\n\r\nEdit: I have come up with this solution. It does what I want, but I feel it's not very elegant\r\n\r\n```python\r\nclass CustomPytorchDataset(Dataset):\r\n def __init__(self):\r\n self.dataset = some_hf_dataset(...)\r\n self.tokenizer = BertTokenizerFast.from_pretrained(\"bert-base-uncased\")\r\n\r\n def __getitem__(self, batch_idx):\r\n instance = self.dataset[text_col][batch_idx]\r\n tokenized_text = self.tokenizer(instance, truncation=True, padding=True)\r\n return tokenized_text\r\n\r\n def __len__(self):\r\n return len(self.dataset)\r\n\r\n @staticmethod\r\n def collate_fn(batch):\r\n # batch is a list, however it will always contain 1 item because we should not use the\r\n # batch_size argument as batch_size is controlled by the sampler\r\n return {k: torch.tensor(v) for k, v in batch[0].items()}\r\n\r\ntorch_ds = CustomPytorchDataset()\r\n\r\n# NOTE: batch_sampler returns list of integers and since here we have SequentialSampler\r\n# it returns: [1, 2, 3], [4, 5, 6], etc. - check calling `list(batch_sampler)`\r\nbatch_sampler = BatchSampler(SequentialSampler(torch_ds), batch_size=3, drop_last=True)\r\n\r\n# NOTE: no `batch_size` as now the it is controlled by the sampler!\r\ndl = DataLoader(dataset=torch_ds, sampler=batch_sampler, collate_fn=torch_ds.collate_fn)\r\n```","comment_length":63,"text":"\u2753 On-the-fly tokenization with datasets, tokenizers, and torch Datasets and Dataloaders\nHi there,\r\n\r\nI have a question regarding \"on-the-fly\" tokenization. This question was elicited by reading the \"How to train a new language model from scratch using Transformers and Tokenizers\" [here](https:\/\/huggingface.co\/blog\/how-to-train). Towards the end there is this sentence: \"If your dataset is very large, you can opt to load and tokenize examples on the fly, rather than as a preprocessing step\". I've tried coming up with a solution that would combine both `datasets` and `tokenizers`, but did not manage to find a good pattern.\r\n\r\nI guess the solution would entail wrapping a dataset into a Pytorch dataset.\r\n\r\nAs a concrete example from the [docs](https:\/\/huggingface.co\/transformers\/custom_datasets.html)\r\n\r\n```python\r\nimport torch\r\n\r\nclass SquadDataset(torch.utils.data.Dataset):\r\n def __init__(self, encodings):\r\n # instead of doing this beforehand, I'd like to do tokenization on the fly\r\n self.encodings = encodings \r\n\r\n def __getitem__(self, idx):\r\n return {key: torch.tensor(val[idx]) for key, val in self.encodings.items()}\r\n\r\n def __len__(self):\r\n return len(self.encodings.input_ids)\r\n\r\ntrain_dataset = SquadDataset(train_encodings)\r\n```\r\n\r\nHow would one implement this with \"on-the-fly\" tokenization exploiting the vectorized capabilities of tokenizers?\r\n\r\n\r\n----\r\n\r\nEdit: I have come up with this solution. It does what I want, but I feel it's not very elegant\r\n\r\n```python\r\nclass CustomPytorchDataset(Dataset):\r\n def __init__(self):\r\n self.dataset = some_hf_dataset(...)\r\n self.tokenizer = BertTokenizerFast.from_pretrained(\"bert-base-uncased\")\r\n\r\n def __getitem__(self, batch_idx):\r\n instance = self.dataset[text_col][batch_idx]\r\n tokenized_text = self.tokenizer(instance, truncation=True, padding=True)\r\n return tokenized_text\r\n\r\n def __len__(self):\r\n return len(self.dataset)\r\n\r\n @staticmethod\r\n def collate_fn(batch):\r\n # batch is a list, however it will always contain 1 item because we should not use the\r\n # batch_size argument as batch_size is controlled by the sampler\r\n return {k: torch.tensor(v) for k, v in batch[0].items()}\r\n\r\ntorch_ds = CustomPytorchDataset()\r\n\r\n# NOTE: batch_sampler returns list of integers and since here we have SequentialSampler\r\n# it returns: [1, 2, 3], [4, 5, 6], etc. - check calling `list(batch_sampler)`\r\nbatch_sampler = BatchSampler(SequentialSampler(torch_ds), batch_size=3, drop_last=True)\r\n\r\n# NOTE: no `batch_size` as now the it is controlled by the sampler!\r\ndl = DataLoader(dataset=torch_ds, sampler=batch_sampler, collate_fn=torch_ds.collate_fn)\r\n```\nWe're working on adding on-the-fly transforms in datasets.\r\nCurrently the only on-the-fly functions that can be applied are in `set_format` in which we transform the data in either numpy\/torch\/tf tensors or pandas.\r\nFor example\r\n```python\r\ndataset.set_format(\"torch\")\r\n```\r\napplies `torch.Tensor` to the dataset entries on-the-fly.\r\n\r\nWe plan to extend this to user-defined formatting transforms.\r\nFor example\r\n```python\r\ndataset.set_format(transform=tokenize)\r\n```\r\n\r\nWhat do you think ?","embeddings":[-0.0571537055,-0.0378616676,0.0511645302,-0.0909028575,0.1780218631,-0.0221669544,0.5987785459,0.0828894675,-0.3151782453,-0.0732152164,0.1041466147,0.2624972165,-0.2840220034,0.0880036131,0.1834016889,-0.0820697471,0.2904809117,0.0297533553,0.1173341721,0.1771769375,0.1671274304,-0.3500298262,-0.2116543353,0.0744798779,-0.4641672373,-0.3110953271,-0.2085753232,-0.3593772352,-0.1359574348,-0.4139526784,0.1105546877,0.321094662,0.5333424807,0.5197013021,-0.0001220133,0.0060748109,-0.0021984389,-0.003848559,-0.2255186588,-0.0814791322,0.5955740213,-0.2086483538,0.0416740924,-0.1967036575,0.1241504401,-0.3029721975,-0.0499146953,-0.3180963397,0.4491649866,-0.1373936236,0.0502252616,0.3842606843,-0.1409321427,0.1993175298,-0.0296621621,0.234946698,-0.0849569961,-0.2966648936,0.3195873201,-0.1833726019,-0.2352818847,0.0610689744,-0.1143956408,-0.3018173277,0.3582299948,-0.089004688,-0.4720630944,-0.2796250582,-0.1780117452,0.0847608447,0.1384000778,-0.4279212952,-0.5727862716,-0.4943040013,-0.0471200868,0.0320064016,-0.3290809393,0.0610616878,-0.0465172306,-0.0024522992,-0.4008657336,-0.1909391135,-0.0380677842,0.1802811027,-0.1636813432,0.3774628043,-0.0297228619,0.2500401437,-0.186584264,-0.241131261,0.3104632497,-0.3286831677,0.1768815666,0.4461234212,-0.2475443184,-0.1767549068,-0.0567022003,-0.388520658,0.2258034199,-0.0604092032,-0.1110222861,0.1192571223,-0.1659703404,-0.0158327799,0.075269565,0.2692531943,0.1779618114,0.2152022719,0.1101557836,-0.4173893631,0.0087661706,0.0452198572,-0.3886713386,0.2294880897,0.0866203457,0.0366172791,-0.3295232952,0.1210333109,0.194229424,-0.0327317379,-0.3979840875,0.0642564446,0.0989573076,-0.0077737379,-0.1188002005,0.0148154534,-0.0021530192,0.1808408201,0.0536755659,-0.0783503056,0.0499117672,0.004377903,-0.1358769089,0.0813597217,-0.0152883288,0.1048120558,-0.0366410352,0.1655039936,0.1909096092,0.0099893352,0.2136267424,-0.0565729141,0.1137497649,-0.1060923561,-0.0452968143,0.1164471805,0.1570580602,0.0914212018,-0.2916375995,0.2853478193,-0.0932707563,-0.2645080686,0.2515757382,0.0332041085,-0.3945994675,-0.2396771014,-0.3217060566,0.3632381856,0.1968056411,0.0486058928,0.0922082514,-0.0612006113,0.1684160531,0.0173153542,0.3184633851,0.2123426944,-0.2312633395,-0.3678816259,0.6276065111,0.1510909647,0.3359800875,0.0259950589,-0.4712953568,0.4807199538,0.0094875405,0.2042948157,0.3170516491,-0.3109449744,-0.0013000687,0.2392292023,-0.1839888245,-0.0250468124,-0.0423632786,0.021418795,0.4540833831,0.096150361,0.0890528262,0.5082175136,-0.0558253229,0.0745907426,-0.0885538831,-0.2710236609,0.3134452403,0.089934729,0.1590591669,0.0109740319,-0.0871456414,0.2193920761,-0.0712971911,-0.1872736961,0.178572759,-0.1451717019,-0.1254775524,0.1383187026,0.019939715,-0.0620585792,-0.1841981709,0.2015856802,-0.0898046345,0.1972798556,-0.1764650941,-0.2715864778,-0.0019044264,0.1634167284,-0.0177414566,-0.0604430363,-0.034226656,-0.217196703,-0.0614103936,0.0688114613,-0.4198598862,-0.0446384512,-0.3517662287,0.1824168563,-0.0726610571,0.1820559502,0.1786254346,-0.0246587824,-0.2389157861,0.1298804581,0.0365737751,-0.2153056264,0.0737867802,0.2548118532,0.0257010628,0.0530781597,-0.1508049965,0.4075115621,0.2054130733,0.2179683,0.064851746,0.0144562926,0.0228054821,-0.011162349,-0.0500633828,0.6243644953,0.1738144457,0.4188345075,0.0587214492,-0.0847589746,0.1361261606,0.1478969902,-0.2176665515,0.1060156077,-0.1907912642,0.1404688805,0.6474385262,-0.0946735144,-0.0212153215,-0.0418491513,-0.1164316386,-0.1920311451,0.0326263793,0.0857618302,-0.4482505023,-0.1540810168,0.0124109471,-0.2460067868,-0.039687328,0.0382439047,0.1952546984,-0.0384145379,0.0311054252,0.0237772651,0.1196404099,0.2907427251,-0.095278427,-0.0264544506,0.0945417434,0.3982783258,-0.1978990138,-0.2764557004,0.1561102569,0.0813878402,-0.0881064311,0.2645244002,-0.1462126374,0.0312485229,-0.0986711904,-0.5169739127,-0.4592525661,-0.0376530886,0.0611969307,-0.1395250559,0.1332175732,0.3374570906,-0.1167318895,0.6314311028,0.0603885092,-0.2420101464,0.2137720436,-0.2343087494,-0.0964066461,-0.0841436535,0.2045982033,-0.3338748217,0.2849550843,0.2845368385,0.1369119436,-0.1833368987,-0.4072165787,0.3090746701,0.0721441135,0.1843277961,-0.0469122864,-0.348541677,-0.0848900974,-0.2920381129,-0.0587293468,-0.2009663731,-0.0947476774,-0.0253325906,-0.0770390704,0.0565334223,0.0295351259,0.089028582,-0.3233362436,0.0183351617,0.0809076726,-0.0284002218,0.0759999156,-0.0228999108,-0.2220890075,0.1977068633,-0.2201230973,-0.061893113,0.0560037978,-0.4251791835,0.3280382752,0.0323309265,-0.1505910605,-0.2655514181,-0.2750265598,0.146408543,0.4522670805,-0.2196106166,-0.0494643524,-0.1388583183,0.1196654662,-0.2595689893,0.1129310578,0.3263929188,0.0525319725,0.1158334464,0.0058208667,0.2495071441,-0.0138157206,-0.255254209,0.0529872328,0.2193771899,0.3225408792,0.2826948464,1.0907583237,-0.2653211951,-0.4417159557,0.1172583252,-0.0038027915,0.0831062868,-0.0358712934,-0.2238188386,-0.1119775623,-0.3827898502,-0.0429907702,0.2087297887,0.1094159707,-0.0801464096,-0.0020278494,-0.0019526454,-0.0970236659,-0.2663558125,0.2293047756,-0.4422942698,0.413271904,-0.0941940099,0.254963398,-0.4347328842,-0.0247466657,0.084849298,0.2052984685,0.4099351168,0.0938752145,-0.5661994219,0.23538436,-0.2332434952,0.1201211214,0.0206235778,0.7507368326,0.3363958895,-0.2519698441,0.1290071011,-0.1288745552,0.3085372448,0.3416008055,-0.1634880006,0.1156839281,-0.2812089324,-0.2312756032,-0.017970996,-0.1686773151,0.248106122,0.1464394927,0.455339849,-0.2253693938,-0.2425400317,-0.3524317443,0.2437627017,0.0014685674,-0.218692407,0.0571224913,0.2255403996,-0.5106260777,0.3068697751,0.2126403153,0.0954997614,-0.0619946271,0.0917518735,-0.4535755217,-0.085020639,0.0770285875,0.0725757033,-0.1561036706,0.0679719225,0.2716444731,0.096051693,-0.1259208918,-0.013309407,-0.0631100461,0.1258790493,-0.617074132,0.0007479958,0.2541421056,0.411850512,0.1819610894,0.024055738,0.4859575927,-0.3031937778,-0.0094072195,-0.3962802887,0.5464116335,0.0242227558,-0.0252992678,-0.5309437513,-0.4034660161,0.3645172715,0.1773575246,-0.1896819025,0.3228040934,0.0861280411,-0.3093383312,0.460732758,0.3549751639,0.8964489102,-0.0236433428,0.3979175389,-0.2223595828,0.24559021,0.3934746087,-0.4777966142,0.0300308596,-0.3676273227,0.104617551,-0.083175689,-0.0768415928,0.0131577123,0.3261489272,0.0544820242,0.005042518,-0.0008298204,0.4058941007,0.0635089576,0.1367975771,-0.0398488492,-0.4966390431,0.0374905467,0.0334417745,-0.0647483692,0.2863084972,-0.1428420842,0.2848939896,0.1318843365,-0.1304996312,-0.0563553534,-0.1178945154,-0.2638394237,0.1771225482,-0.0635788739,-0.362716794,0.225692451,0.1014468297,0.3591664433,0.3173240125,0.1744779348,0.0409534313,-0.1761881411,0.0043446524,-0.1403595507,-0.2208968699,0.3551127911,0.1750175953,0.1721569747,0.0201683622,-0.0730726272,-0.0931456238,-0.3597799242,0.0895616785,0.7337256074,-0.2959562242,0.1770381629,0.1553662419,0.2993141413,-0.0336980894,-0.0082002915,0.1381346285,-0.213906303,0.1079950035,-0.0951231718,-0.0639024228,0.0267541986,0.2608216405,-0.1543871611,-0.4901171625,0.2969224453,0.0133023523,-0.1729755551,0.0567419082,0.5062622428,0.4875359833,-0.2112668604,-0.1471051574,-0.0453536138,0.0765917748,-0.0697290227,0.3390697837,0.2382781059,-0.2658874691,-0.1052818373,-0.1975066811,-0.5801364183,0.0118163107,0.1268416345,0.2358639538,0.3718901873,0.322717607,-0.2153473496,0.2921592593,-0.1673697531,0.0981537923,0.1467703134,0.3098250329,-0.096134685,-0.1994104534,-0.2798697352,0.0072559277,-0.0761616081,0.1189474314,0.1027215347,-0.1377082467,-0.3678577244,0.22484532,0.4112247825,-0.1652942002,0.0293413084,0.1079046652,-0.0838233232,-0.3125033081,0.2672368288,0.2581844926,-0.0553496294,0.0948166028,0.4142946303,0.4423992336,-0.407929033,-0.101942502,0.2915414274,0.2471206784,0.347856611,0.1327596456,0.1289526373,0.0092224907,-0.0622769557,0.3119682372,0.6525678039,0.147920832,0.0786887184,-0.0482202768,0.3042079806,-0.0931295455,0.072872594,0.3677192628,-0.0939882696,-0.2006838918,0.8060765862,0.08405862,-0.0009937359,-0.2954787016,0.0068542445,-0.340272367,-0.017260028,0.2794938684,0.2354652435,0.4467678666,-0.2362058461,-0.1310013086,0.5702825785,-0.07040824,0.1065291017,0.269148469,-0.1352084577,0.0443971381,0.2258825302,0.2633185983,0.1828677505,-0.0025697099,-0.1095884964,0.7173179984,0.1554578245,-0.0087364279,0.103907682,-0.3417069316,-0.1907453239,-0.1912746578,0.487100035,0.2507338524,-0.2644141912,0.4150304198,-0.0215167105,-0.056713488,0.160033673,0.3881199956,-0.0895279199,-0.1384508461,-0.1139998138,-0.0830821022,-0.2152475715,-0.0731716901,-0.0566966496,-0.3669995666,-0.3092119992,-0.0178291779,0.0673569664,-0.1642953753,0.0782455876,0.2478358299,0.2222848088,-0.2409062535,0.0247607697,-0.0885452554,0.2468532324,-0.1093225628,-0.0642320216,0.0747654587,0.1818754226,-0.3244398534,0.082891725,-0.3606896996,0.2744425833,0.0858662203,0.1658396572,-0.2622159421,-0.1543387771,0.1312732399,-0.0612484515,-0.0922188312,-0.1838851422,0.0922465175,-0.3349145651,-0.1584737599,0.5190867782,0.1012964919,-0.095188275,-0.0281848125,0.1516730338,-0.2014421523,0.1195520014,0.3971827626,-0.1025988683,-0.0166327152,-0.0552155972,0.0153927393,0.0182225145,0.3347089291,0.2925766706,0.0867856741,-0.1991381645,0.0263314396,-0.3334692121,0.0017622345,0.2479094565,-0.0928166211,-0.267165035,-0.1437853128,0.159813568,0.219634667,-0.3998706639,0.0179351829,-0.2632228732,0.1405721754,-0.1145288125,0.1985363662,-0.2804019749,0.2988177538,0.0378649309,-0.0321961083,-0.1026839763,0.2618412077,-0.1494312733,0.1269631386,-0.2775554955,0.0672771037,-0.4449775219,0.462349385,0.061738845,0.503215909,0.0640159175,0.1754671335,-0.0248742066,-0.003424665,-0.3323584795,0.1869727075,-0.2027464062,0.0324722007,-0.1831374317,-0.4176796973,-0.2072975039,-0.3010694683,0.0893365368,-0.4719409347,-0.4294627011,0.1154173762,-0.3035817742,0.0210234039,-0.034586288,0.5569841862,0.0185216218,0.3111116588,-0.2615813613,-0.1516501158,0.3447657824,-0.4638013244,-0.2019296139,-0.2953728139,-0.2560539246,0.3270575106,0.1150664166,-0.4861240089,-0.0298658535,0.4179374278,-0.1744677126,0.1982419938,0.0635631233,0.1041086912,-0.2349909097,-0.3590208292,0.5853841901,-0.1504399031,0.1299432665,-0.0261253435,-0.3002541065]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1115","title":"Incorrect URL for MRQA SQuAD train subset","comments":"good catch !","body":"https:\/\/github.com\/huggingface\/datasets\/blob\/4ef4c8f8b7a60e35c6fa21115fca9faae91c9f74\/datasets\/mrqa\/mrqa.py#L53\r\n\r\nThe URL for `train+SQuAD` subset of MRQA points to the dev set instead of train set. It should be `https:\/\/s3.us-east-2.amazonaws.com\/mrqa\/release\/v2\/train\/SQuAD.jsonl.gz`.","comment_length":3,"text":"Incorrect URL for MRQA SQuAD train subset\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4ef4c8f8b7a60e35c6fa21115fca9faae91c9f74\/datasets\/mrqa\/mrqa.py#L53\r\n\r\nThe URL for `train+SQuAD` subset of MRQA points to the dev set instead of train set. It should be `https:\/\/s3.us-east-2.amazonaws.com\/mrqa\/release\/v2\/train\/SQuAD.jsonl.gz`.\ngood catch !","embeddings":[-0.0552403182,-0.166327551,-0.1727203429,-0.0637327582,0.2072728425,-0.0070914794,0.2826894522,0.0058715278,-0.3507852256,0.2180649489,0.0904231966,0.1793125272,0.0837411359,0.2786473036,0.3175214231,-0.0123761678,0.0094548389,-0.1937505305,-0.0140152127,-0.0992432311,-0.0783594549,0.5605151653,-0.2973468304,-0.0290464312,-0.1097000465,0.0327517018,0.113586098,0.2550131679,-0.2352006733,-0.2739777565,0.2358432263,-0.028490996,-0.0124402409,0.3646531999,-0.0000995497,0.0634824559,0.2255186141,-0.0002692446,-0.1769288033,-0.1462752074,-0.0886171088,0.2381099313,0.0108843334,-0.0585447513,-0.2912882268,0.0927694961,0.1109422222,0.0822442174,0.5693241358,0.3872354925,0.2980988324,0.3791337907,-0.1807916015,-0.1960732639,0.0742948204,0.1137739718,-0.1171203777,0.0224134568,0.3136518598,0.0652306974,-0.019535251,0.365175724,0.2343328297,-0.0958289355,0.2819946408,0.2544850111,-0.1159142181,-0.2061252296,0.1252987683,0.0386536419,0.3886876702,-0.1160891205,-0.3243901134,0.1387892216,-0.0330498964,-0.003793353,0.158671841,0.3097504973,-0.1513158679,0.1436237097,-0.3737303615,-0.1032634303,-0.123152785,0.2940821052,-0.0577478521,-0.0308723003,-0.2327200174,-0.1604318619,0.0168047398,-0.0504192114,-0.2183589041,0.227291435,-0.1096555442,0.1440701634,-0.233064279,-0.2809363902,0.1279051751,-0.3587923646,0.0613694564,0.0848114863,0.042283982,0.1692130119,0.087216258,0.0677234828,0.0674557909,0.2575228214,-0.0809514076,0.045349557,0.3059437275,0.2517853379,-0.1283872277,0.058576867,-0.0723256245,-0.1540757716,-0.3388608098,-0.0854250565,0.1630456597,-0.157943517,-0.1150481403,0.0258657113,-0.2419040501,-0.1617324203,-0.0592028275,0.4353732169,-0.0034460733,-0.430795759,-0.0095314998,0.2882061899,-0.1826164126,0.0232082866,-0.3798113167,0.1341443062,-0.1887079477,0.0660412908,0.052648101,0.1080670431,0.4600808918,-0.1397121102,0.2253501117,-0.0114113158,0.1610430926,0.4340673685,-0.0573824942,0.0689048991,0.2259973586,0.0191507898,0.0644679144,-0.022454299,0.051269453,-0.1159109622,-0.2009708732,-0.3352335989,0.0159756355,0.3021095097,-0.1037573293,0.028547246,0.0557489209,0.1501010209,0.01086039,0.0841697156,0.0455283262,0.3302441239,0.0635770857,-0.1118278578,0.0982566774,0.2011521906,0.0898438022,-0.4015395641,0.4042726159,0.105088003,0.0552513488,0.3364303708,-0.321562469,0.0163858682,-0.282237947,-0.0253178272,0.3448059559,-0.234590441,-0.2023515403,0.215350017,-0.435898006,-0.3520142138,-0.0059922594,0.1053779051,-0.0499786586,0.1330472678,0.3951580226,0.3601233661,0.0263667349,0.0947615877,-0.3339544237,-0.2327303439,-0.2730420828,0.2493669689,0.1039048806,0.1565434188,0.0715312511,-0.0730860382,0.341375649,0.0949292332,-0.1600126773,0.2321185768,0.2250039726,-0.190168187,-0.0356905684,0.0304442439,-0.3336405754,-0.0431317501,-0.6602592468,0.4274292588,-0.0797381178,-0.1100947037,-0.1777488291,-0.3464308083,-0.2043916285,-0.4487955272,0.2531247735,-0.009742734,0.2730597258,0.0991960987,-0.2940689325,-0.2685540318,0.0958998203,0.2805000544,-0.3091254532,0.4968037903,-0.3257849813,0.0057510766,0.3150213659,0.3468263447,0.1095383167,-0.2397946268,0.0174365062,0.2098353058,-0.2330329567,0.2567083836,0.4213255048,0.0900400877,0.2157733589,-0.2719132006,-0.0336264819,0.0711703449,-0.0986601189,0.011987946,-0.2870590687,0.2975179553,-0.0822956488,0.1605347693,0.0600144304,0.1667090207,0.2051091194,-0.0191031806,-0.1527727097,-0.1373783946,0.0051380135,0.2125653327,-0.0522059947,0.0070167147,-0.1629202962,0.2086453885,0.310164392,0.0003241964,-0.0262615643,0.0174181331,-0.107245788,-0.0568379238,-0.1262924224,0.2562229335,0.1881125569,0.179491654,0.1131392643,0.1133510396,0.1256258339,-0.3607001007,0.1679818183,0.2028134614,0.0297890287,-0.0337634273,0.1998580247,-0.0628461018,-0.2783399522,0.1908360571,0.0960575268,0.1059952676,-0.2211149186,0.1438060105,-0.2500985265,-0.3494679034,-0.1924685538,-0.3878750801,-0.347975105,-0.3411156535,0.2402498573,-0.256739229,0.0812605545,0.2684015632,0.1831965446,0.2524886727,-0.1170591116,-0.0385477468,0.1008504629,-0.0857447833,-0.2534776628,0.2537674308,0.1993772238,-0.0103415977,0.011773292,-0.3335173726,-0.1434890032,-0.3254334033,-0.4049886167,0.1849654764,0.0126226302,0.3085076809,0.3910367191,0.2020207196,-0.200432092,0.015707491,0.2996630073,0.0991897136,-0.3441697657,-0.1480030417,-0.0929617286,-0.3018954694,-0.1144485027,-0.5709872842,-0.0165065229,-0.181750983,0.1202682927,-0.1196482778,0.0752943978,0.2520777881,-0.0928833634,0.056161765,-0.0675603598,-0.0874127075,-0.297709167,-0.6424868107,0.2836808264,-0.4765506685,-0.4215047061,0.2517982423,0.0471378081,0.3734665513,-0.2247873247,-0.4206837416,-0.5368637443,-0.0718486309,-0.1058978811,0.0690207183,0.0813108832,0.3004348576,-0.3141121268,-0.2519017458,-0.0432022959,-0.4310556352,0.3179524839,0.3133411705,0.4556294382,-0.2171174139,0.3653045297,0.1866107583,0.6803519726,0.1612847298,-0.0500629283,0.1396222711,-0.2578572035,0.1880905032,0.1220863312,-0.1009006798,0.356372714,0.1040409058,-0.0974542722,0.2364839315,0.1227946505,0.1122864485,-0.2405222058,0.1452206224,-0.212740764,-0.3131842613,-0.2218431979,0.2061431706,0.1693541408,0.2989200652,0.2027079761,0.1248862371,0.1014345959,0.2895374,0.32752496,0.0781725869,0.1478476077,-0.7929028869,-0.2510589063,-0.4361379743,0.0729785487,-0.0892474577,0.0639454126,-0.1128380299,-0.0229926631,0.1647299677,-0.038645722,0.7296904922,-0.141218558,-0.1033320278,-0.0479558669,0.0651701689,-0.164995119,-0.1521252245,0.175909549,0.4647174478,0.3038283885,0.1992969513,-0.1635496765,-0.1326139122,0.1242729574,-0.4057498276,-0.1916021556,-0.1299042404,-0.3487387002,-0.3071278036,-0.2073148191,0.0687863156,0.1016919389,0.2247930616,0.2016624361,-0.2846734226,0.3047398031,-0.2521218359,-0.0751159862,0.2870200276,0.3381626606,0.2844432294,-0.0313247852,0.1965001971,0.2414732128,-0.0367712118,0.552799046,0.0187603924,-0.1061570495,0.2186949402,-0.2981671691,0.2395879328,0.4658309519,-0.0647602752,0.0908303708,0.166995883,0.1638801545,0.1063675582,-0.1128830835,0.2369273603,0.0096524144,0.2241993695,-0.1445487738,0.149673,-0.4562055469,0.1058483422,-0.1592703015,0.2357513905,-0.140599817,0.1254597455,0.2778027952,0.8813003898,0.1380236149,0.1250957698,0.1304696053,-0.0348007157,0.1563376933,-0.2026730925,0.2882745564,-0.4168681204,0.1467998922,-0.0266056489,-0.1947803944,-0.0542650409,0.0785172731,-0.1897347569,0.2415062636,0.2073028982,0.2575924098,-0.017173443,-0.0172537491,0.101614587,-0.1038193777,-0.342517823,0.2468546778,0.113675192,0.4774712622,-0.1707638055,-0.2849980891,-0.1673181355,-0.1324050426,-0.0365032256,-0.0021396298,-0.5267632008,-0.0676548108,0.4035550952,-0.0757363141,0.0281788278,0.3880539536,0.3161689341,-0.1025392041,-0.214391306,0.3123577237,-0.0629471242,-0.0321107171,-0.0168785471,0.0636971518,0.038561698,-0.3694817126,-0.0466459468,0.3116798103,0.0150096612,-0.0257834475,-0.2381720543,-0.1263307035,-0.2070901245,-0.1573216915,-0.1093521714,0.0228814259,-0.1939035207,-0.2796188593,0.2231097221,0.0542420372,-0.168439433,0.0168946907,0.3399467468,-0.0028103041,0.0371440798,0.4342284799,-0.0791197047,-0.3127822578,0.3251740634,0.3289222419,0.0151244467,-0.3871148825,-0.1042549238,-0.0003841409,-0.1508663446,0.0920649245,0.0212284699,-0.1761325747,0.2160264254,0.341442287,0.04122651,0.2061235607,0.0230481606,-0.3460125029,-0.1884963363,0.1508173943,-0.1070841253,0.0918476433,0.013870121,0.1923068762,0.3087194562,-0.0090174302,-0.4524793625,0.0718831047,-0.4937857687,-0.0963892937,-0.46747455,-0.3178353906,0.4758331478,0.0221550614,0.1605303437,-0.0754515827,-0.4398806095,-0.2781520486,0.0316097476,0.091667667,0.076730065,-0.0740349218,0.1372193396,0.1795110703,-0.3500447869,0.1880050898,0.1041768342,-0.0579758286,0.0960543081,0.0640828162,0.0069779865,-0.2308334857,-0.2639099061,-0.2711797953,0.2827153206,0.0123548377,0.2723995149,-0.0965712592,-0.2267574221,-0.2381961644,0.1303896755,0.2098062932,-0.055625394,0.2665932178,0.2444494069,-0.4509963393,-0.2021183819,0.2355111539,0.0818291381,0.201151967,-0.343626231,0.2857138216,-0.1800400764,0.3418397307,-0.3264828026,-0.1062926054,-0.1897333115,-0.0043549715,-0.0687891021,0.1521324068,0.1892016828,0.1440150738,-0.0568733513,0.3600403368,0.2159683108,-0.2028829306,0.2094957232,-0.102432549,-0.2266426533,0.1281533837,0.1395117193,-0.0723665729,-0.0906583145,0.1037813649,-0.1771114469,0.3303381503,0.054872591,0.1985722929,0.1292032152,0.0515695922,-0.1204103604,0.2046668082,0.2088893652,-0.0319696218,-0.0213297065,0.4358336926,0.0527313575,-0.26678738,-0.2727211118,-0.0076429844,-0.2327268124,-0.0662887022,-0.3737081885,-0.2736905813,-0.0174594317,-0.0715571865,-0.0848641321,-0.2901794612,-0.1183483154,-0.1985728443,0.0239264295,-0.2777355909,-0.1512964964,0.0952168629,0.0705257133,-0.013401947,0.3199039102,-0.0372345261,-0.1124580279,0.3927389383,-0.0626504123,0.3273686171,0.0725422055,-0.0723003447,0.0523835458,-0.122117281,-0.082171604,0.0750239938,0.0300136432,0.1789321005,-0.3136762679,0.2200569212,0.2279718816,-0.2229955047,0.0838181227,0.0151858656,0.2016946226,0.0813060552,0.2690519094,-0.2985616922,0.0162200462,-0.3949178755,0.1134442985,-0.4264189899,0.0554442741,-0.1274918914,-0.0526814014,0.0138872061,-0.1774678081,0.1886038035,-0.0947459713,0.7417615056,0.1786285043,-0.0198006351,-0.1712446213,-0.3244260848,-0.5071848631,0.2088654786,0.2790502608,0.209845677,0.0596339852,0.5093580484,-0.0977253541,0.2247430384,0.1466082186,0.2703066468,-0.0231401771,0.1160735264,-0.4602560997,0.0739963278,0.012671724,0.2702517509,0.0608213246,0.216794908,0.0811249763,0.0154830068,0.0730634555,0.0358672552,0.0469544604,-0.2638999522,-0.0789574906,0.1911784858,-0.120196797,-0.0758202821,0.0703049079,-0.1863765717,-0.1540343016,-0.3964311779,-0.1614504158,-0.002878384,0.0801784322,0.2939460576,0.1172183752,0.0863174275,-0.3843426406,0.2739161849,0.210734129,-0.0687626004,-0.2071549296,0.2352647334,0.1263173968,0.2198413908,0.2032269984,0.3044282794,0.0119290259,-0.0558587685,-0.2682212591,-0.4133523107,0.5802981853,-0.426022172,-0.0192039888,0.004131915,0.2375363857,0.1663128436,-0.1868737638,-0.549538672,0.0746421367,0.1273457259,0.0074233511,-0.0253609959,0.1669954956,-0.1239387766,0.1139176786,-0.2422735393,0.1139123961,0.1487605274,-0.1930104196,-0.0378204882,-0.1858731955]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1110","title":"Using a feature named \"_type\" fails with certain operations","comments":"Thanks for reporting !\r\n\r\nIndeed this is a keyword in the library that is used to encode\/decode features to a python dictionary that we can save\/load to json.\r\nWe can probably change `_type` to something that is less likely to collide with user feature names.\r\nIn this case we would want something backward compatible though.\r\n\r\nFeel free to try a fix and open a PR, and to ping me if I can help :) ","body":"A column named `_type` leads to a `TypeError: unhashable type: 'dict'` for certain operations:\r\n```python\r\nfrom datasets import Dataset, concatenate_datasets\r\n\r\nds = Dataset.from_dict({\"_type\": [\"whatever\"]}).map()\r\nconcatenate_datasets([ds])\r\n# or simply\r\nDataset(ds._data)\r\n```\r\nContext: We are using datasets to persist data coming from elasticsearch to feed to our pipeline, and elasticsearch has a `_type` field, hence the strange name of the column.\r\n\r\nNot sure if you wish to support this specific column name, but if you do i would be happy to try a fix and provide a PR. I already had a look into it and i think the culprit is the `datasets.features.generate_from_dict` function. It uses the hard coded `_type` string to figure out if it reached the end of the nested feature object from a serialized dict.\r\n\r\nBest wishes and keep up the awesome work!","comment_length":74,"text":"Using a feature named \"_type\" fails with certain operations\nA column named `_type` leads to a `TypeError: unhashable type: 'dict'` for certain operations:\r\n```python\r\nfrom datasets import Dataset, concatenate_datasets\r\n\r\nds = Dataset.from_dict({\"_type\": [\"whatever\"]}).map()\r\nconcatenate_datasets([ds])\r\n# or simply\r\nDataset(ds._data)\r\n```\r\nContext: We are using datasets to persist data coming from elasticsearch to feed to our pipeline, and elasticsearch has a `_type` field, hence the strange name of the column.\r\n\r\nNot sure if you wish to support this specific column name, but if you do i would be happy to try a fix and provide a PR. I already had a look into it and i think the culprit is the `datasets.features.generate_from_dict` function. It uses the hard coded `_type` string to figure out if it reached the end of the nested feature object from a serialized dict.\r\n\r\nBest wishes and keep up the awesome work!\nThanks for reporting !\r\n\r\nIndeed this is a keyword in the library that is used to encode\/decode features to a python dictionary that we can save\/load to json.\r\nWe can probably change `_type` to something that is less likely to collide with user feature names.\r\nIn this case we would want something backward compatible though.\r\n\r\nFeel free to try a fix and open a PR, and to ping me if I can help :) ","embeddings":[0.0078380946,-0.2019331455,0.033574339,-0.0459202118,0.3190200329,0.1368784606,0.4551838934,0.4090704322,0.1596739441,0.0305586196,0.4140094221,0.4669063091,0.0361163728,0.5192950964,-0.3772074878,-0.0860040858,0.1542394161,-0.0560006872,-0.0117228627,0.2065544724,-0.4115031362,0.1005299389,-0.1578256488,0.3828088641,-0.1612658799,0.1273504347,0.2941843271,-0.1226475462,-0.1234727353,-0.1726844013,0.2820243239,-0.1374555975,-0.118606627,0.4063162804,-0.0001096195,-0.0599151626,0.4927601516,0.0518484935,-0.2427797019,-0.601084888,-0.2757981718,-0.3708138764,0.3751625717,-0.4024413824,0.1444876641,-0.0466072261,-0.0005312023,-0.4921488464,0.1850102246,0.2458359301,0.2882341444,0.0291186236,0.2679183781,-0.2092796266,0.3826477528,0.5050954223,-0.3839738965,-0.2925162315,-0.0465633199,-0.1764739156,0.5520861745,0.1724968404,-0.1376736164,-0.1723724306,0.3057540655,0.1554484516,-0.2008398026,-0.2484654188,0.1735776216,0.2742727995,0.4289379716,-0.1145282537,-0.2597034872,0.0091515593,-0.1126078069,-0.2941512764,0.2335266918,-0.2032423168,0.0832203552,-0.0831837952,0.0906343311,-0.0753074586,-0.0665570423,0.2105636448,-0.3794054389,-0.1487878859,-0.1780242026,0.2584201694,-0.149717018,-0.45123595,0.0153090945,0.0681741908,0.0999335572,-0.0704969168,-0.3838429451,-0.025802521,-0.1406578124,-0.4826210737,0.1313471347,-0.1265292466,-0.1564218849,0.0742508769,0.0907491371,0.3832581937,0.0959511548,0.1200556904,0.2974463999,0.2780844867,0.2569692731,-0.0820531994,-0.2062655389,-0.2184764594,0.0611144602,0.0515845977,-0.0175609514,-0.3705222309,0.5927597284,-0.21134983,-0.3760260046,0.1572351456,-0.1641093493,0.0052227918,-0.0133907776,0.2600543499,0.0564046986,0.1037949845,0.1861784458,0.3259424269,0.2374450117,0.114186354,-0.1411479712,-0.0845821053,-0.1519916952,-0.1330151409,-0.1320183724,0.174354434,-0.0461566895,-0.0614385307,-0.020113226,-0.0935074314,0.0563111417,0.0819186494,-0.2206217349,0.129203409,-0.2331810147,0.104369089,0.2374669611,-0.2826281786,-0.1736219227,-0.1181725785,-0.3802196085,-0.1871344,-0.5376911163,0.2363945544,0.0096822493,-0.0684196949,-0.1012085825,0.0138355754,0.397441417,-0.1062710732,0.0043603592,-0.0051320344,0.0021375064,-0.2661719322,-0.1595337987,-0.0022617171,-0.2039359063,-0.0401688404,-0.270314604,0.3082981408,0.2345616221,0.2524323761,-0.0577437878,-0.1316457838,0.017906446,0.2042624205,0.5040896535,0.0275670178,0.1945368797,0.2862772942,0.0232302304,0.1791197062,0.0600974634,-0.189326942,0.1734788716,-0.2156347483,0.1905263364,0.1097171605,-0.0205638483,-0.2330801189,-0.0655529574,-0.1836001575,0.4098141491,-0.0864174366,-0.1034864634,0.2569858432,0.1207137406,-0.1224279404,-0.1491516531,-0.1833721697,-0.0162974652,-0.1180430725,0.4970188141,0.3460541666,0.0392823182,-0.2998205125,-0.4126471877,0.151278913,0.0616792031,0.1455200613,-0.325156033,-0.3413262069,-0.2021323442,0.1106054857,0.059676487,0.3594352007,0.1491472274,-0.0784470141,-0.296739608,0.1645853221,-0.3081689179,0.0445623882,-0.0911244676,-0.0029058566,-0.4897051752,0.0612497516,-0.0507755652,-0.0232210848,-0.3492095768,0.1185407415,0.2980417311,0.1508868635,-0.1748881787,0.233707577,-0.0989083648,0.0890000239,-0.1473782957,0.2635775208,-0.04283797,-0.0335714445,-0.0107726539,0.046681717,0.1778206527,-0.1940222383,-0.0110005988,0.3024355173,0.4619303644,0.3414457738,-0.1291917562,0.341183424,0.107732363,-0.1361240596,-0.221174106,-0.3081472516,0.1069941297,-0.1612017006,-0.6002209783,-0.1709656119,-0.3033189774,0.003332302,0.446339339,0.0418112539,0.3868363202,0.2647478282,0.0387372822,0.1389839351,-0.311075598,-0.037021216,0.4198590815,0.019858066,-0.1243733242,0.0913057178,-0.166259557,-0.0383374691,0.2877165675,0.1083028093,0.0079928841,-0.0387195088,0.229022935,0.0168444831,-0.0823425651,-0.1836352199,0.1096568331,0.0418761447,-0.269926995,0.3570947647,-0.3549265563,0.2029594183,0.0300161671,-0.0922260284,-0.1036202759,-0.4741544127,-0.0895498097,0.4128101766,-0.1608472317,0.1376708895,-0.082268618,0.087516807,0.1381702125,-0.3113170564,-0.1339861751,-0.2423387021,-0.0953630358,0.0726690292,0.1683610082,-0.0640012324,0.2009734511,0.2082196325,-0.1990567446,-0.4550857544,-0.4419567883,-0.0097708404,-0.3552289307,0.2507650554,0.4034199417,-0.0415578857,-0.2243301272,-0.2054418474,0.2690586746,-0.0943883434,-0.1374160349,0.41147542,0.0861622915,-0.0328210704,-0.1460819691,-0.3563118577,-0.1373309791,-0.3598007262,0.3746552467,-0.1349002719,0.1130203828,0.0065148915,-0.1802773923,-0.088161163,0.2241382152,-0.019415956,-0.4994114339,0.0871522725,0.2874124348,-0.1887975037,-0.1905930191,0.0700726062,-0.0170530099,-0.2108181119,0.2952624857,-0.1421574503,0.0074794753,-0.1785200685,0.1600124985,0.0358536579,0.0128940819,0.074634783,0.2611310184,-0.1116136909,-0.0501778722,-0.068865478,0.1660798788,0.4550357461,0.2671754658,0.3056822419,0.4348924756,-0.1675058603,0.0459146574,-0.0589106418,-0.1945316643,0.4369627237,-0.0955308676,-0.059444461,-0.1465328485,-0.2718073428,-0.2213056684,-0.3195917606,-0.0297861025,-0.0941236094,-0.2576058209,-0.1811709106,-0.1336782277,0.0806884468,0.0154542252,-0.2368081212,0.1843155921,-0.1802856177,0.1941886991,-0.0336315632,-0.0165207013,-0.0621149652,-0.0999420211,0.0031905805,0.0962001607,-0.1880307943,-0.2310652733,-0.4673822522,-0.3416421413,-0.2804404497,0.6000746489,-0.1661688536,0.3577224612,-0.0791719779,-0.2142569274,-0.0344756842,0.2675849497,0.5227485895,-0.347858727,-0.3228453994,0.1956053376,-0.1217502281,-0.159473747,-0.0842688605,-0.1011983231,0.2628883719,-0.2334238142,0.430203706,-0.2847671509,-0.0198171102,0.1237878129,0.1419466585,-0.1319362521,0.0143825226,-0.1187301651,-0.3412693143,-0.1650189459,-0.2368637174,0.195519805,0.1813248098,-0.1521898955,-0.1367295384,-0.584181428,0.0194472205,0.3518446982,0.1724542379,0.1786127836,-0.4430750608,0.1557589918,-0.0179377776,0.3297305703,0.1664588451,0.3813909888,0.2173552662,-0.2755932212,0.0208719131,-0.1797060072,0.2482610196,0.1482811421,-0.10596174,0.4874870479,-0.1478772014,0.127898261,-0.1236969754,-0.244906202,0.5630016923,0.1281279773,-0.255431056,-0.3126144707,0.5425533652,0.2307957411,0.0061520338,0.1921421289,0.255572319,-0.3521591127,0.6077802181,-0.0389150269,0.9458515048,0.1161478609,-0.0305242762,0.0317024253,0.1691474766,0.5663940907,-0.1443421245,-0.0564985014,-0.2901569605,-0.2272481769,-0.1399281472,-0.0785673708,0.4586822987,-0.1653611809,-0.4343219101,0.3294343948,-0.5950178504,0.4806118011,-0.069977127,0.0144302119,-0.0347239003,-0.1819639504,-0.0143152,0.1080074385,0.0628185794,0.1485927105,0.143881917,-0.0821312442,-0.0015156051,-0.1846720278,-0.4210825264,-0.0368767865,0.311001569,0.4263428152,0.1029006317,0.056073606,0.3445010185,0.0459113866,0.3193943799,0.0747067109,-0.1326287687,0.1109394208,0.2535710931,0.0141708059,0.0445940644,0.0104792956,0.1669882387,-0.0419618711,-0.2480622083,0.2486104816,0.2197608948,0.1183563098,-0.1384518892,0.0728878528,-0.046113275,-0.5918126106,-0.314491868,-0.2867548466,-0.0953657627,-0.3193062544,0.1779658645,0.0198974442,-0.3198503852,0.5103650093,0.01052937,-0.0948977172,-0.0597015172,0.0650437251,0.1726642996,0.0704749823,0.2992478311,-0.0189851597,-0.1513278186,-0.155100897,0.269524008,0.1913001388,-0.1429394037,0.2461666614,-0.1132729501,-0.4089074433,-0.1264115423,0.4697760642,-0.0649934262,0.3313154578,-0.2421823889,-0.3201699257,0.0758778229,0.268535465,0.3235088587,0.0573450923,-0.4495080113,0.0709632784,-0.2450015098,-0.1629460305,-0.3492009044,0.1159223542,-0.2193783224,0.2333666235,0.0386610329,-0.0822999626,0.4548922181,-0.1379771829,0.2539156079,0.1826857775,-0.0597852878,-0.231190905,-0.1873538196,0.1350816786,0.032893192,0.0615717769,-0.1360838413,-0.2783337533,-0.1146056801,-0.0336243697,-0.0390360765,0.4186953306,-0.1318443716,0.1545301825,-0.177819103,0.0494905114,0.1851139367,0.2085568011,-0.0868912265,0.0844728276,-0.1604657173,0.0181879662,-0.0597079135,-0.0194267929,-0.2476194501,0.155608207,-0.2354651988,0.1066055074,0.3695607781,-0.1048661768,-0.2147055864,0.0280260555,0.2179192305,0.7123606801,-0.1586175561,-0.1734547019,0.2851216197,0.1771644205,-0.1732674837,-0.2116282284,0.2943560183,-0.0705116764,0.139190346,0.3272460103,0.169193849,0.0255920291,-0.4620865285,-0.0384220704,0.1326924413,-0.0908572227,0.2415949255,0.2752794623,0.0864611268,0.0974549502,0.2127746642,0.2671547532,0.2359683812,0.4340647459,0.1615923941,0.076536648,0.2801217139,-0.0458556861,0.4383800626,-0.0370011739,0.6159651279,0.1756719351,-0.2650684118,0.2208493203,0.1655959934,-0.0081437565,-0.2301917821,-0.1368788481,-0.2479957044,0.1035616919,0.0234741047,-0.2508146763,0.0629108548,-0.1710419655,-0.3035826385,-0.4562498927,-0.0010046656,0.1679894179,0.4843472242,-0.1335965097,0.2562808096,-0.1984515041,0.3258999288,-0.2088283598,0.3876576424,-0.0023021672,-0.1747999042,0.2231551707,-0.1649699062,0.1909561455,0.0898351967,0.3307883441,0.3177857399,-0.2704296112,0.1409361362,-0.1850466728,-0.2600796819,0.1971627772,0.2103593647,-0.0544071868,0.0183523465,0.1896165013,0.1789619327,-0.157146275,0.0658843592,-0.0593706965,0.3009968102,-0.1975553036,0.2622719705,-0.1181359142,-0.0158591196,0.0726702511,-0.0384616442,-0.3148462772,-0.4191499352,0.501968503,0.2499378175,0.1965204775,0.0647882819,0.1188406721,-0.0179608297,0.2620566487,0.1194232777,0.0324214734,0.1013923362,0.1399133652,-0.1345528662,0.3168327212,-0.3086352348,-0.3098202944,0.5548021197,0.241625756,0.2531351149,0.0350245833,-0.0504580513,-0.1541440785,-0.4290820956,0.1581777781,-0.3153626025,-0.4558658898,0.1596486121,0.0298872981,0.019590091,-0.2860648334,0.2710367143,-0.0433870405,0.0799036175,-0.3420718908,-0.0643090382,0.0267632604,-0.1564316303,0.323613137,0.2117212266,0.1048996896,-0.046154622,-0.1595768332,0.1725401282,-0.3265886009,-0.2769678831,0.3132359982,0.2126336545,0.1480471939,-0.0744599774,0.1867537946,-0.1731048971,0.3277274966,-0.1185579449,-0.3846750855,-0.3629052341,0.2404216379,0.1940721273,-0.1020048633,-0.2771923542,0.2814822793,-0.063097544,0.3445727825,0.1085697263,-0.4713973999,0.7269507051,0.1071700603,-0.6315425038,0.0681725591,0.0471899956,0.3253974319,0.0877632126,-0.4373453856,-0.1800644696,0.2037195265,-0.0663134381,0.0399643555,-0.1014614329,-0.0084693059,-0.1823565513,-0.0703419521,0.1706643254,-0.1790298223,0.1127607077,0.1726850867,-0.0581836477]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1103","title":"Add support to download kaggle datasets","comments":"Hey, I think this is great idea. Any plan to integrate kaggle private datasets loading to `datasets`?","body":"We can use API key","comment_length":17,"text":"Add support to download kaggle datasets\nWe can use API key\nHey, I think this is great idea. Any plan to integrate kaggle private datasets loading to `datasets`?","embeddings":[-0.0929649547,-0.053584218,-0.3102498949,-0.0035680523,0.1154249161,-0.0199873969,0.2132152468,0.1274612695,0.4255777001,0.1014183462,-0.2042617798,0.6420328021,-0.0812081993,0.8870463967,0.0586065575,0.0357894786,0.1376734525,0.1102098078,0.1666734815,0.201852873,-0.1626320183,0.0076732179,-0.1302725375,-0.1597166955,0.1356511265,-0.2504248619,0.017365545,-0.1539834887,-0.3501514792,-0.4008308351,0.2865821719,0.2992354333,0.2788437009,0.1203535944,-0.0000982298,-0.1139250547,0.3490458429,0.0557616614,-0.210727945,0.1748960465,-0.5735141039,-0.2588055432,0.016866181,-0.1636401713,-0.1123944819,-0.2678085566,-0.1713749319,-0.3603084087,0.2351503372,-0.1250749379,0.2732905746,0.006384558,0.0885759369,-0.1921589077,0.1979539096,0.1911397725,-0.1841082871,-0.1639204174,0.4762825668,0.0860301107,0.0310169384,0.2202863246,0.2117949575,-0.1668585539,0.2060696185,-0.0416042879,-0.2298645228,-0.5424913168,0.0190903172,0.0220291577,1.0136518478,-0.1241830066,-0.2577559054,-0.3066722155,0.032376729,0.0141826691,-0.067809701,0.0927636996,-0.0220896937,0.1796204895,0.1324836463,-0.6074897051,-0.2174334824,0.1237905174,0.1043906137,0.5533359051,-0.0011595344,-0.0430266485,0.1158502027,-0.1654866487,0.435978353,-0.1371934712,-0.1044681221,0.1234971806,-0.1767945886,-0.2896084785,-0.0022104767,0.1693391353,0.3592155576,0.3478613794,0.0657119006,0.3563125134,-0.1603371799,-0.0845618248,0.4710317254,0.0696583688,0.2438338548,-0.0263642445,0.4169931412,0.0564147197,0.2899772525,0.0078798831,-0.0219127778,0.0729435608,-0.275755316,-0.1811768115,0.0029459349,-0.0364105552,0.0814337209,-0.1260784417,0.1859476119,-0.1600484848,-0.1126899794,0.1078306511,-0.1048745662,0.1617765427,-0.1908884645,0.043846406,0.2076887339,-0.0576501302,-0.1485186964,0.0058379769,0.0369170681,0.1093908548,0.0658129975,-0.1096902937,0.1760620326,-0.1862022728,0.277864486,0.19964917,0.1841573417,0.3225196898,-0.2182249427,0.0646085739,0.0479777977,-0.1586374342,-0.305888176,-0.1642429233,-0.087016955,0.1231410578,-0.2252591997,-0.2828967571,-0.1214038357,0.3159552217,-0.329862982,-0.2243256718,-0.5454100966,0.3330480456,-0.1434221566,-0.1627181321,0.0098491665,0.3126566112,-0.0290235244,-0.1342370063,0.0432077125,-0.0810189322,-0.10948053,-0.2283750623,-0.1753743589,-0.1984549761,-0.0973824561,0.0002355361,-0.2214530706,0.094040364,-0.2998362184,0.2068417221,0.6396182775,-0.4268696606,-0.3617960215,0.2395075113,-0.1071419567,-0.1990328133,0.0805609971,0.2385319322,0.253870219,0.0339549147,0.0562470555,0.3365939856,-0.0096011767,-0.0481802523,-0.2135483921,-0.4054665864,-0.0370904468,0.2970249653,0.0025687772,-0.0871418267,0.3029793203,-0.1463075727,0.1826406121,0.0095972428,0.1907448471,-0.1902959645,0.4004264176,-0.1466522664,0.1918909699,0.0919801444,-0.3843583167,0.1946885139,-0.3316366374,0.0416459739,-0.0040103919,-0.345194757,-0.1721148938,0.0983393714,-0.3835199475,0.026156133,0.2060831636,-0.2966518402,0.0946116075,0.1269972771,-0.2230920494,0.2289740443,0.2138634473,0.0540144034,-0.175079897,0.2178053409,-0.2335838228,0.2678087056,0.0801347867,0.0497058854,-0.0649685934,-0.1727432609,0.2545925379,-0.0741074383,0.162464872,0.4052411616,0.4697453678,-0.0727103949,0.4805855453,-0.0431595147,0.1689442396,0.0748146027,0.1771881133,-0.0979500487,-0.2475565374,0.3408476114,0.0691556334,-0.0139367543,-0.176382035,-0.0046359641,0.1185565814,0.0576760508,-0.1215807647,-0.1760537326,-0.0643194169,-0.3711441159,-0.0425792746,-0.1604275256,-0.1076765433,0.4074072242,0.0366183445,-0.1014718637,0.1294517666,0.0963313803,-0.0978769809,-0.1545976698,0.3012616932,0.2223402709,0.2467843145,0.3052209616,-0.0063447845,-0.1875217706,0.4511363506,-0.1555742174,0.1747367829,0.2000945359,0.1127392054,0.2868580818,0.0697410628,-0.2519512773,0.0697324872,0.0750761554,0.1321231574,0.1331682503,-0.0068098707,-0.1906313449,-0.105003342,-0.4177584052,0.2547333241,-0.0595765784,0.2357586026,-0.4169063866,0.188446641,-0.0759593993,-0.0838453621,0.1074507907,-0.2154809386,0.4759658277,-0.1823241413,-0.1267191619,-0.2750762999,-0.3052017689,-0.1172137558,0.3365138471,0.2783371508,-0.212425366,0.5011542439,0.131085366,0.2197770774,-0.3135256767,-0.5158466101,0.0511482544,0.008959285,-0.0700863972,0.053938549,0.0763895363,0.1957163066,-0.0425221622,-0.0450359359,0.0193727016,-0.3334476352,-0.3106675446,-0.0499325395,-0.0490736254,-0.1771842837,-0.3822337687,-0.4056794941,-0.3214925528,0.3853613734,0.113257356,-0.0648361221,-0.2443984598,0.1801357716,0.0929392651,0.0071948017,-0.3171281517,-0.2774167359,-0.1570271999,0.28554672,-0.2521851957,-0.3377675414,0.3224461973,-0.1721737832,0.1879161745,-0.2725309134,-0.3888732791,-0.2476118207,0.1317628771,0.3319598138,0.0334973037,-0.0380640849,0.0672867671,0.1975274533,-0.0695056766,-0.1006402075,-0.2368372083,-0.1417509913,0.0269679315,0.0733378157,0.04553454,-0.1579207778,0.0893397182,0.5135681629,0.0981259048,0.140764907,0.3335723579,0.1886803061,0.1842944771,0.0722925067,-0.1296096146,0.0371074416,-0.1021242514,0.1340874434,0.187505126,0.0932982266,-0.4821219444,-0.4905827045,-0.200535357,0.047692325,-0.1638046205,-0.2082970738,0.0326471888,-0.1167183071,0.0588514246,-0.0233443379,-0.0880275816,-0.0061777788,0.1972371489,0.0132877706,0.1718357801,-0.0843722746,-0.4658823013,-0.1954049319,-0.3334221542,0.3679112196,0.1897567213,-0.2230316252,0.0708073527,-0.0562009774,0.0581816509,-0.2285453677,0.6534587741,-0.3065120578,-0.2575014532,-0.1291676164,0.2158357203,-0.0918693021,0.0106993308,0.0724211261,-0.0184760354,-0.2127920985,0.3907854855,-0.1326831877,-0.0177306794,-0.1910142004,-0.0287533924,-0.0749338865,-0.1725342125,-0.0093584722,-0.2202651501,-0.4252130985,-0.1111151427,-0.0637717471,0.0229962394,-0.2125656009,-0.2726647854,0.0464068055,0.0784119591,0.0158583783,-0.0307470188,0.1023207977,0.1993190944,0.1953769028,-0.05780131,-0.026122475,0.1510788798,0.467923671,-0.0197304748,-0.3996259868,-0.1424735337,-0.1883178502,0.0556115024,0.1531094611,0.1956165135,0.2850226164,0.115560554,0.0761394575,-0.1451361179,0.0567434691,0.0219499338,-0.1884119064,-0.2844206691,-0.5027635694,0.5110705495,0.0027662523,-0.1216208041,0.3103560209,0.2278690338,-0.2142793685,0.0076067969,0.1378335953,0.6620998383,0.1231224537,0.1892496049,0.0632906109,-0.2115328014,0.4805825055,-0.295568794,-0.0145670446,-0.2645292282,0.2212959528,-0.1122009233,0.0509034805,-0.0191401057,-0.2788563371,-0.2523945272,0.0697062388,0.0164146703,0.2677139938,-0.1337601095,0.3727672398,-0.2736597359,-0.1749032587,0.0456944779,0.306294769,-0.1826332062,0.1030726433,-0.1702608019,-0.3906962276,-0.1565624774,0.0879911259,-0.0459345542,0.1115319654,0.0642602295,0.2367167175,-0.2668136358,-0.3333693445,0.2618551552,0.304577142,0.247437641,-0.0958679989,-0.4158816636,0.0696768016,-0.1493025869,-0.2111491114,-0.1502452642,-0.049039185,0.0148762343,-0.3142475486,-0.3047282696,0.2433128059,0.0034381596,-0.2426565439,-0.4107021689,0.3009026349,0.1274304092,-0.2916112542,-0.0882865041,0.2000094503,-0.111989744,-0.1669714004,0.2071650326,0.3160638809,-0.0326032378,0.3302985728,0.231081605,0.1423494071,-0.172202006,0.3114984035,-0.0171106234,-0.2345902026,0.0796131864,-0.1526956856,-0.2538977265,-0.3289506435,0.1357594132,-0.0517473221,0.1926960498,-0.1181408092,0.2647686899,-0.0805021897,-0.063677676,0.1663477719,0.0415413827,-0.1377886385,0.2124141306,-0.1667530835,-0.3571164012,0.4417880476,-0.3826167881,0.2800000012,-0.0061101625,-0.219789505,-0.0043859086,-0.2295763791,-0.4121127725,0.0911054537,0.0332789756,-0.2060200274,0.3590273559,-0.0143123334,-0.075270012,-0.0123583805,0.1216102764,-0.1553171426,-0.3481841683,-0.3103874922,0.228262648,0.0620975718,0.2434818447,0.0406570062,-0.0201954506,-0.2606166005,-0.1771944165,0.1396281272,0.2344973683,-0.1229179204,0.0557387732,0.1445401758,-0.0391530022,0.1076482013,-0.0486273654,0.4111316204,0.2773839235,0.0744770691,0.02592857,-0.0422373749,-0.0040139989,-0.0055713588,0.4392646253,0.1945367306,0.2255241126,0.2343066931,0.3847555518,-0.231657356,0.1163105667,0.2075439394,0.0343271941,0.2719973028,-0.192101717,0.3829947114,0.1658796519,0.2990360856,-0.2118804902,-0.3960698545,0.1848542839,0.0814826861,-0.0197086725,0.05713569,0.3609973192,0.1662778705,0.1145679578,0.2274983376,0.2971511781,-0.1147974283,0.1020116434,0.1916823685,-0.0784905404,-0.049157761,-0.1321175247,0.3963246942,-0.005015838,0.2574884295,-0.4771890044,0.3880511522,0.2483060509,0.2273726612,0.2867911458,-0.4313172698,0.1096467227,-0.0853083432,-0.1545745432,0.1048182175,0.3467981815,0.3886000216,-0.0219791196,-0.0655084476,-0.0124445725,-0.0012877268,-0.0404918864,0.1086618528,-0.1578398198,-0.2432581633,0.2660822868,-0.0331961401,-0.0122045437,-0.3073901236,0.0549852848,-0.0621817894,-0.1678737551,0.1594757289,0.1848970801,0.2497632951,0.3459221125,0.0684132501,-0.1249388307,0.0947970673,-0.1988098174,-0.0518659875,-0.0745468214,0.1415677965,0.0313362926,-0.3935882151,0.0955515355,0.1549432427,-0.266766578,-0.1042864993,0.1845569313,0.329575181,-0.3697730303,0.3326388001,0.2323466986,-0.2163144797,0.0680007786,0.3306400776,0.006103457,-0.1204655766,0.0792485774,-0.0088947257,-0.2567465007,-0.0215248968,0.1850334704,-0.1867246032,-0.1252340525,0.0581543632,0.049330309,0.0190477613,-0.1451479495,0.1632635295,-0.1179371476,0.3923142552,-0.0112604238,0.3060814142,-0.0150120882,0.0862742439,-0.2546750605,0.1000448614,0.3030633032,-0.1152527481,-0.2710487247,0.1032600701,0.205179885,0.2681258321,-0.0179239642,-0.2057392895,0.7407826781,0.1356456131,-0.2567977905,0.1258152723,0.3990848958,0.1358109713,0.1246816963,-0.1193980649,0.2117156684,-0.3204461932,0.1264894307,-0.084613651,0.0422974378,0.1486829966,-0.2377163023,0.629942596,-0.2186717391,0.4631664455,-0.0844897181,0.1559531391,-0.1135104895,-0.0303343944,-0.1064897329,-0.1322488338,0.2037771046,-0.1080566943,-0.1990811676,-0.0052766744,0.0534868501,0.0620469004,0.1513766795,0.3160066903,-0.0170425978,-0.0003463161,-0.0172345582,-0.2140434831,0.1000687554,0.028297158,-0.1393904686,-0.3136107922,-0.1487105489,-0.2686765492,0.2581799924,0.0730011389,-0.0763814449,-0.0326306894,-0.1309964657,0.142015174,-0.0084375935,-0.3998366594,0.109104529,0.2313683629,-0.1604682803,0.1279533505,0.3397709727,0.1807638556,-0.0602821559,-0.3250398636,0.3017235696,0.0389069878,0.0691498145,-0.1928170025,-0.2094547153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1064","title":"Not support links with 302 redirect ","comments":"Hi !\r\nThis kind of links is now supported by the library since #1316","body":"I have an issue adding this download link https:\/\/github.com\/jitkapat\/thailitcorpus\/releases\/download\/v.2.0\/tlc_v.2.0.tar.gz\r\n\r\nit might be because it is not a direct link (it returns 302 and redirects to aws that returns 403 for head requests). \r\n\r\n```\r\nr.head(\"https:\/\/github.com\/jitkapat\/thailitcorpus\/releases\/download\/v.2.0\/tlc_v.2.0.tar.gz\", allow_redirects=True) \r\n# <Response [403]>\r\n```","comment_length":14,"text":"Not support links with 302 redirect \nI have an issue adding this download link https:\/\/github.com\/jitkapat\/thailitcorpus\/releases\/download\/v.2.0\/tlc_v.2.0.tar.gz\r\n\r\nit might be because it is not a direct link (it returns 302 and redirects to aws that returns 403 for head requests). \r\n\r\n```\r\nr.head(\"https:\/\/github.com\/jitkapat\/thailitcorpus\/releases\/download\/v.2.0\/tlc_v.2.0.tar.gz\", allow_redirects=True) \r\n# <Response [403]>\r\n```\nHi !\r\nThis kind of links is now supported by the library since #1316","embeddings":[0.0149221728,-0.3587625921,-0.0337656587,-0.1861969233,0.1563389003,-0.1056114137,-0.1225814372,0.300697118,0.1491116434,-0.017120041,-0.0895056427,0.1395662576,0.020941576,-0.0117675615,0.174247995,-0.1715269983,-0.0050398447,-0.0440110415,-0.0960696414,0.1772267222,-0.0441379137,0.5200989246,0.0112193832,-0.0332711078,0.0578048378,0.1537277997,0.1305814683,0.2227170169,0.0742215216,-0.227287665,0.2602292597,0.0402767546,0.1580749303,0.448610276,-0.0001188066,0.1631999463,0.4703695476,0.0955999419,-0.2145934254,-0.755459547,-0.2289547771,-0.1805663556,-0.0137461917,0.0309577119,-0.0779622421,0.6263099909,-0.0671242625,0.3083412051,-0.0390283428,0.16951105,0.1601454467,0.0760916919,0.2434636801,-0.0887859538,0.9593945742,0.0823956653,-0.1057204828,0.3439059258,-0.0240964573,0.1540397853,0.3436636627,0.0820771009,0.1154093891,0.0235496201,-0.1165730879,-0.1275460273,0.2508887053,-0.2757377625,-0.1942501664,0.183090508,0.5101156235,0.0960139632,-0.4602693319,0.2184269428,-0.1298384964,-0.2513917387,0.3529098034,0.2700514197,0.0908351168,-0.0276834834,0.1483119726,-0.5257793665,-0.1332829744,0.2161276191,0.2922543287,0.2214302868,0.1143787876,0.0359403752,0.348980993,0.213957727,0.189195618,-0.0621066801,-0.3270667493,-0.0554924347,-0.0859632865,-0.0087824361,-0.1158879921,0.1469218135,0.2169439197,0.340203315,0.1216157153,0.0502002463,-0.2024755329,0.2060571015,-0.2922396064,-0.0242700148,-0.4050612152,0.0084896181,0.5359179378,0.2311802506,0.2099915594,-0.0703547969,-0.0266534053,-0.3257949352,-0.713087976,0.0058567962,-0.0479910485,0.0884877369,-0.0064580105,-0.0252977498,-0.3300875723,-0.1098442078,-0.0889113471,0.2359973937,-0.0834131837,-0.0110359294,0.4827166498,0.0255848058,0.0734854639,0.259052664,0.0257302672,0.1379295439,-0.1825551838,-0.262335211,-0.0233850442,0.0338828079,0.129906714,-0.0353261158,0.3255897462,-0.2598709166,0.1309609562,0.0470988192,0.0802972615,0.5251379609,0.0359513946,0.1989576817,-0.1170271486,-0.2067434341,0.1343780905,0.3438480496,-0.6418790817,-0.1358141899,-0.2085443288,0.0551042631,-0.1087356731,-0.182103321,-0.0164167676,-0.1142074168,-0.2552559376,-0.4963825643,-0.0933271423,0.0919502825,-0.0018974659,-0.1339541227,0.0551231131,0.2953075767,-0.2741287351,-0.0126254391,-0.0430625863,0.2482357621,0.0709149241,0.3294225335,-0.1333895028,-0.1289914399,-0.2385500222,-0.2221092582,0.744864285,-0.4539116621,-0.3319433033,0.4254137278,-0.3469317555,-0.1983221322,-0.0581769645,0.2673660517,-0.4240484834,-0.1979289651,-0.0271969233,0.1523573399,0.3585472703,0.0105252843,-0.305593878,-0.1498569846,-0.1069116443,-0.0731272474,-0.0225795284,0.1339569837,0.0428314209,-0.3318567276,0.3232218921,0.1299894452,0.2090632766,0.1217187494,0.1543910354,0.0397308432,0.1365791112,0.1374153197,0.2182255536,-0.097567305,-0.226621449,0.3671724498,0.0055043614,-0.0701578408,-0.2698697746,-0.1951114535,0.0752856433,0.1029955372,0.0105436537,-0.0009184317,-0.0194439944,0.340080291,-0.1013044119,-0.105952248,0.1651076227,0.349620074,0.0560421199,0.377499938,-0.3596326709,0.0817526206,0.1439939141,-0.1088129357,0.2221928388,-0.1723660976,-0.0606623329,0.2650099397,-0.2779941559,0.174278602,0.5346614718,0.0009858293,0.2412486374,0.0720633715,-0.335059613,0.4000975192,0.2222886682,0.102997005,-0.1044804752,0.0343137793,0.2051616311,0.0673960149,-0.0221158285,0.3123553991,0.2319592834,0.0526497774,-0.1994672716,0.2649547756,0.2194305211,0.1030748188,-0.8348568082,-0.6178781986,-0.1216123328,-0.2202077061,0.0820854828,-0.2381602824,0.2093659341,0.2246349454,0.0471040495,0.0851994827,0.0824154988,0.8542968035,0.0241581555,0.1897915751,0.0575653426,0.0643747151,-0.0093688061,-0.2893692553,0.2861531079,0.036004886,-0.2276037633,0.1245198548,0.0807054788,-0.0240866356,-0.1636630595,-0.4051418602,0.1709961891,0.0935482159,-0.1592458189,-0.0820638686,-0.1697734296,-0.208250463,0.1497473866,-0.159143582,-0.0104675107,-0.0833475366,0.1364867538,0.3100847006,-0.1451772153,0.1238734722,0.1029031426,0.0615828037,0.1361964792,-0.4276224673,-0.1786812246,-0.068055205,-0.0878753737,0.0996065438,0.0980220214,-0.571367383,0.0885083079,-0.1221439242,0.2164582461,-0.1174343899,-0.3290936947,-0.0900188088,-0.0047625578,-0.1467423886,0.1649374068,0.0501471534,-0.1439404637,0.1014038846,0.2008356601,0.1796271056,0.1677933633,-0.1586631238,0.1826068908,-0.1610291451,-0.2764309049,-0.2724060118,0.1724614501,-0.3352833986,0.0102301799,0.2769356966,0.1733379066,-0.1706607193,0.2451930642,-0.1196067482,0.261846602,-0.1620912105,-0.1217430681,-0.4162988365,0.5209669471,-0.3937479258,-0.4407958984,0.3352371752,-0.0794197917,0.2530189753,0.0821739957,-0.4106555581,-0.1272304356,0.0903355256,-0.3771960735,0.0127960509,0.3411304355,0.2425238192,0.2121710777,-0.1823771447,0.0958647877,-0.5838413239,-0.0122737186,0.2450440377,0.2829438448,0.1081751063,0.3572807312,0.2435961068,0.370459795,-0.0174127948,0.1945943683,0.4757985473,0.0737787634,0.1145528331,0.1007956713,0.18817316,0.0485620648,0.0973256156,0.0496953428,0.0259729996,0.3236912787,-0.3145161867,-0.4991569817,-0.3594521284,-0.1816527694,-0.1810364574,-0.1228957176,-0.3349279463,-0.0748714581,0.0818249583,-0.2084265351,0.1050831452,0.0098242657,0.3280150294,0.3642404079,0.4486465156,-0.0281834714,-0.207389757,-0.2655781507,0.0099908048,0.2359292358,0.2387791276,-0.0420354865,-0.1699389964,-0.00669399,-0.0102161914,-0.1695068032,0.4211156964,-0.5294987559,0.1060192361,-0.062477842,0.0048560197,-0.3061772883,-0.0669636503,0.1640030295,0.4481180906,-0.0491533652,-0.2841190398,-0.3136630952,0.1215207055,-0.0303293001,-0.3715469241,0.0847368166,-0.0947506279,-0.0854684189,-0.2913258672,-0.6386103034,0.1650147885,0.1099085137,-0.0690509677,-0.0827410892,-0.5481491685,0.3226939142,-0.1312431097,0.1708611846,-0.348328203,0.0780049041,0.045921769,-0.3176087141,-0.0755329281,-0.0723449811,-0.1417149007,0.2557140589,0.1276099086,0.2913628519,0.0143877976,0.1539016366,0.068834424,0.1527580917,0.1594799012,0.1415271908,-0.0590320639,-0.1040921956,0.1912824512,-0.1572487354,0.3774000406,-0.139639318,-0.3807179928,-0.1723499447,0.2174610645,-0.3572964668,-0.0218443125,0.3449890018,0.3979879916,0.1133232862,0.0268894434,-0.2089859992,1.1213121414,0.0389046483,0.0998335704,0.1405498236,-0.7236024141,0.5559365749,-0.3618429303,0.3245793581,-0.0332490727,0.5445761681,-0.205619067,-0.1868577451,-0.2195036709,-0.1054177657,-0.2039785534,0.2071602643,-0.1127347425,-0.0463668853,0.1511155218,0.2018111497,0.0058635813,0.0653600767,0.064453207,0.1634976417,0.0257603936,0.3128051162,-0.0613468513,-0.3913813829,-0.2023896277,-0.0901414603,0.0217841454,0.1669911295,0.0135650467,-0.2704378366,-0.1656670719,-0.6074132919,0.3759601116,0.3296615183,-0.0721083209,0.2466846257,-0.3088755906,0.2908816636,-0.4464882016,-0.0004018444,0.0659294426,0.2586262226,-0.0141986823,-0.2835482657,-0.2831528783,0.0959584042,0.0991025046,-0.279905051,0.4116242826,-0.0384137183,0.081280455,0.0124252979,-0.1445641518,0.1384762079,-0.555333972,-0.1159134954,0.1082022488,0.0861666724,-0.2882773876,-0.1367609352,0.4051542878,0.0787446052,0.0086354017,0.5716797113,-0.2308532745,0.2378365844,-0.0643165261,0.0357955545,-0.2343488634,-0.013190168,-0.363938272,-0.3612615466,-0.4673965871,0.301453054,-0.1370707154,-0.0912452936,-0.3615956604,-0.024396345,-0.1191272214,0.4778690636,0.0230840929,-0.3861107826,-0.0061232024,-0.084404029,-0.3943320513,0.1714624763,0.0264676828,-0.1110173389,0.466378212,-0.2861480415,-0.2098607272,0.4218209088,-0.0295082554,0.0041127889,-0.0731503889,-0.0547841676,0.5686075091,0.0090690628,0.0300838146,0.0746629238,-0.3034638166,-0.0264375098,-0.208979696,0.1273927838,0.1434576809,-0.0502722897,0.0013766527,-0.2847809196,0.0794512704,0.202683717,-0.0499775857,-0.2086305618,-0.1711372584,-0.1060644016,0.0521963015,-0.375084132,-0.0812744722,0.1738920808,-0.0318503417,0.0191154797,0.1120585054,-0.2222242057,-0.0489278063,-0.229378283,-0.4004249871,0.0085454732,0.2917487919,0.1373473108,0.6697180867,-0.3907047212,-0.2150971293,-0.048525814,0.2751260698,-0.1452295333,-0.079298526,0.0912574306,0.0621839352,0.1095026657,0.0005044762,-0.1547709405,-0.0585541427,0.0864699781,0.0949420854,-0.1694361567,0.0790579394,-0.3279075325,0.1553753614,0.3269732893,0.3604196906,-0.456230551,-0.1540232599,0.0431863926,-0.145430848,0.2649382949,-0.1549035311,-0.252844274,0.0461606756,-0.1187387258,0.0487509593,-0.2995419502,0.7237972617,-0.0661339462,0.0385452025,0.1036020815,-0.2656478584,0.0203950871,0.4072826505,-0.1042162925,0.3816079795,0.1787421256,0.3095730841,0.2711795568,-0.0965465009,0.010448236,0.1949106604,-0.0486670136,-0.2441396117,-0.0762302727,-0.0358771272,-0.0271069761,-0.0532716922,0.1328485012,0.0352612995,0.1971565783,0.3468217552,0.0170280095,-0.2164551169,-0.0305095483,0.179225117,-0.1098434329,0.0420880169,0.3216710985,-0.1197561473,0.1076515168,-0.2086541504,0.4965259731,-0.0200679526,-0.0977057666,-0.0812783986,0.005520287,-0.1059229672,0.1791393161,0.4138316512,0.4679416418,0.3052105606,0.0029082571,0.0911721587,-0.1233384386,0.4962015152,-0.2964965701,0.0499218106,-0.0627954155,0.5249295235,-0.1723538041,-0.211750567,-0.3970637918,-0.1881683469,-0.1614333242,0.0077771507,0.2775969505,-0.3011260033,-0.1226676926,-0.1666151583,0.043609187,0.0400107056,0.2352075428,0.1494226158,0.2075204849,-0.127589196,-0.0017603117,0.0965947136,0.1268963218,0.2467985749,0.2358632088,-0.2755610943,0.0302004963,-0.0432720892,0.0794033408,-0.0416977555,-0.6225004196,-0.0103024114,0.2645578682,0.1600485742,0.1626002342,-0.2113098353,0.137758404,0.1415695101,0.0450209416,-0.0807380304,-0.0461603552,0.0247547831,-0.3007130027,-0.0201586466,-0.2140280902,-0.0824822262,0.6554682255,0.2405855954,0.2532428801,-0.1190944612,-0.0530661643,-0.1883915067,-0.0333491266,-0.0132833328,-0.316350311,-0.0327336974,0.3620268703,-0.0512752682,0.2281544507,-0.1256246418,0.0914715528,-0.2493193001,0.5819194317,0.1535123885,-0.2592717707,-0.1864845306,0.0443195589,0.0824954808,-0.0300574526,0.066667892,-0.0591105521,-0.3894324899,-0.3572238982,0.5200697184,-0.0012872342,0.0275644492,0.4367583394,0.2222079486,0.1137195751,0.0156517848,-0.071636118,0.0754678175,0.0667475984,0.305107981,-0.3333064318,0.4045859575,-0.0365518853,0.1410461366,0.0407293662,0.0167458467,0.0379657224,-0.1344156265,-0.2038805634,-0.2467507422]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1064","title":"Not support links with 302 redirect ","comments":"> Hi !\r\n> This kind of links is now supported by the library since #1316\r\n\r\nI updated links in TLC datasets to be the github links in this pull request \r\n https:\/\/github.com\/huggingface\/datasets\/pull\/1737\r\n\r\nEverything works now. Thank you.","body":"I have an issue adding this download link https:\/\/github.com\/jitkapat\/thailitcorpus\/releases\/download\/v.2.0\/tlc_v.2.0.tar.gz\r\n\r\nit might be because it is not a direct link (it returns 302 and redirects to aws that returns 403 for head requests). \r\n\r\n```\r\nr.head(\"https:\/\/github.com\/jitkapat\/thailitcorpus\/releases\/download\/v.2.0\/tlc_v.2.0.tar.gz\", allow_redirects=True) \r\n# <Response [403]>\r\n```","comment_length":37,"text":"Not support links with 302 redirect \nI have an issue adding this download link https:\/\/github.com\/jitkapat\/thailitcorpus\/releases\/download\/v.2.0\/tlc_v.2.0.tar.gz\r\n\r\nit might be because it is not a direct link (it returns 302 and redirects to aws that returns 403 for head requests). \r\n\r\n```\r\nr.head(\"https:\/\/github.com\/jitkapat\/thailitcorpus\/releases\/download\/v.2.0\/tlc_v.2.0.tar.gz\", allow_redirects=True) \r\n# <Response [403]>\r\n```\n> Hi !\r\n> This kind of links is now supported by the library since #1316\r\n\r\nI updated links in TLC datasets to be the github links in this pull request \r\n https:\/\/github.com\/huggingface\/datasets\/pull\/1737\r\n\r\nEverything works now. Thank you.","embeddings":[-0.0889685974,-0.2517869771,-0.0088746836,-0.1504719406,0.1692062467,-0.0146814175,-0.0734717399,0.35192433,0.0370435566,-0.0853566155,-0.1462335736,0.1753974855,-0.0085471682,0.0927394256,0.1950139403,-0.2360422313,-0.0524225272,-0.0261497889,-0.0826820955,0.2028121501,-0.027272163,0.5130646825,-0.0248572268,-0.0624030828,-0.068671234,0.0774284601,0.1210468486,0.2031793594,0.0683747903,-0.2401624471,0.303055048,0.1178529337,0.1594067365,0.5492524505,-0.0001215735,0.1579644233,0.3548876941,0.072055988,-0.2901379168,-0.5894262195,-0.2536681294,-0.2439851612,0.0339293927,0.032435026,-0.1306387931,0.5072250962,-0.0778939947,0.1441019475,-0.0753327459,0.1765078306,0.1485421062,0.2375597805,0.2008063793,-0.0810438469,0.9201811552,0.0543805584,-0.1180588827,0.3259595335,-0.0304333828,0.1545787603,0.3328635097,0.177852422,0.095285818,0.0808029175,-0.111031048,-0.1084039807,-0.0450997874,-0.2801686823,-0.15075472,0.2552185953,0.6863910556,0.0206972808,-0.5041068792,0.1242092252,-0.1229810044,-0.2610161602,0.3652248383,0.2766008675,0.1618135124,0.0424441136,0.0220192727,-0.4948576689,-0.1582905501,0.1913100332,0.1813697368,0.1272976249,0.079785414,0.0317876004,0.2093795091,0.1482400596,0.2320011407,-0.1184864119,-0.3121515214,-0.0051235668,-0.2185737342,-0.079031229,-0.1069963127,0.1383845359,0.2005695701,0.3603540957,0.0971251801,0.0609982759,-0.2779536545,0.1389889568,-0.1637238413,0.0455287695,-0.4307024777,0.1138571426,0.599773109,0.1345745623,0.1671667397,-0.1014247984,0.025814455,-0.3618370593,-0.640655756,-0.0050231772,0.021018425,0.0484001972,-0.0703640431,-0.0238336902,-0.3518248796,-0.1699009836,-0.1177654788,0.2023970634,-0.0567761026,0.0609716028,0.5130020976,0.0460665971,0.0043183723,0.1508783102,-0.0211402364,0.0231467597,-0.130186066,-0.1388564408,0.076084815,-0.0486892276,0.1044013426,-0.0494075343,0.2970653474,-0.1429970115,0.1065954491,0.0482982174,0.0091796611,0.5238859057,0.0449464433,0.2545099258,-0.1337096244,-0.1347231716,0.1427950412,0.3535699844,-0.6124945283,-0.2109988928,-0.1807133853,0.0326900445,-0.1545538604,-0.2067750841,-0.1573908031,-0.0028978975,-0.1474408358,-0.5722486377,-0.1322476417,0.06036878,-0.0241044015,-0.1116486788,0.0952816308,0.3426842093,-0.4351748228,0.0094628846,-0.0647440851,0.2282375544,-0.012002646,0.3981320262,-0.1556462944,-0.0729114115,-0.3013798892,-0.1341858655,0.7303339839,-0.4807989001,-0.3789023459,0.4282314181,-0.2898419797,-0.1877360195,-0.1061226353,0.2756073773,-0.3474136591,-0.1764629483,-0.1268268824,0.1685846895,0.2764305174,0.0358418599,-0.2897286713,-0.1738007516,-0.0571221299,-0.0227623209,-0.0404700972,0.1606975347,0.1272218972,-0.3260907829,0.3235673904,0.0820758864,0.2460441887,0.1628433168,0.2186272591,0.083060056,0.0801443234,0.0311642233,0.0336822607,-0.0903039277,0.001823253,0.3663538694,-0.0354037546,-0.093159847,-0.2621215284,-0.2167806327,0.0052463999,0.1010594741,-0.0144658852,0.0691929385,-0.0824376494,0.3779104948,-0.1585894376,0.0074779531,0.1893827915,0.3136532009,-0.0861897618,0.5394979119,-0.2678464353,0.1029390469,0.1966937482,-0.0428085066,0.1874131113,-0.1979300976,-0.1212188601,0.3034728169,-0.3177319169,0.1804010719,0.6118975282,0.0786386654,0.2613998353,-0.0922379345,-0.313267231,0.3064296544,0.2189106494,0.110688135,-0.1312217265,0.1056995988,0.1801846623,0.0462584086,0.0018849198,0.2894602716,0.2616717219,0.0509676673,-0.227001518,0.2264629304,0.1595851183,0.0993027538,-0.7005435228,-0.574280262,-0.2411242723,-0.0756931528,0.2490627617,-0.2395289093,0.1468952596,0.183181271,-0.0438629016,0.0771969482,0.1577172577,0.8859192729,0.1208991036,0.2381839305,0.0884258673,0.0439855233,-0.0307360832,-0.2622562647,0.3470580876,0.0624578409,-0.2946636379,0.1126286536,0.1000957862,0.0410520211,-0.2012776285,-0.4085431695,0.2061423808,0.14872168,-0.2893052399,-0.0932014212,-0.2989703119,-0.2179988623,0.1421817988,-0.2177566439,-0.0476185866,-0.2104533017,0.082964927,0.3161398768,-0.1778657734,0.1747334749,0.1694303453,0.0775181353,0.2099795938,-0.4418687224,-0.2040369511,-0.1694379896,-0.1004686952,0.081806533,0.1153391376,-0.5353226066,0.1564164609,-0.138467893,0.2705561221,-0.2036029696,-0.3495527506,-0.0268721711,-0.0887323618,-0.1053681001,0.182529673,0.1798851341,-0.1700104028,0.0821453407,0.235492155,0.194015041,0.1734012365,-0.1229033992,0.1601561159,-0.1496061981,-0.2317112237,-0.3762661517,0.0774012581,-0.347966373,-0.0618770272,0.1576521695,0.1705688983,-0.046214655,0.2099301964,-0.0879655629,0.2394791543,-0.2012473792,-0.1868759692,-0.4467816949,0.5848569274,-0.4474649727,-0.487768203,0.4066593349,0.0196296908,0.2492921203,0.082568787,-0.4886935651,-0.1169576198,-0.0007140478,-0.2425427735,0.0936522782,0.4416886568,0.1985341609,0.2711295784,-0.1463364661,0.0718564466,-0.6122390628,-0.043817915,0.2205577493,0.2329094857,0.1641921699,0.3945307732,0.2477852404,0.4924861193,0.0896543935,0.1346900165,0.5550949574,0.0060134712,0.2173865288,0.0753515065,0.1182252541,0.0517434254,0.0758493543,0.1876437515,0.1232973039,0.3488506377,-0.2544369102,-0.5247748494,-0.2646774948,-0.1947950125,-0.2648602426,-0.1564062387,-0.3417546451,0.0060830275,0.1678368002,-0.2579476237,-0.0370162465,-0.0611165874,0.2510456145,0.3608328402,0.4510367215,0.0007097292,-0.2106647044,-0.2279955745,-0.0390673243,0.2530306578,0.1908376515,0.0448911712,-0.1536705792,0.0495476834,0.0041145231,-0.0442487523,0.4536240697,-0.5851208568,0.1177207455,-0.0885075703,-0.0010408077,-0.4180229008,-0.0071931323,0.1050683558,0.4362521768,-0.1284873486,-0.1862546355,-0.3607248366,0.0475909933,-0.0311879758,-0.2774047256,0.0786441043,-0.0558335185,-0.0152126458,-0.3150623143,-0.5795224905,0.0964534134,0.1397798955,0.0141182635,-0.1541859508,-0.5138261914,0.2536278069,-0.1545204371,0.1522517353,-0.3228801191,0.1153495461,0.1447962523,-0.2176791877,-0.0293697231,-0.0311050937,-0.1075266376,0.3889840841,0.0025143253,0.1795887649,0.0742037296,0.1656280756,0.0766892359,0.2557786703,0.1276845485,0.1066344455,-0.0201178137,-0.0473728813,0.1491013765,-0.158490628,0.3536530435,-0.1017190441,-0.3683064282,-0.271253407,0.2432252169,-0.2895277143,-0.0264065601,0.3046959341,0.464310348,0.0440710783,-0.0107846139,-0.2777213454,1.1651057005,0.0195688829,0.1489719152,0.2043876648,-0.6890200377,0.622633338,-0.3744078875,0.2987077832,-0.0720069557,0.36834234,-0.232326135,-0.1634857059,-0.2121368796,-0.0576787032,-0.2465138733,0.2987430096,-0.0814924091,-0.0231853239,0.1697258949,0.191819191,0.0351373255,0.0958983377,-0.0914280713,0.1366746426,-0.0447200574,0.2567814887,-0.0499420501,-0.3706736565,-0.1537326127,-0.0821034238,0.0061929459,0.0796001256,-0.107432276,-0.2448487133,-0.2179087698,-0.6740682125,0.4979631007,0.4169785082,-0.0167945828,0.2131245583,-0.380930692,0.2791494131,-0.3834372163,0.1113098487,0.0327414572,0.2366836071,0.0840962753,-0.2209800482,-0.2883395851,0.0783735812,0.0814278275,-0.2733275592,0.4177575409,-0.0521048717,0.0164521933,-0.1052115336,-0.1408561915,0.1241043508,-0.4260392189,-0.1168865487,0.0737938359,0.1023824364,-0.2741328478,-0.0857418701,0.3911029994,0.0816473141,0.0292867161,0.6203910708,-0.2215279043,0.3158129156,0.0132890539,0.1694729328,-0.2366682291,0.0089695519,-0.2597521544,-0.1960687488,-0.6053310037,0.2689874768,-0.1052271053,0.0188861229,-0.3100356758,0.065442428,-0.1169670075,0.3583503067,0.0188779701,-0.3829031885,0.0224501584,-0.0471163057,-0.3808612823,0.2462105006,0.0276369825,-0.0416691303,0.4045987725,-0.2615495622,-0.2085302472,0.4591498971,-0.0382072143,0.0686478689,0.0407214463,-0.0313891433,0.6047629714,-0.0192077737,-0.0203948598,0.0063335467,-0.3603475988,-0.0000544643,-0.1368187964,0.140026629,0.2005037516,-0.1495243013,-0.0763384774,-0.356962949,-0.0031254471,0.1024742424,0.0521693863,-0.1762252301,-0.1822401583,-0.065651454,-0.0509690754,-0.5027399063,-0.1036820263,0.201268509,0.0270433817,0.0134950392,0.1688159406,-0.182006225,-0.0273711104,-0.222149536,-0.3524934947,0.0822127983,0.339910239,0.1501911432,0.66255337,-0.490031004,-0.1894516796,0.0816828534,0.3682281375,-0.1697556227,-0.0991652161,0.0680611208,0.1000921354,0.0853437185,-0.0724864304,-0.1392832398,0.0850876793,0.1007645503,0.1168415323,-0.1893628538,0.1114645004,-0.2744365931,0.1228788495,0.2569235265,0.4701001942,-0.3736874759,-0.0202095676,0.1007959992,-0.143697381,0.2254040241,-0.1036121547,-0.2153600901,-0.0235446375,-0.0247107577,0.0115562547,-0.2242949903,0.6830972433,-0.0913544521,0.0311853904,-0.066826351,-0.1223293319,0.0060824389,0.3291818202,-0.1530186683,0.3240323961,0.2617713511,0.1593336016,0.1860973388,-0.1107607186,0.0848562792,0.2043670714,0.0001790236,-0.1979276091,-0.0814363435,-0.0171643216,-0.0715727657,-0.1194844469,0.086936526,0.0831442028,0.2113757282,0.368537128,-0.1532917172,-0.1742378473,0.0122790905,0.1659178883,-0.0398956574,0.175587222,0.2483384758,-0.1321436763,0.1575344354,-0.0637352318,0.4261856079,-0.1377837509,-0.0574373566,-0.0198080633,0.0427991785,-0.1093342975,0.2672948837,0.3388477862,0.3575831354,0.2028677762,0.1016949341,0.0715591833,-0.087121971,0.3904490471,-0.2852380872,0.0988241509,-0.1897280663,0.4406181574,-0.2631388903,-0.26309672,-0.3342834413,-0.2541718781,-0.1799211204,0.0261031426,0.3590555191,-0.3362891376,-0.1018824577,-0.2251100093,0.0400168225,0.0043710833,0.2992422581,0.254983902,0.186730817,-0.1178346127,-0.0454623625,-0.0539559722,0.1426371932,0.3550260067,0.1628576815,-0.225479126,0.0045336299,0.0158653613,0.0830284953,0.0428047292,-0.6330807209,0.1068468168,0.1801851094,0.0551547445,0.0485301018,-0.2542408407,0.1518829614,0.0600721091,0.0303701703,-0.0813526735,-0.0359801725,0.0045007537,-0.2740031779,0.0421616621,-0.202282697,-0.1776155382,0.6491750479,0.2948361635,0.3487353325,-0.0349286385,-0.0001385173,-0.1983118653,-0.0706490502,-0.0605613813,-0.2901889682,0.0004361907,0.4371778071,-0.1068338677,0.1789089441,-0.1465785801,0.0755383223,-0.1912952811,0.7068845034,0.1473888457,-0.1841676086,-0.2230777293,-0.0142893987,0.1335858405,-0.0381839797,0.0524566695,0.0219445843,-0.4203704596,-0.4358584583,0.5019134283,-0.1479953527,0.0242779925,0.4396547377,0.2329077274,0.1449548006,-0.058179304,-0.1517632604,0.0923013017,0.1003262624,0.2673763037,-0.3939915001,0.4566243589,-0.021420965,0.196099177,0.0748507529,0.1301731318,0.0484582298,-0.1997936219,-0.2278058529,-0.196218878]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1046","title":"Dataset.map() turns tensors into lists?","comments":"A solution is to have the tokenizer return a list instead of a tensor, and then use `dataset_tok.set_format(type = 'torch')` to convert that list into a tensor. Still not sure if bug.","body":"I apply `Dataset.map()` to a function that returns a dict of torch tensors (like a tokenizer from the repo transformers). However, in the mapped dataset, these tensors have turned to lists!\r\n\r\n```import datasets\r\nimport torch \r\nfrom datasets import load_dataset \r\nprint(\"version datasets\", datasets.__version__)\r\n\r\ndataset = load_dataset(\"snli\", split='train[0:50]') \r\n\r\ndef tokenizer_fn(example):\r\n # actually uses a tokenizer which does something like:\r\n return {'input_ids': torch.tensor([[0, 1, 2]])}\r\n\r\nprint(\"First item in dataset:\\n\", dataset[0])\r\ntokenized = tokenizer_fn(dataset[0])\r\nprint(\"Tokenized hyp:\\n\", tokenized)\r\ndataset_tok = dataset.map(tokenizer_fn, batched=False,\r\n remove_columns=['label', 'premise', 'hypothesis'])\r\nprint(\"Tokenized using map:\\n\", dataset_tok[0])\r\nprint(type(tokenized['input_ids']), type(dataset_tok[0]['input_ids']))\r\ndataset_tok = dataset.map(tokenizer_fn, batched=False,\r\n remove_columns=['label', 'premise', 'hypothesis'])\r\nprint(\"Tokenized using map:\\n\", dataset_tok[0])\r\nprint(type(tokenized['input_ids']), type(dataset_tok[0]['input_ids']))\r\n```\r\n\r\nThe output is:\r\n\r\n```\r\nversion datasets 1.1.3\r\nReusing dataset snli (\/home\/tom\/.cache\/huggingface\/datasets\/snli\/plain_text\/1.0.0\/bb1102591c6230bd78813e229d5dd4c7fbf4fc478cec28f298761eb69e5b537c)\r\nFirst item in dataset:\r\n {'premise': 'A person on a horse jumps over a broken down airplane.', 'hypothesis': 'A person is training his horse for a competition.', 'label': 1}\r\nTokenized hyp:\r\n {'input_ids': tensor([[0, 1, 2]])}\r\nLoading cached processed dataset at \/home\/tom\/.cache\/huggingface\/datasets\/snli\/plain_text\/1.0.0\/bb1102591c6230bd78813e229d5dd4c7fbf4fc478cec28f298761eb69e5b537c\/cache-fe38f449fe9ac46f.arrow\r\nTokenized using map:\r\n {'input_ids': [[0, 1, 2]]}\r\n<class 'torch.Tensor'> <class 'list'>\r\n```\r\n\r\nOr am I doing something wrong?\r\n","comment_length":32,"text":"Dataset.map() turns tensors into lists?\nI apply `Dataset.map()` to a function that returns a dict of torch tensors (like a tokenizer from the repo transformers). However, in the mapped dataset, these tensors have turned to lists!\r\n\r\n```import datasets\r\nimport torch \r\nfrom datasets import load_dataset \r\nprint(\"version datasets\", datasets.__version__)\r\n\r\ndataset = load_dataset(\"snli\", split='train[0:50]') \r\n\r\ndef tokenizer_fn(example):\r\n # actually uses a tokenizer which does something like:\r\n return {'input_ids': torch.tensor([[0, 1, 2]])}\r\n\r\nprint(\"First item in dataset:\\n\", dataset[0])\r\ntokenized = tokenizer_fn(dataset[0])\r\nprint(\"Tokenized hyp:\\n\", tokenized)\r\ndataset_tok = dataset.map(tokenizer_fn, batched=False,\r\n remove_columns=['label', 'premise', 'hypothesis'])\r\nprint(\"Tokenized using map:\\n\", dataset_tok[0])\r\nprint(type(tokenized['input_ids']), type(dataset_tok[0]['input_ids']))\r\ndataset_tok = dataset.map(tokenizer_fn, batched=False,\r\n remove_columns=['label', 'premise', 'hypothesis'])\r\nprint(\"Tokenized using map:\\n\", dataset_tok[0])\r\nprint(type(tokenized['input_ids']), type(dataset_tok[0]['input_ids']))\r\n```\r\n\r\nThe output is:\r\n\r\n```\r\nversion datasets 1.1.3\r\nReusing dataset snli (\/home\/tom\/.cache\/huggingface\/datasets\/snli\/plain_text\/1.0.0\/bb1102591c6230bd78813e229d5dd4c7fbf4fc478cec28f298761eb69e5b537c)\r\nFirst item in dataset:\r\n {'premise': 'A person on a horse jumps over a broken down airplane.', 'hypothesis': 'A person is training his horse for a competition.', 'label': 1}\r\nTokenized hyp:\r\n {'input_ids': tensor([[0, 1, 2]])}\r\nLoading cached processed dataset at \/home\/tom\/.cache\/huggingface\/datasets\/snli\/plain_text\/1.0.0\/bb1102591c6230bd78813e229d5dd4c7fbf4fc478cec28f298761eb69e5b537c\/cache-fe38f449fe9ac46f.arrow\r\nTokenized using map:\r\n {'input_ids': [[0, 1, 2]]}\r\n<class 'torch.Tensor'> <class 'list'>\r\n```\r\n\r\nOr am I doing something wrong?\r\n\nA solution is to have the tokenizer return a list instead of a tensor, and then use `dataset_tok.set_format(type = 'torch')` to convert that list into a tensor. Still not sure if bug.","embeddings":[-0.0905342773,-0.1736622602,-0.1339069456,0.2130297422,0.2162035555,0.2193657011,0.4914948344,0.3712697923,0.257444948,-0.0263979062,-0.1461347193,0.680251956,-0.10030251,-0.4164583683,0.1840583086,0.0253589209,0.294393152,0.0346806832,-0.0507990569,-0.2774459422,-0.1831274182,-0.0675765797,-0.1957766712,0.0791181624,-0.1412031651,-0.1547512561,0.1446099281,-0.3113888204,-0.0627612844,-0.2368612587,0.1946785748,0.0335779414,0.1022556797,0.446719557,-0.0001160152,0.1183271185,-0.037234392,0.2079562694,-0.0176475514,-0.1034466103,0.0076182238,-0.1327908486,0.1470107734,-0.1844788641,-0.0054996279,-0.1479554027,-0.0391307659,-0.6350414753,0.179096505,0.3810082972,0.1443395168,0.2552904189,-0.1439363956,0.2130863816,0.3324943781,0.2167777121,-0.2084973902,-0.2550541461,0.1627887785,-0.0757924542,0.0668914467,0.1129236296,-0.4984409213,-0.0192023627,0.2370801419,0.2221133709,-0.1869863868,-0.5936381817,0.0203761477,0.1143103391,0.346229285,-0.3974611461,-0.1002945304,-0.1565368474,-0.2213947773,0.0423909165,-0.0511560366,0.0540001467,0.0545048267,-0.0863507465,-0.2528419197,-0.0814966038,0.0649670362,0.2210161388,-0.5532370806,0.2875820696,-0.0781859979,0.2344746888,0.1804774553,-0.0505854003,0.1507671028,-0.3792155981,0.0853365734,0.1579334438,-0.3272643685,-0.1862962246,0.1454360634,-0.4385431707,0.0561810508,-0.1962929219,0.1501737237,0.2216709554,-0.397954762,0.1758711338,0.1768113673,0.1449903399,-0.0155394245,0.680030942,-0.0505105294,-0.4166106284,-0.0770037621,0.183756128,0.0966400057,-0.0125526022,0.0351255648,-0.0809561312,0.1311855167,0.0802533701,-0.1056552529,0.1230914518,-0.5634173751,0.1094647199,0.002619667,0.0384143218,-0.186223343,0.1880778372,0.0646297261,-0.0047181626,0.0791275799,0.2578691542,-0.0984069183,-0.2119452506,-0.1304918379,-0.0817815512,-0.0984928384,-0.0916112587,0.0619072318,0.092382431,0.1118398234,-0.0017336064,0.2803901732,-0.0927288532,0.5385236144,0.1707926393,-0.1122141108,0.4933412373,0.2230668664,-0.0651279017,-0.2487444878,0.2211106271,-0.3083320856,0.1185796559,0.0200917516,0.1291921139,-0.0472025983,-0.0984765962,-0.2833484411,0.1163356304,0.2763032913,-0.0582993142,0.2532229424,-0.3349198103,-0.132949248,-0.1963020861,0.1043847948,-0.0451547392,-0.118641369,-0.0467285588,0.044093363,0.0888606682,0.0959444866,0.3504171371,-0.2296563685,0.3976287246,-0.1970900148,0.3188100755,0.3553226292,-0.1605848074,-0.4590785503,0.0325200707,-0.0414366685,0.2008669823,-0.4211300015,0.2581838071,0.5175717473,0.0115962904,0.1081317067,-0.0138313118,0.0394521281,0.0712760985,0.0759738982,0.0893174186,0.6063425541,-0.1942304671,0.0912254751,0.4368440509,-0.0899207368,0.1663245708,0.2264939547,-0.0353974365,-0.0143849719,0.2562342286,-0.0984062627,0.1964918971,0.0711438432,-0.1331272125,-0.3207593262,-0.0673473179,0.2024287283,0.0814435259,-0.0857608393,-0.2460781038,0.0505057834,0.1824002415,-0.1853736043,0.0724808052,0.1081486642,-0.0400793664,-0.2167155892,0.0308499429,-0.0139192659,-0.2510787547,0.1745028645,0.204618454,-0.4255838692,0.2819700539,0.1628724337,0.1375444531,-0.4352928102,0.1270138621,-0.0553200878,-0.0830142573,0.0988000557,0.0349606797,0.1280521005,0.0265865475,-0.0954571441,0.5214632154,0.2866054773,-0.2311563343,0.0751397461,0.4937567711,0.0238809288,-0.229771331,0.1961795241,0.1395006329,0.2197048515,0.3267371356,-0.4468351305,0.2423465103,0.036944773,0.2038594931,-0.3506708741,0.1286530793,-0.4682184458,0.2203543335,0.1738942266,0.1489115953,-0.1910985857,0.0657252818,0.4890259206,-0.1033508554,0.0994033739,-0.0586623512,-0.5665890574,-0.0065713013,0.2327365875,-0.2958332598,0.1299310476,-0.078820765,0.3676331341,0.0134283369,-0.0215824917,0.1092869341,0.0924237296,0.351996392,0.0202198178,-0.0760973096,0.0611620396,0.065281257,-0.3260114789,0.0977090001,-0.0819270909,0.1087761745,-0.1746973693,0.215378195,-0.2485268414,0.114612475,-0.0772436485,-0.4351750314,-0.1941461265,-0.1610156745,-0.0932937786,0.1621305048,0.0375246331,0.1928770095,0.3247022033,0.0016953748,0.3418312073,-0.225582853,-0.1251525581,-0.2449374646,-0.1197988689,0.0738311112,0.0181270093,-0.4264046252,0.083983548,-0.0697512254,-0.0664159432,-0.7368848324,-0.2431028038,0.1734049171,-0.2962734401,-0.2769447267,0.3027134836,0.2004784793,-0.3066086471,0.0971379206,-0.0186878163,0.1475964338,0.0467998534,0.0844613612,-0.1103945002,0.1160804108,-0.1308739036,-0.3476417065,-0.0094152419,-0.0166204292,0.2716757655,0.0838909224,0.1525645256,0.0774168074,0.3029705882,0.1268658638,0.0299354363,-0.1717898846,-0.1643576324,-0.2441188097,0.3611247838,-0.0520239957,-0.22654441,-0.0212237015,-0.0030592845,0.1346479952,0.5783620477,-0.3174059987,0.0344276801,-0.0420033671,0.2637937367,0.0175747406,0.1347250193,0.3775294721,0.2478036731,0.1076655388,-0.0430456065,-0.0288084429,-0.0538102761,-0.0161097497,0.3160053194,0.1498702914,0.3397210538,0.1404833645,0.6328849196,0.0862662196,-0.5870518684,-0.2141459286,-0.3498601615,0.2165563256,-0.1520948708,-0.3102630079,-0.1093504801,-0.28600806,-0.300843358,0.0750686973,-0.1314490288,-0.2657194734,0.2969166636,0.446225971,-0.0882325992,-0.1677335501,0.0572007596,-0.6227615476,0.1875963956,0.0649887696,0.2781692147,-0.3583343327,-0.1568048894,0.1168322787,-0.1129730567,0.3872697353,-0.1263889372,-0.7034638524,0.0063643605,-0.4204152524,0.1873979867,0.0056880182,0.1773734242,0.1652300209,-0.112509191,0.1744450778,0.3250636756,0.7036460638,-0.0586114973,-0.3005672395,0.1336065233,-0.0997832417,-0.4846154153,0.0846258402,-0.2593833804,0.2423414141,0.2093499899,0.3109831214,-0.0476541407,-0.2322614938,-0.3107507229,0.1333297491,-0.053243123,0.3153674006,-0.1331279129,-0.0941951349,-0.127183944,-0.1217167154,0.1963419467,-0.2370506227,-0.0338916741,-0.5592253804,-0.1888412088,-0.0711832419,0.1944610924,0.0427112952,0.2000722885,0.131830126,-0.2557500601,0.2216896415,0.1591052711,0.117755577,0.0787251666,-0.1197154671,-0.3400996625,-0.051767502,0.2129713744,0.1649908274,-0.1602071673,-0.141362384,0.1177048609,-0.2446500808,-0.1058980897,-0.3997228444,0.4764786363,0.5671220422,0.2579691112,-0.3157423139,-0.6125047803,0.2302146554,0.3774290681,-0.2791235447,0.4521859288,-0.5061218143,-0.3920195699,0.2071548253,0.4282876551,0.5739824772,-0.4119100869,0.2498632669,0.0765873864,0.2339149117,0.426009953,-0.4805026054,0.1393812001,-0.3452283144,0.1527659148,-0.2154876143,-0.2566210926,0.0636829063,0.1887102276,-0.0316687711,-0.1242906526,-0.0239557792,0.7876774669,0.0780502781,-0.1464799196,0.3955989182,-0.2671814561,0.0892166644,0.2036619037,-0.1212615892,0.2099370509,0.0092979362,0.0513104908,-0.1965031326,-0.127871111,0.0116996625,-0.0130519327,-0.2299930602,0.2618724704,0.1525951326,-0.201145187,0.1392355412,0.7396448851,0.2234844863,0.2516423762,-0.0744041055,-0.038147714,0.1769199222,0.2847000957,0.1070501581,-0.1848676354,0.0835148692,0.2304860801,-0.0115294596,0.2355246991,0.0631159917,-0.268345505,-0.4846624434,0.2600140274,0.5671141148,-0.3808259666,0.2440679818,0.1625121534,-0.0416189581,0.001362141,0.0675532818,-0.1471410394,-0.1325733364,0.4093293548,-0.1548150033,-0.1608565003,0.1107495502,0.3674654663,0.2076233774,-0.1533612311,0.2480866611,0.056940794,-0.0848241076,-0.2917854786,0.1096708104,0.2019403428,0.1349809766,0.2240548879,-0.3578024209,-0.1645009816,-0.1746764034,0.1499622911,-0.2525676191,-0.0387568548,-0.3988036513,-0.0235320553,-0.3721062243,-0.025249349,-0.0994727314,0.4052252769,0.1330139041,0.033003889,-0.2353828102,-0.2841766775,-0.2150231451,0.0614063554,0.1049758121,0.317512691,-0.0161709879,0.1339095384,-0.0781409293,-0.2015335411,0.080073081,0.1677751988,-0.0815930665,-0.2401295453,0.0324873514,0.0831723288,0.1623704135,-0.2706999779,-0.0207832586,0.0171218216,-0.1772500575,-0.3691253066,0.2644973099,0.3387384713,-0.0548426434,0.2520482838,0.1629133224,0.1922449619,0.0664829612,-0.2389813811,-0.0673396289,-0.12356361,0.0767102763,0.3552877307,0.0898754299,-0.0253588594,-0.2356754243,0.2498126924,0.4343939126,-0.1193523854,0.5966325402,-0.2622196674,0.1074622124,0.1145689413,0.2363798171,0.2327507734,-0.3498212397,-0.2978017032,0.097635977,0.1566759348,0.0503739156,-0.2055505514,0.3785065711,0.1290355772,0.0085398173,0.138445437,0.1084859595,0.1346496791,-0.0434607454,-0.2603903711,0.3331289887,-0.3271982074,0.4942014515,0.397609055,0.0265155472,0.0466422215,0.1061060429,-0.0197215285,-0.088384591,0.4616709054,0.2282730788,0.5523833632,0.30277583,-0.0340121798,0.5073820949,-0.0137486272,0.0007735395,-0.0157656074,-0.2453914583,0.2581062317,0.1902956069,0.0556507185,-0.0475449152,-0.2028522789,0.0577834547,0.1941241026,-0.3690662682,-0.0328410789,-0.1844906807,-0.009945306,0.2395032048,-0.209104225,-0.0376407951,-0.2905730605,0.176477477,-0.0967049152,-0.0216083918,-0.0340525955,-0.0009388835,-0.0679468364,-0.0062976726,0.0717409328,-0.1286428273,-0.4583580494,0.0727507025,-0.1606461555,0.1530968249,0.0665907785,0.2265326381,-0.1436648667,0.2057797313,-0.0782430917,-0.1337422132,-0.0354567468,0.1714806855,-0.2607640326,-0.2001806647,-0.0241287295,0.08529073,-0.0528155193,-0.1415355057,-0.2138057649,0.3650476933,0.0129362866,0.2667193711,-0.1424587518,-0.2983596027,-0.1448249519,-0.1189547405,-0.160856843,-0.1878195852,0.6996372342,-0.03504619,0.1697005779,-0.1595540792,0.0491679423,0.0152545413,0.2613005638,0.0646018758,0.1442367733,-0.0080209337,0.041236002,-0.1432537138,0.3525063097,0.240338102,0.1202170402,-0.0026343146,0.0231717173,0.0901697204,0.2375983447,-0.1209659055,-0.4073794484,0.0233071763,0.1407173723,0.0084185265,-0.0124782268,-0.5062525868,0.2792611718,0.0773458257,-0.1727735847,0.2210201025,0.0803734884,-0.0051450813,-0.3682461977,0.0631453693,-0.2415378541,-0.0329703949,0.4665831029,0.2855281532,0.1157337949,-0.0312674791,0.0886829048,-0.1312686205,-0.0826187134,-0.1515829265,0.1294198036,-0.3124343455,0.105223164,-0.2197997868,-0.0723182783,-0.2349605262,-0.1989329904,0.4463572204,-0.314909637,-0.4419314861,0.1768953651,-0.2677648664,0.0713506564,0.054126054,0.4623350799,0.1842846721,0.2511640489,-0.0895964429,-0.1723601669,0.2852713168,-0.4563175738,-0.3587994277,-0.045180615,-0.068057403,0.1567696035,-0.2037457973,-0.6132336259,-0.3314446807,0.4240120351,-0.3604155481,-0.1031689867,-0.1472704113,0.079053469,0.1174907014,-0.3405918181,0.1946748942,-0.0184210837,0.0094523104,0.2543967366,-0.4694918692]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1046","title":"Dataset.map() turns tensors into lists?","comments":"It is expected behavior, you should set the format to `\"torch\"` as you mentioned to get pytorch tensors back.\r\nBy default datasets returns pure python objects.","body":"I apply `Dataset.map()` to a function that returns a dict of torch tensors (like a tokenizer from the repo transformers). However, in the mapped dataset, these tensors have turned to lists!\r\n\r\n```import datasets\r\nimport torch \r\nfrom datasets import load_dataset \r\nprint(\"version datasets\", datasets.__version__)\r\n\r\ndataset = load_dataset(\"snli\", split='train[0:50]') \r\n\r\ndef tokenizer_fn(example):\r\n # actually uses a tokenizer which does something like:\r\n return {'input_ids': torch.tensor([[0, 1, 2]])}\r\n\r\nprint(\"First item in dataset:\\n\", dataset[0])\r\ntokenized = tokenizer_fn(dataset[0])\r\nprint(\"Tokenized hyp:\\n\", tokenized)\r\ndataset_tok = dataset.map(tokenizer_fn, batched=False,\r\n remove_columns=['label', 'premise', 'hypothesis'])\r\nprint(\"Tokenized using map:\\n\", dataset_tok[0])\r\nprint(type(tokenized['input_ids']), type(dataset_tok[0]['input_ids']))\r\ndataset_tok = dataset.map(tokenizer_fn, batched=False,\r\n remove_columns=['label', 'premise', 'hypothesis'])\r\nprint(\"Tokenized using map:\\n\", dataset_tok[0])\r\nprint(type(tokenized['input_ids']), type(dataset_tok[0]['input_ids']))\r\n```\r\n\r\nThe output is:\r\n\r\n```\r\nversion datasets 1.1.3\r\nReusing dataset snli (\/home\/tom\/.cache\/huggingface\/datasets\/snli\/plain_text\/1.0.0\/bb1102591c6230bd78813e229d5dd4c7fbf4fc478cec28f298761eb69e5b537c)\r\nFirst item in dataset:\r\n {'premise': 'A person on a horse jumps over a broken down airplane.', 'hypothesis': 'A person is training his horse for a competition.', 'label': 1}\r\nTokenized hyp:\r\n {'input_ids': tensor([[0, 1, 2]])}\r\nLoading cached processed dataset at \/home\/tom\/.cache\/huggingface\/datasets\/snli\/plain_text\/1.0.0\/bb1102591c6230bd78813e229d5dd4c7fbf4fc478cec28f298761eb69e5b537c\/cache-fe38f449fe9ac46f.arrow\r\nTokenized using map:\r\n {'input_ids': [[0, 1, 2]]}\r\n<class 'torch.Tensor'> <class 'list'>\r\n```\r\n\r\nOr am I doing something wrong?\r\n","comment_length":26,"text":"Dataset.map() turns tensors into lists?\nI apply `Dataset.map()` to a function that returns a dict of torch tensors (like a tokenizer from the repo transformers). However, in the mapped dataset, these tensors have turned to lists!\r\n\r\n```import datasets\r\nimport torch \r\nfrom datasets import load_dataset \r\nprint(\"version datasets\", datasets.__version__)\r\n\r\ndataset = load_dataset(\"snli\", split='train[0:50]') \r\n\r\ndef tokenizer_fn(example):\r\n # actually uses a tokenizer which does something like:\r\n return {'input_ids': torch.tensor([[0, 1, 2]])}\r\n\r\nprint(\"First item in dataset:\\n\", dataset[0])\r\ntokenized = tokenizer_fn(dataset[0])\r\nprint(\"Tokenized hyp:\\n\", tokenized)\r\ndataset_tok = dataset.map(tokenizer_fn, batched=False,\r\n remove_columns=['label', 'premise', 'hypothesis'])\r\nprint(\"Tokenized using map:\\n\", dataset_tok[0])\r\nprint(type(tokenized['input_ids']), type(dataset_tok[0]['input_ids']))\r\ndataset_tok = dataset.map(tokenizer_fn, batched=False,\r\n remove_columns=['label', 'premise', 'hypothesis'])\r\nprint(\"Tokenized using map:\\n\", dataset_tok[0])\r\nprint(type(tokenized['input_ids']), type(dataset_tok[0]['input_ids']))\r\n```\r\n\r\nThe output is:\r\n\r\n```\r\nversion datasets 1.1.3\r\nReusing dataset snli (\/home\/tom\/.cache\/huggingface\/datasets\/snli\/plain_text\/1.0.0\/bb1102591c6230bd78813e229d5dd4c7fbf4fc478cec28f298761eb69e5b537c)\r\nFirst item in dataset:\r\n {'premise': 'A person on a horse jumps over a broken down airplane.', 'hypothesis': 'A person is training his horse for a competition.', 'label': 1}\r\nTokenized hyp:\r\n {'input_ids': tensor([[0, 1, 2]])}\r\nLoading cached processed dataset at \/home\/tom\/.cache\/huggingface\/datasets\/snli\/plain_text\/1.0.0\/bb1102591c6230bd78813e229d5dd4c7fbf4fc478cec28f298761eb69e5b537c\/cache-fe38f449fe9ac46f.arrow\r\nTokenized using map:\r\n {'input_ids': [[0, 1, 2]]}\r\n<class 'torch.Tensor'> <class 'list'>\r\n```\r\n\r\nOr am I doing something wrong?\r\n\nIt is expected behavior, you should set the format to `\"torch\"` as you mentioned to get pytorch tensors back.\r\nBy default datasets returns pure python objects.","embeddings":[-0.0905342773,-0.1736622602,-0.1339069456,0.2130297422,0.2162035555,0.2193657011,0.4914948344,0.3712697923,0.257444948,-0.0263979062,-0.1461347193,0.680251956,-0.10030251,-0.4164583683,0.1840583086,0.0253589209,0.294393152,0.0346806832,-0.0507990569,-0.2774459422,-0.1831274182,-0.0675765797,-0.1957766712,0.0791181624,-0.1412031651,-0.1547512561,0.1446099281,-0.3113888204,-0.0627612844,-0.2368612587,0.1946785748,0.0335779414,0.1022556797,0.446719557,-0.0001160152,0.1183271185,-0.037234392,0.2079562694,-0.0176475514,-0.1034466103,0.0076182238,-0.1327908486,0.1470107734,-0.1844788641,-0.0054996279,-0.1479554027,-0.0391307659,-0.6350414753,0.179096505,0.3810082972,0.1443395168,0.2552904189,-0.1439363956,0.2130863816,0.3324943781,0.2167777121,-0.2084973902,-0.2550541461,0.1627887785,-0.0757924542,0.0668914467,0.1129236296,-0.4984409213,-0.0192023627,0.2370801419,0.2221133709,-0.1869863868,-0.5936381817,0.0203761477,0.1143103391,0.346229285,-0.3974611461,-0.1002945304,-0.1565368474,-0.2213947773,0.0423909165,-0.0511560366,0.0540001467,0.0545048267,-0.0863507465,-0.2528419197,-0.0814966038,0.0649670362,0.2210161388,-0.5532370806,0.2875820696,-0.0781859979,0.2344746888,0.1804774553,-0.0505854003,0.1507671028,-0.3792155981,0.0853365734,0.1579334438,-0.3272643685,-0.1862962246,0.1454360634,-0.4385431707,0.0561810508,-0.1962929219,0.1501737237,0.2216709554,-0.397954762,0.1758711338,0.1768113673,0.1449903399,-0.0155394245,0.680030942,-0.0505105294,-0.4166106284,-0.0770037621,0.183756128,0.0966400057,-0.0125526022,0.0351255648,-0.0809561312,0.1311855167,0.0802533701,-0.1056552529,0.1230914518,-0.5634173751,0.1094647199,0.002619667,0.0384143218,-0.186223343,0.1880778372,0.0646297261,-0.0047181626,0.0791275799,0.2578691542,-0.0984069183,-0.2119452506,-0.1304918379,-0.0817815512,-0.0984928384,-0.0916112587,0.0619072318,0.092382431,0.1118398234,-0.0017336064,0.2803901732,-0.0927288532,0.5385236144,0.1707926393,-0.1122141108,0.4933412373,0.2230668664,-0.0651279017,-0.2487444878,0.2211106271,-0.3083320856,0.1185796559,0.0200917516,0.1291921139,-0.0472025983,-0.0984765962,-0.2833484411,0.1163356304,0.2763032913,-0.0582993142,0.2532229424,-0.3349198103,-0.132949248,-0.1963020861,0.1043847948,-0.0451547392,-0.118641369,-0.0467285588,0.044093363,0.0888606682,0.0959444866,0.3504171371,-0.2296563685,0.3976287246,-0.1970900148,0.3188100755,0.3553226292,-0.1605848074,-0.4590785503,0.0325200707,-0.0414366685,0.2008669823,-0.4211300015,0.2581838071,0.5175717473,0.0115962904,0.1081317067,-0.0138313118,0.0394521281,0.0712760985,0.0759738982,0.0893174186,0.6063425541,-0.1942304671,0.0912254751,0.4368440509,-0.0899207368,0.1663245708,0.2264939547,-0.0353974365,-0.0143849719,0.2562342286,-0.0984062627,0.1964918971,0.0711438432,-0.1331272125,-0.3207593262,-0.0673473179,0.2024287283,0.0814435259,-0.0857608393,-0.2460781038,0.0505057834,0.1824002415,-0.1853736043,0.0724808052,0.1081486642,-0.0400793664,-0.2167155892,0.0308499429,-0.0139192659,-0.2510787547,0.1745028645,0.204618454,-0.4255838692,0.2819700539,0.1628724337,0.1375444531,-0.4352928102,0.1270138621,-0.0553200878,-0.0830142573,0.0988000557,0.0349606797,0.1280521005,0.0265865475,-0.0954571441,0.5214632154,0.2866054773,-0.2311563343,0.0751397461,0.4937567711,0.0238809288,-0.229771331,0.1961795241,0.1395006329,0.2197048515,0.3267371356,-0.4468351305,0.2423465103,0.036944773,0.2038594931,-0.3506708741,0.1286530793,-0.4682184458,0.2203543335,0.1738942266,0.1489115953,-0.1910985857,0.0657252818,0.4890259206,-0.1033508554,0.0994033739,-0.0586623512,-0.5665890574,-0.0065713013,0.2327365875,-0.2958332598,0.1299310476,-0.078820765,0.3676331341,0.0134283369,-0.0215824917,0.1092869341,0.0924237296,0.351996392,0.0202198178,-0.0760973096,0.0611620396,0.065281257,-0.3260114789,0.0977090001,-0.0819270909,0.1087761745,-0.1746973693,0.215378195,-0.2485268414,0.114612475,-0.0772436485,-0.4351750314,-0.1941461265,-0.1610156745,-0.0932937786,0.1621305048,0.0375246331,0.1928770095,0.3247022033,0.0016953748,0.3418312073,-0.225582853,-0.1251525581,-0.2449374646,-0.1197988689,0.0738311112,0.0181270093,-0.4264046252,0.083983548,-0.0697512254,-0.0664159432,-0.7368848324,-0.2431028038,0.1734049171,-0.2962734401,-0.2769447267,0.3027134836,0.2004784793,-0.3066086471,0.0971379206,-0.0186878163,0.1475964338,0.0467998534,0.0844613612,-0.1103945002,0.1160804108,-0.1308739036,-0.3476417065,-0.0094152419,-0.0166204292,0.2716757655,0.0838909224,0.1525645256,0.0774168074,0.3029705882,0.1268658638,0.0299354363,-0.1717898846,-0.1643576324,-0.2441188097,0.3611247838,-0.0520239957,-0.22654441,-0.0212237015,-0.0030592845,0.1346479952,0.5783620477,-0.3174059987,0.0344276801,-0.0420033671,0.2637937367,0.0175747406,0.1347250193,0.3775294721,0.2478036731,0.1076655388,-0.0430456065,-0.0288084429,-0.0538102761,-0.0161097497,0.3160053194,0.1498702914,0.3397210538,0.1404833645,0.6328849196,0.0862662196,-0.5870518684,-0.2141459286,-0.3498601615,0.2165563256,-0.1520948708,-0.3102630079,-0.1093504801,-0.28600806,-0.300843358,0.0750686973,-0.1314490288,-0.2657194734,0.2969166636,0.446225971,-0.0882325992,-0.1677335501,0.0572007596,-0.6227615476,0.1875963956,0.0649887696,0.2781692147,-0.3583343327,-0.1568048894,0.1168322787,-0.1129730567,0.3872697353,-0.1263889372,-0.7034638524,0.0063643605,-0.4204152524,0.1873979867,0.0056880182,0.1773734242,0.1652300209,-0.112509191,0.1744450778,0.3250636756,0.7036460638,-0.0586114973,-0.3005672395,0.1336065233,-0.0997832417,-0.4846154153,0.0846258402,-0.2593833804,0.2423414141,0.2093499899,0.3109831214,-0.0476541407,-0.2322614938,-0.3107507229,0.1333297491,-0.053243123,0.3153674006,-0.1331279129,-0.0941951349,-0.127183944,-0.1217167154,0.1963419467,-0.2370506227,-0.0338916741,-0.5592253804,-0.1888412088,-0.0711832419,0.1944610924,0.0427112952,0.2000722885,0.131830126,-0.2557500601,0.2216896415,0.1591052711,0.117755577,0.0787251666,-0.1197154671,-0.3400996625,-0.051767502,0.2129713744,0.1649908274,-0.1602071673,-0.141362384,0.1177048609,-0.2446500808,-0.1058980897,-0.3997228444,0.4764786363,0.5671220422,0.2579691112,-0.3157423139,-0.6125047803,0.2302146554,0.3774290681,-0.2791235447,0.4521859288,-0.5061218143,-0.3920195699,0.2071548253,0.4282876551,0.5739824772,-0.4119100869,0.2498632669,0.0765873864,0.2339149117,0.426009953,-0.4805026054,0.1393812001,-0.3452283144,0.1527659148,-0.2154876143,-0.2566210926,0.0636829063,0.1887102276,-0.0316687711,-0.1242906526,-0.0239557792,0.7876774669,0.0780502781,-0.1464799196,0.3955989182,-0.2671814561,0.0892166644,0.2036619037,-0.1212615892,0.2099370509,0.0092979362,0.0513104908,-0.1965031326,-0.127871111,0.0116996625,-0.0130519327,-0.2299930602,0.2618724704,0.1525951326,-0.201145187,0.1392355412,0.7396448851,0.2234844863,0.2516423762,-0.0744041055,-0.038147714,0.1769199222,0.2847000957,0.1070501581,-0.1848676354,0.0835148692,0.2304860801,-0.0115294596,0.2355246991,0.0631159917,-0.268345505,-0.4846624434,0.2600140274,0.5671141148,-0.3808259666,0.2440679818,0.1625121534,-0.0416189581,0.001362141,0.0675532818,-0.1471410394,-0.1325733364,0.4093293548,-0.1548150033,-0.1608565003,0.1107495502,0.3674654663,0.2076233774,-0.1533612311,0.2480866611,0.056940794,-0.0848241076,-0.2917854786,0.1096708104,0.2019403428,0.1349809766,0.2240548879,-0.3578024209,-0.1645009816,-0.1746764034,0.1499622911,-0.2525676191,-0.0387568548,-0.3988036513,-0.0235320553,-0.3721062243,-0.025249349,-0.0994727314,0.4052252769,0.1330139041,0.033003889,-0.2353828102,-0.2841766775,-0.2150231451,0.0614063554,0.1049758121,0.317512691,-0.0161709879,0.1339095384,-0.0781409293,-0.2015335411,0.080073081,0.1677751988,-0.0815930665,-0.2401295453,0.0324873514,0.0831723288,0.1623704135,-0.2706999779,-0.0207832586,0.0171218216,-0.1772500575,-0.3691253066,0.2644973099,0.3387384713,-0.0548426434,0.2520482838,0.1629133224,0.1922449619,0.0664829612,-0.2389813811,-0.0673396289,-0.12356361,0.0767102763,0.3552877307,0.0898754299,-0.0253588594,-0.2356754243,0.2498126924,0.4343939126,-0.1193523854,0.5966325402,-0.2622196674,0.1074622124,0.1145689413,0.2363798171,0.2327507734,-0.3498212397,-0.2978017032,0.097635977,0.1566759348,0.0503739156,-0.2055505514,0.3785065711,0.1290355772,0.0085398173,0.138445437,0.1084859595,0.1346496791,-0.0434607454,-0.2603903711,0.3331289887,-0.3271982074,0.4942014515,0.397609055,0.0265155472,0.0466422215,0.1061060429,-0.0197215285,-0.088384591,0.4616709054,0.2282730788,0.5523833632,0.30277583,-0.0340121798,0.5073820949,-0.0137486272,0.0007735395,-0.0157656074,-0.2453914583,0.2581062317,0.1902956069,0.0556507185,-0.0475449152,-0.2028522789,0.0577834547,0.1941241026,-0.3690662682,-0.0328410789,-0.1844906807,-0.009945306,0.2395032048,-0.209104225,-0.0376407951,-0.2905730605,0.176477477,-0.0967049152,-0.0216083918,-0.0340525955,-0.0009388835,-0.0679468364,-0.0062976726,0.0717409328,-0.1286428273,-0.4583580494,0.0727507025,-0.1606461555,0.1530968249,0.0665907785,0.2265326381,-0.1436648667,0.2057797313,-0.0782430917,-0.1337422132,-0.0354567468,0.1714806855,-0.2607640326,-0.2001806647,-0.0241287295,0.08529073,-0.0528155193,-0.1415355057,-0.2138057649,0.3650476933,0.0129362866,0.2667193711,-0.1424587518,-0.2983596027,-0.1448249519,-0.1189547405,-0.160856843,-0.1878195852,0.6996372342,-0.03504619,0.1697005779,-0.1595540792,0.0491679423,0.0152545413,0.2613005638,0.0646018758,0.1442367733,-0.0080209337,0.041236002,-0.1432537138,0.3525063097,0.240338102,0.1202170402,-0.0026343146,0.0231717173,0.0901697204,0.2375983447,-0.1209659055,-0.4073794484,0.0233071763,0.1407173723,0.0084185265,-0.0124782268,-0.5062525868,0.2792611718,0.0773458257,-0.1727735847,0.2210201025,0.0803734884,-0.0051450813,-0.3682461977,0.0631453693,-0.2415378541,-0.0329703949,0.4665831029,0.2855281532,0.1157337949,-0.0312674791,0.0886829048,-0.1312686205,-0.0826187134,-0.1515829265,0.1294198036,-0.3124343455,0.105223164,-0.2197997868,-0.0723182783,-0.2349605262,-0.1989329904,0.4463572204,-0.314909637,-0.4419314861,0.1768953651,-0.2677648664,0.0713506564,0.054126054,0.4623350799,0.1842846721,0.2511640489,-0.0895964429,-0.1723601669,0.2852713168,-0.4563175738,-0.3587994277,-0.045180615,-0.068057403,0.1567696035,-0.2037457973,-0.6132336259,-0.3314446807,0.4240120351,-0.3604155481,-0.1031689867,-0.1472704113,0.079053469,0.1174907014,-0.3405918181,0.1946748942,-0.0184210837,0.0094523104,0.2543967366,-0.4694918692]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1004","title":"how large datasets are handled under the hood ","comments":"This library uses Apache Arrow under the hood to store datasets on disk.\r\nThe advantage of Apache Arrow is that it allows to memory map the dataset. This allows to load datasets bigger than memory and with almost no RAM usage. It also offers excellent I\/O speed.\r\n\r\nFor example when you access one element or one batch\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nsquad = load_dataset(\"squad\", split=\"train\")\r\nfirst_element = squad[0]\r\none_batch = squad[:8]\r\n```\r\n\r\nthen only this element\/batch is loaded in memory, while the rest of the dataset is memory mapped.","body":"Hi\r\nI want to use multiple large datasets with a mapping style dataloader, where they cannot fit into memory, could you tell me how you handled the datasets under the hood? is this you bring all in memory in case of mapping style ones? or is this some sharding under the hood and you bring in memory when necessary, thanks ","comment_length":90,"text":"how large datasets are handled under the hood \nHi\r\nI want to use multiple large datasets with a mapping style dataloader, where they cannot fit into memory, could you tell me how you handled the datasets under the hood? is this you bring all in memory in case of mapping style ones? or is this some sharding under the hood and you bring in memory when necessary, thanks \nThis library uses Apache Arrow under the hood to store datasets on disk.\r\nThe advantage of Apache Arrow is that it allows to memory map the dataset. This allows to load datasets bigger than memory and with almost no RAM usage. It also offers excellent I\/O speed.\r\n\r\nFor example when you access one element or one batch\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nsquad = load_dataset(\"squad\", split=\"train\")\r\nfirst_element = squad[0]\r\none_batch = squad[:8]\r\n```\r\n\r\nthen only this element\/batch is loaded in memory, while the rest of the dataset is memory mapped.","embeddings":[-0.0588864759,-0.4006500244,-0.0841394514,0.4953401983,-0.0051892721,-0.0410410501,0.2562442422,0.082638748,0.4370158017,0.1028491035,-0.1519440413,0.0303455573,-0.1056119353,-0.0217184518,0.1242289692,0.061641559,0.039550595,-0.1121024564,-0.6995661855,0.1022263393,-0.2493334413,-0.3557489812,-0.1300130635,-0.3272945881,-0.3039337397,-0.3051180542,-0.1829487383,-0.0901290253,0.0206644628,-0.2042738348,-0.0386582613,-0.0815087706,0.2003152072,0.7517547011,-0.0001189604,-0.0367895141,0.1360406429,0.0017194661,0.0376435071,0.2387624979,-0.2333405614,-0.3989806473,-0.0278325807,-0.1681381911,0.2054624259,-0.3424188793,0.0848583505,-0.616220355,0.3852132559,-0.1248321608,0.1235423461,-0.2621011436,-0.0925903544,0.2763326168,0.3111933172,0.1433324367,0.0195619334,-0.1486619264,0.6805699468,-0.0344118811,-0.2302269489,-0.1037611589,0.1673593074,-0.1966172457,0.0919266045,-0.201681003,0.1234455183,-0.3608862162,-0.0673593357,0.1288553774,0.5153530836,-0.1416386366,-0.3262237012,-0.2012563944,-0.2796247602,-0.0023358271,-0.1063352972,0.4049009085,-0.0822648257,0.1108777225,-0.3683152795,-0.3405007124,-0.3337821364,-0.0366813987,-0.3001170754,0.1452180445,-0.062937662,0.0894539431,0.1375580728,-0.2816590667,0.2963627577,-0.4750454724,0.1944923401,0.2549041808,-0.1021594629,-0.2395146191,-0.1099360362,0.0570476651,0.4679745138,0.1335092336,0.2479615808,0.1755761802,-0.2547267079,0.0770526528,0.3933933973,0.1047822088,-0.2252570242,-0.0755659118,0.4390206337,-0.359421432,-0.1419167966,-0.1065338179,-0.062414363,-0.0423801355,-0.0418439098,-0.4785649776,-0.078477487,-0.2720260918,0.2171857357,-0.0676897094,-0.1527037919,-0.1355276257,0.1241803691,0.3209927976,-0.1514753848,0.2444750816,-0.0954088345,-0.0181139726,0.0559465848,0.121969372,-0.0666993782,0.0123048499,-0.3038210869,0.1711558253,0.1640012115,0.0028020029,0.1232638881,0.2098841369,-0.0685096681,-0.0771990791,0.2681392729,-0.2920338213,0.3498936296,0.2049546391,-0.0418092422,0.1418164074,-0.072726801,0.3579365611,-0.2949612737,0.4229598641,-0.4347180128,-0.2101899087,0.4934930205,0.0568359867,-0.1956836134,0.0365932435,-0.5651928782,0.0108646322,0.1202700511,0.005338362,0.0754910111,0.1426695585,-0.0617287979,-0.3355418146,-0.0220437553,0.1639938504,-0.5725691915,0.095140487,0.0164686739,0.1323545277,0.0291997623,0.4890319407,-0.2978397012,0.0071913837,-0.1534463763,-0.178645879,0.4220184684,-0.214909941,-0.4589166343,0.4631193578,0.362381041,0.1702606231,0.2086673528,0.188361302,0.0513136126,-0.1379454285,-0.1611617357,0.6018933058,-0.1137406304,0.1777114868,-0.0230638571,0.0201796833,-0.0609261394,0.0680001155,0.0143553047,-0.2378221154,0.3191002309,-0.2142076045,0.1988714933,-0.0789626762,0.1503863782,0.1449685544,0.0309887771,0.0816588029,0.0755418167,-0.0170645695,-0.4261789918,0.2068165243,-0.1172283739,-0.3329842985,0.0282216612,-0.0069538574,0.2477512956,0.0752912015,-0.0171729643,0.1064273342,-0.1028251648,-0.1765655875,-0.2068213373,-0.0904372931,-0.2467716485,0.2182663381,0.0694199577,0.2246511132,-0.3101911545,0.1304001063,0.2957300246,-0.1103137061,-0.195648402,-0.0994968191,-0.188249588,-0.1264005005,0.0898182988,0.0821364373,0.2953771353,-0.0338168219,0.0958972201,0.3368049562,0.4200541675,-0.1791235209,0.4618538618,-0.0391623415,0.1382342428,-0.3454586565,-0.2119730264,0.3631180823,-0.2875338793,0.5229017138,0.1235602126,-0.1351252198,0.0095715178,0.0218977872,-0.0130687412,0.2025154829,0.1118588671,0.2873811722,0.6247633696,0.2490035146,0.0394603796,-0.2190214843,0.2453068048,-0.0015913476,0.2007402629,0.0249179639,-0.4273910522,-0.0144782541,0.2521755099,-0.0906086937,0.2695960104,0.2748970687,0.1174039394,-0.1638914794,0.4040957689,-0.0336422846,0.1366795748,0.1346002966,0.3476347923,-0.0030691165,0.0086476887,-0.2951585352,0.0046866145,-0.2069446892,0.0752809346,-0.1986514181,-0.1057801768,0.1440148503,-0.2291140109,-0.2856042385,-0.2744174898,-0.2355762571,-0.4793237746,-0.0293898005,-0.2364707589,0.0844094828,0.0439957567,0.0969997048,0.027675895,0.6954143643,-0.0482198,-0.1201428846,0.0571183935,-0.029637143,-0.0915001631,0.0601225682,0.4931307137,-0.2695896626,0.3946196735,0.4713744521,0.0249749385,-0.132005766,-0.0868253186,0.0459545739,0.0567706898,0.0622712336,-0.3290964961,0.1470323205,0.0561653152,0.3827188313,-0.0509234108,0.0700895265,-0.054986503,0.0320945084,0.0162867215,0.3869161904,0.201121524,-0.1517053992,-0.2286734581,-0.3820271492,0.3547494113,0.1659956127,0.1937939376,0.0476233959,0.1845691651,-0.0287558027,-0.0829225555,-0.0636627525,0.1533194631,-0.4200549126,0.4128156006,-0.1698292345,-0.0909518301,-0.0368246101,0.1147509962,0.0962599665,0.4367291033,-0.2818885148,0.1796182841,-0.2459022701,-0.1230282784,0.0438913926,0.6782795191,0.2842325568,-0.1847536266,-0.0193747841,-0.0482371226,0.0945092142,-0.2058205903,-0.2530432343,0.4445666075,0.3438854516,0.3572461307,0.1479998529,0.7315521836,0.0413678251,0.1667902768,0.1807873696,0.2659292519,0.0448410548,-0.0270311162,0.059310481,-0.2050995678,0.0018673115,0.0555118099,0.2514944971,0.4081156254,-0.0600079745,-0.1175139546,-0.3410619795,0.10175284,-0.1194689944,0.5493801832,-0.4264600575,0.0512254052,-0.2740892172,-0.4009952247,-0.2083830833,-0.3367422223,-0.0265997928,-0.222328797,0.0735255033,-0.0055070226,-0.3459632695,0.0601938628,-0.3648677766,0.4680427313,-0.2491692752,-0.1178677902,0.2148746997,-0.159533754,0.0428810716,0.1266005784,0.2923495471,0.0148011614,-0.149348855,-0.2697579861,-0.4613231421,-0.3847885728,0.1961252093,-0.2139132619,0.3192955256,-0.0573563315,0.341753304,-0.0771781579,-0.2102856189,-0.152629301,0.2417679727,-0.1906874627,-0.4075164497,-0.1601244658,0.0599455424,-0.0553111583,-0.037753541,0.2017256618,-0.2686134577,-0.0661658645,-0.3019456565,-0.2867477834,-0.0922346786,0.2089640498,-0.1186999455,-0.0252168737,-0.2162247151,0.0735459849,0.1569156945,0.5434238315,0.3309314251,0.6896024346,0.0474419668,-0.1577114165,-0.0407936312,0.0503258854,0.7631646991,0.1033930182,0.0692786649,-0.047991626,-0.1760788709,-0.0175357182,-0.3799072504,0.0458383262,0.023721436,0.0447602533,-0.3666876853,-0.4932942688,0.749147296,0.4416844845,-0.1586073041,0.6419865489,0.1175266132,-0.3553230166,0.2868682742,0.0879770443,1.0530860424,-0.512976408,0.3421281874,-0.4214243591,0.1534468979,0.2286805958,-0.2978703976,0.0038785655,-0.2961543798,0.1843925118,0.0075096227,-0.2474556118,0.1965359151,0.322840333,-0.03417347,0.4432891011,-0.0440063439,-0.0497646406,-0.1011340693,0.1013288721,-0.2140790224,-0.2364328802,-0.1501041055,0.043413762,0.0592692569,-0.2337747216,-0.3037478626,0.2017507702,0.011406471,0.0741859972,0.1891050488,-0.1522643417,-0.2654876113,0.2479362041,-0.334143281,-0.3019464612,0.3705105186,0.1748022139,-0.4805151224,0.1436814368,0.1476095915,0.0156086879,-0.0116144838,0.0454970561,-0.0424078628,0.0417146757,0.2375325263,0.0139356311,-0.1776860058,0.2402747869,-0.040700458,-0.3315973282,-0.2111346573,0.2733591199,0.40356794,-0.2380135506,-0.1234077141,0.0345751494,-0.0764058754,0.2733542919,0.071725294,0.2668690383,0.0065926337,0.7183684111,0.1413907558,-0.2909387052,-0.0997757316,0.161968708,0.2954937518,-0.1225241274,0.1843900084,0.2407252192,-0.0996517465,-0.1255910993,0.4894376397,0.5721589923,0.0064000082,0.1267372519,-0.1524837166,0.0716017485,-0.0288003162,-0.0673772618,0.2119221985,-0.1597386152,-0.1866055727,-0.2719227076,-0.3603436351,0.4172888398,-0.0010326522,0.2916476727,-0.2325250357,-0.4795269966,-0.1086044908,0.2901135087,-0.2042009681,0.1051967889,0.2014524341,0.1602000594,-0.0514426418,0.3269394934,-0.1814511716,0.1690793931,-0.0361963585,0.1177322492,-0.0470622033,-0.1597383916,0.0736398995,0.2084085941,0.1023759246,-0.2409004122,-0.0610242598,-0.4998578429,0.1702677757,-0.1786268502,0.1662011147,0.0231718309,0.1309781075,-0.1394669861,0.2782172859,-0.2474117577,-0.232793659,0.5235774517,0.1672568768,0.3377898335,0.1600790769,0.1526779979,0.0924206525,-0.3020650744,-0.2376357168,0.0810125843,0.5221099854,-0.0775743499,0.2776810229,-0.1506262273,0.0485768467,0.3529634476,0.162630856,0.2794004381,-0.1803961098,0.0967006311,0.1010616124,0.1215696633,0.0658566281,-0.0772052631,-0.0332974158,-0.0593744367,-0.0464187562,0.3802596033,0.0765118524,0.2080688924,-0.4234376848,-0.163489297,0.2168819904,0.0777112767,-0.2301083207,-0.0167401358,0.01490987,-0.2354390621,0.5816823244,0.5420168638,0.2874979973,0.3710432649,-0.0723345205,0.3936845064,0.0437646247,-0.266138345,0.1471127272,-0.5092186928,-0.0709941909,0.5397107601,-0.0892358199,-0.0474878736,-0.138450861,0.3336668611,0.1360914558,-0.1799747944,0.0582269467,0.0999576598,-0.1680313796,-0.0200085118,-0.0613547377,-0.2377376407,-0.0461174101,0.1220200732,-0.2352475971,-0.3217306137,0.5591568351,0.1718376279,0.2663176358,-0.0406023487,0.3802272975,0.2251405865,0.2252901047,-0.2364737242,0.1622788459,-0.0555886216,0.0773999766,-0.0306830592,0.0617590398,0.1170413941,0.0487930365,-0.1427313387,0.2462531626,-0.2952980399,-0.0831240714,-0.1168901101,0.0976505056,-0.065614298,-0.0480974913,0.1170056015,0.0314985588,-0.0452126227,-0.0958296806,-0.1793905646,-0.1782430261,-0.3034231663,0.0829633847,-0.016464198,-0.381731689,-0.026554225,0.1691401452,-0.3279438019,-0.1619229764,0.8945940733,-0.4270968437,-0.0632167608,0.0319752619,0.0031154752,-0.2406429201,0.4338238537,0.4252313673,0.3689442873,-0.1785404682,-0.2688188851,-0.2837274671,0.1266162694,-0.3786372244,0.2014346719,-0.2451912612,0.3513478339,0.1779806912,0.1477283984,0.1775557101,-0.2818575203,-0.0299362335,0.0541761331,-0.2798920274,-0.148404941,0.0503982492,0.396819979,-0.0759716332,-0.349755466,0.0906404778,-0.214778319,-0.0670952648,-0.1379139572,0.1266917586,0.2633190751,0.1132462993,0.4393740594,-0.0053255223,0.4129464328,0.2855178118,0.0577972382,0.0514244437,-0.0933442339,0.1209214851,0.0183623955,-0.0268532243,-0.1251738518,-0.4177531004,-0.1935304701,0.059753824,-0.1311771572,-0.3699883223,-0.1227177233,0.0791654959,-0.4259049594,-0.0719489008,0.0059338873,0.1723584384,0.4510594606,-0.1976388246,-0.2763511539,-0.2034581751,0.0344838612,0.0280854441,-0.1240314916,-0.2196734399,-0.072981447,0.2027069181,0.0024563198,0.0743601024,-0.2780815065,-0.0264943186,0.4451416731,0.0745479092,-0.2586443126,0.1050809845,-0.0525454953,-0.0332913287,-0.2224711031,0.2652715147,-0.1851905882,0.0087166671,-0.2955729067,-0.1870116591]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1004","title":"how large datasets are handled under the hood ","comments":"How can we change how much data is loaded to memory with Arrow? I think that I am having some performance issue with it. When Arrow loads the data from disk it does it in multiprocess? It's almost twice slower training with arrow than in memory.\r\n\r\nEDIT:\r\nMy fault! I had not seen the `dataloader_num_workers` in `TrainingArguments` ! Now I can parallelize and go fast! Sorry, and thanks.","body":"Hi\r\nI want to use multiple large datasets with a mapping style dataloader, where they cannot fit into memory, could you tell me how you handled the datasets under the hood? is this you bring all in memory in case of mapping style ones? or is this some sharding under the hood and you bring in memory when necessary, thanks ","comment_length":68,"text":"how large datasets are handled under the hood \nHi\r\nI want to use multiple large datasets with a mapping style dataloader, where they cannot fit into memory, could you tell me how you handled the datasets under the hood? is this you bring all in memory in case of mapping style ones? or is this some sharding under the hood and you bring in memory when necessary, thanks \nHow can we change how much data is loaded to memory with Arrow? I think that I am having some performance issue with it. When Arrow loads the data from disk it does it in multiprocess? It's almost twice slower training with arrow than in memory.\r\n\r\nEDIT:\r\nMy fault! I had not seen the `dataloader_num_workers` in `TrainingArguments` ! Now I can parallelize and go fast! Sorry, and thanks.","embeddings":[-0.1215671673,-0.4438164532,-0.1115840673,0.4983528852,0.067883864,-0.1806738824,0.197150439,-0.0075913304,0.4332483113,0.0732882097,-0.1071146503,0.2367630154,-0.1338435113,0.0458660387,-0.0414698794,-0.0014341233,0.0602205321,-0.0184701513,-0.6697032452,0.06769678,-0.1225022003,-0.2739754319,-0.2129993737,-0.2707862258,-0.4327622652,-0.3156321049,-0.0948926508,-0.0439279675,-0.0171791948,-0.1848109812,0.0234301481,-0.1452039033,0.1936464906,0.7684065104,-0.0001201546,-0.1148603484,0.1988204718,0.0495708138,0.1518303752,0.4270415008,-0.2958095372,-0.3880932331,0.0579504557,-0.1263776124,0.1963099837,-0.3373782337,-0.0204911698,-0.3247038424,0.4007909596,-0.0438235365,0.1147308648,-0.3078790605,-0.0085201412,0.2554507554,0.1298278868,0.0812880397,-0.0228819046,0.0341698378,0.6602918506,0.0674037188,-0.3207933307,0.0198224075,0.3828050196,-0.0876138061,0.0810198709,-0.2135655135,0.1960029751,-0.3160541058,-0.1036865786,0.2220429629,0.5307355523,0.0745657608,-0.2137930095,-0.1574528664,-0.1965864152,-0.1126702949,0.0260423962,0.1900892407,-0.1245145053,0.0925197676,-0.4205059111,-0.384206444,-0.3212280273,-0.1058504358,-0.1592171639,0.0602541268,-0.144015789,0.0976159871,0.2512976825,-0.2444238365,0.3467597663,-0.2242797315,0.0637847334,0.2126442492,-0.3167775571,-0.3181909919,-0.3145974576,0.0072492859,0.2140951008,0.0262053572,0.2432012856,0.2331735939,-0.2770914733,0.0579361878,0.373039335,0.1893132627,-0.2583979964,-0.0292445906,0.2394621968,-0.3323382139,-0.2076711804,-0.0278383736,-0.1254138649,-0.1924134493,0.0290484224,-0.45896402,-0.3164376318,-0.22271806,0.1845161468,-0.1283962131,-0.0210293718,-0.2380319089,0.1576445848,0.3425010741,-0.1521487832,0.4732321203,-0.1909871697,-0.1226953715,0.0920948312,0.0799302161,-0.0951354131,-0.0252468791,-0.444096148,0.0891451836,0.1055520698,0.1422843188,0.1316375732,0.1794971377,-0.2031678706,-0.0553410649,0.1711278856,-0.3524627984,0.2662963271,0.085853599,-0.1850575507,0.0771692842,-0.1993013173,0.5173634291,-0.1493874937,0.3545512259,-0.6022629738,-0.2095153332,0.3533734977,0.0607273914,0.0541064776,-0.0319280885,-0.5426729321,0.061998006,0.1068783551,0.0489877872,0.1851813942,0.0778620765,0.0058571994,-0.2848335803,-0.0387132503,-0.0621048734,-0.5078706145,0.0839684904,0.0214739274,0.0379009359,0.3106988072,0.5393097401,-0.334410727,0.0553935207,-0.1138814688,-0.1326754838,0.4459807277,-0.2488920093,-0.3662179112,0.5615829825,0.21416834,0.0004851517,0.253995657,0.170931682,-0.0130625069,-0.0812172294,-0.1040080711,0.4598284662,-0.0227198713,0.1220827773,0.0063937814,-0.054606162,-0.1883065999,0.1092321277,-0.0403846428,-0.2716549933,0.2416463643,-0.208726719,0.2332989424,-0.0532250181,0.0906439573,0.1668754965,-0.017967917,-0.0557855554,0.0768469647,0.1827198118,-0.4248358607,0.1400219202,-0.1323659122,-0.4478213787,0.2587947547,0.1581343561,0.2468056828,0.0421299823,0.0565526597,0.0603841655,-0.1120461002,-0.1753149778,-0.1937236339,-0.2523959875,-0.229202196,0.2160687596,-0.0651035309,0.1158558875,-0.0814065859,0.0038226172,0.3082476854,-0.1531687677,-0.2304404527,-0.1268645078,-0.2324857861,-0.1005854234,-0.0570703186,0.158270821,0.2823875546,0.0536024198,0.2251091003,0.25341326,0.3548586965,-0.2406089157,0.3344473839,-0.065612942,0.0892825872,-0.28173244,-0.2128619254,0.1794786751,-0.3186481595,0.4223964512,0.0850994959,-0.1555410326,0.0558321327,-0.0452408716,0.0845040828,0.252423048,0.1671889126,0.1530891806,0.5587385297,0.3262904286,-0.1059882715,-0.265664041,0.2755075991,0.1176057309,0.0361893289,0.0962628946,-0.4059443772,-0.0126217231,0.3432349861,-0.0570628755,0.322545141,0.270318687,0.1609900147,-0.2113861889,0.4023603499,-0.0841270611,0.153287217,0.043371018,0.5052736402,0.1157467887,-0.0274701584,-0.3022116125,0.0172642786,-0.1175325066,0.161785394,-0.1426556259,-0.0345931686,0.0774320811,-0.1431572139,-0.15403153,-0.2296726108,-0.2347118258,-0.226842016,0.1399940699,-0.2304358035,0.0682878867,0.0375845134,0.1580303162,-0.0489989631,0.6598153114,-0.1480918676,0.0952270404,0.0034780821,0.0267199781,-0.1317080408,0.0593353435,0.357062906,-0.3354431093,0.2742570341,0.4707393646,-0.0013770198,-0.075243555,0.0400803909,0.050390631,0.165208593,-0.0751269087,-0.330455184,0.0211023819,-0.1063793525,0.2256067842,-0.0876437351,-0.1321447343,-0.1368320137,-0.0243298598,-0.0768870562,0.3532197773,0.2285236567,-0.1721686125,-0.2935732901,-0.37499246,0.5281118155,0.1634026319,0.1652739793,0.0051041623,0.2578337193,0.0530497581,-0.1767074615,-0.1510028094,0.1319056451,-0.4742539525,0.3149018586,-0.0819231942,-0.0689846054,-0.1042475328,0.1330755502,0.2088453472,0.488294512,-0.2809638083,0.1956727058,-0.2383020371,-0.1123147234,-0.0271476898,0.5863160491,0.3648005724,-0.2229548246,-0.116202265,-0.0678480715,0.1707149744,0.0376973338,-0.1658644378,0.47704795,0.2928766906,0.2151965797,0.0553326942,0.7928889394,0.0288615357,0.0240357071,0.2115834206,0.2363714278,-0.0143146822,0.0673245713,0.0543309934,-0.0071293735,-0.1185246259,0.1347329468,0.2699111998,0.2387024164,-0.2775375247,-0.0683466122,-0.2236427367,0.0654868558,0.0012613561,0.6162108183,-0.4705193639,0.0677061155,-0.3641810417,-0.3642459512,-0.2563403547,-0.3372337818,0.0286780912,-0.415347755,0.050631728,-0.1354021728,-0.3786631823,0.0277164672,-0.3058374524,0.3697522879,-0.2064961642,0.0653835461,0.2111525238,-0.097457543,0.0683728233,0.1143167764,0.3644967973,0.0273176897,-0.1912029833,-0.3383257389,-0.3897688091,-0.3269128799,0.1274356842,-0.2958537638,0.3514429033,0.0417007133,0.3639027178,-0.0888681188,-0.3185628653,-0.2023007125,-0.0053315945,-0.1367944181,-0.359798193,-0.0470373631,0.3136752546,-0.0042880448,0.1570521593,0.179333806,-0.226377368,-0.0278576799,-0.3649969101,-0.3292686939,-0.0495650209,0.1219874024,-0.0408016741,0.0632385164,-0.0552371591,0.0639039353,0.163278386,0.5767408609,0.380877614,0.6005833149,-0.132429108,-0.0862098932,0.0495775342,0.0185183473,0.7553296685,0.2143443376,0.1131468192,-0.0358715355,-0.0748905763,0.0954969302,-0.3298135102,0.225569129,0.2111706287,0.112803787,-0.2189625949,-0.2797259986,0.636715591,0.4975313544,-0.206094265,0.681525588,0.1078051627,-0.3115997612,0.1510140896,0.198617667,1.073299408,-0.3391585946,0.2563871443,-0.4001900554,0.0107423887,0.2980068028,-0.2192751467,-0.0169277918,-0.2645283341,0.1219550371,0.1109512895,-0.2917241752,0.0577327311,0.4232118726,0.0188758876,0.3320441842,-0.0719197616,-0.2329235375,-0.1749434024,0.1831757426,0.0247573294,-0.3504062891,0.0221556313,0.0375883691,0.0225735977,-0.2787261009,-0.2484208494,0.0610426329,-0.0144863091,0.1264961362,0.193210423,-0.1726915538,-0.4433511794,0.2086639553,-0.0599603876,-0.2894452214,0.2820929587,0.1990351826,-0.4780541062,-0.1115258038,0.0340762101,0.0615906715,0.0477529615,0.0596268512,-0.0487140082,-0.0240328778,0.2609966397,-0.0870554671,-0.183052808,0.2442506403,-0.1225936338,-0.4327818751,-0.0582565665,0.2496171296,0.3004877865,-0.1192242429,-0.2311590165,0.0204099026,-0.151182726,0.2517453432,0.0836270899,0.2774944305,-0.0500974506,0.790972054,0.2494159341,-0.2567221224,0.0019732132,0.130725652,0.2735644281,-0.0689034909,0.2961666882,0.1190427169,-0.0571287498,-0.1272889078,0.2883098423,0.7976261377,-0.1685119271,0.0926268101,-0.0829770714,-0.1534589827,0.0405344591,-0.0706321597,0.3782191277,0.0092466408,-0.2428781986,-0.1243839934,-0.5275825262,0.3954949677,-0.1356111318,0.2484888881,-0.4305975437,-0.4996179044,-0.0290927943,0.5126160383,-0.1911392659,0.0985944569,0.0381435528,0.1594442874,-0.0553248338,0.1617840379,-0.1339916885,0.2139783651,-0.0627117082,0.1075630859,-0.0651446506,-0.1699531525,0.0806177557,0.1879484653,0.12133196,-0.2216005176,-0.0124038327,-0.510140419,0.1077816039,-0.125304684,-0.0457675345,-0.1900532246,0.132308498,-0.1343142837,0.2778827846,-0.216713354,-0.2320339084,0.5406051874,0.291130662,0.3465308547,-0.0482354052,0.0967061967,0.235479936,-0.3710894883,-0.1831847429,0.0099967876,0.5146935582,-0.0649757981,0.0711035132,-0.1131519675,0.0666726753,0.2207450569,0.1693572402,0.3511228263,-0.1985305846,-0.0276741404,0.1518449932,0.1196762547,0.1336817443,-0.1367549151,-0.0212045778,-0.0366332904,-0.0523795299,0.2293815166,0.1747460961,0.1164897084,-0.3455793858,-0.0876054242,-0.15219827,-0.09009552,-0.3171087503,0.0032006081,-0.0952976868,-0.1678313017,0.7194027901,0.4998329878,0.2111177593,0.3042778373,0.0560458116,0.3872670531,0.1095195934,-0.15522255,0.1879204214,-0.407227993,-0.2156740129,0.4330636263,-0.1631530076,-0.08576601,-0.1494219899,0.4096547365,0.1309779435,-0.4170706272,0.0842153728,-0.0093061151,-0.2595573068,-0.0334639102,-0.1247088462,-0.2547791004,-0.063251473,0.1400958449,-0.3406799138,-0.1436422914,0.7731212974,0.253367275,0.1256374419,0.0517148003,0.1903752089,0.3081794977,0.2620790899,-0.2052369714,0.2533287108,-0.1910855323,0.0156280678,-0.0730993971,0.2836798728,0.0875423476,0.1881262511,-0.0988175496,0.1186792329,-0.0183655452,-0.074338764,-0.0485588983,0.2628978193,0.1105057076,-0.1783749014,0.0129785258,0.0354523808,-0.1066150144,-0.0415150076,-0.2173056006,-0.0954137892,-0.1136345863,0.0578153282,0.1335176677,-0.3826259375,0.0070565427,0.1573422551,-0.2919599712,0.0127275325,0.6628470421,-0.6076680422,-0.0422040336,0.0380575769,0.0058590593,-0.2538717091,0.6007269025,0.3087890744,0.440982759,-0.2053620219,-0.291845113,-0.3426283598,0.1966382265,-0.2951281667,0.1858118027,-0.1586432904,0.4188531935,0.1239365488,0.0961690694,0.2513204217,-0.3321963847,0.106666632,0.1493688673,-0.4687696993,0.0747141168,0.2115774304,0.4281837046,0.0526448637,-0.3922716975,0.2519932985,-0.1400500387,-0.1164679006,0.0479285307,0.1025740132,0.1349616349,0.2817673385,0.4300050437,-0.0734965876,0.3949261606,0.2003695816,0.0717505887,0.0910513029,-0.0690013096,0.0167735461,-0.0334710144,0.1069194824,-0.0096137905,-0.3980923891,-0.046117153,0.0858110636,-0.3152442575,-0.4286338985,-0.030279303,0.0765184462,-0.4270525575,0.0555779859,-0.0600565895,0.3090154529,0.458004415,-0.1033366248,-0.4279014766,-0.3046090603,0.0688320994,0.0634723902,-0.1068477482,-0.1119825915,-0.1420947164,0.2086123079,0.1609275639,0.1177033558,-0.2550452948,-0.0825882852,0.3437247574,0.0126334792,-0.2696295381,0.1697088182,-0.0544390567,0.0863038749,-0.1577997059,0.3956686854,-0.1610727459,0.0231102947,-0.4086108506,-0.1885232627]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/1004","title":"how large datasets are handled under the hood ","comments":"> How can we change how much data is loaded to memory with Arrow? I think that I am having some performance issue with it. When Arrow loads the data from disk it does it in multiprocess? It's almost twice slower training with arrow than in memory.\r\n\r\nLoading arrow data from disk is done with memory-mapping. This allows to load huge datasets without filling your RAM.\r\nMemory mapping is almost instantaneous and is done within one process.\r\n\r\nThen, the speed of querying examples from the dataset is I\/O bounded depending on your disk. If it's an SSD then fetching examples from the dataset will be very fast.\r\nBut since the I\/O speed of an SSD is lower than the one of RAM it's expected to be slower to fetch data from disk than from memory.\r\nStill, if you load the dataset in different processes then it can be faster but there will still be the I\/O bottleneck of the disk.\r\n\r\n> EDIT:\r\n> My fault! I had not seen the `dataloader_num_workers` in `TrainingArguments` ! Now I can parallelize and go fast! Sorry, and thanks.\r\n\r\nOk let me know if that helps !\r\n","body":"Hi\r\nI want to use multiple large datasets with a mapping style dataloader, where they cannot fit into memory, could you tell me how you handled the datasets under the hood? is this you bring all in memory in case of mapping style ones? or is this some sharding under the hood and you bring in memory when necessary, thanks ","comment_length":192,"text":"how large datasets are handled under the hood \nHi\r\nI want to use multiple large datasets with a mapping style dataloader, where they cannot fit into memory, could you tell me how you handled the datasets under the hood? is this you bring all in memory in case of mapping style ones? or is this some sharding under the hood and you bring in memory when necessary, thanks \n> How can we change how much data is loaded to memory with Arrow? I think that I am having some performance issue with it. When Arrow loads the data from disk it does it in multiprocess? It's almost twice slower training with arrow than in memory.\r\n\r\nLoading arrow data from disk is done with memory-mapping. This allows to load huge datasets without filling your RAM.\r\nMemory mapping is almost instantaneous and is done within one process.\r\n\r\nThen, the speed of querying examples from the dataset is I\/O bounded depending on your disk. If it's an SSD then fetching examples from the dataset will be very fast.\r\nBut since the I\/O speed of an SSD is lower than the one of RAM it's expected to be slower to fetch data from disk than from memory.\r\nStill, if you load the dataset in different processes then it can be faster but there will still be the I\/O bottleneck of the disk.\r\n\r\n> EDIT:\r\n> My fault! I had not seen the `dataloader_num_workers` in `TrainingArguments` ! Now I can parallelize and go fast! Sorry, and thanks.\r\n\r\nOk let me know if that helps !\r\n","embeddings":[-0.1094140559,-0.4308580458,-0.0871282145,0.4903816283,0.0192652438,-0.2002978623,0.1710616499,0.0366569124,0.4349535108,0.0388859324,-0.1141780913,0.255189091,-0.1232734472,0.0230295472,-0.0365431942,-0.0497361943,0.078658551,-0.032821767,-0.6193565726,0.1037578136,-0.1304108351,-0.2858450413,-0.1606623381,-0.2703860998,-0.374741137,-0.3324694037,-0.0884396583,-0.038149666,-0.0140343411,-0.2108242363,0.0616952069,-0.0937612951,0.2144096345,0.7233284712,-0.0001186684,-0.0918641016,0.1741971672,0.0445404537,0.0798384473,0.3599433601,-0.2897218168,-0.3927341998,0.0601457208,-0.1590089202,0.2115675211,-0.3300986886,0.0054535544,-0.406844914,0.3743906021,-0.0250430312,0.0991554707,-0.2730448246,-0.0589687005,0.2336074263,0.086818032,0.0922659338,-0.010505178,0.0626730323,0.6923478246,0.0525821187,-0.3609701991,0.0302438959,0.3175122142,0.0123824496,0.1165365651,-0.2114973217,0.1849644631,-0.2949663401,-0.1062445641,0.2516591251,0.5105578303,0.0641362891,-0.2390613556,-0.2338179797,-0.1869091243,-0.1160810515,0.029436959,0.2002780735,-0.1415206939,0.0633397698,-0.4293504357,-0.3593241572,-0.2917678356,-0.0438303649,-0.0944022983,0.0374131799,-0.1127544418,0.1373373419,0.2806222737,-0.2216241658,0.3802333474,-0.3126363456,0.1021855101,0.1682952344,-0.34418872,-0.2493542582,-0.3483252227,-0.0298017878,0.262118876,0.0415903516,0.2865476608,0.2560191751,-0.2070029378,0.0754499957,0.3970377147,0.1765200049,-0.2426723391,0.006630295,0.1881208271,-0.2431785613,-0.189987272,0.0183487572,-0.1382554471,-0.1635321677,0.0175347459,-0.4711621404,-0.3381839097,-0.1440287977,0.1884435117,-0.1743113846,-0.0635789186,-0.2412644923,0.1829616427,0.3198435605,-0.1800623834,0.4577937722,-0.1897554398,-0.0614738278,0.0486692786,0.0233896598,-0.0542968847,-0.0407357588,-0.4293087721,0.1759207398,0.1483029872,0.1976806521,0.1073697433,0.1569950879,-0.2365114242,-0.0819059536,0.1956626773,-0.3922347426,0.2660459876,0.0276158527,-0.1685373634,0.114251256,-0.1680119485,0.4740410149,-0.1923685223,0.374212712,-0.5957955718,-0.1878307909,0.3516031504,0.0461341552,-0.0175210331,-0.0504430383,-0.6256650686,0.1540378779,0.1236738339,0.0126045831,0.1617602557,0.0536949225,-0.0882831141,-0.2674564421,0.0252437443,-0.0110351825,-0.4720745087,0.0524091311,-0.00689208,0.1041392237,0.2835587561,0.5517659187,-0.4106472433,0.1135390922,-0.1192628443,-0.1173923016,0.417812109,-0.2137856632,-0.4303643107,0.5667306781,0.209075138,0.1083019972,0.2623871565,0.1652846038,0.0832058266,-0.0930705816,-0.059861958,0.4544083178,0.0080702081,0.1396607757,-0.0041866316,-0.0941515863,-0.159570545,0.1432994306,-0.0489722975,-0.2116727382,0.1907656938,-0.2539235055,0.2201652676,-0.0919652134,0.1123252735,0.1764955819,-0.0526342914,-0.0457840785,0.0468089059,0.2086643279,-0.4808402658,0.2109634727,-0.1099825725,-0.40636608,0.2123639286,0.19595927,0.2658372223,0.0164994448,0.0127253914,0.0896878764,-0.1028103232,-0.1301660538,-0.1081871614,-0.2483939976,-0.1687362045,0.2720019519,-0.0546696186,0.0887358636,-0.1681142151,0.0489379764,0.3127925098,-0.1338700354,-0.231592834,-0.1520959586,-0.2393630892,-0.1030456945,-0.0203980822,0.1859755516,0.23611027,0.0516227223,0.1734180748,0.33000651,0.3617977202,-0.2407522947,0.331685245,-0.059569817,0.1335672885,-0.2932128012,-0.1862156838,0.205185473,-0.2873652577,0.4769780338,0.0904592946,-0.1368916482,0.0462920889,-0.0474574305,0.0434842855,0.256268084,0.1648508459,0.1016660333,0.5824890137,0.292388469,-0.1364488006,-0.287222743,0.2512176037,0.1232348084,0.0444282554,0.1085529923,-0.3892177343,-0.0254145283,0.3360786438,-0.0132308826,0.3325226903,0.2566536963,0.1349617839,-0.2326954603,0.351439476,-0.0757604092,0.1290671974,0.1013587564,0.5716062188,0.0419178791,-0.0346123725,-0.3070819676,0.0590891354,-0.179525584,0.1923035085,-0.1513794363,-0.0766243339,0.1248512045,-0.125127703,-0.1016531289,-0.2586200833,-0.2021533549,-0.2168625146,0.1597631723,-0.2158708721,0.1099507287,0.0296053234,0.1242319793,-0.0525167212,0.5798897147,-0.1888747811,-0.0299105234,0.0334783942,-0.0032975054,-0.1046979055,0.0513123013,0.4138850868,-0.2695017457,0.2428399622,0.403031826,0.0128974235,-0.0930936933,0.0647097081,0.0429363176,0.1651882082,-0.0084680505,-0.3586981595,0.0244836155,-0.1535338759,0.1645013243,-0.0560683608,-0.1228211522,-0.0865571648,-0.0029996345,-0.0506532155,0.3680099249,0.2494110614,-0.1850355566,-0.2905768454,-0.3868479431,0.4570654631,0.1317229271,0.2138234079,-0.0320347212,0.313857168,0.0434479751,-0.1027893052,-0.0727140978,0.1427296996,-0.4498758912,0.3469192982,-0.0438542403,-0.1049363017,-0.1490645409,0.0617099963,0.1873238981,0.5508527756,-0.3123829961,0.2569415569,-0.2491494268,-0.1047134995,-0.0540878586,0.5024775267,0.3338576555,-0.2245719731,-0.0651558042,-0.0494262464,0.1075888574,0.0588983372,-0.1344208568,0.4623301029,0.362972647,0.207198292,0.1127268076,0.8515409231,0.0474615395,-0.0225949753,0.2076239735,0.2847183943,-0.0164002404,0.0514184609,0.0425158739,-0.0239238963,-0.1738159508,0.1217253655,0.2867704332,0.1645157486,-0.3466246724,-0.0741961747,-0.1795201302,0.0329018719,-0.0355861299,0.6288302541,-0.4904952943,0.0746556446,-0.3362239003,-0.3155008256,-0.2783941627,-0.2954708338,0.0321417451,-0.4350835383,0.0736292079,-0.0972021297,-0.414013654,0.0775749758,-0.3163492978,0.3598531485,-0.1669425964,0.080343388,0.2238841355,-0.1482654959,0.0656244829,0.1223185658,0.4185760021,0.0593040287,-0.191244483,-0.3322924972,-0.3795841932,-0.3738748729,0.1434661448,-0.2774178386,0.3311069012,0.0264470056,0.3643664718,-0.1758686155,-0.2965624928,-0.2337972373,0.0115196696,-0.1195291728,-0.4018985927,-0.0110540893,0.2433244735,-0.0414344743,0.1138633788,0.1448019147,-0.1673829257,0.0252785478,-0.3045027256,-0.3236111999,-0.0033874048,0.1566354334,-0.0510360859,0.0762694553,-0.0333356,0.0881538093,0.2125045359,0.5491750836,0.4417771399,0.6041042209,-0.1456149668,-0.1475081295,0.0422990993,0.0767437071,0.627581656,0.1844051033,0.1692971736,-0.0458271466,0.0028007941,0.0699223354,-0.3473648131,0.2020463645,0.1494476646,0.1004863679,-0.258120805,-0.3264258504,0.6474955678,0.4854881763,-0.2168371081,0.646934092,0.0751323178,-0.3495358527,0.1484859437,0.2040815502,0.9574038386,-0.387441963,0.2295381576,-0.3754864335,-0.0699312091,0.2560865879,-0.1462657601,-0.0042834571,-0.2057727575,0.073818326,0.1323818117,-0.2957795262,0.1023078188,0.4680591524,-0.005251199,0.3361812532,-0.0461962335,-0.2531687915,-0.1590358019,0.2207282335,0.0578307286,-0.4079686701,-0.0023678422,0.0437045842,-0.0328821912,-0.3417801857,-0.1971239746,0.0073157265,-0.1261062324,0.1556927413,0.233646661,-0.1472529173,-0.4819092751,0.2061094046,-0.0214655809,-0.2980781794,0.3455943167,0.1853111535,-0.5106198788,-0.1934552193,0.0024701257,0.0441160984,0.0646185726,0.0611228906,-0.0074635809,-0.081553027,0.280911833,-0.0454711206,-0.148222357,0.2299865484,-0.119441852,-0.426666826,-0.0451175682,0.1846520752,0.2980019152,-0.0968153253,-0.2881253064,0.0538215637,-0.1638363749,0.2378149033,0.0783949718,0.2507181168,-0.0493798628,0.769258678,0.2077457607,-0.2335702181,0.0121915601,0.147860691,0.2357034981,-0.0680421069,0.3137345314,0.1327741742,-0.0774321631,-0.0974882618,0.3043731153,0.7953406572,-0.1697614789,0.0399803147,-0.1523754448,-0.0817265958,0.0579848774,-0.0974015668,0.3811481893,0.0190900695,-0.2097109407,-0.0944109783,-0.4866094291,0.4007515013,-0.1191908494,0.1855312586,-0.4160959125,-0.5647240877,-0.0367179774,0.5318692327,-0.1963838488,0.0946612954,0.0556866154,0.1922022104,-0.0515156426,0.1955499649,-0.1304434836,0.213013947,-0.0663412586,0.073248297,-0.1183957234,-0.1491827667,0.1228653938,0.1845219731,0.1135303825,-0.2815809846,-0.0595920905,-0.5127946734,0.1128903106,-0.1944173127,-0.0809625089,-0.1357211471,0.0704331249,-0.1956106573,0.2361652702,-0.193059966,-0.2789949477,0.506896615,0.2091393471,0.305881083,0.0017363444,0.1235039458,0.2212914973,-0.3678734303,-0.1917039305,0.0420166329,0.4997470081,-0.087818332,0.1413146406,-0.1559048295,0.1018661037,0.3003263772,0.207177937,0.3266133666,-0.2178706676,-0.0989935845,0.1522374451,0.1160596609,0.2051440924,-0.1475694627,0.0188564416,-0.0459570736,-0.0052340906,0.2867136598,0.1750594079,0.104387477,-0.3770065308,-0.0878642872,-0.0629248247,-0.0994721502,-0.2742574215,-0.0066032964,-0.098459594,-0.1801307648,0.7090316415,0.4778489769,0.2578676641,0.2521958053,0.0204008017,0.4376070797,0.1607919484,-0.1186862066,0.1933942288,-0.3835197389,-0.2017527521,0.4595516622,-0.1625296474,-0.0583283007,-0.1901446879,0.3574742675,0.1463835686,-0.3823429048,0.1063081995,-0.0545992181,-0.2116843313,-0.0239792485,-0.1792743951,-0.2611463666,-0.098667331,0.1246707961,-0.3771369457,-0.2479954511,0.7554060221,0.2194464803,0.1049712226,0.015877353,0.1960073411,0.2868534625,0.2450678051,-0.2382164896,0.2443639934,-0.2037657052,-0.047804568,-0.048289787,0.2154528648,0.0604507104,0.222289443,-0.0795786604,0.1138989404,0.0227032043,-0.0669767112,-0.0135789728,0.2710331976,0.1326074004,-0.1880733073,0.1035916135,0.0246736109,-0.1113354415,-0.0708182976,-0.1593857259,-0.0402798578,-0.1572140902,-0.017369248,0.1063375473,-0.3502475321,0.0729996562,0.1413803399,-0.2973563373,-0.0231424291,0.6617959142,-0.6750299931,-0.0622528717,0.0818928406,0.010699641,-0.2685936987,0.6212177873,0.2621107101,0.4459910095,-0.2740018964,-0.2401632965,-0.3499602675,0.1486311853,-0.3031346202,0.196233198,-0.1214342639,0.4490489662,0.1144374609,0.1519428343,0.1392837614,-0.3217310607,0.0912250578,0.1589361131,-0.4003050923,0.0434240215,0.1738899797,0.4088532329,0.0576848164,-0.4208478332,0.2522930801,-0.1683635861,-0.1276034862,0.0830053613,0.1131270975,0.1415845901,0.2653016448,0.4083060026,-0.0353764258,0.4078211784,0.2056368142,0.1028342545,0.0268668588,-0.0510273911,-0.0313524269,0.0229757261,0.081712231,0.0683192015,-0.4066522419,-0.0133742057,0.0460571609,-0.2658313215,-0.3916955292,-0.0877719298,0.0716312006,-0.3862326741,0.0770111233,-0.0453293063,0.2807236314,0.431735009,-0.086612381,-0.3715317845,-0.2835671902,0.0814543739,0.0554731861,-0.1319973767,-0.1057255045,-0.1213904694,0.2079126686,0.1666451544,0.1442809254,-0.2687599361,-0.086984396,0.3934926391,0.0091934102,-0.2492121905,0.1441196948,-0.0894979909,0.0986504406,-0.1405495405,0.3531612754,-0.1568920016,-0.0214888491,-0.4095470607,-0.2184598744]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"Looks like the google drive download failed.\r\nI'm getting a `Google Drive - Quota exceeded` error while looking at the downloaded file.\r\n\r\nWe should consider finding a better host than google drive for this dataset imo\r\nrelated : #873 #864 ","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":40,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nLooks like the google drive download failed.\r\nI'm getting a `Google Drive - Quota exceeded` error while looking at the downloaded file.\r\n\r\nWe should consider finding a better host than google drive for this dataset imo\r\nrelated : #873 #864 ","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"It is working now, thank you. \r\n\r\nShould I leave this issue open to address the Quota-exceeded error?","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":17,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nIt is working now, thank you. \r\n\r\nShould I leave this issue open to address the Quota-exceeded error?","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"Yes please. It's been happening several times, we definitely need to address it","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":13,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nYes please. It's been happening several times, we definitely need to address it","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"Any updates on this one? I'm facing a similar issue trying to add CelebA.","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":14,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nAny updates on this one? I'm facing a similar issue trying to add CelebA.","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"I've looked into it and couldn't find a solution. This looks like a Google Drive limitation..\r\nPlease try to use other hosts when possible","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":24,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nI've looked into it and couldn't find a solution. This looks like a Google Drive limitation..\r\nPlease try to use other hosts when possible","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"The original links are google drive links. Would it be feasible for HF to maintain their own servers for this? Also, I think the same issue must also exist with TFDS.","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":31,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nThe original links are google drive links. Would it be feasible for HF to maintain their own servers for this? Also, I think the same issue must also exist with TFDS.","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"It's possible to host data on our side but we should ask the authors. TFDS has the same issue and doesn't have a solution either afaik.\r\nOtherwise you can use the google drive link, but it it's not that convenient because of this quota issue.","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":45,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nIt's possible to host data on our side but we should ask the authors. TFDS has the same issue and doesn't have a solution either afaik.\r\nOtherwise you can use the google drive link, but it it's not that convenient because of this quota issue.","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"Okay. I imagine asking every author who shares their dataset on Google Drive will also be cumbersome.","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":17,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nOkay. I imagine asking every author who shares their dataset on Google Drive will also be cumbersome.","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"I am getting this error as well. Is there a fix?","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":11,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nI am getting this error as well. Is there a fix?","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/996","title":"NotADirectoryError while loading the CNN\/Dailymail dataset","comments":"Not as long as the data is stored on GG drive unfortunately.\r\nMaybe we can ask if there's a mirror ?\r\n\r\nHi @JafferWilson is there a download link to get cnn dailymail from another host than GG drive ?\r\n\r\nTo give you some context, this library provides tools to download and process datasets. For CNN DailyMail the data are downloaded from the link you provide on your github repository. Unfortunately because of GG drive quotas, many users are not able to load this dataset.","body":"\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'","comment_length":84,"text":"NotADirectoryError while loading the CNN\/Dailymail dataset\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\nNot as long as the data is stored on GG drive unfortunately.\r\nMaybe we can ask if there's a mirror ?\r\n\r\nHi @JafferWilson is there a download link to get cnn dailymail from another host than GG drive ?\r\n\r\nTo give you some context, this library provides tools to download and process datasets. For CNN DailyMail the data are downloaded from the link you provide on your github repository. Unfortunately because of GG drive quotas, many users are not able to load this dataset.","embeddings":[-0.1502866894,0.013875572,0.0027152719,0.2634353042,0.4774574339,0.0598452203,0.4901823997,0.3076288998,-0.2510559261,0.2567766011,-0.3245876431,-0.0494811945,-0.4837743342,-0.1979143322,0.0145921567,0.0012005472,-0.1768816113,0.1828713715,-0.1539244652,-0.0876947716,-0.1727184951,0.321044147,-0.0058411579,0.1479859501,-0.5002063513,-0.1975877583,-0.0039216476,0.5093349218,-0.1271757334,-0.2311133146,0.2591030598,-0.2564248741,0.0552018695,0.7316402197,-0.0001171105,0.3691133261,0.3056399524,-0.0785469413,-0.5518268347,-0.4986005127,0.2059125155,-0.2117751688,0.0436902605,-0.2601816654,-0.1276563704,0.0635365993,0.1162231341,0.314124018,0.3930333555,0.4601137042,0.2285237312,0.2302971035,0.2858363986,0.0685345754,-0.1296327561,0.3916880488,-0.1658728719,0.5677597523,-0.4966021776,0.1812898517,0.0389438532,0.305713594,-0.1901411414,-0.0401899554,0.0673795491,0.0520442091,0.2945782244,-0.4586854279,0.0862284899,0.1943815053,0.2952353358,-0.3140712976,-0.352039814,-0.0179468803,0.1240581274,-0.3361694813,0.2545648515,0.2885335088,-0.2310854793,0.0986495689,-0.1670768112,-0.1167311221,-0.3022927642,0.4389296472,-0.1841000915,0.2932372689,-0.1138128117,0.0412949026,0.1631835103,-0.1387998015,0.1639585048,-0.0754344985,0.3370934725,0.0137136988,0.1517517567,-0.1500609815,0.0298330132,-0.2463437021,-0.1883456111,0.1450426131,-0.2204875052,-0.3060177863,-0.0318602733,0.0074542542,-0.0541749112,0.0959956795,-0.1966276169,0.6701014638,0.2965373397,0.5027940869,0.032012783,-0.0403410904,-0.1909714937,-0.2045384943,-0.0870549977,-0.0306533463,0.6625394821,-0.0383561254,-0.3331961334,-0.0131454738,-0.4197606444,0.0783914253,0.3010311425,0.5060918927,0.0743204281,-0.0245896969,0.0031736968,0.1511423886,-0.0656605288,0.3835736215,-0.2604818046,0.3102358878,-0.1780570149,-0.0055621727,0.3036935627,-0.3783838153,0.2278976589,-0.0782359242,0.1879086792,-0.1120337173,0.0400563292,-0.1671946496,-0.0153000206,0.6389673948,-0.3143552542,0.122172676,0.2448890358,-0.0823115781,-0.2242325693,0.1375702471,-0.1204497367,-0.5571912527,0.0483228825,0.0999315456,0.1062155142,-0.0634725392,0.3532803357,-0.3217015862,0.0260888226,-0.1006018892,0.0767360628,-0.0643335655,-0.1164478585,-0.1373271048,0.3215051293,0.5865563154,-0.1676753461,-0.1125146374,-0.1753574014,-0.1381154209,0.2370170057,0.4046929777,-0.0672517568,0.2623784244,-0.283016324,-0.2667835057,0.3034884334,-0.3377234638,-0.4008953273,0.0308544394,-0.2147400379,0.2598990798,0.0527489483,0.1346762031,-0.3762584031,-0.0294961911,0.0617878698,0.1520974636,-0.0357308947,0.0904737264,-0.2340044081,-0.1187213138,0.087981984,0.3039717376,0.4382177889,0.0894154683,0.0313502997,0.2226996422,0.0503849052,-0.2450427115,-0.0719153062,0.3450480998,0.4285292625,-0.0693802163,-0.1211159676,0.1079738438,-0.3771114051,0.0220831204,-0.0822666511,0.3187367916,-0.2467507422,0.0032257261,-0.345186919,-0.1285056919,-0.2752381563,-0.1792377532,0.1339013726,0.1903796345,0.0196383093,0.2382143587,-0.214954406,0.2647375464,-0.1341636926,0.3000068367,-0.328188777,0.3138391674,-0.2469952852,-0.0395054705,-0.0923118144,0.4096641243,0.1486414522,-0.2167054117,-0.063395448,0.2275290042,-0.0740630478,-0.0576760061,0.0219573081,-0.0339813121,0.2657676935,-0.0904394239,0.1729007512,0.4206444323,0.0139150154,0.0184680186,-0.035002321,0.4468706846,0.0479335338,-0.0523554198,0.1870311946,-0.0533488803,0.0944574773,-0.0551234968,-0.0113500115,-0.0475617275,0.2278446257,-0.1467844099,0.5949158072,-0.0756230205,-0.2936546504,-0.498267144,-0.017299084,-0.2320438623,0.1143534184,0.2787705362,-0.2667388916,0.0841669738,-0.181010142,0.181104362,0.475499481,0.056991227,-0.0085043851,0.1831954718,-0.1056299657,-0.1482228041,0.3001134098,0.122813046,0.0768738464,0.2936602235,0.0724526346,0.0703333616,-0.5287972093,-0.3453274369,0.0061527174,0.3515328765,-0.3705409467,-0.1018935218,-0.1132583693,-0.2290444076,-0.031764362,-0.4700891078,-0.2261396796,-0.1386552751,0.0434038565,0.3248308003,0.0040019695,0.164022252,-0.02154289,-0.0796546414,0.1637674123,-0.1426153183,-0.225755021,0.1100877151,0.1099090129,0.0368247889,0.4638560116,-0.3609945476,0.245059073,-0.0506406166,-0.0929013863,-0.3026795685,-0.4553116262,-0.065082185,0.0196095668,0.263384372,0.1738888323,0.353010118,-0.1059392542,-0.3259163201,0.2694613934,0.1050129607,-0.0333887413,0.4007444382,-0.2799060941,0.0285295025,-0.030528551,-0.3856055737,-0.4965451062,-0.2660867572,0.0952375159,-0.0068388553,0.1661554724,0.6047119498,0.1591120362,0.1782280058,-0.2618115246,0.173198998,-0.0134630306,-0.1314901114,0.3829728365,-0.3403701782,-0.2306490541,-0.0014361457,-0.066587925,0.2421359271,-0.313904047,-0.5207077265,-0.2967836559,-0.0093218395,-0.2617911696,-0.0441968888,-0.1472934484,0.1218888909,-0.0831382871,-0.0681381524,0.2673321962,-0.1342415065,-0.1002488732,0.0882345513,0.0051644589,-0.2591240108,0.4475921392,-0.0936799422,0.7807009816,0.096185118,0.0188385956,0.4052089155,-0.016006697,0.4341351092,-0.2269287109,-0.4035912752,0.0421542041,0.3586930037,-0.092974782,0.1946215034,0.1108985469,-0.2174888998,-0.16596964,-0.2698647678,-0.4062983394,-0.1927652806,0.2044945657,0.0324250162,0.0445904918,0.0204262026,-0.0603807718,0.0829848573,-0.248153016,-0.066614829,0.4942092896,0.1149649471,0.0863041505,0.0322949104,-0.3037804663,-0.3901974559,0.0338840224,0.0017105788,0.0812265202,-0.2152146697,0.0576918572,0.1245850176,0.0010683795,0.3473902643,0.0930749029,-0.1501548737,0.3155097067,-0.3130577505,-0.5068178773,-0.147001937,-0.1153768152,0.1861529648,0.0638440028,0.4790127873,-0.3008008301,-0.1790096909,0.5184316039,0.0287056044,-0.073257491,-0.2285499424,-0.3372282088,-0.387275666,-0.4333268404,0.112052761,-0.2079117596,0.1160897389,0.0733146891,-0.1698256284,0.138342455,0.1019176245,0.1633889973,-0.0997598544,0.2785457671,-0.132990092,0.1996034086,0.2281342149,0.2923973799,0.3537794352,0.3676193058,-0.1571993828,-0.1176634282,-0.0099630849,0.133553192,-0.0225231014,0.4273509681,-0.1563072652,-0.3167841136,0.0756197795,0.3467943072,0.1500735879,-0.1691112369,0.4981242716,-0.0588797182,-0.4221140444,-0.0062629022,0.3730384707,-0.1417728662,0.2072502673,0.0918784067,0.0032515556,-0.0316046476,0.0829576477,-0.2496451437,0.9348476529,0.0290452335,0.2436060309,0.5999217629,0.2633118331,0.2116592228,-0.3162898719,0.1946512759,-0.3441984951,-0.1643297076,-0.0195441842,-0.1975620985,0.0800100043,0.2080590129,-0.1675433218,0.350014478,-0.0501817949,-0.0432965085,-0.2124762684,0.2127786279,-0.1947493404,0.0912313461,-0.2391412854,0.1060208455,0.0028010686,0.4187838137,-0.3177286983,-0.0307762865,-0.1151658595,-0.2349917889,-0.0425483026,0.0143389432,-0.369941622,0.2966138721,0.2785767019,-0.2126495391,-0.3221946359,0.1412672848,0.1212881133,-0.190947637,-0.1534513682,0.0655241609,-0.2706925571,-0.1494620293,-0.1236636415,0.0520416647,0.3858026564,-0.2016162276,-0.092235595,-0.069834888,-0.1452213675,-0.2897571623,0.2210944146,-0.0591005422,0.1999624968,-0.2618899643,-0.2983102798,-0.0748918653,-0.2779407203,-0.1450492889,0.1157986,-0.0094909621,-0.1632428765,-0.0443025976,-0.0702364817,-0.3143126965,-0.0948319957,0.1749742925,-0.1100249588,-0.061841324,0.646193862,0.0266936645,-0.3503605723,-0.1812146306,0.1435990036,-0.1017236039,-0.4620892107,0.2607919872,0.1379673779,0.190510124,0.0245131198,0.0751047954,0.2225316018,-0.0063559967,-0.2272480577,-0.4220342636,-0.1887057722,0.2030885369,0.1064955443,0.3375370204,-0.0991909504,0.1013700143,0.2454857677,-0.1607853174,-0.2722640634,0.1723113209,-0.1976251453,0.033101812,0.1233262718,-0.1212876067,0.6023900509,0.1528308392,0.1659272611,-0.0086321291,-0.1814550906,-0.2331738472,-0.0539257228,0.1537873447,0.2020715922,-0.140008375,0.1849014014,0.1343211383,-0.1158257052,-0.0235448536,0.0312279761,0.1810223609,-0.0009693113,-0.1373784244,0.0374226831,-0.1362691224,-0.051865425,0.0387653895,-0.0665324777,0.2540249825,0.1684999764,0.1101963297,-0.142462194,-0.1031715646,-0.2843037844,-0.0124077136,0.0225078557,-0.0433380008,0.3292863965,-0.2255821228,-0.1474162936,-0.0153064998,0.3085732758,0.0639067292,-0.3083703518,0.0931777507,0.1740079224,0.1565859318,-0.2176671773,-0.0170727279,0.226440087,-0.0404809415,0.0061321151,0.1918838769,-0.1909846514,0.0592048317,0.1142963022,0.1501492262,0.2236474752,-0.2713264823,-0.0109922178,0.4282627106,-0.1666841656,0.0820700601,0.3861884475,-0.0831831023,0.2314912528,0.4109051228,-0.3030012846,0.1395142078,-0.3390327692,0.0567451306,0.1770600975,-0.3423680961,0.0567750521,0.2707987428,-0.0504284427,0.0627737567,-0.1856461763,0.331595391,-0.1513169855,0.1580837667,-0.3751071692,0.4998324215,-0.2893129289,0.1727905422,0.0603434369,-0.182962954,-0.1121396869,0.082131438,-0.0408675708,-0.14884305,0.1253468543,0.4593407512,-0.221597895,-0.0026675374,0.2491135001,0.2356093526,-0.0787999704,-0.208149001,0.2621292174,0.5025692582,-0.0162126608,-0.12249659,0.654966116,0.5555667281,0.2684943378,0.0158881377,0.3025692403,0.0192761216,-0.1680324823,0.102583304,0.1173413694,0.1646898687,0.4219744802,0.1354040354,0.1444707215,-0.2612223327,0.0850301012,0.064402014,0.1566947401,-0.0751107037,0.5024009347,-0.2946595848,-0.2122948766,-0.3024602532,-0.1719893813,-0.4507956505,0.2477429956,0.3150810897,-0.0316826142,0.0680089146,-0.0747672245,0.0540815406,-0.0236215144,0.151729539,0.3406074345,-0.0009296077,-0.5548706651,-0.0407999456,-0.5642197132,0.3048743606,-0.2687525749,0.1406704783,-0.0807122141,-0.0204128381,-0.2169798613,0.1230339035,0.2206643522,0.0039660176,-0.1187179312,0.0199877247,-0.2343534678,0.1120592207,0.1332816482,-0.1299860477,0.0518985204,-0.3918950558,0.0636957139,-0.3433345556,-0.0343912393,0.1496935785,-0.011627919,-0.2407974899,0.3184995651,0.2625727355,-0.0410652235,0.4454624057,-0.1143084764,-0.2598309517,-0.0583604611,-0.2245173007,-0.2071983963,0.0214923564,0.1680610776,0.6819016933,-0.1619597673,-0.2586810887,-0.3634815812,0.432546258,-0.1076610684,-0.2463701218,-0.066547811,0.1007145196,-0.1359902173,0.1650627255,0.1683791727,0.1788600981,-0.0089131333,0.1097738296,-0.4074542522,-0.4818341732,0.6003141999,-0.2737822831,-0.096901536,-0.0200534146,0.1188859493,-0.0998277366,0.0116760274,-0.3960606158,0.3837617338,0.2735113204,0.1435285211,-0.1272129565,0.0563923828,0.1020406485,0.000462132,0.071871601,0.3017484546,0.177706942,-0.3659015894,0.0227029789,-0.2111171633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/993","title":"Problem downloading amazon_reviews_multi","comments":"Hi @hfawaz ! This is working fine for me. Is it a repeated occurence? Have you tried from the latest verion?","body":"Thanks for adding the dataset. \r\nAfter trying to load the dataset, I am getting the following error: \r\n`ConnectionError: Couldn't reach https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_fr_train.json\r\n`\r\nI used the following code to load the dataset: \r\n`load_dataset(\r\n dataset_name,\r\n \"all_languages\",\r\n cache_dir=\".data\"\r\n )`\r\n\r\nI am using version 1.1.3 of `datasets`\r\n\r\nNote that I can perform a successfull `wget https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_fr_train.json`","comment_length":21,"text":"Problem downloading amazon_reviews_multi\nThanks for adding the dataset. \r\nAfter trying to load the dataset, I am getting the following error: \r\n`ConnectionError: Couldn't reach https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_fr_train.json\r\n`\r\nI used the following code to load the dataset: \r\n`load_dataset(\r\n dataset_name,\r\n \"all_languages\",\r\n cache_dir=\".data\"\r\n )`\r\n\r\nI am using version 1.1.3 of `datasets`\r\n\r\nNote that I can perform a successfull `wget https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_fr_train.json`\nHi @hfawaz ! This is working fine for me. Is it a repeated occurence? Have you tried from the latest verion?","embeddings":[-0.3760583997,-0.0687171593,-0.1512161046,0.3995557427,0.2218290865,0.0871828422,0.2271226645,0.0012161636,-0.195425272,-0.1700538546,-0.1386815906,0.0604705513,0.1707975268,0.0332740657,0.0018724314,0.0242816843,0.0877761245,-0.1517885327,-0.085342288,0.0247635096,-0.1652069837,0.1678636968,-0.1811462194,-0.0880342796,-0.2276556343,-0.0458784476,-0.063408047,-0.0876697153,-0.3202602863,-0.3831088543,0.4728878736,0.2383781224,0.2648138702,0.3811822236,-0.0001174623,-0.1535703838,0.3675583899,-0.1078284085,-0.2808185518,-0.6039906144,-0.3533784449,-0.2391920239,0.2802509964,0.0287721436,0.0832013786,-0.0687334239,-0.0326403342,-0.2118874639,0.1221739128,0.2553088963,0.1851304471,0.1339330375,0.4523338675,-0.0434270166,0.2750202417,-0.1569391638,-0.0335237533,0.3968005478,0.0909879431,-0.0359309576,0.2279304117,0.1416745931,0.0797833428,0.3172475994,0.1299233735,-0.0447500683,0.2263744026,-0.1134047061,0.091087617,0.422377795,1.0506269932,-0.10850247,-0.5078006387,0.2045578659,0.158776626,-0.0387851559,0.1576090306,0.2874030173,-0.1353876442,-0.1331919432,-0.429851532,-0.3206474781,-0.288641721,0.5616419911,-0.0451895967,0.3056504726,-0.0740531236,0.2216133922,0.3126436174,-0.0657328293,0.5485727787,-0.0852166191,-0.0937298164,0.2087666541,-0.8067884445,0.0056742691,-0.2236727774,-0.1762273163,0.0434960127,0.312823087,0.0840657279,0.0621921755,-0.028283149,0.2298112959,0.4412487447,-0.1112852916,-0.0073343068,0.300162673,0.0472344235,0.4499185979,0.1058038995,-0.1061760634,-0.2239825875,-0.213673383,0.0520493872,0.0940273702,0.0255822558,-0.2895092666,-0.3205344081,0.0411056243,-0.5326039195,-0.1666151434,0.0506155863,0.2480783015,-0.2394607812,0.2531683743,-0.0434296466,0.3617746234,-0.0829924047,0.0746374875,-0.1123440489,0.0865691081,-0.177681908,-0.0614344738,0.2400965989,0.1313207597,0.2572076917,-0.1772686243,0.0769190118,-0.3000606894,0.1120905504,-0.129009515,-0.2409229726,0.2902305126,0.3096740544,0.1613355726,0.0355424248,0.0597349256,0.056611605,0.0393190756,-0.0957612991,-0.1534672529,-0.1657355577,0.1485721022,-0.2256349325,-0.2716696858,-0.2497368306,-0.0944566056,0.0449420847,-0.1431613266,-0.1663995385,-0.0688501447,0.1288469732,-0.0877747387,0.1445049942,0.3304733336,-0.3380998671,0.0293947477,-0.4909790754,-0.1468683332,0.0422965735,0.3341827691,-0.3259694874,-0.0363804698,-0.1774045676,-0.0283069517,0.3972617686,-0.2358672768,-0.391702652,0.5111751556,-0.2891079187,-0.1156889498,-0.0255055241,0.2173869014,-0.114275001,0.1892291307,-0.0317801088,0.2984339297,-0.0319019072,-0.1459318846,-0.0540253595,-0.2005337924,0.2504793704,0.3099494576,0.1478506327,-0.0288840756,0.198020786,0.1882623732,0.3265378177,0.2143276632,0.0855998024,0.3521366715,0.353977561,0.0481052995,0.0110343955,-0.0713915974,-0.5713795424,0.0172049403,-0.0972928256,-0.1698230058,-0.1361767948,-0.0581827387,-0.2476274967,-0.0544587187,-0.2091049403,0.1748756021,0.0531651042,0.1519445628,-0.0390002467,-0.0467892438,-0.0379904509,0.1853621453,0.0951345563,0.1475638151,-0.2620975375,0.2204957902,-0.1911636591,0.1527510732,0.275311172,-0.1027089506,0.2127271593,-0.1334378719,-0.0645937249,0.3175309002,0.0101136528,0.1606999934,0.2440329045,0.1258067489,-0.0531193279,-0.1045050919,0.0152933765,0.259087503,0.2428422123,-0.099616766,-0.2561759651,0.1772407889,0.1053444445,0.2681096196,0.0212707371,0.0990921259,0.2801519632,0.0281583741,-0.0041362522,-0.0167243909,0.3990073502,-0.0466886424,-0.1435838342,-0.3875207901,-0.0578988157,0.009158507,0.1139790118,-0.0975529924,0.1503444016,0.109780401,0.0100891581,-0.2266483903,-0.0137330126,0.4280416369,0.4721235931,0.0101010324,0.215177834,0.0595326573,0.0701398328,-0.1609264612,0.1694002599,0.0959650576,0.3051463366,0.3443479538,-0.0558003485,-0.0531795509,-0.0502258539,-0.1583517045,0.3277203441,0.2827937603,-0.250895381,0.1228301823,-0.482427001,-0.0901075229,-0.2425749749,-0.1869426072,-0.3196102977,-0.3992118537,-0.2374681234,0.3306756318,0.1338935047,-0.0416912399,0.0084600495,-0.0954337195,0.1509998888,-0.037251085,0.0951510444,-0.2162163556,-0.2899645567,0.1005893871,0.2935492992,-0.1091830805,0.1274592131,0.0831154659,0.1468355358,-0.3079131246,-0.1269351244,-0.1508038789,0.1781183779,0.1312027425,0.231682241,0.3100517392,-0.1224841774,0.0356893577,0.0782066062,0.1118622348,0.0035301854,0.125541091,-0.0680983886,-0.175469175,0.2824950218,-0.5468952656,-0.1960801929,-0.1828793883,0.1689725816,-0.2596386969,0.019116411,-0.4372482598,0.0934619159,0.1319027841,-0.1245440021,0.0840198025,-0.2682675123,-0.2069098651,0.2822828293,-0.0779304877,-0.3567316234,0.3560471535,0.2730462849,0.2037633955,0.1154053062,-0.7272118926,0.0916247591,0.2344181687,-0.2698007822,0.1553618312,0.0154137164,0.1727865785,-0.2322534621,-0.0533687733,0.1976912171,-0.0629869029,-0.160548225,0.1174262986,0.7222898006,-0.1894682646,0.2288447469,0.1139828637,0.3751296699,0.2665195763,-0.0288639143,0.2219634205,0.1663761586,0.1095670313,0.2133678049,-0.2673486769,0.0439313017,-0.345387876,0.2778543234,-0.0118747773,-0.1510619521,-0.5303330421,-0.3466280997,0.0501109064,-0.2043256611,-0.1051555946,0.0215278808,-0.4200635552,0.0786083937,0.1402589083,-0.1143857017,0.2780182362,-0.2665728033,-0.1785021871,0.3525506258,-0.0871999934,0.0359369814,-0.4260105789,-0.1308297515,-0.0104482342,0.356335789,-0.1179320589,0.7267031074,-0.1333709806,0.2489183694,0.1951018274,-0.23011823,0.7335609794,-0.586974442,-0.0048653381,0.0623429343,0.0915310308,-0.4636372328,-0.1752259582,-0.2895460129,0.4296636581,0.1251025498,0.2294652164,-0.29459849,0.0513652153,0.1915203631,0.0596601106,-0.1192008331,-0.4214910865,-0.2599736154,-0.1756554842,-0.3072964847,-0.0393195748,-0.2951026559,0.3925751746,-0.0786137059,-0.0391157642,0.1079352647,0.0629970282,-0.1717078239,-0.1695131361,-0.1070077047,0.0813384578,0.1733444035,0.1802344918,0.011370209,0.3162175417,0.7589716315,0.0694131181,-0.5596141815,0.259096384,-0.1655554473,0.1200817004,0.1197094545,0.1849684864,0.1255122572,0.3726021647,0.0107749859,0.2102598101,-0.0218314268,0.3853048086,0.0338166393,-0.1196129769,-0.3037387729,0.2869792879,-0.2414189577,0.0476065204,0.0115100238,-0.2778789699,0.0368820578,-0.1208551452,0.1104170606,0.9052498341,0.113566421,0.0996834114,-0.1198924929,-0.2967720926,0.2970975637,-0.5791083574,0.043830093,-0.3369165361,0.0016262984,-0.1806020439,-0.1437612176,0.1378137916,0.3396259844,0.0060512712,0.3256645501,-0.0057111164,0.043152079,0.1399622411,0.3301214278,-0.1041151583,-0.0245807711,0.0014010461,0.1184773371,-0.2240977138,0.2851725221,-0.201671198,0.0011586754,0.0076762037,-0.2724421024,-0.5300316811,0.102153115,-0.2045398206,0.3161271513,-0.0062511717,-0.3018481731,0.184987694,0.2600186765,-0.0908028856,0.1287351102,-0.216505304,0.1222099736,-0.0555148497,-0.2705999315,-0.0241379905,-0.2159351259,0.0079906769,-0.1819922477,-0.597163856,0.2154240012,0.192094326,-0.17895253,0.1722553223,0.1359446794,-0.3334887624,-0.1371559203,-0.1769361049,0.1274012327,0.0828850567,-0.0977430791,0.08534275,0.2360744923,-0.3618901372,0.16973643,0.2105908692,0.0492083393,-0.0643649176,0.8636333942,0.0286202896,-0.2775164247,0.1623734832,0.4344795048,-0.252655983,-0.3645714223,0.1376476437,0.1207159758,-0.4923096597,-0.1717735529,-0.1336128414,0.295249939,0.0356935039,-0.0159089901,0.2574012876,0.1399097443,0.1712179482,-0.7608023882,0.2117218524,0.1154407412,-0.0505735502,0.1819092333,-0.1716828048,-0.4116681516,0.2183818668,0.2390035838,-0.2321224362,0.147290498,-0.4260753095,0.0291892849,0.0129770003,-0.181736663,0.1943996251,0.0441685542,0.074447386,-0.0411591344,-0.2427626997,-0.1290091276,0.0303493459,0.1401367337,0.1493046284,-0.1141726002,0.045432996,-0.350435853,-0.2139592469,0.307747215,0.1034890264,-0.1074509844,-0.004476191,0.1552470773,0.0591496266,-0.1350941509,-0.3512435555,0.0180092212,-0.0578512438,0.1934486628,0.1699364632,0.1226345524,0.4204266071,0.0044744192,-0.2742658854,-0.0970615372,-0.0810326412,0.0194173232,0.5168263316,-0.3800700307,0.1693310142,-0.0673338622,0.4852585495,0.5259001255,-0.2682291567,0.0795675442,0.3408383727,0.1828825772,-0.3047328889,-0.4977709353,-0.0220139362,0.1146616489,-0.1691058278,-0.0814406425,0.0267349184,-0.1408555359,0.2197670937,0.1466979384,0.3943866789,-0.100010328,0.2413313985,0.1268939972,-0.225724861,0.1006753519,0.2271491289,0.1426261067,-0.1209111512,0.2855171561,-0.0766798407,0.0614101738,0.0915947706,0.3873986006,-0.0108579323,-0.3694122732,0.1577726752,0.1090785712,-0.1153824627,0.0297025871,-0.045318272,0.2133299708,-0.1024141088,0.1619334519,-0.559427917,0.0122129712,-0.0585211292,-0.1331551224,-0.2422489226,-0.2024041414,-0.276275456,0.0247438978,-0.0607907958,-0.2763905823,0.1711342633,0.2096412629,0.0516682379,-0.594412744,-0.0524073988,0.0873768404,-0.0601974428,-0.3464761078,0.1010099575,0.3304134607,0.0336795449,0.0459678248,0.0191223603,0.3294522762,0.1732229888,-0.1279518753,0.1771547049,-0.0083385166,-0.0767915547,0.1825032532,0.2080483884,0.5027759075,-0.1009993702,0.4242614806,0.0497200713,-0.0788146332,0.7516396642,0.1391801685,0.2534099221,-0.1906704307,0.2152339369,-0.2270902991,0.0391906388,-0.3854593933,0.1672207117,-0.6156358123,-0.0862334669,0.3548777997,-0.1562501788,0.0931014717,-0.2295362055,0.0301775746,-0.0798113346,0.8588131666,0.2599476278,0.1833547354,-0.2782604396,-0.0734047741,-0.4318870008,0.139236182,-0.2241561413,0.1223364621,0.0839166865,0.3067053556,-0.1299304813,0.4181704521,0.0848797038,0.0986853093,-0.0958618224,0.6857907772,-0.2925002277,-0.264313221,0.2988100648,0.2186221033,-0.0371114649,-0.1455605626,0.1634250432,-0.1592045724,-0.094672516,-0.2772415876,-0.2932485044,0.0870182812,0.0460379496,0.1496989876,-0.0614807568,0.6655396223,0.0549403802,-0.1532881409,-0.339235127,-0.1613489091,-0.0591119602,0.2846425176,0.358450532,0.4268785715,-0.1081949323,0.1689175516,-0.1911164373,-0.0041979835,-0.1685983986,-0.0218932852,0.0049289335,-0.1939948052,0.1501346231,0.0093858531,0.1331478506,-0.0641481504,0.028533183,-0.0743602067,-0.1731964797,-0.1396244615,0.2915416062,0.1399747729,-0.0243466329,-0.1901143938,0.232648775,-0.1780021936,0.012232339,-0.4893919826,0.3814643919,0.4427547157,0.0313323289,-0.2681514919,0.2902330756,-0.0731367767,-0.0090241292,-0.0591967218,0.2034932226,-0.0358704999,-0.2879573107,-0.1622181982,-0.2576028705]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/993","title":"Problem downloading amazon_reviews_multi","comments":"Hi, it seems a connection problem. \r\nNow it says: \r\n`ConnectionError: Couldn't reach https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_ja_train.json`","body":"Thanks for adding the dataset. \r\nAfter trying to load the dataset, I am getting the following error: \r\n`ConnectionError: Couldn't reach https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_fr_train.json\r\n`\r\nI used the following code to load the dataset: \r\n`load_dataset(\r\n dataset_name,\r\n \"all_languages\",\r\n cache_dir=\".data\"\r\n )`\r\n\r\nI am using version 1.1.3 of `datasets`\r\n\r\nNote that I can perform a successfull `wget https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_fr_train.json`","comment_length":13,"text":"Problem downloading amazon_reviews_multi\nThanks for adding the dataset. \r\nAfter trying to load the dataset, I am getting the following error: \r\n`ConnectionError: Couldn't reach https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_fr_train.json\r\n`\r\nI used the following code to load the dataset: \r\n`load_dataset(\r\n dataset_name,\r\n \"all_languages\",\r\n cache_dir=\".data\"\r\n )`\r\n\r\nI am using version 1.1.3 of `datasets`\r\n\r\nNote that I can perform a successfull `wget https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_fr_train.json`\nHi, it seems a connection problem. \r\nNow it says: \r\n`ConnectionError: Couldn't reach https:\/\/amazon-reviews-ml.s3-us-west-2.amazonaws.com\/json\/train\/dataset_ja_train.json`","embeddings":[-0.440661639,-0.0421525501,-0.180196628,0.3614037335,0.1892244667,0.1924448162,0.2124000043,-0.0044507785,-0.1972015202,-0.1464764327,-0.1976143867,0.0377508067,0.1531439275,0.0971539244,-0.0720924139,0.0213294346,0.0598915182,-0.2066427618,-0.0130367987,0.1234657615,-0.1873295903,0.2317028493,-0.1700606793,-0.0764305145,-0.2450657785,-0.0578406788,-0.0536769107,-0.150131315,-0.3790063262,-0.393214941,0.4189879894,0.2610453069,0.2405183315,0.3830579519,-0.0001122186,-0.1504674107,0.3549053669,-0.0775949806,-0.233854413,-0.5858892798,-0.4332997203,-0.2583684325,0.2224095911,0.0037304533,0.0805023238,-0.0729167685,0.008034288,-0.1661413908,0.1844665259,0.2972430289,0.2394647449,0.0478377007,0.4759506881,-0.0099651869,0.1992232949,-0.1317870319,-0.0510212108,0.4177095294,0.0474772975,-0.0311333574,0.1777214706,0.1282536983,0.085675396,0.2983752787,0.0398989506,-0.1000356898,0.1395142376,-0.1488240957,0.1885825694,0.380489856,1.0128480196,-0.1324625909,-0.4798005223,0.1949474216,0.1820300221,0.050821092,0.1354616284,0.2535705268,-0.1403279901,-0.0082462719,-0.4710298479,-0.2919905782,-0.2809957266,0.598226428,-0.0767178386,0.3625804186,-0.0946455449,0.1868107319,0.2157939076,-0.0573819056,0.5269710422,-0.0770130605,-0.0526941456,0.1407216787,-0.6389880776,0.04572488,-0.1645273119,-0.060463842,0.0262737554,0.214213565,0.0612248108,0.1116949022,0.0707698315,0.2164036334,0.4166511297,-0.0915412009,0.0062137586,0.225363344,0.0026506274,0.3906871974,0.1476271003,-0.1223669276,-0.1879302412,-0.1665714085,0.025643073,0.0482621007,0.0512857661,-0.2647967041,-0.307893455,0.0288290344,-0.4337103367,-0.2023431063,0.0653539076,0.2858324945,-0.2444898337,0.3465029001,-0.020939488,0.3661306798,-0.0472360998,0.0387246758,-0.1676917076,0.1135276705,-0.1621274501,-0.0965998843,0.2482125461,0.2540058494,0.3333954513,-0.2126161903,0.0831840709,-0.2670731246,0.1419405788,-0.1239495501,-0.2174868435,0.2310986072,0.3497374058,0.2209533751,0.0109049128,0.051960282,0.0698629469,0.0246192776,-0.1975236237,-0.1149197221,-0.2183882743,0.209666431,-0.3319366574,-0.2344602942,-0.377127409,-0.0420396514,0.0887420699,-0.1345764399,-0.1252012253,-0.0961787775,0.0981521159,-0.1047096252,0.1536551714,0.3281702995,-0.2963987887,0.0085816495,-0.4491500854,-0.2107775807,0.0513828695,0.215740785,-0.3005242944,0.0639288947,-0.1748476177,0.0596091822,0.4675537944,-0.2087474018,-0.438670814,0.5093540549,-0.3005396724,-0.1079659238,-0.0227083582,0.1568250954,-0.1246630624,0.2524941564,0.0310954992,0.2984033525,-0.036490608,-0.0798618495,-0.0422002636,-0.2047291696,0.1852912754,0.2859449387,0.11120186,-0.0302403439,0.160651505,0.2090693563,0.3353353441,0.2228563726,0.069008939,0.3957965672,0.1918061227,0.0089879381,-0.0508994162,-0.0973049477,-0.5423532128,0.031369213,-0.107634671,-0.1157220751,-0.1929374635,-0.0771879852,-0.2699021995,-0.0584591627,-0.204268232,0.2394461334,0.1260460913,0.0920762047,-0.0327238664,-0.0365516245,-0.0126861325,0.1411378235,0.1106335297,0.0928663984,-0.3241458535,0.1924648285,-0.2239928097,0.1246941984,0.315915823,-0.0857346505,0.1917408258,-0.1509892493,-0.0828753337,0.3241711855,-0.0536321849,0.2173619121,0.172751084,0.1581158489,-0.0410824418,-0.0852235332,0.0306781065,0.2349263281,0.1823924333,-0.1747746021,-0.21346654,0.2787787616,0.1135569066,0.20592843,-0.0065109967,0.0992243737,0.3240623176,0.0037151373,0.0013112015,-0.0323501118,0.3470447958,-0.0835113823,-0.1462997943,-0.3628333211,0.0090043265,0.0800748393,0.1562293172,-0.0644564182,0.0856222659,0.0756700039,0.029052129,-0.2701375782,0.0428286865,0.4612466693,0.5124764442,0.0916243419,0.24558492,0.0493032858,0.1054602712,-0.1978019178,0.1595284194,0.1313164681,0.2810057998,0.2644899786,-0.03918292,-0.0773394555,-0.0168493036,-0.1058157235,0.2511737645,0.3593254387,-0.3429761529,0.1199733913,-0.4734178782,-0.2288341373,-0.1947231889,-0.1906579584,-0.2702898681,-0.424895823,-0.1889460683,0.2956547439,0.1207586154,-0.0979424864,-0.0713549852,-0.1496239454,0.1122409031,-0.0545912236,0.0744970217,-0.182371363,-0.2936092317,0.1620462686,0.4148870111,-0.0376440883,0.1876989007,-0.004967893,0.0769256279,-0.1776151508,-0.067433171,-0.2267477065,0.1485593021,0.0659042969,0.2173073739,0.356574595,-0.1446691602,-0.0706683174,0.082984291,0.0692675933,-0.0513451472,0.0553911775,-0.0382881947,-0.2145678103,0.2567067742,-0.5830492973,-0.1825186163,-0.2220588773,0.217844516,-0.1455401182,0.0136734452,-0.5214996338,0.1326012015,0.1945129335,-0.0121366885,0.1300954372,-0.2675257921,-0.2920618355,0.3096813858,-0.0955046192,-0.3806818724,0.4077480435,0.2706608176,0.2381499708,0.0205425508,-0.6656979918,-0.0103041334,0.2116973102,-0.1562308967,0.1493950039,-0.017417239,0.213972047,-0.2093674988,-0.0809690058,0.2110592127,-0.0527792275,-0.1948176771,0.0816424713,0.7380830646,-0.2327980846,0.2476056814,0.0968621895,0.3843829334,0.1734437495,0.0010284205,0.1565515101,0.1998076737,0.090055272,0.1668444574,-0.2576676309,0.0527835637,-0.2984928191,0.2946266234,0.0577165335,-0.0300939716,-0.5020580292,-0.4192767739,0.0480014309,-0.311329037,-0.1121433303,0.0040116361,-0.364557445,0.0232912563,0.1450235844,-0.1861004382,0.2460611761,-0.2906420529,-0.1679451317,0.3306908011,-0.1795937121,0.0334123112,-0.4635703862,-0.1423354447,-0.1290426999,0.3790090084,-0.0531536229,0.6651069522,-0.155368194,0.2367214561,0.200546205,-0.2115322202,0.7264887691,-0.541878283,-0.0214732941,0.0957005024,0.1831038296,-0.3423242569,-0.1450091153,-0.2911923826,0.4786806107,0.2924930751,0.2267177552,-0.2945902646,0.0048763934,0.2121051252,0.0867203847,-0.1556700468,-0.3682292998,-0.231018886,-0.2736576498,-0.2248019576,-0.1151574701,-0.1824531853,0.3894383013,-0.0502959229,-0.0345846042,0.1309304684,0.0769985914,-0.1521398574,-0.1787942648,-0.1403599679,0.1406129152,0.1131817624,0.1368540525,0.0539150462,0.3864408135,0.708503902,0.0870889127,-0.4951971769,0.2215723395,-0.1156115234,0.0573640615,0.0968167186,0.1661998183,0.1419053227,0.3887968063,0.0128175076,0.2069472522,0.0697037131,0.3587079048,0.0553704612,-0.1028324664,-0.301486522,0.3484570384,-0.2232583761,0.0682591945,0.0234327409,-0.2103714049,0.0239493083,-0.1141626388,0.1050273106,0.9148750305,0.0675369352,0.0199703574,-0.0970879346,-0.34305951,0.2762137055,-0.5420379043,-0.008473915,-0.2567883134,0.0612188987,-0.1753014773,-0.0596881695,0.1409989148,0.3261041343,0.0277027208,0.3113138974,0.0301635992,0.0728225932,0.1781526655,0.3841431439,-0.0706093237,-0.0113270935,0.0018535531,0.1735106558,-0.2027315199,0.1869496256,-0.2098205388,-0.0693471208,-0.0541672893,-0.2305452526,-0.5194786191,0.0761084706,-0.2706916034,0.340960592,-0.0570873432,-0.3308576345,0.2252188623,0.2856422961,-0.108079128,0.1196203902,-0.293009907,0.1215807274,-0.0837983266,-0.2735330462,-0.0026457086,-0.184179917,0.0371953249,-0.2080657035,-0.4792585671,0.2090539783,0.2257266045,-0.1741741002,0.2302820534,0.1341575682,-0.2589701414,-0.2140360922,-0.2679061592,0.0659208223,0.0313607678,-0.1049347445,0.128747046,0.2471747994,-0.3634593189,0.1690279245,0.2204640061,0.0208649021,-0.0899758339,0.9271849394,0.0344196595,-0.3576313257,0.1683180183,0.3516970575,-0.259308815,-0.3425430954,0.1567542106,0.0722377673,-0.386800915,-0.2213102728,-0.0701678321,0.296395421,-0.039074447,-0.1088572517,0.2443478853,0.2028720081,0.1838430315,-0.6791090965,0.1111969575,0.0931189135,-0.0280282628,0.1557468176,-0.1634495556,-0.3308768272,0.2308112532,0.2193477899,-0.287686497,0.0773958042,-0.4478084743,-0.0295320842,0.0317175835,-0.1673192084,0.1748370826,0.0173653457,0.1306153983,-0.0881284252,-0.2234139591,-0.1853520125,0.0953733847,0.1183839366,0.1552373022,-0.1360722184,0.0738608018,-0.2899034917,-0.1657214761,0.3333759308,-0.0276956875,-0.0425839685,-0.0225178357,0.1189133078,0.0420490615,-0.1744041145,-0.3100430369,0.0065769916,-0.0623914152,0.1786175072,0.1024987921,0.1827684045,0.4477968812,-0.0188222602,-0.2513489425,-0.1504745781,-0.1484369338,0.0124785975,0.4902305305,-0.3221192956,0.0973981917,0.0573829822,0.4473293424,0.4150266349,-0.2487100065,0.1160178706,0.3552432358,0.209967792,-0.2582303882,-0.3832267523,0.0482805073,0.0442013964,-0.1552620679,-0.1012749076,0.0435200632,-0.103511326,0.0928539559,0.1150112376,0.3403528631,0.016099656,0.2637943923,0.1711255163,-0.2129700184,0.0858329311,0.2724956274,0.1432519555,-0.1315772533,0.1534402519,-0.0762163699,0.1418017745,0.0111558447,0.3800500929,-0.1564682871,-0.3494597375,0.121199213,0.0242581628,-0.1499291062,0.0565179475,-0.0298024528,0.2573842406,-0.1049557626,0.1147749797,-0.4929705262,-0.0354595669,-0.0713299364,-0.1276001483,-0.2454233468,-0.2206134051,-0.2700885832,-0.0187100377,-0.0447759256,-0.3287270665,0.1647709161,0.1723304689,0.0446299314,-0.5132949352,-0.0735873282,0.0151590817,-0.0396166146,-0.3599068522,0.0723802075,0.2594974041,0.0585508496,0.023302326,0.0636955649,0.3877937794,0.1218707338,-0.0884348229,0.1476139426,0.0031081853,-0.1141007915,0.1398432106,0.239806518,0.524017334,-0.1653729081,0.413626492,0.1157336608,-0.1157828793,0.7268936038,0.1059128121,0.2117364556,-0.1229654551,0.2029301822,-0.3015274107,-0.0705669969,-0.3871929049,0.1213930994,-0.5511074066,-0.1570287645,0.2852011621,-0.2201046497,0.1032995656,-0.2525820136,0.0656626746,-0.0609341487,0.8566226959,0.2983169258,0.2468223423,-0.2307704538,-0.1300488114,-0.4464171529,0.0793334097,-0.1826803833,0.1642921865,0.022689335,0.2820459008,-0.1744336039,0.4298175573,0.0923396051,0.1412178576,-0.0722104535,0.6773055196,-0.2939931452,-0.2450264692,0.2367917746,0.2300055772,-0.0410017222,-0.0544590466,0.1734462082,-0.1863701791,-0.048238758,-0.3316556215,-0.232077837,0.0088832108,0.0349620283,0.1082888395,-0.0255939979,0.6409857869,0.0002662457,-0.0239583533,-0.2579888701,-0.1286123842,-0.0783887208,0.3046498597,0.3044913113,0.4460861683,-0.0482210107,0.1640740335,-0.1222346202,0.0108422879,-0.2047878057,-0.0360313505,-0.002501182,-0.1963063926,0.0693461597,0.0041416334,0.1513000429,-0.0839699879,0.0027169939,-0.0583308637,-0.1633187979,-0.1345888972,0.2782887518,0.1325925738,0.0228500515,-0.1494448036,0.2259258777,-0.2175251544,0.0171872098,-0.490966469,0.3663880229,0.4204707742,0.0141101293,-0.2601905465,0.3800023198,-0.2002423406,0.0039442526,-0.0523405783,0.107850574,-0.0839459226,-0.2194305807,-0.1304418445,-0.2291510999]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/988","title":"making sure datasets are not loaded in memory and distributed training of them","comments":"my implementation of sharding per TPU core: https:\/\/github.com\/google-research\/ruse\/blob\/d4dd58a2d8efe0ffb1a9e9e77e3228d6824d3c3c\/seq2seq\/trainers\/t5_trainer.py#L316 \r\nmy implementation of dataloader for this case https:\/\/github.com\/google-research\/ruse\/blob\/d4dd58a2d8efe0ffb1a9e9e77e3228d6824d3c3c\/seq2seq\/tasks\/tasks.py#L496 ","body":"Hi\r\nI am dealing with large-scale datasets which I need to train distributedly, I used the shard function to divide the dataset across the cores, without any sampler, this does not work for distributed training and does not become any faster than 1 TPU core. 1) how I can make sure data is not loaded in memory 2) in case of distributed training with iterative datasets which measures needs to be taken? Is this all sharding the data only. I was wondering if there can be possibility for me to discuss this with someone with distributed training with iterative datasets using dataset library. thanks ","comment_length":16,"text":"making sure datasets are not loaded in memory and distributed training of them\nHi\r\nI am dealing with large-scale datasets which I need to train distributedly, I used the shard function to divide the dataset across the cores, without any sampler, this does not work for distributed training and does not become any faster than 1 TPU core. 1) how I can make sure data is not loaded in memory 2) in case of distributed training with iterative datasets which measures needs to be taken? Is this all sharding the data only. I was wondering if there can be possibility for me to discuss this with someone with distributed training with iterative datasets using dataset library. thanks \nmy implementation of sharding per TPU core: https:\/\/github.com\/google-research\/ruse\/blob\/d4dd58a2d8efe0ffb1a9e9e77e3228d6824d3c3c\/seq2seq\/trainers\/t5_trainer.py#L316 \r\nmy implementation of dataloader for this case https:\/\/github.com\/google-research\/ruse\/blob\/d4dd58a2d8efe0ffb1a9e9e77e3228d6824d3c3c\/seq2seq\/tasks\/tasks.py#L496 ","embeddings":[-0.285097748,-0.4469243288,-0.1016407087,0.2859593928,0.0856813341,-0.1491084099,0.2383992374,-0.1465865225,0.0240799971,0.4799138606,0.0181809831,-0.3488185406,0.0343001224,0.0275095236,0.0604484193,0.0179703571,0.1120214239,0.3357720077,-0.3774603903,-0.2143622786,0.0459745787,-0.3006141484,-0.223113209,-0.4654530585,-0.2706236243,-0.3876399994,-0.3339650035,0.1683359593,0.0477674939,-0.1257739812,0.0534130745,0.3121866584,0.1771171987,0.3838865757,-0.0001218972,0.0222533122,0.0042732516,-0.2202750742,0.0192178413,0.2641945183,-0.191398263,0.0063295132,0.0003852803,-0.334035188,0.0132982358,0.4170162976,0.1963105798,0.0613025464,0.3701331913,-0.1553578228,0.0558342673,-0.0255187601,-0.2354943156,0.1314513981,-0.1535720378,0.2047358602,-0.0576478876,0.0762736872,0.1306769103,0.2814641595,-0.3883312941,-0.0501429215,0.0898795873,0.3026072979,-0.0810466632,-0.0840713978,-0.295624882,-0.4764955342,-0.1365134865,0.2408643663,0.3387235403,-0.1296498477,-0.263266772,-0.0972090289,-0.2123541981,-0.2463740408,0.2450061738,0.2494798452,-0.1925616115,0.1366336197,-0.4543084204,-0.2199466228,-0.0141443247,0.0019956343,-0.0769488215,0.2673737407,-0.0107317837,0.0682420507,0.1199442446,0.0490622669,0.0894723609,-0.155540809,0.1790196151,0.1713034213,-0.4141792059,-0.351346463,-0.270578593,0.2262998521,0.3593382835,0.2377804369,0.2247718722,-0.1068681702,-0.0023740984,0.0505558774,0.2380449921,0.1071522534,-0.2517019808,0.340919286,0.0513899326,0.0400046036,-0.2665715218,0.0251010731,0.0846540108,-0.0536056608,0.2762925923,0.0320422202,-0.3585418165,-0.2263350636,0.1232634634,-0.3235213757,0.3846230209,-0.3293304145,0.0101514123,0.3466292322,-0.0196163971,0.2146199942,-0.2132746428,-0.1072795838,-0.398860842,-0.3019176722,-0.2076008171,0.2741039097,-0.2674984038,0.4903126955,0.2091442049,-0.0299386587,0.184454903,0.1069870815,0.1533885449,-0.0155791789,-0.0933186859,-0.5966613889,0.4550487101,0.2268831879,-0.1433906555,-0.2938842773,-0.1485484689,0.1741016954,-0.1960165948,0.236036852,-0.4523730874,-0.2490272075,0.5554457903,-0.0028192489,-0.3629242182,0.1157255173,-0.4995554388,0.1332217604,0.2668932676,-0.1161277071,0.1454328001,-0.0052114259,-0.2313591689,-0.2011776865,0.4455416799,0.4342692792,-0.0924743563,-0.1634878963,-0.1232568026,-0.2671778798,0.3049454093,0.5289571285,-0.1256093681,0.2426065207,-0.1033519655,0.1214640588,0.482978791,-0.1648506373,-0.5223394632,0.3366286755,-0.0895818919,-0.0239746459,0.3501639664,0.1028359905,0.363491416,-0.2017167658,-0.3693883121,0.3324264288,-0.0446624793,0.0064770789,-0.3000495434,-0.3167108893,0.0293716565,0.3014536798,0.2228060663,0.3439861834,0.3938890398,0.1369553357,0.1571153998,0.1752231568,0.2316107154,0.1442394257,-0.0722376183,-0.1870475411,0.0008604322,0.1804031879,-0.3296205401,0.1845464706,0.2262245566,-0.106459178,0.1684907079,-0.0674762577,0.2585634887,-0.2612834573,-0.051738482,0.1651479751,-0.1074118838,-0.0291459728,0.2330950648,-0.0241219141,-0.2720944285,0.1472534537,-0.5785939693,-0.0418074019,-0.3138766289,0.3831665218,0.3711929321,-0.2603056729,0.0054376628,-0.0081121624,-0.3401144743,-0.1454956681,-0.0536569804,0.3793819249,0.1350981444,0.2542769313,0.2725787461,0.4957120121,0.2204970121,-0.0001669396,-0.1201718822,-0.1857900918,0.0326541886,-0.2213969976,0.1440976411,0.3712821305,-0.2791765034,0.0334159657,0.3532111049,-0.386836201,0.0125810327,-0.2244392037,0.079587236,-0.0474359952,0.1515684426,-0.1644577682,0.2237015069,0.3324975669,-0.3510330319,-0.2701885104,0.3463738263,-0.0360678397,0.1155010909,-0.0303095877,0.2549030781,-0.1055499539,0.4570446014,0.1863237172,0.5929903984,0.029524602,0.1714456826,0.0755788609,0.4698835909,-0.1659840047,-0.1260812581,0.1193937734,0.0960812047,0.1210762411,-0.2539706528,-0.0385595895,-0.0428546369,-0.2479244471,0.1410574168,0.0665938482,-0.2223700434,-0.3473719954,-0.2748941481,0.171120286,-0.1257777065,0.5265744328,-0.1569702327,0.2490986884,-0.1757230908,0.1859362423,0.1160792559,0.0754176304,0.1315279901,0.5472943187,-0.2812935114,-0.1306928992,-0.101544559,-0.0055974396,0.0378920846,-0.010304885,0.0849505067,0.0203342326,0.3242632151,-0.1828661263,-0.1165828109,0.0160064381,0.2350922972,0.0868623257,-0.0110533815,-0.119467631,-0.2694165409,0.2036110759,-0.3451444805,0.0100246808,-0.0343434177,-0.3875853717,-0.2120893598,0.0211277548,0.2033813,0.0441890731,0.1312756985,-0.3102548122,-0.5511998534,-0.1931799203,0.0596582741,0.0882997811,0.2374936938,-0.3693223298,0.0009117867,-0.2158302367,-0.1285565943,0.0852252916,0.0186598413,-0.7339746356,0.2251438946,-0.1277549416,-0.1065980718,0.1969069242,0.0439924076,0.2969291508,0.6327620745,-0.601549089,0.0256650783,-0.0054410538,-0.0483612008,-0.0260075741,0.3246159554,0.3580394089,-0.2550134659,0.092970565,0.1729644686,0.2414897382,0.218149215,-0.2989733219,0.2201259881,0.2244482785,0.3264575601,0.0751104951,1.064537406,0.1015434861,-0.0853825733,0.1934490353,-0.054232277,-0.2000161409,-0.131140098,0.068760477,0.4237136245,0.2482779175,-0.3482013047,0.3478907943,-0.0717026293,0.067837894,-0.311026752,-0.200659588,-0.189043507,-0.1877678931,0.1117288545,-0.1849602759,0.3963999748,-0.4010534286,-0.1964775473,0.1389084607,-0.4262200296,-0.318005383,0.368234396,0.2384227514,0.1031150073,-0.3052876592,-0.0947430059,-0.0156618468,0.1499421448,-0.0120916702,0.3900527656,-0.2108360231,-0.1154823229,0.2468057722,0.2986290157,0.2268299013,-0.1457247585,-0.0798829719,-0.3076453209,-0.777862072,0.3000763655,-0.0271937139,-0.1035114899,-0.1390631348,-0.0292469133,0.133903414,-0.3538064063,-0.2225175202,-0.0865317807,-0.0404616632,-0.226362437,-0.4271759391,-0.2894743383,0.0326933339,-0.0461131781,0.1753048748,-0.0697796866,-0.1096778214,-0.1639954001,-0.0245052241,-0.3207270205,0.123097606,0.2537720501,-0.0023328685,-0.0198343769,-0.0037336689,0.104982473,0.0772402883,0.1602945775,0.4270332754,0.3155898452,0.0475321151,0.1518766731,0.103822127,0.1812534183,0.3947945535,0.1755104065,-0.0963862166,0.1324894577,0.0933346152,0.1298791319,-0.5946345925,0.1252009422,-0.206732735,-0.1251223534,-0.3810137808,-0.3799036741,0.5109300017,0.2016141862,-0.0772605017,0.2692998648,-0.1859848201,-0.2704081237,0.0910877883,0.2416861206,1.1949558258,-0.2357973158,0.0015875181,0.0235199034,0.6256213784,0.2952644229,-0.142560482,0.0791517049,-0.1567939371,-0.506491363,-0.1396771669,-0.0914026126,0.1993635595,0.6071106791,-0.2471410781,0.5053642392,0.0977537036,0.1230283678,-0.0511082783,0.2899981737,-0.0255787112,-0.2176855356,-0.0754524842,-0.0066130995,-0.0861462131,-0.0675923377,0.0272292141,0.1248542741,0.0706727132,0.3435359597,0.3692074418,-0.0705789179,-0.1748686433,0.1539407223,-0.0577231497,-0.2397858351,0.0510639772,0.3057440519,0.2143629938,-0.0646287724,0.0794181451,0.1858104914,0.0110641271,0.0627135113,0.0538734123,0.1874873787,0.3667631447,-0.0420449376,-0.1687341183,0.1026156917,-0.1083456427,-0.3432527184,-0.3397271931,-0.2703552842,0.2166690528,-0.2415284067,0.0034722271,0.1015732437,-0.1148637235,0.2285501063,0.0456751175,0.106235832,-0.2522290051,0.4409602582,0.1434900165,-0.2262580693,-0.2410607636,0.399733752,0.2230374217,-0.0317891203,0.0558510013,-0.336022824,-0.115932852,-0.2397971302,0.1464861482,0.2557976544,-0.2557865381,0.0343996398,-0.0327229016,0.2196079046,-0.0911332741,0.2653229535,0.3497650027,0.1494254768,-0.1341978461,-0.213292703,-0.3637446761,0.1150746346,-0.0284006055,0.275542438,0.041388575,-0.1445616335,0.0022593508,0.5074510574,-0.1924565881,-0.315613687,-0.3064939976,0.2280227095,0.0856725648,0.1686109006,-0.1790863574,0.0609555244,-0.0969797894,0.0760204494,-0.1741685867,-0.055994492,0.2542454898,0.2356889844,0.1058013737,0.2557393909,-0.2457221001,-0.286653161,-0.0609524101,-0.2496193945,0.3579746187,-0.1908962429,-0.0357806906,0.0814168453,0.5675016642,-0.1763918102,-0.5132965446,0.331522584,-0.1228829697,-0.0873665363,-0.0473595336,0.1017596796,0.0947391689,-0.0541585013,-0.1167320311,0.0673039109,0.4573140442,-0.2007621378,0.0813215971,-0.0831921473,0.2025208771,-0.1187139973,0.2375454158,0.278608799,-0.1989977658,-0.1033812389,0.3501418531,0.0641594604,0.1749477834,0.1949371099,0.0907966048,0.412189424,-0.0727303475,0.3976581097,0.0841873512,-0.1620191038,-0.1039725095,-0.0637226626,0.3649158776,-0.4576648474,0.0706191361,0.0527823456,0.0701726228,0.2187184989,0.2454748452,0.2847749889,-0.1966038495,-0.1076284423,0.0905356556,0.937486887,0.0294712689,0.3309078515,-0.0883608013,-0.7197977304,-0.1805518121,0.3365643024,-0.2061477751,-0.066504918,-0.2090165019,0.1267830282,0.1566277146,-0.3895667493,-0.0767219216,0.2007272691,-0.0612219088,0.0977807567,-0.1993889958,-0.089860186,-0.1976269931,0.0342360586,0.1126263067,-0.1062224805,0.7192896605,0.3560318053,0.0974273458,0.1140450686,0.0710157156,0.2500627041,0.2436406314,-0.2792297006,0.2228551805,-0.1764144897,0.0562716573,0.2046948224,0.2707375586,-0.1226151809,0.2770506144,0.3466807306,-0.2012426108,0.3650785089,0.0360430367,0.099968113,0.4576960504,-0.0863892585,0.241648972,0.1397784799,0.0419722944,0.0460630991,0.1136242151,0.023222547,0.1060885489,-0.3862357736,0.3639751077,0.2190271914,-0.0323519669,0.1867967248,0.1043370366,-0.3539731801,0.2261858135,0.2614781857,-0.7350065708,0.0536823571,0.0051753134,0.0006826246,-0.1476230323,0.4439831078,0.2665707171,0.012703428,-0.2566806078,-0.422954917,-0.2817396522,0.0209111962,-0.0806217715,-0.2530038357,-0.0539070033,0.2732564807,0.0111294445,0.0842351466,-0.4803050756,-0.3653127253,-0.2499962747,0.3177435696,-0.1425902396,-0.0580937937,-0.6569178104,0.1993012726,-0.0205474794,-0.3092359006,0.2699124217,-0.0003713911,-0.1959732622,0.3814072907,0.200753957,0.162461549,-0.0903098211,0.2300457209,-0.0853756517,0.2732608914,-0.0545865335,0.3909990191,-0.1388929933,-0.0620086305,-0.1367463171,0.0556106903,-0.0775226504,0.1302131116,-0.1727822721,0.2934637666,-0.0388475247,-0.1885387301,-0.2850565314,0.154657796,0.2135297656,-0.3440775573,0.2935578227,0.246616438,0.5267214775,0.3046112061,-0.1450859159,-0.247274816,-0.1405438334,-0.0609351248,0.0872265697,-0.2570711374,0.0993138552,-0.3666562736,0.3546512723,-0.0944079161,-0.1182358935,-0.535998106,-0.0756091624,0.3408851624,0.0310136583,-0.1221683025,0.278460294,0.2851260304,-0.1024947912,0.0723014772,-0.1669162959,-0.3003565073,-0.2355331182,-0.127407968,-0.304115057]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/961","title":"sample multiple datasets ","comments":"here I share my dataloader currently for multiple tasks: https:\/\/gist.github.com\/rabeehkarimimahabadi\/39f9444a4fb6f53dcc4fca5d73bf8195 \r\n\r\nI need to train my model distributedly with this dataloader, \"MultiTasksataloader\", currently this does not work in distributed fasion,\r\nto save on memory I tried to use iterative datasets, could you have a look in this dataloader and tell me if this is indeed the case? not sure how to make datasets being iterative to not load them in memory, then I remove the sampler for dataloader, and shard the data per core, could you tell me please how I should implement this case in datasets library? and how do you find my implementation in terms of correctness? thanks \r\n","body":"Hi\r\nI am dealing with multiple datasets, I need to have a dataloader over them with a condition that in each batch data samples are coming from one of the datasets. My main question is: \r\n- I need to have a way to sample the datasets first with some weights, lets say 2x dataset1 1x dataset2, could you point me how I can do it\r\n\r\nsub-questions:\r\n- I want to concat sampled datasets and define one dataloader on it, then I need a way to make sure batches come from 1 dataset in each iteration, could you assist me how I can do?\r\n- I use iterative-type of datasets, but I need a method of shuffling still since it brings accuracy performance issues if not doing it, thanks for the help. ","comment_length":109,"text":"sample multiple datasets \nHi\r\nI am dealing with multiple datasets, I need to have a dataloader over them with a condition that in each batch data samples are coming from one of the datasets. My main question is: \r\n- I need to have a way to sample the datasets first with some weights, lets say 2x dataset1 1x dataset2, could you point me how I can do it\r\n\r\nsub-questions:\r\n- I want to concat sampled datasets and define one dataloader on it, then I need a way to make sure batches come from 1 dataset in each iteration, could you assist me how I can do?\r\n- I use iterative-type of datasets, but I need a method of shuffling still since it brings accuracy performance issues if not doing it, thanks for the help. \nhere I share my dataloader currently for multiple tasks: https:\/\/gist.github.com\/rabeehkarimimahabadi\/39f9444a4fb6f53dcc4fca5d73bf8195 \r\n\r\nI need to train my model distributedly with this dataloader, \"MultiTasksataloader\", currently this does not work in distributed fasion,\r\nto save on memory I tried to use iterative datasets, could you have a look in this dataloader and tell me if this is indeed the case? not sure how to make datasets being iterative to not load them in memory, then I remove the sampler for dataloader, and shard the data per core, could you tell me please how I should implement this case in datasets library? and how do you find my implementation in terms of correctness? thanks \r\n","embeddings":[-0.3916403353,-0.0641560704,-0.0715138465,0.1802322865,0.0308315791,-0.1964396238,0.3828041852,-0.0021056198,0.3755799234,0.3051428497,0.0210494325,0.1671252847,-0.199287191,0.2304519564,-0.0638537854,-0.1857499182,0.1219987422,0.1643287688,-0.5311963558,0.0054712659,-0.0241466686,-0.1971098185,-0.1702240109,-0.2883293927,-0.177111581,-0.2674229145,-0.4406118989,-0.0727713034,-0.1395245045,-0.0913867578,0.1153420433,0.490537107,0.3580513,0.5394567251,-0.000111204,-0.1196052656,-0.044422999,-0.1387528181,-0.1061960831,-0.1808307171,-0.0781026706,-0.0892329291,0.0344417617,-0.3380921483,0.0083269645,0.0254511405,0.1565666646,-0.3575609922,0.2909744084,-0.1612730473,0.1276792139,0.1199493259,-0.1547259986,0.0384347066,-0.0925308093,0.3123201132,0.0935158804,0.4337083697,0.5372362733,-0.159167558,-0.3149550855,0.0160149205,-0.0044768094,0.2461697608,0.0793833658,-0.1820981055,0.0218562614,-0.3477081358,-0.3850466907,0.2288401425,0.4140793383,-0.279899776,-0.3263224959,-0.2925460339,-0.0736813918,0.2130671591,-0.0283798203,0.1663020104,-0.2419202179,-0.0525974296,-0.1114866138,-0.0543073602,-0.1045175418,0.1332090795,-0.023840351,0.1524389386,-0.1014345288,0.1528570652,0.3740716279,0.0841336623,0.2387399971,-0.410461992,0.2212483287,0.2202965915,-0.4802778959,-0.2056568861,-0.24776797,-0.1046733856,0.26503703,0.2081973553,0.121757105,-0.1342131644,-0.1067484468,0.1430408806,0.3041803241,-0.0853135884,-0.0716501623,0.2037613839,-0.0634067059,0.2247578949,-0.0819226354,0.0488003194,-0.3137095273,-0.0129536679,-0.0031802626,0.2531783581,-0.2476243675,0.0603042766,0.2005784065,-0.4978003502,0.1361875534,-0.3567002714,0.2366858125,0.0747649297,-0.1907098591,0.3283049762,-0.1854242682,0.2657660842,-0.0532729663,-0.2072146684,-0.1880469918,0.2148921043,-0.3193081915,0.2454105467,0.1908979118,-0.0126184532,0.1669201255,0.3347459733,0.0716238469,0.0064523933,0.3675651252,-0.2884424031,-0.2200206965,-0.1544591933,0.2525639832,0.2033264935,0.1435806155,0.3965344429,-0.2144623995,0.0902175456,-0.4931745827,-0.0262247566,0.2761395574,0.1008031145,-0.1596169174,0.0404095985,-0.8439462185,0.2114651054,0.1535434127,0.0032143136,0.0099209603,0.0232038572,0.1156395748,-0.2657038569,0.0627910644,0.3633191288,-0.2656930089,-0.1372898519,-0.0091387006,-0.2979201078,0.4705411494,0.0035394614,-0.4522489607,0.3560403585,-0.2815080285,0.1066805348,0.2924124897,-0.4420186579,-0.065670304,0.4262463152,-0.1766885668,-0.0041094092,0.3303823173,0.0544343144,0.2620957494,-0.1768242866,-0.365495652,0.5972786546,-0.2043524086,0.1548202336,-0.0525416061,-0.2917997837,0.1647380888,0.4052251875,0.1216875985,0.0416774824,-0.0769997016,0.0828570426,0.3132277131,-0.231779933,0.3287110329,0.0279812403,0.0492679738,-0.1616756469,-0.1748330593,-0.0466452464,-0.3390178084,0.1256399006,-0.2319261879,-0.0970664993,0.4011271298,-0.0232186895,0.3465255797,-0.2669230103,-0.3869844079,0.0723768026,0.0611793436,0.2023778558,0.1085962802,-0.3398434818,-0.2768117189,0.190814063,-0.2290107012,0.0546378084,-0.3565302789,0.3841884434,0.1405826211,0.1393807679,-0.0097083859,-0.1697499901,-0.1182317436,-0.3064123988,0.0667834952,0.2394201458,0.1556432992,0.1732922941,0.1532119066,0.042723503,0.225937143,0.0417281725,0.2905356884,-0.2535174191,0.1214320585,-0.32268098,0.0727038607,0.3777752221,-0.4089926779,0.3344213963,-0.0807147697,-0.1459089369,0.1602962762,-0.0955844671,-0.0668594614,0.0897979364,0.0799409002,-0.0394896828,0.5106659532,0.033486709,-0.2563400269,-0.0866077095,-0.3283580244,0.0508652553,0.0272149257,-0.0320315845,-0.0541128032,-0.0881682113,0.0534976833,0.2296973765,0.5622861385,0.0916322097,0.1222975105,-0.313041389,0.3364981413,-0.0595748015,-0.0175773036,-0.0588467903,0.2014515847,0.1984524131,-0.2046967447,-0.1562070101,0.1088904887,-0.349308908,0.4324038625,-0.010543013,-0.0821597874,-0.0819676444,-0.1365523487,-0.211372152,-0.378957212,-0.181583181,-0.1165512055,0.2115495056,-0.1122413874,0.4244526625,0.1437133551,0.0697333962,0.0289149098,0.4648036063,-0.2581796348,-0.1165130883,0.1640331894,0.0414396934,-0.0891951248,0.1470765471,0.300648272,0.2237675041,0.2247207165,0.047310032,-0.1249577776,-0.1231014505,0.393351227,-0.0228143204,0.0427180156,0.1739188731,0.0519869141,0.186262086,-0.0528644398,-0.1905214638,-0.0391823873,0.0591024198,-0.1496568918,-0.1275014877,-0.2147630453,0.0657796785,0.0686950237,0.0749981403,-0.6593735218,-0.2051500231,-0.1912441552,-0.0376747064,0.2734798491,-0.0044931122,0.1117944866,-0.1348209977,-0.0110144448,-0.0916977078,0.0539694391,-0.8185967207,0.0981758609,-0.0951287001,-0.0645316541,-0.3637395501,-0.1193406656,0.2723683417,0.3621838391,-0.2142647356,-0.2822706103,0.3100191355,-0.0146709578,0.1291709244,0.3735460937,0.1150702387,-0.1168296933,0.1084925756,-0.0514418557,0.3031534851,0.3980281949,-0.1371177435,0.4599200189,0.2732169032,0.0310575776,-0.1509060562,0.6859967709,0.2674050927,-0.1025428697,0.3714950979,0.2768211365,-0.2127128094,0.2239500135,-0.3388921618,0.2419187874,-0.024466116,-0.0515040196,0.2720037103,-0.2013646364,-0.1227095425,-0.2037574947,-0.1127375364,-0.2855706513,-0.1201566681,0.3364460766,-0.4417082965,0.1837793291,-0.4353539646,-0.4029151201,-0.2996039093,0.0606859103,-0.1537039727,0.1451987773,0.1991253197,-0.0662371963,-0.7584744692,0.0372873172,-0.1249942854,0.0306953341,0.0765898153,0.2014190704,0.3181453645,0.1460332423,-0.018023558,0.1173703447,0.6082801819,-0.0661456734,-0.1419585198,-0.1396322548,-0.2273276895,0.0899065882,-0.2605347633,-0.2128526121,0.3071988225,-0.1260770559,0.1600037664,-0.479688704,-0.2513048649,-0.1834201664,0.2079835087,-0.1182783172,-0.4122638106,-0.0381339975,0.0837408975,-0.0803017095,0.0695662424,-0.0831739902,0.233773157,-0.0652955994,-0.0486731194,-0.1169731915,0.1158371121,0.231575951,-0.0874583051,0.0555868782,0.0780887976,0.3457980752,0.1934350729,-0.0737488866,0.518306613,0.3417626619,-0.490023315,-0.1208452359,0.1475350559,0.0188254304,0.2842287123,-0.0302861128,0.1082385257,0.3041319847,0.0835691988,-0.1507168412,-0.2274364382,0.0888385102,-0.0415253304,0.0231990591,-0.4876147509,-0.6660345197,0.468459934,0.3520278931,0.0157676991,0.5351880193,-0.5151388645,-0.261973381,0.1972192377,0.2440884858,0.6343522072,-0.0645349547,-0.0142659061,-0.0526922308,0.6073738337,0.4276472032,-0.1701568067,0.0359557085,-0.273194015,-0.3259967566,-0.089996174,-0.2492547929,0.1500123441,0.6537644267,-0.0752204433,0.314753592,-0.0180118084,0.0218272638,-0.0847361088,0.2518901527,-0.0768450797,-0.4026192129,0.0569935814,0.0568419695,-0.1508840472,0.1048473492,-0.0717465654,-0.1574088633,0.1687270701,0.2035701871,0.0173581075,0.1116112545,-0.3153598011,0.1035550088,0.106475994,0.1456953585,0.0690371245,0.0525827967,0.1457950771,-0.1208103523,0.0828677788,-0.0783257186,0.039951358,0.0345094949,-0.1295622438,-0.1900818199,0.4435195327,0.0005731077,-0.1457633674,0.0637674555,0.1457603723,-0.1982660443,-0.1122392118,-0.0134078581,0.2863933146,-0.263312757,-0.2650086284,0.2543340921,0.1905943602,0.0370326899,0.0668788254,0.2705124021,-0.190426901,0.6262239218,-0.1872766465,0.09305913,-0.1267874092,0.067933619,0.3257100284,-0.3565046191,0.0873127803,0.1231315956,0.0086442651,-0.1919458658,0.0667832941,0.4896388352,-0.2254796773,0.1107655838,-0.1611425132,0.107313551,0.15497908,0.3132692873,0.2724294662,0.167606011,0.0756264329,-0.3852107823,-0.3145557642,0.1792341471,-0.0731982142,0.3070866764,0.0941191167,-0.0287555903,-0.0472243577,0.4644154608,-0.2387148142,-0.0739164799,-0.0439634323,0.1152409613,0.2772041559,0.1561947316,-0.1621524096,0.3207270503,-0.0046563298,-0.0284529608,-0.0849507377,-0.1576074064,0.0386969,0.1843989193,0.0473490246,-0.0253964625,0.1175850779,-0.2932241559,-0.0927736983,-0.1650468111,0.0020964826,-0.2110262513,-0.0389786102,-0.0483469591,0.5077000856,-0.1548279524,-0.4957728684,0.4022542536,-0.0133389039,-0.2609181106,0.0047041597,0.2325227708,0.1736960709,-0.3700736761,-0.450350672,-0.0439994931,0.4514316916,0.1486096382,-0.0588129014,-0.0423283242,0.0874571875,0.2635795176,0.3449197114,0.2492966354,-0.0706129447,0.0667752922,0.3052440286,0.1638261378,-0.045641318,0.0237994399,0.0184332822,0.0652498379,0.071026057,0.4253252745,0.0888465345,0.0350847803,-0.1052877754,-0.0379476734,0.2695039511,0.1630143523,0.2150264978,0.1660506427,0.0464003123,-0.123306118,0.438277334,0.2847966254,0.0830297992,-0.1347545385,-0.2000863403,0.6424167752,0.2414994091,-0.1970065087,0.0568949729,-0.596843183,0.1614179462,0.4746443927,0.0026906691,-0.052801732,-0.1341400146,0.5436433554,-0.4190953374,-0.4090664983,0.0036268989,0.4454933107,0.0053001773,-0.0089169759,-0.5101049542,-0.1802321523,-0.2166267782,-0.1497143954,-0.0265609343,-0.2004607767,0.4630516768,0.0762895867,-0.0038346166,-0.3279163241,-0.2256793231,0.4454766214,0.008699201,-0.3360503912,0.1276868135,0.1430613399,-0.1460707486,0.0249374472,0.3911396265,0.1470084339,0.4168976545,0.2586797476,-0.0381480977,-0.1158528775,-0.0226655006,0.0162768159,0.0359926261,-0.1415427774,-0.1696101576,0.3932849765,0.1120727509,-0.1219302341,0.147296682,0.2878340781,-0.1451048404,-0.2953323126,-0.0019900347,0.3645460606,0.1936895847,0.0502156951,0.0415512286,-0.5218898654,-0.0013009596,0.1949213147,-0.4592628181,-0.0127273332,-0.0081410175,0.0428719558,-0.089567177,0.2777059674,0.3190700114,0.2470875531,-0.5013747811,-0.3431473076,-0.3294943571,-0.0038282585,-0.1907373667,-0.1383875012,-0.3621275723,0.2742125392,0.0922476575,0.0791240111,-0.2293824703,0.171263814,0.0204745252,0.5848488808,-0.456245929,-0.0630008653,-0.192102924,0.4216589332,0.1097044796,-0.3497252166,0.1359268129,0.1766252667,-0.0787556395,-0.0335157886,0.0060759322,0.4594005942,-0.3696484864,0.2187571675,-0.1088632643,0.5026680827,-0.0190915558,0.223815605,0.0433431901,0.101631403,-0.3354448676,-0.2664397955,0.263309896,0.3186025321,-0.1851017326,0.2326597571,-0.2636696696,-0.0721024275,0.0876754373,-0.0653568581,-0.331482321,-0.1060245931,0.0035361552,-0.096131593,0.3555636704,0.2686107457,-0.0257672649,0.0629180744,-0.2068597078,-0.0556054413,0.0867587849,-0.3383480608,-0.0118225496,-0.4012134373,0.047852993,-0.2957096994,0.2436215132,-0.3869486749,0.1066703722,0.4079364836,-0.0373276137,0.2124281675,0.3201629519,0.3638679385,0.0808595195,-0.2029716372,-0.1619793922,-0.0559309945,-0.1848646998,-0.0263393428,-0.2335924506]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/937","title":"Local machine\/cluster Beam Datasets example\/tutorial","comments":"I tried to make it run once on the SparkRunner but it seems that this runner has some issues when it is run locally.\r\nFrom my experience the DirectRunner is fine though, even if it's clearly not memory efficient.\r\n\r\nIt would be awesome though to make it work locally on a SparkRunner !\r\nDid you manage to make your processing work ?","body":"Hi,\r\n\r\nI'm wondering if https:\/\/huggingface.co\/docs\/datasets\/beam_dataset.html has an non-GCP or non-Dataflow version example\/tutorial? I tried to migrate it to run on DirectRunner and SparkRunner, however, there were way too many runtime errors that I had to fix during the process, and even so I wasn't able to get either runner correctly producing the desired output.\r\n\r\nThanks!\r\nShang","comment_length":62,"text":"Local machine\/cluster Beam Datasets example\/tutorial\nHi,\r\n\r\nI'm wondering if https:\/\/huggingface.co\/docs\/datasets\/beam_dataset.html has an non-GCP or non-Dataflow version example\/tutorial? I tried to migrate it to run on DirectRunner and SparkRunner, however, there were way too many runtime errors that I had to fix during the process, and even so I wasn't able to get either runner correctly producing the desired output.\r\n\r\nThanks!\r\nShang\nI tried to make it run once on the SparkRunner but it seems that this runner has some issues when it is run locally.\r\nFrom my experience the DirectRunner is fine though, even if it's clearly not memory efficient.\r\n\r\nIt would be awesome though to make it work locally on a SparkRunner !\r\nDid you manage to make your processing work ?","embeddings":[-0.3029166758,-0.2597852647,-0.0041649756,0.0888215601,0.0611687601,-0.2294601947,0.2730875015,-0.1252957582,0.1471410692,0.2653598189,0.1793320179,0.292142719,-0.4641712606,0.6944869757,0.2492861748,-0.5270981789,0.3334576786,-0.1199868992,0.0216181409,-0.2315687686,0.2499720454,0.346807003,-0.3306990564,-0.0525378808,-0.0932485908,-0.0328033715,-0.1721203774,0.2677879333,-0.2200568318,0.0841011629,0.2658599615,-0.1088843346,0.1863445193,0.9052969813,-0.0001201803,-0.0497153252,0.1546807736,-0.0539860837,0.0199465435,-0.1867879331,0.1566442847,-0.0154153006,0.1619346738,-0.0727845058,-0.2123981416,-0.0957708508,0.0908535048,-0.0055826623,0.588593483,0.169653371,0.1072054803,0.3378794789,-0.1923525631,-0.1925451756,-0.0544641316,0.3821781278,-0.3049484789,0.4146491885,0.4382084608,-0.0001665803,-0.2150050253,0.2646872401,-0.1692704409,-0.1197135076,0.274279952,-0.1783319712,-0.4092798531,-0.2647215724,0.0094662057,0.3803465664,0.6444175243,-0.3084756732,-0.1788963377,-0.1481508464,0.077381447,0.0355662592,0.1662692428,0.1155457944,-0.2511981428,0.1777534932,-0.7257139683,-0.2467558533,-0.2392481565,-0.2115032971,-0.1159090698,0.3887881935,-0.1884066612,0.0926894322,0.1613622606,0.0715748742,0.4279718101,0.0208769329,0.1395020485,-0.1157980636,-0.4014292955,-0.1895822436,-0.0212012138,0.6500042081,-0.0850925595,0.1844476461,0.1049814001,-0.1101926491,0.0840262249,0.1532219052,-0.1201360673,0.2642698586,0.2677609622,0.0173277669,0.0258307029,-0.2055127919,0.205679372,0.1432573199,-0.451302588,-0.1956639588,-0.7054801583,-0.0198629946,-0.2233604938,-0.1397652924,-0.0853938982,-0.0715436935,-0.114061147,0.0226786807,0.2285529524,0.160407275,0.2587019205,0.1301407516,-0.0917088538,0.0659475401,-0.0490657575,-0.3933473527,-0.0686852261,0.3721349239,-0.5328849554,-0.0297052823,0.0960214064,-0.0564032793,0.0813707113,-0.0479136147,0.1972567439,0.3849090934,0.0602352731,0.1320077926,0.0568427555,0.1362747699,0.0219949577,0.1173633784,0.1448400319,0.1398748755,-0.2038722634,-0.1034587696,-0.2237517089,-0.3094114959,0.0077487589,-0.0295714587,-0.0920754969,-0.232178852,-0.5076836944,0.4833564162,-0.2039633542,0.0057094223,0.2567739785,-0.1527157426,-0.2808699608,-0.2357541919,0.3215939105,0.4583259821,-0.526817739,0.0077194958,0.091197826,-0.3155311048,0.2618651986,0.2542577684,0.0717829168,0.2218497843,-0.3367223144,-0.0444209725,0.3111922443,-0.7092143297,0.1847455055,0.667861104,-0.2511822581,-0.1177572757,-0.0484736189,-0.1916433424,0.0642129108,-0.0097681554,-0.0884431899,0.2391255647,-0.1667373478,0.128468737,-0.0637892038,-0.2918200791,-0.1378310919,0.2049366534,-0.0137708057,0.1322251558,-0.0084460527,-0.1825434417,0.2063750327,-0.0425384566,0.0986142755,0.4210444689,0.0658791214,-0.1860312968,-0.0873794854,-0.0554307215,-0.0324145593,0.1529463977,0.2787356079,-0.0447169393,0.5172027946,-0.1820726544,-0.2667183876,-0.0303762518,0.0600623675,-0.2148182988,-0.0125031192,-0.250058502,0.3179764748,-0.1634434909,-0.3365817666,0.3091777861,-0.2486633211,-0.0815386027,-0.2981159091,0.4468592107,0.1138235703,-0.3273065984,-0.0089702997,-0.1782461256,0.113416031,-0.3471775949,-0.0641544238,0.4955254197,-0.250857234,0.4583969414,0.3250461519,0.5605686903,0.3120700717,-0.0244113207,0.30259642,-0.3043553531,-0.1911411732,0.2306416035,-0.3735177219,0.1239517108,-0.0207574945,0.109043628,0.2474950999,0.4676242173,-0.091564931,0.0088479323,-0.007127889,0.3194851577,0.4472433031,-0.1027063057,0.1383283436,-0.1776744276,-0.5297761559,0.1255438775,0.2470868677,-0.1524432003,0.0014616583,0.0039344076,-0.3945119977,0.0936572403,0.6196673512,0.348968327,0.1256736368,0.2109438479,0.2386404127,0.132665351,0.2032565325,-0.3326336741,0.1098635644,-0.1469198614,0.1641789526,0.0269414857,0.0126337865,0.134936437,-0.2734863162,-0.0991846547,0.0476974174,0.0586126298,-0.1351984292,0.2164726257,-0.3170535266,-0.30358091,-0.2953271866,0.3783070445,-0.1633469611,-0.1004698053,-0.0179955438,0.3528321385,-0.0808843747,0.1908670515,0.3830774426,0.3397631049,-0.2224692851,-0.180575192,-0.3251306117,-0.1658750176,-0.2335842401,0.0170451067,0.2264356017,0.1227668151,0.2665676773,-0.1184477955,0.0159032196,-0.4890853465,-0.2099977732,0.4201814532,-0.0344737321,0.0625678971,0.2083222568,0.0580512509,-0.7962208986,-0.0000914092,0.0758332908,-0.5537876487,-0.2558651567,-0.1597559303,0.1086098775,0.0087943813,-0.2054430246,0.114836596,-0.4072747827,-0.3363671601,0.4378926754,0.0404328108,0.2076618671,-0.3011609912,0.09281075,0.1540543586,-0.0330746472,-0.2489912957,0.128828764,-0.5100076795,-0.0225219335,-0.4240023494,-0.3663930893,0.0597414263,0.1786950827,0.0822686553,0.2870673239,-0.2168940604,-0.2206926197,0.0196215548,-0.1551414579,0.198210299,0.1966361105,0.5560917258,-0.2582683265,0.1138668954,0.0125354026,-0.0091435118,0.0828745365,0.0091216229,0.0329649597,0.3433301449,0.1300072819,-0.0158585813,0.7973929048,-0.0741805136,0.0728162602,-0.1025688499,-0.1284381002,0.2782167792,-0.1638895273,0.0173636209,0.4576426744,0.1414338052,-0.3600635827,0.0164701082,0.2358418852,0.0456860438,-0.2341413945,-0.355784446,-0.0880047157,-0.0365448184,-0.1356152445,-0.0463116318,0.1653000861,-0.1509040445,0.1746188402,0.1001491621,0.0543483943,0.2127901018,0.2468179613,0.2273703963,0.2215748429,-0.1988704056,0.204654187,-0.4602746964,0.1095663235,-0.2576130927,-0.0133429971,-0.3338283896,-0.132216692,-0.0141466465,-0.1376370043,0.7230764031,-0.3389846981,-0.2621293068,0.054561615,-0.3302979767,-0.1781732291,-0.255602628,-0.1431460679,0.0982773975,0.224247098,0.3155584633,0.1999566853,-0.2803316116,0.0975505784,-0.0491103418,0.1015608907,0.0088230977,0.0894233063,-0.1612184793,-0.085391216,0.0859622955,0.0370441861,-0.1177597269,-0.2781384289,0.1649894565,0.008270476,0.0751135051,0.2309519351,0.0013341636,0.1580346227,-0.1394092441,-0.0957103744,0.3423470259,0.5169062614,-0.0716211647,0.1330205649,-0.0208775271,0.1498576552,0.4877234101,0.0558794551,0.2664663196,0.5296105146,-0.2160327733,-0.1808888763,0.0472263247,-0.0519788004,-0.2359547019,0.072097525,-0.1418364495,0.0448876247,-0.1412517726,-0.0976256654,0.4548662603,0.4906442761,-0.2670263648,0.5127561092,0.4088796377,-0.1208225936,0.1734759659,0.0133365113,1.0554234982,0.0016569326,0.1177982688,0.4885128736,-0.1430189908,0.5475766659,-0.5305339694,0.1182217002,-0.334290117,-0.3852117956,-0.1246422082,-0.1634066254,0.3275976181,0.1242474616,-0.1112974286,0.2440290004,-0.0172482971,0.4491384625,-0.2483262122,0.1307474226,-0.1545081139,-0.1395361871,-0.3024851978,0.0793383271,0.1973284036,0.2672301233,-0.2252735198,-0.2027399391,-0.1872768551,-0.0034633707,-0.367454648,0.16070427,-0.2660143971,-0.2253263146,0.2754973173,-0.6139964461,0.3639706671,0.1783306897,0.5937721133,0.0112059563,-0.4948188961,0.2625656128,0.0412666462,0.0233648866,0.062108852,-0.1079570428,0.0566596203,-0.1893245578,0.0783527344,0.0312810838,0.0232159588,-0.073332876,-0.0285961181,-0.1311839521,0.1109094396,0.1184466183,0.0129067712,0.2325351387,0.0766578019,0.096789211,0.037342649,-0.1108589247,-0.0279593561,-0.1504760981,-0.0062421695,0.306094557,-0.1080936715,0.3110361695,0.0753175616,-0.5972081423,0.047655955,0.3871474862,-0.0092671597,-0.239884913,0.024404794,0.5963469148,-0.2447294891,0.0229676422,-0.2787951827,-0.2756192982,0.3374157548,0.0833806098,0.2540742457,0.1235672906,-0.1175528467,0.1041907296,-0.1384351104,0.246155709,-0.0251323394,0.0075644711,-0.1641470939,0.0277096685,0.3490948379,-0.3043750226,-0.2110625952,-0.0164431203,-0.0269931536,-0.2115584314,0.4576094449,0.2009390146,0.1570958197,-0.1562631279,-0.0969046876,0.0014304243,-0.2091068774,-0.0618015565,0.0451362245,0.1672239751,0.072190322,-0.317605406,0.4395858049,-0.0148167266,-0.4041312039,-0.1516283751,0.166075483,0.1654158682,-0.0450248457,-0.1144682243,-0.0547501184,0.3857513666,-0.0112835597,0.0029515342,0.3871547282,0.2254847288,-0.0508905873,0.0972597077,0.328014046,-0.1559069902,0.0257748347,0.199436903,0.0737900063,0.0673026443,0.1660429835,-0.1132616699,-0.0023466039,-0.1303623766,0.5203140974,-0.1079604477,-0.1523617506,-0.0858542174,0.5089148879,0.0968809575,-0.0511811003,-0.0613836795,0.5012268424,0.2788022757,-0.0215615463,0.1282038242,0.3890690506,0.0611621514,-0.289353013,0.0949901417,0.3582286537,-0.1721312702,0.2224830389,0.0266068392,0.0430503003,0.1875685602,-0.1934008002,0.0390869863,0.1262450069,0.4716622829,-0.1522820592,0.305449456,0.1241915077,0.2441392541,0.3796722591,-0.3615206778,-0.1002424508,-0.0757725611,0.0501615368,0.0719821006,-0.2617413402,0.1345465481,-0.2365346551,-0.5163178444,-0.331872344,0.3266518116,-0.0773198083,0.0949672088,-0.4300448,-0.1562930346,-0.1127731577,0.1372811943,-0.0433887579,-0.0536215529,0.4022263885,-0.0932580829,0.3207627833,-0.0274502672,0.2587762177,-0.0812922642,0.2836805582,-0.0014639233,0.2012596428,-0.0261021946,-0.1364035308,-0.113867417,0.2616125941,0.224704504,0.1011153609,0.0732357427,-0.0518705808,-0.1911328733,0.2726092041,0.1300589591,0.1860206127,-0.0665594265,0.007550871,0.0879277587,0.026527999,-0.0052123293,0.2739977539,0.0494311452,0.0081948321,-0.3530938625,0.2339314967,-0.249013871,0.0039735464,-0.1353858858,0.0845595598,-0.3766196966,0.0934632346,0.3412457407,0.3684307039,0.0813928396,-0.0443530679,0.0228583682,-0.0226985496,0.423571825,0.3718879223,0.5538849235,-0.147413671,-0.420601666,-0.335865438,-0.0891872346,0.0809986591,0.1617308855,-0.1413481534,0.017453989,0.2189963907,0.1168738976,-0.1587516516,0.0541568249,-0.2116027921,0.1564374715,-0.071143508,0.0294141062,-0.3237953484,0.1154666767,0.0720893815,-0.1201969981,0.2249583602,-0.1654519439,-0.147086069,0.1159336567,-0.1225319803,0.0044540293,-0.4958156049,0.0246731378,-0.0418882221,0.22358872,0.0366985649,-0.1109812409,-0.1840109974,0.0573640019,-0.3844517469,0.0069520483,-0.2657730579,0.1360099465,-0.3506758213,0.0819669217,-0.2572058439,0.0514893234,0.1000331044,0.3314540684,-0.4565942883,-0.0936173126,-0.1509473622,0.0737447962,0.3138274848,0.2484234571,-0.0530740619,-0.0278220046,-0.5111489892,0.2181857228,0.5034855008,-0.8420357108,-0.0022266603,-0.3200293481,0.0960513353,0.464060843,-0.163667202,-0.430975318,-0.1037833095,-0.0181506723,-0.0661019906,0.3196087778,0.1118425503,-0.2095422,-0.0933535174,-0.1406071186,-0.3227351308,-0.0745263472,-0.279468745,-0.0044083474,-0.1557657719]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/919","title":"wrong length with datasets ","comments":"Also, I cannot first convert it to torch format, since huggingface seq2seq_trainer codes process the datasets afterwards during datacollector function to make it optimize for TPUs. ","body":"Hi\r\nI have a MRPC dataset which I convert it to seq2seq format, then this is of this format:\r\n\r\n`Dataset(features: {'src_texts': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 10)\r\n`\r\n\r\nI feed it to a dataloader:\r\n```\r\ndataloader = DataLoader(\r\n train_dataset,\r\n batch_size=self.args.train_batch_size,\r\n sampler=train_sampler,\r\n collate_fn=self.data_collator,\r\n drop_last=self.args.dataloader_drop_last,\r\n num_workers=self.args.dataloader_num_workers,\r\n )\r\n```\r\n\r\nnow if I type len(dataloader) this is 1, which is wrong, and this needs to be 10. could you assist me please? thanks \r\n","comment_length":26,"text":"wrong length with datasets \nHi\r\nI have a MRPC dataset which I convert it to seq2seq format, then this is of this format:\r\n\r\n`Dataset(features: {'src_texts': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 10)\r\n`\r\n\r\nI feed it to a dataloader:\r\n```\r\ndataloader = DataLoader(\r\n train_dataset,\r\n batch_size=self.args.train_batch_size,\r\n sampler=train_sampler,\r\n collate_fn=self.data_collator,\r\n drop_last=self.args.dataloader_drop_last,\r\n num_workers=self.args.dataloader_num_workers,\r\n )\r\n```\r\n\r\nnow if I type len(dataloader) this is 1, which is wrong, and this needs to be 10. could you assist me please? thanks \r\n\nAlso, I cannot first convert it to torch format, since huggingface seq2seq_trainer codes process the datasets afterwards during datacollector function to make it optimize for TPUs. ","embeddings":[-0.0898990035,-0.3192081451,-0.0323941223,0.6070456505,0.2765787244,-0.0336134471,0.479079783,0.07050246,-0.5328689814,0.2078109235,0.1901033223,-0.1201391816,0.0139271552,0.344342649,0.0675100163,-0.2707992196,0.0643086359,0.1787355542,-0.1016150042,-0.0089108553,-0.2274374962,0.1709845364,-0.2598226666,-0.0706524029,-0.3126019537,-0.0730075464,-0.1010056064,-0.0333041064,-0.3462432921,-0.3525754511,0.0661085322,-0.2506403625,0.3544397354,0.6040885448,-0.0001265039,-0.0722673237,0.1824869663,-0.110704042,-0.091110453,0.097599566,-0.2339126319,-0.0103232116,0.301758647,-0.2914406657,-0.0674708933,0.3816043437,-0.3033182621,-0.1190713048,0.1697501093,0.4160888791,0.1491683275,0.2589475513,-0.1338305324,-0.2041472197,0.2666983306,0.419282943,0.189833343,-0.146262005,0.0996487811,0.4892357886,-0.0897349492,0.1494981796,0.1966224909,-0.1416451186,0.1134943441,-0.0658493713,0.0303079206,-0.0960000977,0.3137546778,0.326462388,0.6222063303,-0.1899978817,-0.0739827827,-0.0441377163,0.0826853886,-0.575581193,0.1940187365,0.201002866,0.0677700862,0.0129151046,-0.6586116552,0.0756467208,0.0332969055,0.1146570817,-0.3273516595,-0.1786706448,-0.2108794302,0.1678252518,-0.0765730292,-0.3343451023,0.2114920765,0.1300551444,0.1223515794,0.185035035,-0.3168521225,0.0040680012,-0.369782269,0.3031198084,0.1807913929,-0.0312010814,-0.0071850996,-0.2769787014,0.165917486,-0.1649957448,-0.0832642689,0.2409600914,0.1471588612,0.0716321245,-0.0593911558,-0.1920505017,-0.074643679,-0.1838715076,-0.0941523612,-0.4005021155,0.1571927667,-0.0017074397,-0.0728077739,-0.170095548,-0.1529698968,0.0509631336,-0.05719845,0.0362973884,0.1326597631,0.1279755682,-0.0890798494,0.2195448726,0.0168394465,0.1168331206,-0.3445405066,-0.3399665058,-0.2275850326,-0.0942784473,-0.2037996054,-0.0455149338,0.0486436971,-0.3687188029,0.1643275172,-0.1625645757,0.0464205816,-0.2440051883,-0.3857925534,-0.0361463539,0.0173320472,-0.0301137101,-0.0740301833,0.1003653258,0.0298842061,-0.0682504699,-0.0919884145,0.1005262658,-0.7326493263,0.0935457647,-0.1725936085,0.0720311105,0.0175444838,0.1992681623,-0.1956207156,0.1166813746,0.5712836385,-0.0981904417,0.2525557876,-0.3171157837,-0.1953944564,-0.0876751989,-0.0140131786,0.0605668128,-0.5668563843,-0.2994161248,0.2563399076,0.0627233759,0.2786988318,0.5221871734,-0.0936203524,0.0594693609,-0.2201630771,0.0260044392,0.5655913353,-0.1227831244,-0.1180530265,0.456328541,-0.1290660203,-0.4169598222,0.1042512879,0.0699568316,0.4696896672,0.1280137002,0.2782408893,-0.038107384,-0.0918015167,0.0789657086,-0.2886632979,-0.1799527109,0.4170443118,-0.0638004839,0.0736436844,0.0247802678,0.1782239527,0.0828145519,0.3927258849,0.1072315499,0.2317766696,0.398463428,-0.1945101768,-0.2083268166,0.09484905,0.0719915852,-0.2795373201,-0.1646472812,-0.1662304699,0.1573649645,0.365321368,0.134268418,-0.2184961289,-0.0270041954,0.0401607342,-0.0757085681,-0.047881823,-0.2674982846,0.1338494867,-0.168636173,0.3453323245,0.3576398194,0.0782868639,0.4306086898,-0.5524819493,0.230559811,0.1218416616,-0.1312159151,0.0046947431,0.1191498041,0.4328717887,0.1964816153,-0.284879297,0.4029537141,0.0604498386,0.0718330666,-0.0674024373,0.0387820527,0.392415911,-0.3995311558,0.3362594247,0.0354069918,0.1262167692,-0.0613594986,-0.4830278456,0.1899389476,-0.0230089892,0.3436667323,-0.2495260984,-0.0264235549,-0.0104308603,0.2588877678,-0.3581843674,0.1657913625,0.2824555933,0.449891448,0.3601811528,0.6867608428,-0.2321612239,0.0525794998,0.437312603,-0.2694606483,-0.3443401456,0.3847820461,-0.2486208826,-0.1497918069,0.1073855981,-0.4560212791,0.1897209585,0.1604328752,0.0731342211,-0.2589393258,0.0459696464,-0.2219903022,-0.0669613555,0.0007336559,0.1374105364,-0.0228539389,0.3271223307,-0.2202792764,0.0269249808,0.0608229972,-0.0663886666,0.204864338,-0.4285590947,0.1715996116,-0.4512310028,-0.2724793553,-0.3519100845,-0.0021687944,-0.532020092,-0.0893990099,-0.0661174357,-0.3157937527,0.3861306906,0.0175728183,-0.1127113327,0.1099761799,0.4287566841,0.0775999054,0.0759254321,0.3114215732,-0.2205259204,-0.030512359,-0.0619396083,-0.0093869185,0.0510534048,-0.2923031449,0.0519282818,-0.1027170122,-0.4211384356,0.0676883683,0.0418813564,0.1687707752,0.3833267391,0.0160383321,0.0991705135,0.0103298062,0.0505603924,0.1298252195,-0.1057745516,0.3102925718,-0.1964636147,-0.0786110535,0.0432806462,-0.0585185587,-0.2840179801,0.0792290494,0.448312223,0.2354335934,0.1060388461,0.058264792,0.1695801467,-0.4198349416,0.1814240366,0.1239705235,-0.2222098559,-0.8714920282,0.137448445,-0.0258041043,-0.1211955249,-0.1383436471,-0.0529831238,0.2503461242,0.2581641078,-0.2691625953,0.1381452531,-0.2374102175,-0.123158589,-0.0599108078,0.1605795324,0.3572884798,-0.2368603945,0.1117625833,-0.0181454699,-0.2933224738,0.3579669595,0.0900374874,0.4553975165,-0.0026253827,0.4152847826,-0.0846623257,0.2809686065,0.3777345717,-0.0929544419,0.0658729225,-0.2463390678,0.1668543518,-0.1257380098,0.0782256946,0.4688602388,0.0359245427,-0.2177512646,0.4141396284,0.033404462,-0.1037968546,0.1519071311,0.0047098561,-0.4924779832,-0.1021351442,0.0319733694,0.1428110152,0.0719536617,-0.0539313443,0.2238917202,-0.2681477666,-0.2980812788,-0.3863811195,0.0775976628,-0.142179817,0.3162004054,-0.4841610789,-0.1177314743,-0.5206753016,0.1896814853,0.1766361147,0.2427069843,0.2643305659,-0.1345770508,0.1728814989,0.1135367975,0.6000973582,0.2969931066,-0.2703236639,0.1222210303,-0.050236959,-0.292277813,-0.1250489205,-0.0280728247,0.2035726309,0.0180919748,0.343769908,-0.3671685159,-0.1061794162,0.1143214628,0.0440746993,-0.3028909862,-0.269080162,0.0075236042,0.063634038,0.1795831174,-0.0644516796,0.036985673,0.4641848207,-0.2313703746,-0.202168107,-0.1667034626,-0.3244054914,0.2900685072,-0.0469802991,0.3544560969,-0.048290126,0.1201767251,0.0185980573,0.3662227094,0.430593431,0.2056291252,-0.3145063818,0.0388343967,0.3624874055,-0.3927770853,0.6368674636,0.4278537631,-0.2170654088,0.1135634184,-0.197917223,0.6306791902,-0.3744517565,0.1094826683,0.2282309979,0.042859219,-0.5182513595,-0.1454633325,0.0448171943,0.1073277593,0.1506494135,-0.3658346236,0.0322116278,-0.3529890776,-0.1641676128,0.1559338272,1.12180233,-0.157575503,0.3598313928,-0.078933388,0.1159812585,0.3529954255,-0.4284681082,-0.0613413788,-0.4853715301,-0.2979169488,-0.0348556451,-0.1774213165,-0.1271034926,0.259364903,-0.2782273889,0.532550931,-0.1582541317,-0.1435043961,-0.2566657066,0.1430366337,0.1784835309,-0.4158133566,-0.3069468141,0.0833541751,-0.1447827667,0.1517049372,-0.0714324713,0.087107338,0.0323585048,-0.3796269,-0.0573397018,-0.2712667584,-0.5653037429,-0.3887813687,0.3015840352,-0.1744165123,-0.0060836668,0.2656547725,0.1588515192,0.3715470731,0.0437523425,-0.0849592686,-0.0386399105,0.366751641,0.0405420624,0.1678374559,0.0041146255,0.1074033007,0.1929979771,0.1859679222,-0.013978296,0.0839177147,0.0472830012,-0.3048883379,-0.0860036537,-0.6301488876,0.0714238659,-0.2834022045,0.1399890929,-0.0646586493,-0.0034211262,0.3126320243,-0.1089058071,0.2990912497,0.2489946187,-0.3930512071,-0.2548384666,0.3060933352,0.225319609,-0.137170732,0.1721099317,0.2746470571,-0.1748773009,-0.0772933215,0.4727818668,0.2911938727,-0.3504849374,-0.0853228644,-0.0691030771,0.200354442,-0.0938858762,0.3242397606,0.1417705715,0.1458901614,0.082610786,-0.0999900401,-0.4517858028,0.0450847521,0.1475600302,0.3111539185,0.1525844634,0.2891599834,0.0875789076,0.0428053848,-0.158815369,0.0802423209,-0.1718364954,0.2328822166,0.036306262,-0.3175066113,0.0090584746,0.1880316585,0.0019684155,0.249470517,0.1264845431,-0.1226279214,-0.2269300073,0.1617534012,0.0589635521,-0.1850125194,0.1396538764,-0.0314571895,-0.0331321061,-0.214924112,0.0110802632,0.1028549522,0.3433351219,0.1548295766,0.493142128,0.0259574298,-0.0572643243,0.5009926558,0.4072121382,0.285082221,-0.045407135,0.2998177409,-0.0715335608,-0.0222387146,0.0152966874,0.4587609172,0.1614380181,-0.0692025498,-0.1355172843,-0.2911246717,0.0823559836,0.3094452322,0.3856016397,0.3570576906,0.0173756536,0.1095139533,0.0526956543,0.054413747,-0.0426371321,0.0549727753,-0.018383434,0.0882370099,-0.2388880998,0.503516078,-0.1174186766,0.2689978182,0.0345472097,-0.0692775548,0.2958171964,0.0733756721,-0.1585284323,0.2268845737,0.014632619,0.1815703809,0.3980835378,0.2304409146,-0.0904003829,0.2487923205,-0.136253491,0.2077596188,-0.1028727069,0.2748848796,0.0996890664,-0.1741975248,-0.4117302299,0.30647403,0.0449565277,-0.3348573744,-0.2127372921,0.6561852694,0.335527122,-0.4087522328,-0.3752572536,0.3956862688,-0.2118831724,-0.3872316778,-0.1680449396,-0.1247528642,-0.4670372307,-0.0503808483,-0.328350246,-0.1068392843,0.2790128887,0.1475426704,0.1368355751,-0.4126327932,-0.3528138697,0.1106700003,-0.2719160318,-0.2629649639,0.4174813032,-0.028082747,0.0383488648,0.5048876405,-0.1395744979,0.2498667091,0.0499683283,0.1436410099,-0.2169841826,0.0738121793,-0.0510925241,0.06596338,0.0085654892,0.1284972578,0.2257042974,0.24484846,0.0181406923,-0.0416092351,0.1521878392,-0.2556640506,0.0173903797,-0.5372476578,-0.0638975129,-0.1842152178,0.1272339374,-0.49594751,-0.0671940893,-0.5162888169,0.229468137,0.2347259969,-0.0822369233,0.3013238013,0.1103514731,0.0058021359,-0.1481136829,0.1696172655,0.2570337653,0.1164728701,-0.4964764714,-0.1613906175,-0.2868890464,0.012698045,-0.0468790233,-0.0013179349,0.0494354255,0.27580598,0.0087071536,-0.0306385048,0.1666119397,0.1644866318,-0.1197647825,0.2888931036,0.0099755665,-0.0915594324,-0.0375657603,-0.0318867937,0.0673323125,-0.0662945583,0.3562545478,0.0759863406,-0.1119587794,-0.3380556107,0.21354343,-0.0625086129,0.1419484913,-0.0070468192,0.137716949,0.1964765787,0.1568265259,0.1664333045,-0.0800620914,-0.1184216663,-0.2754455805,0.1171438619,-0.1821974963,0.2555265129,-0.1760876328,0.1999507695,-0.1618426889,0.1166886017,0.135652557,-0.0746061876,-0.3712223768,0.0905906558,-0.2492763698,0.0067337025,0.3864865005,0.133850798,0.1362559348,0.2491204292,0.1391144842,-0.3154530823,0.4479087889,-0.4412989616,0.0733994916,-0.2331730127,-0.1476220936,0.4630677402,-0.0930573121,-0.2554319501,0.0355448537,0.4829043448,-0.1332327574,-0.024043683,0.1865383685,-0.2772270143,0.0304163881,-0.3819780052,0.1503356844,0.17976062,0.0116301691,-0.0862030461,-0.4198220968]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/919","title":"wrong length with datasets ","comments":"sorry I misunderstood length of dataset with dataloader, closed. thanks ","body":"Hi\r\nI have a MRPC dataset which I convert it to seq2seq format, then this is of this format:\r\n\r\n`Dataset(features: {'src_texts': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 10)\r\n`\r\n\r\nI feed it to a dataloader:\r\n```\r\ndataloader = DataLoader(\r\n train_dataset,\r\n batch_size=self.args.train_batch_size,\r\n sampler=train_sampler,\r\n collate_fn=self.data_collator,\r\n drop_last=self.args.dataloader_drop_last,\r\n num_workers=self.args.dataloader_num_workers,\r\n )\r\n```\r\n\r\nnow if I type len(dataloader) this is 1, which is wrong, and this needs to be 10. could you assist me please? thanks \r\n","comment_length":10,"text":"wrong length with datasets \nHi\r\nI have a MRPC dataset which I convert it to seq2seq format, then this is of this format:\r\n\r\n`Dataset(features: {'src_texts': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 10)\r\n`\r\n\r\nI feed it to a dataloader:\r\n```\r\ndataloader = DataLoader(\r\n train_dataset,\r\n batch_size=self.args.train_batch_size,\r\n sampler=train_sampler,\r\n collate_fn=self.data_collator,\r\n drop_last=self.args.dataloader_drop_last,\r\n num_workers=self.args.dataloader_num_workers,\r\n )\r\n```\r\n\r\nnow if I type len(dataloader) this is 1, which is wrong, and this needs to be 10. could you assist me please? thanks \r\n\nsorry I misunderstood length of dataset with dataloader, closed. thanks ","embeddings":[-0.1484532058,-0.0579237342,-0.1097845137,0.424312979,0.1983342767,0.0155795468,0.6153778434,-0.0300615653,-0.3935883641,0.2385188788,0.2845693231,-0.1365799606,0.001870802,0.2772705555,0.0390551202,-0.1534984708,0.134595409,0.2068183869,-0.1075908169,-0.0113622406,-0.1992981732,0.0875401348,-0.366948992,-0.145993486,-0.2597512007,-0.1478260309,-0.102556251,-0.1015302986,-0.3941307366,-0.3376798928,0.0577428564,-0.1751493216,0.3658725917,0.6645886302,-0.0001230165,-0.0578815006,0.3222329915,-0.109741576,-0.1493114829,0.0369666144,-0.387855798,-0.0762600005,0.2585636973,-0.3467233181,0.1709084958,0.3988198936,-0.2273860276,-0.1813577414,0.0087325526,0.3765812516,0.1826709956,0.1304131299,-0.1490728855,-0.2224144638,0.2772220075,0.3844955266,0.214423731,-0.0511102714,0.1034295186,0.4382268786,-0.1440883577,0.1032710075,0.3008810878,-0.1005020365,0.0535923764,-0.0823540911,0.0923892856,-0.1085596681,0.3164989352,0.2760624588,0.7602072358,-0.1633217633,-0.1288762689,0.0829127282,0.0847775787,-0.477632612,0.1389883906,0.1863850057,-0.0113100065,-0.0631436408,-0.5548979044,0.0612442866,-0.0017479602,0.1209460571,-0.3987205625,-0.10894081,-0.2006841004,0.2249225527,-0.149554044,-0.3475795686,0.2297280878,0.0863837302,0.0785030723,0.1214743033,-0.3062303662,-0.0082161138,-0.3244349957,0.2960521579,0.1426718831,-0.0984225646,0.0209617484,-0.3210240006,0.0934583321,-0.1254028976,-0.0327420421,0.2314243913,0.206742987,0.0216468703,-0.1091681421,-0.1123837978,-0.1864562333,-0.1644041538,-0.090991199,-0.4017698467,0.3005324304,-0.0327762105,0.0876841396,-0.1163536608,-0.1849485934,0.059163671,-0.1180569381,-0.1281404644,0.0928879231,0.1090518534,-0.0819267854,0.2646772861,-0.2143784612,0.1211424023,-0.2698008716,-0.4232234657,-0.2757686973,-0.0290339328,-0.2308087796,-0.1767855436,0.0452850349,-0.3077953458,0.1341857016,-0.2097176164,-0.0314018428,-0.2520030141,-0.2990756929,-0.1713713259,-0.0428184643,-0.0825944692,0.0017072784,0.1013746709,0.0192014482,-0.1683187038,-0.0828836337,0.1461813748,-0.7466750145,0.14268969,-0.1942008436,0.1370267272,0.0257423893,0.1710135788,-0.1948990524,-0.0040943702,0.5577683449,-0.0989805013,0.2930517495,-0.2801348567,-0.0862714201,-0.1504052281,-0.1875623912,0.0717131048,-0.6944797635,-0.1421006322,-0.0186983366,-0.002764764,0.3534892499,0.367751956,-0.1084485054,0.1125148609,-0.1455992311,0.0537980013,0.5884771347,-0.0960233212,-0.244258672,0.5304946303,-0.1576711386,-0.4664910436,0.0664178655,0.1167112067,0.5823686719,0.069826901,0.2579508722,-0.1087369621,-0.0314114876,-0.0128550259,-0.2402967513,-0.2342473418,0.4487700164,-0.0893210024,0.1273365468,0.0030963803,0.2602439821,0.1019973159,0.4729665816,0.2255495936,0.2359513789,0.4539121091,-0.025095785,-0.233451426,0.1161466911,-0.0342524126,-0.2373930514,-0.1723991036,-0.1307647079,0.1349682063,0.3851307333,0.1429201365,-0.2457294017,0.069149673,0.0940936059,-0.0626366436,0.0293557849,-0.2779304981,-0.016406225,-0.203164503,0.3040094376,0.250376761,0.0797053874,0.3208247721,-0.4083916545,0.1794476807,0.1044740751,-0.0440459326,-0.0515855066,0.0251221154,0.3710131049,0.2719392776,-0.2542948723,0.4124064147,0.2074258029,0.0241739191,-0.0648721531,-0.0613224506,0.3062682748,-0.3542502522,0.4422667325,0.0412175916,0.1910346597,-0.0469219871,-0.4780450165,0.0049955556,-0.1459697485,0.3799421787,-0.1969213039,-0.016354477,0.1081055328,0.2022185028,-0.3464079201,0.0673714206,0.2420445383,0.4414482713,0.4111988246,0.6519747972,0.0189744625,0.0473885909,0.3859439492,-0.1815590709,-0.3802646101,0.3481154442,-0.3327201605,-0.1480311602,0.165692687,-0.2175584733,0.2755247355,0.2208128124,0.1265797764,-0.3419341147,0.0361300893,-0.252599746,-0.0504365787,0.0819518492,0.1996147037,-0.0294686053,0.2686838806,-0.2564893663,0.1022034287,-0.0488412157,0.0415717773,0.2532036901,-0.3895573914,0.0944827721,-0.4429273605,-0.2746573985,-0.2133897394,-0.0877320021,-0.5240041018,-0.1904933453,-0.2149902284,-0.3803806305,0.3409196138,-0.0067027393,-0.290663898,0.0094754677,0.4828657806,0.1591576189,0.2070003301,0.3013962805,-0.1760151088,0.0339865461,-0.0676308572,0.0538684875,0.0156066837,-0.1486205459,-0.0442979001,-0.1239488646,-0.3617945313,0.076264821,0.1137032285,0.2266464382,0.3722518086,0.0509699658,0.1533731818,-0.0176596548,0.0068283994,0.1280597299,-0.0489104614,0.2939337194,-0.1297620982,-0.0558201373,0.0952564552,-0.0059205876,-0.337638706,0.0478550382,0.3607355952,0.3372558653,0.0945205912,0.0809400976,0.1468465179,-0.3791817725,0.2955830693,0.0232088976,-0.261921078,-0.7850465178,0.1305869371,-0.0435286649,-0.1370524168,-0.074119769,-0.0495683402,0.2352404147,0.307741791,-0.3117691278,0.2146181613,-0.2207946777,-0.1075189635,0.0199037343,0.1233334988,0.3826772571,-0.2606689632,0.0816687867,-0.0359750092,-0.2120728791,0.3896808624,0.0563026406,0.473811239,-0.0323531814,0.4210056961,-0.0564873666,0.1040275097,0.3675329685,-0.0307814609,0.1001452133,-0.2040292025,0.1042713597,0.0039867386,0.0627705231,0.4992082119,-0.0074891923,-0.3693678975,0.4168255031,-0.0291753244,-0.2197219133,0.2108669132,-0.1458872557,-0.4274154603,-0.0441440493,0.0339580961,0.0950041562,0.0652790368,0.0652035326,0.1169416159,-0.3204629719,-0.3325731456,-0.4748093188,-0.017255852,-0.1644746214,0.2305848598,-0.5444819331,-0.0364839435,-0.4988295138,0.1410461664,0.0976005793,0.3117804527,0.3174705207,-0.1425162405,0.1248481199,0.1080343351,0.5405458808,0.1959581673,-0.1731981635,0.1482587606,0.0934848189,-0.3134649396,-0.1909338087,-0.1671455353,0.3289185166,0.0218242668,0.2301287651,-0.226148814,0.0936975777,0.2032451779,0.0004176538,-0.2948707044,-0.3821448684,0.0531099699,0.1475420147,0.2834562659,-0.1798297614,-0.0899539068,0.4330143034,-0.134327963,-0.2024238706,-0.1883527935,-0.3761198521,0.3070125878,-0.083556883,0.2912072241,-0.084912613,0.1250417084,-0.1701801717,0.3872897327,0.4588620663,0.2545301914,-0.3043520451,-0.0476749726,0.3966949582,-0.4657770693,0.6897573471,0.3189404607,-0.3215453625,0.1683869511,-0.1023211852,0.6286619306,-0.3371999264,0.1329188347,0.2432321012,0.075537771,-0.603275001,-0.1264387965,0.0726132467,0.0230544135,0.1127703935,-0.2686364353,-0.0872821882,-0.3265564144,-0.0902553946,0.2088122815,1.0579308271,-0.0075248145,0.2082349956,-0.0920748636,0.2177958339,0.3460125923,-0.4727055132,-0.163086772,-0.3817359507,-0.4100233316,-0.0458298512,-0.1943518966,-0.0440433957,0.3364103734,-0.327037394,0.470934242,-0.1049432158,-0.056302581,-0.1597361714,0.1243080571,0.0368216485,-0.325468123,-0.1136389673,0.1239988804,-0.1040468365,0.0698868334,-0.065938279,0.1320531219,0.019878922,-0.3468900621,-0.117053777,-0.2552345097,-0.5392011404,-0.2621690631,0.1596604735,-0.0273678172,0.0157307945,0.1416556537,0.0483190417,0.4225271344,0.024864262,-0.0626801699,-0.0416153371,0.2996298373,0.1173140556,0.1013159901,0.112327069,0.0697635189,0.0733677372,0.1829398721,0.0025293422,0.014791904,0.0297122486,-0.1606921405,-0.0425065607,-0.6389211416,-0.0849576518,-0.2471167445,0.0853012577,0.0070347083,0.0221668854,0.2646163106,-0.098661229,0.3483760059,0.2085373551,-0.3603860438,-0.2903544009,0.2889426649,0.3777514398,-0.1470654756,0.2214141488,0.1633721143,-0.1847082824,-0.1164518073,0.3959254324,0.2095462978,-0.293632865,-0.1052083895,-0.1568511724,0.3164390922,-0.1340877414,0.0985527784,0.2052953988,0.3107081056,-0.0064045899,-0.2152390182,-0.4231516123,0.0843211636,0.1008450687,0.2245872915,0.1011847034,0.2013315707,0.1403475255,0.161079511,-0.1880596429,0.019087173,-0.1720429957,0.1932395548,0.0477998741,-0.2853646278,-0.0162156206,0.1315483004,0.0543106571,0.2288517952,0.1696697325,-0.1570046246,-0.1630245,0.1405002773,0.0528855436,-0.134845823,0.2424912602,-0.0862341225,-0.0545063317,-0.1204436049,-0.211841315,0.0560938828,0.3162460625,0.1747783571,0.607683301,-0.0316643342,-0.1387172788,0.5717507601,0.3162107766,0.3496249318,-0.0544903055,0.4046169519,0.0885278583,0.0634054393,-0.1745188087,0.3727692366,0.0375237949,-0.0559214056,-0.0591221191,-0.3590408564,0.1745515168,0.2052376419,0.4466871023,0.3245345056,0.0394601896,0.1409817785,0.0945319459,0.1357515752,-0.1048972979,0.0678579435,0.0346491002,0.1042138785,-0.1952091455,0.4736631513,-0.1621290594,0.1958526969,0.0200464856,-0.0661888421,0.3169709444,0.0407026373,-0.1787770241,0.2456480265,-0.0661211982,0.1673233062,0.3628135622,0.3212835789,-0.061542213,0.3760654926,-0.0969486907,0.1546264589,0.0627940595,0.1737039983,0.0505728014,-0.1368220747,-0.3418731391,0.5096856356,-0.0561405793,-0.3550994694,-0.0862666965,0.476188004,0.4144007564,-0.1955530196,-0.3295387626,0.318669498,-0.2011193186,-0.3335978091,-0.274941802,-0.1082387939,-0.4984323382,-0.0980222598,-0.2993413508,-0.0680143759,0.1353637129,0.1702950448,0.0878846422,-0.3793101907,-0.2436901629,0.0841179863,-0.3221333027,-0.3107858002,0.3732687533,0.0340794586,0.1224880442,0.5273926258,-0.1263452917,0.3814272285,0.130577907,0.2179104239,-0.2215656191,0.1085121334,-0.1158867851,-0.0379200801,-0.0474159308,0.1914303005,0.2838411927,0.2647959888,0.0964494273,-0.0448536724,0.1864156425,-0.1223979071,0.0415183604,-0.4791247547,-0.0860648379,-0.0604880229,0.1173362061,-0.5198062062,-0.0817656666,-0.6760644317,0.2444154322,0.2900956869,-0.1814505458,0.2903415561,0.0431052931,0.0240049679,-0.2712117136,0.1566164047,0.2759867013,0.1976001859,-0.409560591,-0.1936157197,-0.3510187566,0.0224773921,-0.1896546483,-0.0825599059,0.0346439891,0.2906786799,0.0658441931,-0.0471232161,0.2675833404,0.1967502832,-0.2974972129,0.3155091107,-0.0626881644,-0.1351547092,0.0811152011,0.0894282013,0.0267988555,-0.1007116362,0.3999295831,-0.0046695354,-0.0968102738,-0.2991072536,0.1683850139,0.0905734897,0.1882615983,-0.0000646728,0.0806785002,0.2700633109,0.1117945164,0.1662617475,-0.1423642188,-0.0702782497,-0.2887332141,0.1540664732,-0.0391268432,0.2232566327,-0.2389199287,0.1789536029,0.0134316487,0.1113478392,0.0097515425,-0.0872367993,-0.4231829643,0.0366130061,-0.2132935077,0.0710528642,0.3272092938,-0.0300509892,0.168356359,0.2223884463,0.1496954113,-0.34641698,0.3372442126,-0.3589369655,-0.022790974,-0.2529606521,-0.1595913023,0.2846291363,0.0000601782,-0.1710284054,0.0287723169,0.5927268863,-0.1185143292,-0.0037051048,0.2228023708,-0.2039362639,0.0871348083,-0.3711521924,0.1214536875,0.0757028833,-0.0557018146,-0.0852737576,-0.3970738351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/915","title":"Shall we change the hashing to encoding to reduce potential replicated cache files?","comments":"This is an interesting idea !\r\nDo you have ideas about how to approach the decoding and the normalization ?","body":"Hi there. For now, we are using `xxhash` to hash the transformations to fingerprint and we will save a copy of the processed dataset to disk if there is a new hash value. However, there are some transformations that are idempotent or commutative to each other. I think that encoding the transformation chain as the fingerprint may help in those cases, for example, use `base64.urlsafe_b64encode`. In this way, before we want to save a new copy, we can decode the transformation chain and normalize it to prevent omit potential reuse. As the main targets of this project are the really large datasets that cannot be loaded entirely in memory, I believe it would save a lot of time if we can avoid some write.\r\n\r\nIf you have interest in this, I'd love to help :).","comment_length":20,"text":"Shall we change the hashing to encoding to reduce potential replicated cache files?\nHi there. For now, we are using `xxhash` to hash the transformations to fingerprint and we will save a copy of the processed dataset to disk if there is a new hash value. However, there are some transformations that are idempotent or commutative to each other. I think that encoding the transformation chain as the fingerprint may help in those cases, for example, use `base64.urlsafe_b64encode`. In this way, before we want to save a new copy, we can decode the transformation chain and normalize it to prevent omit potential reuse. As the main targets of this project are the really large datasets that cannot be loaded entirely in memory, I believe it would save a lot of time if we can avoid some write.\r\n\r\nIf you have interest in this, I'd love to help :).\nThis is an interesting idea !\r\nDo you have ideas about how to approach the decoding and the normalization ?","embeddings":[0.0579881594,0.1647051126,-0.0206481814,-0.0999954194,0.2292947173,-0.0559269451,0.2166689336,0.4347034395,-0.2473879009,-0.1617708057,0.0097895823,-0.0496323295,-0.1669543087,0.2896748781,0.2092278302,0.1215441376,0.1424438655,0.1846169084,-0.4095783532,-0.067763336,-0.2962715328,-0.0467055179,0.2424422503,-0.1144291312,-0.3984666467,-0.0129985576,-0.2188078761,0.2230956256,0.0136245247,-0.18215774,-0.3285973668,0.3551388979,0.1186941713,0.0364232846,-0.0001102374,0.0145410877,0.1038492545,-0.0967335254,0.197291404,0.2823079526,-0.4415962994,-0.0208728202,-0.3179782033,-0.2204544097,0.2632786334,-0.1618583351,0.2451173812,-0.1386906505,0.3337541521,-0.2657234073,0.2161934227,-0.4380463958,-0.2725088,0.3931819797,0.175028801,0.1780483276,-0.0581744425,-0.0528199486,0.1193539649,0.3262684047,-0.3176146746,0.5141986609,0.1687352955,-0.3415008485,0.1522488445,0.000400933,0.3446982205,-0.0906696543,0.2288550436,0.1792102158,0.1403950006,-0.2407236397,-0.2909703553,0.0616342872,-0.2266462445,-0.0471751541,0.3117389679,0.0513948798,0.1743303239,0.2972970605,-0.8323922157,-0.3791384697,-0.0001243489,-0.2025853395,0.1313333511,-0.172271058,0.0970596224,-0.1211425066,0.2808828652,-0.0671150014,0.3315512836,0.1234548613,-0.0365065522,-0.2305438966,0.0620265193,-0.1533109248,-0.4224923849,0.5188221335,0.2395397723,0.3001599312,0.1627902389,0.308688134,0.1106365621,-0.0680961609,-0.0554917417,0.3370959759,-0.155349791,-0.0187082849,0.2099351734,-0.2234660834,-0.3734851182,0.1732573509,-0.0504439436,-0.2014845908,0.3005166352,0.3068946004,-0.1599647552,-0.2409578413,0.5560629368,-0.1805860251,-0.1228703186,0.1204524264,0.1737584323,0.0580147542,0.0289724879,0.3135021627,0.0868721083,-0.1348686814,-0.066814445,-0.0665935501,-0.1691817492,-0.0237779636,-0.3316341639,0.2602967918,-0.0459179804,-0.2559429407,0.0559978038,-0.0472614802,0.2711273134,0.1667011678,0.1988694072,0.1367662698,0.5526273847,0.0959468484,-0.1163153946,-0.138270691,-0.1938530505,0.0865323618,-0.23593539,0.115888685,-0.1936692744,-0.2252523452,0.3494835794,0.1478482634,-0.2709454,-0.0909559503,-0.079563871,0.0610845424,0.2011087835,-0.0352455787,0.0708694085,0.0607461818,0.0380195044,-0.3379440308,-0.0195818879,0.2535013855,0.0997974277,-0.000055283,0.3916781247,0.1493162066,0.1937788576,0.4858622551,-0.1002652645,-0.1992478818,-0.0508004837,0.1779374778,0.0069970177,-0.3228585422,-0.1532531828,0.4586922824,0.0817386732,-0.0321356691,0.2724073529,0.2414174527,0.1365738958,-0.5183570981,-0.3733369112,0.084889099,-0.226559788,0.1455516666,-0.4432271123,-0.543218255,0.2205468267,-0.1168504581,0.1406552792,0.0731808245,0.0739090294,0.0928900391,-0.0259612594,-0.4790357649,0.1978315264,-0.0261201076,0.0933567435,-0.1285909563,0.1191432551,0.4896504283,0.0705787092,-0.0535092242,-0.4647884667,-0.3019529581,-0.2511307001,-0.3260699213,0.3761866689,-0.4562111795,-0.1690853238,0.1202602684,0.1046998277,-0.0504509099,0.0656179786,-0.2368854433,-0.1092494279,0.2331489027,-0.0064952155,-0.1074649245,-0.4308805168,-0.0559862182,0.0417532511,0.2098347396,-0.0880730078,0.1034951583,0.0917723477,-0.26852718,0.1405608803,0.2021124363,-0.2694943845,0.4485934675,0.4598626494,0.7957437038,0.2327311784,-0.3012223542,0.6025997996,-0.0764033273,-0.1474384218,-0.090851374,0.1016919985,0.3585917354,-0.2511354685,-0.087358579,0.0913746879,-0.1395505667,0.107802093,-0.4175278842,-0.3176829219,-0.2822001576,-0.1522046775,-0.1216212139,0.1933118701,0.3982318938,-0.0867917612,0.2805507183,0.2394749224,-0.1773470044,0.0103848455,0.2432071567,0.0836632326,-0.4002476335,-0.2816604674,0.2078521848,0.0816142932,0.2223655879,0.2411377132,0.2849276364,0.1991673857,-0.1477099359,0.1926002353,-0.0074547324,-0.0329275019,-0.0549311601,-0.1868720502,-0.0027007696,-0.2573947608,-0.0729503185,0.11498449,-0.1399638504,-0.2359175533,0.0989863053,-0.4986063838,-0.0223356523,-0.1428444684,-0.192379877,-0.0621933565,-0.1301116943,0.2478783876,-0.1797421873,-0.2955242395,0.228232339,-0.1217207089,0.2037386745,-0.0943714827,0.0076093031,-0.2146125883,-0.2665739059,-0.0174300019,0.0581853911,0.3697447181,-0.2563674748,0.1824273318,-0.6005426645,0.225244984,-0.4703187943,-0.5034909248,0.0603678264,-0.1038346738,-0.1646051705,0.1006126106,-0.1556524634,-0.3723519146,-0.1416883618,-0.1298087835,-0.2227977961,-0.172168389,0.153026402,-0.039013356,-0.1300099194,0.0559994765,-0.1785848439,-0.1654712707,-0.0548216924,0.1365204901,-0.1140015423,0.239343062,0.2022376657,-0.0854616389,-0.1993618757,-0.2033474296,-0.1849959344,-0.3022694588,-0.9126803279,0.4630579054,-0.2917784154,-0.1275364161,-0.0220049582,-0.0933089182,0.0496542975,0.5415793657,-0.2904457152,-0.4510396123,0.080108121,0.0154578378,0.1255819649,0.0304481164,0.1202549413,0.2174690962,-0.1953010857,-0.0381361023,0.3975361586,0.3179323375,0.036382962,0.3993934989,0.2605863214,0.0236946009,0.3085049987,0.5524952412,0.2742907703,-0.155865252,0.0450664349,0.2183421999,0.1625771672,0.107157886,-0.0146386158,-0.0329932161,-0.1554430425,-0.074449487,0.4130132198,0.1230946183,-0.2062533349,0.0986750945,-0.146907419,-0.3558013141,-0.4585825503,0.4402860701,-0.2552873194,0.2209807336,-0.151128903,-0.2410647869,-0.1631073952,-0.0763136297,0.5368345976,0.2171967179,0.1238927543,0.1359716356,-0.02647366,-0.1222635061,-0.2475917935,0.2743958533,0.3565450013,0.1777852625,-0.0094023906,-0.1717475653,0.038613271,-0.0493147373,0.429173708,-0.2685380876,-0.0485619195,-0.1228233799,-0.5765990019,0.5129473805,0.2474611551,0.0935683772,0.0458177403,0.3048223853,0.1030102819,-0.0661193952,-0.0591352433,-0.1869179904,0.2788245082,-0.1313562095,-0.027404489,0.127781406,-0.3672358096,-0.0058547547,-0.0968874469,0.0139896479,-0.360583365,0.1340368837,-0.2099425793,-0.3357189,-0.1878918111,0.2707710266,0.1370663643,0.033573702,0.0128287124,0.1681177169,0.2732578516,0.2485530078,0.1062948629,0.3191768527,0.0693296269,0.0574338585,-0.2655369937,-0.1567774415,-0.1749814004,0.441670388,-0.1119564623,-0.3950529695,0.3257461786,0.1138638109,-0.1574765742,-0.2019823939,-0.3884449899,0.0846869275,-0.4554491937,-0.1692981869,0.3722611368,0.143645823,-0.2350661457,-0.0437314399,-0.020939637,-0.3833583295,0.1587499082,0.5202263594,1.2037165165,-0.149846524,0.1079215482,-0.2164003104,0.0236983169,-0.2174625546,-0.2415346205,-0.0166024845,-0.2205498666,-0.4065418541,-0.2070002109,0.0081875622,0.1574708819,0.1322930306,-0.3657614887,0.2342297137,0.1607875079,-0.2798333168,0.05224327,0.203773886,-0.1822733879,-0.1449738145,0.1952240467,0.1249397397,0.113779597,-0.3169341385,0.0910163596,-0.4105030298,-0.0143835079,0.1019921899,0.4724983871,-0.0261833593,-0.485563457,0.2704308033,-0.1951266378,0.0226759817,-0.0404835902,-0.1540570855,0.6115783453,-0.0070080999,-0.1267042309,0.0630144849,0.3266372979,0.3293948174,-0.0012059201,-0.2020390779,0.4494643509,0.0397719853,0.1162980869,-0.3359423578,0.295109868,-0.3244663477,-0.1815893203,-0.2578531504,-0.242523402,-0.0376977287,0.3148887157,0.1724087,-0.4297751188,0.1186883226,0.1001664326,0.3503258228,-0.0267004445,0.6805775166,0.253706634,-0.1376314461,0.007445375,0.2244800925,0.3333892822,-0.0904204398,0.0790440664,-0.3230108619,-0.3410330117,-0.1152855679,0.2726310194,-0.2552440763,-0.2441884577,0.1347500831,-0.2213781476,-0.1266428381,-0.0676683038,-0.074829489,0.2005329281,0.1958202124,-0.1287343502,0.0309505351,-0.4009748399,0.3730387688,0.0043665422,0.3889574707,-0.0117394049,0.0754855052,-0.1265766472,0.2835724354,-0.3309337199,0.1714204997,-0.0394417867,0.2735458314,-0.2536267042,-0.3610468507,0.21200867,-0.1412412971,0.0559566319,0.1845545471,-0.2429118454,-0.1468961686,-0.1379620582,0.142119661,0.092569381,-0.1376426965,-0.0097228149,0.2268533558,0.0682705343,-0.0792328119,0.0784937367,0.2135468125,0.0107645895,0.0031717846,0.0671271682,-0.4323627651,0.0211170893,-0.0780686587,0.1021715626,0.1335618943,-0.0566442646,0.1140898094,-0.1146933809,-0.2190972567,0.172509253,0.3052981794,0.1749486476,-0.0663634837,0.3501424491,0.2496856451,0.017591862,0.5870927572,-0.0496079735,-0.0747155994,-0.1433601826,-0.4321393073,0.3590959907,0.1659306586,-0.1730496436,-0.1153542548,0.0590402111,-0.1240815148,-0.3018915355,0.3146198988,0.4946980774,-0.2148975283,0.1783946157,-0.1635513455,0.2595250905,-0.4596236646,0.0658154264,0.116557382,-0.2677550316,-0.0403200909,0.1498768926,0.3782489002,-0.0656573027,0.2406570613,-0.2567142248,0.3606832027,0.074513711,0.3942448497,-0.1373748779,0.0727260485,-0.1543373615,0.2562468946,0.0385161862,0.3209947944,0.2127822936,0.3322633207,0.3905766606,-0.2104560137,-0.0217797887,0.6193832159,0.0776275843,-0.2505090833,0.1396140009,-0.003626422,-0.440571934,0.069588989,-0.0663807839,-0.1549735218,0.4601646662,0.5837759972,0.3065883219,0.0159228146,0.1993261725,-0.1872242093,0.5381706357,-0.4364183545,0.1084178463,0.3436721265,-0.1181031317,0.3846259117,0.1986009777,0.1219824925,0.041189108,-0.1795731932,0.3040752113,-0.0543882623,0.0911794081,-0.1364307553,0.1980258226,0.0535160527,-0.1056820974,0.0543945096,0.0945077613,-0.0241560526,0.0730996951,-0.1337474734,0.0583433025,0.3379209042,0.380582273,0.3612347245,0.0051964116,0.0338188931,0.0447300114,-0.3965477049,-0.1949443221,0.1841637194,-0.4594890475,0.0011475988,0.0883102864,0.0899411365,-0.0223617591,0.500944376,0.3568814695,-0.2505388856,-0.3136978447,0.1042905599,0.077921249,0.0152800241,0.2501473427,0.0461634472,-0.077408649,0.1467861384,-0.0443832651,0.2367869318,-0.0869134516,0.0266541019,-0.0599949993,0.1972402632,-0.0571915172,0.5674791932,0.1685570627,0.3000094593,0.1052265689,-0.2551064789,0.3098328114,0.2597007751,0.1014342308,0.1046239063,-0.130917713,-0.0986829326,0.0510913543,0.195315212,0.2842186093,-0.0392181873,-0.2065333575,-0.0410738178,0.0114777638,-0.2477720231,-0.0681463033,0.145037815,0.3426802754,-0.0684208795,-0.158384487,-0.1008010656,-0.0749195814,0.1233981028,-0.2849733829,0.1807679236,-0.209704563,0.2057795227,-0.1874547452,-0.0099780327,0.0302267969,0.5391848087,-0.0342181772,0.0706415325,-0.1243440509,-0.4154353738,0.4183195233,-0.4434714019,-0.0425947495,0.1918369532,0.2244072407,0.3125598133,0.0835568905,-0.4205691814,-0.0332431793,-0.1036072224,-0.0892922282,-0.0221995525,0.1637200564,-0.1072976887,-0.2310145944,-0.1094599888,-0.2687176466,-0.0031997887,-0.0363401584,-0.2746324241,-0.3152399659]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/915","title":"Shall we change the hashing to encoding to reduce potential replicated cache files?","comments":"@lhoestq\r\nI think we first need to save the transformation chain to a list in `self._fingerprint`. Then we can\r\n- decode all the current saved datasets to see if there is already one that is equivalent to the transformation we need now.\r\n- or, calculate all the possible hash value of the current chain for comparison so that we could continue to use hashing.\r\nIf we find one, we can adjust the list in `self._fingerprint` to it.\r\n\r\nAs for the transformation reordering rules, we can just start with some manual rules, like two sort on the same column should merge to one, filter and select can change orders.\r\n\r\nAnd for encoding and decoding, we can just manually specify `sort` is 0, `shuffling` is 2 and create a base-n number or use some general algorithm like `base64.urlsafe_b64encode`.\r\n\r\nBecause we are not doing lazy evaluation now, we may not be able to normalize the transformation to its minimal form. If we want to support that, we can provde a `Sequential` api and let user input a list or transformation, so that user would not use the intermediate datasets. This would look like tf.data.Dataset.","body":"Hi there. For now, we are using `xxhash` to hash the transformations to fingerprint and we will save a copy of the processed dataset to disk if there is a new hash value. However, there are some transformations that are idempotent or commutative to each other. I think that encoding the transformation chain as the fingerprint may help in those cases, for example, use `base64.urlsafe_b64encode`. In this way, before we want to save a new copy, we can decode the transformation chain and normalize it to prevent omit potential reuse. As the main targets of this project are the really large datasets that cannot be loaded entirely in memory, I believe it would save a lot of time if we can avoid some write.\r\n\r\nIf you have interest in this, I'd love to help :).","comment_length":191,"text":"Shall we change the hashing to encoding to reduce potential replicated cache files?\nHi there. For now, we are using `xxhash` to hash the transformations to fingerprint and we will save a copy of the processed dataset to disk if there is a new hash value. However, there are some transformations that are idempotent or commutative to each other. I think that encoding the transformation chain as the fingerprint may help in those cases, for example, use `base64.urlsafe_b64encode`. In this way, before we want to save a new copy, we can decode the transformation chain and normalize it to prevent omit potential reuse. As the main targets of this project are the really large datasets that cannot be loaded entirely in memory, I believe it would save a lot of time if we can avoid some write.\r\n\r\nIf you have interest in this, I'd love to help :).\n@lhoestq\r\nI think we first need to save the transformation chain to a list in `self._fingerprint`. Then we can\r\n- decode all the current saved datasets to see if there is already one that is equivalent to the transformation we need now.\r\n- or, calculate all the possible hash value of the current chain for comparison so that we could continue to use hashing.\r\nIf we find one, we can adjust the list in `self._fingerprint` to it.\r\n\r\nAs for the transformation reordering rules, we can just start with some manual rules, like two sort on the same column should merge to one, filter and select can change orders.\r\n\r\nAnd for encoding and decoding, we can just manually specify `sort` is 0, `shuffling` is 2 and create a base-n number or use some general algorithm like `base64.urlsafe_b64encode`.\r\n\r\nBecause we are not doing lazy evaluation now, we may not be able to normalize the transformation to its minimal form. If we want to support that, we can provde a `Sequential` api and let user input a list or transformation, so that user would not use the intermediate datasets. This would look like tf.data.Dataset.","embeddings":[0.1285304129,0.0863642022,-0.0490874425,-0.0772112384,0.1253058016,-0.0536653809,0.2214981318,0.4482245445,-0.1249614656,-0.0400376767,-0.1387324035,0.0334732234,-0.1281396449,0.3592914343,0.2445239872,-0.0045850868,-0.0214299057,0.1100231335,-0.4076034129,-0.0112218559,-0.2559686899,-0.0132378452,0.1368180662,-0.0337848403,-0.2687934041,-0.0806391537,-0.1797393262,0.1853831559,0.1048366651,-0.1334937662,-0.3031501472,0.4466129839,0.1472111642,0.2794195116,-0.0001038101,0.0562618412,-0.2488046438,-0.138645649,0.2129288614,0.0744406283,-0.2611119747,0.0785090327,-0.3203867972,-0.2040926963,0.0010415929,-0.2220731676,0.1788083911,-0.0459423847,0.374322623,-0.0652710795,0.2467910647,-0.2013984919,-0.1129855141,0.3312351406,0.2261059135,0.1098318473,-0.1199187562,-0.1880589873,0.1244985312,0.1867548376,-0.217949003,0.5250282288,0.1326779872,-0.3111564815,0.227358982,-0.0156110916,0.1431379169,-0.1239111125,0.0056162598,0.4643925726,0.222835198,-0.270663321,-0.4482094049,-0.0549162105,-0.1963089854,-0.0430078059,0.2514185607,-0.018924946,0.282205224,0.1687414795,-0.5559618473,-0.1699756831,-0.0533313863,-0.110022001,-0.0684079006,-0.0438300967,0.1227401197,-0.122079201,0.0873712301,-0.1470597535,0.2502611578,0.0586046949,0.1833016425,-0.1270232648,-0.1814170927,-0.3422726095,-0.2197387218,0.3131836951,0.2802988589,0.2581949234,0.2559705973,0.2266033292,-0.0624806546,-0.1181505173,-0.1766685396,0.3222263157,-0.3438585401,0.0889629573,0.133767426,-0.2756940722,-0.3974499404,0.1466489136,-0.0085868202,-0.2518129647,0.2226762921,0.282235533,-0.196213007,-0.0995170921,0.4083944261,-0.2009485215,-0.0107703712,0.1300056279,0.2449009269,0.1339378655,-0.0542051569,0.2221426368,0.2067654282,-0.1895075887,-0.0806853324,0.1101926267,-0.1272460222,-0.0858556852,-0.2488651872,0.1876550317,-0.0048876973,-0.1962255687,0.0158583559,-0.0835492313,0.2741125226,0.2346903086,0.3042016327,-0.0417356379,0.4640479684,0.1292725503,-0.0601887666,-0.2695890963,-0.1872320771,0.1579571962,-0.3326706588,0.1288958192,-0.2424294502,-0.2861660421,0.2012259066,0.2426214367,-0.2469905615,-0.1991754472,-0.0448270142,0.0969924554,0.2532232106,-0.1112782061,0.0230899993,0.0169687066,-0.1419422626,-0.3485346138,-0.1842769384,0.2687899768,-0.1900322884,0.124914892,0.3478443921,0.1490198076,-0.090565674,0.3433871269,-0.1107633784,0.0102250446,-0.0510068089,0.2945253551,0.2158773541,-0.3417193294,-0.0383224413,0.3436824679,-0.1598877162,0.0314034075,0.2459237128,0.2952590883,0.2183941603,-0.6586913466,-0.268019855,0.115048714,-0.1457945853,0.2035927474,-0.3782372773,-0.4893104732,0.1551601738,-0.111020565,0.2189465016,-0.0219094958,-0.0264859423,-0.0243930183,0.0211042333,-0.3764765263,-0.009697373,-0.020175932,0.0328957848,0.186211288,0.1014084369,0.2024668306,-0.2186006755,0.1092706621,-0.2190377265,-0.2866873741,0.0303831175,-0.3408524692,0.2135965526,-0.3898274601,-0.0898485854,0.2973106802,0.1639081538,0.0066269063,0.0824806914,-0.3306164443,-0.2006497085,0.1946838051,-0.160166949,-0.1434936672,-0.2852327228,0.0684343204,0.0442206115,0.2410115898,0.041320201,-0.0341020711,0.0668013021,-0.2658017278,0.1290796995,0.317902267,-0.1664431542,0.3309113979,0.2668876648,0.8330717683,0.1407028884,-0.1076701656,0.343884021,-0.084625639,-0.1168700531,-0.0055886093,-0.0805219635,0.4911959469,-0.3341499567,-0.1171824858,0.1367959082,-0.1599956751,0.0559728444,-0.372017473,-0.1977209896,-0.2793736756,-0.0932421237,-0.106302537,0.1200073138,0.3791052997,-0.2798251212,0.4305251539,0.4273661077,-0.1499266326,0.0294489134,0.2169590145,0.2922954559,-0.1389044821,-0.2679871917,0.2191865891,0.1560052037,0.3948587179,0.1579460502,0.0696535185,0.1232545972,-0.1453303844,0.2383618653,-0.1427056044,0.0940891653,0.1048287377,-0.020677615,0.1556965411,-0.3001692891,-0.1245947927,0.1747216582,-0.0403483659,-0.3369621634,0.0005104129,-0.3786216378,0.0028127201,-0.0328971222,-0.2066309601,-0.0539776199,-0.1642453223,0.4976658225,-0.1089979783,-0.3371109664,0.2331423312,-0.2194539607,0.1465765834,-0.0242132917,-0.1001765579,-0.2513731122,-0.1650529355,-0.0147685688,0.1514493525,0.4425896406,-0.1307712495,0.3409085274,-0.3635287881,0.0824580565,-0.3977570236,-0.6287537813,0.0700297505,-0.2315190434,-0.0788373277,0.1244051754,-0.0712286159,-0.2253907174,-0.0993884504,-0.1441022903,-0.2221812606,-0.3359176517,0.0935992822,0.011735484,-0.0142571712,0.0353115685,-0.2946578264,-0.1908772588,-0.1841779798,0.0517575778,0.1154505312,0.3753777146,0.1448075175,-0.0652874708,-0.2465345562,0.0321274027,-0.0179093629,-0.3294393122,-0.8471505046,0.472307235,-0.3325072825,-0.1874354631,-0.0672622472,-0.0376644544,-0.1309810728,0.5486729741,-0.3638745248,-0.6462661624,-0.000509608,0.1221546307,-0.0092846481,-0.0365933031,0.4072078168,0.1895595044,-0.2352834791,-0.1316512972,0.3807606995,0.2739590704,0.0604361072,0.1967510432,0.1090952232,0.161050126,0.3650012314,0.6264784932,0.1853601784,-0.2784263492,0.0466128513,0.0675854459,0.1245018691,-0.0798213631,-0.012351362,-0.0082908291,-0.015417912,0.1357920319,0.3944532275,0.1157402694,-0.0154705187,0.1152288392,0.1230582818,-0.4963377714,-0.5042803884,0.3902656138,-0.31130898,0.2560547292,-0.039640557,-0.2495987266,-0.3893629611,-0.1653741151,0.4180832207,0.2367179394,0.2869227231,0.0459790044,0.0231366716,0.0878517702,-0.1084378809,0.2726279497,0.3196282089,0.0213305224,0.043189317,-0.2385331988,-0.0072184294,-0.2276932448,0.6021747589,-0.1238812059,0.0670818835,-0.066582717,-0.5419083834,0.3541778326,0.2271940708,0.1421805173,-0.1321707815,0.1844017804,0.3718284369,-0.2227366716,-0.2264919579,-0.083950758,0.1937793791,-0.1486981362,-0.0568809174,0.0475336611,-0.3485065401,-0.0548108779,0.0101911789,-0.0789896324,-0.4550441802,0.105661355,-0.2257423252,-0.3692721426,-0.2502522469,0.2358340025,0.1980257779,0.0366616286,0.0707150474,0.1968159229,0.2124982625,0.1011148542,0.0640045702,0.254614234,-0.036829561,0.1430101544,-0.3102518618,0.0349277556,-0.3283665776,0.4785364866,-0.1703670919,-0.2068730891,0.1241940036,0.2493243217,-0.1271365285,-0.1288752407,-0.2221632004,-0.0225475058,-0.3139002621,-0.1667636633,0.3888576627,0.1422588676,-0.2729968131,0.0919478536,0.1733016819,-0.3826983571,0.256011337,0.542014122,1.1761375666,-0.3763305545,0.1005102098,0.0699510202,0.0234090779,-0.0114784772,-0.2966814935,-0.030650625,-0.2049043328,-0.4842042923,-0.223287195,0.0384871662,-0.0360253267,0.2198575288,-0.2503261566,0.1636202633,0.1835589111,-0.119072035,-0.0428816825,0.2936108112,-0.1004273221,-0.2144843936,0.0220693555,0.2072319537,0.1814972758,-0.3016629815,0.0739468411,-0.3371147215,0.0196941905,0.126095131,0.3993551731,-0.0989341959,-0.5738697052,0.1900018007,-0.1759685576,0.0322094932,-0.0943492204,-0.0423776209,0.485293448,0.0370614231,-0.1211243197,0.0495822504,0.366206646,0.3474068046,0.001646137,-0.2287258059,0.5342909694,-0.029244896,0.0950183123,-0.3131179512,0.3360684216,-0.3367793858,-0.2795274556,-0.2785630226,-0.1608463079,-0.080255568,0.3283863366,0.188047722,-0.4775858819,0.0487525426,0.1665292829,0.1995808184,-0.0874186903,0.6254339218,0.1361822337,-0.0647733808,-0.0498282984,0.2343391478,0.2947429121,0.120268777,0.1588822007,-0.0883425027,-0.3206319809,-0.2068024427,0.1803953201,-0.2727230191,-0.2477396429,0.249517113,-0.1612062156,-0.1212345511,0.1711607873,0.1120508984,0.3005467653,0.0668950304,-0.0554795712,0.0352399945,-0.3055217564,0.3674727678,0.1737472266,0.4600826502,-0.1467991322,0.434343487,-0.2452533394,0.3762705922,-0.451000005,0.1424838156,-0.1463257074,0.3262868524,-0.1900706142,-0.2824577987,0.1221834943,-0.1587542295,0.1252053678,0.1285536885,-0.3576830626,-0.2642920613,-0.1343515068,0.1233718023,0.0929591507,-0.0617202781,-0.0241252799,0.0406206921,-0.0930192396,-0.0773363113,0.1555429995,0.2089787573,0.0028009382,-0.0521772057,0.0783906057,-0.4736968875,0.160030961,-0.0281404983,0.1240977794,-0.104107596,-0.0500571467,0.2096000314,-0.1825319231,-0.1774723083,0.1890350282,0.3020366132,0.2633102238,-0.1904022545,0.2855670154,0.1765292436,-0.1778228283,0.4587361515,0.1259514689,0.0384658426,0.0009790672,-0.4334667325,0.2604823709,0.2351166755,-0.4202830791,-0.0899146125,0.120500721,-0.065469265,-0.0204907209,0.2778393328,0.2072882354,-0.0317331329,0.1610590965,0.0001042478,0.1213791072,-0.5231853127,0.0557380095,0.1909553558,-0.2242421657,-0.039082028,0.0846140385,0.2472977787,-0.0045035826,0.1397808343,-0.0607760623,0.410654217,-0.2522783577,0.2449948639,-0.0482307114,-0.2053447813,-0.0295081455,0.2851448059,0.0826415867,0.2242777199,0.3080564737,0.0737210065,0.4339232743,-0.2171858102,-0.0726332963,0.6125522852,0.0919124112,-0.2435761094,-0.0462825075,-0.0841785595,-0.2197406888,0.0218678098,-0.0159906074,-0.1172394529,0.3299681246,0.5798950195,0.3886064291,-0.1233007461,-0.0035299519,-0.2381326556,0.5514826775,-0.3509506881,0.1697096825,0.3621652722,-0.2181718796,0.2207414806,0.242424652,-0.0696321651,-0.0261439867,-0.1438602805,0.3785850108,-0.0473928452,0.1139521301,-0.0769383609,0.1688647717,-0.0400084481,-0.1288118362,0.2110714912,0.1555361301,-0.1085625812,0.0934372097,-0.1807086766,0.0129310302,0.0931245163,0.3506374657,0.3761066496,-0.0662326962,0.0498140417,0.030940881,-0.2851729393,-0.0742973834,0.1877622604,-0.3989115059,0.1143961549,0.0720774382,0.130342707,0.07160303,0.4829993844,0.3806512356,-0.2432487458,-0.3473899364,0.1224918887,-0.1109221503,-0.0021116834,0.0821843743,0.1023038253,0.03736699,0.2220949382,0.1007130593,0.1974875629,-0.0495860428,-0.0793667957,-0.1267407238,0.3335696459,-0.1618368477,0.3978960514,-0.1369896531,0.1682547033,0.0589105897,-0.3917449713,0.1069098637,0.2157736868,0.1770469248,0.0063038962,-0.0130821858,0.0348612331,0.1745424122,0.1908226311,0.1926882118,-0.1297958493,-0.2064408958,0.1018957123,0.0281661674,-0.1197159067,-0.3270350099,0.1273053288,0.3439900279,-0.0424506329,-0.1253280491,-0.2214300334,-0.0630225912,-0.1344345063,-0.2100063115,0.193084076,-0.1583124846,0.1568442881,-0.0833464935,-0.0139825298,-0.0066034961,0.6900027394,0.0189206079,0.203186959,-0.2757936716,-0.368040204,0.5384502411,-0.4374366999,-0.0583665967,0.1495702863,0.2244147807,0.2603414059,-0.0162786916,-0.3034524918,0.0596203767,0.0519409366,-0.0623714663,0.0690092966,0.2875989079,-0.068038106,-0.2214598507,-0.0995096043,-0.133329764,0.0159841981,-0.1004358232,-0.3943923414,-0.2757805884]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/911","title":"datasets module not found","comments":"nvm, I'd made an assumption that the library gets installed with transformers. ","body":"Currently, running `from datasets import load_dataset` will throw a `ModuleNotFoundError: No module named 'datasets'` error.\r\n","comment_length":12,"text":"datasets module not found\nCurrently, running `from datasets import load_dataset` will throw a `ModuleNotFoundError: No module named 'datasets'` error.\r\n\nnvm, I'd made an assumption that the library gets installed with transformers. ","embeddings":[-0.5071018338,0.1617704779,-0.1671233922,0.1035202369,0.140610382,-0.0260318089,0.2918533683,0.2383850962,0.2065815926,0.057521496,0.0492231473,0.3291115761,-0.1274572015,0.1172012389,0.2540237308,0.0836432278,0.0580348559,0.3216485083,-0.3469064534,-0.1845293641,-0.1245576367,0.2255856097,-0.1934998482,-0.0363941193,-0.1254108548,-0.0336664729,0.0655893385,0.2748528719,-0.224902764,-0.3301688731,0.2005624324,0.099587217,0.1826145053,0.3182609379,-0.0000952629,0.0524798259,0.2305037528,-0.1074842066,-0.0970767289,-0.1658412069,-0.0767058581,-0.2284301072,0.1548487246,-0.179061234,-0.0325208567,-0.0479483716,0.0479542278,-0.3380207121,0.0583804473,0.1618400514,0.375384897,0.3236250877,0.0168855693,-0.3422498703,0.2585724592,-0.0452732593,0.0021664717,0.0287007857,0.1289357096,-0.0189784896,-0.0511539876,0.2808100283,-0.0705393255,0.0408040993,0.4888446331,-0.0090691717,-0.1403720081,0.086291559,-0.0097759096,0.1048650816,0.7409256697,-0.3725343347,0.0012503874,0.1397527307,-0.0420020297,-0.0475303009,0.1421617121,-0.1328109056,0.0898640007,0.1617739499,0.0056135771,-0.3213680983,0.0039069476,0.1484774202,-0.1540057659,0.4333933294,-0.1538278311,0.0529026091,0.2533197701,-0.0338754728,0.0134751927,0.1321777403,0.0167345144,0.0422600843,-0.1894995421,0.1030025408,0.3303790689,-0.1028016061,-0.0007505019,0.1180239841,0.0163638256,0.1580959558,-0.1658577472,0.2206495106,0.2908982933,-0.0628436804,-0.0095893759,0.2132614851,0.4497016966,0.0145134665,0.0310614277,-0.0183559563,-0.098706089,-0.1530425549,-0.055195827,-0.0241776481,0.1903494895,-0.2648780942,-0.1732951552,-0.2549210191,0.1967031211,-0.0799351856,0.1686813384,0.3981201053,-0.1296675503,0.1241472363,0.1451105773,-0.0442001857,-0.2422227263,-0.0687173158,-0.2977294922,0.2666596472,-0.3092794418,-0.2503127754,0.4085467458,0.0100091686,0.3717374206,-0.2439638525,-0.0003173907,0.2769038081,0.0815922096,0.1599702686,-0.0496809594,0.3940953016,0.1095014215,0.1575692296,0.0949383304,0.0923011899,-0.3470519781,0.0547775254,-0.2908751369,-0.2516860366,-0.0324297659,0.3703681827,-0.1241430044,-0.2234567851,-0.0442639291,-0.1419681311,-0.1496614814,-0.410002768,-0.2198766172,-0.0868641287,0.1486779898,-0.2146069258,0.2388042659,0.402328819,-0.5935941935,-0.113819994,-0.2774130106,-0.3158791065,-0.0071324161,0.1112396196,-0.3147812784,-0.0103241093,-0.2167398781,-0.0452063158,0.3968484998,-0.3573170602,-0.315202862,-0.0191079024,-0.0497320108,-0.2713224292,-0.1102506444,0.2817439139,-0.0371839963,0.0818935931,0.0938517451,0.3833034039,0.0862578675,-0.1351456493,0.0668269619,-0.2338264436,0.1548213214,0.2722499073,0.3224198818,0.0077554565,0.2527837157,0.1929936707,-0.0319907255,0.041391518,-0.0999145806,0.3686125278,0.342849642,0.0011367664,0.2267829925,-0.311360091,-0.262925297,0.2540789843,0.0651911721,0.0128906211,-0.092173107,0.0355134234,-0.2216910422,0.0830988213,-0.1105089635,-0.0835933313,0.2607947886,0.2113289237,-0.1259750128,-0.0531582646,-0.2787624896,0.0352855995,0.0045452495,0.0158705171,-0.4685574472,0.1848568916,-0.2582829297,-0.0061545959,0.1551525146,0.1031966358,-0.0019137278,-0.1604573429,-0.0646694303,0.4726766646,0.1229470596,0.1844251007,-0.0474331938,0.1201043949,0.0850240588,-0.449310869,0.4171442091,0.1401764005,0.0500995405,0.0624698475,-0.2639983594,-0.0015145194,-0.1738760322,-0.0336539932,0.2134456486,0.0600854494,0.3562760949,0.0565548912,-0.0189402569,-0.2819401622,0.1432222724,0.1694465429,0.2006710172,0.1148854345,-0.0528077483,-0.1379664391,0.5823255181,0.1417146027,0.2571653724,-0.0417336635,-0.3824773729,0.0253093895,0.1703519523,0.1954641044,0.3864998221,0.3329785466,0.072523579,-0.1339379996,-0.075726293,-0.2527935207,0.0286673401,-0.0565303266,0.1559200436,0.0969061255,-0.0917649195,-0.110004425,-0.4811644256,-0.2047089338,-0.1127308384,0.2805505991,-0.1720443368,0.1659642607,-0.5265134573,0.0454396494,-0.0483808443,-0.1713654697,-0.1856464297,-0.2343870699,0.0195244271,-0.0754432604,0.0221583806,0.237895444,-0.0336249806,-0.0510540903,0.0719701722,-0.2702361643,-0.0093969954,-0.2933236361,-0.2144890726,0.2260944396,0.0950596631,0.3321057558,0.2928007841,-0.1013461873,0.3006391227,-0.2206078619,-0.4864536226,0.1204467788,0.0695139989,0.0774461254,0.4811066687,0.0637237951,0.1272225082,-0.1702474505,0.2160419971,-0.3907992542,-0.0269567091,0.0767680332,-0.1908503771,-0.3700143099,-0.2034624517,-0.5828970075,-0.3107132912,-0.5354549885,-0.0460437499,0.2145684958,-0.0587897711,0.3745036423,0.1426668316,0.2761254609,0.2014656514,0.0916599929,-0.0694833249,-0.0081516486,0.4969363511,-0.2875054479,-0.3372792602,0.279604286,-0.0434515215,0.2800657153,0.1655354798,-0.3527469039,0.0970135704,0.072821632,-0.0400887877,0.1933039427,0.3185569942,0.2759097517,0.0462094471,-0.1709285975,-0.2657714486,-0.2192771584,0.0088898493,-0.0540105924,0.1902957112,-0.1057095751,0.2553887069,-0.2281645834,0.3259727061,-0.2913483679,-0.173369348,0.1770007461,-0.0429944657,0.5857387781,0.0167048723,-0.2867473066,-0.0081583876,0.1478550434,0.0134550137,0.2910920084,-0.1106125489,-0.3223560452,-0.3016456962,-0.0145476311,-0.000490715,-0.0680507198,0.129145354,-0.0347189717,0.2423152626,0.1355110407,0.168083176,-0.1787774116,0.0800926238,0.1080637723,0.1287172586,-0.1534990221,-0.0142139588,-0.4375036061,-0.0174356252,-0.1484923065,0.1695044637,-0.195951581,0.242259264,-0.1524689794,-0.0024264625,-0.1022212878,-0.2230393887,0.5168608427,-0.1888913512,-0.0528376438,0.196700111,0.035788063,-0.2234663218,-0.0222658347,-0.2610403895,0.0988768339,-0.0040580505,0.187093854,-0.1711180359,-0.1711376756,0.4019730389,0.0464774743,-0.0739536062,-0.1757095158,-0.4000532031,-0.2000800073,-0.4444824457,-0.2832746506,-0.1960256845,-0.0399041586,0.0960646793,-0.125302583,-0.035620179,-0.2056301236,0.1825617552,0.2699173391,0.3692810237,0.1459405124,0.034295287,0.1813460141,0.1171040535,0.1562164575,0.3423170447,-0.1380964071,-0.1645581722,0.2427629381,-0.0247606765,0.3135809004,0.0043509016,-0.1927183419,0.0344849713,0.0855227038,-0.0624479614,0.0223277416,0.0816825107,0.0911293849,0.1164923161,-0.1351816952,-0.511048317,0.2842982709,0.0420911089,-0.0097144665,0.2251487523,0.007819701,-0.2122168541,0.0352267437,0.081984356,0.6634898782,-0.033357162,-0.0038374648,0.2298600674,0.0163570885,0.4532926679,-0.3480087519,0.0322563238,-0.3162344694,-0.2168401182,-0.0839689672,-0.2124835849,0.2125982493,0.0936316177,-0.4127350152,0.2460869998,0.0645087734,0.2579441369,-0.0711510628,0.2618178129,0.092536591,-0.3170053065,-0.2220102102,0.3099567294,-0.2467071116,0.2836147249,-0.3222348988,-0.0708144456,0.122526288,-0.2632375956,-0.0665598735,0.1662086546,0.0778876916,0.4515031576,0.1890317351,0.1092335209,0.16004017,0.0852562711,0.1807198673,0.2268373072,-0.2343921959,0.4038122892,-0.2706180513,-0.1646209657,-0.0677190647,0.2667576969,0.168699652,-0.1658231914,-0.1850438565,0.1817899644,-0.0659009814,-0.3200293481,0.1316567957,0.0340998285,0.3019717634,-0.0068991189,-0.120126009,0.2264561653,-0.0466211289,-0.0765689164,0.275123477,0.291416049,0.0004035781,0.2744409442,0.2462871671,-0.0925806463,-0.121740967,0.2560034096,0.0458638258,-0.0836529806,0.1561255902,0.2797532678,-0.244492799,-0.3573535383,-0.1802330315,0.152030468,-0.1322962344,0.1005658954,0.0244586598,-0.1371925771,0.3250218928,0.1077443734,0.0822749659,-0.1883013994,-0.2657389343,-0.3804146051,-0.2745209932,0.3335862756,-0.1123385727,0.2686651945,0.1398382932,-0.1494812369,0.1235290468,0.0670109764,-0.4995964766,-0.0334778205,-0.160257116,0.0228213519,0.1386928409,-0.0148203177,-0.023537552,-0.1311663538,0.2218534648,-0.0513333045,-0.3307057321,-0.3133274615,-0.0995962769,0.0762630776,-0.1733574867,0.1592007875,0.1738720834,-0.3024280369,-0.4342029095,-0.1019719541,-0.0059377528,-0.1041654795,-0.0705465078,0.179357022,0.0025964228,-0.0891386792,-0.1511559933,0.0824694782,0.075678587,0.0211684965,-0.0637631118,0.1419050246,-0.0754835904,-0.1708869934,-0.1182878464,0.1603296846,0.1402865648,0.1034037098,0.3474603593,-0.1706517935,0.1869807541,-0.0200766195,0.1612169743,0.2507724166,-0.0142808184,0.1317951679,0.084398523,0.3403864503,-0.5574911833,-0.0741862655,0.1020581648,-0.0381826013,0.0984582528,0.1931410134,0.1766382009,0.0147006018,0.3961431384,0.2771225274,0.417539686,-0.1245120615,0.1971939504,0.4530421197,-0.1799061447,0.0269812234,0.2139957398,0.3048278391,0.1493645608,0.7942054868,-0.290578723,0.2506766021,0.0702869296,0.0887394547,-0.0250274986,-0.4160639644,0.1678220183,0.2443366498,0.1615140885,0.1136516407,0.0205520615,0.2896237671,-0.3237682581,0.1270613074,-0.2600284219,0.2102137804,-0.1562155485,0.0574435741,-0.1821528375,-0.130384773,-0.2670621574,0.0289210938,0.1546356678,-0.0298344158,0.0004085502,0.031285774,0.0419466719,-0.4273507893,0.3212869763,-0.0661781132,-0.2757753432,-0.2098176479,0.1208487451,-0.0140741663,-0.0041499217,0.0377392955,-0.0606133342,0.2153734416,0.2623687387,0.0530401506,0.017606305,0.0313875787,-0.2169709951,-0.1668744236,-0.0148036294,0.0681321919,0.4723003805,0.0426641963,0.304744184,-0.2981826663,-0.116567567,0.0022498323,-0.0246620625,-0.1333911568,0.2099874318,0.0555534251,-0.1134060174,-0.250369221,0.1005625427,-0.5538131595,-0.2323220223,0.4497943521,0.1133892238,-0.0685680062,-0.2156080306,0.1515846997,0.0295723993,0.475181967,0.0409649983,0.3285049498,-0.0272624325,-0.3100897372,-0.9142742157,0.1377665699,0.0695577338,-0.0359932482,0.0721525401,0.070734933,0.3116700053,0.0420295633,0.053741347,0.0740558133,0.0800727457,-0.3163180947,-0.1714575142,-0.2690962255,-0.0916951224,0.01398151,-0.0514406674,-0.3626460433,-0.1758106202,-0.1327332109,0.2041591108,-0.3967482448,-0.1352550834,0.1218039095,0.0241888165,0.5116804242,-0.2210973352,0.5889575481,-0.0679897815,0.0142166466,-0.2251857221,-0.3556123674,-0.0867312402,0.0085110953,0.1275801808,0.3536199331,-0.1160404757,0.0588921718,-0.3372921944,0.2051825821,0.2276365012,0.0539292209,-0.3017591238,0.1769097,-0.2047332376,0.2410845011,0.0647211969,0.0801188499,0.0339541025,-0.1202458963,-0.2132847011,-0.191697374,0.3978063166,-0.0618177652,-0.2801631093,-0.2002094388,-0.0736237839,-0.0327994265,-0.0803984851,-0.3862095177,0.2306971848,0.2089803815,0.0253140554,-0.147590667,0.3085522056,0.0004926344,0.0715374649,-0.1784708798,0.1114276871,0.0336539149,-0.2583512366,-0.1133672595,-0.1599351764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/900","title":"datasets.load_dataset() custom chaching directory bug","comments":"Thanks for reporting ! I'm looking into it.","body":"Hello,\r\nI'm having issue with loading a dataset with a custom `cache_dir`. Despite specifying the output dir, it is still downloaded to \r\n `~\/.cache`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\nfrom pathlib import Path\r\n\r\nvalidation_dataset = datasets.load_dataset(\"natural_questions\", split=\"validation[:5%]\", cache_dir=Path(\".\/data\")) \r\n```\r\n\r\n## The output:\r\n\r\n* The dataset is downloaded to my home directory's `.cache` \r\n* A new empty directory named \"`natural_questions` is created in the specified directory `.data`\r\n* `tree data` in the shell outputs:\r\n```\r\ndata\r\n\u2514\u2500\u2500 natural_questions\r\n \u2514\u2500\u2500 default\r\n \u2514\u2500\u2500 0.0.2\r\n3 directories, 0 files\r\n```\r\n\r\nThe output:\r\n```\r\nDownloading: 8.61kB [00:00, 5.11MB\/s] \r\nDownloading: 13.6kB [00:00, 7.89MB\/s] \r\nUsing custom data configuration default \r\nDownloading and preparing dataset natural_questions\/default (download: 41.97 GiB, generated: 92.95 GiB, post-processed: Unknown size, total: 134.92 GiB) to .\/data\/natural_questions\/default\/0.0.2\/867dbbaf9137c1b8\r\n3ecb19f5eb80559e1002ea26e702c6b919cfa81a17a8c531... \r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 13.6k\/13.6k [00:00<00:00, 1.51MB\/s] \r\nDownloading: 7%|\u2588\u2588\u2588\u258e | 6.70G\/97.4G [03:46<1:37:05, 15.6MB\/s]\r\n```\r\n\r\n## Expected behaviour:\r\nThe dataset \"Natural Questions\" should be downloaded to the directory \".\/data\"\r\n","comment_length":8,"text":"datasets.load_dataset() custom chaching directory bug\nHello,\r\nI'm having issue with loading a dataset with a custom `cache_dir`. Despite specifying the output dir, it is still downloaded to \r\n `~\/.cache`.\r\n\r\n## Environment info\r\n- `datasets` version: 1.1.3\r\n- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1\r\n- Python version: 3.7.3\r\n\r\n## The code I'm running:\r\n```python\r\nimport datasets\r\nfrom pathlib import Path\r\n\r\nvalidation_dataset = datasets.load_dataset(\"natural_questions\", split=\"validation[:5%]\", cache_dir=Path(\".\/data\")) \r\n```\r\n\r\n## The output:\r\n\r\n* The dataset is downloaded to my home directory's `.cache` \r\n* A new empty directory named \"`natural_questions` is created in the specified directory `.data`\r\n* `tree data` in the shell outputs:\r\n```\r\ndata\r\n\u2514\u2500\u2500 natural_questions\r\n \u2514\u2500\u2500 default\r\n \u2514\u2500\u2500 0.0.2\r\n3 directories, 0 files\r\n```\r\n\r\nThe output:\r\n```\r\nDownloading: 8.61kB [00:00, 5.11MB\/s] \r\nDownloading: 13.6kB [00:00, 7.89MB\/s] \r\nUsing custom data configuration default \r\nDownloading and preparing dataset natural_questions\/default (download: 41.97 GiB, generated: 92.95 GiB, post-processed: Unknown size, total: 134.92 GiB) to .\/data\/natural_questions\/default\/0.0.2\/867dbbaf9137c1b8\r\n3ecb19f5eb80559e1002ea26e702c6b919cfa81a17a8c531... \r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 13.6k\/13.6k [00:00<00:00, 1.51MB\/s] \r\nDownloading: 7%|\u2588\u2588\u2588\u258e | 6.70G\/97.4G [03:46<1:37:05, 15.6MB\/s]\r\n```\r\n\r\n## Expected behaviour:\r\nThe dataset \"Natural Questions\" should be downloaded to the directory \".\/data\"\r\n\nThanks for reporting ! I'm looking into it.","embeddings":[-0.100310035,0.1566864401,-0.0138976667,0.3185725212,0.1520786434,0.0216163658,0.2687704563,0.0195329338,0.2911308408,-0.1837452352,-0.1367085278,0.3869104087,-0.1338080019,-0.3240659535,0.0426421054,0.2474383861,-0.0759033784,0.1002800167,-0.0471776463,-0.1533921659,-0.2862761319,0.3169527948,0.0555715859,-0.0547319911,-0.3056131601,0.1543409526,0.0887487829,0.3776969016,0.0426424891,-0.3673247993,0.726052165,0.2212792337,0.314116925,0.6464560628,-0.0001215971,-0.0148937944,0.3063436449,-0.2060140371,-0.4097087681,-0.3351287842,-0.0759110302,-0.0097760335,0.1206249818,-0.0179076344,0.009506708,0.2860504985,-0.1667199582,-0.6777467132,-0.0464590862,0.2521005571,0.0861385092,-0.4244450033,-0.7473320365,0.2661696672,0.082806237,0.4113192558,-0.0619617403,0.3328621089,0.002130951,-0.0263146758,-0.1532160044,0.1160967201,0.0104363374,0.2888596654,0.4086684883,0.2578530908,-0.183906123,0.1464347243,0.1335081607,0.1286077797,0.883110106,-0.4564929903,-0.2414955944,-0.1922520697,-0.2393264174,-0.3878084719,0.331967473,0.2704976797,-0.3042610884,0.2642801106,-0.2381147742,-0.0693234354,-0.1462028772,0.065216966,-0.0264728032,0.044931002,-0.1007825062,-0.0435524434,-0.0109208934,0.0603811108,0.6779462099,-0.3270463049,-0.0874882638,0.2589638233,-0.0136455745,0.1731263399,0.0006245137,0.1011433601,-0.0092381183,0.3575193882,0.1823396087,0.2034995407,-0.3495633304,-0.1536741853,0.3089654446,0.2394989729,-0.0634289831,0.3599571586,0.0272127893,0.2537470162,0.0279552806,-0.1083407775,0.0397811234,-0.0059674173,0.228737101,-0.0933415964,-0.0383427478,-0.2454602867,0.008939364,-0.0694214478,0.3045836389,0.1163822487,0.1297688037,0.2351420522,0.1084585786,0.2255127281,-0.2647976577,0.312155813,-0.0311296098,0.3595801592,-0.1248819754,-0.221042797,-0.346085012,0.0962350667,0.3177661896,0.2229542136,0.1707334965,0.0800085664,0.1727413684,-0.0488837846,-0.2592087686,-0.0260077361,0.2516829669,0.1977748871,-0.3073123097,0.5351738334,0.0951653495,-0.0907133743,-0.2672970295,0.2881723344,-0.2912237942,-0.4989192188,0.3338618279,0.0476522446,-0.3118879199,-0.0336685665,-0.448284775,0.0157757774,0.2454269379,0.0139443064,0.1619272381,-0.0630524233,-0.2865334451,-0.2779092491,0.0469969921,0.5059197545,-0.3776888847,0.0915305465,-0.2293853611,0.1108023822,0.22879228,0.2118332833,-0.1066753194,0.1562856287,-0.5334421396,-0.0346776918,0.3338702619,-0.4785211086,-0.8453207016,0.313013941,-0.0224225614,0.2746495008,0.0632882863,0.2726621032,-0.0099754268,-0.0914382339,0.0420069285,0.4156000912,0.0102838548,0.0931527019,-0.2984859645,-0.0646856725,0.0421404988,0.017560577,-0.0531981289,0.2228942662,0.0520581417,-0.0365635343,0.3624807596,0.1151247174,0.0734879375,0.1296464503,0.2857322395,0.14559412,0.0362929739,0.0350309201,-0.3819657266,0.3323414922,0.0696068406,-0.3697364628,-0.3099512458,-0.0119267656,-0.4464807212,-0.16420196,-0.4505040944,-0.3120310009,-0.0348856449,0.2236723453,0.1508025378,-0.0464094281,0.0524944216,0.5415360332,0.0662625358,-0.0269982666,0.0644986629,0.1347407997,0.2625746131,-0.0803037807,-0.3099212646,0.2334026247,0.1562754065,-0.1227813438,-0.1813182235,0.487319231,0.3435271382,0.0916891545,-0.0214195326,-0.1859931946,0.126938954,-0.2470653653,0.1838678718,0.3062117398,-0.0311064925,-0.1173462942,-0.333240509,0.2258646339,-0.422421664,0.1760955453,-0.1806161255,-0.0969884768,0.3039502501,0.0281189531,0.2815055251,-0.1179810837,-0.0119644273,0.1406895965,0.5359691381,-0.0470201671,0.0888136402,0.0291806161,-0.0882730857,0.0814835951,-0.0442600623,0.1867279261,-0.0160850976,-0.1757408231,-0.0317753814,0.4905295968,0.6581900716,-0.0151117779,0.2403131276,-0.0171288308,0.2095117122,-0.1541909426,0.0634462833,-0.2090675384,-0.1557327658,0.2276001573,0.1284860522,-0.2468304634,-0.0124646649,0.0395339206,-0.0699774697,0.0370174162,-0.2672823966,0.1763571501,-0.2761175632,-0.0485663079,-0.1666098684,-0.3963249922,-0.0045491676,-0.1476577371,-0.1091582999,0.1150322109,0.1297849268,-0.082723707,-0.0726414323,0.0977625921,-0.0263839923,-0.7693251371,-0.3080716729,0.0139962342,-0.3375791609,-0.0927922949,-0.0879394561,-0.0453946367,0.0658181831,-0.1189690381,0.0868887082,-0.3877847791,0.1328192204,-0.1239641681,0.1933169514,0.2174898088,0.2800197303,0.0802810192,0.3303265572,0.2187753022,0.101655826,-0.087982893,0.0741182789,0.0881865993,0.2687190771,0.4017107785,-0.2374659479,-0.2188120633,-0.2041891813,-0.2193749696,0.1708123386,0.2367019951,0.1073880941,0.2860151529,-0.0861401781,-0.206588164,0.2188395411,-0.29227826,-0.2334153056,-0.6056892276,0.4533127844,-0.1807848364,-0.0170180891,-0.116881676,-0.157194227,0.2490514517,0.3172036111,-0.4699003398,0.2374418229,-0.1255251765,0.3220001459,-0.0359948017,-0.2082154304,0.2012636065,-0.2192900479,0.1483707428,-0.1928873211,-0.235575676,0.1047272012,0.0520060621,0.2776826918,0.1375557631,0.0639165565,0.153824985,0.7221413851,0.1539371461,0.0742423832,0.4522139728,0.0303761847,0.5724590421,-0.1169300899,-0.007672139,-0.2306324244,0.1170830652,-0.3017332852,0.2620259821,0.3963634372,-0.1375484765,0.1484493166,0.2293836325,-0.0581068322,-0.0099141421,0.0004128208,-0.4490572512,-0.0360231213,-0.0196599197,0.0369227156,-0.1578114629,-0.2052124441,0.0700267032,0.3217170835,-0.0247585233,0.1586972028,-0.1694205701,0.2370512486,-0.2716444731,-0.0052203061,0.1251841038,0.0811061338,-0.0500415154,0.1385585815,-0.0303219929,0.1891993582,0.3788340092,-0.4624364376,0.3742393255,-0.0133163296,0.177074641,-0.2982981801,-0.0844088048,0.2065370679,0.3438355625,0.1180706173,0.2843798399,-0.1416012496,-0.0317823887,-0.2734389603,0.0930940509,-0.2558269799,-0.3774323761,-0.2003502399,-0.3097074926,0.2808871567,-0.0216332562,0.1363885552,0.1516315192,-0.0326052792,-0.2365138829,0.0598477125,-0.1065173522,0.0269626137,-0.0988513678,0.4053294361,-0.0481843539,0.0644872338,0.0932340547,0.0154844346,0.3688474596,0.6062712669,-0.2902154922,-0.0930352435,-0.196860373,-0.153568998,-0.0112940753,0.112245068,-0.1644544452,-0.224720493,-0.1506807804,-0.2050099522,-0.0658675507,0.2289537638,-0.0348609611,-0.1079313681,-0.1426237524,-0.6423487663,0.6105131507,-0.1019863114,-0.2246486843,0.3990159333,-0.2751591206,-0.4762174785,0.0601588003,0.0049009323,0.5365189314,-0.1949052662,0.0437405556,0.1506889313,-0.4975365698,0.1374094486,-0.3170315921,-0.1448359787,-0.3075854182,-0.0641890839,-0.0905090198,-0.2173981667,-0.0243252534,0.5099092126,0.0517953932,0.3643232286,-0.2680823505,0.0393122248,0.1543271989,0.0964693055,-0.0881887674,-0.2170658857,-0.1301265061,0.068639636,-0.1002887785,0.2139971256,-0.1089007407,-0.1205543652,-0.3574446142,-0.1217590868,-0.2260870337,0.0924165472,-0.1498946249,0.0837113932,-0.0223799236,-0.6317331791,-0.0655814335,0.5021620393,0.2567076683,-0.1353722513,-0.189909175,0.1939696223,-0.1225313023,0.3328512013,0.0580582023,-0.1394952387,-0.2117276341,0.089103207,0.0078134993,-0.3503928483,-0.2574719191,-0.3597339392,0.2472090423,0.0820889547,0.5061628819,-0.2173203677,-0.2124123573,-0.2512455881,0.2360196561,-0.0049491748,0.029265767,0.2241957635,0.0826166421,0.0560235754,-0.4281601906,-0.2743397653,-0.0831524059,0.1023245007,-0.2184900641,-0.0483511984,0.3331488073,-0.2102406025,-0.091687113,-0.0928354859,0.055910483,-0.0566570386,-0.2180189341,0.1766453534,-0.1676133871,-0.0255993828,0.0407505222,-0.1110672653,0.3499626517,-0.0028158315,-0.3172101676,-0.2882134318,-0.2207769752,-0.2910120785,-0.4724444449,0.238265574,-0.1950856,-0.2398250848,0.2391822189,-0.155542329,-0.1269519031,0.2302238494,0.1910094172,-0.0797806084,-0.0104289744,0.1285414845,0.2302985787,0.0027771625,-0.0330820084,0.0045185783,-0.1475675106,-0.1016011089,-0.1545259506,0.1715357155,0.0963872746,0.0213149432,-0.1491516232,-0.1518881023,0.0108513916,-0.3372264802,0.0827274993,0.34728688,0.0024117436,0.2840576768,0.1765421778,0.0365642346,-0.4913540184,0.1411708444,-0.3437809944,0.2318935245,-0.0746468753,0.1349936873,-0.0139980605,-0.0904749185,-0.2865051031,-0.0499793179,0.0694598183,0.0764873177,0.2573838532,-0.361813128,0.2461123317,0.2595525384,0.3133693635,0.0359033979,-0.1139330417,-0.3866613507,0.1298729032,0.1036791503,0.0151253529,-0.2221674472,0.0143837109,0.1379692703,-0.1063722968,0.4658239186,0.2330063581,-0.3086942136,0.0583379716,-0.0941796377,0.6277664304,-0.0338302329,0.1503467262,0.4308443069,-0.1127836704,0.2660207152,0.3019171655,0.2297517061,-0.0406536274,0.7379966378,0.1415832639,0.1891445071,0.1071858779,-0.1197424755,-0.2871803939,-0.5775136352,-0.2379062921,-0.1316013932,-0.0356723256,0.1468628347,0.1408684105,0.3643674254,-0.2864255607,0.1463699341,-0.0987172946,-0.4558994174,-0.0508669056,-0.1998109072,0.4144226015,-0.084016867,0.2777834535,0.0072708684,-0.1241524369,-0.1577518731,0.1624831408,-0.0399608798,0.0564061217,-0.2194122374,0.2750935256,0.1800235212,-0.1220553368,-0.0685844645,0.2392012775,-0.1734965444,-0.0869195759,0.3120979965,-0.0627679452,0.0780171081,0.4837380052,-0.0813343003,-0.0249847565,0.0591058508,0.000807804,-0.18935588,0.2493908405,-0.116148673,0.0127551658,0.0862139463,0.1372099817,-0.0353697874,0.2734617591,0.2379638404,0.4238523245,-0.1790697426,0.4305942655,0.0257795975,0.0230144821,-0.0027406213,0.2846668959,-0.1170066148,-0.1875480562,0.2559602857,0.0138052665,0.3057967126,0.2002982795,0.0026322608,0.0319202207,0.355125159,0.0579897799,0.1783721298,-0.2604323626,-0.1325942874,-0.6559234262,0.1622914374,-0.0748841539,-0.2789477408,-0.1963876486,0.2335993797,-0.1601109803,0.0303598568,0.1746539474,0.1548372805,-0.0314239562,0.3068628013,-0.1913923621,-0.198256731,-0.2722246051,0.2340079695,0.0449352302,-0.5114966631,-0.16333808,0.2584909201,0.0440616906,0.2399416566,0.019314561,0.1846775115,0.3545166552,0.4034712017,-0.0068516252,0.1315156072,0.2316978127,-0.026290914,-0.4371985197,-0.0559247918,0.0062545938,0.0259486437,-0.097233355,0.4291609824,-0.5795593262,0.1051491797,-0.4876827598,0.1325143576,0.3703945577,-0.0520584993,-0.0363186672,0.056012895,-0.2457572073,0.1781115979,0.1219949499,0.2843018472,-0.1871739477,0.2649449408,0.0188363381,-0.1472875625,0.3051881194,-0.2745055556,-0.4347991943,0.2511443496,0.3194212914,0.0825503692,-0.1097271517,-0.2869985998,0.1181311384,0.4543891251,-0.0903284326,-0.0694086924,-0.1755545735,-0.3217994571,0.0846159905,-0.0101030152,0.4636053145,0.1898632497,-0.0211057682,0.0442877486,-0.2492629141]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/897","title":"Dataset viewer issues","comments":"Thanks for reporting !\r\ncc @srush for the empty feature list issue and the encoding issue\r\ncc @julien-c maybe we can update the url and just have a redirection from the old url to the new one ?","body":"I was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n","comment_length":38,"text":"Dataset viewer issues\nI was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n\nThanks for reporting !\r\ncc @srush for the empty feature list issue and the encoding issue\r\ncc @julien-c maybe we can update the url and just have a redirection from the old url to the new one ?","embeddings":[-0.2186932266,0.2789271474,-0.026432747,0.2816513479,-0.0023903828,0.1115044877,0.2981258929,0.2988632917,-0.1956236809,0.0647588521,-0.0076386253,0.2664061487,-0.3950020671,0.0988750979,-0.1122838408,-0.2007787526,0.0220978931,-0.0146634579,0.0232188925,-0.0943433568,-0.1718729287,-0.2318453193,-0.3681103587,0.0472469144,0.1730451733,-0.1287799329,-0.0984149203,-0.0248546954,-0.1657533944,-0.6458482146,0.0453227796,0.2832351029,0.1305533051,0.4277637899,-0.0001141002,-0.0952439457,0.5284169316,0.2207832187,-0.4246532321,-0.1656228304,-0.1231052652,-0.3184152842,0.4010361731,-0.1876373738,0.2975274324,-0.34791255,0.0207551885,-0.1845204532,0.0901871175,0.469602257,0.1956328005,0.1108126119,-0.1084176004,0.1752130687,0.3652741909,0.2491191179,-0.2797198892,-0.1015249267,0.1632782519,-0.0791658387,-0.025709914,0.3140527904,-0.1640222073,-0.0861883536,0.3573250473,-0.1882148087,-0.0757901296,-0.4510447383,-0.0953087807,0.2241510004,0.7972296476,-0.3495242298,-0.3797179759,-0.3601836562,0.0131202983,-0.1577013135,0.1619466543,0.0079309354,-0.1782668978,0.0668451935,-0.2203083187,-0.3423940837,-0.1549729556,0.0965098739,-0.2113793194,0.6253469586,0.1779181063,0.0807632133,0.1139025018,-0.1835037321,0.5544168949,-0.1228946298,-0.1087577194,0.1986276656,-0.2692966461,-0.1737065315,0.461104244,-0.1622395366,-0.1637828648,0.4153255224,0.2714703977,0.0435075462,-0.3141144812,0.009632552,0.4544065893,0.1671871543,0.147337541,0.1572187394,0.3739819229,-0.038006708,0.3260328472,0.1013467759,0.0978917032,-0.0086792316,-0.2973321676,-0.116244182,0.4545963407,-0.2330589294,-0.303227216,0.1526127011,-0.1009672284,-0.263060987,0.1532100588,0.219112128,-0.0706097782,0.0922878385,-0.1988677233,0.1861469299,-0.2814712524,-0.1896958649,-0.1201233566,-0.1157088652,-0.1896719784,-0.3266671598,0.1634646803,-0.2876240313,0.119710207,-0.1003617421,-0.1547210664,0.194593668,0.0908226743,-0.1636734307,-0.0009145842,0.0688804686,-0.2302842587,0.1360697597,0.1197247058,-0.2654082775,0.0479442663,0.2560563684,0.2586598098,-0.1782319695,-0.1489072144,0.1788307279,-0.2246099263,0.0703034028,0.0440880843,0.3644196093,-0.2849558592,-0.2395537049,-0.1893162578,-0.2350794524,-0.1941615939,-0.187835604,0.1720422506,0.5530490279,-0.4830772579,-0.0465624295,-0.4554136395,-0.0557204336,0.0420507528,0.059990108,-0.1993482709,0.1053986698,-0.3029534817,-0.0673145503,0.5750009418,-0.219153285,-0.4035130143,0.2735553682,-0.172700122,0.0276703667,0.0239073969,0.1319134533,0.1925355643,-0.3088919222,-0.3873092234,-0.0334397294,-0.0646378994,-0.2213612795,-0.3128040135,-0.1074304804,0.5276122689,0.036488194,0.13503021,-0.0540118702,0.2751266658,0.0094174203,0.3650549352,0.1358383149,0.0362217166,0.0742885768,-0.0089339912,0.1816091686,-0.0343146659,-0.1788937896,-0.370862782,0.0983280316,0.1578837186,0.0319679044,0.0578098372,-0.0395850949,-0.0441822745,0.176623553,-0.2002382427,0.0091595007,0.1059279591,0.0371709205,-0.228386879,0.150171414,-0.3601384461,-0.3355703354,-0.0756464526,0.1033099219,-0.2184205651,0.4458243251,-0.0775557607,-0.0546324924,0.2532230914,0.3302969038,0.2176555693,-0.1338980347,-0.1547198892,0.225530237,0.3120882213,0.0964817703,0.2038223445,0.2202825546,0.4631510973,-0.4166270792,0.4512762427,0.4248676896,0.0157839339,-0.0111361668,-0.4655109346,0.1208941489,-0.4010038972,0.3006582558,0.1231233552,0.0035366071,0.3432372808,0.0029793673,0.0387485735,-0.1294725239,0.2690471411,0.0817266405,-0.0146125136,-0.0484843403,-0.4041103125,0.2343586981,-0.1110977903,0.142692402,-0.2079056054,0.1828098595,-0.2422186732,-0.1872489154,0.2078428715,0.1649779528,0.1381793618,0.1974901557,0.00771394,0.1722227633,0.1840904504,-0.1922398806,0.2213065922,0.2634327114,-0.2253491729,0.1050514877,0.11574696,0.0659233034,-0.2038090527,-0.3943380713,0.1596047878,0.3535887003,-0.2750309408,-0.172228083,-0.383197397,-0.708920002,0.0878656879,-0.0594570749,-0.3682579696,-0.550560534,0.1306333691,0.3110088706,-0.3960461915,0.2758713663,-0.5261030793,0.0930711403,-0.0833637863,0.377148807,0.029483242,-0.315015018,-0.1494927406,0.1138873547,0.3704943359,0.3368929923,0.3391319811,0.2285357863,0.0902660713,-0.5124884248,-0.3946644664,0.1821422726,-0.0811305791,0.3599928916,0.0334619135,-0.0304460786,-0.0033648976,-0.0145208053,0.1848327667,-0.0334529392,0.0808227807,0.0377494097,0.2265972346,-0.0568405129,0.0043872967,-0.2615486085,-0.1642236412,-0.5408262014,0.0232084766,0.1773352772,0.2984770536,0.2128200233,-0.110032782,0.0811095014,0.1700314879,-0.1284519434,0.08159215,0.2522673607,0.2351938635,-0.1482935697,-0.3752676547,0.1431304067,0.1678874493,-0.1806288958,-0.0289230738,-0.6907394528,0.2365356535,0.0004746777,0.0933537111,0.0852360874,-0.1197598875,0.2346420884,0.3941362798,-0.0285920184,0.0604873337,-0.1572939903,-0.1623924971,-0.0272981729,0.2487436384,0.0613057353,0.2927386463,-0.0788261741,0.4196695387,0.1608722657,0.1330552101,0.3300288022,-0.1292558014,0.3863185942,0.0880506858,-0.364426285,0.3801051974,0.2277207226,-0.2473003566,0.3891529143,-0.0468089357,-0.2272368819,-0.1322384924,0.111875169,-0.0174210239,-0.2235319912,0.0054794597,-0.3366178274,0.0723116398,0.1084828675,0.1603690684,-0.087550588,-0.4027015269,-0.0639728084,0.0354260877,0.1260525137,-0.397769779,-0.6519774199,-0.1440824121,-0.4036369324,0.3362410069,0.1335719079,0.2592405677,-0.1384121776,-0.0428604856,0.0075490596,0.3046002984,0.3338590562,-0.2926357687,0.0308585726,0.2173912674,0.2529658377,-0.3621837795,-0.1252756864,-0.1247658134,0.468627125,-0.0056123338,0.1088866964,0.1177692637,-0.1460490376,0.1244101599,0.0838265717,-0.3346751928,-0.0230384395,0.0858213305,-0.1470863372,-0.1134802401,-0.2595456243,-0.2903983891,-0.0807918385,-0.1117307097,-0.0160472374,0.0051927869,-0.171682179,0.1276297867,0.1367404908,0.4477116764,-0.0723016411,0.2226013392,0.2707365453,0.2623255849,0.3942559958,0.3254241347,0.1623783857,-0.1076942012,0.1056443229,-0.1841582358,0.3969781995,0.2862301171,0.0432810932,0.0279178899,-0.0078068012,0.3893857896,-0.1983547658,0.0260479581,0.3512432873,-0.0967423618,-0.1660416126,-0.6548130512,0.17357862,-0.1757688969,-0.2077556103,0.4948207438,0.0051999725,-0.1810278147,0.5770657659,0.0360416919,0.9330322742,0.2672248483,-0.1094423905,0.2783369422,-0.1210113466,0.612965703,0.0159467701,0.277343899,-0.3774955273,-0.3241224885,-0.0243557822,0.0032225286,0.0388527885,-0.1578301191,-0.1630984992,0.238726899,-0.233989805,-0.0200340506,-0.0329130441,0.3232554197,0.2424539328,0.1481516361,-0.1084030494,0.1145466119,-0.0608149692,-0.1897985637,-0.1467810422,0.0128434738,0.1084817946,0.0014939425,-0.3961191773,0.032109648,-0.1198762879,0.1515937299,-0.3794673681,-0.3329626322,0.2918035388,0.2802639008,-0.1389145106,0.1128718928,-0.156388551,0.1829782724,0.0113465497,0.181289658,0.0638161004,0.2905876935,0.2833430469,-0.0767332315,-0.2789282501,0.1410662979,0.0836604014,-0.0352023803,-0.2029493302,-0.185880959,0.1857665479,-0.277907759,0.1268288493,-0.1243294179,-0.3048667312,-0.1910587847,0.1262686253,0.2097568065,-0.0669730082,0.2531118393,0.1537329853,-0.0373004451,-0.1281880587,0.2323104441,0.0470716655,0.0388124958,0.5362433791,0.4200657308,-0.1564955562,-0.1853557229,0.1352237463,-0.1985569149,-0.1128123924,0.0336666778,-0.4092645347,0.0790371373,0.0843415558,0.1507214308,0.1443125755,0.1743270755,-0.4357252121,-0.2476724833,-0.0197396055,0.2915032208,-0.1539634466,0.1664569527,-0.4321200252,-0.141108498,0.1247308552,-0.1702901423,-0.3006685972,-0.190025717,0.0198921841,0.181043312,0.0269484352,0.1825227886,-0.2085029036,-0.4424799681,0.0355275087,-0.2638809681,-0.0845437199,-0.1637734473,-0.0852388814,0.1769664884,0.0740286261,-0.441613853,-0.3148105443,-0.2439830303,-0.0189493522,-0.063104406,-0.3076319993,-0.047821641,0.0404856764,0.0820248723,-0.0307593718,-0.2753333151,-0.0077422154,0.2412666827,0.1193006262,0.541461885,0.2563714385,0.2153195143,0.1639128923,-0.1250135452,-0.2580597401,0.1643506736,-0.3087143302,0.1652209312,0.3549098074,-0.3322285712,0.2556830645,0.3808780015,0.4932388961,0.4313769042,-0.1432711631,-0.2789634764,0.4424249232,0.1640637368,-0.4123229384,-0.1763328463,0.2795851529,0.2438185513,0.0173587725,0.1030137315,0.136197865,-0.1607579887,-0.2585865855,0.0680326298,0.4175924957,-0.203919515,0.3002223074,0.4241982698,-0.212717548,-0.0639845207,0.1778074652,0.2166756988,0.1435504109,0.6396567822,0.1417242885,0.0758812279,0.4841813743,0.1691903025,0.1241489053,-0.2833169997,-0.1818279028,0.2028390914,-0.099501811,0.0930603817,0.259293735,0.04682054,0.063155517,-0.3688143194,0.0459530018,0.2396122366,-0.0075236442,0.2120814323,-0.3550190926,-0.2509007454,-0.0494616255,-0.0261533819,-0.0940736085,-0.0751662403,-0.1726794243,-0.0033784211,-0.3168993294,-0.4275234044,0.3732615113,0.1240848452,0.2268958837,-0.2028679252,-0.0114721479,0.3269617856,0.1735690534,-0.0246689357,0.4546066821,0.2917703092,0.1306448132,-0.2347610295,0.0310389008,-0.0664144084,-0.1432176083,-0.1319899261,0.132255286,0.0302545149,0.1951593608,0.1808878481,0.076801993,-0.053962335,0.4734232128,0.2337242514,0.0584048405,-0.2925067544,0.3399984241,0.145849213,-0.5043171048,0.0774904564,-0.3538134694,-0.2001726329,0.1857924014,0.7511013746,0.031302385,0.1374648809,-0.0585534424,0.061800804,0.0923976228,0.3376110196,0.4668429494,0.0355729833,-0.1303791851,-0.2312413007,-0.8589652181,0.0914343521,-0.2171116024,-0.2172402143,0.5133656263,0.1280394047,-0.0502474234,0.2213602513,0.3961088061,-0.0102151269,0.0270923283,0.0515248701,-0.3876031041,-0.1328259557,-0.1321590841,0.3818224669,-0.3627648056,-0.1248889267,0.2687458694,-0.4400665462,-0.0286197234,-0.0540294871,0.0377760753,-0.0401768088,0.1479733884,0.1913590431,0.1345200837,0.1358108371,0.2102367878,-0.0262315515,-0.4110262394,-0.2733960152,-0.2440471798,0.2069380134,0.098176308,0.310220629,-0.3268450499,0.035486497,-0.2343169451,0.2626631558,-0.0487470105,0.192589432,-0.4621894062,-0.0339333825,-0.1455089748,0.2107549608,0.0786109194,0.1603516042,-0.0111095784,0.0877976194,-0.1460526586,-0.301722914,0.4168035388,-0.389662683,-0.1787187904,-0.0096131312,0.1982851475,0.3603422344,0.0374178886,-0.327151835,-0.1917618066,0.2938884199,-0.1424470246,0.0055394652,-0.0835715681,-0.1831873506,-0.0577196181,0.1430443674,-0.0101630688,0.1501576304,-0.2010464221,0.1263785064,0.0989648774]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/897","title":"Dataset viewer issues","comments":"Ok, I redirected on our side to a new url. \u26a0\ufe0f @srush: if you update the Streamlit config too to `\/datasets\/viewer`, let me know because I'll need to change our nginx config at the same time","body":"I was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n","comment_length":36,"text":"Dataset viewer issues\nI was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n\nOk, I redirected on our side to a new url. \u26a0\ufe0f @srush: if you update the Streamlit config too to `\/datasets\/viewer`, let me know because I'll need to change our nginx config at the same time","embeddings":[-0.2186932266,0.2789271474,-0.026432747,0.2816513479,-0.0023903828,0.1115044877,0.2981258929,0.2988632917,-0.1956236809,0.0647588521,-0.0076386253,0.2664061487,-0.3950020671,0.0988750979,-0.1122838408,-0.2007787526,0.0220978931,-0.0146634579,0.0232188925,-0.0943433568,-0.1718729287,-0.2318453193,-0.3681103587,0.0472469144,0.1730451733,-0.1287799329,-0.0984149203,-0.0248546954,-0.1657533944,-0.6458482146,0.0453227796,0.2832351029,0.1305533051,0.4277637899,-0.0001141002,-0.0952439457,0.5284169316,0.2207832187,-0.4246532321,-0.1656228304,-0.1231052652,-0.3184152842,0.4010361731,-0.1876373738,0.2975274324,-0.34791255,0.0207551885,-0.1845204532,0.0901871175,0.469602257,0.1956328005,0.1108126119,-0.1084176004,0.1752130687,0.3652741909,0.2491191179,-0.2797198892,-0.1015249267,0.1632782519,-0.0791658387,-0.025709914,0.3140527904,-0.1640222073,-0.0861883536,0.3573250473,-0.1882148087,-0.0757901296,-0.4510447383,-0.0953087807,0.2241510004,0.7972296476,-0.3495242298,-0.3797179759,-0.3601836562,0.0131202983,-0.1577013135,0.1619466543,0.0079309354,-0.1782668978,0.0668451935,-0.2203083187,-0.3423940837,-0.1549729556,0.0965098739,-0.2113793194,0.6253469586,0.1779181063,0.0807632133,0.1139025018,-0.1835037321,0.5544168949,-0.1228946298,-0.1087577194,0.1986276656,-0.2692966461,-0.1737065315,0.461104244,-0.1622395366,-0.1637828648,0.4153255224,0.2714703977,0.0435075462,-0.3141144812,0.009632552,0.4544065893,0.1671871543,0.147337541,0.1572187394,0.3739819229,-0.038006708,0.3260328472,0.1013467759,0.0978917032,-0.0086792316,-0.2973321676,-0.116244182,0.4545963407,-0.2330589294,-0.303227216,0.1526127011,-0.1009672284,-0.263060987,0.1532100588,0.219112128,-0.0706097782,0.0922878385,-0.1988677233,0.1861469299,-0.2814712524,-0.1896958649,-0.1201233566,-0.1157088652,-0.1896719784,-0.3266671598,0.1634646803,-0.2876240313,0.119710207,-0.1003617421,-0.1547210664,0.194593668,0.0908226743,-0.1636734307,-0.0009145842,0.0688804686,-0.2302842587,0.1360697597,0.1197247058,-0.2654082775,0.0479442663,0.2560563684,0.2586598098,-0.1782319695,-0.1489072144,0.1788307279,-0.2246099263,0.0703034028,0.0440880843,0.3644196093,-0.2849558592,-0.2395537049,-0.1893162578,-0.2350794524,-0.1941615939,-0.187835604,0.1720422506,0.5530490279,-0.4830772579,-0.0465624295,-0.4554136395,-0.0557204336,0.0420507528,0.059990108,-0.1993482709,0.1053986698,-0.3029534817,-0.0673145503,0.5750009418,-0.219153285,-0.4035130143,0.2735553682,-0.172700122,0.0276703667,0.0239073969,0.1319134533,0.1925355643,-0.3088919222,-0.3873092234,-0.0334397294,-0.0646378994,-0.2213612795,-0.3128040135,-0.1074304804,0.5276122689,0.036488194,0.13503021,-0.0540118702,0.2751266658,0.0094174203,0.3650549352,0.1358383149,0.0362217166,0.0742885768,-0.0089339912,0.1816091686,-0.0343146659,-0.1788937896,-0.370862782,0.0983280316,0.1578837186,0.0319679044,0.0578098372,-0.0395850949,-0.0441822745,0.176623553,-0.2002382427,0.0091595007,0.1059279591,0.0371709205,-0.228386879,0.150171414,-0.3601384461,-0.3355703354,-0.0756464526,0.1033099219,-0.2184205651,0.4458243251,-0.0775557607,-0.0546324924,0.2532230914,0.3302969038,0.2176555693,-0.1338980347,-0.1547198892,0.225530237,0.3120882213,0.0964817703,0.2038223445,0.2202825546,0.4631510973,-0.4166270792,0.4512762427,0.4248676896,0.0157839339,-0.0111361668,-0.4655109346,0.1208941489,-0.4010038972,0.3006582558,0.1231233552,0.0035366071,0.3432372808,0.0029793673,0.0387485735,-0.1294725239,0.2690471411,0.0817266405,-0.0146125136,-0.0484843403,-0.4041103125,0.2343586981,-0.1110977903,0.142692402,-0.2079056054,0.1828098595,-0.2422186732,-0.1872489154,0.2078428715,0.1649779528,0.1381793618,0.1974901557,0.00771394,0.1722227633,0.1840904504,-0.1922398806,0.2213065922,0.2634327114,-0.2253491729,0.1050514877,0.11574696,0.0659233034,-0.2038090527,-0.3943380713,0.1596047878,0.3535887003,-0.2750309408,-0.172228083,-0.383197397,-0.708920002,0.0878656879,-0.0594570749,-0.3682579696,-0.550560534,0.1306333691,0.3110088706,-0.3960461915,0.2758713663,-0.5261030793,0.0930711403,-0.0833637863,0.377148807,0.029483242,-0.315015018,-0.1494927406,0.1138873547,0.3704943359,0.3368929923,0.3391319811,0.2285357863,0.0902660713,-0.5124884248,-0.3946644664,0.1821422726,-0.0811305791,0.3599928916,0.0334619135,-0.0304460786,-0.0033648976,-0.0145208053,0.1848327667,-0.0334529392,0.0808227807,0.0377494097,0.2265972346,-0.0568405129,0.0043872967,-0.2615486085,-0.1642236412,-0.5408262014,0.0232084766,0.1773352772,0.2984770536,0.2128200233,-0.110032782,0.0811095014,0.1700314879,-0.1284519434,0.08159215,0.2522673607,0.2351938635,-0.1482935697,-0.3752676547,0.1431304067,0.1678874493,-0.1806288958,-0.0289230738,-0.6907394528,0.2365356535,0.0004746777,0.0933537111,0.0852360874,-0.1197598875,0.2346420884,0.3941362798,-0.0285920184,0.0604873337,-0.1572939903,-0.1623924971,-0.0272981729,0.2487436384,0.0613057353,0.2927386463,-0.0788261741,0.4196695387,0.1608722657,0.1330552101,0.3300288022,-0.1292558014,0.3863185942,0.0880506858,-0.364426285,0.3801051974,0.2277207226,-0.2473003566,0.3891529143,-0.0468089357,-0.2272368819,-0.1322384924,0.111875169,-0.0174210239,-0.2235319912,0.0054794597,-0.3366178274,0.0723116398,0.1084828675,0.1603690684,-0.087550588,-0.4027015269,-0.0639728084,0.0354260877,0.1260525137,-0.397769779,-0.6519774199,-0.1440824121,-0.4036369324,0.3362410069,0.1335719079,0.2592405677,-0.1384121776,-0.0428604856,0.0075490596,0.3046002984,0.3338590562,-0.2926357687,0.0308585726,0.2173912674,0.2529658377,-0.3621837795,-0.1252756864,-0.1247658134,0.468627125,-0.0056123338,0.1088866964,0.1177692637,-0.1460490376,0.1244101599,0.0838265717,-0.3346751928,-0.0230384395,0.0858213305,-0.1470863372,-0.1134802401,-0.2595456243,-0.2903983891,-0.0807918385,-0.1117307097,-0.0160472374,0.0051927869,-0.171682179,0.1276297867,0.1367404908,0.4477116764,-0.0723016411,0.2226013392,0.2707365453,0.2623255849,0.3942559958,0.3254241347,0.1623783857,-0.1076942012,0.1056443229,-0.1841582358,0.3969781995,0.2862301171,0.0432810932,0.0279178899,-0.0078068012,0.3893857896,-0.1983547658,0.0260479581,0.3512432873,-0.0967423618,-0.1660416126,-0.6548130512,0.17357862,-0.1757688969,-0.2077556103,0.4948207438,0.0051999725,-0.1810278147,0.5770657659,0.0360416919,0.9330322742,0.2672248483,-0.1094423905,0.2783369422,-0.1210113466,0.612965703,0.0159467701,0.277343899,-0.3774955273,-0.3241224885,-0.0243557822,0.0032225286,0.0388527885,-0.1578301191,-0.1630984992,0.238726899,-0.233989805,-0.0200340506,-0.0329130441,0.3232554197,0.2424539328,0.1481516361,-0.1084030494,0.1145466119,-0.0608149692,-0.1897985637,-0.1467810422,0.0128434738,0.1084817946,0.0014939425,-0.3961191773,0.032109648,-0.1198762879,0.1515937299,-0.3794673681,-0.3329626322,0.2918035388,0.2802639008,-0.1389145106,0.1128718928,-0.156388551,0.1829782724,0.0113465497,0.181289658,0.0638161004,0.2905876935,0.2833430469,-0.0767332315,-0.2789282501,0.1410662979,0.0836604014,-0.0352023803,-0.2029493302,-0.185880959,0.1857665479,-0.277907759,0.1268288493,-0.1243294179,-0.3048667312,-0.1910587847,0.1262686253,0.2097568065,-0.0669730082,0.2531118393,0.1537329853,-0.0373004451,-0.1281880587,0.2323104441,0.0470716655,0.0388124958,0.5362433791,0.4200657308,-0.1564955562,-0.1853557229,0.1352237463,-0.1985569149,-0.1128123924,0.0336666778,-0.4092645347,0.0790371373,0.0843415558,0.1507214308,0.1443125755,0.1743270755,-0.4357252121,-0.2476724833,-0.0197396055,0.2915032208,-0.1539634466,0.1664569527,-0.4321200252,-0.141108498,0.1247308552,-0.1702901423,-0.3006685972,-0.190025717,0.0198921841,0.181043312,0.0269484352,0.1825227886,-0.2085029036,-0.4424799681,0.0355275087,-0.2638809681,-0.0845437199,-0.1637734473,-0.0852388814,0.1769664884,0.0740286261,-0.441613853,-0.3148105443,-0.2439830303,-0.0189493522,-0.063104406,-0.3076319993,-0.047821641,0.0404856764,0.0820248723,-0.0307593718,-0.2753333151,-0.0077422154,0.2412666827,0.1193006262,0.541461885,0.2563714385,0.2153195143,0.1639128923,-0.1250135452,-0.2580597401,0.1643506736,-0.3087143302,0.1652209312,0.3549098074,-0.3322285712,0.2556830645,0.3808780015,0.4932388961,0.4313769042,-0.1432711631,-0.2789634764,0.4424249232,0.1640637368,-0.4123229384,-0.1763328463,0.2795851529,0.2438185513,0.0173587725,0.1030137315,0.136197865,-0.1607579887,-0.2585865855,0.0680326298,0.4175924957,-0.203919515,0.3002223074,0.4241982698,-0.212717548,-0.0639845207,0.1778074652,0.2166756988,0.1435504109,0.6396567822,0.1417242885,0.0758812279,0.4841813743,0.1691903025,0.1241489053,-0.2833169997,-0.1818279028,0.2028390914,-0.099501811,0.0930603817,0.259293735,0.04682054,0.063155517,-0.3688143194,0.0459530018,0.2396122366,-0.0075236442,0.2120814323,-0.3550190926,-0.2509007454,-0.0494616255,-0.0261533819,-0.0940736085,-0.0751662403,-0.1726794243,-0.0033784211,-0.3168993294,-0.4275234044,0.3732615113,0.1240848452,0.2268958837,-0.2028679252,-0.0114721479,0.3269617856,0.1735690534,-0.0246689357,0.4546066821,0.2917703092,0.1306448132,-0.2347610295,0.0310389008,-0.0664144084,-0.1432176083,-0.1319899261,0.132255286,0.0302545149,0.1951593608,0.1808878481,0.076801993,-0.053962335,0.4734232128,0.2337242514,0.0584048405,-0.2925067544,0.3399984241,0.145849213,-0.5043171048,0.0774904564,-0.3538134694,-0.2001726329,0.1857924014,0.7511013746,0.031302385,0.1374648809,-0.0585534424,0.061800804,0.0923976228,0.3376110196,0.4668429494,0.0355729833,-0.1303791851,-0.2312413007,-0.8589652181,0.0914343521,-0.2171116024,-0.2172402143,0.5133656263,0.1280394047,-0.0502474234,0.2213602513,0.3961088061,-0.0102151269,0.0270923283,0.0515248701,-0.3876031041,-0.1328259557,-0.1321590841,0.3818224669,-0.3627648056,-0.1248889267,0.2687458694,-0.4400665462,-0.0286197234,-0.0540294871,0.0377760753,-0.0401768088,0.1479733884,0.1913590431,0.1345200837,0.1358108371,0.2102367878,-0.0262315515,-0.4110262394,-0.2733960152,-0.2440471798,0.2069380134,0.098176308,0.310220629,-0.3268450499,0.035486497,-0.2343169451,0.2626631558,-0.0487470105,0.192589432,-0.4621894062,-0.0339333825,-0.1455089748,0.2107549608,0.0786109194,0.1603516042,-0.0111095784,0.0877976194,-0.1460526586,-0.301722914,0.4168035388,-0.389662683,-0.1787187904,-0.0096131312,0.1982851475,0.3603422344,0.0374178886,-0.327151835,-0.1917618066,0.2938884199,-0.1424470246,0.0055394652,-0.0835715681,-0.1831873506,-0.0577196181,0.1430443674,-0.0101630688,0.1501576304,-0.2010464221,0.1263785064,0.0989648774]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/897","title":"Dataset viewer issues","comments":"9","body":"I was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n","comment_length":1,"text":"Dataset viewer issues\nI was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n\n9","embeddings":[-0.2186932266,0.2789271474,-0.026432747,0.2816513479,-0.0023903828,0.1115044877,0.2981258929,0.2988632917,-0.1956236809,0.0647588521,-0.0076386253,0.2664061487,-0.3950020671,0.0988750979,-0.1122838408,-0.2007787526,0.0220978931,-0.0146634579,0.0232188925,-0.0943433568,-0.1718729287,-0.2318453193,-0.3681103587,0.0472469144,0.1730451733,-0.1287799329,-0.0984149203,-0.0248546954,-0.1657533944,-0.6458482146,0.0453227796,0.2832351029,0.1305533051,0.4277637899,-0.0001141002,-0.0952439457,0.5284169316,0.2207832187,-0.4246532321,-0.1656228304,-0.1231052652,-0.3184152842,0.4010361731,-0.1876373738,0.2975274324,-0.34791255,0.0207551885,-0.1845204532,0.0901871175,0.469602257,0.1956328005,0.1108126119,-0.1084176004,0.1752130687,0.3652741909,0.2491191179,-0.2797198892,-0.1015249267,0.1632782519,-0.0791658387,-0.025709914,0.3140527904,-0.1640222073,-0.0861883536,0.3573250473,-0.1882148087,-0.0757901296,-0.4510447383,-0.0953087807,0.2241510004,0.7972296476,-0.3495242298,-0.3797179759,-0.3601836562,0.0131202983,-0.1577013135,0.1619466543,0.0079309354,-0.1782668978,0.0668451935,-0.2203083187,-0.3423940837,-0.1549729556,0.0965098739,-0.2113793194,0.6253469586,0.1779181063,0.0807632133,0.1139025018,-0.1835037321,0.5544168949,-0.1228946298,-0.1087577194,0.1986276656,-0.2692966461,-0.1737065315,0.461104244,-0.1622395366,-0.1637828648,0.4153255224,0.2714703977,0.0435075462,-0.3141144812,0.009632552,0.4544065893,0.1671871543,0.147337541,0.1572187394,0.3739819229,-0.038006708,0.3260328472,0.1013467759,0.0978917032,-0.0086792316,-0.2973321676,-0.116244182,0.4545963407,-0.2330589294,-0.303227216,0.1526127011,-0.1009672284,-0.263060987,0.1532100588,0.219112128,-0.0706097782,0.0922878385,-0.1988677233,0.1861469299,-0.2814712524,-0.1896958649,-0.1201233566,-0.1157088652,-0.1896719784,-0.3266671598,0.1634646803,-0.2876240313,0.119710207,-0.1003617421,-0.1547210664,0.194593668,0.0908226743,-0.1636734307,-0.0009145842,0.0688804686,-0.2302842587,0.1360697597,0.1197247058,-0.2654082775,0.0479442663,0.2560563684,0.2586598098,-0.1782319695,-0.1489072144,0.1788307279,-0.2246099263,0.0703034028,0.0440880843,0.3644196093,-0.2849558592,-0.2395537049,-0.1893162578,-0.2350794524,-0.1941615939,-0.187835604,0.1720422506,0.5530490279,-0.4830772579,-0.0465624295,-0.4554136395,-0.0557204336,0.0420507528,0.059990108,-0.1993482709,0.1053986698,-0.3029534817,-0.0673145503,0.5750009418,-0.219153285,-0.4035130143,0.2735553682,-0.172700122,0.0276703667,0.0239073969,0.1319134533,0.1925355643,-0.3088919222,-0.3873092234,-0.0334397294,-0.0646378994,-0.2213612795,-0.3128040135,-0.1074304804,0.5276122689,0.036488194,0.13503021,-0.0540118702,0.2751266658,0.0094174203,0.3650549352,0.1358383149,0.0362217166,0.0742885768,-0.0089339912,0.1816091686,-0.0343146659,-0.1788937896,-0.370862782,0.0983280316,0.1578837186,0.0319679044,0.0578098372,-0.0395850949,-0.0441822745,0.176623553,-0.2002382427,0.0091595007,0.1059279591,0.0371709205,-0.228386879,0.150171414,-0.3601384461,-0.3355703354,-0.0756464526,0.1033099219,-0.2184205651,0.4458243251,-0.0775557607,-0.0546324924,0.2532230914,0.3302969038,0.2176555693,-0.1338980347,-0.1547198892,0.225530237,0.3120882213,0.0964817703,0.2038223445,0.2202825546,0.4631510973,-0.4166270792,0.4512762427,0.4248676896,0.0157839339,-0.0111361668,-0.4655109346,0.1208941489,-0.4010038972,0.3006582558,0.1231233552,0.0035366071,0.3432372808,0.0029793673,0.0387485735,-0.1294725239,0.2690471411,0.0817266405,-0.0146125136,-0.0484843403,-0.4041103125,0.2343586981,-0.1110977903,0.142692402,-0.2079056054,0.1828098595,-0.2422186732,-0.1872489154,0.2078428715,0.1649779528,0.1381793618,0.1974901557,0.00771394,0.1722227633,0.1840904504,-0.1922398806,0.2213065922,0.2634327114,-0.2253491729,0.1050514877,0.11574696,0.0659233034,-0.2038090527,-0.3943380713,0.1596047878,0.3535887003,-0.2750309408,-0.172228083,-0.383197397,-0.708920002,0.0878656879,-0.0594570749,-0.3682579696,-0.550560534,0.1306333691,0.3110088706,-0.3960461915,0.2758713663,-0.5261030793,0.0930711403,-0.0833637863,0.377148807,0.029483242,-0.315015018,-0.1494927406,0.1138873547,0.3704943359,0.3368929923,0.3391319811,0.2285357863,0.0902660713,-0.5124884248,-0.3946644664,0.1821422726,-0.0811305791,0.3599928916,0.0334619135,-0.0304460786,-0.0033648976,-0.0145208053,0.1848327667,-0.0334529392,0.0808227807,0.0377494097,0.2265972346,-0.0568405129,0.0043872967,-0.2615486085,-0.1642236412,-0.5408262014,0.0232084766,0.1773352772,0.2984770536,0.2128200233,-0.110032782,0.0811095014,0.1700314879,-0.1284519434,0.08159215,0.2522673607,0.2351938635,-0.1482935697,-0.3752676547,0.1431304067,0.1678874493,-0.1806288958,-0.0289230738,-0.6907394528,0.2365356535,0.0004746777,0.0933537111,0.0852360874,-0.1197598875,0.2346420884,0.3941362798,-0.0285920184,0.0604873337,-0.1572939903,-0.1623924971,-0.0272981729,0.2487436384,0.0613057353,0.2927386463,-0.0788261741,0.4196695387,0.1608722657,0.1330552101,0.3300288022,-0.1292558014,0.3863185942,0.0880506858,-0.364426285,0.3801051974,0.2277207226,-0.2473003566,0.3891529143,-0.0468089357,-0.2272368819,-0.1322384924,0.111875169,-0.0174210239,-0.2235319912,0.0054794597,-0.3366178274,0.0723116398,0.1084828675,0.1603690684,-0.087550588,-0.4027015269,-0.0639728084,0.0354260877,0.1260525137,-0.397769779,-0.6519774199,-0.1440824121,-0.4036369324,0.3362410069,0.1335719079,0.2592405677,-0.1384121776,-0.0428604856,0.0075490596,0.3046002984,0.3338590562,-0.2926357687,0.0308585726,0.2173912674,0.2529658377,-0.3621837795,-0.1252756864,-0.1247658134,0.468627125,-0.0056123338,0.1088866964,0.1177692637,-0.1460490376,0.1244101599,0.0838265717,-0.3346751928,-0.0230384395,0.0858213305,-0.1470863372,-0.1134802401,-0.2595456243,-0.2903983891,-0.0807918385,-0.1117307097,-0.0160472374,0.0051927869,-0.171682179,0.1276297867,0.1367404908,0.4477116764,-0.0723016411,0.2226013392,0.2707365453,0.2623255849,0.3942559958,0.3254241347,0.1623783857,-0.1076942012,0.1056443229,-0.1841582358,0.3969781995,0.2862301171,0.0432810932,0.0279178899,-0.0078068012,0.3893857896,-0.1983547658,0.0260479581,0.3512432873,-0.0967423618,-0.1660416126,-0.6548130512,0.17357862,-0.1757688969,-0.2077556103,0.4948207438,0.0051999725,-0.1810278147,0.5770657659,0.0360416919,0.9330322742,0.2672248483,-0.1094423905,0.2783369422,-0.1210113466,0.612965703,0.0159467701,0.277343899,-0.3774955273,-0.3241224885,-0.0243557822,0.0032225286,0.0388527885,-0.1578301191,-0.1630984992,0.238726899,-0.233989805,-0.0200340506,-0.0329130441,0.3232554197,0.2424539328,0.1481516361,-0.1084030494,0.1145466119,-0.0608149692,-0.1897985637,-0.1467810422,0.0128434738,0.1084817946,0.0014939425,-0.3961191773,0.032109648,-0.1198762879,0.1515937299,-0.3794673681,-0.3329626322,0.2918035388,0.2802639008,-0.1389145106,0.1128718928,-0.156388551,0.1829782724,0.0113465497,0.181289658,0.0638161004,0.2905876935,0.2833430469,-0.0767332315,-0.2789282501,0.1410662979,0.0836604014,-0.0352023803,-0.2029493302,-0.185880959,0.1857665479,-0.277907759,0.1268288493,-0.1243294179,-0.3048667312,-0.1910587847,0.1262686253,0.2097568065,-0.0669730082,0.2531118393,0.1537329853,-0.0373004451,-0.1281880587,0.2323104441,0.0470716655,0.0388124958,0.5362433791,0.4200657308,-0.1564955562,-0.1853557229,0.1352237463,-0.1985569149,-0.1128123924,0.0336666778,-0.4092645347,0.0790371373,0.0843415558,0.1507214308,0.1443125755,0.1743270755,-0.4357252121,-0.2476724833,-0.0197396055,0.2915032208,-0.1539634466,0.1664569527,-0.4321200252,-0.141108498,0.1247308552,-0.1702901423,-0.3006685972,-0.190025717,0.0198921841,0.181043312,0.0269484352,0.1825227886,-0.2085029036,-0.4424799681,0.0355275087,-0.2638809681,-0.0845437199,-0.1637734473,-0.0852388814,0.1769664884,0.0740286261,-0.441613853,-0.3148105443,-0.2439830303,-0.0189493522,-0.063104406,-0.3076319993,-0.047821641,0.0404856764,0.0820248723,-0.0307593718,-0.2753333151,-0.0077422154,0.2412666827,0.1193006262,0.541461885,0.2563714385,0.2153195143,0.1639128923,-0.1250135452,-0.2580597401,0.1643506736,-0.3087143302,0.1652209312,0.3549098074,-0.3322285712,0.2556830645,0.3808780015,0.4932388961,0.4313769042,-0.1432711631,-0.2789634764,0.4424249232,0.1640637368,-0.4123229384,-0.1763328463,0.2795851529,0.2438185513,0.0173587725,0.1030137315,0.136197865,-0.1607579887,-0.2585865855,0.0680326298,0.4175924957,-0.203919515,0.3002223074,0.4241982698,-0.212717548,-0.0639845207,0.1778074652,0.2166756988,0.1435504109,0.6396567822,0.1417242885,0.0758812279,0.4841813743,0.1691903025,0.1241489053,-0.2833169997,-0.1818279028,0.2028390914,-0.099501811,0.0930603817,0.259293735,0.04682054,0.063155517,-0.3688143194,0.0459530018,0.2396122366,-0.0075236442,0.2120814323,-0.3550190926,-0.2509007454,-0.0494616255,-0.0261533819,-0.0940736085,-0.0751662403,-0.1726794243,-0.0033784211,-0.3168993294,-0.4275234044,0.3732615113,0.1240848452,0.2268958837,-0.2028679252,-0.0114721479,0.3269617856,0.1735690534,-0.0246689357,0.4546066821,0.2917703092,0.1306448132,-0.2347610295,0.0310389008,-0.0664144084,-0.1432176083,-0.1319899261,0.132255286,0.0302545149,0.1951593608,0.1808878481,0.076801993,-0.053962335,0.4734232128,0.2337242514,0.0584048405,-0.2925067544,0.3399984241,0.145849213,-0.5043171048,0.0774904564,-0.3538134694,-0.2001726329,0.1857924014,0.7511013746,0.031302385,0.1374648809,-0.0585534424,0.061800804,0.0923976228,0.3376110196,0.4668429494,0.0355729833,-0.1303791851,-0.2312413007,-0.8589652181,0.0914343521,-0.2171116024,-0.2172402143,0.5133656263,0.1280394047,-0.0502474234,0.2213602513,0.3961088061,-0.0102151269,0.0270923283,0.0515248701,-0.3876031041,-0.1328259557,-0.1321590841,0.3818224669,-0.3627648056,-0.1248889267,0.2687458694,-0.4400665462,-0.0286197234,-0.0540294871,0.0377760753,-0.0401768088,0.1479733884,0.1913590431,0.1345200837,0.1358108371,0.2102367878,-0.0262315515,-0.4110262394,-0.2733960152,-0.2440471798,0.2069380134,0.098176308,0.310220629,-0.3268450499,0.035486497,-0.2343169451,0.2626631558,-0.0487470105,0.192589432,-0.4621894062,-0.0339333825,-0.1455089748,0.2107549608,0.0786109194,0.1603516042,-0.0111095784,0.0877976194,-0.1460526586,-0.301722914,0.4168035388,-0.389662683,-0.1787187904,-0.0096131312,0.1982851475,0.3603422344,0.0374178886,-0.327151835,-0.1917618066,0.2938884199,-0.1424470246,0.0055394652,-0.0835715681,-0.1831873506,-0.0577196181,0.1430443674,-0.0101630688,0.1501576304,-0.2010464221,0.1263785064,0.0989648774]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/897","title":"Dataset viewer issues","comments":"\u200f\u2800\u200f\u200f\u200f\u2800\u200f\u200f\u200f\u2800 \u200f\u2800 ","body":"I was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n","comment_length":2,"text":"Dataset viewer issues\nI was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n\n\u200f\u2800\u200f\u200f\u200f\u2800\u200f\u200f\u200f\u2800 \u200f\u2800 ","embeddings":[-0.2186932266,0.2789271474,-0.026432747,0.2816513479,-0.0023903828,0.1115044877,0.2981258929,0.2988632917,-0.1956236809,0.0647588521,-0.0076386253,0.2664061487,-0.3950020671,0.0988750979,-0.1122838408,-0.2007787526,0.0220978931,-0.0146634579,0.0232188925,-0.0943433568,-0.1718729287,-0.2318453193,-0.3681103587,0.0472469144,0.1730451733,-0.1287799329,-0.0984149203,-0.0248546954,-0.1657533944,-0.6458482146,0.0453227796,0.2832351029,0.1305533051,0.4277637899,-0.0001141002,-0.0952439457,0.5284169316,0.2207832187,-0.4246532321,-0.1656228304,-0.1231052652,-0.3184152842,0.4010361731,-0.1876373738,0.2975274324,-0.34791255,0.0207551885,-0.1845204532,0.0901871175,0.469602257,0.1956328005,0.1108126119,-0.1084176004,0.1752130687,0.3652741909,0.2491191179,-0.2797198892,-0.1015249267,0.1632782519,-0.0791658387,-0.025709914,0.3140527904,-0.1640222073,-0.0861883536,0.3573250473,-0.1882148087,-0.0757901296,-0.4510447383,-0.0953087807,0.2241510004,0.7972296476,-0.3495242298,-0.3797179759,-0.3601836562,0.0131202983,-0.1577013135,0.1619466543,0.0079309354,-0.1782668978,0.0668451935,-0.2203083187,-0.3423940837,-0.1549729556,0.0965098739,-0.2113793194,0.6253469586,0.1779181063,0.0807632133,0.1139025018,-0.1835037321,0.5544168949,-0.1228946298,-0.1087577194,0.1986276656,-0.2692966461,-0.1737065315,0.461104244,-0.1622395366,-0.1637828648,0.4153255224,0.2714703977,0.0435075462,-0.3141144812,0.009632552,0.4544065893,0.1671871543,0.147337541,0.1572187394,0.3739819229,-0.038006708,0.3260328472,0.1013467759,0.0978917032,-0.0086792316,-0.2973321676,-0.116244182,0.4545963407,-0.2330589294,-0.303227216,0.1526127011,-0.1009672284,-0.263060987,0.1532100588,0.219112128,-0.0706097782,0.0922878385,-0.1988677233,0.1861469299,-0.2814712524,-0.1896958649,-0.1201233566,-0.1157088652,-0.1896719784,-0.3266671598,0.1634646803,-0.2876240313,0.119710207,-0.1003617421,-0.1547210664,0.194593668,0.0908226743,-0.1636734307,-0.0009145842,0.0688804686,-0.2302842587,0.1360697597,0.1197247058,-0.2654082775,0.0479442663,0.2560563684,0.2586598098,-0.1782319695,-0.1489072144,0.1788307279,-0.2246099263,0.0703034028,0.0440880843,0.3644196093,-0.2849558592,-0.2395537049,-0.1893162578,-0.2350794524,-0.1941615939,-0.187835604,0.1720422506,0.5530490279,-0.4830772579,-0.0465624295,-0.4554136395,-0.0557204336,0.0420507528,0.059990108,-0.1993482709,0.1053986698,-0.3029534817,-0.0673145503,0.5750009418,-0.219153285,-0.4035130143,0.2735553682,-0.172700122,0.0276703667,0.0239073969,0.1319134533,0.1925355643,-0.3088919222,-0.3873092234,-0.0334397294,-0.0646378994,-0.2213612795,-0.3128040135,-0.1074304804,0.5276122689,0.036488194,0.13503021,-0.0540118702,0.2751266658,0.0094174203,0.3650549352,0.1358383149,0.0362217166,0.0742885768,-0.0089339912,0.1816091686,-0.0343146659,-0.1788937896,-0.370862782,0.0983280316,0.1578837186,0.0319679044,0.0578098372,-0.0395850949,-0.0441822745,0.176623553,-0.2002382427,0.0091595007,0.1059279591,0.0371709205,-0.228386879,0.150171414,-0.3601384461,-0.3355703354,-0.0756464526,0.1033099219,-0.2184205651,0.4458243251,-0.0775557607,-0.0546324924,0.2532230914,0.3302969038,0.2176555693,-0.1338980347,-0.1547198892,0.225530237,0.3120882213,0.0964817703,0.2038223445,0.2202825546,0.4631510973,-0.4166270792,0.4512762427,0.4248676896,0.0157839339,-0.0111361668,-0.4655109346,0.1208941489,-0.4010038972,0.3006582558,0.1231233552,0.0035366071,0.3432372808,0.0029793673,0.0387485735,-0.1294725239,0.2690471411,0.0817266405,-0.0146125136,-0.0484843403,-0.4041103125,0.2343586981,-0.1110977903,0.142692402,-0.2079056054,0.1828098595,-0.2422186732,-0.1872489154,0.2078428715,0.1649779528,0.1381793618,0.1974901557,0.00771394,0.1722227633,0.1840904504,-0.1922398806,0.2213065922,0.2634327114,-0.2253491729,0.1050514877,0.11574696,0.0659233034,-0.2038090527,-0.3943380713,0.1596047878,0.3535887003,-0.2750309408,-0.172228083,-0.383197397,-0.708920002,0.0878656879,-0.0594570749,-0.3682579696,-0.550560534,0.1306333691,0.3110088706,-0.3960461915,0.2758713663,-0.5261030793,0.0930711403,-0.0833637863,0.377148807,0.029483242,-0.315015018,-0.1494927406,0.1138873547,0.3704943359,0.3368929923,0.3391319811,0.2285357863,0.0902660713,-0.5124884248,-0.3946644664,0.1821422726,-0.0811305791,0.3599928916,0.0334619135,-0.0304460786,-0.0033648976,-0.0145208053,0.1848327667,-0.0334529392,0.0808227807,0.0377494097,0.2265972346,-0.0568405129,0.0043872967,-0.2615486085,-0.1642236412,-0.5408262014,0.0232084766,0.1773352772,0.2984770536,0.2128200233,-0.110032782,0.0811095014,0.1700314879,-0.1284519434,0.08159215,0.2522673607,0.2351938635,-0.1482935697,-0.3752676547,0.1431304067,0.1678874493,-0.1806288958,-0.0289230738,-0.6907394528,0.2365356535,0.0004746777,0.0933537111,0.0852360874,-0.1197598875,0.2346420884,0.3941362798,-0.0285920184,0.0604873337,-0.1572939903,-0.1623924971,-0.0272981729,0.2487436384,0.0613057353,0.2927386463,-0.0788261741,0.4196695387,0.1608722657,0.1330552101,0.3300288022,-0.1292558014,0.3863185942,0.0880506858,-0.364426285,0.3801051974,0.2277207226,-0.2473003566,0.3891529143,-0.0468089357,-0.2272368819,-0.1322384924,0.111875169,-0.0174210239,-0.2235319912,0.0054794597,-0.3366178274,0.0723116398,0.1084828675,0.1603690684,-0.087550588,-0.4027015269,-0.0639728084,0.0354260877,0.1260525137,-0.397769779,-0.6519774199,-0.1440824121,-0.4036369324,0.3362410069,0.1335719079,0.2592405677,-0.1384121776,-0.0428604856,0.0075490596,0.3046002984,0.3338590562,-0.2926357687,0.0308585726,0.2173912674,0.2529658377,-0.3621837795,-0.1252756864,-0.1247658134,0.468627125,-0.0056123338,0.1088866964,0.1177692637,-0.1460490376,0.1244101599,0.0838265717,-0.3346751928,-0.0230384395,0.0858213305,-0.1470863372,-0.1134802401,-0.2595456243,-0.2903983891,-0.0807918385,-0.1117307097,-0.0160472374,0.0051927869,-0.171682179,0.1276297867,0.1367404908,0.4477116764,-0.0723016411,0.2226013392,0.2707365453,0.2623255849,0.3942559958,0.3254241347,0.1623783857,-0.1076942012,0.1056443229,-0.1841582358,0.3969781995,0.2862301171,0.0432810932,0.0279178899,-0.0078068012,0.3893857896,-0.1983547658,0.0260479581,0.3512432873,-0.0967423618,-0.1660416126,-0.6548130512,0.17357862,-0.1757688969,-0.2077556103,0.4948207438,0.0051999725,-0.1810278147,0.5770657659,0.0360416919,0.9330322742,0.2672248483,-0.1094423905,0.2783369422,-0.1210113466,0.612965703,0.0159467701,0.277343899,-0.3774955273,-0.3241224885,-0.0243557822,0.0032225286,0.0388527885,-0.1578301191,-0.1630984992,0.238726899,-0.233989805,-0.0200340506,-0.0329130441,0.3232554197,0.2424539328,0.1481516361,-0.1084030494,0.1145466119,-0.0608149692,-0.1897985637,-0.1467810422,0.0128434738,0.1084817946,0.0014939425,-0.3961191773,0.032109648,-0.1198762879,0.1515937299,-0.3794673681,-0.3329626322,0.2918035388,0.2802639008,-0.1389145106,0.1128718928,-0.156388551,0.1829782724,0.0113465497,0.181289658,0.0638161004,0.2905876935,0.2833430469,-0.0767332315,-0.2789282501,0.1410662979,0.0836604014,-0.0352023803,-0.2029493302,-0.185880959,0.1857665479,-0.277907759,0.1268288493,-0.1243294179,-0.3048667312,-0.1910587847,0.1262686253,0.2097568065,-0.0669730082,0.2531118393,0.1537329853,-0.0373004451,-0.1281880587,0.2323104441,0.0470716655,0.0388124958,0.5362433791,0.4200657308,-0.1564955562,-0.1853557229,0.1352237463,-0.1985569149,-0.1128123924,0.0336666778,-0.4092645347,0.0790371373,0.0843415558,0.1507214308,0.1443125755,0.1743270755,-0.4357252121,-0.2476724833,-0.0197396055,0.2915032208,-0.1539634466,0.1664569527,-0.4321200252,-0.141108498,0.1247308552,-0.1702901423,-0.3006685972,-0.190025717,0.0198921841,0.181043312,0.0269484352,0.1825227886,-0.2085029036,-0.4424799681,0.0355275087,-0.2638809681,-0.0845437199,-0.1637734473,-0.0852388814,0.1769664884,0.0740286261,-0.441613853,-0.3148105443,-0.2439830303,-0.0189493522,-0.063104406,-0.3076319993,-0.047821641,0.0404856764,0.0820248723,-0.0307593718,-0.2753333151,-0.0077422154,0.2412666827,0.1193006262,0.541461885,0.2563714385,0.2153195143,0.1639128923,-0.1250135452,-0.2580597401,0.1643506736,-0.3087143302,0.1652209312,0.3549098074,-0.3322285712,0.2556830645,0.3808780015,0.4932388961,0.4313769042,-0.1432711631,-0.2789634764,0.4424249232,0.1640637368,-0.4123229384,-0.1763328463,0.2795851529,0.2438185513,0.0173587725,0.1030137315,0.136197865,-0.1607579887,-0.2585865855,0.0680326298,0.4175924957,-0.203919515,0.3002223074,0.4241982698,-0.212717548,-0.0639845207,0.1778074652,0.2166756988,0.1435504109,0.6396567822,0.1417242885,0.0758812279,0.4841813743,0.1691903025,0.1241489053,-0.2833169997,-0.1818279028,0.2028390914,-0.099501811,0.0930603817,0.259293735,0.04682054,0.063155517,-0.3688143194,0.0459530018,0.2396122366,-0.0075236442,0.2120814323,-0.3550190926,-0.2509007454,-0.0494616255,-0.0261533819,-0.0940736085,-0.0751662403,-0.1726794243,-0.0033784211,-0.3168993294,-0.4275234044,0.3732615113,0.1240848452,0.2268958837,-0.2028679252,-0.0114721479,0.3269617856,0.1735690534,-0.0246689357,0.4546066821,0.2917703092,0.1306448132,-0.2347610295,0.0310389008,-0.0664144084,-0.1432176083,-0.1319899261,0.132255286,0.0302545149,0.1951593608,0.1808878481,0.076801993,-0.053962335,0.4734232128,0.2337242514,0.0584048405,-0.2925067544,0.3399984241,0.145849213,-0.5043171048,0.0774904564,-0.3538134694,-0.2001726329,0.1857924014,0.7511013746,0.031302385,0.1374648809,-0.0585534424,0.061800804,0.0923976228,0.3376110196,0.4668429494,0.0355729833,-0.1303791851,-0.2312413007,-0.8589652181,0.0914343521,-0.2171116024,-0.2172402143,0.5133656263,0.1280394047,-0.0502474234,0.2213602513,0.3961088061,-0.0102151269,0.0270923283,0.0515248701,-0.3876031041,-0.1328259557,-0.1321590841,0.3818224669,-0.3627648056,-0.1248889267,0.2687458694,-0.4400665462,-0.0286197234,-0.0540294871,0.0377760753,-0.0401768088,0.1479733884,0.1913590431,0.1345200837,0.1358108371,0.2102367878,-0.0262315515,-0.4110262394,-0.2733960152,-0.2440471798,0.2069380134,0.098176308,0.310220629,-0.3268450499,0.035486497,-0.2343169451,0.2626631558,-0.0487470105,0.192589432,-0.4621894062,-0.0339333825,-0.1455089748,0.2107549608,0.0786109194,0.1603516042,-0.0111095784,0.0877976194,-0.1460526586,-0.301722914,0.4168035388,-0.389662683,-0.1787187904,-0.0096131312,0.1982851475,0.3603422344,0.0374178886,-0.327151835,-0.1917618066,0.2938884199,-0.1424470246,0.0055394652,-0.0835715681,-0.1831873506,-0.0577196181,0.1430443674,-0.0101630688,0.1501576304,-0.2010464221,0.1263785064,0.0989648774]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/897","title":"Dataset viewer issues","comments":"\u200f\u2800\u200f\u200f\u200f\u2800\u200f\u200f\u200f\u2800 \u200f\u2800 ","body":"I was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n","comment_length":2,"text":"Dataset viewer issues\nI was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues\/bugs though:\r\n\r\n- the URL is still under `nlp`, perhaps an alias for `datasets` can be made\r\n- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user\r\n\r\n```bash\r\nIndexError: list index out of range\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 316, in <module>\r\n st.table(style)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 122, in wrapped_method\r\n return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 367, in _enqueue_new_element_delta\r\n rv = marshall_element(msg.delta.new_element)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 120, in marshall_element\r\n return method(dg, element, *args, **kwargs)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/DeltaGenerator.py\", line 2944, in table\r\n data_frame_proto.marshall_data_frame(data, element.table)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 54, in marshall_data_frame\r\n _marshall_styles(proto_df.style, df, styler)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/elements\/data_frame_proto.py\", line 73, in _marshall_styles\r\n translated_style = styler._translate()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/pandas\/io\/formats\/style.py\", line 351, in _translate\r\n * (len(clabels[0]) - len(hidden_columns))\r\n```\r\n\r\n- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wmt19&config=cs-en). This problem goes away when you enable \"List view\", because then some syntax highlighteris used, and the special characters are coded correctly.\r\n\n\u200f\u2800\u200f\u200f\u200f\u2800\u200f\u200f\u200f\u2800 \u200f\u2800 ","embeddings":[-0.2186932266,0.2789271474,-0.026432747,0.2816513479,-0.0023903828,0.1115044877,0.2981258929,0.2988632917,-0.1956236809,0.0647588521,-0.0076386253,0.2664061487,-0.3950020671,0.0988750979,-0.1122838408,-0.2007787526,0.0220978931,-0.0146634579,0.0232188925,-0.0943433568,-0.1718729287,-0.2318453193,-0.3681103587,0.0472469144,0.1730451733,-0.1287799329,-0.0984149203,-0.0248546954,-0.1657533944,-0.6458482146,0.0453227796,0.2832351029,0.1305533051,0.4277637899,-0.0001141002,-0.0952439457,0.5284169316,0.2207832187,-0.4246532321,-0.1656228304,-0.1231052652,-0.3184152842,0.4010361731,-0.1876373738,0.2975274324,-0.34791255,0.0207551885,-0.1845204532,0.0901871175,0.469602257,0.1956328005,0.1108126119,-0.1084176004,0.1752130687,0.3652741909,0.2491191179,-0.2797198892,-0.1015249267,0.1632782519,-0.0791658387,-0.025709914,0.3140527904,-0.1640222073,-0.0861883536,0.3573250473,-0.1882148087,-0.0757901296,-0.4510447383,-0.0953087807,0.2241510004,0.7972296476,-0.3495242298,-0.3797179759,-0.3601836562,0.0131202983,-0.1577013135,0.1619466543,0.0079309354,-0.1782668978,0.0668451935,-0.2203083187,-0.3423940837,-0.1549729556,0.0965098739,-0.2113793194,0.6253469586,0.1779181063,0.0807632133,0.1139025018,-0.1835037321,0.5544168949,-0.1228946298,-0.1087577194,0.1986276656,-0.2692966461,-0.1737065315,0.461104244,-0.1622395366,-0.1637828648,0.4153255224,0.2714703977,0.0435075462,-0.3141144812,0.009632552,0.4544065893,0.1671871543,0.147337541,0.1572187394,0.3739819229,-0.038006708,0.3260328472,0.1013467759,0.0978917032,-0.0086792316,-0.2973321676,-0.116244182,0.4545963407,-0.2330589294,-0.303227216,0.1526127011,-0.1009672284,-0.263060987,0.1532100588,0.219112128,-0.0706097782,0.0922878385,-0.1988677233,0.1861469299,-0.2814712524,-0.1896958649,-0.1201233566,-0.1157088652,-0.1896719784,-0.3266671598,0.1634646803,-0.2876240313,0.119710207,-0.1003617421,-0.1547210664,0.194593668,0.0908226743,-0.1636734307,-0.0009145842,0.0688804686,-0.2302842587,0.1360697597,0.1197247058,-0.2654082775,0.0479442663,0.2560563684,0.2586598098,-0.1782319695,-0.1489072144,0.1788307279,-0.2246099263,0.0703034028,0.0440880843,0.3644196093,-0.2849558592,-0.2395537049,-0.1893162578,-0.2350794524,-0.1941615939,-0.187835604,0.1720422506,0.5530490279,-0.4830772579,-0.0465624295,-0.4554136395,-0.0557204336,0.0420507528,0.059990108,-0.1993482709,0.1053986698,-0.3029534817,-0.0673145503,0.5750009418,-0.219153285,-0.4035130143,0.2735553682,-0.172700122,0.0276703667,0.0239073969,0.1319134533,0.1925355643,-0.3088919222,-0.3873092234,-0.0334397294,-0.0646378994,-0.2213612795,-0.3128040135,-0.1074304804,0.5276122689,0.036488194,0.13503021,-0.0540118702,0.2751266658,0.0094174203,0.3650549352,0.1358383149,0.0362217166,0.0742885768,-0.0089339912,0.1816091686,-0.0343146659,-0.1788937896,-0.370862782,0.0983280316,0.1578837186,0.0319679044,0.0578098372,-0.0395850949,-0.0441822745,0.176623553,-0.2002382427,0.0091595007,0.1059279591,0.0371709205,-0.228386879,0.150171414,-0.3601384461,-0.3355703354,-0.0756464526,0.1033099219,-0.2184205651,0.4458243251,-0.0775557607,-0.0546324924,0.2532230914,0.3302969038,0.2176555693,-0.1338980347,-0.1547198892,0.225530237,0.3120882213,0.0964817703,0.2038223445,0.2202825546,0.4631510973,-0.4166270792,0.4512762427,0.4248676896,0.0157839339,-0.0111361668,-0.4655109346,0.1208941489,-0.4010038972,0.3006582558,0.1231233552,0.0035366071,0.3432372808,0.0029793673,0.0387485735,-0.1294725239,0.2690471411,0.0817266405,-0.0146125136,-0.0484843403,-0.4041103125,0.2343586981,-0.1110977903,0.142692402,-0.2079056054,0.1828098595,-0.2422186732,-0.1872489154,0.2078428715,0.1649779528,0.1381793618,0.1974901557,0.00771394,0.1722227633,0.1840904504,-0.1922398806,0.2213065922,0.2634327114,-0.2253491729,0.1050514877,0.11574696,0.0659233034,-0.2038090527,-0.3943380713,0.1596047878,0.3535887003,-0.2750309408,-0.172228083,-0.383197397,-0.708920002,0.0878656879,-0.0594570749,-0.3682579696,-0.550560534,0.1306333691,0.3110088706,-0.3960461915,0.2758713663,-0.5261030793,0.0930711403,-0.0833637863,0.377148807,0.029483242,-0.315015018,-0.1494927406,0.1138873547,0.3704943359,0.3368929923,0.3391319811,0.2285357863,0.0902660713,-0.5124884248,-0.3946644664,0.1821422726,-0.0811305791,0.3599928916,0.0334619135,-0.0304460786,-0.0033648976,-0.0145208053,0.1848327667,-0.0334529392,0.0808227807,0.0377494097,0.2265972346,-0.0568405129,0.0043872967,-0.2615486085,-0.1642236412,-0.5408262014,0.0232084766,0.1773352772,0.2984770536,0.2128200233,-0.110032782,0.0811095014,0.1700314879,-0.1284519434,0.08159215,0.2522673607,0.2351938635,-0.1482935697,-0.3752676547,0.1431304067,0.1678874493,-0.1806288958,-0.0289230738,-0.6907394528,0.2365356535,0.0004746777,0.0933537111,0.0852360874,-0.1197598875,0.2346420884,0.3941362798,-0.0285920184,0.0604873337,-0.1572939903,-0.1623924971,-0.0272981729,0.2487436384,0.0613057353,0.2927386463,-0.0788261741,0.4196695387,0.1608722657,0.1330552101,0.3300288022,-0.1292558014,0.3863185942,0.0880506858,-0.364426285,0.3801051974,0.2277207226,-0.2473003566,0.3891529143,-0.0468089357,-0.2272368819,-0.1322384924,0.111875169,-0.0174210239,-0.2235319912,0.0054794597,-0.3366178274,0.0723116398,0.1084828675,0.1603690684,-0.087550588,-0.4027015269,-0.0639728084,0.0354260877,0.1260525137,-0.397769779,-0.6519774199,-0.1440824121,-0.4036369324,0.3362410069,0.1335719079,0.2592405677,-0.1384121776,-0.0428604856,0.0075490596,0.3046002984,0.3338590562,-0.2926357687,0.0308585726,0.2173912674,0.2529658377,-0.3621837795,-0.1252756864,-0.1247658134,0.468627125,-0.0056123338,0.1088866964,0.1177692637,-0.1460490376,0.1244101599,0.0838265717,-0.3346751928,-0.0230384395,0.0858213305,-0.1470863372,-0.1134802401,-0.2595456243,-0.2903983891,-0.0807918385,-0.1117307097,-0.0160472374,0.0051927869,-0.171682179,0.1276297867,0.1367404908,0.4477116764,-0.0723016411,0.2226013392,0.2707365453,0.2623255849,0.3942559958,0.3254241347,0.1623783857,-0.1076942012,0.1056443229,-0.1841582358,0.3969781995,0.2862301171,0.0432810932,0.0279178899,-0.0078068012,0.3893857896,-0.1983547658,0.0260479581,0.3512432873,-0.0967423618,-0.1660416126,-0.6548130512,0.17357862,-0.1757688969,-0.2077556103,0.4948207438,0.0051999725,-0.1810278147,0.5770657659,0.0360416919,0.9330322742,0.2672248483,-0.1094423905,0.2783369422,-0.1210113466,0.612965703,0.0159467701,0.277343899,-0.3774955273,-0.3241224885,-0.0243557822,0.0032225286,0.0388527885,-0.1578301191,-0.1630984992,0.238726899,-0.233989805,-0.0200340506,-0.0329130441,0.3232554197,0.2424539328,0.1481516361,-0.1084030494,0.1145466119,-0.0608149692,-0.1897985637,-0.1467810422,0.0128434738,0.1084817946,0.0014939425,-0.3961191773,0.032109648,-0.1198762879,0.1515937299,-0.3794673681,-0.3329626322,0.2918035388,0.2802639008,-0.1389145106,0.1128718928,-0.156388551,0.1829782724,0.0113465497,0.181289658,0.0638161004,0.2905876935,0.2833430469,-0.0767332315,-0.2789282501,0.1410662979,0.0836604014,-0.0352023803,-0.2029493302,-0.185880959,0.1857665479,-0.277907759,0.1268288493,-0.1243294179,-0.3048667312,-0.1910587847,0.1262686253,0.2097568065,-0.0669730082,0.2531118393,0.1537329853,-0.0373004451,-0.1281880587,0.2323104441,0.0470716655,0.0388124958,0.5362433791,0.4200657308,-0.1564955562,-0.1853557229,0.1352237463,-0.1985569149,-0.1128123924,0.0336666778,-0.4092645347,0.0790371373,0.0843415558,0.1507214308,0.1443125755,0.1743270755,-0.4357252121,-0.2476724833,-0.0197396055,0.2915032208,-0.1539634466,0.1664569527,-0.4321200252,-0.141108498,0.1247308552,-0.1702901423,-0.3006685972,-0.190025717,0.0198921841,0.181043312,0.0269484352,0.1825227886,-0.2085029036,-0.4424799681,0.0355275087,-0.2638809681,-0.0845437199,-0.1637734473,-0.0852388814,0.1769664884,0.0740286261,-0.441613853,-0.3148105443,-0.2439830303,-0.0189493522,-0.063104406,-0.3076319993,-0.047821641,0.0404856764,0.0820248723,-0.0307593718,-0.2753333151,-0.0077422154,0.2412666827,0.1193006262,0.541461885,0.2563714385,0.2153195143,0.1639128923,-0.1250135452,-0.2580597401,0.1643506736,-0.3087143302,0.1652209312,0.3549098074,-0.3322285712,0.2556830645,0.3808780015,0.4932388961,0.4313769042,-0.1432711631,-0.2789634764,0.4424249232,0.1640637368,-0.4123229384,-0.1763328463,0.2795851529,0.2438185513,0.0173587725,0.1030137315,0.136197865,-0.1607579887,-0.2585865855,0.0680326298,0.4175924957,-0.203919515,0.3002223074,0.4241982698,-0.212717548,-0.0639845207,0.1778074652,0.2166756988,0.1435504109,0.6396567822,0.1417242885,0.0758812279,0.4841813743,0.1691903025,0.1241489053,-0.2833169997,-0.1818279028,0.2028390914,-0.099501811,0.0930603817,0.259293735,0.04682054,0.063155517,-0.3688143194,0.0459530018,0.2396122366,-0.0075236442,0.2120814323,-0.3550190926,-0.2509007454,-0.0494616255,-0.0261533819,-0.0940736085,-0.0751662403,-0.1726794243,-0.0033784211,-0.3168993294,-0.4275234044,0.3732615113,0.1240848452,0.2268958837,-0.2028679252,-0.0114721479,0.3269617856,0.1735690534,-0.0246689357,0.4546066821,0.2917703092,0.1306448132,-0.2347610295,0.0310389008,-0.0664144084,-0.1432176083,-0.1319899261,0.132255286,0.0302545149,0.1951593608,0.1808878481,0.076801993,-0.053962335,0.4734232128,0.2337242514,0.0584048405,-0.2925067544,0.3399984241,0.145849213,-0.5043171048,0.0774904564,-0.3538134694,-0.2001726329,0.1857924014,0.7511013746,0.031302385,0.1374648809,-0.0585534424,0.061800804,0.0923976228,0.3376110196,0.4668429494,0.0355729833,-0.1303791851,-0.2312413007,-0.8589652181,0.0914343521,-0.2171116024,-0.2172402143,0.5133656263,0.1280394047,-0.0502474234,0.2213602513,0.3961088061,-0.0102151269,0.0270923283,0.0515248701,-0.3876031041,-0.1328259557,-0.1321590841,0.3818224669,-0.3627648056,-0.1248889267,0.2687458694,-0.4400665462,-0.0286197234,-0.0540294871,0.0377760753,-0.0401768088,0.1479733884,0.1913590431,0.1345200837,0.1358108371,0.2102367878,-0.0262315515,-0.4110262394,-0.2733960152,-0.2440471798,0.2069380134,0.098176308,0.310220629,-0.3268450499,0.035486497,-0.2343169451,0.2626631558,-0.0487470105,0.192589432,-0.4621894062,-0.0339333825,-0.1455089748,0.2107549608,0.0786109194,0.1603516042,-0.0111095784,0.0877976194,-0.1460526586,-0.301722914,0.4168035388,-0.389662683,-0.1787187904,-0.0096131312,0.1982851475,0.3603422344,0.0374178886,-0.327151835,-0.1917618066,0.2938884199,-0.1424470246,0.0055394652,-0.0835715681,-0.1831873506,-0.0577196181,0.1430443674,-0.0101630688,0.1501576304,-0.2010464221,0.1263785064,0.0989648774]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/888","title":"Nested lists are zipped unexpectedly","comments":"Yes following the Tensorflow Datasets convention, objects with type `Sequence of a Dict` are actually stored as a `dictionary of lists`.\r\nSee the [documentation](https:\/\/huggingface.co\/docs\/datasets\/features.html?highlight=features) for more details","body":"I might misunderstand something, but I expect that if I define:\r\n```python\r\n\"top\": datasets.features.Sequence({\r\n \"middle\": datasets.features.Sequence({\r\n \"bottom\": datasets.Value(\"int32\")\r\n })\r\n})\r\n```\r\n\r\nAnd I then create an example:\r\n```python\r\nyield 1, {\r\n \"top\": [{\r\n \"middle\": [\r\n {\"bottom\": 1},\r\n {\"bottom\": 2}\r\n ]\r\n }]\r\n}\r\n```\r\n\r\nI then load my dataset:\r\n```python\r\ntrain = load_dataset(\"my dataset\")[\"train\"]\r\n```\r\n\r\nand expect to be able to access `data[0][\"top\"][0][\"middle\"][0]`.\r\n\r\nThat is not the case. Here is `data[0]` as JSON:\r\n\r\n```json\r\n{\"top\": {\"middle\": [{\"bottom\": [1, 2]}]}}\r\n```\r\n\r\nClearly different than the thing I inputted.\r\n```json\r\n{\"top\": [{\"middle\": [{\"bottom\": 1},{\"bottom\": 2}]}]}\r\n```","comment_length":27,"text":"Nested lists are zipped unexpectedly\nI might misunderstand something, but I expect that if I define:\r\n```python\r\n\"top\": datasets.features.Sequence({\r\n \"middle\": datasets.features.Sequence({\r\n \"bottom\": datasets.Value(\"int32\")\r\n })\r\n})\r\n```\r\n\r\nAnd I then create an example:\r\n```python\r\nyield 1, {\r\n \"top\": [{\r\n \"middle\": [\r\n {\"bottom\": 1},\r\n {\"bottom\": 2}\r\n ]\r\n }]\r\n}\r\n```\r\n\r\nI then load my dataset:\r\n```python\r\ntrain = load_dataset(\"my dataset\")[\"train\"]\r\n```\r\n\r\nand expect to be able to access `data[0][\"top\"][0][\"middle\"][0]`.\r\n\r\nThat is not the case. Here is `data[0]` as JSON:\r\n\r\n```json\r\n{\"top\": {\"middle\": [{\"bottom\": [1, 2]}]}}\r\n```\r\n\r\nClearly different than the thing I inputted.\r\n```json\r\n{\"top\": [{\"middle\": [{\"bottom\": 1},{\"bottom\": 2}]}]}\r\n```\nYes following the Tensorflow Datasets convention, objects with type `Sequence of a Dict` are actually stored as a `dictionary of lists`.\r\nSee the [documentation](https:\/\/huggingface.co\/docs\/datasets\/features.html?highlight=features) for more details","embeddings":[0.2298384905,-0.4676939547,-0.1091862842,0.3930577338,0.0112235183,-0.0073927417,0.2414543033,0.0436257534,0.1704822183,0.1277209669,-0.1480687261,0.4359344244,0.2361756712,0.2945679426,-0.0669526309,-0.0988031253,0.2758648992,-0.0747677162,-0.1930995435,-0.1146473587,-0.3136231005,0.2664879262,-0.3605991602,0.0551948212,0.0843064338,-0.0568167605,-0.1831727624,0.1451124549,0.1657565385,-0.0842644498,0.5213048458,0.0379811637,-0.0886196643,0.5056124926,-0.0001140757,0.2123831511,0.1992186159,-0.0585278496,-0.2496746927,-0.3481115401,0.0021727781,-0.4233009219,0.1910582185,-0.2739611268,-0.0578965172,-0.2668310702,-0.027284259,-0.4505228996,0.5443418026,0.0647106767,0.2045087218,0.2837791443,0.0236122441,0.0481437519,0.2327332348,0.0205130894,-0.1190875769,0.0153840426,0.1555838287,0.0709386021,0.1159702092,-0.0186715648,0.0423263907,0.0694481134,0.2614141405,0.4643959403,-0.2956536412,-0.4514957666,-0.3163237274,0.2530359328,0.0560796261,-0.1317227781,-0.385786891,-0.5079496503,-0.3157915473,-0.4794180989,0.2364197224,0.1448518634,-0.0973740146,0.2124937028,-0.1825155169,0.1744580567,-0.0692022443,0.3213108182,-0.1747741252,0.3411718011,-0.2386367768,0.1546947062,-0.1007574424,-0.1128670499,0.1495955139,-0.4561371207,0.2408811301,0.1273013055,-0.1851060539,-0.0407877788,0.0541589707,-0.3483132124,0.0675065592,0.104037106,-0.0081319045,0.1614092886,-0.3872076869,-0.0663330257,0.2979258597,0.2695173025,-0.0865092576,-0.0295973197,-0.1483498067,0.296833396,0.1175824329,0.0159885343,0.2882220745,0.0699860975,-0.1379981041,-0.0307000056,0.3500602841,0.1445537806,-0.336206913,0.101761207,-0.4738681316,-0.0749661848,-0.0336969607,0.0638875291,-0.3206653297,0.1525014639,-0.0071569891,0.2471447438,-0.134557724,0.3446595669,-0.2371739298,-0.1904279739,-0.0652847886,0.0663499311,-0.0420984179,-0.0953788757,0.165487349,0.2125787437,0.1812343299,-0.267642647,0.039244622,0.0152299171,0.3261408508,0.1474461704,-0.1008216143,0.3499751091,0.2230881602,-0.181228146,-0.1014808938,0.0566560216,-0.5195723176,-0.0518508218,0.1581280828,0.133820042,-0.4378278255,0.0968860388,-0.0804307014,0.1471621841,0.3309530914,0.3669150174,0.0343742371,0.0342494622,-0.508404851,-0.2293902636,0.0003260465,-0.073894307,-0.2071796358,-0.0262299776,-0.101927571,0.0716583878,0.2804749012,0.3513951302,-0.1295993477,0.2629665136,-0.3151412308,0.2650033236,0.1025323644,0.0171958786,-0.3716728389,0.216081202,-0.1512710452,0.5964806676,0.0167001151,0.0097154761,0.0972158089,-0.1860406846,0.4022799134,0.1987628192,0.3353220224,-0.2042614371,-0.1204482689,-0.0844999179,0.4401874244,-0.2315003872,-0.5066453815,0.0658121705,-0.2022055387,-0.0535663217,0.2737059295,-0.0846181214,-0.022978913,0.426255703,-0.050876908,0.1780268103,0.0066015054,0.057343334,-0.5211535692,0.1728512347,-0.1369978935,-0.1579048485,-0.3792924881,-0.1082615778,-0.1152307987,0.3877604902,-0.1813657284,-0.0864765346,0.1596705765,0.1504893452,-0.0718193054,-0.126271829,-0.1037976593,0.1233520582,0.0714035407,0.0906172246,-0.6953419447,0.3920565248,0.2789391875,0.0548487045,-0.2176354676,0.2317307889,-0.0456582345,-0.2940310836,0.0523777828,0.2249441743,0.23389557,-0.1588659734,-0.4249760807,0.0298046209,0.3272547722,-0.0199340992,-0.1682194918,0.4148456156,-0.0735301897,-0.0860977098,-0.4220732152,0.4835673273,-0.0940231383,0.4029453993,-0.1599096209,-0.0625059307,0.0480617508,0.1892828047,-0.203675285,-0.2096896619,-0.0529219136,0.2995162308,0.1877167523,0.2752779722,-0.243581444,0.4470909238,0.7448371053,0.0350882672,-0.0720682368,0.2781617045,-0.0195389204,-0.24792023,-0.0599321611,0.0952594429,0.3652765453,0.0646614954,0.2216386199,0.115891628,-0.1229742169,-0.2536591589,-0.0077809398,0.2639900446,0.1566016376,-0.0100727268,0.2818867564,-0.0750234798,-0.25094679,0.1176291183,-0.0121422997,0.0026538027,-0.1509614885,0.3371892571,-0.3686792254,0.0113473218,0.0587112792,-0.5471538901,-0.094660528,-0.314432025,-0.1246987954,0.2293175161,-0.0363037214,-0.0210861433,0.5416872501,0.0074516344,0.2658097446,0.0552181304,-0.0712552443,0.0766024962,-0.2392053306,0.0723662525,-0.0737971365,-0.2192000002,-0.0197297335,0.2536250353,-0.2125402242,-0.2188146561,-0.20200409,0.0020514855,-0.4176460207,0.2868644893,0.124497503,0.3577858508,0.0863532647,-0.1363054365,0.0286891311,0.7422843575,-0.2904424369,0.0333607905,0.1572047919,-0.0130648529,-0.2704990804,-0.3338178396,-0.2083853483,-0.0691349134,0.6958264112,0.2305352539,0.1344957203,0.1984081417,-0.0606457628,-0.2014248818,0.3159995675,-0.1195729449,-0.2812665403,-0.3180609345,0.290892154,-0.150292635,-0.4329236448,-0.1716616005,-0.2320409268,0.1632248759,-0.0522457212,-0.2677069902,-0.0519596711,-0.3353410065,0.2634570897,0.1147342622,-0.0183236897,0.3230091929,0.0054677785,0.0031281051,-0.0969485864,-0.0848343,0.2513024211,0.0311405659,0.1827379018,0.1495402455,0.2881786227,0.0694663823,0.3716077805,0.4487586319,-0.0575874783,0.0708739385,-0.2050477564,0.0590154119,-0.3577525318,-0.0285773091,-0.3248988688,-0.0913258642,-0.1134553924,0.1470746398,-0.0445475578,-0.1424783915,0.2277165204,0.2406812459,0.0588044971,-0.1793935001,0.0593523122,-0.138631478,-0.0676757172,0.125119254,0.1005209386,-0.2591460049,-0.2106327415,-0.1890942454,-0.2704479694,0.3228024542,-0.1335511208,-0.4832397103,-0.3275099099,-0.2935387492,0.0865744948,0.0658592507,0.3359498084,0.1210119873,-0.3332602382,0.1720486879,0.0042104875,0.6275284886,0.1835338622,-0.1038911641,0.2392939776,-0.3792848587,-0.5593377948,-0.2233467698,-0.3037458062,0.0582429953,-0.073396869,0.7011970878,-0.4322921038,-0.2735138535,-0.1542567015,0.0225943811,-0.1823845953,-0.110319525,-0.1726929545,-0.0383672528,-0.0809846073,-0.3157395422,0.1211033612,0.0220624804,-0.2013719082,-0.5030103326,-0.0302700642,-0.1497213989,0.0327285901,-0.1705443859,0.2045018673,0.1852382272,-0.0533424877,0.1829633713,0.3905874789,-0.3235560954,0.5041468143,0.1633858681,-0.4202218354,-0.0567698888,-0.052258417,-0.0387408175,0.1707356572,-0.2789490223,0.3413133919,-0.2047440112,0.1857427955,-0.4233269989,0.026189791,0.5656720996,0.1640046835,-0.1857060045,-0.4488417804,0.4633074701,-0.0362390652,-0.2361240685,0.2130220234,-0.4771503806,-0.2249504328,0.0851177573,-0.0977168083,0.6535152197,-0.0141142858,0.3276809752,0.7303436995,0.2532294095,0.651691854,-0.2009021491,0.1017264277,-0.037872009,-0.2553576231,-0.2321520746,-0.0990923047,-0.2437690645,0.0722096711,-0.167147696,0.3249192834,-0.2017386407,-0.09809044,-0.1091667861,-0.2246543318,0.3676775694,-0.1916199178,-0.2470280081,0.0923593566,0.1439547986,0.2766925693,-0.0290464498,-0.1788530499,-0.3586546779,-0.0818017125,0.0169012118,0.0519188866,-0.0419865996,0.2930719852,0.5869737267,-0.0869002342,-0.1537225991,0.3690270185,0.1423909068,-0.1004729643,-0.0946219563,-0.1521724463,-0.2162119299,0.1140700653,0.1403345913,-0.1992438585,0.18698892,0.1408914328,-0.067196086,-0.0583789013,-0.1235271096,-0.0281103589,-0.1539426297,0.3628600538,-0.1732292175,-0.3580111265,-0.1335516274,0.1337673962,-0.0743304342,0.0093414271,0.1406382173,-0.072063677,-0.1272349954,0.1793529689,-0.1711459607,-0.384334594,-0.1006320268,0.4531825185,0.3465633988,0.2332885861,0.5220939517,0.2117365748,-0.0131213116,-0.0413363129,0.1071981117,0.24509345,0.0429080129,0.2102357447,-0.2296631783,0.0036050405,-0.1185464039,0.4313834608,0.1401435584,-0.00680803,-0.3198307157,-0.2256010175,-0.1199583709,0.1146233603,0.1380531639,0.0869319886,-0.3789420128,0.1209981367,0.0172732845,-0.2959543765,-0.2442490608,0.0440332741,-0.1746987402,0.0734402016,0.1650625765,-0.0672527179,0.2527103722,0.0798484907,0.1422857195,0.1194660366,-0.2147552967,-0.2483876795,-0.0779388845,0.0247311052,-0.1127661094,-0.1158621311,0.0402208604,0.0732932314,0.0024652905,-0.222398892,0.062097773,0.0523642451,0.1028145924,0.0934442505,0.118789345,0.1133369654,-0.0887405276,0.0219954699,-0.1139061898,0.3333363533,-0.0880424306,-0.0082744835,-0.3272545636,-0.0375976674,-0.328681469,-0.0047411034,-0.0873387754,0.0518627837,0.3860618472,-0.1540907174,-0.1757969111,-0.147393465,0.3297917843,0.4187699258,-0.1839772463,-0.1148336604,-0.3160582781,0.1524474323,-0.0430010222,-0.1944517195,0.204142049,0.4142587483,0.0478667244,0.3599616289,0.0795380995,0.1255003363,0.0107442038,-0.2078155428,0.0908526853,-0.0719288066,0.3282035887,0.294871062,-0.0125399651,0.1877646148,0.1393261999,-0.107622005,0.1366218179,0.5006206036,0.4455509782,0.1572834104,0.4649832845,0.2576405406,0.162031889,-0.3960626125,-0.0934355333,0.2119640112,-0.0068730377,0.1135138348,0.2420565188,0.1350420862,0.0317590833,-0.2465137392,-0.0710508972,0.137797758,0.0352619998,-0.3907749653,0.0854904354,-0.0234658942,-0.0725790933,-0.190702036,-0.0247699711,-0.3372789621,0.2577609718,-0.1021224707,0.1690778285,-0.1788423061,0.0724069253,-0.0390333794,0.1608751118,-0.011252597,0.3989444971,0.3208033741,-0.1631928682,0.0096530924,0.0057606259,0.1689160019,0.0739874542,-0.0025229785,0.1851235032,0.0849192142,-0.1691342145,-0.0921691433,0.1738840491,-0.2618695498,-0.1227688491,0.3215523958,0.1802615076,-0.0950470045,0.3788063824,0.1822907478,0.3223470747,-0.4141272604,0.3234879375,-0.3690626025,0.069021076,-0.1374792755,-0.214028582,-0.4857179821,-0.1849492937,0.4549840093,0.0622962452,0.38710168,0.0771670565,0.1050898284,-0.1980644166,0.56604141,0.1562877446,0.0848102048,0.2197218388,-0.0166593958,-0.1285703182,0.1937224418,0.0282121003,0.0746909827,0.0411252789,0.2085569054,-0.0126786288,0.1919088811,-0.2487942278,-0.3145215213,-0.2395996004,0.1238393858,-0.6030551195,-0.2136266977,0.0546103567,0.1322104633,0.0227826722,-0.1303688139,0.3184503615,-0.0531227365,-0.0405780524,0.0055196844,0.2425608635,-0.3423747718,0.1814104021,0.0602163225,0.3718806505,-0.1478212476,-0.0252268091,-0.1264715791,-0.2578124702,-0.1313626766,0.0753141344,0.5303570628,-0.1008528396,0.3721912503,-0.1619537026,0.2664503753,-0.0218831059,-0.1117807776,0.2469821721,-0.4401483834,-0.3289555609,0.2583472133,-0.2655967772,0.1213144213,0.1467083544,0.6770501137,-0.0092603583,0.1449607164,0.229273811,-0.1296216697,0.4818060994,-0.2091841996,-0.4057359099,0.1518884897,-0.1456567496,0.1526296884,-0.2212257087,-0.0083837677,-0.090432331,0.4702816606,-0.3118601441,-0.3027620912,0.0298772044,-0.0429780371,-0.0823716819,-0.2434902787,0.3457022309,0.1218633205,-0.2046686411,0.3793985546,-0.2259826958]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/888","title":"Nested lists are zipped unexpectedly","comments":"Thanks.\r\nThis is a bit (very) confusing, but I guess if its intended, I'll just work with it as if its how my data was originally structured :) \r\n","body":"I might misunderstand something, but I expect that if I define:\r\n```python\r\n\"top\": datasets.features.Sequence({\r\n \"middle\": datasets.features.Sequence({\r\n \"bottom\": datasets.Value(\"int32\")\r\n })\r\n})\r\n```\r\n\r\nAnd I then create an example:\r\n```python\r\nyield 1, {\r\n \"top\": [{\r\n \"middle\": [\r\n {\"bottom\": 1},\r\n {\"bottom\": 2}\r\n ]\r\n }]\r\n}\r\n```\r\n\r\nI then load my dataset:\r\n```python\r\ntrain = load_dataset(\"my dataset\")[\"train\"]\r\n```\r\n\r\nand expect to be able to access `data[0][\"top\"][0][\"middle\"][0]`.\r\n\r\nThat is not the case. Here is `data[0]` as JSON:\r\n\r\n```json\r\n{\"top\": {\"middle\": [{\"bottom\": [1, 2]}]}}\r\n```\r\n\r\nClearly different than the thing I inputted.\r\n```json\r\n{\"top\": [{\"middle\": [{\"bottom\": 1},{\"bottom\": 2}]}]}\r\n```","comment_length":28,"text":"Nested lists are zipped unexpectedly\nI might misunderstand something, but I expect that if I define:\r\n```python\r\n\"top\": datasets.features.Sequence({\r\n \"middle\": datasets.features.Sequence({\r\n \"bottom\": datasets.Value(\"int32\")\r\n })\r\n})\r\n```\r\n\r\nAnd I then create an example:\r\n```python\r\nyield 1, {\r\n \"top\": [{\r\n \"middle\": [\r\n {\"bottom\": 1},\r\n {\"bottom\": 2}\r\n ]\r\n }]\r\n}\r\n```\r\n\r\nI then load my dataset:\r\n```python\r\ntrain = load_dataset(\"my dataset\")[\"train\"]\r\n```\r\n\r\nand expect to be able to access `data[0][\"top\"][0][\"middle\"][0]`.\r\n\r\nThat is not the case. Here is `data[0]` as JSON:\r\n\r\n```json\r\n{\"top\": {\"middle\": [{\"bottom\": [1, 2]}]}}\r\n```\r\n\r\nClearly different than the thing I inputted.\r\n```json\r\n{\"top\": [{\"middle\": [{\"bottom\": 1},{\"bottom\": 2}]}]}\r\n```\nThanks.\r\nThis is a bit (very) confusing, but I guess if its intended, I'll just work with it as if its how my data was originally structured :) \r\n","embeddings":[0.2184679955,-0.3172184825,-0.1545520872,0.4096399546,0.0198083632,0.0022643409,0.1813217551,0.1284821928,0.12042135,0.1107622907,-0.0894270167,0.5233137608,0.330974251,0.2051908076,-0.109711796,-0.0653213114,0.3139220476,-0.0546331406,-0.0289824065,-0.0713038146,-0.3160270452,0.2633286417,-0.3739173114,0.0464868322,0.1322630197,-0.052191142,-0.257597208,0.3128820956,0.1092591286,-0.0722756535,0.4127613306,-0.0793377608,-0.1082954109,0.4087022543,-0.0001130303,0.1514425874,0.2343340516,-0.1096897945,-0.2292565107,-0.2896882892,0.1124610975,-0.4846199155,0.1144394651,-0.254737854,-0.0396972746,-0.2845827341,-0.1264148057,-0.6383929849,0.4745340943,-0.1257840693,0.2245387584,0.2374026626,0.066508241,0.0422530994,0.2652522027,-0.0834824145,-0.0373349898,0.1073116064,0.2651552558,-0.0105168568,0.1435327828,-0.0846476629,-0.1027298719,0.0434745848,0.142403692,0.3884346187,-0.3656547666,-0.4055756927,-0.31646505,0.1316068769,0.1293771714,-0.080391638,-0.3327471316,-0.4394020736,-0.3403945863,-0.4365845323,0.1443232149,0.2944245338,-0.0351609923,0.2033887357,-0.1429800838,0.1454972625,0.026514465,0.3943552077,-0.0928072035,0.5669482946,-0.1781391799,0.2075331062,-0.1415402293,-0.0978918895,0.3217330575,-0.5392986536,0.1195542812,0.0827537253,-0.0811452046,0.0289473999,-0.0254871789,-0.2950782776,0.0209243447,0.2430077046,0.0097585898,0.0969810039,-0.2464405894,-0.0963637009,0.3439159393,0.230506897,-0.1567772627,0.0480112173,-0.1993891001,0.2459215224,0.0642057806,-0.0743023232,0.3313893378,-0.0181406587,0.0041822796,0.0940826386,0.3247089982,-0.006613283,-0.3000763655,-0.0172231663,-0.6085450053,-0.1270057708,-0.0157090407,-0.021211097,-0.218605265,0.2345955968,-0.066056788,0.3846606612,-0.1284175813,0.2158971429,-0.1964947134,-0.1969259232,-0.1403606832,0.0498289317,-0.0228432324,-0.0731570721,0.1988915205,0.3086185753,0.0985997394,-0.3232749999,0.0642548949,0.0588632487,0.299834162,0.0770312697,0.04181787,0.3301210701,0.2961995006,-0.1713803411,-0.0828469694,0.1917276978,-0.6464284658,-0.0623244494,0.1348978579,0.153404057,-0.4481498897,0.0445792265,0.0383603871,-0.0374039449,0.3544823825,0.2428795695,0.0145594385,0.0994814262,-0.3080238998,-0.1964315772,-0.1299649328,-0.061015334,-0.2842946351,0.1140612587,-0.2470975369,0.1278833151,0.4151358306,0.1677791476,-0.1544118971,0.2156068981,-0.3377684653,0.2923148572,0.1259207875,0.134586677,-0.4144970775,0.3028047085,-0.1807539761,0.5522437096,0.0382716246,-0.0058981394,-0.1193735525,-0.1312676668,0.4894069135,0.1428713351,0.3969982266,-0.1976763159,-0.2331183404,-0.0353870355,0.4713866413,-0.1843084842,-0.5262537599,0.1478794217,-0.1520498842,-0.1689659506,0.3451805711,-0.127233699,0.0350561962,0.4289444685,0.0043613054,-0.0003774321,-0.0241305437,0.0816469565,-0.3355894983,0.0989226699,-0.2258457243,-0.2860853672,-0.4463803172,-0.0183767788,-0.1477211714,0.2564418614,-0.1700654775,-0.0308613479,0.1736151874,0.2090605497,-0.1163886189,-0.3199609518,-0.0105879959,0.0195043832,0.3265604079,0.0976136401,-0.6830198765,0.4054053724,0.287594229,-0.0113849714,-0.1458497047,0.0695441589,0.1524824053,-0.2553286552,0.0180172846,0.3207508028,0.1593113393,-0.1835816354,-0.3426262736,-0.1545603424,0.152085945,-0.0245116409,0.0242283363,0.5047954321,-0.0688403696,-0.037720222,-0.2710852027,0.5477904677,-0.2554773986,0.4314802885,-0.2103954256,-0.0547998399,-0.0019476638,0.1101335883,-0.1451644599,-0.1782767326,-0.0519486926,0.3697811365,0.1320610195,0.2217051834,-0.127253592,0.5368802547,0.6226235628,0.0149432728,-0.1085233986,0.2366561741,-0.0172710251,-0.2339497209,-0.0827031359,0.1318300515,0.4769052267,0.1663245559,0.2383282781,0.0633735284,-0.0466573946,-0.2988387048,0.0415332839,0.294847697,0.1417426616,0.0189413652,0.2471589893,-0.1103716344,-0.1890186518,0.1310887337,-0.0676205233,-0.0555508547,-0.1847279221,0.4284857213,-0.4148808122,0.1286701411,0.008959488,-0.6086062193,-0.1259289086,-0.367431581,-0.0962766632,0.1630174369,-0.0332720689,-0.0793730319,0.3183373809,-0.050881777,0.2074662745,0.0344276726,-0.0979534984,0.0549232662,-0.3040453494,0.1165449396,-0.0992763788,-0.1788055301,-0.1328109205,0.2860119045,-0.0628293157,-0.0983806923,-0.1677879244,-0.0613331199,-0.3298401833,0.4358928502,-0.0053036991,0.2740277946,0.1420147717,-0.16843234,-0.0573893785,0.8114554286,-0.2242003679,-0.0530012213,0.1051107273,-0.1221103892,-0.2619016171,-0.2897172868,-0.2417907417,-0.0876349509,0.5448132753,0.2349026799,0.0652129278,0.0126600154,-0.0915844589,-0.4041164815,0.4660033584,-0.2062158734,-0.2937187254,-0.362541914,0.3837715983,-0.1704190671,-0.307312429,-0.141105026,-0.2317969948,0.2073756009,-0.0551498756,-0.2060276866,0.0694100633,-0.2908574641,0.177852422,0.0170119833,0.0154639743,0.252792716,0.1166265607,-0.0228671432,-0.1234941036,-0.0833317712,0.1336052865,-0.1174999177,0.3187896907,-0.0142855374,0.1677422523,0.0150169525,0.2484380007,0.3785092831,0.1109360531,0.175862506,-0.1604481488,0.041153077,-0.2754459083,-0.0336143374,-0.4888683558,-0.0638975352,-0.1470775306,0.1049498841,-0.0306321029,-0.2773481309,0.2061710656,0.0542604551,-0.0645456389,-0.1242480576,0.0656335577,-0.1353945434,-0.0298898369,0.1548280567,0.0780918598,-0.2866789699,-0.280462116,-0.2260524333,-0.3523418307,0.3581950068,-0.1251737624,-0.5212160945,-0.3446735442,-0.0681027323,0.2324893028,0.0155182155,0.2490625829,0.1768508255,-0.377173692,0.0962257534,-0.0448231623,0.5204252005,0.2217597961,-0.0332037956,0.32608217,-0.1375302374,-0.4786706865,-0.2850190699,-0.3057845831,0.2125431895,-0.1146560386,0.6859157681,-0.4032107294,-0.1047900692,-0.0662889332,0.1085298881,-0.1328310817,-0.2047206461,-0.206184119,-0.010784314,-0.0190015212,-0.2465892583,0.0428602844,0.1672843546,-0.3029408455,-0.4507920146,0.0202640072,-0.1985384971,0.0193104707,-0.2084351629,0.2844068706,0.1069713831,-0.0047229584,0.1485110074,0.4633721113,-0.3736360371,0.6574385762,0.2314502299,-0.2134834379,-0.0199563056,-0.0827863961,0.0931335464,0.1565194428,-0.2462546229,0.3424846828,-0.1902332604,0.0940129682,-0.2823220491,-0.1878391057,0.4364049137,0.024759315,-0.1664128006,-0.4040624201,0.4862018824,-0.152612716,-0.2920736969,0.1725669056,-0.7546376586,-0.1507408917,0.0666646212,-0.0745668337,0.7946743965,0.1333986968,0.1628906131,0.5988457203,0.1569936872,0.3440333307,-0.2301672548,0.1290615648,-0.1200245246,-0.1831686348,-0.1518256068,-0.1016812399,-0.2299713045,0.2175675631,-0.1713269055,0.3684467077,-0.3066546321,-0.2079588175,-0.0820804536,-0.2884979844,0.2763154209,-0.139738068,-0.36778,0.0831405893,0.0889605135,0.1820562184,0.0162746385,-0.2232854366,-0.4019068182,-0.1828648001,0.1694707125,0.188186124,0.0737712681,0.238206476,0.5669982433,-0.090839237,-0.2376575917,0.2308873534,0.0378217623,0.0083171576,-0.0510930717,-0.0868951827,-0.0498087443,0.1095718294,0.1393981278,-0.0793702975,0.001978857,0.1429879069,-0.1889206618,-0.0380919091,-0.1322387159,-0.0750016794,-0.2670205832,0.303619355,-0.0664990842,-0.4930697083,-0.0863872617,0.162557289,0.0961073935,-0.0442151502,0.1409608424,-0.0584478118,-0.0600010939,0.2154303789,-0.2095663846,-0.3898904622,-0.1302578002,0.4818537235,0.35420838,0.3314028978,0.5364958048,0.1173828095,-0.0052481727,-0.0615891255,0.1774370521,0.1731762141,0.0555573925,0.2308419645,-0.2614545524,0.1389486194,0.0255602319,0.2495402098,0.1538052559,0.0933203101,-0.2747829854,-0.2578544617,-0.0123965982,0.0752705038,0.2469940484,0.0264129154,-0.3694917858,0.0166562293,-0.0655556545,-0.2943484485,-0.2588206232,0.1187275201,-0.073504068,0.0706627518,0.0632948279,-0.0498926751,0.0983068421,0.1755087078,0.1788015366,0.0973351821,-0.0939834565,-0.2423440367,-0.0893190876,0.0399382599,-0.099445954,-0.0512975678,0.0312251598,0.0619443171,0.1177683398,-0.1408179104,-0.0800768435,-0.0501600467,0.1151590869,0.0494760796,0.1752440035,0.0092204632,-0.1670494974,0.0650653765,-0.1628144234,0.35219419,-0.0734155029,0.022996353,-0.204765141,-0.0357837826,-0.3333109021,-0.1181888878,-0.0601646788,0.2736842632,0.442330569,-0.1754452437,-0.0991264805,-0.1710648984,0.3002393842,0.4621215761,-0.1071062088,-0.058222279,-0.4228375852,0.1897702068,-0.0301655252,-0.2206501514,0.0477977693,0.4067502022,-0.0115044881,0.3411348164,0.0762362182,0.004459037,0.187585026,-0.1609860808,0.2129374295,0.0691064745,0.2861249745,0.445766151,-0.0188227836,0.1424773782,0.1934843659,-0.1898761243,0.07587412,0.403838098,0.4355384707,0.1002639234,0.5404435992,0.1917512268,-0.050772842,-0.386857301,-0.1148916557,0.1924684644,0.1613169014,0.1440419108,0.3425574601,0.1328895241,-0.023340635,-0.1878120452,-0.1028043553,0.163994953,0.1770245433,-0.4165566266,0.0156628639,-0.1586011648,-0.1736982912,-0.0960783064,-0.0284376964,-0.2167844921,0.2008898407,-0.197361365,0.1110691279,-0.2086458504,0.1426243931,0.0291805863,0.1135149226,-0.1219931692,0.3748128712,0.4225357175,-0.1691365391,0.1362073869,-0.0214618631,0.1988118589,0.1122122407,-0.0623035394,0.0575759262,0.1041241512,-0.2180199474,-0.0522047952,0.1724111587,-0.2057880312,-0.1832584739,0.2945329249,0.2070864886,-0.1300415099,0.5459814668,0.2538106143,0.3727188408,-0.328188628,0.2841492295,-0.2704302073,-0.030711377,-0.1845007837,-0.356141746,-0.4568126798,-0.2268985361,0.348900646,0.1265204251,0.3013012111,0.0669780225,0.1108610779,-0.1432696581,0.5278849602,0.1273645461,-0.0721038803,0.1525624096,-0.0574445166,-0.0945012048,0.3024398386,-0.1291500926,0.0792855099,-0.0166698117,0.0780356079,0.0833573192,0.1977619827,-0.2263824046,-0.2716867924,-0.3187078834,0.2024095803,-0.6554205418,-0.24554573,0.3589972258,0.2067427933,0.0032625813,-0.1457839608,0.2539188862,-0.029918747,0.0071007395,-0.018424565,0.044192262,-0.2437457889,0.1884052157,0.157117486,0.4281648099,-0.0677081123,-0.0735549852,-0.1380256861,-0.2773789167,-0.195383057,0.1449756622,0.5468021035,0.0324326344,0.3450416923,-0.1000140384,0.3605937958,-0.1093713343,-0.1459473819,0.1021506488,-0.2745310366,-0.2356798351,0.2147416621,-0.2724907398,0.0901660845,0.0718507767,0.6004976034,0.0255684834,0.2015389353,0.2285086215,-0.0928081796,0.3809424937,-0.1058680415,-0.4406690001,0.2755937874,-0.13677001,0.1099884585,-0.2000233978,0.2015327811,-0.0702717379,0.57451123,-0.2451062649,-0.1149065495,0.0288585294,-0.0563111342,-0.1113771945,-0.1828141958,0.2170536965,0.1465842873,-0.1766922325,0.2814049423,-0.1616722345]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Yes right now `ArrayXD` can only be used as a column feature type, not a subtype.\r\nWith the current Arrow limitations I don't think we'll be able to make it work as a subtype, however it should be possible to allow dimensions of dynamic sizes (`Array3D(shape=(None, 137, 2), dtype=\"float32\")` for example since the [underlying arrow type](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L236) allows dynamic sizes.\r\n\r\nFor now I'd suggest the use of nested `Sequence` types. Once we have the dynamic sizes you can update the dataset.\r\nWhat do you think ?","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":85,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nYes right now `ArrayXD` can only be used as a column feature type, not a subtype.\r\nWith the current Arrow limitations I don't think we'll be able to make it work as a subtype, however it should be possible to allow dimensions of dynamic sizes (`Array3D(shape=(None, 137, 2), dtype=\"float32\")` for example since the [underlying arrow type](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/features.py#L236) allows dynamic sizes.\r\n\r\nFor now I'd suggest the use of nested `Sequence` types. Once we have the dynamic sizes you can update the dataset.\r\nWhat do you think ?","embeddings":[-0.1616059095,0.0108419182,-0.066619046,0.1772314459,0.3133355081,0.0057711992,0.5568307638,0.1153363064,-0.1397421658,0.0790204331,0.1618123502,0.0611892194,-0.2482876331,0.2882796824,0.1433374882,-0.4168105423,0.1650112569,0.257226795,0.0553811863,0.1130094305,-0.1897820085,-0.2030609548,-0.2229629159,0.1924257725,-0.2032823563,-0.1325577945,-0.231277436,-0.1746644974,-0.2154562771,-0.4873974323,0.3294047117,0.1431981474,0.2426173389,0.1376820207,-0.0001154924,-0.0107514635,0.3926018178,0.0132955424,-0.1953313798,0.0277076904,0.0773622021,-0.4880388379,0.1142667159,-0.3838246763,0.1063902453,-0.9378988147,-0.0416392684,-0.089719303,0.1050876603,0.0500929318,0.1922921687,0.1715771407,0.3899645805,-0.0105450368,0.4836086333,0.3242554367,-0.2707018256,-0.0386977233,0.3831052482,0.1397040933,0.0771353468,-0.1748768538,-0.0284891389,0.0574248694,0.4720866382,0.1713972539,0.2727361321,-0.1684577018,-0.1026954353,0.4105275571,0.4952409863,-0.2319118083,-0.3075928688,-0.3115698397,0.0032654202,-0.1753804535,-0.107309036,0.213774845,-0.1296340227,-0.0140746748,-0.153499797,0.1548554152,-0.4418312311,0.3143471777,0.107572265,0.1329267472,0.0735217705,0.2417716831,-0.2476113737,-0.5056387782,0.4316933155,-0.2360915989,-0.0270423125,0.1407017559,0.007250621,-0.0080920141,-0.1698652208,-0.3400796354,0.4770568311,-0.1000495404,0.3479025066,0.0111999288,0.1662002504,0.0931350738,0.0409678333,0.3405994475,-0.1212562546,-0.274788022,-0.1511282027,0.1171398014,0.1713352799,0.0727716759,-0.2241659313,-0.1283685118,0.2919893861,-0.1056135818,0.2182835788,0.3313554227,0.0621668063,-0.0202995911,-0.5062046051,-0.1305086911,-0.0367953256,0.1205356345,0.0674209297,0.134990871,-0.0809327886,0.3865687251,0.1314042807,-0.005581412,-0.096765779,-0.1221151724,-0.1639343798,-0.0432264544,-0.0311276615,0.266010344,-0.0362326242,-0.0843537003,0.1206867918,0.151942417,-0.0673132837,0.1036842763,0.1126583666,0.0875582546,-0.1935007125,-0.2413788289,-0.2568321824,0.0092555583,-0.2337166965,0.3463324308,0.0196913555,-0.2524262071,-0.4545671642,0.1253437996,-0.2485116571,-0.0374918729,-0.2303138226,0.2204686403,0.2725118697,0.0192481298,0.1092601791,-0.0748157352,0.0403708667,-0.4584237337,-0.0161776636,-0.0629338846,-0.279412061,0.1130259931,0.0052774181,-0.0103954813,-0.1350494027,-0.0232585538,-0.1822223067,0.0009784076,-0.0583328903,0.199472785,0.4691610038,0.2001069784,-0.044287011,0.0683372617,0.0993991643,-0.184038803,0.3197771907,0.0600170381,-0.0458100401,-0.1721129268,-0.2826547027,0.3145986497,-0.0964618996,-0.169590503,-0.0800119117,-0.5706949234,0.3061468303,0.1484428793,0.0984555408,-0.0734474957,0.0419839844,-0.4109460711,0.2504271865,-0.4095450342,0.2487358302,-0.0755418465,0.2601501048,-0.0907999352,0.1170577183,-0.278133601,-0.8328418136,0.2076082975,-0.3454922438,-0.1067001,0.0295050927,-0.1889729053,-0.1704300195,0.1456988603,-0.1304069459,0.1564953774,0.0873014778,0.003468764,0.056150198,-0.2968775928,-0.2516316772,-0.3527005613,-0.0608819425,-0.0207408238,-0.4056304097,0.4219984412,0.1468448043,-0.2653646767,-0.2945557535,0.2871393263,-0.0391316265,-0.0812378004,-0.0405562706,0.2517087162,-0.0943240002,0.0002011115,-0.5369089246,0.1763547063,0.0346698388,-0.1743862033,0.3629363775,0.0060334634,0.0265163053,-0.1310422868,0.2302036434,0.4568025768,-0.00220561,0.4285460413,-0.1744944006,-0.0404056124,-0.0721176341,0.1969555914,-0.0736600012,-0.1658544093,-0.2773365378,0.2361840755,0.1594388634,-0.0260374323,-0.202352047,0.1595479995,0.2601911128,-0.1021904275,0.0460606888,0.1400887966,-0.2551288307,0.0729365945,0.1663903445,0.1524063498,0.1684515923,0.2051693499,-0.1223241985,-0.0231505018,-0.1620201916,-0.0988160968,0.2174188793,0.1760276556,0.4924125671,0.0949758366,0.4513032734,-0.1852277368,-0.1904083043,-0.0040292214,0.0273078512,0.0714368448,-0.1541869342,0.0563471504,-0.095658578,-0.0615262166,-0.0848443359,-0.6617180109,-0.0479578748,0.0101695135,0.0034061878,0.0023362103,-0.1764207333,0.2182947099,0.0707682073,-0.1285052747,0.2033283412,-0.1506586671,-0.1665368378,0.0061287363,-0.0245495308,0.063650392,0.2478206456,-0.0018959439,0.0085574938,0.2448384613,0.0823839381,-0.331317693,-0.31513533,-0.008297408,-0.1528569609,0.232638672,0.431912303,-0.0495447852,0.1705277413,-0.3125085533,0.3104337454,0.0129286656,-0.2742746472,0.1388646811,-0.2198174149,0.1748547554,-0.1059625894,0.007749442,-0.1792992055,-0.3620901704,0.3248399198,0.4152484536,0.1981838644,0.2461229265,0.3520291448,0.171993956,-0.0005932474,0.0170385297,0.1446620226,0.2558503151,0.3828643262,-0.0266483743,-0.1819356531,-0.1355111152,-0.9175083041,0.0974397585,0.3725898564,0.0600478686,-0.1703146398,-0.1563492268,0.2750999928,0.0853943676,-0.0136378491,0.4456554055,0.1230401099,0.0161780324,-0.0962931663,-0.0493287668,0.3227924705,0.2541518509,-0.0290216487,0.2366008013,0.3795795441,-0.116150856,0.6229423285,0.0335682631,-0.4343720973,0.3152050376,-0.2801405191,-0.1684267521,-0.1951777041,-0.0541303828,-0.1407594681,-0.030709805,-0.0905178413,-0.0600013062,-0.1186711714,-0.3501573801,0.1830917299,0.1964474916,0.0443265624,0.0143344728,0.1271090657,0.0562775247,0.3102790117,-0.1323706806,0.0103131486,-0.392272383,-0.234360531,-0.2627184093,-0.2672703862,0.1622469127,-0.0061258981,-0.2194530517,0.0716567487,-0.0161595419,0.3092053831,0.5854972601,0.1012936011,0.3075446486,-0.2115700692,0.0656297728,-0.0478953719,0.3690636754,-0.1495689005,-0.2604703903,0.46193555,0.1115684062,-0.4515560269,-0.1744281948,-0.0168555081,-0.1202256083,-0.2905308902,0.1574334651,-0.3936344385,-0.0258700233,0.3197914958,0.2921334207,-0.0364387296,-0.1393869966,-0.1364899278,-0.0923476517,-0.1867286861,-0.2591003776,0.0299786758,0.3263767958,0.158237651,-0.3826505244,-0.4629874527,-0.1317451894,-0.2740682065,0.0501399636,0.4171847105,-0.1065525338,0.1590996087,-0.3115161657,0.0425376669,0.2721488476,0.1653776616,0.1622942537,0.0412532613,0.004972287,0.1017699391,0.4132732451,0.1775956154,-0.1318428516,0.1368516535,0.0340915583,0.0879227221,-0.2265187055,-0.2132219374,0.4135329425,-0.3542912602,-0.0208450183,-0.341904968,0.6716285944,-0.0275889952,0.1143835932,0.2011250854,0.3124499023,-0.1880145073,0.2436715811,0.627974689,0.8101684451,0.082661055,0.0925122574,0.6586751938,0.0395746604,-0.0368232056,0.4076533616,0.0628157258,-0.4858413339,0.1058629528,-0.1523804218,-0.2838106453,0.2352132201,0.0837633312,-0.1277507991,-0.0044286442,-0.064094387,0.4120748341,0.105051294,0.0133968638,-0.3504088819,-0.2208383232,-0.1791244447,0.0456856005,0.0559170954,-0.460331291,-0.0490153655,-0.3513252437,0.025259858,0.0259790719,-0.1247499436,-0.2373873889,-0.3729759157,-0.0640467927,0.4174675643,-0.5670108199,0.1183716506,0.2152516842,-0.0177167393,-0.2802428603,-0.1261241585,-0.0766262263,0.1672941893,-0.2428412139,-0.2445152253,-0.3188093901,0.2764324546,0.0674628019,-0.0918487087,-0.1010529697,0.0350776464,-0.0383982584,-0.1744727939,-0.0141755259,0.2645861506,-0.3808227777,-0.0809327662,-0.2153043747,-0.0524789542,-0.1254979968,0.0665172711,0.0475825071,0.0788212866,-0.0933834463,0.1883073598,0.1496848464,-0.0993907899,-0.0204069316,0.5038687587,0.3624753654,0.3046237528,0.2016715109,0.0599654876,-0.1735873967,0.348520726,0.4494480491,-0.50545609,0.1632677466,0.1418393254,-0.2478163689,-0.1363990754,0.5404826999,0.0394274108,0.2122329921,-0.078790158,0.0117737539,-0.4203208983,0.4531821012,0.0842344612,0.317855835,-0.2962430418,0.0066832155,-0.0164402127,-0.0033700881,-0.2555531561,0.0823914781,0.1399247944,-0.1206921935,-0.4355022609,-0.0697562918,0.0899367407,0.1302221864,0.0567968972,0.043143861,-0.4732172489,-0.1366363317,-0.063060604,0.1465227455,0.1113865823,-0.0750959665,0.1088625044,-0.0356663167,-0.3903193474,-0.1736459583,-0.0385061316,-0.180611074,-0.1987410039,-0.0751566067,0.2943124473,0.0307447407,-0.3263241053,0.0314623564,-0.1110586524,0.2427466214,0.0838690251,0.1172857806,-0.0286928974,-0.0710214153,-0.077790834,0.280321449,-0.0209854227,-0.2457625717,0.3363734186,-0.2781675756,-0.4002944827,-0.1711622775,0.555498898,0.2221169174,-0.1055669636,0.0775413066,0.0982765779,0.1608135253,-0.3934295774,0.0250374395,0.14382945,0.0519386157,0.2361174226,0.4788057804,0.3344214559,-0.0777623132,-0.0287034791,0.1193264127,-0.1326758862,-0.0002069844,0.0437338129,0.6355384588,-0.1258766949,0.048518464,0.4811274111,0.1939938515,0.4050464928,0.0457966998,0.0016180743,0.3823404312,0.356831193,-0.1043540016,0.2455842793,-0.0511926636,0.0729859546,0.1793749034,0.492575109,-0.0442359634,0.2195580751,0.7067701221,0.233474344,-0.0554624945,-0.2771449089,-0.1155789122,0.0063219452,-0.082209602,-0.0751169026,-0.1691101491,-0.3798797727,-0.3740451038,-0.2321600467,-0.4588375688,0.2907000482,-0.233600527,0.2934204936,-0.4083950222,0.5953157544,0.2168373615,0.0516919084,-0.1279551685,0.5140501261,0.4316692054,-0.0781318471,-0.0360895768,0.4701982439,0.2200416476,0.2157692313,-0.3217063844,0.0058283256,-0.0964448601,-0.2204750925,0.1055470854,0.0603390783,0.147679463,0.1796262264,0.2984699905,0.0671435148,-0.0987448171,0.217395708,0.1905705482,-0.0963219628,-0.270799458,0.3887585402,-0.2764686942,-0.0880610049,0.0133799966,-0.0852077976,-0.3422196805,-0.0359235778,0.2539328039,-0.1686334163,0.0642235577,0.0385524519,0.0512697063,-0.2165198922,0.5640590787,0.0377694033,0.3752739131,-0.0008870538,0.0180552658,-0.3669318557,0.2662675083,-0.284376204,0.0349879116,0.3976560533,0.15150626,-0.0319140702,0.0935205221,0.2348974049,0.2483495921,-0.0476074778,0.3987116814,-0.3823180497,-0.0468158722,0.3674177825,0.0457460769,-0.0166746844,-0.3289561868,0.231316179,-0.3987184465,-0.0398683809,0.1851515174,0.0911636949,0.0733312145,0.0101804873,0.455673039,-0.0521393791,-0.0057957266,0.1570114642,-0.0347008184,-0.3226483464,-0.3023774922,-0.210318774,0.2482535541,0.168811053,0.4968717396,-0.2597642541,0.167003721,-0.011452727,-0.139312163,-0.0138446279,0.1429246366,-0.2275400609,0.5355846882,-0.4553392231,-0.0262939725,-0.0327978134,-0.06463065,0.0457299575,0.0954742208,-0.1081473455,-0.5193561316,0.6752308607,-0.0691897571,-0.4102226496,0.1890443712,-0.1185931787,0.5248762369,-0.0722400695,-0.2589216232,0.03436606,0.3283545673,0.0191486161,0.0881619751,0.0410502665,0.0724670589,-0.0939228013,-0.2172298282,0.1776498705,-0.1318302155,-0.1210477278,-0.3250689507,-0.5128965378]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"> Yes right now ArrayXD can only be used as a column feature type, not a subtype. \r\n\r\nMeaning it can't be nested under `Sequence`?\r\nIf so, for now I'll just make it a python list and make it with the nested `Sequence` type you suggested.","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":45,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\n> Yes right now ArrayXD can only be used as a column feature type, not a subtype. \r\n\r\nMeaning it can't be nested under `Sequence`?\r\nIf so, for now I'll just make it a python list and make it with the nested `Sequence` type you suggested.","embeddings":[-0.1496909857,0.0466833152,-0.0832266435,0.1190683097,0.2730078399,0.1201471463,0.6236190796,0.1325292736,0.0004305769,0.1169990227,0.2726795971,0.2137202024,-0.148215428,0.0873367488,0.1494974643,-0.3523188233,0.2304175645,0.3752020895,0.1984266639,0.0808257163,-0.1512219608,-0.096542038,-0.2307991832,0.3625943065,-0.2430000156,-0.1153224781,-0.1199117154,-0.083274208,-0.1397193968,-0.5142061114,0.400708884,0.0255213007,0.2322932035,0.1146209612,-0.0001181501,-0.0146637578,0.4348638058,0.0591648482,-0.1229830682,-0.0048914473,0.1011948138,-0.4042523205,0.2644628286,-0.4277358651,-0.0240509287,-0.9039796591,0.0427302048,0.0066637788,0.1252934635,0.2079314888,0.1442271024,0.1297579557,0.4764501154,-0.1806749701,0.4600847661,0.1405650675,-0.294495523,-0.0807835236,0.5120022297,0.1438986361,0.0992064923,-0.3135210872,-0.1309571117,0.070738256,0.4406239986,0.341997236,0.2460750192,-0.258565098,-0.2623868287,0.3397233784,0.4132310152,-0.1750988066,-0.1234787703,-0.1843845248,0.035751652,-0.2876292765,0.0165764224,0.108738631,-0.2263008356,0.0154627264,-0.0096945362,0.1207381189,-0.4009537697,0.3173990548,0.167513445,0.2928384542,0.0707616508,0.2630764842,-0.1044254303,-0.3481960297,0.5756608844,-0.2738814652,0.1003337279,0.2177470624,-0.0703516454,0.0229680296,-0.1750971228,-0.3643649518,0.4070279598,-0.0397160687,0.2340948582,-0.0654597878,0.262444824,0.1516705006,-0.0850233287,0.3085604012,-0.0135916527,-0.2896936536,-0.0956176296,0.135140568,0.069339782,0.0423803516,-0.2960188091,-0.0798206255,0.2877986133,-0.0363064967,0.2192562371,0.2294505835,0.0576040633,0.0001428305,-0.4910034239,-0.2163024694,-0.0469095036,0.2278482318,0.1335302591,-0.0912338495,-0.1209771037,0.3931637108,0.1145647317,0.0539095663,-0.1054858863,-0.0203385726,-0.1385608315,0.0261314381,-0.0925683603,0.149864018,0.1324961931,-0.0177042931,0.1824238598,0.1176808625,0.1183420047,0.1672090739,0.1050793156,0.1216110811,-0.0390887484,0.0213914569,0.0607711226,-0.0322818048,-0.2146696597,0.3445396423,-0.1502037346,-0.2812199295,-0.4611911476,0.0951247141,-0.2679341137,0.0268113408,-0.0855621696,0.0531741716,0.3206283748,-0.0092907809,0.0700603575,0.0084868576,0.0941199511,-0.3910106719,0.0102527263,-0.2070121318,-0.3107109666,0.1050077155,-0.18326886,-0.0511369854,-0.1033404246,-0.0341761559,-0.1830640733,0.211112693,-0.2438436449,0.3546352983,0.4485312104,0.1238437071,0.0761409104,0.0937400237,0.0458837189,-0.1258334666,0.2754472494,0.1234984547,-0.0420816056,-0.2332231998,-0.1978224665,0.13695544,-0.0650237277,-0.3227218688,0.0924911499,-0.3614104986,0.4532861412,0.0346874967,-0.0556673594,0.1533393115,-0.145034492,-0.3850358129,0.1831584722,-0.4505046904,0.1036106423,0.1198770404,0.2142672837,-0.0773840323,0.0693560019,-0.3480018079,-0.7058081627,0.1983203292,-0.331862241,-0.0510807261,-0.0323284976,-0.1590729654,-0.2161258608,0.1869294643,-0.2282211334,0.123631902,0.0654980689,-0.0101885274,0.1269925386,-0.2844240367,-0.3284108341,-0.3324742615,-0.027413439,-0.0462289862,-0.5676476359,0.4337337911,-0.0152600538,-0.3834446073,-0.2677718699,0.2581305206,0.0449853316,-0.0654633492,0.0735676512,0.2084098458,-0.0886835456,-0.1237079054,-0.5375239849,0.0925365314,-0.0326056667,-0.4755496085,0.148510769,0.1747869402,0.0008851513,-0.2654135227,0.1246099547,0.3514433205,0.2007004917,0.5677340031,-0.162301302,0.0468121879,-0.0505783074,0.1966940612,-0.0660024434,-0.082470648,-0.0508012064,0.2496977746,0.0501037538,-0.0416222885,-0.1632547975,0.2374595553,0.3121314943,-0.0719148219,0.0712543279,0.0841196105,-0.1205454245,0.0141477119,0.1988032311,-0.226331979,0.1835254133,0.1453007609,-0.1091471016,0.1152717471,-0.3458954692,-0.0947077125,0.1184055805,0.0416260101,0.4965810776,0.0626622066,0.2471486926,-0.216076836,-0.1896668822,-0.1980236322,-0.0887104049,0.1040930152,-0.2519029975,0.0450727008,-0.2113893479,-0.0883209705,0.0093189608,-0.5655494332,0.0671063662,-0.0881417468,-0.1248616129,0.1353561282,-0.0841241106,0.2511493564,-0.0522039458,-0.2421683669,0.2563661337,-0.0964112952,-0.253352344,-0.0717587173,0.026772311,0.0615802445,0.1554649621,0.0690751523,-0.0449589565,0.2691813111,-0.0338182524,-0.3550164402,-0.2752417028,0.0499213599,-0.2036171556,0.2274008393,0.4706684947,0.1386533529,0.0706332773,-0.5315939188,0.3367731869,0.0535758436,-0.3053088486,0.2561750412,-0.1780288666,0.0555123426,-0.0216427539,-0.0892763287,-0.138226822,-0.3535017371,0.1806749254,0.5079566836,0.2637710571,0.0704964027,0.2626504898,0.2246547639,0.0375612006,-0.0808727145,0.181336835,0.2828784287,0.459724158,-0.174524352,-0.2502916753,-0.0227665305,-0.8539394736,0.03768811,0.1683426648,0.1771163195,-0.1344535053,-0.1827836931,0.1993626356,0.1049528271,-0.0358290114,0.366969198,0.1059952453,0.0051781307,-0.1473027319,-0.0526276715,0.4297555089,0.0648394153,0.0059892545,0.2077502757,0.3540019095,-0.2904821336,0.4884471297,0.2339154929,-0.2760825455,0.3280070722,-0.2727936804,-0.1918717474,-0.3395034969,-0.0488788709,-0.259424895,0.0347659551,-0.132966429,-0.0561419986,-0.1326052845,-0.3195640445,0.1813720465,0.3871634007,0.0519081913,0.0253867153,-0.0014423424,-0.0121569941,0.2845231593,-0.0957220569,-0.029859297,-0.2548687756,-0.1367805749,-0.2171656042,-0.2505752146,0.1864921302,-0.030670274,-0.478213191,-0.1128227264,-0.0605324432,0.2148715109,0.5360186696,0.0449313521,0.1140574813,-0.0836285576,0.1271998137,-0.1206835806,0.4702155292,-0.2056218237,-0.2539267838,0.5924776196,0.1768164933,-0.4166941345,-0.1383584291,-0.0237036049,-0.0376471803,-0.3803611994,0.1894265711,-0.4217669964,0.0324786976,0.3612165153,0.2312781364,-0.083988905,-0.1792088151,-0.1439408511,-0.2551181018,-0.1262041777,-0.3026898503,0.0754062012,0.3523685634,0.1189749911,-0.4888960421,-0.3445180953,-0.0403813533,-0.2198516428,-0.0591379702,0.4252808392,-0.2475018948,0.0428386293,-0.3103660047,-0.0387659296,0.0678389817,0.1222123131,0.1696278155,0.0364888497,0.0777315795,-0.1879545301,0.4526891112,0.0788120329,-0.2287487835,0.2178769261,0.0710788816,0.0160121117,-0.1328125149,-0.2347794026,0.4364160001,-0.3234702349,-0.0671139359,-0.3596933782,0.5855863094,0.1039955765,0.1038193032,0.2081757188,0.1932806075,-0.2010702193,0.2021732032,0.6190626025,0.8599440455,-0.0062218825,0.1288317591,0.7438949347,0.1930784434,0.2325137109,0.4236086905,0.0557883121,-0.4851095378,-0.1173646376,-0.2180395424,-0.3682934642,0.2132682949,0.1600597501,-0.2500894368,0.1049613655,-0.2131554782,0.406516403,0.2036922127,-0.0523841791,-0.1543254852,-0.1582649648,-0.2367913425,0.0568629019,0.1128934324,-0.5091958642,-0.107365869,-0.2736949921,-0.0362202898,-0.0406348184,-0.1031135991,-0.0497532599,-0.2219358832,0.0116089862,0.3405354321,-0.4945535064,0.0969383046,0.2401857376,0.0545359179,-0.3938040733,-0.0949656591,-0.1582011729,0.1575785279,-0.2412554622,-0.2249206901,-0.3560344279,0.3116673529,-0.0391350985,0.0634183511,0.0238667391,-0.0319007374,-0.1545086205,-0.0965021774,0.0643355027,0.1521964222,-0.25870049,-0.2742073238,-0.0816378146,-0.1232521236,-0.1828743219,0.0355989709,0.1239328831,-0.1015351638,-0.0728861839,0.0388308205,0.1103446409,0.0355498604,-0.0067557003,0.6729286313,0.3418900073,0.3343169093,0.3273963034,0.0259483457,-0.1585261375,0.1565001011,0.6011670232,-0.4475329518,0.2417572141,0.1094401851,-0.2066471279,-0.2043613046,0.5340682864,-0.100726746,0.2458524108,-0.211316973,-0.0559875555,-0.3951442838,0.4105908573,0.1197804511,0.2570126355,-0.3223991096,0.0689127967,0.00303749,-0.2505087256,-0.2125860453,-0.0127696898,-0.0930164382,-0.0676522478,-0.3429581523,0.0151142431,0.1378581673,0.1709837765,0.0657344013,0.0698623285,-0.3820566833,-0.1314588785,-0.038152352,0.1883569062,0.1633758396,-0.1374686807,0.233101368,0.0005478095,-0.2975275815,-0.1349496096,-0.1046613231,-0.1948812753,-0.0096147498,-0.1015764996,0.293063581,0.0848869905,-0.4199877083,-0.0363450795,-0.2228241116,0.1653892845,0.0611912757,-0.0188489351,-0.0809586719,-0.0923190638,-0.0486192591,0.258980155,0.0017538444,-0.1703344285,0.4338004887,-0.2293497175,-0.3992435634,-0.2450799495,0.5093851089,0.23236157,-0.193242535,0.1219989136,0.015303757,0.1013645083,-0.3427643478,0.0777716339,0.278752327,0.0246660411,0.1570969522,0.3537639081,0.2022586167,0.0182434991,0.048677206,0.0217560474,0.0099515235,-0.0906862766,0.0676557124,0.6141085029,-0.0121820113,0.1486715823,0.4665028453,0.0863925368,0.3322013021,0.1201978847,0.0987811312,0.2952759862,0.4559447467,-0.1217546389,0.3147561848,0.0493678823,-0.0089232437,0.216412887,0.6006954908,-0.137131542,0.1976409256,0.511880815,0.1762336791,-0.0658931285,-0.2277881056,0.016727699,0.0745887682,-0.0821542442,-0.0925730616,-0.0477430038,-0.2798786759,-0.4828260839,-0.1167276651,-0.3559363484,0.1859839112,-0.2634468377,0.2670085728,-0.4153425694,0.7211953998,0.2131902725,-0.0373687372,0.0074034184,0.5196146965,0.2977152169,-0.0227331184,-0.0710393637,0.4569800794,0.3191640377,0.2489789575,-0.2854171395,0.0154572334,-0.1617582887,-0.2667803168,0.0596194565,0.1258026361,0.1170554534,0.2419991642,0.292016536,0.0533625185,-0.0957194567,0.2782729268,0.1897234619,0.0001534062,-0.168174997,0.4383731782,-0.3106478453,-0.2009638995,0.1728694588,0.0415561311,-0.4118335545,0.0932750925,0.2761673927,-0.1568493694,0.081644617,0.0230174698,0.0471564308,-0.28802827,0.504727304,-0.0302452296,0.5685747266,-0.0571383722,-0.0395858251,-0.3826924562,0.3545950651,-0.3610605896,-0.0181553699,0.4315899611,0.1460476071,0.0048919776,0.1987656057,0.1994733065,0.1266619563,-0.2016122341,0.3863148093,-0.3414880633,-0.1408742517,0.4099110961,0.1535303444,-0.2102467567,-0.2267037332,0.2206877768,-0.2635422945,-0.0387196951,0.2351827323,-0.0236656852,0.0491422862,-0.0218447242,0.4545924664,0.1139241681,0.1612928808,0.1480777264,0.0239156336,-0.2510904968,-0.2117914408,-0.2025730014,0.2957013845,0.1632513553,0.4352879226,-0.2515105605,0.2785423696,-0.149145782,0.0109390384,0.0757076219,0.0707297176,-0.3185372353,0.6374915838,-0.5367581844,-0.0255097523,0.0312038381,-0.2014327347,0.0274927337,0.1525222063,-0.0613231435,-0.5150296688,0.6163955331,0.0207465366,-0.4586499333,0.1995038092,-0.0950660929,0.4964280427,-0.1267415136,-0.1346389055,-0.0836292952,0.3756230175,-0.0157663338,-0.0287526604,0.0455979072,0.0340927579,-0.0245656576,-0.1855210364,0.0254752431,-0.0755705014,-0.1420820355,-0.3219659626,-0.5487014055]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Yea unfortunately..\r\nThat's a current limitation with Arrow ExtensionTypes that can't be used in the default Arrow Array objects.\r\nWe already have an ExtensionArray that allows us to use them as column types but not for subtypes.\r\nMaybe we can extend it, I haven't experimented with that yet","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":48,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nYea unfortunately..\r\nThat's a current limitation with Arrow ExtensionTypes that can't be used in the default Arrow Array objects.\r\nWe already have an ExtensionArray that allows us to use them as column types but not for subtypes.\r\nMaybe we can extend it, I haven't experimented with that yet","embeddings":[-0.0212208629,0.1621364951,0.0191255398,0.1933747083,0.3534964919,0.14655599,0.6769052148,0.1001415253,-0.1353637576,0.169750154,0.1754417866,0.2483867407,-0.2187324613,-0.0489068329,0.1682210267,-0.4851907492,0.1346689463,0.301538527,0.0228541978,0.0704353228,-0.0469424687,-0.1379909664,-0.2132520974,0.2988721728,-0.3213111758,0.0069461325,-0.0415838324,-0.2609827816,-0.2811144292,-0.5396837592,0.4548950791,-0.1367884874,0.2157600969,0.0541257896,-0.0001180573,0.1391656101,0.4717529714,0.1158182248,-0.2676101625,-0.0728083625,0.11657352,-0.3765859902,0.2254805863,-0.2949297726,0.0746296048,-0.8747572899,-0.0137047069,-0.0480413474,-0.0390084349,0.2231349796,0.1390294582,0.1530657113,0.3917603493,-0.1380421072,0.5562932491,0.202924341,-0.3445002139,-0.033176966,0.2575600147,0.1419353336,0.0736942366,-0.2520451844,0.0317469165,0.1535783261,0.3329479098,0.3048067093,0.3153316975,-0.214182049,-0.0936915204,0.358394742,0.4189319015,-0.1743405461,-0.2196032852,-0.1580248773,0.0351172648,-0.0825668797,-0.0719243437,0.1402474791,-0.1435849816,-0.0595130511,-0.0816430151,0.234032765,-0.4148140252,0.2877807319,0.1884303987,0.0627020672,0.1177320853,0.2468917817,-0.1241737679,-0.2687945068,0.3728429973,-0.2962972522,-0.0879029259,0.1672547311,0.0104683796,0.1018036157,-0.17798087,-0.295091331,0.3978040218,-0.0078654578,0.262419045,0.0521328934,0.2533870041,0.1738923937,-0.0227853209,0.2581659257,-0.0524807721,-0.0677274689,-0.1461710781,-0.0010091361,0.0735208914,0.0296862274,-0.26029405,-0.1566109955,0.4041567445,-0.1065996066,0.1935001612,0.3161326945,0.0838794187,-0.0544140004,-0.5178289413,-0.126952678,0.0819739774,0.0538603775,0.1410972327,-0.0200638101,-0.0995899886,0.4002565444,0.0466507711,-0.0912169069,-0.1167473197,-0.0064966544,-0.2495865077,-0.1215882674,0.0517685488,0.1578974575,-0.0178110804,-0.060043335,0.1222077906,0.1848501265,0.1443341523,0.0823199823,0.1457604319,0.1453525871,-0.117805317,-0.1048186645,0.0018953665,-0.0324768573,-0.2453708351,0.1876764297,-0.035435237,-0.2638182938,-0.5372961164,0.0882840231,-0.2288085073,0.0409309901,-0.302485615,-0.0318419971,0.176060617,0.0075882496,0.076031886,-0.1696111411,0.2447959632,-0.3598736525,0.1270972341,-0.1972584873,-0.2495065182,0.0388288461,-0.2804268897,-0.0530519858,-0.0234772153,-0.0567796342,-0.2708203197,0.0650436729,-0.1856776327,0.3091666698,0.535697639,0.3405401111,0.0292397663,-0.1499933153,0.1001170278,-0.1054030806,0.2279826403,0.0375015363,-0.0667348355,-0.1072203442,-0.2658880651,0.2532443702,-0.1829283386,-0.2341534942,0.0166667402,-0.4594129026,0.2168775946,0.0980007425,-0.0008795772,0.0604747236,0.0081520816,-0.4529452324,0.0485944264,-0.4307850301,0.2427630126,-0.0922999978,0.3001604676,0.0703751594,0.1670927554,-0.3774493337,-0.6568467021,0.2559567392,-0.3469610512,0.0396645591,0.0744251609,-0.1671468765,-0.1128228977,0.2092345059,-0.0401513577,0.212995261,0.0444583595,-0.0040363651,0.1563403308,-0.2169308066,-0.2036586553,-0.4877384901,-0.0552401952,0.0020829339,-0.4064585865,0.4541429877,-0.0363356397,-0.2464057058,-0.3527885973,0.2259245664,-0.1012142375,-0.0033876027,0.0250052679,0.170704633,0.0352447741,-0.1194292605,-0.4558752477,0.0306929294,-0.0753997043,-0.3779909015,0.1492902189,0.1526350528,0.0179340709,-0.1737755835,0.1094519496,0.3874122202,0.2484331727,0.5152691007,-0.0953628421,0.1582252085,-0.031422928,0.2399685532,-0.1496783346,-0.1498088986,-0.1329908222,0.1323253363,0.0241526403,0.0498933308,-0.0974258333,0.0543422885,0.2293841243,-0.0608253032,0.0662455857,0.1479123682,-0.2643353045,0.1247423813,0.2334783971,-0.0696420521,0.0761213675,0.1735021472,0.0079988455,-0.0898217037,-0.3006952703,-0.1095322669,0.1824299991,0.1665159762,0.5337377191,0.0058295014,0.3214497268,-0.1366489083,-0.1477596313,0.0287922323,-0.0188665222,0.0288343504,-0.2176436186,0.0899148285,-0.2739956379,0.0458775051,-0.0408506729,-0.5973742008,0.167632103,-0.0625094175,-0.108652994,-0.0140208555,-0.2081925869,0.2753811479,0.0204892047,-0.1851257086,0.1897025704,-0.1098251715,-0.1761683077,-0.0481249504,-0.1175917536,0.0242518932,0.2732262313,-0.0044652619,-0.0865117088,0.2622177601,-0.07133995,-0.3759306669,-0.3747505546,0.0583744124,-0.1334777772,0.2712836266,0.6011125445,-0.088197656,0.05690483,-0.487717241,0.3660814464,-0.049885422,-0.2371715307,0.4311589301,-0.2244034261,0.1103974655,-0.101535663,-0.0475695133,-0.0823078677,-0.2928113043,0.2672373354,0.4423721731,0.2078857571,0.1900187135,0.367729336,0.2935208976,-0.0595102161,-0.0418136157,0.210139513,0.3851242661,0.5554254651,-0.1326040775,-0.2390658557,0.0469727889,-0.91827178,0.0598658696,0.2745571733,0.0965736359,-0.053191565,-0.166264385,0.246754691,0.151778698,0.0734153092,0.236335665,0.0730143115,0.0306824204,-0.0971598029,-0.0720546618,0.3589542508,0.0592509732,-0.0544411652,0.0599308871,0.2749343216,-0.2500484288,0.4926742613,0.032344237,-0.3953185081,0.2900138795,-0.2813062072,-0.1571723223,-0.2730333507,-0.0383777022,-0.2033661604,0.0202534366,-0.0747557804,-0.0579001904,-0.0469464026,-0.370583564,0.2049007714,0.3417923152,0.0460331924,-0.0593374074,0.0681157261,-0.1048078611,0.1422339231,-0.0960889086,0.0151937027,-0.3160892129,-0.1001574546,-0.2141135484,-0.2398774773,0.1352550536,0.036657434,-0.3534593284,-0.0899078697,0.1876164377,0.2133792788,0.5379564762,0.0069758692,0.2843595147,-0.124366045,0.144852668,-0.0040669758,0.3810842931,-0.2507451475,-0.2140431851,0.5493073463,0.2064726353,-0.4260236919,-0.1997698545,0.0115089575,0.0045287209,-0.3076283634,-0.0146603826,-0.3175295889,0.0350648612,0.4345046282,0.2985549569,-0.0186325796,-0.0140904579,-0.1972004473,-0.1879315823,-0.0682576224,-0.3189417422,-0.0060396525,0.3794977963,0.0550575219,-0.3710640967,-0.457519412,-0.0366516896,-0.2041070312,0.0219986606,0.4698269963,-0.1820261627,0.1130624563,-0.389331758,0.0157567058,0.38531968,0.1386947185,0.1158117801,-0.0339307077,0.1972127706,-0.0521699935,0.3916822672,-0.0113974642,-0.1500463337,0.1345305592,0.0340181477,-0.0290015191,-0.0184129588,-0.1875638962,0.376622051,-0.3450867236,-0.0627885014,-0.3435282111,0.6501119733,0.0876190737,0.1313520074,0.239979431,0.4366544783,-0.1913464665,0.3129681945,0.7067781687,0.7163453102,0.0190930758,0.1417754591,0.5875766873,0.0486635789,0.283980608,0.3927360773,0.129973352,-0.4477541745,0.0877565369,-0.1177495793,-0.34262532,0.2784101367,0.1346614063,-0.1673137695,-0.1764114946,-0.2258796394,0.5874405503,0.1703427732,0.0179734509,-0.3542063236,-0.1725925952,-0.0608988442,0.0736890733,0.1389015764,-0.5042874217,-0.0324833021,-0.2297136188,-0.1975996345,0.0037450462,-0.2245842665,-0.0636293888,-0.2012909651,-0.0964099169,0.4480982721,-0.4837576747,0.2040230334,0.222660318,0.1445991099,-0.273784548,-0.1472534537,-0.06846302,0.2030745596,-0.3200408518,-0.1794199646,-0.323414892,0.1967076361,0.0785892904,0.0055461633,0.0469653383,0.0079794964,-0.1857351959,-0.0939311236,0.0207626615,0.2710928023,-0.3104768395,-0.076296635,-0.2866956592,-0.1470494717,-0.1473860741,0.0409676768,0.0230464656,-0.1762244999,-0.0833945051,0.0549644902,0.169942081,0.0378514901,-0.04407572,0.6143606305,0.4108807445,0.3430806696,0.2616029084,0.0466913283,-0.1633810252,0.158455506,0.7001410723,-0.5357528925,0.1756707877,0.0948995799,-0.297806263,-0.1640963703,0.5314976573,-0.0296763331,0.2333259135,-0.2450135797,-0.0202229246,-0.3762752116,0.4165404141,0.0881201401,0.3043981791,-0.2610390782,-0.0439732857,-0.0204675384,-0.0843439251,-0.2042539716,0.0000113159,0.0268676747,-0.0468635187,-0.4833246171,-0.0449608788,0.1146534681,-0.0281990543,0.0764419958,0.0469549857,-0.3073564768,-0.0938722044,-0.0695838705,0.1445753127,0.0750076249,-0.1609966308,0.1316702068,-0.0184961744,-0.2922177315,-0.1102906615,-0.1173489466,-0.1716517508,-0.188548252,-0.100364089,0.3707653284,0.0218675472,-0.4096215367,-0.0622744262,-0.2591798007,0.2143736631,0.1431512684,0.0104114786,0.0338706858,-0.0140721612,0.0108552128,0.240366444,-0.0489341095,-0.2167767882,0.3667788804,-0.2843901217,-0.3287570775,-0.2931697965,0.3823815882,0.3120449185,-0.1300273836,0.1466228366,0.1218251586,0.0957396552,-0.2635981739,0.0599469431,0.2136284858,-0.1237899587,0.1957714111,0.3809493482,0.1733323485,0.0293947589,-0.0499514863,0.0747183859,-0.0583281592,0.0404808596,0.158037737,0.7727530599,-0.0889632329,0.046870362,0.346378088,0.0572017394,0.3092584312,-0.1376058906,-0.0499248654,0.4082030356,0.1194057018,-0.1470287293,0.2353643179,0.0365151912,0.1436935365,0.290863663,0.599858284,-0.0655349493,0.1546136886,0.6458664536,0.1295801848,-0.0608993508,-0.3079247773,-0.1705600917,0.0009383429,-0.0170823447,-0.3197802007,-0.0718168318,-0.3400319219,-0.3777693808,-0.2001766562,-0.3219610453,0.1966742426,-0.2414500415,0.1399931759,-0.2682490349,0.6058333516,0.1733976305,0.0186112095,-0.0114507414,0.4665213227,0.1931339204,-0.1103843972,-0.1041186452,0.3926157057,0.2352757156,0.3206702769,-0.2402933091,-0.1084050834,-0.0384166166,-0.2632315159,0.0478270948,0.1419234127,0.2050636858,0.1260255426,0.2356912047,0.0549224503,-0.0762815848,0.1294129044,0.1068353206,-0.0240246393,-0.1870777011,0.4450000226,-0.2076493502,-0.1497956216,0.1466751695,-0.0625944585,-0.2974499166,-0.141936779,0.184987843,-0.1132821143,0.0068519013,-0.0043787789,0.0315437354,-0.1552274674,0.5530236363,0.0463124178,0.4742758274,-0.0836474225,-0.0495651141,-0.2923718989,0.2667783201,-0.4952572584,-0.0020595165,0.4960637987,0.2562362552,-0.006528615,0.1396308243,0.1487074345,0.3494867682,-0.0630578026,0.4208671451,-0.2486196011,-0.1834609658,0.4288113415,0.0162885096,-0.1165725887,-0.3974457979,0.2201226652,-0.2266689986,-0.0249236375,0.2209421247,0.0251276232,0.1664776504,0.1008820534,0.5126074553,0.0213207863,0.0937749669,0.2348668128,0.0071402397,-0.3104788661,-0.2480721772,-0.241710797,0.1454455107,0.0429885909,0.4380657077,-0.226176247,0.1597023606,-0.0859451815,-0.0857119262,0.0723846629,0.1899730265,-0.2099248767,0.5638707876,-0.6787708402,-0.0031191672,-0.0248711985,-0.2467943132,0.0584506802,0.0574176274,-0.1613289714,-0.3593488932,0.6176453829,-0.1556895673,-0.4157746434,0.1446668804,-0.0950243622,0.4629308283,-0.1521492749,-0.2348794043,-0.1205972582,0.3565760851,0.0417902879,0.0196836405,0.0153246848,0.087272577,-0.009389976,-0.2324377,0.1348852366,-0.0955719352,-0.0258563235,-0.361965239,-0.5236225724]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Cool\r\nSo please consider this issue as a feature request for:\r\n```\r\nArray3D(shape=(None, 137, 2), dtype=\"float32\")\r\n```\r\n\r\nits a way to represent videos, poses, and other cool sequences","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":28,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nCool\r\nSo please consider this issue as a feature request for:\r\n```\r\nArray3D(shape=(None, 137, 2), dtype=\"float32\")\r\n```\r\n\r\nits a way to represent videos, poses, and other cool sequences","embeddings":[-0.1195075139,0.2424518317,-0.0825230032,0.0961937606,0.3023026884,-0.0807917342,0.6678081155,0.1274508834,-0.4523451328,0.2668219209,0.3651503921,0.1533161998,-0.2551758289,0.1488408297,0.2744103074,-0.4257724583,0.137148574,0.5718728304,0.1896610856,0.1134701744,-0.0811235532,-0.0910715684,-0.1586165726,0.1871943772,-0.3031569421,-0.0180845149,-0.1779543012,-0.329739064,-0.1345255077,-0.4514861703,0.2623895109,0.0389636084,0.1677659899,0.1494594365,-0.0001254906,-0.0167882256,0.5140053034,0.0055480083,-0.1965205371,0.0710476041,0.0390492938,-0.5094625354,0.2315442562,-0.261415422,-0.0486275256,-0.5891057849,-0.0373304486,-0.0661952943,-0.0366015285,0.1377487034,0.0999122933,0.2449054867,0.1895420998,-0.06588687,0.5248126388,0.610652566,-0.4015274942,-0.0032514634,0.4093408287,0.0484816991,0.0090829208,-0.0446680859,-0.1138379797,0.1528004557,0.3713417053,0.1253393888,0.415722698,-0.3516628444,-0.1197506264,0.3804672956,0.5132594109,-0.2269854546,-0.1633286476,-0.182523191,0.1375932992,-0.1191869304,-0.1681579649,0.2776908875,-0.3058533072,-0.1067285091,-0.3200063407,0.1812176108,-0.4519753158,0.4415179193,0.2319643945,-0.0324618071,0.0265332405,0.1550103873,-0.1543131471,-0.209668532,0.6051813364,-0.3144930005,0.1713800877,0.0266731288,0.2315782309,0.0007797396,-0.1781047583,-0.2584551573,0.2240985185,0.0240561981,0.1896695644,0.0009135625,0.3361068964,0.0549349301,-0.2350885868,0.1551093906,0.0196652636,-0.2865149081,-0.1086950898,0.1641214192,0.0523839481,0.0812797323,-0.0928827748,-0.1581509113,0.2144554853,0.1445916593,0.2246973515,0.332344532,0.0989628211,-0.0212183725,-0.6472465992,-0.1190168485,0.0493511483,0.1676117927,0.1162116081,0.0217959359,-0.1420206875,0.4201963842,0.0413306095,-0.1068359762,-0.1327309608,0.0284793191,-0.2020907849,0.0345518552,0.1474196017,0.2719275951,0.0682382137,-0.109878473,0.2920950949,0.11498826,0.2518875003,0.1304156184,0.2680571973,0.1843969524,-0.1766609102,-0.0291833673,-0.0108279614,0.0239786617,-0.201913923,0.2528160214,0.0359996557,-0.2911665738,-0.4017091393,0.067572996,-0.1332117319,0.114795804,0.0110675618,0.0392675735,0.1291614175,-0.0920967236,0.0919329748,-0.0653633699,0.0920566171,-0.4038434625,0.1100912169,0.0587433577,-0.3762679994,-0.006013969,-0.178534314,0.059890978,-0.0581475683,-0.0933830068,-0.2455004007,0.1491202414,-0.2815334499,0.0978553221,0.595564425,0.0683713183,-0.0665700361,0.1286036968,0.0778094828,-0.1911045015,0.2700201869,0.1773519665,-0.0744975433,-0.0917330533,-0.3189812899,0.3159507811,-0.1740139723,-0.2062491626,0.0853026509,-0.4805605412,0.5045505762,0.1227777302,0.0911306366,-0.0121926675,-0.0017257027,-0.3954717815,0.1035296321,-0.5231149197,0.226359427,0.0648657605,0.3272498846,-0.1017292067,0.0142545803,-0.2904191017,-0.7621379495,0.111858502,-0.3817775846,0.032397341,-0.0928597674,-0.0941649526,-0.0098006679,0.1435677707,-0.1588366777,0.1365599483,0.0186942574,0.2234544903,0.1755224764,-0.0927112252,-0.1479680538,-0.4886057973,0.1605400294,-0.0935594961,-0.3995197415,0.6104867458,-0.0304089449,-0.3412742615,-0.2333664894,0.3026459217,0.1116786525,-0.1268317699,0.0759551004,0.0232768264,-0.0467142202,-0.0072606667,-0.5716792345,0.145181343,0.1695394218,-0.5125933886,0.3018430769,0.2380748391,-0.0208656117,-0.2693944573,0.2650481761,0.5354461074,0.0164004322,0.5599130988,-0.1488922536,0.1069105789,-0.0834948793,0.3042312562,-0.171058476,-0.2814382017,-0.077081494,0.3215998113,0.2348310798,-0.0816382766,-0.1202621683,0.1673327982,0.3450403512,-0.1063046157,0.1170570776,0.1044826806,-0.167239517,0.0305953287,0.2194487154,-0.2486446351,-0.0095984172,0.1133675948,-0.1509042084,-0.0167223141,-0.1790010631,-0.1775204539,0.2399976403,0.1978987306,0.5377187133,0.1734980047,0.2530478239,-0.2342146337,-0.2563255131,-0.2838896215,0.0782159492,-0.0029493568,-0.1662907004,0.0819261819,-0.0985972136,-0.1753582805,-0.0538215861,-0.8407202363,-0.054962907,-0.0161434188,-0.1332004666,-0.0183025189,-0.1250449717,0.2904634476,0.13499704,-0.1161116585,0.188772127,-0.1279420555,-0.2394805998,-0.0208785441,0.0010634139,-0.0478813611,0.1938603371,0.1102154404,-0.0285726655,0.0893598944,0.1956140846,-0.4436903,-0.4140672684,0.076156199,-0.054647658,0.3043238223,0.6237299442,-0.028849924,0.1416362822,-0.4335573018,0.3720016479,0.0307335332,-0.1980158538,0.2701029778,-0.1913002133,0.1311505884,-0.0774391294,0.1274390668,-0.0436305888,-0.2877110243,0.1742714942,0.4535890222,0.2285217792,0.0646868199,0.3853501379,0.3064377606,0.0319394469,0.0617693998,0.2502890527,0.3139868975,0.3440012336,-0.1534818113,-0.1516786218,-0.079578571,-0.81973207,-0.0414601192,0.1995376348,0.0056264522,-0.3204469383,-0.1587313265,0.2815923393,0.0212312546,0.0466254018,0.3734765351,0.1232596338,0.1193044111,-0.0999317393,-0.247372061,0.320019871,0.1477468759,-0.1601671278,0.0104148667,0.423867017,-0.2608790696,0.6008725762,-0.0103291906,-0.365965575,0.3163512945,-0.168075338,-0.0308003966,-0.2145516574,-0.0248499848,-0.1443897337,0.0874093398,-0.0968138874,-0.0515102968,-0.1967119575,-0.2967807055,0.1497734487,0.2748077512,-0.0645302311,0.0163017251,-0.0485325493,-0.1632586122,0.1895655543,-0.170474261,0.0903790221,-0.2627589107,-0.1558353156,-0.1304805577,-0.2118925899,0.2559705973,-0.0261652227,-0.3173459172,-0.0732128322,0.0702209175,0.1907731742,0.6411178708,0.0179135818,0.1241759211,-0.0189886726,0.082742162,-0.093616195,0.7690871358,-0.2436898947,-0.1289536506,0.4463673532,0.0362390652,-0.4153819382,-0.1920944303,-0.1146053448,-0.0710530877,-0.3726539314,0.1269796342,-0.204790473,-0.0257862285,0.3133607805,0.3239357471,-0.0185655467,-0.0981940404,-0.1191840544,-0.1535016447,-0.1967300922,-0.3282932043,-0.0400127359,0.3237630725,0.2426332086,-0.4854758084,-0.3750821352,-0.1122343093,-0.067532219,-0.1278328896,0.3981783092,-0.0636543259,0.0789537355,-0.4502410293,0.0193289705,0.2227458358,-0.1319336593,0.2497077733,0.0249829702,0.1944114715,-0.0653873235,0.5268790126,0.2793349028,0.007096034,0.0411041118,0.0846434608,0.0907714814,-0.187256366,-0.1035773754,0.3837707341,-0.4485935271,-0.1014221609,-0.2877463698,0.5306910872,-0.055160746,0.1810458004,0.0935235098,0.4269598722,-0.2141290605,0.2697048783,0.6378160715,0.9567376375,0.0340466537,0.3521620631,0.6656085253,-0.0162935015,0.2430692911,0.372667402,0.145352304,-0.4971989691,-0.0267079957,-0.1719360948,-0.4330444634,0.1819643825,0.1326776147,-0.2565492392,-0.0561421327,-0.148504585,0.284632355,0.0706547871,-0.0362541899,-0.1029658169,-0.1446872056,-0.0948462114,0.0010900197,0.0123769473,-0.382865876,-0.1385523975,-0.346088022,-0.0228880644,0.0327561386,-0.1455654949,0.0021425216,-0.282784909,0.0419152901,0.361723423,-0.6093167067,0.0306673981,0.0879783183,0.0515802577,-0.2078803927,-0.1185667887,-0.1019458324,0.3350895345,-0.2927417457,-0.2024456412,-0.3291128874,0.3069601357,-0.0366927199,0.0343520455,0.1367973983,0.0006990238,-0.0908616483,-0.1594605297,0.0150706256,0.245378986,-0.1630985886,-0.342320472,-0.2061365098,-0.1085418239,-0.0423592962,-0.0009979639,0.0876539126,-0.2088279873,-0.2218299061,0.1476569921,0.1027667895,-0.0497641712,-0.0066635017,0.4624760747,0.2039846629,0.3091699779,0.3508840203,-0.0702256709,-0.1864085197,0.2011597306,0.4627959132,-0.5874415636,0.1854407191,0.1240335777,-0.2766520977,-0.0849191099,0.4894613028,-0.1367443949,0.3455241919,-0.1024455726,-0.0035817139,-0.303003937,0.1414215714,0.0851135552,0.3611210585,-0.2015336603,-0.120584026,0.002042823,-0.1557764113,-0.1603776067,0.0274455175,0.1507695913,0.0718493983,-0.6480786204,-0.1603157669,-0.1307603419,0.1700796485,0.042788513,0.0692403316,-0.4720149338,-0.0495758019,-0.0446767248,0.2160999477,0.0868034437,-0.0737600178,0.1676158607,0.0494847074,-0.3284142017,-0.1556387395,-0.0908654928,-0.1441147625,-0.1505562663,-0.063020885,0.3078474998,-0.0997765586,-0.3593606651,-0.1216495261,-0.1980055869,0.3116026223,0.1084488332,0.0544850975,-0.0120753301,-0.0124390088,-0.0697836205,0.2384720594,-0.0767380893,-0.1882880181,0.3575385213,-0.2864848971,-0.3052308261,-0.240088746,0.3885750771,0.0224890113,-0.244790718,0.0734746456,0.1066968888,0.0882798359,-0.2088195533,-0.0326344222,0.1386853456,-0.0958517492,0.1512499154,0.4127186835,0.1199402362,0.047712218,-0.0144565487,0.0668140799,-0.2037325054,0.0214628316,0.1346934736,0.5152418613,-0.0964197069,0.0465311669,0.3629019856,0.1197186559,0.303296566,-0.0718875453,-0.0999749973,0.2413749099,0.1897230148,-0.148165077,0.2805444598,0.0797760114,0.0472671427,0.2441509068,0.5410906076,0.0388212577,0.0845872611,0.2596602142,0.1191754118,-0.0498792417,-0.4547811151,-0.0564660914,0.009520079,0.0025748648,-0.0519653037,-0.0345162302,-0.2187305838,-0.2139027864,-0.1241588667,-0.3498775959,0.1636026502,-0.0278084613,0.230952099,-0.059436243,0.7911989093,0.1474613845,0.0351061746,-0.0823311061,0.3654716015,0.1892609894,-0.0842181817,-0.2363669127,0.3424333036,0.3028413951,0.3057920933,-0.4137303233,-0.0027461108,-0.0482072309,-0.251401782,0.000754809,0.1104696393,0.1996539384,0.3655871749,0.2430269569,0.0529301167,-0.0775848553,0.1095648184,0.2621611059,-0.059065979,-0.2608366907,0.4748169482,-0.3683914244,-0.0772834048,0.0200594142,-0.0214008577,-0.441303134,-0.0922143534,0.3433691561,-0.2222929448,-0.0314979143,0.0078837303,0.0003661563,-0.4076475203,0.4685801566,0.0565804094,0.5203981996,-0.1426042467,-0.0570298098,-0.1997388154,0.2256310135,-0.1827675998,0.0631262362,0.3602494299,0.2822048366,-0.0702932104,0.1351900846,0.0817965269,0.3429289162,-0.009585117,0.3274740875,-0.3078692257,0.0721469373,0.3332428336,0.0698347166,-0.0840914249,-0.3582053781,0.1551063955,-0.3811786473,-0.0411695838,0.1865786612,0.0476588458,0.0602167808,0.3429683447,0.4439255297,0.0021102007,0.0794409737,0.1255158484,-0.1302105039,-0.1910497397,-0.1506793201,0.0171831083,0.4100233316,-0.030391369,0.5122289062,-0.2526367307,0.4674161375,-0.159375608,0.1406831443,0.1890547872,0.2505687475,-0.1429097503,0.473892957,-0.6883052588,0.0614361204,0.1171912551,-0.2447527498,0.1054180637,0.1444072127,-0.0944648013,-0.3961619437,0.6354135275,-0.236148566,-0.5173558593,0.1891828924,-0.0922156423,0.3242469132,-0.0486646779,-0.3283878267,-0.083972618,0.4084095359,0.0294566136,0.0013381647,0.1020154729,-0.0959281996,-0.010799041,-0.1552886218,0.1843918264,-0.0152670406,-0.1727772802,-0.426997155,-0.5472983122]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"@lhoestq well, so sequence of sequences doesn't work either...\r\n\r\n```\r\npyarrow.lib.ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\n\r\n\r\n","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":23,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\n@lhoestq well, so sequence of sequences doesn't work either...\r\n\r\n```\r\npyarrow.lib.ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\n\r\n\r\n","embeddings":[-0.1600914001,0.1551725566,-0.1365799904,0.2011847347,0.2378053665,0.0024272436,0.5615333319,0.0813985392,-0.1561107635,0.1868873239,0.3679437339,0.1477245986,-0.1952098012,-0.0751516744,0.1982174069,-0.3740194738,0.1051326022,0.4515889585,0.0958153531,0.08156012,-0.1296787113,-0.1000082046,-0.2531840801,0.2315312177,-0.254568696,-0.1155647039,-0.1887791753,-0.2444766909,-0.06532906,-0.6357085109,0.2661866248,0.0415746868,0.1111066714,0.2200846672,-0.0001201842,-0.0430640951,0.4993039668,0.0553212203,-0.186354965,-0.0483908392,-0.0165861212,-0.5161793232,0.2466993332,-0.2830962837,0.0256015882,-0.6335045099,-0.0507780239,-0.1111695543,0.0991790667,0.2162883282,0.1357580125,0.1012992039,0.2948573232,-0.137925908,0.4939403832,0.1856482029,-0.2645460367,-0.0064463499,0.4243836999,0.007798085,0.0039838003,-0.2260001004,-0.0914998204,0.1946606338,0.3239576519,0.254425168,0.1728006154,-0.2684161663,-0.1167435572,0.4633981884,0.5487830043,-0.1992025524,-0.2062028348,-0.1623517722,0.0251365583,-0.2495518625,-0.0168566722,0.2230231464,-0.2625147998,-0.0679104626,-0.2960222065,0.2252050191,-0.363474369,0.3561138809,0.2739037275,0.0741850212,0.0362597667,0.1922661662,-0.0426773988,-0.1601042449,0.4481900334,-0.2737982273,0.0411335751,0.2004844248,-0.0477827378,0.0812530294,-0.178754583,-0.2598590553,0.3866290748,0.1006254405,0.2853740156,-0.0497447625,0.2407920212,0.0112909209,-0.1565422118,0.2287863046,-0.0354375876,-0.1955488473,-0.1534389257,0.1481019109,0.1131824628,0.1070677564,-0.0841125026,-0.2671606541,0.342169106,-0.0100373942,0.1621028483,0.2628884912,-0.0015815891,0.1584796607,-0.7266274095,-0.227558881,-0.0102352845,0.234435901,0.067370452,-0.0787664577,-0.1443527788,0.320643723,0.069072336,0.0175569374,-0.2220751792,-0.0282847118,-0.1574119329,0.1480099559,-0.0056370324,0.1456133872,0.1314304173,-0.0021543058,0.1279088408,0.0469316244,0.2403779626,-0.0591416769,0.1808516383,0.1860529482,-0.1897394061,0.0794378445,-0.0132834725,-0.1411844492,-0.2069685161,0.3373622596,-0.0748132393,-0.2489276081,-0.2396021038,0.1195377484,-0.246433869,0.1436358094,-0.0567751117,0.0221291538,0.3504425287,0.0329586603,0.0972870141,-0.0463144891,0.1343153715,-0.3639497161,0.0316104814,0.029170379,-0.4040642977,0.1027748659,-0.0170007963,0.0420777649,0.0010878013,0.0002231555,-0.2483725697,0.316739589,-0.3398919702,0.3151932657,0.4771946669,0.2314243913,-0.1476486921,0.0780263394,0.0087051541,-0.1805401891,0.2836229205,0.1247067675,-0.0706822649,-0.1788303554,-0.2814875245,0.1967943013,-0.0682694316,-0.2683036625,-0.0474518985,-0.3563322127,0.3282252252,0.0095096314,-0.030009063,-0.0579109564,-0.0611507185,-0.3634927571,0.0984668285,-0.3985061944,0.1682198346,0.31026721,0.2540777922,-0.107006073,0.0180024803,-0.2215562314,-0.6752679944,0.1367509216,-0.3360004723,-0.0714466199,-0.1608760059,-0.0421856195,-0.1771296114,0.2556402981,-0.2550107837,0.1528830677,0.0767279789,0.1722394973,0.2379298359,-0.2839889526,-0.129172653,-0.2462118715,0.0235670023,-0.0346851312,-0.4833646715,0.6345498562,-0.0643043891,-0.3637027442,-0.2196398973,0.134125039,0.0457702875,-0.1106671616,0.0889573693,0.2254216522,0.063523598,-0.1646242142,-0.4993397295,0.0038288655,0.1241193116,-0.5690355301,0.0915195569,0.2302256972,0.0061184354,-0.2539628148,0.1037810966,0.3309844732,0.009149394,0.5892533064,-0.1080661267,-0.0582965054,-0.0223790538,0.2074624002,-0.1052798182,-0.1419218183,0.0479141511,0.364918381,0.3332372308,-0.0593755059,-0.0809152201,0.0729771703,0.2331445366,-0.0524769053,0.0030045197,0.0871590674,-0.1870843023,-0.0404211357,0.1665886045,-0.2382638752,0.1528133303,0.1505664587,-0.1506057233,-0.0140659194,-0.3169110119,-0.2260744423,0.1385513544,0.1558432877,0.5817113519,0.1286941022,0.3334386051,-0.1319292635,-0.1932003796,-0.37912485,0.0576524436,0.1512158662,-0.1938259155,0.1737959981,-0.1454391927,0.0365481786,0.0281192232,-0.6938303113,0.0446573012,-0.0963609591,-0.1944794655,0.053863287,-0.1052140892,0.216292426,0.120070532,-0.2124529779,0.2809821069,-0.0610840246,-0.1656045765,-0.0961052477,-0.0122706117,0.0360585079,0.2140233964,0.0470585413,0.0029294493,0.1734709293,-0.0841613412,-0.4023230672,-0.3385605514,0.072231628,-0.0891835764,0.207783699,0.5260062814,0.0793336332,-0.0433284454,-0.5001631379,0.3418544531,0.171259746,-0.2787890136,0.3581455946,-0.2093979269,0.1884738058,0.0264000911,-0.0835735872,-0.0133491298,-0.3030293882,0.2011641413,0.5015564561,0.2515633106,0.2823497057,0.3574547768,0.2633704841,-0.0483494699,-0.0047425479,0.1647399962,0.3353207409,0.3585027158,-0.1083962098,-0.2562660575,-0.1058744043,-0.8942086101,0.0657701492,0.3306562304,-0.0534470007,-0.2664452791,-0.2960111499,0.2652873993,-0.0444679111,0.1249464676,0.3958416879,0.1515047103,-0.0277538504,-0.0751148835,-0.0660347939,0.4360360801,0.0146460207,0.0330353975,0.0122602638,0.3850975037,-0.1809913367,0.6154463291,0.2386411726,-0.2948965132,0.2468746454,-0.2687768042,-0.0160998311,-0.3804985583,0.1405596286,-0.2149468511,0.1324998438,-0.13962695,-0.0848143771,-0.1557293534,-0.1856856346,0.2108891904,0.3257881403,-0.0560119413,0.027062431,-0.0124887908,-0.0062739439,0.2157729864,-0.0722607598,0.0085482197,-0.1657004058,-0.1219093576,-0.2832225561,-0.3324129581,0.2549269199,-0.0420733765,-0.4164330363,-0.069573909,0.0470713563,0.1746811271,0.5719996691,0.0572096519,0.2247126848,-0.1525773406,0.1133042723,-0.1367263049,0.5029539466,-0.1614838839,-0.2173504829,0.5037879944,0.1487830877,-0.3988546729,-0.1728272885,-0.1450513154,0.1050176173,-0.3108483851,0.0374075621,-0.3851986229,0.0350427702,0.2270036638,0.4519897103,0.0308633298,-0.1545875371,-0.1730750501,-0.2341377437,-0.1679702252,-0.1769971251,-0.0772145987,0.3325337768,0.1185417622,-0.5217426419,-0.2831084728,-0.2218744159,-0.2166454643,-0.1168450788,0.4292851686,-0.0959719345,0.1976205558,-0.424410969,-0.0217128526,0.2756628692,0.0755055696,0.2074224353,0.0471863598,0.1186169833,-0.1547592282,0.3812936544,0.1365758926,-0.0977920964,0.1208830923,0.0873417333,0.1731405705,-0.0309425127,-0.1182438433,0.2858746648,-0.3870583475,-0.2409185171,-0.2526113391,0.6065798402,0.0345119163,0.1306643486,0.2011975795,0.3789948225,-0.1907278299,0.1544510424,0.5385934114,0.8276067376,-0.1091832817,0.1095595881,0.6449869871,0.106607236,0.3623753786,0.4651365876,0.1103194579,-0.4643544853,-0.1454156935,-0.0950389653,-0.4753380716,0.1898429394,0.2201741934,-0.2813163102,0.1027400941,-0.2025315613,0.4250156581,0.1137101799,-0.1704331785,-0.1086489186,-0.1236268282,-0.2000635564,0.0582709312,0.0476624891,-0.352681756,-0.1718628556,-0.1864745021,-0.0950225368,-0.1046453044,-0.0701149628,-0.0149997538,-0.3250381649,-0.0454031788,0.4836613834,-0.7281644344,0.019851096,0.1408112794,0.0578351244,-0.1766629815,-0.0700274855,-0.1826733649,0.0984091014,-0.2828502357,-0.1620808989,-0.3218930364,0.1863593757,0.0097959619,0.0841342956,0.1313340366,-0.1087404191,-0.1377047747,0.0513093323,0.1794709712,0.1363498718,-0.2245845646,-0.2015951425,-0.0501844995,-0.2770042121,-0.1344671845,0.0315409824,0.1442112476,-0.1530608982,-0.1088711917,0.0898811445,0.0480570309,-0.0880135372,0.0267976113,0.6479645371,0.2005579472,0.4696872532,0.2697415948,-0.0549195707,-0.1910224706,0.0595995076,0.586403966,-0.6140030026,0.1377328038,0.0855309591,-0.1088808104,-0.0170557722,0.5587212443,-0.146316424,0.3311334848,-0.0994507,-0.0858525485,-0.3952617347,0.2944930196,0.0345837288,0.271448344,-0.2411224842,-0.0056884149,-0.0490241311,-0.0177653339,-0.2013214976,-0.0274353027,-0.1023831442,0.0754278749,-0.551682651,-0.0063281348,0.0060596718,0.1183264703,0.0839471519,0.127722472,-0.4031296074,-0.166673854,0.0490459763,0.1636104137,0.1434490532,-0.1364783645,0.1057919264,0.0164429862,-0.2059158683,-0.1904213578,-0.0890773758,-0.1587847173,-0.1524307728,-0.1798937619,0.3732468486,0.0133501794,-0.4500874281,-0.0993072987,-0.3206630945,0.3353337646,0.0838312656,0.0338701829,0.051851552,-0.0675971135,-0.1365988404,0.2506708503,0.0252409335,-0.282659471,0.3494510353,-0.4587415159,-0.3821809888,-0.1372277737,0.4560286999,0.282740891,-0.1941174865,0.0783236995,0.0601532198,0.1380978227,-0.2400970906,-0.0345075615,0.0729424879,-0.0124859093,0.1442500651,0.3384786546,0.2148675621,-0.1368960738,0.0849252492,0.1068417281,-0.0443745479,-0.0057366113,0.0059639141,0.6895576715,0.0221744031,0.0649674907,0.4045622349,0.0822872296,0.3356209695,-0.0252461229,-0.0928708389,0.3353502452,0.2887383401,-0.0284225773,0.1966649443,-0.0920879543,-0.0685998797,0.2863985002,0.5743163228,-0.2170487791,0.1065245569,0.3571174145,0.2653896213,0.0638024434,-0.2555142343,-0.1167544946,0.076863952,-0.0006834709,-0.1098319069,-0.0275489651,-0.1941812634,-0.3827090263,-0.1505114585,-0.3897601366,0.2327124476,-0.1659683436,0.217774421,-0.3669434488,0.6742689013,0.2101667672,-0.0369131491,-0.1713701189,0.5313136578,0.3248063922,-0.0792083368,-0.109372057,0.3668127358,0.3572769761,0.3015879095,-0.2407060564,0.0795765072,-0.1918934137,-0.2566197813,-0.0545774847,0.1561476439,0.1720080376,0.0037681966,0.2482557595,0.0709493607,-0.1256276667,0.3689872921,0.2804300487,0.0370747373,-0.1932657063,0.3348279893,-0.2718710303,-0.1903453022,0.1311919838,0.0063216169,-0.4402512908,0.1030826345,0.3191308975,-0.1223135218,0.0201656986,0.0450456142,0.0341216549,-0.3694950938,0.4734950662,0.1577860117,0.6076859236,-0.0797691569,-0.1412038654,-0.2672116756,0.2897810042,-0.4417595565,0.0740942881,0.3080429435,0.2579082847,-0.0487484224,0.2920187712,0.0772694871,0.2233107686,-0.0479865447,0.370135814,-0.3534585536,-0.1368518323,0.3459899426,0.0485213697,-0.2119519264,-0.2849324942,0.2408120632,-0.3483982682,0.0187243968,0.1598386467,0.1610459387,0.0116113117,0.2485992014,0.4437896311,0.1199980304,0.0778709948,0.1355816424,0.0182929877,-0.2765599489,-0.2072043121,-0.0377060212,0.3501323462,0.1534585506,0.4672709107,-0.2480130494,0.3416921794,-0.0683879927,0.0491894558,0.0788117573,0.3267088234,-0.2487181872,0.4114189446,-0.5436223149,0.0665598288,0.0299387965,-0.1942564696,0.1354703754,0.1080585867,-0.1219415665,-0.5837486982,0.6118191481,-0.1655677557,-0.4106805921,0.1020886526,-0.0144821117,0.3352402747,-0.1222288311,-0.1329847127,0.0158910025,0.4949216545,-0.0330866501,-0.2015378773,0.0128793484,0.0020365329,0.036506325,-0.1622644216,0.2119116932,-0.007995436,-0.1454565674,-0.3897952735,-0.5064119101]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Working with Arrow can be quite fun sometimes.\r\nYou can fix this issue by trying to reduce the writer batch size (same trick than the one used to reduce the RAM usage in https:\/\/github.com\/huggingface\/datasets\/issues\/741).\r\n\r\nLet me know if it works.\r\nI haven't investigated yet on https:\/\/github.com\/huggingface\/datasets\/issues\/741 since I was preparing this week's sprint to add datasets but this is in my priority list for early next week.","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":67,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nWorking with Arrow can be quite fun sometimes.\r\nYou can fix this issue by trying to reduce the writer batch size (same trick than the one used to reduce the RAM usage in https:\/\/github.com\/huggingface\/datasets\/issues\/741).\r\n\r\nLet me know if it works.\r\nI haven't investigated yet on https:\/\/github.com\/huggingface\/datasets\/issues\/741 since I was preparing this week's sprint to add datasets but this is in my priority list for early next week.","embeddings":[-0.2213128656,0.1979756653,-0.0157015547,0.364423275,0.3536314368,0.0121849468,0.5489980578,0.1765438765,-0.2264501303,0.2063268274,0.2152808607,0.2304821461,-0.1810271591,0.0037143859,0.2849797308,-0.3763326705,0.152505219,0.3805985153,0.0335737132,0.1360763907,-0.039774064,-0.0720350146,-0.1213939786,0.1114183366,-0.4059671164,-0.0048565464,-0.174713254,-0.1200739592,-0.2310266197,-0.5286946297,0.1810473055,0.0685255677,0.0716023594,0.2211990356,-0.0001212513,0.087012656,0.5184174776,0.0738940537,-0.3144264519,-0.025351407,0.1473908573,-0.4104197919,0.2535746694,-0.2232877612,0.0524131991,-0.6484952569,0.0717870221,-0.0252834801,0.1690187305,0.2796278894,0.1555815488,0.2720468044,0.3203712106,-0.1293947846,0.4958250821,0.4110741317,-0.2912885845,-0.0375179946,0.2439015657,0.0963079259,-0.2048285753,-0.0812970996,-0.0461048447,0.1272631735,0.4374938905,0.2534179091,0.1801085174,-0.2713246047,0.1292905211,0.3118749857,0.5142441392,-0.2984479666,-0.168679297,-0.2406481653,0.0696059838,-0.1214201823,0.0500409193,0.2096702605,-0.3758709431,-0.1750664711,-0.2559000552,0.0531944819,-0.3825436234,0.1938744485,0.3322298229,-0.0729309767,-0.0771721229,0.1957787424,0.0948959142,-0.0844441727,0.4489100277,-0.327493161,-0.0603616014,0.1682132185,-0.0260267965,0.1914516836,-0.0618921444,-0.2489123344,0.3557296395,-0.0146377338,0.0907957405,-0.0260442365,0.4090178311,-0.0300462581,-0.1666974127,0.2424432188,0.0268453769,-0.0564475209,-0.0126626221,0.1458721906,0.1368909776,0.0023782705,-0.3036974669,-0.3543183804,0.4389374256,-0.1597704291,0.2262702584,0.2719847262,0.0580549575,0.0527417213,-0.3872817159,-0.1634101421,-0.0512050129,0.2605142593,0.0079418793,-0.1695230752,-0.0749827847,0.3475945294,-0.1134954691,-0.0771844387,-0.2006068528,0.014128658,-0.1872189641,-0.0355556086,0.1856772751,-0.0006738221,0.1712860465,-0.1005151942,0.1102587208,0.0364879072,0.2294089049,-0.003272105,0.0745189786,0.2431539297,-0.1059862971,0.0759956688,-0.1007448137,-0.1035404727,-0.1938961297,0.3595401347,-0.0072725825,-0.3163636327,-0.360494107,0.0677369311,-0.1884798706,0.1095320061,-0.333424747,0.0241517015,0.2464962453,0.1929247975,0.0468096547,-0.0788814947,0.1197319925,-0.4186955988,0.255220294,0.1506827325,-0.3296794593,-0.0764193609,-0.1774583757,-0.1453157067,0.014685709,0.140763998,-0.2277489454,0.316801995,-0.1969990581,0.1646838039,0.589875102,0.1421812475,-0.1942748576,0.0489282757,-0.0359419808,-0.1330453902,0.273894757,0.1337053478,-0.0858033672,0.0003619304,-0.2265037745,0.2427281588,-0.2060781866,-0.1284230053,-0.102871716,-0.4257862568,0.2570198774,0.1826580465,0.1883201748,-0.2002343237,-0.0763456523,-0.4452912211,0.1509934217,-0.4555953145,0.3199744821,0.2012659609,0.2738685608,-0.0686623827,0.0919819102,-0.1971551031,-0.6865983605,0.2245500386,-0.3239831626,0.1136917248,-0.1173114628,-0.038934432,-0.0470360108,0.2342349738,-0.1890112758,0.0169434957,0.0304380786,0.0079185683,0.2468836457,-0.1551266462,-0.1208433658,-0.3237175345,-0.0724272728,0.0879729837,-0.4804650843,0.3551844358,-0.0268486403,-0.325533092,-0.2417758107,0.1261120737,-0.0964428559,-0.0699635223,-0.0347927138,0.1379459947,-0.1718249172,0.0094862087,-0.4415675104,0.0668799579,0.0340707116,-0.3649432659,0.3318942189,0.0976708978,0.0095571307,-0.2681993842,0.023220323,0.3331259191,0.1053549051,0.5135986805,-0.0469888523,-0.0121937897,-0.0900004432,0.2590038776,-0.0001805388,-0.0954834595,0.0781409889,0.1912837923,0.3023990393,0.1007427201,-0.0900859535,0.018186193,0.3622381687,-0.0750440434,-0.0154785076,0.2164575905,-0.3582879603,-0.0443061814,0.1146616265,-0.2236914337,0.0958220214,0.1896508783,-0.0962158144,-0.0030712364,-0.1216244921,-0.2456038296,0.2308207005,0.1090443581,0.4995390177,0.0859795809,0.2474468946,-0.2025018185,-0.295427382,-0.1155417562,0.1100827605,0.211879313,-0.3103191555,-0.0364510939,0.0036005985,-0.0263739657,-0.0472575314,-0.4187700748,-0.0772805512,-0.0021739975,-0.1159006581,0.0387827083,-0.139152199,0.2060878724,0.0957118347,-0.1754145473,0.1691356301,-0.0827374086,-0.1555674076,-0.0700948238,-0.0628747791,0.0273880363,0.4325107038,-0.0062161619,0.1022777036,0.0694009662,0.007533317,-0.386561662,-0.3180573881,0.0014939316,-0.1885012686,0.3855703175,0.5089985728,0.0129553992,0.0704964399,-0.2351247966,0.4283284843,-0.0620499067,-0.2789503932,0.2951141894,-0.2495291084,0.0132227372,-0.0754204839,-0.0625773817,0.0871267617,-0.3013954461,0.4161198735,0.389233321,0.2587939799,0.2775218785,0.4537557662,0.3612791598,-0.1578188688,0.0766769573,0.2665534019,0.0884502903,0.4034970403,-0.0317742452,-0.2591501772,-0.0606958941,-0.6752923131,0.0791358799,0.1878122836,-0.1335006803,-0.0855476558,-0.3491368592,0.3653412759,-0.0175711326,0.2512726486,0.320278734,0.1309190243,0.0220842119,-0.0302399304,-0.153926909,0.4097852707,-0.0565460175,-0.116648674,-0.0444961376,0.3685621619,-0.2781518102,0.7187972069,0.1594279408,-0.1908810735,0.1735194027,-0.1888525188,0.0868747234,-0.2385756671,-0.0506617054,0.0356894806,0.0839146897,-0.0330270156,-0.0050410959,-0.0146538522,-0.1548367441,0.24451828,0.0944903418,0.0179521143,-0.0938047469,0.2726668119,0.0028944542,0.1238576844,-0.149591431,0.0174952671,-0.3889083564,-0.1801965088,-0.1647979021,-0.3463932276,0.2637693584,-0.0110869445,-0.2853621542,0.0391526967,-0.3850506246,0.1164442077,0.4242253006,-0.1564018428,0.2217938155,-0.020631751,0.1937055439,-0.1511175483,0.6794896722,-0.1187513247,-0.3409122527,0.4274685681,0.1553490609,-0.5605745316,-0.0846424997,-0.1652865559,0.2461470217,-0.3158963025,0.1086696014,-0.4143292308,-0.143332541,0.4140113592,0.3913716078,-0.0545659177,-0.1851378083,-0.2121137828,-0.3040747941,-0.185867548,-0.1297897249,0.0128500536,0.3727321029,0.1485012174,-0.3051449358,-0.2967606783,0.0164086558,-0.1248364672,-0.0524399094,0.4077995121,-0.2468923777,0.1785911322,-0.219857946,-0.066586487,0.3268235326,0.2026312947,0.0536713153,-0.1761280149,0.1134072244,-0.0393439829,0.4988567531,0.2018640488,-0.1145921573,0.1512622535,0.1214178726,0.2522668242,-0.1870196611,-0.0860506073,0.4117688835,-0.2626804709,-0.2254761159,-0.3236449957,0.551341176,0.0427404344,0.2184253186,0.3525519073,0.4389819205,-0.2235934883,0.240086332,0.48352319,0.889210701,-0.1507897824,0.2404503524,0.6072924733,0.0055065746,0.2729428709,0.2393169552,0.2225950211,-0.5379677415,-0.0511485264,0.0709353015,-0.351935029,0.2847338319,0.1038029492,-0.113371864,-0.041759029,-0.1383735389,0.4122714698,0.0663561374,0.1267137676,-0.2973128557,-0.3045106232,-0.0827562883,0.0312231369,0.0454034135,-0.3888436854,-0.059233781,-0.1134186536,-0.0788849071,-0.0733949542,-0.2180921733,-0.0616668798,-0.5097700357,-0.0959413424,0.4721806347,-0.4421561658,0.2003863305,0.133344233,0.125382334,-0.0804616883,-0.1025497913,-0.0346536338,0.1324410588,-0.1809815466,-0.0668337345,-0.3247702718,0.1169551611,0.0500529744,-0.0929776728,0.0070038922,-0.1274048686,-0.0782844797,-0.0345480293,-0.0546572208,0.1990982145,-0.2799147964,-0.2094139904,-0.2535100281,-0.1570219547,-0.1931712776,0.0391988233,0.1604587138,-0.1208101138,-0.0720693618,-0.0012566973,0.0288351532,-0.0694504529,-0.0455084145,0.545869112,0.2170066535,0.5177696347,0.3589176238,-0.0498760529,-0.1814766526,-0.0332918577,0.6822755933,-0.7731728554,0.2939394712,0.1818484813,-0.2421348989,-0.1677031517,0.5106953382,-0.1247626171,0.1342896074,-0.1064627394,0.1037916765,-0.4808640778,0.2990931571,0.2030528039,0.3578311801,-0.1786769032,-0.0927868485,0.0718785152,0.1234466881,-0.1977482289,-0.0050352453,0.0958986282,0.0161493141,-0.4035745263,-0.0425034128,0.244857952,0.0397890583,0.0327779874,0.0198103879,-0.3607736528,-0.1390213519,-0.0378615893,0.1455412954,-0.0371487439,-0.1967011839,0.2865374088,-0.0224830154,-0.2330085188,-0.2979308665,0.1548816115,-0.0779568106,-0.185362801,-0.2841272652,0.3411080837,-0.0969538465,-0.3457170725,0.0179397427,-0.2194333822,0.2988488972,0.0271401107,0.1437520981,0.1295877695,-0.0372080356,-0.0095057487,0.2139303833,0.107495755,-0.2173411846,0.2024222016,-0.4366179407,-0.4164586961,-0.1429621875,0.3953709006,0.3089133501,-0.061444547,0.1281326264,0.1037924513,0.1279309541,-0.2023244798,0.0143392477,0.1377169788,-0.1896725446,0.1461174488,0.3554198146,0.0605182238,-0.0066237124,-0.1027262807,0.1862227917,-0.0740233734,0.0466209091,0.1086216718,0.449513793,0.0012580429,-0.0275560394,0.4771541357,0.0175012164,0.294362247,-0.071062021,-0.0856120735,0.3696001172,0.1361246556,0.0128993234,0.0857531577,-0.1288852394,0.1021619514,0.23273471,0.3544176221,-0.055894129,-0.0186507814,0.7463912964,0.2487951219,-0.0134416185,-0.3547645509,-0.0538713075,-0.1985649467,-0.1277568191,-0.079432711,-0.0791804641,-0.2595430911,-0.0793813318,-0.2000027448,-0.4033377767,0.2291912585,-0.1226261854,0.0847561508,-0.3299498558,0.6227366328,0.2193329781,0.0030638345,-0.1611844897,0.4744717777,0.2397120148,-0.2144979835,-0.3211129904,0.6006444693,0.446781069,0.3391548395,-0.2090889215,0.0397164747,0.0275916774,-0.1698438078,-0.068292506,0.0773958638,0.2278274596,0.1914871484,0.1424110532,0.0518179238,-0.1375400573,0.0341773294,0.2019423246,-0.0764239132,-0.1230496764,0.2156752646,-0.3101531267,-0.239506349,-0.0163392406,-0.1258933097,-0.3975880146,-0.062582612,0.4661124051,-0.1991132647,-0.0436265245,-0.2300644219,0.0351974592,-0.1980310678,0.5869739652,0.1621029526,0.4991671443,-0.3921733797,-0.0479744226,-0.3389751017,0.2338691205,-0.4670828879,0.1145958081,0.2771801353,0.310375303,-0.122502625,0.0611137338,0.2008219659,0.2744983733,0.0680852607,0.3729521334,-0.2622807026,0.0341850892,0.3085058033,-0.0292830318,-0.0148930056,-0.4345167279,0.266192019,-0.3839818835,-0.0235680938,0.0823846757,0.21931988,0.2361361831,0.0669240206,0.4159300327,-0.006545987,0.0268067718,0.1588429362,-0.2340559363,-0.3672115803,-0.1426383853,-0.2464659214,0.2377554029,0.1127895415,0.496814996,-0.3303936422,0.1746419966,-0.2260749489,0.0220769793,0.0435158089,0.2901391983,-0.1415717155,0.4183501303,-0.6275558472,0.1206127107,0.1010946408,-0.1684278399,0.0512659326,0.0253991466,-0.297852248,-0.4806821644,0.5629156828,-0.3771930635,-0.4562425315,0.0751595721,0.0318003558,0.4164333344,-0.2171494812,-0.4832150042,0.0156035768,0.3150599599,0.0608460791,-0.0468185581,0.1661626995,-0.0293317288,0.1080473289,-0.1505972147,0.3135237694,-0.1069214866,-0.1743168086,-0.2626368105,-0.5996109843]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"The batch size fix doesn't work... not for #741 and not for this dataset I'm trying (DGS corpus)\r\nLoading the DGS corpus takes 400GB of RAM, which is fine with me as my machine is large enough\r\n","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":37,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nThe batch size fix doesn't work... not for #741 and not for this dataset I'm trying (DGS corpus)\r\nLoading the DGS corpus takes 400GB of RAM, which is fine with me as my machine is large enough\r\n","embeddings":[-0.3745762408,0.1280584186,-0.0476197153,0.2082384676,0.2405514866,-0.0033118401,0.498398006,0.1474630982,-0.2932840288,0.1866018623,0.2814418674,-0.0539645143,-0.1909173727,-0.1175011471,0.2229365259,-0.2851468623,0.1425438225,0.3723789752,0.0049171061,0.0282003712,-0.0874818787,-0.1740174294,-0.2596130073,-0.1141795963,-0.2867548764,-0.1924320161,-0.2385251522,-0.1416583657,-0.1820704788,-0.5740127563,0.1513036191,0.002134026,0.2676509023,0.41988343,-0.0001258403,0.0140420245,0.5327542424,-0.0296705477,-0.3078179061,0.1804806888,0.1160388887,-0.5670775771,0.0755023509,-0.276815325,0.0667396337,-0.6155688763,0.0792861879,-0.0935679004,0.081132412,0.268692255,0.0758143887,0.114766039,0.2332258224,-0.2019833624,0.45135355,0.2915461063,-0.193495363,0.0987177491,0.479788363,0.2244311869,-0.1344648898,-0.0934096575,-0.2257816494,0.278609246,0.2249961346,0.1619924605,0.502917707,-0.2946100831,0.1333659887,0.3447367847,0.5489017963,-0.1115242913,-0.2093063295,-0.2974889278,-0.0686773434,-0.1175813824,-0.2431563884,0.2419317216,-0.2852624655,-0.1375110447,-0.2829046547,0.1252440214,-0.3688933253,0.3038164377,0.2715136111,0.0086959489,0.0880789384,0.2428213358,-0.0325247385,-0.147288844,0.3926494122,-0.1554118097,-0.1184123755,0.1431395561,0.0588827133,0.1550054699,-0.0380720459,-0.2888388336,0.4321164489,-0.1063440517,0.0138710439,-0.0473383963,0.5867905021,-0.0926304832,-0.1287658811,0.3904908001,-0.197296083,-0.0416502208,0.104577817,0.1069969684,-0.0620186403,0.2088286281,-0.2877637148,-0.4063176513,0.4093298316,-0.1046169549,0.0710790157,0.1694402844,0.0670005009,0.1073107868,-0.3212196827,-0.0598258451,-0.1791732311,0.3538230062,-0.0039107813,-0.0253145602,-0.067893751,0.2610295117,-0.0380551033,-0.0714618713,-0.1640246958,-0.0808365121,-0.2309175581,-0.0260763094,0.2522598505,0.1547982097,0.2087245136,-0.2251219302,0.2305276841,0.0526594669,0.1099237651,-0.1192197204,0.2041054815,0.3068583906,-0.0317756459,0.0919987112,-0.0652691722,0.0603402667,-0.2398638576,0.4755078852,-0.1224317998,-0.2511698008,-0.2756825387,0.0325488411,-0.234231621,0.0859296247,-0.3321462274,0.0839202031,0.4152879715,0.0884541422,0.1165011972,-0.1233680993,0.0668459609,-0.4432203174,0.0363993905,0.1962483078,-0.3947774768,0.0962427631,-0.017629914,0.1693833172,0.2834205925,0.1204254329,-0.0769041181,0.2557766438,-0.1914673746,0.244873032,0.5503516197,0.037440028,-0.3248765469,0.2341147214,-0.0235389136,-0.279196918,0.3945928514,0.1149656698,0.0964222103,-0.0518599004,-0.1384651363,0.3612681031,-0.2345642298,0.073998265,-0.1379666179,-0.5268688798,0.0947980955,0.1622679532,0.3325763345,-0.1004156545,0.0353725329,-0.3560432196,0.164801836,-0.4008250535,0.3126735985,0.1492701322,0.2288143933,-0.1841290146,0.2083989829,-0.1813565493,-0.6962065697,0.2876124978,-0.5014748573,0.0739164948,0.0391726717,-0.0429699719,-0.1620541066,0.0703147948,-0.06924849,0.0174353831,-0.0559540018,0.0669509843,0.3730543256,-0.1842379272,-0.0645482466,-0.3771052957,0.0573678315,-0.0027467362,-0.3606955409,0.204564333,0.0023691175,-0.4256267846,-0.1967066675,0.07872767,-0.0671877041,-0.1091160476,-0.0298987515,0.1906698495,-0.1057278588,-0.1265948117,-0.5945059657,0.1133438274,-0.0075300592,-0.1709633023,0.3066866398,0.1203050762,-0.0383864716,-0.2761614025,0.1548981369,0.2250410914,0.0956047401,0.4443617761,-0.0483179651,-0.1787980348,0.0536852293,0.1091341153,0.0232802946,-0.1872337908,0.0573429167,0.1311627775,0.2436996251,0.017907491,-0.1004728675,-0.0674117953,0.2330175042,-0.0413283221,0.1955351233,0.1248196363,-0.3651072979,-0.0716262013,0.1410181373,-0.2624342144,0.1857797205,0.2264382243,0.0025504262,0.0076846564,-0.0249636211,-0.2826313972,0.1699322462,0.0370539054,0.5360158086,0.2710428834,0.2685303688,-0.1529060304,-0.259624213,-0.0702255815,0.1242118999,0.1814390123,-0.2733660638,0.0103877597,0.0000498165,-0.1357454062,-0.0412943549,-0.39041695,0.0741709098,0.0874913856,-0.03729425,0.0703357086,-0.1533456594,0.1652028412,-0.0397780389,-0.0418311208,0.2344289571,-0.2710824609,-0.093743071,-0.0093460726,-0.2818738222,-0.0323999524,0.3056126237,-0.0038354909,0.1419288516,0.0992972031,0.0966010988,-0.2507053018,-0.2788382769,-0.0341021232,0.0217122845,0.3230953217,0.4574573636,0.1499737054,0.1179577932,-0.1752307415,0.3148290813,-0.1135184541,-0.2176906615,0.1710073501,-0.3160590231,0.0326278284,-0.12103533,-0.0494649559,0.0112170922,-0.3729518056,0.0696581304,0.4055189788,0.2205289751,0.3084494472,0.4828367829,0.2367551923,0.1243309602,0.1658313274,0.1984265894,0.2151188701,0.4572960436,-0.0293485895,-0.2974740267,-0.1902780086,-0.6981368661,0.0184096936,0.3738331199,-0.2694817185,-0.0245692376,-0.3942727447,0.1062362418,0.0583179519,0.2845515907,0.3782826662,-0.0474195741,-0.0254100319,-0.070316568,-0.0584218055,0.2494764179,0.1239886656,0.1173087806,-0.030874934,0.4462179244,-0.1945865005,0.7147114277,0.0947718918,-0.3038935959,0.2707913518,-0.1401407123,0.0170624387,-0.2703028619,0.0372008979,0.0961655378,0.1277515888,-0.2376380712,0.147577107,-0.0591038987,-0.3861612082,0.1775975823,0.0471052751,-0.079025887,-0.0444594808,0.2812482715,-0.2039060742,0.1338864267,-0.0792910755,0.0368071906,-0.3127509654,-0.2526505888,-0.1902708858,-0.4711126387,0.1410615444,0.0433599874,-0.1022919491,0.1238356009,-0.3025414348,0.2792179883,0.4346424043,0.0007252608,0.239561975,-0.0787819773,0.2065567672,-0.0970239639,0.6973791122,-0.1804163754,-0.0816429183,0.3259969652,0.1955783814,-0.5518755913,-0.2422317266,-0.0672410876,0.0954443887,-0.1372193843,0.1183144376,-0.4352432191,0.1217412353,0.3166222572,0.4247947931,-0.0536097847,-0.2296970636,-0.2966372073,-0.33488819,-0.2174382508,-0.0390176289,0.1008377224,0.3269643188,0.1987349838,-0.3782702684,-0.4008757174,-0.0408067107,-0.0739450455,0.0705563203,0.3742887676,0.0263964348,0.0172824562,-0.2732777297,0.1370510608,0.3507034481,0.1135034859,-0.0043431744,0.0875767395,0.1709027439,0.0145839108,0.451887995,0.307462275,-0.0481159203,-0.0361696519,0.1742116958,0.3180819452,-0.156324178,-0.0163770784,0.2593966722,-0.3316501975,-0.4587730765,-0.2922759056,0.5466743708,0.1049683988,0.1723932177,0.296672076,0.2643334866,-0.3065364063,0.4464043677,0.5022369623,0.8925898671,-0.1167143658,0.2179514915,0.3722965717,0.2857913077,0.2620509565,0.3009260893,0.1885157079,-0.4703862965,0.0611447133,0.1075822785,-0.4274890125,0.1937937587,0.3318332136,-0.2136283815,0.0555977784,-0.2017822564,0.3583788574,-0.0195149053,0.0569366962,-0.0770813897,-0.2509154081,-0.0300210342,-0.0220964756,0.1072167903,-0.309563607,-0.178305164,-0.1149400324,-0.1804322004,-0.1578804851,-0.3014234304,-0.1294432729,-0.3661155701,-0.026383616,0.3523427844,-0.5593370199,0.0127384728,0.0975904837,0.0904765427,0.0514149405,-0.0709397942,-0.0683585927,0.2398952693,-0.2562881112,0.1004847512,-0.3581423759,0.1616695076,-0.0170577168,-0.0635131821,0.0105686607,-0.1707508564,-0.1043334305,-0.1473851949,0.0010245384,0.0947536901,-0.2414893955,-0.3821020424,-0.1939329058,-0.0871561617,-0.087881431,-0.0287432559,0.0897853151,-0.0756546706,0.2411969304,0.07993716,-0.1455174387,-0.0500123762,0.0268446617,0.652841568,-0.0308869947,0.4302845597,0.3163838089,-0.0479409806,-0.1371314228,-0.0058023715,0.4690588117,-0.6159567237,0.3625900447,0.1511606723,-0.274306953,-0.005990793,0.4451167881,-0.0727246031,0.2775433064,-0.1605949402,-0.0049943812,-0.4703550041,0.2354563028,0.1205837205,0.3705757856,-0.1522597969,-0.2207000256,0.0976174697,0.1611382365,-0.1268955171,-0.0327624455,0.0083875842,0.016170729,-0.6082546711,0.1483255774,0.0897213966,0.1500920355,-0.0480168127,0.0465208255,-0.2260247022,-0.0851690322,0.0268543046,0.2054857463,-0.0310735926,-0.0522642061,0.3520442843,-0.1590041667,-0.3431046009,-0.1962460577,0.1736917198,-0.0258746352,-0.2029256821,-0.2717243135,0.3582440913,0.0095413029,-0.1717250943,0.1633916944,-0.282449156,0.2918496132,0.0643572807,0.2059909552,0.1921402216,-0.0676647723,-0.0345501341,0.1483297795,0.2114979178,-0.2168189287,0.1892520189,-0.4090180397,-0.4071270227,-0.0274290591,0.3321344554,0.3150841296,-0.0426493622,0.0057695634,0.072525546,0.1130340099,-0.2808313966,0.044814311,0.1133657843,-0.0556092113,0.1534712762,0.4482033253,0.0966031402,-0.0977462307,-0.1922983527,0.2072163969,0.0730999932,0.0531509891,-0.0388383307,0.4976275265,0.0606212132,-0.1085691899,0.3637855053,-0.0256396662,0.4311863482,-0.0151953828,0.0178498328,0.4252934754,0.2894132137,-0.0267558768,0.0356651805,-0.1030520648,-0.0276338998,0.4369136095,0.4502414763,-0.0880770609,0.0422735326,0.4286779165,0.3360227644,0.0045400448,-0.3861785531,-0.1956740022,-0.1396336108,-0.125927195,-0.2316329479,-0.0897772834,-0.0855601579,-0.1662969589,-0.2277532518,-0.4116145372,0.0615376234,-0.1246627942,0.0069563841,-0.2559812367,0.5384840965,0.0866563171,-0.0274610091,-0.2318942249,0.4704699814,0.2490822524,-0.1112724319,-0.1051201373,0.5898385644,0.5599634647,0.4387229085,-0.2553031743,-0.0112182293,-0.020123601,-0.1263954192,-0.3179320693,0.1776315123,0.2475596368,0.0826786757,0.1629334539,-0.0125130126,-0.1205988526,0.2289250493,0.1614982784,-0.2385055423,0.0332439505,0.3079956174,-0.2352716625,-0.1989794374,0.0437851474,-0.0636349991,-0.453760922,-0.0903050303,0.5740747452,-0.3744279444,0.0108039975,-0.0825747699,-0.0180426743,-0.3596270084,0.6653097868,0.2563318312,0.6802735925,-0.2463729829,0.0150946248,-0.3281954229,0.3073232472,-0.5743768811,0.1913206279,0.1857649088,0.4029053748,-0.0787028745,0.0760385543,0.1396751404,0.2906526029,-0.0944431722,0.5938670039,-0.2642401159,0.0183152631,0.013586428,0.0321019217,-0.0200353526,-0.4920800626,0.273561269,-0.4048537016,-0.0611601658,0.0724635422,0.0960552841,0.3531217575,0.3399708271,0.5330666304,-0.0138518726,0.1522526294,0.097739324,-0.0338156261,-0.4907117784,-0.0232549459,-0.2535423636,0.1672681421,0.1176789701,0.355897218,-0.3410948217,0.3061569035,-0.045797579,-0.1281310022,-0.0186487529,0.2860833406,-0.0882527009,0.2683669031,-0.5508411527,0.0947744474,0.0382199995,-0.1866860986,0.0787468031,0.0134666031,-0.215300709,-0.5623304844,0.6698253155,-0.3410619497,-0.5061424971,-0.1133759767,0.0440777987,0.229251951,-0.1240592971,-0.2394167185,0.0347790867,0.351703465,0.162270844,-0.2105356753,0.1108048409,-0.1335238516,-0.0080626942,-0.1948655993,0.266715765,-0.0892977715,-0.2301196307,-0.612136662,-0.6397601962]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Sorry it doesn't work. Will let you know once I fixed it","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":12,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nSorry it doesn't work. Will let you know once I fixed it","embeddings":[-0.1042423174,0.1494897753,-0.0652579218,0.2138959467,0.3500129879,-0.0066353721,0.6622277498,0.0979357362,-0.2908475697,0.2196952552,0.2490588278,0.287930727,-0.1786040217,-0.0591259561,0.2324267328,-0.4135220349,0.1507681012,0.5165317655,0.0655505732,0.0367594995,-0.0783932656,-0.1241633743,-0.2830764949,0.2892333865,-0.4049880803,-0.0120881228,-0.2528217435,-0.2124157995,-0.2097512037,-0.4955278933,0.2838594615,0.0099740522,0.1375893503,0.212060377,-0.0001218631,0.0493992493,0.4537374079,0.0437113456,-0.2346890569,-0.1213211939,0.0618161187,-0.4057792425,0.2430216223,-0.3820168078,-0.0682533607,-0.7827058434,-0.0319672413,0.0351395309,0.0884595439,0.2648015022,0.1328193694,0.1410133392,0.3335062265,-0.1731497943,0.4526646733,0.3020344675,-0.2434293628,-0.054144837,0.2915045023,0.0918472409,0.0409064293,-0.191209957,-0.1024708748,0.1941273212,0.3450611234,0.3153828084,0.3589240909,-0.2414564937,-0.0579791069,0.3783745468,0.5469770432,-0.1866198331,-0.1442067921,-0.0628271922,0.091143772,-0.1086797789,-0.0690670088,0.2505178452,-0.257725358,-0.0385440663,-0.2309804261,0.158730641,-0.4949021637,0.3876017332,0.2566405237,0.0851879865,-0.0207222365,0.2255711257,-0.0897867978,-0.2255309969,0.5516836643,-0.3473108113,0.0486570336,0.2318571955,0.0827329308,0.1066797897,-0.166638732,-0.1987201273,0.4221950173,-0.023132842,0.173382625,-0.0773789436,0.3187099397,0.1046820134,-0.2318506688,0.2378740162,-0.0257181711,-0.0825930014,-0.0961566269,0.1280433685,0.0368795916,0.049977161,-0.2147968113,-0.2905360758,0.4110459089,-0.0488075949,0.2666971982,0.2427157313,0.0843957141,-0.003783779,-0.5956351161,-0.1771803051,-0.029583957,0.2699269652,0.069362767,-0.1498063654,-0.1373371035,0.4213271141,0.0225136857,-0.0932965875,-0.216162324,0.021413669,-0.1818009168,-0.021878114,0.0809443444,0.1148112193,0.1569646746,0.0084600607,0.0308328681,0.0427750684,0.2255206704,0.0210629087,0.1202523112,0.1228224635,-0.1525766551,-0.0533961877,-0.026844494,-0.1625831872,-0.1855115294,0.3810545504,-0.1145145819,-0.2496401966,-0.4235814512,0.0905585662,-0.2301340252,0.0767316595,-0.0620885603,0.0555699654,0.2552243769,0.0564651527,0.0185704194,-0.0203765649,0.2212602049,-0.413069278,0.1647655964,-0.0048194989,-0.3202501237,0.0171205197,-0.0664875582,-0.0698628351,0.0029893664,-0.0578231327,-0.2618609965,0.2961896062,-0.2571356893,0.3081174493,0.67667979,0.2144323885,-0.057288304,0.0318613686,0.04511749,-0.2321721166,0.2360155135,0.0920905992,-0.1430412382,-0.0031376283,-0.3496956229,0.2302065045,-0.1230952293,-0.2149841636,-0.0290801618,-0.3707619309,0.4673070312,0.1280809343,0.0726557299,0.0192497112,-0.0305020474,-0.4463748038,0.112353161,-0.4989665151,0.226630643,0.1264299601,0.3409397304,-0.0951159522,0.1041450426,-0.2370554358,-0.6823999882,0.126211375,-0.3840541244,0.0433072411,-0.0819167346,-0.0213620421,-0.2010470331,0.1590017825,-0.2332262099,0.0959815234,0.0389826931,0.0950985625,0.24640809,-0.1624483466,-0.2224087864,-0.3990084827,0.0119844899,-0.0375619568,-0.4042716622,0.6548698545,-0.0762032866,-0.3788359165,-0.3228643835,0.1265250593,0.1178471446,-0.1121335179,-0.0124491723,0.1694529802,-0.0795501322,-0.0992590934,-0.4899818897,0.0571678765,0.0318024606,-0.5039455891,0.2374043763,0.168480888,0.006601661,-0.2473383844,0.1107012182,0.393543601,0.080084376,0.532846272,-0.1658350229,0.0232006591,0.0140833305,0.1508431882,-0.0177235268,-0.1430934221,-0.0480171852,0.2181899399,0.2341907173,-0.0728161484,-0.0611552671,-0.0160615221,0.2888770103,-0.090434812,0.0287534241,0.0674503297,-0.218656987,0.0204512458,0.1498185396,-0.2979003191,0.14235048,0.1564988494,-0.1343573183,0.0762708634,-0.2636730671,-0.1826357394,0.1995032579,0.0816728696,0.5558643937,0.1662814617,0.2331930697,-0.2625217736,-0.2054173797,-0.1770325005,0.088491112,0.2168957442,-0.2369953841,0.0544642247,-0.1243369281,-0.0115495529,-0.0616944917,-0.6133178473,0.0459784344,0.0210774224,-0.1535223722,0.0125600388,-0.0554758944,0.2956106961,0.1760962009,-0.2092721313,0.2567285299,-0.0724273697,-0.2502189279,-0.0418419652,-0.1074557155,-0.0008581424,0.175860703,0.1160600334,-0.092841953,0.0956529379,0.0420361981,-0.4288928807,-0.3519209623,0.0687781274,-0.086468339,0.3462564051,0.5517397523,0.0186016038,0.0608366132,-0.3329231441,0.4385393858,0.0024643801,-0.2445051521,0.3374000192,-0.2827532291,0.0700658113,-0.0291785896,-0.1097240075,-0.059399534,-0.252112031,0.1481532604,0.4817167819,0.2209787667,0.1858062148,0.4675873816,0.3232325613,-0.0590015352,0.0429519229,0.2400598526,0.2605691552,0.4415315688,-0.1270758808,-0.2261185944,-0.0606506094,-0.8689013124,0.101275906,0.2098588943,0.0377010554,-0.214000091,-0.1395468712,0.3403987586,-0.0759417936,0.1568249613,0.317420274,0.1196548119,0.0463266186,-0.1227449924,-0.0593924597,0.4407801032,0.0991576537,-0.0067234254,0.0271843988,0.3386137486,-0.2569153905,0.5897928476,0.1842786968,-0.3790029585,0.2258927971,-0.2735379338,-0.0539824925,-0.2137079388,0.0312220901,-0.1858869642,0.1825532019,-0.075505361,-0.0438391268,-0.1115397811,-0.2608655393,0.1191682369,0.1855006218,-0.1237858608,0.0230836999,0.0038079817,0.0078341644,0.195575431,-0.1175934598,0.0584966987,-0.2927107215,-0.1150560901,-0.1988292038,-0.2486795187,0.2094255835,0.0133989593,-0.2967778444,0.0255552828,0.0710445195,0.1681192964,0.4809155762,-0.1024649888,0.2408785671,-0.0104085477,0.1579610109,-0.1295667589,0.5917640328,-0.188012287,-0.2083677202,0.5140587687,0.2346076071,-0.4516089261,-0.1905746013,-0.1291822195,0.1535137743,-0.3279176652,0.1928777099,-0.3795130551,-0.0259410087,0.380841732,0.3698164523,-0.0365386531,-0.192098394,-0.1639513969,-0.2755339742,-0.1130756065,-0.2581480742,-0.1281175613,0.3883526325,0.1119009629,-0.4445601404,-0.3629311323,-0.0200014524,-0.212393865,-0.0041328655,0.4714739025,-0.2344204783,0.0497542992,-0.349037379,-0.0121418545,0.230843097,0.095860742,0.1503242552,-0.0297546629,0.0940816179,-0.1162950918,0.4623720348,0.0451872945,-0.1239412129,0.1041290164,0.1243943051,0.1493957639,0.0000177394,-0.0880213156,0.4232066274,-0.3177558482,-0.1677952111,-0.3313556015,0.6141457558,-0.0153242955,0.2298217416,0.2243813574,0.3999930322,-0.2052365243,0.238558948,0.553630352,0.8731020093,0.0056031197,0.1328386068,0.7003201246,0.0406143479,0.2746665478,0.3372698724,0.1013573483,-0.5313410759,0.0277788024,-0.0891580433,-0.4420184791,0.2304415405,0.2557449043,-0.2885767519,0.0125594819,-0.2196217775,0.4555341899,0.1282096803,-0.0397424847,-0.1500552595,-0.1191975474,-0.1185309812,0.0486553162,-0.0274226144,-0.3595374227,-0.1401020139,-0.2395864874,-0.1219776645,-0.0516836196,-0.0289813001,0.0170698222,-0.360280931,0.0061849621,0.4852213264,-0.5853769779,0.0548268259,0.18109712,0.0848624781,-0.2175229937,-0.1298929006,-0.0863509923,0.215733245,-0.2925462127,-0.1897024959,-0.258700043,0.2475111783,0.0101474067,-0.110336937,0.0727997273,-0.1061458066,-0.0832086205,-0.0676517859,-0.0352056734,0.2163004726,-0.2449264228,-0.1741998494,-0.1373084486,-0.1462848485,-0.1316023767,0.0266893897,0.1466733515,-0.1921153665,-0.0804582313,0.1464282423,0.0436457358,-0.0060348576,-0.0160652157,0.5854399204,0.2642018497,0.4217185676,0.2241900712,-0.0255763512,-0.1950495541,0.0912097543,0.5432104468,-0.5992942452,0.2000695467,0.1801600605,-0.1144539416,-0.0913233832,0.5799960494,-0.1164459884,0.3449296653,-0.1186771691,0.0247704666,-0.4236699343,0.2134523541,0.1717179567,0.3398872316,-0.2058713585,0.0113266474,0.0328602009,-0.0793184415,-0.1966310143,0.0599478371,0.0093688099,0.0012848572,-0.5196859241,-0.0755675584,-0.0084685925,0.1311113089,0.0853559971,-0.0001029967,-0.3869801462,-0.1297311634,-0.0636149347,0.181463182,0.1371341944,-0.1112714037,0.1298023015,-0.053581275,-0.3181972802,-0.1872293502,-0.0678441599,-0.1058214158,-0.1602517962,-0.0775573403,0.3789927959,-0.030997185,-0.4346858859,-0.102814436,-0.2543013692,0.2477870136,0.0533856973,0.0123101249,0.0812722743,-0.063896209,-0.1222940683,0.2027356923,0.063475281,-0.2295316458,0.3533005714,-0.4551365376,-0.2800445855,-0.2566644549,0.3914858699,0.2121289074,-0.2006789744,0.1433118582,0.1633383483,0.12417496,-0.3156222999,0.0804652497,0.1883450747,-0.0575197674,0.1459337473,0.3051912487,0.1101395488,-0.0369439088,-0.0827536359,0.0789490044,-0.0936325341,0.017328525,0.0675049573,0.616733551,-0.0927979201,0.067754969,0.4268684983,0.076080136,0.2377434671,-0.0730623454,-0.130031243,0.3514395654,0.0867541581,-0.1169045419,0.2104790807,0.0273880102,0.0031120284,0.196421966,0.5162181854,-0.1697843969,0.0648010448,0.5700086355,0.2077957988,0.0048532505,-0.3459285796,-0.0861323476,-0.0347062647,0.0019954147,-0.1697525382,-0.0931589529,-0.2346098572,-0.3670601845,-0.1256728619,-0.3793820143,0.1523205936,-0.081701532,0.1555516422,-0.303660363,0.7454209328,0.2429089844,-0.009942621,-0.1145155877,0.4819870889,0.3174130321,-0.0816397667,-0.2133921087,0.4112561643,0.3334978521,0.3796687126,-0.301527679,-0.0742741898,-0.0924416035,-0.2509911358,0.0770320594,0.1086592972,0.2241354436,0.1007253677,0.2217304409,0.046628803,-0.1051282659,0.2414906174,0.2547978759,0.0310281646,-0.1877031922,0.3347968757,-0.2769152224,-0.187775448,0.020956872,-0.0649400651,-0.4437890947,-0.0413559303,0.3066322505,-0.1131286472,0.0182381142,-0.0650386959,0.0315653645,-0.279425174,0.423778832,0.0691494718,0.6451105475,-0.2346090674,-0.1121217906,-0.3026794493,0.3413717747,-0.3799895346,0.0724562407,0.2692596912,0.3026476502,-0.1315508038,0.106810227,0.1322534382,0.2624329031,0.0413084365,0.381860435,-0.355368346,-0.0862121135,0.2618177235,0.0571657121,-0.0836311653,-0.3268481195,0.1574676037,-0.3711979985,-0.023712825,0.1848099083,0.1702366471,0.1748687923,0.2106842548,0.4791782796,-0.0213982798,0.1326126605,0.1599947661,0.0215879586,-0.2195277214,-0.18659468,-0.1516507119,0.2846142054,0.1083976328,0.4165272117,-0.188606903,0.290880233,-0.1661307216,-0.1229582429,0.0915653631,0.2103203833,-0.1864602864,0.4493142664,-0.5623297095,-0.010725189,0.0213229898,-0.2507137656,0.0732515156,0.1180432588,-0.1379968673,-0.4534033835,0.7265468836,-0.2294420898,-0.4847109616,0.0746463984,-0.0158723071,0.3869391978,-0.1429222226,-0.2762978077,-0.0327704661,0.5050794482,-0.0003569242,0.0019933037,0.092422165,0.0267879143,0.1032109484,-0.1370240301,0.1586841792,-0.042845279,-0.0944519415,-0.4036180675,-0.554736197]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Hi @lhoestq , any update on dynamic sized arrays?\r\n(`Array3D(shape=(None, 137, 2), dtype=\"float32\")`)","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":13,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nHi @lhoestq , any update on dynamic sized arrays?\r\n(`Array3D(shape=(None, 137, 2), dtype=\"float32\")`)","embeddings":[-0.1287701279,0.0186553523,-0.1019617021,0.1574513912,0.3658857048,-0.1406299919,0.6375443935,0.0635337606,-0.4259898961,0.2961487174,0.267958492,0.0535992682,-0.2795879841,0.1703053266,0.3127899766,-0.345171392,0.1452298909,0.4214511514,0.0743447468,0.1137337387,-0.1247543022,-0.114022091,-0.3191619813,0.1045017764,-0.1188817471,-0.0341387056,-0.3823333383,-0.2468222082,-0.1176284328,-0.4203189909,0.3294174075,0.0966103524,0.1098231301,0.1425026804,-0.0001261372,0.0113340542,0.499147445,-0.044858899,-0.2167017311,0.0826310813,0.2181711495,-0.5079609156,0.170320645,-0.2754124999,0.079237543,-0.8136113882,-0.0832489878,-0.1218054593,-0.0201679952,0.0094954241,0.1033869237,0.1089051664,0.4258924127,-0.0637480393,0.3811818361,0.4665982127,-0.3054684401,-0.0382352322,0.3515105546,0.035251122,0.0713145137,-0.2429521829,-0.0806690156,0.208883211,0.3401163518,0.1136933938,0.2787699699,-0.2211209983,0.0379115157,0.3831120431,0.6318704486,-0.2174209058,-0.2096032053,-0.2581042349,0.0336498879,-0.0039631417,-0.1974021345,0.3042964935,-0.2740191817,-0.1150243208,-0.414334625,0.1051476225,-0.4532974064,0.4496032894,0.1584250033,0.0439995825,-0.0120462757,0.1349015087,-0.0755653083,-0.4344663918,0.6348824501,-0.124558948,-0.0460127108,0.1963317543,0.2353021502,0.01650347,-0.0907946154,-0.2533193827,0.3737744093,-0.1791183949,0.2409239262,-0.0866086632,0.3398280442,0.0598953553,-0.1902873665,0.3192541301,-0.1290430129,-0.1568002254,-0.1635065973,0.0848586634,0.1824169606,0.0722908154,-0.2106084079,-0.2262529284,0.2396393567,-0.1170315295,0.2018922269,0.3245718777,0.1130031124,-0.0331543982,-0.6734422445,-0.1144786999,-0.0774619952,0.1960643679,0.1078790426,-0.0189946201,-0.0955628678,0.3130709529,0.0777236298,-0.0504442528,-0.1298517734,-0.0393960774,-0.1109872162,-0.0835304186,0.1050193161,0.3919271231,0.0330290534,-0.182032764,0.241908133,0.0535032973,0.0085625378,0.0783467665,0.2236736715,0.1762494147,-0.2363653779,-0.1998924762,-0.2285877019,0.0200432613,-0.1820211411,0.5171078444,-0.0322762728,-0.2744094431,-0.4156139791,0.0678007156,-0.2482108921,-0.040047463,-0.0023681389,0.1804259121,0.2630549371,-0.0753227025,0.0672108606,-0.1447685361,0.1246533021,-0.396823734,0.1105346382,-0.0284290425,-0.488573879,0.0781656206,-0.0599838085,0.128510952,-0.0718861595,-0.0683876351,-0.2366809249,0.1923705637,-0.1724205166,0.0216629822,0.545814991,0.1559459716,-0.0872531757,0.1066133454,0.1558508426,-0.2858664989,0.2646385729,-0.009224561,-0.1528193206,-0.0288699456,-0.4380586445,0.3427395821,-0.0975316092,-0.1724283695,-0.0332217552,-0.6252562404,0.3917537332,0.1115146726,0.227333948,-0.0546956547,0.1469982415,-0.3311154246,0.1880865693,-0.3731940985,0.2344378978,0.0858219415,0.3480222225,-0.1919447929,0.1989389509,-0.2129873335,-0.790011704,0.1432184428,-0.4485105872,-0.1169092283,0.0234397873,-0.0037462919,-0.1047214195,0.1310029626,-0.0486833602,0.0783058032,0.0053050206,0.0439526029,0.3169800341,-0.1948535293,-0.1058025137,-0.4455670714,0.0986746997,-0.1220757291,-0.2726283967,0.5380514264,0.1132996827,-0.3931030631,-0.247272104,0.2332564145,-0.0461429991,-0.1039488614,-0.1179587841,0.1711436808,-0.0515180193,-0.0074724541,-0.5071985722,0.1577339321,0.0106719052,-0.1677958071,0.4845966101,0.0675657466,-0.0716510415,-0.1920900643,0.2350781858,0.5024756789,-0.0487351939,0.4012362063,-0.192133829,-0.0936560631,-0.0437685139,0.2133524865,0.0982919931,-0.2562173605,-0.1955235898,0.3372577727,0.3485880196,-0.1196103767,-0.100035578,0.0613835156,0.2615087926,-0.1062096953,0.0172848478,0.1300597936,-0.1922769099,0.0251976419,0.1317035258,0.0028584832,0.1393509209,0.2016618252,-0.1611419767,0.0199680422,-0.0528016351,-0.1986819059,0.2649274766,0.1548123956,0.5386959314,0.1317648739,0.4557793736,-0.2649108171,-0.2715694904,-0.1157159433,0.0837133676,0.0607895963,-0.1490996927,0.0811008662,-0.0553978421,-0.1221789196,-0.0849398226,-0.7820144296,-0.1587238461,0.0170831308,-0.0952348858,-0.0178292021,-0.0510668643,0.163937822,0.2144906074,-0.0023880645,0.2651560903,-0.0611699931,-0.2339700162,0.0814073607,-0.0731025487,0.0151374703,0.1560858637,-0.033525221,-0.0449494533,0.1925107241,0.2397195399,-0.2804154158,-0.3629063368,-0.0058647781,-0.0044745756,0.2989165187,0.486987263,-0.1354885995,0.0915219858,-0.1415153444,0.3228246272,0.0728689134,-0.1954722703,0.173753649,-0.1644379348,0.1834281534,-0.1391686052,0.1529304832,-0.1441302598,-0.2166685164,0.2321152687,0.491771251,0.1886978298,0.2038684785,0.5268548727,0.2954704165,-0.0545371883,-0.0383926593,0.2793904841,0.2340346873,0.3586117923,-0.0659145936,-0.1700883657,-0.1986999065,-0.7936996222,0.1113068759,0.2543061078,-0.0180235449,-0.2454684079,-0.16548796,0.2038526386,-0.0362840518,0.1424084008,0.4113785028,0.0521760173,0.0948181748,-0.1475494355,-0.0860408321,0.3395673335,0.2388890237,-0.0263151098,0.0766285732,0.3914115727,-0.1869532317,0.6602121592,-0.1230893657,-0.4135858715,0.1955326498,-0.243300274,0.010741883,-0.2050338238,0.0671157986,-0.0919282287,0.1320828199,-0.235784933,-0.0601574108,-0.0838046521,-0.3523299694,0.1602477878,0.208177045,-0.024454318,0.1218655482,-0.0186867993,-0.0835763812,0.2097479999,-0.1542604715,0.1045625806,-0.384780705,-0.2735451758,-0.3125109375,-0.3800286651,0.1861594319,-0.010942461,-0.0020402381,0.0594504476,0.1103940457,0.2551222742,0.5049304962,0.119706206,0.2276295573,-0.1421836913,0.1579415202,-0.0854485482,0.5165900588,-0.0927556381,-0.2316025496,0.4981088638,0.1386691034,-0.4042032063,-0.3039893508,-0.1389647573,-0.0136534786,-0.2055276334,0.0884961262,-0.2167532593,0.0057071028,0.2920795679,0.3042383194,-0.0655794963,-0.2040680349,-0.2293923795,-0.0953162163,-0.1556747556,-0.2176050246,-0.1179289594,0.3328302205,0.1743379384,-0.4158086181,-0.3981755078,-0.1684646159,-0.2232286036,0.0062804157,0.3584988415,-0.1213186756,0.0389804468,-0.2902752757,0.0236667451,0.2323712111,-0.0940731168,0.2246317565,0.1605124474,0.0792133436,0.0586772561,0.5390751958,0.1791692376,-0.0636906475,0.0911789685,0.1145909876,0.1912495196,-0.1090987399,-0.1707193702,0.3839963078,-0.4455029666,-0.0737368912,-0.35773772,0.574540019,-0.0706913918,0.225395754,0.2694799006,0.3877087831,-0.1973439455,0.2579213083,0.5493543148,0.7992324829,-0.0323528722,0.1136971265,0.6316646338,-0.0438139141,-0.0665456355,0.4273270965,0.139669463,-0.5126766562,0.2507558763,-0.0617326945,-0.4036190212,0.2303837091,0.1240720898,-0.1631240547,-0.1042168066,-0.0908374116,0.3602855504,0.0629372001,-0.0416921228,-0.168328017,-0.142787531,-0.1191569567,0.0092220698,-0.0208421219,-0.3586651683,-0.1101725921,-0.2872814536,0.014968358,-0.040598426,-0.0483116768,-0.0320153497,-0.363524586,0.0683269948,0.5614314079,-0.622392714,0.0514132604,0.1855808645,-0.0240105037,-0.1401941925,-0.2260898203,-0.0906763002,0.2950280905,-0.3198117912,-0.1384154707,-0.2453231663,0.2284658551,0.0743334666,-0.1106830761,0.0482379645,-0.0114254849,0.0022454499,-0.168334797,-0.0524242111,0.3001860082,-0.3319578171,-0.1096336022,-0.2356263399,-0.0400966257,-0.028643528,-0.0089133596,0.145584017,0.0109160272,-0.1406305432,0.2478250265,0.1006921902,-0.0863928944,0.0758142471,0.5533052087,0.1591317654,0.3127593994,0.1867321581,-0.0396161303,-0.1320022643,0.2988877296,0.2448761612,-0.5574204922,0.2118809521,0.2210920304,-0.1626013517,-0.0106305927,0.4478331506,-0.078882508,0.2852898538,-0.1779521108,0.080390811,-0.3457203507,0.2551289201,0.0628827363,0.3550233841,-0.1593718529,-0.1278428435,0.0124461921,0.0071348739,-0.1515844613,0.1123264804,0.2829765081,-0.1047691405,-0.6746771932,-0.1411705017,-0.0314464755,0.1920497715,0.0140584521,-0.0794576034,-0.4690225124,-0.0577677377,-0.0251487214,0.1825230271,0.0891747773,0.0689480677,-0.0107072024,-0.1233614683,-0.3302122951,-0.2078675628,-0.0145939272,-0.1347303987,-0.2184298635,-0.1503964812,0.4366420805,-0.0796623752,-0.2481102645,-0.0400624424,-0.1567326337,0.2132183015,0.0837093666,0.0931207687,0.0855255127,-0.0176527947,-0.0557464398,0.2178376019,0.0235567354,-0.3295560777,0.3664884567,-0.3945344985,-0.2868810594,-0.3271918595,0.4587500393,0.1658554673,-0.1550716013,0.0321363471,0.1273084134,0.0846865773,-0.2686041892,-0.0150286127,0.1085989773,0.1369284242,0.0916032121,0.4959739745,0.2843778729,-0.1441245675,-0.1320724487,0.1613452286,-0.1172083318,0.1023047194,-0.0705155954,0.6021248698,-0.1069101095,0.0186881032,0.4834986925,0.0754333809,0.3712845147,0.0861627683,-0.0465874933,0.3896912634,0.2483340949,-0.0800695419,0.1863161325,-0.0364515968,0.0941547602,0.1051780432,0.5284581184,-0.1010380462,0.0086251413,0.5886664987,0.2558294535,0.1105617359,-0.491414398,-0.1963398159,-0.0464426689,-0.040739052,0.1087075695,-0.1596527547,-0.1805653125,-0.2612041831,-0.1394476444,-0.4332994819,0.2028988004,-0.1698097885,0.2046634257,-0.1548360586,0.7041541934,0.1438018829,0.0387740135,-0.2392888814,0.4324243963,0.2403900921,-0.0535907969,-0.2465344816,0.4275678992,0.3783705235,0.354503423,-0.372375071,-0.0949476585,-0.0179411843,-0.1940891445,0.0467228033,0.1657107025,0.1697601378,0.2692265511,0.1810587049,0.0399799719,-0.0701337904,0.1623847932,0.3432019949,0.028402403,-0.293312043,0.3041587472,-0.3685290217,-0.040964175,-0.0299755558,0.0028812797,-0.4895504117,-0.0878656283,0.3479340672,-0.1739916652,-0.0830126926,0.0305689368,-0.0157127436,-0.2540929019,0.5254688263,0.1615406871,0.4966217279,0.0015724144,-0.1186037734,-0.1873991787,0.3363170624,-0.3246521056,0.1561742574,0.2239054888,0.2720383406,-0.0695528612,0.0086227795,0.1154975146,0.3065669239,-0.105939649,0.3609136939,-0.3164866269,-0.0066881324,0.1750170887,-0.0079103736,-0.0573026538,-0.2872265279,0.1292375773,-0.5795044303,-0.0654704049,0.139438197,0.1809034348,0.1044568345,0.3043366969,0.366669625,-0.0992176011,0.0431436561,0.2242401093,-0.0624863282,-0.3074066639,-0.2504257262,0.0705013499,0.3298051059,0.0342654027,0.3621095419,-0.2851155102,0.2744007707,0.0472367406,-0.1316520721,0.0097281067,0.3762563169,-0.0911201239,0.2945158482,-0.5456206799,0.0093006091,0.1046643406,-0.2259228975,0.0919336379,0.0982557386,-0.0797771513,-0.3693528473,0.651860714,-0.238523066,-0.4645755887,0.1019965187,-0.0920029208,0.2271754295,-0.142829299,-0.3381186128,0.0890455246,0.4280821085,0.0713156387,0.1513066292,0.0765411481,-0.0418128222,-0.0334986709,-0.1498270482,0.1905301958,-0.0712510943,-0.0916746929,-0.535823822,-0.4275725782]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Not yet, I've been pretty busy with the dataset sprint lately but this is something that's been asked several times already. So I'll definitely work on this as soon as I'm done with the sprint and with the RAM issue you reported.","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":42,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nNot yet, I've been pretty busy with the dataset sprint lately but this is something that's been asked several times already. So I'll definitely work on this as soon as I'm done with the sprint and with the RAM issue you reported.","embeddings":[-0.1500287801,0.2067138553,-0.0467009619,0.2796334326,0.4458692074,-0.0149022257,0.6093963981,0.1298004836,-0.3220183551,0.1937067658,0.2713231444,0.3150215149,-0.1662629992,-0.1340918243,0.284039706,-0.3528666794,0.19875063,0.4137451351,-0.026127778,0.0539427474,-0.0769715011,-0.1808713377,-0.2342819422,0.1500301063,-0.4875097275,0.0162202325,-0.2609251738,-0.0898402855,-0.2083282918,-0.6032204032,0.254519701,0.0392498374,0.0257065631,0.2163897157,-0.0001187129,0.0717893392,0.5233740211,0.100117974,-0.3757022321,-0.0207123328,0.1298973262,-0.4313429296,0.292747587,-0.3118649125,-0.0444274358,-0.6730905175,-0.0052330205,-0.0925311223,0.0851977468,0.2867689133,0.1531415731,0.1496409923,0.2610460818,-0.2167924196,0.5045078993,0.3086963296,-0.2349268496,-0.0965385884,0.2660935819,0.0270716902,-0.086784184,-0.141119346,-0.0499659292,0.2337395102,0.3484751582,0.3037255406,0.2073807418,-0.2104097009,0.0256282426,0.3553407788,0.5374174118,-0.2118532956,-0.1425718665,-0.1533841938,0.0074660117,-0.135996148,-0.0367606618,0.2109224051,-0.3065979481,-0.0860812441,-0.272334069,0.1510215849,-0.3894477487,0.2965713441,0.1956054121,0.0183366071,0.0433375016,0.2343143076,-0.0197654553,-0.1694684625,0.520832777,-0.2911010981,-0.0225759018,0.1274242401,0.0479965396,0.159519285,-0.0755240172,-0.309481889,0.3864966333,-0.0554514565,0.1326502264,0.0440819077,0.3791761994,0.0212110206,-0.155546248,0.2336513996,-0.1281942576,-0.0899409801,0.0055330736,0.0701004863,0.0022220521,0.047043819,-0.2628852427,-0.2715009749,0.4232479632,-0.0934406668,0.2694977522,0.1873334795,0.027152203,0.0392056406,-0.5091846585,-0.1334914267,-0.0782195926,0.2353959531,0.0806331336,-0.0823203847,-0.0616264939,0.3054972291,-0.117213659,-0.1475103796,-0.2355750948,0.016607523,-0.2514528334,-0.0253880639,0.1708405167,0.2288855016,0.1449002475,-0.0470355786,0.0650947466,0.1135158092,0.2895406187,-0.0105913784,0.1357876062,0.1952750534,-0.2065051794,0.0302209035,-0.0129400818,-0.1132825986,-0.1792579889,0.383002609,0.0182225816,-0.3663841188,-0.287332505,0.0946431831,-0.2151788771,0.0613822043,-0.0709390715,0.1718257815,0.2112199366,0.1510830671,0.0608303361,-0.0333512835,0.0417920388,-0.4401707947,0.2193191946,0.1395140141,-0.4058801234,0.0554630384,-0.16317752,-0.0664396361,-0.0156456679,-0.024564052,-0.2742761672,0.2772711217,-0.1955377012,0.2298749089,0.5872292519,0.1661988199,-0.2438392788,-0.036898382,0.003327277,-0.1664946824,0.2262385786,0.1600516289,-0.1171486527,0.0681513846,-0.1770196408,0.3029491007,-0.1542997658,-0.1596271843,-0.1055895463,-0.3996223211,0.3469384313,0.2101228684,0.1483471245,-0.1116844714,-0.1029537246,-0.3682799041,0.1174836382,-0.452626735,0.2587610781,0.131894663,0.3844100237,-0.159466669,0.0933059677,-0.2613682449,-0.6854189634,0.2985851765,-0.2492964268,0.1730052531,-0.0901348516,0.0227009468,-0.0997130573,0.1692358255,-0.2032765299,-0.0019785985,0.0679667667,0.0506005473,0.2336383909,-0.1973650455,-0.1905950308,-0.3107894957,-0.0612241738,0.0008915419,-0.3421287537,0.5485239029,-0.0172521248,-0.3967372775,-0.3598187268,0.0743663013,-0.0085200118,-0.045928441,-0.0238599069,0.192988649,-0.0130882757,-0.1298053563,-0.4362171888,0.1046651453,0.0309572704,-0.3927151859,0.2777435482,0.2070194632,-0.0343700536,-0.2954361439,0.040611852,0.3032567203,0.0700180531,0.5143806338,-0.0675568953,0.0170665272,0.0247860625,0.2132641077,-0.0073072454,-0.1172537357,-0.0014551623,0.2322852463,0.2657836676,0.0138674714,-0.0900225639,0.0586162247,0.3243603408,-0.0800146908,0.0507921055,0.0836850479,-0.2945884168,-0.0193244703,0.2128085494,-0.2105057836,0.1979473382,0.2074630111,-0.0482013412,0.1102696732,-0.2756565809,-0.148309499,0.2355589122,0.1429792047,0.5560609102,0.0222477783,0.2309238911,-0.2626541853,-0.2959458828,-0.1668593436,0.1668504477,0.2569403648,-0.3018983305,-0.0585598759,-0.1459394544,-0.0343174115,-0.0355405957,-0.4733985066,0.0393411368,0.0014645186,-0.1478918493,0.0221957322,-0.1176974103,0.2584047318,0.0480031744,-0.0906381011,0.2260757089,-0.0672281161,-0.1370675266,-0.1112420186,-0.1663928926,0.0230181832,0.3057001531,0.0607741438,-0.0255741272,0.1304396689,0.0086157033,-0.3728959858,-0.3643664122,0.0700161085,-0.0963850766,0.3734642565,0.4846831858,0.0340089314,0.0707842708,-0.2421431392,0.4771783352,-0.0035104363,-0.2795153856,0.3224967718,-0.2145095915,0.0431248248,-0.0884386525,-0.2002503425,0.0433548018,-0.338442862,0.2144085169,0.3770569861,0.2181912661,0.2402545661,0.4826954901,0.3846856654,-0.0262592789,0.0725125968,0.1978988498,0.2519872487,0.3446090519,-0.0619258769,-0.2083565742,-0.088966094,-0.7822172046,0.0658636242,0.3343043029,-0.132839486,-0.1330103725,-0.3091554344,0.368486762,-0.0566361919,0.1405151188,0.378064692,0.1708713621,-0.0217635259,-0.0804595947,-0.1139997169,0.3861759305,0.0908233076,-0.0778868273,-0.0079130335,0.422504127,-0.1440055519,0.67773664,0.098717466,-0.2711271644,0.2037224323,-0.2157235891,0.0048970757,-0.2505938709,0.0265117045,-0.1217844039,0.0888629109,-0.1523335278,-0.0191508774,-0.0363733061,-0.3188762963,0.1218132079,0.1965554655,-0.0624401867,-0.0397178121,0.1316151619,-0.0308444034,0.1022257954,-0.0918362066,0.1050003469,-0.2875980139,-0.0756792277,-0.190887779,-0.4204811752,0.1240359321,-0.0575267822,-0.3283354938,0.1398129016,-0.0885654837,0.1524424702,0.3744049072,-0.1042877957,0.1758559644,-0.1414589137,0.1569481194,-0.1547943354,0.6634982824,-0.1758404821,-0.2710128427,0.5152444243,0.1407134086,-0.541521132,-0.2129684687,-0.2095394433,0.245808512,-0.2725162208,0.1520115584,-0.4102877975,-0.0327011682,0.3773133457,0.3596919179,-0.032730367,-0.094484821,-0.2216612846,-0.3411521614,-0.0681905895,-0.1154122576,-0.0727886334,0.3152891695,0.123610273,-0.3971572518,-0.3735953569,-0.0400240943,-0.1721286476,-0.0104516558,0.458894968,-0.2380856872,0.1649463028,-0.2397783995,0.0009235816,0.2648705244,0.1921317428,0.117578499,-0.0643796548,0.1052885577,-0.0572617166,0.4494801462,0.259275645,-0.054085508,0.0574306622,0.1639574915,0.1375256926,-0.1249851957,-0.0268663559,0.4133701921,-0.3308499455,-0.1898858994,-0.3715833127,0.6200917959,0.0414904021,0.1514876336,0.3555671573,0.3764851391,-0.2424073517,0.316528827,0.5446491241,0.966724515,-0.1289045662,0.2134443223,0.6495653987,0.0029645807,0.2828020751,0.2540852129,0.2277838141,-0.5002316833,-0.0339990966,-0.0065798215,-0.3927591443,0.2069499195,0.2149785757,-0.1865026802,-0.0012378546,-0.2182621509,0.4358303845,0.1036747843,0.0344063081,-0.2134681791,-0.1178780198,-0.0730039552,0.0895677209,0.0058275359,-0.438196063,-0.1154489219,-0.1017157584,-0.1978418678,-0.0135276075,-0.1254939139,-0.0239034817,-0.349571228,0.0696776882,0.3701092303,-0.5165644288,0.020866219,0.1090424657,0.0572875105,-0.1402059197,-0.096969761,0.0260196235,0.2428457588,-0.1842741072,-0.0130588254,-0.3214933872,0.263061285,0.0328300744,-0.2051189691,0.022953961,-0.1353123635,-0.1692550778,-0.1750240922,-0.0551151037,0.247951299,-0.2530693412,-0.1780581623,-0.0815899298,-0.1463174075,-0.1738775671,0.0666027442,0.1170256734,-0.1410746127,0.0025063904,0.1300409883,0.0235495009,-0.0709221512,0.0205199756,0.5894396901,0.2913070917,0.430231601,0.309469521,-0.0467939563,-0.1781943887,0.0038344599,0.6479730606,-0.541528523,0.2230454981,0.1170604602,-0.2020018399,-0.1218762994,0.4708729088,-0.1145159751,0.2816265821,-0.1862773299,0.1173035055,-0.473161757,0.2905726433,0.2220406681,0.4258405268,-0.2673299909,-0.1688767225,-0.005042498,-0.0191145465,-0.2316870391,0.0789486989,0.027930893,0.0678843409,-0.5748067498,-0.105927065,0.1317717135,0.0811185986,0.1017664447,0.0013396768,-0.388156563,-0.1580252498,-0.0020905635,0.1558316499,0.058741726,-0.17484276,0.2019313723,-0.1383804679,-0.2560036182,-0.2296722233,0.0389318019,-0.0213918593,-0.2182002068,-0.1809265465,0.3834403455,-0.0254353303,-0.3117885292,0.0269560851,-0.2520413399,0.2981004417,0.1187542379,0.0703945011,0.1229051054,-0.0584069341,-0.0751527026,0.1965300292,0.0404650383,-0.2731575668,0.2590331435,-0.4374717772,-0.3664401472,-0.2270938009,0.4581240714,0.3234955966,-0.1110689342,0.0736070871,0.1590511054,0.1698397398,-0.2220702022,0.0490026213,0.2103671134,-0.1193469539,0.1806747913,0.3013379276,0.0832025856,-0.0033884523,-0.1054539829,0.117128484,-0.0229333304,0.0252824035,0.0834056288,0.4652062356,-0.1098438725,-0.0518161096,0.3878831863,-0.0196096189,0.2227765173,0.0079953941,-0.0850417912,0.4147613049,0.0869290084,-0.0517750569,0.1325203776,-0.0293117799,0.1116987318,0.2715786994,0.3833861947,-0.1204078943,-0.0267836172,0.6152066588,0.2223729491,-0.0460368842,-0.4197398722,-0.1864475012,-0.1455493867,-0.0340117402,-0.0827055052,-0.1281300336,-0.2359882295,-0.154085353,-0.1639438868,-0.4509734511,0.2512566745,-0.0805564672,0.069225207,-0.2903128564,0.7235395908,0.182648316,0.0155358417,-0.1364161223,0.4172535837,0.3453892171,-0.1296582669,-0.1810429245,0.4945087135,0.3849455118,0.4238958061,-0.4110917747,-0.0324110314,0.0056650275,-0.2146723568,-0.0559282638,0.0593395606,0.2370568365,0.1880389005,0.1770807058,0.0776376724,-0.1500835121,0.1009813696,0.2514288425,-0.0357965231,-0.2235600501,0.2157305628,-0.2251520455,-0.2205971628,0.0589822419,-0.0766528845,-0.4713689983,-0.143974632,0.4179427922,-0.1683094949,-0.0066465549,-0.0709945932,0.042357441,-0.2739167511,0.486087054,0.0101106558,0.5837244987,-0.2782611549,-0.1044880524,-0.3904719651,0.3186706603,-0.5196862817,0.1071946025,0.2680416107,0.3691579103,-0.075701654,0.1105256304,0.1447155774,0.1774957925,0.0062260064,0.4266397059,-0.3925250173,-0.0103615383,0.25644207,0.0076438976,-0.0866416991,-0.3860512376,0.1612875313,-0.3479377925,-0.0182349104,0.176424697,0.148156032,0.2485917658,0.2258594185,0.5414074659,-0.0579420961,0.1531686336,0.2425011694,-0.1587215662,-0.3644285798,-0.1778422296,-0.1589808166,0.2417852581,0.075546518,0.4335434437,-0.2603600323,0.2352063954,-0.163804099,-0.0019351648,0.0276691876,0.2400591075,-0.1374566406,0.4033572674,-0.6006198525,0.0910600573,-0.0216425825,-0.1658451706,0.0927805752,0.0565389767,-0.2232379615,-0.4527819157,0.6323030591,-0.3164665699,-0.4803242683,0.0255677179,-0.0108190514,0.379121393,-0.1154550686,-0.3446508944,-0.0904824287,0.3934347928,0.0028524392,-0.1030599847,0.1104017496,0.0531178601,0.0731430501,-0.1679363251,0.2442335784,-0.1923168749,-0.1297082752,-0.384249568,-0.5637407303]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Hi @lhoestq,\r\nAny chance you have some updates on the supporting `ArrayXD` as a subtype or support of dynamic sized arrays?\r\n\r\ne.g.:\r\n`datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))`\r\n`Array3D(shape=(None, 137, 2), dtype=\"float32\")`","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":29,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nHi @lhoestq,\r\nAny chance you have some updates on the supporting `ArrayXD` as a subtype or support of dynamic sized arrays?\r\n\r\ne.g.:\r\n`datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))`\r\n`Array3D(shape=(None, 137, 2), dtype=\"float32\")`","embeddings":[-0.2084440738,0.0039850092,-0.0901311412,0.0411780179,0.3527282774,-0.0363005772,0.6182537675,0.0921264216,-0.1956912279,0.1928671002,0.2420260608,0.0475220531,-0.279655844,0.0491814986,0.2426549196,-0.3445935845,0.2176743299,0.4056478441,0.1302125007,0.0160453673,-0.1540563405,-0.1357140988,-0.3179493546,0.2507115304,-0.1650814414,-0.0063748928,-0.2581344247,-0.119082734,-0.0953019857,-0.479681313,0.4352333546,0.1003657132,0.2172332257,0.1735836864,-0.0001256443,-0.0151909478,0.4540738761,0.0053071789,-0.227201283,0.1061744466,0.0516524911,-0.4238549471,0.1683063805,-0.3253711462,0.0930252373,-1.0158853531,0.037417639,-0.1650810093,-0.0047545969,-0.0251165349,0.0971729681,0.11753501,0.574431777,-0.0544573478,0.4038918316,0.3394396007,-0.2964442074,0.0017574386,0.6183679104,0.1803751737,0.1633254141,-0.3020035326,-0.1555918902,0.108475998,0.3668763041,0.090071328,0.2777218521,-0.1963364184,-0.103932485,0.3525259197,0.4946058095,-0.2098657191,-0.2475662529,-0.2931713462,0.0062063825,-0.0968711451,-0.1634802818,0.1315524876,-0.2634223998,-0.0381583236,-0.2493409365,0.0278524123,-0.4791485667,0.3923698962,0.1832090616,0.3168774843,0.0780540109,0.2102219164,-0.1305958778,-0.4713119566,0.5399202108,-0.194673717,-0.0393474698,0.1770695597,-0.0109686926,0.0382324494,-0.1436097324,-0.3914031684,0.3296518028,-0.1347400844,0.2233513296,0.0036931592,0.293533504,0.1309708953,-0.1120029986,0.4439811409,-0.0050440785,-0.2573388517,-0.1000133753,0.0176749211,0.1494630575,0.1185548976,-0.3885904849,-0.0665612966,0.2135720402,-0.0072955526,0.1125636622,0.3052555621,0.0839365646,-0.0825337246,-0.5444916487,-0.1952552646,-0.0342573859,0.1130028889,0.1320813447,0.0644626394,-0.0958098546,0.3358771801,0.0190796219,-0.0274252389,-0.0492887571,-0.0743581057,-0.1696981192,-0.2090104669,-0.0742272735,0.242568031,0.0230555441,-0.2454911023,0.2963340878,0.2589597404,-0.100914754,0.1745902151,0.1355239749,0.0865416005,-0.1246086657,-0.1421930939,-0.1530184895,0.0659832358,-0.2035204768,0.4575885832,0.0349642374,-0.2632192671,-0.519004643,0.0634810776,-0.2189638764,-0.0599216819,-0.0134624243,0.1540388167,0.2465145588,-0.0262611266,0.0705753192,-0.1021020785,0.1038182303,-0.3753688335,0.0550073683,-0.1984353215,-0.4159475565,0.0520624742,0.0362304784,0.0176687427,-0.2496100962,-0.1657966375,-0.1368048042,0.1306780875,-0.1476325989,0.1554267555,0.4196239114,0.0659340173,0.0262140669,0.1470692605,0.1074054241,-0.2000729293,0.2734329104,0.0588099398,0.0231610555,-0.0664216205,-0.3663851917,0.2728359997,-0.1010288522,-0.2770082951,0.0349043347,-0.5597991943,0.3022977412,0.2005216032,0.1087435186,0.0053187301,0.0642158836,-0.3475567997,0.3698147833,-0.4612921476,0.2046328038,-0.045162335,0.3303063214,-0.1753949374,0.1733091474,-0.3176588714,-0.7671214342,0.2427631915,-0.3001486063,-0.0834676474,0.1464914978,0.0340713076,-0.1214431673,0.1729816794,-0.2208021134,0.0893120319,-0.0316909812,-0.0686633363,0.1665673107,-0.2547790408,-0.2116602212,-0.4258862436,0.0152585693,-0.0417573117,-0.5153887868,0.3571023345,0.119301036,-0.3739496768,-0.3116869032,0.2690061033,-0.1038860083,-0.2080880255,-0.1168521494,0.2687589824,-0.108457692,0.0143572027,-0.5090368986,0.1212968975,-0.0007314586,-0.3114989996,0.4452406764,0.1880426556,0.0074210605,-0.3056252897,0.2982736826,0.4102921784,0.1303010136,0.4531944692,-0.1391331255,-0.0143084815,-0.0755283907,0.2140805274,0.0493239611,-0.1115580276,-0.1117410287,0.28411448,0.1176197901,-0.0952014923,-0.1254863739,0.1526382864,0.1783764213,-0.0880279541,0.061760243,0.156608209,-0.1742018312,0.1263123602,0.1600427032,0.0594885349,0.1768950969,0.1391759664,-0.2293741256,0.0531166196,-0.1666702628,-0.1580823064,0.2809893787,0.0568841062,0.4773663878,0.1018739268,0.2846907377,-0.2070886791,-0.1774286032,-0.1415572464,-0.032472983,0.080820322,-0.1407948434,0.0441586338,-0.1467220485,-0.1975245476,0.0022291201,-0.6770448685,0.0133736972,-0.023938952,-0.1055031866,0.0230710264,-0.0805765614,0.2785808146,-0.0787080303,-0.1286106259,0.2623370588,-0.153232947,-0.1476906389,0.0146163749,-0.0939416289,-0.0071486169,0.1849808991,-0.0640511289,-0.1356012374,0.23879309,0.1510271579,-0.1945213377,-0.325247705,-0.0284230523,-0.0677536502,0.2654829323,0.4885601401,-0.1751176864,0.1404495686,-0.3283764124,0.3227990568,-0.1191454157,-0.249133423,0.1754204929,-0.1840374321,0.010525967,-0.102536954,0.0885618776,-0.1960027218,-0.2493997961,0.2090706229,0.4657217264,0.2165712565,0.2988514602,0.456590414,0.2107476443,-0.0091061844,-0.1195418462,0.3286369443,0.2301234156,0.4304123521,-0.0030972138,-0.1493749022,-0.1261469573,-0.8431602716,0.1197349057,0.3775215447,0.1666362584,-0.1006572992,-0.1861110777,0.1386382729,0.0683850944,0.0345074013,0.3450479209,0.0837456211,0.0808112323,-0.208127141,-0.0666603222,0.3870494068,0.2004280537,-0.0055897529,0.2823007107,0.3034777641,-0.1905017644,0.6329155564,-0.0674083084,-0.3457585871,0.2688483298,-0.157796964,-0.1502286792,-0.1161775738,-0.2045220733,-0.0719859451,0.0357283913,-0.2112000734,-0.0722618923,-0.0487272926,-0.3421905339,0.1218046471,0.3561455905,0.047599487,0.1159227788,-0.0661759526,-0.0563893244,0.26703462,-0.1015590653,0.1103118211,-0.3185107708,-0.2099430114,-0.3018078208,-0.3100989759,0.234204933,-0.0515378825,-0.2744898796,-0.034687981,0.0665809736,0.2780244052,0.5870445967,0.145005703,0.1673178673,-0.0304364432,0.1368951499,-0.0556073859,0.3812724054,-0.093761988,-0.2223682404,0.5127313137,0.0989441425,-0.4337401092,-0.2174158543,-0.0203591604,-0.0968287513,-0.2593232691,0.0813826397,-0.2699007392,-0.0149402181,0.3794332743,0.1416913867,-0.0776130632,-0.2296660393,-0.1584009081,-0.0875891447,-0.1614018232,-0.2980894744,-0.098220028,0.3584465981,0.119739674,-0.4149910808,-0.3605048954,-0.1243129149,-0.1632251441,-0.0213893168,0.3761022687,-0.2070841491,0.0470635444,-0.2696406543,-0.0585361943,0.2221458405,-0.0487723164,0.1563634723,0.1389496624,0.0472326577,-0.0450554863,0.5575083494,0.1805845201,-0.1562365144,0.1604666561,0.0501110218,0.1121299043,-0.1579119712,-0.300976634,0.521943748,-0.4097984731,0.0791214332,-0.3341868818,0.5171593428,-0.0643040016,0.2047123164,0.3343685269,0.2912309468,-0.1984198689,0.1640222073,0.6330119371,0.8277635574,0.1401899904,0.1410465837,0.7423071861,-0.0447884761,-0.0904026106,0.5768278241,0.1131712049,-0.4875541925,0.1066653803,-0.2029244751,-0.3054341972,0.2331068367,0.151863873,-0.1239454672,-0.0216851495,-0.0988500416,0.4097416103,0.2667947412,-0.0106266346,-0.2763104737,-0.165661335,-0.1189560741,-0.0065223435,0.0528804176,-0.5251213312,-0.1581431925,-0.3651210964,0.0084713353,-0.0222081933,-0.1836152673,-0.0284262411,-0.225871861,0.045062542,0.4800754488,-0.5878506899,0.1917489767,0.1592961103,-0.015608673,-0.3468287885,-0.1109324023,-0.0694319159,0.1753302366,-0.1501413435,-0.2098661065,-0.3819771111,0.3388650119,0.0652789995,-0.1567866057,-0.0076092244,-0.046829056,-0.1408230066,-0.253927052,-0.0392384045,0.2380220592,-0.2631183267,-0.1170007139,-0.2332717329,-0.0605960563,-0.1205428094,-0.0294410046,0.177678749,0.0005440895,0.0130404215,0.1819693744,0.1543811113,-0.0411549732,-0.0601760447,0.5626733303,0.1886489838,0.3105585277,0.2667617798,0.0261357613,-0.0932110697,0.264655143,0.4370450974,-0.5262246728,0.1717756689,0.1161476895,-0.1427187473,-0.0936782584,0.3504407704,-0.0658153668,0.3271068335,-0.1283333898,-0.0530625135,-0.2728982568,0.3411972523,-0.0420773849,0.3002476692,-0.2621001303,-0.0368269458,0.0563180819,-0.0335124545,-0.1439363211,0.0840521529,0.2423205227,-0.1925822943,-0.4776723981,-0.0777852759,0.0831142142,0.152728036,-0.0344700739,0.005705046,-0.5085963011,-0.0119309891,-0.037454389,0.2035380453,0.1438117623,-0.1473115683,0.1234725714,-0.0211333297,-0.3804351687,-0.1049260125,-0.1208062395,-0.1715066433,-0.1165284738,-0.0932291001,0.3691178858,0.1329875439,-0.2661266923,0.0869141221,-0.1710867733,0.2645001113,0.1452868879,0.0637351796,0.0402694233,-0.1136700436,-0.122701779,0.2589553893,0.0126045411,-0.2605251074,0.5132203698,-0.3607145548,-0.3177291155,-0.246616289,0.6039426327,0.2200224251,-0.10873539,0.0422896147,0.0518625937,0.0542955399,-0.2933723629,0.0083419364,0.362006247,0.1683652699,0.0796842799,0.3961844444,0.3050346076,-0.0819174945,0.0781159475,0.1511453986,-0.0342172645,0.02294112,0.0221414175,0.6143956184,-0.0979317874,0.0784122944,0.4165047109,0.0984421298,0.4063203931,0.2499708086,0.0153698092,0.4456933141,0.4248219728,-0.078811042,0.2008180916,-0.0426151417,0.0157026779,0.1423566341,0.4935704768,-0.0558742508,0.1126008183,0.7638303041,0.1266028583,0.0897480845,-0.2980299294,-0.1383945197,0.0386781581,-0.0417295247,-0.0584022664,-0.0854912177,-0.3334433734,-0.3653235137,-0.1893310547,-0.3462080956,0.1235830113,-0.2630028129,0.1551785469,-0.4477919638,0.6662041545,0.2066176087,0.0486719646,-0.0889627337,0.4410880804,0.3619861603,0.0626945868,-0.1430427283,0.582398653,0.3293176889,0.3508200049,-0.3735731244,-0.1025461406,-0.1375536919,-0.183868587,0.1246770248,0.1995573789,0.0957780033,0.2726633847,0.2463157326,-0.0009986383,-0.0173517838,0.1782152206,0.3353331685,-0.1091607288,-0.2478856444,0.2219826728,-0.3379712105,-0.0672515854,0.092932485,-0.0029861121,-0.4135436714,-0.0170294922,0.277439028,-0.2001987994,-0.0512875617,0.0032046975,-0.0143806413,-0.2667317688,0.5101758838,0.0598900691,0.4724583626,0.0209749918,-0.1303228438,-0.356361568,0.3754887581,-0.2177902162,0.0949427038,0.3324357271,0.1496599764,0.0329762325,-0.029118875,0.2620603442,0.2070948184,-0.0250281598,0.4323697686,-0.3074161708,0.0020208911,0.3492791057,0.0656791776,-0.1466333717,-0.2338234037,0.1958294362,-0.4173709452,-0.1105689704,0.1605216265,0.0218837988,0.1033970937,0.0660815388,0.4318191409,-0.0104291271,0.1300344169,0.2182446122,-0.0676099733,-0.2745562196,-0.2089975178,-0.123242557,0.271892637,0.1555342376,0.4557086825,-0.2516758144,0.2150686979,-0.0657660887,-0.0497675762,0.0068621882,0.1790613979,-0.2423692346,0.5438681245,-0.549284339,-0.0779446214,0.1637871563,-0.254853487,0.0568843558,0.2144898623,-0.1388094127,-0.4605132341,0.5378773212,0.0551779233,-0.4742545485,0.0790166408,-0.1847565025,0.3659106195,-0.0947691575,-0.2403983772,-0.0366393737,0.3571832776,0.0730747506,0.1886267066,0.0995163396,0.1005748287,-0.1019120887,-0.2168260962,-0.0144391116,-0.0712287053,-0.0771795213,-0.3374569416,-0.4207014441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Hi ! We haven't worked in this lately and it's not in our very short-term roadmap since it requires a bit a work to make it work with arrow. Though this will definitely be added at one point.","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":38,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nHi ! We haven't worked in this lately and it's not in our very short-term roadmap since it requires a bit a work to make it work with arrow. Though this will definitely be added at one point.","embeddings":[-0.098730363,0.2157691866,-0.036189571,0.1980810314,0.3335241973,-0.0148143889,0.6729161739,0.0941125751,-0.2653336525,0.1920526028,0.3300935626,0.2847850323,-0.1851584017,-0.0952648148,0.2423299104,-0.3900923729,0.1759411246,0.4636602998,0.0826879442,0.0519561619,-0.0520088933,-0.1384513229,-0.2443876714,0.291831702,-0.3977594078,-0.0074219075,-0.2625413537,-0.2283953875,-0.1936278641,-0.5090429187,0.3147083819,-0.0006440361,0.1157892793,0.2204848826,-0.0001211161,0.0560126528,0.4980730712,0.0563406199,-0.2676818371,-0.1266442239,0.0979174301,-0.3960363865,0.2492384017,-0.3438536227,-0.0336581133,-0.8026284575,0.0193171967,0.022197593,0.0784239694,0.2514157891,0.1347982287,0.1327703595,0.2914691865,-0.1992772669,0.4904192388,0.3490455449,-0.2730592489,-0.0843790248,0.3083592057,0.0497198217,-0.019599501,-0.2024357468,-0.0691877753,0.1527654529,0.3510464132,0.3043106496,0.3325360119,-0.2280716896,-0.0403411388,0.4299192131,0.5088960528,-0.1575274915,-0.1708115935,-0.1280062795,0.0862528756,-0.0945466086,-0.0624217018,0.23521173,-0.2872251272,-0.070620805,-0.1919656992,0.1625069529,-0.4509160817,0.3606544733,0.2118666619,0.0809227228,-0.011420276,0.224977985,-0.0935353264,-0.2292497754,0.5371082425,-0.3507045805,0.0374394953,0.2259349227,0.1172810718,0.1355792731,-0.1068528518,-0.2385307401,0.4191494882,-0.0049818107,0.1828813404,-0.0639445111,0.3001132309,0.1084017679,-0.1901930124,0.2092389464,-0.0208750684,-0.1174462512,-0.079273425,0.1655405909,0.0466907024,0.0618224926,-0.2769748271,-0.2844107449,0.4106661081,-0.070956558,0.2312156707,0.2673474848,0.0908107162,0.029430246,-0.5808674097,-0.1497225612,-0.0254861526,0.2121632546,0.0922296867,-0.1061180383,-0.1419930756,0.4051830173,0.0228594691,-0.0956141427,-0.1998205781,0.0070374119,-0.2048146278,-0.0589816011,0.1027753428,0.1123301014,0.1261414587,-0.0109216636,0.0562070161,0.1010504514,0.2348321527,0.0499260314,0.1728387773,0.1402012706,-0.1890899837,-0.0646460578,-0.033547692,-0.1814259142,-0.2189383358,0.3710366189,-0.074490577,-0.2519074678,-0.4225301147,0.0955078527,-0.2765147984,0.0651352033,-0.0594705343,0.0752608404,0.2303730994,0.062844269,0.040109355,-0.0310946181,0.2473570555,-0.3945892751,0.187360853,-0.0015768883,-0.317786932,-0.0184286162,-0.147383675,-0.0497858599,-0.0766056105,-0.08500579,-0.2564000189,0.2750615478,-0.2200451642,0.2653540075,0.6616533995,0.2300476432,-0.027522454,0.0093542282,0.0203854386,-0.1692549139,0.2073658705,0.0774759054,-0.1663610935,-0.037582919,-0.3147429228,0.2284341455,-0.1725175679,-0.2374633849,-0.0014495193,-0.3701037169,0.4319508374,0.1311580986,0.0757123306,-0.025696978,-0.0593152493,-0.4242015183,0.0617760122,-0.4753856063,0.2388728857,0.0956538469,0.3680794835,-0.1174954548,0.1033731177,-0.1987990141,-0.6633157134,0.1479938328,-0.3269465566,0.0684972331,-0.1035158411,-0.0127287619,-0.116716668,0.1993882358,-0.1957894266,0.0680562183,0.0465568937,0.0832817182,0.1992708445,-0.1604417413,-0.22584261,-0.4588735998,0.0166465584,-0.0229313765,-0.3702343404,0.5743254423,-0.0487583391,-0.3515618443,-0.3371126056,0.1342283189,0.0599165261,-0.0795552805,0.0301119033,0.1492354721,-0.0830535591,-0.0787754729,-0.4900359809,0.086402446,0.0316077992,-0.473190099,0.2193963528,0.1813652962,-0.0090986267,-0.2457866669,0.0635049418,0.3747840226,0.0879565179,0.5813252926,-0.126971975,0.0660553575,-0.0166170597,0.1873176247,-0.0637854636,-0.0988989323,-0.0413624495,0.1875483096,0.2195616513,-0.0412394181,-0.0814983547,0.0124033531,0.2831691802,-0.0763910264,0.0485755615,0.0943504274,-0.2622021139,0.024129929,0.2071482092,-0.2704463303,0.0933376402,0.1391813755,-0.074250184,0.0506263003,-0.2858203053,-0.1754847914,0.1957800239,0.0846200436,0.5732079148,0.1150958687,0.2738003433,-0.2092379332,-0.2014389485,-0.176131919,0.0851689354,0.150538072,-0.2391188145,0.0694116428,-0.0960010886,0.0080925869,-0.0627086237,-0.5734021664,0.0562896356,-0.0410986617,-0.1251586378,-0.0071785632,-0.1493553221,0.2693190575,0.0940229446,-0.1845419854,0.2232490033,-0.1377822757,-0.2320767939,-0.0810120031,-0.1397719979,0.0229129959,0.1893306226,0.031169109,-0.0940592363,0.0999367386,0.0354998484,-0.4118802845,-0.4540100992,0.0996510759,-0.1142487898,0.3344694078,0.5833544135,-0.0149075463,0.0700243488,-0.3229906559,0.4294372499,-0.0011500459,-0.2277644128,0.351760596,-0.2792389095,0.078852579,-0.1059107035,-0.0940299928,0.0376971699,-0.2489798814,0.1979440749,0.4887146056,0.2029841542,0.2031731009,0.4810780883,0.3484107554,-0.080846265,0.0429802574,0.2867523134,0.2725436985,0.4623089731,-0.1302213222,-0.2261158228,-0.0421507992,-0.9024546146,0.0799669623,0.2271867096,0.0157992914,-0.1956704706,-0.1324646175,0.3634037971,-0.0211039837,0.1189388111,0.2944111526,0.1568292975,0.0415894873,-0.1517076492,-0.0737201124,0.4391624928,0.1008504704,-0.0229888428,-0.0276293308,0.2658874094,-0.1881069988,0.6245250106,0.1190605685,-0.3782255352,0.1852205098,-0.2673841417,-0.0269855857,-0.2219008505,0.0707336441,-0.1824989468,0.1632906348,-0.0858732015,-0.06857723,-0.0720269829,-0.2815803587,0.1158354059,0.2448932379,-0.0577874891,-0.0230543893,0.0227087289,-0.02644963,0.1611940265,-0.1319911927,0.0804307833,-0.2733572125,-0.1079016328,-0.1891680062,-0.282892853,0.2439753562,0.017773835,-0.2827561498,0.0709548146,0.0676847845,0.1395320743,0.4566805661,-0.1347286999,0.2383640856,-0.1145222113,0.0929659158,-0.176099807,0.543880403,-0.1566630751,-0.231419459,0.5165029168,0.2314660996,-0.523229301,-0.1886627823,-0.1555230021,0.1535919905,-0.3465435803,0.1494383812,-0.3906955719,-0.011744393,0.3368811607,0.3738782704,0.0040631583,-0.1206001043,-0.1647748649,-0.2754901946,-0.0847474933,-0.2419990897,-0.0858827904,0.3476139903,0.0977022275,-0.419130832,-0.3573639989,0.0367634743,-0.1770736724,-0.0007152061,0.4346479774,-0.2874872684,0.0712671876,-0.339414984,-0.0203110483,0.2301027626,0.1398201883,0.1775071919,-0.0662317351,0.1435071081,-0.0926415771,0.4353049994,0.0863309279,-0.0882268548,0.0783949941,0.1191977933,0.1082923487,-0.0133696292,-0.0798894241,0.4191329181,-0.3305972219,-0.161999926,-0.3191149533,0.6340940595,-0.0104079833,0.1801792085,0.2604023516,0.421644479,-0.1894237697,0.2551965117,0.5969882011,0.8620143533,0.0378392152,0.1722197533,0.691613853,0.0355377905,0.2889745533,0.3305392563,0.1108352393,-0.5170720816,0.0400558636,-0.0893094912,-0.4294856191,0.239800036,0.2657113075,-0.2476804554,-0.0125882914,-0.2438177317,0.4807460606,0.1565134376,-0.0536971986,-0.2015184909,-0.1308676302,-0.0795704126,0.0604366884,0.0184548069,-0.3900193572,-0.1430811435,-0.2199503183,-0.1441862434,-0.0366119035,-0.0495154373,0.0460662544,-0.3448441923,-0.0502130836,0.5167239308,-0.5515484214,0.0668505281,0.1653504819,0.0559453741,-0.2133409232,-0.117941238,-0.036618989,0.2471693903,-0.2361063361,-0.1293750107,-0.3045616448,0.2676637769,0.0444943644,-0.0840871483,0.053834822,-0.1256597787,-0.0603569597,-0.0545581281,-0.029210411,0.1936974674,-0.2170751095,-0.1509849727,-0.1145187095,-0.159256056,-0.1218208,0.0389195643,0.1215787083,-0.141083315,-0.137894243,0.1177484617,0.0704649389,-0.0119492663,-0.0516499504,0.5510166287,0.3104273975,0.3901571333,0.2568704784,-0.0192213375,-0.1990412921,0.0780669823,0.5724360347,-0.6061903238,0.1683155,0.1329140216,-0.178997159,-0.1007020473,0.5450142622,-0.1170513555,0.2916316688,-0.1273502856,0.0724846199,-0.3985959291,0.2515298128,0.1833909005,0.3132033348,-0.2412899137,-0.0718392953,0.0583634973,-0.1033632532,-0.2108917981,0.0564168394,0.0543428585,0.0073620318,-0.5352551937,-0.0172982235,0.0381671004,0.0722812116,0.0940530524,0.0146775683,-0.3673615754,-0.1302560717,-0.0836908296,0.173405394,0.1289810985,-0.1058810353,0.2010436803,-0.0167974196,-0.2995280027,-0.2105608284,-0.0431372151,-0.1308129281,-0.1920118183,-0.0990229025,0.3399970829,0.0207304507,-0.4380847812,-0.1024568677,-0.2397677898,0.2741346657,0.0725254342,0.0206980966,0.0892329961,-0.0455881357,-0.0507865176,0.2219917923,0.0105400756,-0.2401548028,0.3798941374,-0.4209012389,-0.2788291574,-0.3206575811,0.386651516,0.2373379022,-0.2053187937,0.1342788786,0.1600240767,0.1335510463,-0.2931455076,0.112754181,0.1565288305,-0.0677808225,0.1565395147,0.3240006864,0.1053992361,-0.0171625223,-0.082428813,0.096317023,-0.1141307428,0.0136931455,0.1066901684,0.6419899464,-0.1201371253,0.1007379144,0.4103653133,-0.0068511008,0.2283302397,-0.1362492293,-0.1021320075,0.3618030846,0.1366477758,-0.11503499,0.2637448609,0.0144052124,0.0423780344,0.194503054,0.5667410493,-0.1246595904,0.0776446983,0.6018323898,0.2360682636,0.0012238524,-0.357630372,-0.1152864844,-0.0390124395,-0.002881682,-0.1766206622,-0.1126499921,-0.2524791658,-0.3445046842,-0.1269434988,-0.398576498,0.1758284569,-0.0852212235,0.1613777131,-0.3082702756,0.7386052608,0.192874223,0.0191336926,-0.0908377096,0.4798517823,0.3387975395,-0.1544819772,-0.1970342696,0.4266251922,0.3237750828,0.3594305217,-0.3366144896,-0.0549653545,-0.0887424797,-0.2488773614,0.0408422127,0.1050989628,0.2121212929,0.092043668,0.2440501153,0.0547367632,-0.1042803153,0.2355066091,0.2357189655,0.0055963225,-0.1912838072,0.3488586545,-0.2889256775,-0.1870106161,0.0504454151,-0.0754319504,-0.3536975384,-0.1060263887,0.3406725824,-0.108630307,-0.0013783076,-0.0808870345,0.0409444608,-0.2456537932,0.447625339,0.073339276,0.5915705562,-0.1646002084,-0.0669567734,-0.3540522456,0.3694659472,-0.427302599,0.0692674816,0.2883570492,0.272444576,-0.0647480339,0.1407485902,0.0931721628,0.2419687808,0.0484821387,0.3780895472,-0.3498033881,-0.111547038,0.2903210521,0.0275091883,-0.1034033746,-0.3440865576,0.148264721,-0.3978516161,-0.0133104986,0.2130671889,0.1051585078,0.1991250366,0.2104683518,0.5235130191,-0.0204211492,0.0953864977,0.1667882502,-0.0646306053,-0.2480611503,-0.172142446,-0.1831062436,0.2971852422,0.044929143,0.405739367,-0.1876797974,0.2667944133,-0.1443157792,-0.1357725114,0.1033096015,0.244129464,-0.2240691781,0.4508425593,-0.6143883467,0.0196529385,0.0096426122,-0.2210223824,0.0637934059,0.0643513277,-0.1437482834,-0.4331988096,0.6617772579,-0.2992739677,-0.4578563869,0.0730499998,-0.0448698439,0.4074344635,-0.1394216865,-0.2819718421,-0.0614761747,0.4609859288,0.0146077657,-0.0193116684,0.0172541589,0.0047944579,0.0918335468,-0.1730270982,0.1973406523,-0.0937364623,-0.1125746742,-0.4181849957,-0.5695380569]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"@lhoestq, thanks for the update.\r\n\r\nI actually tried to modify some piece of code to make it work. Can you please tell if I missing anything here?\r\nI think that for vast majority of cases it's enough to make first dimension of the array dynamic i.e. `shape=(None, 100, 100)`. For that, it's enough to modify class [ArrayExtensionArray](https:\/\/github.com\/huggingface\/datasets\/blob\/9ca24250ea44e7611c4dabd01ecf9415a7f0be6c\/src\/datasets\/features.py#L397) to output list of arrays of different sizes instead of list of arrays of same sizes (current version)\r\nBelow are my modifications of this class.\r\n\r\n```\r\nclass ArrayExtensionArray(pa.ExtensionArray):\r\n def __array__(self):\r\n zero_copy_only = _is_zero_copy_only(self.storage.type)\r\n return self.to_numpy(zero_copy_only=zero_copy_only)\r\n\r\n def __getitem__(self, i):\r\n return self.storage[i]\r\n\r\n def to_numpy(self, zero_copy_only=True):\r\n storage: pa.ListArray = self.storage\r\n size = 1\r\n for i in range(self.type.ndims):\r\n size *= self.type.shape[i]\r\n storage = storage.flatten()\r\n numpy_arr = storage.to_numpy(zero_copy_only=zero_copy_only)\r\n numpy_arr = numpy_arr.reshape(len(self), *self.type.shape)\r\n return numpy_arr\r\n\r\n def to_list_of_numpy(self, zero_copy_only=True):\r\n storage: pa.ListArray = self.storage\r\n shape = self.type.shape\r\n arrays = []\r\n for dim in range(1, self.type.ndims):\r\n assert shape[dim] is not None, f\"Support only dynamic size on first dimension. Got: {shape}\"\r\n\r\n first_dim_offsets = np.array([off.as_py() for off in storage.offsets])\r\n for i in range(len(storage)):\r\n storage_el = storage[i:i+1]\r\n first_dim = first_dim_offsets[i+1] - first_dim_offsets[i]\r\n # flatten storage\r\n for dim in range(self.type.ndims):\r\n storage_el = storage_el.flatten()\r\n\r\n numpy_arr = storage_el.to_numpy(zero_copy_only=zero_copy_only)\r\n arrays.append(numpy_arr.reshape(first_dim, *shape[1:]))\r\n\r\n return arrays\r\n\r\n def to_pylist(self):\r\n zero_copy_only = _is_zero_copy_only(self.storage.type)\r\n if self.type.shape[0] is None:\r\n return self.to_list_of_numpy(zero_copy_only=zero_copy_only)\r\n else:\r\n return self.to_numpy(zero_copy_only=zero_copy_only).tolist()\r\n```\r\n\r\nI ran few tests and it works as expected. Let me know what you think.","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":224,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\n@lhoestq, thanks for the update.\r\n\r\nI actually tried to modify some piece of code to make it work. Can you please tell if I missing anything here?\r\nI think that for vast majority of cases it's enough to make first dimension of the array dynamic i.e. `shape=(None, 100, 100)`. For that, it's enough to modify class [ArrayExtensionArray](https:\/\/github.com\/huggingface\/datasets\/blob\/9ca24250ea44e7611c4dabd01ecf9415a7f0be6c\/src\/datasets\/features.py#L397) to output list of arrays of different sizes instead of list of arrays of same sizes (current version)\r\nBelow are my modifications of this class.\r\n\r\n```\r\nclass ArrayExtensionArray(pa.ExtensionArray):\r\n def __array__(self):\r\n zero_copy_only = _is_zero_copy_only(self.storage.type)\r\n return self.to_numpy(zero_copy_only=zero_copy_only)\r\n\r\n def __getitem__(self, i):\r\n return self.storage[i]\r\n\r\n def to_numpy(self, zero_copy_only=True):\r\n storage: pa.ListArray = self.storage\r\n size = 1\r\n for i in range(self.type.ndims):\r\n size *= self.type.shape[i]\r\n storage = storage.flatten()\r\n numpy_arr = storage.to_numpy(zero_copy_only=zero_copy_only)\r\n numpy_arr = numpy_arr.reshape(len(self), *self.type.shape)\r\n return numpy_arr\r\n\r\n def to_list_of_numpy(self, zero_copy_only=True):\r\n storage: pa.ListArray = self.storage\r\n shape = self.type.shape\r\n arrays = []\r\n for dim in range(1, self.type.ndims):\r\n assert shape[dim] is not None, f\"Support only dynamic size on first dimension. Got: {shape}\"\r\n\r\n first_dim_offsets = np.array([off.as_py() for off in storage.offsets])\r\n for i in range(len(storage)):\r\n storage_el = storage[i:i+1]\r\n first_dim = first_dim_offsets[i+1] - first_dim_offsets[i]\r\n # flatten storage\r\n for dim in range(self.type.ndims):\r\n storage_el = storage_el.flatten()\r\n\r\n numpy_arr = storage_el.to_numpy(zero_copy_only=zero_copy_only)\r\n arrays.append(numpy_arr.reshape(first_dim, *shape[1:]))\r\n\r\n return arrays\r\n\r\n def to_pylist(self):\r\n zero_copy_only = _is_zero_copy_only(self.storage.type)\r\n if self.type.shape[0] is None:\r\n return self.to_list_of_numpy(zero_copy_only=zero_copy_only)\r\n else:\r\n return self.to_numpy(zero_copy_only=zero_copy_only).tolist()\r\n```\r\n\r\nI ran few tests and it works as expected. Let me know what you think.","embeddings":[0.0081767803,0.1172873601,-0.1061084569,0.2801861167,0.4025137126,-0.0643157065,0.5933448672,0.1092592403,-0.2332408428,0.2207361907,0.1275905818,0.3022358418,-0.2472839952,-0.021195896,0.2019220144,-0.3135286272,0.0890397951,0.5067413449,0.0488858111,-0.027319083,-0.0532906801,-0.0695483908,-0.1607426554,0.1468552947,-0.2513417602,0.068744354,-0.2443502545,-0.2243448496,-0.1773382425,-0.4697862864,0.2879074514,0.1888695657,0.0287835617,0.0859840885,-0.0001230811,0.0541417636,0.3321122825,-0.000784546,-0.271482408,-0.0819979534,0.2015638053,-0.4400664866,0.2384080887,-0.3255049884,-0.0139813889,-0.8362496495,-0.0494648106,-0.1110906601,0.0708753094,0.2057912499,0.1044426039,0.0736096427,0.2843008041,-0.0789362416,0.469932735,0.3682914674,-0.2882514894,-0.1275915951,0.3975459039,0.1431577504,-0.0400086902,-0.2101088613,-0.1313411742,0.1417175084,0.3835519552,0.2264853567,0.21871306,-0.2353864908,0.0733531862,0.3944592178,0.5115253925,-0.2453649491,-0.1943132877,-0.2347998172,0.0897710398,-0.0467550159,-0.0856359825,0.3283034563,-0.2202517986,-0.0487610213,-0.3536956906,0.228640601,-0.5123696923,0.3541974127,0.1902458519,0.0773997903,-0.0146469232,0.2034287453,-0.1073710397,-0.2308212966,0.5585191846,-0.3572503328,0.0319317989,0.2399896532,0.0688930228,0.1028445438,-0.0690795258,-0.4354409575,0.3606022298,-0.0460961498,0.1935750544,-0.0213646404,0.2743874788,0.0304053184,-0.1026975885,0.2985353172,-0.0004932333,-0.0596963055,-0.2731168568,0.1101117358,0.1455025822,0.0948338583,-0.1704499573,-0.1599857956,0.4061376154,-0.0382157266,0.2606358528,0.2152996212,0.0485563874,0.01908659,-0.7071174383,-0.1580285132,-0.0421594679,0.1278482974,0.1143367141,0.0015046305,-0.0514350832,0.3788039982,0.0120271919,-0.130340308,-0.1523470432,-0.0806081817,-0.1551577151,0.0050094398,0.0776838809,0.242716983,0.0285711735,0.0373322554,0.1099322289,0.0145733897,0.0978115797,0.0748503357,0.2582392693,0.1224490479,-0.1893648952,-0.0665747374,-0.0766481608,-0.0325968675,-0.2722557485,0.4397105277,-0.0544199869,-0.2422068566,-0.3447396457,0.0649461374,-0.2117099464,0.118435584,-0.1155644432,0.0989569426,0.243027553,0.0399205908,0.0517658889,-0.0657433197,0.1976664066,-0.4457695484,0.1396119446,0.0001105489,-0.4662371278,0.0617593825,-0.0835714191,-0.0497667231,-0.0255610775,-0.1201092005,-0.2007638514,0.1728928983,-0.2748225033,0.3207069933,0.5103976727,0.3285774887,-0.1169237345,0.0830684155,0.0326129645,-0.0853299946,0.2711330354,0.0790556371,-0.1332363039,0.0536823235,-0.3695650697,0.1610988081,-0.0928770304,-0.1600335687,-0.0935422108,-0.4882471263,0.3123804331,0.1447314769,0.1749598235,0.0073294099,-0.0438041948,-0.4492785633,0.2552002966,-0.5342483521,0.2359112054,0.1078927666,0.3874282539,-0.0732296407,0.0397920944,-0.1910843849,-0.6725085974,0.1904433221,-0.4558854997,0.048146721,-0.0301428977,-0.0505475476,-0.1543769836,0.1480968446,-0.173468411,0.1258011758,0.0503978841,0.1491036564,0.2341405749,-0.2919446528,-0.0407898687,-0.3832749724,-0.1340226382,-0.1048874259,-0.5120505095,0.6572319865,0.0724064708,-0.2904246747,-0.2972910106,0.1471006125,0.0280328784,-0.0370860957,-0.1188070551,0.1229283139,-0.0658304915,-0.113117449,-0.470267117,0.1696010679,-0.0339044966,-0.2905175388,0.2976221144,0.1144750789,-0.0782995895,-0.1841138899,0.1907569319,0.5881355405,0.0851957574,0.5646958351,-0.1479355544,-0.0133569725,-0.123678565,0.1781947762,-0.0113226762,-0.1554098874,-0.1618798077,0.230711773,0.2857266963,-0.0325835124,-0.1471522748,0.1662853062,0.2829980254,-0.0573412813,-0.0830609724,0.0654538125,-0.1417272687,0.068353273,0.1191514879,-0.1040920317,0.2394607067,0.1961164773,-0.1819210649,-0.0090215402,-0.1736881733,-0.1677853614,0.2829214633,0.1613493711,0.5691132545,0.140334785,0.3767085373,-0.3264777064,-0.2785777748,-0.1017967686,0.0143978046,0.1471296251,-0.2031050473,0.0580046773,-0.1077750176,-0.0069000553,-0.1237576082,-0.5938737988,-0.0299383029,0.0092836525,-0.0754430518,-0.00920224,-0.1901368946,0.1717361659,0.1932961196,-0.1336842775,0.2393604219,-0.0947372764,-0.1887859106,0.0840677246,-0.1397699714,-0.006381908,0.2208733112,0.0207979623,-0.09176559,0.1303753853,-0.0254858453,-0.4571220875,-0.3635931313,0.105608426,-0.0876275823,0.282962501,0.5301013589,-0.0609327778,0.0696673542,-0.2881963849,0.3887398541,0.0011954469,-0.2087062746,0.2950402498,-0.2056496292,0.1029592007,-0.0882220715,-0.0808877349,-0.0251198038,-0.2163183838,0.2204162329,0.3758439422,0.2948091328,0.2302731723,0.4909337461,0.2809514701,-0.0408370122,0.0959874541,0.1638640463,0.2621394992,0.3690192997,-0.0638757721,-0.1603313237,-0.1058410406,-0.8136040568,0.087740168,0.3370889127,-0.0335913338,-0.3418251276,-0.0918516293,0.1671680957,-0.0780736059,0.0688280165,0.3388420939,0.0913430005,0.0552683398,-0.0304660685,-0.044307258,0.414275825,0.1505993754,-0.0082934257,0.0106465947,0.2831360102,-0.198879391,0.5975487828,0.0907361731,-0.2871260941,0.1008346081,-0.2542867661,-0.0541666187,-0.2191137969,-0.0126045523,-0.1798820794,0.1291854829,-0.231093809,0.0599529445,-0.1282235384,-0.3053784668,0.2517749369,0.3372273147,-0.0572942458,-0.0584359467,0.1691236794,0.1370233744,0.2135795355,-0.1607351899,0.1220996678,-0.356007576,-0.1773287952,-0.3337774277,-0.3037228584,0.1417999715,-0.0402957387,-0.2624799013,0.006457191,0.0919445083,0.2578709722,0.5154531598,-0.0265330598,0.2461932898,-0.1622951925,0.1618169546,-0.0537190586,0.5656299591,-0.002727424,-0.2775247097,0.5173031092,0.1740545928,-0.3937618136,-0.3545571864,-0.1032175273,0.0811747089,-0.2107179314,0.0577996038,-0.3799524903,-0.1724500656,0.2668898702,0.3965167999,-0.0091542527,-0.1572368592,-0.1660692692,-0.116889067,-0.1453478485,-0.1957619041,-0.185888052,0.4336944222,-0.0070196777,-0.4450423717,-0.3357054889,-0.1189310402,-0.1390493065,-0.0002583457,0.5432819724,-0.2023106962,-0.0015573518,-0.3428708613,-0.0080447476,0.2560172379,0.1139485314,0.182446003,0.0979418457,0.2349186689,-0.0910520777,0.3822708726,0.2372794002,-0.2198994905,0.044897031,0.1027145386,0.1267090291,-0.0041624969,-0.1936790198,0.3796470463,-0.4180856943,-0.0466983728,-0.3388864696,0.6189413071,0.0008023975,0.1976180524,0.296982497,0.4021168649,-0.2434062511,0.1732521057,0.5898041129,0.8234207034,0.021046089,0.0911853388,0.7183238864,0.0649314299,0.0476040281,0.382261008,0.2309387475,-0.4490117133,0.0541267134,-0.0949690565,-0.3679797947,0.2082755864,0.3048765957,-0.1094015539,-0.0617870763,-0.0698586255,0.4589512348,0.0870966986,-0.0568038933,-0.247720167,-0.1235439405,-0.1430539191,-0.016434418,-0.1008479148,-0.4252541661,-0.0255964827,-0.3447966874,-0.1840120852,0.0054284702,-0.0751822367,-0.075714536,-0.4166588783,0.0070467941,0.5495271087,-0.5068945289,0.0692056194,0.0717687085,-0.1224000081,-0.2166785151,-0.1111352518,-0.2361381799,0.274156183,-0.1455589384,-0.221679315,-0.3660665452,0.2566587329,0.135408923,0.0125843836,-0.0158618595,-0.1012873724,-0.0187194049,-0.1451930553,-0.0238190517,0.4159697592,-0.3938892484,-0.1620072722,-0.1356797367,-0.1929825395,-0.0884041116,-0.00104245,0.1095190421,-0.1016231179,-0.0975384489,0.1391428411,0.0800201148,-0.0463301055,-0.0389971249,0.6464511156,0.2786627412,0.3994380534,0.1047930047,0.0850064009,-0.1724823564,0.1371434182,0.4868006706,-0.5715532303,0.2576581836,0.0907599851,-0.1696768701,-0.0290163662,0.5399835706,-0.0442121662,0.4526693821,-0.1568921953,0.1109001562,-0.3084824681,0.1464322954,0.2114505172,0.4577489197,-0.2875778377,0.0632194951,0.0201824773,-0.041405987,-0.1726886034,-0.0439859703,0.0355866663,-0.0942763388,-0.5580289364,-0.028081825,0.0223017521,0.0794299096,0.0192790758,-0.0443857312,-0.5251638889,-0.0822562277,0.0110797621,0.1850548387,0.1398024708,0.0243031811,0.0904350951,-0.0721930042,-0.2565634847,-0.2234506905,-0.0548698939,-0.1382227242,-0.1830091923,-0.1036143228,0.3470047414,-0.0156557765,-0.4039346576,-0.0975511223,-0.2694081068,0.2371115237,0.1720994264,0.0714603439,0.0395612717,-0.0362390354,-0.1881852448,0.2440897226,-0.0303860847,-0.2020336688,0.2280722558,-0.4099785984,-0.4220006168,-0.21879372,0.4399959147,0.23592709,-0.1673677862,0.0056287064,-0.0047206767,0.1214899272,-0.2195072174,0.0944188833,0.2432648987,0.0230192468,0.1590060592,0.4403980076,0.2849532366,-0.1559406072,-0.0112904981,0.1214797944,-0.0477193929,0.0541858226,0.1368535161,0.6701667905,-0.0493996963,0.0855326653,0.4568313956,-0.0183634553,0.22316508,-0.075039126,-0.1158412918,0.4804654121,0.0057458282,-0.1140032262,0.1891218722,-0.0114356019,0.0559552722,0.2198228985,0.4599031508,-0.0305869598,0.0386232883,0.6613844037,0.079027839,-0.0843454674,-0.316994369,-0.128635928,-0.0705109164,0.0905375108,-0.0533489659,-0.1738108099,-0.2111714929,-0.3472498655,-0.1079620868,-0.4550631642,0.2416435778,-0.1753717959,0.2323980182,-0.3419376612,0.5808370709,0.2017829716,0.0361171551,-0.1317173094,0.4449385405,0.3095203042,-0.0174006745,-0.180003047,0.3680502772,0.2668584883,0.4084316492,-0.1282460839,-0.1556953043,0.0157008339,-0.2543372512,0.0782622844,0.0555277057,0.1852087975,0.0283581764,0.1995549649,0.0424828865,-0.0792540014,0.2689209282,0.2696076632,0.0073093502,-0.1913244873,0.3418929279,-0.1809462309,-0.193979606,-0.0263808034,-0.2045589536,-0.4663594961,-0.0944080204,0.2306022197,-0.0828590691,-0.0167258605,0.0332151316,0.0149346609,-0.190934509,0.306932956,0.0963119715,0.4915101528,-0.2389019877,-0.0760149509,-0.2282328159,0.3368020654,-0.3284419775,0.0892068297,0.3115128875,0.3918777406,-0.0411182195,-0.0048771137,0.0992718488,0.129320696,0.0404654443,0.4305920899,-0.3858338892,-0.0226858165,0.3656060696,-0.0340372138,-0.0408562422,-0.2874153256,0.1530673951,-0.3587368131,-0.0401052199,0.195693478,0.2613576651,0.1265033782,0.2545158267,0.3684670329,-0.086130321,-0.0251062848,0.1655199677,-0.1269243509,-0.294483304,-0.1134409085,-0.1281651706,0.3101491928,0.2008766532,0.5762239099,-0.1929154545,0.2245916426,-0.1427208036,-0.124385044,0.1041136682,0.2011827976,-0.2228565216,0.4451680481,-0.5870544314,0.1153985113,-0.0236126706,-0.162664637,0.2037955821,0.1254850775,-0.1950470209,-0.4592522085,0.6747375727,-0.3088298738,-0.4959323406,0.130236432,-0.0085577052,0.2961165011,-0.107486777,-0.3079210818,-0.0651789755,0.4423903823,-0.0427193642,0.2331910431,0.0388664901,0.0046465574,0.0975326821,-0.1714746207,0.1417770684,0.0340842716,-0.1422705054,-0.4382562041,-0.5738328695]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/887","title":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comments":"Thanks for diving into this !\r\n\r\nIndeed focusing on making the first dimensions dynamic make total sense (and users could still re-order their dimensions to match this constraint).\r\nYour code looks great :) I think it can even be extended to support several dynamic dimensions if we want to.\r\n\r\nFeel free to open a PR to include these changes, then we can update our test suite to make sure it works in all use cases.\r\nIn particular I think we might need a few tweaks to allow it to be converted to pandas (though I haven't tested yet):\r\n\r\n```python\r\nfrom datasets import Dataset, Features, Array3D\r\n\r\n# this works\r\nmatrix = [[1, 0], [0, 1]]\r\nfeatures = Features({\"a\": Array3D(dtype=\"int32\", shape=(1, 2, 2))})\r\nd = Dataset.from_dict({\"a\": [[matrix], [matrix]]})\r\nprint(d.to_pandas())\r\n\r\n# this should work as well\r\nmatrix = [[1, 0], [0, 1]]\r\nfeatures = Features({\"a\": Array3D(dtype=\"int32\", shape=(None, 2, 2))})\r\nd = Dataset.from_dict({\"a\": [[matrix], [matrix] * 2]})\r\nprint(d.to_pandas())\r\n```\r\n\r\nI'll be happy to help you on this :)","body":"I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>","comment_length":164,"text":"pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nI set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) \r\n\r\n```python\r\n def _info(self):\r\n return datasets.DatasetInfo(\r\n description=_DESCRIPTION,\r\n # This defines the different columns of the dataset and their types\r\n features=datasets.Features(\r\n {\r\n \"pose\": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype=\"float32\"))\r\n }\r\n ),\r\n homepage=_HOMEPAGE,\r\n citation=_CITATION,\r\n )\r\n def _generate_examples(self):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n yield 1, {\r\n \"pose\": [np.zeros(shape=(137, 2), dtype=np.float32)]\r\n }\r\n```\r\n\r\nBut this doesn't work -\r\n> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>\nThanks for diving into this !\r\n\r\nIndeed focusing on making the first dimensions dynamic make total sense (and users could still re-order their dimensions to match this constraint).\r\nYour code looks great :) I think it can even be extended to support several dynamic dimensions if we want to.\r\n\r\nFeel free to open a PR to include these changes, then we can update our test suite to make sure it works in all use cases.\r\nIn particular I think we might need a few tweaks to allow it to be converted to pandas (though I haven't tested yet):\r\n\r\n```python\r\nfrom datasets import Dataset, Features, Array3D\r\n\r\n# this works\r\nmatrix = [[1, 0], [0, 1]]\r\nfeatures = Features({\"a\": Array3D(dtype=\"int32\", shape=(1, 2, 2))})\r\nd = Dataset.from_dict({\"a\": [[matrix], [matrix]]})\r\nprint(d.to_pandas())\r\n\r\n# this should work as well\r\nmatrix = [[1, 0], [0, 1]]\r\nfeatures = Features({\"a\": Array3D(dtype=\"int32\", shape=(None, 2, 2))})\r\nd = Dataset.from_dict({\"a\": [[matrix], [matrix] * 2]})\r\nprint(d.to_pandas())\r\n```\r\n\r\nI'll be happy to help you on this :)","embeddings":[0.0811185241,0.0660867691,-0.0749657378,0.1997890472,0.5059955716,-0.0620560125,0.6662253141,0.2228145152,-0.2351686358,0.2103948593,0.2033907324,0.3141406775,-0.1679162234,0.15190956,0.1614730805,-0.52645123,0.1352110356,0.4473284483,-0.0370457545,0.0868214816,-0.0864016786,-0.1268898994,-0.2347302288,0.1998275518,-0.3560606539,0.043541424,-0.4179327786,-0.1319022328,-0.1962919384,-0.3397125602,0.2675036788,0.1394725293,0.1546587497,0.1940463483,-0.0001162566,-0.0201859809,0.2754315734,0.0027224848,-0.1653280109,-0.1504532695,0.1756362766,-0.4560697675,0.1426978111,-0.3767650723,-0.0597968884,-0.7914659381,-0.1558022797,-0.0565776117,0.0584394522,0.1819483638,0.186089769,0.1970723122,0.3445558846,0.0070053488,0.3372129798,0.2662027776,-0.3337921798,-0.0743540004,0.2205041051,-0.1364297271,-0.0197567232,-0.0812024102,-0.0935630351,0.1253886223,0.3703870177,0.2559622526,0.4412609935,-0.3059556186,-0.0385078937,0.4866024852,0.5266446471,-0.226383552,-0.1486890614,-0.210560903,0.0326052234,-0.0720221996,-0.1134072766,0.3932394087,-0.2839558423,-0.0259063467,-0.1617712677,0.2671583593,-0.4738789499,0.3436041474,0.0153585486,0.1768891215,0.0600757487,0.1067782864,-0.0788151026,-0.3343683183,0.6473375559,-0.2683770061,0.1695421636,0.1576764137,0.1260790229,0.0547992699,0.0399517789,-0.3056690097,0.2611988783,-0.0662692338,0.2565307021,-0.054056149,0.3141371012,0.07778164,-0.0961987153,0.29313609,-0.1910168082,-0.0353444107,-0.1915338784,0.2000769228,0.1407672167,0.1348167658,-0.0958224013,-0.337719202,0.3894747496,-0.0704433247,0.3181086481,0.2844154835,0.0503381267,0.0295106042,-0.7314228415,-0.1056453735,-0.0794327408,0.1333228797,0.0632770285,0.0234068334,-0.0665666834,0.391638726,0.0048370548,0.0344884247,-0.1772149205,0.1383274943,-0.2407189608,-0.1582002491,0.0580462553,0.3111123741,0.032390967,0.0656857565,0.1681861281,-0.0492571257,0.0473771356,0.0943843052,0.1924476475,0.1907957196,-0.2183337212,-0.2042362839,-0.0482183471,-0.2408762276,-0.2660052776,0.3408511579,0.1237652227,-0.1147489101,-0.296264708,0.1257819682,-0.2576963902,0.033282306,0.018987475,0.1656404883,0.1230323464,0.1861085445,0.0670752078,-0.016318094,0.2285517305,-0.4390902221,0.131014064,0.1434193701,-0.3788299263,0.1839784831,0.0902065486,0.0472827926,-0.0204891823,-0.1859839708,-0.1958567351,0.2340254188,-0.109226726,0.2045662999,0.4382779598,0.1031704918,-0.0508325957,0.0396897495,-0.1008548811,-0.0574911609,0.1844580919,0.0679804683,-0.2325879633,0.0086382972,-0.3313021064,0.361754179,-0.0468136743,-0.1340918988,-0.1193844229,-0.4566427171,0.5288200974,0.1604706049,0.2312584668,-0.0293570794,-0.0378944725,-0.4135747552,0.2131325752,-0.5370700955,0.2064066082,-0.0110484222,0.3182369173,-0.2316179872,0.1017261222,-0.1497078836,-0.6831347942,0.209557429,-0.2615545988,0.1348551661,-0.162729606,0.0161487162,-0.213775456,0.0864838734,-0.1614551246,0.0513470024,0.1555673778,-0.0517706648,0.1015730947,-0.2328872681,-0.1256848276,-0.2842238843,0.0373976976,-0.0718302056,-0.2994231582,0.6068703532,-0.0240522046,-0.2256797701,-0.2027260512,0.1907726824,0.1596813351,-0.0174941216,-0.0039477651,0.164190501,-0.1085744053,-0.0108289039,-0.5107385516,0.220136553,-0.0088732252,-0.2927829921,0.3259554207,0.148439914,0.0312676504,-0.2186182886,0.1461091936,0.5942308903,-0.2054375261,0.5001148582,-0.1696748435,0.1078420877,-0.0440092832,0.1070577204,-0.1206412092,-0.1674749255,-0.1490993351,0.2324400395,0.1000096276,-0.050064683,-0.2019974142,0.1650609225,0.2209520638,-0.0789225101,-0.0382381827,0.0389137268,-0.1205159947,0.0863598883,0.2220526785,-0.0977577046,0.1082824916,0.2069818825,-0.1347233057,0.0805680454,-0.3435122669,-0.1386656314,0.2859655321,0.1962883174,0.5069098473,0.0309313815,0.3337845802,-0.2485228628,-0.2182988524,-0.0744913518,0.0497499816,0.1341309547,-0.3379507363,0.1335220337,-0.1574705541,-0.0023977,0.0058325911,-0.7622504234,-0.092170015,-0.0124773914,-0.0817484781,-0.0018158609,-0.206468761,0.1749795824,0.1516536772,-0.1810883433,0.0989143699,-0.182350114,-0.2228329331,-0.0240917429,-0.1154094189,0.0595060512,0.1391379982,0.1685658544,-0.0194698088,0.1155475602,0.1271314174,-0.2960381806,-0.3022587299,0.0435510539,-0.2427384853,0.376385361,0.4350397885,0.0141049484,0.0395605043,-0.1887970567,0.3545383215,0.2314631343,-0.1280812174,0.3346897364,-0.1032501608,0.0845849141,-0.11869663,0.015143007,0.0281195957,-0.2963921428,0.1846939623,0.3997448981,0.2637634873,0.2296736687,0.5827411413,0.3116678298,0.05216676,0.1223438829,0.1999587268,0.2072532177,0.4295158386,0.0321406238,-0.2279698551,-0.1815953851,-0.8064343333,0.1123274863,0.2651630938,-0.0003437024,-0.4003822505,-0.1192323938,0.4195517004,-0.0658382103,0.1292748302,0.4766238332,0.1627043784,-0.0051324065,-0.1331644207,-0.0762428194,0.3462710679,0.2557983696,-0.0854397789,-0.0197138004,0.2788517177,-0.0859717503,0.5770151019,-0.054513976,-0.3353949785,0.2743908763,-0.2290357798,-0.0486153513,-0.1206581518,-0.0830183178,-0.2085779011,-0.0000325124,-0.0883704424,-0.1022764891,-0.1255783439,-0.3067598641,0.0067855581,0.2312723249,-0.2400392592,-0.0507022925,0.0194869656,0.0140849464,0.1986011714,-0.1047889441,0.1462582201,-0.3106016815,-0.0770361871,-0.1376651824,-0.2555592358,0.104993999,-0.0008804805,-0.0141871925,0.0247313287,0.06883201,0.0896678716,0.3966972828,-0.1401860267,0.3456253707,-0.1802478284,0.1216173023,-0.1736610383,0.6282705665,-0.0847250596,-0.1932678372,0.5385679007,0.1090378761,-0.4343969226,-0.2760931849,-0.2283091992,0.0245924965,-0.2875922918,0.1180344895,-0.368552357,-0.1481019855,0.2178628594,0.3199230731,0.0165560581,-0.1950828582,-0.2308904529,-0.2040196955,-0.0511280224,-0.1320187002,-0.0458167754,0.3967013061,0.0591019094,-0.3044266105,-0.3976419568,-0.1071246341,-0.1380683184,0.1546686292,0.3969093263,-0.1583366394,0.042383831,-0.3037187755,-0.0078442991,0.0820224509,0.0775681362,0.1756754816,0.0458687693,0.0215854906,0.027589025,0.3561809063,0.239168182,-0.114433974,-0.0643039942,-0.0221674163,0.1271671802,-0.1965864599,-0.0822141618,0.4506602585,-0.3617124856,0.0082826223,-0.2644386292,0.6715801358,-0.0901378244,0.1170948446,0.2661953568,0.5690854192,-0.1716607213,0.2838432789,0.6202139854,1.0007245541,-0.0295079779,0.1052313447,0.7753417492,-0.1138477623,0.064445734,0.3265844285,0.2164056003,-0.4182457626,0.1595389694,-0.1114615127,-0.3246540129,0.1179885864,0.1220963523,-0.2034106851,-0.0714606568,-0.2789859474,0.2990578711,0.1234778985,-0.0442035049,-0.1172316372,-0.2051873356,-0.2319262028,0.0663266778,-0.0482186265,-0.3505477309,-0.1049225405,-0.3317467868,-0.2237375528,-0.0289108567,0.0008953165,0.0034473373,-0.333873719,0.0088118622,0.3812509477,-0.5121396184,0.0015977038,0.1844581217,-0.0920751542,-0.1640257537,-0.2030434608,-0.1268924922,0.4598894715,-0.2843433619,-0.0914331228,-0.3732326925,0.3492224514,0.0423916355,-0.1033636704,-0.0897815749,-0.024545135,0.0914483592,-0.1866288632,-0.0215145852,0.2234338373,-0.218254447,-0.0323791206,-0.0599063374,0.0157264844,-0.1066974178,0.09592136,0.0441115908,-0.0428542309,-0.0544345155,0.1728221476,0.0448396318,-0.067849651,-0.0043524015,0.4437872469,0.3373041451,0.2604781687,0.2011245191,-0.0618997812,-0.2048474401,0.113385506,0.3587192595,-0.4849697649,0.1792154312,0.1019558534,-0.2370810062,-0.1328746229,0.4973852336,0.0548735298,0.2220710963,-0.0465807542,0.1459174603,-0.4149115682,0.1146582589,0.1664492339,0.2982103229,-0.3258861303,0.0879805908,0.0275404397,-0.0100676557,-0.288970679,0.0506106131,0.1050652862,-0.0090650348,-0.5112304091,-0.1208126098,0.0193717275,0.0840833411,0.1421371847,-0.0621778965,-0.4542089105,-0.1741423309,-0.173061192,0.1461533755,0.1153010204,-0.1177983508,0.1603147388,-0.1115378737,-0.3431859612,-0.2813438177,-0.072650142,0.0215305723,-0.1532002687,-0.1249315068,0.3240789473,-0.011987119,-0.2783163786,-0.2467869371,-0.2576208413,0.1958533973,0.1499989778,0.0449920706,0.0442360491,-0.0411367342,-0.1221293285,0.2578855753,-0.0548803248,-0.2784707248,0.2966708243,-0.3331163228,-0.3128942847,-0.3269756436,0.4821080565,0.2051306665,-0.1748419255,0.0102816578,0.1069050506,0.2002280056,-0.3558768928,0.0239392724,0.0264556743,0.0195045918,0.2280591875,0.4076539576,0.1806114614,-0.0915517434,-0.1612572819,0.0598357506,-0.1176615879,0.0223851353,0.0006820582,0.562587142,-0.2342285961,0.1041017696,0.4643881023,0.0835714787,0.1918973476,-0.1058519557,-0.1186733395,0.2647427022,0.0816802382,-0.0826890245,0.0555424951,0.0383592248,0.0832171962,0.0390605628,0.3929712772,-0.0885397047,0.0344258808,0.5620694757,0.1830459386,0.1093764156,-0.2962388992,-0.1132536456,-0.0964517668,-0.0015239341,0.0734219104,-0.2210583538,-0.3147691786,-0.3613390326,-0.080677107,-0.4869595766,0.1853488386,-0.0947150141,0.1995816678,-0.2643700838,0.6058086157,0.1335456669,0.1624041796,-0.2395661771,0.276404649,0.3889525235,-0.1408864856,-0.3185476065,0.3108559549,0.3091473281,0.3820615709,-0.4403640926,-0.1288670003,-0.0075539076,-0.2716631293,0.1383078992,0.038080018,0.1786568314,0.2999198139,0.3266952336,0.1114859283,-0.1785046607,0.2261472493,0.2892929614,0.0628218129,-0.2804661393,0.3225165009,-0.2205843329,-0.0536752194,-0.0161147323,-0.2147413492,-0.377214998,-0.128320381,0.2484835535,-0.079629004,0.0010218892,-0.0482817627,0.0747316405,-0.0962634087,0.3985402286,0.0026096113,0.5584189296,-0.2954240143,0.0728157461,-0.4015295804,0.2960191965,-0.3695742786,-0.0063605439,0.2887953222,0.3542706668,-0.016345391,0.1101768464,0.0010079899,0.2976796031,0.0083115529,0.4255388081,-0.3503302932,0.0875596404,0.1951819956,0.0048938668,-0.0131515246,-0.2848476768,0.1286351532,-0.4223635197,0.0601827092,0.1569891572,0.1705658883,0.1910827458,0.327172339,0.4762114584,-0.0726426914,-0.0076853614,0.1358398646,-0.0648116022,-0.3627668023,0.0060715177,-0.1850656122,0.2274803966,0.1393572241,0.4237051904,-0.1974121928,0.116933316,-0.14795959,-0.0916852206,0.0730620995,0.1957907379,-0.1012750044,0.3477635682,-0.4079725444,0.1093747243,-0.1123022884,-0.1291372627,0.0720417425,0.179442063,-0.166844517,-0.5900039077,0.7276579738,-0.2650013268,-0.5538643003,0.1625079513,-0.0311573669,0.3296854794,-0.0272085108,-0.3190827966,-0.0786864683,0.3554775417,-0.0164469201,0.0774405003,0.1284802258,-0.0175587814,0.0584818274,-0.1163030043,0.0996357128,-0.0530601293,-0.0559013113,-0.4643076062,-0.5050966144]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/885","title":"Very slow cold-start","comments":"Good point!","body":"Hi,\r\nI expect when importing `datasets` that nothing major happens in the background, and so the import should be insignificant.\r\nWhen I load a metric, or a dataset, its fine that it takes time.\r\n\r\nThe following ranges from 3 to 9 seconds:\r\n```\r\npython -m timeit -n 1 -r 1 'from datasets import load_dataset'\r\n```\r\n\r\nedit:\r\nsorry for the mis-tag, not sure how I added it.","comment_length":2,"text":"Very slow cold-start\nHi,\r\nI expect when importing `datasets` that nothing major happens in the background, and so the import should be insignificant.\r\nWhen I load a metric, or a dataset, its fine that it takes time.\r\n\r\nThe following ranges from 3 to 9 seconds:\r\n```\r\npython -m timeit -n 1 -r 1 'from datasets import load_dataset'\r\n```\r\n\r\nedit:\r\nsorry for the mis-tag, not sure how I added it.\nGood point!","embeddings":[-0.60203439,-0.3001388311,-0.1774234176,0.2726907134,0.1111319661,0.1556578726,0.2611881196,-0.0906468257,0.2295375764,-0.0899170861,-0.2638136148,0.0229987483,0.068844445,0.1952751428,-0.1143709198,0.1114253849,0.0574728698,0.0863008276,-0.2199211866,-0.1543319225,-0.340741545,0.0687089562,-0.4001384676,-0.2398431152,-0.0658351332,-0.0993470699,-0.0685147047,0.2544725239,-0.3526309133,-0.376944989,0.0211639535,0.155079335,0.0192186162,0.3430170715,-0.0001071061,-0.1745954901,0.2723818719,0.1438901275,-0.0478050262,0.0911347046,-0.081728667,-0.6443762779,0.1917965263,0.0202946831,-0.0348953195,-0.1638610661,-0.1892261654,-0.1703682542,0.104845129,0.0610222779,0.2250679731,0.4700413346,-0.302847743,-0.08919058,-0.1151185185,-0.2418320924,0.0668528304,0.2785087824,0.249446243,-0.1587840766,0.1959967166,0.1174623296,-0.372515738,0.3402546942,0.3903544247,-0.0238954239,0.1026232466,-0.3341920376,-0.0540977605,0.2898876667,0.5343538523,0.134877488,-0.115758568,-0.1831371486,0.0760414675,-0.6805512905,0.0228260066,-0.0056772484,0.131288588,-0.0198751055,0.0582145974,-0.0110369362,0.0528646596,-0.0011423543,-0.2775913477,0.3385930359,-0.0288685169,0.051985275,0.1572544873,-0.0750763118,0.2081601322,-0.3665187657,0.031879466,0.3300020695,-0.5311107039,0.0341973379,0.2136602402,0.3316475451,-0.1010920554,0.246525988,-0.1472571492,0.2328058779,0.068950817,0.0654722229,0.1522335708,0.1406521946,0.0875112787,-0.0912602395,0.4177553356,-0.2665183842,-0.0522785261,0.0691880509,-0.1856914461,-0.3749068081,0.3235384524,-0.1299705952,-0.0366474502,-0.1592369825,-0.2901844978,-0.0138446959,0.2563132942,-0.1209104061,-0.0087912958,0.3228690624,-0.3602033854,0.3206612468,-0.0384322107,-0.1048020422,-0.266526103,-0.0632444993,-0.2433300018,-0.0661533102,-0.0265030451,0.0530923046,0.040190652,-0.338942647,0.1716019362,0.0990846977,0.1348050386,0.1875940859,0.4104666412,-0.1541067511,0.0815631747,0.4263131618,-0.0213114358,0.2480795383,0.1807498634,-0.1439567953,-0.27300331,0.0460036136,-0.3834046721,-0.2424009293,-0.1539972723,0.2569669187,-0.3560239673,-0.068637073,-0.2940933108,0.1973431855,-0.3073172867,-0.1205872223,-0.3238448203,0.0067139138,-0.1379615963,-0.0079734353,0.2687847316,0.468662858,-0.5317761898,0.1369440258,-0.2238775343,-0.0670680404,0.2234587967,0.2385476977,-0.2472918779,0.2547321022,-0.0526052304,-0.0730651617,0.1281453371,-0.1686891466,-0.0688349903,0.1635387391,-0.2819069326,0.0453465506,-0.0629018098,0.3721788228,0.1544890404,-0.0119016515,0.409763962,0.4939131141,-0.0178617444,0.0213468876,-0.3975462317,-0.0582852587,0.0636473969,0.3491016626,-0.1340911835,-0.0605234839,-0.0509531423,-0.2053479999,-0.0331244543,0.1028529182,-0.2231173813,0.4321230352,-0.0209951457,0.1401133686,-0.052554816,0.0664725304,-0.2302241772,0.3152040243,0.5464602113,0.1842130274,0.1367646456,0.0725806281,-0.0662404373,0.1846736521,-0.2317476422,0.1756404489,0.1235978156,-0.121152699,-0.163933441,0.1547098309,-0.1007265896,0.3425988853,0.0473284125,0.0949370265,-0.3025710881,0.0715590343,0.1969492733,-0.0209174026,0.1915250421,-0.2429059148,0.0557637066,-0.2005761117,-0.0918993354,0.2914743423,0.1088660881,0.5422956347,0.1256845891,0.082670711,0.1357611716,0.0505247265,0.2161809653,-0.1420118213,0.3168997467,-0.3441056013,-0.1066438556,0.6420802474,-0.1319904625,0.1872759759,-0.1135331988,-0.1313971281,0.1552505046,0.2236092389,-0.2313645482,0.0867992342,0.3727408946,-0.05923694,-0.0064200028,0.2690320313,-0.1353455484,0.0266806334,0.389524132,-0.0542468466,-0.2578853369,-0.0617320612,-0.270008862,-0.2132203877,0.0980476215,-0.0203654133,0.5546646714,0.1948672682,0.0153918909,0.054210294,0.0195922647,-0.203476876,0.3438090384,0.1354932338,0.0063207559,0.1699090004,0.1534942985,-0.0021746599,-0.2089171559,0.0587440357,0.0586210676,0.3136481047,-0.1989046931,0.0702587217,-0.0539495163,0.1572200805,0.1783878803,0.0409980752,0.1257680506,-0.5635374784,-0.1650237292,0.0984390005,0.1954262108,0.0065944321,0.0205113012,0.1450737268,-0.0037121575,-0.2964527905,-0.1621313393,-0.349606514,0.1995120496,0.0986392498,0.1567856073,0.5048321486,0.4841914177,-0.3276697099,0.0138521884,0.1555716395,-0.3514622152,-0.1716386229,0.1325016022,0.3121512234,-0.1099068597,0.2440956533,-0.0007282505,0.0493158177,0.0639758706,-0.0344897695,0.0158580337,0.1498448998,-0.2967799604,-0.215517208,0.1514324397,0.0070063947,-0.2056843936,-0.2705686986,-0.0404320396,0.0361554995,-0.3085926473,0.0951288939,0.0305850171,0.194892168,0.1775690168,0.0470423996,-0.1469632983,-0.3508835137,0.6676402688,-0.3441329896,-0.2898915708,0.2434624285,0.0727175474,0.1023314819,-0.2380320579,-0.3350301981,-0.2578153312,-0.4791706502,0.2559577227,0.0530027188,0.1853966117,0.2087030113,-0.1021345183,-0.0774433538,-0.0038616972,-0.2509997189,-0.2294419706,0.0424399935,-0.2214907259,-0.2405650765,0.573915422,-0.3439781666,0.4527095854,0.0254198294,-0.0440301225,-0.0404899903,-0.0421784781,0.4969454706,-0.2332018763,-0.304456383,-0.1255338043,0.0612231009,-0.2627058923,0.237290591,-0.0171800349,-0.1901524067,-0.2785839438,0.0983834714,0.1541176885,-0.0520523749,-0.0624742061,-0.1377028674,-0.1642140001,0.2520971,0.0952316672,-0.351885885,-0.2829274237,-0.0007191646,0.2249854505,-0.1335535645,-0.2112915963,-0.3701244593,0.1797317713,-0.4448298514,0.1028090417,-0.1564099193,0.0369764455,0.1660514325,-0.115727596,0.2550345957,-0.1012000293,0.2483492643,0.0616876185,-0.2025478333,0.1617474407,-0.1769581735,-0.4268444479,-0.0970497057,-0.1382817626,0.1106836274,0.284476757,0.5335561037,-0.3332775235,-0.2585601807,0.0364414342,0.3765615225,-0.0501750261,-0.0005555699,-0.3749991059,-0.5280005932,-0.1813882142,-0.1260286123,0.1385634243,0.1456074119,-0.0267005619,0.0325953402,-0.2308302373,0.2180839181,0.4137191474,0.3343873024,0.1164789647,-0.0183127001,0.7292149663,0.2397995442,-0.0998789072,-0.1276174635,0.4910134971,-0.0340739973,-0.3517422974,0.0605401881,0.0396430492,0.2732017934,0.3979995847,-0.0237096548,0.0369685739,-0.286452949,0.3286936283,-0.3159598708,0.2070382088,0.3706466854,0.3109523356,-0.3729905784,-0.4796614349,0.2868121862,0.1296544075,-0.1244769394,0.5882565975,-0.0127740065,0.2048601508,0.1468917131,-0.1641554087,0.8067724705,-0.2577364445,0.2067988366,0.2670609057,0.0037959169,0.3509605229,-0.0747553408,-0.0391489938,-0.1782001853,-0.2750571072,0.2112232745,-0.0481331833,0.0846720412,-0.0752959922,0.0923591852,0.052941069,-0.0673592687,0.0748715624,-0.3300383687,0.4108041525,0.1153482497,-0.3128204048,0.0462052412,0.2583569288,-0.0579561368,0.0658795163,-0.1517101824,0.0258723572,0.1072246805,0.1746988147,-0.1244854778,-0.3572093546,-0.0032770466,0.1878542751,-0.0652331486,-0.2228443772,-0.0424180105,0.3162085414,0.3347946405,0.0726272389,-0.0309576225,0.3964682519,-0.1870466918,0.1145816296,0.0910525545,-0.1015668809,-0.0328878239,-0.2175671905,0.0852859765,-0.093218267,-0.1860541403,-0.0509766117,0.0270763189,0.1776986867,0.1641135812,-0.0729068518,-0.3710871041,0.2800578475,0.240159899,-0.0677358583,0.2080439329,0.3574302495,0.0629345551,0.5694745183,0.1248324886,-0.3255695105,-0.1812822372,0.2910829186,-0.0655668527,-0.1003309861,0.3448939621,0.1924186051,-0.4358762801,-0.0917264521,0.4713911712,0.0547657013,-0.1208446175,0.2527191043,0.2153061628,0.0033337839,0.1989400834,0.201775685,0.0008836906,-0.5157071352,0.046844393,-0.2409404069,-0.00010326,0.0700584203,0.0615363568,0.1766564548,-0.1086313725,-0.2008621544,-0.2077881992,-0.0040637492,-0.3790800869,0.2078917325,-0.1523995548,-0.0559451878,0.097472772,-0.1178605258,0.0978769362,-0.0584803149,0.1947587132,-0.090611808,-0.0924225897,-0.2954230011,-0.1414564252,0.1340321004,-0.2794909477,-0.3319440186,0.46550861,-0.3307195902,-0.3060807288,-0.1559122801,0.0689029172,0.1570551991,0.01924197,-0.201419428,0.0500908196,0.228697449,-0.0780724511,0.2869752049,-0.0498481058,0.3567430973,-0.2062717974,-0.0503806993,0.1815207005,-0.1240547001,0.0936692208,0.1058463156,0.0655746162,-0.0797557831,0.3062538803,-0.3578977585,0.2378913015,0.0729722381,0.2155153155,0.6081091762,-0.1232376695,-0.0527414978,0.1704931259,0.2288909107,-0.2227239907,-0.187622726,-0.0973760635,-0.0721872225,0.1067677587,-0.0358957686,0.148223713,0.2655563056,-0.0545147285,0.1165780947,0.4380982518,-0.0152248805,0.0415561087,0.1060593873,0.0162988752,-0.0697323456,0.4055394828,0.3843295574,-0.1309156269,0.5091711879,0.0495150052,0.3111949265,0.0172777791,0.3768233657,-0.2196886539,-0.3871998489,0.2867502868,0.1408707649,-0.3184437156,-0.0679019913,0.1126885861,0.1326077133,-0.4279852509,-0.1797191501,-0.2867845595,0.2800872922,0.0520098433,-0.0942455754,-0.0221423171,0.0574121624,0.0271224175,0.2476748079,-0.0826443881,0.1126899496,-0.0093697477,0.1205743253,0.1393701136,-0.1081193089,-0.0240195971,-0.2388944179,0.1634635329,-0.1587848812,-0.2806481719,-0.0048460579,-0.0555243678,0.0262458175,0.1443884969,0.4395475984,0.2318207622,-0.1235968322,0.1214884445,-0.1110588089,-0.313049525,-0.0495830886,0.3518740237,-0.1432612389,0.1365970671,0.1254636794,0.1697177738,-0.3748100996,-0.0902542844,-0.068692632,-0.2657524943,-0.3736445308,0.1810675263,-0.127317667,-0.0227318853,-0.3927401602,-0.0117733236,-0.1561660171,-0.1519793868,0.358305335,-0.118689917,-0.0511127338,-0.1823003739,0.1001808196,-0.1424398273,0.4733500183,0.2354709655,0.2415966988,-0.1804796755,-0.0970067084,-1.065408349,-0.0985288993,-0.0603331178,-0.0220126212,0.1146084219,0.1548025161,0.1641484499,0.339166671,-0.1551431865,0.1788660437,-0.0621203072,0.0126708364,-0.3222080469,-0.0697570518,-0.1623311341,-0.071921967,-0.1044153869,-0.2425858825,0.1426813006,-0.3087442517,0.1714071184,-0.271376282,0.0284274705,0.1419676393,0.5104142427,0.31175825,0.0791208297,0.3100606501,0.2060794234,0.1359825581,0.0371276028,-0.2613273263,-0.0055380245,-0.0519541316,-0.0737708062,-0.0141279632,-0.0369744264,-0.3154831529,0.0888263732,0.0629786402,-0.1721173227,-0.1114706919,-0.3681301177,0.1242327094,-0.0466473959,0.3660989702,0.1134794056,0.3256784379,0.0707094073,-0.1045493484,-0.1514672488,-0.1625346541,0.46764642,-0.2492280751,-0.1270369738,-0.1879206598,0.0671189129,0.2954213023,0.1068210825,-0.0475036204,0.0641502216,0.1877527684,-0.1741340458,-0.2356522083,0.6979936957,0.2165116668,-0.1069486737,-0.1529236287,-0.0912762806,0.05877259,-0.2743103206,-0.2917266488,0.0359083973]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/885","title":"Very slow cold-start","comments":"Yes indeed. We can probably improve that by using lazy imports","body":"Hi,\r\nI expect when importing `datasets` that nothing major happens in the background, and so the import should be insignificant.\r\nWhen I load a metric, or a dataset, its fine that it takes time.\r\n\r\nThe following ranges from 3 to 9 seconds:\r\n```\r\npython -m timeit -n 1 -r 1 'from datasets import load_dataset'\r\n```\r\n\r\nedit:\r\nsorry for the mis-tag, not sure how I added it.","comment_length":11,"text":"Very slow cold-start\nHi,\r\nI expect when importing `datasets` that nothing major happens in the background, and so the import should be insignificant.\r\nWhen I load a metric, or a dataset, its fine that it takes time.\r\n\r\nThe following ranges from 3 to 9 seconds:\r\n```\r\npython -m timeit -n 1 -r 1 'from datasets import load_dataset'\r\n```\r\n\r\nedit:\r\nsorry for the mis-tag, not sure how I added it.\nYes indeed. We can probably improve that by using lazy imports","embeddings":[-0.5925875902,-0.2534993887,-0.1908845454,0.2816770971,0.1464059055,0.1379763037,0.2646388412,-0.0451315045,0.2316522449,-0.1083664224,-0.2794908285,0.0491926111,0.0483701266,0.214217186,-0.1124559343,0.0953868181,0.065960817,0.0731302351,-0.212594673,-0.1679442227,-0.3407582343,0.0477048345,-0.4150152504,-0.2240625918,-0.060381081,-0.0978036076,-0.0657732263,0.202290222,-0.3653015196,-0.3793236613,0.0423676334,0.1579445004,0.0184304249,0.3116640747,-0.0001061617,-0.1743324548,0.3004815876,0.1214670017,-0.0744866058,0.0857710391,-0.0885937214,-0.6683532,0.1768217832,-0.0395070389,-0.0440437309,-0.1993699074,-0.1967883408,-0.1833809465,0.1255805194,0.0674901307,0.2435587645,0.4415156245,-0.2615938187,-0.0655341297,-0.0896932259,-0.2311548144,0.0571453869,0.2664393187,0.2409889549,-0.1475556344,0.1836443841,0.1103624031,-0.3705660701,0.2956441641,0.4018332064,-0.0165391564,0.132823512,-0.317417413,-0.0946289822,0.2774747908,0.5253145695,0.0923311859,-0.170176208,-0.2294833511,0.1210832745,-0.6835532784,-0.0048043174,0.0068856874,0.1179233268,0.0311148874,0.0655342489,-0.0566635951,0.0201267805,0.0058387453,-0.2364623547,0.2983241081,-0.0388645977,0.0477611721,0.1439597011,-0.0719143227,0.2540329695,-0.3888659179,0.0431908742,0.3026014566,-0.5251953602,0.0302097723,0.1942340136,0.3470771015,-0.101720579,0.2482113838,-0.1484954953,0.2446787208,0.0582745038,0.0861781538,0.1918877214,0.1574894935,0.0635531172,-0.0696862713,0.4129786193,-0.2564128935,-0.0528333075,0.0941038355,-0.1690013856,-0.4061451256,0.3210658431,-0.0878913477,-0.0455273688,-0.1071170494,-0.2704916,-0.0207518339,0.2502713203,-0.1518703699,-0.0318032578,0.2971466184,-0.3594429195,0.3434172571,-0.0574341714,-0.0925567746,-0.2434558272,-0.1033649966,-0.2510086298,-0.0321275964,-0.0110292817,0.0552634858,0.0121642221,-0.3473739028,0.1512191892,0.0676746219,0.1308228076,0.1897843182,0.4134910405,-0.1786821634,0.0473913215,0.4175024033,-0.0306906346,0.2030414641,0.1951648444,-0.1530443132,-0.3201892078,0.018205801,-0.3704948425,-0.2572314143,-0.1247038022,0.2694185078,-0.37645486,-0.0577565469,-0.3390626013,0.1959043443,-0.282798171,-0.0894269943,-0.3268613517,0.0183837879,-0.148791492,-0.0161565989,0.2568677068,0.4654300809,-0.4893730581,0.1240103245,-0.1980686635,-0.0899429023,0.2029488236,0.2026887387,-0.2614350319,0.2778733671,-0.0700312555,-0.1274839789,0.1573629081,-0.1748062074,-0.0341285802,0.1654344648,-0.2694562376,0.0547982529,-0.0142684765,0.3330360651,0.1879055053,-0.0295884106,0.4072860777,0.4998137653,-0.0391921774,0.0319456048,-0.3734217286,-0.0831274167,0.0911686271,0.3653967977,-0.118959561,-0.0597677901,-0.0375708565,-0.1950114518,-0.0355052613,0.0988006517,-0.212468341,0.41240713,-0.003849688,0.1347583085,-0.0679128319,0.050239224,-0.2736624479,0.3026242554,0.5202704668,0.1680863053,0.125649333,0.0209981613,-0.0586694032,0.1670411825,-0.2242581993,0.2025723904,0.1531952471,-0.1378941983,-0.1771403998,0.1572544575,-0.1201050058,0.3725882173,0.0238470118,0.0708577186,-0.284291029,0.0896958485,0.1770458519,0.0408891961,0.1782681048,-0.199075371,0.0529161245,-0.1955977678,-0.1123791188,0.2610100806,0.1215333641,0.5515108109,0.1432825625,0.1190087348,0.1447596699,0.0580319948,0.1770840883,-0.1443289071,0.3020992577,-0.3517437577,-0.0843629166,0.6781523824,-0.1289116591,0.213531509,-0.0817453042,-0.113296032,0.1987755299,0.2133754194,-0.248799026,0.0836772621,0.3240392506,-0.0475979,0.0242174584,0.2852467895,-0.1294702142,0.0250923317,0.3594790995,-0.052504044,-0.2396522611,-0.0372226574,-0.2304405272,-0.2070433497,0.1244242713,0.0076669063,0.5331953764,0.2225809991,-0.0166741144,0.0878755376,0.0581914783,-0.1747906059,0.3739052713,0.1645681113,0.0347997248,0.1721442789,0.1227868274,-0.0074970881,-0.2421823591,0.0485069118,0.0821972787,0.324637115,-0.2157706767,0.0660825223,-0.0447913595,0.1443436742,0.1359424889,0.040012721,0.1642837822,-0.5571197271,-0.1563406289,0.0901294947,0.1396428645,0.0330740735,0.0138185285,0.1919759363,-0.0284395926,-0.3080407381,-0.1676023602,-0.3505666554,0.1971523315,0.1037878767,0.1823804229,0.5026218295,0.469451189,-0.3358836174,0.0264197197,0.1116977781,-0.3518889248,-0.1414852738,0.1326482743,0.3126125932,-0.0615626164,0.1816811711,-0.0065628025,0.0365100726,0.0412196405,-0.0457442552,0.0051954035,0.1364577711,-0.3296897709,-0.1990407258,0.1338093579,-0.0769754052,-0.2318758965,-0.3141288757,-0.0507811978,0.0037525936,-0.2978521585,0.090750128,0.0725199804,0.1553863585,0.1506896317,0.0667313039,-0.111037083,-0.3732533157,0.6770269275,-0.3247504234,-0.2904395759,0.2102508545,0.0423700996,0.1149249151,-0.2528562844,-0.3326010108,-0.2827486992,-0.4079503119,0.2651265562,0.0597687364,0.2134715021,0.1594474912,-0.0828729123,-0.0744379163,0.0284717642,-0.2034832686,-0.2313767374,0.012953192,-0.2481335253,-0.2382221371,0.5572842956,-0.3204289377,0.4273020327,0.0345223211,-0.0243859552,-0.0145630967,-0.0296033826,0.5247789025,-0.1791105717,-0.3119039536,-0.1373861283,0.036365997,-0.2915403545,0.241137296,-0.0389710516,-0.1565484703,-0.2748866379,0.0822249874,0.1416663229,-0.0735094696,-0.0608638898,-0.135316819,-0.1299604177,0.2403079271,0.0792461559,-0.3531514406,-0.2424631417,-0.0057256604,0.2029283196,-0.0838643983,-0.198043406,-0.3530620039,0.1821668595,-0.4472873211,0.0959905759,-0.1332827359,0.0469428115,0.2251823843,-0.104558371,0.212950632,-0.0803296641,0.224345237,0.0791726857,-0.2390847802,0.1591321528,-0.2053531706,-0.4069598615,-0.0775570869,-0.1119056791,0.1123760641,0.2967158258,0.5045583248,-0.3925017118,-0.2595597208,0.0398184657,0.4059911966,-0.0730023831,-0.0381918587,-0.3471061885,-0.5230740309,-0.2212367803,-0.1270320565,0.1606568545,0.1446753144,-0.0320632905,0.0412970968,-0.2361859232,0.2400386631,0.3987216651,0.3066966534,0.1213483959,-0.0237548891,0.7368592024,0.2122177631,-0.1062500253,-0.1163738519,0.5154114962,-0.0879944414,-0.3945025206,0.0097256992,0.0489322767,0.2923960984,0.3657761514,-0.0466332883,0.0436516367,-0.2978369594,0.2986538112,-0.3132973313,0.1909629852,0.3489670157,0.2924430668,-0.3369265795,-0.5218687654,0.3178559244,0.1301282793,-0.1168417111,0.5345883965,0.0164060723,0.1915544271,0.1796200871,-0.1294951886,0.7814317346,-0.2475083023,0.2058160454,0.2683568299,-0.0056805811,0.3896656632,-0.1359904706,-0.0662906393,-0.1916659623,-0.2746461332,0.2144646943,-0.0406115577,0.09005218,-0.0517657697,0.1075814068,0.1183499917,-0.0367291085,0.0440372229,-0.3171846867,0.4147066772,0.1296007186,-0.3316685855,0.0506476834,0.2602449954,-0.0535937995,0.0479897,-0.151949659,0.0215019528,0.1355526149,0.1517442763,-0.1136551425,-0.3823691607,0.0136453034,0.2082216889,-0.0578834116,-0.2127549797,-0.0109164286,0.3057300746,0.3449802995,0.0798466429,-0.0344980173,0.3663045466,-0.1638701707,0.1472089291,0.1022965685,-0.0981960595,-0.0128784236,-0.2068327665,0.0736038387,-0.0803165063,-0.1855614632,-0.0916898474,0.0149937952,0.1492569447,0.1765592247,-0.0621736161,-0.3654595017,0.3004775941,0.2206166536,-0.1166512817,0.2239808291,0.3777399957,0.0421457365,0.6064774394,0.1231036782,-0.323815465,-0.1971776038,0.2970294058,-0.0473271348,-0.131207481,0.3219345808,0.1777614504,-0.4358015954,-0.1143908575,0.4583107829,0.0607700944,-0.1392610073,0.228698507,0.2205284685,-0.0128062479,0.1397704333,0.2014525086,0.0248464495,-0.4576748013,0.0588647276,-0.2094558477,-0.023969749,0.101610817,0.090307802,0.1625692844,-0.0474003889,-0.2200209349,-0.2481890917,0.0132933687,-0.3941420317,0.1752072275,-0.1507738829,-0.0992513001,0.0945874974,-0.0932875201,0.0706326738,-0.0482944511,0.2208076119,-0.1015319228,-0.0892424285,-0.3083395958,-0.1503976285,0.1282693148,-0.259552896,-0.2601573765,0.4489615858,-0.3261109591,-0.327224642,-0.1675859541,0.0160971303,0.1933513582,0.0130519792,-0.173225984,0.0260184743,0.1735249162,-0.1173996031,0.2708916962,-0.0179770757,0.370570302,-0.2282685041,-0.0397137515,0.1767336875,-0.1151811928,0.1004157886,0.0699763745,0.0732898787,-0.0543221757,0.341368258,-0.3302643299,0.2424525768,0.1233799085,0.204967916,0.6053807139,-0.1557853371,-0.0217632502,0.2031158358,0.2352588475,-0.280262351,-0.1993740648,-0.1154573634,-0.0789682344,0.1322929263,-0.0177552421,0.1669326574,0.2686676085,-0.0351455733,0.112371996,0.4310762584,-0.0015404022,0.0502676032,0.1057326123,0.025453588,-0.0871459171,0.3975332975,0.3654499352,-0.1003664061,0.5127022266,0.0369189382,0.3305672705,-0.0090368018,0.3658376932,-0.195587337,-0.3651329875,0.2994041741,0.1456952691,-0.2691249549,-0.0716580302,0.1232441291,0.1281138361,-0.3738012612,-0.1699070781,-0.2836357057,0.2531766891,0.0361432657,-0.0610268936,-0.0016364575,0.0889811888,0.0335742831,0.2417921424,-0.0871280134,0.0909363702,-0.0193141904,0.1488876939,0.1411452144,-0.1107941568,-0.0178764258,-0.2295384556,0.1915459931,-0.1595290601,-0.2794153988,0.0151488511,-0.0795058459,-0.0281960368,0.1772700101,0.412812978,0.2465439439,-0.1117185429,0.0928735733,-0.1066131741,-0.2714257538,-0.0650298223,0.3095237315,-0.1613773406,0.1028870717,0.1291127205,0.1746082157,-0.3831197619,-0.0534720346,-0.0639893934,-0.2899237871,-0.3825587928,0.2428262383,-0.0737441257,-0.029371243,-0.3927705586,0.0147798983,-0.1602503806,-0.1434235424,0.3721470237,-0.1219509616,-0.0553529523,-0.2089991868,0.1054223999,-0.1324262619,0.4810526669,0.2211628407,0.257035166,-0.1930351108,-0.0696993247,-1.0758597851,-0.0655327365,-0.0786400139,0.0077684862,0.1661422849,0.1098765656,0.18001993,0.3877029419,-0.1688269526,0.1857691109,-0.0358013324,-0.0024828468,-0.3438432515,-0.1121108457,-0.1539691836,-0.0688699335,-0.0871139243,-0.2810838521,0.1593410522,-0.3072564006,0.1815533638,-0.27219069,0.0589229241,0.1561419815,0.4925622344,0.3242911994,0.0746162459,0.2725034654,0.1690384299,0.1639840901,0.0370144807,-0.2876471579,-0.0294668451,-0.0266166404,-0.0912362486,-0.0553544275,-0.0554515943,-0.3327808678,0.066861406,0.0287124626,-0.1652916223,-0.1274882406,-0.3973036408,0.1539572179,-0.0652680621,0.3583585322,0.0800907984,0.3224331141,0.055702541,-0.1021420732,-0.1466718018,-0.1693280339,0.4715800881,-0.2293431014,-0.1210393161,-0.1957930177,0.0917303488,0.3308492601,0.1014412493,-0.0317102298,0.0850085393,0.1899064183,-0.1398709416,-0.2147209644,0.6773117781,0.2192211747,-0.0840151906,-0.1424789131,-0.0780890137,0.0264874157,-0.2549606562,-0.279522419,-0.0037353192]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/885","title":"Very slow cold-start","comments":"#1690 added fast start-up of the library ","body":"Hi,\r\nI expect when importing `datasets` that nothing major happens in the background, and so the import should be insignificant.\r\nWhen I load a metric, or a dataset, its fine that it takes time.\r\n\r\nThe following ranges from 3 to 9 seconds:\r\n```\r\npython -m timeit -n 1 -r 1 'from datasets import load_dataset'\r\n```\r\n\r\nedit:\r\nsorry for the mis-tag, not sure how I added it.","comment_length":7,"text":"Very slow cold-start\nHi,\r\nI expect when importing `datasets` that nothing major happens in the background, and so the import should be insignificant.\r\nWhen I load a metric, or a dataset, its fine that it takes time.\r\n\r\nThe following ranges from 3 to 9 seconds:\r\n```\r\npython -m timeit -n 1 -r 1 'from datasets import load_dataset'\r\n```\r\n\r\nedit:\r\nsorry for the mis-tag, not sure how I added it.\n#1690 added fast start-up of the library ","embeddings":[-0.602571547,-0.211007461,-0.151371941,0.3197123706,0.1022797972,0.1545116007,0.2709058821,-0.0459072515,0.2117199302,-0.134920761,-0.2209196091,0.0993037522,0.0299887769,0.1454749256,-0.0819741637,0.0898783058,0.0883857012,0.1213310137,-0.1827077866,-0.1224152148,-0.2441505492,0.1657155454,-0.4087193608,-0.2181066126,0.0550400354,-0.0865702778,0.0383734368,0.2148941606,-0.3697673976,-0.4055235684,0.0364410281,0.1578925252,0.0051567894,0.2761823535,-0.0001103514,-0.2184170932,0.3041707873,0.1524942368,-0.0731763095,0.0261609685,0.0289134216,-0.6502453089,0.2010892332,0.0304943155,-0.0188538264,-0.116281651,-0.2276204973,-0.2381216288,-0.0809081793,0.0136293769,0.1896730959,0.4453216195,-0.3051299751,0.0059765461,0.0177641883,-0.2047889829,0.0469535887,0.2333336025,0.2930360734,-0.1300432384,0.214069277,0.095975019,-0.4504042864,0.2660440505,0.4658872187,-0.0326749608,-0.0170801766,-0.2556320131,-0.0455167554,0.3477578163,0.5974785686,0.0448746011,-0.074109368,-0.1728643924,0.0659376308,-0.6625239849,0.0556617267,-0.0222687609,0.186950013,-0.059675619,0.0870027468,-0.0495181493,0.0780406147,0.0493710041,-0.1248856932,0.3651601076,-0.038179215,0.1000888273,0.2592972219,-0.0812064335,0.2238399833,-0.343611449,-0.047764603,0.379343152,-0.4913768768,0.0338852555,0.196609661,0.3342951238,-0.0322562791,0.3031908274,-0.2107491344,0.2228454947,0.0558112822,0.0386811532,0.1658383906,0.1444265991,0.0957567096,-0.0653790012,0.4894381762,-0.199372679,-0.0161814336,0.0811965242,-0.195545733,-0.382363826,0.2997480035,-0.1021356955,-0.0837663114,-0.1909400225,-0.2448705286,-0.0149357896,0.2252373546,-0.1259168237,0.0898187086,0.2942535579,-0.4263310432,0.351290375,0.0094356975,-0.2094779611,-0.2862427235,0.0165541414,-0.1880870014,0.038553264,-0.027737068,0.0855263248,0.0804195032,-0.3941794336,0.1787950248,0.0690699443,0.1496739835,0.1827543974,0.3997801244,-0.1837646812,0.0968490392,0.4454523027,-0.1219945699,0.3247007132,0.1562653929,-0.1563398838,-0.3915389478,0.0324178711,-0.413818419,-0.2526403368,-0.1885108352,0.1981045008,-0.3492929637,-0.052202709,-0.3377743363,0.2034139782,-0.2970047295,-0.2027387321,-0.2495371848,0.0300909206,-0.1194850504,-0.037912719,0.2686445415,0.4742113054,-0.5602862835,0.1305045933,-0.3237309158,-0.0536415726,0.2156618088,0.1795023531,-0.309353441,0.2314313501,-0.1014679894,-0.0888464674,0.1760209054,-0.2454822958,-0.1185000092,0.20029369,-0.26831159,0.0132934609,-0.0841483101,0.5086934566,0.0917161852,-0.0249935295,0.3442605436,0.5224611759,0.0019452479,0.0648834258,-0.3354314268,-0.1146101207,0.1588848382,0.3503253162,-0.1662995368,-0.1156443954,-0.0639003962,-0.1015596613,0.033418484,0.1542697698,-0.2036692202,0.3878430128,0.0420627519,0.1543320864,0.0012390498,0.0584604256,-0.1774624735,0.3168016076,0.6126400828,0.1744609624,0.1332306415,0.0702076331,-0.0289812237,0.2169528306,-0.1794779152,0.1719321758,0.069942005,-0.1443098485,-0.2355839908,0.2001042664,-0.133084625,0.3819099367,0.0142849376,0.073684983,-0.4076305926,0.0477952808,0.2445872277,0.0154439155,0.1947240829,-0.2395606339,0.0265864432,-0.2114695013,-0.1482853293,0.2994533777,0.1203082427,0.5419535041,0.072860904,0.1017299145,0.1552065909,0.0397973508,0.2307513952,-0.1869127601,0.3371538818,-0.3244896233,-0.0353003144,0.6556994319,-0.2090411484,0.1605956256,-0.1009838134,-0.1963698864,0.1224210784,0.2426916063,-0.2325881869,0.1193887591,0.4414366782,-0.0592895672,0.0669571608,0.266579628,-0.1082887799,-0.0086445818,0.4239743054,-0.0465735234,-0.3139166534,-0.0304405205,-0.231829077,-0.2016129643,0.145744279,0.0021507849,0.5909540653,0.1374234557,-0.0819740295,0.0263348315,0.0557249896,-0.1853698939,0.3417536914,0.1191203743,0.0261071678,0.1271059811,0.1392278075,0.006467096,-0.2197554708,0.0020573437,0.0519850925,0.3009493947,-0.1711178273,0.1160180941,-0.0430218801,0.0901157185,0.1663554311,0.0325561613,0.1178978458,-0.4830349982,-0.1986387819,0.1155933291,0.2795440257,-0.0126924329,0.062590003,0.0901667699,-0.0721614212,-0.4371313453,-0.2350986153,-0.3359422684,0.1807667613,0.058373753,0.1674613059,0.55694592,0.4753953218,-0.2997868061,-0.0133090243,0.0860642046,-0.4264891744,-0.2077337056,0.1318716258,0.3212265968,0.0228063688,0.192919001,-0.0154154049,0.0353053212,0.0213336088,-0.0447382592,0.0593291037,0.1050857455,-0.3856213391,-0.2057312876,0.1393604875,-0.0207841955,-0.1895751208,-0.249840796,0.0210419931,-0.0230513923,-0.2947294712,0.0955823958,-0.0476959608,0.2170064151,0.1943314373,0.0720428377,-0.1486469209,-0.3670857549,0.5977367759,-0.3132078946,-0.2381418496,0.2604372501,0.0798457116,0.1309761852,-0.2863627076,-0.4329994917,-0.1942496598,-0.4834936559,0.258653909,0.027351588,0.1744901687,0.181731388,-0.1299026012,-0.0076593626,-0.0182468817,-0.3560900688,-0.2389338017,-0.0053519825,-0.1787311137,-0.1928805411,0.56616503,-0.39475137,0.4639852643,-0.0155391628,-0.0692958459,-0.0394339189,0.0042859316,0.5095997453,-0.1894470304,-0.3374353051,-0.0735797137,0.1065206081,-0.2981089354,0.2369118631,-0.0024013394,-0.2425149381,-0.3083766401,0.0791329518,0.2155647725,-0.048032701,-0.0932584852,-0.167838946,-0.2658427656,0.2241580188,0.101901643,-0.3928024173,-0.3408774137,0.0363534801,0.3432288468,-0.1667483002,-0.1681527942,-0.3419036269,0.0736242384,-0.4481399357,0.048656486,-0.1529129595,0.0383855961,0.1878057122,-0.1018706635,0.2800030708,-0.1097937599,0.2408899367,0.1185790002,-0.2372305691,0.231599316,-0.1931909621,-0.3364624977,-0.1001290381,-0.0526327975,0.1401833892,0.2535772026,0.5181686878,-0.3703635633,-0.3176739514,0.0010665057,0.3977069557,0.0046710614,-0.0931895748,-0.3443008065,-0.5439814925,-0.1714048237,-0.1508938819,0.1068549305,0.1494020373,-0.0101554571,-0.0195736736,-0.1713344753,0.2293621004,0.4312007725,0.2777052224,0.126251936,-0.0314278193,0.7831503749,0.1541159451,-0.0607645139,-0.1577389091,0.4004118145,-0.0785242915,-0.2891189456,0.0259881504,0.1384990811,0.3096545041,0.3592306077,-0.0847017914,0.1554333866,-0.2619049251,0.2596198022,-0.2925323844,0.1989690959,0.3572862148,0.2712266743,-0.3927634954,-0.5481680036,0.3665040731,0.2001556605,-0.1234103367,0.7008314729,-0.0445305891,0.2480921745,0.1001345068,-0.1137090027,0.8527988195,-0.3117338717,0.2037096471,0.2453543991,-0.0066594766,0.5185530186,-0.1289287359,-0.046379216,-0.2939523757,-0.2382114828,0.2090293765,-0.0720509067,0.0925909281,-0.1514421999,0.0984491259,0.0538494103,-0.0404227413,0.0385870561,-0.3994137347,0.475055486,0.0519873239,-0.3578966558,-0.007327666,0.2098941356,-0.1050000861,0.1192438975,-0.1535383314,0.0279945284,0.0900728703,0.1356311142,-0.1457735747,-0.3684045374,0.0078066778,0.1711133569,-0.0174298286,-0.2281907499,0.0483424217,0.2959441543,0.3394856453,0.1466648877,-0.0332572609,0.4172423482,-0.2411818057,0.0984830782,0.1410487145,-0.0987650007,-0.0677845329,-0.2016370296,0.1069861501,-0.1084092408,-0.2224805951,-0.050932955,0.0564636327,0.176627472,0.269562155,-0.0101063438,-0.3807669878,0.3147930503,0.254912287,-0.0372239724,0.1685167849,0.4544308782,0.1048802361,0.5576230288,0.0829145312,-0.3571509719,-0.1812167913,0.2696716487,-0.043594785,-0.1401346326,0.2849886715,0.2418843806,-0.4188238978,-0.0274310429,0.4618595839,0.0551936924,-0.1639839858,0.247304678,0.2780380249,-0.0302170832,0.269937247,0.1918000579,-0.0895359516,-0.4548493624,0.0199171323,-0.2280524969,-0.05917437,0.0675776824,-0.0405442603,0.1588089168,-0.0284826159,-0.2652177811,-0.2643187642,-0.0143432394,-0.3190997541,0.2370579392,-0.160147801,-0.0796381682,0.0379643813,-0.0509256683,0.0915654302,0.0146899745,0.1578222066,-0.0510049723,-0.0877792761,-0.2352872044,-0.1746252775,0.166529581,-0.265517354,-0.2977370918,0.4559206069,-0.389313668,-0.3188762665,-0.1790541261,0.0900600329,0.1677060723,0.0398259871,-0.2473649681,-0.0005577523,0.1712877899,-0.1480482519,0.3486731052,-0.0657163039,0.345710218,-0.2075016797,-0.022876909,0.1922330111,-0.1092559546,0.0640378296,0.0631913692,0.0827209949,-0.0991599038,0.32936728,-0.2550427318,0.2113534063,0.092956841,0.185723111,0.6067054868,-0.0870661139,-0.0421088748,0.2106573582,0.1590953171,-0.1777430773,-0.2338440418,-0.1752177775,-0.102182284,0.0363200791,-0.0216319915,0.1533679813,0.291190654,-0.0841875374,0.0751623958,0.5047087669,0.0368509106,0.0133718178,0.1685574502,-0.0239439886,-0.0653523207,0.3846467435,0.37406528,-0.0948247537,0.5020701289,0.0400639325,0.3288763165,0.0730920732,0.4250144064,-0.2787449062,-0.3514546454,0.2537782192,0.2278036028,-0.2702299058,-0.0875164717,0.1376724839,0.2448903918,-0.402949363,-0.1130218133,-0.231855765,0.3073064387,0.0389543884,-0.0474421121,0.0923039988,0.1023484766,0.022566827,0.2238012403,-0.1131879315,0.1670860499,0.0148119023,0.1161490232,0.1273556501,-0.1602743864,-0.0284266807,-0.2444956899,0.1179583147,-0.1578824967,-0.2971569002,-0.0956543684,-0.0683346242,-0.0186130516,0.1085612029,0.3543396294,0.2641469538,0.0078939097,0.0414982587,-0.0907191187,-0.2239303738,0.0377083831,0.3460166752,-0.0992773324,0.1818576157,0.0964610055,0.1117104664,-0.3139173985,-0.079786323,-0.1091104671,-0.2525316179,-0.3587174714,0.1884839088,-0.1468092054,0.0259712841,-0.4306676388,0.014845334,-0.105295524,-0.2080581039,0.357642144,-0.0739327669,-0.0778264776,-0.240304023,0.0796272159,-0.1438844353,0.4281851351,0.1988985837,0.252666086,-0.159515962,-0.1640518606,-1.0985774994,-0.0735193044,-0.0504138395,0.0297184866,0.1209196672,0.0790633261,0.150608182,0.2584552169,-0.1747710258,0.1503992826,0.0257961545,-0.0482591763,-0.3066459298,-0.1403674334,-0.2334703356,-0.1142414287,-0.1282785237,-0.1777509898,0.0719250366,-0.2821331322,0.1015777886,-0.3859752715,0.0773281455,0.1675958633,0.5630706549,0.3486433923,0.0450865701,0.2577917278,0.2160315365,0.0931226537,0.0635947287,-0.2505141795,0.0199555997,-0.0680502951,-0.1076423973,-0.0123104211,-0.0158132277,-0.2051880211,0.030645771,0.0926611125,-0.2150408179,-0.0804366246,-0.4490271509,0.1128825769,-0.1288319826,0.3945900202,0.1139233336,0.3476682305,0.1206391603,-0.104387641,-0.1895102262,-0.1286709905,0.5081007481,-0.2645910978,-0.0774770901,-0.1380358934,0.0104184905,0.2829547524,0.1033336669,-0.1215137616,0.0407907963,0.1751049757,-0.1051948816,-0.2129803151,0.6788527966,0.2015587389,-0.0345855951,-0.1348697394,-0.1009135172,0.0732375085,-0.2196242809,-0.201125592,0.0549074747]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/883","title":"Downloading\/caching only a part of a datasets' dataset.","comments":"Not at the moment but we could likely support this feature.","body":"Hi,\r\nI want to use the validation data *only* (of natural question).\r\nI don't want to have the whole dataset cached in my machine, just the dev set.\r\nIs this possible? I can't find a way to do it in the docs.\r\n\r\nThank you,\r\nSapir","comment_length":11,"text":"Downloading\/caching only a part of a datasets' dataset.\nHi,\r\nI want to use the validation data *only* (of natural question).\r\nI don't want to have the whole dataset cached in my machine, just the dev set.\r\nIs this possible? I can't find a way to do it in the docs.\r\n\r\nThank you,\r\nSapir\nNot at the moment but we could likely support this feature.","embeddings":[-0.1721919328,-0.0785837397,-0.1101517826,0.1496574581,-0.0266026482,0.0321875773,-0.009482801,0.3326726556,-0.3031471968,0.1801902801,-0.1334840208,-0.1585270464,0.1139857098,0.3811725676,0.1864495277,0.2583482862,-0.0862012357,0.3838580549,0.1213866174,0.1218435839,0.1287989169,-0.2711049616,0.1405358613,0.1316489726,-0.1419464797,0.0805708244,0.2840859592,-0.1750894934,-0.3893735111,-0.2463962287,0.5170015693,0.1349228173,0.0742571056,-0.387973249,-0.0001252517,-0.0412728637,0.1077560484,-0.3247351646,-0.1863010824,0.1280506253,-0.5343056321,0.2603619695,-0.1771430671,-0.0470350832,-0.4029557109,0.1003951877,0.1171863973,-0.1273756027,0.4954646528,0.3370750546,0.0640953332,0.2468543649,-0.2601716518,0.1065130234,-0.4099748135,0.0921439007,-0.0464430787,0.1984981,0.3917531371,0.1116437018,0.0816467404,0.317121923,0.1741402596,0.4878472984,0.4795396328,-0.1241252422,0.1041153371,-0.5178179741,0.4865063429,0.3966907859,0.8627646565,-0.2840754688,-0.1959988773,0.0505747423,-0.0946474299,-0.3427264392,0.0241497364,0.3780059516,-0.3084818125,0.5162419081,-0.5401997566,-0.395622313,-0.3904267251,0.0190462861,-0.1720171124,-0.0871438235,0.0105412295,-0.1455799192,0.1530164778,0.1460443437,0.418779254,-0.0477824919,-0.337402612,0.224400267,-0.1201153547,-0.4483332336,-0.5021434426,0.0908239707,0.2265194356,0.1798329055,-0.1862124801,0.2354723811,0.066235669,0.0698804557,0.3200733066,0.1411610246,0.423417151,0.3131199181,0.1075624004,-0.4265489578,0.1067730859,-0.0410148613,0.0982689112,0.1274333,-0.2020550519,0.2316394746,-0.1456046402,-0.5380343199,-0.0343383029,-0.2452111542,0.0420579463,-0.1104858741,-0.0258755907,0.4035154283,-0.1675199717,-0.1550711691,-0.4510666132,0.18785128,0.0630291477,-0.7456681728,-0.1019594744,0.1776832491,-0.0708105937,0.1799603701,0.3832868934,-0.2576662302,0.2517768741,-0.1807743162,-0.1463032216,-0.0594721809,0.3550841212,-0.319460243,0.3798952103,0.3150658607,-0.1114131436,0.1592617929,-0.0649160743,0.0124246636,-0.2758476436,0.1934392601,-0.0832060874,-0.5966411829,0.0895427242,0.04194602,-0.3469689786,-0.0039182813,0.0126517471,0.2906639576,0.2063959539,-0.0802086219,-0.0584508702,0.1671793014,-0.1164166853,0.0871850327,0.2228513807,0.0954848826,-0.3683046699,-0.0803324729,0.0525998622,-0.1591074914,0.0491105802,0.2000938803,-0.2829194367,0.2232723981,-0.043890696,0.1213945374,0.5505468249,-0.053583432,-0.333812207,0.0630787835,0.0643922687,-0.4127536118,0.0302952956,0.1457807273,0.3395308852,0.0904299021,-0.4790005088,0.5126757622,0.0085213706,-0.1791003048,-0.0845978931,-0.0884876102,-0.0144575983,-0.0377419665,0.3669629693,0.5608133078,0.1335727125,0.1593114883,-0.0525063537,-0.1787158549,0.2388892621,0.2362822145,0.2058106065,-0.2173732221,-0.1256145239,-0.3919209838,-0.2709895968,0.2555086315,-0.0818464458,-0.5978168845,0.4778692722,-0.2583063245,-0.2612285614,-0.301145643,0.2260324508,-0.3769453764,-0.0717824697,0.0178305171,0.4209887683,-0.2558437288,-0.4518943429,0.3033281565,0.132202059,0.0490965471,-0.0911229998,0.0624668375,0.016453892,0.1251921505,0.0299502946,-0.2367448658,0.0181071609,-0.0343427211,0.0501807816,0.2111343741,-0.0756804794,0.0804320276,-0.1178140342,0.3071854711,0.1701345295,-0.1060704738,0.0013298367,-0.1242687404,-0.0506846383,0.0312568918,-0.2849458158,0.427323699,0.0295932982,0.0342788622,0.04062704,-0.0121682873,0.2551823556,-0.4470511973,-0.0299693905,-0.2776483595,0.067414321,-0.178284511,0.3106676936,0.0469453596,-0.1613187045,-0.0527300723,0.1247884855,0.1174355969,0.0591896065,-0.0129036559,-0.269972831,-0.0324840285,0.2312888801,0.2289917171,0.3496314287,0.115698345,0.3512847722,0.0342467166,0.2262532413,-0.0512010455,0.0902003199,-0.3052617311,-0.2149517387,-0.0048807352,-0.2654453814,-0.1836682558,-0.1550815701,0.4538541436,0.1572224796,-0.029182272,-0.1685611755,0.0010863844,-0.4902951419,-0.2800119817,-0.0975242257,0.0815566257,-0.1654750705,0.2053429633,0.1156303436,0.2112517655,-0.0332498439,0.1312853396,-0.0300707929,0.4173390865,-0.3789163232,-0.1348476708,-0.2340608537,-0.0924062282,0.0131628197,0.0842126682,0.0436279997,0.252548337,0.4446316063,-0.1340670437,-0.1215796024,-0.3680643737,0.0678073838,0.0943639651,0.4574839771,-0.1171515882,-0.1442151517,0.4991394281,0.1297985613,0.084979929,-0.2086957693,-0.0758643746,-0.3315315545,-0.1449801773,0.1801486611,0.3122531474,-0.1543768644,-0.1788300127,-0.3968042135,-0.1747997552,0.180310294,-0.1156640276,0.2182236016,0.0924047828,-0.0676007643,-0.3087828457,-0.3239448667,-0.0161743592,-0.206396535,-0.3550776839,0.2109493166,-0.0881979167,-0.2693279684,0.0925171822,-0.0282172877,0.016177604,0.3049800694,-0.2734812796,-0.2682304084,-0.0042057978,-0.0481314398,-0.0980097055,-0.0583930314,0.3750461936,-0.4643383622,0.0714336708,0.0093641905,-0.1184322834,-0.0229049493,-0.0059864121,0.2117903829,0.0783641636,-0.0368255675,0.3925867677,0.8176041245,-0.1708510071,-0.2922508121,0.0071513071,0.0765939057,0.4609477818,0.0371164158,-0.1185649708,0.2269795835,-0.0316408351,-0.2075950354,-0.0661003888,-0.1603615433,-0.0048721205,-0.3039003611,0.0275688972,0.027729379,-0.000963409,0.059073884,-0.0452957973,0.1002086923,0.4998924136,0.0994199663,0.0546422079,-0.1810015887,0.1940510869,0.1275216043,0.1373193711,0.154809922,-0.5627899766,-0.0277308822,-0.1342883855,0.1363339424,-0.1989897341,-0.1131240577,0.0008902828,0.1268898845,0.2319722474,0.0536368787,0.4444595575,-0.1728119254,0.3386341631,-0.2411594838,-0.371055603,0.0571912676,-0.0406888016,0.1190116778,-0.0647267774,0.2721293867,0.0490347855,-0.2388035059,-0.009640458,-0.1223076582,0.2375264466,-0.3879928589,-0.0015040449,0.0159350466,-0.109943755,0.0131367519,0.39192608,0.1378644407,0.0135249374,0.4097759426,0.1442152262,-0.497607857,0.0785207599,0.0322838165,0.2788361907,0.1922329366,-0.1292794347,0.0909295976,0.1111607254,0.2234663665,0.1591204405,0.1979227513,-0.0487259254,-0.2857816815,-0.0351845436,-0.1305648685,0.1572665125,0.1777327508,-0.0596756749,-0.211930126,0.3234419823,-0.1633157581,-0.2870645523,0.4875818193,-0.4552233815,-0.1169766337,-0.2067614198,-0.261526674,0.1276055276,0.1201593801,-0.135349229,0.2356792539,-0.1499542445,-0.0892257765,-0.1631143093,0.46385023,0.9512078762,-0.110257946,0.1544109434,-0.4548836946,-0.2424463332,0.3881334364,-0.6364178061,0.1230963916,-0.3222511709,-0.6446196437,-0.2769861817,-0.2548324764,0.1701149344,0.4244231582,0.1449197531,0.2662792206,0.2202853858,0.3925861716,0.1395750642,0.2188519388,0.1050671563,-0.0499099977,0.0382157005,-0.0187378526,-0.1779113114,0.3877008557,-0.1563107073,0.090573132,0.1381307542,-0.2021086514,-0.3683967292,0.0349752717,-0.4835196733,-0.3193583488,-0.1983293891,-0.1050007343,-0.1611587256,0.6106063724,0.5191224813,0.2554518282,-0.2751760483,0.2046951503,0.0499762222,0.1072300524,-0.2390394956,0.0409159251,-0.0047395756,-0.3190503418,0.1609255522,0.4274270236,-0.1077920496,-0.2211221308,0.0800050125,0.2500099242,-0.2230263352,-0.1024609432,-0.0285336077,0.0582537502,0.3181501031,-0.0467495173,0.0466799326,0.410015285,-0.3039351702,0.2527983487,-0.1804301888,-0.3592645228,-0.0745124519,0.6200005412,0.2009725869,-0.3169282675,0.2317246497,-0.6443600655,-0.1544136107,-0.2400146276,0.0046451427,0.0634560138,-0.2399082482,-0.0872924179,-0.2366236895,-0.5042473674,-0.2811076343,-0.0148047954,0.3477854431,0.19227162,-0.1549737304,-0.2511276007,-0.2489345074,-0.0213063713,-0.0368133299,0.281445086,-0.0855257437,0.0221631546,-0.0623625927,0.4397819638,-0.1768968403,-0.0505925901,0.0129041467,0.2811278105,-0.0726351738,0.2149667293,0.0193275493,-0.2057608962,-0.1547146738,0.0692065284,-0.3085489869,-0.106896691,-0.0977821499,0.1999124438,-0.0651383772,-0.019518692,0.0739022791,0.0767337382,-0.0162347388,0.0969004035,0.1849387288,0.1663274169,0.1635080129,0.3371215463,0.417263478,0.0450641364,-0.5580723882,0.0045634103,0.0776347369,0.3335310221,-0.0398052558,0.443864435,0.0589767173,-0.0536442176,-0.1682883203,-0.2278661579,0.162588492,-0.1700044721,-0.0123476349,0.0889406502,0.072947219,-0.1642906964,0.211525768,0.6892214417,-0.1270240247,0.2658444047,0.5344479084,0.1247840896,-0.0779512301,-0.1968764216,0.2698137462,-0.2581944168,-0.1192479357,0.1488577425,0.1045387089,0.1528265029,0.4976006448,-0.0638405383,0.129744485,-0.1781682521,-0.0362895988,0.3569216728,-0.0781722665,-0.1497523189,0.1435687989,0.2890768349,0.1037418917,0.0205131434,-0.1031879112,0.3308336139,0.1263918281,0.076552771,-0.0850741938,-0.2613575757,-0.0313375741,0.1608713716,-0.037491627,0.1059536785,-0.1038472876,-0.2617317736,0.2561286986,-0.4189781249,-0.2652271092,-0.0451036505,0.1198685095,-0.0386008024,-0.0583102591,-0.0120711941,0.2708316445,0.2669641078,0.1111741364,-0.3568680286,0.3783860803,0.0905447528,0.0198655371,0.2002969831,0.3367417455,0.048800718,-0.177995801,-0.1519586891,0.2164179385,-0.2991585135,-0.0324565396,0.3661028147,0.2261829674,-0.2345114946,-0.0149076022,0.2408696711,-0.0516367406,-0.046261318,-0.0225156359,-0.3220124245,0.3685225844,0.1428087205,-0.0230738763,0.3599671125,0.0242867973,0.0276294053,0.3552056551,-0.3362348676,0.1001110822,-0.4896653593,0.2341651171,0.6023523211,-0.2003487498,0.2339518517,-0.0601427369,0.0062149987,-0.0488332957,-0.1157034263,-0.1481423229,0.1988975853,0.1247574762,-0.0034724788,-0.2451101691,0.5002832413,0.1018412784,0.0953216106,-0.1783418357,0.0402329378,-0.2901412845,0.1853389889,-0.0358838513,-0.0577067845,-0.0647070482,0.1328808814,0.1529143751,0.5575082302,-0.2293194383,-0.0373077765,0.0284239538,0.5112212896,-0.1240158454,0.0761894882,0.1746072024,0.5827258229,0.1541285962,-0.2528559566,0.3929022253,0.1215763241,-0.085987255,0.3141989112,-0.1304077059,0.4672880769,0.1673900336,0.4954754114,-0.0085944878,0.2672285438,0.0012572634,0.1383724809,0.097738333,0.1736004055,-0.4166921973,-0.0643259883,0.2751037478,-0.153191641,-0.4899983704,0.0420864709,-0.2413119227,0.3103083074,-0.0376972891,0.0449836291,0.1035349891,-0.2690927386,0.0909479558,0.148105517,0.3021665812,0.2165759206,-0.2832598686,-0.1835466921,-0.4077525437,0.1065463051,0.4210262597,-0.2014581114,-0.1317655295,-0.2105410099,0.4319619834,0.4071934819,-0.1818083078,-0.3839523494,0.1498399824,0.2985139787,-0.3370352685,-0.1010269225,0.0965486616,0.0452614576,0.0278763454,-0.2077089846,0.4524623454,0.2232903689,-0.306347549,-0.4392747283,-0.4033889174]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/883","title":"Downloading\/caching only a part of a datasets' dataset.","comments":"?","body":"Hi,\r\nI want to use the validation data *only* (of natural question).\r\nI don't want to have the whole dataset cached in my machine, just the dev set.\r\nIs this possible? I can't find a way to do it in the docs.\r\n\r\nThank you,\r\nSapir","comment_length":1,"text":"Downloading\/caching only a part of a datasets' dataset.\nHi,\r\nI want to use the validation data *only* (of natural question).\r\nI don't want to have the whole dataset cached in my machine, just the dev set.\r\nIs this possible? I can't find a way to do it in the docs.\r\n\r\nThank you,\r\nSapir\n?","embeddings":[-0.1374467611,-0.0426160395,-0.141087532,0.1024580672,-0.0408133119,0.0648453757,-0.0121131353,0.3359820843,-0.2723958492,0.1254544109,-0.1265009642,-0.2351125479,0.0974971578,0.3681795299,0.181126684,0.3194242716,-0.0627782345,0.4418856502,0.0889103115,0.1366407126,0.1183996126,-0.2997474968,0.0909952149,0.1433088332,-0.0615132079,0.0096344361,0.3426644802,-0.1270690411,-0.4054439366,-0.2210817486,0.508469522,0.1572905183,0.073552303,-0.3696883917,-0.0001228316,-0.0318253413,0.0752060786,-0.3035333753,-0.1024793684,0.1627182662,-0.517436862,0.2403925657,-0.1568187028,-0.1071446165,-0.3103720844,0.1122122854,0.1275636852,-0.1583422273,0.4761701226,0.3376511335,0.0787040219,0.1859824359,-0.2700878084,0.1042646915,-0.3700443506,-0.0401240624,-0.0917354971,0.1628636271,0.4165641963,0.1856683493,0.0712484866,0.3408580422,0.1432282031,0.4570239484,0.5227971077,-0.1314655244,0.1303727776,-0.5452263355,0.478299737,0.4119470119,0.9071882367,-0.250955075,-0.1308246851,0.1060938686,-0.1466452926,-0.3169739544,0.0534848087,0.3323171437,-0.2783376575,0.4988292158,-0.62631917,-0.4391212761,-0.4325579107,0.0141700925,-0.1454232931,-0.0686153471,-0.019549707,-0.1016099006,0.179340899,0.1878785491,0.4861334264,-0.0349499322,-0.3248578012,0.2488519847,-0.0958415568,-0.4637936354,-0.4852438569,0.1451468915,0.2966068983,0.2268617749,-0.1171468943,0.2044386119,0.0162711199,0.058999788,0.3115120232,0.1409805268,0.3578493595,0.283824414,0.1568547636,-0.3509333134,0.0994940326,-0.0399884321,0.054066442,0.1293051541,-0.1050073653,0.2131259292,-0.118384257,-0.6321311593,0.0207788292,-0.2109308094,0.0650768429,-0.0598816574,-0.0191945974,0.3762590885,-0.1715584695,-0.1063481197,-0.47386989,0.152756542,0.0543978065,-0.6790841222,-0.1179624945,0.2149178684,-0.0769704133,0.1502093524,0.3990958333,-0.2848175764,0.3069519997,-0.1370978355,-0.0759940371,-0.010884773,0.4047518075,-0.2640735805,0.3685115278,0.3398521543,-0.0566371307,0.1933007538,-0.052007772,0.0457630791,-0.294903338,0.2102558911,-0.0797557086,-0.6121847034,0.04486477,0.0467117392,-0.34981668,-0.0024292523,0.0592459776,0.2316785157,0.1725096107,-0.0947780237,0.0103178332,0.1664558798,-0.0138502894,0.0101089012,0.2632687092,-0.0003314269,-0.3484489322,-0.0181803722,0.0424299315,-0.2017668486,0.0218397621,0.1978917718,-0.2650619745,0.2062421441,-0.0211427696,0.0744974688,0.5971490741,-0.1140482277,-0.344722569,0.0710783303,0.1004632562,-0.415718168,-0.0322776772,0.1460054219,0.2701596022,0.0720762536,-0.5158414841,0.5135374069,-0.0162908398,-0.2199502289,-0.063721098,-0.0968876407,0.0079720123,-0.0509548895,0.3238458633,0.5501176715,0.1531887203,0.1884586513,-0.1440668851,-0.2171712071,0.2572123408,0.1831329614,0.2408896238,-0.26440081,-0.0850991234,-0.3341043591,-0.37259835,0.236064285,-0.0682310537,-0.6367200613,0.4715855718,-0.2042511255,-0.2771148086,-0.3674243689,0.2397251278,-0.4086696506,-0.0631608218,-0.0320049748,0.4450316727,-0.3287798464,-0.4518418908,0.2736391425,0.1365437806,0.0322911441,-0.0560454801,-0.049032677,-0.0034353475,0.1451181918,0.0082641058,-0.2690135241,0.0159483254,-0.0819392577,0.057065405,0.2657315135,-0.0854317695,0.0423620343,-0.0401026085,0.4129811227,0.1607373804,-0.1386441439,0.0715291202,-0.1249858513,-0.0310649686,0.0153563777,-0.2902079821,0.4223224521,-0.0933532864,0.054379493,-0.0209952071,-0.0281281956,0.2735019028,-0.4489250481,-0.0301996823,-0.2573123574,0.0331830606,-0.1417350173,0.2891089022,0.0148774423,-0.1733635962,-0.0426876955,0.1342742443,0.0957644582,0.024761593,-0.0699215457,-0.2100890279,-0.0180694293,0.1931414902,0.2020051479,0.3368987739,0.0947055295,0.3695683777,0.0914395005,0.1567519456,-0.1025972664,0.0932619348,-0.3187775314,-0.1920813322,-0.0027053603,-0.3635240197,-0.1590875685,-0.1404148042,0.3883660138,0.1040587351,-0.0225094501,-0.1600258201,-0.0442952067,-0.4688511193,-0.225802362,-0.0323610492,0.0539622828,-0.1789652407,0.2022025883,0.085823141,0.1962666959,-0.0509961136,0.1344579905,-0.2207099646,0.4085652232,-0.3327551186,-0.0179870613,-0.181783244,-0.1099309772,-0.0397527069,0.1218764037,0.0517673008,0.2202750295,0.4303820729,-0.1373547316,-0.0658623278,-0.3770914972,0.071472913,0.0300357267,0.4007458985,-0.085519217,-0.1211743876,0.5655697584,0.1051134691,0.0830382034,-0.2331193835,-0.1972286701,-0.3242508769,-0.1633091271,0.1400143802,0.2757110298,-0.1900677681,-0.1904013604,-0.3398293853,-0.1679541171,0.1173124462,-0.0613575727,0.2415679991,0.1865021884,-0.0792717338,-0.2454088926,-0.2725312114,0.0276040118,-0.267896831,-0.3968152404,0.2685887516,-0.0838987678,-0.2650083899,0.1844808161,-0.0204751436,0.0405625813,0.3595046103,-0.2926294804,-0.2033503652,0.019185951,-0.0165040512,-0.1308446079,0.0053278981,0.3708942831,-0.5016208887,0.0578617714,-0.0780358911,-0.0861175582,-0.0468225181,0.0330719128,0.2427095473,0.1267855465,0.0060205488,0.3394410014,0.8367421627,-0.1562753767,-0.1451544464,0.0468373559,0.1235922575,0.4819706976,0.0237503164,-0.0599255748,0.2448618263,-0.0098639466,-0.1979780793,-0.0227937717,-0.1697829515,-0.0401639566,-0.3260614574,0.0161697231,-0.021365786,0.0177161004,0.0595101714,-0.1068871766,0.0490466654,0.4871895611,0.0409284197,0.0646501333,-0.1720582545,0.2007102072,0.1052323878,0.1145450175,0.1393191814,-0.5544501543,-0.08879105,-0.1391596943,0.1617945433,-0.2232851386,-0.1059561074,-0.0004469543,0.1410416365,0.2496241331,0.0491020232,0.4754675925,-0.1890699714,0.3007536829,-0.1983014494,-0.2632834017,0.0801516771,-0.064339079,0.0827695727,-0.0453801267,0.1880314797,0.0369495079,-0.213186115,0.0026454034,-0.1168536916,0.2085748762,-0.3043315411,0.020347748,0.0857347921,-0.2044940591,-0.0110946139,0.4005877078,0.1486458778,-0.0077359029,0.376612395,0.0820831582,-0.3784015179,0.0910139233,0.0358769298,0.3117738962,0.1376159042,-0.0415282622,0.0577494502,0.1335817873,0.1560987085,0.067692481,0.2202865481,-0.0023568082,-0.1942608654,-0.0524848551,-0.1055566967,0.1327354759,0.1960754991,-0.0448829122,-0.2435298562,0.3584085107,-0.1220319793,-0.2181425244,0.4580951631,-0.4418253005,-0.1854168177,-0.256526202,-0.3057307601,0.1724599749,0.1566273421,-0.1587400734,0.3580463529,-0.1211837977,-0.1771604866,-0.2602133751,0.4858151972,0.9656774402,-0.0824785456,0.1716563702,-0.4369552732,-0.2747071385,0.4116189778,-0.5839255452,-0.0022938652,-0.2434084713,-0.7081192732,-0.3047324419,-0.2651164532,0.2082592845,0.3617419004,0.1382876635,0.2809907496,0.1361981481,0.4389011562,0.1327794045,0.2159486711,0.0590404868,-0.0687291399,0.0660021305,0.0080125621,-0.1922214031,0.3214334249,-0.1782647371,0.0792063475,0.0770836249,-0.188220948,-0.3145408332,0.0466819592,-0.4325347245,-0.2972433865,-0.1925108433,-0.0478646308,-0.1209909096,0.4438857734,0.5885183811,0.2205909789,-0.3257596493,0.2219022661,0.1563003063,0.0752307922,-0.2569479644,0.0430275388,-0.0074365204,-0.3587322831,0.126035586,0.4123756886,-0.1065580323,-0.2264293581,0.0868154988,0.2273316979,-0.2253081352,-0.1141295135,-0.1107213646,0.0477021411,0.318979919,0.0104048997,0.0576559529,0.4329645634,-0.3249253333,0.2550159097,-0.1262630075,-0.3172800243,-0.0714946911,0.4745688736,0.2743753195,-0.3237265646,0.1772327423,-0.6621029377,-0.132888779,-0.2224541306,0.018223241,0.0993283391,-0.1720969677,-0.0905845389,-0.2307333499,-0.4553498924,-0.2947902977,0.0546590835,0.391005367,0.2297551185,-0.1474802941,-0.309730649,-0.2799613476,0.0368717201,-0.0937420353,0.3394378126,-0.0650113225,-0.0323174037,-0.0451625548,0.445165962,-0.1900268793,-0.0815002173,-0.003022165,0.298040539,-0.0918201283,0.1778992862,-0.0303280596,-0.1808772832,-0.1483724862,0.0462594256,-0.3433184028,-0.1339532882,-0.1163734123,0.2160454839,-0.0050806138,-0.11031048,0.1084719449,0.0492346771,-0.042007871,0.1125307232,0.1900823563,0.064357318,0.2413266599,0.2987461984,0.4387895763,0.0308948588,-0.5843533874,0.0012794648,0.0732101575,0.2805948555,-0.0698784888,0.4520714581,0.0059578065,-0.0783169866,-0.1579786092,-0.1881336421,0.142772004,-0.1586769372,-0.0538808107,0.1139403656,0.1377262771,-0.0833973438,0.2055320442,0.7441473603,-0.1581216156,0.2178369164,0.5463070273,0.1350397617,-0.0323027484,-0.1739357114,0.3269474804,-0.2734062374,-0.1742238849,0.0945894122,0.1193305552,0.1205757484,0.4585033059,-0.0782593265,0.217984274,-0.2780933678,0.0096005751,0.3457688391,-0.1150410846,-0.1852799058,0.153897047,0.3198654056,0.1203401089,0.0590040199,-0.094193764,0.1704739034,0.1642772853,0.1129316166,-0.1096979976,-0.2603097856,-0.0254041087,0.2459260076,-0.0731129199,0.0914040506,-0.0972566679,-0.3186215758,0.2541065514,-0.404066354,-0.2275491655,0.0003432054,0.191698283,-0.0545702204,-0.1464837044,-0.0006390892,0.2672331035,0.250022769,0.1231922507,-0.3893027902,0.3186857104,0.1060644984,-0.0341998935,0.1553311199,0.3613994122,0.0751968697,-0.188309148,-0.148131296,0.2250049412,-0.2318807691,-0.0306779426,0.3434047103,0.2200312167,-0.2050026357,-0.0452892408,0.2053169459,0.0108029135,0.0226544254,-0.0571590513,-0.2804427743,0.3539022207,0.170593515,-0.0089791967,0.282041043,0.0248416234,0.0198285431,0.3590813279,-0.3500386775,0.0627183691,-0.4856595099,0.1470258981,0.5868788362,-0.1691010743,0.2690508366,-0.0078771897,0.0127172135,-0.0906349197,-0.1570570171,-0.1845003664,0.1900912076,0.1308382004,-0.0000771253,-0.2316782176,0.528866291,0.0937855616,0.1363275647,-0.1662082076,0.0012829792,-0.3343859613,0.1980863065,-0.0352335311,-0.0618153773,0.0300384164,0.1113355011,0.1149401739,0.5425885916,-0.1419861019,0.0364263058,0.0559761189,0.4926174581,-0.0487380028,0.1260616928,0.1735846102,0.5694362521,0.1088765934,-0.2390999049,0.4452870786,0.0827021897,-0.1038113683,0.3452756107,-0.1192124784,0.5597140193,0.2299899161,0.4492059052,0.0016438774,0.2780327499,0.0143390773,0.1992880255,0.0642677471,0.2131054997,-0.4620919228,-0.0988513008,0.2637323439,-0.2011560649,-0.4939234853,0.0667466745,-0.1801367253,0.3621132672,-0.1035579294,0.0447167046,0.161713019,-0.300301522,0.0634083375,0.1214228049,0.3071911633,0.1459642351,-0.2854562998,-0.1960706413,-0.4336859286,0.092421934,0.466198504,-0.1405248195,-0.2004921287,-0.1580229551,0.4047879577,0.3079943955,-0.1773512661,-0.3957813978,0.1822151244,0.3118024468,-0.3804439604,-0.1844836473,0.1428719312,0.0525922291,-0.043532934,-0.1755223572,0.4481946826,0.2222404629,-0.3349874914,-0.5048503876,-0.3779147267]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/883","title":"Downloading\/caching only a part of a datasets' dataset.","comments":"I think it would be a very helpful feature, because sometimes one only wants to evaluate models on the dev set, and the whole training data may be many times bigger.\r\nThis makes the task impossible with limited memory resources.","body":"Hi,\r\nI want to use the validation data *only* (of natural question).\r\nI don't want to have the whole dataset cached in my machine, just the dev set.\r\nIs this possible? I can't find a way to do it in the docs.\r\n\r\nThank you,\r\nSapir","comment_length":40,"text":"Downloading\/caching only a part of a datasets' dataset.\nHi,\r\nI want to use the validation data *only* (of natural question).\r\nI don't want to have the whole dataset cached in my machine, just the dev set.\r\nIs this possible? I can't find a way to do it in the docs.\r\n\r\nThank you,\r\nSapir\nI think it would be a very helpful feature, because sometimes one only wants to evaluate models on the dev set, and the whole training data may be many times bigger.\r\nThis makes the task impossible with limited memory resources.","embeddings":[-0.2596469522,0.0176629163,-0.1444136202,0.0560275577,0.0052094632,0.1932338327,0.0714360848,0.424452126,-0.2096630186,0.1225835308,-0.1673961878,-0.3112493753,0.0382805541,0.3431017399,0.2733544707,0.3157155216,-0.0449594185,0.4445394576,-0.013587472,0.1797878444,0.1321712285,-0.3596689105,0.0900435597,0.1798119247,-0.0343141593,-0.0130379051,0.3524547517,-0.1729032844,-0.3025227785,-0.2646118999,0.507656455,0.1568506509,0.098767817,-0.3624978662,-0.0001216029,0.0493053906,0.0157325901,-0.259304136,-0.1948634237,0.1747968346,-0.3646432161,0.2211199254,-0.1016264111,-0.1109533757,-0.2629934251,0.1496164203,0.0997373387,-0.1148649603,0.3868528008,0.3819859028,0.0726678595,0.2527618706,-0.2582719028,0.0818846226,-0.4386486113,-0.120596759,-0.1113695875,0.2293516546,0.3401831985,0.1742604077,0.1508786678,0.3609142303,0.203758955,0.3365459442,0.5206016898,-0.105901733,0.2213973999,-0.5290890932,0.3528997898,0.3634751141,0.8179839849,-0.2111045569,-0.1709374338,0.0381576978,-0.1108286008,-0.324495703,-0.0276461076,0.3498649001,-0.3038274348,0.5577995777,-0.6025673747,-0.5060825944,-0.4514364898,0.0211165417,-0.0719960853,-0.1111353412,0.0267906878,-0.0793382227,0.2038470954,0.3404905498,0.4757578373,-0.1010633186,-0.3275689483,0.2050732076,-0.1556142569,-0.4630132914,-0.4531757832,0.1124547273,0.2769366503,0.2156210542,-0.0281268787,0.1360063851,0.0196452886,0.0326850899,0.3324066103,0.2571540177,0.2097781897,0.3000056744,0.126241982,-0.2836258709,0.0435491428,-0.0383794494,0.0166266803,0.1793973297,-0.0352480635,0.2005891949,-0.1922214776,-0.57342875,0.0282574501,-0.162422657,-0.026856754,-0.1042123809,0.0117447656,0.3720725477,-0.2337694913,-0.0338172726,-0.414483577,0.1290370375,-0.0244037621,-0.7614021897,-0.1353474855,0.2001847923,-0.1518820971,0.1537026912,0.3087083995,-0.2576307654,0.2571436763,-0.2327064127,-0.0536342002,-0.0144959623,0.3822110593,-0.2405206561,0.3597934842,0.3008718193,-0.1003428549,0.1436013728,-0.0037028547,0.0677643716,-0.2702770233,0.2140781879,-0.0240437184,-0.7160574198,0.132977128,0.0306058656,-0.291547507,0.0498247035,-0.0186125264,0.2691911757,0.1003063768,-0.0424363464,-0.028077485,0.1042831391,-0.0931085125,0.0569567047,0.3539173901,0.0054562553,-0.3496731222,-0.0455453433,0.0757640079,-0.2223428041,-0.0782371387,0.179689303,-0.3526026309,0.2832117081,-0.0349932089,-0.0515899062,0.6107296348,-0.033008486,-0.341673106,0.0503783599,0.0374969244,-0.3896382749,-0.107169047,0.1974391341,0.1391672492,0.0782243237,-0.5371806026,0.4953878224,-0.0045664981,-0.2572207153,-0.0902148783,-0.1335968673,0.0509743616,-0.0344618112,0.3187843263,0.5672199726,0.093451567,0.2109842896,-0.0654322132,-0.1826470941,0.1779056787,0.1441976428,0.1808380336,-0.2567463815,-0.1036055014,-0.2643276751,-0.3934879899,0.2896128893,0.0190090984,-0.6332793832,0.5450124741,-0.1466409117,-0.2126225233,-0.2811501026,0.1985191107,-0.3874562681,-0.0683606789,-0.0787717327,0.5183345079,-0.2221153527,-0.4332073927,0.2459861338,-0.0388423614,-0.0085976645,-0.0347101055,-0.0773481131,-0.0136265745,0.2332274318,-0.0780256838,-0.2326035351,-0.1438825428,-0.0770930946,0.0010289117,0.2435720712,-0.1894330233,-0.0382388569,-0.0579957701,0.4480165541,0.1133444309,-0.1220167428,0.0389686115,-0.1264074892,0.037879277,0.0592170879,-0.2691627443,0.445589602,-0.0425802842,0.1304669082,0.0402546898,-0.0788122565,0.2034360319,-0.3745237887,0.0144097637,-0.2602585554,0.0538863651,-0.1644445062,0.335213393,-0.038239155,-0.2119158804,-0.1100293547,0.1415116936,0.0741901323,0.0180712882,-0.1057825983,-0.1695173681,-0.0013913501,0.1838152558,0.187406376,0.3733811378,0.0554433055,0.3742053807,0.1727607846,0.0835776925,-0.0322370641,0.0589875244,-0.2422545105,-0.1820131689,-0.03363695,-0.3563263416,-0.1824299991,-0.1795852184,0.4001220763,0.1201160178,-0.0454229973,-0.1748720109,0.0162675995,-0.4271829426,-0.0491978042,-0.0482637696,0.1393751502,-0.1046889052,0.2175350934,0.063642472,0.1797725111,-0.0330221541,0.0858021379,-0.2827939391,0.4851481616,-0.3380614221,0.0753311589,-0.0537016839,-0.1475924999,-0.069603391,0.101518549,0.1213861555,0.1434981525,0.3591328263,-0.0873147398,-0.1552483439,-0.3242338002,0.1137347221,-0.020188326,0.4564116299,-0.0426918343,-0.1663059741,0.513420701,0.16451253,0.1184290349,-0.180262655,-0.2282163352,-0.3151454329,-0.1713221967,0.2047231942,0.2288086414,-0.1834838986,-0.2206598818,-0.3584470153,-0.1597438604,0.1136493832,-0.1076956987,0.1931927055,0.2225603163,-0.0169393755,-0.2397378832,-0.3206496239,0.0403390974,-0.2267978936,-0.3674768507,0.2391419262,-0.041413974,-0.2598355114,0.1799654961,-0.0711047426,0.0322008021,0.3513366282,-0.3724110126,-0.1565231383,0.0650557131,-0.0956741422,-0.1702403426,-0.0697042122,0.2953811288,-0.5932688713,0.1054323688,0.0074931323,-0.0587730929,-0.0502597913,0.0245569292,0.2246670425,0.0605238751,0.1298893839,0.350484848,0.8538551927,-0.1514459699,-0.1626930386,0.0154224522,0.0880191401,0.4468539357,-0.0379184447,-0.0748696849,0.2885404825,-0.049391441,-0.2889501154,0.0268131737,-0.1392236799,-0.104868196,-0.3318148255,0.0505775996,-0.0113214133,-0.0139270797,0.0929991007,-0.1182702184,0.0811954737,0.5327092409,0.0187404919,0.0137389842,-0.1102430895,0.2959983051,0.0583859682,0.1100257114,0.1178510487,-0.523583889,-0.1630560905,-0.0981968269,0.0888543874,-0.227131322,-0.0714122206,0.034540534,0.0316852368,0.2360617518,0.1014387682,0.5184095502,-0.2011044919,0.2330281287,-0.1885452569,-0.3180351257,0.0645318776,-0.190653801,0.0945238397,-0.0763200447,0.0763737559,0.0073571913,-0.193640545,-0.0187628083,-0.0699421018,0.2333287746,-0.3665068746,0.0682157204,0.047265362,-0.011143256,0.0290325861,0.4540618956,0.1134135276,-0.0058890972,0.3645147383,0.0731299371,-0.3548815548,0.0975846499,0.0919066966,0.3425664008,0.0917205513,0.0567645356,0.1430366784,0.1277785748,0.1461908221,0.0764795765,0.2214029729,0.011715007,-0.1909357756,-0.0623567402,0.0023007775,0.1911739558,0.1707263887,-0.0837877318,-0.2795104086,0.3804681897,-0.0915223807,-0.2260641009,0.4861368835,-0.3498319387,-0.2019871175,-0.2566847205,-0.3647646606,0.1765506417,0.0798219889,-0.0859941393,0.3306180239,-0.252682358,-0.1009435281,-0.2397096306,0.4438515604,0.8610088229,-0.1227932051,0.2062957287,-0.3994804621,-0.2029975802,0.3680358827,-0.568880558,0.0569614954,-0.2196463645,-0.7584602833,-0.2906113267,-0.2542025149,0.1865492761,0.3905203044,0.2456165105,0.2405982465,0.1417122781,0.4928078353,0.1205115095,0.1755267531,0.1043293178,-0.0288764983,0.0713439807,0.0225885529,-0.1270377636,0.2708648443,-0.1556711495,0.0918053985,0.0571949296,-0.0372974873,-0.3012590706,0.0489365421,-0.460675329,-0.2496441901,-0.1814269722,-0.0514715016,-0.013131653,0.4394465089,0.5612179041,0.22548531,-0.2542156577,0.2422557175,0.0651296228,0.0886115059,-0.2462749332,0.0181908924,0.0016432061,-0.3460488915,0.1401317567,0.4257399142,-0.0995908156,-0.2508583069,0.0373264477,0.1967259049,-0.0806845352,-0.1119176224,-0.161089927,0.1151348278,0.3356603086,-0.0060500419,0.0489645042,0.3836348653,-0.3482622802,0.2078300416,-0.1432594806,-0.3118987679,-0.0696334764,0.4085510969,0.2608778477,-0.2065868825,0.2444152087,-0.6133481264,-0.0704692006,-0.2092334479,-0.0219698865,0.191212967,-0.1649276018,0.0004837792,-0.2133574486,-0.4010840058,-0.3402512074,0.0490704626,0.3917354047,0.1911246479,-0.1881880909,-0.2655119002,-0.231969744,0.0511179492,-0.1082250327,0.3742311001,-0.062148083,-0.01692193,-0.1252294779,0.4188044965,-0.1844319701,-0.1776658297,-0.0053341785,0.2531560361,-0.1674052924,0.1499563903,-0.0022157354,-0.1374547333,-0.1754706353,0.0281307474,-0.3456308842,-0.1469108164,-0.0746241212,0.1991707534,-0.0085584493,0.0384336263,0.117102623,0.0402875431,-0.0568367615,0.052002836,0.2014924437,-0.0122299753,0.257540077,0.2684284747,0.4537934065,-0.018109858,-0.5937926769,-0.009499778,0.0744519681,0.2208894938,0.0398879908,0.4148598313,0.0169251896,-0.0538877659,-0.3319833875,-0.1668200195,0.0532803014,-0.1499504745,-0.0398057699,0.118595317,0.1468976289,-0.0237996355,0.1988660991,0.8052158356,-0.0924879387,0.2291515321,0.5309456587,0.1136762649,0.0040293285,-0.1860904396,0.3704554439,-0.304171741,-0.1823699325,0.1436482817,0.1134742349,0.1958086193,0.4399095178,-0.0908521935,0.1677696109,-0.2348217517,0.0168016572,0.4365864694,-0.0261416491,-0.1989732683,0.1153116822,0.2590424716,0.2069092095,0.03517906,-0.1256256551,0.1561603546,0.167011112,0.2003984749,-0.1934622377,-0.235666573,-0.0692513138,0.1692312658,-0.0888706148,0.1284372211,-0.1844213456,-0.3254106343,0.2570571005,-0.4299590588,-0.2986115515,0.0170880128,0.1921701133,-0.0724868104,-0.2084243,0.0067041209,0.2563123107,0.3134522438,0.1623129696,-0.4159726501,0.2470381111,0.1064635888,-0.0972429514,0.2171068043,0.2817168236,0.1689370424,-0.0912872776,-0.1346125603,0.1551804543,-0.2595524788,-0.0017357422,0.3158125579,0.2392761558,-0.2397789955,-0.0678152815,0.2292584181,0.0141973011,0.0887075216,-0.0198534187,-0.2765794098,0.2951282561,0.1854898632,0.0411084853,0.2854758203,0.017252421,0.014039143,0.4403685331,-0.390126735,0.2003467977,-0.5645316839,0.0401735753,0.6362974644,-0.1602789909,0.1644346863,0.0662632063,0.0568207875,-0.02891054,-0.099383533,-0.2132300287,0.1308529228,0.1542663276,-0.0039088451,-0.2629947364,0.5772280097,0.067066133,0.0950344577,-0.1848518699,0.0580332093,-0.3419104815,0.1647651643,-0.0758660734,0.0435415432,0.0781059563,0.123653397,0.1260083467,0.543223381,-0.0659056604,-0.0162429102,-0.0320679285,0.5045503974,-0.1067461297,0.1457104534,0.0873923451,0.5286428928,0.1306842715,-0.2284795046,0.4167973697,0.11578352,-0.15635176,0.3952567577,-0.1051854342,0.43802917,0.3276388347,0.5421128869,-0.056315653,0.2171078324,0.0971882716,0.187504977,-0.0357793719,0.1916661263,-0.4345735312,-0.043586269,0.190625906,-0.1363946348,-0.5238884687,0.0310886316,-0.1524832249,0.2835089564,-0.0245359056,-0.1028685346,0.2182343304,-0.3127633333,0.047814358,0.1807873696,0.424695015,0.1729243845,-0.2599115968,-0.2585413456,-0.3770530522,0.1391886175,0.429576546,-0.0886319801,-0.2041847408,-0.1264173388,0.3098582327,0.2169163227,-0.211459741,-0.435421139,0.0923217162,0.2990783751,-0.3669798672,-0.1788091362,0.2019109726,0.091581203,0.0200916603,-0.1414799988,0.4481548071,0.0611790679,-0.3696348667,-0.4982996881,-0.3521561623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/880","title":"Add SQA","comments":"I\u2019ll take this one to test the workflow for the sprint next week cc @yjernite @lhoestq ","body":"## Adding a Dataset\r\n- **Name:** SQA (Sequential Question Answering) by Microsoft. \r\n- **Description:** The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. It has 6,066 sequences with 17,553 questions in total.\r\n- **Paper:** https:\/\/www.microsoft.com\/en-us\/research\/publication\/search-based-neural-structured-learning-sequential-question-answering\/\r\n- **Data:** https:\/\/www.microsoft.com\/en-us\/download\/details.aspx?id=54253\r\n- **Motivation:** currently, the [Tapas](https:\/\/ai.googleblog.com\/2020\/04\/using-neural-networks-to-find-answers.html) algorithm by Google AI is being added to the Transformers library (see https:\/\/github.com\/huggingface\/transformers\/pull\/8113). It would be great to use that model in combination with this dataset, on which it achieves SOTA results (average question accuracy of 0.71).\r\n\r\nNote 1: this dataset actually consists of 2 types of files: \r\n1) TSV files, containing the questions, answer coordinates and answer texts (for training, dev and test)\r\n2) a folder of csv files, which contain the actual tabular data\r\n\r\nNote 2: if you download the dataset straight from the download link above, then you will see that the `answer_coordinates` and `answer_text` columns are string lists of string tuples and strings respectively, which is not ideal. It would be better to make them true Python lists of tuples and strings respectively (using `ast.literal_eval`), before uploading them to the HuggingFace hub.\r\n\r\nAdding this would be great! Then we could possibly also add [WTQ (WikiTable Questions)](https:\/\/github.com\/ppasupat\/WikiTableQuestions) and [TabFact (Tabular Fact Checking)](https:\/\/github.com\/wenhuchen\/Table-Fact-Checking) on which TAPAS also achieves state-of-the-art results. Note that the TAPAS algorithm requires these datasets to first be converted into the SQA format.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":16,"text":"Add SQA\n## Adding a Dataset\r\n- **Name:** SQA (Sequential Question Answering) by Microsoft. \r\n- **Description:** The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. It has 6,066 sequences with 17,553 questions in total.\r\n- **Paper:** https:\/\/www.microsoft.com\/en-us\/research\/publication\/search-based-neural-structured-learning-sequential-question-answering\/\r\n- **Data:** https:\/\/www.microsoft.com\/en-us\/download\/details.aspx?id=54253\r\n- **Motivation:** currently, the [Tapas](https:\/\/ai.googleblog.com\/2020\/04\/using-neural-networks-to-find-answers.html) algorithm by Google AI is being added to the Transformers library (see https:\/\/github.com\/huggingface\/transformers\/pull\/8113). It would be great to use that model in combination with this dataset, on which it achieves SOTA results (average question accuracy of 0.71).\r\n\r\nNote 1: this dataset actually consists of 2 types of files: \r\n1) TSV files, containing the questions, answer coordinates and answer texts (for training, dev and test)\r\n2) a folder of csv files, which contain the actual tabular data\r\n\r\nNote 2: if you download the dataset straight from the download link above, then you will see that the `answer_coordinates` and `answer_text` columns are string lists of string tuples and strings respectively, which is not ideal. It would be better to make them true Python lists of tuples and strings respectively (using `ast.literal_eval`), before uploading them to the HuggingFace hub.\r\n\r\nAdding this would be great! Then we could possibly also add [WTQ (WikiTable Questions)](https:\/\/github.com\/ppasupat\/WikiTableQuestions) and [TabFact (Tabular Fact Checking)](https:\/\/github.com\/wenhuchen\/Table-Fact-Checking) on which TAPAS also achieves state-of-the-art results. Note that the TAPAS algorithm requires these datasets to first be converted into the SQA format.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nI\u2019ll take this one to test the workflow for the sprint next week cc @yjernite @lhoestq ","embeddings":[-0.0640092716,-0.2312078327,-0.2099176347,-0.1210695952,0.0999142751,-0.1975832283,0.0337971561,0.3069660664,0.1483602226,0.0562699474,-0.1675558388,-0.002632767,-0.0712603852,0.5443822742,0.1500162035,-0.1045532525,0.0272174533,0.1428327858,-0.2944614291,0.0711858869,-0.196293667,0.0914863199,0.0215712246,0.031229699,-0.2368902415,-0.0844550505,-0.1073180363,-0.1923377961,-0.1904111356,-0.1972694844,0.0054088389,0.072247088,0.1686420888,0.2730415761,-0.0001059325,-0.1343121231,-0.0751999691,-0.0975328833,-0.0939193591,0.0506609641,-0.1607964188,0.0926124603,-0.1110516489,-0.0771295801,-0.5252555013,0.2041900754,-0.0748102888,-0.3007471859,0.5820878148,0.2628435194,0.2186131328,0.0928574726,0.0751176327,-0.2174711227,0.3752230704,0.0049580853,-0.3586767614,-0.0065896218,0.4021380246,-0.348051697,0.0548400171,0.4530360699,0.265696764,-0.2276320606,0.1525066942,0.1255038083,0.2736350596,0.0334795527,0.0674656704,0.2694988251,0.3335775435,-0.0927536935,-0.1821043491,-0.1886140257,0.0144181484,0.1235252768,-0.2833782434,0.118624717,-0.3490382135,0.0167793985,-0.1807740331,-0.2453861386,-0.349671483,0.101798743,0.0179150905,0.5167180896,-0.0192335062,-0.1948671192,-0.0831076577,-0.1331273913,-0.0427691936,-0.0472527817,0.2443597317,0.2530235648,-0.5382785201,-0.2720508277,0.2136117667,0.1150445491,0.2435956448,0.1921216249,0.4720071256,-0.0723904297,0.0314646773,-0.2965055406,0.1672351956,0.4781259894,-0.003513505,0.0426871032,-0.1635618806,0.1145740077,0.069656536,0.0185354222,0.0113641415,0.4373466372,-0.2889096141,0.0332337134,0.121735774,0.0777647123,-0.3187230825,0.2000280917,-0.2216603905,-0.0858714953,-0.0461771674,0.3067342937,-0.2722217441,0.0963958725,0.1617859751,0.1525750607,0.0376236141,-0.1205838397,-0.0992913097,-0.1381375194,-0.1389065683,0.327624768,0.2592454553,0.0756572187,-0.118240878,-0.1105898842,0.1236674637,0.0671117604,-0.1190187186,-0.1257779002,0.2719293535,0.0063779862,-0.1196178198,-0.2474647164,0.006988266,-0.1855606288,-0.0826086476,0.2046098411,0.086625129,0.0822594985,-0.0458444618,0.2935351431,-0.0351122767,-0.3378953934,0.0560649224,0.3847602308,0.0544066653,-0.0786112845,0.1612036526,0.0770690739,-0.1640680432,-0.0699188039,0.2680043876,0.1623664647,0.0437387638,0.0179177821,-0.0602712929,0.2952128351,-0.4049222767,0.1558105201,0.1623992175,0.4924613237,0.1605460048,0.4469804764,-0.3020619452,0.042209927,-0.051084321,-0.1068252027,-0.0611654036,0.2619523108,-0.0233064312,0.222698912,0.1795921922,-0.0981692225,0.124762632,0.2569143176,-0.243451342,-0.0364894234,-0.2605640292,-0.320944041,-0.1499199122,0.5569957495,-0.0081572412,-0.0947261527,-0.1503633857,-0.2269732505,0.0897196904,-0.2418155223,-0.1053456739,-0.202632457,0.418138355,-0.1816864908,0.0669904649,-0.5491580367,-0.5756059885,-0.1482614726,-0.1463883817,0.2786490321,0.0339066721,-0.4509789348,-0.3733327091,0.1327016652,0.0075063659,-0.0343044698,0.2479703277,0.1869211048,0.2428165525,-0.0500193909,-0.0937463418,0.1550153494,-0.0424523093,-0.1147119105,-0.1329626888,0.198689729,-0.1225987598,0.0456116945,-0.0185674727,0.314791292,-0.1338375211,0.1751381606,0.3347406685,0.1550746709,-0.13191697,-0.2625659704,0.0805614889,0.4734991193,0.0326889567,-0.0527552143,-0.1923760474,0.4993881285,0.1238283738,-0.0980774164,-0.2853282094,0.4043613374,-0.1816473901,0.2681885958,0.1508674324,0.0346501954,-0.0117136929,-0.0824667439,-0.131069541,0.1497388333,0.4146442711,0.1367218494,-0.2725359797,-0.0683397576,-0.3361949921,0.3972062767,0.2918215394,-0.2563173175,0.3381398618,0.0930465311,-0.0448115245,-0.099624306,0.0174711868,0.1898336112,0.1904856265,0.2655975521,-0.1340573281,0.132565096,-0.0351311676,-0.1304928362,0.0047888304,0.0729855746,-0.1076816693,0.3593825102,0.3060602844,0.1129134521,-0.0875147507,-0.081166327,-0.2993248403,-0.0468394086,-0.0217849854,0.1793799549,0.2182229459,-0.2264639288,0.0334007703,-0.3110869825,0.0157671552,0.0116476575,0.6521463394,-0.1362127811,-0.1908007115,0.2219664752,0.074260965,0.910795629,-0.1697585285,-0.0477678366,-0.2234660834,-0.2786903083,0.1275860667,0.3061983287,0.0774174184,-0.0461224802,0.5052061081,-0.1818101406,0.3880173266,-0.0880887881,-0.3654263318,0.0920002386,-0.0208264478,0.0868983418,-0.0135212597,0.0519566387,-0.2462124228,-0.311757654,0.0962447152,-0.2440950423,-0.1063671485,-0.0403895676,0.2175629735,-0.090775229,-0.1126224622,-0.4963790476,-0.0617724843,-0.3034152091,0.4760821164,0.2297993004,-0.0288607217,-0.0194287114,-0.0775105655,0.0343687013,-0.1526608765,0.1068070829,-0.2156098038,0.1638543606,0.2227652371,-0.1240256354,-0.211823523,-0.4943149984,-0.3110443056,-0.0237979498,0.0763129145,-0.1710079461,-0.0684284568,-0.2200554162,0.0027596692,-0.1519955397,0.1327385157,0.4515957832,-0.0766791403,-0.2075061649,-0.0513131246,0.2085881829,0.5510808229,0.0616792701,-0.1596296579,0.1330932677,0.2265027463,0.0326379724,0.6488921046,0.0589807928,0.0846870765,0.2585291862,0.0645583197,-0.11768765,-0.2783655524,0.0752346739,-0.1934869885,-0.1577434689,0.0985051915,0.4758580923,0.2683967352,0.0449928492,-0.1529213935,0.1253436506,-0.2520808578,-0.1979157031,0.4055438638,0.0370066762,0.4625874758,-0.0875199139,-0.1534026265,0.007968137,-0.355032146,0.0913981497,-0.0922903121,0.0258747675,-0.4211962223,-0.5453454852,-0.2182356119,-0.2534332573,0.3696666658,0.4543704987,-0.0103512472,-0.0700463057,-0.4352989197,-0.0819312185,-0.0756269917,0.1700690091,0.062229462,-0.351492703,0.0121095553,-0.2442807257,0.0773436651,0.1962990165,-0.3358927071,-0.3367526531,-0.0537999533,0.3874261975,-0.3562621772,0.0581251197,0.2264221758,-0.0486682057,-0.0562434755,0.0086918687,-0.0854227766,-0.1264292002,0.0051181843,0.2333197594,0.3687148988,0.036403697,0.2235229313,-0.0396039337,-0.0924399942,-0.0361280143,0.0263197273,0.1368793249,0.0830826312,0.0011648067,0.1247434616,-0.1430703998,-0.1977378577,-0.0488720126,0.4001817107,-0.2085921615,-0.0698705316,-0.1505306065,-0.1544058472,-0.0211568605,0.2488559783,0.1011996493,0.3039302826,0.0348631851,0.1376393139,-0.2829020023,-0.1738834679,-0.1508074105,-0.0058538243,0.1136687249,-0.1343204975,0.4686171412,-0.2742679119,-0.2376136333,-0.0322412364,0.4682995975,-0.1962357312,0.1028005555,0.0156817846,1.0317857265,-0.1658943444,0.3711087406,0.141588971,-0.0974734575,0.2048790306,-0.350300014,0.144818902,-0.1310677826,-0.3593126833,-0.0561568215,0.0399527773,-0.0805622116,-0.004561604,-0.3912659585,0.1389491707,0.175603494,0.3666844964,0.2946521938,0.2584328055,0.069652155,-0.0152626121,-0.5002795458,0.2108229697,0.1116845459,-0.2397036254,-0.0502777211,0.0481445976,-0.1371403933,-0.1277022213,-0.1942481101,0.0707471296,-0.4300709963,-0.0462026186,0.2861474156,-0.1057446077,0.1106237322,0.2124155015,-0.0267282948,-0.0604788736,-0.3943512142,0.0101640364,-0.2656233609,0.2279188782,0.0728195086,-0.0291024782,0.292478174,-0.2039584965,-0.1170177087,0.2273564041,0.3363114595,-0.4757345021,-0.4612273872,0.0721802711,0.3284933567,-0.1256040484,0.0085265655,0.1294416487,-0.1191570759,-0.1417362243,0.2308954149,0.0601020046,0.0594073758,0.1053040102,0.4885711968,-0.1789277941,-0.0357294269,0.117661871,0.4107621014,-0.0385524631,0.2471322268,0.3028377593,-0.2766368091,-0.4391760826,-0.1043767333,0.2383774668,-0.4756156504,-0.029951755,-0.1643128544,-0.3447369039,-0.0333019234,0.5254306793,0.4061190486,-0.0993947014,-0.060709253,-0.0520864055,0.103038922,0.627266705,0.3372285366,0.1448047608,0.0963591561,-0.18211326,-0.0878586471,0.244761616,-0.4542951584,-0.0925646946,-0.3687016666,0.2434976846,-0.0720842332,-0.2631471455,0.2722587287,0.0614762381,0.1426138282,-0.1043646857,-0.3949477077,-0.4038264453,0.0597070605,0.0971546397,0.1380205601,-0.0552518591,-0.150805831,-0.1437973827,-0.0928228199,-0.1315386742,0.0542685091,-0.3340456784,0.0110975206,-0.1788978279,0.0363318175,-0.1240276173,-0.1814766228,0.0099526485,-0.2088931352,-0.1822603345,0.134678334,-0.0392420478,0.0597429164,-0.0656927228,0.0799402148,0.5429310203,0.4998341203,-0.1578968018,0.0591222942,0.0836744308,-0.315489471,-0.0443155728,0.2773605585,-0.0831806138,-0.1181732267,0.0842384174,0.1406008303,0.3385837972,-0.4173093438,-0.0755336955,-0.079886049,-0.202534005,0.3330042958,0.2637990117,-0.0864126757,-0.0999054983,0.3682368994,0.2149712294,-0.0356021039,-0.0385455564,0.1731214374,0.193765074,0.2517060637,0.0641473308,-0.0184769016,-0.0144703863,-0.1412945241,-0.0334264822,-0.2027126998,0.1448781788,-0.0525627658,0.0754862502,0.2267361134,0.3809486032,-0.0089712143,0.1959056407,-0.1053644419,0.1468940973,0.2529247701,0.0774802417,0.1793923378,-0.0227171537,-0.4215112627,-0.1569180936,0.1076810807,-0.1497696191,0.0696460009,-0.2426839173,-0.3709783554,-0.1268849373,-0.0896277204,-0.0408991612,0.5435215235,-0.1635982841,0.1739099771,-0.0686831921,0.047919713,0.0670228973,0.3332659006,-0.0928836912,0.2201660573,0.2191977054,0.3239123523,0.0336722955,0.4749951065,-0.0025974251,0.2313491255,-0.4086587131,0.3700209558,-0.2205614895,-0.152435258,-0.0604496039,0.2455563992,0.283398211,-0.0461558625,0.2276707739,0.2533397079,-0.294352591,0.251737535,-0.0129227191,0.0322241671,-0.2853594422,0.1191600487,-0.0543493889,-0.0460737683,-0.0495115854,-0.0612844713,-0.5117421746,0.1801471859,-0.2863976061,-0.2968678474,-0.0125351027,-0.0736484304,0.1288129985,-0.1876115501,0.2755152285,0.2395881265,0.4237524867,-0.0510504805,-0.3606442213,-0.0462603159,-0.1418848932,0.1449588984,-0.1167037413,0.1562411189,-0.0363338664,0.2164872289,0.3477029502,-0.1118953675,-0.3725241423,-0.3810081184,0.2364702672,-0.4007804692,0.3359997869,0.1742295921,0.0977543145,0.1720022112,-0.3217593431,-0.0427544042,0.0127450619,0.1785071045,0.0582197644,0.1562899351,0.0390236154,-0.2107220292,-0.1147495955,0.0297924131,0.3679977655,-0.1752372533,-0.1896431297,0.2272825539,-0.2485176027,-0.1982803047,0.0425052382,0.3440340161,0.3117296994,-0.0415380783,-0.3080849648,0.3648127913,0.0051476732,0.1454783678,0.4213912785,-0.1411507726,0.0980362371,0.0003456514,-0.0700884908,-0.0728384033,0.1219618991,-0.042959556,0.0887330994,0.0602393448,-0.201068446,0.3448042572,-0.0921594203,-0.2158393711,0.0264809281,0.0432168245,0.0877085403,-0.1586978436,-0.5564057827,-0.0904892311,0.1962942779,-0.0127187781,-0.0732560754,0.6972959042,0.0856290236,-0.1479186416,-0.129885301,0.1628668308,0.0622371659,-0.1279256791,-0.3026589453,-0.5541504025]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/880","title":"Add SQA","comments":"@thomwolf here's a slightly adapted version of the code from the [official Tapas repository](https:\/\/github.com\/google-research\/tapas\/blob\/master\/tapas\/utils\/interaction_utils.py) that is used to turn the `answer_coordinates` and `answer_texts` columns into true Python lists of tuples\/strings:\r\n\r\n```\r\nimport pandas as pd\r\nimport ast\r\n\r\ndata = pd.read_csv(\"\/content\/sqa_data\/random-split-1-dev.tsv\", sep='\\t')\r\n\r\ndef _parse_answer_coordinates(answer_coordinate_str):\r\n \"\"\"Parses the answer_coordinates of a question.\r\n Args:\r\n answer_coordinate_str: A string representation of a Python list of tuple\r\n strings.\r\n For example: \"['(1, 4)','(1, 3)', ...]\"\r\n \"\"\"\r\n\r\n try:\r\n answer_coordinates = []\r\n # make a list of strings\r\n coords = ast.literal_eval(answer_coordinate_str)\r\n # parse each string as a tuple\r\n for row_index, column_index in sorted(\r\n ast.literal_eval(coord) for coord in coords):\r\n answer_coordinates.append((row_index, column_index))\r\n except SyntaxError:\r\n raise ValueError('Unable to evaluate %s' % answer_coordinate_str)\r\n \r\n return answer_coordinates\r\n\r\n\r\ndef _parse_answer_text(answer_text):\r\n \"\"\"Populates the answer_texts field of `answer` by parsing `answer_text`.\r\n Args:\r\n answer_text: A string representation of a Python list of strings.\r\n For example: \"[u'test', u'hello', ...]\"\r\n \"\"\"\r\n try:\r\n answer = []\r\n for value in ast.literal_eval(answer_text):\r\n answer.append(value)\r\n except SyntaxError:\r\n raise ValueError('Unable to evaluate %s' % answer_text)\r\n\r\n return answer\r\n\r\ndata['answer_coordinates'] = data['answer_coordinates'].apply(lambda coords_str: _parse_answer_coordinates(coords_str))\r\ndata['answer_text'] = data['answer_text'].apply(lambda txt: _parse_answer_text(txt))\r\n```\r\n\r\nHere I'm using Pandas to read in one of the TSV files (the dev set). \r\n\r\n","body":"## Adding a Dataset\r\n- **Name:** SQA (Sequential Question Answering) by Microsoft. \r\n- **Description:** The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. It has 6,066 sequences with 17,553 questions in total.\r\n- **Paper:** https:\/\/www.microsoft.com\/en-us\/research\/publication\/search-based-neural-structured-learning-sequential-question-answering\/\r\n- **Data:** https:\/\/www.microsoft.com\/en-us\/download\/details.aspx?id=54253\r\n- **Motivation:** currently, the [Tapas](https:\/\/ai.googleblog.com\/2020\/04\/using-neural-networks-to-find-answers.html) algorithm by Google AI is being added to the Transformers library (see https:\/\/github.com\/huggingface\/transformers\/pull\/8113). It would be great to use that model in combination with this dataset, on which it achieves SOTA results (average question accuracy of 0.71).\r\n\r\nNote 1: this dataset actually consists of 2 types of files: \r\n1) TSV files, containing the questions, answer coordinates and answer texts (for training, dev and test)\r\n2) a folder of csv files, which contain the actual tabular data\r\n\r\nNote 2: if you download the dataset straight from the download link above, then you will see that the `answer_coordinates` and `answer_text` columns are string lists of string tuples and strings respectively, which is not ideal. It would be better to make them true Python lists of tuples and strings respectively (using `ast.literal_eval`), before uploading them to the HuggingFace hub.\r\n\r\nAdding this would be great! Then we could possibly also add [WTQ (WikiTable Questions)](https:\/\/github.com\/ppasupat\/WikiTableQuestions) and [TabFact (Tabular Fact Checking)](https:\/\/github.com\/wenhuchen\/Table-Fact-Checking) on which TAPAS also achieves state-of-the-art results. Note that the TAPAS algorithm requires these datasets to first be converted into the SQA format.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":185,"text":"Add SQA\n## Adding a Dataset\r\n- **Name:** SQA (Sequential Question Answering) by Microsoft. \r\n- **Description:** The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. It has 6,066 sequences with 17,553 questions in total.\r\n- **Paper:** https:\/\/www.microsoft.com\/en-us\/research\/publication\/search-based-neural-structured-learning-sequential-question-answering\/\r\n- **Data:** https:\/\/www.microsoft.com\/en-us\/download\/details.aspx?id=54253\r\n- **Motivation:** currently, the [Tapas](https:\/\/ai.googleblog.com\/2020\/04\/using-neural-networks-to-find-answers.html) algorithm by Google AI is being added to the Transformers library (see https:\/\/github.com\/huggingface\/transformers\/pull\/8113). It would be great to use that model in combination with this dataset, on which it achieves SOTA results (average question accuracy of 0.71).\r\n\r\nNote 1: this dataset actually consists of 2 types of files: \r\n1) TSV files, containing the questions, answer coordinates and answer texts (for training, dev and test)\r\n2) a folder of csv files, which contain the actual tabular data\r\n\r\nNote 2: if you download the dataset straight from the download link above, then you will see that the `answer_coordinates` and `answer_text` columns are string lists of string tuples and strings respectively, which is not ideal. It would be better to make them true Python lists of tuples and strings respectively (using `ast.literal_eval`), before uploading them to the HuggingFace hub.\r\n\r\nAdding this would be great! Then we could possibly also add [WTQ (WikiTable Questions)](https:\/\/github.com\/ppasupat\/WikiTableQuestions) and [TabFact (Tabular Fact Checking)](https:\/\/github.com\/wenhuchen\/Table-Fact-Checking) on which TAPAS also achieves state-of-the-art results. Note that the TAPAS algorithm requires these datasets to first be converted into the SQA format.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\n@thomwolf here's a slightly adapted version of the code from the [official Tapas repository](https:\/\/github.com\/google-research\/tapas\/blob\/master\/tapas\/utils\/interaction_utils.py) that is used to turn the `answer_coordinates` and `answer_texts` columns into true Python lists of tuples\/strings:\r\n\r\n```\r\nimport pandas as pd\r\nimport ast\r\n\r\ndata = pd.read_csv(\"\/content\/sqa_data\/random-split-1-dev.tsv\", sep='\\t')\r\n\r\ndef _parse_answer_coordinates(answer_coordinate_str):\r\n \"\"\"Parses the answer_coordinates of a question.\r\n Args:\r\n answer_coordinate_str: A string representation of a Python list of tuple\r\n strings.\r\n For example: \"['(1, 4)','(1, 3)', ...]\"\r\n \"\"\"\r\n\r\n try:\r\n answer_coordinates = []\r\n # make a list of strings\r\n coords = ast.literal_eval(answer_coordinate_str)\r\n # parse each string as a tuple\r\n for row_index, column_index in sorted(\r\n ast.literal_eval(coord) for coord in coords):\r\n answer_coordinates.append((row_index, column_index))\r\n except SyntaxError:\r\n raise ValueError('Unable to evaluate %s' % answer_coordinate_str)\r\n \r\n return answer_coordinates\r\n\r\n\r\ndef _parse_answer_text(answer_text):\r\n \"\"\"Populates the answer_texts field of `answer` by parsing `answer_text`.\r\n Args:\r\n answer_text: A string representation of a Python list of strings.\r\n For example: \"[u'test', u'hello', ...]\"\r\n \"\"\"\r\n try:\r\n answer = []\r\n for value in ast.literal_eval(answer_text):\r\n answer.append(value)\r\n except SyntaxError:\r\n raise ValueError('Unable to evaluate %s' % answer_text)\r\n\r\n return answer\r\n\r\ndata['answer_coordinates'] = data['answer_coordinates'].apply(lambda coords_str: _parse_answer_coordinates(coords_str))\r\ndata['answer_text'] = data['answer_text'].apply(lambda txt: _parse_answer_text(txt))\r\n```\r\n\r\nHere I'm using Pandas to read in one of the TSV files (the dev set). \r\n\r\n","embeddings":[-0.0640092716,-0.2312078327,-0.2099176347,-0.1210695952,0.0999142751,-0.1975832283,0.0337971561,0.3069660664,0.1483602226,0.0562699474,-0.1675558388,-0.002632767,-0.0712603852,0.5443822742,0.1500162035,-0.1045532525,0.0272174533,0.1428327858,-0.2944614291,0.0711858869,-0.196293667,0.0914863199,0.0215712246,0.031229699,-0.2368902415,-0.0844550505,-0.1073180363,-0.1923377961,-0.1904111356,-0.1972694844,0.0054088389,0.072247088,0.1686420888,0.2730415761,-0.0001059325,-0.1343121231,-0.0751999691,-0.0975328833,-0.0939193591,0.0506609641,-0.1607964188,0.0926124603,-0.1110516489,-0.0771295801,-0.5252555013,0.2041900754,-0.0748102888,-0.3007471859,0.5820878148,0.2628435194,0.2186131328,0.0928574726,0.0751176327,-0.2174711227,0.3752230704,0.0049580853,-0.3586767614,-0.0065896218,0.4021380246,-0.348051697,0.0548400171,0.4530360699,0.265696764,-0.2276320606,0.1525066942,0.1255038083,0.2736350596,0.0334795527,0.0674656704,0.2694988251,0.3335775435,-0.0927536935,-0.1821043491,-0.1886140257,0.0144181484,0.1235252768,-0.2833782434,0.118624717,-0.3490382135,0.0167793985,-0.1807740331,-0.2453861386,-0.349671483,0.101798743,0.0179150905,0.5167180896,-0.0192335062,-0.1948671192,-0.0831076577,-0.1331273913,-0.0427691936,-0.0472527817,0.2443597317,0.2530235648,-0.5382785201,-0.2720508277,0.2136117667,0.1150445491,0.2435956448,0.1921216249,0.4720071256,-0.0723904297,0.0314646773,-0.2965055406,0.1672351956,0.4781259894,-0.003513505,0.0426871032,-0.1635618806,0.1145740077,0.069656536,0.0185354222,0.0113641415,0.4373466372,-0.2889096141,0.0332337134,0.121735774,0.0777647123,-0.3187230825,0.2000280917,-0.2216603905,-0.0858714953,-0.0461771674,0.3067342937,-0.2722217441,0.0963958725,0.1617859751,0.1525750607,0.0376236141,-0.1205838397,-0.0992913097,-0.1381375194,-0.1389065683,0.327624768,0.2592454553,0.0756572187,-0.118240878,-0.1105898842,0.1236674637,0.0671117604,-0.1190187186,-0.1257779002,0.2719293535,0.0063779862,-0.1196178198,-0.2474647164,0.006988266,-0.1855606288,-0.0826086476,0.2046098411,0.086625129,0.0822594985,-0.0458444618,0.2935351431,-0.0351122767,-0.3378953934,0.0560649224,0.3847602308,0.0544066653,-0.0786112845,0.1612036526,0.0770690739,-0.1640680432,-0.0699188039,0.2680043876,0.1623664647,0.0437387638,0.0179177821,-0.0602712929,0.2952128351,-0.4049222767,0.1558105201,0.1623992175,0.4924613237,0.1605460048,0.4469804764,-0.3020619452,0.042209927,-0.051084321,-0.1068252027,-0.0611654036,0.2619523108,-0.0233064312,0.222698912,0.1795921922,-0.0981692225,0.124762632,0.2569143176,-0.243451342,-0.0364894234,-0.2605640292,-0.320944041,-0.1499199122,0.5569957495,-0.0081572412,-0.0947261527,-0.1503633857,-0.2269732505,0.0897196904,-0.2418155223,-0.1053456739,-0.202632457,0.418138355,-0.1816864908,0.0669904649,-0.5491580367,-0.5756059885,-0.1482614726,-0.1463883817,0.2786490321,0.0339066721,-0.4509789348,-0.3733327091,0.1327016652,0.0075063659,-0.0343044698,0.2479703277,0.1869211048,0.2428165525,-0.0500193909,-0.0937463418,0.1550153494,-0.0424523093,-0.1147119105,-0.1329626888,0.198689729,-0.1225987598,0.0456116945,-0.0185674727,0.314791292,-0.1338375211,0.1751381606,0.3347406685,0.1550746709,-0.13191697,-0.2625659704,0.0805614889,0.4734991193,0.0326889567,-0.0527552143,-0.1923760474,0.4993881285,0.1238283738,-0.0980774164,-0.2853282094,0.4043613374,-0.1816473901,0.2681885958,0.1508674324,0.0346501954,-0.0117136929,-0.0824667439,-0.131069541,0.1497388333,0.4146442711,0.1367218494,-0.2725359797,-0.0683397576,-0.3361949921,0.3972062767,0.2918215394,-0.2563173175,0.3381398618,0.0930465311,-0.0448115245,-0.099624306,0.0174711868,0.1898336112,0.1904856265,0.2655975521,-0.1340573281,0.132565096,-0.0351311676,-0.1304928362,0.0047888304,0.0729855746,-0.1076816693,0.3593825102,0.3060602844,0.1129134521,-0.0875147507,-0.081166327,-0.2993248403,-0.0468394086,-0.0217849854,0.1793799549,0.2182229459,-0.2264639288,0.0334007703,-0.3110869825,0.0157671552,0.0116476575,0.6521463394,-0.1362127811,-0.1908007115,0.2219664752,0.074260965,0.910795629,-0.1697585285,-0.0477678366,-0.2234660834,-0.2786903083,0.1275860667,0.3061983287,0.0774174184,-0.0461224802,0.5052061081,-0.1818101406,0.3880173266,-0.0880887881,-0.3654263318,0.0920002386,-0.0208264478,0.0868983418,-0.0135212597,0.0519566387,-0.2462124228,-0.311757654,0.0962447152,-0.2440950423,-0.1063671485,-0.0403895676,0.2175629735,-0.090775229,-0.1126224622,-0.4963790476,-0.0617724843,-0.3034152091,0.4760821164,0.2297993004,-0.0288607217,-0.0194287114,-0.0775105655,0.0343687013,-0.1526608765,0.1068070829,-0.2156098038,0.1638543606,0.2227652371,-0.1240256354,-0.211823523,-0.4943149984,-0.3110443056,-0.0237979498,0.0763129145,-0.1710079461,-0.0684284568,-0.2200554162,0.0027596692,-0.1519955397,0.1327385157,0.4515957832,-0.0766791403,-0.2075061649,-0.0513131246,0.2085881829,0.5510808229,0.0616792701,-0.1596296579,0.1330932677,0.2265027463,0.0326379724,0.6488921046,0.0589807928,0.0846870765,0.2585291862,0.0645583197,-0.11768765,-0.2783655524,0.0752346739,-0.1934869885,-0.1577434689,0.0985051915,0.4758580923,0.2683967352,0.0449928492,-0.1529213935,0.1253436506,-0.2520808578,-0.1979157031,0.4055438638,0.0370066762,0.4625874758,-0.0875199139,-0.1534026265,0.007968137,-0.355032146,0.0913981497,-0.0922903121,0.0258747675,-0.4211962223,-0.5453454852,-0.2182356119,-0.2534332573,0.3696666658,0.4543704987,-0.0103512472,-0.0700463057,-0.4352989197,-0.0819312185,-0.0756269917,0.1700690091,0.062229462,-0.351492703,0.0121095553,-0.2442807257,0.0773436651,0.1962990165,-0.3358927071,-0.3367526531,-0.0537999533,0.3874261975,-0.3562621772,0.0581251197,0.2264221758,-0.0486682057,-0.0562434755,0.0086918687,-0.0854227766,-0.1264292002,0.0051181843,0.2333197594,0.3687148988,0.036403697,0.2235229313,-0.0396039337,-0.0924399942,-0.0361280143,0.0263197273,0.1368793249,0.0830826312,0.0011648067,0.1247434616,-0.1430703998,-0.1977378577,-0.0488720126,0.4001817107,-0.2085921615,-0.0698705316,-0.1505306065,-0.1544058472,-0.0211568605,0.2488559783,0.1011996493,0.3039302826,0.0348631851,0.1376393139,-0.2829020023,-0.1738834679,-0.1508074105,-0.0058538243,0.1136687249,-0.1343204975,0.4686171412,-0.2742679119,-0.2376136333,-0.0322412364,0.4682995975,-0.1962357312,0.1028005555,0.0156817846,1.0317857265,-0.1658943444,0.3711087406,0.141588971,-0.0974734575,0.2048790306,-0.350300014,0.144818902,-0.1310677826,-0.3593126833,-0.0561568215,0.0399527773,-0.0805622116,-0.004561604,-0.3912659585,0.1389491707,0.175603494,0.3666844964,0.2946521938,0.2584328055,0.069652155,-0.0152626121,-0.5002795458,0.2108229697,0.1116845459,-0.2397036254,-0.0502777211,0.0481445976,-0.1371403933,-0.1277022213,-0.1942481101,0.0707471296,-0.4300709963,-0.0462026186,0.2861474156,-0.1057446077,0.1106237322,0.2124155015,-0.0267282948,-0.0604788736,-0.3943512142,0.0101640364,-0.2656233609,0.2279188782,0.0728195086,-0.0291024782,0.292478174,-0.2039584965,-0.1170177087,0.2273564041,0.3363114595,-0.4757345021,-0.4612273872,0.0721802711,0.3284933567,-0.1256040484,0.0085265655,0.1294416487,-0.1191570759,-0.1417362243,0.2308954149,0.0601020046,0.0594073758,0.1053040102,0.4885711968,-0.1789277941,-0.0357294269,0.117661871,0.4107621014,-0.0385524631,0.2471322268,0.3028377593,-0.2766368091,-0.4391760826,-0.1043767333,0.2383774668,-0.4756156504,-0.029951755,-0.1643128544,-0.3447369039,-0.0333019234,0.5254306793,0.4061190486,-0.0993947014,-0.060709253,-0.0520864055,0.103038922,0.627266705,0.3372285366,0.1448047608,0.0963591561,-0.18211326,-0.0878586471,0.244761616,-0.4542951584,-0.0925646946,-0.3687016666,0.2434976846,-0.0720842332,-0.2631471455,0.2722587287,0.0614762381,0.1426138282,-0.1043646857,-0.3949477077,-0.4038264453,0.0597070605,0.0971546397,0.1380205601,-0.0552518591,-0.150805831,-0.1437973827,-0.0928228199,-0.1315386742,0.0542685091,-0.3340456784,0.0110975206,-0.1788978279,0.0363318175,-0.1240276173,-0.1814766228,0.0099526485,-0.2088931352,-0.1822603345,0.134678334,-0.0392420478,0.0597429164,-0.0656927228,0.0799402148,0.5429310203,0.4998341203,-0.1578968018,0.0591222942,0.0836744308,-0.315489471,-0.0443155728,0.2773605585,-0.0831806138,-0.1181732267,0.0842384174,0.1406008303,0.3385837972,-0.4173093438,-0.0755336955,-0.079886049,-0.202534005,0.3330042958,0.2637990117,-0.0864126757,-0.0999054983,0.3682368994,0.2149712294,-0.0356021039,-0.0385455564,0.1731214374,0.193765074,0.2517060637,0.0641473308,-0.0184769016,-0.0144703863,-0.1412945241,-0.0334264822,-0.2027126998,0.1448781788,-0.0525627658,0.0754862502,0.2267361134,0.3809486032,-0.0089712143,0.1959056407,-0.1053644419,0.1468940973,0.2529247701,0.0774802417,0.1793923378,-0.0227171537,-0.4215112627,-0.1569180936,0.1076810807,-0.1497696191,0.0696460009,-0.2426839173,-0.3709783554,-0.1268849373,-0.0896277204,-0.0408991612,0.5435215235,-0.1635982841,0.1739099771,-0.0686831921,0.047919713,0.0670228973,0.3332659006,-0.0928836912,0.2201660573,0.2191977054,0.3239123523,0.0336722955,0.4749951065,-0.0025974251,0.2313491255,-0.4086587131,0.3700209558,-0.2205614895,-0.152435258,-0.0604496039,0.2455563992,0.283398211,-0.0461558625,0.2276707739,0.2533397079,-0.294352591,0.251737535,-0.0129227191,0.0322241671,-0.2853594422,0.1191600487,-0.0543493889,-0.0460737683,-0.0495115854,-0.0612844713,-0.5117421746,0.1801471859,-0.2863976061,-0.2968678474,-0.0125351027,-0.0736484304,0.1288129985,-0.1876115501,0.2755152285,0.2395881265,0.4237524867,-0.0510504805,-0.3606442213,-0.0462603159,-0.1418848932,0.1449588984,-0.1167037413,0.1562411189,-0.0363338664,0.2164872289,0.3477029502,-0.1118953675,-0.3725241423,-0.3810081184,0.2364702672,-0.4007804692,0.3359997869,0.1742295921,0.0977543145,0.1720022112,-0.3217593431,-0.0427544042,0.0127450619,0.1785071045,0.0582197644,0.1562899351,0.0390236154,-0.2107220292,-0.1147495955,0.0297924131,0.3679977655,-0.1752372533,-0.1896431297,0.2272825539,-0.2485176027,-0.1982803047,0.0425052382,0.3440340161,0.3117296994,-0.0415380783,-0.3080849648,0.3648127913,0.0051476732,0.1454783678,0.4213912785,-0.1411507726,0.0980362371,0.0003456514,-0.0700884908,-0.0728384033,0.1219618991,-0.042959556,0.0887330994,0.0602393448,-0.201068446,0.3448042572,-0.0921594203,-0.2158393711,0.0264809281,0.0432168245,0.0877085403,-0.1586978436,-0.5564057827,-0.0904892311,0.1962942779,-0.0127187781,-0.0732560754,0.6972959042,0.0856290236,-0.1479186416,-0.129885301,0.1628668308,0.0622371659,-0.1279256791,-0.3026589453,-0.5541504025]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/880","title":"Add SQA","comments":"Closing since SQA was added in #1566 ","body":"## Adding a Dataset\r\n- **Name:** SQA (Sequential Question Answering) by Microsoft. \r\n- **Description:** The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. It has 6,066 sequences with 17,553 questions in total.\r\n- **Paper:** https:\/\/www.microsoft.com\/en-us\/research\/publication\/search-based-neural-structured-learning-sequential-question-answering\/\r\n- **Data:** https:\/\/www.microsoft.com\/en-us\/download\/details.aspx?id=54253\r\n- **Motivation:** currently, the [Tapas](https:\/\/ai.googleblog.com\/2020\/04\/using-neural-networks-to-find-answers.html) algorithm by Google AI is being added to the Transformers library (see https:\/\/github.com\/huggingface\/transformers\/pull\/8113). It would be great to use that model in combination with this dataset, on which it achieves SOTA results (average question accuracy of 0.71).\r\n\r\nNote 1: this dataset actually consists of 2 types of files: \r\n1) TSV files, containing the questions, answer coordinates and answer texts (for training, dev and test)\r\n2) a folder of csv files, which contain the actual tabular data\r\n\r\nNote 2: if you download the dataset straight from the download link above, then you will see that the `answer_coordinates` and `answer_text` columns are string lists of string tuples and strings respectively, which is not ideal. It would be better to make them true Python lists of tuples and strings respectively (using `ast.literal_eval`), before uploading them to the HuggingFace hub.\r\n\r\nAdding this would be great! Then we could possibly also add [WTQ (WikiTable Questions)](https:\/\/github.com\/ppasupat\/WikiTableQuestions) and [TabFact (Tabular Fact Checking)](https:\/\/github.com\/wenhuchen\/Table-Fact-Checking) on which TAPAS also achieves state-of-the-art results. Note that the TAPAS algorithm requires these datasets to first be converted into the SQA format.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":7,"text":"Add SQA\n## Adding a Dataset\r\n- **Name:** SQA (Sequential Question Answering) by Microsoft. \r\n- **Description:** The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. It has 6,066 sequences with 17,553 questions in total.\r\n- **Paper:** https:\/\/www.microsoft.com\/en-us\/research\/publication\/search-based-neural-structured-learning-sequential-question-answering\/\r\n- **Data:** https:\/\/www.microsoft.com\/en-us\/download\/details.aspx?id=54253\r\n- **Motivation:** currently, the [Tapas](https:\/\/ai.googleblog.com\/2020\/04\/using-neural-networks-to-find-answers.html) algorithm by Google AI is being added to the Transformers library (see https:\/\/github.com\/huggingface\/transformers\/pull\/8113). It would be great to use that model in combination with this dataset, on which it achieves SOTA results (average question accuracy of 0.71).\r\n\r\nNote 1: this dataset actually consists of 2 types of files: \r\n1) TSV files, containing the questions, answer coordinates and answer texts (for training, dev and test)\r\n2) a folder of csv files, which contain the actual tabular data\r\n\r\nNote 2: if you download the dataset straight from the download link above, then you will see that the `answer_coordinates` and `answer_text` columns are string lists of string tuples and strings respectively, which is not ideal. It would be better to make them true Python lists of tuples and strings respectively (using `ast.literal_eval`), before uploading them to the HuggingFace hub.\r\n\r\nAdding this would be great! Then we could possibly also add [WTQ (WikiTable Questions)](https:\/\/github.com\/ppasupat\/WikiTableQuestions) and [TabFact (Tabular Fact Checking)](https:\/\/github.com\/wenhuchen\/Table-Fact-Checking) on which TAPAS also achieves state-of-the-art results. Note that the TAPAS algorithm requires these datasets to first be converted into the SQA format.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nClosing since SQA was added in #1566 ","embeddings":[-0.0640092716,-0.2312078327,-0.2099176347,-0.1210695952,0.0999142751,-0.1975832283,0.0337971561,0.3069660664,0.1483602226,0.0562699474,-0.1675558388,-0.002632767,-0.0712603852,0.5443822742,0.1500162035,-0.1045532525,0.0272174533,0.1428327858,-0.2944614291,0.0711858869,-0.196293667,0.0914863199,0.0215712246,0.031229699,-0.2368902415,-0.0844550505,-0.1073180363,-0.1923377961,-0.1904111356,-0.1972694844,0.0054088389,0.072247088,0.1686420888,0.2730415761,-0.0001059325,-0.1343121231,-0.0751999691,-0.0975328833,-0.0939193591,0.0506609641,-0.1607964188,0.0926124603,-0.1110516489,-0.0771295801,-0.5252555013,0.2041900754,-0.0748102888,-0.3007471859,0.5820878148,0.2628435194,0.2186131328,0.0928574726,0.0751176327,-0.2174711227,0.3752230704,0.0049580853,-0.3586767614,-0.0065896218,0.4021380246,-0.348051697,0.0548400171,0.4530360699,0.265696764,-0.2276320606,0.1525066942,0.1255038083,0.2736350596,0.0334795527,0.0674656704,0.2694988251,0.3335775435,-0.0927536935,-0.1821043491,-0.1886140257,0.0144181484,0.1235252768,-0.2833782434,0.118624717,-0.3490382135,0.0167793985,-0.1807740331,-0.2453861386,-0.349671483,0.101798743,0.0179150905,0.5167180896,-0.0192335062,-0.1948671192,-0.0831076577,-0.1331273913,-0.0427691936,-0.0472527817,0.2443597317,0.2530235648,-0.5382785201,-0.2720508277,0.2136117667,0.1150445491,0.2435956448,0.1921216249,0.4720071256,-0.0723904297,0.0314646773,-0.2965055406,0.1672351956,0.4781259894,-0.003513505,0.0426871032,-0.1635618806,0.1145740077,0.069656536,0.0185354222,0.0113641415,0.4373466372,-0.2889096141,0.0332337134,0.121735774,0.0777647123,-0.3187230825,0.2000280917,-0.2216603905,-0.0858714953,-0.0461771674,0.3067342937,-0.2722217441,0.0963958725,0.1617859751,0.1525750607,0.0376236141,-0.1205838397,-0.0992913097,-0.1381375194,-0.1389065683,0.327624768,0.2592454553,0.0756572187,-0.118240878,-0.1105898842,0.1236674637,0.0671117604,-0.1190187186,-0.1257779002,0.2719293535,0.0063779862,-0.1196178198,-0.2474647164,0.006988266,-0.1855606288,-0.0826086476,0.2046098411,0.086625129,0.0822594985,-0.0458444618,0.2935351431,-0.0351122767,-0.3378953934,0.0560649224,0.3847602308,0.0544066653,-0.0786112845,0.1612036526,0.0770690739,-0.1640680432,-0.0699188039,0.2680043876,0.1623664647,0.0437387638,0.0179177821,-0.0602712929,0.2952128351,-0.4049222767,0.1558105201,0.1623992175,0.4924613237,0.1605460048,0.4469804764,-0.3020619452,0.042209927,-0.051084321,-0.1068252027,-0.0611654036,0.2619523108,-0.0233064312,0.222698912,0.1795921922,-0.0981692225,0.124762632,0.2569143176,-0.243451342,-0.0364894234,-0.2605640292,-0.320944041,-0.1499199122,0.5569957495,-0.0081572412,-0.0947261527,-0.1503633857,-0.2269732505,0.0897196904,-0.2418155223,-0.1053456739,-0.202632457,0.418138355,-0.1816864908,0.0669904649,-0.5491580367,-0.5756059885,-0.1482614726,-0.1463883817,0.2786490321,0.0339066721,-0.4509789348,-0.3733327091,0.1327016652,0.0075063659,-0.0343044698,0.2479703277,0.1869211048,0.2428165525,-0.0500193909,-0.0937463418,0.1550153494,-0.0424523093,-0.1147119105,-0.1329626888,0.198689729,-0.1225987598,0.0456116945,-0.0185674727,0.314791292,-0.1338375211,0.1751381606,0.3347406685,0.1550746709,-0.13191697,-0.2625659704,0.0805614889,0.4734991193,0.0326889567,-0.0527552143,-0.1923760474,0.4993881285,0.1238283738,-0.0980774164,-0.2853282094,0.4043613374,-0.1816473901,0.2681885958,0.1508674324,0.0346501954,-0.0117136929,-0.0824667439,-0.131069541,0.1497388333,0.4146442711,0.1367218494,-0.2725359797,-0.0683397576,-0.3361949921,0.3972062767,0.2918215394,-0.2563173175,0.3381398618,0.0930465311,-0.0448115245,-0.099624306,0.0174711868,0.1898336112,0.1904856265,0.2655975521,-0.1340573281,0.132565096,-0.0351311676,-0.1304928362,0.0047888304,0.0729855746,-0.1076816693,0.3593825102,0.3060602844,0.1129134521,-0.0875147507,-0.081166327,-0.2993248403,-0.0468394086,-0.0217849854,0.1793799549,0.2182229459,-0.2264639288,0.0334007703,-0.3110869825,0.0157671552,0.0116476575,0.6521463394,-0.1362127811,-0.1908007115,0.2219664752,0.074260965,0.910795629,-0.1697585285,-0.0477678366,-0.2234660834,-0.2786903083,0.1275860667,0.3061983287,0.0774174184,-0.0461224802,0.5052061081,-0.1818101406,0.3880173266,-0.0880887881,-0.3654263318,0.0920002386,-0.0208264478,0.0868983418,-0.0135212597,0.0519566387,-0.2462124228,-0.311757654,0.0962447152,-0.2440950423,-0.1063671485,-0.0403895676,0.2175629735,-0.090775229,-0.1126224622,-0.4963790476,-0.0617724843,-0.3034152091,0.4760821164,0.2297993004,-0.0288607217,-0.0194287114,-0.0775105655,0.0343687013,-0.1526608765,0.1068070829,-0.2156098038,0.1638543606,0.2227652371,-0.1240256354,-0.211823523,-0.4943149984,-0.3110443056,-0.0237979498,0.0763129145,-0.1710079461,-0.0684284568,-0.2200554162,0.0027596692,-0.1519955397,0.1327385157,0.4515957832,-0.0766791403,-0.2075061649,-0.0513131246,0.2085881829,0.5510808229,0.0616792701,-0.1596296579,0.1330932677,0.2265027463,0.0326379724,0.6488921046,0.0589807928,0.0846870765,0.2585291862,0.0645583197,-0.11768765,-0.2783655524,0.0752346739,-0.1934869885,-0.1577434689,0.0985051915,0.4758580923,0.2683967352,0.0449928492,-0.1529213935,0.1253436506,-0.2520808578,-0.1979157031,0.4055438638,0.0370066762,0.4625874758,-0.0875199139,-0.1534026265,0.007968137,-0.355032146,0.0913981497,-0.0922903121,0.0258747675,-0.4211962223,-0.5453454852,-0.2182356119,-0.2534332573,0.3696666658,0.4543704987,-0.0103512472,-0.0700463057,-0.4352989197,-0.0819312185,-0.0756269917,0.1700690091,0.062229462,-0.351492703,0.0121095553,-0.2442807257,0.0773436651,0.1962990165,-0.3358927071,-0.3367526531,-0.0537999533,0.3874261975,-0.3562621772,0.0581251197,0.2264221758,-0.0486682057,-0.0562434755,0.0086918687,-0.0854227766,-0.1264292002,0.0051181843,0.2333197594,0.3687148988,0.036403697,0.2235229313,-0.0396039337,-0.0924399942,-0.0361280143,0.0263197273,0.1368793249,0.0830826312,0.0011648067,0.1247434616,-0.1430703998,-0.1977378577,-0.0488720126,0.4001817107,-0.2085921615,-0.0698705316,-0.1505306065,-0.1544058472,-0.0211568605,0.2488559783,0.1011996493,0.3039302826,0.0348631851,0.1376393139,-0.2829020023,-0.1738834679,-0.1508074105,-0.0058538243,0.1136687249,-0.1343204975,0.4686171412,-0.2742679119,-0.2376136333,-0.0322412364,0.4682995975,-0.1962357312,0.1028005555,0.0156817846,1.0317857265,-0.1658943444,0.3711087406,0.141588971,-0.0974734575,0.2048790306,-0.350300014,0.144818902,-0.1310677826,-0.3593126833,-0.0561568215,0.0399527773,-0.0805622116,-0.004561604,-0.3912659585,0.1389491707,0.175603494,0.3666844964,0.2946521938,0.2584328055,0.069652155,-0.0152626121,-0.5002795458,0.2108229697,0.1116845459,-0.2397036254,-0.0502777211,0.0481445976,-0.1371403933,-0.1277022213,-0.1942481101,0.0707471296,-0.4300709963,-0.0462026186,0.2861474156,-0.1057446077,0.1106237322,0.2124155015,-0.0267282948,-0.0604788736,-0.3943512142,0.0101640364,-0.2656233609,0.2279188782,0.0728195086,-0.0291024782,0.292478174,-0.2039584965,-0.1170177087,0.2273564041,0.3363114595,-0.4757345021,-0.4612273872,0.0721802711,0.3284933567,-0.1256040484,0.0085265655,0.1294416487,-0.1191570759,-0.1417362243,0.2308954149,0.0601020046,0.0594073758,0.1053040102,0.4885711968,-0.1789277941,-0.0357294269,0.117661871,0.4107621014,-0.0385524631,0.2471322268,0.3028377593,-0.2766368091,-0.4391760826,-0.1043767333,0.2383774668,-0.4756156504,-0.029951755,-0.1643128544,-0.3447369039,-0.0333019234,0.5254306793,0.4061190486,-0.0993947014,-0.060709253,-0.0520864055,0.103038922,0.627266705,0.3372285366,0.1448047608,0.0963591561,-0.18211326,-0.0878586471,0.244761616,-0.4542951584,-0.0925646946,-0.3687016666,0.2434976846,-0.0720842332,-0.2631471455,0.2722587287,0.0614762381,0.1426138282,-0.1043646857,-0.3949477077,-0.4038264453,0.0597070605,0.0971546397,0.1380205601,-0.0552518591,-0.150805831,-0.1437973827,-0.0928228199,-0.1315386742,0.0542685091,-0.3340456784,0.0110975206,-0.1788978279,0.0363318175,-0.1240276173,-0.1814766228,0.0099526485,-0.2088931352,-0.1822603345,0.134678334,-0.0392420478,0.0597429164,-0.0656927228,0.0799402148,0.5429310203,0.4998341203,-0.1578968018,0.0591222942,0.0836744308,-0.315489471,-0.0443155728,0.2773605585,-0.0831806138,-0.1181732267,0.0842384174,0.1406008303,0.3385837972,-0.4173093438,-0.0755336955,-0.079886049,-0.202534005,0.3330042958,0.2637990117,-0.0864126757,-0.0999054983,0.3682368994,0.2149712294,-0.0356021039,-0.0385455564,0.1731214374,0.193765074,0.2517060637,0.0641473308,-0.0184769016,-0.0144703863,-0.1412945241,-0.0334264822,-0.2027126998,0.1448781788,-0.0525627658,0.0754862502,0.2267361134,0.3809486032,-0.0089712143,0.1959056407,-0.1053644419,0.1468940973,0.2529247701,0.0774802417,0.1793923378,-0.0227171537,-0.4215112627,-0.1569180936,0.1076810807,-0.1497696191,0.0696460009,-0.2426839173,-0.3709783554,-0.1268849373,-0.0896277204,-0.0408991612,0.5435215235,-0.1635982841,0.1739099771,-0.0686831921,0.047919713,0.0670228973,0.3332659006,-0.0928836912,0.2201660573,0.2191977054,0.3239123523,0.0336722955,0.4749951065,-0.0025974251,0.2313491255,-0.4086587131,0.3700209558,-0.2205614895,-0.152435258,-0.0604496039,0.2455563992,0.283398211,-0.0461558625,0.2276707739,0.2533397079,-0.294352591,0.251737535,-0.0129227191,0.0322241671,-0.2853594422,0.1191600487,-0.0543493889,-0.0460737683,-0.0495115854,-0.0612844713,-0.5117421746,0.1801471859,-0.2863976061,-0.2968678474,-0.0125351027,-0.0736484304,0.1288129985,-0.1876115501,0.2755152285,0.2395881265,0.4237524867,-0.0510504805,-0.3606442213,-0.0462603159,-0.1418848932,0.1449588984,-0.1167037413,0.1562411189,-0.0363338664,0.2164872289,0.3477029502,-0.1118953675,-0.3725241423,-0.3810081184,0.2364702672,-0.4007804692,0.3359997869,0.1742295921,0.0977543145,0.1720022112,-0.3217593431,-0.0427544042,0.0127450619,0.1785071045,0.0582197644,0.1562899351,0.0390236154,-0.2107220292,-0.1147495955,0.0297924131,0.3679977655,-0.1752372533,-0.1896431297,0.2272825539,-0.2485176027,-0.1982803047,0.0425052382,0.3440340161,0.3117296994,-0.0415380783,-0.3080849648,0.3648127913,0.0051476732,0.1454783678,0.4213912785,-0.1411507726,0.0980362371,0.0003456514,-0.0700884908,-0.0728384033,0.1219618991,-0.042959556,0.0887330994,0.0602393448,-0.201068446,0.3448042572,-0.0921594203,-0.2158393711,0.0264809281,0.0432168245,0.0877085403,-0.1586978436,-0.5564057827,-0.0904892311,0.1962942779,-0.0127187781,-0.0732560754,0.6972959042,0.0856290236,-0.1479186416,-0.129885301,0.1628668308,0.0622371659,-0.1279256791,-0.3026589453,-0.5541504025]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/879","title":"boolq does not load ","comments":"Hi ! It runs on my side without issues. I tried\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"boolq\")\r\n```\r\n\r\nWhat version of datasets and tensorflow are your runnning ?\r\nAlso if you manage to get a minimal reproducible script (on google colab for example) that would be useful.","body":"Hi\r\nI am getting these errors trying to load boolq thanks \r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 5, in <module>\r\n data = AutoTask().get(\"boolq\").get_dataset(\"train\", n_obs=10)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 42, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 38, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 150, in download_custom\r\n get_from_cache(url, cache_dir=cache_dir, local_files_only=True, use_etag=False)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 472, in get_from_cache\r\n f\"Cannot find the requested files in the cached path at {cache_path} and outgoing traffic has been\"\r\nFileNotFoundError: Cannot find the requested files in the cached path at \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/eaee069e38f6ceaa84de02ad088c34e63ec97671f2cd1910ddb16b10dc60808c and outgoing traffic has been disabled. To enable file online look-ups, set 'local_files_only' to False.\r\n","comment_length":47,"text":"boolq does not load \nHi\r\nI am getting these errors trying to load boolq thanks \r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 5, in <module>\r\n data = AutoTask().get(\"boolq\").get_dataset(\"train\", n_obs=10)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 42, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 38, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 150, in download_custom\r\n get_from_cache(url, cache_dir=cache_dir, local_files_only=True, use_etag=False)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 472, in get_from_cache\r\n f\"Cannot find the requested files in the cached path at {cache_path} and outgoing traffic has been\"\r\nFileNotFoundError: Cannot find the requested files in the cached path at \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/eaee069e38f6ceaa84de02ad088c34e63ec97671f2cd1910ddb16b10dc60808c and outgoing traffic has been disabled. To enable file online look-ups, set 'local_files_only' to False.\r\n\nHi ! It runs on my side without issues. I tried\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"boolq\")\r\n```\r\n\r\nWhat version of datasets and tensorflow are your runnning ?\r\nAlso if you manage to get a minimal reproducible script (on google colab for example) that would be useful.","embeddings":[-0.3142423332,-0.1296663433,-0.1398147941,0.2243659049,0.0898041949,0.0106670726,0.4253870547,0.3031967878,0.4457686841,-0.1245959029,-0.086008586,0.3309848905,-0.1855678558,0.4824474752,0.0417104885,-0.0254136994,-0.1236411631,0.0837756172,-0.0132116294,-0.0677919537,-0.4335164726,0.3662593663,-0.4190507531,0.1728812605,-0.1200258434,-0.2764460444,-0.0112306615,0.2354567647,-0.3227624893,-0.3402307034,0.129193157,0.0399898887,0.2798753381,0.1975854039,-0.0001004217,0.04742755,0.4136013389,0.0912821293,-0.2201259285,-0.290461421,-0.1725928336,-0.3197637498,0.1770071685,-0.1704840362,-0.0781291723,-0.054515522,-0.0554295033,0.0769731924,0.4125249982,0.4600992799,0.3143282831,0.0757981241,0.1170568392,-0.1951514184,0.1931924373,-0.1626982242,0.1515527815,0.2015698403,0.0419378281,-0.2938958704,0.1587524712,0.2637588382,-0.183339566,-0.246029079,0.0121768303,-0.2074029446,-0.1017384678,-0.3344747722,0.2351708412,0.0141455783,0.5186159015,-0.1224103197,-0.2733826637,-0.056684956,0.1429380774,-0.1925470382,0.1893985271,0.0914991349,0.0534336679,-0.0635306165,0.1850988269,-0.4896081984,-0.0376428179,-0.0354207493,-0.1383907795,0.2759750187,-0.1215392649,0.051884301,0.1912807226,-0.1711804718,0.1133473963,0.2080259323,-0.152289167,0.1584958732,-0.5179923177,0.0358057208,0.1561286449,-0.0271783192,-0.0363988988,0.129627496,-0.189453423,0.0187248792,-0.091720216,0.2068092227,0.0046222373,0.5135238171,0.3842213452,0.2736550868,0.1583575159,0.1075289547,-0.2233469784,0.0342296846,-0.5201335549,-0.3621052206,-0.1302121133,0.2492164373,0.2366273552,-0.0848474279,-0.3538365662,-0.1739944071,-0.3224677145,-0.1347561181,0.3432625234,0.6987859011,0.0430136062,-0.1265625507,0.0525118038,0.2696886957,-0.3483470976,-0.0706117153,-0.2877412438,0.3017656803,-0.0307151303,-0.1554742754,0.2976013422,-0.1689217985,0.352542758,0.0606402606,-0.249807626,-0.0990925953,0.2741260231,-0.2546375394,-0.373259753,0.2511183619,0.0047492059,0.0151632084,0.2008347362,0.2303849608,-0.0411107466,0.2055432796,-0.2143663764,-0.2891750336,0.0416452251,0.3372689188,-0.0229818821,-0.2775761187,-0.1418446451,0.1758444458,0.1150734574,-0.1472827941,-0.2639817595,-0.327144146,-0.1046774015,-0.2185691744,0.3993265629,0.4546142817,-0.3941679597,-0.264450103,-0.3861935735,-0.2117394805,0.005800161,0.2955780327,-0.1370882541,0.5168119073,-0.2074039429,0.0861331522,0.2646962404,-0.2111936361,-0.1844504923,0.1970254332,-0.0149663696,-0.0348743387,-0.2750920653,-0.044528529,0.1203418449,0.1673647314,0.1317070127,0.1415139586,0.2968439162,-0.2430085838,-0.1465197951,-0.1223876178,0.0710638985,0.3457807302,0.2301300913,-0.0535534918,-0.0584518202,-0.4113751054,0.0685673952,0.1532153785,-0.2991673946,0.1951564103,0.3404016793,-0.1151792407,0.1968043298,0.0254542511,-0.3794722259,0.1544528753,-0.2339846492,-0.0378329121,-0.062096227,0.1533769965,-0.326677829,0.0734995157,-0.1862191111,-0.0524057746,0.1931232661,-0.1655142456,-0.0569789149,0.0857940465,-0.308848381,0.2691085041,-0.0235381071,0.1346359849,-0.2137397379,0.1054314226,-0.1465704441,-0.0927517265,0.2968972921,0.2993911803,0.1980737299,0.0299892873,-0.0212003067,0.2432890236,0.1288769841,0.1043745577,-0.0412735492,-0.5058262348,0.1112553775,-0.0384226665,0.307051599,0.1330125481,0.0742572993,0.0808739066,0.0900950283,0.2717500031,-0.0859902874,0.1937272102,0.1444682777,0.1817198992,0.2400854379,0.0292750653,-0.006893957,-0.1292017549,0.4810894728,-0.0400450267,-0.0255702455,0.0285188742,-0.1477422416,0.0691427216,0.4380694628,0.1077947244,0.122457467,0.0266056359,-0.1225930899,0.1235340685,0.1176418364,0.1531540751,0.2710778117,0.2076154053,-0.4852933288,-0.1021019965,-0.0421196967,-0.2693491876,0.1733107269,-0.1938354522,0.0288225207,0.2061602622,-0.0693069771,-0.1248023659,-0.1851713657,-0.2250053734,0.09095148,0.2774633765,-0.1973669082,-0.0649487823,-0.0835349858,-0.3896842897,0.191729635,-0.1083577797,-0.0304460432,-0.35120821,0.1287216693,-0.029635014,-0.0343716405,0.2476255149,0.0612640604,0.3148314655,0.107014671,-0.0850824639,-0.1364320368,-0.2685342133,-0.2266000211,0.1310845762,0.1870205998,0.0818480328,0.2532462776,-0.0990140066,0.0883723646,-0.0534891039,-0.2225491405,-0.08780545,-0.0944324955,0.3347641528,0.1184832379,0.0666116774,-0.0564614795,-0.2238010466,0.3523494303,-0.1633935571,-0.0211878568,0.1725939363,0.1716831923,-0.0578616336,-0.220501855,-0.5058780313,-0.5898679495,-0.4911761284,0.0963694528,0.0636325926,0.1966700107,0.3625305295,-0.0294897985,0.1960614175,0.0712995976,-0.0495073088,0.0055860337,-0.1599078923,0.262078315,-0.0651985332,-0.2783256769,0.0434393212,-0.0880052298,0.4329284132,-0.2539993227,-0.3081473708,0.003639091,-0.0644266978,-0.0977022052,-0.0620605685,-0.1536531448,-0.1169783846,0.0611614808,0.0138719501,0.0663510412,-0.3257706761,0.227166757,0.0191833954,0.1968749166,-0.0655772835,0.3469649255,-0.3650504053,0.6071892977,0.1826376915,0.1984719187,0.2066986561,0.2602954507,0.0882441327,0.022590179,-0.7111612558,0.0119960429,0.1972316355,-0.1737873405,0.0223199911,-0.0742926896,0.0048973546,-0.2321588546,-0.1811950803,-0.1740103066,-0.2393470556,-0.055500295,-0.3643978238,0.022894742,-0.0218463298,0.3673841655,-0.2108635604,0.0025235156,0.0858273208,0.1972029209,0.1464118659,-0.0942063183,-0.5303607583,-0.5024215579,-0.3433577418,0.2414335608,-0.268050611,0.2392213047,-0.1371821761,0.0209193882,0.1031649038,-0.1103468761,0.4958151877,-0.1310851723,-0.1198505163,0.3111153841,-0.0852036104,-0.2169436067,-0.1560447812,-0.2145740539,0.2487731874,0.0919606164,0.4482049048,-0.250259459,0.0183465444,0.0149936453,-0.1805451065,-0.2933079898,-0.4980681539,-0.4371125698,-0.1474029571,-0.3785982728,-0.0684283376,-0.2436530143,0.2505195737,-0.2380888164,-0.0173401665,-0.0446072482,-0.2353481203,0.1815394759,0.1636779308,0.5122842193,-0.1921356916,0.0965679884,0.0463171192,0.0618032813,0.1100911647,0.3877335489,-0.1745917797,-0.2596477866,-0.0316261984,-0.0194504261,0.2970832288,-0.1034843102,-0.2948684096,0.3071787655,-0.0516149625,0.1102492511,0.025938293,0.0147437481,0.2532850504,0.166193068,-0.0491432957,-0.4161556065,0.1546282619,-0.026426319,0.0593896918,0.2063326389,-0.2129352987,-0.0297328606,0.2763859034,-0.0648007765,0.5007886887,-0.1812511981,-0.1933797598,0.3407919109,0.2174998671,0.4038965702,-0.1669141054,0.2915231884,-0.5527960062,-0.0196560118,0.2265730053,-0.1374719739,-0.0402738154,-0.2586255074,-0.5010080338,0.2584686875,-0.1938874722,0.3548452258,0.0251688473,0.2396081388,-0.1991093904,0.0686027333,-0.2652857304,0.1974008232,-0.1181023642,0.3159292936,-0.1748003811,-0.0673146844,0.338904947,-0.1213414446,-0.0972986966,0.3927971721,-0.1331145167,0.249198094,0.1869202554,-0.0859444141,0.1371118277,0.4802416861,0.0360942744,0.2273453474,-0.1351352185,0.065763317,-0.019321125,-0.094333306,-0.112178646,0.1603537798,0.2917366624,-0.1176791191,-0.3513068259,0.4982581437,-0.053962864,0.0808963627,-0.2077497095,0.0966773257,-0.0197356436,-0.2792844772,-0.0582650863,-0.0777066201,0.0835340992,-0.1127732247,0.1955122054,0.1503963768,-0.045955576,0.2849721611,-0.1004797891,-0.2767522335,-0.0368325561,0.1631646305,-0.0202607755,0.2678747773,0.7444613576,0.3148196042,-0.0434075892,-0.4068887234,-0.1119854152,0.1022955924,-0.0354892612,0.2343204767,0.2025268972,-0.0975747928,0.2175710946,0.0701927096,0.0332487226,-0.0487874486,-0.0780247897,-0.4976789057,-0.2899364829,0.5084173083,-0.2384429127,-0.0717616007,-0.0226836149,-0.0233008359,0.0253997985,0.1786234975,-0.4123268425,-0.0047474536,-0.5003218055,-0.1386016607,-0.1944634318,-0.2909421921,0.1350518465,0.1839277148,0.2573776543,-0.0973946154,-0.1259430051,-0.2820370495,0.0279131066,0.018612029,-0.0506161414,0.1341328174,-0.0356099978,-0.2192932963,0.026943488,0.124546729,-0.019082034,-0.1418261826,0.0428549983,0.2309853882,0.0664203838,0.1498230696,-0.1048577055,0.3925117552,-0.1542077214,-0.1091762707,0.0892825425,0.0404633768,0.1454324126,0.0041160379,-0.0036404058,-0.0570225157,0.0763983354,0.0788752958,0.1567761004,-0.2010095417,0.2653385103,-0.1736503392,0.0509918667,0.6181201339,-0.2227316052,0.1117159277,0.0968501717,0.3101870716,-0.227864027,0.1903332025,-0.226213932,-0.0269351378,0.0044274949,0.1368973404,-0.2643563449,-0.1858036667,0.1944193244,0.1868375391,0.2378834039,0.1207540259,0.0021422226,0.2828045785,-0.2137751877,-0.1657215059,0.1804838777,0.1804220527,0.0042540804,0.3604365289,-0.2713555992,-0.0449317358,0.0594246946,0.2507373989,0.1973168105,-0.3961364329,-0.1421324164,0.3903949559,-0.0077800611,-0.2976878583,-0.2608277202,0.4449037015,-0.1044107005,-0.0567574911,-0.1178352311,0.1353802085,-0.3331618309,-0.0949439034,-0.5158820748,-0.0901945084,-0.0454395972,-0.1312888116,0.0252012815,0.0420153625,-0.1148137525,0.0730800107,-0.1637796462,-0.3743667305,0.2009001225,0.2530694306,-0.2437724918,-0.1150775999,0.3914286196,0.1533192545,0.3194291294,0.1970059723,0.0829290003,0.3079617321,0.5380836129,-0.3283626139,-0.0000046035,0.0199086033,-0.1313573867,0.0346756577,0.3137899935,0.2148800194,0.3973927796,0.2980228961,0.2429427803,-0.1719399244,0.1943987459,0.0471306443,0.1650931686,0.0040556081,0.6533156633,-0.0515634567,0.0107027795,-0.2028621137,-0.0492494442,-0.5103918314,-0.0879737213,0.2999315262,0.2265715152,0.0560057573,-0.2306601256,0.1365998983,-0.3244097531,0.4204716682,0.4334087074,0.464944005,-0.4140831232,0.1457706243,-0.4007861614,0.0466136932,0.0658597574,-0.0020465965,-0.1097378805,0.0314681605,-0.0425010435,-0.0867101401,0.131939888,0.0399009325,0.2227889597,-0.1774658263,-0.4814176559,0.1328634173,0.0461931862,0.1684636027,0.0244556442,-0.1332152337,0.1925159246,-0.4258807003,0.099328503,-0.0376955234,-0.2557370663,0.1841211617,-0.2434110194,0.2420452237,0.0028272639,0.2686461508,0.0102750724,-0.0241375919,-0.1938576102,-0.3289037347,-0.1430860758,0.0125012193,0.3162784874,0.2012417167,-0.0406675674,0.2181183398,-0.1372948736,0.1671044677,-0.0101538477,-0.1750993431,-0.1615086496,-0.0480243564,0.2521775365,-0.1583052278,0.0523845926,0.145988524,0.0466804728,0.1564805657,-0.4045481682,-0.0697857961,0.490942657,-0.018943686,-0.0651996136,-0.0544002727,0.2422242314,0.1531124562,0.0249590147,-0.3776175678,0.0245507509,0.3022921383,-0.0252069011,0.077616632,0.3028816879,0.4011399448,0.0964677855,-0.0990418643,-0.283208698,0.1615866125,0.09360677,0.0778960735,0.0205555446]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/879","title":"boolq does not load ","comments":"hey\ni do the exact same commands. for me it fails i guess might be issues with\ncaching maybe?\nthanks\nbest\nrabeeh\n\nOn Tue, Nov 24, 2020, 10:24 AM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> Hi ! It runs on my side without issues. I tried\n>\n> from datasets import load_datasetload_dataset(\"boolq\")\n>\n> What version of datasets and tensorflow are your runnning ?\n> Also if you manage to get a minimal reproducible script (on google colab\n> for example) that would be useful.\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/879#issuecomment-732769114>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ABP4ZCGGDR2FUMRKZTIY5CTSRN3VXANCNFSM4T7R3U6A>\n> .\n>\n","body":"Hi\r\nI am getting these errors trying to load boolq thanks \r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 5, in <module>\r\n data = AutoTask().get(\"boolq\").get_dataset(\"train\", n_obs=10)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 42, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 38, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 150, in download_custom\r\n get_from_cache(url, cache_dir=cache_dir, local_files_only=True, use_etag=False)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 472, in get_from_cache\r\n f\"Cannot find the requested files in the cached path at {cache_path} and outgoing traffic has been\"\r\nFileNotFoundError: Cannot find the requested files in the cached path at \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/eaee069e38f6ceaa84de02ad088c34e63ec97671f2cd1910ddb16b10dc60808c and outgoing traffic has been disabled. To enable file online look-ups, set 'local_files_only' to False.\r\n","comment_length":117,"text":"boolq does not load \nHi\r\nI am getting these errors trying to load boolq thanks \r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 5, in <module>\r\n data = AutoTask().get(\"boolq\").get_dataset(\"train\", n_obs=10)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 42, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 38, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 150, in download_custom\r\n get_from_cache(url, cache_dir=cache_dir, local_files_only=True, use_etag=False)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 472, in get_from_cache\r\n f\"Cannot find the requested files in the cached path at {cache_path} and outgoing traffic has been\"\r\nFileNotFoundError: Cannot find the requested files in the cached path at \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/eaee069e38f6ceaa84de02ad088c34e63ec97671f2cd1910ddb16b10dc60808c and outgoing traffic has been disabled. To enable file online look-ups, set 'local_files_only' to False.\r\n\nhey\ni do the exact same commands. for me it fails i guess might be issues with\ncaching maybe?\nthanks\nbest\nrabeeh\n\nOn Tue, Nov 24, 2020, 10:24 AM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> Hi ! It runs on my side without issues. I tried\n>\n> from datasets import load_datasetload_dataset(\"boolq\")\n>\n> What version of datasets and tensorflow are your runnning ?\n> Also if you manage to get a minimal reproducible script (on google colab\n> for example) that would be useful.\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/879#issuecomment-732769114>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ABP4ZCGGDR2FUMRKZTIY5CTSRN3VXANCNFSM4T7R3U6A>\n> .\n>\n","embeddings":[-0.3142423332,-0.1296663433,-0.1398147941,0.2243659049,0.0898041949,0.0106670726,0.4253870547,0.3031967878,0.4457686841,-0.1245959029,-0.086008586,0.3309848905,-0.1855678558,0.4824474752,0.0417104885,-0.0254136994,-0.1236411631,0.0837756172,-0.0132116294,-0.0677919537,-0.4335164726,0.3662593663,-0.4190507531,0.1728812605,-0.1200258434,-0.2764460444,-0.0112306615,0.2354567647,-0.3227624893,-0.3402307034,0.129193157,0.0399898887,0.2798753381,0.1975854039,-0.0001004217,0.04742755,0.4136013389,0.0912821293,-0.2201259285,-0.290461421,-0.1725928336,-0.3197637498,0.1770071685,-0.1704840362,-0.0781291723,-0.054515522,-0.0554295033,0.0769731924,0.4125249982,0.4600992799,0.3143282831,0.0757981241,0.1170568392,-0.1951514184,0.1931924373,-0.1626982242,0.1515527815,0.2015698403,0.0419378281,-0.2938958704,0.1587524712,0.2637588382,-0.183339566,-0.246029079,0.0121768303,-0.2074029446,-0.1017384678,-0.3344747722,0.2351708412,0.0141455783,0.5186159015,-0.1224103197,-0.2733826637,-0.056684956,0.1429380774,-0.1925470382,0.1893985271,0.0914991349,0.0534336679,-0.0635306165,0.1850988269,-0.4896081984,-0.0376428179,-0.0354207493,-0.1383907795,0.2759750187,-0.1215392649,0.051884301,0.1912807226,-0.1711804718,0.1133473963,0.2080259323,-0.152289167,0.1584958732,-0.5179923177,0.0358057208,0.1561286449,-0.0271783192,-0.0363988988,0.129627496,-0.189453423,0.0187248792,-0.091720216,0.2068092227,0.0046222373,0.5135238171,0.3842213452,0.2736550868,0.1583575159,0.1075289547,-0.2233469784,0.0342296846,-0.5201335549,-0.3621052206,-0.1302121133,0.2492164373,0.2366273552,-0.0848474279,-0.3538365662,-0.1739944071,-0.3224677145,-0.1347561181,0.3432625234,0.6987859011,0.0430136062,-0.1265625507,0.0525118038,0.2696886957,-0.3483470976,-0.0706117153,-0.2877412438,0.3017656803,-0.0307151303,-0.1554742754,0.2976013422,-0.1689217985,0.352542758,0.0606402606,-0.249807626,-0.0990925953,0.2741260231,-0.2546375394,-0.373259753,0.2511183619,0.0047492059,0.0151632084,0.2008347362,0.2303849608,-0.0411107466,0.2055432796,-0.2143663764,-0.2891750336,0.0416452251,0.3372689188,-0.0229818821,-0.2775761187,-0.1418446451,0.1758444458,0.1150734574,-0.1472827941,-0.2639817595,-0.327144146,-0.1046774015,-0.2185691744,0.3993265629,0.4546142817,-0.3941679597,-0.264450103,-0.3861935735,-0.2117394805,0.005800161,0.2955780327,-0.1370882541,0.5168119073,-0.2074039429,0.0861331522,0.2646962404,-0.2111936361,-0.1844504923,0.1970254332,-0.0149663696,-0.0348743387,-0.2750920653,-0.044528529,0.1203418449,0.1673647314,0.1317070127,0.1415139586,0.2968439162,-0.2430085838,-0.1465197951,-0.1223876178,0.0710638985,0.3457807302,0.2301300913,-0.0535534918,-0.0584518202,-0.4113751054,0.0685673952,0.1532153785,-0.2991673946,0.1951564103,0.3404016793,-0.1151792407,0.1968043298,0.0254542511,-0.3794722259,0.1544528753,-0.2339846492,-0.0378329121,-0.062096227,0.1533769965,-0.326677829,0.0734995157,-0.1862191111,-0.0524057746,0.1931232661,-0.1655142456,-0.0569789149,0.0857940465,-0.308848381,0.2691085041,-0.0235381071,0.1346359849,-0.2137397379,0.1054314226,-0.1465704441,-0.0927517265,0.2968972921,0.2993911803,0.1980737299,0.0299892873,-0.0212003067,0.2432890236,0.1288769841,0.1043745577,-0.0412735492,-0.5058262348,0.1112553775,-0.0384226665,0.307051599,0.1330125481,0.0742572993,0.0808739066,0.0900950283,0.2717500031,-0.0859902874,0.1937272102,0.1444682777,0.1817198992,0.2400854379,0.0292750653,-0.006893957,-0.1292017549,0.4810894728,-0.0400450267,-0.0255702455,0.0285188742,-0.1477422416,0.0691427216,0.4380694628,0.1077947244,0.122457467,0.0266056359,-0.1225930899,0.1235340685,0.1176418364,0.1531540751,0.2710778117,0.2076154053,-0.4852933288,-0.1021019965,-0.0421196967,-0.2693491876,0.1733107269,-0.1938354522,0.0288225207,0.2061602622,-0.0693069771,-0.1248023659,-0.1851713657,-0.2250053734,0.09095148,0.2774633765,-0.1973669082,-0.0649487823,-0.0835349858,-0.3896842897,0.191729635,-0.1083577797,-0.0304460432,-0.35120821,0.1287216693,-0.029635014,-0.0343716405,0.2476255149,0.0612640604,0.3148314655,0.107014671,-0.0850824639,-0.1364320368,-0.2685342133,-0.2266000211,0.1310845762,0.1870205998,0.0818480328,0.2532462776,-0.0990140066,0.0883723646,-0.0534891039,-0.2225491405,-0.08780545,-0.0944324955,0.3347641528,0.1184832379,0.0666116774,-0.0564614795,-0.2238010466,0.3523494303,-0.1633935571,-0.0211878568,0.1725939363,0.1716831923,-0.0578616336,-0.220501855,-0.5058780313,-0.5898679495,-0.4911761284,0.0963694528,0.0636325926,0.1966700107,0.3625305295,-0.0294897985,0.1960614175,0.0712995976,-0.0495073088,0.0055860337,-0.1599078923,0.262078315,-0.0651985332,-0.2783256769,0.0434393212,-0.0880052298,0.4329284132,-0.2539993227,-0.3081473708,0.003639091,-0.0644266978,-0.0977022052,-0.0620605685,-0.1536531448,-0.1169783846,0.0611614808,0.0138719501,0.0663510412,-0.3257706761,0.227166757,0.0191833954,0.1968749166,-0.0655772835,0.3469649255,-0.3650504053,0.6071892977,0.1826376915,0.1984719187,0.2066986561,0.2602954507,0.0882441327,0.022590179,-0.7111612558,0.0119960429,0.1972316355,-0.1737873405,0.0223199911,-0.0742926896,0.0048973546,-0.2321588546,-0.1811950803,-0.1740103066,-0.2393470556,-0.055500295,-0.3643978238,0.022894742,-0.0218463298,0.3673841655,-0.2108635604,0.0025235156,0.0858273208,0.1972029209,0.1464118659,-0.0942063183,-0.5303607583,-0.5024215579,-0.3433577418,0.2414335608,-0.268050611,0.2392213047,-0.1371821761,0.0209193882,0.1031649038,-0.1103468761,0.4958151877,-0.1310851723,-0.1198505163,0.3111153841,-0.0852036104,-0.2169436067,-0.1560447812,-0.2145740539,0.2487731874,0.0919606164,0.4482049048,-0.250259459,0.0183465444,0.0149936453,-0.1805451065,-0.2933079898,-0.4980681539,-0.4371125698,-0.1474029571,-0.3785982728,-0.0684283376,-0.2436530143,0.2505195737,-0.2380888164,-0.0173401665,-0.0446072482,-0.2353481203,0.1815394759,0.1636779308,0.5122842193,-0.1921356916,0.0965679884,0.0463171192,0.0618032813,0.1100911647,0.3877335489,-0.1745917797,-0.2596477866,-0.0316261984,-0.0194504261,0.2970832288,-0.1034843102,-0.2948684096,0.3071787655,-0.0516149625,0.1102492511,0.025938293,0.0147437481,0.2532850504,0.166193068,-0.0491432957,-0.4161556065,0.1546282619,-0.026426319,0.0593896918,0.2063326389,-0.2129352987,-0.0297328606,0.2763859034,-0.0648007765,0.5007886887,-0.1812511981,-0.1933797598,0.3407919109,0.2174998671,0.4038965702,-0.1669141054,0.2915231884,-0.5527960062,-0.0196560118,0.2265730053,-0.1374719739,-0.0402738154,-0.2586255074,-0.5010080338,0.2584686875,-0.1938874722,0.3548452258,0.0251688473,0.2396081388,-0.1991093904,0.0686027333,-0.2652857304,0.1974008232,-0.1181023642,0.3159292936,-0.1748003811,-0.0673146844,0.338904947,-0.1213414446,-0.0972986966,0.3927971721,-0.1331145167,0.249198094,0.1869202554,-0.0859444141,0.1371118277,0.4802416861,0.0360942744,0.2273453474,-0.1351352185,0.065763317,-0.019321125,-0.094333306,-0.112178646,0.1603537798,0.2917366624,-0.1176791191,-0.3513068259,0.4982581437,-0.053962864,0.0808963627,-0.2077497095,0.0966773257,-0.0197356436,-0.2792844772,-0.0582650863,-0.0777066201,0.0835340992,-0.1127732247,0.1955122054,0.1503963768,-0.045955576,0.2849721611,-0.1004797891,-0.2767522335,-0.0368325561,0.1631646305,-0.0202607755,0.2678747773,0.7444613576,0.3148196042,-0.0434075892,-0.4068887234,-0.1119854152,0.1022955924,-0.0354892612,0.2343204767,0.2025268972,-0.0975747928,0.2175710946,0.0701927096,0.0332487226,-0.0487874486,-0.0780247897,-0.4976789057,-0.2899364829,0.5084173083,-0.2384429127,-0.0717616007,-0.0226836149,-0.0233008359,0.0253997985,0.1786234975,-0.4123268425,-0.0047474536,-0.5003218055,-0.1386016607,-0.1944634318,-0.2909421921,0.1350518465,0.1839277148,0.2573776543,-0.0973946154,-0.1259430051,-0.2820370495,0.0279131066,0.018612029,-0.0506161414,0.1341328174,-0.0356099978,-0.2192932963,0.026943488,0.124546729,-0.019082034,-0.1418261826,0.0428549983,0.2309853882,0.0664203838,0.1498230696,-0.1048577055,0.3925117552,-0.1542077214,-0.1091762707,0.0892825425,0.0404633768,0.1454324126,0.0041160379,-0.0036404058,-0.0570225157,0.0763983354,0.0788752958,0.1567761004,-0.2010095417,0.2653385103,-0.1736503392,0.0509918667,0.6181201339,-0.2227316052,0.1117159277,0.0968501717,0.3101870716,-0.227864027,0.1903332025,-0.226213932,-0.0269351378,0.0044274949,0.1368973404,-0.2643563449,-0.1858036667,0.1944193244,0.1868375391,0.2378834039,0.1207540259,0.0021422226,0.2828045785,-0.2137751877,-0.1657215059,0.1804838777,0.1804220527,0.0042540804,0.3604365289,-0.2713555992,-0.0449317358,0.0594246946,0.2507373989,0.1973168105,-0.3961364329,-0.1421324164,0.3903949559,-0.0077800611,-0.2976878583,-0.2608277202,0.4449037015,-0.1044107005,-0.0567574911,-0.1178352311,0.1353802085,-0.3331618309,-0.0949439034,-0.5158820748,-0.0901945084,-0.0454395972,-0.1312888116,0.0252012815,0.0420153625,-0.1148137525,0.0730800107,-0.1637796462,-0.3743667305,0.2009001225,0.2530694306,-0.2437724918,-0.1150775999,0.3914286196,0.1533192545,0.3194291294,0.1970059723,0.0829290003,0.3079617321,0.5380836129,-0.3283626139,-0.0000046035,0.0199086033,-0.1313573867,0.0346756577,0.3137899935,0.2148800194,0.3973927796,0.2980228961,0.2429427803,-0.1719399244,0.1943987459,0.0471306443,0.1650931686,0.0040556081,0.6533156633,-0.0515634567,0.0107027795,-0.2028621137,-0.0492494442,-0.5103918314,-0.0879737213,0.2999315262,0.2265715152,0.0560057573,-0.2306601256,0.1365998983,-0.3244097531,0.4204716682,0.4334087074,0.464944005,-0.4140831232,0.1457706243,-0.4007861614,0.0466136932,0.0658597574,-0.0020465965,-0.1097378805,0.0314681605,-0.0425010435,-0.0867101401,0.131939888,0.0399009325,0.2227889597,-0.1774658263,-0.4814176559,0.1328634173,0.0461931862,0.1684636027,0.0244556442,-0.1332152337,0.1925159246,-0.4258807003,0.099328503,-0.0376955234,-0.2557370663,0.1841211617,-0.2434110194,0.2420452237,0.0028272639,0.2686461508,0.0102750724,-0.0241375919,-0.1938576102,-0.3289037347,-0.1430860758,0.0125012193,0.3162784874,0.2012417167,-0.0406675674,0.2181183398,-0.1372948736,0.1671044677,-0.0101538477,-0.1750993431,-0.1615086496,-0.0480243564,0.2521775365,-0.1583052278,0.0523845926,0.145988524,0.0466804728,0.1564805657,-0.4045481682,-0.0697857961,0.490942657,-0.018943686,-0.0651996136,-0.0544002727,0.2422242314,0.1531124562,0.0249590147,-0.3776175678,0.0245507509,0.3022921383,-0.0252069011,0.077616632,0.3028816879,0.4011399448,0.0964677855,-0.0990418643,-0.283208698,0.1615866125,0.09360677,0.0778960735,0.0205555446]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/879","title":"boolq does not load ","comments":"Could you check if it works on the master branch ?\r\nYou can use `load_dataset(\"boolq\", script_version=\"master\")` to do so.\r\nWe did some changes recently in boolq to remove the TF dependency and we changed the way the data files are downloaded in https:\/\/github.com\/huggingface\/datasets\/pull\/881","body":"Hi\r\nI am getting these errors trying to load boolq thanks \r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 5, in <module>\r\n data = AutoTask().get(\"boolq\").get_dataset(\"train\", n_obs=10)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 42, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 38, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 150, in download_custom\r\n get_from_cache(url, cache_dir=cache_dir, local_files_only=True, use_etag=False)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 472, in get_from_cache\r\n f\"Cannot find the requested files in the cached path at {cache_path} and outgoing traffic has been\"\r\nFileNotFoundError: Cannot find the requested files in the cached path at \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/eaee069e38f6ceaa84de02ad088c34e63ec97671f2cd1910ddb16b10dc60808c and outgoing traffic has been disabled. To enable file online look-ups, set 'local_files_only' to False.\r\n","comment_length":43,"text":"boolq does not load \nHi\r\nI am getting these errors trying to load boolq thanks \r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 5, in <module>\r\n data = AutoTask().get(\"boolq\").get_dataset(\"train\", n_obs=10)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 42, in get_dataset\r\n dataset = self.load_dataset(split=split)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks\/tasks.py\", line 38, in load_dataset\r\n return datasets.load_dataset(self.task.name, split=split)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 150, in download_custom\r\n get_from_cache(url, cache_dir=cache_dir, local_files_only=True, use_etag=False)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 472, in get_from_cache\r\n f\"Cannot find the requested files in the cached path at {cache_path} and outgoing traffic has been\"\r\nFileNotFoundError: Cannot find the requested files in the cached path at \/idiap\/home\/rkarimi\/.cache\/huggingface\/datasets\/eaee069e38f6ceaa84de02ad088c34e63ec97671f2cd1910ddb16b10dc60808c and outgoing traffic has been disabled. To enable file online look-ups, set 'local_files_only' to False.\r\n\nCould you check if it works on the master branch ?\r\nYou can use `load_dataset(\"boolq\", script_version=\"master\")` to do so.\r\nWe did some changes recently in boolq to remove the TF dependency and we changed the way the data files are downloaded in https:\/\/github.com\/huggingface\/datasets\/pull\/881","embeddings":[-0.3142423332,-0.1296663433,-0.1398147941,0.2243659049,0.0898041949,0.0106670726,0.4253870547,0.3031967878,0.4457686841,-0.1245959029,-0.086008586,0.3309848905,-0.1855678558,0.4824474752,0.0417104885,-0.0254136994,-0.1236411631,0.0837756172,-0.0132116294,-0.0677919537,-0.4335164726,0.3662593663,-0.4190507531,0.1728812605,-0.1200258434,-0.2764460444,-0.0112306615,0.2354567647,-0.3227624893,-0.3402307034,0.129193157,0.0399898887,0.2798753381,0.1975854039,-0.0001004217,0.04742755,0.4136013389,0.0912821293,-0.2201259285,-0.290461421,-0.1725928336,-0.3197637498,0.1770071685,-0.1704840362,-0.0781291723,-0.054515522,-0.0554295033,0.0769731924,0.4125249982,0.4600992799,0.3143282831,0.0757981241,0.1170568392,-0.1951514184,0.1931924373,-0.1626982242,0.1515527815,0.2015698403,0.0419378281,-0.2938958704,0.1587524712,0.2637588382,-0.183339566,-0.246029079,0.0121768303,-0.2074029446,-0.1017384678,-0.3344747722,0.2351708412,0.0141455783,0.5186159015,-0.1224103197,-0.2733826637,-0.056684956,0.1429380774,-0.1925470382,0.1893985271,0.0914991349,0.0534336679,-0.0635306165,0.1850988269,-0.4896081984,-0.0376428179,-0.0354207493,-0.1383907795,0.2759750187,-0.1215392649,0.051884301,0.1912807226,-0.1711804718,0.1133473963,0.2080259323,-0.152289167,0.1584958732,-0.5179923177,0.0358057208,0.1561286449,-0.0271783192,-0.0363988988,0.129627496,-0.189453423,0.0187248792,-0.091720216,0.2068092227,0.0046222373,0.5135238171,0.3842213452,0.2736550868,0.1583575159,0.1075289547,-0.2233469784,0.0342296846,-0.5201335549,-0.3621052206,-0.1302121133,0.2492164373,0.2366273552,-0.0848474279,-0.3538365662,-0.1739944071,-0.3224677145,-0.1347561181,0.3432625234,0.6987859011,0.0430136062,-0.1265625507,0.0525118038,0.2696886957,-0.3483470976,-0.0706117153,-0.2877412438,0.3017656803,-0.0307151303,-0.1554742754,0.2976013422,-0.1689217985,0.352542758,0.0606402606,-0.249807626,-0.0990925953,0.2741260231,-0.2546375394,-0.373259753,0.2511183619,0.0047492059,0.0151632084,0.2008347362,0.2303849608,-0.0411107466,0.2055432796,-0.2143663764,-0.2891750336,0.0416452251,0.3372689188,-0.0229818821,-0.2775761187,-0.1418446451,0.1758444458,0.1150734574,-0.1472827941,-0.2639817595,-0.327144146,-0.1046774015,-0.2185691744,0.3993265629,0.4546142817,-0.3941679597,-0.264450103,-0.3861935735,-0.2117394805,0.005800161,0.2955780327,-0.1370882541,0.5168119073,-0.2074039429,0.0861331522,0.2646962404,-0.2111936361,-0.1844504923,0.1970254332,-0.0149663696,-0.0348743387,-0.2750920653,-0.044528529,0.1203418449,0.1673647314,0.1317070127,0.1415139586,0.2968439162,-0.2430085838,-0.1465197951,-0.1223876178,0.0710638985,0.3457807302,0.2301300913,-0.0535534918,-0.0584518202,-0.4113751054,0.0685673952,0.1532153785,-0.2991673946,0.1951564103,0.3404016793,-0.1151792407,0.1968043298,0.0254542511,-0.3794722259,0.1544528753,-0.2339846492,-0.0378329121,-0.062096227,0.1533769965,-0.326677829,0.0734995157,-0.1862191111,-0.0524057746,0.1931232661,-0.1655142456,-0.0569789149,0.0857940465,-0.308848381,0.2691085041,-0.0235381071,0.1346359849,-0.2137397379,0.1054314226,-0.1465704441,-0.0927517265,0.2968972921,0.2993911803,0.1980737299,0.0299892873,-0.0212003067,0.2432890236,0.1288769841,0.1043745577,-0.0412735492,-0.5058262348,0.1112553775,-0.0384226665,0.307051599,0.1330125481,0.0742572993,0.0808739066,0.0900950283,0.2717500031,-0.0859902874,0.1937272102,0.1444682777,0.1817198992,0.2400854379,0.0292750653,-0.006893957,-0.1292017549,0.4810894728,-0.0400450267,-0.0255702455,0.0285188742,-0.1477422416,0.0691427216,0.4380694628,0.1077947244,0.122457467,0.0266056359,-0.1225930899,0.1235340685,0.1176418364,0.1531540751,0.2710778117,0.2076154053,-0.4852933288,-0.1021019965,-0.0421196967,-0.2693491876,0.1733107269,-0.1938354522,0.0288225207,0.2061602622,-0.0693069771,-0.1248023659,-0.1851713657,-0.2250053734,0.09095148,0.2774633765,-0.1973669082,-0.0649487823,-0.0835349858,-0.3896842897,0.191729635,-0.1083577797,-0.0304460432,-0.35120821,0.1287216693,-0.029635014,-0.0343716405,0.2476255149,0.0612640604,0.3148314655,0.107014671,-0.0850824639,-0.1364320368,-0.2685342133,-0.2266000211,0.1310845762,0.1870205998,0.0818480328,0.2532462776,-0.0990140066,0.0883723646,-0.0534891039,-0.2225491405,-0.08780545,-0.0944324955,0.3347641528,0.1184832379,0.0666116774,-0.0564614795,-0.2238010466,0.3523494303,-0.1633935571,-0.0211878568,0.1725939363,0.1716831923,-0.0578616336,-0.220501855,-0.5058780313,-0.5898679495,-0.4911761284,0.0963694528,0.0636325926,0.1966700107,0.3625305295,-0.0294897985,0.1960614175,0.0712995976,-0.0495073088,0.0055860337,-0.1599078923,0.262078315,-0.0651985332,-0.2783256769,0.0434393212,-0.0880052298,0.4329284132,-0.2539993227,-0.3081473708,0.003639091,-0.0644266978,-0.0977022052,-0.0620605685,-0.1536531448,-0.1169783846,0.0611614808,0.0138719501,0.0663510412,-0.3257706761,0.227166757,0.0191833954,0.1968749166,-0.0655772835,0.3469649255,-0.3650504053,0.6071892977,0.1826376915,0.1984719187,0.2066986561,0.2602954507,0.0882441327,0.022590179,-0.7111612558,0.0119960429,0.1972316355,-0.1737873405,0.0223199911,-0.0742926896,0.0048973546,-0.2321588546,-0.1811950803,-0.1740103066,-0.2393470556,-0.055500295,-0.3643978238,0.022894742,-0.0218463298,0.3673841655,-0.2108635604,0.0025235156,0.0858273208,0.1972029209,0.1464118659,-0.0942063183,-0.5303607583,-0.5024215579,-0.3433577418,0.2414335608,-0.268050611,0.2392213047,-0.1371821761,0.0209193882,0.1031649038,-0.1103468761,0.4958151877,-0.1310851723,-0.1198505163,0.3111153841,-0.0852036104,-0.2169436067,-0.1560447812,-0.2145740539,0.2487731874,0.0919606164,0.4482049048,-0.250259459,0.0183465444,0.0149936453,-0.1805451065,-0.2933079898,-0.4980681539,-0.4371125698,-0.1474029571,-0.3785982728,-0.0684283376,-0.2436530143,0.2505195737,-0.2380888164,-0.0173401665,-0.0446072482,-0.2353481203,0.1815394759,0.1636779308,0.5122842193,-0.1921356916,0.0965679884,0.0463171192,0.0618032813,0.1100911647,0.3877335489,-0.1745917797,-0.2596477866,-0.0316261984,-0.0194504261,0.2970832288,-0.1034843102,-0.2948684096,0.3071787655,-0.0516149625,0.1102492511,0.025938293,0.0147437481,0.2532850504,0.166193068,-0.0491432957,-0.4161556065,0.1546282619,-0.026426319,0.0593896918,0.2063326389,-0.2129352987,-0.0297328606,0.2763859034,-0.0648007765,0.5007886887,-0.1812511981,-0.1933797598,0.3407919109,0.2174998671,0.4038965702,-0.1669141054,0.2915231884,-0.5527960062,-0.0196560118,0.2265730053,-0.1374719739,-0.0402738154,-0.2586255074,-0.5010080338,0.2584686875,-0.1938874722,0.3548452258,0.0251688473,0.2396081388,-0.1991093904,0.0686027333,-0.2652857304,0.1974008232,-0.1181023642,0.3159292936,-0.1748003811,-0.0673146844,0.338904947,-0.1213414446,-0.0972986966,0.3927971721,-0.1331145167,0.249198094,0.1869202554,-0.0859444141,0.1371118277,0.4802416861,0.0360942744,0.2273453474,-0.1351352185,0.065763317,-0.019321125,-0.094333306,-0.112178646,0.1603537798,0.2917366624,-0.1176791191,-0.3513068259,0.4982581437,-0.053962864,0.0808963627,-0.2077497095,0.0966773257,-0.0197356436,-0.2792844772,-0.0582650863,-0.0777066201,0.0835340992,-0.1127732247,0.1955122054,0.1503963768,-0.045955576,0.2849721611,-0.1004797891,-0.2767522335,-0.0368325561,0.1631646305,-0.0202607755,0.2678747773,0.7444613576,0.3148196042,-0.0434075892,-0.4068887234,-0.1119854152,0.1022955924,-0.0354892612,0.2343204767,0.2025268972,-0.0975747928,0.2175710946,0.0701927096,0.0332487226,-0.0487874486,-0.0780247897,-0.4976789057,-0.2899364829,0.5084173083,-0.2384429127,-0.0717616007,-0.0226836149,-0.0233008359,0.0253997985,0.1786234975,-0.4123268425,-0.0047474536,-0.5003218055,-0.1386016607,-0.1944634318,-0.2909421921,0.1350518465,0.1839277148,0.2573776543,-0.0973946154,-0.1259430051,-0.2820370495,0.0279131066,0.018612029,-0.0506161414,0.1341328174,-0.0356099978,-0.2192932963,0.026943488,0.124546729,-0.019082034,-0.1418261826,0.0428549983,0.2309853882,0.0664203838,0.1498230696,-0.1048577055,0.3925117552,-0.1542077214,-0.1091762707,0.0892825425,0.0404633768,0.1454324126,0.0041160379,-0.0036404058,-0.0570225157,0.0763983354,0.0788752958,0.1567761004,-0.2010095417,0.2653385103,-0.1736503392,0.0509918667,0.6181201339,-0.2227316052,0.1117159277,0.0968501717,0.3101870716,-0.227864027,0.1903332025,-0.226213932,-0.0269351378,0.0044274949,0.1368973404,-0.2643563449,-0.1858036667,0.1944193244,0.1868375391,0.2378834039,0.1207540259,0.0021422226,0.2828045785,-0.2137751877,-0.1657215059,0.1804838777,0.1804220527,0.0042540804,0.3604365289,-0.2713555992,-0.0449317358,0.0594246946,0.2507373989,0.1973168105,-0.3961364329,-0.1421324164,0.3903949559,-0.0077800611,-0.2976878583,-0.2608277202,0.4449037015,-0.1044107005,-0.0567574911,-0.1178352311,0.1353802085,-0.3331618309,-0.0949439034,-0.5158820748,-0.0901945084,-0.0454395972,-0.1312888116,0.0252012815,0.0420153625,-0.1148137525,0.0730800107,-0.1637796462,-0.3743667305,0.2009001225,0.2530694306,-0.2437724918,-0.1150775999,0.3914286196,0.1533192545,0.3194291294,0.1970059723,0.0829290003,0.3079617321,0.5380836129,-0.3283626139,-0.0000046035,0.0199086033,-0.1313573867,0.0346756577,0.3137899935,0.2148800194,0.3973927796,0.2980228961,0.2429427803,-0.1719399244,0.1943987459,0.0471306443,0.1650931686,0.0040556081,0.6533156633,-0.0515634567,0.0107027795,-0.2028621137,-0.0492494442,-0.5103918314,-0.0879737213,0.2999315262,0.2265715152,0.0560057573,-0.2306601256,0.1365998983,-0.3244097531,0.4204716682,0.4334087074,0.464944005,-0.4140831232,0.1457706243,-0.4007861614,0.0466136932,0.0658597574,-0.0020465965,-0.1097378805,0.0314681605,-0.0425010435,-0.0867101401,0.131939888,0.0399009325,0.2227889597,-0.1774658263,-0.4814176559,0.1328634173,0.0461931862,0.1684636027,0.0244556442,-0.1332152337,0.1925159246,-0.4258807003,0.099328503,-0.0376955234,-0.2557370663,0.1841211617,-0.2434110194,0.2420452237,0.0028272639,0.2686461508,0.0102750724,-0.0241375919,-0.1938576102,-0.3289037347,-0.1430860758,0.0125012193,0.3162784874,0.2012417167,-0.0406675674,0.2181183398,-0.1372948736,0.1671044677,-0.0101538477,-0.1750993431,-0.1615086496,-0.0480243564,0.2521775365,-0.1583052278,0.0523845926,0.145988524,0.0466804728,0.1564805657,-0.4045481682,-0.0697857961,0.490942657,-0.018943686,-0.0651996136,-0.0544002727,0.2422242314,0.1531124562,0.0249590147,-0.3776175678,0.0245507509,0.3022921383,-0.0252069011,0.077616632,0.3028816879,0.4011399448,0.0964677855,-0.0990418643,-0.283208698,0.1615866125,0.09360677,0.0778960735,0.0205555446]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"This would be a neat feature","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":6,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nThis would be a neat feature","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"> neat feature\r\n\r\nI dint get these clearly, can you please elaborate like how to work on these ","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":18,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\n> neat feature\r\n\r\nI dint get these clearly, can you please elaborate like how to work on these ","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"It could maybe work almost out of the box just by using `cached_path` in the text\/csv\/json scripts, no?","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":18,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nIt could maybe work almost out of the box just by using `cached_path` in the text\/csv\/json scripts, no?","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"Thanks thomwolf and julien-c\r\n\r\nI'm still confusion on what you guys said, \r\n\r\nI have solved the problem as follows:\r\n\r\n1. read the csv file using pandas from s3 \r\n2. Convert to dictionary key as column name and values as list column data\r\n3. convert it to Dataset using \r\n`from datasets import Dataset`\r\n`train_dataset = Dataset.from_dict(train_dict)`","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":55,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nThanks thomwolf and julien-c\r\n\r\nI'm still confusion on what you guys said, \r\n\r\nI have solved the problem as follows:\r\n\r\n1. read the csv file using pandas from s3 \r\n2. Convert to dictionary key as column name and values as list column data\r\n3. convert it to Dataset using \r\n`from datasets import Dataset`\r\n`train_dataset = Dataset.from_dict(train_dict)`","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"We were brainstorming around your use-case.\r\n\r\nLet's keep the issue open for now, I think this is an interesting question to think about.","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":23,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nWe were brainstorming around your use-case.\r\n\r\nLet's keep the issue open for now, I think this is an interesting question to think about.","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"> We were brainstorming around your use-case.\r\n> \r\n> Let's keep the issue open for now, I think this is an interesting question to think about.\r\n\r\nSure thomwolf, Thanks for your concern ","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":32,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\n> We were brainstorming around your use-case.\r\n> \r\n> Let's keep the issue open for now, I think this is an interesting question to think about.\r\n\r\nSure thomwolf, Thanks for your concern ","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"I agree it would be cool to have that feature. Also that's good to know that pandas supports this.\r\nFor the moment I'd suggest to first download the files locally as thom suggested and then load the dataset by providing paths to the local files","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":45,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nI agree it would be cool to have that feature. Also that's good to know that pandas supports this.\r\nFor the moment I'd suggest to first download the files locally as thom suggested and then load the dataset by providing paths to the local files","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"Don't get\n","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":2,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nDon't get\n","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"Any updates on this issue?\r\nI face a similar issue. I have many parquet files in S3 and I would like to train on them. \r\nTo be honest I even face issues with only getting the last layer embedding out of them.","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":42,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nAny updates on this issue?\r\nI face a similar issue. I have many parquet files in S3 and I would like to train on them. \r\nTo be honest I even face issues with only getting the last layer embedding out of them.","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"Hi dorlavie, \r\nYou can find one solution that i have mentioned above, that can help you. \r\nAnd there is one more solution also which is downloading files locally\r\n","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":28,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nHi dorlavie, \r\nYou can find one solution that i have mentioned above, that can help you. \r\nAnd there is one more solution also which is downloading files locally\r\n","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"> Hi dorlavie,\r\n> You can find one solution that i have mentioned above, that can help you.\r\n> And there is one more solution also which is downloading files locally\r\n\r\nmahesh1amour, thanks for the fast reply\r\n\r\nUnfortunately, in my case I can not read with pandas. The dataset is too big (50GB). \r\nIn addition, due to security concerns I am not allowed to save the data locally","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":68,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\n> Hi dorlavie,\r\n> You can find one solution that i have mentioned above, that can help you.\r\n> And there is one more solution also which is downloading files locally\r\n\r\nmahesh1amour, thanks for the fast reply\r\n\r\nUnfortunately, in my case I can not read with pandas. The dataset is too big (50GB). \r\nIn addition, due to security concerns I am not allowed to save the data locally","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"@dorlavie could use `boto3` to download the data to your local machine and then load it with `dataset`\r\n\r\nboto3 example [documentation](https:\/\/boto3.amazonaws.com\/v1\/documentation\/api\/latest\/guide\/s3-example-download-file.html)\r\n```python\r\nimport boto3\r\n\r\ns3 = boto3.client('s3')\r\ns3.download_file('BUCKET_NAME', 'OBJECT_NAME', 'FILE_NAME')\r\n```\r\n\r\ndatasets example [documentation](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html)\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files=['my_file_1.csv', 'my_file_2.csv', 'my_file_3.csv'])\r\n```\r\n","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":46,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\n@dorlavie could use `boto3` to download the data to your local machine and then load it with `dataset`\r\n\r\nboto3 example [documentation](https:\/\/boto3.amazonaws.com\/v1\/documentation\/api\/latest\/guide\/s3-example-download-file.html)\r\n```python\r\nimport boto3\r\n\r\ns3 = boto3.client('s3')\r\ns3.download_file('BUCKET_NAME', 'OBJECT_NAME', 'FILE_NAME')\r\n```\r\n\r\ndatasets example [documentation](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html)\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files=['my_file_1.csv', 'my_file_2.csv', 'my_file_3.csv'])\r\n```\r\n","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"Thanks @philschmid for the suggestion.\r\nAs I mentioned in the previous comment, due to security issues I can not save the data locally.\r\nI need to read it from S3 and process it directly.\r\n\r\nI guess that many other people try to train \/ fit those models on huge datasets (e.g entire Wiki), what is the best practice in those cases?","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":61,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nThanks @philschmid for the suggestion.\r\nAs I mentioned in the previous comment, due to security issues I can not save the data locally.\r\nI need to read it from S3 and process it directly.\r\n\r\nI guess that many other people try to train \/ fit those models on huge datasets (e.g entire Wiki), what is the best practice in those cases?","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"If I understand correctly you're not allowed to write data on disk that you downloaded from S3 for example ?\r\nOr is it the use of the `boto3` library that is not allowed in your case ?","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":37,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\nIf I understand correctly you're not allowed to write data on disk that you downloaded from S3 for example ?\r\nOr is it the use of the `boto3` library that is not allowed in your case ?","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"@lhoestq yes you are correct.\r\nI am not allowed to save the \"raw text\" locally - The \"raw text\" must be saved only on S3.\r\nI am allowed to save the output of any model locally. \r\nIt doesn't matter how I do it boto3\/pandas\/pyarrow, it is forbidden","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":47,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\n@lhoestq yes you are correct.\r\nI am not allowed to save the \"raw text\" locally - The \"raw text\" must be saved only on S3.\r\nI am allowed to save the output of any model locally. \r\nIt doesn't matter how I do it boto3\/pandas\/pyarrow, it is forbidden","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/878","title":"Loading Data From S3 Path in Sagemaker","comments":"@dorlavie are you using sagemaker for training too? Then you could use S3 URI, for example `s3:\/\/my-bucket\/my-training-data` and pass it within the `.fit()` function when you start the sagemaker training job. Sagemaker would then download the data from s3 into the training runtime and you could load it from disk\r\n\r\n**sagemaker start training job**\r\n```python\r\npytorch_estimator.fit({'train':'s3:\/\/my-bucket\/my-training-data','eval':'s3:\/\/my-bucket\/my-evaluation-data'})\r\n```\r\n\r\n**in the train.py script**\r\n```python\r\nfrom datasets import load_from_disk\r\n\r\ntrain_dataset = load_from_disk(os.environ['SM_CHANNEL_TRAIN'])\r\n```\r\n\r\nI have created an example of how to use transformers and datasets with sagemaker. \r\nhttps:\/\/github.com\/philschmid\/huggingface-sagemaker-example\/tree\/main\/03_huggingface_sagemaker_trainer_with_data_from_s3\r\n\r\nThe example contains a jupyter notebook `sagemaker-example.ipynb` and an `src\/` folder. The sagemaker-example is a jupyter notebook that is used to create the training job on AWS Sagemaker. The `src\/` folder contains the `train.py`, our training script, and `requirements.txt` for additional dependencies.\r\n\r\n","body":"In Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load","comment_length":127,"text":"Loading Data From S3 Path in Sagemaker\nIn Sagemaker Im tring to load the data set from S3 path as follows\r\n\r\n`train_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/train.csv'\r\n valid_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/validation.csv'\r\n test_path = 's3:\/\/xxxxxxxxxx\/xxxxxxxxxx\/test.csv'\r\n \r\n data_files = {}\r\n data_files[\"train\"] = train_path\r\n data_files[\"validation\"] = valid_path\r\n data_files[\"test\"] = test_path\r\n extension = train_path.split(\".\")[-1]\r\n datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)\r\n print(datasets)`\r\n\r\n\r\nI getting an error of\r\n\r\n`algo-1-7plil_1 | File \"main.py\", line 21, in <module>\r\nalgo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 603, in load_dataset\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 155, in __init__\r\nalgo-1-7plil_1 | **config_kwargs,\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 305, in _create_builder_config\r\nalgo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))\r\nalgo-1-7plil_1 | File \"\/opt\/conda\/lib\/python3.6\/genericpath.py\", line 55, in getmtime\r\nalgo-1-7plil_1 | return os.stat(filename).st_mtime\r\nalgo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3:\/\/lsmv-sagemaker\/pubmedbert\/test.csv`\r\n\r\nBut when im trying with pandas , it is able to load from S3\r\n\r\nDoes the datasets library support S3 path to load\n@dorlavie are you using sagemaker for training too? Then you could use S3 URI, for example `s3:\/\/my-bucket\/my-training-data` and pass it within the `.fit()` function when you start the sagemaker training job. Sagemaker would then download the data from s3 into the training runtime and you could load it from disk\r\n\r\n**sagemaker start training job**\r\n```python\r\npytorch_estimator.fit({'train':'s3:\/\/my-bucket\/my-training-data','eval':'s3:\/\/my-bucket\/my-evaluation-data'})\r\n```\r\n\r\n**in the train.py script**\r\n```python\r\nfrom datasets import load_from_disk\r\n\r\ntrain_dataset = load_from_disk(os.environ['SM_CHANNEL_TRAIN'])\r\n```\r\n\r\nI have created an example of how to use transformers and datasets with sagemaker. \r\nhttps:\/\/github.com\/philschmid\/huggingface-sagemaker-example\/tree\/main\/03_huggingface_sagemaker_trainer_with_data_from_s3\r\n\r\nThe example contains a jupyter notebook `sagemaker-example.ipynb` and an `src\/` folder. The sagemaker-example is a jupyter notebook that is used to create the training job on AWS Sagemaker. The `src\/` folder contains the `train.py`, our training script, and `requirements.txt` for additional dependencies.\r\n\r\n","embeddings":[-0.2508122325,-0.2407749146,-0.0531521216,0.5038032532,0.2426776439,0.0379796848,0.4234098494,0.230222702,0.0534563586,0.0122716874,-0.0213181283,0.3855955303,-0.236418426,0.408721298,-0.0733910277,0.1519470513,0.0036897035,0.1174415499,0.241575554,0.1663119793,-0.0428829901,0.2773869038,-0.0571360439,0.1090650633,-0.1840185374,0.0136175472,0.0146284783,0.3506195247,-0.075076893,-0.5084875822,0.010816888,-0.0190767832,0.4783017635,0.3833350539,-0.0001139213,-0.0286531113,0.3157631457,0.0450190939,-0.422922045,-0.384821564,-0.5591976643,0.1110935286,0.1526586562,-0.0374609157,-0.3507244587,-0.0653413385,-0.3116801381,-0.2656855285,0.4028450549,0.6623779535,0.1782039702,0.1914703399,-0.073105447,-0.3078445792,0.2821199894,0.0203729924,0.0018274108,0.2220327705,-0.2715070248,0.1779980212,0.2951638997,0.2709996998,-0.4019239247,0.1139450222,0.5365952253,0.0304873791,0.3513378203,0.0784727633,0.127464205,0.1705706865,0.5800996423,-0.1361254901,-0.4543988407,-0.184919998,-0.2141834646,-0.2824923396,0.1915267557,0.0894623622,0.0374890529,0.028242277,-0.0464704409,-0.0466800258,-0.0988201797,0.2466035932,0.0339196883,0.440559864,-0.278003633,0.0633563548,0.2615089417,0.0703382492,0.3421872258,-0.1132721528,0.0181237478,0.0466640703,-0.2519853413,0.16740264,-0.0153579842,0.2122866958,-0.0699794069,0.1938651204,0.3034305573,-0.0739920959,0.3996638358,0.2401907742,0.3883311152,-0.1680295616,0.0913824365,0.0003885696,-0.0994113535,0.4869207144,0.0785982981,-0.1934865415,-0.3133118749,-0.5367276073,0.1240956336,0.268817991,0.3971948624,-0.3510361314,-0.2658778727,-0.0860199854,-0.1192131713,-0.1965766996,0.2694832683,0.4972227812,-0.3471298218,-0.1139851063,-0.1179710552,0.2110883445,-0.2143538892,0.0854033828,-0.1983184814,0.2434528023,-0.319057256,0.0227509346,0.148680076,-0.5799320936,0.5168523192,-0.0738639161,0.0051261182,-0.3440077305,0.1492902786,-0.2164640427,0.1777615249,0.181439817,0.1935322434,0.2570756674,-0.0537814647,-0.0760165527,-0.1820334494,0.0855752379,-0.1849054843,-0.4797708988,0.0412797295,0.2240626961,-0.2657831907,-0.1602720469,-0.0594209172,-0.1859486103,0.2781676948,-0.0984312296,-0.3886997104,0.0453326479,-0.1181020886,-0.2207810432,0.4690745473,0.3192168772,-0.3537423909,0.1085963026,-0.471652329,0.113847062,0.0835102424,0.21051687,-0.3659238815,0.4489594996,-0.3048421741,0.1328942031,0.5610044003,-0.2565937042,0.0060540144,0.0853612125,-0.0421512462,-0.1606775224,0.1342204362,-0.0630417317,-0.1340707988,-0.1085644215,0.0167824328,0.3441396952,0.0382880904,0.0388525389,-0.1566989571,-0.0247691069,0.1257411838,0.2443982661,0.1100840494,-0.0220544655,-0.0373709686,-0.1964025348,-0.0556683019,-0.1184008718,-0.0434881225,0.4679802656,0.273306489,0.301110357,-0.0028269866,0.1516498625,-0.2742566764,0.1333042681,0.0448954403,-0.2772715986,-0.0668589994,0.0106648421,-0.2895232141,0.0034126535,-0.4520739615,0.0989072919,0.1053083465,0.3138039708,0.2326270342,-0.0971172452,-0.1190223917,-0.305985868,0.3305369616,-0.0151232248,-0.3088082671,0.113530919,-0.4051382542,-0.0548965894,0.0869040713,0.0439935476,0.1716851592,0.1176360101,0.0150821852,0.1787704825,0.1243599057,0.2274493873,-0.2144693881,0.1562106162,-0.1199650764,0.030693043,0.1797361076,-0.1334500164,0.1356184781,0.0835730806,-0.044440113,0.4771928191,-0.0267096646,0.189440012,0.1397075802,0.0150925107,0.2696094215,-0.1694880128,-0.0204154514,-0.13537094,0.0476758145,-0.1200017855,0.1146692112,0.0762749314,-0.0729279891,-0.129548803,0.3834590316,0.0959832966,-0.0108043309,-0.2696371973,-0.2296590656,-0.1449694782,0.065362744,0.495529741,0.3665240705,0.22105892,0.0965472981,-0.035579741,-0.0539808832,-0.1855220199,0.2379354239,0.4237586856,0.4380967915,0.3556293249,-0.3781282902,-0.0747246221,-0.2402070016,-0.1607340276,0.1642100364,0.2822508812,-0.4002408981,0.1890033185,-0.0598931015,-0.1696401238,-0.298271358,-0.354865402,-0.1032038108,-0.2554685473,-0.1192265376,0.2769101858,-0.2668563426,-0.027291365,0.0960367247,0.0128378468,-0.0740059763,-0.0552989915,-0.3418391049,-0.2471806109,-0.1704055071,0.0550118871,0.2183091938,0.0958847925,0.2658967972,-0.1865487844,0.1083012894,-0.463339597,-0.0825944766,-0.0176458135,-0.00809725,0.3024953902,0.2742291689,0.0802076086,0.2997995019,-0.2465780228,0.4653754234,-0.012098128,0.0204350967,0.2364755422,-0.0213384144,-0.1121434867,-0.0333958156,-0.3809584379,0.0699729174,-0.4534810483,-0.3755663633,0.0112922871,0.1440520436,0.2753456831,0.3328946829,0.2299250364,-0.2099558264,0.3361966312,-0.2096320391,-0.2037051469,0.1267781556,-0.2523744106,-0.2639803588,0.3304711878,-0.0450135954,0.8056797385,-0.1035317406,-0.2581118941,-0.028181551,0.054154627,-0.3135477901,-0.1422804147,-0.2467391789,0.3080139756,0.208803907,-0.0856468529,-0.0078174854,-0.3258495927,0.0911117196,-0.2148199975,0.2122523487,-0.2017768174,0.8904483914,-0.0051942291,0.4429087043,0.1470208913,-0.2118462175,0.4529938698,-0.1063801572,0.1824462712,-0.1048233956,-0.2602479756,-0.1180830076,0.1664527059,-0.2900656164,0.1482967585,-0.2593997419,-0.0632343963,-0.2945415378,0.0631955415,-0.3335372508,-0.2166402042,0.0108930338,-0.2645999491,0.1334776431,-0.1084317788,0.037141528,-0.0522479564,0.1782280356,0.1964562982,0.27402547,0.1798409969,-0.0360960402,-0.2892716527,-0.0228562299,-0.3024620116,0.3309329748,-0.044929307,0.1445596367,-0.1725998223,-0.2284811884,0.1085917056,-0.212063536,0.5687652826,-0.2886083424,0.0759901106,0.256680727,-0.1874071658,-0.4711462259,-0.2658157051,-0.0675028712,0.2460006773,0.0289376229,0.4087799788,-0.1191496626,-0.4835366011,0.1473179758,-0.056742277,-0.0862192437,-0.2047346383,-0.3479643762,-0.1231660545,-0.3823904991,-0.0772973001,-0.0697868168,0.3470997214,0.0993612707,-0.2050059736,0.0540416799,0.0844748989,0.0261744261,-0.0463366471,0.4726723433,0.0330599062,0.2310328335,0.272346586,0.3769771159,0.1094535515,0.3990899622,0.203102082,-0.4686415493,-0.103952691,0.0307143126,0.4628865719,-0.0234796014,-0.1841691434,0.1056576818,0.1430654079,0.1544755101,-0.1691709906,0.1093408093,0.4040108025,0.0435304232,-0.305380255,-0.5429391265,-0.0136748198,0.0826764032,0.0919229835,-0.0626312867,-0.3203953505,-0.2080414742,0.2609343529,0.4015892446,0.6901736259,-0.0723919272,0.2966907024,0.5587425232,0.0910604149,0.001688217,0.2774413526,-0.1042702571,-0.5139327645,0.1954909563,0.0833371654,-0.1815756261,-0.0093930447,0.2339333445,-0.1757297069,0.2052436024,-0.2196788192,0.0933761597,-0.1733642817,0.1602142602,-0.1688349843,-0.0502103083,-0.1218755692,0.0497322902,-0.069273226,0.0820502788,-0.1036420241,-0.2097410262,0.0337738469,-0.2309396267,0.0748144165,0.2753441632,0.0081000198,0.0343265422,-0.2805510461,-0.1951717585,0.1213368252,0.2447195798,0.3164099157,0.0351313315,-0.0893453956,0.1010994837,-0.3682298958,-0.2763568461,-0.2608578503,-0.0956428275,0.1849868596,-0.2108019888,-0.04998051,0.1240625754,-0.0613715723,-0.3762870431,-0.0335865058,-0.0350460485,0.2754879892,-0.4317034483,-0.2545199692,-0.1319119632,-0.11611256,-0.1919201314,0.160815075,0.2287708521,-0.1201416254,0.015217443,-0.1504315287,-0.27451846,0.0584029406,0.5248114467,-0.1016796082,0.1521701664,0.6307948828,0.0710356385,-0.1138501987,-0.2894229293,-0.2996490896,0.0183575936,-0.2909430861,0.1833910793,0.2434943318,-0.0689215511,0.2441017926,0.307755053,0.1714057475,-0.0989844948,-0.0508271605,-0.5483646393,-0.028405169,0.0474096015,-0.1534883827,0.3620046973,0.1767349392,-0.3220783472,-0.0817603543,-0.1770448834,-0.2993300557,0.0011034457,-0.1197999939,0.2405624986,-0.2157948613,-0.2780376673,0.2113571763,0.0620969944,0.162225619,-0.3558205962,-0.0777116194,-0.2533762157,-0.0681318119,0.1221702024,0.1441825479,0.0095417993,0.1336410791,-0.144476667,-0.3320429623,-0.1306843013,0.0670906752,0.1010915935,-0.0039907838,0.2840555608,0.0924806446,0.0219094437,-0.1482167095,-0.1065009087,-0.1994542927,-0.1743012816,0.0085589848,0.2015719116,-0.1814142168,-0.2146724612,0.0438419692,0.0030675931,0.0793969631,0.0792408884,0.312815696,-0.4647563994,0.0515677333,0.1602487862,0.5055231452,0.3906174004,-0.079903841,0.2140353769,0.3118725419,0.2381579727,-0.1135245264,-0.274047792,-0.1960725039,-0.2340989709,-0.0069741891,-0.0698648021,-0.0677564368,0.0582613237,0.2001817375,-0.0457281396,0.1784252077,0.1479998976,0.1270190179,0.2794730365,-0.3211747706,0.040614251,0.2804034948,0.1732441634,0.1449828893,0.3777576089,-0.1555615962,0.1443804204,-0.2752650082,0.1954726726,-0.1280406713,-0.6779032946,0.1011462808,0.0413988568,-0.0967220813,-0.089592196,-0.1370323449,0.3286107183,0.0310608484,0.0625444576,-0.3678517938,0.0839882866,-0.1466532499,-0.0880457684,-0.3406411111,0.0076265386,-0.0325514823,0.1551521122,0.1516607702,-0.0349882431,-0.0857885927,0.0875550285,-0.04003511,-0.043664366,0.0709544718,0.396187067,-0.231902957,-0.4020034671,0.188653484,-0.1219384372,-0.0789997056,0.0003442157,0.071505405,0.5281403065,0.3324438035,0.1592496336,0.0043607485,-0.0478780158,-0.183707878,-0.0530143976,0.0596844293,0.1682320237,0.3104560077,0.4630704522,0.170382455,-0.1467406899,0.1241716072,-0.1453259885,0.3208077252,-0.0304010194,0.1715204716,0.1750354171,-0.2403035313,-0.2071768194,0.2140195072,-0.5623790622,-0.2455465198,0.3704132438,0.0733886212,0.0577127934,-0.116495803,0.0590368956,0.1694280654,0.5225659013,0.1972418427,0.3079318702,-0.4567182362,-0.0008990353,-0.5817424059,-0.0127128242,0.1645321548,0.2716315091,0.1598724872,0.1856330782,-0.2448574156,-0.0757827833,0.2462042868,0.2971832156,0.1082021371,0.2338159084,-0.1574125886,-0.1347535402,0.151497826,-0.1274897605,0.089990139,-0.1791538,0.0747646689,-0.0472137891,0.0362290181,-0.2751492858,0.0237008911,0.001294943,-0.1667191684,0.5104475617,-0.2578533888,0.4464894533,-0.0321037807,0.1099810302,-0.4661693871,0.0489507988,-0.1495843828,0.1149457395,0.0738409683,0.2687285244,-0.0252671968,0.3849056065,-0.409668833,0.2747506797,-0.160855785,-0.3426374793,-0.0414932184,0.1178343967,0.0767927542,0.3193479776,0.0821502432,-0.0379422046,0.0348787531,0.0631606579,-0.5235251784,-0.2149447054,0.601724565,-0.4399197698,-0.0485644154,-0.0070455624,0.2270691246,0.1117750183,-0.1416441649,-0.4197171628,0.1078704968,0.2636763155,0.1084360704,-0.0447175838,0.3393390775,0.2886273563,0.14588359,-0.0916723683,0.1939832866,0.1492439508,-0.1321775019,-0.1302170008,-0.3663122058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/877","title":"DataLoader(datasets) become more and more slowly within iterations","comments":"Hi ! Thanks for reporting.\r\nDo you have the same slowdown when you iterate through the raw dataset object as well ? (no dataloader)\r\nIt would be nice to know whether it comes from the dataloader or not","body":"Hello, when I for loop my dataloader, the loading speed is becoming more and more slowly!\r\n```\r\ndataset = load_from_disk(dataset_path) # around 21,000,000 lines\r\n\r\nlineloader = tqdm(DataLoader(dataset, batch_size=1))\r\nfor idx, line in enumerate(lineloader):\r\n # do some thing for each line\r\n```\r\nIn the begining, the loading speed is around 2000it\/s, but after 1 minutes later, the speed is much slower, just around 800it\/s.\r\n\r\nAnd when I set `num_workers=4` in DataLoader, the loading speed is much lower, just 130it\/s.\r\n\r\nCould you please help me with this problem?\r\nThanks a lot!","comment_length":38,"text":"DataLoader(datasets) become more and more slowly within iterations\nHello, when I for loop my dataloader, the loading speed is becoming more and more slowly!\r\n```\r\ndataset = load_from_disk(dataset_path) # around 21,000,000 lines\r\n\r\nlineloader = tqdm(DataLoader(dataset, batch_size=1))\r\nfor idx, line in enumerate(lineloader):\r\n # do some thing for each line\r\n```\r\nIn the begining, the loading speed is around 2000it\/s, but after 1 minutes later, the speed is much slower, just around 800it\/s.\r\n\r\nAnd when I set `num_workers=4` in DataLoader, the loading speed is much lower, just 130it\/s.\r\n\r\nCould you please help me with this problem?\r\nThanks a lot!\nHi ! Thanks for reporting.\r\nDo you have the same slowdown when you iterate through the raw dataset object as well ? (no dataloader)\r\nIt would be nice to know whether it comes from the dataloader or not","embeddings":[-0.3241671026,-0.0680928603,-0.1207572892,0.2375429422,-0.0479865931,-0.0551257432,0.4634110332,-0.0104469657,0.27750054,0.1358924508,-0.057494428,0.4406970739,-0.0379497446,-0.0338122584,-0.268433094,0.0123529946,-0.0058121388,0.1865820438,-0.492305249,-0.124488838,0.0327855013,0.0130969109,-0.2001439035,-0.2224330306,-0.1072914302,-0.1239836439,0.1369379908,0.1350974143,-0.1098117679,-0.4247672856,0.0017357268,0.1950880736,0.3713124394,0.6450659037,-0.0001175916,0.0595057197,0.0318348072,0.2282907963,-0.2663004696,0.4214517474,0.0467983857,-0.4410398006,0.1777765304,-0.150221318,0.3183623552,-0.1672293693,-0.362226963,-0.1270672083,0.0363598131,-0.0522836782,0.1114333048,0.5388416052,-0.5062484145,0.0144917658,-0.042780146,-0.2250587642,-0.1712978631,0.3582552969,0.7214533091,0.0632588342,-0.4866904318,0.267052263,0.0660408586,0.3211706579,0.1695450246,0.1509487182,0.4048445225,-0.2836863697,-0.0620002225,0.3078357577,0.7381132841,-0.0354181193,-0.2990169227,-0.0910646394,0.0105067668,-0.1903959215,0.0654632226,-0.3881118298,-0.3574895263,-0.1338220686,-0.3717273772,-0.026922144,0.0679100603,-0.0472422279,-0.1119692475,-0.0657399148,0.0455352738,-0.1303454787,0.1895777434,-0.234932512,0.2986094952,-0.1954666078,0.1711438149,0.1489230841,-0.6994066834,-0.0902779102,-0.2848401964,0.0521829464,-0.0891148224,-0.0181694608,0.354893595,0.1102723554,-0.0667482987,0.0600802004,0.1825924367,0.1178168654,0.004976037,-0.242470935,0.1138005108,0.093819961,-0.1968663633,0.0588540249,-0.2202976346,-0.3105726242,0.601760745,-0.1301610917,-0.1297000051,0.1186170429,-0.2490780056,-0.1723696291,0.1275794506,-0.2305172682,0.3537150025,0.2471130937,-0.1536796242,0.4567603171,-0.2397268414,-0.0503765978,-0.2576823831,-0.4324019253,-0.1131179407,-0.2418783158,-0.1623128206,0.3175446987,-0.0372215956,-0.2119458616,0.0773528814,0.4499238133,0.2396336347,0.0114685306,0.0637501553,-0.2298670262,-0.1895278543,-0.0006900873,-0.0946097225,0.3075334728,0.0369768403,0.1128361449,-0.068998538,0.2015663236,-0.6437550187,0.0785501301,-0.0394436494,0.1380124688,0.2078983933,-0.192004472,-0.274797529,-0.0646827519,-0.1752679199,0.0844000801,0.0575299002,-0.3610057533,-0.0583952814,-0.089168705,-0.0122568877,0.3467795849,-0.2351538241,0.020275604,-0.1796611398,0.0347963311,0.4447409511,0.432410121,-0.3443000317,0.6609274149,-0.4009302855,-0.2704447806,0.4817548394,-0.2173739374,-0.4320207536,0.7465626597,-0.1581312269,0.2618145049,0.3124278784,0.4026639462,0.4636730254,-0.2787517607,0.2423096746,0.1920707524,0.1954514235,0.217764169,-0.147203207,0.1382672936,0.2342145294,0.2169909775,-0.0854829103,0.3467286229,-0.0216153041,-0.2039886862,0.4737628996,-0.2093394548,-0.0713676661,0.3619624972,-0.1518404633,-0.1550887376,0.1403793544,0.1130172685,-0.5651847124,0.2188805491,0.0174571536,-0.2443373054,0.3122852445,0.1191148087,-0.1705151647,0.0810036212,-0.2775758803,0.0633910224,0.0383196808,0.1155604497,-0.1098965108,-0.3375312388,0.0054834839,0.3593555093,-0.0542127639,0.0724648312,0.1110631377,0.2941327691,0.4734765887,-0.1357281953,0.2266172022,-0.3497341275,-0.0022912719,-0.1050759554,-0.1363227516,0.2549382448,0.2548448741,0.4663394392,0.4570462704,0.0839059278,0.1968622655,0.0920431912,0.151091218,-0.1182600558,0.2299347669,-0.1504813284,0.3272106647,0.339900285,-0.4314762056,0.304664731,-0.1557522416,-0.2725868225,0.0748399794,0.0128783425,-0.0994906873,0.3636085093,0.4137886167,0.1891222298,0.5586000681,0.2135505229,-0.1636326313,0.1299816072,0.08290001,0.0867921859,-0.2338926792,0.1753518283,-0.2699453235,-0.0051416652,-0.1918001622,-0.0197301209,0.4462364018,0.0328504369,0.1807245463,-0.1930119544,0.2182891071,-0.1475970298,-0.0196376462,0.2065653205,0.2737132013,0.4008261561,-0.029416373,-0.0897580832,-0.3164697289,-0.2832614183,0.0854961202,0.3603068888,-0.099567987,0.2097769529,-0.2502031028,0.1953954697,-0.1894832402,-0.355763793,-0.1991962045,-0.260418117,-0.1925880611,0.0787446126,0.2964641452,-0.009786549,0.1002823412,-0.0192539357,0.2309795916,0.073518686,-0.1692913473,0.0327541828,-0.0956841335,-0.0191796403,0.2025862932,-0.177978158,0.0635963455,-0.093472667,0.184092015,-0.3834203482,0.1968765855,-0.143396616,0.0445303321,0.1653608978,-0.0252622887,0.3253192306,-0.0218567811,-0.1016788557,0.1156654209,-0.1382929534,-0.0316268504,-0.0245847311,-0.1935520768,0.1150001362,0.3658194244,0.0710399821,-0.2702754438,-0.1006651372,-0.2287169546,0.0387516953,0.132837981,-0.1525261998,-0.1126026064,0.0776150525,0.2564549148,0.026456818,-0.0234877411,-0.7568570971,0.4094848931,0.0923793614,0.0480863713,-0.2590308785,-0.1115682498,0.2096382082,0.2576956153,-0.4277701974,0.0233919453,-0.1314118207,0.2236766219,-0.1926167011,0.1394828409,-0.1026301458,-0.16370444,-0.0642289147,0.0713584647,-0.0732139423,-0.1602997333,-0.2642502487,0.0712503344,-0.1342391074,0.2918333113,-0.2233045697,0.4822228551,0.2165642083,-0.1205234081,0.0885198265,-0.0922472477,0.0981974676,0.1598916799,-0.3891573548,-0.2941411436,-0.1147342622,-0.3326261342,0.3235489726,-0.3457531929,0.0190701447,0.0377804339,-0.0053882767,-0.1034277081,-0.0096180048,0.1407832801,-0.1608001292,0.285082072,-0.1276008487,0.0551184788,-0.3122484088,-0.1316484809,0.0756107122,0.1663702428,0.0100021577,-0.3491051793,-0.4216745496,0.1617165506,-0.4815859795,0.0578870326,-0.0331894159,0.5251989365,0.3719904721,0.0860175565,0.35447523,0.0799848214,0.5257925391,0.2051191628,-0.2232925892,0.0023491015,-0.4758287072,0.014172066,-0.1697371453,-0.2489891499,0.2695690393,0.3460761905,0.4094494879,-0.0674000755,-0.1872821599,-0.0953926072,0.1143275052,-0.2102629095,-0.5684310794,-0.15205203,0.1021388918,0.1626297235,0.1056239456,-0.5356791615,0.2143605649,0.0322541334,-0.0189002734,-0.0836556479,-0.2037841529,0.3269469142,0.0747586936,0.1129341945,0.1472709477,0.3608943224,0.0147418203,-0.0684394091,0.125515908,0.1568820477,-0.0220568776,-0.1695779115,0.1808070093,0.0972108468,0.3712503016,0.2193909585,-0.0518936217,0.1839528531,-0.1259784102,0.1605490297,-0.181938678,-0.2642449737,0.5638887286,0.0309711024,-0.4069402218,-0.1719620675,0.4834611714,0.0989408717,-0.0105490759,0.3498977423,-0.5904158354,0.0142255509,-0.1594046056,0.1420179605,0.7446190715,-0.1264875829,-0.0125972675,0.0659486279,0.3584729433,0.0503184684,-0.1736754179,-0.0291847456,-0.2662471533,-0.427446574,0.0013964836,-0.3462823629,0.4885743558,0.157131061,-0.071206592,0.1693335623,-0.1653468758,-0.0267518684,-0.182356745,0.1607087404,-0.0941554531,-0.4092692435,-0.3305321932,0.050294742,-0.1155602708,0.1641436517,-0.1359096467,-0.196217224,0.0800795108,0.297790885,-0.0549249165,-0.1854807884,-0.308108747,-0.0060290061,0.5209762454,-0.2394126654,-0.1247554421,0.0293194409,0.0641796738,-0.1590401977,0.0430512652,0.2172895968,-0.0277936347,0.3600897491,0.1343170553,-0.2849557996,0.1039281711,-0.0048290705,0.2251033038,0.3471117616,0.0937860385,-0.2201289386,0.0377965756,0.1709920466,0.252641499,-0.2016204,-0.1696676463,0.2142750323,0.1497723609,-0.1001045406,0.0529837832,0.4944340289,-0.2453262508,0.3979700506,0.048455663,-0.3939450085,-0.1342613101,0.2439271808,0.3966488242,-0.0035280609,0.5114946961,0.0457233973,-0.2285662293,-0.1265178472,0.2642816603,0.084812969,-0.2579109967,0.2370849848,0.0847057849,0.382350415,0.1951826513,0.0365418009,-0.0463561565,-0.0795826018,-0.4196228981,-0.3935624659,-0.1383419484,-0.1341870874,-0.0344085358,0.0254788958,-0.2641662359,-0.1282237768,-0.0912749693,0.2852681875,-0.19153063,-0.1164884344,-0.0878033713,-0.0669835731,0.1230996102,-0.232482776,0.0989254266,0.315607518,0.033267267,0.1362623423,-0.0383093469,-0.1720198542,0.093509376,0.1423203051,-0.0538630038,-0.1565545946,0.0857352838,-0.0315394178,0.1166315973,-0.1002720073,-0.059548974,-0.0775006711,0.0313649774,0.0510883629,0.3887100816,-0.1345026791,-0.1091198325,0.2731111348,0.0256584585,0.2094879448,0.2166260928,-0.0087043829,0.1175247282,-0.2616748512,-0.2864661813,0.1258468479,0.0932440683,-0.1184830666,0.3263697922,-0.5417141318,0.2881821096,0.046942886,0.1405947357,0.4909071326,-0.112205863,-0.3595325947,-0.0469776839,0.1188165471,-0.055132661,-0.3106612265,-0.0435383692,0.1552539319,-0.0671742409,0.1591504365,-0.0264544226,0.0291598197,0.1033742204,-0.1961595863,-0.0410267748,-0.2628024817,-0.1473148912,0.120651871,0.263515085,0.1223751754,0.3378017247,0.4258939624,-0.2413260937,0.1406067163,0.0870904028,0.2866999805,0.333116442,0.157248795,-0.1672667563,-0.4285086691,-0.1022029221,0.5111817718,-0.2431716472,-0.1594283581,0.0612277389,-0.0594525523,0.1479033828,-0.1062038839,-0.2463625818,0.2721638381,-0.1473351121,-0.2630119622,0.0548407845,0.0015625667,-0.1366144717,0.0878090188,-0.2911334336,0.0286304317,-0.2513045073,0.257212162,0.1657460928,-0.4683098197,-0.4442843199,0.146336779,-0.0904181302,-0.2134361118,0.0612117127,-0.1709281653,0.275265187,0.1769126654,0.2811591625,0.2175331712,0.3926663101,0.0840732902,-0.1159082204,-0.3157945275,-0.106000185,0.1331929266,0.1759579182,-0.2666062117,0.0788069218,0.2236634493,0.0109628132,-0.2197423875,-0.2131666839,-0.1389034837,-0.1174120158,-0.420017153,0.2254895717,0.1522699445,0.2054167688,-0.2552960515,0.0101229222,-0.3215107918,-0.269421339,0.3446243405,-0.471599102,-0.0606432669,-0.004403797,0.0155245941,-0.4134431779,0.73725456,0.4419795275,0.398958832,-0.3557165861,-0.1895863265,-0.4855272174,-0.1053841934,-0.2492283136,-0.1253550053,0.3149228096,0.4505525231,0.090956524,0.3465802968,0.4759301245,0.0007378706,-0.2330504358,0.1060464531,-0.3805701733,0.0859563947,-0.2519362569,0.1973508,-0.2169899642,-0.2561098039,0.50605762,0.0261227861,0.0850094706,0.1327456832,0.0064479914,0.2202890366,0.5054414868,0.1820296943,-0.0491195619,0.0127558177,0.1249501407,-0.0120063527,-0.0260486994,-0.2467246503,0.0929745436,0.0209720321,0.0895190984,-0.0480322316,-0.0645180866,0.1016282886,-0.1225496009,0.3272275031,-0.1230127513,-0.083908692,-0.105365172,-0.0240380522,0.0994610488,0.2704182565,0.0243564826,0.3296841085,0.3303027153,0.2231872976,-0.3987430334,-0.0195249915,0.1854535043,-0.452390343,-0.2727073133,-0.2303163409,0.2011426985,0.1696782112,0.259323597,-0.1572715491,0.1248508021,0.5009222627,-0.0768934488,0.1301141679,0.3328208327,-0.0222498011,0.3303352594,-0.0723400787,0.1694293469,-0.0651862696,-0.196044147,-0.2223620862,-0.0946179628]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/877","title":"DataLoader(datasets) become more and more slowly within iterations","comments":"> Hi ! Thanks for reporting.\r\n> Do you have the same slowdown when you iterate through the raw dataset object as well ? (no dataloader)\r\n> It would be nice to know whether it comes from the dataloader or not\r\n\r\nI did not iter data from raw dataset, maybe I will test later. Now I iter all files directly from `open(file)`, around 20000it\/s.","body":"Hello, when I for loop my dataloader, the loading speed is becoming more and more slowly!\r\n```\r\ndataset = load_from_disk(dataset_path) # around 21,000,000 lines\r\n\r\nlineloader = tqdm(DataLoader(dataset, batch_size=1))\r\nfor idx, line in enumerate(lineloader):\r\n # do some thing for each line\r\n```\r\nIn the begining, the loading speed is around 2000it\/s, but after 1 minutes later, the speed is much slower, just around 800it\/s.\r\n\r\nAnd when I set `num_workers=4` in DataLoader, the loading speed is much lower, just 130it\/s.\r\n\r\nCould you please help me with this problem?\r\nThanks a lot!","comment_length":64,"text":"DataLoader(datasets) become more and more slowly within iterations\nHello, when I for loop my dataloader, the loading speed is becoming more and more slowly!\r\n```\r\ndataset = load_from_disk(dataset_path) # around 21,000,000 lines\r\n\r\nlineloader = tqdm(DataLoader(dataset, batch_size=1))\r\nfor idx, line in enumerate(lineloader):\r\n # do some thing for each line\r\n```\r\nIn the begining, the loading speed is around 2000it\/s, but after 1 minutes later, the speed is much slower, just around 800it\/s.\r\n\r\nAnd when I set `num_workers=4` in DataLoader, the loading speed is much lower, just 130it\/s.\r\n\r\nCould you please help me with this problem?\r\nThanks a lot!\n> Hi ! Thanks for reporting.\r\n> Do you have the same slowdown when you iterate through the raw dataset object as well ? (no dataloader)\r\n> It would be nice to know whether it comes from the dataloader or not\r\n\r\nI did not iter data from raw dataset, maybe I will test later. Now I iter all files directly from `open(file)`, around 20000it\/s.","embeddings":[-0.3166600168,-0.1029369235,-0.0979052261,0.2440703958,-0.0499429554,-0.0601618662,0.4438633025,0.0588657148,0.2712089717,0.1299746037,-0.1063351855,0.4185294211,-0.0331584513,-0.027514616,-0.2983728051,-0.0049316194,0.0067096082,0.1882975101,-0.4768670201,-0.1194707602,0.0179645866,0.0545865968,-0.2014028579,-0.1911975294,-0.1009984538,-0.1206041723,0.1409040242,0.1710790247,-0.1337466389,-0.4246099591,-0.0138660697,0.1935267895,0.3952450156,0.6482269168,-0.0001178202,0.0566183403,0.0264995378,0.2120862305,-0.2565659285,0.4005936384,0.0597716197,-0.4337366223,0.1683360487,-0.1513475329,0.2930934727,-0.196649611,-0.3613282144,-0.1367076486,0.0631073266,-0.0252258107,0.0939410254,0.5212462544,-0.4703530371,0.0297146291,-0.0337093323,-0.1853243411,-0.1692756563,0.3814854324,0.7359797955,0.0564532205,-0.505846858,0.2526673377,0.0517945513,0.3203217089,0.1774702668,0.1548087746,0.393853277,-0.2973325253,-0.0632712618,0.3236049116,0.7466252446,-0.0210778806,-0.294334203,-0.0697091371,-0.0124599244,-0.2196054161,0.0809603259,-0.3400804996,-0.3324059844,-0.1357454956,-0.323839426,-0.0309910066,0.0235060193,-0.0490880758,-0.0860463753,-0.06609191,0.0649642795,-0.1187759265,0.2368847877,-0.2670112252,0.3284733295,-0.1962855458,0.1320453435,0.1515492648,-0.6910883784,-0.1019920409,-0.3191577792,0.0319953449,-0.0595742948,0.0018932206,0.3282820582,0.1334093362,-0.0833956823,0.0504404306,0.1658605188,0.1284782588,-0.0052224086,-0.2074887305,0.1097568646,0.1336804777,-0.2376725227,0.0282058809,-0.2098190337,-0.3414177597,0.5878216028,-0.1503876597,-0.1432618797,0.1301599294,-0.2243592888,-0.2226841599,0.1200108826,-0.27054739,0.3614211679,0.242932111,-0.1394592673,0.4287482798,-0.2386578768,-0.0481600575,-0.2818553448,-0.4302238226,-0.0740945339,-0.2263042182,-0.1469932944,0.3179359436,-0.0038822857,-0.2084015161,0.06032864,0.4574333727,0.2425083965,-0.0128223496,0.1101611778,-0.2686156332,-0.1184235737,-0.0063191769,-0.1206203401,0.306160897,0.0516659394,0.0760855898,-0.0969517753,0.183252722,-0.6864604354,0.0834369436,-0.0672795996,0.1188875064,0.2085925788,-0.1886491328,-0.2581002414,-0.0979585946,-0.2360369116,-0.0075425874,0.0347404853,-0.3694814742,-0.0533458479,-0.069252193,0.0106374407,0.3819206655,-0.1973311156,0.0350575633,-0.1582673639,0.0454134271,0.454456389,0.4650477171,-0.3375992179,0.6549207568,-0.4523156583,-0.2464648485,0.4832217693,-0.2251610309,-0.4284912944,0.8008061051,-0.1540120691,0.2838916183,0.3193062246,0.4010230005,0.4309623837,-0.2910649478,0.183638826,0.1751568019,0.2073602378,0.2159058154,-0.1610917747,0.1502369046,0.2100468129,0.2198768556,-0.0866752714,0.373167485,0.0138059147,-0.24458462,0.5300358534,-0.2338070571,-0.0645176992,0.3581281304,-0.127601862,-0.1772902906,0.1632803679,0.1161284298,-0.529248476,0.2247512341,-0.0182533301,-0.2675389946,0.3040397465,0.1315128207,-0.1732224077,0.0576077998,-0.2893494666,0.0605174378,0.0344241522,0.1301120073,-0.1397275627,-0.3061434627,-0.0136180939,0.3415113389,-0.0361140966,0.0500603504,0.0910876095,0.2567869723,0.4768992364,-0.1074484885,0.2059673667,-0.3815588355,0.0034207159,-0.1245022044,-0.1431287974,0.2594467103,0.2545198798,0.4919049442,0.4801602364,0.1059655696,0.2099206597,0.0794372186,0.1212286279,-0.1313035786,0.2305027843,-0.1494422704,0.3068238497,0.3459582925,-0.4667885303,0.3317511082,-0.135515973,-0.2834477425,0.0963179097,-0.0036895322,-0.094732523,0.4250706732,0.4221171737,0.212761268,0.5241928101,0.2097854763,-0.16578044,0.1091376692,0.1247464344,0.0873474255,-0.2630177438,0.1912245601,-0.247846663,0.0131325778,-0.1887078434,-0.0204513147,0.4574693441,0.0307617225,0.1961123496,-0.1940348148,0.2421053499,-0.1629879922,-0.0085568558,0.2014562041,0.2979935706,0.4072522223,-0.0514678173,-0.0824852288,-0.2856568992,-0.3075772822,0.0966855139,0.3380295038,-0.1108183116,0.2201793492,-0.2431992739,0.183147788,-0.183497265,-0.3231659234,-0.2017690688,-0.233312577,-0.2080877274,0.1413586289,0.3014655709,-0.0248409063,0.1472413242,-0.020452857,0.2273660451,0.0493646488,-0.1851484179,0.0364046991,-0.0934502929,-0.0383997746,0.259582907,-0.1724710464,0.0450633243,-0.1190327331,0.1774958372,-0.3923738301,0.2000944167,-0.139960736,0.0718207061,0.1724396199,-0.0584421083,0.3595056236,-0.0697539747,-0.0975473076,0.1009128094,-0.1151335835,-0.0050505064,0.0129506458,-0.2145968378,0.1132043973,0.376596719,0.0537094511,-0.2767662108,-0.1288035065,-0.1705857217,0.0340348072,0.1489662081,-0.1220617071,-0.1305592656,0.0655792654,0.2450949997,0.0501666069,0.015743088,-0.7621064186,0.3951426744,0.0677169934,0.0155773954,-0.2487180531,-0.0788237378,0.1925945282,0.242400825,-0.4438734353,0.0044752099,-0.1068583354,0.2340218425,-0.2126053274,0.1059335694,-0.1001092046,-0.1356114,-0.0600721799,0.0671863779,-0.0468254797,-0.1423873454,-0.259316653,0.0523374714,-0.0820427537,0.2709999084,-0.2003933787,0.4641516209,0.2466202825,-0.0860853866,0.135313198,-0.1073710546,0.0927095786,0.1337198913,-0.3944262266,-0.2607373893,-0.1174351722,-0.3128916323,0.3140819967,-0.3539556563,0.0119833639,0.0010052606,0.0014412608,-0.0945533887,-0.0288457721,0.154630512,-0.1546781957,0.2852312028,-0.1361411214,0.0237185732,-0.3096563518,-0.1658522487,0.109975934,0.2008674741,0.0345290415,-0.3326033056,-0.4072929323,0.1362869889,-0.4540002346,0.0484832674,-0.0090089999,0.5288022757,0.3669130802,0.0660330206,0.343811959,0.0605065078,0.5548296571,0.2545396984,-0.2080112547,0.0195860304,-0.458950758,0.0202270411,-0.1629423499,-0.2597132027,0.2689759135,0.3551837504,0.4644612074,-0.055768922,-0.2231532633,-0.1256972402,0.1310108155,-0.1777798384,-0.5715376139,-0.1179521829,0.1032797396,0.1491025537,0.1027582213,-0.5400915146,0.2180642635,0.0086474223,-0.0137991216,-0.0727658272,-0.2045610845,0.3114061356,0.0704570636,0.131979242,0.1789388657,0.3548609018,0.0108647402,-0.0264908522,0.1380889416,0.1556929201,-0.0258878879,-0.1709597111,0.1659111381,0.0766546056,0.3214378059,0.2069893479,-0.0495263524,0.1642781645,-0.1335684657,0.1671289057,-0.1898436695,-0.2364575118,0.569737494,0.0143375862,-0.4116796255,-0.1541273445,0.4692337215,0.0824372619,-0.0329309255,0.3657379746,-0.5970932841,0.0303878859,-0.1592686027,0.118207261,0.7345691919,-0.1399648935,-0.0110860933,0.0630709305,0.3613538444,0.0445493013,-0.199779287,-0.0245236233,-0.2633059621,-0.4968056381,-0.0064577102,-0.3365733922,0.4635772109,0.1743907332,-0.0789043605,0.1565501243,-0.1058643609,-0.0809654593,-0.1944639683,0.2345713526,-0.0844627246,-0.4102780819,-0.2810539305,0.0388456173,-0.1339755356,0.1421446353,-0.113242276,-0.2350192368,0.0851001143,0.3290156722,-0.0351834856,-0.196254909,-0.3124464452,-0.012079116,0.4914206266,-0.2618899345,-0.1324980855,0.0706860572,0.0515630879,-0.1660013795,0.0156612657,0.2186153978,-0.0118653467,0.3512452245,0.1358561814,-0.2737425268,0.1267365515,-0.0248623807,0.2071804553,0.2973391414,0.0734464899,-0.2179531455,0.067145206,0.1551449001,0.2528369427,-0.1812253445,-0.1727756709,0.1670875251,0.0898407698,-0.0780785605,0.0423072353,0.4858026803,-0.2343535721,0.4053570628,0.0650556535,-0.4244103134,-0.1394928843,0.2559439242,0.3345218599,-0.0433806032,0.5527670383,0.054233972,-0.210147053,-0.1225174963,0.3144050241,0.0625266582,-0.2677164972,0.2483632118,0.1136023104,0.425268203,0.2227356583,0.0260569267,-0.0250558816,-0.0971149132,-0.3878438473,-0.3976692557,-0.1520534307,-0.1583001614,-0.0333079137,0.0078741852,-0.2520018816,-0.1751621813,-0.0461986475,0.2650782466,-0.1847981215,-0.091290608,-0.0756388828,-0.0717384145,0.1688823253,-0.1944395751,0.1162437946,0.3230673969,0.0221841075,0.0959104672,-0.0278770849,-0.1538805962,0.0664542392,0.1495305896,-0.0937712342,-0.1992664337,0.082797192,-0.0799876079,0.0872301012,-0.1243393421,-0.1106038839,-0.1067114398,-0.0078495322,0.0647785887,0.3694907129,-0.1574409902,-0.1222184822,0.2708654404,0.0151601126,0.1746544689,0.1952191293,-0.0044967397,0.1171736792,-0.2872149944,-0.2666620314,0.1080245525,0.1165789217,-0.1188655645,0.3841861486,-0.5312282443,0.2938057482,0.0472392142,0.1531573981,0.4938795567,-0.1071816906,-0.3221667707,-0.0204020645,0.108744204,-0.0412736721,-0.3164399564,-0.0432352833,0.139094919,-0.0700434595,0.1413055807,-0.0209870506,0.0084888525,0.120906882,-0.1787742227,-0.0318037309,-0.2754971385,-0.183291316,0.1137369275,0.2706340849,0.0924995095,0.3110259771,0.4216326475,-0.2320677638,0.1466452777,0.0508887731,0.2746365964,0.3467852473,0.1904754639,-0.1923951805,-0.4684841335,-0.1191157773,0.4923173785,-0.238751471,-0.1527888924,0.0457612984,-0.0554562099,0.1530507803,-0.1237033308,-0.2178711295,0.2652499378,-0.1303158998,-0.2375983745,0.0564872511,-0.0068063214,-0.1462867856,0.0895872563,-0.2862944007,0.0816506147,-0.2494117022,0.316223532,0.1634910405,-0.4218569398,-0.4552081227,0.1364046782,-0.0514430031,-0.2444344759,0.0437283255,-0.1586106867,0.236404568,0.1942924559,0.2766426504,0.2307552695,0.4069311023,0.0945903808,-0.0864987522,-0.3239963055,-0.0940492302,0.1422664672,0.1759307086,-0.2421970069,0.0937006176,0.2388682067,-0.0122274542,-0.2029804885,-0.2131847143,-0.1822449863,-0.1011536047,-0.4517961144,0.1729628891,0.190404132,0.1621837318,-0.2738578916,-0.0217474271,-0.3191103637,-0.2476677299,0.3361480236,-0.479113102,-0.0334634967,-0.0005312168,0.0150803849,-0.3787364662,0.7244365811,0.426589191,0.3731871247,-0.37124753,-0.2031199783,-0.4963703454,-0.1356551796,-0.1927957386,-0.092963621,0.3313573599,0.4202614129,0.0891369656,0.2848303616,0.4489074051,-0.0367921069,-0.2154839039,0.1458655447,-0.3762745857,0.0610195473,-0.2251276821,0.2461049408,-0.2304222882,-0.2388395667,0.5200161338,-0.0023021034,0.0758837238,0.1667208076,-0.0106620798,0.238466382,0.5113667846,0.1818321645,-0.014339596,0.0355592333,0.0796925649,0.0058667823,-0.0101934085,-0.2185796052,0.0669599921,0.0282925107,0.0977744311,-0.0449026525,-0.0518781655,0.073815085,-0.150827527,0.3138608932,-0.1692913473,-0.0898571163,-0.1428136081,-0.0491733849,0.1282466799,0.2565408349,0.0382504947,0.3413078189,0.3397577405,0.2115655243,-0.4215071499,-0.0126900924,0.1704146713,-0.4388912916,-0.2576790452,-0.1882519573,0.213484183,0.1332682967,0.298181653,-0.1307428032,0.150523752,0.5181916952,-0.059444759,0.1413767636,0.3339850008,-0.021269789,0.3372211158,-0.0530973449,0.1800500005,-0.0183500871,-0.2241745442,-0.22409679,-0.0788606405]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/876","title":"imdb dataset cannot be loaded ","comments":"It looks like there was an issue while building the imdb dataset.\r\nCould you provide more information about your OS and the version of python and `datasets` ?\r\n\r\nAlso could you try again with \r\n```python\r\ndataset = datasets.load_dataset(\"imdb\", split=\"train\", download_mode=\"force_redownload\")\r\n```\r\nto make sure it's not a corrupted file issue ?","body":"Hi\r\nI am trying to load the imdb train dataset\r\n\r\n`dataset = datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\ngetting following errors, thanks for your help \r\n```\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=32660064, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='test', num_bytes=26476338, num_examples=20316, dataset_name='imdb')}, {'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=0, num_examples=0, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n>>> dataset = datasets.load_dataset(\"imdb\", split=\"train\")\r\n\r\n```\r\n","comment_length":51,"text":"imdb dataset cannot be loaded \nHi\r\nI am trying to load the imdb train dataset\r\n\r\n`dataset = datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\ngetting following errors, thanks for your help \r\n```\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=32660064, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='test', num_bytes=26476338, num_examples=20316, dataset_name='imdb')}, {'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=0, num_examples=0, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n>>> dataset = datasets.load_dataset(\"imdb\", split=\"train\")\r\n\r\n```\r\n\nIt looks like there was an issue while building the imdb dataset.\r\nCould you provide more information about your OS and the version of python and `datasets` ?\r\n\r\nAlso could you try again with \r\n```python\r\ndataset = datasets.load_dataset(\"imdb\", split=\"train\", download_mode=\"force_redownload\")\r\n```\r\nto make sure it's not a corrupted file issue ?","embeddings":[-0.5215904713,-0.0397978574,-0.1209275573,0.4138287604,0.3087592721,0.3442183137,0.4162842035,0.4116177261,0.1893083304,-0.1259101331,-0.2459378541,-0.0363801979,-0.2387997508,0.1189185232,-0.1006124243,-0.1933890283,-0.1093897745,-0.0083492799,0.0082321288,0.0248102322,-0.1954829097,0.2631110549,-0.4074177742,-0.0073810266,-0.5533467531,-0.2076044679,0.0664415807,0.2464725524,-0.1878924072,-0.3777330518,0.4374702871,0.0338272862,0.2985222936,0.4265396595,-0.0001084289,0.0661215633,0.5560089946,-0.1408037841,-0.3377951086,-0.2855277359,-0.2910868526,-0.0734336898,0.2554953098,-0.2121904343,-0.0610393658,-0.2735612988,-0.0876790956,-0.4681127667,0.3000148833,0.4594834447,0.2551243603,-0.0305228662,0.076037012,-0.1456561089,0.0994302854,0.1825692356,0.009692302,0.3030386865,0.0760118738,0.0173887219,0.2145227194,0.0415253453,-0.2540514171,-0.0379882827,0.2962742746,0.0274888687,0.067300871,-0.4677696824,0.0347380377,0.2485592067,0.6612049937,-0.0836040899,-0.3514135182,-0.0572897382,0.1594483256,-0.2821145952,0.1125696599,0.2813580036,-0.1202666163,0.1045412347,-0.4093435705,-0.066220209,-0.0858248621,0.1770364344,0.0300608855,-0.0702761412,0.0590686873,0.0750905573,0.2223216593,-0.1004089639,0.3060327768,0.0287062842,0.0114890561,0.1601479799,-0.2722389698,0.0616545677,-0.309935689,0.0013539037,-0.0696112737,0.2496410906,-0.2069759518,-0.0210834984,-0.0292110536,0.2595463991,0.2834863961,0.1375840008,0.0388361663,0.4064239264,0.1448393166,0.2383134514,-0.2738069892,-0.1059613079,-0.2260224223,-0.1501700431,0.0980359986,0.1302989721,0.0258586034,-0.1813311875,-0.5039566755,-0.0179434661,-0.1861053258,0.0018605727,0.3757007718,0.3940535188,-0.1082345173,0.1711023748,0.1874571294,0.3105209172,-0.1181692928,-0.063041687,-0.1855241805,0.2731193006,-0.1848761886,0.0785208493,0.4066706896,-0.1432269812,0.4070621133,0.0138801718,0.0988444313,-0.2059773356,0.1962580979,-0.3265180588,0.0474708453,0.3069472015,0.2882384956,0.1749825925,0.1750798076,-0.0222159196,-0.1219259053,0.0434901752,-0.2463379949,-0.4131351709,-0.1829316765,0.2787935436,-0.0036364328,-0.0550396219,-0.4261500239,-0.0292973146,0.0431886576,-0.1881007403,-0.1758461297,-0.3051405251,-0.1755626351,-0.1236034855,0.4283623397,0.4742207825,-0.2819431126,-0.0677016973,-0.3545875847,-0.1019600853,0.2071660906,0.2522236705,-0.3016596138,0.2245974392,-0.3644298613,0.1026928723,0.3114507794,-0.4896250367,-0.1315897405,0.307010144,0.1519509554,0.1178987846,0.0510700904,0.1178216711,0.3234882951,0.0152814304,0.2373962253,0.4272217155,-0.0333052389,-0.1566784978,-0.0633250624,-0.2016087919,0.1954357326,0.3601022959,0.3082616925,0.00211513,0.0012145919,0.2097969502,0.2742595971,0.117574349,-0.0408722498,0.2063197494,-0.0526263677,0.3531843424,-0.0541510358,-0.1990605146,-0.4411433637,0.1967228204,0.2666048408,-0.0465099066,-0.1574426889,-0.1011871099,-0.275845468,-0.1793883443,-0.2391406745,0.0034423762,0.1849972755,-0.0831256956,-0.0139205651,0.058201801,-0.437218219,0.396702677,-0.1066199988,0.1430695206,-0.3454021513,0.3588088751,-0.1727034897,-0.2236957699,0.1571170837,0.1187447309,0.1863096654,-0.2496664673,-0.0465861037,0.4384959638,0.3271714747,-0.028225705,-0.0712756068,-0.2227297127,0.1980135292,-0.3737979829,0.1602865607,-0.1303503811,0.0517517813,-0.1354398578,-0.0591690466,0.1639060974,-0.2817702293,0.2440556437,0.2280354053,0.1386639625,0.2511508763,-0.0994582698,-0.0211281814,-0.440608412,0.1827893555,-0.0190275162,0.018330995,-0.0145888012,-0.2629710436,-0.2127719522,0.4115481675,0.1476501226,-0.0533790998,0.1388183981,-0.1717075109,0.1022975519,-0.0761989653,0.1512963623,0.2494870126,0.1761391312,-0.0901043043,-0.073893562,0.1371012181,-0.1112428233,0.2687686086,0.063081257,0.1386984587,0.3271507621,-0.1574336737,-0.0432163775,-0.3005732298,-0.3626558483,0.1164298654,0.266146481,-0.3761330843,-0.0721555799,-0.3137325346,-0.1708401293,-0.0889626667,-0.3290142715,-0.0407060795,-0.3265829384,-0.0876567066,0.3447710276,0.3259420991,0.1502281427,-0.2954137623,-0.1498266309,-0.0587890856,-0.0474688821,-0.0757709593,-0.2140558362,-0.1474243701,0.1250645518,0.4519128203,0.1995556653,0.2789734304,-0.3943610191,-0.0972729251,0.0210898109,0.15831393,-0.0527739674,-0.1372921914,0.1901933104,0.1164055914,0.2480758727,0.1137991101,-0.5119875669,0.3313744664,-0.1434600055,0.0764830038,0.1595186889,0.0767459869,-0.0790889338,0.0146051319,-0.47806409,-0.4822545052,-0.3378787637,-0.0881939232,0.1694431752,0.0430157892,0.2023380548,0.136626929,0.1154308394,0.318382442,0.1276842356,-0.0295815244,-0.3274250031,0.3202656507,-0.2303406298,-0.1603825837,0.2412955314,0.1050819606,0.1461465657,0.0087437248,-0.414042294,0.0979003534,-0.0617495067,0.1573085785,-0.0945536718,-0.146949634,-0.0430615917,0.0699405968,-0.0419445448,0.0470371619,-0.0949973911,-0.03719889,-0.0491980128,0.1601986736,-0.1255298108,0.5187346935,-0.3258313537,0.5013358593,0.2936226428,-0.2053146213,0.3749496639,-0.192737937,0.1645800173,-0.2501033247,-0.6196154952,-0.1546026468,-0.1297275871,-0.1464299858,-0.0048476099,-0.112047635,-0.1894546747,-0.2324501574,0.126452595,-0.3696002066,-0.1551600397,0.001215992,-0.3850667775,0.0758725405,-0.2124606967,0.1771603525,-0.0284011792,-0.2188705653,0.0555678606,0.3343017995,0.1610343307,-0.0290782768,-0.151420787,-0.1663630158,-0.0766227245,0.2000465989,0.0912995115,0.3703827262,-0.28136006,0.1310257614,0.0968395919,0.2025268376,0.7217801213,-0.3861365914,0.0293872692,0.2499885112,-0.0360618681,-0.278817445,-0.0903183892,-0.1689852923,0.2853344083,0.3045784533,0.4047124982,-0.0590221398,-0.1351769716,0.3005590141,0.2268385142,-0.1510554999,-0.2041571736,-0.2009697407,-0.0961608663,-0.3041917682,-0.1563256532,0.0761824101,0.1681107283,-0.1662715673,-0.1548368484,-0.1071051955,-0.0860015675,0.2872059345,0.1429075003,0.4447770715,0.136456877,0.2428423017,0.2942498326,-0.0182972923,0.608650744,0.5828627944,-0.3432210088,-0.1498763114,-0.0054401075,-0.0642554089,0.2022987455,0.0719376206,-0.1482065767,-0.3828922212,-0.0234472528,0.0470136739,0.0367247127,0.0739893094,0.4998264909,-0.1790411323,-0.3552518487,-0.722640872,0.2665428817,0.2397559732,0.1121797189,0.2370267659,0.0124276951,-0.2850595117,0.3875684738,0.013414952,0.6175737977,-0.1111615896,-0.1354353428,-0.0062209,0.2396580577,0.1921861321,-0.2774076462,0.1478308141,-0.3572452068,-0.2386856824,-0.1130825281,-0.1766702384,-0.0648922101,0.1164605245,-0.2836247981,0.2537152469,-0.4340718985,-0.1987107545,-0.3366388083,0.5193693638,-0.1824017465,-0.07923197,-0.060076464,0.2007259279,-0.0260536633,0.0932925567,0.0542489327,0.066475369,-0.1279193461,0.0140652452,-0.4498793185,0.1206536815,-0.0912440866,0.4674479961,-0.236301437,-0.181772992,0.1474528164,0.2513224781,0.1848677844,-0.1039214581,-0.061609745,0.1473174095,0.1996781379,-0.2431143075,-0.0718803033,0.0659915209,0.5382035375,0.0189234354,-0.2080607861,0.1704944074,-0.0192907508,-0.0903050303,0.2119589299,-0.1407053322,0.2336817086,-0.395804286,-0.3060357571,-0.1800658256,-0.0167639609,-0.0842181221,0.158750385,-0.0818678513,-0.0336174145,0.2271247804,0.1778802574,-0.2174728364,0.0068268306,0.3988774717,0.0557020493,0.2617917955,0.605458796,0.3335589468,-0.1888597757,-0.2647025287,0.1452559829,0.1888812184,-0.3561069965,0.196836561,0.0737197101,0.2766790986,0.176698029,0.086505577,0.1420460939,0.0673733875,-0.1388965398,-0.543253541,-0.2428753972,0.2320494801,-0.1529144347,0.1940087974,-0.1489850432,0.0385877602,0.0362786315,-0.089737393,-0.3076259196,0.1704565436,-0.3536539972,0.023795696,0.0332673714,-0.1266146451,0.1435002238,-0.1172337532,0.2098435462,-0.1444977671,0.0421803258,-0.2270584106,0.006778684,0.0830583721,-0.0757992417,-0.2000370175,-0.2402272373,-0.1800373346,-0.2467716634,-0.1115229055,-0.1279415935,0.2296342105,0.1045649871,0.2553446293,0.1092190295,-0.1440081745,-0.1522473991,0.1293580681,-0.0784709156,0.1014202833,0.0305777118,0.0742831305,0.1241870895,-0.0104298824,-0.2565321326,0.0256586932,-0.0393743701,0.0003746384,0.3196094334,-0.2303964049,0.2110101283,0.2362058312,0.2677215338,0.6366466284,-0.1434469521,0.1193625033,-0.1094079167,0.2458394617,-0.2158834934,0.0160182789,0.2568395138,0.018162556,-0.1134101078,0.1555421799,0.0622501075,0.0334846787,-0.1858018488,0.0147244912,0.49768278,-0.0226809308,0.3811484575,0.2872027457,-0.0353963189,-0.1042967364,0.0545421541,0.1543368995,0.2103146017,0.5497243404,-0.3222808242,0.1640164405,-0.1561699957,0.0331005417,-0.0107062459,-0.3065340221,0.0461900197,0.073716864,0.0256276149,-0.1397123039,-0.1254886538,0.3516054451,-0.2902526259,-0.120076485,-0.2268692106,0.0749211535,-0.1423197687,0.109563522,-0.1731152534,-0.1698090583,-0.3732798398,-0.0568979494,-0.0800222605,-0.0284512993,0.1086794138,0.3877878487,-0.1657390594,-0.309850812,0.0844726861,0.0971010476,0.1705610156,-0.3105942011,0.1344516724,0.2724173367,0.1277645379,0.3165692389,0.1486421078,0.5456299782,0.3749960959,-0.2448311299,0.0116230771,0.1487063915,-0.1571664661,0.0753748193,0.299941957,0.1794322133,0.3049530387,0.2285931557,0.2683782578,-0.1066691279,0.0680550709,-0.0231997147,0.2483193576,-0.154508695,0.3747867048,-0.1099103838,-0.0839409381,-0.3794899583,-0.0722261146,-0.3162464201,-0.1649577171,0.3867693245,-0.1232076138,0.003352049,0.004056118,0.0844310448,-0.190573588,0.4992541671,0.3944143951,0.1096335724,-0.4755222499,-0.0928499401,-0.5908115506,-0.0558132194,-0.0984433591,-0.0759910643,0.2061026841,0.2298832238,-0.120437853,0.0457117595,0.0933319852,0.1367587298,-0.0401351228,0.2172304243,-0.2901271582,-0.0901622027,-0.2090145797,0.055877056,-0.166366756,-0.1674050391,0.0828185007,-0.5049977899,0.036049448,0.0008963113,-0.3720471859,0.1805410087,0.3742025197,0.2840799391,-0.0968510807,0.799004674,0.0698330402,-0.0349647626,-0.2617948651,-0.3925072253,-0.1967744976,0.4153388739,0.2645186782,0.3011763096,-0.1691698134,0.0210932996,-0.3125961125,0.3212580979,-0.0824553296,-0.1678817868,-0.3540570736,-0.1923818737,-0.0200940855,0.0330966897,0.2503875196,0.1431867927,-0.067863971,0.1911814213,-0.2078012675,-0.3176259398,0.5762025118,-0.0904495269,-0.1553604603,-0.1684870124,0.096538648,0.2750171423,0.0661562607,-0.5203388333,0.09408281,0.3871851563,-0.1895416379,-0.1051006988,0.1640696377,0.2043433636,0.1031652838,-0.0863084644,-0.076200366,0.0498594269,-0.1477047354,-0.0565747246,0.0410751328]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/876","title":"imdb dataset cannot be loaded ","comments":"I was using version 1.1.2 and this resolved with version 1.1.3, thanks. ","body":"Hi\r\nI am trying to load the imdb train dataset\r\n\r\n`dataset = datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\ngetting following errors, thanks for your help \r\n```\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=32660064, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='test', num_bytes=26476338, num_examples=20316, dataset_name='imdb')}, {'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=0, num_examples=0, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n>>> dataset = datasets.load_dataset(\"imdb\", split=\"train\")\r\n\r\n```\r\n","comment_length":12,"text":"imdb dataset cannot be loaded \nHi\r\nI am trying to load the imdb train dataset\r\n\r\n`dataset = datasets.load_dataset(\"imdb\", split=\"train\")`\r\n\r\ngetting following errors, thanks for your help \r\n```\r\nTraceback (most recent call last): \r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 558, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 73, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=32660064, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='test', num_bytes=26476338, num_examples=20316, dataset_name='imdb')}, {'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=0, num_examples=0, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n>>> dataset = datasets.load_dataset(\"imdb\", split=\"train\")\r\n\r\n```\r\n\nI was using version 1.1.2 and this resolved with version 1.1.3, thanks. ","embeddings":[-0.5215904713,-0.0397978574,-0.1209275573,0.4138287604,0.3087592721,0.3442183137,0.4162842035,0.4116177261,0.1893083304,-0.1259101331,-0.2459378541,-0.0363801979,-0.2387997508,0.1189185232,-0.1006124243,-0.1933890283,-0.1093897745,-0.0083492799,0.0082321288,0.0248102322,-0.1954829097,0.2631110549,-0.4074177742,-0.0073810266,-0.5533467531,-0.2076044679,0.0664415807,0.2464725524,-0.1878924072,-0.3777330518,0.4374702871,0.0338272862,0.2985222936,0.4265396595,-0.0001084289,0.0661215633,0.5560089946,-0.1408037841,-0.3377951086,-0.2855277359,-0.2910868526,-0.0734336898,0.2554953098,-0.2121904343,-0.0610393658,-0.2735612988,-0.0876790956,-0.4681127667,0.3000148833,0.4594834447,0.2551243603,-0.0305228662,0.076037012,-0.1456561089,0.0994302854,0.1825692356,0.009692302,0.3030386865,0.0760118738,0.0173887219,0.2145227194,0.0415253453,-0.2540514171,-0.0379882827,0.2962742746,0.0274888687,0.067300871,-0.4677696824,0.0347380377,0.2485592067,0.6612049937,-0.0836040899,-0.3514135182,-0.0572897382,0.1594483256,-0.2821145952,0.1125696599,0.2813580036,-0.1202666163,0.1045412347,-0.4093435705,-0.066220209,-0.0858248621,0.1770364344,0.0300608855,-0.0702761412,0.0590686873,0.0750905573,0.2223216593,-0.1004089639,0.3060327768,0.0287062842,0.0114890561,0.1601479799,-0.2722389698,0.0616545677,-0.309935689,0.0013539037,-0.0696112737,0.2496410906,-0.2069759518,-0.0210834984,-0.0292110536,0.2595463991,0.2834863961,0.1375840008,0.0388361663,0.4064239264,0.1448393166,0.2383134514,-0.2738069892,-0.1059613079,-0.2260224223,-0.1501700431,0.0980359986,0.1302989721,0.0258586034,-0.1813311875,-0.5039566755,-0.0179434661,-0.1861053258,0.0018605727,0.3757007718,0.3940535188,-0.1082345173,0.1711023748,0.1874571294,0.3105209172,-0.1181692928,-0.063041687,-0.1855241805,0.2731193006,-0.1848761886,0.0785208493,0.4066706896,-0.1432269812,0.4070621133,0.0138801718,0.0988444313,-0.2059773356,0.1962580979,-0.3265180588,0.0474708453,0.3069472015,0.2882384956,0.1749825925,0.1750798076,-0.0222159196,-0.1219259053,0.0434901752,-0.2463379949,-0.4131351709,-0.1829316765,0.2787935436,-0.0036364328,-0.0550396219,-0.4261500239,-0.0292973146,0.0431886576,-0.1881007403,-0.1758461297,-0.3051405251,-0.1755626351,-0.1236034855,0.4283623397,0.4742207825,-0.2819431126,-0.0677016973,-0.3545875847,-0.1019600853,0.2071660906,0.2522236705,-0.3016596138,0.2245974392,-0.3644298613,0.1026928723,0.3114507794,-0.4896250367,-0.1315897405,0.307010144,0.1519509554,0.1178987846,0.0510700904,0.1178216711,0.3234882951,0.0152814304,0.2373962253,0.4272217155,-0.0333052389,-0.1566784978,-0.0633250624,-0.2016087919,0.1954357326,0.3601022959,0.3082616925,0.00211513,0.0012145919,0.2097969502,0.2742595971,0.117574349,-0.0408722498,0.2063197494,-0.0526263677,0.3531843424,-0.0541510358,-0.1990605146,-0.4411433637,0.1967228204,0.2666048408,-0.0465099066,-0.1574426889,-0.1011871099,-0.275845468,-0.1793883443,-0.2391406745,0.0034423762,0.1849972755,-0.0831256956,-0.0139205651,0.058201801,-0.437218219,0.396702677,-0.1066199988,0.1430695206,-0.3454021513,0.3588088751,-0.1727034897,-0.2236957699,0.1571170837,0.1187447309,0.1863096654,-0.2496664673,-0.0465861037,0.4384959638,0.3271714747,-0.028225705,-0.0712756068,-0.2227297127,0.1980135292,-0.3737979829,0.1602865607,-0.1303503811,0.0517517813,-0.1354398578,-0.0591690466,0.1639060974,-0.2817702293,0.2440556437,0.2280354053,0.1386639625,0.2511508763,-0.0994582698,-0.0211281814,-0.440608412,0.1827893555,-0.0190275162,0.018330995,-0.0145888012,-0.2629710436,-0.2127719522,0.4115481675,0.1476501226,-0.0533790998,0.1388183981,-0.1717075109,0.1022975519,-0.0761989653,0.1512963623,0.2494870126,0.1761391312,-0.0901043043,-0.073893562,0.1371012181,-0.1112428233,0.2687686086,0.063081257,0.1386984587,0.3271507621,-0.1574336737,-0.0432163775,-0.3005732298,-0.3626558483,0.1164298654,0.266146481,-0.3761330843,-0.0721555799,-0.3137325346,-0.1708401293,-0.0889626667,-0.3290142715,-0.0407060795,-0.3265829384,-0.0876567066,0.3447710276,0.3259420991,0.1502281427,-0.2954137623,-0.1498266309,-0.0587890856,-0.0474688821,-0.0757709593,-0.2140558362,-0.1474243701,0.1250645518,0.4519128203,0.1995556653,0.2789734304,-0.3943610191,-0.0972729251,0.0210898109,0.15831393,-0.0527739674,-0.1372921914,0.1901933104,0.1164055914,0.2480758727,0.1137991101,-0.5119875669,0.3313744664,-0.1434600055,0.0764830038,0.1595186889,0.0767459869,-0.0790889338,0.0146051319,-0.47806409,-0.4822545052,-0.3378787637,-0.0881939232,0.1694431752,0.0430157892,0.2023380548,0.136626929,0.1154308394,0.318382442,0.1276842356,-0.0295815244,-0.3274250031,0.3202656507,-0.2303406298,-0.1603825837,0.2412955314,0.1050819606,0.1461465657,0.0087437248,-0.414042294,0.0979003534,-0.0617495067,0.1573085785,-0.0945536718,-0.146949634,-0.0430615917,0.0699405968,-0.0419445448,0.0470371619,-0.0949973911,-0.03719889,-0.0491980128,0.1601986736,-0.1255298108,0.5187346935,-0.3258313537,0.5013358593,0.2936226428,-0.2053146213,0.3749496639,-0.192737937,0.1645800173,-0.2501033247,-0.6196154952,-0.1546026468,-0.1297275871,-0.1464299858,-0.0048476099,-0.112047635,-0.1894546747,-0.2324501574,0.126452595,-0.3696002066,-0.1551600397,0.001215992,-0.3850667775,0.0758725405,-0.2124606967,0.1771603525,-0.0284011792,-0.2188705653,0.0555678606,0.3343017995,0.1610343307,-0.0290782768,-0.151420787,-0.1663630158,-0.0766227245,0.2000465989,0.0912995115,0.3703827262,-0.28136006,0.1310257614,0.0968395919,0.2025268376,0.7217801213,-0.3861365914,0.0293872692,0.2499885112,-0.0360618681,-0.278817445,-0.0903183892,-0.1689852923,0.2853344083,0.3045784533,0.4047124982,-0.0590221398,-0.1351769716,0.3005590141,0.2268385142,-0.1510554999,-0.2041571736,-0.2009697407,-0.0961608663,-0.3041917682,-0.1563256532,0.0761824101,0.1681107283,-0.1662715673,-0.1548368484,-0.1071051955,-0.0860015675,0.2872059345,0.1429075003,0.4447770715,0.136456877,0.2428423017,0.2942498326,-0.0182972923,0.608650744,0.5828627944,-0.3432210088,-0.1498763114,-0.0054401075,-0.0642554089,0.2022987455,0.0719376206,-0.1482065767,-0.3828922212,-0.0234472528,0.0470136739,0.0367247127,0.0739893094,0.4998264909,-0.1790411323,-0.3552518487,-0.722640872,0.2665428817,0.2397559732,0.1121797189,0.2370267659,0.0124276951,-0.2850595117,0.3875684738,0.013414952,0.6175737977,-0.1111615896,-0.1354353428,-0.0062209,0.2396580577,0.1921861321,-0.2774076462,0.1478308141,-0.3572452068,-0.2386856824,-0.1130825281,-0.1766702384,-0.0648922101,0.1164605245,-0.2836247981,0.2537152469,-0.4340718985,-0.1987107545,-0.3366388083,0.5193693638,-0.1824017465,-0.07923197,-0.060076464,0.2007259279,-0.0260536633,0.0932925567,0.0542489327,0.066475369,-0.1279193461,0.0140652452,-0.4498793185,0.1206536815,-0.0912440866,0.4674479961,-0.236301437,-0.181772992,0.1474528164,0.2513224781,0.1848677844,-0.1039214581,-0.061609745,0.1473174095,0.1996781379,-0.2431143075,-0.0718803033,0.0659915209,0.5382035375,0.0189234354,-0.2080607861,0.1704944074,-0.0192907508,-0.0903050303,0.2119589299,-0.1407053322,0.2336817086,-0.395804286,-0.3060357571,-0.1800658256,-0.0167639609,-0.0842181221,0.158750385,-0.0818678513,-0.0336174145,0.2271247804,0.1778802574,-0.2174728364,0.0068268306,0.3988774717,0.0557020493,0.2617917955,0.605458796,0.3335589468,-0.1888597757,-0.2647025287,0.1452559829,0.1888812184,-0.3561069965,0.196836561,0.0737197101,0.2766790986,0.176698029,0.086505577,0.1420460939,0.0673733875,-0.1388965398,-0.543253541,-0.2428753972,0.2320494801,-0.1529144347,0.1940087974,-0.1489850432,0.0385877602,0.0362786315,-0.089737393,-0.3076259196,0.1704565436,-0.3536539972,0.023795696,0.0332673714,-0.1266146451,0.1435002238,-0.1172337532,0.2098435462,-0.1444977671,0.0421803258,-0.2270584106,0.006778684,0.0830583721,-0.0757992417,-0.2000370175,-0.2402272373,-0.1800373346,-0.2467716634,-0.1115229055,-0.1279415935,0.2296342105,0.1045649871,0.2553446293,0.1092190295,-0.1440081745,-0.1522473991,0.1293580681,-0.0784709156,0.1014202833,0.0305777118,0.0742831305,0.1241870895,-0.0104298824,-0.2565321326,0.0256586932,-0.0393743701,0.0003746384,0.3196094334,-0.2303964049,0.2110101283,0.2362058312,0.2677215338,0.6366466284,-0.1434469521,0.1193625033,-0.1094079167,0.2458394617,-0.2158834934,0.0160182789,0.2568395138,0.018162556,-0.1134101078,0.1555421799,0.0622501075,0.0334846787,-0.1858018488,0.0147244912,0.49768278,-0.0226809308,0.3811484575,0.2872027457,-0.0353963189,-0.1042967364,0.0545421541,0.1543368995,0.2103146017,0.5497243404,-0.3222808242,0.1640164405,-0.1561699957,0.0331005417,-0.0107062459,-0.3065340221,0.0461900197,0.073716864,0.0256276149,-0.1397123039,-0.1254886538,0.3516054451,-0.2902526259,-0.120076485,-0.2268692106,0.0749211535,-0.1423197687,0.109563522,-0.1731152534,-0.1698090583,-0.3732798398,-0.0568979494,-0.0800222605,-0.0284512993,0.1086794138,0.3877878487,-0.1657390594,-0.309850812,0.0844726861,0.0971010476,0.1705610156,-0.3105942011,0.1344516724,0.2724173367,0.1277645379,0.3165692389,0.1486421078,0.5456299782,0.3749960959,-0.2448311299,0.0116230771,0.1487063915,-0.1571664661,0.0753748193,0.299941957,0.1794322133,0.3049530387,0.2285931557,0.2683782578,-0.1066691279,0.0680550709,-0.0231997147,0.2483193576,-0.154508695,0.3747867048,-0.1099103838,-0.0839409381,-0.3794899583,-0.0722261146,-0.3162464201,-0.1649577171,0.3867693245,-0.1232076138,0.003352049,0.004056118,0.0844310448,-0.190573588,0.4992541671,0.3944143951,0.1096335724,-0.4755222499,-0.0928499401,-0.5908115506,-0.0558132194,-0.0984433591,-0.0759910643,0.2061026841,0.2298832238,-0.120437853,0.0457117595,0.0933319852,0.1367587298,-0.0401351228,0.2172304243,-0.2901271582,-0.0901622027,-0.2090145797,0.055877056,-0.166366756,-0.1674050391,0.0828185007,-0.5049977899,0.036049448,0.0008963113,-0.3720471859,0.1805410087,0.3742025197,0.2840799391,-0.0968510807,0.799004674,0.0698330402,-0.0349647626,-0.2617948651,-0.3925072253,-0.1967744976,0.4153388739,0.2645186782,0.3011763096,-0.1691698134,0.0210932996,-0.3125961125,0.3212580979,-0.0824553296,-0.1678817868,-0.3540570736,-0.1923818737,-0.0200940855,0.0330966897,0.2503875196,0.1431867927,-0.067863971,0.1911814213,-0.2078012675,-0.3176259398,0.5762025118,-0.0904495269,-0.1553604603,-0.1684870124,0.096538648,0.2750171423,0.0661562607,-0.5203388333,0.09408281,0.3871851563,-0.1895416379,-0.1051006988,0.1640696377,0.2043433636,0.1031652838,-0.0863084644,-0.076200366,0.0498594269,-0.1477047354,-0.0565747246,0.0410751328]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/875","title":"bug in boolq dataset loading","comments":"I just opened a PR to fix this.\r\nThanks for reporting !","body":"Hi\r\nI am trying to load boolq dataset:\r\n\r\n```\r\nimport datasets\r\ndatasets.load_dataset(\"boolq\")\r\n```\r\n\r\nI am getting the following errors, thanks for your help \r\n\r\n```\r\n>>> import datasets\r\n2020-11-22 09:16:30.070332: W tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:60] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudart.so.11.0: cannot open shared object file: No such file or directory\r\n2020-11-22 09:16:30.070389: I tensorflow\/stream_executor\/cuda\/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\n>>> datasets.load_dataset(\"boolq\")\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home\/datasets\r\nUsing custom data configuration default\r\nDownloading and preparing dataset boolq\/default (download: 8.36 MiB, generated: 7.47 MiB, post-processed: Unknown size, total: 15.83 MiB) to \/idiap\/temp\/rkarimi\/cache_home\/datasets\/boolq\/default\/0.1.0\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\ntensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists\r\n\r\n\r\n```","comment_length":12,"text":"bug in boolq dataset loading\nHi\r\nI am trying to load boolq dataset:\r\n\r\n```\r\nimport datasets\r\ndatasets.load_dataset(\"boolq\")\r\n```\r\n\r\nI am getting the following errors, thanks for your help \r\n\r\n```\r\n>>> import datasets\r\n2020-11-22 09:16:30.070332: W tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:60] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudart.so.11.0: cannot open shared object file: No such file or directory\r\n2020-11-22 09:16:30.070389: I tensorflow\/stream_executor\/cuda\/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\n>>> datasets.load_dataset(\"boolq\")\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home\/datasets\r\nUsing custom data configuration default\r\nDownloading and preparing dataset boolq\/default (download: 8.36 MiB, generated: 7.47 MiB, post-processed: Unknown size, total: 15.83 MiB) to \/idiap\/temp\/rkarimi\/cache_home\/datasets\/boolq\/default\/0.1.0\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11...\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home\/datasets\r\ncahce dir \/idiap\/temp\/rkarimi\/cache_home\/datasets\/downloads\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/boolq\/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11\/boolq.py\", line 74, in _split_generators\r\n downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 149, in download_custom\r\n custom_download(url, path)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/tensorflow\/python\/lib\/io\/file_io.py\", line 516, in copy_v2\r\n compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)\r\ntensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists\r\n\r\n\r\n```\nI just opened a PR to fix this.\r\nThanks for reporting !","embeddings":[-0.3869605362,-0.0838912427,-0.0665007457,0.2883241773,0.0438405238,0.0618771352,0.4376253188,0.2201648802,0.3304823041,-0.0377961509,-0.2085753083,0.337156564,-0.2383551747,0.283223778,0.0568275675,0.0874554142,-0.0236246847,0.1608762443,-0.0729175881,-0.0738498867,-0.3257868588,0.3313914537,-0.4082230926,0.0091998093,0.1021860242,-0.1036428362,0.0438121296,0.2151902914,-0.1025481895,-0.1810048968,0.3564610779,-0.0007975249,0.2222795784,0.2786723971,-0.0001078239,0.1229071915,0.2594000995,0.0040256525,-0.2295629978,-0.4981689155,-0.2177509964,-0.2001857013,0.1152237207,-0.0788688138,-0.0354141854,-0.0083218738,-0.131290108,-0.4074631035,0.2145261317,0.3374620974,0.2536766827,-0.1183824912,0.0978849158,0.0120429629,0.2631524205,-0.1907168627,-0.0591488257,0.261844337,0.2268665284,-0.0124748154,0.1214487255,0.0596169792,-0.0548957959,0.0289294031,0.1567766219,0.0136594195,-0.2450666577,-0.4116235673,0.0817192644,0.2084687501,0.730246067,-0.3719807267,-0.2936655581,-0.0101190824,0.1574753076,-0.349697262,0.3866807222,0.2698636651,-0.0986044183,0.037988238,0.0036120331,-0.2632412612,-0.1383769959,0.1547578424,-0.146501869,0.150723502,-0.1364014447,0.1147832423,-0.0955759063,-0.1457810104,0.3629250526,0.0639710277,-0.21387963,0.1867697537,-0.4694334269,0.1608710289,-0.0982677415,-0.0921665952,-0.0381351672,-0.0255363639,0.1382977664,-0.0167334788,-0.3334625959,0.2271488309,0.0417479686,0.411075294,0.1430176497,0.1700753868,0.1913610846,0.167780742,-0.1363297552,0.1314877868,-0.286013186,-0.2058374733,-0.2072819471,-0.0166096631,0.4206533432,-0.1571507156,-0.5354081392,-0.01278676,-0.2416020185,0.2071383595,0.2026649415,0.5684051514,-0.1979948431,-0.2000176609,0.2322383821,-0.0983272493,-0.2184522301,-0.2230280191,-0.1672997475,0.1514298916,0.0447223298,-0.1544265449,0.0828656033,-0.2505772114,0.1074656621,0.198867619,-0.0038592562,-0.1586820185,0.1559875906,-0.1280894578,-0.1569602937,0.332752794,-0.0228124484,0.1676560044,0.0281753652,-0.0714224651,-0.1167929992,0.3247256279,-0.4475924671,-0.2047373503,-0.3761184514,0.2280641049,-0.1747250408,-0.0255616046,-0.1292053759,0.2398463339,-0.0284499358,-0.4059328735,-0.1741387099,-0.2199857533,-0.4171388149,-0.2545000613,0.4067872465,0.379737407,-0.721835196,-0.0355733074,-0.2776166201,0.0705956444,0.1980144083,0.1514276266,-0.3038983047,0.2256023735,-0.2927027643,-0.215186134,0.3147742748,-0.1603867263,-0.6935407519,0.1530070901,0.0025058687,0.1899314076,-0.2540836334,0.2007584721,0.0795062706,0.0731834993,0.0166072864,0.3496389091,0.0605257228,-0.058256045,-0.196218431,-0.2025472075,0.1786857992,0.1600362211,0.0435940661,0.0001039102,0.0987449139,-0.0077889818,0.1124231517,0.0595210046,-0.075460583,0.2122410238,0.4286994934,-0.3330630064,0.1749678552,-0.0053272983,-0.7156538367,0.2441764176,-0.004363507,-0.0022171536,0.0064462852,0.0866244584,-0.1742620766,0.1450309306,-0.2025068849,0.0555320345,0.0220845453,-0.0405623727,-0.1508659571,0.1786490381,-0.3294427991,0.62746346,-0.1952595264,0.1320889145,-0.1291275918,0.1880652308,0.1591094881,0.0398843624,0.176959157,0.2990739644,-0.0685007945,-0.104123503,-0.2320394218,0.3005132675,0.0707616955,0.2634524405,-0.2713550031,-0.168611899,0.2498421818,0.0589981563,0.1932523102,0.0781690404,0.084757708,0.0251313951,0.1747727841,0.2414655089,-0.1293069869,0.0208872613,-0.0686473399,0.1026418507,0.2241619378,0.0811002105,0.184528634,-0.1741169542,0.4500230551,0.0767691359,0.4679001272,0.1908534914,-0.1517343521,0.0387746319,0.2942797542,0.2066695988,0.0228063762,0.1856391579,-0.1557674557,0.0374003984,0.2078783959,0.3661500812,0.3415323496,-0.029207984,-0.3238050342,0.1508284807,-0.0420628414,-0.168706134,0.077173762,-0.1357836723,-0.0268956795,0.085288927,0.073813118,0.155342564,0.0124269649,-0.3200937808,0.2480069995,0.3516390026,-0.331722945,0.036611896,-0.1515666246,-0.2471154928,0.0437065847,-0.2348983437,-0.020119682,-0.1890294254,-0.0683524385,0.2175014317,0.0283255149,0.2553221881,0.2362817973,0.4084107578,0.1851765066,-0.4096200466,-0.1020152718,-0.0308746714,-0.3739933372,0.0503384434,0.4008745551,0.0808714405,0.2665649354,-0.1803814024,0.0994646028,-0.2947653532,-0.2167720646,-0.0895123333,0.0580171011,0.4669362009,0.0687248707,-0.0000539693,-0.1010652333,-0.2400234342,0.218405813,0.043314144,-0.1410662383,0.0692570508,-0.0351482593,0.0872873738,-0.188394174,-0.4688681364,-0.3779191971,-0.4696047902,-0.0345970877,-0.0452780798,0.2412164956,0.2227881402,0.0747132823,0.0977527052,0.3064665496,-0.0297804587,0.0407440104,-0.4055051506,0.2691354752,-0.1998114288,-0.4638816118,-0.0674405769,0.0333010368,0.655474782,0.1136731058,-0.6542869806,0.1961147934,-0.0128032444,0.039688915,-0.1549566239,-0.3205379844,-0.1733205616,-0.1037629992,0.1638691276,0.0088503845,-0.2736274898,0.049788259,-0.0490213335,0.3482614458,0.2130174637,0.4302854538,-0.2201614082,0.5505217314,-0.0259402432,-0.1205433905,0.2185139954,0.0305833891,0.3502043784,0.0262686368,-0.4317833185,0.0945264101,0.2031854391,-0.294924438,0.1396898627,-0.0526577979,-0.0800796077,-0.2784977257,-0.0230365675,-0.1645695865,-0.1816980988,0.0615495704,-0.5242297649,-0.0701089948,0.0701786503,-0.02802963,-0.1056012064,-0.2954127789,0.0473906249,0.1989612281,-0.1231795847,0.0129833128,-0.1661007255,-0.3063174784,-0.2590009868,0.2794209421,-0.1435547024,0.4720006287,-0.0869446546,0.1545425206,0.1280600727,0.1406049877,0.3688179255,0.134359017,-0.2023051232,0.2342370898,-0.0991191342,-0.2998874485,-0.1829418987,-0.0496329553,0.4129618406,0.279135704,0.1300392151,-0.0277312677,0.1153523177,-0.1765911728,0.1164236963,-0.2459073663,-0.4588683844,-0.3118872643,-0.1315522939,-0.2180456966,-0.3292867541,-0.2932395041,0.2531573772,-0.1232661679,-0.2432574481,-0.0769650936,-0.233198002,0.2070476711,0.1073518991,0.3580560684,-0.1367629915,0.4462005198,-0.0989645049,0.193307966,0.2017917037,0.3720154762,-0.0385853499,-0.3568095565,-0.0320025347,0.0970401466,0.0076784957,-0.1862906069,-0.3368903995,0.1357554048,-0.1504209936,0.0963873193,0.1518227309,0.3111180961,0.1622388512,0.3080874979,-0.3741059601,-0.5317149162,0.4520749748,0.0116269179,-0.0758813173,0.7785291672,-0.1770778,-0.1833170801,0.3040297329,-0.0680690259,0.4870553017,-0.3155319989,-0.1467308551,0.4302545786,0.1481091231,0.7454965711,-0.2681266069,0.1758441776,-0.5234609246,-0.1224377453,0.0317396522,-0.1860247701,0.1239816919,-0.3128448427,-0.2137595862,0.2281452417,-0.2579331994,0.458470434,0.1016780511,0.0796271116,-0.1870939136,-0.1504049003,-0.2342005372,0.0840261653,-0.147655651,0.2856167257,-0.1601477861,0.0702477694,0.0083797118,-0.0526283719,-0.3777528107,0.219699502,-0.3511234522,0.1904224455,0.1974768341,-0.3003546,0.2309161574,0.7164726853,0.2337072939,0.2223954797,-0.1795072556,0.1739663035,-0.0318723843,0.014408106,0.1212390736,0.1058340594,0.3010649681,-0.1038330644,-0.0993799791,0.5033387542,-0.0490804985,-0.0379643515,0.0604371466,0.2475333065,0.3814709187,-0.1753364801,-0.2627196908,-0.3102827668,0.3199895322,-0.0583604649,0.11773967,0.1340431273,-0.0884171054,0.0655003116,-0.0649370104,-0.4327146709,-0.0953590795,0.152655378,0.065600805,0.0429134183,0.6548742056,0.1696946174,0.0670233816,-0.085906595,-0.0406555794,0.1385495961,-0.1292442232,0.1778517663,-0.0831774399,-0.0098825414,0.0847202465,0.3258040249,0.1707681566,0.221321255,-0.1010223478,-0.5384134054,-0.1645409316,0.1909805685,-0.3423134685,0.072420612,0.0140940892,-0.1038377881,-0.0899931639,-0.0688796043,-0.2565254569,0.1209125817,-0.3521261811,-0.2284742594,-0.1361025721,0.2427481264,0.0544835627,0.0781450197,0.1097099036,0.0793967173,-0.1620194018,-0.2115866393,-0.2000622749,0.0880006999,-0.0115651367,0.0124970628,-0.1334087402,-0.3030138016,-0.03668084,-0.0779174119,0.1170442402,0.0662101582,0.0295617562,-0.0192995332,0.089964956,0.0897111595,-0.2693090439,0.5472531915,-0.1145167798,0.1592840999,0.096821472,0.1754212081,0.0535404049,0.0856809616,-0.152259782,-0.1399876028,-0.2098907977,-0.0946500897,0.2548997402,-0.2929824889,0.4273592234,0.1290460527,0.0547694117,0.3316515982,-0.4680135548,-0.1119233221,0.0752506033,0.1782181263,-0.177643165,0.0105295777,-0.2052609324,0.1555999219,-0.0466004945,0.1336104274,0.1298520714,-0.420360595,-0.035352312,0.0252878405,0.2934327126,-0.0043115709,0.1327414811,0.1555710137,-0.2704759538,0.2117436528,-0.0707832873,0.2413289845,-0.0845678747,0.6476597786,-0.2700987756,0.047854241,0.2105769813,0.2644408941,0.1408813745,-0.6917639971,0.10763859,0.4476806223,-0.2913639843,-0.1531372368,-0.0769903585,0.2691467702,-0.1563839018,0.0657491386,-0.0107785454,-0.0028937785,-0.2302712351,0.0282554682,-0.0174616501,-0.0478897579,0.0813106969,-0.0215899087,0.0183066186,-0.188697964,-0.1430656016,0.1184736788,-0.3856545091,-0.2799485922,0.179797262,0.2837855816,-0.2470885962,-0.3161188662,0.4311686456,-0.0754805952,0.0753436908,0.2144082636,-0.1462122947,0.2707770765,0.5777344108,0.07740058,0.0786655694,0.1424739957,0.029782068,0.1078163087,0.2315504104,0.2563355267,0.5436984301,0.2439774722,0.1708475351,-0.0798749998,0.0944577157,0.1520403028,0.5004565716,-0.2842152119,0.416269362,-0.3972594738,0.1071118861,-0.3402261138,0.0856290981,-0.1754276156,-0.2803033292,0.3753528893,0.0565649234,0.0669605881,-0.0294803381,0.0759802237,-0.1911455244,0.3084782362,0.4819156528,0.496370852,-0.29153952,-0.2681752443,-0.5836856365,0.415697366,0.2240816802,0.0240750667,-0.1111974567,0.2286956012,-0.1878726184,-0.0298718493,-0.0401478522,0.3125475347,-0.0114744361,-0.2909247875,-0.5411710143,-0.3103784919,-0.3781166673,0.0875968486,-0.0697732493,-0.2966274023,0.2563684583,-0.2631497979,-0.0013345645,-0.0938514769,-0.0863939971,0.1231533214,0.0603667386,0.331864804,-0.108577922,0.3283698261,0.2203752249,-0.0265298877,-0.3869496882,-0.3125711381,-0.1110213324,0.1596502364,0.0254753344,0.0969140381,-0.094989188,0.1534350663,-0.1550619006,0.3719188869,0.115152441,-0.1157672703,-0.2083053142,-0.1021851376,0.0360071734,0.1936639249,0.2598772943,0.3127132952,0.0664655194,0.1501929313,-0.1569650322,-0.1465139538,0.5321074724,-0.3778974116,0.0073250495,-0.0293177906,0.2695159018,0.2042550147,-0.224300921,-0.5164439082,-0.0665705279,0.4917192757,-0.2397497147,0.0612734035,0.0444139726,0.1526078135,0.3518846631,-0.2122843862,0.1752812564,0.1546386927,-0.1291125119,0.2478343993,-0.122139588]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/874","title":"trec dataset unavailable ","comments":"This was fixed in #740 \r\nCould you try to update `datasets` and try again ?","body":"Hi\r\nwhen I try to load the trec dataset I am getting these errors, thanks for your help\r\n\r\n`datasets.load_dataset(\"trec\", split=\"train\")\r\n`\r\n```\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```","comment_length":15,"text":"trec dataset unavailable \nHi\r\nwhen I try to load the trec dataset I am getting these errors, thanks for your help\r\n\r\n`datasets.load_dataset(\"trec\", split=\"train\")\r\n`\r\n```\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\nThis was fixed in #740 \r\nCould you try to update `datasets` and try again ?","embeddings":[-0.2240759432,-0.1891383827,-0.1560605913,0.5127977133,0.4812908173,0.0757965222,0.1706149876,0.2018767893,-0.3105413914,-0.0070777978,-0.0842718333,0.2694479525,-0.1539143622,0.0265443679,-0.129180938,-0.2997629046,-0.2681818604,-0.0244284403,-0.2405125946,0.0483351834,-0.0587802902,0.4065684378,-0.125723958,0.0979769975,-0.2362714857,-0.2480489016,0.1409946978,0.4598864019,-0.2826002836,-0.5274050832,0.2849692404,-0.0177245606,0.3709703982,0.3073472977,-0.0001021283,0.1509710401,0.1813474149,-0.2111720741,-0.306563735,-0.3233765662,-0.3610441685,-0.0866183117,0.0757732168,-0.1620913446,-0.3169558644,0.0146298017,-0.1352517307,-0.3862551749,0.3613158762,0.6899943352,0.3637973666,0.1514606923,-0.0283245947,-0.3910897076,0.2264177203,0.0206688475,-0.0893178806,0.1343048513,0.233579874,0.0518326722,0.2063576579,0.1370676905,-0.0454090945,-0.0706335083,0.1702017337,-0.1739807129,0.2600903809,-0.4757940173,0.1216312274,0.1714842319,0.5936956406,-0.0747715533,-0.0821766257,0.1962061673,0.0936465487,-0.4280080199,0.0998888537,0.1892330945,-0.1239952892,0.1831748188,-0.2033871263,-0.0163598824,-0.1373423338,0.0775788426,-0.1374014914,-0.0550735593,-0.1529302597,-0.0545405708,0.091843769,-0.0502964929,0.3392096758,0.0922845006,-0.1810020953,0.0584329255,-0.401509434,0.1006721556,-0.2733131051,-0.2098054737,-0.1319612861,0.40281111,-0.0079075145,-0.1210628301,-0.10228993,0.2246786356,-0.0860977769,0.2009397745,-0.0673919916,0.3941071033,0.1311525851,0.0160408467,0.0133325662,-0.0768501014,-0.419998914,-0.0603024885,-0.048858013,0.3520775139,-0.0871054381,-0.2397856116,-0.4846762121,-0.0441869199,-0.1038339362,0.0582611002,0.1301860064,0.4268060029,0.0628136843,0.1867752671,-0.0311150383,0.1859532297,-0.1286757588,-0.1014356911,-0.1982307583,0.2059634328,-0.2534955442,-0.0435546301,0.2879181206,-0.2600408494,0.469299078,-0.0436072983,-0.1114706025,-0.1143935472,0.0916975215,-0.2650456131,-0.1463737041,0.4750537574,0.0369325168,0.0798024759,0.1752609611,-0.1059956476,-0.0116745317,0.0161533784,-0.348282367,-0.4820892811,-0.2043677568,0.3273501992,0.0279786438,-0.0244218633,-0.0788691342,-0.0661339089,0.199517563,0.0056747179,-0.0995236412,-0.3260039389,-0.3151400685,-0.2239082754,0.4097644389,0.2326906621,-0.5721760988,0.0684436485,-0.1717181653,-0.1442878991,0.2108448744,0.1516979933,-0.2628960013,0.2333089113,-0.2781080306,-0.0175106935,0.2961311936,-0.363389641,-0.010497327,-0.07272131,0.0905746892,-0.0577664264,0.036992956,-0.071947813,0.2615342438,0.0130102979,0.2852233052,0.2011217922,-0.0262266323,-0.1376431435,-0.2602847219,-0.1254175901,-0.0128316609,0.2704815865,0.3017410338,0.0612269752,0.0983132273,-0.080491595,0.184846893,0.0936481729,-0.0455087647,0.3449452817,0.2431674451,0.032414306,-0.0899966732,-0.0994115695,-0.1914825737,0.2174130678,-0.0126739247,0.158986479,-0.2006436437,-0.022018984,-0.453796953,-0.0773894265,-0.1460566968,-0.0995203331,0.1834416091,0.0587672219,0.2581504881,0.049402371,-0.1740527153,0.3094017208,0.0333382487,-0.003760231,-0.0576821044,0.4963739812,-0.1554124653,-0.1004945338,0.0874522105,0.173438102,0.1461244226,-0.3698685765,-0.1382739693,0.2650900185,0.3832599819,-0.1713000387,-0.052605927,0.030284496,0.2669124007,-0.3482354581,0.1235647574,-0.0553770885,0.1726935804,0.0458114929,0.2732149065,0.2183525413,-0.2407518178,0.1193748415,0.2427416891,0.1041708365,0.3852187991,0.022029303,0.2470780164,-0.0348203033,0.238078773,0.1400066018,0.0248525571,-0.1059380621,-0.3515887558,0.0206237752,0.1517464072,0.2041469216,-0.1789947003,-0.1551303118,-0.239075616,0.271448493,0.0613146275,0.3359176815,0.3362541795,0.2429400384,-0.2055936605,-0.1046907157,0.1783907562,-0.2023523301,0.2118508369,0.0598972477,0.4868586063,0.4026939571,0.0727825165,-0.0813554823,-0.2903768718,-0.2248474509,0.1750530601,0.4474345148,-0.1128532663,0.1110542715,-0.1393177658,-0.3932133615,0.0561419725,-0.33960253,-0.115590468,-0.3334448338,0.115237847,0.1213995367,0.1581493914,0.1088576168,-0.3941227794,-0.0427042618,-0.082400687,0.1095313206,-0.0745449066,-0.1658824533,-0.1335544735,0.2228135914,0.301273495,0.0544815399,0.2698025107,-0.2179455906,0.0221887883,-0.0934360772,0.0266391803,0.1765350699,-0.0060986467,0.1582209021,0.0704213232,0.1427804232,0.0026559343,-0.182052508,0.3320333064,0.0289808232,0.077508986,0.189229697,-0.0161327273,-0.1276092082,-0.1767940074,-0.59174788,-0.7456588149,-0.4230188429,-0.2232384235,0.0632297471,0.2144909054,0.3098082244,-0.0765922517,0.0556808151,-0.0816435292,-0.0721956715,-0.111635223,-0.2436315864,0.173043862,-0.3071800172,-0.3240697682,0.1565102488,0.0230668057,0.3781657517,-0.1507125199,-0.4343943,-0.0599716976,-0.1219854951,0.2636187077,-0.1777512729,-0.254512161,0.2447733879,0.057920482,-0.1834361404,-0.0212377906,0.0650949627,-0.0434069894,-0.0293859988,0.2543842793,-0.2415582836,0.3700798154,-0.0753223002,0.6227113605,0.2297004461,-0.2114059776,0.3377933204,-0.1926215291,0.1812559515,0.0414117128,-0.4108155072,-0.2863261104,-0.0922190771,-0.0821291283,0.1032686606,0.0800325498,-0.0198808406,-0.2535728216,0.2554444671,-0.2055705637,-0.1657082289,-0.1372539699,-0.0200855806,0.2368067801,-0.0006997248,0.2393251806,0.0769370571,-0.1933849901,-0.0880350173,0.3441258967,0.0349233598,-0.0086388141,-0.2222765833,-0.1602435857,-0.4073936939,0.1254253983,0.0805624798,0.3283106685,-0.2838206589,0.0634717569,0.1014071256,0.0413852371,0.4455864727,-0.2319888175,-0.09219791,0.3710724413,-0.0727267712,-0.1205944046,-0.2734822035,-0.1236978471,0.1143569201,0.321775049,0.421639353,0.0664762035,-0.2635364532,0.1733637303,0.1499916613,-0.2662637234,-0.0861174315,-0.2409995049,-0.2350275367,-0.213279739,-0.1840509474,-0.1287232637,0.3332981765,-0.1160959527,-0.1492751539,0.0617740601,-0.1191463172,0.2661498189,0.145618856,0.3197060227,0.0047665625,0.209876731,0.0833957717,0.0161231533,0.3467456996,0.6243855357,-0.2091498077,-0.0162975602,0.1002544314,-0.0692851543,0.3107989728,-0.0380380377,-0.1746347845,0.0229388904,-0.236232996,0.1823988706,-0.0068228319,0.1409838051,0.1954028308,-0.1706965715,-0.4670507312,-0.2723291218,0.0211106297,-0.0779283345,0.1352162063,-0.0424259789,-0.0543259084,-0.1606990695,0.134313032,-0.135290876,0.6496952772,-0.2068393975,-0.0852998644,0.0328672379,0.2123313993,0.1706420481,-0.5275483727,0.2523849905,-0.3724553585,-0.292627275,-0.1440812498,-0.1282721311,-0.1043935418,-0.0642356053,-0.1302622259,0.2709556818,-0.1872669607,0.0630025342,-0.16591537,0.1166357771,0.0235563982,0.1235782877,-0.0039678542,0.2780224085,-0.0691929385,0.2756775022,-0.0721848086,-0.0977892429,0.0090726474,-0.027373001,-0.2633858025,0.2005570978,-0.4239985347,0.2991537154,-0.147605896,-0.1582648456,0.2772197723,0.0550048389,0.1889999062,0.1087882817,-0.1314546019,0.1115829274,0.0729560554,-0.3029617965,-0.2043101937,0.2553197443,0.1872167885,-0.1521196514,-0.2120906115,0.4443241656,-0.0822089985,0.0438034274,0.0990850925,-0.1392192841,0.2406107485,-0.1944972426,0.0844609439,0.0300548431,-0.0680467412,-0.1539527774,0.2276661545,0.0920309424,-0.1159081459,0.1464741677,0.1236235723,-0.2496787757,0.0093015423,0.223060593,0.1015426964,0.1094123498,0.5173196793,0.4673837423,-0.1717287302,-0.4369762838,-0.0689091384,0.0427414402,-0.5380026102,0.1862756759,0.185779199,0.1668301523,0.204739809,0.4199910462,0.2051145583,-0.2273218036,-0.0388151966,-0.5593430996,-0.3340878189,0.0558359362,0.0169092696,0.1571891606,-0.1007332653,-0.1817000955,-0.092874907,-0.0292596053,-0.4078367949,0.1411808729,-0.422072649,0.0793454349,-0.0796759725,-0.1749951243,0.3485801816,0.0072437087,0.2479034364,-0.0682774708,-0.3510243297,-0.3303359449,0.0342241824,0.0976689979,0.053904023,-0.3677451313,0.0418416075,-0.0827764049,-0.1233590618,0.1301679611,0.0536865667,0.1750522852,-0.0552242137,0.2377417237,-0.1670637578,0.0867047235,-0.1421962082,0.2236778438,0.1769455969,0.2199096978,0.1091810167,0.2470529675,0.0758924261,-0.1457904726,-0.3984348476,0.0044678608,-0.0388165116,-0.1446317434,0.2778137922,-0.3273468912,0.2341758013,-0.0474432819,0.1914362609,0.1739644706,0.0257309172,0.2252423018,-0.0168688055,0.3185099661,-0.3731405139,0.0445112996,-0.2408952862,0.1968016773,0.0352598578,0.1043796316,0.2929666638,-0.0592706054,0.1254289895,0.1163959131,0.3555493057,0.0351339616,0.3022553027,0.391459614,0.1272570789,-0.0984137654,0.0458012633,-0.1744153351,-0.0577814095,0.4059921205,-0.3803471029,0.406065017,-0.0386299454,-0.2481488585,0.1508409977,-0.2902107537,0.1829164177,0.0570016503,0.0243060663,-0.1505505294,-0.0094099203,0.4734877646,-0.0363058895,-0.0708057508,-0.3587545156,0.2241030186,-0.1937678456,0.1288270354,-0.0884512514,-0.3128544092,-0.2070668936,0.019660918,0.0014104446,-0.0254246835,0.1243443415,0.2644387782,-0.1285712719,-0.3196577728,0.197651729,0.3489187658,-0.1421086937,-0.1306139678,0.2791151702,0.4111112058,-0.0362854451,0.0708668157,0.1855561733,0.438146919,0.223391369,0.0246023946,0.0509225875,-0.0630650446,-0.0144276368,0.2165222168,0.294498533,0.0152958455,0.2285864204,0.2288653702,0.2380897254,-0.1208187193,-0.1731190979,0.0598139614,-0.1312666237,0.108987838,0.4257694781,0.0151976002,-0.0962107927,-0.227216199,0.094420366,-0.251883924,0.3239656687,0.4686419964,0.0327762067,0.1891157478,0.0083695417,0.1263009608,-0.3402481079,0.3257192969,0.3519151211,0.1580166519,-0.3806728721,-0.1893755645,-0.2770898044,-0.0100819729,-0.081626147,-0.0681137666,0.0071093915,-0.0619149059,0.0630834848,0.1590077281,0.2134357542,0.2474719435,0.1426685005,0.2584801018,-0.482984513,0.0333073623,-0.0678560361,-0.0743075833,-0.0781259313,-0.3803732991,-0.0781450123,-0.4235623479,0.1393739432,-0.2467064112,-0.053698279,0.0493010879,0.0719870031,0.1433852762,-0.0796139464,0.5888292789,0.1127403453,0.093510367,-0.0285529532,-0.467715621,0.0088279098,0.1077240556,0.3931409419,0.2104096264,-0.0528077297,-0.1432992816,-0.3667438328,0.3225114942,0.0123727908,0.0671888441,-0.157173723,0.0714857057,-0.115926303,0.0566350371,0.2818861008,-0.0422167815,0.1495508999,0.052116856,-0.2895562649,-0.3224889636,0.5122329593,-0.230109334,-0.033263348,-0.2177451402,0.2154729962,0.3125780821,-0.0308132265,-0.5867868066,0.0577223636,0.3870896995,-0.2206560224,-0.0162981637,0.3754563034,0.0091362372,-0.0368317105,0.0090297284,0.2376734763,-0.0515198186,-0.2084713727,-0.1225912422,-0.1195775941]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/874","title":"trec dataset unavailable ","comments":"This has been fixed in datasets 1.1.3","body":"Hi\r\nwhen I try to load the trec dataset I am getting these errors, thanks for your help\r\n\r\n`datasets.load_dataset(\"trec\", split=\"train\")\r\n`\r\n```\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```","comment_length":7,"text":"trec dataset unavailable \nHi\r\nwhen I try to load the trec dataset I am getting these errors, thanks for your help\r\n\r\n`datasets.load_dataset(\"trec\", split=\"train\")\r\n`\r\n```\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \" \/idiap\/home\/rkarimi\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 477, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\nThis has been fixed in datasets 1.1.3","embeddings":[-0.2240759432,-0.1891383827,-0.1560605913,0.5127977133,0.4812908173,0.0757965222,0.1706149876,0.2018767893,-0.3105413914,-0.0070777978,-0.0842718333,0.2694479525,-0.1539143622,0.0265443679,-0.129180938,-0.2997629046,-0.2681818604,-0.0244284403,-0.2405125946,0.0483351834,-0.0587802902,0.4065684378,-0.125723958,0.0979769975,-0.2362714857,-0.2480489016,0.1409946978,0.4598864019,-0.2826002836,-0.5274050832,0.2849692404,-0.0177245606,0.3709703982,0.3073472977,-0.0001021283,0.1509710401,0.1813474149,-0.2111720741,-0.306563735,-0.3233765662,-0.3610441685,-0.0866183117,0.0757732168,-0.1620913446,-0.3169558644,0.0146298017,-0.1352517307,-0.3862551749,0.3613158762,0.6899943352,0.3637973666,0.1514606923,-0.0283245947,-0.3910897076,0.2264177203,0.0206688475,-0.0893178806,0.1343048513,0.233579874,0.0518326722,0.2063576579,0.1370676905,-0.0454090945,-0.0706335083,0.1702017337,-0.1739807129,0.2600903809,-0.4757940173,0.1216312274,0.1714842319,0.5936956406,-0.0747715533,-0.0821766257,0.1962061673,0.0936465487,-0.4280080199,0.0998888537,0.1892330945,-0.1239952892,0.1831748188,-0.2033871263,-0.0163598824,-0.1373423338,0.0775788426,-0.1374014914,-0.0550735593,-0.1529302597,-0.0545405708,0.091843769,-0.0502964929,0.3392096758,0.0922845006,-0.1810020953,0.0584329255,-0.401509434,0.1006721556,-0.2733131051,-0.2098054737,-0.1319612861,0.40281111,-0.0079075145,-0.1210628301,-0.10228993,0.2246786356,-0.0860977769,0.2009397745,-0.0673919916,0.3941071033,0.1311525851,0.0160408467,0.0133325662,-0.0768501014,-0.419998914,-0.0603024885,-0.048858013,0.3520775139,-0.0871054381,-0.2397856116,-0.4846762121,-0.0441869199,-0.1038339362,0.0582611002,0.1301860064,0.4268060029,0.0628136843,0.1867752671,-0.0311150383,0.1859532297,-0.1286757588,-0.1014356911,-0.1982307583,0.2059634328,-0.2534955442,-0.0435546301,0.2879181206,-0.2600408494,0.469299078,-0.0436072983,-0.1114706025,-0.1143935472,0.0916975215,-0.2650456131,-0.1463737041,0.4750537574,0.0369325168,0.0798024759,0.1752609611,-0.1059956476,-0.0116745317,0.0161533784,-0.348282367,-0.4820892811,-0.2043677568,0.3273501992,0.0279786438,-0.0244218633,-0.0788691342,-0.0661339089,0.199517563,0.0056747179,-0.0995236412,-0.3260039389,-0.3151400685,-0.2239082754,0.4097644389,0.2326906621,-0.5721760988,0.0684436485,-0.1717181653,-0.1442878991,0.2108448744,0.1516979933,-0.2628960013,0.2333089113,-0.2781080306,-0.0175106935,0.2961311936,-0.363389641,-0.010497327,-0.07272131,0.0905746892,-0.0577664264,0.036992956,-0.071947813,0.2615342438,0.0130102979,0.2852233052,0.2011217922,-0.0262266323,-0.1376431435,-0.2602847219,-0.1254175901,-0.0128316609,0.2704815865,0.3017410338,0.0612269752,0.0983132273,-0.080491595,0.184846893,0.0936481729,-0.0455087647,0.3449452817,0.2431674451,0.032414306,-0.0899966732,-0.0994115695,-0.1914825737,0.2174130678,-0.0126739247,0.158986479,-0.2006436437,-0.022018984,-0.453796953,-0.0773894265,-0.1460566968,-0.0995203331,0.1834416091,0.0587672219,0.2581504881,0.049402371,-0.1740527153,0.3094017208,0.0333382487,-0.003760231,-0.0576821044,0.4963739812,-0.1554124653,-0.1004945338,0.0874522105,0.173438102,0.1461244226,-0.3698685765,-0.1382739693,0.2650900185,0.3832599819,-0.1713000387,-0.052605927,0.030284496,0.2669124007,-0.3482354581,0.1235647574,-0.0553770885,0.1726935804,0.0458114929,0.2732149065,0.2183525413,-0.2407518178,0.1193748415,0.2427416891,0.1041708365,0.3852187991,0.022029303,0.2470780164,-0.0348203033,0.238078773,0.1400066018,0.0248525571,-0.1059380621,-0.3515887558,0.0206237752,0.1517464072,0.2041469216,-0.1789947003,-0.1551303118,-0.239075616,0.271448493,0.0613146275,0.3359176815,0.3362541795,0.2429400384,-0.2055936605,-0.1046907157,0.1783907562,-0.2023523301,0.2118508369,0.0598972477,0.4868586063,0.4026939571,0.0727825165,-0.0813554823,-0.2903768718,-0.2248474509,0.1750530601,0.4474345148,-0.1128532663,0.1110542715,-0.1393177658,-0.3932133615,0.0561419725,-0.33960253,-0.115590468,-0.3334448338,0.115237847,0.1213995367,0.1581493914,0.1088576168,-0.3941227794,-0.0427042618,-0.082400687,0.1095313206,-0.0745449066,-0.1658824533,-0.1335544735,0.2228135914,0.301273495,0.0544815399,0.2698025107,-0.2179455906,0.0221887883,-0.0934360772,0.0266391803,0.1765350699,-0.0060986467,0.1582209021,0.0704213232,0.1427804232,0.0026559343,-0.182052508,0.3320333064,0.0289808232,0.077508986,0.189229697,-0.0161327273,-0.1276092082,-0.1767940074,-0.59174788,-0.7456588149,-0.4230188429,-0.2232384235,0.0632297471,0.2144909054,0.3098082244,-0.0765922517,0.0556808151,-0.0816435292,-0.0721956715,-0.111635223,-0.2436315864,0.173043862,-0.3071800172,-0.3240697682,0.1565102488,0.0230668057,0.3781657517,-0.1507125199,-0.4343943,-0.0599716976,-0.1219854951,0.2636187077,-0.1777512729,-0.254512161,0.2447733879,0.057920482,-0.1834361404,-0.0212377906,0.0650949627,-0.0434069894,-0.0293859988,0.2543842793,-0.2415582836,0.3700798154,-0.0753223002,0.6227113605,0.2297004461,-0.2114059776,0.3377933204,-0.1926215291,0.1812559515,0.0414117128,-0.4108155072,-0.2863261104,-0.0922190771,-0.0821291283,0.1032686606,0.0800325498,-0.0198808406,-0.2535728216,0.2554444671,-0.2055705637,-0.1657082289,-0.1372539699,-0.0200855806,0.2368067801,-0.0006997248,0.2393251806,0.0769370571,-0.1933849901,-0.0880350173,0.3441258967,0.0349233598,-0.0086388141,-0.2222765833,-0.1602435857,-0.4073936939,0.1254253983,0.0805624798,0.3283106685,-0.2838206589,0.0634717569,0.1014071256,0.0413852371,0.4455864727,-0.2319888175,-0.09219791,0.3710724413,-0.0727267712,-0.1205944046,-0.2734822035,-0.1236978471,0.1143569201,0.321775049,0.421639353,0.0664762035,-0.2635364532,0.1733637303,0.1499916613,-0.2662637234,-0.0861174315,-0.2409995049,-0.2350275367,-0.213279739,-0.1840509474,-0.1287232637,0.3332981765,-0.1160959527,-0.1492751539,0.0617740601,-0.1191463172,0.2661498189,0.145618856,0.3197060227,0.0047665625,0.209876731,0.0833957717,0.0161231533,0.3467456996,0.6243855357,-0.2091498077,-0.0162975602,0.1002544314,-0.0692851543,0.3107989728,-0.0380380377,-0.1746347845,0.0229388904,-0.236232996,0.1823988706,-0.0068228319,0.1409838051,0.1954028308,-0.1706965715,-0.4670507312,-0.2723291218,0.0211106297,-0.0779283345,0.1352162063,-0.0424259789,-0.0543259084,-0.1606990695,0.134313032,-0.135290876,0.6496952772,-0.2068393975,-0.0852998644,0.0328672379,0.2123313993,0.1706420481,-0.5275483727,0.2523849905,-0.3724553585,-0.292627275,-0.1440812498,-0.1282721311,-0.1043935418,-0.0642356053,-0.1302622259,0.2709556818,-0.1872669607,0.0630025342,-0.16591537,0.1166357771,0.0235563982,0.1235782877,-0.0039678542,0.2780224085,-0.0691929385,0.2756775022,-0.0721848086,-0.0977892429,0.0090726474,-0.027373001,-0.2633858025,0.2005570978,-0.4239985347,0.2991537154,-0.147605896,-0.1582648456,0.2772197723,0.0550048389,0.1889999062,0.1087882817,-0.1314546019,0.1115829274,0.0729560554,-0.3029617965,-0.2043101937,0.2553197443,0.1872167885,-0.1521196514,-0.2120906115,0.4443241656,-0.0822089985,0.0438034274,0.0990850925,-0.1392192841,0.2406107485,-0.1944972426,0.0844609439,0.0300548431,-0.0680467412,-0.1539527774,0.2276661545,0.0920309424,-0.1159081459,0.1464741677,0.1236235723,-0.2496787757,0.0093015423,0.223060593,0.1015426964,0.1094123498,0.5173196793,0.4673837423,-0.1717287302,-0.4369762838,-0.0689091384,0.0427414402,-0.5380026102,0.1862756759,0.185779199,0.1668301523,0.204739809,0.4199910462,0.2051145583,-0.2273218036,-0.0388151966,-0.5593430996,-0.3340878189,0.0558359362,0.0169092696,0.1571891606,-0.1007332653,-0.1817000955,-0.092874907,-0.0292596053,-0.4078367949,0.1411808729,-0.422072649,0.0793454349,-0.0796759725,-0.1749951243,0.3485801816,0.0072437087,0.2479034364,-0.0682774708,-0.3510243297,-0.3303359449,0.0342241824,0.0976689979,0.053904023,-0.3677451313,0.0418416075,-0.0827764049,-0.1233590618,0.1301679611,0.0536865667,0.1750522852,-0.0552242137,0.2377417237,-0.1670637578,0.0867047235,-0.1421962082,0.2236778438,0.1769455969,0.2199096978,0.1091810167,0.2470529675,0.0758924261,-0.1457904726,-0.3984348476,0.0044678608,-0.0388165116,-0.1446317434,0.2778137922,-0.3273468912,0.2341758013,-0.0474432819,0.1914362609,0.1739644706,0.0257309172,0.2252423018,-0.0168688055,0.3185099661,-0.3731405139,0.0445112996,-0.2408952862,0.1968016773,0.0352598578,0.1043796316,0.2929666638,-0.0592706054,0.1254289895,0.1163959131,0.3555493057,0.0351339616,0.3022553027,0.391459614,0.1272570789,-0.0984137654,0.0458012633,-0.1744153351,-0.0577814095,0.4059921205,-0.3803471029,0.406065017,-0.0386299454,-0.2481488585,0.1508409977,-0.2902107537,0.1829164177,0.0570016503,0.0243060663,-0.1505505294,-0.0094099203,0.4734877646,-0.0363058895,-0.0708057508,-0.3587545156,0.2241030186,-0.1937678456,0.1288270354,-0.0884512514,-0.3128544092,-0.2070668936,0.019660918,0.0014104446,-0.0254246835,0.1243443415,0.2644387782,-0.1285712719,-0.3196577728,0.197651729,0.3489187658,-0.1421086937,-0.1306139678,0.2791151702,0.4111112058,-0.0362854451,0.0708668157,0.1855561733,0.438146919,0.223391369,0.0246023946,0.0509225875,-0.0630650446,-0.0144276368,0.2165222168,0.294498533,0.0152958455,0.2285864204,0.2288653702,0.2380897254,-0.1208187193,-0.1731190979,0.0598139614,-0.1312666237,0.108987838,0.4257694781,0.0151976002,-0.0962107927,-0.227216199,0.094420366,-0.251883924,0.3239656687,0.4686419964,0.0327762067,0.1891157478,0.0083695417,0.1263009608,-0.3402481079,0.3257192969,0.3519151211,0.1580166519,-0.3806728721,-0.1893755645,-0.2770898044,-0.0100819729,-0.081626147,-0.0681137666,0.0071093915,-0.0619149059,0.0630834848,0.1590077281,0.2134357542,0.2474719435,0.1426685005,0.2584801018,-0.482984513,0.0333073623,-0.0678560361,-0.0743075833,-0.0781259313,-0.3803732991,-0.0781450123,-0.4235623479,0.1393739432,-0.2467064112,-0.053698279,0.0493010879,0.0719870031,0.1433852762,-0.0796139464,0.5888292789,0.1127403453,0.093510367,-0.0285529532,-0.467715621,0.0088279098,0.1077240556,0.3931409419,0.2104096264,-0.0528077297,-0.1432992816,-0.3667438328,0.3225114942,0.0123727908,0.0671888441,-0.157173723,0.0714857057,-0.115926303,0.0566350371,0.2818861008,-0.0422167815,0.1495508999,0.052116856,-0.2895562649,-0.3224889636,0.5122329593,-0.230109334,-0.033263348,-0.2177451402,0.2154729962,0.3125780821,-0.0308132265,-0.5867868066,0.0577223636,0.3870896995,-0.2206560224,-0.0162981637,0.3754563034,0.0091362372,-0.0368317105,0.0090297284,0.2376734763,-0.0515198186,-0.2084713727,-0.1225912422,-0.1195775941]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/873","title":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error","comments":"I get the same error. It was fixed some days ago, but again it appears","body":"```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab","comment_length":15,"text":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab\nI get the same error. It was fixed some days ago, but again it appears","embeddings":[-0.224293083,0.1419460326,-0.0225758813,0.2384324819,0.3819946051,0.1261650622,0.6168980598,0.2147594988,-0.0231766459,0.1334128678,-0.163756609,0.0582367294,-0.3759562671,-0.123077713,0.0085837077,0.114968501,-0.0964015499,0.1582960486,-0.2051095515,-0.1422860175,-0.3901512027,0.1864792407,-0.0768181756,0.0159925465,-0.2544403374,-0.2072729915,-0.1191695929,0.4782720804,-0.2281916291,-0.3415712714,0.291224122,-0.1937130839,0.2596129179,0.6446203589,-0.0001103697,0.1348124146,0.257455349,-0.0459846482,-0.5990345478,-0.6107162833,-0.1344918758,-0.1810776293,0.2224275321,-0.2857463658,-0.0545736887,0.1120457277,0.0625397936,0.019521933,0.3589360118,0.5676545501,0.2382051796,0.2500875592,-0.0214656368,-0.1423246413,-0.1394215375,0.3917793632,-0.08329615,0.5298015475,-0.3784783185,-0.0074536237,0.3009919226,0.2560061216,-0.3467578292,0.2481284142,0.3105844557,0.0931708962,0.2014932483,-0.318579495,0.0102561293,0.052064389,0.462816,-0.377645731,-0.2139322609,-0.0260961484,0.0874807909,-0.4519509375,0.3440143466,0.3034414649,-0.1303296834,0.0531827994,-0.146174401,-0.1438385248,-0.1814096421,0.4520489275,-0.1718828231,0.2167272717,-0.0570031144,0.0317016952,0.0401946232,0.0549006499,0.5230491161,-0.2012925595,0.1109838113,0.1689763218,-0.0259168874,0.0619172007,0.200421989,-0.1987337023,-0.1959023029,0.1629692465,-0.0814728886,-0.2725152671,-0.0308208447,0.0420363732,0.0562603734,0.1875984669,-0.0870422944,0.5960316658,0.2050020099,0.2587226927,-0.0126039628,-0.1288865656,-0.3524653614,-0.249772191,0.1027377322,0.0556450076,0.6368337274,0.0808701962,-0.262622416,-0.1706470996,-0.2875089049,0.0266700387,0.1824934632,0.3770372868,0.0879202038,-0.0015779705,0.1018526927,0.2980035841,-0.0611399859,0.3207466304,-0.2227891982,0.2121761739,-0.1194054484,-0.03893435,0.270580411,-0.2202292234,0.1217843741,-0.1732522994,-0.0137373274,-0.0063968683,0.1550607979,-0.1759008318,0.074225083,0.5078181028,-0.1087166741,0.2788080275,0.270821631,0.0265562665,-0.3172303438,0.1326729208,-0.1793326437,-0.4073957503,-0.0890357345,0.191819638,0.0433196686,-0.1540216804,0.0512391925,-0.2044744641,-0.1510185897,-0.1646996737,-0.0049300003,-0.1914943755,-0.0878131837,-0.3597047925,0.294339478,0.6152353883,-0.3911423981,0.0004790114,-0.4250386953,-0.1540962905,0.1708229333,0.2825909257,-0.1810852587,0.418251425,-0.2191811651,-0.2109978348,0.4664492309,-0.4039042294,-0.4069173932,0.1619243324,-0.1690849811,-0.0231488161,0.0143251596,0.1950932592,-0.1637684554,0.0874652714,0.0317935087,0.3440898061,-0.0140823061,0.0791541114,-0.095202975,-0.2891000509,0.1494426131,0.1394836754,0.2537146509,0.2065376937,0.0017706892,0.057444334,0.0345365591,-0.1340152472,-0.0199387986,0.193029359,0.4896597266,0.0706251934,-0.3033567369,-0.0080354828,-0.5152108073,0.1154644713,0.0506806336,0.2688595951,-0.2443437576,0.0475811958,-0.2807430923,-0.1213214844,-0.2518719435,-0.0720805973,0.1413440108,0.362624079,0.0602164939,0.06581752,-0.2875800133,0.2650077939,0.0555366054,0.1482611448,-0.1362416446,0.4147193134,-0.1979314238,-0.0624308884,-0.0436080284,0.1352329403,0.2345956862,-0.1446532458,-0.1301756352,0.3645673692,-0.1508782059,-0.0773257464,0.0398339257,-0.1054766253,0.0736475289,-0.0363827124,0.2342951298,0.3620508015,0.1599349976,-0.0431439169,-0.1002390012,0.2904100418,0.161951825,0.0390076637,0.1538697779,-0.0371430255,0.0395891629,-0.0555200279,0.1082931161,-0.041275423,0.2116338313,0.0997428,0.4739992917,0.0227582809,-0.1307650357,-0.3711324632,0.2705956399,-0.134705767,0.1444204599,0.1506156623,-0.2888510227,0.0992454216,-0.1900030524,0.1472333819,0.6052938104,0.1164713353,0.0325042419,0.1258146018,-0.0870899335,-0.0704523772,0.2077787369,0.0883645713,0.2569261789,0.2160902619,0.0695267469,0.0620280765,-0.3969856501,-0.2876626253,0.0426680557,0.4161446691,-0.3931626678,0.059591908,-0.1244833395,-0.2166050375,-0.0977737382,-0.5352479219,-0.0786898881,-0.3155716062,-0.1555617154,0.1735693216,0.1367603093,0.1662632525,-0.1210425571,-0.1262292862,0.1058900803,-0.4859184921,-0.2068847418,-0.0484767519,-0.0174885709,0.0651339069,0.2804500163,-0.1032850295,0.2476293594,-0.1312977076,-0.1217798218,-0.1863985956,-0.0907368809,-0.0835310295,-0.1042266935,0.2084366828,0.2644519508,0.3542820513,0.0467904173,-0.1947212368,0.3619304001,-0.1828137487,0.0223103501,0.3297629058,-0.1260294914,-0.0134168044,-0.0603099614,-0.5458139777,-0.5458031297,-0.3574637175,-0.1193348393,0.128277272,0.1643208861,0.439052999,0.149543047,0.1939637363,-0.1065731868,0.0041575646,-0.0272339098,-0.1540678293,0.4063127637,-0.2877522409,-0.2439707518,0.1713533998,-0.0405313782,0.1901890039,-0.1609371156,-0.4939940274,-0.2556856275,-0.0731515959,0.038228672,-0.1987717301,-0.051362209,0.2149719,0.0962023139,-0.100954771,0.1519619524,-0.1301398873,-0.0496761128,0.1687473506,0.0203161482,-0.2199278027,0.5267876983,-0.2912866473,0.5844215155,0.1298206747,-0.0976593196,0.3287144303,-0.2017037868,0.435521692,-0.2107226998,-0.5223431587,-0.0787455291,0.2892427742,-0.1790048331,0.1800332218,0.0565606989,-0.2591916919,-0.2643011212,-0.0444018207,-0.3779843152,-0.0743462443,0.1186384708,-0.0179859605,0.0781653449,-0.059277568,0.0965221897,-0.1504167318,-0.1698943824,-0.1216434762,0.3782083094,0.0190812517,0.1019870639,-0.0288795531,0.0270932689,-0.3145630658,0.2298073471,-0.0520729199,0.071546942,-0.2187217623,0.0831159204,-0.0285105985,0.0293108635,0.3612942696,-0.0923879296,0.0112150013,0.4276471138,0.072345525,-0.485065043,-0.0946589336,-0.2761999965,0.1676738858,-0.001272671,0.2047504038,-0.1170452461,-0.1677084416,0.3938706517,0.2733164728,-0.0797540098,-0.3248061538,-0.4682425559,-0.3894968033,-0.4203393757,-0.0765345469,-0.1250401139,0.0756565779,0.1052619666,-0.1257726848,-0.0364609472,0.0204616189,0.1145620719,-0.0140965423,0.2421565503,0.0633371398,0.1261455715,0.1231676489,0.1575559825,0.3707537055,0.5872366428,-0.1024381369,-0.3017361462,-0.0854127482,0.1388856322,0.0639618933,0.2151001841,-0.2014811933,-0.3429370821,0.0943885595,0.1914982945,0.0416608639,-0.0360612422,0.5686709881,-0.1586035043,-0.3970258832,-0.2941640317,0.4028087556,-0.132741496,0.1745825559,0.2641446888,-0.1621219516,-0.193986252,0.1144844517,-0.2864453197,0.7768808007,-0.112184912,0.2149889618,0.6759076715,0.1097218022,0.4076379538,-0.0114143025,0.137216568,-0.311796248,-0.1835328639,-0.1531409025,-0.2621161938,0.0990275741,0.2857246995,-0.1570022702,0.4138245583,-0.2599776685,0.1028068587,-0.127301529,0.1603608727,-0.0852782652,0.0861265436,-0.3097000718,0.1809026301,0.0234545097,0.3886294961,-0.2298669666,0.1079559177,-0.151260823,-0.2345502526,0.0025140594,0.1233494729,-0.4199483693,0.2769888043,-0.0176260546,-0.1646072417,-0.1869401336,0.247104615,-0.0704445019,-0.0975803882,-0.1174648926,0.0991975963,-0.1049690396,-0.0900420919,-0.1319454163,0.0436581001,0.2979975641,-0.0978417844,-0.2941854596,0.0657169744,-0.1450731903,-0.279520601,0.2277370542,0.0566369779,0.2674985826,-0.3583118021,-0.2808163762,-0.0401816852,-0.0521988161,-0.0776873454,0.1419314146,0.0039220243,0.019712897,0.0714475289,-0.2435005307,-0.3559811711,-0.1612131447,0.2835132182,-0.2098854929,0.0607408918,0.5739116073,0.1368587017,-0.2089036703,-0.3013285995,0.1953066587,0.0391629376,-0.1753415763,0.2251401544,0.2365752906,0.2650004625,-0.0125541659,0.255510658,0.2237352729,-0.0629739016,-0.1974655241,-0.3072934151,-0.1711298227,0.0286055654,0.0978401601,0.2520788312,0.0883995891,0.0631099939,0.3833882511,-0.2658512294,-0.3185986578,0.0119840642,-0.016286904,-0.0612216629,0.0742244571,-0.0595226027,0.5541501045,0.0120948935,0.2147975415,-0.3351586461,-0.0658874363,-0.2823093235,-0.0192318764,0.1273972094,0.1241839752,-0.1153774783,0.0820352361,-0.257568568,-0.2182910442,-0.1288654506,-0.0556795076,0.238222301,-0.063959077,0.1493335813,0.0589087121,-0.0790562183,-0.1343450695,0.0602044202,-0.2575852871,0.0918631107,0.0379097909,0.1405527741,0.2113551497,-0.1355248392,-0.4185724258,-0.0494291224,0.0334853791,0.0262343343,0.2199243456,-0.1869715601,-0.0101036849,-0.0072820913,0.3588445187,0.1904233992,-0.2865005136,0.0540695675,0.2150326967,0.2182925791,-0.2886954546,0.0796895772,-0.0444070697,0.0619237162,0.1027720124,0.1913689822,-0.3624694645,0.0646900684,-0.196090281,0.0315474756,0.3888284564,-0.197690323,0.0516497865,0.3998382986,-0.1716947705,0.0677291825,0.2848182023,0.0445199311,0.1066730469,0.5403015018,-0.169164896,0.027433278,-0.2749078274,-0.115499109,0.0093393354,-0.3157514334,0.0527968816,0.1499107182,-0.0275917388,-0.0061320551,-0.1501961946,0.2725358903,-0.45939973,0.1213818267,-0.3764194548,0.3777875602,-0.1425220072,0.1625410467,-0.1579572111,0.0105228955,-0.012090886,0.0446171537,0.0195148569,-0.2471570969,0.2412227839,0.4903680086,-0.1194794476,-0.1413522214,0.3903640211,0.2625901401,-0.1352918744,-0.2359022647,0.3025464416,0.3034740388,0.1200008765,-0.0492572039,0.4870747924,0.4269422591,0.4090619385,0.0039016933,0.2013225257,-0.0100115081,-0.1889312714,-0.0214954503,0.0882004872,-0.0550458506,0.5318027735,0.1638848484,0.2567686439,-0.2468539625,0.185496375,0.0133697847,-0.0159491897,-0.1292088777,0.691208601,-0.2564795911,-0.1056898162,-0.2331102788,0.0544373505,-0.5289071202,0.2520841956,0.3535762727,0.0812641606,0.1096357256,0.0725061074,0.0778281316,0.0445873886,0.2782047987,0.342743665,0.30538252,-0.5661407709,0.0243918225,-0.6365788579,0.1711902022,-0.28920421,-0.0434493534,-0.1806510538,-0.015653681,-0.136431843,0.1296598315,0.1261664033,0.0536839589,-0.0534693822,-0.105646655,-0.2898376286,-0.0953034908,-0.0626657307,0.0257232394,0.015950067,-0.4950198531,0.0248199925,-0.413179934,0.0006352767,0.027058905,-0.133508116,0.0396981873,0.3611627519,0.4533376694,-0.026389746,0.5661256313,-0.0605582707,-0.1965637505,-0.1913531125,-0.2649549842,-0.1598000079,0.2294674814,0.1461038738,0.3851700723,-0.3064142168,-0.0087604532,-0.4270081222,0.2901292741,-0.0378126204,-0.067394942,-0.2009867877,0.0023696769,-0.1316203475,0.1983623058,0.1464611888,-0.0151976235,-0.0840773508,0.1883648485,-0.2635794878,-0.3884920478,0.46128425,-0.2634094059,-0.2436618209,0.0428768806,0.1699992418,-0.0931862071,0.1788178533,-0.4839810431,0.3811461627,0.438337028,0.1640297323,-0.2190377563,0.1378456354,0.0010955908,0.1383843273,0.0792405829,0.4123993814,0.0075344145,-0.3825779259,-0.0340820029,-0.1186552644]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/873","title":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error","comments":"Hi @mrm8488 it's working again today without any fix so I am closing this issue.","body":"```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab","comment_length":15,"text":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab\nHi @mrm8488 it's working again today without any fix so I am closing this issue.","embeddings":[-0.224293083,0.1419460326,-0.0225758813,0.2384324819,0.3819946051,0.1261650622,0.6168980598,0.2147594988,-0.0231766459,0.1334128678,-0.163756609,0.0582367294,-0.3759562671,-0.123077713,0.0085837077,0.114968501,-0.0964015499,0.1582960486,-0.2051095515,-0.1422860175,-0.3901512027,0.1864792407,-0.0768181756,0.0159925465,-0.2544403374,-0.2072729915,-0.1191695929,0.4782720804,-0.2281916291,-0.3415712714,0.291224122,-0.1937130839,0.2596129179,0.6446203589,-0.0001103697,0.1348124146,0.257455349,-0.0459846482,-0.5990345478,-0.6107162833,-0.1344918758,-0.1810776293,0.2224275321,-0.2857463658,-0.0545736887,0.1120457277,0.0625397936,0.019521933,0.3589360118,0.5676545501,0.2382051796,0.2500875592,-0.0214656368,-0.1423246413,-0.1394215375,0.3917793632,-0.08329615,0.5298015475,-0.3784783185,-0.0074536237,0.3009919226,0.2560061216,-0.3467578292,0.2481284142,0.3105844557,0.0931708962,0.2014932483,-0.318579495,0.0102561293,0.052064389,0.462816,-0.377645731,-0.2139322609,-0.0260961484,0.0874807909,-0.4519509375,0.3440143466,0.3034414649,-0.1303296834,0.0531827994,-0.146174401,-0.1438385248,-0.1814096421,0.4520489275,-0.1718828231,0.2167272717,-0.0570031144,0.0317016952,0.0401946232,0.0549006499,0.5230491161,-0.2012925595,0.1109838113,0.1689763218,-0.0259168874,0.0619172007,0.200421989,-0.1987337023,-0.1959023029,0.1629692465,-0.0814728886,-0.2725152671,-0.0308208447,0.0420363732,0.0562603734,0.1875984669,-0.0870422944,0.5960316658,0.2050020099,0.2587226927,-0.0126039628,-0.1288865656,-0.3524653614,-0.249772191,0.1027377322,0.0556450076,0.6368337274,0.0808701962,-0.262622416,-0.1706470996,-0.2875089049,0.0266700387,0.1824934632,0.3770372868,0.0879202038,-0.0015779705,0.1018526927,0.2980035841,-0.0611399859,0.3207466304,-0.2227891982,0.2121761739,-0.1194054484,-0.03893435,0.270580411,-0.2202292234,0.1217843741,-0.1732522994,-0.0137373274,-0.0063968683,0.1550607979,-0.1759008318,0.074225083,0.5078181028,-0.1087166741,0.2788080275,0.270821631,0.0265562665,-0.3172303438,0.1326729208,-0.1793326437,-0.4073957503,-0.0890357345,0.191819638,0.0433196686,-0.1540216804,0.0512391925,-0.2044744641,-0.1510185897,-0.1646996737,-0.0049300003,-0.1914943755,-0.0878131837,-0.3597047925,0.294339478,0.6152353883,-0.3911423981,0.0004790114,-0.4250386953,-0.1540962905,0.1708229333,0.2825909257,-0.1810852587,0.418251425,-0.2191811651,-0.2109978348,0.4664492309,-0.4039042294,-0.4069173932,0.1619243324,-0.1690849811,-0.0231488161,0.0143251596,0.1950932592,-0.1637684554,0.0874652714,0.0317935087,0.3440898061,-0.0140823061,0.0791541114,-0.095202975,-0.2891000509,0.1494426131,0.1394836754,0.2537146509,0.2065376937,0.0017706892,0.057444334,0.0345365591,-0.1340152472,-0.0199387986,0.193029359,0.4896597266,0.0706251934,-0.3033567369,-0.0080354828,-0.5152108073,0.1154644713,0.0506806336,0.2688595951,-0.2443437576,0.0475811958,-0.2807430923,-0.1213214844,-0.2518719435,-0.0720805973,0.1413440108,0.362624079,0.0602164939,0.06581752,-0.2875800133,0.2650077939,0.0555366054,0.1482611448,-0.1362416446,0.4147193134,-0.1979314238,-0.0624308884,-0.0436080284,0.1352329403,0.2345956862,-0.1446532458,-0.1301756352,0.3645673692,-0.1508782059,-0.0773257464,0.0398339257,-0.1054766253,0.0736475289,-0.0363827124,0.2342951298,0.3620508015,0.1599349976,-0.0431439169,-0.1002390012,0.2904100418,0.161951825,0.0390076637,0.1538697779,-0.0371430255,0.0395891629,-0.0555200279,0.1082931161,-0.041275423,0.2116338313,0.0997428,0.4739992917,0.0227582809,-0.1307650357,-0.3711324632,0.2705956399,-0.134705767,0.1444204599,0.1506156623,-0.2888510227,0.0992454216,-0.1900030524,0.1472333819,0.6052938104,0.1164713353,0.0325042419,0.1258146018,-0.0870899335,-0.0704523772,0.2077787369,0.0883645713,0.2569261789,0.2160902619,0.0695267469,0.0620280765,-0.3969856501,-0.2876626253,0.0426680557,0.4161446691,-0.3931626678,0.059591908,-0.1244833395,-0.2166050375,-0.0977737382,-0.5352479219,-0.0786898881,-0.3155716062,-0.1555617154,0.1735693216,0.1367603093,0.1662632525,-0.1210425571,-0.1262292862,0.1058900803,-0.4859184921,-0.2068847418,-0.0484767519,-0.0174885709,0.0651339069,0.2804500163,-0.1032850295,0.2476293594,-0.1312977076,-0.1217798218,-0.1863985956,-0.0907368809,-0.0835310295,-0.1042266935,0.2084366828,0.2644519508,0.3542820513,0.0467904173,-0.1947212368,0.3619304001,-0.1828137487,0.0223103501,0.3297629058,-0.1260294914,-0.0134168044,-0.0603099614,-0.5458139777,-0.5458031297,-0.3574637175,-0.1193348393,0.128277272,0.1643208861,0.439052999,0.149543047,0.1939637363,-0.1065731868,0.0041575646,-0.0272339098,-0.1540678293,0.4063127637,-0.2877522409,-0.2439707518,0.1713533998,-0.0405313782,0.1901890039,-0.1609371156,-0.4939940274,-0.2556856275,-0.0731515959,0.038228672,-0.1987717301,-0.051362209,0.2149719,0.0962023139,-0.100954771,0.1519619524,-0.1301398873,-0.0496761128,0.1687473506,0.0203161482,-0.2199278027,0.5267876983,-0.2912866473,0.5844215155,0.1298206747,-0.0976593196,0.3287144303,-0.2017037868,0.435521692,-0.2107226998,-0.5223431587,-0.0787455291,0.2892427742,-0.1790048331,0.1800332218,0.0565606989,-0.2591916919,-0.2643011212,-0.0444018207,-0.3779843152,-0.0743462443,0.1186384708,-0.0179859605,0.0781653449,-0.059277568,0.0965221897,-0.1504167318,-0.1698943824,-0.1216434762,0.3782083094,0.0190812517,0.1019870639,-0.0288795531,0.0270932689,-0.3145630658,0.2298073471,-0.0520729199,0.071546942,-0.2187217623,0.0831159204,-0.0285105985,0.0293108635,0.3612942696,-0.0923879296,0.0112150013,0.4276471138,0.072345525,-0.485065043,-0.0946589336,-0.2761999965,0.1676738858,-0.001272671,0.2047504038,-0.1170452461,-0.1677084416,0.3938706517,0.2733164728,-0.0797540098,-0.3248061538,-0.4682425559,-0.3894968033,-0.4203393757,-0.0765345469,-0.1250401139,0.0756565779,0.1052619666,-0.1257726848,-0.0364609472,0.0204616189,0.1145620719,-0.0140965423,0.2421565503,0.0633371398,0.1261455715,0.1231676489,0.1575559825,0.3707537055,0.5872366428,-0.1024381369,-0.3017361462,-0.0854127482,0.1388856322,0.0639618933,0.2151001841,-0.2014811933,-0.3429370821,0.0943885595,0.1914982945,0.0416608639,-0.0360612422,0.5686709881,-0.1586035043,-0.3970258832,-0.2941640317,0.4028087556,-0.132741496,0.1745825559,0.2641446888,-0.1621219516,-0.193986252,0.1144844517,-0.2864453197,0.7768808007,-0.112184912,0.2149889618,0.6759076715,0.1097218022,0.4076379538,-0.0114143025,0.137216568,-0.311796248,-0.1835328639,-0.1531409025,-0.2621161938,0.0990275741,0.2857246995,-0.1570022702,0.4138245583,-0.2599776685,0.1028068587,-0.127301529,0.1603608727,-0.0852782652,0.0861265436,-0.3097000718,0.1809026301,0.0234545097,0.3886294961,-0.2298669666,0.1079559177,-0.151260823,-0.2345502526,0.0025140594,0.1233494729,-0.4199483693,0.2769888043,-0.0176260546,-0.1646072417,-0.1869401336,0.247104615,-0.0704445019,-0.0975803882,-0.1174648926,0.0991975963,-0.1049690396,-0.0900420919,-0.1319454163,0.0436581001,0.2979975641,-0.0978417844,-0.2941854596,0.0657169744,-0.1450731903,-0.279520601,0.2277370542,0.0566369779,0.2674985826,-0.3583118021,-0.2808163762,-0.0401816852,-0.0521988161,-0.0776873454,0.1419314146,0.0039220243,0.019712897,0.0714475289,-0.2435005307,-0.3559811711,-0.1612131447,0.2835132182,-0.2098854929,0.0607408918,0.5739116073,0.1368587017,-0.2089036703,-0.3013285995,0.1953066587,0.0391629376,-0.1753415763,0.2251401544,0.2365752906,0.2650004625,-0.0125541659,0.255510658,0.2237352729,-0.0629739016,-0.1974655241,-0.3072934151,-0.1711298227,0.0286055654,0.0978401601,0.2520788312,0.0883995891,0.0631099939,0.3833882511,-0.2658512294,-0.3185986578,0.0119840642,-0.016286904,-0.0612216629,0.0742244571,-0.0595226027,0.5541501045,0.0120948935,0.2147975415,-0.3351586461,-0.0658874363,-0.2823093235,-0.0192318764,0.1273972094,0.1241839752,-0.1153774783,0.0820352361,-0.257568568,-0.2182910442,-0.1288654506,-0.0556795076,0.238222301,-0.063959077,0.1493335813,0.0589087121,-0.0790562183,-0.1343450695,0.0602044202,-0.2575852871,0.0918631107,0.0379097909,0.1405527741,0.2113551497,-0.1355248392,-0.4185724258,-0.0494291224,0.0334853791,0.0262343343,0.2199243456,-0.1869715601,-0.0101036849,-0.0072820913,0.3588445187,0.1904233992,-0.2865005136,0.0540695675,0.2150326967,0.2182925791,-0.2886954546,0.0796895772,-0.0444070697,0.0619237162,0.1027720124,0.1913689822,-0.3624694645,0.0646900684,-0.196090281,0.0315474756,0.3888284564,-0.197690323,0.0516497865,0.3998382986,-0.1716947705,0.0677291825,0.2848182023,0.0445199311,0.1066730469,0.5403015018,-0.169164896,0.027433278,-0.2749078274,-0.115499109,0.0093393354,-0.3157514334,0.0527968816,0.1499107182,-0.0275917388,-0.0061320551,-0.1501961946,0.2725358903,-0.45939973,0.1213818267,-0.3764194548,0.3777875602,-0.1425220072,0.1625410467,-0.1579572111,0.0105228955,-0.012090886,0.0446171537,0.0195148569,-0.2471570969,0.2412227839,0.4903680086,-0.1194794476,-0.1413522214,0.3903640211,0.2625901401,-0.1352918744,-0.2359022647,0.3025464416,0.3034740388,0.1200008765,-0.0492572039,0.4870747924,0.4269422591,0.4090619385,0.0039016933,0.2013225257,-0.0100115081,-0.1889312714,-0.0214954503,0.0882004872,-0.0550458506,0.5318027735,0.1638848484,0.2567686439,-0.2468539625,0.185496375,0.0133697847,-0.0159491897,-0.1292088777,0.691208601,-0.2564795911,-0.1056898162,-0.2331102788,0.0544373505,-0.5289071202,0.2520841956,0.3535762727,0.0812641606,0.1096357256,0.0725061074,0.0778281316,0.0445873886,0.2782047987,0.342743665,0.30538252,-0.5661407709,0.0243918225,-0.6365788579,0.1711902022,-0.28920421,-0.0434493534,-0.1806510538,-0.015653681,-0.136431843,0.1296598315,0.1261664033,0.0536839589,-0.0534693822,-0.105646655,-0.2898376286,-0.0953034908,-0.0626657307,0.0257232394,0.015950067,-0.4950198531,0.0248199925,-0.413179934,0.0006352767,0.027058905,-0.133508116,0.0396981873,0.3611627519,0.4533376694,-0.026389746,0.5661256313,-0.0605582707,-0.1965637505,-0.1913531125,-0.2649549842,-0.1598000079,0.2294674814,0.1461038738,0.3851700723,-0.3064142168,-0.0087604532,-0.4270081222,0.2901292741,-0.0378126204,-0.067394942,-0.2009867877,0.0023696769,-0.1316203475,0.1983623058,0.1464611888,-0.0151976235,-0.0840773508,0.1883648485,-0.2635794878,-0.3884920478,0.46128425,-0.2634094059,-0.2436618209,0.0428768806,0.1699992418,-0.0931862071,0.1788178533,-0.4839810431,0.3811461627,0.438337028,0.1640297323,-0.2190377563,0.1378456354,0.0010955908,0.1383843273,0.0792405829,0.4123993814,0.0075344145,-0.3825779259,-0.0340820029,-0.1186552644]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/873","title":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error","comments":"I see the issue happening again today - \r\n\r\n[nltk_data] Downloading package stopwords to \/root\/nltk_data...\r\n[nltk_data] Package stopwords is already up-to-date!\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\nCan someone please take a look ?","body":"```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab","comment_length":108,"text":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab\nI see the issue happening again today - \r\n\r\n[nltk_data] Downloading package stopwords to \/root\/nltk_data...\r\n[nltk_data] Package stopwords is already up-to-date!\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-9-cd4bf8bea840> in <module>()\r\n 22 \r\n 23 \r\n---> 24 train = load_dataset('cnn_dailymail', '3.0.0', split='train')\r\n 25 validation = load_dataset('cnn_dailymail', '3.0.0', split='validation')\r\n 26 test = load_dataset('cnn_dailymail', '3.0.0', split='test')\r\n\r\n5 frames\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\nCan someone please take a look ?","embeddings":[-0.224293083,0.1419460326,-0.0225758813,0.2384324819,0.3819946051,0.1261650622,0.6168980598,0.2147594988,-0.0231766459,0.1334128678,-0.163756609,0.0582367294,-0.3759562671,-0.123077713,0.0085837077,0.114968501,-0.0964015499,0.1582960486,-0.2051095515,-0.1422860175,-0.3901512027,0.1864792407,-0.0768181756,0.0159925465,-0.2544403374,-0.2072729915,-0.1191695929,0.4782720804,-0.2281916291,-0.3415712714,0.291224122,-0.1937130839,0.2596129179,0.6446203589,-0.0001103697,0.1348124146,0.257455349,-0.0459846482,-0.5990345478,-0.6107162833,-0.1344918758,-0.1810776293,0.2224275321,-0.2857463658,-0.0545736887,0.1120457277,0.0625397936,0.019521933,0.3589360118,0.5676545501,0.2382051796,0.2500875592,-0.0214656368,-0.1423246413,-0.1394215375,0.3917793632,-0.08329615,0.5298015475,-0.3784783185,-0.0074536237,0.3009919226,0.2560061216,-0.3467578292,0.2481284142,0.3105844557,0.0931708962,0.2014932483,-0.318579495,0.0102561293,0.052064389,0.462816,-0.377645731,-0.2139322609,-0.0260961484,0.0874807909,-0.4519509375,0.3440143466,0.3034414649,-0.1303296834,0.0531827994,-0.146174401,-0.1438385248,-0.1814096421,0.4520489275,-0.1718828231,0.2167272717,-0.0570031144,0.0317016952,0.0401946232,0.0549006499,0.5230491161,-0.2012925595,0.1109838113,0.1689763218,-0.0259168874,0.0619172007,0.200421989,-0.1987337023,-0.1959023029,0.1629692465,-0.0814728886,-0.2725152671,-0.0308208447,0.0420363732,0.0562603734,0.1875984669,-0.0870422944,0.5960316658,0.2050020099,0.2587226927,-0.0126039628,-0.1288865656,-0.3524653614,-0.249772191,0.1027377322,0.0556450076,0.6368337274,0.0808701962,-0.262622416,-0.1706470996,-0.2875089049,0.0266700387,0.1824934632,0.3770372868,0.0879202038,-0.0015779705,0.1018526927,0.2980035841,-0.0611399859,0.3207466304,-0.2227891982,0.2121761739,-0.1194054484,-0.03893435,0.270580411,-0.2202292234,0.1217843741,-0.1732522994,-0.0137373274,-0.0063968683,0.1550607979,-0.1759008318,0.074225083,0.5078181028,-0.1087166741,0.2788080275,0.270821631,0.0265562665,-0.3172303438,0.1326729208,-0.1793326437,-0.4073957503,-0.0890357345,0.191819638,0.0433196686,-0.1540216804,0.0512391925,-0.2044744641,-0.1510185897,-0.1646996737,-0.0049300003,-0.1914943755,-0.0878131837,-0.3597047925,0.294339478,0.6152353883,-0.3911423981,0.0004790114,-0.4250386953,-0.1540962905,0.1708229333,0.2825909257,-0.1810852587,0.418251425,-0.2191811651,-0.2109978348,0.4664492309,-0.4039042294,-0.4069173932,0.1619243324,-0.1690849811,-0.0231488161,0.0143251596,0.1950932592,-0.1637684554,0.0874652714,0.0317935087,0.3440898061,-0.0140823061,0.0791541114,-0.095202975,-0.2891000509,0.1494426131,0.1394836754,0.2537146509,0.2065376937,0.0017706892,0.057444334,0.0345365591,-0.1340152472,-0.0199387986,0.193029359,0.4896597266,0.0706251934,-0.3033567369,-0.0080354828,-0.5152108073,0.1154644713,0.0506806336,0.2688595951,-0.2443437576,0.0475811958,-0.2807430923,-0.1213214844,-0.2518719435,-0.0720805973,0.1413440108,0.362624079,0.0602164939,0.06581752,-0.2875800133,0.2650077939,0.0555366054,0.1482611448,-0.1362416446,0.4147193134,-0.1979314238,-0.0624308884,-0.0436080284,0.1352329403,0.2345956862,-0.1446532458,-0.1301756352,0.3645673692,-0.1508782059,-0.0773257464,0.0398339257,-0.1054766253,0.0736475289,-0.0363827124,0.2342951298,0.3620508015,0.1599349976,-0.0431439169,-0.1002390012,0.2904100418,0.161951825,0.0390076637,0.1538697779,-0.0371430255,0.0395891629,-0.0555200279,0.1082931161,-0.041275423,0.2116338313,0.0997428,0.4739992917,0.0227582809,-0.1307650357,-0.3711324632,0.2705956399,-0.134705767,0.1444204599,0.1506156623,-0.2888510227,0.0992454216,-0.1900030524,0.1472333819,0.6052938104,0.1164713353,0.0325042419,0.1258146018,-0.0870899335,-0.0704523772,0.2077787369,0.0883645713,0.2569261789,0.2160902619,0.0695267469,0.0620280765,-0.3969856501,-0.2876626253,0.0426680557,0.4161446691,-0.3931626678,0.059591908,-0.1244833395,-0.2166050375,-0.0977737382,-0.5352479219,-0.0786898881,-0.3155716062,-0.1555617154,0.1735693216,0.1367603093,0.1662632525,-0.1210425571,-0.1262292862,0.1058900803,-0.4859184921,-0.2068847418,-0.0484767519,-0.0174885709,0.0651339069,0.2804500163,-0.1032850295,0.2476293594,-0.1312977076,-0.1217798218,-0.1863985956,-0.0907368809,-0.0835310295,-0.1042266935,0.2084366828,0.2644519508,0.3542820513,0.0467904173,-0.1947212368,0.3619304001,-0.1828137487,0.0223103501,0.3297629058,-0.1260294914,-0.0134168044,-0.0603099614,-0.5458139777,-0.5458031297,-0.3574637175,-0.1193348393,0.128277272,0.1643208861,0.439052999,0.149543047,0.1939637363,-0.1065731868,0.0041575646,-0.0272339098,-0.1540678293,0.4063127637,-0.2877522409,-0.2439707518,0.1713533998,-0.0405313782,0.1901890039,-0.1609371156,-0.4939940274,-0.2556856275,-0.0731515959,0.038228672,-0.1987717301,-0.051362209,0.2149719,0.0962023139,-0.100954771,0.1519619524,-0.1301398873,-0.0496761128,0.1687473506,0.0203161482,-0.2199278027,0.5267876983,-0.2912866473,0.5844215155,0.1298206747,-0.0976593196,0.3287144303,-0.2017037868,0.435521692,-0.2107226998,-0.5223431587,-0.0787455291,0.2892427742,-0.1790048331,0.1800332218,0.0565606989,-0.2591916919,-0.2643011212,-0.0444018207,-0.3779843152,-0.0743462443,0.1186384708,-0.0179859605,0.0781653449,-0.059277568,0.0965221897,-0.1504167318,-0.1698943824,-0.1216434762,0.3782083094,0.0190812517,0.1019870639,-0.0288795531,0.0270932689,-0.3145630658,0.2298073471,-0.0520729199,0.071546942,-0.2187217623,0.0831159204,-0.0285105985,0.0293108635,0.3612942696,-0.0923879296,0.0112150013,0.4276471138,0.072345525,-0.485065043,-0.0946589336,-0.2761999965,0.1676738858,-0.001272671,0.2047504038,-0.1170452461,-0.1677084416,0.3938706517,0.2733164728,-0.0797540098,-0.3248061538,-0.4682425559,-0.3894968033,-0.4203393757,-0.0765345469,-0.1250401139,0.0756565779,0.1052619666,-0.1257726848,-0.0364609472,0.0204616189,0.1145620719,-0.0140965423,0.2421565503,0.0633371398,0.1261455715,0.1231676489,0.1575559825,0.3707537055,0.5872366428,-0.1024381369,-0.3017361462,-0.0854127482,0.1388856322,0.0639618933,0.2151001841,-0.2014811933,-0.3429370821,0.0943885595,0.1914982945,0.0416608639,-0.0360612422,0.5686709881,-0.1586035043,-0.3970258832,-0.2941640317,0.4028087556,-0.132741496,0.1745825559,0.2641446888,-0.1621219516,-0.193986252,0.1144844517,-0.2864453197,0.7768808007,-0.112184912,0.2149889618,0.6759076715,0.1097218022,0.4076379538,-0.0114143025,0.137216568,-0.311796248,-0.1835328639,-0.1531409025,-0.2621161938,0.0990275741,0.2857246995,-0.1570022702,0.4138245583,-0.2599776685,0.1028068587,-0.127301529,0.1603608727,-0.0852782652,0.0861265436,-0.3097000718,0.1809026301,0.0234545097,0.3886294961,-0.2298669666,0.1079559177,-0.151260823,-0.2345502526,0.0025140594,0.1233494729,-0.4199483693,0.2769888043,-0.0176260546,-0.1646072417,-0.1869401336,0.247104615,-0.0704445019,-0.0975803882,-0.1174648926,0.0991975963,-0.1049690396,-0.0900420919,-0.1319454163,0.0436581001,0.2979975641,-0.0978417844,-0.2941854596,0.0657169744,-0.1450731903,-0.279520601,0.2277370542,0.0566369779,0.2674985826,-0.3583118021,-0.2808163762,-0.0401816852,-0.0521988161,-0.0776873454,0.1419314146,0.0039220243,0.019712897,0.0714475289,-0.2435005307,-0.3559811711,-0.1612131447,0.2835132182,-0.2098854929,0.0607408918,0.5739116073,0.1368587017,-0.2089036703,-0.3013285995,0.1953066587,0.0391629376,-0.1753415763,0.2251401544,0.2365752906,0.2650004625,-0.0125541659,0.255510658,0.2237352729,-0.0629739016,-0.1974655241,-0.3072934151,-0.1711298227,0.0286055654,0.0978401601,0.2520788312,0.0883995891,0.0631099939,0.3833882511,-0.2658512294,-0.3185986578,0.0119840642,-0.016286904,-0.0612216629,0.0742244571,-0.0595226027,0.5541501045,0.0120948935,0.2147975415,-0.3351586461,-0.0658874363,-0.2823093235,-0.0192318764,0.1273972094,0.1241839752,-0.1153774783,0.0820352361,-0.257568568,-0.2182910442,-0.1288654506,-0.0556795076,0.238222301,-0.063959077,0.1493335813,0.0589087121,-0.0790562183,-0.1343450695,0.0602044202,-0.2575852871,0.0918631107,0.0379097909,0.1405527741,0.2113551497,-0.1355248392,-0.4185724258,-0.0494291224,0.0334853791,0.0262343343,0.2199243456,-0.1869715601,-0.0101036849,-0.0072820913,0.3588445187,0.1904233992,-0.2865005136,0.0540695675,0.2150326967,0.2182925791,-0.2886954546,0.0796895772,-0.0444070697,0.0619237162,0.1027720124,0.1913689822,-0.3624694645,0.0646900684,-0.196090281,0.0315474756,0.3888284564,-0.197690323,0.0516497865,0.3998382986,-0.1716947705,0.0677291825,0.2848182023,0.0445199311,0.1066730469,0.5403015018,-0.169164896,0.027433278,-0.2749078274,-0.115499109,0.0093393354,-0.3157514334,0.0527968816,0.1499107182,-0.0275917388,-0.0061320551,-0.1501961946,0.2725358903,-0.45939973,0.1213818267,-0.3764194548,0.3777875602,-0.1425220072,0.1625410467,-0.1579572111,0.0105228955,-0.012090886,0.0446171537,0.0195148569,-0.2471570969,0.2412227839,0.4903680086,-0.1194794476,-0.1413522214,0.3903640211,0.2625901401,-0.1352918744,-0.2359022647,0.3025464416,0.3034740388,0.1200008765,-0.0492572039,0.4870747924,0.4269422591,0.4090619385,0.0039016933,0.2013225257,-0.0100115081,-0.1889312714,-0.0214954503,0.0882004872,-0.0550458506,0.5318027735,0.1638848484,0.2567686439,-0.2468539625,0.185496375,0.0133697847,-0.0159491897,-0.1292088777,0.691208601,-0.2564795911,-0.1056898162,-0.2331102788,0.0544373505,-0.5289071202,0.2520841956,0.3535762727,0.0812641606,0.1096357256,0.0725061074,0.0778281316,0.0445873886,0.2782047987,0.342743665,0.30538252,-0.5661407709,0.0243918225,-0.6365788579,0.1711902022,-0.28920421,-0.0434493534,-0.1806510538,-0.015653681,-0.136431843,0.1296598315,0.1261664033,0.0536839589,-0.0534693822,-0.105646655,-0.2898376286,-0.0953034908,-0.0626657307,0.0257232394,0.015950067,-0.4950198531,0.0248199925,-0.413179934,0.0006352767,0.027058905,-0.133508116,0.0396981873,0.3611627519,0.4533376694,-0.026389746,0.5661256313,-0.0605582707,-0.1965637505,-0.1913531125,-0.2649549842,-0.1598000079,0.2294674814,0.1461038738,0.3851700723,-0.3064142168,-0.0087604532,-0.4270081222,0.2901292741,-0.0378126204,-0.067394942,-0.2009867877,0.0023696769,-0.1316203475,0.1983623058,0.1464611888,-0.0151976235,-0.0840773508,0.1883648485,-0.2635794878,-0.3884920478,0.46128425,-0.2634094059,-0.2436618209,0.0428768806,0.1699992418,-0.0931862071,0.1788178533,-0.4839810431,0.3811461627,0.438337028,0.1640297323,-0.2190377563,0.1378456354,0.0010955908,0.1383843273,0.0792405829,0.4123993814,0.0075344145,-0.3825779259,-0.0340820029,-0.1186552644]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/873","title":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error","comments":"Sometimes happens. Try in a while","body":"```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab","comment_length":6,"text":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab\nSometimes happens. Try in a while","embeddings":[-0.224293083,0.1419460326,-0.0225758813,0.2384324819,0.3819946051,0.1261650622,0.6168980598,0.2147594988,-0.0231766459,0.1334128678,-0.163756609,0.0582367294,-0.3759562671,-0.123077713,0.0085837077,0.114968501,-0.0964015499,0.1582960486,-0.2051095515,-0.1422860175,-0.3901512027,0.1864792407,-0.0768181756,0.0159925465,-0.2544403374,-0.2072729915,-0.1191695929,0.4782720804,-0.2281916291,-0.3415712714,0.291224122,-0.1937130839,0.2596129179,0.6446203589,-0.0001103697,0.1348124146,0.257455349,-0.0459846482,-0.5990345478,-0.6107162833,-0.1344918758,-0.1810776293,0.2224275321,-0.2857463658,-0.0545736887,0.1120457277,0.0625397936,0.019521933,0.3589360118,0.5676545501,0.2382051796,0.2500875592,-0.0214656368,-0.1423246413,-0.1394215375,0.3917793632,-0.08329615,0.5298015475,-0.3784783185,-0.0074536237,0.3009919226,0.2560061216,-0.3467578292,0.2481284142,0.3105844557,0.0931708962,0.2014932483,-0.318579495,0.0102561293,0.052064389,0.462816,-0.377645731,-0.2139322609,-0.0260961484,0.0874807909,-0.4519509375,0.3440143466,0.3034414649,-0.1303296834,0.0531827994,-0.146174401,-0.1438385248,-0.1814096421,0.4520489275,-0.1718828231,0.2167272717,-0.0570031144,0.0317016952,0.0401946232,0.0549006499,0.5230491161,-0.2012925595,0.1109838113,0.1689763218,-0.0259168874,0.0619172007,0.200421989,-0.1987337023,-0.1959023029,0.1629692465,-0.0814728886,-0.2725152671,-0.0308208447,0.0420363732,0.0562603734,0.1875984669,-0.0870422944,0.5960316658,0.2050020099,0.2587226927,-0.0126039628,-0.1288865656,-0.3524653614,-0.249772191,0.1027377322,0.0556450076,0.6368337274,0.0808701962,-0.262622416,-0.1706470996,-0.2875089049,0.0266700387,0.1824934632,0.3770372868,0.0879202038,-0.0015779705,0.1018526927,0.2980035841,-0.0611399859,0.3207466304,-0.2227891982,0.2121761739,-0.1194054484,-0.03893435,0.270580411,-0.2202292234,0.1217843741,-0.1732522994,-0.0137373274,-0.0063968683,0.1550607979,-0.1759008318,0.074225083,0.5078181028,-0.1087166741,0.2788080275,0.270821631,0.0265562665,-0.3172303438,0.1326729208,-0.1793326437,-0.4073957503,-0.0890357345,0.191819638,0.0433196686,-0.1540216804,0.0512391925,-0.2044744641,-0.1510185897,-0.1646996737,-0.0049300003,-0.1914943755,-0.0878131837,-0.3597047925,0.294339478,0.6152353883,-0.3911423981,0.0004790114,-0.4250386953,-0.1540962905,0.1708229333,0.2825909257,-0.1810852587,0.418251425,-0.2191811651,-0.2109978348,0.4664492309,-0.4039042294,-0.4069173932,0.1619243324,-0.1690849811,-0.0231488161,0.0143251596,0.1950932592,-0.1637684554,0.0874652714,0.0317935087,0.3440898061,-0.0140823061,0.0791541114,-0.095202975,-0.2891000509,0.1494426131,0.1394836754,0.2537146509,0.2065376937,0.0017706892,0.057444334,0.0345365591,-0.1340152472,-0.0199387986,0.193029359,0.4896597266,0.0706251934,-0.3033567369,-0.0080354828,-0.5152108073,0.1154644713,0.0506806336,0.2688595951,-0.2443437576,0.0475811958,-0.2807430923,-0.1213214844,-0.2518719435,-0.0720805973,0.1413440108,0.362624079,0.0602164939,0.06581752,-0.2875800133,0.2650077939,0.0555366054,0.1482611448,-0.1362416446,0.4147193134,-0.1979314238,-0.0624308884,-0.0436080284,0.1352329403,0.2345956862,-0.1446532458,-0.1301756352,0.3645673692,-0.1508782059,-0.0773257464,0.0398339257,-0.1054766253,0.0736475289,-0.0363827124,0.2342951298,0.3620508015,0.1599349976,-0.0431439169,-0.1002390012,0.2904100418,0.161951825,0.0390076637,0.1538697779,-0.0371430255,0.0395891629,-0.0555200279,0.1082931161,-0.041275423,0.2116338313,0.0997428,0.4739992917,0.0227582809,-0.1307650357,-0.3711324632,0.2705956399,-0.134705767,0.1444204599,0.1506156623,-0.2888510227,0.0992454216,-0.1900030524,0.1472333819,0.6052938104,0.1164713353,0.0325042419,0.1258146018,-0.0870899335,-0.0704523772,0.2077787369,0.0883645713,0.2569261789,0.2160902619,0.0695267469,0.0620280765,-0.3969856501,-0.2876626253,0.0426680557,0.4161446691,-0.3931626678,0.059591908,-0.1244833395,-0.2166050375,-0.0977737382,-0.5352479219,-0.0786898881,-0.3155716062,-0.1555617154,0.1735693216,0.1367603093,0.1662632525,-0.1210425571,-0.1262292862,0.1058900803,-0.4859184921,-0.2068847418,-0.0484767519,-0.0174885709,0.0651339069,0.2804500163,-0.1032850295,0.2476293594,-0.1312977076,-0.1217798218,-0.1863985956,-0.0907368809,-0.0835310295,-0.1042266935,0.2084366828,0.2644519508,0.3542820513,0.0467904173,-0.1947212368,0.3619304001,-0.1828137487,0.0223103501,0.3297629058,-0.1260294914,-0.0134168044,-0.0603099614,-0.5458139777,-0.5458031297,-0.3574637175,-0.1193348393,0.128277272,0.1643208861,0.439052999,0.149543047,0.1939637363,-0.1065731868,0.0041575646,-0.0272339098,-0.1540678293,0.4063127637,-0.2877522409,-0.2439707518,0.1713533998,-0.0405313782,0.1901890039,-0.1609371156,-0.4939940274,-0.2556856275,-0.0731515959,0.038228672,-0.1987717301,-0.051362209,0.2149719,0.0962023139,-0.100954771,0.1519619524,-0.1301398873,-0.0496761128,0.1687473506,0.0203161482,-0.2199278027,0.5267876983,-0.2912866473,0.5844215155,0.1298206747,-0.0976593196,0.3287144303,-0.2017037868,0.435521692,-0.2107226998,-0.5223431587,-0.0787455291,0.2892427742,-0.1790048331,0.1800332218,0.0565606989,-0.2591916919,-0.2643011212,-0.0444018207,-0.3779843152,-0.0743462443,0.1186384708,-0.0179859605,0.0781653449,-0.059277568,0.0965221897,-0.1504167318,-0.1698943824,-0.1216434762,0.3782083094,0.0190812517,0.1019870639,-0.0288795531,0.0270932689,-0.3145630658,0.2298073471,-0.0520729199,0.071546942,-0.2187217623,0.0831159204,-0.0285105985,0.0293108635,0.3612942696,-0.0923879296,0.0112150013,0.4276471138,0.072345525,-0.485065043,-0.0946589336,-0.2761999965,0.1676738858,-0.001272671,0.2047504038,-0.1170452461,-0.1677084416,0.3938706517,0.2733164728,-0.0797540098,-0.3248061538,-0.4682425559,-0.3894968033,-0.4203393757,-0.0765345469,-0.1250401139,0.0756565779,0.1052619666,-0.1257726848,-0.0364609472,0.0204616189,0.1145620719,-0.0140965423,0.2421565503,0.0633371398,0.1261455715,0.1231676489,0.1575559825,0.3707537055,0.5872366428,-0.1024381369,-0.3017361462,-0.0854127482,0.1388856322,0.0639618933,0.2151001841,-0.2014811933,-0.3429370821,0.0943885595,0.1914982945,0.0416608639,-0.0360612422,0.5686709881,-0.1586035043,-0.3970258832,-0.2941640317,0.4028087556,-0.132741496,0.1745825559,0.2641446888,-0.1621219516,-0.193986252,0.1144844517,-0.2864453197,0.7768808007,-0.112184912,0.2149889618,0.6759076715,0.1097218022,0.4076379538,-0.0114143025,0.137216568,-0.311796248,-0.1835328639,-0.1531409025,-0.2621161938,0.0990275741,0.2857246995,-0.1570022702,0.4138245583,-0.2599776685,0.1028068587,-0.127301529,0.1603608727,-0.0852782652,0.0861265436,-0.3097000718,0.1809026301,0.0234545097,0.3886294961,-0.2298669666,0.1079559177,-0.151260823,-0.2345502526,0.0025140594,0.1233494729,-0.4199483693,0.2769888043,-0.0176260546,-0.1646072417,-0.1869401336,0.247104615,-0.0704445019,-0.0975803882,-0.1174648926,0.0991975963,-0.1049690396,-0.0900420919,-0.1319454163,0.0436581001,0.2979975641,-0.0978417844,-0.2941854596,0.0657169744,-0.1450731903,-0.279520601,0.2277370542,0.0566369779,0.2674985826,-0.3583118021,-0.2808163762,-0.0401816852,-0.0521988161,-0.0776873454,0.1419314146,0.0039220243,0.019712897,0.0714475289,-0.2435005307,-0.3559811711,-0.1612131447,0.2835132182,-0.2098854929,0.0607408918,0.5739116073,0.1368587017,-0.2089036703,-0.3013285995,0.1953066587,0.0391629376,-0.1753415763,0.2251401544,0.2365752906,0.2650004625,-0.0125541659,0.255510658,0.2237352729,-0.0629739016,-0.1974655241,-0.3072934151,-0.1711298227,0.0286055654,0.0978401601,0.2520788312,0.0883995891,0.0631099939,0.3833882511,-0.2658512294,-0.3185986578,0.0119840642,-0.016286904,-0.0612216629,0.0742244571,-0.0595226027,0.5541501045,0.0120948935,0.2147975415,-0.3351586461,-0.0658874363,-0.2823093235,-0.0192318764,0.1273972094,0.1241839752,-0.1153774783,0.0820352361,-0.257568568,-0.2182910442,-0.1288654506,-0.0556795076,0.238222301,-0.063959077,0.1493335813,0.0589087121,-0.0790562183,-0.1343450695,0.0602044202,-0.2575852871,0.0918631107,0.0379097909,0.1405527741,0.2113551497,-0.1355248392,-0.4185724258,-0.0494291224,0.0334853791,0.0262343343,0.2199243456,-0.1869715601,-0.0101036849,-0.0072820913,0.3588445187,0.1904233992,-0.2865005136,0.0540695675,0.2150326967,0.2182925791,-0.2886954546,0.0796895772,-0.0444070697,0.0619237162,0.1027720124,0.1913689822,-0.3624694645,0.0646900684,-0.196090281,0.0315474756,0.3888284564,-0.197690323,0.0516497865,0.3998382986,-0.1716947705,0.0677291825,0.2848182023,0.0445199311,0.1066730469,0.5403015018,-0.169164896,0.027433278,-0.2749078274,-0.115499109,0.0093393354,-0.3157514334,0.0527968816,0.1499107182,-0.0275917388,-0.0061320551,-0.1501961946,0.2725358903,-0.45939973,0.1213818267,-0.3764194548,0.3777875602,-0.1425220072,0.1625410467,-0.1579572111,0.0105228955,-0.012090886,0.0446171537,0.0195148569,-0.2471570969,0.2412227839,0.4903680086,-0.1194794476,-0.1413522214,0.3903640211,0.2625901401,-0.1352918744,-0.2359022647,0.3025464416,0.3034740388,0.1200008765,-0.0492572039,0.4870747924,0.4269422591,0.4090619385,0.0039016933,0.2013225257,-0.0100115081,-0.1889312714,-0.0214954503,0.0882004872,-0.0550458506,0.5318027735,0.1638848484,0.2567686439,-0.2468539625,0.185496375,0.0133697847,-0.0159491897,-0.1292088777,0.691208601,-0.2564795911,-0.1056898162,-0.2331102788,0.0544373505,-0.5289071202,0.2520841956,0.3535762727,0.0812641606,0.1096357256,0.0725061074,0.0778281316,0.0445873886,0.2782047987,0.342743665,0.30538252,-0.5661407709,0.0243918225,-0.6365788579,0.1711902022,-0.28920421,-0.0434493534,-0.1806510538,-0.015653681,-0.136431843,0.1296598315,0.1261664033,0.0536839589,-0.0534693822,-0.105646655,-0.2898376286,-0.0953034908,-0.0626657307,0.0257232394,0.015950067,-0.4950198531,0.0248199925,-0.413179934,0.0006352767,0.027058905,-0.133508116,0.0396981873,0.3611627519,0.4533376694,-0.026389746,0.5661256313,-0.0605582707,-0.1965637505,-0.1913531125,-0.2649549842,-0.1598000079,0.2294674814,0.1461038738,0.3851700723,-0.3064142168,-0.0087604532,-0.4270081222,0.2901292741,-0.0378126204,-0.067394942,-0.2009867877,0.0023696769,-0.1316203475,0.1983623058,0.1464611888,-0.0151976235,-0.0840773508,0.1883648485,-0.2635794878,-0.3884920478,0.46128425,-0.2634094059,-0.2436618209,0.0428768806,0.1699992418,-0.0931862071,0.1788178533,-0.4839810431,0.3811461627,0.438337028,0.1640297323,-0.2190377563,0.1378456354,0.0010955908,0.1383843273,0.0792405829,0.4123993814,0.0075344145,-0.3825779259,-0.0340820029,-0.1186552644]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/873","title":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error","comments":"It is working now, thank you. ","body":"```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab","comment_length":6,"text":"load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('cnn_dailymail', '3.0.0')\r\n```\r\nStack trace:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nNotADirectoryError Traceback (most recent call last)\r\n\r\n<ipython-input-6-2e06a8332652> in <module>()\r\n 1 from datasets import load_dataset\r\n----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n5 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 608 download_config=download_config,\r\n 609 download_mode=download_mode,\r\n--> 610 ignore_verifications=ignore_verifications,\r\n 611 )\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 513 if not downloaded_from_gcs:\r\n 514 self._download_and_prepare(\r\n--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 516 )\r\n 517 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 568 split_dict = SplitDict(dataset_name=self.name)\r\n 569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 571 \r\n 572 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n```\r\nI have ran the code on Google Colab\nIt is working now, thank you. ","embeddings":[-0.224293083,0.1419460326,-0.0225758813,0.2384324819,0.3819946051,0.1261650622,0.6168980598,0.2147594988,-0.0231766459,0.1334128678,-0.163756609,0.0582367294,-0.3759562671,-0.123077713,0.0085837077,0.114968501,-0.0964015499,0.1582960486,-0.2051095515,-0.1422860175,-0.3901512027,0.1864792407,-0.0768181756,0.0159925465,-0.2544403374,-0.2072729915,-0.1191695929,0.4782720804,-0.2281916291,-0.3415712714,0.291224122,-0.1937130839,0.2596129179,0.6446203589,-0.0001103697,0.1348124146,0.257455349,-0.0459846482,-0.5990345478,-0.6107162833,-0.1344918758,-0.1810776293,0.2224275321,-0.2857463658,-0.0545736887,0.1120457277,0.0625397936,0.019521933,0.3589360118,0.5676545501,0.2382051796,0.2500875592,-0.0214656368,-0.1423246413,-0.1394215375,0.3917793632,-0.08329615,0.5298015475,-0.3784783185,-0.0074536237,0.3009919226,0.2560061216,-0.3467578292,0.2481284142,0.3105844557,0.0931708962,0.2014932483,-0.318579495,0.0102561293,0.052064389,0.462816,-0.377645731,-0.2139322609,-0.0260961484,0.0874807909,-0.4519509375,0.3440143466,0.3034414649,-0.1303296834,0.0531827994,-0.146174401,-0.1438385248,-0.1814096421,0.4520489275,-0.1718828231,0.2167272717,-0.0570031144,0.0317016952,0.0401946232,0.0549006499,0.5230491161,-0.2012925595,0.1109838113,0.1689763218,-0.0259168874,0.0619172007,0.200421989,-0.1987337023,-0.1959023029,0.1629692465,-0.0814728886,-0.2725152671,-0.0308208447,0.0420363732,0.0562603734,0.1875984669,-0.0870422944,0.5960316658,0.2050020099,0.2587226927,-0.0126039628,-0.1288865656,-0.3524653614,-0.249772191,0.1027377322,0.0556450076,0.6368337274,0.0808701962,-0.262622416,-0.1706470996,-0.2875089049,0.0266700387,0.1824934632,0.3770372868,0.0879202038,-0.0015779705,0.1018526927,0.2980035841,-0.0611399859,0.3207466304,-0.2227891982,0.2121761739,-0.1194054484,-0.03893435,0.270580411,-0.2202292234,0.1217843741,-0.1732522994,-0.0137373274,-0.0063968683,0.1550607979,-0.1759008318,0.074225083,0.5078181028,-0.1087166741,0.2788080275,0.270821631,0.0265562665,-0.3172303438,0.1326729208,-0.1793326437,-0.4073957503,-0.0890357345,0.191819638,0.0433196686,-0.1540216804,0.0512391925,-0.2044744641,-0.1510185897,-0.1646996737,-0.0049300003,-0.1914943755,-0.0878131837,-0.3597047925,0.294339478,0.6152353883,-0.3911423981,0.0004790114,-0.4250386953,-0.1540962905,0.1708229333,0.2825909257,-0.1810852587,0.418251425,-0.2191811651,-0.2109978348,0.4664492309,-0.4039042294,-0.4069173932,0.1619243324,-0.1690849811,-0.0231488161,0.0143251596,0.1950932592,-0.1637684554,0.0874652714,0.0317935087,0.3440898061,-0.0140823061,0.0791541114,-0.095202975,-0.2891000509,0.1494426131,0.1394836754,0.2537146509,0.2065376937,0.0017706892,0.057444334,0.0345365591,-0.1340152472,-0.0199387986,0.193029359,0.4896597266,0.0706251934,-0.3033567369,-0.0080354828,-0.5152108073,0.1154644713,0.0506806336,0.2688595951,-0.2443437576,0.0475811958,-0.2807430923,-0.1213214844,-0.2518719435,-0.0720805973,0.1413440108,0.362624079,0.0602164939,0.06581752,-0.2875800133,0.2650077939,0.0555366054,0.1482611448,-0.1362416446,0.4147193134,-0.1979314238,-0.0624308884,-0.0436080284,0.1352329403,0.2345956862,-0.1446532458,-0.1301756352,0.3645673692,-0.1508782059,-0.0773257464,0.0398339257,-0.1054766253,0.0736475289,-0.0363827124,0.2342951298,0.3620508015,0.1599349976,-0.0431439169,-0.1002390012,0.2904100418,0.161951825,0.0390076637,0.1538697779,-0.0371430255,0.0395891629,-0.0555200279,0.1082931161,-0.041275423,0.2116338313,0.0997428,0.4739992917,0.0227582809,-0.1307650357,-0.3711324632,0.2705956399,-0.134705767,0.1444204599,0.1506156623,-0.2888510227,0.0992454216,-0.1900030524,0.1472333819,0.6052938104,0.1164713353,0.0325042419,0.1258146018,-0.0870899335,-0.0704523772,0.2077787369,0.0883645713,0.2569261789,0.2160902619,0.0695267469,0.0620280765,-0.3969856501,-0.2876626253,0.0426680557,0.4161446691,-0.3931626678,0.059591908,-0.1244833395,-0.2166050375,-0.0977737382,-0.5352479219,-0.0786898881,-0.3155716062,-0.1555617154,0.1735693216,0.1367603093,0.1662632525,-0.1210425571,-0.1262292862,0.1058900803,-0.4859184921,-0.2068847418,-0.0484767519,-0.0174885709,0.0651339069,0.2804500163,-0.1032850295,0.2476293594,-0.1312977076,-0.1217798218,-0.1863985956,-0.0907368809,-0.0835310295,-0.1042266935,0.2084366828,0.2644519508,0.3542820513,0.0467904173,-0.1947212368,0.3619304001,-0.1828137487,0.0223103501,0.3297629058,-0.1260294914,-0.0134168044,-0.0603099614,-0.5458139777,-0.5458031297,-0.3574637175,-0.1193348393,0.128277272,0.1643208861,0.439052999,0.149543047,0.1939637363,-0.1065731868,0.0041575646,-0.0272339098,-0.1540678293,0.4063127637,-0.2877522409,-0.2439707518,0.1713533998,-0.0405313782,0.1901890039,-0.1609371156,-0.4939940274,-0.2556856275,-0.0731515959,0.038228672,-0.1987717301,-0.051362209,0.2149719,0.0962023139,-0.100954771,0.1519619524,-0.1301398873,-0.0496761128,0.1687473506,0.0203161482,-0.2199278027,0.5267876983,-0.2912866473,0.5844215155,0.1298206747,-0.0976593196,0.3287144303,-0.2017037868,0.435521692,-0.2107226998,-0.5223431587,-0.0787455291,0.2892427742,-0.1790048331,0.1800332218,0.0565606989,-0.2591916919,-0.2643011212,-0.0444018207,-0.3779843152,-0.0743462443,0.1186384708,-0.0179859605,0.0781653449,-0.059277568,0.0965221897,-0.1504167318,-0.1698943824,-0.1216434762,0.3782083094,0.0190812517,0.1019870639,-0.0288795531,0.0270932689,-0.3145630658,0.2298073471,-0.0520729199,0.071546942,-0.2187217623,0.0831159204,-0.0285105985,0.0293108635,0.3612942696,-0.0923879296,0.0112150013,0.4276471138,0.072345525,-0.485065043,-0.0946589336,-0.2761999965,0.1676738858,-0.001272671,0.2047504038,-0.1170452461,-0.1677084416,0.3938706517,0.2733164728,-0.0797540098,-0.3248061538,-0.4682425559,-0.3894968033,-0.4203393757,-0.0765345469,-0.1250401139,0.0756565779,0.1052619666,-0.1257726848,-0.0364609472,0.0204616189,0.1145620719,-0.0140965423,0.2421565503,0.0633371398,0.1261455715,0.1231676489,0.1575559825,0.3707537055,0.5872366428,-0.1024381369,-0.3017361462,-0.0854127482,0.1388856322,0.0639618933,0.2151001841,-0.2014811933,-0.3429370821,0.0943885595,0.1914982945,0.0416608639,-0.0360612422,0.5686709881,-0.1586035043,-0.3970258832,-0.2941640317,0.4028087556,-0.132741496,0.1745825559,0.2641446888,-0.1621219516,-0.193986252,0.1144844517,-0.2864453197,0.7768808007,-0.112184912,0.2149889618,0.6759076715,0.1097218022,0.4076379538,-0.0114143025,0.137216568,-0.311796248,-0.1835328639,-0.1531409025,-0.2621161938,0.0990275741,0.2857246995,-0.1570022702,0.4138245583,-0.2599776685,0.1028068587,-0.127301529,0.1603608727,-0.0852782652,0.0861265436,-0.3097000718,0.1809026301,0.0234545097,0.3886294961,-0.2298669666,0.1079559177,-0.151260823,-0.2345502526,0.0025140594,0.1233494729,-0.4199483693,0.2769888043,-0.0176260546,-0.1646072417,-0.1869401336,0.247104615,-0.0704445019,-0.0975803882,-0.1174648926,0.0991975963,-0.1049690396,-0.0900420919,-0.1319454163,0.0436581001,0.2979975641,-0.0978417844,-0.2941854596,0.0657169744,-0.1450731903,-0.279520601,0.2277370542,0.0566369779,0.2674985826,-0.3583118021,-0.2808163762,-0.0401816852,-0.0521988161,-0.0776873454,0.1419314146,0.0039220243,0.019712897,0.0714475289,-0.2435005307,-0.3559811711,-0.1612131447,0.2835132182,-0.2098854929,0.0607408918,0.5739116073,0.1368587017,-0.2089036703,-0.3013285995,0.1953066587,0.0391629376,-0.1753415763,0.2251401544,0.2365752906,0.2650004625,-0.0125541659,0.255510658,0.2237352729,-0.0629739016,-0.1974655241,-0.3072934151,-0.1711298227,0.0286055654,0.0978401601,0.2520788312,0.0883995891,0.0631099939,0.3833882511,-0.2658512294,-0.3185986578,0.0119840642,-0.016286904,-0.0612216629,0.0742244571,-0.0595226027,0.5541501045,0.0120948935,0.2147975415,-0.3351586461,-0.0658874363,-0.2823093235,-0.0192318764,0.1273972094,0.1241839752,-0.1153774783,0.0820352361,-0.257568568,-0.2182910442,-0.1288654506,-0.0556795076,0.238222301,-0.063959077,0.1493335813,0.0589087121,-0.0790562183,-0.1343450695,0.0602044202,-0.2575852871,0.0918631107,0.0379097909,0.1405527741,0.2113551497,-0.1355248392,-0.4185724258,-0.0494291224,0.0334853791,0.0262343343,0.2199243456,-0.1869715601,-0.0101036849,-0.0072820913,0.3588445187,0.1904233992,-0.2865005136,0.0540695675,0.2150326967,0.2182925791,-0.2886954546,0.0796895772,-0.0444070697,0.0619237162,0.1027720124,0.1913689822,-0.3624694645,0.0646900684,-0.196090281,0.0315474756,0.3888284564,-0.197690323,0.0516497865,0.3998382986,-0.1716947705,0.0677291825,0.2848182023,0.0445199311,0.1066730469,0.5403015018,-0.169164896,0.027433278,-0.2749078274,-0.115499109,0.0093393354,-0.3157514334,0.0527968816,0.1499107182,-0.0275917388,-0.0061320551,-0.1501961946,0.2725358903,-0.45939973,0.1213818267,-0.3764194548,0.3777875602,-0.1425220072,0.1625410467,-0.1579572111,0.0105228955,-0.012090886,0.0446171537,0.0195148569,-0.2471570969,0.2412227839,0.4903680086,-0.1194794476,-0.1413522214,0.3903640211,0.2625901401,-0.1352918744,-0.2359022647,0.3025464416,0.3034740388,0.1200008765,-0.0492572039,0.4870747924,0.4269422591,0.4090619385,0.0039016933,0.2013225257,-0.0100115081,-0.1889312714,-0.0214954503,0.0882004872,-0.0550458506,0.5318027735,0.1638848484,0.2567686439,-0.2468539625,0.185496375,0.0133697847,-0.0159491897,-0.1292088777,0.691208601,-0.2564795911,-0.1056898162,-0.2331102788,0.0544373505,-0.5289071202,0.2520841956,0.3535762727,0.0812641606,0.1096357256,0.0725061074,0.0778281316,0.0445873886,0.2782047987,0.342743665,0.30538252,-0.5661407709,0.0243918225,-0.6365788579,0.1711902022,-0.28920421,-0.0434493534,-0.1806510538,-0.015653681,-0.136431843,0.1296598315,0.1261664033,0.0536839589,-0.0534693822,-0.105646655,-0.2898376286,-0.0953034908,-0.0626657307,0.0257232394,0.015950067,-0.4950198531,0.0248199925,-0.413179934,0.0006352767,0.027058905,-0.133508116,0.0396981873,0.3611627519,0.4533376694,-0.026389746,0.5661256313,-0.0605582707,-0.1965637505,-0.1913531125,-0.2649549842,-0.1598000079,0.2294674814,0.1461038738,0.3851700723,-0.3064142168,-0.0087604532,-0.4270081222,0.2901292741,-0.0378126204,-0.067394942,-0.2009867877,0.0023696769,-0.1316203475,0.1983623058,0.1464611888,-0.0151976235,-0.0840773508,0.1883648485,-0.2635794878,-0.3884920478,0.46128425,-0.2634094059,-0.2436618209,0.0428768806,0.1699992418,-0.0931862071,0.1788178533,-0.4839810431,0.3811461627,0.438337028,0.1640297323,-0.2190377563,0.1378456354,0.0010955908,0.1383843273,0.0792405829,0.4123993814,0.0075344145,-0.3825779259,-0.0340820029,-0.1186552644]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/871","title":"terminate called after throwing an instance of 'google::protobuf::FatalException'","comments":"Loading the iwslt2017-en-nl config of iwslt2017 works fine on my side. \r\nMaybe you can open an issue on transformers as well ? And also add more details about your environment (OS, python version, version of transformers and datasets etc.)","body":"Hi\r\nI am using the dataset \"iwslt2017-en-nl\", and after downloading it I am getting this error when trying to evaluate it on T5-base with seq2seq_trainer.py in the huggingface repo could you assist me please? thanks \r\n\r\n\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 63\/63 [02:47<00:00, 2.18s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nrun_t5_base_eval.sh: line 19: 5795 Aborted ","comment_length":39,"text":"terminate called after throwing an instance of 'google::protobuf::FatalException'\nHi\r\nI am using the dataset \"iwslt2017-en-nl\", and after downloading it I am getting this error when trying to evaluate it on T5-base with seq2seq_trainer.py in the huggingface repo could you assist me please? thanks \r\n\r\n\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 63\/63 [02:47<00:00, 2.18s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nrun_t5_base_eval.sh: line 19: 5795 Aborted \nLoading the iwslt2017-en-nl config of iwslt2017 works fine on my side. \r\nMaybe you can open an issue on transformers as well ? And also add more details about your environment (OS, python version, version of transformers and datasets etc.)","embeddings":[-0.2756947875,-0.6896218061,0.1162040085,0.3686614335,0.3344204426,-0.0612295717,0.3348029554,0.1536107957,-0.2083397657,0.1179825291,-0.0654007196,-0.3324613571,-0.0295326822,0.4058206081,-0.1832788736,-0.2990669608,0.0189545434,0.1540884376,-0.4621327221,-0.0021639555,-0.229054451,0.3899804056,-0.2134581506,0.3483884335,-0.1130772457,0.1178065538,0.0829797462,0.0127721429,-0.0949840471,-0.2316043675,0.2211651206,-0.0909958556,0.16676566,0.3702025115,-0.0001311232,-0.0367903896,0.2655953467,0.0570688918,-0.144013986,0.1104688868,-0.0511542857,0.1461777091,0.0536198951,0.0361517482,0.0598792806,-0.2049441636,0.2089884281,0.231264174,0.4897398949,0.0151624773,0.0531786382,0.3063324094,0.3312010467,0.0816985816,-0.0270447396,-0.1165635586,-0.2888156772,-0.0628107935,-0.1118589938,-0.0160784628,0.3052893579,0.0109918034,0.0765329003,-0.1810525209,-0.165804401,-0.046485927,0.4919874072,-0.3176007271,0.2732107639,0.3026655614,0.5659933686,-0.2089445889,-0.320820719,0.2825088203,-0.162963748,-0.663364172,0.029516371,0.0157640278,0.0368270725,0.2872439623,-0.1765947342,-0.0747157559,-0.35607481,0.0618637055,-0.2483198345,0.3457750976,-0.0724097714,0.3870842457,0.1909076124,0.0661901161,-0.3279819787,0.2797852159,-0.1643176973,0.0391894393,-0.3566974699,-0.2677064836,-0.07975474,0.1459319293,0.1702948064,0.1213659793,-0.0884903446,0.0574843735,0.250775665,0.1139543504,0.4480936527,0.0888853446,-0.0909259915,0.0955569446,0.263713032,0.1925384849,-0.0199964084,0.0608912818,0.1707051396,-0.2056468129,0.2166097313,-0.0390899517,0.5442737341,-0.4371444881,-0.3980947733,0.3282663822,-0.5851676464,0.0734328553,0.1095122471,0.4647295177,-0.0218743067,0.1591077894,0.2288995832,0.0017637751,-0.2170954794,-0.3298825622,-0.0325273424,0.1248777285,0.013593358,0.0792466924,-0.0938060656,-0.1105452105,0.0189604871,-0.1687659621,0.0570333302,-0.4175057113,-0.0315140933,-0.2017413676,-0.2151169777,0.0476760156,0.0230274014,0.0913014933,0.0626819059,0.2338235974,-0.108844474,-0.0194759425,0.2328732312,-0.4418900907,-0.1168154776,0.0209349543,-0.3718620241,0.0700107217,0.0511764847,0.1008900106,0.0117201507,0.0647251308,0.0112039009,-0.1178448871,-0.601600647,0.0955598652,0.2137631625,0.5098158121,-0.2983199954,-0.28317222,0.0214242265,-0.2185611278,0.1330840141,0.3510248661,-0.1074053645,0.4740375578,-0.1680228561,0.0771273598,0.4050397873,-0.0754855275,-0.4074357152,0.2993320227,-0.1043747813,-0.0264380835,-0.2313578129,-0.3312267959,-0.2108638287,-0.0910302997,-0.4682708383,0.0834774077,0.0042564496,0.0229098741,-0.3265223205,-0.2786833346,0.2017011046,0.1292589903,0.084895663,-0.1155647039,0.0439050086,0.5326945782,0.0079243751,-0.1599257588,-0.0717023686,0.2296440005,0.3724676967,0.0957855359,0.0016297285,-0.0008592252,-0.3926899433,0.1919946223,-0.1509628445,0.3499168158,-0.0339801721,0.1316320002,-0.3984308839,0.2384351492,-0.3872943819,-0.2096565515,0.0423250608,0.1670191139,-0.1135266274,-0.0217385683,-0.0022783617,0.2230875194,-0.3238983154,0.1648083925,-0.1945526004,0.2180779129,-0.1290382594,-0.1867525429,-0.0920874253,0.2005738467,0.1242625192,-0.2472160012,-0.0812027007,0.5033457279,-0.4497034252,0.3622533381,0.1256987303,0.2714719474,-0.1072640643,-0.7133020163,-0.1491653323,0.0605793186,0.0442922227,0.0842485502,0.2721653879,0.0209644437,0.36870116,0.1002801135,0.0422450602,-0.1372859478,0.0769415498,-0.1270656139,-0.3090764582,-0.0567719229,0.196080789,0.0975053906,-0.097971715,0.092575267,-0.3171671331,-0.1859455705,0.5001758337,-0.3895013034,0.1055529118,0.1940921694,0.221313715,0.0158279464,-0.0167704094,-0.3721936345,0.4321252108,-0.1188456342,-0.0447104648,0.2925679684,-0.2349233478,0.3068990409,0.2109820098,0.2268821746,-0.0990797579,0.2470541894,0.1511300802,0.1275759935,-0.2835476398,0.1347997189,-0.0292089619,0.3193389773,-0.3493123651,0.3340668976,-0.2501600385,0.0029524611,-0.0844357237,0.1695369184,-0.2327766865,-0.1175029948,-0.0777457878,0.3478622139,0.2846853137,0.2375227958,0.136313796,0.2873690426,0.3123163581,0.0504829735,0.0776460469,0.0606766343,-0.3715503216,-0.1625667959,0.2789064944,-0.0084201917,0.1755388528,-0.0387269333,-0.2875962555,-0.1948439926,-0.2629094422,0.4366433024,-0.2573680282,0.3862488568,0.2549961507,0.3238667548,0.0508058593,-0.2321354598,0.3769733608,-0.0804985315,-0.1112752408,0.1476732641,0.0080181323,0.0012472833,-0.1322028935,-0.1545511931,-0.0031232389,-0.2535025775,0.3643809855,-0.1252323091,-0.0454624519,0.1597647965,0.0816032141,0.0978151113,0.0324481204,0.0944973305,-0.2424069941,0.071293667,0.1028503701,0.170868665,-0.1095625237,0.0842222348,0.1628800929,0.4006054103,0.0270047523,-0.5209590197,-0.0805877149,-0.0152732851,-0.2542364895,-0.0713413805,0.1167048663,0.33683303,0.0828818902,0.163786754,0.1131120771,-0.0577146374,0.3626656234,0.1209022775,0.4364433587,0.1704380363,0.6243668199,0.03596754,0.4471463859,0.2275732607,-0.094705753,0.3119553328,0.093693763,0.0500957854,-0.0640045106,-0.2120324373,0.0056936136,-0.1506774127,-0.1362486482,0.3213931024,-0.1273608655,-0.052493155,-0.2199124992,0.3802466989,-0.2665280402,-0.383638829,-0.1950196475,0.4791100025,0.4469768405,-0.211784929,-0.1293497533,0.0210550521,0.1487210095,0.2695375383,0.0164810456,0.0836106166,-0.0299992282,0.1247652248,-0.1271421611,-0.3214202523,0.2202692479,0.0824960172,0.3695675731,-0.2289727032,-0.0523838587,0.222915858,-0.0894602388,0.3142283261,-0.1110014245,0.0958874002,-0.0223501697,-0.3808889091,-0.0712860227,-0.1707492918,-0.2458643913,-0.0534696914,0.1116184667,0.196276024,-0.722766459,-0.1465538144,0.6654101014,0.0148802157,-0.2597118318,-0.0307936873,-0.4269811213,-0.4244686365,-0.3164515495,0.026061669,0.1898845881,-0.0228520781,-0.1808922291,0.1003893167,0.1215846911,-0.3787500262,0.1677027345,0.243192777,-0.0777752399,-0.2946285903,-0.141963318,0.1522611678,-0.2428970039,-0.2091581225,0.3184257746,0.2315772325,-0.3591951728,0.3091997206,0.193026647,0.1260813028,0.4352583885,-0.237452507,0.0241909176,-0.3933628201,0.1062252373,-0.0747986063,0.1896966547,0.6916982532,0.1641761363,-0.3732667863,0.1904458553,0.462097317,-0.1921073347,0.0380767509,0.22014153,0.3627300262,-0.1466461718,0.2382568866,-0.1088859141,1.2690070868,0.0817548186,-0.1839938015,0.3657687306,0.1786314547,0.2574061751,-0.2753503323,0.2142233402,-0.2846076488,-0.2586662471,-0.0980635807,-0.0324810036,-0.0331729576,-0.111914061,-0.467831403,0.3271196485,0.1595585048,-0.1385068744,0.2032443583,0.2743907273,0.0084879482,-0.0066031311,-0.2490342259,-0.0122612668,-0.0838413164,0.547954917,-0.4007770419,-0.0408161134,0.1884874552,0.069105424,-0.5751296282,0.1962982565,-0.2301561087,0.0088640861,0.16889368,-0.1305916309,0.3103165925,0.1113410369,0.3887819946,0.150553152,-0.3223505914,0.0906874314,-0.2385719717,-0.1092514098,-0.143951416,0.2629018128,0.063634254,-0.1279810071,-0.4963575602,0.1836188585,-0.0275697261,-0.1637552083,0.0390904248,-0.4867976904,-0.1143436357,-0.4812121987,-0.0499083549,0.0644627661,-0.3322571516,-0.191035077,-0.0241218321,-0.0708206147,-0.4919304848,-0.0446474999,0.1353206187,-0.3358820081,-0.0327948332,0.2914811671,-0.4647303224,0.495544672,0.5175926685,0.4059613347,0.0345027,-0.1226497591,-0.41053316,-0.105709292,-0.7558622956,0.1090186089,-0.0462948494,-0.1276709437,0.0046852822,0.6766906977,0.1893719584,0.0832461119,0.0375354327,-0.0291810036,-0.1648816317,0.2903730869,0.0359393433,0.392903775,0.0263908468,0.6678969264,-0.1816685796,-0.2087010741,-0.1348351538,-0.1140431836,-0.5952382088,0.0431545153,0.6293122768,-0.4968104959,0.2551211715,0.2187540084,0.0470926687,0.365015924,-0.1530413032,-0.0085513489,0.0009548252,0.2152924985,0.1432794631,-0.2293219268,-0.0499960333,0.0648234263,0.1181975976,0.0223332252,0.1977396607,-0.0345476493,0.1043082625,0.1668601483,0.4337745607,0.0893760547,-0.328541398,-0.0826574713,0.1569119394,0.3206041455,0.4231345356,0.1758082956,-0.0221177656,-0.1343773156,-0.009829565,0.2518425882,0.5083553195,-0.0776105672,0.3535529971,-0.2344874591,-0.1737599969,-0.2526756227,0.4068441391,0.43582201,-0.3030586243,-0.077644892,-0.0429386199,-0.0174998119,-0.1370070577,0.0602756292,0.2170641124,-0.0761528015,-0.1653425992,0.5258344412,0.0706920102,0.0977590084,0.3721161485,0.2665567994,0.2728874087,-0.1531066298,-0.0537603907,0.270961225,-0.3218219876,0.2332793623,0.0879884884,-0.0149972206,0.3673686087,-0.1052370593,-0.1762795448,-0.0077566765,-0.6213701367,0.5529290438,0.0994614512,-0.411948204,-0.1608340889,0.0955712274,0.2579673529,0.1640248448,-0.1661908776,0.3881980181,0.0901009589,-0.3478374779,-0.4924358428,0.5627099872,-0.122069329,-0.3420293033,0.1224452406,0.0005514151,-0.4981306791,0.2230651826,-0.1528074443,-0.3076917231,-0.1538954377,0.1734644324,-0.1786789894,-0.2412707806,-0.3105516136,0.1349073797,0.0829802081,-0.168107897,0.3439385295,0.253600657,-0.230373174,-0.2507804632,0.4816981554,0.0894158855,-0.0768770427,-0.2228053957,0.2002964765,0.0184817482,0.1735889763,0.1793337613,-0.0232881773,-0.0084643327,0.2301135659,0.2180700898,-0.0970742255,-0.030791644,-0.0214373767,-0.0897680521,0.1906954646,-0.5353975296,0.0491416529,-0.1127958298,0.0020704046,-0.4266152084,0.2310675532,-0.4342006147,0.2886883914,0.2736751139,0.0018204541,0.0686449856,0.1550022662,-0.037596751,-0.1003341377,0.3217399716,0.4034511745,0.0491667986,-0.3347398937,-0.2030802518,-0.2273234278,0.352452606,0.0174698327,0.3788674474,0.3489823937,0.0774133652,-0.1255188286,0.2282455564,0.1392574012,0.2732212245,-0.1463273466,0.2943367064,-0.1593057364,-0.2022354007,-0.3895681798,-0.0320488587,-0.0684848949,-0.1479368806,0.1528607011,-0.0483459681,-0.09997648,0.0195354894,-0.4915774763,-0.3992326558,-0.1382072568,0.0471328162,0.2893938422,-0.0539369993,-0.1106616259,-0.1935112476,-0.1631006747,-0.3601306677,-0.1371573508,0.1329978108,-0.2041321546,0.1272538751,-0.0116046891,0.3329965472,-0.3921409845,-0.0241247024,0.2137951255,-0.1817460507,-0.3372456729,-0.0258327369,-0.2492695302,-0.0717044398,0.1051634625,0.2207347602,0.1219839528,0.0383794904,-0.4720091224,0.0056419955,0.6514028907,-0.0394012071,-0.1021966934,0.025127491,0.0307157077,0.3078186214,-0.1596795022,-0.5445809364,0.0636720657,-0.0219524652,0.0791893154,-0.2491616756,0.1685010344,0.1020446122,-0.0360069275,0.1172407418,0.3234377205,0.2480837852,0.0809629038,0.3928499222,-0.0464752987]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/871","title":"terminate called after throwing an instance of 'google::protobuf::FatalException'","comments":"closing now, figured out this is because the max length of decoder was set smaller than the input_dimensions. thanks ","body":"Hi\r\nI am using the dataset \"iwslt2017-en-nl\", and after downloading it I am getting this error when trying to evaluate it on T5-base with seq2seq_trainer.py in the huggingface repo could you assist me please? thanks \r\n\r\n\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 63\/63 [02:47<00:00, 2.18s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nrun_t5_base_eval.sh: line 19: 5795 Aborted ","comment_length":19,"text":"terminate called after throwing an instance of 'google::protobuf::FatalException'\nHi\r\nI am using the dataset \"iwslt2017-en-nl\", and after downloading it I am getting this error when trying to evaluate it on T5-base with seq2seq_trainer.py in the huggingface repo could you assist me please? thanks \r\n\r\n\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 63\/63 [02:47<00:00, 2.18s\/it][libprotobuf FATAL \/sentencepiece\/src\/..\/third_party\/protobuf-lite\/google\/protobuf\/repeated_field.h:1505] CHECK failed: (index) >= (0): \r\nterminate called after throwing an instance of 'google::protobuf::FatalException'\r\n what(): CHECK failed: (index) >= (0): \r\nrun_t5_base_eval.sh: line 19: 5795 Aborted \nclosing now, figured out this is because the max length of decoder was set smaller than the input_dimensions. thanks ","embeddings":[-0.1825481504,-0.7557119727,0.0836407691,0.556730926,0.3312155604,-0.0441040806,0.2286126763,0.2201018035,-0.3796162307,0.3128049076,-0.0100409836,-0.4250835478,0.0028816557,0.5060398579,-0.2287921458,-0.4332334399,-0.0710474998,0.3094232976,-0.3874340355,0.1011399552,-0.2264460176,0.418671906,-0.2191978246,0.1852833182,0.0081758648,0.1957097799,-0.0577186234,-0.0667141601,-0.3531189263,-0.3468537331,0.0182195511,-0.2039844096,0.0581342317,0.3137004673,-0.0001306237,-0.0284277499,0.2318881005,-0.0366353914,-0.0024561405,0.1110143065,-0.0404468663,0.0954321474,-0.0432409495,0.0571707711,-0.0214736089,-0.2490238994,0.1340977252,0.0824278742,0.4000383317,0.1778144687,0.0598794147,0.0708278567,0.5268979669,0.1011874452,0.1303852797,-0.1859424114,-0.2451243103,-0.0544138551,-0.0872224346,0.254393518,0.2519197166,0.0127651393,0.2343612909,-0.10397394,-0.1582995653,-0.1667215228,0.4694933891,-0.3594505787,0.3739234507,0.5036091208,0.4249293506,-0.1596073657,-0.2389924079,0.1978807002,-0.0463761911,-0.7821172476,0.0214621071,0.1177389845,-0.0314023718,0.3177411556,-0.1956370324,0.0250428114,-0.3913386464,-0.0535499491,-0.3420303464,0.1048160717,-0.1563069522,0.2693475485,0.3107014596,-0.012363188,-0.3689581454,0.4468303621,-0.2233463228,0.0046288338,-0.3167666197,-0.2840055227,-0.087674655,0.0885874704,0.3425703347,0.0983420759,0.0299788564,-0.044186756,0.3686290681,-0.0731782168,0.4484470189,0.0698964819,-0.1750067025,-0.047461804,-0.0304171108,0.2391154468,-0.0833925828,0.0399034657,0.0878183767,-0.386467576,0.0334503576,-0.0612443797,0.5407340527,-0.4030388296,-0.3742876649,0.5508589745,-0.5646084547,0.1382816583,-0.0152639402,0.3928346932,0.0368884876,0.3064093888,0.1208616048,-0.025470525,-0.1815385073,-0.4603445232,-0.0324585438,0.0571843795,0.068704173,-0.0381941721,-0.2290016264,-0.1918761581,0.0274250377,-0.1162856147,0.2301355898,-0.4824358821,-0.0526085831,-0.3860574365,-0.1272218376,0.0582951158,0.1070084795,-0.0941504464,0.0039941398,0.1228219047,-0.136877954,0.1021425724,0.2030085027,-0.3419500291,-0.2019581199,0.0443180725,-0.2757993937,0.2491314113,0.0974390805,0.2676698267,0.0701395497,0.1285983175,0.037547905,-0.2970994115,-0.5591106415,0.0333798081,0.2109642178,0.3538913429,-0.4521012306,-0.1101317555,0.1289368123,-0.162707895,0.2067390829,0.3624174297,0.0659419522,0.2911633551,-0.2080686092,0.1842984855,0.3846414685,0.0041619176,-0.4475169182,0.4741547704,-0.1718921214,0.0268619098,-0.0815850273,-0.5575348139,-0.0763312876,0.0625408515,-0.3875132203,0.0609403141,-0.050428085,-0.1020278111,-0.3786960244,-0.3348831534,0.2869663835,0.0521849319,0.0931096971,-0.2644413114,-0.0289088674,0.3470962942,0.1054541841,-0.1164799556,-0.0546410456,0.2943658829,0.1599955708,-0.0123672755,0.0500191785,-0.0715421513,-0.059885744,0.1266557723,-0.1815430522,0.2302283794,0.1811086833,0.0477650389,-0.4110559523,0.2406524271,-0.1958540082,-0.1214052588,0.0813282952,0.0905070826,0.0031745289,0.0335053764,0.0377168022,0.0373364352,-0.279001534,0.1965822279,-0.1421760172,0.0688369796,-0.2133826017,-0.1122256517,0.0000704748,0.2261541337,0.0785465315,-0.1326743513,-0.13583529,0.4077900052,-0.1459801346,0.1788313091,-0.0823335871,0.0799966678,-0.1061856225,-0.6050518155,0.0727640837,0.0006517002,-0.0793634504,0.0802322552,0.4008986652,0.0033712611,0.4482319057,0.1287207454,0.0407877825,-0.2953241169,0.0407613441,-0.0307979509,-0.2658254504,-0.1130768657,0.2272973061,0.0492021553,0.0264593475,0.0901864544,-0.3722547591,0.0433302931,0.5868806243,-0.3244768381,-0.0204519518,0.3520979285,0.1912356168,-0.0975401103,-0.1523132473,-0.2997414768,0.2898995876,-0.0278740041,0.1170780435,0.2232637703,-0.1730557382,0.1739894599,0.2037848681,0.2063530385,-0.0424582995,0.3433316946,0.1607792675,0.0747347847,-0.194274202,0.1378559768,-0.0842173025,0.2802475095,-0.2942640781,0.2067055255,-0.3464756608,-0.1027304381,-0.036015369,0.2028561831,-0.1097783148,-0.0775026605,-0.0121806301,0.1491119415,0.2092546076,0.0943256617,0.0159660671,0.3456902802,0.3015296757,0.3043162525,0.1930469275,0.2721781135,-0.3297426701,-0.0792557523,0.2436175197,-0.09949781,0.1018210649,-0.0045349356,-0.2779570222,0.047878366,-0.3190452158,0.3975229859,-0.2260155231,0.4261747599,0.2910575271,0.3495925069,0.162409544,-0.1286505461,0.2713697255,0.1310101002,-0.2149520963,0.1189751178,0.0433747992,0.1060262918,-0.1928852499,0.0047683287,0.0756090879,-0.3440089822,0.597823441,0.0725695342,0.0980785564,0.1002419367,0.2643187344,0.1980092078,0.1320688128,0.2236942202,-0.3274862468,-0.0945414826,0.070234336,0.0601463728,-0.1019025594,0.0203388724,0.1251716018,0.552069962,0.0094988542,-0.6542897224,-0.0641237572,-0.0393796414,-0.3225176632,-0.0548027568,0.0118522821,0.2194900066,-0.0144388573,0.1663329154,0.1242133677,0.0081098229,0.2320028096,0.1027007401,0.3276295662,0.1301310956,0.6841685176,0.1434123963,0.4717115164,0.3242788911,0.0999211371,0.2292104363,-0.0142609617,0.1526381522,-0.1874094158,0.0287796371,0.2010574043,-0.1487109363,-0.228917703,0.3736218214,-0.1404163688,-0.0357154459,-0.0309773013,0.2981751859,-0.1749610156,-0.3061987162,-0.0185549166,0.4317148328,0.1991914809,-0.363768965,0.0199164394,0.0947428048,0.1148888022,0.1037391573,-0.1432925463,0.0509277359,0.106900461,0.1287240386,-0.2539620996,-0.453119725,0.2327690125,0.1363760531,0.2795752585,-0.2190567702,-0.1029563397,0.2003332227,-0.1413599849,0.2534681559,-0.0616666973,0.0041800593,0.087018773,-0.2812114954,-0.0106085846,-0.1489334702,-0.1682466716,-0.0887436643,0.0999071896,0.0629592836,-0.6216032505,-0.0018601259,0.6708574295,0.0052203136,-0.1165489107,0.1024371609,-0.3825119436,-0.4406279922,-0.2710041404,0.0987983644,0.3051098585,0.1490618736,-0.1904389113,-0.0208010338,0.0858972147,-0.3469175696,0.2035472691,0.2855482697,-0.0550692864,-0.2784851789,-0.0909604877,0.0618520118,-0.2808569372,-0.0092509724,0.3237591386,0.1873803735,-0.4014208615,0.3799914122,0.1589251906,0.2553518116,0.5518909693,-0.1993993223,0.1199700534,-0.361264348,0.3271133602,-0.2000967711,0.340396136,0.6970121861,-0.0088739647,-0.2793234885,0.1268748045,0.354111284,-0.3083726466,0.1861292869,0.2749052048,0.4879734516,-0.1317798048,0.2013567388,-0.0295752808,1.3459961414,0.1415599138,-0.2212045193,0.362362951,0.0220959336,0.0692190751,-0.3213464916,0.222608462,-0.2752582133,-0.2632488906,-0.0801570415,-0.0565181002,-0.0217673983,-0.1965672225,-0.5029943585,0.1471654326,0.146159187,-0.2665548623,0.0257626437,0.2183863074,-0.0723987147,0.0217419546,-0.0787748173,0.004437129,-0.127093181,0.2550251186,-0.2989115417,0.0173455887,0.0929535031,-0.0203148387,-0.7395150065,0.1649557501,-0.3972647488,-0.2317931354,0.0192548949,-0.2671796978,0.2511187792,0.1190080494,0.2909447551,0.2553223073,-0.3112871647,0.056603346,-0.2090291381,-0.2243221849,0.0117058679,0.1983579993,0.0477385558,-0.1979283988,-0.3976749778,0.1828247458,0.1018944532,0.0594495758,0.1003047451,-0.5291588902,-0.090665251,-0.5466359258,-0.0517877117,-0.123738572,-0.1631921381,-0.1833691746,-0.0224830918,-0.0587468781,-0.497446388,-0.2320779562,0.127278626,-0.3527326882,-0.0955886245,0.1552886367,-0.2879287601,0.5394779444,0.4464948177,0.3756778538,0.0185434557,-0.0896206573,-0.2411793321,-0.1186503768,-0.5672418475,0.1138521507,-0.1296156198,-0.1808696538,-0.1257193983,0.4178597331,0.2591121197,0.072935164,0.0418938138,-0.0680543259,-0.2588287294,0.2252996713,-0.0678906068,0.442528218,-0.016238099,0.6142128706,-0.1967774928,-0.0419657677,-0.1338166445,0.024128342,-0.4663615525,-0.0486499816,0.4135287106,-0.5600551963,0.0643675625,0.1984225065,0.0650205761,0.5820106864,-0.1316850036,-0.034659639,-0.0889136866,0.1771107763,0.0784016699,-0.2593673766,0.0305792838,0.1024134383,-0.0138144577,0.0227826089,0.1982651502,-0.113809891,0.190584451,0.0753315091,0.452208221,0.0531573258,-0.2297775149,-0.063360326,0.2922243774,0.5698916316,0.4143060744,0.2580898404,-0.047366295,-0.1122656018,0.0235160943,0.3457762301,0.4345481396,-0.2133328766,0.3456012905,-0.2215126902,-0.1793737113,-0.174184233,0.3166123629,0.4049066007,-0.349421382,-0.0612707995,-0.0223750845,-0.0171981081,-0.1205972284,0.0780511647,0.2076966465,-0.147224158,-0.1990442872,0.4944213629,0.0151153123,0.0951936245,0.2049716413,0.2440080196,0.043497324,-0.0081223901,-0.145504415,0.3248820007,-0.3696915507,0.1776641756,0.0745541826,0.0668813363,0.5588196516,0.0386203378,-0.2440882325,-0.0359904617,-0.5670961142,0.5346621871,0.1142034158,-0.3848816156,-0.2093838453,0.2809021175,0.0738079697,0.229156062,-0.0532865338,0.3404663503,0.2014734894,-0.4261129498,-0.5190367103,0.4846898317,-0.2030854076,-0.3480958343,0.1934646368,0.0305538215,-0.5891631842,0.3082311451,-0.0769739226,-0.2484859824,-0.2079977542,0.0530775115,-0.1518977731,-0.0550663508,-0.2682462335,0.070760414,0.1284526587,-0.3370711505,0.4778930843,0.4300207198,-0.1742856354,-0.1102564931,0.4660412073,0.2497323155,-0.0895262882,-0.2848259211,0.0940386802,0.1176966652,0.079322584,0.0723173246,-0.0033886547,0.1967161447,0.4727398157,0.356105417,-0.076764822,-0.0420947522,-0.0304094814,0.0197085235,0.1798907518,-0.6338071823,-0.0884627923,-0.1270017326,0.0667711124,-0.3211508393,0.0748252869,-0.5564795136,0.2530530989,0.1631408185,-0.0004767778,0.0540471002,0.1432605833,-0.0327448137,-0.0677551553,0.2711188197,0.4403904974,0.0064048748,-0.2855249047,-0.2524416149,-0.1323001087,0.3613683283,-0.2400722951,0.1405515373,0.3154822588,0.1597657055,-0.1240986213,0.1682173312,0.2631752789,0.3169566989,-0.062899828,0.3524359465,-0.1335162222,-0.2348483354,-0.3311816156,-0.1356596649,-0.0208175369,-0.1364220977,0.2971139252,-0.0657794476,-0.1463050246,0.0958926678,-0.2550741732,-0.3955486715,-0.2489908934,0.0886138678,0.2972622514,-0.092504777,-0.1601975113,-0.2199968845,-0.2032061219,-0.2314457595,-0.2321934849,0.2132405937,-0.1535039991,0.1309258193,-0.0800581649,0.1188635454,-0.21908696,0.0467709601,0.2452969849,-0.0743778944,-0.3315639496,-0.0217457414,-0.2387637049,-0.029810464,0.2515655458,0.2170667201,0.1479577422,-0.019263912,-0.319655627,-0.080541186,0.551391542,0.021487046,-0.0822739974,0.052417852,0.0193322059,0.391785264,-0.1135360375,-0.6487380862,0.1016550735,0.0261529107,0.1344753951,-0.2566714585,0.2677623332,-0.1109553948,-0.0772487894,0.0469754636,0.3199744225,0.2324951291,0.050279323,0.3381661773,-0.1106577069]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/870","title":"[Feature Request] Add optional parameter in text loading script to preserve linebreaks","comments":"Hi ! Thanks for your message.\r\nIndeed it's a free feature we can add and that can be useful.\r\nIf you want to contribute, feel free to open a PR to add it to the text dataset script :)","body":"I'm working on a project about rhyming verse using phonetic poetry and song lyrics, and line breaks are a vital part of the data. \r\n\r\nI recently switched over to use the datasets library when my various corpora grew larger than my computer's memory. And so far, it is SO great. \r\n\r\nBut the first time I processed all of my data into a dataset, I hadn't realized the text loader script was processing the source files line-by-line and stripping off the newlines. \r\n\r\nOnce I caught the issue, I made my own data loader by modifying one line in the default text loader (changing `batch = batch.splitlines()` to `batch = batch.splitlines(True)` inside `_generate_tables`). And so I'm all set as far as my project is concerned.\r\n\r\nBut if my use case is more general, it seems like it'd be pretty trivial to add a kwarg to the default text loader called keeplinebreaks or something, which would default to False and get passed to `splitlines()`. ","comment_length":39,"text":"[Feature Request] Add optional parameter in text loading script to preserve linebreaks\nI'm working on a project about rhyming verse using phonetic poetry and song lyrics, and line breaks are a vital part of the data. \r\n\r\nI recently switched over to use the datasets library when my various corpora grew larger than my computer's memory. And so far, it is SO great. \r\n\r\nBut the first time I processed all of my data into a dataset, I hadn't realized the text loader script was processing the source files line-by-line and stripping off the newlines. \r\n\r\nOnce I caught the issue, I made my own data loader by modifying one line in the default text loader (changing `batch = batch.splitlines()` to `batch = batch.splitlines(True)` inside `_generate_tables`). And so I'm all set as far as my project is concerned.\r\n\r\nBut if my use case is more general, it seems like it'd be pretty trivial to add a kwarg to the default text loader called keeplinebreaks or something, which would default to False and get passed to `splitlines()`. \nHi ! Thanks for your message.\r\nIndeed it's a free feature we can add and that can be useful.\r\nIf you want to contribute, feel free to open a PR to add it to the text dataset script :)","embeddings":[-0.4354711771,0.2950844765,0.0567216426,-0.2441309839,-0.0972463861,-0.2072373778,0.3331260979,0.1758846045,0.298045516,0.3277394772,0.503433764,0.2391882837,-0.0176374037,0.1306846887,-0.0111618992,-0.1541550457,-0.16735439,0.3164179623,0.0948439389,0.0943592563,-0.1742428243,0.2289418578,-0.235174939,-0.3131567836,-0.0942863822,0.1758474857,0.0615477115,-0.1762807816,0.0486642383,-0.2417834848,0.0536904819,0.4617238045,-0.2017268538,-0.0629813299,-0.0001233125,-0.1519173682,0.1834264398,-0.1822315753,-0.6022976637,-0.1147578061,-0.3729275763,-0.2914463282,-0.0992282629,-0.2274686992,0.0622691773,0.3424104452,-0.0458786301,-0.4362998009,0.2800565064,0.3078560531,0.0495161414,0.1074072942,-0.5713555813,-0.0663225502,0.5268213749,0.2910938561,-0.0629604682,0.0548023283,0.0162094031,0.0578812249,-0.4066486061,0.3086510599,-0.3627738655,-0.2398989648,0.3153908551,0.0333261937,-0.1040520519,0.0191558599,0.1258228421,0.6455950141,0.60443151,-0.5228234529,-0.1811380684,-0.5374609232,0.3158695698,-0.2599709928,0.0773754343,0.0897537693,-0.4800499976,0.1028653905,-0.3286261261,-0.5799759626,-0.2724969685,0.0282563232,0.0515560955,0.0071636853,0.0852803141,-0.1294995248,0.2139755487,0.2667275369,-0.1155733764,-0.3966223001,-0.1672593951,-0.0120494142,0.186608389,-0.1383511573,-0.1339076161,0.0295246951,0.2464732081,0.1903955638,0.2343104035,-0.0728933737,0.1659848839,-0.0096090594,-0.1466227919,0.0010832801,0.6396608949,-0.1263293624,0.3962599933,-0.0598768182,0.2541896701,-0.0147522138,0.3621482253,-0.0553385541,0.2506429851,0.2168622315,0.5353254676,0.0507847294,-0.1015287712,-0.0084517095,0.1577020437,-0.2738271654,0.0231101047,0.0231772289,0.2529730201,0.2497906089,0.0939324647,0.339779228,-0.0245637652,-0.4520849586,-0.011584973,-0.0739066601,0.0690934211,0.4735104442,0.4798795879,-0.1322537512,0.0188573785,0.3098440468,-0.3272877038,0.2125819623,-0.2599174976,0.0107391179,-0.0624036416,0.0897311941,0.0502958484,-0.0514470898,-0.2781385481,0.1292174906,-0.1442714781,0.3982302845,-0.2081502825,-0.2423781157,-0.0120961415,0.1186780483,-0.1746381819,-0.1164849475,-0.119011566,0.3408353925,0.2329739481,-0.2190415263,0.3950002193,-0.0677531585,-0.4148365855,-0.1905252039,0.1208384708,0.6844409108,-0.5018539429,-0.3941883147,-0.2446784824,-0.161777854,0.1972188801,-0.0845851451,-0.2083405107,0.3146696985,-0.0943591967,0.4131858051,0.5370357037,0.2126027495,-0.2113257647,0.4115028977,0.1002037302,-0.026850421,0.302057296,0.1166809946,0.4532318711,0.0867427513,-0.3238539994,0.505616188,0.1082096696,0.2016529441,0.0791892409,-0.1329990327,0.3330931067,0.1691438109,-0.1261501461,-0.0380360596,-0.0463945642,-0.0436888225,0.1735132337,-0.1540231556,0.0981511027,-0.0570635386,0.1134195551,0.2344139367,-0.0318807997,-0.0984189659,-0.5309476852,-0.1450162679,0.1449734271,-0.1053956747,-0.2306468189,-0.3572915196,0.1532925963,0.0899948329,0.0845068693,0.3819425404,-0.0544856638,0.076240249,-0.4913656712,0.1855785102,-0.2050102353,0.1258591115,-0.059617836,0.2973459065,-0.0412375517,0.1600450277,0.4320734441,-0.1415886432,-0.0284115095,0.2037593275,0.0357373431,0.0757633299,-0.1862323135,0.5891005993,0.2256730944,0.3013256192,-0.35499385,-0.3149200082,-0.0073471186,0.078584373,0.0901804417,0.0804409906,-0.1342437565,-0.1915708333,-0.2021723688,0.3526658714,-0.1503386796,0.335311085,-0.1181824058,-0.1217780635,-0.2494113445,-0.3147170246,-0.2909689844,-0.4226794541,0.2018985897,-0.1425284743,0.3798127174,0.2567830086,-0.0651082024,-0.0855781361,0.579664588,-0.1368783265,0.3069625795,0.3238114715,0.0211424418,-0.1402036101,0.1488457471,0.2698791921,0.3828237057,0.2634972036,-0.1487598717,0.066809997,-0.0059781624,-0.384434104,0.2802663445,0.2170626372,-0.0857429132,0.373426646,0.0165433735,-0.1488485634,-0.3234377503,-0.0239392072,0.021003224,-0.0865441561,-0.3517746031,-0.1167265996,-0.2842797637,-0.2587321103,-0.1788249165,-0.2389427423,-0.3589533269,-0.2414561361,0.1967178732,-0.1388000548,-0.3496547043,0.183488369,-0.2629195452,0.7408051491,-0.1833309233,-0.5621784925,0.0269698948,-0.1942996532,0.0812375247,-0.0659784749,-0.0138453506,0.1965544224,0.2640571594,-0.0036518862,0.0909801573,-0.069748193,0.1031589136,0.0497970171,0.0387121104,-0.2021922767,0.4928569496,0.1726311892,0.3285779655,-0.2593170106,-0.1793032438,-0.2124928087,0.0418997332,0.1382410973,0.11571455,-0.0770108402,-0.3744924068,-0.1818037778,0.0186165143,-0.2738801837,-0.0273665339,-0.4108988047,0.0061276481,0.2384626418,0.0889581889,0.0378395692,-0.0128541291,-0.0526321232,0.0465871394,0.1360524893,0.3777827024,0.018419534,0.1504377127,-0.0623602904,-0.0568684004,0.001330325,0.3807697594,-0.4775208533,0.2629410625,-0.120035246,0.495228976,-0.122098282,-0.0351204202,0.3664514124,0.13640064,0.0815182477,-0.2246571034,-0.0147726443,0.2787934244,-0.3612312675,0.0486786589,0.2344342768,0.5663764477,-0.0025382354,0.3888105452,-0.0068308953,0.1409278512,0.2664133608,-0.1572481096,0.3855630159,0.3046710491,-0.2800631225,-0.3151492476,0.079511039,0.0812318623,0.2062059939,0.0704425648,0.0193441808,-0.1236396357,-0.4614212513,0.2398138642,-0.5338416696,0.1624890268,-0.0830802321,0.2392730564,-0.1354087144,0.2915031314,-0.2403940409,-0.3213288486,-0.013779493,0.3209162354,0.3365838826,-0.0723714679,-0.2709951997,0.3177291751,-0.2754435539,0.1121079326,0.1583696902,-0.0539841019,0.067231752,-0.1683012992,-0.0166092422,0.2334416807,0.5474967957,-0.1255040318,-0.3826481998,-0.1610995084,0.2949236631,-0.0890731215,-0.0021835673,0.1107509285,0.2331686765,0.280511111,-0.0554789416,-0.262471199,-0.0135296257,0.2837168574,0.1578983963,-0.0368033946,-0.6033733487,-0.0771294981,0.0821998417,-0.091673404,0.0179504417,-0.0288762823,0.0184796453,-0.2994555831,0.1300362051,0.3077903688,-0.37880674,0.364957422,-0.093725957,0.1118697748,-0.0028544297,-0.0436347947,-0.217523858,-0.1295164526,-0.0625527129,0.2717804015,-0.2432946861,-0.3551975191,-0.4424709082,-0.4403403699,0.6828674078,0.2383652478,-0.2397620827,0.2966234684,-0.0111270016,0.0083185025,-0.0566013418,-0.0455776639,-0.1543446779,-0.5080237985,-0.1981165409,-0.4083796144,0.3745973408,-0.0850753188,-0.1184439585,0.0197544321,-0.0309661813,-0.4312166572,0.150943473,-0.1620285809,0.7613752484,0.4389469624,0.2339311093,-0.0483150817,-0.0530293919,0.4091447294,0.2154637128,0.0468552783,-0.1680984199,0.160315156,-0.1227327585,-0.0489166491,0.2017664164,0.0660158172,-0.4209389091,0.1925767511,0.1253014207,0.0905520841,-0.1094526052,0.3929077983,-0.5889460444,-0.4246123135,0.1515378505,-0.0910883248,-0.1265282035,0.010897452,0.2151646763,-0.1984631717,0.4758714139,-0.2690285742,-0.3731586039,0.1091191024,-0.1277019978,-0.1077003926,-0.0946270227,-0.1670501083,0.3970530331,0.4145120382,0.1724785566,0.3433482349,-0.1255666763,0.0487629697,-0.2599712014,0.1427637339,-0.0769660771,0.1011255682,0.252468437,-0.0155111551,-0.3237215579,0.0645443648,0.202799812,-0.4637390077,-0.2452293336,-0.3908312023,0.2647171021,-0.4940481484,-0.0714089945,-0.0571230166,0.357041806,-0.181701377,-0.0178062003,0.0578142665,0.3217685819,0.0182358734,-0.2762846649,-0.1267777383,-0.1952646822,0.3264217079,-0.0292975307,0.0016730555,0.4528479576,0.2491423041,-0.1543955207,-0.2483579069,0.0330738835,-0.0227095801,-0.6799131036,0.201711446,0.1961639225,0.1710075289,-0.0467553101,0.2048946917,-0.2973155379,0.2577685416,0.0099280244,-0.2536346614,0.0011220912,-0.0372331552,0.2559600174,0.4014256597,0.2199046016,-0.1678390652,0.0035412763,0.0085902037,-0.1561026424,-0.0239541121,0.4870351851,-0.1765925288,0.0112182358,0.2093708366,-0.0970261842,-0.2056295276,-0.1125248671,0.0031667855,0.043721322,-0.0393633619,-0.3089464605,0.2235017717,-0.1048402712,0.0451110601,-0.1107986644,-0.0085431226,0.1096312404,-0.1315191537,0.3208100498,0.1152769476,-0.2157720625,0.3483763039,0.1890638918,0.3747319281,-0.2061555535,0.096937649,-0.3119312525,0.2090507448,-0.0529515557,0.2815997005,-0.078202717,0.0784908757,-0.1537644714,0.1794761419,0.5239456296,0.1282903552,0.0408470742,-0.0844517574,-0.2382967621,0.0513686724,-0.048997201,0.0801132619,-0.4318627119,0.021955885,0.1042867079,0.086374566,-0.2357271463,-0.0589278825,0.0660034865,-0.0113795465,0.1313068569,0.2672941387,0.0319642089,-0.1237803176,0.19129017,-0.1409002692,0.4049847126,0.237860769,0.1618035436,0.5519504547,-0.0044507035,-0.0477991514,-0.1967972666,0.3016058803,0.0184472613,-0.0318382382,-0.0680063814,0.3881103396,0.1991785616,0.0285223462,-0.1200755686,-0.3986965716,0.1948861033,0.2730499506,-0.2493599504,0.218257919,0.1928058118,0.2495420575,0.0528857522,0.0230406132,-0.2613331378,0.2050992399,-0.0662043616,-0.1579679251,-0.5553824902,-0.0218588375,-0.1779490709,-0.0033020345,0.0341413058,-0.0601884238,-0.030757444,0.1199648306,0.0815442502,-0.003784783,0.3470280468,-0.3495810628,0.227890715,-0.3045914769,-0.1719600856,-0.1140139028,0.3842436969,0.5464944243,-0.0421134569,0.1087838039,0.1423364878,0.2631820738,-0.0332070142,-0.187361747,-0.0052228244,0.093494378,0.2236047983,0.1216042414,0.5366201401,0.3224514425,-0.0000479928,0.0238543097,0.4961556196,0.1896340698,-0.0232980195,-0.405671984,0.400192678,0.4818977416,-0.1691308171,-0.4080249965,0.0576710626,0.1050148606,-0.267621398,0.2647188902,-0.0559796877,-0.170176357,-0.2995608151,0.0201151799,0.0405171923,0.1210685521,0.3503684103,0.0009221981,-0.1106901318,-0.0419925153,-0.2783929706,-0.2308301777,-0.0526416749,-0.0632622167,0.2438689917,-0.2401642799,0.0626297593,0.1671016067,-0.3163401186,0.3353950679,-0.3051980436,-0.3688963354,-0.1019046605,-0.0181719344,0.164908126,-0.0723795146,-0.167579487,-0.084037073,0.07134597,-0.1059036404,-0.1353935599,-0.1731180698,0.2033342719,0.3793602586,0.317494601,-0.1166924462,-0.1601062566,0.185725227,-0.1130660027,0.1530615389,0.0698579401,0.2486235797,-0.2387923896,-0.0182065833,0.2196348608,0.2118971646,0.0287920404,-0.0389988869,0.0500122719,0.5180223584,-0.0331451371,0.1850028336,-0.1818422824,0.0306759719,-0.1150219291,0.0227195062,-0.107134968,-0.2259223908,-0.182999596,0.1417904198,0.0761193559,-0.1698001325,0.2956995964,-0.238693729,0.0191464704,-0.3457777798,0.4418111742,0.2836984396,0.043562483,-0.7789452672,-0.1735749543,0.3784147799,-0.0504149944,0.0143381096,0.5416885018,0.1780784875,0.2196512669,-0.0907408595,0.2851370871,-0.0125172772,-0.3764247894,-0.1094918624,-0.3844896257]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/866","title":"OSCAR from Inria group","comments":"PR is already open here : #348 \r\nThe only thing remaining is to compute the metadata of each subdataset (one per language + shuffled\/unshuffled).\r\nAs soon as #863 is merged we can start computing them. This will take a bit of time though","body":"## Adding a Dataset\r\n- **Name:** *OSCAR* (Open Super-large Crawled ALMAnaCH coRpus), multilingual parsing of Common Crawl (separate crawls for many different languages), [here](https:\/\/oscar-corpus.com\/).\r\n- **Description:** *OSCAR or Open Super-large Crawled ALMAnaCH coRpus is a huge multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the goclassy architecture.*\r\n- **Paper:** *[here](https:\/\/hal.inria.fr\/hal-02148693)*\r\n- **Data:** *[here](https:\/\/oscar-corpus.com\/)*\r\n- **Motivation:** *useful for unsupervised tasks in separate languages. In an ideal world, your team would be able to obtain the unshuffled version, that could be used to train GPT-2-like models (the shuffled version, I suppose, could be used for translation).*\r\n\r\nI am aware that you do offer the \"colossal\" Common Crawl dataset already, but this has the advantage to be available in many subcorpora for different languages.\r\n","comment_length":43,"text":"OSCAR from Inria group\n## Adding a Dataset\r\n- **Name:** *OSCAR* (Open Super-large Crawled ALMAnaCH coRpus), multilingual parsing of Common Crawl (separate crawls for many different languages), [here](https:\/\/oscar-corpus.com\/).\r\n- **Description:** *OSCAR or Open Super-large Crawled ALMAnaCH coRpus is a huge multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the goclassy architecture.*\r\n- **Paper:** *[here](https:\/\/hal.inria.fr\/hal-02148693)*\r\n- **Data:** *[here](https:\/\/oscar-corpus.com\/)*\r\n- **Motivation:** *useful for unsupervised tasks in separate languages. In an ideal world, your team would be able to obtain the unshuffled version, that could be used to train GPT-2-like models (the shuffled version, I suppose, could be used for translation).*\r\n\r\nI am aware that you do offer the \"colossal\" Common Crawl dataset already, but this has the advantage to be available in many subcorpora for different languages.\r\n\nPR is already open here : #348 \r\nThe only thing remaining is to compute the metadata of each subdataset (one per language + shuffled\/unshuffled).\r\nAs soon as #863 is merged we can start computing them. This will take a bit of time though","embeddings":[-0.0789998397,-0.0689798594,-0.0829853863,0.0613858551,-0.0888234153,0.0940621272,-0.0122487498,0.2985754311,0.0025459034,0.0863161162,-0.5969402194,-0.0512048565,-0.2450727522,0.0754724219,0.1506562084,-0.3321431279,0.0520217121,-0.1897595078,-0.2447093278,-0.2940576375,-0.199752599,0.1900043935,0.1084448248,-0.070598796,0.2079501152,-0.2153771073,-0.434027344,-0.1118402407,-0.1303550303,-0.1936031133,-0.0030159694,0.2002108693,0.2905132473,0.6354877949,-0.0000960657,-0.1057058871,-0.0594696738,-0.239071697,0.0947269425,-0.1166666821,0.2534609437,-0.0826734975,-0.1983726919,-0.3360138237,-0.4045934379,-0.3702588379,-0.0277564283,-0.4808144867,0.1629257798,0.1797285974,0.280683279,-0.227109924,-0.1751101911,-0.2170567811,-0.0248893499,-0.2869288921,-0.3192919791,0.0610993914,0.5420654416,-0.0749456733,0.0386708193,0.380599916,0.1031532437,-0.2377462238,-0.1179090589,-0.0452904068,0.0805145651,-0.3604671061,0.1537543535,0.337295413,0.01977247,-0.342140913,-0.3169185221,-0.1763942689,-0.049127724,-0.0205627326,-0.2518903911,0.3864118159,0.0364644006,0.2546266019,0.2060189247,-0.2413934618,-0.1193951666,0.1010256633,0.031734243,0.5971215963,0.0801759139,-0.1237552091,-0.0247247443,0.0153800361,-0.0398108587,-0.0067003742,0.0270718131,0.1904661953,-0.3079119325,-0.3250470757,0.2066055089,-0.1593383849,0.4159701467,-0.1509889215,0.2482505888,0.1740213186,-0.1795057654,0.1214326546,0.2831611335,-0.0899356827,-0.1744023114,-0.1156539544,0.0723321214,-0.4601128399,0.082082212,0.1743639112,-0.1357317269,0.1313313097,-0.2064629942,0.0541563295,-0.0775346607,-0.0458718017,0.0806471184,-0.053738445,-0.229937017,0.0137936892,-0.0986439809,0.0768383518,-0.0258921981,0.0120433979,-0.0909743235,0.067946665,-0.1071033701,-0.2761722207,-0.2034252882,0.217635721,-0.3801766038,0.128845647,0.172749579,0.0190114211,0.0119643249,-0.2340984046,0.2328408957,0.2133850306,0.3611582518,-0.320294857,0.1609599441,-0.1494008154,0.2117096037,-0.2598241866,0.0502515323,-0.2192108184,-0.4714737535,-0.02956287,0.0450573191,-0.3023353219,-0.1582363695,0.2464541197,-0.231522426,-0.1368895769,-0.2449243069,0.7459754348,0.007385409,0.1247611344,-0.0491997711,-0.0011200133,-0.3634642661,-0.0202040933,0.0891351327,0.2649673522,0.1860181391,-0.0088046864,0.0476065725,0.1841085404,0.1200442836,0.3375632167,-0.1249205098,-0.1493848115,0.0020371596,0.1753962636,0.2010474354,-0.2509192228,0.0566660091,0.0413590707,0.0060923966,-0.0718067288,0.1888000369,-0.1069103032,0.3648880124,-0.3289222419,0.1453191936,0.6325167418,-0.246833697,0.1829524487,-0.2372281253,-0.0358664356,-0.0515489802,0.266964227,0.3047321737,-0.3177728355,0.1110945493,-0.3176100552,0.1368640959,-0.1907478273,0.2715631127,-0.019909827,0.1988722235,0.0542283766,0.2405931354,-0.4054969251,-0.0055573247,0.0321984887,-0.0676580444,0.5373077989,-0.0850452781,-0.3493946791,-0.1995297819,-0.2065919489,-0.0381338969,-0.0213455074,0.2677493393,0.3026916981,0.0885429606,0.0233812835,-0.1373600811,-0.0861660987,-0.1342484504,-0.3086870313,-0.0033678757,0.3649312854,-0.092291303,-0.0811056048,0.2129443586,0.4055532217,-0.1997703165,-0.2372059375,0.1052786708,0.1152985618,-0.1646990031,0.0917293653,0.5061905384,0.3447551131,0.1893038005,-0.5001240969,0.1349046379,0.114430137,0.1833438426,-0.0649386346,-0.1615116894,0.2966725528,0.1831851155,-0.0219672918,0.2072990388,-0.0891565457,0.3712120354,-0.0440208018,-0.0954973996,-0.0847255811,0.0343358442,-0.01430998,0.0468290932,0.1122884527,-0.5806649327,0.1995316148,0.5494887829,0.0785557106,0.3099135458,0.1553822309,-0.0023567595,-0.1778399795,-0.0857951343,0.0435632914,0.1141144857,0.3094471693,0.3973644674,0.0615208037,0.1920935512,-0.0087894751,0.173266083,-0.1201455519,0.2182255983,0.2752735019,0.2165376097,0.0397535898,-0.2015638053,-0.1670654267,0.086631842,0.1049868613,0.3335330784,0.1044118851,0.162094906,-0.2809582949,-0.2595311105,-0.1624915302,-0.1026665196,-0.0025337348,0.2718780935,-0.28386724,-0.1953326166,0.25302881,-0.0899115428,0.3012341261,-0.1137074754,-0.0297509097,0.1293421686,-0.4761375785,-0.0963706374,0.2489105463,0.2857629955,0.0243882071,0.4065279365,0.1019363329,-0.0001941636,-0.0554574728,-0.4803338349,0.197908029,-0.2193891853,-0.1993477046,-0.0839121789,-0.3282859921,-0.0089863446,-0.2560660541,0.1163565591,-0.1906362772,-0.1358748525,-0.02395395,-0.1252659708,-0.0306453183,-0.1082518473,-0.6249163151,-0.5535086989,-0.3931676745,0.1162367687,-0.0869350731,0.1905634701,-0.2222539932,-0.2236909419,-0.0532186143,0.2386676222,0.1823453456,-0.2849721313,-0.055927936,0.2615026534,-0.3868191838,-0.2336463481,-0.2682065368,-0.3120085001,0.2669037282,0.1276145428,-0.0902532935,0.0230467692,-0.0722815096,-0.3246203661,0.2294089049,-0.0798891932,0.2951599956,0.2661080062,-0.2085749805,-0.0683946311,-0.0480041467,0.0669269115,0.1893723011,-0.0352248847,-0.069174476,0.2103768736,0.241316095,0.3895183802,-0.0755582377,0.1789115518,0.3257670999,0.3153468966,0.0676872954,-0.0875455365,-0.200642854,0.1844980568,0.185775578,0.0273675509,0.5681686997,0.2440765798,-0.7737677693,-0.2368361652,-0.0665207282,-0.2774050832,-0.2823314667,0.2267484814,-0.0064303912,0.1672701836,-0.0694505349,-0.3338105381,-0.2970200181,-0.0867692232,0.1868282706,0.0179583188,-0.0448423438,0.1390143186,-0.1333761215,0.1693529487,-0.3628246784,0.1929843724,0.2051903307,-0.0744497478,-0.0547928736,-0.1376662701,0.0149120037,-0.0647610277,-0.0724482983,-0.3927128911,-0.0003968728,0.1283618808,-0.3000940084,0.0754092336,0.0153135043,-0.2824622095,0.0676113218,0.3900348842,-0.0661965013,-0.4734046459,-0.2287934721,0.1623782665,-0.0098263556,-0.098170124,-0.1904400885,-0.0223836321,-0.4972299039,-0.0556998327,0.1005588099,0.1771858335,0.1592842937,0.0268543214,-0.026848007,-0.114692688,-0.0503914729,0.3639598787,0.1062973365,-0.1620108932,0.354858011,0.1845196635,-0.1037703529,0.2475249171,0.2776955664,0.3752758205,-0.2040381879,-0.2674690485,0.2303122282,-0.4009726346,0.3091101348,0.0440560319,0.3894363344,0.2537849844,-0.1759382337,-0.0548438393,-0.1668071002,-0.0963999555,-0.0252583306,0.247674033,-0.4372764826,-0.6188765764,0.3124141395,0.1915918887,-0.2940199375,0.325756073,0.1227358356,-0.4092984796,0.6036086679,-0.0085063856,1.00682199,-0.2064820677,0.019268116,-0.0757431462,0.2425942868,0.1657004654,-0.3214389682,-0.0249558073,-0.0408603922,-0.1627236307,0.1012695357,0.179309532,0.0055644293,0.1615138501,-0.2621531188,0.0929739475,0.0894047692,-0.0839616582,0.0868630931,0.3980652392,0.0476872548,-0.3139051795,-0.2840451002,0.2409484982,-0.0037068308,-0.0164636374,0.0375621021,-0.1747716367,-0.0627153367,0.3782950044,-0.4576453865,0.0030915115,0.2986183763,0.0420789123,-0.1318152547,-0.2317119837,0.5097091794,0.0554343127,0.4480832517,-0.1488118172,-0.0856939852,0.0614238493,-0.199829638,-0.1961732507,0.1059418619,-0.1727459431,0.4053301811,-0.0920225084,-0.3007202446,0.2931477427,0.3604834676,-0.1609971821,-0.2587643266,-0.0515532494,0.0858496204,0.0663367212,0.188725397,0.3918248713,-0.3024207056,-0.0377420895,0.2599419951,0.2304695547,-0.2249955535,0.1714369953,-0.0128719267,-0.1304267049,-0.0046184487,-0.0113621205,0.5888926983,-0.1910806745,0.0283400305,0.2820191979,-0.0597929098,-0.2995474339,-0.2678392231,0.1736902297,0.0272039585,-0.0603510141,-0.1044183895,-0.240527764,0.1992271096,-0.0005390092,0.2522384226,0.1426198632,-0.0704540163,0.0334543288,0.1015530452,0.1490963697,-0.0322877131,0.018511517,-0.1188637093,0.047874324,0.068877548,0.3429550827,-0.5057381392,0.0078538582,-0.2016778737,0.2185337991,0.0528443642,-0.0939101726,-0.0618631728,0.1001634002,0.117612794,-0.0188256893,-0.1256082356,-0.2954027057,-0.1370369792,0.0935640186,-0.089944154,0.2373555005,0.1993775815,-0.0618282966,-0.1603852957,-0.3016292453,0.1719177812,0.1017071977,0.0034259646,0.0895932615,0.3138928711,-0.1009608209,-0.0453012548,-0.0164861679,0.0591555126,0.1452383995,-0.0343684889,0.1639671028,0.0371304713,-0.226897493,0.1959364712,0.2424283177,0.0517467223,0.1004216522,0.2397506833,0.2974244356,0.1709429175,0.3149266243,0.1969351321,-0.1982324123,0.014286832,-0.0125678247,0.0564443693,0.3172782958,-0.372684747,0.0981694162,0.1270092726,0.0227463003,0.3888136446,0.3959035873,0.2513133883,0.1040050089,0.0570982955,-0.0961457342,0.2420562059,0.0094872331,0.2537920773,-0.1055883616,0.1361368597,-0.0255210288,0.1322354376,-0.0362820104,0.4412903488,0.1181591675,-0.1980966181,0.4327004254,0.3672033846,0.0002728836,0.1847949475,-0.0257602427,0.1654509902,0.2912642658,0.2248007655,0.1123098433,0.0766996518,0.4137377739,0.1545703709,0.0604716763,-0.1232117042,0.3448936939,0.1426739544,0.1009161398,-0.2385988683,-0.185368225,-0.3306390047,-0.0049492321,-0.1086821258,-0.1603017002,0.2862168252,0.012314464,0.1106619909,-0.0545540936,-0.0300406571,0.0132005159,0.3707948923,-0.1473266631,0.263261646,0.3792475164,-0.3474470079,-0.1610162705,0.4022833407,-0.0734456703,0.0082244268,-0.330733031,0.3771045804,-0.016593501,-0.0604395978,0.001569006,0.1722233146,0.1657084078,-0.1233995706,0.1630363613,0.1844747514,-0.1700631529,-0.0438157693,0.1857714057,-0.147670418,0.1212134734,0.2190967649,0.210152775,0.1303414255,-0.1294061095,0.0292535163,-0.5958346128,-0.2451606989,0.299020797,-0.0082033025,0.012321,-0.1301578879,0.121096395,0.200407356,0.4456836879,0.1922936291,0.2045442313,-0.0138139073,-0.3460650146,-0.3378744423,0.2569002509,-0.1420868933,0.0148800621,-0.2077053487,0.2883600295,0.4104836583,0.320004195,-0.3424785137,0.3709461689,0.0179372821,0.0065466538,-0.5773762465,0.1793444008,-0.0044489861,0.1575571895,0.0918135121,-0.0819355622,-0.1166497692,-0.1669201404,0.2160651535,-0.2341342121,-0.1001128256,0.0029197061,0.1120485663,0.1357707679,0.0802974254,0.2219418883,-0.0308195204,-0.0328293815,0.071403414,-0.3288218975,-0.1540770531,0.3053357899,0.1469471008,-0.1198242307,-0.2654133439,-0.3528785408,0.2071141154,0.2779453695,-0.1227898821,0.0493452549,-0.2114328742,-0.0391650647,0.0466241948,-0.1417345852,-0.1302024573,0.3917131126,-0.0472890735,-0.0689288154,-0.0754265338,-0.2966424525,0.2450676709,-0.1027344987,-0.3108024597,-0.3666956425,0.0727659091,0.2543421686,0.2759603262,-0.2505302429,0.2675186396,0.2665630579,0.2953965962,-0.081642203,0.1215063035,0.0683361739,-0.2200340629,-0.2534703612,-0.1043637246,-0.0112960748,-0.2040100098,-0.3458048701,-0.2720399797]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/866","title":"OSCAR from Inria group","comments":"Grand, thanks for this!","body":"## Adding a Dataset\r\n- **Name:** *OSCAR* (Open Super-large Crawled ALMAnaCH coRpus), multilingual parsing of Common Crawl (separate crawls for many different languages), [here](https:\/\/oscar-corpus.com\/).\r\n- **Description:** *OSCAR or Open Super-large Crawled ALMAnaCH coRpus is a huge multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the goclassy architecture.*\r\n- **Paper:** *[here](https:\/\/hal.inria.fr\/hal-02148693)*\r\n- **Data:** *[here](https:\/\/oscar-corpus.com\/)*\r\n- **Motivation:** *useful for unsupervised tasks in separate languages. In an ideal world, your team would be able to obtain the unshuffled version, that could be used to train GPT-2-like models (the shuffled version, I suppose, could be used for translation).*\r\n\r\nI am aware that you do offer the \"colossal\" Common Crawl dataset already, but this has the advantage to be available in many subcorpora for different languages.\r\n","comment_length":4,"text":"OSCAR from Inria group\n## Adding a Dataset\r\n- **Name:** *OSCAR* (Open Super-large Crawled ALMAnaCH coRpus), multilingual parsing of Common Crawl (separate crawls for many different languages), [here](https:\/\/oscar-corpus.com\/).\r\n- **Description:** *OSCAR or Open Super-large Crawled ALMAnaCH coRpus is a huge multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the goclassy architecture.*\r\n- **Paper:** *[here](https:\/\/hal.inria.fr\/hal-02148693)*\r\n- **Data:** *[here](https:\/\/oscar-corpus.com\/)*\r\n- **Motivation:** *useful for unsupervised tasks in separate languages. In an ideal world, your team would be able to obtain the unshuffled version, that could be used to train GPT-2-like models (the shuffled version, I suppose, could be used for translation).*\r\n\r\nI am aware that you do offer the \"colossal\" Common Crawl dataset already, but this has the advantage to be available in many subcorpora for different languages.\r\n\nGrand, thanks for this!","embeddings":[-0.1445858032,0.016484851,-0.1031186804,-0.00934033,-0.0510738678,0.1384983957,0.0371580943,0.2859366536,0.0971412733,0.1239481419,-0.4872651994,0.0026186323,-0.2676612437,0.1161381379,0.2122641951,-0.3277106881,-0.0200689081,-0.1677759737,-0.2419612557,-0.3077515066,-0.2028297782,0.1192580983,0.1390772909,-0.0513661876,0.1979219168,-0.2105436772,-0.3486982584,-0.1936535835,-0.0885706693,-0.2174150497,-0.0340564512,0.1081350297,0.264904201,0.6240710616,-0.0000967589,-0.0597904548,-0.0014506138,-0.1695243567,0.0683292449,-0.2048475295,0.2476925403,-0.1030816138,-0.2127399296,-0.289398402,-0.4396190345,-0.2817305028,0.0654796585,-0.4842118621,0.1355609149,0.1664846838,0.2888611555,-0.2562481761,-0.2101947665,-0.2617106438,-0.0527719483,-0.324524343,-0.3707594275,0.0598616973,0.5367722511,-0.0559340194,0.0862058103,0.3613136709,0.0600395165,-0.2086071819,-0.1219967306,-0.0668849498,0.0904935151,-0.308655113,0.138198331,0.4060927927,0.0226227734,-0.293243289,-0.2717109025,-0.197235018,0.0629564002,-0.0253641624,-0.3164671361,0.4696978331,0.0074475114,0.3794468045,0.1585885286,-0.3281267285,-0.1443434954,0.2026724964,-0.0913020149,0.485823065,0.0840211138,-0.0736347213,-0.0302360076,0.129370138,-0.1137201041,0.0156790689,0.0724136829,0.1774158925,-0.3323958516,-0.3173817694,0.2251686752,-0.1196720302,0.4310801923,-0.1872955412,0.2686153054,0.1515803635,-0.1656484008,0.1314605474,0.3548539877,-0.1252739429,-0.2485191822,-0.140223965,0.170664981,-0.4276833832,0.0600685291,0.1763889194,-0.0987589881,0.1996543854,-0.1489375979,0.0324692205,0.0032796827,-0.0486624762,0.0221459046,0.0461629219,-0.2402831167,0.0245909579,-0.1517216563,0.1192118227,-0.0111214099,0.0430792756,-0.0991123915,0.0953590274,-0.0726402029,-0.303416729,-0.187131837,0.2235997468,-0.358305335,0.1634706259,0.1736025661,0.0233962238,-0.0341855623,-0.2246085107,0.2925863862,0.1596542001,0.3738742769,-0.3004074097,0.1316725314,-0.1726049334,0.1738442332,-0.2080658674,0.0994779915,-0.1797598451,-0.4599420726,0.0721159056,-0.0723252222,-0.2930959463,-0.1162974685,0.2389659733,-0.2222577333,-0.2172658145,-0.1892539114,0.6963754296,-0.0376748368,0.1191370338,0.083238475,0.0040602922,-0.3908231556,0.0045028189,0.0170208272,0.2493008524,0.1737061888,-0.0906266198,-0.0049576242,0.2804631889,0.1302437335,0.3575429916,-0.1491763741,-0.1267976314,0.0366629362,0.160831362,0.2407505363,-0.2303552628,-0.0149251949,0.0860929862,-0.0899991989,-0.0687466264,0.1904790998,-0.0505736768,0.306760639,-0.2740561068,0.156795904,0.641854465,-0.1818187982,0.1776449531,-0.2068531811,-0.0548980273,-0.0056405682,0.3230901957,0.3055527508,-0.311732769,0.1383364499,-0.2599675357,0.1598313451,-0.1982628405,0.243225053,-0.0015655924,0.228230387,-0.0312549286,0.1705017686,-0.3139799535,0.0175502282,0.0251002535,-0.0819569826,0.4946747422,-0.0188827273,-0.3380518258,-0.212963894,-0.2357354015,-0.0776234493,0.0126833366,0.220571056,0.2531768084,0.0415374227,0.1310700178,-0.0924139097,-0.0666417554,-0.1640547216,-0.3190417588,0.0196606684,0.3861542046,-0.0701467022,-0.0534980558,0.1425224096,0.4113716781,-0.2831939459,-0.287545979,0.1034685597,0.0394168757,-0.2475566268,0.1497770101,0.3936319649,0.3910166621,0.2580038309,-0.5360629559,0.0957437083,0.1707190722,0.1508589238,-0.0477024838,-0.1522431523,0.3210076392,0.248837024,-0.0397803262,0.1438562274,-0.0989069268,0.3314937651,-0.0863516405,-0.0989320353,-0.0821596757,0.1419727057,-0.0240692571,0.0884431601,0.1244020611,-0.6120539308,0.1538811624,0.6174884439,0.171649456,0.4529703856,0.1347225755,0.0011205048,-0.1140217409,-0.0942676663,0.0270402282,0.1720230281,0.312877506,0.333969146,0.1024120897,0.2660679817,-0.0302466303,0.1338129193,-0.1663380265,0.2226225734,0.3287149668,0.2378990054,0.0036039387,-0.1862039268,-0.1111879796,0.0787485316,0.0660390109,0.3761610985,0.0991128981,0.1619069129,-0.241372332,-0.2768630087,-0.1323563308,-0.092607893,0.0306355525,0.3057220578,-0.2812845409,-0.1388320923,0.2140440345,-0.2493514568,0.3250258267,-0.1305247694,-0.1342095882,0.1186188832,-0.4936399162,-0.0721784085,0.2454951853,0.1697326005,-0.1055274457,0.386957854,0.1704349518,0.0150791248,-0.0705282763,-0.4740937054,0.207608521,-0.1882716119,-0.2048534602,-0.0744485483,-0.293040216,-0.067681551,-0.2161619663,0.1208323315,-0.3457308412,-0.0863053501,-0.0465815328,-0.1104046032,0.0136717949,-0.0990703627,-0.662391305,-0.4860935807,-0.4594108164,0.125407964,-0.109137699,0.0265706107,-0.1947520375,-0.1571363211,-0.0786314383,0.2422841936,0.1922285855,-0.3297122121,-0.0518235117,0.2705526054,-0.3627394736,-0.2344500571,-0.3599881828,-0.2991578877,0.2228305489,0.1197379231,-0.2133834809,0.0499425605,-0.1330438703,-0.3288289607,0.1673636883,-0.1026222184,0.2277552783,0.1971339881,-0.2268678993,-0.0615539066,-0.0205542594,-0.0007756046,0.2694598734,0.0143564539,-0.0222032052,0.241910547,0.268961072,0.3808377683,-0.1451183259,0.1867309958,0.4341460466,0.2888916433,0.09986642,-0.19208996,-0.2322518229,0.1998123974,0.2050095797,-0.0000312325,0.5013803244,0.2287256867,-0.788174808,-0.1793417633,-0.1913844347,-0.1969180554,-0.2960776091,0.2714386284,0.0881389529,0.1826538295,-0.0913858041,-0.3516804278,-0.2225639373,-0.1838569641,0.2425142527,-0.0106536448,-0.0434073359,0.1631295383,-0.1429233849,0.1312039793,-0.4391806722,0.1842797995,0.1886852235,-0.0313511416,-0.156327486,-0.1529278904,0.0669949651,-0.005520381,-0.0440397076,-0.3390207589,-0.041638501,0.0628244057,-0.2689154446,0.0839894712,0.0484985895,-0.2936020195,0.1112798899,0.3255601525,-0.1141980588,-0.4281307757,-0.1339552253,0.1858039349,-0.0046106572,-0.1323847324,-0.1886339784,-0.0306441896,-0.4565509558,-0.116736412,0.1545260549,0.2376204133,0.1856621355,0.0931764692,-0.0231464412,-0.092398122,-0.0083168978,0.3479495943,0.0390034318,-0.2274333239,0.4003280997,0.1534712315,-0.1179183647,0.2055165321,0.1923165619,0.4146375656,-0.2164549828,-0.3111155331,0.2346285135,-0.4457848668,0.415355742,0.0595133342,0.408198595,0.2835743129,-0.1605643481,-0.0666455254,-0.1404797882,-0.0926352963,-0.0563353971,0.1705968231,-0.5125516653,-0.6277090311,0.2743141055,0.2264434844,-0.2456716001,0.3576985002,0.1557896733,-0.3822216988,0.4642174542,-0.1078245044,0.9954947829,-0.1772863716,0.1374628842,-0.1453820616,0.2498279661,0.17723158,-0.2784645855,0.0109646525,-0.0685996413,-0.1123351082,0.0774536654,0.1417655349,0.0351982452,0.1755021363,-0.2676954269,0.1381658018,0.0718989,-0.1495690346,0.0453148149,0.4261248112,-0.0287379138,-0.2648002803,-0.2675659359,0.2217811197,0.0110488245,0.0580643974,0.000365538,-0.115742445,0.0547513776,0.3314219415,-0.4564264119,-0.0240350924,0.3056653738,0.0505652688,-0.0804992989,-0.3284572661,0.5977244377,-0.0501499102,0.324072957,-0.102749899,-0.0638605803,0.064030692,-0.2806101143,-0.1777107418,0.0932473093,-0.1812385321,0.3702580929,-0.1091510579,-0.337517947,0.3809428215,0.3500444591,-0.2528614402,-0.2582578361,0.040642187,0.0903496072,0.0765040964,0.1210619435,0.4218381047,-0.24272798,-0.0684977621,0.2593055069,0.2026486248,-0.1217351034,0.0913833305,0.066779919,-0.2427987605,-0.0384005345,0.0208900254,0.5612045527,-0.1932151467,0.1064482853,0.2140948921,-0.0148233511,-0.3455882967,-0.1730260998,0.2961411178,0.0027231942,-0.0424858294,-0.0884865075,-0.1630763263,0.2121956646,-0.0011760732,0.1917809844,0.0859177858,-0.0975163653,0.0226650797,0.2450116426,0.1200327501,-0.0044701439,0.0220102612,-0.0545468479,0.0055212737,0.0895166323,0.1702607274,-0.4833020568,0.0293092448,-0.1697764993,0.1930232048,0.0293356255,-0.0412312523,-0.1005541757,0.1259784997,0.1051170304,-0.0616782084,-0.139346689,-0.3243767619,-0.1957515478,0.0993182287,-0.0954288542,0.3085191846,0.2305142432,-0.0768186375,-0.0865974277,-0.2524496019,0.2147750705,0.1750876307,0.0074499049,0.0932724923,0.2528125048,-0.193242684,-0.076216504,0.0374025851,0.0104800453,0.1862716675,0.0073816734,0.1778144985,0.1316207051,-0.1654668152,0.128538087,0.220834896,0.0902260542,0.1298959404,0.1588992923,0.2710493803,0.1962890923,0.3606760204,0.1180472374,-0.2469601929,-0.0957260206,0.0437305346,0.0772786438,0.3159349263,-0.3316794932,0.0182804558,0.1367146373,0.0447512865,0.3985548913,0.4019498527,0.3228544891,0.1140583158,0.0536393151,-0.1485764831,0.2540562153,0.0215741396,0.2547355592,-0.1186730415,0.137586236,-0.0889369249,0.1455702037,-0.0784622878,0.3785454333,0.0670762062,-0.2016979903,0.3752337694,0.4323550463,0.014296459,0.1547323614,0.0330990814,0.2067587376,0.2224243581,0.1097240672,0.1216400564,0.1183817685,0.4285281301,0.1841545999,0.1259358376,-0.1901759803,0.3173555434,0.185716033,0.0828518718,-0.1668772697,-0.1437027007,-0.3310287893,0.0439597368,-0.0711804628,-0.1567970961,0.3402096927,0.0663632154,0.1044912413,-0.0900916159,0.0655655414,0.0114736604,0.3465658724,-0.1737174988,0.2030332088,0.2755797505,-0.3496954441,-0.1292739362,0.3973204195,-0.0489212349,-0.094808504,-0.2490786761,0.296181798,-0.09254767,0.009775524,-0.0733742937,0.1399688721,0.1784007847,-0.1987353265,0.1345016658,0.1930098981,-0.159701854,0.0415743366,0.2192446291,-0.1346370131,-0.0307876971,0.2232432663,0.2015758902,0.2033009678,-0.226861313,0.0979544818,-0.6043023467,-0.1542328894,0.4349019825,-0.0984693393,-0.0444815606,-0.1777411848,0.1217633709,0.1516460627,0.3734312952,0.2057112902,0.2305817306,-0.0356981568,-0.2548215389,-0.308655709,0.3644093871,-0.1924664825,0.117240034,-0.1977429539,0.3131814301,0.3301109374,0.4188324213,-0.2654492855,0.3270304799,-0.1181362569,0.004330501,-0.5115629435,0.2137742937,0.102616787,0.2079676986,0.0486172549,-0.0754549578,-0.1284623742,-0.1868003309,0.189596653,-0.305587858,-0.1391993314,-0.083169274,0.0972729623,0.2352818996,0.0434098765,0.1591630876,0.0021873708,-0.0028714826,0.0562832244,-0.3959495723,-0.1000919044,0.2860251069,0.1439290345,-0.1969027519,-0.2754175365,-0.394536674,0.2587830722,0.2915534079,-0.1231786832,0.0516428649,-0.1919390559,-0.083309561,0.0089204879,-0.126779899,-0.2465782464,0.3960870504,-0.0539318547,-0.1439455599,-0.0038603554,-0.3008363247,0.2342068255,-0.1358793229,-0.3669730723,-0.3989069164,0.094960399,0.1821185201,0.2074428052,-0.2814953625,0.2637113631,0.2803593576,0.2846938074,-0.0973827913,0.1487986892,0.0439326614,-0.1878337264,-0.2188906372,-0.0391786881,-0.0890343115,-0.2112196088,-0.2771580517,-0.2465275973]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/865","title":"Have Trouble importing `datasets`","comments":"I'm sorry, this was a problem with my environment.\r\nNow that I have identified the cause of environmental dependency, I would like to fix it and try it.\r\nExcuse me for making a noise.","body":"I'm failing to import transformers (v4.0.0-dev), and tracing the cause seems to be failing to import datasets.\r\n\r\nI cloned the newest version of datasets (master branch), and do `pip install -e .`.\r\n\r\nThen, `import datasets` causes the error below.\r\n\r\n```\r\n~\/workspace\/Clone\/datasets\/src\/datasets\/utils\/file_utils.py in <module>\r\n 116 sys.path.append(str(HF_MODULES_CACHE))\r\n 117 \r\n--> 118 os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n 119 if not os.path.exists(os.path.join(HF_MODULES_CACHE, \"__init__.py\")):\r\n 120 with open(os.path.join(HF_MODULES_CACHE, \"__init__.py\"), \"w\"):\r\n\r\n~\/.pyenv\/versions\/anaconda3-2020.07\/lib\/python3.8\/os.py in makedirs(name, mode, exist_ok)\r\n 221 return\r\n 222 try:\r\n--> 223 mkdir(name, mode)\r\n 224 except OSError:\r\n 225 # Cannot rely on checking for EEXIST, since the operating system \r\n\r\nFileNotFoundError: [Errno 2] No such file or directory: '<MY_HOME_DIRECTORY>\/.cache\/huggingface\/modules'\r\n```\r\n\r\nThe error occurs in `os.makedirs` in `file_utils.py`, even though `exist_ok = True` option is set.\r\n(I use Python 3.8, so `exist_ok` is expected to work.)\r\n\r\nI've checked some environment variables, and they are set as below.\r\n\r\n```\r\n*** NameError: name 'HF_MODULES_CACHE' is not defined\r\n*** NameError: name 'hf_cache_home' is not defined\r\n*** NameError: name 'XDG_CACHE_HOME' is not defined\r\n```\r\n\r\nShould I set some environment variables before using this library?\r\nAnd, do you have any idea why \"No such file or directory\" occurs even though the `exist_ok = True` option is set?\r\n\r\nThank you in advance.","comment_length":34,"text":"Have Trouble importing `datasets`\nI'm failing to import transformers (v4.0.0-dev), and tracing the cause seems to be failing to import datasets.\r\n\r\nI cloned the newest version of datasets (master branch), and do `pip install -e .`.\r\n\r\nThen, `import datasets` causes the error below.\r\n\r\n```\r\n~\/workspace\/Clone\/datasets\/src\/datasets\/utils\/file_utils.py in <module>\r\n 116 sys.path.append(str(HF_MODULES_CACHE))\r\n 117 \r\n--> 118 os.makedirs(HF_MODULES_CACHE, exist_ok=True)\r\n 119 if not os.path.exists(os.path.join(HF_MODULES_CACHE, \"__init__.py\")):\r\n 120 with open(os.path.join(HF_MODULES_CACHE, \"__init__.py\"), \"w\"):\r\n\r\n~\/.pyenv\/versions\/anaconda3-2020.07\/lib\/python3.8\/os.py in makedirs(name, mode, exist_ok)\r\n 221 return\r\n 222 try:\r\n--> 223 mkdir(name, mode)\r\n 224 except OSError:\r\n 225 # Cannot rely on checking for EEXIST, since the operating system \r\n\r\nFileNotFoundError: [Errno 2] No such file or directory: '<MY_HOME_DIRECTORY>\/.cache\/huggingface\/modules'\r\n```\r\n\r\nThe error occurs in `os.makedirs` in `file_utils.py`, even though `exist_ok = True` option is set.\r\n(I use Python 3.8, so `exist_ok` is expected to work.)\r\n\r\nI've checked some environment variables, and they are set as below.\r\n\r\n```\r\n*** NameError: name 'HF_MODULES_CACHE' is not defined\r\n*** NameError: name 'hf_cache_home' is not defined\r\n*** NameError: name 'XDG_CACHE_HOME' is not defined\r\n```\r\n\r\nShould I set some environment variables before using this library?\r\nAnd, do you have any idea why \"No such file or directory\" occurs even though the `exist_ok = True` option is set?\r\n\r\nThank you in advance.\nI'm sorry, this was a problem with my environment.\r\nNow that I have identified the cause of environmental dependency, I would like to fix it and try it.\r\nExcuse me for making a noise.","embeddings":[-0.2681194544,0.1308584958,0.0091978218,0.1722479761,0.2753852904,0.0738350973,0.2072007209,0.1833841354,0.1528996229,-0.103595838,-0.2849372625,-0.1424152702,-0.1180940643,-0.3181892335,0.030698495,0.1511705667,0.2635430694,0.3381048739,-0.3553359807,-0.2657526731,-0.2365347594,0.1956792325,0.0699337795,0.0383725874,-0.3075052202,0.1586766541,-0.1923209578,0.3546025157,-0.0874144137,-0.3079354763,0.312802583,0.0767817199,0.2757256031,0.7110565901,-0.0001215231,0.1625322849,0.1903069764,-0.0628716424,-0.4191234708,-0.3236556053,-0.089015618,-0.3284656107,0.1100215167,-0.1870281249,0.1990170926,-0.2734282017,0.0801438391,-0.2908483744,0.499045819,0.3661040962,0.1287855655,0.3562702835,0.158652395,-0.0425901823,0.1484103799,0.4451641142,-0.0781542957,-0.0663885251,-0.0085933041,-0.1054819077,0.2778817713,0.0019224144,-0.1829766482,0.0140808029,0.4155776203,-0.0752482414,0.0377338007,-0.3184424639,0.0598807037,0.2000163943,0.7275674343,-0.4121959507,-0.7552509308,-0.1253420413,-0.1917922646,-0.1638842523,0.2845663726,-0.0826183408,0.0983769298,0.2858668268,0.0040534018,-0.344030112,-0.1661034375,0.3062963784,-0.0775494576,0.1118347645,0.0083073387,0.0248556882,0.2521101534,-0.189615652,0.3435187638,-0.34113428,0.0948805362,0.3831880391,-0.3209574819,-0.0606265664,0.0867931843,0.1025330126,-0.066327408,0.0996712446,-0.4005199969,-0.0226232875,-0.2901917994,0.1602175683,0.1484983265,0.2043939829,-0.0641826391,0.3525402546,0.2106776237,-0.1175974756,0.1762677282,-0.0752191767,-0.2436424047,-0.2041593492,0.0367049463,-0.07335186,0.1752852947,-0.3583666682,-0.0762620121,0.0499932393,-0.0009142692,0.0625434071,-0.0196596943,0.3692307174,-0.0863489211,0.1779126227,0.1272700727,0.336876899,-0.2937982082,0.1131299511,-0.1551246345,0.3129038215,-0.2285295278,0.0236860625,0.292273283,-0.0684456974,0.2617128491,-0.120366618,0.3084441125,0.246784091,0.0894458815,-0.2138560414,0.0618057363,0.600042522,-0.2023255676,0.4967263043,0.3052872717,-0.2590166032,-0.1908972561,-0.0503731258,-0.1729454249,-0.2369495183,-0.1975753754,0.090082705,-0.4204051793,0.0599871017,-0.1369484663,-0.4408088326,0.287068665,-0.2488075197,-0.0929095671,-0.274868995,0.2901176512,-0.1786625534,0.163043052,0.7134696841,-0.4537554085,-0.0906329527,-0.4043835998,-0.0830353349,-0.1028399915,0.3096979558,-0.1342532337,0.1072888374,-0.2472750247,-0.1418412924,0.2915894985,-0.4425392151,-0.3745623231,-0.0354103893,0.0048852852,0.0591549091,0.1987728775,-0.0395493209,0.1241699755,-0.2175826728,-0.1797391921,0.2130536437,0.030654503,-0.1623436958,-0.1035052165,-0.2806387842,0.0464729294,0.3110309839,0.4101851583,0.0462240204,0.0579210073,0.1423965693,-0.0716323927,-0.0558816716,0.1470709443,0.5463572741,0.4187244177,0.2150306851,0.1226576716,-0.0435194634,-0.2793141007,0.3250387311,0.0944105759,-0.0312895589,-0.3206380308,0.0869084224,-0.1255264282,0.0822432265,-0.2658863664,-0.0612874404,0.0229207948,0.0132843051,0.042952776,0.0041286382,-0.1376996487,0.5887189507,0.0841866434,0.2985939384,-0.3340809643,0.4976209402,-0.2197433263,-0.0243328996,-0.2272533327,0.2132330835,0.0356290005,-0.09588743,-0.1653893888,0.2277174294,0.0509410389,0.0403989963,-0.0894364864,0.4067900479,0.1685875058,-0.3530105352,0.3245938718,0.2188595831,0.0722006634,0.0125812823,-0.1360511929,-0.1728676111,-0.1118316501,0.2684084773,0.0629780516,0.0753322914,0.159749493,-0.0182622466,-0.0926279575,-0.297341764,0.2764791846,0.1400915086,0.5950938463,0.0195134785,0.2839184999,-0.0588102527,0.0648782477,-0.0461956933,0.0504941344,0.029605398,-0.2449672669,0.0941413566,0.2393054068,0.3468308449,0.5564562082,0.1072781309,-0.1874959618,0.104786694,-0.1752255857,0.0151931047,0.0236833971,0.1832286865,0.1447095424,0.1225250736,-0.1115251556,0.0541960895,-0.1200408712,-0.5026333332,0.0624662638,0.3078273833,-0.5334969163,0.214196831,-0.3632020056,-0.1724309027,-0.1713004112,-0.0200897194,-0.2282012403,-0.3269524276,-0.1747828424,0.1958208531,-0.0569681153,0.1376822889,-0.0231506284,-0.1850917339,-0.1616015285,-0.5011033416,-0.1654448956,-0.1189043298,-0.0615781955,-0.0852289796,0.1452154219,-0.0186589975,0.1159555316,-0.106114693,0.1091464683,-0.3176578581,-0.3510705531,0.0525581427,0.0505925156,0.3023043573,0.125305593,0.2575555146,0.1992651969,-0.2578475773,0.1370526105,0.0607150346,-0.0799539387,0.0862371251,0.0300319437,-0.3280070126,-0.1532604992,-0.1224608868,-0.4666886032,-0.4357229471,-0.2288575917,0.0394917913,0.070153892,0.6503170729,0.2038682848,0.1293322593,0.108779259,-0.0912211537,-0.0634966046,-0.0144536514,0.5791630149,-0.2120685279,-0.2635461092,0.0121335378,-0.1841260195,0.2727899849,0.1857423633,-0.2984972298,-0.0178074744,-0.1328408718,0.138414979,0.0182921272,0.3956168294,0.4145404994,0.1749931872,0.0981788114,-0.2287124693,0.0139991287,-0.1211025566,0.1775335371,0.3473450243,-0.1626528949,0.3198258877,-0.293930918,0.8454340696,0.0806806609,0.0163488109,0.416377306,0.1902275085,0.7380416989,-0.0096346894,-0.473370105,-0.0255538244,-0.0094589153,-0.2399977148,-0.0010805603,-0.0055481135,-0.1845030785,-0.1395790726,0.1722677797,-0.1525453478,-0.1586577743,-0.0561994053,-0.0725013539,0.1048395559,0.1237601712,0.0919111148,-0.2340692133,-0.0683836937,0.1189659983,0.5389794707,-0.1123923361,0.2590022981,-0.4309729636,-0.3115486205,-0.1010295749,0.1086865887,-0.1497281045,0.3363617659,-0.1191828996,0.1415855438,-0.1787700802,0.1255773604,0.4687953591,-0.2233209908,-0.3136986494,0.2532336414,-0.0265995283,-0.55150491,-0.2780077457,0.0786727071,0.1461157203,-0.0212887116,0.1174365208,-0.3697504997,-0.2067412287,-0.0211539585,0.4966008961,-0.1066760942,-0.1665017754,-0.1915779412,-0.2636027038,-0.3931652009,-0.3731764257,-0.0536193028,-0.0174947772,0.042743165,0.2692256868,-0.0728069097,-0.0048158485,0.3909718096,0.0288445074,0.3221440315,-0.1840127707,0.2099011391,0.2420088202,-0.1991626471,0.3177755177,0.5349607468,-0.2972735763,-0.3691241145,0.1290199459,0.0433830284,0.1929130107,0.1176803708,-0.0890938342,-0.1491372734,-0.0108161494,-0.0472197868,0.0685800016,0.0497183762,0.0365896486,0.1812492013,-0.1543532014,-0.3426198363,0.4528335333,0.1937424392,0.072144568,0.37910074,0.3884422779,-0.1684701294,0.4540009499,-0.0736417621,0.8807650805,-0.0256998558,0.1092430353,0.1145276353,-0.1691682488,0.5640375018,-0.3705225587,-0.079240799,-0.3676967621,-0.1644208729,-0.1051587686,-0.2876475453,0.3911509216,0.2009360194,-0.1806271821,0.3654473126,-0.0542882755,0.0644487888,0.0157216229,0.201832056,0.0514678098,-0.4242671728,-0.0983051211,0.1415809095,-0.0429061167,0.557328403,-0.1223058328,0.1680155247,-0.3432843685,-0.1380122751,-0.2611137331,0.1359822154,-0.1427855045,0.4621758461,0.1932129711,-0.2737782598,0.261238575,0.201859355,0.5882260203,0.2044986933,-0.2599535882,0.0772635937,-0.1663910151,0.016582137,0.0612667054,-0.0391545519,0.4917527437,0.0534949489,-0.2163823247,0.0601257347,-0.2685372233,-0.1303708255,0.225776121,-0.2135467827,0.0380448401,-0.1937054843,-0.2546373606,0.1674422473,0.0846446902,0.0727841407,0.0500445999,0.1630938053,0.0611617044,0.3679528534,-0.1991782486,-0.1089945883,-0.2001005709,0.399002105,-0.1630835086,-0.0530780777,0.5511952639,0.0485677756,-0.2752061188,-0.0091852667,-0.0508705489,0.0294106677,-0.4710206091,-0.2380772382,-0.188812837,-0.37086308,0.0413851365,-0.2812014222,0.3264870942,-0.1140056476,-0.1113222837,-0.3906497359,-0.2503590882,0.2954976261,-0.1786894947,0.1940517277,-0.0487760939,-0.0523174107,0.1656801552,-0.1656511575,-0.218109414,-0.0774855688,-0.2516181767,-0.0239506997,0.2636090815,-0.0844351649,0.0436901748,-0.3919887245,0.0042982562,-0.1659097224,-0.3986682296,-0.0611190088,-0.1735190302,0.1971190125,0.121475175,-0.1493666172,0.2229337841,-0.1989724785,0.0006004007,-0.2565852404,-0.0021219829,0.1197713241,-0.1184921563,0.1176442876,0.1455771327,-0.1439989805,-0.3486425281,0.2148783952,0.1121492684,0.1845344752,0.0702422261,0.2187034637,0.0399331115,-0.043969743,-0.5120844841,0.0981448963,-0.2063385546,-0.0965810344,0.3869278431,-0.3597470522,0.2260470837,0.2410660684,0.211378485,0.3813197017,-0.1031669974,0.1548477262,0.4258761704,0.0828904659,-0.2561053932,-0.218585819,0.1594657302,0.0705732405,0.0581146255,0.3408959806,0.1401667595,-0.1198265478,0.7606736422,0.231188193,0.6941436529,-0.110095419,0.2643289864,0.3837802112,-0.0834088773,0.1918819398,0.2336905897,0.0849544555,0.2238956541,0.3608672023,-0.2763069272,0.1486975402,-0.0203119647,0.1499193758,0.0451034121,-0.5368813276,0.0119824875,-0.2342727184,0.2228981256,-0.0501374714,-0.0394915901,0.0400026515,-0.3863478899,0.0247562435,-0.0632812083,0.2031027079,-0.0334807821,0.1203935146,0.4617051184,0.0134358322,-0.0666046068,-0.1801098883,0.0200307816,-0.4102992415,0.1239869669,0.0687172785,-0.1802632958,-0.2426968217,0.2413455844,0.1321543455,-0.0655998737,-0.2045089006,0.1677087992,0.1210537702,-0.1471865326,-0.1756991446,0.4764433503,0.4472573996,0.3650378883,0.034752164,0.0656483248,0.0429321043,-0.0583051033,0.1567665786,0.0065881098,-0.3767542839,0.2104239911,0.224456802,0.0903666988,-0.0939449146,-0.3274742365,-0.1402044594,-0.0610648133,-0.2975260615,0.3151406348,0.0074527417,-0.1436066777,-0.0054718386,-0.0202253312,-0.4903227389,0.0424956828,0.4755943716,0.023722874,-0.028277427,-0.0399075672,-0.0054830881,0.0109305438,0.3377835155,0.5781381726,0.2742245495,-0.4089942575,0.0974651575,-0.8250724077,-0.0210974906,-0.2424007952,-0.130093351,0.1927946955,0.0657977611,0.0001380579,0.0886073038,0.2619172335,0.1530327499,-0.0672080591,-0.071751982,-0.329986304,-0.2705994844,0.0108857732,-0.0221074335,-0.0670914501,-0.3362775147,0.0149737187,0.1458157301,-0.0369930938,-0.1294059306,-0.0858344063,0.2077594101,-0.157101512,0.4208461046,-0.1390828639,0.3130939305,0.1830291897,-0.1593535393,-0.2420644909,-0.2784920037,-0.0913389027,-0.0075334432,0.035672009,0.5254508257,-0.420216769,-0.3511502147,-0.3888190985,0.0676918328,0.2071896642,-0.0218446087,-0.1191696525,0.3861375153,-0.2201074213,0.3155391514,0.2222313136,0.0527676046,-0.0292808451,0.3702807128,-0.2489839643,-0.3444296122,0.4452055991,-0.1271151304,-0.1187377721,-0.1291197389,0.264762789,-0.0421509743,-0.2734029889,-0.4115774333,0.0109540345,0.147251457,-0.0865049958,-0.3207423389,0.0393424705,0.1452866495,0.0905788988,-0.0591901429,0.4109852612,-0.1266404837,-0.0084436266,0.1631889641,0.1122632176]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/864","title":"Unable to download cnn_dailymail dataset","comments":"Same error here!\r\n","body":"### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.","comment_length":3,"text":"Unable to download cnn_dailymail dataset\n### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.\nSame error here!\r\n","embeddings":[-0.3053483963,0.2343734205,-0.0792272016,0.1706803292,0.3605651855,0.1738087684,0.5607900023,0.2436558902,-0.1719244719,0.0751596391,-0.1179388687,0.0767754763,-0.2897081375,-0.037752647,-0.0130565865,-0.029159829,-0.2127254903,0.1067626104,-0.1973434091,0.0291639101,-0.1067140326,0.0961273983,-0.152200684,-0.0697495267,-0.2929644287,-0.1320359707,0.0363866985,0.1792989373,-0.3712784052,-0.3521047533,0.3593557477,-0.0825211555,0.1873489618,0.4354524016,-0.0001112958,0.2995791733,0.3638949096,-0.0500410311,-0.4619030058,-0.4736372828,-0.179699868,-0.1351321191,0.0417744182,-0.2966795862,-0.0017572836,0.0122687863,-0.0355194137,0.1481749117,0.3508039415,0.6557962298,0.2708093822,0.4034875631,0.3544582725,-0.1669905931,-0.2214082927,0.1594429463,-0.0742193013,0.5186958313,-0.4013257325,0.1032769531,0.2056567669,0.1207840592,-0.1278402507,0.1728732437,0.0507546216,0.0189495403,0.1369116306,-0.4564376771,0.1248682067,0.2598636746,0.4185577333,-0.2014641464,-0.2431451827,0.1196377352,0.1434075534,-0.2662436664,0.1472920924,0.3098779321,-0.1643135697,0.0582009703,-0.2776976824,0.0111331921,-0.2915741205,0.4140602648,-0.2303673029,0.1816995293,0.0403025933,0.074070029,0.2228647918,-0.1520415843,0.4015417099,-0.2636485696,0.2471809238,0.0726078078,-0.0523370765,-0.1367853284,0.0545635037,-0.2483980805,0.0137203373,0.013428431,-0.1899162829,-0.2480668873,0.0034492572,0.0725217238,-0.0031762752,0.0848950818,-0.2378142029,0.6527371407,0.1373988241,0.4710148573,-0.1147683114,0.0423018746,-0.231807664,-0.1314209402,0.1342493296,0.1773545742,0.5618001223,-0.1913481504,-0.4141942859,0.2526679039,-0.5342707038,0.0010653652,0.1967067122,0.3225836754,-0.1696945727,0.0354947671,0.0004754911,0.1084765121,0.0281692632,0.0804795772,-0.269304961,0.2859299481,-0.1196614057,-0.0974230468,0.3719027638,-0.3562030494,0.1796881706,-0.022808535,0.1048488617,-0.1313501447,0.1455949098,-0.2708338499,-0.1464384943,0.4110406041,-0.1056434289,0.1318189353,0.0854265168,0.0459011495,-0.1711771935,0.1168934256,-0.0236167368,-0.39915061,0.0757410899,0.2404704243,-0.0782731622,-0.3346488178,0.1967336833,-0.2113056481,-0.0441895574,0.009028675,-0.1316437125,-0.274735719,-0.0488870777,-0.1960804462,0.3570636213,0.6024298072,-0.4387532175,0.0300037079,-0.1770686358,-0.221620962,0.1865835041,0.2859837115,-0.3314290345,0.286572963,-0.0801039413,-0.2004273683,0.4411747754,-0.4087324739,-0.5403160453,0.2019555569,-0.0457448848,0.0747404993,-0.0577663146,0.0530315936,0.0841376334,0.0774072558,0.0953814834,0.4915386736,-0.1939899623,-0.0277162641,-0.1510462016,-0.329621315,0.1893856525,0.3296568394,0.4307672381,0.1475800574,-0.0082081985,0.1020928398,0.1281474233,0.0709251389,-0.0517404899,0.1613204777,0.3219523132,-0.0353380404,-0.1970408112,-0.0764020458,-0.3915072083,0.1054196283,0.1194396839,0.1985470504,-0.0941805393,-0.0577527583,-0.4808587134,-0.1143843681,-0.2268015891,0.0408287831,0.2129354179,0.2092737854,0.1277357638,0.0437873676,-0.2601549923,0.3888755739,-0.0502965711,0.206739068,-0.1509432048,0.4017488956,-0.2926222682,-0.0586007535,0.0658875853,0.0477816202,0.2527953982,-0.1254135668,-0.1062384844,0.1132924706,0.0106233535,-0.126910314,-0.0642558783,-0.0739669502,0.0886368826,-0.1603420526,0.0321264453,0.3180967569,0.0794667676,0.0362760201,-0.0129773337,0.2884828448,0.1633599401,-0.042649705,0.1151296869,0.1497045159,0.2333556563,-0.2054045051,0.0016976002,-0.0741698295,0.25011307,-0.0294035058,0.3837204874,-0.1167592481,-0.3227631152,-0.2946901619,0.2018496543,-0.223611176,-0.0194849409,0.0601015314,-0.1977682114,0.104586333,-0.2133276165,0.3237728477,0.4441053867,0.0796773285,0.0288690459,0.1061182842,-0.2178812474,-0.1060986742,0.1595469862,0.2282201052,0.1528194398,0.1916666478,-0.0428625122,0.2319728434,-0.2976652682,-0.4345216155,0.2509975433,0.4884727299,-0.3115290105,-0.029125981,-0.3028720617,-0.1064913571,-0.1622152328,-0.4730811119,-0.0811016932,-0.2360571027,-0.1359013468,0.2614663541,0.0638850853,0.0723204911,-0.3064315617,-0.1044544727,0.1325635612,-0.322935015,-0.0856684521,0.0551765487,-0.0396370254,0.1476238221,0.3995283544,-0.0969782919,0.3422371447,-0.0154319182,-0.1305357367,-0.2630587816,-0.1927981675,-0.0004327538,0.0221684016,0.2420882732,0.1672898382,0.430962652,0.0226662755,-0.193605423,0.1404919475,0.0584454164,-0.026869243,0.2278819531,-0.2574721575,-0.1057041809,-0.010171094,-0.4860118926,-0.5390897393,-0.326920718,-0.1599415839,-0.0482898466,0.19598943,0.423708111,0.116052188,0.1385355592,-0.1566384435,0.0500074252,-0.1222570613,-0.1248526126,0.4896658361,-0.3741426468,-0.3817529976,0.1049950793,-0.1745428294,0.2857010067,-0.089311935,-0.4778813422,-0.2060608715,-0.1840455681,-0.0779386759,-0.1294920743,-0.2238213867,0.1238588989,-0.0531830192,-0.0448054187,0.121457316,-0.1170100793,-0.0215034578,0.1581927538,0.0776544288,-0.1886182427,0.5733031034,0.0269378722,0.6518713236,0.2029567659,-0.2608668506,0.3176005781,-0.0788142756,0.2220961154,-0.0406563729,-0.4731614292,-0.0460523441,0.2134987712,0.034807004,-0.059652403,-0.2256636322,-0.3095944822,-0.2722097635,-0.1301361918,-0.3081510365,-0.2545178831,0.0712788478,0.1031405553,0.2053652704,0.0995039642,0.191415906,0.0591029339,-0.1681022048,-0.0150375487,0.431399405,0.0771581084,0.0633863285,0.1295619309,-0.0592789985,-0.330547601,0.0920389891,-0.0928845704,0.3071429133,-0.0619968139,0.1866856068,-0.0089119561,-0.0765375122,0.2313290685,0.0687429532,0.1700016409,0.2777774334,-0.1218906119,-0.3610595167,-0.2361999005,-0.3534475267,0.0337870456,0.2101772279,0.183933273,-0.3156815171,-0.0661932454,0.615386188,0.1403870285,-0.0556864068,-0.3950105906,-0.2080629021,-0.3233417273,-0.491627574,0.0996384248,-0.1292090118,0.2446268946,0.0665892959,-0.1874731183,-0.0348469503,0.0217926241,0.18876414,0.0004602246,0.1229709759,0.0819305703,0.393582195,0.1386470348,-0.0997420549,0.4361158907,0.4650555253,0.0479854606,-0.41444695,-0.088219732,0.1417153776,-0.1368325055,0.1666573733,-0.2181945294,-0.2737380266,0.2523176968,0.2460319698,0.1622442454,0.1378744692,0.4610220492,-0.2430559099,-0.5729042292,-0.3615541756,0.189625755,-0.179726705,0.1413583755,0.1366215795,-0.2497364134,0.0172254704,0.0104196006,-0.0525608733,0.7789852023,0.1548537612,0.0441102199,0.4680342674,-0.0037566889,0.3014906645,-0.0559014827,0.0838298053,-0.3072827458,-0.2321264148,-0.1228302494,-0.2282546759,0.1536598504,0.2682596445,-0.2145921886,0.3666175008,-0.0568442829,0.1451283544,0.0025314549,0.3421612084,-0.0584325343,0.0655459985,-0.0626575202,0.2040625513,-0.1444291919,0.409722656,-0.1356842667,-0.149004966,-0.0768997669,-0.2596896887,-0.2632807493,0.1581942588,-0.3894182444,0.3009178042,0.0940063596,-0.0541631989,-0.0775868669,-0.0720009059,0.0379504673,0.0387830436,-0.1752779186,0.0844960958,-0.1288580447,-0.2365467399,-0.0595169775,-0.1094142944,0.4144064486,-0.2647612989,-0.2428983152,0.0964698642,0.0492814519,-0.2107218653,0.1066624746,0.0112686176,0.2041609436,-0.3407523632,-0.2518103719,-0.0327805951,-0.1775391549,-0.12553069,0.186204195,0.0360444263,-0.0364982747,0.2830505073,-0.0620818809,-0.1947600096,-0.1506979316,0.351208955,-0.1067584306,-0.1233872399,0.4347908795,0.072054632,-0.3199335039,-0.2920496762,0.0793827996,-0.0465265624,-0.4147839248,0.0896514952,0.0486419089,0.3403031528,0.0655375123,0.1039589494,0.1910346895,0.044651486,-0.0263227336,-0.5710043907,-0.2213813961,-0.0354618803,0.035441488,0.2289005071,-0.094972007,0.3262064159,0.2034397423,-0.0655390695,-0.3433675766,0.1364937276,-0.095421575,-0.0452936105,0.1359447986,-0.1490375251,0.4430599213,0.1009041294,0.1978885531,0.0278638974,-0.1176836565,-0.2749330401,-0.1282685101,0.1190886796,0.167985335,-0.1261828542,0.0420026705,-0.1996611506,-0.3425234854,0.1667084098,0.0233251974,0.1777889729,-0.0187482443,0.0477432758,0.1527056545,0.057696756,-0.1524341851,0.0197509918,-0.0933409557,0.3424445093,0.2201880962,0.2526732683,0.1092485264,-0.028304074,-0.4353369474,-0.0889778137,0.0215206128,-0.1089897975,0.3970861137,-0.2267069072,-0.0844079256,0.1168391258,0.2218981236,0.3427518606,-0.2445510477,0.0473774672,0.3164501786,0.2502194941,-0.2942929864,-0.074464418,0.1661240906,0.1107305214,0.1122423261,0.0778133348,-0.2546724975,0.0717090666,0.0237519108,0.1373420209,0.3921376169,-0.0400031582,0.1250549704,0.3877003193,-0.1304763705,-0.0067147673,0.1958841681,-0.0121206986,0.070871979,0.4315577447,-0.2908119261,-0.0119646117,-0.172923103,-0.0734783411,0.2268629521,-0.2598978579,-0.0039406689,0.2269203067,-0.118860811,0.0853064358,-0.0197697077,0.1014457345,-0.0209625904,0.1261386722,-0.4866469204,0.3916065395,-0.0998357832,0.3023129702,-0.2309093624,-0.132981807,-0.2292218655,0.1452821046,0.1043042019,-0.1691759378,0.2273527682,0.3998037577,-0.2776856422,-0.1550411731,0.0752838925,0.3108185828,-0.0813329294,-0.2761153579,0.1822858006,0.3667353988,-0.0344139636,-0.0020702961,0.4475148022,0.3937963545,0.2824236155,0.0474153645,0.2371765971,0.1083737686,-0.1588353515,0.0991288722,0.1113982573,0.2268317044,0.347774893,0.3200223744,0.2253325135,-0.2515419424,0.0480179787,-0.0931240693,-0.0826394707,-0.2318333983,0.5614549518,0.0496760122,0.0397591777,-0.2407839596,-0.0536701046,-0.4856576324,-0.0348074809,0.3326095641,0.1133724004,0.0982405171,-0.1042847708,0.0643889904,-0.0975297689,0.3120363653,0.3321331441,0.2571236491,-0.6201792955,0.004288333,-0.6339927316,0.1701634973,-0.1985664666,0.0743850023,-0.0166364368,0.0412023179,-0.2311744988,0.3115764856,0.2193800658,0.0395457298,0.0116452165,0.2519503236,-0.1994519979,-0.0294276495,0.1661611497,-0.1432815492,0.0830325782,-0.4240460098,0.0994042978,-0.4774453342,0.017368475,-0.0443782322,-0.3066776097,-0.0353295766,0.2529799342,0.4154708385,0.0541495681,0.7115378976,-0.0803386495,-0.2179797292,-0.2073084563,-0.0156711414,-0.2390040755,0.1551925689,0.1466782242,0.3737408221,-0.0241130292,-0.1898607612,-0.417525202,0.3508703113,-0.107716091,-0.1765215695,-0.2167667896,-0.0179202445,-0.1924640536,0.2424509078,0.0509917289,0.0850015953,-0.0730701983,0.1792785823,-0.3576475084,-0.4977722764,0.6089873314,-0.2351760268,-0.2108120024,-0.2193452865,0.1239553168,-0.117237106,0.1133626476,-0.497730881,0.3211047649,0.3238481283,-0.0129207922,-0.1095672026,0.2009022683,0.2233956754,0.2048543543,0.0640877187,0.3913771808,0.1503724754,-0.3911642432,-0.0575917885,-0.2036331147]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/864","title":"Unable to download cnn_dailymail dataset","comments":"Same here! My kaggle notebook stopped working like yesterday. It's strange because I have fixed version of datasets==1.1.2","body":"### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.","comment_length":18,"text":"Unable to download cnn_dailymail dataset\n### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.\nSame here! My kaggle notebook stopped working like yesterday. It's strange because I have fixed version of datasets==1.1.2","embeddings":[-0.3053483963,0.2343734205,-0.0792272016,0.1706803292,0.3605651855,0.1738087684,0.5607900023,0.2436558902,-0.1719244719,0.0751596391,-0.1179388687,0.0767754763,-0.2897081375,-0.037752647,-0.0130565865,-0.029159829,-0.2127254903,0.1067626104,-0.1973434091,0.0291639101,-0.1067140326,0.0961273983,-0.152200684,-0.0697495267,-0.2929644287,-0.1320359707,0.0363866985,0.1792989373,-0.3712784052,-0.3521047533,0.3593557477,-0.0825211555,0.1873489618,0.4354524016,-0.0001112958,0.2995791733,0.3638949096,-0.0500410311,-0.4619030058,-0.4736372828,-0.179699868,-0.1351321191,0.0417744182,-0.2966795862,-0.0017572836,0.0122687863,-0.0355194137,0.1481749117,0.3508039415,0.6557962298,0.2708093822,0.4034875631,0.3544582725,-0.1669905931,-0.2214082927,0.1594429463,-0.0742193013,0.5186958313,-0.4013257325,0.1032769531,0.2056567669,0.1207840592,-0.1278402507,0.1728732437,0.0507546216,0.0189495403,0.1369116306,-0.4564376771,0.1248682067,0.2598636746,0.4185577333,-0.2014641464,-0.2431451827,0.1196377352,0.1434075534,-0.2662436664,0.1472920924,0.3098779321,-0.1643135697,0.0582009703,-0.2776976824,0.0111331921,-0.2915741205,0.4140602648,-0.2303673029,0.1816995293,0.0403025933,0.074070029,0.2228647918,-0.1520415843,0.4015417099,-0.2636485696,0.2471809238,0.0726078078,-0.0523370765,-0.1367853284,0.0545635037,-0.2483980805,0.0137203373,0.013428431,-0.1899162829,-0.2480668873,0.0034492572,0.0725217238,-0.0031762752,0.0848950818,-0.2378142029,0.6527371407,0.1373988241,0.4710148573,-0.1147683114,0.0423018746,-0.231807664,-0.1314209402,0.1342493296,0.1773545742,0.5618001223,-0.1913481504,-0.4141942859,0.2526679039,-0.5342707038,0.0010653652,0.1967067122,0.3225836754,-0.1696945727,0.0354947671,0.0004754911,0.1084765121,0.0281692632,0.0804795772,-0.269304961,0.2859299481,-0.1196614057,-0.0974230468,0.3719027638,-0.3562030494,0.1796881706,-0.022808535,0.1048488617,-0.1313501447,0.1455949098,-0.2708338499,-0.1464384943,0.4110406041,-0.1056434289,0.1318189353,0.0854265168,0.0459011495,-0.1711771935,0.1168934256,-0.0236167368,-0.39915061,0.0757410899,0.2404704243,-0.0782731622,-0.3346488178,0.1967336833,-0.2113056481,-0.0441895574,0.009028675,-0.1316437125,-0.274735719,-0.0488870777,-0.1960804462,0.3570636213,0.6024298072,-0.4387532175,0.0300037079,-0.1770686358,-0.221620962,0.1865835041,0.2859837115,-0.3314290345,0.286572963,-0.0801039413,-0.2004273683,0.4411747754,-0.4087324739,-0.5403160453,0.2019555569,-0.0457448848,0.0747404993,-0.0577663146,0.0530315936,0.0841376334,0.0774072558,0.0953814834,0.4915386736,-0.1939899623,-0.0277162641,-0.1510462016,-0.329621315,0.1893856525,0.3296568394,0.4307672381,0.1475800574,-0.0082081985,0.1020928398,0.1281474233,0.0709251389,-0.0517404899,0.1613204777,0.3219523132,-0.0353380404,-0.1970408112,-0.0764020458,-0.3915072083,0.1054196283,0.1194396839,0.1985470504,-0.0941805393,-0.0577527583,-0.4808587134,-0.1143843681,-0.2268015891,0.0408287831,0.2129354179,0.2092737854,0.1277357638,0.0437873676,-0.2601549923,0.3888755739,-0.0502965711,0.206739068,-0.1509432048,0.4017488956,-0.2926222682,-0.0586007535,0.0658875853,0.0477816202,0.2527953982,-0.1254135668,-0.1062384844,0.1132924706,0.0106233535,-0.126910314,-0.0642558783,-0.0739669502,0.0886368826,-0.1603420526,0.0321264453,0.3180967569,0.0794667676,0.0362760201,-0.0129773337,0.2884828448,0.1633599401,-0.042649705,0.1151296869,0.1497045159,0.2333556563,-0.2054045051,0.0016976002,-0.0741698295,0.25011307,-0.0294035058,0.3837204874,-0.1167592481,-0.3227631152,-0.2946901619,0.2018496543,-0.223611176,-0.0194849409,0.0601015314,-0.1977682114,0.104586333,-0.2133276165,0.3237728477,0.4441053867,0.0796773285,0.0288690459,0.1061182842,-0.2178812474,-0.1060986742,0.1595469862,0.2282201052,0.1528194398,0.1916666478,-0.0428625122,0.2319728434,-0.2976652682,-0.4345216155,0.2509975433,0.4884727299,-0.3115290105,-0.029125981,-0.3028720617,-0.1064913571,-0.1622152328,-0.4730811119,-0.0811016932,-0.2360571027,-0.1359013468,0.2614663541,0.0638850853,0.0723204911,-0.3064315617,-0.1044544727,0.1325635612,-0.322935015,-0.0856684521,0.0551765487,-0.0396370254,0.1476238221,0.3995283544,-0.0969782919,0.3422371447,-0.0154319182,-0.1305357367,-0.2630587816,-0.1927981675,-0.0004327538,0.0221684016,0.2420882732,0.1672898382,0.430962652,0.0226662755,-0.193605423,0.1404919475,0.0584454164,-0.026869243,0.2278819531,-0.2574721575,-0.1057041809,-0.010171094,-0.4860118926,-0.5390897393,-0.326920718,-0.1599415839,-0.0482898466,0.19598943,0.423708111,0.116052188,0.1385355592,-0.1566384435,0.0500074252,-0.1222570613,-0.1248526126,0.4896658361,-0.3741426468,-0.3817529976,0.1049950793,-0.1745428294,0.2857010067,-0.089311935,-0.4778813422,-0.2060608715,-0.1840455681,-0.0779386759,-0.1294920743,-0.2238213867,0.1238588989,-0.0531830192,-0.0448054187,0.121457316,-0.1170100793,-0.0215034578,0.1581927538,0.0776544288,-0.1886182427,0.5733031034,0.0269378722,0.6518713236,0.2029567659,-0.2608668506,0.3176005781,-0.0788142756,0.2220961154,-0.0406563729,-0.4731614292,-0.0460523441,0.2134987712,0.034807004,-0.059652403,-0.2256636322,-0.3095944822,-0.2722097635,-0.1301361918,-0.3081510365,-0.2545178831,0.0712788478,0.1031405553,0.2053652704,0.0995039642,0.191415906,0.0591029339,-0.1681022048,-0.0150375487,0.431399405,0.0771581084,0.0633863285,0.1295619309,-0.0592789985,-0.330547601,0.0920389891,-0.0928845704,0.3071429133,-0.0619968139,0.1866856068,-0.0089119561,-0.0765375122,0.2313290685,0.0687429532,0.1700016409,0.2777774334,-0.1218906119,-0.3610595167,-0.2361999005,-0.3534475267,0.0337870456,0.2101772279,0.183933273,-0.3156815171,-0.0661932454,0.615386188,0.1403870285,-0.0556864068,-0.3950105906,-0.2080629021,-0.3233417273,-0.491627574,0.0996384248,-0.1292090118,0.2446268946,0.0665892959,-0.1874731183,-0.0348469503,0.0217926241,0.18876414,0.0004602246,0.1229709759,0.0819305703,0.393582195,0.1386470348,-0.0997420549,0.4361158907,0.4650555253,0.0479854606,-0.41444695,-0.088219732,0.1417153776,-0.1368325055,0.1666573733,-0.2181945294,-0.2737380266,0.2523176968,0.2460319698,0.1622442454,0.1378744692,0.4610220492,-0.2430559099,-0.5729042292,-0.3615541756,0.189625755,-0.179726705,0.1413583755,0.1366215795,-0.2497364134,0.0172254704,0.0104196006,-0.0525608733,0.7789852023,0.1548537612,0.0441102199,0.4680342674,-0.0037566889,0.3014906645,-0.0559014827,0.0838298053,-0.3072827458,-0.2321264148,-0.1228302494,-0.2282546759,0.1536598504,0.2682596445,-0.2145921886,0.3666175008,-0.0568442829,0.1451283544,0.0025314549,0.3421612084,-0.0584325343,0.0655459985,-0.0626575202,0.2040625513,-0.1444291919,0.409722656,-0.1356842667,-0.149004966,-0.0768997669,-0.2596896887,-0.2632807493,0.1581942588,-0.3894182444,0.3009178042,0.0940063596,-0.0541631989,-0.0775868669,-0.0720009059,0.0379504673,0.0387830436,-0.1752779186,0.0844960958,-0.1288580447,-0.2365467399,-0.0595169775,-0.1094142944,0.4144064486,-0.2647612989,-0.2428983152,0.0964698642,0.0492814519,-0.2107218653,0.1066624746,0.0112686176,0.2041609436,-0.3407523632,-0.2518103719,-0.0327805951,-0.1775391549,-0.12553069,0.186204195,0.0360444263,-0.0364982747,0.2830505073,-0.0620818809,-0.1947600096,-0.1506979316,0.351208955,-0.1067584306,-0.1233872399,0.4347908795,0.072054632,-0.3199335039,-0.2920496762,0.0793827996,-0.0465265624,-0.4147839248,0.0896514952,0.0486419089,0.3403031528,0.0655375123,0.1039589494,0.1910346895,0.044651486,-0.0263227336,-0.5710043907,-0.2213813961,-0.0354618803,0.035441488,0.2289005071,-0.094972007,0.3262064159,0.2034397423,-0.0655390695,-0.3433675766,0.1364937276,-0.095421575,-0.0452936105,0.1359447986,-0.1490375251,0.4430599213,0.1009041294,0.1978885531,0.0278638974,-0.1176836565,-0.2749330401,-0.1282685101,0.1190886796,0.167985335,-0.1261828542,0.0420026705,-0.1996611506,-0.3425234854,0.1667084098,0.0233251974,0.1777889729,-0.0187482443,0.0477432758,0.1527056545,0.057696756,-0.1524341851,0.0197509918,-0.0933409557,0.3424445093,0.2201880962,0.2526732683,0.1092485264,-0.028304074,-0.4353369474,-0.0889778137,0.0215206128,-0.1089897975,0.3970861137,-0.2267069072,-0.0844079256,0.1168391258,0.2218981236,0.3427518606,-0.2445510477,0.0473774672,0.3164501786,0.2502194941,-0.2942929864,-0.074464418,0.1661240906,0.1107305214,0.1122423261,0.0778133348,-0.2546724975,0.0717090666,0.0237519108,0.1373420209,0.3921376169,-0.0400031582,0.1250549704,0.3877003193,-0.1304763705,-0.0067147673,0.1958841681,-0.0121206986,0.070871979,0.4315577447,-0.2908119261,-0.0119646117,-0.172923103,-0.0734783411,0.2268629521,-0.2598978579,-0.0039406689,0.2269203067,-0.118860811,0.0853064358,-0.0197697077,0.1014457345,-0.0209625904,0.1261386722,-0.4866469204,0.3916065395,-0.0998357832,0.3023129702,-0.2309093624,-0.132981807,-0.2292218655,0.1452821046,0.1043042019,-0.1691759378,0.2273527682,0.3998037577,-0.2776856422,-0.1550411731,0.0752838925,0.3108185828,-0.0813329294,-0.2761153579,0.1822858006,0.3667353988,-0.0344139636,-0.0020702961,0.4475148022,0.3937963545,0.2824236155,0.0474153645,0.2371765971,0.1083737686,-0.1588353515,0.0991288722,0.1113982573,0.2268317044,0.347774893,0.3200223744,0.2253325135,-0.2515419424,0.0480179787,-0.0931240693,-0.0826394707,-0.2318333983,0.5614549518,0.0496760122,0.0397591777,-0.2407839596,-0.0536701046,-0.4856576324,-0.0348074809,0.3326095641,0.1133724004,0.0982405171,-0.1042847708,0.0643889904,-0.0975297689,0.3120363653,0.3321331441,0.2571236491,-0.6201792955,0.004288333,-0.6339927316,0.1701634973,-0.1985664666,0.0743850023,-0.0166364368,0.0412023179,-0.2311744988,0.3115764856,0.2193800658,0.0395457298,0.0116452165,0.2519503236,-0.1994519979,-0.0294276495,0.1661611497,-0.1432815492,0.0830325782,-0.4240460098,0.0994042978,-0.4774453342,0.017368475,-0.0443782322,-0.3066776097,-0.0353295766,0.2529799342,0.4154708385,0.0541495681,0.7115378976,-0.0803386495,-0.2179797292,-0.2073084563,-0.0156711414,-0.2390040755,0.1551925689,0.1466782242,0.3737408221,-0.0241130292,-0.1898607612,-0.417525202,0.3508703113,-0.107716091,-0.1765215695,-0.2167667896,-0.0179202445,-0.1924640536,0.2424509078,0.0509917289,0.0850015953,-0.0730701983,0.1792785823,-0.3576475084,-0.4977722764,0.6089873314,-0.2351760268,-0.2108120024,-0.2193452865,0.1239553168,-0.117237106,0.1133626476,-0.497730881,0.3211047649,0.3238481283,-0.0129207922,-0.1095672026,0.2009022683,0.2233956754,0.2048543543,0.0640877187,0.3913771808,0.1503724754,-0.3911642432,-0.0575917885,-0.2036331147]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/864","title":"Unable to download cnn_dailymail dataset","comments":"I'm looking at it right now","body":"### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.","comment_length":6,"text":"Unable to download cnn_dailymail dataset\n### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.\nI'm looking at it right now","embeddings":[-0.3053483963,0.2343734205,-0.0792272016,0.1706803292,0.3605651855,0.1738087684,0.5607900023,0.2436558902,-0.1719244719,0.0751596391,-0.1179388687,0.0767754763,-0.2897081375,-0.037752647,-0.0130565865,-0.029159829,-0.2127254903,0.1067626104,-0.1973434091,0.0291639101,-0.1067140326,0.0961273983,-0.152200684,-0.0697495267,-0.2929644287,-0.1320359707,0.0363866985,0.1792989373,-0.3712784052,-0.3521047533,0.3593557477,-0.0825211555,0.1873489618,0.4354524016,-0.0001112958,0.2995791733,0.3638949096,-0.0500410311,-0.4619030058,-0.4736372828,-0.179699868,-0.1351321191,0.0417744182,-0.2966795862,-0.0017572836,0.0122687863,-0.0355194137,0.1481749117,0.3508039415,0.6557962298,0.2708093822,0.4034875631,0.3544582725,-0.1669905931,-0.2214082927,0.1594429463,-0.0742193013,0.5186958313,-0.4013257325,0.1032769531,0.2056567669,0.1207840592,-0.1278402507,0.1728732437,0.0507546216,0.0189495403,0.1369116306,-0.4564376771,0.1248682067,0.2598636746,0.4185577333,-0.2014641464,-0.2431451827,0.1196377352,0.1434075534,-0.2662436664,0.1472920924,0.3098779321,-0.1643135697,0.0582009703,-0.2776976824,0.0111331921,-0.2915741205,0.4140602648,-0.2303673029,0.1816995293,0.0403025933,0.074070029,0.2228647918,-0.1520415843,0.4015417099,-0.2636485696,0.2471809238,0.0726078078,-0.0523370765,-0.1367853284,0.0545635037,-0.2483980805,0.0137203373,0.013428431,-0.1899162829,-0.2480668873,0.0034492572,0.0725217238,-0.0031762752,0.0848950818,-0.2378142029,0.6527371407,0.1373988241,0.4710148573,-0.1147683114,0.0423018746,-0.231807664,-0.1314209402,0.1342493296,0.1773545742,0.5618001223,-0.1913481504,-0.4141942859,0.2526679039,-0.5342707038,0.0010653652,0.1967067122,0.3225836754,-0.1696945727,0.0354947671,0.0004754911,0.1084765121,0.0281692632,0.0804795772,-0.269304961,0.2859299481,-0.1196614057,-0.0974230468,0.3719027638,-0.3562030494,0.1796881706,-0.022808535,0.1048488617,-0.1313501447,0.1455949098,-0.2708338499,-0.1464384943,0.4110406041,-0.1056434289,0.1318189353,0.0854265168,0.0459011495,-0.1711771935,0.1168934256,-0.0236167368,-0.39915061,0.0757410899,0.2404704243,-0.0782731622,-0.3346488178,0.1967336833,-0.2113056481,-0.0441895574,0.009028675,-0.1316437125,-0.274735719,-0.0488870777,-0.1960804462,0.3570636213,0.6024298072,-0.4387532175,0.0300037079,-0.1770686358,-0.221620962,0.1865835041,0.2859837115,-0.3314290345,0.286572963,-0.0801039413,-0.2004273683,0.4411747754,-0.4087324739,-0.5403160453,0.2019555569,-0.0457448848,0.0747404993,-0.0577663146,0.0530315936,0.0841376334,0.0774072558,0.0953814834,0.4915386736,-0.1939899623,-0.0277162641,-0.1510462016,-0.329621315,0.1893856525,0.3296568394,0.4307672381,0.1475800574,-0.0082081985,0.1020928398,0.1281474233,0.0709251389,-0.0517404899,0.1613204777,0.3219523132,-0.0353380404,-0.1970408112,-0.0764020458,-0.3915072083,0.1054196283,0.1194396839,0.1985470504,-0.0941805393,-0.0577527583,-0.4808587134,-0.1143843681,-0.2268015891,0.0408287831,0.2129354179,0.2092737854,0.1277357638,0.0437873676,-0.2601549923,0.3888755739,-0.0502965711,0.206739068,-0.1509432048,0.4017488956,-0.2926222682,-0.0586007535,0.0658875853,0.0477816202,0.2527953982,-0.1254135668,-0.1062384844,0.1132924706,0.0106233535,-0.126910314,-0.0642558783,-0.0739669502,0.0886368826,-0.1603420526,0.0321264453,0.3180967569,0.0794667676,0.0362760201,-0.0129773337,0.2884828448,0.1633599401,-0.042649705,0.1151296869,0.1497045159,0.2333556563,-0.2054045051,0.0016976002,-0.0741698295,0.25011307,-0.0294035058,0.3837204874,-0.1167592481,-0.3227631152,-0.2946901619,0.2018496543,-0.223611176,-0.0194849409,0.0601015314,-0.1977682114,0.104586333,-0.2133276165,0.3237728477,0.4441053867,0.0796773285,0.0288690459,0.1061182842,-0.2178812474,-0.1060986742,0.1595469862,0.2282201052,0.1528194398,0.1916666478,-0.0428625122,0.2319728434,-0.2976652682,-0.4345216155,0.2509975433,0.4884727299,-0.3115290105,-0.029125981,-0.3028720617,-0.1064913571,-0.1622152328,-0.4730811119,-0.0811016932,-0.2360571027,-0.1359013468,0.2614663541,0.0638850853,0.0723204911,-0.3064315617,-0.1044544727,0.1325635612,-0.322935015,-0.0856684521,0.0551765487,-0.0396370254,0.1476238221,0.3995283544,-0.0969782919,0.3422371447,-0.0154319182,-0.1305357367,-0.2630587816,-0.1927981675,-0.0004327538,0.0221684016,0.2420882732,0.1672898382,0.430962652,0.0226662755,-0.193605423,0.1404919475,0.0584454164,-0.026869243,0.2278819531,-0.2574721575,-0.1057041809,-0.010171094,-0.4860118926,-0.5390897393,-0.326920718,-0.1599415839,-0.0482898466,0.19598943,0.423708111,0.116052188,0.1385355592,-0.1566384435,0.0500074252,-0.1222570613,-0.1248526126,0.4896658361,-0.3741426468,-0.3817529976,0.1049950793,-0.1745428294,0.2857010067,-0.089311935,-0.4778813422,-0.2060608715,-0.1840455681,-0.0779386759,-0.1294920743,-0.2238213867,0.1238588989,-0.0531830192,-0.0448054187,0.121457316,-0.1170100793,-0.0215034578,0.1581927538,0.0776544288,-0.1886182427,0.5733031034,0.0269378722,0.6518713236,0.2029567659,-0.2608668506,0.3176005781,-0.0788142756,0.2220961154,-0.0406563729,-0.4731614292,-0.0460523441,0.2134987712,0.034807004,-0.059652403,-0.2256636322,-0.3095944822,-0.2722097635,-0.1301361918,-0.3081510365,-0.2545178831,0.0712788478,0.1031405553,0.2053652704,0.0995039642,0.191415906,0.0591029339,-0.1681022048,-0.0150375487,0.431399405,0.0771581084,0.0633863285,0.1295619309,-0.0592789985,-0.330547601,0.0920389891,-0.0928845704,0.3071429133,-0.0619968139,0.1866856068,-0.0089119561,-0.0765375122,0.2313290685,0.0687429532,0.1700016409,0.2777774334,-0.1218906119,-0.3610595167,-0.2361999005,-0.3534475267,0.0337870456,0.2101772279,0.183933273,-0.3156815171,-0.0661932454,0.615386188,0.1403870285,-0.0556864068,-0.3950105906,-0.2080629021,-0.3233417273,-0.491627574,0.0996384248,-0.1292090118,0.2446268946,0.0665892959,-0.1874731183,-0.0348469503,0.0217926241,0.18876414,0.0004602246,0.1229709759,0.0819305703,0.393582195,0.1386470348,-0.0997420549,0.4361158907,0.4650555253,0.0479854606,-0.41444695,-0.088219732,0.1417153776,-0.1368325055,0.1666573733,-0.2181945294,-0.2737380266,0.2523176968,0.2460319698,0.1622442454,0.1378744692,0.4610220492,-0.2430559099,-0.5729042292,-0.3615541756,0.189625755,-0.179726705,0.1413583755,0.1366215795,-0.2497364134,0.0172254704,0.0104196006,-0.0525608733,0.7789852023,0.1548537612,0.0441102199,0.4680342674,-0.0037566889,0.3014906645,-0.0559014827,0.0838298053,-0.3072827458,-0.2321264148,-0.1228302494,-0.2282546759,0.1536598504,0.2682596445,-0.2145921886,0.3666175008,-0.0568442829,0.1451283544,0.0025314549,0.3421612084,-0.0584325343,0.0655459985,-0.0626575202,0.2040625513,-0.1444291919,0.409722656,-0.1356842667,-0.149004966,-0.0768997669,-0.2596896887,-0.2632807493,0.1581942588,-0.3894182444,0.3009178042,0.0940063596,-0.0541631989,-0.0775868669,-0.0720009059,0.0379504673,0.0387830436,-0.1752779186,0.0844960958,-0.1288580447,-0.2365467399,-0.0595169775,-0.1094142944,0.4144064486,-0.2647612989,-0.2428983152,0.0964698642,0.0492814519,-0.2107218653,0.1066624746,0.0112686176,0.2041609436,-0.3407523632,-0.2518103719,-0.0327805951,-0.1775391549,-0.12553069,0.186204195,0.0360444263,-0.0364982747,0.2830505073,-0.0620818809,-0.1947600096,-0.1506979316,0.351208955,-0.1067584306,-0.1233872399,0.4347908795,0.072054632,-0.3199335039,-0.2920496762,0.0793827996,-0.0465265624,-0.4147839248,0.0896514952,0.0486419089,0.3403031528,0.0655375123,0.1039589494,0.1910346895,0.044651486,-0.0263227336,-0.5710043907,-0.2213813961,-0.0354618803,0.035441488,0.2289005071,-0.094972007,0.3262064159,0.2034397423,-0.0655390695,-0.3433675766,0.1364937276,-0.095421575,-0.0452936105,0.1359447986,-0.1490375251,0.4430599213,0.1009041294,0.1978885531,0.0278638974,-0.1176836565,-0.2749330401,-0.1282685101,0.1190886796,0.167985335,-0.1261828542,0.0420026705,-0.1996611506,-0.3425234854,0.1667084098,0.0233251974,0.1777889729,-0.0187482443,0.0477432758,0.1527056545,0.057696756,-0.1524341851,0.0197509918,-0.0933409557,0.3424445093,0.2201880962,0.2526732683,0.1092485264,-0.028304074,-0.4353369474,-0.0889778137,0.0215206128,-0.1089897975,0.3970861137,-0.2267069072,-0.0844079256,0.1168391258,0.2218981236,0.3427518606,-0.2445510477,0.0473774672,0.3164501786,0.2502194941,-0.2942929864,-0.074464418,0.1661240906,0.1107305214,0.1122423261,0.0778133348,-0.2546724975,0.0717090666,0.0237519108,0.1373420209,0.3921376169,-0.0400031582,0.1250549704,0.3877003193,-0.1304763705,-0.0067147673,0.1958841681,-0.0121206986,0.070871979,0.4315577447,-0.2908119261,-0.0119646117,-0.172923103,-0.0734783411,0.2268629521,-0.2598978579,-0.0039406689,0.2269203067,-0.118860811,0.0853064358,-0.0197697077,0.1014457345,-0.0209625904,0.1261386722,-0.4866469204,0.3916065395,-0.0998357832,0.3023129702,-0.2309093624,-0.132981807,-0.2292218655,0.1452821046,0.1043042019,-0.1691759378,0.2273527682,0.3998037577,-0.2776856422,-0.1550411731,0.0752838925,0.3108185828,-0.0813329294,-0.2761153579,0.1822858006,0.3667353988,-0.0344139636,-0.0020702961,0.4475148022,0.3937963545,0.2824236155,0.0474153645,0.2371765971,0.1083737686,-0.1588353515,0.0991288722,0.1113982573,0.2268317044,0.347774893,0.3200223744,0.2253325135,-0.2515419424,0.0480179787,-0.0931240693,-0.0826394707,-0.2318333983,0.5614549518,0.0496760122,0.0397591777,-0.2407839596,-0.0536701046,-0.4856576324,-0.0348074809,0.3326095641,0.1133724004,0.0982405171,-0.1042847708,0.0643889904,-0.0975297689,0.3120363653,0.3321331441,0.2571236491,-0.6201792955,0.004288333,-0.6339927316,0.1701634973,-0.1985664666,0.0743850023,-0.0166364368,0.0412023179,-0.2311744988,0.3115764856,0.2193800658,0.0395457298,0.0116452165,0.2519503236,-0.1994519979,-0.0294276495,0.1661611497,-0.1432815492,0.0830325782,-0.4240460098,0.0994042978,-0.4774453342,0.017368475,-0.0443782322,-0.3066776097,-0.0353295766,0.2529799342,0.4154708385,0.0541495681,0.7115378976,-0.0803386495,-0.2179797292,-0.2073084563,-0.0156711414,-0.2390040755,0.1551925689,0.1466782242,0.3737408221,-0.0241130292,-0.1898607612,-0.417525202,0.3508703113,-0.107716091,-0.1765215695,-0.2167667896,-0.0179202445,-0.1924640536,0.2424509078,0.0509917289,0.0850015953,-0.0730701983,0.1792785823,-0.3576475084,-0.4977722764,0.6089873314,-0.2351760268,-0.2108120024,-0.2193452865,0.1239553168,-0.117237106,0.1133626476,-0.497730881,0.3211047649,0.3238481283,-0.0129207922,-0.1095672026,0.2009022683,0.2233956754,0.2048543543,0.0640877187,0.3913771808,0.1503724754,-0.3911642432,-0.0575917885,-0.2036331147]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/864","title":"Unable to download cnn_dailymail dataset","comments":"I couldn't reproduce unfortunately. I tried\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nload_dataset(\"cnn_dailymail\", \"3.0.0\", download_mode=\"force_redownload\")\r\n```\r\nand it worked fine on both my env (python 3.7.2) and colab (python 3.6.9)\r\n\r\nMaybe there was an issue with the google drive download link of the dataset ?\r\nAre you still having the issue ? If so could your give me more info about your python and requests version ?","body":"### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.","comment_length":66,"text":"Unable to download cnn_dailymail dataset\n### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.\nI couldn't reproduce unfortunately. I tried\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nload_dataset(\"cnn_dailymail\", \"3.0.0\", download_mode=\"force_redownload\")\r\n```\r\nand it worked fine on both my env (python 3.7.2) and colab (python 3.6.9)\r\n\r\nMaybe there was an issue with the google drive download link of the dataset ?\r\nAre you still having the issue ? If so could your give me more info about your python and requests version ?","embeddings":[-0.3053483963,0.2343734205,-0.0792272016,0.1706803292,0.3605651855,0.1738087684,0.5607900023,0.2436558902,-0.1719244719,0.0751596391,-0.1179388687,0.0767754763,-0.2897081375,-0.037752647,-0.0130565865,-0.029159829,-0.2127254903,0.1067626104,-0.1973434091,0.0291639101,-0.1067140326,0.0961273983,-0.152200684,-0.0697495267,-0.2929644287,-0.1320359707,0.0363866985,0.1792989373,-0.3712784052,-0.3521047533,0.3593557477,-0.0825211555,0.1873489618,0.4354524016,-0.0001112958,0.2995791733,0.3638949096,-0.0500410311,-0.4619030058,-0.4736372828,-0.179699868,-0.1351321191,0.0417744182,-0.2966795862,-0.0017572836,0.0122687863,-0.0355194137,0.1481749117,0.3508039415,0.6557962298,0.2708093822,0.4034875631,0.3544582725,-0.1669905931,-0.2214082927,0.1594429463,-0.0742193013,0.5186958313,-0.4013257325,0.1032769531,0.2056567669,0.1207840592,-0.1278402507,0.1728732437,0.0507546216,0.0189495403,0.1369116306,-0.4564376771,0.1248682067,0.2598636746,0.4185577333,-0.2014641464,-0.2431451827,0.1196377352,0.1434075534,-0.2662436664,0.1472920924,0.3098779321,-0.1643135697,0.0582009703,-0.2776976824,0.0111331921,-0.2915741205,0.4140602648,-0.2303673029,0.1816995293,0.0403025933,0.074070029,0.2228647918,-0.1520415843,0.4015417099,-0.2636485696,0.2471809238,0.0726078078,-0.0523370765,-0.1367853284,0.0545635037,-0.2483980805,0.0137203373,0.013428431,-0.1899162829,-0.2480668873,0.0034492572,0.0725217238,-0.0031762752,0.0848950818,-0.2378142029,0.6527371407,0.1373988241,0.4710148573,-0.1147683114,0.0423018746,-0.231807664,-0.1314209402,0.1342493296,0.1773545742,0.5618001223,-0.1913481504,-0.4141942859,0.2526679039,-0.5342707038,0.0010653652,0.1967067122,0.3225836754,-0.1696945727,0.0354947671,0.0004754911,0.1084765121,0.0281692632,0.0804795772,-0.269304961,0.2859299481,-0.1196614057,-0.0974230468,0.3719027638,-0.3562030494,0.1796881706,-0.022808535,0.1048488617,-0.1313501447,0.1455949098,-0.2708338499,-0.1464384943,0.4110406041,-0.1056434289,0.1318189353,0.0854265168,0.0459011495,-0.1711771935,0.1168934256,-0.0236167368,-0.39915061,0.0757410899,0.2404704243,-0.0782731622,-0.3346488178,0.1967336833,-0.2113056481,-0.0441895574,0.009028675,-0.1316437125,-0.274735719,-0.0488870777,-0.1960804462,0.3570636213,0.6024298072,-0.4387532175,0.0300037079,-0.1770686358,-0.221620962,0.1865835041,0.2859837115,-0.3314290345,0.286572963,-0.0801039413,-0.2004273683,0.4411747754,-0.4087324739,-0.5403160453,0.2019555569,-0.0457448848,0.0747404993,-0.0577663146,0.0530315936,0.0841376334,0.0774072558,0.0953814834,0.4915386736,-0.1939899623,-0.0277162641,-0.1510462016,-0.329621315,0.1893856525,0.3296568394,0.4307672381,0.1475800574,-0.0082081985,0.1020928398,0.1281474233,0.0709251389,-0.0517404899,0.1613204777,0.3219523132,-0.0353380404,-0.1970408112,-0.0764020458,-0.3915072083,0.1054196283,0.1194396839,0.1985470504,-0.0941805393,-0.0577527583,-0.4808587134,-0.1143843681,-0.2268015891,0.0408287831,0.2129354179,0.2092737854,0.1277357638,0.0437873676,-0.2601549923,0.3888755739,-0.0502965711,0.206739068,-0.1509432048,0.4017488956,-0.2926222682,-0.0586007535,0.0658875853,0.0477816202,0.2527953982,-0.1254135668,-0.1062384844,0.1132924706,0.0106233535,-0.126910314,-0.0642558783,-0.0739669502,0.0886368826,-0.1603420526,0.0321264453,0.3180967569,0.0794667676,0.0362760201,-0.0129773337,0.2884828448,0.1633599401,-0.042649705,0.1151296869,0.1497045159,0.2333556563,-0.2054045051,0.0016976002,-0.0741698295,0.25011307,-0.0294035058,0.3837204874,-0.1167592481,-0.3227631152,-0.2946901619,0.2018496543,-0.223611176,-0.0194849409,0.0601015314,-0.1977682114,0.104586333,-0.2133276165,0.3237728477,0.4441053867,0.0796773285,0.0288690459,0.1061182842,-0.2178812474,-0.1060986742,0.1595469862,0.2282201052,0.1528194398,0.1916666478,-0.0428625122,0.2319728434,-0.2976652682,-0.4345216155,0.2509975433,0.4884727299,-0.3115290105,-0.029125981,-0.3028720617,-0.1064913571,-0.1622152328,-0.4730811119,-0.0811016932,-0.2360571027,-0.1359013468,0.2614663541,0.0638850853,0.0723204911,-0.3064315617,-0.1044544727,0.1325635612,-0.322935015,-0.0856684521,0.0551765487,-0.0396370254,0.1476238221,0.3995283544,-0.0969782919,0.3422371447,-0.0154319182,-0.1305357367,-0.2630587816,-0.1927981675,-0.0004327538,0.0221684016,0.2420882732,0.1672898382,0.430962652,0.0226662755,-0.193605423,0.1404919475,0.0584454164,-0.026869243,0.2278819531,-0.2574721575,-0.1057041809,-0.010171094,-0.4860118926,-0.5390897393,-0.326920718,-0.1599415839,-0.0482898466,0.19598943,0.423708111,0.116052188,0.1385355592,-0.1566384435,0.0500074252,-0.1222570613,-0.1248526126,0.4896658361,-0.3741426468,-0.3817529976,0.1049950793,-0.1745428294,0.2857010067,-0.089311935,-0.4778813422,-0.2060608715,-0.1840455681,-0.0779386759,-0.1294920743,-0.2238213867,0.1238588989,-0.0531830192,-0.0448054187,0.121457316,-0.1170100793,-0.0215034578,0.1581927538,0.0776544288,-0.1886182427,0.5733031034,0.0269378722,0.6518713236,0.2029567659,-0.2608668506,0.3176005781,-0.0788142756,0.2220961154,-0.0406563729,-0.4731614292,-0.0460523441,0.2134987712,0.034807004,-0.059652403,-0.2256636322,-0.3095944822,-0.2722097635,-0.1301361918,-0.3081510365,-0.2545178831,0.0712788478,0.1031405553,0.2053652704,0.0995039642,0.191415906,0.0591029339,-0.1681022048,-0.0150375487,0.431399405,0.0771581084,0.0633863285,0.1295619309,-0.0592789985,-0.330547601,0.0920389891,-0.0928845704,0.3071429133,-0.0619968139,0.1866856068,-0.0089119561,-0.0765375122,0.2313290685,0.0687429532,0.1700016409,0.2777774334,-0.1218906119,-0.3610595167,-0.2361999005,-0.3534475267,0.0337870456,0.2101772279,0.183933273,-0.3156815171,-0.0661932454,0.615386188,0.1403870285,-0.0556864068,-0.3950105906,-0.2080629021,-0.3233417273,-0.491627574,0.0996384248,-0.1292090118,0.2446268946,0.0665892959,-0.1874731183,-0.0348469503,0.0217926241,0.18876414,0.0004602246,0.1229709759,0.0819305703,0.393582195,0.1386470348,-0.0997420549,0.4361158907,0.4650555253,0.0479854606,-0.41444695,-0.088219732,0.1417153776,-0.1368325055,0.1666573733,-0.2181945294,-0.2737380266,0.2523176968,0.2460319698,0.1622442454,0.1378744692,0.4610220492,-0.2430559099,-0.5729042292,-0.3615541756,0.189625755,-0.179726705,0.1413583755,0.1366215795,-0.2497364134,0.0172254704,0.0104196006,-0.0525608733,0.7789852023,0.1548537612,0.0441102199,0.4680342674,-0.0037566889,0.3014906645,-0.0559014827,0.0838298053,-0.3072827458,-0.2321264148,-0.1228302494,-0.2282546759,0.1536598504,0.2682596445,-0.2145921886,0.3666175008,-0.0568442829,0.1451283544,0.0025314549,0.3421612084,-0.0584325343,0.0655459985,-0.0626575202,0.2040625513,-0.1444291919,0.409722656,-0.1356842667,-0.149004966,-0.0768997669,-0.2596896887,-0.2632807493,0.1581942588,-0.3894182444,0.3009178042,0.0940063596,-0.0541631989,-0.0775868669,-0.0720009059,0.0379504673,0.0387830436,-0.1752779186,0.0844960958,-0.1288580447,-0.2365467399,-0.0595169775,-0.1094142944,0.4144064486,-0.2647612989,-0.2428983152,0.0964698642,0.0492814519,-0.2107218653,0.1066624746,0.0112686176,0.2041609436,-0.3407523632,-0.2518103719,-0.0327805951,-0.1775391549,-0.12553069,0.186204195,0.0360444263,-0.0364982747,0.2830505073,-0.0620818809,-0.1947600096,-0.1506979316,0.351208955,-0.1067584306,-0.1233872399,0.4347908795,0.072054632,-0.3199335039,-0.2920496762,0.0793827996,-0.0465265624,-0.4147839248,0.0896514952,0.0486419089,0.3403031528,0.0655375123,0.1039589494,0.1910346895,0.044651486,-0.0263227336,-0.5710043907,-0.2213813961,-0.0354618803,0.035441488,0.2289005071,-0.094972007,0.3262064159,0.2034397423,-0.0655390695,-0.3433675766,0.1364937276,-0.095421575,-0.0452936105,0.1359447986,-0.1490375251,0.4430599213,0.1009041294,0.1978885531,0.0278638974,-0.1176836565,-0.2749330401,-0.1282685101,0.1190886796,0.167985335,-0.1261828542,0.0420026705,-0.1996611506,-0.3425234854,0.1667084098,0.0233251974,0.1777889729,-0.0187482443,0.0477432758,0.1527056545,0.057696756,-0.1524341851,0.0197509918,-0.0933409557,0.3424445093,0.2201880962,0.2526732683,0.1092485264,-0.028304074,-0.4353369474,-0.0889778137,0.0215206128,-0.1089897975,0.3970861137,-0.2267069072,-0.0844079256,0.1168391258,0.2218981236,0.3427518606,-0.2445510477,0.0473774672,0.3164501786,0.2502194941,-0.2942929864,-0.074464418,0.1661240906,0.1107305214,0.1122423261,0.0778133348,-0.2546724975,0.0717090666,0.0237519108,0.1373420209,0.3921376169,-0.0400031582,0.1250549704,0.3877003193,-0.1304763705,-0.0067147673,0.1958841681,-0.0121206986,0.070871979,0.4315577447,-0.2908119261,-0.0119646117,-0.172923103,-0.0734783411,0.2268629521,-0.2598978579,-0.0039406689,0.2269203067,-0.118860811,0.0853064358,-0.0197697077,0.1014457345,-0.0209625904,0.1261386722,-0.4866469204,0.3916065395,-0.0998357832,0.3023129702,-0.2309093624,-0.132981807,-0.2292218655,0.1452821046,0.1043042019,-0.1691759378,0.2273527682,0.3998037577,-0.2776856422,-0.1550411731,0.0752838925,0.3108185828,-0.0813329294,-0.2761153579,0.1822858006,0.3667353988,-0.0344139636,-0.0020702961,0.4475148022,0.3937963545,0.2824236155,0.0474153645,0.2371765971,0.1083737686,-0.1588353515,0.0991288722,0.1113982573,0.2268317044,0.347774893,0.3200223744,0.2253325135,-0.2515419424,0.0480179787,-0.0931240693,-0.0826394707,-0.2318333983,0.5614549518,0.0496760122,0.0397591777,-0.2407839596,-0.0536701046,-0.4856576324,-0.0348074809,0.3326095641,0.1133724004,0.0982405171,-0.1042847708,0.0643889904,-0.0975297689,0.3120363653,0.3321331441,0.2571236491,-0.6201792955,0.004288333,-0.6339927316,0.1701634973,-0.1985664666,0.0743850023,-0.0166364368,0.0412023179,-0.2311744988,0.3115764856,0.2193800658,0.0395457298,0.0116452165,0.2519503236,-0.1994519979,-0.0294276495,0.1661611497,-0.1432815492,0.0830325782,-0.4240460098,0.0994042978,-0.4774453342,0.017368475,-0.0443782322,-0.3066776097,-0.0353295766,0.2529799342,0.4154708385,0.0541495681,0.7115378976,-0.0803386495,-0.2179797292,-0.2073084563,-0.0156711414,-0.2390040755,0.1551925689,0.1466782242,0.3737408221,-0.0241130292,-0.1898607612,-0.417525202,0.3508703113,-0.107716091,-0.1765215695,-0.2167667896,-0.0179202445,-0.1924640536,0.2424509078,0.0509917289,0.0850015953,-0.0730701983,0.1792785823,-0.3576475084,-0.4977722764,0.6089873314,-0.2351760268,-0.2108120024,-0.2193452865,0.1239553168,-0.117237106,0.1133626476,-0.497730881,0.3211047649,0.3238481283,-0.0129207922,-0.1095672026,0.2009022683,0.2233956754,0.2048543543,0.0640877187,0.3913771808,0.1503724754,-0.3911642432,-0.0575917885,-0.2036331147]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/864","title":"Unable to download cnn_dailymail dataset","comments":"No, It's working fine now. Very strange. Here are my python and request versions\r\n\r\nrequests 2.24.0\r\nPython 3.8.2","body":"### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.","comment_length":18,"text":"Unable to download cnn_dailymail dataset\n### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.\nNo, It's working fine now. Very strange. Here are my python and request versions\r\n\r\nrequests 2.24.0\r\nPython 3.8.2","embeddings":[-0.3053483963,0.2343734205,-0.0792272016,0.1706803292,0.3605651855,0.1738087684,0.5607900023,0.2436558902,-0.1719244719,0.0751596391,-0.1179388687,0.0767754763,-0.2897081375,-0.037752647,-0.0130565865,-0.029159829,-0.2127254903,0.1067626104,-0.1973434091,0.0291639101,-0.1067140326,0.0961273983,-0.152200684,-0.0697495267,-0.2929644287,-0.1320359707,0.0363866985,0.1792989373,-0.3712784052,-0.3521047533,0.3593557477,-0.0825211555,0.1873489618,0.4354524016,-0.0001112958,0.2995791733,0.3638949096,-0.0500410311,-0.4619030058,-0.4736372828,-0.179699868,-0.1351321191,0.0417744182,-0.2966795862,-0.0017572836,0.0122687863,-0.0355194137,0.1481749117,0.3508039415,0.6557962298,0.2708093822,0.4034875631,0.3544582725,-0.1669905931,-0.2214082927,0.1594429463,-0.0742193013,0.5186958313,-0.4013257325,0.1032769531,0.2056567669,0.1207840592,-0.1278402507,0.1728732437,0.0507546216,0.0189495403,0.1369116306,-0.4564376771,0.1248682067,0.2598636746,0.4185577333,-0.2014641464,-0.2431451827,0.1196377352,0.1434075534,-0.2662436664,0.1472920924,0.3098779321,-0.1643135697,0.0582009703,-0.2776976824,0.0111331921,-0.2915741205,0.4140602648,-0.2303673029,0.1816995293,0.0403025933,0.074070029,0.2228647918,-0.1520415843,0.4015417099,-0.2636485696,0.2471809238,0.0726078078,-0.0523370765,-0.1367853284,0.0545635037,-0.2483980805,0.0137203373,0.013428431,-0.1899162829,-0.2480668873,0.0034492572,0.0725217238,-0.0031762752,0.0848950818,-0.2378142029,0.6527371407,0.1373988241,0.4710148573,-0.1147683114,0.0423018746,-0.231807664,-0.1314209402,0.1342493296,0.1773545742,0.5618001223,-0.1913481504,-0.4141942859,0.2526679039,-0.5342707038,0.0010653652,0.1967067122,0.3225836754,-0.1696945727,0.0354947671,0.0004754911,0.1084765121,0.0281692632,0.0804795772,-0.269304961,0.2859299481,-0.1196614057,-0.0974230468,0.3719027638,-0.3562030494,0.1796881706,-0.022808535,0.1048488617,-0.1313501447,0.1455949098,-0.2708338499,-0.1464384943,0.4110406041,-0.1056434289,0.1318189353,0.0854265168,0.0459011495,-0.1711771935,0.1168934256,-0.0236167368,-0.39915061,0.0757410899,0.2404704243,-0.0782731622,-0.3346488178,0.1967336833,-0.2113056481,-0.0441895574,0.009028675,-0.1316437125,-0.274735719,-0.0488870777,-0.1960804462,0.3570636213,0.6024298072,-0.4387532175,0.0300037079,-0.1770686358,-0.221620962,0.1865835041,0.2859837115,-0.3314290345,0.286572963,-0.0801039413,-0.2004273683,0.4411747754,-0.4087324739,-0.5403160453,0.2019555569,-0.0457448848,0.0747404993,-0.0577663146,0.0530315936,0.0841376334,0.0774072558,0.0953814834,0.4915386736,-0.1939899623,-0.0277162641,-0.1510462016,-0.329621315,0.1893856525,0.3296568394,0.4307672381,0.1475800574,-0.0082081985,0.1020928398,0.1281474233,0.0709251389,-0.0517404899,0.1613204777,0.3219523132,-0.0353380404,-0.1970408112,-0.0764020458,-0.3915072083,0.1054196283,0.1194396839,0.1985470504,-0.0941805393,-0.0577527583,-0.4808587134,-0.1143843681,-0.2268015891,0.0408287831,0.2129354179,0.2092737854,0.1277357638,0.0437873676,-0.2601549923,0.3888755739,-0.0502965711,0.206739068,-0.1509432048,0.4017488956,-0.2926222682,-0.0586007535,0.0658875853,0.0477816202,0.2527953982,-0.1254135668,-0.1062384844,0.1132924706,0.0106233535,-0.126910314,-0.0642558783,-0.0739669502,0.0886368826,-0.1603420526,0.0321264453,0.3180967569,0.0794667676,0.0362760201,-0.0129773337,0.2884828448,0.1633599401,-0.042649705,0.1151296869,0.1497045159,0.2333556563,-0.2054045051,0.0016976002,-0.0741698295,0.25011307,-0.0294035058,0.3837204874,-0.1167592481,-0.3227631152,-0.2946901619,0.2018496543,-0.223611176,-0.0194849409,0.0601015314,-0.1977682114,0.104586333,-0.2133276165,0.3237728477,0.4441053867,0.0796773285,0.0288690459,0.1061182842,-0.2178812474,-0.1060986742,0.1595469862,0.2282201052,0.1528194398,0.1916666478,-0.0428625122,0.2319728434,-0.2976652682,-0.4345216155,0.2509975433,0.4884727299,-0.3115290105,-0.029125981,-0.3028720617,-0.1064913571,-0.1622152328,-0.4730811119,-0.0811016932,-0.2360571027,-0.1359013468,0.2614663541,0.0638850853,0.0723204911,-0.3064315617,-0.1044544727,0.1325635612,-0.322935015,-0.0856684521,0.0551765487,-0.0396370254,0.1476238221,0.3995283544,-0.0969782919,0.3422371447,-0.0154319182,-0.1305357367,-0.2630587816,-0.1927981675,-0.0004327538,0.0221684016,0.2420882732,0.1672898382,0.430962652,0.0226662755,-0.193605423,0.1404919475,0.0584454164,-0.026869243,0.2278819531,-0.2574721575,-0.1057041809,-0.010171094,-0.4860118926,-0.5390897393,-0.326920718,-0.1599415839,-0.0482898466,0.19598943,0.423708111,0.116052188,0.1385355592,-0.1566384435,0.0500074252,-0.1222570613,-0.1248526126,0.4896658361,-0.3741426468,-0.3817529976,0.1049950793,-0.1745428294,0.2857010067,-0.089311935,-0.4778813422,-0.2060608715,-0.1840455681,-0.0779386759,-0.1294920743,-0.2238213867,0.1238588989,-0.0531830192,-0.0448054187,0.121457316,-0.1170100793,-0.0215034578,0.1581927538,0.0776544288,-0.1886182427,0.5733031034,0.0269378722,0.6518713236,0.2029567659,-0.2608668506,0.3176005781,-0.0788142756,0.2220961154,-0.0406563729,-0.4731614292,-0.0460523441,0.2134987712,0.034807004,-0.059652403,-0.2256636322,-0.3095944822,-0.2722097635,-0.1301361918,-0.3081510365,-0.2545178831,0.0712788478,0.1031405553,0.2053652704,0.0995039642,0.191415906,0.0591029339,-0.1681022048,-0.0150375487,0.431399405,0.0771581084,0.0633863285,0.1295619309,-0.0592789985,-0.330547601,0.0920389891,-0.0928845704,0.3071429133,-0.0619968139,0.1866856068,-0.0089119561,-0.0765375122,0.2313290685,0.0687429532,0.1700016409,0.2777774334,-0.1218906119,-0.3610595167,-0.2361999005,-0.3534475267,0.0337870456,0.2101772279,0.183933273,-0.3156815171,-0.0661932454,0.615386188,0.1403870285,-0.0556864068,-0.3950105906,-0.2080629021,-0.3233417273,-0.491627574,0.0996384248,-0.1292090118,0.2446268946,0.0665892959,-0.1874731183,-0.0348469503,0.0217926241,0.18876414,0.0004602246,0.1229709759,0.0819305703,0.393582195,0.1386470348,-0.0997420549,0.4361158907,0.4650555253,0.0479854606,-0.41444695,-0.088219732,0.1417153776,-0.1368325055,0.1666573733,-0.2181945294,-0.2737380266,0.2523176968,0.2460319698,0.1622442454,0.1378744692,0.4610220492,-0.2430559099,-0.5729042292,-0.3615541756,0.189625755,-0.179726705,0.1413583755,0.1366215795,-0.2497364134,0.0172254704,0.0104196006,-0.0525608733,0.7789852023,0.1548537612,0.0441102199,0.4680342674,-0.0037566889,0.3014906645,-0.0559014827,0.0838298053,-0.3072827458,-0.2321264148,-0.1228302494,-0.2282546759,0.1536598504,0.2682596445,-0.2145921886,0.3666175008,-0.0568442829,0.1451283544,0.0025314549,0.3421612084,-0.0584325343,0.0655459985,-0.0626575202,0.2040625513,-0.1444291919,0.409722656,-0.1356842667,-0.149004966,-0.0768997669,-0.2596896887,-0.2632807493,0.1581942588,-0.3894182444,0.3009178042,0.0940063596,-0.0541631989,-0.0775868669,-0.0720009059,0.0379504673,0.0387830436,-0.1752779186,0.0844960958,-0.1288580447,-0.2365467399,-0.0595169775,-0.1094142944,0.4144064486,-0.2647612989,-0.2428983152,0.0964698642,0.0492814519,-0.2107218653,0.1066624746,0.0112686176,0.2041609436,-0.3407523632,-0.2518103719,-0.0327805951,-0.1775391549,-0.12553069,0.186204195,0.0360444263,-0.0364982747,0.2830505073,-0.0620818809,-0.1947600096,-0.1506979316,0.351208955,-0.1067584306,-0.1233872399,0.4347908795,0.072054632,-0.3199335039,-0.2920496762,0.0793827996,-0.0465265624,-0.4147839248,0.0896514952,0.0486419089,0.3403031528,0.0655375123,0.1039589494,0.1910346895,0.044651486,-0.0263227336,-0.5710043907,-0.2213813961,-0.0354618803,0.035441488,0.2289005071,-0.094972007,0.3262064159,0.2034397423,-0.0655390695,-0.3433675766,0.1364937276,-0.095421575,-0.0452936105,0.1359447986,-0.1490375251,0.4430599213,0.1009041294,0.1978885531,0.0278638974,-0.1176836565,-0.2749330401,-0.1282685101,0.1190886796,0.167985335,-0.1261828542,0.0420026705,-0.1996611506,-0.3425234854,0.1667084098,0.0233251974,0.1777889729,-0.0187482443,0.0477432758,0.1527056545,0.057696756,-0.1524341851,0.0197509918,-0.0933409557,0.3424445093,0.2201880962,0.2526732683,0.1092485264,-0.028304074,-0.4353369474,-0.0889778137,0.0215206128,-0.1089897975,0.3970861137,-0.2267069072,-0.0844079256,0.1168391258,0.2218981236,0.3427518606,-0.2445510477,0.0473774672,0.3164501786,0.2502194941,-0.2942929864,-0.074464418,0.1661240906,0.1107305214,0.1122423261,0.0778133348,-0.2546724975,0.0717090666,0.0237519108,0.1373420209,0.3921376169,-0.0400031582,0.1250549704,0.3877003193,-0.1304763705,-0.0067147673,0.1958841681,-0.0121206986,0.070871979,0.4315577447,-0.2908119261,-0.0119646117,-0.172923103,-0.0734783411,0.2268629521,-0.2598978579,-0.0039406689,0.2269203067,-0.118860811,0.0853064358,-0.0197697077,0.1014457345,-0.0209625904,0.1261386722,-0.4866469204,0.3916065395,-0.0998357832,0.3023129702,-0.2309093624,-0.132981807,-0.2292218655,0.1452821046,0.1043042019,-0.1691759378,0.2273527682,0.3998037577,-0.2776856422,-0.1550411731,0.0752838925,0.3108185828,-0.0813329294,-0.2761153579,0.1822858006,0.3667353988,-0.0344139636,-0.0020702961,0.4475148022,0.3937963545,0.2824236155,0.0474153645,0.2371765971,0.1083737686,-0.1588353515,0.0991288722,0.1113982573,0.2268317044,0.347774893,0.3200223744,0.2253325135,-0.2515419424,0.0480179787,-0.0931240693,-0.0826394707,-0.2318333983,0.5614549518,0.0496760122,0.0397591777,-0.2407839596,-0.0536701046,-0.4856576324,-0.0348074809,0.3326095641,0.1133724004,0.0982405171,-0.1042847708,0.0643889904,-0.0975297689,0.3120363653,0.3321331441,0.2571236491,-0.6201792955,0.004288333,-0.6339927316,0.1701634973,-0.1985664666,0.0743850023,-0.0166364368,0.0412023179,-0.2311744988,0.3115764856,0.2193800658,0.0395457298,0.0116452165,0.2519503236,-0.1994519979,-0.0294276495,0.1661611497,-0.1432815492,0.0830325782,-0.4240460098,0.0994042978,-0.4774453342,0.017368475,-0.0443782322,-0.3066776097,-0.0353295766,0.2529799342,0.4154708385,0.0541495681,0.7115378976,-0.0803386495,-0.2179797292,-0.2073084563,-0.0156711414,-0.2390040755,0.1551925689,0.1466782242,0.3737408221,-0.0241130292,-0.1898607612,-0.417525202,0.3508703113,-0.107716091,-0.1765215695,-0.2167667896,-0.0179202445,-0.1924640536,0.2424509078,0.0509917289,0.0850015953,-0.0730701983,0.1792785823,-0.3576475084,-0.4977722764,0.6089873314,-0.2351760268,-0.2108120024,-0.2193452865,0.1239553168,-0.117237106,0.1133626476,-0.497730881,0.3211047649,0.3238481283,-0.0129207922,-0.1095672026,0.2009022683,0.2233956754,0.2048543543,0.0640877187,0.3913771808,0.1503724754,-0.3911642432,-0.0575917885,-0.2036331147]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/864","title":"Unable to download cnn_dailymail dataset","comments":"It's working as expected. Closing the issue \r\n\r\nThanks everybody.","body":"### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.","comment_length":9,"text":"Unable to download cnn_dailymail dataset\n### Script to reproduce the error\r\n```\r\nfrom datasets import load_dataset\r\n\r\ntrain_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\nvalid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n```\r\n\r\n\r\n### Error\r\n```\r\n---------------------------------------------------------------------------\r\nNotADirectoryError Traceback (most recent call last)\r\n<ipython-input-8-47c39c228935> in <module>()\r\n 1 from datasets import load_dataset\r\n 2 \r\n----> 3 train_dataset = load_dataset(\"cnn_dailymail\", \"3.0.0\", split= 'train[:10%')\r\n 4 valid_dataset = load_dataset(\"cnn_dailymail\",\"3.0.0\", split=\"validation[:5%]\")\r\n\r\n5 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 469 if not downloaded_from_gcs:\r\n 470 self._download_and_prepare(\r\n--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 472 )\r\n 473 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 524 split_dict = SplitDict(dataset_name=self.name)\r\n 525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 527 \r\n 528 # Checksums verification\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _split_generators(self, dl_manager)\r\n 252 def _split_generators(self, dl_manager):\r\n 253 dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)\r\n 255 # Generate shared vocabulary\r\n 256 \r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _subset_filenames(dl_paths, split)\r\n 153 else:\r\n 154 logging.fatal(\"Unsupported split: %s\", split)\r\n--> 155 cnn = _find_files(dl_paths, \"cnn\", urls)\r\n 156 dm = _find_files(dl_paths, \"dm\", urls)\r\n 157 return cnn + dm\r\n\r\n\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/cnn_dailymail\/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)\r\n 132 else:\r\n 133 logging.fatal(\"Unsupported publisher: %s\", publisher)\r\n--> 134 files = sorted(os.listdir(top_dir))\r\n 135 \r\n 136 ret_files = []\r\n\r\nNotADirectoryError: [Errno 20] Not a directory: '\/root\/.cache\/huggingface\/datasets\/downloads\/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b\/cnn\/stories'\r\n\r\n```\r\n\r\nThanks for any suggestions.\nIt's working as expected. Closing the issue \r\n\r\nThanks everybody.","embeddings":[-0.3053483963,0.2343734205,-0.0792272016,0.1706803292,0.3605651855,0.1738087684,0.5607900023,0.2436558902,-0.1719244719,0.0751596391,-0.1179388687,0.0767754763,-0.2897081375,-0.037752647,-0.0130565865,-0.029159829,-0.2127254903,0.1067626104,-0.1973434091,0.0291639101,-0.1067140326,0.0961273983,-0.152200684,-0.0697495267,-0.2929644287,-0.1320359707,0.0363866985,0.1792989373,-0.3712784052,-0.3521047533,0.3593557477,-0.0825211555,0.1873489618,0.4354524016,-0.0001112958,0.2995791733,0.3638949096,-0.0500410311,-0.4619030058,-0.4736372828,-0.179699868,-0.1351321191,0.0417744182,-0.2966795862,-0.0017572836,0.0122687863,-0.0355194137,0.1481749117,0.3508039415,0.6557962298,0.2708093822,0.4034875631,0.3544582725,-0.1669905931,-0.2214082927,0.1594429463,-0.0742193013,0.5186958313,-0.4013257325,0.1032769531,0.2056567669,0.1207840592,-0.1278402507,0.1728732437,0.0507546216,0.0189495403,0.1369116306,-0.4564376771,0.1248682067,0.2598636746,0.4185577333,-0.2014641464,-0.2431451827,0.1196377352,0.1434075534,-0.2662436664,0.1472920924,0.3098779321,-0.1643135697,0.0582009703,-0.2776976824,0.0111331921,-0.2915741205,0.4140602648,-0.2303673029,0.1816995293,0.0403025933,0.074070029,0.2228647918,-0.1520415843,0.4015417099,-0.2636485696,0.2471809238,0.0726078078,-0.0523370765,-0.1367853284,0.0545635037,-0.2483980805,0.0137203373,0.013428431,-0.1899162829,-0.2480668873,0.0034492572,0.0725217238,-0.0031762752,0.0848950818,-0.2378142029,0.6527371407,0.1373988241,0.4710148573,-0.1147683114,0.0423018746,-0.231807664,-0.1314209402,0.1342493296,0.1773545742,0.5618001223,-0.1913481504,-0.4141942859,0.2526679039,-0.5342707038,0.0010653652,0.1967067122,0.3225836754,-0.1696945727,0.0354947671,0.0004754911,0.1084765121,0.0281692632,0.0804795772,-0.269304961,0.2859299481,-0.1196614057,-0.0974230468,0.3719027638,-0.3562030494,0.1796881706,-0.022808535,0.1048488617,-0.1313501447,0.1455949098,-0.2708338499,-0.1464384943,0.4110406041,-0.1056434289,0.1318189353,0.0854265168,0.0459011495,-0.1711771935,0.1168934256,-0.0236167368,-0.39915061,0.0757410899,0.2404704243,-0.0782731622,-0.3346488178,0.1967336833,-0.2113056481,-0.0441895574,0.009028675,-0.1316437125,-0.274735719,-0.0488870777,-0.1960804462,0.3570636213,0.6024298072,-0.4387532175,0.0300037079,-0.1770686358,-0.221620962,0.1865835041,0.2859837115,-0.3314290345,0.286572963,-0.0801039413,-0.2004273683,0.4411747754,-0.4087324739,-0.5403160453,0.2019555569,-0.0457448848,0.0747404993,-0.0577663146,0.0530315936,0.0841376334,0.0774072558,0.0953814834,0.4915386736,-0.1939899623,-0.0277162641,-0.1510462016,-0.329621315,0.1893856525,0.3296568394,0.4307672381,0.1475800574,-0.0082081985,0.1020928398,0.1281474233,0.0709251389,-0.0517404899,0.1613204777,0.3219523132,-0.0353380404,-0.1970408112,-0.0764020458,-0.3915072083,0.1054196283,0.1194396839,0.1985470504,-0.0941805393,-0.0577527583,-0.4808587134,-0.1143843681,-0.2268015891,0.0408287831,0.2129354179,0.2092737854,0.1277357638,0.0437873676,-0.2601549923,0.3888755739,-0.0502965711,0.206739068,-0.1509432048,0.4017488956,-0.2926222682,-0.0586007535,0.0658875853,0.0477816202,0.2527953982,-0.1254135668,-0.1062384844,0.1132924706,0.0106233535,-0.126910314,-0.0642558783,-0.0739669502,0.0886368826,-0.1603420526,0.0321264453,0.3180967569,0.0794667676,0.0362760201,-0.0129773337,0.2884828448,0.1633599401,-0.042649705,0.1151296869,0.1497045159,0.2333556563,-0.2054045051,0.0016976002,-0.0741698295,0.25011307,-0.0294035058,0.3837204874,-0.1167592481,-0.3227631152,-0.2946901619,0.2018496543,-0.223611176,-0.0194849409,0.0601015314,-0.1977682114,0.104586333,-0.2133276165,0.3237728477,0.4441053867,0.0796773285,0.0288690459,0.1061182842,-0.2178812474,-0.1060986742,0.1595469862,0.2282201052,0.1528194398,0.1916666478,-0.0428625122,0.2319728434,-0.2976652682,-0.4345216155,0.2509975433,0.4884727299,-0.3115290105,-0.029125981,-0.3028720617,-0.1064913571,-0.1622152328,-0.4730811119,-0.0811016932,-0.2360571027,-0.1359013468,0.2614663541,0.0638850853,0.0723204911,-0.3064315617,-0.1044544727,0.1325635612,-0.322935015,-0.0856684521,0.0551765487,-0.0396370254,0.1476238221,0.3995283544,-0.0969782919,0.3422371447,-0.0154319182,-0.1305357367,-0.2630587816,-0.1927981675,-0.0004327538,0.0221684016,0.2420882732,0.1672898382,0.430962652,0.0226662755,-0.193605423,0.1404919475,0.0584454164,-0.026869243,0.2278819531,-0.2574721575,-0.1057041809,-0.010171094,-0.4860118926,-0.5390897393,-0.326920718,-0.1599415839,-0.0482898466,0.19598943,0.423708111,0.116052188,0.1385355592,-0.1566384435,0.0500074252,-0.1222570613,-0.1248526126,0.4896658361,-0.3741426468,-0.3817529976,0.1049950793,-0.1745428294,0.2857010067,-0.089311935,-0.4778813422,-0.2060608715,-0.1840455681,-0.0779386759,-0.1294920743,-0.2238213867,0.1238588989,-0.0531830192,-0.0448054187,0.121457316,-0.1170100793,-0.0215034578,0.1581927538,0.0776544288,-0.1886182427,0.5733031034,0.0269378722,0.6518713236,0.2029567659,-0.2608668506,0.3176005781,-0.0788142756,0.2220961154,-0.0406563729,-0.4731614292,-0.0460523441,0.2134987712,0.034807004,-0.059652403,-0.2256636322,-0.3095944822,-0.2722097635,-0.1301361918,-0.3081510365,-0.2545178831,0.0712788478,0.1031405553,0.2053652704,0.0995039642,0.191415906,0.0591029339,-0.1681022048,-0.0150375487,0.431399405,0.0771581084,0.0633863285,0.1295619309,-0.0592789985,-0.330547601,0.0920389891,-0.0928845704,0.3071429133,-0.0619968139,0.1866856068,-0.0089119561,-0.0765375122,0.2313290685,0.0687429532,0.1700016409,0.2777774334,-0.1218906119,-0.3610595167,-0.2361999005,-0.3534475267,0.0337870456,0.2101772279,0.183933273,-0.3156815171,-0.0661932454,0.615386188,0.1403870285,-0.0556864068,-0.3950105906,-0.2080629021,-0.3233417273,-0.491627574,0.0996384248,-0.1292090118,0.2446268946,0.0665892959,-0.1874731183,-0.0348469503,0.0217926241,0.18876414,0.0004602246,0.1229709759,0.0819305703,0.393582195,0.1386470348,-0.0997420549,0.4361158907,0.4650555253,0.0479854606,-0.41444695,-0.088219732,0.1417153776,-0.1368325055,0.1666573733,-0.2181945294,-0.2737380266,0.2523176968,0.2460319698,0.1622442454,0.1378744692,0.4610220492,-0.2430559099,-0.5729042292,-0.3615541756,0.189625755,-0.179726705,0.1413583755,0.1366215795,-0.2497364134,0.0172254704,0.0104196006,-0.0525608733,0.7789852023,0.1548537612,0.0441102199,0.4680342674,-0.0037566889,0.3014906645,-0.0559014827,0.0838298053,-0.3072827458,-0.2321264148,-0.1228302494,-0.2282546759,0.1536598504,0.2682596445,-0.2145921886,0.3666175008,-0.0568442829,0.1451283544,0.0025314549,0.3421612084,-0.0584325343,0.0655459985,-0.0626575202,0.2040625513,-0.1444291919,0.409722656,-0.1356842667,-0.149004966,-0.0768997669,-0.2596896887,-0.2632807493,0.1581942588,-0.3894182444,0.3009178042,0.0940063596,-0.0541631989,-0.0775868669,-0.0720009059,0.0379504673,0.0387830436,-0.1752779186,0.0844960958,-0.1288580447,-0.2365467399,-0.0595169775,-0.1094142944,0.4144064486,-0.2647612989,-0.2428983152,0.0964698642,0.0492814519,-0.2107218653,0.1066624746,0.0112686176,0.2041609436,-0.3407523632,-0.2518103719,-0.0327805951,-0.1775391549,-0.12553069,0.186204195,0.0360444263,-0.0364982747,0.2830505073,-0.0620818809,-0.1947600096,-0.1506979316,0.351208955,-0.1067584306,-0.1233872399,0.4347908795,0.072054632,-0.3199335039,-0.2920496762,0.0793827996,-0.0465265624,-0.4147839248,0.0896514952,0.0486419089,0.3403031528,0.0655375123,0.1039589494,0.1910346895,0.044651486,-0.0263227336,-0.5710043907,-0.2213813961,-0.0354618803,0.035441488,0.2289005071,-0.094972007,0.3262064159,0.2034397423,-0.0655390695,-0.3433675766,0.1364937276,-0.095421575,-0.0452936105,0.1359447986,-0.1490375251,0.4430599213,0.1009041294,0.1978885531,0.0278638974,-0.1176836565,-0.2749330401,-0.1282685101,0.1190886796,0.167985335,-0.1261828542,0.0420026705,-0.1996611506,-0.3425234854,0.1667084098,0.0233251974,0.1777889729,-0.0187482443,0.0477432758,0.1527056545,0.057696756,-0.1524341851,0.0197509918,-0.0933409557,0.3424445093,0.2201880962,0.2526732683,0.1092485264,-0.028304074,-0.4353369474,-0.0889778137,0.0215206128,-0.1089897975,0.3970861137,-0.2267069072,-0.0844079256,0.1168391258,0.2218981236,0.3427518606,-0.2445510477,0.0473774672,0.3164501786,0.2502194941,-0.2942929864,-0.074464418,0.1661240906,0.1107305214,0.1122423261,0.0778133348,-0.2546724975,0.0717090666,0.0237519108,0.1373420209,0.3921376169,-0.0400031582,0.1250549704,0.3877003193,-0.1304763705,-0.0067147673,0.1958841681,-0.0121206986,0.070871979,0.4315577447,-0.2908119261,-0.0119646117,-0.172923103,-0.0734783411,0.2268629521,-0.2598978579,-0.0039406689,0.2269203067,-0.118860811,0.0853064358,-0.0197697077,0.1014457345,-0.0209625904,0.1261386722,-0.4866469204,0.3916065395,-0.0998357832,0.3023129702,-0.2309093624,-0.132981807,-0.2292218655,0.1452821046,0.1043042019,-0.1691759378,0.2273527682,0.3998037577,-0.2776856422,-0.1550411731,0.0752838925,0.3108185828,-0.0813329294,-0.2761153579,0.1822858006,0.3667353988,-0.0344139636,-0.0020702961,0.4475148022,0.3937963545,0.2824236155,0.0474153645,0.2371765971,0.1083737686,-0.1588353515,0.0991288722,0.1113982573,0.2268317044,0.347774893,0.3200223744,0.2253325135,-0.2515419424,0.0480179787,-0.0931240693,-0.0826394707,-0.2318333983,0.5614549518,0.0496760122,0.0397591777,-0.2407839596,-0.0536701046,-0.4856576324,-0.0348074809,0.3326095641,0.1133724004,0.0982405171,-0.1042847708,0.0643889904,-0.0975297689,0.3120363653,0.3321331441,0.2571236491,-0.6201792955,0.004288333,-0.6339927316,0.1701634973,-0.1985664666,0.0743850023,-0.0166364368,0.0412023179,-0.2311744988,0.3115764856,0.2193800658,0.0395457298,0.0116452165,0.2519503236,-0.1994519979,-0.0294276495,0.1661611497,-0.1432815492,0.0830325782,-0.4240460098,0.0994042978,-0.4774453342,0.017368475,-0.0443782322,-0.3066776097,-0.0353295766,0.2529799342,0.4154708385,0.0541495681,0.7115378976,-0.0803386495,-0.2179797292,-0.2073084563,-0.0156711414,-0.2390040755,0.1551925689,0.1466782242,0.3737408221,-0.0241130292,-0.1898607612,-0.417525202,0.3508703113,-0.107716091,-0.1765215695,-0.2167667896,-0.0179202445,-0.1924640536,0.2424509078,0.0509917289,0.0850015953,-0.0730701983,0.1792785823,-0.3576475084,-0.4977722764,0.6089873314,-0.2351760268,-0.2108120024,-0.2193452865,0.1239553168,-0.117237106,0.1133626476,-0.497730881,0.3211047649,0.3238481283,-0.0129207922,-0.1095672026,0.2009022683,0.2233956754,0.2048543543,0.0640877187,0.3913771808,0.1503724754,-0.3911642432,-0.0575917885,-0.2036331147]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/861","title":"Possible Bug: Small training\/dataset file creates gigantic output","comments":"The preprocessing tokenizes the input text. Tokenization outputs `input_ids`, `attention_mask`, `token_type_ids` and `special_tokens_mask`. All those are of length`max_seq_length` because of padding. Therefore for each sample it generate 4 *`max_seq_length` integers. Currently they're all saved as int64. This is why the tokenization takes so much space.\r\n\r\nI'm sure we can optimize that though\r\nWhat do you think @sgugger ?","body":"Hey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n","comment_length":58,"text":"Possible Bug: Small training\/dataset file creates gigantic output\nHey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n\nThe preprocessing tokenizes the input text. Tokenization outputs `input_ids`, `attention_mask`, `token_type_ids` and `special_tokens_mask`. All those are of length`max_seq_length` because of padding. Therefore for each sample it generate 4 *`max_seq_length` integers. Currently they're all saved as int64. This is why the tokenization takes so much space.\r\n\r\nI'm sure we can optimize that though\r\nWhat do you think @sgugger ?","embeddings":[-0.1432341933,-0.3876414895,0.1133134663,0.3227741718,0.5927131176,-0.1981907338,0.2874234021,0.3909299076,-0.3091718554,0.1720577329,-0.0339277796,-0.1797885746,-0.3273913562,0.1897948235,0.1685712636,-0.0655538663,0.1466006935,0.116178818,-0.0373176374,-0.2512279153,0.0122301606,0.0331541412,0.2052446604,0.2363262922,-0.5892032385,0.116232723,-0.0976605043,0.0225226693,-0.4522168636,-0.0920315832,0.0688617229,-0.3121908009,0.1892007589,0.5060113072,-0.000128655,-0.0801775157,0.1424444616,-0.0695752501,0.0205701813,0.2090911865,0.2144756615,-0.0632570237,-0.2133893222,0.1053180844,0.215238288,-0.1143107042,-0.0709828362,-0.0340408124,0.3668509126,0.1840243191,0.0656992346,0.1355391443,0.0148969349,0.1036687791,0.2429664731,0.3396823704,-0.0371804014,-0.0100553706,0.2122915089,-0.1989565939,0.0148402276,0.1700541526,0.0983981639,-0.1982290298,0.3284576833,0.211391449,-0.0749181658,-0.4113597274,0.2120729089,0.3883528709,-0.0406982489,-0.3539565206,-0.3082945943,-0.3972704113,0.2411764115,-0.5351740122,0.0933327749,0.1817525029,-0.1299932003,0.0571895428,-0.2757267952,-0.4147567749,-0.2204075903,-0.0975380242,-0.2143557817,0.0497394353,-0.2284279913,-0.0945073366,-0.0247093644,-0.1514883786,-0.3679102659,-0.0261253119,-0.0072875586,0.1992163956,-0.0928649604,-0.324272722,-0.1126561165,-0.3281558454,0.2717485428,-0.4419617057,-0.2412327826,-0.4030223191,-0.1295828819,-0.4189319313,0.3993144631,0.45951581,-0.3760156333,0.1799371392,0.129983902,0.0144979181,-0.1388451308,-0.0527608991,-0.1634542048,0.1329072416,0.0710657537,-0.4047501087,0.0313030481,-0.2290522456,-0.1003042832,0.4591183364,-0.3784509897,0.2226309627,0.0510771945,0.3274946213,-0.0760810301,0.1053190678,0.0109074526,0.051346615,-0.4180237353,-0.0177939795,-0.0611290112,0.0742200837,-0.2151125818,0.1869876385,-0.0288434997,0.1044516861,0.021352781,-0.0179804843,0.3231276572,-0.099340573,0.0395014994,-0.2510479093,0.1726971269,0.1685138196,-0.2115899771,0.4402742386,-0.032869909,0.0573067553,-0.057333827,0.0731936917,-0.2025119215,-0.4613577127,-0.0575413331,0.0954975709,-0.0324038714,0.3140635192,-0.0619527102,0.3278669119,0.4565284252,0.0198546946,0.0250289068,-0.2284952551,-0.2265143692,0.0091150738,0.3995687664,0.3902240396,-0.484636277,-0.1758751422,-0.2054727226,0.3122827709,0.3894993663,0.5888673663,0.0077837189,0.5783097744,-0.2811872363,0.2357924432,-0.2949523628,0.2546403706,-0.4308692813,0.3521903157,-0.0578786768,-0.0545474775,0.0877296999,-0.0587389097,0.1824905276,0.0208971072,0.3823952079,0.0333050042,-0.1517231613,0.0272625796,-0.3604698479,-0.4016648233,-0.2655518949,0.0143587757,0.1724800318,-0.1572998762,-0.3328549564,0.5829175711,0.1297849119,-0.1000100225,0.1598347723,0.3174715042,0.0870409384,-0.038315177,0.0524265729,-0.0793230012,-0.1097159609,-0.0088055804,-0.307941258,0.2040724009,0.1063894778,-0.2711715996,-0.0457132347,0.0320105255,0.0785083622,-0.206470862,-0.0180337559,-0.1957940906,-0.1108462065,0.0106169786,-0.0733258352,-0.1052491441,-0.0842952281,0.0720166713,-0.4379627407,-0.0457763746,-0.0231428873,-0.1467807889,-0.0945440531,0.360575676,-0.1125401333,0.0251169559,-0.0708715022,0.1594514549,0.1305894256,0.0670296475,-0.0969280303,0.2377001047,0.2820757926,0.0817722231,0.1023699269,0.1841624379,0.0422055312,-0.0524386503,-0.1178935096,0.2136938721,0.2088421732,0.1476619989,0.241948694,-0.2180727869,-0.1030561328,0.1184765771,-0.044392433,-0.2775182724,0.3078273535,0.0634442493,0.2897989154,0.1936533898,-0.199255988,-0.0915348753,0.2885442376,-0.1878994107,0.0241267122,0.4312776625,-0.3436952829,-0.1197063774,-0.1063292995,-0.2838623822,0.2830734253,0.0803629383,0.1233097166,0.1092917994,0.1420129687,-0.0746385753,0.2112692147,0.2753331065,-0.1940709949,0.1610940397,0.1346510351,0.0592170246,-0.1989697665,0.0861475393,-0.1964784116,0.3385744989,-0.4070677757,0.3386442363,0.0577274486,0.0137761049,-0.3941378295,-0.2747454345,-0.1809253544,0.0264378972,0.1926738918,-0.2345577329,-0.1110428274,-0.052663371,0.3810812831,0.2989987433,0.1577854604,0.2653697729,0.0186494663,0.1821882874,-0.3386882246,-0.0294872876,0.1913497746,-0.1869173199,0.1518467218,0.0983013734,-0.2298426032,-0.0564885587,-0.3236949444,0.2155354023,-0.1799170524,0.279482007,0.2024963349,0.1363641471,-0.4355506599,0.094604522,0.0197804589,-0.0027538224,-0.1579777598,-0.014179674,-0.1234321445,0.1277320832,-0.0837785304,-0.2641698122,-0.0601658933,0.0195474718,0.3254421949,-0.0098445956,0.0360103473,0.211654976,0.1585821211,0.1004245207,-0.0514147468,-0.1972807795,-0.0010389336,-0.1854279786,0.2125998437,0.0062263282,-0.1962724924,-0.3562273085,-0.053124927,0.0255862307,-0.1569030285,-0.5457936525,0.0518393554,-0.4050366879,-0.3002316356,-0.2334708869,0.3785063624,0.1156734899,-0.107770294,0.0849830359,0.0984227061,0.1134190038,-0.0752596483,-0.2457754016,0.2301739305,-0.4163112342,0.3523648381,-0.0792582184,0.878788054,0.3173368573,-0.0099722948,-0.0350926481,-0.0326418132,0.399187237,-0.4704627395,-0.1981317103,0.2872303426,-0.0592223033,-0.3348690271,0.3239846528,0.1225906536,0.4575767219,0.2133401185,0.061285805,0.280878216,-0.3745425642,0.4108827412,0.133876279,0.176882565,0.0014608199,0.1307318211,0.0162938442,-0.3149764836,0.0631069988,-0.0538524538,0.2175052911,0.0166281685,0.0180319045,-0.3413485885,-0.6975656748,-0.0455707125,0.1327458173,0.0729151219,0.1507203579,-0.1237143651,0.1899592876,0.046843525,0.7578899264,0.2813235819,-0.1726810485,0.0055504846,-0.0681674704,-0.1512740552,0.0470158905,-0.3141689897,-0.0336300023,0.565553546,0.6516923904,-0.3451510966,-0.0619588532,0.1178143099,0.3456683755,-0.1228976697,0.0067761713,-0.42242226,0.0064462442,-0.2393390685,0.2443727553,0.0683157295,0.0679180324,-0.0780402645,-0.0828614309,0.0585738942,-0.0831706747,0.3588627279,0.2406857163,0.3210001886,-0.2953753471,0.3806355596,-0.0379775725,0.1155733913,0.3630248308,0.1936297864,-0.1670265049,-0.2968226969,0.2599287033,-0.1010311842,0.3370382786,0.5717677474,-0.0340764746,0.1766499728,0.1356079429,0.4558725655,-0.4732834399,0.5112178326,0.4902334809,0.0128841186,-0.5175443292,-0.2263826579,0.3698660433,0.0831365436,0.1231708154,0.1651781052,0.3095340431,-0.3091195524,0.2911875844,-0.001634889,1.1803641319,-0.1970944852,0.5995185375,0.1163959801,0.2842214406,0.585283339,-0.2833365202,0.1925698221,-0.3745793402,-0.062271297,0.1641652733,0.0342524238,-0.1442031413,0.0297324173,-0.1081283987,0.1598674208,-0.1128819212,0.2599046826,-0.2608391643,0.3257431984,0.0613674447,-0.3731700778,-0.23922351,-0.0527997129,0.0120127527,0.0464326032,-0.1548966765,0.0905369297,-0.0296992082,-0.3042659461,-0.2594524026,-0.1731944829,-0.0260345396,-0.038991753,0.1994987726,-0.3362631202,0.3272384405,-0.0302609019,0.2560785711,0.659665823,-0.2298511565,0.0539064072,-0.1808632463,0.0355850644,-0.0185966324,-0.0847667977,0.0327786952,0.0537502952,0.070880793,-0.0461536907,0.0136144022,0.2896091044,-0.3708911836,-0.1865696609,0.594525218,-0.480150044,0.0140863629,-0.1068132073,-0.0510032065,-0.1402089149,0.0055736033,0.1723662764,-0.2537513375,0.0729223713,0.0433618911,-0.3664639592,-0.2403925657,0.1738371402,0.2803501785,-0.0036435905,0.7379029393,0.2823022604,-0.071955815,-0.0889389291,0.4287666976,0.5030834675,-0.4444898069,0.2955849469,0.1780548096,-0.0375426374,-0.0009512292,0.1282506883,0.2379667759,-0.3074365556,-0.1623658389,-0.0494091921,-0.4165500104,0.0472992547,-0.0431369953,0.1312137395,0.1436967105,0.348347038,-0.0929587632,0.1567374617,-0.1273813844,0.0535807721,-0.0168821495,0.0394775234,-0.2568358183,-0.0548397191,0.1413848847,-0.1418828964,-0.0923596397,0.2089784145,-0.019302398,-0.0735148787,-0.2762777805,0.1742604077,0.0180754606,-0.0971492305,0.1658904403,0.1317505389,-0.1845954508,-0.2557170689,0.3654860854,0.1042429507,0.1710370928,-0.2266407311,0.1208520979,-0.0682145804,-0.1481811404,0.0164354593,0.2539683878,0.3333453834,0.0642728731,0.1704205573,0.2489991039,-0.0927020609,-0.0580304191,0.259157002,0.2553392947,0.0652257279,0.118090786,-0.222453773,-0.0711682811,0.0400911979,0.1688091755,0.4920586944,-0.2982497215,-0.2490655184,0.4635601044,0.0708697662,-0.0373559669,-0.077276893,-0.1433793753,-0.2336226255,-0.2046160251,0.3197961748,-0.0375347398,0.0652030706,-0.1347514391,0.1632535458,-0.1233683974,0.1600246578,-0.0292368606,0.5991111398,0.0031000432,0.1194957271,0.2187775373,0.1010295376,0.2389773428,0.6983104348,0.1785127074,0.1925281286,-0.3714151978,0.1088145822,-0.1927821636,-0.4904711545,-0.461263746,0.1844991595,-0.1447879523,0.2918328643,-0.1545111388,0.58831954,-0.0762196705,-0.249321565,-0.1546923071,-0.0103586763,-0.2898415327,-0.1191217527,0.1923538595,-0.1257757396,0.1729236692,0.2596201599,-0.1238780469,0.0102518322,-0.0933727399,0.0794774368,-0.1036153063,0.0493975617,-0.1820998341,0.1839226186,0.1136898473,-0.2876952887,0.2269536257,-0.0806936249,0.0367100462,-0.2104424536,0.2947309315,0.4753099084,0.3873186707,-0.2240899354,0.16996108,0.1405054778,0.0864484981,-0.0693234876,0.1638728827,0.3289234638,0.120653078,0.0141667435,-0.0846549943,-0.1228396893,-0.1579105109,0.2446879447,0.1013684794,-0.1489238292,0.2284968942,-0.1582212597,-0.1319342107,-0.2944997847,0.0270951744,-0.7009481192,0.5352689028,0.4764958322,0.0252955984,0.123450987,-0.0915325209,-0.0095249433,-0.1323713362,0.5158675313,0.4690684676,-0.0139554748,-0.3866828084,0.0308354758,-0.2977653146,0.2726365626,-0.2909190059,0.314070493,-0.1308755428,-0.0755323321,0.0208769012,-0.139403671,-0.0827983394,-0.2343416214,-0.6251960993,0.1800921857,-0.1407461911,-0.1719839722,-0.2263513803,0.0863991156,-0.0605694279,-0.2572997212,0.4457692504,-0.0461422764,-0.1705084741,0.0293919723,0.2279556096,-0.2168888301,0.1738213003,0.2571803331,0.2815208137,0.0494486503,0.0635209978,-0.2289598137,-0.2911761403,0.1392115802,-0.0718203187,0.1263869852,0.2461519092,0.2943047881,-0.3145042956,-0.264855653,-0.3899097443,0.0929283351,0.3281577229,-0.2242101431,-0.2443577051,0.1015164852,-0.1149020717,0.0344204642,0.0135651594,0.4777696133,-0.1209277958,-0.0470704995,-0.3605847657,-0.2439097911,0.4634321928,-0.3155981302,-0.4158999622,-0.2619010806,0.0782352015,0.3039796948,-0.2966359258,-0.7923548222,0.101564981,0.1591156572,0.0253923647,-0.4467414021,0.1350894868,-0.1159248501,-0.1337082386,-0.1870725751,0.7428531051,0.0427238494,-0.1907565296,0.4248633087,-0.2602861524]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/861","title":"Possible Bug: Small training\/dataset file creates gigantic output","comments":"First I think we should disable padding in the dataset processing and let the data collator do it.\r\n\r\nThen I'm wondering if you need attention_mask and token_type_ids at this point ?\r\n\r\nFinally we can also specify the output feature types at this line https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py#L280 to use more optimized integer precisions for the output. Maybe something like:\r\n- input_ids: uint16 or uint32\r\n- token_type_ids: uint8 or bool\r\n- attention_mask: bool\r\n- special_tokens_mask: bool\r\n\r\nAlso IMO these changes are all on the `transformers` side. Maybe we should discuss on the `transformers` repo","body":"Hey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n","comment_length":90,"text":"Possible Bug: Small training\/dataset file creates gigantic output\nHey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n\nFirst I think we should disable padding in the dataset processing and let the data collator do it.\r\n\r\nThen I'm wondering if you need attention_mask and token_type_ids at this point ?\r\n\r\nFinally we can also specify the output feature types at this line https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py#L280 to use more optimized integer precisions for the output. Maybe something like:\r\n- input_ids: uint16 or uint32\r\n- token_type_ids: uint8 or bool\r\n- attention_mask: bool\r\n- special_tokens_mask: bool\r\n\r\nAlso IMO these changes are all on the `transformers` side. Maybe we should discuss on the `transformers` repo","embeddings":[-0.2299046516,-0.3667338192,0.1243235469,0.2771853507,0.6030983925,-0.1695275754,0.3121454716,0.3734357655,-0.2599174082,0.1522065699,-0.058923617,-0.1322191358,-0.2868515849,0.2487872392,0.1245998815,-0.1173472181,0.1013758779,0.1097865552,-0.0822892934,-0.2580442429,-0.0382415242,-0.0007749674,0.2506649494,0.2208165973,-0.5668442249,0.0628852993,-0.0407152809,0.0689602792,-0.4490863383,-0.0693647936,0.0581910834,-0.2989962697,0.2023249865,0.5197603703,-0.0001248772,-0.050877206,0.1890182644,-0.0757151917,-0.0064578718,0.1778805852,0.1570471227,-0.0995637178,-0.1501622796,0.0919623673,0.2281863689,-0.0761860162,-0.0426351726,-0.0638567582,0.3386057615,0.2224294841,0.0923961699,0.1321176738,0.0814070776,0.0844403207,0.2095718682,0.3288491666,-0.052218359,0.014561533,0.2090313584,-0.1524130702,-0.0214012172,0.1811145246,0.1139105558,-0.1699478179,0.2978738248,0.218269974,-0.1149848625,-0.4121504724,0.1769650728,0.4269205034,-0.058446575,-0.376488179,-0.3386244774,-0.4122276306,0.2050081193,-0.5156775713,0.0491871461,0.1984972656,-0.1357265562,0.0747207999,-0.2738565207,-0.4056916833,-0.2355403304,-0.0613149367,-0.1954496652,0.0406941809,-0.2122223973,-0.0865625739,0.0166964587,-0.1147670746,-0.3613003492,-0.0058366884,-0.0146025326,0.2444557399,-0.0902943313,-0.3005383015,-0.0907191262,-0.3367122412,0.2940711975,-0.5009034872,-0.1907384247,-0.3996056914,-0.1979431659,-0.3644928932,0.379765898,0.4768549502,-0.3775405288,0.2085641325,0.1635924876,0.0616293512,-0.1025520265,-0.0461832546,-0.107098639,0.1027833298,0.0321718752,-0.3345995247,0.0909971446,-0.2671134472,-0.1278442144,0.4031009674,-0.3146030605,0.2281466275,0.041138608,0.3260876536,-0.0870726928,0.1651624888,0.0686943606,0.0581925884,-0.4440341294,-0.0256729517,-0.0942217112,0.0886550695,-0.1856420338,0.1809190959,0.0117867105,0.0643380433,0.0366569385,0.0092921248,0.2664726973,-0.0247753523,0.0805954635,-0.2521767914,0.1330581009,0.169172138,-0.1863246858,0.4540976584,0.0054919026,0.0517551638,-0.0362077318,0.045383174,-0.210335061,-0.4931018651,-0.0973958001,0.1252601594,-0.0389679,0.2960734665,-0.0794365406,0.3503779471,0.4486837983,0.0162848309,0.0150420386,-0.192922011,-0.2002844214,-0.0125840828,0.3824052215,0.4193097055,-0.5442303419,-0.1573411673,-0.2385899127,0.2531915307,0.3889272809,0.6155259609,-0.0056070806,0.5517367721,-0.2353394628,0.1642785668,-0.2852040827,0.2683778107,-0.4222464859,0.3572139442,-0.0525044538,-0.0388309173,0.1061370522,-0.0538562946,0.174445644,0.0127832154,0.3944090009,0.0238885619,-0.2038847804,0.0473412313,-0.3813453913,-0.4033684731,-0.2660674453,0.0800017491,0.1581137478,-0.1353046745,-0.2735057175,0.5514494777,0.1155518964,-0.1455298513,0.1355006248,0.3262093663,0.1113239229,-0.0687134787,0.019378813,-0.0532411262,-0.2125861496,0.0344204605,-0.2865603566,0.2489629984,0.0793814436,-0.2721630633,-0.0573521294,-0.0195421576,0.0813331082,-0.2065415829,0.0124215931,-0.2158961296,-0.091592364,0.031171117,-0.0479382686,-0.0457130112,-0.102979362,0.048843503,-0.502892971,-0.0453934222,-0.0081499685,-0.1882385015,-0.0324201174,0.3290528059,-0.0995146036,0.0423071869,-0.1000375897,0.1843878776,0.1257511675,0.0820540935,-0.0610663593,0.230011031,0.2973825634,0.1116479784,0.0934734419,0.2055562884,0.0610838011,-0.051709231,-0.1138145626,0.2268062383,0.1831572205,0.1398509592,0.2676385939,-0.2479642332,-0.0863999426,0.1264736503,-0.0272574909,-0.2584272325,0.311812073,0.0601300485,0.307054013,0.1462773681,-0.205597192,-0.0656072423,0.3099732995,-0.1693716198,0.0509300604,0.4050944149,-0.3019103706,-0.0469115749,-0.0779378414,-0.2446383387,0.3169588745,0.1027565002,0.0909688696,0.130417943,0.0760696381,-0.08735165,0.1489487588,0.2738137245,-0.1550520211,0.1503685415,0.1463555843,-0.0123092178,-0.2240320295,0.0660324916,-0.2190813124,0.3789514303,-0.389223367,0.3302790225,0.090871118,0.0197698418,-0.4125824571,-0.284348011,-0.2321334779,0.0014972417,0.1670883149,-0.1848983616,-0.1268641502,-0.068473503,0.3420519233,0.2327810973,0.1725662351,0.2960615158,-0.0134932948,0.1116621047,-0.330593437,0.001932239,0.1982810497,-0.1490035057,0.1594868749,0.10087277,-0.2657164037,-0.0178066939,-0.3181884289,0.1642902642,-0.1982602626,0.2966389656,0.2247881889,0.1824064404,-0.3565532267,0.0378920175,0.0556856766,-0.0080807311,-0.1318658739,-0.030173799,-0.1717743576,0.171892643,-0.0972635821,-0.3041326106,-0.1180305183,-0.0744101256,0.3672360182,-0.0105409715,0.0147801014,0.2039791197,0.1889281869,0.1171235368,0.0248135868,-0.167905122,-0.0129318396,-0.1366127878,0.2170595676,-0.0020629296,-0.217152968,-0.3731190264,-0.0392452739,0.0291202646,-0.1878812313,-0.5029156804,0.0255237203,-0.4563643038,-0.2719499767,-0.2287575155,0.4256253541,0.0850238502,-0.1087955683,0.0785249397,0.082861878,0.135474056,-0.0338128507,-0.2376572937,0.2106582224,-0.4013062119,0.4056994319,-0.1035357341,0.888209939,0.3483371735,-0.0970775336,-0.0007676132,0.0034505122,0.3610590994,-0.4532393813,-0.2227277309,0.2996627986,-0.0941919982,-0.3264152408,0.2661916614,0.0962236002,0.4384837151,0.1695129573,0.0531374402,0.2463840544,-0.3661045134,0.4002577066,0.1699473858,0.1683425456,-0.0398625024,0.0964051634,-0.0057650735,-0.3088213801,0.0813731328,-0.0367562883,0.190827325,0.0290936269,0.11779809,-0.3415733278,-0.7532606125,-0.0089538433,0.1532117426,0.0726144537,0.1456365883,-0.1499405652,0.1320814788,0.0780924931,0.7480297089,0.2280015051,-0.1866868287,-0.0243703742,-0.0419677943,-0.1549034417,0.0467049442,-0.3488087356,-0.0722114816,0.6041032672,0.6499783993,-0.4103456736,-0.0432791151,0.1184371486,0.3192984462,-0.1154529825,0.017846996,-0.4534131587,-0.027014995,-0.2239370793,0.2107045949,0.0687585473,0.0523454323,-0.1507504731,-0.0843717679,0.0828976184,-0.0923526436,0.3055901527,0.3103858531,0.2941124737,-0.3023830652,0.3669596612,-0.041695483,0.1407231241,0.3650065958,0.3272884488,-0.2009644359,-0.3620548546,0.2555139363,-0.1160360724,0.3236405253,0.5161673427,0.0053906804,0.1600106359,0.1557781994,0.457516402,-0.4653674364,0.5355374813,0.4863606691,-0.0177372526,-0.5349086523,-0.2112374306,0.4121993184,0.0721789524,0.1378423274,0.1906249374,0.2771161199,-0.3334349096,0.2811171114,0.0121667329,1.1012880802,-0.2145240605,0.5295492411,0.1170042232,0.2662311196,0.6018490791,-0.2290400565,0.1664569527,-0.3895483315,-0.1080549955,0.1780197173,0.0635723323,-0.1201989353,0.0220747553,-0.1803472787,0.2107702196,-0.0939509049,0.1979261786,-0.2349267155,0.414270997,0.0643136427,-0.3886007965,-0.2565069199,-0.0330815017,-0.0308762174,0.0132635646,-0.1502638608,0.0457156785,-0.0300234556,-0.3175837994,-0.2670261562,-0.1795191169,-0.1207991019,-0.0250809509,0.1849127412,-0.3685482442,0.3284722865,-0.0686188713,0.2967181802,0.594273746,-0.241095975,0.0316297412,-0.1764055938,0.0335868858,-0.0160011947,-0.0686834604,0.0461450331,0.057219293,0.0298611801,-0.0124811186,-0.0009000049,0.3430385292,-0.3757948577,-0.2448651791,0.4503363669,-0.5107018948,0.0249905027,-0.1506493837,-0.0612496249,-0.1385045797,0.0316618942,0.1808363199,-0.2773392797,0.0250360705,0.022024624,-0.3429327905,-0.2632741034,0.1784669012,0.2554321289,0.0376856141,0.654103756,0.3423447907,-0.0658519641,-0.0643092692,0.3937629759,0.5364499688,-0.5111446977,0.2903211415,0.193658188,-0.0504794121,0.0055523473,0.1892943829,0.2149124891,-0.2882865369,-0.1783059686,-0.0321891047,-0.426137507,0.0959350318,-0.0664637238,0.1580213159,0.1405061334,0.3553335965,-0.0714857876,0.2126518935,-0.1578660458,0.0235701036,-0.0105698984,0.0625283048,-0.1994101852,-0.0494636483,0.1744731069,-0.1104822457,-0.0572031066,0.2021813542,-0.0392469503,-0.1036099643,-0.2420643419,0.145062834,0.0128965676,-0.0629804954,0.1344017982,0.1039877012,-0.16123721,-0.2328151911,0.407618314,0.0781859159,0.1133873537,-0.1191850528,0.0942145288,-0.0924133435,-0.0830872804,0.006390485,0.1768653542,0.3059254289,0.0771718547,0.1720285118,0.2513033748,-0.084110111,0.0164577197,0.2219178975,0.1959113032,0.1082604527,0.1113377213,-0.2297803015,-0.0865420401,0.0216388684,0.2315419316,0.4675869048,-0.2486834824,-0.2054344714,0.433788836,0.1126068383,-0.0412768573,-0.0754648298,-0.1049243584,-0.2165251076,-0.0906258747,0.3524930179,-0.0557722189,0.0519348197,-0.1145399436,0.1457198262,-0.1238848716,0.1627272367,-0.0535479225,0.6416904926,-0.0037555725,0.0996095985,0.2106966227,0.0525346659,0.2626218796,0.6772089601,0.1848816723,0.1508556604,-0.421551913,0.0804427117,-0.1494599283,-0.5071752071,-0.4876213968,0.1039029658,-0.1718595326,0.3066627979,-0.1484480053,0.5831757188,-0.0919552818,-0.2220142484,-0.1316652298,-0.03544062,-0.2864599824,-0.0965777934,0.1660562009,-0.1335314512,0.1358466893,0.2186131626,-0.1328592896,0.0446251556,-0.0361390747,0.0860942379,-0.0728239492,0.0552888326,-0.2137135565,0.1781378984,0.1519741714,-0.2392749488,0.2329532355,0.001659583,-0.0107983621,-0.2549520731,0.319134593,0.4827312529,0.3858772516,-0.2447324842,0.2269487232,0.1933878213,0.0481451154,-0.0690272152,0.1335735023,0.3105468154,0.1057750061,0.0640176684,-0.0337164663,-0.1393623054,-0.1620503068,0.1474623084,0.087148793,-0.1255704612,0.2617918253,-0.1687384546,-0.137326926,-0.275339663,0.0100973481,-0.7297048569,0.5182319283,0.4826716483,-0.0365395732,0.1011706218,-0.0945149064,0.0048824186,-0.1210666969,0.5459296703,0.4561127722,0.04383073,-0.4208503067,0.0229861643,-0.2733936906,0.2850801647,-0.3483309746,0.2294561565,-0.1512725204,-0.0749797076,0.0398242399,-0.0809344947,-0.0586804859,-0.2328859866,-0.6350175142,0.2164501548,-0.1353760809,-0.2508300841,-0.3114378154,-0.0002751876,-0.0719832405,-0.2702170014,0.433116734,0.0030057742,-0.1167681292,0.0144516835,0.2491118312,-0.188437745,0.1153704301,0.2229571044,0.3003044724,0.0588106401,0.0025346812,-0.2231749892,-0.2673762143,0.1154301167,-0.1128506586,0.1092515737,0.2529754341,0.3821110427,-0.2819197476,-0.2846087217,-0.4440238774,0.160567075,0.3504857719,-0.1855471581,-0.2507892251,0.1337047815,-0.1008124426,-0.0683623403,0.0340645537,0.4943032861,-0.0824924856,-0.04165411,-0.3408871591,-0.2351451814,0.4602756202,-0.3019508719,-0.3991840184,-0.237896502,0.0669847056,0.2551595569,-0.2525520623,-0.8035458922,0.1077653319,0.1308983117,-0.009748105,-0.4651971757,0.1387013942,-0.0748908967,-0.1406367868,-0.1919123083,0.6681408882,0.0604233854,-0.1583125591,0.4259201586,-0.25872612]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/861","title":"Possible Bug: Small training\/dataset file creates gigantic output","comments":"> First I think we should disable padding in the dataset processing and let the data collator do it.\r\n\r\nNo, you can't do that on TPUs as dynamic shapes will result in a very slow training. The script can however be tweaked to use the `PaddingDataCollator` with a fixed max length instead of dynamic batching.\r\n\r\nFor the other optimizations, they can be done by changing the script directly for each user's use case. Not sure we can find something that is general enough to be in transformers or the examples script.","body":"Hey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n","comment_length":91,"text":"Possible Bug: Small training\/dataset file creates gigantic output\nHey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n\n> First I think we should disable padding in the dataset processing and let the data collator do it.\r\n\r\nNo, you can't do that on TPUs as dynamic shapes will result in a very slow training. The script can however be tweaked to use the `PaddingDataCollator` with a fixed max length instead of dynamic batching.\r\n\r\nFor the other optimizations, they can be done by changing the script directly for each user's use case. Not sure we can find something that is general enough to be in transformers or the examples script.","embeddings":[-0.2449227124,-0.4394875467,0.1173569635,0.318944186,0.5710244775,-0.1765770465,0.2565920949,0.4291751087,-0.3390790522,0.1573361307,-0.0675725713,-0.2011723965,-0.3087227643,0.1915093958,0.1266338229,-0.1478894055,0.0698192343,0.173173815,-0.1252346039,-0.221469,-0.0003152027,0.0349694043,0.2422268838,0.1785005182,-0.5778369904,0.0762562156,-0.07158079,0.0110261906,-0.3461946249,-0.0348516442,0.0994135439,-0.2865832448,0.2010674626,0.5267562866,-0.0001265749,-0.0411316752,0.1517601311,-0.0826687366,0.0461400636,0.1951688379,0.1975021064,-0.085682258,-0.1636204869,0.0772056431,0.1840523332,-0.0889649913,-0.0251297597,-0.003594059,0.3703834713,0.2273509502,0.081475459,0.221255213,-0.0237424318,0.0711170435,0.2784283757,0.2887702882,-0.0818410516,-0.0509197935,0.2467285991,-0.2162994444,-0.0755180046,0.1138721183,0.0970574841,-0.1382569373,0.2942581475,0.2418761551,-0.0901601389,-0.4337415099,0.2268631905,0.4442629814,-0.0038127259,-0.3651985526,-0.3235093951,-0.4126268923,0.2002189159,-0.4695293903,0.0931720212,0.1572839916,-0.1708563268,0.0714475214,-0.3042104244,-0.3376942277,-0.2605009377,-0.0602699406,-0.1891828179,0.0072341273,-0.2047071904,-0.129764989,0.0258300323,-0.0898103267,-0.2813793123,0.0086461799,-0.0054782415,0.2329305857,-0.087534532,-0.2953803837,-0.0820568353,-0.3209120929,0.2784362435,-0.480804354,-0.1911471188,-0.4234406352,-0.1217624322,-0.4122720957,0.4198234081,0.4721465409,-0.4405255318,0.1479226053,0.1813026667,0.0140898079,-0.1194204241,-0.0408408679,-0.1277310997,0.0868289918,0.092828609,-0.4029773772,0.0393005759,-0.2354254425,-0.0472667664,0.3343820572,-0.2421362698,0.1646773815,-0.0149446558,0.3203733265,-0.0953860432,0.2072838396,0.0915634334,0.0109717743,-0.4334857762,0.0280755404,-0.0920762643,0.0704810768,-0.2314606011,0.1855055392,0.0562622212,0.0775650963,-0.0132145351,-0.0600127131,0.3360011876,-0.0606572628,0.0315354392,-0.244994998,0.141784519,0.218581751,-0.1720943898,0.3873316646,-0.0372590832,0.1247835383,-0.0676249266,0.1088969857,-0.1717208922,-0.5269572735,-0.0558348522,0.0992493704,-0.0632847399,0.2949822247,-0.104798317,0.3964322209,0.46083197,-0.0342460908,-0.0337650031,-0.2015198618,-0.2432101816,-0.0263377503,0.4121342599,0.49381423,-0.5112314224,-0.1573763043,-0.2400308698,0.2583684623,0.3261887133,0.6362030506,-0.0559865832,0.6048205495,-0.2477885038,0.1249576658,-0.307009846,0.2827396691,-0.4184560478,0.3741445243,-0.0461004451,-0.0473696217,0.1415730417,-0.0500427261,0.2129478604,-0.0024274585,0.3216819167,0.048708979,-0.2102934867,0.1044703424,-0.3638079464,-0.4317641556,-0.2797992527,0.090648137,0.1936855912,-0.1161678135,-0.3184915483,0.6250885725,0.0562413037,-0.1004456729,0.1612388492,0.3503362238,0.0153473923,-0.0499008372,0.0480900519,-0.0607255809,-0.2662008107,0.0175641309,-0.3042603433,0.2425423265,0.1202826723,-0.2763026357,-0.0203578267,-0.0617138743,0.1419868618,-0.1604048908,0.0006388248,-0.186659798,-0.0840154663,-0.0231113005,-0.0585225001,-0.0471781753,0.0113554131,0.0236854777,-0.4993109703,-0.0146626448,0.0077614873,-0.1825955361,-0.0356433503,0.3575810492,-0.1342759579,0.0389887318,-0.1232834235,0.1663417816,0.0892147869,0.137729004,-0.0722378939,0.3133111,0.295452863,0.0509734638,0.084175393,0.2468766719,0.0595441759,-0.0747777969,-0.1448894143,0.1593147814,0.1643046737,0.1164806783,0.2938349545,-0.2498119026,-0.1497595906,0.0634250045,-0.0376486331,-0.2155546397,0.2923839092,0.0413802043,0.3077195585,0.2018053383,-0.1722927541,-0.0777101815,0.3362633884,-0.200277701,0.0266961176,0.4785175323,-0.276144892,-0.0927360579,-0.1404372156,-0.2291949242,0.3158210516,0.0973858014,0.0808683187,0.1155241802,0.1317080259,-0.0829708651,0.1752370894,0.330427289,-0.1765456051,0.1896955669,0.1255268455,0.1003875956,-0.1847053319,0.0316197723,-0.2409152985,0.3497812748,-0.3809463084,0.2891239226,0.0935084447,0.0197599977,-0.4174820483,-0.2146106511,-0.2094571888,0.0536932051,0.2030449659,-0.184506461,-0.1346839219,-0.0525902212,0.4098627567,0.2963763177,0.1396318376,0.2047356367,-0.0346682929,0.1242037714,-0.3263192177,-0.0076295468,0.1608230472,-0.1400020123,0.1128738225,0.0829424113,-0.2021440566,0.0533419177,-0.3232709765,0.179868117,-0.1953159571,0.2358195633,0.1865128875,0.2158547193,-0.3509802222,0.0346098617,-0.0180248544,-0.0835414454,-0.0908870921,0.0012711278,-0.141614601,0.1193026155,-0.0918076336,-0.3632541597,-0.080646202,-0.0925011858,0.3361850381,-0.0578352958,0.0124794133,0.1788610369,0.220102638,0.1198034286,-0.0008885673,-0.1502713114,0.0651718229,-0.0914852098,0.2097014189,0.0222972371,-0.2280718237,-0.3884096742,-0.049505569,0.0034498556,-0.1507952511,-0.5319606066,0.0428444967,-0.4780923426,-0.2615787983,-0.2581435144,0.4361437857,0.1785502136,-0.0556233749,0.0651727989,0.0948266014,0.085528791,-0.0653801411,-0.2383454442,0.1657000333,-0.4090308547,0.3104115129,-0.1278923601,0.8918733597,0.2818318009,-0.0301536024,-0.0232942551,-0.0228208452,0.3791375756,-0.4229274392,-0.1471819133,0.2681609094,-0.0813390836,-0.3361057639,0.3171370924,0.1355081797,0.4041448534,0.1233596578,0.0580145754,0.2295855731,-0.3638523519,0.4420793951,0.1802274436,0.1680265069,-0.0428529195,0.0800591931,0.0187664889,-0.3227300942,0.0402565375,-0.0541116744,0.2084042877,0.0055822614,0.143393293,-0.3114244342,-0.7348177433,-0.0694180205,0.1514322609,0.0041259793,0.129867956,-0.1483124644,0.1331444234,0.0591772906,0.7799109817,0.2442352772,-0.1808483601,-0.0171397831,-0.1189842373,-0.1900437921,0.0765746832,-0.3629523814,-0.1096825376,0.5771887898,0.6488320231,-0.3822695315,-0.059770029,0.0983514711,0.3369762003,-0.1064748541,-0.0028545617,-0.4241302311,-0.0482366979,-0.2051329613,0.2068946809,0.1253680736,0.0427046083,-0.1501823068,-0.0668411106,0.1060286835,-0.0546177104,0.2829441726,0.3270207942,0.291377455,-0.3086840808,0.3943769038,-0.0321291797,0.0407237187,0.3370806575,0.2501644194,-0.2270700186,-0.3491237462,0.2843481302,-0.0604727119,0.2991709113,0.5100072622,-0.0280362293,0.1072089449,0.1304300874,0.4783917964,-0.4216321707,0.4475492239,0.4779143035,-0.0455986559,-0.4816351831,-0.2087474465,0.405030638,0.0920074731,0.1492944807,0.1679892987,0.2624354064,-0.3083886206,0.3386710286,-0.0276184939,1.1405240297,-0.257012248,0.6073176861,0.1081267074,0.2599499524,0.6181569099,-0.2622877955,0.1768555194,-0.3545400202,-0.1569010913,0.1905536503,0.0545929819,-0.1465180516,0.096696943,-0.2047160417,0.2035128027,-0.0937369391,0.1888563633,-0.228587985,0.3655287325,0.1225678325,-0.4232102334,-0.2664298415,-0.0459932536,-0.0511844158,0.0639954507,-0.1379386038,0.0754685551,-0.0181395318,-0.2961391509,-0.2930018902,-0.153223902,-0.0735534653,-0.0920340642,0.2486364096,-0.2898269296,0.3593746126,0.0114060352,0.2215628326,0.6132361293,-0.265958786,0.0643401369,-0.1660613418,0.0261283088,-0.0231769197,-0.0207893979,0.0484747551,0.0487469584,0.021406902,-0.040928483,-0.0156557094,0.3132303655,-0.3416118026,-0.2126064301,0.4788935184,-0.4908950031,0.0345226526,-0.0967927426,-0.1015707627,-0.1302843243,0.0321828909,0.137683019,-0.2899424136,0.0958396196,0.0277614109,-0.3693692088,-0.2427587509,0.2325755805,0.2811689973,0.0055731782,0.6309413314,0.3198364377,-0.0506042205,-0.0940011889,0.3591878116,0.4548475444,-0.587146759,0.3231056333,0.1827057898,-0.0776099637,0.0416631848,0.1862029433,0.2471877486,-0.2686728835,-0.1792808771,0.0041799219,-0.4433617294,0.0834011361,0.0541787744,0.1633237302,0.1913104504,0.3379992247,0.0115272524,0.2028885782,-0.1476166695,0.0184988379,-0.0158751085,0.056517411,-0.1889320314,-0.0843075067,0.2142502367,-0.1100474373,-0.0787368491,0.2463212907,-0.0365482196,-0.100357905,-0.2585129142,0.1644328237,0.0116173802,-0.0932760015,0.0746606439,0.1224260181,-0.1858924627,-0.2925647497,0.4265333414,0.1071489155,0.1057100296,-0.1591506153,0.1500169933,-0.1114604175,-0.145363763,0.025575703,0.201366365,0.2704308331,0.0875747353,0.169248715,0.2338264138,-0.1096136048,-0.0159221068,0.2243997753,0.2848624587,0.1556613296,0.1020764261,-0.1707968861,-0.102998957,-0.0018230536,0.2042332888,0.469201833,-0.253243804,-0.2259342223,0.4052749276,0.0992898867,-0.0526457429,-0.0993525609,-0.1344956756,-0.1887051165,-0.1091618687,0.3316679299,-0.08643426,0.0454815067,-0.0495170653,0.152903378,-0.0607315749,0.1428847015,-0.0211196449,0.6312286854,-0.0057855872,0.0852711648,0.1575893909,0.0630870536,0.2630429268,0.6118055582,0.2405248284,0.1666042358,-0.4917708635,0.0864614919,-0.1838754714,-0.5297768116,-0.4431778789,0.1267947853,-0.1018827483,0.3069047928,-0.1597961783,0.5299904346,-0.0897259489,-0.2319508791,-0.1653169543,-0.0583481602,-0.2386588454,-0.1073422581,0.3041739464,-0.099670358,0.1822091043,0.2335484624,-0.115566887,0.0317357741,-0.0119836228,0.1258921027,-0.0749742016,0.1250470728,-0.2241402268,0.153832525,0.1388193667,-0.2075572461,0.2348943502,-0.0260949284,0.0195629504,-0.2605033815,0.3408118188,0.4472007155,0.3692032695,-0.2102328539,0.2285743356,0.1935970485,0.0688949153,-0.0467609763,0.1994188875,0.2860611975,0.1885344684,0.0434780493,-0.0506954789,-0.1486740559,-0.118226409,0.1432494819,0.078599669,-0.1758379936,0.2510385513,-0.1143715233,-0.2028599977,-0.2750482261,0.0261966866,-0.6956096888,0.5344954133,0.4925582707,-0.0480542444,0.101113759,-0.0474883728,-0.0031143881,-0.0562915504,0.5935714245,0.4852705896,0.0755331293,-0.3850583136,0.0492330492,-0.2965462804,0.2927154601,-0.3322428167,0.319203794,-0.0713101327,-0.1095998064,-0.014767834,-0.0869967192,-0.0745086074,-0.2496883571,-0.6370695829,0.1584707946,-0.1358510256,-0.2381409109,-0.2515378594,0.0269781109,-0.0475627109,-0.3367347121,0.3873045743,0.0477775633,-0.1069088727,-0.0323372968,0.2124532163,-0.1892139316,0.1731825024,0.1958434731,0.2702770531,0.0202771015,0.016923286,-0.1743377,-0.3229994774,0.1347472221,-0.0919105187,0.0575073026,0.2862854004,0.3036929965,-0.3394980729,-0.2640425861,-0.4243774116,0.0537348166,0.3286544681,-0.164001435,-0.1837852299,0.0568171367,-0.0863060132,-0.0345544405,0.058342956,0.5320749879,-0.1458185613,-0.0805742145,-0.3762450814,-0.21002841,0.4452861249,-0.3400377631,-0.438200444,-0.2218047827,0.0361188091,0.1998680234,-0.2338713259,-0.7568818331,0.1211131364,0.126661405,0.0193045605,-0.4797683656,0.1573762447,-0.1383927613,-0.1603676826,-0.1602217555,0.7219979763,0.0823606104,-0.1749937683,0.4088820815,-0.2712994218]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/861","title":"Possible Bug: Small training\/dataset file creates gigantic output","comments":"Oh yes right..\r\nDo you think that a lazy map feature on the `datasets` side could help to avoid storing padded tokenized texts then ?","body":"Hey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n","comment_length":25,"text":"Possible Bug: Small training\/dataset file creates gigantic output\nHey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n\nOh yes right..\r\nDo you think that a lazy map feature on the `datasets` side could help to avoid storing padded tokenized texts then ?","embeddings":[-0.1584593505,-0.322750479,0.1142651811,0.2783169448,0.6050810218,-0.0893944949,0.3665043712,0.3757137358,-0.2504401803,0.1644323915,-0.0610655993,-0.0809407011,-0.3737191558,0.1472707838,0.1473814845,0.0324401334,0.1391411573,0.1354866177,-0.0773729756,-0.2471373975,0.0194773749,-0.0309278648,0.2424664944,0.1948526949,-0.5476983786,0.0997936502,-0.0319239534,-0.0307262652,-0.4810985029,-0.1251824498,0.1005678922,-0.2483758777,0.1287238598,0.424087584,-0.0001295393,-0.0181988981,0.1412843764,-0.0814446658,0.0510074385,0.1926465183,0.0713473484,-0.0778828338,-0.1597842276,0.0811405629,0.2066449672,-0.1023944914,0.0354038849,-0.079418689,0.3247146308,0.1447516978,0.063819319,0.1143500283,-0.0348902978,0.0747108832,0.3187438548,0.2770513296,-0.0755639598,0.0055221049,0.1870386451,-0.2282137126,-0.0746483877,0.235629946,0.0775171071,-0.2015115768,0.3477509022,0.2555222213,-0.0118268114,-0.3913118243,0.2150571942,0.3343567848,-0.0059385668,-0.369189173,-0.281550616,-0.4759232402,0.1507344693,-0.3983701468,0.0676338822,0.1264641881,-0.0999671966,0.1164510176,-0.3256670237,-0.4626322091,-0.2151950002,-0.0421484858,-0.1925977319,0.0499017164,-0.2066259682,-0.1162705347,-0.0609063059,-0.1237846464,-0.2689934075,-0.0728556141,0.0181449167,0.2470308542,0.0355407819,-0.3357768357,-0.0486233234,-0.400313288,0.3179071546,-0.5198046565,-0.1670588255,-0.3840146661,-0.2822938859,-0.3092981875,0.4190124869,0.4008267522,-0.3786330521,0.2864236236,0.147435993,0.0143901473,-0.0981879681,-0.0872362033,-0.1584558487,0.1529669315,0.0889035314,-0.366333425,0.0794737265,-0.2428524643,-0.11945007,0.4220166206,-0.3422700167,0.2382763326,0.0211726893,0.3271232843,-0.1095505804,0.0815232024,-0.0186989699,0.0428276025,-0.3623221517,0.069616802,-0.0575441085,-0.0081365826,-0.1674923748,0.1830439121,-0.0043521412,0.0869902298,0.0923545286,-0.0637386516,0.2453142405,-0.0269418415,0.074988164,-0.2383559346,0.2199847698,0.1590303481,-0.2199200541,0.4460888505,-0.0356886499,0.0270396918,-0.06633766,0.1065040454,-0.2845102549,-0.5047079921,-0.0448392741,0.0848665237,0.01339396,0.2874855995,-0.1068297699,0.3798874617,0.4936864674,-0.0647113547,0.0465578251,-0.1642074883,-0.3196155131,-0.0120980768,0.3641358614,0.3811829984,-0.4837587774,-0.1946118027,-0.1694867313,0.2451892048,0.2841442823,0.6231352687,-0.0581989922,0.5801355243,-0.2897653282,0.2861112654,-0.260382086,0.2289155871,-0.4580262303,0.3011760712,-0.0231810678,-0.0101977661,0.0377300903,-0.1035020128,0.2780748606,-0.0187087432,0.4887704849,-0.0405674987,-0.1937264651,0.048390232,-0.3471644223,-0.3909625411,-0.2233014405,-0.0301923789,0.1755296439,-0.1729311347,-0.2678878307,0.5790894628,0.1735107601,-0.0645666271,0.2249490172,0.3249577284,0.1268682331,-0.0376047418,0.0377345607,-0.1063976064,-0.1840966493,0.000745947,-0.2624854147,0.1973665208,0.0605289787,-0.3366334438,-0.0279029291,-0.0303604268,0.0522277951,-0.1490053236,-0.0096160518,-0.2432951331,-0.1688456088,0.1220776215,-0.0486761443,-0.1587258875,-0.0792643651,0.0563898794,-0.3761591017,-0.0059617641,0.0287979059,-0.1611763984,-0.0967428312,0.3295545876,-0.1723220646,-0.0049660024,-0.0797172859,0.1359841675,0.1103423238,0.1088261977,-0.1641490757,0.3137359917,0.3106914163,0.0941727832,0.0482255332,0.2186592072,0.0120981317,-0.0983668417,-0.1908928603,0.1803031266,0.208724305,0.0652376935,0.1750696898,-0.2777831554,-0.0779210106,0.1187643185,-0.0812474042,-0.3290862441,0.2585925758,0.047444284,0.2946568429,0.1983209252,-0.1492181271,-0.075496234,0.3165712655,-0.1617223918,0.0371678807,0.3789635301,-0.3250766397,-0.0943503231,-0.0045814626,-0.3105895519,0.2905078232,0.0462883227,0.1345673352,0.0940559953,0.1150725111,-0.0617256872,0.1856593937,0.2106611729,-0.1262027025,0.2095306963,0.192082569,-0.0009269142,-0.2589702904,0.1365609318,-0.1572662592,0.2898987234,-0.3594797254,0.2282149941,0.0755934566,0.0265272297,-0.3885186911,-0.2284081131,-0.1528207213,0.0331248231,0.2542636096,-0.1518006921,-0.151860863,-0.0158998445,0.3029646575,0.3114877939,0.2414972484,0.2074394822,0.0153080598,0.1233328879,-0.2894179225,-0.031588655,0.1572476476,-0.2357091755,0.1222896278,0.1127193421,-0.2993734181,-0.0817451477,-0.2953290343,0.2014806569,-0.1648557335,0.196104303,0.2442874312,0.125231877,-0.3874123991,0.1048204973,0.0128359301,0.0034779874,-0.1348334402,-0.1136595905,-0.0770547688,0.1489091218,-0.080621928,-0.2752304077,-0.0724572614,-0.0015945281,0.3967808187,-0.0382245816,0.0059397155,0.198438257,0.2022285312,0.0955072492,-0.0675063059,-0.1960124373,-0.0069532352,-0.2221462876,0.2693296671,-0.0200770982,-0.1803517789,-0.3509810567,-0.0367721766,-0.0391584337,-0.1439628601,-0.5498438478,0.1042803302,-0.3475675583,-0.2124298364,-0.1869568676,0.3236688375,0.162125513,-0.0278233867,0.0954016596,0.0770545453,0.0741135105,-0.046077922,-0.153002888,0.2259287834,-0.367066294,0.3084632754,-0.0015540273,0.8413866758,0.258456707,0.0591780581,0.0175117403,-0.0364140868,0.4315395355,-0.4216063023,-0.1953714192,0.2533493936,-0.1266723126,-0.3854464293,0.3003961742,0.1515319943,0.3678123951,0.2488987595,0.0220231246,0.3477158248,-0.3686490357,0.4326383173,0.0086077927,0.1201054305,0.0394562408,0.1065802574,-0.0276057404,-0.3450220823,0.1210093051,-0.1188098565,0.2471742183,-0.0138743473,0.0747002959,-0.2114319801,-0.8055394888,-0.0793854967,0.1710191518,0.0181142725,0.1024311483,-0.0702355653,0.1438985914,0.1094579697,0.7522368431,0.1485324502,-0.1466356516,-0.0731038302,-0.0320918858,-0.1479646713,-0.00733164,-0.317501545,-0.0080263969,0.554476738,0.805159688,-0.3837490082,-0.0145990252,0.0127897821,0.2712612748,-0.1630632132,0.0465953313,-0.3782891333,0.033482857,-0.1805645823,0.1646660864,0.1506462246,0.0484342538,-0.0364954062,-0.1383151561,0.0361431167,-0.0462700054,0.3272375166,0.1801160872,0.30792135,-0.269018352,0.3792641759,0.022727998,0.1455383748,0.3632508218,0.2782917321,-0.2656795382,-0.3319880664,0.2026930302,-0.1705929339,0.3661507666,0.5202790499,-0.0199191105,0.1646699905,0.1409870386,0.3879474699,-0.5066109896,0.5397726893,0.5001416206,-0.0155579699,-0.5173014998,-0.260522306,0.3242104948,0.089011021,0.1086992025,0.2183576822,0.1754108518,-0.4198832512,0.3800750077,0.058238782,1.0927495956,-0.2820734978,0.6457759142,0.0856931806,0.347442627,0.6195512414,-0.3919829726,0.1663229018,-0.3371164501,-0.1650778204,0.1239652634,0.087960355,-0.1708421856,0.0517651923,-0.1440128982,0.1905903965,-0.0315781087,0.3619603515,-0.2650906444,0.2576601207,0.1245743334,-0.3357917666,-0.2187579572,-0.0675895661,0.0815504119,0.0012560491,-0.1550394595,0.0662308857,0.0139890332,-0.3324243724,-0.2732451856,-0.1858699024,-0.1023099348,0.0334131271,0.2606478631,-0.2193477154,0.2665244639,0.0149169788,0.1957245767,0.6118080616,-0.2747984529,-0.0073970631,-0.141610384,0.0322249718,-0.0554138832,-0.1050668806,0.0278455094,0.1215406358,0.1015194803,0.0101793166,-0.051777482,0.2496397793,-0.4359840155,-0.1743823439,0.5185425282,-0.5724605322,-0.0609520152,-0.1661489904,-0.0553824157,-0.1067349166,0.003389868,0.1229849458,-0.2376911342,0.0719602704,0.0481570922,-0.3586072028,-0.1936806291,0.308465898,0.1919232011,0.0500711352,0.6962034702,0.1818852723,-0.0695303008,-0.046462696,0.4055944383,0.5893694162,-0.3776433766,0.2560918629,0.1496012211,-0.0949784815,-0.1004093736,0.1920970976,0.2429114282,-0.259929657,-0.1791065931,-0.0494264811,-0.4665798545,0.0721068233,0.0032246276,0.1664883494,0.1396561116,0.3237324059,-0.066635415,0.1534444541,-0.1248238087,0.0931608751,-0.0238917228,0.1016378179,-0.2898565531,0.0022013672,0.1709153205,-0.1663370579,-0.0985280052,0.2696664631,0.0147397835,-0.0643740222,-0.2574064434,0.1716156006,0.0513091125,-0.0414519757,0.1521832049,0.0773045868,-0.1261708587,-0.3096643984,0.3914384246,0.1916747093,0.140350759,-0.2274481505,0.1455268115,-0.0863011852,-0.1304282993,0.0412223153,0.1899661273,0.3233479261,-0.0304895602,0.1629435718,0.1836392879,-0.076837115,0.0709079728,0.2680519819,0.286149472,0.0453732386,0.1284097135,-0.2139104754,-0.0652390346,0.0621737018,0.2593594193,0.4585362375,-0.3202498853,-0.2236282825,0.4694548547,0.0802864656,-0.0413801596,-0.1314872652,-0.0876242071,-0.1502095014,-0.12259835,0.2824838161,-0.0133888079,-0.0183291472,-0.1305375695,0.1619996279,-0.1548793316,0.1399544924,-0.0480199978,0.56574893,0.004432545,0.0701994598,0.2165225148,0.0875182077,0.2833935618,0.6908547282,0.2469200641,0.1544189751,-0.3074209094,0.0846028775,-0.2320080549,-0.4808971882,-0.3943101764,0.2658729553,-0.119941406,0.3273163438,-0.1321462095,0.5771371722,-0.048140619,-0.2353112549,-0.1017271727,0.0392377526,-0.323209852,-0.1748595387,0.2453921884,-0.1198596805,0.1695698798,0.2694952488,-0.1514987499,0.0249420777,0.0607271083,0.0578390546,-0.0696149245,0.0819531679,-0.151566416,0.14337717,0.1724611968,-0.2777451575,0.2601990998,-0.184916079,0.0236287471,-0.2479348332,0.3573226929,0.4614425302,0.4569038749,-0.2203868479,0.1961307079,0.1398974806,0.0503160208,-0.0649370998,0.2028531283,0.3066430688,0.0474392511,0.0513346829,-0.0806986317,-0.1091971919,-0.0626922026,0.232647568,0.0353197381,-0.074703671,0.1730511338,-0.1069738418,-0.1755283773,-0.1765588522,-0.02387283,-0.6561957002,0.4798920751,0.5330019593,-0.0657745451,0.2312941849,-0.0600782372,-0.0091700898,-0.1038944572,0.4950033724,0.4832202196,-0.0837986469,-0.4275384545,0.0257522389,-0.2784149945,0.3173959851,-0.3863589168,0.2983151674,-0.1307020783,-0.0844775587,0.0218279697,-0.1150521636,-0.1574872881,-0.2555812001,-0.653503716,0.1622949094,-0.1024790332,-0.2134438008,-0.2421798259,0.0815437138,-0.0284336191,-0.3112556636,0.4596813917,-0.0351956524,-0.1514107138,0.0133029828,0.3289444447,-0.2615923584,0.1339915991,0.2557195723,0.2501488626,0.0000682016,0.0081561683,-0.1990966946,-0.2667861581,0.0887241364,-0.0846682414,0.127519086,0.2306878716,0.2665814459,-0.3857631385,-0.2505602837,-0.4074324667,0.1125191227,0.3654773831,-0.1307971179,-0.2475523949,0.1453557611,-0.1500486732,-0.0236987844,-0.0205412321,0.4520379007,-0.1429158449,-0.0449985787,-0.2473853379,-0.2518764436,0.4671111405,-0.3677014709,-0.406357497,-0.2025919706,0.0299033783,0.385264039,-0.2789250612,-0.8607033491,0.1042303592,0.2206311971,0.0156902727,-0.5150167942,0.0325448923,-0.0204755515,-0.1085937619,-0.2627114952,0.8137919903,0.0751594827,-0.1930020899,0.432377547,-0.3210276961]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/861","title":"Possible Bug: Small training\/dataset file creates gigantic output","comments":"I think I can do the tweak mentioned above with the data collator as short fix (but fully focused on v4 right now so that will be for later this week, beginning of next week :-) ).\r\nIf it doesn't hurt performance to tokenize on the fly, that would clearly be the long-term solution however!","body":"Hey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n","comment_length":55,"text":"Possible Bug: Small training\/dataset file creates gigantic output\nHey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n\nI think I can do the tweak mentioned above with the data collator as short fix (but fully focused on v4 right now so that will be for later this week, beginning of next week :-) ).\r\nIf it doesn't hurt performance to tokenize on the fly, that would clearly be the long-term solution however!","embeddings":[-0.2022947818,-0.2920219898,0.1189169437,0.2300961465,0.6037492156,-0.1722533703,0.315689683,0.3898852468,-0.2658645213,0.18251656,-0.0329228267,-0.147778973,-0.3427737355,0.2524911165,0.1107931584,-0.0713135824,0.1064804345,0.1568867862,-0.0434261188,-0.2282034308,0.0541994162,0.0182766169,0.2426915318,0.2236170024,-0.5773249865,0.093116641,-0.0897391662,0.0289621837,-0.4984022975,-0.1051620916,0.0200684927,-0.2774215639,0.2171355188,0.488702625,-0.0001255141,-0.0723906457,0.1390117407,-0.0683199093,0.0042039501,0.1563264132,0.203809917,-0.0963017419,-0.1436719298,0.1250668019,0.2383717299,-0.1081832573,-0.0434214957,-0.0000365168,0.3469856083,0.169578895,0.0850071013,0.1917706728,0.0978405029,0.0513320677,0.2168005258,0.2441504598,-0.0491969772,0.0492714457,0.2134055197,-0.2436895669,-0.0475986265,0.1766494215,0.1195391566,-0.1437061578,0.3066144884,0.192255035,-0.0794912949,-0.411546886,0.1700009108,0.4091199338,-0.0398110822,-0.3814209998,-0.2852917016,-0.3907469213,0.2280966341,-0.5098490119,0.0361433998,0.1651209593,-0.1153219119,0.0947942585,-0.2568562329,-0.391516,-0.2216070294,-0.0695686936,-0.2767575085,0.0675148293,-0.2188895941,-0.0988881961,-0.0032573708,-0.0955056176,-0.3700415492,0.0121714342,-0.0498173758,0.1920453161,-0.1057280302,-0.3005862236,-0.093510747,-0.3335511088,0.2913614511,-0.4924598038,-0.2083922178,-0.4083329141,-0.1743303239,-0.3648548722,0.4070279896,0.4913434088,-0.4073669612,0.1588030905,0.1524096578,0.036944177,-0.1634155959,-0.0812346712,-0.1595999002,0.1207202673,0.016142847,-0.3720576167,0.0639492795,-0.2456080317,-0.1266844869,0.3839028776,-0.3510911465,0.2377408147,0.0550510064,0.3134867251,-0.0650647357,0.1613825709,0.0347572714,0.041105397,-0.4139357805,-0.0494072996,-0.0716322884,0.0595841482,-0.1605283618,0.1914116293,-0.0187105779,0.100606747,0.0448962785,-0.0284892917,0.3421201706,-0.0739843398,0.0721718669,-0.255746454,0.1020974442,0.1747199595,-0.1739896983,0.4763528407,0.0112040155,0.0109734619,-0.0122268423,0.0318185911,-0.1611828953,-0.5222705007,-0.0953116566,0.1208966076,-0.0136681022,0.2884064913,-0.0481704585,0.3760944009,0.395816952,0.0022844994,-0.0270233732,-0.2176161557,-0.1882951558,-0.0019311025,0.4034152627,0.3783380091,-0.5633426905,-0.1548845172,-0.2199615538,0.284835279,0.3699565232,0.6364403963,-0.0472258776,0.5527011156,-0.2490289807,0.2100589722,-0.3150477707,0.2535911202,-0.4138020277,0.3480509222,-0.1281874627,-0.072096169,0.0379310586,-0.0920354947,0.247273311,0.0253312513,0.4554362297,0.0424429737,-0.1859478354,0.0485504866,-0.3573648632,-0.3754622042,-0.2417564243,0.0276660081,0.1375680268,-0.1434081644,-0.2986965477,0.5591665506,0.1224410012,-0.1235059127,0.1712397784,0.3022281826,0.1504045725,-0.0457573496,0.0347059667,-0.0240929928,-0.1081381887,-0.0064385473,-0.2265459001,0.2554838955,0.0755123645,-0.3105176985,-0.0878991336,-0.0096517168,0.10226053,-0.1911879331,0.0169620737,-0.221573621,-0.1310765445,0.0813742056,-0.0826464519,-0.1217534319,-0.1119900718,0.0957406461,-0.4077610672,-0.0668910071,-0.0375665426,-0.1556822062,-0.0447244309,0.3314082921,-0.1087693125,0.022935316,-0.0908818096,0.2122023255,0.0596889518,0.0822513476,-0.0839946866,0.232918784,0.2856232226,0.1345306933,0.1035177633,0.1799343228,0.0648086518,-0.0836119652,-0.0658560619,0.2375005335,0.231471926,0.112771228,0.2546750605,-0.2077405453,-0.0510378331,0.1223885641,-0.0303190965,-0.2649427354,0.2836706042,0.0495792888,0.2589974105,0.1513907462,-0.190586105,-0.0424954854,0.2515276074,-0.1678776145,0.0247898456,0.3735732138,-0.3690002561,-0.0784586444,-0.0579495206,-0.2319737077,0.2978201509,0.0852164626,0.1362763196,0.120152764,0.0970116854,-0.0916482285,0.1554721147,0.282751292,-0.1316686273,0.1620893776,0.2225505114,0.0028682395,-0.2590218186,0.0297040921,-0.2065397054,0.3683235943,-0.367923528,0.29502666,0.0566736832,0.0482108817,-0.3758403361,-0.276639998,-0.2192476839,0.0057147471,0.2062768638,-0.1668457985,-0.1179317534,-0.0931925625,0.3834417164,0.263286829,0.1543856561,0.3253230453,0.0347320177,0.0900794715,-0.3657518327,-0.0053994451,0.1525135487,-0.1867826581,0.1173915789,0.1443026811,-0.2773666978,-0.0515350252,-0.3002146184,0.1842243075,-0.1805002689,0.2566883266,0.2256781608,0.1575404704,-0.4216123223,0.0344796591,0.0324531607,-0.0248775072,-0.1283336878,-0.0536395386,-0.1409126818,0.1573622078,-0.0745352432,-0.2841409445,-0.1479458958,-0.0517506041,0.361946106,-0.0264673084,-0.0296914093,0.1829232126,0.1623683572,0.0922787413,0.0150932139,-0.2145968676,0.0251060873,-0.1410413831,0.1920464486,0.013445612,-0.1889305711,-0.3331333995,-0.0168802962,0.0008043608,-0.1568006575,-0.5277712941,0.0539353974,-0.434366107,-0.2976064682,-0.2196919769,0.34762308,0.0865405574,-0.0860693753,0.0464070737,0.1269993782,0.1192756593,-0.0778975412,-0.2258176357,0.2165210247,-0.4335952401,0.4292582273,-0.0345524959,0.8483142257,0.2267303169,-0.0550652854,0.0104465755,-0.0209994148,0.3937900364,-0.4736229479,-0.2124661505,0.3288615942,-0.1232093722,-0.380184412,0.2800569236,0.122731097,0.4507929683,0.1759644002,0.0481790453,0.2780033648,-0.3748352826,0.3768453598,0.1353834122,0.1944564432,0.0062738084,0.0723600611,0.0427688733,-0.3326547146,0.073512882,-0.0231747907,0.1662969291,-0.0096783433,0.0337760895,-0.2803366184,-0.7721146345,-0.0122382967,0.0897061676,0.0998128355,0.1731178612,-0.1408755332,0.164035514,0.0309019219,0.8085677624,0.3027648926,-0.1947133988,-0.0275229234,-0.0687405393,-0.1778626889,0.0168134961,-0.3545091152,-0.035879638,0.5801979899,0.6342161298,-0.3662616313,-0.0494977087,0.1119946465,0.3816509843,-0.1101432815,0.0391918384,-0.3996424079,-0.0195072219,-0.2422991842,0.2102530897,0.0761092976,0.0797600895,-0.1154864654,-0.0831390843,0.1009412855,-0.1210438684,0.3523439467,0.2720105648,0.3041644096,-0.2905327678,0.3746730685,-0.04268758,0.1138568893,0.3892744482,0.1802374423,-0.1690205038,-0.3310913444,0.2556257844,-0.0677469522,0.3350260854,0.5346246362,-0.0161222164,0.2118690312,0.0786076114,0.4518143833,-0.4136651754,0.5587219596,0.4952727258,-0.0402752087,-0.5326898098,-0.1620940566,0.3384355903,0.073851794,0.1286016554,0.1966747642,0.2628086805,-0.3021442592,0.3217558265,-0.0430060737,1.1316547394,-0.2781016529,0.5939578414,0.0634623468,0.2603398263,0.6253473163,-0.3204675913,0.1985353231,-0.3777290583,-0.111315079,0.161072433,0.084492892,-0.1474875212,0.0647373646,-0.1699178517,0.1965812743,-0.1071451157,0.2627726793,-0.2470771223,0.3641862571,0.035758812,-0.3776301146,-0.2614325285,-0.0329858474,0.0006420949,0.0289849024,-0.1743228287,0.0873858407,0.0621008947,-0.2995586097,-0.2748017609,-0.1961102933,-0.082653828,-0.0117116151,0.2052902579,-0.3377436996,0.2742099762,-0.0674555674,0.2621939778,0.6511217952,-0.2098906785,0.0959377289,-0.1948786229,0.0624819137,-0.0203914437,-0.0850728229,0.0024808072,0.0770368278,0.0503250323,0.0032787607,-0.0120321997,0.2906377614,-0.3991993666,-0.1737021506,0.5477033257,-0.4731478095,0.0558952391,-0.0999977812,-0.0465255417,-0.1736382842,0.0261042919,0.1446434557,-0.2854093313,0.0365473703,0.0392333902,-0.3762051165,-0.2354313582,0.2242616117,0.2178137749,0.0336708277,0.6945822239,0.3077147007,-0.0296190474,-0.0777299032,0.3860263526,0.5473822355,-0.433883965,0.3099583387,0.1825859845,-0.0441234,0.0502261966,0.2061123699,0.2624817789,-0.3000706732,-0.1796600521,-0.0724706575,-0.4185904562,0.064185597,-0.0437605567,0.1417154372,0.1978532225,0.4324569106,-0.1058502048,0.1831015348,-0.1512870044,0.0984663218,-0.0487737879,0.0988991782,-0.2720241249,-0.0635882467,0.157898128,-0.1114619598,-0.0566094853,0.255033195,0.0044631022,-0.1042548716,-0.2301461548,0.1626790017,0.0316880606,-0.0372919179,0.1404599696,0.1088925526,-0.167083025,-0.2456863225,0.3883174956,0.144298926,0.1313142478,-0.2201205939,0.1369384527,-0.0714999363,-0.0951592922,-0.0318625048,0.2092144191,0.2916563749,0.1146193221,0.1534420252,0.2614181936,-0.0702032521,-0.012366809,0.2207598388,0.2243562043,0.0791246369,0.0622730032,-0.1821434051,-0.1318075061,-0.0142753571,0.2707621157,0.4881924093,-0.2587164938,-0.2257805914,0.5012185574,0.1117135212,-0.0246025994,-0.0911834538,-0.1167160273,-0.2040233612,-0.1240226775,0.3177524507,-0.0207639206,0.0352480635,-0.1154177412,0.1799301058,-0.0869194046,0.1717033386,-0.0511527583,0.5434497595,-0.0404454581,0.0851744339,0.1827305853,0.0224991329,0.2841389179,0.6910934448,0.180780828,0.1545212716,-0.3504056036,0.1210544333,-0.1919400841,-0.5043106079,-0.4453705847,0.1366318315,-0.088642478,0.2801234722,-0.1445302963,0.5805773139,-0.1093649641,-0.270963639,-0.1186148673,0.0088290833,-0.2598236203,-0.1174584329,0.1925292015,-0.1576167792,0.1352836937,0.2644309998,-0.0990419239,0.0462278314,-0.0953775421,0.0562901944,-0.0968344286,-0.0287598558,-0.2080896348,0.178024292,0.125055477,-0.2428969145,0.2051354796,-0.071095936,0.0525449328,-0.2004516572,0.3062068522,0.4847804904,0.4154979885,-0.2384962291,0.1954756826,0.1275564134,0.0669913441,-0.0642051697,0.1136269271,0.3247196674,0.1053210273,0.0340534449,-0.0433576293,-0.1434876472,-0.1536737978,0.2062390745,0.0756683946,-0.1636309177,0.197009474,-0.1018453836,-0.1498575658,-0.2699945271,0.0260947924,-0.7529441118,0.5340988636,0.4614876211,-0.0610114112,0.1280629784,-0.0869826898,0.0034485755,-0.1110579446,0.4745289683,0.4845038354,-0.0024140298,-0.4102847874,-0.0407221541,-0.3033803701,0.2930364311,-0.3688573837,0.2744470239,-0.1734830588,-0.082260184,0.0568435714,-0.1173155084,-0.0912353769,-0.1900202334,-0.703112781,0.1316700727,-0.1559220254,-0.1953705847,-0.2384415269,0.0313235223,-0.0565427542,-0.2429548353,0.438921541,0.0333598629,-0.1227197424,-0.0055753202,0.2359008342,-0.2267874181,0.10968557,0.2416948676,0.281845659,0.0293617509,0.055632595,-0.198058784,-0.2926268876,0.0976225138,-0.116041027,0.1269226074,0.2477734387,0.2843007147,-0.2559836209,-0.2809932232,-0.4314659834,0.1640501767,0.3631734252,-0.1304235011,-0.2610964477,0.1225452274,-0.1037977412,-0.015727669,0.0005375079,0.5165205598,-0.1023203209,-0.0567015037,-0.3643942177,-0.2662373781,0.4903406203,-0.3295814097,-0.4211165309,-0.2379918396,0.0514787547,0.3020356297,-0.2552241087,-0.7636173964,0.0847047642,0.1540557295,0.0148217976,-0.5005554557,0.1322122514,-0.0590818375,-0.1500605941,-0.205328092,0.6773766279,0.0786697045,-0.2040115595,0.3991594911,-0.2519074976]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/861","title":"Possible Bug: Small training\/dataset file creates gigantic output","comments":"> Hey guys,\r\n> \r\n> I was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n> \r\n> I've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n> \r\n> I've used the following CMD:\r\n> `python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\nIt's actually because of the parameter 'preprocessing_num_worker' when using TPU. \r\nI am also planning to have my model trained on the google TPU with a 11gb text corpus. With x8 cores enabled, each TPU core has its own dataset. When not using distributed training, the preprocessed file is about 77gb. On the opposite, if enable xla, the file produced will easily consume all my free space(more than 220gb, I think it will be, in the end, around 600gb ). \r\nSo I think that's maybe where the problem came from. \r\n\r\nIs there any possibility that all of the cores share the same preprocess dataset?\r\n\r\n@sgugger @RammMaschine ","body":"Hey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n","comment_length":273,"text":"Possible Bug: Small training\/dataset file creates gigantic output\nHey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n\n> Hey guys,\r\n> \r\n> I was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n> \r\n> I've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n> \r\n> I've used the following CMD:\r\n> `python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\nIt's actually because of the parameter 'preprocessing_num_worker' when using TPU. \r\nI am also planning to have my model trained on the google TPU with a 11gb text corpus. With x8 cores enabled, each TPU core has its own dataset. When not using distributed training, the preprocessed file is about 77gb. On the opposite, if enable xla, the file produced will easily consume all my free space(more than 220gb, I think it will be, in the end, around 600gb ). \r\nSo I think that's maybe where the problem came from. \r\n\r\nIs there any possibility that all of the cores share the same preprocess dataset?\r\n\r\n@sgugger @RammMaschine ","embeddings":[-0.1276512891,-0.4218369722,0.1068674028,0.2943581343,0.6427578926,-0.1283721924,0.3398812711,0.4012147486,-0.2470448911,0.1549628377,-0.0381345004,-0.1218298003,-0.298771292,0.2704769373,0.1444567144,-0.0848587081,0.0800594166,0.156734705,-0.0432324745,-0.3624180257,0.0178007986,0.0119267926,0.2649854124,0.2487947494,-0.5399374366,0.116436705,-0.0890409946,0.0878877193,-0.4778915346,-0.0461199917,0.0388478674,-0.3701862097,0.1601385921,0.5560237169,-0.0001261977,-0.0579311587,0.1979992837,-0.1043423191,0.0430375598,0.1906141341,0.0401332714,0.0164479893,-0.1522930861,0.1174175739,0.1633697003,-0.074980475,-0.0505147576,-0.0368019789,0.3213077784,0.2368453145,0.088009201,0.0191166382,0.0510077886,0.0527154952,0.2443502098,0.2661501467,-0.0326643512,0.0410114676,0.1123430356,-0.0585210361,0.0381150506,0.1659110487,0.1789548993,-0.0714272782,0.3485538661,0.2815019786,-0.062659204,-0.4381529391,0.1757108867,0.4649783075,-0.0341458023,-0.3490568697,-0.3314692676,-0.3132906258,0.2559036911,-0.5074098706,0.0019555108,0.3379586935,-0.0914493799,0.0786473528,-0.2632108033,-0.4374252856,-0.2496211976,-0.1110474318,-0.2281236947,-0.0159756113,-0.2338900566,-0.0931955799,-0.0396680608,-0.0589147583,-0.4165319204,-0.0705715269,-0.0816852972,0.2201217264,-0.031972155,-0.2654673457,-0.1090493724,-0.2294502705,0.2409886718,-0.5617985725,-0.0786131322,-0.4566052556,-0.2560954094,-0.3459122479,0.3546568453,0.4483872056,-0.4579661191,0.1858590692,0.1445351243,0.0826460943,-0.2117664665,-0.06054103,-0.1389766037,0.1224850789,0.0102628861,-0.4655169547,0.0591996051,-0.2968020737,-0.1397833079,0.3230869174,-0.3538083136,0.202371493,0.0499953069,0.3438809514,-0.1023268476,0.041925963,0.0515126586,0.0887674317,-0.4091444612,-0.0151635017,-0.1164165661,0.1181161553,-0.1676016748,0.2057823837,-0.057826519,-0.0049614981,0.0498314612,0.0084664635,0.2533488572,-0.0428125262,0.0828226656,-0.2464299947,0.2176022828,0.1504771411,-0.1683322638,0.4425061643,-0.0380499996,0.122501269,-0.001926252,0.1240713298,-0.206124261,-0.4667069316,-0.0831879154,0.1185633466,-0.0553817675,0.2808890939,-0.0317334644,0.3154496253,0.450430721,-0.0002377933,0.0170981605,-0.1776507199,-0.235056445,-0.0263187177,0.3950234056,0.3266683817,-0.5285282135,-0.1483890265,-0.2407324165,0.277894944,0.4014596641,0.6155525446,-0.0071397009,0.5933489203,-0.2661556602,0.1274461448,-0.3081243634,0.2741969228,-0.37720415,0.3744081259,-0.0902812034,-0.0957927927,0.0843489021,-0.0470152386,0.0716257989,0.0765350312,0.394489944,0.0010770552,-0.1426581889,-0.0059017921,-0.3934575319,-0.37250337,-0.3202274442,-0.0293919351,0.1224965379,-0.0589129403,-0.2856299579,0.5653201938,0.1582309604,-0.1308507472,0.1914218366,0.2992986441,0.1277524978,-0.089828141,-0.0050906278,0.0295594726,-0.0824714825,0.0331872217,-0.4174715579,0.1905743778,0.1240222678,-0.253105998,-0.0447313562,-0.0687903091,0.0282635409,-0.2317643464,-0.0192760415,-0.154654637,-0.0446011312,0.0154301254,-0.02482762,-0.1252895743,-0.0736534745,0.0106474645,-0.5096874833,-0.0241763014,0.0002805335,-0.1650916338,-0.039744731,0.3676860034,-0.1405473799,0.0562488437,-0.0805534273,0.1306506544,0.1284166425,0.0748591498,-0.0634482503,0.1455746442,0.3039922118,0.1224542037,0.113545835,0.2218716592,0.1184036955,-0.0786923841,-0.1911046505,0.2072176337,0.2488701791,0.0800248906,0.2820250094,-0.2214490324,-0.089483358,0.0987152383,-0.0532893464,-0.2998576164,0.2517929673,0.1236835271,0.2842864096,0.1805296391,-0.1856545955,-0.1370201707,0.323943764,-0.1951678842,-0.0396147296,0.4245702326,-0.3088125885,-0.0758593604,-0.0875334069,-0.2740561068,0.3153855503,0.1101151928,0.1338965595,0.1187676415,0.1388584077,-0.0974701717,0.1322098076,0.2498320192,-0.1262380779,0.1747721285,0.2392382026,-0.0798345506,-0.1743391454,0.1291180253,-0.1983764023,0.3336581886,-0.4030173719,0.3130139112,0.1204990372,0.112706393,-0.4343594015,-0.2917586863,-0.1530718505,0.0325137526,0.1591103077,-0.2382534146,-0.0750620961,-0.1352491528,0.3465173542,0.1934902072,0.1789177805,0.264380008,0.0568552725,0.1604308635,-0.3675194085,0.0146151548,0.1229770482,-0.1204927117,0.1066781431,0.0952569544,-0.2957550287,-0.0349997655,-0.2590642869,0.1933892071,-0.2113824338,0.2909787893,0.1516253352,0.1146207973,-0.368178606,0.0680634752,0.0882991701,-0.0322857238,-0.1498581469,-0.0441937521,-0.1486745775,0.228110075,-0.0874723494,-0.3000989854,-0.0941368267,-0.0312904865,0.4148477018,0.0837838948,-0.0100268302,0.2177245766,0.1786605418,0.1272266209,0.0197736006,-0.1597074717,-0.0354675986,-0.1911401302,0.1912483722,-0.0144756231,-0.2573117614,-0.3347419202,-0.1144602671,0.1040080488,-0.2010481656,-0.5380892158,0.0351849049,-0.4793241024,-0.2589476705,-0.216630429,0.3814693391,0.0344931707,-0.1614193022,0.0466761738,0.0747114271,0.1408394873,-0.0054685799,-0.2377617657,0.2027298063,-0.4108911455,0.3287277818,-0.1822104007,0.8339398503,0.3465123177,-0.0555108003,-0.0644268394,-0.0809496045,0.402064383,-0.4588729143,-0.1653172374,0.306430608,-0.1217416003,-0.3759494722,0.2950263917,0.1579595357,0.4908567965,0.1650063097,0.1103336066,0.2498965263,-0.3243836462,0.3450452685,0.1185866147,0.1294510365,-0.0090000276,0.0719856322,0.0478034131,-0.3362256587,0.0543728657,-0.0476700068,0.2388566136,0.0033361907,0.1545103043,-0.3303192556,-0.6893522143,-0.0198045671,0.1392314881,-0.0248521697,0.145348832,-0.1320491582,0.2101353407,0.0516631939,0.7545105219,0.247542426,-0.0989424363,0.0239295214,-0.0505445451,-0.1498227119,0.0122501003,-0.3787526786,-0.0279884879,0.5659729242,0.5356939435,-0.3186971247,-0.0022773431,0.1669673175,0.3624534607,-0.1111406162,-0.0095177665,-0.4973680079,-0.0270319544,-0.1853753626,0.1172206849,0.0561154112,0.0575089045,-0.1731967479,-0.1451406628,0.088765651,-0.0695768371,0.3130294383,0.209191069,0.3291224539,-0.2285303324,0.3800683022,-0.0179799479,0.1245592907,0.3865334392,0.3226422369,-0.1884766966,-0.2584013939,0.1959349066,-0.20280011,0.3252988756,0.4745903909,-0.0183884911,0.1406528503,0.170586735,0.4246020615,-0.3947393596,0.4877224267,0.552847743,-0.0566165037,-0.52134794,-0.2778941095,0.4880212843,0.1094363704,0.1318005323,0.2507802546,0.2139130086,-0.2760674655,0.2151191533,-0.0630096123,1.0572487116,-0.1979905367,0.5809550285,0.1332606971,0.2443154603,0.5600246787,-0.1634085625,0.1860708743,-0.371760577,-0.0683659837,0.1798360646,0.0233997349,-0.1514533609,0.0723548979,-0.2003833503,0.1706084311,-0.1733764261,0.2452854812,-0.2034247667,0.3449352384,0.1053252593,-0.2919416726,-0.2884564102,-0.0556589551,0.0399784893,0.0736346841,-0.1359112114,0.0102362344,-0.0451818183,-0.2965848446,-0.2285731584,-0.1514571458,-0.061606735,-0.0180228353,0.1356382966,-0.4155600667,0.2852833867,-0.0061043925,0.2018768042,0.6037316322,-0.2552428842,0.0815706551,-0.1647108644,0.0555000901,-0.0609474778,-0.0629587024,0.0183575992,0.0395532064,-0.0112470845,0.0081242891,-0.0311512724,0.3303093016,-0.3478429317,-0.1418488175,0.4647269547,-0.4978819191,-0.0072217686,-0.0802576467,-0.1155118644,-0.1249568015,0.0196050126,0.1544744223,-0.3114489019,0.0475622788,0.1161858588,-0.4666772485,-0.2527616024,0.2053713202,0.3127787709,0.1111268029,0.7269276977,0.2397605628,-0.0011833927,-0.0601239353,0.3727790415,0.5438687801,-0.4371438026,0.3184968233,0.262195617,0.0488700904,0.0291104503,0.1672333777,0.2023548484,-0.2320089191,-0.1916528046,-0.0115763303,-0.3460534811,0.0280330088,-0.0723961368,0.1560414881,0.1169248298,0.4359501898,0.0357437804,0.1918995529,-0.141484648,0.070005253,-0.0046818741,0.0214007664,-0.2310763597,-0.0026917052,0.1514332891,-0.0814433917,-0.0647879615,0.1667025536,0.0020320294,-0.0980186462,-0.2285377532,0.1536141187,0.0398370177,-0.0385811292,0.1680365056,0.1496417373,-0.1911123991,-0.2004627883,0.3976661563,0.14277637,0.1690921485,-0.1483489573,0.1470240504,-0.1166025996,-0.0988809764,-0.0595455244,0.1867140681,0.3183196485,0.0611324646,0.1301338077,0.3052387834,-0.0736712292,-0.0272994693,0.1290051788,0.1368476301,0.112856105,0.0538338274,-0.2543921769,-0.004885905,-0.0355647095,0.297403872,0.4269484282,-0.2330292463,-0.1478859335,0.3677387238,0.092793934,-0.0772649646,-0.0211797208,-0.1828501374,-0.2182419002,-0.120305188,0.3224663734,-0.0430026278,0.0732126683,-0.1927617341,0.12278869,-0.1977942884,0.1827686727,-0.0105105285,0.6458750963,-0.0254396852,0.1201389208,0.2480580807,0.0898621529,0.2171668112,0.6840233207,0.2106300592,0.155953005,-0.4127493203,0.0459773913,-0.147639215,-0.5137033463,-0.4766752422,0.1807593554,-0.1385578364,0.25104478,-0.1436317861,0.5851089358,-0.0925023481,-0.2286136001,-0.1587093025,-0.022502508,-0.2247796655,-0.1209856719,0.1957602054,-0.1729843765,0.171004042,0.2518329918,-0.0850014761,0.1018372327,-0.0150037985,0.0656477809,-0.0919593796,0.0620727353,-0.124301143,0.1526820064,0.0915936828,-0.255581826,0.2083453685,-0.0374238417,0.062365409,-0.1423091739,0.3047384322,0.504411459,0.4020273983,-0.1960876882,0.1983146518,0.1214170605,0.032037314,-0.0764960945,0.1660143435,0.342246145,0.0307880845,0.057471551,-0.0357736014,-0.1374777853,-0.0107414732,0.207541585,0.159353897,-0.2030375153,0.3092693686,-0.1445065141,-0.1366320252,-0.332480222,0.0277769882,-0.7808368802,0.6018417478,0.3939931393,-0.0382943489,0.0751705244,-0.0575839467,-0.0058173514,-0.1773789823,0.5535485744,0.4052788615,0.0527980253,-0.4022138417,0.0409857817,-0.2695578933,0.287891686,-0.4112736583,0.2815774381,-0.2419728339,-0.019810278,-0.0062522246,-0.1268972903,-0.0319104977,-0.2795478404,-0.735722959,0.1375220567,-0.1183147207,-0.2483353615,-0.2567407489,0.0827602595,-0.0633302778,-0.2646411359,0.3999221027,-0.0384019092,-0.1106514856,0.0319809504,0.1788473427,-0.236438632,0.1792990863,0.2317092717,0.3124322593,0.1329686195,-0.010536206,-0.2296808958,-0.2905624211,0.0933325961,-0.1113018245,0.1991880536,0.2345941216,0.3189929724,-0.21726951,-0.3325841129,-0.4254434407,0.2387919426,0.3529764712,-0.1453873664,-0.2179762572,0.0919542611,-0.075562261,-0.0770610869,0.0453809164,0.4668253064,-0.1706966758,-0.1004317477,-0.3269614875,-0.2557926476,0.5335482955,-0.3489524126,-0.3766444325,-0.2509153783,0.0626160577,0.2268562913,-0.3144404292,-0.7230035663,0.1928254962,0.2091497779,-0.0305041075,-0.4149523079,0.0981287062,-0.1121780127,-0.1236162484,-0.1959390044,0.6790009141,0.0530063771,-0.2023358494,0.3828852177,-0.2743701339]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/861","title":"Possible Bug: Small training\/dataset file creates gigantic output","comments":"Hi @NebelAI, we have optimized Datasets' disk usage in the latest release v1.5.\r\n\r\nFeel free to update your Datasets version\r\n```shell\r\npip install -U datasets\r\n```\r\nand see if it better suits your needs.","body":"Hey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n","comment_length":34,"text":"Possible Bug: Small training\/dataset file creates gigantic output\nHey guys,\r\n\r\nI was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.\r\n\r\nI've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?\r\n\r\nI've used the following CMD:\r\n`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`\r\n\r\n\nHi @NebelAI, we have optimized Datasets' disk usage in the latest release v1.5.\r\n\r\nFeel free to update your Datasets version\r\n```shell\r\npip install -U datasets\r\n```\r\nand see if it better suits your needs.","embeddings":[-0.2338341922,-0.3642718196,0.1098738834,0.327167213,0.5802890658,-0.1227116883,0.2948497832,0.405277729,-0.1822131574,0.1775166541,-0.0482908227,-0.1247077584,-0.3384362161,0.2299035192,0.0976041406,-0.0641692579,0.1237757429,0.1533718109,-0.104493238,-0.2381588221,0.0190756135,0.0444056056,0.2407272607,0.2485654652,-0.5509336591,0.0961136147,-0.0530444272,0.065711841,-0.4600453079,-0.0649351627,0.0930358097,-0.3468028009,0.218477875,0.5662022829,-0.0001273608,-0.0424782932,0.158666268,-0.0371259227,-0.036269132,0.151588738,0.1211322546,-0.0592242442,-0.1560094357,0.1039409339,0.2607800663,-0.0881411657,-0.0348140784,-0.0693408921,0.3328700364,0.2546795309,0.0841450766,0.1687070429,0.0530854799,0.0208149012,0.2235208601,0.277895391,-0.0392047167,0.0236582737,0.1884796172,-0.1935082525,-0.0342083648,0.2210821062,0.1108082533,-0.112880215,0.3315674961,0.1934318841,-0.0573355258,-0.3944152594,0.2410113364,0.3753354549,0.0406103432,-0.4069215953,-0.2597789168,-0.4071704149,0.2153960615,-0.4601975381,0.0786264762,0.1565641165,-0.1745781153,0.0869323239,-0.2822416127,-0.433242768,-0.2624081075,-0.0644187182,-0.1952349842,-0.0204993654,-0.2531405985,-0.0878002122,0.03168533,-0.0513358004,-0.3014719188,0.001802813,-0.0131004192,0.1772996336,-0.0929567218,-0.2810669541,-0.0840085447,-0.3132231832,0.2787923813,-0.5234856606,-0.1843097955,-0.4352833331,-0.2017560005,-0.3546653688,0.4170531631,0.472019732,-0.396017313,0.1702489257,0.166572541,0.0548957549,-0.1536719203,-0.0758734569,-0.1547504961,0.0704373196,0.0707222223,-0.4381209016,0.129961893,-0.2795091271,-0.1710017174,0.389624536,-0.2968327403,0.203411743,0.0244191531,0.3456146717,-0.0571381301,0.1504266262,0.0753576159,0.0295034647,-0.4357024431,-0.0140295429,-0.0874336883,0.0690611601,-0.187155813,0.1723109186,-0.0044256677,0.0444649197,0.0975162014,-0.0309927966,0.2879881859,-0.0614806116,0.0658553243,-0.2731784284,0.1098375544,0.1949203312,-0.1669240147,0.4897101223,-0.0056195566,0.0467573926,0.0041758274,0.0890709981,-0.2138061821,-0.5087085366,-0.1334372163,0.1073392332,-0.0225376859,0.270331651,-0.1141858697,0.3702563047,0.4396837354,-0.0440382175,-0.0226347893,-0.1944944859,-0.2229566425,-0.0126110548,0.3754176795,0.4236915112,-0.5573843122,-0.1374476403,-0.2831021845,0.2760488689,0.3170405924,0.6052901745,-0.0366573595,0.5938890576,-0.2680997849,0.1369596124,-0.2820636332,0.2830556929,-0.4317550957,0.3631533384,-0.0558017679,-0.0812832415,0.0675941855,-0.0509304926,0.1601727456,0.0494954251,0.4206693172,0.0105388872,-0.1910239309,0.0278292429,-0.3491926789,-0.3933972418,-0.3029136062,0.0391489714,0.2074746937,-0.1426802725,-0.2428856492,0.5842377543,0.1493231952,-0.1176010892,0.1443619877,0.4034582973,0.1061514243,-0.0489939563,0.0309899431,-0.1546383053,-0.152195558,0.0177603085,-0.2466843575,0.1808103472,0.0917116851,-0.2361067832,-0.0985658318,-0.0183471236,0.0784693137,-0.1859901994,-0.0025018868,-0.1868593842,-0.1341238618,0.0562628359,-0.0534956567,-0.0828902647,-0.1111693531,0.0557697192,-0.4824645817,-0.021635972,-0.0438356549,-0.1878069937,-0.0588857941,0.3047100008,-0.1567812711,0.0496466309,-0.1076477319,0.2003514916,0.0655244663,0.0901458561,-0.0104257148,0.1849321127,0.264739275,0.0936757922,0.1331100613,0.1993817389,0.0828021467,-0.0678429306,-0.14168109,0.1253068745,0.1942510158,0.1309911758,0.2353687584,-0.238740474,-0.0969014019,0.1526818573,-0.0067462875,-0.2505146265,0.2871016264,0.0482300855,0.2550387383,0.1577132046,-0.1824488491,-0.0940706655,0.3434639573,-0.1666010469,0.0557361171,0.4261432886,-0.3578202128,-0.0915519223,-0.0463074259,-0.1888587028,0.3729067147,0.0803322345,0.1341005117,0.1067790464,0.0644714683,-0.0672613978,0.1499306411,0.2874512672,-0.117579177,0.1728916913,0.1884485483,-0.0295738932,-0.255150497,0.0547025092,-0.2172520757,0.3717318475,-0.3848313391,0.3306095004,0.094664447,0.032725431,-0.4031649232,-0.2052146494,-0.1755549014,0.0127516724,0.1642492265,-0.1578899324,-0.0829308257,-0.0999551937,0.3912376165,0.2013758719,0.1962990761,0.3108693659,0.033097636,0.138301909,-0.3434259295,-0.0045561604,0.1663041711,-0.1633000374,0.1363902241,0.1126272157,-0.2667165399,-0.062426392,-0.2402979136,0.1260363311,-0.1923039705,0.2783963084,0.2113269866,0.2293050587,-0.3828122318,0.0926501378,0.0435232259,-0.0196592789,-0.1070588678,-0.0421098135,-0.1186993942,0.1387941241,-0.0932442546,-0.3279576302,-0.1222926825,-0.0531825721,0.3685314655,-0.0001866183,-0.0081697125,0.2316599637,0.2695368826,0.132894665,0.0196138173,-0.1554873884,0.0083823735,-0.1846472472,0.1803590357,0.0460972153,-0.1881256253,-0.2821445167,0.0419535525,-0.0011371947,-0.1777221411,-0.5618310571,0.0184513349,-0.4612651169,-0.2686636448,-0.2251045257,0.4086788595,0.1059861481,-0.0596177578,0.0595238023,0.0867204517,0.0746403039,-0.0934038609,-0.2141722739,0.2236485779,-0.4218251705,0.4128449261,-0.0376652479,0.8130553961,0.3064289987,-0.068814449,-0.0068479721,-0.0082486682,0.4584790468,-0.455973357,-0.2210274935,0.2777840793,-0.1431837827,-0.3980737031,0.2701790929,0.130034551,0.4550913572,0.1515551358,0.0478054993,0.2745588124,-0.3100079,0.3527386785,0.1191464737,0.1723787487,0.001176604,0.0965537727,0.0363406464,-0.2952890694,0.0445205905,-0.0182025824,0.1653150171,0.0148821529,0.0854096785,-0.3076306581,-0.7416739464,-0.0337135345,0.0653304756,0.0161201917,0.1323540658,-0.0856465623,0.1822294891,0.0500860736,0.7840266228,0.2417460084,-0.1469583362,-0.0631196201,-0.0090904422,-0.2243033051,0.0502407365,-0.3604033291,-0.0077724159,0.5267219543,0.6359656453,-0.33792454,-0.0449704528,0.1372144967,0.381801486,-0.1095816046,0.0351184793,-0.4170676768,-0.0714577064,-0.2245296389,0.1731090993,0.0825726166,0.0398782678,-0.1176660657,-0.1058454216,0.1136167198,-0.0607748218,0.3254383802,0.2567200363,0.3178471327,-0.3010819852,0.3417423666,-0.0462817885,0.1264616102,0.3970063627,0.2715195119,-0.2576500475,-0.2769014537,0.2349772006,-0.1083866656,0.3079812825,0.5227623582,-0.0646281689,0.1311268955,0.155381307,0.4302624762,-0.4125352204,0.5151005387,0.5527086854,0.0177975986,-0.4765746295,-0.2469357997,0.4360094666,0.1092831343,0.1275287271,0.210526675,0.2606101036,-0.3030824065,0.270116508,-0.0946194679,1.0859620571,-0.2421150804,0.5881166458,0.1212818995,0.2371953875,0.5970579386,-0.2419917583,0.1563251913,-0.3379874229,-0.131324634,0.1505072415,0.0553208441,-0.1317776293,0.039905481,-0.1674168557,0.2230981141,-0.0800758898,0.1945078671,-0.2192201614,0.3541292548,0.0505176,-0.3207402229,-0.2754777372,-0.0320471451,-0.0303048175,0.0507145897,-0.1687881947,0.0587943904,0.040565487,-0.2991409004,-0.3170065582,-0.1916313767,-0.0658529997,0.0150368623,0.1493944377,-0.3508225679,0.2959386408,0.051182583,0.2471508235,0.6219081283,-0.2682354748,0.0588832051,-0.2696535289,0.0439972579,-0.0632771924,-0.0463334471,-0.0081132175,0.0676463097,0.0005962919,0.0085469354,-0.0352821387,0.2803077102,-0.3483643234,-0.1616436094,0.4436450303,-0.467615664,-0.0234429091,-0.1549752057,-0.1121089309,-0.1517965645,0.0233200304,0.1797081381,-0.2624877691,0.0978789702,0.0117113739,-0.3891355097,-0.2371117175,0.2602454424,0.2394696027,0.0908640698,0.680362463,0.3069350719,-0.043292474,-0.0754632652,0.3743827045,0.4730662405,-0.4497230649,0.31705302,0.1860165149,-0.0346311368,0.0334120244,0.1253682822,0.2458349913,-0.3215655684,-0.1962847561,-0.0428497978,-0.4025847912,0.0923330486,-0.042891752,0.1559311002,0.1785841137,0.2970409095,-0.0137361158,0.1401788592,-0.1373620182,0.1233335584,-0.0231050272,0.0687610134,-0.1933572888,-0.0587361716,0.2272243202,-0.1636864096,-0.0683894753,0.204264462,0.0142689692,-0.1004520282,-0.1986792982,0.1495772749,0.0027418183,-0.0232982282,0.1306261122,0.0538234189,-0.1501656473,-0.2443066984,0.4213886857,0.1724332273,0.1414135247,-0.139263019,0.1098979414,-0.1006514654,-0.0634010956,-0.0140845012,0.1469547153,0.2605257034,0.0540863946,0.1588280052,0.3157699406,-0.0877612457,-0.1094087288,0.1993462741,0.1971243471,0.1085846275,0.0951880664,-0.2654798031,-0.0809269845,0.0378724001,0.2864855826,0.4713871777,-0.2842045426,-0.1716521978,0.3851933181,0.1027003527,-0.0477528907,-0.1189641356,-0.1335844398,-0.209462136,-0.139550969,0.2914474905,-0.0526711233,0.0205770489,-0.1172215715,0.1852547973,-0.1063658074,0.1524030268,-0.0206125025,0.653932333,0.015221539,0.0670527816,0.178647235,0.0688594505,0.2759259045,0.8238028884,0.1972072423,0.1680946797,-0.3683773875,0.0733839646,-0.2112159729,-0.4943948686,-0.4449974597,0.1355960518,-0.1487525403,0.280988425,-0.1661428213,0.5747609735,-0.1352188736,-0.2540167868,-0.1453139633,-0.078365773,-0.2735462487,-0.1423636973,0.2345374376,-0.1449487358,0.1949322075,0.2462967038,-0.1218235344,0.0105039813,-0.0223572142,0.0733818188,-0.1523852646,0.0285175983,-0.1688957661,0.1849359423,0.0798945427,-0.2207345217,0.2122030854,-0.0599278919,0.0100177517,-0.2495431155,0.3357707858,0.4979715645,0.384826988,-0.2373092324,0.2271024734,0.1799070239,0.0169183854,-0.0835810676,0.0812611133,0.2790589929,0.1057369336,0.0313186534,-0.0387332253,-0.1393174082,-0.064955458,0.1780905128,0.1297256649,-0.2238082886,0.2087999731,-0.178438291,-0.1754206121,-0.2495660633,0.028815316,-0.7493742108,0.5085187554,0.5131714344,-0.0313962363,0.1146012992,-0.0797065496,-0.0068541849,-0.1509418488,0.5040619373,0.4640610218,-0.0103108482,-0.3956050873,0.0211509857,-0.3162700236,0.2524128258,-0.3745406568,0.3190282583,-0.1406810284,-0.0834925026,-0.0082902079,-0.1462576836,-0.0114829959,-0.2426197231,-0.6652166843,0.1791588217,-0.1702089459,-0.2547819614,-0.2493761629,0.0223898571,-0.0184739735,-0.2822555304,0.4183259308,-0.0032186839,-0.1322531551,-0.0118236244,0.2694914937,-0.2186582685,0.1853106469,0.266577363,0.2303338945,0.0351795293,0.035174869,-0.2640859485,-0.3061712384,0.1012737378,-0.0681265444,0.1514075547,0.2569340467,0.352964133,-0.3432300687,-0.2458403707,-0.3924256563,0.1620886326,0.3589940667,-0.1316670477,-0.2357919663,0.076217778,-0.1069594398,-0.0141407661,0.0605250783,0.4861750305,-0.1284477115,-0.0805444494,-0.3727774918,-0.2169008702,0.4692784548,-0.2248285711,-0.4343772829,-0.2305406332,0.0410190709,0.2724094391,-0.2714159191,-0.8018815517,0.1316687316,0.1668115705,-0.0057348125,-0.5056576729,0.135008648,-0.087055631,-0.1299732625,-0.1699074507,0.7421054244,0.0822327882,-0.2132929713,0.4542503059,-0.2582867444]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"Hi,I also posted it to the forum, but this is a bug, perhaps it needs to be reported here? thanks ","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":20,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nHi,I also posted it to the forum, but this is a bug, perhaps it needs to be reported here? thanks ","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"It looks like the official OPUS server for WMT16 doesn't provide the data files anymore (503 error).\r\nI searched a bit and couldn't find a mirror except maybe http:\/\/nlp.ffzg.hr\/resources\/corpora\/setimes\/ (the data are a cleaned version of the original ones though)\r\nShould we consider replacing the old urls with these ones even though it's not the exact same data ?","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":59,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nIt looks like the official OPUS server for WMT16 doesn't provide the data files anymore (503 error).\r\nI searched a bit and couldn't find a mirror except maybe http:\/\/nlp.ffzg.hr\/resources\/corpora\/setimes\/ (the data are a cleaned version of the original ones though)\r\nShould we consider replacing the old urls with these ones even though it's not the exact same data ?","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"The data storage is down at the moment. Sorry. Hopefully, it will come back soon. Apologies for the inconvenience ...","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":20,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nThe data storage is down at the moment. Sorry. Hopefully, it will come back soon. Apologies for the inconvenience ...","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"Dear great huggingface team, this is not working yet, I really appreciate some temporary fix on this, I need this for my project and this is time sensitive and I will be grateful for your help on this. ","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":38,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nDear great huggingface team, this is not working yet, I really appreciate some temporary fix on this, I need this for my project and this is time sensitive and I will be grateful for your help on this. ","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"We have reached out to the OPUS team which is currently working on making the data available again. Cc @jorgtied ","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":20,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nWe have reached out to the OPUS team which is currently working on making the data available again. Cc @jorgtied ","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"thank you @thomwolf and HuggingFace team for the help. ","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":9,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nthank you @thomwolf and HuggingFace team for the help. ","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"OPUS is still down - hopefully back tomorrow.","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":8,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nOPUS is still down - hopefully back tomorrow.","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"Hi, this is still down, I would be really grateful if you could ping them one more time. thank you so much. ","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":22,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nHi, this is still down, I would be really grateful if you could ping them one more time. thank you so much. ","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"Hi\r\nI am trying with multiple setting of wmt datasets and all failed so far, I need to have at least one dataset working for testing somecodes, and this is really time sensitive, I greatly appreciate letting me know of one translation datasets currently working. thanks ","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":46,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nHi\r\nI am trying with multiple setting of wmt datasets and all failed so far, I need to have at least one dataset working for testing somecodes, and this is really time sensitive, I greatly appreciate letting me know of one translation datasets currently working. thanks ","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"It is still down, unfortunately. I'm sorry for that. It should come up again later today or tomorrow at the latest if no additional complications will happen.","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":27,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nIt is still down, unfortunately. I'm sorry for that. It should come up again later today or tomorrow at the latest if no additional complications will happen.","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"Hi all, \r\nI pulled a request that fix this issue by replacing urls. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/1901\r\n\r\nThanks!\r\n","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":15,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nHi all, \r\nI pulled a request that fix this issue by replacing urls. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/1901\r\n\r\nThanks!\r\n","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/854","title":"wmt16 does not download ","comments":"It's still down for the wmt.","body":"Hi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz","comment_length":6,"text":"wmt16 does not download \nHi, I appreciate your help with the following error, thanks \r\n\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"wmt16\", \"ro-en\", split=\"train\")\r\nDownloading and preparing dataset wmt16\/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wmt16\/ro-en\/1.0.0\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/root\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wmt16\/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308\/wmt_utils.py\", line 755, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in _single_map_nested\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 181, in <listcomp>\r\n mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/root\/anaconda3\/envs\/pytorch\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/opus.nlpl.eu\/download.php?f=SETIMES\/v2\/tmx\/en-ro.tmx.gz\nIt's still down for the wmt.","embeddings":[-0.444955498,-0.4224762022,-0.0815040842,0.4179413617,0.5046890974,0.1244975701,0.1281453669,0.2139965892,0.2786702514,0.001354157,0.0727131143,-0.110461086,-0.2351618558,0.097207725,0.2837482691,-0.0405813567,-0.1238627285,-0.002883099,-0.4859333932,0.0390384048,-0.2367140353,0.1167636067,-0.1118449643,-0.0652594045,-0.3488698602,-0.0671071485,-0.1394528598,0.0584430955,-0.360514909,-0.2884312868,0.3591679037,0.0484933145,0.2195616364,0.4216263294,-0.0001045874,-0.0344254412,0.2609169781,-0.1103322431,-0.1643058062,-0.3919683993,-0.3435014486,-0.1982588619,-0.0863412246,-0.0891985819,-0.0184502378,-0.1176216602,-0.1129943654,-0.3425154686,0.2524401546,0.3540210128,0.3467800915,0.2697255611,0.3545668721,-0.1914296448,0.1382765472,0.1699798554,0.0555037595,0.1284032315,-0.0070466073,-0.1400783062,0.1166838035,0.1394858658,-0.0372346565,0.2416531146,0.104287751,-0.1586447954,0.0307699218,-0.254928261,0.2687593997,0.1392050236,0.3950155675,-0.1157429069,-0.1314078271,0.1513412595,0.065062128,-0.5738540292,0.2200950086,0.5009623766,-0.1863734424,0.1921614856,-0.2714767158,-0.0870952085,0.0238122586,0.1413062066,-0.2197024077,0.2921693921,-0.213550061,0.0183895659,0.1849008203,-0.1017094776,0.0880924761,-0.0139018279,0.1450838894,0.2526480556,-0.2700490355,0.1261817366,-0.1145764664,0.1292577088,0.0806544647,0.0086127445,-0.2520152926,-0.1507218033,0.2881473005,-0.0259473715,0.0269001275,0.1888409704,-0.2194881141,0.0834902674,0.0834250972,0.1954315603,-0.0369618386,-0.0546250939,-0.3145231903,-0.2925898135,0.0326478705,-0.0937117338,0.1151885688,0.0685743988,-0.2826316357,0.0161389802,0.0049903886,-0.1779205352,0.0023902771,0.3310005367,-0.2150157988,0.148575604,0.2978157997,0.0600120388,-0.0970932543,-0.106148757,-0.2628639042,-0.0803785697,-0.1983224154,-0.056407325,0.3051498234,-0.0921546817,0.3761104941,0.0229630414,-0.179779619,-0.1802678704,0.0199732501,-0.1830584705,0.1112451032,0.4338179529,0.0693689287,0.237574324,0.0718808621,0.254779309,-0.0013334155,0.2246186733,-0.0462478027,-0.4587911963,0.1238833144,0.370313406,-0.1037134379,-0.115419358,0.0248291045,-0.1448108256,0.0704493821,0.2599559426,-0.1440329552,-0.2033017129,-0.1653168947,-0.2190619856,0.5277351141,0.393963933,0.0342416279,-0.0421035253,-0.4067505896,-0.2453195155,0.3608443141,0.2823581398,0.0014633266,0.1401390284,-0.3047563136,0.0974318907,0.2519396544,-0.2148155421,-0.2565926909,0.211815387,-0.2131187767,-0.0246797614,-0.0519014411,0.0910042152,0.0106451157,0.0145032462,0.267085135,0.1857897788,-0.3312119544,0.0249924865,-0.1970980912,-0.3113334775,0.4462002218,0.2143390477,0.2483334541,-0.0707214475,0.0157011021,0.114467077,0.3708969951,0.2580592632,0.0792705491,-0.0338332728,0.2125306726,0.0079512121,-0.2015179247,-0.1089143082,-0.2836967707,0.1032906994,0.2451010346,-0.0778647214,0.0508962013,-0.0072366488,-0.4273627698,0.0413815603,-0.168958351,-0.1493412256,0.1657439917,-0.0761240497,0.1448943615,-0.0492947139,-0.0764248297,-0.0518814288,-0.2385359704,0.0705099478,0.1058792993,-0.0204066541,-0.3056635261,-0.0091247633,0.2450969666,-0.0581195839,0.1331098229,-0.0372367017,-0.135418728,0.2523603141,0.123330079,0.2528036237,-0.0802929029,0.0251827948,0.1523814946,-0.452506125,0.0430596173,0.0419443399,0.1044071913,-0.1576236039,0.2183504254,-0.0314418413,-0.0910328627,0.0418949425,0.174456507,0.1563680768,0.1477839649,-0.024936242,0.0901833549,-0.1331830174,0.3215939999,-0.0938108116,0.2526182532,-0.0654185861,-0.0649184287,-0.1171566918,0.4020731449,0.152085036,0.0896184817,0.044230748,-0.3025979102,0.0285023507,-0.2157823443,0.3114795089,0.4959137142,0.1446499079,0.1639534533,0.1199491099,0.1569378823,-0.2816541195,0.4310475886,-0.040102955,0.2351522595,0.2522007525,-0.1961644143,-0.0563999377,-0.2167151421,0.2778975368,0.0175765604,0.2078112811,-0.1578265578,-0.2426187694,-0.2548194826,-0.0673511997,-0.4623289704,-0.5436559319,-0.2178747207,-0.1321282089,-0.2382991165,0.0240342431,-0.0367041901,0.1426422745,-0.0420636795,-0.1025992855,0.0817945004,0.1268719584,0.0026955113,-0.0534274839,-0.0635396317,0.1841251254,0.2524006665,-0.3535847664,0.5754960179,-0.3072552681,0.0299873985,-0.2428164035,-0.1333234608,-0.0243879445,-0.0370058864,0.0867503658,0.3703727424,0.3781490922,0.2589347661,-0.0355772786,0.369383812,-0.2827515304,-0.0627758875,-0.121466957,0.1225737482,0.0374437831,-0.2173156589,-0.562472403,-0.4598293304,-0.4110196233,-0.0917023495,0.0304594226,0.1562970281,0.1512025297,0.0725564435,0.0382239632,0.1439446807,-0.0554896332,-0.3277630508,-0.3739012778,0.3905153573,-0.2670379281,-0.4456063807,0.2986950874,0.0598832257,0.4613804221,0.2046910077,-0.4112800956,0.3908862174,-0.1190442592,-0.0669095516,0.0473832563,0.1677639037,0.1566010565,-0.0419885479,-0.1558275223,-0.1271512806,-0.1122654378,-0.1918485314,-0.1983194798,0.5375465751,-0.1967826039,0.337072134,0.0668314472,0.7139240503,0.141764015,-0.167829141,0.2005022764,-0.10505981,0.1060315967,0.1251528114,-0.4757870138,0.1009271666,-0.1338579655,0.171341151,0.1957198381,0.0078043779,0.1225353926,-0.2567455769,0.009698797,-0.2643204927,-0.246010825,0.0622865409,0.0964218676,0.4988240302,0.1776275784,0.0379241705,-0.0200700909,-0.0173667688,-0.013074779,0.3110693395,0.1599546522,0.1306657195,0.1260280162,-0.0890352353,-0.3551563621,0.0651445761,-0.0090178279,0.448325485,-0.0488910787,-0.1861808598,-0.0250149313,-0.0229497831,0.5050638318,0.1352038532,0.2418664545,-0.0990693942,-0.1132145077,-0.3881953359,-0.0088242749,-0.0477997065,-0.0792361721,0.1792103499,0.0337592065,-0.2113889903,-0.1805664301,0.3363735676,0.196585983,-0.2405355871,-0.3452710211,-0.214828521,-0.3769052029,-0.375888586,-0.0632985607,0.0156026492,0.2370364666,-0.2710612416,-0.0545754619,0.0127762519,-0.0314765349,0.081289731,-0.0627886951,0.2074970007,0.0170503519,0.3276123405,0.2472558618,0.2483445853,0.3023757339,0.4499738812,-0.0790659189,0.0300896149,-0.045030944,-0.0318384729,0.1685838848,0.0883079171,-0.1288238913,0.0313172899,0.3365079165,0.1797917783,-0.5687059164,0.1187111884,0.4142961204,-0.0442772955,-0.2248915434,-0.2964838147,0.1844127625,0.0234355778,0.1980578154,0.2240127176,0.0846090689,-0.102600269,-0.3729058504,0.0741106123,0.7303836346,0.0040986263,-0.0063374615,0.0541395694,-0.245799005,0.0493166,0.1965435743,0.0595950671,-0.2025648504,0.0823460743,0.0715467781,-0.1753578931,0.0899460316,0.1638875753,-0.2038238496,0.1945634633,-0.0903947279,0.3699932396,0.1237482578,0.2732431889,-0.299500525,-0.0765024051,-0.1843225658,0.2496200651,-0.1579498351,0.5097196698,-0.2337878197,-0.0218787454,0.2522149682,-0.3380146623,-0.3348367214,-0.0194519646,-0.6515519023,0.1543282568,-0.2069599479,0.0544424467,0.2665858865,0.0717395023,0.0452236794,0.3972819746,-0.1099756584,0.1481288671,-0.2182461917,-0.1837649047,-0.149944365,0.1043707058,0.0388212949,-0.2055919766,-0.242231831,-0.0304818135,-0.0293192118,0.101859495,0.1618238091,-0.0849094391,0.2629016638,-0.1588936299,-0.131407097,-0.0529144853,-0.0808756351,-0.1313951463,0.2391680479,0.0814331844,-0.105900459,0.35980317,0.1204868108,-0.0357933044,-0.1756919324,0.4182898104,-0.121393837,0.0133032547,0.400534302,0.2816832364,-0.1847721487,-0.4077315331,0.0011175231,-0.228067264,-0.2547617257,0.2138407081,0.1572362483,0.2700095475,-0.0490143336,0.1123867556,0.2817123532,-0.2257385105,0.1532507688,-0.4374891818,-0.3613281548,-0.0451176725,-0.0801632926,0.1647503376,-0.0839295685,0.3007494211,0.1840868741,0.3149897456,-0.3982215524,0.1844789237,-0.0554504916,0.0410329774,0.408641547,-0.1523307562,0.2701260448,-0.0001028442,0.2006694078,0.0359568447,-0.3535069823,-0.2827748656,-0.1789458692,0.0797092766,0.0645149276,-0.3131317198,0.2315576375,-0.2747566104,-0.1528648585,0.1786975265,0.047386013,0.1009499058,-0.0488116294,0.1522327811,0.0410715044,0.0885967836,-0.2128453255,0.070073843,-0.0519008376,0.3334442377,0.0832397267,-0.0684862062,0.3025777042,-0.1539804339,-0.099962607,-0.2193925977,0.1600324064,-0.2081458569,0.3932882249,-0.184799999,-0.0114570362,0.1140617579,0.3009463251,0.3354029655,-0.0449945591,0.1436927915,0.1372206807,0.3516526818,-0.408049345,0.1773460805,-0.1444799006,0.0148590049,-0.070278801,0.000802085,0.0222215485,0.0371155627,-0.1540627331,0.2090328038,0.442977041,-0.2516424358,0.1013131216,0.2070240825,-0.1443167478,0.1512601376,0.0537341237,0.0145270033,0.0511470661,0.3569306433,-0.3507205844,0.3197171688,-0.1200648174,0.0391064771,0.1024587452,-0.2943322659,-0.0493203923,0.262611717,-0.2538187206,-0.0398402773,-0.1569648981,0.2941789627,-0.1178480983,0.0144354785,-0.1574142128,0.2634153962,-0.0796028227,-0.0266712792,-0.224519372,-0.1667223275,-0.2416519076,-0.0081914496,0.1487244517,-0.1794112027,0.292416513,0.265994072,-0.021894034,-0.5418660641,-0.1938093305,0.1615170091,-0.0660706311,-0.3661060035,0.0962023363,0.426284194,-0.1073656753,0.1673747003,0.4767377377,0.3966048956,0.3219347894,-0.1423855722,0.0870662034,0.0576878488,-0.1005892903,-0.1879336387,0.2051906437,0.0790675804,-0.0485189669,0.2097937614,0.2723872364,-0.2374074608,-0.19492203,0.0235634204,-0.1338632256,-0.063516207,0.2793864012,-0.1369129717,0.0430594012,-0.4318418801,0.2305633873,-0.2953834236,0.3097433448,0.5763674974,-0.0516445786,0.167992264,-0.2874497473,0.114385657,-0.0044967607,0.6133633852,0.2889218628,0.0559175722,-0.4034251869,-0.4311852753,-0.7414616346,0.0106799258,0.0747843981,0.0936226174,-0.1725310534,0.0881135091,-0.0538026541,0.0651678815,0.249034375,0.0820928887,0.1977203637,0.0279185697,-0.0566339307,-0.0552846454,-0.200275138,-0.0143969934,0.1361457258,-0.3531438708,0.1582952589,-0.3870320916,0.129818365,-0.4399925172,0.2706383765,-0.0487770773,0.2809252143,0.2876272202,0.1505339742,0.3705016971,-0.0907473862,-0.0546107553,-0.195024237,-0.0690794438,-0.1504274607,0.1500723362,0.2182554752,0.17057845,-0.001808534,-0.3095676303,-0.44950822,0.0844907314,0.0031600106,-0.0372306816,-0.1925239712,0.0436136872,-0.1081488356,-0.1006138846,0.3356007338,0.229915604,-0.1436970085,0.0228848476,-0.3451318145,-0.3848659694,0.44604671,-0.1559346765,-0.2160535306,-0.2820884287,0.0934079513,-0.0535831824,-0.0098088281,-0.4851154387,0.3032738268,0.2533223331,-0.0807626098,-0.200013116,0.4165595174,-0.1124202013,-0.0350138843,0.0745051801,0.425993979,0.0531243235,-0.2661990821,-0.2081848979,-0.1793360114]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"Unfortunately `concatenate_datasets` only supports concatenating the rows, while what you want to achieve is concatenate the columns.\r\nCurrently to add more columns to a dataset, one must use `map`.\r\nWhat you can do is somehting like this:\r\n```python\r\n# suppose you have datasets d1, d2, d3\r\ndef add_columns(example, index):\r\n example.update(d2[index])\r\n example.update(d3[index])\r\n return example\r\n\r\nfull_dataset = d1.map(add_columns, with_indices=True)\r\n```","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":58,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\nUnfortunately `concatenate_datasets` only supports concatenating the rows, while what you want to achieve is concatenate the columns.\r\nCurrently to add more columns to a dataset, one must use `map`.\r\nWhat you can do is somehting like this:\r\n```python\r\n# suppose you have datasets d1, d2, d3\r\ndef add_columns(example, index):\r\n example.update(d2[index])\r\n example.update(d3[index])\r\n return example\r\n\r\nfull_dataset = d1.map(add_columns, with_indices=True)\r\n```","embeddings":[-0.4226087928,-0.0862603486,-0.1545989662,0.0781621411,0.1073093116,0.4110481143,0.3115082681,0.4648504257,0.11830201,0.1705961972,-0.1758232564,0.3464903831,-0.0750733614,0.5408211946,-0.2195973098,-0.2845952511,0.158336103,0.5231662393,-0.5183831453,0.0794194117,-0.436337471,-0.114001967,-0.3252285123,-0.3519526124,-0.330509007,0.2272312343,-0.3263256848,0.0496895649,-0.2420312911,-0.2773002088,0.1845785081,0.1565621346,0.173753351,0.1870673746,-0.0001018656,-0.0065462426,-0.014074605,-0.0987925455,-0.2388120592,-0.2386993021,-0.4274660349,-0.5326142907,0.0191351399,-0.1855815798,0.00708585,-0.54960531,-0.1515820473,-0.1952258646,-0.0363274477,0.2691266239,0.2765330672,-0.0024136822,0.1438422054,-0.2205383331,-0.032708846,0.0477842875,-0.1488682181,-0.2383367568,-0.1028846502,-0.0304823443,0.2678528428,0.0924807116,-0.0786667019,-0.0955473408,0.023014795,0.242949158,0.3289216161,-0.4637416899,0.2783742845,0.3595761955,0.5390436053,-0.1761268824,-0.3088946939,-0.0107760755,0.2730102837,-0.2251131088,-0.1153179258,0.1722409278,0.1841141433,0.1055816859,0.0051624589,-0.0969100967,-0.23845011,0.3530945182,-0.1239473671,0.5723182559,0.0937582031,0.1246292293,0.1647108048,-0.2323429585,0.2123437822,-0.23055996,-0.0263885669,0.3826236725,-0.5797169805,-0.0849746242,-0.1941434741,-0.2402858585,0.1860107481,0.0934349298,0.0127488635,-0.0230270587,-0.1148875579,0.2161030173,0.1035968661,0.1430449039,0.1178156883,0.3845241368,-0.1129865721,-0.2188611925,0.3609628081,-0.0352086946,-0.3074696958,-0.4276658595,-0.0928756297,-0.0125978654,0.0828213617,0.2491242439,0.0911288038,-0.1014141664,-0.1131042466,-0.0271373149,-0.0245385617,0.0541519746,0.1044036224,0.7092756629,-0.0496406145,0.2461360544,0.2259809971,0.0167850908,-0.2002800703,0.3086336255,0.1093921065,-0.0477167293,0.1133465692,-0.202474311,-0.0465146936,0.3516930938,0.1328005195,0.1754063815,0.0056137159,-0.2470530123,0.1815116853,0.2198554426,0.1323290914,-0.04526278,0.1345876306,-0.1955263466,0.0179670826,0.3099166751,-0.1724102497,-0.0889016166,-0.5252352357,0.2082276791,-0.0423238724,-0.0700515583,-0.3134088218,0.362336725,-0.0771497115,0.053949222,0.1155345589,-0.0474661477,0.0621779561,-0.1661249548,0.2009577751,0.1113442332,-0.2669385672,0.008026164,0.1536028087,-0.1589449644,0.144599691,-0.060014613,0.1177267805,-0.2345389277,-0.0067726024,0.005962213,0.3625376821,-0.3298892677,-0.092698358,-0.0321074724,-0.119783029,-0.0059565259,0.2458178103,0.0969633609,0.1838972867,0.2839229703,0.2015565485,0.048203133,0.03500719,-0.0099596903,0.0151676936,-0.1204555184,0.4265736043,0.2494770288,-0.0432004891,0.2766696513,-0.0150050763,-0.458845675,0.3301061988,-0.3999517858,0.2924720943,0.1318642795,0.1267351508,-0.2035218477,0.0067223655,-0.1177159026,-0.5746800303,0.0519519411,0.2862970233,0.2127360255,0.0785789639,-0.1286859661,-0.1028125957,-0.1928750426,0.0099845137,0.0941524804,0.1835021228,0.041317977,0.0519455858,-0.2493308485,-0.3810605705,0.3606599867,0.0950973034,-0.043297831,-0.150848031,0.2130998522,0.1889882833,0.0291418806,-0.0165353753,0.2718974352,0.094373554,-0.1173461825,0.1646354496,0.3794598281,-0.3452474475,0.0082483059,0.0471475646,0.2545304894,0.1226964965,0.2721693516,-0.1793629527,-0.3759014606,0.1136744767,0.0921133831,-0.0816449523,0.4765714705,-0.2874643803,0.2087194175,-0.0956577808,0.1742875129,0.2577313781,-0.0465931073,-0.2239287049,-0.3543518484,-0.0163601711,0.2291424125,-0.1916588992,-0.023656385,-0.0687824488,0.1239473447,0.2656056583,0.0214643311,-0.1022831798,-0.0943579972,-0.1167930365,0.1243985221,0.2267947644,0.1167610437,0.2644933164,0.3999997377,-0.1086042523,0.2540391684,-0.2752259374,0.0746818334,0.0916333795,0.037317358,-0.077711314,0.4481419325,0.2226282358,-0.0754357949,-0.0185630322,-0.2296893299,0.1100940779,0.5023943782,-0.0948243812,-0.1374462992,-0.2418773025,-0.0132372165,0.0900166482,-0.3467955887,0.0228301026,-0.2478994131,-0.1796188504,0.1272639781,-0.1465993971,0.3609495461,-0.3012213111,0.130200386,-0.0414414369,-0.023027841,-0.0322708227,0.071534358,0.1925762147,0.1740308702,-0.1611603647,-0.08097969,-0.0969500318,-0.119551599,0.2034372389,-0.1111137196,0.0296608675,0.0209119581,-0.1271328479,-0.4854318798,0.0185908508,0.0776493698,-0.2564775348,-0.1758208275,0.1687879264,0.3261306882,0.0632399768,0.1794681698,0.0166821964,0.0185164176,-0.1775396615,-0.2074855417,-0.2333163023,-0.4139026701,0.1558990926,0.2039990872,0.2865823507,-0.038900733,0.4930993617,-0.0612504482,0.0392677672,-0.0590723567,-0.1255935133,-0.336705327,0.4871846437,-0.1394262612,-0.0531917028,0.1305378675,-0.2593684494,0.1299346089,0.281085521,0.0949886888,-0.2102482617,-0.0393762663,0.1496009827,-0.3974593878,0.2691262066,0.4252674878,0.3284216821,-0.1030907184,0.0182091072,-0.0089028059,0.0823037326,0.2164365053,-0.2202183753,-0.2070922703,0.1523664743,-0.0258811712,-0.1935560107,0.4212956131,0.0882193223,0.17858693,-0.2154992223,0.3437775671,0.0687642395,-0.2970072329,-0.242259264,0.0927130431,0.4133452475,-0.1866027266,0.0552480817,0.0985620245,-0.1663855016,0.0215987209,-0.1545245498,-0.0231146105,0.2027694881,-0.3458365202,0.290559262,-0.3137407601,-0.028523149,-0.4044006765,0.0856050551,-0.1631050408,0.1187716722,-0.0330521092,-0.2274551094,-0.2630125284,-0.2444104254,-0.4903455079,0.3173393011,0.1552322209,0.2717688084,0.291292429,-0.0228429362,-0.2171895951,-0.0164477564,0.6867682338,-0.3689185679,-0.6013423204,0.3722738326,-0.0072731962,-0.5260193944,0.0412569568,-0.1601488739,0.1185230836,-0.1868173927,0.6114800572,-0.1121608689,-0.1128566712,0.5385574102,0.1910730898,-0.0030897555,-0.0647648051,-0.0275243968,0.146508351,-0.0947306007,-0.3211616576,-0.130618766,0.5407404304,-0.1205691993,-0.1492382139,-0.2390724272,-0.3590199351,0.113465406,0.1685439348,0.1686153561,0.323690027,-0.0559849627,0.1164250821,-0.2771925628,0.3140989542,0.2233119309,-0.1735615283,-0.4095315337,-0.0674382076,-0.1244138107,0.3769440949,-0.1998220235,0.0715272129,0.227482751,-0.2681818902,0.0985781327,-0.1385202855,0.0678735226,0.2422307432,0.0868622512,-0.4898300469,-0.4183366299,0.2966909707,0.1976842582,0.0894350037,0.4377852678,0.1015229225,-0.2574076951,-0.0000240971,-0.1051816493,0.6077110767,0.1193211973,0.0889272839,0.0494117439,-0.1816971302,0.4047291577,0.021520365,0.372857511,-0.1871741116,-0.2410685122,-0.1781828851,-0.015463572,0.1891345084,0.0828680694,-0.4538230896,0.0597165413,-0.4398391545,0.1877653897,-0.0441069044,0.096449621,-0.0235197246,-0.5262392163,-0.3846259713,0.1481364071,0.0433426723,-0.373168081,-0.06815584,0.0462673642,-0.1016042382,0.0921850428,-0.1371257007,-0.0093368581,0.1608572602,0.0250581317,0.1002049744,-0.3422575891,-0.1159635931,0.0210103504,0.0521620773,0.016562162,-0.0197315495,-0.1912325174,-0.0184391178,0.0611676425,-0.023227023,-0.1652085036,-0.0057612523,0.231161356,-0.3075413704,0.3474169075,0.1665046066,-0.3117278814,0.0602800399,0.1185574457,-0.1237676367,-0.1787308604,-0.0517202355,0.2353450358,0.2143881917,-0.2692739367,0.1917435527,0.0625466406,-0.2333200276,0.0978003517,-0.1885360926,-0.2763433754,-0.1605476588,0.5698921084,0.0657669678,0.1076139808,0.2477426827,0.1160033345,-0.0971588194,-0.1662722528,0.1173512414,0.3057340086,-0.1103130355,0.1017763242,0.1677981913,-0.110181205,-0.1801377684,0.1938293874,-0.0277388506,0.4791873395,-0.1644944996,-0.2471840978,-0.4967111647,0.1797678024,0.3701281846,0.1714560241,-0.0239216387,0.1592522562,-0.151713118,-0.0794297829,-0.360830009,-0.1363074332,-0.0717432275,0.0462656766,0.28365767,-0.0406303853,-0.068503134,-0.0753762573,0.3054109812,-0.0074987193,-0.0163174812,-0.3289112151,-0.1638776213,0.0028933471,0.1814591289,-0.2591577172,0.0272619799,-0.072944887,-0.4046710432,-0.2934731245,0.0532366969,0.4485887885,-0.2827629745,0.0978394225,-0.0520512275,-0.1574319601,-0.0596139729,-0.0769662336,0.0502068065,-0.1973648965,-0.0968634039,0.2935681641,0.0562770627,0.0049738097,-0.1948285848,0.1838008314,0.0269653965,-0.0468933098,0.0476152115,-0.0602111071,0.0672953129,0.257935971,0.3072111011,0.1623544544,0.0628555194,0.2926856279,0.2124693394,0.3209311366,-0.2774735391,-0.3693932891,0.0605046786,0.3878197372,0.4583673179,0.1108511165,-0.3141108751,-0.2479771525,0.0807526857,0.0079259714,0.0242100097,-0.0927616432,0.3725390136,0.5311440825,0.073262617,0.0134726139,0.0544654205,0.1699898988,0.1813458651,0.1594892442,-0.0579946041,0.0507518239,0.3965899944,-0.1946714222,-0.1240460053,-0.2143039405,0.0232700482,0.110735707,-0.1414573193,0.0860746354,0.1904232353,-0.1014309525,-0.0167745445,0.0004359156,-0.0118415272,0.0746932402,-0.0561026633,-0.1204990447,-0.0210177749,-0.2038307637,-0.0772635639,0.0052640233,0.1286929846,-0.283600986,0.4928397238,-0.1913950443,0.070456095,-0.150291115,0.4490296543,0.3303478658,0.066988267,0.0462754853,-0.0922060162,0.3998709619,0.0101639554,0.0721107349,0.2076397836,0.3230277002,0.1880320162,0.2334593683,0.0811585337,-0.1455555409,-0.1438239962,0.0895117894,0.006291402,0.0821914375,-0.12053895,0.123559311,0.3070297837,-0.2196787596,0.1558573842,-0.1300047189,0.0915828571,-0.5912762284,0.0805234164,0.1916843057,-0.3469381928,0.0281240828,0.0422351174,-0.0697695687,-0.2851134837,0.0619487762,0.0495807901,-0.0367397293,0.0849198848,0.1784985214,0.0829802975,0.406406194,-0.1365164816,-0.0137446737,-0.2608768642,-0.0508247204,-0.1332627535,-0.1348143816,0.0357732326,-0.186597839,-0.1200979725,0.1328771859,0.087583065,-0.0422041453,0.3020148277,0.2018318474,0.1086597666,0.2069045156,-0.1476822495,0.0071725366,0.2105503827,-0.0537947826,0.0318596177,-0.3354479074,0.3440541923,0.2687551677,0.1323976517,0.0037177175,0.1429714411,-0.1260971129,-0.184671849,0.2557794452,0.1176211387,0.1147153825,-0.2001407892,0.303607583,-0.3600517213,-0.1929202527,-0.3687105179,0.3780510724,0.1165059358,0.2029487193,-0.3130413592,-0.1715016812,0.1105526909,-0.0124347853,0.0706337318,0.2122298032,-0.1443947256,0.4930886328,0.222259298,-0.2137944549,-0.1979366541,-0.0625333935,0.0300316252,0.0159356594,0.0925036743,-0.2083307356,0.0662314966,0.0441695116,-0.4783132672,-0.252684623,-0.095286347,0.3018608093,0.0910614952,-0.0933064669,-0.099399291,0.2757215202,-0.0609045587,-0.3030002117,0.3159589171,-0.2151515335,0.0517357998,-0.0920515135,0.0167081654,-0.015443448,-0.1852019429,-0.0837773979,-0.3880162835]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"Closing this one, feel free to re-open if you have other questions about this issue","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":15,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\nClosing this one, feel free to re-open if you have other questions about this issue","embeddings":[-0.5589522719,-0.0219471473,-0.1970526576,0.1196180731,0.1910416931,0.17136693,0.4774155319,0.4065475166,-0.0855716914,0.3177537322,-0.1674809754,0.3331536353,-0.0070962594,0.4035311937,-0.5135407448,-0.137831673,0.1243322045,0.7238878012,-0.5702276826,-0.0542898513,-0.2762099504,0.0659739897,-0.3504694104,-0.2556045055,-0.5830298662,0.131641686,-0.0848130807,0.1766971946,-0.3436987102,-0.131626904,0.2923163176,0.2983741462,-0.0042471346,0.1029173136,-0.0001038613,0.0328198746,-0.0205337182,-0.2030233443,-0.1378456354,-0.3093399405,-0.5436580181,-0.2830734253,-0.0657638013,-0.0730352402,-0.0722616389,-0.3067986965,-0.005400849,-0.3397389948,0.0196795687,0.1951270401,0.2863464653,-0.0466452278,-0.1496233046,-0.2937594652,-0.0467361659,0.2329248935,-0.0415957086,-0.1943009496,-0.2115375698,0.0572937317,0.0852863714,0.2676559687,0.1399693638,-0.0069204103,-0.1093418598,0.1562287211,0.4153472781,-0.4487882853,0.2936053574,0.3532115519,0.386395067,-0.1014842689,-0.0451277457,0.0531152785,0.0682914779,-0.1235768571,0.054115396,0.199627921,0.0955230445,0.2653749585,-0.0846680328,-0.0812496841,-0.3452096283,0.2850145102,-0.1677802056,0.5754256248,0.0496565327,0.1649676412,0.078405425,-0.238799125,0.1117017716,-0.2194503099,-0.0375737175,0.2384165525,-0.4280178249,-0.1063370258,-0.3525575697,-0.2526483536,0.2456942797,0.25399369,0.0088784536,0.0345451795,-0.04490095,0.2609994113,0.1154588237,-0.0025774208,0.390106529,0.2242020667,0.0090648755,-0.2757575512,0.2019952238,0.1477575898,-0.3148320615,-0.6203756332,-0.2400444448,0.0400030985,0.0383469351,0.0928744003,-0.093299076,-0.0849352702,-0.1556033939,0.0816893131,-0.0578253381,0.0663352683,0.153515473,0.6336736083,-0.0473588221,0.2102044672,0.1695601344,-0.1221097782,-0.245687753,0.2192939967,0.11140652,-0.0545237064,0.1200805083,-0.2169322222,0.0248917192,0.2492700517,0.3037640154,0.1437592804,0.0695822537,-0.214308396,0.2221267968,0.1788457632,0.1842616647,-0.0767031983,0.1260351688,-0.1722683311,0.0109446244,0.2778958976,-0.1258340031,0.0841127783,-0.464938581,0.2669240534,-0.1407072544,-0.0174539201,-0.0666696206,0.2642068863,-0.1075211689,0.0465729386,-0.0025339779,0.0010852838,0.113712728,-0.1683018506,0.1260785758,0.0365425497,-0.2921462655,-0.0690272897,0.1387468129,-0.3543215096,0.1236032844,0.071645245,-0.0378018916,-0.3159668148,-0.0988088399,0.246797353,0.4512539506,-0.3765801489,-0.0650160909,-0.0275819488,-0.2550020218,-0.1439220607,0.2206656635,0.0859462768,0.1102736592,0.1495333165,0.1460573822,0.1284083277,0.0017528749,-0.0514392853,-0.0223947577,-0.2257310748,0.4379812181,0.1697297841,-0.0478985533,0.1202811673,0.1079323441,-0.5172353387,0.4305879176,-0.3578746617,0.2536372542,0.1325771213,0.2268005162,-0.472737819,-0.0046582292,-0.005191124,-0.4701038897,-0.0249537826,0.0217990447,0.2642014325,-0.0155351367,-0.1890500188,-0.1215959936,-0.2424745709,-0.0665803105,0.0685076788,0.2348789275,0.0683590099,-0.0159636959,-0.1704014689,-0.3112568259,0.4256345034,0.1786870658,-0.1251479238,-0.1614399254,0.2169053704,0.0350712612,0.0697263777,0.0179671999,0.2247159332,0.0988956019,-0.1492453814,0.2431645542,0.4697552025,-0.3239512444,0.0433980413,0.2475423515,-0.0106481202,-0.0067719542,-0.1096979603,-0.206324026,-0.4322458208,0.1129576415,-0.0006850107,-0.0397226438,0.553422451,-0.2631407678,0.0402467735,-0.0862194896,0.1589965522,0.276155889,-0.1324076951,-0.3190905452,-0.389878124,0.0004482642,0.3210680485,-0.0923746526,0.2119948268,-0.232667014,0.0844902918,0.3680706918,0.0589054115,-0.1171433181,-0.1730790287,-0.1670372933,0.1566119641,0.2975708544,-0.0511336587,0.2696861327,0.3809529543,-0.165123865,0.1744092852,-0.282857269,-0.0855967849,0.1393811107,-0.0380023941,0.1936631203,0.4958361089,0.2703158259,-0.1777202487,-0.2166058421,-0.2777330279,-0.02060581,0.3299668133,-0.1469975263,-0.0985574722,-0.3520730138,-0.067052722,0.0552513041,-0.276948005,-0.2293403149,-0.128987208,0.0319375619,0.0853675529,-0.047311116,0.6076834202,-0.1991473138,0.3171981573,-0.081270799,0.0575190336,-0.1464976072,0.1321217418,0.0824735984,0.2338362038,-0.1970041692,-0.103603214,-0.0059301551,-0.1984150261,0.2179051191,-0.0693602487,0.0168961771,0.0599916168,0.0037248051,-0.5207155347,0.098707132,0.0927398205,-0.2479969412,-0.0795180947,0.1201987937,0.3922540247,-0.1098785028,-0.0802016109,-0.1605359167,0.0491794124,-0.3705626726,-0.1032693684,-0.1874578893,-0.3255110085,0.0859034434,0.2596586645,0.2253918946,0.1166663542,0.5379615426,-0.1889897883,-0.0355601348,-0.0921946093,-0.2045497745,-0.4197784066,0.4666752517,-0.0234711803,-0.0600306205,0.1292587817,-0.2024140656,0.174821943,0.3789291978,0.0094471853,-0.2182095796,0.1151662767,0.242251575,-0.4077408016,0.1181117147,0.3421899378,0.4056375921,-0.1530716866,-0.1850049198,0.0224720202,0.3252284527,-0.0286486,-0.1708370596,-0.2083592564,-0.1081953049,0.1648619771,-0.1017878503,0.3574621379,-0.1978254318,0.1102277264,-0.2180105001,0.3860122263,0.2757179737,-0.0921215117,-0.0628282055,0.3671753705,0.4562272131,-0.1358759999,0.073490262,0.1227255464,-0.2883257568,0.1936931461,-0.2394684404,-0.0395241193,0.137210533,-0.2611524463,0.1319364905,-0.1881864518,-0.045652803,-0.2658253312,0.1215333119,-0.1064684093,0.0917101949,-0.0130568519,-0.097541742,-0.1772100776,-0.0742330626,-0.2656185627,0.2797128558,0.173674345,0.1610487401,0.2312579751,0.0171000641,-0.1457936466,0.0012984002,0.8527836204,-0.252381146,-0.3581684232,0.3279907703,0.1433203369,-0.2738674581,0.1198266149,-0.0900230259,0.1458140165,-0.123748444,0.6246834397,-0.1252326816,-0.1859117746,0.3782856762,-0.0403644405,-0.0000606227,0.0275629889,0.0243768543,0.0369214788,-0.2163733691,-0.3414357603,-0.0450160988,0.5182338357,-0.1622333974,-0.1460369676,-0.2046862245,-0.4217590392,0.1110610366,0.1652890891,0.2409183681,0.091536507,-0.197613731,0.1244649142,-0.0794638395,0.4462946653,0.2429394126,0.0186475534,-0.5014885664,0.0476040468,-0.1876970679,0.3297657967,0.0731339678,-0.0934568644,0.0645353422,-0.184826538,0.0052749477,0.0874954015,0.2745679021,0.0845828429,-0.0303442925,-0.5429161191,-0.1594686061,0.1743781716,0.1902106702,0.0484394357,0.3779054284,-0.1339030862,-0.295769304,-0.1375360489,-0.0433211476,0.7728338242,0.2786627114,0.1641221344,-0.0027516542,-0.0874260291,0.4298829436,0.0662602112,0.2571220696,-0.2552756965,-0.026973784,-0.228431493,-0.1518204361,0.2176896185,0.2237551361,-0.2951362133,-0.0293315817,-0.3719416559,0.1393828839,-0.0370202884,-0.0078997146,-0.0887474492,-0.4041402042,-0.2286013365,0.1418777108,-0.0915499926,-0.1919761449,-0.2513388097,-0.1222918928,-0.1409667134,-0.0585064925,-0.0433323234,0.0641729608,-0.1200496852,0.1233180165,0.1631138772,-0.3551661968,-0.2459085286,0.0840715691,0.02448502,-0.0384628959,-0.0335329734,-0.0041603339,0.140831694,0.0556917824,-0.0500035621,-0.0511449464,0.0611445047,0.0299177337,-0.2060109377,0.141185984,0.1423656046,-0.2220191807,0.0629867315,0.0876217932,-0.1823130548,0.0725938752,-0.0068297097,0.1353507936,0.1813863516,-0.1766286045,0.1613514423,0.1631725878,-0.1139628068,0.0527075417,-0.0499648489,-0.2757784724,-0.124335058,0.5621169209,0.0759358779,-0.08583875,0.0641252846,0.0783807039,-0.1039564162,-0.2599696815,0.1831183881,0.2027694881,-0.3626724482,0.1580256522,0.1460494846,-0.204631716,-0.2407395691,0.4096333086,0.076660119,0.4840569496,0.0857702494,-0.3121174872,-0.4867979288,0.0466494299,0.1051542908,0.2338466048,-0.0304240342,0.2891672254,-0.1414143592,-0.3373048306,-0.4158650041,-0.0829816163,-0.068991825,0.0320588201,-0.0863106102,0.0511662625,-0.0945731178,-0.0607229955,0.3206768632,0.0777202472,-0.1882850379,-0.3412645459,-0.0452640615,0.0643922836,0.0696268305,-0.352391541,0.0537991039,0.0563686006,-0.4233012497,-0.2111660093,0.0516348332,0.4079569578,-0.2178716958,0.0210301615,-0.0309754945,-0.0909365863,0.1186757013,-0.2260744423,0.1114635915,-0.0965624228,-0.1132600307,0.236212045,-0.1043361053,-0.0872695968,0.0590554178,0.1115730554,0.1316951066,0.0807865709,-0.0860703066,-0.1471372396,-0.069883801,0.0771827176,0.389893949,0.051099021,-0.0929351449,0.1949859262,0.1221033856,0.3612775505,-0.2071386278,-0.2619972527,0.1662655026,0.4146547318,0.3261677921,0.0224062987,-0.1987842619,-0.242963329,0.0958135054,0.0988378227,0.255743593,-0.1724435985,0.3138059378,0.2850887477,-0.0771850795,0.1416695714,0.0095632896,0.3490308821,0.392803967,0.1610404998,-0.1267355233,0.1214735284,0.4473414421,-0.1725574285,-0.136803627,-0.185215801,0.0226264615,0.2671969831,-0.169110328,0.1264968812,0.3146825135,-0.1205715239,-0.0326393545,-0.058796335,0.1113173589,0.1374451816,-0.1187968999,-0.0806644484,-0.2819737196,-0.2665122449,-0.1829949617,-0.0002764633,0.1864655912,-0.2009147853,0.5084832311,-0.1087752655,-0.0108050909,-0.1695479155,0.4768933952,0.213911131,0.0462682247,0.0897573233,0.0441994146,0.4171486497,-0.0904969871,0.3240760565,0.2699109018,0.3835984766,0.1813391149,0.2432879061,-0.0927135795,-0.1920068115,-0.2198747098,-0.0751480237,-0.045415666,0.2287876308,-0.1665354222,0.2545788586,0.3654244244,-0.1956764758,0.2480123639,-0.0455141924,0.0205709804,-0.4085699618,-0.1045295596,-0.0794483274,-0.2451295853,0.0997890681,-0.1314628571,-0.0049884985,-0.3538551033,0.0012929819,0.0803651884,0.0101848729,-0.0372269824,0.1542009711,0.0803024098,0.2464262992,-0.1792532802,0.1213413626,-0.1778199524,-0.2104966789,-0.1651737839,0.0662283674,0.1760769188,-0.1439035088,-0.2122052312,0.1498461813,0.193024531,0.0051113446,0.4120056331,0.3153792918,0.3690012097,0.1395631731,-0.2414654195,0.1462325752,0.271751523,-0.1465563476,0.1161292493,-0.2902487814,0.2951092124,0.2362074852,0.2193357348,-0.0274537504,0.0364617035,-0.0838651657,-0.1561774164,0.2165358067,0.2623246908,0.2008309662,-0.3044827878,0.2970706224,-0.1473299563,-0.1749085486,-0.3200679719,0.2374628931,0.1388292462,0.3039537668,-0.3158653378,-0.1155157164,0.1076135337,-0.060958907,0.1168887392,0.3355900943,-0.2001686543,0.5600337386,0.1067360938,-0.1790841818,-0.2906063795,-0.166596815,0.0668034256,-0.0549455397,-0.0609131232,-0.2927676737,0.3036876023,0.1394965649,-0.5392445326,-0.2155703306,0.0170973763,0.3193677068,0.1466684788,-0.1096066087,-0.0866847783,0.2998965979,0.0433502123,-0.2628442943,0.2205141038,-0.2223234326,-0.0054158275,-0.0232392345,0.0322750136,0.069322288,-0.2502540648,-0.0729132295,-0.4435074329]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"That's not really difficult to add, though, no?\r\nI think it can be done without copy.\r\nMaybe let's add it to the roadmap?","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":23,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\nThat's not really difficult to add, though, no?\r\nI think it can be done without copy.\r\nMaybe let's add it to the roadmap?","embeddings":[-0.5632335544,0.1566246152,-0.1052774861,-0.0830112472,0.0874820575,0.1624974906,0.3028231859,0.4194337726,-0.224942416,0.3868966103,-0.1537489593,0.2858173251,-0.1336852908,0.4265962541,-0.3774040043,-0.1137892231,0.1690777391,0.6831735969,-0.4871610105,-0.0461216047,-0.2769603431,0.0618332699,-0.1913713217,-0.3543604612,-0.5695503354,0.2912069857,-0.1852046996,0.1471563578,-0.3962182403,-0.1970464885,0.2927441597,0.3368051052,-0.0439014137,0.1087513044,-0.0001123584,0.0200279392,0.0133769969,-0.158865124,-0.2239809185,-0.2859130204,-0.4628483057,-0.3097318411,-0.0573883057,-0.0591023043,0.1034685671,-0.1293453574,0.0893115848,-0.23392196,0.0351497196,0.1296929866,0.1886484325,-0.0460463874,-0.1494620144,-0.3029183149,0.0352646708,0.3872572184,-0.1596915424,-0.1119441763,0.036842376,0.0510122664,0.1473965943,0.1967135519,0.0847492069,-0.1048685014,-0.0312238317,0.127853781,0.2851846814,-0.5738030672,0.1894932836,0.4373262525,0.3858180642,-0.2070233077,-0.2310138941,-0.106377244,0.1256219,-0.0378942862,-0.0085131256,0.2486876398,0.0003842789,0.2998109758,-0.1266189814,-0.3214491308,-0.3337571323,0.219253853,0.0035825574,0.5095900893,0.1821104288,0.1283189505,0.247975871,-0.0896193013,0.1514654011,-0.1121116281,-0.1872598082,0.3319087327,-0.3466578126,-0.1094931662,-0.42353037,-0.3127010465,0.3242367506,0.3647496104,0.2044358253,-0.0514718108,-0.2810693681,0.2224729955,0.1368426979,-0.0281804111,0.2627055049,0.2246279866,0.1004936025,-0.4747583866,0.4068492949,0.1419351399,-0.2541023493,-0.6352835894,-0.3118085861,-0.0495493114,0.0398108363,0.1275728941,-0.0244059376,-0.2763108909,-0.0880281478,0.0547675602,-0.1994536072,-0.0870600417,0.1845135838,0.5468416214,-0.0291948318,0.1838525087,0.1613679528,-0.2034956664,-0.1228639185,0.1137353703,0.2657358646,0.1090382114,0.0932540372,-0.3103024662,-0.0534344316,0.101030387,0.4630501568,0.2935289145,0.0899410695,-0.112579301,0.1572803408,0.2680854797,0.1508649588,-0.1060420647,0.0711883977,-0.2159586251,-0.1277613044,0.3148569763,-0.2388080806,-0.078952089,-0.6027963758,0.1250679642,-0.1908835173,-0.1287215203,-0.2735530734,0.401488781,-0.108865045,0.0152949858,0.1486211866,0.1504068524,0.1207897589,-0.1274710149,0.1286760271,0.1386974752,-0.2228637338,-0.125365898,0.1494992673,-0.3760162294,-0.0257564671,-0.0534549877,-0.0917768329,-0.3919664025,-0.1096220985,0.238018766,0.5789764524,-0.3724956214,-0.0461445339,-0.0229790378,-0.3579958677,-0.1780030578,0.3134596646,0.2745227218,0.182815522,0.0905676335,0.0883903801,0.285232842,-0.1022025645,0.0373513959,0.0420043692,-0.2569622695,0.2918095887,0.2295566052,-0.0237759054,0.1568799764,0.2578094304,-0.5314164162,0.2531365454,-0.4744189382,0.1922059208,0.065575771,0.4364443719,-0.458805263,-0.027983278,0.0510564335,-0.4954914749,0.0699076429,0.1260803938,0.2886126041,0.1175459027,-0.3634769917,0.0863496736,-0.1281976551,-0.0332222879,0.0490106083,0.0638270751,-0.0172972493,-0.0022057504,-0.095239453,-0.3850867748,0.4718657434,0.1025447175,-0.0925958082,-0.1465745121,0.2257517725,0.0452562012,0.122484073,-0.0722053275,0.2096060961,-0.0679520518,-0.2650655806,0.2195994556,0.5704522729,-0.4314958155,0.1467092335,0.2623591721,0.1093144417,0.266836971,-0.1554703563,-0.1848710626,-0.5273680687,-0.0256932471,-0.0429723822,-0.0176549312,0.5327882767,-0.3147445917,0.030716097,0.0439040363,0.2274836302,0.3138510883,-0.0138027193,-0.3958470523,-0.3787468076,-0.0729582459,0.4171087444,0.0160726439,0.1849319935,-0.189302817,0.1207722872,0.344766587,0.0078030094,-0.0184842981,-0.0078622419,-0.1470701247,0.1143816933,0.2931669652,0.027650645,0.3205238283,0.3793960214,-0.1209535748,0.1799788475,-0.2484185994,-0.0468228161,0.1409377009,-0.0139968572,0.2253265232,0.3300787807,0.3414494395,-0.0892062932,-0.2700644135,-0.3984596431,0.0046033515,0.2516726255,-0.1310018152,-0.2252396643,-0.2523091435,0.0098989401,0.0092809079,-0.2826526165,-0.241720289,-0.2430527955,0.128206104,0.1327527463,-0.0508079194,0.570241034,-0.2493341863,0.5388227105,-0.2218421847,-0.0355609357,-0.1266869754,0.1360903531,0.2055342644,0.0919156224,-0.1181469411,-0.3078886271,0.096773915,-0.1567498296,0.2157622278,-0.2193115652,-0.2720019221,0.092466116,0.0982192457,-0.5494916439,0.1595659405,0.0025911804,-0.1303557754,-0.0801043808,0.0279196408,0.3592705131,-0.2882347405,-0.003315567,-0.2200141698,0.0260150228,-0.4219441712,0.0597501583,-0.2398126274,-0.3552928865,0.2629978359,0.3009651601,0.2693175077,0.1727238595,0.5507326722,-0.1992111206,-0.0610369667,-0.1994949132,-0.0663796365,-0.4897853136,0.5394616127,-0.0980456769,0.0153267551,0.0874516964,-0.1252419502,0.0777762309,0.4468548298,-0.0799492449,-0.3376495242,0.1285665631,0.3639609218,-0.3346345723,0.1178630292,0.4186065793,0.4637767375,-0.047855366,-0.1099586934,0.1922911853,0.2692982852,0.1375145763,-0.2730713785,-0.2173589319,-0.1211927906,0.3359744847,0.0764826462,0.2960463166,-0.0274231024,0.2046491653,-0.2270864844,0.5847343206,0.2745037675,-0.0230415035,-0.001479847,0.3178342879,0.3991067111,-0.2302646339,0.1185982376,0.0865187049,-0.343840301,0.0654726923,-0.0959149152,-0.0179026145,0.175866589,-0.2882618606,0.173236385,-0.275051713,-0.1769231409,-0.3953577876,0.0536460243,-0.0832129419,0.1600789726,0.0639592037,-0.1055450812,-0.2958188057,-0.1693474501,-0.3520299494,0.2836629748,0.2666339576,0.0805597901,0.3079217672,-0.0387127697,-0.2518890798,-0.0421437249,0.8338670731,-0.3788590133,-0.4378479421,0.224266082,0.1127663329,-0.1577361971,0.097095795,0.1057941541,0.1632111222,-0.0759411976,0.596118331,-0.1054953188,-0.1591183543,0.5606845617,-0.0316053405,-0.0854043588,0.0484694429,0.027789576,0.0359925516,-0.196729973,-0.4320863783,-0.0555669032,0.3386975229,-0.1358344853,-0.1920320839,-0.307298094,-0.3666205704,0.0244276188,0.2055243403,0.179495126,0.0950459242,-0.0820064098,0.0414194316,-0.2957623601,0.5316928029,0.3902170658,-0.1822921634,-0.5838048458,0.0246762875,-0.1763226241,0.4705657959,0.1444323808,0.0501422919,0.0679472461,-0.1717259735,-0.0821689442,-0.0484124422,0.1831500232,0.0039011771,-0.1674989164,-0.5671281815,-0.1095215827,0.3092727065,0.278362155,0.0711731538,0.4819407761,-0.0583449677,-0.3044939637,-0.157521233,0.0009393194,0.885058701,0.1260218471,0.3311336339,-0.2262195498,0.0032261407,0.5197994113,0.0478335805,0.2047701776,-0.1481336951,-0.1241938695,-0.2646896839,-0.1712024361,0.2445557863,0.2108684629,-0.3462064564,-0.0158087388,-0.1570598185,0.2430428565,-0.0885988995,0.1055435836,-0.1538318247,-0.4903747141,-0.0216167811,0.0488249995,-0.0245370995,-0.3043811619,-0.1461356431,-0.1761859804,-0.1279936731,-0.0854801014,-0.1804280281,0.0423263535,-0.1056613475,0.0711872727,0.3174977005,-0.3614898026,-0.254360348,0.0293074977,0.1480719298,-0.0064477017,-0.0709308535,0.0422564708,0.062302459,0.1289054602,0.0310269184,-0.0537451692,0.1355312318,0.1384451091,-0.0861510187,0.2090974152,0.1216172576,-0.4662896395,-0.0472945794,0.0197886229,-0.1885576993,0.3184712827,0.0496443845,0.2659671903,0.0081745721,-0.2457351983,0.0697664469,0.161101535,-0.1090481281,-0.0136900218,-0.0945163816,-0.2226863205,-0.1102864966,0.5008261204,-0.0163830817,-0.0746424124,0.0070938431,-0.0292049889,-0.0497542918,-0.1372649223,0.1912911236,0.1772451252,-0.4085436165,0.0378795154,0.0131661454,-0.32524845,-0.2265711874,0.4109407365,0.1060938612,0.4780811071,-0.0085215308,-0.2397455722,-0.3905131221,0.0692639798,0.1329522282,0.3109382987,-0.0479527228,0.3263036311,-0.1558870971,-0.2524098456,-0.3186053634,-0.1857459545,0.050137084,-0.0410273187,-0.0346874408,0.1839322746,-0.1158590764,-0.1639681756,0.2049784213,0.1577210426,-0.0985302478,-0.1968848556,-0.1340304762,0.1485393643,0.1073062271,-0.284013778,0.1500672996,-0.0432475246,-0.4669212401,-0.2316419184,0.0755483806,0.3219165504,-0.2924268544,-0.0770319775,-0.0338765979,-0.1470321119,0.2508573234,-0.1052213684,0.1271600872,0.2297836393,-0.1485428661,0.3913184106,-0.0452922247,-0.0981235653,0.081267342,0.162668258,0.1188078597,0.1618898809,-0.1216425598,-0.1106599569,-0.0631339476,0.055422686,0.4006836116,0.0205759779,-0.1196003482,0.261610508,0.2204762101,0.2098541558,-0.1823874265,-0.3811215162,0.1406416893,0.2541769743,0.342220366,0.161204353,-0.0517608337,-0.0710424632,0.2106883079,0.0698185191,0.2964717746,-0.2202169597,0.3369923532,0.2586953044,0.0071606869,0.1303485483,-0.0890477672,0.3342763186,0.4624047279,-0.0069994656,-0.2291438878,0.1734943539,0.4945875406,-0.1835909486,0.1120678782,-0.1738142818,0.0264861565,0.2452165782,0.0525414608,0.3209815323,0.3081978261,-0.1145000383,-0.0815669894,-0.0159730781,0.0095273526,0.1592116803,0.0778552592,-0.0503283031,-0.2130485624,-0.2653894424,-0.2309087217,0.0890113413,0.1550233215,-0.3399500549,0.5834512711,-0.0975017026,0.1491421461,-0.1184301153,0.5540007353,0.1860547066,0.0896638259,0.0470757596,0.0783339515,0.3469671905,-0.2284468412,0.2433462143,0.4025321603,0.2854547799,0.2740370333,0.2811916471,0.0306160189,-0.1462983787,-0.2365971953,-0.0283689126,-0.1163258478,0.1971512139,-0.301626116,0.2150505632,0.2173686624,-0.0716040954,0.2614808977,-0.0788992643,-0.0456440262,-0.4272860289,-0.0193385649,0.0338472761,-0.2840410769,0.1452043205,-0.0406887159,-0.0088728983,-0.3922979534,0.116033569,-0.0790721998,-0.0983598232,0.0247813128,0.0787742287,0.2413949519,0.3060618341,-0.1230449229,0.1278811395,-0.0186882261,-0.1987858862,-0.2602223754,0.0422803462,0.0991177037,-0.1305374205,-0.1737665385,0.1883452535,0.2380334884,0.0759845003,0.4585739672,0.2235715091,0.3961382508,0.16712071,-0.1709620506,0.0779600814,0.4045486748,-0.2439547032,0.0588640235,-0.2913271487,0.3434559405,0.2600004375,0.0656880811,-0.0603650957,0.0335985385,-0.1045143381,-0.1272336841,0.265838027,0.1440152079,0.0529984981,-0.3452483714,0.1761558354,-0.1152969822,-0.2272214592,-0.4654476345,0.1702618301,0.1033772528,0.1734022349,-0.4425382018,-0.2223483771,0.0486699082,-0.1315545291,0.1597049087,0.1992393583,-0.3781602681,0.5593159795,-0.0075792279,-0.1205114499,-0.2674332857,-0.2030246258,0.1023359075,-0.0259287041,0.0333861448,-0.2070146799,0.1993539929,0.0469621383,-0.2782863975,-0.1676057875,-0.0560256168,0.419996053,0.2059347928,-0.1958507448,-0.0618317649,0.3006953597,0.1444235444,-0.2298732251,0.4396008849,-0.1456954032,-0.075433515,-0.1232633665,0.043304082,-0.0170352682,-0.2158042341,-0.0351694264,-0.516391933]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"Actually it's doable but requires to update the `Dataset._data_files` schema to support this.\r\nI'm re-opening this since we may want to add this in the future","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":26,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\nActually it's doable but requires to update the `Dataset._data_files` schema to support this.\r\nI'm re-opening this since we may want to add this in the future","embeddings":[-0.6200821996,0.1132390276,-0.1497742683,0.1142357141,0.1257235706,0.203243047,0.3956650198,0.3959012628,-0.003955096,0.2456055582,-0.2038999349,0.382650435,-0.0755913034,0.5006220341,-0.4717248976,-0.0797574967,0.1835412532,0.7113130093,-0.5424261093,-0.0117648598,-0.2366658449,0.0113508236,-0.339661628,-0.3187690377,-0.5360070467,0.231191203,-0.0610606186,0.0987927616,-0.3007909656,-0.2530787289,0.3102920949,0.3644462526,0.0571097322,0.1387692094,-0.0001069398,0.0727541521,-0.0874109194,-0.1837273687,-0.268895328,-0.2693293095,-0.5620741248,-0.2944664955,-0.0757344291,-0.0576736629,-0.0406845585,-0.3894323409,0.0240067728,-0.3615780473,-0.0405684486,0.1573858559,0.2330461591,-0.019054044,-0.1294506192,-0.2183726579,-0.0201024469,0.318313092,-0.0768049359,-0.1730319262,-0.1243906766,-0.0206792969,-0.0104138469,0.1871278286,0.1481388658,-0.1268914491,0.0444814377,0.1418748647,0.3391287923,-0.4659833014,0.1633160859,0.4161893725,0.465918988,-0.1859806478,-0.2017848194,-0.0357140675,0.0992443562,-0.1593410969,0.063998051,0.2853879035,0.1368727088,0.3058355749,-0.0108416639,-0.1034041122,-0.3154926598,0.2407753915,-0.2459397018,0.5976780057,0.1353767961,0.092905879,0.0829346702,-0.2302642018,0.1642215401,-0.1872150451,-0.1841837913,0.3217138648,-0.541516304,-0.1199429631,-0.3733865917,-0.1350853443,0.3298960924,0.2463191748,0.2044765651,0.0405509956,-0.1924936622,0.2452629358,0.2398508042,0.0245794728,0.4051347375,0.1368649751,0.0434726961,-0.2887252569,0.219042778,0.0950889662,-0.3077245653,-0.6520453095,-0.3247913718,0.0252014529,0.0389828049,0.1931320876,-0.1030317545,-0.143552348,-0.0916431695,0.1084327847,0.0559918582,-0.0860122368,0.2221773416,0.7616563439,-0.0534542948,0.192138344,0.1734066308,-0.2751089931,-0.1755377501,0.1871213317,0.1160215512,0.0174796227,0.1904693991,-0.1842016429,-0.0449708812,0.314483881,0.33914572,0.1827120632,0.1533740461,-0.2209471911,0.2149470448,0.2782842815,0.0714068413,-0.0948672742,0.0562412553,-0.168221429,-0.014185952,0.3145468235,-0.1227612421,0.0353895612,-0.4891705513,0.2089896053,-0.200812608,-0.0463645123,-0.2499530911,0.3336359859,-0.1040295139,-0.0500911064,0.0291792974,-0.0175051223,0.0748191103,-0.1127745807,0.1682960838,0.106441766,-0.2814204097,-0.0691355243,0.1847222745,-0.4150511324,-0.0711692423,-0.0134752281,-0.0360353626,-0.3152643144,-0.1420895606,0.2441543192,0.4369239807,-0.4175728261,-0.0972495824,-0.0296995398,-0.2238773406,-0.1285273433,0.2407663614,0.1232236624,0.1626571566,0.1246631965,0.1226341277,0.1622515619,0.0343255587,0.0087945648,-0.0435843021,-0.2390632927,0.3819728494,0.1639908552,-0.0783886909,0.1224677637,0.1311313808,-0.4837779701,0.4867559373,-0.3978932798,0.2967704833,0.1409301013,0.2017504126,-0.4265710413,-0.0140514541,-0.0838253051,-0.5214062333,0.0832028612,0.0862970725,0.2248170972,0.0663842112,-0.2890563011,0.0348978378,-0.2273228317,-0.0602604412,0.0564953536,0.2055269331,-0.004535391,-0.1216683313,-0.1537702531,-0.3518807292,0.376627475,0.1587485224,-0.1154476553,-0.117099151,0.2699806392,0.1768978089,0.13533023,-0.0552103892,0.2202481329,-0.0084873801,-0.1499313563,0.2518579066,0.5035156012,-0.2917832732,0.0624000318,0.2550005317,0.207112357,0.0982297361,0.0081180157,-0.1666467339,-0.4425747097,0.0584744625,0.0448831022,-0.1063391268,0.6201360226,-0.3699854612,0.0314688012,-0.014405868,0.1793006361,0.3834830523,-0.0434765667,-0.4043259323,-0.3500224948,-0.0427922569,0.3511182666,-0.1354760677,0.2653811872,-0.1887793243,0.110713914,0.2804255486,0.0704487041,-0.0874780715,-0.2053280771,-0.2089479268,0.127779603,0.3520625532,0.1063212454,0.3652781844,0.3536134362,-0.1055600941,0.160537228,-0.2328376472,-0.0367885716,0.1600503922,-0.1095190272,0.1294082999,0.5694359541,0.3345155418,-0.1547619253,-0.2027948201,-0.3446968496,0.054579135,0.3210361898,-0.1644705981,-0.1237167418,-0.3376923203,-0.0289667007,-0.0205559023,-0.2282090187,-0.1973005384,-0.1747505516,0.0097843781,0.120202966,-0.0833294019,0.5366846323,-0.2834307849,0.3671422899,-0.1475298107,0.0455883108,-0.0810189918,0.1155287623,0.1502861232,0.1537218839,-0.0949315578,-0.0600872748,-0.0542418137,-0.1520798057,0.1689743996,-0.2086485624,-0.0228293501,0.0750807822,0.0956626236,-0.4917036295,0.1373925656,0.0581366643,-0.1652506292,-0.031514287,0.1185321361,0.4089565873,-0.1036080047,-0.0026436097,-0.1406343281,0.1124781594,-0.3971733451,-0.0300063509,-0.2158603221,-0.3724113405,0.2106262147,0.1946964711,0.2574153244,0.132876575,0.4780389369,-0.1940965056,-0.0909743607,-0.1697713584,-0.03095611,-0.4245989323,0.4616805613,-0.0298407376,-0.0371070504,0.1607216746,-0.1619771123,0.1664905548,0.4515373707,0.0763074011,-0.2016095966,0.1102224067,0.237504676,-0.3646521568,0.0101901228,0.3265556097,0.3919829726,-0.0877716616,-0.1209307536,0.1506007016,0.3027242124,0.0087928418,-0.2549477518,-0.2459114194,-0.187236473,0.2631952465,-0.1249638051,0.3277338743,-0.1082651168,0.1255526096,-0.1994958669,0.4069361687,0.3109946549,-0.0276982412,-0.1179938987,0.2570595741,0.3770459294,-0.0890035331,0.1365223825,0.1633624583,-0.2239509076,0.082133159,-0.1582340896,-0.1031018719,0.1620392501,-0.3506244719,0.2318838537,-0.2672050297,-0.1196489632,-0.2770112157,0.1028707698,-0.1344162226,0.2083797455,0.0226844773,-0.1044847667,-0.2123042941,-0.1270283312,-0.2655828297,0.2473722696,0.1132121533,0.2012472153,0.3306742907,-0.0520230085,-0.2196616083,-0.0646101013,0.8043146729,-0.1964742541,-0.2649796605,0.2764642537,0.1832418442,-0.2816002071,0.0588716418,-0.0003236154,0.1411295831,-0.1500121057,0.6463127136,-0.1053609774,-0.2349709868,0.3337987065,0.0356731936,-0.0269500986,0.0496501662,0.0313549079,0.0973667502,-0.1070194393,-0.3596585989,0.0260892436,0.4266490936,-0.2164449543,-0.1124901101,-0.2984729409,-0.3305152059,0.0920702964,0.2017970979,0.2636724114,0.0238040462,-0.1854524314,0.0843251869,-0.1085094139,0.5909478068,0.2734281719,-0.0902699083,-0.5555426478,0.0161080863,-0.2124795914,0.3017392457,0.0445831269,-0.0492893234,0.0913477838,-0.2318295836,-0.0863752961,-0.0627549291,0.2212533653,0.0731515139,-0.0394286737,-0.5384452343,-0.1915698946,0.2679322958,0.2278105766,-0.0005602105,0.3381001353,-0.0664309338,-0.3769412041,-0.0578928888,0.0016405585,0.7042246461,0.196083039,0.2802120447,-0.0452646576,-0.0734535232,0.4230068624,-0.1159614474,0.3017453253,-0.322917521,-0.1712749004,-0.233822003,-0.1116468608,0.2787314057,0.318213433,-0.309191674,-0.0227297153,-0.3546568453,0.2063690275,-0.0992080197,0.0684022829,-0.1414733231,-0.3997264802,-0.2292437553,0.093290858,-0.0462397374,-0.3588957489,-0.2243019044,-0.1141460314,-0.0944366306,0.0363028646,-0.1164585128,-0.016697837,-0.0416434593,0.0535579063,0.1420464665,-0.2721469104,-0.1693543792,0.1830169261,-0.0191532131,-0.0161071122,-0.0831818506,-0.0362561904,0.0723044202,0.1616737545,-0.0199310891,-0.0701218173,0.0358036682,0.1406971961,-0.1232453063,0.084394671,0.1165631711,-0.222160995,0.0868036523,0.130158782,-0.2665665746,0.0771867335,0.0197352208,0.0916165784,0.1281281859,-0.1698982418,0.1342815161,0.167147398,-0.1286481619,-0.0165018868,-0.1649949402,-0.2615962029,-0.1177315488,0.608930409,0.0195696466,-0.0570774786,0.0621193834,0.0503900312,-0.1109855026,-0.229891181,0.2489146739,0.2228277773,-0.3753678501,0.154132098,0.1678534597,-0.2025003582,-0.1807569116,0.3313976228,0.0607091486,0.466770947,0.0213509519,-0.3434015214,-0.3806184828,0.1042640433,0.1155208796,0.2782459855,-0.0278321076,0.2225751281,-0.1813865155,-0.3636708558,-0.3565867245,-0.0849390402,-0.0003999163,-0.0124318404,0.0026943991,0.1050576866,-0.1492236704,-0.0949683487,0.2796123326,0.0400895365,-0.1153401136,-0.2759155631,-0.0686293468,0.0815151706,0.1414373368,-0.2460020036,0.1420935541,0.0204804782,-0.4042542875,-0.3431597352,0.0698396564,0.4484346509,-0.2318343371,0.1061019227,0.0223233644,-0.0160909146,0.0974606499,-0.1148715839,0.1849149913,-0.1200097725,-0.1341986507,0.3616339266,-0.0978333429,-0.0388589762,-0.0022234782,0.1870062798,0.0968815684,0.0798408017,-0.0928237289,-0.0740216151,-0.0172479935,0.1759510785,0.4340113997,0.1533678472,-0.0640267581,0.2420134544,0.1094633117,0.3031440973,-0.2127266973,-0.3532363176,0.2220484912,0.3588953912,0.3141605854,0.0558181889,-0.0908034444,-0.2506698072,0.0678107962,0.0501982905,0.2047774643,-0.1698725671,0.3834694922,0.43582353,-0.0294311866,0.1261448115,-0.0136492234,0.2723965049,0.4389003217,0.3103128374,-0.114363052,0.1330091804,0.5196546912,-0.2122596949,-0.178476572,-0.2729001641,-0.0640007183,0.2034423053,-0.0655438229,0.0750014484,0.2460419536,-0.1240104437,-0.0699607655,-0.0658298954,0.0619355291,0.0639645159,-0.0822219476,-0.1166552976,-0.2260518968,-0.2981688082,-0.1549359262,-0.0022589287,0.0914168507,-0.1644561589,0.40760988,-0.1368981898,0.0725946352,-0.1749768555,0.5676493049,0.3037233949,0.0145088006,0.0808504373,-0.0271641538,0.4270546436,-0.0663243905,0.307372272,0.2212711722,0.2742750347,0.2930245697,0.2621847987,-0.1128536984,-0.2860121727,-0.1186675802,-0.1190197319,-0.1276032627,0.1903668791,-0.2001357526,0.2976948917,0.284093231,-0.136968255,0.19182311,-0.1167658344,0.0572090186,-0.4939636588,-0.0611628182,0.0196014661,-0.3098632693,0.1819347292,-0.0679013953,0.051944647,-0.4709716737,0.0641248673,0.1147002727,0.0574925132,0.0310806185,0.1373212636,0.1497326195,0.2665427327,-0.1925782561,0.0749784335,-0.1077983975,-0.1571920514,-0.3030640781,-0.0121721001,0.1811234653,-0.1273795813,-0.2198212892,0.1420223862,0.1640564054,-0.0186317191,0.3866995573,0.3028850257,0.4080854356,0.1749006063,-0.206069544,0.0816563591,0.2137496173,-0.1575513035,0.0261627547,-0.3309252858,0.267406553,0.3030881286,0.1588629335,0.0156217068,0.0574570373,-0.037379086,-0.2359866351,0.1922109574,0.1937492192,0.1685361713,-0.2342081219,0.2491455674,-0.2890502512,-0.164546296,-0.4667681158,0.2209827155,0.0836230665,0.2395302653,-0.3569457531,-0.1498156339,0.081729576,-0.1276066899,0.0964068174,0.3045399785,-0.3017326891,0.5580824018,0.1114962101,-0.1556446105,-0.2687898874,-0.1030099466,0.0534994565,-0.0644027591,0.0059632305,-0.1904580593,0.1649006605,0.0615110956,-0.3933854699,-0.2425923496,0.0094971191,0.2682802379,0.1611192822,-0.0859815478,-0.1158311963,0.300006628,0.0728874058,-0.3006667197,0.1787554175,-0.2094936967,-0.0793691874,-0.1268153787,0.0100359786,0.011526702,-0.251770705,-0.1031510457,-0.4564321339]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"Hi @lhoestq, I would love to help and add this feature if still needed. My plan is to add an axis variable in the `concatenate_datasets` function in `arrow_dataset.py` and when that is set to 1 concatenate columns instead of rows. ","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":40,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\nHi @lhoestq, I would love to help and add this feature if still needed. My plan is to add an axis variable in the `concatenate_datasets` function in `arrow_dataset.py` and when that is set to 1 concatenate columns instead of rows. ","embeddings":[-0.3484048843,-0.0351736769,-0.1219938248,0.0604969226,0.1725979745,0.091855064,0.3840520084,0.3609426618,0.0169334374,0.3619816005,0.003005679,0.4973300993,-0.1567098498,0.5172737837,-0.3565698862,-0.1661897302,0.1776000559,0.7238048315,-0.4239696562,0.0119400825,-0.2513479292,0.0003018853,-0.4031876624,-0.1789040715,-0.3692486584,0.2398511916,-0.0626367256,-0.1479697078,-0.1530625671,-0.3112292886,0.3284269869,0.0878706872,0.1952238232,0.01228886,-0.0001191998,-0.0108588291,0.0415536053,-0.1125788316,-0.2453662753,-0.3221568167,-0.4053148627,-0.4979775846,0.2271569669,-0.0378433801,-0.0390658863,-0.5209636092,-0.1310438812,-0.4387720525,-0.0303994082,0.2519806623,0.1772928685,-0.0298096929,-0.0708883554,-0.0822864547,0.0069344579,0.2928192317,-0.2658262253,-0.2799834013,-0.1053103581,-0.1982269287,0.12269485,0.2715041935,0.115182139,-0.0590899698,0.1243962422,0.1927006692,0.4056869149,-0.5003891587,0.1288584173,0.3248573542,0.5749339461,-0.2733677328,-0.1551249325,-0.0996268541,0.4653525054,-0.0741728991,-0.073067084,0.1629952937,0.0500215851,0.1906056404,0.135131225,-0.1057225838,-0.382907927,0.3393296897,-0.0798391178,0.5351185203,0.0546734594,0.1881549805,0.2290105671,-0.1005355939,0.2734649181,-0.0220015496,-0.0882297456,0.4876595736,-0.7353643179,-0.0565521345,-0.184257701,-0.1495024711,0.0795142502,0.2421560585,0.2245164514,0.092650868,0.021393111,0.0912516937,0.2350979298,0.2246278971,0.3509427607,0.3787029982,-0.0325114056,-0.243798539,0.3597438931,0.0425394922,-0.3643557131,-0.5424804091,0.0170116648,0.0181894172,0.2261976153,0.1226506233,0.0239369348,-0.2209314406,-0.1189647689,0.0870715529,0.0508961566,0.0251886342,0.1460366696,0.9004566073,0.0120824613,0.3414290249,0.2295050472,0.0137569755,-0.1139784828,0.0569169708,0.1438239515,-0.0556241572,0.2048874795,-0.0913899764,-0.2196103781,0.3884479403,0.2302772105,0.0550045744,0.1491926014,-0.2047059983,0.3219252527,0.2805392444,-0.0968061686,-0.1261657029,0.2055715173,0.0623703524,-0.0557909422,0.4711827338,-0.2314359099,-0.1000727639,-0.4651686847,0.0941773355,-0.104697369,0.0153276101,-0.1906225383,0.2959982157,-0.185953021,-0.0435081497,0.1876229942,-0.0357849076,0.2972150445,-0.1553348303,0.1612838656,0.0049321968,-0.3409697115,-0.147139281,0.0910132527,-0.3876318634,0.0534614585,-0.2109619975,-0.1383642107,-0.3526219428,0.0315908268,0.0757282376,0.5826779008,-0.3446740508,-0.0554095171,-0.0860994086,-0.2328720838,-0.1188211739,0.2510834336,0.084835887,0.1563534141,0.3511503339,0.0567248054,0.1157439426,0.1549999714,0.0601842366,-0.0310866311,-0.2949569523,0.3611492217,0.0845942125,-0.1838752031,0.1872298419,0.1451583952,-0.6566550136,0.4483216405,-0.2873119712,0.2504684329,0.0672571734,0.2859602869,-0.3821868598,0.1760794371,0.0115003968,-0.5103507638,0.009330268,0.070922479,0.0939126238,-0.1279547662,-0.1896890551,0.0881058946,-0.0501625948,-0.0020471499,0.1332370192,0.0746651217,-0.1537557691,-0.0137527147,-0.0632727742,-0.5141731501,0.2657358348,0.0369735621,-0.0585437343,0.0747816414,0.1293304265,0.3989734054,0.0104729328,-0.050749559,0.251303196,-0.0305072051,0.002641377,0.1037912816,0.4819204807,-0.2660456598,0.0595243163,0.0792405754,0.1076504067,-0.0076668193,0.1580852568,-0.0221203119,-0.4409920573,0.124280639,0.076231204,-0.0412997752,0.6403236389,-0.3575990796,0.3346841931,-0.1100436673,0.1808681339,0.161535278,0.0231884997,-0.4258250892,-0.4088206291,-0.0126217203,0.0484391227,-0.0436074883,0.055849731,-0.1431513876,0.1709879786,0.1125739664,0.1262000054,0.0519082919,0.0012856096,-0.1522818506,0.1239795759,0.3801649511,0.1782371104,0.3607864082,0.2986896038,-0.2555094659,0.0794027448,-0.2633956671,0.0075087352,0.1795023829,-0.1009885445,0.1762393117,0.4263388216,0.3166389763,-0.247816205,-0.0401934311,-0.2858068943,-0.0806196481,0.3229048848,-0.1934489906,-0.0563695021,-0.2910402715,0.0493003875,-0.0500068516,-0.2946996689,-0.0211710576,-0.1120547205,-0.155431062,-0.0364156663,-0.0193312857,0.3741746843,-0.4930709898,0.3559175134,-0.2469777316,-0.1339384913,-0.137953341,0.0090173688,0.1622431874,0.0472500101,-0.3723032176,-0.0139805805,-0.1844158769,0.1461020708,0.3058330119,-0.0948380828,0.1029902697,0.0089045297,0.0370648578,-0.5876755714,0.1067023203,0.026941251,-0.2069617808,-0.0409313813,0.1177265719,0.2338819057,0.112556152,0.2097989172,-0.0109643955,0.1404884607,-0.2592791915,0.0035885668,-0.2987793684,-0.2682057619,0.2431948036,0.1682780981,0.2570115626,-0.0250568539,0.57996732,-0.0892570093,-0.1376332492,-0.1850079894,-0.0082032597,-0.255728662,0.485234946,0.110486947,0.102568984,0.0995438173,-0.086270079,0.0467193015,0.4133327603,0.0534076095,-0.1445586234,0.0527915396,0.2172708064,-0.5132225156,0.1616336554,0.4702358842,0.5106543899,-0.0310231987,-0.0433092937,0.1809465885,0.21718283,0.0877058432,-0.2157236636,-0.1919551492,-0.0750128403,0.0424115248,-0.1768639535,0.3741515577,0.0440558083,0.1584657133,-0.2632229924,0.4001315832,0.226540193,-0.1207538396,-0.2325463444,0.1497762054,0.435962826,-0.2546740472,0.0710949972,-0.0624023788,-0.2544453144,-0.0384765677,-0.0993881673,-0.0369387567,0.2689318657,-0.4602848291,0.3100158572,-0.3693746626,0.057233654,-0.3677945435,0.0736623704,-0.165150851,0.0835070908,-0.0003400298,-0.1691704839,-0.0650746897,-0.3578392863,-0.3593356013,0.2574385107,0.0695168674,0.3801535368,0.3835252821,-0.1553505212,-0.0204755459,-0.0836835057,0.5714021325,-0.1562053561,-0.4705401361,0.2725018263,0.2294002026,-0.3887176812,0.0548177771,-0.0952998325,0.1669649035,-0.5454449058,0.5598384738,-0.0703092888,-0.2177533209,0.4023017883,0.115829654,0.0991184935,0.0258430429,-0.0698378086,0.3210980594,-0.0830855519,-0.2661032677,-0.1717464179,0.3820981085,-0.1190672666,-0.0910445452,-0.398578465,-0.409141928,-0.0884825811,0.2239565253,0.3877208531,-0.064055711,-0.2556831241,0.2702801824,-0.3332705498,0.3518729806,0.1304448247,0.0350073613,-0.5236518383,-0.0468707494,-0.1196565405,0.4337215722,-0.0220799018,-0.2208117098,0.277023375,-0.2694500983,0.1267730296,0.0521558039,-0.1338442266,0.1665182114,-0.0097457273,-0.4315814078,-0.2191574126,0.3486344516,0.1353699118,0.0084242169,0.6973356605,0.0833866224,-0.2565927505,-0.0252843294,-0.0733765066,0.7857358456,0.3151414692,0.1289798766,0.1136041507,-0.177608192,0.3352881968,-0.0529930368,0.3411026597,-0.2741169631,0.0570811518,-0.2420927733,-0.0834379271,0.4021309912,0.1842034459,-0.3846099675,-0.0006986171,-0.4704162478,0.0434357636,-0.0910095423,-0.0349008106,-0.0610849671,-0.4337596297,-0.3539843857,0.0192390755,-0.0836376548,-0.2131679058,-0.147509709,-0.0658616945,-0.0981746241,0.0158095062,-0.1265794337,0.0224882048,-0.1242732853,0.0267803986,0.2064050436,-0.4968129098,-0.1666460782,0.0748856217,-0.195094347,-0.0776572824,-0.1020540744,-0.144298777,0.2181645483,0.1547259688,0.0675877333,-0.1068940014,-0.0902711898,0.1711313426,-0.4106583297,0.295350045,0.0785985142,-0.3563297391,0.2157660574,0.1351617277,-0.211405009,0.0175465085,-0.0476081856,0.1775506139,0.196191147,-0.1308623403,0.1118998006,0.2456721812,-0.0513118207,0.0592196137,-0.0775819793,-0.4195982218,-0.1293430328,0.5261601806,0.0439899266,0.0712044165,0.1582731307,-0.0439693518,-0.1154846922,-0.1995777488,0.2184592485,0.3122900724,-0.4885493219,0.1014291346,0.3493881822,-0.1719631404,-0.1115613952,0.2508732975,-0.100279294,0.5149899721,-0.3174563348,-0.3425139189,-0.2884968817,0.1409843713,0.261449635,0.1416516155,-0.1929025352,0.0557606108,-0.0528504997,-0.3349052668,-0.2467558682,-0.2235088497,0.1084795743,0.0367374122,-0.0551151447,0.170184046,-0.0315227099,-0.1321485043,0.2110210806,-0.0566484854,0.0468339622,-0.2108976841,-0.0269656703,0.119766362,0.0379402339,-0.1967041939,-0.0256897993,-0.2114143074,-0.4215812683,-0.3005755842,-0.0577735268,0.3900688887,-0.2933142185,0.1927296519,0.0073891645,-0.1262215525,0.0584373958,-0.1333044171,0.1063779965,-0.2697783113,-0.1290733218,0.3115583956,-0.0146954702,-0.0239918977,-0.1153497621,0.1420282871,-0.0804168507,0.1508470327,-0.1009420231,-0.0178982951,0.1060809344,0.0043393341,0.2145279497,0.1985800117,0.069589816,0.1378557533,0.1196546406,0.1944483668,-0.2550897002,-0.3299837708,-0.1221532151,0.4712211192,0.2662366927,0.0889464095,-0.3071503937,-0.171468392,0.2530725598,-0.0350031592,-0.0061068647,-0.3235417306,0.3857967257,0.5572224855,0.1179555357,0.0595886335,0.039698042,0.1406227052,0.3361616135,0.1338004768,-0.1947600543,0.0111881625,0.6011144519,-0.1956798136,-0.324470669,-0.2275511175,-0.1771327406,0.0576401241,-0.3520765603,0.0259014331,0.2534103692,0.0475108325,0.0037972061,0.0382119566,0.0047176429,-0.049100887,-0.0333116166,-0.1532093138,-0.1339405775,-0.1807447225,-0.039009437,-0.1634123623,-0.0081553925,-0.0364505872,0.5291033983,-0.0122985924,0.1284839511,-0.2611061335,0.6372094154,0.4024971724,0.296676755,0.1472109705,-0.0721088424,0.2510962188,0.122637108,0.1405629814,0.100942187,0.2281176001,0.3508400619,0.1949286461,-0.0105882594,-0.1840802133,-0.1523503661,0.1153996885,-0.0542159751,0.2688622773,-0.1554080397,0.0373204798,0.2149942815,-0.1238097697,0.192307353,-0.2610242665,0.0407705382,-0.5898842812,0.2387279123,0.0548744537,-0.3407072723,0.0447938405,0.1309806108,0.1003146917,-0.2478526533,0.104454346,0.1209603772,-0.2225829065,-0.0895208046,0.0841973573,0.1150964573,0.3001424372,-0.3738722503,0.1910102516,-0.1773017496,-0.0435479544,-0.1761844754,-0.1203372106,0.0298220571,-0.2808218002,-0.0348400325,0.1566493213,0.1548455954,-0.0142836347,0.491192311,0.35643062,0.2708083689,0.1197316721,-0.1729989648,0.1104134843,0.1908469647,0.0002530392,0.011730657,-0.4289298952,0.3420016766,0.2795124948,0.0535209887,0.0645616353,0.0777075216,0.0041037784,-0.1786299497,0.4160437584,-0.0091126841,0.022723414,-0.0563520379,0.3888106048,-0.3106720746,-0.124453865,-0.2836521864,0.3805933297,0.1252901107,0.0776660666,-0.3122369647,-0.1756624281,0.0684749559,-0.1056556776,0.04105068,0.3822819293,-0.0377193503,0.5629076362,0.2832452357,-0.1616649032,-0.3342673182,-0.2283573747,0.1190135181,-0.0508765951,0.0048652058,-0.18593961,0.044282265,0.0811175033,-0.400842905,-0.0934761465,0.1942816079,0.2948791683,0.045638971,-0.3090932667,-0.1178348213,0.2098563612,0.1142397076,-0.2594255507,0.1607903242,-0.1314876825,0.126108557,-0.0699984059,0.1405727416,-0.0546757989,0.0712310597,-0.1971196979,-0.4769255519]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"Hi ! I would love to see this feature implemented as well :) Thank you for proposing your help !\r\n\r\nHere is a few things about the current implementation:\r\n- A dataset object is a wrapper of one `pyarrow.Table` that contains the data\r\n- Pyarrow offers an API that allows to transform Table objects. For example there are functions like `concat_tables`, `Table.rename_columns`, `Table.add_column` etc.\r\n\r\nTherefore adding columns from another dataset is possible thanks to the pyarrow API and in particular `Table.add_column` :) \r\n\r\nHowever this breaks some features we have regarding pickle. A dataset object can be pickled and unpickled without loading all the data in memory. It is useful for multiprocessing for example. Pickling a dataset object is possible thanks to the `Dataset._data_files` which defines the list of arrow files that will be used to form the final Table (basically all the data from each files are concatenated on axis 0).\r\n\r\nTherefore to be able to add columns to a Dataset and still be able to work with it in a multiprocessing setup, we need to extend this last aspect to be able to reconstruct a Table object from multiple arrow files that are combined in both axis 0 and 1. Currently this reconstruction mechanism only supports axis 0.\r\n\r\nI'm sure we can figure something out that enables users to add columns from another dataset while keeping the multiprocessing support.","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":230,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\nHi ! I would love to see this feature implemented as well :) Thank you for proposing your help !\r\n\r\nHere is a few things about the current implementation:\r\n- A dataset object is a wrapper of one `pyarrow.Table` that contains the data\r\n- Pyarrow offers an API that allows to transform Table objects. For example there are functions like `concat_tables`, `Table.rename_columns`, `Table.add_column` etc.\r\n\r\nTherefore adding columns from another dataset is possible thanks to the pyarrow API and in particular `Table.add_column` :) \r\n\r\nHowever this breaks some features we have regarding pickle. A dataset object can be pickled and unpickled without loading all the data in memory. It is useful for multiprocessing for example. Pickling a dataset object is possible thanks to the `Dataset._data_files` which defines the list of arrow files that will be used to form the final Table (basically all the data from each files are concatenated on axis 0).\r\n\r\nTherefore to be able to add columns to a Dataset and still be able to work with it in a multiprocessing setup, we need to extend this last aspect to be able to reconstruct a Table object from multiple arrow files that are combined in both axis 0 and 1. Currently this reconstruction mechanism only supports axis 0.\r\n\r\nI'm sure we can figure something out that enables users to add columns from another dataset while keeping the multiprocessing support.","embeddings":[-0.5330207348,0.3033456802,-0.0301639233,0.0675024018,-0.0625750944,-0.0582219549,0.2094470412,0.4255707562,-0.0745848566,0.1827805936,-0.1869088858,0.7275902033,-0.1599380672,0.5724496841,-0.2195644528,-0.2587185502,0.1644585729,0.6102408171,-0.5832628608,0.2564370036,-0.2590903342,-0.1035273075,-0.3908792138,-0.2035314292,-0.479586333,0.2016779631,-0.2737205327,-0.0490148738,-0.2796463966,-0.3253386915,0.3125083447,0.2397160828,0.1502655894,0.2005506605,-0.0001199671,0.1369438618,-0.1574162394,-0.1425992846,-0.2051132619,-0.3425643742,-0.2079406083,-0.4237650633,0.1109852642,0.0214358438,0.0509263799,-0.3994389176,0.0115839047,-0.2094816118,-0.084717989,0.0106306169,0.106538333,0.2030961365,0.1404246688,-0.0321108177,0.1800280213,0.2532159686,-0.2782604396,-0.0167417303,0.1530687958,-0.2789544463,-0.0680217445,0.0427841656,0.0780213028,-0.1596894115,-0.1154882833,0.2264118791,-0.017162431,-0.4759059548,0.00997792,0.3201736808,0.5716279149,-0.4608500004,-0.3677833676,-0.2338828146,0.2534549832,-0.2218376845,0.0001270876,0.3750638664,0.0167706646,0.2371695787,0.1775349975,-0.0572856665,-0.339548111,0.241171658,-0.0190362707,0.4345147014,0.191094175,0.2638529241,0.2380313724,0.0664346367,0.1477770507,-0.0472184569,0.0093137389,0.3843119144,-0.5604681969,-0.0574878342,-0.0918807611,-0.2331230193,0.3174954355,0.3900786936,0.1542620808,0.1787298173,-0.2172387391,0.1800462157,0.3443218768,0.0806481913,0.1206246316,0.1088648587,0.0886459723,-0.35766536,0.2269135863,0.1656297147,-0.2526624203,-0.4635510743,-0.1971271634,0.008335501,0.0876557231,0.3238953352,0.0359356739,-0.1849306822,-0.1439826936,-0.0253474452,-0.0368295535,-0.0154865012,0.0473193936,0.6687628031,-0.0877524465,0.354726851,0.0581446216,-0.2263710499,0.0241897553,0.1549757868,-0.1172148809,0.0183615871,0.2820980251,-0.2173440307,-0.1762043983,0.3253812492,0.2179025859,0.3356548846,0.1447256207,-0.3521080613,0.208059147,0.2606727183,0.0259189941,-0.2390604615,0.1745023578,-0.2554290593,-0.112465091,0.507558465,-0.0809986293,-0.1647176743,-0.5950244069,-0.0452269912,-0.2724877,0.0383551754,-0.4368534386,0.4114597142,-0.1237173229,-0.1192981228,-0.0507039875,-0.0770139992,0.1889921278,-0.3673183322,0.1948676109,0.1620839685,-0.4798499644,0.0667726472,0.2755307257,-0.2665080726,0.1613331735,-0.0931987539,-0.136908859,-0.1783189774,-0.025109157,0.4001131058,0.3169251978,-0.1422618926,-0.1286444962,-0.2277919054,-0.1653326452,0.0983732268,0.2961516678,0.1404905915,0.0227770675,0.2862357795,0.030266637,0.1991989464,0.0826271102,0.0302405972,-0.0266129151,-0.2715085745,0.4335519671,0.230537042,-0.0026214032,0.0165927131,0.2031267136,-0.8441972136,0.4641341567,-0.3334346116,0.3690649867,-0.1239213347,0.2949068248,-0.2296962589,-0.0010065569,-0.2239664048,-0.5165885687,0.2283263057,0.0090546226,0.306769222,-0.0804099813,-0.2215365618,0.233523488,-0.0020000336,-0.0799393803,0.0808485299,-0.0164996739,-0.054875873,0.169963032,-0.1656709611,-0.481654197,0.2382019907,0.1083152145,-0.0715284348,-0.0773396641,0.3713358641,0.1570248157,-0.0870870426,-0.1743693799,0.3615190685,-0.0181467645,-0.0206053741,0.3296248913,0.4635334313,-0.1377492249,-0.1679189503,0.1001984477,0.1087247431,-0.0209423788,0.0529409871,-0.027797332,-0.624504745,0.1870194972,0.0502292551,0.0945491716,0.7123307586,-0.3065618575,0.0765161961,0.0154870041,0.0544215031,0.2053551823,0.0487106331,-0.2358563393,-0.2040682584,-0.0640039369,0.4038100541,-0.1281858236,0.1696962565,-0.5598098636,0.1875688583,0.331804961,0.0576810651,0.0605530925,-0.0258674938,-0.0626168251,-0.0101345079,0.2175338119,-0.0440558381,0.4475494325,0.2890944183,-0.0274714101,0.0343553089,-0.3288681507,0.0302778147,0.1246185154,-0.1124888659,0.1410586089,0.4801807404,0.3467616439,-0.0932761654,-0.0006989819,-0.2700641155,0.1327155381,0.2761547565,-0.2611801624,-0.1646780521,-0.546102047,0.1018718928,-0.1480001807,-0.4510300159,-0.0061001205,-0.0652492046,-0.0083421925,0.1660646498,-0.1294544041,0.3992390335,-0.3121035993,0.3103449941,-0.4550032616,0.0091779148,-0.1386479139,-0.0753175393,0.0793842375,0.0230257344,0.0066539124,-0.1259336025,-0.1095997095,0.0941338688,0.1009102687,-0.1167372316,-0.0924804211,0.04725435,-0.1176140606,-0.2593384683,0.1583619267,-0.0850253329,-0.1846009344,-0.1618708521,0.0596693531,0.2899071276,-0.1702336967,0.1635434926,-0.0480611362,0.141576767,-0.2636120915,-0.1541386396,-0.3244239688,-0.2585850656,0.4762250483,0.081650205,0.2547367215,0.0297357086,0.4941406846,-0.0699383467,0.0058190618,-0.0565872081,0.0685968623,-0.2480999827,0.4400029778,-0.0790254474,-0.1334135085,0.0509739965,-0.2691562772,0.0301825721,0.3969061375,0.0643414259,-0.0442860685,-0.0039547672,0.5810075998,-0.2796831131,0.2266358882,0.4895676672,0.5526223779,0.0019504568,-0.0365976393,-0.0019033587,0.1531602591,0.214816317,-0.3528675735,-0.0444393679,-0.20587641,0.2725043595,0.223283112,0.0741394758,-0.0553186908,0.4579606652,-0.2235922217,0.4109124541,0.1331116557,-0.1093586162,-0.1176005602,-0.1932198554,0.4728285968,-0.2240560055,0.0294254292,-0.0382619873,-0.2882134616,0.073822245,-0.3293465674,0.0006597345,0.3874143362,-0.2322069257,0.1698437333,-0.2407621443,-0.1499233544,-0.3715638816,0.1180474758,-0.2668705881,0.0060680099,0.1096710339,-0.2099918723,-0.1868218333,-0.4137237668,-0.4379819036,0.2904361486,0.2337206453,0.1537478119,0.4324122965,-0.2906364202,-0.0364421308,-0.1638887525,0.8776168823,-0.0471252128,-0.4652521312,0.3030637503,0.0765909031,-0.3506373763,0.047269959,0.0481738225,0.3198086023,-0.3462214768,0.3192034364,-0.2881579995,-0.2275656015,0.3936283886,0.1990343928,0.0815921873,0.0893720686,-0.1340238899,0.0799694136,-0.0584960431,-0.2685229778,-0.0878795236,0.4243321121,-0.3054555058,-0.0423197486,-0.3029095531,-0.2251659781,-0.067994535,0.2495992035,0.3446760476,0.0142231565,0.0559494533,-0.0384294949,-0.0930851847,0.5758818388,0.3656627238,-0.1264256686,-0.3438638151,-0.0423652604,-0.079172641,0.4678758979,-0.1007185578,0.0841262192,0.3410396278,-0.288405329,-0.0456498787,-0.1305135489,-0.0745454356,0.0298489537,-0.0139244897,-0.4618703127,-0.255763948,0.3674506545,0.1260503381,-0.0256794281,0.463991493,0.08298105,-0.1904276907,0.0207840335,0.0098596159,0.8966764808,0.0596232563,0.2757004201,0.256747365,-0.2981203794,0.3558058441,-0.1812604666,0.3097745776,-0.3066489398,-0.029555982,-0.2604186535,-0.1996923387,0.3407333195,0.2249901891,-0.3719777167,-0.0644298196,-0.3056955934,-0.2230795473,0.0173065178,0.0808954686,-0.023418868,-0.4417957366,-0.1849945933,-0.0637986809,-0.1060140282,-0.3592424393,-0.1693647802,-0.0410866737,-0.1509803683,0.0985207334,-0.2436625063,0.0995113924,-0.02471059,0.16521281,0.3316884637,-0.3736480176,0.0339357406,0.2207980454,-0.2576705515,-0.2243197411,-0.0132425055,-0.09695407,0.1727022231,0.1069507375,0.1969632059,-0.157610774,0.2014959306,0.1474806666,-0.4610093832,0.1778361797,-0.0069199149,-0.4613776803,0.1420334131,0.2087974995,-0.0012906205,0.0477046855,-0.0021934889,0.3976413608,0.2397092879,-0.263053149,0.0156452581,0.317511797,-0.0865342617,0.290056318,-0.326157093,-0.2962747216,-0.1808796972,0.4714464843,0.0306737442,0.3516027033,0.2016851306,0.2436854392,-0.1698751003,-0.1079997271,0.0497925654,0.3055420518,-0.3950815797,-0.0473992899,0.2083682269,-0.0649423301,-0.1075998694,0.2953146994,-0.0114643108,0.402772218,-0.243648693,-0.2460116595,-0.3364085853,0.2480018288,0.31662184,0.1127138212,0.2214525789,0.2741640508,0.1236607358,-0.0652220026,-0.2084725648,-0.042140808,0.0053929468,0.0677642748,0.0958331302,0.1007972658,-0.1477750987,-0.1431111842,0.1375611722,-0.0252182875,0.0167367831,-0.1208443195,-0.1074450091,0.1239082739,0.090865992,-0.2139456123,-0.0985748619,-0.1487617195,-0.3023738861,-0.3176002204,0.0574127883,0.4234733582,-0.3227651715,0.1867679209,0.1893169731,0.010110761,-0.0515126958,0.0707517788,0.1303385794,-0.1048759967,-0.0562458001,0.409755975,-0.0604097433,-0.0842280164,-0.1349344254,0.0976625085,0.0329356752,-0.0573479235,0.0110330656,-0.1591788679,-0.1203738153,-0.1842466891,0.3884129226,0.1567456573,0.2498006821,0.19566378,0.3727380335,0.1325202882,-0.3204218447,-0.2983633578,0.0081938254,0.2753521502,0.2538135946,0.0434173681,-0.1207035929,0.0566852018,-0.0781914815,-0.0479852408,-0.0221239273,0.035925813,0.2558054328,0.3084744215,0.0704802051,0.0720238984,0.0009227169,0.1066492647,0.2689950168,-0.068648003,-0.2647779584,0.2673962414,0.2540494204,-0.2251475006,-0.1445079893,-0.3573763072,-0.0025077201,0.2004942,0.0413181335,0.0645491183,0.145562917,0.4730015695,-0.0086351857,-0.1192494333,0.0373827666,0.2118678838,0.0130721526,-0.1205486655,-0.3114872575,-0.3096059561,0.0150656365,0.0557751358,-0.0459837057,-0.1940574348,0.6483001709,-0.0610023513,0.0668933243,-0.2883142829,0.5695890188,0.3132674098,0.1505685747,0.0833427981,-0.0736339763,0.3232869506,-0.1268063486,-0.0176824685,0.1162416339,0.410545826,0.2340721786,0.071700491,-0.1317362487,-0.1633090079,-0.0918440521,-0.1385730207,-0.063571617,0.1402385682,-0.0525459498,0.327152431,0.1570770293,-0.0177535452,-0.0164922979,0.1322992295,0.311196208,-0.549384892,-0.0683372691,0.1063666195,-0.4318335652,0.3776520491,0.0480137877,0.0639222413,-0.1442487687,0.1444073766,0.0125828963,-0.1712986976,-0.0671480075,0.0333336666,0.3119398057,0.3572604954,-0.0917434469,0.2588999867,-0.2507966161,-0.0364117622,-0.2457784861,-0.0583240464,0.0326447301,-0.3628958166,-0.1146352813,0.1966533661,0.3155165613,0.0640326589,0.2317622602,0.2420227826,0.3450902104,0.4110006988,-0.2325463146,0.1246387959,0.1620644629,-0.0696848184,-0.0361638665,-0.3552899659,0.3243925571,0.4109692872,-0.0147687513,-0.171863541,0.1330033243,0.1479050517,0.0294813383,0.5030781627,0.0799614564,0.247620672,-0.1521263868,0.1969065517,-0.3423469663,-0.1439630985,-0.3880272806,0.2662152648,0.0031732083,0.1038075238,-0.3213914633,0.0565838143,0.0116899861,-0.0113983676,0.0905898809,0.5292708874,0.0320646204,0.2856718302,0.2315205485,-0.3014598489,-0.204269141,-0.0537179522,0.0942554548,-0.0446446948,-0.1167248785,-0.0606504492,0.0480540097,-0.0629007295,-0.3654365242,-0.2624554932,-0.0996065661,0.0968552306,0.0432380438,-0.2877285182,-0.1671684831,0.2455510795,0.1739555895,-0.2317970544,0.217087701,-0.1821328551,-0.0675741509,-0.2149738222,0.283803761,-0.1606426984,-0.228373304,-0.1463682503,-0.5240388513]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"@lhoestq, we have two Pull Requests to implement:\r\n- Dataset.add_item: #1870\r\n- Dataset.add_column: #2145\r\nwhich add a single row or column, repectively.\r\n\r\nThe request here is to implement the concatenation of *multiple* rows\/columns. Am I right?\r\n\r\nWe should agree on the API:\r\n- `concatenate_datasets` with `axis`?\r\n- other Dataset method name?","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":51,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\n@lhoestq, we have two Pull Requests to implement:\r\n- Dataset.add_item: #1870\r\n- Dataset.add_column: #2145\r\nwhich add a single row or column, repectively.\r\n\r\nThe request here is to implement the concatenation of *multiple* rows\/columns. Am I right?\r\n\r\nWe should agree on the API:\r\n- `concatenate_datasets` with `axis`?\r\n- other Dataset method name?","embeddings":[-0.2659380436,-0.0951724425,-0.0820606872,0.0408310741,-0.2382574379,0.1075186431,0.2930018604,0.2532637119,0.0705212206,0.163655743,-0.0048813536,0.3495405316,-0.0274940208,0.506087184,-0.0220796485,-0.0006613127,0.2938616872,0.5536045432,-0.258451134,0.0008193748,-0.5191210508,0.0612342656,-0.4006459713,-0.2011353225,-0.1511581987,0.1710902005,-0.4236633182,-0.092278868,-0.3538229465,-0.0868456736,0.3892123699,0.1871355176,0.0247681364,0.2834436893,-0.0001184696,-0.0120507609,0.0087371422,-0.122200653,0.0314648636,-0.2648957372,-0.4881327152,-0.503651917,0.0892496109,-0.176972419,0.1351442635,-0.4490820169,0.1138924435,-0.3092614412,0.0400503054,0.1761848927,0.1028396562,-0.2188607305,0.0002189216,-0.1063739136,0.0210726708,0.3390488625,-0.0424190722,-0.2795723081,0.1721025556,0.1112681478,0.0802450702,0.0127466992,0.1822037399,-0.0479208753,0.1250523925,0.1816170216,0.0806788653,-0.4402333498,0.1858932525,0.2143524438,0.7385683656,-0.1596595347,-0.1948232204,0.1284783781,0.3589209914,-0.0335197598,-0.1378203481,0.3503243625,0.0804880932,0.016247917,-0.1092110351,-0.3514394164,-0.2742884755,0.4549536109,-0.2489991784,0.4160044789,0.0517960228,0.2449574769,-0.1435271204,-0.2731948793,0.1568623036,-0.2587737441,-0.0133946165,0.393481344,-0.7762804031,-0.0654555783,-0.0358732305,-0.172175318,0.352060318,0.2876513898,0.2437452674,0.0126422876,-0.1284730881,0.1759790182,0.1850314587,0.1046386287,0.3357360065,0.1796384752,0.0372422561,-0.0798572674,0.3408164382,0.1116783842,-0.3545680642,-0.2287241668,-0.2912501097,-0.0033484448,0.1097575054,0.219615221,-0.0571818277,-0.1750199497,-0.4987612665,-0.2761734128,-0.0422026739,-0.2724586427,0.3980245888,0.7039450407,0.0774429813,0.1547517627,0.460747838,-0.026079705,-0.1366769224,0.3382978439,0.0647643954,0.1297107935,0.1435089111,-0.3264356256,-0.1615809798,0.3666673899,0.2250947058,-0.0098874541,0.2446798235,-0.2090383172,-0.02384842,0.261744976,0.1008512378,-0.2530786693,0.0368566662,-0.1425586194,-0.1416334063,0.5437421799,-0.455088228,-0.0620337501,-0.3632274866,0.0947867706,0.1142199412,0.0765894502,-0.3987362683,0.5860467553,0.0410762765,-0.0614060536,0.2056023479,0.06568297,0.2846110761,-0.2194785476,0.1627058238,0.112755008,-0.3627958894,-0.0764591545,-0.0711542815,-0.1191039681,0.0906974301,0.1103571653,-0.3048974872,-0.0153968679,-0.1455277652,0.3676785827,0.7453736663,-0.5776740313,0.0677253753,-0.0179841202,-0.1502817422,-0.2375476807,0.3531585932,-0.0343224369,0.2178145498,0.211429894,0.4623912573,-0.0348031484,-0.1017920449,-0.1677041203,0.1560549289,-0.4027188718,0.2917557359,0.0737123415,-0.080206044,0.093509309,0.106322147,-0.5713726878,0.3373341262,-0.2997717261,0.1135528386,-0.1664852202,0.1833057106,-0.425771445,-0.0045941123,-0.0861032158,-0.7237885594,-0.1307581961,0.053054437,0.2005598247,0.218866244,-0.2594639659,0.051768396,-0.2207985669,0.0086278226,0.2065585405,0.050066229,0.1679950505,-0.0919817537,-0.2912178934,-0.5517224669,0.2889827192,0.0675501004,0.0585341826,-0.2895825803,0.2176003307,0.2340444773,0.1096145958,-0.0363398381,0.1233324409,0.0766651556,-0.2965348065,0.3771733046,0.4547506869,-0.2495809346,-0.1006749198,0.2262954265,0.2702122927,0.2350303382,0.402156204,0.0361059047,-0.6281508803,0.2069537342,-0.1773712784,-0.0195848253,0.257620275,-0.1827200502,0.0336572975,-0.112759538,0.0090819988,0.1381788999,0.15898785,-0.4199895561,-0.2442210019,0.0635697916,0.4059892297,-0.0317375101,0.0387631282,-0.2467802465,0.0434407927,0.5327420235,-0.141611129,-0.0219513848,0.1453937143,0.0774851739,-0.1550809443,0.2595832646,0.1713085771,0.5673298836,0.1690229475,-0.1723489165,0.0535901785,0.0660697222,-0.0508411601,0.147713989,0.0979552716,0.1257692724,0.2814124227,0.4114007652,-0.1037703753,-0.0634790808,-0.355222106,0.3785521984,0.1729941666,0.0084063746,-0.1615497917,-0.2478617877,0.0271673221,-0.1061720178,-0.3447228372,-0.0536144786,-0.2709853351,-0.1538495868,-0.2257503718,0.0189992078,0.4302059412,-0.3312753141,0.5477534533,-0.3603266478,-0.06783171,-0.0629120544,0.2237073183,0.1445543766,0.0330756679,-0.2828635871,-0.2202087194,0.161260739,-0.2060239166,0.3179126978,-0.1629032046,-0.1811393052,0.1974108368,-0.1518647671,-0.5007161498,0.3635136783,-0.0827935189,-0.1428014636,-0.154132247,0.0581260435,0.3354324698,-0.1542829424,-0.0468135811,-0.0196095519,0.1223357692,-0.3268173039,-0.1804267764,-0.2262930125,-0.2809428573,0.2300112098,0.1175999716,0.2725910842,-0.1568998545,-0.0127175162,-0.0007027829,0.1691474915,-0.1735597998,-0.1779267341,-0.4013747573,0.2534333169,0.0225359574,-0.0000554147,0.0847217068,0.0013698537,0.1821198165,0.5443748236,0.0448849164,-0.5154663324,0.0607238635,0.4005960524,-0.182083562,0.05265899,0.642696619,0.3233398199,-0.0403671376,-0.0673003197,-0.2063944042,0.2793902457,0.4032505453,-0.3642904758,0.2013050318,0.2751306593,0.0929688215,0.0622564219,0.5054622889,0.0393628962,0.2922310531,-0.0776181668,0.199205339,0.1672183573,-0.136530146,-0.1519128382,0.1184378117,0.3083663285,-0.0467519946,-0.0385840014,-0.0922700241,-0.326947242,0.1583937854,-0.2337283045,0.0420070961,0.1047855988,-0.4960407019,0.4481529593,-0.299162209,-0.1178069711,-0.3616155684,0.1685625464,-0.2411234677,0.3434290588,0.0083737103,-0.1559028476,-0.1463203728,-0.2115322053,-0.5169849992,0.6049105525,0.1331953704,0.4692654312,0.2555139065,0.0806022957,-0.0036345029,-0.2486986816,0.6828721166,-0.1374824941,-0.7627074718,0.1882392466,0.1565132886,0.0762536377,0.0738666728,-0.0890095234,0.2012547851,0.0860771462,0.5689386129,-0.4552724659,-0.1842133105,0.3902847171,-0.093749769,0.1354747266,-0.0241969172,0.0389217474,-0.2280540317,-0.0203228407,-0.5453959107,-0.0347625129,0.3650994599,0.0949463025,-0.0065666009,-0.0906306654,-0.3901824653,0.0862176865,0.0129139628,0.0156074781,0.1319618374,-0.1676281989,-0.0249622669,-0.188012585,0.5056007504,0.3024645448,-0.0728693455,-0.5819800496,0.0669580102,-0.3535189033,0.3753157854,-0.0340739787,0.0915389881,0.4277943671,-0.2331534177,-0.0033340745,-0.0182279665,-0.0767424405,0.3826944828,-0.046935562,-0.5235780478,-0.4147223532,0.2101319283,0.2485021502,-0.0165551566,0.4987654388,0.2634552419,-0.3296020031,-0.2017481476,-0.0129103977,0.9365925789,0.1939231157,0.180941239,-0.0044170641,-0.3974725306,0.560654819,-0.2324373275,0.3922207057,-0.1786491573,0.0300494563,-0.3281524479,-0.188546136,0.4826352,0.1770555824,-0.4571590126,0.1632849723,-0.2388965786,0.184555918,-0.0881109685,-0.0314198248,-0.0060571469,-0.4417263269,-0.332988292,0.0666843131,-0.287360549,-0.3360757232,-0.3041713536,-0.0794162154,-0.0443299413,-0.0642534196,-0.3397365808,-0.138821736,0.1110100299,0.1522612125,0.1347534359,-0.6152999997,-0.0619004294,-0.0827064365,-0.1349840909,-0.272752136,0.1539119333,-0.1507541388,0.0338161476,0.0909772292,-0.0046526515,-0.2258452177,0.2538472712,0.0882655829,-0.1879055798,0.3396713734,0.1121064201,-0.4694495201,-0.1654673964,0.3670762181,-0.1243226156,-0.0913394839,-0.0802346244,0.4779292643,0.2731666863,-0.1057237163,0.0286959335,0.3009194732,-0.0024810762,0.2158420384,-0.3237312436,-0.130543381,-0.0994138643,0.3208955228,0.1321807504,-0.1705539674,0.1659011096,-0.0440683477,-0.116954729,-0.0541420504,0.313800931,0.2610627115,-0.3240880966,0.0037105265,0.0574519709,-0.0841064528,0.0361734182,0.3860148191,-0.0811924264,0.4857614338,-0.0453150496,-0.2722169459,-0.1927320808,0.1752653569,0.2161035389,0.1199374497,0.0992690921,0.0257990118,0.0873121247,-0.090391539,-0.2035096884,-0.1518657804,-0.2255321145,0.0990368798,0.0344717205,0.1019615382,-0.1322094053,0.0433800519,0.1176022813,-0.0309013836,-0.0362746865,-0.0832625926,0.0134010119,0.1989131123,0.271892339,-0.0829382762,0.1289587766,-0.1453686357,-0.4257693291,-0.1130372509,0.1239675358,0.1796738803,-0.2157672793,0.1030599773,0.1339569688,-0.0609334968,-0.1370384246,0.0208140705,0.3245593309,0.0707513019,-0.2378291488,0.3518879414,0.029448634,-0.0522331744,-0.2871178985,0.3617918789,0.2377800643,0.1915775985,-0.0393922552,-0.1410298496,0.1022868976,-0.0410287678,0.2688397467,0.3218462169,-0.1569656432,0.1806854457,-0.0366642326,0.0841390491,-0.1616769433,-0.3302039802,0.111653246,0.4873654544,0.1295969784,0.1054629385,-0.3884117305,-0.0021566553,0.0917226225,0.005239292,0.2883105874,-0.276376456,0.3092471957,0.1430706382,-0.0911045745,0.1455274373,0.1306857616,0.3028081656,0.0655435249,0.02915393,-0.0292403642,0.0241987761,0.6329202056,-0.3214028478,0.330799073,-0.1272871792,-0.0666886792,0.2555521727,-0.0257827062,0.1582731754,0.3990779221,-0.1910302937,-0.0708009377,0.0098810503,-0.0617143363,0.1584403515,-0.0908686742,0.0154387513,-0.1980733126,-0.1592763662,-0.0526956432,-0.0575690083,0.0449729674,0.0679866076,0.4651589692,-0.0106762005,0.0533904396,-0.1582463831,0.4188797772,0.0925083831,0.160216555,0.0117973303,0.1006492823,0.4103654325,-0.1588496864,0.27220276,0.1824039668,0.0696911588,0.1628680825,0.4065343738,0.0344981365,-0.2243819833,-0.0364555269,0.0498354025,0.0103007704,0.1431871355,-0.2174358666,0.0759617686,0.1773578078,-0.0004712512,0.4957228005,-0.0077623846,-0.0601394698,-0.793012917,0.4609390795,-0.2092262655,-0.4095269144,0.2161488235,0.1652542651,-0.0257308353,-0.1591013819,-0.0952442884,0.1630519629,-0.0859139562,-0.1366682798,0.0605926961,0.0856018737,0.2372969687,-0.1111216396,0.1397966444,-0.1057889909,-0.2706924379,-0.1531235725,-0.1015585363,0.2189246267,-0.387680769,-0.1370244473,0.1394798458,0.3020131886,-0.1061061621,0.2027369887,0.1105291396,0.0659361556,0.1947049797,-0.0948869511,0.0733061731,0.3137875795,-0.1603407115,0.0028227835,-0.2452968955,0.2166268378,0.2222867161,-0.0723541081,-0.2109267265,0.0426267833,0.2366086096,-0.3100658059,0.1615151018,0.1329053342,0.1775909662,-0.2898869514,0.2474562824,-0.0665620938,-0.0518042818,-0.5047516227,0.3855798542,-0.2173956335,0.0169956852,-0.2887913883,-0.1012295112,0.1427825242,-0.0787410364,-0.053963773,0.2523762286,-0.1317421347,0.3597705066,0.0979177877,-0.4467119277,0.0543239787,0.0797116682,0.1406457126,-0.0075273812,-0.0072467728,-0.0595965572,0.0752680376,-0.049971275,-0.2651265562,-0.2647545338,0.0607160181,0.0697256923,-0.0122640207,-0.2821732163,-0.3014077544,0.4539093673,0.0419733748,-0.4050902128,0.3852284551,-0.028943561,-0.272331059,-0.258952558,0.1456723064,-0.0210318472,-0.1197115108,-0.1088757589,-0.2165898979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"For the API, I like `concatenate_datasets` with `axis` personally :)\r\nFrom a list of `Dataset` objects, it would concatenate them to a new `Dataset` object backed by a `ConcatenationTable`, that is the concatenation of the tables of each input dataset. The concatenation is either on axis=0 (append rows) or on axis=1 (append columns).\r\n\r\nRegarding what we need to implement:\r\nThe axis=0 is already supported and is the current behavior of `concatenate_datasets`.\r\nAlso `add_item` is not needed to implement axis=1 (though it's an awesome addition to this library).\r\n\r\nTo implement axis=1, we either need `add_column` or a `ConcatenationTable` constructor to concatenate tables horizontally.\r\nI have a preference for using a `ConcatenationTable` constructor because this way we can end up with a `ConcatenationTable` with only 1 additional block per table, while `add_column` would add 1 block per new column.\r\n\r\nMaybe we can simply have an equivalent of `ConcatenationTable.from_tables` but for axis=1 ?\r\n`axis` could also be an argument of `ConcatenationTable.from_tables`","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":158,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\nFor the API, I like `concatenate_datasets` with `axis` personally :)\r\nFrom a list of `Dataset` objects, it would concatenate them to a new `Dataset` object backed by a `ConcatenationTable`, that is the concatenation of the tables of each input dataset. The concatenation is either on axis=0 (append rows) or on axis=1 (append columns).\r\n\r\nRegarding what we need to implement:\r\nThe axis=0 is already supported and is the current behavior of `concatenate_datasets`.\r\nAlso `add_item` is not needed to implement axis=1 (though it's an awesome addition to this library).\r\n\r\nTo implement axis=1, we either need `add_column` or a `ConcatenationTable` constructor to concatenate tables horizontally.\r\nI have a preference for using a `ConcatenationTable` constructor because this way we can end up with a `ConcatenationTable` with only 1 additional block per table, while `add_column` would add 1 block per new column.\r\n\r\nMaybe we can simply have an equivalent of `ConcatenationTable.from_tables` but for axis=1 ?\r\n`axis` could also be an argument of `ConcatenationTable.from_tables`","embeddings":[-0.3482715786,0.0645761117,-0.0892973468,0.1464073062,-0.0047059562,0.0687122494,0.1781945527,0.4301660955,-0.0550013296,0.1688200533,-0.0407027751,0.5322623849,-0.1796880811,0.445717454,-0.1766492575,-0.1298689544,0.2669727206,0.6772677898,-0.5987274647,0.1521145254,-0.2508710027,0.0997436792,-0.3959632218,-0.3213891685,-0.2488771528,0.1981462091,-0.3766765893,-0.1862781644,-0.3892164826,-0.2522885203,0.4810886681,0.4199801087,0.1927005947,0.2435198277,-0.0001130875,0.1142905056,-0.1581038088,-0.1443047673,-0.1761360765,-0.4564717412,-0.3388725519,-0.4732193351,0.0550094843,-0.0381946638,-0.0079040946,-0.4393771887,-0.0983280316,-0.4240880311,-0.1667995751,0.067817919,0.1913835555,-0.1983626038,-0.0467882939,-0.1468116045,0.0341700539,0.2906013131,-0.2897627354,-0.3481404483,0.0273524895,-0.0086381566,0.1154628769,-0.0258336309,-0.0689706877,-0.1557124555,0.2706198692,0.0808820054,0.1776652187,-0.4447636604,0.0939608514,0.343719542,0.6369028687,-0.4048181474,-0.2229255438,-0.0605255365,0.2747723758,-0.1232588738,-0.0366237573,0.2869488001,0.206185624,0.1165096238,0.0360166915,-0.1216584221,-0.4184626341,0.4105969965,-0.0651765838,0.5333595872,0.0679957345,0.1643005311,0.2092807889,-0.1621576995,0.1780517995,-0.305097878,-0.1483661681,0.4382909536,-0.598508656,-0.0461570807,0.0286519378,-0.1468940675,0.221202448,0.2496879697,0.2240441889,0.0773980767,-0.2034948766,0.124492839,0.3155554533,0.2442362607,0.3259379864,0.1582871377,0.0588031709,-0.0576618239,0.4282644689,0.0426718444,-0.3903122544,-0.2664996684,-0.3373597562,-0.0756568685,0.2000364661,0.1580793411,0.0427088365,-0.0701212063,-0.1635327041,0.0718540326,0.1799079776,-0.0388043784,0.1471805274,0.8159406185,0.1902590543,0.1429032832,0.222755,-0.049682945,-0.0719223693,0.2015489042,0.0534341633,-0.015657058,0.2549004853,-0.2117113769,-0.2964214385,0.3922960758,0.3107230663,0.201591745,0.1524663419,-0.2357833087,0.272909075,0.2159264982,-0.0233637188,-0.1713620275,0.1085500717,-0.1895643473,-0.1724314541,0.52756989,-0.1246685386,-0.0434619971,-0.4908125997,0.0946800932,-0.2702818215,0.1111802086,-0.324477762,0.5614660978,-0.2449736446,-0.1677212864,0.0688402876,0.0336241573,0.1603230089,-0.1857771873,0.2226322889,0.1458013803,-0.3933649659,-0.0158849023,0.2394766212,-0.2605885267,-0.0569060519,-0.1669490188,-0.1067134887,-0.1044214815,-0.0215456896,0.1586321592,0.389970392,-0.2733126283,-0.1256055385,-0.0999528766,-0.1640653759,0.0103304069,0.1373250484,0.1862504929,0.2845201492,0.1511421502,0.2225961536,0.1819189638,0.0488866791,0.0880408809,-0.0409181193,-0.2720239162,0.4463224113,0.1754126102,-0.2321843356,0.1784433573,0.0155665707,-0.4601062834,0.323685497,-0.3310925364,0.2429708987,-0.0784440413,0.2711414695,-0.3207565844,0.0282503609,-0.070944123,-0.5652903914,0.0586757548,0.0997295529,0.3890404999,0.1061437726,-0.2458092123,0.3670020401,-0.0809095055,-0.0522159934,0.2115101367,0.1009573936,-0.1020019799,0.0345257558,-0.1555290669,-0.4602502584,0.3170199394,0.0911786109,0.0120880818,-0.3484279513,0.2383431494,0.2912606299,0.122279644,-0.0443316698,0.3409978151,-0.0028972286,-0.1761046499,0.2134356052,0.4273332655,-0.2389496714,-0.1755453348,0.0812274441,0.4070699215,0.1399201155,0.3254657388,0.0133524667,-0.6059426069,0.1656654775,-0.005250657,0.0453851297,0.7022058964,-0.3599378467,0.1289962977,0.0048014605,0.1523850262,0.2226151377,0.1210525185,-0.4439998567,-0.3551975191,0.0241794195,0.3818527162,-0.3455274105,0.0015524292,-0.1329493076,0.1672097296,0.2806382477,-0.036435362,0.0220139213,-0.0107343197,0.0183313824,0.0007221641,0.341783762,0.1739911288,0.4883176982,0.2531154156,-0.1698952168,0.1381896138,-0.2280168682,0.0345700458,0.2451861054,0.0025361185,-0.0327581502,0.1938692927,0.2984042466,-0.0517041683,0.0257179327,-0.5200487971,0.0546143986,0.4540601373,-0.1175141931,-0.1232595444,-0.4112281501,-0.0146782696,0.0131857749,-0.5087953806,-0.0605513453,-0.1007205024,-0.0289612841,-0.0368107595,-0.1229571477,0.3387333155,-0.2723270357,0.4934965074,-0.5590958595,-0.0255259685,-0.0199112277,0.0346751064,0.1099816561,0.0477813445,-0.031433776,0.0257086307,0.030616682,-0.0037529406,0.2519108653,0.0522177108,-0.1747587174,0.1148420125,-0.0004915578,-0.5256105065,0.173251465,-0.0341614597,-0.109292224,-0.0681305975,0.050896287,0.1972895563,-0.0247644298,0.1738230735,-0.2129161805,-0.1244230717,-0.4129621387,0.0100566652,-0.1492166966,-0.2941409051,0.3062925041,0.2049193531,0.276792109,-0.0660553724,0.3649310172,0.0285790451,0.1179981604,0.0248038545,0.0412423164,-0.3984266818,0.3619605601,-0.0162214395,0.0040751467,0.1638403833,-0.1292437911,0.231779024,0.3739016652,-0.0349322706,-0.1287055761,0.0903959796,0.321022898,-0.2032411993,0.1163638383,0.5452889204,0.4964998364,0.0389782526,-0.0805067495,-0.1451353282,0.2195563167,0.1137303486,-0.409794271,-0.0398146175,0.0830532014,0.2609753609,-0.1843200475,0.3280723691,-0.0582596287,0.348760128,-0.0385393165,0.4031772316,0.0968077183,-0.0685751215,-0.2857970297,0.1226112917,0.3362408876,-0.1062162742,0.1177512184,-0.1936092973,-0.4214165211,0.0606006347,-0.1784741431,-0.0546933897,0.2122351825,-0.452472806,0.1912060976,-0.3075186908,-0.0593355373,-0.4267512858,0.0754699111,-0.2327066958,0.272051394,-0.0361563489,-0.2596208751,-0.2545233965,-0.433667928,-0.3955099881,0.3501536846,0.1118818521,0.4040623605,0.4325247407,-0.0526388958,-0.1201394349,-0.2049402297,0.4980341494,-0.074016735,-0.5626016259,0.307575345,0.0144362859,-0.2854399383,0.0122462977,0.0523534305,0.2708072066,-0.1133352146,0.3593371809,-0.2529783249,-0.2239244133,0.3291352689,0.0781999975,0.1901568919,-0.036209248,-0.1000168324,0.0483827181,-0.0891680866,-0.4129238427,-0.1794504672,0.4076736569,-0.1807337999,-0.1760284752,-0.1747318357,-0.3464696109,0.073598817,0.178296566,0.2868493497,0.045461826,-0.0356042422,0.0390238538,-0.2527460456,0.4640751779,0.2757255435,0.0017459657,-0.4743174016,0.1603029966,-0.1189007014,0.1470323652,0.0156090334,0.0655274242,0.3621349335,-0.2586050928,-0.0546791628,-0.0449764542,-0.1711930037,0.0605065264,-0.0700813755,-0.6451605558,-0.4660741985,0.2414062172,0.1642179936,0.0081462199,0.7264255881,0.0391053595,-0.3268624842,-0.0575176068,0.0129785938,0.8818545938,0.2684469223,0.1046683043,0.1382033676,-0.2416412085,0.4139749706,-0.213951394,0.3939244747,-0.3427498341,-0.1126441285,-0.2992990613,-0.1103008389,0.4037828743,0.1506044269,-0.4877853692,-0.0114333471,-0.2502489388,0.1278433651,-0.1170871109,0.0348281153,-0.0709373578,-0.5233120322,-0.4722076654,0.0572582223,-0.1092985719,-0.4260680676,-0.3047958016,-0.0353248008,-0.0910702199,0.0351721607,-0.2306586355,-0.0044141468,0.0521646254,0.0739053637,0.2463569194,-0.2035380751,-0.0238776505,0.0880607069,-0.0804830194,-0.1443761587,-0.1169088036,-0.1940747201,-0.0046441532,0.1477806121,0.0605574287,-0.2075451463,0.2673376799,0.1429786384,-0.2555795312,0.1882784069,0.1570274979,-0.3402674794,0.151824981,0.1354855299,-0.0706053749,0.0696972758,-0.030659711,0.4619453549,0.2831305563,-0.0958659351,0.1021937653,0.4088977277,-0.1689614803,0.1002548486,-0.1308337152,-0.2835275531,-0.1844327897,0.4703077972,0.1327184737,-0.0097242696,-0.0830394775,0.1672370881,-0.1261287779,-0.1417265236,0.2227784097,0.3376014233,-0.3897215128,-0.0286747552,0.1679208875,-0.1050640568,-0.0428482331,0.274360925,-0.1265005767,0.4867375493,-0.1858577728,-0.2188507617,-0.1902679503,0.1972333044,0.2149624676,0.2364257723,-0.0278679356,0.1071326956,-0.0475700758,-0.2605208158,-0.265489608,-0.2106081396,0.0451844819,-0.0754061118,-0.0286610406,0.1914942265,-0.2067786306,-0.2142188698,0.1541944295,-0.0128007578,-0.1399383992,-0.1516808271,-0.1325609237,0.1478936225,0.1064864323,-0.1686020643,-0.1028909385,-0.2250379175,-0.5166934133,-0.3138574362,0.1047011167,0.2757297158,-0.3435199857,0.1125442833,0.1932657063,0.0849787295,0.0095063196,-0.0532588214,0.2293924838,-0.0755905658,-0.0246028323,0.2878011763,-0.1329402775,-0.0555085018,-0.2768466473,0.355145961,-0.0233925208,0.1009764597,0.0635814369,-0.0852632672,0.2157887071,0.2338927686,0.1922274381,0.186818108,0.2035005242,0.1437075734,0.118969433,0.1874693483,-0.0867769569,-0.3905170858,0.1752238721,0.4561102986,0.2235665321,0.0587785877,-0.2400045097,-0.0298411418,0.0576095022,-0.040813081,0.1074297726,-0.1897708327,0.5861267447,0.2604874372,-0.0298043918,0.1763006151,-0.1901127696,0.0819437727,0.2492521405,0.1269637346,-0.2105136365,0.2456100136,0.5388047099,-0.2231216431,-0.0810673609,-0.3095415831,-0.1911907643,0.1502434909,-0.0186622497,0.129054606,0.2990817726,-0.0238250177,-0.2618978024,-0.017926937,-0.0979856849,0.0311884712,-0.0290089585,0.0915174559,-0.2588435411,-0.245770961,-0.0579135194,-0.0626990199,0.0663744435,0.0114244483,0.3781792521,-0.0919896588,0.0840891674,-0.3125244975,0.6007604003,0.2733028829,0.1719360054,0.1691137999,0.0040533836,0.3385770023,-0.017898513,0.096020788,0.0880056843,0.1707542837,0.4111485779,0.2453007102,-0.003423647,-0.2591455877,-0.0920120999,-0.0122419083,0.0898456201,0.2527466714,-0.047232721,0.1304782033,0.2112451792,-0.0388472378,0.2479958981,-0.0010517798,0.194560498,-0.5377138257,0.3259516954,-0.0153190857,-0.2476265728,0.1876193881,-0.0025637683,0.0135935573,-0.3884643018,0.0154285021,-0.0024482301,-0.2050782293,-0.1724731177,0.0693037733,0.2210496962,0.174883157,-0.2789320052,0.2301529199,-0.1784125119,-0.1908657849,-0.1957604736,-0.11202351,0.1847130209,-0.1209947839,-0.1233276129,0.1535859108,0.2016492486,-0.0556820035,0.2447181642,0.1807334125,0.2482660264,0.185690105,-0.1890784651,0.1793857962,0.1156917512,-0.1145537049,-0.0593516529,-0.2988677919,0.0319709927,0.1862419695,-0.0040521477,-0.0816064626,0.181711182,-0.0542002097,-0.23271887,0.3671843708,0.1295995116,0.1132440418,-0.1426472217,0.2120760381,-0.2643704712,-0.0600606948,-0.4451530874,0.3055152893,-0.0275903139,0.1804619133,-0.2197709233,-0.1825060695,0.1628463119,-0.1611245126,0.112172462,0.5696960688,-0.2790895998,0.4884003103,0.1081150025,-0.1315839291,-0.2470436692,-0.0565738119,0.2272877544,0.0492838174,-0.014837428,0.0623115972,-0.0095032379,-0.0775154009,-0.262572974,-0.1371796727,-0.0412405767,0.1345881522,-0.0638451949,-0.2394396961,-0.197496295,0.2346857935,0.0215343926,-0.2805652022,0.2788971364,-0.1155546978,-0.0043689469,-0.2056774944,0.1086844802,-0.0163389295,-0.2483726293,0.0064750244,-0.414254874]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"@lhoestq I think I guessed your suggestions in advance... \ud83d\ude09 #2151","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":11,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\n@lhoestq I think I guessed your suggestions in advance... \ud83d\ude09 #2151","embeddings":[-0.5413237214,-0.1003052071,-0.1948980391,0.1004194096,0.0090673072,0.1702324897,0.4748986959,0.379992038,-0.0324481726,0.2948921323,-0.1000558808,0.2593750358,-0.0725789592,0.4482992589,-0.4687573612,-0.1475338936,0.204702422,0.7700719237,-0.4917495847,-0.0932759047,-0.3471489847,-0.0572348088,-0.3503410816,-0.2354000658,-0.5935836434,0.1543764025,-0.160999313,0.246626243,-0.3593087494,-0.1385876387,0.2981300652,0.2804862559,0.0293613356,0.0763263553,-0.0001092072,-0.0172426756,-0.0381877571,-0.210913375,-0.0772033334,-0.3425039649,-0.6630764604,-0.2920962274,-0.0813337639,-0.0741614029,-0.0895887241,-0.1983622313,0.0325804576,-0.2941661477,0.0225813445,0.1890414059,0.2530962527,-0.1772570461,-0.2069817185,-0.2106200159,-0.0449669659,0.2774246931,-0.0514751114,-0.249887988,-0.1196273491,0.1053017974,0.1317644566,0.2691669166,0.1250127405,-0.1088462546,-0.025147168,0.1826514006,0.3971336186,-0.5081047416,0.275456965,0.3899284005,0.4473588169,0.0371308401,-0.0053826082,0.1005500332,0.0439605825,-0.0923884213,0.0504218824,0.2917488515,0.1725241989,0.2450819463,-0.2068620175,-0.0814758018,-0.3889192641,0.3107754588,-0.3372690082,0.5472215414,0.0924597532,0.2091780305,0.0850314125,-0.2149663717,0.0522600636,-0.2802857459,-0.0091147963,0.2735034823,-0.4403130114,-0.0956466347,-0.4049751759,-0.2847208977,0.2227303982,0.2549239993,-0.0498397276,0.0032614283,-0.0515766256,0.243800506,0.0237077773,-0.0698240399,0.3717694283,0.1926803738,-0.0117151979,-0.2604424655,0.1617744267,0.1492048502,-0.4056868553,-0.5465149879,-0.2631780803,-0.0347245894,-0.070922561,0.0687550455,-0.1131505072,-0.0121216262,-0.1750293672,-0.0024070917,-0.0478333384,0.0571330935,0.2084116787,0.6669795513,-0.0451468118,0.1653524935,0.3062640131,-0.202891767,-0.1880452931,0.2469790131,0.074143365,-0.0368288495,0.0636902377,-0.2338926345,-0.0183996651,0.2025794536,0.2475450337,0.0525212474,0.1546894461,-0.3017824888,0.310118705,0.1843409389,0.1224833578,-0.1084531546,0.0886099115,-0.1603548527,-0.009255656,0.4350531101,-0.1540715098,0.0238943528,-0.4001111984,0.2183384448,-0.1007686257,0.0128012784,-0.1173458919,0.3618926108,-0.2325457335,0.0951410085,0.1154094413,0.0625104159,0.1847585887,-0.2185913324,0.1071977541,-0.0213704929,-0.3310712576,-0.0423188619,0.1018601209,-0.2769185603,0.128582105,0.1242673695,-0.0199311078,-0.3177000284,-0.027804438,0.1978758574,0.6085445285,-0.4483624995,-0.0192597657,-0.0225118455,-0.1771875918,-0.18525213,0.2128385305,0.1594665796,0.0461106412,0.0975892916,0.2276992053,0.0335393734,-0.0389300585,-0.0574467815,0.031697344,-0.2708642483,0.4530325234,0.2252358496,0.0485467017,0.0371546894,-0.014306928,-0.5980929136,0.3700184822,-0.3047553003,0.1748097092,0.074373275,0.2700891793,-0.5048536658,-0.0824422911,-0.0032245272,-0.4140111804,-0.0361536518,-0.0453846939,0.2385606021,0.034964852,-0.2083636373,-0.115628235,-0.2490006834,-0.1024479344,0.1066158786,0.162634328,0.0202419627,0.0709644631,-0.1992158443,-0.4464944899,0.5152198672,0.1768602431,-0.0498799309,-0.1450470835,0.2496038079,0.0584474951,0.086550504,-0.0504301488,0.2078648955,0.1596392542,-0.2056347281,0.3678059876,0.47829023,-0.4052699804,0.1212107241,0.2574973404,-0.0044402699,0.0910858959,-0.1485591382,-0.1892578006,-0.4678788483,0.1214648485,-0.0173648205,0.0358460583,0.4558909535,-0.2713360786,-0.0325651579,-0.1280659139,0.0449764542,0.0781876668,-0.1429164708,-0.3601724207,-0.3655131459,0.0152502088,0.3750608563,-0.1162293851,0.3154688776,-0.2617191076,0.0528885052,0.4928072393,-0.0330175497,-0.2248377949,-0.0574842021,-0.125877887,0.0907401294,0.1909987926,-0.0463851839,0.2550848126,0.3386761844,-0.1492943615,0.1548433155,-0.2208678573,-0.0445988476,0.2310600132,-0.1150544956,0.2029323727,0.3886837363,0.2747204006,-0.106125176,-0.1338111162,-0.292191416,-0.0128163127,0.3565496802,-0.0224604383,-0.125749141,-0.2099318504,-0.0707132593,0.036554236,-0.3831885755,-0.3366951048,-0.1273772418,-0.0129004754,0.0226550847,-0.006530025,0.4958512783,-0.3203112483,0.3203988969,-0.2031337172,0.0734246969,-0.1515676826,0.245251298,0.0925296769,0.1853172928,-0.2057447135,-0.1914178282,0.0343974866,-0.312092483,0.2161119878,-0.0538155772,-0.0066667725,0.0706637651,0.0355312862,-0.5833345056,0.1203363836,0.1055272147,-0.2207583189,-0.0524261743,0.189788878,0.3685263097,-0.112311013,-0.0967121869,-0.1503447443,0.1087661907,-0.2126114219,-0.0482996441,-0.2319453657,-0.2622636557,0.0612174682,0.3086572886,0.2773959637,0.0741458312,0.4565211535,-0.165667668,0.0310828425,-0.057802394,-0.2094094157,-0.5378594398,0.3911677003,0.1064619869,0.0379580595,0.0402058475,-0.1784639806,0.1459362209,0.4697676301,0.0022052284,-0.2300561666,0.0350184441,0.3097981513,-0.4242521822,0.0061484165,0.3858377337,0.3840447664,-0.1435409039,-0.1512637734,0.0680567846,0.3085819185,0.0759602189,-0.1491535902,-0.0950693861,-0.1237328351,0.2151140124,-0.0503986515,0.5197184086,-0.2379965633,0.0945190862,-0.0909191966,0.2578978837,0.2850297689,-0.0826978907,-0.0051408526,0.4165293574,0.4652913809,-0.083024174,0.051198639,0.0847347528,-0.4091944993,0.1038559228,-0.3321980834,0.0637285486,0.1913819909,-0.295255661,0.2457906306,-0.3245898187,-0.148776859,-0.2755119205,0.0720842555,-0.1732793599,0.245058924,0.0672198907,-0.0965071768,-0.1442984641,-0.0775850862,-0.3451694548,0.2751735449,0.1787978709,0.1979744434,0.2517862022,0.033836171,-0.119148396,-0.0638392717,0.9373391271,-0.2427912652,-0.4007003307,0.1892488301,0.2144692391,-0.1797751784,0.1952916235,-0.076766558,0.2037720829,-0.0817862302,0.6227956414,-0.1216454282,-0.2445168942,0.5103909373,0.0327063948,0.1019441038,-0.0243886244,0.0172948502,0.0538743846,-0.0512201898,-0.3392217755,-0.0615914203,0.4431960881,-0.0759375021,-0.2331362814,-0.129599452,-0.3963052928,0.1359055638,0.1441354305,0.1703930795,0.0656756535,-0.1828836799,0.1467134207,-0.1217928752,0.5604591966,0.1922353953,0.0714063197,-0.5288614631,-0.0588053092,-0.1475750655,0.4146777391,0.122561492,-0.0445311405,0.0894699022,-0.2477893531,0.0389989316,0.0373911671,0.0971110612,0.1238880903,-0.0025193554,-0.658577323,-0.130291149,0.1683579981,0.244731307,0.0218154192,0.3498449028,-0.0205625985,-0.3401846886,-0.2555495203,0.039576035,0.9742177725,0.1969002783,0.2020006031,-0.0682547241,-0.1248191372,0.3875368834,0.015093876,0.2669775486,-0.2329411209,-0.0839165673,-0.2550908327,-0.1761299223,0.3061948419,0.2572872639,-0.3125679195,0.0432480462,-0.3326020837,0.1023469493,0.0034519592,0.0184214376,-0.0658964664,-0.3935294449,-0.2261766493,0.1042756289,-0.1777770072,-0.1405581832,-0.3017210662,-0.0801482946,-0.1858669221,0.0012231488,-0.0454975143,-0.0527239107,-0.0686838105,0.0126823755,0.2269649506,-0.4333641529,-0.2621794939,0.1766919494,0.1112042814,-0.0993260145,0.0489172228,-0.0449822098,0.096293807,0.121410206,0.0234970748,-0.0736528337,0.1469210088,-0.0373260118,-0.2390606403,0.1672302634,0.1293886304,-0.1683927476,0.0366863534,0.1374218166,-0.0995605588,0.0762531757,-0.1446534097,0.2085071206,0.1589946002,-0.0845115855,0.1181945205,0.2408798188,-0.1158316135,0.0808439553,-0.0520529598,-0.3235879242,-0.1195144281,0.5304312706,0.0958694667,-0.1162992269,0.0173932351,-0.0135014998,-0.0908678472,-0.2355183214,0.3449670374,0.1376710534,-0.3006293476,0.1046736315,0.1252598315,-0.1686789542,-0.170454219,0.4229601622,0.1375764012,0.475440979,0.1593709588,-0.292016387,-0.5532192588,0.06560684,0.1958108544,0.2014091313,-0.0033210341,0.2528998256,-0.0156621188,-0.3432905376,-0.3722118139,0.011239008,-0.118638657,0.1061370671,-0.0974335223,-0.0152003774,-0.1794530451,-0.1399595439,0.3034422398,0.0054222238,-0.2012146413,-0.288705349,-0.0571748503,0.1143432409,0.0592943504,-0.4133202732,0.068917267,0.0251099095,-0.4330668151,-0.1564998031,0.088507846,0.383539021,-0.2166298628,0.029669743,-0.1255039275,-0.0733133256,-0.0423585549,-0.1460123062,0.0920557603,-0.0652790144,-0.1895789802,0.3129301369,0.0192448162,-0.1310908347,0.0242481064,0.1568793505,0.0518534109,0.1310886592,-0.0572997481,-0.1294837594,-0.0013743087,-0.0072886497,0.3505624533,0.0040148259,-0.108822234,0.1971595734,0.1260202676,0.3233270645,-0.1977558136,-0.2385207266,0.1490721405,0.3754609227,0.28024441,0.0515161343,-0.2023123801,-0.1406437308,0.1318602711,0.069531545,0.3885891736,-0.0850331336,0.3761941791,0.2415130883,-0.0427045003,0.2032974362,-0.0143730761,0.3909630179,0.2621408403,0.0214235298,-0.1477411091,0.100749366,0.5039468408,-0.2157059163,-0.0583520122,-0.1326247156,-0.0285516158,0.3359545767,-0.2765117884,0.179289788,0.3484642208,-0.0050271316,-0.0956941098,-0.1428019702,0.2195358574,0.1262497455,-0.0642031729,-0.0548223332,-0.3060777485,-0.2935318053,-0.3178439438,0.0055639418,0.1509181857,-0.2146021575,0.3955417275,0.104535155,0.1079611629,-0.1423147619,0.4736536145,0.2347743511,0.1739435196,0.0721948147,0.1364260763,0.5018185973,-0.0874682441,0.451769501,0.2335474789,0.333652705,0.1314662397,0.2379849851,0.0442182384,-0.219221428,-0.2292195112,-0.0835902169,-0.0365064777,0.3010550439,-0.1441752911,0.1702702641,0.3354735076,-0.1568963826,0.2065784186,-0.1188434884,-0.0477695614,-0.4927087426,-0.1352444738,-0.1374194473,-0.356644392,0.0688883886,-0.1172553226,-0.0279487018,-0.3460725546,-0.0570285246,-0.01651554,-0.023762878,0.0865494758,0.1242566779,0.0428028367,0.2615102828,-0.1889528483,0.1582454294,-0.1775405258,-0.3004092276,-0.0714422092,0.0352164358,0.2242366225,-0.2693081796,-0.2211596519,0.0614113808,0.2386067808,0.0105825132,0.4490698874,0.3104816377,0.3719747663,0.2803595364,-0.1308874339,0.1326898783,0.335782975,-0.0643200427,-0.0006827469,-0.262021184,0.3436221182,0.3731600344,0.1626788825,-0.0387312658,0.0557173193,-0.0854384452,-0.1504700929,0.1416978389,0.2044542283,0.2211107761,-0.307027638,0.2750555873,0.0506389625,0.0104639083,-0.2943790555,0.298751533,0.1620498002,0.3137441874,-0.3003535569,-0.1019016951,0.1443708539,-0.0812624842,0.1299780309,0.3508080244,-0.3162516356,0.4376812577,0.1395700276,-0.293269366,-0.2147998661,-0.021445293,0.0630878806,-0.114420481,-0.1407759786,-0.1803686917,0.2713367045,0.1461749971,-0.4752266109,-0.2457989454,0.0014920739,0.3491264284,0.2876042128,-0.1407812536,-0.1777435392,0.3525556624,-0.0054267189,-0.3030340672,0.2684340477,-0.148743704,-0.0871536657,-0.0644877329,-0.0151966019,0.0075569954,-0.2097451538,-0.1420717537,-0.4172588885]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/853","title":"concatenate_datasets support axis=0 or 1 \uff1f","comments":"Cool ! Sorry I missed this one ^^\r\nI'm taking a look ;)","body":"I want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n","comment_length":13,"text":"concatenate_datasets support axis=0 or 1 \uff1f\nI want to achieve the following result\r\n![image](https:\/\/user-images.githubusercontent.com\/12437751\/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)\r\n\nCool ! Sorry I missed this one ^^\r\nI'm taking a look ;)","embeddings":[-0.5155815482,-0.0580864772,-0.2239764482,0.1333018392,0.1151226759,0.1795205921,0.4404043257,0.4526895583,-0.0764536783,0.3060360551,-0.1705072671,0.3478773236,-0.0317183174,0.4408389926,-0.4537091553,-0.1625219136,0.1577768326,0.7125646472,-0.5475831032,-0.0517587587,-0.2784138024,0.0670344979,-0.3731887043,-0.2570608854,-0.5555547476,0.1851278692,-0.1247621924,0.1754702181,-0.360753417,-0.1189828813,0.3124313056,0.314042449,-0.0122038079,0.1143880859,-0.0001054569,-0.0073451204,-0.0326705575,-0.1953460723,-0.1307891011,-0.2881548405,-0.5886066556,-0.2761183083,-0.1042415425,-0.0521479398,-0.0406038836,-0.3204924166,0.0080311429,-0.3775686324,0.0053415019,0.1436096877,0.2810679376,-0.0583702475,-0.2020744979,-0.2869739234,-0.0387280956,0.2547492385,-0.0642213449,-0.2072035521,-0.0989705175,0.0385011956,0.0924641266,0.2502973974,0.1130794436,-0.0813729391,-0.1092053577,0.1400460154,0.3620582223,-0.4693070054,0.2398084104,0.390340656,0.3796524405,-0.0730064958,-0.0360091999,0.0489338338,0.0631609112,-0.0950636342,0.0120756011,0.215985924,0.0659664273,0.2830992937,-0.1176809445,-0.1073049754,-0.3424516618,0.2544249296,-0.1782375127,0.6519129872,0.043262206,0.156318754,0.1207979918,-0.1971937269,0.1217685714,-0.2609863877,-0.051391121,0.277339071,-0.4040619731,-0.1155929491,-0.3758709431,-0.2019822598,0.2654398978,0.2708342075,0.0324139968,0.0013985961,-0.0724627897,0.2591264844,0.1073816195,-0.064022921,0.3961734772,0.1866090149,-0.0273467638,-0.3011013865,0.2041162401,0.163511619,-0.323561728,-0.6249569654,-0.2708283961,0.0077811261,0.028802434,0.0897952691,-0.0467708297,-0.0768613368,-0.2134748548,0.0535083264,-0.0759218633,0.0245978869,0.1481781155,0.6182163358,-0.0608843416,0.1585977077,0.2402513921,-0.1491344422,-0.2170333713,0.1983865052,0.11600326,-0.0652399212,0.1297375709,-0.2598021924,-0.0125878928,0.2612368464,0.3123256266,0.1482710093,0.1054779813,-0.1578204632,0.2607877254,0.1848374158,0.1406504512,-0.1078134775,0.1641249955,-0.2154131979,-0.0400218293,0.3296026886,-0.1188287959,0.0778603554,-0.49772802,0.2496272475,-0.1370572597,-0.042851571,-0.0765912533,0.3684305549,-0.1102505028,0.0472010933,0.0812427104,0.0158115849,0.1301126629,-0.1621741205,0.1435157508,0.0563312769,-0.3401089311,-0.0482003726,0.1348537207,-0.3508056402,0.0660915524,0.0462109335,-0.0519623756,-0.3198770583,-0.1261694282,0.227063477,0.5038756132,-0.3785028458,-0.0287208669,-0.0206634048,-0.232281059,-0.0844698623,0.256848067,0.095091708,0.0731759369,0.1313910931,0.16983594,0.1413858682,-0.0403484628,-0.0665356442,0.0378431156,-0.1913472861,0.4538066983,0.1565574855,-0.0750372112,0.1041307673,0.0827458054,-0.5132255554,0.4854120016,-0.3297025859,0.2201822549,0.1046254262,0.196526885,-0.4596834779,0.010965256,-0.0332809985,-0.4613925517,-0.0406115316,0.0457751304,0.2568705678,-0.0091816308,-0.2181274593,-0.1060495973,-0.2303851843,-0.0974937603,0.0347821712,0.2123486102,0.0178027656,0.0192666706,-0.1275993288,-0.3686252534,0.4711649716,0.1991760582,-0.1310565323,-0.2193186432,0.2412691414,0.0724534988,0.1186994091,-0.0595979318,0.2203546762,0.0741464272,-0.1684537977,0.2591335475,0.4711952209,-0.3480706513,0.0650249198,0.2556901276,0.0198858399,0.0200209152,-0.1637205184,-0.1903708875,-0.4855538309,0.088739343,-0.0501743779,-0.0176422335,0.5476641059,-0.303321749,0.0730747506,-0.1112863719,0.1527916044,0.2127372026,-0.1140283123,-0.3792066574,-0.3479141295,0.0112204617,0.3094614446,-0.1036533341,0.1978872418,-0.219772473,0.1153494492,0.3679582775,0.0417100228,-0.158367604,-0.1960139722,-0.1648881882,0.1474483609,0.2933868766,-0.0524960756,0.2759038508,0.3383800089,-0.1285715103,0.1632322073,-0.2567316294,-0.1046339124,0.1654088497,-0.0605001114,0.1733321398,0.4324951172,0.2742013335,-0.1643018872,-0.2458649129,-0.3495108187,-0.0515626073,0.3064709902,-0.0912161395,-0.0603533313,-0.3502790332,-0.074013181,0.0479928963,-0.3351544142,-0.22559762,-0.1466371119,0.0335551351,0.092845507,0.0028862585,0.58536309,-0.210146606,0.3882623911,-0.1028584763,0.044280529,-0.1700212359,0.1159076765,0.1318573058,0.2278045416,-0.198851198,-0.1447396874,-0.0086302012,-0.199608475,0.2139638364,-0.0863781571,-0.0356238596,0.0573706739,-0.014866625,-0.5546426177,0.1137610152,0.0818865299,-0.2609014809,-0.0444472581,0.0687497258,0.4143261611,-0.1457023323,-0.1303435564,-0.1443804204,0.0713059679,-0.3425168395,-0.058025308,-0.1735663861,-0.3033429384,0.1405387521,0.2759819627,0.1929913461,0.1069172993,0.5854634643,-0.1125856787,-0.0462631285,-0.117812492,-0.2137790173,-0.4836705327,0.4525052011,-0.0230637435,-0.0125465542,0.1023955122,-0.1814521253,0.0974466503,0.3665453196,0.0018967907,-0.1933666617,0.0714212582,0.2549127638,-0.4106974602,0.1080183685,0.336835593,0.4182127118,-0.1129806414,-0.174665302,0.0864139199,0.2740675509,-0.0409549773,-0.168802321,-0.2342901975,-0.1756316125,0.2114259452,-0.0984527841,0.3557653129,-0.1660605371,0.0789516568,-0.2177328914,0.3769802451,0.3104101419,-0.0965782851,-0.051084958,0.3956594169,0.4602273107,-0.1563824564,0.0829537138,0.1216803268,-0.3008328676,0.1822717488,-0.1991916001,0.0297114644,0.1430083662,-0.3090876341,0.0754571557,-0.2241558433,-0.0868519545,-0.2493718565,0.0833460093,-0.112798281,0.1385931671,0.0148137128,-0.103670679,-0.1905154884,-0.0799644813,-0.206426993,0.2853164673,0.1770788431,0.096094653,0.2708644271,-0.0032520164,-0.125346601,0.004335979,0.9094236493,-0.2637699246,-0.3977023959,0.2957358658,0.1617089808,-0.2308389395,0.1260682791,-0.0781748369,0.1510690302,-0.1484979689,0.6653321981,-0.0190898534,-0.179801479,0.4242911041,-0.0441941842,0.0029175752,-0.0234359633,0.0116065396,0.0556089468,-0.1774528027,-0.3348138332,-0.0198352262,0.4887776971,-0.201514557,-0.1818109006,-0.177590996,-0.3570861816,0.1022010893,0.1906695664,0.1723886281,0.0893501863,-0.249853313,0.109648481,-0.0978539214,0.4168504179,0.2159688771,-0.0179012567,-0.5185494423,0.0412826352,-0.144165948,0.4223347008,0.0831556022,-0.0617762357,0.0654947013,-0.1649211943,-0.007906761,0.0673000738,0.240260005,0.0902411044,-0.0457724258,-0.5940477848,-0.2082719356,0.2303511798,0.2107039243,0.0431239903,0.4256564677,-0.0779037401,-0.2623772621,-0.1738905013,-0.0669622347,0.8383826017,0.290522337,0.1781001091,-0.0325614326,-0.1252667606,0.481572032,0.0361296497,0.2628106773,-0.2362004817,-0.0489717387,-0.268288821,-0.2078687698,0.2621250153,0.220512256,-0.336192131,-0.0556067973,-0.4040441215,0.1430475265,-0.0181322861,-0.042655699,-0.1115783155,-0.3820610046,-0.1892844439,0.1469641328,-0.1001920402,-0.1469146311,-0.3112929761,-0.1077252552,-0.1447879225,-0.0144855622,-0.116501987,0.0747029558,-0.0432367139,0.1288874745,0.2091342211,-0.3754896522,-0.2398540825,0.1047353819,0.0633689612,-0.0418561399,-0.0195300076,0.0011416698,0.1717018336,0.0710474402,-0.0116168354,-0.0712626874,0.0880544409,0.0483885184,-0.2202010751,0.165109545,0.1857271641,-0.2358253002,0.0226342622,0.1215433329,-0.1458854675,0.1114265546,-0.0288346577,0.1488301903,0.1992562115,-0.1611275375,0.1475212276,0.1680375934,-0.1236445382,0.0374844819,-0.0435325615,-0.2828967273,-0.1082370505,0.5214458704,0.0514512211,-0.132752642,0.0183189567,0.0842834264,-0.0836781263,-0.2614895105,0.2542391419,0.1829019785,-0.3181785941,0.1598466337,0.1086567268,-0.1869520694,-0.2286645621,0.4116674066,0.083272025,0.4725841284,0.1200263351,-0.3151127398,-0.4627394974,0.0326800384,0.1235268191,0.2468420714,-0.0312839076,0.2325875163,-0.115396902,-0.310564369,-0.4047044516,-0.0831072628,-0.0157247484,-0.0306020696,-0.082306765,0.0786600858,-0.1425515562,-0.1091148853,0.3138448894,0.1361850202,-0.181476444,-0.3330233991,-0.0759677738,0.081963487,0.0873983502,-0.3256025612,0.0608023405,0.0864466876,-0.4033417702,-0.222102046,0.0310033578,0.4001174271,-0.165881291,0.0495835766,-0.0012419367,-0.0924865454,0.0830143243,-0.224375084,0.1293615699,-0.0800678879,-0.0690841004,0.2346087247,-0.0784607232,-0.0848467797,0.078937456,0.1398114711,0.1245181486,0.1013177857,-0.0638325661,-0.1082671657,-0.0473794676,0.0398701578,0.3719714284,0.0728811026,-0.0798444301,0.1799773425,0.1176268309,0.3359162807,-0.1710351706,-0.2652395368,0.1846456081,0.4127750099,0.2804492712,0.03493727,-0.2378046066,-0.1935523152,0.047812894,0.0974794701,0.2963288426,-0.1378981024,0.3774010241,0.2653287947,-0.0540864021,0.1050586551,-0.0059436392,0.3334804773,0.4020539522,0.1033150628,-0.1396164894,0.097836785,0.4503250122,-0.1818425506,-0.0436583012,-0.1415113509,0.0294158012,0.2986090481,-0.2061505765,0.1806076765,0.3545348048,-0.1003586948,-0.056552697,-0.0440882072,0.1366898417,0.1603876054,-0.0848610476,-0.0618534647,-0.2673289776,-0.297369957,-0.1938901842,-0.0127563905,0.2254209965,-0.2387561053,0.5563297868,-0.0900093988,0.0111590223,-0.1981933117,0.5220184326,0.2117370516,0.0856912285,0.1101217717,0.0088752769,0.4086787701,-0.1446930319,0.3135301769,0.281953752,0.3434752524,0.1812379956,0.227287367,-0.0355274938,-0.2511698008,-0.2268112898,-0.0663600564,-0.0278465394,0.2472525686,-0.209269613,0.2070513815,0.3559572995,-0.1685532033,0.2728010714,-0.0133067286,-0.0123498794,-0.4368600249,-0.1232352182,-0.0990845188,-0.2485724241,0.1147099063,-0.0519135706,-0.0178159326,-0.3994470835,-0.000383508,0.0659300461,-0.0257204622,0.0020662891,0.1420654356,0.1367396861,0.261534363,-0.1357182711,0.1176768467,-0.1525500417,-0.2352029234,-0.1951711476,0.0709466636,0.1758554131,-0.1458296776,-0.2360496223,0.1795117855,0.2412057668,0.0473681465,0.4139926732,0.3235191107,0.3671124876,0.1350892931,-0.1947491467,0.1342658103,0.3285612762,-0.1465614736,0.0442721993,-0.258679688,0.2977811992,0.2315460294,0.1989944875,-0.0391249806,0.0357389711,-0.0948108584,-0.1561581641,0.1990955621,0.2619300783,0.1887194514,-0.2953293324,0.29363662,-0.095789291,-0.1535141468,-0.2890127003,0.291372776,0.0836669058,0.2285156399,-0.3356886208,-0.1344081461,0.1340022087,-0.0625590906,0.141674608,0.2946102917,-0.2629126906,0.5242142081,0.101231724,-0.2020090222,-0.2933758199,-0.14353095,0.0630900264,-0.058693897,-0.0489204079,-0.2605824769,0.3117169142,0.1425978392,-0.5548177958,-0.2099341005,0.0025965236,0.3300035,0.186989367,-0.1180755794,-0.1357410997,0.3360280097,0.0430969745,-0.210325703,0.2589012086,-0.2115746439,0.0006275388,-0.0748652369,0.0150379883,0.088133283,-0.2164473385,-0.0594494492,-0.4478659332]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/851","title":"Add support for other languages for rouge","comments":"@alexyalunin \r\n\r\nI did something similar for others languages.\r\n\r\n[Repo: rouge-metric](https:\/\/github.com\/m3hrdadfi\/rouge-metric)","body":"I calculate rouge with\r\n```\r\nfrom datasets import load_metric\r\nrouge = load_metric(\"rouge\")\r\nrouge_output = rouge.compute(predictions=['\u0442\u0435\u0441\u0442 \u0442\u0435\u0441\u0442 \u043f\u0440\u0438\u0432\u0435\u0442'], references=['\u0442\u0435\u0441\u0442 \u0442\u0435\u0441\u0442 \u043f\u043e\u043a\u0430'], rouge_types=[\r\n \"rouge2\"])[\"rouge2\"].mid\r\nprint(rouge_output)\r\n```\r\nthe result is\r\n`Score(precision=0.0, recall=0.0, fmeasure=0.0)`\r\nIt seems like the `rouge_score` library that this metric uses filters all non-alphanueric latin characters \r\nin `rouge_scorer\/tokenize.py` with `text = re.sub(r\"[^a-z0-9]+\", \" \", six.ensure_str(text))`.\r\nPlease add support for other languages. ","comment_length":10,"text":"Add support for other languages for rouge\nI calculate rouge with\r\n```\r\nfrom datasets import load_metric\r\nrouge = load_metric(\"rouge\")\r\nrouge_output = rouge.compute(predictions=['\u0442\u0435\u0441\u0442 \u0442\u0435\u0441\u0442 \u043f\u0440\u0438\u0432\u0435\u0442'], references=['\u0442\u0435\u0441\u0442 \u0442\u0435\u0441\u0442 \u043f\u043e\u043a\u0430'], rouge_types=[\r\n \"rouge2\"])[\"rouge2\"].mid\r\nprint(rouge_output)\r\n```\r\nthe result is\r\n`Score(precision=0.0, recall=0.0, fmeasure=0.0)`\r\nIt seems like the `rouge_score` library that this metric uses filters all non-alphanueric latin characters \r\nin `rouge_scorer\/tokenize.py` with `text = re.sub(r\"[^a-z0-9]+\", \" \", six.ensure_str(text))`.\r\nPlease add support for other languages. \n@alexyalunin \r\n\r\nI did something similar for others languages.\r\n\r\n[Repo: rouge-metric](https:\/\/github.com\/m3hrdadfi\/rouge-metric)","embeddings":[-0.1975666285,-0.1996216774,-0.0618753806,0.1323693395,0.3951304853,-0.3681798577,-0.2110817283,0.0119296852,-0.0544838905,0.4690480828,-0.2052647024,-0.0287141725,-0.0242667198,0.0488729887,0.0174612906,0.0532266349,0.0248597171,0.0744016841,0.3439333737,-0.3068408668,-0.1955282688,0.4024336636,0.0439972058,0.183745116,0.0121665187,0.2464933693,0.3911622465,-0.0277388208,-0.4155447781,-0.2775441408,-0.1769976467,0.0741563812,-0.0785336047,0.464209795,-0.0001092498,-0.2014050037,0.2390644848,-0.3072464466,0.0389284715,-0.440207094,0.0053247949,-0.0209817383,0.0712653771,0.0579086989,-0.0821590349,0.1259775758,-0.4007093012,-0.4302278757,0.218223691,0.5451992154,0.2035252601,0.095208548,-0.0212372933,-0.0208682138,0.2856881917,-0.0776511207,-0.0697572902,0.3765059114,0.1493060291,0.1755555272,0.0687481537,0.3037921786,0.1223899275,-0.2009081692,-0.1292490512,-0.0354446135,0.2640488148,-0.2034990788,-0.0025658328,0.2447784245,0.184852913,-0.1724867523,-0.4331069291,0.2365184128,0.1076366454,-0.7995042205,-0.0817373693,-0.0995959044,-0.0398950316,-0.1187453866,0.0050444608,0.2691437602,-0.0395534933,0.0036241154,-0.1162126884,0.1588865966,-0.122290954,-0.0532621369,0.0554723889,0.01126575,-0.3704309464,0.2188408375,-0.1317212433,0.1577842832,-0.4892687201,0.0378112532,-0.0412859581,0.2219455391,0.2480162829,0.0273516178,0.3466551602,0.4280377626,0.0005559556,0.0877345726,-0.2989966273,0.5079794526,0.1249455512,0.1890327185,0.3438434601,-0.0492251813,0.0069011794,-0.1512029618,0.3055439293,-0.292404145,0.0862295926,0.2419241667,-0.1788023859,-0.3959760964,-0.6018204093,0.1168890223,-0.2879897356,-0.1139950156,0.1281738728,0.2527341545,-0.1142799631,0.4247303903,-0.0651597977,0.2039280087,-0.1871563196,-0.1841606647,-0.2318840772,0.0188032091,-0.269185394,0.1315092593,-0.009236238,0.0669622868,0.0595875792,-0.1147663519,0.5793462992,-0.0499637499,0.1193769723,-0.3021790385,0.1222472191,-0.1107287779,-0.205411762,0.0156531446,0.2732788026,-0.4688226581,-0.0128545184,-0.2419839948,-0.1985593438,-0.0072813206,-0.0204627048,0.237384975,-0.0170887504,-0.044593595,-0.1200552732,0.2201262861,0.0117405672,-0.09848205,0.2534069717,-0.1302761883,-0.1629296541,-0.0890071839,0.405564338,-0.1022474393,-0.3046510816,-0.0753154308,0.3594304025,-0.0181679074,0.5161855221,0.3134860098,0.3155922592,-0.02707633,-0.1300076991,-0.2198492438,0.5051718354,-0.6645435691,-0.167777285,-0.2900456786,0.0390516184,-0.0440725647,0.1601073444,-0.1530565768,-0.057223741,0.0324388966,0.2193163335,0.0715662986,-0.0701984391,-0.1903066933,-0.2069900334,0.2397997975,0.391417861,0.1456720382,0.258099556,-0.1370952874,0.0235938691,-0.2760108411,0.2315428853,-0.1321610808,-0.0257633217,0.0135537004,0.1062324941,0.2308056504,0.1926734,-0.2618959546,0.4493220747,0.1448684335,-0.7048096061,0.4194575548,-0.1444995701,-0.0764315426,-0.3683908284,-0.0954541117,-0.0892067626,-0.2053843588,0.2087349743,0.1267928928,0.15487881,0.5415862799,-0.0209742524,0.119030267,-0.1187605485,-0.0078776935,-0.1455576271,-0.2077372372,-0.1280347556,-0.1368233413,0.0129870046,0.6449344754,0.2656037807,0.1522060633,-0.2444246262,0.2478715181,0.1986221373,-0.151444748,0.0508069955,0.1260529011,0.139705345,0.0793660134,-0.0919397473,0.6127666235,0.363452673,-0.0101814391,0.1158078313,0.5160871744,-0.3609585464,0.0793743506,-0.0558519699,-0.1030010507,0.2650699317,0.2836160064,-0.2701507509,-0.2781109512,0.489349395,-0.0931820571,0.3736890852,-0.057383474,-0.2104810923,-0.4544012845,0.4172363877,0.0192860644,0.3348826766,0.1364224255,0.0937930942,0.0390118696,-0.0677286088,-0.2031253278,0.0250885561,0.1029624641,0.0900532529,-0.142136693,0.0440219827,-0.2969490588,-0.0140441116,-0.1887357086,-0.1117651165,0.212662816,0.2226190865,0.034207657,-0.4290671051,-0.1572298259,-0.3795201778,-0.0189395584,-0.077033326,0.0473708436,-0.1510138363,0.2671286464,-0.0847673863,-0.3527912199,-0.3352108896,-0.2481893748,0.2217553854,-0.2667916417,0.1777866334,0.0872376934,0.121478945,0.0954322517,0.0061583854,-0.1476211995,-0.1464134604,-0.214355126,-0.2266206741,0.1450522393,-0.2689492702,-0.1138906181,0.3191297948,-0.4624062777,0.0266751908,0.2263182253,-0.3188216984,0.0963369384,-0.1537021697,-0.1105794013,0.1372851133,-0.3897536099,-0.2152041048,0.3034488261,0.3806331754,-0.3338530064,-0.1820965856,0.4587659538,-0.109372668,0.0481311455,0.0129303066,-0.0852119848,-0.2527599037,-0.2325787842,0.1091336682,0.162351504,-0.2286106199,0.0279093813,-0.1064244285,0.1455371827,0.0599427037,0.2569841146,-0.2288132459,-0.1780769825,0.4847200811,-0.2373673916,-0.417054981,0.0196039956,0.1314415485,0.1675889343,0.0294872206,0.1347781569,-0.2530069649,0.0115926769,-0.0785633773,-0.1138659939,0.1590440422,0.5170838833,-0.1263070107,-0.1043065414,-0.1111763194,-0.1088801697,-0.2234170735,-0.1353065223,0.1945600361,-0.3147878945,0.4256430268,0.1111440063,0.1845232248,0.0490930155,0.1862823516,0.2725143731,0.0604293793,0.2467903793,-0.2542189062,-0.085050188,0.3745290935,0.0022255094,-0.1934871823,0.1124425083,-0.0797633305,0.5023851991,-0.2670534253,-0.0339830182,-0.1076892167,-0.018684959,-0.0144061428,-0.3469944596,0.1185861453,-0.1066944972,0.0960224941,-0.0039601889,-0.1401072443,0.282772392,-0.0478205942,0.1479985118,0.036601264,-0.6129229665,0.108471401,-0.3170462251,0.4604831934,0.080975689,0.3339834809,-0.027549291,0.0002640889,-0.0852389634,-0.0593780354,0.240359813,-0.3082799911,-0.0273791105,0.0850267559,-0.3454485834,-0.5695742965,0.0178238265,-0.2557504773,0.0666152686,-0.1082455292,-0.0018944093,-0.185272783,0.0179219786,-0.1906075776,0.2635688782,-0.1175922826,0.161916852,-0.5067321658,0.0284669641,-0.0477447696,0.2392520756,0.1028550714,0.4289150834,0.0940821543,0.0335679539,-0.1120951846,-0.1008762941,0.2903498411,0.2952880859,0.1897213757,-0.1713788211,-0.219705224,-0.0455469191,-0.2577818334,-0.2592520118,0.2834260464,-0.0188425519,-1.0625962019,-0.0648350492,-0.195430696,0.3173086643,-0.0246498398,-0.1060158908,-0.0279875118,-0.4211496115,0.0564444773,-0.2418707758,0.0847592503,0.2004930228,0.1969815046,-0.0214236826,-0.1015521288,0.0840758979,0.1715693921,-0.0702102557,0.0817709938,0.1047332734,-0.2958290577,0.5322651863,0.2290433943,0.7910509706,-0.0379863903,-0.0365457125,-0.0657203048,-0.062351577,0.1357663423,-0.3338634968,0.0428878963,-0.3183025122,-0.0922372118,-0.1185459718,0.1085467562,0.0473483503,0.0024215481,-0.089836292,0.37751472,-0.0480040051,0.2147511095,0.1238911301,0.1698364764,0.0530607887,-0.3393788934,-0.1808009148,0.195829168,-0.0361091606,-0.138395384,-0.0718754157,0.1940261126,-0.2068691999,-0.1910692602,-0.0816207752,-0.0778215602,-0.4477725923,-0.0955840796,-0.0252746176,-0.6008175015,-0.1413097084,0.5418083072,0.6607747674,0.3075968027,0.044371821,0.1414561868,-0.1531281471,0.2506037951,0.0484588705,0.0175871756,-0.1973829716,-0.0751694962,-0.1492000222,0.1715042889,0.1428431422,-0.166340068,-0.0922758207,0.0823371932,0.1064516827,0.0145158432,0.1288437396,-0.3394491971,0.088268429,-0.234861359,0.1583710313,0.2935744524,-0.318784833,-0.0288248435,-0.0202019159,-0.2216687351,-0.1209074855,0.1186444089,-0.0813441798,-0.3281901181,0.0649512187,-0.0055970135,-0.0572819747,-0.2401269525,-0.0912197828,0.1201633587,-0.2057271451,0.1298383623,0.1380511373,-0.3015813529,-0.0622074343,0.5735049844,0.1255970597,-0.165979743,0.2185986638,-0.4442996979,-0.2323514819,0.0721813217,0.269061029,0.0134407338,0.1324671954,-0.1457728595,0.0261458941,0.2935348153,-0.3694892526,-0.258382827,-0.273252517,0.1525232792,-0.2013515979,-0.072402291,0.0879817754,0.1553324908,0.1626900584,0.3535944223,-0.0910130292,-0.1985587776,-0.156080097,0.078504473,0.0320286565,0.3314733803,0.1748376489,0.25274086,0.0454202779,-0.3798587024,0.0209107678,0.5152859688,0.0887020379,0.5211655498,0.051667165,0.2445174903,0.0198700894,-0.278416723,0.1219739541,0.1068126708,0.0691035911,-0.0541930534,0.1683248729,-0.0663522258,-0.0441120379,0.0526834913,0.0489781946,0.1107495055,0.059142407,-0.1153851897,0.0761439949,0.0653905123,0.2958738208,0.2116316557,-0.0903511047,-0.2737523615,0.0414002948,0.174375385,-0.3215384483,-0.1068115011,-0.0949797407,0.0376204289,0.1498739868,0.3660729229,0.1189887971,-0.1977428496,0.2220004648,-0.0637873188,-0.0765317604,-0.2729205191,0.2483175695,-0.3020533025,0.0324030071,0.0458153076,0.3567244411,-0.1413780898,0.0158974007,0.4557009339,0.0051218807,0.3973891139,-0.0977627486,0.3166027665,-0.1116759032,0.2679521739,-0.0496335141,0.0984447226,-0.0469038002,0.0046953601,0.117970854,0.0246097706,0.0531760342,0.0136702936,-0.0466562621,-0.3870701492,-0.0393778868,-0.0879193246,-0.2264567465,-0.1276361346,-0.1538048089,-0.2229202837,-0.0098734945,0.2636827528,0.0608066805,0.0897722617,-0.0088046659,-0.1008595601,-0.3344277143,0.0244045593,0.3619347513,-0.138747707,0.1349378228,0.3402404189,0.3166129291,0.1068567559,0.1234064549,0.1588608027,0.3081885576,-0.3514321744,0.2110801339,-0.3409923911,-0.0804751962,0.0371652134,0.472440809,0.3296273351,-0.0731164664,0.1959508061,0.1633362919,-0.1723304689,0.1296252906,0.0478284545,0.1229501814,-0.0961186588,0.3986921012,0.1737415195,0.2071497589,-0.5398373604,0.0507044941,-0.3321765065,0.0941691399,-0.3074989021,0.1959497482,-0.0490147397,-0.3666323721,0.1211956963,0.3527683914,0.1619022191,0.0391130112,0.1953913569,-0.270969063,-0.0571050011,-0.4009109735,0.1845006198,0.1242390499,0.0799296349,-0.3103607595,0.1453627944,0.2083415836,0.4607051313,0.0605839305,0.5777207613,-0.0711152181,0.0770509169,-0.1541181058,0.1447992921,-0.2478025109,-0.0084939133,0.078426823,-0.0243899189,-0.0921005234,-0.2238451093,0.18694143,-0.1736181527,-0.2770746052,-0.0139221502,0.4028490782,0.4931404889,0.459692806,0.3084006608,-0.2260137051,-0.0044139796,0.1009266227,-0.1670191586,-0.243208468,0.4605926573,0.0315325223,0.2264266163,0.0032170513,-0.2348322421,-0.2389853299,0.6837211847,-0.0561465807,-0.2158748806,-0.1027358696,0.2263979316,0.4558701813,-0.2406397015,-0.342371136,0.2849819064,0.0052873543,0.1242789403,-0.2443761677,-0.2648325562,0.235898599,-0.1594115198,-0.1784211844,0.0204137415,0.2867368162,0.2807131708,-0.1557201445,-0.5182507038,0.2579732239,0.1795469224,0.2588748038,-0.0879395679,0.0803226158,0.1072661653,0.0448458083,-0.1447995156,0.2863199413,0.1919552386,0.2517579198,0.2032451779,-0.3782761991]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/849","title":"Load amazon dataset","comments":"Thanks for reporting !\r\nWe plan to show information about the different configs of the datasets on the website, with the corresponding `load_dataset` calls.\r\n\r\nAlso I think the bullet points formatting has been fixed","body":"Hi,\r\nI was going through amazon_us_reviews dataset and found that example API usage given on website is different from the API usage while loading dataset. \r\n\r\nEg. what API usage is on the [website](https:\/\/huggingface.co\/datasets\/amazon_us_reviews) \r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"amazon_us_reviews\")\r\n```\r\nHow it is when I tried (the error generated does point me to the right direction though)\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"amazon_us_reviews\", 'Books_v1_00')\r\n``` \r\nAlso, there is some issue with formatting as it's not showing bullet list in description with new line. Can I work on it?","comment_length":34,"text":"Load amazon dataset\nHi,\r\nI was going through amazon_us_reviews dataset and found that example API usage given on website is different from the API usage while loading dataset. \r\n\r\nEg. what API usage is on the [website](https:\/\/huggingface.co\/datasets\/amazon_us_reviews) \r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"amazon_us_reviews\")\r\n```\r\nHow it is when I tried (the error generated does point me to the right direction though)\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset(\"amazon_us_reviews\", 'Books_v1_00')\r\n``` \r\nAlso, there is some issue with formatting as it's not showing bullet list in description with new line. Can I work on it?\nThanks for reporting !\r\nWe plan to show information about the different configs of the datasets on the website, with the corresponding `load_dataset` calls.\r\n\r\nAlso I think the bullet points formatting has been fixed","embeddings":[-0.0789317265,-0.1892752051,-0.2111489326,0.5182409286,0.1775332242,0.2812187374,0.2807130814,0.1095822901,0.1075348184,-0.2891573012,-0.1656707525,0.1336098462,0.3731780946,0.3808050156,0.1173837855,-0.0786431283,-0.0114683844,0.0120945619,-0.0104108388,0.0372485481,-0.0519340113,0.3416516781,-0.4777825177,-0.1001210883,0.0012968603,-0.1514359862,0.0161404442,0.0154884551,-0.0385072865,-0.3352609277,0.2822167873,0.1228939518,0.2670375109,0.4127134681,-0.000120029,-0.1779257655,0.4665377736,-0.1020854488,-0.2560844719,-0.2737341821,-0.2082238495,-0.3046475351,0.4594644904,-0.3098630607,-0.0713830516,-0.5172998905,-0.3546373248,-0.262871176,0.323926121,0.2284378558,0.1925453097,0.1057268605,-0.0450246073,-0.1168593541,0.0473390706,0.2259967178,0.1323072016,0.1022474021,0.2680190206,0.151785329,-0.0303034801,0.2312604487,0.0532070957,0.2929938138,0.3227801621,-0.0937235281,-0.16267021,-0.1474439055,0.4255650342,0.2147620618,1.3290979862,-0.2534140348,-0.2807647288,-0.1864213943,-0.2302003056,-0.1689096987,-0.016869057,0.3455489874,-0.007727148,0.0436370149,-0.2335562557,-0.2231645286,-0.186931625,0.3701874614,0.059278477,0.063065514,-0.3054240942,0.1380738467,-0.1763880998,-0.1328852624,0.1571980864,-0.1911980659,0.0317064747,0.1797028482,-0.5834522247,0.0873515531,0.0808451697,0.1540183574,0.06331826,0.4426725507,0.3057207763,0.2691638768,-0.0215794686,0.1393393576,0.5947414041,0.2712326646,0.3216514289,0.0154353594,-0.0527829342,0.3021752536,-0.0381989889,-0.1961116791,-0.0054876166,-0.1151610389,0.097782284,-0.2972743511,0.237949118,-0.5695005059,-0.2049052119,-0.0927838236,-0.2847169638,-0.3555089235,0.0098460838,0.4803138971,-0.1018202677,0.1047831625,0.2056426853,0.2912029028,-0.1113749444,-0.1213556975,-0.1241604239,-0.0199554022,-0.2060885578,0.0916363001,-0.0304607861,-0.0041035721,0.358540833,0.0954123884,-0.001444246,-0.3614833057,0.3334643543,-0.1039520055,0.0891807377,0.0643302128,0.4270684123,0.2381878495,0.1438618004,0.1221372634,-0.2038876414,0.270342797,-0.3434373736,-0.2165079266,-0.183988899,0.1285462826,-0.3012284636,-0.30059129,-0.5243891478,0.1205802262,0.1504328996,-0.1128383353,-0.0485668294,0.1281505376,-0.1047201306,-0.1302075237,0.0923537239,0.4687439799,-0.4306244254,-0.2646574676,-0.3125124574,-0.1004269049,0.2145790309,0.2239419222,-0.3063053787,0.354963094,-0.1731179953,0.0351228304,0.559214294,-0.3355660141,-0.1034140214,0.4995976686,0.216415599,-0.0355051793,0.0961955935,-0.2372708172,-0.1126170531,0.1777271032,0.1222919226,0.2151847035,0.3439027667,0.127610296,-0.2076399773,-0.2062740177,0.2997550368,0.1547664702,-0.1266524643,-0.0012366591,-0.0236076526,-0.1026685312,0.3608762026,-0.0464179069,-0.1341619045,0.3027929366,-0.0695767403,0.2714541554,0.1779069304,0.1316058636,-0.4466254711,-0.0358647406,0.000468363,0.1481631696,0.1683010608,-0.1922745109,-0.5272638798,0.0090979794,-0.1991000026,-0.1610389799,0.0777426958,0.0320809744,0.0412667245,-0.0335679464,-0.0199565012,-0.0856475905,0.0163855385,0.0283043571,-0.4079386592,0.165841192,0.0833528191,0.0614559948,0.2739598155,0.0993565172,0.2622644603,-0.011710952,-0.1645417064,0.2622707188,0.0659875572,0.6024624705,0.151269421,0.302249521,0.0752668306,0.0555782318,0.2502038777,0.277800709,0.3070844412,-0.1422146708,-0.3737635016,0.3783277571,0.0334415659,0.4875419438,0.1482855678,-0.1668079346,0.2705942392,0.0136275319,-0.0556318685,-0.2422356755,0.2620974779,0.0789038315,0.6111764312,-0.1333584934,-0.1279371083,0.2783301771,0.1441027522,-0.0556796454,0.0739202648,0.0867314711,-0.2402981073,-0.4528817832,0.5009841919,-0.0215052739,0.2161624581,0.2434825152,0.0983895957,0.1076892763,0.2769545317,-0.2266352177,0.302154839,0.3179554045,0.2120098323,-0.1057116687,0.1732845455,-0.166939795,-0.1414024532,-0.1537816226,0.0471223928,0.2280318588,-0.5924049616,0.0342507362,-0.3620118201,-0.4074659646,-0.3450549841,-0.131496191,0.0816858485,-0.3198181093,-0.2920684218,-0.0485499203,-0.04713016,0.0176134277,-0.1387409121,0.0507323742,-0.1904655844,0.054074578,-0.2068430781,-0.2030698508,-0.1652269661,0.0597145557,0.473749429,0.2471573204,0.3921639919,-0.1970732659,0.092662245,0.0477438457,-0.2660235465,0.2005766779,0.1971172392,0.1751429141,-0.1349904239,0.055997923,0.1437902302,0.1499388069,0.1842006296,0.119969666,0.1417253464,-0.0464305691,-0.1063116565,-0.1500596702,0.1808391362,-0.5233334303,-0.1948813796,-0.3031464517,-0.0537587479,0.2007278651,0.2991801798,-0.1856728494,-0.0715600774,0.1559025347,0.3398942351,0.2962896526,-0.3032110631,-0.4763090611,0.161378935,-0.2892211676,-0.275064826,0.1589958966,0.3177672625,0.242174089,0.0612025522,-0.2267962992,-0.0413053893,-0.0851614624,-0.1112039015,-0.0724887326,0.1672209352,0.2617268264,0.0364038274,-0.1051956266,0.0054173698,-0.2416447401,-0.0716852769,-0.1896223277,0.0281640124,-0.0354814641,0.1203640103,-0.1563309282,0.5662083626,0.255155921,-0.0545405857,0.3984052241,-0.103493534,0.1799929589,-0.0499388427,-0.0158129949,-0.3308517337,-0.1166108623,0.004270799,0.4625917673,0.1697825789,-0.2965262234,-0.2413651794,-0.1513515115,-0.8929014206,-0.1086798236,0.1384853125,-0.1324648261,0.2511935532,0.1188588142,0.0773566514,-0.0727128908,-0.101469487,-0.0233823266,0.088534534,-0.0062200679,-0.1664228439,-0.4538201392,-0.101599656,-0.1419453621,0.5074074268,-0.2162258476,0.1196993068,0.0532301143,-0.0910085142,0.1657986194,-0.2087830603,0.4329097271,-0.3500380516,-0.1027416438,0.1852796078,0.0078395838,-0.3345591128,0.0840518251,-0.2710379362,0.4508850276,0.1022003889,0.0443188176,-0.1706228554,-0.2756689489,-0.1051070541,0.3202067018,-0.2772405446,-0.2040268332,-0.0663655326,-0.2064158171,-0.3689192235,-0.3099520206,-0.2958849967,0.1530736983,-0.0443872102,-0.0725075155,-0.0855379999,0.0618659668,0.0961783454,0.2213361412,0.4586628079,-0.1938783675,-0.077220872,0.1081203073,-0.1309430897,0.4480089247,0.7521959543,-0.0088666761,-0.4299919605,-0.0108978534,-0.1588894278,0.1221937016,0.1990404725,0.0710833222,0.1092334017,0.0085381046,-0.0651765764,-0.1920917928,-0.2146843076,0.1032841131,0.1821847558,-0.1044165641,-0.6993977427,0.6753956676,0.0755910575,-0.1975606233,0.2609327137,-0.0029875452,-0.214088425,0.2761699259,0.0100667141,1.1095734835,0.1581323892,-0.0145885572,0.1266873181,-0.2598596215,0.2006963342,-0.1367522478,-0.1212267578,-0.4516961575,0.3673309982,-0.0736472532,-0.1184357852,0.1955995113,0.2501854599,-0.1443833858,0.3782751262,-0.2910306752,-0.0442530327,-0.0205134135,0.4625371695,-0.0387585387,-0.2825290263,-0.6071410775,0.0899346173,-0.0510194041,0.1447156519,-0.2772305012,0.089092575,-0.2481803447,-0.091366142,-0.3456159234,0.0245125331,-0.0853706524,0.0931589678,-0.1741664112,-0.4508790672,0.1298786104,0.3845647573,-0.2622558475,-0.0901250616,-0.4244490564,0.1190184131,-0.0972173885,0.0301195402,0.035664767,0.0210345741,0.371547699,-0.1616903394,-0.3018017411,0.0079999659,0.0854749531,0.062539883,-0.1689022034,0.0588683933,-0.0031789185,-0.5724694133,-0.5873054266,0.1221195906,0.0422567166,-0.1212282255,0.0923035964,0.1828105897,-0.1142607331,0.2109425664,0.2463114709,0.0369306616,-0.0788510889,0.4332987666,0.3101987541,-0.0244959723,0.4668748081,0.2951889634,-0.1460087001,-0.1963519007,0.1923954934,0.1957504898,-0.2520182133,0.0509447269,-0.0368923806,0.3241746724,-0.2070686817,0.1904633045,0.0188210756,0.2794863582,-0.0220122356,-0.4668632448,0.2508147955,0.2537477911,0.193042472,-0.0113745332,-0.1099514514,-0.3284414411,0.1592364013,0.0505198762,-0.2618047893,0.321079731,-0.0695698783,0.0781370476,0.1321983337,0.1443041265,0.1451752186,-0.2790772319,-0.0119590992,-0.1267284751,-0.1100216731,-0.1981800199,-0.1354358494,0.2127933651,0.1015968695,-0.2201015502,-0.0115799401,0.008608724,0.1161865592,-0.1793661416,0.0562731586,-0.0937526375,0.0999777615,0.0836096555,0.3907763958,0.0669819415,-0.6115540266,0.0721727014,-0.0410912633,0.1988336295,0.241068542,-0.0440358333,0.3422813118,-0.0221054386,-0.0257915501,-0.0306738634,0.2908673286,0.2202176899,0.1947539449,-0.0818215311,0.2536888123,0.2407731116,0.3208608925,0.5258144736,-0.1758219004,0.1646322459,-0.0215727519,0.142804563,-0.2438157648,0.0389110669,-0.1467023045,-0.2376057953,-0.183507964,0.0038751627,0.1829794049,0.3325914741,0.1130231991,-0.120325245,0.4508987963,-0.2245201617,0.4222685993,0.0590375662,-0.1689968705,0.145917803,0.1762697995,0.1485859454,-0.0028411336,0.0183643475,0.0689573362,0.5400498509,0.4455858171,0.1596898288,-0.2230799794,-0.5270540714,0.3603376448,0.1754484475,0.3199091852,0.2928749323,0.0271711387,0.4435742497,0.1476008594,0.0913557187,-0.0670745522,0.0020864711,-0.0573351458,-0.2027501464,-0.1516113579,-0.2909406126,-0.1086124778,-0.1345993578,-0.20566383,-0.4516636133,0.054415185,-0.1240047067,-0.1323204041,-0.5200738311,0.0575002357,-0.0345600843,0.1957113445,-0.1927078366,0.0774271265,0.168881014,-0.0196263883,0.0776727572,0.2913623154,0.5396184921,0.1511488408,-0.102262415,-0.0486478433,0.0775669217,-0.1285819411,0.0892522112,0.1390936971,0.2173725069,0.0814578608,0.1464356333,0.1171539947,-0.1644436568,0.366730094,0.1373812407,-0.0216105245,-0.2827694416,0.2542513013,-0.3644189835,-0.3706579208,-0.2655566037,-0.1761494279,-0.4030531347,-0.0331169777,0.2407069951,-0.1102062166,-0.1653468609,-0.1996424049,0.0526959077,0.0668963194,0.6728174686,0.0392680317,0.3681770861,-0.0654766113,-0.062787585,-0.6017756462,-0.0151448427,-0.0510748476,-0.1354796141,0.1072410494,0.193096742,0.1050080061,0.2610845268,0.173890233,-0.1343341023,0.1025074124,0.0447393656,-0.4506543577,0.0176677704,0.019242147,0.0695834979,-0.1925379783,-0.2314060181,0.0460730977,-0.2032757103,-0.102360487,-0.3747596145,-0.2316420823,0.0956798121,0.0854067057,0.6618387103,-0.0471609011,0.5717574954,-0.2162524313,-0.0034676846,-0.0478658453,-0.2418791354,-0.0051128697,0.6518115401,0.0552234165,0.0082456628,-0.1022308171,-0.2083972245,-0.0936172456,0.101900667,-0.2653567195,-0.2419253141,-0.0490176827,-0.2247557342,-0.0065895547,-0.2708671093,0.0671397299,0.2326692641,-0.0741141364,0.0605399013,-0.1462936848,-0.1081118882,0.4086982906,-0.0487561747,-0.0113418158,0.0078062401,0.1532592773,0.0871675685,-0.180296734,-0.6323314905,0.1697620898,0.5076562762,-0.0154825328,0.1548660994,0.4629185796,0.0157343745,0.0307730604,-0.1776595414,0.3021737337,0.0158303697,-0.2255867273,-0.3437719345,-0.2972668409]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/848","title":"Error when concatenate_datasets","comments":"As you can see in the error the test checks if `indices_mappings_in_memory` is True or not, which is different from the test you do in your script. In a dataset, both the data and the indices mapping can be either on disk or in memory.\r\n\r\nThe indices mapping correspond to a mapping on top of the data table that is used to re-order\/select a sample of the original data table. For example if you do `dataset.train_test_split`, then the resulting train and test datasets will have both an indices mapping to tell which examples are in train and which ones in test.\r\n\r\nBefore saving your datasets on disk, you should call `dataset.flatten_indices()` to remove the indices mapping. It should fix your issue. Under the hood it will create a new data table using the indices mapping. The new data table is going to be a subset of the old one (for example taking only the test set examples), and since the indices mapping will be gone you'll be able to concatenate your datasets.\r\n","body":"Hello, when I concatenate two dataset loading from disk, I encountered a problem:\r\n```\r\ntest_dataset = load_from_disk('data\/test_dataset')\r\ntrn_dataset = load_from_disk('data\/train_dataset')\r\n\r\ntrain_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n```\r\nAnd it reported ValueError blow:\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-38-74fa525512ca> in <module>\r\n----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n\r\n\/opt\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in concatenate_datasets(dsets, info, split)\r\n 2547 \"However datasets' indices {} come from memory and datasets' indices {} come from disk.\".format(\r\n 2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],\r\n-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],\r\n 2550 )\r\n 2551 )\r\n\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\n\r\nBut it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:\r\n```\r\ntrn_dataset._data_files\r\n# output\r\n[{'filename': 'data\/train_dataset\/csv-train.arrow', 'skip': 0, 'take': 593264}]\r\n\r\ntest_dataset._data_files\r\n# output\r\n[{'filename': 'data\/test_dataset\/csv-test.arrow', 'skip': 0, 'take': 424383}]\r\n\r\nprint([not dset._data_files for dset in [trn_dataset, test_dataset]])\r\n# [False, False]\r\n\r\n# And I tested the code the same as arrow_dataset, but nothing happened\r\ndsets = [trn_dataset, test_dataset]\r\ndsets_in_memory = [not dset._data_files for dset in dsets]\r\nif any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):\r\n raise ValueError(\r\n \"Datasets should ALL come from memory, or should ALL come from disk.\\n\"\r\n \"However datasets {} come from memory and datasets {} come from disk.\".format(\r\n [i for i in range(len(dsets)) if dsets_in_memory[i]],\r\n [i for i in range(len(dsets)) if not dsets_in_memory[i]],\r\n )\r\n )\r\n```\r\n\r\nAny suggestions would be greatly appreciated! \r\nThanks!","comment_length":172,"text":"Error when concatenate_datasets\nHello, when I concatenate two dataset loading from disk, I encountered a problem:\r\n```\r\ntest_dataset = load_from_disk('data\/test_dataset')\r\ntrn_dataset = load_from_disk('data\/train_dataset')\r\n\r\ntrain_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n```\r\nAnd it reported ValueError blow:\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-38-74fa525512ca> in <module>\r\n----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n\r\n\/opt\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in concatenate_datasets(dsets, info, split)\r\n 2547 \"However datasets' indices {} come from memory and datasets' indices {} come from disk.\".format(\r\n 2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],\r\n-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],\r\n 2550 )\r\n 2551 )\r\n\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\n\r\nBut it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:\r\n```\r\ntrn_dataset._data_files\r\n# output\r\n[{'filename': 'data\/train_dataset\/csv-train.arrow', 'skip': 0, 'take': 593264}]\r\n\r\ntest_dataset._data_files\r\n# output\r\n[{'filename': 'data\/test_dataset\/csv-test.arrow', 'skip': 0, 'take': 424383}]\r\n\r\nprint([not dset._data_files for dset in [trn_dataset, test_dataset]])\r\n# [False, False]\r\n\r\n# And I tested the code the same as arrow_dataset, but nothing happened\r\ndsets = [trn_dataset, test_dataset]\r\ndsets_in_memory = [not dset._data_files for dset in dsets]\r\nif any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):\r\n raise ValueError(\r\n \"Datasets should ALL come from memory, or should ALL come from disk.\\n\"\r\n \"However datasets {} come from memory and datasets {} come from disk.\".format(\r\n [i for i in range(len(dsets)) if dsets_in_memory[i]],\r\n [i for i in range(len(dsets)) if not dsets_in_memory[i]],\r\n )\r\n )\r\n```\r\n\r\nAny suggestions would be greatly appreciated! \r\nThanks!\nAs you can see in the error the test checks if `indices_mappings_in_memory` is True or not, which is different from the test you do in your script. In a dataset, both the data and the indices mapping can be either on disk or in memory.\r\n\r\nThe indices mapping correspond to a mapping on top of the data table that is used to re-order\/select a sample of the original data table. For example if you do `dataset.train_test_split`, then the resulting train and test datasets will have both an indices mapping to tell which examples are in train and which ones in test.\r\n\r\nBefore saving your datasets on disk, you should call `dataset.flatten_indices()` to remove the indices mapping. It should fix your issue. Under the hood it will create a new data table using the indices mapping. The new data table is going to be a subset of the old one (for example taking only the test set examples), and since the indices mapping will be gone you'll be able to concatenate your datasets.\r\n","embeddings":[-0.0911899135,-0.1483132094,-0.0447487496,0.6809360385,0.1068756282,0.2218791395,0.3129051626,0.2291087657,-0.1234838143,0.1246045008,-0.1083224565,0.2736025453,-0.0816895962,-0.1234090403,-0.30136621,-0.0974017531,0.1984925121,0.1485707313,-0.5059524179,-0.031675972,-0.3506394327,0.1602455676,-0.3358776569,0.0398673192,-0.4097848237,0.049245391,-0.1317905039,0.2761821151,-0.0312033258,-0.1838645488,0.4585869312,-0.3488870859,0.1864263117,0.5364499688,-0.0001176379,0.1700864136,0.2740953863,-0.1723825485,-0.3783839941,-0.256719321,-0.3621298075,-0.1004148722,-0.0029407758,-0.1234583333,0.207650885,-0.1481209397,-0.2583855689,-0.5043874979,0.33071509,0.2264483124,0.1104794741,0.3941425681,0.0912052542,-0.0980597213,0.0378851555,0.1533982009,0.0659453794,0.4203861058,-0.1898589581,-0.0617092587,0.2108293176,0.0670610666,-0.1246672496,0.1450287402,0.0780371949,0.3099108636,0.1576000452,-0.3968754411,-0.1252778769,0.1079114303,0.525313437,-0.3321693838,-0.3178471923,-0.0861591324,0.1388193816,-0.3995229304,0.217888236,0.2979132831,-0.147090584,0.1174623221,-0.0907389298,0.0771459118,-0.2191733122,0.1846294105,-0.1806519479,0.114519842,-0.1610969752,0.2407015562,0.138170585,-0.0730411187,0.5286436677,-0.3571856916,-0.1269245148,0.1591957361,-0.3422968984,0.0321878828,-0.1896209419,-0.704597652,0.0890091434,0.055820778,0.1037015319,-0.2204585522,0.0773697644,0.2818921804,0.4017777145,0.1492893547,-0.037381921,0.5497024655,-0.0202757102,-0.0616969094,0.199370116,0.009052719,-0.2867643833,-0.2114093006,0.1154302359,-0.1646887511,0.0836554542,-0.0013334746,-0.4971266091,-0.1518924385,-0.1278606951,-0.1404941082,0.1965599209,0.0301942825,0.123899661,0.5308156013,0.130738169,0.4409759939,0.217348516,0.0174215604,-0.226509124,0.0222139601,-0.0451169945,0.2104356736,0.103413932,0.0044444473,0.0821225643,0.1775865853,0.1087558568,0.006450789,0.0768852085,-0.4948595166,0.0583267137,0.3114701509,0.0211642738,0.0750039294,0.2738741934,-0.0503142029,-0.1439029574,0.3170154095,-0.4437072873,-0.2742709517,-0.513563931,0.0968817398,0.0989558399,0.2329158783,-0.3203627765,-0.1113318652,0.6928232908,0.1045668423,-0.0683576986,-0.0662886873,-0.1984153092,-0.3195584714,0.3204246461,0.2007118911,-0.2164676934,-0.0330451019,-0.0184931085,-0.0677970722,0.3604027629,0.5054668188,-0.3153637946,0.1992003024,-0.3855018318,-0.0331822969,0.3084054589,-0.4366770387,-0.1402199119,0.2154022306,0.0032159954,-0.073586531,0.2376121879,0.0082833301,0.1314229518,0.1472160071,0.8280869722,0.3132544458,0.0639004782,-0.1781871915,-0.097612448,-0.1383359432,0.2659764588,0.1000099033,-0.2323941588,0.0350794122,-0.1362989247,-0.4177925885,0.3237215579,-0.1016945988,0.0396771654,0.3689118922,0.0351548083,-0.343180269,-0.1154859886,0.0461800434,-0.1954729706,0.0483524054,0.0379616804,-0.0852145106,-0.2998000979,0.0915440768,-0.1228940859,0.0769214854,-0.072335422,0.0498545468,0.0176120866,0.235278517,-0.1725991517,-0.2480318099,-0.1653319597,0.4237135053,-0.2777830064,-0.026918387,-0.2073603868,0.5686868429,-0.0645174757,-0.0320977345,-0.1763465405,-0.0393736251,0.0589343719,0.1164020374,-0.0541011542,0.5149708986,0.0520663075,-0.1684773117,0.025941968,-0.2057821304,0.2290216684,-0.0772498548,0.139911443,-0.1201847419,0.1309275776,-0.0969888195,0.0723317713,0.4227220416,-0.1323702931,0.2828258872,-0.0215705317,0.0900893733,0.1695275009,0.1452977657,-0.058325775,-0.4177126288,-0.0745221898,0.1675421298,0.2188819051,0.2504171431,-0.4655806124,-0.2347148806,0.338824898,0.0246771201,0.0422085784,0.1224695444,-0.3912708461,-0.0406097658,-0.0284233205,0.3491848409,0.7707187533,0.1276440471,0.0334592573,-0.1203363314,-0.0359245241,0.0076954216,0.3499154747,0.0637109801,0.3463784754,0.3114224076,-0.0173665993,0.0689949468,-0.0166558977,-0.0879927427,0.0875865072,0.0622759163,-0.403681308,0.0528634526,-0.1899863631,-0.0543130003,-0.304048568,-0.246319145,-0.1040120274,-0.2119144052,-0.409406215,0.227412343,-0.0545448326,0.195634678,-0.1933861077,0.0545213036,0.150928095,-0.0917485282,0.0003248953,0.0094417734,-0.2065684795,-0.0326042064,0.2210373729,-0.1053390354,0.036282707,-0.1857272536,0.030104639,-0.1180914491,0.1154467016,0.0138302073,-0.0758283138,0.1963147074,0.0516775548,0.2055033743,-0.0473079644,-0.6516988873,0.1352370232,0.5362832546,-0.2745609879,0.3799952567,0.1490535438,-0.0947601125,-0.1053224504,-0.2667202652,-0.4566354156,-0.3636389971,-0.065149948,-0.1452127844,0.2417424619,0.2856628299,0.2368689626,0.0091699623,0.1459453553,0.1538192779,-0.0961045325,-0.2391644567,0.4765647054,0.0441288911,-0.1764593273,-0.1454289705,-0.0685137585,0.2939199209,0.2050095499,-0.3704074025,0.1182424426,-0.0839300379,0.0721743554,-0.3724434078,0.0816907659,0.0955052301,0.3071263731,-0.0495688766,-0.1056536734,-0.1409310997,0.1213522106,0.0509188175,0.3641844094,-0.026990803,0.3982563913,-0.0897532776,0.4547502995,0.2624274492,-0.1617570817,0.2130926996,0.1749910712,0.5188266635,0.009832982,-0.4894423485,-0.3085491657,-0.1053094938,-0.0222776867,-0.1453578919,-0.1314470023,-0.2135678381,-0.1715726703,0.1677533388,-0.2021043301,-0.0977446586,0.0934714526,-0.4890373945,0.2789086998,-0.2671627402,-0.0172277242,-0.2255419195,0.2368049771,-0.3388524354,-0.0028135933,0.153073892,-0.1081698015,-0.07584133,-0.0532242581,-0.1671222746,0.2889293432,0.0708025694,0.6286579967,0.0871338695,-0.3619214296,-0.141357556,0.1061918512,0.6050748229,0.373142153,-0.352982223,0.3440543711,0.1444435716,-0.5278424621,-0.1323237866,-0.143072769,0.2472813576,-0.0800599456,0.6192361116,-0.2810607851,-0.2102868855,0.0483553894,0.3565573692,-0.0551361591,-0.0701752603,-0.3375290036,-0.0948550701,-0.2657559514,-0.1750979275,-0.0085992655,0.3727573156,-0.1177621409,-0.0307829622,-0.1024160236,-0.2032846063,0.0942054838,-0.0615632124,0.573679626,-0.1058080271,0.338200748,-0.0404294394,0.2728076577,0.7298618555,0.3896467686,0.0485650674,-0.2766025066,0.2084406763,0.066528216,0.3524948359,0.0325588882,-0.0774381459,0.0276742242,-0.1396045983,0.1097321287,-0.0665155947,0.0800471306,0.3577901721,-0.0214205831,-0.6502224207,-0.2415112257,0.3491684198,0.2181298733,-0.0446061864,0.4857558608,-0.1855886877,-0.4610419273,0.1322818846,0.0488302335,0.917632699,0.07885582,0.338803947,0.1819430143,-0.2325960696,0.2211383432,0.1656601578,0.166991666,-0.2519565523,-0.2050489336,-0.1160061955,-0.3928782344,-0.0402875803,0.1229912043,-0.2947121561,0.1419768631,-0.4592790306,0.1178659424,-0.1265632212,-0.0132099185,-0.1652893126,-0.2160494328,-0.1619814932,0.0130732972,-0.0312642641,-0.1588718444,-0.0141595202,-0.2484316528,-0.3144509196,-0.0913043618,-0.1530977935,0.2241207659,-0.2308596224,0.711448729,-0.0641593337,-0.2984419763,-0.2077656239,0.1909829974,0.1160420105,-0.2021065652,0.0624314547,-0.1067817435,-0.0023861239,0.0436084978,-0.0302293636,-0.1206146553,0.1903314143,0.103524074,-0.2395177782,0.1121623516,-0.0587251335,-0.3359271586,0.0498029776,0.3367282152,0.0734798312,-0.1848011017,-0.1993735135,0.2047347128,0.0342360474,-0.0893299803,0.0972223878,0.1633282453,-0.1373355389,0.1153959483,-0.2287455648,-0.34221524,-0.0366114564,0.536349833,0.3310490549,0.1768986583,0.5639303923,-0.2513001263,-0.0115110436,-0.1210900098,0.3575093746,-0.0484554805,-0.1171174571,0.2420293838,0.0464915559,0.0316427536,-0.1686171591,0.0865352824,0.4298751652,0.2080887109,-0.1115064472,-0.3559649289,-0.5394133329,0.384875387,-0.1338525862,0.2295488119,-0.0943683907,0.01880938,-0.2178017348,-0.1428943872,-0.2319716066,0.0040239966,-0.2592309415,0.0460578129,0.14327389,-0.0542205013,0.1375245154,-0.0019388734,0.117102541,-0.0530867018,-0.0559874289,-0.1022240222,0.1451755464,0.1476070732,0.1074886918,-0.2296527028,-0.1094962209,-0.3554100692,-0.0964422226,-0.1427858174,-0.194521606,0.1473959982,-0.0696068481,0.0959886387,0.0284263659,0.1697631925,-0.0842252448,0.1773672402,-0.0078186998,0.1223417073,-0.1652569473,0.3226857185,-0.1020188332,-0.1035023034,-0.5782651901,0.0080413828,-0.113838926,-0.1187391132,0.1946616322,-0.2355369329,-0.182467401,-0.0658958927,0.1693889797,0.4238764942,-0.2421946824,0.0538665093,0.0671151131,0.1156093776,-0.1454008371,-0.2378463894,-0.437729001,-0.1679386497,-0.2898409069,0.017974183,0.0264651068,0.0214356221,-0.1476328969,0.0471716784,0.2570031881,-0.0331265628,0.3051680624,0.4617432654,0.1223694012,0.0497953072,0.1322663426,0.2484991997,0.3821695149,0.3037095666,-0.1484187543,0.2287614495,0.1643528491,-0.2151861638,0.0003148272,-0.295201987,0.0082641225,0.2186150998,-0.1552518755,-0.1198685765,-0.0192838293,0.1178453267,-0.3018823266,-0.2792961895,-0.1589518487,0.1722860932,-0.0979042053,-0.0384445526,-0.0908661634,-0.1682472229,-0.057713449,-0.1306224465,0.1583766788,-0.1463904381,0.318644166,0.1499689072,-0.2276839167,-0.1426099241,0.2273935825,0.6780184507,-0.0640060231,-0.3532020152,0.1663103849,0.4771778286,0.0767008439,0.1096612886,0.2268028259,0.6265037656,0.6114814878,0.0625266284,-0.0997550935,0.2163287103,-0.1871844083,0.1355879158,0.1679725051,0.0278323274,-0.2090847343,0.2076055259,0.1935129166,-0.0455948152,0.1615617275,0.1468857527,0.2199598253,-0.3095570803,0.1947542429,-0.3178147376,-0.2008012086,0.0813222602,0.0192717444,-0.3569807112,0.049722556,0.6105672717,0.0852061212,0.1231747791,0.3197076321,0.0237061493,0.0234231185,0.3911937773,0.4590095878,0.1405980736,-0.4542918801,-0.1931885332,-0.3783898652,-0.05332423,0.2400224358,0.0984772667,-0.1206514537,0.283275187,0.256362915,-0.0056377598,0.1536625028,0.2694154084,-0.0883773789,-0.0958834291,-0.4344916642,-0.132951349,0.1246934235,-0.0194771346,0.0221357178,-0.5010998249,0.1718220264,0.1736254692,0.0116492044,-0.0122560086,0.2122923136,0.2748070955,0.5003048778,0.3791132569,0.1555559188,0.4756854475,-0.0263230372,-0.1617094427,0.0252983123,-0.6679952741,-0.1731608957,0.1567950547,0.003010558,0.2358827144,-0.332893312,-0.0505152382,-0.1813413501,0.2213235945,-0.118419081,-0.2370862067,-0.1663625836,-0.0157356467,0.0436302871,-0.2296957672,0.1340019554,0.119735375,0.0235281307,0.4017446041,-0.0975404531,-0.2917836308,0.4567213953,-0.0764693841,-0.1883582324,-0.0190353505,0.0982667431,0.0895537511,0.2853969932,-0.5559753776,-0.0551319532,0.4927321076,-0.1836723834,-0.4319855869,0.0786193237,0.081673272,0.0435456187,-0.0173979606,0.6414551735,-0.2835347056,-0.5346691012,-0.0249203686,-0.2045637965]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/848","title":"Error when concatenate_datasets","comments":"> As you can see in the error the test checks if `indices_mappings_in_memory` is True or not, which is different from the test you do in your script. In a dataset, both the data and the indices mapping can be either on disk or in memory.\r\n> \r\n> The indices mapping correspond to a mapping on top of the data table that is used to re-order\/select a sample of the original data table. For example if you do `dataset.train_test_split`, then the resulting train and test datasets will have both an indices mapping to tell which examples are in train and which ones in test.\r\n> \r\n> Before saving your datasets on disk, you should call `dataset.flatten_indices()` to remove the indices mapping. It should fix your issue. Under the hood it will create a new data table using the indices mapping. The new data table is going to be a subset of the old one (for example taking only the test set examples), and since the indices mapping will be gone you'll be able to concatenate your datasets.\r\n\r\n`dataset.flatten_indices()` solved my problem, thanks so much!","body":"Hello, when I concatenate two dataset loading from disk, I encountered a problem:\r\n```\r\ntest_dataset = load_from_disk('data\/test_dataset')\r\ntrn_dataset = load_from_disk('data\/train_dataset')\r\n\r\ntrain_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n```\r\nAnd it reported ValueError blow:\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-38-74fa525512ca> in <module>\r\n----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n\r\n\/opt\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in concatenate_datasets(dsets, info, split)\r\n 2547 \"However datasets' indices {} come from memory and datasets' indices {} come from disk.\".format(\r\n 2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],\r\n-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],\r\n 2550 )\r\n 2551 )\r\n\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\n\r\nBut it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:\r\n```\r\ntrn_dataset._data_files\r\n# output\r\n[{'filename': 'data\/train_dataset\/csv-train.arrow', 'skip': 0, 'take': 593264}]\r\n\r\ntest_dataset._data_files\r\n# output\r\n[{'filename': 'data\/test_dataset\/csv-test.arrow', 'skip': 0, 'take': 424383}]\r\n\r\nprint([not dset._data_files for dset in [trn_dataset, test_dataset]])\r\n# [False, False]\r\n\r\n# And I tested the code the same as arrow_dataset, but nothing happened\r\ndsets = [trn_dataset, test_dataset]\r\ndsets_in_memory = [not dset._data_files for dset in dsets]\r\nif any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):\r\n raise ValueError(\r\n \"Datasets should ALL come from memory, or should ALL come from disk.\\n\"\r\n \"However datasets {} come from memory and datasets {} come from disk.\".format(\r\n [i for i in range(len(dsets)) if dsets_in_memory[i]],\r\n [i for i in range(len(dsets)) if not dsets_in_memory[i]],\r\n )\r\n )\r\n```\r\n\r\nAny suggestions would be greatly appreciated! \r\nThanks!","comment_length":184,"text":"Error when concatenate_datasets\nHello, when I concatenate two dataset loading from disk, I encountered a problem:\r\n```\r\ntest_dataset = load_from_disk('data\/test_dataset')\r\ntrn_dataset = load_from_disk('data\/train_dataset')\r\n\r\ntrain_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n```\r\nAnd it reported ValueError blow:\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-38-74fa525512ca> in <module>\r\n----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n\r\n\/opt\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in concatenate_datasets(dsets, info, split)\r\n 2547 \"However datasets' indices {} come from memory and datasets' indices {} come from disk.\".format(\r\n 2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],\r\n-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],\r\n 2550 )\r\n 2551 )\r\n\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\n\r\nBut it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:\r\n```\r\ntrn_dataset._data_files\r\n# output\r\n[{'filename': 'data\/train_dataset\/csv-train.arrow', 'skip': 0, 'take': 593264}]\r\n\r\ntest_dataset._data_files\r\n# output\r\n[{'filename': 'data\/test_dataset\/csv-test.arrow', 'skip': 0, 'take': 424383}]\r\n\r\nprint([not dset._data_files for dset in [trn_dataset, test_dataset]])\r\n# [False, False]\r\n\r\n# And I tested the code the same as arrow_dataset, but nothing happened\r\ndsets = [trn_dataset, test_dataset]\r\ndsets_in_memory = [not dset._data_files for dset in dsets]\r\nif any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):\r\n raise ValueError(\r\n \"Datasets should ALL come from memory, or should ALL come from disk.\\n\"\r\n \"However datasets {} come from memory and datasets {} come from disk.\".format(\r\n [i for i in range(len(dsets)) if dsets_in_memory[i]],\r\n [i for i in range(len(dsets)) if not dsets_in_memory[i]],\r\n )\r\n )\r\n```\r\n\r\nAny suggestions would be greatly appreciated! \r\nThanks!\n> As you can see in the error the test checks if `indices_mappings_in_memory` is True or not, which is different from the test you do in your script. In a dataset, both the data and the indices mapping can be either on disk or in memory.\r\n> \r\n> The indices mapping correspond to a mapping on top of the data table that is used to re-order\/select a sample of the original data table. For example if you do `dataset.train_test_split`, then the resulting train and test datasets will have both an indices mapping to tell which examples are in train and which ones in test.\r\n> \r\n> Before saving your datasets on disk, you should call `dataset.flatten_indices()` to remove the indices mapping. It should fix your issue. Under the hood it will create a new data table using the indices mapping. The new data table is going to be a subset of the old one (for example taking only the test set examples), and since the indices mapping will be gone you'll be able to concatenate your datasets.\r\n\r\n`dataset.flatten_indices()` solved my problem, thanks so much!","embeddings":[-0.0911899135,-0.1483132094,-0.0447487496,0.6809360385,0.1068756282,0.2218791395,0.3129051626,0.2291087657,-0.1234838143,0.1246045008,-0.1083224565,0.2736025453,-0.0816895962,-0.1234090403,-0.30136621,-0.0974017531,0.1984925121,0.1485707313,-0.5059524179,-0.031675972,-0.3506394327,0.1602455676,-0.3358776569,0.0398673192,-0.4097848237,0.049245391,-0.1317905039,0.2761821151,-0.0312033258,-0.1838645488,0.4585869312,-0.3488870859,0.1864263117,0.5364499688,-0.0001176379,0.1700864136,0.2740953863,-0.1723825485,-0.3783839941,-0.256719321,-0.3621298075,-0.1004148722,-0.0029407758,-0.1234583333,0.207650885,-0.1481209397,-0.2583855689,-0.5043874979,0.33071509,0.2264483124,0.1104794741,0.3941425681,0.0912052542,-0.0980597213,0.0378851555,0.1533982009,0.0659453794,0.4203861058,-0.1898589581,-0.0617092587,0.2108293176,0.0670610666,-0.1246672496,0.1450287402,0.0780371949,0.3099108636,0.1576000452,-0.3968754411,-0.1252778769,0.1079114303,0.525313437,-0.3321693838,-0.3178471923,-0.0861591324,0.1388193816,-0.3995229304,0.217888236,0.2979132831,-0.147090584,0.1174623221,-0.0907389298,0.0771459118,-0.2191733122,0.1846294105,-0.1806519479,0.114519842,-0.1610969752,0.2407015562,0.138170585,-0.0730411187,0.5286436677,-0.3571856916,-0.1269245148,0.1591957361,-0.3422968984,0.0321878828,-0.1896209419,-0.704597652,0.0890091434,0.055820778,0.1037015319,-0.2204585522,0.0773697644,0.2818921804,0.4017777145,0.1492893547,-0.037381921,0.5497024655,-0.0202757102,-0.0616969094,0.199370116,0.009052719,-0.2867643833,-0.2114093006,0.1154302359,-0.1646887511,0.0836554542,-0.0013334746,-0.4971266091,-0.1518924385,-0.1278606951,-0.1404941082,0.1965599209,0.0301942825,0.123899661,0.5308156013,0.130738169,0.4409759939,0.217348516,0.0174215604,-0.226509124,0.0222139601,-0.0451169945,0.2104356736,0.103413932,0.0044444473,0.0821225643,0.1775865853,0.1087558568,0.006450789,0.0768852085,-0.4948595166,0.0583267137,0.3114701509,0.0211642738,0.0750039294,0.2738741934,-0.0503142029,-0.1439029574,0.3170154095,-0.4437072873,-0.2742709517,-0.513563931,0.0968817398,0.0989558399,0.2329158783,-0.3203627765,-0.1113318652,0.6928232908,0.1045668423,-0.0683576986,-0.0662886873,-0.1984153092,-0.3195584714,0.3204246461,0.2007118911,-0.2164676934,-0.0330451019,-0.0184931085,-0.0677970722,0.3604027629,0.5054668188,-0.3153637946,0.1992003024,-0.3855018318,-0.0331822969,0.3084054589,-0.4366770387,-0.1402199119,0.2154022306,0.0032159954,-0.073586531,0.2376121879,0.0082833301,0.1314229518,0.1472160071,0.8280869722,0.3132544458,0.0639004782,-0.1781871915,-0.097612448,-0.1383359432,0.2659764588,0.1000099033,-0.2323941588,0.0350794122,-0.1362989247,-0.4177925885,0.3237215579,-0.1016945988,0.0396771654,0.3689118922,0.0351548083,-0.343180269,-0.1154859886,0.0461800434,-0.1954729706,0.0483524054,0.0379616804,-0.0852145106,-0.2998000979,0.0915440768,-0.1228940859,0.0769214854,-0.072335422,0.0498545468,0.0176120866,0.235278517,-0.1725991517,-0.2480318099,-0.1653319597,0.4237135053,-0.2777830064,-0.026918387,-0.2073603868,0.5686868429,-0.0645174757,-0.0320977345,-0.1763465405,-0.0393736251,0.0589343719,0.1164020374,-0.0541011542,0.5149708986,0.0520663075,-0.1684773117,0.025941968,-0.2057821304,0.2290216684,-0.0772498548,0.139911443,-0.1201847419,0.1309275776,-0.0969888195,0.0723317713,0.4227220416,-0.1323702931,0.2828258872,-0.0215705317,0.0900893733,0.1695275009,0.1452977657,-0.058325775,-0.4177126288,-0.0745221898,0.1675421298,0.2188819051,0.2504171431,-0.4655806124,-0.2347148806,0.338824898,0.0246771201,0.0422085784,0.1224695444,-0.3912708461,-0.0406097658,-0.0284233205,0.3491848409,0.7707187533,0.1276440471,0.0334592573,-0.1203363314,-0.0359245241,0.0076954216,0.3499154747,0.0637109801,0.3463784754,0.3114224076,-0.0173665993,0.0689949468,-0.0166558977,-0.0879927427,0.0875865072,0.0622759163,-0.403681308,0.0528634526,-0.1899863631,-0.0543130003,-0.304048568,-0.246319145,-0.1040120274,-0.2119144052,-0.409406215,0.227412343,-0.0545448326,0.195634678,-0.1933861077,0.0545213036,0.150928095,-0.0917485282,0.0003248953,0.0094417734,-0.2065684795,-0.0326042064,0.2210373729,-0.1053390354,0.036282707,-0.1857272536,0.030104639,-0.1180914491,0.1154467016,0.0138302073,-0.0758283138,0.1963147074,0.0516775548,0.2055033743,-0.0473079644,-0.6516988873,0.1352370232,0.5362832546,-0.2745609879,0.3799952567,0.1490535438,-0.0947601125,-0.1053224504,-0.2667202652,-0.4566354156,-0.3636389971,-0.065149948,-0.1452127844,0.2417424619,0.2856628299,0.2368689626,0.0091699623,0.1459453553,0.1538192779,-0.0961045325,-0.2391644567,0.4765647054,0.0441288911,-0.1764593273,-0.1454289705,-0.0685137585,0.2939199209,0.2050095499,-0.3704074025,0.1182424426,-0.0839300379,0.0721743554,-0.3724434078,0.0816907659,0.0955052301,0.3071263731,-0.0495688766,-0.1056536734,-0.1409310997,0.1213522106,0.0509188175,0.3641844094,-0.026990803,0.3982563913,-0.0897532776,0.4547502995,0.2624274492,-0.1617570817,0.2130926996,0.1749910712,0.5188266635,0.009832982,-0.4894423485,-0.3085491657,-0.1053094938,-0.0222776867,-0.1453578919,-0.1314470023,-0.2135678381,-0.1715726703,0.1677533388,-0.2021043301,-0.0977446586,0.0934714526,-0.4890373945,0.2789086998,-0.2671627402,-0.0172277242,-0.2255419195,0.2368049771,-0.3388524354,-0.0028135933,0.153073892,-0.1081698015,-0.07584133,-0.0532242581,-0.1671222746,0.2889293432,0.0708025694,0.6286579967,0.0871338695,-0.3619214296,-0.141357556,0.1061918512,0.6050748229,0.373142153,-0.352982223,0.3440543711,0.1444435716,-0.5278424621,-0.1323237866,-0.143072769,0.2472813576,-0.0800599456,0.6192361116,-0.2810607851,-0.2102868855,0.0483553894,0.3565573692,-0.0551361591,-0.0701752603,-0.3375290036,-0.0948550701,-0.2657559514,-0.1750979275,-0.0085992655,0.3727573156,-0.1177621409,-0.0307829622,-0.1024160236,-0.2032846063,0.0942054838,-0.0615632124,0.573679626,-0.1058080271,0.338200748,-0.0404294394,0.2728076577,0.7298618555,0.3896467686,0.0485650674,-0.2766025066,0.2084406763,0.066528216,0.3524948359,0.0325588882,-0.0774381459,0.0276742242,-0.1396045983,0.1097321287,-0.0665155947,0.0800471306,0.3577901721,-0.0214205831,-0.6502224207,-0.2415112257,0.3491684198,0.2181298733,-0.0446061864,0.4857558608,-0.1855886877,-0.4610419273,0.1322818846,0.0488302335,0.917632699,0.07885582,0.338803947,0.1819430143,-0.2325960696,0.2211383432,0.1656601578,0.166991666,-0.2519565523,-0.2050489336,-0.1160061955,-0.3928782344,-0.0402875803,0.1229912043,-0.2947121561,0.1419768631,-0.4592790306,0.1178659424,-0.1265632212,-0.0132099185,-0.1652893126,-0.2160494328,-0.1619814932,0.0130732972,-0.0312642641,-0.1588718444,-0.0141595202,-0.2484316528,-0.3144509196,-0.0913043618,-0.1530977935,0.2241207659,-0.2308596224,0.711448729,-0.0641593337,-0.2984419763,-0.2077656239,0.1909829974,0.1160420105,-0.2021065652,0.0624314547,-0.1067817435,-0.0023861239,0.0436084978,-0.0302293636,-0.1206146553,0.1903314143,0.103524074,-0.2395177782,0.1121623516,-0.0587251335,-0.3359271586,0.0498029776,0.3367282152,0.0734798312,-0.1848011017,-0.1993735135,0.2047347128,0.0342360474,-0.0893299803,0.0972223878,0.1633282453,-0.1373355389,0.1153959483,-0.2287455648,-0.34221524,-0.0366114564,0.536349833,0.3310490549,0.1768986583,0.5639303923,-0.2513001263,-0.0115110436,-0.1210900098,0.3575093746,-0.0484554805,-0.1171174571,0.2420293838,0.0464915559,0.0316427536,-0.1686171591,0.0865352824,0.4298751652,0.2080887109,-0.1115064472,-0.3559649289,-0.5394133329,0.384875387,-0.1338525862,0.2295488119,-0.0943683907,0.01880938,-0.2178017348,-0.1428943872,-0.2319716066,0.0040239966,-0.2592309415,0.0460578129,0.14327389,-0.0542205013,0.1375245154,-0.0019388734,0.117102541,-0.0530867018,-0.0559874289,-0.1022240222,0.1451755464,0.1476070732,0.1074886918,-0.2296527028,-0.1094962209,-0.3554100692,-0.0964422226,-0.1427858174,-0.194521606,0.1473959982,-0.0696068481,0.0959886387,0.0284263659,0.1697631925,-0.0842252448,0.1773672402,-0.0078186998,0.1223417073,-0.1652569473,0.3226857185,-0.1020188332,-0.1035023034,-0.5782651901,0.0080413828,-0.113838926,-0.1187391132,0.1946616322,-0.2355369329,-0.182467401,-0.0658958927,0.1693889797,0.4238764942,-0.2421946824,0.0538665093,0.0671151131,0.1156093776,-0.1454008371,-0.2378463894,-0.437729001,-0.1679386497,-0.2898409069,0.017974183,0.0264651068,0.0214356221,-0.1476328969,0.0471716784,0.2570031881,-0.0331265628,0.3051680624,0.4617432654,0.1223694012,0.0497953072,0.1322663426,0.2484991997,0.3821695149,0.3037095666,-0.1484187543,0.2287614495,0.1643528491,-0.2151861638,0.0003148272,-0.295201987,0.0082641225,0.2186150998,-0.1552518755,-0.1198685765,-0.0192838293,0.1178453267,-0.3018823266,-0.2792961895,-0.1589518487,0.1722860932,-0.0979042053,-0.0384445526,-0.0908661634,-0.1682472229,-0.057713449,-0.1306224465,0.1583766788,-0.1463904381,0.318644166,0.1499689072,-0.2276839167,-0.1426099241,0.2273935825,0.6780184507,-0.0640060231,-0.3532020152,0.1663103849,0.4771778286,0.0767008439,0.1096612886,0.2268028259,0.6265037656,0.6114814878,0.0625266284,-0.0997550935,0.2163287103,-0.1871844083,0.1355879158,0.1679725051,0.0278323274,-0.2090847343,0.2076055259,0.1935129166,-0.0455948152,0.1615617275,0.1468857527,0.2199598253,-0.3095570803,0.1947542429,-0.3178147376,-0.2008012086,0.0813222602,0.0192717444,-0.3569807112,0.049722556,0.6105672717,0.0852061212,0.1231747791,0.3197076321,0.0237061493,0.0234231185,0.3911937773,0.4590095878,0.1405980736,-0.4542918801,-0.1931885332,-0.3783898652,-0.05332423,0.2400224358,0.0984772667,-0.1206514537,0.283275187,0.256362915,-0.0056377598,0.1536625028,0.2694154084,-0.0883773789,-0.0958834291,-0.4344916642,-0.132951349,0.1246934235,-0.0194771346,0.0221357178,-0.5010998249,0.1718220264,0.1736254692,0.0116492044,-0.0122560086,0.2122923136,0.2748070955,0.5003048778,0.3791132569,0.1555559188,0.4756854475,-0.0263230372,-0.1617094427,0.0252983123,-0.6679952741,-0.1731608957,0.1567950547,0.003010558,0.2358827144,-0.332893312,-0.0505152382,-0.1813413501,0.2213235945,-0.118419081,-0.2370862067,-0.1663625836,-0.0157356467,0.0436302871,-0.2296957672,0.1340019554,0.119735375,0.0235281307,0.4017446041,-0.0975404531,-0.2917836308,0.4567213953,-0.0764693841,-0.1883582324,-0.0190353505,0.0982667431,0.0895537511,0.2853969932,-0.5559753776,-0.0551319532,0.4927321076,-0.1836723834,-0.4319855869,0.0786193237,0.081673272,0.0435456187,-0.0173979606,0.6414551735,-0.2835347056,-0.5346691012,-0.0249203686,-0.2045637965]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/848","title":"Error when concatenate_datasets","comments":"@lhoestq we can add a mention of `dataset.flatten_indices()` in the error message (no rush, just put it on your TODO list or I can do it when I come at it)","body":"Hello, when I concatenate two dataset loading from disk, I encountered a problem:\r\n```\r\ntest_dataset = load_from_disk('data\/test_dataset')\r\ntrn_dataset = load_from_disk('data\/train_dataset')\r\n\r\ntrain_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n```\r\nAnd it reported ValueError blow:\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-38-74fa525512ca> in <module>\r\n----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n\r\n\/opt\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in concatenate_datasets(dsets, info, split)\r\n 2547 \"However datasets' indices {} come from memory and datasets' indices {} come from disk.\".format(\r\n 2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],\r\n-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],\r\n 2550 )\r\n 2551 )\r\n\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\n\r\nBut it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:\r\n```\r\ntrn_dataset._data_files\r\n# output\r\n[{'filename': 'data\/train_dataset\/csv-train.arrow', 'skip': 0, 'take': 593264}]\r\n\r\ntest_dataset._data_files\r\n# output\r\n[{'filename': 'data\/test_dataset\/csv-test.arrow', 'skip': 0, 'take': 424383}]\r\n\r\nprint([not dset._data_files for dset in [trn_dataset, test_dataset]])\r\n# [False, False]\r\n\r\n# And I tested the code the same as arrow_dataset, but nothing happened\r\ndsets = [trn_dataset, test_dataset]\r\ndsets_in_memory = [not dset._data_files for dset in dsets]\r\nif any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):\r\n raise ValueError(\r\n \"Datasets should ALL come from memory, or should ALL come from disk.\\n\"\r\n \"However datasets {} come from memory and datasets {} come from disk.\".format(\r\n [i for i in range(len(dsets)) if dsets_in_memory[i]],\r\n [i for i in range(len(dsets)) if not dsets_in_memory[i]],\r\n )\r\n )\r\n```\r\n\r\nAny suggestions would be greatly appreciated! \r\nThanks!","comment_length":31,"text":"Error when concatenate_datasets\nHello, when I concatenate two dataset loading from disk, I encountered a problem:\r\n```\r\ntest_dataset = load_from_disk('data\/test_dataset')\r\ntrn_dataset = load_from_disk('data\/train_dataset')\r\n\r\ntrain_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n```\r\nAnd it reported ValueError blow:\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-38-74fa525512ca> in <module>\r\n----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n\r\n\/opt\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in concatenate_datasets(dsets, info, split)\r\n 2547 \"However datasets' indices {} come from memory and datasets' indices {} come from disk.\".format(\r\n 2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],\r\n-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],\r\n 2550 )\r\n 2551 )\r\n\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\n\r\nBut it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:\r\n```\r\ntrn_dataset._data_files\r\n# output\r\n[{'filename': 'data\/train_dataset\/csv-train.arrow', 'skip': 0, 'take': 593264}]\r\n\r\ntest_dataset._data_files\r\n# output\r\n[{'filename': 'data\/test_dataset\/csv-test.arrow', 'skip': 0, 'take': 424383}]\r\n\r\nprint([not dset._data_files for dset in [trn_dataset, test_dataset]])\r\n# [False, False]\r\n\r\n# And I tested the code the same as arrow_dataset, but nothing happened\r\ndsets = [trn_dataset, test_dataset]\r\ndsets_in_memory = [not dset._data_files for dset in dsets]\r\nif any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):\r\n raise ValueError(\r\n \"Datasets should ALL come from memory, or should ALL come from disk.\\n\"\r\n \"However datasets {} come from memory and datasets {} come from disk.\".format(\r\n [i for i in range(len(dsets)) if dsets_in_memory[i]],\r\n [i for i in range(len(dsets)) if not dsets_in_memory[i]],\r\n )\r\n )\r\n```\r\n\r\nAny suggestions would be greatly appreciated! \r\nThanks!\n@lhoestq we can add a mention of `dataset.flatten_indices()` in the error message (no rush, just put it on your TODO list or I can do it when I come at it)","embeddings":[-0.0911899135,-0.1483132094,-0.0447487496,0.6809360385,0.1068756282,0.2218791395,0.3129051626,0.2291087657,-0.1234838143,0.1246045008,-0.1083224565,0.2736025453,-0.0816895962,-0.1234090403,-0.30136621,-0.0974017531,0.1984925121,0.1485707313,-0.5059524179,-0.031675972,-0.3506394327,0.1602455676,-0.3358776569,0.0398673192,-0.4097848237,0.049245391,-0.1317905039,0.2761821151,-0.0312033258,-0.1838645488,0.4585869312,-0.3488870859,0.1864263117,0.5364499688,-0.0001176379,0.1700864136,0.2740953863,-0.1723825485,-0.3783839941,-0.256719321,-0.3621298075,-0.1004148722,-0.0029407758,-0.1234583333,0.207650885,-0.1481209397,-0.2583855689,-0.5043874979,0.33071509,0.2264483124,0.1104794741,0.3941425681,0.0912052542,-0.0980597213,0.0378851555,0.1533982009,0.0659453794,0.4203861058,-0.1898589581,-0.0617092587,0.2108293176,0.0670610666,-0.1246672496,0.1450287402,0.0780371949,0.3099108636,0.1576000452,-0.3968754411,-0.1252778769,0.1079114303,0.525313437,-0.3321693838,-0.3178471923,-0.0861591324,0.1388193816,-0.3995229304,0.217888236,0.2979132831,-0.147090584,0.1174623221,-0.0907389298,0.0771459118,-0.2191733122,0.1846294105,-0.1806519479,0.114519842,-0.1610969752,0.2407015562,0.138170585,-0.0730411187,0.5286436677,-0.3571856916,-0.1269245148,0.1591957361,-0.3422968984,0.0321878828,-0.1896209419,-0.704597652,0.0890091434,0.055820778,0.1037015319,-0.2204585522,0.0773697644,0.2818921804,0.4017777145,0.1492893547,-0.037381921,0.5497024655,-0.0202757102,-0.0616969094,0.199370116,0.009052719,-0.2867643833,-0.2114093006,0.1154302359,-0.1646887511,0.0836554542,-0.0013334746,-0.4971266091,-0.1518924385,-0.1278606951,-0.1404941082,0.1965599209,0.0301942825,0.123899661,0.5308156013,0.130738169,0.4409759939,0.217348516,0.0174215604,-0.226509124,0.0222139601,-0.0451169945,0.2104356736,0.103413932,0.0044444473,0.0821225643,0.1775865853,0.1087558568,0.006450789,0.0768852085,-0.4948595166,0.0583267137,0.3114701509,0.0211642738,0.0750039294,0.2738741934,-0.0503142029,-0.1439029574,0.3170154095,-0.4437072873,-0.2742709517,-0.513563931,0.0968817398,0.0989558399,0.2329158783,-0.3203627765,-0.1113318652,0.6928232908,0.1045668423,-0.0683576986,-0.0662886873,-0.1984153092,-0.3195584714,0.3204246461,0.2007118911,-0.2164676934,-0.0330451019,-0.0184931085,-0.0677970722,0.3604027629,0.5054668188,-0.3153637946,0.1992003024,-0.3855018318,-0.0331822969,0.3084054589,-0.4366770387,-0.1402199119,0.2154022306,0.0032159954,-0.073586531,0.2376121879,0.0082833301,0.1314229518,0.1472160071,0.8280869722,0.3132544458,0.0639004782,-0.1781871915,-0.097612448,-0.1383359432,0.2659764588,0.1000099033,-0.2323941588,0.0350794122,-0.1362989247,-0.4177925885,0.3237215579,-0.1016945988,0.0396771654,0.3689118922,0.0351548083,-0.343180269,-0.1154859886,0.0461800434,-0.1954729706,0.0483524054,0.0379616804,-0.0852145106,-0.2998000979,0.0915440768,-0.1228940859,0.0769214854,-0.072335422,0.0498545468,0.0176120866,0.235278517,-0.1725991517,-0.2480318099,-0.1653319597,0.4237135053,-0.2777830064,-0.026918387,-0.2073603868,0.5686868429,-0.0645174757,-0.0320977345,-0.1763465405,-0.0393736251,0.0589343719,0.1164020374,-0.0541011542,0.5149708986,0.0520663075,-0.1684773117,0.025941968,-0.2057821304,0.2290216684,-0.0772498548,0.139911443,-0.1201847419,0.1309275776,-0.0969888195,0.0723317713,0.4227220416,-0.1323702931,0.2828258872,-0.0215705317,0.0900893733,0.1695275009,0.1452977657,-0.058325775,-0.4177126288,-0.0745221898,0.1675421298,0.2188819051,0.2504171431,-0.4655806124,-0.2347148806,0.338824898,0.0246771201,0.0422085784,0.1224695444,-0.3912708461,-0.0406097658,-0.0284233205,0.3491848409,0.7707187533,0.1276440471,0.0334592573,-0.1203363314,-0.0359245241,0.0076954216,0.3499154747,0.0637109801,0.3463784754,0.3114224076,-0.0173665993,0.0689949468,-0.0166558977,-0.0879927427,0.0875865072,0.0622759163,-0.403681308,0.0528634526,-0.1899863631,-0.0543130003,-0.304048568,-0.246319145,-0.1040120274,-0.2119144052,-0.409406215,0.227412343,-0.0545448326,0.195634678,-0.1933861077,0.0545213036,0.150928095,-0.0917485282,0.0003248953,0.0094417734,-0.2065684795,-0.0326042064,0.2210373729,-0.1053390354,0.036282707,-0.1857272536,0.030104639,-0.1180914491,0.1154467016,0.0138302073,-0.0758283138,0.1963147074,0.0516775548,0.2055033743,-0.0473079644,-0.6516988873,0.1352370232,0.5362832546,-0.2745609879,0.3799952567,0.1490535438,-0.0947601125,-0.1053224504,-0.2667202652,-0.4566354156,-0.3636389971,-0.065149948,-0.1452127844,0.2417424619,0.2856628299,0.2368689626,0.0091699623,0.1459453553,0.1538192779,-0.0961045325,-0.2391644567,0.4765647054,0.0441288911,-0.1764593273,-0.1454289705,-0.0685137585,0.2939199209,0.2050095499,-0.3704074025,0.1182424426,-0.0839300379,0.0721743554,-0.3724434078,0.0816907659,0.0955052301,0.3071263731,-0.0495688766,-0.1056536734,-0.1409310997,0.1213522106,0.0509188175,0.3641844094,-0.026990803,0.3982563913,-0.0897532776,0.4547502995,0.2624274492,-0.1617570817,0.2130926996,0.1749910712,0.5188266635,0.009832982,-0.4894423485,-0.3085491657,-0.1053094938,-0.0222776867,-0.1453578919,-0.1314470023,-0.2135678381,-0.1715726703,0.1677533388,-0.2021043301,-0.0977446586,0.0934714526,-0.4890373945,0.2789086998,-0.2671627402,-0.0172277242,-0.2255419195,0.2368049771,-0.3388524354,-0.0028135933,0.153073892,-0.1081698015,-0.07584133,-0.0532242581,-0.1671222746,0.2889293432,0.0708025694,0.6286579967,0.0871338695,-0.3619214296,-0.141357556,0.1061918512,0.6050748229,0.373142153,-0.352982223,0.3440543711,0.1444435716,-0.5278424621,-0.1323237866,-0.143072769,0.2472813576,-0.0800599456,0.6192361116,-0.2810607851,-0.2102868855,0.0483553894,0.3565573692,-0.0551361591,-0.0701752603,-0.3375290036,-0.0948550701,-0.2657559514,-0.1750979275,-0.0085992655,0.3727573156,-0.1177621409,-0.0307829622,-0.1024160236,-0.2032846063,0.0942054838,-0.0615632124,0.573679626,-0.1058080271,0.338200748,-0.0404294394,0.2728076577,0.7298618555,0.3896467686,0.0485650674,-0.2766025066,0.2084406763,0.066528216,0.3524948359,0.0325588882,-0.0774381459,0.0276742242,-0.1396045983,0.1097321287,-0.0665155947,0.0800471306,0.3577901721,-0.0214205831,-0.6502224207,-0.2415112257,0.3491684198,0.2181298733,-0.0446061864,0.4857558608,-0.1855886877,-0.4610419273,0.1322818846,0.0488302335,0.917632699,0.07885582,0.338803947,0.1819430143,-0.2325960696,0.2211383432,0.1656601578,0.166991666,-0.2519565523,-0.2050489336,-0.1160061955,-0.3928782344,-0.0402875803,0.1229912043,-0.2947121561,0.1419768631,-0.4592790306,0.1178659424,-0.1265632212,-0.0132099185,-0.1652893126,-0.2160494328,-0.1619814932,0.0130732972,-0.0312642641,-0.1588718444,-0.0141595202,-0.2484316528,-0.3144509196,-0.0913043618,-0.1530977935,0.2241207659,-0.2308596224,0.711448729,-0.0641593337,-0.2984419763,-0.2077656239,0.1909829974,0.1160420105,-0.2021065652,0.0624314547,-0.1067817435,-0.0023861239,0.0436084978,-0.0302293636,-0.1206146553,0.1903314143,0.103524074,-0.2395177782,0.1121623516,-0.0587251335,-0.3359271586,0.0498029776,0.3367282152,0.0734798312,-0.1848011017,-0.1993735135,0.2047347128,0.0342360474,-0.0893299803,0.0972223878,0.1633282453,-0.1373355389,0.1153959483,-0.2287455648,-0.34221524,-0.0366114564,0.536349833,0.3310490549,0.1768986583,0.5639303923,-0.2513001263,-0.0115110436,-0.1210900098,0.3575093746,-0.0484554805,-0.1171174571,0.2420293838,0.0464915559,0.0316427536,-0.1686171591,0.0865352824,0.4298751652,0.2080887109,-0.1115064472,-0.3559649289,-0.5394133329,0.384875387,-0.1338525862,0.2295488119,-0.0943683907,0.01880938,-0.2178017348,-0.1428943872,-0.2319716066,0.0040239966,-0.2592309415,0.0460578129,0.14327389,-0.0542205013,0.1375245154,-0.0019388734,0.117102541,-0.0530867018,-0.0559874289,-0.1022240222,0.1451755464,0.1476070732,0.1074886918,-0.2296527028,-0.1094962209,-0.3554100692,-0.0964422226,-0.1427858174,-0.194521606,0.1473959982,-0.0696068481,0.0959886387,0.0284263659,0.1697631925,-0.0842252448,0.1773672402,-0.0078186998,0.1223417073,-0.1652569473,0.3226857185,-0.1020188332,-0.1035023034,-0.5782651901,0.0080413828,-0.113838926,-0.1187391132,0.1946616322,-0.2355369329,-0.182467401,-0.0658958927,0.1693889797,0.4238764942,-0.2421946824,0.0538665093,0.0671151131,0.1156093776,-0.1454008371,-0.2378463894,-0.437729001,-0.1679386497,-0.2898409069,0.017974183,0.0264651068,0.0214356221,-0.1476328969,0.0471716784,0.2570031881,-0.0331265628,0.3051680624,0.4617432654,0.1223694012,0.0497953072,0.1322663426,0.2484991997,0.3821695149,0.3037095666,-0.1484187543,0.2287614495,0.1643528491,-0.2151861638,0.0003148272,-0.295201987,0.0082641225,0.2186150998,-0.1552518755,-0.1198685765,-0.0192838293,0.1178453267,-0.3018823266,-0.2792961895,-0.1589518487,0.1722860932,-0.0979042053,-0.0384445526,-0.0908661634,-0.1682472229,-0.057713449,-0.1306224465,0.1583766788,-0.1463904381,0.318644166,0.1499689072,-0.2276839167,-0.1426099241,0.2273935825,0.6780184507,-0.0640060231,-0.3532020152,0.1663103849,0.4771778286,0.0767008439,0.1096612886,0.2268028259,0.6265037656,0.6114814878,0.0625266284,-0.0997550935,0.2163287103,-0.1871844083,0.1355879158,0.1679725051,0.0278323274,-0.2090847343,0.2076055259,0.1935129166,-0.0455948152,0.1615617275,0.1468857527,0.2199598253,-0.3095570803,0.1947542429,-0.3178147376,-0.2008012086,0.0813222602,0.0192717444,-0.3569807112,0.049722556,0.6105672717,0.0852061212,0.1231747791,0.3197076321,0.0237061493,0.0234231185,0.3911937773,0.4590095878,0.1405980736,-0.4542918801,-0.1931885332,-0.3783898652,-0.05332423,0.2400224358,0.0984772667,-0.1206514537,0.283275187,0.256362915,-0.0056377598,0.1536625028,0.2694154084,-0.0883773789,-0.0958834291,-0.4344916642,-0.132951349,0.1246934235,-0.0194771346,0.0221357178,-0.5010998249,0.1718220264,0.1736254692,0.0116492044,-0.0122560086,0.2122923136,0.2748070955,0.5003048778,0.3791132569,0.1555559188,0.4756854475,-0.0263230372,-0.1617094427,0.0252983123,-0.6679952741,-0.1731608957,0.1567950547,0.003010558,0.2358827144,-0.332893312,-0.0505152382,-0.1813413501,0.2213235945,-0.118419081,-0.2370862067,-0.1663625836,-0.0157356467,0.0436302871,-0.2296957672,0.1340019554,0.119735375,0.0235281307,0.4017446041,-0.0975404531,-0.2917836308,0.4567213953,-0.0764693841,-0.1883582324,-0.0190353505,0.0982667431,0.0895537511,0.2853969932,-0.5559753776,-0.0551319532,0.4927321076,-0.1836723834,-0.4319855869,0.0786193237,0.081673272,0.0435456187,-0.0173979606,0.6414551735,-0.2835347056,-0.5346691012,-0.0249203686,-0.2045637965]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/848","title":"Error when concatenate_datasets","comments":"Yup I agree ! And in the docs as well","body":"Hello, when I concatenate two dataset loading from disk, I encountered a problem:\r\n```\r\ntest_dataset = load_from_disk('data\/test_dataset')\r\ntrn_dataset = load_from_disk('data\/train_dataset')\r\n\r\ntrain_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n```\r\nAnd it reported ValueError blow:\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-38-74fa525512ca> in <module>\r\n----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n\r\n\/opt\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in concatenate_datasets(dsets, info, split)\r\n 2547 \"However datasets' indices {} come from memory and datasets' indices {} come from disk.\".format(\r\n 2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],\r\n-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],\r\n 2550 )\r\n 2551 )\r\n\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\n\r\nBut it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:\r\n```\r\ntrn_dataset._data_files\r\n# output\r\n[{'filename': 'data\/train_dataset\/csv-train.arrow', 'skip': 0, 'take': 593264}]\r\n\r\ntest_dataset._data_files\r\n# output\r\n[{'filename': 'data\/test_dataset\/csv-test.arrow', 'skip': 0, 'take': 424383}]\r\n\r\nprint([not dset._data_files for dset in [trn_dataset, test_dataset]])\r\n# [False, False]\r\n\r\n# And I tested the code the same as arrow_dataset, but nothing happened\r\ndsets = [trn_dataset, test_dataset]\r\ndsets_in_memory = [not dset._data_files for dset in dsets]\r\nif any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):\r\n raise ValueError(\r\n \"Datasets should ALL come from memory, or should ALL come from disk.\\n\"\r\n \"However datasets {} come from memory and datasets {} come from disk.\".format(\r\n [i for i in range(len(dsets)) if dsets_in_memory[i]],\r\n [i for i in range(len(dsets)) if not dsets_in_memory[i]],\r\n )\r\n )\r\n```\r\n\r\nAny suggestions would be greatly appreciated! \r\nThanks!","comment_length":10,"text":"Error when concatenate_datasets\nHello, when I concatenate two dataset loading from disk, I encountered a problem:\r\n```\r\ntest_dataset = load_from_disk('data\/test_dataset')\r\ntrn_dataset = load_from_disk('data\/train_dataset')\r\n\r\ntrain_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n```\r\nAnd it reported ValueError blow:\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-38-74fa525512ca> in <module>\r\n----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])\r\n\r\n\/opt\/miniconda3\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in concatenate_datasets(dsets, info, split)\r\n 2547 \"However datasets' indices {} come from memory and datasets' indices {} come from disk.\".format(\r\n 2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],\r\n-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],\r\n 2550 )\r\n 2551 )\r\n\r\nValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.\r\nHowever datasets' indices [1] come from memory and datasets' indices [0] come from disk.\r\n```\r\n\r\nBut it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:\r\n```\r\ntrn_dataset._data_files\r\n# output\r\n[{'filename': 'data\/train_dataset\/csv-train.arrow', 'skip': 0, 'take': 593264}]\r\n\r\ntest_dataset._data_files\r\n# output\r\n[{'filename': 'data\/test_dataset\/csv-test.arrow', 'skip': 0, 'take': 424383}]\r\n\r\nprint([not dset._data_files for dset in [trn_dataset, test_dataset]])\r\n# [False, False]\r\n\r\n# And I tested the code the same as arrow_dataset, but nothing happened\r\ndsets = [trn_dataset, test_dataset]\r\ndsets_in_memory = [not dset._data_files for dset in dsets]\r\nif any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):\r\n raise ValueError(\r\n \"Datasets should ALL come from memory, or should ALL come from disk.\\n\"\r\n \"However datasets {} come from memory and datasets {} come from disk.\".format(\r\n [i for i in range(len(dsets)) if dsets_in_memory[i]],\r\n [i for i in range(len(dsets)) if not dsets_in_memory[i]],\r\n )\r\n )\r\n```\r\n\r\nAny suggestions would be greatly appreciated! \r\nThanks!\nYup I agree ! And in the docs as well","embeddings":[-0.0911899135,-0.1483132094,-0.0447487496,0.6809360385,0.1068756282,0.2218791395,0.3129051626,0.2291087657,-0.1234838143,0.1246045008,-0.1083224565,0.2736025453,-0.0816895962,-0.1234090403,-0.30136621,-0.0974017531,0.1984925121,0.1485707313,-0.5059524179,-0.031675972,-0.3506394327,0.1602455676,-0.3358776569,0.0398673192,-0.4097848237,0.049245391,-0.1317905039,0.2761821151,-0.0312033258,-0.1838645488,0.4585869312,-0.3488870859,0.1864263117,0.5364499688,-0.0001176379,0.1700864136,0.2740953863,-0.1723825485,-0.3783839941,-0.256719321,-0.3621298075,-0.1004148722,-0.0029407758,-0.1234583333,0.207650885,-0.1481209397,-0.2583855689,-0.5043874979,0.33071509,0.2264483124,0.1104794741,0.3941425681,0.0912052542,-0.0980597213,0.0378851555,0.1533982009,0.0659453794,0.4203861058,-0.1898589581,-0.0617092587,0.2108293176,0.0670610666,-0.1246672496,0.1450287402,0.0780371949,0.3099108636,0.1576000452,-0.3968754411,-0.1252778769,0.1079114303,0.525313437,-0.3321693838,-0.3178471923,-0.0861591324,0.1388193816,-0.3995229304,0.217888236,0.2979132831,-0.147090584,0.1174623221,-0.0907389298,0.0771459118,-0.2191733122,0.1846294105,-0.1806519479,0.114519842,-0.1610969752,0.2407015562,0.138170585,-0.0730411187,0.5286436677,-0.3571856916,-0.1269245148,0.1591957361,-0.3422968984,0.0321878828,-0.1896209419,-0.704597652,0.0890091434,0.055820778,0.1037015319,-0.2204585522,0.0773697644,0.2818921804,0.4017777145,0.1492893547,-0.037381921,0.5497024655,-0.0202757102,-0.0616969094,0.199370116,0.009052719,-0.2867643833,-0.2114093006,0.1154302359,-0.1646887511,0.0836554542,-0.0013334746,-0.4971266091,-0.1518924385,-0.1278606951,-0.1404941082,0.1965599209,0.0301942825,0.123899661,0.5308156013,0.130738169,0.4409759939,0.217348516,0.0174215604,-0.226509124,0.0222139601,-0.0451169945,0.2104356736,0.103413932,0.0044444473,0.0821225643,0.1775865853,0.1087558568,0.006450789,0.0768852085,-0.4948595166,0.0583267137,0.3114701509,0.0211642738,0.0750039294,0.2738741934,-0.0503142029,-0.1439029574,0.3170154095,-0.4437072873,-0.2742709517,-0.513563931,0.0968817398,0.0989558399,0.2329158783,-0.3203627765,-0.1113318652,0.6928232908,0.1045668423,-0.0683576986,-0.0662886873,-0.1984153092,-0.3195584714,0.3204246461,0.2007118911,-0.2164676934,-0.0330451019,-0.0184931085,-0.0677970722,0.3604027629,0.5054668188,-0.3153637946,0.1992003024,-0.3855018318,-0.0331822969,0.3084054589,-0.4366770387,-0.1402199119,0.2154022306,0.0032159954,-0.073586531,0.2376121879,0.0082833301,0.1314229518,0.1472160071,0.8280869722,0.3132544458,0.0639004782,-0.1781871915,-0.097612448,-0.1383359432,0.2659764588,0.1000099033,-0.2323941588,0.0350794122,-0.1362989247,-0.4177925885,0.3237215579,-0.1016945988,0.0396771654,0.3689118922,0.0351548083,-0.343180269,-0.1154859886,0.0461800434,-0.1954729706,0.0483524054,0.0379616804,-0.0852145106,-0.2998000979,0.0915440768,-0.1228940859,0.0769214854,-0.072335422,0.0498545468,0.0176120866,0.235278517,-0.1725991517,-0.2480318099,-0.1653319597,0.4237135053,-0.2777830064,-0.026918387,-0.2073603868,0.5686868429,-0.0645174757,-0.0320977345,-0.1763465405,-0.0393736251,0.0589343719,0.1164020374,-0.0541011542,0.5149708986,0.0520663075,-0.1684773117,0.025941968,-0.2057821304,0.2290216684,-0.0772498548,0.139911443,-0.1201847419,0.1309275776,-0.0969888195,0.0723317713,0.4227220416,-0.1323702931,0.2828258872,-0.0215705317,0.0900893733,0.1695275009,0.1452977657,-0.058325775,-0.4177126288,-0.0745221898,0.1675421298,0.2188819051,0.2504171431,-0.4655806124,-0.2347148806,0.338824898,0.0246771201,0.0422085784,0.1224695444,-0.3912708461,-0.0406097658,-0.0284233205,0.3491848409,0.7707187533,0.1276440471,0.0334592573,-0.1203363314,-0.0359245241,0.0076954216,0.3499154747,0.0637109801,0.3463784754,0.3114224076,-0.0173665993,0.0689949468,-0.0166558977,-0.0879927427,0.0875865072,0.0622759163,-0.403681308,0.0528634526,-0.1899863631,-0.0543130003,-0.304048568,-0.246319145,-0.1040120274,-0.2119144052,-0.409406215,0.227412343,-0.0545448326,0.195634678,-0.1933861077,0.0545213036,0.150928095,-0.0917485282,0.0003248953,0.0094417734,-0.2065684795,-0.0326042064,0.2210373729,-0.1053390354,0.036282707,-0.1857272536,0.030104639,-0.1180914491,0.1154467016,0.0138302073,-0.0758283138,0.1963147074,0.0516775548,0.2055033743,-0.0473079644,-0.6516988873,0.1352370232,0.5362832546,-0.2745609879,0.3799952567,0.1490535438,-0.0947601125,-0.1053224504,-0.2667202652,-0.4566354156,-0.3636389971,-0.065149948,-0.1452127844,0.2417424619,0.2856628299,0.2368689626,0.0091699623,0.1459453553,0.1538192779,-0.0961045325,-0.2391644567,0.4765647054,0.0441288911,-0.1764593273,-0.1454289705,-0.0685137585,0.2939199209,0.2050095499,-0.3704074025,0.1182424426,-0.0839300379,0.0721743554,-0.3724434078,0.0816907659,0.0955052301,0.3071263731,-0.0495688766,-0.1056536734,-0.1409310997,0.1213522106,0.0509188175,0.3641844094,-0.026990803,0.3982563913,-0.0897532776,0.4547502995,0.2624274492,-0.1617570817,0.2130926996,0.1749910712,0.5188266635,0.009832982,-0.4894423485,-0.3085491657,-0.1053094938,-0.0222776867,-0.1453578919,-0.1314470023,-0.2135678381,-0.1715726703,0.1677533388,-0.2021043301,-0.0977446586,0.0934714526,-0.4890373945,0.2789086998,-0.2671627402,-0.0172277242,-0.2255419195,0.2368049771,-0.3388524354,-0.0028135933,0.153073892,-0.1081698015,-0.07584133,-0.0532242581,-0.1671222746,0.2889293432,0.0708025694,0.6286579967,0.0871338695,-0.3619214296,-0.141357556,0.1061918512,0.6050748229,0.373142153,-0.352982223,0.3440543711,0.1444435716,-0.5278424621,-0.1323237866,-0.143072769,0.2472813576,-0.0800599456,0.6192361116,-0.2810607851,-0.2102868855,0.0483553894,0.3565573692,-0.0551361591,-0.0701752603,-0.3375290036,-0.0948550701,-0.2657559514,-0.1750979275,-0.0085992655,0.3727573156,-0.1177621409,-0.0307829622,-0.1024160236,-0.2032846063,0.0942054838,-0.0615632124,0.573679626,-0.1058080271,0.338200748,-0.0404294394,0.2728076577,0.7298618555,0.3896467686,0.0485650674,-0.2766025066,0.2084406763,0.066528216,0.3524948359,0.0325588882,-0.0774381459,0.0276742242,-0.1396045983,0.1097321287,-0.0665155947,0.0800471306,0.3577901721,-0.0214205831,-0.6502224207,-0.2415112257,0.3491684198,0.2181298733,-0.0446061864,0.4857558608,-0.1855886877,-0.4610419273,0.1322818846,0.0488302335,0.917632699,0.07885582,0.338803947,0.1819430143,-0.2325960696,0.2211383432,0.1656601578,0.166991666,-0.2519565523,-0.2050489336,-0.1160061955,-0.3928782344,-0.0402875803,0.1229912043,-0.2947121561,0.1419768631,-0.4592790306,0.1178659424,-0.1265632212,-0.0132099185,-0.1652893126,-0.2160494328,-0.1619814932,0.0130732972,-0.0312642641,-0.1588718444,-0.0141595202,-0.2484316528,-0.3144509196,-0.0913043618,-0.1530977935,0.2241207659,-0.2308596224,0.711448729,-0.0641593337,-0.2984419763,-0.2077656239,0.1909829974,0.1160420105,-0.2021065652,0.0624314547,-0.1067817435,-0.0023861239,0.0436084978,-0.0302293636,-0.1206146553,0.1903314143,0.103524074,-0.2395177782,0.1121623516,-0.0587251335,-0.3359271586,0.0498029776,0.3367282152,0.0734798312,-0.1848011017,-0.1993735135,0.2047347128,0.0342360474,-0.0893299803,0.0972223878,0.1633282453,-0.1373355389,0.1153959483,-0.2287455648,-0.34221524,-0.0366114564,0.536349833,0.3310490549,0.1768986583,0.5639303923,-0.2513001263,-0.0115110436,-0.1210900098,0.3575093746,-0.0484554805,-0.1171174571,0.2420293838,0.0464915559,0.0316427536,-0.1686171591,0.0865352824,0.4298751652,0.2080887109,-0.1115064472,-0.3559649289,-0.5394133329,0.384875387,-0.1338525862,0.2295488119,-0.0943683907,0.01880938,-0.2178017348,-0.1428943872,-0.2319716066,0.0040239966,-0.2592309415,0.0460578129,0.14327389,-0.0542205013,0.1375245154,-0.0019388734,0.117102541,-0.0530867018,-0.0559874289,-0.1022240222,0.1451755464,0.1476070732,0.1074886918,-0.2296527028,-0.1094962209,-0.3554100692,-0.0964422226,-0.1427858174,-0.194521606,0.1473959982,-0.0696068481,0.0959886387,0.0284263659,0.1697631925,-0.0842252448,0.1773672402,-0.0078186998,0.1223417073,-0.1652569473,0.3226857185,-0.1020188332,-0.1035023034,-0.5782651901,0.0080413828,-0.113838926,-0.1187391132,0.1946616322,-0.2355369329,-0.182467401,-0.0658958927,0.1693889797,0.4238764942,-0.2421946824,0.0538665093,0.0671151131,0.1156093776,-0.1454008371,-0.2378463894,-0.437729001,-0.1679386497,-0.2898409069,0.017974183,0.0264651068,0.0214356221,-0.1476328969,0.0471716784,0.2570031881,-0.0331265628,0.3051680624,0.4617432654,0.1223694012,0.0497953072,0.1322663426,0.2484991997,0.3821695149,0.3037095666,-0.1484187543,0.2287614495,0.1643528491,-0.2151861638,0.0003148272,-0.295201987,0.0082641225,0.2186150998,-0.1552518755,-0.1198685765,-0.0192838293,0.1178453267,-0.3018823266,-0.2792961895,-0.1589518487,0.1722860932,-0.0979042053,-0.0384445526,-0.0908661634,-0.1682472229,-0.057713449,-0.1306224465,0.1583766788,-0.1463904381,0.318644166,0.1499689072,-0.2276839167,-0.1426099241,0.2273935825,0.6780184507,-0.0640060231,-0.3532020152,0.1663103849,0.4771778286,0.0767008439,0.1096612886,0.2268028259,0.6265037656,0.6114814878,0.0625266284,-0.0997550935,0.2163287103,-0.1871844083,0.1355879158,0.1679725051,0.0278323274,-0.2090847343,0.2076055259,0.1935129166,-0.0455948152,0.1615617275,0.1468857527,0.2199598253,-0.3095570803,0.1947542429,-0.3178147376,-0.2008012086,0.0813222602,0.0192717444,-0.3569807112,0.049722556,0.6105672717,0.0852061212,0.1231747791,0.3197076321,0.0237061493,0.0234231185,0.3911937773,0.4590095878,0.1405980736,-0.4542918801,-0.1931885332,-0.3783898652,-0.05332423,0.2400224358,0.0984772667,-0.1206514537,0.283275187,0.256362915,-0.0056377598,0.1536625028,0.2694154084,-0.0883773789,-0.0958834291,-0.4344916642,-0.132951349,0.1246934235,-0.0194771346,0.0221357178,-0.5010998249,0.1718220264,0.1736254692,0.0116492044,-0.0122560086,0.2122923136,0.2748070955,0.5003048778,0.3791132569,0.1555559188,0.4756854475,-0.0263230372,-0.1617094427,0.0252983123,-0.6679952741,-0.1731608957,0.1567950547,0.003010558,0.2358827144,-0.332893312,-0.0505152382,-0.1813413501,0.2213235945,-0.118419081,-0.2370862067,-0.1663625836,-0.0157356467,0.0436302871,-0.2296957672,0.1340019554,0.119735375,0.0235281307,0.4017446041,-0.0975404531,-0.2917836308,0.4567213953,-0.0764693841,-0.1883582324,-0.0190353505,0.0982667431,0.0895537511,0.2853969932,-0.5559753776,-0.0551319532,0.4927321076,-0.1836723834,-0.4319855869,0.0786193237,0.081673272,0.0435456187,-0.0173979606,0.6414551735,-0.2835347056,-0.5346691012,-0.0249203686,-0.2045637965]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/847","title":"multiprocessing in dataset map \"can only test a child process\"","comments":"It looks like an issue with wandb\/tqdm here.\r\nWe're using the `multiprocess` library instead of the `multiprocessing` builtin python package to support various types of mapping functions. Maybe there's some sort of incompatibility.\r\n\r\nCould you make a minimal script to reproduce or a google colab ?","body":"Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```","comment_length":46,"text":"multiprocessing in dataset map \"can only test a child process\"\nUsing a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```\nIt looks like an issue with wandb\/tqdm here.\r\nWe're using the `multiprocess` library instead of the `multiprocessing` builtin python package to support various types of mapping functions. Maybe there's some sort of incompatibility.\r\n\r\nCould you make a minimal script to reproduce or a google colab ?","embeddings":[-0.3956997097,0.0156290922,-0.1538687646,-0.1850299388,0.141797632,-0.0578565337,0.5068988204,0.3400265574,-0.0834040865,0.1681774557,-0.0184273254,0.3370628655,0.0475015864,0.0918332115,-0.2264437228,0.1021468788,-0.0123605253,0.0946920067,0.1290109754,0.0473636836,-0.1753095239,0.1040760726,-0.2302506417,0.2859587967,-0.4516958296,-0.285730809,-0.0163212903,-0.0623164959,-0.2643237412,-0.4696436226,-0.149113372,0.2851142585,-0.1861115098,0.6030860543,-0.0001044559,0.0936724842,0.2627740204,0.0446500145,-0.054140687,-0.030070236,0.095131956,-0.0515859984,-0.0310001038,-0.4035981894,-0.0044792355,-0.2674306333,0.1779136211,-0.4431346357,0.2878316641,0.3189612627,0.2665320337,0.5655280352,-0.2092227638,-0.0047211065,-0.201201424,-0.0783623531,-0.0357206613,-0.1625539958,0.2167165279,-0.2403003126,-0.3567937613,0.2491746992,-0.0430285074,0.1743139625,-0.3290867805,0.0007867881,0.277649343,-0.237424776,0.3172336221,0.079235658,0.1507846713,-0.2674021721,-0.3465930223,-0.1820102185,-0.2675664425,-0.1380905807,-0.0896170512,0.0866119117,-0.3090318441,0.0237945151,-0.231428802,0.0673615932,-0.0069425311,-0.0971344262,-0.20974648,0.6421571374,0.0204102993,0.2258118987,-0.058426097,0.0421580598,0.1321844459,-0.1236515045,0.1550007761,-0.0283196401,-0.1767590791,-0.1339012235,0.1732696593,-0.3608185351,0.1934940219,-0.0518135242,-0.2702677846,0.1569519043,0.1210899875,0.1657909304,0.1612778902,0.0505134016,0.1559104919,0.2796486914,0.2925396562,-0.196995303,-0.3908066452,-0.0121860905,0.0276292767,-0.2749152184,0.19010818,0.1587806493,0.1859900355,0.0203872453,-0.316894412,-0.0674285442,-0.1310207993,-0.0011093016,0.0109488498,0.4655701518,-0.0625771657,0.210119307,-0.1869856268,0.2766140103,-0.1955184191,-0.1335654557,-0.2674530447,0.1599751413,-0.105241254,-0.0912755728,0.0864879191,0.2284782082,0.3958972394,0.057422515,0.0732819885,-0.0852145255,0.4325413704,-0.219634518,0.2467176765,0.0557839759,0.068621248,0.1359149069,0.26744017,-0.3098887801,-0.1797690243,-0.0748893917,0.0085688289,-0.1883713305,-0.0060650436,0.2934336662,0.0234699752,0.0294385776,-0.1553809643,0.1325114965,0.2363231629,0.0417060144,0.0084723802,-0.1836157143,-0.2125665992,-0.1611631811,0.1659645736,0.2332042754,-0.1246359199,-0.0227329582,0.0209111851,-0.1451625675,0.2747372985,0.2273871899,-0.0913355052,0.6480855346,-0.1730043739,0.5878829956,0.1861093491,-0.5310037136,-0.1901540011,0.2092227191,-0.3177124858,-0.0111948214,0.1317908615,-0.1686787903,0.3851614892,0.080286935,0.1898801774,0.2464483976,-0.1104762852,0.2880237997,-0.2860673666,-0.0096036047,0.0849479064,-0.071969904,0.2393062264,-0.2704798877,-0.0424101911,-0.1621235162,0.2372229397,-0.1482038051,0.2230832577,-0.0443854667,0.0625272393,0.0818433464,-0.0392007828,-0.1215575859,0.0788098946,0.2157844901,0.0433690101,0.0418133661,-0.1905940175,-0.1984119862,-0.0963724405,0.2760661244,-0.2941963971,-0.3430737555,0.3505134284,0.3571797907,-0.0198514294,-0.0224457197,-0.0754135773,0.1201187074,0.1530826986,0.0062083742,0.0715006441,-0.0785518736,-0.1991024315,-0.1974657774,-0.0115251811,0.0726843029,0.1600964069,0.003486624,-0.0784358084,0.4022406936,0.2385911644,0.1009370983,0.2098905593,-0.1776235253,-0.0964289233,0.0071378755,-0.0728402957,0.111530818,0.0388300754,-0.0264948942,0.1444305331,0.4932917356,0.2365600765,0.1932385713,0.2239498198,-0.0101361321,0.2388229668,-0.1069113389,-0.0609275028,-0.1341272146,0.057798285,0.0031344036,0.1119551882,-0.1359223723,-0.2041955888,0.1788984239,0.314935714,-0.0856286287,0.1352984309,-0.1058834121,-0.0881043524,-0.1929530203,0.0814027041,0.1926788986,0.3943273127,0.2004811913,0.1856659055,-0.0386409611,-0.0774591267,-0.0734680966,0.1406773478,0.0299724508,0.1926366687,0.2765003741,0.2475029826,-0.2172683924,-0.3548972011,-0.2435459346,0.1768790036,0.1148345843,-0.1390989274,0.0299772602,-0.3175421357,0.0918787643,-0.0323445313,-0.0665194169,0.2071175724,-0.3488472402,0.1094204113,0.1852676868,-0.1694183499,0.3220192492,-0.1171384901,0.0898978636,-0.0134830419,-0.0217510052,-0.0988093317,-0.1474416703,-0.1254175901,0.1327666491,0.1662156433,0.1449765265,0.4963173568,0.0013359481,-0.2189733684,-0.5480391383,-0.2451718748,0.0971291885,-0.1633996367,0.1954907775,0.2778953016,0.0427568778,-0.0250689685,-0.3409390152,0.3002497554,-0.4462933838,-0.373233676,-0.2133507133,0.0948979482,-0.1722041965,-0.0335693397,-0.3764518201,-0.3236232698,-0.1813369691,0.3523524106,-0.0931495354,0.1716113091,-0.155457288,-0.1054516658,0.1455910206,-0.0203673467,0.1121426821,-0.2187279463,-0.1448161304,-0.0194471348,-0.1510415971,-0.283069551,-0.1560443342,-0.140906021,0.4374369085,-0.10340821,-0.1784181744,-0.0407508872,-0.1612140536,0.4367106259,0.0242293067,-0.0490040779,0.3442159593,0.2696896195,-0.348139286,-0.0873824731,-0.16058743,0.0496429652,-0.2982184887,-0.1103075147,-0.185685128,0.4214314222,0.2078024745,0.5972415209,0.028514469,-0.0613885559,0.1324319392,-0.4377176166,-0.0172750056,-0.0337198712,-0.4259907305,0.1626303345,-0.2667844892,-0.1014861315,0.1350760311,-0.1677293479,-0.115086928,0.1514069438,0.0009905943,-0.3835220933,-0.209966138,0.0377299711,-0.1925776899,0.0860434994,0.0603309199,0.0141091933,-0.4112586379,0.2260724902,-0.0721928179,-0.0218361951,-0.1237445623,-0.3398272991,-0.3464430869,0.0175716821,-0.6085758209,0.3544566035,0.1318598092,0.4851483107,0.1367789209,0.0052999631,0.0414658636,-0.2825492918,0.6524977684,-0.2151976973,0.0749009252,0.183681041,-0.3277136385,-0.2910260558,-0.0436595753,-0.3557530344,0.4460589588,0.3172533214,0.4865555763,-0.1918331534,-0.3518019319,0.1224313006,-0.0199515596,0.0334758908,0.0118676918,-0.28859815,-0.3060879111,-0.4088720381,0.3470336199,0.088713713,0.1813630015,-0.0699506551,-0.1245866269,-0.1805427074,-0.0352943242,0.1069817096,0.1873820275,-0.0616821796,-0.1572833955,0.1408728659,-0.0219112542,0.2116450667,0.2355050445,0.0500611067,-0.0951047987,-0.1150436103,0.0940205231,0.0208330434,0.4114652574,0.2416535616,-0.1685579568,0.3699742556,-0.1364515871,0.288634181,-0.110836409,0.4739766717,0.5543159246,0.1899577975,-0.1046064273,-0.1071350798,0.0384807661,0.1414999962,-0.1566073895,0.3062249124,-0.2532045543,-0.2288146019,0.1476330608,0.1779163629,0.5770257115,-0.0932503566,0.0875760093,0.0692830384,-0.0339656845,0.066425845,-0.322596252,0.1247427389,-0.2881054282,-0.122368142,0.0637505651,-0.0489776097,0.1619583219,0.1876238585,-0.3256686628,0.1354865581,0.2791850567,0.1854793131,0.0244828667,0.4316500723,0.0075259595,-0.2300830781,-0.023363851,0.2279523462,0.1796824038,-0.0562872812,-0.0279301107,0.100480549,-0.0691840649,-0.376296699,-0.193999365,0.0208728425,-0.4312960207,0.3942311704,0.0237804763,0.0594658144,0.0281341858,0.1458708793,0.1447972655,0.2050364465,0.010779175,0.1170280799,0.0932478532,0.0910386071,0.0838550776,-0.2934494615,0.2623491287,-0.1249917671,-0.4529618621,0.0569270551,-0.0704900026,-0.2504338026,0.0244053863,0.1059229895,0.2930746675,-0.2247793376,0.0423178449,-0.0464648195,-0.2383497357,-0.3965626061,0.2039445192,-0.1147942171,-0.198367089,0.2978394926,-0.1070739329,-0.3055641949,-0.0280588754,0.2836199701,0.0299851205,0.152165398,0.5996189117,-0.0773305595,-0.1156411916,-0.4316452146,0.0979604349,0.3189346194,0.1494747549,0.2477394789,-0.1895447522,-0.2431792468,0.0721193999,0.1885505766,0.037892025,0.1453275084,-0.2402581722,-0.2921876609,-0.4588644207,0.0046025463,-0.1703901291,0.2114720941,0.194389537,0.4413281083,-0.1352238059,0.236654982,-0.4279208481,-0.037296284,-0.4685045481,0.2483720928,-0.0478404127,-0.0903501287,0.0450875238,0.0513905063,0.3058929145,0.2208818644,-0.0576614812,-0.3870592415,-0.0921496823,0.0931526944,0.1092805713,0.0143305482,0.1976386756,0.2131322324,-0.2241609246,-0.2637848258,0.1043587327,0.1285088062,-0.0872338116,0.0787764937,0.1816832423,0.0326143913,-0.0256909542,-0.33613047,-0.0561857261,0.0320777707,-0.0667607933,0.3304432631,0.0313246921,-0.0545088127,0.0997803807,0.0593816601,0.2828300595,0.2285246551,0.0306528434,-0.1030482054,-0.3491645157,-0.0028985492,0.3846751153,0.2440446466,-0.2425766736,-0.0939171612,0.2190344483,0.3167302907,-0.231924966,-0.0727749914,0.2596981823,-0.380011797,0.3790821135,0.2261145115,0.0072065424,0.1852706224,0.1105445027,0.0050976626,0.2265835255,-0.3973679543,-0.0119793043,0.0900445506,-0.1410699338,0.1985625029,0.44924137,0.0734671876,0.2462897003,0.341006577,0.014755778,0.5114738345,-0.1246856526,0.0724941492,0.0589090176,-0.0415682569,0.3311864436,-0.0005669777,-0.0347976312,0.0089654345,-0.0880122706,0.3054791391,-0.3336259723,-0.0528675728,-0.2124442756,0.368789047,-0.1914831251,-0.3856400847,-0.23804757,-0.0818293467,-0.1413496882,-0.0785161257,-0.0190696828,0.1973137259,0.0451892503,-0.0820419714,0.0131340185,-0.2212535143,-0.0192149449,-0.0783587024,0.130819723,-0.2626004219,0.2001495957,0.1314703077,0.1404423863,-0.1671113372,0.6095541716,0.5709316134,0.3046991527,-0.1988588274,-0.1260323524,-0.3344715834,-0.1386544555,-0.2002345026,0.286361903,0.0003168109,-0.1747994572,0.2625156939,0.2772244811,-0.2658777535,-0.1564970165,0.4190347791,-0.1771449149,0.1225867122,0.1675085574,0.1353608817,-0.2269658446,0.0732962564,0.1038703024,-0.3389841914,-0.0345561467,0.2996349633,-0.2357952893,0.3266506493,-0.131075263,0.1637846828,0.0463033915,0.4507513344,0.0503231846,0.1739298105,-0.3142849505,-0.1331529915,-0.5782245398,0.2144250721,-0.1688652784,-0.3648037016,-0.1931368709,0.1571838707,0.1029490456,0.1065549478,0.2409303337,0.0778954029,-0.1090169325,0.2187099457,-0.3619211912,0.0869541839,0.1755951494,0.0519847982,0.204318434,-0.4335597754,0.2095237821,-0.0014672085,0.2383017242,-0.0144974636,0.0347771682,0.1814197898,-0.3401248157,0.2746292949,0.1093128175,0.5257284641,-0.4383835793,0.091044113,-0.1269073635,0.1606307179,-0.4419291317,0.0574725047,0.2414098531,0.1892359853,0.0872847959,-0.0311474353,-0.1315664053,0.1188577861,0.2362552434,-0.0922232419,-0.0139229549,0.3029530942,0.1060066298,0.0748861805,-0.1064884812,0.3399718106,-0.0225495938,0.2731781006,-0.4903123081,-0.4749884307,0.3061774373,-0.5121244192,-0.4921657443,-0.3859884441,0.288736403,-0.0939683691,0.0935930386,-0.3449763656,-0.0112365661,0.288376838,0.1060355008,-0.5911830068,0.1808962375,0.088660121,-0.0450419672,-0.0889836028,0.3249361813,0.0077151936,-0.1511936933,-0.0211812947,-0.1020610854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/847","title":"multiprocessing in dataset map \"can only test a child process\"","comments":"hi facing the same issue here - \r\n\r\n`AssertionError: Caught AssertionError in DataLoader worker process 0.\r\nOriginal Traceback (most recent call last):\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 996, in emit\r\n stream.write(msg)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/lib\/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/wandb_run.py\", line 723, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 153, in publish_output\r\n self._publish_output(o)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 158, in _publish_output\r\n self._publish(rec)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 456, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/_utils\/worker.py\", line 198, in _worker_loop\r\n data = fetcher.fetch(index)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"<ipython-input-8-a4d9a08d114e>\", line 20, in __getitem__\r\n return_token_type_ids=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/tokenization_utils_base.py\", line 2405, in encode_plus\r\n **kwargs,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/tokenization_utils_base.py\", line 2125, in _get_padding_truncation_strategies\r\n \"Truncation was not explicitly activated but `max_length` is provided a specific value, \"\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1320, in warning\r\n self._log(WARNING, msg, args, **kwargs)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1444, in _log\r\n self.handle(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1454, in handle\r\n self.callHandlers(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1516, in callHandlers\r\n hdlr.handle(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 865, in handle\r\n self.emit(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1000, in emit\r\n self.handleError(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 917, in handleError\r\n sys.stderr.write('--- Logging error ---\\n')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/lib\/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/wandb_run.py\", line 723, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 153, in publish_output\r\n self._publish_output(o)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 158, in _publish_output\r\n self._publish(rec)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 456, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process`\r\n","body":"Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```","comment_length":293,"text":"multiprocessing in dataset map \"can only test a child process\"\nUsing a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```\nhi facing the same issue here - \r\n\r\n`AssertionError: Caught AssertionError in DataLoader worker process 0.\r\nOriginal Traceback (most recent call last):\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 996, in emit\r\n stream.write(msg)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/lib\/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/wandb_run.py\", line 723, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 153, in publish_output\r\n self._publish_output(o)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 158, in _publish_output\r\n self._publish(rec)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 456, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/_utils\/worker.py\", line 198, in _worker_loop\r\n data = fetcher.fetch(index)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"<ipython-input-8-a4d9a08d114e>\", line 20, in __getitem__\r\n return_token_type_ids=True\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/tokenization_utils_base.py\", line 2405, in encode_plus\r\n **kwargs,\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/transformers\/tokenization_utils_base.py\", line 2125, in _get_padding_truncation_strategies\r\n \"Truncation was not explicitly activated but `max_length` is provided a specific value, \"\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1320, in warning\r\n self._log(WARNING, msg, args, **kwargs)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1444, in _log\r\n self.handle(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1454, in handle\r\n self.callHandlers(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1516, in callHandlers\r\n hdlr.handle(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 865, in handle\r\n self.emit(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 1000, in emit\r\n self.handleError(record)\r\n File \"\/usr\/lib\/python3.6\/logging\/__init__.py\", line 917, in handleError\r\n sys.stderr.write('--- Logging error ---\\n')\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/lib\/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/wandb_run.py\", line 723, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 153, in publish_output\r\n self._publish_output(o)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 158, in _publish_output\r\n self._publish(rec)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/wandb\/sdk\/interface\/interface.py\", line 456, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process`\r\n","embeddings":[-0.3956997097,0.0156290922,-0.1538687646,-0.1850299388,0.141797632,-0.0578565337,0.5068988204,0.3400265574,-0.0834040865,0.1681774557,-0.0184273254,0.3370628655,0.0475015864,0.0918332115,-0.2264437228,0.1021468788,-0.0123605253,0.0946920067,0.1290109754,0.0473636836,-0.1753095239,0.1040760726,-0.2302506417,0.2859587967,-0.4516958296,-0.285730809,-0.0163212903,-0.0623164959,-0.2643237412,-0.4696436226,-0.149113372,0.2851142585,-0.1861115098,0.6030860543,-0.0001044559,0.0936724842,0.2627740204,0.0446500145,-0.054140687,-0.030070236,0.095131956,-0.0515859984,-0.0310001038,-0.4035981894,-0.0044792355,-0.2674306333,0.1779136211,-0.4431346357,0.2878316641,0.3189612627,0.2665320337,0.5655280352,-0.2092227638,-0.0047211065,-0.201201424,-0.0783623531,-0.0357206613,-0.1625539958,0.2167165279,-0.2403003126,-0.3567937613,0.2491746992,-0.0430285074,0.1743139625,-0.3290867805,0.0007867881,0.277649343,-0.237424776,0.3172336221,0.079235658,0.1507846713,-0.2674021721,-0.3465930223,-0.1820102185,-0.2675664425,-0.1380905807,-0.0896170512,0.0866119117,-0.3090318441,0.0237945151,-0.231428802,0.0673615932,-0.0069425311,-0.0971344262,-0.20974648,0.6421571374,0.0204102993,0.2258118987,-0.058426097,0.0421580598,0.1321844459,-0.1236515045,0.1550007761,-0.0283196401,-0.1767590791,-0.1339012235,0.1732696593,-0.3608185351,0.1934940219,-0.0518135242,-0.2702677846,0.1569519043,0.1210899875,0.1657909304,0.1612778902,0.0505134016,0.1559104919,0.2796486914,0.2925396562,-0.196995303,-0.3908066452,-0.0121860905,0.0276292767,-0.2749152184,0.19010818,0.1587806493,0.1859900355,0.0203872453,-0.316894412,-0.0674285442,-0.1310207993,-0.0011093016,0.0109488498,0.4655701518,-0.0625771657,0.210119307,-0.1869856268,0.2766140103,-0.1955184191,-0.1335654557,-0.2674530447,0.1599751413,-0.105241254,-0.0912755728,0.0864879191,0.2284782082,0.3958972394,0.057422515,0.0732819885,-0.0852145255,0.4325413704,-0.219634518,0.2467176765,0.0557839759,0.068621248,0.1359149069,0.26744017,-0.3098887801,-0.1797690243,-0.0748893917,0.0085688289,-0.1883713305,-0.0060650436,0.2934336662,0.0234699752,0.0294385776,-0.1553809643,0.1325114965,0.2363231629,0.0417060144,0.0084723802,-0.1836157143,-0.2125665992,-0.1611631811,0.1659645736,0.2332042754,-0.1246359199,-0.0227329582,0.0209111851,-0.1451625675,0.2747372985,0.2273871899,-0.0913355052,0.6480855346,-0.1730043739,0.5878829956,0.1861093491,-0.5310037136,-0.1901540011,0.2092227191,-0.3177124858,-0.0111948214,0.1317908615,-0.1686787903,0.3851614892,0.080286935,0.1898801774,0.2464483976,-0.1104762852,0.2880237997,-0.2860673666,-0.0096036047,0.0849479064,-0.071969904,0.2393062264,-0.2704798877,-0.0424101911,-0.1621235162,0.2372229397,-0.1482038051,0.2230832577,-0.0443854667,0.0625272393,0.0818433464,-0.0392007828,-0.1215575859,0.0788098946,0.2157844901,0.0433690101,0.0418133661,-0.1905940175,-0.1984119862,-0.0963724405,0.2760661244,-0.2941963971,-0.3430737555,0.3505134284,0.3571797907,-0.0198514294,-0.0224457197,-0.0754135773,0.1201187074,0.1530826986,0.0062083742,0.0715006441,-0.0785518736,-0.1991024315,-0.1974657774,-0.0115251811,0.0726843029,0.1600964069,0.003486624,-0.0784358084,0.4022406936,0.2385911644,0.1009370983,0.2098905593,-0.1776235253,-0.0964289233,0.0071378755,-0.0728402957,0.111530818,0.0388300754,-0.0264948942,0.1444305331,0.4932917356,0.2365600765,0.1932385713,0.2239498198,-0.0101361321,0.2388229668,-0.1069113389,-0.0609275028,-0.1341272146,0.057798285,0.0031344036,0.1119551882,-0.1359223723,-0.2041955888,0.1788984239,0.314935714,-0.0856286287,0.1352984309,-0.1058834121,-0.0881043524,-0.1929530203,0.0814027041,0.1926788986,0.3943273127,0.2004811913,0.1856659055,-0.0386409611,-0.0774591267,-0.0734680966,0.1406773478,0.0299724508,0.1926366687,0.2765003741,0.2475029826,-0.2172683924,-0.3548972011,-0.2435459346,0.1768790036,0.1148345843,-0.1390989274,0.0299772602,-0.3175421357,0.0918787643,-0.0323445313,-0.0665194169,0.2071175724,-0.3488472402,0.1094204113,0.1852676868,-0.1694183499,0.3220192492,-0.1171384901,0.0898978636,-0.0134830419,-0.0217510052,-0.0988093317,-0.1474416703,-0.1254175901,0.1327666491,0.1662156433,0.1449765265,0.4963173568,0.0013359481,-0.2189733684,-0.5480391383,-0.2451718748,0.0971291885,-0.1633996367,0.1954907775,0.2778953016,0.0427568778,-0.0250689685,-0.3409390152,0.3002497554,-0.4462933838,-0.373233676,-0.2133507133,0.0948979482,-0.1722041965,-0.0335693397,-0.3764518201,-0.3236232698,-0.1813369691,0.3523524106,-0.0931495354,0.1716113091,-0.155457288,-0.1054516658,0.1455910206,-0.0203673467,0.1121426821,-0.2187279463,-0.1448161304,-0.0194471348,-0.1510415971,-0.283069551,-0.1560443342,-0.140906021,0.4374369085,-0.10340821,-0.1784181744,-0.0407508872,-0.1612140536,0.4367106259,0.0242293067,-0.0490040779,0.3442159593,0.2696896195,-0.348139286,-0.0873824731,-0.16058743,0.0496429652,-0.2982184887,-0.1103075147,-0.185685128,0.4214314222,0.2078024745,0.5972415209,0.028514469,-0.0613885559,0.1324319392,-0.4377176166,-0.0172750056,-0.0337198712,-0.4259907305,0.1626303345,-0.2667844892,-0.1014861315,0.1350760311,-0.1677293479,-0.115086928,0.1514069438,0.0009905943,-0.3835220933,-0.209966138,0.0377299711,-0.1925776899,0.0860434994,0.0603309199,0.0141091933,-0.4112586379,0.2260724902,-0.0721928179,-0.0218361951,-0.1237445623,-0.3398272991,-0.3464430869,0.0175716821,-0.6085758209,0.3544566035,0.1318598092,0.4851483107,0.1367789209,0.0052999631,0.0414658636,-0.2825492918,0.6524977684,-0.2151976973,0.0749009252,0.183681041,-0.3277136385,-0.2910260558,-0.0436595753,-0.3557530344,0.4460589588,0.3172533214,0.4865555763,-0.1918331534,-0.3518019319,0.1224313006,-0.0199515596,0.0334758908,0.0118676918,-0.28859815,-0.3060879111,-0.4088720381,0.3470336199,0.088713713,0.1813630015,-0.0699506551,-0.1245866269,-0.1805427074,-0.0352943242,0.1069817096,0.1873820275,-0.0616821796,-0.1572833955,0.1408728659,-0.0219112542,0.2116450667,0.2355050445,0.0500611067,-0.0951047987,-0.1150436103,0.0940205231,0.0208330434,0.4114652574,0.2416535616,-0.1685579568,0.3699742556,-0.1364515871,0.288634181,-0.110836409,0.4739766717,0.5543159246,0.1899577975,-0.1046064273,-0.1071350798,0.0384807661,0.1414999962,-0.1566073895,0.3062249124,-0.2532045543,-0.2288146019,0.1476330608,0.1779163629,0.5770257115,-0.0932503566,0.0875760093,0.0692830384,-0.0339656845,0.066425845,-0.322596252,0.1247427389,-0.2881054282,-0.122368142,0.0637505651,-0.0489776097,0.1619583219,0.1876238585,-0.3256686628,0.1354865581,0.2791850567,0.1854793131,0.0244828667,0.4316500723,0.0075259595,-0.2300830781,-0.023363851,0.2279523462,0.1796824038,-0.0562872812,-0.0279301107,0.100480549,-0.0691840649,-0.376296699,-0.193999365,0.0208728425,-0.4312960207,0.3942311704,0.0237804763,0.0594658144,0.0281341858,0.1458708793,0.1447972655,0.2050364465,0.010779175,0.1170280799,0.0932478532,0.0910386071,0.0838550776,-0.2934494615,0.2623491287,-0.1249917671,-0.4529618621,0.0569270551,-0.0704900026,-0.2504338026,0.0244053863,0.1059229895,0.2930746675,-0.2247793376,0.0423178449,-0.0464648195,-0.2383497357,-0.3965626061,0.2039445192,-0.1147942171,-0.198367089,0.2978394926,-0.1070739329,-0.3055641949,-0.0280588754,0.2836199701,0.0299851205,0.152165398,0.5996189117,-0.0773305595,-0.1156411916,-0.4316452146,0.0979604349,0.3189346194,0.1494747549,0.2477394789,-0.1895447522,-0.2431792468,0.0721193999,0.1885505766,0.037892025,0.1453275084,-0.2402581722,-0.2921876609,-0.4588644207,0.0046025463,-0.1703901291,0.2114720941,0.194389537,0.4413281083,-0.1352238059,0.236654982,-0.4279208481,-0.037296284,-0.4685045481,0.2483720928,-0.0478404127,-0.0903501287,0.0450875238,0.0513905063,0.3058929145,0.2208818644,-0.0576614812,-0.3870592415,-0.0921496823,0.0931526944,0.1092805713,0.0143305482,0.1976386756,0.2131322324,-0.2241609246,-0.2637848258,0.1043587327,0.1285088062,-0.0872338116,0.0787764937,0.1816832423,0.0326143913,-0.0256909542,-0.33613047,-0.0561857261,0.0320777707,-0.0667607933,0.3304432631,0.0313246921,-0.0545088127,0.0997803807,0.0593816601,0.2828300595,0.2285246551,0.0306528434,-0.1030482054,-0.3491645157,-0.0028985492,0.3846751153,0.2440446466,-0.2425766736,-0.0939171612,0.2190344483,0.3167302907,-0.231924966,-0.0727749914,0.2596981823,-0.380011797,0.3790821135,0.2261145115,0.0072065424,0.1852706224,0.1105445027,0.0050976626,0.2265835255,-0.3973679543,-0.0119793043,0.0900445506,-0.1410699338,0.1985625029,0.44924137,0.0734671876,0.2462897003,0.341006577,0.014755778,0.5114738345,-0.1246856526,0.0724941492,0.0589090176,-0.0415682569,0.3311864436,-0.0005669777,-0.0347976312,0.0089654345,-0.0880122706,0.3054791391,-0.3336259723,-0.0528675728,-0.2124442756,0.368789047,-0.1914831251,-0.3856400847,-0.23804757,-0.0818293467,-0.1413496882,-0.0785161257,-0.0190696828,0.1973137259,0.0451892503,-0.0820419714,0.0131340185,-0.2212535143,-0.0192149449,-0.0783587024,0.130819723,-0.2626004219,0.2001495957,0.1314703077,0.1404423863,-0.1671113372,0.6095541716,0.5709316134,0.3046991527,-0.1988588274,-0.1260323524,-0.3344715834,-0.1386544555,-0.2002345026,0.286361903,0.0003168109,-0.1747994572,0.2625156939,0.2772244811,-0.2658777535,-0.1564970165,0.4190347791,-0.1771449149,0.1225867122,0.1675085574,0.1353608817,-0.2269658446,0.0732962564,0.1038703024,-0.3389841914,-0.0345561467,0.2996349633,-0.2357952893,0.3266506493,-0.131075263,0.1637846828,0.0463033915,0.4507513344,0.0503231846,0.1739298105,-0.3142849505,-0.1331529915,-0.5782245398,0.2144250721,-0.1688652784,-0.3648037016,-0.1931368709,0.1571838707,0.1029490456,0.1065549478,0.2409303337,0.0778954029,-0.1090169325,0.2187099457,-0.3619211912,0.0869541839,0.1755951494,0.0519847982,0.204318434,-0.4335597754,0.2095237821,-0.0014672085,0.2383017242,-0.0144974636,0.0347771682,0.1814197898,-0.3401248157,0.2746292949,0.1093128175,0.5257284641,-0.4383835793,0.091044113,-0.1269073635,0.1606307179,-0.4419291317,0.0574725047,0.2414098531,0.1892359853,0.0872847959,-0.0311474353,-0.1315664053,0.1188577861,0.2362552434,-0.0922232419,-0.0139229549,0.3029530942,0.1060066298,0.0748861805,-0.1064884812,0.3399718106,-0.0225495938,0.2731781006,-0.4903123081,-0.4749884307,0.3061774373,-0.5121244192,-0.4921657443,-0.3859884441,0.288736403,-0.0939683691,0.0935930386,-0.3449763656,-0.0112365661,0.288376838,0.1060355008,-0.5911830068,0.1808962375,0.088660121,-0.0450419672,-0.0889836028,0.3249361813,0.0077151936,-0.1511936933,-0.0211812947,-0.1020610854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/847","title":"multiprocessing in dataset map \"can only test a child process\"","comments":"It looks like this warning : \r\n\"Truncation was not explicitly activated but max_length is provided a specific value, \"\r\nis not handled well by wandb.\r\n\r\nThe error occurs when calling the tokenizer.\r\nMaybe you can try to specify `truncation=True` when calling the tokenizer to remove the warning ?\r\nOtherwise I don't know why wandb would fail on a warning. Maybe one of its logging handlers have some issues with the logging of tokenizers. Maybe @n1t0 knows more about this ?","body":"Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```","comment_length":80,"text":"multiprocessing in dataset map \"can only test a child process\"\nUsing a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```\nIt looks like this warning : \r\n\"Truncation was not explicitly activated but max_length is provided a specific value, \"\r\nis not handled well by wandb.\r\n\r\nThe error occurs when calling the tokenizer.\r\nMaybe you can try to specify `truncation=True` when calling the tokenizer to remove the warning ?\r\nOtherwise I don't know why wandb would fail on a warning. Maybe one of its logging handlers have some issues with the logging of tokenizers. Maybe @n1t0 knows more about this ?","embeddings":[-0.3956997097,0.0156290922,-0.1538687646,-0.1850299388,0.141797632,-0.0578565337,0.5068988204,0.3400265574,-0.0834040865,0.1681774557,-0.0184273254,0.3370628655,0.0475015864,0.0918332115,-0.2264437228,0.1021468788,-0.0123605253,0.0946920067,0.1290109754,0.0473636836,-0.1753095239,0.1040760726,-0.2302506417,0.2859587967,-0.4516958296,-0.285730809,-0.0163212903,-0.0623164959,-0.2643237412,-0.4696436226,-0.149113372,0.2851142585,-0.1861115098,0.6030860543,-0.0001044559,0.0936724842,0.2627740204,0.0446500145,-0.054140687,-0.030070236,0.095131956,-0.0515859984,-0.0310001038,-0.4035981894,-0.0044792355,-0.2674306333,0.1779136211,-0.4431346357,0.2878316641,0.3189612627,0.2665320337,0.5655280352,-0.2092227638,-0.0047211065,-0.201201424,-0.0783623531,-0.0357206613,-0.1625539958,0.2167165279,-0.2403003126,-0.3567937613,0.2491746992,-0.0430285074,0.1743139625,-0.3290867805,0.0007867881,0.277649343,-0.237424776,0.3172336221,0.079235658,0.1507846713,-0.2674021721,-0.3465930223,-0.1820102185,-0.2675664425,-0.1380905807,-0.0896170512,0.0866119117,-0.3090318441,0.0237945151,-0.231428802,0.0673615932,-0.0069425311,-0.0971344262,-0.20974648,0.6421571374,0.0204102993,0.2258118987,-0.058426097,0.0421580598,0.1321844459,-0.1236515045,0.1550007761,-0.0283196401,-0.1767590791,-0.1339012235,0.1732696593,-0.3608185351,0.1934940219,-0.0518135242,-0.2702677846,0.1569519043,0.1210899875,0.1657909304,0.1612778902,0.0505134016,0.1559104919,0.2796486914,0.2925396562,-0.196995303,-0.3908066452,-0.0121860905,0.0276292767,-0.2749152184,0.19010818,0.1587806493,0.1859900355,0.0203872453,-0.316894412,-0.0674285442,-0.1310207993,-0.0011093016,0.0109488498,0.4655701518,-0.0625771657,0.210119307,-0.1869856268,0.2766140103,-0.1955184191,-0.1335654557,-0.2674530447,0.1599751413,-0.105241254,-0.0912755728,0.0864879191,0.2284782082,0.3958972394,0.057422515,0.0732819885,-0.0852145255,0.4325413704,-0.219634518,0.2467176765,0.0557839759,0.068621248,0.1359149069,0.26744017,-0.3098887801,-0.1797690243,-0.0748893917,0.0085688289,-0.1883713305,-0.0060650436,0.2934336662,0.0234699752,0.0294385776,-0.1553809643,0.1325114965,0.2363231629,0.0417060144,0.0084723802,-0.1836157143,-0.2125665992,-0.1611631811,0.1659645736,0.2332042754,-0.1246359199,-0.0227329582,0.0209111851,-0.1451625675,0.2747372985,0.2273871899,-0.0913355052,0.6480855346,-0.1730043739,0.5878829956,0.1861093491,-0.5310037136,-0.1901540011,0.2092227191,-0.3177124858,-0.0111948214,0.1317908615,-0.1686787903,0.3851614892,0.080286935,0.1898801774,0.2464483976,-0.1104762852,0.2880237997,-0.2860673666,-0.0096036047,0.0849479064,-0.071969904,0.2393062264,-0.2704798877,-0.0424101911,-0.1621235162,0.2372229397,-0.1482038051,0.2230832577,-0.0443854667,0.0625272393,0.0818433464,-0.0392007828,-0.1215575859,0.0788098946,0.2157844901,0.0433690101,0.0418133661,-0.1905940175,-0.1984119862,-0.0963724405,0.2760661244,-0.2941963971,-0.3430737555,0.3505134284,0.3571797907,-0.0198514294,-0.0224457197,-0.0754135773,0.1201187074,0.1530826986,0.0062083742,0.0715006441,-0.0785518736,-0.1991024315,-0.1974657774,-0.0115251811,0.0726843029,0.1600964069,0.003486624,-0.0784358084,0.4022406936,0.2385911644,0.1009370983,0.2098905593,-0.1776235253,-0.0964289233,0.0071378755,-0.0728402957,0.111530818,0.0388300754,-0.0264948942,0.1444305331,0.4932917356,0.2365600765,0.1932385713,0.2239498198,-0.0101361321,0.2388229668,-0.1069113389,-0.0609275028,-0.1341272146,0.057798285,0.0031344036,0.1119551882,-0.1359223723,-0.2041955888,0.1788984239,0.314935714,-0.0856286287,0.1352984309,-0.1058834121,-0.0881043524,-0.1929530203,0.0814027041,0.1926788986,0.3943273127,0.2004811913,0.1856659055,-0.0386409611,-0.0774591267,-0.0734680966,0.1406773478,0.0299724508,0.1926366687,0.2765003741,0.2475029826,-0.2172683924,-0.3548972011,-0.2435459346,0.1768790036,0.1148345843,-0.1390989274,0.0299772602,-0.3175421357,0.0918787643,-0.0323445313,-0.0665194169,0.2071175724,-0.3488472402,0.1094204113,0.1852676868,-0.1694183499,0.3220192492,-0.1171384901,0.0898978636,-0.0134830419,-0.0217510052,-0.0988093317,-0.1474416703,-0.1254175901,0.1327666491,0.1662156433,0.1449765265,0.4963173568,0.0013359481,-0.2189733684,-0.5480391383,-0.2451718748,0.0971291885,-0.1633996367,0.1954907775,0.2778953016,0.0427568778,-0.0250689685,-0.3409390152,0.3002497554,-0.4462933838,-0.373233676,-0.2133507133,0.0948979482,-0.1722041965,-0.0335693397,-0.3764518201,-0.3236232698,-0.1813369691,0.3523524106,-0.0931495354,0.1716113091,-0.155457288,-0.1054516658,0.1455910206,-0.0203673467,0.1121426821,-0.2187279463,-0.1448161304,-0.0194471348,-0.1510415971,-0.283069551,-0.1560443342,-0.140906021,0.4374369085,-0.10340821,-0.1784181744,-0.0407508872,-0.1612140536,0.4367106259,0.0242293067,-0.0490040779,0.3442159593,0.2696896195,-0.348139286,-0.0873824731,-0.16058743,0.0496429652,-0.2982184887,-0.1103075147,-0.185685128,0.4214314222,0.2078024745,0.5972415209,0.028514469,-0.0613885559,0.1324319392,-0.4377176166,-0.0172750056,-0.0337198712,-0.4259907305,0.1626303345,-0.2667844892,-0.1014861315,0.1350760311,-0.1677293479,-0.115086928,0.1514069438,0.0009905943,-0.3835220933,-0.209966138,0.0377299711,-0.1925776899,0.0860434994,0.0603309199,0.0141091933,-0.4112586379,0.2260724902,-0.0721928179,-0.0218361951,-0.1237445623,-0.3398272991,-0.3464430869,0.0175716821,-0.6085758209,0.3544566035,0.1318598092,0.4851483107,0.1367789209,0.0052999631,0.0414658636,-0.2825492918,0.6524977684,-0.2151976973,0.0749009252,0.183681041,-0.3277136385,-0.2910260558,-0.0436595753,-0.3557530344,0.4460589588,0.3172533214,0.4865555763,-0.1918331534,-0.3518019319,0.1224313006,-0.0199515596,0.0334758908,0.0118676918,-0.28859815,-0.3060879111,-0.4088720381,0.3470336199,0.088713713,0.1813630015,-0.0699506551,-0.1245866269,-0.1805427074,-0.0352943242,0.1069817096,0.1873820275,-0.0616821796,-0.1572833955,0.1408728659,-0.0219112542,0.2116450667,0.2355050445,0.0500611067,-0.0951047987,-0.1150436103,0.0940205231,0.0208330434,0.4114652574,0.2416535616,-0.1685579568,0.3699742556,-0.1364515871,0.288634181,-0.110836409,0.4739766717,0.5543159246,0.1899577975,-0.1046064273,-0.1071350798,0.0384807661,0.1414999962,-0.1566073895,0.3062249124,-0.2532045543,-0.2288146019,0.1476330608,0.1779163629,0.5770257115,-0.0932503566,0.0875760093,0.0692830384,-0.0339656845,0.066425845,-0.322596252,0.1247427389,-0.2881054282,-0.122368142,0.0637505651,-0.0489776097,0.1619583219,0.1876238585,-0.3256686628,0.1354865581,0.2791850567,0.1854793131,0.0244828667,0.4316500723,0.0075259595,-0.2300830781,-0.023363851,0.2279523462,0.1796824038,-0.0562872812,-0.0279301107,0.100480549,-0.0691840649,-0.376296699,-0.193999365,0.0208728425,-0.4312960207,0.3942311704,0.0237804763,0.0594658144,0.0281341858,0.1458708793,0.1447972655,0.2050364465,0.010779175,0.1170280799,0.0932478532,0.0910386071,0.0838550776,-0.2934494615,0.2623491287,-0.1249917671,-0.4529618621,0.0569270551,-0.0704900026,-0.2504338026,0.0244053863,0.1059229895,0.2930746675,-0.2247793376,0.0423178449,-0.0464648195,-0.2383497357,-0.3965626061,0.2039445192,-0.1147942171,-0.198367089,0.2978394926,-0.1070739329,-0.3055641949,-0.0280588754,0.2836199701,0.0299851205,0.152165398,0.5996189117,-0.0773305595,-0.1156411916,-0.4316452146,0.0979604349,0.3189346194,0.1494747549,0.2477394789,-0.1895447522,-0.2431792468,0.0721193999,0.1885505766,0.037892025,0.1453275084,-0.2402581722,-0.2921876609,-0.4588644207,0.0046025463,-0.1703901291,0.2114720941,0.194389537,0.4413281083,-0.1352238059,0.236654982,-0.4279208481,-0.037296284,-0.4685045481,0.2483720928,-0.0478404127,-0.0903501287,0.0450875238,0.0513905063,0.3058929145,0.2208818644,-0.0576614812,-0.3870592415,-0.0921496823,0.0931526944,0.1092805713,0.0143305482,0.1976386756,0.2131322324,-0.2241609246,-0.2637848258,0.1043587327,0.1285088062,-0.0872338116,0.0787764937,0.1816832423,0.0326143913,-0.0256909542,-0.33613047,-0.0561857261,0.0320777707,-0.0667607933,0.3304432631,0.0313246921,-0.0545088127,0.0997803807,0.0593816601,0.2828300595,0.2285246551,0.0306528434,-0.1030482054,-0.3491645157,-0.0028985492,0.3846751153,0.2440446466,-0.2425766736,-0.0939171612,0.2190344483,0.3167302907,-0.231924966,-0.0727749914,0.2596981823,-0.380011797,0.3790821135,0.2261145115,0.0072065424,0.1852706224,0.1105445027,0.0050976626,0.2265835255,-0.3973679543,-0.0119793043,0.0900445506,-0.1410699338,0.1985625029,0.44924137,0.0734671876,0.2462897003,0.341006577,0.014755778,0.5114738345,-0.1246856526,0.0724941492,0.0589090176,-0.0415682569,0.3311864436,-0.0005669777,-0.0347976312,0.0089654345,-0.0880122706,0.3054791391,-0.3336259723,-0.0528675728,-0.2124442756,0.368789047,-0.1914831251,-0.3856400847,-0.23804757,-0.0818293467,-0.1413496882,-0.0785161257,-0.0190696828,0.1973137259,0.0451892503,-0.0820419714,0.0131340185,-0.2212535143,-0.0192149449,-0.0783587024,0.130819723,-0.2626004219,0.2001495957,0.1314703077,0.1404423863,-0.1671113372,0.6095541716,0.5709316134,0.3046991527,-0.1988588274,-0.1260323524,-0.3344715834,-0.1386544555,-0.2002345026,0.286361903,0.0003168109,-0.1747994572,0.2625156939,0.2772244811,-0.2658777535,-0.1564970165,0.4190347791,-0.1771449149,0.1225867122,0.1675085574,0.1353608817,-0.2269658446,0.0732962564,0.1038703024,-0.3389841914,-0.0345561467,0.2996349633,-0.2357952893,0.3266506493,-0.131075263,0.1637846828,0.0463033915,0.4507513344,0.0503231846,0.1739298105,-0.3142849505,-0.1331529915,-0.5782245398,0.2144250721,-0.1688652784,-0.3648037016,-0.1931368709,0.1571838707,0.1029490456,0.1065549478,0.2409303337,0.0778954029,-0.1090169325,0.2187099457,-0.3619211912,0.0869541839,0.1755951494,0.0519847982,0.204318434,-0.4335597754,0.2095237821,-0.0014672085,0.2383017242,-0.0144974636,0.0347771682,0.1814197898,-0.3401248157,0.2746292949,0.1093128175,0.5257284641,-0.4383835793,0.091044113,-0.1269073635,0.1606307179,-0.4419291317,0.0574725047,0.2414098531,0.1892359853,0.0872847959,-0.0311474353,-0.1315664053,0.1188577861,0.2362552434,-0.0922232419,-0.0139229549,0.3029530942,0.1060066298,0.0748861805,-0.1064884812,0.3399718106,-0.0225495938,0.2731781006,-0.4903123081,-0.4749884307,0.3061774373,-0.5121244192,-0.4921657443,-0.3859884441,0.288736403,-0.0939683691,0.0935930386,-0.3449763656,-0.0112365661,0.288376838,0.1060355008,-0.5911830068,0.1808962375,0.088660121,-0.0450419672,-0.0889836028,0.3249361813,0.0077151936,-0.1511936933,-0.0211812947,-0.1020610854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/847","title":"multiprocessing in dataset map \"can only test a child process\"","comments":"I'm having a similar issue but when I try to do multiprocessing with the `DataLoader`\r\n\r\nCode to reproduce:\r\n\r\n```\r\nfrom datasets import load_dataset\r\n\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='\/home\/ad\/Desktop\/bookcorpus', split='train[:1%]')\r\nbook_corpus = book_corpus.map(encode, batched=True, num_proc=20, load_from_cache_file=True, batch_size=5000)\r\nbook_corpus.set_format(type='torch', columns=['text', \"input_ids\", \"attention_mask\", \"token_type_ids\"])\r\n\r\nfrom transformers import DataCollatorForWholeWordMask\r\nfrom transformers import Trainer, TrainingArguments\r\n\r\ndata_collator = DataCollatorForWholeWordMask(\r\n tokenizer=tokenizer, mlm=True, mlm_probability=0.15)\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/mobile_linear_att_8L_128_128_03layerdrop_shared\",\r\n overwrite_output_dir=True,\r\n num_train_epochs=1,\r\n per_device_train_batch_size=64,\r\n save_steps=50,\r\n save_total_limit=2,\r\n logging_first_step=True,\r\n warmup_steps=100,\r\n logging_steps=50,\r\n gradient_accumulation_steps=1,\r\n fp16=True,\r\n **dataloader_num_workers=10**,\r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=book_corpus,\r\n tokenizer=tokenizer)\r\n\r\ntrainer.train()\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<timed eval> in <module>\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/transformers\/trainer.py in train(self, model_path, trial)\r\n 869 self.control = self.callback_handler.on_epoch_begin(self.args, self.state, self.control)\r\n 870 \r\n--> 871 for step, inputs in enumerate(epoch_iterator):\r\n 872 \r\n 873 # Skip past any already trained steps if resuming training\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 433 if self._sampler_iter is None:\r\n 434 self._reset()\r\n--> 435 data = self._next_data()\r\n 436 self._num_yielded += 1\r\n 437 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 1083 else:\r\n 1084 del self._task_info[idx]\r\n-> 1085 return self._process_data(data)\r\n 1086 \r\n 1087 def _try_put_index(self):\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in _process_data(self, data)\r\n 1109 self._try_put_index()\r\n 1110 if isinstance(data, ExceptionWrapper):\r\n-> 1111 data.reraise()\r\n 1112 return data\r\n 1113 \r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/_utils.py in reraise(self)\r\n 426 # have message field\r\n 427 raise self.exc_type(message=msg)\r\n--> 428 raise self.exc_type(msg)\r\n 429 \r\n 430 \r\n\r\nAssertionError: Caught AssertionError in DataLoader worker process 0.\r\nOriginal Traceback (most recent call last):\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/worker.py\", line 198, in _worker_loop\r\n data = fetcher.fetch(index)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1087, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1074, in _getitem\r\n format_kwargs=format_kwargs,\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/warnings.py\", line 101, in _showwarnmsg\r\n _showwarnmsg_impl(msg)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/warnings.py\", line 30, in _showwarnmsg_impl\r\n file.write(text)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 723, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 153, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 158, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 456, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n```\r\n\r\nAs a workaround I have commented line 456 and 457 in `\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py`","body":"Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```","comment_length":383,"text":"multiprocessing in dataset map \"can only test a child process\"\nUsing a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```\nI'm having a similar issue but when I try to do multiprocessing with the `DataLoader`\r\n\r\nCode to reproduce:\r\n\r\n```\r\nfrom datasets import load_dataset\r\n\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='\/home\/ad\/Desktop\/bookcorpus', split='train[:1%]')\r\nbook_corpus = book_corpus.map(encode, batched=True, num_proc=20, load_from_cache_file=True, batch_size=5000)\r\nbook_corpus.set_format(type='torch', columns=['text', \"input_ids\", \"attention_mask\", \"token_type_ids\"])\r\n\r\nfrom transformers import DataCollatorForWholeWordMask\r\nfrom transformers import Trainer, TrainingArguments\r\n\r\ndata_collator = DataCollatorForWholeWordMask(\r\n tokenizer=tokenizer, mlm=True, mlm_probability=0.15)\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/mobile_linear_att_8L_128_128_03layerdrop_shared\",\r\n overwrite_output_dir=True,\r\n num_train_epochs=1,\r\n per_device_train_batch_size=64,\r\n save_steps=50,\r\n save_total_limit=2,\r\n logging_first_step=True,\r\n warmup_steps=100,\r\n logging_steps=50,\r\n gradient_accumulation_steps=1,\r\n fp16=True,\r\n **dataloader_num_workers=10**,\r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=book_corpus,\r\n tokenizer=tokenizer)\r\n\r\ntrainer.train()\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<timed eval> in <module>\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/transformers\/trainer.py in train(self, model_path, trial)\r\n 869 self.control = self.callback_handler.on_epoch_begin(self.args, self.state, self.control)\r\n 870 \r\n--> 871 for step, inputs in enumerate(epoch_iterator):\r\n 872 \r\n 873 # Skip past any already trained steps if resuming training\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 433 if self._sampler_iter is None:\r\n 434 self._reset()\r\n--> 435 data = self._next_data()\r\n 436 self._num_yielded += 1\r\n 437 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 1083 else:\r\n 1084 del self._task_info[idx]\r\n-> 1085 return self._process_data(data)\r\n 1086 \r\n 1087 def _try_put_index(self):\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in _process_data(self, data)\r\n 1109 self._try_put_index()\r\n 1110 if isinstance(data, ExceptionWrapper):\r\n-> 1111 data.reraise()\r\n 1112 return data\r\n 1113 \r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/_utils.py in reraise(self)\r\n 426 # have message field\r\n 427 raise self.exc_type(message=msg)\r\n--> 428 raise self.exc_type(msg)\r\n 429 \r\n 430 \r\n\r\nAssertionError: Caught AssertionError in DataLoader worker process 0.\r\nOriginal Traceback (most recent call last):\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/worker.py\", line 198, in _worker_loop\r\n data = fetcher.fetch(index)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1087, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1074, in _getitem\r\n format_kwargs=format_kwargs,\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 890, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 851, in command\r\n return torch.tensor(x, **format_kwargs)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/warnings.py\", line 101, in _showwarnmsg\r\n _showwarnmsg_impl(msg)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/warnings.py\", line 30, in _showwarnmsg_impl\r\n file.write(text)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 723, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 153, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 158, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 456, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n```\r\n\r\nAs a workaround I have commented line 456 and 457 in `\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py`","embeddings":[-0.3956997097,0.0156290922,-0.1538687646,-0.1850299388,0.141797632,-0.0578565337,0.5068988204,0.3400265574,-0.0834040865,0.1681774557,-0.0184273254,0.3370628655,0.0475015864,0.0918332115,-0.2264437228,0.1021468788,-0.0123605253,0.0946920067,0.1290109754,0.0473636836,-0.1753095239,0.1040760726,-0.2302506417,0.2859587967,-0.4516958296,-0.285730809,-0.0163212903,-0.0623164959,-0.2643237412,-0.4696436226,-0.149113372,0.2851142585,-0.1861115098,0.6030860543,-0.0001044559,0.0936724842,0.2627740204,0.0446500145,-0.054140687,-0.030070236,0.095131956,-0.0515859984,-0.0310001038,-0.4035981894,-0.0044792355,-0.2674306333,0.1779136211,-0.4431346357,0.2878316641,0.3189612627,0.2665320337,0.5655280352,-0.2092227638,-0.0047211065,-0.201201424,-0.0783623531,-0.0357206613,-0.1625539958,0.2167165279,-0.2403003126,-0.3567937613,0.2491746992,-0.0430285074,0.1743139625,-0.3290867805,0.0007867881,0.277649343,-0.237424776,0.3172336221,0.079235658,0.1507846713,-0.2674021721,-0.3465930223,-0.1820102185,-0.2675664425,-0.1380905807,-0.0896170512,0.0866119117,-0.3090318441,0.0237945151,-0.231428802,0.0673615932,-0.0069425311,-0.0971344262,-0.20974648,0.6421571374,0.0204102993,0.2258118987,-0.058426097,0.0421580598,0.1321844459,-0.1236515045,0.1550007761,-0.0283196401,-0.1767590791,-0.1339012235,0.1732696593,-0.3608185351,0.1934940219,-0.0518135242,-0.2702677846,0.1569519043,0.1210899875,0.1657909304,0.1612778902,0.0505134016,0.1559104919,0.2796486914,0.2925396562,-0.196995303,-0.3908066452,-0.0121860905,0.0276292767,-0.2749152184,0.19010818,0.1587806493,0.1859900355,0.0203872453,-0.316894412,-0.0674285442,-0.1310207993,-0.0011093016,0.0109488498,0.4655701518,-0.0625771657,0.210119307,-0.1869856268,0.2766140103,-0.1955184191,-0.1335654557,-0.2674530447,0.1599751413,-0.105241254,-0.0912755728,0.0864879191,0.2284782082,0.3958972394,0.057422515,0.0732819885,-0.0852145255,0.4325413704,-0.219634518,0.2467176765,0.0557839759,0.068621248,0.1359149069,0.26744017,-0.3098887801,-0.1797690243,-0.0748893917,0.0085688289,-0.1883713305,-0.0060650436,0.2934336662,0.0234699752,0.0294385776,-0.1553809643,0.1325114965,0.2363231629,0.0417060144,0.0084723802,-0.1836157143,-0.2125665992,-0.1611631811,0.1659645736,0.2332042754,-0.1246359199,-0.0227329582,0.0209111851,-0.1451625675,0.2747372985,0.2273871899,-0.0913355052,0.6480855346,-0.1730043739,0.5878829956,0.1861093491,-0.5310037136,-0.1901540011,0.2092227191,-0.3177124858,-0.0111948214,0.1317908615,-0.1686787903,0.3851614892,0.080286935,0.1898801774,0.2464483976,-0.1104762852,0.2880237997,-0.2860673666,-0.0096036047,0.0849479064,-0.071969904,0.2393062264,-0.2704798877,-0.0424101911,-0.1621235162,0.2372229397,-0.1482038051,0.2230832577,-0.0443854667,0.0625272393,0.0818433464,-0.0392007828,-0.1215575859,0.0788098946,0.2157844901,0.0433690101,0.0418133661,-0.1905940175,-0.1984119862,-0.0963724405,0.2760661244,-0.2941963971,-0.3430737555,0.3505134284,0.3571797907,-0.0198514294,-0.0224457197,-0.0754135773,0.1201187074,0.1530826986,0.0062083742,0.0715006441,-0.0785518736,-0.1991024315,-0.1974657774,-0.0115251811,0.0726843029,0.1600964069,0.003486624,-0.0784358084,0.4022406936,0.2385911644,0.1009370983,0.2098905593,-0.1776235253,-0.0964289233,0.0071378755,-0.0728402957,0.111530818,0.0388300754,-0.0264948942,0.1444305331,0.4932917356,0.2365600765,0.1932385713,0.2239498198,-0.0101361321,0.2388229668,-0.1069113389,-0.0609275028,-0.1341272146,0.057798285,0.0031344036,0.1119551882,-0.1359223723,-0.2041955888,0.1788984239,0.314935714,-0.0856286287,0.1352984309,-0.1058834121,-0.0881043524,-0.1929530203,0.0814027041,0.1926788986,0.3943273127,0.2004811913,0.1856659055,-0.0386409611,-0.0774591267,-0.0734680966,0.1406773478,0.0299724508,0.1926366687,0.2765003741,0.2475029826,-0.2172683924,-0.3548972011,-0.2435459346,0.1768790036,0.1148345843,-0.1390989274,0.0299772602,-0.3175421357,0.0918787643,-0.0323445313,-0.0665194169,0.2071175724,-0.3488472402,0.1094204113,0.1852676868,-0.1694183499,0.3220192492,-0.1171384901,0.0898978636,-0.0134830419,-0.0217510052,-0.0988093317,-0.1474416703,-0.1254175901,0.1327666491,0.1662156433,0.1449765265,0.4963173568,0.0013359481,-0.2189733684,-0.5480391383,-0.2451718748,0.0971291885,-0.1633996367,0.1954907775,0.2778953016,0.0427568778,-0.0250689685,-0.3409390152,0.3002497554,-0.4462933838,-0.373233676,-0.2133507133,0.0948979482,-0.1722041965,-0.0335693397,-0.3764518201,-0.3236232698,-0.1813369691,0.3523524106,-0.0931495354,0.1716113091,-0.155457288,-0.1054516658,0.1455910206,-0.0203673467,0.1121426821,-0.2187279463,-0.1448161304,-0.0194471348,-0.1510415971,-0.283069551,-0.1560443342,-0.140906021,0.4374369085,-0.10340821,-0.1784181744,-0.0407508872,-0.1612140536,0.4367106259,0.0242293067,-0.0490040779,0.3442159593,0.2696896195,-0.348139286,-0.0873824731,-0.16058743,0.0496429652,-0.2982184887,-0.1103075147,-0.185685128,0.4214314222,0.2078024745,0.5972415209,0.028514469,-0.0613885559,0.1324319392,-0.4377176166,-0.0172750056,-0.0337198712,-0.4259907305,0.1626303345,-0.2667844892,-0.1014861315,0.1350760311,-0.1677293479,-0.115086928,0.1514069438,0.0009905943,-0.3835220933,-0.209966138,0.0377299711,-0.1925776899,0.0860434994,0.0603309199,0.0141091933,-0.4112586379,0.2260724902,-0.0721928179,-0.0218361951,-0.1237445623,-0.3398272991,-0.3464430869,0.0175716821,-0.6085758209,0.3544566035,0.1318598092,0.4851483107,0.1367789209,0.0052999631,0.0414658636,-0.2825492918,0.6524977684,-0.2151976973,0.0749009252,0.183681041,-0.3277136385,-0.2910260558,-0.0436595753,-0.3557530344,0.4460589588,0.3172533214,0.4865555763,-0.1918331534,-0.3518019319,0.1224313006,-0.0199515596,0.0334758908,0.0118676918,-0.28859815,-0.3060879111,-0.4088720381,0.3470336199,0.088713713,0.1813630015,-0.0699506551,-0.1245866269,-0.1805427074,-0.0352943242,0.1069817096,0.1873820275,-0.0616821796,-0.1572833955,0.1408728659,-0.0219112542,0.2116450667,0.2355050445,0.0500611067,-0.0951047987,-0.1150436103,0.0940205231,0.0208330434,0.4114652574,0.2416535616,-0.1685579568,0.3699742556,-0.1364515871,0.288634181,-0.110836409,0.4739766717,0.5543159246,0.1899577975,-0.1046064273,-0.1071350798,0.0384807661,0.1414999962,-0.1566073895,0.3062249124,-0.2532045543,-0.2288146019,0.1476330608,0.1779163629,0.5770257115,-0.0932503566,0.0875760093,0.0692830384,-0.0339656845,0.066425845,-0.322596252,0.1247427389,-0.2881054282,-0.122368142,0.0637505651,-0.0489776097,0.1619583219,0.1876238585,-0.3256686628,0.1354865581,0.2791850567,0.1854793131,0.0244828667,0.4316500723,0.0075259595,-0.2300830781,-0.023363851,0.2279523462,0.1796824038,-0.0562872812,-0.0279301107,0.100480549,-0.0691840649,-0.376296699,-0.193999365,0.0208728425,-0.4312960207,0.3942311704,0.0237804763,0.0594658144,0.0281341858,0.1458708793,0.1447972655,0.2050364465,0.010779175,0.1170280799,0.0932478532,0.0910386071,0.0838550776,-0.2934494615,0.2623491287,-0.1249917671,-0.4529618621,0.0569270551,-0.0704900026,-0.2504338026,0.0244053863,0.1059229895,0.2930746675,-0.2247793376,0.0423178449,-0.0464648195,-0.2383497357,-0.3965626061,0.2039445192,-0.1147942171,-0.198367089,0.2978394926,-0.1070739329,-0.3055641949,-0.0280588754,0.2836199701,0.0299851205,0.152165398,0.5996189117,-0.0773305595,-0.1156411916,-0.4316452146,0.0979604349,0.3189346194,0.1494747549,0.2477394789,-0.1895447522,-0.2431792468,0.0721193999,0.1885505766,0.037892025,0.1453275084,-0.2402581722,-0.2921876609,-0.4588644207,0.0046025463,-0.1703901291,0.2114720941,0.194389537,0.4413281083,-0.1352238059,0.236654982,-0.4279208481,-0.037296284,-0.4685045481,0.2483720928,-0.0478404127,-0.0903501287,0.0450875238,0.0513905063,0.3058929145,0.2208818644,-0.0576614812,-0.3870592415,-0.0921496823,0.0931526944,0.1092805713,0.0143305482,0.1976386756,0.2131322324,-0.2241609246,-0.2637848258,0.1043587327,0.1285088062,-0.0872338116,0.0787764937,0.1816832423,0.0326143913,-0.0256909542,-0.33613047,-0.0561857261,0.0320777707,-0.0667607933,0.3304432631,0.0313246921,-0.0545088127,0.0997803807,0.0593816601,0.2828300595,0.2285246551,0.0306528434,-0.1030482054,-0.3491645157,-0.0028985492,0.3846751153,0.2440446466,-0.2425766736,-0.0939171612,0.2190344483,0.3167302907,-0.231924966,-0.0727749914,0.2596981823,-0.380011797,0.3790821135,0.2261145115,0.0072065424,0.1852706224,0.1105445027,0.0050976626,0.2265835255,-0.3973679543,-0.0119793043,0.0900445506,-0.1410699338,0.1985625029,0.44924137,0.0734671876,0.2462897003,0.341006577,0.014755778,0.5114738345,-0.1246856526,0.0724941492,0.0589090176,-0.0415682569,0.3311864436,-0.0005669777,-0.0347976312,0.0089654345,-0.0880122706,0.3054791391,-0.3336259723,-0.0528675728,-0.2124442756,0.368789047,-0.1914831251,-0.3856400847,-0.23804757,-0.0818293467,-0.1413496882,-0.0785161257,-0.0190696828,0.1973137259,0.0451892503,-0.0820419714,0.0131340185,-0.2212535143,-0.0192149449,-0.0783587024,0.130819723,-0.2626004219,0.2001495957,0.1314703077,0.1404423863,-0.1671113372,0.6095541716,0.5709316134,0.3046991527,-0.1988588274,-0.1260323524,-0.3344715834,-0.1386544555,-0.2002345026,0.286361903,0.0003168109,-0.1747994572,0.2625156939,0.2772244811,-0.2658777535,-0.1564970165,0.4190347791,-0.1771449149,0.1225867122,0.1675085574,0.1353608817,-0.2269658446,0.0732962564,0.1038703024,-0.3389841914,-0.0345561467,0.2996349633,-0.2357952893,0.3266506493,-0.131075263,0.1637846828,0.0463033915,0.4507513344,0.0503231846,0.1739298105,-0.3142849505,-0.1331529915,-0.5782245398,0.2144250721,-0.1688652784,-0.3648037016,-0.1931368709,0.1571838707,0.1029490456,0.1065549478,0.2409303337,0.0778954029,-0.1090169325,0.2187099457,-0.3619211912,0.0869541839,0.1755951494,0.0519847982,0.204318434,-0.4335597754,0.2095237821,-0.0014672085,0.2383017242,-0.0144974636,0.0347771682,0.1814197898,-0.3401248157,0.2746292949,0.1093128175,0.5257284641,-0.4383835793,0.091044113,-0.1269073635,0.1606307179,-0.4419291317,0.0574725047,0.2414098531,0.1892359853,0.0872847959,-0.0311474353,-0.1315664053,0.1188577861,0.2362552434,-0.0922232419,-0.0139229549,0.3029530942,0.1060066298,0.0748861805,-0.1064884812,0.3399718106,-0.0225495938,0.2731781006,-0.4903123081,-0.4749884307,0.3061774373,-0.5121244192,-0.4921657443,-0.3859884441,0.288736403,-0.0939683691,0.0935930386,-0.3449763656,-0.0112365661,0.288376838,0.1060355008,-0.5911830068,0.1808962375,0.088660121,-0.0450419672,-0.0889836028,0.3249361813,0.0077151936,-0.1511936933,-0.0211812947,-0.1020610854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/847","title":"multiprocessing in dataset map \"can only test a child process\"","comments":"Isn't it more the pytorch warning on the use of non-writable memory for tensor that trigger this here @lhoestq? (since it seems to be a warning triggered in `torch.tensor()`","body":"Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```","comment_length":29,"text":"multiprocessing in dataset map \"can only test a child process\"\nUsing a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```\nIsn't it more the pytorch warning on the use of non-writable memory for tensor that trigger this here @lhoestq? (since it seems to be a warning triggered in `torch.tensor()`","embeddings":[-0.3956997097,0.0156290922,-0.1538687646,-0.1850299388,0.141797632,-0.0578565337,0.5068988204,0.3400265574,-0.0834040865,0.1681774557,-0.0184273254,0.3370628655,0.0475015864,0.0918332115,-0.2264437228,0.1021468788,-0.0123605253,0.0946920067,0.1290109754,0.0473636836,-0.1753095239,0.1040760726,-0.2302506417,0.2859587967,-0.4516958296,-0.285730809,-0.0163212903,-0.0623164959,-0.2643237412,-0.4696436226,-0.149113372,0.2851142585,-0.1861115098,0.6030860543,-0.0001044559,0.0936724842,0.2627740204,0.0446500145,-0.054140687,-0.030070236,0.095131956,-0.0515859984,-0.0310001038,-0.4035981894,-0.0044792355,-0.2674306333,0.1779136211,-0.4431346357,0.2878316641,0.3189612627,0.2665320337,0.5655280352,-0.2092227638,-0.0047211065,-0.201201424,-0.0783623531,-0.0357206613,-0.1625539958,0.2167165279,-0.2403003126,-0.3567937613,0.2491746992,-0.0430285074,0.1743139625,-0.3290867805,0.0007867881,0.277649343,-0.237424776,0.3172336221,0.079235658,0.1507846713,-0.2674021721,-0.3465930223,-0.1820102185,-0.2675664425,-0.1380905807,-0.0896170512,0.0866119117,-0.3090318441,0.0237945151,-0.231428802,0.0673615932,-0.0069425311,-0.0971344262,-0.20974648,0.6421571374,0.0204102993,0.2258118987,-0.058426097,0.0421580598,0.1321844459,-0.1236515045,0.1550007761,-0.0283196401,-0.1767590791,-0.1339012235,0.1732696593,-0.3608185351,0.1934940219,-0.0518135242,-0.2702677846,0.1569519043,0.1210899875,0.1657909304,0.1612778902,0.0505134016,0.1559104919,0.2796486914,0.2925396562,-0.196995303,-0.3908066452,-0.0121860905,0.0276292767,-0.2749152184,0.19010818,0.1587806493,0.1859900355,0.0203872453,-0.316894412,-0.0674285442,-0.1310207993,-0.0011093016,0.0109488498,0.4655701518,-0.0625771657,0.210119307,-0.1869856268,0.2766140103,-0.1955184191,-0.1335654557,-0.2674530447,0.1599751413,-0.105241254,-0.0912755728,0.0864879191,0.2284782082,0.3958972394,0.057422515,0.0732819885,-0.0852145255,0.4325413704,-0.219634518,0.2467176765,0.0557839759,0.068621248,0.1359149069,0.26744017,-0.3098887801,-0.1797690243,-0.0748893917,0.0085688289,-0.1883713305,-0.0060650436,0.2934336662,0.0234699752,0.0294385776,-0.1553809643,0.1325114965,0.2363231629,0.0417060144,0.0084723802,-0.1836157143,-0.2125665992,-0.1611631811,0.1659645736,0.2332042754,-0.1246359199,-0.0227329582,0.0209111851,-0.1451625675,0.2747372985,0.2273871899,-0.0913355052,0.6480855346,-0.1730043739,0.5878829956,0.1861093491,-0.5310037136,-0.1901540011,0.2092227191,-0.3177124858,-0.0111948214,0.1317908615,-0.1686787903,0.3851614892,0.080286935,0.1898801774,0.2464483976,-0.1104762852,0.2880237997,-0.2860673666,-0.0096036047,0.0849479064,-0.071969904,0.2393062264,-0.2704798877,-0.0424101911,-0.1621235162,0.2372229397,-0.1482038051,0.2230832577,-0.0443854667,0.0625272393,0.0818433464,-0.0392007828,-0.1215575859,0.0788098946,0.2157844901,0.0433690101,0.0418133661,-0.1905940175,-0.1984119862,-0.0963724405,0.2760661244,-0.2941963971,-0.3430737555,0.3505134284,0.3571797907,-0.0198514294,-0.0224457197,-0.0754135773,0.1201187074,0.1530826986,0.0062083742,0.0715006441,-0.0785518736,-0.1991024315,-0.1974657774,-0.0115251811,0.0726843029,0.1600964069,0.003486624,-0.0784358084,0.4022406936,0.2385911644,0.1009370983,0.2098905593,-0.1776235253,-0.0964289233,0.0071378755,-0.0728402957,0.111530818,0.0388300754,-0.0264948942,0.1444305331,0.4932917356,0.2365600765,0.1932385713,0.2239498198,-0.0101361321,0.2388229668,-0.1069113389,-0.0609275028,-0.1341272146,0.057798285,0.0031344036,0.1119551882,-0.1359223723,-0.2041955888,0.1788984239,0.314935714,-0.0856286287,0.1352984309,-0.1058834121,-0.0881043524,-0.1929530203,0.0814027041,0.1926788986,0.3943273127,0.2004811913,0.1856659055,-0.0386409611,-0.0774591267,-0.0734680966,0.1406773478,0.0299724508,0.1926366687,0.2765003741,0.2475029826,-0.2172683924,-0.3548972011,-0.2435459346,0.1768790036,0.1148345843,-0.1390989274,0.0299772602,-0.3175421357,0.0918787643,-0.0323445313,-0.0665194169,0.2071175724,-0.3488472402,0.1094204113,0.1852676868,-0.1694183499,0.3220192492,-0.1171384901,0.0898978636,-0.0134830419,-0.0217510052,-0.0988093317,-0.1474416703,-0.1254175901,0.1327666491,0.1662156433,0.1449765265,0.4963173568,0.0013359481,-0.2189733684,-0.5480391383,-0.2451718748,0.0971291885,-0.1633996367,0.1954907775,0.2778953016,0.0427568778,-0.0250689685,-0.3409390152,0.3002497554,-0.4462933838,-0.373233676,-0.2133507133,0.0948979482,-0.1722041965,-0.0335693397,-0.3764518201,-0.3236232698,-0.1813369691,0.3523524106,-0.0931495354,0.1716113091,-0.155457288,-0.1054516658,0.1455910206,-0.0203673467,0.1121426821,-0.2187279463,-0.1448161304,-0.0194471348,-0.1510415971,-0.283069551,-0.1560443342,-0.140906021,0.4374369085,-0.10340821,-0.1784181744,-0.0407508872,-0.1612140536,0.4367106259,0.0242293067,-0.0490040779,0.3442159593,0.2696896195,-0.348139286,-0.0873824731,-0.16058743,0.0496429652,-0.2982184887,-0.1103075147,-0.185685128,0.4214314222,0.2078024745,0.5972415209,0.028514469,-0.0613885559,0.1324319392,-0.4377176166,-0.0172750056,-0.0337198712,-0.4259907305,0.1626303345,-0.2667844892,-0.1014861315,0.1350760311,-0.1677293479,-0.115086928,0.1514069438,0.0009905943,-0.3835220933,-0.209966138,0.0377299711,-0.1925776899,0.0860434994,0.0603309199,0.0141091933,-0.4112586379,0.2260724902,-0.0721928179,-0.0218361951,-0.1237445623,-0.3398272991,-0.3464430869,0.0175716821,-0.6085758209,0.3544566035,0.1318598092,0.4851483107,0.1367789209,0.0052999631,0.0414658636,-0.2825492918,0.6524977684,-0.2151976973,0.0749009252,0.183681041,-0.3277136385,-0.2910260558,-0.0436595753,-0.3557530344,0.4460589588,0.3172533214,0.4865555763,-0.1918331534,-0.3518019319,0.1224313006,-0.0199515596,0.0334758908,0.0118676918,-0.28859815,-0.3060879111,-0.4088720381,0.3470336199,0.088713713,0.1813630015,-0.0699506551,-0.1245866269,-0.1805427074,-0.0352943242,0.1069817096,0.1873820275,-0.0616821796,-0.1572833955,0.1408728659,-0.0219112542,0.2116450667,0.2355050445,0.0500611067,-0.0951047987,-0.1150436103,0.0940205231,0.0208330434,0.4114652574,0.2416535616,-0.1685579568,0.3699742556,-0.1364515871,0.288634181,-0.110836409,0.4739766717,0.5543159246,0.1899577975,-0.1046064273,-0.1071350798,0.0384807661,0.1414999962,-0.1566073895,0.3062249124,-0.2532045543,-0.2288146019,0.1476330608,0.1779163629,0.5770257115,-0.0932503566,0.0875760093,0.0692830384,-0.0339656845,0.066425845,-0.322596252,0.1247427389,-0.2881054282,-0.122368142,0.0637505651,-0.0489776097,0.1619583219,0.1876238585,-0.3256686628,0.1354865581,0.2791850567,0.1854793131,0.0244828667,0.4316500723,0.0075259595,-0.2300830781,-0.023363851,0.2279523462,0.1796824038,-0.0562872812,-0.0279301107,0.100480549,-0.0691840649,-0.376296699,-0.193999365,0.0208728425,-0.4312960207,0.3942311704,0.0237804763,0.0594658144,0.0281341858,0.1458708793,0.1447972655,0.2050364465,0.010779175,0.1170280799,0.0932478532,0.0910386071,0.0838550776,-0.2934494615,0.2623491287,-0.1249917671,-0.4529618621,0.0569270551,-0.0704900026,-0.2504338026,0.0244053863,0.1059229895,0.2930746675,-0.2247793376,0.0423178449,-0.0464648195,-0.2383497357,-0.3965626061,0.2039445192,-0.1147942171,-0.198367089,0.2978394926,-0.1070739329,-0.3055641949,-0.0280588754,0.2836199701,0.0299851205,0.152165398,0.5996189117,-0.0773305595,-0.1156411916,-0.4316452146,0.0979604349,0.3189346194,0.1494747549,0.2477394789,-0.1895447522,-0.2431792468,0.0721193999,0.1885505766,0.037892025,0.1453275084,-0.2402581722,-0.2921876609,-0.4588644207,0.0046025463,-0.1703901291,0.2114720941,0.194389537,0.4413281083,-0.1352238059,0.236654982,-0.4279208481,-0.037296284,-0.4685045481,0.2483720928,-0.0478404127,-0.0903501287,0.0450875238,0.0513905063,0.3058929145,0.2208818644,-0.0576614812,-0.3870592415,-0.0921496823,0.0931526944,0.1092805713,0.0143305482,0.1976386756,0.2131322324,-0.2241609246,-0.2637848258,0.1043587327,0.1285088062,-0.0872338116,0.0787764937,0.1816832423,0.0326143913,-0.0256909542,-0.33613047,-0.0561857261,0.0320777707,-0.0667607933,0.3304432631,0.0313246921,-0.0545088127,0.0997803807,0.0593816601,0.2828300595,0.2285246551,0.0306528434,-0.1030482054,-0.3491645157,-0.0028985492,0.3846751153,0.2440446466,-0.2425766736,-0.0939171612,0.2190344483,0.3167302907,-0.231924966,-0.0727749914,0.2596981823,-0.380011797,0.3790821135,0.2261145115,0.0072065424,0.1852706224,0.1105445027,0.0050976626,0.2265835255,-0.3973679543,-0.0119793043,0.0900445506,-0.1410699338,0.1985625029,0.44924137,0.0734671876,0.2462897003,0.341006577,0.014755778,0.5114738345,-0.1246856526,0.0724941492,0.0589090176,-0.0415682569,0.3311864436,-0.0005669777,-0.0347976312,0.0089654345,-0.0880122706,0.3054791391,-0.3336259723,-0.0528675728,-0.2124442756,0.368789047,-0.1914831251,-0.3856400847,-0.23804757,-0.0818293467,-0.1413496882,-0.0785161257,-0.0190696828,0.1973137259,0.0451892503,-0.0820419714,0.0131340185,-0.2212535143,-0.0192149449,-0.0783587024,0.130819723,-0.2626004219,0.2001495957,0.1314703077,0.1404423863,-0.1671113372,0.6095541716,0.5709316134,0.3046991527,-0.1988588274,-0.1260323524,-0.3344715834,-0.1386544555,-0.2002345026,0.286361903,0.0003168109,-0.1747994572,0.2625156939,0.2772244811,-0.2658777535,-0.1564970165,0.4190347791,-0.1771449149,0.1225867122,0.1675085574,0.1353608817,-0.2269658446,0.0732962564,0.1038703024,-0.3389841914,-0.0345561467,0.2996349633,-0.2357952893,0.3266506493,-0.131075263,0.1637846828,0.0463033915,0.4507513344,0.0503231846,0.1739298105,-0.3142849505,-0.1331529915,-0.5782245398,0.2144250721,-0.1688652784,-0.3648037016,-0.1931368709,0.1571838707,0.1029490456,0.1065549478,0.2409303337,0.0778954029,-0.1090169325,0.2187099457,-0.3619211912,0.0869541839,0.1755951494,0.0519847982,0.204318434,-0.4335597754,0.2095237821,-0.0014672085,0.2383017242,-0.0144974636,0.0347771682,0.1814197898,-0.3401248157,0.2746292949,0.1093128175,0.5257284641,-0.4383835793,0.091044113,-0.1269073635,0.1606307179,-0.4419291317,0.0574725047,0.2414098531,0.1892359853,0.0872847959,-0.0311474353,-0.1315664053,0.1188577861,0.2362552434,-0.0922232419,-0.0139229549,0.3029530942,0.1060066298,0.0748861805,-0.1064884812,0.3399718106,-0.0225495938,0.2731781006,-0.4903123081,-0.4749884307,0.3061774373,-0.5121244192,-0.4921657443,-0.3859884441,0.288736403,-0.0939683691,0.0935930386,-0.3449763656,-0.0112365661,0.288376838,0.1060355008,-0.5911830068,0.1808962375,0.088660121,-0.0450419672,-0.0889836028,0.3249361813,0.0077151936,-0.1511936933,-0.0211812947,-0.1020610854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/847","title":"multiprocessing in dataset map \"can only test a child process\"","comments":"Yep this time this is a warning from pytorch that causes wandb to not work properly.\r\nCould this by a wandb issue ?","body":"Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```","comment_length":23,"text":"multiprocessing in dataset map \"can only test a child process\"\nUsing a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```\nYep this time this is a warning from pytorch that causes wandb to not work properly.\r\nCould this by a wandb issue ?","embeddings":[-0.3956997097,0.0156290922,-0.1538687646,-0.1850299388,0.141797632,-0.0578565337,0.5068988204,0.3400265574,-0.0834040865,0.1681774557,-0.0184273254,0.3370628655,0.0475015864,0.0918332115,-0.2264437228,0.1021468788,-0.0123605253,0.0946920067,0.1290109754,0.0473636836,-0.1753095239,0.1040760726,-0.2302506417,0.2859587967,-0.4516958296,-0.285730809,-0.0163212903,-0.0623164959,-0.2643237412,-0.4696436226,-0.149113372,0.2851142585,-0.1861115098,0.6030860543,-0.0001044559,0.0936724842,0.2627740204,0.0446500145,-0.054140687,-0.030070236,0.095131956,-0.0515859984,-0.0310001038,-0.4035981894,-0.0044792355,-0.2674306333,0.1779136211,-0.4431346357,0.2878316641,0.3189612627,0.2665320337,0.5655280352,-0.2092227638,-0.0047211065,-0.201201424,-0.0783623531,-0.0357206613,-0.1625539958,0.2167165279,-0.2403003126,-0.3567937613,0.2491746992,-0.0430285074,0.1743139625,-0.3290867805,0.0007867881,0.277649343,-0.237424776,0.3172336221,0.079235658,0.1507846713,-0.2674021721,-0.3465930223,-0.1820102185,-0.2675664425,-0.1380905807,-0.0896170512,0.0866119117,-0.3090318441,0.0237945151,-0.231428802,0.0673615932,-0.0069425311,-0.0971344262,-0.20974648,0.6421571374,0.0204102993,0.2258118987,-0.058426097,0.0421580598,0.1321844459,-0.1236515045,0.1550007761,-0.0283196401,-0.1767590791,-0.1339012235,0.1732696593,-0.3608185351,0.1934940219,-0.0518135242,-0.2702677846,0.1569519043,0.1210899875,0.1657909304,0.1612778902,0.0505134016,0.1559104919,0.2796486914,0.2925396562,-0.196995303,-0.3908066452,-0.0121860905,0.0276292767,-0.2749152184,0.19010818,0.1587806493,0.1859900355,0.0203872453,-0.316894412,-0.0674285442,-0.1310207993,-0.0011093016,0.0109488498,0.4655701518,-0.0625771657,0.210119307,-0.1869856268,0.2766140103,-0.1955184191,-0.1335654557,-0.2674530447,0.1599751413,-0.105241254,-0.0912755728,0.0864879191,0.2284782082,0.3958972394,0.057422515,0.0732819885,-0.0852145255,0.4325413704,-0.219634518,0.2467176765,0.0557839759,0.068621248,0.1359149069,0.26744017,-0.3098887801,-0.1797690243,-0.0748893917,0.0085688289,-0.1883713305,-0.0060650436,0.2934336662,0.0234699752,0.0294385776,-0.1553809643,0.1325114965,0.2363231629,0.0417060144,0.0084723802,-0.1836157143,-0.2125665992,-0.1611631811,0.1659645736,0.2332042754,-0.1246359199,-0.0227329582,0.0209111851,-0.1451625675,0.2747372985,0.2273871899,-0.0913355052,0.6480855346,-0.1730043739,0.5878829956,0.1861093491,-0.5310037136,-0.1901540011,0.2092227191,-0.3177124858,-0.0111948214,0.1317908615,-0.1686787903,0.3851614892,0.080286935,0.1898801774,0.2464483976,-0.1104762852,0.2880237997,-0.2860673666,-0.0096036047,0.0849479064,-0.071969904,0.2393062264,-0.2704798877,-0.0424101911,-0.1621235162,0.2372229397,-0.1482038051,0.2230832577,-0.0443854667,0.0625272393,0.0818433464,-0.0392007828,-0.1215575859,0.0788098946,0.2157844901,0.0433690101,0.0418133661,-0.1905940175,-0.1984119862,-0.0963724405,0.2760661244,-0.2941963971,-0.3430737555,0.3505134284,0.3571797907,-0.0198514294,-0.0224457197,-0.0754135773,0.1201187074,0.1530826986,0.0062083742,0.0715006441,-0.0785518736,-0.1991024315,-0.1974657774,-0.0115251811,0.0726843029,0.1600964069,0.003486624,-0.0784358084,0.4022406936,0.2385911644,0.1009370983,0.2098905593,-0.1776235253,-0.0964289233,0.0071378755,-0.0728402957,0.111530818,0.0388300754,-0.0264948942,0.1444305331,0.4932917356,0.2365600765,0.1932385713,0.2239498198,-0.0101361321,0.2388229668,-0.1069113389,-0.0609275028,-0.1341272146,0.057798285,0.0031344036,0.1119551882,-0.1359223723,-0.2041955888,0.1788984239,0.314935714,-0.0856286287,0.1352984309,-0.1058834121,-0.0881043524,-0.1929530203,0.0814027041,0.1926788986,0.3943273127,0.2004811913,0.1856659055,-0.0386409611,-0.0774591267,-0.0734680966,0.1406773478,0.0299724508,0.1926366687,0.2765003741,0.2475029826,-0.2172683924,-0.3548972011,-0.2435459346,0.1768790036,0.1148345843,-0.1390989274,0.0299772602,-0.3175421357,0.0918787643,-0.0323445313,-0.0665194169,0.2071175724,-0.3488472402,0.1094204113,0.1852676868,-0.1694183499,0.3220192492,-0.1171384901,0.0898978636,-0.0134830419,-0.0217510052,-0.0988093317,-0.1474416703,-0.1254175901,0.1327666491,0.1662156433,0.1449765265,0.4963173568,0.0013359481,-0.2189733684,-0.5480391383,-0.2451718748,0.0971291885,-0.1633996367,0.1954907775,0.2778953016,0.0427568778,-0.0250689685,-0.3409390152,0.3002497554,-0.4462933838,-0.373233676,-0.2133507133,0.0948979482,-0.1722041965,-0.0335693397,-0.3764518201,-0.3236232698,-0.1813369691,0.3523524106,-0.0931495354,0.1716113091,-0.155457288,-0.1054516658,0.1455910206,-0.0203673467,0.1121426821,-0.2187279463,-0.1448161304,-0.0194471348,-0.1510415971,-0.283069551,-0.1560443342,-0.140906021,0.4374369085,-0.10340821,-0.1784181744,-0.0407508872,-0.1612140536,0.4367106259,0.0242293067,-0.0490040779,0.3442159593,0.2696896195,-0.348139286,-0.0873824731,-0.16058743,0.0496429652,-0.2982184887,-0.1103075147,-0.185685128,0.4214314222,0.2078024745,0.5972415209,0.028514469,-0.0613885559,0.1324319392,-0.4377176166,-0.0172750056,-0.0337198712,-0.4259907305,0.1626303345,-0.2667844892,-0.1014861315,0.1350760311,-0.1677293479,-0.115086928,0.1514069438,0.0009905943,-0.3835220933,-0.209966138,0.0377299711,-0.1925776899,0.0860434994,0.0603309199,0.0141091933,-0.4112586379,0.2260724902,-0.0721928179,-0.0218361951,-0.1237445623,-0.3398272991,-0.3464430869,0.0175716821,-0.6085758209,0.3544566035,0.1318598092,0.4851483107,0.1367789209,0.0052999631,0.0414658636,-0.2825492918,0.6524977684,-0.2151976973,0.0749009252,0.183681041,-0.3277136385,-0.2910260558,-0.0436595753,-0.3557530344,0.4460589588,0.3172533214,0.4865555763,-0.1918331534,-0.3518019319,0.1224313006,-0.0199515596,0.0334758908,0.0118676918,-0.28859815,-0.3060879111,-0.4088720381,0.3470336199,0.088713713,0.1813630015,-0.0699506551,-0.1245866269,-0.1805427074,-0.0352943242,0.1069817096,0.1873820275,-0.0616821796,-0.1572833955,0.1408728659,-0.0219112542,0.2116450667,0.2355050445,0.0500611067,-0.0951047987,-0.1150436103,0.0940205231,0.0208330434,0.4114652574,0.2416535616,-0.1685579568,0.3699742556,-0.1364515871,0.288634181,-0.110836409,0.4739766717,0.5543159246,0.1899577975,-0.1046064273,-0.1071350798,0.0384807661,0.1414999962,-0.1566073895,0.3062249124,-0.2532045543,-0.2288146019,0.1476330608,0.1779163629,0.5770257115,-0.0932503566,0.0875760093,0.0692830384,-0.0339656845,0.066425845,-0.322596252,0.1247427389,-0.2881054282,-0.122368142,0.0637505651,-0.0489776097,0.1619583219,0.1876238585,-0.3256686628,0.1354865581,0.2791850567,0.1854793131,0.0244828667,0.4316500723,0.0075259595,-0.2300830781,-0.023363851,0.2279523462,0.1796824038,-0.0562872812,-0.0279301107,0.100480549,-0.0691840649,-0.376296699,-0.193999365,0.0208728425,-0.4312960207,0.3942311704,0.0237804763,0.0594658144,0.0281341858,0.1458708793,0.1447972655,0.2050364465,0.010779175,0.1170280799,0.0932478532,0.0910386071,0.0838550776,-0.2934494615,0.2623491287,-0.1249917671,-0.4529618621,0.0569270551,-0.0704900026,-0.2504338026,0.0244053863,0.1059229895,0.2930746675,-0.2247793376,0.0423178449,-0.0464648195,-0.2383497357,-0.3965626061,0.2039445192,-0.1147942171,-0.198367089,0.2978394926,-0.1070739329,-0.3055641949,-0.0280588754,0.2836199701,0.0299851205,0.152165398,0.5996189117,-0.0773305595,-0.1156411916,-0.4316452146,0.0979604349,0.3189346194,0.1494747549,0.2477394789,-0.1895447522,-0.2431792468,0.0721193999,0.1885505766,0.037892025,0.1453275084,-0.2402581722,-0.2921876609,-0.4588644207,0.0046025463,-0.1703901291,0.2114720941,0.194389537,0.4413281083,-0.1352238059,0.236654982,-0.4279208481,-0.037296284,-0.4685045481,0.2483720928,-0.0478404127,-0.0903501287,0.0450875238,0.0513905063,0.3058929145,0.2208818644,-0.0576614812,-0.3870592415,-0.0921496823,0.0931526944,0.1092805713,0.0143305482,0.1976386756,0.2131322324,-0.2241609246,-0.2637848258,0.1043587327,0.1285088062,-0.0872338116,0.0787764937,0.1816832423,0.0326143913,-0.0256909542,-0.33613047,-0.0561857261,0.0320777707,-0.0667607933,0.3304432631,0.0313246921,-0.0545088127,0.0997803807,0.0593816601,0.2828300595,0.2285246551,0.0306528434,-0.1030482054,-0.3491645157,-0.0028985492,0.3846751153,0.2440446466,-0.2425766736,-0.0939171612,0.2190344483,0.3167302907,-0.231924966,-0.0727749914,0.2596981823,-0.380011797,0.3790821135,0.2261145115,0.0072065424,0.1852706224,0.1105445027,0.0050976626,0.2265835255,-0.3973679543,-0.0119793043,0.0900445506,-0.1410699338,0.1985625029,0.44924137,0.0734671876,0.2462897003,0.341006577,0.014755778,0.5114738345,-0.1246856526,0.0724941492,0.0589090176,-0.0415682569,0.3311864436,-0.0005669777,-0.0347976312,0.0089654345,-0.0880122706,0.3054791391,-0.3336259723,-0.0528675728,-0.2124442756,0.368789047,-0.1914831251,-0.3856400847,-0.23804757,-0.0818293467,-0.1413496882,-0.0785161257,-0.0190696828,0.1973137259,0.0451892503,-0.0820419714,0.0131340185,-0.2212535143,-0.0192149449,-0.0783587024,0.130819723,-0.2626004219,0.2001495957,0.1314703077,0.1404423863,-0.1671113372,0.6095541716,0.5709316134,0.3046991527,-0.1988588274,-0.1260323524,-0.3344715834,-0.1386544555,-0.2002345026,0.286361903,0.0003168109,-0.1747994572,0.2625156939,0.2772244811,-0.2658777535,-0.1564970165,0.4190347791,-0.1771449149,0.1225867122,0.1675085574,0.1353608817,-0.2269658446,0.0732962564,0.1038703024,-0.3389841914,-0.0345561467,0.2996349633,-0.2357952893,0.3266506493,-0.131075263,0.1637846828,0.0463033915,0.4507513344,0.0503231846,0.1739298105,-0.3142849505,-0.1331529915,-0.5782245398,0.2144250721,-0.1688652784,-0.3648037016,-0.1931368709,0.1571838707,0.1029490456,0.1065549478,0.2409303337,0.0778954029,-0.1090169325,0.2187099457,-0.3619211912,0.0869541839,0.1755951494,0.0519847982,0.204318434,-0.4335597754,0.2095237821,-0.0014672085,0.2383017242,-0.0144974636,0.0347771682,0.1814197898,-0.3401248157,0.2746292949,0.1093128175,0.5257284641,-0.4383835793,0.091044113,-0.1269073635,0.1606307179,-0.4419291317,0.0574725047,0.2414098531,0.1892359853,0.0872847959,-0.0311474353,-0.1315664053,0.1188577861,0.2362552434,-0.0922232419,-0.0139229549,0.3029530942,0.1060066298,0.0748861805,-0.1064884812,0.3399718106,-0.0225495938,0.2731781006,-0.4903123081,-0.4749884307,0.3061774373,-0.5121244192,-0.4921657443,-0.3859884441,0.288736403,-0.0939683691,0.0935930386,-0.3449763656,-0.0112365661,0.288376838,0.1060355008,-0.5911830068,0.1808962375,0.088660121,-0.0450419672,-0.0889836028,0.3249361813,0.0077151936,-0.1511936933,-0.0211812947,-0.1020610854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/847","title":"multiprocessing in dataset map \"can only test a child process\"","comments":"Hi @timothyjlaurent @gaceladri \r\nIf you're running `transformers` from `master` you can try setting the env var `WAND_DISABLE=true` (from https:\/\/github.com\/huggingface\/transformers\/pull\/9896) and try again ?\r\nThis issue might be related to https:\/\/github.com\/huggingface\/transformers\/issues\/9623 ","body":"Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```","comment_length":30,"text":"multiprocessing in dataset map \"can only test a child process\"\nUsing a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```\nHi @timothyjlaurent @gaceladri \r\nIf you're running `transformers` from `master` you can try setting the env var `WAND_DISABLE=true` (from https:\/\/github.com\/huggingface\/transformers\/pull\/9896) and try again ?\r\nThis issue might be related to https:\/\/github.com\/huggingface\/transformers\/issues\/9623 ","embeddings":[-0.3956997097,0.0156290922,-0.1538687646,-0.1850299388,0.141797632,-0.0578565337,0.5068988204,0.3400265574,-0.0834040865,0.1681774557,-0.0184273254,0.3370628655,0.0475015864,0.0918332115,-0.2264437228,0.1021468788,-0.0123605253,0.0946920067,0.1290109754,0.0473636836,-0.1753095239,0.1040760726,-0.2302506417,0.2859587967,-0.4516958296,-0.285730809,-0.0163212903,-0.0623164959,-0.2643237412,-0.4696436226,-0.149113372,0.2851142585,-0.1861115098,0.6030860543,-0.0001044559,0.0936724842,0.2627740204,0.0446500145,-0.054140687,-0.030070236,0.095131956,-0.0515859984,-0.0310001038,-0.4035981894,-0.0044792355,-0.2674306333,0.1779136211,-0.4431346357,0.2878316641,0.3189612627,0.2665320337,0.5655280352,-0.2092227638,-0.0047211065,-0.201201424,-0.0783623531,-0.0357206613,-0.1625539958,0.2167165279,-0.2403003126,-0.3567937613,0.2491746992,-0.0430285074,0.1743139625,-0.3290867805,0.0007867881,0.277649343,-0.237424776,0.3172336221,0.079235658,0.1507846713,-0.2674021721,-0.3465930223,-0.1820102185,-0.2675664425,-0.1380905807,-0.0896170512,0.0866119117,-0.3090318441,0.0237945151,-0.231428802,0.0673615932,-0.0069425311,-0.0971344262,-0.20974648,0.6421571374,0.0204102993,0.2258118987,-0.058426097,0.0421580598,0.1321844459,-0.1236515045,0.1550007761,-0.0283196401,-0.1767590791,-0.1339012235,0.1732696593,-0.3608185351,0.1934940219,-0.0518135242,-0.2702677846,0.1569519043,0.1210899875,0.1657909304,0.1612778902,0.0505134016,0.1559104919,0.2796486914,0.2925396562,-0.196995303,-0.3908066452,-0.0121860905,0.0276292767,-0.2749152184,0.19010818,0.1587806493,0.1859900355,0.0203872453,-0.316894412,-0.0674285442,-0.1310207993,-0.0011093016,0.0109488498,0.4655701518,-0.0625771657,0.210119307,-0.1869856268,0.2766140103,-0.1955184191,-0.1335654557,-0.2674530447,0.1599751413,-0.105241254,-0.0912755728,0.0864879191,0.2284782082,0.3958972394,0.057422515,0.0732819885,-0.0852145255,0.4325413704,-0.219634518,0.2467176765,0.0557839759,0.068621248,0.1359149069,0.26744017,-0.3098887801,-0.1797690243,-0.0748893917,0.0085688289,-0.1883713305,-0.0060650436,0.2934336662,0.0234699752,0.0294385776,-0.1553809643,0.1325114965,0.2363231629,0.0417060144,0.0084723802,-0.1836157143,-0.2125665992,-0.1611631811,0.1659645736,0.2332042754,-0.1246359199,-0.0227329582,0.0209111851,-0.1451625675,0.2747372985,0.2273871899,-0.0913355052,0.6480855346,-0.1730043739,0.5878829956,0.1861093491,-0.5310037136,-0.1901540011,0.2092227191,-0.3177124858,-0.0111948214,0.1317908615,-0.1686787903,0.3851614892,0.080286935,0.1898801774,0.2464483976,-0.1104762852,0.2880237997,-0.2860673666,-0.0096036047,0.0849479064,-0.071969904,0.2393062264,-0.2704798877,-0.0424101911,-0.1621235162,0.2372229397,-0.1482038051,0.2230832577,-0.0443854667,0.0625272393,0.0818433464,-0.0392007828,-0.1215575859,0.0788098946,0.2157844901,0.0433690101,0.0418133661,-0.1905940175,-0.1984119862,-0.0963724405,0.2760661244,-0.2941963971,-0.3430737555,0.3505134284,0.3571797907,-0.0198514294,-0.0224457197,-0.0754135773,0.1201187074,0.1530826986,0.0062083742,0.0715006441,-0.0785518736,-0.1991024315,-0.1974657774,-0.0115251811,0.0726843029,0.1600964069,0.003486624,-0.0784358084,0.4022406936,0.2385911644,0.1009370983,0.2098905593,-0.1776235253,-0.0964289233,0.0071378755,-0.0728402957,0.111530818,0.0388300754,-0.0264948942,0.1444305331,0.4932917356,0.2365600765,0.1932385713,0.2239498198,-0.0101361321,0.2388229668,-0.1069113389,-0.0609275028,-0.1341272146,0.057798285,0.0031344036,0.1119551882,-0.1359223723,-0.2041955888,0.1788984239,0.314935714,-0.0856286287,0.1352984309,-0.1058834121,-0.0881043524,-0.1929530203,0.0814027041,0.1926788986,0.3943273127,0.2004811913,0.1856659055,-0.0386409611,-0.0774591267,-0.0734680966,0.1406773478,0.0299724508,0.1926366687,0.2765003741,0.2475029826,-0.2172683924,-0.3548972011,-0.2435459346,0.1768790036,0.1148345843,-0.1390989274,0.0299772602,-0.3175421357,0.0918787643,-0.0323445313,-0.0665194169,0.2071175724,-0.3488472402,0.1094204113,0.1852676868,-0.1694183499,0.3220192492,-0.1171384901,0.0898978636,-0.0134830419,-0.0217510052,-0.0988093317,-0.1474416703,-0.1254175901,0.1327666491,0.1662156433,0.1449765265,0.4963173568,0.0013359481,-0.2189733684,-0.5480391383,-0.2451718748,0.0971291885,-0.1633996367,0.1954907775,0.2778953016,0.0427568778,-0.0250689685,-0.3409390152,0.3002497554,-0.4462933838,-0.373233676,-0.2133507133,0.0948979482,-0.1722041965,-0.0335693397,-0.3764518201,-0.3236232698,-0.1813369691,0.3523524106,-0.0931495354,0.1716113091,-0.155457288,-0.1054516658,0.1455910206,-0.0203673467,0.1121426821,-0.2187279463,-0.1448161304,-0.0194471348,-0.1510415971,-0.283069551,-0.1560443342,-0.140906021,0.4374369085,-0.10340821,-0.1784181744,-0.0407508872,-0.1612140536,0.4367106259,0.0242293067,-0.0490040779,0.3442159593,0.2696896195,-0.348139286,-0.0873824731,-0.16058743,0.0496429652,-0.2982184887,-0.1103075147,-0.185685128,0.4214314222,0.2078024745,0.5972415209,0.028514469,-0.0613885559,0.1324319392,-0.4377176166,-0.0172750056,-0.0337198712,-0.4259907305,0.1626303345,-0.2667844892,-0.1014861315,0.1350760311,-0.1677293479,-0.115086928,0.1514069438,0.0009905943,-0.3835220933,-0.209966138,0.0377299711,-0.1925776899,0.0860434994,0.0603309199,0.0141091933,-0.4112586379,0.2260724902,-0.0721928179,-0.0218361951,-0.1237445623,-0.3398272991,-0.3464430869,0.0175716821,-0.6085758209,0.3544566035,0.1318598092,0.4851483107,0.1367789209,0.0052999631,0.0414658636,-0.2825492918,0.6524977684,-0.2151976973,0.0749009252,0.183681041,-0.3277136385,-0.2910260558,-0.0436595753,-0.3557530344,0.4460589588,0.3172533214,0.4865555763,-0.1918331534,-0.3518019319,0.1224313006,-0.0199515596,0.0334758908,0.0118676918,-0.28859815,-0.3060879111,-0.4088720381,0.3470336199,0.088713713,0.1813630015,-0.0699506551,-0.1245866269,-0.1805427074,-0.0352943242,0.1069817096,0.1873820275,-0.0616821796,-0.1572833955,0.1408728659,-0.0219112542,0.2116450667,0.2355050445,0.0500611067,-0.0951047987,-0.1150436103,0.0940205231,0.0208330434,0.4114652574,0.2416535616,-0.1685579568,0.3699742556,-0.1364515871,0.288634181,-0.110836409,0.4739766717,0.5543159246,0.1899577975,-0.1046064273,-0.1071350798,0.0384807661,0.1414999962,-0.1566073895,0.3062249124,-0.2532045543,-0.2288146019,0.1476330608,0.1779163629,0.5770257115,-0.0932503566,0.0875760093,0.0692830384,-0.0339656845,0.066425845,-0.322596252,0.1247427389,-0.2881054282,-0.122368142,0.0637505651,-0.0489776097,0.1619583219,0.1876238585,-0.3256686628,0.1354865581,0.2791850567,0.1854793131,0.0244828667,0.4316500723,0.0075259595,-0.2300830781,-0.023363851,0.2279523462,0.1796824038,-0.0562872812,-0.0279301107,0.100480549,-0.0691840649,-0.376296699,-0.193999365,0.0208728425,-0.4312960207,0.3942311704,0.0237804763,0.0594658144,0.0281341858,0.1458708793,0.1447972655,0.2050364465,0.010779175,0.1170280799,0.0932478532,0.0910386071,0.0838550776,-0.2934494615,0.2623491287,-0.1249917671,-0.4529618621,0.0569270551,-0.0704900026,-0.2504338026,0.0244053863,0.1059229895,0.2930746675,-0.2247793376,0.0423178449,-0.0464648195,-0.2383497357,-0.3965626061,0.2039445192,-0.1147942171,-0.198367089,0.2978394926,-0.1070739329,-0.3055641949,-0.0280588754,0.2836199701,0.0299851205,0.152165398,0.5996189117,-0.0773305595,-0.1156411916,-0.4316452146,0.0979604349,0.3189346194,0.1494747549,0.2477394789,-0.1895447522,-0.2431792468,0.0721193999,0.1885505766,0.037892025,0.1453275084,-0.2402581722,-0.2921876609,-0.4588644207,0.0046025463,-0.1703901291,0.2114720941,0.194389537,0.4413281083,-0.1352238059,0.236654982,-0.4279208481,-0.037296284,-0.4685045481,0.2483720928,-0.0478404127,-0.0903501287,0.0450875238,0.0513905063,0.3058929145,0.2208818644,-0.0576614812,-0.3870592415,-0.0921496823,0.0931526944,0.1092805713,0.0143305482,0.1976386756,0.2131322324,-0.2241609246,-0.2637848258,0.1043587327,0.1285088062,-0.0872338116,0.0787764937,0.1816832423,0.0326143913,-0.0256909542,-0.33613047,-0.0561857261,0.0320777707,-0.0667607933,0.3304432631,0.0313246921,-0.0545088127,0.0997803807,0.0593816601,0.2828300595,0.2285246551,0.0306528434,-0.1030482054,-0.3491645157,-0.0028985492,0.3846751153,0.2440446466,-0.2425766736,-0.0939171612,0.2190344483,0.3167302907,-0.231924966,-0.0727749914,0.2596981823,-0.380011797,0.3790821135,0.2261145115,0.0072065424,0.1852706224,0.1105445027,0.0050976626,0.2265835255,-0.3973679543,-0.0119793043,0.0900445506,-0.1410699338,0.1985625029,0.44924137,0.0734671876,0.2462897003,0.341006577,0.014755778,0.5114738345,-0.1246856526,0.0724941492,0.0589090176,-0.0415682569,0.3311864436,-0.0005669777,-0.0347976312,0.0089654345,-0.0880122706,0.3054791391,-0.3336259723,-0.0528675728,-0.2124442756,0.368789047,-0.1914831251,-0.3856400847,-0.23804757,-0.0818293467,-0.1413496882,-0.0785161257,-0.0190696828,0.1973137259,0.0451892503,-0.0820419714,0.0131340185,-0.2212535143,-0.0192149449,-0.0783587024,0.130819723,-0.2626004219,0.2001495957,0.1314703077,0.1404423863,-0.1671113372,0.6095541716,0.5709316134,0.3046991527,-0.1988588274,-0.1260323524,-0.3344715834,-0.1386544555,-0.2002345026,0.286361903,0.0003168109,-0.1747994572,0.2625156939,0.2772244811,-0.2658777535,-0.1564970165,0.4190347791,-0.1771449149,0.1225867122,0.1675085574,0.1353608817,-0.2269658446,0.0732962564,0.1038703024,-0.3389841914,-0.0345561467,0.2996349633,-0.2357952893,0.3266506493,-0.131075263,0.1637846828,0.0463033915,0.4507513344,0.0503231846,0.1739298105,-0.3142849505,-0.1331529915,-0.5782245398,0.2144250721,-0.1688652784,-0.3648037016,-0.1931368709,0.1571838707,0.1029490456,0.1065549478,0.2409303337,0.0778954029,-0.1090169325,0.2187099457,-0.3619211912,0.0869541839,0.1755951494,0.0519847982,0.204318434,-0.4335597754,0.2095237821,-0.0014672085,0.2383017242,-0.0144974636,0.0347771682,0.1814197898,-0.3401248157,0.2746292949,0.1093128175,0.5257284641,-0.4383835793,0.091044113,-0.1269073635,0.1606307179,-0.4419291317,0.0574725047,0.2414098531,0.1892359853,0.0872847959,-0.0311474353,-0.1315664053,0.1188577861,0.2362552434,-0.0922232419,-0.0139229549,0.3029530942,0.1060066298,0.0748861805,-0.1064884812,0.3399718106,-0.0225495938,0.2731781006,-0.4903123081,-0.4749884307,0.3061774373,-0.5121244192,-0.4921657443,-0.3859884441,0.288736403,-0.0939683691,0.0935930386,-0.3449763656,-0.0112365661,0.288376838,0.1060355008,-0.5911830068,0.1808962375,0.088660121,-0.0450419672,-0.0889836028,0.3249361813,0.0077151936,-0.1511936933,-0.0211812947,-0.1020610854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/847","title":"multiprocessing in dataset map \"can only test a child process\"","comments":"I have commented the lines that cause my code break. I'm now seeing my reports on Wandb and my code does not break. I am training now, so I will check probably in 6 hours. I suppose that setting wandb disable will work as well.","body":"Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```","comment_length":45,"text":"multiprocessing in dataset map \"can only test a child process\"\nUsing a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.\r\n\r\n``` \r\ndef tokenizer_fn(example):\r\n return tokenizer.batch_encode_plus(example['text'])\r\n\r\nds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])\r\n```\r\n\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/multiprocess\/pool.py\", line 119, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 156, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1510, in _map_single\r\n for i in pbar:\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 228, in __iter__\r\n for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1186, in __iter__\r\n self.close()\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/notebook.py\", line 251, in close\r\n super(tqdm_notebook, self).close(*args, **kwargs)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1291, in close\r\n fp_write('')\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/tqdm\/std.py\", line 1288, in fp_write\r\n self.fp.write(_unicode(s))\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 91, in new_write\r\n cb(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 598, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 146, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 151, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/jovyan\/share\/users\/tlaurent\/invitae-bert\/ve\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 431, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/usr\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n\"\"\"\r\n```\nI have commented the lines that cause my code break. I'm now seeing my reports on Wandb and my code does not break. I am training now, so I will check probably in 6 hours. I suppose that setting wandb disable will work as well.","embeddings":[-0.3956997097,0.0156290922,-0.1538687646,-0.1850299388,0.141797632,-0.0578565337,0.5068988204,0.3400265574,-0.0834040865,0.1681774557,-0.0184273254,0.3370628655,0.0475015864,0.0918332115,-0.2264437228,0.1021468788,-0.0123605253,0.0946920067,0.1290109754,0.0473636836,-0.1753095239,0.1040760726,-0.2302506417,0.2859587967,-0.4516958296,-0.285730809,-0.0163212903,-0.0623164959,-0.2643237412,-0.4696436226,-0.149113372,0.2851142585,-0.1861115098,0.6030860543,-0.0001044559,0.0936724842,0.2627740204,0.0446500145,-0.054140687,-0.030070236,0.095131956,-0.0515859984,-0.0310001038,-0.4035981894,-0.0044792355,-0.2674306333,0.1779136211,-0.4431346357,0.2878316641,0.3189612627,0.2665320337,0.5655280352,-0.2092227638,-0.0047211065,-0.201201424,-0.0783623531,-0.0357206613,-0.1625539958,0.2167165279,-0.2403003126,-0.3567937613,0.2491746992,-0.0430285074,0.1743139625,-0.3290867805,0.0007867881,0.277649343,-0.237424776,0.3172336221,0.079235658,0.1507846713,-0.2674021721,-0.3465930223,-0.1820102185,-0.2675664425,-0.1380905807,-0.0896170512,0.0866119117,-0.3090318441,0.0237945151,-0.231428802,0.0673615932,-0.0069425311,-0.0971344262,-0.20974648,0.6421571374,0.0204102993,0.2258118987,-0.058426097,0.0421580598,0.1321844459,-0.1236515045,0.1550007761,-0.0283196401,-0.1767590791,-0.1339012235,0.1732696593,-0.3608185351,0.1934940219,-0.0518135242,-0.2702677846,0.1569519043,0.1210899875,0.1657909304,0.1612778902,0.0505134016,0.1559104919,0.2796486914,0.2925396562,-0.196995303,-0.3908066452,-0.0121860905,0.0276292767,-0.2749152184,0.19010818,0.1587806493,0.1859900355,0.0203872453,-0.316894412,-0.0674285442,-0.1310207993,-0.0011093016,0.0109488498,0.4655701518,-0.0625771657,0.210119307,-0.1869856268,0.2766140103,-0.1955184191,-0.1335654557,-0.2674530447,0.1599751413,-0.105241254,-0.0912755728,0.0864879191,0.2284782082,0.3958972394,0.057422515,0.0732819885,-0.0852145255,0.4325413704,-0.219634518,0.2467176765,0.0557839759,0.068621248,0.1359149069,0.26744017,-0.3098887801,-0.1797690243,-0.0748893917,0.0085688289,-0.1883713305,-0.0060650436,0.2934336662,0.0234699752,0.0294385776,-0.1553809643,0.1325114965,0.2363231629,0.0417060144,0.0084723802,-0.1836157143,-0.2125665992,-0.1611631811,0.1659645736,0.2332042754,-0.1246359199,-0.0227329582,0.0209111851,-0.1451625675,0.2747372985,0.2273871899,-0.0913355052,0.6480855346,-0.1730043739,0.5878829956,0.1861093491,-0.5310037136,-0.1901540011,0.2092227191,-0.3177124858,-0.0111948214,0.1317908615,-0.1686787903,0.3851614892,0.080286935,0.1898801774,0.2464483976,-0.1104762852,0.2880237997,-0.2860673666,-0.0096036047,0.0849479064,-0.071969904,0.2393062264,-0.2704798877,-0.0424101911,-0.1621235162,0.2372229397,-0.1482038051,0.2230832577,-0.0443854667,0.0625272393,0.0818433464,-0.0392007828,-0.1215575859,0.0788098946,0.2157844901,0.0433690101,0.0418133661,-0.1905940175,-0.1984119862,-0.0963724405,0.2760661244,-0.2941963971,-0.3430737555,0.3505134284,0.3571797907,-0.0198514294,-0.0224457197,-0.0754135773,0.1201187074,0.1530826986,0.0062083742,0.0715006441,-0.0785518736,-0.1991024315,-0.1974657774,-0.0115251811,0.0726843029,0.1600964069,0.003486624,-0.0784358084,0.4022406936,0.2385911644,0.1009370983,0.2098905593,-0.1776235253,-0.0964289233,0.0071378755,-0.0728402957,0.111530818,0.0388300754,-0.0264948942,0.1444305331,0.4932917356,0.2365600765,0.1932385713,0.2239498198,-0.0101361321,0.2388229668,-0.1069113389,-0.0609275028,-0.1341272146,0.057798285,0.0031344036,0.1119551882,-0.1359223723,-0.2041955888,0.1788984239,0.314935714,-0.0856286287,0.1352984309,-0.1058834121,-0.0881043524,-0.1929530203,0.0814027041,0.1926788986,0.3943273127,0.2004811913,0.1856659055,-0.0386409611,-0.0774591267,-0.0734680966,0.1406773478,0.0299724508,0.1926366687,0.2765003741,0.2475029826,-0.2172683924,-0.3548972011,-0.2435459346,0.1768790036,0.1148345843,-0.1390989274,0.0299772602,-0.3175421357,0.0918787643,-0.0323445313,-0.0665194169,0.2071175724,-0.3488472402,0.1094204113,0.1852676868,-0.1694183499,0.3220192492,-0.1171384901,0.0898978636,-0.0134830419,-0.0217510052,-0.0988093317,-0.1474416703,-0.1254175901,0.1327666491,0.1662156433,0.1449765265,0.4963173568,0.0013359481,-0.2189733684,-0.5480391383,-0.2451718748,0.0971291885,-0.1633996367,0.1954907775,0.2778953016,0.0427568778,-0.0250689685,-0.3409390152,0.3002497554,-0.4462933838,-0.373233676,-0.2133507133,0.0948979482,-0.1722041965,-0.0335693397,-0.3764518201,-0.3236232698,-0.1813369691,0.3523524106,-0.0931495354,0.1716113091,-0.155457288,-0.1054516658,0.1455910206,-0.0203673467,0.1121426821,-0.2187279463,-0.1448161304,-0.0194471348,-0.1510415971,-0.283069551,-0.1560443342,-0.140906021,0.4374369085,-0.10340821,-0.1784181744,-0.0407508872,-0.1612140536,0.4367106259,0.0242293067,-0.0490040779,0.3442159593,0.2696896195,-0.348139286,-0.0873824731,-0.16058743,0.0496429652,-0.2982184887,-0.1103075147,-0.185685128,0.4214314222,0.2078024745,0.5972415209,0.028514469,-0.0613885559,0.1324319392,-0.4377176166,-0.0172750056,-0.0337198712,-0.4259907305,0.1626303345,-0.2667844892,-0.1014861315,0.1350760311,-0.1677293479,-0.115086928,0.1514069438,0.0009905943,-0.3835220933,-0.209966138,0.0377299711,-0.1925776899,0.0860434994,0.0603309199,0.0141091933,-0.4112586379,0.2260724902,-0.0721928179,-0.0218361951,-0.1237445623,-0.3398272991,-0.3464430869,0.0175716821,-0.6085758209,0.3544566035,0.1318598092,0.4851483107,0.1367789209,0.0052999631,0.0414658636,-0.2825492918,0.6524977684,-0.2151976973,0.0749009252,0.183681041,-0.3277136385,-0.2910260558,-0.0436595753,-0.3557530344,0.4460589588,0.3172533214,0.4865555763,-0.1918331534,-0.3518019319,0.1224313006,-0.0199515596,0.0334758908,0.0118676918,-0.28859815,-0.3060879111,-0.4088720381,0.3470336199,0.088713713,0.1813630015,-0.0699506551,-0.1245866269,-0.1805427074,-0.0352943242,0.1069817096,0.1873820275,-0.0616821796,-0.1572833955,0.1408728659,-0.0219112542,0.2116450667,0.2355050445,0.0500611067,-0.0951047987,-0.1150436103,0.0940205231,0.0208330434,0.4114652574,0.2416535616,-0.1685579568,0.3699742556,-0.1364515871,0.288634181,-0.110836409,0.4739766717,0.5543159246,0.1899577975,-0.1046064273,-0.1071350798,0.0384807661,0.1414999962,-0.1566073895,0.3062249124,-0.2532045543,-0.2288146019,0.1476330608,0.1779163629,0.5770257115,-0.0932503566,0.0875760093,0.0692830384,-0.0339656845,0.066425845,-0.322596252,0.1247427389,-0.2881054282,-0.122368142,0.0637505651,-0.0489776097,0.1619583219,0.1876238585,-0.3256686628,0.1354865581,0.2791850567,0.1854793131,0.0244828667,0.4316500723,0.0075259595,-0.2300830781,-0.023363851,0.2279523462,0.1796824038,-0.0562872812,-0.0279301107,0.100480549,-0.0691840649,-0.376296699,-0.193999365,0.0208728425,-0.4312960207,0.3942311704,0.0237804763,0.0594658144,0.0281341858,0.1458708793,0.1447972655,0.2050364465,0.010779175,0.1170280799,0.0932478532,0.0910386071,0.0838550776,-0.2934494615,0.2623491287,-0.1249917671,-0.4529618621,0.0569270551,-0.0704900026,-0.2504338026,0.0244053863,0.1059229895,0.2930746675,-0.2247793376,0.0423178449,-0.0464648195,-0.2383497357,-0.3965626061,0.2039445192,-0.1147942171,-0.198367089,0.2978394926,-0.1070739329,-0.3055641949,-0.0280588754,0.2836199701,0.0299851205,0.152165398,0.5996189117,-0.0773305595,-0.1156411916,-0.4316452146,0.0979604349,0.3189346194,0.1494747549,0.2477394789,-0.1895447522,-0.2431792468,0.0721193999,0.1885505766,0.037892025,0.1453275084,-0.2402581722,-0.2921876609,-0.4588644207,0.0046025463,-0.1703901291,0.2114720941,0.194389537,0.4413281083,-0.1352238059,0.236654982,-0.4279208481,-0.037296284,-0.4685045481,0.2483720928,-0.0478404127,-0.0903501287,0.0450875238,0.0513905063,0.3058929145,0.2208818644,-0.0576614812,-0.3870592415,-0.0921496823,0.0931526944,0.1092805713,0.0143305482,0.1976386756,0.2131322324,-0.2241609246,-0.2637848258,0.1043587327,0.1285088062,-0.0872338116,0.0787764937,0.1816832423,0.0326143913,-0.0256909542,-0.33613047,-0.0561857261,0.0320777707,-0.0667607933,0.3304432631,0.0313246921,-0.0545088127,0.0997803807,0.0593816601,0.2828300595,0.2285246551,0.0306528434,-0.1030482054,-0.3491645157,-0.0028985492,0.3846751153,0.2440446466,-0.2425766736,-0.0939171612,0.2190344483,0.3167302907,-0.231924966,-0.0727749914,0.2596981823,-0.380011797,0.3790821135,0.2261145115,0.0072065424,0.1852706224,0.1105445027,0.0050976626,0.2265835255,-0.3973679543,-0.0119793043,0.0900445506,-0.1410699338,0.1985625029,0.44924137,0.0734671876,0.2462897003,0.341006577,0.014755778,0.5114738345,-0.1246856526,0.0724941492,0.0589090176,-0.0415682569,0.3311864436,-0.0005669777,-0.0347976312,0.0089654345,-0.0880122706,0.3054791391,-0.3336259723,-0.0528675728,-0.2124442756,0.368789047,-0.1914831251,-0.3856400847,-0.23804757,-0.0818293467,-0.1413496882,-0.0785161257,-0.0190696828,0.1973137259,0.0451892503,-0.0820419714,0.0131340185,-0.2212535143,-0.0192149449,-0.0783587024,0.130819723,-0.2626004219,0.2001495957,0.1314703077,0.1404423863,-0.1671113372,0.6095541716,0.5709316134,0.3046991527,-0.1988588274,-0.1260323524,-0.3344715834,-0.1386544555,-0.2002345026,0.286361903,0.0003168109,-0.1747994572,0.2625156939,0.2772244811,-0.2658777535,-0.1564970165,0.4190347791,-0.1771449149,0.1225867122,0.1675085574,0.1353608817,-0.2269658446,0.0732962564,0.1038703024,-0.3389841914,-0.0345561467,0.2996349633,-0.2357952893,0.3266506493,-0.131075263,0.1637846828,0.0463033915,0.4507513344,0.0503231846,0.1739298105,-0.3142849505,-0.1331529915,-0.5782245398,0.2144250721,-0.1688652784,-0.3648037016,-0.1931368709,0.1571838707,0.1029490456,0.1065549478,0.2409303337,0.0778954029,-0.1090169325,0.2187099457,-0.3619211912,0.0869541839,0.1755951494,0.0519847982,0.204318434,-0.4335597754,0.2095237821,-0.0014672085,0.2383017242,-0.0144974636,0.0347771682,0.1814197898,-0.3401248157,0.2746292949,0.1093128175,0.5257284641,-0.4383835793,0.091044113,-0.1269073635,0.1606307179,-0.4419291317,0.0574725047,0.2414098531,0.1892359853,0.0872847959,-0.0311474353,-0.1315664053,0.1188577861,0.2362552434,-0.0922232419,-0.0139229549,0.3029530942,0.1060066298,0.0748861805,-0.1064884812,0.3399718106,-0.0225495938,0.2731781006,-0.4903123081,-0.4749884307,0.3061774373,-0.5121244192,-0.4921657443,-0.3859884441,0.288736403,-0.0939683691,0.0935930386,-0.3449763656,-0.0112365661,0.288376838,0.1060355008,-0.5911830068,0.1808962375,0.088660121,-0.0450419672,-0.0889836028,0.3249361813,0.0077151936,-0.1511936933,-0.0211812947,-0.1020610854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/846","title":"Add HoVer multi-hop fact verification dataset","comments":"Hi @yjernite I'm new but wanted to contribute. Has anyone already taken this problem and do you think it is suitable for newbies?","body":"## Adding a Dataset\r\n- **Name:** HoVer\r\n- **Description:** https:\/\/twitter.com\/YichenJiang9\/status\/1326954363806429186 contains 20K claim verification examples\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2011.03088\r\n- **Data:** https:\/\/hover-nlp.github.io\/\r\n- **Motivation:** There are still few multi-hop information extraction benchmarks (HotpotQA, which dataset wase based off, notwithstanding)\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":23,"text":"Add HoVer multi-hop fact verification dataset\n## Adding a Dataset\r\n- **Name:** HoVer\r\n- **Description:** https:\/\/twitter.com\/YichenJiang9\/status\/1326954363806429186 contains 20K claim verification examples\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2011.03088\r\n- **Data:** https:\/\/hover-nlp.github.io\/\r\n- **Motivation:** There are still few multi-hop information extraction benchmarks (HotpotQA, which dataset wase based off, notwithstanding)\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nHi @yjernite I'm new but wanted to contribute. Has anyone already taken this problem and do you think it is suitable for newbies?","embeddings":[-0.2272989005,-0.0823319703,0.0003385669,0.1499359757,-0.4059508145,-0.034600284,0.1919758469,0.1057019457,0.0961782411,-0.1447629035,-0.0471815988,-0.1035467982,-0.2635313869,0.2344664782,0.2132848501,-0.0595738515,-0.1592912972,-0.194790706,0.1115603521,0.009324884,-0.0119550144,0.0130747324,-0.2015615404,0.0045057177,-0.5718672276,0.0330752693,-0.2534197867,-0.1548031569,0.085829176,-0.3601928651,0.2060377002,0.6282486916,0.1120479181,0.5196282864,-0.0001165368,-0.3225642741,0.1629251093,0.0532238297,-0.2220108807,-0.0875658318,-0.23451294,0.0085838782,0.0981310904,-0.1646174341,0.0027158612,0.0003690502,-0.3450667858,0.0169514809,0.2611086667,0.1450025737,0.1791539639,0.2080793083,-0.1005195454,0.0022978534,0.2288694829,-0.0532467775,0.0240708869,0.5337884426,0.1278332323,0.0082168104,-0.3305610418,0.2556033432,-0.0482346192,0.0952407345,0.0296689831,-0.2165821046,0.2072823197,-0.0248800274,-0.1847686172,0.6821916103,0.1113199666,-0.1358933747,-0.3626839221,-0.1780867875,0.1415117979,0.1910554618,0.0108547676,0.0837788507,-0.3718945682,0.2523231208,-0.2055734992,-0.4319069088,-0.0509221964,0.0619594045,0.3457696438,0.1788576692,-0.0137493173,0.1511368901,0.2833220661,0.1685571223,0.1721879393,-0.3932555318,0.1634166688,-0.0612804182,-0.2462203205,-0.3360615373,0.3016603291,0.4368488193,0.3293212354,0.408415705,0.1978220791,0.1490226984,-0.3179811835,0.0085508069,0.1067508161,-0.1359966099,-0.365855515,-0.4572265744,0.3480314314,0.0410776399,0.1714973897,0.0878176093,0.089294225,0.2867174149,-0.3137625158,0.1096377447,0.1290668398,-0.2129604816,-0.1475043297,0.1170882434,-0.4780736566,-0.4129661322,0.1277894229,-0.1038535833,-0.0501338355,0.3127092421,0.2355261147,0.1694095731,0.1646916717,-0.3821560442,-0.1032017544,0.3580859303,-0.0851727203,0.1719325185,0.2503606379,0.1378546506,-0.1598970145,-0.0078856191,0.2475170344,0.1192437336,0.1588481665,-0.2928923666,-0.0950281546,-0.2073489875,-0.2217977345,-0.0861623213,0.040520221,-0.1773965061,-0.0580103956,0.1205490008,0.1289972961,-0.2766111791,-0.0955953449,0.113270469,-0.5047513247,-0.076026611,-0.4323448241,0.5820619464,-0.0895041451,0.2344906628,-0.0069962996,0.1698775291,0.0283586793,-0.0424429923,0.0730624348,0.4558426142,0.0693438277,-0.0632667169,-0.1861064583,-0.0203272663,-0.1498075277,-0.1419146657,-0.0407370627,0.3462034762,-0.186228931,0.074893102,0.1125301272,-0.4129200876,-0.0766671523,0.2678289115,-0.3645820618,0.0779021084,0.3461769223,0.1696006507,0.0150969187,-0.3515820205,-0.2433547378,0.5789194703,-0.0191441067,0.0034195944,-0.457888335,-0.0963017642,0.0129377283,0.2265038788,0.0787469819,-0.3378924131,-0.0535279885,0.3043085337,0.3111548126,-0.0393290445,0.2238323092,-0.0580610447,0.2108476311,-0.6122569442,-0.0277998205,-0.1950296462,-0.1623071879,-0.0837937295,0.2959181964,0.4300269783,0.3091977835,-0.1348819733,-0.1295538247,-0.0997592211,-0.2761282623,-0.4181807339,0.1464835405,0.3087125719,0.1575443745,0.0618326738,-0.0519811809,-0.0381782576,0.0885732621,0.0160052553,0.1007430702,0.4683502913,-0.0116548184,0.0949854255,0.008958105,0.3744881749,-0.1441869289,-0.2735956013,0.0491674617,0.4318183064,-0.2593183219,-0.0692741424,0.7123370767,0.2589343488,0.3165448904,-0.4310328662,0.2104399949,-0.0167118199,0.1257604659,-0.1446498036,-0.2468677312,0.345398128,0.1053264886,0.1579203159,-0.0618477389,0.1428497583,0.0796139091,-0.14160873,-0.288089484,-0.0173713434,0.4400113523,0.0126462849,0.2279019505,-0.2999400496,-0.0706615523,0.2726605535,-0.0812865794,0.0399397314,0.2333283722,0.1611472666,-0.0433438905,-0.2864331603,-0.0294447709,0.1646221578,0.3358062208,0.1740187705,0.2849569321,-0.0659761131,0.0629301071,-0.0905710459,0.209753409,-0.1903973669,-0.2169682235,0.0436335318,0.062459413,0.0254511628,-0.1361393481,-0.4984517992,0.2366655469,0.3610859811,-0.1582579017,-0.1105240285,-0.1688098758,-0.307507813,-0.1922272742,-0.341814816,-0.2188437879,-0.4317831695,0.0095069604,0.0510069206,0.2749845088,0.361848712,0.2475446165,0.4854172468,-0.0482300669,0.4010957479,-0.1088409349,-0.1563973129,-0.0341116935,0.1446636766,0.2256753743,0.141429171,0.4724613428,0.0736723244,0.1783833057,-0.2351920009,-0.3586514294,0.2105641514,-0.0159761328,0.2286697179,0.2341108918,0.0499635339,0.2000391185,0.2142780572,0.1577644348,0.0015248955,-0.1417124718,0.1305211782,-0.1321739703,0.0097809955,-0.1054687798,0.1849376857,-0.2513540387,-0.1153514013,0.3771843016,0.0288773943,-0.0870286673,0.0213880111,-0.114605166,0.324867785,-0.1924210936,0.0294510014,-0.287555486,-0.5863478184,-0.0235058051,-0.081751585,-0.0255939867,-0.3916584849,-0.3142075539,0.1521116644,-0.1375472695,-0.5253841877,-0.3399840891,0.0022224619,0.2595570683,0.1521728635,0.0431391187,0.2412624806,-0.1067442372,-0.1025691703,0.0414261408,0.0672295243,0.0882845446,-0.1683748215,0.3496380746,0.2270532101,-0.1118773222,0.5127782822,0.4072456658,0.106169574,0.063926585,0.2201278061,-0.0686427876,0.0777291283,0.1895662546,-0.5597390532,0.4821096361,-0.031890139,-0.0508244745,-0.0369958654,-0.1315049529,0.2448635995,-0.2971071303,0.0702997372,-0.067023851,-0.2937935293,0.1264007241,-0.3778006434,0.1239682138,0.1089379936,-0.3209062219,-0.1242931932,-0.2189759016,0.1418031603,0.2942288518,0.2277389318,-0.2954146564,-0.066876255,-0.095034264,-0.3934743404,0.3894406557,0.0661580339,-0.0900334045,0.1126883999,-0.0044272318,-0.0039529144,0.0922147483,0.4180674553,-0.5143921375,0.1353335828,0.1167920083,0.0336884558,-0.5341309905,0.0339397453,-0.4302006066,-0.0627645254,0.1773236096,0.0435868166,-0.455832094,-0.1332000345,0.7585170269,0.044862736,0.0005283657,-0.3360905349,0.1109617949,0.0661185235,-0.2460403144,0.0561073609,-0.0099069783,0.0208702683,-0.0661032721,0.0346743576,0.2253526449,0.1161696389,0.3921644092,0.0235507917,-0.1376076788,0.2696407735,0.0850262195,-0.2122835815,-0.3036352098,0.1736400425,0.1538458765,-0.0102861812,-0.0846934542,0.0796292648,-0.0378926396,0.2571399212,0.3785805106,0.0597668849,0.3685817719,0.0722899809,0.2682338357,-0.217176646,0.1767383665,-0.1100782976,0.0010996115,0.0037900966,-0.1943553537,0.4577750564,0.0886094272,-0.1785667241,-0.1837354302,-0.1082381904,-0.0236830842,-0.0018278053,0.0532352664,1.2627148628,0.245486632,-0.0133655863,0.2612295747,0.188864097,0.4842988253,-0.4983178675,-0.2737855017,-0.1867612749,-0.0294728,-0.1651691347,-0.3271545768,-0.0046534557,0.1926646382,-0.3592123091,-0.1146055609,0.1274693012,0.0210137926,0.044356823,0.5635526776,-0.0875811428,-0.2835756838,0.2453342229,0.16947712,-0.2909670174,0.0991518795,-0.1385642588,-0.1920528859,-0.1444829106,-0.2057260871,-0.4938208461,-0.130606249,-0.1088442653,0.114552319,-0.2344889492,-0.3886846304,-0.0501266345,-0.5567728281,0.1295872033,0.0299222637,-0.2018659711,-0.1130383015,-0.1547299474,-0.0867039859,0.0413505062,-0.2835637331,0.2741181552,0.0151725356,-0.4493738413,0.1459130198,0.0339933634,-0.1373792142,0.4455838799,-0.0426314846,0.3611053228,0.0761184469,0.2114387155,0.4344063401,-0.2108479738,-0.1954820901,0.1394939125,0.3742080629,-0.30207479,0.149873659,0.0783270076,0.047498472,-0.1404799223,0.3055311739,0.2763256133,-0.1267758012,0.1218034774,0.1743708998,-0.1480489075,-0.268643558,0.4009828568,0.0840239674,-0.4870324731,0.0370533429,-0.1743946373,-0.1999868751,-0.2373298854,0.2377757579,0.4147127569,0.4652632475,0.0448788553,-0.5292447209,0.0094288997,0.2511830032,-0.3439130187,-0.0446378589,-0.5093989968,0.6212291121,0.2525544167,0.2308319658,-0.3422257006,0.2347232848,-0.1392382979,-0.110717006,0.4395376444,-0.2561744153,0.1268961877,0.0996727869,0.0984075889,0.1235120147,-0.1038625091,-0.1879123002,-0.2445531338,0.1251425296,0.0526084304,-0.0673981532,0.1634922028,0.0472759344,-0.0400858484,-0.0242090803,-0.098493062,-0.1040563583,0.0353754275,0.0872734562,-0.0539797284,-0.0468093231,-0.2468536496,-0.0526104234,0.1595726162,0.185394153,0.008584829,0.0589541234,-0.0294489935,-0.0878489763,-0.06073137,0.2267066687,0.025972493,0.0632998869,-0.0865543112,-0.0082771182,-0.0700101629,0.4451744258,0.4849706292,0.2325761169,-0.1189172566,0.1047340035,-0.1195660755,0.1804613769,-0.3028199077,0.2973518968,0.1752165705,0.1731589586,0.1861558408,0.478384316,-0.0383911543,0.0364910252,0.1809012294,-0.0525320955,0.2358426303,-0.2555924058,-0.1147483662,-0.1154726893,-0.2953887284,0.1107535213,0.2694611847,0.1786248386,0.1760028154,-0.1221665069,0.0060345167,0.0549355745,0.0906938165,0.089446336,0.2672556341,-0.2129654735,0.4202230573,0.2588844895,0.0499809384,-0.099419266,0.0852739662,0.6111462712,-0.234055236,-0.033997383,-0.3608177304,0.224563852,0.0237089097,0.118540816,0.0140386913,-0.2033748627,-0.3663735092,-0.3091671765,-0.074329026,-0.1315102428,0.4406821728,0.1155263409,0.2183003277,-0.2886620462,-0.3951531053,0.0844946504,0.21691221,-0.3521037996,0.0441318341,0.4230369329,0.0523838811,0.0111109037,0.6570655704,0.1866727769,-0.1337443441,-0.3192817867,0.1201313138,-0.3357859552,0.0300090555,-0.0276111402,0.172644943,0.5718494058,-0.0983301252,0.197761938,0.0498352833,-0.2692709863,0.4436641037,0.2408137769,0.0371607356,-0.5864850283,0.0541561358,0.0381033234,0.1615028679,-0.1707799882,-0.0078864899,-0.3688827455,-0.1426107138,0.0498989597,-0.2634572387,-0.0825020969,-0.1683405638,0.0525500514,-0.1341020614,0.2285887748,0.4410325885,-0.009619792,-0.3868284523,-0.2921031713,-0.6059029102,0.0989548489,0.0267807301,-0.2023603916,-0.4342101514,0.266454041,0.1798341274,0.7058999538,0.1808881164,-0.2668347061,0.142433852,0.1624396443,-0.3843021691,0.1687568277,0.0906043276,-0.1243207827,0.1892859489,0.1048799604,-0.0175444689,0.333563,0.0561727136,-0.0172391552,-0.5416662693,0.3225411177,0.2993495166,0.3092262745,-0.0227931309,0.4958721399,0.0934569985,-0.1722733974,-0.0667198226,0.0959021077,0.1008073613,0.574200809,-0.0160351507,0.1934215575,-0.0118904216,-0.0537695959,-0.0987725034,-0.2474557012,-0.0965986103,0.0109381918,-0.4421472549,0.0059922677,-0.0498701744,0.3267240524,0.0018914273,0.1454037875,-0.0550496355,0.3461010456,-0.0550591834,-0.198527351,0.3473351896,0.157650575,-0.004852545,-0.2775719464,0.381052047,-0.8294070363,-0.1915791184,-0.5885876417,-0.3144218624,0.3645707965,0.0971504971,-0.0741953999,0.1580603421,-0.2104311287,0.0272302702,0.0301884357,0.3414421976,-0.1916599423,-0.1499206722,-0.070100978,-0.0429637171]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/846","title":"Add HoVer multi-hop fact verification dataset","comments":"Hi @tenjjin! This dataset is still up for grabs! Here's the link with the guide to add it. You should play around with the library first (download and look at a few datasets), then follow the steps here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md","body":"## Adding a Dataset\r\n- **Name:** HoVer\r\n- **Description:** https:\/\/twitter.com\/YichenJiang9\/status\/1326954363806429186 contains 20K claim verification examples\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2011.03088\r\n- **Data:** https:\/\/hover-nlp.github.io\/\r\n- **Motivation:** There are still few multi-hop information extraction benchmarks (HotpotQA, which dataset wase based off, notwithstanding)\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":39,"text":"Add HoVer multi-hop fact verification dataset\n## Adding a Dataset\r\n- **Name:** HoVer\r\n- **Description:** https:\/\/twitter.com\/YichenJiang9\/status\/1326954363806429186 contains 20K claim verification examples\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2011.03088\r\n- **Data:** https:\/\/hover-nlp.github.io\/\r\n- **Motivation:** There are still few multi-hop information extraction benchmarks (HotpotQA, which dataset wase based off, notwithstanding)\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nHi @tenjjin! This dataset is still up for grabs! Here's the link with the guide to add it. You should play around with the library first (download and look at a few datasets), then follow the steps here:\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/ADD_NEW_DATASET.md","embeddings":[-0.2867436111,-0.2775808871,-0.0820240155,0.0229686629,-0.2790857255,-0.1241937578,0.1645177305,0.114708744,0.0981632844,0.1487334222,-0.0963040292,0.0598273389,-0.0950224623,0.2981519401,0.2122938484,-0.1031539589,-0.1313484758,-0.0484599657,-0.0788757727,0.0388446972,0.033641424,0.1323769987,-0.1098015383,-0.0514863804,-0.3951320946,-0.011697568,-0.1528687775,-0.1421919912,0.1767664999,-0.347092092,0.0790029094,0.4557312131,-0.0380147658,0.352250129,-0.0001060055,-0.1624501348,0.1336018741,0.0436690412,-0.2459265739,0.0830911249,-0.080947347,0.0728994682,0.1475116313,-0.097418502,-0.1465436071,-0.0416963473,-0.2358896434,0.137357682,0.3529121578,0.151205942,0.2692036629,0.3744874299,-0.0909703895,-0.1768823266,0.1466149986,-0.1151665598,0.0339635424,0.4747840166,0.0599792004,-0.0117558846,-0.4476830065,0.4065052867,0.0009208469,0.1696484983,-0.1179278716,-0.1064522639,0.1529274732,-0.0746432841,-0.0437307246,0.6290504932,-0.1349589825,-0.146775946,-0.3264868855,-0.2491106391,0.0428118929,0.1343468279,0.0443006642,0.1490805745,-0.2735669315,0.2022499293,-0.2149324715,-0.0940118581,-0.0298706461,0.052622743,0.2132790685,0.1690588295,-0.1091086045,-0.0234358348,0.3585716486,0.1595171243,-0.035753388,-0.2431963384,0.1687718183,-0.1860033572,-0.0606004298,-0.2975720167,0.2898409963,0.3649698496,0.3671050668,0.3012627363,0.265204519,0.1582503766,-0.2483098954,0.0941536129,-0.0687469319,-0.0953088403,-0.2874856293,-0.5833560824,0.3475053012,0.2314156741,0.1296987236,0.0780098587,0.1783207357,0.2515965402,-0.3900646865,0.0836937949,-0.1611859053,-0.1440847665,-0.1861333102,0.106196709,-0.3253612816,-0.1250450313,0.0548867285,0.0711383373,-0.100818716,0.3313430846,0.1849272549,0.0917672366,0.0834687352,-0.3077295721,-0.2107056379,0.3382160366,-0.0327012278,0.1933873296,0.2491162419,0.0862507001,0.1802133322,0.1143538803,0.221690774,0.0304584447,0.1632221341,-0.2651650608,0.0008842012,-0.0709963441,-0.0951058418,-0.2264544666,0.0763855502,-0.1893618554,-0.1272076815,-0.1004316881,0.10269209,-0.1465356201,-0.1066593155,0.2135045826,-0.3886550665,-0.1913703233,-0.2606023848,0.4759401679,-0.1730584055,0.2680049837,0.0965695009,0.2344008088,0.1171851829,0.0083590727,0.1605289727,0.368747443,0.1212783903,-0.2506896257,0.0331295729,-0.0858479887,-0.1767145693,-0.0707864091,-0.0486028716,0.2160945088,-0.2160632759,0.1856982559,-0.0379081219,-0.3163784444,-0.1167093366,0.0720605105,-0.3015666902,0.0878662542,0.2680449188,0.0802268386,0.0543415174,-0.1052969322,-0.166125074,0.5198565125,0.096768342,0.1907670647,-0.3441918194,-0.1677937806,-0.2322195321,0.1796374023,-0.0135582099,-0.320284158,0.011876368,0.2393240035,0.3109693825,-0.0990885794,0.1330364794,0.0712411627,0.1044498459,-0.4205801487,-0.1068930253,-0.2172247916,-0.0872856006,-0.1155741513,0.0521446429,0.4067137837,0.1621482968,-0.2223690748,-0.2138919085,-0.0603819042,-0.1263387203,-0.418602109,0.3150726259,0.4057452381,0.1784080714,0.1347752362,0.0722763464,-0.0547275767,0.0883349776,0.1120368987,0.1067325398,0.286919117,-0.0464472398,-0.0571552925,0.2328546941,0.2236227095,-0.1332375705,-0.2088885605,0.1052233577,0.4635672569,-0.2398419827,0.1055540591,0.550093472,0.3264041841,0.172630921,-0.3131439388,0.1095872596,0.0789438784,0.0664948821,0.0865849629,-0.3467356861,0.4270170927,0.2828143835,0.0556462407,0.0929596871,0.1265733242,0.2059932351,-0.1193459108,-0.2036150545,-0.17944628,0.363371402,-0.025536593,0.2383512706,-0.2011653781,-0.1728343815,0.269005388,0.0963138416,0.1139984503,0.3939310908,0.1579205692,-0.3027701676,-0.3743290305,0.0889573619,0.0322211869,0.0094886217,0.2932662666,0.3736000955,-0.1231614351,-0.063248381,-0.234593749,0.2791765928,-0.2151177377,-0.2966951132,0.1072228476,0.175737381,-0.0794863477,-0.4073752761,-0.411487788,-0.0374751128,0.0819425881,-0.0628039539,-0.0727675557,-0.1820836216,-0.3367846906,-0.1073109284,-0.205092743,-0.2028737217,-0.3738683164,0.1655452102,0.184712708,0.079681173,0.3150455654,0.1682979465,0.5256705284,0.048397731,0.4300487041,-0.1132222041,-0.1184878275,-0.0463088639,0.2272655666,-0.0351149738,0.1383318752,0.4301140606,0.0323062539,0.1055170074,-0.134631902,-0.3381764889,0.3274391294,0.0189793892,0.0784658045,0.249945581,0.0561926737,0.235697642,0.1594017595,0.1428957134,-0.1569969803,-0.1630942971,0.1478888392,-0.1265541911,0.0657547563,-0.2258962095,0.172743544,-0.1313379556,-0.1951169372,0.4276386499,0.0411098972,-0.0647484288,0.1986626685,-0.0287462622,0.3513500988,-0.3559431434,0.2393335849,-0.2662965059,-0.5580601692,0.0462880805,-0.2378893197,-0.0749971643,-0.4700532854,-0.3998803794,0.1343472302,-0.1288474947,-0.4724076688,-0.4922143221,-0.2050073892,0.233960703,0.1317834705,0.0074967425,0.3199823499,-0.1177764162,-0.2865698934,-0.0135590257,0.0779701322,0.1763718575,-0.2182677537,0.1791594476,0.0688348636,-0.1795098037,0.5941285491,0.3341059387,0.0510026067,-0.0625781491,0.22476013,-0.1615021378,0.0383760408,0.1382564753,-0.2809799612,0.2556060851,-0.0379055291,0.1196631491,0.1581821889,-0.1470266879,0.2490223944,-0.2720151246,-0.0091464855,-0.1886363626,-0.3408610821,0.0859107599,-0.1938691735,0.1196765155,-0.0584904589,-0.2399746776,-0.0151589438,-0.1219424233,0.2562094331,0.1171791628,0.167104885,-0.2866930366,-0.0003817461,-0.0684205964,-0.3204259276,0.3787927926,0.1180958822,-0.1601747274,0.1036235765,-0.1899847239,-0.069833152,0.0220604632,0.2806836367,-0.3590142429,0.1685786247,-0.0736776739,-0.1262996495,-0.4930105805,0.1724653244,-0.6214956641,-0.2309940457,0.2284553647,0.0947990343,-0.5830438733,-0.1353436261,0.6615108848,-0.0321313031,0.0462495051,-0.1048676819,0.0395186245,0.0320983902,-0.2198536843,0.1494941562,-0.0236578099,-0.0890543982,-0.0653772205,-0.0596348085,0.1529756784,-0.0461476445,0.1430097818,0.2116665393,-0.0876234919,0.2434218526,-0.1006888077,-0.2109390497,-0.134845376,0.0749133453,0.0712182298,0.0836840048,-0.0305534955,-0.1203543916,0.0264120288,0.058853928,0.3391480148,0.0564361215,0.4760086238,0.035745468,0.3672853112,-0.2276158333,0.2897334397,-0.0722121149,-0.0454164296,0.2744023502,0.0004257601,0.5121608973,0.0745410621,-0.1095569953,-0.2221313715,-0.0729920939,-0.1039227471,0.0380488485,0.1324604452,1.1257487535,0.328962177,0.0377922431,0.322550863,0.1011457443,0.4174523652,-0.3723428547,-0.0780912116,-0.1175964177,-0.1119009778,-0.0754860714,-0.1083521098,-0.0685053393,0.0830523223,-0.4119778574,-0.0901867151,0.0935385376,-0.0378630273,0.0057590241,0.3582372069,-0.2234948426,-0.3446009755,0.0054323357,0.2527285814,-0.2597345114,0.122023277,-0.0510405116,-0.1156575233,-0.0420904048,-0.2122900039,-0.3557643294,-0.0791419595,-0.1158489585,0.0163981784,-0.1187106147,-0.2603164017,-0.3931949437,-0.4722172916,0.1464754045,0.1668835431,-0.2364411503,-0.0338361263,-0.2105513662,-0.1473634541,0.0603980348,-0.1346817017,0.1946250498,-0.0242949091,-0.3363328278,0.0735561177,0.0514823608,-0.089957431,0.4749714434,-0.1214457825,0.1072713956,0.1345890313,0.3538151979,0.2668706775,-0.3107757568,-0.2847304642,0.2133523226,0.2541810572,-0.2631848454,0.0771743059,0.0967084765,-0.0011073686,-0.2568236291,0.4368759096,-0.0553010739,-0.0261058155,0.0106996782,0.2215566784,-0.1990110576,-0.3756071031,0.0922975093,0.0463563241,-0.440669179,0.2439425737,-0.2361664474,-0.3502997756,-0.2419025898,0.3663056195,0.3758274913,0.4089523852,-0.0247769486,-0.4372684062,-0.0649759397,0.3385904133,-0.2752910852,0.0909104943,-0.2320914865,0.5329762697,0.2194464803,0.1724179089,-0.4702297747,0.0667384267,-0.3435592949,-0.0625151321,0.2846618891,-0.3762759268,0.2563636005,0.2098731846,0.2285183966,0.1995832622,-0.0418576971,-0.31905514,-0.2192121595,0.041325476,-0.0831858367,0.0582533069,0.1292321831,0.269472301,-0.0234930515,-0.1886211932,0.1112307608,-0.148851648,0.0650067553,0.1702756137,-0.1654929072,0.0777486712,-0.0685701072,-0.2459513992,0.0568082258,0.0679345801,-0.0006955434,-0.0113131953,-0.1898345798,-0.1794884056,0.0908634067,0.3038333654,0.1592198461,0.0141779846,-0.0317457877,0.1759515554,-0.2716871798,0.3827173114,0.4669646323,0.1605249196,-0.0365919471,0.0663274005,-0.123364985,0.3364463151,-0.3178904653,0.2393034995,0.0291433781,0.0190666448,0.4145539999,0.4714474678,-0.0747157484,-0.1081274152,0.236681655,0.0667110085,0.1551903784,-0.2402079105,-0.2382585257,-0.0876715034,-0.0859266892,0.0475506149,0.1945165694,0.1122501642,0.4746484458,-0.2354247719,0.0376126915,0.1002870947,0.0406000391,0.1428982466,0.327990979,-0.0409104563,0.2567054033,0.2584797144,0.0455380417,-0.0206385199,0.0381927304,0.597851336,-0.1981186867,-0.0071810195,-0.4543906152,0.2818242013,0.1027126461,-0.0286312811,-0.1455942392,-0.192822516,-0.4093618691,-0.2367829382,0.1066768244,0.0303995889,0.4083534479,0.0520018972,0.2306715697,-0.3958830535,-0.39821738,0.1973355114,0.2946583331,-0.3110479116,0.0026076161,0.5717426538,0.0518730544,-0.0503759831,0.7172040939,0.2138891816,-0.2227067649,-0.2931574881,0.0956832841,-0.3083190918,-0.0713371411,-0.2046506107,0.0046785632,0.4872709811,-0.0667352304,0.2789957523,0.2149055302,-0.3893035054,0.3618430495,-0.0075209169,0.091741845,-0.3114089966,-0.0006609976,0.1626613736,0.2265348434,-0.2334202528,-0.0882823542,-0.4648963511,-0.1227074191,-0.1091867313,-0.2664310932,-0.028084157,-0.1699664742,0.1203376129,-0.0369608551,0.3108156025,0.4283604324,0.0154275103,-0.1131104901,-0.3041285574,-0.3745430112,0.0971514881,0.0803198442,-0.2185714543,-0.483202368,0.3267868161,0.0939989835,0.5921640396,0.0417667851,-0.2411854118,0.2161262035,0.0370694511,-0.3163021505,0.2047961503,0.0088106748,-0.1555593014,0.3016020358,0.0475177653,0.0113609964,0.1880793273,0.2030011564,0.1140051484,-0.3534924686,0.1826294512,0.0313193053,0.3012471497,0.0594014674,0.2661758363,-0.1705606133,-0.2438613325,-0.0775653645,0.0824897215,-0.0444263518,0.3249352574,0.0264936872,0.2693081498,0.101801917,-0.0790498033,0.0145309884,-0.0539229773,0.1040873304,-0.0490860753,-0.2996728122,0.0251365062,-0.0413082726,0.1729389131,-0.059239585,0.2082289159,-0.0681930631,0.1640979499,-0.0508985817,-0.5344406962,0.1976064742,0.1305535734,0.0211022627,-0.2160770595,0.4518178999,-0.6408717036,-0.1828158647,-0.4566699266,-0.12735416,0.25054726,0.1949226409,-0.1603253633,0.1837486923,-0.188582167,0.002334191,-0.0384718068,0.3554107249,0.1156807914,-0.1133876517,-0.1092248186,-0.1269770116]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/846","title":"Add HoVer multi-hop fact verification dataset","comments":"Closed by #1399 ","body":"## Adding a Dataset\r\n- **Name:** HoVer\r\n- **Description:** https:\/\/twitter.com\/YichenJiang9\/status\/1326954363806429186 contains 20K claim verification examples\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2011.03088\r\n- **Data:** https:\/\/hover-nlp.github.io\/\r\n- **Motivation:** There are still few multi-hop information extraction benchmarks (HotpotQA, which dataset wase based off, notwithstanding)\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":3,"text":"Add HoVer multi-hop fact verification dataset\n## Adding a Dataset\r\n- **Name:** HoVer\r\n- **Description:** https:\/\/twitter.com\/YichenJiang9\/status\/1326954363806429186 contains 20K claim verification examples\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2011.03088\r\n- **Data:** https:\/\/hover-nlp.github.io\/\r\n- **Motivation:** There are still few multi-hop information extraction benchmarks (HotpotQA, which dataset wase based off, notwithstanding)\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nClosed by #1399 ","embeddings":[-0.2926411331,-0.1824046522,-0.0832223669,0.0094794845,-0.2821558416,-0.0880276859,0.2173709869,0.175518319,0.0569905005,0.1214196235,-0.0253052227,0.0805580094,-0.0879793838,0.2176565528,0.2109585851,-0.0655959621,-0.156846866,-0.0501364134,-0.0371176004,0.0134783098,-0.048302047,0.1000273973,-0.1147566214,-0.0549463704,-0.4133689702,0.0432088971,-0.1005345881,-0.1333541572,0.1836408526,-0.3441006243,0.0701467097,0.4664562941,-0.0346626639,0.3261966109,-0.0001062558,-0.1653868854,0.1518639326,0.0597814582,-0.3415099084,0.0861944854,-0.1334052384,0.0455760993,0.1221976653,-0.1493278742,-0.1062269583,-0.0654234067,-0.2154440582,0.1172586009,0.2446358204,0.1495280564,0.2732451856,0.3261539936,-0.1651806682,-0.1800213754,0.2050244212,-0.1409967989,-0.0358669981,0.4638676941,0.0870537236,-0.0016045957,-0.4450533688,0.3908061087,-0.0228549372,0.164638564,-0.1309421211,-0.1386982799,0.2436672002,-0.0033331951,-0.0151494425,0.6459630132,-0.0770192593,-0.126639396,-0.2680917382,-0.2150569856,0.0391991399,0.1102074087,0.0098436298,0.0812515467,-0.2807851136,0.2027350515,-0.1613394916,-0.1892645657,-0.0108404746,0.1054849774,0.2533622384,0.2799039781,-0.0414538756,0.0126071004,0.3020769656,0.2004768103,0.0435484797,-0.2540850639,0.1921586096,-0.2317866981,-0.0772964209,-0.276389569,0.335634768,0.3323394954,0.3419082761,0.3325184584,0.2849826217,0.1739403903,-0.2147066146,0.0759190917,-0.0388166606,-0.0821781158,-0.3042308688,-0.5394927859,0.3314715028,0.2093583792,0.1393019408,0.0732675344,0.1967903078,0.2431105971,-0.3780615032,0.075476326,-0.0756325573,-0.1090321317,-0.2511191666,0.1304467767,-0.3375341296,-0.1614359319,-0.0078624729,0.0560452454,-0.0748350695,0.3618521094,0.1850007176,0.0904989839,0.0636816323,-0.3660285473,-0.2255158275,0.3185940087,-0.0697508305,0.1649014503,0.2010704279,0.1498136967,0.1380792856,0.0711133331,0.2513661683,0.0571643375,0.1719981283,-0.2277094722,0.0121060004,-0.1018400192,-0.0837786198,-0.1610542387,0.1013650969,-0.1900025755,-0.1064477041,-0.0077266651,0.0684797689,-0.1782903522,-0.1655776501,0.2320199162,-0.3756120205,-0.1787757725,-0.2864339948,0.4785857797,-0.1104278639,0.1732627451,0.1400043219,0.1509780437,0.1037250236,0.0167229343,0.1142221466,0.3217819333,0.1054617465,-0.2500014603,-0.0386625156,0.0094469544,-0.1135753095,-0.0049866736,-0.0870264471,0.2435325384,-0.2019051313,0.1776392013,0.0441367179,-0.3320504427,-0.1551666409,0.0962302312,-0.3663288057,0.1026533693,0.3016909957,0.1048454642,0.0760781839,-0.1495029777,-0.1274667084,0.5180132389,0.0721448511,0.1635618955,-0.377644062,-0.1656468064,-0.1623027474,0.1406258196,0.0182690341,-0.31602332,-0.0168884266,0.3055976927,0.3754940629,-0.079085581,0.2006211877,-0.0576195531,0.11770726,-0.5269863605,-0.0929024369,-0.2241512537,-0.0423688218,-0.117778331,0.1116555408,0.4301849008,0.225897789,-0.2319374084,-0.2312806547,-0.0583699718,-0.1534368694,-0.4802298248,0.2975873053,0.3965680599,0.1181276068,0.1219788343,0.0482598022,-0.0892243609,0.0585446954,0.0833967999,0.1626946628,0.2885733545,-0.0377165414,-0.0528239124,0.1587053537,0.2636120915,-0.1495578438,-0.1889282763,0.0764694214,0.4326452017,-0.229974106,0.0727872103,0.6216334105,0.2610376775,0.1621242911,-0.290725708,0.133721754,0.0953080654,0.0740979016,0.0464605391,-0.3078958988,0.3551093042,0.3471755385,0.0870099142,0.0755628273,0.158232674,0.1772206873,-0.1266625375,-0.1983170062,-0.1843211055,0.3688652217,0.0563780814,0.2344776839,-0.1823656261,-0.1634538025,0.3425341547,0.1362421364,0.0929805189,0.3897684813,0.1387778372,-0.2577393949,-0.4348027408,0.1220139638,0.0524802133,0.0628677309,0.3192191124,0.3993506432,-0.1272809952,-0.0456153192,-0.2745091021,0.2912228107,-0.2035769075,-0.2642000318,0.12545681,0.1698744297,-0.0771499574,-0.3629429042,-0.4166957736,0.0355778411,0.1225857735,-0.0472918674,-0.0750511736,-0.2316682786,-0.3038294911,-0.0875205621,-0.2160865515,-0.1893460602,-0.4182238579,0.1892332584,0.1910335571,0.1118303835,0.3215334415,0.1268539578,0.5644375682,0.0714393556,0.4004858434,-0.1620598286,-0.1486043632,-0.0696453676,0.2398556471,-0.037719477,0.1117368937,0.442755729,0.0793735757,0.100865528,-0.1778471768,-0.3010581732,0.2926563919,-0.0034021419,0.0895146057,0.2311865538,0.0075819283,0.2178614289,0.1646504551,0.1306245774,-0.1834563017,-0.163323015,0.1134348065,-0.1144768745,0.0619083382,-0.2108767927,0.1469052136,-0.0836219788,-0.2078428417,0.3805053532,0.0030046385,-0.0508528091,0.1565502584,-0.0271927956,0.3139796555,-0.2894527614,0.1491618901,-0.3468467593,-0.530800283,0.0241660401,-0.2234942615,-0.0700462535,-0.4743170738,-0.3644640148,0.1619862616,-0.1221678033,-0.4787297249,-0.415874511,-0.1525086164,0.2586300969,0.1364514381,-0.0155739561,0.339792937,-0.0972908661,-0.2970924675,0.0108086439,0.0699150413,0.1157884002,-0.1576492637,0.1742073596,0.0572112538,-0.1680077314,0.6028123498,0.390684396,0.0330010056,0.0036482522,0.2284851819,-0.1273395568,0.0545681454,0.0860243738,-0.3464310467,0.3340741396,-0.0204961002,0.0773340687,0.1444112808,-0.1779718548,0.2180064619,-0.2173922956,-0.0050110482,-0.1240520552,-0.3253725171,0.0551863983,-0.249353081,0.1008529589,0.0213351082,-0.2159358859,-0.0788289011,-0.1040667593,0.2173995376,0.1068974808,0.156926617,-0.3065971732,0.001655621,-0.0687965006,-0.3599436581,0.387208432,0.1023299545,-0.1503808796,0.0726346746,-0.1889680475,-0.0419929959,0.0567869879,0.3021114171,-0.448274225,0.1528776437,-0.0080255866,-0.0183347296,-0.5127301216,0.1578028202,-0.5497812033,-0.171788916,0.219239369,0.0307222884,-0.5071614981,-0.1390031129,0.7380422354,-0.0498485044,0.0323680751,-0.1460302919,0.0568998568,0.0340607837,-0.1995957345,0.085742414,-0.0416008048,-0.053786803,-0.0696529374,-0.0457066,0.1812526137,-0.0314376317,0.1379967928,0.1586990207,-0.1753229797,0.2660672069,-0.1137791574,-0.2634181678,-0.1695485115,0.0557754412,0.0580326393,0.0928694159,-0.0316542052,-0.0401786678,-0.0013019032,0.0741067305,0.3671077788,0.0544949211,0.4564806223,0.0105627971,0.3550541103,-0.1633567363,0.2972740531,-0.0561872981,-0.0360135026,0.2371344715,-0.0671853647,0.4585021734,0.0534109883,-0.1132172197,-0.2410182953,-0.1395718008,-0.0922075808,0.035684213,0.0932822078,1.110045433,0.3499253392,0.0677855462,0.3256233633,0.1653955728,0.4293482304,-0.3959325254,-0.0789203867,-0.0904708877,-0.0771625638,-0.0834302008,-0.0955791771,-0.054032594,0.0752859637,-0.3713331223,-0.0873905942,0.1503661126,-0.0272022933,0.0725958571,0.3750090003,-0.1897038072,-0.3312288225,0.0646157339,0.2458931506,-0.2726120055,0.1076574773,-0.0637468994,-0.1256402433,-0.0938426703,-0.199967742,-0.3753089607,-0.0634484813,-0.0670211613,0.0561224557,-0.2070159167,-0.3501097262,-0.3718960881,-0.5318844318,0.1176308095,0.1426368803,-0.2231043428,-0.0549948439,-0.2192800641,-0.1365193278,0.0836897716,-0.1987645924,0.2259029895,-0.0132939629,-0.3489095867,0.1598082334,0.090778023,-0.1875606775,0.4708525836,-0.0941180661,0.1214375272,0.102966845,0.3022011518,0.3264908791,-0.2829475403,-0.3771141768,0.2108428329,0.2758306265,-0.2589705884,0.0977740139,0.0961903483,-0.0336216353,-0.2423160821,0.4428671002,0.0296923816,-0.0252199192,0.0246209316,0.1900991946,-0.1838174015,-0.3592560887,0.1569516063,0.0632385984,-0.4485059083,0.2036780566,-0.2691105306,-0.3128500581,-0.284746021,0.3043771982,0.361384213,0.4721865654,-0.0627232715,-0.428717643,-0.0464917794,0.3014599383,-0.3301718235,0.0692737848,-0.2257284373,0.5716406107,0.2014207542,0.1406110227,-0.4661981463,0.064139232,-0.2762031853,-0.0859555006,0.2975970507,-0.411087662,0.2215637416,0.2125908583,0.2262466848,0.1491268873,-0.0332412831,-0.3264566064,-0.2059484571,0.0427443087,-0.0253273901,0.0552084818,0.1157696918,0.2069705129,-0.0268331487,-0.1755869389,0.0857886747,-0.1939745843,0.0496791899,0.1456490904,-0.1456302851,0.1077964455,-0.058332745,-0.1744917184,0.0483881645,0.1252936423,-0.0147349192,-0.0197608415,-0.1375834793,-0.148309499,0.0654821917,0.3075504601,0.1389796585,0.0848063678,-0.0037175089,0.1529565305,-0.2777144313,0.3814915717,0.4607661068,0.1595104933,-0.0552059859,0.0325841606,-0.1616338044,0.3275392056,-0.3352267742,0.226932019,0.0556793362,0.0656557977,0.4157958031,0.4898492396,-0.0798765495,-0.1163081452,0.2600768805,0.085689202,0.1578803807,-0.2927993536,-0.2362254262,-0.1031217352,-0.0844578147,0.0784662217,0.2423159331,0.089490436,0.4658797681,-0.1873736084,0.0797054395,0.1484080702,0.1677635014,0.1516531408,0.3191920221,-0.0263853632,0.2330100685,0.2018990964,0.012087618,-0.0503574684,0.078769125,0.4890813828,-0.2354154587,0.0348007865,-0.4394783378,0.242094785,0.0504569896,-0.0198738482,-0.1371614635,-0.1953598857,-0.3676930666,-0.2554170191,0.0452381112,0.0759950653,0.4089242816,0.0152284382,0.1646805257,-0.3564487994,-0.4142535031,0.1287582964,0.2621890306,-0.3257737756,-0.0424622521,0.4978591204,0.0408282541,-0.0981042907,0.7255489826,0.2239058614,-0.1994127929,-0.3121958971,0.0372934714,-0.299929738,-0.0786783844,-0.2465233654,0.0392437242,0.5122929215,-0.0554352775,0.2736164927,0.2115096152,-0.3936723471,0.4222272038,0.1272728592,0.0739962831,-0.3778429031,-0.0114312395,0.1973060668,0.1731378734,-0.2036616355,-0.0142332437,-0.4663616717,-0.1502051353,-0.0714300647,-0.2582857907,-0.0475511327,-0.2526482344,0.1191664636,-0.0912918225,0.2835497856,0.3870516121,-0.0117723718,-0.1602223814,-0.2760059237,-0.3572070301,0.1364608407,0.0237098075,-0.2260553241,-0.4798444808,0.3550781608,0.1227398068,0.6072481275,0.0803868473,-0.2222627699,0.2077378333,0.0748928338,-0.3103011847,0.181585446,0.0989558548,-0.1668448597,0.2572969198,0.0060011507,0.0016127968,0.2089764625,0.2014874667,0.0622316487,-0.4426154196,0.3020258844,0.1138672531,0.3493315279,0.0550334193,0.2400385141,-0.1564810425,-0.214614287,-0.1314392984,0.0597249307,0.0012862607,0.3640187681,-0.0232007001,0.2192474753,0.0885743573,-0.0557472073,-0.0027368281,-0.059517432,0.0620766059,0.0299466513,-0.2984604537,0.0768836588,-0.0684006587,0.1839211434,-0.0903267339,0.1426745355,-0.0613854155,0.16258277,-0.0229382962,-0.523955822,0.1875503957,0.1613318324,-0.0330145098,-0.2677288651,0.4870825112,-0.7016089559,-0.1750898063,-0.3998643458,-0.1867863089,0.31049034,0.2185500711,-0.1664726287,0.1975096315,-0.2184741646,0.024353724,-0.0273728222,0.3610826731,0.0201462582,-0.1401724666,-0.0978901908,-0.1643689424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/843","title":"use_custom_baseline still produces errors for bertscore","comments":"Thanks for reporting ! That's a bug indeed\r\nIf you want to contribute, feel free to fix this issue and open a PR :)","body":"`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2","comment_length":24,"text":"use_custom_baseline still produces errors for bertscore\n`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2\nThanks for reporting ! That's a bug indeed\r\nIf you want to contribute, feel free to fix this issue and open a PR :)","embeddings":[-0.0237094034,0.1297198385,0.1280906647,0.0422289521,0.1812677085,-0.0610478111,0.4962110817,0.0543297529,-0.0781167671,0.1093732566,0.0923896208,0.2173016816,-0.2150024921,-0.0573218279,-0.3110678792,-0.2098553926,-0.1186002046,0.2547813654,0.0980242491,-0.0632236227,-0.5826330781,0.0948995575,-0.5061203837,-0.215433374,-0.35657686,0.0306723919,-0.1378583461,0.0974576175,-0.1915955991,-0.4349971116,0.3886699378,0.2518529296,-0.0504473448,0.4357790351,-0.0001216621,0.0103914374,0.1038189605,-0.2130182087,0.0852774009,-0.2657131553,-0.3200367689,0.2057951838,-0.1055373028,-0.2700328529,-0.0187032651,-0.0896984786,-0.3070335388,-0.2126887143,0.2421084493,0.4283301532,0.1078686416,0.1945304722,-0.0735884681,-0.3550617695,0.0813606381,-0.3017669916,-0.1456321925,0.4351503253,-0.4142250419,0.0377761535,-0.1057628989,0.0243913326,0.0382915474,0.2849423885,0.4474446476,0.1052068099,0.1581571102,0.038551081,0.1076110601,0.3610633314,-0.0474509597,-0.2185909897,-0.2287302911,0.0338249914,0.1287768334,-0.7086322904,0.0648087263,-0.2146535367,-0.1373078823,0.0183905475,-0.1703177691,-0.2219579816,-0.3350681961,0.1999560148,0.0099008847,-0.0556113496,-0.1081215888,-0.0112577705,0.5676791072,-0.0950888395,0.1574656516,-0.2250990719,0.0535053648,0.0574578121,-0.2062042654,-0.1097372621,0.12056005,0.0120413145,0.054992035,-0.4186506271,0.2580140531,0.0277164709,-0.0427124538,0.0072117271,0.0113446796,0.4146568477,0.0075685591,0.0421947539,0.0975388438,0.191309765,0.0582834631,0.0420947596,0.1064490676,-0.2987893522,0.4436658919,0.1426459551,0.0499458313,-0.0650884509,-0.0797337145,-0.0866107643,-0.3678119183,-0.1821782291,0.123251684,0.1614996195,0.1791384518,0.3282141685,-0.2257362157,0.3017466664,-0.0212723557,-0.0919406414,-0.1466501504,0.1026794463,-0.4289891124,0.2630151808,0.3088880479,-0.231842652,0.338822633,0.3215659857,0.3396804333,0.0668376684,0.0925773531,-0.2689784467,0.1435692757,-0.1173270866,-0.0588499606,-0.1988247335,0.1535756886,0.2831663787,-0.1306258589,0.1588048041,-0.2481677085,-0.2222669274,0.2047489434,0.0787288472,-0.2645999789,-0.0504632033,-0.4324563146,0.1234535053,0.2307007313,-0.0401564389,-0.1669592857,-0.212514475,-0.5594500899,-0.1909614503,0.2443027347,0.1580370814,-0.6430358887,0.0087558338,-0.327750057,0.3320314586,-0.0049079126,0.0230638236,-0.1867527962,0.524061203,-0.1446579248,0.0781480223,0.1206864864,-0.3428070843,-0.3460653424,-0.3002776504,0.2792723775,-0.0698436797,-0.0094367675,-0.0923496783,0.1101372689,-0.0958160833,0.2726924419,-0.2665002644,0.2159829736,0.048586145,-0.1672807783,-0.1360288858,-0.0962842554,-0.0488965586,0.0996153727,-0.0844486505,-0.081754446,0.2319123149,-0.2356778234,-0.279645294,-0.0172303971,0.1953890473,0.1852337569,-0.3643390834,0.2012469918,-0.033185754,-0.0130387833,0.2165409625,-0.0341511108,0.3013232946,0.1570087224,-0.0586365052,-0.414134711,-0.1540224254,-0.1648738533,-0.1367589384,0.020873813,0.052580867,0.0798623934,0.0717615858,-0.2632439435,-0.6167683601,-0.2063766569,-0.0747958198,-0.1205357388,0.1412699074,0.0095126759,-0.111390762,0.2324660718,0.4710415006,0.1814607084,0.1329210252,-0.1492077559,0.3699563742,0.0696664825,-0.0689894408,0.1675907969,0.2383378595,-0.0214524176,-0.3603651524,0.1581872851,0.0130310114,0.0414637662,-0.0580951311,0.0373458974,0.4591887891,0.1649273485,0.2114328593,-0.1492764056,-0.0983438566,0.0560314693,-0.0891301706,-0.0203597043,-0.4252582192,0.0593474247,0.0462406427,0.1611888856,-0.0780128688,0.1707072854,-0.1522732377,0.6477606297,0.2304110974,0.3287092149,0.2288495749,0.0277477484,-0.1844192892,-0.2881701887,-0.2612560987,0.6102736592,0.1198472455,-0.1168618053,0.300278455,-0.0058482424,-0.1454414725,0.1483454257,0.2962324023,-0.151988253,-0.0261059254,0.2358247638,0.0336825959,-0.0323423557,0.0176370312,-0.0364517421,0.2214458287,-0.4314708114,0.1298580766,-0.3040905297,0.1482007354,0.0089214835,-0.4882210791,-0.0718791708,-0.2311621457,0.0693006143,-0.0674946308,0.0923007056,0.5150328875,0.1193876714,0.7285951972,0.2180926949,-0.0683989078,-0.0394961499,-0.355155766,-0.0989579856,0.01308465,-0.2014216632,-0.1084744856,0.243563354,-0.195292443,-0.1802336425,0.1880682409,-0.3187426627,0.028164126,-0.0842724815,0.3037753999,0.521917522,-0.1194943562,-0.1522707939,0.1433849931,0.2648014426,0.0178211331,0.2042805403,0.0669968277,-0.2079891413,0.0298557431,-0.0894269943,-0.2768165767,0.3283346295,-0.1472149789,-0.1486426443,0.0162934661,-0.0162279829,-0.0395782441,0.3637632728,0.2727850378,0.1033838913,-0.2550582886,-0.1482810676,-0.6655891538,0.4515167773,-0.2910750508,-0.1359456033,0.0890357345,-0.2381780744,0.547031045,0.0908265561,-0.2057355195,-0.5882261992,0.0714426935,0.2592980862,-0.0418433808,0.0959583148,0.1591807753,0.0861903727,0.0080002183,-0.1678431779,-0.1875739247,0.1888466477,-0.3293327689,0.1942361444,0.1593272686,0.2722424567,0.133979097,0.8989217877,0.2725177705,-0.273058027,0.4387993217,-0.0669018626,0.3824104071,-0.394960165,-0.3603945374,-0.1386916637,0.1852064878,0.0897334889,0.2308573872,-0.0889833495,0.3890800774,-0.0727471411,-0.2514277101,-0.0915350616,-0.3395074904,-0.0338156782,-0.0550406836,0.5190787911,0.0533272326,0.3071940243,-0.3714312017,-0.2084727734,-0.1186090559,-0.1633488834,0.0669531152,-0.1605889201,0.2298687398,0.3610686958,-0.4554145336,0.4417236447,0.0863184631,0.102569744,0.0685720146,0.4200751483,-0.1246188506,0.0207336545,0.2103983164,-0.4105684459,0.0395163521,0.0080110235,0.1348177642,-0.4824214578,0.1323272586,-0.0598307215,0.0011058162,0.2754400074,0.6075905561,-0.1760204732,0.0478164628,0.3029841483,-0.1043117791,-0.079675816,-0.2913924754,-0.337318927,-0.1893279403,-0.0420688093,-0.0433572195,0.0202027094,0.2483237386,-0.0108463326,0.22800228,0.1602766365,-0.3473769724,0.1340127736,0.2677551806,0.1726616025,-0.1401166171,0.101267986,0.1739390194,-0.6338841319,0.5574402213,0.2815368772,-0.397859484,-0.612365067,-0.0742356777,-0.0695870072,0.3827637434,0.1446547508,-0.2513596714,0.0405860804,-0.2011119574,0.0878574625,-0.1828409284,0.3273381293,0.4565599263,-0.1059441194,-0.1726165563,-0.502494216,0.4625495672,0.0861486495,0.0143510271,0.0512884557,0.2718212008,-0.3436954021,0.268666476,0.0771053955,0.5343260765,0.3586908579,0.0433405302,0.2417575568,0.2244484723,0.2928492129,0.2980546951,0.0608589984,-0.1527870893,-0.0186553188,-0.0219232365,-0.1315648258,0.1438341588,0.1072168276,-0.2754093111,0.5928207636,-0.2126577497,-0.1861461103,0.0355654284,-0.1029080451,0.1143590063,-0.2550799847,-0.0573831648,0.0981801525,0.0167520922,-0.1190953627,-0.0321326628,0.0632062852,0.1672267318,-0.2652075291,-0.259388119,0.0015917486,0.0619140044,-0.1357498765,0.0750088021,0.0207689721,0.3008498847,0.1344204098,0.6360820532,0.2668307424,-0.1996303201,0.1343509853,-0.1055977568,0.0545005165,0.0288880058,0.0366788171,-0.1748997867,-0.0665103942,-0.2871817946,0.1855507344,0.1030964628,-0.2627031505,-0.3800094128,-0.2893660367,0.0602564923,-0.1784206182,0.1185602248,-0.1847486645,0.1300378442,-0.2566103041,0.0846286044,-0.1447120905,0.0288334731,0.1587762535,-0.3751531541,-0.2457477599,0.0785041749,0.3984597623,-0.1668800712,0.1826784462,0.5197712183,-0.1384222358,0.0691487119,-0.0718600601,0.1067755669,0.0612329803,-0.1613891721,0.3675586581,0.2718496621,0.2357190698,0.1165106818,0.7036588192,0.0834739953,-0.2088410258,0.1764168739,-0.1835187674,0.1219914705,-0.174578473,0.0451837964,0.0757613555,0.0343478508,0.1804419905,-0.0717571899,0.0012457321,-0.1992887408,0.2200905383,-0.0626442581,-0.1084517762,-0.0790426508,-0.1588965803,0.2469670326,0.0427795574,-0.0239221007,0.3228087425,-0.2335000336,-0.1334548295,-0.4277101457,0.1314189285,0.055531051,-0.1027385592,0.030966742,-0.0170362554,-0.2394918799,-0.1901258826,-0.0479831472,0.4341006577,0.0197228026,0.426517576,0.0594023392,0.2494660765,0.1284283102,-0.0400906131,0.1799040884,0.3275191784,0.0259820111,0.3166768551,-0.0752786621,-0.1005652174,-0.0495298877,0.0780814588,0.3835342228,-0.0928287655,0.1734530181,0.0398061313,0.0129743228,0.1488582492,0.3509458303,0.0870776623,-0.421803534,-0.3142795861,-0.0661010295,0.0702042952,-0.4328536391,0.0043686288,0.0172031522,0.4553099573,0.0943941101,0.2588648498,0.1548332721,-0.1870606095,-0.0921345949,-0.0252041742,0.1609390974,-0.2590942979,0.292109102,0.0573968291,-0.1557390243,0.1425219923,0.590616703,0.1923966855,0.2686830163,0.6254937649,-0.1610739976,0.5122528076,-0.2839603424,-0.1504513919,-0.2018979341,-0.211322993,0.1975490451,0.2666290998,0.1487216204,0.1288258135,-0.1132487208,0.5228831172,-0.0128005464,-0.0169586577,-0.4258069694,-0.0685613081,0.0128795002,-0.2093350887,-0.3719137311,-0.2758212686,0.015103763,-0.0126633886,-0.0385220498,0.0434074998,0.1811032593,-0.0445868932,-0.0594086163,-0.0848025605,-0.121227257,0.1452345848,0.1957995445,-0.014512145,0.1591163427,0.3156150281,0.1271768957,0.1770684272,0.4339847565,0.2479565293,0.2180333138,-0.0452142358,-0.2701665461,-0.054088179,0.0116945701,0.1586028188,0.1665058136,0.253525883,0.301089406,0.2748164237,0.0601846725,-0.1429127008,0.5313389301,-0.1550447047,0.6881144047,-0.5048012733,0.5134906173,0.0508231185,0.1847126186,0.2451091707,0.0496638305,-0.1778466105,0.1560327709,0.2319923341,0.1972460449,-0.1477131248,-0.3273929358,0.0459554344,-0.1239377558,0.4184467494,0.3465740383,-0.1592112631,-0.6492037177,-0.032610409,-0.6483089924,0.0520336814,-0.114040643,0.2721144259,0.0522238053,-0.0635034069,-0.1088919193,0.4449110329,0.1597812325,-0.2190127373,-0.3276512027,0.0375858359,-0.4445097744,-0.093409054,-0.2396303117,-0.4727941453,0.1905007213,-0.2825447619,-0.2204064876,-0.1620347947,0.0217777994,-0.0148912482,-0.1854910254,0.3216646612,0.380898267,0.4474574924,0.1773633659,0.3423072994,0.0542104915,0.0294941887,-0.0818478614,0.0103957821,-0.1400308758,-0.018311929,-0.0683735162,0.094009459,-0.3902923167,-0.2314219326,-0.3763729036,0.13938573,0.1452989131,-0.2436527908,-0.4069201648,0.151377663,0.0800674334,-0.1785117537,-0.1112259775,0.1306920797,-0.1412380636,0.4892010391,-0.1556322873,0.0389573686,0.5096229911,-0.3619689047,-0.3078297973,-0.0988688171,0.275654763,-0.1014277264,-0.3412373662,-0.7951288819,0.1672400385,0.5032975078,0.1859323382,-0.1084079519,0.1340212375,-0.291313976,0.1869578958,0.0937094018,0.2266325355,-0.0618694797,-0.2960381508,0.0938289091,-0.191813305]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/843","title":"use_custom_baseline still produces errors for bertscore","comments":"This error is because of a mismatch between `datasets` and `bert_score`. With `datasets=1.1.2` and `bert_score>=0.3.6` it works ok. So `pip install -U bert_score` should fix the problem. ","body":"`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2","comment_length":27,"text":"use_custom_baseline still produces errors for bertscore\n`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2\nThis error is because of a mismatch between `datasets` and `bert_score`. With `datasets=1.1.2` and `bert_score>=0.3.6` it works ok. So `pip install -U bert_score` should fix the problem. ","embeddings":[-0.047540348,0.1264257878,0.1010241136,0.0786020234,0.1025783569,-0.1123297587,0.3708122969,0.0062376615,-0.0859037638,0.0964994654,0.0731694847,0.327162832,-0.1698618382,-0.0304797478,-0.3144399524,-0.2213500589,-0.0507825762,0.281804949,0.0639780164,-0.0864623636,-0.5255195498,0.1529728025,-0.4210203886,-0.1924541593,-0.2500153184,0.057941366,-0.0525206365,0.1060039327,-0.2861156464,-0.5162747502,0.4373293817,0.2641012669,-0.0446660072,0.51725775,-0.0001244289,-0.009829429,0.0957094356,-0.1925257444,0.0883710682,-0.2875192165,-0.3482615352,0.1504689157,-0.0692221671,-0.2116601765,-0.0044016042,-0.0118120238,-0.2781254649,-0.1751513332,0.3080380559,0.3524618149,0.087459974,0.2012179941,0.0653075576,-0.3420537114,0.1931594759,-0.3299062252,-0.1392865628,0.396102488,-0.2709240019,0.0564828888,-0.0680724606,0.0319634862,0.0056583192,0.2695562541,0.4815789163,0.0913681462,0.1096116006,-0.0055419882,0.0920879766,0.4091941118,-0.0803010538,-0.2674737573,-0.3128249645,0.0369421206,0.0314069428,-0.6414815187,0.0790718198,-0.3048743904,-0.1437421143,-0.0087263035,-0.1705687344,-0.2099008262,-0.3400878906,0.2357411385,0.0795747936,0.0174178965,-0.1047035903,-0.0014914243,0.5478115082,-0.1566693038,0.2536180317,-0.2071748376,0.0818214938,0.0793578997,-0.1765011549,-0.0984943062,0.0903500095,0.0529368296,0.0696575493,-0.4432585835,0.2492204756,0.031260334,-0.0459455326,0.0243309587,-0.0558708087,0.4640091956,0.0278092958,0.117042847,0.0476976894,0.2404187471,0.1034623459,0.0706096441,0.1022755131,-0.2756897807,0.352683872,0.1294337809,-0.0083996756,-0.0959280729,-0.1331208944,-0.0568191297,-0.3427043259,-0.1830422878,0.1160185039,0.1568683684,0.1579164416,0.3257536888,-0.2200782448,0.333032161,-0.040930856,0.0152143082,-0.0834589899,0.1074834988,-0.3994547129,0.1989299655,0.3128276765,-0.2242771536,0.3718512356,0.2852113843,0.3254767656,0.1721015424,0.1374445558,-0.2168230414,0.1365477294,-0.1047816947,-0.0621742457,-0.1410234571,0.1644819975,0.3025271297,-0.1387709528,0.1754385531,-0.2223198265,-0.1910640001,0.1717914045,0.0680136234,-0.2613267899,-0.1120781973,-0.42405352,0.0528318025,0.2243285775,-0.1529645026,-0.1727681011,-0.2411655188,-0.4576372504,-0.1593731344,0.2582949102,0.1398139447,-0.6280488372,0.0148948627,-0.3402602077,0.3336730003,0.0314761363,0.0371164419,-0.1415841877,0.486243397,-0.0757541806,0.062522389,0.1722272187,-0.4790654182,-0.4099445045,-0.2716494501,0.3522036672,-0.0814557672,0.0634876043,-0.1295196265,0.1314167827,-0.0889102668,0.228276208,-0.2969485223,0.2183344662,0.0493387505,-0.20322375,-0.2176333964,-0.194954142,-0.0615558401,0.1458945423,-0.1838209331,-0.1019391194,0.2684340775,-0.2421149611,-0.2514140606,-0.0068490189,0.1727921665,0.2394328862,-0.2319909036,0.206085816,-0.0433927774,-0.0264950935,0.2278805822,-0.1440651417,0.2700972855,0.0971954092,-0.0890802145,-0.477510035,-0.1666538417,-0.1690966636,-0.1864432544,0.0201113988,0.0931874737,0.1828804165,-0.0067810942,-0.2632796466,-0.6373126507,-0.1353193223,-0.0122356331,-0.180956468,0.0817297697,-0.0653166324,-0.1508963406,0.2251847386,0.4985397756,0.1647060215,0.093275629,-0.1004620716,0.3833450675,0.0325173065,-0.0487697646,0.1908474416,0.1934936345,-0.0505404882,-0.3579322696,0.1553335935,0.0782532468,0.0530722439,0.0505020022,-0.0071354667,0.3536441922,0.1620366573,0.2635678947,-0.052743312,-0.1715869904,0.0601253361,-0.0859342143,-0.0406900905,-0.4568100572,0.1455317885,0.0303775761,0.2230425775,-0.1036999673,0.2699281871,-0.0982603431,0.5954118371,0.1917045116,0.2943177819,0.1994555593,0.0498729832,-0.1664097756,-0.2326803654,-0.1959205866,0.5780509114,0.1562895626,-0.0870840326,0.2507531345,-0.0461638346,-0.1074400544,0.1028246507,0.2760710418,-0.2259881347,0.0223335922,0.2249092609,0.0746869966,0.0045519606,0.0407139063,-0.0784461126,0.2430475801,-0.4302924275,0.1707729697,-0.2851997912,0.1154578626,-0.0365211815,-0.5909368396,-0.0689788982,-0.2016040832,0.098447606,-0.0695415139,0.0941003263,0.5627202988,0.1588648558,0.6345853806,0.1455806941,-0.1373713613,-0.0077177379,-0.3456173539,-0.1424933821,-0.0156793501,-0.1727605611,-0.0790925995,0.2614364624,-0.2617178559,-0.2088693529,0.2103558183,-0.4115682244,0.0435671434,-0.0928409919,0.3466618359,0.5264642239,-0.1340174824,-0.173584491,0.1042517945,0.2667131722,-0.0165022183,0.1917164028,-0.0075533018,-0.2139127105,-0.027662728,-0.0042331172,-0.3713999093,0.3031690419,-0.1856302619,-0.1432873011,0.0525251664,-0.0169995427,-0.1507223099,0.3766866028,0.3246974349,0.0195614249,-0.1791644841,-0.0949535593,-0.6065406799,0.4005813301,-0.251252979,-0.1187706888,0.099317506,-0.2435000241,0.5438969135,0.0890605301,-0.1721716076,-0.6287647486,0.0189840663,0.1671383828,0.0131574627,0.153705582,0.1817462444,0.1393284351,0.0057116766,-0.2012548894,-0.1687321663,0.1920139492,-0.4219176769,0.1910488755,0.168325454,0.2982525527,0.082905747,0.9984369278,0.2949477732,-0.2306022644,0.4176668227,-0.0261317436,0.3826902211,-0.413040489,-0.3121481538,-0.1641314477,0.2399426103,0.1465793103,0.242016241,-0.061296206,0.4389861822,-0.0926675573,-0.1720948964,-0.1663753986,-0.3272832036,-0.0104576685,0.0201057643,0.5321362019,0.0834752694,0.2833065093,-0.3268345594,-0.2500029504,-0.1058009565,-0.1618316025,-0.0333570726,-0.0744611397,0.2539100349,0.2720233798,-0.4521148801,0.5129059553,0.0967228115,0.1286413223,0.0144802965,0.3394744098,-0.0900992453,-0.0172878709,0.1277976781,-0.4057281017,0.0249430723,-0.0481578708,0.1609790772,-0.4846993387,0.121554181,-0.0886702314,-0.0307448264,0.3386254013,0.5947799087,-0.2976599932,0.0695525035,0.3657417893,-0.0862273425,-0.0149191264,-0.2503311336,-0.3436200917,-0.1520916224,-0.1540435851,-0.1198301688,-0.0043402561,0.2229527682,0.0521177612,0.1928667873,0.1507243514,-0.3360343277,0.1529764831,0.3328763843,0.1639515013,-0.1490710825,0.0832159668,0.1456437707,-0.6445592642,0.5847187042,0.3334153593,-0.5383549333,-0.6002219319,-0.11169976,-0.035782963,0.2527258694,0.1808994561,-0.2176814526,0.0380877629,-0.1042931601,0.1029699594,-0.0830947533,0.3479178548,0.4639826715,-0.0874484777,-0.1575486511,-0.5294520259,0.4799644351,0.1455224901,-0.022885453,-0.0147880055,0.365658462,-0.3212712407,0.323394686,0.0711866617,0.584533155,0.3510916531,0.0616562143,0.2266031206,0.180706501,0.3489794433,0.3082762361,0.0471376777,-0.1556598395,0.0970692933,-0.0514413007,-0.0707825199,0.1444441825,0.0954973251,-0.3130029142,0.6615698934,-0.0801577792,-0.1753310114,0.0459352508,-0.1420898736,0.0693012699,-0.2309725881,-0.092320554,0.0870451257,0.1109674647,-0.0641193688,-0.0626290441,0.0776703432,0.0323701203,-0.3183478415,-0.3015075922,-0.0515706912,-0.0457399078,-0.160887897,0.100705348,-0.1219303533,0.2481879592,0.0742624626,0.6118413806,0.2196636349,-0.2011515349,0.1219270304,-0.0803492293,-0.0363658704,0.0418829359,0.0539523065,-0.18730326,-0.0847094283,-0.1997782588,0.1623312831,0.0749458373,-0.2091068178,-0.2846113443,-0.2393483371,0.2234270573,-0.1772479713,0.0648167133,-0.2225438952,0.113662675,-0.2210837454,0.076035805,-0.1922015995,0.0370546617,0.1392563283,-0.3807074726,-0.2317741662,0.0547251441,0.3735398054,-0.1737287939,0.1404579282,0.5156523585,-0.171253264,0.0694394112,-0.0331016481,0.0473871157,0.0579795614,-0.22301507,0.2284890264,0.2958238721,0.1286980361,0.1665138602,0.6771206856,0.1922251731,-0.2087531835,0.1540137231,-0.2043345571,0.1462817788,-0.120679915,0.0931998938,0.0807235539,0.0545607321,0.0711248964,-0.0874379352,-0.001255392,-0.1929501742,0.1930587739,-0.0121723823,-0.0783282667,-0.0290332362,-0.1146666035,0.2602442801,0.0055253902,-0.0417127758,0.2775156498,-0.2428808063,-0.1258268058,-0.38765499,0.1283793598,0.0745543018,-0.1210663617,-0.0004356652,0.0485573635,-0.1436318904,-0.1750052124,-0.0119156875,0.3709057868,0.0360195376,0.3890729547,0.0001861664,0.2678272426,0.1356675327,-0.0006349397,0.1431619823,0.3406485617,0.0472090952,0.293812722,-0.0505764671,-0.1108507663,0.0956293195,0.0604146756,0.508589983,-0.0917667523,0.1378829926,0.0066509061,0.0412440524,0.161800161,0.3347967565,0.2061110586,-0.3880255222,-0.3362266719,0.0433190204,0.0623811707,-0.414481163,-0.0175596587,-0.0311296936,0.3678812683,0.0829413608,0.2416049838,0.1816250086,-0.2028316855,-0.0254024602,0.0475308932,0.2192849964,-0.2425729483,0.3168146014,0.1429430693,-0.1601440161,0.1709577441,0.5377205014,0.146738857,0.30654338,0.5560675859,-0.1749820411,0.5516892672,-0.3667337298,-0.0827362835,-0.2097264677,-0.2243664861,0.1176189259,0.1864117831,0.2387371063,0.1637248844,-0.008943025,0.5908870101,-0.0506694689,0.0171236172,-0.42791152,-0.076276198,0.074365139,-0.2974324524,-0.3414063752,-0.259840548,0.0050450778,-0.0187735148,-0.0942322388,0.0429615602,0.1572687924,-0.0256230645,-0.098420687,-0.1696415544,-0.191411525,0.1319750249,0.1011875793,0.0063191261,0.2313850671,0.3034433126,0.0641277134,0.1433519572,0.5091142654,0.3321654797,0.2770749331,-0.0855097324,-0.2339012027,-0.1135610193,0.054075662,0.0989252776,0.1439590305,0.210880518,0.2965940535,0.2634874284,0.0378999859,-0.1765951961,0.4662085474,-0.2152259052,0.6183242798,-0.4149276912,0.519543767,-0.0124508403,0.2282214612,0.2758104205,0.1006528437,-0.3554702997,0.2301421911,0.1334057003,0.2126860917,-0.1203248352,-0.3135837913,0.0347558856,-0.1234557852,0.4652172625,0.3406949341,-0.0994321853,-0.6055421233,-0.0359825417,-0.6710897088,-0.0210807566,-0.1803076416,0.2058196366,-0.0289071314,-0.1119681001,-0.1211819649,0.476606518,0.1699824631,-0.2435964644,-0.2885995209,0.091237314,-0.4174422026,-0.0837603956,-0.1503177732,-0.4924095571,0.2141640633,-0.2705977261,-0.2216986716,-0.1842904836,-0.0247655846,0.014156105,-0.2034019679,0.3416087031,0.2944447994,0.4395036697,0.2000734359,0.4376264811,0.0445015132,0.0343798362,-0.0662430152,0.0148658697,-0.1716544032,0.0311927851,-0.0418829098,0.0817328393,-0.442838937,-0.3534297943,-0.3720622659,0.1289663911,0.1001348197,-0.2365226001,-0.3508346975,0.1414265037,0.1127920225,-0.1639038026,-0.1411022842,0.1018549874,-0.0969401971,0.4392127693,-0.2368866652,-0.0387350395,0.5047754049,-0.388880223,-0.2466631681,-0.1457577646,0.2097215205,-0.068789281,-0.3297351599,-0.8442142606,0.2420658618,0.4301360846,0.2046894282,-0.127221182,0.1177575737,-0.2999602258,0.1608386189,0.056592375,0.3699767888,-0.004196431,-0.268149823,0.0867065191,-0.1939450651]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/843","title":"use_custom_baseline still produces errors for bertscore","comments":"Thanks for the heads up @pvl and for the PR as well :)","body":"`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2","comment_length":13,"text":"use_custom_baseline still produces errors for bertscore\n`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2\nThanks for the heads up @pvl and for the PR as well :)","embeddings":[-0.0218941793,0.1136413068,0.112302199,0.0570567325,0.1592639089,-0.0578251593,0.4743404388,0.0436180383,-0.0856555477,0.0781214908,0.0820973739,0.2076596171,-0.2462242991,-0.0410770774,-0.3367952108,-0.2097294033,-0.1138471812,0.2605832815,0.1124425605,-0.0634572506,-0.596329391,0.0944170803,-0.5166925192,-0.2089489102,-0.3638880551,0.0089309709,-0.1511327028,0.1041851714,-0.2014132291,-0.4369574487,0.3743726015,0.2792129517,-0.0177721214,0.4813259244,-0.0001232784,0.0089111999,0.0725463778,-0.2114634216,0.104233861,-0.2636010051,-0.309502244,0.1964526176,-0.110593155,-0.2633963525,-0.027444642,-0.0954083279,-0.3114771545,-0.2165827453,0.2758494914,0.4349390864,0.0988326073,0.1866126806,-0.0645147637,-0.3469269276,0.0693414509,-0.2854550183,-0.1582451016,0.4327476621,-0.3868806362,0.0286703389,-0.0774104893,-0.0064484365,0.0298947785,0.2750663459,0.4685847759,0.0898507982,0.1353470981,0.0165152289,0.1270869076,0.3683955371,-0.0263542589,-0.207177639,-0.230287239,0.0362715907,0.1042927802,-0.726387918,0.0655155405,-0.2413693666,-0.1364225447,0.0227344781,-0.1771344095,-0.230724737,-0.3604672253,0.2085285485,0.0297667589,-0.039715901,-0.1032571346,-0.0141784046,0.5806558132,-0.135253638,0.1817461848,-0.2254330665,0.0892734826,0.0765444487,-0.1945025176,-0.1120496988,0.1237638965,0.0161883794,0.0410013832,-0.4241900742,0.2551020086,0.0518831015,-0.0241838936,0.0161622614,-0.0346031152,0.4124668539,0.0222809687,0.0769869983,0.106606476,0.2062343359,0.0347386487,0.0527523309,0.0749365464,-0.305881083,0.4417506456,0.1354361773,0.0409283489,-0.0709537268,-0.07110852,-0.0771070793,-0.3721138239,-0.1870043129,0.1295906454,0.16317904,0.2003038675,0.3283219934,-0.2298715413,0.2978575528,-0.0200391114,-0.072374478,-0.1290157884,0.1187709942,-0.4198554456,0.2668346167,0.3302771747,-0.219974488,0.3323104382,0.3090642095,0.3211795092,0.0651097,0.0859830752,-0.271407336,0.1166786849,-0.1478212923,-0.0716028661,-0.1981462091,0.1602655649,0.2868601382,-0.1430522949,0.1544095725,-0.2267408073,-0.2376000136,0.2086211145,0.0678298846,-0.2635498941,-0.0741982386,-0.4286122322,0.1351036429,0.2275343835,-0.0362660028,-0.1721635759,-0.2053703517,-0.5445809364,-0.1897035837,0.2408191413,0.1640048772,-0.6624581218,0.0249800254,-0.2985251248,0.3442323804,-0.0124302749,0.0346627347,-0.1729978621,0.5283489823,-0.1204582676,0.061828725,0.1422985196,-0.3676961362,-0.3364787996,-0.3080311716,0.3230183125,-0.0855855569,0.0160622932,-0.1044687927,0.1167748347,-0.085985899,0.2731749415,-0.2802529037,0.2296978384,0.0253234878,-0.159799099,-0.1368355453,-0.1091356799,-0.0260084569,0.1039797887,-0.0982104838,-0.0813206211,0.2296898812,-0.2626010478,-0.2891021073,-0.0163193699,0.1953529119,0.1827475727,-0.3408950567,0.1918164641,-0.0180603825,-0.0330677554,0.2095744759,-0.0214827321,0.3010982573,0.1420589536,-0.0496485345,-0.4447845519,-0.1568556875,-0.1669616997,-0.1412568241,0.0099147353,0.0559448265,0.1167961061,0.0565204956,-0.2903281748,-0.6320902705,-0.1908553392,-0.0367804207,-0.1114452854,0.1402647793,-0.0014442542,-0.1417408735,0.2513108552,0.4908497632,0.2035687566,0.1256274879,-0.132595107,0.3688109517,0.042504847,-0.0703443065,0.1904079914,0.2616536617,-0.0247750394,-0.3786281645,0.1851952374,0.0075344779,0.0189405419,-0.0573757254,0.0464042164,0.4574383199,0.1629690826,0.2236258686,-0.129406184,-0.1138168871,0.0345552675,-0.0860925838,0.0015101718,-0.4130072594,0.0983453169,0.0369585194,0.151216507,-0.0825689137,0.1815174371,-0.1523884982,0.6337943077,0.220052585,0.3320113122,0.2181577384,0.0381784253,-0.1850304604,-0.2790032029,-0.2729865909,0.5786850452,0.1237926036,-0.1061576605,0.3212693632,0.0038928762,-0.1291131377,0.141755417,0.2751808465,-0.1621743888,-0.0304770134,0.210294053,0.0498226583,-0.0208723061,-0.0048262347,-0.0404571593,0.2664393485,-0.4259026945,0.1111001745,-0.3015710413,0.1249610856,0.0289600473,-0.504755497,-0.0801235512,-0.2289238721,0.0683135614,-0.0688522607,0.1172926798,0.5084078312,0.1089787558,0.7309294939,0.2221571952,-0.0884833783,-0.0418619104,-0.3706047237,-0.0938981026,0.0051324926,-0.1675924212,-0.1081163287,0.248287335,-0.1957300305,-0.1817724109,0.2134941071,-0.3316074312,0.0054062358,-0.0847581327,0.291046828,0.5130856037,-0.0982805565,-0.1604834944,0.1313720345,0.2841677666,-0.0001518991,0.2412525713,0.0516485646,-0.2282531559,0.0180473197,-0.0589938499,-0.272904247,0.319388777,-0.155060336,-0.142799899,0.0349994786,-0.0177338552,-0.0386879072,0.370239675,0.3034929037,0.1188180223,-0.2300212979,-0.1194092333,-0.6648519039,0.4553357661,-0.2929923832,-0.1087906286,0.1177509949,-0.2278100997,0.5461059213,0.0821604133,-0.1991848797,-0.5965135694,0.0502404161,0.250220567,-0.0362006277,0.097188592,0.1329384148,0.0957233831,0.0078016487,-0.1460781842,-0.212894991,0.184142828,-0.3416827917,0.1926825941,0.1651636362,0.2795262635,0.1398905367,0.9020349383,0.2709238827,-0.281452775,0.4462845922,-0.0629040226,0.3629128337,-0.3857726455,-0.3537685275,-0.1327574998,0.2067406029,0.0979138166,0.2283212692,-0.0798870027,0.3991346955,-0.0931990519,-0.2578041553,-0.1166767254,-0.3271866441,-0.0445062555,-0.0414846539,0.5182970166,0.0314048901,0.2818897665,-0.3684830964,-0.2209340185,-0.1256369203,-0.1468351334,0.0447547473,-0.1376474202,0.2064362615,0.3601918519,-0.4478794336,0.4855196774,0.0488484949,0.0957093835,0.0371524394,0.4254160225,-0.1135928854,0.0117070395,0.1905833334,-0.4265570343,0.0347687937,0.006311324,0.1297144443,-0.4860115647,0.1416696608,-0.0598239601,0.0118681639,0.2834944725,0.6065133214,-0.1750784516,0.0454032421,0.2988995016,-0.1233688891,-0.0613542087,-0.3018509746,-0.3138075471,-0.2025614083,-0.0685620531,-0.0641749054,0.0039455104,0.2559005618,0.0143890912,0.2161925882,0.1685958803,-0.3550197184,0.1805219799,0.2913928032,0.155244112,-0.1508333087,0.071813345,0.190815255,-0.6225715876,0.5667641163,0.2896650732,-0.4220636785,-0.6125307083,-0.0960294902,-0.0508674532,0.3875898719,0.143642053,-0.2398140877,0.0524121113,-0.2107733041,0.1070803106,-0.1609295905,0.3424239755,0.4375675917,-0.1072244346,-0.1995589882,-0.5116867423,0.4413838983,0.094638966,0.0334274545,0.0409586467,0.2835838497,-0.3390710652,0.2662268281,0.057126414,0.5559182763,0.3571514785,0.0288760681,0.2227933705,0.2120592296,0.2918720245,0.286837846,0.0456911884,-0.1450481415,0.0168374367,-0.0371734723,-0.1158001348,0.1779960394,0.1113035455,-0.2892237902,0.6147172451,-0.2025346607,-0.2017870247,0.0118796965,-0.0943582281,0.1071691662,-0.2357636541,-0.0885599107,0.0887356177,0.0230943859,-0.111722514,-0.039442569,0.0824054405,0.1448679566,-0.2616629004,-0.2599175572,-0.0353086628,0.0371974967,-0.1349534392,0.0810837746,0.0181644503,0.3237316012,0.1488489658,0.653203845,0.2418645322,-0.207136035,0.1182952896,-0.1193675324,0.0064764498,0.0263738483,0.0393317789,-0.1799152344,-0.0715444311,-0.2894842923,0.1836419255,0.1256276816,-0.2663265467,-0.3684746623,-0.3204212487,0.0713494793,-0.1727733463,0.1073026136,-0.1944182068,0.0886707082,-0.2228709757,0.0771448687,-0.150942117,0.033999186,0.1720388234,-0.3711317778,-0.2299392223,0.0759704337,0.3704081178,-0.2015377134,0.1773388535,0.5261321068,-0.1230134368,0.0628305376,-0.0600657426,0.1064030528,0.0654751286,-0.1386009008,0.3454103172,0.2945848107,0.2183434665,0.1430258304,0.7127477527,0.0977629647,-0.20019117,0.2122234106,-0.1841594428,0.1099696979,-0.1727627665,0.057094641,0.0928730518,0.0101971226,0.17662853,-0.0503209494,0.0149428435,-0.1861610264,0.2273892015,-0.0684597492,-0.1279036701,-0.0565339439,-0.1750913113,0.2398561239,0.0384891368,-0.0406209938,0.3173969984,-0.2455488443,-0.1224795952,-0.411349684,0.1346178949,0.078740932,-0.1166668385,0.0218821745,-0.0185908396,-0.2206945419,-0.1853694767,-0.0701066703,0.4343701899,0.0213522874,0.4307079613,0.0698842555,0.2463932484,0.1116877347,-0.0364817157,0.1741282791,0.33659181,0.0569783412,0.3018293679,-0.072365202,-0.1096206978,-0.0391750783,0.0956133679,0.4132085741,-0.1316749305,0.1792759746,0.0403222777,0.0137238996,0.1524851471,0.3480333686,0.1198687926,-0.4086858928,-0.2935936451,-0.0599623397,0.0492643565,-0.4219584763,0.0167807713,0.0277254134,0.4374590814,0.0878485739,0.2638888359,0.1583175063,-0.1762225628,-0.1202416793,-0.0211931914,0.182082817,-0.2566046715,0.3037871122,0.0547501743,-0.1580067575,0.1335220933,0.5958923101,0.2054216415,0.2644688785,0.6123894453,-0.1803975701,0.5311422348,-0.3062996566,-0.1444173604,-0.1915825456,-0.242513746,0.2101691067,0.2746715844,0.1775108427,0.120166406,-0.1270647794,0.5288738012,0.0042936546,-0.0066058035,-0.4124710858,-0.0633299276,0.0470052958,-0.2205009609,-0.3566699028,-0.272878617,0.0021743996,-0.0159856826,-0.0539781302,0.0160163809,0.1668214053,-0.049576588,-0.0439200252,-0.0974004865,-0.122740306,0.1251526177,0.1903985739,0.0028790757,0.1741584688,0.3312364519,0.1200811416,0.1952888966,0.4346508086,0.2796159089,0.2170420438,-0.0819662213,-0.2309222221,-0.0824272409,0.007078704,0.1619892865,0.1596878022,0.2534364164,0.3066961765,0.2544464171,0.0441780053,-0.1342924982,0.5340619087,-0.178024143,0.6815495491,-0.4717414677,0.4967933595,0.0681216866,0.1819931418,0.2473811358,0.0696363077,-0.1978520006,0.1430838257,0.2249484658,0.1974614263,-0.1443538517,-0.3264790773,0.0344181843,-0.1260913908,0.4268822968,0.3506712914,-0.12943694,-0.6421672702,-0.037459936,-0.6559430361,0.019743016,-0.1101392359,0.2305752635,0.0502443872,-0.0739753172,-0.1047328636,0.4412332177,0.1627909392,-0.2250572443,-0.3156934977,0.0508506857,-0.4535089731,-0.0678714663,-0.2392272055,-0.4561624825,0.1878072172,-0.2697881162,-0.2112484127,-0.171210289,-0.0017721967,-0.0239220075,-0.1900016665,0.3145197332,0.3738414049,0.4422281981,0.1602003574,0.3671755195,0.077450797,0.0392942801,-0.0766876414,0.0218123365,-0.1268318444,-0.002756858,-0.040699251,0.0743444264,-0.420794934,-0.2326158136,-0.3469344079,0.1172550842,0.1225599125,-0.2500376403,-0.376244843,0.1451450586,0.1201128066,-0.1879861951,-0.09150967,0.1361382604,-0.1393854767,0.4888540804,-0.18744196,0.044963792,0.5023432374,-0.3556188047,-0.3082362115,-0.0916114002,0.2599363625,-0.1059316769,-0.316693157,-0.8160826564,0.1901779473,0.4983621538,0.1807351112,-0.0991933346,0.1552737802,-0.2834812105,0.1785763204,0.0969762653,0.2136149108,-0.0676537678,-0.2712716162,0.0590189435,-0.1713263243]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/843","title":"use_custom_baseline still produces errors for bertscore","comments":"Hello everyone,\r\n\r\nI think the problem is not solved: \r\n\r\n```\r\nfrom datasets import load_metric\r\nmetric=load_metric('bertscore')\r\nmetric.compute(\r\n predictions=predictions,\r\n references=references,\r\n lang='fr',\r\n rescale_with_baseline=True\r\n)\r\nTypeError: get_hash() missing 2 required positional arguments: 'use_custom_baseline' and 'use_fast_tokenizer'\r\n```\r\nThis code is produced using `Python 3.6.9 datasets==1.1.2 and bert_score==0.3.10`","body":"`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2","comment_length":42,"text":"use_custom_baseline still produces errors for bertscore\n`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2\nHello everyone,\r\n\r\nI think the problem is not solved: \r\n\r\n```\r\nfrom datasets import load_metric\r\nmetric=load_metric('bertscore')\r\nmetric.compute(\r\n predictions=predictions,\r\n references=references,\r\n lang='fr',\r\n rescale_with_baseline=True\r\n)\r\nTypeError: get_hash() missing 2 required positional arguments: 'use_custom_baseline' and 'use_fast_tokenizer'\r\n```\r\nThis code is produced using `Python 3.6.9 datasets==1.1.2 and bert_score==0.3.10`","embeddings":[0.0139922751,0.0406964272,0.0927704349,0.0760170147,0.1578616649,-0.0484319031,0.4595745802,0.0406530872,-0.0702814609,0.0988220498,0.057240773,0.248223722,-0.2429824471,-0.091258958,-0.3137350082,-0.2044166178,-0.1234866977,0.2484205514,0.1250584424,-0.0682218224,-0.6134792566,0.1149746105,-0.5267378688,-0.2245671898,-0.3061106801,-0.0073447996,-0.1297489554,0.0940015838,-0.1675110012,-0.4430608451,0.3566359282,0.2593745887,-0.0273150597,0.5043865442,-0.0001236519,0.0151397856,0.0650538281,-0.2212443501,0.1006062552,-0.2754280567,-0.3003541529,0.2056778818,-0.1089255288,-0.3084519506,-0.0567066446,-0.0974938422,-0.3291552365,-0.2222906947,0.2909580171,0.4247182906,0.0997443795,0.1678161025,-0.0701237246,-0.3143260777,0.0707339644,-0.3041083813,-0.132868439,0.4761042893,-0.3822268844,0.0905164182,-0.0475983918,0.0151075721,0.0084916111,0.275585413,0.4619062841,0.0610387363,0.1685637087,0.0290288627,0.1067181006,0.3407550752,-0.0626366436,-0.1964442432,-0.1896905005,0.0891402587,0.1015939116,-0.7081446648,0.0732918009,-0.2468524128,-0.1026242673,0.0150160482,-0.1463925391,-0.2310788929,-0.3840640485,0.2057585269,0.0964826643,-0.0224360973,-0.0936609283,-0.010377693,0.5541225672,-0.1331993788,0.2133925557,-0.2867097855,0.0761622712,0.0739566684,-0.2188491374,-0.0774903074,0.1112971604,-0.0142223453,0.0371265672,-0.4702761471,0.2807561159,0.0519439876,0.0180908851,0.0164009593,-0.0885795504,0.4158307016,0.0310362186,0.064934589,0.0825008005,0.2168201357,-0.0032398945,0.0564349405,0.0582802892,-0.3148384094,0.4590700567,0.2077643722,0.0701418519,-0.0728394613,-0.0429898798,-0.1028405949,-0.4027306736,-0.1958101392,0.1490715593,0.17960006,0.1914663017,0.3129924834,-0.1987798959,0.2753800154,-0.0125335623,-0.0601523481,-0.1233491302,0.1453828961,-0.38491714,0.2618605494,0.2926675975,-0.260150671,0.3255074918,0.30894804,0.3424107432,-0.0041429466,0.0734393448,-0.2821101844,0.1013380215,-0.1857380718,-0.0533379838,-0.2254092246,0.1736967713,0.2982823551,-0.1498628855,0.1796702445,-0.2692430317,-0.2601077855,0.2153955698,0.0630152002,-0.2491658181,-0.0688862652,-0.4689850807,0.1305702031,0.237238735,0.0045827674,-0.1681928933,-0.1690101027,-0.5841546655,-0.186730504,0.2247252613,0.1487060636,-0.6226087213,0.0103871049,-0.2680073082,0.3499619961,0.0292986743,0.0206672512,-0.1841612458,0.5567415357,-0.1364111602,0.0483741127,0.1785369813,-0.4073660076,-0.3432235122,-0.3021988869,0.3096629381,-0.1078904644,0.0489362925,-0.1057818085,0.118581593,-0.0633806363,0.2802256048,-0.2763059139,0.2693374157,0.049765192,-0.1741134077,-0.1504568309,-0.1033492908,-0.0554999784,0.1072779223,-0.0930515602,-0.0682105869,0.2482923865,-0.2434185147,-0.2698500454,-0.0465709418,0.2116320878,0.194063589,-0.3016175926,0.1966770291,0.0066732019,-0.0459150858,0.2049425393,-0.085631296,0.3094464839,0.1192712188,-0.0633344203,-0.4761702716,-0.1659834236,-0.1861745715,-0.1153101027,0.0121974889,0.0760244727,0.1235741228,0.0723399892,-0.255281955,-0.6563645005,-0.2443365902,-0.0664544851,-0.1146557406,0.1561340839,-0.0055224099,-0.1433352381,0.2601358891,0.460088104,0.2633832395,0.0992789343,-0.1747127175,0.35264045,0.0372519083,-0.0551031791,0.1486088187,0.2489783615,-0.0254417956,-0.4127577245,0.1762646288,0.0228236318,0.0518608727,-0.0410667434,0.113983281,0.4497624934,0.1684157401,0.2445272952,-0.1405307204,-0.13733311,0.0315355137,-0.0921100304,0.0284693372,-0.4044124484,0.1003224105,0.069021903,0.1901570112,-0.0592134222,0.2248275429,-0.1967794895,0.7020145655,0.2295797765,0.3146533966,0.2003013492,0.0612428933,-0.1737088114,-0.3064579368,-0.275839597,0.5699107647,0.1226788387,-0.1319181621,0.3084480464,0.0509160534,-0.1044948846,0.152882427,0.2951426506,-0.1105292141,0.0119593767,0.1853530556,0.0484520607,0.002255708,-0.0102937007,-0.0005142058,0.276298821,-0.4457083941,0.1012327224,-0.2893078327,0.1160955131,0.0153841209,-0.4470415413,-0.0500866547,-0.2175884396,0.0222832039,-0.1053767353,0.1690021008,0.5220665336,0.1414784938,0.7223024368,0.226603508,-0.1214421019,-0.0468008667,-0.3608556986,-0.0696485639,-0.0147932395,-0.1647163928,-0.0889836773,0.2091378272,-0.2166952789,-0.1947316378,0.2374855876,-0.3260768652,0.0331060514,-0.0415863246,0.2737569809,0.4932181835,-0.1144722924,-0.1439746171,0.1149252057,0.2683179975,-0.019764347,0.2143016309,0.0747958496,-0.2519958913,0.0253388956,-0.0508959368,-0.31077829,0.3045133352,-0.1827806085,-0.198859036,0.0107530486,0.0226688292,-0.0432986468,0.3582718372,0.3304134607,0.1229671165,-0.2094842196,-0.1147248447,-0.6629092693,0.438267529,-0.2981931269,-0.1130236089,0.1229114309,-0.254455626,0.5852690935,0.039028883,-0.203691557,-0.6184414625,0.0393268503,0.2566467226,-0.0809669569,0.0743938088,0.0919255465,0.0779964924,0.0086050136,-0.1452542841,-0.2045074403,0.222530663,-0.3108867705,0.1650232971,0.2436050326,0.3170932233,0.1539705098,0.9103382826,0.2706152499,-0.2894462049,0.4798105955,-0.0593984537,0.3143587708,-0.3554131687,-0.3691224754,-0.16725339,0.2391471416,0.1340104789,0.2440320402,-0.1102440134,0.433088094,-0.0725063607,-0.3233250678,-0.1608930528,-0.3306833506,-0.0512797534,-0.0570974536,0.5392923951,-0.0012505561,0.2285223752,-0.367079407,-0.2014560699,-0.0900196433,-0.1214061975,0.0307160076,-0.118936263,0.2083640248,0.4051836431,-0.4139282107,0.4988673925,0.0669777542,0.107929945,0.0215001274,0.4737037718,-0.0893367827,0.0336133316,0.1686109304,-0.3760922253,0.0861467198,0.0106591145,0.1086212844,-0.4380438328,0.1527949572,-0.0400891453,0.0207400694,0.3356541097,0.6264057755,-0.1839401871,0.032359343,0.2444843501,-0.1218855754,-0.1041042283,-0.3233880997,-0.3281171024,-0.2228057683,-0.0893908516,-0.0504238904,0.0312821083,0.3153294921,0.0235158205,0.1831658185,0.1488301307,-0.4003753662,0.1612762809,0.2591054142,0.1627115309,-0.0894839168,0.0577943437,0.1674821824,-0.6602146626,0.5150834322,0.3028129339,-0.4229430258,-0.6607756615,-0.120942615,-0.054575488,0.4059806466,0.0957078561,-0.2597759068,0.1247484684,-0.1759486049,0.0940787345,-0.0982142463,0.314596653,0.4231685102,-0.1008176059,-0.1693933606,-0.5441931486,0.4014719725,0.0640226975,0.0178908315,0.034311194,0.2680589855,-0.3483158052,0.2627154291,0.0803082511,0.5774639249,0.3167763054,-0.0194891822,0.2353925109,0.2079371065,0.3009339869,0.3254628479,0.0168542601,-0.1357762963,0.0258887466,-0.0251999423,-0.1251005679,0.1769699454,0.1260727346,-0.3157387376,0.6133509278,-0.2259572297,-0.2325965911,-0.0105406651,-0.0436598808,0.1495301574,-0.2235094607,-0.1270403415,0.0796343535,-0.0221545082,-0.0884341672,-0.0165913384,0.052665934,0.1608530283,-0.2936869562,-0.2373049706,-0.0411991924,0.0128375348,-0.1123118773,0.0519582592,0.0126609625,0.3148292899,0.1824474484,0.6069731712,0.2493256778,-0.1622727364,0.0922045633,-0.1178657338,-0.0127976285,0.0018348253,0.0610976219,-0.2082540542,-0.0953393802,-0.3335289061,0.2035053819,0.1401425898,-0.2949586213,-0.3115858734,-0.2978537679,0.1001946405,-0.1511565447,0.05816558,-0.1985634714,0.0589671768,-0.2253695428,0.0792949423,-0.152747795,-0.0013031301,0.1458332539,-0.3612082899,-0.2589081526,0.0928633139,0.3415715098,-0.1735960692,0.2014159113,0.5621205568,-0.1293609589,0.0718692914,-0.0616955459,0.1020109206,0.013544932,-0.1509646326,0.3395422399,0.3096989393,0.2640450895,0.1182970852,0.6846894026,0.1009590179,-0.1395335197,0.1983098388,-0.1839973629,0.0912184268,-0.2021648288,0.0457404368,0.083053194,0.0208785757,0.1670601964,-0.0462190099,0.0603764988,-0.183685407,0.2324300557,-0.1228204817,-0.1392485946,-0.0423926525,-0.1706882715,0.2101248652,0.0735417977,-0.0369080119,0.2965259254,-0.2218365818,-0.1250485927,-0.4624553323,0.1367423236,0.0681846589,-0.0925628245,0.0059866644,-0.038207747,-0.2127822042,-0.1412305087,-0.0843899846,0.4205981791,0.0277719051,0.4663683474,0.0805818513,0.2660737634,0.0592218228,-0.0173186343,0.1376077384,0.3475292921,0.0743340999,0.2801558673,-0.0333191343,-0.1043333113,-0.1088938639,0.0123448074,0.4550784826,-0.1400425732,0.1288688332,0.0833876655,0.0280177891,0.1437490284,0.2597427964,0.0604500212,-0.4357894063,-0.2595280111,-0.0562424883,0.0386524126,-0.4195451438,-0.0001196319,-0.0074307648,0.3897435963,0.086579144,0.2663342655,0.1477300227,-0.1425555348,-0.1165732741,-0.0469071493,0.1650988907,-0.2278248668,0.3007052243,0.0460492373,-0.1321206987,0.1493014693,0.5626954436,0.2178757638,0.2259547114,0.5427420735,-0.203547731,0.5215120912,-0.3308874071,-0.1211791113,-0.1570494026,-0.2285763919,0.1866047382,0.2842144966,0.1555813104,0.1448968947,-0.1207382083,0.5115415454,0.040925961,0.0316457488,-0.3718785644,-0.0203670971,0.0784443244,-0.2142840922,-0.3626483083,-0.2173562795,-0.0206147265,-0.033714164,-0.0153516466,0.0180398896,0.1275290847,0.0207724236,-0.0563838072,-0.0406958908,-0.1477530748,0.18017295,0.209284246,-0.0137684001,0.1770896912,0.326333046,0.1509558409,0.1939722598,0.4427273571,0.229570955,0.2412535399,-0.0601646863,-0.2420249879,-0.0385427363,0.024976721,0.1657063812,0.1447427422,0.2192070633,0.3229822218,0.2372628599,0.0401301868,-0.1470800489,0.508295536,-0.1372559965,0.6505014896,-0.4317146838,0.5439696908,0.0632281527,0.1906373501,0.2060276717,0.1321190745,-0.2425420433,0.1380615383,0.2112990618,0.2143058926,-0.1341810822,-0.3158295453,0.0322323255,-0.15167436,0.4194208384,0.3295184374,-0.0575855114,-0.6801505089,-0.0239645168,-0.6489636302,0.0207918249,-0.0896082744,0.1807385683,0.0731624663,-0.093354322,-0.1565409899,0.4147487283,0.1558232754,-0.2584223747,-0.2880349457,0.0568971559,-0.4772015214,-0.0626385957,-0.2028372586,-0.4374268651,0.1991520822,-0.285133481,-0.2248172313,-0.1306184083,-0.0007863853,-0.0512953922,-0.199453041,0.3091574311,0.4073768556,0.4424705207,0.1255680919,0.3649844229,0.034182623,0.054356724,-0.0113631291,0.0168516319,-0.1058610156,-0.0111449528,-0.0192588046,0.046386186,-0.417824924,-0.2572529316,-0.3420444131,0.10455966,0.0824924186,-0.3039464355,-0.3829138279,0.1456281245,0.0989883468,-0.1760291308,-0.086481303,0.1349320263,-0.1208005697,0.4523110986,-0.205382362,0.0930544958,0.5157268047,-0.3611714244,-0.3002435863,-0.0522417314,0.3067159653,-0.1005391032,-0.3011502028,-0.7532256842,0.1969093382,0.5441911221,0.1701501757,-0.0642267838,0.1828490198,-0.2694274485,0.2287347913,0.0970768183,0.1186893955,-0.083960712,-0.2456703335,0.0267213192,-0.1984726042]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/843","title":"use_custom_baseline still produces errors for bertscore","comments":"Hi ! This has been fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/2770, we'll do a new release soon to make the fix available :)\r\n\r\nIn the meantime please use an older version of `bert_score`","body":"`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2","comment_length":30,"text":"use_custom_baseline still produces errors for bertscore\n`metric = load_metric('bertscore')`\r\n`a1 = \"random sentences\"`\r\n`b1 = \"random sentences\"`\r\n`metric.compute(predictions = [a1], references = [b1], lang = 'en')`\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\n File \"\/home\/stephen_chan\/.cache\/huggingface\/modules\/datasets_modules\/metrics\/bertscore\/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363\/bertscore.py\", line 108, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`\r\n\r\nAdding 'use_custom_baseline = False' as an argument produces this error\r\n\r\n`Traceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/stephen_chan\/.local\/lib\/python3.6\/site-packages\/datasets\/metric.py\", line 393, in compute\r\n output = self._compute(predictions=predictions, references=references, **kwargs)\r\nTypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`\r\n\r\nThis is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2\nHi ! This has been fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/2770, we'll do a new release soon to make the fix available :)\r\n\r\nIn the meantime please use an older version of `bert_score`","embeddings":[-0.065399088,0.1055707484,0.1252271235,0.0791326985,0.1758203357,-0.1042810902,0.4214691222,0.0186607651,-0.0667726249,0.118460685,0.030669719,0.2369848192,-0.1979214549,0.030478498,-0.3029206395,-0.2030736357,-0.1065387726,0.2290034145,0.0602318011,-0.0018443018,-0.5456944108,0.1240426227,-0.4763626754,-0.1773430258,-0.3232955635,0.0166038554,-0.0931779444,0.0963591561,-0.1938426793,-0.4644752443,0.3707922101,0.2937927246,-0.0783435032,0.4273879826,-0.0001202581,-0.0173141062,0.0906589329,-0.1919967681,0.0823990405,-0.3026836216,-0.3079280853,0.2507736683,-0.0756759867,-0.1881365031,-0.0568038225,-0.0254872702,-0.277130127,-0.2293009609,0.2610580921,0.3748821318,0.1189197302,0.248761192,-0.0141959656,-0.3966115117,0.1391546875,-0.285646826,-0.1170233935,0.4653153718,-0.4163216352,0.0323894061,-0.1525924653,0.0655753314,0.0691507012,0.2131879628,0.4943492413,0.1090632901,0.0412167087,0.0574481115,0.1284718812,0.3740675449,-0.050068941,-0.2580922544,-0.268222183,-0.0349042676,0.1512615681,-0.6960839033,0.08812695,-0.298895061,-0.0544674546,0.0102438703,-0.213337943,-0.1494624466,-0.3620051742,0.1846598238,0.0540706404,-0.1190564036,-0.1737605929,-0.0061601447,0.6393912435,-0.1262257397,0.0312464442,-0.160253346,-0.0081798183,0.0703401491,-0.2226918787,-0.1142823473,0.1105880961,-0.0201596767,0.0919506475,-0.360073328,0.2725138068,0.030992236,-0.0768296495,0.0010996127,0.0304140672,0.4342128336,0.0872670263,-0.0564139113,0.1268499494,0.2700208724,0.1153942868,0.0258651953,0.0955619141,-0.2423228472,0.3633945584,0.1154833809,0.0006458133,-0.1070115343,-0.1145095229,-0.048915565,-0.2870645225,-0.1943377703,0.1714108884,0.2480600476,0.1765980273,0.2997084558,-0.2521529198,0.2725555301,0.0030375775,-0.0681247488,-0.1418930292,0.0871898234,-0.3522244394,0.2969470918,0.3173727393,-0.2396328151,0.355274111,0.2791146636,0.3951649666,0.0837464109,0.0532858297,-0.2689545155,0.1519267857,-0.1623131782,-0.0842929482,-0.2084585726,0.1395534575,0.2767295539,-0.1474154145,0.1042362824,-0.2027218491,-0.2549835443,0.2148388773,0.091043137,-0.2995553613,-0.0900899097,-0.4372721612,0.1694345325,0.1358772367,-0.094205454,-0.1295930594,-0.1712069809,-0.5282446146,-0.1445283443,0.3029254973,0.2183142304,-0.5863320231,-0.0307365377,-0.2681738734,0.2509887815,-0.0577307492,0.0461670123,-0.1169253811,0.4635560811,-0.0958911777,0.09203621,0.0361051597,-0.3284228742,-0.3902496099,-0.3276280165,0.3103570938,-0.0577141792,0.0274140816,-0.09317334,0.031585183,-0.0997008607,0.2024668157,-0.2675183713,0.1949836761,0.0600165091,-0.1490323842,-0.1651210934,-0.1675140411,-0.0138966935,0.0807847232,-0.1848622262,-0.0777184218,0.3040081561,-0.2377139926,-0.2887169421,-0.0250663795,0.1667808443,0.1623036414,-0.295412451,0.178111583,-0.0408695564,-0.0651674643,0.2309506088,-0.084703967,0.2836992443,0.1728590578,-0.0925022289,-0.4605918825,-0.1143953279,-0.1628159136,-0.1544780731,0.0337178856,0.0203069765,0.1430742592,0.1113379225,-0.2873969078,-0.6143479347,-0.2077045888,-0.055125121,-0.1873973012,0.0940355137,0.0287742354,-0.1254122704,0.2686022818,0.5271698833,0.1975507736,0.1344726682,-0.1280670911,0.4088070095,0.0464214347,-0.0431450307,0.1315714419,0.2197876722,0.0151624121,-0.344604671,0.0824566633,0.0367514528,0.0169554148,0.0046336357,-0.0471122898,0.4674701989,0.0854474008,0.2155937552,-0.0656195581,-0.1301858872,0.0143371029,-0.0257110745,-0.0032305294,-0.4469573796,0.1062192619,-0.0156325996,0.1746184677,-0.1073334962,0.1675826609,-0.1494366825,0.5737681985,0.1685099155,0.3206823766,0.2710583806,0.0261102356,-0.203624621,-0.2102162242,-0.3017736971,0.5065487027,0.1265647411,-0.1132989973,0.2986457944,0.0085949944,-0.1254892051,0.1335160434,0.2480135262,-0.1131374687,-0.0160017945,0.2157008797,0.0475166589,-0.0509593226,0.0090919938,-0.1127000153,0.253324002,-0.4297865331,0.1104360744,-0.3132552803,0.0675640702,-0.002046505,-0.4873185754,-0.1534600854,-0.2266827524,0.1109821647,-0.0218976811,0.0506076179,0.4909879565,0.1721702218,0.6954312325,0.1758468151,-0.0201313645,-0.0392874293,-0.4039442241,-0.0806418955,0.0235443413,-0.1637841314,-0.1326147914,0.2936668694,-0.2023090869,-0.2084415406,0.1652985364,-0.3238097131,0.0610867515,-0.0988676026,0.3026757836,0.5101534724,-0.1756213307,-0.248159796,0.1170610934,0.2749007046,-0.0380744934,0.2095594704,0.0698156431,-0.2160654664,0.066987589,-0.0378154255,-0.2673087418,0.3557494581,-0.125671044,0.0374212675,0.0129873194,0.0083139688,0.034192998,0.3668732345,0.3727366924,0.0499798581,-0.1662344933,-0.176174894,-0.6425845027,0.3792101741,-0.3096800745,-0.1714020222,0.0285001304,-0.1941157877,0.4604353607,0.0138922427,-0.1628634483,-0.6429674625,0.0338853747,0.2201700956,-0.0202882253,0.1086328477,0.1279480308,0.0701922327,-0.0310323592,-0.1363003552,-0.1708536446,0.2175342739,-0.4094260633,0.1471411288,0.1625185311,0.2354807109,0.1228960231,0.851375699,0.2858025134,-0.1723722219,0.4137632549,-0.0692917556,0.4069245756,-0.4315741658,-0.3084430397,-0.0779511333,0.1519485116,0.1701080054,0.2565909028,-0.0720668882,0.4978993535,-0.0773520619,-0.2730075121,-0.0725355893,-0.347966224,-0.0067151855,0.0119993147,0.462151885,0.0531114042,0.3101488352,-0.3184792697,-0.2585998774,-0.0800905526,-0.192386806,0.0593601689,-0.1621022224,0.2538943887,0.3043074012,-0.5376888514,0.4907963872,0.1324540973,0.0841787383,0.0115160551,0.3476767242,-0.1114816666,-0.0016953881,0.161225155,-0.3205763102,0.051981315,0.0078486279,0.0893843025,-0.4751473069,0.19583413,-0.0915481597,-0.0450603291,0.2846351266,0.6144083738,-0.2590414286,0.0209020022,0.3384500444,-0.0882209465,-0.0520394072,-0.289038837,-0.3144123852,-0.1584334522,-0.0857423618,-0.0695382059,0.0334752128,0.2594649494,0.0098926267,0.2843878567,0.1427084953,-0.3517608345,0.1875328571,0.3193560541,0.1910389513,-0.0949842036,0.0976454392,0.23997733,-0.6102179885,0.5049683452,0.3000902534,-0.4140455425,-0.6082895398,-0.0720816702,-0.02378067,0.3444721699,0.1610903442,-0.2345910668,0.0968234465,-0.1430865824,0.1310018897,-0.2406667918,0.3252991438,0.4248114526,-0.1367213577,-0.1481276602,-0.422729522,0.5183262229,0.1018850282,0.0071887132,-0.0038972024,0.3827879429,-0.3182097971,0.2791959941,0.0846856311,0.4901058078,0.2753661275,0.0245344155,0.290043503,0.1830847114,0.3743003309,0.3055674434,0.0335659198,-0.1801935285,-0.0225130245,0.036251802,-0.0294745415,0.1254560351,0.0717088282,-0.2818481028,0.6017307043,-0.1307419688,-0.1639166027,-0.0053627179,-0.0650169924,0.075887084,-0.3066626191,-0.1337852776,0.1088937819,0.0349916741,-0.1130534261,-0.0032097474,0.1066663414,0.1519921273,-0.287434876,-0.2890518308,-0.021241853,0.070057869,-0.1893222779,0.1021169797,-0.0341689587,0.3054367304,0.1074922606,0.6377574205,0.2460401058,-0.2074770778,0.1262596995,-0.1689180285,0.0338990316,0.0147839934,0.0716975406,-0.1978403926,-0.0963170156,-0.2559721172,0.1424469352,0.1260792762,-0.1926216185,-0.3124896586,-0.3262650967,0.0765812844,-0.14884606,0.137048468,-0.2489936799,0.1425117105,-0.2460162789,0.0948802978,-0.1482515037,0.0463664159,0.0553110167,-0.33520329,-0.2087284625,0.0580328479,0.4250167608,-0.2554201186,0.1469242722,0.4983087182,-0.0873437896,0.0471854806,-0.0636236072,0.1186024845,0.1683365703,-0.2462573647,0.2997986972,0.3104067147,0.152751416,0.0915719047,0.7438477278,0.1266752332,-0.2194368094,0.2081769854,-0.1710689813,0.0963266194,-0.1248525754,0.1101642251,0.0800838843,0.0851042867,0.1306932271,-0.1125175059,0.0472443439,-0.2121935487,0.2070274949,-0.1134936139,-0.1501967907,-0.0465591885,-0.1763799787,0.2995878458,0.0289711263,-0.003154288,0.3445771039,-0.2078132331,-0.1346688122,-0.439357698,0.1065452322,0.0074745268,-0.094700031,-0.0095291687,0.0501481816,-0.223576054,-0.2341237962,-0.0193774067,0.4133852124,0.0612044111,0.4334385097,0.0264657084,0.2277649194,0.1173272133,-0.0473477207,0.2234375626,0.3463303149,-0.0125476383,0.2922861874,-0.1184536219,-0.11452315,0.0115731889,0.115279004,0.4361112118,-0.1534221172,0.1395267248,0.0549851209,-0.0531127043,0.1615715921,0.3761327565,0.0671465024,-0.3552521169,-0.3730352223,-0.0390151627,0.0675002486,-0.412142545,0.0141472677,-0.0619879402,0.3827046156,0.0940744653,0.2849509418,0.1263643205,-0.1524021775,-0.0832103491,-0.0063977907,0.1320806146,-0.1703872383,0.2363165319,0.0356845409,-0.1092456356,0.101977706,0.5675435066,0.1736680418,0.3719374239,0.4825721979,-0.2171816081,0.4797876179,-0.4271298349,-0.0366534814,-0.1633065641,-0.2361468375,0.1992263347,0.2894110382,0.1696046591,0.094002448,-0.1354033053,0.6976354122,-0.0564677753,-0.0413106792,-0.399312824,-0.0470608026,0.0599515736,-0.2644565701,-0.3217134178,-0.2762301862,-0.0094794733,-0.0000548496,-0.0773511678,0.0037153675,0.168800354,-0.0301580057,-0.0176368672,-0.1436504275,-0.196263209,0.1265273094,0.2020414621,0.0204988271,0.1888605207,0.369397521,0.1587786973,0.2183806002,0.4889178574,0.2557505369,0.1460046321,-0.024648359,-0.1974700242,-0.120705612,0.0464288183,0.1615689695,0.1537140161,0.2954897881,0.2286320925,0.3217760921,0.0555734783,-0.14090918,0.4595594704,-0.1289280802,0.6855304241,-0.4920084178,0.4960341752,0.020603247,0.2143351436,0.2529295385,0.024840612,-0.2106886357,0.255523771,0.1016033143,0.2023917139,-0.1319218725,-0.2983578146,0.0570166856,-0.0494170152,0.4805620313,0.2433172762,-0.1569368243,-0.6128852963,-0.0559757873,-0.7017802596,0.0165389534,-0.0322359242,0.2955349088,0.0617189966,-0.1260977387,-0.1343746632,0.4902203679,0.1386624873,-0.2364033163,-0.3039498031,0.0087614991,-0.3811794817,-0.060671974,-0.1999618262,-0.4957592189,0.1941894591,-0.2205463201,-0.1901516914,-0.1369559616,-0.0039148517,-0.0027140782,-0.1556366682,0.1915475726,0.308513701,0.4104911983,0.1946674734,0.3023005724,0.0211828612,0.0073801922,-0.0617012121,-0.0012756417,-0.1275002062,-0.0074593914,-0.0871501043,0.0897248387,-0.3511870205,-0.2627788782,-0.3653207421,0.1832738221,0.1086656079,-0.2378021479,-0.4509291649,0.1449363977,0.1193899214,-0.2026013285,-0.1660802364,0.1367261261,-0.1208843961,0.5144050121,-0.1965547651,0.0217105653,0.5324780345,-0.4166596234,-0.2275541276,-0.0542858429,0.254973352,-0.0397064537,-0.3813864887,-0.8497495651,0.2573449016,0.4507564008,0.2204394192,-0.1265248805,0.0851902664,-0.3977725804,0.1467456222,0.0594753772,0.3238841891,0.0100373309,-0.2801128626,0.1192414761,-0.1822182238]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/842","title":"How to enable `.map()` pre-processing pipelines to support multi-node parallelism?","comments":"Right now multiprocessing only runs on single node.\r\n\r\nHowever it's probably possible to extend it to support multi nodes. Indeed we're using the `multiprocess` library from the `pathos` project to do multiprocessing in `datasets`, and `pathos` is made to support parallelism on several nodes. More info about pathos [on the pathos repo](https:\/\/github.com\/uqfoundation\/pathos).\r\n\r\nIf you're familiar with pathos or if you want to give it a try, it could be a nice addition to the library :)","body":"Hi,\r\n\r\nCurrently, multiprocessing can be enabled for the `.map()` stages on a single node. However, in the case of multi-node training, (since more than one node would be available) I'm wondering if it's possible to extend the parallel processing among nodes, instead of only 1 node running the `.map()` while the other node is waiting for it to finish?\r\n\r\nThanks!","comment_length":76,"text":"How to enable `.map()` pre-processing pipelines to support multi-node parallelism?\nHi,\r\n\r\nCurrently, multiprocessing can be enabled for the `.map()` stages on a single node. However, in the case of multi-node training, (since more than one node would be available) I'm wondering if it's possible to extend the parallel processing among nodes, instead of only 1 node running the `.map()` while the other node is waiting for it to finish?\r\n\r\nThanks!\nRight now multiprocessing only runs on single node.\r\n\r\nHowever it's probably possible to extend it to support multi nodes. Indeed we're using the `multiprocess` library from the `pathos` project to do multiprocessing in `datasets`, and `pathos` is made to support parallelism on several nodes. More info about pathos [on the pathos repo](https:\/\/github.com\/uqfoundation\/pathos).\r\n\r\nIf you're familiar with pathos or if you want to give it a try, it could be a nice addition to the library :)","embeddings":[-0.3369057477,-0.31968382,-0.1686351299,-0.0785499811,-0.1322283894,0.0334650092,0.0507945754,-0.0447819233,0.23892124,0.2358291,0.254429847,0.5422918797,-0.157852158,0.322792232,0.0088176709,-0.4016127288,0.0359969996,-0.0369444191,-0.0063834358,0.2683166265,-0.0824367628,0.0330719277,-0.3741092682,0.0420585163,-0.0999500751,-0.1551183015,-0.1367379725,-0.0298173353,-0.0436844453,-0.2958486378,-0.2203299701,0.3011392653,0.0312907211,0.3520682752,-0.0001225286,0.0020044774,0.2021490186,0.1364621222,0.2638634145,-0.2030464858,-0.2922287583,-0.1116604879,0.207213819,-0.141263485,-0.0572501421,0.1294005662,0.2296898514,-0.4878687263,0.0606242493,-0.1914961189,0.0673572645,0.3336282372,0.0711796507,0.2239813656,-0.5225394368,0.1676170975,-0.0575988702,-0.1018259302,0.6528367996,-0.3768672943,-0.1162374765,-0.0478885807,0.0505254529,0.1481447369,0.0178733636,-0.2197743803,0.5481138229,-0.6776022911,-0.136391893,0.1981161386,-0.0967679918,0.0911058486,-0.2966392934,-0.0853985548,-0.3229728639,-0.2332477868,-0.1775843799,-0.0001219448,0.0031385261,-0.1174954176,-0.5662781596,-0.4102555811,0.0705289096,0.2957291305,0.1834556013,0.6211115718,0.2272422761,0.2678335011,0.3702197969,0.1326148659,0.0390931107,0.0203601159,0.3610491753,0.2500997484,-0.3338703215,0.0637047216,0.0589844957,-0.1692203581,-0.2133903801,-0.2581843734,0.2163925916,0.2993526757,-0.2664533854,0.4457849562,0.1577317864,-0.1810770035,0.087563701,0.361533612,0.1224431545,-0.2412858903,-0.0981195047,0.0740630776,-0.036469657,-0.4829166234,-0.4006948173,0.3822456598,-0.3058327436,0.3852480948,0.3548592031,-0.3782683611,-0.1410485357,-0.2402350605,0.3023352325,-0.21192348,0.44224751,0.3982145786,0.063173674,0.2084145248,-0.1993974596,-0.2483129799,0.0638077632,-0.0258909427,-0.1408359259,0.1586971581,-0.0259823594,0.0150852073,0.0209516305,-0.1133035198,0.1074199229,-0.0729566813,0.4114916921,0.0906192586,0.0863064602,0.3343430459,0.0364167951,-0.0127970399,0.0399881005,0.6484883428,-0.1837859601,0.2720239758,-0.459177047,-0.3027990162,0.2662397325,-0.0613082014,0.2564393878,0.0969900116,-0.3469530642,0.1863732338,0.3003367782,-0.2851771116,0.269323945,-0.1820218712,-0.6244847775,-0.1700150669,0.3108026981,0.020038614,0.1818705648,-0.023772506,-0.1502303481,-0.0657917038,0.2736561894,0.0446197055,-0.2955271006,0.1507144123,0.1166545451,0.3167734146,0.065670602,-0.3526523113,0.4236588776,0.3865174949,-0.3203398287,0.0235287379,-0.2047762722,0.0141240684,-0.1535019875,0.1601706147,-0.0434379615,0.2964762747,-0.0211057644,0.0839059427,-0.1153804138,-0.2237194777,0.1503989398,0.0188092142,0.229651615,0.1636524647,-0.0292276144,-0.4754525423,0.3516525626,-0.0050842268,0.3887659907,0.0353224427,-0.2026080489,0.1578772813,-0.15698798,-0.2968689799,0.2742410898,0.2267602682,-0.3845553398,-0.2667267025,0.1944010258,-0.0281353705,0.658357203,-0.0203725323,0.0132354638,-0.0133071709,0.0218380522,-0.1558619142,-0.1386217475,-0.2770654559,-0.1314100325,0.1661051363,0.1653211266,-0.0123661188,-0.1140139401,0.2490856946,0.0076229563,0.1042492613,-0.3774382472,0.0114934901,0.0409500487,-0.0934703946,0.0581272058,0.2879985571,0.139239341,0.3358106017,0.2157691866,0.3359017372,0.3863933086,-0.1082795709,-0.0061820247,-0.1044881791,-0.1360841393,-0.252900213,0.466763556,0.2141513228,-0.0421636961,0.263661176,0.0598158538,0.1984933019,-0.0791036189,0.1946500689,-0.1260084361,0.159426868,0.1235274523,0.0014310838,-0.1964187473,-0.1850953996,-0.2523138821,-0.0365091898,-0.1916247606,0.0773274526,-0.1236756593,-0.1794439703,0.0905999169,0.1549218446,0.2646670043,-0.2878614068,0.2626061738,-0.108295761,0.3512683213,0.080504477,0.0600422546,-0.0556469597,-0.0219643414,0.1413629204,0.4599143267,-0.1237834543,0.0615026392,-0.0394493081,0.1551724821,-0.1666533202,0.3832467794,-0.1452605277,-0.08250238,0.3208311796,-0.2554083169,0.0117312223,0.0800678134,-0.4001930654,0.0283896513,0.0566958003,0.0959389582,0.2010514438,-0.1413591504,0.2351785004,0.2716403008,0.1544082314,-0.1668892801,-0.2972889841,0.1566224247,-0.0717362985,0.3108200133,-0.0209002178,0.0974181816,0.0560167171,-0.1543952078,0.3162545562,-0.172154665,-0.1625145078,-0.1071026102,-0.1085598096,0.1113179922,-0.1270810813,-0.0457958728,-0.1687175333,0.0315191001,-0.1978416145,0.1965869218,-0.2807557881,-0.0775249004,-0.1319009513,0.0424582027,-0.3148796856,-0.289426744,-0.0161795262,-0.3274700642,-0.2385780662,0.5306853652,-0.113502562,0.2678406537,-0.3076747954,0.0607952587,-0.2777210772,-0.2799748182,-0.2050078809,0.306293577,-0.4266135097,0.0779460371,-0.2861332297,0.0420378782,0.0888571739,0.1036917716,0.3386378884,0.422760278,0.1103589237,0.2011974454,0.1311051697,0.1691947728,0.0740518644,0.0106884902,0.4992435575,0.0048932615,0.1549522281,0.0984938592,0.1453436613,0.2018730789,0.3149352968,-0.0234792847,0.3149955571,-0.0544724055,0.0026231229,0.6437373757,0.1863555312,-0.0420929976,0.1919645518,0.0683293268,-0.3384121358,0.1102453917,-0.1958668083,-0.0822985396,-0.2924408615,-0.2481647581,0.0276500303,-0.0030350438,-0.4998773336,-0.292126745,0.0563718937,-0.3272611797,0.0083568525,-0.1388545632,-0.2200656682,0.7672846913,-0.2289755493,-0.348597765,-0.2488140017,0.3278869987,-0.2051020414,0.1145918369,0.4850689173,-0.1215671003,-0.2425573319,-0.2685330212,-0.1603889912,0.1594322324,0.2481433898,0.2303038985,-0.144482255,0.0974029452,0.2635603249,-0.085503079,0.3454560935,-0.4574623704,-0.2319710553,-0.024531452,-0.5518503189,0.1544320136,-0.0578177497,0.1378304511,0.4147225916,0.4971220493,0.0621995814,0.1970852911,-0.0779509023,-0.2383117974,-0.3341007233,-0.0130303446,-0.2394776344,-0.1139989048,0.1875666082,0.0187088586,0.1372369975,0.0922420472,0.0444072597,-0.0825210661,-0.0536532067,0.2551801503,0.0913646072,-0.1172084585,0.0785811245,-0.2555147111,-0.1166598722,-0.0385083035,0.3323026299,-0.2792900503,0.2440059185,-0.0023078432,-0.2696370184,0.1186039001,0.0186826251,0.0575738288,0.5479559302,-0.0063478653,0.0025869859,0.2468775362,0.3662625253,-0.0753516406,-0.4279352129,-0.1515331715,0.2849560082,0.1896314621,-0.080733791,-0.1994790137,0.2332254052,0.3173645437,-0.1335046738,0.5590898395,-0.7157691717,-0.151461482,-0.0580990538,0.1186355278,0.4146730602,-0.3008106649,-0.1821462661,-0.5093060136,-0.0542032458,0.0405051075,-0.6150126457,0.4420182705,-0.0092679309,-0.1447686404,-0.1018301696,-0.0206127055,0.5045963526,0.572945416,-0.0295715518,0.1001844779,0.3352402151,-0.4507008493,-0.1239355281,0.0944952071,0.554297626,-0.1666269898,0.5189483166,-0.0140087325,0.1069434658,0.1352590173,-0.0473027714,-0.1290603429,0.1321744621,0.070765391,-0.2802150249,-0.2274742723,-0.2805252969,0.0318155847,0.0266726874,0.2813259959,0.4023773372,0.417666018,-0.1105033532,-0.0568218492,-0.1543670595,-0.171269536,0.2499729842,-0.1161378026,-0.4350855052,-0.1170087308,-0.208559379,-0.0701491535,-0.3550237715,-0.0110196816,-0.0848327205,-0.5259230733,-0.2172741294,0.4013301134,0.1704325974,0.5439360142,0.0414080992,0.3589699268,-0.0272845812,0.0484286286,0.0243726689,0.2353884578,-0.3888373375,0.4202180505,-0.4228858054,-0.1100676358,0.1594162583,0.2093492448,0.0689661503,0.0010660177,-0.3470276296,0.4052503109,-0.1667878628,-0.0888366699,-0.3233777583,-0.129602626,0.6013835669,0.0267301016,0.3127380013,-0.0706743374,0.1087552831,-0.1204054132,-0.1222475916,0.3277972639,-0.269164592,-0.20201011,-0.1639289707,0.0102074258,-0.4468435049,0.038955681,0.2557687461,0.1791652292,0.3046794534,0.5270444155,-0.1095570922,0.097232081,0.1379773766,0.1390132755,0.2224263102,0.0652060807,-0.2121952325,0.0151992934,-0.0929197073,0.0756125152,-0.0860999823,-0.0377708189,0.2506555915,0.1767220795,0.1925657392,-0.1069147959,0.3903740048,-0.1686281711,-0.2094806433,-0.1731235087,-0.0312537588,0.0309492461,-0.1018214449,0.2018516362,0.2333864719,-0.0485743172,-0.3915390074,0.3528502285,0.1057595611,-0.1072908416,-0.1105403677,0.1556724757,0.2883764207,-0.1442365348,-0.0850943103,-0.2678103149,0.200562492,-0.0508220904,0.3649409711,0.150748834,-0.0474670529,-0.4286777675,0.3698693514,0.1183608398,0.0610660687,0.3107891083,-0.0347310789,0.002482662,-0.0003462076,-0.0497612432,0.388887018,0.2213274688,0.2612667382,0.2090934962,0.4744567275,0.1417669356,-0.1024058461,0.1506480575,-0.2302343845,-0.3846206367,0.1483996361,-0.0840582624,0.1118762791,0.0653534457,0.1206891984,0.1489319354,-0.1694583893,-0.0262508765,0.0463692024,0.6140065193,0.4706126451,-0.0234555844,0.2348855138,-0.2129795849,-0.3037589192,-0.0854920521,0.0686824918,0.2077355981,-0.171606347,0.2382114679,-0.2456057966,-0.3759762049,-0.0106182778,0.4144327343,-0.141971752,-0.1881616861,-0.3213080466,0.0009074361,-0.031622909,-0.1063456386,0.1898759305,0.1517588049,1.0623277426,0.0633837655,0.0991474837,-0.0641375408,0.0515113436,-0.0943848491,0.2204993814,-0.0740344003,0.2023748308,-0.2133171111,-0.3748236299,-0.1249232814,0.143462047,0.0070969458,0.1084009781,-0.1320801079,0.1315677613,-0.2157431692,0.10912624,0.0746295974,0.2510552406,0.1244946197,-0.0163330548,-0.0062672244,-0.0468387417,-0.0448912606,0.2159854621,0.0012289876,0.5676503181,-0.0460304394,0.2820247114,0.0842241347,0.0618720986,0.1470472217,0.0061163013,-0.2501160204,0.1226870939,0.2114715427,-0.4649175406,-0.2161206454,-0.0205475949,-0.0445127189,-0.1639263779,0.2474397719,0.3628328741,0.1625466943,-0.2349158674,-0.2758197188,-0.1981358677,0.0529950298,-0.3341674805,0.1772064418,-0.1068423316,0.0287184566,0.0604207851,0.1646991819,0.2909303904,-0.6244517565,-0.0101242764,0.6283121705,-0.0689728633,0.319591701,-0.1501970738,0.0043783835,0.2291322052,-0.0275657438,0.2200334966,0.1052830815,-0.1801231802,-0.0374857597,-0.0583973527,0.0808415338,-0.0990723968,0.1811187416,0.0638208762,0.2293646485,-0.1535037607,0.260083735,-0.1809124053,0.0871979818,-0.1413410008,0.0614007972,-0.1681456268,0.2067477554,-0.5416621566,-0.2201779932,-0.1180699989,-0.2511607111,-0.271753937,-0.2261439115,-0.1266873926,-0.1736239344,-0.1696582437,0.0124385431,0.2903865576,0.3010908961,0.1507919431,-0.2055586874,-0.2148837298,0.0619840138,-0.1120367348,-0.4172698259,0.0203563683,-0.5614672899,-0.1228953153,0.0636729673,0.2879074812,-0.2366262972,-0.4211460352,0.2118504196,0.0256063901,-0.0919334441,0.2861262262,-0.3022787571,-0.2326252759,-0.2141355127,0.0205481183,0.0349209122,0.1182156801,-0.1045978963,-0.2701588571]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/841","title":"Can not reuse datasets already downloaded","comments":"It seems the process needs '\/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py'\r\nWhere and how to assign this ```wikipedia.py``` after I manually download it ?","body":"Hello,\r\nI need to connect to a frontal node (with http proxy, no gpu) before connecting to a gpu node (but no http proxy, so can not use wget so on).\r\nI successfully downloaded and reuse the wikipedia datasets in a frontal node. \r\nWhen I connect to the gpu node, I supposed to use the downloaded datasets from cache, but failed and end with time out error.\r\n\r\nOn frontal node:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('wikipedia', '20200501.en')\r\nReusing dataset wikipedia (\/linkhome\/rech\/genini01\/uua34ms\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/f92599dfccab29832c442b82870fa8f6983e5b4ebbf5e6e2dcbe894e325339cd)\r\n\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/torch\/cuda\/__init__.py:52: UserWarning: CUDA initialization: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http:\/\/www.nvidia.com\/Download\/index.aspx (Triggered internally at \/pytorch\/c10\/cuda\/CUDAFunctions.cpp:100.)\r\n return torch._C._cuda_getDeviceCount() > 0\r\n```\r\n\r\nOn gpu node:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('wikipedia', '20200501.en')\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 160, in _new_conn\r\n (self._dns_host, self.port), self.timeout, **extra_kw\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 84, in create_connection\r\n raise err\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 74, in create_connection\r\n sock.connect(sa)\r\nTimeoutError: [Errno 110] Connection timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 677, in urlopen\r\n chunked=chunked,\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 381, in _make_request\r\n self._validate_conn(conn)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 978, in _validate_conn\r\n conn.connect()\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 309, in connect\r\n conn = self._new_conn()\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 172, in _new_conn\r\n self, \"Failed to establish a new connection: %s\" % e\r\nurllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 449, in send\r\n timeout=timeout\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 727, in urlopen\r\n method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py\", line 446, in increment\r\n raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\nurllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 590, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 264, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return requests.head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/sessions.py\", line 530, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/sessions.py\", line 643, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 516, in send\r\n raise ConnectionError(e, request=request)\r\nrequests.exceptions.ConnectionError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\n```\r\n\r\nAny advice?Thanks!\r\n","comment_length":19,"text":"Can not reuse datasets already downloaded\nHello,\r\nI need to connect to a frontal node (with http proxy, no gpu) before connecting to a gpu node (but no http proxy, so can not use wget so on).\r\nI successfully downloaded and reuse the wikipedia datasets in a frontal node. \r\nWhen I connect to the gpu node, I supposed to use the downloaded datasets from cache, but failed and end with time out error.\r\n\r\nOn frontal node:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('wikipedia', '20200501.en')\r\nReusing dataset wikipedia (\/linkhome\/rech\/genini01\/uua34ms\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/f92599dfccab29832c442b82870fa8f6983e5b4ebbf5e6e2dcbe894e325339cd)\r\n\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/torch\/cuda\/__init__.py:52: UserWarning: CUDA initialization: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http:\/\/www.nvidia.com\/Download\/index.aspx (Triggered internally at \/pytorch\/c10\/cuda\/CUDAFunctions.cpp:100.)\r\n return torch._C._cuda_getDeviceCount() > 0\r\n```\r\n\r\nOn gpu node:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('wikipedia', '20200501.en')\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 160, in _new_conn\r\n (self._dns_host, self.port), self.timeout, **extra_kw\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 84, in create_connection\r\n raise err\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 74, in create_connection\r\n sock.connect(sa)\r\nTimeoutError: [Errno 110] Connection timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 677, in urlopen\r\n chunked=chunked,\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 381, in _make_request\r\n self._validate_conn(conn)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 978, in _validate_conn\r\n conn.connect()\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 309, in connect\r\n conn = self._new_conn()\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 172, in _new_conn\r\n self, \"Failed to establish a new connection: %s\" % e\r\nurllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 449, in send\r\n timeout=timeout\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 727, in urlopen\r\n method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py\", line 446, in increment\r\n raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\nurllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 590, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 264, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return requests.head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/sessions.py\", line 530, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/sessions.py\", line 643, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 516, in send\r\n raise ConnectionError(e, request=request)\r\nrequests.exceptions.ConnectionError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\n```\r\n\r\nAny advice?Thanks!\r\n\nIt seems the process needs '\/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py'\r\nWhere and how to assign this ```wikipedia.py``` after I manually download it ?","embeddings":[-0.1548725963,-0.237658456,-0.0935400799,0.247308448,0.3116977513,0.0938448086,0.2028710246,0.0549309067,0.4361793399,-0.1159729883,0.0571257547,-0.0696771219,0.4684705436,-0.0135196988,0.0379944071,-0.0326445065,-0.1300401986,-0.1004902571,0.043639604,-0.1173638478,-0.067842707,0.0512391962,0.098733671,-0.1973604411,-0.5170297027,-0.0366622321,0.0695867091,0.0257131644,0.1279812306,-0.3292092085,0.4963114858,0.1519474685,0.3188914657,0.3908818066,-0.0001227586,0.0331632793,0.030148387,-0.0086721852,-0.2577647865,-0.2175235897,-0.0583401732,-0.0221800189,0.2983060181,-0.3426147997,-0.3325791955,0.101670213,0.1009476855,-0.5310968757,0.5326627493,0.0265148096,0.0990964547,-0.2245471179,0.1964668036,-0.1530518681,-0.1416291595,0.1641683578,0.0021096452,0.1725648493,0.1944148242,-0.0351392329,0.240674302,0.0909061506,-0.0575655103,0.292018801,0.4881071448,-0.041038204,0.1593190581,-0.4436327815,-0.0056614853,0.1268529743,0.7601932287,-0.1796581596,-0.0926003754,-0.1281972378,-0.2547182739,-0.2147118151,-0.0247817393,0.0598493181,-0.2936771214,0.1041671634,-0.0512807183,-0.3261173368,-0.1129007787,0.3657031059,0.0524880961,0.0805803537,0.2111367285,0.2992266715,0.1300238222,0.0384743251,0.1402243525,-0.200508222,0.1321433336,-0.0326064713,-0.4005512297,0.0403882489,0.0198649764,-0.019808352,0.0059408899,0.4356174171,-0.0049160216,0.1558227986,0.0689676479,0.0348820761,-0.1209750921,0.0979312733,0.0121463193,0.06792593,0.2295104861,0.0980924442,-0.3003791869,-0.1996251047,0.214803949,-0.0826328993,-0.1759484708,0.045841869,0.2712121308,-0.1657319218,-0.0148507813,0.0250686854,-0.1129281819,-0.0103006111,-0.0408026613,0.3777107,-0.1183998212,0.2515930831,0.3486973643,0.0073920749,-0.2008040547,-0.1518076509,-0.0998892114,-0.0000952289,-0.1795948297,0.2472524494,0.3716264069,-0.02905076,0.1782625765,-0.0607331134,0.1072972268,-0.1576452255,0.5114760995,-0.2132519633,0.1165531278,-0.0030914692,0.1830020398,0.4008692503,0.2319138795,0.0943756551,0.0228220876,0.0235882774,-0.2802010179,-0.3750861585,0.0085785277,0.1051453203,0.1237072051,0.1330554336,0.0396871641,-0.1955207735,0.3279145062,-0.2095109075,-0.0007530835,-0.0024339631,-0.416903764,-0.2892304063,0.2441191375,0.5860952139,-0.4188292325,0.0107058454,0.1581711322,0.1157769933,0.1240282804,0.6433213353,-0.3510301709,0.1319203973,-0.1617866009,-0.1739807278,0.2686309516,-0.5118985176,-0.3102938235,0.1303102374,-0.0157408956,0.1743181646,0.0765928924,0.2913169861,0.078338787,0.0681622103,0.1149008647,0.4836628437,-0.0171200335,-0.2183826119,-0.1767849475,-0.2592068017,0.3600456715,-0.100005284,0.1763042957,0.130079329,0.054526329,0.1995591968,0.2683362961,-0.1596238613,0.2831754684,0.231387049,0.1733906716,0.2606773973,0.0332907476,0.0171749331,-0.4030640721,0.3302550912,-0.2461273968,0.0058934791,-0.2333052158,-0.3232369423,-0.1935282648,-0.2746445835,-0.003013182,-0.0217538923,0.0367822424,0.2466294467,0.2087594718,0.1449085325,0.2628288567,0.4253348708,-0.2703686059,0.04579065,-0.3987375796,0.3691842854,-0.2255245149,0.0189247094,-0.2020832896,-0.0271259695,0.3046060205,0.0082265893,0.2096119374,0.0491676554,-0.1407170594,0.3016272485,0.0209527593,0.1291457713,0.3272279203,-0.3400678337,0.0546231046,0.3461517692,0.3550490737,-0.3573242426,-0.1212392151,0.0814285874,0.2531200349,0.1774753183,-0.0738468692,0.0116527285,0.1081959307,-0.0070220106,0.1595490128,-0.0215573218,0.1896750331,0.4042560756,-0.2426960766,-0.0646580383,-0.2098242044,0.2700272202,0.0560824908,0.1939546764,0.0376007371,-0.1223338842,0.0595591776,-0.2091467232,0.3500083387,-0.3235579431,-0.0773916319,0.0955946445,0.3279133141,0.1028089672,-0.1100912392,-0.1628506631,-0.0124461604,-0.1816113442,0.2727844715,-0.0513619222,0.01331932,-0.0920112804,0.0433616899,-0.0553808548,0.3955430388,0.1600077003,-0.3600956798,0.139388293,-0.3536379039,-0.4845449924,0.1353780627,-0.172681585,-0.0969468877,-0.3000202477,-0.1184427962,0.5529574752,0.2341260016,0.2936708629,-0.0186400749,0.2825274765,0.24666439,-0.2299642414,-0.2326804996,-0.3672699332,0.0690651238,0.0263998099,0.3404126465,0.199619025,0.2153250128,-0.1309523284,-0.2550795078,-0.4913638532,-0.08555682,-0.0909787863,0.1733043194,0.1972339004,-0.3326894939,0.349003315,-0.0626110807,-0.2103337348,0.19828628,-0.0772249624,0.0963688567,-0.2657545507,-0.2588169873,0.1132068262,-0.0567036606,-0.346736908,-0.3991658986,-0.3085103929,-0.0470423885,0.1668826342,0.0726270974,-0.068381764,0.178662762,0.0022527329,0.1858491302,0.2069580704,-0.3619476259,-0.3047581613,0.2433561087,-0.1396142542,-0.3152446151,0.1756567806,0.1172586679,-0.0212443564,0.471175909,-0.4773329496,-0.3291167021,-0.0574132986,0.3427009284,0.0243760739,-0.1468636394,0.3705311716,-0.3419768512,-0.0292724743,0.0876091123,0.0092929425,0.0055176825,-0.307690084,0.4178673625,0.1014434099,0.3604357243,0.0691400543,0.9451534152,0.1241173595,-0.2306533456,0.3185611665,-0.1047408879,0.2298144251,-0.103616789,-0.1561472863,0.0744232088,-0.3390268087,0.2327618897,-0.3644365072,-0.1278510392,-0.2709164321,-0.334579885,0.1986713558,-0.3212337196,-0.4758189619,-0.138314411,-0.2197270989,0.382599622,0.3686401248,-0.0122793978,-0.2355505973,-0.1893039644,0.0968700871,0.1850917488,0.1408934444,0.1471264064,-0.4159779251,0.2112641484,-0.3479153216,0.4394700527,0.0835312307,0.1060061902,-0.1756965816,0.0594754443,0.1006443948,-0.1219907179,0.7283738852,-0.3127952814,0.1315129846,0.207166627,-0.2820923626,-0.4684227407,0.2966364622,0.0124673536,0.1171653867,0.3558496237,-0.0800430551,-0.203017652,-0.0769134834,-0.3034848571,0.2625688314,-0.2390353978,0.1314582676,0.0752761438,-0.2807820737,-0.6343670487,-0.0619315468,-0.0307715777,0.2057969719,-0.0782400295,-0.0400362089,-0.292501986,0.4315606952,0.0750034899,-0.0184841119,0.2736267745,0.0548881553,0.2037198097,0.0949679017,0.1674259156,0.1388572007,0.2600415647,0.2361792326,-0.0487389714,0.0585296676,0.1607573181,-0.4420765042,0.4043967426,0.1165127978,0.3205904663,0.1022855863,-0.0798602626,-0.2230350077,0.1919568777,-0.0529369414,0.2260248214,-0.3468127549,-0.602173686,0.0737718865,0.136892423,-0.1481115073,0.4638576806,0.1307070702,-0.1034088284,-0.0100934533,0.1330152154,1.1369920969,-0.1097073629,0.2494181544,0.0549144484,0.0543845519,0.7935971022,-0.3770438731,0.3242078722,-0.2323517352,0.148858428,-0.1180851609,-0.1100877076,0.1324557662,-0.1406783611,-0.048584085,0.1924432218,0.1312217563,0.0881239846,-0.1138620228,0.3052190542,-0.0495613441,-0.1473941058,0.2064673752,0.0588750541,-0.0345370136,0.3890476525,-0.3410145044,0.2255520523,0.1073000729,-0.1429103762,-0.1898394972,0.0651275143,-0.3222646117,0.5679987073,-0.4035392106,-0.587012291,0.0778376237,0.2357517034,0.295404166,0.0649521276,-0.3694161475,0.0318314321,-0.3367700875,-0.2194244266,-0.2801853418,-0.1022374481,0.2282085568,-0.0931406692,-0.4943189323,0.4344157577,0.1626363099,-0.3901789486,-0.1335012168,0.0061755725,-0.3375343084,-0.061242193,-0.2839338481,0.2013020962,-0.1708076745,-0.1155114919,0.0372127257,-0.0539513603,-0.1507027,0.0140875578,0.1188190132,-0.1676081419,0.0437566265,0.6288123727,0.3283686638,-0.1268274486,0.4865684211,0.3852147758,-0.3451159894,-0.0897152424,-0.5433734059,-0.2771345377,-0.4499161839,-0.1004412025,-0.0599107668,0.2556528151,-0.071073316,-0.1618792862,0.1116944328,0.1838758588,0.2296092212,-0.4139637351,-0.2296734452,0.1353534758,-0.2252458334,0.225109309,0.3666955829,-0.066802673,0.1139815599,0.1445028782,-0.2293465883,0.3229386806,-0.1981832385,0.2599373758,-0.0411989726,-0.2617835402,-0.0257216711,-0.1522387862,-0.1032204404,0.1068646461,-0.2564244568,-0.1204989031,-0.0193754751,0.168319568,0.0862177014,-0.3023171127,0.0452374369,-0.5413496494,0.1897811741,-0.0564400107,0.3283282518,0.2563540936,-0.0059872204,-0.0656555891,0.1460813284,0.0260093641,-0.4261065125,0.2824684381,0.1789683402,0.0737882629,0.2832489312,0.1818433106,0.0182467308,0.0927433968,-0.3297591805,0.0606947653,0.355280906,-0.1543047726,0.3506964147,-0.1640052497,-0.1087649688,-0.017438598,0.183109507,0.0715587959,-0.1704185158,0.2969181538,0.003957775,0.0945935175,-0.2185349911,0.1525517404,0.271368891,-0.1579953879,0.0503200889,0.113421388,0.4863115251,0.0114199212,0.2574236691,0.1771299541,0.3620299101,-0.251676023,0.2153895199,-0.1647552252,0.0180758275,-0.1410150826,0.1426822394,0.4029867947,0.0652771369,0.1033280045,-0.1458620429,0.3220045567,-0.1225010008,0.3613238037,-0.1696137786,-0.4061296582,-0.088118583,0.0352235883,-0.2489658296,-0.0163800735,-0.2956756055,-0.130460158,0.173795104,-0.2261281908,-0.0401272587,0.4048238993,-0.2687307298,-0.2827675343,-0.3352327347,-0.3463445902,0.1784393936,-0.0429823659,0.2153524458,-0.33094877,0.2442252785,0.0466660075,-0.4038794935,-0.3088327944,0.3640608191,-0.0036338333,0.108394444,-0.0482486151,0.0521883778,0.1386573315,-0.2849012911,-0.0889848843,0.305285275,0.4353722334,0.0898911878,-0.1928082407,0.5962461829,0.1944009215,0.0021676521,-0.1304047555,0.0647217408,0.412422061,-0.0204030741,0.2085270733,0.1300189048,-0.0924625099,-0.0421472415,0.1148763746,0.4866264462,-0.3733814359,0.2650318444,-0.1798631847,-0.0689876601,-0.0901721641,0.2335721403,-0.2867179215,0.2885174751,0.611315608,-0.0639151633,0.1486181766,-0.1438118368,0.0186846796,-0.2901233435,0.4194196165,0.1110554338,0.3019726574,-0.3612966239,-0.1454018503,-0.4733638465,0.2039208859,-0.4668929577,0.0593985133,-0.0580026992,0.4289111793,0.0649042875,0.1286662966,-0.1016300321,-0.0497974455,0.1085102186,0.1072006524,-0.2101535648,-0.2865597904,-0.3860230744,-0.0287037976,-0.0604817718,-0.4726716876,0.0891435221,0.063199684,-0.0475048758,-0.0783382729,-0.1557792127,0.1945841759,0.2042075247,0.4846568704,0.1691206843,0.6330028772,0.2342400253,0.0923510566,-0.2478438765,-0.3968434632,0.0741073564,0.1058295146,0.1868032664,0.0478784889,-0.5410246253,-0.2097579539,-0.28719455,0.3450892568,-0.0004355547,-0.1266297549,-0.0814173073,-0.2556202412,-0.1839204282,0.2147217691,0.1197051257,0.335378468,0.0070096436,0.2591794133,-0.2700031698,-0.1904765069,0.7102396488,-0.1347930282,-0.1052471325,-0.3297406733,0.0763674825,0.1047487557,-0.0031945943,-0.3488744795,0.0254382845,0.2682200968,-0.1837798059,-0.1111019328,0.0281186514,-0.2183413059,0.1376701593,-0.2694945931,0.0691198856,0.0182127785,-0.2271564752,-0.3523175418,-0.2803575099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/841","title":"Can not reuse datasets already downloaded","comments":"\r\ndownload the ```wikipedia.py``` at the working directory and go with ```dataset = load_dataset('wikipedia.py', '20200501.en')``` works.","body":"Hello,\r\nI need to connect to a frontal node (with http proxy, no gpu) before connecting to a gpu node (but no http proxy, so can not use wget so on).\r\nI successfully downloaded and reuse the wikipedia datasets in a frontal node. \r\nWhen I connect to the gpu node, I supposed to use the downloaded datasets from cache, but failed and end with time out error.\r\n\r\nOn frontal node:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('wikipedia', '20200501.en')\r\nReusing dataset wikipedia (\/linkhome\/rech\/genini01\/uua34ms\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/f92599dfccab29832c442b82870fa8f6983e5b4ebbf5e6e2dcbe894e325339cd)\r\n\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/torch\/cuda\/__init__.py:52: UserWarning: CUDA initialization: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http:\/\/www.nvidia.com\/Download\/index.aspx (Triggered internally at \/pytorch\/c10\/cuda\/CUDAFunctions.cpp:100.)\r\n return torch._C._cuda_getDeviceCount() > 0\r\n```\r\n\r\nOn gpu node:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('wikipedia', '20200501.en')\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 160, in _new_conn\r\n (self._dns_host, self.port), self.timeout, **extra_kw\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 84, in create_connection\r\n raise err\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 74, in create_connection\r\n sock.connect(sa)\r\nTimeoutError: [Errno 110] Connection timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 677, in urlopen\r\n chunked=chunked,\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 381, in _make_request\r\n self._validate_conn(conn)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 978, in _validate_conn\r\n conn.connect()\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 309, in connect\r\n conn = self._new_conn()\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 172, in _new_conn\r\n self, \"Failed to establish a new connection: %s\" % e\r\nurllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 449, in send\r\n timeout=timeout\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 727, in urlopen\r\n method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py\", line 446, in increment\r\n raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\nurllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 590, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 264, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return requests.head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/sessions.py\", line 530, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/sessions.py\", line 643, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 516, in send\r\n raise ConnectionError(e, request=request)\r\nrequests.exceptions.ConnectionError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\n```\r\n\r\nAny advice?Thanks!\r\n","comment_length":15,"text":"Can not reuse datasets already downloaded\nHello,\r\nI need to connect to a frontal node (with http proxy, no gpu) before connecting to a gpu node (but no http proxy, so can not use wget so on).\r\nI successfully downloaded and reuse the wikipedia datasets in a frontal node. \r\nWhen I connect to the gpu node, I supposed to use the downloaded datasets from cache, but failed and end with time out error.\r\n\r\nOn frontal node:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('wikipedia', '20200501.en')\r\nReusing dataset wikipedia (\/linkhome\/rech\/genini01\/uua34ms\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/f92599dfccab29832c442b82870fa8f6983e5b4ebbf5e6e2dcbe894e325339cd)\r\n\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/torch\/cuda\/__init__.py:52: UserWarning: CUDA initialization: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http:\/\/www.nvidia.com\/Download\/index.aspx (Triggered internally at \/pytorch\/c10\/cuda\/CUDAFunctions.cpp:100.)\r\n return torch._C._cuda_getDeviceCount() > 0\r\n```\r\n\r\nOn gpu node:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset('wikipedia', '20200501.en')\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 160, in _new_conn\r\n (self._dns_host, self.port), self.timeout, **extra_kw\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 84, in create_connection\r\n raise err\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 74, in create_connection\r\n sock.connect(sa)\r\nTimeoutError: [Errno 110] Connection timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 677, in urlopen\r\n chunked=chunked,\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 381, in _make_request\r\n self._validate_conn(conn)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 978, in _validate_conn\r\n conn.connect()\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 309, in connect\r\n conn = self._new_conn()\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 172, in _new_conn\r\n self, \"Failed to establish a new connection: %s\" % e\r\nurllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 449, in send\r\n timeout=timeout\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 727, in urlopen\r\n method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py\", line 446, in increment\r\n raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\nurllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 590, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 264, in prepare_module\r\n head_hf_s3(path, filename=name, dataset=dataset)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 200, in head_hf_s3\r\n return requests.head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/api.py\", line 104, in head\r\n return request('head', url, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/api.py\", line 61, in request\r\n return session.request(method=method, url=url, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/sessions.py\", line 530, in request\r\n resp = self.send(prep, **send_kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/sessions.py\", line 643, in send\r\n r = adapter.send(request, **kwargs)\r\n File \"\/linkhome\/rech\/genini01\/uua34ms\/work\/anaconda3\/envs\/pytorch_pip170_cuda102\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 516, in send\r\n raise ConnectionError(e, request=request)\r\nrequests.exceptions.ConnectionError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/wikipedia\/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\n```\r\n\r\nAny advice?Thanks!\r\n\n\r\ndownload the ```wikipedia.py``` at the working directory and go with ```dataset = load_dataset('wikipedia.py', '20200501.en')``` works.","embeddings":[-0.1548725963,-0.237658456,-0.0935400799,0.247308448,0.3116977513,0.0938448086,0.2028710246,0.0549309067,0.4361793399,-0.1159729883,0.0571257547,-0.0696771219,0.4684705436,-0.0135196988,0.0379944071,-0.0326445065,-0.1300401986,-0.1004902571,0.043639604,-0.1173638478,-0.067842707,0.0512391962,0.098733671,-0.1973604411,-0.5170297027,-0.0366622321,0.0695867091,0.0257131644,0.1279812306,-0.3292092085,0.4963114858,0.1519474685,0.3188914657,0.3908818066,-0.0001227586,0.0331632793,0.030148387,-0.0086721852,-0.2577647865,-0.2175235897,-0.0583401732,-0.0221800189,0.2983060181,-0.3426147997,-0.3325791955,0.101670213,0.1009476855,-0.5310968757,0.5326627493,0.0265148096,0.0990964547,-0.2245471179,0.1964668036,-0.1530518681,-0.1416291595,0.1641683578,0.0021096452,0.1725648493,0.1944148242,-0.0351392329,0.240674302,0.0909061506,-0.0575655103,0.292018801,0.4881071448,-0.041038204,0.1593190581,-0.4436327815,-0.0056614853,0.1268529743,0.7601932287,-0.1796581596,-0.0926003754,-0.1281972378,-0.2547182739,-0.2147118151,-0.0247817393,0.0598493181,-0.2936771214,0.1041671634,-0.0512807183,-0.3261173368,-0.1129007787,0.3657031059,0.0524880961,0.0805803537,0.2111367285,0.2992266715,0.1300238222,0.0384743251,0.1402243525,-0.200508222,0.1321433336,-0.0326064713,-0.4005512297,0.0403882489,0.0198649764,-0.019808352,0.0059408899,0.4356174171,-0.0049160216,0.1558227986,0.0689676479,0.0348820761,-0.1209750921,0.0979312733,0.0121463193,0.06792593,0.2295104861,0.0980924442,-0.3003791869,-0.1996251047,0.214803949,-0.0826328993,-0.1759484708,0.045841869,0.2712121308,-0.1657319218,-0.0148507813,0.0250686854,-0.1129281819,-0.0103006111,-0.0408026613,0.3777107,-0.1183998212,0.2515930831,0.3486973643,0.0073920749,-0.2008040547,-0.1518076509,-0.0998892114,-0.0000952289,-0.1795948297,0.2472524494,0.3716264069,-0.02905076,0.1782625765,-0.0607331134,0.1072972268,-0.1576452255,0.5114760995,-0.2132519633,0.1165531278,-0.0030914692,0.1830020398,0.4008692503,0.2319138795,0.0943756551,0.0228220876,0.0235882774,-0.2802010179,-0.3750861585,0.0085785277,0.1051453203,0.1237072051,0.1330554336,0.0396871641,-0.1955207735,0.3279145062,-0.2095109075,-0.0007530835,-0.0024339631,-0.416903764,-0.2892304063,0.2441191375,0.5860952139,-0.4188292325,0.0107058454,0.1581711322,0.1157769933,0.1240282804,0.6433213353,-0.3510301709,0.1319203973,-0.1617866009,-0.1739807278,0.2686309516,-0.5118985176,-0.3102938235,0.1303102374,-0.0157408956,0.1743181646,0.0765928924,0.2913169861,0.078338787,0.0681622103,0.1149008647,0.4836628437,-0.0171200335,-0.2183826119,-0.1767849475,-0.2592068017,0.3600456715,-0.100005284,0.1763042957,0.130079329,0.054526329,0.1995591968,0.2683362961,-0.1596238613,0.2831754684,0.231387049,0.1733906716,0.2606773973,0.0332907476,0.0171749331,-0.4030640721,0.3302550912,-0.2461273968,0.0058934791,-0.2333052158,-0.3232369423,-0.1935282648,-0.2746445835,-0.003013182,-0.0217538923,0.0367822424,0.2466294467,0.2087594718,0.1449085325,0.2628288567,0.4253348708,-0.2703686059,0.04579065,-0.3987375796,0.3691842854,-0.2255245149,0.0189247094,-0.2020832896,-0.0271259695,0.3046060205,0.0082265893,0.2096119374,0.0491676554,-0.1407170594,0.3016272485,0.0209527593,0.1291457713,0.3272279203,-0.3400678337,0.0546231046,0.3461517692,0.3550490737,-0.3573242426,-0.1212392151,0.0814285874,0.2531200349,0.1774753183,-0.0738468692,0.0116527285,0.1081959307,-0.0070220106,0.1595490128,-0.0215573218,0.1896750331,0.4042560756,-0.2426960766,-0.0646580383,-0.2098242044,0.2700272202,0.0560824908,0.1939546764,0.0376007371,-0.1223338842,0.0595591776,-0.2091467232,0.3500083387,-0.3235579431,-0.0773916319,0.0955946445,0.3279133141,0.1028089672,-0.1100912392,-0.1628506631,-0.0124461604,-0.1816113442,0.2727844715,-0.0513619222,0.01331932,-0.0920112804,0.0433616899,-0.0553808548,0.3955430388,0.1600077003,-0.3600956798,0.139388293,-0.3536379039,-0.4845449924,0.1353780627,-0.172681585,-0.0969468877,-0.3000202477,-0.1184427962,0.5529574752,0.2341260016,0.2936708629,-0.0186400749,0.2825274765,0.24666439,-0.2299642414,-0.2326804996,-0.3672699332,0.0690651238,0.0263998099,0.3404126465,0.199619025,0.2153250128,-0.1309523284,-0.2550795078,-0.4913638532,-0.08555682,-0.0909787863,0.1733043194,0.1972339004,-0.3326894939,0.349003315,-0.0626110807,-0.2103337348,0.19828628,-0.0772249624,0.0963688567,-0.2657545507,-0.2588169873,0.1132068262,-0.0567036606,-0.346736908,-0.3991658986,-0.3085103929,-0.0470423885,0.1668826342,0.0726270974,-0.068381764,0.178662762,0.0022527329,0.1858491302,0.2069580704,-0.3619476259,-0.3047581613,0.2433561087,-0.1396142542,-0.3152446151,0.1756567806,0.1172586679,-0.0212443564,0.471175909,-0.4773329496,-0.3291167021,-0.0574132986,0.3427009284,0.0243760739,-0.1468636394,0.3705311716,-0.3419768512,-0.0292724743,0.0876091123,0.0092929425,0.0055176825,-0.307690084,0.4178673625,0.1014434099,0.3604357243,0.0691400543,0.9451534152,0.1241173595,-0.2306533456,0.3185611665,-0.1047408879,0.2298144251,-0.103616789,-0.1561472863,0.0744232088,-0.3390268087,0.2327618897,-0.3644365072,-0.1278510392,-0.2709164321,-0.334579885,0.1986713558,-0.3212337196,-0.4758189619,-0.138314411,-0.2197270989,0.382599622,0.3686401248,-0.0122793978,-0.2355505973,-0.1893039644,0.0968700871,0.1850917488,0.1408934444,0.1471264064,-0.4159779251,0.2112641484,-0.3479153216,0.4394700527,0.0835312307,0.1060061902,-0.1756965816,0.0594754443,0.1006443948,-0.1219907179,0.7283738852,-0.3127952814,0.1315129846,0.207166627,-0.2820923626,-0.4684227407,0.2966364622,0.0124673536,0.1171653867,0.3558496237,-0.0800430551,-0.203017652,-0.0769134834,-0.3034848571,0.2625688314,-0.2390353978,0.1314582676,0.0752761438,-0.2807820737,-0.6343670487,-0.0619315468,-0.0307715777,0.2057969719,-0.0782400295,-0.0400362089,-0.292501986,0.4315606952,0.0750034899,-0.0184841119,0.2736267745,0.0548881553,0.2037198097,0.0949679017,0.1674259156,0.1388572007,0.2600415647,0.2361792326,-0.0487389714,0.0585296676,0.1607573181,-0.4420765042,0.4043967426,0.1165127978,0.3205904663,0.1022855863,-0.0798602626,-0.2230350077,0.1919568777,-0.0529369414,0.2260248214,-0.3468127549,-0.602173686,0.0737718865,0.136892423,-0.1481115073,0.4638576806,0.1307070702,-0.1034088284,-0.0100934533,0.1330152154,1.1369920969,-0.1097073629,0.2494181544,0.0549144484,0.0543845519,0.7935971022,-0.3770438731,0.3242078722,-0.2323517352,0.148858428,-0.1180851609,-0.1100877076,0.1324557662,-0.1406783611,-0.048584085,0.1924432218,0.1312217563,0.0881239846,-0.1138620228,0.3052190542,-0.0495613441,-0.1473941058,0.2064673752,0.0588750541,-0.0345370136,0.3890476525,-0.3410145044,0.2255520523,0.1073000729,-0.1429103762,-0.1898394972,0.0651275143,-0.3222646117,0.5679987073,-0.4035392106,-0.587012291,0.0778376237,0.2357517034,0.295404166,0.0649521276,-0.3694161475,0.0318314321,-0.3367700875,-0.2194244266,-0.2801853418,-0.1022374481,0.2282085568,-0.0931406692,-0.4943189323,0.4344157577,0.1626363099,-0.3901789486,-0.1335012168,0.0061755725,-0.3375343084,-0.061242193,-0.2839338481,0.2013020962,-0.1708076745,-0.1155114919,0.0372127257,-0.0539513603,-0.1507027,0.0140875578,0.1188190132,-0.1676081419,0.0437566265,0.6288123727,0.3283686638,-0.1268274486,0.4865684211,0.3852147758,-0.3451159894,-0.0897152424,-0.5433734059,-0.2771345377,-0.4499161839,-0.1004412025,-0.0599107668,0.2556528151,-0.071073316,-0.1618792862,0.1116944328,0.1838758588,0.2296092212,-0.4139637351,-0.2296734452,0.1353534758,-0.2252458334,0.225109309,0.3666955829,-0.066802673,0.1139815599,0.1445028782,-0.2293465883,0.3229386806,-0.1981832385,0.2599373758,-0.0411989726,-0.2617835402,-0.0257216711,-0.1522387862,-0.1032204404,0.1068646461,-0.2564244568,-0.1204989031,-0.0193754751,0.168319568,0.0862177014,-0.3023171127,0.0452374369,-0.5413496494,0.1897811741,-0.0564400107,0.3283282518,0.2563540936,-0.0059872204,-0.0656555891,0.1460813284,0.0260093641,-0.4261065125,0.2824684381,0.1789683402,0.0737882629,0.2832489312,0.1818433106,0.0182467308,0.0927433968,-0.3297591805,0.0606947653,0.355280906,-0.1543047726,0.3506964147,-0.1640052497,-0.1087649688,-0.017438598,0.183109507,0.0715587959,-0.1704185158,0.2969181538,0.003957775,0.0945935175,-0.2185349911,0.1525517404,0.271368891,-0.1579953879,0.0503200889,0.113421388,0.4863115251,0.0114199212,0.2574236691,0.1771299541,0.3620299101,-0.251676023,0.2153895199,-0.1647552252,0.0180758275,-0.1410150826,0.1426822394,0.4029867947,0.0652771369,0.1033280045,-0.1458620429,0.3220045567,-0.1225010008,0.3613238037,-0.1696137786,-0.4061296582,-0.088118583,0.0352235883,-0.2489658296,-0.0163800735,-0.2956756055,-0.130460158,0.173795104,-0.2261281908,-0.0401272587,0.4048238993,-0.2687307298,-0.2827675343,-0.3352327347,-0.3463445902,0.1784393936,-0.0429823659,0.2153524458,-0.33094877,0.2442252785,0.0466660075,-0.4038794935,-0.3088327944,0.3640608191,-0.0036338333,0.108394444,-0.0482486151,0.0521883778,0.1386573315,-0.2849012911,-0.0889848843,0.305285275,0.4353722334,0.0898911878,-0.1928082407,0.5962461829,0.1944009215,0.0021676521,-0.1304047555,0.0647217408,0.412422061,-0.0204030741,0.2085270733,0.1300189048,-0.0924625099,-0.0421472415,0.1148763746,0.4866264462,-0.3733814359,0.2650318444,-0.1798631847,-0.0689876601,-0.0901721641,0.2335721403,-0.2867179215,0.2885174751,0.611315608,-0.0639151633,0.1486181766,-0.1438118368,0.0186846796,-0.2901233435,0.4194196165,0.1110554338,0.3019726574,-0.3612966239,-0.1454018503,-0.4733638465,0.2039208859,-0.4668929577,0.0593985133,-0.0580026992,0.4289111793,0.0649042875,0.1286662966,-0.1016300321,-0.0497974455,0.1085102186,0.1072006524,-0.2101535648,-0.2865597904,-0.3860230744,-0.0287037976,-0.0604817718,-0.4726716876,0.0891435221,0.063199684,-0.0475048758,-0.0783382729,-0.1557792127,0.1945841759,0.2042075247,0.4846568704,0.1691206843,0.6330028772,0.2342400253,0.0923510566,-0.2478438765,-0.3968434632,0.0741073564,0.1058295146,0.1868032664,0.0478784889,-0.5410246253,-0.2097579539,-0.28719455,0.3450892568,-0.0004355547,-0.1266297549,-0.0814173073,-0.2556202412,-0.1839204282,0.2147217691,0.1197051257,0.335378468,0.0070096436,0.2591794133,-0.2700031698,-0.1904765069,0.7102396488,-0.1347930282,-0.1052471325,-0.3297406733,0.0763674825,0.1047487557,-0.0031945943,-0.3488744795,0.0254382845,0.2682200968,-0.1837798059,-0.1111019328,0.0281186514,-0.2183413059,0.1376701593,-0.2694945931,0.0691198856,0.0182127785,-0.2271564752,-0.3523175418,-0.2803575099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/836","title":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas","comments":"Which version of pyarrow do you have ? Could you try to update pyarrow and try again ?","body":"Hi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n","comment_length":18,"text":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas\nHi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n\nWhich version of pyarrow do you have ? Could you try to update pyarrow and try again ?","embeddings":[-0.24118343,-0.3198650777,-0.0681558475,0.4316074848,0.3625251055,0.0191550385,0.5210206509,0.437713474,0.2949123979,0.0327753536,0.0169624221,-0.1187045723,0.0905919969,0.1802892983,0.0350663252,0.1130092442,0.0730282664,0.3988213837,0.0317200907,0.0699720979,-0.218860805,0.1208694875,0.005867396,0.0575151071,-0.1244225353,0.0723639429,-0.008208585,0.1169146225,-0.2343517244,-0.4013206661,0.4197234511,0.0784542188,0.3328505158,0.5207661986,-0.0001169051,0.1564733684,0.2654486001,-0.1856259704,-0.0441993587,-0.5195806026,0.0301294662,-0.5329574347,0.2776663601,-0.1672024429,-0.0144887799,-0.1786803156,-0.1109988168,-0.0358052403,0.3014124632,0.3175454736,0.1804958135,-0.2607497573,0.0598921739,-0.0100186532,0.2052922398,0.0371788666,-0.0156855155,0.4175470769,0.1173958927,-0.1267608404,-0.0487308204,-0.193666935,-0.2392427474,0.3581996262,0.3196350932,0.1399399787,0.0709672496,-0.1089668348,0.2680699527,0.1055052057,0.6067355871,-0.307642132,-0.1485474855,-0.0154003808,-0.0202969182,-0.4886516333,0.2558154762,0.404494822,-0.1521599442,0.1173331738,-0.2122125626,0.2484468818,-0.2663929462,0.2028620988,-0.0673474893,-0.3531901836,-0.2522074878,0.2693178952,0.2523397803,0.0737477168,0.3336401582,-0.0937501863,0.0268538464,0.1806966811,-0.5556491017,0.2796493471,-0.1176608056,-0.0904364362,0.1637407988,0.096218653,0.4014219046,0.1370620281,0.0423445031,0.1946978569,0.3191299736,0.1861328632,-0.0428146571,0.0691050887,0.0273857322,0.3462172449,-0.2297836244,-0.2338579148,-0.3881369233,-0.5049571991,0.2381427437,-0.1538383663,0.2674439549,-0.2704333961,-0.2479329556,-0.0960200727,-0.0361324735,-0.0094233667,-0.0528344922,0.5738207698,0.215833962,0.3802042603,-0.054209996,0.2572092116,-0.2112757564,-0.09865468,-0.0803525597,0.0201157145,-0.0576021709,0.0522914641,0.3777484894,-0.2166886032,0.1214607805,0.1762953401,-0.043517001,-0.2938742042,-0.1455057859,-0.4177222252,0.0367241055,0.283308208,0.2024739832,0.0979529768,0.2341527939,-0.2812092602,-0.0606879592,0.3805413246,-0.2301190495,-0.1685689539,0.0972641706,0.0963025913,-0.0661173835,0.1656782478,-0.3956311643,0.1563510746,0.3519503474,-0.4384936988,0.0334676392,-0.2134845853,-0.2411233634,-0.3524810374,0.2018900961,0.2790597677,-0.7854661942,0.3429474831,0.0539065748,-0.0317488685,0.1149988398,0.4237150252,-0.1483171433,0.0611786693,-0.2709869146,0.0921530128,0.1406241804,0.007542707,-0.6293674707,0.3307564855,-0.04602734,0.0895455927,0.3120644391,0.0645528063,0.0512152202,0.0891771689,-0.1080223322,0.0185540114,0.0231345072,0.0440054424,-0.2607476711,-0.1291924864,0.3187506497,0.142625019,-0.0647743195,-0.0676729232,0.2388889641,-0.4087391198,0.2025453746,-0.0180878602,0.1416161656,0.1752468795,0.1791311353,-0.1968236417,0.0349740647,0.0104333451,-0.4493494034,0.1433825642,0.2848802209,-0.1687007099,-0.2159446925,-0.0797307715,-0.384118408,-0.0297010913,-0.2681604922,-0.1241994351,-0.0246524885,0.0725153983,0.0595197789,-0.1440529078,-0.1857402176,0.1161194369,-0.2237771302,-0.0112759126,-0.0158436764,0.1422190964,0.058799319,-0.1249890551,0.1064720228,-0.0629007593,0.207167998,-0.0483339541,-0.2763522565,0.2495912611,0.1456864476,0.0456155054,-0.2857958674,-0.1747876853,0.0830201879,-0.2415053844,-0.0872012973,0.0237385109,0.3663716614,-0.1825469434,-0.3392808735,0.2706229687,-0.1567003578,0.3015136421,0.1879752129,-0.3580714464,0.3065021634,0.016720105,0.1781937331,0.0704920888,0.1380855441,0.0736546293,0.4947431087,0.2995411456,-0.0968417376,-0.1562502384,0.5009906292,-0.0470440127,-0.2185632885,0.3112592995,-0.1181778163,-0.1945351362,0.0684309453,0.0887451172,0.3798215091,0.1293763965,-0.1915307492,0.0146643994,0.1435023844,-0.2500295639,0.1204057485,-0.2204343826,0.1473618895,0.3965650201,0.1270093769,-0.1272856295,-0.195144549,0.0377438515,0.0875855386,0.3257235587,-0.4119557142,-0.0675459281,-0.0734540001,-0.0052061281,-0.1802845895,0.0986031592,-0.0555794202,0.0971230865,-0.3996109366,0.3328239024,0.2897825539,-0.0790665746,-0.2431259155,0.08863388,0.2158128768,-0.3370558321,-0.0034607018,-0.0489910357,-0.2919403613,-0.0261766538,0.3307105303,0.1098190099,-0.0008771374,-0.0217662286,0.0280251838,-0.0548448637,-0.0925327018,-0.0439247601,-0.0732297674,0.0209216662,-0.0511663184,0.4193309844,0.0076079252,-0.1446308047,0.4017484784,-0.2536608279,-0.0489250831,0.350135982,0.0636930913,-0.0958797708,-0.1087071598,-0.3498898149,-0.139888227,-0.4293371439,0.3910627365,0.0847561061,0.0419315621,0.0926315561,0.4331049323,-0.0260847546,0.3013546467,-0.0312164128,-0.0848050937,-0.2977054417,0.3645742834,-0.0561163239,-0.4676565826,0.011941432,0.0828354955,0.3591901362,0.1638266891,-0.4252471328,0.3267471194,-0.15090774,0.1628198922,0.0011049408,-0.0242274385,0.0609546043,-0.0872276798,-0.0218234472,0.0364301093,-0.2417160571,-0.0141803464,-0.0320445746,0.1475835294,0.0513685606,0.458429724,-0.0371929593,0.6513361931,0.0874645486,-0.0882313102,0.4855556488,-0.308239907,0.3818774223,-0.2096708715,-0.3737454712,-0.0855747312,-0.3519817293,-0.1924462765,0.2677486241,-0.0335841477,-0.0334150307,-0.1257756203,0.0035179411,-0.4046861231,-0.0377031192,0.183632195,0.0306968763,-0.0148301804,-0.1005954891,-0.0133011667,-0.0116456645,-0.237362951,-0.1942582428,0.2739638686,-0.1513422579,0.0633870587,-0.2242484242,-0.0615803823,-0.3029371798,0.3292051852,-0.0136234136,0.3102864921,0.0106146866,-0.0176951028,0.0972969308,-0.1252222657,0.7449785471,0.1892860085,-0.1550699919,0.1278296113,-0.0241136737,-0.441583544,-0.121761322,0.1906314343,0.3073361516,0.0744616762,0.3548655212,-0.0703261346,-0.0096933683,-0.1443433017,0.3546780944,0.0432527624,-0.3951280713,-0.5194913745,-0.2349791825,-0.383053571,-0.2121350318,-0.0898666456,0.2882787883,-0.3244074285,-0.1423869133,0.1183612794,-0.2990536094,0.0601567216,0.1533810496,0.2973894477,-0.1129143313,0.1448963135,0.0468351059,0.0698044524,0.376911819,0.7987186313,-0.0785759091,-0.6220775843,0.0848431289,-0.2050266713,0.3853479028,0.0478198007,-0.2032050937,0.1289862841,0.0907342061,0.0620335788,0.0379479825,0.0144376121,0.3036864102,0.0496549308,-0.3454539776,-0.5439224243,0.4874189198,-0.1259078234,0.2102593035,0.302079767,-0.2119999677,-0.1099810228,0.1555226296,-0.3346246481,0.6592135429,-0.4218300581,0.1938303411,0.3073363602,0.056336008,0.2584289312,-0.4451968074,0.1323073953,-0.3919684887,-0.210903883,-0.1043123081,-0.0712032765,0.2393857986,0.432190001,-0.1212453321,0.3118637204,-0.0741497204,0.3325560987,0.11313425,0.2438252568,-0.1005161256,-0.0864683241,-0.3010203838,0.0104239564,-0.1430807561,-0.0349519551,0.028959142,0.0544647723,-0.1805675626,-0.2016452104,-0.2461119592,0.2200246751,-0.2911679149,0.1154046133,-0.1109549478,-0.4686649442,0.1963640451,0.2380675226,-0.2087023854,-0.0851794481,-0.1240850538,0.1550540775,0.1596347541,-0.0401756726,-0.2083743513,-0.0536952429,-0.0158856176,0.0761261061,-0.2351237983,0.1545512527,-0.122278817,-0.0956747606,0.2780734897,0.2376574725,-0.0848071724,-0.5452085137,-0.5326197147,-0.4284158647,0.0759811029,-0.1957320869,0.0651982948,-0.0420785286,0.0983413011,-0.1616345942,0.1859346777,-0.4688693583,-0.1020267904,0.3918125927,-0.1116865948,0.0939545855,0.5907301903,0.2193497866,-0.0168925617,-0.1411525458,0.0642407984,0.3081057072,-0.3531885743,0.1170660481,0.1141491011,0.3159501255,0.1392528862,0.2917454839,0.1565803736,-0.0776725486,0.0744550973,-0.5115905404,-0.3249694705,-0.0165290646,-0.0015315027,0.2479578257,0.1672193259,0.08186692,0.1726779193,0.2032965571,-0.1562302262,0.1004625708,0.1059293076,-0.0869575143,0.2106330097,0.1420163661,-0.1338726133,-0.1038192213,0.0934978351,0.1776407212,-0.0534449145,-0.1887511909,-0.0798553601,0.1500999779,0.0983605087,0.0232860353,-0.2034453899,-0.2526888847,-0.1497063339,-0.19280532,0.0664090961,0.0297824815,-0.0981118679,0.0717605129,-0.2987648547,-0.0846538544,-0.5091862082,0.0271216799,-0.5075390935,0.2231322825,0.29926911,0.0485824719,0.1217573285,0.2004558295,-0.1724265963,-0.0311586745,0.1516266763,-0.1366193742,0.3009471297,-0.4014762044,0.2035068721,-0.0510654896,0.3403376043,0.4848605692,-0.2076416463,0.2197109312,0.0375863202,0.1341150701,-0.3825282753,0.2754767835,0.1172315627,-0.0616297424,-0.027484389,0.1696746349,0.1992117763,-0.2083067745,-0.0957768857,0.0069104712,0.4759882987,-0.0078224344,-0.0339511037,0.4135632217,0.0027141657,0.1789693832,0.2403351814,0.1356374919,0.019620385,0.7002619505,-0.2602286339,-0.0039646165,-0.2352686673,0.3014819026,-0.3278697431,-0.7641029358,0.1616647542,0.2355505377,-0.3533098698,-0.1676512659,0.1258980781,0.6017500162,-0.0233745389,0.0114930086,-0.1145932153,-0.0448354296,0.0140551319,0.0483895428,-0.0662138388,-0.2536709607,-0.0708237216,0.1856833249,-0.0595556647,0.0145596312,0.2749812305,0.1876641512,-0.1494839638,-0.4457591772,0.0657755956,0.1728491187,0.0665428638,-0.2748388648,0.3666117489,0.2653487027,-0.0859915987,0.1805052459,-0.0413444005,0.6402472854,0.504191637,0.0924113542,-0.0617402829,-0.1215721816,-0.1139907762,-0.1794179529,0.1778173298,0.3488176763,0.4212691784,0.1485440582,0.1135984808,-0.0401850268,0.316088289,0.2067891061,0.0996007472,-0.545976758,-0.1189198121,-0.2083118558,-0.2577385902,-0.2266330868,0.0241517406,-0.2890593708,0.0209434181,0.4277476966,0.0739154965,0.2772270143,-0.02682258,0.0345675685,0.0743338466,0.3142196834,0.239776656,0.40228948,-0.4218083918,-0.0354187004,-0.7509756088,0.1165279523,-0.3764895499,0.009592602,-0.0318588242,0.1585383266,0.1377418339,0.0838609263,0.1591721773,0.2994238138,0.116723083,0.2220783383,-0.0561847016,-0.1846203357,-0.3380708694,0.0264856666,-0.1750924289,-0.4655142426,0.2213603109,-0.1279327124,0.009210824,-0.2251862437,0.2239884585,0.2559674084,-0.0059870374,0.4777292609,-0.1120114401,0.6255970597,-0.0850180835,0.3190019131,-0.3009388149,-0.0801894665,-0.1928057671,0.5200991035,0.2061499506,0.271399498,-0.177993089,-0.107916005,-0.2153041661,0.1574100554,-0.0065483749,-0.0705471635,0.0667275563,-0.1510605514,-0.0018241928,-0.058490295,0.1522316486,0.0920479745,-0.2229390591,-0.0781728476,-0.1745537966,-0.180185318,0.1810824573,-0.4043286145,-0.3411037624,0.1487179548,0.1261260211,0.1018662825,0.1396923661,-0.4524213076,0.1217898354,0.424475193,-0.1280173361,-0.4174263477,0.1340316832,-0.2116182446,0.1005702689,-0.1218523532,0.4765018821,0.1411091983,-0.370509088,-0.1615719497,-0.3583221138]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/836","title":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas","comments":"Thanks for the fast response. I have the latest version '2.0.0' (I tried to update)\r\nI am working with Python 3.8.5","body":"Hi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n","comment_length":21,"text":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas\nHi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n\nThanks for the fast response. I have the latest version '2.0.0' (I tried to update)\r\nI am working with Python 3.8.5","embeddings":[-0.24118343,-0.3198650777,-0.0681558475,0.4316074848,0.3625251055,0.0191550385,0.5210206509,0.437713474,0.2949123979,0.0327753536,0.0169624221,-0.1187045723,0.0905919969,0.1802892983,0.0350663252,0.1130092442,0.0730282664,0.3988213837,0.0317200907,0.0699720979,-0.218860805,0.1208694875,0.005867396,0.0575151071,-0.1244225353,0.0723639429,-0.008208585,0.1169146225,-0.2343517244,-0.4013206661,0.4197234511,0.0784542188,0.3328505158,0.5207661986,-0.0001169051,0.1564733684,0.2654486001,-0.1856259704,-0.0441993587,-0.5195806026,0.0301294662,-0.5329574347,0.2776663601,-0.1672024429,-0.0144887799,-0.1786803156,-0.1109988168,-0.0358052403,0.3014124632,0.3175454736,0.1804958135,-0.2607497573,0.0598921739,-0.0100186532,0.2052922398,0.0371788666,-0.0156855155,0.4175470769,0.1173958927,-0.1267608404,-0.0487308204,-0.193666935,-0.2392427474,0.3581996262,0.3196350932,0.1399399787,0.0709672496,-0.1089668348,0.2680699527,0.1055052057,0.6067355871,-0.307642132,-0.1485474855,-0.0154003808,-0.0202969182,-0.4886516333,0.2558154762,0.404494822,-0.1521599442,0.1173331738,-0.2122125626,0.2484468818,-0.2663929462,0.2028620988,-0.0673474893,-0.3531901836,-0.2522074878,0.2693178952,0.2523397803,0.0737477168,0.3336401582,-0.0937501863,0.0268538464,0.1806966811,-0.5556491017,0.2796493471,-0.1176608056,-0.0904364362,0.1637407988,0.096218653,0.4014219046,0.1370620281,0.0423445031,0.1946978569,0.3191299736,0.1861328632,-0.0428146571,0.0691050887,0.0273857322,0.3462172449,-0.2297836244,-0.2338579148,-0.3881369233,-0.5049571991,0.2381427437,-0.1538383663,0.2674439549,-0.2704333961,-0.2479329556,-0.0960200727,-0.0361324735,-0.0094233667,-0.0528344922,0.5738207698,0.215833962,0.3802042603,-0.054209996,0.2572092116,-0.2112757564,-0.09865468,-0.0803525597,0.0201157145,-0.0576021709,0.0522914641,0.3777484894,-0.2166886032,0.1214607805,0.1762953401,-0.043517001,-0.2938742042,-0.1455057859,-0.4177222252,0.0367241055,0.283308208,0.2024739832,0.0979529768,0.2341527939,-0.2812092602,-0.0606879592,0.3805413246,-0.2301190495,-0.1685689539,0.0972641706,0.0963025913,-0.0661173835,0.1656782478,-0.3956311643,0.1563510746,0.3519503474,-0.4384936988,0.0334676392,-0.2134845853,-0.2411233634,-0.3524810374,0.2018900961,0.2790597677,-0.7854661942,0.3429474831,0.0539065748,-0.0317488685,0.1149988398,0.4237150252,-0.1483171433,0.0611786693,-0.2709869146,0.0921530128,0.1406241804,0.007542707,-0.6293674707,0.3307564855,-0.04602734,0.0895455927,0.3120644391,0.0645528063,0.0512152202,0.0891771689,-0.1080223322,0.0185540114,0.0231345072,0.0440054424,-0.2607476711,-0.1291924864,0.3187506497,0.142625019,-0.0647743195,-0.0676729232,0.2388889641,-0.4087391198,0.2025453746,-0.0180878602,0.1416161656,0.1752468795,0.1791311353,-0.1968236417,0.0349740647,0.0104333451,-0.4493494034,0.1433825642,0.2848802209,-0.1687007099,-0.2159446925,-0.0797307715,-0.384118408,-0.0297010913,-0.2681604922,-0.1241994351,-0.0246524885,0.0725153983,0.0595197789,-0.1440529078,-0.1857402176,0.1161194369,-0.2237771302,-0.0112759126,-0.0158436764,0.1422190964,0.058799319,-0.1249890551,0.1064720228,-0.0629007593,0.207167998,-0.0483339541,-0.2763522565,0.2495912611,0.1456864476,0.0456155054,-0.2857958674,-0.1747876853,0.0830201879,-0.2415053844,-0.0872012973,0.0237385109,0.3663716614,-0.1825469434,-0.3392808735,0.2706229687,-0.1567003578,0.3015136421,0.1879752129,-0.3580714464,0.3065021634,0.016720105,0.1781937331,0.0704920888,0.1380855441,0.0736546293,0.4947431087,0.2995411456,-0.0968417376,-0.1562502384,0.5009906292,-0.0470440127,-0.2185632885,0.3112592995,-0.1181778163,-0.1945351362,0.0684309453,0.0887451172,0.3798215091,0.1293763965,-0.1915307492,0.0146643994,0.1435023844,-0.2500295639,0.1204057485,-0.2204343826,0.1473618895,0.3965650201,0.1270093769,-0.1272856295,-0.195144549,0.0377438515,0.0875855386,0.3257235587,-0.4119557142,-0.0675459281,-0.0734540001,-0.0052061281,-0.1802845895,0.0986031592,-0.0555794202,0.0971230865,-0.3996109366,0.3328239024,0.2897825539,-0.0790665746,-0.2431259155,0.08863388,0.2158128768,-0.3370558321,-0.0034607018,-0.0489910357,-0.2919403613,-0.0261766538,0.3307105303,0.1098190099,-0.0008771374,-0.0217662286,0.0280251838,-0.0548448637,-0.0925327018,-0.0439247601,-0.0732297674,0.0209216662,-0.0511663184,0.4193309844,0.0076079252,-0.1446308047,0.4017484784,-0.2536608279,-0.0489250831,0.350135982,0.0636930913,-0.0958797708,-0.1087071598,-0.3498898149,-0.139888227,-0.4293371439,0.3910627365,0.0847561061,0.0419315621,0.0926315561,0.4331049323,-0.0260847546,0.3013546467,-0.0312164128,-0.0848050937,-0.2977054417,0.3645742834,-0.0561163239,-0.4676565826,0.011941432,0.0828354955,0.3591901362,0.1638266891,-0.4252471328,0.3267471194,-0.15090774,0.1628198922,0.0011049408,-0.0242274385,0.0609546043,-0.0872276798,-0.0218234472,0.0364301093,-0.2417160571,-0.0141803464,-0.0320445746,0.1475835294,0.0513685606,0.458429724,-0.0371929593,0.6513361931,0.0874645486,-0.0882313102,0.4855556488,-0.308239907,0.3818774223,-0.2096708715,-0.3737454712,-0.0855747312,-0.3519817293,-0.1924462765,0.2677486241,-0.0335841477,-0.0334150307,-0.1257756203,0.0035179411,-0.4046861231,-0.0377031192,0.183632195,0.0306968763,-0.0148301804,-0.1005954891,-0.0133011667,-0.0116456645,-0.237362951,-0.1942582428,0.2739638686,-0.1513422579,0.0633870587,-0.2242484242,-0.0615803823,-0.3029371798,0.3292051852,-0.0136234136,0.3102864921,0.0106146866,-0.0176951028,0.0972969308,-0.1252222657,0.7449785471,0.1892860085,-0.1550699919,0.1278296113,-0.0241136737,-0.441583544,-0.121761322,0.1906314343,0.3073361516,0.0744616762,0.3548655212,-0.0703261346,-0.0096933683,-0.1443433017,0.3546780944,0.0432527624,-0.3951280713,-0.5194913745,-0.2349791825,-0.383053571,-0.2121350318,-0.0898666456,0.2882787883,-0.3244074285,-0.1423869133,0.1183612794,-0.2990536094,0.0601567216,0.1533810496,0.2973894477,-0.1129143313,0.1448963135,0.0468351059,0.0698044524,0.376911819,0.7987186313,-0.0785759091,-0.6220775843,0.0848431289,-0.2050266713,0.3853479028,0.0478198007,-0.2032050937,0.1289862841,0.0907342061,0.0620335788,0.0379479825,0.0144376121,0.3036864102,0.0496549308,-0.3454539776,-0.5439224243,0.4874189198,-0.1259078234,0.2102593035,0.302079767,-0.2119999677,-0.1099810228,0.1555226296,-0.3346246481,0.6592135429,-0.4218300581,0.1938303411,0.3073363602,0.056336008,0.2584289312,-0.4451968074,0.1323073953,-0.3919684887,-0.210903883,-0.1043123081,-0.0712032765,0.2393857986,0.432190001,-0.1212453321,0.3118637204,-0.0741497204,0.3325560987,0.11313425,0.2438252568,-0.1005161256,-0.0864683241,-0.3010203838,0.0104239564,-0.1430807561,-0.0349519551,0.028959142,0.0544647723,-0.1805675626,-0.2016452104,-0.2461119592,0.2200246751,-0.2911679149,0.1154046133,-0.1109549478,-0.4686649442,0.1963640451,0.2380675226,-0.2087023854,-0.0851794481,-0.1240850538,0.1550540775,0.1596347541,-0.0401756726,-0.2083743513,-0.0536952429,-0.0158856176,0.0761261061,-0.2351237983,0.1545512527,-0.122278817,-0.0956747606,0.2780734897,0.2376574725,-0.0848071724,-0.5452085137,-0.5326197147,-0.4284158647,0.0759811029,-0.1957320869,0.0651982948,-0.0420785286,0.0983413011,-0.1616345942,0.1859346777,-0.4688693583,-0.1020267904,0.3918125927,-0.1116865948,0.0939545855,0.5907301903,0.2193497866,-0.0168925617,-0.1411525458,0.0642407984,0.3081057072,-0.3531885743,0.1170660481,0.1141491011,0.3159501255,0.1392528862,0.2917454839,0.1565803736,-0.0776725486,0.0744550973,-0.5115905404,-0.3249694705,-0.0165290646,-0.0015315027,0.2479578257,0.1672193259,0.08186692,0.1726779193,0.2032965571,-0.1562302262,0.1004625708,0.1059293076,-0.0869575143,0.2106330097,0.1420163661,-0.1338726133,-0.1038192213,0.0934978351,0.1776407212,-0.0534449145,-0.1887511909,-0.0798553601,0.1500999779,0.0983605087,0.0232860353,-0.2034453899,-0.2526888847,-0.1497063339,-0.19280532,0.0664090961,0.0297824815,-0.0981118679,0.0717605129,-0.2987648547,-0.0846538544,-0.5091862082,0.0271216799,-0.5075390935,0.2231322825,0.29926911,0.0485824719,0.1217573285,0.2004558295,-0.1724265963,-0.0311586745,0.1516266763,-0.1366193742,0.3009471297,-0.4014762044,0.2035068721,-0.0510654896,0.3403376043,0.4848605692,-0.2076416463,0.2197109312,0.0375863202,0.1341150701,-0.3825282753,0.2754767835,0.1172315627,-0.0616297424,-0.027484389,0.1696746349,0.1992117763,-0.2083067745,-0.0957768857,0.0069104712,0.4759882987,-0.0078224344,-0.0339511037,0.4135632217,0.0027141657,0.1789693832,0.2403351814,0.1356374919,0.019620385,0.7002619505,-0.2602286339,-0.0039646165,-0.2352686673,0.3014819026,-0.3278697431,-0.7641029358,0.1616647542,0.2355505377,-0.3533098698,-0.1676512659,0.1258980781,0.6017500162,-0.0233745389,0.0114930086,-0.1145932153,-0.0448354296,0.0140551319,0.0483895428,-0.0662138388,-0.2536709607,-0.0708237216,0.1856833249,-0.0595556647,0.0145596312,0.2749812305,0.1876641512,-0.1494839638,-0.4457591772,0.0657755956,0.1728491187,0.0665428638,-0.2748388648,0.3666117489,0.2653487027,-0.0859915987,0.1805052459,-0.0413444005,0.6402472854,0.504191637,0.0924113542,-0.0617402829,-0.1215721816,-0.1139907762,-0.1794179529,0.1778173298,0.3488176763,0.4212691784,0.1485440582,0.1135984808,-0.0401850268,0.316088289,0.2067891061,0.0996007472,-0.545976758,-0.1189198121,-0.2083118558,-0.2577385902,-0.2266330868,0.0241517406,-0.2890593708,0.0209434181,0.4277476966,0.0739154965,0.2772270143,-0.02682258,0.0345675685,0.0743338466,0.3142196834,0.239776656,0.40228948,-0.4218083918,-0.0354187004,-0.7509756088,0.1165279523,-0.3764895499,0.009592602,-0.0318588242,0.1585383266,0.1377418339,0.0838609263,0.1591721773,0.2994238138,0.116723083,0.2220783383,-0.0561847016,-0.1846203357,-0.3380708694,0.0264856666,-0.1750924289,-0.4655142426,0.2213603109,-0.1279327124,0.009210824,-0.2251862437,0.2239884585,0.2559674084,-0.0059870374,0.4777292609,-0.1120114401,0.6255970597,-0.0850180835,0.3190019131,-0.3009388149,-0.0801894665,-0.1928057671,0.5200991035,0.2061499506,0.271399498,-0.177993089,-0.107916005,-0.2153041661,0.1574100554,-0.0065483749,-0.0705471635,0.0667275563,-0.1510605514,-0.0018241928,-0.058490295,0.1522316486,0.0920479745,-0.2229390591,-0.0781728476,-0.1745537966,-0.180185318,0.1810824573,-0.4043286145,-0.3411037624,0.1487179548,0.1261260211,0.1018662825,0.1396923661,-0.4524213076,0.1217898354,0.424475193,-0.1280173361,-0.4174263477,0.1340316832,-0.2116182446,0.1005702689,-0.1218523532,0.4765018821,0.1411091983,-0.370509088,-0.1615719497,-0.3583221138]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/836","title":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas","comments":"I think that the issue is similar to this one:https:\/\/issues.apache.org\/jira\/browse\/ARROW-9612\r\nThe problem is in arrow when the column data contains long strings.\r\nAny ideas on how to bypass this?","body":"Hi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n","comment_length":29,"text":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas\nHi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n\nI think that the issue is similar to this one:https:\/\/issues.apache.org\/jira\/browse\/ARROW-9612\r\nThe problem is in arrow when the column data contains long strings.\r\nAny ideas on how to bypass this?","embeddings":[-0.24118343,-0.3198650777,-0.0681558475,0.4316074848,0.3625251055,0.0191550385,0.5210206509,0.437713474,0.2949123979,0.0327753536,0.0169624221,-0.1187045723,0.0905919969,0.1802892983,0.0350663252,0.1130092442,0.0730282664,0.3988213837,0.0317200907,0.0699720979,-0.218860805,0.1208694875,0.005867396,0.0575151071,-0.1244225353,0.0723639429,-0.008208585,0.1169146225,-0.2343517244,-0.4013206661,0.4197234511,0.0784542188,0.3328505158,0.5207661986,-0.0001169051,0.1564733684,0.2654486001,-0.1856259704,-0.0441993587,-0.5195806026,0.0301294662,-0.5329574347,0.2776663601,-0.1672024429,-0.0144887799,-0.1786803156,-0.1109988168,-0.0358052403,0.3014124632,0.3175454736,0.1804958135,-0.2607497573,0.0598921739,-0.0100186532,0.2052922398,0.0371788666,-0.0156855155,0.4175470769,0.1173958927,-0.1267608404,-0.0487308204,-0.193666935,-0.2392427474,0.3581996262,0.3196350932,0.1399399787,0.0709672496,-0.1089668348,0.2680699527,0.1055052057,0.6067355871,-0.307642132,-0.1485474855,-0.0154003808,-0.0202969182,-0.4886516333,0.2558154762,0.404494822,-0.1521599442,0.1173331738,-0.2122125626,0.2484468818,-0.2663929462,0.2028620988,-0.0673474893,-0.3531901836,-0.2522074878,0.2693178952,0.2523397803,0.0737477168,0.3336401582,-0.0937501863,0.0268538464,0.1806966811,-0.5556491017,0.2796493471,-0.1176608056,-0.0904364362,0.1637407988,0.096218653,0.4014219046,0.1370620281,0.0423445031,0.1946978569,0.3191299736,0.1861328632,-0.0428146571,0.0691050887,0.0273857322,0.3462172449,-0.2297836244,-0.2338579148,-0.3881369233,-0.5049571991,0.2381427437,-0.1538383663,0.2674439549,-0.2704333961,-0.2479329556,-0.0960200727,-0.0361324735,-0.0094233667,-0.0528344922,0.5738207698,0.215833962,0.3802042603,-0.054209996,0.2572092116,-0.2112757564,-0.09865468,-0.0803525597,0.0201157145,-0.0576021709,0.0522914641,0.3777484894,-0.2166886032,0.1214607805,0.1762953401,-0.043517001,-0.2938742042,-0.1455057859,-0.4177222252,0.0367241055,0.283308208,0.2024739832,0.0979529768,0.2341527939,-0.2812092602,-0.0606879592,0.3805413246,-0.2301190495,-0.1685689539,0.0972641706,0.0963025913,-0.0661173835,0.1656782478,-0.3956311643,0.1563510746,0.3519503474,-0.4384936988,0.0334676392,-0.2134845853,-0.2411233634,-0.3524810374,0.2018900961,0.2790597677,-0.7854661942,0.3429474831,0.0539065748,-0.0317488685,0.1149988398,0.4237150252,-0.1483171433,0.0611786693,-0.2709869146,0.0921530128,0.1406241804,0.007542707,-0.6293674707,0.3307564855,-0.04602734,0.0895455927,0.3120644391,0.0645528063,0.0512152202,0.0891771689,-0.1080223322,0.0185540114,0.0231345072,0.0440054424,-0.2607476711,-0.1291924864,0.3187506497,0.142625019,-0.0647743195,-0.0676729232,0.2388889641,-0.4087391198,0.2025453746,-0.0180878602,0.1416161656,0.1752468795,0.1791311353,-0.1968236417,0.0349740647,0.0104333451,-0.4493494034,0.1433825642,0.2848802209,-0.1687007099,-0.2159446925,-0.0797307715,-0.384118408,-0.0297010913,-0.2681604922,-0.1241994351,-0.0246524885,0.0725153983,0.0595197789,-0.1440529078,-0.1857402176,0.1161194369,-0.2237771302,-0.0112759126,-0.0158436764,0.1422190964,0.058799319,-0.1249890551,0.1064720228,-0.0629007593,0.207167998,-0.0483339541,-0.2763522565,0.2495912611,0.1456864476,0.0456155054,-0.2857958674,-0.1747876853,0.0830201879,-0.2415053844,-0.0872012973,0.0237385109,0.3663716614,-0.1825469434,-0.3392808735,0.2706229687,-0.1567003578,0.3015136421,0.1879752129,-0.3580714464,0.3065021634,0.016720105,0.1781937331,0.0704920888,0.1380855441,0.0736546293,0.4947431087,0.2995411456,-0.0968417376,-0.1562502384,0.5009906292,-0.0470440127,-0.2185632885,0.3112592995,-0.1181778163,-0.1945351362,0.0684309453,0.0887451172,0.3798215091,0.1293763965,-0.1915307492,0.0146643994,0.1435023844,-0.2500295639,0.1204057485,-0.2204343826,0.1473618895,0.3965650201,0.1270093769,-0.1272856295,-0.195144549,0.0377438515,0.0875855386,0.3257235587,-0.4119557142,-0.0675459281,-0.0734540001,-0.0052061281,-0.1802845895,0.0986031592,-0.0555794202,0.0971230865,-0.3996109366,0.3328239024,0.2897825539,-0.0790665746,-0.2431259155,0.08863388,0.2158128768,-0.3370558321,-0.0034607018,-0.0489910357,-0.2919403613,-0.0261766538,0.3307105303,0.1098190099,-0.0008771374,-0.0217662286,0.0280251838,-0.0548448637,-0.0925327018,-0.0439247601,-0.0732297674,0.0209216662,-0.0511663184,0.4193309844,0.0076079252,-0.1446308047,0.4017484784,-0.2536608279,-0.0489250831,0.350135982,0.0636930913,-0.0958797708,-0.1087071598,-0.3498898149,-0.139888227,-0.4293371439,0.3910627365,0.0847561061,0.0419315621,0.0926315561,0.4331049323,-0.0260847546,0.3013546467,-0.0312164128,-0.0848050937,-0.2977054417,0.3645742834,-0.0561163239,-0.4676565826,0.011941432,0.0828354955,0.3591901362,0.1638266891,-0.4252471328,0.3267471194,-0.15090774,0.1628198922,0.0011049408,-0.0242274385,0.0609546043,-0.0872276798,-0.0218234472,0.0364301093,-0.2417160571,-0.0141803464,-0.0320445746,0.1475835294,0.0513685606,0.458429724,-0.0371929593,0.6513361931,0.0874645486,-0.0882313102,0.4855556488,-0.308239907,0.3818774223,-0.2096708715,-0.3737454712,-0.0855747312,-0.3519817293,-0.1924462765,0.2677486241,-0.0335841477,-0.0334150307,-0.1257756203,0.0035179411,-0.4046861231,-0.0377031192,0.183632195,0.0306968763,-0.0148301804,-0.1005954891,-0.0133011667,-0.0116456645,-0.237362951,-0.1942582428,0.2739638686,-0.1513422579,0.0633870587,-0.2242484242,-0.0615803823,-0.3029371798,0.3292051852,-0.0136234136,0.3102864921,0.0106146866,-0.0176951028,0.0972969308,-0.1252222657,0.7449785471,0.1892860085,-0.1550699919,0.1278296113,-0.0241136737,-0.441583544,-0.121761322,0.1906314343,0.3073361516,0.0744616762,0.3548655212,-0.0703261346,-0.0096933683,-0.1443433017,0.3546780944,0.0432527624,-0.3951280713,-0.5194913745,-0.2349791825,-0.383053571,-0.2121350318,-0.0898666456,0.2882787883,-0.3244074285,-0.1423869133,0.1183612794,-0.2990536094,0.0601567216,0.1533810496,0.2973894477,-0.1129143313,0.1448963135,0.0468351059,0.0698044524,0.376911819,0.7987186313,-0.0785759091,-0.6220775843,0.0848431289,-0.2050266713,0.3853479028,0.0478198007,-0.2032050937,0.1289862841,0.0907342061,0.0620335788,0.0379479825,0.0144376121,0.3036864102,0.0496549308,-0.3454539776,-0.5439224243,0.4874189198,-0.1259078234,0.2102593035,0.302079767,-0.2119999677,-0.1099810228,0.1555226296,-0.3346246481,0.6592135429,-0.4218300581,0.1938303411,0.3073363602,0.056336008,0.2584289312,-0.4451968074,0.1323073953,-0.3919684887,-0.210903883,-0.1043123081,-0.0712032765,0.2393857986,0.432190001,-0.1212453321,0.3118637204,-0.0741497204,0.3325560987,0.11313425,0.2438252568,-0.1005161256,-0.0864683241,-0.3010203838,0.0104239564,-0.1430807561,-0.0349519551,0.028959142,0.0544647723,-0.1805675626,-0.2016452104,-0.2461119592,0.2200246751,-0.2911679149,0.1154046133,-0.1109549478,-0.4686649442,0.1963640451,0.2380675226,-0.2087023854,-0.0851794481,-0.1240850538,0.1550540775,0.1596347541,-0.0401756726,-0.2083743513,-0.0536952429,-0.0158856176,0.0761261061,-0.2351237983,0.1545512527,-0.122278817,-0.0956747606,0.2780734897,0.2376574725,-0.0848071724,-0.5452085137,-0.5326197147,-0.4284158647,0.0759811029,-0.1957320869,0.0651982948,-0.0420785286,0.0983413011,-0.1616345942,0.1859346777,-0.4688693583,-0.1020267904,0.3918125927,-0.1116865948,0.0939545855,0.5907301903,0.2193497866,-0.0168925617,-0.1411525458,0.0642407984,0.3081057072,-0.3531885743,0.1170660481,0.1141491011,0.3159501255,0.1392528862,0.2917454839,0.1565803736,-0.0776725486,0.0744550973,-0.5115905404,-0.3249694705,-0.0165290646,-0.0015315027,0.2479578257,0.1672193259,0.08186692,0.1726779193,0.2032965571,-0.1562302262,0.1004625708,0.1059293076,-0.0869575143,0.2106330097,0.1420163661,-0.1338726133,-0.1038192213,0.0934978351,0.1776407212,-0.0534449145,-0.1887511909,-0.0798553601,0.1500999779,0.0983605087,0.0232860353,-0.2034453899,-0.2526888847,-0.1497063339,-0.19280532,0.0664090961,0.0297824815,-0.0981118679,0.0717605129,-0.2987648547,-0.0846538544,-0.5091862082,0.0271216799,-0.5075390935,0.2231322825,0.29926911,0.0485824719,0.1217573285,0.2004558295,-0.1724265963,-0.0311586745,0.1516266763,-0.1366193742,0.3009471297,-0.4014762044,0.2035068721,-0.0510654896,0.3403376043,0.4848605692,-0.2076416463,0.2197109312,0.0375863202,0.1341150701,-0.3825282753,0.2754767835,0.1172315627,-0.0616297424,-0.027484389,0.1696746349,0.1992117763,-0.2083067745,-0.0957768857,0.0069104712,0.4759882987,-0.0078224344,-0.0339511037,0.4135632217,0.0027141657,0.1789693832,0.2403351814,0.1356374919,0.019620385,0.7002619505,-0.2602286339,-0.0039646165,-0.2352686673,0.3014819026,-0.3278697431,-0.7641029358,0.1616647542,0.2355505377,-0.3533098698,-0.1676512659,0.1258980781,0.6017500162,-0.0233745389,0.0114930086,-0.1145932153,-0.0448354296,0.0140551319,0.0483895428,-0.0662138388,-0.2536709607,-0.0708237216,0.1856833249,-0.0595556647,0.0145596312,0.2749812305,0.1876641512,-0.1494839638,-0.4457591772,0.0657755956,0.1728491187,0.0665428638,-0.2748388648,0.3666117489,0.2653487027,-0.0859915987,0.1805052459,-0.0413444005,0.6402472854,0.504191637,0.0924113542,-0.0617402829,-0.1215721816,-0.1139907762,-0.1794179529,0.1778173298,0.3488176763,0.4212691784,0.1485440582,0.1135984808,-0.0401850268,0.316088289,0.2067891061,0.0996007472,-0.545976758,-0.1189198121,-0.2083118558,-0.2577385902,-0.2266330868,0.0241517406,-0.2890593708,0.0209434181,0.4277476966,0.0739154965,0.2772270143,-0.02682258,0.0345675685,0.0743338466,0.3142196834,0.239776656,0.40228948,-0.4218083918,-0.0354187004,-0.7509756088,0.1165279523,-0.3764895499,0.009592602,-0.0318588242,0.1585383266,0.1377418339,0.0838609263,0.1591721773,0.2994238138,0.116723083,0.2220783383,-0.0561847016,-0.1846203357,-0.3380708694,0.0264856666,-0.1750924289,-0.4655142426,0.2213603109,-0.1279327124,0.009210824,-0.2251862437,0.2239884585,0.2559674084,-0.0059870374,0.4777292609,-0.1120114401,0.6255970597,-0.0850180835,0.3190019131,-0.3009388149,-0.0801894665,-0.1928057671,0.5200991035,0.2061499506,0.271399498,-0.177993089,-0.107916005,-0.2153041661,0.1574100554,-0.0065483749,-0.0705471635,0.0667275563,-0.1510605514,-0.0018241928,-0.058490295,0.1522316486,0.0920479745,-0.2229390591,-0.0781728476,-0.1745537966,-0.180185318,0.1810824573,-0.4043286145,-0.3411037624,0.1487179548,0.1261260211,0.1018662825,0.1396923661,-0.4524213076,0.1217898354,0.424475193,-0.1280173361,-0.4174263477,0.1340316832,-0.2116182446,0.1005702689,-0.1218523532,0.4765018821,0.1411091983,-0.370509088,-0.1615719497,-0.3583221138]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/836","title":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas","comments":"We should expose the [`block_size` argument](https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.csv.ReadOptions.html#pyarrow.csv.ReadOptions) of Apache Arrow csv `ReadOptions` in the [script](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/csv\/csv.py).\r\n\r\n\r\nIn the meantime you can specify yourself the `ReadOptions` config like this:\r\n```python\r\nimport pyarrow.csv as pac # PyArrow is installed with `datasets`\r\n\r\nread_options = pac.ReadOptions(block_size=1e9) # try to find the right value for your use-case\r\ndataset = load_dataset('csv', data_files=files, read_options=read_options)\r\n```\r\n","body":"Hi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n","comment_length":56,"text":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas\nHi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n\nWe should expose the [`block_size` argument](https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.csv.ReadOptions.html#pyarrow.csv.ReadOptions) of Apache Arrow csv `ReadOptions` in the [script](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/csv\/csv.py).\r\n\r\n\r\nIn the meantime you can specify yourself the `ReadOptions` config like this:\r\n```python\r\nimport pyarrow.csv as pac # PyArrow is installed with `datasets`\r\n\r\nread_options = pac.ReadOptions(block_size=1e9) # try to find the right value for your use-case\r\ndataset = load_dataset('csv', data_files=files, read_options=read_options)\r\n```\r\n","embeddings":[-0.24118343,-0.3198650777,-0.0681558475,0.4316074848,0.3625251055,0.0191550385,0.5210206509,0.437713474,0.2949123979,0.0327753536,0.0169624221,-0.1187045723,0.0905919969,0.1802892983,0.0350663252,0.1130092442,0.0730282664,0.3988213837,0.0317200907,0.0699720979,-0.218860805,0.1208694875,0.005867396,0.0575151071,-0.1244225353,0.0723639429,-0.008208585,0.1169146225,-0.2343517244,-0.4013206661,0.4197234511,0.0784542188,0.3328505158,0.5207661986,-0.0001169051,0.1564733684,0.2654486001,-0.1856259704,-0.0441993587,-0.5195806026,0.0301294662,-0.5329574347,0.2776663601,-0.1672024429,-0.0144887799,-0.1786803156,-0.1109988168,-0.0358052403,0.3014124632,0.3175454736,0.1804958135,-0.2607497573,0.0598921739,-0.0100186532,0.2052922398,0.0371788666,-0.0156855155,0.4175470769,0.1173958927,-0.1267608404,-0.0487308204,-0.193666935,-0.2392427474,0.3581996262,0.3196350932,0.1399399787,0.0709672496,-0.1089668348,0.2680699527,0.1055052057,0.6067355871,-0.307642132,-0.1485474855,-0.0154003808,-0.0202969182,-0.4886516333,0.2558154762,0.404494822,-0.1521599442,0.1173331738,-0.2122125626,0.2484468818,-0.2663929462,0.2028620988,-0.0673474893,-0.3531901836,-0.2522074878,0.2693178952,0.2523397803,0.0737477168,0.3336401582,-0.0937501863,0.0268538464,0.1806966811,-0.5556491017,0.2796493471,-0.1176608056,-0.0904364362,0.1637407988,0.096218653,0.4014219046,0.1370620281,0.0423445031,0.1946978569,0.3191299736,0.1861328632,-0.0428146571,0.0691050887,0.0273857322,0.3462172449,-0.2297836244,-0.2338579148,-0.3881369233,-0.5049571991,0.2381427437,-0.1538383663,0.2674439549,-0.2704333961,-0.2479329556,-0.0960200727,-0.0361324735,-0.0094233667,-0.0528344922,0.5738207698,0.215833962,0.3802042603,-0.054209996,0.2572092116,-0.2112757564,-0.09865468,-0.0803525597,0.0201157145,-0.0576021709,0.0522914641,0.3777484894,-0.2166886032,0.1214607805,0.1762953401,-0.043517001,-0.2938742042,-0.1455057859,-0.4177222252,0.0367241055,0.283308208,0.2024739832,0.0979529768,0.2341527939,-0.2812092602,-0.0606879592,0.3805413246,-0.2301190495,-0.1685689539,0.0972641706,0.0963025913,-0.0661173835,0.1656782478,-0.3956311643,0.1563510746,0.3519503474,-0.4384936988,0.0334676392,-0.2134845853,-0.2411233634,-0.3524810374,0.2018900961,0.2790597677,-0.7854661942,0.3429474831,0.0539065748,-0.0317488685,0.1149988398,0.4237150252,-0.1483171433,0.0611786693,-0.2709869146,0.0921530128,0.1406241804,0.007542707,-0.6293674707,0.3307564855,-0.04602734,0.0895455927,0.3120644391,0.0645528063,0.0512152202,0.0891771689,-0.1080223322,0.0185540114,0.0231345072,0.0440054424,-0.2607476711,-0.1291924864,0.3187506497,0.142625019,-0.0647743195,-0.0676729232,0.2388889641,-0.4087391198,0.2025453746,-0.0180878602,0.1416161656,0.1752468795,0.1791311353,-0.1968236417,0.0349740647,0.0104333451,-0.4493494034,0.1433825642,0.2848802209,-0.1687007099,-0.2159446925,-0.0797307715,-0.384118408,-0.0297010913,-0.2681604922,-0.1241994351,-0.0246524885,0.0725153983,0.0595197789,-0.1440529078,-0.1857402176,0.1161194369,-0.2237771302,-0.0112759126,-0.0158436764,0.1422190964,0.058799319,-0.1249890551,0.1064720228,-0.0629007593,0.207167998,-0.0483339541,-0.2763522565,0.2495912611,0.1456864476,0.0456155054,-0.2857958674,-0.1747876853,0.0830201879,-0.2415053844,-0.0872012973,0.0237385109,0.3663716614,-0.1825469434,-0.3392808735,0.2706229687,-0.1567003578,0.3015136421,0.1879752129,-0.3580714464,0.3065021634,0.016720105,0.1781937331,0.0704920888,0.1380855441,0.0736546293,0.4947431087,0.2995411456,-0.0968417376,-0.1562502384,0.5009906292,-0.0470440127,-0.2185632885,0.3112592995,-0.1181778163,-0.1945351362,0.0684309453,0.0887451172,0.3798215091,0.1293763965,-0.1915307492,0.0146643994,0.1435023844,-0.2500295639,0.1204057485,-0.2204343826,0.1473618895,0.3965650201,0.1270093769,-0.1272856295,-0.195144549,0.0377438515,0.0875855386,0.3257235587,-0.4119557142,-0.0675459281,-0.0734540001,-0.0052061281,-0.1802845895,0.0986031592,-0.0555794202,0.0971230865,-0.3996109366,0.3328239024,0.2897825539,-0.0790665746,-0.2431259155,0.08863388,0.2158128768,-0.3370558321,-0.0034607018,-0.0489910357,-0.2919403613,-0.0261766538,0.3307105303,0.1098190099,-0.0008771374,-0.0217662286,0.0280251838,-0.0548448637,-0.0925327018,-0.0439247601,-0.0732297674,0.0209216662,-0.0511663184,0.4193309844,0.0076079252,-0.1446308047,0.4017484784,-0.2536608279,-0.0489250831,0.350135982,0.0636930913,-0.0958797708,-0.1087071598,-0.3498898149,-0.139888227,-0.4293371439,0.3910627365,0.0847561061,0.0419315621,0.0926315561,0.4331049323,-0.0260847546,0.3013546467,-0.0312164128,-0.0848050937,-0.2977054417,0.3645742834,-0.0561163239,-0.4676565826,0.011941432,0.0828354955,0.3591901362,0.1638266891,-0.4252471328,0.3267471194,-0.15090774,0.1628198922,0.0011049408,-0.0242274385,0.0609546043,-0.0872276798,-0.0218234472,0.0364301093,-0.2417160571,-0.0141803464,-0.0320445746,0.1475835294,0.0513685606,0.458429724,-0.0371929593,0.6513361931,0.0874645486,-0.0882313102,0.4855556488,-0.308239907,0.3818774223,-0.2096708715,-0.3737454712,-0.0855747312,-0.3519817293,-0.1924462765,0.2677486241,-0.0335841477,-0.0334150307,-0.1257756203,0.0035179411,-0.4046861231,-0.0377031192,0.183632195,0.0306968763,-0.0148301804,-0.1005954891,-0.0133011667,-0.0116456645,-0.237362951,-0.1942582428,0.2739638686,-0.1513422579,0.0633870587,-0.2242484242,-0.0615803823,-0.3029371798,0.3292051852,-0.0136234136,0.3102864921,0.0106146866,-0.0176951028,0.0972969308,-0.1252222657,0.7449785471,0.1892860085,-0.1550699919,0.1278296113,-0.0241136737,-0.441583544,-0.121761322,0.1906314343,0.3073361516,0.0744616762,0.3548655212,-0.0703261346,-0.0096933683,-0.1443433017,0.3546780944,0.0432527624,-0.3951280713,-0.5194913745,-0.2349791825,-0.383053571,-0.2121350318,-0.0898666456,0.2882787883,-0.3244074285,-0.1423869133,0.1183612794,-0.2990536094,0.0601567216,0.1533810496,0.2973894477,-0.1129143313,0.1448963135,0.0468351059,0.0698044524,0.376911819,0.7987186313,-0.0785759091,-0.6220775843,0.0848431289,-0.2050266713,0.3853479028,0.0478198007,-0.2032050937,0.1289862841,0.0907342061,0.0620335788,0.0379479825,0.0144376121,0.3036864102,0.0496549308,-0.3454539776,-0.5439224243,0.4874189198,-0.1259078234,0.2102593035,0.302079767,-0.2119999677,-0.1099810228,0.1555226296,-0.3346246481,0.6592135429,-0.4218300581,0.1938303411,0.3073363602,0.056336008,0.2584289312,-0.4451968074,0.1323073953,-0.3919684887,-0.210903883,-0.1043123081,-0.0712032765,0.2393857986,0.432190001,-0.1212453321,0.3118637204,-0.0741497204,0.3325560987,0.11313425,0.2438252568,-0.1005161256,-0.0864683241,-0.3010203838,0.0104239564,-0.1430807561,-0.0349519551,0.028959142,0.0544647723,-0.1805675626,-0.2016452104,-0.2461119592,0.2200246751,-0.2911679149,0.1154046133,-0.1109549478,-0.4686649442,0.1963640451,0.2380675226,-0.2087023854,-0.0851794481,-0.1240850538,0.1550540775,0.1596347541,-0.0401756726,-0.2083743513,-0.0536952429,-0.0158856176,0.0761261061,-0.2351237983,0.1545512527,-0.122278817,-0.0956747606,0.2780734897,0.2376574725,-0.0848071724,-0.5452085137,-0.5326197147,-0.4284158647,0.0759811029,-0.1957320869,0.0651982948,-0.0420785286,0.0983413011,-0.1616345942,0.1859346777,-0.4688693583,-0.1020267904,0.3918125927,-0.1116865948,0.0939545855,0.5907301903,0.2193497866,-0.0168925617,-0.1411525458,0.0642407984,0.3081057072,-0.3531885743,0.1170660481,0.1141491011,0.3159501255,0.1392528862,0.2917454839,0.1565803736,-0.0776725486,0.0744550973,-0.5115905404,-0.3249694705,-0.0165290646,-0.0015315027,0.2479578257,0.1672193259,0.08186692,0.1726779193,0.2032965571,-0.1562302262,0.1004625708,0.1059293076,-0.0869575143,0.2106330097,0.1420163661,-0.1338726133,-0.1038192213,0.0934978351,0.1776407212,-0.0534449145,-0.1887511909,-0.0798553601,0.1500999779,0.0983605087,0.0232860353,-0.2034453899,-0.2526888847,-0.1497063339,-0.19280532,0.0664090961,0.0297824815,-0.0981118679,0.0717605129,-0.2987648547,-0.0846538544,-0.5091862082,0.0271216799,-0.5075390935,0.2231322825,0.29926911,0.0485824719,0.1217573285,0.2004558295,-0.1724265963,-0.0311586745,0.1516266763,-0.1366193742,0.3009471297,-0.4014762044,0.2035068721,-0.0510654896,0.3403376043,0.4848605692,-0.2076416463,0.2197109312,0.0375863202,0.1341150701,-0.3825282753,0.2754767835,0.1172315627,-0.0616297424,-0.027484389,0.1696746349,0.1992117763,-0.2083067745,-0.0957768857,0.0069104712,0.4759882987,-0.0078224344,-0.0339511037,0.4135632217,0.0027141657,0.1789693832,0.2403351814,0.1356374919,0.019620385,0.7002619505,-0.2602286339,-0.0039646165,-0.2352686673,0.3014819026,-0.3278697431,-0.7641029358,0.1616647542,0.2355505377,-0.3533098698,-0.1676512659,0.1258980781,0.6017500162,-0.0233745389,0.0114930086,-0.1145932153,-0.0448354296,0.0140551319,0.0483895428,-0.0662138388,-0.2536709607,-0.0708237216,0.1856833249,-0.0595556647,0.0145596312,0.2749812305,0.1876641512,-0.1494839638,-0.4457591772,0.0657755956,0.1728491187,0.0665428638,-0.2748388648,0.3666117489,0.2653487027,-0.0859915987,0.1805052459,-0.0413444005,0.6402472854,0.504191637,0.0924113542,-0.0617402829,-0.1215721816,-0.1139907762,-0.1794179529,0.1778173298,0.3488176763,0.4212691784,0.1485440582,0.1135984808,-0.0401850268,0.316088289,0.2067891061,0.0996007472,-0.545976758,-0.1189198121,-0.2083118558,-0.2577385902,-0.2266330868,0.0241517406,-0.2890593708,0.0209434181,0.4277476966,0.0739154965,0.2772270143,-0.02682258,0.0345675685,0.0743338466,0.3142196834,0.239776656,0.40228948,-0.4218083918,-0.0354187004,-0.7509756088,0.1165279523,-0.3764895499,0.009592602,-0.0318588242,0.1585383266,0.1377418339,0.0838609263,0.1591721773,0.2994238138,0.116723083,0.2220783383,-0.0561847016,-0.1846203357,-0.3380708694,0.0264856666,-0.1750924289,-0.4655142426,0.2213603109,-0.1279327124,0.009210824,-0.2251862437,0.2239884585,0.2559674084,-0.0059870374,0.4777292609,-0.1120114401,0.6255970597,-0.0850180835,0.3190019131,-0.3009388149,-0.0801894665,-0.1928057671,0.5200991035,0.2061499506,0.271399498,-0.177993089,-0.107916005,-0.2153041661,0.1574100554,-0.0065483749,-0.0705471635,0.0667275563,-0.1510605514,-0.0018241928,-0.058490295,0.1522316486,0.0920479745,-0.2229390591,-0.0781728476,-0.1745537966,-0.180185318,0.1810824573,-0.4043286145,-0.3411037624,0.1487179548,0.1261260211,0.1018662825,0.1396923661,-0.4524213076,0.1217898354,0.424475193,-0.1280173361,-0.4174263477,0.1340316832,-0.2116182446,0.1005702689,-0.1218523532,0.4765018821,0.1411091983,-0.370509088,-0.1615719497,-0.3583221138]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/836","title":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas","comments":"This did help to load the data. But the problem now is that I get:\r\nArrowInvalid: CSV parse error: Expected 5 columns, got 187\r\n\r\nIt seems that this change the parsing so I changed the table to tab-separated and tried to load it directly from pyarrow\r\nBut I got a similar error, again it loaded fine in pandas so I am not sure what to do.\r\n\r\n\r\n\r\n","body":"Hi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n","comment_length":66,"text":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas\nHi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n\nThis did help to load the data. But the problem now is that I get:\r\nArrowInvalid: CSV parse error: Expected 5 columns, got 187\r\n\r\nIt seems that this change the parsing so I changed the table to tab-separated and tried to load it directly from pyarrow\r\nBut I got a similar error, again it loaded fine in pandas so I am not sure what to do.\r\n\r\n\r\n\r\n","embeddings":[-0.24118343,-0.3198650777,-0.0681558475,0.4316074848,0.3625251055,0.0191550385,0.5210206509,0.437713474,0.2949123979,0.0327753536,0.0169624221,-0.1187045723,0.0905919969,0.1802892983,0.0350663252,0.1130092442,0.0730282664,0.3988213837,0.0317200907,0.0699720979,-0.218860805,0.1208694875,0.005867396,0.0575151071,-0.1244225353,0.0723639429,-0.008208585,0.1169146225,-0.2343517244,-0.4013206661,0.4197234511,0.0784542188,0.3328505158,0.5207661986,-0.0001169051,0.1564733684,0.2654486001,-0.1856259704,-0.0441993587,-0.5195806026,0.0301294662,-0.5329574347,0.2776663601,-0.1672024429,-0.0144887799,-0.1786803156,-0.1109988168,-0.0358052403,0.3014124632,0.3175454736,0.1804958135,-0.2607497573,0.0598921739,-0.0100186532,0.2052922398,0.0371788666,-0.0156855155,0.4175470769,0.1173958927,-0.1267608404,-0.0487308204,-0.193666935,-0.2392427474,0.3581996262,0.3196350932,0.1399399787,0.0709672496,-0.1089668348,0.2680699527,0.1055052057,0.6067355871,-0.307642132,-0.1485474855,-0.0154003808,-0.0202969182,-0.4886516333,0.2558154762,0.404494822,-0.1521599442,0.1173331738,-0.2122125626,0.2484468818,-0.2663929462,0.2028620988,-0.0673474893,-0.3531901836,-0.2522074878,0.2693178952,0.2523397803,0.0737477168,0.3336401582,-0.0937501863,0.0268538464,0.1806966811,-0.5556491017,0.2796493471,-0.1176608056,-0.0904364362,0.1637407988,0.096218653,0.4014219046,0.1370620281,0.0423445031,0.1946978569,0.3191299736,0.1861328632,-0.0428146571,0.0691050887,0.0273857322,0.3462172449,-0.2297836244,-0.2338579148,-0.3881369233,-0.5049571991,0.2381427437,-0.1538383663,0.2674439549,-0.2704333961,-0.2479329556,-0.0960200727,-0.0361324735,-0.0094233667,-0.0528344922,0.5738207698,0.215833962,0.3802042603,-0.054209996,0.2572092116,-0.2112757564,-0.09865468,-0.0803525597,0.0201157145,-0.0576021709,0.0522914641,0.3777484894,-0.2166886032,0.1214607805,0.1762953401,-0.043517001,-0.2938742042,-0.1455057859,-0.4177222252,0.0367241055,0.283308208,0.2024739832,0.0979529768,0.2341527939,-0.2812092602,-0.0606879592,0.3805413246,-0.2301190495,-0.1685689539,0.0972641706,0.0963025913,-0.0661173835,0.1656782478,-0.3956311643,0.1563510746,0.3519503474,-0.4384936988,0.0334676392,-0.2134845853,-0.2411233634,-0.3524810374,0.2018900961,0.2790597677,-0.7854661942,0.3429474831,0.0539065748,-0.0317488685,0.1149988398,0.4237150252,-0.1483171433,0.0611786693,-0.2709869146,0.0921530128,0.1406241804,0.007542707,-0.6293674707,0.3307564855,-0.04602734,0.0895455927,0.3120644391,0.0645528063,0.0512152202,0.0891771689,-0.1080223322,0.0185540114,0.0231345072,0.0440054424,-0.2607476711,-0.1291924864,0.3187506497,0.142625019,-0.0647743195,-0.0676729232,0.2388889641,-0.4087391198,0.2025453746,-0.0180878602,0.1416161656,0.1752468795,0.1791311353,-0.1968236417,0.0349740647,0.0104333451,-0.4493494034,0.1433825642,0.2848802209,-0.1687007099,-0.2159446925,-0.0797307715,-0.384118408,-0.0297010913,-0.2681604922,-0.1241994351,-0.0246524885,0.0725153983,0.0595197789,-0.1440529078,-0.1857402176,0.1161194369,-0.2237771302,-0.0112759126,-0.0158436764,0.1422190964,0.058799319,-0.1249890551,0.1064720228,-0.0629007593,0.207167998,-0.0483339541,-0.2763522565,0.2495912611,0.1456864476,0.0456155054,-0.2857958674,-0.1747876853,0.0830201879,-0.2415053844,-0.0872012973,0.0237385109,0.3663716614,-0.1825469434,-0.3392808735,0.2706229687,-0.1567003578,0.3015136421,0.1879752129,-0.3580714464,0.3065021634,0.016720105,0.1781937331,0.0704920888,0.1380855441,0.0736546293,0.4947431087,0.2995411456,-0.0968417376,-0.1562502384,0.5009906292,-0.0470440127,-0.2185632885,0.3112592995,-0.1181778163,-0.1945351362,0.0684309453,0.0887451172,0.3798215091,0.1293763965,-0.1915307492,0.0146643994,0.1435023844,-0.2500295639,0.1204057485,-0.2204343826,0.1473618895,0.3965650201,0.1270093769,-0.1272856295,-0.195144549,0.0377438515,0.0875855386,0.3257235587,-0.4119557142,-0.0675459281,-0.0734540001,-0.0052061281,-0.1802845895,0.0986031592,-0.0555794202,0.0971230865,-0.3996109366,0.3328239024,0.2897825539,-0.0790665746,-0.2431259155,0.08863388,0.2158128768,-0.3370558321,-0.0034607018,-0.0489910357,-0.2919403613,-0.0261766538,0.3307105303,0.1098190099,-0.0008771374,-0.0217662286,0.0280251838,-0.0548448637,-0.0925327018,-0.0439247601,-0.0732297674,0.0209216662,-0.0511663184,0.4193309844,0.0076079252,-0.1446308047,0.4017484784,-0.2536608279,-0.0489250831,0.350135982,0.0636930913,-0.0958797708,-0.1087071598,-0.3498898149,-0.139888227,-0.4293371439,0.3910627365,0.0847561061,0.0419315621,0.0926315561,0.4331049323,-0.0260847546,0.3013546467,-0.0312164128,-0.0848050937,-0.2977054417,0.3645742834,-0.0561163239,-0.4676565826,0.011941432,0.0828354955,0.3591901362,0.1638266891,-0.4252471328,0.3267471194,-0.15090774,0.1628198922,0.0011049408,-0.0242274385,0.0609546043,-0.0872276798,-0.0218234472,0.0364301093,-0.2417160571,-0.0141803464,-0.0320445746,0.1475835294,0.0513685606,0.458429724,-0.0371929593,0.6513361931,0.0874645486,-0.0882313102,0.4855556488,-0.308239907,0.3818774223,-0.2096708715,-0.3737454712,-0.0855747312,-0.3519817293,-0.1924462765,0.2677486241,-0.0335841477,-0.0334150307,-0.1257756203,0.0035179411,-0.4046861231,-0.0377031192,0.183632195,0.0306968763,-0.0148301804,-0.1005954891,-0.0133011667,-0.0116456645,-0.237362951,-0.1942582428,0.2739638686,-0.1513422579,0.0633870587,-0.2242484242,-0.0615803823,-0.3029371798,0.3292051852,-0.0136234136,0.3102864921,0.0106146866,-0.0176951028,0.0972969308,-0.1252222657,0.7449785471,0.1892860085,-0.1550699919,0.1278296113,-0.0241136737,-0.441583544,-0.121761322,0.1906314343,0.3073361516,0.0744616762,0.3548655212,-0.0703261346,-0.0096933683,-0.1443433017,0.3546780944,0.0432527624,-0.3951280713,-0.5194913745,-0.2349791825,-0.383053571,-0.2121350318,-0.0898666456,0.2882787883,-0.3244074285,-0.1423869133,0.1183612794,-0.2990536094,0.0601567216,0.1533810496,0.2973894477,-0.1129143313,0.1448963135,0.0468351059,0.0698044524,0.376911819,0.7987186313,-0.0785759091,-0.6220775843,0.0848431289,-0.2050266713,0.3853479028,0.0478198007,-0.2032050937,0.1289862841,0.0907342061,0.0620335788,0.0379479825,0.0144376121,0.3036864102,0.0496549308,-0.3454539776,-0.5439224243,0.4874189198,-0.1259078234,0.2102593035,0.302079767,-0.2119999677,-0.1099810228,0.1555226296,-0.3346246481,0.6592135429,-0.4218300581,0.1938303411,0.3073363602,0.056336008,0.2584289312,-0.4451968074,0.1323073953,-0.3919684887,-0.210903883,-0.1043123081,-0.0712032765,0.2393857986,0.432190001,-0.1212453321,0.3118637204,-0.0741497204,0.3325560987,0.11313425,0.2438252568,-0.1005161256,-0.0864683241,-0.3010203838,0.0104239564,-0.1430807561,-0.0349519551,0.028959142,0.0544647723,-0.1805675626,-0.2016452104,-0.2461119592,0.2200246751,-0.2911679149,0.1154046133,-0.1109549478,-0.4686649442,0.1963640451,0.2380675226,-0.2087023854,-0.0851794481,-0.1240850538,0.1550540775,0.1596347541,-0.0401756726,-0.2083743513,-0.0536952429,-0.0158856176,0.0761261061,-0.2351237983,0.1545512527,-0.122278817,-0.0956747606,0.2780734897,0.2376574725,-0.0848071724,-0.5452085137,-0.5326197147,-0.4284158647,0.0759811029,-0.1957320869,0.0651982948,-0.0420785286,0.0983413011,-0.1616345942,0.1859346777,-0.4688693583,-0.1020267904,0.3918125927,-0.1116865948,0.0939545855,0.5907301903,0.2193497866,-0.0168925617,-0.1411525458,0.0642407984,0.3081057072,-0.3531885743,0.1170660481,0.1141491011,0.3159501255,0.1392528862,0.2917454839,0.1565803736,-0.0776725486,0.0744550973,-0.5115905404,-0.3249694705,-0.0165290646,-0.0015315027,0.2479578257,0.1672193259,0.08186692,0.1726779193,0.2032965571,-0.1562302262,0.1004625708,0.1059293076,-0.0869575143,0.2106330097,0.1420163661,-0.1338726133,-0.1038192213,0.0934978351,0.1776407212,-0.0534449145,-0.1887511909,-0.0798553601,0.1500999779,0.0983605087,0.0232860353,-0.2034453899,-0.2526888847,-0.1497063339,-0.19280532,0.0664090961,0.0297824815,-0.0981118679,0.0717605129,-0.2987648547,-0.0846538544,-0.5091862082,0.0271216799,-0.5075390935,0.2231322825,0.29926911,0.0485824719,0.1217573285,0.2004558295,-0.1724265963,-0.0311586745,0.1516266763,-0.1366193742,0.3009471297,-0.4014762044,0.2035068721,-0.0510654896,0.3403376043,0.4848605692,-0.2076416463,0.2197109312,0.0375863202,0.1341150701,-0.3825282753,0.2754767835,0.1172315627,-0.0616297424,-0.027484389,0.1696746349,0.1992117763,-0.2083067745,-0.0957768857,0.0069104712,0.4759882987,-0.0078224344,-0.0339511037,0.4135632217,0.0027141657,0.1789693832,0.2403351814,0.1356374919,0.019620385,0.7002619505,-0.2602286339,-0.0039646165,-0.2352686673,0.3014819026,-0.3278697431,-0.7641029358,0.1616647542,0.2355505377,-0.3533098698,-0.1676512659,0.1258980781,0.6017500162,-0.0233745389,0.0114930086,-0.1145932153,-0.0448354296,0.0140551319,0.0483895428,-0.0662138388,-0.2536709607,-0.0708237216,0.1856833249,-0.0595556647,0.0145596312,0.2749812305,0.1876641512,-0.1494839638,-0.4457591772,0.0657755956,0.1728491187,0.0665428638,-0.2748388648,0.3666117489,0.2653487027,-0.0859915987,0.1805052459,-0.0413444005,0.6402472854,0.504191637,0.0924113542,-0.0617402829,-0.1215721816,-0.1139907762,-0.1794179529,0.1778173298,0.3488176763,0.4212691784,0.1485440582,0.1135984808,-0.0401850268,0.316088289,0.2067891061,0.0996007472,-0.545976758,-0.1189198121,-0.2083118558,-0.2577385902,-0.2266330868,0.0241517406,-0.2890593708,0.0209434181,0.4277476966,0.0739154965,0.2772270143,-0.02682258,0.0345675685,0.0743338466,0.3142196834,0.239776656,0.40228948,-0.4218083918,-0.0354187004,-0.7509756088,0.1165279523,-0.3764895499,0.009592602,-0.0318588242,0.1585383266,0.1377418339,0.0838609263,0.1591721773,0.2994238138,0.116723083,0.2220783383,-0.0561847016,-0.1846203357,-0.3380708694,0.0264856666,-0.1750924289,-0.4655142426,0.2213603109,-0.1279327124,0.009210824,-0.2251862437,0.2239884585,0.2559674084,-0.0059870374,0.4777292609,-0.1120114401,0.6255970597,-0.0850180835,0.3190019131,-0.3009388149,-0.0801894665,-0.1928057671,0.5200991035,0.2061499506,0.271399498,-0.177993089,-0.107916005,-0.2153041661,0.1574100554,-0.0065483749,-0.0705471635,0.0667275563,-0.1510605514,-0.0018241928,-0.058490295,0.1522316486,0.0920479745,-0.2229390591,-0.0781728476,-0.1745537966,-0.180185318,0.1810824573,-0.4043286145,-0.3411037624,0.1487179548,0.1261260211,0.1018662825,0.1396923661,-0.4524213076,0.1217898354,0.424475193,-0.1280173361,-0.4174263477,0.1340316832,-0.2116182446,0.1005702689,-0.1218523532,0.4765018821,0.1411091983,-0.370509088,-0.1615719497,-0.3583221138]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/836","title":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas","comments":"Got almost the same error loading a ~5GB TSV file, first got the same error as OP, then tried giving it my own ReadOptions and also got the same CSV parse error.","body":"Hi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n","comment_length":32,"text":"load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas\nHi All\r\nI am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:\r\ndataset = load_dataset('csv', data_files=files)\r\nWhen I run it I get:\r\n\r\nDownloading and preparing dataset csv\/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache\/huggingface\/datasets\/csv\/default-35575a1051604c88\/0.0.0\/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...\r\n\r\nI am getting this error:\r\n6a4ac4\/csv.py in _generate_tables(self, files)\r\n 78 def _generate_tables(self, files):\r\n 79 for i, file in enumerate(files):\r\n---> 80 pa_table = pac.read_csv(\r\n 81 file,\r\n 82 read_options=self.config.pa_read_options,\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/anaconda2\/envs\/nlp\/lib\/python3.8\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\n**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\n\r\n\r\nThe size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.\r\nThere is no issue reading the file with pandas. any idea what could be the issue?\r\nWhen I am running a different CSV I do not get this line:\r\n (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)\r\n\r\nAny ideas?\r\n\nGot almost the same error loading a ~5GB TSV file, first got the same error as OP, then tried giving it my own ReadOptions and also got the same CSV parse error.","embeddings":[-0.24118343,-0.3198650777,-0.0681558475,0.4316074848,0.3625251055,0.0191550385,0.5210206509,0.437713474,0.2949123979,0.0327753536,0.0169624221,-0.1187045723,0.0905919969,0.1802892983,0.0350663252,0.1130092442,0.0730282664,0.3988213837,0.0317200907,0.0699720979,-0.218860805,0.1208694875,0.005867396,0.0575151071,-0.1244225353,0.0723639429,-0.008208585,0.1169146225,-0.2343517244,-0.4013206661,0.4197234511,0.0784542188,0.3328505158,0.5207661986,-0.0001169051,0.1564733684,0.2654486001,-0.1856259704,-0.0441993587,-0.5195806026,0.0301294662,-0.5329574347,0.2776663601,-0.1672024429,-0.0144887799,-0.1786803156,-0.1109988168,-0.0358052403,0.3014124632,0.3175454736,0.1804958135,-0.2607497573,0.0598921739,-0.0100186532,0.2052922398,0.0371788666,-0.0156855155,0.4175470769,0.1173958927,-0.1267608404,-0.0487308204,-0.193666935,-0.2392427474,0.3581996262,0.3196350932,0.1399399787,0.0709672496,-0.1089668348,0.2680699527,0.1055052057,0.6067355871,-0.307642132,-0.1485474855,-0.0154003808,-0.0202969182,-0.4886516333,0.2558154762,0.404494822,-0.1521599442,0.1173331738,-0.2122125626,0.2484468818,-0.2663929462,0.2028620988,-0.0673474893,-0.3531901836,-0.2522074878,0.2693178952,0.2523397803,0.0737477168,0.3336401582,-0.0937501863,0.0268538464,0.1806966811,-0.5556491017,0.2796493471,-0.1176608056,-0.0904364362,0.1637407988,0.096218653,0.4014219046,0.1370620281,0.0423445031,0.1946978569,0.3191299736,0.1861328632,-0.0428146571,0.0691050887,0.0273857322,0.3462172449,-0.2297836244,-0.2338579148,-0.3881369233,-0.5049571991,0.2381427437,-0.1538383663,0.2674439549,-0.2704333961,-0.2479329556,-0.0960200727,-0.0361324735,-0.0094233667,-0.0528344922,0.5738207698,0.215833962,0.3802042603,-0.054209996,0.2572092116,-0.2112757564,-0.09865468,-0.0803525597,0.0201157145,-0.0576021709,0.0522914641,0.3777484894,-0.2166886032,0.1214607805,0.1762953401,-0.043517001,-0.2938742042,-0.1455057859,-0.4177222252,0.0367241055,0.283308208,0.2024739832,0.0979529768,0.2341527939,-0.2812092602,-0.0606879592,0.3805413246,-0.2301190495,-0.1685689539,0.0972641706,0.0963025913,-0.0661173835,0.1656782478,-0.3956311643,0.1563510746,0.3519503474,-0.4384936988,0.0334676392,-0.2134845853,-0.2411233634,-0.3524810374,0.2018900961,0.2790597677,-0.7854661942,0.3429474831,0.0539065748,-0.0317488685,0.1149988398,0.4237150252,-0.1483171433,0.0611786693,-0.2709869146,0.0921530128,0.1406241804,0.007542707,-0.6293674707,0.3307564855,-0.04602734,0.0895455927,0.3120644391,0.0645528063,0.0512152202,0.0891771689,-0.1080223322,0.0185540114,0.0231345072,0.0440054424,-0.2607476711,-0.1291924864,0.3187506497,0.142625019,-0.0647743195,-0.0676729232,0.2388889641,-0.4087391198,0.2025453746,-0.0180878602,0.1416161656,0.1752468795,0.1791311353,-0.1968236417,0.0349740647,0.0104333451,-0.4493494034,0.1433825642,0.2848802209,-0.1687007099,-0.2159446925,-0.0797307715,-0.384118408,-0.0297010913,-0.2681604922,-0.1241994351,-0.0246524885,0.0725153983,0.0595197789,-0.1440529078,-0.1857402176,0.1161194369,-0.2237771302,-0.0112759126,-0.0158436764,0.1422190964,0.058799319,-0.1249890551,0.1064720228,-0.0629007593,0.207167998,-0.0483339541,-0.2763522565,0.2495912611,0.1456864476,0.0456155054,-0.2857958674,-0.1747876853,0.0830201879,-0.2415053844,-0.0872012973,0.0237385109,0.3663716614,-0.1825469434,-0.3392808735,0.2706229687,-0.1567003578,0.3015136421,0.1879752129,-0.3580714464,0.3065021634,0.016720105,0.1781937331,0.0704920888,0.1380855441,0.0736546293,0.4947431087,0.2995411456,-0.0968417376,-0.1562502384,0.5009906292,-0.0470440127,-0.2185632885,0.3112592995,-0.1181778163,-0.1945351362,0.0684309453,0.0887451172,0.3798215091,0.1293763965,-0.1915307492,0.0146643994,0.1435023844,-0.2500295639,0.1204057485,-0.2204343826,0.1473618895,0.3965650201,0.1270093769,-0.1272856295,-0.195144549,0.0377438515,0.0875855386,0.3257235587,-0.4119557142,-0.0675459281,-0.0734540001,-0.0052061281,-0.1802845895,0.0986031592,-0.0555794202,0.0971230865,-0.3996109366,0.3328239024,0.2897825539,-0.0790665746,-0.2431259155,0.08863388,0.2158128768,-0.3370558321,-0.0034607018,-0.0489910357,-0.2919403613,-0.0261766538,0.3307105303,0.1098190099,-0.0008771374,-0.0217662286,0.0280251838,-0.0548448637,-0.0925327018,-0.0439247601,-0.0732297674,0.0209216662,-0.0511663184,0.4193309844,0.0076079252,-0.1446308047,0.4017484784,-0.2536608279,-0.0489250831,0.350135982,0.0636930913,-0.0958797708,-0.1087071598,-0.3498898149,-0.139888227,-0.4293371439,0.3910627365,0.0847561061,0.0419315621,0.0926315561,0.4331049323,-0.0260847546,0.3013546467,-0.0312164128,-0.0848050937,-0.2977054417,0.3645742834,-0.0561163239,-0.4676565826,0.011941432,0.0828354955,0.3591901362,0.1638266891,-0.4252471328,0.3267471194,-0.15090774,0.1628198922,0.0011049408,-0.0242274385,0.0609546043,-0.0872276798,-0.0218234472,0.0364301093,-0.2417160571,-0.0141803464,-0.0320445746,0.1475835294,0.0513685606,0.458429724,-0.0371929593,0.6513361931,0.0874645486,-0.0882313102,0.4855556488,-0.308239907,0.3818774223,-0.2096708715,-0.3737454712,-0.0855747312,-0.3519817293,-0.1924462765,0.2677486241,-0.0335841477,-0.0334150307,-0.1257756203,0.0035179411,-0.4046861231,-0.0377031192,0.183632195,0.0306968763,-0.0148301804,-0.1005954891,-0.0133011667,-0.0116456645,-0.237362951,-0.1942582428,0.2739638686,-0.1513422579,0.0633870587,-0.2242484242,-0.0615803823,-0.3029371798,0.3292051852,-0.0136234136,0.3102864921,0.0106146866,-0.0176951028,0.0972969308,-0.1252222657,0.7449785471,0.1892860085,-0.1550699919,0.1278296113,-0.0241136737,-0.441583544,-0.121761322,0.1906314343,0.3073361516,0.0744616762,0.3548655212,-0.0703261346,-0.0096933683,-0.1443433017,0.3546780944,0.0432527624,-0.3951280713,-0.5194913745,-0.2349791825,-0.383053571,-0.2121350318,-0.0898666456,0.2882787883,-0.3244074285,-0.1423869133,0.1183612794,-0.2990536094,0.0601567216,0.1533810496,0.2973894477,-0.1129143313,0.1448963135,0.0468351059,0.0698044524,0.376911819,0.7987186313,-0.0785759091,-0.6220775843,0.0848431289,-0.2050266713,0.3853479028,0.0478198007,-0.2032050937,0.1289862841,0.0907342061,0.0620335788,0.0379479825,0.0144376121,0.3036864102,0.0496549308,-0.3454539776,-0.5439224243,0.4874189198,-0.1259078234,0.2102593035,0.302079767,-0.2119999677,-0.1099810228,0.1555226296,-0.3346246481,0.6592135429,-0.4218300581,0.1938303411,0.3073363602,0.056336008,0.2584289312,-0.4451968074,0.1323073953,-0.3919684887,-0.210903883,-0.1043123081,-0.0712032765,0.2393857986,0.432190001,-0.1212453321,0.3118637204,-0.0741497204,0.3325560987,0.11313425,0.2438252568,-0.1005161256,-0.0864683241,-0.3010203838,0.0104239564,-0.1430807561,-0.0349519551,0.028959142,0.0544647723,-0.1805675626,-0.2016452104,-0.2461119592,0.2200246751,-0.2911679149,0.1154046133,-0.1109549478,-0.4686649442,0.1963640451,0.2380675226,-0.2087023854,-0.0851794481,-0.1240850538,0.1550540775,0.1596347541,-0.0401756726,-0.2083743513,-0.0536952429,-0.0158856176,0.0761261061,-0.2351237983,0.1545512527,-0.122278817,-0.0956747606,0.2780734897,0.2376574725,-0.0848071724,-0.5452085137,-0.5326197147,-0.4284158647,0.0759811029,-0.1957320869,0.0651982948,-0.0420785286,0.0983413011,-0.1616345942,0.1859346777,-0.4688693583,-0.1020267904,0.3918125927,-0.1116865948,0.0939545855,0.5907301903,0.2193497866,-0.0168925617,-0.1411525458,0.0642407984,0.3081057072,-0.3531885743,0.1170660481,0.1141491011,0.3159501255,0.1392528862,0.2917454839,0.1565803736,-0.0776725486,0.0744550973,-0.5115905404,-0.3249694705,-0.0165290646,-0.0015315027,0.2479578257,0.1672193259,0.08186692,0.1726779193,0.2032965571,-0.1562302262,0.1004625708,0.1059293076,-0.0869575143,0.2106330097,0.1420163661,-0.1338726133,-0.1038192213,0.0934978351,0.1776407212,-0.0534449145,-0.1887511909,-0.0798553601,0.1500999779,0.0983605087,0.0232860353,-0.2034453899,-0.2526888847,-0.1497063339,-0.19280532,0.0664090961,0.0297824815,-0.0981118679,0.0717605129,-0.2987648547,-0.0846538544,-0.5091862082,0.0271216799,-0.5075390935,0.2231322825,0.29926911,0.0485824719,0.1217573285,0.2004558295,-0.1724265963,-0.0311586745,0.1516266763,-0.1366193742,0.3009471297,-0.4014762044,0.2035068721,-0.0510654896,0.3403376043,0.4848605692,-0.2076416463,0.2197109312,0.0375863202,0.1341150701,-0.3825282753,0.2754767835,0.1172315627,-0.0616297424,-0.027484389,0.1696746349,0.1992117763,-0.2083067745,-0.0957768857,0.0069104712,0.4759882987,-0.0078224344,-0.0339511037,0.4135632217,0.0027141657,0.1789693832,0.2403351814,0.1356374919,0.019620385,0.7002619505,-0.2602286339,-0.0039646165,-0.2352686673,0.3014819026,-0.3278697431,-0.7641029358,0.1616647542,0.2355505377,-0.3533098698,-0.1676512659,0.1258980781,0.6017500162,-0.0233745389,0.0114930086,-0.1145932153,-0.0448354296,0.0140551319,0.0483895428,-0.0662138388,-0.2536709607,-0.0708237216,0.1856833249,-0.0595556647,0.0145596312,0.2749812305,0.1876641512,-0.1494839638,-0.4457591772,0.0657755956,0.1728491187,0.0665428638,-0.2748388648,0.3666117489,0.2653487027,-0.0859915987,0.1805052459,-0.0413444005,0.6402472854,0.504191637,0.0924113542,-0.0617402829,-0.1215721816,-0.1139907762,-0.1794179529,0.1778173298,0.3488176763,0.4212691784,0.1485440582,0.1135984808,-0.0401850268,0.316088289,0.2067891061,0.0996007472,-0.545976758,-0.1189198121,-0.2083118558,-0.2577385902,-0.2266330868,0.0241517406,-0.2890593708,0.0209434181,0.4277476966,0.0739154965,0.2772270143,-0.02682258,0.0345675685,0.0743338466,0.3142196834,0.239776656,0.40228948,-0.4218083918,-0.0354187004,-0.7509756088,0.1165279523,-0.3764895499,0.009592602,-0.0318588242,0.1585383266,0.1377418339,0.0838609263,0.1591721773,0.2994238138,0.116723083,0.2220783383,-0.0561847016,-0.1846203357,-0.3380708694,0.0264856666,-0.1750924289,-0.4655142426,0.2213603109,-0.1279327124,0.009210824,-0.2251862437,0.2239884585,0.2559674084,-0.0059870374,0.4777292609,-0.1120114401,0.6255970597,-0.0850180835,0.3190019131,-0.3009388149,-0.0801894665,-0.1928057671,0.5200991035,0.2061499506,0.271399498,-0.177993089,-0.107916005,-0.2153041661,0.1574100554,-0.0065483749,-0.0705471635,0.0667275563,-0.1510605514,-0.0018241928,-0.058490295,0.1522316486,0.0920479745,-0.2229390591,-0.0781728476,-0.1745537966,-0.180185318,0.1810824573,-0.4043286145,-0.3411037624,0.1487179548,0.1261260211,0.1018662825,0.1396923661,-0.4524213076,0.1217898354,0.424475193,-0.1280173361,-0.4174263477,0.1340316832,-0.2116182446,0.1005702689,-0.1218523532,0.4765018821,0.1411091983,-0.370509088,-0.1615719497,-0.3583221138]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/835","title":"Wikipedia postprocessing","comments":"Hi @bminixhofer ! Parsing WikiMedia is notoriously difficult: this processing used [mwparserfromhell](https:\/\/github.com\/earwig\/mwparserfromhell) which is pretty good but not perfect.\r\n\r\nAs an alternative, you can also use the Wiki40b dataset which was pre-processed using an un-released Google internal tool","body":"Hi, thanks for this library!\r\n\r\nRunning this code:\r\n\r\n```py\r\nimport datasets\r\nwikipedia = datasets.load_dataset(\"wikipedia\", \"20200501.de\")\r\nprint(wikipedia['train']['text'][0])\r\n```\r\n\r\nI get:\r\n\r\n```\r\nmini|Ricardo Flores Mag\u00f3n\r\nmini|Mexikanische Revolution\u00e4re, Mag\u00f3n in der Mitte anf\u00fchrend, gegen die Diktatur von Porfirio Diaz, Ausschnitt des Gem\u00e4lde \u201eTierra y Libertad\u201c von Idelfonso Carrara (?) von 1930.\r\n\r\nRicardo Flores Mag\u00f3n (* 16. September 1874 in San Antonio Eloxochitl\u00e1n im mexikanischen Bundesstaat Oaxaca; \u2020 22. November 1922 im Bundesgef\u00e4ngnis Leavenworth im US-amerikanischen Bundesstaat Kansas) war als Journalist, Gewerkschafter und Literat ein f\u00fchrender anarchistischer Theoretiker und Aktivist, der die revolution\u00e4re mexikanische Bewegung radikal beeinflusste. Mag\u00f3n war Gr\u00fcnder der Partido Liberal Mexicano und Mitglied der Industrial Workers of the World.\r\n\r\nPolitische Biografie \r\nJournalistisch und politisch k\u00e4mpfte er und sein Bruder sehr kompromisslos gegen die Diktatur Porfirio Diaz. Philosophisch und politisch orientiert an radikal anarchistischen Idealen und den Erfahrungen seiner indigenen Vorfahren bei der gemeinschaftlichen Bewirtschaftung des Gemeindelandes, machte er die Forderung \u201eLand und Freiheit\u201c (Tierra y Libertad) popul\u00e4r. Besonders Francisco Villa und Emiliano Zapata griffen die Forderung Land und Freiheit auf. Seine Philosophie hatte gro\u00dfen Einfluss auf die Landarbeiter. 1904 floh er in die USA und gr\u00fcndete 1906 die Partido Liberal Mexicano. Im Exil lernte er u. a. Emma Goldman kennen. Er verbrachte die meiste Zeit seines Lebens in Gef\u00e4ngnissen und im Exil und wurde 1918 in den USA wegen \u201eBehinderung der Kriegsanstrengungen\u201c zu zwanzig Jahren Gef\u00e4ngnis verurteilt. Zu seinem Tod gibt es drei verschiedene Theorien. Offiziell starb er an Herzversagen. Librado Rivera, der die Leiche mit eigenen Augen gesehen hat, geht davon aus, dass Mag\u00f3n von einem Mitgefangenen erdrosselt wurde. Die staatstreue Gewerkschaftszeitung CROM ver\u00f6ffentlichte 1923 einen Beitrag, nachdem Mag\u00f3n von einem Gef\u00e4ngnisw\u00e4rter erschlagen wurde.\r\nmini|Die Br\u00fcder Ricardo (links) und Enrique Flores Mag\u00f3n (rechts) vor dem Los Angeles County Jail, 1917\r\n\r\n[...]\r\n```\r\n\r\nso some Markup like `mini|` is still left. Should I run another parser on this text before feeding it to an ML model or is this a known imperfection of parsing Wiki markup?\r\n\r\nApologies if this has been asked before.","comment_length":38,"text":"Wikipedia postprocessing\nHi, thanks for this library!\r\n\r\nRunning this code:\r\n\r\n```py\r\nimport datasets\r\nwikipedia = datasets.load_dataset(\"wikipedia\", \"20200501.de\")\r\nprint(wikipedia['train']['text'][0])\r\n```\r\n\r\nI get:\r\n\r\n```\r\nmini|Ricardo Flores Mag\u00f3n\r\nmini|Mexikanische Revolution\u00e4re, Mag\u00f3n in der Mitte anf\u00fchrend, gegen die Diktatur von Porfirio Diaz, Ausschnitt des Gem\u00e4lde \u201eTierra y Libertad\u201c von Idelfonso Carrara (?) von 1930.\r\n\r\nRicardo Flores Mag\u00f3n (* 16. September 1874 in San Antonio Eloxochitl\u00e1n im mexikanischen Bundesstaat Oaxaca; \u2020 22. November 1922 im Bundesgef\u00e4ngnis Leavenworth im US-amerikanischen Bundesstaat Kansas) war als Journalist, Gewerkschafter und Literat ein f\u00fchrender anarchistischer Theoretiker und Aktivist, der die revolution\u00e4re mexikanische Bewegung radikal beeinflusste. Mag\u00f3n war Gr\u00fcnder der Partido Liberal Mexicano und Mitglied der Industrial Workers of the World.\r\n\r\nPolitische Biografie \r\nJournalistisch und politisch k\u00e4mpfte er und sein Bruder sehr kompromisslos gegen die Diktatur Porfirio Diaz. Philosophisch und politisch orientiert an radikal anarchistischen Idealen und den Erfahrungen seiner indigenen Vorfahren bei der gemeinschaftlichen Bewirtschaftung des Gemeindelandes, machte er die Forderung \u201eLand und Freiheit\u201c (Tierra y Libertad) popul\u00e4r. Besonders Francisco Villa und Emiliano Zapata griffen die Forderung Land und Freiheit auf. Seine Philosophie hatte gro\u00dfen Einfluss auf die Landarbeiter. 1904 floh er in die USA und gr\u00fcndete 1906 die Partido Liberal Mexicano. Im Exil lernte er u. a. Emma Goldman kennen. Er verbrachte die meiste Zeit seines Lebens in Gef\u00e4ngnissen und im Exil und wurde 1918 in den USA wegen \u201eBehinderung der Kriegsanstrengungen\u201c zu zwanzig Jahren Gef\u00e4ngnis verurteilt. Zu seinem Tod gibt es drei verschiedene Theorien. Offiziell starb er an Herzversagen. Librado Rivera, der die Leiche mit eigenen Augen gesehen hat, geht davon aus, dass Mag\u00f3n von einem Mitgefangenen erdrosselt wurde. Die staatstreue Gewerkschaftszeitung CROM ver\u00f6ffentlichte 1923 einen Beitrag, nachdem Mag\u00f3n von einem Gef\u00e4ngnisw\u00e4rter erschlagen wurde.\r\nmini|Die Br\u00fcder Ricardo (links) und Enrique Flores Mag\u00f3n (rechts) vor dem Los Angeles County Jail, 1917\r\n\r\n[...]\r\n```\r\n\r\nso some Markup like `mini|` is still left. Should I run another parser on this text before feeding it to an ML model or is this a known imperfection of parsing Wiki markup?\r\n\r\nApologies if this has been asked before.\nHi @bminixhofer ! Parsing WikiMedia is notoriously difficult: this processing used [mwparserfromhell](https:\/\/github.com\/earwig\/mwparserfromhell) which is pretty good but not perfect.\r\n\r\nAs an alternative, you can also use the Wiki40b dataset which was pre-processed using an un-released Google internal tool","embeddings":[-0.0376301855,0.1160008237,-0.1909410208,0.4500383735,0.2954652011,-0.2197374552,0.3712628186,0.2625176609,-0.0028850846,0.0670773685,0.163732484,0.3684876263,0.0403140783,-0.0634338036,-0.0525207818,-0.2800580859,0.0010123396,0.1088692024,-0.2758776546,-0.2141533643,-0.2710891366,0.172482416,-0.2704933286,0.1990950406,-0.3585416675,0.0835002214,0.0380281396,-0.1664292365,-0.1340274662,0.0589905679,-0.1230926737,-0.0113907522,-0.2323265374,0.5959078074,-0.0001038565,0.093801789,0.2659810781,-0.14833574,0.1651000977,-0.2740632892,0.0381591879,0.1320966035,-0.018274622,-0.4883696437,-0.1506610513,-0.0973137245,-0.102646552,-0.0080199372,0.1642289311,0.1089652106,0.2779662907,-0.2608303428,0.0990448594,-0.0815536305,0.4071276188,-0.0391591825,0.0021229023,-0.056793265,-0.1048759297,0.0740148202,0.0853269771,0.4533440173,-0.0669130534,-0.1527017206,-0.0390591733,-0.0801407471,0.1149623245,-0.2299122661,0.3598327935,0.3970839679,0.4823433757,-0.237019226,-0.0033172935,0.1349339485,-0.0562789328,-0.0627484247,0.1382213533,0.2485168725,-0.2463987023,0.1998824328,0.108613342,-0.0425026454,-0.1570647359,0.3849435747,-0.3486666977,0.7615579367,-0.1470627636,0.0890357867,-0.0713838562,-0.0618577339,-0.2655573785,0.0647552907,0.406740129,0.1151141599,0.2293580621,0.0623623431,0.0952006355,-0.1879187226,-0.1581106782,-0.2297959328,-0.3073724508,-0.0305526275,-0.0009028913,0.0312914886,0.0951704159,-0.0677764118,-0.1882221997,0.1448563337,0.1271027476,0.1956168562,-0.3205268085,0.1341058016,-0.0726409927,-0.05885702,-0.0109663522,0.0493914671,0.1880798042,-0.1510051042,-0.0501707979,0.1197864413,-0.4061352015,-0.1329094023,-0.2625003159,0.3644146621,-0.3537258208,0.3217244744,0.1339441836,-0.0410714298,0.0427075326,0.2019065171,-0.3462704122,0.3184956312,-0.2574843168,0.0214953125,-0.0806169435,-0.114391923,0.326862514,0.0183948055,-0.0553089455,-0.3175844848,-0.0900209397,-0.4657154083,0.3773736358,0.1164557636,-0.137309745,0.085368216,0.1821402758,-0.4203641117,-0.0611058399,0.0739957839,-0.0117768524,-0.0430926643,-0.1845549345,0.2914001942,0.1253378987,0.114219144,0.2324087024,0.3484160304,0.3461151123,0.1279000342,0.2048683316,0.0798774436,-0.2936294675,-0.1988723129,0.3602199852,0.1398384422,-0.4822925329,-0.1239366159,0.1542090476,0.1174046993,0.4012049735,0.090257898,-0.0307218339,0.5387797952,-0.0575563423,0.2503638566,0.0693088174,-0.1390769929,0.0177722927,-0.0094972653,-0.0056565078,-0.0330239944,0.2078623325,-0.0513774194,0.193570137,0.2120351493,0.5239936113,0.0795123503,0.2428410947,0.1865055859,-0.1770028919,-0.1695266068,0.1912652701,-0.0131387552,0.178401649,-0.0867691785,-0.065780662,0.2785589397,0.3917039037,-0.0177074671,0.03367972,0.2748847008,-0.1571976691,-0.0402402021,0.1508070529,-0.1868496239,0.2586314678,-0.0947320834,0.3009767234,0.3739810288,-0.2793432772,-0.321506232,-0.4420488477,-0.0306167658,-0.0859380737,-0.125920251,0.3621974289,-0.0075938813,-0.1433638334,0.1667322516,0.1475275755,-0.04814687,-0.4452185035,-0.0623425022,-0.3474645615,0.2008005828,-0.2842546701,-0.0703957006,-0.1780652106,0.3738847673,0.0864047855,0.0987277552,0.0366893709,0.0681396648,0.1614260674,0.1951446533,0.0872016698,0.1368025839,-0.1354448497,-0.3788892925,0.1041647419,0.4465124011,0.2120099366,-0.087688081,-0.2107884735,0.4056066871,0.2071245313,-0.0240108278,0.1008186489,-0.2186525911,0.1274811625,0.0669870749,-0.1606174409,-0.3407888114,-0.0228196643,0.353531152,-0.1206595004,0.2333276272,-0.3192092776,0.0747786462,0.5386906266,0.134352088,0.4133019745,-0.0901700556,-0.5303305387,-0.1579623073,-0.0161011554,0.1799594164,-0.1050404757,0.187538743,0.1714820713,0.0786534324,-0.0189090725,-0.1408509463,0.3554648459,0.2964965403,0.4085935354,0.271145761,0.0359766521,0.168243885,-0.251855582,-0.0341589823,-0.209574759,0.379539758,-0.2574423552,0.1205544397,-0.1207627803,-0.0712015107,-0.0884926394,-0.2095499933,-0.2182871401,-0.137300387,-0.0413503908,-0.1236971617,0.0427570492,0.2243424654,-0.1203943491,-0.0427134782,0.250151068,-0.1579011232,-0.0263582114,-0.2779904008,-0.0228662621,0.1925877333,-0.192516908,0.1851531714,0.1140914857,-0.1133456752,-0.3316794038,0.0247598588,-0.4591087997,0.3076533079,-0.2841238976,0.0178090855,0.1315817088,0.2630504668,-0.1862816513,-0.298184514,0.1260712594,-0.1022754908,-0.0718508959,-0.0729202703,-0.2129737437,0.1151741818,-0.0549499691,-0.6069825292,-0.184398517,-0.2522272468,0.1753420532,0.4133389592,0.1765688062,0.4491932094,0.2899242342,0.1499410123,0.2895770371,0.1858320087,-0.36908108,-0.0820946619,0.3819562197,0.1831923574,-0.1003113538,-0.0403756425,-0.1214303225,-0.1555097997,0.03503149,-0.3284410238,-0.0780139491,0.1545450389,0.009956385,0.0603399575,0.0701203197,0.3381595612,0.0599646047,-0.237380594,-0.0741835982,-0.0543027446,-0.0672510192,-0.1804206967,0.1228020489,-0.2215511054,0.1971711069,0.0745376498,0.56469661,-0.0219051447,-0.0945554003,-0.1301510632,-0.1098464206,0.0467870198,0.0142612727,-0.1725403517,0.1183864325,-0.0657217801,-0.1128841639,0.285828799,0.0218980443,-0.3113709688,-0.0823392123,0.2434457988,-0.5951212049,-0.233756721,0.073885493,0.2072135806,0.3250167668,-0.2242445201,0.0181401148,-0.1529426575,-0.1985205412,-0.0743087381,0.2720814943,-0.1294915974,-0.1800871342,-0.003993832,0.0455366112,-0.3112497926,0.0897599682,0.106302321,-0.2374592274,0.0194009636,-0.3281236887,0.1842596233,-0.1281688064,0.2043621242,-0.164046824,-0.1016448587,0.2306401879,0.0266481508,-0.2165665627,-0.017609857,-0.3630180061,0.1319010556,0.5669853091,0.4043781161,-0.4217731953,-0.1530512571,0.5155639052,-0.0920582935,-0.0368700549,-0.1138465926,-0.2131608129,-0.1064262241,-0.2825685143,0.2626232505,0.0399504974,0.1282545775,-0.0002701624,-0.144147411,0.3049248755,0.3409052193,0.1048147157,0.1139119118,0.1659893095,0.1508635283,-0.2560137212,-0.2157394439,0.0827042684,-0.3168414533,0.1227055416,0.0964087173,-0.0023286266,-0.0933877677,-0.4582925141,0.1420152485,0.2253702581,-0.1954173446,0.2062242776,0.1236633286,0.0487761311,0.0111880777,0.1006363705,0.3577772677,-0.1785044521,-0.0160647426,-0.3824625015,0.3698531985,-0.0442784615,-0.1138214618,0.0079461094,-0.3191575706,-0.2880116105,0.2350889146,-0.0233705901,1.0900639296,0.1916282028,0.0601801537,0.0798477754,0.3357889652,0.4070530236,0.0503851138,-0.0357324816,-0.4293996394,0.183091566,-0.109221518,0.1870466024,-0.0323119424,-0.0063557001,-0.3201174736,-0.0288609788,-0.2233071029,0.0416371636,-0.0022941369,-0.2610445321,0.2794048488,-0.0714469925,-0.1940616369,0.1843516082,-0.0085481983,0.0132442368,0.008396632,0.1159456819,-0.2418378592,-0.1455354989,-0.3945600092,-0.0565445051,-0.3439761996,0.2061203867,0.1736727655,-0.4040267467,-0.123506479,0.0324687175,0.1290130615,0.3851498663,-0.0401921049,0.1245268807,-0.0149954855,-0.115345493,0.1051421538,0.12752074,0.302090615,-0.059637025,-0.3325281739,0.0474285409,-0.0684778765,-0.2840799987,-0.1140778735,-0.1202758998,0.1880212575,-0.0775806159,-0.0944027007,0.2715235651,-0.165989086,-0.3825518191,0.220366165,-0.122352846,-0.2146441489,0.1128939092,0.2487442642,-0.4793298244,0.0701411888,0.267894268,0.3993580937,0.2682992518,0.3634387255,0.4402483702,-0.3178363144,-0.4193708599,-0.3453825712,-0.344388932,-0.3034403026,-0.2273005992,0.127565816,-0.0442080237,0.0322743431,0.2528116107,0.1267941445,-0.0456003658,-0.1809589863,-0.4442609549,-0.1378889382,0.1769991368,-0.0078211483,0.1069375947,0.26016891,-0.0003516951,0.1314752996,0.0718079805,-0.4596363008,0.0954610333,-0.401098907,0.2911189795,0.249866128,-0.2424764186,-0.1911943257,0.0079129813,0.3404124975,-0.0843356848,-0.1227236539,-0.4186241329,-0.1275819242,0.0791985765,0.0390139632,-0.2021325827,0.1440882385,0.1327942461,-0.2354942858,-0.1377093643,-0.0916096568,0.1327189654,0.1528631598,0.2610372007,-0.0577553473,0.3609778583,-0.2998996377,-0.2252101302,0.0453321673,0.2198771834,0.0875917152,0.1392415464,0.114196144,-0.0657399222,-0.3645977676,0.0850177333,-0.0203922708,0.3905954957,0.3512126803,-0.0503478386,-0.0427762307,0.3029699922,0.2377179116,0.2442759573,-0.2000593245,-0.0955429673,-0.0412150286,0.3443334997,-0.3381033838,0.2190651596,0.4438180029,-0.2796862423,0.1534242183,0.3051404059,0.1139929369,0.0646672398,0.1722955555,0.1230759844,-0.0289411955,-0.3710559309,0.1493567377,0.0802782178,0.1446579099,0.0820689425,0.0083300704,0.3275237381,0.1921262294,0.4157094359,0.5534650087,0.0458230227,0.0485229902,0.0934075266,-0.2124032825,-0.1737228483,0.3782322705,0.250852555,-0.3872358203,-0.0425766632,0.0690187216,-0.0957765654,0.1958178878,0.220619455,-0.1258994639,0.2072698921,-0.1061266288,-0.2802641094,-0.2669960558,-0.216238603,-0.1254568994,-0.2240758389,0.1985271573,0.0819865987,0.0428587534,0.0007927907,-0.1652704477,-0.3171619773,0.1963714957,-0.3757813871,0.3776463568,-0.247577399,0.102629438,-0.1802186221,0.2640243471,-0.3552978337,0.4572235644,0.1867239326,-0.0397551209,-0.3567960858,-0.1345352679,-0.1819654256,-0.1360786259,-0.1003686562,-0.0508694798,0.4067136049,0.0049415873,0.1646077186,0.2986797392,-0.3479203284,0.0352781601,0.3535931706,0.033140827,0.0874756426,-0.0871276259,0.1985788047,-0.1533524394,-0.1870585382,0.0449051037,-0.3499832749,0.2425706387,0.3185038567,0.1421524733,0.2089875787,-0.1796865165,0.1499054283,0.0874758661,0.1393532902,0.4506242573,0.1468863636,0.0160898175,-0.1277967989,-0.2883241475,0.3457967937,0.0065063708,0.038130641,0.0090687796,0.3338782489,0.2970461547,0.1066390052,0.0200609546,-0.255394429,0.0714398623,0.0563794933,-0.1783145666,0.084132269,-0.0173060037,-0.1360272318,-0.0498393103,-0.1777641773,-0.1465191692,-0.2832279503,0.303167522,-0.0624727495,-0.3920107782,0.0241401959,0.1590516716,0.3841212988,-0.1318282038,0.1754274219,-0.3827268779,0.0820533484,-0.0026540465,-0.0795917809,0.0105235148,0.5436720848,0.1163988337,0.1178999245,0.1743744314,-0.0268047359,-0.0146710789,0.2737036645,-0.0396735817,-0.3405407071,-0.1938305199,-0.0594838709,0.0699940398,0.0680811852,-0.2632058263,0.1269263476,-0.1516225487,0.0331990793,-0.2077036351,-0.5855808258,0.2162085176,-0.0780524611,-0.5391718149,-0.3936989605,0.1647264957,0.0663961694,0.4991380274,-0.4331305325,0.1830520779,0.3660574555,-0.1315494329,-0.3380313814,-0.0578141324,-0.1394308209,0.1607730985,-0.1068724319,0.0511503443,-0.2077844143,-0.1916946918,-0.259087801,-0.3279494941]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/835","title":"Wikipedia postprocessing","comments":"Ok, thanks! I'll try the Wiki40b dataset.","body":"Hi, thanks for this library!\r\n\r\nRunning this code:\r\n\r\n```py\r\nimport datasets\r\nwikipedia = datasets.load_dataset(\"wikipedia\", \"20200501.de\")\r\nprint(wikipedia['train']['text'][0])\r\n```\r\n\r\nI get:\r\n\r\n```\r\nmini|Ricardo Flores Mag\u00f3n\r\nmini|Mexikanische Revolution\u00e4re, Mag\u00f3n in der Mitte anf\u00fchrend, gegen die Diktatur von Porfirio Diaz, Ausschnitt des Gem\u00e4lde \u201eTierra y Libertad\u201c von Idelfonso Carrara (?) von 1930.\r\n\r\nRicardo Flores Mag\u00f3n (* 16. September 1874 in San Antonio Eloxochitl\u00e1n im mexikanischen Bundesstaat Oaxaca; \u2020 22. November 1922 im Bundesgef\u00e4ngnis Leavenworth im US-amerikanischen Bundesstaat Kansas) war als Journalist, Gewerkschafter und Literat ein f\u00fchrender anarchistischer Theoretiker und Aktivist, der die revolution\u00e4re mexikanische Bewegung radikal beeinflusste. Mag\u00f3n war Gr\u00fcnder der Partido Liberal Mexicano und Mitglied der Industrial Workers of the World.\r\n\r\nPolitische Biografie \r\nJournalistisch und politisch k\u00e4mpfte er und sein Bruder sehr kompromisslos gegen die Diktatur Porfirio Diaz. Philosophisch und politisch orientiert an radikal anarchistischen Idealen und den Erfahrungen seiner indigenen Vorfahren bei der gemeinschaftlichen Bewirtschaftung des Gemeindelandes, machte er die Forderung \u201eLand und Freiheit\u201c (Tierra y Libertad) popul\u00e4r. Besonders Francisco Villa und Emiliano Zapata griffen die Forderung Land und Freiheit auf. Seine Philosophie hatte gro\u00dfen Einfluss auf die Landarbeiter. 1904 floh er in die USA und gr\u00fcndete 1906 die Partido Liberal Mexicano. Im Exil lernte er u. a. Emma Goldman kennen. Er verbrachte die meiste Zeit seines Lebens in Gef\u00e4ngnissen und im Exil und wurde 1918 in den USA wegen \u201eBehinderung der Kriegsanstrengungen\u201c zu zwanzig Jahren Gef\u00e4ngnis verurteilt. Zu seinem Tod gibt es drei verschiedene Theorien. Offiziell starb er an Herzversagen. Librado Rivera, der die Leiche mit eigenen Augen gesehen hat, geht davon aus, dass Mag\u00f3n von einem Mitgefangenen erdrosselt wurde. Die staatstreue Gewerkschaftszeitung CROM ver\u00f6ffentlichte 1923 einen Beitrag, nachdem Mag\u00f3n von einem Gef\u00e4ngnisw\u00e4rter erschlagen wurde.\r\nmini|Die Br\u00fcder Ricardo (links) und Enrique Flores Mag\u00f3n (rechts) vor dem Los Angeles County Jail, 1917\r\n\r\n[...]\r\n```\r\n\r\nso some Markup like `mini|` is still left. Should I run another parser on this text before feeding it to an ML model or is this a known imperfection of parsing Wiki markup?\r\n\r\nApologies if this has been asked before.","comment_length":7,"text":"Wikipedia postprocessing\nHi, thanks for this library!\r\n\r\nRunning this code:\r\n\r\n```py\r\nimport datasets\r\nwikipedia = datasets.load_dataset(\"wikipedia\", \"20200501.de\")\r\nprint(wikipedia['train']['text'][0])\r\n```\r\n\r\nI get:\r\n\r\n```\r\nmini|Ricardo Flores Mag\u00f3n\r\nmini|Mexikanische Revolution\u00e4re, Mag\u00f3n in der Mitte anf\u00fchrend, gegen die Diktatur von Porfirio Diaz, Ausschnitt des Gem\u00e4lde \u201eTierra y Libertad\u201c von Idelfonso Carrara (?) von 1930.\r\n\r\nRicardo Flores Mag\u00f3n (* 16. September 1874 in San Antonio Eloxochitl\u00e1n im mexikanischen Bundesstaat Oaxaca; \u2020 22. November 1922 im Bundesgef\u00e4ngnis Leavenworth im US-amerikanischen Bundesstaat Kansas) war als Journalist, Gewerkschafter und Literat ein f\u00fchrender anarchistischer Theoretiker und Aktivist, der die revolution\u00e4re mexikanische Bewegung radikal beeinflusste. Mag\u00f3n war Gr\u00fcnder der Partido Liberal Mexicano und Mitglied der Industrial Workers of the World.\r\n\r\nPolitische Biografie \r\nJournalistisch und politisch k\u00e4mpfte er und sein Bruder sehr kompromisslos gegen die Diktatur Porfirio Diaz. Philosophisch und politisch orientiert an radikal anarchistischen Idealen und den Erfahrungen seiner indigenen Vorfahren bei der gemeinschaftlichen Bewirtschaftung des Gemeindelandes, machte er die Forderung \u201eLand und Freiheit\u201c (Tierra y Libertad) popul\u00e4r. Besonders Francisco Villa und Emiliano Zapata griffen die Forderung Land und Freiheit auf. Seine Philosophie hatte gro\u00dfen Einfluss auf die Landarbeiter. 1904 floh er in die USA und gr\u00fcndete 1906 die Partido Liberal Mexicano. Im Exil lernte er u. a. Emma Goldman kennen. Er verbrachte die meiste Zeit seines Lebens in Gef\u00e4ngnissen und im Exil und wurde 1918 in den USA wegen \u201eBehinderung der Kriegsanstrengungen\u201c zu zwanzig Jahren Gef\u00e4ngnis verurteilt. Zu seinem Tod gibt es drei verschiedene Theorien. Offiziell starb er an Herzversagen. Librado Rivera, der die Leiche mit eigenen Augen gesehen hat, geht davon aus, dass Mag\u00f3n von einem Mitgefangenen erdrosselt wurde. Die staatstreue Gewerkschaftszeitung CROM ver\u00f6ffentlichte 1923 einen Beitrag, nachdem Mag\u00f3n von einem Gef\u00e4ngnisw\u00e4rter erschlagen wurde.\r\nmini|Die Br\u00fcder Ricardo (links) und Enrique Flores Mag\u00f3n (rechts) vor dem Los Angeles County Jail, 1917\r\n\r\n[...]\r\n```\r\n\r\nso some Markup like `mini|` is still left. Should I run another parser on this text before feeding it to an ML model or is this a known imperfection of parsing Wiki markup?\r\n\r\nApologies if this has been asked before.\nOk, thanks! I'll try the Wiki40b dataset.","embeddings":[-0.0376301855,0.1160008237,-0.1909410208,0.4500383735,0.2954652011,-0.2197374552,0.3712628186,0.2625176609,-0.0028850846,0.0670773685,0.163732484,0.3684876263,0.0403140783,-0.0634338036,-0.0525207818,-0.2800580859,0.0010123396,0.1088692024,-0.2758776546,-0.2141533643,-0.2710891366,0.172482416,-0.2704933286,0.1990950406,-0.3585416675,0.0835002214,0.0380281396,-0.1664292365,-0.1340274662,0.0589905679,-0.1230926737,-0.0113907522,-0.2323265374,0.5959078074,-0.0001038565,0.093801789,0.2659810781,-0.14833574,0.1651000977,-0.2740632892,0.0381591879,0.1320966035,-0.018274622,-0.4883696437,-0.1506610513,-0.0973137245,-0.102646552,-0.0080199372,0.1642289311,0.1089652106,0.2779662907,-0.2608303428,0.0990448594,-0.0815536305,0.4071276188,-0.0391591825,0.0021229023,-0.056793265,-0.1048759297,0.0740148202,0.0853269771,0.4533440173,-0.0669130534,-0.1527017206,-0.0390591733,-0.0801407471,0.1149623245,-0.2299122661,0.3598327935,0.3970839679,0.4823433757,-0.237019226,-0.0033172935,0.1349339485,-0.0562789328,-0.0627484247,0.1382213533,0.2485168725,-0.2463987023,0.1998824328,0.108613342,-0.0425026454,-0.1570647359,0.3849435747,-0.3486666977,0.7615579367,-0.1470627636,0.0890357867,-0.0713838562,-0.0618577339,-0.2655573785,0.0647552907,0.406740129,0.1151141599,0.2293580621,0.0623623431,0.0952006355,-0.1879187226,-0.1581106782,-0.2297959328,-0.3073724508,-0.0305526275,-0.0009028913,0.0312914886,0.0951704159,-0.0677764118,-0.1882221997,0.1448563337,0.1271027476,0.1956168562,-0.3205268085,0.1341058016,-0.0726409927,-0.05885702,-0.0109663522,0.0493914671,0.1880798042,-0.1510051042,-0.0501707979,0.1197864413,-0.4061352015,-0.1329094023,-0.2625003159,0.3644146621,-0.3537258208,0.3217244744,0.1339441836,-0.0410714298,0.0427075326,0.2019065171,-0.3462704122,0.3184956312,-0.2574843168,0.0214953125,-0.0806169435,-0.114391923,0.326862514,0.0183948055,-0.0553089455,-0.3175844848,-0.0900209397,-0.4657154083,0.3773736358,0.1164557636,-0.137309745,0.085368216,0.1821402758,-0.4203641117,-0.0611058399,0.0739957839,-0.0117768524,-0.0430926643,-0.1845549345,0.2914001942,0.1253378987,0.114219144,0.2324087024,0.3484160304,0.3461151123,0.1279000342,0.2048683316,0.0798774436,-0.2936294675,-0.1988723129,0.3602199852,0.1398384422,-0.4822925329,-0.1239366159,0.1542090476,0.1174046993,0.4012049735,0.090257898,-0.0307218339,0.5387797952,-0.0575563423,0.2503638566,0.0693088174,-0.1390769929,0.0177722927,-0.0094972653,-0.0056565078,-0.0330239944,0.2078623325,-0.0513774194,0.193570137,0.2120351493,0.5239936113,0.0795123503,0.2428410947,0.1865055859,-0.1770028919,-0.1695266068,0.1912652701,-0.0131387552,0.178401649,-0.0867691785,-0.065780662,0.2785589397,0.3917039037,-0.0177074671,0.03367972,0.2748847008,-0.1571976691,-0.0402402021,0.1508070529,-0.1868496239,0.2586314678,-0.0947320834,0.3009767234,0.3739810288,-0.2793432772,-0.321506232,-0.4420488477,-0.0306167658,-0.0859380737,-0.125920251,0.3621974289,-0.0075938813,-0.1433638334,0.1667322516,0.1475275755,-0.04814687,-0.4452185035,-0.0623425022,-0.3474645615,0.2008005828,-0.2842546701,-0.0703957006,-0.1780652106,0.3738847673,0.0864047855,0.0987277552,0.0366893709,0.0681396648,0.1614260674,0.1951446533,0.0872016698,0.1368025839,-0.1354448497,-0.3788892925,0.1041647419,0.4465124011,0.2120099366,-0.087688081,-0.2107884735,0.4056066871,0.2071245313,-0.0240108278,0.1008186489,-0.2186525911,0.1274811625,0.0669870749,-0.1606174409,-0.3407888114,-0.0228196643,0.353531152,-0.1206595004,0.2333276272,-0.3192092776,0.0747786462,0.5386906266,0.134352088,0.4133019745,-0.0901700556,-0.5303305387,-0.1579623073,-0.0161011554,0.1799594164,-0.1050404757,0.187538743,0.1714820713,0.0786534324,-0.0189090725,-0.1408509463,0.3554648459,0.2964965403,0.4085935354,0.271145761,0.0359766521,0.168243885,-0.251855582,-0.0341589823,-0.209574759,0.379539758,-0.2574423552,0.1205544397,-0.1207627803,-0.0712015107,-0.0884926394,-0.2095499933,-0.2182871401,-0.137300387,-0.0413503908,-0.1236971617,0.0427570492,0.2243424654,-0.1203943491,-0.0427134782,0.250151068,-0.1579011232,-0.0263582114,-0.2779904008,-0.0228662621,0.1925877333,-0.192516908,0.1851531714,0.1140914857,-0.1133456752,-0.3316794038,0.0247598588,-0.4591087997,0.3076533079,-0.2841238976,0.0178090855,0.1315817088,0.2630504668,-0.1862816513,-0.298184514,0.1260712594,-0.1022754908,-0.0718508959,-0.0729202703,-0.2129737437,0.1151741818,-0.0549499691,-0.6069825292,-0.184398517,-0.2522272468,0.1753420532,0.4133389592,0.1765688062,0.4491932094,0.2899242342,0.1499410123,0.2895770371,0.1858320087,-0.36908108,-0.0820946619,0.3819562197,0.1831923574,-0.1003113538,-0.0403756425,-0.1214303225,-0.1555097997,0.03503149,-0.3284410238,-0.0780139491,0.1545450389,0.009956385,0.0603399575,0.0701203197,0.3381595612,0.0599646047,-0.237380594,-0.0741835982,-0.0543027446,-0.0672510192,-0.1804206967,0.1228020489,-0.2215511054,0.1971711069,0.0745376498,0.56469661,-0.0219051447,-0.0945554003,-0.1301510632,-0.1098464206,0.0467870198,0.0142612727,-0.1725403517,0.1183864325,-0.0657217801,-0.1128841639,0.285828799,0.0218980443,-0.3113709688,-0.0823392123,0.2434457988,-0.5951212049,-0.233756721,0.073885493,0.2072135806,0.3250167668,-0.2242445201,0.0181401148,-0.1529426575,-0.1985205412,-0.0743087381,0.2720814943,-0.1294915974,-0.1800871342,-0.003993832,0.0455366112,-0.3112497926,0.0897599682,0.106302321,-0.2374592274,0.0194009636,-0.3281236887,0.1842596233,-0.1281688064,0.2043621242,-0.164046824,-0.1016448587,0.2306401879,0.0266481508,-0.2165665627,-0.017609857,-0.3630180061,0.1319010556,0.5669853091,0.4043781161,-0.4217731953,-0.1530512571,0.5155639052,-0.0920582935,-0.0368700549,-0.1138465926,-0.2131608129,-0.1064262241,-0.2825685143,0.2626232505,0.0399504974,0.1282545775,-0.0002701624,-0.144147411,0.3049248755,0.3409052193,0.1048147157,0.1139119118,0.1659893095,0.1508635283,-0.2560137212,-0.2157394439,0.0827042684,-0.3168414533,0.1227055416,0.0964087173,-0.0023286266,-0.0933877677,-0.4582925141,0.1420152485,0.2253702581,-0.1954173446,0.2062242776,0.1236633286,0.0487761311,0.0111880777,0.1006363705,0.3577772677,-0.1785044521,-0.0160647426,-0.3824625015,0.3698531985,-0.0442784615,-0.1138214618,0.0079461094,-0.3191575706,-0.2880116105,0.2350889146,-0.0233705901,1.0900639296,0.1916282028,0.0601801537,0.0798477754,0.3357889652,0.4070530236,0.0503851138,-0.0357324816,-0.4293996394,0.183091566,-0.109221518,0.1870466024,-0.0323119424,-0.0063557001,-0.3201174736,-0.0288609788,-0.2233071029,0.0416371636,-0.0022941369,-0.2610445321,0.2794048488,-0.0714469925,-0.1940616369,0.1843516082,-0.0085481983,0.0132442368,0.008396632,0.1159456819,-0.2418378592,-0.1455354989,-0.3945600092,-0.0565445051,-0.3439761996,0.2061203867,0.1736727655,-0.4040267467,-0.123506479,0.0324687175,0.1290130615,0.3851498663,-0.0401921049,0.1245268807,-0.0149954855,-0.115345493,0.1051421538,0.12752074,0.302090615,-0.059637025,-0.3325281739,0.0474285409,-0.0684778765,-0.2840799987,-0.1140778735,-0.1202758998,0.1880212575,-0.0775806159,-0.0944027007,0.2715235651,-0.165989086,-0.3825518191,0.220366165,-0.122352846,-0.2146441489,0.1128939092,0.2487442642,-0.4793298244,0.0701411888,0.267894268,0.3993580937,0.2682992518,0.3634387255,0.4402483702,-0.3178363144,-0.4193708599,-0.3453825712,-0.344388932,-0.3034403026,-0.2273005992,0.127565816,-0.0442080237,0.0322743431,0.2528116107,0.1267941445,-0.0456003658,-0.1809589863,-0.4442609549,-0.1378889382,0.1769991368,-0.0078211483,0.1069375947,0.26016891,-0.0003516951,0.1314752996,0.0718079805,-0.4596363008,0.0954610333,-0.401098907,0.2911189795,0.249866128,-0.2424764186,-0.1911943257,0.0079129813,0.3404124975,-0.0843356848,-0.1227236539,-0.4186241329,-0.1275819242,0.0791985765,0.0390139632,-0.2021325827,0.1440882385,0.1327942461,-0.2354942858,-0.1377093643,-0.0916096568,0.1327189654,0.1528631598,0.2610372007,-0.0577553473,0.3609778583,-0.2998996377,-0.2252101302,0.0453321673,0.2198771834,0.0875917152,0.1392415464,0.114196144,-0.0657399222,-0.3645977676,0.0850177333,-0.0203922708,0.3905954957,0.3512126803,-0.0503478386,-0.0427762307,0.3029699922,0.2377179116,0.2442759573,-0.2000593245,-0.0955429673,-0.0412150286,0.3443334997,-0.3381033838,0.2190651596,0.4438180029,-0.2796862423,0.1534242183,0.3051404059,0.1139929369,0.0646672398,0.1722955555,0.1230759844,-0.0289411955,-0.3710559309,0.1493567377,0.0802782178,0.1446579099,0.0820689425,0.0083300704,0.3275237381,0.1921262294,0.4157094359,0.5534650087,0.0458230227,0.0485229902,0.0934075266,-0.2124032825,-0.1737228483,0.3782322705,0.250852555,-0.3872358203,-0.0425766632,0.0690187216,-0.0957765654,0.1958178878,0.220619455,-0.1258994639,0.2072698921,-0.1061266288,-0.2802641094,-0.2669960558,-0.216238603,-0.1254568994,-0.2240758389,0.1985271573,0.0819865987,0.0428587534,0.0007927907,-0.1652704477,-0.3171619773,0.1963714957,-0.3757813871,0.3776463568,-0.247577399,0.102629438,-0.1802186221,0.2640243471,-0.3552978337,0.4572235644,0.1867239326,-0.0397551209,-0.3567960858,-0.1345352679,-0.1819654256,-0.1360786259,-0.1003686562,-0.0508694798,0.4067136049,0.0049415873,0.1646077186,0.2986797392,-0.3479203284,0.0352781601,0.3535931706,0.033140827,0.0874756426,-0.0871276259,0.1985788047,-0.1533524394,-0.1870585382,0.0449051037,-0.3499832749,0.2425706387,0.3185038567,0.1421524733,0.2089875787,-0.1796865165,0.1499054283,0.0874758661,0.1393532902,0.4506242573,0.1468863636,0.0160898175,-0.1277967989,-0.2883241475,0.3457967937,0.0065063708,0.038130641,0.0090687796,0.3338782489,0.2970461547,0.1066390052,0.0200609546,-0.255394429,0.0714398623,0.0563794933,-0.1783145666,0.084132269,-0.0173060037,-0.1360272318,-0.0498393103,-0.1777641773,-0.1465191692,-0.2832279503,0.303167522,-0.0624727495,-0.3920107782,0.0241401959,0.1590516716,0.3841212988,-0.1318282038,0.1754274219,-0.3827268779,0.0820533484,-0.0026540465,-0.0795917809,0.0105235148,0.5436720848,0.1163988337,0.1178999245,0.1743744314,-0.0268047359,-0.0146710789,0.2737036645,-0.0396735817,-0.3405407071,-0.1938305199,-0.0594838709,0.0699940398,0.0680811852,-0.2632058263,0.1269263476,-0.1516225487,0.0331990793,-0.2077036351,-0.5855808258,0.2162085176,-0.0780524611,-0.5391718149,-0.3936989605,0.1647264957,0.0663961694,0.4991380274,-0.4331305325,0.1830520779,0.3660574555,-0.1315494329,-0.3380313814,-0.0578141324,-0.1394308209,0.1607730985,-0.1068724319,0.0511503443,-0.2077844143,-0.1916946918,-0.259087801,-0.3279494941]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/835","title":"Wikipedia postprocessing","comments":"If anyone else is concerned about this, `wiki40b` does indeed seem very well cleaned.","body":"Hi, thanks for this library!\r\n\r\nRunning this code:\r\n\r\n```py\r\nimport datasets\r\nwikipedia = datasets.load_dataset(\"wikipedia\", \"20200501.de\")\r\nprint(wikipedia['train']['text'][0])\r\n```\r\n\r\nI get:\r\n\r\n```\r\nmini|Ricardo Flores Mag\u00f3n\r\nmini|Mexikanische Revolution\u00e4re, Mag\u00f3n in der Mitte anf\u00fchrend, gegen die Diktatur von Porfirio Diaz, Ausschnitt des Gem\u00e4lde \u201eTierra y Libertad\u201c von Idelfonso Carrara (?) von 1930.\r\n\r\nRicardo Flores Mag\u00f3n (* 16. September 1874 in San Antonio Eloxochitl\u00e1n im mexikanischen Bundesstaat Oaxaca; \u2020 22. November 1922 im Bundesgef\u00e4ngnis Leavenworth im US-amerikanischen Bundesstaat Kansas) war als Journalist, Gewerkschafter und Literat ein f\u00fchrender anarchistischer Theoretiker und Aktivist, der die revolution\u00e4re mexikanische Bewegung radikal beeinflusste. Mag\u00f3n war Gr\u00fcnder der Partido Liberal Mexicano und Mitglied der Industrial Workers of the World.\r\n\r\nPolitische Biografie \r\nJournalistisch und politisch k\u00e4mpfte er und sein Bruder sehr kompromisslos gegen die Diktatur Porfirio Diaz. Philosophisch und politisch orientiert an radikal anarchistischen Idealen und den Erfahrungen seiner indigenen Vorfahren bei der gemeinschaftlichen Bewirtschaftung des Gemeindelandes, machte er die Forderung \u201eLand und Freiheit\u201c (Tierra y Libertad) popul\u00e4r. Besonders Francisco Villa und Emiliano Zapata griffen die Forderung Land und Freiheit auf. Seine Philosophie hatte gro\u00dfen Einfluss auf die Landarbeiter. 1904 floh er in die USA und gr\u00fcndete 1906 die Partido Liberal Mexicano. Im Exil lernte er u. a. Emma Goldman kennen. Er verbrachte die meiste Zeit seines Lebens in Gef\u00e4ngnissen und im Exil und wurde 1918 in den USA wegen \u201eBehinderung der Kriegsanstrengungen\u201c zu zwanzig Jahren Gef\u00e4ngnis verurteilt. Zu seinem Tod gibt es drei verschiedene Theorien. Offiziell starb er an Herzversagen. Librado Rivera, der die Leiche mit eigenen Augen gesehen hat, geht davon aus, dass Mag\u00f3n von einem Mitgefangenen erdrosselt wurde. Die staatstreue Gewerkschaftszeitung CROM ver\u00f6ffentlichte 1923 einen Beitrag, nachdem Mag\u00f3n von einem Gef\u00e4ngnisw\u00e4rter erschlagen wurde.\r\nmini|Die Br\u00fcder Ricardo (links) und Enrique Flores Mag\u00f3n (rechts) vor dem Los Angeles County Jail, 1917\r\n\r\n[...]\r\n```\r\n\r\nso some Markup like `mini|` is still left. Should I run another parser on this text before feeding it to an ML model or is this a known imperfection of parsing Wiki markup?\r\n\r\nApologies if this has been asked before.","comment_length":14,"text":"Wikipedia postprocessing\nHi, thanks for this library!\r\n\r\nRunning this code:\r\n\r\n```py\r\nimport datasets\r\nwikipedia = datasets.load_dataset(\"wikipedia\", \"20200501.de\")\r\nprint(wikipedia['train']['text'][0])\r\n```\r\n\r\nI get:\r\n\r\n```\r\nmini|Ricardo Flores Mag\u00f3n\r\nmini|Mexikanische Revolution\u00e4re, Mag\u00f3n in der Mitte anf\u00fchrend, gegen die Diktatur von Porfirio Diaz, Ausschnitt des Gem\u00e4lde \u201eTierra y Libertad\u201c von Idelfonso Carrara (?) von 1930.\r\n\r\nRicardo Flores Mag\u00f3n (* 16. September 1874 in San Antonio Eloxochitl\u00e1n im mexikanischen Bundesstaat Oaxaca; \u2020 22. November 1922 im Bundesgef\u00e4ngnis Leavenworth im US-amerikanischen Bundesstaat Kansas) war als Journalist, Gewerkschafter und Literat ein f\u00fchrender anarchistischer Theoretiker und Aktivist, der die revolution\u00e4re mexikanische Bewegung radikal beeinflusste. Mag\u00f3n war Gr\u00fcnder der Partido Liberal Mexicano und Mitglied der Industrial Workers of the World.\r\n\r\nPolitische Biografie \r\nJournalistisch und politisch k\u00e4mpfte er und sein Bruder sehr kompromisslos gegen die Diktatur Porfirio Diaz. Philosophisch und politisch orientiert an radikal anarchistischen Idealen und den Erfahrungen seiner indigenen Vorfahren bei der gemeinschaftlichen Bewirtschaftung des Gemeindelandes, machte er die Forderung \u201eLand und Freiheit\u201c (Tierra y Libertad) popul\u00e4r. Besonders Francisco Villa und Emiliano Zapata griffen die Forderung Land und Freiheit auf. Seine Philosophie hatte gro\u00dfen Einfluss auf die Landarbeiter. 1904 floh er in die USA und gr\u00fcndete 1906 die Partido Liberal Mexicano. Im Exil lernte er u. a. Emma Goldman kennen. Er verbrachte die meiste Zeit seines Lebens in Gef\u00e4ngnissen und im Exil und wurde 1918 in den USA wegen \u201eBehinderung der Kriegsanstrengungen\u201c zu zwanzig Jahren Gef\u00e4ngnis verurteilt. Zu seinem Tod gibt es drei verschiedene Theorien. Offiziell starb er an Herzversagen. Librado Rivera, der die Leiche mit eigenen Augen gesehen hat, geht davon aus, dass Mag\u00f3n von einem Mitgefangenen erdrosselt wurde. Die staatstreue Gewerkschaftszeitung CROM ver\u00f6ffentlichte 1923 einen Beitrag, nachdem Mag\u00f3n von einem Gef\u00e4ngnisw\u00e4rter erschlagen wurde.\r\nmini|Die Br\u00fcder Ricardo (links) und Enrique Flores Mag\u00f3n (rechts) vor dem Los Angeles County Jail, 1917\r\n\r\n[...]\r\n```\r\n\r\nso some Markup like `mini|` is still left. Should I run another parser on this text before feeding it to an ML model or is this a known imperfection of parsing Wiki markup?\r\n\r\nApologies if this has been asked before.\nIf anyone else is concerned about this, `wiki40b` does indeed seem very well cleaned.","embeddings":[-0.0376301855,0.1160008237,-0.1909410208,0.4500383735,0.2954652011,-0.2197374552,0.3712628186,0.2625176609,-0.0028850846,0.0670773685,0.163732484,0.3684876263,0.0403140783,-0.0634338036,-0.0525207818,-0.2800580859,0.0010123396,0.1088692024,-0.2758776546,-0.2141533643,-0.2710891366,0.172482416,-0.2704933286,0.1990950406,-0.3585416675,0.0835002214,0.0380281396,-0.1664292365,-0.1340274662,0.0589905679,-0.1230926737,-0.0113907522,-0.2323265374,0.5959078074,-0.0001038565,0.093801789,0.2659810781,-0.14833574,0.1651000977,-0.2740632892,0.0381591879,0.1320966035,-0.018274622,-0.4883696437,-0.1506610513,-0.0973137245,-0.102646552,-0.0080199372,0.1642289311,0.1089652106,0.2779662907,-0.2608303428,0.0990448594,-0.0815536305,0.4071276188,-0.0391591825,0.0021229023,-0.056793265,-0.1048759297,0.0740148202,0.0853269771,0.4533440173,-0.0669130534,-0.1527017206,-0.0390591733,-0.0801407471,0.1149623245,-0.2299122661,0.3598327935,0.3970839679,0.4823433757,-0.237019226,-0.0033172935,0.1349339485,-0.0562789328,-0.0627484247,0.1382213533,0.2485168725,-0.2463987023,0.1998824328,0.108613342,-0.0425026454,-0.1570647359,0.3849435747,-0.3486666977,0.7615579367,-0.1470627636,0.0890357867,-0.0713838562,-0.0618577339,-0.2655573785,0.0647552907,0.406740129,0.1151141599,0.2293580621,0.0623623431,0.0952006355,-0.1879187226,-0.1581106782,-0.2297959328,-0.3073724508,-0.0305526275,-0.0009028913,0.0312914886,0.0951704159,-0.0677764118,-0.1882221997,0.1448563337,0.1271027476,0.1956168562,-0.3205268085,0.1341058016,-0.0726409927,-0.05885702,-0.0109663522,0.0493914671,0.1880798042,-0.1510051042,-0.0501707979,0.1197864413,-0.4061352015,-0.1329094023,-0.2625003159,0.3644146621,-0.3537258208,0.3217244744,0.1339441836,-0.0410714298,0.0427075326,0.2019065171,-0.3462704122,0.3184956312,-0.2574843168,0.0214953125,-0.0806169435,-0.114391923,0.326862514,0.0183948055,-0.0553089455,-0.3175844848,-0.0900209397,-0.4657154083,0.3773736358,0.1164557636,-0.137309745,0.085368216,0.1821402758,-0.4203641117,-0.0611058399,0.0739957839,-0.0117768524,-0.0430926643,-0.1845549345,0.2914001942,0.1253378987,0.114219144,0.2324087024,0.3484160304,0.3461151123,0.1279000342,0.2048683316,0.0798774436,-0.2936294675,-0.1988723129,0.3602199852,0.1398384422,-0.4822925329,-0.1239366159,0.1542090476,0.1174046993,0.4012049735,0.090257898,-0.0307218339,0.5387797952,-0.0575563423,0.2503638566,0.0693088174,-0.1390769929,0.0177722927,-0.0094972653,-0.0056565078,-0.0330239944,0.2078623325,-0.0513774194,0.193570137,0.2120351493,0.5239936113,0.0795123503,0.2428410947,0.1865055859,-0.1770028919,-0.1695266068,0.1912652701,-0.0131387552,0.178401649,-0.0867691785,-0.065780662,0.2785589397,0.3917039037,-0.0177074671,0.03367972,0.2748847008,-0.1571976691,-0.0402402021,0.1508070529,-0.1868496239,0.2586314678,-0.0947320834,0.3009767234,0.3739810288,-0.2793432772,-0.321506232,-0.4420488477,-0.0306167658,-0.0859380737,-0.125920251,0.3621974289,-0.0075938813,-0.1433638334,0.1667322516,0.1475275755,-0.04814687,-0.4452185035,-0.0623425022,-0.3474645615,0.2008005828,-0.2842546701,-0.0703957006,-0.1780652106,0.3738847673,0.0864047855,0.0987277552,0.0366893709,0.0681396648,0.1614260674,0.1951446533,0.0872016698,0.1368025839,-0.1354448497,-0.3788892925,0.1041647419,0.4465124011,0.2120099366,-0.087688081,-0.2107884735,0.4056066871,0.2071245313,-0.0240108278,0.1008186489,-0.2186525911,0.1274811625,0.0669870749,-0.1606174409,-0.3407888114,-0.0228196643,0.353531152,-0.1206595004,0.2333276272,-0.3192092776,0.0747786462,0.5386906266,0.134352088,0.4133019745,-0.0901700556,-0.5303305387,-0.1579623073,-0.0161011554,0.1799594164,-0.1050404757,0.187538743,0.1714820713,0.0786534324,-0.0189090725,-0.1408509463,0.3554648459,0.2964965403,0.4085935354,0.271145761,0.0359766521,0.168243885,-0.251855582,-0.0341589823,-0.209574759,0.379539758,-0.2574423552,0.1205544397,-0.1207627803,-0.0712015107,-0.0884926394,-0.2095499933,-0.2182871401,-0.137300387,-0.0413503908,-0.1236971617,0.0427570492,0.2243424654,-0.1203943491,-0.0427134782,0.250151068,-0.1579011232,-0.0263582114,-0.2779904008,-0.0228662621,0.1925877333,-0.192516908,0.1851531714,0.1140914857,-0.1133456752,-0.3316794038,0.0247598588,-0.4591087997,0.3076533079,-0.2841238976,0.0178090855,0.1315817088,0.2630504668,-0.1862816513,-0.298184514,0.1260712594,-0.1022754908,-0.0718508959,-0.0729202703,-0.2129737437,0.1151741818,-0.0549499691,-0.6069825292,-0.184398517,-0.2522272468,0.1753420532,0.4133389592,0.1765688062,0.4491932094,0.2899242342,0.1499410123,0.2895770371,0.1858320087,-0.36908108,-0.0820946619,0.3819562197,0.1831923574,-0.1003113538,-0.0403756425,-0.1214303225,-0.1555097997,0.03503149,-0.3284410238,-0.0780139491,0.1545450389,0.009956385,0.0603399575,0.0701203197,0.3381595612,0.0599646047,-0.237380594,-0.0741835982,-0.0543027446,-0.0672510192,-0.1804206967,0.1228020489,-0.2215511054,0.1971711069,0.0745376498,0.56469661,-0.0219051447,-0.0945554003,-0.1301510632,-0.1098464206,0.0467870198,0.0142612727,-0.1725403517,0.1183864325,-0.0657217801,-0.1128841639,0.285828799,0.0218980443,-0.3113709688,-0.0823392123,0.2434457988,-0.5951212049,-0.233756721,0.073885493,0.2072135806,0.3250167668,-0.2242445201,0.0181401148,-0.1529426575,-0.1985205412,-0.0743087381,0.2720814943,-0.1294915974,-0.1800871342,-0.003993832,0.0455366112,-0.3112497926,0.0897599682,0.106302321,-0.2374592274,0.0194009636,-0.3281236887,0.1842596233,-0.1281688064,0.2043621242,-0.164046824,-0.1016448587,0.2306401879,0.0266481508,-0.2165665627,-0.017609857,-0.3630180061,0.1319010556,0.5669853091,0.4043781161,-0.4217731953,-0.1530512571,0.5155639052,-0.0920582935,-0.0368700549,-0.1138465926,-0.2131608129,-0.1064262241,-0.2825685143,0.2626232505,0.0399504974,0.1282545775,-0.0002701624,-0.144147411,0.3049248755,0.3409052193,0.1048147157,0.1139119118,0.1659893095,0.1508635283,-0.2560137212,-0.2157394439,0.0827042684,-0.3168414533,0.1227055416,0.0964087173,-0.0023286266,-0.0933877677,-0.4582925141,0.1420152485,0.2253702581,-0.1954173446,0.2062242776,0.1236633286,0.0487761311,0.0111880777,0.1006363705,0.3577772677,-0.1785044521,-0.0160647426,-0.3824625015,0.3698531985,-0.0442784615,-0.1138214618,0.0079461094,-0.3191575706,-0.2880116105,0.2350889146,-0.0233705901,1.0900639296,0.1916282028,0.0601801537,0.0798477754,0.3357889652,0.4070530236,0.0503851138,-0.0357324816,-0.4293996394,0.183091566,-0.109221518,0.1870466024,-0.0323119424,-0.0063557001,-0.3201174736,-0.0288609788,-0.2233071029,0.0416371636,-0.0022941369,-0.2610445321,0.2794048488,-0.0714469925,-0.1940616369,0.1843516082,-0.0085481983,0.0132442368,0.008396632,0.1159456819,-0.2418378592,-0.1455354989,-0.3945600092,-0.0565445051,-0.3439761996,0.2061203867,0.1736727655,-0.4040267467,-0.123506479,0.0324687175,0.1290130615,0.3851498663,-0.0401921049,0.1245268807,-0.0149954855,-0.115345493,0.1051421538,0.12752074,0.302090615,-0.059637025,-0.3325281739,0.0474285409,-0.0684778765,-0.2840799987,-0.1140778735,-0.1202758998,0.1880212575,-0.0775806159,-0.0944027007,0.2715235651,-0.165989086,-0.3825518191,0.220366165,-0.122352846,-0.2146441489,0.1128939092,0.2487442642,-0.4793298244,0.0701411888,0.267894268,0.3993580937,0.2682992518,0.3634387255,0.4402483702,-0.3178363144,-0.4193708599,-0.3453825712,-0.344388932,-0.3034403026,-0.2273005992,0.127565816,-0.0442080237,0.0322743431,0.2528116107,0.1267941445,-0.0456003658,-0.1809589863,-0.4442609549,-0.1378889382,0.1769991368,-0.0078211483,0.1069375947,0.26016891,-0.0003516951,0.1314752996,0.0718079805,-0.4596363008,0.0954610333,-0.401098907,0.2911189795,0.249866128,-0.2424764186,-0.1911943257,0.0079129813,0.3404124975,-0.0843356848,-0.1227236539,-0.4186241329,-0.1275819242,0.0791985765,0.0390139632,-0.2021325827,0.1440882385,0.1327942461,-0.2354942858,-0.1377093643,-0.0916096568,0.1327189654,0.1528631598,0.2610372007,-0.0577553473,0.3609778583,-0.2998996377,-0.2252101302,0.0453321673,0.2198771834,0.0875917152,0.1392415464,0.114196144,-0.0657399222,-0.3645977676,0.0850177333,-0.0203922708,0.3905954957,0.3512126803,-0.0503478386,-0.0427762307,0.3029699922,0.2377179116,0.2442759573,-0.2000593245,-0.0955429673,-0.0412150286,0.3443334997,-0.3381033838,0.2190651596,0.4438180029,-0.2796862423,0.1534242183,0.3051404059,0.1139929369,0.0646672398,0.1722955555,0.1230759844,-0.0289411955,-0.3710559309,0.1493567377,0.0802782178,0.1446579099,0.0820689425,0.0083300704,0.3275237381,0.1921262294,0.4157094359,0.5534650087,0.0458230227,0.0485229902,0.0934075266,-0.2124032825,-0.1737228483,0.3782322705,0.250852555,-0.3872358203,-0.0425766632,0.0690187216,-0.0957765654,0.1958178878,0.220619455,-0.1258994639,0.2072698921,-0.1061266288,-0.2802641094,-0.2669960558,-0.216238603,-0.1254568994,-0.2240758389,0.1985271573,0.0819865987,0.0428587534,0.0007927907,-0.1652704477,-0.3171619773,0.1963714957,-0.3757813871,0.3776463568,-0.247577399,0.102629438,-0.1802186221,0.2640243471,-0.3552978337,0.4572235644,0.1867239326,-0.0397551209,-0.3567960858,-0.1345352679,-0.1819654256,-0.1360786259,-0.1003686562,-0.0508694798,0.4067136049,0.0049415873,0.1646077186,0.2986797392,-0.3479203284,0.0352781601,0.3535931706,0.033140827,0.0874756426,-0.0871276259,0.1985788047,-0.1533524394,-0.1870585382,0.0449051037,-0.3499832749,0.2425706387,0.3185038567,0.1421524733,0.2089875787,-0.1796865165,0.1499054283,0.0874758661,0.1393532902,0.4506242573,0.1468863636,0.0160898175,-0.1277967989,-0.2883241475,0.3457967937,0.0065063708,0.038130641,0.0090687796,0.3338782489,0.2970461547,0.1066390052,0.0200609546,-0.255394429,0.0714398623,0.0563794933,-0.1783145666,0.084132269,-0.0173060037,-0.1360272318,-0.0498393103,-0.1777641773,-0.1465191692,-0.2832279503,0.303167522,-0.0624727495,-0.3920107782,0.0241401959,0.1590516716,0.3841212988,-0.1318282038,0.1754274219,-0.3827268779,0.0820533484,-0.0026540465,-0.0795917809,0.0105235148,0.5436720848,0.1163988337,0.1178999245,0.1743744314,-0.0268047359,-0.0146710789,0.2737036645,-0.0396735817,-0.3405407071,-0.1938305199,-0.0594838709,0.0699940398,0.0680811852,-0.2632058263,0.1269263476,-0.1516225487,0.0331990793,-0.2077036351,-0.5855808258,0.2162085176,-0.0780524611,-0.5391718149,-0.3936989605,0.1647264957,0.0663961694,0.4991380274,-0.4331305325,0.1830520779,0.3660574555,-0.1315494329,-0.3380313814,-0.0578141324,-0.1394308209,0.1607730985,-0.1068724319,0.0511503443,-0.2077844143,-0.1916946918,-0.259087801,-0.3279494941]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/834","title":"[GEM] add WikiLingua cross-lingual abstractive summarization dataset","comments":"Hey @yjernite. This is a very interesting dataset. Would love to work on adding it but I see that the link to the data is to a gdrive folder. Can I just confirm wether dlmanager can handle gdrive urls or would this have to be a manual dl?","body":"## Adding a Dataset\r\n- **Name:** WikiLingua\r\n- **Description:** The dataset includes ~770k article and summary pairs in 18 languages from WikiHow. The gold-standard article-summary alignments across languages were extracted by aligning the images that are used to describe each how-to step in an article.\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2010.03093.pdf\r\n- **Data:** https:\/\/github.com\/esdurmus\/Wikilingua\r\n- **Motivation:** Included in the GEM shared task. Multilingual.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":48,"text":"[GEM] add WikiLingua cross-lingual abstractive summarization dataset\n## Adding a Dataset\r\n- **Name:** WikiLingua\r\n- **Description:** The dataset includes ~770k article and summary pairs in 18 languages from WikiHow. The gold-standard article-summary alignments across languages were extracted by aligning the images that are used to describe each how-to step in an article.\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2010.03093.pdf\r\n- **Data:** https:\/\/github.com\/esdurmus\/Wikilingua\r\n- **Motivation:** Included in the GEM shared task. Multilingual.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nHey @yjernite. This is a very interesting dataset. Would love to work on adding it but I see that the link to the data is to a gdrive folder. Can I just confirm wether dlmanager can handle gdrive urls or would this have to be a manual dl?","embeddings":[-0.2052895278,0.319988668,0.0722148195,0.5395573378,0.1300372928,0.2193807662,0.1277997494,-0.1864241362,-0.0289536398,-0.0699264407,0.2743760049,0.0008884232,-0.4298549891,0.2057960778,0.3081344664,-0.2922178805,0.0054489104,-0.2033134252,0.020582499,-0.1820490211,-0.1673908681,0.4652565718,0.1199145317,0.0503368936,-0.4685993195,-0.0266561564,-0.221419096,0.4315164089,-0.0926346853,-0.1953790188,-0.0345566086,0.039644599,0.1316577494,0.4354016185,-0.0001247795,-0.157182321,0.2606930733,-0.2081924975,-0.0927850381,-0.2235533744,-0.3300047517,-0.1996063143,-0.1916079819,-0.1280442029,0.175924316,0.128310591,-0.0804150552,0.1652437001,-0.2902439833,0.0426329561,0.1184955612,-0.0546926484,0.177083537,-0.0281240065,0.3376811445,-0.2312233597,-0.0272024591,0.5546996593,-0.0733492374,0.1199260205,-0.1226686612,0.4144370556,-0.1018918827,-0.1896880269,0.0083972318,0.2109396905,0.0017848539,-0.0774430335,0.1375340819,0.2818035483,0.816493392,-0.23187159,-0.0623929389,-0.1128107756,-0.1559919119,-0.1765374243,0.4124311805,0.161280483,-0.0271767341,0.0534955487,0.0535362288,-0.5035980344,-0.1352059394,0.2228758782,-0.0204972029,0.4011122882,0.065344125,0.0235374309,0.4362442195,0.2480707169,-0.1870647073,-0.096568279,-0.3789028823,0.1565636396,0.0187848099,-0.2509078085,0.0238386672,0.2325567156,0.114207454,-0.0979505479,0.0392102003,0.0481207408,-0.0954174697,0.0808626562,0.3494334817,0.2302668393,-0.2853442132,-0.0323552191,0.2578540742,-0.1746076494,0.1903271526,-0.0247632414,-0.1124453545,-0.1519078612,-0.3518644869,-0.1581168622,-0.3366399705,0.1224362776,0.0758734867,0.1092616916,-0.224985078,-0.2128604949,0.1006018296,0.0871745571,0.2883945405,-0.2428627759,-0.0612188093,0.3998378217,0.2928891778,0.1777856797,0.0584300682,0.2694648504,-0.0828705728,0.2098586559,-0.1889800131,0.084674865,0.124035351,-0.085408546,-0.2084121704,-0.1505706161,0.0108843315,0.0665663704,-0.183389768,0.1558983624,0.3799422681,0.3868770897,0.0069782999,-0.6324939132,-0.3096870482,0.1332536936,-0.171611011,-0.3060252368,-0.2727497518,-0.0214119256,-0.2076001018,-0.3338611126,-0.2829799056,0.5032006502,0.0978399888,-0.3781301975,0.0974009335,0.425876081,0.0381935351,-0.1741587222,-0.2004806399,0.3255687058,-0.2859554291,0.0491206385,-0.2233881056,-0.0063545676,0.3607916832,0.2327745706,-0.0493528396,0.0860492662,-0.2695146501,0.4979097545,0.2523216903,-0.1120953634,-0.1419781446,0.1514694542,-0.4574857056,0.0134938266,0.192811653,0.353248328,-0.07077948,-0.1276389062,-0.1220430434,0.2550140917,0.1740980744,0.0441155881,-0.3670468628,-0.436099261,0.227699846,0.1296114326,0.2096010894,-0.141670242,0.3847234547,0.3507202268,0.2013375163,-0.25459373,0.1432922035,0.1356197894,0.2010868341,0.0470831208,-0.070095621,0.0878504664,-0.4258346856,0.0398113281,0.484170258,0.4287417233,0.2425023317,-0.0437106267,0.1110672206,-0.0640633255,-0.21433492,-0.1703340262,-0.0034962082,0.39950791,-0.0788112357,0.2376496047,-0.0732419491,-0.2341186255,-0.2306239307,-0.1214824617,0.0381963402,0.3170720041,-0.1187933236,0.0798900649,-0.1165798977,0.2972486317,-0.2057968974,-0.0472994857,-0.0345139541,0.1623503566,-0.1037028581,0.3616391122,0.6724396348,0.7141319513,0.0822548419,-0.2385071814,0.3458833992,-0.0627142712,0.2449794859,-0.1494480968,-0.0729832724,0.2659508288,-0.0039396649,0.0668013692,0.0737567842,0.2355251014,0.2932620049,-0.019925477,-0.0440586656,-0.3305786848,0.3704812825,0.2180639505,-0.2417214364,0.1267597973,-0.2977869809,0.0076664374,0.4142896533,-0.0502846688,0.4785454273,0.218574062,-0.3042879999,-0.3060279489,-0.0249244981,0.610376358,0.3237772584,0.0808550045,0.378413111,0.0617275313,0.1076251939,-0.1882869005,0.365808636,0.0872485787,-0.0782542452,0.5172120333,-0.1013433263,-0.0657736734,-0.2807046473,0.1297214925,0.0327551328,0.2047456503,-0.149987638,-0.0216756314,-0.0512133278,-0.1157659367,-0.1355130523,-0.7797605991,-0.6879051328,-0.3252259791,0.1452034861,-0.1766125262,-0.1035038456,0.0233518053,0.1382453591,0.1964084208,-0.4809713662,-0.3989130259,-0.087769106,-0.4259753823,-0.1295738816,-0.0361106209,0.4199083149,-0.0313053317,0.291648984,0.019221358,0.1689298153,-0.2439244241,-0.4900278747,0.3283202648,-0.1685169786,0.2541861534,0.0756706521,-0.1180241033,0.0898889452,-0.4692749083,0.1869605184,0.255725354,0.1016540229,-0.0497036763,0.0046757781,-0.0229755584,0.045001559,0.0411935225,-0.294365108,-0.2600292265,-0.0858576968,0.0822593868,0.1040048003,-0.2154649347,-0.0400587469,0.2880854905,0.0635915697,-0.1582787782,-0.3831860423,-0.1378221214,0.3207927346,-0.1182926521,-0.246581316,0.272031188,-0.2453189045,0.3434216082,-0.0815165266,-0.4744024277,-0.2974785864,0.1456634253,-0.0116080921,0.4878692925,0.1211170778,0.3266866505,0.1165993288,-0.0131175695,-0.0340050459,-0.2883881927,0.1065922976,0.3647940159,0.2817761898,0.1308726817,0.3322602212,0.1952828765,0.6535807848,-0.0746081769,0.23401995,0.2576474547,-0.2271134704,0.3407444358,-0.0939052701,-0.101073496,-0.0377458967,-0.1238515228,-0.4158323705,0.3140842915,0.2167859524,-0.5161979198,-0.0897718295,0.0016283659,-0.2928319275,-0.2946752906,0.2758501768,0.0945888162,0.0241470281,-0.1234209687,-0.2413083464,-0.0503906533,-0.2646073699,0.1166996211,0.8497219682,-0.2143952996,0.0782408342,-0.3703143001,-0.2512573898,-0.1082792729,0.1750403345,0.2574891746,-0.0486550592,-0.3461049795,-0.1142122298,0.1585990638,-0.3126897812,0.1028868333,-0.5930868387,-0.1834145337,0.0585017912,-0.1729306281,0.4593811333,0.2165704519,0.0346569158,-0.2776293755,0.5621849298,-0.231465891,-0.231191054,0.0351758599,0.2557694018,0.0068298881,-0.2277309,0.171949774,-0.0879821181,-0.4606232047,-0.4878263175,-0.2078489661,-0.0419357456,0.0494095981,0.2344489992,0.1014161408,0.0381707251,-0.1455045044,0.0297031924,-0.1796230823,0.0404421575,0.1885642707,-0.2332445234,0.2494195551,0.1618937254,-0.0151094561,0.2166644037,0.1174060777,0.3896990716,0.1741852909,-0.5150285363,0.6288019419,0.3609160483,0.0863166377,-0.0762051791,0.1040381193,-0.2789347172,-0.3891699612,0.0016888053,0.0844886675,0.2216775417,-0.2243632227,-0.2750700414,0.5395458341,0.0345623344,-0.2513055205,0.1492883712,0.2535794377,-0.4851569235,0.0559551977,0.1228714585,1.2584296465,0.1731718779,0.2715183198,0.1733859181,-0.1096202284,0.0859105587,0.0561419614,0.1998311877,-0.3550165892,0.555221796,-0.2270050645,0.0451040044,-0.1872153282,-0.0858415514,-0.1615121067,-0.0376487896,-0.0489198565,0.0934512988,-0.1644868106,0.3035354018,-0.2491308749,-0.2379184812,0.0487553105,-0.0177878663,0.2321211249,0.2364140004,-0.1232843921,-0.4004274607,-0.0926724896,-0.3229565918,-0.7480203509,-0.1462262571,0.016425889,-0.0247239172,-0.4567170143,-0.5324651003,-0.1003825441,0.3304934502,-0.2378312945,-0.1338420212,-0.1963004321,0.0218865965,-0.5718806386,-0.0543903857,0.2409208417,-0.0025459428,0.1420661062,-0.1023655757,0.1303026378,0.0192631967,0.0046669771,-0.4703747332,-0.1594414115,-0.1892676055,0.0361338519,0.245529145,-0.0258021224,0.4970343411,-0.1906968802,-0.2556156218,0.0335374288,0.0721829534,0.1186013222,-0.0151517848,-0.0999378711,0.0129912626,-0.1618505716,0.2606792748,-0.0799250528,-0.2677699625,0.1365800947,0.2739425302,-0.253998071,-0.1561606377,0.0496803112,0.0289192591,-0.2983618677,-0.1093276516,0.1146746352,-0.1720047295,-0.1853491068,0.2532095909,0.305318594,0.009273354,-0.3213846684,-0.4174228907,-0.1386594623,0.3139130175,-0.1347519308,0.3784938157,-0.7157737017,-0.3194902837,0.2560022175,0.0739576444,-0.1883509457,0.0577641092,-0.1561155915,-0.1509699225,0.6616249084,0.1179491654,0.3081755936,0.1688100696,-0.0610070825,-0.4879081547,-0.1559155136,0.0641437173,-0.1242694557,0.214111194,0.4191162288,-0.0905523673,0.0314500779,-0.3898132741,-0.133720845,0.1204042435,0.2108095884,-0.1571482718,-0.2858671844,0.0173228178,-0.1569104046,0.2881869674,-0.1998731792,0.0762178078,0.1188787669,0.5915831923,0.1583921164,-0.031415984,0.0030102241,-0.0456664972,0.4624197781,0.4269456267,0.1892706305,0.0728677884,0.1712717712,0.0155881746,0.1197737381,0.3151093423,0.1932163984,0.0145007381,-0.1390069872,-0.0433963612,0.184990555,0.0005184761,0.0293462947,-0.3166526854,-0.2396707833,-0.1344020814,-0.0602646843,0.228508845,0.4845857918,0.0489013642,-0.1327949613,-0.1768658161,0.5035765171,-0.2242107689,-0.1081035212,0.113253139,-0.479501754,0.5232340693,0.1773428619,0.3323850036,0.2191063315,0.1033021212,-0.1309674978,0.0549307801,0.3881404996,0.0142446021,-0.0708748177,-0.1977723986,0.0558009818,0.2556618154,0.3525903225,-0.0198867023,0.094900921,0.0365249775,0.2033006251,0.4621016085,-0.2480401546,0.2013392448,-0.0392635129,-0.0765961632,-0.2275441587,-0.2805992663,-0.325170666,0.4158680737,-0.2221699208,-0.1813543588,0.0556773692,-0.0235716291,0.2005225718,-0.1309473962,0.2222487926,-0.0271433853,-0.3678202331,-0.2760508358,0.0957401022,0.1604192108,0.1875728369,0.0119545609,0.0505943261,0.1883041263,0.0912700668,-0.0713561922,-0.0708181113,-0.2897806466,0.188542366,-0.1639314145,0.5140028596,0.1508898437,0.0002554808,0.1288577169,-0.0240112413,-0.083444804,0.1777041256,0.3646956682,-0.0815741941,0.2900409102,0.4989517033,-0.0069330824,0.0616487302,-0.1075539216,0.0119593488,-0.2995429635,-0.1839977503,0.4080570042,-0.0861745179,0.1035358012,-0.0198865682,-0.0276099313,0.2268510163,0.346698761,0.165162921,0.2325510681,-0.1531082094,-0.1460170448,-0.5308949947,-0.0676825121,0.3016186059,0.0427079648,-0.190417394,0.1468811631,0.2609976232,0.3723931909,0.1010961607,0.2840654552,-0.0358070061,-0.0369493626,0.0060530249,0.2983292639,0.2434145957,0.1828011274,-0.2755824625,0.0127733238,-0.0651872009,-0.1498203278,-0.0901640505,-0.2980970144,0.1432217658,-0.3748915493,0.1640306115,0.3807445765,-0.1049274877,0.5162321925,-0.2271387875,0.0607254766,-0.3134492934,-0.0631872714,-0.1761236042,-0.1492343247,-0.025673015,0.0878186449,-0.2892702222,0.4012246728,-0.1729557812,-0.1698646843,-0.0885790437,0.5634937286,-0.1865979731,0.0859000385,0.0004585758,0.0216975566,0.4482543766,0.2392441779,0.1635235101,0.0456233732,-0.2102030069,0.0588907339,0.0938397869,0.045802664,-0.0469336472,0.0148063609,0.1403072476,0.3675567806,0.1528910249,-0.3563582003,-0.099365212,0.2028649002,0.2868802845,-0.188631013,0.1483278126,0.2838778794,-0.3120300472,0.0277509205,-0.3410733044,-0.05031313,-0.2796905637,-0.1944033802,-0.260997504]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/834","title":"[GEM] add WikiLingua cross-lingual abstractive summarization dataset","comments":"Hi @KMFODA ! A version of WikiLingua is actually already accessible in the [GEM dataset](https:\/\/huggingface.co\/datasets\/gem)\r\n\r\nYou can use it for example to load the French to English translation with:\r\n```python\r\nfrom datasets import load_dataset\r\nwikilingua = load_dataset(\"gem\", \"wiki_lingua_french_fr\")\r\n```\r\n\r\nClosed by https:\/\/github.com\/huggingface\/datasets\/pull\/1807","body":"## Adding a Dataset\r\n- **Name:** WikiLingua\r\n- **Description:** The dataset includes ~770k article and summary pairs in 18 languages from WikiHow. The gold-standard article-summary alignments across languages were extracted by aligning the images that are used to describe each how-to step in an article.\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2010.03093.pdf\r\n- **Data:** https:\/\/github.com\/esdurmus\/Wikilingua\r\n- **Motivation:** Included in the GEM shared task. Multilingual.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":42,"text":"[GEM] add WikiLingua cross-lingual abstractive summarization dataset\n## Adding a Dataset\r\n- **Name:** WikiLingua\r\n- **Description:** The dataset includes ~770k article and summary pairs in 18 languages from WikiHow. The gold-standard article-summary alignments across languages were extracted by aligning the images that are used to describe each how-to step in an article.\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2010.03093.pdf\r\n- **Data:** https:\/\/github.com\/esdurmus\/Wikilingua\r\n- **Motivation:** Included in the GEM shared task. Multilingual.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nHi @KMFODA ! A version of WikiLingua is actually already accessible in the [GEM dataset](https:\/\/huggingface.co\/datasets\/gem)\r\n\r\nYou can use it for example to load the French to English translation with:\r\n```python\r\nfrom datasets import load_dataset\r\nwikilingua = load_dataset(\"gem\", \"wiki_lingua_french_fr\")\r\n```\r\n\r\nClosed by https:\/\/github.com\/huggingface\/datasets\/pull\/1807","embeddings":[-0.2764140666,0.0026880144,-0.1001025662,0.3107095659,-0.1134149954,0.1402342021,0.0106717795,0.2736697793,-0.0497234203,0.142544508,-0.0666124299,0.2388452888,0.0332172699,0.3290542066,0.1007773429,-0.5720511079,0.0124257961,0.0506400205,-0.189469412,-0.0938588232,0.0496565029,0.2917680144,-0.1541137397,0.034482073,-0.0742902011,0.1673282683,-0.2920971811,-0.0621645786,0.0580645055,-0.1694517136,-0.049450133,0.1079933122,0.1147315726,0.1120782569,-0.0000970769,-0.235713914,0.1173490658,-0.0934773833,-0.0832296833,-0.3354904652,0.0568946898,-0.4867471457,-0.0898417607,-0.152729094,-0.0286800899,-0.0799354091,-0.0225389898,0.3618356287,0.1711872816,-0.0163235459,0.30158481,-0.2413649559,0.1877074689,-0.0982502475,0.0409793444,-0.3352949619,-0.187696889,0.1640687734,-0.0243242867,-0.0562591292,-0.2111863643,0.5608845353,-0.0509362333,-0.2531584799,-0.2900612652,-0.0126946662,-0.1052725241,-0.0649602041,0.1165637448,0.4845497906,0.529859364,-0.1681496799,-0.2733721137,-0.0119785434,-0.1028654203,-0.1336417496,-0.0689065158,0.1995877922,0.0682092905,0.2821740508,0.1197725609,-0.25279212,-0.1005490199,0.2251769602,0.1247845069,0.4762441218,-0.0582085103,-0.0441962518,0.2988864183,-0.1675044298,-0.5653138161,0.1359974742,-0.107461445,0.2373199612,-0.0855726525,-0.1235274225,0.1528774947,0.0799808875,0.1966228336,-0.2533028424,-0.152831316,0.4855698049,0.0506215021,0.1805730313,0.2176923007,0.175059244,0.0759467632,-0.0547261015,-0.0148441475,0.1278166026,-0.0411525331,0.0292952079,-0.0243418477,0.0643674061,-0.2445610017,-0.0819883272,-0.4378207028,0.0686352327,0.2060537487,0.2339699119,0.1268783659,-0.1293237656,-0.1584433019,0.3530856967,0.0957632139,0.2372762561,-0.0781104341,0.2695079446,-0.2140510082,-0.1261201203,-0.123630695,0.2545527816,-0.1758843064,0.1946148276,-0.1769746095,0.0841224939,0.2207692266,0.0119759105,0.2674368322,-0.191990748,0.0756382048,0.0522559807,-0.0806918144,-0.2325192541,0.2944804132,0.0688738376,0.253816098,-0.4139411151,-0.2517581582,-0.1296180189,-0.2286814153,-0.0899842232,-0.3403100967,0.2785220444,0.0378572978,-0.2425411195,-0.2779177129,0.7183287144,0.1214745641,-0.3903105557,0.0970283002,0.2184337378,-0.2372532934,0.02628747,0.1079149693,0.4262573719,-0.0857505426,-0.1368868351,-0.0094694961,-0.0878258795,0.0817494914,0.1771855205,-0.0713140368,-0.1179260164,-0.1324967295,0.4285255075,0.1194244251,-0.1509241015,-0.1736158282,0.0201856829,-0.1053154841,0.2885639369,-0.0036446585,0.0764425918,0.0263945311,-0.0283590816,0.1301962733,0.1024601087,0.1560315788,0.1353257,-0.2220541686,-0.327780813,0.0901897177,0.1936070323,0.0324617438,-0.1908340007,0.2195291519,0.1991110891,0.2937733829,-0.2792626619,0.1236492842,0.2130652368,-0.1529013813,0.1465049088,0.0546806157,-0.2641773224,0.0251854416,0.1113230437,0.3206298053,0.3576340377,0.2289250493,-0.2382941544,-0.0464861356,-0.1705206931,-0.3059718907,-0.2131436914,0.3340744972,0.1207557693,0.0298732203,0.3504917026,0.0571480058,-0.2722002268,-0.3950471282,-0.2795887887,-0.0788680017,0.0020740025,-0.1584918946,-0.0642237365,-0.0997120813,0.3893690705,-0.2712367773,0.0983206928,0.0154334633,0.1517489552,-0.0550842546,0.4982587397,0.4569424391,0.2020403147,-0.0717084259,-0.5005124807,0.1014631912,-0.1269534379,0.0405575447,0.0141751384,0.0026556414,0.4599095285,0.2683020234,0.1284926683,0.1728938073,0.077689968,0.438174516,0.064910844,-0.1373220682,-0.3299251497,0.3048838079,0.0782761425,-0.4333210886,-0.122109659,-0.1712439656,0.1732613593,0.5744444132,0.1377800256,0.5119507909,0.2164384127,-0.0974679589,-0.204381749,-0.0144237485,-0.0558188856,-0.173418045,0.2701702416,0.3225536346,0.1384960711,0.1292924732,0.0410367325,0.2512614727,-0.0856256187,0.0452633686,0.1973093301,0.0913208649,-0.1187370792,-0.3444887996,0.0685833469,-0.2377658188,-0.0307473838,0.1356806457,0.1075929031,-0.0717532188,-0.0445475504,-0.1838493645,-0.3029926717,-0.2344042808,-0.161650151,0.2162353396,-0.2200760543,-0.240059346,0.128452763,0.1167839691,-0.1129344478,-0.2529805005,-0.1024315432,-0.01384304,-0.5788555741,-0.2555541098,0.1801484823,0.048232194,-0.0450079776,0.3994951248,-0.4012573063,-0.1552133411,0.0356057584,-0.6449292898,0.2720446587,-0.1622140706,0.020054061,-0.0106310388,-0.0697962418,-0.3605292141,-0.3130759299,0.3076897562,0.0325359292,-0.0289600249,0.1261402518,0.0377021208,0.0117119923,-0.035280969,-0.5705169439,-0.1980738342,-0.407093972,0.3077458143,0.0545626767,0.1486679316,-0.3104326427,0.2955928445,0.1908233464,-0.1759011447,0.3428914249,-0.3420238495,-0.0923003629,0.1120934337,-0.1527304798,-0.3790644407,0.0731656253,-0.2005584538,0.3650987446,-0.1057616845,-0.2588934004,-0.1058176011,0.179116264,0.0326339118,0.1819611192,0.2368079126,0.2737877071,0.1304390579,-0.2074815631,0.0725155026,0.0377643704,-0.0318166912,-0.1206884682,0.0855492949,-0.2100822926,0.3565778136,0.2325437665,0.6987831593,0.1690097153,0.3995890915,0.3031805456,-0.1493190527,0.121412836,-0.3727869987,-0.1183328182,-0.066765435,-0.1618153006,-0.0017637017,0.4374344945,0.2114104331,-0.6284598708,-0.1920642406,-0.2594239414,-0.4223432243,-0.2361973822,0.3057610393,0.0923373923,0.3545749485,-0.1009262279,-0.2588641346,0.0151338661,-0.1767124087,0.1695002019,0.3979317546,-0.2794201672,0.1074101627,-0.4918294251,-0.1902907342,0.067383945,0.1728888303,0.4474493265,-0.0807947293,-0.220544368,-0.1300716847,-0.1302406788,-0.2977638245,-0.0314537585,-0.2745253444,-0.2566510737,-0.0925144255,-0.299352318,0.0558251403,0.0835799575,-0.1684217155,-0.3951658309,0.5567393899,0.1419668347,-0.3097411692,0.0260452069,0.1145018786,0.3811562061,-0.2633333802,0.222710833,-0.2694185078,-0.1932817847,-0.5281088948,0.1691386551,-0.0399694592,0.0465843268,0.1690196544,0.0321973525,0.0884623602,0.0804295093,-0.059489686,0.061630886,0.2686176896,0.123026602,-0.1082789749,-0.0406362042,-0.2364170402,-0.1076548547,0.2783064842,0.0775391385,-0.1844505072,-0.3306642473,-0.2572511137,0.3101328909,0.0035287873,0.2262663394,0.0107805999,0.1983656138,-0.1305654943,-0.3597178757,-0.0355608426,-0.0116043789,0.3058981299,-0.1058426648,-0.431660533,0.4495398104,-0.1073316634,0.0158915892,-0.026343191,0.2421527356,-0.2652864158,0.230311498,0.0778751299,0.927844286,0.2354873121,0.2578746676,0.103415288,0.1018179134,0.1726311445,0.0618221872,0.02043291,-0.250228256,0.2777409852,-0.0039720368,0.3077612817,-0.2442103624,-0.0057328171,-0.4587951303,0.1628588289,-0.2138059288,-0.0694860518,-0.035729792,0.0389157571,-0.3852998912,-0.2546166182,-0.0749216974,0.2158120424,0.1608697176,-0.1427343637,-0.0968836769,-0.1639650613,0.0675171614,-0.1155510768,-0.5750609636,0.2037638426,0.3043476343,0.0976548493,-0.2666789591,-0.1327326298,0.050123889,0.3451713026,-0.1938077658,0.0252293013,-0.082395494,-0.0341518708,-0.4138200581,-0.1523543447,0.1776979566,0.0573248193,0.2658069432,-0.1814178079,-0.2510643899,0.0641665533,0.1339982599,0.0557067767,-0.2997832,-0.1333728284,0.385440439,0.1851493418,-0.0951647386,0.4663962126,-0.2612830997,-0.3155130744,0.2619303465,0.0702879354,-0.2286221832,-0.110629648,0.1430110633,0.0036537987,0.0739965588,0.2212085724,0.2505519092,-0.3267936409,0.2621975243,0.3297332525,-0.2598975897,-0.3272093832,-0.2419425845,-0.1252944171,-0.3335345089,-0.1387812495,0.074893795,-0.1159901768,-0.1452168673,0.0949209854,0.031352371,0.0480561443,-0.2062579989,-0.2742919922,0.0370925441,0.3391432166,0.1259647012,0.1663532406,-0.0384915322,-0.2375499755,-0.0325014293,0.2613565028,-0.4619376063,-0.1443137974,-0.1348393559,-0.0458178408,0.4608716667,0.1276094913,-0.0003698246,0.1995986849,0.2001710534,-0.0017002134,-0.2089054286,-0.3422905207,-0.311621964,0.0522611402,0.0538801029,0.2088842243,0.1513405442,-0.1388468295,0.1814887673,-0.3803017139,0.1357378364,-0.1391914338,-0.1668485254,0.3987881839,-0.215623498,0.4115794599,-0.0930071473,-0.1541317403,0.0210989825,0.2426481694,0.0362112485,-0.0763922483,0.0233234651,-0.09440355,0.1507288069,0.2031555176,0.1740852296,0.1767477691,0.1257218719,0.2538172901,0.2154756933,0.1199661195,0.2261169851,-0.0065142247,-0.1190208644,-0.1153842136,0.1623837203,0.3639622033,-0.2695581317,-0.1094488502,-0.1899956018,-0.3089713156,0.1934010386,0.246803686,0.1579925716,-0.0653852746,0.0884058997,-0.1375381202,-0.0879687369,-0.3478445709,0.1863392442,0.1399736106,-0.0900238231,0.1239619181,-0.1945773363,-0.0240204353,0.2098167986,0.1813307256,-0.1363976598,0.3350525498,0.1722569019,0.3454723358,-0.1401737481,0.0112580182,0.3753678203,0.2206597328,0.3333486021,0.2926024795,0.0376373231,0.0302176848,0.3496354222,0.2469122112,-0.190176785,0.3346485496,-0.1067723855,-0.2746526897,-0.0019901502,-0.2964924574,-0.240092501,0.1327789873,0.1189873889,0.0631737635,0.084314689,0.1264618486,0.1882159859,-0.246555537,0.071119599,0.0024900974,0.1469144821,-0.2575463951,-0.003993996,0.3881651759,-0.015042658,-0.2881391644,0.165201962,0.2157081068,0.0387002826,-0.112126492,0.221543625,-0.0579874888,0.0988768488,-0.1475172043,0.0860487223,0.1966527104,0.2050224841,0.1241289228,0.2433704138,-0.3616892397,0.0829413906,0.2663657367,-0.0936807245,0.0070117922,0.1793234199,0.0499934815,0.3326556087,-0.244281143,0.1301928759,-0.4037368596,-0.0927733555,0.3473378122,0.1063144878,-0.0324112624,-0.1338119507,0.1636234671,0.3008716702,0.2863396704,0.2391320169,0.1717404425,0.0551128648,-0.1467788666,-0.3187097013,0.0749259517,0.1380318552,0.0064695855,0.0353381224,0.3323947787,0.2515556514,0.4170928001,-0.0292966198,-0.1962983906,-0.0174175743,0.1513733417,-0.0338280573,0.1229170635,0.1477125436,0.1513022929,0.050928127,-0.1168858632,-0.2503301203,-0.078698881,0.1806658208,-0.2451578677,-0.0248237923,-0.1587809622,-0.3033137619,0.2724943757,-0.0848014876,0.2059204578,-0.3882678449,-0.0711868703,0.109319672,-0.2837794125,-0.0184021257,0.3539358675,-0.0492296219,0.213359639,-0.0602180883,0.0124979112,0.0551262423,-0.1149764955,0.0985493511,0.1641899347,-0.1521481127,-0.010428898,0.4137108028,-0.2293925285,-0.0354563966,0.6886970997,0.0779253915,-0.0594686531,-0.0233438984,-0.1827031225,-0.0837537348,-0.1309295446,-0.2704118192,0.0565037839,0.0346853733,0.1069268808,-0.0441022366,-0.4928299189,0.0725593194,0.2490632683,0.4061501622,-0.0792971775,0.208754003,0.1567536741,-0.1536522061,-0.0504048988,-0.080055818,0.0436577573,-0.1664244831,-0.0209540259,-0.3597687185]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/830","title":"[GEM] add ToTTo Table-to-text dataset","comments":"closed via #1098 ","body":"## Adding a Dataset\r\n- **Name:** ToTTo\r\n- **Description:** ToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled generation task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2004.14373\r\n- **Data:** https:\/\/github.com\/google-research-datasets\/totto\r\n- **Motivation:** Included in the GEM shared task\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":3,"text":"[GEM] add ToTTo Table-to-text dataset\n## Adding a Dataset\r\n- **Name:** ToTTo\r\n- **Description:** ToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled generation task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2004.14373\r\n- **Data:** https:\/\/github.com\/google-research-datasets\/totto\r\n- **Motivation:** Included in the GEM shared task\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nclosed via #1098 ","embeddings":[-0.0273684431,0.1511737555,-0.1203573272,0.0459154844,0.1650583893,0.0658841804,0.2097620219,0.4703023732,-0.2653895319,0.0291339271,0.1412174702,0.4117406905,-0.0769067779,0.1481701732,0.2141355574,-0.0381730311,0.2410978973,0.0928752273,-0.2889253497,0.0190391988,0.0902766213,0.179858163,0.1408910155,-0.0626551062,-0.5825004578,0.0662437528,0.0600659326,-0.0553481318,-0.1092827469,-0.1506040245,-0.0700694025,0.2686460018,-0.2755880952,0.2215086073,-0.0001057034,-0.0157502703,0.0828187317,-0.1861425638,-0.2771243453,-0.3999198973,-0.1647598594,-0.1567646116,-0.070574142,-0.1915213317,-0.1927236915,0.4742588699,0.2011318952,0.2431647778,0.0090455143,0.0654646009,0.307929486,-0.0984701514,0.0454800911,-0.2659132779,0.4525997043,0.3961320817,-0.314366281,-0.1684805155,0.0421958826,0.0258450881,0.0590524524,0.4648965299,-0.0885527581,-0.0139328754,0.0755281001,0.1069055051,-0.3391492665,-0.0851649195,0.1862735301,0.541434586,0.469748795,-0.4126270115,-0.0823484808,0.1189172268,-0.1298076957,-0.2768389285,0.0043377848,0.2339904457,-0.1358437091,0.1911451817,0.0393925868,-0.5426699519,-0.1043765843,-0.2187154889,0.176854834,0.1644093692,-0.1577719152,-0.2073693424,0.1289696842,-0.1817041487,-0.4301338196,0.0105359629,-0.0754361451,0.0836476982,-0.1934067756,-0.3590834439,0.1566174924,-0.1044853628,0.3218886554,0.0606944785,0.1604447067,0.1759866625,-0.2662603855,-0.0301776063,-0.0109874113,-0.296440661,0.3933424056,0.0277081598,-0.1239003688,-0.0610983558,-0.1702145785,0.1460494399,-0.0566203743,0.1725394726,-0.0726233944,-0.3172284961,0.0298650376,-0.0489234813,-0.2562882304,0.3682172894,0.3315542936,-0.0239307676,-0.3630932271,0.4765580893,0.1272850186,-0.2233539522,0.0688178241,0.0503018498,-0.0046313899,-0.3582448661,-0.1178738326,0.0627911314,-0.0366172642,0.2131000906,0.0353216417,-0.2138940543,0.070053868,-0.0027298545,-0.1821658462,0.0213164967,-0.277535826,-0.1256885082,0.0956330895,0.1032300144,0.0692580566,-0.0862075016,0.1052422747,-0.4264554679,-0.2147160023,0.3555814624,0.051653143,-0.1548305303,-0.4031311572,0.2743148208,-0.2059075832,-0.34286955,-0.0876871645,0.7321739197,0.237085849,-0.212563023,0.0599825755,0.2877841592,-0.0741565898,-0.0539031886,0.0754004195,0.5199614167,-0.0316368267,-0.0667117983,0.0784641877,-0.0669503957,0.3748269081,0.2022559494,0.0893583521,0.3822464645,0.3547000885,0.2544623911,0.3761446774,0.3576936424,-0.1832738221,-0.0330463164,-0.2598951459,0.0286680739,0.1698327363,0.1543174237,0.330680728,-0.0177829973,0.1834346652,0.0091117872,-0.0550150201,0.0061148368,-0.4415458441,-0.0835697055,0.2845737338,0.0723998174,-0.0030256088,0.0318704173,-0.0728490874,0.2582190931,0.3519338071,-0.0700471029,0.2536584735,0.0427159443,-0.0717216805,0.1619864553,-0.1501300037,-0.5334439874,0.3026926816,-0.1066366434,0.4574030638,0.6266674399,0.1017384604,-0.3626384735,-0.1141485572,0.1309620738,-0.0554659478,-0.1788823605,0.2358155847,-0.1058754921,-0.0825932249,-0.0085666152,-0.1629523486,-0.2731958926,-0.4728883803,0.0767379403,0.1192253083,0.0960396305,-0.1302262694,-0.1887910366,-0.0048234537,0.2026980966,-0.2043079436,0.0986413583,0.0885612965,0.1616818607,-0.1315848678,-0.0061398377,0.4108643234,0.1390835345,-0.1665814966,-0.2770116627,0.0139405588,0.0095711024,0.1446238607,-0.1234318465,-0.3131465912,-0.0232597813,0.3800584376,0.1240849346,-0.1344082654,0.1354094595,0.3141591549,0.1311094314,0.0093049789,-0.2472665012,0.2174110413,0.2712011337,-0.5922092199,0.0147933215,-0.2766612172,0.3490387201,0.5501685143,0.007749232,0.4651241601,0.2857610881,0.0095577873,-0.3727356791,-0.2555267215,-0.1243509501,-0.1396030933,0.2677642703,0.1719058603,-0.1290390491,-0.2173302025,-0.0948465988,0.193602398,-0.1267535836,0.1561992764,0.168606922,0.1103106216,-0.0498778559,-0.3834073246,-0.2473418266,-0.0508276634,0.164636597,-0.0686244369,0.0010582254,-0.0450568162,0.141196996,-0.0122424131,-0.3772462904,-0.1268491894,-0.1285016835,0.4708686471,-0.0577265657,-0.0055256127,0.1414255947,0.2219111621,0.1650451571,-0.0575126037,-0.0924267545,-0.3087554276,-0.6229859591,0.1146941781,0.2116523981,0.0244752746,-0.207779929,0.5435648561,-0.4458746314,0.0581375659,-0.0936751813,-0.2992414832,0.1016631424,-0.4998219907,0.1622820199,0.169631198,0.1956583411,0.1346250921,-0.0950767845,0.1283337027,-0.0213813726,-0.0275122952,0.0839674547,-0.1403794289,-0.2595466077,0.0095590344,-0.4604827166,0.1345309913,-0.3566787839,0.2501229644,0.1101710349,0.3489561081,0.0631596446,0.2264976352,0.1081456318,-0.0103542302,0.3249147236,-0.2448162436,-0.1530455053,0.1325310022,-0.2032836229,-0.146553129,-0.0370611325,0.0255770329,0.1653888077,-0.0594214015,-0.1640657336,0.2898806632,-0.0571218207,-0.0292187482,0.0959936976,-0.1556544155,0.3203541934,0.3608259857,-0.323368907,-0.1277830154,-0.1393213421,0.3005742729,-0.1271823645,-0.0336162373,-0.0549268462,0.2390419692,0.018715173,0.5291284919,0.0341061987,-0.0153650474,0.134374246,-0.1689169109,0.2697131634,-0.2346494347,-0.0925755054,-0.2286020368,-0.0048998343,-0.11488875,0.1607830524,0.0025077409,-0.0715700537,0.0525253117,-0.0024309892,-0.3095356226,-0.2200315148,0.0150452228,-0.1498307586,0.2609986961,0.0487068072,-0.0817911997,-0.0618528947,-0.117672205,0.0393016934,0.2676590681,0.0295861959,-0.2685284019,-0.4405417442,-0.2893298566,-0.2391783446,0.0323151983,0.3208422959,-0.054685954,-0.3447121084,-0.3483787775,0.0268662218,-0.0373081788,0.276596278,-0.2384776026,0.0264820289,0.1826663315,0.1320007294,0.0022258058,0.2334503382,-0.0972593352,0.2654801607,0.2467637509,0.1480636001,-0.3726954162,0.0269201435,0.5315345526,-0.0132593047,-0.0906410813,0.1905854046,-0.1291369796,-0.1743909568,-0.3904857337,0.1949473768,0.0490189716,0.2345085889,0.0590002015,-0.0270240307,0.1536387354,0.2973504961,0.2407647371,0.052780699,0.1001152322,-0.0664742664,0.0427116081,-0.3014538586,-0.3016818762,0.0629281551,0.4964072406,-0.0984224975,-0.0937341303,-0.1838716269,-0.1373693049,0.256615907,0.2037686855,0.3517951071,0.4980363846,-0.0902909487,-0.0344541296,-0.0286775958,0.029989345,0.0347785056,0.1147361845,-0.3315271139,-0.4223519564,0.1673106402,-0.2731217444,-0.0397185013,0.1541344225,0.1347600222,0.0963369831,0.1906631738,0.0931879282,1.2108489275,-0.0564042144,0.4080526829,0.0988678932,0.0734630674,0.1593521833,-0.3770824969,0.1256880313,-0.3403652012,0.2211360037,-0.0227327365,0.197715342,-0.0934434608,0.1047364473,-0.1862933338,-0.0423749983,-0.0904861167,0.1629123837,0.2484714985,-0.0977061167,-0.1256242394,-0.2822242677,-0.2487204671,0.2500095963,0.1077614129,-0.2345747203,-0.0203658007,-0.1734341979,-0.105499208,0.0219828971,-0.5087255239,0.1759409159,0.0582401007,-0.1581332833,-0.1070632935,-0.2997943461,0.0974189118,0.3252301514,-0.1909096539,0.3439103067,-0.2919045091,-0.021847764,-0.3868194818,-0.009673507,0.048302982,0.22208637,0.4559492767,-0.051047273,-0.3100934923,0.1705653071,0.0016172162,-0.2022618502,-0.054920055,-0.1606494933,0.3719721138,-0.1491977572,0.1189011857,0.580999434,-0.2362263203,-0.4276350737,0.2584320903,0.106059365,-0.1041668355,0.0945983008,0.2284525186,-0.213984713,-0.0069225137,0.2884706855,0.278244555,0.3051139116,0.0521886423,0.2168047577,-0.3701768816,-0.4442611039,-0.0270674713,-0.0643298775,-0.3463463187,0.0714597404,-0.0367516167,-0.066123955,-0.1040002033,0.2321003377,0.1238160506,0.0221239049,-0.4511495233,0.120376572,-0.0649489984,0.165061444,0.0861953646,0.0656384975,0.0842761472,0.1837783009,-0.0319781154,0.2571857274,-0.4677874744,-0.1082119867,-0.1316693127,0.1399478614,0.5001212955,0.0727989301,0.1781305969,-0.0030994276,0.175899595,0.0198873151,-0.146920532,-0.4027986228,-0.1265345663,0.0586024262,0.0455125906,0.1174867526,-0.1898629218,0.0480011664,-0.0894880742,-0.2494620085,0.2931416035,-0.0424604788,-0.0598942935,0.1893616766,0.2281966507,0.3348003626,-0.0507198498,-0.2912103534,-0.0236544497,0.2507018745,-0.1550400704,-0.2885837555,0.0633031651,-0.2579932213,-0.1338054389,0.4720132351,0.4272881746,-0.2092519999,0.2347492874,0.0699178129,-0.0423683189,0.1527816951,0.2934681773,-0.1496689022,-0.1223615929,-0.1611938775,0.1480340958,0.4016269445,-0.2410201132,-0.1168166324,-0.3034080267,-0.3202054203,0.1450942457,0.2394766361,-0.1468034834,-0.0116455778,0.0976330265,0.0753566474,-0.0537546724,-0.2743573189,-0.2063909173,0.0860017762,0.0747517273,0.2472122461,0.3010364473,-0.1047464982,0.1949842423,0.2550906539,0.1903479248,0.5294098258,0.2952967882,-0.050148394,0.220783636,0.3157145679,0.1261717528,0.0454320982,0.2117093951,0.1033368632,0.0418910757,0.2532936633,0.0532404967,0.1031543091,-0.2978188396,0.1878958493,-0.1603634506,-0.1499089897,0.0720814839,-0.431035459,-0.0646451861,-0.1902017444,0.0895031244,0.2261160612,0.0458817035,-0.1359938085,0.2347641885,-0.375012517,0.4892819226,-0.0599416904,0.1551002413,-0.3245519102,0.0601202212,-0.0781089216,-0.1209972575,0.0667617843,0.4986949563,0.2809194028,0.1298241913,0.0574917085,-0.0689391494,-0.506968677,-0.1374689043,-0.1225559264,0.3048804104,0.1851514578,0.0021871082,0.1208130419,0.218743965,-0.2764625847,0.0446066633,0.3658650815,-0.0348321833,0.0672699288,0.0676590651,0.1294835806,-0.2585326135,-0.1138335541,0.1088423878,-0.5237907171,-0.1552445143,0.2836376727,-0.0167588275,-0.0255321171,-0.321714431,0.1563199311,0.351685673,0.2485607266,-0.1418109387,0.0507725552,-0.063922666,-0.1858907342,-0.1359014213,-0.0052690776,0.1410516053,-0.1193078682,-0.1394418031,0.1098374352,0.3044404387,0.2461598068,-0.0137109868,-0.3559066355,-0.464048028,0.1793326139,-0.0024049906,0.2249222249,0.112525709,0.0024729541,0.0066928794,-0.1813453734,-0.1733565032,-0.0810440034,0.131626755,-0.4532071948,0.0197478738,0.2463756353,0.0226951484,0.2262905985,-0.0995949358,0.2657625079,-0.2122879326,-0.1345541626,-0.0236710552,-0.1245663092,-0.1385743171,0.2478076667,0.0491604581,0.053052485,0.175332725,-0.1573432535,0.2364658117,-0.0046610814,-0.0658319667,0.5287590027,-0.0727890134,-0.1020260155,0.0013478688,0.0005244028,-0.180835247,0.1911252439,0.1074999869,0.1370232254,-0.2294063568,-0.2048784792,0.1577636302,0.1735467464,-0.3399541378,-0.208470732,0.1461176127,0.5066061616,-0.0968665481,-0.4703118503,-0.2162419111,0.1998507231,0.1342915595,-0.2421725839,0.474293083,-0.1838775873,-0.0305248108,-0.0745542645,0.1559154391,-0.2725869715,-0.1366281807,-0.2833069563,-0.5181700587]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/827","title":"[GEM] MultiWOZ dialogue dataset","comments":"Hi @yjernite can I help in adding this dataset? \r\n\r\nI am excited about this because this will be my first contribution to the datasets library as well as to hugginface.","body":"## Adding a Dataset\r\n- **Name:** MultiWOZ (Multi-Domain Wizard-of-Oz)\r\n- **Description:** 10k annotated human-human dialogues. Each dialogue consists of a goal, multiple user and system utterances as well as a belief state. Only system utterances are annotated with dialogue acts \u2013 there are no annotations from the user side.\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2007.12720.pdf\r\n- **Data:** https:\/\/github.com\/budzianowski\/multiwoz\r\n- **Motivation:** Will likely be part of the GEM shared task\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":30,"text":"[GEM] MultiWOZ dialogue dataset\n## Adding a Dataset\r\n- **Name:** MultiWOZ (Multi-Domain Wizard-of-Oz)\r\n- **Description:** 10k annotated human-human dialogues. Each dialogue consists of a goal, multiple user and system utterances as well as a belief state. Only system utterances are annotated with dialogue acts \u2013 there are no annotations from the user side.\r\n- **Paper:** https:\/\/arxiv.org\/pdf\/2007.12720.pdf\r\n- **Data:** https:\/\/github.com\/budzianowski\/multiwoz\r\n- **Motivation:** Will likely be part of the GEM shared task\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nHi @yjernite can I help in adding this dataset? \r\n\r\nI am excited about this because this will be my first contribution to the datasets library as well as to hugginface.","embeddings":[-0.0495593362,-0.1261757463,0.0484424531,0.5216318965,0.0280474238,0.2442046106,0.2257897556,-0.129513517,-0.0764305368,-0.0352300182,-0.3937834501,0.0283759013,-0.4157284796,0.3712361455,0.2753158212,-0.4878649712,-0.0401315726,-0.1847281158,-0.0573805906,-0.0662159994,0.0544462986,0.2652646005,-0.0837500691,-0.0141651276,-0.7082762122,-0.2576463521,-0.2459636778,0.5358034968,-0.0593674779,-0.2135654688,-0.1025887504,-0.0253923126,-0.1053599641,0.4787712991,-0.0001166607,-0.2765088081,0.2415370196,-0.0704951137,-0.158460483,-0.1901423484,-0.303958714,-0.1010894477,-0.1578002423,0.0485258624,-0.0672851503,0.1110400036,-0.2357476354,0.072080344,0.1161287352,0.2136203647,0.1436246634,0.1713851988,0.229274109,-0.1413156241,-0.0795614347,0.1754795611,-0.1360081434,0.1217542589,0.1660677642,0.1298537403,0.13394095,0.2862460911,-0.2868770659,-0.3387582302,0.0378700607,-0.1565734595,-0.1137382835,-0.2021169215,-0.1208337992,0.6377723813,0.782586813,-0.0460340902,-0.4587578773,-0.1823887378,0.1662186831,0.1890712678,-0.0151091991,0.2074808776,-0.1902395189,0.0880809948,-0.499795258,-0.3024016023,-0.1117864102,0.1520489454,0.1694430709,0.4894135296,-0.0048377253,0.1498208344,0.0406989902,-0.0129358061,-0.6592090726,0.0461831838,-0.1632288098,-0.0782562196,-0.5269982815,-0.2313212007,0.0054978612,0.2764656842,0.2792702913,0.2418639213,0.1933353096,0.0343860462,-0.4858166873,-0.0963608399,0.3565450907,-0.0357551947,0.2355093807,-0.4836921692,-0.0768123865,-0.3012522161,0.2371323407,0.1238581017,-0.2489851564,0.0568620972,-0.0454359427,-0.191259414,0.0778548494,0.0098750964,0.1015383452,-0.0046854638,-0.1881654263,-0.1904006451,0.3987607658,0.3084577918,-0.2176786363,0.0310238171,-0.2466755211,0.5173609853,0.214096576,-0.275044769,-0.1006629393,0.20847027,-0.0185947064,0.2577991784,0.0416524336,-0.1372463703,0.3089649081,0.0768450275,0.4037554562,0.0251382645,0.0357506275,0.0723679066,0.0309202764,0.0611653328,0.0265694745,0.2006107271,0.1279187799,0.1385942101,-0.0259206686,0.0555952191,-0.0541707613,-0.4240857959,-0.8219958544,0.0419629551,-0.3433516324,-0.4522158206,-0.0966314226,0.2801698446,0.0758952275,-0.1119188294,0.0898485929,0.1927165836,-0.2006132454,-0.2603945136,0.1181053296,0.4660356641,-0.0739355162,0.0199907217,-0.0051787714,0.0489323474,0.1514672339,0.2265951782,-0.2079218924,0.0020900264,0.0370475911,0.5209835768,0.0974760428,-0.1354452372,-0.0091419891,0.1261940747,-0.1579438746,0.455842495,0.2269834727,0.4218324125,0.0401256382,-0.0735240206,0.2379349768,0.2138475478,0.0857971832,0.0093416246,-0.1274498552,-0.2149721831,0.354528904,0.1450643539,-0.1484981477,-0.271540463,0.0218770802,-0.0533044823,0.3994185627,-0.0728087649,0.214067772,0.1887446791,-0.0459010825,0.3278737664,0.0064844755,-0.3588349223,-0.0401546247,0.0463247076,0.4060170949,0.3642266393,0.2133593559,-0.0913456082,-0.0593798496,0.0209012106,-0.1971767396,-0.3014435172,-0.0209976994,0.3040892184,0.0192148145,-0.2110359818,-0.3215515018,0.6143251657,0.2347304672,-0.0238543488,-0.3885957897,0.1535160542,0.0642829984,-0.0954658687,-0.0493885837,0.4657987952,-0.1018567234,-0.1015431583,0.0256350245,0.3633405268,-0.3810467422,-0.0191664658,0.2527374923,0.211559996,0.1455547661,-0.0376944318,0.3808533847,-0.3574172854,0.0935362279,-0.1666322201,-0.0316772014,0.0237861816,0.0688074902,-0.1138606817,-0.3425866663,0.3334616721,0.1323535591,-0.0600715019,-0.2378442287,-0.0501989573,0.2843133211,-0.3968089819,-0.2087936252,0.0861881599,-0.5717890859,0.2553270161,0.2885805368,-0.2681139708,0.3120868206,-0.0075467913,-0.2782268226,-0.2280256003,-0.0677193478,-0.2392433584,0.1310912669,0.3073764741,0.0268663969,-0.2262406647,0.0077417539,0.0461377352,0.1179578006,0.1477182209,0.1421055347,0.2155932486,0.1317403018,0.1953988522,-0.0600356273,-0.1858831495,0.0917344838,0.0612705499,-0.2202705741,-0.3777098358,-0.2315757275,0.0206582509,-0.0031795367,-0.8598554134,-0.3776175082,0.0210866872,0.1854080409,-0.208636567,0.1864878833,0.140997529,0.1710365862,0.090553388,-0.1613760293,-0.0967269167,-0.2473046035,-0.474103272,0.0599152967,0.070634678,0.1254862398,-0.0712724626,0.3366229832,0.1479424685,0.158692345,-0.1248507723,-0.3355082273,-0.0475235246,-0.3214035332,0.3632976115,0.1861050874,0.0045008925,0.1736129075,-0.5391616821,0.0940193236,0.0636401251,-0.1032526642,0.0796101019,-0.0586554632,-0.3287820816,0.0343878902,-0.1330083609,-0.2042253911,-0.446413964,0.2932297885,-0.0822253153,0.068107672,0.1388568729,-0.2183091044,0.3125792742,-0.228241384,0.2930274308,-0.2374432385,-0.4391099513,0.386726439,-0.1543562263,-0.1579540074,-0.0408226177,-0.2597220838,0.27492854,-0.2068925351,-0.1263917089,-0.2367945462,-0.1196150109,-0.1223814338,0.1548680514,0.0074252626,0.3379835784,0.1504380554,-0.1130660027,-0.0759113729,-0.4116906524,0.10021016,0.4797812402,0.3220478594,-0.0224163495,0.0151808262,0.027829811,0.4263081849,0.259511143,0.1940469444,0.1456073821,-0.0416595452,-0.050338082,-0.1276602894,-0.6021878123,0.3620354831,-0.1072281078,0.1803506464,0.3086300194,0.3506394327,0.029751325,0.0051747519,-0.3613120317,-0.4042972922,-0.248694703,0.0308375768,-0.1375452131,-0.165692091,-0.3035289347,-0.3996590376,-0.2994038165,-0.2451196909,0.2551681995,0.4556032717,-0.187185511,0.0674218237,-0.0554055274,0.2080793828,-0.1167583019,0.2883276641,0.2834438682,-0.24261491,-0.3276434839,-0.00680744,0.1010645777,0.270768106,0.8676868677,-0.3521382213,-0.4406842887,-0.0189918745,-0.0773843154,-0.1104205102,0.2062216699,0.067893438,0.215457812,0.22410281,0.4616443813,-0.0805999115,-0.0421269462,0.4925209284,0.2451322526,0.0402824543,-0.0942880064,0.1154118404,-0.2007338852,-0.4446724951,0.048264008,-0.1206622869,0.2248320729,0.4122959077,-0.2588495016,0.1395419389,0.2476587892,0.4659259021,-0.2856921256,-0.0763509348,0.4765966535,0.1237192973,0.1571492851,-0.0069006397,-0.0467528999,0.2670635283,0.3491562605,-0.0073567955,-0.2886116207,-0.3671667278,0.6399357319,0.2458284497,0.2599239051,-0.0043925,-0.0496905781,0.2666392922,-0.070273377,0.0905362293,-0.0380105972,0.2193392217,-0.3569739759,-0.2911620438,0.2542299926,-0.0566715598,-0.1089179143,0.348128587,0.1393656284,0.091753453,0.0575123131,-0.0708093271,1.2208231688,-0.12197458,0.2385089397,0.1832055449,0.118587628,0.3062672317,-0.3133832514,0.0434007347,-0.0887928531,0.4918666482,-0.1058600843,0.0500412881,0.0773831159,0.0689166784,-0.2723092437,-0.1200606078,-0.1255691499,-0.3180556893,-0.1728464365,0.1763084233,0.1679090708,-0.43968606,0.1948031932,0.0914048329,0.0268137604,0.099116832,0.1932019293,-0.4706449509,-0.2478196323,-0.0061870557,-0.5756494999,-0.1030137986,-0.1777904034,-0.0751584396,0.1510763615,-0.4125762582,0.2650757134,0.6566467881,-0.0458359234,-0.0364848673,-0.1906546801,-0.0270595346,-0.3097652495,-0.1800451279,-0.0290715061,0.0146051338,0.5625340939,-0.0856467634,-0.4616304934,0.1913685501,-0.0405458026,-0.1574316174,-0.2143808752,0.0939354524,0.3513994813,-0.123413913,0.0840732902,0.4758815467,0.1632349193,-0.1882021129,0.0587330833,0.3752679229,-0.2726207078,0.1003443077,-0.0824425593,-0.200220421,-0.192137301,0.1064523384,-0.0451688841,0.0649108142,0.3358596563,0.2078392953,0.0046557873,-0.1652280688,-0.1171681955,0.0126073845,-0.0297698639,-0.2118089944,0.5123296976,-0.186995104,-0.0449880771,0.0160912853,0.1603281051,0.1677426398,-0.0814915895,-0.3418284357,-0.1718787402,0.1577097476,0.0150298206,0.0357606709,-0.2085574865,0.1139455885,0.2710703313,0.4699586928,-0.2757550776,0.0167700425,-0.2806571722,-0.0290530547,0.5896809697,-0.2300732583,0.1488163322,-0.0789915174,0.0771974549,-0.4200710654,-0.1101441458,-0.0885407478,-0.3610190153,0.1684495807,0.0748721361,-0.1540771425,0.2706817687,0.0158064738,-0.0894804671,-0.225072071,0.332005769,0.2376986593,-0.0703223348,0.0462866165,-0.2484468371,-0.0196932722,-0.3048243523,0.1348234564,0.1987129748,0.0725642443,-0.03848188,-0.0703204945,0.0490027592,-0.294041872,0.2146622986,0.2510720491,0.141667679,-0.05680307,-0.0330819935,0.1440103054,-0.0253857393,0.3045450151,0.3379099369,-0.296710521,0.1120469719,0.2842157483,0.0707964972,0.043912109,-0.2023289651,-0.1432787329,0.1610299647,-0.2441746145,0.0529661626,0.1319922656,-0.0511856191,0.1343575865,0.0085505489,-0.3371321857,0.0945689976,-0.1736166924,-0.0558875166,0.0412960574,-0.4468115866,-0.1386161,0.3172749281,0.2480266988,0.2894787788,0.1697863787,0.0131895328,0.235025093,0.0140649918,0.0810088068,0.022241842,-0.1308628917,0.272714287,-0.0621148013,0.1389015317,-0.0964663923,-0.0127952686,0.8489482403,-0.0531655289,0.0939543843,-0.1823534667,0.409339726,-0.0385651477,0.0198610183,0.1863223463,-0.1902027428,-0.2819322646,-0.1889741272,0.0304585584,-0.1939524859,0.5753381252,0.1022952721,0.1599065512,-0.5247256756,-0.1194863692,0.0622979365,-0.2186884731,-0.3328735232,0.127347663,0.4176219106,-0.2335387915,0.2680813372,0.2843910158,0.5901280642,0.058166407,0.059219256,0.324536562,-0.2579360008,-0.0714358985,0.0402381495,0.3600096405,0.194201082,0.1206989214,0.1378458589,-0.0050475285,-0.1075205654,0.2780385017,-0.0980773121,-0.0631899834,-0.3260617554,0.2701668739,-0.1711437106,-0.3203284144,-0.2823599577,0.009438308,-0.455919683,-0.404370755,0.3326562643,0.0235002283,-0.2665002644,-0.2315282822,0.0525779091,0.041016899,0.4059999883,0.3472782373,0.179275319,0.0704644918,0.2959874868,-0.6454319954,-0.1585429013,0.3647211492,-0.0800439194,-0.0159734283,-0.1301050782,0.5115667582,0.1902517825,0.081014201,-0.3565253317,-0.0754356012,0.0700324103,-0.1761051267,0.4418860078,0.4003624618,0.1796965003,0.2914295495,0.232322529,0.0274969544,-0.0136616854,-0.0643403307,-0.0957337767,-0.1059933752,0.0085339611,-0.3982722759,0.2332936674,-0.1317209899,0.6390528679,-0.1394646317,0.1082861647,-0.0935809985,-0.0011770361,-0.1551577598,0.1179217696,0.0444159955,0.6158026457,-0.1510683298,0.0670225099,0.0121991504,-0.1279336065,-0.2693753242,0.337331295,0.0121930009,0.1458608061,-0.0913503766,-0.1188278422,0.309186697,0.2624584734,-0.0521305911,0.1051345542,0.0185578167,-0.079529658,0.5583524704,-0.0703388602,-0.2576852143,-0.4120702446,0.1972282827,0.0915482193,-0.0519298129,-0.5948369503,0.0284350384,0.180210039,0.3053228557,-0.2246882468,0.4031971097,0.1277081072,-0.1863012165,-0.0024883871,0.0475102961,-0.0397526547,0.0446500517,0.2181625217,-0.0562602505]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/824","title":"Discussion using datasets in offline mode","comments":"No comments ?","body":"`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n","comment_length":3,"text":"Discussion using datasets in offline mode\n`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n\nNo comments ?","embeddings":[-0.4933066368,0.2524318397,-0.0220922027,0.1568014473,0.2215198725,-0.1686570346,0.5425215364,0.042939011,0.2951898873,0.2246187627,0.0599192642,-0.0534892827,0.0234732274,0.3996051252,-0.095134303,-0.0656923205,-0.1387800127,0.0495659448,-0.1556704938,0.135307461,-0.1299726665,-0.1212218478,-0.4417335093,-0.0718617961,-0.1173430979,-0.2039083093,-0.0650552437,0.1827041805,-0.192239657,-0.3585914969,0.2215988934,0.1480474919,0.1775918603,0.4751347601,-0.0001058719,0.0230443776,0.0620841645,0.051758714,-0.2133131474,-0.401139766,-0.282608062,-0.4664886296,0.1922245324,-0.3922984898,-0.1885079145,-0.0210282691,-0.0783914998,-0.4714707732,0.3156807423,0.3838435709,0.24149625,0.6029465199,-0.1457916498,-0.3019393384,-0.1656311154,-0.1101765037,-0.163140595,0.4276005626,0.0751697347,-0.0206379071,-0.0712380409,0.1007647589,-0.1281983405,-0.0219588596,0.3216148317,0.0209862124,-0.1397964358,-0.1123327389,0.218405962,0.4208197594,0.4184082747,-0.2919480801,-0.2635616958,-0.3355534971,-0.0108078923,-0.4421505332,0.0973920599,0.155895263,0.0163611863,0.2374585271,-0.4316322803,-0.3646235764,-0.3455609083,0.1233967543,-0.3376245499,-0.02359985,-0.270880729,0.0455762148,-0.0820924118,0.1323668957,-0.1986339688,-0.2720161378,0.036093045,-0.1327680647,-0.2699775696,-0.0022249832,0.0939113572,0.1123581529,0.1933836043,0.078358613,0.2112049162,-0.0305778347,0.0887852162,0.0302313995,0.4626075625,0.2498055249,-0.0198664423,-0.1941179186,0.572101891,0.0713566765,-0.0052483832,-0.3090769053,0.3155036271,-0.3124776781,0.123578757,0.0141127966,0.3457062244,-0.3079696596,-0.130490154,0.0860361829,0.1276158094,-0.047168199,0.195558548,0.2239585519,-0.2267397791,0.0101274969,0.0207190756,0.4521712661,-0.2595635951,-0.2919174731,-0.0560963489,-0.168199271,-0.1676646471,0.1140057445,0.1903556883,-0.0070676729,0.0758215412,0.0265823826,0.2115077972,-0.0630335435,0.3294494748,-0.1293949634,0.0791260749,0.2282637656,0.1955180764,-0.0809123144,0.0424205922,-0.1981299073,0.0976364762,0.0557508431,-0.1009421647,-0.3952343762,0.1471506655,0.2337478399,-0.6416087747,-0.1394525468,-0.2261295021,0.1055918559,-0.3148970008,-0.1563836187,0.0244722851,-0.0357205421,-0.1736389846,-0.2243375033,0.2679877579,0.5881357193,-0.4116489291,0.1247192547,0.1473379582,-0.2155241221,-0.1905390471,0.0440312698,-0.339963764,0.5210672617,-0.0051524742,-0.2699905634,0.52447474,-0.334168911,-0.2518340647,0.2847811282,-0.1962419599,-0.1360980719,-0.0448308177,0.0704183355,0.462128222,0.0958058387,-0.0481142029,0.2037850767,-0.0698206052,-0.0061580581,-0.2521780729,-0.2548526227,0.1646097451,0.2539943159,0.3010041118,0.1337380856,0.272659421,-0.2314561605,0.0115257492,-0.2322752029,0.1102632582,0.1487744749,0.1880429238,0.0702731907,-0.0771012977,-0.1872040033,-0.462149322,0.250354439,0.0657768026,0.1604584008,-0.1991915554,-0.1493580937,0.0013562463,-0.2152406722,-0.1537901014,-0.0057076751,0.1181790382,-0.186362654,0.1349956244,0.1802526712,-0.3134862185,0.2052322626,-0.0615581349,0.0189323127,-0.0849768668,0.3531087935,-0.0569319203,0.031540677,-0.0390327834,0.0920801163,0.041979,-0.154771179,-0.0754115433,0.4924159646,-0.0046384176,0.0191542488,0.3357336521,0.402839303,0.349696219,0.0179140698,0.548206687,-0.0897624195,0.3306947649,-0.1125491038,-0.2958375216,0.727725625,-0.1408099979,0.2367670834,0.1103549898,-0.0027518526,0.2078945935,-0.0183882937,-0.4496650696,-0.0211210232,0.1286108047,-0.0060404879,0.2088193744,-0.0317348242,-0.378756851,-0.0398309082,0.3447001874,-0.0118219638,0.1677264124,-0.0646092594,-0.0297307651,0.1202487051,-0.0349112824,0.0608304404,0.5366030931,0.2184972167,0.0458083898,0.2538666427,-0.0439100191,-0.2942561805,0.1189216077,0.2135707289,0.0259694196,0.1078604609,0.0689221621,-0.0163394157,-0.2666737139,-0.0513092764,-0.0296778791,0.0325317346,-0.316646874,0.2817128301,-0.258345753,-0.0302088745,0.0437725522,0.1776666194,-0.4932307005,-0.2147786468,0.0775242373,0.0815686062,-0.1933612525,0.1323656142,-0.146904856,0.4614443481,-0.2636246681,-0.154857561,0.2218280882,-0.5266567469,-0.0910279453,0.0826990083,0.3359416425,0.1201439276,0.49711743,0.0185741521,-0.1277314126,-0.3264021575,-0.1801318079,0.0520243943,-0.0520270839,0.3036504686,0.0738120824,0.0636909008,-0.0442282483,0.3457593322,0.3820469677,-0.2871483862,0.1930681616,0.1283769608,-0.0787296593,-0.1153176129,-0.327337712,-0.4236873984,-0.3910315335,-0.3972489536,0.205739677,-0.0504819341,-0.097445339,0.1532876492,0.1434328407,0.1030545086,-0.0874698535,0.1496271193,-0.1879969835,-0.4750307798,0.1180900782,-0.1620282233,-0.58878088,0.2941455245,0.2351694256,0.150486812,0.3871750534,-0.4629498422,-0.3662821352,0.1006098166,0.4340940416,0.0750323236,0.32946679,0.3646383882,-0.0668853596,-0.0202237293,-0.074608624,0.0579903387,0.1952864826,0.0252261665,-0.0951528549,0.0633210242,0.2719270587,-0.0238344688,0.4337469637,0.0237140656,0.0674959794,0.236680761,-0.1461587548,0.7218088508,-0.1768418252,-0.1463215649,0.0303369127,-0.0431501903,-0.3235523105,0.1734118164,0.0018309732,-0.003422868,-0.6454547644,-0.0766717196,-0.3651378453,-0.3913440704,0.1411714852,0.1404760778,0.3954807818,0.122136578,0.2962167859,-0.2375854999,-0.025561627,-0.1006201431,0.276650399,0.1206144542,-0.1047402918,-0.2381327599,0.0176162552,-0.3912906945,0.3382920027,-0.1352334768,0.2526281178,-0.1117689013,0.0197446663,0.0154671809,-0.1137966588,0.4152953029,-0.4107178152,-0.1126563996,0.2961502373,-0.062455304,-0.2572464347,0.0975401029,-0.2104826868,0.079355143,0.1779461354,0.1908947378,-0.0766474754,-0.3310196996,0.0305376202,0.4057390988,-0.2477792501,-0.1999259293,-0.204371646,-0.0918231606,-0.2920825481,0.0659123212,-0.2056976259,-0.0279439315,-0.0952322185,0.1331879646,-0.2401023954,0.1881027073,-0.0662739873,0.1122351885,0.1192715466,-0.2772941887,0.2798005641,0.2437320054,0.0584180392,0.3556734324,0.3104756773,-0.0845445022,-0.2784185708,0.1972304285,-0.0529773422,0.4277353585,0.4833644927,-0.1102514341,-0.0911981314,0.1663771421,0.2840402722,-0.4301401079,-0.1851825714,0.2204681337,0.1762772352,-0.1650571525,-0.356742084,0.1954751462,0.1313639432,-0.0314057097,-0.0359575339,0.1533774734,-0.1841651052,0.1673785299,-0.1274889112,0.7628161311,-0.0491121486,0.19868128,0.2266466618,-0.0023502298,0.4669119418,-0.2524152398,-0.0245646313,-0.0483880043,-0.2462437749,-0.0198394507,-0.1269959509,0.4598327875,-0.0259256009,-0.084996514,0.576674819,-0.0576908551,0.3026741445,-0.3154513538,0.2089506984,-0.049419336,-0.1615305692,-0.5147092342,0.1982733607,0.1120822728,0.0854094401,-0.2225240022,-0.2555074096,0.085963212,0.064757809,-0.0303362031,-0.1312019229,-0.0322327726,0.0290215313,-0.1984352022,-0.023475064,0.083584398,0.1054834872,0.4154070914,0.181367293,-0.3255632818,0.0459973998,-0.4018482864,0.1409970522,0.0468029939,0.2634520829,0.144577235,-0.2265856862,-0.2533637583,0.0494967401,-0.0797083601,0.0248373635,0.4239843488,-0.0687179714,-0.045586843,-0.2636452615,0.1718321741,0.2122199088,0.2264752835,-0.120852679,0.1541417688,0.2662613988,0.0979206711,0.3438726366,-0.1222857609,-0.1791863739,-0.2606613934,0.330686748,-0.1380280256,-0.2132194489,-0.0854203254,0.1875199676,-0.2072871923,-0.3044223487,0.2852861285,0.0458308421,-0.1402495056,0.2280561477,0.3780412078,0.3516243994,-0.1552801877,0.2376644611,0.0499191694,-0.2627231479,-0.0556258261,-0.341675967,-0.2863191366,0.2138006389,0.4600594044,0.1125384495,0.0868829042,-0.1011995748,0.0535498112,-0.2750267982,-0.3924860358,0.0645789206,-0.0849328935,0.0482678637,0.2691722512,-0.1769080758,0.378655225,-0.2806554735,0.0780818611,-0.1999795288,-0.3577020168,-0.2534793913,0.0228292476,0.1752212197,0.0018367281,-0.2831081152,-0.0082761869,-0.6014600992,-0.3430717289,-0.0998379812,0.2361384183,0.4565203488,-0.1453760862,-0.0288690571,-0.2372286767,0.4448953569,-0.1520394534,0.1317467541,0.0674255267,0.2695014477,0.0059559578,0.2536630929,-0.025859138,0.07535889,-0.03214553,0.1277179718,0.136867255,0.1186827198,0.4098312557,-0.2472625375,0.1569513232,0.2645841837,0.5581196547,0.2460851371,-0.0855736658,0.2051826119,0.1521235853,0.2568421066,-0.3573365808,0.0528199226,0.0268688444,0.1425834894,0.3763659,0.106513381,0.4058741927,0.0929308608,-0.0808149576,-0.0025142883,0.2996039093,-0.0783417448,0.2358202934,0.0867297351,0.0966128185,0.1267880052,0.051988747,0.4973748326,-0.0507934652,0.306931138,-0.3401525915,0.2431389093,0.0895504951,0.1299022734,-0.2388595492,-0.4137021601,0.2915698886,0.0172721092,-0.0630377457,-0.2593877017,-0.2947163284,0.2426262647,-0.4775572121,-0.1492782533,-0.161766693,0.0547038876,-0.1185432747,0.0804379582,-0.4077640176,-0.3162714541,-0.0106916474,0.1170381233,0.0549074486,-0.0228854604,0.0233087223,0.0075506298,0.1047930568,-0.2285488248,0.1750072837,0.0475088507,0.0957758576,-0.0206110273,0.0499546006,0.123297438,0.0859024823,0.1452385634,0.4987824261,0.3502499759,0.0387498289,-0.1653532833,0.1410956234,-0.0266319606,-0.0964997038,-0.064035587,0.1293409914,0.0435687527,0.1336808801,0.3595692813,0.1503888071,-0.1812263578,0.4055191278,-0.1362619102,-0.154629156,-0.4655710459,0.2220834494,0.1459266096,-0.1377104372,-0.0977088287,-0.3576329947,-0.1625605822,-0.1034983099,0.3943767548,0.0248626154,0.1261225641,-0.2269939482,0.1148464754,0.0879190639,0.3894189,0.1644449532,0.3074153066,-0.3373091817,0.1070023477,-0.7802565694,-0.1503261477,-0.1485504359,0.3222281337,-0.0157778207,0.0801694393,0.0437735692,0.4615562856,0.0706163049,0.1592685133,0.1419586092,-0.186975956,-0.3136603832,-0.1003363207,-0.0893155113,-0.1453506947,-0.0491982214,-0.2281465828,0.116271764,-0.3354311585,-0.0193079896,0.03706811,0.2027772069,-0.2085835189,-0.0927863643,0.4465796351,0.0557947643,0.4139313102,0.2298982143,-0.0895214379,-0.0551356003,-0.0034939789,-0.0318071209,0.1843704432,-0.0343292765,-0.064143829,-0.1329948455,-0.1524425745,-0.1678947806,0.191934526,0.0529338866,0.0359339789,-0.1394106746,-0.0219196863,0.1097123474,0.1400276124,0.0205062032,0.0275904331,-0.1313932687,0.1999552995,-0.1885472238,0.0577039011,0.3298625648,-0.0150046349,0.0562844761,-0.3699326217,0.3722217679,0.2172840387,-0.0352433808,-0.5318301916,0.1642120481,0.2944125533,-0.2089621574,-0.1515275985,0.2880146503,-0.2621638179,-0.22131975,0.0086071603,0.2698963284,-0.293240726,-0.2728482783,0.1057705581,-0.1122144535]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/824","title":"Discussion using datasets in offline mode","comments":"I think it would be very cool. I'm currently working on a cluster from Compute Canada, and I have internet access only when I'm not in the nodes where I run the scripts. So I was expecting to be able to use the wmt14 dataset until I realized I needed internet connection even if I downloaded the data already. I'm going to try option 2 you mention for now though! Thanks ;)","body":"`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n","comment_length":72,"text":"Discussion using datasets in offline mode\n`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n\nI think it would be very cool. I'm currently working on a cluster from Compute Canada, and I have internet access only when I'm not in the nodes where I run the scripts. So I was expecting to be able to use the wmt14 dataset until I realized I needed internet connection even if I downloaded the data already. I'm going to try option 2 you mention for now though! Thanks ;)","embeddings":[-0.6300708055,0.2058906406,-0.0102498597,0.1388324946,0.1934469193,-0.1737716347,0.5314203501,0.0831640139,0.3500735462,0.2367119044,0.1306480765,-0.0646166503,-0.114948757,0.4892579913,-0.0613240823,-0.0358768292,-0.166447714,-0.0370642841,-0.2075223029,0.1665765494,-0.1922239512,-0.163463071,-0.526426971,-0.1563569605,-0.2378030717,-0.2821217477,-0.0507955141,0.143443346,-0.067549102,-0.3244419992,0.2889086008,0.1382351369,0.2235022783,0.444934994,-0.0001075604,0.04597155,0.0817122832,0.0424997136,-0.2193595916,-0.3220709562,-0.2740172744,-0.5323532224,0.1792140752,-0.3614763319,-0.2261399925,-0.106020093,-0.00228727,-0.484377414,0.3558473587,0.2895818949,0.1922075748,0.6198381782,-0.2299235016,-0.2431981862,-0.1677201837,-0.165620774,-0.1867387891,0.3589761257,0.164250344,-0.038412191,-0.0706448629,0.0921199098,-0.1044342965,0.0972447917,0.4349209964,0.0993571058,-0.1420596242,-0.0717146248,0.1741680056,0.5084940195,0.5811343789,-0.2377759814,-0.2352240235,-0.2079222351,-0.1121013388,-0.478767693,0.088491641,0.1763013899,0.0440583453,0.3034045696,-0.4777004421,-0.2545405626,-0.326336205,0.1942960173,-0.4016208053,0.2013929486,-0.1917235851,0.1068929732,-0.1827457398,0.1533904225,-0.0075839888,-0.2337021381,0.1550296992,-0.1302921772,-0.316718787,-0.0140245194,0.0825168639,0.2733768523,0.0394199155,0.1726535559,0.1644296497,0.0296572708,0.0703102201,0.0388146229,0.3789533377,0.2198379338,-0.1209222078,-0.3116755486,0.6292068362,0.0068749459,0.0290250573,-0.2422123849,0.3572416008,-0.3322311044,0.1250274777,-0.0004763916,0.3401116431,-0.0963580981,-0.1281303465,0.0624006987,0.183156684,-0.1575584859,0.0885406286,0.216064781,-0.2617760301,0.0305719264,0.142048806,0.4152132869,-0.2527972758,-0.3033896089,-0.0704956651,-0.1785693914,-0.0896871686,0.1340269446,0.3163670897,0.2598673105,0.0254875217,-0.0343847089,0.1714944392,-0.0133727537,0.424575448,-0.0901880264,0.1361578554,0.3018315732,0.1157079861,-0.0600553267,0.0591588579,-0.1181263402,0.0598678291,0.0127167394,-0.1156592965,-0.3404823542,0.2708459795,0.2129337788,-0.5129674673,-0.1099461019,-0.3253318369,0.0677997172,-0.3322136402,-0.1121459752,-0.0031902401,-0.158851862,-0.3020486534,-0.2183876187,0.2850363255,0.6506155133,-0.5364612937,0.1531349868,-0.0336903036,-0.1371347159,-0.1496427357,0.0943927839,-0.3362303972,0.6654826403,-0.0118889064,-0.3396773338,0.4533733726,-0.4085681736,-0.2588607073,0.2776843905,-0.328661263,-0.0729339048,-0.065404959,0.20602718,0.4211373627,0.065777339,0.0929549411,0.1874692887,-0.0519748516,-0.0244046748,-0.1945666522,-0.2632915974,0.2911950648,0.2045401782,0.3346414864,0.0669414848,0.3081213832,-0.1736802757,0.043517448,-0.1846490353,0.2528410852,0.0154083641,0.1728463769,-0.0217746403,-0.1127807647,-0.1736566424,-0.6063697934,0.3064254224,0.0231700111,0.0968806744,-0.1869288385,-0.1187956631,0.024096081,-0.1919628531,-0.0954886898,0.0504775718,0.1024771109,-0.2254904956,0.0808495507,0.1283679903,-0.2564831972,0.2029672563,-0.1189566478,-0.0099110864,0.0438441113,0.2811492383,-0.0489588417,0.029021278,0.0381036885,0.0005989124,0.0214393232,-0.2149713635,-0.01718162,0.4893764555,0.0006060577,0.1398847103,0.2435282469,0.4766898453,0.3864599466,-0.1545752287,0.5698558688,-0.0209588613,0.2697749436,-0.2345478833,-0.2590694129,0.5913993716,-0.1949974298,0.2110204697,0.0851286948,-0.003755586,0.1491420269,-0.0165282283,-0.4646927416,-0.071834065,-0.0075024227,0.0016719986,0.1255860031,-0.1001415178,-0.3027392924,0.0762121156,0.3309549093,0.0889344886,0.2562539876,-0.1848618984,-0.0806870162,0.0944451541,-0.0143336561,0.0207728483,0.5388502479,0.1315514594,0.048894424,0.2817274034,-0.0158035774,-0.3472720683,0.124608539,0.219894737,-0.0228709579,0.090892829,0.0311373491,-0.0293929577,-0.1287688315,0.0264611691,0.0120821949,-0.0286060069,-0.255479902,0.2454203814,-0.3986744583,0.0395875461,-0.0565145239,0.0932429656,-0.3712323606,-0.2391112149,0.0468526557,0.1336499006,-0.2514076233,0.1489571333,-0.1666212976,0.520655632,-0.2418760508,-0.1473505348,0.1968584806,-0.5571706295,-0.114932023,0.0846704319,0.3601271212,0.2178107351,0.464566052,-0.0489885807,-0.0849459469,-0.3521950245,-0.1436619163,0.0259494949,-0.0372475497,0.3027525842,0.0437212214,0.2129589617,-0.030940542,0.216339618,0.3268577158,-0.3476834297,0.2610098124,0.0145970285,0.0120825963,-0.1428596228,-0.3249204159,-0.5815706253,-0.5232006311,-0.3354818523,0.1599170119,0.0059763142,-0.0602069832,0.04779258,0.0959361866,0.0486622006,0.0634140968,0.0541641712,-0.1572210044,-0.4133062065,0.1801300049,-0.1736979783,-0.5983689427,0.4090825915,0.1445647925,0.1866304278,0.4446810782,-0.4329784811,-0.1786331981,0.0999474004,0.4149424136,0.2194602638,0.3698682487,0.4159147143,-0.0263587236,-0.0044901734,-0.1017857566,0.1604266018,0.2131545246,-0.0501594208,-0.0727311,0.1946499497,0.1170222908,-0.0909304172,0.5493084192,-0.0598432086,0.0626043305,0.2284747958,-0.1173968017,0.6396966577,-0.1641614437,-0.0831010863,0.1386887729,-0.0655948892,-0.2300423384,0.2599748075,-0.0657806396,0.0406438261,-0.6198848486,0.0020187062,-0.3696756661,-0.4392199218,0.0987968296,0.2668218017,0.4354861081,0.2188317627,0.23445189,-0.2026198655,-0.1401429921,-0.0432291105,0.2950327694,0.1143773422,-0.00951178,-0.16620107,0.0718225315,-0.3588623405,0.3179502189,-0.1387809366,0.3281266987,-0.0474771038,-0.0183208976,-0.0609111898,-0.0483756848,0.3537316918,-0.429656148,0.0080434922,0.2310298234,-0.089441821,-0.2590506077,0.0481312275,-0.3023554981,-0.1070722044,0.1232942194,0.2312655598,-0.009061466,-0.2890507281,-0.093575187,0.4394928813,-0.3306919336,-0.1491357833,-0.1300390214,-0.0264644325,-0.2965613008,0.1153249741,-0.1198114231,-0.0172980968,-0.2313773483,0.0964094028,-0.1704923958,0.1686469615,-0.0581250899,0.0691087246,0.0691001639,-0.2343954891,0.2624781132,0.2558904886,0.1509255618,0.5121373534,0.2453927696,-0.0556901693,-0.0811072364,0.1570569575,-0.0704071522,0.4112352431,0.4923603237,-0.112567462,-0.1765354872,0.2434657961,0.2588063776,-0.5872603655,-0.0939202011,0.0892567933,0.1319001317,-0.1381255388,-0.2807336152,0.214585945,0.1436351985,-0.0498341918,-0.0280276369,0.275250107,-0.2484537214,0.0456356592,-0.0597792417,0.7471334338,-0.0503546819,0.2299505472,0.1710410863,0.1345073879,0.3820542395,-0.3570559919,0.0242576711,0.0083441082,-0.2609067559,-0.1121664047,-0.1517976075,0.3989054859,0.119651705,-0.1589447558,0.5752491355,-0.0488136113,0.3787566125,-0.3056920171,0.2353721708,-0.0690017715,-0.150935486,-0.280444473,0.1945443451,0.0841047987,0.1325060427,-0.2965494692,-0.1999388933,0.0805014819,0.1032053158,0.0199742783,-0.1605793983,-0.0908032581,0.0750892311,-0.2080814689,0.0110155353,0.1845130622,0.0723550171,0.448882997,0.2354969084,-0.4055701792,0.0239274185,-0.4745929241,0.1689142734,-0.0147644449,0.2257876843,0.2119621038,-0.2073981017,-0.2758950293,0.0584933423,-0.0901271701,0.052082289,0.4216811657,-0.0912183225,-0.1135954708,-0.2078427672,0.1400360465,0.2337010652,0.1270939559,-0.118896544,0.1302799582,0.1750443578,0.1143170595,0.4348134696,-0.0489926748,-0.115358457,-0.2252298594,0.3077353239,-0.1198249981,-0.1528263986,-0.0878993347,0.2735413611,-0.2031891644,-0.3384055197,0.196448788,0.0465169474,0.0449065752,0.2258599848,0.4077033103,0.4182564318,-0.1193952337,0.4594063163,0.0348663665,-0.1733819842,-0.1316818744,-0.3336318135,-0.3115442395,0.2858285308,0.4760401547,0.1319065392,0.0429589078,0.0135140764,0.080083482,-0.1690980494,-0.3756701052,-0.0080091581,-0.0301890466,0.0805658475,0.1734397709,-0.1646102965,0.318439573,-0.1392024606,-0.0031448016,-0.1964249462,-0.407360971,-0.2196913213,-0.0056017246,0.2171227038,0.0729132071,-0.4410377145,0.0773469657,-0.5827118158,-0.2432707697,-0.1612302959,0.2887277007,0.4088554382,-0.198624447,0.0132579235,-0.0323876664,0.3897836804,-0.198643297,0.1378962696,0.0811308101,0.2517086565,-0.1156207025,0.2089472413,0.1144660488,0.0455652922,0.0876552239,0.0893079638,0.1664662957,0.1618378609,0.4194568694,-0.2242739201,0.0255389996,0.151209563,0.5265103579,0.2531048656,-0.0403761007,0.1189205572,0.1235127822,0.1960496753,-0.3990579545,0.0341197513,0.1256488562,0.1630678326,0.3431506753,0.0698585436,0.4620226324,-0.0408508517,-0.2096596211,0.0178091284,0.2600048184,-0.2919554412,0.2190141231,0.0231899973,0.0135690477,0.1464996487,0.0203751456,0.3932596147,-0.0602547638,0.2468483895,-0.3339198828,0.2350604832,0.0747700185,0.1230946183,-0.1912600845,-0.4348840415,0.2412766069,-0.0731848106,-0.1855632961,-0.2716336548,-0.2943277061,0.0332474858,-0.3944120705,-0.0690758079,-0.1865726858,0.0916663632,-0.1238563731,0.0107898619,-0.3834721446,-0.3246234655,0.0348961055,0.1055618227,0.0962771103,-0.0437326357,-0.0064712348,0.0819927007,0.0759535581,-0.1092398837,0.1654802561,-0.0799267516,0.0434889235,-0.030207064,-0.0392674692,0.2523209453,0.0701688081,0.1473389119,0.358079344,0.2618344724,0.0386816859,-0.2227414399,0.2019043416,-0.0103037534,-0.1117861345,-0.1631402969,0.1722758859,0.0063027446,0.0302214883,0.3716366589,0.1245407686,-0.1198315695,0.3440977931,-0.133833006,-0.1742625386,-0.4563521147,0.2229207456,0.1796982288,-0.1642237604,-0.1059800684,-0.1711052656,-0.0731739476,0.0550526232,0.4335769117,0.0114540998,0.1584357768,-0.2045156509,0.0911283046,0.0008950661,0.5557250977,0.1665425599,0.2724840939,-0.2918747067,0.0434722118,-0.8117441535,-0.1683024168,-0.2191194445,0.4150376022,-0.0744278878,0.0899982974,0.0912720263,0.4602128863,0.092196174,0.164307192,0.0427024737,-0.1950075328,-0.3067815602,-0.0896473899,-0.125637427,-0.044790484,-0.0425439402,-0.25630036,0.1380438507,-0.3528144658,-0.025443105,0.0278754104,0.1930948645,-0.159822464,-0.0996007174,0.4608584046,0.1199982688,0.4414413869,0.1976628751,-0.0157539211,-0.1962549686,0.0738468468,-0.071666874,0.1275081486,-0.0289347097,-0.005290688,-0.1757265776,-0.1376700252,-0.1885929555,0.1871832311,0.029352732,0.0823310688,-0.0889895409,-0.1222604886,-0.0735001788,0.1361116618,0.0541609451,0.0281361546,-0.1415182352,0.1670806855,-0.1907015741,-0.0634676591,0.3272378743,-0.0947411284,0.0126258796,-0.5545053482,0.3333742917,0.0527845174,0.0571754053,-0.3271178305,0.1505368501,0.269619137,-0.214347899,-0.1104709804,0.2254099995,-0.2465933859,-0.2582942843,-0.0356632881,0.286562562,-0.2709173262,-0.2224069834,-0.1685857326,-0.1292221546]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/824","title":"Discussion using datasets in offline mode","comments":"Requiring online connection is a deal breaker in some cases unfortunately so it'd be great if offline mode is added similar to how `transformers` loads models offline fine.\r\n\r\n@mandubian's second bullet point suggests that there's a workaround allowing you to use your offline (custom?) dataset with `datasets`. Could you please elaborate on how that should look like?","body":"`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n","comment_length":57,"text":"Discussion using datasets in offline mode\n`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n\nRequiring online connection is a deal breaker in some cases unfortunately so it'd be great if offline mode is added similar to how `transformers` loads models offline fine.\r\n\r\n@mandubian's second bullet point suggests that there's a workaround allowing you to use your offline (custom?) dataset with `datasets`. Could you please elaborate on how that should look like?","embeddings":[-0.4731812477,0.2457838953,-0.0126306163,0.1412141472,0.2833554745,-0.1470207572,0.6012274623,0.0127757993,0.2687810659,0.1312778145,-0.0240368135,-0.0239853207,-0.0246622115,0.4021067917,-0.0912403315,-0.1021380574,-0.1722334921,0.0388772562,-0.1540046185,0.1325951368,-0.1408108175,-0.1281190962,-0.3727094233,-0.0282363463,-0.1628144085,-0.1865869015,0.009141881,0.1420212984,-0.157231614,-0.2762964368,0.2854104936,0.1805244088,0.2238958478,0.4791246653,-0.0001042308,0.1437665671,0.0545553416,-0.0134886838,-0.2069027573,-0.3381868601,-0.1917754114,-0.3979368508,0.2337578535,-0.4007375836,-0.2150847316,0.0365372971,-0.0734726861,-0.395838052,0.3150739074,0.3079384267,0.2526885569,0.6633725762,-0.14806135,-0.3228698373,-0.1723310798,0.010202216,-0.2392340451,0.3925467134,0.0617909804,-0.0708869025,-0.1110986471,0.1163179576,-0.119291313,0.0119442893,0.3592294753,0.0990151986,-0.0436064601,-0.0926574692,0.1286549419,0.4549899101,0.3857674897,-0.2256464809,-0.3233912885,-0.4115032256,-0.1213282347,-0.3992854655,0.0777640641,0.0915819705,-0.0518016815,0.2859633863,-0.4508610964,-0.2770608366,-0.3688122928,0.0866334587,-0.3436934948,0.129058823,-0.2209807485,0.0518599302,-0.0594523624,0.1764986664,-0.1783763617,-0.2810511887,0.1342646629,-0.131892249,-0.2653567493,-0.0710257813,0.1660244018,0.0489503071,0.1325022876,0.0835703537,0.245603919,0.0134868138,0.0630616322,0.0728721172,0.4931008518,0.2701188326,-0.066404894,-0.1654520333,0.590166986,0.1680280119,0.0154071264,-0.2984622121,0.34388116,-0.247408554,0.0511436015,-0.0044536674,0.3934408128,-0.2904764712,-0.0705588907,-0.028938571,0.2609572411,-0.0422502495,0.2302477062,0.2531334758,-0.200735867,0.0239093136,0.0023878447,0.3707030416,-0.2583317161,-0.287779361,-0.0691777095,-0.2514779568,-0.1524575353,0.0909147337,0.1942048073,0.0988478512,0.0931645483,-0.0352786593,0.2799161375,0.0193171445,0.2784254849,-0.0613344237,-0.0014951745,0.3023600876,0.1102334112,-0.0681090355,0.1214372739,-0.1861234307,0.0257518515,0.0565423518,-0.0653337613,-0.4622017443,0.1761785448,0.219720602,-0.6732071042,-0.2223996818,-0.2730343044,0.1555661261,-0.2369358987,-0.2124957442,0.0235868525,0.018460894,-0.2118948549,-0.211705938,0.2878770828,0.5173381567,-0.4603885412,0.0129500646,0.0447110832,-0.2136073858,-0.2071649134,0.1019181535,-0.414919436,0.6289384961,0.0107537126,-0.3816751838,0.5595380068,-0.3048187792,-0.3140288293,0.2726414204,-0.1605727226,-0.0045871525,-0.0534324944,0.1087546423,0.4603524208,0.084867388,-0.0908238813,0.2521170378,-0.128145203,-0.0130102765,-0.2615109086,-0.317256242,0.1956141591,0.3260177076,0.3225422502,0.170808658,0.2575958967,-0.1861371994,-0.0085249459,-0.1622692347,0.1202049702,0.0423874669,0.2255617529,0.0671963319,-0.0102309389,-0.1653408855,-0.5502773523,0.2796308994,-0.0287088715,0.2188271582,-0.2158027589,-0.210777849,-0.027634386,-0.1952420026,-0.181086868,-0.0795596987,0.1171106398,-0.1978259683,0.0859222189,0.0982890129,-0.2900714576,0.1137299016,-0.011981139,0.0217427425,-0.1584365219,0.3012230992,-0.0498387516,0.0836622119,-0.1551096439,0.0562893637,-0.0545988455,-0.1622434705,-0.0954587534,0.5110428333,-0.02388593,0.0563261919,0.2349392772,0.5590957999,0.3679666519,0.0043761171,0.495628953,-0.050213553,0.3213089108,-0.1114702821,-0.3984197676,0.7357347608,-0.144050926,0.2574732006,0.0989937186,-0.0164562874,0.2910700142,-0.0265043136,-0.4488827586,-0.0220366903,0.0183654837,0.0145753147,0.2139872313,-0.0351547152,-0.3072082996,0.0073910407,0.3660686314,0.0653327256,0.1509138197,-0.0016662013,-0.0714008063,0.0618565455,-0.0847105533,0.0961386189,0.5059415102,0.222510621,-0.0020846452,0.222269088,-0.0752183199,-0.2918922305,0.1622450352,0.2784020007,0.0117790736,0.0948213711,0.1106504574,-0.0135444626,-0.2606648505,-0.0851194039,0.0104148071,0.0521303751,-0.3171434402,0.3023553789,-0.192141518,0.0446793325,0.0116658304,0.1038789749,-0.503385365,-0.1916292906,0.1738934666,0.0866980851,-0.2768389881,0.1611506343,-0.0816928521,0.5301485658,-0.1862419546,-0.172270909,0.1785868406,-0.587223351,-0.0618414544,0.0992661193,0.3291293085,0.031592764,0.4697214663,0.1105318367,-0.100488849,-0.2267502248,-0.2593652904,0.0775288045,-0.0444324464,0.3455691338,0.0793805569,0.0427691117,0.0755403414,0.2921114266,0.4025548398,-0.2950235307,0.2515189648,0.1761236638,-0.124143742,-0.0401732139,-0.3764849007,-0.432044059,-0.3803494871,-0.4162005484,0.2004627436,-0.0402469784,-0.0421941429,0.2894975543,0.1804338396,0.0699566826,-0.0712620169,0.0905176699,-0.1396664679,-0.4403935373,0.2246841788,-0.1572482735,-0.5371839404,0.208421275,0.1666212529,0.1439089775,0.2761147022,-0.5148002505,-0.3084290028,0.0591682941,0.3579508364,0.1813364923,0.2270392478,0.3989451826,-0.1022249237,-0.0224113967,-0.0809667036,0.069430992,0.2551833987,0.0415008217,-0.0146295261,0.0407630838,0.3204085529,0.0124633554,0.4728292227,-0.065537177,0.0620023683,0.2407371551,-0.0948777795,0.6474258304,-0.2596338689,-0.1485426873,-0.0732285753,-0.0469047278,-0.3458383381,0.2646626234,0.0285727102,-0.0157961119,-0.5687811971,-0.0163188912,-0.2486623079,-0.4440139234,0.1849075556,0.2254671901,0.341771543,0.1493411511,0.2687122822,-0.2899604738,-0.0377983525,-0.0554851852,0.2830599546,0.0635943115,-0.1171806082,-0.2402947396,0.0881897137,-0.3629257381,0.2781030834,-0.2016147673,0.1761330813,-0.0393386632,-0.0541603342,0.0559929349,-0.0758442134,0.4782177806,-0.381919235,-0.1164454073,0.2526444495,-0.1034650579,-0.3154055178,0.0472196713,-0.2072025537,-0.0110109895,0.1880609989,0.2789724767,-0.093525067,-0.2863284945,-0.0143172154,0.2607756257,-0.3398856819,-0.225745365,-0.2240486741,-0.0473369732,-0.3530589938,0.0501264073,-0.167606011,-0.0237873774,-0.0988303274,0.1071026474,-0.1596325636,0.1846583635,-0.0727059543,0.1555313468,0.0937974826,-0.3015151918,0.2017156482,0.2891573906,0.0554315485,0.2439215332,0.3130555749,-0.0757674202,-0.2507785261,0.2184928507,-0.0449131168,0.3743802011,0.447499752,-0.0075464849,-0.1503677517,0.2359789461,0.2909380794,-0.4307907522,-0.132485345,0.1148859784,0.0783318803,-0.1366835535,-0.4067745507,0.2473443598,-0.0049635419,-0.0512084402,-0.0943640172,0.158539176,-0.239334777,0.3174614906,-0.1068656743,0.7930578589,-0.0637973845,0.2442911565,0.2223540246,-0.0119604915,0.4988227487,-0.3252683282,-0.0229490343,-0.0340016484,-0.178710714,-0.031381879,-0.0793323442,0.4078509808,0.0548717044,-0.1353465617,0.5168308616,-0.033894185,0.3856006563,-0.2857428491,0.1290824413,0.0501796305,-0.2691266239,-0.5489073992,0.2034656703,0.0918537602,0.0732088536,-0.2224258482,-0.2794910669,0.0674428195,0.0627634749,-0.0266709588,-0.1881917119,-0.0325168557,0.0757756308,-0.1396136135,0.0518587716,0.0764454827,0.1227718964,0.4003405869,0.1215793565,-0.4199986458,0.0589224473,-0.3985019624,0.1547273844,0.022546811,0.1779079884,0.1813083887,-0.1888163686,-0.2497040778,-0.0088854646,-0.1091714054,-0.1067306623,0.4405538738,-0.1216345727,0.0073922765,-0.3226769865,0.1737302244,0.1859423071,0.2177406698,-0.1645352542,0.1630951017,0.2032280415,0.0897041857,0.3784289956,-0.1469400674,-0.1935924441,-0.2416878939,0.2845247388,-0.1109834984,-0.1488573253,-0.0757598206,0.2245462835,-0.1876361519,-0.3061795533,0.2268287987,0.0737866685,-0.2808648944,0.2850721776,0.3767939806,0.2331891805,-0.1165045202,0.2991285622,0.0643195137,-0.2871190012,-0.1611160934,-0.3281604946,-0.3837020099,0.2428532243,0.4140175879,0.1610078514,0.0781943873,-0.0397630334,-0.0063381726,-0.1876193434,-0.4073794782,0.0217937957,-0.0592033528,0.0651988685,0.2616648972,-0.2107882202,0.3878107071,-0.2093546391,0.0535878912,-0.1886049062,-0.4114201069,-0.2793820798,-0.0371242762,0.1872642636,0.0342983678,-0.3216848671,-0.0484384447,-0.4539426565,-0.39113608,-0.1426026672,0.2971313596,0.4522984922,-0.140037477,0.0119625991,-0.1794686913,0.4436798394,-0.1515535265,0.0933163539,0.0538431741,0.2722334862,0.0794811845,0.2418547422,-0.1242847592,0.0662921071,0.0052196677,0.1862123907,0.2457852364,0.1565163732,0.4803642929,-0.1563785523,0.1423408836,0.3202059567,0.5329152346,0.1267689168,-0.0554246865,0.189650327,0.0750887319,0.2618720829,-0.418261826,0.0555132851,0.0559120215,0.0692081377,0.3451641798,0.1656796932,0.4651870131,0.1580532342,-0.0120311966,0.044956658,0.3792149425,-0.0836964548,0.2313958108,0.1390156299,0.0545487143,0.146652475,0.0978635177,0.3678914309,0.027954692,0.2849600911,-0.3214758039,0.1772391945,0.0446814969,0.1018272489,-0.1581350565,-0.4124343395,0.2898585498,-0.0497147217,-0.0546015091,-0.182363838,-0.2525885701,0.1863817126,-0.4326297641,0.0116935298,-0.2156715542,0.064670369,-0.0798990875,0.0103112217,-0.372207731,-0.3459404707,-0.021677224,0.11335437,0.0338323787,-0.0128650684,0.0488108546,-0.0818488076,0.0918953419,-0.2111901045,0.1460550278,0.0580454729,0.0776614994,-0.0307539124,0.0314861163,0.2307014912,0.0295002423,0.1551377624,0.492482245,0.2937014699,0.0306279305,-0.2166259438,0.1912288964,-0.0251139607,-0.0947622508,-0.1152831689,0.1165185496,0.0572031513,0.1724907607,0.3164797723,0.1543086618,-0.2009875178,0.3622323275,-0.2038220763,-0.0695498586,-0.4959087074,0.316302985,0.1731223017,-0.1483541131,-0.0404534265,-0.262960732,-0.1618154198,-0.2425553203,0.3993487656,-0.0187275298,0.1477475613,-0.1969573498,0.1060572341,0.0271289703,0.445312053,0.1525288522,0.2580768168,-0.270478785,0.2232396156,-0.7864069343,-0.1592854857,-0.0789298192,0.3429844975,-0.0179707427,0.1001323238,0.1175370142,0.4013324082,0.0578604564,0.0657564029,0.1440416723,-0.1705752611,-0.2648237944,-0.087211445,-0.1376920193,-0.0596197918,-0.0228805896,-0.2512314916,0.1176674142,-0.2488218993,-0.0115263062,0.0107077006,0.1454161704,-0.1141222119,-0.0654080883,0.4388478696,0.1017874703,0.3600196242,0.254740119,-0.0651025251,-0.1099723056,0.0138451476,-0.0983546004,0.0416437387,-0.0822127983,0.0232169107,-0.1789954305,-0.1889515519,-0.1935424209,0.1395343542,0.0901735947,-0.0838337243,-0.2044851184,0.022094436,0.0032965345,0.1341684759,0.0018572641,0.0741954967,-0.1065519899,0.1723715365,-0.1570728719,0.0643467456,0.3616630137,-0.0158102773,-0.0513042994,-0.3619368374,0.2845944464,0.2454409152,0.0160163436,-0.6187485456,0.0592007823,0.2603121698,-0.1704741567,-0.1079792902,0.2965720594,-0.1877588481,-0.2338690609,-0.0189625323,0.2834101915,-0.2633384168,-0.2472620606,0.0018668523,-0.1285084486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/824","title":"Discussion using datasets in offline mode","comments":"here is my way to load a dataset offline, but it **requires** an online machine\r\n1. (online machine)\r\n```\r\nimport datasets\r\ndata = datasets.load_dataset(...)\r\ndata.save_to_disk(\/YOUR\/DATASET\/DIR)\r\n```\r\n2. copy the dir from online to the offline machine\r\n3. (offline machine)\r\n```\r\nimport datasets\r\ndata = datasets.load_from_disk(\/SAVED\/DATA\/DIR)\r\n```\r\n\r\nHTH.","body":"`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n","comment_length":47,"text":"Discussion using datasets in offline mode\n`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n\nhere is my way to load a dataset offline, but it **requires** an online machine\r\n1. (online machine)\r\n```\r\nimport datasets\r\ndata = datasets.load_dataset(...)\r\ndata.save_to_disk(\/YOUR\/DATASET\/DIR)\r\n```\r\n2. copy the dir from online to the offline machine\r\n3. (offline machine)\r\n```\r\nimport datasets\r\ndata = datasets.load_from_disk(\/SAVED\/DATA\/DIR)\r\n```\r\n\r\nHTH.","embeddings":[-0.4902576208,0.2288961858,-0.0332210772,0.138873294,0.2363725007,-0.0891150385,0.5481942892,0.0673723891,0.2955965102,0.2450511903,-0.0101743611,-0.0694922432,0.0276259091,0.3827281892,-0.1057184339,-0.0218463782,-0.1530392021,0.0536614954,-0.1772286743,0.0896224678,-0.1652288735,-0.0921332538,-0.4337235987,-0.0704064891,-0.0181331802,-0.1504255235,-0.0832125545,0.219438836,-0.1661252081,-0.2807519138,0.238820821,0.1633139551,0.2291972488,0.475061357,-0.0001033812,0.0263173375,0.0082090031,-0.0013799195,-0.2269736379,-0.4025881588,-0.2158048302,-0.5285046697,0.1796488315,-0.4550701976,-0.2032037377,-0.0582043938,-0.0745971277,-0.4929271638,0.3164803088,0.3498669565,0.2593776882,0.5757408142,-0.1358263642,-0.2754445672,-0.1606792808,-0.1514874995,-0.1493050307,0.4366157353,-0.0145771503,-0.0082134157,-0.0025773346,0.0952334329,-0.1961562037,0.0413857102,0.3163404763,0.0586813726,-0.1016099751,-0.1018178985,0.2198021561,0.3930848241,0.4261620939,-0.2737801671,-0.2470030934,-0.339004308,-0.0023413347,-0.4567470551,0.0788485259,0.1836712658,0.0242283177,0.224884212,-0.3797212541,-0.3182393014,-0.3313245773,0.1638136953,-0.32146281,0.0111691775,-0.2536790669,0.0693017393,-0.0612447411,0.0913351476,-0.1768202484,-0.28515324,0.073911421,-0.0619464032,-0.2594501376,0.0268554222,0.1183790341,0.0684494004,0.1312052011,0.0652054697,0.2398014665,-0.018818114,0.0791069493,0.0497344695,0.4407645464,0.2978351414,0.0075793457,-0.176871866,0.5290125012,0.091033712,-0.0241946559,-0.29890728,0.2565964162,-0.3073076606,0.1321211159,0.0137979519,0.3654094338,-0.3048083484,-0.158787936,-0.0139988959,0.1392609626,-0.0424629487,0.1689319164,0.2522922754,-0.2716249526,0.0451420024,0.0697742701,0.3969351351,-0.2989778817,-0.2657657862,-0.0589482971,-0.1259195954,-0.1128110662,0.0637708753,0.178990826,0.0000603757,0.0520434454,-0.0411280394,0.2421651185,-0.0735087693,0.2981470525,-0.0951727554,0.0823780447,0.2353083044,0.2750588059,-0.0874181688,0.0769162402,-0.2003214508,0.0379160382,0.038790673,-0.1426270306,-0.3582569659,0.170656845,0.2765568197,-0.6289308071,-0.1753666699,-0.321626395,0.0367598273,-0.3161136806,-0.1942735314,-0.0308537446,-0.0616714284,-0.1780046523,-0.2453662008,0.3240109086,0.5668384433,-0.383969754,0.1050266996,0.0844061449,-0.160434112,-0.1781923324,0.023398824,-0.3148404658,0.5233699083,-0.0130562568,-0.3010381162,0.5681838393,-0.3550915122,-0.2651605904,0.2487461865,-0.1917605847,-0.105965592,-0.1038612276,0.0901149809,0.440076679,0.0750485212,-0.0843940079,0.289010942,-0.0351713262,0.0004676559,-0.2186113894,-0.2369781137,0.1673176885,0.2837593257,0.255562067,0.1129005253,0.3047425449,-0.1972569376,0.0391629934,-0.2821209729,0.1191658303,0.1654856652,0.1697198749,0.1023969799,-0.044208128,-0.1828145236,-0.48932302,0.2718419433,0.0478837043,0.1571957916,-0.199201867,-0.1379158646,-0.020388905,-0.1831310242,-0.1769697517,0.0272823572,0.1410640925,-0.1107874289,0.1724625081,0.1293686181,-0.2598986626,0.1782922298,0.0107880915,0.002460724,-0.1300739944,0.2903401554,-0.0492371321,0.0646001175,-0.0289290342,0.0881469399,0.0581811368,-0.1937225014,-0.1185834855,0.5259695649,-0.0300909895,0.0616296493,0.26878497,0.3818560541,0.306464076,-0.0120455539,0.5338500142,-0.0427261144,0.3334380388,-0.0925104991,-0.3228542805,0.7542521954,-0.1525749713,0.2171272337,0.1025260761,0.0183239989,0.2599797547,0.001278999,-0.4040777385,-0.0359521806,0.1184480637,0.0022458544,0.181137234,-0.0850736648,-0.3064594567,-0.0152912978,0.3312467039,-0.0188398156,0.1778849959,-0.0660747066,-0.0304849781,0.1306045651,-0.0831728801,0.0887685046,0.516928196,0.2506965101,0.0519572198,0.2298254222,-0.0313782431,-0.2562146187,0.132566765,0.2321359664,0.0492413528,0.0684337541,0.1004821807,-0.0237272847,-0.2063807398,-0.052083157,-0.0547190756,0.0232721549,-0.2868103087,0.2843016386,-0.2516606748,-0.0163154081,0.0470518991,0.1107893586,-0.4484026432,-0.2002999336,0.0647204295,0.077696614,-0.2318820804,0.1336179227,-0.1268714517,0.3907523155,-0.2403125763,-0.2388580739,0.2622436583,-0.4872460365,-0.1227765456,0.1072288454,0.3064813912,0.1311014891,0.5087970495,0.0214199591,-0.0971539766,-0.3305742443,-0.1572176069,0.0155851031,-0.065936245,0.2778904438,0.0809208751,0.0578192733,-0.0453803204,0.3248309493,0.3875049353,-0.3374926746,0.2036764622,0.1821809709,-0.0564875826,-0.0917128548,-0.3432077765,-0.431303978,-0.4415458739,-0.3976587355,0.1741933823,0.0349772051,-0.0529258959,0.1827891618,0.1519728601,0.1029269621,0.0166772045,0.1853231043,-0.182843402,-0.4674518108,0.1784860343,-0.1808056384,-0.6038360596,0.2909515798,0.2311562598,0.1528016329,0.3490706384,-0.4436715245,-0.3765623868,0.0681066141,0.4156066775,0.1085773632,0.3436664641,0.343449086,-0.0702814162,-0.0382861346,-0.0923792347,0.0928403884,0.1769469231,0.0589718558,-0.1342711449,0.0336064994,0.2543356419,-0.0294251926,0.3790815473,-0.0375484191,0.0918622464,0.2364589721,-0.157876581,0.6929191351,-0.1827820092,-0.1525170207,-0.039478302,-0.0537111871,-0.3249545097,0.2491013706,0.0245234203,0.0068162805,-0.6647506952,-0.0420180112,-0.3744178116,-0.4079521298,0.0988641307,0.1220592335,0.366121918,0.1131866798,0.2727554142,-0.2575809956,-0.0872171074,-0.0868246779,0.2532112002,0.0775293484,-0.0718222409,-0.2494612336,0.0245864484,-0.3649164438,0.3807895482,-0.1454290003,0.1998325139,-0.0346336514,-0.0230422541,-0.0126947565,-0.1682794541,0.3438601196,-0.3169399798,-0.0129384911,0.2967363298,-0.0679403841,-0.3190099597,0.0656508207,-0.2115930915,-0.0245799068,0.1876072735,0.1661459804,-0.1067496911,-0.3034740686,0.0126637649,0.4440042675,-0.2934578657,-0.1976154596,-0.2340571731,-0.1652711779,-0.3448308706,-0.0077203051,-0.2022228986,0.0696706101,-0.120433487,0.1846712679,-0.2544874251,0.1687718034,-0.0028257163,0.1041491404,0.1279095113,-0.2341517359,0.2522114217,0.2547443509,0.1065212265,0.3232454658,0.3494312167,-0.0530188195,-0.2862736583,0.1681595892,-0.0160192922,0.35865435,0.3766712248,-0.0966337919,-0.1404381096,0.1375189573,0.2744402885,-0.4191712141,-0.1013082415,0.1470344663,0.179600358,-0.1190664992,-0.3914581239,0.2641130686,0.1082138494,-0.0258151069,-0.0286947861,0.1335439533,-0.2077489644,0.1910682172,-0.1117649153,0.7465305328,-0.0877654478,0.2242861986,0.2535328865,0.0000774055,0.4418720305,-0.2145643085,-0.0392789468,-0.0411001146,-0.2530001998,-0.0048275688,-0.0833426937,0.4119864106,0.0042823022,-0.0889123157,0.5612853765,-0.0295867305,0.2540517151,-0.3048312664,0.2157215625,-0.0231642257,-0.220427081,-0.5347032547,0.2236427069,0.1193529144,0.1017456502,-0.1956091076,-0.2288061231,0.0823684409,0.073649317,0.0393219106,-0.0716270506,-0.0051057399,0.0971567258,-0.2511774302,-0.0116195045,-0.0075248764,0.142170012,0.3566609919,0.1451168209,-0.301761061,0.0237517711,-0.3742330968,0.1623167694,0.0489318818,0.2389555126,0.171368286,-0.2298683375,-0.2295824289,0.0770820379,-0.0900077,-0.0092648016,0.42446208,-0.0489098914,-0.009291999,-0.234820202,0.1417715251,0.1884103119,0.190955773,-0.1208170429,0.1698993146,0.2535117865,0.0531004742,0.3266043067,-0.113379851,-0.222156018,-0.2290591747,0.32467556,-0.1461772025,-0.21636419,-0.028165346,0.2071454823,-0.1757817715,-0.3009601235,0.2828659415,-0.0069940356,-0.125973165,0.2331747711,0.3803782463,0.3927211165,-0.1148344055,0.2829953134,0.018474292,-0.2470531166,-0.0675074756,-0.3831626177,-0.2631471753,0.204870075,0.479160428,0.1200163513,0.1498101652,-0.0835262686,0.0403611585,-0.2564301491,-0.4158100784,0.0577494279,-0.0461971648,0.0455120951,0.2434913665,-0.1222786531,0.3542178869,-0.2156258523,0.099288784,-0.23531425,-0.3157812357,-0.2837586999,-0.0089733656,0.1735137999,-0.0462745689,-0.3089570999,-0.0029457209,-0.6132439971,-0.3331042826,-0.1570519507,0.2411421537,0.4147737026,-0.1820590794,-0.0197102241,-0.1957209855,0.4264013469,-0.190115422,0.1095879674,0.0468604527,0.2768315375,-0.0304182582,0.2757185996,-0.0449677818,0.0424675085,-0.0480308793,0.0965553895,0.1639403999,0.0903308988,0.3923719227,-0.2577670515,0.1543769687,0.2673698068,0.5688072443,0.197502777,-0.0605365895,0.2026238889,0.1057217196,0.2705929875,-0.3872043192,0.0592557527,-0.006890045,0.1027035862,0.3957237601,0.0998353064,0.3355102241,0.1089522466,-0.0963628218,-0.012912211,0.3361456394,-0.1054278761,0.2378420085,0.1243169531,0.0390205272,0.1391032487,0.0136831002,0.419791311,-0.0107136974,0.3140262961,-0.3290991783,0.242822513,0.1165318862,0.0871019736,-0.297940582,-0.4573815465,0.2896075547,0.021437848,-0.0385113545,-0.1695443839,-0.2580637634,0.1817516237,-0.5244922638,-0.0815284476,-0.1498389244,0.1035999358,-0.0662675425,0.0504695885,-0.4011798203,-0.3024998605,-0.0331322253,0.1197491661,0.0892832279,-0.0883325413,-0.0261482205,0.0244443361,0.1078766584,-0.2080112696,0.1284511536,0.0660775825,0.0618913285,-0.0348128304,0.0230818335,0.2032921165,0.0634403527,0.1143965125,0.4303237498,0.3389603198,0.0386465937,-0.178898558,0.1777808219,-0.01888909,-0.123788394,-0.0961964726,0.0878961831,0.0156530626,0.1670670807,0.372803092,0.1951121539,-0.2173009068,0.3377450109,-0.0706461817,-0.1390172094,-0.4647113383,0.2669894099,0.191283986,-0.1077181026,-0.1194827706,-0.2801829278,-0.1833825558,-0.1093611494,0.3616840541,0.0254851151,0.1069909036,-0.1845504493,0.1272707283,0.0930400193,0.4290935099,0.1821925789,0.3078639507,-0.285238713,0.1113601699,-0.8087354302,-0.150158748,-0.1427772343,0.2350052297,-0.0523448922,0.1085456163,0.0847963467,0.4387237132,0.0659141392,0.1579321772,0.1626468599,-0.1813364178,-0.2886351347,-0.1446584165,-0.0816260204,-0.0860896483,-0.0314259566,-0.3293494284,0.0789925158,-0.385258764,0.026291145,0.0775821805,0.1974054575,-0.2024395615,-0.1328171492,0.4599530399,0.0243595764,0.4064572155,0.1564127356,-0.1124192998,-0.0652244613,0.0353197679,-0.057773985,0.1724690795,-0.0338675082,-0.0269051567,-0.1379747987,-0.1405127794,-0.16621086,0.2008037567,0.0666892827,-0.0119728008,-0.1137939543,-0.0359998457,0.086795181,0.1418627352,0.0072061275,0.0431739576,-0.1489287317,0.2132104337,-0.2026586831,0.045329053,0.339382261,-0.0350239351,0.0127118817,-0.3722335696,0.392375648,0.1644823104,0.00088806,-0.4804071486,0.2365405709,0.3327222168,-0.1532103419,-0.1472786814,0.3244405091,-0.2522740066,-0.2001632452,0.0071670003,0.2544192374,-0.270077318,-0.3051465154,0.0650704429,-0.1409677416]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/824","title":"Discussion using datasets in offline mode","comments":"> here is my way to load a dataset offline, but it **requires** an online machine\n> \n> 1. (online machine)\n> \n> ```\n> \n> import datasets\n> \n> data = datasets.load_dataset(...)\n> \n> data.save_to_disk(\/YOUR\/DATASET\/DIR)\n> \n> ```\n> \n> 2. copy the dir from online to the offline machine\n> \n> 3. (offline machine)\n> \n> ```\n> \n> import datasets\n> \n> data = datasets.load_from_disk(\/SAVED\/DATA\/DIR)\n> \n> ```\n> \n> \n> \n> HTH.\n\n","body":"`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n","comment_length":76,"text":"Discussion using datasets in offline mode\n`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n\n> here is my way to load a dataset offline, but it **requires** an online machine\n> \n> 1. (online machine)\n> \n> ```\n> \n> import datasets\n> \n> data = datasets.load_dataset(...)\n> \n> data.save_to_disk(\/YOUR\/DATASET\/DIR)\n> \n> ```\n> \n> 2. copy the dir from online to the offline machine\n> \n> 3. (offline machine)\n> \n> ```\n> \n> import datasets\n> \n> data = datasets.load_from_disk(\/SAVED\/DATA\/DIR)\n> \n> ```\n> \n> \n> \n> HTH.\n\n","embeddings":[-0.4992605448,0.2269978523,-0.0324693248,0.1418719441,0.2369505018,-0.1029172167,0.5442941189,0.0744111389,0.2753627598,0.2442882061,-0.0088339234,-0.0665396452,0.0280860681,0.3756225705,-0.0988998637,-0.0419552475,-0.1623334885,0.0563546792,-0.183036536,0.1076543331,-0.1633780748,-0.0869517475,-0.4333003163,-0.0625077635,-0.0159387887,-0.1606817096,-0.0798981264,0.2115724087,-0.1717491895,-0.2845186889,0.2436344028,0.1665008962,0.1988918632,0.4731360674,-0.0001034733,0.0370239355,0.0119215008,0.0104895122,-0.2277320027,-0.3849576116,-0.2243474573,-0.5146489739,0.1638061702,-0.4546643198,-0.2070107609,-0.0379570685,-0.0688914806,-0.5084757209,0.3095602393,0.3518121243,0.25706622,0.5904164314,-0.1282708943,-0.2792427242,-0.1558488756,-0.15226309,-0.1556978971,0.4283836484,-0.0172956381,0.0013684349,-0.0222441554,0.1041298062,-0.1752347797,0.0346602872,0.3222211003,0.0672788918,-0.1097400188,-0.0845898315,0.2206331044,0.3967526555,0.4247180521,-0.2884344757,-0.25191921,-0.3396162391,-0.0075891577,-0.4678492844,0.0960984454,0.1565518826,0.0217524823,0.2291809618,-0.3944315314,-0.3091626763,-0.3443603218,0.1524468064,-0.3242220581,0.0023712981,-0.2609032691,0.0710143372,-0.0611248761,0.0929871574,-0.1979121715,-0.2828268111,0.0606515519,-0.0742605627,-0.2604793012,0.0141299367,0.1201808229,0.0637012869,0.1402420402,0.058837384,0.2486954182,-0.0098348763,0.0979686975,0.0510386862,0.4422774613,0.2957024574,0.0074487762,-0.18953906,0.5416590571,0.0993699357,-0.0075667072,-0.2980296016,0.2820797563,-0.3101067841,0.1336489916,0.0073752431,0.3609673977,-0.2954883873,-0.1674440801,-0.0125406291,0.1405458748,-0.0426560454,0.1835252494,0.248358503,-0.2873734534,0.0299902372,0.0637684539,0.3856902122,-0.2997954786,-0.2546656728,-0.0616325028,-0.1359128952,-0.1291581094,0.0727540553,0.1864255965,0.0080418475,0.0445535854,-0.0334911197,0.2402406931,-0.0691183209,0.3054921031,-0.1039336249,0.0785027891,0.2336844802,0.2520386577,-0.0957668126,0.0624969415,-0.1969115734,0.0382462256,0.041058708,-0.1074667722,-0.3596408069,0.178369388,0.2726848423,-0.6226016879,-0.1710650325,-0.3079641163,0.0581852198,-0.306432128,-0.1984430552,-0.0153495949,-0.0460300557,-0.193595171,-0.2405310273,0.3376948535,0.5559729338,-0.398294121,0.0968975052,0.0791635215,-0.1771891415,-0.1863670498,0.0195536334,-0.3082571626,0.5199754834,-0.0187853388,-0.2941873968,0.5648753643,-0.3655905724,-0.2701430023,0.2469820678,-0.1869648099,-0.1120411009,-0.0934418514,0.0912888423,0.444742173,0.0852880627,-0.0959229767,0.2809376121,-0.0458619855,-0.0005543638,-0.2211628258,-0.2394605875,0.1544378698,0.2881146371,0.2723816335,0.0969029889,0.3091706038,-0.1952723414,0.0280246921,-0.2787548006,0.1350964159,0.1634455919,0.1533000767,0.0920097008,-0.0438158251,-0.1929366887,-0.4884028733,0.2647079527,0.041078344,0.1632880867,-0.2042591572,-0.1388626993,-0.0210518315,-0.1890713125,-0.1622339189,0.0133329192,0.1397613436,-0.1002759784,0.1611437947,0.1137790605,-0.2563475668,0.1799523532,0.0201561693,0.0074801487,-0.138511315,0.269367069,-0.0576035604,0.0668179542,-0.0368147902,0.0932565108,0.0424022302,-0.1958928406,-0.1061084792,0.529029727,-0.0309691075,0.0468532704,0.2650736272,0.3803910315,0.3248943985,-0.011337583,0.5266875029,-0.0440164357,0.3241901398,-0.0970292091,-0.3409502506,0.7555660009,-0.1524325758,0.2127019763,0.1112290621,0.0115109617,0.265591383,0.0024921908,-0.4044695199,-0.0297388695,0.1032060906,0.0016839154,0.1952834725,-0.0857538134,-0.3268732727,-0.0021086955,0.3489390612,-0.0138921747,0.1751540899,-0.0613358766,-0.0318150036,0.1201424897,-0.0734539181,0.0859299004,0.5219720006,0.2544083893,0.0612387247,0.2307022661,-0.0414699912,-0.2559099793,0.136794433,0.2250438631,0.0342362486,0.0703126267,0.0828345418,-0.0354552269,-0.2182721943,-0.0521009415,-0.0413502678,0.0165150035,-0.2833811343,0.2871290445,-0.2482628822,-0.028776994,0.0571506321,0.1327002645,-0.4499962926,-0.1922508031,0.0758107826,0.0834263191,-0.2347825021,0.1367863566,-0.1147691905,0.4157854319,-0.2417856604,-0.2162092179,0.2557033598,-0.489272207,-0.1254084259,0.1139194816,0.2904767692,0.1211071759,0.5174885988,0.018219661,-0.0972281322,-0.3355351686,-0.1610652059,0.0355822183,-0.0712829605,0.2829785645,0.0933507159,0.0530732386,-0.067163527,0.3253516853,0.3937312961,-0.3271920383,0.1934456229,0.1792645156,-0.0497172922,-0.0871885717,-0.3529951572,-0.4282158017,-0.432069838,-0.3919803202,0.192620337,0.0132437786,-0.0574363023,0.1898374707,0.1482215226,0.116527997,0.0058472878,0.1888998151,-0.1883189976,-0.4760995209,0.1688492447,-0.189299643,-0.6016700268,0.2806211114,0.2295092046,0.1657632887,0.3582723439,-0.4448154569,-0.3781495988,0.0732920542,0.4401058555,0.110242337,0.3342652023,0.3533639014,-0.0593588874,-0.0430468284,-0.0973874703,0.0781206414,0.1909287125,0.0516173132,-0.1234366819,0.0378183462,0.2629092932,-0.0286334138,0.4023732245,-0.0331205502,0.0967799574,0.2285624743,-0.1566263735,0.7003893256,-0.1845089942,-0.1487710476,-0.0188667681,-0.0512376875,-0.3176402152,0.2477181852,0.0202709325,0.0034449229,-0.6579137444,-0.0308890101,-0.3640883863,-0.410166651,0.1023242176,0.1237966865,0.3612010479,0.1131152362,0.2846204937,-0.2655318081,-0.0806796551,-0.0837902278,0.2637239099,0.0705902278,-0.0821330845,-0.2368568778,0.0353909135,-0.3795093596,0.3788214922,-0.1252358854,0.1993347704,-0.0494945385,-0.0205659028,-0.0158610009,-0.160190627,0.3575387895,-0.3219276667,-0.0184350386,0.3051732481,-0.0723782331,-0.3048911095,0.0750038773,-0.2185464352,-0.0187089425,0.1801496446,0.1695200056,-0.1195449531,-0.3004764616,0.0167263858,0.4310730994,-0.2822128832,-0.1958411187,-0.222024858,-0.1508185863,-0.3353959024,0.0113308132,-0.1939946711,0.0724307597,-0.1042586938,0.1709124744,-0.2531684935,0.1641309857,-0.016891066,0.1121283397,0.1209591553,-0.2463969886,0.2492660582,0.2544650137,0.1075447947,0.3158960342,0.336989969,-0.0578443483,-0.2881180644,0.1851859838,-0.0216391962,0.3540468812,0.3804876208,-0.1084250435,-0.1297910362,0.1562836319,0.277231276,-0.4221793115,-0.1058231443,0.1520667076,0.1899351627,-0.1318843067,-0.3830702305,0.2421082407,0.1058469042,-0.0428113677,-0.0279097427,0.1405662596,-0.2162882835,0.165775463,-0.1204747185,0.7463241816,-0.0853946432,0.2261953205,0.2716045976,-0.01059891,0.4589013457,-0.2182703614,-0.0172342658,-0.0299437828,-0.2580789626,-0.0018667637,-0.0838018432,0.4105696082,0.0120970933,-0.101506032,0.5569252372,0.0023533541,0.2621752322,-0.3156482875,0.206828326,-0.0262680966,-0.2241601944,-0.5295377374,0.2227849364,0.111693196,0.0908092111,-0.1965965778,-0.2296914458,0.089205496,0.0733072758,0.0209829565,-0.0881580412,-0.0147264618,0.0910438821,-0.2471253723,-0.011324605,0.0023111124,0.1168786138,0.3662894964,0.1540874392,-0.3008590639,0.0407698527,-0.3916719556,0.1758881211,0.043192368,0.2353020012,0.1732371002,-0.2316584438,-0.2294980288,0.0711231455,-0.0932983607,-0.0229832288,0.44661358,-0.0534590594,-0.0299995374,-0.2270329148,0.1701972634,0.1872000545,0.1742009073,-0.1272588819,0.1690127701,0.2469747216,0.0646355599,0.3324843943,-0.118562609,-0.2234360427,-0.2319291681,0.3323866129,-0.1481749266,-0.2163781822,-0.0340101458,0.2127753645,-0.1911338121,-0.3022423387,0.2718286514,0.0050803022,-0.1282936931,0.234399274,0.3882580698,0.3762431145,-0.1151657403,0.2702775002,0.0319874138,-0.2423068881,-0.0731427148,-0.370336175,-0.2799700499,0.2077266574,0.4752341807,0.1235135943,0.1477679312,-0.0800835192,0.0471662097,-0.2568988204,-0.4148227274,0.0427973084,-0.0579886287,0.0490955599,0.2588075697,-0.1258111298,0.3629033267,-0.2199745029,0.0944314376,-0.2352956831,-0.3346167207,-0.2838246226,-0.0065519013,0.1733448058,-0.0442159623,-0.3215044141,-0.0092092017,-0.5947599411,-0.3334679902,-0.1524946094,0.2498431951,0.4137348235,-0.1798789948,-0.0268013272,-0.1973176599,0.4317834973,-0.1764412373,0.0949151292,0.0630059615,0.2877397835,-0.034239132,0.279807806,-0.0504023544,0.038233202,-0.0291264988,0.1064660996,0.1639818251,0.079293333,0.3910473883,-0.2540489435,0.1430637687,0.2583361566,0.5726851821,0.1912139505,-0.0539067648,0.1875460744,0.098394908,0.271559,-0.3790971041,0.060038127,-0.0015358452,0.1076111495,0.4134904146,0.0845966339,0.3291352391,0.1103388518,-0.0963395759,-0.0093084294,0.3304803967,-0.1119166166,0.223280713,0.1307686865,0.0439182855,0.141795665,0.0311193783,0.4378814995,-0.0079678781,0.3239933848,-0.3390367031,0.2351375073,0.1155501306,0.1127718315,-0.2816655934,-0.4408522844,0.2878967822,0.0305653717,-0.0485544093,-0.182086587,-0.2464107424,0.181925714,-0.5260383487,-0.0845423713,-0.1453720629,0.0936665833,-0.0773525685,0.0639610142,-0.4059838057,-0.3086449504,-0.0317532122,0.1227741316,0.077229917,-0.0736874714,-0.0024293466,0.0145098828,0.1101319268,-0.2059514523,0.1357246786,0.0530860573,0.0674321577,-0.0314456075,0.0315850228,0.2038819045,0.0759631619,0.1162949875,0.4558703005,0.325094223,0.0279592015,-0.1732877791,0.1791213155,-0.0174094439,-0.1257012337,-0.1085021645,0.0967774764,0.0271152724,0.1316803843,0.3657478094,0.1895966381,-0.2076842785,0.3444927931,-0.0761378333,-0.1504829824,-0.4568811357,0.2505191863,0.1846465319,-0.1169918552,-0.1064551771,-0.2874945402,-0.1889007986,-0.1083047837,0.3725979328,0.0194765814,0.1105332673,-0.1914427429,0.1280117333,0.101879932,0.4391471446,0.1733891517,0.2939247191,-0.2876727879,0.1159935668,-0.8045651913,-0.1493818462,-0.125921458,0.2573935091,-0.0518087484,0.1091890633,0.0875607282,0.4373385608,0.0744644329,0.1585405469,0.1727100909,-0.1801653653,-0.3003027141,-0.1347003281,-0.062439505,-0.1002140418,-0.0315278955,-0.3196187615,0.0919976309,-0.3633929789,0.0229150914,0.0673181713,0.20422481,-0.2061326206,-0.1297549754,0.4416780174,0.0433072075,0.395085454,0.1441107243,-0.1122508049,-0.052895885,0.037975058,-0.0639386475,0.1701036543,-0.0361089408,-0.0328885131,-0.1144766957,-0.141364187,-0.1595003456,0.2147668302,0.0632851869,-0.0111295348,-0.1094945967,-0.0361579955,0.077591002,0.1335782111,0.000827905,0.0445996784,-0.1451621205,0.2125579268,-0.2044009417,0.0314986482,0.3339629769,-0.034064088,0.0181125905,-0.3719019294,0.3834601045,0.1742333025,-0.006817848,-0.49317047,0.2285157591,0.3217993677,-0.1685982198,-0.1405643076,0.3053504825,-0.2599490583,-0.1895595193,0.0004531168,0.2614339888,-0.2595774233,-0.2974052727,0.0570863485,-0.1473935992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/824","title":"Discussion using datasets in offline mode","comments":"I opened a PR that allows to reload modules that have already been loaded once even if there's no internet.\r\n\r\nLet me know if you know other ways that can make the offline mode experience better. I'd be happy to add them :) \r\n\r\nI already note the \"freeze\" modules option, to prevent local modules updates. It would be a cool feature.\r\n\r\n----------\r\n\r\n> @mandubian's second bullet point suggests that there's a workaround allowing you to use your offline (custom?) dataset with `datasets`. Could you please elaborate on how that should look like?\r\n\r\nIndeed `load_dataset` allows to load remote dataset script (squad, glue, etc.) but also you own local ones.\r\nFor example if you have a dataset script at `.\/my_dataset\/my_dataset.py` then you can do\r\n```python\r\nload_dataset(\".\/my_dataset\")\r\n```\r\nand the dataset script will generate your dataset once and for all.\r\n\r\n----------\r\n\r\nAbout I'm looking into having `csv`, `json`, `text`, `pandas` dataset builders already included in the `datasets` package, so that they are available offline by default, as opposed to the other datasets that require the script to be downloaded.\r\ncf #1724 ","body":"`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n","comment_length":179,"text":"Discussion using datasets in offline mode\n`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n\nI opened a PR that allows to reload modules that have already been loaded once even if there's no internet.\r\n\r\nLet me know if you know other ways that can make the offline mode experience better. I'd be happy to add them :) \r\n\r\nI already note the \"freeze\" modules option, to prevent local modules updates. It would be a cool feature.\r\n\r\n----------\r\n\r\n> @mandubian's second bullet point suggests that there's a workaround allowing you to use your offline (custom?) dataset with `datasets`. Could you please elaborate on how that should look like?\r\n\r\nIndeed `load_dataset` allows to load remote dataset script (squad, glue, etc.) but also you own local ones.\r\nFor example if you have a dataset script at `.\/my_dataset\/my_dataset.py` then you can do\r\n```python\r\nload_dataset(\".\/my_dataset\")\r\n```\r\nand the dataset script will generate your dataset once and for all.\r\n\r\n----------\r\n\r\nAbout I'm looking into having `csv`, `json`, `text`, `pandas` dataset builders already included in the `datasets` package, so that they are available offline by default, as opposed to the other datasets that require the script to be downloaded.\r\ncf #1724 ","embeddings":[-0.471647799,0.2902263701,-0.0476719923,0.1334424019,0.2106889337,-0.2122251689,0.5858260393,0.0534165502,0.2833412886,0.1841130108,0.031059403,0.0326385237,0.0011814381,0.3333258331,-0.0721439496,-0.0585422292,-0.190542534,-0.005104905,-0.1658585221,0.0827916339,-0.0870347396,-0.1724681854,-0.4247154593,-0.1195316389,-0.0675603822,-0.1688151807,-0.0242116675,0.2116727829,-0.1926144809,-0.445373714,0.2666605413,0.2000291497,0.1570937186,0.422985673,-0.000104336,0.0591886006,0.0797744915,0.0126057863,-0.2829338312,-0.3505377769,-0.237617746,-0.4724885821,0.2454893738,-0.417991966,-0.1765559912,-0.0319712609,-0.1149641797,-0.4964652956,0.2549802959,0.3332003653,0.2521955967,0.5795013309,-0.1773819327,-0.2577656806,-0.0930708274,-0.0455761217,-0.2170029879,0.4130172431,0.1451813281,-0.0119967405,-0.0917475745,0.0427968465,-0.2056658864,0.0555355139,0.3426394463,0.0692112818,-0.101287663,-0.0757476985,0.1654934734,0.4096831083,0.4384514689,-0.2788181007,-0.2337877899,-0.3661537766,-0.0599909835,-0.4158058465,0.0579965152,0.0842632875,-0.0501417816,0.2703290284,-0.4178495407,-0.2782533765,-0.3631697297,0.1871719807,-0.2958251834,0.0800979808,-0.219046101,0.0571911149,-0.0345727429,0.1217101961,-0.1603948921,-0.287496984,0.1231968477,-0.055233486,-0.2458191365,-0.0483222455,0.1733665019,0.0262083262,0.1455963254,0.1075297371,0.1663131565,0.055470001,0.0069582094,0.0516370237,0.4685066044,0.2116460055,0.0110625401,-0.2097331285,0.5494056344,0.148972258,0.0090585183,-0.234075278,0.2657729685,-0.2524652481,0.0911977068,0.0700877458,0.3765212595,-0.32149297,-0.0229850113,0.0353144482,0.1597326398,-0.0621036999,0.1616376191,0.1747798473,-0.2847248614,0.0129403984,-0.0295246858,0.3609687686,-0.277133733,-0.3468526006,-0.0389768779,-0.2118643224,-0.1780509055,0.0920449644,0.2455183864,0.0309363343,0.0510609336,-0.0147727989,0.2424403429,-0.0368465111,0.3018389046,-0.0809502825,0.0659481734,0.2120589316,0.1196169853,-0.0765699595,0.124859646,-0.1560155302,0.0067118825,0.0666678101,-0.0514350384,-0.4386138618,0.2374287397,0.2237864435,-0.6274784207,-0.1871648133,-0.1969017088,0.0777899697,-0.3017685115,-0.1738761067,0.0214956682,0.075584963,-0.1690935194,-0.2382545173,0.2987488508,0.6010077,-0.5180035233,0.0819053575,0.04580345,-0.1816399693,-0.1889690161,0.0283215363,-0.3945119679,0.5864564776,0.003940382,-0.3889356256,0.5678287745,-0.3095942438,-0.3093517423,0.2875798941,-0.1802513003,-0.0595034733,-0.0382034406,0.1196887791,0.4773570895,0.0809495598,-0.0078142164,0.2666749954,-0.0770242438,-0.0095415451,-0.2956813872,-0.2896500826,0.0985429436,0.296525687,0.2704698741,0.1280689538,0.3454582393,-0.236692667,0.0276600327,-0.1362006068,0.1363507509,0.0990994647,0.2330277115,0.1239424273,-0.0695232451,-0.1321175992,-0.5115550756,0.32426247,0.0397507697,0.1621360332,-0.2056600153,-0.1533112377,-0.0401652306,-0.2169201821,-0.1385666728,-0.0377867445,0.1234516278,-0.0895907059,0.0869746506,0.1763064712,-0.2977676988,0.135161221,-0.0182938278,-0.0051473621,-0.1500218958,0.3400147259,-0.0314372331,0.0565065779,-0.1015016213,0.0904149786,0.0123003358,-0.1740998775,-0.1055455878,0.483792752,0.0415284075,0.1183546484,0.2707455456,0.4232814312,0.3186860383,0.0245525986,0.560336709,-0.1271901876,0.3274110556,-0.1185178757,-0.3417674303,0.7027890682,-0.1783438474,0.272815913,0.1150313839,-0.0340052322,0.2038816661,-0.0940221325,-0.4044412374,-0.0952030122,0.0663549379,0.0030854121,0.1993399262,0.0168769918,-0.2931219935,0.0504621863,0.3374177814,0.0135319028,0.1364228576,-0.0604639389,-0.0171772875,0.1130490005,-0.0254898518,0.137879774,0.513273716,0.2308206111,0.0430370569,0.3077920377,0.0352877602,-0.2593407035,0.1188974679,0.1687578112,-0.0185087677,0.1212383509,0.0963387787,-0.0117987394,-0.1399939209,-0.0319385789,0.003812724,0.05903548,-0.3146699965,0.2902598977,-0.3545603156,-0.027913373,0.0130221266,0.1309276223,-0.4807070792,-0.2102815807,0.1170057431,0.0569216907,-0.2447342277,0.171292603,-0.0877068639,0.4514835179,-0.2693450451,-0.2608876526,0.2427781075,-0.5937603116,-0.0655437261,0.0848006383,0.3401410282,0.1822122484,0.565823853,0.0593990535,-0.1411339194,-0.2710398138,-0.1447022408,0.0892937556,-0.0116074029,0.3152152598,0.1083986461,0.0203317963,0.027070595,0.4002498984,0.385861218,-0.2979997694,0.1902170032,0.1272397041,-0.0945452154,-0.0322720334,-0.3383330703,-0.4963891208,-0.399233371,-0.467761457,0.1700818539,-0.0527774207,0.010702298,0.229778707,0.2038754672,0.1109455898,-0.0705158859,0.1104866639,-0.1722420603,-0.4613300562,0.145051986,-0.1792815924,-0.4937444925,0.253762275,0.2192489356,0.1917006224,0.3427464068,-0.5899965167,-0.3177880943,0.1043311656,0.3914316297,0.1040238962,0.2679328918,0.3933205605,-0.0794315711,-0.0585553907,-0.0903849155,0.0574980229,0.2076742053,0.0564016998,-0.1091034859,0.149455905,0.2685927749,0.0167623274,0.5026164651,-0.0798343346,0.109816812,0.2587999701,-0.1174223348,0.7056586742,-0.1825235933,-0.1671553105,-0.0413852073,0.0148815038,-0.324709475,0.1728413254,-0.017982034,-0.015906956,-0.6336125731,-0.0798164308,-0.2484522909,-0.4550782144,0.1834675372,0.1597409397,0.3247781396,0.1058614105,0.2193622291,-0.270796597,-0.0256444104,-0.0919202268,0.2632635832,-0.0149521809,-0.0774901137,-0.2507244349,0.1369996965,-0.3115711808,0.2955721915,-0.1352505237,0.2162646949,-0.0439071096,-0.0313427635,0.0826732889,-0.0391843058,0.2991420925,-0.4082177579,-0.0943692774,0.3293497562,-0.0216435902,-0.3531905413,0.0764372274,-0.1722090989,0.001943219,0.1599160284,0.1581934988,-0.0679426342,-0.3483632505,-0.0060396907,0.3597085178,-0.2774301469,-0.2177672088,-0.2001278698,-0.0423309393,-0.4275080264,-0.0052138278,-0.2545889616,-0.0984356925,-0.0803567842,0.1182883829,-0.1544917673,0.2191759646,-0.1380103678,0.0739968941,0.1102335379,-0.3203080297,0.2273410559,0.2977680564,0.0553014316,0.3309388459,0.3506483138,-0.118118152,-0.1482797861,0.2086242735,-0.0433464833,0.4113816321,0.4854654074,-0.0095938276,-0.084601678,0.1447081268,0.2892356217,-0.4037901759,-0.1615210921,0.1078801081,0.0592983514,-0.1000850946,-0.4673404396,0.1834896952,-0.0405343883,-0.0738386586,-0.003525208,0.129417181,-0.1851621419,0.241601482,-0.0919818804,0.8122492433,0.0263648778,0.2219541073,0.1739808619,-0.0078909574,0.4734134972,-0.3028861284,-0.0455839634,-0.0602652282,-0.0436436422,0.0182211399,-0.0760406628,0.4410826862,0.0106894718,-0.2054695338,0.493493259,-0.0749406517,0.2824131548,-0.3375301957,0.1076018065,0.0202240534,-0.226479426,-0.5203982592,0.2028240561,0.1081517041,0.1028201804,-0.1851934046,-0.2435837388,0.0342530869,0.14930363,0.009566362,-0.1898495555,-0.0526054613,0.0308099072,-0.2059034705,0.0560679063,0.0032336819,0.0450264774,0.3648864329,0.1468626708,-0.3398707509,0.0138868121,-0.3556911051,0.1691973358,0.0360772088,0.1752617806,0.1566658765,-0.198594898,-0.334464848,0.0011104419,-0.1116639972,-0.0255522933,0.4310037494,-0.0548414439,0.0942337736,-0.2805888951,0.1767488271,0.2303110808,0.2608217895,-0.0935614482,0.1746456772,0.2110819817,0.0750605166,0.388412267,-0.1253615171,-0.2212532014,-0.2144706845,0.3347358108,-0.1247460991,-0.1553388685,-0.0833659321,0.2226356715,-0.1695810258,-0.3365006447,0.3025309145,-0.0054206541,-0.2230083346,0.198640734,0.350471735,0.2963169515,-0.0260181986,0.2046554238,0.010253449,-0.188002646,-0.1679405272,-0.275377512,-0.3178842962,0.1660073996,0.4533553123,0.1563894898,0.0473574474,-0.0970104113,-0.0148069076,-0.2212059796,-0.4241659045,0.0660497919,-0.0683547705,0.0864292756,0.2857186198,-0.1539194435,0.3005248308,-0.2875233293,0.07204853,-0.2119683027,-0.427126646,-0.2759729028,-0.0565468296,0.1780339777,-0.037533544,-0.3255496621,-0.0270778779,-0.5326864719,-0.3123993576,-0.1282408237,0.2727524638,0.440995425,-0.129640609,0.0053064087,-0.2117775977,0.5039588809,-0.1466331184,0.1623123437,0.0296293125,0.3317891359,0.0089618471,0.203832075,-0.0991154462,0.076196909,0.0818035007,0.1574381292,0.2122733891,0.1228736266,0.4452152848,-0.1462512314,0.1563674062,0.2498918921,0.5009359121,0.1629018635,-0.0704720095,0.1216586828,0.2080479711,0.2638338506,-0.3340935111,0.0065753055,0.0738723353,0.1430096924,0.3401826322,0.1444491595,0.4526937604,0.095358707,0.0051787593,0.0126380492,0.3435204923,-0.0753496513,0.223722294,0.0738916397,0.0469854102,0.0596651919,-0.0378163308,0.3901125193,0.0471249372,0.2782624066,-0.3217486143,0.2200015634,0.0860395506,0.1509748101,-0.2448892444,-0.4754575789,0.4114567935,0.0712680668,-0.101576224,-0.1877399981,-0.2861892283,0.2244449705,-0.3701642454,-0.0536599606,-0.0775001645,0.0845382065,-0.0699018091,0.0885946602,-0.3811097443,-0.3510326743,-0.073594071,0.1183509603,0.0528446287,-0.024111243,0.027261626,-0.034705922,0.0938971341,-0.3102188706,0.1447045803,0.0402173996,0.0982588902,-0.0307618193,0.0310811531,0.1860680133,0.0218679663,0.140045166,0.4324544966,0.2805573046,0.0885334164,-0.1821982712,0.1789683253,0.0096463151,-0.0614575334,-0.1308431476,0.0728377178,0.0065968642,0.1147893816,0.3294236064,0.1586290896,-0.1999017596,0.3588386178,-0.1305038482,-0.2217881829,-0.4970718622,0.239600271,0.2392929494,-0.1237450317,-0.0584078692,-0.2656486928,-0.2500069737,-0.2632139027,0.4473289251,-0.009919527,0.1495080888,-0.2325070351,0.1104773805,-0.0039669094,0.4872244895,0.1359924376,0.2626902461,-0.2467202246,0.1591033489,-0.8487069607,-0.1010251194,-0.1350577772,0.2625231445,-0.0060989154,0.0703933612,0.1294197887,0.497467041,0.0732407272,0.1348395944,0.1989409477,-0.1726145595,-0.3112728596,-0.1594673693,-0.0758786127,-0.0788231269,-0.0683218241,-0.2402743995,0.0617132336,-0.3475409746,-0.0013597806,0.0275617987,0.1004274711,-0.0879153013,-0.0120245488,0.5188177824,0.0286140516,0.3453910053,0.3027181625,-0.1201030314,-0.1079946533,0.0716166794,-0.023258131,0.0583465472,-0.0380428508,-0.0517743118,-0.1984514147,-0.2123390585,-0.2314856797,0.2240794897,0.0803294182,0.023162175,-0.0777491182,-0.0508500077,0.0900446922,0.1569570899,-0.0631006435,0.0342967696,-0.0961916223,0.2404280007,-0.2528859079,0.0804953873,0.3342810869,0.0179550536,0.0175155345,-0.3761808872,0.3429668546,0.2321786433,0.0176304244,-0.5226225853,0.0929991752,0.3145461977,-0.1564721614,-0.0433372706,0.3258460462,-0.2541510761,-0.1967098266,0.0114058265,0.2230953723,-0.2774820626,-0.2227725387,0.0650163442,-0.1179777831]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/824","title":"Discussion using datasets in offline mode","comments":"The local dataset builders (csv, text , json and pandas) are now part of the `datasets` package since #1726 :)\r\nYou can now use them offline\r\n```python\r\ndatasets = load_dataset('text', data_files=data_files)\r\n```\r\n\r\nWe'll do a new release soon","body":"`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n","comment_length":38,"text":"Discussion using datasets in offline mode\n`datasets.load_dataset(\"csv\", ...)` breaks if you have no connection (There is already this issue https:\/\/github.com\/huggingface\/datasets\/issues\/761 about it). It seems to be the same for metrics too.\r\n\r\nI create this ticket to discuss a bit and gather what you have in mind or other propositions.\r\n\r\nHere are some points to open discussion:\r\n- if you want to prepare your code\/datasets on your machine (having internet connexion) but run it on another offline machine (not having internet connexion), it won't work as is, even if you have all files locally on this machine.\r\n- AFAIK, you can make it work if you manually put the python files (csv.py for example) on this offline machine and change your code to `datasets.load_dataset(\"MY_PATH\/csv.py\", ...)`. But it would be much better if you could run ths same code without modification if files are available locally.\r\n- I've also been considering the requirement of downloading Python code and execute on your machine to use datasets. This can be an issue in a professional context. Downloading a CSV\/H5 file is acceptable, downloading an executable script can open many security issues. We certainly need a mechanism to at least \"freeze\" the dataset code you retrieved once so that you can review it if you want and then be sure you use this one everywhere and not a version dowloaded from internet.\r\n \r\nWDYT? (thks)\r\n\r\n\nThe local dataset builders (csv, text , json and pandas) are now part of the `datasets` package since #1726 :)\r\nYou can now use them offline\r\n```python\r\ndatasets = load_dataset('text', data_files=data_files)\r\n```\r\n\r\nWe'll do a new release soon","embeddings":[-0.4490848482,0.2095090002,-0.0598193891,0.129349485,0.2621940374,-0.1312862486,0.5469678044,0.0949262381,0.3154391944,0.2294326127,0.0510399193,-0.0031574785,0.0479425304,0.4036746323,-0.1032212973,-0.1191765517,-0.1096814498,0.0806255639,-0.170926407,0.0925025642,-0.1816644222,-0.0897327736,-0.3750758767,-0.0255609434,-0.104983367,-0.1746587455,-0.1071397066,0.1566956937,-0.2103743553,-0.3327201605,0.2491134703,0.1175882071,0.1634298414,0.4241681695,-0.0000999822,0.0421361551,0.1069040447,0.0379941911,-0.2203476429,-0.419916749,-0.1981761009,-0.4793075919,0.2418385744,-0.40180251,-0.1757818609,-0.1275016814,-0.0598839149,-0.4807877541,0.2974709272,0.3426465392,0.2918207049,0.5356094241,-0.08034008,-0.2813901603,-0.1216623858,-0.0586182959,-0.1778581738,0.4252844453,0.0733571202,-0.0123791518,-0.0838829204,0.0888442323,-0.1531640589,0.0956401974,0.3111796975,0.0713066906,-0.0924655721,-0.1528321356,0.2028551847,0.3944026828,0.4603175521,-0.3109174371,-0.2949747741,-0.3411391079,-0.0479014292,-0.4361881316,0.1274563819,0.1441726834,0.0090271737,0.233983919,-0.3632723987,-0.2975689471,-0.3418593109,0.169493258,-0.2816976607,0.0181066487,-0.2516751885,0.0413689315,-0.0832428709,0.0752860978,-0.1551486254,-0.3205278516,0.098621361,-0.0793635026,-0.1888277978,-0.0165316593,0.1534282714,0.008187145,0.1217974946,0.0751493126,0.1525371522,-0.0060451161,0.0926281363,0.0680228099,0.4427141249,0.3181465268,0.0188554749,-0.2447724193,0.5638321638,0.0607491359,-0.0046479688,-0.2882612944,0.3297551274,-0.2873273194,0.0361570455,-0.016085932,0.3358884156,-0.2712710202,-0.0916830376,0.0604369603,0.1495210975,-0.0821654052,0.1536692828,0.2071387023,-0.2203721106,0.0295776036,0.0346205868,0.3958796859,-0.2415048033,-0.2347968221,-0.1089479625,-0.1066183075,-0.1055866778,0.0744700357,0.1888174713,-0.0131090106,0.1214088798,-0.0402394049,0.2105225325,-0.0341542251,0.278637141,-0.1574495286,0.0479996912,0.2500382364,0.1872593313,-0.106234327,0.0567811839,-0.2355721295,0.0381199308,-0.0035781907,-0.0436274111,-0.404987365,0.0963111818,0.2798317969,-0.5623047352,-0.1680555195,-0.2548446953,0.1297697127,-0.2823038101,-0.0971270129,0.0206746962,-0.0097434577,-0.1449509412,-0.1983079612,0.239082709,0.5462014675,-0.4452953041,0.0876879916,0.1217823178,-0.1871830672,-0.17599307,0.0642042682,-0.343101263,0.5343926549,0.051759094,-0.2778843641,0.5229828954,-0.3582603335,-0.2345213741,0.2888397574,-0.1610287726,-0.1441398114,-0.0667064488,0.0498279259,0.3899838328,0.1192973927,0.0001437401,0.2430983931,-0.0564485043,0.0195687898,-0.2536431253,-0.323527962,0.14072375,0.2494011074,0.3065206707,0.1239916235,0.2792335451,-0.2578680515,-0.0405986905,-0.2623461485,0.165442884,0.1579164416,0.2099319845,0.1023168489,-0.0483671613,-0.2059078664,-0.4760749638,0.2252617329,0.0337006971,0.2320879996,-0.230220288,-0.1496614814,-0.008372074,-0.179708451,-0.1597651243,-0.0425286368,0.1691499352,-0.1388658583,0.1849676222,0.1760512888,-0.3189759254,0.1439083815,-0.0562663488,0.0259863827,-0.1155401766,0.2796734273,-0.0809465349,0.0232010745,0.0426407829,0.0838658288,-0.0070740283,-0.1859913468,-0.1073740348,0.4920392036,-0.0600119606,0.050861761,0.315263927,0.361379236,0.3560869098,0.0397625342,0.5095092654,-0.1218395978,0.3111970127,-0.0770422667,-0.3356893063,0.6987401247,-0.1456841528,0.2038489729,0.0956835374,0.0029137544,0.3142004907,-0.0028089855,-0.3935572803,-0.0653379932,0.0797696859,-0.044440072,0.2148961127,-0.0678070262,-0.3889352679,-0.0229981318,0.3756242096,-0.0001716741,0.1845305413,-0.0001421404,0.0194953308,0.062744543,-0.0596937574,0.0841197968,0.5141689777,0.2386389077,0.0379683487,0.2522847354,-0.0703066364,-0.3152319491,0.1395313442,0.2154895812,0.0092474939,0.0789768994,0.089975588,-0.003388139,-0.3099704087,-0.0535693206,-0.0292135775,0.0814710259,-0.2326799631,0.2459449321,-0.3154218197,-0.0520201139,0.0422418192,0.111955598,-0.4437055588,-0.2312417775,0.0868251696,0.0358285643,-0.2842279077,0.1475006491,-0.0897418633,0.4524809718,-0.2049701959,-0.138762638,0.1279346496,-0.5095137954,-0.1033318937,0.133870855,0.3391273618,0.1479579806,0.4992528558,0.0086958241,-0.0772898719,-0.2907142639,-0.1874965131,0.0661994219,-0.0880849659,0.3189185262,0.0668629855,0.0481230058,-0.0487571284,0.2254338264,0.4193473756,-0.3061971366,0.1707129776,0.085586451,-0.0520983674,-0.1553375125,-0.3459178209,-0.487205416,-0.4130060375,-0.4035458863,0.2704867423,0.0267385636,-0.0292487256,0.1243112534,0.158038646,0.1001027748,-0.0176117718,0.2050555199,-0.2188558429,-0.4933129549,0.1240409464,-0.2316770852,-0.5781071782,0.2055875212,0.1968098283,0.1813315898,0.2394980788,-0.4866481721,-0.374794066,0.0422560088,0.3973865807,0.1870754659,0.3177512586,0.3660067022,-0.0246600416,-0.076631479,-0.0784322396,0.0246607829,0.2000732422,0.0979012176,-0.1254032403,0.0741901919,0.3168487847,-0.0595339797,0.4317931533,-0.0111093363,0.0496152043,0.2642003894,-0.1114155203,0.6812868118,-0.2694692016,-0.1440427005,0.0132880118,-0.0587873459,-0.2512556314,0.2106068581,0.0157646891,-0.0362582467,-0.6542541385,-0.0097049242,-0.403083384,-0.3856451511,0.0840909854,0.1315927356,0.3249271214,0.1220792904,0.2325794697,-0.2801973224,0.007008655,-0.0742462799,0.2477307767,0.06175901,-0.0998949334,-0.2681442499,0.009503277,-0.3775528371,0.336648196,-0.1258386672,0.2580756843,-0.0771125928,-0.0305225961,0.0071699228,-0.1293199062,0.3594410419,-0.3762827218,-0.1614654362,0.2891313434,-0.0968559235,-0.2732861936,0.0844901577,-0.1972345114,0.0496110693,0.1447974443,0.2760510445,-0.0842916593,-0.3075995743,0.0610624291,0.3504799008,-0.2497546226,-0.1882384717,-0.1717914045,-0.1545330435,-0.3299390376,0.0528551564,-0.2317132801,0.0491187498,-0.0869815275,0.1136053428,-0.1949661523,0.1558368653,-0.0642601103,0.0974733755,0.0733715743,-0.2716390789,0.227278918,0.2307690233,0.1161052585,0.3351896107,0.3377739787,-0.123373732,-0.2471155375,0.1811089069,-0.0564749502,0.36204198,0.4465497136,-0.0729906633,-0.097519964,0.1527477503,0.2758108675,-0.4200819135,-0.1456189007,0.2055243254,0.1781510711,-0.0686922893,-0.3760152757,0.2442426682,0.1129909828,-0.0533210784,0.0150246732,0.123032473,-0.1884620935,0.2214028686,-0.0593033656,0.7759805322,-0.0680378303,0.1573288292,0.228532061,-0.04602639,0.4807931781,-0.3189853132,0.0172787793,-0.0401388332,-0.1941192299,0.0036332707,-0.1070229858,0.4337182343,-0.0415739231,-0.0733070523,0.5114344954,0.0493919328,0.3181025982,-0.3153334856,0.2438227534,-0.0738768727,-0.2079118639,-0.5192394257,0.2445125729,0.1184825748,0.0833035335,-0.2140755206,-0.2424930036,0.0730268881,0.0449495874,-0.0016644809,-0.1683880389,0.0244122129,0.0900919437,-0.1606284231,-0.0268593617,-0.010151092,0.1060651988,0.4375319481,0.136662215,-0.3299772143,0.0077851443,-0.3949714303,0.157768622,0.0260520019,0.2085138261,0.2068130821,-0.2194934636,-0.234115079,0.0194601081,-0.1111916676,-0.043162968,0.3716807365,-0.0979472324,-0.0760398433,-0.2462724596,0.1823972017,0.2029304206,0.1959369332,-0.1174157038,0.201196298,0.2146023512,0.1222187579,0.3525808454,-0.1376956701,-0.2014677674,-0.2937906384,0.3205314577,-0.0845624432,-0.1777475625,-0.0722320378,0.2106570303,-0.2353446931,-0.3138158619,0.2293109298,0.0463183448,-0.1134389266,0.2572028637,0.3315884471,0.2462130487,-0.1465516686,0.2690930963,0.0718549639,-0.1856536567,-0.0777779296,-0.3274802566,-0.3413894176,0.2560734153,0.4513781667,0.155822888,0.1147615016,-0.0552717224,0.0866375118,-0.197172761,-0.4408754706,0.0679855272,-0.1029251963,0.0660291687,0.2493936867,-0.1730806381,0.3766629398,-0.1902855635,0.1050744727,-0.1715284884,-0.3376734853,-0.3105429113,0.040069446,0.1459278464,-0.0123596154,-0.2967282832,-0.0252775047,-0.55437392,-0.3203285635,-0.1349170655,0.2085369974,0.4151882231,-0.1913521439,0.0276035368,-0.186750263,0.3740739822,-0.0389068611,0.0819933936,0.0325910673,0.3030642867,-0.0030833385,0.2433012277,-0.0295130871,0.0430576727,0.0253942627,0.107260935,0.0851695836,0.1314177215,0.415476799,-0.2244153023,0.1076865196,0.2965520918,0.5483056903,0.2145422846,-0.086429283,0.1892021447,0.1614557356,0.3001289666,-0.3635789156,0.0032816371,0.0243402701,0.13692379,0.3610396981,0.0615084581,0.4126155078,0.0223261081,-0.0799297616,0.0220958758,0.3285107613,-0.1095594987,0.2245313376,0.1235160381,0.0635393932,0.1075255424,0.0822855756,0.4752013385,0.0161640663,0.3380511403,-0.2761106193,0.1901878715,0.0358073674,0.1610104293,-0.1996608078,-0.3803336322,0.3735830188,0.0126131345,-0.0909559578,-0.2275025547,-0.227736026,0.2517718971,-0.450640142,-0.0610096119,-0.1715638489,0.1333201528,-0.1218138188,0.0248794165,-0.397905916,-0.3176402748,0.0156230135,0.1247259527,0.0711785108,-0.0480278656,0.0062527247,-0.0422267914,0.0772870183,-0.1869489998,0.1966571063,-0.0445921905,0.096211195,-0.0109037282,0.0259179585,0.2208557278,-0.0055067185,0.0531305745,0.4975852668,0.3380316794,0.0415282734,-0.2303503007,0.1456147581,-0.0522004478,-0.1474351436,-0.1049373001,0.1791226864,0.0373310484,0.1159182414,0.3794879317,0.2081171572,-0.2256062925,0.3958656192,-0.1164210066,-0.115716733,-0.428954035,0.236761272,0.2151343971,-0.1547234505,-0.0418341868,-0.3053732216,-0.1934447289,-0.1661620438,0.4117180407,-0.000499127,0.1332321167,-0.2520887852,0.149787426,0.1007194966,0.4505885541,0.1846740246,0.2423756272,-0.2392034382,0.0792959109,-0.6985385418,-0.1670776159,-0.1281275749,0.3263763189,-0.0735832378,0.1057553142,0.1115700305,0.4124391079,-0.0063100792,0.1760124266,0.0596971326,-0.2131690383,-0.3156398535,-0.0660587549,-0.1222000718,-0.1291182786,0.0074628936,-0.227692917,0.1430811137,-0.3369763196,0.0381980278,0.0014569109,0.2213789821,-0.1770854592,-0.1268351227,0.4185497463,0.1040790156,0.3364282846,0.1634527147,-0.0826330334,-0.0467543453,-0.0029851513,-0.0648404509,0.1943533868,-0.0771029964,-0.0347661339,-0.16950351,-0.174856931,-0.1651425958,0.1906249523,0.040541403,0.0237386394,-0.1428376138,-0.049061764,0.0587364808,0.0616846085,0.0528427511,0.0825138688,-0.1166033447,0.2132470459,-0.1673659533,0.0156505313,0.3196119368,-0.0693766549,0.0099102575,-0.3681201637,0.2837967277,0.2504253387,-0.0342554003,-0.4602249563,0.1561269015,0.2961423695,-0.1832142919,-0.1591356695,0.2970778048,-0.2199114263,-0.18535842,-0.0506612174,0.1897984892,-0.2244891077,-0.2536653876,0.0138357617,-0.1592106521]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/823","title":"how processing in batch works in datasets ","comments":"Hi I don\u2019t think this is a request for a dataset like you labeled it.\r\n\r\nI also think this would be better suited for the forum at https:\/\/discuss.huggingface.co. we try to keep the issue for the repo for bug reports and new features\/dataset requests and have usage questions discussed on the forum. Thanks.","body":"Hi,\r\nI need to process my datasets before it is passed to dataloader in batch, \r\nhere is my codes \r\n\r\n```\r\nclass AbstractTask(ABC):\r\n task_name: str = NotImplemented\r\n preprocessor: Callable = NotImplemented\r\n split_to_data_split: Mapping[str, str] = NotImplemented\r\n tokenizer: Callable = NotImplemented\r\n max_source_length: str = NotImplemented\r\n max_target_length: str = NotImplemented\r\n # TODO: should not be a task item, but cannot see other ways.\r\n tpu_num_cores: int = None\r\n\r\n # The arguments set are for all tasks and needs to be kept common.\r\n def __init__(self, config):\r\n self.max_source_length = config['max_source_length']\r\n self.max_target_length = config['max_target_length']\r\n self.tokenizer = config['tokenizer']\r\n self.tpu_num_cores = config['tpu_num_cores']\r\n\r\n def _encode(self, batch) -> Dict[str, torch.Tensor]:\r\n batch_encoding = self.tokenizer.prepare_seq2seq_batch(\r\n [x[\"src_texts\"] for x in batch],\r\n tgt_texts=[x[\"tgt_texts\"] for x in batch],\r\n max_length=self.max_source_length,\r\n max_target_length=self.max_target_length,\r\n padding=\"max_length\" if self.tpu_num_cores is not None else \"longest\", # TPU hack\r\n return_tensors=\"pt\"\r\n )\r\n return batch_encoding.data\r\n\r\n\r\n def data_split(self, split):\r\n return self.split_to_data_split[split]\r\n\r\n def get_dataset(self, split, n_obs=None):\r\n split = self.data_split(split)\r\n if n_obs is not None:\r\n split = split+\"[:{}]\".format(n_obs)\r\n dataset = load_dataset(self.task_name, split=split)\r\n dataset = dataset.map(self.preprocessor, remove_columns=dataset.column_names)\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n dataset.set_format(type=\"torch\", columns=['input_ids', 'token_type_ids', 'attention_mask', 'label'])\r\n return dataset\r\n\r\n```\r\n\r\nI call it like \r\n\r\n`AutoTask.get(task, train_dataset_config).get_dataset(split=\"train\", n_obs=data_args.n_train) \r\n`\r\n\r\nThis gives the following error, to me because the data inside the dataset = dataset.map(lambda batch: self._encode(batch), batched=True) is not processed in batch, could you tell me how I can process dataset in batch inside my function? thanks \r\n\r\n File \"finetune_multitask_trainer.py\", line 192, in main\r\n if training_args.do_train else None\r\n File \"finetune_multitask_trainer.py\", line 191, in <dictcomp>\r\n split=\"train\", n_obs=data_args.n_train) for task in data_args.task}\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in get_dataset\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1236, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1207, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in <lambda>\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in _encode\r\n [x[\"src_texts\"] for x in batch],\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in <listcomp>\r\n [x[\"src_texts\"] for x in batch],\r\nTypeError: string indices must be integers\r\n\r\n","comment_length":53,"text":"how processing in batch works in datasets \nHi,\r\nI need to process my datasets before it is passed to dataloader in batch, \r\nhere is my codes \r\n\r\n```\r\nclass AbstractTask(ABC):\r\n task_name: str = NotImplemented\r\n preprocessor: Callable = NotImplemented\r\n split_to_data_split: Mapping[str, str] = NotImplemented\r\n tokenizer: Callable = NotImplemented\r\n max_source_length: str = NotImplemented\r\n max_target_length: str = NotImplemented\r\n # TODO: should not be a task item, but cannot see other ways.\r\n tpu_num_cores: int = None\r\n\r\n # The arguments set are for all tasks and needs to be kept common.\r\n def __init__(self, config):\r\n self.max_source_length = config['max_source_length']\r\n self.max_target_length = config['max_target_length']\r\n self.tokenizer = config['tokenizer']\r\n self.tpu_num_cores = config['tpu_num_cores']\r\n\r\n def _encode(self, batch) -> Dict[str, torch.Tensor]:\r\n batch_encoding = self.tokenizer.prepare_seq2seq_batch(\r\n [x[\"src_texts\"] for x in batch],\r\n tgt_texts=[x[\"tgt_texts\"] for x in batch],\r\n max_length=self.max_source_length,\r\n max_target_length=self.max_target_length,\r\n padding=\"max_length\" if self.tpu_num_cores is not None else \"longest\", # TPU hack\r\n return_tensors=\"pt\"\r\n )\r\n return batch_encoding.data\r\n\r\n\r\n def data_split(self, split):\r\n return self.split_to_data_split[split]\r\n\r\n def get_dataset(self, split, n_obs=None):\r\n split = self.data_split(split)\r\n if n_obs is not None:\r\n split = split+\"[:{}]\".format(n_obs)\r\n dataset = load_dataset(self.task_name, split=split)\r\n dataset = dataset.map(self.preprocessor, remove_columns=dataset.column_names)\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n dataset.set_format(type=\"torch\", columns=['input_ids', 'token_type_ids', 'attention_mask', 'label'])\r\n return dataset\r\n\r\n```\r\n\r\nI call it like \r\n\r\n`AutoTask.get(task, train_dataset_config).get_dataset(split=\"train\", n_obs=data_args.n_train) \r\n`\r\n\r\nThis gives the following error, to me because the data inside the dataset = dataset.map(lambda batch: self._encode(batch), batched=True) is not processed in batch, could you tell me how I can process dataset in batch inside my function? thanks \r\n\r\n File \"finetune_multitask_trainer.py\", line 192, in main\r\n if training_args.do_train else None\r\n File \"finetune_multitask_trainer.py\", line 191, in <dictcomp>\r\n split=\"train\", n_obs=data_args.n_train) for task in data_args.task}\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in get_dataset\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1236, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1207, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in <lambda>\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in _encode\r\n [x[\"src_texts\"] for x in batch],\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in <listcomp>\r\n [x[\"src_texts\"] for x in batch],\r\nTypeError: string indices must be integers\r\n\r\n\nHi I don\u2019t think this is a request for a dataset like you labeled it.\r\n\r\nI also think this would be better suited for the forum at https:\/\/discuss.huggingface.co. we try to keep the issue for the repo for bug reports and new features\/dataset requests and have usage questions discussed on the forum. Thanks.","embeddings":[-0.5023794174,-0.1743950695,-0.229502216,0.1301572621,0.1780355573,0.0317946561,0.2955617607,0.1713047624,-0.1484707892,0.1780768037,0.0091214189,0.1247809753,0.0641202703,0.2456238419,-0.0561650358,-0.0358917713,-0.0745975822,0.0483772866,0.1096090525,0.0072343689,-0.3676092625,0.0689976066,-0.404307127,0.0522343665,-0.2230468243,-0.2693564892,-0.061245136,-0.0392570309,-0.0044130813,-0.1525889933,-0.0022256428,0.2334949821,0.2088395506,0.4326501787,-0.0001049565,0.0785311237,0.246775642,0.0393823534,-0.2903354466,0.2226282358,-0.0494002365,-0.1695528775,-0.106298469,-0.3042087555,-0.100836426,-0.0860008001,-0.1385592818,-0.3374301493,0.3848311007,0.1753834486,0.2844721079,0.2802553177,-0.1476055533,0.141417712,0.0610040873,0.071453087,0.085112974,-0.3902092874,0.0682323202,0.147927478,-0.0484061763,0.149620384,-0.2720299363,0.0859410837,-0.2550970316,0.0218879282,0.0260984451,-0.3847026229,0.0884854123,-0.1349758655,0.2846595347,-0.2123625129,-0.244074598,-0.3842577636,-0.2724657655,-0.4653172493,-0.0269787684,0.1131320298,-0.0684819147,0.1012712568,-0.0785877854,0.2408464402,-0.0737230629,-0.1666946113,0.1114271581,0.256811142,-0.2119238675,0.2198818326,0.0700407103,0.0887886882,0.1648149788,-0.2350720912,-0.0004340044,0.2723601758,-0.490796864,-0.0871194154,-0.111774452,-0.1519913375,-0.1408956349,0.0541603342,-0.1284128726,0.1272303164,0.3161275685,0.1564184427,-0.0295962766,0.1643566489,0.1187213957,0.2562070191,-0.077118583,-0.3050593734,-0.2260431051,0.0277573764,-0.0785889626,-0.3298466206,0.1471393257,0.3610413373,0.2499244362,0.0302780811,-0.4871647358,-0.090392217,-0.0190420989,-0.0580975227,0.1420931816,0.2845916152,-0.2407628298,0.2508462965,-0.0810260475,-0.1523166597,-0.2926350832,-0.0461959541,-0.3027382791,-0.1152567863,-0.1496196091,0.0702788755,0.1349772662,-0.1721600443,0.3239180446,0.0755829066,0.0962865055,-0.1746606827,0.5014520288,-0.3949423134,0.2224124223,0.2599134743,0.1795815676,0.2426499426,0.0532012805,0.2097658068,0.0399616584,0.07378117,-0.6257678866,-0.1791180819,0.2474631667,0.3138492405,-0.050551679,0.0014612001,-0.5935842991,-0.0869221464,0.3213269711,0.0906271413,0.1197098047,-0.1754258871,-0.1859953701,-0.1946797967,0.2003967762,0.0852582231,-0.4504095614,0.067932725,-0.1153643429,-0.3227443099,0.5454286933,0.1242681667,-0.1138667688,0.6199551225,-0.3095553815,0.378302753,0.1937190294,-0.318939209,-0.2815083861,0.5390690565,-0.2575122714,0.0180868749,0.0398210399,0.0325068124,0.558662951,-0.0602873936,0.0362305008,0.3787432015,-0.1192286387,0.0902983919,-0.0995698944,0.1156369671,0.0859533623,-0.017408317,0.103893064,0.0438636467,-0.2167430967,-0.0445535667,0.1737817377,0.0534989499,0.2965995669,0.2496741712,0.0114939697,0.2087981552,0.1932462901,-0.045412153,-0.2964651883,0.0957323313,-0.0203324519,-0.1832336485,0.0405643061,-0.3565836847,-0.1427385509,0.1464838088,-0.4080851972,-0.170314461,0.2187933773,0.0912246779,0.0223609656,-0.3133271337,0.0051412471,0.3289259076,-0.018437,0.0320123062,-0.1938534528,-0.0220454745,0.0078472476,-0.1507639438,-0.0613341182,-0.0791995898,0.2228215635,0.0074299281,-0.0493339114,0.3533655703,0.2747455239,0.0639955476,-0.0365163907,-0.1238466576,-0.010248973,0.1141474769,-0.0905264094,-0.0358541906,0.2171731889,-0.1256070137,0.2478310466,0.4252819717,-0.0469338372,0.2657586038,-0.0382558815,-0.2736453712,0.2264523953,-0.0702164769,-0.0428678021,0.1976911724,-0.0458724052,0.0015260566,0.3408337235,0.0440044962,-0.0559479967,0.1770871282,0.427303493,-0.0043448089,-0.1692032218,-0.0192307252,-0.2083047777,0.1080089733,0.1594958156,-0.0258413963,0.1219083592,0.1693717986,0.1196489036,-0.1023140699,0.104523696,-0.0379634686,0.0820302144,0.3170749545,0.180519715,0.1618741006,-0.0916261226,-0.0561815985,-0.0036794732,-0.1100691184,0.1913622171,0.2398844659,-0.3654334843,0.2132201493,-0.145929724,-0.08981058,-0.1511178762,-0.0314461552,0.1613285244,-0.1398269385,-0.0764070004,0.2173885703,-0.1582910717,0.0283092428,-0.0862446353,0.0393373929,0.2715219557,-0.3807459176,-0.1608629972,-0.0596176609,-0.1818827242,0.2018623352,0.0678684339,0.1080862805,0.4683231413,0.0514345989,0.0387846157,-0.1155499518,-0.1522178799,0.0778788254,-0.1976525038,0.1767100245,0.0169449784,0.1401751041,-0.1079600304,-0.3839942813,0.1391080171,-0.1839413941,-0.3313497603,-0.0855730399,-0.0173469968,0.040500477,-0.2087071538,-0.4318087101,-0.2330217361,-0.1791113019,0.0216235537,-0.0695498586,0.2032481134,-0.0093688983,-0.0178895891,0.1497216076,0.2623847723,-0.1572469622,-0.0877216086,-0.4881262481,0.1246121675,-0.2836306691,-0.0197322294,-0.0922569484,-0.1024708152,0.4223552346,0.3086383641,-0.1903934777,0.1180599406,-0.0889263675,0.2408317477,-0.0977540761,-0.0265352037,0.1954074502,-0.0789332092,-0.099664703,-0.1173348725,0.0736229196,0.1176979691,-0.2414382696,0.0647149384,-0.017784683,0.6492415667,-0.0192825571,0.9116159081,0.0870948359,-0.0874783471,0.132925272,0.0044225426,-0.1579394042,-0.0129763782,-0.3009167314,0.0907453895,0.0279778596,-0.1869365871,0.1928884685,-0.1721884608,-0.0085576698,0.0702767372,-0.0171480607,-0.6442946196,-0.0674669221,0.0769532323,0.0041395295,0.343569845,-0.1565953642,0.0682594031,-0.2833535373,-0.0590694696,-0.0908283219,0.0464698821,0.2381592095,-0.1836909056,-0.359834075,0.0888359025,-0.2709055543,0.295625627,0.066839695,0.3335528672,0.0276030544,-0.0400166623,0.1112626567,-0.1317254603,0.7860842943,-0.044302132,0.0675168633,0.0571944974,-0.2433980554,-0.1615096331,-0.084504284,-0.33915326,0.2084002197,0.5520215034,0.294418484,-0.2110376358,-0.3628798425,0.0173130538,-0.1886417866,-0.1648462117,-0.2995319068,-0.3377781808,-0.1493666619,-0.3823865652,0.3109837472,0.2597950399,0.1134826764,-0.4368878603,-0.3712888062,-0.3057021499,-0.089634411,0.1953645051,-0.0665384829,0.0286609922,0.0238977037,0.2161179781,0.0219960399,0.2875387371,0.3828946054,0.2126956433,-0.1729637533,-0.3372835517,0.182780847,-0.0983873978,0.4843111634,0.0285968147,-0.2987975478,0.3437284529,-0.037379548,0.1920201927,-0.4507605731,0.5018005371,0.4323754013,0.1806503981,-0.5321657658,-0.5739417672,0.1534674019,0.3390219808,-0.098034218,0.1371526122,-0.2328913659,-0.2770265639,0.1009882987,0.195102036,0.6205477118,-0.0256134924,-0.0886720344,0.101748161,0.2892380357,0.0657999516,-0.206457451,0.1026582122,-0.3231612742,0.0041053961,0.0323343873,-0.1772434562,0.2091562748,0.4043450356,-0.1352706403,0.197356239,-0.1509763002,0.2151631564,-0.1551909298,0.3089804053,0.0421411842,-0.2329288721,-0.0506951697,0.2430498898,-0.131561324,0.2198929936,-0.1560653001,-0.1121162102,0.1223985925,-0.104833737,0.2029970586,0.0285311434,-0.4526079595,-0.0327905193,-0.183313787,-0.1597459316,0.0133769931,-0.0820379257,0.3468180001,0.2518924177,0.1860014796,-0.0041455068,0.0332504734,0.1427504271,-0.00881988,-0.3308890164,0.347242564,-0.0457745157,-0.1752589345,0.0488177985,0.0530347899,-0.2700374424,-0.0171610638,-0.0162138138,0.4081456363,-0.4375675619,-0.0731346756,-0.0617685728,-0.1597323567,-0.0974127799,0.1880708933,-0.0819994956,-0.2776878774,0.4453329742,-0.1918442398,-0.204532966,-0.0458533578,0.1701379865,0.272202611,-0.2661238909,0.5271493793,-0.0081695635,-0.2039593458,-0.2986089587,0.1181670129,0.2406757623,-0.1789259911,0.3072325885,0.1491332501,0.1801615208,0.23993285,0.0652863905,-0.2452689707,0.0456208475,-0.1946120709,-0.25122118,-0.4587342143,-0.2272527665,-0.1453618705,0.2936918736,0.3126306832,0.2828490436,-0.0498391464,0.3358579576,-0.4065794945,-0.1149407104,-0.1531325132,0.2728170753,0.1524819732,0.0128615918,-0.0067069111,0.2280720025,0.1662236303,0.2789082825,-0.0558065735,-0.3985133767,0.0245181639,0.0459400862,-0.1330243647,-0.0621523336,0.0371135734,-0.021833716,-0.2458432466,-0.1073626652,0.2516538203,0.0131633338,0.1318154633,0.2092590034,0.4427997768,0.3391921222,0.1198219731,0.1046171337,-0.1400850415,0.0519433282,0.0390864424,0.2915998697,0.0728554055,-0.0097600026,-0.0087768361,-0.2714516819,0.61431849,0.0637843087,0.0193376075,-0.393550694,-0.2791869938,0.128147617,0.0886587873,0.2924180329,0.057500381,-0.2814278901,0.3145102859,0.3561895788,0.0374487862,0.3104781806,0.169152528,-0.1325735748,0.1291972548,0.139757812,-0.0616449676,0.2524917126,0.1008177623,0.1690559834,0.2947505713,-0.2797976434,0.1288770884,0.1937782168,0.0881134272,0.0643601045,0.3061844707,0.0950236693,0.0919646323,0.4118394852,0.1421423703,0.5869659185,0.156269744,0.1330235153,0.0360658169,-0.4071056545,-0.3413540423,0.209771499,-0.3046946526,0.1884255558,-0.0142169474,0.2182088345,0.1494460404,-0.1891223937,-0.087864399,0.3838881552,-0.4658757448,-0.119666256,-0.527037859,-0.0618733913,0.0470664948,-0.1525252163,0.1422010809,0.0003348004,-0.0268428586,0.0209393594,-0.0446058884,0.03610982,-0.287540555,0.2877539098,-0.0674866289,-0.1569443941,0.0797995627,-0.0260496885,0.1523937732,-0.0068784403,0.2188947499,0.3761075139,0.3229938745,-0.1922142655,-0.0940963626,0.0517296307,-0.0628716573,-0.1619847715,0.2549487054,-0.1922801584,-0.0053928769,0.3815495968,0.2739574909,-0.2499833256,-0.334929347,0.0257082433,-0.0218391176,-0.0413478799,0.4966604412,0.1191578358,-0.1086807922,-0.2861167789,-0.0260956082,-0.4254947901,-0.0493845232,0.5210596919,-0.4102794826,0.1081943363,-0.1843963861,0.1237969697,-0.2985062003,0.4248542786,0.3058854938,0.2850418985,-0.467228055,0.0159201417,-0.2836147249,0.2286422253,-0.0501086563,0.0008188674,-0.2460899651,0.2427743077,-0.2073601484,0.0642418042,-0.0658618659,-0.1411859244,0.0898486227,0.3890626729,-0.3444704413,0.0781766847,-0.3071103692,-0.027144555,0.1391095668,-0.3164841235,0.1617375463,0.1141605154,0.1574425101,-0.1564975381,0.0271079689,0.2952474356,-0.1897252351,0.251014024,0.217400685,0.356393218,-0.2116020322,0.4840889871,-0.0992439166,0.0562532954,-0.3012344539,-0.1634408683,0.1430549324,0.1765762269,0.0225723367,0.0269494597,-0.2492116839,-0.0863267928,0.0803805813,-0.1846029311,0.033137355,0.0671218187,-0.0580933653,0.0357012153,0.0681237876,0.1903722286,0.2587812543,0.2300123721,-0.4224675,-0.4245306253,0.4023638964,-0.3883866966,-0.3013541996,-0.3631491065,0.2955272496,0.0590143502,0.0962256864,-0.157304287,0.0010297932,0.4445219934,-0.0583704785,-0.0071650837,0.3775203526,0.1706832051,0.2029901147,-0.1777066737,-0.11001537,0.1193484887,-0.0026773766,0.07206247,-0.3162924945]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/823","title":"how processing in batch works in datasets ","comments":"Hi Thomas,\nwhat I do not get from documentation is that why when you set batched=True,\nthis is processed in batch, while data is not divided to batched\nbeforehand, basically this is a question on the documentation and I do not\nget the batched=True, but sure, if you think this is more appropriate in\nforum I will post it there.\nthanks\nBest\nRabeeh\n\nOn Tue, Nov 10, 2020 at 12:21 PM Thomas Wolf <notifications@github.com>\nwrote:\n\n> Hi I don\u2019t think this is a request for a dataset like you labeled it.\n>\n> I also think this would be better suited for the forum at\n> https:\/\/discuss.huggingface.co. we try to keep the issue for the repo for\n> bug reports and new features\/dataset requests and have usage questions\n> discussed on the forum. Thanks.\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/823#issuecomment-724639476>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ARPXHH4FIPFHVVUHANAE4F3SPEO2JANCNFSM4TQQVEXQ>\n> .\n>\n","body":"Hi,\r\nI need to process my datasets before it is passed to dataloader in batch, \r\nhere is my codes \r\n\r\n```\r\nclass AbstractTask(ABC):\r\n task_name: str = NotImplemented\r\n preprocessor: Callable = NotImplemented\r\n split_to_data_split: Mapping[str, str] = NotImplemented\r\n tokenizer: Callable = NotImplemented\r\n max_source_length: str = NotImplemented\r\n max_target_length: str = NotImplemented\r\n # TODO: should not be a task item, but cannot see other ways.\r\n tpu_num_cores: int = None\r\n\r\n # The arguments set are for all tasks and needs to be kept common.\r\n def __init__(self, config):\r\n self.max_source_length = config['max_source_length']\r\n self.max_target_length = config['max_target_length']\r\n self.tokenizer = config['tokenizer']\r\n self.tpu_num_cores = config['tpu_num_cores']\r\n\r\n def _encode(self, batch) -> Dict[str, torch.Tensor]:\r\n batch_encoding = self.tokenizer.prepare_seq2seq_batch(\r\n [x[\"src_texts\"] for x in batch],\r\n tgt_texts=[x[\"tgt_texts\"] for x in batch],\r\n max_length=self.max_source_length,\r\n max_target_length=self.max_target_length,\r\n padding=\"max_length\" if self.tpu_num_cores is not None else \"longest\", # TPU hack\r\n return_tensors=\"pt\"\r\n )\r\n return batch_encoding.data\r\n\r\n\r\n def data_split(self, split):\r\n return self.split_to_data_split[split]\r\n\r\n def get_dataset(self, split, n_obs=None):\r\n split = self.data_split(split)\r\n if n_obs is not None:\r\n split = split+\"[:{}]\".format(n_obs)\r\n dataset = load_dataset(self.task_name, split=split)\r\n dataset = dataset.map(self.preprocessor, remove_columns=dataset.column_names)\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n dataset.set_format(type=\"torch\", columns=['input_ids', 'token_type_ids', 'attention_mask', 'label'])\r\n return dataset\r\n\r\n```\r\n\r\nI call it like \r\n\r\n`AutoTask.get(task, train_dataset_config).get_dataset(split=\"train\", n_obs=data_args.n_train) \r\n`\r\n\r\nThis gives the following error, to me because the data inside the dataset = dataset.map(lambda batch: self._encode(batch), batched=True) is not processed in batch, could you tell me how I can process dataset in batch inside my function? thanks \r\n\r\n File \"finetune_multitask_trainer.py\", line 192, in main\r\n if training_args.do_train else None\r\n File \"finetune_multitask_trainer.py\", line 191, in <dictcomp>\r\n split=\"train\", n_obs=data_args.n_train) for task in data_args.task}\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in get_dataset\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1236, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1207, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in <lambda>\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in _encode\r\n [x[\"src_texts\"] for x in batch],\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in <listcomp>\r\n [x[\"src_texts\"] for x in batch],\r\nTypeError: string indices must be integers\r\n\r\n","comment_length":167,"text":"how processing in batch works in datasets \nHi,\r\nI need to process my datasets before it is passed to dataloader in batch, \r\nhere is my codes \r\n\r\n```\r\nclass AbstractTask(ABC):\r\n task_name: str = NotImplemented\r\n preprocessor: Callable = NotImplemented\r\n split_to_data_split: Mapping[str, str] = NotImplemented\r\n tokenizer: Callable = NotImplemented\r\n max_source_length: str = NotImplemented\r\n max_target_length: str = NotImplemented\r\n # TODO: should not be a task item, but cannot see other ways.\r\n tpu_num_cores: int = None\r\n\r\n # The arguments set are for all tasks and needs to be kept common.\r\n def __init__(self, config):\r\n self.max_source_length = config['max_source_length']\r\n self.max_target_length = config['max_target_length']\r\n self.tokenizer = config['tokenizer']\r\n self.tpu_num_cores = config['tpu_num_cores']\r\n\r\n def _encode(self, batch) -> Dict[str, torch.Tensor]:\r\n batch_encoding = self.tokenizer.prepare_seq2seq_batch(\r\n [x[\"src_texts\"] for x in batch],\r\n tgt_texts=[x[\"tgt_texts\"] for x in batch],\r\n max_length=self.max_source_length,\r\n max_target_length=self.max_target_length,\r\n padding=\"max_length\" if self.tpu_num_cores is not None else \"longest\", # TPU hack\r\n return_tensors=\"pt\"\r\n )\r\n return batch_encoding.data\r\n\r\n\r\n def data_split(self, split):\r\n return self.split_to_data_split[split]\r\n\r\n def get_dataset(self, split, n_obs=None):\r\n split = self.data_split(split)\r\n if n_obs is not None:\r\n split = split+\"[:{}]\".format(n_obs)\r\n dataset = load_dataset(self.task_name, split=split)\r\n dataset = dataset.map(self.preprocessor, remove_columns=dataset.column_names)\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n dataset.set_format(type=\"torch\", columns=['input_ids', 'token_type_ids', 'attention_mask', 'label'])\r\n return dataset\r\n\r\n```\r\n\r\nI call it like \r\n\r\n`AutoTask.get(task, train_dataset_config).get_dataset(split=\"train\", n_obs=data_args.n_train) \r\n`\r\n\r\nThis gives the following error, to me because the data inside the dataset = dataset.map(lambda batch: self._encode(batch), batched=True) is not processed in batch, could you tell me how I can process dataset in batch inside my function? thanks \r\n\r\n File \"finetune_multitask_trainer.py\", line 192, in main\r\n if training_args.do_train else None\r\n File \"finetune_multitask_trainer.py\", line 191, in <dictcomp>\r\n split=\"train\", n_obs=data_args.n_train) for task in data_args.task}\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in get_dataset\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1236, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1207, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in <lambda>\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in _encode\r\n [x[\"src_texts\"] for x in batch],\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in <listcomp>\r\n [x[\"src_texts\"] for x in batch],\r\nTypeError: string indices must be integers\r\n\r\n\nHi Thomas,\nwhat I do not get from documentation is that why when you set batched=True,\nthis is processed in batch, while data is not divided to batched\nbeforehand, basically this is a question on the documentation and I do not\nget the batched=True, but sure, if you think this is more appropriate in\nforum I will post it there.\nthanks\nBest\nRabeeh\n\nOn Tue, Nov 10, 2020 at 12:21 PM Thomas Wolf <notifications@github.com>\nwrote:\n\n> Hi I don\u2019t think this is a request for a dataset like you labeled it.\n>\n> I also think this would be better suited for the forum at\n> https:\/\/discuss.huggingface.co. we try to keep the issue for the repo for\n> bug reports and new features\/dataset requests and have usage questions\n> discussed on the forum. Thanks.\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/823#issuecomment-724639476>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ARPXHH4FIPFHVVUHANAE4F3SPEO2JANCNFSM4TQQVEXQ>\n> .\n>\n","embeddings":[-0.5023794174,-0.1743950695,-0.229502216,0.1301572621,0.1780355573,0.0317946561,0.2955617607,0.1713047624,-0.1484707892,0.1780768037,0.0091214189,0.1247809753,0.0641202703,0.2456238419,-0.0561650358,-0.0358917713,-0.0745975822,0.0483772866,0.1096090525,0.0072343689,-0.3676092625,0.0689976066,-0.404307127,0.0522343665,-0.2230468243,-0.2693564892,-0.061245136,-0.0392570309,-0.0044130813,-0.1525889933,-0.0022256428,0.2334949821,0.2088395506,0.4326501787,-0.0001049565,0.0785311237,0.246775642,0.0393823534,-0.2903354466,0.2226282358,-0.0494002365,-0.1695528775,-0.106298469,-0.3042087555,-0.100836426,-0.0860008001,-0.1385592818,-0.3374301493,0.3848311007,0.1753834486,0.2844721079,0.2802553177,-0.1476055533,0.141417712,0.0610040873,0.071453087,0.085112974,-0.3902092874,0.0682323202,0.147927478,-0.0484061763,0.149620384,-0.2720299363,0.0859410837,-0.2550970316,0.0218879282,0.0260984451,-0.3847026229,0.0884854123,-0.1349758655,0.2846595347,-0.2123625129,-0.244074598,-0.3842577636,-0.2724657655,-0.4653172493,-0.0269787684,0.1131320298,-0.0684819147,0.1012712568,-0.0785877854,0.2408464402,-0.0737230629,-0.1666946113,0.1114271581,0.256811142,-0.2119238675,0.2198818326,0.0700407103,0.0887886882,0.1648149788,-0.2350720912,-0.0004340044,0.2723601758,-0.490796864,-0.0871194154,-0.111774452,-0.1519913375,-0.1408956349,0.0541603342,-0.1284128726,0.1272303164,0.3161275685,0.1564184427,-0.0295962766,0.1643566489,0.1187213957,0.2562070191,-0.077118583,-0.3050593734,-0.2260431051,0.0277573764,-0.0785889626,-0.3298466206,0.1471393257,0.3610413373,0.2499244362,0.0302780811,-0.4871647358,-0.090392217,-0.0190420989,-0.0580975227,0.1420931816,0.2845916152,-0.2407628298,0.2508462965,-0.0810260475,-0.1523166597,-0.2926350832,-0.0461959541,-0.3027382791,-0.1152567863,-0.1496196091,0.0702788755,0.1349772662,-0.1721600443,0.3239180446,0.0755829066,0.0962865055,-0.1746606827,0.5014520288,-0.3949423134,0.2224124223,0.2599134743,0.1795815676,0.2426499426,0.0532012805,0.2097658068,0.0399616584,0.07378117,-0.6257678866,-0.1791180819,0.2474631667,0.3138492405,-0.050551679,0.0014612001,-0.5935842991,-0.0869221464,0.3213269711,0.0906271413,0.1197098047,-0.1754258871,-0.1859953701,-0.1946797967,0.2003967762,0.0852582231,-0.4504095614,0.067932725,-0.1153643429,-0.3227443099,0.5454286933,0.1242681667,-0.1138667688,0.6199551225,-0.3095553815,0.378302753,0.1937190294,-0.318939209,-0.2815083861,0.5390690565,-0.2575122714,0.0180868749,0.0398210399,0.0325068124,0.558662951,-0.0602873936,0.0362305008,0.3787432015,-0.1192286387,0.0902983919,-0.0995698944,0.1156369671,0.0859533623,-0.017408317,0.103893064,0.0438636467,-0.2167430967,-0.0445535667,0.1737817377,0.0534989499,0.2965995669,0.2496741712,0.0114939697,0.2087981552,0.1932462901,-0.045412153,-0.2964651883,0.0957323313,-0.0203324519,-0.1832336485,0.0405643061,-0.3565836847,-0.1427385509,0.1464838088,-0.4080851972,-0.170314461,0.2187933773,0.0912246779,0.0223609656,-0.3133271337,0.0051412471,0.3289259076,-0.018437,0.0320123062,-0.1938534528,-0.0220454745,0.0078472476,-0.1507639438,-0.0613341182,-0.0791995898,0.2228215635,0.0074299281,-0.0493339114,0.3533655703,0.2747455239,0.0639955476,-0.0365163907,-0.1238466576,-0.010248973,0.1141474769,-0.0905264094,-0.0358541906,0.2171731889,-0.1256070137,0.2478310466,0.4252819717,-0.0469338372,0.2657586038,-0.0382558815,-0.2736453712,0.2264523953,-0.0702164769,-0.0428678021,0.1976911724,-0.0458724052,0.0015260566,0.3408337235,0.0440044962,-0.0559479967,0.1770871282,0.427303493,-0.0043448089,-0.1692032218,-0.0192307252,-0.2083047777,0.1080089733,0.1594958156,-0.0258413963,0.1219083592,0.1693717986,0.1196489036,-0.1023140699,0.104523696,-0.0379634686,0.0820302144,0.3170749545,0.180519715,0.1618741006,-0.0916261226,-0.0561815985,-0.0036794732,-0.1100691184,0.1913622171,0.2398844659,-0.3654334843,0.2132201493,-0.145929724,-0.08981058,-0.1511178762,-0.0314461552,0.1613285244,-0.1398269385,-0.0764070004,0.2173885703,-0.1582910717,0.0283092428,-0.0862446353,0.0393373929,0.2715219557,-0.3807459176,-0.1608629972,-0.0596176609,-0.1818827242,0.2018623352,0.0678684339,0.1080862805,0.4683231413,0.0514345989,0.0387846157,-0.1155499518,-0.1522178799,0.0778788254,-0.1976525038,0.1767100245,0.0169449784,0.1401751041,-0.1079600304,-0.3839942813,0.1391080171,-0.1839413941,-0.3313497603,-0.0855730399,-0.0173469968,0.040500477,-0.2087071538,-0.4318087101,-0.2330217361,-0.1791113019,0.0216235537,-0.0695498586,0.2032481134,-0.0093688983,-0.0178895891,0.1497216076,0.2623847723,-0.1572469622,-0.0877216086,-0.4881262481,0.1246121675,-0.2836306691,-0.0197322294,-0.0922569484,-0.1024708152,0.4223552346,0.3086383641,-0.1903934777,0.1180599406,-0.0889263675,0.2408317477,-0.0977540761,-0.0265352037,0.1954074502,-0.0789332092,-0.099664703,-0.1173348725,0.0736229196,0.1176979691,-0.2414382696,0.0647149384,-0.017784683,0.6492415667,-0.0192825571,0.9116159081,0.0870948359,-0.0874783471,0.132925272,0.0044225426,-0.1579394042,-0.0129763782,-0.3009167314,0.0907453895,0.0279778596,-0.1869365871,0.1928884685,-0.1721884608,-0.0085576698,0.0702767372,-0.0171480607,-0.6442946196,-0.0674669221,0.0769532323,0.0041395295,0.343569845,-0.1565953642,0.0682594031,-0.2833535373,-0.0590694696,-0.0908283219,0.0464698821,0.2381592095,-0.1836909056,-0.359834075,0.0888359025,-0.2709055543,0.295625627,0.066839695,0.3335528672,0.0276030544,-0.0400166623,0.1112626567,-0.1317254603,0.7860842943,-0.044302132,0.0675168633,0.0571944974,-0.2433980554,-0.1615096331,-0.084504284,-0.33915326,0.2084002197,0.5520215034,0.294418484,-0.2110376358,-0.3628798425,0.0173130538,-0.1886417866,-0.1648462117,-0.2995319068,-0.3377781808,-0.1493666619,-0.3823865652,0.3109837472,0.2597950399,0.1134826764,-0.4368878603,-0.3712888062,-0.3057021499,-0.089634411,0.1953645051,-0.0665384829,0.0286609922,0.0238977037,0.2161179781,0.0219960399,0.2875387371,0.3828946054,0.2126956433,-0.1729637533,-0.3372835517,0.182780847,-0.0983873978,0.4843111634,0.0285968147,-0.2987975478,0.3437284529,-0.037379548,0.1920201927,-0.4507605731,0.5018005371,0.4323754013,0.1806503981,-0.5321657658,-0.5739417672,0.1534674019,0.3390219808,-0.098034218,0.1371526122,-0.2328913659,-0.2770265639,0.1009882987,0.195102036,0.6205477118,-0.0256134924,-0.0886720344,0.101748161,0.2892380357,0.0657999516,-0.206457451,0.1026582122,-0.3231612742,0.0041053961,0.0323343873,-0.1772434562,0.2091562748,0.4043450356,-0.1352706403,0.197356239,-0.1509763002,0.2151631564,-0.1551909298,0.3089804053,0.0421411842,-0.2329288721,-0.0506951697,0.2430498898,-0.131561324,0.2198929936,-0.1560653001,-0.1121162102,0.1223985925,-0.104833737,0.2029970586,0.0285311434,-0.4526079595,-0.0327905193,-0.183313787,-0.1597459316,0.0133769931,-0.0820379257,0.3468180001,0.2518924177,0.1860014796,-0.0041455068,0.0332504734,0.1427504271,-0.00881988,-0.3308890164,0.347242564,-0.0457745157,-0.1752589345,0.0488177985,0.0530347899,-0.2700374424,-0.0171610638,-0.0162138138,0.4081456363,-0.4375675619,-0.0731346756,-0.0617685728,-0.1597323567,-0.0974127799,0.1880708933,-0.0819994956,-0.2776878774,0.4453329742,-0.1918442398,-0.204532966,-0.0458533578,0.1701379865,0.272202611,-0.2661238909,0.5271493793,-0.0081695635,-0.2039593458,-0.2986089587,0.1181670129,0.2406757623,-0.1789259911,0.3072325885,0.1491332501,0.1801615208,0.23993285,0.0652863905,-0.2452689707,0.0456208475,-0.1946120709,-0.25122118,-0.4587342143,-0.2272527665,-0.1453618705,0.2936918736,0.3126306832,0.2828490436,-0.0498391464,0.3358579576,-0.4065794945,-0.1149407104,-0.1531325132,0.2728170753,0.1524819732,0.0128615918,-0.0067069111,0.2280720025,0.1662236303,0.2789082825,-0.0558065735,-0.3985133767,0.0245181639,0.0459400862,-0.1330243647,-0.0621523336,0.0371135734,-0.021833716,-0.2458432466,-0.1073626652,0.2516538203,0.0131633338,0.1318154633,0.2092590034,0.4427997768,0.3391921222,0.1198219731,0.1046171337,-0.1400850415,0.0519433282,0.0390864424,0.2915998697,0.0728554055,-0.0097600026,-0.0087768361,-0.2714516819,0.61431849,0.0637843087,0.0193376075,-0.393550694,-0.2791869938,0.128147617,0.0886587873,0.2924180329,0.057500381,-0.2814278901,0.3145102859,0.3561895788,0.0374487862,0.3104781806,0.169152528,-0.1325735748,0.1291972548,0.139757812,-0.0616449676,0.2524917126,0.1008177623,0.1690559834,0.2947505713,-0.2797976434,0.1288770884,0.1937782168,0.0881134272,0.0643601045,0.3061844707,0.0950236693,0.0919646323,0.4118394852,0.1421423703,0.5869659185,0.156269744,0.1330235153,0.0360658169,-0.4071056545,-0.3413540423,0.209771499,-0.3046946526,0.1884255558,-0.0142169474,0.2182088345,0.1494460404,-0.1891223937,-0.087864399,0.3838881552,-0.4658757448,-0.119666256,-0.527037859,-0.0618733913,0.0470664948,-0.1525252163,0.1422010809,0.0003348004,-0.0268428586,0.0209393594,-0.0446058884,0.03610982,-0.287540555,0.2877539098,-0.0674866289,-0.1569443941,0.0797995627,-0.0260496885,0.1523937732,-0.0068784403,0.2188947499,0.3761075139,0.3229938745,-0.1922142655,-0.0940963626,0.0517296307,-0.0628716573,-0.1619847715,0.2549487054,-0.1922801584,-0.0053928769,0.3815495968,0.2739574909,-0.2499833256,-0.334929347,0.0257082433,-0.0218391176,-0.0413478799,0.4966604412,0.1191578358,-0.1086807922,-0.2861167789,-0.0260956082,-0.4254947901,-0.0493845232,0.5210596919,-0.4102794826,0.1081943363,-0.1843963861,0.1237969697,-0.2985062003,0.4248542786,0.3058854938,0.2850418985,-0.467228055,0.0159201417,-0.2836147249,0.2286422253,-0.0501086563,0.0008188674,-0.2460899651,0.2427743077,-0.2073601484,0.0642418042,-0.0658618659,-0.1411859244,0.0898486227,0.3890626729,-0.3444704413,0.0781766847,-0.3071103692,-0.027144555,0.1391095668,-0.3164841235,0.1617375463,0.1141605154,0.1574425101,-0.1564975381,0.0271079689,0.2952474356,-0.1897252351,0.251014024,0.217400685,0.356393218,-0.2116020322,0.4840889871,-0.0992439166,0.0562532954,-0.3012344539,-0.1634408683,0.1430549324,0.1765762269,0.0225723367,0.0269494597,-0.2492116839,-0.0863267928,0.0803805813,-0.1846029311,0.033137355,0.0671218187,-0.0580933653,0.0357012153,0.0681237876,0.1903722286,0.2587812543,0.2300123721,-0.4224675,-0.4245306253,0.4023638964,-0.3883866966,-0.3013541996,-0.3631491065,0.2955272496,0.0590143502,0.0962256864,-0.157304287,0.0010297932,0.4445219934,-0.0583704785,-0.0071650837,0.3775203526,0.1706832051,0.2029901147,-0.1777066737,-0.11001537,0.1193484887,-0.0026773766,0.07206247,-0.3162924945]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/823","title":"how processing in batch works in datasets ","comments":"Yes the forum is perfect for that. You can post in the `datasets` section.\r\nThanks a lot!","body":"Hi,\r\nI need to process my datasets before it is passed to dataloader in batch, \r\nhere is my codes \r\n\r\n```\r\nclass AbstractTask(ABC):\r\n task_name: str = NotImplemented\r\n preprocessor: Callable = NotImplemented\r\n split_to_data_split: Mapping[str, str] = NotImplemented\r\n tokenizer: Callable = NotImplemented\r\n max_source_length: str = NotImplemented\r\n max_target_length: str = NotImplemented\r\n # TODO: should not be a task item, but cannot see other ways.\r\n tpu_num_cores: int = None\r\n\r\n # The arguments set are for all tasks and needs to be kept common.\r\n def __init__(self, config):\r\n self.max_source_length = config['max_source_length']\r\n self.max_target_length = config['max_target_length']\r\n self.tokenizer = config['tokenizer']\r\n self.tpu_num_cores = config['tpu_num_cores']\r\n\r\n def _encode(self, batch) -> Dict[str, torch.Tensor]:\r\n batch_encoding = self.tokenizer.prepare_seq2seq_batch(\r\n [x[\"src_texts\"] for x in batch],\r\n tgt_texts=[x[\"tgt_texts\"] for x in batch],\r\n max_length=self.max_source_length,\r\n max_target_length=self.max_target_length,\r\n padding=\"max_length\" if self.tpu_num_cores is not None else \"longest\", # TPU hack\r\n return_tensors=\"pt\"\r\n )\r\n return batch_encoding.data\r\n\r\n\r\n def data_split(self, split):\r\n return self.split_to_data_split[split]\r\n\r\n def get_dataset(self, split, n_obs=None):\r\n split = self.data_split(split)\r\n if n_obs is not None:\r\n split = split+\"[:{}]\".format(n_obs)\r\n dataset = load_dataset(self.task_name, split=split)\r\n dataset = dataset.map(self.preprocessor, remove_columns=dataset.column_names)\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n dataset.set_format(type=\"torch\", columns=['input_ids', 'token_type_ids', 'attention_mask', 'label'])\r\n return dataset\r\n\r\n```\r\n\r\nI call it like \r\n\r\n`AutoTask.get(task, train_dataset_config).get_dataset(split=\"train\", n_obs=data_args.n_train) \r\n`\r\n\r\nThis gives the following error, to me because the data inside the dataset = dataset.map(lambda batch: self._encode(batch), batched=True) is not processed in batch, could you tell me how I can process dataset in batch inside my function? thanks \r\n\r\n File \"finetune_multitask_trainer.py\", line 192, in main\r\n if training_args.do_train else None\r\n File \"finetune_multitask_trainer.py\", line 191, in <dictcomp>\r\n split=\"train\", n_obs=data_args.n_train) for task in data_args.task}\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in get_dataset\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1236, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1207, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in <lambda>\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in _encode\r\n [x[\"src_texts\"] for x in batch],\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in <listcomp>\r\n [x[\"src_texts\"] for x in batch],\r\nTypeError: string indices must be integers\r\n\r\n","comment_length":17,"text":"how processing in batch works in datasets \nHi,\r\nI need to process my datasets before it is passed to dataloader in batch, \r\nhere is my codes \r\n\r\n```\r\nclass AbstractTask(ABC):\r\n task_name: str = NotImplemented\r\n preprocessor: Callable = NotImplemented\r\n split_to_data_split: Mapping[str, str] = NotImplemented\r\n tokenizer: Callable = NotImplemented\r\n max_source_length: str = NotImplemented\r\n max_target_length: str = NotImplemented\r\n # TODO: should not be a task item, but cannot see other ways.\r\n tpu_num_cores: int = None\r\n\r\n # The arguments set are for all tasks and needs to be kept common.\r\n def __init__(self, config):\r\n self.max_source_length = config['max_source_length']\r\n self.max_target_length = config['max_target_length']\r\n self.tokenizer = config['tokenizer']\r\n self.tpu_num_cores = config['tpu_num_cores']\r\n\r\n def _encode(self, batch) -> Dict[str, torch.Tensor]:\r\n batch_encoding = self.tokenizer.prepare_seq2seq_batch(\r\n [x[\"src_texts\"] for x in batch],\r\n tgt_texts=[x[\"tgt_texts\"] for x in batch],\r\n max_length=self.max_source_length,\r\n max_target_length=self.max_target_length,\r\n padding=\"max_length\" if self.tpu_num_cores is not None else \"longest\", # TPU hack\r\n return_tensors=\"pt\"\r\n )\r\n return batch_encoding.data\r\n\r\n\r\n def data_split(self, split):\r\n return self.split_to_data_split[split]\r\n\r\n def get_dataset(self, split, n_obs=None):\r\n split = self.data_split(split)\r\n if n_obs is not None:\r\n split = split+\"[:{}]\".format(n_obs)\r\n dataset = load_dataset(self.task_name, split=split)\r\n dataset = dataset.map(self.preprocessor, remove_columns=dataset.column_names)\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n dataset.set_format(type=\"torch\", columns=['input_ids', 'token_type_ids', 'attention_mask', 'label'])\r\n return dataset\r\n\r\n```\r\n\r\nI call it like \r\n\r\n`AutoTask.get(task, train_dataset_config).get_dataset(split=\"train\", n_obs=data_args.n_train) \r\n`\r\n\r\nThis gives the following error, to me because the data inside the dataset = dataset.map(lambda batch: self._encode(batch), batched=True) is not processed in batch, could you tell me how I can process dataset in batch inside my function? thanks \r\n\r\n File \"finetune_multitask_trainer.py\", line 192, in main\r\n if training_args.do_train else None\r\n File \"finetune_multitask_trainer.py\", line 191, in <dictcomp>\r\n split=\"train\", n_obs=data_args.n_train) for task in data_args.task}\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in get_dataset\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1236, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/idiap\/user\/rkarimi\/libs\/anaconda3\/envs\/internship\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py\", line 1207, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 56, in <lambda>\r\n dataset = dataset.map(lambda batch: self._encode(batch), batched=True)\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in _encode\r\n [x[\"src_texts\"] for x in batch],\r\n File \"\/remote\/idiap.svm\/user.active\/rkarimi\/dev\/internship\/seq2seq\/tasks.py\", line 37, in <listcomp>\r\n [x[\"src_texts\"] for x in batch],\r\nTypeError: string indices must be integers\r\n\r\n\nYes the forum is perfect for that. You can post in the `datasets` section.\r\nThanks a lot!","embeddings":[-0.5023794174,-0.1743950695,-0.229502216,0.1301572621,0.1780355573,0.0317946561,0.2955617607,0.1713047624,-0.1484707892,0.1780768037,0.0091214189,0.1247809753,0.0641202703,0.2456238419,-0.0561650358,-0.0358917713,-0.0745975822,0.0483772866,0.1096090525,0.0072343689,-0.3676092625,0.0689976066,-0.404307127,0.0522343665,-0.2230468243,-0.2693564892,-0.061245136,-0.0392570309,-0.0044130813,-0.1525889933,-0.0022256428,0.2334949821,0.2088395506,0.4326501787,-0.0001049565,0.0785311237,0.246775642,0.0393823534,-0.2903354466,0.2226282358,-0.0494002365,-0.1695528775,-0.106298469,-0.3042087555,-0.100836426,-0.0860008001,-0.1385592818,-0.3374301493,0.3848311007,0.1753834486,0.2844721079,0.2802553177,-0.1476055533,0.141417712,0.0610040873,0.071453087,0.085112974,-0.3902092874,0.0682323202,0.147927478,-0.0484061763,0.149620384,-0.2720299363,0.0859410837,-0.2550970316,0.0218879282,0.0260984451,-0.3847026229,0.0884854123,-0.1349758655,0.2846595347,-0.2123625129,-0.244074598,-0.3842577636,-0.2724657655,-0.4653172493,-0.0269787684,0.1131320298,-0.0684819147,0.1012712568,-0.0785877854,0.2408464402,-0.0737230629,-0.1666946113,0.1114271581,0.256811142,-0.2119238675,0.2198818326,0.0700407103,0.0887886882,0.1648149788,-0.2350720912,-0.0004340044,0.2723601758,-0.490796864,-0.0871194154,-0.111774452,-0.1519913375,-0.1408956349,0.0541603342,-0.1284128726,0.1272303164,0.3161275685,0.1564184427,-0.0295962766,0.1643566489,0.1187213957,0.2562070191,-0.077118583,-0.3050593734,-0.2260431051,0.0277573764,-0.0785889626,-0.3298466206,0.1471393257,0.3610413373,0.2499244362,0.0302780811,-0.4871647358,-0.090392217,-0.0190420989,-0.0580975227,0.1420931816,0.2845916152,-0.2407628298,0.2508462965,-0.0810260475,-0.1523166597,-0.2926350832,-0.0461959541,-0.3027382791,-0.1152567863,-0.1496196091,0.0702788755,0.1349772662,-0.1721600443,0.3239180446,0.0755829066,0.0962865055,-0.1746606827,0.5014520288,-0.3949423134,0.2224124223,0.2599134743,0.1795815676,0.2426499426,0.0532012805,0.2097658068,0.0399616584,0.07378117,-0.6257678866,-0.1791180819,0.2474631667,0.3138492405,-0.050551679,0.0014612001,-0.5935842991,-0.0869221464,0.3213269711,0.0906271413,0.1197098047,-0.1754258871,-0.1859953701,-0.1946797967,0.2003967762,0.0852582231,-0.4504095614,0.067932725,-0.1153643429,-0.3227443099,0.5454286933,0.1242681667,-0.1138667688,0.6199551225,-0.3095553815,0.378302753,0.1937190294,-0.318939209,-0.2815083861,0.5390690565,-0.2575122714,0.0180868749,0.0398210399,0.0325068124,0.558662951,-0.0602873936,0.0362305008,0.3787432015,-0.1192286387,0.0902983919,-0.0995698944,0.1156369671,0.0859533623,-0.017408317,0.103893064,0.0438636467,-0.2167430967,-0.0445535667,0.1737817377,0.0534989499,0.2965995669,0.2496741712,0.0114939697,0.2087981552,0.1932462901,-0.045412153,-0.2964651883,0.0957323313,-0.0203324519,-0.1832336485,0.0405643061,-0.3565836847,-0.1427385509,0.1464838088,-0.4080851972,-0.170314461,0.2187933773,0.0912246779,0.0223609656,-0.3133271337,0.0051412471,0.3289259076,-0.018437,0.0320123062,-0.1938534528,-0.0220454745,0.0078472476,-0.1507639438,-0.0613341182,-0.0791995898,0.2228215635,0.0074299281,-0.0493339114,0.3533655703,0.2747455239,0.0639955476,-0.0365163907,-0.1238466576,-0.010248973,0.1141474769,-0.0905264094,-0.0358541906,0.2171731889,-0.1256070137,0.2478310466,0.4252819717,-0.0469338372,0.2657586038,-0.0382558815,-0.2736453712,0.2264523953,-0.0702164769,-0.0428678021,0.1976911724,-0.0458724052,0.0015260566,0.3408337235,0.0440044962,-0.0559479967,0.1770871282,0.427303493,-0.0043448089,-0.1692032218,-0.0192307252,-0.2083047777,0.1080089733,0.1594958156,-0.0258413963,0.1219083592,0.1693717986,0.1196489036,-0.1023140699,0.104523696,-0.0379634686,0.0820302144,0.3170749545,0.180519715,0.1618741006,-0.0916261226,-0.0561815985,-0.0036794732,-0.1100691184,0.1913622171,0.2398844659,-0.3654334843,0.2132201493,-0.145929724,-0.08981058,-0.1511178762,-0.0314461552,0.1613285244,-0.1398269385,-0.0764070004,0.2173885703,-0.1582910717,0.0283092428,-0.0862446353,0.0393373929,0.2715219557,-0.3807459176,-0.1608629972,-0.0596176609,-0.1818827242,0.2018623352,0.0678684339,0.1080862805,0.4683231413,0.0514345989,0.0387846157,-0.1155499518,-0.1522178799,0.0778788254,-0.1976525038,0.1767100245,0.0169449784,0.1401751041,-0.1079600304,-0.3839942813,0.1391080171,-0.1839413941,-0.3313497603,-0.0855730399,-0.0173469968,0.040500477,-0.2087071538,-0.4318087101,-0.2330217361,-0.1791113019,0.0216235537,-0.0695498586,0.2032481134,-0.0093688983,-0.0178895891,0.1497216076,0.2623847723,-0.1572469622,-0.0877216086,-0.4881262481,0.1246121675,-0.2836306691,-0.0197322294,-0.0922569484,-0.1024708152,0.4223552346,0.3086383641,-0.1903934777,0.1180599406,-0.0889263675,0.2408317477,-0.0977540761,-0.0265352037,0.1954074502,-0.0789332092,-0.099664703,-0.1173348725,0.0736229196,0.1176979691,-0.2414382696,0.0647149384,-0.017784683,0.6492415667,-0.0192825571,0.9116159081,0.0870948359,-0.0874783471,0.132925272,0.0044225426,-0.1579394042,-0.0129763782,-0.3009167314,0.0907453895,0.0279778596,-0.1869365871,0.1928884685,-0.1721884608,-0.0085576698,0.0702767372,-0.0171480607,-0.6442946196,-0.0674669221,0.0769532323,0.0041395295,0.343569845,-0.1565953642,0.0682594031,-0.2833535373,-0.0590694696,-0.0908283219,0.0464698821,0.2381592095,-0.1836909056,-0.359834075,0.0888359025,-0.2709055543,0.295625627,0.066839695,0.3335528672,0.0276030544,-0.0400166623,0.1112626567,-0.1317254603,0.7860842943,-0.044302132,0.0675168633,0.0571944974,-0.2433980554,-0.1615096331,-0.084504284,-0.33915326,0.2084002197,0.5520215034,0.294418484,-0.2110376358,-0.3628798425,0.0173130538,-0.1886417866,-0.1648462117,-0.2995319068,-0.3377781808,-0.1493666619,-0.3823865652,0.3109837472,0.2597950399,0.1134826764,-0.4368878603,-0.3712888062,-0.3057021499,-0.089634411,0.1953645051,-0.0665384829,0.0286609922,0.0238977037,0.2161179781,0.0219960399,0.2875387371,0.3828946054,0.2126956433,-0.1729637533,-0.3372835517,0.182780847,-0.0983873978,0.4843111634,0.0285968147,-0.2987975478,0.3437284529,-0.037379548,0.1920201927,-0.4507605731,0.5018005371,0.4323754013,0.1806503981,-0.5321657658,-0.5739417672,0.1534674019,0.3390219808,-0.098034218,0.1371526122,-0.2328913659,-0.2770265639,0.1009882987,0.195102036,0.6205477118,-0.0256134924,-0.0886720344,0.101748161,0.2892380357,0.0657999516,-0.206457451,0.1026582122,-0.3231612742,0.0041053961,0.0323343873,-0.1772434562,0.2091562748,0.4043450356,-0.1352706403,0.197356239,-0.1509763002,0.2151631564,-0.1551909298,0.3089804053,0.0421411842,-0.2329288721,-0.0506951697,0.2430498898,-0.131561324,0.2198929936,-0.1560653001,-0.1121162102,0.1223985925,-0.104833737,0.2029970586,0.0285311434,-0.4526079595,-0.0327905193,-0.183313787,-0.1597459316,0.0133769931,-0.0820379257,0.3468180001,0.2518924177,0.1860014796,-0.0041455068,0.0332504734,0.1427504271,-0.00881988,-0.3308890164,0.347242564,-0.0457745157,-0.1752589345,0.0488177985,0.0530347899,-0.2700374424,-0.0171610638,-0.0162138138,0.4081456363,-0.4375675619,-0.0731346756,-0.0617685728,-0.1597323567,-0.0974127799,0.1880708933,-0.0819994956,-0.2776878774,0.4453329742,-0.1918442398,-0.204532966,-0.0458533578,0.1701379865,0.272202611,-0.2661238909,0.5271493793,-0.0081695635,-0.2039593458,-0.2986089587,0.1181670129,0.2406757623,-0.1789259911,0.3072325885,0.1491332501,0.1801615208,0.23993285,0.0652863905,-0.2452689707,0.0456208475,-0.1946120709,-0.25122118,-0.4587342143,-0.2272527665,-0.1453618705,0.2936918736,0.3126306832,0.2828490436,-0.0498391464,0.3358579576,-0.4065794945,-0.1149407104,-0.1531325132,0.2728170753,0.1524819732,0.0128615918,-0.0067069111,0.2280720025,0.1662236303,0.2789082825,-0.0558065735,-0.3985133767,0.0245181639,0.0459400862,-0.1330243647,-0.0621523336,0.0371135734,-0.021833716,-0.2458432466,-0.1073626652,0.2516538203,0.0131633338,0.1318154633,0.2092590034,0.4427997768,0.3391921222,0.1198219731,0.1046171337,-0.1400850415,0.0519433282,0.0390864424,0.2915998697,0.0728554055,-0.0097600026,-0.0087768361,-0.2714516819,0.61431849,0.0637843087,0.0193376075,-0.393550694,-0.2791869938,0.128147617,0.0886587873,0.2924180329,0.057500381,-0.2814278901,0.3145102859,0.3561895788,0.0374487862,0.3104781806,0.169152528,-0.1325735748,0.1291972548,0.139757812,-0.0616449676,0.2524917126,0.1008177623,0.1690559834,0.2947505713,-0.2797976434,0.1288770884,0.1937782168,0.0881134272,0.0643601045,0.3061844707,0.0950236693,0.0919646323,0.4118394852,0.1421423703,0.5869659185,0.156269744,0.1330235153,0.0360658169,-0.4071056545,-0.3413540423,0.209771499,-0.3046946526,0.1884255558,-0.0142169474,0.2182088345,0.1494460404,-0.1891223937,-0.087864399,0.3838881552,-0.4658757448,-0.119666256,-0.527037859,-0.0618733913,0.0470664948,-0.1525252163,0.1422010809,0.0003348004,-0.0268428586,0.0209393594,-0.0446058884,0.03610982,-0.287540555,0.2877539098,-0.0674866289,-0.1569443941,0.0797995627,-0.0260496885,0.1523937732,-0.0068784403,0.2188947499,0.3761075139,0.3229938745,-0.1922142655,-0.0940963626,0.0517296307,-0.0628716573,-0.1619847715,0.2549487054,-0.1922801584,-0.0053928769,0.3815495968,0.2739574909,-0.2499833256,-0.334929347,0.0257082433,-0.0218391176,-0.0413478799,0.4966604412,0.1191578358,-0.1086807922,-0.2861167789,-0.0260956082,-0.4254947901,-0.0493845232,0.5210596919,-0.4102794826,0.1081943363,-0.1843963861,0.1237969697,-0.2985062003,0.4248542786,0.3058854938,0.2850418985,-0.467228055,0.0159201417,-0.2836147249,0.2286422253,-0.0501086563,0.0008188674,-0.2460899651,0.2427743077,-0.2073601484,0.0642418042,-0.0658618659,-0.1411859244,0.0898486227,0.3890626729,-0.3444704413,0.0781766847,-0.3071103692,-0.027144555,0.1391095668,-0.3164841235,0.1617375463,0.1141605154,0.1574425101,-0.1564975381,0.0271079689,0.2952474356,-0.1897252351,0.251014024,0.217400685,0.356393218,-0.2116020322,0.4840889871,-0.0992439166,0.0562532954,-0.3012344539,-0.1634408683,0.1430549324,0.1765762269,0.0225723367,0.0269494597,-0.2492116839,-0.0863267928,0.0803805813,-0.1846029311,0.033137355,0.0671218187,-0.0580933653,0.0357012153,0.0681237876,0.1903722286,0.2587812543,0.2300123721,-0.4224675,-0.4245306253,0.4023638964,-0.3883866966,-0.3013541996,-0.3631491065,0.2955272496,0.0590143502,0.0962256864,-0.157304287,0.0010297932,0.4445219934,-0.0583704785,-0.0071650837,0.3775203526,0.1706832051,0.2029901147,-0.1777066737,-0.11001537,0.1193484887,-0.0026773766,0.07206247,-0.3162924945]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/822","title":"datasets freezes ","comments":"Pytorch is unable to convert strings to tensors unfortunately.\r\nYou can use `set_format(type=\"torch\")` on columns that can be converted to tensors, such as token ids.\r\n\r\nThis makes me think that we should probably raise an error or at least a warning when one tries to create pytorch tensors out of text columns","body":"Hi, I want to load these two datasets and convert them to Dataset format in torch and the code freezes for me, could you have a look please? thanks \r\n\r\ndataset1 = load_dataset(\"squad\", split=\"train[:10]\")\r\ndataset1 = dataset1.set_format(type='torch', columns=['context', 'answers', 'question'])\r\n\r\ndataset2 = load_dataset(\"imdb\", split=\"train[:10]\")\r\ndataset2 = dataset2.set_format(type=\"torch\", columns=[\"text\", \"label\"])\r\nprint(len(dataset1))\r\n","comment_length":52,"text":"datasets freezes \nHi, I want to load these two datasets and convert them to Dataset format in torch and the code freezes for me, could you have a look please? thanks \r\n\r\ndataset1 = load_dataset(\"squad\", split=\"train[:10]\")\r\ndataset1 = dataset1.set_format(type='torch', columns=['context', 'answers', 'question'])\r\n\r\ndataset2 = load_dataset(\"imdb\", split=\"train[:10]\")\r\ndataset2 = dataset2.set_format(type=\"torch\", columns=[\"text\", \"label\"])\r\nprint(len(dataset1))\r\n\nPytorch is unable to convert strings to tensors unfortunately.\r\nYou can use `set_format(type=\"torch\")` on columns that can be converted to tensors, such as token ids.\r\n\r\nThis makes me think that we should probably raise an error or at least a warning when one tries to create pytorch tensors out of text columns","embeddings":[-0.1619098037,-0.3347924054,-0.0465797186,0.5367450714,0.3581036329,0.2125209719,0.507258594,0.3903239369,-0.0151281459,0.1207219064,-0.2165184468,0.2929005921,-0.1185355037,-0.163084209,-0.1131302863,-0.5349528193,0.1154309735,-0.0951829627,-0.3404101431,0.099049516,0.0500869565,0.0638206452,-0.1515897661,-0.1478492171,-0.4358990192,0.0075123678,0.0953070819,0.0503921248,-0.1359757781,-0.0677020252,0.1145445183,-0.0718163922,0.0865185112,0.7834653854,-0.0001211927,0.0531437434,0.304533422,0.1574761122,-0.1029468924,-0.1995556355,0.3838464916,-0.1363200545,0.1754233837,-0.2181679606,-0.13459225,0.010250886,-0.2882260978,-0.695276022,0.3244414032,0.4434750676,0.0937087387,0.5506082177,0.0422830433,0.0736018717,0.3115720153,-0.0100868698,-0.209312886,-0.1175049171,0.3612826467,0.1951310337,0.1336320788,0.2532553673,-0.4077238142,-0.0682554767,0.0949572027,0.073289074,-0.6647979617,-0.4644117653,0.0864163265,0.3095062077,0.5834822059,-0.459107399,-0.1308524907,-0.1118449643,0.0003552453,0.0143291298,0.0995948836,0.1460790634,-0.1873875856,-0.0970856175,-0.3416729867,-0.1862387955,-0.244726643,0.2388877869,-0.1802271158,-0.1747247279,0.0037158506,0.1924472749,0.1358708888,-0.0270289164,0.4467356503,-0.3042343855,0.1853563637,-0.0435124636,-0.3819194436,-0.0245090611,-0.5991339684,-0.3232549131,-0.0219162125,-0.0832411498,-0.0257245861,0.1993785948,0.0594956875,0.1835234612,-0.0041054711,0.2055943012,0.0165229831,0.3655209541,0.1646527648,-0.4713938236,-0.0830719993,0.1285180002,-0.0653167292,-0.5972931981,-0.0566868968,-0.1479667127,0.1452046931,-0.2821500599,-0.2286584377,0.0399436429,-0.2544407845,0.0160606895,0.1840375215,0.3908998072,-0.2808794975,0.2990472019,0.3615897,0.1452337205,-0.2739693224,-0.0120231789,-0.0326119326,0.1297482997,-0.1249793321,-0.0447367541,0.2628989518,-0.4223605096,-0.0743070692,0.3290851712,0.1034980193,0.1094256118,0.2540254295,-0.243856743,0.2168108374,0.0404201262,0.0672834069,0.2015821189,0.0806757435,0.3792657554,-0.0123764686,0.3443981409,-0.2613652945,0.1523290128,-0.2981989384,0.0603747405,0.0591149107,0.0105317133,0.0715914145,0.1019400582,0.3803873658,-0.136848405,0.213163957,-0.2690629065,-0.122754328,-0.0398496948,0.1607143283,0.3656430542,-0.2027439773,0.080682762,0.5047172904,0.0049646464,-0.0276198518,0.3383366168,-0.166179657,-0.0340987258,0.0272393059,-0.1163115501,0.3451951146,-0.2792834044,0.0360599644,0.082947053,-0.0611122884,-0.0769845322,0.1463359892,0.1638709307,0.1401912272,-0.0225377791,0.2774900198,0.1859399676,-0.1286250055,0.0115317963,-0.1566434503,-0.0132685956,0.5077599287,0.0647813752,0.0822657272,-0.0090032117,-0.254273206,0.2098658383,-0.062928237,-0.1881982088,0.0560245216,0.4037212729,-0.3440684676,0.0713159442,0.0562156551,-0.4236242771,-0.4435555041,0.0634101629,0.2090815157,0.154961288,0.1222557649,-0.010005923,-0.0060081375,-0.1404091865,0.0333748497,0.2474482059,-0.0741904899,-0.3382622898,0.0578820817,0.0056778826,-0.0865047723,0.3033380806,-0.1110071093,0.254724741,-0.2904978395,0.3031731248,-0.056193836,-0.3207971752,-0.0883349702,0.2030960917,0.0999128968,-0.0701328889,-0.0903616995,0.2229854614,-0.0041757375,0.0624920875,-0.4371133149,-0.1511400491,0.2752560377,-0.1057287306,0.3364081085,0.0070383325,-0.0124842981,-0.1347418278,-0.1144411191,0.1100080237,-0.1437335908,0.2196324319,-0.0382271335,0.0104148416,-0.2166306674,0.1087430939,-0.1233973876,0.037359409,0.3091332912,-0.1157442108,0.2992348969,0.0862626061,-0.6952055097,-0.0541114882,0.3723339438,-0.0515537821,0.0357348733,0.0595976487,-0.2202146947,0.1280035228,0.1781383604,-0.3011554778,0.237342909,0.1993241459,0.0374270044,-0.0060503324,-0.3070225716,-0.1291005611,0.2157094181,0.1472150981,0.0153890336,0.0693671256,0.0225959811,-0.1770961136,-0.0239738394,0.1697230041,0.107528761,0.0790479556,-0.3709747493,0.1485443711,-0.0851233751,0.0055117817,-0.2897947133,-0.0226648543,-0.2290740311,-0.2103304267,-0.0848406702,0.3650323451,0.2018142045,0.0982009545,0.1191345453,-0.2002101839,0.2110097259,-0.164093554,-0.1957546026,-0.2156160027,-0.1579663157,-0.0199520588,0.4141778648,-0.0392508619,0.2661200166,-0.0810979456,0.1195303053,-0.3094357848,-0.1037665382,0.1392344981,-0.2589545846,-0.1265085787,-0.1212610528,0.2966690063,-0.017536452,-0.0650339201,0.1117540449,0.0458006449,-0.0447583608,0.2886176407,0.0960974023,-0.0350133441,-0.0744962767,-0.5995054841,-0.2112913281,-0.1445088536,-0.0621451512,-0.3863690197,-0.0292695723,-0.120325923,0.460427165,0.2548903227,0.0141043458,0.1126452461,-0.029850319,-0.2954290211,0.5208236575,-0.2749049962,-0.4628773332,-0.0779241547,0.2729069293,-0.0477543809,0.3154073954,-0.1507891417,0.0254723467,-0.4094622731,0.045414146,-0.1422857046,-0.0549072437,0.4183910787,0.1074017733,0.0682937577,-0.084711656,0.0460318625,-0.0318503417,-0.1916361302,0.0971079543,-0.0728925318,0.3707443476,0.0162022281,0.4566970766,-0.1325985938,-0.477281332,0.1253287792,-0.1024665684,0.3399639726,-0.0632666945,-0.4580647051,0.0325973742,-0.1638097912,0.0203840174,-0.137748152,-0.0733874142,-0.1564056724,-0.2065341622,0.2494827211,0.0467860103,-0.1162741482,0.2099995762,-0.5037884116,0.0899967924,-0.1194486767,0.3503584564,-0.220053792,-0.2446114272,0.0894979164,-0.0262189358,0.0932111591,0.0452532507,-0.523216486,-0.0482228249,-0.3011490405,0.3275445998,0.2851087451,0.3945766687,-0.1878953427,0.0880636722,0.1405413598,-0.0121411504,0.5735449195,0.0773106962,-0.3306096792,0.0714651123,-0.0654986501,-0.6979705691,0.1823859066,-0.1450955719,0.1250785738,0.0778574795,0.3985333443,-0.2446841598,-0.5145688653,0.1661031097,0.5513159633,-0.067939505,-0.1278381944,-0.1222823784,-0.0074194362,-0.373198837,-0.0467737913,0.3043912053,0.1292806268,-0.3400334716,-0.3042585254,-0.1497847885,0.0717579424,0.4278435707,-0.0078843143,0.1984150857,-0.1825616658,0.3117378354,0.1805599928,0.3363945484,0.204188183,0.1429825574,-0.2641713619,0.0028305354,0.197644949,0.157817781,0.3143092394,0.138880074,-0.0275352728,0.0444601066,-0.3225722611,0.1053727418,-0.1478787661,0.3770979047,0.4117918909,0.4101708531,-0.5693951845,-0.4581964612,0.1601796299,0.2949399352,0.1253014356,0.514218986,-0.139828369,-0.1602346897,0.2174318582,0.1858221143,0.9097225666,-0.5160053372,0.3435388207,0.1018021852,-0.1608796567,0.3924465775,-0.4085098505,0.0474883616,-0.401460439,-0.0857328102,-0.0788028091,-0.3112160265,0.314555347,-0.1096248701,-0.2194159925,0.1111756563,-0.2852615118,0.362826556,-0.1312673241,0.4157393277,-0.2564929426,-0.3052507341,0.0096839089,0.1124138013,-0.1420692503,0.0391997099,0.1307408661,0.0171858463,-0.2372531146,0.0918162167,-0.2067209482,0.1955279708,-0.3046233356,0.0834035799,-0.2852322161,-0.6679525375,0.2832608521,0.8073207736,0.3981665373,0.0669004917,-0.0834553018,0.0971856117,-0.1695283651,-0.2504493594,0.0269056708,-0.3444750905,0.2013790458,0.0798741728,-0.1297061294,0.2350297272,0.057570897,-0.1077376455,-0.0895990357,-0.0086019877,0.2334491909,-0.4762621522,-0.2683629692,0.0145069864,0.0258033071,-0.0851543248,0.0587984882,0.0282173157,0.0383639485,0.1562390774,-0.0542327277,-0.1352987885,0.0846428126,0.6563454866,-0.0061267428,-0.096102342,0.5771403909,0.6213268638,-0.2204687893,-0.2095316797,0.204873085,0.4002889395,-0.3191522658,0.068903625,-0.1601971835,-0.1529139727,-0.1304018945,0.0528423525,-0.170387432,-0.1908480525,-0.0382954925,-0.0024241277,-0.5649744868,0.0948399827,0.2658156455,0.4143565595,0.1766462624,0.0676173046,-0.2299854159,-0.1003998667,-0.1700629145,0.1461408734,0.0186311733,0.1811944395,0.169264257,-0.1351484805,0.1340319663,-0.4327129722,-0.0117486641,0.2693085372,0.1207040548,-0.1611533463,0.0037209417,0.0609984472,-0.1493924409,-0.2178912163,-0.222203657,-0.3654772341,0.1486224532,-0.1958627105,0.2625582814,0.538805604,-0.1317013353,0.0577047691,-0.1244095713,0.2113130987,-0.0680030733,0.0010776214,-0.1793113798,0.0337107591,0.0958347321,0.0973706022,0.3113963902,-0.0363971628,-0.3601101637,0.1446569115,0.4241945744,-0.0860697627,0.0985026434,-0.2135945708,0.0952600092,-0.0058070025,0.1979549527,0.3417714238,-0.1832294166,-0.0298404228,0.2780558467,0.0876779631,0.0256874412,-0.0857880637,-0.3014292121,-0.1910788715,-0.0677775666,0.0360645428,-0.0907765478,0.4408115447,-0.1841945499,0.2449506223,0.2201438993,-0.053762719,0.1670485735,0.3690186739,0.1142244413,-0.032390099,0.2292596549,0.1114542484,-0.0382730663,-0.0421534963,0.1451996565,0.6161945462,0.3098063767,0.2715212107,0.3532732129,0.0654802471,-0.2112428397,-0.0687883273,0.039396707,-0.0224070214,-0.048443675,0.471244514,-0.1950107664,-0.4782605767,-0.079380542,0.1935923994,-0.2054057717,-0.0167619698,-0.1558363587,-0.1668985486,0.1748680919,-0.0170220472,-0.0580653735,-0.0894870982,0.3186838627,0.3166029453,-0.0515858829,-0.0481539443,0.1059055701,0.2006661892,-0.0935683474,0.0305728335,0.1262204349,0.1343733221,-0.1311130077,0.0214785133,0.2105252445,0.3595518768,0.1974035054,-0.3775435388,0.1038366407,0.0788815692,-0.1375031173,0.4146793485,0.2447693944,-0.0010706459,0.0035324502,0.3206641078,0.111569494,-0.1362182796,-0.1674811393,-0.0720978603,0.0024070842,-0.228616029,0.3632534742,-0.1428265721,-0.2521868646,-0.2247458845,-0.2958553433,-0.3985684812,-0.0584642366,0.6664867997,-0.1330010146,-0.1282684952,-0.0228909682,0.0111746648,0.2890182137,0.5228798389,0.0561596155,0.2366770357,-0.3511782289,0.0257967003,-0.2219412476,-0.0386843421,-0.0867254362,0.0581281036,-0.0493377224,0.0415677913,0.1033102199,0.3067891598,0.0808424801,-0.087951906,0.1135655791,0.2893975973,-0.3699816167,-0.0638180673,-0.0227779988,0.1017897427,0.1345224828,-0.3856550157,0.2077293545,0.1359437853,-0.024448704,-0.3676212728,0.1786266714,0.1885702014,-0.0352404229,0.2021131665,-0.0608666092,0.6809895039,0.1546275765,0.0014239688,0.1001373231,-0.2080869228,-0.0296401009,0.0288342275,0.0372903831,0.2259258628,-0.1935019046,-0.2897948623,-0.2599354684,0.4156822264,-0.0505487509,-0.0398924164,-0.1345122606,0.1093737558,0.0320105962,0.1095596105,0.2725985646,0.5703676343,0.0921333581,0.4067799151,-0.3857578933,-0.1645225883,0.5223591328,-0.2746139467,-0.3560431898,-0.0018261961,-0.0017388201,0.5570203066,0.0881890953,-0.7385710478,-0.0915500373,-0.0065587135,-0.2329978198,0.0450816639,0.208785519,0.182759285,0.1478168666,-0.1287043542,0.3528142273,0.075620085,-0.0020419117,-0.2567732334,-0.2446459085]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/817","title":"Add MRQA dataset","comments":"Done! cf #1117 and #1022","body":"## Adding a Dataset\r\n- **Name:** MRQA\r\n- **Description:** Collection of different (subsets of) QA datasets all converted to the same format to evaluate out-of-domain generalization (the datasets come from different domains, distributions, etc.). Some datasets are used for training and others are used for evaluation. This dataset was collected as part of MRQA 2019's shared task \r\n- **Paper:** https:\/\/arxiv.org\/abs\/1910.09753\r\n- **Data:** https:\/\/github.com\/mrqa\/MRQA-Shared-Task-2019\r\n- **Motivation:** Out-of-domain generalization is becoming (has become) a de-factor evaluation for NLU systems\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).","comment_length":5,"text":"Add MRQA dataset\n## Adding a Dataset\r\n- **Name:** MRQA\r\n- **Description:** Collection of different (subsets of) QA datasets all converted to the same format to evaluate out-of-domain generalization (the datasets come from different domains, distributions, etc.). Some datasets are used for training and others are used for evaluation. This dataset was collected as part of MRQA 2019's shared task \r\n- **Paper:** https:\/\/arxiv.org\/abs\/1910.09753\r\n- **Data:** https:\/\/github.com\/mrqa\/MRQA-Shared-Task-2019\r\n- **Motivation:** Out-of-domain generalization is becoming (has become) a de-factor evaluation for NLU systems\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\nDone! cf #1117 and #1022","embeddings":[-0.3379291296,-0.2314092815,-0.1203742698,-0.0300259683,-0.0403992161,-0.0229204502,0.2539535463,0.3597253561,-0.0810624361,0.288100183,-0.2137913704,0.1627328098,-0.043995712,0.4142344594,0.2651360333,-0.0954021811,-0.0707546696,0.1376918107,-0.2828525305,-0.008358201,-0.1140417457,0.3614771664,-0.068791002,-0.0657266006,0.2294557393,-0.2739301324,-0.1435822994,-0.1535597146,0.0984469503,-0.2097584605,-0.307341218,0.0807675719,0.0208870694,0.0269577224,-0.0000959789,-0.1600842327,-0.0307716187,-0.0323698148,-0.1009855121,0.2765292823,-0.2225584537,0.129726097,-0.0700436011,0.0416208878,-0.2557869256,-0.2212925404,-0.153503865,-0.0409474634,0.2427387387,0.181117177,0.340159446,0.3674889505,-0.2893643975,-0.4059957564,-0.4239031076,-0.081427604,-0.1550893337,0.0865431055,0.3566902578,-0.1077354625,-0.1909460276,0.3263984323,0.1536389291,0.0299654305,-0.0900519192,-0.159235999,0.3744971752,-0.0292295627,0.107344158,0.4105097651,0.2631832063,-0.0465812609,-0.3531058133,-0.1951436549,0.0704900697,-0.1821657866,-0.0928804353,0.0620762333,-0.3613476455,0.1523222625,-0.1931378692,0.2035882026,-0.015180937,0.0227988157,-0.2607189119,0.2774297595,-0.2471418232,-0.226174593,0.1664784551,-0.0654361397,-0.3561445177,0.1364005357,-0.1637908965,-0.2880037129,-0.3712352514,-0.4995357394,0.2813028395,0.023594033,0.2546811998,0.2123573273,0.2925755978,0.1636602432,-0.0038634702,0.0692492276,0.1663413644,0.4039610028,-0.2219952345,-0.192112267,0.1895534843,0.208284989,-0.1469023675,0.187347576,0.1003925428,0.1334969997,-0.25304389,-0.0150684481,-0.2466462702,-0.1852458864,-0.3833523095,-0.1135929748,-0.0365342796,0.1433398277,-0.0861154273,0.2908374071,0.0823445544,-0.0463198014,-0.2654726803,0.3775545359,-0.1813080609,-0.3199292123,-0.254866004,0.0117407721,-0.0827007964,0.0573413149,0.0858379304,0.2928692997,0.3532237709,-0.1085598022,0.2278307825,0.0061730621,0.2663857937,0.0581049696,0.2235987186,0.0747802556,0.1213533357,-0.2920752764,0.1235811636,0.0432914123,-0.1935803592,-0.2713438272,0.2505704761,-0.1460158676,-0.2224564701,0.304572612,0.0748012811,0.0113745155,0.1157138348,0.5489933491,-0.3620863557,-0.0944209471,-0.0554565154,0.1007563844,0.0368919522,0.000385037,0.189528808,0.3353548348,-0.0003263598,-0.133159101,0.0335387699,0.071202673,-0.0359048918,0.2940556109,-0.2784415483,0.1695641279,-0.3699987233,0.0783908814,0.4032755196,-0.1867573857,-0.0074202549,0.3002349138,-0.2871706784,-0.1359691769,0.1756822765,0.1299977005,0.1245957986,0.0122871082,0.1817288995,0.5127796531,-0.2431005836,0.0920020789,-0.2167281359,-0.2370366156,-0.0599852726,0.0024839011,-0.0122573562,0.1335996687,0.2170537859,-0.2014193386,0.1739390939,-0.2022508532,-0.0050009964,-0.07848683,0.0024382295,-0.314950943,0.0090162847,-0.2671385109,-0.1517367512,-0.0967749655,-0.4534310997,0.3336873353,0.152449891,-0.3023792207,-0.1769931465,-0.1299985796,-0.0807174817,-0.4765872657,0.3982811272,0.3701138496,0.0698428527,0.1260946244,-0.0193110146,-0.1419375688,0.2912904322,-0.1270203888,0.0586061105,0.0004814067,-0.1180670634,0.0492629744,0.3360090256,0.5060277581,-0.152502954,0.055595465,0.0420875773,0.057534337,-0.3870841563,0.2160793543,0.4085708261,0.4257628322,-0.0705949292,-0.0956271812,0.0596628636,0.0558207706,-0.0314811692,0.0494980328,-0.1383434683,0.5979952812,0.2053761035,0.1295676678,0.1104024649,0.1516973823,0.1000224128,-0.0982379839,-0.1745696217,-0.4162879586,0.1442571431,0.1955443025,-0.0902796984,-0.0845643878,-0.2908942997,0.2544463575,0.303571701,-0.035298489,0.3126989007,-0.2187599987,-0.1859422028,-0.0933285207,0.2261098474,0.1316975504,0.1161721274,0.2954978347,0.2874610424,0.1670740843,0.0065192385,-0.2544041872,0.1500364542,0.0344854891,-0.0771378651,0.2270834446,0.2665913999,0.0180686619,-0.3005491495,0.1340085119,0.0143347755,-0.0845263451,-0.1613460481,0.234033823,-0.1788584888,-0.1587464213,-0.1644724011,0.0549763814,-0.262589395,-0.4955067635,0.3357520401,-0.0196520034,-0.2489840388,0.3200473785,0.1504140496,0.330916822,-0.1537104994,-0.0463568568,0.2283955365,-0.0282618571,-0.049723886,0.2890926301,0.2149559259,0.1373776346,0.5632379651,-0.0051970701,0.2478790432,-0.3795193136,-0.4318851233,0.1999313235,-0.2370245904,0.46453619,0.1981160343,-0.2267215699,-0.3076577485,-0.0213166792,0.0398454815,-0.1413614005,-0.1201088279,0.0812540203,-0.2712190747,0.0648097992,-0.2831617296,-0.5327270627,0.0566346683,-0.5068886876,0.1568512619,-0.3603903949,0.0797085315,0.1042668,-0.278903693,0.0790781826,-0.3446101546,0.057152573,-0.3253599107,0.2067345679,0.120967567,-0.4968489408,-0.1777054071,-0.1131485328,-0.289747566,0.6386483312,0.0175543912,-0.2199180573,-0.5145491958,0.091030784,0.0478203148,0.0455327854,-0.0793416351,0.5354915261,-0.3646611869,-0.30148983,0.090597108,-0.0708136111,0.3101285696,0.1899701208,-0.1019580886,-0.1841344684,0.2542762756,0.1451863945,0.2713384628,-0.014211609,0.0986515433,0.2500343621,-0.2172840834,0.221716404,0.0079408241,0.0170631967,0.30557096,0.2769009173,-0.0806008652,0.5731272101,0.0445235446,-0.0770695359,-0.0783169866,-0.0349746235,-0.3344036937,-0.2794383764,0.0421802737,0.0059060683,0.3819884956,0.1042717248,-0.0086072143,-0.0501840077,0.1014809385,0.3408029675,0.0646784306,0.0330703855,-0.2390542626,-0.4494090974,-0.0519387759,-0.1070117429,0.3143009245,0.088457115,0.0845557004,-0.2683357298,-0.2516252995,-0.0059056208,0.0094149942,0.3992017508,-0.2652344704,-0.0371277928,0.1431508809,-0.190765962,0.0229456108,0.0642932877,-0.1543306112,-0.1891494542,-0.0831179321,0.148750484,-0.2157453001,-0.2229653746,0.1949570328,0.1350525022,-0.1360062361,-0.0228717532,-0.212086916,-0.2212324739,-0.3085677326,0.2361793965,-0.0843450278,0.2148912698,0.0483340211,-0.0253457502,-0.0657322556,-0.2121558487,0.2480683774,0.4201636612,0.1350887865,0.2734489143,0.0316348411,-0.0239642803,0.0513742268,-0.1004037112,0.0666554868,0.0955391452,-0.0019402178,0.0648261309,-0.0584723726,0.3847210109,0.1819701344,-0.0516946763,0.0875364915,-0.1132542789,0.0825315714,-0.1884168833,-0.0495997146,0.123963885,0.0365410373,0.1141567752,0.0228123777,0.3281755447,0.10594327,-0.2612778842,-0.0456396602,0.2692076564,-0.122112155,0.3295139968,0.1742990017,0.7732360363,-0.00738732,0.2962048948,0.507866323,0.0742957667,0.1719227731,-0.05234088,0.264147222,-0.3744692802,0.0218851008,0.0224463474,0.0845637247,0.1778323948,-0.0121033583,-0.2749878168,0.2346036583,0.1852474809,0.2586102486,0.0086384863,-0.2089555711,-0.0746501237,-0.231914416,-0.320638001,0.2541207969,0.0574209057,0.0008051437,-0.0982521698,-0.1986857951,0.0869774222,0.2444111705,0.0079419008,0.0152874347,-0.2702739537,-0.2923884094,0.0848292932,-0.0352098681,0.1236102059,0.0856775269,0.4375536442,0.1367622465,-0.2987443209,0.0737567917,-0.168396309,0.1220315397,-0.0198270902,-0.127543658,0.3879493177,-0.2640428841,-0.0366139896,0.337669462,0.2487759739,-0.0805271938,-0.2423264533,-0.1083673537,0.0431746431,-0.2186345309,0.2081348598,0.1455328614,0.0227565169,-0.2166585624,0.2507808506,0.0271257032,-0.212068215,0.1143487021,0.2771986723,-0.1140616462,0.0634751692,-0.0949055105,0.068479754,-0.1131098643,0.1498427242,-0.1032908782,0.029100474,-0.4956218302,-0.389580369,-0.0401243009,-0.1255233139,0.2596913874,0.0942500085,-0.4623063505,0.2361895591,0.3717957139,-0.0267874207,-0.0050118989,-0.0231809039,-0.1663092822,-0.1284462214,0.2513446212,-0.3950454295,0.0899802372,0.083878696,0.234011814,-0.2627492249,0.1831223518,-0.5330997109,-0.0528200231,-0.3341145515,0.1367405653,-0.3619538248,-0.2586230338,0.2134346515,0.3897889555,0.1823923141,0.1123305187,-0.4024713635,-0.3335260153,-0.0408478267,0.0598119162,0.0134127419,0.2518759072,0.3463186026,0.0766262189,-0.2939297557,-0.2102373391,0.4332271516,-0.4351553619,-0.0284865554,0.0468369126,0.1445864588,0.017490752,-0.1418463141,-0.1096979305,0.1319603026,0.069185026,-0.0779290646,-0.0397776403,-0.280892849,-0.1149958819,-0.1823314428,0.4243626595,0.1585606337,0.1382246017,0.0737145916,0.1361849904,-0.250330627,0.2441385835,0.246151045,-0.3611325026,-0.1352060288,0.1034217402,-0.0021865931,0.4197442532,-0.2955122292,0.2833989561,-0.1688549221,-0.252312839,0.3496871591,0.3437298238,0.7140422463,-0.2160160244,0.1593971997,0.0191928782,-0.0988915712,-0.2327389568,0.0466485582,0.2293528914,0.0145184984,0.061266318,0.1720989496,-0.0708465055,0.1064439639,-0.0608466007,-0.174925372,0.2813668251,-0.0850485414,0.0577613972,0.1048266962,-0.0926742256,0.4100750983,0.2115384638,0.0777445808,0.1147978231,0.0403574817,-0.0528895445,-0.0495488495,-0.1817494929,-0.438192755,0.2212312073,-0.2966368496,-0.041611921,-0.0835660249,-0.1663713306,-0.2637891769,0.0549680404,-0.0012559416,-0.042954661,0.0015603114,-0.1181472465,0.0465965383,-0.1638748795,-0.1439262033,0.1463808268,0.3316173255,-0.2232591808,0.1344809532,0.1284247935,-0.143198967,0.2475660741,0.541010499,0.3279054761,-0.2567958236,-0.369956255,-0.0236471407,-0.3931616545,0.1143628582,-0.1911545545,-0.0090124113,0.0593721457,0.1580133438,0.3008077443,0.2331299782,-0.257209599,-0.0355623811,-0.0106464559,0.0515799746,-0.3458046019,0.1545700133,0.1072595641,0.1238118857,-0.1845238656,0.0931959003,-0.3634851873,-0.004605866,0.1393557191,-0.0707540289,-0.0176638495,-0.1975844502,0.1898914725,0.1384199262,0.3405684531,0.1527792364,0.0655080751,-0.1138006896,-0.2793235183,-0.5851952434,0.2916977406,0.3878126442,0.0169865079,-0.0012361372,0.1820773333,0.2382829785,0.23461923,-0.3237170875,0.2201520801,0.0310054943,-0.2409292608,-0.3662896752,0.0083114989,-0.1913269311,0.0332180038,0.0994150341,-0.1449591368,0.0379742272,-0.1285688877,0.1693518609,0.0261826944,0.1942167431,0.0283026341,0.2376377434,0.2198460102,0.239895612,0.0573148504,0.0239224732,-0.2906651497,-0.1991188079,-0.1112013906,-0.1484753937,-0.0114264535,0.026457997,0.2559542656,0.0933018774,-0.0391783863,-0.0014736926,0.03460069,0.1273926347,0.0743485242,-0.0746531412,0.295666188,0.1225992367,-0.0058943746,0.2172319442,0.5124965906,0.0587392338,0.0064912066,-0.1283379793,-0.4462196827,0.2543745637,-0.2712899148,-0.2026142329,-0.0249022543,0.2888183296,-0.1842331737,-0.1248505488,-0.4972246885,0.1747484058,0.2441172004,0.156426996,0.2203083038,0.4398518801,0.0314947292,-0.1901859939,-0.1355281919,0.0718063042,0.1401340365,-0.2506893277,0.0180959124,-0.1105015352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/816","title":"[Caching] Dill globalvars() output order is not deterministic and can cause cache issues.","comments":"To show the issue:\r\n```\r\npython -c \"from datasets.fingerprint import Hasher; a=[]; func = lambda : len(a); print(Hasher.hash(func))\"\r\n```\r\ndoesn't always return the same ouput since `globs` is a dictionary with \"a\" and \"len\" as keys but sometimes not in the same order","body":"Dill uses `dill.detect.globalvars` to get the globals used by a function in a recursive dump. `globalvars` returns a dictionary of all the globals that a dumped function needs. However the order of the keys in this dict is not deterministic and can cause caching issues.\r\n\r\nTo fix that one could register an implementation of dill's `save_function` in the `datasets` pickler that sorts the globals keys before dumping a function.","comment_length":43,"text":"[Caching] Dill globalvars() output order is not deterministic and can cause cache issues.\nDill uses `dill.detect.globalvars` to get the globals used by a function in a recursive dump. `globalvars` returns a dictionary of all the globals that a dumped function needs. However the order of the keys in this dict is not deterministic and can cause caching issues.\r\n\r\nTo fix that one could register an implementation of dill's `save_function` in the `datasets` pickler that sorts the globals keys before dumping a function.\nTo show the issue:\r\n```\r\npython -c \"from datasets.fingerprint import Hasher; a=[]; func = lambda : len(a); print(Hasher.hash(func))\"\r\n```\r\ndoesn't always return the same ouput since `globs` is a dictionary with \"a\" and \"len\" as keys but sometimes not in the same order","embeddings":[-0.081926316,-0.0495188832,-0.0879312232,0.1216140687,0.0892920494,-0.156362012,0.1671694219,0.2433263063,-0.0996844098,0.0803640485,-0.0827869698,0.2674162388,-0.1859903485,-0.2689157426,0.0077195074,0.3598428965,-0.0797575489,0.0291419644,-0.4379902482,-0.2746317387,-0.2619422078,-0.0908107534,-0.120361276,-0.1932392567,-0.421674639,0.001889275,-0.0498695895,0.2467167825,0.0397862568,-0.3421670198,-0.1833326221,0.1572594643,0.0308243632,0.189659521,-0.0001114966,-0.1850806028,0.2640141845,-0.0081459675,-0.2726581693,0.0902210996,0.0154270325,0.1384180486,-0.0409753881,-0.2122771889,0.1143219844,-0.1681281775,0.0286574792,-0.6539846659,0.032668259,0.0999311507,0.277249068,0.1996293217,-0.1389131397,0.1366502345,0.3288570344,-0.0815244392,-0.0022259969,0.2885014713,0.0104087703,-0.0936940387,-0.307117939,0.2813210189,-0.1913660467,0.0808453411,0.3518430293,0.1796078533,0.0172839649,0.0677473098,0.1912310272,0.0597434379,0.3375328183,-0.5688877702,-0.4461806715,-0.2534594238,-0.1146053672,-0.0763154179,0.1631677449,-0.1536449641,0.1536260694,0.1698363125,-0.1345395297,0.0991548821,0.2495270967,0.0985691771,-0.0092676794,0.1577008814,0.067767933,-0.0419676155,0.0160904657,-0.1232828647,0.2864695787,-0.2080323696,0.1085584015,-0.0559339486,0.016890632,-0.0224733464,0.3788667619,0.1954948902,0.1477339119,0.0638351738,0.0911780298,0.0182211325,0.1135687977,0.0352341868,0.1287693828,0.2473442107,0.1520804614,0.357517302,0.3471331298,-0.5027477145,-0.3569695652,0.1133540645,0.286718756,-0.0380505733,0.6074771285,0.1303846091,-0.0743750557,-0.0495635979,-0.031439364,0.32664904,-0.0200539809,-0.2020178139,0.2717580199,0.4664558172,-0.3419175148,0.1982015371,-0.0471672416,-0.2786376774,-0.1859294325,0.3082937896,-0.371824652,-0.1541631967,-0.140581131,0.2501672804,0.0095407413,-0.132905364,0.1580966711,0.1285112649,0.2983812392,0.1937909126,0.0873249173,-0.3041604757,0.3930503428,0.1353822351,-0.0953307822,0.2159696817,0.047216665,-0.0636445656,-0.4521884024,0.0486170091,-0.0159860663,-0.2103940994,-0.0230899435,0.2847323716,-0.2386864275,0.0968033373,0.1912357807,-0.2342095971,0.2490407377,-0.1804463863,-0.0749651641,-0.3261179626,-0.325134933,-0.4191172719,0.0944070593,0.4102349579,-0.3341307342,-0.0940667167,-0.1776915193,0.259762466,0.14738594,0.1098356172,-0.0282693114,-0.0454402789,-0.3166434467,0.2087910026,0.0650046766,-0.3961066008,-0.5688530803,0.0614080541,-0.1132695153,0.232750237,-0.0886431858,0.1138274446,0.4595196545,-0.0851631612,0.1303600222,0.2083604187,0.2607237995,0.2012596577,-0.3684207201,-0.0376232415,0.0532560572,-0.0696235448,-0.0592403524,-0.0664013252,0.1056605056,-0.1016225219,0.0801990852,-0.0615177713,0.0204895474,0.1380243897,0.3963465095,0.087246865,0.2272403538,-0.0892871395,-0.6030647159,0.3717261553,-0.0349171273,-0.4906652868,-0.2042220831,-0.3759244382,0.1311759353,0.1593645662,-0.033455722,0.0759956986,0.0885329396,0.3169272244,0.2658632994,0.2659104466,0.0573767908,0.131419003,-0.1487334222,-0.2696956098,0.0122145247,0.046974726,-0.1958996803,-0.2563337386,0.0352156013,0.001453958,0.042727299,0.0691585541,-0.1159347668,0.2138011605,0.1877532899,-0.0765466914,0.0036100738,0.3865779936,-0.1713905782,-0.0223198179,0.2390533537,0.1847987026,0.0641399324,-0.1483221352,-0.1509419531,0.6806604862,0.117683731,0.0892539471,-0.0601302199,-0.1575537622,-0.0774331987,0.0911973268,0.0173605606,-0.4390444756,0.0523103252,-0.1941667348,0.6369011402,0.3142585456,-0.2441952229,0.4189313948,0.6483618617,0.0400007777,0.239436999,0.1224158034,0.0175779648,-0.2655003667,0.1121724993,-0.1683278084,0.4186325371,0.2538229525,0.1602876782,-0.0879485831,0.0060967859,-0.1408472955,0.0432099923,-0.2433653921,-0.0574139655,0.2297235429,0.1816812009,-0.0787309483,-0.1304650009,0.0715684369,-0.0922231972,0.1561158746,-0.1333819628,0.0872975811,-0.4846898913,0.1538639814,0.0726690218,-0.1994608939,-0.2668467164,-0.2821635902,0.1082778648,0.0481430776,-0.1991004497,0.1922544241,-0.2088033408,0.1449298859,0.205556795,-0.322463274,-0.2208684832,-0.1914170086,-0.2088921368,0.0058677318,0.1395238936,-0.1580912024,0.3791800141,0.0502017066,-0.1432881057,-0.2470655739,-0.1958899796,-0.0793744698,0.1651477814,0.0886351392,0.0485241003,-0.2080036104,-0.2918452621,-0.3191621602,0.1711027324,-0.3942877948,-0.3049253523,-0.2680914104,0.1064395681,0.0693447366,-0.0969776213,-0.0972774997,-0.101793848,-0.2493078709,-0.1194640025,0.0802558288,0.2488237768,-0.0363949165,-0.1088806763,-0.0146394297,0.0594106242,0.1286454201,-0.4220730662,-0.3723148406,0.3654733002,-0.240479961,-0.201571852,-0.2894231677,-0.2820841074,-0.2084700614,0.1274233311,-0.4713217318,-0.4080052674,-0.2187334597,0.3893478811,0.1095756665,0.0051509412,0.1764832735,-0.0399254449,-0.0563617945,-0.2230707556,-0.2313829958,0.0313620791,-0.0464049615,0.1364597231,0.026224792,-0.0058429455,0.0798951909,0.8998237848,-0.0113259749,-0.3779034317,0.0146307014,0.1756103486,0.3247360289,-0.2775584161,-0.2364738733,0.1491016001,0.0812693536,0.1021386832,0.3682431579,-0.0651276335,-0.2316619158,0.0626738667,0.2769613862,-0.4035154879,-0.2864977121,0.1481481344,-0.3117341101,-0.1262226254,0.1735555381,0.3351280987,-0.2934386432,0.0429755077,0.0211327281,-0.1494346708,0.2826304138,-0.0590810254,-0.0108063873,0.2990707457,-0.3614232242,0.3434768319,0.2358406931,0.1108579189,0.0882397741,-0.4010428786,0.1434253007,-0.1349484026,0.5495229959,-0.164112702,0.1260450482,0.1582793295,-0.4842356741,-0.095623523,-0.0304003078,0.1184852421,0.1968019605,0.255135864,0.0388445631,-0.3402619362,0.0498503521,-0.0500083528,0.0486910567,-0.0687908605,-0.3214802742,-0.2081477046,-0.0372431763,0.0294931512,0.0547013246,-0.2114487141,0.1144878268,-0.140261054,-0.0664147884,-0.2791723311,-0.2111039311,0.1920104325,0.0755909979,0.3900791109,-0.1225335971,0.414575696,0.2395905852,0.4274957776,0.0329538547,0.1387422979,-0.3923276365,-0.1195821986,0.1017704308,0.0948023498,-0.1582092643,0.3111718595,-0.0407913141,-0.0222082268,0.0554257557,0.0218422879,-0.118723534,0.1140557155,0.0108846733,0.1072777212,-0.3166214526,-0.4029178917,0.2768453956,0.2825394869,-0.1167238578,0.5915082693,-0.1340258867,-0.1336925328,0.193659097,0.4005237818,1.1834952831,-0.2322788835,0.2383719385,0.0206897669,0.1786907166,0.0663913861,0.2275196612,-0.0020093357,-0.4197832644,-0.227865696,0.1450518519,-0.1376361102,-0.0428065322,-0.1463010311,-0.2276194245,0.0268211756,-0.3041321635,0.3562989533,-0.1215877011,0.2969777286,-0.0509468392,-0.0515769273,0.3112762868,0.2058666795,0.1518610269,0.1217871383,-0.0988250002,0.1393097341,-0.1224689633,-0.1717007905,0.0069994819,-0.0000602422,-0.0879221633,0.4257004559,0.0290248413,-0.396663934,-0.4203565121,0.47770679,0.1334199458,0.21346201,0.0355224386,0.1281349361,0.0422874987,0.341768533,0.0404703282,-0.2964973748,0.1377899945,0.1243945435,-0.092903845,-0.0122524546,0.2514988184,-0.4036676586,-0.2988367975,0.4665337205,-0.1979624033,-0.0991262645,-0.0215114616,0.0649689585,-0.1109249294,-0.1502815038,0.1958770752,0.3595410883,-0.0462892167,0.5922560096,-0.2799726725,-0.1550388634,-0.2139163762,0.6729285121,0.2095045596,0.3016417921,0.0368688405,0.0176520031,-0.4765994549,-0.1334564239,-0.1271628886,-0.422075361,0.1366964132,0.4294389188,-0.0695429817,-0.0391993374,-0.1124983132,-0.0852755532,-0.0871281847,0.1773782223,-0.3132680357,-0.0296300445,-0.1486731321,-0.0450386927,-0.1579332203,0.2434314936,-0.0159741193,0.138918817,-0.3936751187,0.333286196,-0.405120492,-0.1230917424,-0.2080714405,0.0440106839,-0.1888354719,-0.2565595806,0.1808721721,-0.1447148323,0.2279924601,0.275018841,-0.1062995121,-0.2835400999,-0.0280256215,0.0919839218,-0.0844925269,0.1104622856,-0.0389811657,-0.0929505751,-0.007598334,-0.3970496953,0.2798281312,0.5578253865,0.030945085,0.0972242504,0.0877612457,-0.1142396778,0.3257340193,0.2858262956,-0.2543868423,0.1293895096,-0.2279172987,0.331422776,-0.0536840819,-0.082950756,0.3215911686,0.0306039117,-0.0861040428,-0.1676231772,0.2974962294,0.0400652811,-0.2399720699,0.238781631,0.4261125922,0.1939098537,-0.0205077417,-0.4133689404,0.3929366171,0.2304688841,-0.1804399937,-0.3763190806,0.0591882616,0.0784869567,0.1060817391,0.1327662021,-0.089582473,-0.0144724315,-0.0005604242,-0.044650808,0.231490612,-0.1061653048,0.149729237,0.1725641191,0.3157944679,0.0230832696,0.4607911706,0.1874253303,0.1809663773,0.3305386305,-0.1241100952,0.1913288683,-0.182096079,0.0102817388,0.008898397,-0.5474196076,0.4308394194,0.3629772067,-0.1097008511,0.2956851721,0.0735180676,0.1832118928,0.0164281446,0.1066217497,-0.1848502457,0.1507455111,-0.2467526346,-0.0344877504,0.1113545522,-0.0215220414,0.1639777571,0.0704097301,-0.0889811292,-0.2862012684,0.3403049111,0.1530380249,-0.0340164304,-0.2817849219,0.2028592229,-0.3229221106,0.2571031153,-0.17045632,-0.0473237112,0.1258998215,0.1614992321,-0.1658188105,0.1366126537,0.2450376749,0.5118284822,-0.0608608983,0.106567286,-0.0077795233,-0.0586204752,-0.4952119589,0.343331188,-0.312584877,-0.3157385886,0.3530361056,0.2092964649,-0.2533220053,0.0642441064,0.0696771294,0.2718545794,0.2033101469,0.1590228826,0.0150677124,0.1130046621,0.0240574293,0.2794403732,-0.2663087249,-0.2240269035,0.3107811809,0.0132212937,0.1472039223,-0.2555626035,0.0890769884,-0.1660210639,0.3152881563,0.3831397891,0.3086492419,-0.1163453162,0.0531297773,-0.3521826267,0.3837222457,-0.1386062354,-0.0015883838,-0.0800752565,0.4194467366,-0.1131014526,-0.1083755121,0.0712577477,0.3192202151,-0.0017824856,0.3947654366,-0.4328179955,0.1371633857,-0.5196845531,-0.2913024127,0.139479056,-0.5391646624,0.1958771646,-0.0349057727,0.2048135549,-0.2840933204,0.1139306948,-0.0007824018,0.163715288,0.5047900081,0.3265343308,0.2440895885,-0.1214174852,0.0880948678,-0.0405598208,-0.100420624,-0.0561725423,-0.2300751209,0.0645581782,0.1127056703,-0.3109338582,-0.3468679786,-0.1534579843,0.6476136446,0.0630400404,-0.0504759364,0.0664480627,-0.0019149266,-0.2814874053,0.0068398044,-0.191513747,0.3433923125,0.0192119163,0.1441071928,-0.2024853379,-0.3411545455,0.6150742769,-0.0810100362,0.0345462933,-0.2443350852,0.2873256803,0.2199004143,-0.1681070477,-0.236532867,-0.15946953,0.29433465,0.1065888628,-0.1817377806,0.0841092169,0.0291775782,0.2167295069,-0.1901470125,-0.06646882,-0.034695521,-0.2794094086,0.1684294492,-0.350638926]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/815","title":"Is dataset iterative or not?","comments":"Hello !\r\nCould you give more details ?\r\n\r\nIf you mean iter through one dataset then yes, `Dataset` object does implement the `__iter__` method so you can use \r\n```python\r\nfor example in dataset:\r\n # do something\r\n```\r\n\r\nIf you want to iter through several datasets you can first concatenate them\r\n```python\r\nfrom datasets import concatenate_datasets\r\n\r\nnew_dataset = concatenate_datasets([dataset1, dataset2])\r\n```\r\nLet me know if this helps !","body":"Hi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks","comment_length":67,"text":"Is dataset iterative or not?\nHi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks\nHello !\r\nCould you give more details ?\r\n\r\nIf you mean iter through one dataset then yes, `Dataset` object does implement the `__iter__` method so you can use \r\n```python\r\nfor example in dataset:\r\n # do something\r\n```\r\n\r\nIf you want to iter through several datasets you can first concatenate them\r\n```python\r\nfrom datasets import concatenate_datasets\r\n\r\nnew_dataset = concatenate_datasets([dataset1, dataset2])\r\n```\r\nLet me know if this helps !","embeddings":[-0.2620126605,-0.2916283309,-0.1804153472,0.1139712036,0.0606512427,-0.0265371073,0.2819753587,0.1687740684,0.1109021083,0.0845518932,0.1554597318,0.2481859922,-0.3554365635,0.348952949,0.1711002141,-0.0656197593,0.0205348637,0.136970818,-0.6233143806,-0.0821494088,0.0825396925,-0.1274391562,-0.2666273415,-0.1741281897,0.0523192361,-0.3682792485,-0.3957517743,-0.142244488,-0.2242219597,-0.3630472422,0.2143723816,0.2408522069,0.4066430032,0.554397583,-0.000103148,-0.1346856058,0.1404353678,-0.0824473649,-0.1723602265,-0.2457790971,-0.1541205347,-0.2673062086,0.0849707127,-0.4197580814,-0.0327179879,-0.1310338676,0.0873626098,-0.3289755881,0.2696433663,0.1327248216,0.2795183361,-0.2198545337,0.1302630305,-0.0211496111,-0.031763576,0.0726195425,-0.0583847053,0.220533058,0.1717952639,0.1565460712,0.121111922,0.028726371,0.1655883044,-0.003596331,0.3876419663,-0.0965651274,-0.0442575067,-0.4276583195,-0.1730998605,0.0660026446,0.8125045896,-0.4412543774,-0.3938426971,0.0490858927,0.0430765748,-0.5168247819,-0.152706787,0.1491913646,0.0641432405,0.0586709566,-0.1348238289,-0.2621481419,-0.2698768973,0.1240683198,-0.1202691644,0.1766415834,-0.1396183521,0.0977383852,0.1554127485,-0.1787713915,0.3191452026,-0.0058359746,0.1843648553,0.2849902213,-0.56351614,-0.3246938884,0.0088891173,0.3077014387,0.3426326215,0.3845371902,0.0610330328,-0.0531119891,-0.1638740897,0.2153864801,0.5052185655,-0.0575832464,-0.0967656747,0.0773621872,0.1371305436,-0.074623026,-0.1787652373,-0.0850120857,-0.2036229223,-0.2351184189,0.1354383528,-0.1673399955,-0.0108033558,-0.0898151025,-0.0006082836,-0.2141327262,-0.2479062378,-0.2665740252,0.1642467827,0.235846296,-0.2475819737,0.1832304746,-0.271872133,-0.1247223839,0.0381813161,0.2522245646,-0.1004244685,0.4507007003,-0.3770989776,0.0321811177,-0.0641270578,-0.3635902405,0.0666291267,0.0946309268,0.1086149588,-0.0891537145,0.1355137974,-0.2728854418,0.3756600916,-0.003879535,-0.0490063578,-0.0022548011,-0.1385650188,0.0565218776,-0.4262951016,0.0191203542,-0.3938416243,-0.0465501696,-0.0372907184,0.2154168636,-0.1831136048,-0.2176530361,0.079800874,0.1534947306,0.0734745041,-0.112355262,0.0158816241,-0.1077264473,-0.0338811763,-0.2993712425,0.0883344263,0.1641410291,-0.35705176,0.0776655599,0.0804291293,0.0491958223,0.1362267435,0.2218888253,-0.2568956912,0.2136597186,-0.212691471,-0.0707973689,0.5765584707,-0.2613782585,-0.1435530037,0.2921441495,0.0746691748,-0.0027985352,0.1206455603,0.1994356066,-0.0524061844,-0.0169341825,-0.1686793566,0.2380015701,-0.011463956,-0.1006743237,-0.0052550002,0.1009077653,0.1932469457,0.0401523896,0.0893867463,0.2351981401,0.0582371019,0.1127639264,0.455421716,-0.2984415591,0.108294256,0.2104291022,0.3090749681,0.114899829,0.1462791264,-0.3679890633,0.1484366357,0.0986365899,0.2785969377,0.0520479679,0.2340695709,-0.069348231,-0.0442877933,0.0764071643,-0.2159165442,-0.0578871146,0.1884937882,0.1673408449,0.1742748469,-0.1189610735,-0.5899415016,0.218558684,-0.2163225114,-0.118447803,-0.5084574819,0.2147316337,0.1542232484,-0.0108168935,0.0965031907,0.116957441,-0.1846950948,-0.434378773,0.2691549659,0.2770812511,0.0441000424,-0.0713122487,0.0792737752,0.3585073054,0.2771754563,-0.1215388402,0.0665503368,-0.1913222969,0.3170205057,-0.0304633435,0.244586587,0.4754381776,-0.1599055827,0.1892339438,0.4515470266,-0.1398740113,0.0697892085,-0.0783411115,-0.3781239092,0.1251234412,-0.1108881757,-0.164718017,0.432398051,-0.1913584173,-0.4387002289,-0.2600629032,0.4845893979,-0.0010402949,-0.0046966672,0.0654750764,-0.073386617,0.0072842892,-0.0527888536,0.1374273598,0.3874976337,0.0551788807,0.2038333267,0.0640795827,0.0119077144,0.1259277463,0.1537786573,0.0687423125,0.3130619526,-0.0194043946,0.0347604007,0.0564903691,-0.163587302,-0.3934365511,-0.0763192996,0.1618542373,-0.021135876,0.3687207401,-0.3297156096,-0.3494110107,-0.3988712132,-0.2384474277,0.1140254661,0.0361569747,-0.2951516509,-0.0410355069,0.3395088315,0.1355220973,0.14926745,0.1712168604,0.0856718794,0.0247138273,0.0559491813,-0.0249677356,-0.0153802279,0.1965170205,0.2742664218,-0.1589580178,0.5873372555,0.0007301636,-0.0804165229,-0.3680339158,-0.1861910671,0.1703963727,-0.0128429551,0.0351405255,0.2513346076,-0.0045819697,-0.048999127,-0.0918228626,0.0235354751,-0.3367699981,0.1729476899,-0.2204838693,-0.1017430574,0.1009223163,0.0909778923,-0.2924436629,-0.5083232522,-0.2475090325,0.2429561317,0.4355327189,0.2685111463,0.1604335457,0.0689342543,0.2123544365,0.1988351494,0.0773472935,0.0226623565,-0.299783498,0.18126297,-0.3456683755,-0.2615365088,-0.0005994889,-0.0124733187,0.4389668107,0.1449292451,-0.2703377306,-0.1131626964,0.0078622205,-0.016493693,0.069344826,0.5041059852,0.1270898581,-0.2268275321,-0.083556585,-0.0664393157,0.0531284325,-0.066472061,-0.143053785,0.2096247226,0.4548493922,0.3592880964,-0.0460021123,0.3433163762,0.2341688424,-0.4142976105,0.3587226868,-0.3138966858,-0.1621472687,-0.0567742474,-0.2992808819,0.0955502167,-0.328994751,-0.2866768241,0.0765292943,-0.1328617334,-0.3404943645,-0.3377614617,0.1142513901,0.0144964531,-0.0170974787,-0.0081660086,-0.3336505294,0.252032578,-0.2462818176,-0.18163459,-0.1159571409,0.0789188519,-0.0262018312,0.0999785364,0.1092674658,-0.1281919181,-0.5397815108,-0.1871839911,-0.2831958532,0.0810429528,-0.109438099,0.2423329353,0.3701023757,-0.1744752973,0.0719158426,-0.3205095828,0.3552826941,-0.0110201621,-0.5156714916,0.0702781156,-0.4106995165,0.0827037022,0.1170744598,-0.3059648573,0.522639513,0.1802407801,-0.0700212419,-0.1232783645,-0.2890223861,0.3584613502,0.1638960242,0.0033585851,-0.2747871578,-0.0254922658,-0.2516489625,-0.2371177822,-0.1153583676,-0.1130126119,0.0015419454,-0.0715376064,-0.0634526834,-0.2532050014,0.0591780655,0.2296279818,-0.0585731789,0.0305904839,-0.0187710207,0.1887828261,0.0589883141,0.3323000669,0.3582285345,0.3531622887,-0.1257246882,-0.0742667988,0.1878255308,0.2395172268,0.4375710785,0.0094461404,-0.0741961598,0.1853927821,-0.2198527902,-0.0161875673,-0.3805890977,-0.0386221632,0.3348107934,-0.001857016,-0.6414475441,-0.8713588119,0.5688750744,0.2613633275,0.0040651578,0.3360829651,-0.1226273626,-0.1023518741,-0.0902932733,0.1467006803,0.6982848048,0.0836788937,0.0944305733,0.1957822144,0.291090399,0.4742543101,-0.0231766347,0.0957147703,-0.4712153971,-0.2424999326,-0.1576360315,-0.1352897286,0.3274860978,0.0973742157,-0.2673665583,0.4617575407,0.08613199,-0.0191617776,-0.0781333894,0.2627310455,-0.2957575619,-0.1491945088,-0.3420797288,0.2087102383,0.1375456452,0.1305044591,-0.1679983288,-0.1962712556,0.3250442743,0.1217645556,-0.025997797,-0.1470038146,-0.1014485061,0.0157757513,0.1271650642,-0.3621403277,0.2924298346,0.1912252456,0.2234527022,-0.0133158909,-0.0412635729,0.0281855669,-0.1452393085,-0.0371000804,-0.0228268225,0.1572776586,0.5331594348,-0.053782966,-0.0648752898,0.2946961224,0.0965827182,-0.0767126232,-0.0837324783,0.2173759937,0.2105670124,-0.1378240883,-0.1368601173,0.4520658851,-0.1138685569,-0.073050037,0.1984200627,0.1927951127,-0.2502839267,0.4345503449,-0.0265247449,-0.1206346825,-0.1684548557,0.1747996509,0.2953974307,-0.1548342258,0.2711715996,0.0277815424,-0.2030160576,-0.2583884597,0.1134043932,0.1887515187,0.0470324121,0.0357752107,0.0563853048,0.3641643226,0.2256297618,0.1616132259,-0.004046035,0.0757428706,-0.3522455394,-0.1699042916,-0.2105127126,0.0448403507,0.1735902876,0.3234499991,-0.0506315939,-0.0821368769,0.1715502143,0.0824963003,-0.3810793459,-0.0656668842,-0.1518850178,0.1120758206,0.0901951641,0.0344073623,-0.2411402762,0.3060914278,0.1729075909,0.1513153315,-0.2065102607,-0.2799615562,-0.0027321228,0.0741402432,-0.0406979322,-0.0506379232,0.1484757066,-0.1919181496,-0.2478047162,-0.083961308,0.1618690342,-0.0411012024,0.004889606,0.215501219,0.1861739308,-0.38000983,-0.0785850585,-0.0580966212,0.1399797499,-0.0120565416,0.1560516506,0.3166588843,0.0141459042,-0.2948203683,-0.2338852733,0.1436720341,0.2659362555,-0.1882646829,0.3587215543,-0.2616411448,0.1691550165,0.0873796567,0.4440041184,0.6819087267,-0.1831477731,0.1529594809,0.2660349905,0.2514120042,-0.131991148,-0.064438425,-0.0242324378,-0.09626735,0.0915644616,0.2693862021,0.1511513293,0.112254791,-0.2158710957,-0.2792433798,0.1278318167,-0.2928752005,0.1434127092,0.299303174,-0.0822099969,-0.112462461,0.2549159527,0.4423461556,0.1981951296,0.5453619361,-0.0141148241,0.6302165985,0.106754981,-0.1112340987,-0.0250927638,-0.1928500533,-0.0242173672,0.207698077,0.1615584046,0.0700048134,-0.031154247,0.2256401777,0.1823238134,-0.1027310342,0.0183735397,0.3125155866,0.082324028,-0.1238415241,-0.2100811601,-0.1635658294,-0.1809418797,-0.0708314106,-0.1897087842,-0.1125659421,0.2526173294,0.0850832686,0.2983179092,-0.4108554125,0.3159436882,-0.1058568954,0.3703703582,-0.2970692813,0.2014585882,-0.0724241659,0.1050315872,-0.3055816591,0.1716096103,0.3079659939,0.4848974347,0.1029262319,0.0477106087,-0.2340180576,-0.1359864771,0.1141855493,0.0032014346,-0.0817966908,-0.051416643,0.4961348176,0.1336660087,-0.0962805972,0.4675194621,0.0762164444,-0.0275229905,-0.8000189066,0.0478105731,0.083297722,-0.0378124453,-0.0930335671,0.2179276496,-0.4655721784,-0.0468703248,0.4806190133,-0.4441665709,-0.0073589226,-0.1756755859,0.1024015024,-0.3210001886,0.6157894731,0.1939952075,0.5805901885,-0.0145388041,-0.3817857504,-0.3062667847,-0.0336714759,-0.0866715163,-0.1495523602,-0.0763069019,0.1704975665,0.1077193767,0.0987849459,-0.1735310107,-0.4373109043,-0.2029021829,0.4054436684,-0.2589812279,-0.244092077,-0.4298059642,0.3638330698,0.0726839006,-0.1971220225,-0.0425106511,-0.4353247583,0.0534014255,0.0012334511,0.1778694838,0.1927584261,-0.2071614116,0.4288679063,-0.0192674045,0.3466538191,-0.1509357393,0.0377650969,-0.1243556887,-0.1971428543,-0.1997792274,0.3828832805,0.1541639268,0.0247666501,-0.0930310413,-0.0408334136,-0.0006952345,0.0192726981,0.0890781805,0.0188278239,-0.0209812932,-0.2271886766,0.1004725099,-0.0382760838,0.0966302976,0.6581110954,-0.1470127553,0.0655413792,-0.201807484,-0.0885169506,0.2630481422,-0.1844277829,-0.1997631192,-0.3186928034,0.0300829858,0.0652605593,-0.1534787565,-0.447735101,0.0825767964,0.4290879071,-0.0004237942,-0.2846854627,0.3960919678,0.3302383423,0.0554769561,-0.0614908971,0.1827528477,-0.0785411,-0.1147850603,-0.0152916526,-0.2455781847]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/815","title":"Is dataset iterative or not?","comments":"Hi Huggingface\/Datasets team,\nI want to use the datasets inside Seq2SeqDataset here\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py\nand there I need to return back each line from the datasets and I am not\nsure how to access each line and implement this?\nIt seems it also has get_item attribute? so I was not sure if this is\niterative dataset? or if this is non-iterable datasets?\nthanks.\n\n\n\nOn Mon, Nov 9, 2020 at 10:18 AM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> Hello !\n> Could you give more details ?\n>\n> If you mean iter through one dataset then yes, Dataset object does\n> implement the __iter__ method so you can use\n>\n> for example in dataset:\n> # do something\n>\n> If you want to iter through several datasets you can first concatenate them\n>\n> from datasets import concatenate_datasets\n> new_dataset = concatenate_datasets([dataset1, dataset2])\n>\n> Let me know if this helps !\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/815#issuecomment-723881199>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ARPXHHYRLSSYW6NZN2HYDBTSO6XV5ANCNFSM4TPB7OWA>\n> .\n>\n","body":"Hi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks","comment_length":185,"text":"Is dataset iterative or not?\nHi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks\nHi Huggingface\/Datasets team,\nI want to use the datasets inside Seq2SeqDataset here\nhttps:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py\nand there I need to return back each line from the datasets and I am not\nsure how to access each line and implement this?\nIt seems it also has get_item attribute? so I was not sure if this is\niterative dataset? or if this is non-iterable datasets?\nthanks.\n\n\n\nOn Mon, Nov 9, 2020 at 10:18 AM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> Hello !\n> Could you give more details ?\n>\n> If you mean iter through one dataset then yes, Dataset object does\n> implement the __iter__ method so you can use\n>\n> for example in dataset:\n> # do something\n>\n> If you want to iter through several datasets you can first concatenate them\n>\n> from datasets import concatenate_datasets\n> new_dataset = concatenate_datasets([dataset1, dataset2])\n>\n> Let me know if this helps !\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/815#issuecomment-723881199>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ARPXHHYRLSSYW6NZN2HYDBTSO6XV5ANCNFSM4TPB7OWA>\n> .\n>\n","embeddings":[-0.0225438979,-0.3792024851,-0.071719721,0.3793151379,0.0794595927,-0.1549821049,0.2763752937,0.0828172937,0.0711623803,-0.1428877562,-0.064362593,0.0111193135,-0.2454968393,0.3429043889,0.1579365581,-0.0702992827,-0.1000306979,0.0074630883,-0.7639597058,-0.0446302965,0.0603365041,0.2321568131,-0.2711442709,-0.0109251523,-0.0369475,-0.2411496639,-0.326510787,-0.062773332,-0.2232872099,-0.3062738776,0.3705847859,0.2676627934,0.3039007485,0.5059751868,-0.0001149827,-0.0865792185,-0.0136720212,-0.1119051799,-0.260540694,-0.2602420747,-0.2470697761,-0.1210107729,0.1278484464,-0.3559494913,-0.04556977,0.037800625,0.1667557806,-0.3080901802,0.6174594164,0.1739711314,0.163666904,0.1257338077,0.0991589129,-0.0202110875,-0.0610699952,0.2306855172,0.0680697858,0.1013273224,0.2460479736,0.161670208,0.0873173177,0.1419124752,0.1690963805,-0.1322489083,0.2284310758,0.0251351129,-0.1097571179,-0.3999998271,-0.0769667327,0.2307674885,0.7655302286,-0.4036310911,-0.493411839,-0.2389011681,-0.0845850855,-0.3491231799,-0.0607267022,0.087834686,0.061579302,0.2074181736,-0.2750288248,-0.1799227744,-0.2871975899,-0.0714753866,-0.0842036381,-0.1233078465,-0.3571073711,0.0375433192,0.1368417293,-0.2389593422,-0.1284435242,0.1822870374,0.111900121,0.3583672643,-0.6574568152,-0.1644616723,0.0324652568,0.2509856522,0.5292328596,0.5199728608,0.0708161891,-0.0728314742,-0.059563607,0.1355994046,0.3547314107,0.2167765945,0.1332976073,0.014836791,-0.0468348674,-0.0156760421,0.0080921166,-0.1874767244,-0.1975459754,-0.263307482,-0.0288977232,-0.1812646687,0.1243735775,-0.1571443975,-0.0886652917,-0.1423647404,-0.2186679989,-0.2480700761,0.2155747712,0.1797961891,0.0147842159,-0.0092717009,-0.1977039576,0.1596263349,0.0233290289,0.1618883312,-0.08999376,0.293906033,-0.3323544264,0.3278651536,-0.106336765,-0.516843617,0.1241302714,-0.0498067364,-0.0145229921,-0.1895644367,0.0131153511,-0.1561181247,0.2389704734,0.0416877829,-0.135061428,0.1681132317,-0.1038133204,-0.006343523,-0.4026481211,-0.3620334864,-0.2837927043,-0.2724316716,-0.0047699902,0.0828314647,-0.2892344296,-0.0493358523,0.0017668679,0.2995201349,0.042851679,-0.0531391948,0.1576129645,-0.0021380899,-0.2632719576,-0.0274778921,0.1387465,0.5703583956,-0.5171878338,-0.1693527848,0.1901704669,-0.0613867976,0.0647265315,0.4910531044,-0.1845665574,0.2792844176,-0.2950097322,0.1451631933,0.4489835203,-0.1201361865,-0.3379767239,0.3601585329,-0.0514758937,0.1209129244,0.2169084847,-0.0967275575,0.1411954463,0.0771336481,-0.1310061663,0.0307137091,0.101206392,-0.2649412751,-0.0493882522,-0.076473698,0.1218608767,-0.0727327168,-0.0181674603,0.3623619676,0.080653362,0.1890706122,0.4062772989,-0.2153889239,0.2619225383,0.2413795739,0.1841720939,0.1566097438,0.1263941675,-0.3120626807,-0.2084096819,0.1497214884,0.1003217399,-0.0974685103,-0.0270919185,-0.0313607678,-0.2901252508,0.0899075717,-0.2494388521,-0.2793979645,0.1004119292,0.2482919246,0.3779501319,-0.0617013089,-0.4771336019,0.5368739963,0.0398987681,0.1137709469,-0.820933938,0.2269504964,0.2916396856,-0.0419201218,0.0092738755,0.2299204469,-0.0443532504,-0.3739892244,0.4070418477,0.2918508053,-0.2554703653,0.1345612854,0.1751822829,0.413870424,0.2163657695,-0.0493225977,-0.0306187849,-0.3302228451,0.2782608271,-0.0702292398,0.1413688362,0.4477908909,-0.2019180506,0.4001391828,0.2018059194,-0.0399867333,-0.0388885327,-0.0268118605,-0.6704294682,0.1082902253,0.0874225572,-0.1647338122,0.4941927493,-0.13775222,-0.276053071,-0.1087179184,0.4138204455,-0.0944038481,-0.0200498216,0.2225458175,-0.1451559663,-0.0746287853,0.0574685447,-0.0752753913,0.2247612029,-0.0499064773,0.2173964232,0.0197042469,0.152012229,0.0490660034,0.0343685336,0.1375614554,0.2676272094,-0.063083604,0.0223439671,0.0916071907,-0.1821548045,-0.3705839217,-0.1461513489,0.035750553,-0.2060860544,0.4430924356,-0.2589530051,-0.2633575797,-0.4498876631,-0.0666277185,0.0132170264,0.0169128012,-0.3413527906,-0.0877528265,0.3287211359,0.0274921115,0.1229677051,0.3827255666,0.2354795486,0.0622025579,-0.2230383456,-0.0625786334,0.140451476,0.0504576527,0.1174330488,-0.1744202971,0.4147769511,-0.0110876467,0.0203940421,-0.3350255191,-0.2326262146,0.3629404604,-0.0303245783,0.2029055059,0.2526981533,0.0050781057,-0.0727120712,-0.2006012499,-0.0021753481,-0.1680450141,-0.024172185,0.0551481694,-0.0846828222,0.222228542,0.1161014065,-0.2689585388,-0.3376810253,-0.1086261794,0.7068123221,0.0823258534,0.1625310481,0.2160707414,-0.1467021406,0.3742017448,-0.1792030782,0.2425270379,-0.0822483376,-0.4794171154,0.2723617554,-0.3086797595,-0.2070455551,-0.0615933686,-0.111495778,0.5452197194,0.0474165976,-0.2879161239,-0.2418958098,-0.0889611468,-0.0576903969,0.1043202877,0.2547335625,0.3147212863,-0.218001619,0.0530307032,0.0392929614,-0.0398627333,0.1431075484,-0.029665783,0.1179028824,0.3721633554,0.6116313934,0.2033930123,0.5283328295,0.5597794056,0.0441531688,0.1339010894,-0.203017965,-0.1482434124,-0.0539473034,-0.4497843087,0.1008930057,-0.3520355225,-0.224203527,0.1032586768,-0.0820935592,-0.0072762272,-0.1764746159,0.2343404442,-0.3153306842,-0.1751325727,-0.1159481928,-0.0524545051,0.2276574373,-0.1514528543,-0.1010068804,-0.0028498759,0.1166383475,-0.062475767,0.2495337278,0.2291749418,0.1404865086,-0.487324357,-0.119173415,-0.5276031494,0.0991174802,-0.0658045858,0.0816221535,0.1764950901,-0.1703934371,0.2090613246,-0.1972821802,0.7068558335,0.3709081411,-0.4413893223,0.0167110208,-0.6154154539,-0.0944966152,0.2515738606,-0.3018533885,0.2464167178,0.4119485021,-0.1115325391,-0.4199147522,-0.2527623773,0.1832794994,-0.0398676619,-0.061524082,-0.1657912284,0.123714827,-0.1315439492,-0.1779130101,-0.0346643031,-0.0297942627,0.1431575269,-0.0613068119,0.0167554617,-0.1951410174,-0.1009072065,0.2924517095,0.0021703453,0.13552095,-0.0717054009,0.1020827293,0.3970294595,0.1857032776,0.1829592437,0.5120188594,0.0045756991,-0.0163418949,0.2443136424,0.1149492413,0.3713564873,0.2297402918,-0.2243171185,0.1311879605,-0.2082903534,0.0624608733,-0.48769629,-0.0646539927,0.3882619739,0.0042077508,-0.4003357887,-0.5670481324,0.4350757301,0.0636786819,-0.0679373667,-0.1132916436,0.2199296355,-0.21935004,-0.0635295883,-0.0480271801,0.8930822015,-0.0798639953,0.1776343733,0.2711480856,0.3125710487,0.6791244149,-0.1045409292,0.0373732485,-0.3593168557,-0.128375262,-0.176179558,-0.1432954073,0.2393951267,0.2416950762,-0.3168786168,0.4740023017,0.2364821136,-0.0876341388,-0.2001952082,0.1341818422,-0.2505304515,-0.1713343859,-0.5428107381,0.0797572359,0.1976353377,0.3496427834,-0.203543365,-0.1589895934,0.387213856,0.0108768567,-0.1957822293,-0.1961762607,-0.1912960112,-0.1198799163,0.3108175099,-0.4503106475,0.248372227,0.2719883621,0.3074179888,-0.0202801973,-0.233743906,-0.1706657708,-0.1715364009,-0.0575992987,-0.0443540439,0.1191496626,0.5094353557,0.0862159282,0.0098276604,0.2524541914,0.1227700934,0.0552144535,-0.1233928874,0.0279917754,0.0247118436,-0.3799900711,0.0445732921,0.4037043452,0.0138999792,-0.075177148,0.0709157661,0.3548552692,-0.1962822229,0.2054766417,-0.0224687755,-0.1437734514,-0.1721856296,0.3500187695,-0.1183037609,-0.1491501629,0.4499125183,0.2053413242,-0.2025939077,-0.1555675715,0.1078349724,0.3016616404,-0.5539578795,-0.0551479794,0.0158119183,0.295509994,0.1231885776,0.2468172014,0.1575181335,-0.1966639608,-0.1688225567,-0.3144056499,-0.1128086746,0.2498926371,0.1568762064,0.2673359811,0.0133146867,-0.096287027,0.0466276892,0.2190751731,-0.2746019661,0.0008876108,-0.1824958175,0.0270250291,0.3159902692,-0.1311344951,0.1340765506,0.2994025946,0.0049855248,0.2990652323,-0.286002934,-0.1493279785,-0.1430832595,0.1100989133,-0.0432761088,-0.1182494387,0.1256170869,0.0538481325,-0.1396158189,-0.1766372323,0.373131752,0.1063202992,0.2922565937,0.2873691022,0.3195651472,-0.188318193,-0.1875496507,0.0244979057,0.2193179876,-0.0319584869,0.1633568853,0.3162524104,-0.3384663761,-0.2343709171,-0.1301619112,0.4080592096,0.4407764077,-0.2145994008,0.3446981609,-0.3372783363,0.0115616508,0.0162495878,0.5546602607,0.5000157952,-0.1114776582,0.1232729778,0.2385492623,0.1289215684,-0.0131409587,0.1448390931,-0.1646961272,-0.0155051369,0.0392157026,0.3914724588,-0.0636789501,0.1908528507,0.3073041439,-0.1546216011,0.2058760524,-0.2311472744,0.0248272158,0.1402018368,-0.0052893125,-0.0466807783,0.3599811494,0.5259152651,0.2698194385,0.3038617074,-0.1987416297,0.3367417157,-0.2960053086,0.3056201935,0.1545064449,-0.1887047887,-0.2669288814,0.2299317122,0.1297982335,0.0332546756,-0.320492059,0.5874302983,0.2111532986,0.0050562001,-0.0595758073,0.4466045499,0.0898115188,-0.4591878653,-0.0746890679,-0.2179273367,-0.1943539977,-0.1487868279,-0.2237690687,-0.1877889633,0.2310134023,-0.0714439303,0.4418852031,-0.5277567506,0.0358835757,-0.0264716782,0.0527190417,-0.2840577364,0.244479388,-0.0745943934,0.1100014225,-0.0017723112,0.3280362487,0.2134834528,0.1244835854,0.0427425057,0.1268680245,-0.2345499843,-0.10807129,-0.0262487698,0.1887279004,-0.2101010978,-0.1202227697,0.5229051113,-0.0081963819,-0.0220614523,0.2686904669,-0.1294374019,0.0905608684,-0.9172279239,0.2284286767,-0.1415837109,0.0130808856,-0.1373815536,0.094060868,-0.4923450351,0.0472625308,0.1562886238,-0.4805273116,0.1494233906,-0.0902927518,0.0486967452,-0.2270434946,0.4418410659,0.3467173576,0.3968277574,-0.0014269379,-0.1602833271,-0.2863873243,-0.1142125726,0.1662473828,-0.0105512589,0.0262269694,0.3324878216,-0.0263188221,0.0976142362,-0.2443809509,-0.5556641221,-0.4309262037,0.2382337153,-0.1723215282,-0.1835770309,-0.3301457763,0.2728576958,0.0861759782,0.0668267533,0.1457109302,-0.3383136392,-0.096451059,-0.0013659162,0.2898555398,-0.0888065919,-0.2830108404,0.128914848,0.1943982989,0.0226772483,-0.0549044646,0.1512881219,-0.3607619107,-0.3699644506,-0.0088428278,0.2426394969,-0.0101731932,0.2483728379,-0.2000075132,-0.2119711637,-0.0647171363,0.0107861357,0.1808038205,-0.1111468673,-0.0816977695,-0.0053945966,-0.0315476581,-0.0651746094,0.0553358048,0.6036369205,-0.020929357,0.2141369879,-0.0475206301,-0.1800524145,0.4160072803,-0.4784716368,-0.0407220349,-0.0385174677,0.0972358733,0.320835948,-0.4703532159,-0.4702593982,0.1263055056,0.3914020956,0.0702919737,-0.1530092508,0.1964986771,0.2747024298,-0.0594079904,-0.262642771,0.302857995,0.1639413536,0.0323696993,0.1523394585,-0.2388470322]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/815","title":"Is dataset iterative or not?","comments":"could you tell me please if datasets also has __getitem__ any idea on how\nto integrate it with Seq2SeqDataset is appreciated thanks\n\nOn Mon, Nov 9, 2020 at 10:22 AM Rabeeh Karimi Mahabadi <rabeeh@google.com>\nwrote:\n\n> Hi Huggingface\/Datasets team,\n> I want to use the datasets inside Seq2SeqDataset here\n> https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py\n> and there I need to return back each line from the datasets and I am not\n> sure how to access each line and implement this?\n> It seems it also has get_item attribute? so I was not sure if this is\n> iterative dataset? or if this is non-iterable datasets?\n> thanks.\n>\n>\n>\n> On Mon, Nov 9, 2020 at 10:18 AM Quentin Lhoest <notifications@github.com>\n> wrote:\n>\n>> Hello !\n>> Could you give more details ?\n>>\n>> If you mean iter through one dataset then yes, Dataset object does\n>> implement the __iter__ method so you can use\n>>\n>> for example in dataset:\n>> # do something\n>>\n>> If you want to iter through several datasets you can first concatenate\n>> them\n>>\n>> from datasets import concatenate_datasets\n>> new_dataset = concatenate_datasets([dataset1, dataset2])\n>>\n>> Let me know if this helps !\n>>\n>> \u2014\n>> You are receiving this because you authored the thread.\n>> Reply to this email directly, view it on GitHub\n>> <https:\/\/github.com\/huggingface\/datasets\/issues\/815#issuecomment-723881199>,\n>> or unsubscribe\n>> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ARPXHHYRLSSYW6NZN2HYDBTSO6XV5ANCNFSM4TPB7OWA>\n>> .\n>>\n>\n","body":"Hi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks","comment_length":236,"text":"Is dataset iterative or not?\nHi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks\ncould you tell me please if datasets also has __getitem__ any idea on how\nto integrate it with Seq2SeqDataset is appreciated thanks\n\nOn Mon, Nov 9, 2020 at 10:22 AM Rabeeh Karimi Mahabadi <rabeeh@google.com>\nwrote:\n\n> Hi Huggingface\/Datasets team,\n> I want to use the datasets inside Seq2SeqDataset here\n> https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py\n> and there I need to return back each line from the datasets and I am not\n> sure how to access each line and implement this?\n> It seems it also has get_item attribute? so I was not sure if this is\n> iterative dataset? or if this is non-iterable datasets?\n> thanks.\n>\n>\n>\n> On Mon, Nov 9, 2020 at 10:18 AM Quentin Lhoest <notifications@github.com>\n> wrote:\n>\n>> Hello !\n>> Could you give more details ?\n>>\n>> If you mean iter through one dataset then yes, Dataset object does\n>> implement the __iter__ method so you can use\n>>\n>> for example in dataset:\n>> # do something\n>>\n>> If you want to iter through several datasets you can first concatenate\n>> them\n>>\n>> from datasets import concatenate_datasets\n>> new_dataset = concatenate_datasets([dataset1, dataset2])\n>>\n>> Let me know if this helps !\n>>\n>> \u2014\n>> You are receiving this because you authored the thread.\n>> Reply to this email directly, view it on GitHub\n>> <https:\/\/github.com\/huggingface\/datasets\/issues\/815#issuecomment-723881199>,\n>> or unsubscribe\n>> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ARPXHHYRLSSYW6NZN2HYDBTSO6XV5ANCNFSM4TPB7OWA>\n>> .\n>>\n>\n","embeddings":[-0.1425883025,-0.3471777141,-0.1183143407,0.351855129,0.0803463459,-0.0877772421,0.223645106,0.2374293953,0.0332362615,-0.1268832535,-0.0385066755,0.0500974245,-0.2673036456,0.3630913198,0.0759754926,-0.0330848284,-0.0817347839,-0.0110200336,-0.7120335102,-0.0541849993,0.0637001842,0.2144561112,-0.278190881,-0.0741706416,-0.0047890185,-0.2388824522,-0.3776046038,-0.0251137242,-0.2068046629,-0.3644319177,0.2285363674,0.259321481,0.2990615964,0.5007333159,-0.0001083229,-0.1844812483,0.0241037235,-0.0698913485,-0.3096701801,-0.2250895202,-0.2824634612,-0.2135332972,0.0689910948,-0.3665153384,-0.0540316068,0.0273773614,0.1581238508,-0.3098548055,0.5706965923,0.1549828649,0.2261646688,0.135545373,0.0920443982,-0.0505273603,-0.0746740252,0.1321913302,0.0607593432,0.107791394,0.2328056991,0.1501888037,0.0214257874,0.1485862136,0.0968904793,-0.119904913,0.2620963752,-0.0519879013,-0.1156976596,-0.3440950513,-0.0209760815,0.1512996405,0.7606995106,-0.3858964145,-0.5661017895,-0.3009552956,-0.0400618352,-0.4118648171,-0.0546102487,0.1472489387,0.0820052549,0.1540421695,-0.2121725231,-0.1637293696,-0.2180507481,-0.138769567,-0.0414787196,-0.0969732553,-0.3794801235,-0.0211918242,0.0490676872,-0.2237613946,-0.1534072012,0.1326575428,0.1087393835,0.3603591025,-0.6488003731,-0.2567496002,0.0532763414,0.3321332037,0.5177853107,0.5682513118,0.0786654949,-0.0219275337,-0.0438606963,0.1380574256,0.36561656,0.1424448043,0.1061025709,-0.0619186237,-0.0528920069,0.0391137227,0.0796753243,-0.1260007024,-0.1406913102,-0.2216004282,-0.0771193728,-0.1565037966,0.0832067356,-0.1770687997,-0.102325581,-0.1652231067,-0.1618963778,-0.216027841,0.114815414,0.1496847272,-0.1081649587,-0.0370219424,-0.2981789112,0.0841588825,-0.0231003538,0.1277363598,-0.1204520166,0.3016714156,-0.2454103082,0.2856014073,-0.1388820857,-0.5510953665,0.1653596759,0.0034501124,-0.0500821993,-0.0972503349,-0.0386770107,-0.2035280168,0.133711949,0.0107535673,-0.0971942544,0.175756678,-0.1511567682,-0.0482361279,-0.4053089619,-0.3546987474,-0.2803290188,-0.1745544076,0.0428336449,0.1455173641,-0.2266248018,-0.0543788932,0.0425452627,0.343660742,0.0179279372,-0.1024764851,0.1268294603,0.0494637229,-0.3142506778,-0.0543956235,0.1424339116,0.5140600204,-0.447607398,-0.2488467246,0.193901211,-0.0189503431,0.0248288605,0.446115315,-0.1179488525,0.2248840034,-0.2795751691,0.1410651207,0.367336154,-0.1072311252,-0.2978147864,0.3310321867,-0.1225782707,0.1277452707,0.2151694298,-0.1188951135,0.1472864747,0.0209828187,-0.1125127152,0.0276411753,0.0095553789,-0.2397180051,-0.1244156882,-0.0534946397,0.0133727426,-0.0317953303,-0.0155017227,0.2256134152,0.1291157305,0.1652711183,0.4320666194,-0.1759159714,0.2170918435,0.2566938102,0.222090289,0.0498518646,0.0886928588,-0.2924880981,-0.26236099,0.1929330379,0.1236289218,-0.0143335843,-0.0734221488,-0.017405197,-0.2349372804,0.0079411762,-0.2194015086,-0.3230405152,0.1734453142,0.2221041918,0.3988912404,-0.053792119,-0.4686821401,0.5173083544,0.0457804874,0.1597618163,-0.8009337187,0.1625424325,0.2433849126,-0.0265652779,0.0728494972,0.1985827833,-0.0902727172,-0.3639824688,0.3573990166,0.2823091745,-0.1799612343,0.1475442648,0.1906578988,0.4468785524,0.2049324363,0.025963014,-0.0192388874,-0.3458394408,0.2969488502,-0.0436750725,0.0461218357,0.4764792621,-0.1672108173,0.3203335404,0.2574346662,-0.0481926575,0.0035503039,-0.0402696878,-0.5831554532,0.1241913959,0.1108983904,0.0170067437,0.4671846032,-0.1074706391,-0.2827883363,-0.0736338049,0.3593038917,-0.1044112071,-0.0366754904,0.1507020593,-0.1847528368,-0.1070031226,0.107776314,-0.0739282519,0.2477795929,0.0068557751,0.2435870022,0.0406483598,0.1717321277,0.021413913,0.0621021688,0.1553298533,0.1925956756,-0.0365203433,0.0750450119,0.1274712533,-0.155453071,-0.3518773317,-0.1105525941,0.057045754,-0.1778412014,0.3865460157,-0.2391075045,-0.2566678822,-0.3890227079,0.1047529653,-0.0565256588,0.0039648698,-0.2241475731,-0.074211061,0.3099526763,0.0557001233,0.114118889,0.4600075483,0.2347175181,0.0880562961,-0.1795245409,-0.0620844886,0.154369235,0.1218862236,0.1859986782,-0.1343820393,0.5079180002,-0.0053067748,-0.0006818027,-0.3153592646,-0.2890056968,0.3099226952,-0.1156775504,0.232262373,0.1817793101,-0.0603307225,-0.0614840649,-0.132613495,-0.0467311852,-0.1158417314,-0.0954230651,-0.040212974,-0.0446160547,0.2140877694,0.1073508188,-0.2305363864,-0.2717323005,-0.2117005438,0.7207010388,0.1371042579,0.2246389687,0.1908034682,-0.1179903522,0.3346859217,-0.1943351775,0.3252695501,-0.0625202358,-0.4787774384,0.3021920323,-0.3026356101,-0.1847880334,-0.1038351059,-0.1253803819,0.6247452497,0.0419879518,-0.3094746768,-0.2259999067,-0.05416568,0.0536925234,0.1094723418,0.2407761961,0.3135147989,-0.1626960039,-0.032929875,0.037129119,0.0101286797,0.0426246934,-0.04738383,0.0893518254,0.3445410132,0.5481717587,0.2190328091,0.5223556161,0.5356540084,0.0138916271,0.1400536597,-0.0983513892,-0.0980428457,-0.0277523268,-0.4531847239,0.1973807812,-0.3463818431,-0.2529503703,0.157041505,-0.0951182246,-0.0656500831,-0.2182156742,0.194153145,-0.2358365953,-0.1959584653,-0.1057221442,0.0122242663,0.2226586789,-0.2190306932,-0.1003992707,0.0521242768,0.1372745633,0.0047713104,0.2026717365,0.1884460747,0.0563495532,-0.5617639422,-0.0057134349,-0.4896618724,0.1085178852,-0.0363337062,0.0070988848,0.2339688689,-0.2241960615,0.1268604398,-0.2134832591,0.5830471516,0.3160966635,-0.4566282928,-0.0202447791,-0.5475603342,-0.0474631265,0.2805719972,-0.2953293025,0.2513093054,0.3468487859,-0.0295147672,-0.4579987526,-0.3302728832,0.1355598718,0.0209405497,-0.046124164,-0.1347164959,0.1114867628,-0.1327063441,-0.2398449183,0.0322486944,-0.0017581393,0.1818426251,-0.070636034,0.0369971246,-0.2385934889,-0.1215209439,0.2413257957,-0.0531454496,0.1067716181,-0.0217647552,0.2209533155,0.3449824452,0.1826763749,0.1949136555,0.5341263413,-0.0401168801,-0.0637314841,0.2419627607,0.1099615991,0.3433791399,0.2787808478,-0.1187918931,0.1542900503,-0.1969991326,0.1371665895,-0.4949762821,-0.0650090203,0.3207852542,0.0214663129,-0.3879270256,-0.5668643117,0.3979650736,0.0945351794,-0.0297925733,-0.111482136,0.1785356253,-0.1891283244,-0.1138752252,-0.0083277402,0.8772637248,0.0319275334,0.1329398453,0.2381796539,0.2841379941,0.6234858632,-0.0280070528,0.0178284552,-0.3746920824,-0.2490864545,-0.1163340881,-0.1235605702,0.2895082831,0.2170466632,-0.3202975094,0.5073506236,0.2475793362,-0.1227073446,-0.1606146842,0.213112697,-0.2810669243,-0.2357784957,-0.5983777642,0.1458412111,0.1600373834,0.2836305499,-0.1796882153,-0.1390869766,0.4076881111,-0.0017842427,-0.1486614645,-0.2349316329,-0.2477785647,-0.1041986346,0.2241984755,-0.4196967781,0.2544489503,0.2264540195,0.2770259976,0.0333958529,-0.2222714573,-0.1569137424,-0.1830388308,0.0029032656,-0.0053308699,0.0556920208,0.6071686149,0.0503938198,0.1515834033,0.2582307458,0.0890820548,0.0571655929,-0.0774367377,-0.0198148713,-0.1026986316,-0.3565374911,0.0067695347,0.3971951604,0.0090592364,-0.0701593459,0.1332820356,0.3546117842,-0.1463988423,0.236544013,0.0434919931,-0.1403742582,-0.1912482232,0.4071800411,-0.1181037351,-0.1983419359,0.4554192722,0.1661762595,-0.2287354022,-0.2077703625,0.1140217409,0.322791338,-0.4184871614,-0.0639821067,-0.0201242529,0.2953800559,0.0926590189,0.2044045925,0.1547838897,-0.1619636267,-0.0641450137,-0.2246106118,-0.1052446887,0.315466404,0.2270198464,0.2550222874,-0.0240134113,-0.152289167,0.0329786986,0.2615962923,-0.3584605455,-0.0105181467,-0.1733997315,0.0665032119,0.3252342045,-0.0954422355,0.1232602522,0.2710687816,0.1061671823,0.2823154628,-0.3206465244,-0.2339215279,-0.1526694894,0.0869125128,0.0169801302,-0.0633454025,0.1054919511,0.0557909012,-0.1318688095,-0.210532099,0.3006531298,0.0091102617,0.2976890802,0.2314928919,0.3964725435,-0.1185308546,-0.1576303393,0.0600554831,0.1846354306,-0.0466666073,0.1568503231,0.2884025276,-0.3015138507,-0.2241450697,-0.0560048446,0.392347157,0.4073381126,-0.233911708,0.3088851273,-0.2933986783,-0.0103805363,0.1303736418,0.5188375711,0.5250545144,-0.1241682991,0.0555375814,0.2879222333,0.2092597187,-0.0727561265,0.1112759858,-0.2497937828,-0.0984958112,0.1296977699,0.383281976,-0.0850634202,0.1669869125,0.2132769525,-0.1150757745,0.1987613589,-0.2142683864,0.0188857466,0.0797779486,0.0300601758,-0.0799141973,0.3738003969,0.4228509963,0.2139608264,0.2633865476,-0.2759991884,0.397324115,-0.2264548838,0.2418713123,0.1715239733,-0.1857383251,-0.1389621496,0.2813614309,0.1532009244,-0.0042248927,-0.2404959351,0.5388868451,0.1678975821,-0.0543511659,0.0097788014,0.4228827953,0.0503141545,-0.3581236899,-0.071141161,-0.243021667,-0.2250370383,-0.1238202527,-0.2629260421,-0.2353131473,0.2613812089,-0.0566111282,0.3644625247,-0.5372337103,0.0038962755,-0.0219980367,0.1132595912,-0.3350177109,0.1712703705,-0.0511972308,0.0864297673,-0.0958902463,0.3389082253,0.2542077005,0.1117471531,0.0073345648,0.0347144864,-0.2352551818,-0.1012960672,-0.0062682019,0.1274973899,-0.254877001,-0.1391127408,0.5069288015,0.0564196296,-0.0876500458,0.2528473735,-0.0787241608,0.0096901236,-0.9196891785,0.2116187066,-0.0865175053,0.0649556443,-0.1183385104,0.0958193988,-0.505110383,0.0817726478,0.1601700783,-0.4940124154,0.0981379524,-0.0870595276,0.1044276357,-0.1923288405,0.4898715317,0.3311973214,0.3033388853,-0.0329315364,-0.1535871327,-0.3243837655,-0.1424014568,0.1704669744,0.0346859582,-0.0449092425,0.3771384358,0.0564841926,0.1602362543,-0.2984938323,-0.5570547581,-0.3035137653,0.24983868,-0.2531627417,-0.1206758767,-0.3716829419,0.2428581715,0.1048691422,0.0782123208,0.1468791515,-0.3805567026,-0.028953325,-0.0192323085,0.2883636653,-0.0651243851,-0.320183903,0.1772939265,0.1361568868,0.005562257,-0.1212323681,0.0008988651,-0.3256331384,-0.3225668073,-0.1099896803,0.3711661696,-0.0431344509,0.2471926808,-0.1225269139,-0.2282951176,-0.0202582963,0.1438214332,0.1343943477,-0.1148288995,-0.059149988,-0.0152185764,0.0624645688,-0.0228195805,-0.0204009581,0.525351584,-0.0225597229,0.2398401946,-0.1348939091,-0.2718342841,0.3910017908,-0.4392785728,0.0391687639,-0.1211602837,0.1023644656,0.254471153,-0.4025706649,-0.40008986,0.1132172421,0.4312736988,0.0453593507,-0.1451373249,0.2790007293,0.284240514,-0.0681636781,-0.2841490507,0.3456366956,0.104129985,-0.0624755621,0.187496379,-0.2553837299]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/815","title":"Is dataset iterative or not?","comments":"`datasets.Dataset` objects implement indeed `__getitem__`. It returns a dictionary with one field per column.\r\n\r\nWe've not added the integration of the datasets library for the seq2seq utilities yet. The current seq2seq utilities are based on text files.\r\n\r\nHowever as soon as you have a `datasets.Dataset` with columns \"tgt_texts\" (str), \"src_texts\" (str), and \"id\" (int) you should be able to implement your own Seq2SeqDataset class that wraps your dataset object. Does that make sense to you ?","body":"Hi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks","comment_length":76,"text":"Is dataset iterative or not?\nHi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks\n`datasets.Dataset` objects implement indeed `__getitem__`. It returns a dictionary with one field per column.\r\n\r\nWe've not added the integration of the datasets library for the seq2seq utilities yet. The current seq2seq utilities are based on text files.\r\n\r\nHowever as soon as you have a `datasets.Dataset` with columns \"tgt_texts\" (str), \"src_texts\" (str), and \"id\" (int) you should be able to implement your own Seq2SeqDataset class that wraps your dataset object. Does that make sense to you ?","embeddings":[-0.1060156971,-0.0653585941,-0.0744617656,0.2728957236,0.0212031826,-0.0095434887,0.2350491583,0.2607481778,0.0178500302,-0.1811223626,0.2007772624,0.2117641866,-0.3875654638,0.19634749,0.0909050927,0.0099149477,-0.1221981049,0.0177189317,-0.6584701538,-0.1331171393,-0.0125520015,0.1090283915,-0.3733929098,-0.0125812609,-0.0573374778,-0.2997436523,-0.3547765613,-0.0941223279,-0.3227387667,-0.3790437877,0.1680089086,0.3244619071,0.4530280828,0.3990264237,-0.000104547,-0.3836460412,0.0271534603,-0.0992269292,-0.2510547936,-0.1770192385,-0.2945179641,-0.2617836297,0.1364851445,-0.4748810828,-0.0804626793,-0.0584186241,0.0909327939,-0.3703354597,0.2763306499,0.0961010903,0.2760343254,-0.424844265,-0.0344233364,-0.0648907349,0.0225603599,0.049664747,-0.0703690499,-0.0418418087,0.3111680746,0.3059273064,-0.0155926682,-0.0311851595,0.1218771935,-0.1981253922,0.3405269384,-0.0772939995,-0.0879147723,-0.3361750245,0.0023595735,0.0840367898,1.1303713322,-0.2262818068,-0.306194514,0.0234375037,-0.015231166,-0.4028499424,-0.1640854478,0.29465729,0.1456294358,0.1291965842,-0.2035063654,-0.3969239295,-0.2363087386,-0.1007134318,-0.126223281,-0.0489157997,-0.1681756973,0.0840758383,-0.0203479566,-0.2884940207,0.1532578617,0.1695320457,0.0152814379,0.2250156701,-0.6152793169,-0.3189679086,0.0359667391,0.3883178532,0.4129840136,0.3702382445,0.0624873154,0.0317518041,0.002943628,0.2047530264,0.4332893193,0.0117595373,0.0783103853,0.0479110144,0.0488203801,-0.1845354289,-0.205026418,-0.0807033256,-0.2166119516,-0.1597901881,0.1061509401,-0.164549768,-0.1567519009,-0.1790119857,-0.0017529721,0.0000938252,-0.2421870977,-0.1376674771,0.1298843771,0.1093622819,-0.0902501866,-0.0274752136,-0.1273261607,0.1498544961,0.1142807603,-0.0207859315,-0.0622128136,0.474968195,-0.2980031371,0.0847253576,-0.0686666444,-0.3925188482,0.0734553486,0.0530897193,-0.1771769673,-0.0755128488,0.0849153474,-0.2322214246,0.1744033247,-0.1072451621,-0.168242231,0.1184785441,-0.1914855093,-0.1652150154,-0.4168971181,-0.1530209631,-0.3378536105,-0.1186842546,-0.1220381483,0.1512054056,-0.1720935106,-0.190055728,0.1995152831,0.4862645566,-0.0187182296,-0.1701807976,0.04553831,-0.1141480654,-0.2942237258,-0.1201193109,-0.0273045525,0.4135944843,-0.7152242064,0.112702921,0.0177283678,0.1990044415,0.0440053158,0.3204875886,-0.1433266252,0.2875398695,-0.1735821366,0.0832265913,0.4689073563,0.0261819661,-0.2216172963,0.1995010227,-0.1015404016,-0.1679414064,0.1796113551,0.0820492953,0.2748194933,-0.0586654022,-0.113525264,0.0100040371,0.0261893589,-0.2152169049,-0.1232686713,-0.0396792367,0.1559778601,0.0144439712,0.0755374804,0.2684825361,0.1918976307,0.2799932361,0.2079637647,-0.1532561034,0.2448382527,0.0787206218,0.3107308149,-0.0092805242,0.1097956449,-0.182276234,-0.0991682038,0.1989352703,0.3425956368,-0.0336224139,0.0973454714,0.0093338443,-0.1203001216,-0.0338788666,-0.1323507577,-0.1853948683,0.193635568,-0.0194624607,0.3257060051,-0.1215667352,-0.361246258,0.2648516297,-0.0278109107,-0.071654588,-0.635613203,0.1184780896,0.2372146845,-0.0631385371,0.1188148335,0.1731708795,-0.1699392051,-0.2727327943,0.3001478314,0.3359160125,-0.0359949581,-0.0734905005,0.2832366228,0.5574188828,0.1177820191,0.0026029074,0.2196706831,-0.4144702256,0.2916940749,0.0419196561,-0.0218471196,0.4437411129,-0.1784731895,0.1090888977,0.1369903237,-0.0261689257,0.074157767,0.0160719939,-0.5682403445,0.0849422514,-0.2583654225,0.1411114633,0.253107518,-0.0921327323,-0.3893495798,-0.1187008172,0.4135907888,0.024784619,0.0798389092,0.1774080992,-0.111607179,-0.1065041423,-0.0296097398,-0.0075327279,0.3253003657,-0.0390895978,0.1885775924,-0.0195131265,0.0985074043,0.0713533461,0.1269074082,0.073755078,0.1753561497,-0.1058323383,0.0945662931,0.0835992545,-0.1805022657,-0.2936784327,0.0505964272,0.2180814445,-0.0557901785,0.3049521744,-0.4185894132,-0.1184837818,-0.3756096363,0.2154709846,0.2625839412,0.087883763,-0.0428275131,-0.1942445636,0.40513134,0.0120761963,-0.0770671666,0.5587694645,0.2348256409,0.0640678704,-0.0663646385,-0.1118966937,0.0384163037,0.1578057408,0.1735128313,-0.0893606544,0.5077780485,0.0487739891,0.2402329594,-0.4060040414,-0.3096460104,0.3122244179,-0.0110638198,0.2463253886,0.2539362311,-0.1717471629,-0.0561871156,-0.1773165464,-0.033527866,-0.2286292464,0.0861981958,0.0113951238,-0.0456722192,0.0788047239,0.0978551954,-0.3836063445,-0.1660971045,-0.1279617548,0.4235076606,0.2191938311,0.3020019829,-0.0476010256,-0.0579888113,0.2834451795,-0.0312266331,0.1959004402,-0.1284116507,-0.4388583899,0.3763958514,-0.2919609845,-0.1978156716,0.0604937226,-0.10550908,0.6279435158,0.0679761171,-0.2470659614,0.1047998741,0.1767358631,0.1371100247,0.252369225,0.1300071329,0.266561836,-0.1598167568,-0.0430794172,0.0375887752,-0.1077144817,0.2246945947,0.2038492709,0.2265102714,0.3720066249,0.6611022949,0.036840599,0.4294275641,0.3031008244,-0.2027745992,0.229659453,-0.0535757914,-0.2370571792,-0.0487279333,-0.2013127208,0.2020185143,-0.3556555808,-0.4386449456,0.2825660408,-0.1113413796,-0.3914494216,-0.2349856943,0.4713566005,-0.3060874641,-0.0637127236,-0.0891982466,-0.2912811935,0.1065605134,-0.1391033083,-0.1137614027,0.090913333,0.0534704328,-0.0448542908,0.0933407024,0.1892964393,0.0122022079,-0.5341045856,-0.1590789855,-0.2637723088,0.0319079496,-0.3149481118,0.1258900315,0.3008846343,-0.2130642682,0.2733726203,-0.2154854387,0.3534606993,0.0385159366,-0.5208890438,0.022639215,-0.3992438316,0.0477931052,0.0145367999,-0.2505373359,0.2993895411,0.3013772964,-0.1759857237,-0.4163496792,-0.2557186484,0.1383341402,-0.0987203568,-0.1172995865,-0.2326556593,0.1805446446,-0.0896567106,-0.0827388912,-0.0147186518,-0.1037463322,0.0314406268,-0.1423358321,-0.1396666616,-0.2243328542,-0.0871972069,0.1617321521,-0.0342006981,0.0619011745,0.0406257994,0.2935373485,0.2934241891,0.2970147729,0.1757422686,0.3449501693,0.0146712177,0.3328717649,0.5411565304,-0.0869915932,0.4087324739,0.0927115679,-0.1379282624,0.1165258288,-0.4787021279,-0.052269876,-0.4371348917,-0.2746163309,0.4635876417,-0.1199735701,-0.3364788294,-0.6614847779,0.4097927213,0.0774947032,-0.1066132113,-0.130663991,-0.0899975076,-0.1915473193,-0.0262279008,0.3491604924,0.7590566278,-0.1307473034,-0.0448020585,0.0450963974,0.2970717847,0.6364308596,-0.2961922586,0.0231918059,-0.6081212163,-0.1643237323,-0.1351261586,-0.0963244513,0.2675540745,0.3350670934,-0.4736582935,0.4287880063,0.127219528,-0.152836889,-0.109224245,0.3182109892,-0.1011966094,-0.1116529927,-0.2465977222,0.1743552089,0.0811897814,0.0533640869,-0.2265841067,-0.0237983689,0.3006130755,0.1477136761,-0.1413565874,-0.2302380502,0.1132761613,-0.013439809,0.1637122184,-0.316362083,0.458355546,0.2964681685,0.1053092107,0.1151626483,-0.2490949035,-0.0775178298,0.0167330503,0.0995309055,0.1977814883,0.0973373726,0.5548225641,0.0285493881,0.1249640658,0.3597328067,0.0011494143,0.0049217865,-0.1626740098,-0.0178962275,0.1485939324,-0.3025823236,-0.1348324865,0.3531946242,0.0456089973,-0.0448981263,0.1735304892,0.3285720348,-0.0790313408,0.4723214805,0.0573697537,-0.1232923344,-0.1456577927,0.2969469428,0.0554350801,-0.0328127332,0.3297345936,0.0898153409,-0.1387702376,-0.2093993425,0.2170038074,0.2161769122,-0.2025470585,-0.1739452481,-0.0904433802,0.2153176814,0.371304512,0.1287776083,0.0677009076,-0.1450610906,-0.2804511786,-0.1708846092,-0.1562746018,0.1291252375,0.1656689644,0.1216776669,-0.1157604083,-0.3182606697,0.0446493402,0.2772842944,-0.3766229749,0.0246583391,-0.1277281195,0.2063361704,0.0793087557,0.0711804107,-0.2723812759,0.3149991632,0.0793944597,0.1331261694,-0.2244430184,-0.1894849688,-0.0707060397,0.0689680725,0.0498472154,0.036801029,0.1204812005,-0.174567014,-0.3809264898,-0.0627046376,0.1292874068,-0.0888311192,0.1195102483,0.274607569,0.4228079021,-0.3086164892,-0.1767037958,-0.0247148462,0.1111768708,-0.0180243384,0.1713849902,0.3880265355,-0.152714774,-0.3117508888,-0.1822988689,0.4856969118,0.1304291487,-0.1920969784,0.5589895248,-0.2203160822,0.2540809214,0.2254094929,0.3742898405,0.7063232064,-0.105304487,-0.1500674337,0.2866018116,0.2389710397,-0.0559889786,0.0108430507,-0.3334939182,0.0441766642,-0.0118429093,0.3476697803,0.0535372384,0.0318178758,-0.0273017455,-0.2289721817,0.1374920458,-0.2728151083,0.0267365407,0.316413641,-0.0272724312,-0.0846009403,0.4505916834,0.2930499613,0.1773281991,0.4294733405,0.0442479588,0.4079272151,0.0812658444,0.2050414085,0.0839120746,-0.1520217061,-0.0152878445,0.2997791171,0.2001311034,-0.0682498664,-0.1755220294,0.408197552,0.3399285674,-0.0029968859,-0.2276376933,0.2862155735,-0.0447308496,-0.3045302033,-0.282445401,-0.4127629101,-0.2404183745,0.0070386976,-0.4302239716,-0.0106578181,0.1558199227,-0.0350746997,0.3782491982,-0.4300780892,0.2995273769,-0.1636081338,0.0305792261,-0.351619035,0.2129738033,-0.1583889276,0.0427319333,0.1224057376,0.1994773597,0.0034866808,0.4258717597,-0.1050929725,-0.1495064944,-0.2400678992,-0.1211316213,0.0517933927,0.2002233267,-0.0645294562,-0.2436395586,0.5186992884,0.0383190364,-0.0564184301,0.3666834235,0.0617587864,0.0538375378,-0.8294041753,0.0074729207,0.2230871469,-0.030852994,0.0967528224,0.0852524713,-0.2967817783,-0.0864661634,0.1451617628,-0.4347529709,0.0169294886,-0.1616193503,0.118241705,-0.2994230688,0.5220739245,0.1555450261,0.1369382143,-0.0087492839,-0.2753685117,-0.2950528264,-0.0729633272,0.138455227,0.0468282215,0.0191169139,0.4880802333,0.1408645958,0.1854006648,-0.4162905216,-0.4003131092,-0.4876222908,0.4227484167,-0.2096909732,-0.0771628469,-0.447594136,0.2959057391,-0.0930420235,0.112606518,0.0568128414,-0.381595999,0.0128305824,-0.0074272319,0.123792246,0.0027805679,-0.3025190234,0.3909682333,0.0033504725,-0.0014867488,-0.0579905845,0.1699957848,-0.2926506996,-0.2134749889,-0.0951056629,0.3746216595,0.0295181032,0.1459097564,0.0756729245,0.1762552708,0.0559029281,0.0100220516,0.0112853413,-0.0517374724,-0.210752055,-0.1740847975,-0.0035121792,0.0619442314,-0.0053608529,0.5795051455,0.0160581898,0.1983938962,-0.1722986549,-0.062805362,0.2013785541,-0.172105968,-0.0550857373,-0.2402300239,0.1379254013,0.2016530633,-0.2732510269,-0.1337060779,-0.0182908606,0.411808461,0.0264307745,-0.0824638233,0.144028753,0.361060977,0.059081208,-0.2883608341,0.2852184772,-0.025662953,0.0656940192,-0.084488824,-0.2586916685]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/815","title":"Is dataset iterative or not?","comments":"Hi\nI am sorry for asking it multiple times but I am not getting the dataloader\ntype, could you confirm if the dataset library returns back an iterable\ntype dataloader or a mapping type one where one has access to __getitem__,\nin the former case, one can iterate with __iter__, and how I can configure\nit to return the data back as the iterative type? I am dealing with\nlarge-scale datasets and I do not want to bring all in memory\nthanks for your help\nBest regards\nRabeeh\n\nOn Mon, Nov 9, 2020 at 11:17 AM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> datasets.Dataset objects implement indeed __getitem__. It returns a\n> dictionary with one field per column.\n>\n> We've not added the integration of the datasets library for the seq2seq\n> utilities yet. The current seq2seq utilities are based on text files.\n>\n> However as soon as you have a datasets.Dataset with columns \"tgt_texts\"\n> (str), \"src_texts\" (str), and \"id\" (int) you should be able to implement\n> your own Seq2SeqDataset class that wraps your dataset object. Does that\n> make sense ?\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/815#issuecomment-723915556>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ARPXHHYOC22EM7F666BZSOTSO66R3ANCNFSM4TPB7OWA>\n> .\n>\n","body":"Hi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks","comment_length":217,"text":"Is dataset iterative or not?\nHi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks\nHi\nI am sorry for asking it multiple times but I am not getting the dataloader\ntype, could you confirm if the dataset library returns back an iterable\ntype dataloader or a mapping type one where one has access to __getitem__,\nin the former case, one can iterate with __iter__, and how I can configure\nit to return the data back as the iterative type? I am dealing with\nlarge-scale datasets and I do not want to bring all in memory\nthanks for your help\nBest regards\nRabeeh\n\nOn Mon, Nov 9, 2020 at 11:17 AM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> datasets.Dataset objects implement indeed __getitem__. It returns a\n> dictionary with one field per column.\n>\n> We've not added the integration of the datasets library for the seq2seq\n> utilities yet. The current seq2seq utilities are based on text files.\n>\n> However as soon as you have a datasets.Dataset with columns \"tgt_texts\"\n> (str), \"src_texts\" (str), and \"id\" (int) you should be able to implement\n> your own Seq2SeqDataset class that wraps your dataset object. Does that\n> make sense ?\n>\n> \u2014\n> You are receiving this because you authored the thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/815#issuecomment-723915556>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ARPXHHYOC22EM7F666BZSOTSO66R3ANCNFSM4TPB7OWA>\n> .\n>\n","embeddings":[-0.1699585021,-0.1714058369,-0.0689312667,0.29238832,0.1009495035,-0.0845248029,0.3338168263,0.2466901839,0.179701373,-0.1265341341,0.068009831,0.1442486793,-0.3788862526,0.412889421,0.1700596809,-0.0318089277,-0.0919883251,0.0115778297,-0.6868837476,-0.1174983755,0.0290254131,0.1487227678,-0.2845766842,-0.039639689,-0.0925771147,-0.2904197276,-0.2832887471,0.0088117104,-0.3676557541,-0.3626948595,0.3347042799,0.2895670235,0.494183898,0.5043627024,-0.000105784,-0.2772066593,0.0528096072,-0.1300345659,-0.2246703207,-0.2106571794,-0.2412293106,-0.2051681876,0.1279392838,-0.3685281277,-0.044530645,-0.085979417,0.0379860997,-0.3243009448,0.3284459412,0.07557372,0.2515381873,-0.2350952774,0.0400567502,-0.0884437189,-0.0125634493,0.1745035052,-0.0649378449,0.0631418675,0.2634143531,0.1847220361,-0.1005844921,-0.0201257877,0.1368308514,-0.120177418,0.3518400192,-0.1117137149,-0.1345442384,-0.3502725661,-0.0687426254,0.1795235276,1.1247750521,-0.2266197652,-0.3315274715,-0.1051943004,-0.1077040732,-0.384303391,0.0519696921,0.1613957435,0.1376384348,0.0931484699,-0.2176344097,-0.2320619076,-0.2463390976,-0.0989342481,-0.1977272481,-0.218505457,-0.3083755374,0.0284801517,0.1063963398,-0.3117009699,-0.0058239074,0.1124444008,0.1506894529,0.284599334,-0.6823297739,-0.3560846746,-0.0164678078,0.2891712487,0.4668389559,0.3724458814,0.0903385282,-0.0460713729,-0.1429632455,0.1361737102,0.4857326448,0.1536505073,0.091422677,-0.0331862569,0.0548191965,-0.0040666908,0.0111732017,-0.0994318202,-0.2033264786,-0.1721980423,0.0453471877,-0.1871928871,-0.0844531655,-0.2436035872,0.0123192566,-0.0572085008,-0.0694758371,-0.1390123516,0.1348976046,0.1547203958,-0.0590074472,0.1231463104,-0.112008132,0.0874819085,0.0184106734,0.0013463529,-0.0628624037,0.3680286705,-0.3997497857,0.1252013147,-0.0149138141,-0.4068548381,0.0481101125,0.1276565492,-0.194136247,0.0171589013,-0.0430658795,-0.2238801867,0.1048341841,-0.0725570247,-0.2010730803,0.2333637476,-0.2167590559,0.0247235019,-0.3907789588,-0.1400688738,-0.3842135966,-0.0907068476,-0.1005614698,0.1538315862,-0.2043938637,-0.230973646,-0.1431592256,0.3670678735,-0.0654408634,-0.1174171567,0.0704515204,-0.1015896052,-0.2147193104,-0.1085753292,0.0136708729,0.4157608151,-0.7381816506,-0.0744674206,0.0410949215,-0.0164117254,0.1620524973,0.4142245948,-0.147715345,0.2711117566,-0.2453812361,0.1480748802,0.3742136061,0.0069465423,-0.2698088884,0.2698873281,-0.0603346266,0.0054400959,0.1192872524,0.0076572672,0.2325147092,-0.0825404748,-0.1676668227,0.1485361308,0.080385901,-0.1396999657,-0.1522193402,-0.1051234007,0.1615737826,0.0591262095,0.0720561668,0.2700530887,0.1509655267,0.2775504589,0.2225801945,-0.1702853441,0.2985440195,0.0976856202,0.294963181,0.0548568815,0.0704542994,-0.2696400583,-0.2262954265,0.2209364027,0.1445796192,-0.0016945195,0.1903557777,0.0931611657,-0.1611178815,-0.0879651383,-0.2035260648,-0.1719710231,0.176610738,0.0991632566,0.284912318,-0.0842207223,-0.4242133498,0.3340341747,-0.1232871562,-0.0024321913,-0.7031782269,0.1637267619,0.3130322397,-0.0872310475,0.0873475075,0.1098693162,-0.1471638083,-0.2829943597,0.2350479662,0.3226821125,-0.0962361544,0.1296237409,0.2510299683,0.4881451428,0.1757104993,0.0233464204,0.2103522718,-0.4388235807,0.2086124569,-0.0578279197,0.0817303732,0.4162714481,-0.2419126928,0.1187950522,0.1114322618,-0.0289150774,0.0098034479,-0.0222697835,-0.4933255613,0.2212381065,-0.0613504834,0.0618069582,0.2879597247,-0.0901600569,-0.2579862773,-0.1866322458,0.4187476039,-0.0632652566,0.0046470636,0.2310644686,-0.2082143426,-0.0975415632,0.0669202954,-0.033318989,0.3083370328,-0.0466552228,0.2622880638,-0.0572665557,0.1278201789,0.0857335702,0.1289166063,0.0169786401,0.2707156241,0.0105032828,0.0580160804,0.0779035836,-0.2738215029,-0.219468981,-0.0109609812,0.1313216537,-0.0874417871,0.1833934635,-0.4029295146,-0.1049549431,-0.3309220076,0.1238508001,0.0387210064,0.0642085597,-0.118480444,-0.0965833813,0.4058251679,0.0656917691,-0.0229723603,0.5276075006,0.2944499552,0.0137874298,-0.1694500744,0.0357918106,0.0432924218,0.1796344221,0.2374723703,-0.2035424262,0.4684326351,0.0262028147,0.0783145949,-0.4120460153,-0.2755690515,0.3278045058,-0.1076430827,0.2515527606,0.3064386547,0.0044726939,-0.0206184741,-0.0463249534,0.0341679901,-0.3438085616,-0.0050206608,-0.0264046974,-0.0640740618,0.2453982681,0.0882140025,-0.4056174755,-0.2220086604,-0.2096966058,0.5581911802,0.2181431651,0.2245459855,0.1635544896,-0.0632682517,0.3151679635,-0.0617828034,0.2107881606,-0.2140313387,-0.5315765142,0.387829423,-0.2561224997,-0.2005531937,0.0406557843,-0.0047406703,0.5716431737,0.0240638386,-0.3571440279,0.0166935623,0.0419940725,0.1119483784,0.1873508692,0.2401827276,0.2688873708,-0.2422915995,-0.0287387297,0.0596929789,-0.0924317911,0.158922255,0.1621715128,0.3196926713,0.4330098927,0.5967088938,0.0411427133,0.5284713507,0.3443607986,-0.1735187322,0.2482471615,-0.0325032473,-0.2144196332,-0.1131134257,-0.2672350407,0.2950757444,-0.3456632197,-0.2737661302,0.1929452866,-0.1302106977,-0.307544291,-0.1988605559,0.3118447661,-0.1982480735,-0.0239245892,0.0483292267,-0.108372435,0.1205379292,-0.24764359,-0.1920785308,0.0079594748,0.0128113711,0.060469076,0.1921095699,0.2455522716,0.0608974025,-0.6099678278,0.0582396276,-0.3910817206,0.0739468113,-0.280895561,0.0219321437,0.1650058329,-0.241587013,0.2220959961,-0.1012185663,0.5630609989,0.1098441109,-0.5177435875,-0.0360055901,-0.4384682178,0.0871369615,0.0322222225,-0.2800098062,0.2983269691,0.2503449321,-0.1287993342,-0.4448174238,-0.242533654,0.133397609,-0.0708023757,-0.0454372466,-0.3161968291,0.1628025621,-0.0371813476,-0.104187943,0.0087770112,-0.0863753632,0.0807396099,-0.1742275506,-0.1564788967,-0.1778128147,-0.0205598753,0.2238779515,-0.0342711248,0.1213206351,0.0417350791,0.2217037976,0.40376091,0.3132401109,0.2758372128,0.4043897092,0.010641057,0.2390816063,0.5203079581,0.0507808365,0.3208163977,0.1847978383,-0.1532656401,0.045321513,-0.2774815559,0.0296793804,-0.6391438842,-0.2000180781,0.4938615263,-0.0547903553,-0.379122436,-0.6322649717,0.5206336379,0.1117025167,-0.0500369892,-0.0322705396,0.0409660973,-0.209024623,-0.1775488853,0.2071647644,0.8239881396,-0.2135829031,-0.0438237078,0.1120209694,0.4150190949,0.6200096607,-0.2753099501,0.0017928216,-0.4699938297,-0.185735181,-0.1635929942,-0.1298053861,0.2541496754,0.2704184055,-0.3658440709,0.4825980663,0.19027403,-0.0750206634,-0.1667434275,0.3217437565,-0.1960278153,-0.2429683357,-0.4122357368,0.1731429547,0.0524774343,0.1378443539,-0.1709503531,-0.0485196933,0.3666971624,0.0627637506,-0.1055116728,-0.2299222052,-0.1149456725,-0.1376873106,0.2727746964,-0.2581344843,0.4700640142,0.2415486872,0.0990084112,0.0317004398,-0.2596076727,-0.1158435121,-0.0806154013,0.1089257672,0.1270158142,0.0883720368,0.6001086235,0.0133093158,0.1390313059,0.224947229,0.0109505691,-0.0401244015,-0.1303885579,-0.057948187,0.1187091172,-0.4455755353,-0.067079477,0.280908972,0.076958634,0.0015685993,0.1574689895,0.3391580582,-0.0816180557,0.3666170835,0.104087621,-0.0705436096,-0.1703120619,0.3324107826,0.069265455,-0.140418604,0.421854645,0.15576455,-0.2226468772,-0.151833564,0.1414209157,0.3399426937,-0.2813914418,-0.0539524704,-0.0102005098,0.1473529637,0.2052129507,0.2366079688,0.1382711679,-0.1978302002,-0.2085977793,-0.1866783947,-0.1838696897,0.1925513297,0.0223175939,0.2390881777,-0.017047558,-0.29211694,0.1026852876,0.1451331079,-0.3657821417,0.0849661827,-0.1126861796,0.1498846412,0.1255359054,-0.0142158819,-0.0405603983,0.2896203697,0.0711742118,0.1406238079,-0.2981957793,-0.2043917626,-0.0519887023,0.0834468529,0.0235458594,-0.052292604,0.1546593159,-0.1482755393,-0.2841723859,-0.100371778,0.2105144858,-0.153662771,0.0982102081,0.2074412107,0.3586407602,-0.3746783137,-0.1585952193,0.0784563571,0.1911972463,-0.081700094,0.1215364933,0.3406088054,-0.1940483302,-0.3026096821,-0.2282734811,0.5108462572,0.2922608852,-0.1336012483,0.4713009,-0.3304161429,0.0564011,0.2789524198,0.5351672769,0.6012006998,-0.1514584869,-0.0880291611,0.1812178791,0.2208258063,-0.0255824011,0.1370265931,-0.1873174459,0.033784125,0.0326965675,0.3433177173,0.1192837507,0.0253992267,0.1265500188,-0.1984067857,0.1594917327,-0.2759272158,0.0862493664,0.2314596027,-0.130520463,-0.0285996627,0.5951685309,0.3472752571,0.0456608571,0.3939045668,-0.025083147,0.2953464985,-0.0612823255,0.2569728792,0.0952456966,-0.3001562655,-0.0845786184,0.3915069699,0.0978897735,0.0029346554,-0.2082814276,0.5517612696,0.1346995234,-0.1131887659,-0.2175141722,0.2635158598,-0.135924086,-0.2410647124,-0.283913821,-0.3782042265,-0.2725360692,-0.019667577,-0.4187793136,-0.0917318761,0.284294337,-0.0472545959,0.2933647335,-0.520666182,0.1271156967,-0.1305072159,0.0793392733,-0.3140775561,0.3320922256,-0.196868822,0.0830753222,0.0245195329,0.3627671301,0.0794949085,0.3440088332,0.0241480172,-0.0531602725,-0.126097858,-0.0811889246,0.1318767518,0.2378996909,-0.1564376652,-0.2195758671,0.4602530599,0.0706408024,-0.0613158867,0.2296070457,-0.110305883,0.1432420462,-0.806281805,-0.0414233133,0.1084203422,0.1404685229,-0.0649568737,0.0720158294,-0.4094016254,-0.0231505744,0.1770357937,-0.4590133429,0.0971073508,-0.1491707414,0.1216287464,-0.2849751115,0.440184176,0.2593486905,0.0721053481,-0.0150859933,-0.160132736,-0.2753868103,-0.0681062043,0.1776989549,0.0814498961,-0.0326995626,0.4383941889,0.1972638816,0.1426799744,-0.2345995307,-0.5380961895,-0.3567995429,0.3172581494,-0.2102690786,-0.0598603971,-0.3488931656,0.3244600594,-0.0278825779,0.0478275903,0.1658544987,-0.2729411423,-0.0488028899,0.0191644002,0.2362481654,-0.0749064088,-0.3555975258,0.3192619681,0.1284681708,0.0383892432,-0.0466247983,0.0327393413,-0.3662337959,-0.3125418425,-0.1299943179,0.2862896025,-0.0575804673,0.2802180052,-0.0103685763,0.0820516646,-0.0783562809,0.0465470888,-0.0317770168,-0.0440901257,-0.2918261886,-0.1778579354,-0.0064857951,-0.1045860052,0.0302177109,0.5602522492,0.022652626,0.1567508429,-0.1889245361,-0.18420057,0.2402456701,-0.2860487998,-0.0586623773,-0.0946224406,0.0349200219,0.2588537633,-0.1969830841,-0.3583437204,0.1509940475,0.3538052738,0.0106140273,-0.038441997,0.190885365,0.2333613038,0.0560336187,-0.2807218134,0.2072654217,0.0476629771,0.0948215798,0.0134481909,-0.2773856521]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/815","title":"Is dataset iterative or not?","comments":"`datasets.Dataset` objects are both iterative and mapping types: it has both `__iter__` and `__getitem__`\r\nFor example you can do\r\n```python\r\nfor example in dataset:\r\n # do something\r\n```\r\nor\r\n```python\r\nfor i in range(len(dataset)):\r\n example = dataset[i]\r\n # do something\r\n```\r\nWhen you do that, one and only one example is loaded into memory at a time.","body":"Hi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks","comment_length":57,"text":"Is dataset iterative or not?\nHi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks\n`datasets.Dataset` objects are both iterative and mapping types: it has both `__iter__` and `__getitem__`\r\nFor example you can do\r\n```python\r\nfor example in dataset:\r\n # do something\r\n```\r\nor\r\n```python\r\nfor i in range(len(dataset)):\r\n example = dataset[i]\r\n # do something\r\n```\r\nWhen you do that, one and only one example is loaded into memory at a time.","embeddings":[-0.1967024505,-0.4120611846,-0.1522585452,0.1939141899,0.0503938869,0.0091604656,0.2521877587,0.0610394888,0.1378219873,-0.0013328588,0.2277509272,0.2708903253,-0.4278086424,0.1769471169,0.2024409622,-0.088955082,0.0129831638,0.0472291596,-0.6187960505,-0.0078373151,0.1369362026,-0.0778279081,-0.3001040816,-0.1672134846,0.0571090728,-0.4246666431,-0.310051322,-0.1855567694,-0.1906789541,-0.4267381132,0.2655524015,0.1786016971,0.4127590954,0.5190775394,-0.0001068608,-0.2087260485,0.1724295765,-0.0724189803,-0.1570537984,-0.2525888681,-0.1945832968,-0.1432822496,-0.0132758319,-0.3930830359,-0.0778554678,-0.2569579482,0.1618881971,-0.3726496398,0.2177618146,0.0823897272,0.2613908052,-0.2673082352,0.0140859205,-0.0343472809,0.0291362312,0.1420092136,-0.1895926595,0.3146386445,0.2581835389,0.2592709064,0.0707307905,-0.0838772431,0.1835328341,-0.0293516405,0.5176247358,-0.1810382158,-0.0284579135,-0.3781050742,-0.1316801459,0.1800668538,0.9738004804,-0.3325405419,-0.180568099,0.092509523,0.07839486,-0.4161494374,-0.0789299309,0.2087811977,0.0914726183,0.1689748615,-0.3774550557,-0.2158584297,-0.2636704743,0.15692693,-0.0114712836,0.0202188417,-0.1483395249,0.1454471201,0.1469690949,-0.2227167934,0.3494539261,-0.0374085978,0.095672451,0.2934467196,-0.354888618,-0.3146653175,-0.0452603213,0.4222278595,0.4786589742,0.3207339346,0.2023726702,-0.1110485569,-0.2233713269,0.2842499614,0.4751828015,0.0039901678,-0.0489843749,0.0934275985,0.1161991507,-0.0192664526,-0.1462731361,-0.0210601743,-0.2141522914,-0.078951031,0.3076207042,-0.2920466959,-0.1230150163,-0.0697307587,0.0273173954,-0.1093935445,-0.2524139285,-0.1942876875,0.2612032592,0.2329009324,-0.1484914273,0.1340863705,-0.3233181238,-0.0912753865,0.1094379202,0.2910441458,-0.0449736118,0.4422123134,-0.5005634427,0.0238534585,-0.0357486568,-0.2278592587,0.2085994482,0.1532032192,-0.1350090951,-0.107849583,0.1483562887,-0.3669768274,0.4331787527,-0.0354758203,-0.1922498643,0.0064364015,-0.2999882102,0.1547482908,-0.3955385685,0.0480349436,-0.2260352522,0.0580670498,-0.0429606624,0.1644735336,-0.2056888938,-0.2527205944,-0.0254761502,0.3109799922,0.121580705,-0.1092042178,0.1143796593,-0.2127530575,-0.1631635427,-0.3115358651,0.0765097141,0.2239909023,-0.3857717216,0.1330495179,0.0314944386,0.0469468571,0.1477143019,0.2904067338,-0.2406658679,0.2274698466,-0.1616248339,-0.118223168,0.418687731,-0.1354309469,-0.1791634858,0.3351472616,0.201632306,-0.0980946347,0.0681597963,0.29716748,-0.0122090001,0.0224211458,-0.2033890784,0.2134364694,0.0243260618,-0.1471536309,-0.0023533842,-0.0215294175,0.1723941714,0.0104990965,0.0406940207,0.2140854299,0.0902528614,0.1390072256,0.2731764019,-0.2391561717,0.1900429875,0.2386129946,0.3220140636,0.0492032021,0.0945582017,-0.4822582603,0.0254878756,0.155633688,0.186152339,-0.1578350812,0.2956633568,-0.0067588552,-0.1262248307,0.0798645169,-0.1618826538,-0.056561932,0.125680536,0.0663363412,0.2056292892,-0.0901576057,-0.5047391057,0.0998199582,-0.2030547559,-0.2080510706,-0.5096104145,0.1290673614,0.2270847112,-0.0779467896,0.0327608138,0.015286563,-0.2983108163,-0.497923255,0.2777377367,0.3969455063,0.1560833603,-0.007922153,0.1019600481,0.3819927275,0.3212147653,-0.1746238023,0.1462098658,-0.1991002262,0.2882044315,-0.0181854982,0.1007042602,0.4194480777,-0.2194589972,0.0578731149,0.3707637489,-0.1672394127,0.0571276546,-0.0573669299,-0.4465524256,0.1243522912,-0.1040654853,-0.1959182024,0.4206411839,-0.1265025288,-0.3818365335,-0.3234954476,0.4590500295,-0.0309022348,0.0291666482,0.0501575768,-0.1086960807,-0.0667860508,0.0552263297,0.0299283322,0.3643385172,-0.0167238824,0.2228473276,0.0211025793,0.0027691489,0.1936891824,0.1838862598,-0.021365894,0.3514572382,-0.167028293,-0.0330475196,0.023264816,-0.2684541643,-0.3299869895,0.0165297929,0.1501984149,0.0575455762,0.2963264883,-0.4058513641,-0.1907162964,-0.3694969416,-0.1114165112,0.2052114159,0.1090968251,-0.3205469847,0.042428676,0.4008634686,0.1220249757,0.0850847438,0.2148122936,0.058388222,0.0247809552,-0.0603260696,0.0091481637,-0.1000981703,0.1860867292,0.287830472,-0.2173032165,0.6166129708,0.0909842849,-0.0911396667,-0.4059427977,-0.1621184647,0.1468191445,0.0502733923,0.0921218991,0.2490787804,-0.0362272821,-0.1754505932,-0.0418330804,0.0035600148,-0.4967641532,0.1590124071,-0.1841693074,-0.1201962382,0.0413623303,0.0619924329,-0.1926466972,-0.3554748595,-0.1489828676,0.3245443404,0.394795239,0.179288134,0.1702627689,0.0526030511,0.3306016028,0.1137496009,0.0347363278,-0.0312977396,-0.3998983204,0.1823969632,-0.3657858968,-0.2789263725,0.0530234016,0.0544123314,0.4394408166,0.1919809729,-0.3579131067,-0.026459612,-0.0615793876,0.0533848852,0.0929170847,0.394934237,0.0913500488,-0.2753692567,-0.0266986489,-0.0448271036,-0.0533393584,0.0777137876,-0.0599412173,0.3174922168,0.5418897867,0.3369597793,-0.0612844191,0.5106337667,0.3213931024,-0.4664492011,0.381926477,-0.2949079573,-0.2295010686,-0.1006760746,-0.1804185808,0.1238948107,-0.253370434,-0.2882227004,0.137610808,-0.0853164792,-0.50462991,-0.2860321701,0.1899845749,0.0075041526,0.0854132921,-0.0680201277,-0.4584963918,0.2238840163,-0.3123108447,-0.1672614366,-0.083959043,-0.035122592,-0.003793183,0.1022480577,0.1063194647,0.0255276337,-0.6367225647,-0.0780711919,-0.2195491344,-0.0160044897,-0.2293444127,0.228407383,0.2031735778,-0.1550876796,0.2090588808,-0.2704061568,0.346450597,0.025589332,-0.369551152,-0.0154665448,-0.3388346136,0.1033637822,0.1139863953,-0.2592841089,0.502161622,0.1661722064,-0.0499618091,-0.0905011445,-0.2015234083,0.2595628798,-0.0000613956,0.0583487712,-0.4264299572,0.091269955,-0.1913363338,-0.0521197766,-0.1042881459,-0.0256593246,-0.1569271088,0.0598533899,-0.2274881899,-0.2387855798,0.0925383791,0.2368409783,-0.0191253014,-0.0704301596,0.0218055043,0.1799086481,0.1561618298,0.2901681364,0.2923647165,0.2780869603,-0.0688935816,0.0968551561,0.2517821193,0.2262899429,0.329223305,-0.0746600032,-0.0640478879,0.0011868814,-0.0950278789,-0.0137512879,-0.4295496941,-0.0702778548,0.3990368247,-0.028546704,-0.68493402,-0.8001820445,0.5862554908,0.2033589035,-0.0424136072,0.4010373354,-0.1833683252,-0.1230541617,-0.0882995725,0.1937609166,0.7383261919,-0.1139420941,0.0490077212,0.1444425881,0.3099934459,0.4590798616,-0.0163733307,-0.053081803,-0.4539552331,-0.1515319645,-0.1391632408,-0.1917885691,0.2685103714,0.1863999516,-0.2638259232,0.3906111121,0.1017133668,0.0696840733,-0.157968089,0.3008527756,-0.2609703243,-0.0858927295,-0.332444787,0.2069453746,0.1299761385,0.1009472087,-0.1899981201,-0.0860715061,0.3169947565,0.2821399271,-0.1255443841,-0.1276728362,-0.1094578132,0.0689666495,0.1239277273,-0.2988598347,0.4001433849,0.2613204122,0.1422267407,-0.060414955,-0.1662522852,0.0632447675,-0.0975134,-0.0383406021,0.0007142883,0.1794935614,0.4892555475,-0.0808953494,-0.0325543098,0.2623422146,0.1185246333,-0.035216473,-0.1141913608,0.1386711299,0.24337852,-0.1255544573,-0.2266969234,0.4080357552,-0.0530634634,0.0059413891,0.1651062816,0.2601028681,-0.1499896646,0.4505726993,-0.0252892524,-0.0643577203,-0.1876547039,0.1663488299,0.2450580001,-0.0924245119,0.222571969,0.0322860815,-0.177978605,-0.2087043524,0.2494393587,0.2526237071,-0.04920201,0.0156585909,0.0679647401,0.2703877091,0.2758464217,0.1312400848,0.1232998148,-0.1012083292,-0.3609986007,-0.1127125621,-0.3696655333,0.010942528,0.0922776163,0.2632908821,-0.0785031021,-0.2278162539,0.2112400085,0.0889503509,-0.3270403445,0.0319067091,-0.1110570356,0.1568697542,0.0402873084,-0.0438270867,-0.2527572513,0.2932388186,0.1048413664,-0.0036788182,-0.3037315607,-0.2058586925,-0.0077897813,0.0933289453,-0.0342214629,-0.1223050356,0.127613917,-0.3852511644,-0.2536253631,-0.1441391706,0.16458112,-0.0073971925,0.0032501484,0.2268621325,0.230196923,-0.3531453013,-0.0700021163,-0.0088679353,0.1202048361,-0.1476109475,0.106346339,0.2929705977,0.1295568347,-0.3645478189,-0.4189191163,0.2302254438,0.2601175606,-0.226623252,0.4598793387,-0.1836043,0.2835589647,0.1864102036,0.3899705708,0.5834115148,-0.1510751992,0.0721962675,0.1697528958,0.2021749914,-0.0104207713,-0.061800845,-0.0680201575,0.0693430081,0.0310970694,0.1741318256,0.1325970888,0.139788568,-0.3049253523,-0.3196588159,0.1743733138,-0.4326548874,0.0685199797,0.3156374693,-0.0505417325,-0.0920173898,0.3491539955,0.5144851804,0.2512197495,0.4923084974,0.2163974196,0.5837105513,0.1487667859,0.0795086473,0.1027701572,-0.2135764211,-0.0669477358,0.3401177227,0.1138995513,0.1707929522,-0.0026380857,0.2510223389,0.2171026766,-0.1438946575,-0.1152108535,0.1553004086,0.0560763255,-0.2364278734,-0.2033428401,-0.1576825529,-0.1710404158,-0.0708007142,-0.3268954456,-0.0628488734,0.3587143123,0.0965186507,0.2263881713,-0.3642051816,0.4352072775,-0.1042165458,0.3328238726,-0.3009596765,0.2857184112,-0.2362217158,0.1262673289,-0.3357145786,0.1477476209,0.1616248637,0.4522745013,0.0784211829,0.0637700111,-0.1715228707,-0.1346343458,0.0348259322,0.1807406545,0.0055234395,-0.1503407061,0.4830392599,0.0787823498,-0.0299641844,0.4317764044,0.0392986611,-0.0490114912,-0.7619783282,0.0308075231,-0.0738832951,-0.104082182,0.0386139229,0.2058888078,-0.3588980436,0.0310157631,0.3494985402,-0.5118250251,0.0112846233,-0.1281151623,0.0908081457,-0.4438216388,0.580640316,0.2358208299,0.3683755696,0.0646326169,-0.3547824621,-0.2148440033,-0.0000025318,-0.0646088049,0.033011809,0.0023924427,0.2600568533,0.0479090884,0.0815375671,-0.2101940215,-0.4015119374,-0.3318651021,0.471547097,-0.1407267004,-0.1185669601,-0.5041664839,0.4238241613,0.0688322559,-0.1372979581,0.0028389469,-0.4769554734,-0.0566718727,-0.0586554334,0.1339170039,0.1711046845,-0.2130333483,0.4344698191,0.0502153039,0.1733374596,-0.0409561358,0.0506237596,-0.2002869099,-0.2317830175,-0.0560711771,0.4207938313,0.0936154053,0.0182340853,-0.1480523348,0.0300531555,0.150640294,0.0072694384,-0.0172868595,0.025136292,-0.0987890363,-0.4087807536,-0.0384174548,-0.014157746,0.1612665951,0.722605288,-0.1982648373,-0.0007436985,-0.1763432622,0.055588495,0.1908268183,-0.1202918291,-0.1176186353,-0.1891219467,-0.0056447238,0.0931235477,-0.2571721375,-0.4581366479,0.1177974939,0.4690542519,-0.0787140802,-0.1997425556,0.3155782819,0.2660444975,0.0640991032,-0.0137256635,0.3568969667,-0.1605868042,-0.1009712517,0.0214620419,-0.2646002769]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/815","title":"Is dataset iterative or not?","comments":"Hi there, \r\nHere is what I am trying, this is not working for me in map-style datasets, could you please tell me how to use datasets with being able to access ___getitem__ ? could you assist me please correcting this example? I need map-style datasets which is formed from concatenation of two datasets from your library. thanks \r\n\r\n\r\n```\r\nimport datasets\r\ndataset1 = load_dataset(\"squad\", split=\"train[:10]\")\r\ndataset1 = dataset1.map(lambda example: {\"src_texts\": \"question: {0} context: {1} \".format(\r\n example[\"question\"], example[\"context\"]),\r\n \"tgt_texts\": example[\"answers\"][\"text\"][0]}, remove_columns=dataset1.column_names)\r\ndataset2 = load_dataset(\"imdb\", split=\"train[:10]\")\r\ndataset2 = dataset2.map(lambda example: {\"src_texts\": \"imdb: \" + example[\"text\"],\r\n \"tgt_texts\": str(example[\"label\"])}, remove_columns=dataset2.column_names)\r\ntrain_dataset = datasets.concatenate_datasets([dataset1, dataset2])\r\ntrain_dataset.set_format(type='torch', columns=['src_texts', 'tgt_texts'])\r\ndataloader = torch.utils.data.DataLoader(train_dataset, batch_size=32)\r\nfor id, batch in enumerate(dataloader):\r\n print(batch)\r\n\r\n```","body":"Hi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks","comment_length":113,"text":"Is dataset iterative or not?\nHi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks\nHi there, \r\nHere is what I am trying, this is not working for me in map-style datasets, could you please tell me how to use datasets with being able to access ___getitem__ ? could you assist me please correcting this example? I need map-style datasets which is formed from concatenation of two datasets from your library. thanks \r\n\r\n\r\n```\r\nimport datasets\r\ndataset1 = load_dataset(\"squad\", split=\"train[:10]\")\r\ndataset1 = dataset1.map(lambda example: {\"src_texts\": \"question: {0} context: {1} \".format(\r\n example[\"question\"], example[\"context\"]),\r\n \"tgt_texts\": example[\"answers\"][\"text\"][0]}, remove_columns=dataset1.column_names)\r\ndataset2 = load_dataset(\"imdb\", split=\"train[:10]\")\r\ndataset2 = dataset2.map(lambda example: {\"src_texts\": \"imdb: \" + example[\"text\"],\r\n \"tgt_texts\": str(example[\"label\"])}, remove_columns=dataset2.column_names)\r\ntrain_dataset = datasets.concatenate_datasets([dataset1, dataset2])\r\ntrain_dataset.set_format(type='torch', columns=['src_texts', 'tgt_texts'])\r\ndataloader = torch.utils.data.DataLoader(train_dataset, batch_size=32)\r\nfor id, batch in enumerate(dataloader):\r\n print(batch)\r\n\r\n```","embeddings":[-0.3016710877,-0.3916234076,-0.1244997233,0.404502511,0.1218660474,0.0815260336,0.2345808595,0.1949106157,-0.044544179,-0.0258444548,0.0440256149,0.3346882463,-0.3416885436,0.019357549,-0.0354218148,-0.1522007585,-0.1350665241,-0.1091265082,-0.5712974668,-0.0941947401,0.164425537,-0.0607798435,-0.3945014477,-0.1852220446,-0.0154445795,-0.4847174585,-0.2174984962,0.0037641965,-0.1276628673,-0.3302404881,0.1825329959,0.0282143466,0.4424453378,0.6334045529,-0.0001107979,-0.1362617463,0.1161465049,-0.1768816561,-0.1535315514,-0.3136429191,-0.0109618222,-0.1165110692,-0.0174039751,-0.4819852412,-0.1775075346,-0.129210785,0.0617713556,-0.5960788131,0.3605302274,0.174767822,0.1965260357,-0.1608913243,0.0373760648,-0.0517343171,0.1932127476,0.1550517082,-0.088112399,0.2196961045,0.4312990904,0.2817698121,-0.015033531,0.0472457521,0.0915912762,-0.0721105114,0.1525731236,0.0000492522,-0.2120264024,-0.4927812517,-0.0110825384,0.1845875531,0.9366436005,-0.2746395469,-0.2969922721,-0.0351296812,0.016817661,-0.1741650999,-0.0689116716,0.2412091196,0.0703459904,0.1441503614,-0.3334394693,-0.1051563323,-0.1860531867,0.2216388881,0.0751473308,0.2418799847,0.0030301763,0.2312335819,0.0462208837,-0.2195009887,0.4654043615,-0.1329292804,0.0892364904,0.3992209733,-0.4541883767,-0.1844446361,-0.0349401422,0.2144804597,0.4375275671,0.261620611,0.1651032269,-0.1079955995,-0.2650260627,0.2307070792,0.2453352213,0.0775087997,-0.0132842967,0.2721357644,0.0922990367,-0.1410165727,-0.010839927,0.0221032053,-0.2615159154,-0.4489365816,0.0846842527,0.0368227996,-0.1015720591,0.0349570476,-0.0030915847,-0.2123387456,-0.2890574336,-0.2483493537,0.3508866429,0.2104179114,-0.211345911,0.0223543681,-0.1101625934,0.1763292104,0.2847446799,0.3477299511,-0.0633524805,0.4023623466,-0.4622748196,-0.0274971724,0.0432804003,-0.3923403323,0.1161287427,0.1986756623,-0.1707923263,0.0677612275,0.1151254922,-0.3227116168,0.4243600965,-0.1387716681,0.0229630731,0.0071933516,-0.0757018775,0.2180230319,-0.3226849139,0.0705003589,-0.4101574123,-0.0100664478,-0.1824006885,0.1245952249,-0.0685806647,-0.1054999307,-0.1791855246,0.1854764372,0.2596293688,-0.144537881,0.2509702146,0.0507896431,-0.1488247514,-0.2530644536,0.0971306488,0.3796824217,-0.3504396975,0.0436405204,0.3202689886,0.1515729278,0.154348731,0.3483071029,-0.3904181421,0.2539191842,-0.3402405083,-0.1352609396,0.7417567372,-0.363509506,-0.1803584248,0.2364930212,-0.0599689409,-0.2498018891,0.3100492656,0.0061494638,-0.0729998127,0.0717940256,0.192838043,0.1711915284,0.0321159698,-0.160643518,0.0101480642,-0.0285592396,0.3351587951,-0.025440501,0.0790984035,0.1570300907,0.2761316895,-0.0619236082,0.4041700065,-0.2901450396,0.1881703138,0.2839535177,0.0180364363,-0.1514426917,0.051079087,-0.1703205407,-0.2152455151,0.2737480998,0.0589009672,-0.1949934661,0.110076502,0.016197836,0.0288132206,-0.1400697231,-0.2077917606,-0.0491939299,0.0650475249,0.123596929,0.2050937563,-0.2752660811,-0.3432314694,0.0536302775,-0.2269364744,-0.1522424072,-0.5957727432,0.2830938101,0.1569339335,-0.1839175373,0.0886456445,0.1052316949,-0.0220429972,-0.3450835049,0.1771088541,0.3911457062,-0.0403560884,0.1695431471,0.1876871288,0.250310719,0.3719840646,-0.043576315,0.1544352621,-0.3036029041,0.2019958645,-0.1644172966,0.1088795736,0.4175557792,-0.178999722,0.1390865594,0.2945196033,-0.162000522,0.0500433519,0.0338672101,-0.3456018567,0.095110774,-0.1028606743,0.0285397004,0.4338846505,-0.2218484432,-0.4662246704,-0.2694395185,0.3125096262,0.0944652632,0.0498280376,-0.0394720137,-0.1615007073,0.0070869224,0.2433333844,-0.2581147552,0.4800901115,0.0130743021,0.2423738688,0.0065550869,0.1503530294,0.0643304437,0.1935148835,0.0452768989,0.3508191407,0.1702409238,0.2046918422,0.1170993447,-0.0401206091,-0.3137160242,0.0163484886,0.0010911541,-0.1328433156,0.2385167927,-0.2977777123,0.0241626892,-0.4486703873,-0.1598946303,0.1747912467,0.0770025253,-0.3366315067,0.0492514819,0.4261993468,0.1866766065,0.1212245077,0.0822213888,0.3265725672,-0.1102489829,-0.0354747139,-0.1402110308,0.0021169186,0.146857366,0.2564623952,-0.3135670125,0.51144135,0.0465419665,-0.0001485249,-0.5351338983,-0.1205405667,0.3685362935,-0.1301970929,-0.0973266214,0.293846637,-0.1601385474,-0.1209717691,-0.1281894147,0.0786629468,-0.2453061193,0.0866097286,-0.1101557016,-0.2211600393,-0.0738406703,0.0910160169,-0.334584415,-0.513856709,-0.2599232495,0.2739360332,0.3042732179,0.1782877147,0.1947641522,0.1513937712,0.4163945913,0.2766897976,0.1746632308,-0.1345507801,-0.5078074932,0.4365925789,-0.443395555,-0.3223357201,-0.0671115518,-0.0891141891,0.4015447795,0.2927371264,-0.2275419384,-0.0919914469,0.064163357,0.2196813971,-0.0644264668,0.4492762685,0.3778802454,-0.1747938544,-0.0332650058,-0.0464871079,0.1320675015,-0.018378688,0.0769087747,0.5140583515,0.4622370303,0.3192089796,0.056447994,0.7233349681,0.236876145,-0.4231692851,0.4203618467,-0.291253835,-0.2060451359,-0.1417576224,-0.3200324178,0.1163159609,-0.267215699,-0.2039099485,0.0829527453,-0.2075923234,-0.4504363835,-0.3630998731,0.1615891606,-0.1345400363,-0.0364124663,-0.000896136,-0.4431040585,0.1519376934,-0.3746412396,-0.0247338451,-0.1444494277,-0.0664899126,0.0708446205,-0.0481183827,0.0268070456,-0.0074907318,-0.531866312,-0.0776732713,-0.2226397693,0.2868033648,0.0426202193,0.0928513408,0.3301154673,-0.1033716798,0.1380605698,-0.1811048239,0.5042600036,-0.0617236346,-0.2357168943,-0.0273928698,-0.2021422833,-0.019855177,0.0730180442,-0.2566608191,0.4214020371,0.3614584804,0.1236881018,-0.1362921894,-0.2726887763,0.1096011624,-0.2279167026,0.1190546751,-0.2491511405,0.0845934302,0.0185749456,-0.2972394228,-0.1471707076,0.0812163949,0.01962924,0.0006558812,-0.1931228936,-0.2530002892,0.0150098549,0.1761227846,-0.0630650818,0.1344305873,0.1655200571,0.1640261114,0.3385930657,0.4019482434,0.3491041362,0.45746243,-0.0497171953,0.1240924522,0.384007901,0.1247086674,0.345828265,0.1229038239,-0.1199513301,-0.0857483,-0.3532609046,-0.0542011708,-0.3154107332,-0.0091281589,0.3837417662,-0.0855122879,-0.6794344187,-0.7689771056,0.4015054405,0.2226498574,-0.1336444318,0.3440219164,-0.3819896579,-0.31832847,-0.0910502672,0.2710523009,0.8402193189,0.0322991833,0.152525425,0.0226314813,0.3366403878,0.2843594849,0.030588029,0.0468822271,-0.5196399093,-0.1365142167,-0.2416654974,-0.3081583679,0.2897690535,0.2748749852,-0.3071196377,0.4338349998,-0.0896795616,0.2778542042,-0.1443514079,0.2908053398,-0.1449635774,-0.212260142,-0.3896154165,0.0986081809,0.1021098793,0.1109267622,-0.1567333341,-0.1271937042,0.0333908498,0.0483811982,-0.0917539895,-0.0442286134,-0.5501815677,0.2739236355,0.1598562002,-0.5914463401,0.29468894,0.6760756969,0.1154297292,-0.0119798779,-0.1408447325,0.0384407081,0.2060934901,-0.1562191844,-0.0725082979,0.0975352079,0.4758128524,0.0528442487,0.0048448704,0.4223836064,0.139708057,-0.0095873922,-0.1689722091,0.113389343,0.2828720808,-0.278445363,-0.2333352119,0.3246432543,-0.0805046856,0.0151668675,0.0909759253,0.2167088091,-0.167538166,0.2051581442,0.1029292569,-0.0848167986,-0.0763310269,0.4706707299,0.1202263013,-0.0981323645,0.2479032129,0.1295380294,-0.1596693248,-0.123474516,0.2534741163,0.2145289779,-0.1498476416,-0.1539314538,0.0741664171,0.1245174855,0.2328958362,0.2849791348,0.1200711653,0.0664470196,-0.1610537618,-0.1804712117,-0.3888456523,0.0610883385,0.1004843637,0.3253470063,0.1052879691,-0.2224452496,0.0293283928,0.0768356174,-0.2823282778,0.0249238163,0.0109968726,0.1962828636,0.1364853978,-0.1093957126,-0.1034339741,0.1355661303,0.0608205162,0.1372398585,-0.2650090754,-0.1954613328,-0.0620849133,0.0954936966,-0.010230164,-0.1716787964,0.0155682229,-0.4823229611,-0.1556379646,-0.1742751151,0.2260721326,0.097296007,0.1488031298,0.0438317582,0.2419628203,-0.2362684608,-0.0809502155,-0.05541116,0.0508827381,-0.1584217846,0.1208490133,0.3051259518,-0.0701842606,-0.2784437239,-0.2976560593,0.1620580703,0.4519222081,0.067350179,0.2465281934,-0.2574517429,0.0997181907,0.0909810215,0.2638572454,0.4230271876,-0.1600360721,0.0520504974,0.1256247908,0.1248597354,-0.0970332772,-0.0173532106,-0.2267062664,0.0293075219,0.1059300974,0.2214537412,0.0797630027,0.1818278134,-0.3765723407,-0.2088661045,0.2212145627,-0.3223404884,0.1888606101,0.3121754527,-0.0248018466,-0.1189256459,0.3131517172,0.4781965911,0.1840549111,0.1127870381,-0.0069750352,0.5928164721,0.2599864602,-0.0700435415,0.1810557991,-0.2384407818,-0.1235950738,0.1346114427,0.0786680952,0.176468119,-0.1358975917,0.5692288876,0.2992441654,-0.1604916453,0.0906637385,0.3686591089,0.1135782823,-0.2999249399,-0.3345406651,-0.23540245,-0.13830176,-0.1145283207,-0.2687354982,-0.0343326256,0.3106895387,0.152648598,0.305891633,-0.2136776745,0.3738525808,-0.0776967034,0.1008843929,-0.314216733,0.348085165,-0.1641646475,0.1437532604,-0.2153377384,0.0832309648,0.2506156564,0.3648097813,-0.1423474699,-0.0112884566,-0.0685293525,-0.1197921261,0.1105622202,0.0994916186,0.0256878063,-0.3752378821,0.5009596944,0.0727348551,-0.0831472203,0.336227268,0.0254892893,-0.0860902071,-0.5171759725,0.0285144951,-0.0760780275,-0.1616054326,-0.00296538,0.0322173126,-0.3494994938,-0.0534799919,0.3596718907,-0.4814479649,0.1096550003,-0.0712440312,0.0676145256,-0.2086927742,0.5480344892,0.2112808377,0.3716069758,-0.0438836478,-0.170191139,-0.1988397986,-0.0692152902,0.1777981371,0.0427609794,-0.1237589344,0.2416922599,0.0362842046,0.1109809428,-0.0082404613,-0.5410219431,-0.2205155641,0.4227229357,-0.1820371002,-0.2194939256,-0.3366210759,0.4020093679,0.0884030461,-0.3076234162,-0.0263846181,-0.3074908257,-0.0361731313,-0.2024129629,0.0237072352,0.0249545053,-0.274312526,0.5059764981,-0.1371536404,0.3993719816,-0.1935700029,0.1127053276,-0.2125445455,-0.3714738786,0.0528383665,0.3700617552,0.2980939448,0.2023140937,-0.2437048256,-0.0731872395,0.0878390148,0.0888534114,0.1197973341,-0.074353531,-0.1931867152,-0.2361485809,0.0032977795,-0.0590877421,-0.0359152518,0.6751527786,-0.1143776327,0.2119201422,-0.1191649884,-0.349486798,0.432163924,-0.2436609864,-0.1623469144,-0.0897048935,-0.1091445759,0.2427253872,0.0134386914,-0.4573644102,0.0917274505,0.4157389104,-0.0691374615,-0.1967609078,0.1873580962,0.2847016454,0.0736471266,-0.1682554781,0.1744580269,-0.1274974495,-0.1221860796,-0.1616955101,-0.3481247723]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/815","title":"Is dataset iterative or not?","comments":"closed since I found this response on the issue https:\/\/github.com\/huggingface\/datasets\/issues\/469","body":"Hi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks","comment_length":10,"text":"Is dataset iterative or not?\nHi\r\nI want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not?\r\ncould you provide me with example how I can use datasets as iterative datasets?\r\nthanks\nclosed since I found this response on the issue https:\/\/github.com\/huggingface\/datasets\/issues\/469","embeddings":[-0.2152758986,-0.5996425152,-0.1612304151,0.1126579121,0.1488802284,-0.1371172965,0.3341105878,0.0631924421,0.2045614421,0.0755053163,0.0885693133,0.0022733158,-0.3480103314,0.4717803001,0.1587853879,-0.1499307901,0.0093081165,-0.0134526547,-0.5748156309,-0.0561601147,0.184924677,0.0650710762,-0.2007485032,-0.1678477824,-0.1560272872,-0.4805601239,-0.2219743133,-0.0610372871,-0.2367068082,-0.2581875026,0.2977351248,0.2689963281,0.2297496647,0.5465207696,-0.0001047937,-0.2161935568,0.1410544217,-0.0625231043,-0.0325906165,-0.1083294377,-0.1091677099,-0.0637971386,-0.0796224996,-0.2628715038,-0.2246769667,-0.0043112552,0.0901952386,-0.2066560686,0.5462217927,0.0391305313,0.297187835,0.0629623085,0.0768744797,-0.0921791345,-0.0433107354,0.2242321074,-0.2352082729,0.2959604263,0.2492104471,0.1954205185,-0.0449999757,0.1281408072,0.2909298837,0.0309585985,0.4114084244,-0.1356785744,-0.1754536033,-0.4160466492,-0.0771513358,0.2759281695,0.7258728147,-0.2228942513,-0.3173404634,-0.0662528425,0.0717386231,-0.3945492506,0.0459965207,0.1769479364,0.093819581,0.1604771316,-0.3702470064,-0.1908382177,-0.2346430868,-0.0495805778,-0.1212572455,-0.0138891982,-0.3397886157,0.0281387027,0.2128381729,-0.2271791399,-0.058534719,0.0707031563,0.1181562096,0.2583300471,-0.4387018681,-0.3300882578,-0.0370231718,0.4651817083,0.5778976083,0.4266967773,-0.0103043728,-0.0903238729,-0.2511773109,0.2119076848,0.4284005165,0.0880163684,-0.0245499816,-0.1571933031,0.2834160626,0.0642002821,-0.0413722061,0.0346936844,-0.0617175065,-0.0963122025,-0.059508305,-0.3324483931,0.0101454472,-0.1420283467,-0.0294421297,-0.0610429309,0.0302253831,-0.2405381352,0.2299057692,0.2534013689,-0.183554545,-0.0468193777,-0.2542334795,0.025128223,0.0660174713,0.231658563,-0.1997627169,0.290564239,-0.4116564691,0.1604239643,-0.068009302,-0.4969812632,0.2917555571,0.123172082,-0.0017924911,-0.0585003272,-0.1029027775,-0.2051713318,0.1504358053,0.1062185317,-0.1772257835,-0.0374935456,-0.2601611018,0.1801558882,-0.2292867452,-0.0973706916,-0.2034556419,0.0198882557,-0.0688633919,0.1673200577,-0.2797970176,-0.2076657265,-0.1374002993,0.3610751033,0.0364848413,0.0176999047,0.2064676434,0.0000392107,-0.1937201023,-0.1724933386,0.1075326279,0.2720303535,-0.1356485933,-0.1060031876,0.2185147107,-0.0792981535,0.0488869175,0.4418260157,-0.1821234077,0.092704542,-0.150569424,0.0765118003,0.3107453883,-0.1268969923,-0.2871600986,0.1667310447,-0.0245886073,-0.1326002777,0.0474980064,0.120219931,-0.0764306188,-0.0486267284,-0.237215206,0.2070730925,0.0786925852,-0.1144505963,-0.0866860375,-0.0909329653,0.0313999467,0.1205913126,0.0520440862,0.1835703254,0.1397544742,0.0574111305,0.2184494585,-0.2012766451,0.1954911351,0.2131354511,0.3367874324,0.0918544009,0.0428990349,-0.3053662479,-0.159826681,0.1047062352,0.07639613,-0.0549799129,0.2955843806,-0.1340450197,-0.0931983888,-0.0504129864,-0.1226054057,-0.1080619618,0.1436538249,0.1470201761,0.2288975418,-0.0355649516,-0.6270238757,0.2605917454,-0.3618001938,0.0347756259,-0.5840166807,0.2425440103,0.2053309083,-0.1069587171,0.1685209274,0.1320337951,-0.196978122,-0.5065418482,0.2801898122,0.271517992,-0.1400927305,0.0883978009,0.1582023948,0.2880131006,0.3519719541,-0.0704012066,0.0227929167,-0.3130035102,0.1572205275,-0.044308871,0.0485763215,0.3968121111,-0.176180467,-0.1274739951,0.3424718082,-0.1971954107,0.1346330345,-0.1325530261,-0.4134204388,0.1394375712,0.0140238078,-0.2272151113,0.5324184299,-0.1400465816,-0.4881659448,-0.2001370341,0.5166694522,-0.1071473956,0.0478416048,0.1821647882,-0.0070034885,-0.057058312,0.0852095038,-0.0064423834,0.3936980069,0.0842812508,0.208902657,0.1078577265,0.0080041196,0.0162533391,0.0649631694,-0.1074616686,0.1978759021,0.0242406335,0.037690457,0.0059045404,-0.4861201346,-0.1867602021,-0.0349035822,0.131951347,0.0128021268,0.1357516795,-0.3203154504,-0.2432745695,-0.3807289004,-0.085512884,-0.0773845017,0.1594887227,-0.1935440749,0.0439176485,0.2525418997,0.1888554543,0.1936633885,0.4678739905,0.0472121723,0.2341289818,-0.2019315809,0.0988618359,0.0166083034,0.1767798364,0.2016150951,-0.3005748689,0.6726691723,-0.119818382,-0.0291075185,-0.420493573,-0.3258062303,0.3297124505,-0.1312265545,0.1147004887,0.2118051201,0.0674428865,-0.1427493393,-0.0361435786,0.0815010369,-0.4771389663,-0.0114272702,-0.3485866487,-0.0247614793,0.0252385158,-0.0052458267,-0.2058702856,-0.3190422654,-0.1892245412,0.5626165867,0.2507486045,0.0750961229,0.303812921,-0.0977249742,0.2998054028,-0.1765409112,0.0401917025,-0.1801097095,-0.5018218756,0.1360659152,-0.3350024521,-0.3934424222,0.0568163432,0.0882465988,0.3864459097,0.0517952703,-0.4431164265,-0.2531614602,-0.0644650981,0.1172849312,0.0925663486,0.4538231492,0.2060895562,-0.413882643,-0.0695814118,-0.032604292,-0.0915604606,0.0597669296,0.006238075,0.2893308699,0.3880003989,0.3343083858,-0.0721737146,0.5659970641,0.3288727403,-0.3465389907,0.4355629086,-0.300444752,0.0455703251,-0.1686657518,-0.1924561411,0.3106909394,-0.1614097655,-0.0461112522,0.215578258,0.0102795484,-0.1519818604,-0.270739466,0.1217404306,0.0727388114,-0.0058980398,-0.2357281595,0.0082330564,0.2034551203,-0.1889852881,-0.1183208972,0.1036600322,0.0517987907,0.1109300405,0.2961596549,0.2073237747,0.0269493889,-0.5813977122,0.0190993715,-0.4757787287,0.0779286772,-0.1373280585,0.0605312251,0.066035457,-0.1008697003,0.2000738084,-0.2983605564,0.4371704161,0.078101553,-0.3054777086,-0.1143417433,-0.4849513173,0.0681651309,0.1534840763,-0.2067388296,0.2580727041,0.2838666439,0.1905462891,-0.248862952,-0.3588851392,0.3213588297,-0.1427290291,0.0920993686,-0.3375757635,-0.069648087,-0.2756515145,-0.1206185892,0.0901718363,-0.0750861689,-0.0552516356,0.0744487345,-0.1018267423,-0.2015114278,-0.0227678362,0.1020028442,0.0136936931,-0.0155274086,0.1245412454,0.092205435,0.2931227386,0.28480497,0.182130903,0.4279242158,-0.1679588258,0.0609929003,0.2178580016,0.217253536,0.179065913,0.1742561907,-0.0561315455,0.0838849247,0.0384474918,0.1715800613,-0.4720980227,0.0168716405,0.4675679505,0.0829041377,-0.5042950511,-0.5683656931,0.5331327319,0.1404020488,-0.0144992117,0.3822051287,0.2326595336,-0.0061923685,-0.2702794075,0.1103547961,0.7483811975,-0.0063605295,0.0684411004,0.3408314884,0.2270455807,0.5024478436,0.0102231307,-0.0388353914,-0.4723572433,-0.2351185083,-0.108044669,-0.1239193305,0.3379961848,-0.0213004518,-0.1228228435,0.3338159621,0.2805690169,0.2206442356,-0.2130232155,0.2480570972,-0.2527926564,-0.1742379218,-0.5637907982,0.2042285651,0.0995948166,0.3210177422,-0.2327610999,-0.1312089413,0.4170705378,0.0501756817,-0.0038307144,-0.2219552398,-0.2185896486,-0.0752016082,0.3964672685,-0.309821099,0.3217893541,0.1691262871,0.3949807286,-0.0303205252,-0.238571316,0.0254791696,-0.2678332627,-0.1116521806,-0.1306156367,0.1742914319,0.4693537652,-0.1304797381,0.0347726792,0.0390830822,0.1193678826,-0.044996243,-0.1293717474,0.083126992,-0.0686781928,-0.2312103361,-0.0800122097,0.4272654653,0.0364331976,-0.1092041582,0.1894350946,0.2181386948,-0.1039916575,0.2017710358,0.1709693223,-0.0997718573,-0.2415685952,0.3306237757,0.1071775779,-0.1345017999,0.1203261092,0.0240730699,-0.2163225412,-0.2326271981,0.1474823207,0.3765490353,-0.2411607504,0.0611914471,0.1424747407,0.10474433,0.1540437043,0.3729206026,0.1216738448,-0.1576112956,-0.1575553268,-0.1243809164,-0.4211832583,0.0944519937,-0.0360268056,0.2111878693,0.1396916807,-0.1330992281,0.2211217582,0.1105002016,-0.3716565967,0.0186927468,-0.1685322821,-0.0187470093,0.2743391395,-0.2121207565,-0.0111511564,0.2377203703,0.1201309264,0.1228639185,-0.5037324429,-0.2700255513,-0.1003501117,0.0772251263,-0.0669826567,-0.095119454,0.1544835269,-0.2416230291,-0.3176363707,-0.0531137846,0.3666579425,0.0371990912,0.0317752212,0.0455102585,0.1200426817,-0.3969019949,-0.0263632443,-0.0861771628,0.2709459662,-0.1462123841,0.0124408128,0.0888751671,-0.018255813,-0.2418557107,-0.2061897069,0.2545720041,0.4276503026,-0.2344181538,0.3216700852,-0.2949612141,0.1306278408,0.1029652506,0.537135601,0.4382813275,-0.2165292799,0.1006556079,0.2405245453,0.2370661497,-0.1676015109,-0.0299068168,0.06803561,0.134623006,0.1526638269,0.2165888697,0.2483472377,0.1454282552,-0.1190618053,-0.1917932779,0.1568861604,-0.3904901445,-0.0190637056,0.1671920419,-0.1251214445,0.0030781983,0.5188227892,0.4682491422,0.168138653,0.352836132,0.0563288443,0.503099978,-0.0675380379,0.1381747723,0.0976854786,-0.2370652556,-0.0319677778,0.2255084962,0.0488209315,0.0504369512,-0.1456961781,0.5162401199,0.0266897175,-0.2156781703,-0.0859624073,0.3521921039,-0.0761197582,-0.2322155237,-0.1569359601,-0.1396391392,-0.0203533396,0.0834697038,-0.3198937476,-0.0003298144,0.3314307332,-0.0216580275,0.2442578822,-0.3455786705,0.1964780837,-0.0479417481,0.3219679594,-0.144851312,0.5080808997,-0.1460637599,0.0954758748,-0.1539785415,0.4688002467,0.2224887609,0.1814534068,0.2138636112,0.056171868,-0.2036697417,-0.0323775224,0.0786232054,0.3047070503,0.0438766293,-0.126001209,0.4570685625,0.0875641257,-0.143826589,0.3937549591,-0.1479538083,-0.0158204064,-0.7462180853,0.0541969277,-0.2868190706,-0.0511669554,-0.1648920625,0.1908653527,-0.3865896165,0.1470862925,0.1975928396,-0.4294434786,0.075079605,-0.1887974739,0.1062294319,-0.3058880568,0.6627482176,0.155408591,0.2302305102,0.1140995771,-0.3322228789,-0.3297733366,0.1006860659,0.1582124531,0.0115292734,-0.0422525033,0.0417300202,0.0690551698,0.0792736411,-0.0595629141,-0.518104434,-0.0914431885,0.0861059129,-0.2337244004,0.0706972331,-0.4293944538,0.2771669924,0.1030828878,-0.0062660933,0.2162783146,-0.2650651932,-0.0624121465,0.064779751,0.3720040023,-0.1322385669,-0.3581859469,0.2701107264,0.1017169505,0.1403928697,-0.1037809402,0.0039719632,-0.2366079539,-0.2998535633,-0.1468645483,0.277499795,0.1045488492,0.2303896099,-0.0928283632,-0.2631868422,-0.0005328789,0.1299269795,0.1876884401,0.0939373821,-0.0572456233,-0.2315015495,0.0304753035,-0.1648397297,0.2725942731,0.7305060625,-0.1139355525,-0.0975998119,-0.2024051249,-0.0796523616,0.5304425955,-0.2731823027,-0.1362747401,-0.1427194476,0.0424706228,0.4062981606,-0.5290573835,-0.6381687522,0.1681138426,0.2157697529,-0.0160336718,-0.2622496784,0.35910815,0.0679656565,-0.0859552175,0.0005401273,0.4063514471,0.0506887436,-0.1391741931,0.1243764684,-0.2375473529]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/814","title":"Joining multiple datasets ","comments":"found a solution here https:\/\/discuss.pytorch.org\/t\/train-simultaneously-on-two-datasets\/649\/35, closed for now, thanks ","body":"Hi\r\nI have multiple iterative datasets from your library with different size and I want to join them in a way that each datasets is sampled equally, so smaller datasets more, larger one less, could you tell me how to implement this in pytorch? thanks ","comment_length":9,"text":"Joining multiple datasets \nHi\r\nI have multiple iterative datasets from your library with different size and I want to join them in a way that each datasets is sampled equally, so smaller datasets more, larger one less, could you tell me how to implement this in pytorch? thanks \nfound a solution here https:\/\/discuss.pytorch.org\/t\/train-simultaneously-on-two-datasets\/649\/35, closed for now, thanks ","embeddings":[-0.4590135515,-0.2569155395,-0.1316360831,0.1535700411,-0.146100089,-0.0245793071,-0.1302862465,0.3299368024,0.2405078262,0.0473525859,-0.2769912779,0.0925440118,-0.0419583432,0.0891952738,-0.2336437553,-0.4374606609,-0.1186729446,0.1990205646,-0.6851429939,0.0283445269,0.1618569493,-0.050370194,-0.2565020323,-0.3829409182,-0.2362877131,0.0615891963,-0.3117247224,-0.2885652184,-0.1660903394,0.0514835194,0.3491666317,0.3854669333,0.249366492,0.4770977795,-0.0001164486,-0.1989736259,-0.0553743578,-0.1021457314,0.2152364254,-0.2081171125,0.2312368453,-0.1746554673,-0.0831424519,-0.1464643031,-0.2639166415,-0.3018490374,0.180520609,-0.3493152559,0.1506890357,-0.3101222217,0.1196403801,0.2033813149,0.0344788097,-0.0136588691,-0.0344211459,0.1247327924,0.1694005877,0.4164702296,0.5858606696,0.0091664717,-0.0028337904,0.2086452544,0.0264770146,0.1187485009,0.1092470214,-0.0491002724,-0.0306188744,-0.6200716496,-0.3909494877,0.1465875208,0.4665240645,-0.1925663501,-0.0494746901,-0.2848847806,0.0045905975,0.0622103624,-0.3598687053,0.3050809801,0.0166883431,-0.1453465372,-0.3210138083,0.0286287814,-0.2514358759,0.4043123424,0.1598320007,0.3258551061,0.1248598546,0.2239608318,0.1408023238,-0.0728942975,0.1571153849,-0.1926288903,0.2154241949,0.1570694447,-0.5519158244,-0.4109590948,-0.4993471205,-0.2340147048,0.5322319865,0.3471311331,-0.187056005,-0.0425541289,-0.0106803719,0.1553040743,0.493912518,-0.2302529365,-0.2479529679,0.2176077068,-0.0150042642,-0.1472533047,-0.4452621341,0.136173293,0.0788439736,-0.463506937,-0.0643219575,-0.0870474726,-0.0138893556,0.0515833609,0.1186468303,-0.1892642826,-0.374073714,-0.2427833378,0.0456833281,0.0697074756,-0.290289104,0.0961262286,-0.1152992249,0.233320877,-0.2035884857,0.0054381904,-0.0790842995,0.1522406936,-0.0872852951,-0.0453104228,-0.1197358295,-0.029166542,0.0998595655,0.4736708999,0.5969826579,0.0311333314,0.4532226026,-0.2542555928,0.0366581269,-0.1512842774,0.2842718065,-0.0766827688,-0.0781816468,0.4247112274,-0.2761655748,0.2330382019,-0.3877304196,-0.0946155414,-0.2822934687,0.1199702919,-0.0326744616,-0.2474805415,0.1684399694,0.3227289915,0.3061816394,-0.2528181076,0.1183935106,-0.1277761757,-0.1849128455,-0.0440664403,0.3142673373,-0.0091187051,-0.0216904003,-0.0837211087,0.4510033429,0.1642971635,0.4342624247,0.3265576065,-0.2220416963,-0.0788533688,-0.0924038813,0.1061179191,0.455411315,-0.0873496085,-0.3888515234,0.3634167314,-0.0454780422,-0.0825033858,0.0426882617,0.4256159365,0.1711993515,-0.0739692152,0.0481572747,0.5271052718,-0.3739636242,0.0747983381,-0.0343921743,-0.180992946,0.2601332366,0.3026630282,0.2131896168,-0.135275051,0.0176016912,-0.1681253612,0.4683858454,-0.2271416038,0.1655310839,0.2188633978,0.2946617901,-0.0671575591,0.0300156139,-0.562414825,-0.0695618689,0.0436928831,-0.1190819815,0.0793749169,0.3829195499,-0.0134391049,-0.0250632111,-0.2939044535,-0.2948085666,0.2365103811,-0.003748656,0.1354453564,0.3266372979,-0.4320945144,-0.2283997089,0.1857239306,-0.1022268981,-0.2248753011,-0.4264550209,0.8482157588,0.1254719049,0.0202612095,-0.2163470238,0.410366416,-0.055178456,-0.2960664332,0.1137739494,0.0762057975,-0.2945655584,0.0384500399,0.1549415588,-0.0540652573,0.4266708493,-0.2555561662,-0.0285247881,-0.1260786057,-0.1716243178,-0.1258491725,0.5143619776,0.5528585911,-0.2977360189,0.2221729457,-0.1204144359,-0.0968597159,-0.2046644241,-0.163235873,-0.2405589223,0.1665933579,0.1620083451,-0.2134016305,0.2667861581,0.0517468415,-0.6002414823,-0.0013453455,-0.0609622151,0.0906637013,-0.005993431,-0.2595489621,-0.2178518176,0.2565464377,-0.208522439,0.1145234033,0.4101043344,0.2012648731,0.2420346737,-0.1378100663,0.147369802,0.0811074376,-0.0040912582,0.0383574106,0.101084657,0.4634788632,0.112573117,0.1298897266,0.0568559282,-0.1999635398,0.2216187567,-0.1057793573,0.0837027803,-0.0629427955,0.0910041258,-0.404992342,-0.580503881,-0.3800750971,-0.0159037244,0.2527272105,-0.1305503994,0.3656016886,-0.0044832504,0.4892418087,-0.2300687134,0.3726339936,-0.0304049607,-0.2489747107,0.3056527078,0.1697184592,-0.0830142274,0.110234946,0.2524161935,0.0305253267,0.4819402099,-0.0216745939,-0.0769546181,-0.0806724653,-0.1975476742,-0.0715604797,0.0200668387,-0.1097575426,-0.1743241996,-0.0264538508,0.0653014779,-0.3805688024,-0.0076584951,0.000338317,-0.0582237132,-0.2161251605,-0.2526504695,0.0195645858,-0.0997225195,-0.176742062,-0.4895607233,-0.1905343086,-0.0019720818,0.0319719315,0.2032260597,-0.215494886,0.0192336701,-0.3219474256,-0.1027286798,0.078559652,-0.0221646987,-0.2300736606,0.2679518163,-0.2715791464,-0.1101538539,-0.1774645597,-0.2409712374,0.0959181339,0.0505235456,-0.1432676762,-0.2983009517,-0.1149298549,-0.0284855049,-0.0570444465,0.2852809131,-0.0979376361,-0.1978987455,0.1445046812,0.0659551099,0.1332424879,-0.0285454411,0.2458305955,0.316980958,-0.1584664136,0.068559058,-0.0419910476,0.6387034655,0.6891579032,-0.0133643998,0.1680553854,-0.1482053995,-0.1481812298,0.0435330495,-0.415268153,0.1311030537,-0.1617359668,-0.0605989583,0.1035539508,-0.1726259589,-0.175399676,-0.1913670897,0.078159824,0.070716083,-0.1065017879,0.1812945604,-0.3102547824,0.2101420313,-0.3964982629,-0.3729995191,-0.1092768982,-0.0225323159,-0.1925275475,0.076864861,-0.0970566347,0.1011722535,-0.5187768936,-0.1826519817,-0.2368700355,0.238622576,0.2347919494,0.1025319919,0.3105902076,0.0099596689,0.1163884774,0.0760153159,0.4537573755,0.2854245901,-0.4366332293,-0.2657541335,-0.0393688865,0.0862662196,-0.082317397,0.0658927187,0.2204269916,0.3258359134,0.1089678183,-0.5154570937,-0.1038356945,0.3712570071,0.2485854328,0.0737254024,-0.1921334714,-0.1039717048,-0.1186515987,-0.1777072996,0.2236542106,0.177540347,0.3533562422,-0.0048292554,-0.1435520202,0.0248831082,-0.0708927885,0.3075836003,-0.0662422404,0.1911936998,0.2691352963,0.2734242976,0.1028128788,-0.1503737271,0.1508290172,0.587798059,-0.2100385278,-0.2995359004,0.2010277361,0.2647014856,0.4908477962,-0.0178818703,0.3963025808,0.3406713903,0.1137794107,0.1676844209,-0.0321622454,0.085529834,0.1392858028,0.0904778987,-0.5944808722,-0.6396594644,0.2805025876,0.0542973578,0.2604873478,0.1954322159,-0.2748428881,0.1328367293,-0.0366755426,-0.081562914,0.7256222367,0.2588929832,0.0138490526,-0.1044961587,0.0132402601,0.554042697,0.1455581039,0.2158870101,-0.4279939234,-0.1397251189,-0.091365315,-0.1926000118,-0.2772850692,0.0651671514,-0.0455008782,0.1545953006,0.2299617082,-0.0740974694,-0.1390682608,0.0820553824,-0.131459102,-0.0219845455,0.3188331425,-0.0486223251,-0.2373458743,0.2746399939,-0.1130905822,-0.2959321141,0.3506889343,0.2053000182,0.0067688706,0.0093784481,-0.1496304721,0.0566617958,0.2304524332,-0.5618503094,0.1928917766,0.0697874874,0.1151835397,-0.1081553325,-0.0231616087,0.0732601508,0.1590302587,-0.0655935556,-0.0442397371,-0.1434615403,0.0828067288,0.1305462122,-0.1820186675,-0.0324097387,0.2569283247,-0.3210048079,-0.1817279309,0.1722391248,0.1321440935,-0.1000902802,-0.3920053542,0.5171938539,0.1707257479,-0.0562872626,0.037631698,0.0991983041,-0.258326292,0.4298416078,-0.1064654589,-0.2911566794,-0.1219814867,0.0782055855,0.3611403704,-0.2054260969,0.3913815618,0.1222496554,0.1411295831,-0.0008446894,-0.0902630836,0.2419382632,-0.4024803936,0.0324757621,-0.170641467,0.3743256629,0.1830885708,0.2574179471,0.2789665461,0.139702633,0.1234444454,-0.2993406355,-0.3776442409,0.2612786293,0.1417685598,0.4772782326,0.1080549508,0.2794001102,-0.174595952,0.2990301549,-0.1973573864,0.0904071257,-0.1085323691,0.0774191096,0.0926128253,-0.0657533556,-0.4132948518,0.287997067,-0.0850354955,0.1419041455,-0.0592560433,-0.1021972224,0.1137162298,0.0400133394,0.0755651593,0.0298298169,-0.1040558591,-0.3269004524,-0.0163751487,-0.2206658423,0.2849809825,-0.0973825678,-0.0475429967,0.0215734597,0.1174881384,-0.3671900034,-0.3646775782,0.2974102199,0.2369008958,-0.2097958177,0.0863906816,0.2235998809,0.0423293971,-0.1422901452,-0.2750069499,-0.0734731928,0.4901685417,-0.0633864999,-0.0746469796,-0.0521578975,0.0817006603,0.0486902706,0.5045059323,0.421425283,-0.0967121422,0.1336624622,0.1483854353,0.1701591462,-0.1763801575,0.0077837706,-0.067671448,0.1545149982,0.1697342694,0.2456253618,0.0950164124,0.3434695005,-0.0204802621,0.1516300589,-0.0263392311,0.0610443056,0.3186326921,0.4350640178,0.357224375,-0.2507896423,0.2445887327,0.2167702615,0.187227726,-0.1318941116,-0.2033217996,0.6029132009,0.1382052898,-0.4360667765,0.1805845201,-0.1511229873,-0.0242383927,-0.0370609052,0.2130129188,0.038286496,-0.008799349,0.4180286229,-0.389847368,-0.5202153325,-0.00603934,0.3325134218,0.0526299402,-0.0511745214,-0.1600895375,-0.1617425084,0.0139823919,-0.0818414763,0.0798626095,-0.2376319021,0.2072477341,0.0527861565,0.0634618402,-0.2742441893,-0.0311949421,0.5907273889,0.0763112977,-0.2171730548,0.1595056653,0.065948002,-0.1383862495,-0.0548686273,0.305344373,0.1527727097,0.1308168173,-0.2613473535,0.0349870138,0.045588959,-0.0022485536,0.1264483631,-0.136405617,0.241447553,-0.5410106778,0.4235744774,0.0481930077,-0.0998542681,0.2208713293,-0.01523015,-0.1444849819,-0.6419488192,0.3325235844,-0.0860545337,-0.0371143147,-0.0168096982,-0.1230401918,-0.3494381309,-0.3440912962,0.1656602621,-0.5379909277,-0.1796839237,-0.0250334535,0.0086216917,-0.1740727276,0.3852669299,0.279588908,0.2488848418,-0.2261171192,-0.1871798187,-0.2359597832,0.2867124379,-0.0036478068,-0.0015555449,-0.1920395941,0.281722784,0.0728874505,-0.0352619924,0.0437657535,-0.0451631695,0.1437934786,0.4484840035,-0.2158867568,-0.0629339442,-0.3151337504,0.186258629,0.2024557739,-0.037040703,-0.0153200114,-0.0225348249,-0.1281269044,0.1334404796,-0.0225430038,0.5126491189,-0.2624083757,-0.0492696911,-0.0562293865,0.6781873107,-0.1141251475,0.3172460496,0.1932741553,-0.2778713703,-0.2863939404,0.0729600713,0.0909321681,0.3113380373,-0.1171345115,0.2142642438,-0.1012699232,-0.1189801767,0.0412625782,-0.0918112397,0.034752015,-0.1608827561,-0.0348336659,0.2261978388,0.0669839382,0.5594609976,0.1551665962,0.1317995638,-0.0048616766,-0.1357942522,0.2963977456,-0.1188660637,-0.4440218806,-0.5441219807,0.1625768244,0.1110063344,0.0994877592,-0.4045838416,-0.0227594413,0.2584452629,-0.0263849851,0.1454419047,0.4294940233,0.2199430615,0.0293258335,-0.0326822698,0.5706136227,-0.0609181114,-0.2834323645,-0.0984794796,-0.3663148582]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/813","title":"How to implement DistributedSampler with datasets ","comments":"Hi Apparently I need to shard the data and give one host a chunk, could you provide me please with examples on how to do it? I want to use it jointly with finetune_trainer.py in huggingface repo seq2seq examples. thanks. ","body":"Hi,\r\nI am using your datasets to define my dataloaders, and I am training finetune_trainer.py in huggingface repo on them.\r\nI need a distributedSampler to be able to train the models on TPUs being able to distribute the load across the TPU cores. Could you tell me how I can implement the distribued sampler when using datasets in which datasets are iterative? To give you more context, I have multiple of datasets and I need to write sampler for this case. thanks. ","comment_length":40,"text":"How to implement DistributedSampler with datasets \nHi,\r\nI am using your datasets to define my dataloaders, and I am training finetune_trainer.py in huggingface repo on them.\r\nI need a distributedSampler to be able to train the models on TPUs being able to distribute the load across the TPU cores. Could you tell me how I can implement the distribued sampler when using datasets in which datasets are iterative? To give you more context, I have multiple of datasets and I need to write sampler for this case. thanks. \nHi Apparently I need to shard the data and give one host a chunk, could you provide me please with examples on how to do it? I want to use it jointly with finetune_trainer.py in huggingface repo seq2seq examples. thanks. ","embeddings":[-0.1762313098,-0.2129835486,0.0418279395,0.1458279192,0.2960063815,-0.2145047784,0.2066301256,-0.0596603192,0.0736611038,0.3631319404,-0.1494355202,0.2770134211,-0.326344341,0.1506611556,-0.0292026214,-0.5932819843,-0.0871201083,0.2308092415,-0.2494449914,-0.0329345912,0.2496520281,-0.0033963125,-0.1133745164,-0.0734546185,-0.1053177342,-0.1739124805,-0.359015584,-0.0182641707,-0.1748254895,-0.2580248713,0.4371857941,0.3987306952,0.6609314084,0.4599104524,-0.0001257404,0.0744623244,0.1018331274,-0.1535160094,-0.2574951351,-0.1424911767,0.0545339696,0.2479480505,0.1367184818,-0.2322056442,-0.1759753823,0.0377360918,0.1900282204,0.026471369,0.5198253989,-0.1474679559,0.0699756369,-0.0524421185,-0.2422241271,-0.2215243727,-0.1993658096,0.3077515066,-0.0410113595,0.2077390999,0.3981792331,-0.0075883656,-0.4323901832,-0.0242201909,0.0330282822,0.2175633162,0.0072184382,-0.1358048767,-0.2036307305,-0.2351312786,-0.3969820738,0.1602642834,0.6051869392,-0.25707376,-0.3713671267,-0.1774278432,-0.2501272261,0.0500185415,-0.1925211549,0.0510461219,-0.183778286,0.2823157609,-0.153006494,-0.1160468906,-0.3117052913,0.1274695694,-0.1568293422,0.4186398983,-0.1393884569,0.1281813085,0.0936493352,0.2098372877,0.1063390896,0.0415877439,0.2194653004,0.1438554674,-0.223254025,-0.3640882969,0.0095435483,0.274744004,0.4882165492,0.5775046349,0.0499158986,-0.0942468047,-0.0080486406,-0.007061468,0.3743906021,-0.0001486485,-0.1673378646,0.4795022011,0.0228307936,0.2750387192,-0.026658345,-0.1354807466,-0.1654914021,0.0668157712,0.0090902438,-0.1441898793,0.000844351,-0.1314394176,0.0769860521,-0.4496330917,0.2652007341,-0.2379103899,-0.1790024787,0.1799650192,-0.1000174955,0.0708163828,-0.1266990751,0.1342818737,-0.0206947736,-0.1132476702,-0.1113281175,0.1297118366,-0.2131318301,0.2274442911,-0.0074115768,-0.3920825422,0.3760806024,-0.0981843174,0.3529301286,0.0478642061,0.1421647519,-0.129148975,-0.1584916115,-0.0873794779,-0.0718234181,0.1139059812,-0.1525042802,0.3168794811,-0.2817803919,0.0391251482,-0.4134108424,-0.3401998878,0.1583733112,-0.0231166948,-0.2817355692,-0.22678186,-0.2778107226,0.2035528272,0.0177877527,0.009934308,-0.0463818349,-0.057802029,-0.1742765903,-0.1528570801,0.2357306778,0.4027376175,-0.2060623914,-0.2215589136,-0.0437944233,-0.1560462117,0.3560594022,0.2536260188,-0.1780752689,0.3056147695,-0.1869759709,0.1814290136,0.4121002257,-0.4824197888,-0.4623402059,0.3475470245,-0.1774097532,0.0841415152,0.3403612971,-0.0765789449,0.4955591857,-0.199061662,-0.3674882948,0.5588446856,-0.2115409821,-0.1284406036,-0.1992321014,-0.2120351642,0.1296761632,0.3994773924,0.096390225,0.3884046674,0.0329116583,0.0927874893,0.3912806809,-0.0111880135,0.5302360654,-0.0398245864,0.1653404534,0.0933416188,0.1260007918,-0.3339460492,0.0291721094,0.0882046148,0.1346919835,0.1638704687,0.1431144774,-0.0069237747,0.0181160346,-0.2439237386,-0.4402566552,-0.1933107227,-0.0643392801,0.218007192,0.6305688024,-0.3245465159,-0.3223573864,-0.0564459972,0.0957111642,0.0224031936,-0.3555355668,0.4854133129,0.1750119925,-0.1378860623,-0.062399175,-0.1549349874,-0.2792490721,-0.3736248016,0.1042341217,0.2754532099,-0.0492897965,0.1374732107,0.2444846779,-0.1144604161,0.1196562573,-0.2800710499,0.144788146,0.1517023593,0.0262994524,-0.0369204916,0.1056541502,0.4488400519,-0.1263534427,0.1596807092,-0.0317609347,0.1220786348,0.2391726375,-0.0615853257,-0.2905206084,0.0944519117,0.0107972296,-0.2886224389,0.2336177975,-0.180255726,-0.3940914571,-0.0089541143,-0.1615924835,-0.1977666318,0.0055999644,0.0927682668,-0.0583694316,0.221334517,0.115395844,0.2206609249,0.4545042515,0.0143219512,0.1267328411,0.0229571797,0.2172505409,0.032864809,-0.0622270182,-0.0104037626,-0.0782881603,0.0800554454,-0.2320338786,0.0293563213,0.0692374855,-0.3264921904,-0.0312445015,0.1035008803,-0.0159063227,-0.0558410436,-0.2748684287,-0.1262640506,-0.4773798585,-0.0857991874,0.1158196554,0.284018755,-0.234359175,0.1120974347,0.2257270068,0.2018771023,0.0999606997,0.6861427426,-0.068113856,0.2938063145,-0.0379981808,-0.0865904689,0.0861778259,0.0325731114,-0.0054661119,0.0172466263,0.3180850446,-0.1071855873,-0.1507676244,-0.142157048,0.1396576315,0.1572736949,-0.006113145,0.1908628494,0.2596526444,0.0942714289,0.0147717735,0.301555723,-0.0144246528,-0.1850948781,-0.0725712776,0.0200595725,-0.0220017619,0.1053396761,0.0714206919,0.1350694448,-0.5741364956,-0.1103194728,0.1514456272,0.4341512024,0.1620614678,-0.308114022,0.0741549432,-0.204681918,-0.0837833509,0.0611390248,0.1934665442,-0.6808118224,0.2450155169,-0.2307176292,-0.171098128,-0.2517700195,-0.0022165319,0.0215864293,0.3401453793,-0.2149410099,-0.3724907637,0.1669088453,0.1288245916,0.1506884545,0.4997434318,0.3359586,0.0185606442,0.246110633,-0.1261478066,-0.0985610187,0.3742190003,-0.0322942585,0.1080243811,0.3546367884,0.2176999152,-0.0627698302,0.6847801208,0.3673390746,-0.0072791479,0.2519361079,0.0604620501,-0.3771803677,0.2624773383,0.0532165393,0.3872335851,-0.0692632943,-0.0042585488,0.3896640241,0.0688651279,0.2607900798,-0.3597668707,-0.1988830119,-0.010938731,0.0732276067,-0.1000319347,-0.2645080984,0.2502071261,-0.3910618126,-0.2263169438,-0.0937713608,0.1206526905,-0.0165997576,0.2696183324,0.4421617687,0.2345451713,-0.3730179369,0.2232339531,-0.2388771772,-0.1237919629,-0.0817800313,-0.0156753585,0.2256255448,0.0217451248,0.3156937361,-0.0126428949,0.3779911995,-0.2337822616,-0.0938331112,-0.4844174981,-0.4311335981,0.112226136,-0.0873304829,-0.3538045883,0.4462782741,-0.5912209153,0.1568501145,-0.3123340309,-0.1518695652,0.1150375754,-0.0043515484,-0.3764227331,-0.2770949304,-0.1194541529,-0.4327888191,-0.0640134066,-0.1120810881,-0.0450904295,-0.0622289255,-0.0481266715,0.1054449752,-0.2978600562,0.1372386366,0.1504854262,-0.0514174588,0.0223482531,-0.1955296695,0.1062995121,0.3705880642,0.3044517934,0.3851693571,0.448928386,-0.4912441969,-0.3860999644,-0.0408703573,0.1253970414,0.2308127433,0.0087957373,0.0827494636,0.3002053201,-0.0572643951,-0.0678587109,-0.5000570416,-0.2672558427,0.024462251,0.0200952925,-0.4728281796,-0.7147187591,0.2241015732,0.1522457302,-0.1368087828,-0.0321770906,-0.0694763735,-0.2612022161,0.2622130215,0.2249399722,0.8707539439,-0.1021602377,-0.0420661047,0.073840335,0.274867177,0.7067088485,0.319917202,0.1001652256,-0.2827221155,-0.3105245233,-0.1979127228,-0.1109063849,-0.0057181483,0.516413331,-0.5429206491,0.3605951667,0.0219280664,-0.0996981189,0.1444530934,0.093671456,0.1357545406,-0.1352725625,-0.3015867174,0.0253904536,-0.1955166459,0.2485632747,-0.138993293,-0.2632388473,0.1933846921,0.2051689774,0.207427606,-0.0668180212,-0.0121262716,-0.104702495,0.4981981814,-0.084952116,0.245443061,0.0036709597,0.3181904256,-0.2570709884,0.0013374763,-0.0576543212,-0.1081556082,-0.1082593128,-0.0360560156,0.0641892627,0.4372723997,0.0569119342,-0.2256896645,0.2304375619,0.1575966626,-0.283833921,-0.1787339598,-0.2039135396,0.1599760652,-0.3195153475,-0.1481682807,-0.0226087384,0.0651266575,-0.0348767713,-0.0800833777,0.344161123,-0.3170127869,0.352671653,0.0027462768,0.1089169756,-0.1081154644,0.2389671504,0.2448527217,-0.0683445334,0.0795950368,-0.122829847,0.0393767729,0.0055262269,-0.3198083639,0.1380521953,-0.5434182882,0.2638340294,-0.1851090789,0.247944966,-0.1067715809,0.2094822377,0.263399899,-0.0644766241,-0.2166621089,-0.3531405926,-0.3530889452,-0.1570403129,-0.044242952,0.1284519583,0.0706517473,0.0463198759,-0.0158347599,0.2757929862,-0.1610549241,-0.1002331972,-0.0341599397,0.1060389206,0.3885426819,0.2384246141,-0.3457787633,0.2648670375,-0.1020126343,0.1118480042,-0.0135226976,0.0236105174,-0.0651644319,0.2168444246,0.2107196152,-0.2215079516,-0.0826413408,0.0160091911,-0.1435122341,-0.0845793933,0.181099996,-0.0206929017,-0.1497682482,0.2486905158,0.7305862308,-0.4384990335,-0.4540959895,0.4001855254,0.2096713185,-0.1351144016,0.0673504174,0.3071457446,0.175489679,-0.205328837,-0.0024770042,0.2844804525,0.6360450387,0.1042025611,0.3212516308,-0.2410329133,0.3320279419,-0.0104443626,0.4907310605,0.4158389866,-0.1445629001,0.0061384914,0.3453768492,0.0719081461,-0.0637940913,0.1239284202,0.1777506769,0.1627659053,-0.1147236675,0.5678812861,-0.2439210713,-0.0187723823,-0.3228301406,0.0691759884,0.4023536742,0.1933407634,-0.0950644985,0.3029857278,0.248390913,0.0796130523,0.2700076401,0.3157448173,0.1950952113,-0.0959001034,-0.2032369822,0.7602377534,-0.220154345,-0.1591709554,-0.0294414684,-0.4686174691,0.3586615026,0.3634355366,0.2123480439,-0.0684422553,-0.1370842457,0.5740570426,-0.1270174682,-0.2321818024,-0.0025766222,0.3248003721,0.0606992356,-0.3778528273,-0.340952456,-0.1854120344,-0.1240251586,0.0476576723,-0.1081500202,0.0553416684,0.5609665513,-0.1190458387,0.0711848214,-0.5261405706,-0.345315963,0.3194201589,0.1149639264,-0.0742405429,0.258495748,-0.0626595616,0.0744223818,0.130841136,0.4096456468,0.0194164049,0.2499273568,0.009659321,0.0382866636,-0.0724936575,0.1148360968,0.0892678723,0.1026840284,-0.2961399853,-0.1933686733,0.4360482395,-0.0594253354,0.0231242236,0.1796997935,0.1417747736,-0.1073037907,-0.4729669988,0.1520291269,0.021126885,-0.0499812625,0.2113032192,0.1267897785,-0.396258384,-0.0587534271,0.0734542385,-0.5963245034,0.0379444845,-0.153463766,-0.0325333737,-0.1356429756,0.4952053726,0.4738103449,0.1462251991,-0.112451531,-0.3779447377,-0.2596486807,0.2536073029,-0.1999996006,-0.3475928009,-0.2860644758,-0.0203980617,0.0625812039,0.1093226373,-0.5165166855,0.2676765919,0.015500457,0.6406424642,-0.0078481846,-0.3341740072,-0.2320216,0.4174197912,0.1265353709,-0.1967198402,-0.0851983577,0.1007861793,-0.1088243052,0.0788634643,0.2305402905,0.1126676947,-0.3531213403,0.1094886214,0.1274303198,0.3620076478,-0.0593735725,0.1948195398,0.075658232,0.1606430709,-0.3826026917,-0.1971896589,0.0656110644,0.0794163421,-0.1215264872,0.1512836367,-0.3343378901,-0.2985701561,0.2322646528,0.1794825792,-0.1885072738,0.017602738,-0.2611690462,-0.0382782109,0.451354444,0.278659761,-0.0830281451,-0.0083432663,-0.1770680249,-0.0689453483,0.2432762384,-0.3367719054,-0.0019335501,-0.2429523617,-0.0359771512,-0.3963621259,-0.3386980593,-0.6620282531,0.1587570906,0.2287367731,-0.0071009854,0.0644005984,0.2964572608,0.2667241096,-0.0461757481,-0.0840350762,0.1205634847,0.0632555708,-0.0628655255,0.0216287766,-0.1302149296]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/813","title":"How to implement DistributedSampler with datasets ","comments":"Hey @rabeehkarimimahabadi I'm actually looking for the same feature. Did you manage to get somewhere?","body":"Hi,\r\nI am using your datasets to define my dataloaders, and I am training finetune_trainer.py in huggingface repo on them.\r\nI need a distributedSampler to be able to train the models on TPUs being able to distribute the load across the TPU cores. Could you tell me how I can implement the distribued sampler when using datasets in which datasets are iterative? To give you more context, I have multiple of datasets and I need to write sampler for this case. thanks. ","comment_length":15,"text":"How to implement DistributedSampler with datasets \nHi,\r\nI am using your datasets to define my dataloaders, and I am training finetune_trainer.py in huggingface repo on them.\r\nI need a distributedSampler to be able to train the models on TPUs being able to distribute the load across the TPU cores. Could you tell me how I can implement the distribued sampler when using datasets in which datasets are iterative? To give you more context, I have multiple of datasets and I need to write sampler for this case. thanks. \nHey @rabeehkarimimahabadi I'm actually looking for the same feature. Did you manage to get somewhere?","embeddings":[-0.2165309042,-0.2023570538,0.020388063,0.1215565279,0.3843171597,-0.148787871,0.186389029,-0.0463320054,0.0909096599,0.3809642792,-0.0927010551,0.4012136757,-0.3919049799,0.126225397,-0.0306729656,-0.5674625635,-0.0996232703,0.2849842012,-0.2662015259,-0.01740182,0.1651191562,-0.0660950318,-0.1672663391,-0.1096204817,-0.1321873069,-0.1176306158,-0.3399848342,-0.0739170313,-0.1684099138,-0.2888582945,0.5182573795,0.4490874112,0.6537900567,0.4402197599,-0.000124823,0.071545817,0.0709306747,-0.1909096837,-0.3362434208,-0.1468860507,0.1308752447,0.2465892583,0.1811210513,-0.1895624399,-0.2026922256,-0.000441181,0.1944827884,-0.0889957994,0.4009197056,-0.1674772203,0.0885548294,0.0375816561,-0.2528292537,-0.2085051984,-0.3615519106,0.2448582649,-0.097490564,0.2348291278,0.4601494968,-0.1212566718,-0.4466724396,-0.0353154093,0.0963242725,0.2270764858,0.0847311988,-0.1446755081,-0.1061014533,-0.262953341,-0.4032715857,0.1285451204,0.6994565129,-0.3501695991,-0.3488595188,-0.2288467139,-0.1941349357,0.0415006652,-0.2030966282,0.0501772575,-0.2512930036,0.2718913853,-0.1817533076,-0.043110922,-0.2872264981,0.0815197453,-0.2323960215,0.4525696933,-0.0668860003,0.0729217008,0.1494348347,0.2264926136,0.1342888325,0.0596250109,0.2317754924,0.2535398901,-0.2298528552,-0.3702285588,0.0702756941,0.1285797954,0.4616214037,0.5662394762,0.0761739239,-0.1521155685,-0.0448619016,-0.0280365758,0.4356670082,0.0770428926,-0.1707180291,0.5068981051,0.0651540682,0.2797270417,-0.0722907633,-0.0992460549,-0.1708536595,0.058975324,0.1392942071,-0.1172517687,-0.0319345258,-0.086732097,0.0175495818,-0.4746623933,0.1759720296,-0.2886090279,-0.1607275605,0.1076636165,-0.0595859326,0.2143725753,-0.0691009909,0.0708629042,-0.0429354683,-0.015893301,-0.0998504683,0.0154092899,-0.268389225,0.2149849683,0.0389417596,-0.4428074658,0.3086644709,-0.1108377576,0.2469436228,0.1018884629,0.1739122421,-0.0973974094,-0.1672606915,-0.0707488954,-0.1003448218,0.0889528543,-0.1787589341,0.2649844885,-0.277920872,0.1031192765,-0.3787964582,-0.35572052,0.1932729483,0.0127837611,-0.2983446419,-0.3314011395,-0.278249234,0.2571376264,0.0561525784,0.0630708933,0.0124973012,-0.1091790721,-0.2096542865,-0.0970473513,0.1125250012,0.3716898859,-0.2783739865,-0.3031387925,-0.0656665042,-0.2069983482,0.2628522813,0.2090155035,-0.2164518982,0.3393015862,-0.2185666859,0.0735788867,0.4029093385,-0.5532113314,-0.4664339423,0.2351519912,-0.1345618665,0.1386264414,0.2546123862,-0.0702549219,0.4162773192,-0.2432773858,-0.3529928923,0.5552344322,-0.1871447265,-0.1033111215,-0.151576817,-0.2774964571,0.2634341717,0.3917880356,0.1487579793,0.4014329612,0.0287085585,0.0867536888,0.3887357116,-0.034541823,0.4507935047,-0.0422860757,0.1860589236,0.1903915554,0.135160923,-0.3149318397,0.0139392028,0.1383338273,0.1758470237,0.2098708451,0.1540810019,-0.0241910703,0.0765709952,-0.2050461024,-0.4359622598,-0.2129650116,-0.0386844948,0.2436619401,0.6493265033,-0.2880854309,-0.3387707174,-0.0465635359,0.0428923965,-0.0099182073,-0.2881137133,0.5171958804,0.2572255731,-0.096985653,-0.1045919731,-0.0844627842,-0.2886455059,-0.3541497886,0.049752485,0.2313253731,-0.0688014776,0.1071298644,0.1882388294,-0.0307362862,0.160217002,-0.2609453201,0.0494860522,0.1976170987,0.0139576225,-0.0011788661,0.1816427708,0.4584166408,-0.091283448,0.1618430912,-0.0176227354,0.0689404085,0.2482760102,-0.049674172,-0.3117604256,0.0996694341,-0.009035727,-0.3414749801,0.283056289,-0.1612679809,-0.4496816993,0.0055579981,-0.1178175658,-0.1339013278,0.0290130209,0.0919223055,-0.0237082206,0.2973592579,0.076852642,0.2244843394,0.5295805931,0.0130644925,0.0149418954,0.0156922396,0.1311888844,0.0431485921,-0.0322031677,-0.0124325305,-0.0034537339,0.1446638703,-0.1527252197,0.0032505805,-0.0407077223,-0.2775951624,-0.0094986456,0.1618284285,0.0082186041,0.026368672,-0.3325667679,-0.2401412874,-0.4659965634,-0.096452415,0.170759365,0.3262687027,-0.1937349737,0.1140824482,0.238782987,0.2413820922,-0.0037935043,0.7553138137,-0.0628123283,0.1748380661,-0.0445116349,-0.0154870069,0.133641392,0.051479321,0.0187079087,-0.147751376,0.3465191126,-0.0748949498,-0.1254164577,-0.1330322325,0.1165987104,0.1613781154,-0.030916784,0.1130733863,0.2598000169,0.04591351,0.10404782,0.2717871964,-0.0342350155,-0.2606757581,-0.0983254611,-0.0877068788,0.0423041582,0.1600300968,0.0190616008,0.0917886943,-0.5680099726,-0.1370946616,0.1410538405,0.4544929862,0.1400138289,-0.1759437174,0.0463739373,-0.2308117747,0.0500022136,0.0713107213,0.1658339053,-0.5199517608,0.2664794624,-0.1599735916,-0.1007036492,-0.2666448057,-0.0395482555,-0.0002217202,0.3269921243,-0.2514092326,-0.407549113,0.2008335441,0.1957855076,0.175312832,0.470122546,0.3266563714,0.01749257,0.2450100631,-0.1407221258,-0.1195285022,0.2879768312,-0.0182522926,0.1556462049,0.3016829193,0.2099719644,-0.0794219524,0.5440557599,0.3152562082,-0.0208351165,0.2822468579,-0.0004051457,-0.2996735573,0.2345888615,0.0304631833,0.2967295945,-0.0885276124,-0.0089330031,0.3630433381,0.0517778918,0.1827423722,-0.3566708863,-0.0506790504,0.0959517956,0.1358996928,-0.1313022077,-0.1887299418,0.3018294275,-0.33078444,-0.1355901957,-0.1245617494,0.1063714996,-0.0169843771,0.2894280851,0.4251808822,0.2479918599,-0.303231895,0.2204300165,-0.1927861869,-0.1803206652,-0.0391703993,0.0421569012,0.2574430108,0.0536107495,0.2958943844,-0.0323585905,0.3720577657,-0.2531808019,-0.0881546661,-0.387788415,-0.4554795921,0.1427450925,-0.1549454927,-0.3248292208,0.3821784556,-0.6419743299,0.18650572,-0.2228659838,-0.1695716828,0.129749611,-0.0175287146,-0.4033493996,-0.313508749,-0.1300603002,-0.4538510442,-0.0398575738,-0.1058635414,-0.0831257999,-0.0273511745,-0.0757266879,0.1400482506,-0.3459309042,0.1115922034,0.1059642136,-0.0160707869,0.0294209514,-0.1020464525,0.1468328238,0.4120019376,0.1901023388,0.4654660821,0.4581186473,-0.4826382995,-0.4147562385,-0.0251680259,0.1807683855,0.1995957643,-0.042510666,0.0823885649,0.3270778954,-0.0985214263,-0.1893451065,-0.5728685856,-0.3072676957,0.0553162359,0.0399379134,-0.4472287595,-0.828864336,0.2628258765,0.1971182525,-0.1026557535,-0.0123224296,-0.2157283127,-0.3253523409,0.1409413815,0.2372234464,0.7245416641,-0.0711941496,-0.0696348697,0.1125445142,0.2200796902,0.6619564295,0.2771749794,0.1636957079,-0.3156518638,-0.3655219674,-0.1956333518,-0.0939021632,0.0071910387,0.4505413175,-0.4866737723,0.3560767174,0.0267366096,-0.2320420891,0.1834115386,0.0985544845,0.0968170241,-0.1445322782,-0.3794889152,0.0195361618,-0.1668569148,0.2978041768,-0.1896472573,-0.2747825086,0.2784691751,0.2467145622,0.1726403832,-0.1421983242,-0.0956913084,-0.0721727982,0.5814096928,-0.0081383353,0.2799599171,-0.0672553852,0.3201732039,-0.2238219678,-0.0023396071,-0.0889147893,-0.044199273,-0.0409556963,-0.017867066,0.0779831335,0.5492190719,0.0895267501,-0.2007618397,0.1741208434,0.144699946,-0.3615266979,-0.1827022433,-0.180143714,0.2135112584,-0.2193838954,-0.0604780912,-0.0182258412,0.0767516792,-0.0652689934,-0.0450295135,0.3054356575,-0.2713140249,0.4087233245,-0.0337427221,0.0767839178,-0.105837889,0.2480822206,0.2521826625,-0.0844725966,0.0603301078,-0.1019937769,0.0162474476,-0.0056954976,-0.3071508408,0.1561245918,-0.5409187078,0.3603943586,-0.1043569967,0.1988104433,-0.0793464705,0.3388211429,0.2067394257,-0.0519281626,-0.3318500221,-0.3180299997,-0.3616573513,-0.1885628104,0.015309222,0.1196057051,0.0807890147,0.0875288397,0.0575975664,0.1337056309,-0.1778936237,-0.1881615222,-0.0250500198,0.0391960442,0.3825622499,0.2514796257,-0.393999666,0.2616467178,-0.0969747975,0.1353472173,-0.029548021,-0.0121472878,-0.0393366255,0.1990844905,0.1935470998,-0.2127173841,-0.0331040435,-0.0286111869,-0.1400754303,-0.0957015008,0.1598298699,0.0238163453,-0.169951871,0.2665828168,0.6310212612,-0.472913444,-0.3854475617,0.380698204,0.2784909308,-0.1534259766,0.1367963701,0.2742290795,0.1634165496,-0.1100248024,-0.058394976,0.2710906565,0.4807316661,0.1365775764,0.318138212,-0.3487918973,0.342513442,0.0476930365,0.4564115107,0.4464416504,-0.1985292435,0.0193899926,0.3287819028,0.0666717663,-0.1363462508,0.0640694648,0.0896370783,0.1934843361,-0.1114683524,0.566208005,-0.1832181364,-0.1035463884,-0.3213105202,0.0728238896,0.4356056154,0.1994755566,-0.0369374603,0.4037885368,0.2490032017,0.0503894128,0.2538397908,0.3099383414,0.1343287528,-0.1804672778,-0.1194534525,0.6694839597,-0.1510232389,-0.2257778049,-0.0938858911,-0.3789962828,0.3747065961,0.2666992843,0.1315345913,-0.0050789877,-0.0650269166,0.5010744333,-0.1491955221,-0.2812371254,-0.0338017531,0.2625479996,0.0129797906,-0.3595876694,-0.2773423195,-0.1435772181,-0.113453269,0.0897729397,-0.1305410266,0.0529896282,0.4278294146,-0.1990597695,0.058999341,-0.5179473162,-0.2872245312,0.3905977607,0.1770571917,-0.0413198806,0.2715479136,-0.1167307273,0.0959770754,0.117737107,0.4054320455,-0.0025524737,0.2817520797,0.092740193,-0.0026026156,-0.1493595988,0.1466333568,0.1110376418,0.1560707241,-0.2693027556,-0.2722213566,0.4461485744,-0.0516965091,-0.0037645225,0.2431702763,0.1774901599,-0.1076511294,-0.5061547756,0.1216299683,0.0795477778,0.0240765326,0.2514518201,0.1841474324,-0.3306311667,-0.0173088014,0.1254643649,-0.561395824,0.0346092544,-0.1497036666,-0.0287389606,-0.1044995859,0.4370346963,0.4682660103,0.0955482274,0.0081476457,-0.3187167943,-0.3783536851,0.2116574198,-0.1433489025,-0.3909026682,-0.2231076956,-0.0087309368,0.0790762529,0.1404812336,-0.4083209336,0.2986154854,0.1431134343,0.5958378315,-0.0369535089,-0.3885062635,-0.2007407546,0.4217309952,0.1064384729,-0.2215176523,-0.1565141231,-0.0545761511,-0.0711888894,0.1027025804,0.2336345017,0.0654388964,-0.2267396003,0.1581861973,0.0575243086,0.3877450228,-0.0147543652,0.2386939377,0.0008708637,0.1192953661,-0.3936811388,-0.1663524508,0.0619325601,0.0669439062,-0.1383593678,0.1431300193,-0.3414560854,-0.3071427941,0.3422917724,0.1460476518,-0.1964162588,0.0435399748,-0.2608368397,-0.0817055106,0.3979532123,0.2920573652,-0.0611217767,-0.0211903788,-0.147805661,-0.0750260577,0.2409567684,-0.3575324416,-0.0821160674,-0.3256584406,-0.0498342067,-0.332767725,-0.361515075,-0.6557494402,0.1716835052,0.2423206866,-0.0019663223,0.0450806953,0.3479655981,0.1675272733,0.0232135765,-0.1013783589,0.0529873557,0.0330953114,0.0333361179,0.096164763,-0.109282352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/812","title":"Too much logging ","comments":"Hi ! Thanks for reporting :) \r\nI agree these one should be hidden when the logging level is warning, we'll fix that","body":"I'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2","comment_length":22,"text":"Too much logging \nI'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2\nHi ! Thanks for reporting :) \r\nI agree these one should be hidden when the logging level is warning, we'll fix that","embeddings":[-0.0939558521,-0.1382928491,-0.110982053,0.1179051399,0.3646460772,0.1386169642,0.3069940209,0.5121560693,0.1688756645,-0.0100423116,0.0955123603,0.0601680726,-0.4101086259,0.0572798029,-0.2360527515,0.2184815109,0.0992703438,-0.1177201867,-0.231096819,-0.1383837909,-0.0463390015,-0.0264285505,-0.0634166002,0.0502451658,-0.7488369346,-0.1222361922,0.2166901678,0.0171178672,-0.1799691617,-0.465010494,-0.0918776095,0.2049916536,0.1833454669,0.2377516925,-0.0001085636,-0.20621261,0.3049531281,0.0561281964,-0.2948624492,0.3130479455,-0.1765792817,-0.4536098838,0.264634788,-0.1287505031,-0.0320021063,-0.5764784813,0.1200485975,-0.5471832156,0.1971037537,0.2420777082,0.2531926036,0.1721661538,-0.1171072647,0.1836513132,0.1604662985,0.1954940259,-0.0464435145,-0.0426941253,0.2407290637,-0.1449797302,-0.2897666097,0.5247582197,-0.2106172442,-0.093664065,0.2563079298,0.0742956996,-0.0066664438,-0.4359248877,0.2002311647,0.2293611169,0.2505168915,-0.4348904788,-0.2396631539,-0.5395965576,0.1525105983,-0.2023011297,-0.0393041484,0.1188085526,0.0259445962,0.2967513204,-0.3168110251,-0.1253167987,0.0727278888,-0.1946348846,0.2814538479,0.0904524699,-0.0954422504,-0.0199597813,0.242654562,-0.2031737417,-0.005536681,0.0190858953,-0.1287662536,0.0733042508,-0.368399024,-0.1760264486,0.2714226842,0.4546540976,-0.0106762527,0.1569292098,-0.5877223611,0.0308099519,0.1449273527,-0.0687160939,0.2483414859,0.0682765171,0.3815250397,0.4842720628,0.0695502087,0.0023660441,0.2182523906,0.1636326611,0.3279185295,-0.3859716058,0.5891591907,0.0934159234,0.4714000821,-0.243321389,-0.0503412001,-0.0017934728,-0.0907608345,-0.1449495852,0.0998376757,0.2186496705,0.257361263,-0.112491928,-0.0209494438,0.1804551929,-0.1787154227,-0.1113092601,-0.3608377874,0.2265654355,-0.0386886448,0.1814368069,0.1555698663,-0.0132067855,-0.0337050185,0.1645651013,-0.0999963954,-0.0064752758,0.2545646429,-0.0888237432,0.1127234027,0.4412614405,-0.2111655325,0.1293465942,0.0744356066,-0.1091437489,-0.2172542065,-0.0095765982,0.0266517997,-0.344096452,-0.117465958,0.2434980124,-0.3380615711,0.0631735772,0.0415719822,0.0605500452,0.2659909129,0.2676562965,0.2104022354,-0.1764074117,-0.1235794947,-0.160682857,0.046920523,0.7568493485,-0.1010416821,-0.1524249017,-0.1544768959,-0.34261325,0.3341272175,0.2843775749,0.0356809534,0.2831241786,-0.31321612,0.1170730367,-0.072375685,-0.3900976181,-0.0120040989,0.3230906129,-0.2881067395,0.1007575542,0.5861548781,0.1028016508,0.0012948104,-0.2383528054,0.009368659,-0.1693769395,0.0935047865,-0.057938844,-0.2075730115,-0.1830702871,-0.0157597847,0.1808310598,-0.1068214923,-0.2202105224,-0.0464311466,0.0053200815,0.3438035548,-0.1056678444,0.0014571475,0.2216452956,0.2044258267,-0.1354949027,-0.1259181052,-0.1356040984,-0.3863966763,-0.0477210879,-0.0831304491,0.0528704599,-0.4278614521,-0.5052879453,0.0825892389,0.0003729831,-0.1854661852,-0.2281099409,0.1308967173,0.0035779478,-0.2127634585,-0.1032315567,-0.2443549931,0.3279069662,-0.0935143903,0.2245454639,-0.3384838402,-0.3394961655,0.0967411026,-0.123986423,0.3080798388,0.0880163983,0.3376420736,0.0235521495,0.2304944992,0.279966712,-0.0227968302,0.6912989616,0.0908789188,0.3066889644,0.051868137,0.1061581969,0.1213443875,-0.091763407,-0.2069844604,-0.0415721275,0.0172877163,0.3455808163,0.090901643,0.0299074426,-0.2113509476,-0.0423978716,-0.1206108779,-0.072812587,-0.124246195,-0.5477374196,0.3308843672,0.0519312583,0.0606073514,0.2649485469,0.0394691005,0.336020112,0.4038300812,-0.2107298374,0.1244593039,0.08450073,0.0159387402,0.0463408493,0.1757119298,0.1541772634,0.1422488838,0.3217694163,0.0495252907,0.2399304509,0.2244525552,-0.2068880647,0.2346292585,0.0124481553,-0.1978023648,-0.0443737507,-0.2063257992,-0.2044216543,-0.1408551931,-0.1601755619,-0.3429880142,0.2901280224,-0.3985350728,0.0035822436,-0.3858218491,-0.2859234214,-0.1935075521,0.0809366927,0.0804829001,-0.2955826819,0.2598420084,-0.0968647227,-0.3056240678,0.2036596239,-0.1606090814,0.060657829,0.1127745435,0.1230950132,-0.183965072,0.1099292412,-0.0364184715,0.0853303894,0.0721569955,-0.0706835389,0.4171567559,-0.1239209101,-0.0536239594,-0.197516039,-0.2242767662,-0.0322817005,-0.2084121108,0.0298854411,0.2623637021,0.0961000696,0.1462890059,-0.0539728366,-0.3215299845,0.029573366,0.1425770223,-0.0307257473,-0.1538363844,0.0601688214,-0.1685929298,-0.3298369646,-0.2052047402,-0.2479126751,0.1057197899,-0.0258241557,0.0914529413,0.3211137354,-0.1065982133,0.1568311304,-0.3661248386,0.3012650013,-0.028612826,-0.2743656337,0.2313615531,0.0794463158,-0.0137224877,0.0065798308,0.000336167,-0.3040707111,-0.2560657263,-0.5025682449,-0.102837719,-0.4806154668,-0.081789583,-0.1125073209,0.1075822636,0.2545316815,0.2832551301,-0.2098629475,0.0634870753,0.1032883972,-0.0417309552,-0.4492866099,0.1921845526,-0.2249032706,0.2574488521,0.1716227233,0.2437792718,0.6294993162,0.0763812959,-0.0625268146,-0.066285871,0.5207753778,-0.0275091138,-0.275418371,-0.0688880458,0.2339812964,-0.2732108235,0.0331291333,0.028368108,0.3133287728,0.0314498134,-0.0786375254,0.119645901,-0.3498458266,-0.0324782394,0.1352698952,-0.1261870861,-0.1991607696,0.2758159637,0.0850864649,-0.1263507456,-0.0174963027,0.3585094213,0.184747681,-0.0199920908,-0.2938165367,0.1961055994,-0.6184142828,0.2171833366,-0.1861789227,-0.1756950766,-0.2718885243,-0.2114417553,0.0554847382,0.3562108874,0.2229135633,-0.1450013667,0.1960818022,0.0035949715,0.1036259159,-0.2364573032,0.0014578033,-0.3731634021,0.2473729253,0.5612368584,0.2165387571,-0.1743339747,-0.2276818007,0.2084886581,0.2237481326,0.0535887554,-0.306856215,-0.2949361205,-0.1452511698,-0.3326519132,0.0871562883,0.1866647452,-0.0343150832,-0.2662407458,-0.0546183139,0.2502052188,-0.2360361069,0.4416201115,-0.2512198687,0.1626155227,0.0465070531,0.2130648196,0.05702959,0.0740484819,0.1904066503,0.4802784622,0.0943292975,-0.2119494528,-0.1666476279,-0.0299108699,0.2081236541,0.4218257368,-0.283120811,0.2068608105,0.2284541428,0.4665275812,-0.2072075605,0.0717768744,0.2487655133,0.0010954966,-0.262617588,-0.023935942,0.2868354321,0.3003081977,-0.1352033615,0.3484828472,0.2236117125,-0.3694554567,0.0369406492,0.0067659374,0.8249832988,-0.1158920079,0.1240962669,0.0340900719,-0.236386016,0.4835709035,0.1011432782,-0.0739333853,-0.1816164106,-0.2403105348,0.1836619228,-0.0452959277,0.2229619026,-0.2261366844,-0.1083042547,0.2367768288,-0.5038361549,-0.0548334122,-0.3217775226,0.5082201362,-0.4011695385,-0.1209337413,-0.1482985765,0.1056726798,-0.0943693444,0.2023945302,-0.0155990142,-0.1245626509,0.3061549962,-0.1772098839,-0.1929209977,-0.1682961285,0.192841962,-0.0087257214,-0.2535127103,-0.0676524788,-0.021644162,0.1862488389,0.2338095158,0.3983854055,-0.074814409,-0.1149246693,0.0396525636,0.157065779,0.0170801766,0.0275553279,0.3977653384,0.0857748464,-0.2672382593,0.0658211336,0.1803515553,-0.1707699597,-0.2267663181,-0.1777756065,-0.0832262114,-0.266202271,0.1297423542,0.0929617435,-0.2065989375,0.0840089992,0.1887381226,0.4445774853,-0.0216287971,0.1824272722,-0.0548384897,-0.4765105546,-0.3952400386,0.5040408373,0.1510437578,-0.1266809553,0.5556841493,0.2504146397,-0.3811905086,-0.3552457392,0.1111969948,0.3435372114,-0.0773275718,0.0663196817,0.2360484153,0.0889012367,0.0891178995,-0.171262607,0.0418535359,-0.0376884043,-0.0416651927,-0.1587856561,-0.6313868761,0.2076010257,-0.2095308006,0.2792848647,-0.1171589643,-0.0032092934,-0.0582863651,0.0690880492,-0.3741519451,-0.277766645,0.0515773483,-0.1663176417,0.2583635449,0.0489937626,0.3095758855,-0.0568151325,0.1622138619,-0.1277163029,-0.294611454,-0.2139981091,0.0317863561,0.0622518174,-0.3852056563,0.0340220965,-0.0096546086,-0.1907473207,0.0100172106,-0.0825330317,0.2864011824,0.5032940507,0.2410665751,0.3299827874,-0.065027386,-0.1737877131,0.0901242793,-0.3782750964,-0.0790195838,-0.018916009,-0.1019782573,-0.0092230299,-0.08756046,0.0033159219,0.0400350541,0.0569200255,0.0982404649,0.2117705643,0.0139172217,0.163358584,-0.1465501785,0.436992377,0.3946341276,0.0505321063,-0.2457547635,-0.2233489752,-0.158120513,0.2508950233,-0.1108826697,-0.1131797805,-0.1043809652,0.3248178363,-0.0688688233,0.3823311031,-0.1952057332,-0.0537763648,0.4082465172,0.1821787804,0.042700395,-0.3116479814,0.1041914299,-0.1105656922,-0.248974964,-0.2402355373,0.3837442994,0.2099575549,0.187479049,0.3255286515,-0.2500393987,0.4886870086,-0.0836315379,0.0282432213,-0.1258113831,-0.1916559488,0.1283935905,0.4202002585,-0.3408842385,-0.2257042676,0.1094433665,0.100306429,0.0910665393,0.1164527014,-0.0592876971,0.0206260383,-0.1105685383,0.0652008951,0.527222991,0.146957919,-0.0223207399,-0.14546974,0.0002111085,-0.1507511735,0.0020221397,0.0691605583,-0.0758545101,-0.0888675153,0.3956035376,-0.157543689,0.5541149378,-0.2218865305,-0.1620530635,0.0254402719,0.1349115819,0.2545056641,0.2312451452,0.6053708792,0.3216749728,0.2831219733,0.2745065689,-0.0613174103,0.0074724234,-0.0845505595,0.4589730501,-0.1363325715,0.1556657702,0.1265349835,0.1685230732,-0.1802562028,0.1641699225,0.0789167508,-0.0232052114,-0.197009638,0.5414173603,0.1035415232,-0.2683021724,-0.2237721831,0.0505935028,-0.266741842,-0.1599801779,0.3896869719,0.1084455103,0.0784001797,-0.4151429236,0.0978069827,-0.1514095068,0.0709749088,0.2777141035,-0.0226129889,-0.3228414357,0.2279046327,-0.3644594252,0.1353074014,-0.0462399758,-0.2501854897,-0.1109027192,-0.0260771308,-0.1400716156,0.2400045246,0.3284691274,-0.2272968739,-0.0551453233,-0.4129098058,-0.1118052825,-0.0292401686,0.0132133365,-0.0522257984,0.0636382252,-0.0837535858,0.4917401671,-0.5058588386,0.1113729179,-0.1033639163,0.0078775343,0.1151105464,0.2492211163,0.2712682188,0.1212184429,0.2057715803,-0.1138903275,-0.0524764881,-0.3292315304,0.0383195877,-0.2527835965,0.1710240692,0.1203139648,0.1378357857,-0.3307656348,-0.3593438864,-0.2284506261,0.4239183664,-0.0458421037,-0.2112400681,-0.0408494025,0.094330065,0.2376451343,0.0177990589,0.0140665546,0.1596431732,0.0941284671,0.2866966724,-0.0830889866,-0.6573038697,0.5842885971,-0.3260941803,-0.0806082785,0.0214554183,0.6829879284,0.2588315308,-0.2425112873,-0.6747079492,0.1540486962,0.1601794213,-0.1886596978,0.0084413476,0.4398554862,-0.2275979221,0.017099645,0.1593380719,0.2672632933,0.1233646199,0.145019576,0.0384272747,-0.3869270682]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/812","title":"Too much logging ","comments":"+1, the amount of logging is excessive.\r\n\r\nMost of it indeed comes from `filelock.py`, though there are occasionally messages from other sources too. Below is an example (all of these messages were logged after I already called `datasets.logging.set_verbosity_error()`)\r\n\r\n```\r\nI1109 21:26:01.742688 139785006901056 filelock.py:318] Lock 139778216292192 released on \/home\/kitaev\/.cache\/huggingface\/datasets\/9ed4f2e133395826175a892c70611f68522c7bc61a35476e8b51a31afb76e4bf.e6f3e3f3e3875a07469d1cfd32e16e1d06b149616b11eef2d081c43d515b492d.py.lock\r\nI1109 21:26:01.747898 139785006901056 filelock.py:274] Lock 139778216290176 acquired on \/home\/kitaev\/.cache\/huggingface\/datasets\/_home_kitaev_.cache_huggingface_datasets_glue_mnli_1.0.0_7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4.lock\r\nI1109 21:26:01.748258 139785006901056 filelock.py:318] Lock 139778216290176 released on \/home\/kitaev\/.cache\/huggingface\/datasets\/_home_kitaev_.cache_huggingface_datasets_glue_mnli_1.0.0_7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4.lock\r\nI1109 21:26:01.748412 139785006901056 filelock.py:274] Lock 139778215853024 acquired on \/home\/kitaev\/.cache\/huggingface\/datasets\/_home_kitaev_.cache_huggingface_datasets_glue_mnli_1.0.0_7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4.lock\r\nI1109 21:26:01.748497 139785006901056 filelock.py:318] Lock 139778215853024 released on \/home\/kitaev\/.cache\/huggingface\/datasets\/_home_kitaev_.cache_huggingface_datasets_glue_mnli_1.0.0_7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4.lock\r\nI1109 21:07:17.029001 140301730502464 filelock.py:274] Lock 140289479304360 acquired on \/home\/kitaev\/.cache\/huggingface\/datasets\/b16d3a04bf2cad1346896852bf120ba846ea1bebb1cd60255bb3a1a2bbcc3a67.ec871b06a00118091ec63eff0a641fddcb8d3c7cd52e855bbb2be28944df4b82.py.lock\r\nI1109 21:07:17.029341 140301730502464 filelock.py:318] Lock 140289479304360 released on \/home\/kitaev\/.cache\/huggingface\/datasets\/b16d3a04bf2cad1346896852bf120ba846ea1bebb1cd60255bb3a1a2bbcc3a67.ec871b06a00118091ec63eff0a641fddcb8d3c7cd52e855bbb2be28944df4b82.py.lock\r\nI1109 21:07:17.058964 140301730502464 filelock.py:274] Lock 140251889388120 acquired on \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow.lock\r\nI1109 21:07:17.060933 140301730502464 filelock.py:318] Lock 140251889388120 released on \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow.lock\r\nI1109 21:07:17.061067 140301730502464 filelock.py:274] Lock 140296072521488 acquired on \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow.lock\r\nI1109 21:07:17.069736 140301730502464 metric.py:400] Removing \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow\r\nI1109 21:07:17.069949 140301730502464 filelock.py:318] Lock 140296072521488 released on \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow.lock\r\n```","body":"I'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2","comment_length":145,"text":"Too much logging \nI'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2\n+1, the amount of logging is excessive.\r\n\r\nMost of it indeed comes from `filelock.py`, though there are occasionally messages from other sources too. Below is an example (all of these messages were logged after I already called `datasets.logging.set_verbosity_error()`)\r\n\r\n```\r\nI1109 21:26:01.742688 139785006901056 filelock.py:318] Lock 139778216292192 released on \/home\/kitaev\/.cache\/huggingface\/datasets\/9ed4f2e133395826175a892c70611f68522c7bc61a35476e8b51a31afb76e4bf.e6f3e3f3e3875a07469d1cfd32e16e1d06b149616b11eef2d081c43d515b492d.py.lock\r\nI1109 21:26:01.747898 139785006901056 filelock.py:274] Lock 139778216290176 acquired on \/home\/kitaev\/.cache\/huggingface\/datasets\/_home_kitaev_.cache_huggingface_datasets_glue_mnli_1.0.0_7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4.lock\r\nI1109 21:26:01.748258 139785006901056 filelock.py:318] Lock 139778216290176 released on \/home\/kitaev\/.cache\/huggingface\/datasets\/_home_kitaev_.cache_huggingface_datasets_glue_mnli_1.0.0_7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4.lock\r\nI1109 21:26:01.748412 139785006901056 filelock.py:274] Lock 139778215853024 acquired on \/home\/kitaev\/.cache\/huggingface\/datasets\/_home_kitaev_.cache_huggingface_datasets_glue_mnli_1.0.0_7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4.lock\r\nI1109 21:26:01.748497 139785006901056 filelock.py:318] Lock 139778215853024 released on \/home\/kitaev\/.cache\/huggingface\/datasets\/_home_kitaev_.cache_huggingface_datasets_glue_mnli_1.0.0_7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4.lock\r\nI1109 21:07:17.029001 140301730502464 filelock.py:274] Lock 140289479304360 acquired on \/home\/kitaev\/.cache\/huggingface\/datasets\/b16d3a04bf2cad1346896852bf120ba846ea1bebb1cd60255bb3a1a2bbcc3a67.ec871b06a00118091ec63eff0a641fddcb8d3c7cd52e855bbb2be28944df4b82.py.lock\r\nI1109 21:07:17.029341 140301730502464 filelock.py:318] Lock 140289479304360 released on \/home\/kitaev\/.cache\/huggingface\/datasets\/b16d3a04bf2cad1346896852bf120ba846ea1bebb1cd60255bb3a1a2bbcc3a67.ec871b06a00118091ec63eff0a641fddcb8d3c7cd52e855bbb2be28944df4b82.py.lock\r\nI1109 21:07:17.058964 140301730502464 filelock.py:274] Lock 140251889388120 acquired on \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow.lock\r\nI1109 21:07:17.060933 140301730502464 filelock.py:318] Lock 140251889388120 released on \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow.lock\r\nI1109 21:07:17.061067 140301730502464 filelock.py:274] Lock 140296072521488 acquired on \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow.lock\r\nI1109 21:07:17.069736 140301730502464 metric.py:400] Removing \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow\r\nI1109 21:07:17.069949 140301730502464 filelock.py:318] Lock 140296072521488 released on \/home\/kitaev\/.cache\/huggingface\/metrics\/glue\/mnli\/default_experiment-1-0.arrow.lock\r\n```","embeddings":[-0.0745943114,-0.0315374509,-0.051372882,0.1314180642,0.2844710648,0.2179289162,0.261808306,0.4838854373,0.0517462902,-0.1740336418,0.10809917,0.013671496,-0.4307381511,-0.0791786313,-0.1534981877,0.1896948665,0.0866655931,-0.2138257474,-0.2665444911,-0.0811803937,0.0343078077,0.0029079311,0.0677583516,0.0736543462,-0.8763465285,-0.1577260941,0.1634814441,0.0411982201,-0.1345582753,-0.3958104551,-0.1170414612,0.2172026038,0.1443828791,0.2253048569,-0.0001129851,-0.2467205822,0.2506766617,0.0904491022,-0.3506259024,0.3797374666,-0.0783602819,-0.4808190465,0.2577619851,-0.0923956782,0.11044994,-0.5984712243,0.1523717493,-0.5589926839,0.2384316325,0.2718780041,0.183426097,0.1516076922,0.0547525808,0.0827177241,0.1596712321,0.3057552278,0.0517393015,0.0208021719,0.3077190518,-0.122384347,-0.2875323892,0.4429542422,-0.1540036052,-0.1464923471,0.3328672349,-0.0242481306,-0.0548646934,-0.3220509887,0.2032457441,0.303786397,0.253004998,-0.4261766374,-0.3371694684,-0.6361056566,0.1488002837,-0.1059594676,-0.0772842988,0.1271360219,-0.0251316931,0.3121364117,-0.1953437924,-0.1719778031,0.1232566983,-0.2216964662,0.2597734928,-0.0053385668,-0.0639947951,-0.0270456579,0.3266189694,-0.1445752531,-0.0301965568,0.0324756689,-0.1313667446,-0.04204401,-0.4276967943,-0.0721251965,0.3259894252,0.3903737664,0.0073607657,0.1643509418,-0.5164461136,0.0500925705,0.1268525273,-0.1393424422,0.3606739342,0.0856678858,0.2658696175,0.5355189443,0.011528274,-0.077660352,0.1887794137,0.1849962026,0.3528625071,-0.4127818942,0.6333889365,-0.0439091772,0.3881655931,-0.1832707226,-0.056088537,0.0786956623,-0.1461774707,-0.0709004775,0.1323032528,0.1157304496,0.189390257,-0.1116280556,-0.0326280668,0.2161399722,-0.1818705797,-0.0952515155,-0.2570808232,0.2912867963,-0.0619749203,0.237243861,0.1315514147,-0.0919226632,-0.1203432009,0.1508066207,-0.1984521151,-0.1433123052,0.1449324787,-0.0918433368,0.0544101372,0.2532407343,-0.1634290814,0.1583191007,0.1093846411,-0.0930436179,-0.2132111937,-0.1570729762,-0.0029680019,-0.4197719395,-0.1508747041,0.1769854575,-0.3609291315,0.1228154525,0.0573315248,0.0382686593,0.2015438676,0.2192128897,0.1270129085,-0.1874304712,0.047402408,-0.0919263363,0.068847388,0.8149482608,-0.1338517964,-0.114942193,-0.1992361248,-0.2495595217,0.211494863,0.3670786619,0.0883117616,0.453595072,-0.2758108377,0.1275876164,-0.1310860217,-0.3799781203,-0.0529377274,0.260345459,-0.3433253169,0.1315962523,0.5908987522,0.1757712066,0.0965019837,-0.1554005593,0.195778802,-0.1910143644,0.125952825,-0.1410339177,-0.132521987,-0.1077744141,-0.1486362815,0.2916974127,-0.1531303078,-0.358929038,0.0563393645,0.1525655985,0.2883655429,-0.1712488532,-0.0581985675,0.3116374016,0.1786777973,0.0267673787,-0.1798519492,-0.1012191325,-0.3404593468,-0.0530149043,-0.122672379,-0.0383121595,-0.4733019173,-0.3894544542,0.0773143247,0.08421202,-0.1481729448,-0.1231567934,0.0824628696,0.1649744511,-0.2223937958,-0.1105730385,-0.173396185,0.537933588,-0.2286855727,0.2063637227,-0.3914621472,-0.3222681284,0.1621666551,-0.2050782144,0.3768683672,0.2321710289,0.1237566248,-0.0448712409,0.3035086989,0.2935694754,0.1105188653,0.7735791802,0.1506030709,0.2828683257,0.0450517088,0.0371117815,0.0559085794,-0.1027713045,-0.2085751444,-0.0622852445,-0.1143948883,0.2769285142,0.0641203597,0.0131075913,-0.1927532107,0.017503988,-0.0604067445,-0.0479669385,-0.1571537703,-0.4694757462,0.3884260654,0.0354194604,0.1513438672,0.3073644042,0.0933588371,0.2709863782,0.3807005584,-0.3204109967,0.2158672959,0.1895162165,-0.0337895788,-0.0532429703,0.0926937237,0.1728969663,0.1485681534,0.2563018203,0.0606386848,0.2134078592,0.2705394328,-0.1500982791,0.2070964128,-0.0006093183,-0.2361578047,-0.0310836062,-0.1758471876,-0.1036933511,-0.0950704217,-0.1284287721,-0.3363715708,0.3529498875,-0.4624248147,0.0696306601,-0.4349549413,-0.1110905111,-0.2090227604,0.2291614264,0.0648022369,-0.4513124526,0.1271877289,-0.0272102673,-0.2379029542,0.0569796674,-0.0679085702,0.0843495801,0.2101684809,0.0175130218,-0.0816577002,-0.088481389,-0.1031881422,0.0086281681,0.1440275013,-0.0204745978,0.37675789,-0.179528594,0.0304742623,-0.2623321116,-0.2368261367,-0.0319889188,-0.1196215153,0.1111499295,0.3146394789,0.1130473241,0.0980317444,-0.1044176146,-0.3348392546,0.0784268156,0.1560448259,0.0541005991,-0.2534635365,-0.0549040586,0.0237653293,-0.2233605385,-0.2945464253,-0.2709878385,0.2620931566,-0.0507086478,0.0254697483,0.3799551427,-0.0819750875,0.2585391104,-0.485540539,0.4720431864,0.0324536301,-0.1948691607,0.2467245758,0.157118544,-0.0181050766,-0.0250264369,0.0343120098,-0.360928297,-0.1947155744,-0.5805231333,-0.0669255108,-0.4184484482,-0.0779099762,-0.054952912,0.1938615143,0.1924898922,0.2107670456,-0.1480478197,0.1688166708,0.0595319569,0.0140637951,-0.3548664749,0.1852608472,-0.2460646629,0.1779491901,0.2643882036,0.1877130568,0.7277320027,0.0353037119,-0.0418432355,-0.1223061457,0.6265460849,-0.1064636558,-0.3333909214,-0.0343919098,0.1379605383,-0.2627511621,0.0293822531,0.0989605859,0.4334070981,0.0009445297,-0.1464277506,0.0377942435,-0.5074543357,0.0541752949,0.1758879572,0.0315823369,-0.206546396,0.3023816645,0.0383984819,-0.1740318388,0.0583890751,0.3313588798,0.1977048665,0.0018873953,-0.2859816253,0.1939880848,-0.4752548039,0.3170709312,-0.1866372824,-0.0570918061,-0.3711293638,-0.1716537476,-0.0279303305,0.2896545827,0.1175735518,-0.013753254,0.1020945013,-0.1267561764,0.1449100375,-0.37119627,0.0183532424,-0.3257525563,0.1876354516,0.5697363019,0.2430301011,-0.2658623457,-0.0286307018,0.0451286472,0.2988791168,-0.0259166732,-0.254248172,-0.0148700271,-0.218520835,-0.2875644863,0.0141411005,0.050219804,0.032784164,0.1181927398,-0.0466401763,0.2007246763,-0.2478118688,0.4309013784,-0.2870562375,0.2200922519,-0.0109676225,0.1372255236,0.0054598651,0.0665270314,0.2136292309,0.3995425701,0.0992327183,-0.1674853861,-0.1256913543,-0.1075529307,0.1943861246,0.4759275913,-0.2531186342,0.1103116274,0.2523988485,0.3625326157,-0.2590714693,0.1302040666,0.3083066642,0.0562577881,-0.2572055459,-0.0783733577,0.1777040511,0.3756892979,-0.0566731468,0.3477583528,0.2891749144,-0.3465955257,0.0400702134,0.0503163598,0.8999294639,-0.0925032198,0.1147378609,0.1190014705,-0.2670382559,0.5149236917,0.0153374309,-0.1342628002,-0.0883746445,-0.2309800386,0.2134760618,-0.1205796972,0.227093786,-0.2072024345,-0.0472171642,0.1816925108,-0.5165395141,-0.1573135853,-0.3318089545,0.5256655216,-0.4885504842,-0.2017862648,-0.2265624553,0.0999484956,-0.129472062,0.0825484842,0.0132374046,-0.1196901426,0.3383235037,-0.1066318378,-0.3534635305,-0.243192628,0.2296741903,-0.0642429814,-0.1778324544,0.0051144841,-0.0952356756,0.0843713954,0.1877574325,0.4290836155,-0.0039185118,-0.0873904452,-0.0480194986,0.1664818823,0.1477641165,-0.0788520277,0.4667822123,0.1206838191,-0.142326057,0.0436774343,0.1445870996,-0.2751358151,-0.1225852668,-0.2105627507,-0.1163175404,-0.1738652289,0.0548232719,0.1270008385,-0.3204524517,0.1629696339,0.1537241042,0.5195235014,-0.0784757212,0.2295350432,-0.3311522901,-0.4457042515,-0.3362948596,0.4338272512,-0.0239050519,-0.179146558,0.544835031,0.3414300978,-0.3074029088,-0.2703459263,0.2222338617,0.4122694135,-0.1188803762,-0.0120397257,0.120239079,0.1167800128,0.139866665,-0.1841656268,0.2481611818,-0.0416449495,-0.0311380215,-0.2110672742,-0.5866543055,0.223788783,-0.079590477,0.2755915523,-0.0428285748,-0.0358392633,-0.1454812884,0.1687009037,-0.3160315752,-0.3670797646,0.0413454659,-0.1166434363,0.3834860921,-0.1087869778,0.2567990422,-0.1550718695,0.1498803943,-0.0877146646,-0.2334448397,-0.1553080827,-0.0479397215,0.0670776293,-0.2971937954,0.0807246715,0.0234919451,-0.3290579617,0.0254591405,-0.2064293772,0.1208832189,0.5310435891,0.1693323106,0.1067400426,-0.1382099837,-0.1064066887,0.0611085333,-0.4175789654,-0.0476971865,0.0317199305,-0.0102500292,-0.0424824916,-0.0830200016,-0.0113731148,-0.1795108169,0.2315699458,0.2190890312,0.0727261975,0.0272215568,0.1846976876,-0.1117703915,0.4301354587,0.4489691854,0.1083129272,-0.1653653085,-0.1348163784,-0.1361067742,0.1662262529,-0.0152433291,-0.1649344265,-0.1768233478,0.2942756414,-0.1706872582,0.3220429718,-0.1476409733,-0.115913108,0.2302764654,0.1660150737,0.0898009166,-0.2617785633,0.1857284307,-0.0085953521,-0.2405253947,-0.2137368321,0.3052318692,0.2049966156,0.1911841184,0.3457424641,-0.3534468412,0.4051069021,-0.1604712754,0.0363295302,0.0316598415,-0.1979002804,0.2081106901,0.4416001737,-0.1858691275,-0.0913227648,-0.0101755178,0.1298867911,-0.0519712903,0.1838617921,-0.1689810455,0.1603252292,-0.0887483209,0.1397987306,0.5731919408,0.1216398552,-0.0264405031,-0.0852919295,0.0082869194,-0.2131767273,-0.1158265844,0.144696489,-0.1249281317,-0.1700254232,0.2997601926,-0.2238064408,0.5274563432,-0.2927848101,-0.1664265543,0.0705403462,0.0523051582,0.2441108972,0.201549679,0.6094136834,0.2762210369,0.283693701,0.3810275793,-0.0556572564,0.0009149224,-0.0881970078,0.3948673606,-0.204892695,0.0218597632,0.1150421873,0.1665735245,-0.1844943762,0.1279639453,-0.0217363387,0.0872501358,-0.2614971697,0.559635818,0.0769857988,-0.1684877276,-0.2072385699,0.1250965148,-0.2793933749,-0.2924645841,0.4331855476,0.1771564484,0.0425744727,-0.3949587941,0.0702256262,-0.1825059354,0.1641780734,0.3775790036,-0.1003876477,-0.3024157882,0.2173939198,-0.3926461637,0.0719692111,-0.0692621619,-0.3006956875,-0.0450732969,-0.0540904105,-0.0868351758,0.278655231,0.2691422999,-0.2731445134,-0.1238444597,-0.3865241408,-0.0790502578,-0.0171064734,0.1688852757,-0.0770209283,0.001185986,-0.0158319324,0.5374282598,-0.5230344534,0.1071825549,-0.0257074535,-0.0605855957,0.0098460419,0.2962516546,0.2023197711,0.0959273502,0.1849054992,-0.0711838901,-0.128766045,-0.4769387841,-0.006059079,-0.1987172514,0.2148760408,0.0223255567,0.155968532,-0.2797217369,-0.4530621171,-0.2887827456,0.4550157785,-0.0072684544,-0.3219816685,-0.074883759,0.0224791076,0.1311861426,0.0952370539,0.2487843335,0.2188135684,0.1311004609,0.3162855506,-0.1064393669,-0.5914745331,0.4044466913,-0.4396308064,-0.044922851,0.0510893986,0.639175415,0.2168328464,-0.2373143584,-0.4576916695,0.1311213821,0.1609735936,-0.2020823956,0.0410868376,0.3844127357,-0.2066688091,-0.0029250386,0.1318027824,0.1915700138,0.1770692617,0.0624664649,0.0541637205,-0.3314648271]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/812","title":"Too much logging ","comments":"So how to solve this problem?","body":"I'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2","comment_length":6,"text":"Too much logging \nI'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2\nSo how to solve this problem?","embeddings":[-0.1878589541,-0.1036341041,-0.1044046208,0.1018371284,0.3685900569,0.2144313306,0.2110944986,0.5391820669,0.0571364909,-0.0906509757,0.0979582444,0.0468162186,-0.4088027477,0.0735092536,-0.2086857557,0.127010718,0.1074681059,-0.0577916466,-0.2702092528,-0.0782236159,-0.0063196793,-0.0898019075,-0.0523255393,-0.0303916633,-0.8235750794,-0.1335223317,0.2154569924,0.0166498013,-0.2009541839,-0.3829723299,-0.0932982117,0.2164336145,0.2495422065,0.2292892635,-0.0001107674,-0.2579602301,0.3243685067,0.0677138716,-0.2945058346,0.3333757222,-0.0969246328,-0.4279743433,0.2649733126,-0.0942369625,-0.0160213355,-0.5159631371,0.1347922385,-0.4803008735,0.3317019343,0.214900136,0.2423609793,0.020179959,0.0035318283,0.0719569921,0.055719357,0.1587313861,0.0614980869,-0.0574216396,0.299995929,-0.1029080376,-0.1897394359,0.4440111518,-0.1416754276,-0.0412217826,0.1874067336,-0.0278309975,-0.1045982167,-0.3955661058,0.3357219994,0.2324328721,0.3424425125,-0.4386640191,-0.2906722426,-0.48333776,0.1688079834,-0.1978705674,-0.1634287834,0.1490425467,0.0089976434,0.3369559348,-0.3041281104,-0.2491946369,0.0480632447,-0.1049224436,0.2701355219,0.0637856647,-0.0797284469,-0.0068425024,0.2586849332,-0.1693670601,-0.0256734323,0.158174783,-0.1686599255,0.013861171,-0.4815194607,-0.0968032405,0.259275794,0.5239707828,0.0372759588,0.2317657471,-0.627810359,0.0470555797,0.2001330405,-0.1258195639,0.2598703504,0.0815095976,0.3885673285,0.5638594627,0.1349381953,-0.0608493648,0.053247191,0.1304953992,0.2879612148,-0.3486465216,0.6954981685,-0.010575572,0.4556394219,-0.2884269953,0.0687360242,-0.0292726401,-0.0218149964,-0.081405215,0.0797144175,0.3375496566,0.1197745129,-0.1394094229,-0.0512704067,0.1660452038,-0.1835499853,-0.1496657133,-0.3178687692,0.2646766305,-0.0341241769,0.1845475137,0.1482800543,-0.1268782318,0.0017571612,0.1712577343,-0.1901259571,-0.1360518634,0.2511174977,-0.1268980056,-0.0182981361,0.353805095,-0.0760270134,0.1909252107,0.0217806734,-0.0323584005,-0.174420476,-0.1074781045,-0.0218976587,-0.399854362,-0.1624929309,0.2292402387,-0.2793109417,0.1188075766,0.0882805735,0.0197223034,0.2232675254,0.2304334939,0.0715512037,-0.1633440256,-0.0718359575,-0.1827680618,0.1552116722,0.7043803334,-0.1114440262,-0.0555285551,-0.2297691554,-0.2338757664,0.2781081796,0.3710434735,0.085241653,0.3880321681,-0.2677523494,0.0232292693,0.0161801297,-0.4117331505,-0.1612965614,0.2652516663,-0.2392580956,-0.0240994859,0.6234040856,0.1940577328,0.1798762828,-0.1666600108,0.0496257022,-0.1085335016,0.0122059025,-0.1181223318,-0.095288232,-0.0576929934,-0.0985728353,0.2387145162,-0.1280400604,-0.2601753771,0.0651919395,0.0701741576,0.2627762556,-0.1805922091,-0.0717934072,0.3756757975,0.2317004353,-0.0654206201,-0.1441975981,-0.1678650528,-0.4466333985,-0.1125073731,-0.0770775154,-0.0223392397,-0.398024559,-0.464561224,-0.0819985271,0.0392339937,-0.1288459599,-0.130652979,0.1099246889,0.0725133494,-0.1107246131,-0.0962189808,-0.1237519011,0.3803432286,-0.1346494704,0.1669492871,-0.4387069046,-0.3064742982,0.0784584433,-0.2657610178,0.3870931864,0.1349257678,0.2432437092,-0.0490192696,0.2083634436,0.283878386,0.0301167425,0.7291030288,0.1365781724,0.2848594785,-0.0325508714,0.0237695128,0.0284413639,-0.0827121437,-0.1488040239,-0.1100915894,-0.008463081,0.311124146,0.1043321416,0.0219269097,-0.2049291432,-0.0262817126,-0.0797975808,-0.1524565071,-0.0560311973,-0.4086869955,0.4224969745,-0.073427096,0.1568718404,0.2967724502,0.0521235652,0.2301136553,0.4601042569,-0.2454961091,0.2484289259,0.1642582119,-0.1416084915,0.0818946809,0.1126174182,0.1232502535,0.0427242368,0.315777421,0.086112611,0.2402341664,0.3008235097,-0.1466016471,0.276756376,-0.0183760598,-0.2106478512,-0.090726018,-0.1999837607,-0.2262797207,-0.0776029751,-0.1409638375,-0.3436534703,0.4145711064,-0.3780755401,-0.0889762044,-0.514546454,-0.2294922024,-0.1943328232,0.143674463,0.1445472091,-0.3308719695,0.1436034441,0.0640192553,-0.1637030542,0.1828568876,-0.078287214,-0.1319345385,0.2262049168,0.1023923531,-0.1155284196,-0.0547220334,-0.0947055966,0.053245157,0.153093338,0.0215565618,0.4178295135,-0.2154920995,-0.0317153186,-0.1926358938,-0.2442732751,-0.1289579421,-0.0935931802,0.0583967157,0.2677095234,0.2632389963,0.112056233,-0.0173218679,-0.2964196801,-0.0015223565,0.138573885,0.0703050047,-0.2626362443,-0.0254009813,-0.0042201988,-0.4214776158,-0.2225143313,-0.3172715008,0.1818211675,0.019150462,0.1018653512,0.2492626011,-0.0154554388,0.1136704013,-0.4128081799,0.3973900676,0.019878244,-0.2308399677,0.350573957,0.1537650228,-0.0881727263,0.0603782237,0.0165023934,-0.3299206197,-0.2907500267,-0.468072772,-0.143269524,-0.4349012375,-0.0679804161,-0.1490468681,0.2383823097,0.2412441373,0.17141141,-0.1701709032,0.1218362078,0.0448099077,-0.0607373007,-0.3937222958,0.1870473474,-0.1888151169,0.3252043128,0.195554629,0.2594074607,0.665776372,0.0236845892,-0.0280892476,-0.1768313795,0.4944063127,-0.0639406294,-0.3408125937,-0.0741281435,0.2562773824,-0.2144793123,0.0454405844,0.0918221101,0.4195421636,-0.0491156504,-0.1527400166,0.007843175,-0.4033507407,0.0125116687,0.11780487,-0.0471235663,-0.226312384,0.2386758775,0.0888541117,-0.1602493674,0.0955449343,0.2809744775,0.0332897119,0.0198510252,-0.2973652482,0.1338821799,-0.5591213107,0.3430153728,-0.1690996587,-0.0702083483,-0.2407557666,-0.0516058579,0.055274453,0.3645495772,0.1982012838,-0.0962245837,0.1989068091,-0.1740894616,0.0654193684,-0.2608310878,0.0815297589,-0.380117625,0.2637079656,0.5808417201,0.2142953873,-0.2654865384,-0.0923083648,0.1285781562,0.3016326129,-0.0371734723,-0.2965574861,-0.1809970289,-0.2461902648,-0.3428689539,0.0429153107,0.1184793934,0.0395401418,0.009085468,-0.1649234593,0.1562068909,-0.2907391191,0.4395494163,-0.2896514535,0.2015053183,0.0022370692,0.24702622,-0.0707607865,0.0638109818,0.1865650862,0.3900348842,0.1435192525,-0.2154705673,-0.2089958042,0.0146771669,0.28285411,0.267847538,-0.2713182867,0.1991046369,0.2155406624,0.4522902071,-0.2454987913,0.1314472258,0.2000246197,0.1020803005,-0.3109882474,-0.0824193135,0.2697012126,0.2748360932,-0.0914922282,0.3391312063,0.2385041863,-0.3399777114,0.0269357488,-0.0678750277,0.8009650111,-0.0983563364,0.0159791932,0.0364456587,-0.3075370789,0.4925940037,0.1397099495,-0.1278346479,-0.2091457397,-0.2514148355,0.213112548,-0.0629834309,0.2333697379,-0.2393573374,-0.1109230146,0.3117627501,-0.5727350116,-0.1116698086,-0.3081026971,0.5659891963,-0.4143065512,-0.1945654303,-0.1879529804,0.1219123751,-0.2200802118,0.0968975052,-0.0258719288,-0.0617518462,0.3228131533,-0.1328971833,-0.2411687076,-0.2225192338,0.1639807373,-0.0940937474,-0.1785970032,-0.0769694895,-0.0023673838,0.1628973037,0.220419392,0.4763893485,-0.0759224072,-0.0318377651,-0.0865626261,0.0871795416,0.0742124692,0.0477233194,0.3861101866,0.1000741348,-0.2182918638,0.1162267402,0.1808728278,-0.1384647787,-0.2111575007,-0.2440658361,-0.0712001696,-0.2504601181,-0.0137189273,0.096382305,-0.2698377967,0.0553012453,0.1864824891,0.466668278,-0.1370449066,0.3210439384,-0.0796411559,-0.4528504312,-0.335255444,0.4567136168,0.1780349761,-0.0973560736,0.6087590456,0.3077740967,-0.3386906683,-0.3302572966,0.1132151857,0.324809283,-0.1370310634,0.0727140307,0.1941923052,0.1728783399,0.1499607712,-0.2116266936,0.0990785733,-0.037350297,0.0175762828,-0.2333384752,-0.6530663967,0.2036525309,-0.1098815724,0.2779542804,-0.0393491201,0.0241972376,-0.139368996,0.2454649806,-0.3583432436,-0.3506204188,-0.0067056296,-0.0317686461,0.3579742312,-0.0167981144,0.1551988572,-0.0867018849,0.1900973618,-0.0782499239,-0.2754216194,-0.2058374882,-0.0359891504,0.0597278588,-0.3369916379,0.143892467,-0.0533091538,-0.260445416,-0.0260588098,-0.1304038167,0.1607435942,0.4848219454,0.1779703349,0.2384786904,-0.0359934904,-0.1900612712,0.0429781079,-0.3624407053,-0.1130556241,0.0304419138,-0.1487849951,0.018843893,0.0374402031,-0.0355806015,-0.1174370423,-0.007863692,0.2306164801,0.0537378266,0.0781588033,0.1827899814,-0.0827989057,0.4895751178,0.438947618,0.0256031118,-0.1897257566,-0.119970195,0.002799608,0.2263938338,-0.0861058533,-0.0895369127,-0.050210759,0.2337936908,-0.1301301271,0.3400767744,-0.1929971129,-0.0588924326,0.2365812361,0.1145133749,0.0990993753,-0.1988421232,0.1348812431,-0.0807392523,-0.1785717905,-0.3023529649,0.3346436024,0.2490024269,0.14488855,0.3417281806,-0.3027155399,0.4308220148,-0.1163826734,0.0335121714,-0.0752864331,-0.1949884295,0.0821982995,0.4469098747,-0.2531750202,-0.105563432,-0.0394356735,0.1324369609,0.0879876018,0.1847637296,-0.0655937791,0.1596029103,-0.0596900918,0.0725525469,0.5416229963,0.1970666647,0.0177034587,-0.0905037224,0.0923383236,-0.2359404713,-0.036811199,0.1797041297,-0.0802234188,-0.2119424194,0.3877393007,-0.1347739697,0.5194368362,-0.2928852737,-0.0926572308,-0.0159333404,0.0502920784,0.1685719341,0.1963849068,0.5575364828,0.2454877198,0.2644581497,0.3982055187,0.0168962665,-0.0327887125,-0.0594467446,0.3462132514,-0.1613755971,0.1701637805,0.0275316332,0.1828459054,-0.1884641349,0.156885922,-0.024892088,-0.014478392,-0.2404857874,0.4978381693,0.092603527,-0.2969304621,-0.2676284313,0.1231225282,-0.2476146519,-0.2549476326,0.3540368974,0.0495316871,0.0116406353,-0.4107545912,0.0888517946,-0.2737697065,0.1499309838,0.281503886,0.034647081,-0.3381720185,0.0800865218,-0.3323548734,0.1162737459,-0.0169758257,-0.3572063446,-0.0350071527,-0.0261938684,-0.2280015647,0.2126128674,0.3512223661,-0.2378356904,-0.0079774205,-0.2709765732,-0.0287588779,-0.0457537882,-0.0024129967,-0.0729316249,0.0525335036,-0.0724381059,0.5365501642,-0.4799968898,0.0990315676,-0.0891709626,0.0334169269,0.0580087081,0.1876063645,0.1980206966,0.0716910362,0.2010100931,-0.1422899514,0.1052196771,-0.2134283334,0.0491996855,-0.2612791359,0.153617695,0.1177540421,0.1204641536,-0.2780637443,-0.3314649165,-0.2199072987,0.4389554262,-0.0522329435,-0.2927259207,-0.0653645098,0.0723538324,0.1807265729,0.0870919079,0.1625378877,0.1299353987,0.0181755386,0.239957124,-0.1517448574,-0.6257597208,0.4860573113,-0.3297597468,-0.1410671324,0.0117129525,0.6609065533,0.2803503573,-0.2258117646,-0.536575079,0.1924078166,0.172111094,-0.1644552201,0.0537644215,0.4785366058,-0.2407764941,0.0551900454,0.1784808338,0.2521177232,0.1596432924,0.0608136281,0.0253462363,-0.3134700358]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/812","title":"Too much logging ","comments":"In the latest version of the lib the logs about locks are at the DEBUG level so you won't see them by default.\r\nAlso `set_verbosity_warning` does take into account these logs now.\r\nCan you try to update the lib ?\r\n```\r\npip install --upgrade datasets\r\n```","body":"I'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2","comment_length":46,"text":"Too much logging \nI'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2\nIn the latest version of the lib the logs about locks are at the DEBUG level so you won't see them by default.\r\nAlso `set_verbosity_warning` does take into account these logs now.\r\nCan you try to update the lib ?\r\n```\r\npip install --upgrade datasets\r\n```","embeddings":[-0.3842003942,-0.1210572794,-0.0782302096,0.1213070825,0.3027052581,0.0185212623,0.1953247935,0.3992016315,0.2092471421,-0.0036375662,0.0869948938,0.2293002307,-0.3797929883,-0.0543083921,-0.1946718246,0.1247082725,0.0595848374,-0.0120906904,-0.296502471,-0.0548450947,-0.0047584106,0.1723802239,-0.0426926166,-0.0422268845,-0.6513229609,-0.0713218078,0.2296185344,0.0982798114,-0.3177295625,-0.5123000741,0.0021837505,0.3213411868,0.0933954567,0.3480690718,-0.0001114561,-0.200807631,0.3850497007,0.0813151449,-0.3874275386,0.1584529728,-0.1273747236,-0.6889070868,0.3180123866,-0.1518696845,-0.0132954363,-0.3914661407,0.1741176695,-0.4228611588,0.3276722729,0.2203636169,0.2372916043,0.2560481131,0.195072785,-0.0656680092,0.1696548164,0.1276334077,-0.1257700175,0.0613509566,0.395629108,0.0216282569,-0.2153387815,0.481187582,-0.2172435373,-0.0182276201,0.1597385854,0.0684910342,-0.1942418665,-0.2886531353,0.2093299776,0.3352796733,0.297159642,-0.4165317714,-0.4168954194,-0.49858284,0.1036061794,-0.1400542855,0.01207841,0.0399671569,-0.0432757102,0.2632344067,-0.2367275953,-0.2274615765,0.1178487316,-0.0880375504,0.3674965203,0.0152489664,-0.1200143695,-0.0643429831,0.3860506415,-0.1922379583,-0.0838978812,0.2320225686,-0.1648689359,-0.0848813429,-0.3075753152,-0.0059677563,0.2617579699,0.5459194183,0.0173493121,0.2318050414,-0.4413781762,0.1197401211,0.1755082458,-0.0772828385,0.1295022219,0.0951479599,0.2877977788,0.4277957082,0.1300466657,0.1100871563,0.0666892081,0.0680088252,0.3630872965,-0.4080975056,0.486507982,0.0096886158,0.4269075096,-0.2582693398,0.0138986977,0.0117206462,0.0211146027,-0.140564844,0.0885047466,0.3236785233,0.0389576517,-0.1044979841,0.0325594507,0.1472505033,-0.3150570095,-0.0125614721,-0.2725060284,0.1832267046,-0.1048332155,0.1756085008,0.1836132258,-0.1565751731,0.2222394794,0.2806104124,-0.0470447801,-0.0109790359,0.080832094,0.0555732138,-0.0364440866,0.3621894717,-0.1134297177,0.1094324514,-0.0041413005,-0.1216425374,-0.161744833,-0.0908530354,0.0200258847,-0.3312015235,-0.2264630497,0.1802598089,-0.3200673759,0.0009690747,0.1256016046,0.0405621193,0.1751790643,0.0196825489,0.1185606271,-0.2449597716,0.0220148806,-0.1212942973,0.1891782135,0.7124804258,-0.1160551086,-0.1059542596,-0.2146634758,-0.2846231461,0.1850438267,0.2023046762,0.077396296,0.1450667828,-0.161702469,0.0507675409,-0.0373744033,-0.4824723601,-0.2755785286,0.1520289034,-0.1536350995,0.0235888343,0.5440300107,0.148409456,0.1424225867,-0.0662036836,-0.0997482762,-0.2115711123,0.193767935,-0.1589092463,-0.1838151515,-0.1617066115,-0.1949259788,0.1878437847,-0.1351660937,-0.2437298298,0.0579747744,0.2588115335,0.2267549187,-0.1032233164,-0.1416472644,0.3722043335,0.3256737292,0.1242635623,-0.1412417442,-0.2158426493,-0.4205309451,0.0231691673,-0.0652115867,0.0638560727,-0.3778139353,-0.3627589345,-0.0519610122,0.0710157827,-0.1508546025,-0.2055117786,0.1037454158,0.1855564564,-0.0640465394,-0.008680488,-0.2316903919,0.4891944528,-0.1248555556,0.2356144786,-0.4816490114,-0.2807700336,-0.0231933761,-0.2656398714,0.2748295665,0.2632425725,0.1486994773,-0.0271445643,0.175705567,0.3804515302,0.0298848189,0.642467618,-0.013993212,0.1570060998,0.0326066762,0.1046307236,0.1406440735,-0.1251662672,-0.2521862984,0.030247815,-0.0197069347,0.2708444595,0.0190177541,0.0220347401,-0.1444375366,-0.0326223113,-0.0307910256,-0.1097344682,-0.1263531148,-0.4837382734,0.4376625419,-0.0284979697,0.2203833312,0.1991180629,-0.0573324375,0.1922785342,0.5509569645,-0.1026852354,0.2122148424,0.1574656218,0.0063915611,0.0618393235,0.1904491186,0.1744951308,0.2084707916,0.3503152132,0.0116701359,0.2073846161,0.1112729833,-0.1860414147,0.1864984185,-0.0509207323,-0.3102308214,0.0356198326,-0.1938985735,-0.1716840267,-0.2291991562,-0.0639284998,-0.4013713598,0.3197189867,-0.3461950123,-0.0856330395,-0.4943998456,-0.0459201634,-0.1702392399,-0.0059617315,0.0087136477,-0.3549367487,0.2253211886,0.0546040572,-0.1508049667,0.2046785206,-0.0311180037,-0.0979538187,0.0676514283,0.0417160057,-0.1886265278,-0.016510332,-0.1442102343,0.0259317867,0.0992549062,0.0115960026,0.3698418438,-0.1825221628,0.0071974611,-0.2161287814,-0.332898736,-0.086498335,-0.3166137636,0.2226391882,0.3451011181,0.2186995894,0.1186477467,-0.0733521432,-0.0937786549,-0.0263979044,0.0352336876,-0.0686816052,-0.1568099409,-0.0560356006,-0.1092876494,-0.4498687685,-0.2963137627,-0.406060338,0.3338097632,-0.0031321058,0.0861224458,0.3134583533,0.0736862049,0.2679020464,-0.4539487958,0.4992460608,0.0025510641,-0.2084095478,0.2432997972,0.0237041265,-0.0945892259,-0.0157246515,0.1261738837,-0.203374356,-0.2979955971,-0.4741601646,-0.2152491212,-0.4398393929,0.0482872352,-0.0393980779,0.2068232894,0.3207935393,0.1873241067,-0.1971336901,-0.0205138903,0.0013091677,-0.0330726765,-0.3698883951,0.206555292,-0.1743269265,0.2902251482,0.0709295571,0.2890484631,0.6024174094,-0.0093561644,0.0520627424,-0.188550666,0.6065215468,-0.2101409137,-0.4003718197,-0.1404462904,0.142853573,-0.1776547879,0.0030570487,0.0743748993,0.400626272,-0.0758654624,0.0622789115,0.0110465987,-0.4428913295,-0.0294366013,0.2006862462,-0.0829342604,-0.1694946736,0.4223032296,0.1286731064,-0.1203921586,0.0783363059,0.2934119701,-0.0003821663,-0.0087834336,-0.2260134667,0.0914289728,-0.5675858855,0.3230774403,-0.1862260252,-0.0358644165,-0.3785839379,-0.1563174725,0.0676026493,0.2389382273,0.1644756347,-0.1255974025,0.0340729952,-0.0593374558,0.071685791,-0.3205618858,0.0409767814,-0.4133747518,0.1680272371,0.5538190603,0.2187894285,-0.4402222037,-0.0963387042,0.1497095376,0.2032132,0.0424953476,-0.1779584438,-0.2302888185,-0.1540991664,-0.3952036202,0.1186630204,0.0868065655,0.0124113075,-0.0390131213,-0.063446112,0.1977074891,-0.245882228,0.3211677372,-0.1554299444,0.2738792002,-0.0506216437,0.1528329998,0.0439742468,0.0103644161,0.1142114252,0.5003361702,-0.1269228756,-0.0673242286,-0.2192166448,-0.0036701893,0.1662024856,0.3981063366,-0.2110255808,0.115288049,0.3053016961,0.4079304636,-0.1593983471,0.1065626517,0.3676128387,0.1567260176,-0.1973277777,-0.0805218667,0.3228188157,0.3746466339,-0.080031231,0.3842606544,0.2945862114,-0.2395421863,0.1231454983,0.0106877815,0.8338342905,-0.0529665053,0.0665994212,0.0949242488,-0.2706331611,0.4882984161,0.0560973324,-0.1770417392,-0.1328653246,-0.1968459338,0.1891279817,0.0212027393,0.2278796285,-0.380538851,-0.3188482523,0.209636122,-0.3555574119,-0.0265578553,-0.3779291809,0.4703406096,-0.513992846,-0.1669333875,-0.1779045612,0.1617203802,-0.107590057,0.2170080096,-0.0518067554,-0.1411412209,0.2018680423,-0.1970671713,-0.5171783566,-0.1222815886,0.0933080837,-0.048053097,-0.0954968706,-0.180089429,-0.0442679636,0.1278613359,0.189040646,0.4562103152,-0.1590247899,0.0397505127,-0.1158875227,0.0600805096,0.0554581359,-0.0308993459,0.3943157196,0.0699739009,-0.2352161556,-0.0509001575,0.0684571415,-0.2278548181,-0.0601518899,-0.0872573704,-0.1193817332,-0.2786221802,-0.0296499878,0.1004617065,-0.229188025,0.023648845,0.1905824989,0.351918757,-0.0079435445,0.3124556243,-0.0720813349,-0.4334874749,-0.3517424762,0.5259714723,-0.0411589369,-0.1152185351,0.5203582048,0.3307602406,-0.3719975352,-0.2682920992,-0.001906915,0.269818455,-0.2570617497,0.059938103,0.1953255087,0.0193518605,0.3073073626,-0.1209063008,0.1388232559,-0.1146341339,-0.191272229,-0.2003658414,-0.5947332978,0.2515937388,-0.1341845244,0.3578064442,-0.0507515445,-0.1388092935,-0.0105763432,0.1137606502,-0.3678404093,-0.297586143,-0.0328447707,-0.1245165914,0.4510896504,-0.0954400748,0.3745184541,-0.0978108868,0.1554637104,0.1099663079,-0.359421432,-0.1913679987,-0.0146622863,0.0748744458,-0.2906979322,0.1637742221,-0.1543096006,-0.2334357798,0.0883499086,-0.1557351947,0.1895461828,0.4521877468,0.1326584816,0.2487798035,-0.1210598424,0.0438912027,0.2011245042,-0.3045765162,-0.0797946081,-0.0337633155,-0.0070997826,-0.0129314782,-0.0079213465,0.0274974722,-0.0652920082,0.1129453182,0.2750234008,0.0945133492,0.0813548192,0.1038772315,-0.06388399,0.4160508513,0.5061018467,0.1523867995,-0.0532419197,-0.1462579519,-0.019896619,0.237439841,-0.1275088191,-0.0834043249,-0.013637471,0.3474406004,-0.0645245761,0.2793582976,-0.0807393938,-0.0690294653,0.3258931637,0.2143790573,0.1791695356,-0.321613282,0.2039733678,0.0435351916,-0.1933200061,-0.1793675274,0.3450177312,0.1838213503,0.1449818164,0.3770676851,-0.3118774295,0.5273200274,-0.1330574006,0.1501945108,-0.0078298543,-0.1290216893,0.094642505,0.2455661297,-0.0418726504,-0.1048130617,-0.0254211128,0.3819457591,-0.1224865615,0.1320406348,-0.1750387847,0.2013734579,-0.1989213973,-0.078199476,0.5095502734,0.134643212,-0.0471679419,-0.0738073066,0.0545244664,-0.1464519501,0.0552892499,0.1259679347,-0.1458950341,-0.3475178182,0.2371222228,-0.1874280125,0.433835268,-0.1879196912,0.0313360319,0.0992283672,-0.0288842283,0.1070844606,0.3526528478,0.6822696328,0.2854704857,0.2022926211,0.2540796101,-0.1048102304,0.0227650683,-0.1309242398,0.4743940532,-0.1250692457,0.1397269666,0.1326144487,0.1673960239,-0.2666597962,0.1741353869,-0.0964728668,0.0576641075,-0.1789549589,0.561660707,0.0397817977,-0.2426114678,-0.2239872068,0.086728707,-0.384768784,-0.1459619999,0.3230116963,0.1246517077,-0.0189330168,-0.5199956894,0.0988149196,-0.0991550758,0.2348687351,0.335375011,0.0044651087,-0.2047855854,0.1303916276,-0.5033686757,0.0234639142,-0.0397754386,-0.2664725482,-0.1572218537,0.0197522752,-0.1950364262,0.298410207,0.3773124218,-0.3138506413,0.0501940995,-0.2426927388,-0.1993445903,0.0837258846,0.1342658997,-0.1894589663,0.1656033993,0.031353686,0.3460461497,-0.4697655737,0.0481717028,-0.1129009873,-0.0412015133,0.1014007553,0.2144018561,0.2249889672,0.1070279777,0.2256126851,-0.1252801269,-0.1326645166,-0.426047653,-0.0651868507,-0.1742409617,0.1416390389,0.0835433081,0.2157322168,-0.3106633127,-0.3851441443,-0.3129428327,0.4378101528,-0.0415195748,-0.2413826436,0.0373814069,-0.0521590486,0.1979452968,0.0089824451,0.1308052987,0.1892101467,0.0927538574,0.3730069995,-0.2739179432,-0.6905542612,0.5269720554,-0.3244993687,0.0147657432,0.0947725996,0.4946198165,0.248242721,-0.3155159056,-0.7181566358,0.1357464194,0.0534834899,-0.1045395508,0.0733515099,0.412781775,-0.1984711438,0.0815456212,0.0882646888,0.3735580742,0.1968112588,0.0937994197,0.1009653211,-0.2316855192]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/812","title":"Too much logging ","comments":"Thanks. For some reason I have to use the older version. Is that possible I can fix this by some surface-level trick?\r\n\r\nI'm still using 1.13 version datasets.","body":"I'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2","comment_length":28,"text":"Too much logging \nI'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2\nThanks. For some reason I have to use the older version. Is that possible I can fix this by some surface-level trick?\r\n\r\nI'm still using 1.13 version datasets.","embeddings":[-0.1845477223,-0.0496453494,-0.062777333,0.1152714267,0.3269049823,0.223525539,0.2482613325,0.5285163522,0.0655084327,-0.1148342267,0.0331851132,0.0734856799,-0.3805565834,0.0945336744,-0.1415010095,0.0496978089,0.0995935127,-0.134514451,-0.2368705869,-0.0040861694,0.0368101299,-0.0554204546,-0.0283102822,-0.0276657771,-0.6987543106,-0.057560008,0.1499656737,-0.0849991739,-0.1258638203,-0.4327806532,0.0435505845,0.2825653851,0.347121954,0.2078094333,-0.0001142335,-0.172021687,0.4031152725,0.091663219,-0.3014632165,0.3300531507,-0.0656788871,-0.4610651135,0.2376310527,-0.0499527194,-0.0254255533,-0.5238725543,0.1842403263,-0.4831919372,0.3442483842,0.145630613,0.2353027761,0.1066778302,0.0380543396,0.0139943054,0.0903736874,0.3014698625,-0.0019081363,0.0512065217,0.346989274,-0.134948045,-0.1935210079,0.4127218127,-0.1736093909,-0.104301393,0.3182494938,0.0128091397,-0.0742239431,-0.2959928513,0.3224023581,0.2155657709,0.2681922019,-0.3811312616,-0.3995742798,-0.5169018507,0.0863062814,-0.2460647076,-0.0615998879,0.0412022956,-0.0013887151,0.3254362345,-0.2808995247,-0.3249802887,0.1087714955,-0.1722458899,0.2617557943,0.0027784535,-0.0368907973,-0.0374573171,0.3654305935,-0.1419900358,0.0854225233,0.2243274599,-0.186613366,-0.1186130047,-0.4022973776,-0.0864095986,0.2720977664,0.5041313767,-0.0741478279,0.2894924581,-0.6020694971,0.0333600156,0.1272958368,-0.2467955947,0.2720279694,0.0735769942,0.373291105,0.5536032915,0.0765608624,-0.0217355452,0.1850701272,0.0759661347,0.2913796008,-0.4003849924,0.629802227,-0.0334342271,0.4733229578,-0.2939380109,0.1994019747,-0.1041646525,-0.0936498716,-0.0824798197,0.0064308448,0.2301989943,0.2105492204,-0.1513177454,0.0179953892,0.2042636275,-0.1932067126,-0.2217479199,-0.2325340956,0.2259944081,0.0028698135,0.2076526731,0.2626849711,-0.1796609014,-0.144416824,0.1959083974,-0.2313893288,-0.0759731308,0.1347895414,-0.070946984,-0.0010419055,0.2148303539,-0.1866118759,0.1901022941,-0.0359654501,-0.1311169416,-0.2601169944,-0.0466415808,-0.0500824451,-0.3688070476,-0.1743072122,0.1516400874,-0.2515847087,0.0005147704,-0.0462866426,0.1082532927,0.1905434579,0.2031271458,0.058156319,-0.205994308,-0.2135687172,-0.0273733623,0.0847042352,0.7543078065,-0.1837232411,-0.1074110493,-0.3210379779,-0.244593218,0.1847121269,0.2475205213,0.1200597882,0.3297619522,-0.2883557379,0.0214670021,-0.0718074515,-0.4368183911,-0.215370357,0.1418339163,-0.1801400334,0.0469524451,0.5603656769,0.14654167,0.2135159224,-0.2289565057,0.0564918034,-0.177875936,0.0707222447,-0.1444955915,-0.1745427698,-0.1741807461,-0.0578502901,0.2597694099,-0.0377275422,-0.1251596361,0.010942257,0.0561600626,0.301389575,-0.2052971572,0.0062508853,0.3207732737,0.2069016695,-0.1019963995,-0.1961970329,-0.1875396073,-0.4523920119,-0.0530892164,-0.0805827156,0.0437465049,-0.3083762228,-0.5036979318,-0.0187294688,-0.028476391,-0.1009290442,-0.1177567244,0.038482707,0.1760460734,-0.0963705406,-0.1906676888,-0.0979926139,0.3941920996,-0.1013596356,0.1274788827,-0.4881790876,-0.2294750959,0.0685021281,-0.1784319133,0.4208115637,0.1858517081,0.187051788,-0.0038346234,0.2205534279,0.2407078147,-0.0329660252,0.7907809019,0.0644707307,0.337153852,-0.0882408917,0.1161919013,0.0621470176,-0.0583439618,-0.2887877822,-0.0298366267,-0.0413102433,0.4230466783,0.2049775571,0.1118844748,-0.1287044883,0.0438720994,-0.1169309765,-0.150756672,-0.0841367841,-0.50324893,0.4043030739,-0.0699333474,0.1761554778,0.2005905062,-0.0713346675,0.3584004343,0.4714872539,-0.2176748961,0.2357916981,0.2873411179,-0.0208740979,0.0213779658,0.1399319023,0.1880250573,0.1088555008,0.2872410119,0.0148908999,0.3430375755,0.3121269941,-0.0554930381,0.300398469,-0.0318197086,-0.244365558,-0.1732458025,-0.0298624411,-0.1476175487,-0.0066398615,0.0383540168,-0.4003477991,0.3763830364,-0.4999678731,0.0595466942,-0.6568230987,-0.2970498204,-0.2544571459,0.1902570277,0.039268665,-0.3844943941,0.1392332911,0.0904712453,-0.145260185,0.1740479767,-0.0440053269,0.1154812723,0.0897040889,0.1575214565,-0.3321352303,-0.0229365006,-0.1803100407,0.0299765542,0.2657260895,-0.005944049,0.4097063541,-0.200605467,-0.051324062,-0.2849327624,-0.3416579366,-0.0627895817,-0.1004976705,0.1743464172,0.2368714213,0.1116315499,0.0671489239,0.0330287442,-0.1913238019,-0.0941478387,0.0959202647,-0.0322462767,-0.2914229035,0.1275749952,-0.0760287493,-0.327640295,-0.2210877985,-0.2438619435,0.3612037003,-0.0425977185,0.1103075221,0.3691274524,-0.1180830151,0.0815081075,-0.413200736,0.422019273,-0.0115909111,-0.2237794101,0.254720211,0.0998540521,-0.0389037132,-0.0693699345,0.0822910964,-0.3534398973,-0.2201089859,-0.5487343669,-0.2341194451,-0.4150407314,-0.0939439908,-0.0649746358,0.185556367,0.2899772227,0.1923263669,-0.0875974,0.1645504981,-0.0551616661,-0.0478195921,-0.3874906301,0.1405472755,-0.0966754854,0.2842620909,0.1976348758,0.3441001773,0.5574746132,0.0424780734,-0.0754356161,-0.2141036689,0.5718830228,-0.2818314731,-0.3039311767,-0.1314197332,0.115601778,-0.1821359247,-0.0945821851,0.0998164713,0.343189925,-0.0591255911,-0.1497662514,0.0663819239,-0.5078800321,-0.0191528071,0.2249501199,-0.0265607033,-0.1209883168,0.2878616154,0.0156217366,-0.2363460809,0.1701835841,0.2459603399,0.1029064506,0.0323910452,-0.4344898164,0.1664490253,-0.7213892937,0.4466557801,-0.1041037217,0.0341234431,-0.210439682,-0.1561363339,0.0077646463,0.3723122776,0.2193949968,-0.1493701041,0.2264133096,-0.0196223035,0.0373189487,-0.327455014,0.015356686,-0.3497527838,0.2822166085,0.5570352674,0.1623637229,-0.1792654246,-0.1803703904,0.1766922474,0.2699890435,-0.0373566449,-0.1949124783,-0.1092564687,-0.1963038146,-0.2780587673,-0.0031816827,0.1013177857,0.0660798699,-0.0230639968,0.0405233614,0.0406473577,-0.310487926,0.3731943667,-0.3377861679,0.2481720597,0.0325406045,0.2477478236,0.0192487575,0.1079976186,0.2164223194,0.3639289737,0.0886834487,-0.2177188694,-0.1563559175,-0.0123328352,0.2017846704,0.4631585181,-0.2448465377,0.2231055349,0.1815955788,0.3818926215,-0.3454277813,0.0008877561,0.1865532398,0.1166730821,-0.2835581005,-0.0918461457,0.2894295156,0.3648074567,-0.1354424953,0.4223409891,0.2825657427,-0.4037981033,0.0595988706,-0.013427753,1.0228235722,-0.0958488733,0.0820837468,0.0964177698,-0.2569852769,0.4764894843,0.1928961277,-0.0476426445,-0.2166042924,-0.3129029572,0.1542139798,-0.0362482406,0.1545402557,-0.315677762,0.0204742551,0.3621283472,-0.4943402112,-0.0015035503,-0.2654562294,0.4618159533,-0.3782171309,-0.157431066,-0.3005764186,0.0780272484,-0.0972107351,0.0793013647,-0.0983290598,-0.0861012191,0.325499326,-0.1586666703,-0.3358669579,-0.2780641615,0.2462984473,-0.1289514303,-0.0743101686,-0.094782941,-0.0116019994,0.0951548591,0.2536318004,0.3788901269,-0.1345601827,-0.0864739493,-0.1825804561,0.1769082546,0.177061379,-0.0751286224,0.2750737965,0.0701155365,-0.1578247845,-0.0040239724,0.1588584781,-0.306646347,-0.1811248511,-0.1831434667,-0.0824746341,-0.1402903497,0.0432122461,0.0777512491,-0.2249747217,0.0777299851,0.1457720101,0.2886354029,0.0971757397,0.1494158059,-0.1460623741,-0.3254294693,-0.3056714833,0.5781704783,0.0622143075,-0.0449513458,0.5925279856,0.3270190954,-0.298874259,-0.2708675861,0.0871448889,0.3117555082,-0.2120977342,0.0858569965,0.0864734203,0.1492685825,0.2055312544,-0.1347446889,0.0223297775,-0.0177576803,-0.0332336277,-0.2165286988,-0.576238513,0.2602948844,-0.1569847316,0.2617209852,-0.0903237835,-0.0142189302,-0.1747016311,0.1199701801,-0.3274497092,-0.3087001741,0.2129466832,-0.1259738207,0.3004691899,-0.0567873903,0.2129829526,-0.1967760026,0.0944234431,0.0855095163,-0.2549470961,-0.1676031202,0.0385931805,0.0978062525,-0.4013964832,0.0355880372,0.0576630719,-0.2081139237,-0.0697721243,-0.18695952,0.1242233515,0.5811994672,0.178498283,0.2072706819,-0.1150246635,-0.0837635696,0.0948124751,-0.267118752,0.0116609391,0.1067254394,-0.073956646,0.0972134173,-0.039752923,0.0958201215,-0.0369275995,0.1255935729,0.1979068816,0.085253045,0.0882780775,0.2324172556,-0.1626403481,0.532746911,0.4566958249,0.0835588723,-0.2391476184,-0.1167793348,-0.0306135118,0.1834724993,-0.1348665059,-0.0700273365,-0.035843242,0.2857700884,-0.2206103951,0.4170994759,-0.1020311639,-0.1228812188,0.2743918896,0.1356431693,0.0889701545,-0.2586748898,0.2499249727,-0.011815263,-0.1307216436,-0.3220930099,0.4062694311,0.2319776863,0.1693901569,0.3233453929,-0.2802147865,0.3817972243,-0.1167324707,0.2175147086,-0.0006153039,-0.2654291689,0.1276232302,0.51958251,-0.1552280486,-0.0818763599,-0.0869217813,0.1188352257,0.1133855209,0.0546952449,-0.2093761265,0.1882802546,-0.015124009,-0.0388146155,0.5575550199,0.1163374037,-0.0407957099,-0.0006743617,-0.0272410922,-0.2548867166,-0.0662177876,0.1104628593,-0.0752438009,-0.2503231764,0.3213796616,-0.1419885159,0.6471573114,-0.2481482029,-0.1184149534,-0.0314656459,-0.0207130611,0.2452705055,0.2481629401,0.6147253513,0.1973527521,0.2949056327,0.383320272,-0.0775427297,0.0347493924,-0.0457638726,0.3528981805,-0.0421327911,0.0435266607,0.0527079813,0.1348374337,-0.0971958935,0.271974802,-0.1069201604,0.1292623729,-0.4247208536,0.5578889251,0.1249479875,-0.1892279088,-0.1546586752,0.0768140405,-0.1417121738,-0.2965764403,0.4082415998,0.1766599417,0.0283533689,-0.3209650218,0.0760010406,-0.2978374064,0.1563452929,0.3977246583,-0.0668805391,-0.2303667814,0.065509893,-0.3557760417,0.0461469926,-0.0248779282,-0.1748913825,-0.0113251442,-0.0486216284,-0.088965483,0.1923786849,0.3781264722,-0.3197563291,0.0550173819,-0.3638294637,-0.0571337417,-0.0237722751,0.0704446957,-0.091581516,0.1109790429,-0.1278048307,0.4835258722,-0.5077933669,0.0508920364,-0.0503235459,0.0080067441,0.0118280761,0.2130625993,0.2893228531,0.1443718672,0.160549134,-0.0331975967,-0.0812246129,-0.448946625,-0.0131015368,-0.30248034,0.3007672727,-0.0522003509,0.1398361176,-0.399676621,-0.3157457113,-0.3180856407,0.4130194187,-0.0654371455,-0.2322542816,-0.1681165546,0.0659503788,0.0782139376,0.073844716,0.1094259098,0.1592604965,0.0564104207,0.3194153309,-0.2590790987,-0.5449332595,0.5200135112,-0.398548007,-0.0734979883,0.000620353,0.6537173986,0.2560561597,-0.3464394212,-0.5534003973,0.0777347311,0.1594492942,-0.1088631377,0.1241916642,0.4242925644,-0.2983648777,0.1354717016,0.1167238504,0.3301056623,0.1954041123,0.0571990199,0.0780385137,-0.3463290036]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/812","title":"Too much logging ","comments":"On older versions you can use\r\n```python\r\nimport logging\r\n\r\nlogging.getLogger(\"filelock\").setLevel(logging.WARNING)\r\n```","body":"I'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2","comment_length":11,"text":"Too much logging \nI'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2\nOn older versions you can use\r\n```python\r\nimport logging\r\n\r\nlogging.getLogger(\"filelock\").setLevel(logging.WARNING)\r\n```","embeddings":[-0.2052911371,-0.0674033687,-0.119885169,0.0261179078,0.3543666005,0.2235410213,0.2691494226,0.4916390777,0.0920032337,-0.0940798298,0.0488945544,0.0491936915,-0.3852418363,-0.0098099886,-0.2292221934,0.1742514074,0.0487484187,-0.1101985723,-0.1879782826,-0.0881527141,-0.0236166213,-0.0306639764,-0.0316638499,-0.035746377,-0.8591839671,-0.1206995025,0.2094052136,0.078721717,-0.1901960224,-0.387853533,-0.1441777945,0.2157269418,0.1974353045,0.2661120296,-0.0001077709,-0.2456912249,0.2903308272,0.0974795446,-0.3914390504,0.3382302821,-0.0948767364,-0.4716241062,0.2750936747,-0.1480242312,0.0238754023,-0.5220648646,0.1866169274,-0.5077090263,0.2385904193,0.2121922821,0.2447423339,0.0290352572,0.0104631763,0.1233247742,0.0585141666,0.1750897914,-0.0395898446,0.055735752,0.3464447856,-0.1580902934,-0.3009896874,0.4053758979,-0.1806652695,-0.0254221577,0.2453324348,0.0438961126,-0.1568254381,-0.3299164474,0.3078750372,0.2866147161,0.2949096859,-0.4150785506,-0.3199409544,-0.5474219322,0.1366468519,-0.2422443777,-0.1237397194,0.1534412056,-0.0442631394,0.3107481599,-0.2536311448,-0.1431477964,0.0948471427,-0.140265286,0.2359163314,0.1795505583,-0.0962135494,-0.0756151527,0.3022182882,-0.1272724271,-0.0861706436,0.1045397446,-0.126535058,0.0143611999,-0.3803229034,-0.0829464495,0.2871839404,0.5144857168,-0.046429012,0.1930834949,-0.5309262276,0.049067907,0.1602815241,-0.1059748828,0.2731178999,0.0215187296,0.3004045784,0.4944974184,0.0678989664,-0.0240597259,0.0919358879,0.0988412574,0.4038672447,-0.3697163165,0.6681054235,0.005490059,0.4709145129,-0.2167280614,0.0258946735,0.0318361558,-0.0980834737,-0.1841957867,0.0421987064,0.2609640062,0.1804686487,-0.1471904665,-0.0578806065,0.167887345,-0.207514897,-0.0584326945,-0.3261314034,0.2736908793,-0.0682803318,0.1911122948,0.1911198497,-0.0389404669,-0.0394774452,0.1786063462,-0.1210414767,-0.0653684959,0.2355707288,-0.1179703176,0.061893139,0.3063004315,-0.141970396,0.144340381,0.0703555718,-0.1401538551,-0.2088998407,-0.1020398438,0.0254393183,-0.2744003236,-0.1777303815,0.2452001274,-0.2904959619,0.0383916236,0.0456523933,-0.0019884303,0.2191553712,0.2577910423,0.1721871644,-0.1634376943,-0.021830529,-0.1198580861,0.0613975041,0.7136005163,-0.1406108737,-0.0612709746,-0.2232079208,-0.1770913154,0.27233392,0.301540941,0.0914888829,0.4092086256,-0.2884482443,0.0914635062,-0.0242925733,-0.3505485356,-0.1128188372,0.2251680344,-0.256980896,0.0286707226,0.5846853852,0.1595441401,0.1166213974,-0.1799877584,0.1486265659,-0.1919081658,0.0788967311,-0.0978533179,-0.084117502,-0.057288114,-0.103600271,0.2237543762,-0.110809274,-0.3064451516,0.0011332531,0.1274972856,0.3215455413,-0.1198241562,-0.080709137,0.3425498009,0.21953246,-0.0026181862,-0.2119027972,-0.2028751224,-0.3771063089,-0.0685674995,-0.0142751681,0.0317124873,-0.4119102359,-0.4207937419,-0.0091442214,0.0024284124,-0.1501370221,-0.1845276505,0.1472315043,0.1410260201,-0.175382033,-0.0802797005,-0.2215278745,0.3817545772,-0.1176060811,0.1940952986,-0.3741809726,-0.2902238965,0.0950006619,-0.2654646039,0.3345760405,0.1419785917,0.196651563,0.0163917895,0.1734812558,0.2513225079,0.0081911152,0.6989960074,0.1413880587,0.2840221524,0.0124078235,0.070318222,0.0856867433,-0.0593318753,-0.1497045308,-0.0185351651,-0.0543748923,0.2915139198,0.0158990752,-0.0122037549,-0.1908539236,-0.0394252054,-0.1074514538,-0.0972823203,-0.1188357845,-0.4248659015,0.3917116821,-0.0227243248,0.091087088,0.2557282448,0.070487462,0.2213722467,0.4501315653,-0.2595800161,0.2041078955,0.1227677539,-0.0703012645,0.0675456524,0.1384501755,0.182089895,0.1059339941,0.3608317375,0.1439634413,0.1780830622,0.2151442468,-0.1966661364,0.2226483375,0.0135309845,-0.2912634909,-0.0468873456,-0.251270026,-0.1495673209,-0.1519480795,-0.1899175048,-0.3866452575,0.3634433746,-0.4052681029,-0.0215754677,-0.4155617058,-0.1765869111,-0.1730588973,0.1272765994,0.1071418226,-0.4094682336,0.1484881192,0.0675379112,-0.2004635036,0.1349698752,-0.2615050077,-0.1057177112,0.2344011962,0.022318339,-0.1873330325,0.0099762511,-0.0523935333,0.082644701,0.1466089338,-0.0008216842,0.4095440507,-0.1574639082,-0.05686168,-0.2097771615,-0.1979964674,-0.1058199778,-0.1711387634,0.0685409158,0.2625426948,0.2421712577,0.040310178,-0.0677326843,-0.2765601575,0.0499565713,0.1491463929,-0.0267375205,-0.1715223491,0.008407996,-0.0303460583,-0.2893099785,-0.1621551067,-0.3108730018,0.1741632968,-0.0122196991,0.0150214359,0.3338028491,-0.0718572438,0.1824508607,-0.4238116741,0.4110456705,-0.0375076644,-0.2073211521,0.3041426837,0.1170670316,-0.0384315699,-0.0149679976,0.0178918019,-0.3436717689,-0.3159582019,-0.4588204622,-0.1894664466,-0.4915267527,-0.1048420817,-0.1306759119,0.1983649284,0.2312520295,0.1591812968,-0.2149416208,0.1216331497,0.0719046965,-0.0522797406,-0.3733249307,0.1597897559,-0.2500826418,0.2503430843,0.1877085119,0.2284595072,0.5753924251,0.0286650378,0.039310798,-0.1938136071,0.5494670868,-0.0732914284,-0.3828102648,-0.0203469489,0.2238148153,-0.2552374601,-0.0031074344,0.0308050066,0.3650179803,-0.0141379125,-0.0577924661,-0.0195244718,-0.4414480329,0.0051580123,0.1507383287,-0.135242328,-0.2521697581,0.3135064542,0.0883492231,-0.1805983484,0.0555496849,0.3046924174,0.1055313349,-0.0145727536,-0.2515746355,0.1752711535,-0.5737931728,0.3016898632,-0.1598232985,-0.1436605304,-0.2975322902,-0.172819823,0.0010856653,0.2873573303,0.1271380931,-0.0614881776,0.241984576,-0.0728488863,0.0950351357,-0.2677615285,0.0442486294,-0.3960515857,0.250407517,0.5809012055,0.2315653265,-0.1886342913,-0.1298293322,0.0927034914,0.1789142042,0.0265206769,-0.2303434759,-0.1741147488,-0.2127622068,-0.336563319,0.0746911615,0.1448844671,0.0484301187,-0.0289346892,-0.0975381061,0.1843580008,-0.2744374275,0.450622946,-0.2750322521,0.1772591025,0.0711571574,0.1514905393,-0.0084604369,0.1354798973,0.110032104,0.4120410979,0.0963724479,-0.1480823159,-0.1218812093,0.0081085665,0.2394683063,0.3745937347,-0.2147532254,0.1219083518,0.2852028012,0.4761559367,-0.3245251775,0.1343444735,0.2337331027,0.1137404367,-0.3017899394,-0.0645839274,0.2865785658,0.3083015978,-0.0619991049,0.3510041833,0.3237488866,-0.3354731798,-0.0188172162,0.02013213,0.7923347354,-0.0959773883,0.100572668,0.0691165328,-0.2878498435,0.4445904195,0.1003975049,-0.1186616123,-0.160439074,-0.2591606081,0.1819228381,-0.0694870353,0.2141683847,-0.2515714169,-0.1304880679,0.1838679761,-0.4925447106,-0.0948043689,-0.3065469861,0.5243554711,-0.518548429,-0.1631577313,-0.202571556,0.1556368619,-0.1579226106,0.1386088878,-0.023911776,-0.0766728371,0.2492086291,-0.1003989056,-0.2648750842,-0.2477259189,0.2388975322,-0.1034254879,-0.1977125406,-0.0722496361,-0.1207097024,0.126955986,0.2078359574,0.4396882653,-0.0588757694,-0.0323872417,-0.1203342602,0.1173409596,0.0812968165,-0.0105331233,0.4351775944,0.12093018,-0.2034368068,0.0745191649,0.2095646113,-0.2569591999,-0.1573169529,-0.1175010428,-0.1314014792,-0.2844331264,0.0347232856,0.0669918358,-0.2847720981,0.101640366,0.2059863508,0.4314607382,-0.0852984488,0.2422533482,-0.1321609169,-0.4453549683,-0.3656235635,0.4957299829,0.0589576177,-0.1128170863,0.6250342727,0.2887277007,-0.3323740065,-0.3286766708,0.1114606336,0.3105466366,-0.042999588,0.0648219362,0.1706939936,0.1371764392,0.2033063173,-0.1835624725,0.1260010898,-0.0158590879,-0.0247001871,-0.1896881312,-0.6308752298,0.187172696,-0.1378542781,0.3121017814,-0.008396809,-0.0077292472,-0.0403372645,0.1170772761,-0.3853532374,-0.3442631066,0.0146210529,-0.1145388633,0.3640073836,-0.0985443518,0.2244706303,-0.0631977171,0.2339946479,-0.0719024912,-0.2389070094,-0.2462877929,-0.014631751,0.0450959019,-0.3285864592,0.1089471504,-0.0362000726,-0.2692704201,0.0241824873,-0.1728686094,0.1187333167,0.5027704239,0.2037312388,0.2752032578,-0.0163002722,-0.1588032693,0.1102264374,-0.4120043814,-0.0980221331,0.043193344,-0.1205744967,0.0506316274,-0.0095266635,0.0594364628,-0.0861610845,0.0971565917,0.160669595,0.0871660933,0.1081006303,0.1780662537,-0.1130297184,0.4991088212,0.4967198968,0.0526673198,-0.1197878048,-0.1896932721,-0.0873161703,0.252687186,-0.0542753264,-0.0900192186,-0.001088188,0.2914004922,-0.0433525555,0.3514550328,-0.2846549749,-0.1178453565,0.2153025717,0.147905767,0.1038818285,-0.267963469,0.1419242024,-0.0878695101,-0.2079296857,-0.2518969476,0.3713594675,0.2192092389,0.2050760537,0.356240958,-0.2227756679,0.398483932,-0.1663629562,0.0649116635,-0.1073104218,-0.1730525792,0.1630834937,0.4357622564,-0.2621763349,-0.0610284247,0.0328391641,0.0824291185,0.0352626182,0.1538039148,-0.0893297344,0.1391720325,-0.1221469566,0.0869240612,0.5570699573,0.139046371,0.0353924669,-0.0963648632,0.0854533762,-0.1925298721,-0.039169617,0.0963581055,-0.0617299192,-0.0877865329,0.3450489342,-0.1874966919,0.5355802774,-0.257119447,-0.1899001151,0.0379681326,0.0464519188,0.1575238109,0.2484600395,0.562938869,0.2373564839,0.2402896583,0.3968312442,-0.0180671979,-0.0031424121,-0.0874573067,0.3700244427,-0.1491349041,0.0608517863,0.101213038,0.2322010398,-0.2137668133,0.1918164194,-0.0529387332,0.0453448407,-0.1904869676,0.4809634387,0.1108806506,-0.211804539,-0.2611335218,0.0638843775,-0.2851133049,-0.2423845828,0.3022717834,0.0944204181,0.057040628,-0.3771576285,0.1023174822,-0.1553271711,0.0725713149,0.3641506433,-0.0016304883,-0.3082050085,0.1846642196,-0.3723882735,0.0658434257,-0.0118991826,-0.3621377051,-0.1415131092,0.002866115,-0.1383272111,0.251907438,0.3637071252,-0.2765346467,-0.0092001641,-0.2454735339,-0.0479857139,-0.0542927235,0.103854388,-0.0846109167,0.0232375432,-0.0278572105,0.4371882677,-0.5121451616,0.1470005661,-0.0617735945,-0.063308686,0.0244483817,0.1825588346,0.1866242141,0.078722544,0.1696152985,-0.1394826323,-0.0242107678,-0.3028003871,0.1036247462,-0.228605777,0.0860873759,0.0625764877,0.2195679545,-0.2863083184,-0.3200887442,-0.2294332236,0.4890381396,-0.0822902545,-0.3174558878,-0.0765438229,0.0277426634,0.2096515745,0.0491269082,0.0689726099,0.2031684965,0.0811403543,0.3661491573,-0.1318529993,-0.6378974915,0.5040845871,-0.4026662409,-0.1234646663,0.0273628701,0.6393898129,0.1898580343,-0.1647217721,-0.5054775476,0.1541882157,0.1813654155,-0.1996001154,0.0606229678,0.443575263,-0.1760131121,0.0025991804,0.1528061777,0.249320522,0.1794584244,0.1351074725,-0.0021051383,-0.2886384726]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/812","title":"Too much logging ","comments":"Whoa Thank you! It works!","body":"I'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2","comment_length":5,"text":"Too much logging \nI'm doing this in the beginning of my script:\r\n\r\nfrom datasets.utils import logging as datasets_logging\r\ndatasets_logging.set_verbosity_warning()\r\n\r\nbut I'm still getting these logs:\r\n\r\n[2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\n[2020-11-07 15:45:41,909][filelock][INFO] - Lock 139958278886176 released on \/home\/username\/.cache\/huggingface\/datasets\/cfe20ffaa80ef1c145a0a210d5b9cdce2b60002831e6ed0edc7ab9275d6f0d48.1bd4ccbce9de3dad0698d84674a19d6cc66a84db736a6398110bd196795dde7e.py.lock\r\n\r\nusing datasets version = 1.1.2\nWhoa Thank you! It works!","embeddings":[-0.2072971165,-0.0421515889,-0.1270598173,-0.0109144067,0.3720549047,0.0928947404,0.341820389,0.4875230491,0.1003893539,-0.0070675928,0.0728404745,0.093262054,-0.3375290036,0.0610220693,-0.1712955087,0.2014773637,0.0876208544,-0.0337064676,-0.2310762554,-0.0748668462,-0.0272957496,0.0074133645,-0.0612335987,-0.0842504576,-0.7995836735,0.0034260973,0.1343772113,0.0446257479,-0.2030854523,-0.3777699769,-0.0619091801,0.3231148124,0.1703692675,0.2964219749,-0.0001068483,-0.1933900714,0.3340385854,0.0699511319,-0.3568067551,0.3221870661,-0.0767343715,-0.4496650398,0.2405939251,-0.1199066639,-0.0795073882,-0.472446084,0.1666785032,-0.5273306966,0.2656652927,0.236262098,0.2749391198,0.1379924864,0.0212783664,-0.0037271623,0.0695199817,0.1534227729,-0.0534397699,-0.0277649388,0.3035541177,-0.1726299524,-0.2662975788,0.4415304363,-0.1708492041,-0.0428262651,0.2097836882,0.0529422723,-0.1599876434,-0.3221246302,0.3115822375,0.2414227128,0.2862044275,-0.4357619286,-0.3030057549,-0.4975042343,0.0939615369,-0.2403682023,-0.1162571311,0.1005813405,-0.0206672642,0.3584142029,-0.2936783433,-0.1512664706,0.0589108057,-0.1432711035,0.1929270923,0.1539541632,-0.0894465446,-0.0735480562,0.3013433218,-0.120378457,-0.0640685633,0.1339464188,-0.1426076293,-0.0314800777,-0.3686454296,-0.0847615823,0.3426159918,0.4948659539,0.0190342478,0.3103237748,-0.5614261031,0.1316720545,0.1827991009,-0.0989862978,0.2083897144,0.0791142657,0.3795782924,0.474488914,0.1612660736,-0.0120131914,0.0282223597,0.0894550234,0.3541923165,-0.3547449112,0.561634481,0.0382157788,0.4424483478,-0.2282796651,0.0490816385,-0.0210184585,-0.0164068826,-0.1363579631,0.0207493529,0.3832043111,0.1829597056,-0.1327358037,-0.0187563933,0.2207411975,-0.2393440902,-0.1073806211,-0.3708740771,0.262242049,-0.1087296531,0.1594960093,0.2036720216,-0.0563584343,0.0204671212,0.1050315201,-0.0550772622,-0.0621464439,0.1771032363,-0.0191611946,0.0917800814,0.3259897828,-0.097680904,0.1286621243,0.0873195753,-0.1565110385,-0.1923043132,-0.0825317726,0.0904871523,-0.3325644135,-0.1454396993,0.2712183595,-0.3111817539,0.012723458,0.1527120024,0.030307088,0.2072375864,0.2086056024,0.1110117361,-0.1515553743,-0.0536906719,-0.1643035561,0.1639148891,0.6789157987,-0.1487446129,-0.0376678668,-0.2206890881,-0.2317341715,0.2635372281,0.3276049495,0.0599535704,0.3056932986,-0.2579527497,0.1149241552,-0.0715271533,-0.3724373579,-0.1373443156,0.2326922864,-0.240813911,0.0207340159,0.5656962991,0.121892564,0.1428390592,-0.1676103026,0.0324903056,-0.1258678734,0.0662775561,-0.0950739607,-0.0841894522,-0.080484584,-0.0875310227,0.2190512866,-0.0676827356,-0.2174096256,0.0403949209,0.1049749479,0.3169876039,-0.1625739783,-0.1086572185,0.3037968874,0.2670803368,-0.0861581117,-0.1515015066,-0.1527800709,-0.3918775916,-0.029622186,-0.0074394215,0.1183358058,-0.3846380711,-0.4320325553,-0.0599063188,0.0288126543,-0.1600007117,-0.2787984312,0.1612262428,0.1693100184,-0.12410108,-0.0612521991,-0.2399566025,0.3718621135,-0.0759162679,0.1581296325,-0.3817763031,-0.3282312155,0.0348796211,-0.2451961488,0.2778156996,0.1857216358,0.1715683341,-0.0370041206,0.2152755111,0.3252077997,0.0343080238,0.6975376606,0.1060427129,0.2060071081,-0.0282751415,0.0707596689,0.0361081026,-0.0731381252,-0.1471807808,-0.1059285179,-0.0669256896,0.3904894888,0.0607495457,-0.0070909918,-0.1604170501,-0.0401563942,-0.1113884076,-0.1303786337,-0.1524119079,-0.4319553375,0.3726623058,-0.0872874036,0.0843809023,0.2195233703,0.0216540471,0.2733835578,0.4780654013,-0.2018724084,0.2096547782,0.0737715214,-0.0832980946,0.0678053349,0.1549528986,0.147464782,0.1252212971,0.3761994541,0.1453146636,0.2125097513,0.1867551059,-0.2206881493,0.2781381309,-0.0480276234,-0.256149441,-0.0866105631,-0.2563628554,-0.2087423056,-0.1912757158,-0.1265256703,-0.3982666731,0.3201007843,-0.402880013,-0.0699002594,-0.4855633974,-0.1926820576,-0.1259313077,0.0872338861,0.1459980011,-0.4008987248,0.1871897131,0.0644466132,-0.2420402914,0.1871481389,-0.2011657804,-0.1157811731,0.1559434533,0.0228556246,-0.2002104968,-0.0683199242,-0.0904259384,0.1140703037,0.1284175664,-0.0272658635,0.3838983178,-0.1683210433,-0.0180447809,-0.2206968963,-0.2916073203,-0.0971559584,-0.2028935403,0.04410851,0.2730210125,0.2157381028,0.1180011034,-0.0138853285,-0.2336576581,-0.0298386682,0.0694775656,0.0166608449,-0.1693614572,-0.0591710508,-0.1562369019,-0.3680280447,-0.1253471673,-0.3466905355,0.1917095482,0.0248790029,0.0471346639,0.3521454334,0.0146990176,0.1928367764,-0.3979170918,0.3489866257,-0.0427863151,-0.195734933,0.3483536839,0.0601102412,-0.1239596754,0.0278483406,0.0421812385,-0.2777071595,-0.2983964682,-0.4040144384,-0.1371520609,-0.4503412843,-0.0691232383,-0.0086018816,0.2261509895,0.2848710716,0.1242704093,-0.2314003557,0.0496885069,0.0014713894,-0.0656005442,-0.387693733,0.1468873918,-0.2541107237,0.2840368748,0.1954375356,0.2454082668,0.5262454748,0.0271463394,0.0521527641,-0.2421007603,0.5873789787,-0.1416986138,-0.3544979692,-0.0769999549,0.1910565048,-0.2407175303,0.0377028175,0.058770366,0.3360291421,-0.0862289146,0.0453688502,-0.0159330778,-0.4496228695,0.0073270374,0.1529909968,-0.1942164004,-0.2058907002,0.3534441292,0.1319040209,-0.1034213901,0.0658456534,0.2426899374,0.0709318221,-0.0169200543,-0.3286457062,0.2058852762,-0.6358062625,0.3432625532,-0.2477063239,-0.1295188367,-0.2432977855,-0.2572724819,0.0263375062,0.2290343195,0.2653454244,-0.0752578825,0.1824516207,-0.0882386714,0.0602201857,-0.239326641,0.0654156432,-0.3593332767,0.1756491214,0.5247288942,0.2169123143,-0.2306214273,-0.113462165,0.1147004962,0.1456314027,0.0651927963,-0.2302237302,-0.2207626998,-0.2609400451,-0.3673142195,0.083751291,0.1492739618,0.0491405949,-0.0814662576,-0.0743503645,0.1636988819,-0.2827504873,0.4548284709,-0.2183733582,0.1948742419,-0.0040143752,0.1502344757,-0.005087086,0.0855090246,0.0662921593,0.4681385756,0.1149360761,-0.157589823,-0.1528250575,-0.0366553776,0.2557964921,0.4408506453,-0.2438920289,0.0925325677,0.275580436,0.4398546219,-0.255513221,0.1326848567,0.2358936518,0.0549297892,-0.2694128454,-0.0691900551,0.2647117674,0.2216093391,-0.0645202845,0.3086703718,0.3270541131,-0.3197074831,0.079205811,-0.060474772,0.8406259418,-0.089592427,0.1123336405,0.090547204,-0.270766139,0.4847816825,0.1065820903,-0.0943225026,-0.2330712676,-0.2279144824,0.1711331606,-0.0525421686,0.2119799554,-0.2519904971,-0.1686886251,0.2249435782,-0.4618059099,-0.022187518,-0.2802208066,0.4757481515,-0.4430201352,-0.1490007937,-0.1369291842,0.1894237995,-0.0941581726,0.214441359,-0.0284998566,-0.0893548504,0.2286561579,-0.1864808053,-0.2867892683,-0.117096819,0.2136545777,-0.1138760746,-0.191687271,-0.0531553403,-0.1205135137,0.1205999479,0.2392141521,0.3897264898,-0.1100529954,-0.0055549373,-0.0371728837,0.1632100195,0.1074266583,-0.0600676239,0.438649863,0.0736427903,-0.1675469726,0.0468620658,0.196427539,-0.2550866306,-0.2176436037,-0.0729614198,-0.1076303199,-0.2656870782,-0.0196116697,0.1115134582,-0.2958101928,0.0296231546,0.2175111622,0.3719476461,-0.0752796531,0.2943819165,-0.1342274249,-0.4270266294,-0.3509364426,0.5156496763,0.0327279381,-0.1010064706,0.538118422,0.3198406696,-0.3227784932,-0.3465844989,0.0212629084,0.3510243595,-0.1045146734,0.0618090108,0.1484718323,0.0644823834,0.2346545905,-0.1586324573,0.0949811712,-0.0510688424,-0.0268018115,-0.2014493346,-0.6026070714,0.1564708203,-0.1330656111,0.3580900729,-0.0186480824,0.0389538519,-0.0652523786,0.0989039168,-0.4083816409,-0.3238801956,0.0281659476,-0.0993476883,0.3314874768,-0.0510052927,0.2674718797,-0.093008779,0.2400191426,0.0020457546,-0.3016398549,-0.2559558749,-0.0078774616,0.0617695935,-0.3060799837,0.196380049,-0.0229267273,-0.2392114997,-0.0054270844,-0.1772144884,0.1677032113,0.4789828062,0.150842458,0.221645847,0.0031102241,-0.0976249203,0.1465762854,-0.4305880964,-0.076234214,0.0003710097,-0.1262813956,0.0287395101,-0.0511285216,0.0066744038,0.000118538,0.0892125368,0.1682917774,0.1137302518,0.1235962287,0.1415705532,-0.1275117695,0.4035620987,0.5423166752,0.090196006,-0.1507866681,-0.1787838489,-0.0469495505,0.2773491442,-0.1001856849,-0.0510925576,-0.0471261926,0.2866619229,-0.0547112525,0.3634022474,-0.2030724585,-0.0379603356,0.2640381455,0.1484770477,0.1415473521,-0.2986109853,0.1169168949,-0.0980272666,-0.2521897852,-0.2285898477,0.3685877919,0.2144796401,0.1361896247,0.4119732082,-0.2335571945,0.4428212047,-0.1332263499,0.0577428378,-0.1041793004,-0.1745724082,0.1568118334,0.379219681,-0.2009409815,-0.0779096037,0.0256101936,0.1553150266,-0.0002731617,0.1813223064,-0.0846460015,0.1327633411,-0.1077181548,0.0296958014,0.4685586989,0.0815439299,0.0075508622,-0.1307707876,0.0630295351,-0.181604594,0.0101379575,0.0790170133,-0.0672504455,-0.1634048074,0.297126621,-0.3002113998,0.5440413356,-0.3017387986,-0.1053767055,-0.0016213665,0.0279271044,0.205831036,0.1878135949,0.6059619188,0.2629193068,0.2073188573,0.3068871498,-0.0818183497,-0.0815026686,-0.1238088608,0.4002744257,-0.0981098339,0.082080923,0.1439480633,0.229719311,-0.2266088277,0.1767390817,-0.0125021543,0.0282511972,-0.2033626437,0.5581729412,0.1074460521,-0.2445364296,-0.2710513771,0.0942877084,-0.2703098953,-0.2345186919,0.2849091887,0.0736301318,0.0538234524,-0.470166862,0.1161281019,-0.0972346961,0.1266868562,0.3145243227,0.0090690749,-0.2851844132,0.1689267457,-0.4242037237,0.1233613864,-0.0172124542,-0.3290288746,-0.1450751275,0.0750231519,-0.0750782266,0.2492833436,0.3248847723,-0.2792486846,0.0261166058,-0.2550193071,-0.0436829925,0.0230305102,0.0242618732,-0.0899434686,0.0789784715,-0.0686001107,0.4532530904,-0.5257182717,0.1607640982,-0.0623750798,0.0185792074,0.1256386936,0.1732529849,0.2345741242,0.0895607248,0.204596594,-0.1528449804,0.0039869081,-0.3418217301,0.0876211748,-0.2788594961,0.1292668879,0.0688835159,0.1891236603,-0.2743892372,-0.3496080041,-0.2311889678,0.4637450278,-0.0555268452,-0.2339568585,-0.0095175207,0.078488417,0.1628813893,0.0530864596,0.1177381054,0.211753428,0.000607522,0.3405151367,-0.1944128424,-0.7077479959,0.5256866217,-0.3543000519,-0.1893302947,0.0968135744,0.6071866751,0.2119102627,-0.2117572576,-0.5580085516,0.1347529441,0.116975382,-0.1461832374,0.0582034588,0.4478223622,-0.202415958,0.0305164382,0.1105089039,0.2779237926,0.2106987089,0.0945011452,-0.0406196192,-0.3620716035]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/811","title":"nlp viewer error","comments":"and also for 'blog_authorship_corpus'\r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=blog_authorship_corpus\r\n![image](https:\/\/user-images.githubusercontent.com\/30210529\/98557329-5c182800-22a4-11eb-9b01-5b910fb8fcd4.png)\r\n","body":"Hello, \r\nwhen I select amazon_us_reviews in nlp viewer, it shows error.\r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=amazon_us_reviews\r\n![image](https:\/\/user-images.githubusercontent.com\/30210529\/98447334-4aa81200-2124-11eb-9dca-82c3ab34ccc2.png)\r\n","comment_length":6,"text":"nlp viewer error\nHello, \r\nwhen I select amazon_us_reviews in nlp viewer, it shows error.\r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=amazon_us_reviews\r\n![image](https:\/\/user-images.githubusercontent.com\/30210529\/98447334-4aa81200-2124-11eb-9dca-82c3ab34ccc2.png)\r\n\nand also for 'blog_authorship_corpus'\r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=blog_authorship_corpus\r\n![image](https:\/\/user-images.githubusercontent.com\/30210529\/98557329-5c182800-22a4-11eb-9b01-5b910fb8fcd4.png)\r\n","embeddings":[-0.1825665832,0.0957985595,-0.1462980062,0.0771785751,0.091613315,0.1810390055,0.0066385334,0.1151603684,-0.0829188973,-0.0309123192,-0.077577278,0.0548243001,-0.0459580198,-0.0133008454,0.0272896346,-0.2342224568,-0.091261141,0.1647726148,0.1516455412,0.0272852816,-0.1843310446,0.285782814,-0.2277496904,0.2714023888,-0.0895864442,-0.1787430495,0.1502074599,-0.2478437424,-0.2005053312,-0.4086381197,0.2321772575,0.0413870439,-0.0414073281,-0.0605833791,-0.000106674,0.0449138582,0.5803546309,0.0132858967,0.0879850909,-0.5440975428,0.1008808687,0.0416048691,0.0850938782,-0.1619734317,-0.0436260179,-0.3603275716,0.3084620833,0.4110165238,-0.0036558574,0.3728628755,0.3271819949,0.1920292377,0.2094535679,-0.1873826385,0.0385575704,0.0977322012,-0.0692125857,0.0799713731,0.0895411596,-0.1429609358,-0.0501141548,0.306308955,-0.0143487081,-0.0773746818,-0.0745503455,-0.0761744827,0.3562167883,-0.0312033761,0.281309545,0.0259252619,0.6786999106,0.0157161579,0.0447243974,0.1944553256,0.1933368742,0.2479806542,0.1133357659,0.4597432315,-0.1104982868,-0.0787468255,-0.412648201,-0.1476917565,-0.2275767475,0.2025402337,0.0183921307,0.1508202851,0.0183884371,0.1456496418,0.238008067,-0.0128425797,0.1839567572,0.3275102675,-0.0688070431,-0.0976118743,-0.2788859606,-0.2325759977,0.200483188,0.2470029294,-0.2276696116,0.2938127816,0.1241552383,-0.066729866,0.1127929837,0.3037103415,0.1273401678,0.1140986085,0.2106921226,0.3541476727,-0.1132814884,0.2238243073,-0.0056592338,0.0447100438,0.0117173875,-0.4341824353,-0.381966114,-0.2042743415,-0.0143008176,-0.3738906085,-0.5338623524,0.073188439,-0.2940172255,0.0959473252,0.0640055463,0.5672889352,-0.057642933,-0.0226875637,0.103539601,0.1745617241,-0.2112367004,-0.0419372655,-0.1841945499,0.4919338822,-0.2035572827,-0.185360387,0.0986609086,-0.1469860524,0.4239937067,-0.2314012796,0.2135293186,-0.189665705,-0.0136854043,-0.2147449404,-0.1462827474,0.134383902,0.1347830892,-0.1423915625,0.0627067611,-0.073048085,-0.0378420018,0.1621421576,-0.0075000152,-0.1669643968,-0.2253580242,0.2710672021,-0.1049102917,-0.3737588525,-0.0482072495,0.138818711,0.0017103547,-0.1693854034,0.0717033222,-0.0369459353,-0.082334049,0.0123756211,-0.2495905012,0.078172572,0.0590218864,-0.1532998234,-0.2732342482,-0.1902891845,0.1659972072,0.2621859908,-0.1763107926,0.0143780438,-0.1002788618,0.3809202611,0.4203782082,-0.3325335085,-0.1113525406,0.2488748133,-0.1366476119,-0.3971332908,0.1733846664,-0.0257025659,-0.1895457804,-0.3249613345,-0.440790236,0.0819530264,0.0234918483,0.0594345592,-0.2670687139,-0.2087540478,0.219004035,0.3076311648,0.1577041447,-0.0479475334,0.083289966,0.3810984194,0.1320216954,0.166741848,-0.1935451031,0.3279053569,0.0695236847,0.0434742272,-0.0225947089,-0.0262266081,0.0501569659,-0.146075204,-0.5252537727,0.1796812862,0.1507000923,-0.0536815524,-0.5101611018,-0.0622689873,0.0162140112,-0.1301204711,0.2977062762,-0.0495638698,0.0864796489,0.1305490285,0.0903099179,-0.2062642872,-0.1790338755,-0.0192669462,-0.2055384815,0.3475040793,-0.4210826159,-0.0439767428,0.5000002384,0.1449655592,0.3338680565,0.0691970289,0.0774110854,0.1069530398,-0.5652883649,0.5978654623,0.1542705446,-0.1013850942,-0.0176489484,-0.6275677085,-0.0870429948,0.2369992137,0.1773915887,0.0069189076,-0.0594249666,0.0562490635,0.0291542392,-0.1039993316,-0.0355609097,0.3078770638,0.1508753598,-0.2659685612,0.2467293441,-0.2811996341,0.607986331,-0.0468964726,-0.1634049118,-0.2128950655,-0.2576547265,0.0195475053,0.1593132913,0.2411200702,0.0207181256,0.1986235231,0.2207705081,-0.1540320665,0.0649884045,0.0245009549,0.2503027916,0.1575832963,-0.1797471195,0.1247219741,-0.2695562541,-0.3744449019,0.2322755605,0.0075143836,0.0711591393,0.2216119766,0.0979104042,-0.1160347089,-0.3240016699,-0.0471325703,0.2471790314,0.169006452,-0.1707308739,-0.1680963635,-0.3135831952,-0.6968457103,-0.1674931645,0.2041619569,-0.0389065482,-0.2985683084,0.2482632697,0.2658659816,-0.1981299967,0.2364682406,-0.2502886653,0.2545999885,-0.2911274731,0.6735090017,-0.2491139323,-0.2653599679,-0.1449729353,0.223122254,0.246337533,0.2122549117,0.4256932735,-0.0055552479,0.1085690558,-0.2478007078,-0.3401330709,-0.0156990085,-0.0388598219,0.1713903993,-0.022624528,-0.178196758,-0.2033302784,-0.291749239,0.1904279739,0.0473685451,-0.2281743884,0.1528982967,0.2184109688,-0.2253878117,0.0005367745,-0.130324766,-0.0182784516,-0.4309303164,0.0181807056,0.1976568252,0.1923655868,-0.1018810719,-0.1841416955,0.1030850708,-0.0233471543,0.2588749826,-0.2077780813,-0.0202977862,0.179380253,-0.316467464,-0.5435106158,0.046339754,0.2157483846,0.1036941409,-0.1616990268,-0.4887571335,-0.0397516787,0.0758630708,-0.4696342349,0.2138734609,-0.3266378641,0.1919414103,0.0268415641,-0.234275341,0.079067722,-0.1753830463,0.0982681811,0.0425465889,0.324296236,-0.2362303436,-0.0010163189,0.1802991927,0.2010120451,0.4291472435,0.2170517296,0.4457309544,0.2712135315,-0.0176342651,0.1529547423,-0.2266081274,0.4317142665,0.0370702893,0.2571949065,0.3716417551,-0.1155767068,-0.3734424114,-0.2116262317,0.0092148278,-0.4784247279,-0.089666225,0.0394268893,-0.3203672767,0.2842433453,-0.0447905473,-0.1656544656,0.1834592968,-0.4417850375,0.0155467661,0.1039842516,-0.0338003933,-0.0993859991,-0.2854256034,0.0116816862,-0.2031699717,0.2864902318,-0.0550195947,0.2856150866,-0.452236861,0.2000240684,0.3530741334,-0.1079584956,0.4213023782,-0.1414791495,0.0962356478,0.04868057,0.0615285449,-0.1367511153,0.0373438671,-0.2464044392,0.4075390399,-0.0732832253,-0.136828661,-0.1689240932,0.1191413924,0.3491515219,-0.3517277539,-0.2772349715,-0.3379975259,0.0369782671,-0.3008764684,-0.1056493372,-0.1918329298,-0.062206734,0.3322469294,-0.1173475683,0.2271380574,-0.062657766,-0.082908608,0.0779117122,-0.289234668,0.0061493316,-0.1615341157,-0.0086662071,0.0253287256,0.0703222603,0.0013403178,0.3976664543,0.2108407766,-0.308467567,0.0230411999,-0.0446577221,0.1039862707,0.335654974,0.2389217466,0.2558197081,0.1286765188,0.0268322844,0.1807296723,-0.0406233631,0.2774070203,-0.1208268479,0.0317548886,-0.0400159061,0.4364601076,-0.0371190012,0.0257538129,-0.1203959435,0.2057988942,-0.0590293445,0.120770216,0.2723385692,0.926227212,0.3573255837,-0.1934802979,-0.1381358206,-0.3493309617,0.3499742746,-0.1011262313,0.2965338528,-0.323040247,0.0344858803,-0.0098209772,-0.072005406,0.2681917846,0.1104635969,-0.1519093364,0.0132251037,0.2016169578,-0.4078493714,0.0023740584,0.1255775243,-0.2024443448,0.0887537524,-0.1019634381,0.195788905,-0.3366136551,0.2888072133,-0.0241127219,-0.0143023664,0.0745375007,-0.0958946273,-0.6498867273,0.1196941212,0.1529039294,0.2965019643,0.4366156459,-0.3487069309,0.5024280548,0.3479771912,0.0978226736,-0.134469077,-0.1817012131,0.2670594454,-0.0081972564,-0.3333131075,0.1345228255,0.2164837867,0.431242764,-0.248711139,-0.1208236888,0.0887061656,0.2242548466,-0.092918925,-0.2742377818,0.0666302219,-0.1082142219,0.0566688739,-0.5281037688,-0.0680338889,-0.2862310708,-0.1474752873,0.211326018,0.0121115157,-0.1223869845,-0.0541188158,0.2479371876,0.0337132737,0.219680801,0.4317296743,0.1767745018,-0.2623867691,0.3625331223,0.0983792692,-0.3140345216,-0.2832293212,-0.1632616371,-0.1882076412,-0.5992369056,0.1116821393,0.0500462912,0.0485353582,0.0918528959,0.2327153981,0.0671167597,0.2704228163,-0.0103946272,-0.4409473538,0.152132228,0.2068588287,-0.0900095329,-0.1959535629,-0.0554312505,0.1327091604,0.3474212587,0.2478316277,-0.3705713451,0.0219979621,-0.8066151142,0.2578496933,-0.1387397051,-0.1527324617,-0.2663748264,-0.0131353196,0.1881993711,0.2927480042,-0.2539350092,-0.2508189678,-0.0855917558,0.0745598078,0.4079851806,-0.1990192384,-0.0052760099,0.1281678081,0.1731999069,0.3432820141,-0.3793751597,-0.1080289856,-0.0163916126,0.1289957017,-0.150328517,-0.1926799417,0.0309938993,0.0365639925,-0.1324367225,0.0785758719,0.2087307721,-0.0207994655,0.1791902333,-0.0710981637,-0.1945794374,-0.0077179326,-0.0070424173,0.0500332937,0.2638768256,-0.233075127,0.2598917782,0.0361592658,0.2810717225,0.0642258823,-0.2384266555,-0.1236941367,0.5562167168,0.2389519662,-0.4434122741,0.0247947704,-0.041409459,0.1561344415,-0.1421895027,-0.1950068027,0.0676419437,0.0601218082,0.0985980257,0.0732264519,0.2367685288,-0.3288160861,0.0298168696,-0.0262298323,-0.0691185966,-0.030788837,0.3207075298,0.3133562505,0.1652621031,-0.1061988622,0.113481164,0.3593474925,-0.175672397,0.3844712973,0.2292941809,-0.0982699394,0.0683204085,0.1208594665,0.1840697527,0.1911423206,-0.0612222739,0.4018625617,0.1221966073,0.0818434507,-0.2159696519,0.3743452728,-0.1083503589,-0.0966768414,-0.4809775949,0.1010922045,-0.2817732394,-0.0311669763,0.041001685,-0.1225676164,-0.1113910601,0.0851919651,-0.4288272858,-0.5297067165,-0.0370004438,-0.0540391766,0.1134243608,-0.1695820987,0.2004293203,0.1226982176,0.1507233828,0.1994429827,0.5213519335,0.3164956272,0.0817890912,0.1480291635,-0.2558347583,-0.1826345623,-0.0593755282,0.2018956244,0.2581671774,0.6251091361,0.1552933455,0.2553520501,0.1537106335,-0.2273415774,0.4738101363,0.286277771,0.2470715493,-0.1064791679,0.2302937806,-0.2687031925,-0.3250296712,-0.3542169929,-0.1551422924,-0.5146823525,0.1194320321,-0.0579605587,-0.2122407407,-0.090247862,-0.3688592911,0.1332190037,0.2878714502,0.5181463957,0.3140351474,0.0281723067,-0.1002033651,-0.3143658042,-0.3044556379,0.2352886498,0.2170859128,-0.0171915498,0.199650228,0.061818108,-0.3078075051,0.2227214724,-0.2102542669,0.2524964809,-0.0569489114,0.4409445226,-0.1339469105,0.1448565722,0.2532391846,-0.0394067429,-0.1570731401,0.1408738941,0.0112109464,-0.1642329842,0.1693493426,-0.2809847593,-0.1521471739,-0.1949172914,0.2860403359,-0.0703022555,0.0059249778,0.2876330316,-0.0813861787,0.3241367936,0.0612080395,-0.3352140188,-0.2148204297,-0.039935492,0.241878584,0.405169189,-0.1323652565,0.1548966914,0.120609723,0.0342248268,-0.4314126074,-0.0469938666,-0.1981748194,-0.2105949819,-0.05684698,-0.1373497844,0.1902591735,0.138040036,0.1421836913,-0.0857410803,-0.3737942874,-0.4902588427,0.5093217492,-0.1371855438,-0.0673301369,0.2751485109,0.2710306942,0.0795527324,-0.2651308775,-0.4097693861,-0.0491695963,0.048973646,0.0505426265,0.3461513221,0.0052484926,-0.127980262,-0.0475301109,0.0350609981,-0.1687660664,0.174179107,-0.0770711377,0.0458970182,-0.0913835987]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/811","title":"nlp viewer error","comments":"Is this the problem of my local computer or ??","body":"Hello, \r\nwhen I select amazon_us_reviews in nlp viewer, it shows error.\r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=amazon_us_reviews\r\n![image](https:\/\/user-images.githubusercontent.com\/30210529\/98447334-4aa81200-2124-11eb-9dca-82c3ab34ccc2.png)\r\n","comment_length":10,"text":"nlp viewer error\nHello, \r\nwhen I select amazon_us_reviews in nlp viewer, it shows error.\r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=amazon_us_reviews\r\n![image](https:\/\/user-images.githubusercontent.com\/30210529\/98447334-4aa81200-2124-11eb-9dca-82c3ab34ccc2.png)\r\n\nIs this the problem of my local computer or ??","embeddings":[-0.1987897456,-0.1020159721,-0.1214239001,0.2379014045,0.1104616523,0.1728497148,-0.0879265741,0.1648270637,0.0597546771,0.0210140049,0.0550654568,0.0001926277,-0.0605219863,0.0265961643,0.0084900577,-0.2845063806,-0.0601014197,0.1796880811,0.181009382,0.0123821106,-0.3255371153,0.3067645133,-0.2570572793,0.3296793997,-0.0796237737,-0.0979961976,-0.0028168862,-0.1678301543,-0.1798129082,-0.3221288323,0.2803300321,-0.198895961,-0.0145223141,-0.0477520861,-0.0001095439,0.0709594339,0.5964125395,0.0700487122,-0.015552938,-0.3259989619,0.0852964744,-0.0499395095,0.1671976,-0.1806904525,-0.0293180794,-0.4166063368,0.3244630396,0.3307714164,-0.0117645944,0.2804626226,0.336163491,0.3549226522,0.0409128368,-0.0687595829,-0.0645767823,0.1754659712,-0.0823175758,0.2274681777,-0.0209270995,-0.0615156926,0.1201094612,0.2212761939,0.1172553301,0.0368306935,-0.0498984866,-0.0105741667,0.2768278718,0.0630239919,0.2655268908,0.0351424925,0.4925955236,0.0272620823,0.1558447033,0.1835956872,0.1619307101,0.0882922858,0.1778741479,0.5837090015,-0.1361789256,0.0176679231,-0.4156633019,-0.2243994474,-0.2024909854,0.3012946844,-0.1068461761,0.0835320204,0.0052664452,0.1806150973,0.2595107853,0.0427416302,0.0563427471,0.201358825,-0.1262103766,-0.1089712381,-0.3487134576,-0.1586433351,0.0914831683,0.4056123197,-0.2505092919,0.0358989201,0.1651288718,-0.0518943109,0.0978100672,0.3349222839,0.1854756624,0.1356137097,0.0450822562,0.2484783381,-0.1168763861,0.1541486233,-0.0249354355,0.0599463657,-0.058957316,-0.4647994041,-0.3621354699,-0.3384676278,0.0055085886,-0.3176698387,-0.5032426119,0.1718368828,-0.3052618504,0.1638716757,0.1114097387,0.4048556983,-0.0812926367,0.0500819795,0.0263384338,0.1340509057,-0.2000989914,-0.1130745187,-0.176247105,0.3659964204,-0.2417262197,-0.1270311475,-0.0803848729,-0.0470477752,0.3126187921,-0.2675246298,0.2038186789,-0.174344644,-0.1175699309,-0.3327024579,-0.2133033127,0.2377537489,0.0946140662,-0.0707744732,0.0462598428,-0.1122724339,-0.0385274775,0.3303393722,-0.0365408324,-0.1480118632,-0.371555537,0.2320607007,-0.1308811903,-0.380979985,-0.0643263459,0.0747826323,0.009332058,-0.0474643596,0.0392158292,-0.0432627723,-0.1593568474,0.058424484,-0.2680964172,0.0688590854,-0.2564611435,-0.112371929,-0.4486200213,-0.3078768253,0.1722121239,0.4753860533,-0.2530342042,0.0184651855,-0.1730571091,0.336111635,0.4967111945,-0.1427901387,-0.1981010586,0.5017074943,-0.2281194776,-0.4676410556,0.0846229345,0.0419359505,-0.4079915285,-0.018417703,-0.3194664717,0.1505768746,-0.0284661893,0.0515072942,-0.4014267623,-0.3618067801,0.3987001181,0.2657704055,0.2306019217,-0.0373657718,0.0279067662,0.2587144971,0.1486828327,0.2901639938,-0.3188863099,0.2557081282,0.2582989335,0.0695057884,-0.0089620827,-0.0587961227,0.0641666204,-0.1299743056,-0.507268846,0.1895333678,0.1644989848,-0.0224869251,-0.5260702968,0.0635319576,-0.182973057,-0.0748849437,0.2354860455,0.0138595151,0.0093440395,0.1157601625,0.004547047,-0.175853014,-0.1911057383,-0.0875756443,-0.0515006483,0.368588239,-0.4001048803,0.014291103,0.28655231,-0.0160529688,0.24324812,0.0159056317,-0.0188496392,0.1614917666,-0.336027354,0.4253384471,0.2187680006,-0.1391947567,0.0879673138,-0.6081095934,0.0002598217,0.3743945956,0.2143124044,0.1222658828,-0.0927272439,-0.0789372623,0.0098781623,-0.0627007857,-0.0676109195,0.242494911,0.1212378666,-0.1361769587,0.2850449085,-0.2209814042,0.4167427123,0.1468586624,0.0681284964,-0.1107932925,-0.3936543167,-0.0960383564,0.2194964439,0.2832303941,0.0408264808,0.3336362243,0.1408521682,-0.1606905758,0.1470450312,0.096648179,0.2975856662,0.126350984,-0.2292798311,0.293068856,-0.3214799166,-0.370516628,0.186270237,0.1332468688,0.1490830332,0.1362140626,0.0916512161,-0.1203297824,-0.423612237,-0.018258214,0.205297187,0.2230519503,-0.1895302534,0.0206698477,-0.4347986877,-0.8505862951,-0.3099903464,0.352848947,-0.1197673678,-0.341393888,0.1398344487,0.1541447937,-0.1670692861,0.1376443505,-0.2371535152,0.3374137878,-0.2075006366,0.3984903395,-0.1524285376,-0.2009066045,-0.3348246515,0.1576841027,0.3777292371,0.2166478187,0.3773035109,-0.1129204854,0.0538585335,-0.1474130303,-0.2788812816,0.0399149954,0.1117383242,0.2541387081,0.0013841806,0.0532369949,-0.4220574498,-0.2520497143,0.3088472188,0.0749272183,-0.1840538681,0.0120791094,0.3091150522,-0.2018915564,-0.0372093692,-0.0629330873,-0.1501759738,-0.3181959987,0.0541252345,0.159392193,0.2265370041,-0.1881326884,-0.2324261069,0.1735100299,-0.0509771891,0.0458893143,-0.2131382823,-0.1038076058,0.1090509966,-0.2692761719,-0.5786737204,-0.0079469876,0.220523417,0.0904906318,-0.1084431037,-0.5167387128,0.0429354608,-0.0974332914,-0.4246419072,0.2871352136,-0.3087614179,0.2813307643,-0.0541824438,-0.1857566535,0.1199277863,-0.2121789157,0.1458140612,0.2402873933,0.3349895179,-0.0773806199,0.1237992197,0.2280439734,0.3252545893,0.2589250505,0.0620009936,0.684490025,0.0450397134,-0.0097268084,-0.0200206582,-0.1519299895,0.3090478778,0.0898282751,-0.0122504449,0.3250747919,-0.1668097228,-0.4392983615,-0.2875856757,0.0283335168,-0.5272568464,-0.0342942365,0.0189713165,-0.1125821099,0.2483530492,-0.0661077499,-0.1888007671,0.3282043934,-0.3642038703,-0.005782072,0.1156799421,-0.1494796425,-0.0285360329,-0.3481908143,-0.2081970125,0.0796618164,0.2059632689,-0.0352255069,0.6040620208,-0.5417113304,0.037382666,0.3268418908,-0.0764528662,0.4167796373,0.1050997525,0.0870163441,-0.0013136799,-0.0215281472,-0.1622481048,-0.0930285007,-0.1348185688,0.3604333103,-0.1441756934,-0.0367860869,-0.0124601098,0.1366301477,0.2289532125,0.0212776419,-0.423712194,-0.3373801112,-0.1268016398,-0.1766765267,-0.0225656182,-0.1477002352,-0.2320348471,0.2900252342,0.0199607275,0.3034321666,-0.1083508283,-0.1645624936,-0.1220377013,-0.2277792394,0.0909842625,-0.2639654577,-0.1050470993,0.0062851151,0.2728734016,-0.1278005093,0.4018045962,0.3409976363,-0.2408271432,0.1447532624,-0.0180084892,0.1375054717,0.3851905465,0.1479708105,0.0145754041,0.2615849078,0.0067692101,0.1313785762,-0.0004778996,0.3649685979,-0.1940362751,0.0722490698,0.0407412313,0.4579313099,-0.0568434186,-0.0051357327,-0.0720727965,0.2027864307,-0.0013460835,0.1321179867,0.2905712724,1.024662137,0.1972058415,-0.0451527089,-0.0130882757,-0.4143305123,0.3441779912,0.0327278338,0.2115590274,-0.3819638789,0.0934240073,-0.0251322147,-0.0797966272,0.411030829,0.0226190258,0.0615243278,0.1112640351,0.0416502953,-0.3797847331,0.0480320752,0.1258895844,-0.0661786497,0.2126483619,-0.1156470552,0.1774586886,-0.279098779,0.408811003,-0.1118604764,0.0210869219,-0.0180954672,-0.0644965544,-0.6476470828,0.0716705173,0.1697659791,0.2844538689,0.2961400151,-0.3109409809,0.2880986035,0.3809334338,0.0121446075,-0.0725795105,-0.2544805408,0.2662549615,-0.0683110878,-0.1916368306,0.2214734852,0.2256707102,0.3086060882,-0.3420212269,-0.214468658,-0.0627909303,0.2959940434,-0.0395863652,-0.2915011942,0.0342124626,-0.0705318153,0.0851628184,-0.7650272846,-0.2457284182,-0.1577587277,-0.0923513621,0.1850159615,-0.0444724187,-0.0607949682,-0.0781661868,0.3109746277,-0.0331458896,0.0750736073,0.6412118673,0.1276559681,-0.2676475644,0.2465200275,0.0097395862,-0.3206157684,-0.2651545703,0.1152965128,-0.240226686,-0.5567371249,-0.1463367343,-0.0564852729,0.0926174447,-0.0913720503,0.206356883,0.1869278103,0.2202462107,-0.134350419,-0.4492996931,0.0972168595,0.0927124545,-0.0630108789,-0.164819181,-0.0167700201,0.0601540469,0.5125763416,0.0230931193,-0.3180078268,0.0345034041,-0.6429222226,0.1880017519,-0.274310261,0.0392749459,-0.2183953226,-0.0224184934,0.1946029216,0.2189692855,-0.1616226882,-0.2346202731,-0.0473934524,0.1108910441,0.3962883949,-0.228739351,0.0137219923,0.0226092506,0.1171204671,0.4111250937,-0.4800537229,-0.166214481,-0.1020842791,0.2149140239,-0.0299918093,-0.3310098648,0.0191923771,0.166054979,-0.1829364598,0.2810203731,0.305077225,0.032301534,0.3952176571,-0.039992068,-0.3402510881,-0.1709104776,-0.1446912289,0.1726570725,0.1428661942,-0.3739759028,0.2626537383,0.0054652621,0.3569093049,0.2481911927,-0.2363668829,-0.2042905092,0.7604151368,0.2009469271,-0.5002407432,-0.0405428,0.198140651,0.0825175717,-0.216825664,-0.1780615598,0.1206280366,0.1814350784,0.0765762478,-0.0309945624,0.3155522645,-0.2535791099,0.1529458165,0.2210922688,-0.1612709612,0.151014924,0.0725273192,0.3176131845,0.260029912,0.2292955071,-0.0081242193,0.1430979222,-0.0733926147,0.3181270063,0.3171592951,-0.1444299668,0.0929904282,0.1548915952,0.0767355412,0.085332118,0.0740738511,0.4085163176,0.1077808514,0.0038870429,-0.1955781281,0.2853591144,-0.1960166842,0.0653418079,-0.3929649591,0.0663932115,-0.3529050648,0.0947610885,0.0087616788,-0.1361580491,-0.0543110482,0.1597988904,-0.3866367042,-0.3733399808,-0.0870994031,0.0409437567,0.0430903435,-0.2577069998,0.1253837943,0.0712988004,0.1289310902,0.1396966577,0.4657792151,0.0874912739,0.2509115338,-0.0029958657,-0.3256694973,-0.1670639515,0.0417422689,0.0048748618,0.0794549659,0.5179741383,0.3311058879,0.3957586884,0.0999637321,-0.1682367921,0.6408744454,0.1561263651,0.2326239794,-0.2395827621,0.2159273922,-0.3679936528,-0.2486103475,-0.269703567,-0.1832102686,-0.5100034475,0.1704045534,-0.0887309611,-0.0940898359,0.0007999636,-0.3563904464,0.1097224057,0.1545478702,0.7151358724,0.423214376,0.1799439788,-0.0670333952,-0.3876479268,-0.4529993832,0.2161379755,0.2137225121,0.0195346978,0.2587960362,0.0010313202,-0.4304713607,0.3969390988,-0.1878450364,0.4170365036,-0.237500459,0.4457722902,-0.0895908326,0.061357785,0.1582353264,0.0239872243,-0.1748429537,-0.0113761248,0.1676660925,-0.3066346347,0.1320980191,-0.3619790971,-0.1959368885,-0.4241220653,0.2073742896,0.064699769,0.048849266,0.2762438059,-0.0055871848,0.2081439197,0.1106789336,-0.2577343583,-0.1756824553,0.042969279,0.1696546525,0.2703640461,-0.0964027122,0.2509181201,0.0344599932,0.1058658957,-0.3758413196,-0.0164831188,-0.090595454,-0.3862689435,0.0283938367,-0.1980367154,0.3001006246,0.0936789885,0.120542191,-0.0870606601,-0.2372166663,-0.2274197042,0.4554105997,-0.1207192689,0.0153914895,0.2055585086,0.2284128815,0.1713996232,-0.3069841266,-0.1216015518,-0.1054427922,0.190554142,0.0246477574,0.3098843396,0.0809995979,-0.0285495855,-0.071891956,0.0248485915,-0.0196672883,0.0731088817,-0.0529731251,0.0493200868,-0.1683243066]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/809","title":"Add Google Taskmaster dataset","comments":"Hey @yjernite. Was going to start working on this but found taskmaster 1,2 & 3 in the datasets library already so think this can be closed now?","body":"## Adding a Dataset\r\n- **Name:** Taskmaster\r\n- **Description:** A large dataset of task-oriented dialogue with annotated goals (55K dialogues covering entertainment and travel reservations)\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1909.05358\r\n- **Data:** https:\/\/github.com\/google-research-datasets\/Taskmaster\r\n- **Motivation:** One of few annotated datasets of this size for goal-oriented dialogue\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":27,"text":"Add Google Taskmaster dataset\n## Adding a Dataset\r\n- **Name:** Taskmaster\r\n- **Description:** A large dataset of task-oriented dialogue with annotated goals (55K dialogues covering entertainment and travel reservations)\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1909.05358\r\n- **Data:** https:\/\/github.com\/google-research-datasets\/Taskmaster\r\n- **Motivation:** One of few annotated datasets of this size for goal-oriented dialogue\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nHey @yjernite. Was going to start working on this but found taskmaster 1,2 & 3 in the datasets library already so think this can be closed now?","embeddings":[-0.2614889145,0.0552474037,-0.1595136374,0.2156586349,0.1674283445,-0.0379510708,0.3336213231,0.0670085549,0.132318899,0.1240423247,-0.2129537016,0.2662527859,-0.48644045,0.5283167362,0.1895260811,-0.0497275516,0.0831428096,-0.087164484,0.1714651287,0.1660844684,-0.1717765331,0.0900838003,-0.1450598985,0.023761902,-0.1984927356,-0.2001335025,0.027907839,-0.0120169437,-0.1924094111,-0.3131717443,-0.030132398,0.3213685751,-0.0966233164,0.7043562531,-0.0001007273,-0.3078626096,0.1542027891,-0.1997465938,-0.3623100519,-0.0376223251,-0.1035003513,0.0397048853,-0.1769136488,-0.0062243836,-0.2134982497,-0.1862116456,0.0373942629,0.0631220043,0.1247560009,0.4792369604,0.2868831754,0.0314498469,0.0237882473,-0.3139028549,-0.0892795473,0.1673174053,-0.325029403,0.1835339516,0.2450786382,0.0302086752,-0.1205400303,0.3014375567,-0.1818485111,-0.1398768127,0.205429107,-0.2605779767,-0.1797824502,-0.2628004551,0.1561324,0.3817877769,0.578802824,-0.1634546965,-0.129009068,-0.4247144163,-0.0288697425,-0.1403267533,-0.1643438786,0.2142509669,-0.1045960262,0.1321996301,-0.2390735745,-0.3919066191,-0.0509699248,0.0245202649,0.0453483388,0.1810941994,-0.2719650269,-0.1165957153,0.1177231744,-0.2911083996,-0.1091829017,0.0246927124,-0.1397085935,-0.0661734268,-0.5199308991,-0.3846782446,0.3971691728,0.248669222,0.3383380771,0.1867453456,0.1201629192,-0.0119864037,0.0094055673,-0.061529085,0.4166882336,0.1291999519,0.0020518445,-0.0619577542,0.1798332483,0.2364335507,0.1209973842,0.2560901642,-0.172796011,-0.0051418608,-0.0672673434,-0.0164523423,-0.0275292192,0.1008455828,-0.3065927327,0.2058406472,-0.1520988345,0.0444887653,0.1388910264,0.2740323842,-0.2231860757,0.0721610114,-0.0989489108,0.1962263733,0.0454923175,-0.0089532947,-0.0678549334,0.1156092584,-0.28828758,0.3044955134,0.1619835049,0.224510029,0.0406731032,-0.1672491878,-0.0642289221,-0.094134897,0.1624991298,-0.1709158123,-0.0306092966,-0.1118817329,0.0123427762,-0.3497440219,-0.2084514499,0.0089686243,-0.1720944494,-0.0218165051,0.184971571,-0.2141790539,-0.1163008586,0.2941468954,-0.159095183,-0.2060000747,-0.3234015107,0.7503315806,-0.0713446587,-0.0585385524,0.2222454846,0.2418702245,-0.3724497855,-0.1413063854,-0.0511270277,0.3489097655,0.0621128753,-0.0115338294,-0.1397532374,-0.1462148875,-0.4228977561,0.3107075989,0.0636393279,0.4232715368,-0.1566381156,0.0577303506,-0.0616485551,-0.2638650537,-0.2786205709,0.1703373045,-0.5045643449,0.0620137304,0.2612501681,0.1390815079,0.4737354517,-0.4069030583,0.5480496287,0.235035345,-0.1516464949,0.02583923,-0.1084815115,-0.3002912104,0.0096795065,0.3743725121,0.1809770018,-0.2353051156,-0.1113969535,-0.1721189767,0.0077166096,-0.1511417925,0.0177359395,0.1571633518,0.3781970143,0.1448082924,-0.0746482313,-0.35207811,-0.4401951134,-0.2714777887,0.0580170304,0.1237808242,0.2761168778,-0.4302568138,-0.163715601,-0.0106257489,-0.0130670462,-0.146705851,0.2628631294,0.1287089437,0.0623064376,-0.1343763471,-0.1763136238,0.5124223828,-0.1791616082,-0.0888966843,0.1272492558,0.0461914353,0.0050449008,-0.052316092,0.2747652829,0.4075893462,-0.0309089236,0.1027138978,0.2287934124,0.1635100693,-0.0730133355,0.2508262694,0.2525398135,0.4002345502,-0.0040863007,-0.2029565871,-0.0149969924,0.2142480314,0.0235825907,-0.0960364863,-0.2542673647,0.268188864,0.2542662621,0.0910768807,0.0821855739,0.1735062152,0.1134506762,-0.209338516,-0.1832897663,0.0297998786,0.1873932481,0.213800773,-0.0630107,0.1144268438,-0.691962719,0.1650878191,0.3536649942,-0.0179386362,0.4091228545,0.183029905,-0.1275765747,-0.3061305285,-0.0035227491,-0.3534864485,0.104459025,0.4246437848,0.0630821511,-0.0029152632,0.2378720194,-0.1665383875,0.1324298829,0.0892466307,0.0002087655,0.2678321302,0.2840036154,-0.0470410027,-0.2962804437,-0.2191294581,-0.0053095068,0.0315587185,-0.1969404519,-0.199775815,0.1438695788,-0.2817783356,-0.0144887408,-0.0204758402,-0.0104614664,-0.0066536441,0.2491914332,0.1484094262,-0.0625197962,0.1598895639,0.2966145277,0.5640652776,-0.2380785495,0.0970344841,-0.1008222625,-0.5919599533,0.003471192,0.2379782796,0.1573915333,-0.2912105918,0.5987552404,-0.0476847626,0.1234845892,-0.5213546157,-0.4303777218,0.2126313001,-0.3136650622,0.2960092425,0.1633021832,0.066306673,-0.0380839072,0.0635015443,-0.2227798551,0.1621677428,-0.06548962,-0.2262498736,-0.1913326383,0.0895263776,-0.0155229745,-0.314037919,-0.079275474,-0.6464050412,0.5182401538,-0.1646692306,0.1542172432,0.3452112377,-0.1528852731,0.1042081341,-0.1129702479,0.0673101693,-0.3176529706,0.1281550825,0.1276369393,-0.2890620232,-0.1419183612,-0.2176164985,-0.2709980309,0.1267265081,-0.0954340026,-0.3981805146,0.1359130889,-0.0661595017,0.095359318,-0.0531415045,-0.2207897604,0.2568977177,-0.016948536,-0.2660772502,-0.0653651655,-0.3890554309,0.1412203312,0.2797262073,0.0679741129,0.0680811852,0.0780960247,0.0350673497,0.2983905077,-0.0397373848,0.2279607505,0.1530387551,0.0089080166,0.3236825168,-0.0899442136,-0.2734912038,0.5499250889,0.2022094429,-0.0500392243,0.4199289382,0.0643064082,-0.1487182081,-0.0776420385,-0.1573595107,-0.0776632428,-0.1441313624,0.117025651,-0.1653676033,0.2310357392,-0.2109772712,-0.1682575345,-0.2618766427,-0.0410391316,-0.0105372425,0.1600434631,0.0578889921,-0.2461485863,-0.1157540753,0.336930871,-0.3425078392,0.1058330759,-0.0220841095,-0.0903136209,-0.2629334033,0.1058379784,-0.1920100152,0.036721155,0.4593257308,-0.0843126476,-0.2236302793,-0.0058133672,0.3202736378,-0.0442813709,0.30113554,0.0137200691,0.095212087,0.2549005151,0.6940997243,-0.3379404545,0.0444254242,0.4166782498,-0.0531407483,0.1003895029,-0.075676553,0.0354476869,-0.2737960815,-0.0657213852,0.4045336843,0.1714744866,0.1080613807,0.4278351963,-0.1285207719,0.1964720041,-0.0749866143,0.2083271295,-0.2814505994,0.0158280879,0.2705775499,0.1865020245,-0.1502009332,-0.0852071866,0.2100267559,0.3960526586,0.165918842,-0.1972730756,0.144133389,-0.2553134263,0.3726992309,0.2760243714,0.2314091176,0.1951169074,-0.0132763796,0.4930595458,-0.2513540387,0.106901966,0.0910186172,-0.0045255744,-0.5179160833,-0.4419856369,0.1323390305,0.1544008404,-0.0653373674,0.4278345704,-0.0870077685,-0.1938176155,-0.2065756619,0.0495647229,0.9224182367,0.0042038867,0.2119583935,0.4051855803,-0.5795304179,0.0694973022,-0.2566628456,0.1229210272,-0.0451139025,0.2835305929,0.0615990497,0.052759584,-0.1864240021,0.1910013109,-0.4328390658,0.0930375904,0.0615276583,0.1465491354,0.1344944686,0.4971313179,-0.1569572389,-0.3890841305,-0.264988333,0.2539279461,-0.1405627728,-0.0713890418,-0.0638067052,-0.1614766568,0.0310061369,-0.0530793443,-0.5341101885,-0.1280454248,-0.2671846151,-0.198919788,0.0845625699,-0.519002676,0.2743303776,-0.1211940944,0.0168517847,0.1505001336,-0.3063873053,0.1029704958,0.0461027175,0.0080742445,-0.0130423401,-0.076959908,0.447604686,-0.0306448601,-0.0582419559,0.0667864829,0.2132108063,-0.1285063326,-0.5471840501,0.09721753,0.0421941802,-0.1174715385,0.0141498744,0.1093639582,-0.1167789102,-0.2550396025,0.2261781394,0.1129067689,0.0214293003,-0.070040293,0.0169781391,-0.132522881,-0.2255842388,0.0563980155,0.137014091,-0.1952793598,0.1854117215,0.4506282508,-0.1530923545,-0.5877164006,0.0059715812,0.2940577269,0.0000028473,0.269641608,0.0914230496,-0.2956960499,0.046439752,0.3118742108,0.023787789,-0.0398308747,0.0400148071,-0.0107819401,-0.4449403286,0.2384040058,-0.3241703808,0.1604297459,-0.1441944838,-0.0213203114,0.2203150988,0.0423804931,-0.4860435128,-0.0637272447,-0.4109106958,0.0673712268,0.3359673619,-0.1816318929,-0.0566214807,0.0057620583,0.2004772425,-0.1914741397,-0.2832296789,-0.3187803328,-0.1769904494,0.0831012353,0.0346599631,0.220999375,0.0055383057,0.0053614047,-0.3474757373,-0.0585813075,0.3122031689,0.3366632164,-0.0110059846,-0.1279553324,-0.332520932,-0.0652844086,0.2533950806,-0.1236490235,0.1806525141,0.107703194,0.0346107222,-0.0085139479,0.2483477741,-0.0588270426,-0.1101661995,0.4326218665,0.4296225309,-0.0656212047,0.1655566692,0.0395488441,-0.0317774639,0.2534838021,0.1656468213,-0.0127339661,-0.0342913084,0.0483630002,0.1087709144,0.3111405671,-0.0967460796,0.2382243127,0.1319253147,-0.3640677333,0.2468385696,0.1856057793,-0.0028982533,0.2447800338,0.0875357538,0.1429648697,0.0931355953,0.0603960194,0.1990287006,-0.153215766,0.0971820503,-0.064112924,0.5701426268,0.1783618629,0.3212884665,0.0907506198,0.2958139181,0.2464682311,0.0606437325,0.1194164306,0.4031062722,-0.0200183839,-0.104433015,-0.1590248048,-0.1592031717,-0.0486527719,-0.0868871734,0.5234045386,-0.2786833346,0.1393802017,-0.001949818,0.0124193132,-0.2311098576,0.1900731921,-0.0240693334,-0.2697418928,-0.1429315209,0.0223107133,0.0209671855,-0.116647236,0.1573189795,-0.0206569508,0.0962159485,-0.2094026953,0.1180355921,0.2326318771,0.1462402344,-0.3240089715,0.0697671771,-0.2487801015,0.1303092539,0.098571904,0.5581969619,0.1926463544,0.1127704382,-0.331320405,0.0588083602,-0.2364318222,-0.1330974102,-0.0621248707,0.1594278365,0.2089761198,0.0670753643,0.2844001055,0.2394436747,-0.2812752724,0.2649773955,0.2016084939,-0.413731426,-0.3584219515,-0.0328574702,-0.0273715295,-0.2570313513,-0.2707479,-0.0777750984,-0.4774102271,-0.1650235355,0.1326689869,-0.0494088903,0.0154928649,-0.3788840473,0.1565183401,-0.1011172533,0.3919014037,0.3475658298,0.4866895378,-0.0984477773,-0.1428841501,-0.3140197992,0.1523294747,-0.0063523753,-0.1410083622,-0.0102803735,-0.0432209782,0.0756920725,0.3448818326,-0.1129556298,-0.0117366211,-0.0404542759,-0.0593646616,-0.3465914726,0.2285216749,0.3738496602,0.2461614907,0.1255486608,-0.0747673288,-0.0023522847,0.0123156691,0.1539683044,-0.2139319927,-0.0713592321,0.0829248577,-0.1694505066,0.2467193753,-0.2851561606,0.3618652523,-0.1637970805,-0.1217438281,-0.1098735034,0.1035010889,-0.1392450482,-0.1859431118,0.1150813252,0.3691797853,-0.1166030988,-0.0576775372,-0.0611319691,-0.0294606611,-0.0412592962,0.286552161,0.0700432211,0.1787757874,0.0487965196,-0.0715590417,-0.0751321092,0.1126115546,-0.0251286738,-0.0660492033,-0.0930076391,-0.3622236252,0.2840488851,-0.0295285992,-0.402916044,-0.1293289065,0.2246692181,0.0481192693,-0.1693311483,-0.5357699394,0.1112424135,0.2190642804,0.2421075404,0.0319202207,0.3021369874,-0.1388214678,-0.0703438297,-0.0344765,0.0103817796,0.0269469414,0.164232403,-0.2071315944,-0.3559950888]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/809","title":"Add Google Taskmaster dataset","comments":"You are absolutely right :) \r\n\r\nClosed by https:\/\/github.com\/huggingface\/datasets\/pull\/1193 https:\/\/github.com\/huggingface\/datasets\/pull\/1197 https:\/\/github.com\/huggingface\/datasets\/pull\/1213","body":"## Adding a Dataset\r\n- **Name:** Taskmaster\r\n- **Description:** A large dataset of task-oriented dialogue with annotated goals (55K dialogues covering entertainment and travel reservations)\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1909.05358\r\n- **Data:** https:\/\/github.com\/google-research-datasets\/Taskmaster\r\n- **Motivation:** One of few annotated datasets of this size for goal-oriented dialogue\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":10,"text":"Add Google Taskmaster dataset\n## Adding a Dataset\r\n- **Name:** Taskmaster\r\n- **Description:** A large dataset of task-oriented dialogue with annotated goals (55K dialogues covering entertainment and travel reservations)\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1909.05358\r\n- **Data:** https:\/\/github.com\/google-research-datasets\/Taskmaster\r\n- **Motivation:** One of few annotated datasets of this size for goal-oriented dialogue\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nYou are absolutely right :) \r\n\r\nClosed by https:\/\/github.com\/huggingface\/datasets\/pull\/1193 https:\/\/github.com\/huggingface\/datasets\/pull\/1197 https:\/\/github.com\/huggingface\/datasets\/pull\/1213","embeddings":[-0.1486774236,-0.1168125495,-0.1777431518,0.0818352252,0.1488179862,-0.1384498328,0.3084007204,-0.0116919475,0.088527903,0.2229492664,-0.2601377368,0.1787379831,-0.3945517242,0.5397735834,0.1601211578,-0.0852422416,0.1037803814,-0.0817156807,0.0303243287,0.1934656352,-0.1099081114,0.1228096932,-0.0549686328,0.0865845457,-0.2438826263,-0.2118401229,-0.0247966964,0.0338977836,-0.0849566162,-0.1897770017,-0.046462968,0.2223304361,-0.2096952945,0.6409406662,-0.0001015492,-0.1763068289,0.0812904462,-0.2486270666,-0.3439397514,-0.0253416803,-0.0147431511,0.1260441393,-0.1455426365,-0.0232101977,-0.3155143261,-0.037505284,0.0617971458,0.1407047957,0.3061194718,0.420212388,0.306619525,0.201226607,-0.0433079675,-0.2759847343,-0.1874778867,0.1674204767,-0.2356048822,0.2802784145,0.1328943223,0.0200553518,-0.2188149691,0.3347781897,-0.0594030581,-0.0586676598,0.1540442854,-0.1779773235,-0.165504083,-0.2582339048,0.1136142015,0.3226458728,0.3170390427,-0.1311855465,-0.164116621,-0.4935322106,-0.040846087,-0.1515084207,-0.1565878987,0.2364171147,-0.1530518681,0.129201144,-0.2551758885,-0.2297334224,-0.021648474,-0.0957880393,0.0319332592,0.067768313,-0.3118757308,-0.1467959285,0.0355801024,-0.2003589869,-0.3029810488,0.0481947921,-0.0822551847,-0.1222676039,-0.4355589449,-0.3369538188,0.3516315818,0.2763344646,0.2409502417,0.1620264947,0.2463459969,-0.023474995,0.0350546539,-0.0353260972,0.3697343767,0.1534780115,0.0137540586,-0.1326120645,0.0073753428,0.25048545,0.1126033217,0.1783397943,-0.0603913553,0.1437201798,-0.2297850698,-0.0737881958,-0.1767164022,0.2148326486,-0.3849655986,0.1854685694,-0.1303329021,0.0986747444,0.1251199543,0.3616945446,-0.3003456891,-0.0755580887,-0.0892916918,0.2209354788,-0.0018654483,-0.0377666615,-0.0987459868,0.0495547168,-0.3094791472,0.3151774108,0.180680871,0.1006628424,0.1997964978,-0.1297807395,0.0835857689,-0.1823503226,0.1058739349,-0.1866567582,0.0710909441,-0.1138284132,0.0840816349,-0.3722891808,-0.1092889085,-0.0003103285,-0.1774820387,-0.1010580286,0.114383705,-0.2002323717,0.0682482645,0.2849469483,-0.0923410505,-0.1163572297,-0.1920369267,0.7094219327,0.0132868094,-0.0044992757,0.2674138546,0.341295898,-0.3000336289,-0.0759661198,-0.0717562288,0.3681134284,0.2543711066,-0.1003336832,0.1938265562,-0.1679988205,-0.450181216,0.3534466326,0.0283656064,0.4094014764,-0.1437741965,0.1480450779,-0.1075363681,-0.25692451,-0.2005336136,0.1074079424,-0.4778175652,0.1194409132,0.3048751056,0.0241774935,0.3998761773,-0.3044526279,0.5459654927,0.1989825219,-0.0343829654,0.1438728869,-0.0477011874,-0.298364222,-0.1467670202,0.3103402555,0.087934278,-0.1249221116,-0.1355827451,-0.1270587295,0.0428939648,-0.1590363085,-0.0212086495,0.2418187857,0.2144366354,0.1783032119,-0.0266421661,-0.2673570514,-0.3087198138,-0.3272142112,-0.0622871555,0.1642393917,0.3470399082,-0.4632014632,-0.1519921273,-0.0684423968,-0.0075768526,-0.147874698,0.3303474486,0.0899449512,0.0667820945,-0.1961293072,-0.1333874017,0.4857077301,-0.1483169645,-0.0185047165,0.0146084959,-0.0100073759,-0.0849985108,-0.0121034887,0.3503021002,0.3481801748,-0.147026509,0.152389884,0.301038146,0.1874920428,-0.094749935,0.2896752656,0.2073620558,0.3775236011,-0.014939623,-0.0892076939,-0.0277482159,0.2008956969,0.0239288732,-0.03577419,-0.1778584719,0.4172142148,0.2992171049,0.0973074585,0.1945118904,0.0892364755,0.1965234727,-0.2151186913,-0.2611706555,-0.0525691435,0.2030813694,0.1002793908,0.0915323868,0.1669195592,-0.7182570696,0.1688445807,0.3693903387,0.0356472768,0.4104315639,0.1211670414,-0.2099620849,-0.4008717835,0.0142320357,-0.3904336095,0.0850127265,0.4688844085,0.1304170936,-0.0512862168,0.1730974019,-0.1987777203,0.1700057834,0.0705700964,-0.0046772719,0.2537758648,0.1842461228,-0.0547375046,-0.3267530501,-0.1410115957,-0.150959745,-0.0336169787,-0.284991771,-0.104645744,0.2345139086,-0.3308471143,0.0088138627,-0.0035753059,-0.0527707823,-0.0246671699,0.3183798194,0.0771244094,-0.1215085462,0.061925333,0.3813651502,0.5323468447,-0.1680963337,0.0914322063,-0.0786927789,-0.5320340991,0.0732183456,0.2628412843,-0.0265609547,-0.1758841723,0.5278176665,-0.1748012751,0.1134656891,-0.3513397574,-0.3424438238,0.3987934887,-0.3070379794,0.2723332942,0.1617754102,-0.0006497441,-0.030705547,0.0948018208,-0.1921155006,0.1183491647,-0.1701486856,-0.2493833154,-0.1407118291,0.1336362362,-0.1370781958,-0.1705452204,-0.1028497368,-0.5850685239,0.6043180227,-0.0754799396,0.1288394779,0.3910015821,-0.0707401782,0.1081150398,-0.2968107462,0.2221711278,-0.4029000401,0.0835501701,0.0553298444,-0.3446361423,-0.1533868164,-0.4101108909,-0.278051883,0.1154039726,-0.0591136515,-0.3420913815,-0.0969380885,-0.0793490037,0.0112507148,-0.0455871075,-0.1465541869,0.2573902905,-0.0915693864,-0.3233367205,-0.1000387073,-0.2372240722,0.2600184083,0.2086611688,-0.0328066871,0.0551895276,0.020663403,0.0617133379,0.3350654542,0.0524671674,0.2418738455,0.2151213288,-0.1000076532,0.3840008676,-0.1371093094,-0.0998773128,0.5427916646,0.1825716197,-0.0450695045,0.492020309,0.1138087511,-0.0145657761,-0.0852388293,-0.241003722,-0.2665622532,-0.2111562937,0.1684267521,-0.0401576236,0.2590797842,-0.1590491682,-0.1023636833,-0.1809122115,-0.029006768,0.0230320618,0.1419869512,0.1317794323,-0.2534529865,-0.0658974648,0.4191565812,-0.4013429284,0.0769411027,-0.0152436038,-0.1855010539,-0.226987794,-0.1193098426,-0.1556486487,-0.0350843742,0.3388222456,0.1566068828,-0.1888478994,-0.2679095268,0.0893727243,-0.1012669504,0.3929765821,-0.122931987,0.0218506046,0.3440587223,0.627763629,-0.4671883881,-0.0080291023,0.2531717718,-0.0529459603,0.0062386887,-0.0657287836,-0.047491625,-0.2791265249,-0.1629523486,0.3414847553,0.1435141712,0.0337793529,0.3522765636,-0.0415278673,0.1871512532,-0.0529967248,0.2060221136,-0.1675686687,0.1054372489,0.1065675095,0.0418695547,-0.1654476374,0.1196858138,0.1103754491,0.4710234404,0.1673512161,-0.100708805,0.0898650289,-0.2686274946,0.2431518584,0.3165464401,0.1706565768,0.3012593091,0.0372199155,0.4596886933,-0.2736476362,0.0776704401,0.1015099585,-0.0454941541,-0.3199837208,-0.3121949732,0.1338554323,0.0748197287,-0.0434500873,0.2626242042,-0.0093426863,-0.1865013987,-0.1830162406,0.1334325522,0.9006747007,0.1377498806,0.4408392906,0.5497931242,-0.5019984841,0.1525714248,-0.1046704799,0.1050334126,-0.0604266711,0.2277275026,0.0512151234,0.0949567333,-0.2204239815,0.1721685827,-0.348733604,0.0867383406,0.1284541786,0.1169781461,0.0136159537,0.307867229,-0.3134974837,-0.2988511622,-0.3081389666,0.2354575545,-0.0686220974,0.0336984657,-0.0025279371,-0.1604079008,-0.0980638787,-0.1056988463,-0.3598782122,-0.2356658876,-0.313241601,-0.3187948167,0.1590685248,-0.4310519397,0.101844199,-0.1267764568,0.0189493317,0.1471029073,-0.3365393579,0.0490631238,-0.0092229042,0.1035993248,0.0009101587,0.063634865,0.49278,-0.0205701422,-0.0237489156,-0.102331765,0.1586466432,-0.1075182483,-0.48279351,0.1368319839,-0.1784220636,-0.1276743263,-0.032500349,0.1614907086,-0.2083560973,-0.1707879454,0.2283608764,0.0233398881,0.0590933599,-0.0793033466,0.0765039921,-0.1601316333,-0.3129166663,0.0766487122,0.0076527316,-0.2411463708,0.2017475665,0.2397690564,-0.2392582446,-0.5495161414,-0.1317816079,0.3220489919,-0.1358320862,0.1799326092,0.1316051632,-0.3111355305,0.0742909387,0.33510378,0.138957113,0.0838917196,-0.0805449039,0.0554404221,-0.4396031499,0.3921708167,-0.1954184324,0.1529154927,-0.055879917,0.0372926295,0.2945009172,0.0196497254,-0.5043628216,-0.1964242011,-0.4029707313,0.1316829026,0.3855597973,-0.1867576689,0.0181519669,0.0724847838,0.2196518034,-0.1464715302,-0.3305475116,-0.3692562282,-0.2027831227,0.0592645928,0.0081160525,0.1901622415,0.0038463136,0.170506537,-0.339466244,-0.1975737214,0.2917498052,0.3176445663,0.0045947987,-0.1763946265,-0.3349295855,0.0292260628,0.2054787278,-0.1817026734,0.1102117449,0.1219891682,0.0432396419,-0.0571928211,0.0622255839,-0.0126638832,-0.0013795139,0.4550516307,0.4050164521,-0.017371593,0.1306354851,0.0719519407,-0.0607240573,0.1290985197,0.3532839119,-0.0447659083,0.0044879843,-0.0107078925,0.1775710583,0.364623338,-0.1066929847,0.2285068184,0.0884890035,-0.4381400645,0.3215654492,0.3542021215,-0.0063129305,0.2068440616,0.0809525251,0.1004090458,0.0498997122,0.1131373793,0.0536954068,-0.1876080781,0.1256555021,0.0352015123,0.5047416687,0.0901188999,0.3820851147,-0.0125207892,0.2182122767,0.3050793409,-0.0515845455,0.1242791936,0.3796379566,0.1069767773,-0.1325060576,-0.17325598,-0.1401889473,0.0656683594,-0.0863613561,0.5531579852,-0.3594873548,0.1761234403,-0.0252715554,0.0155380601,-0.2075194865,0.1156167239,-0.0101812063,-0.2340443581,-0.1374648809,0.0419467427,0.1073163897,-0.0765556991,0.1472314596,-0.0752702802,0.0879727975,-0.2707345784,-0.0300583746,0.1473641396,0.1989302486,-0.2406490445,0.1577951312,-0.1911440194,0.1453449726,0.1179863289,0.6047141552,0.1764122099,0.0879123881,-0.1439388543,-0.0396089852,-0.1588793397,-0.0621509477,-0.2436464727,0.1019179448,0.3501388431,0.0200589858,0.3121349812,0.2432205379,-0.340798974,0.2294017971,0.1568113267,-0.3278697133,-0.2404827178,0.0163803902,-0.0790226087,-0.2129143178,-0.1953279674,-0.1297970116,-0.6484265327,-0.0315082632,-0.0284983739,-0.0709746033,-0.0143255983,-0.3570800424,0.1412190795,-0.0385862887,0.4797766805,0.3477046192,0.4837758839,-0.0562714003,-0.1044960618,-0.2803616822,0.2316094339,0.0504151322,-0.0583593287,-0.101103954,-0.0200972445,0.0637486503,0.3419718444,-0.1128386557,-0.0520363897,-0.0463994369,-0.1110120416,-0.3171482682,0.2340447903,0.2192608714,0.2617943287,0.2061155289,-0.0260098912,-0.0730681419,-0.0093469499,0.1999869645,-0.1962021589,0.0170157943,-0.0147135453,-0.2330646962,0.1620523483,-0.2522595525,0.2681183219,-0.2589634359,-0.2337541878,0.0178143419,0.1263518482,-0.2000928819,-0.2345462292,0.1345463395,0.4337698221,-0.0351724885,-0.223355487,0.0367753468,-0.0116765602,0.1672279835,0.1652598977,0.1013478115,0.1490751952,0.0692816749,-0.1898072809,-0.0878707618,0.3368713558,0.0202232916,-0.0608740784,-0.0430856906,-0.3414903283,0.2654541433,-0.059778858,-0.2779369652,-0.100176096,0.1740525812,0.2172857225,-0.2961661518,-0.555601418,0.1488337666,0.1625119001,0.2117244601,0.1097287759,0.2616046071,-0.2263737768,-0.1538455635,-0.0800893009,0.0599352606,0.0743968189,0.0676851794,-0.2104592323,-0.3643310666]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"Hi !\r\nThe url works on my side.\r\n\r\nIs the url working in your navigator ?\r\nAre you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":30,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\nHi !\r\nThe url works on my side.\r\n\r\nIs the url working in your navigator ?\r\nAre you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"> Hi !\r\n> The url works on my side.\r\n> \r\n> Is the url working in your navigator ?\r\n> Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n\r\nI tried another server, it's working now. Thanks a lot.\r\n\r\nAnd I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":69,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\n> Hi !\r\n> The url works on my side.\r\n> \r\n> Is the url working in your navigator ?\r\n> Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n\r\nI tried another server, it's working now. Thanks a lot.\r\n\r\nAnd I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"It seems my network frequently crashed so most time it cannot work.","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":12,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\nIt seems my network frequently crashed so most time it cannot work.","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"\r\n\r\n\r\n> > Hi !\r\n> > The url works on my side.\r\n> > Is the url working in your navigator ?\r\n> > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> \r\n> I tried another server, it's working now. Thanks a lot.\r\n> \r\n> And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n\r\nI download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f \r\n\r\nThanks :D","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":103,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\n\r\n\r\n\r\n> > Hi !\r\n> > The url works on my side.\r\n> > Is the url working in your navigator ?\r\n> > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> \r\n> I tried another server, it's working now. Thanks a lot.\r\n> \r\n> And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n\r\nI download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f \r\n\r\nThanks :D","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"hello, how did you solve this problems?\r\n\r\n> > > Hi !\r\n> > > The url works on my side.\r\n> > > Is the url working in your navigator ?\r\n> > > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> > \r\n> > \r\n> > I tried another server, it's working now. Thanks a lot.\r\n> > And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n> \r\n> I download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f\r\n> \r\n> Thanks :D\r\n\r\nhello, I tried this. but it still failed. how do you fix this error?","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":136,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\nhello, how did you solve this problems?\r\n\r\n> > > Hi !\r\n> > > The url works on my side.\r\n> > > Is the url working in your navigator ?\r\n> > > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> > \r\n> > \r\n> > I tried another server, it's working now. Thanks a lot.\r\n> > And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n> \r\n> I download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f\r\n> \r\n> Thanks :D\r\n\r\nhello, I tried this. but it still failed. how do you fix this error?","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"> hello, how did you solve this problems?\r\n> \r\n> > > > Hi !\r\n> > > > The url works on my side.\r\n> > > > Is the url working in your navigator ?\r\n> > > > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> > > \r\n> > > \r\n> > > I tried another server, it's working now. Thanks a lot.\r\n> > > And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n> > \r\n> > \r\n> > I download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f\r\n> > Thanks :D\r\n> \r\n> hello, I tried this. but it still failed. how do you fix this error?\r\n\r\n\u4f60\u628a\u90a3\u4e2a\u811a\u672c\u4e0b\u8f7d\u5230\u4f60\u672c\u5730\u5b89\u88c5\u76ee\u5f55\u4e0b\uff0c\u7136\u540e `load_dataset(csv_script_path, data_fiels)`\r\n\r\n","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":155,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\n> hello, how did you solve this problems?\r\n> \r\n> > > > Hi !\r\n> > > > The url works on my side.\r\n> > > > Is the url working in your navigator ?\r\n> > > > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> > > \r\n> > > \r\n> > > I tried another server, it's working now. Thanks a lot.\r\n> > > And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n> > \r\n> > \r\n> > I download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f\r\n> > Thanks :D\r\n> \r\n> hello, I tried this. but it still failed. how do you fix this error?\r\n\r\n\u4f60\u628a\u90a3\u4e2a\u811a\u672c\u4e0b\u8f7d\u5230\u4f60\u672c\u5730\u5b89\u88c5\u76ee\u5f55\u4e0b\uff0c\u7136\u540e `load_dataset(csv_script_path, data_fiels)`\r\n\r\n","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"> > hello, how did you solve this problems?\r\n> > > > > Hi !\r\n> > > > > The url works on my side.\r\n> > > > > Is the url working in your navigator ?\r\n> > > > > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> > > > \r\n> > > > \r\n> > > > I tried another server, it's working now. Thanks a lot.\r\n> > > > And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n> > > \r\n> > > \r\n> > > I download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f\r\n> > > Thanks :D\r\n> > \r\n> > \r\n> > hello, I tried this. but it still failed. how do you fix this error?\r\n> \r\n> \u4f60\u628a\u90a3\u4e2a\u811a\u672c\u4e0b\u8f7d\u5230\u4f60\u672c\u5730\u5b89\u88c5\u76ee\u5f55\u4e0b\uff0c\u7136\u540e `load_dataset(csv_script_path, data_fiels)`\r\n\r\n\u597d\u7684\u597d\u7684\uff01\u89e3\u51b3\u4e86\uff0c\u611f\u8c22\u611f\u8c22\uff01\uff01\uff01","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":174,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\n> > hello, how did you solve this problems?\r\n> > > > > Hi !\r\n> > > > > The url works on my side.\r\n> > > > > Is the url working in your navigator ?\r\n> > > > > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> > > > \r\n> > > > \r\n> > > > I tried another server, it's working now. Thanks a lot.\r\n> > > > And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n> > > \r\n> > > \r\n> > > I download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f\r\n> > > Thanks :D\r\n> > \r\n> > \r\n> > hello, I tried this. but it still failed. how do you fix this error?\r\n> \r\n> \u4f60\u628a\u90a3\u4e2a\u811a\u672c\u4e0b\u8f7d\u5230\u4f60\u672c\u5730\u5b89\u88c5\u76ee\u5f55\u4e0b\uff0c\u7136\u540e `load_dataset(csv_script_path, data_fiels)`\r\n\r\n\u597d\u7684\u597d\u7684\uff01\u89e3\u51b3\u4e86\uff0c\u611f\u8c22\u611f\u8c22\uff01\uff01\uff01","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"> \r\n> \r\n> > hello, how did you solve this problems?\r\n> > > > > Hi !\r\n> > > > > The url works on my side.\r\n> > > > > Is the url working in your navigator ?\r\n> > > > > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> > > > \r\n> > > > \r\n> > > > I tried another server, it's working now. Thanks a lot.\r\n> > > > And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n> > > \r\n> > > \r\n> > > I download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f\r\n> > > Thanks :D\r\n> > \r\n> > \r\n> > hello, I tried this. but it still failed. how do you fix this error?\r\n> \r\n> \u4f60\u628a\u90a3\u4e2a\u811a\u672c\u4e0b\u8f7d\u5230\u4f60\u672c\u5730\u5b89\u88c5\u76ee\u5f55\u4e0b\uff0c\u7136\u540e `load_dataset(csv_script_path, data_fiels)`\r\n\r\n\u6211\u7167\u7740\u505a\u4e86\uff0c\u7136\u540e\u62a5\u9519\u3002\r\nValueError: unable to parse C:\/Software\/Anaconda\/envs\/ptk_gpu2\/Lib\/site-packages\/datasets\\dataset_infos.json as a URL or as a local path\r\n\r\n`---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-5-fd2106a3f053> in <module>\r\n----> 1 dataset = load_dataset('C:\/Software\/Anaconda\/envs\/ptk_gpu2\/Lib\/site-packages\/datasets\/csv.py', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\nC:\\Software\\Anaconda\\envs\\ptk_gpu2\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\nC:\\Software\\Anaconda\\envs\\ptk_gpu2\\lib\\site-packages\\datasets\\load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 296 local_dataset_infos_path = cached_path(\r\n 297 dataset_infos,\r\n--> 298 download_config=download_config,\r\n 299 )\r\n 300 except (FileNotFoundError, ConnectionError):\r\n\r\nC:\\Software\\Anaconda\\envs\\ptk_gpu2\\lib\\site-packages\\datasets\\utils\\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 316 else:\r\n 317 # Something unknown\r\n--> 318 raise ValueError(\"unable to parse {} as a URL or as a local path\".format(url_or_filename))\r\n 319 \r\n 320 if download_config.extract_compressed_file and output_path is not None:\r\n\r\nValueError: unable to parse C:\/Software\/Anaconda\/envs\/ptk_gpu2\/Lib\/site-packages\/datasets\\dataset_infos.json as a URL or as a local path\r\n\r\n`","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":316,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\n> \r\n> \r\n> > hello, how did you solve this problems?\r\n> > > > > Hi !\r\n> > > > > The url works on my side.\r\n> > > > > Is the url working in your navigator ?\r\n> > > > > Are you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?\r\n> > > > \r\n> > > > \r\n> > > > I tried another server, it's working now. Thanks a lot.\r\n> > > > And I'm curious about why download things from \"github\" when I load dataset from local files ? Dose datasets work if my network crashed?\r\n> > > \r\n> > > \r\n> > > I download the scripts `https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py` and move it to the package dir `*\/datasets\/` solved the problem. Could you please put the file `datasets\/datasets\/csv\/csv.py` to `datasets\/src\/datasets\/`\uff1f\r\n> > > Thanks :D\r\n> > \r\n> > \r\n> > hello, I tried this. but it still failed. how do you fix this error?\r\n> \r\n> \u4f60\u628a\u90a3\u4e2a\u811a\u672c\u4e0b\u8f7d\u5230\u4f60\u672c\u5730\u5b89\u88c5\u76ee\u5f55\u4e0b\uff0c\u7136\u540e `load_dataset(csv_script_path, data_fiels)`\r\n\r\n\u6211\u7167\u7740\u505a\u4e86\uff0c\u7136\u540e\u62a5\u9519\u3002\r\nValueError: unable to parse C:\/Software\/Anaconda\/envs\/ptk_gpu2\/Lib\/site-packages\/datasets\\dataset_infos.json as a URL or as a local path\r\n\r\n`---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-5-fd2106a3f053> in <module>\r\n----> 1 dataset = load_dataset('C:\/Software\/Anaconda\/envs\/ptk_gpu2\/Lib\/site-packages\/datasets\/csv.py', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\nC:\\Software\\Anaconda\\envs\\ptk_gpu2\\lib\\site-packages\\datasets\\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\nC:\\Software\\Anaconda\\envs\\ptk_gpu2\\lib\\site-packages\\datasets\\load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 296 local_dataset_infos_path = cached_path(\r\n 297 dataset_infos,\r\n--> 298 download_config=download_config,\r\n 299 )\r\n 300 except (FileNotFoundError, ConnectionError):\r\n\r\nC:\\Software\\Anaconda\\envs\\ptk_gpu2\\lib\\site-packages\\datasets\\utils\\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 316 else:\r\n 317 # Something unknown\r\n--> 318 raise ValueError(\"unable to parse {} as a URL or as a local path\".format(url_or_filename))\r\n 319 \r\n 320 if download_config.extract_compressed_file and output_path is not None:\r\n\r\nValueError: unable to parse C:\/Software\/Anaconda\/envs\/ptk_gpu2\/Lib\/site-packages\/datasets\\dataset_infos.json as a URL or as a local path\r\n\r\n`","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"I also experienced this issue this morning. Looks like something specific to windows.\r\nI'm working on a fix","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":18,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\nI also experienced this issue this morning. Looks like something specific to windows.\r\nI'm working on a fix","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"I opened a PR @wn1652400018","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":5,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\nI opened a PR @wn1652400018","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/807","title":"load_dataset for LOCAL CSV files report CONNECTION ERROR","comments":"> \r\n> \r\n> I opened a PR @wn1652400018\r\n\r\nThanks you!, It works very well.","body":"## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```","comment_length":14,"text":"load_dataset for LOCAL CSV files report CONNECTION ERROR\n## load_dataset for LOCAL CSV files report CONNECTION ERROR\r\n- **Description:** \r\nA local demo csv file:\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom datasets import load_dataset\r\nimport torch\r\nimport transformers\r\n\r\ndf = pd.DataFrame(np.arange(1200).reshape(300,4))\r\ndf.to_csv('test.csv', header=False, index=False)\r\n\r\nprint('datasets version: ', datasets.__version__)\r\nprint('pytorch version: ', torch.__version__)\r\nprint('transformers version: ', transformers.__version__)\r\n\r\n# output:\r\ndatasets version: 1.1.2\r\npytorch version: 1.5.0\r\ntransformers version: 3.2.0\r\n```\r\n\r\nwhen I load data through `dataset`:\r\n```\r\ndataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n```\r\nError infos:\r\n```\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-17-bbdadb9a0c78> in <module>\r\n----> 1 dataset = load_dataset('csv', data_files='.\/test.csv', delimiter=',', autogenerate_column_names=False)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 588 # Download\/copy dataset processing script\r\n 589 module_path, hash = prepare_module(\r\n--> 590 path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n 591 )\r\n 592 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs)\r\n 266 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version)\r\n 267 try:\r\n--> 268 local_path = cached_path(file_path, download_config=download_config)\r\n 269 except FileNotFoundError:\r\n 270 if script_version is not None:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 306 user_agent=download_config.user_agent,\r\n 307 local_files_only=download_config.local_files_only,\r\n--> 308 use_etag=download_config.use_etag,\r\n 309 )\r\n 310 elif os.path.exists(url_or_filename):\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\r\n```\r\n\r\nAnd I try to connect to the site with requests:\r\n```\r\nimport requests\r\n\r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n```\r\n\r\nSimilarly Error occurs:\r\n```\r\n---------------------------------------------------------------------------\r\nConnectionRefusedError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 159 conn = connection.create_connection(\r\n--> 160 (self._dns_host, self.port), self.timeout, **extra_kw\r\n 161 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 83 if err is not None:\r\n---> 84 raise err\r\n 85 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py in create_connection(address, timeout, source_address, socket_options)\r\n 73 sock.bind(source_address)\r\n---> 74 sock.connect(sa)\r\n 75 return sock\r\n\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nNewConnectionError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 676 headers=headers,\r\n--> 677 chunked=chunked,\r\n 678 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _make_request(self, conn, method, url, timeout, chunked, **httplib_request_kw)\r\n 380 try:\r\n--> 381 self._validate_conn(conn)\r\n 382 except (SocketTimeout, BaseSSLError) as e:\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in _validate_conn(self, conn)\r\n 975 if not getattr(conn, \"sock\", None): # AppEngine might not have `.sock`\r\n--> 976 conn.connect()\r\n 977 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in connect(self)\r\n 307 # Add certificate verification\r\n--> 308 conn = self._new_conn()\r\n 309 hostname = self.host\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connection.py in _new_conn(self)\r\n 171 raise NewConnectionError(\r\n--> 172 self, \"Failed to establish a new connection: %s\" % e\r\n 173 )\r\n\r\nNewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nMaxRetryError Traceback (most recent call last)\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 448 retries=self.max_retries,\r\n--> 449 timeout=timeout\r\n 450 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py in urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, **response_kw)\r\n 724 retries = retries.increment(\r\n--> 725 method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n 726 )\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py in increment(self, method, url, response, error, _pool, _stacktrace)\r\n 438 if new_retry.is_exhausted():\r\n--> 439 raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\n 440 \r\n\r\nMaxRetryError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-20-18cc3eb4a049> in <module>\r\n 1 import requests\r\n 2 \r\n----> 3 requests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py\")\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in head(url, **kwargs)\r\n 102 \r\n 103 kwargs.setdefault('allow_redirects', False)\r\n--> 104 return request('head', url, **kwargs)\r\n 105 \r\n 106 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/api.py in request(method, url, **kwargs)\r\n 59 # cases, and look like a memory leak in others.\r\n 60 with sessions.Session() as session:\r\n---> 61 return session.request(method=method, url=url, **kwargs)\r\n 62 \r\n 63 \r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\r\n 528 }\r\n 529 send_kwargs.update(settings)\r\n--> 530 resp = self.send(prep, **send_kwargs)\r\n 531 \r\n 532 return resp\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/sessions.py in send(self, request, **kwargs)\r\n 641 \r\n 642 # Send the request\r\n--> 643 r = adapter.send(request, **kwargs)\r\n 644 \r\n 645 # Total elapsed time of the request (approximately)\r\n\r\n~\/.conda\/envs\/py36\/lib\/python3.6\/site-packages\/requests\/adapters.py in send(self, request, stream, timeout, verify, cert, proxies)\r\n 514 raise SSLError(e, request=request)\r\n 515 \r\n--> 516 raise ConnectionError(e, request=request)\r\n 517 \r\n 518 except ClosedPoolError as e:\r\n\r\nConnectionError: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: \/huggingface\/datasets\/1.1.2\/datasets\/csv\/csv.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f3cceda5e48>: Failed to establish a new connection: [Errno 111] Connection refused',))\r\n```\n> \r\n> \r\n> I opened a PR @wn1652400018\r\n\r\nThanks you!, It works very well.","embeddings":[-0.324712038,0.0161048565,-0.1088049561,0.0260043461,0.2420431376,0.0198685937,0.7173692584,0.3756367266,0.2509218454,0.1337042451,-0.0342321098,0.1497044712,0.0797994062,0.0211837217,-0.1269313991,-0.1772823036,-0.1271574497,0.2743321061,-0.4187993109,0.0144969616,-0.1248850673,-0.0023575695,-0.0349380858,0.1405339241,-0.2024638206,-0.0062384652,0.1157821938,0.1312774718,-0.0276923925,-0.493919909,0.3365218639,0.0402496904,0.323772788,0.3632176518,-0.0001060309,0.2135237306,0.2517406344,-0.0288667344,-0.4279088378,-0.4611105323,-0.1089922264,-0.3690283597,0.4228135347,-0.3986177444,0.0005737995,-0.1198977008,-0.0673584118,-0.3161996007,0.4313887954,0.4785009623,0.2786398232,0.4154756367,0.0165828858,-0.1110344306,0.0718218759,0.0555340908,-0.0624923073,0.4061276317,0.0516244806,-0.1849907339,-0.242678538,0.0846120864,-0.1926984191,0.3955322206,0.1906039566,0.0830963105,0.0377710871,-0.1966335922,0.1252265424,0.107683368,0.4764897823,-0.2314554602,-0.2321467549,-0.0584997945,-0.1994815022,-0.1881569773,0.1660740525,0.0867301002,-0.1683380306,0.1474405825,-0.2638413012,0.0990802348,-0.3907521665,0.1494552046,-0.1105123758,0.1547572166,-0.067184031,0.1947448403,0.158264637,-0.056483563,0.4036888182,-0.2070344985,0.1916099489,0.1276624054,-0.3377639651,0.1311576068,0.1394052505,-0.3861347735,-0.1064571813,0.0498938598,-0.0361735485,-0.0535907373,0.1335505396,0.2916464508,0.1983619481,0.0791434944,-0.1308648437,0.2947734892,0.2687248588,0.1325622052,0.042261485,-0.0624847747,-0.0800571069,-0.3558065295,0.1634977609,0.1999797225,0.4894587696,-0.0474283807,-0.2816003859,0.1326408088,0.0129102021,-0.1073727086,0.0189691093,0.3589312434,-0.0879683271,0.2043156773,0.0424528793,0.0136886993,-0.2235278487,0.0845431536,-0.2120842189,0.0379552618,-0.1193305477,-0.0695323423,0.3813319206,-0.0790898874,0.0466334634,-0.0146686006,0.0989310592,0.0503491461,0.2611702383,-0.2932826877,0.0491831452,0.4129322171,0.0561541021,0.2478840053,0.4095256925,-0.1842343062,-0.1060605422,0.0865907371,-0.2045428902,-0.3371692896,-0.3385475874,0.2402119339,0.0368901119,0.1089414582,-0.0184665043,-0.3302738965,0.0986591801,-0.2747161686,-0.1541592777,-0.3237557411,-0.0389051214,-0.2159913927,0.3081872761,0.6463362575,-0.4303185344,0.0429565385,0.0566363595,-0.1299165338,0.1414598227,0.2508479655,-0.3076960742,0.0851259753,-0.2284607738,-0.0966767818,0.2314821184,-0.3093754947,-0.3032815158,0.5005193353,-0.2373457253,-0.072943598,0.2169772685,0.0126159824,0.291056484,0.1858281493,0.2615151703,0.3000478446,-0.1007597595,0.1175532788,-0.0799232721,-0.222560361,0.1973579973,0.1164559647,0.0098163048,0.2231837362,0.3249672055,-0.2579794228,0.1741140336,-0.0768722668,0.1447919756,0.1345079839,0.1739715636,0.0419842489,-0.0606005527,-0.021664409,-0.508518219,0.1381557435,0.1310921013,0.0903006941,-0.4376563728,-0.0162693374,-0.3771235943,0.0051721488,-0.2671605647,-0.030241888,0.2100175917,0.3415310383,-0.0191593226,-0.0523174517,-0.0488567986,0.3256911337,-0.1166130304,-0.0192557331,-0.4156301022,0.3395552039,-0.3544683456,-0.0690031797,0.052354306,-0.0314197689,0.1626497358,-0.1825555712,-0.2719682157,0.356123805,-0.0955856889,0.2182246149,0.0170497447,-0.0044389288,0.0526204556,-0.2450993061,0.0323775932,0.0311139729,0.1533319503,0.0431826673,-0.261313647,0.3525983989,-0.0953487977,0.163369596,0.0506540611,0.0704118088,0.4514212012,0.0119942082,-0.1140478626,-0.1085091978,0.1054950804,0.2389213145,0.2262360007,0.0491035245,-0.1802719235,-0.1886238307,0.2018595934,-0.1532273442,0.0116079831,0.0923530832,-0.1086221784,-0.0442817397,0.214684695,0.2319191694,0.5149133801,0.1026022732,-0.1526015699,-0.0183210652,-0.0503232926,-0.1338230819,0.1102777719,-0.0386922434,0.0448774025,0.3272635043,0.0759821385,0.1069803238,-0.2410078943,-0.3630864322,-0.0419490486,0.2638888955,-0.3493005633,0.2479265183,-0.1645680964,0.0182821918,-0.2659018636,-0.0503300428,-0.1105000004,-0.178074345,-0.3695805371,0.3607570827,-0.0763827115,0.2372517288,-0.279001534,-0.0516414121,0.3879840374,-0.4498327672,-0.0730597377,-0.2081662267,-0.1601461768,0.0924806148,0.0935094729,-0.1568920612,0.303164959,-0.0989919826,0.0205262117,-0.16106233,-0.0238344092,-0.0821166858,0.0125281876,0.1283056736,0.1878541559,0.4475226104,0.1206022948,-0.5035608411,0.3184547424,-0.0834972411,0.0481938235,0.1964564323,-0.0537022725,-0.2032553554,0.0006876013,-0.6345347762,-0.5991461873,-0.469863981,0.1375343353,0.0665089488,0.2132315189,0.295524776,0.2269022167,0.2099236399,0.0866580531,0.0757485032,-0.0596129522,-0.1591841877,0.3499985933,-0.2349603325,-0.3570283055,-0.1133962646,-0.1152546331,0.1219444349,0.1364867091,-0.444775492,-0.0236161184,-0.1746839583,0.3565398753,-0.0467238724,0.0517050326,0.4662779868,0.1686569452,-0.1706291884,0.1126605943,-0.244187057,-0.1408112943,0.1942563504,0.0843700692,0.0044104136,0.4593695402,-0.158959344,0.653288424,-0.0798089057,-0.1897096932,0.2884309888,-0.1362253278,0.342966646,-0.1845075488,-0.6308505535,-0.0917780697,-0.3062294424,-0.2018836737,0.0107533596,-0.2297288328,-0.2305867374,-0.3641992807,0.0186996888,-0.3049333394,-0.3137823343,0.1363256425,-0.1196979135,0.3635419607,0.0468819551,-0.0130359363,-0.2529374659,-0.0600578748,-0.248627305,0.3659865558,-0.0062223538,-0.0931663066,-0.2668786943,0.1069135219,-0.2810814381,0.2432359606,0.08004985,0.5622868538,-0.2027084678,-0.1867957264,-0.0377930477,0.058157701,0.4028224051,0.0884445012,0.0874638855,0.256387651,-0.0515910201,-0.312060982,-0.0862115249,-0.0886839032,0.2211953551,0.2276988029,0.4607937634,0.059115015,-0.126977399,0.3169068098,0.2359134257,-0.0872560814,-0.1790539473,-0.283875376,-0.0508114174,-0.3215522468,-0.1053736582,-0.1118415967,0.1790318936,-0.1732904315,-0.1281275451,0.0737234429,-0.1710194945,0.0011048433,0.1193199083,-0.1456994712,0.0959774107,0.2179179043,0.1906979978,0.262887001,0.2021609247,0.4429315627,-0.2094504088,-0.5551645756,0.1772612929,-0.0145549346,0.1078310609,0.1052381322,-0.1369356513,-0.1487584263,0.1752744764,0.1808559,-0.1539573073,0.1938527226,0.2498667836,-0.1166543588,-0.3529865742,-0.4148065448,0.1314441711,-0.2005926818,-0.0967517495,0.195102796,-0.3385471404,-0.0635157302,0.259523809,-0.1583108604,0.6695193052,-0.1268935204,0.1809173077,0.1628531516,-0.1345138997,0.4483827055,-0.1574918926,0.1192620471,-0.3441147804,-0.2360038906,-0.0994971171,-0.2498575151,0.3218544722,0.009868511,-0.0969610214,0.2364213765,-0.0614171065,0.2813484669,0.0652795583,0.2203813642,0.061279159,-0.230269298,-0.1830087751,0.1901203841,-0.0931362361,0.4134555757,0.0299711712,-0.1137412786,-0.0124852862,-0.3615665734,-0.0431393459,0.0593215972,-0.2830664814,0.4136624932,0.0488292351,-0.3536547124,-0.0551416725,0.1381697208,-0.1533754468,0.078329809,-0.1498511434,-0.0794565529,-0.1080912128,-0.0611959435,-0.1169011444,-0.0029088703,0.1473994702,0.0824840888,-0.2685492337,0.2755724192,-0.1178311408,-0.3403720558,0.182836175,0.2773401737,0.2622925043,-0.5065495968,-0.1399478465,-0.3102684021,-0.1075605899,-0.1293204278,0.2005319595,-0.1123933569,0.0256724488,0.0055950279,-0.0878886059,-0.310552001,-0.0419633985,0.601483345,-0.0718142614,-0.0663499683,0.327154398,0.0624912158,-0.1843172908,-0.2428676784,0.1990402192,-0.0044321637,-0.4677977264,-0.0930079147,0.0687799975,0.0586960427,-0.1138907298,0.0717758015,0.2206759155,0.1031339318,0.0357833989,-0.4803100824,-0.5924569368,-0.0901230797,-0.0120879766,0.1493128836,0.4447919726,0.2230695486,0.0557788797,0.1090672836,-0.3599543571,-0.067561321,-0.1895690709,0.1576154083,0.5117642879,-0.2647098005,0.1770654321,-0.1552710533,0.2409113348,-0.1429233551,-0.126287058,-0.3133459687,0.0018988226,0.0618451461,0.0202442314,-0.3433455229,-0.3376550972,-0.2684760094,-0.0148096764,-0.0740636587,-0.00034111,0.2130772471,-0.1187313497,0.2121608406,-0.2453927547,0.0810147896,-0.134008348,0.1350554675,-0.2799666524,0.2402763814,0.0270824414,0.282739222,-0.071864225,-0.0194420535,-0.4871887565,-0.0540029034,0.0109267998,0.0986874476,0.3070954382,-0.180137828,-0.1373213828,0.0053289649,0.3991233408,0.2806532383,-0.2605662048,0.0164483134,0.1206053644,0.3077172041,-0.2169092447,-0.112417683,0.1596361846,0.0255127009,0.2056991458,-0.0233398471,0.0817393437,-0.194660157,0.2731689811,0.0354823582,0.4447309375,-0.0881810635,0.2025933713,0.6118386984,0.0197391156,-0.0029399211,0.1617922485,0.1720683426,0.1941984296,0.4270905852,-0.1792921275,-0.0874075443,-0.0603713617,0.1699082404,-0.1782527715,-0.4733141959,0.1432808638,-0.1597068757,-0.2988284528,-0.0670229122,0.0496748798,0.2803224623,-0.2598395646,0.074986957,-0.1056554988,0.1608567089,-0.0922896937,-0.1408545077,0.0897708088,-0.0730973333,-0.1290258914,-0.0425563492,0.1132715121,0.0838872045,0.277579993,0.2630503476,-0.0881881118,-0.2006155103,0.1970882565,0.2208368033,-0.0557488687,-0.1138677299,0.1557369083,0.3445366025,0.1739408821,-0.1311293989,0.1761197746,0.631509304,0.3470619619,0.0025090496,0.2030489296,-0.1124481931,-0.1424657702,-0.0864446312,0.0251590088,-0.0924748555,0.2070282847,0.3563943505,0.3213755488,-0.1901991218,0.3339128792,-0.0463945791,-0.1453464031,-0.1938874871,0.3707821965,0.0753017589,-0.2841669023,-0.1254215688,0.0150403026,-0.402141571,0.0693432093,0.4656215906,0.0106836986,0.2924721539,0.0018875439,0.1094438508,-0.0905618295,0.4269729257,0.1445115358,0.3763491213,-0.4452260137,0.0112880319,-0.5027255416,0.1829465777,-0.039179761,0.0534199737,-0.0442362353,0.1701687276,0.0105173448,0.2027028352,-0.14749901,-0.0566704385,0.0178523082,0.1256342381,0.0269402266,-0.1631550789,-0.2020689696,0.0583003312,-0.0021004321,-0.3904865086,0.2174809128,0.0210836846,0.1405798793,-0.2631103992,-0.063724272,0.1954773217,0.1273897439,0.3907392919,0.1600240171,0.6441075802,-0.1056483835,0.0829369128,-0.0683100522,-0.2543075085,0.0151442094,0.1953172088,0.2678668201,0.2144363374,-0.0889868662,-0.3141361475,-0.4714734852,0.3997180462,-0.0541135259,-0.0857382044,-0.0444330052,-0.1015130058,-0.0295408033,0.0428922437,0.0381943546,-0.0330935158,0.011734766,0.2108486891,-0.1760046184,-0.4096626043,0.4183122814,-0.264362067,-0.250485599,-0.1710999161,0.278178215,-0.0463287495,0.2893785238,-0.2773315012,0.2958230078,0.4524029195,-0.2110679597,-0.2147776037,0.2234395742,0.0190695953,0.2123969346,-0.0396618769,0.4695009291,0.1008590683,-0.2427970618,-0.1551304758,-0.2195901573]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/806","title":"Quail dataset urls are out of date","comments":"Hi ! Thanks for reporting.\r\nWe should fix the urls and use quail 1.3.\r\nIf you want to contribute feel free to fix the urls and open a PR :) ","body":"<h3>Code<\/h3>\r\n\r\n```\r\nfrom datasets import load_dataset\r\nquail = load_dataset('quail')\r\n```\r\n\r\n<h3>Error<\/h3>\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/text-machine-lab\/quail\/master\/quail_v1.2\/xml\/ordered\/quail_1.2_train.xml\r\n```\r\n\r\n\r\nAs per [quail v1.3 commit](https:\/\/github.com\/text-machine-lab\/quail\/commit\/506501cfa34d9ec6c042d31026ba6fea6bcec8ff) it looks like the location and suggested ordering has changed. In [https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58) the quail v1.2 datasets are being pointed to, which don't exist anymore.","comment_length":30,"text":"Quail dataset urls are out of date\n<h3>Code<\/h3>\r\n\r\n```\r\nfrom datasets import load_dataset\r\nquail = load_dataset('quail')\r\n```\r\n\r\n<h3>Error<\/h3>\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/text-machine-lab\/quail\/master\/quail_v1.2\/xml\/ordered\/quail_1.2_train.xml\r\n```\r\n\r\n\r\nAs per [quail v1.3 commit](https:\/\/github.com\/text-machine-lab\/quail\/commit\/506501cfa34d9ec6c042d31026ba6fea6bcec8ff) it looks like the location and suggested ordering has changed. In [https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58) the quail v1.2 datasets are being pointed to, which don't exist anymore.\nHi ! Thanks for reporting.\r\nWe should fix the urls and use quail 1.3.\r\nIf you want to contribute feel free to fix the urls and open a PR :) ","embeddings":[0.1114190444,0.2604739666,-0.0615226775,0.0869145319,0.0552914701,0.1165681556,-0.1039384753,0.2843024433,-0.1458044499,-0.0269254986,-0.0500934944,-0.1096021086,0.0617608912,0.0529492572,0.0366902463,0.0992366895,0.0957510546,-0.1073763967,-0.4846735895,-0.1207500622,-0.2975239754,0.3327788711,-0.1629837453,0.076441057,-0.3429378867,0.136003539,-0.2239770144,0.2833048999,-0.2790636718,-0.4246107936,0.2327839285,0.026277829,0.2541755438,0.4004095793,-0.0001059034,-0.0236908998,0.2455452532,0.0927897096,-0.1204445884,-0.129503876,-0.3578025997,-0.1204459667,0.0305389483,-0.0633287355,-0.0723091885,-0.0293328855,-0.0699053034,-0.121835269,0.2148514092,0.4633688629,0.2901187539,0.0463867038,0.0949821472,-0.2083571106,0.2387119532,-0.3106752932,-0.064139843,0.3128821254,0.1509216726,0.0517900139,0.2624897957,0.4207628071,0.0320330672,-0.217201516,0.1415033787,0.0620824508,0.1603154242,-0.2267244309,-0.1460566819,0.0553170592,0.2769092917,-0.1575435698,-0.5258876681,-0.0571358465,0.092865333,-0.5297967196,0.164857924,-0.0738296136,0.2400810421,0.2522781193,-0.1442538947,-0.0496900193,-0.0288186669,0.2217445225,0.0636483431,0.1838532537,-0.0920953602,-0.1047559232,-0.0881478116,-0.0482810028,0.2248369753,0.1113290936,-0.0977199599,0.0737553164,-0.2285220474,-0.0461671762,0.0385255218,0.1469060481,-0.0280780178,0.2144791037,0.0104714381,-0.12327189,0.0661567152,-0.173008129,0.1355981231,0.3790611029,0.1448471099,-0.3184601068,0.1518921405,0.4618674517,-0.1842029989,-0.0007187212,0.0595940426,-0.5326232314,-0.287617594,0.0397352874,0.2821303606,-0.3005323708,-0.0254514571,-0.0513354763,-0.2908760011,-0.0297414735,0.019801477,0.2492847443,-0.2216188461,0.1071917713,0.0233608577,0.1979822665,-0.0525340401,-0.2585784495,-0.2975825369,0.0306347068,-0.0917572156,0.0891402811,-0.2212432623,-0.1822705418,0.2391889989,0.0033655411,0.2054988742,-0.0999543667,-0.1102365926,0.1654513329,-0.1795851439,0.1693439782,0.0045816698,0.2769616842,0.2565434277,-0.0907028392,-0.1499936581,-0.0387393571,-0.5298555493,-0.4499754906,0.0872474536,0.2952367663,-0.1409645826,-0.2585271001,-0.0653968155,-0.038510479,-0.1615040451,-0.0347939953,0.0278654508,-0.2688954473,-0.1562509537,-0.1804731786,0.0838652104,0.2679862678,0.0515943058,0.0013741155,-0.0037766609,0.0785496533,-0.0651074499,0.1783442199,-0.1268854737,-0.0905198306,-0.2397110015,-0.1852680594,0.3440431058,-0.4733642042,-0.1089774296,-0.0084120063,-0.1969914883,-0.0493408106,-0.2570844293,0.0429476947,-0.1960624158,-0.1369914263,0.1692588925,0.0898358747,0.2092838287,-0.1185128838,-0.3340954781,-0.4222731292,-0.3132748902,0.039876692,-0.0930715874,0.1723689139,0.0642236024,-0.1694266945,0.2052768469,-0.0010919468,-0.0983717367,0.7106881142,0.4781780243,0.3526978195,-0.0494658798,-0.0677527711,-0.1924735457,0.3061388135,-0.0722022131,0.2881502211,0.1848900318,-0.1361595541,-0.1398892701,-0.0058653932,-0.2005144507,-0.2296420932,0.169844538,-0.1380506307,0.2528872192,0.3219493926,-0.1977851987,0.2368819863,-0.1812588125,0.2085904479,-0.2242492437,0.2781150937,-0.0638112649,0.0624587052,0.2114021778,0.2979295552,0.2014215589,0.0791990384,0.0119290724,0.3588083088,-0.0972498134,0.1709809452,0.331030488,0.3559348881,0.0828729793,-0.2367120683,-0.1930480003,0.089020215,-0.2096580565,0.290805608,0.1785555333,0.4742392898,-0.1554953009,0.129475981,0.2113695592,0.1168810129,0.0017113158,-0.0978079885,0.0756443515,-0.3420624733,0.159460023,0.2844244838,0.2542977333,0.0259374883,-0.1137982905,0.3069618642,0.403326422,0.0230391752,-0.0536017939,0.0925995335,-0.1973938346,0.0209067073,0.0125498576,0.4818214178,0.0471954681,0.2399191558,0.0522000976,0.1905371249,-0.025710823,-0.3361629248,0.2026152313,0.0240997244,0.2085719407,-0.1078710109,0.2298175544,0.0393148661,-0.3651339412,-0.0157523397,-0.0204770491,0.1602081954,-0.213016361,-0.107608445,-0.4700704813,-0.2990796268,-0.4573799074,-0.419302702,-0.509884119,-0.465508759,0.3162053525,-0.1503544301,-0.1422873139,0.249020949,-0.2312072664,-0.0797580332,-0.093357034,-0.0661017224,-0.00522422,0.0906276479,-0.3036705852,0.1049536467,0.0305690467,-0.0140885627,0.0448407456,-0.1542916447,0.1043544337,-0.3426955342,-0.4059768915,-0.0775824413,-0.0843533501,0.5491737723,0.101750575,0.0014882456,-0.1084427834,0.0805077478,0.3233987093,-0.0675478056,-0.1165401712,-0.054307837,-0.0253549218,0.2058884352,0.1120981127,-0.3937988579,-0.0122762527,-0.2286999673,0.0206471831,-0.0292603374,-0.0596240312,0.1615993828,0.0350462869,0.1821860671,-0.1096384823,-0.0502591878,-0.0775308982,-0.6021636724,0.3979029953,-0.1612032652,-0.2727681398,0.0399369188,0.2259169072,0.1621509492,-0.1656929702,-0.5440132022,-0.4649567008,-0.1229071021,0.1300368309,0.327532649,-0.1609509438,0.3047889769,-0.0547415614,-0.2606646717,-0.0574705377,-0.0353730209,0.0941730812,0.1546818167,0.25333336,-0.0691736266,0.6103911996,0.0846163183,0.8092582822,0.2597023845,0.3382929862,0.2470415235,-0.1527681202,0.2351429611,-0.016896069,-0.0565723591,0.0515862964,-0.1241086945,-0.0555761978,0.0963983908,0.378765583,0.6186669469,-0.0697959885,0.2015707046,-0.2937103212,-0.0371193811,-0.1755230129,0.2090346068,0.2424319237,0.0591366626,-0.1537003964,0.0078309514,-0.1437823623,0.214652136,0.2838850915,-0.1120321453,0.1823240668,-0.68541044,-0.11316856,-0.2481709421,0.4097759128,-0.0106774298,0.0811437741,0.0955095813,-0.0941074491,0.0324128866,-0.1074968874,0.3626109362,-0.0696678534,0.068116948,0.1899088919,-0.3169659972,-0.1392084658,-0.1725101918,0.0236054137,-0.2791005969,0.410992384,0.3146064579,-0.0197659694,-0.3156173527,0.0546966121,-0.0899841338,-0.1812010258,0.1167710871,-0.3629798591,-0.3824225366,-0.0023280287,-0.3053698242,-0.0203203838,0.116042085,-0.2191761881,-0.0269888993,0.1659958512,-0.129185304,-0.0373842716,-0.0013794032,0.2292927802,0.1055949032,0.0071709868,0.1888490766,-0.0121671492,0.40288046,0.6721405983,-0.1860485226,-0.590639174,-0.2913392186,0.2249020785,0.2447508425,0.3056868017,-0.0362485498,0.324711293,0.200916037,0.0997397974,-0.0396568105,0.0445981286,0.4839009345,-0.0689599738,-0.1399420649,-0.4995937943,0.6568712592,0.0978092253,0.0137364864,0.266835779,0.2461110651,-0.1783276051,-0.1669573039,0.1724602282,0.8291383386,-0.0409687199,0.2310605347,0.3960908353,-0.4848233759,0.5719251037,0.1202627495,0.1941404045,-0.3730404079,-0.0633786097,-0.0080906255,-0.0919923857,0.0135273971,0.0094932057,-0.0071857874,0.3337569535,-0.1416927576,0.0603106618,-0.1129814163,0.0118250968,-0.1428671032,0.2773576677,-0.5153910518,0.2148798257,0.1652460694,0.238272503,-0.199923113,-0.2056623548,-0.1037570164,-0.0922961459,0.0242446885,0.1859457195,-0.3815821111,0.2116108686,0.0849454477,-0.2099084556,-0.3028818071,0.4327218533,0.1628586799,-0.1900701076,-0.2635912299,0.3623173535,0.2731158435,-0.0681035817,0.1615024209,-0.0301416013,0.2988241017,-0.303568244,-0.0553888604,0.2854469717,0.0279562771,-0.5376611948,-0.2894960642,-0.0722663924,-0.3765801787,-0.1507779956,-0.1321435273,0.1279702783,0.2519581318,-0.3033635318,0.1605209559,0.1552579403,0.155630216,-0.264598757,0.0567703471,-0.2391585857,-0.171948567,0.3282781541,-0.1123785675,0.001801656,0.4858092368,0.3341837823,0.0392555408,-0.1379558444,0.0123779494,0.2258206159,-0.2358326614,0.1973347962,-0.0433264002,0.2427881062,0.2512147725,0.4615600705,0.1652021706,-0.1867063642,0.0350043587,-0.1815743744,0.009483736,0.4817069173,0.1577066481,-0.008812652,-0.1952926219,0.1823574007,-0.0750710368,-0.2257032543,-0.3950367272,0.2038487643,-0.143901974,-0.2771145701,-0.3582419753,0.0014216809,0.2592718005,0.1789799035,0.0973576605,-0.2925738096,-0.3537393808,-0.1660478413,-0.1708533317,0.0708608106,0.0334238559,-0.0888589248,0.0984105617,-0.2000894397,-0.0627467111,0.0128013799,-0.1373790503,0.3463505805,0.0812528506,0.1052270085,-0.0027638585,-0.118260555,0.1221798211,0.0165883098,0.0341541916,-0.0321345031,0.2820424139,0.0988186076,-0.0475563146,0.0455970317,0.139219597,-0.0188551415,-0.2877013981,-0.0232752822,0.2466729581,-0.3490711749,0.0197905544,0.4390015304,0.4129159451,0.3339143097,-0.233678177,-0.065942131,0.0339589901,0.2542688549,-0.2860068679,-0.0536648184,-0.3654851317,0.1874474436,-0.1943534911,-0.013109602,0.0715931058,0.0668575466,0.1099224836,0.3007769287,0.2003918737,-0.0254672859,0.3398095071,-0.1550866961,0.074948974,0.1195508614,0.3186765909,-0.0308236051,0.1462728679,0.1211034954,0.1183244213,0.2409096211,0.1301383227,0.1085388884,-0.0022804495,-0.1870730966,0.0084574716,0.2893747389,-0.14138937,-0.0005288682,-0.0831539631,0.4072683156,-0.3353337646,-0.0297645666,-0.371899873,0.1679225564,-0.0418686308,-0.1575572044,0.1652294397,-0.1897244602,0.1079778001,0.0576378442,-0.1431668103,-0.2384098172,-0.0464773066,-0.0566875674,-0.2598979771,-0.5065022111,-0.2689235508,0.172284171,0.3049836755,0.0092845373,0.1346585453,0.2707107663,-0.2390353531,0.284573853,0.4439421594,0.3326486051,0.1932355165,0.1335198879,-0.1142110527,0.0171820968,-0.0114429519,0.0906790644,0.1613948643,0.181246981,0.1034271866,0.3300352097,0.2458969802,-0.2124292552,0.1631970257,-0.2599750161,0.2762046754,-0.418376863,0.4726519287,-0.1954813898,0.0301966723,-0.1352811158,0.2122774422,-0.3365432024,0.0922621489,0.3247300088,-0.1356580704,-0.1607272476,-0.123722218,0.1383812279,0.1564608216,0.3751399815,0.6169623733,0.1007048562,-0.1028800607,-0.4628229737,-0.6122589111,0.0200038813,-0.1122749895,0.1951148808,-0.0145686083,-0.0792071,-0.0571742356,-0.0858108923,0.2085535526,-0.2471515834,-0.3349532187,-0.2094249725,-0.3358879387,0.1395668536,-0.1603657007,0.128080681,-0.0205316395,0.0371926688,0.0284698512,0.0694417506,0.0452216379,0.1195852011,0.0844078735,0.008262231,-0.3566032946,0.2902091444,0.1418778002,0.0494369045,0.1012411937,-0.2747142017,-0.3510787785,-0.6006173491,-0.1752645224,0.2137157917,-0.3993287086,0.3152039945,-0.1192762107,-0.1663545966,-0.2479841262,0.3971478343,0.0729333386,-0.3265804052,-0.198848784,0.1087142304,0.1168078855,0.0223372784,0.2624645531,0.4446199238,0.0412533768,0.1693067253,-0.1308510453,-0.2479740083,0.4088899195,-0.5502557755,0.1254830807,-0.1424403042,0.2096132785,0.3931019902,-0.0679179803,-0.4312036932,-0.1493133456,0.4588046372,0.4916499853,-0.2030235082,0.4504626691,-0.0185017083,-0.0781897008,-0.0022150294,0.4493299723,-0.0151743097,-0.323926419,0.1513320655,-0.0947426036]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/806","title":"Quail dataset urls are out of date","comments":"Done! PR [https:\/\/github.com\/huggingface\/datasets\/pull\/820](https:\/\/github.com\/huggingface\/datasets\/pull\/820)\r\n\r\nUpdated links and also regenerated the metadata and dummy data for v1.3 in order to pass verifications as described here: [https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html#adding-tests-and-metadata-to-the-dataset](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html#adding-tests-and-metadata-to-the-dataset). ","body":"<h3>Code<\/h3>\r\n\r\n```\r\nfrom datasets import load_dataset\r\nquail = load_dataset('quail')\r\n```\r\n\r\n<h3>Error<\/h3>\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/text-machine-lab\/quail\/master\/quail_v1.2\/xml\/ordered\/quail_1.2_train.xml\r\n```\r\n\r\n\r\nAs per [quail v1.3 commit](https:\/\/github.com\/text-machine-lab\/quail\/commit\/506501cfa34d9ec6c042d31026ba6fea6bcec8ff) it looks like the location and suggested ordering has changed. In [https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58) the quail v1.2 datasets are being pointed to, which don't exist anymore.","comment_length":24,"text":"Quail dataset urls are out of date\n<h3>Code<\/h3>\r\n\r\n```\r\nfrom datasets import load_dataset\r\nquail = load_dataset('quail')\r\n```\r\n\r\n<h3>Error<\/h3>\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/text-machine-lab\/quail\/master\/quail_v1.2\/xml\/ordered\/quail_1.2_train.xml\r\n```\r\n\r\n\r\nAs per [quail v1.3 commit](https:\/\/github.com\/text-machine-lab\/quail\/commit\/506501cfa34d9ec6c042d31026ba6fea6bcec8ff) it looks like the location and suggested ordering has changed. In [https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58) the quail v1.2 datasets are being pointed to, which don't exist anymore.\nDone! PR [https:\/\/github.com\/huggingface\/datasets\/pull\/820](https:\/\/github.com\/huggingface\/datasets\/pull\/820)\r\n\r\nUpdated links and also regenerated the metadata and dummy data for v1.3 in order to pass verifications as described here: [https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html#adding-tests-and-metadata-to-the-dataset](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html#adding-tests-and-metadata-to-the-dataset). ","embeddings":[0.1814386547,0.2633910477,-0.0307960361,0.0592453666,0.0418911017,0.123002179,-0.1239677742,0.2763623297,-0.1936781257,-0.0339004695,-0.0796771422,-0.0646383688,0.0496295094,0.0543348417,0.0547172092,0.1128324419,0.0695692897,-0.1595680565,-0.4316189587,-0.108742781,-0.1913979053,0.2826838195,-0.1356844902,-0.0009308354,-0.3394687772,0.1508263946,-0.2323251665,0.3646855652,-0.33223176,-0.4802744687,0.2540858388,0.0976130366,0.1320038438,0.4148575962,-0.0001096011,-0.032976944,0.2606880367,0.0720774084,-0.3000728786,-0.2538606822,-0.3303827047,-0.0534781553,0.0642531738,0.0051971148,-0.1115038767,0.0087263426,-0.1095507964,-0.1580018699,0.2263865769,0.4945200384,0.2406266332,0.1653037816,0.1251366138,-0.2052194178,0.1658790708,-0.2156875432,-0.0230033901,0.4426383078,0.1209790781,0.1489258558,0.1896133274,0.4562257826,0.0909186006,-0.1802686304,0.1911189258,-0.0242547467,0.0599368885,-0.1483265162,-0.1277471185,0.0273970161,0.2083010674,-0.1990526468,-0.5881258249,-0.1760208905,0.0165027212,-0.4174698293,0.1901533604,-0.0808428749,0.1551347673,0.2480660975,-0.2717453539,-0.0539406575,-0.0046209027,0.17169258,0.0638782084,0.1718876511,-0.1085073724,-0.0988065526,0.0124231363,-0.0263068974,0.1257371753,0.0749671459,-0.1273311079,0.0849021599,-0.1319328845,-0.035978999,0.1138864383,0.1978076696,0.1343403906,0.2857433558,-0.038501583,-0.1474226117,0.0258244965,-0.1968909651,0.0618473925,0.3826511502,0.1576802582,-0.2702056468,0.0801089779,0.5797711015,-0.15998891,-0.0159528591,0.1051403359,-0.4348548949,-0.3297359347,0.0925482512,0.2865137756,-0.3187016249,-0.0632401109,0.0108160209,-0.3502397537,-0.0768397525,0.0638162866,0.2661098838,-0.1927780658,0.0190294627,0.0102394121,0.178371489,-0.0718514547,-0.2222543359,-0.2589594126,0.093697153,-0.0632134378,0.1031194702,-0.0797800869,-0.2712112367,0.3014896512,0.0433486141,0.2006927133,-0.198533237,-0.0881214514,0.1705405861,-0.0275040474,0.1518930048,-0.0345388837,0.2503409982,0.2580696344,-0.1316429228,-0.1506108195,-0.1688246578,-0.4042610228,-0.4505808651,0.1156923696,0.2677429914,-0.2115439177,-0.230075866,-0.0241174772,-0.0581057742,-0.2264197767,-0.0449959375,0.059538886,-0.0966746658,-0.1148602739,-0.1025136486,0.1910263896,0.4267811477,0.0651886463,-0.026120089,0.0257128887,0.0380406678,-0.2048698515,0.1064180955,-0.0868134499,-0.0654594749,-0.3349542618,-0.2091577053,0.197440207,-0.5608186722,-0.0942026377,0.048544582,-0.1807417423,0.0936727449,-0.1627358198,-0.1018608809,-0.2574153244,-0.1729729176,0.0969286188,0.0059844474,0.1815550476,-0.1735564321,-0.3533761501,-0.3301094472,-0.3004466593,0.016160652,-0.0834332332,0.1149524301,0.0764567852,-0.1423191428,0.2333596498,-0.075267978,-0.0558656082,0.6734699607,0.4449435771,0.3045356274,-0.121087946,-0.0760172904,-0.2076108605,0.3075023293,-0.0486543067,0.3698556125,0.0529563092,-0.2197634727,-0.2623218894,-0.0322502181,-0.1872368455,-0.3644112647,0.1521928161,0.0017652938,0.3293638825,0.3688015044,-0.130861178,0.2453601211,-0.092699863,0.2844386995,-0.3363263309,0.2050417215,-0.0753416643,0.1298635155,0.1803190559,0.3316286504,0.1685502678,-0.0470929407,0.0621582121,0.4803514183,0.0241101645,0.2260791957,0.2993496954,0.3887963891,0.140816614,-0.1528508663,-0.2659077942,0.0504740886,-0.1981072128,0.2852492332,0.0042884829,0.5125676394,-0.2272345871,0.0437138677,0.3117083013,0.0802916586,-0.0253980309,-0.1287056953,-0.0630539507,-0.339102447,0.1343839467,0.1905379891,0.255597949,0.069410488,-0.0865645111,0.2544049025,0.3082250059,0.0007834662,-0.1199911013,0.0439022109,-0.1633097082,-0.0302391313,0.0362190194,0.4918468297,0.046304021,0.1890508682,0.1436035335,0.2670886517,-0.1065996215,-0.3105117083,0.2072129399,-0.0025301275,0.1240766421,-0.0794553459,0.1605026126,0.0790012628,-0.3271408975,-0.0104960641,0.0667402521,0.1396940351,-0.3313041925,-0.0737123266,-0.4184131026,-0.3264892995,-0.4596149623,-0.4717081487,-0.5129228234,-0.449932605,0.2855216265,-0.0565387122,-0.161754325,0.2672257721,-0.2155947089,-0.0531651303,-0.0735489354,-0.1525691152,-0.0285651647,0.1614376307,-0.2533707023,0.080409795,0.0947549716,0.0047854227,0.0835869014,-0.1914671659,0.0742064491,-0.3723446727,-0.4635862112,-0.0564412661,-0.1408981532,0.5550060868,0.1813007891,-0.0215270817,-0.1195011884,0.0860508084,0.2670639753,-0.0964375362,-0.1470891833,-0.0960969254,0.0041292692,0.2165307552,-0.0191949401,-0.3963406384,0.092317991,-0.1952121258,0.1519493759,0.0069098896,-0.1381198764,0.2225102484,0.0484431908,0.2165599167,-0.1757255495,-0.0094174715,-0.1496471316,-0.5822513103,0.2848044336,-0.1767052114,-0.2390970588,0.0257048644,0.1659633666,0.2770978212,-0.1885909736,-0.4924695194,-0.5709946752,-0.064425312,0.1746460199,0.2372259051,-0.1355831325,0.3587382436,-0.0854976103,-0.2194231898,-0.0985168442,-0.0537883379,0.061486911,-0.0216504391,0.2505014241,-0.1885724068,0.4891974628,0.1769682169,0.8431837559,0.2600773275,0.2609488368,0.236229226,-0.196809426,0.3244535923,0.047212813,-0.0810236782,0.0727616176,-0.1016949862,-0.0037913718,0.0429978035,0.3066650331,0.7679783702,-0.0914558396,0.1827277839,-0.3247429132,-0.1310320497,-0.1817190498,0.0356880128,0.3020510674,0.1169660166,-0.0425446145,0.0171983168,-0.1026650369,0.2317436039,0.3316006064,-0.0135305477,0.1732324511,-0.6399108171,-0.0365072265,-0.3518028557,0.37970981,0.0645191371,0.139647752,0.055626411,-0.1233761981,-0.0507147796,-0.1160815433,0.4085843563,-0.0691760853,0.1059043482,0.1350666881,-0.3020004332,-0.2461178899,-0.0885841995,-0.0272302348,-0.3619995415,0.45751068,0.4446379542,-0.1656659842,-0.3098143339,0.0562441237,-0.1417027116,-0.0532336198,0.2052564472,-0.3486960828,-0.443046838,-0.0546428673,-0.2301966548,-0.0071737613,0.1470791996,-0.2678348124,0.0868049189,0.1913797557,-0.067724064,0.0639547035,0.0318150185,0.2438021749,0.1417733878,0.0198607817,0.2720101774,0.0153753245,0.4065671563,0.6914363503,-0.1354546994,-0.4697496593,-0.3509455919,0.2945723236,0.2590725422,0.3263111413,-0.047658626,0.3299767673,0.1912799627,0.0777612552,-0.2093716115,0.1048560962,0.4534516335,-0.0409369618,-0.0853037685,-0.3975668252,0.5780667663,0.0859423727,0.0281032939,0.2531371117,0.3308798373,-0.2443995029,-0.1855503917,0.220032692,0.8425858021,0.027327992,0.2498832345,0.4652484357,-0.4865772724,0.6284450293,0.1138913408,0.1167908758,-0.3304570913,-0.0622638986,-0.0057840124,-0.1061193496,0.0082025854,-0.0639033094,-0.1377879381,0.2658024728,-0.0602786057,0.0859859362,-0.0679703876,0.1801139712,-0.2645602822,0.2733731568,-0.4418807924,0.2128523886,0.2451349497,0.297265619,-0.2224360257,-0.2004102916,-0.1561022699,-0.1980498582,-0.0180062987,0.1851615459,-0.3557877839,0.1390531957,0.1724858731,-0.1021339297,-0.2756612301,0.3033572733,0.2223786116,-0.0762572587,-0.2680274248,0.41688627,0.1282349229,-0.0431414731,0.1861746311,-0.0340025984,0.3404322565,-0.2331407815,-0.0630678907,0.1382152736,-0.0767260268,-0.4391551018,-0.2390266508,-0.1183928773,-0.4430551827,-0.205671221,-0.0422481969,0.069417797,0.1149649918,-0.2710414231,0.129942745,0.0479974784,0.1040008813,-0.3076749742,0.0699527413,-0.2854262292,-0.1601449251,0.4254313111,-0.3218722045,-0.0067443284,0.4813893735,0.25023821,0.1123087481,-0.1110353172,-0.03719908,0.2380972803,-0.2929171622,0.1611120403,-0.0172081254,0.2694159746,0.1723062694,0.4040856659,0.2376521379,-0.0955927595,0.0836971402,-0.206609726,0.0278003719,0.4206292629,0.1170642152,0.0566004738,-0.2326499969,0.3210024238,0.0009796299,-0.1152566969,-0.3622203469,0.2478493899,-0.1676051021,-0.1988889426,-0.2702735066,-0.057368502,0.2942203581,0.082320258,0.0673331767,-0.1960427016,-0.3853549361,-0.1269192398,-0.2115200013,0.0900653154,0.0398109369,-0.1216631681,0.0621508025,-0.0562659763,-0.0648768097,-0.0032027422,-0.1305010021,0.3166624904,0.0318104252,0.089210242,-0.013512563,-0.0400491506,0.1037707925,-0.1322861314,0.101128377,-0.0129041784,0.2341834903,0.0758784041,-0.195282042,0.0338546894,0.1749709994,0.1117260084,-0.1557743251,-0.0663064569,0.2309958041,-0.1855957508,-0.053636618,0.3081515729,0.4672926366,0.3777966499,-0.2285105884,0.0026175117,0.0609021224,0.2245687693,-0.2025837004,0.0307427142,-0.3618759215,0.1536624134,-0.0639061481,0.0053173406,0.1485762894,0.091324769,0.1864552051,0.2460995466,0.2154447138,0.0055522076,0.3046572208,-0.1607410461,-0.034944959,0.080044575,0.3350788057,-0.1089570522,0.1698316485,-0.0041717365,0.1120941788,0.2956157327,-0.0068129967,0.1529407054,0.0102919368,-0.1873321384,-0.0332226977,0.2469589263,-0.133867979,0.0300430469,-0.0903745219,0.4863058329,-0.4023507237,-0.0568594895,-0.3628669977,0.1842980385,0.036933057,-0.1936806738,0.1198690832,-0.2513340712,0.0177058149,0.0370140895,-0.0421491489,-0.2770990431,0.0040444564,-0.0828764439,-0.2485540956,-0.5583950877,-0.3507036865,0.2220550776,0.3133892715,-0.0206808001,0.196670562,0.3626984656,-0.2110574543,0.3159107268,0.4462771416,0.3888225555,0.0895378739,0.1971937567,-0.0478763692,-0.0910196006,0.0438549519,0.1091186479,0.2886358798,0.1897922158,-0.0025092079,0.418423295,0.2127280682,-0.1955218911,0.0928409472,-0.290156275,0.2178242803,-0.4855865538,0.5161002278,-0.2369038314,0.091499947,-0.182555154,0.2203063518,-0.3714239299,0.1440792084,0.260411799,-0.1047089398,-0.0537453368,-0.1721446663,0.1095535532,0.2026277632,0.4219171405,0.6534841061,0.0887694955,-0.1786471009,-0.5586004853,-0.6866915226,-0.0761817396,-0.0874831378,0.1036491692,-0.0977806672,-0.0571174882,-0.0961705074,-0.0689977333,0.0769328773,-0.2398112416,-0.2191071212,-0.2407482117,-0.3112721741,0.1616437435,-0.2100399733,0.0585403405,0.0036475842,0.0539024286,0.0561277643,0.0510280877,0.0360844284,0.1960180998,0.1249701753,-0.0187993757,-0.2710784972,0.2645252645,0.1476807147,0.129509002,0.0423919149,-0.3390609622,-0.4437514544,-0.430364579,-0.1474568397,0.2299000919,-0.3755775988,0.3490486741,-0.0557292886,-0.2273749411,-0.2384174615,0.4388746917,0.1461503655,-0.3155905306,-0.2409459502,0.1113784239,-0.0185967553,0.110448271,0.2772140801,0.5209017992,-0.0310770664,0.2682860792,-0.1945824325,-0.2983020246,0.5188693404,-0.5274541974,0.1273461133,-0.1431820989,0.244284749,0.2939406633,-0.1605311483,-0.509229064,-0.0787197649,0.4858420193,0.4363695681,-0.2180989236,0.4078384042,-0.0699497536,-0.0490429252,0.0053985626,0.505484283,0.0244077109,-0.2426117212,0.2663950026,-0.0170518085]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/806","title":"Quail dataset urls are out of date","comments":"Closing since #820 is merged.\r\nThanks again for fixing the urls :)","body":"<h3>Code<\/h3>\r\n\r\n```\r\nfrom datasets import load_dataset\r\nquail = load_dataset('quail')\r\n```\r\n\r\n<h3>Error<\/h3>\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/text-machine-lab\/quail\/master\/quail_v1.2\/xml\/ordered\/quail_1.2_train.xml\r\n```\r\n\r\n\r\nAs per [quail v1.3 commit](https:\/\/github.com\/text-machine-lab\/quail\/commit\/506501cfa34d9ec6c042d31026ba6fea6bcec8ff) it looks like the location and suggested ordering has changed. In [https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58) the quail v1.2 datasets are being pointed to, which don't exist anymore.","comment_length":12,"text":"Quail dataset urls are out of date\n<h3>Code<\/h3>\r\n\r\n```\r\nfrom datasets import load_dataset\r\nquail = load_dataset('quail')\r\n```\r\n\r\n<h3>Error<\/h3>\r\n\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/text-machine-lab\/quail\/master\/quail_v1.2\/xml\/ordered\/quail_1.2_train.xml\r\n```\r\n\r\n\r\nAs per [quail v1.3 commit](https:\/\/github.com\/text-machine-lab\/quail\/commit\/506501cfa34d9ec6c042d31026ba6fea6bcec8ff) it looks like the location and suggested ordering has changed. In [https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/quail\/quail.py#L52-L58) the quail v1.2 datasets are being pointed to, which don't exist anymore.\nClosing since #820 is merged.\r\nThanks again for fixing the urls :)","embeddings":[0.0608125404,0.2439417839,-0.0611303784,0.0762046501,0.0497849025,0.1419590563,-0.1301600039,0.3215988576,-0.1584335417,-0.0602704957,-0.1239197105,-0.126266852,0.1249212697,0.0073083355,0.0321758464,0.1735522151,0.1159169003,-0.107053563,-0.4563395679,-0.1196224019,-0.3680807054,0.3644761145,-0.1979371756,0.0517667197,-0.3658695817,0.2047368735,-0.2805333734,0.3136483133,-0.3325053751,-0.3547169566,0.1868315041,0.0869247913,0.1670171916,0.4381591082,-0.0001096554,-0.0201498941,0.2491810769,0.1029701903,-0.1611931324,-0.1531977355,-0.2958647311,-0.123489745,0.0093054809,-0.036712639,0.0142409969,-0.071164757,-0.0753808767,-0.1051382199,0.2076723278,0.4126666188,0.2421684116,0.0266619492,0.1162429005,-0.1949675232,0.1614004523,-0.2863026857,0.0216299891,0.3187121749,0.0924764201,0.0131599633,0.3983415663,0.4108308256,0.0162540302,-0.3327758014,0.2063134462,-0.0143321324,0.1244537383,-0.2153394669,-0.1038106754,0.0089181988,0.248769179,-0.1204470247,-0.4530266523,-0.0272075832,0.0275411066,-0.4989040494,0.1698428541,-0.082349509,0.2686726153,0.2721535265,-0.1274896711,-0.0724369884,0.0085143289,0.2580993772,0.01755693,0.1994552314,-0.028394416,-0.1215151474,-0.0699935481,-0.0889457017,0.1578140855,-0.0010911011,-0.1220190376,0.1636734158,-0.2282510549,-0.0258534774,0.0339261219,0.1115412414,-0.0482293703,0.2303728759,0.0332921334,-0.1684958488,0.1012689099,-0.1027768627,0.1757727414,0.38980937,0.1516541541,-0.3897966146,0.1508863419,0.4182161987,-0.2334889323,0.0165507477,0.0375992954,-0.5652372241,-0.3289867342,-0.050243523,0.3769861758,-0.2954468131,-0.0087388651,-0.064364925,-0.3402599394,-0.0740261525,0.0456425734,0.2343992293,-0.2254991382,0.1168196127,0.1217640415,0.2103755772,-0.0651686117,-0.2179455906,-0.2744155824,0.0060539395,-0.0477056988,0.0014050903,-0.2159507275,-0.1766604036,0.1966626644,0.0353782065,0.2393166274,-0.1694635004,-0.1196025237,0.1791964769,-0.1205636635,0.207317397,0.0857561827,0.2341933995,0.2682727277,-0.1473615468,-0.1152432188,-0.0237729624,-0.5673645139,-0.4897109866,0.0401016921,0.2761493027,-0.1384191364,-0.2271456122,-0.0674050525,-0.0175180305,-0.1523087323,-0.0783401951,-0.0356418043,-0.198191002,-0.1205832958,-0.1758615673,0.1244933307,0.2951903343,0.0138488021,-0.0455477983,-0.025320923,0.0941665396,-0.1437648982,0.1966792941,-0.1583739966,-0.0929597467,-0.250369817,-0.203046456,0.3976624608,-0.4594070911,-0.1004327089,-0.0187675674,-0.2415303141,-0.0419169068,-0.250043571,0.0156270657,-0.1917584091,-0.1614187062,0.2436414361,0.097727716,0.1768900901,-0.158284843,-0.3172541261,-0.3190061152,-0.2724656165,0.067408286,-0.0764214694,0.1407879293,0.0888835937,-0.1105729267,0.2500568628,0.0035294888,-0.0931316838,0.719599247,0.4827637672,0.3798237741,0.0107699512,-0.0962176546,-0.1952908188,0.2270637304,-0.1341461539,0.3028438985,0.1651276797,-0.191754669,-0.1555230916,-0.0146091264,-0.2351896316,-0.2504706979,0.1417284906,-0.1698066592,0.2027419806,0.3192597628,-0.1925409883,0.2727345228,-0.2023184448,0.2664336562,-0.2788659036,0.3281864226,-0.0583420582,0.0825451463,0.1255296022,0.3761167824,0.1994559169,0.0047314139,0.0108119063,0.4017755091,-0.133265391,0.1862432957,0.3415004313,0.3788930476,0.1713565588,-0.1930596977,-0.2212100625,0.065099977,-0.2039444894,0.1965913475,0.1522451341,0.4314886332,-0.1575832516,0.156116426,0.2484513223,0.0805658028,-0.0864963606,-0.0447745733,0.0217463598,-0.2385277301,0.1940385848,0.30359146,0.2805869877,0.0908759907,-0.0547964722,0.3073366582,0.3748028278,0.0371330753,-0.1436056495,0.0721240565,-0.2248403281,0.0015295569,-0.0906073228,0.4210731387,0.10903918,0.2281354815,0.0393054932,0.2637420297,-0.0278564766,-0.2838598788,0.2487983555,0.0683581606,0.1334620565,-0.0792915896,0.2124491781,0.0558762103,-0.3418454826,-0.03984081,-0.0052301791,0.1022577807,-0.2077612281,-0.0951981321,-0.4465119243,-0.3369686306,-0.5042153597,-0.3889101744,-0.5796838999,-0.5151513219,0.2758876979,-0.1551429927,-0.140726611,0.2846846581,-0.1882501096,-0.0208370537,-0.0587791726,-0.1437847167,0.0613437891,0.1257584393,-0.1785027534,0.0638216063,0.0523089878,-0.0672562271,0.0076515395,-0.122242637,0.126248911,-0.2840857506,-0.4763778448,-0.1255199015,-0.1290104836,0.4056300819,0.0650121942,0.0316179134,-0.072774671,0.0501813702,0.3368803263,-0.0408129841,-0.0768445507,-0.1186293289,0.0367279351,0.1373916119,0.0901592821,-0.3980770409,-0.0249775033,-0.2638337016,0.0504608303,-0.0588229448,-0.0744344965,0.1265251637,0.0213511251,0.117466256,-0.0902896896,-0.1083075479,-0.0509891212,-0.5974401236,0.3126035333,-0.1857345849,-0.219980225,0.1001303941,0.2836296558,0.1801570505,-0.1932674944,-0.4172979891,-0.4270552695,-0.1057781503,0.105941169,0.3191044033,-0.1371427476,0.2942426503,-0.0762361512,-0.2379261702,-0.0524159893,-0.0348752514,0.0750740021,0.16668652,0.2506992221,-0.093117252,0.5803831816,0.0453859046,0.8213196397,0.2451740801,0.3631568849,0.2677175105,-0.1538499296,0.2998325229,-0.0268103089,-0.1048725992,0.026006883,-0.1257091314,-0.0450347364,0.1118215919,0.45588395,0.6698787808,-0.1070732698,0.1657040715,-0.3315210044,-0.0207152404,-0.2004474103,0.2481220961,0.1774083823,0.1309639066,-0.2518577278,0.0109890336,-0.0821263716,0.2334739566,0.4102272391,-0.1594102979,0.1837031394,-0.6835808158,-0.1355096698,-0.2099058777,0.4548007846,-0.0473371893,0.0006088428,0.0879736245,-0.1613029391,-0.0146759171,-0.0685513914,0.4484163225,-0.0410067774,0.0000545048,0.1721254885,-0.2983295023,-0.0968399495,-0.2111673802,0.1289598495,-0.3056159914,0.4923567474,0.3414303958,-0.0516244657,-0.3067576587,0.1094947755,-0.1022459194,-0.1776358336,0.1113072559,-0.3838786483,-0.4259047806,-0.0646413043,-0.3577254117,0.0077701476,0.2142622769,-0.2651001215,-0.0087637808,0.1754736006,-0.2054205686,0.0565132722,-0.0810675472,0.2387697548,0.0886986256,0.0200405885,0.1780194491,0.0522353947,0.4461933076,0.7061529756,-0.1968788058,-0.6313840151,-0.2847265005,0.1729107946,0.3313187361,0.2879415452,-0.0665434226,0.3119157553,0.2300645411,0.1134865358,-0.0467267297,0.0524668954,0.4876787066,-0.0153785786,-0.1832272857,-0.4994898736,0.6665754318,0.1457444727,0.0453699715,0.3399605155,0.2386343181,-0.1857230216,-0.1668860763,0.0852377191,0.9309900403,0.0086146034,0.2677005827,0.382348001,-0.4867438376,0.532223165,0.0133220013,0.1692694128,-0.3380068839,-0.0296916943,-0.0132066896,-0.1190063134,0.0074790977,0.0140332943,-0.0517516024,0.2886382043,-0.1794355512,0.1179559827,-0.1567513049,0.0804674029,-0.1192434728,0.3257146478,-0.5649220347,0.1757887751,0.2232554406,0.2874653637,-0.2624220252,-0.2738950849,-0.143671304,-0.130887121,0.0371147916,0.1986589432,-0.2840956151,0.2184991688,0.0253535695,-0.218932882,-0.3476625383,0.432066232,0.1750815511,-0.1637108177,-0.2092903703,0.3669317067,0.1858635843,-0.0847661197,0.0496629588,-0.0797371864,0.283539325,-0.3309056163,-0.0593589172,0.2919869423,0.0032418612,-0.5397484899,-0.3052287698,0.0029814201,-0.3912092447,-0.1923713684,-0.1498516798,0.1289605051,0.1668189019,-0.3400083482,0.1212668195,0.16955553,0.1138698757,-0.2977008224,0.0710720494,-0.2806676626,-0.1798656881,0.3678746521,-0.1319399029,-0.0992200375,0.5448451042,0.3123117387,0.0595194213,-0.0758955181,-0.0090709059,0.2484535426,-0.2793519199,0.2445382625,-0.070295684,0.3006738424,0.2476569414,0.4083356559,0.1407580227,-0.0789563134,0.0605637245,-0.2162350416,0.0281336065,0.4710331261,0.1078474596,0.0343749039,-0.1814461201,0.261343658,-0.0011427574,-0.1981902868,-0.3564346135,0.2835791707,-0.139524579,-0.252509594,-0.3754749596,-0.0085755792,0.2361700535,0.1638934314,0.0580493771,-0.2439827025,-0.3358672559,-0.1308905631,-0.1562719643,0.0966843814,0.0185678173,-0.0987530872,0.0960830152,-0.1665711254,-0.0123618264,0.0733595192,-0.0876849666,0.2301725894,0.0890312642,0.0150706861,0.0628239438,-0.1438847184,0.073898375,0.0716006309,0.097011447,0.0659617335,0.284397155,0.0628188625,-0.1276236922,0.0317582823,0.1254763305,0.0011970444,-0.2486014962,-0.0619260781,0.2746948898,-0.3538470566,0.0129381539,0.3894807696,0.4585849643,0.4054380655,-0.2547419071,0.0074674101,-0.0426289663,0.2009973824,-0.270586282,-0.0332952552,-0.2660391331,0.2383152544,-0.1976889372,-0.019308662,0.0620221123,0.097435385,0.1006693169,0.3636987209,0.2432046235,0.016916113,0.4397617877,-0.1443358958,0.0954884142,0.148388356,0.2881957591,0.0447418876,0.1378100961,0.093889758,0.0620160252,0.2424452752,0.1852243394,0.0073951241,0.0628260076,-0.2279082537,0.0028432135,0.3268042207,-0.1775955558,-0.0219152104,-0.0382986106,0.3764135838,-0.3403954506,-0.0644504502,-0.3195076287,0.3117974102,-0.0679795966,-0.1204730943,0.2533704937,-0.2487552017,0.111777693,0.0405036695,-0.0447249003,-0.2511742413,-0.1120082363,-0.0748244151,-0.203724727,-0.4887378216,-0.2500604689,0.1774589717,0.3191745281,0.0200696923,0.0524764843,0.3462753892,-0.195792228,0.3558186293,0.4478347003,0.4705027938,0.2149937749,0.1381885409,-0.0572223738,0.0189070497,-0.0120728295,0.1538337171,0.1664831936,0.2038850635,0.0279007833,0.3325169683,0.2368467748,-0.1887479424,0.1309105009,-0.2539086342,0.3011049032,-0.4800184965,0.5105648637,-0.2233593464,-0.0258164071,-0.1737922728,0.2007531822,-0.3118858337,0.0826557502,0.3412695229,-0.070790261,-0.1436940283,-0.1380105168,0.1252918541,0.2120564431,0.3860343695,0.6636459827,0.1432115138,-0.1222169176,-0.5054961443,-0.6517893672,0.0094609438,-0.154744193,0.2385486364,-0.0343375392,-0.0520378128,-0.0560296066,-0.1233821213,0.2809163928,-0.2777066231,-0.257175535,-0.2585817277,-0.3363167048,0.1588512659,-0.0823896453,0.1069965512,-0.0023992488,0.0230883099,0.0572210141,0.0107903378,0.0245506894,0.1272045672,0.0633631572,-0.0262505691,-0.3820322752,0.2086174041,0.1341861039,0.0720930919,0.1193577647,-0.3306094408,-0.3511256576,-0.5532330275,-0.2030267864,0.2234605402,-0.3612075448,0.303065002,-0.1742681563,-0.1935170442,-0.243896842,0.4328301847,0.0439809561,-0.3991267681,-0.2341060191,0.0182330292,0.1064148694,0.0519982092,0.2665888667,0.4902418852,0.0253397375,0.2217826396,-0.1594917029,-0.2655332983,0.4738945365,-0.4630522132,0.0700498894,-0.1982783675,0.2040320635,0.3320183456,0.0108951069,-0.4229315519,-0.154703632,0.4983106256,0.4719289839,-0.2222270668,0.4720397294,-0.0385535173,-0.1091837808,0.001023088,0.3369314075,-0.0032441237,-0.2709363103,0.1793895066,-0.1309637427]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/805","title":"On loading a metric from datasets, I get the following error","comments":"Hi ! We support only pyarrow > 0.17.1 so that we have access to the `PyExtensionType` object.\r\nCould you update pyarrow and try again ?\r\n```\r\npip install --upgrade pyarrow\r\n```","body":"`from datasets import load_metric`\r\n\r\n`metric = load_metric('bleurt')`\r\n\r\nTraceback:\r\n210 class _ArrayXDExtensionType(pa.PyExtensionType):\r\n 211 \r\n 212 ndims: int = None\r\n\r\nAttributeError: module 'pyarrow' has no attribute 'PyExtensionType'\r\n\r\nAny help will be appreciated. Thank you. ","comment_length":31,"text":"On loading a metric from datasets, I get the following error\n`from datasets import load_metric`\r\n\r\n`metric = load_metric('bleurt')`\r\n\r\nTraceback:\r\n210 class _ArrayXDExtensionType(pa.PyExtensionType):\r\n 211 \r\n 212 ndims: int = None\r\n\r\nAttributeError: module 'pyarrow' has no attribute 'PyExtensionType'\r\n\r\nAny help will be appreciated. Thank you. \nHi ! We support only pyarrow > 0.17.1 so that we have access to the `PyExtensionType` object.\r\nCould you update pyarrow and try again ?\r\n```\r\npip install --upgrade pyarrow\r\n```","embeddings":[-0.4477339387,0.0432356484,-0.0295897797,0.4208563864,0.5251889825,0.1035732403,0.1647764891,0.1940906644,0.1096883938,0.1362633109,-0.1096516475,0.3368699849,-0.0627767667,-0.1668589115,0.1410338879,-0.2463784814,-0.0271944776,0.1920602918,0.0922093615,-0.0036332002,-0.3685493469,-0.1364815086,-0.3558763266,0.1456970572,-0.1629061401,0.2254023403,0.1931549758,0.0090093873,-0.398291707,-0.7526149154,0.3747636974,-0.1119994223,0.3816016912,0.3474768698,-0.0001240613,-0.053221643,0.469494462,0.0690529719,-0.0325647779,-0.1555888951,-0.0472494997,-0.5924620628,0.4746727645,-0.2663748264,0.1151427776,-0.3819155693,-0.0803709626,0.0106822941,0.0000424386,0.3255780041,0.1146777421,0.262067914,0.1725226194,-0.0434034839,0.2312049717,-0.2000409216,-0.0261914544,0.7022871375,0.3933603764,-0.0741885453,0.0453938395,-0.1326176524,-0.2683163583,0.3054684401,0.6438337564,0.0471058302,0.7276001573,0.1125335544,-0.0025623052,0.2077432424,0.5605256557,-0.3992996216,-0.1872151494,-0.0055953818,0.1198050156,-0.5990043879,-0.005785937,0.1554776579,-0.0863231942,-0.0297744703,-0.1559571475,-0.2771827579,-0.3049781024,0.3827385306,-0.4256869256,0.1999412924,-0.0028616376,0.1904506534,0.0421300307,0.1036916971,0.3612973988,0.1235068142,-0.0343147144,0.3206458986,-0.1970831752,0.3396155238,-0.0133926738,0.0045510773,0.1632942855,0.0130464789,-0.1188294515,0.01834419,0.2028711885,0.3485454321,0.1614825428,0.3077957034,0.167246446,0.0110294614,0.0559720658,-0.0040248726,0.0830651447,-0.0670681298,-0.216040954,-0.4661629498,0.5968541503,0.066800788,0.301246047,-0.1160360202,-0.4341298044,-0.1129990816,-0.0865911841,-0.0943883881,0.0497982688,0.2513809502,-0.1997378767,0.091239281,0.067767866,0.3696974516,-0.2501341403,-0.1857693344,-0.195272699,0.106611982,-0.2625359595,0.000029409,0.0891798884,-0.0568173267,0.0568081327,0.0030256815,0.0976196378,-0.1197636873,-0.0368676521,0.055588901,-0.3641157746,0.2943823338,-0.0207618456,0.0673542768,0.155113399,-0.2252222896,-0.2445803434,0.4579228759,-0.6061689258,-0.2087029815,-0.1568139642,0.1332367659,-0.3736676574,-0.1092547998,-0.3528388441,0.0116287572,0.0077765966,-0.4129943848,-0.0713235363,-0.311319828,-0.1520630121,-0.3541125059,0.2045817971,-0.0227164198,-0.520308733,-0.3517714143,-0.0129674748,-0.2946338356,0.2751324475,-0.0584545434,-0.1662703305,0.3013266623,0.0159817003,0.1851962358,0.6382400393,-0.4752629399,-0.3443492949,-0.1950326562,-0.0009902224,-0.2643800974,-0.1600805372,0.0409099795,0.1501144022,0.0641754866,0.2980703712,0.2062173188,-0.0031768342,-0.1594294161,-0.1733082384,-0.2832826376,0.4397923946,0.3413206339,0.3044853508,-0.0591276512,0.169913739,-0.1738015413,0.1014139801,-0.0718534738,-0.1542688012,0.1848923564,0.3217561841,-0.035723228,0.0550460368,-0.4485323727,-0.1258915663,0.1892145276,-0.3106262982,-0.1679628491,0.0042414125,0.1094830856,-0.4346903563,0.2367288172,-0.1108597293,0.3339415193,-0.0271152575,-0.2589212358,-0.2148229182,0.1870917529,-0.2771652937,-0.1958680898,-0.0949810445,0.1448462158,0.0309311766,0.3782994747,-0.0787290782,-0.2414984405,-0.0593729168,0.256108284,0.3077491224,0.0221542884,-0.153763026,0.2603686154,0.1600511521,0.0808942989,-0.2138433903,0.3093136251,0.182526201,-0.4685382545,0.1197991669,-0.0668745339,0.0494302139,-0.0901229233,0.4214505255,0.3279112577,0.2351516187,0.1385500431,-0.2111467719,0.0676800162,-0.0596081987,0.1601516604,0.1424189508,-0.0854180828,0.2626128495,0.0419947952,0.1380295008,-0.1082081273,0.2867133915,-0.3000459969,0.289277941,-0.1676278859,0.3101268113,0.0500693247,-0.4673430324,-0.049384214,-0.0922976583,-0.4501459002,0.3805409968,0.1928175837,-0.1406975836,0.2333019227,-0.0476165079,-0.0548883602,0.1000475436,0.1550489217,0.1539150923,0.2955878377,0.137259081,-0.0522830151,-0.2871318161,0.0765533745,0.007502296,0.3762489557,-0.4939357638,0.002011366,-0.2372105122,0.3236038983,0.1041835174,-0.6289691329,-0.2238147259,-0.2210235596,-0.008297204,0.0773931369,0.2758723795,0.2762283683,0.0756507516,0.0263556875,0.1077225134,-0.3396800458,-0.0453541875,-0.4383217096,-0.2347650528,-0.0768533647,0.0777075887,0.119253017,0.209879443,-0.3348416388,0.2730451524,-0.1901924014,-0.2635513842,-0.1309440881,0.0219958052,0.4793193042,0.1622716486,0.172949627,0.0853078142,0.1829311997,0.5639775991,-0.3476026058,-0.0682466552,0.3482559025,-0.4411282837,-0.0449686944,0.0662383065,-0.0566097982,-0.3273491263,-0.2449299246,-0.1256792247,0.222770974,0.0177656729,0.0613550209,0.2782217264,0.2475695908,0.1272646785,0.0482373424,-0.2141440064,-0.0603616834,0.2959457338,-0.1059392169,-0.3464952111,0.2384506315,0.0630744696,0.1560287178,0.1219247431,-0.207655713,-0.5688388944,0.0659226999,0.4641953409,-0.1530641466,0.0602020621,0.2451774925,0.0405853204,0.0811165944,-0.1358038485,-0.1196376011,-0.1869875491,0.0505612865,0.202673614,-0.2258238047,0.078733027,-0.3294830918,0.5184357762,-0.0982304513,-0.0923547223,0.285625577,-0.2308761477,0.2227141857,0.0457650647,-0.2934775651,0.1164181307,0.0845765397,-0.0986220539,0.1352085471,-0.093272984,-0.1219238639,-0.0720164403,-0.1641139686,-0.1699750423,-0.0842036158,-0.0153134251,0.1153476387,0.2311875224,-0.0752974153,0.0969202593,-0.375187695,-0.1673185825,0.1090276539,0.1622146517,-0.0680233538,-0.0152940219,0.0940636396,-0.3395543993,-0.1821203232,0.3705577552,0.1456337273,0.0941573754,-0.0666492656,0.2181899995,0.2654320002,0.0250836518,0.4870398343,-0.1097779647,0.0895300061,0.187118277,0.2904042602,-0.7096180916,-0.1157720834,-0.0955756307,-0.1859760284,-0.1485857368,0.0202713721,-0.2645119727,-0.0277644955,0.1452292055,0.4343498349,-0.2057578713,-0.1839661747,-0.378351897,-0.3793309331,-0.1207828298,-0.2427337319,-0.1333236992,0.5662702918,0.224322632,0.0947568417,-0.249936983,-0.0046907919,0.1537576467,0.075743638,0.232502535,-0.3217101693,-0.1023777351,0.1328735203,-0.2103895396,0.4325802624,0.3238072693,0.1668313742,-0.6007301807,-0.0192839559,0.2132887244,0.4657883048,-0.0689148903,-0.1712119877,-0.0001081121,-0.243932873,0.0699209422,-0.0859471858,0.0014571475,0.4349576831,0.1507766694,-0.2472898215,-0.1691262126,0.7192626595,0.2362060547,-0.0063843532,0.2447104007,0.1861101538,-0.1508256793,0.0386650562,0.2608062923,0.8751792908,-0.1485056132,-0.0356448442,0.6681911349,0.065314725,-0.0434384122,-0.2077663839,-0.0441976152,-0.3461956382,-0.0706757382,-0.1325065047,-0.1745797396,0.2734085917,-0.1651379317,-0.0012432819,0.1631556004,-0.3736902177,0.1283740997,0.2024357915,-0.1762660444,0.0211012959,0.0203247406,-0.1474990696,0.0733399391,-0.317696631,0.0513553508,-0.2012727857,0.0791224688,0.0257730763,0.0428309366,-0.1739718467,-0.1566652805,-0.2334020883,0.0920030549,0.1668938994,-0.3158762157,-0.0182062015,0.3153294027,0.1252102852,-0.1986960024,-0.2998599708,0.3306165636,-0.3592638969,-0.1649204344,0.0045819022,0.1117534116,0.0692865551,-0.13951011,-0.3670636714,0.3593726158,-0.013040673,-0.0593047775,0.1710724235,0.1315307617,0.2078836858,-0.2934139073,-0.1184198931,-0.373143822,0.4853237569,-0.1437186748,0.0763513446,0.2060138285,-0.1366907209,0.2870102823,0.1414489746,-0.2749725878,0.0407083109,0.2111573517,0.0784192979,-0.1541565955,0.354511261,-0.0106181158,-0.0145592671,-0.1102425456,0.0509982705,0.1905712485,-0.2955195308,0.1542095244,0.3862246573,0.3326420486,-0.0159647483,0.4168637991,0.2138174176,-0.2462377399,-0.1240968257,-0.5434438586,-0.1233552918,0.3222772181,0.0183347557,0.024310872,-0.0960983038,0.1499977261,-0.0457147136,-0.0219500307,-0.2024070174,0.2412581891,0.0232405532,0.0797862634,-0.1143792868,0.0139664533,-0.0038551907,-0.0383789316,0.0483501293,-0.2340011448,-0.1978696436,-0.1549497396,-0.1765660644,0.21694462,0.0259698499,0.050173007,0.1963224262,-0.1254276186,-0.1540233791,-0.0930952951,-0.0482307784,0.1811447293,-0.1759063303,0.270262897,0.3743322194,0.1791905016,-0.2139276117,0.1777554154,0.0246975571,0.4639771581,-0.0739620104,0.0156645514,0.0910841599,0.1506349146,-0.0454974622,0.0857112259,0.3387261331,-0.2733119428,0.3574752212,-0.3903368413,0.3122903705,-0.1171356812,0.1881558895,0.2142785937,-0.3070135117,0.1832173318,0.018659791,0.12433061,-0.3278883994,-0.0417937972,0.3054873049,0.0874369815,-0.1000354663,0.1018767804,0.1763714105,0.010102204,0.0436565652,-0.0303970817,0.3371593654,0.0731966421,-0.0185371563,0.4552449286,0.032499142,0.0711443722,0.5597202778,0.2887055576,0.1078417748,0.4497839212,-0.3665194809,0.3961811662,-0.1166441515,0.1072618887,-0.1334351748,-0.3418035805,-0.0803092644,0.0999020636,0.146904707,-0.2909873128,-0.1095117852,0.5688402057,-0.123311393,-0.1263167709,-0.2314775288,0.2086852491,-0.0223884061,-0.0030338641,-0.2800158262,-0.0393800624,0.063828297,-0.1271588057,-0.0191892404,-0.2286751717,0.0651937202,0.1237080917,-0.1528433859,-0.4439595044,0.1534468681,0.3328965604,-0.0695677847,0.0009243129,0.1420506984,0.0221315846,-0.0431055278,0.1155198738,0.3692893684,0.4402990937,0.2888293862,-0.2557513118,-0.164584443,-0.2727389634,-0.1934591979,0.0214600712,0.2022660673,0.2829779387,0.1208724976,0.0573234521,0.1422393769,-0.117505908,0.2880812585,-0.0509856567,0.0716106072,-0.1648777127,0.3716299236,-0.1357873827,-0.2858300209,-0.3438405693,-0.2297124714,-0.1930754185,0.1820337176,0.4044196904,0.2193543613,-0.0469588898,-0.0238010362,-0.00903646,-0.1081417426,0.3285022378,0.3346221149,0.4529925287,-0.3661454022,-0.0021701709,-0.748519063,0.3343604505,-0.1507086754,0.0019647081,0.4247776568,-0.0297553502,-0.258800745,0.3020583391,0.2716748118,0.465641737,0.263826102,-0.0160762165,-0.472394824,-0.2278305441,0.0790031925,-0.0950911567,-0.2266480178,-0.2480527163,0.1562419683,-0.643368423,-0.0659800693,-0.1106954366,0.0169838034,0.0531066954,0.3780280054,0.6841909289,-0.0489702299,0.8348818421,0.1760093719,0.3336000144,-0.0811977461,-0.2964093387,-0.1233666316,0.1439043432,-0.18884857,0.0180383697,-0.2053116411,-0.0396404937,-0.1567749232,0.4555687904,-0.0951712728,-0.2828136384,-0.2444699407,0.1347707957,-0.0988137573,0.1209565252,-0.0876736641,-0.1292894483,0.1170700267,0.33846578,-0.1873058677,-0.2736490369,0.5236207247,-0.2739265859,-0.3627237976,0.1666277796,0.1733132005,0.1041228399,-0.1606012881,-0.5872601867,0.0850027874,0.5146576166,0.0195272472,-0.2608084977,0.2799982727,-0.1740917861,0.0932867229,-0.0409778617,0.313917309,-0.1361715496,-0.0201246124,-0.205331862,-0.1722045541]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/804","title":"Empty output\/answer in TriviaQA test set (both in 'kilt_tasks' and 'trivia_qa')","comments":"cc @yjernite is this expected ?","body":"# The issue\r\n\r\nIt's all in the title, it appears to be fine on the train and validation sets.\r\n\r\nIs there some kind of mapping to do like for the questions (see https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md) ? \r\n\r\n# How to reproduce\r\n```py\r\nfrom datasets import load_dataset\r\nkilt_tasks = load_dataset(\"kilt_tasks\")\r\ntrivia_qa = load_dataset('trivia_qa', 'unfiltered.nocontext')\r\n# both in \"kilt_tasks\"\r\nIn [18]: any([output['answer'] for output in kilt_tasks['test_triviaqa']['output']]) \r\nOut[18]: False\r\n# and \"trivia_qa\"\r\nIn [13]: all([answer['value'] == '<unk>' for answer in trivia_qa['test']['answer']]) \r\nOut[13]: True\r\n# appears to be fine on the train and validation sets.\r\nIn [14]: all([answer['value'] == '<unk>' for answer in trivia_qa['train']['answer']]) \r\nOut[14]: False\r\n\r\nIn [15]: all([answer['value'] == '<unk>' for answer in trivia_qa['validation']['answer']]) \r\nOut[15]: False\r\n\r\nIn [16]: any([output['answer'] for output in kilt_tasks['train_triviaqa']['output']]) \r\nOut[16]: True\r\n\r\nIn [17]: any([output['answer'] for output in kilt_tasks['validation_triviaqa']['output']]) \r\nOut[17]: True\r\n\r\n```","comment_length":6,"text":"Empty output\/answer in TriviaQA test set (both in 'kilt_tasks' and 'trivia_qa')\n# The issue\r\n\r\nIt's all in the title, it appears to be fine on the train and validation sets.\r\n\r\nIs there some kind of mapping to do like for the questions (see https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md) ? \r\n\r\n# How to reproduce\r\n```py\r\nfrom datasets import load_dataset\r\nkilt_tasks = load_dataset(\"kilt_tasks\")\r\ntrivia_qa = load_dataset('trivia_qa', 'unfiltered.nocontext')\r\n# both in \"kilt_tasks\"\r\nIn [18]: any([output['answer'] for output in kilt_tasks['test_triviaqa']['output']]) \r\nOut[18]: False\r\n# and \"trivia_qa\"\r\nIn [13]: all([answer['value'] == '<unk>' for answer in trivia_qa['test']['answer']]) \r\nOut[13]: True\r\n# appears to be fine on the train and validation sets.\r\nIn [14]: all([answer['value'] == '<unk>' for answer in trivia_qa['train']['answer']]) \r\nOut[14]: False\r\n\r\nIn [15]: all([answer['value'] == '<unk>' for answer in trivia_qa['validation']['answer']]) \r\nOut[15]: False\r\n\r\nIn [16]: any([output['answer'] for output in kilt_tasks['train_triviaqa']['output']]) \r\nOut[16]: True\r\n\r\nIn [17]: any([output['answer'] for output in kilt_tasks['validation_triviaqa']['output']]) \r\nOut[17]: True\r\n\r\n```\ncc @yjernite is this expected ?","embeddings":[0.470046103,-0.4707074761,-0.1075564846,0.119669348,0.3161605895,-0.0143134976,0.5024516582,0.43479985,0.1569281071,0.28908059,0.1453078091,0.2302017063,-0.0752013922,0.3938882947,0.0054924078,0.2260351479,-0.0565773509,0.1879831403,-0.2112382501,-0.2724748254,-0.2652142346,0.0140367523,-0.2291554958,-0.1191356257,-0.3718957603,0.0364993177,-0.0163232498,-0.0340056717,-0.0027413897,-0.2856480181,0.3308844268,0.0400861241,-0.514667213,0.2728256285,-0.0001258685,-0.0103399735,0.1982664019,-0.1415138096,-0.3723581731,-0.2981100976,-0.5491599441,0.0122515699,0.2224714011,-0.2390356511,-0.3647924066,0.271248281,-0.0211377181,-0.2958577275,0.3061316013,0.3962603211,0.0836722776,0.3363839686,-0.351175189,-0.1004119217,0.0782362521,0.1827397496,-0.1868788749,0.0799213573,0.3239274323,-0.1741539091,0.1030163243,0.4104625583,0.1057259068,0.1019604579,-0.009865704,0.0874379724,0.1973355114,-0.4942217469,0.2395603955,0.4707058668,0.1272066981,0.0174103882,-0.4457537532,-0.396748215,-0.0347599797,-0.1832487881,0.2236087918,0.2811259627,-0.3021056056,0.0474173389,-0.3934358358,0.0438106395,-0.1686397791,0.0055102953,-0.5558186173,0.3411976099,-0.0574244596,0.0811748207,-0.2847796977,0.0257879496,-0.0974241048,-0.2869355679,0.0706435367,0.1689620614,-0.4941113591,-0.0407925285,0.2595440447,-0.1485185921,-0.0117396088,-0.1723079085,0.1077351198,-0.0070350552,-0.2377846092,0.1717830449,0.4710021317,0.3623896539,0.4358532727,0.1125941053,0.1049214229,-0.0060239197,-0.179536283,0.0467837416,-0.0174206328,0.1253738105,-0.179111138,0.0559331365,0.4148238003,-0.3581530154,-0.5705876946,0.0764494315,-0.3337844312,0.0449087732,-0.0965514109,0.3889575899,-0.2781950533,0.3332197368,-0.1912667751,0.5201530457,-0.2272969931,0.1342340559,-0.1845477968,-0.0310200062,-0.2691880763,0.0450489037,0.1864545196,0.0324998386,0.1601480842,0.0806090161,-0.0903711468,-0.0822831541,0.1727371365,-0.288472265,0.5071529746,0.1978860795,0.1181834862,0.1843982935,0.3236449957,-0.1403573751,0.0008154366,0.4704214931,-0.0974565819,-0.0895353854,0.1787093729,0.1369044036,-0.4601016045,0.0092966389,-0.0062864074,0.3752223253,0.2645501196,0.0765052587,0.1074672863,-0.1584148556,-0.0430232324,-0.1662305593,0.1935497522,0.2128446996,-0.371735841,-0.1891481131,0.2749037445,0.0330352597,-0.1405436099,0.0911584198,-0.0245922655,0.3979967237,-0.2747407258,0.0493222103,0.1446451098,-0.6225361228,0.0299887825,-0.201361239,-0.0072999047,0.3978205621,-0.0205094386,-0.1343879998,0.1917884797,0.219424963,0.2321366817,-0.0246691927,-0.0132375248,-0.1208119392,-0.2493067235,0.013842498,0.2646315396,0.0539087616,0.247583285,-0.0099763777,-0.1797798425,-0.0960119441,0.2337658107,-0.0304874294,0.1139031649,0.1219114214,0.4234983325,0.0240004081,-0.0515378825,0.0465664007,-0.2907760143,0.199155435,-0.010206935,0.4291464388,-0.0339556783,-0.2332825661,-0.2357919812,-0.0800162479,-0.4987868965,-0.5502659678,0.0632421672,0.0771128163,-0.0495685898,0.0897676423,-0.1309790462,0.0157175437,0.0596020818,-0.1172991246,-0.3313575983,0.3278958797,0.0816954896,0.0437066443,-0.3814686537,0.3813838363,-0.0341968052,0.0992077962,0.1005259305,-0.0786303133,0.12573044,-0.0509878024,0.0139121152,0.3072133064,0.3162116408,-0.1393978894,-0.3494017124,0.2730677128,0.1508397609,-0.2726196647,-0.5103601813,0.5709990263,-0.0161707476,0.450881958,0.0766283497,0.0047294521,0.0057516424,-0.2856660187,-0.2212484479,-0.1818615496,-0.0015447924,-0.305464685,0.1646820009,0.1642266363,-0.3667792082,0.1001939476,0.3992271423,0.0374091342,0.2036823332,-0.3033897579,-0.0087337391,-0.0095175933,0.1405608207,-0.0625931546,0.4369859695,0.1655027419,0.2154754698,0.0035496207,0.2316295505,-0.2829145491,0.1838825047,-0.0201991834,0.1908873618,0.4695929289,0.3519487381,-0.1561600715,-0.1707765907,0.2944715917,0.0798154548,0.1067897752,-0.3135199249,-0.2326579839,-0.1617079228,0.0162663627,-0.1390534192,0.0200822707,0.0780735016,-0.3897257149,0.3137888014,-0.0021857324,-0.3013088405,0.2260514796,-0.0786505789,0.1568452418,-0.020977404,-0.3953561783,-0.0494896956,-0.4904035628,-0.237537995,0.1418428421,-0.2333806753,-0.1647131443,0.108165428,-0.1420109868,-0.5023095608,-0.1643257439,-0.4955572784,0.1767603159,-0.2250026315,0.2891466618,0.1090339795,0.4650039673,-0.3373751044,-0.0198076069,0.4969296157,0.0384968072,-0.388518393,0.0214972179,0.1770920157,0.0573991463,-0.1517627388,-0.7311222553,-0.078869991,-0.3435739875,0.2690479755,-0.1643610597,0.1446876526,0.4247790277,-0.0994940549,-0.0679050386,-0.2066884637,-0.1697100401,-0.1584306806,0.0259056017,0.1509226859,-0.0642626137,-0.3531973064,0.0170902926,-0.2293687612,0.3519566953,0.0741365328,-0.4251024723,-0.3854763508,-0.0241226573,0.455834657,0.2231852412,-0.1219860092,0.2559668124,0.0093715219,-0.0576592572,-0.2526750565,-0.0342711657,0.5560509562,0.3253906369,0.0661792904,0.0929436907,-0.0265373159,0.2776345015,0.5226078033,0.328876853,0.3128817379,0.1879550219,-0.235994637,0.3911180496,-0.2682639956,-0.3924846649,0.1946950257,0.2037714571,-0.1287907958,0.2727068961,0.0182134844,0.3206771612,-0.2564151883,0.1822143048,-0.5257818103,-0.264254719,0.1040931493,-0.3858921826,0.1111393198,0.170662269,0.1868489087,0.1279434413,-0.0776631981,0.0419892482,0.0143784722,0.1995806843,-0.0833605975,-0.5689995885,-0.2428134978,-0.1847887933,0.1079054624,0.0509483144,0.0795515925,-0.1375652999,-0.250235647,0.2331012785,0.1292876303,1.0130125284,-0.0740614384,-0.0319090299,-0.0090550538,-0.1558130085,-0.3871887922,-0.0829621479,-0.4501017034,-0.0646816865,0.151218757,0.4560911059,-0.2827369273,-0.1700228155,-0.0841897577,-0.0285576861,-0.1451736838,-0.0644074976,-0.2939057052,-0.2241567671,0.0412888862,0.1370394975,0.1754738092,0.4539821744,-0.2620460391,-0.3331947327,-0.1384079754,0.1488732994,0.3353715539,0.3199372888,0.2879779637,-0.0597172379,0.1021569446,-0.0761029348,-0.0671868995,0.0528622232,0.5140840411,-0.1037818938,-0.0244908296,0.2118850946,-0.2695404589,0.4100121558,0.3712159693,0.1504175812,0.1329444796,-0.2385359704,-0.0380749516,-0.1043800563,0.0725621954,0.1283679456,0.0275403447,-0.045035053,-0.4008689523,0.1207633466,-0.0233024117,-0.1708291173,0.2370844185,0.0667487681,-0.2501732111,0.081417568,0.042625092,0.8845587969,-0.0745196417,0.3844709694,0.640950799,0.1782277077,0.393357724,-0.0173435342,0.2568463385,-0.2962932289,-0.1554338932,0.0147393281,-0.1158924922,-0.0188899338,0.387160778,-0.1849626154,0.5355915427,-0.0022763365,0.5156623125,-0.0309016481,-0.3544171751,-0.0346076712,0.2444928885,-0.5778464079,0.0128702633,0.0629854053,0.4759119153,-0.1577835083,-0.0793719664,-0.3572486639,-0.1709271371,-0.0124499416,0.2494312525,-0.3965124786,-0.0075495904,0.3820751011,-0.2730431557,-0.2918457687,0.4644520879,-0.069094643,0.2014669329,-0.1726857424,0.2123502791,0.0767409131,0.3337996602,0.1881392449,0.2351187319,0.2821857333,-0.1157104895,-0.4258422256,0.1755250692,-0.2977012992,0.2410799712,-0.2195341438,0.4140775502,0.1423128247,-0.2434030771,-0.511896193,0.0023160435,0.245759055,-0.1987230629,0.0597085506,-0.1596779376,-0.1974715143,0.2519940138,0.0654676706,-0.3421114385,0.1124440953,0.2364288867,0.2396625876,-0.0501065552,0.7466733456,-0.0705334842,-0.0520626828,-0.292070955,0.0635502338,0.2165629715,-0.1665552706,0.3654384911,0.1800346822,-0.0996985584,0.0190232079,0.4681164324,0.0677305087,0.1628259122,-0.0848461688,-0.4124251008,0.0324059613,0.3345588446,-0.0023687573,0.2419579476,-0.0836079866,-0.0140135409,0.2187827229,-0.1911319196,-0.2652292252,-0.2245965004,-0.37805444,0.4298984706,0.0174768101,0.0494727716,0.2688782513,-0.1162112281,0.1065473109,-0.3401480019,-0.3123820722,-0.2255500257,0.0367251709,0.199727878,0.1910889745,0.1579169035,0.0174453352,0.0058834516,-0.0014658808,-0.2527084649,0.3498785198,0.3625028431,-0.0101927388,0.2557857037,0.0266031008,0.3602588475,-0.2013679594,0.031796135,0.0370561332,-0.1721069068,0.2100248188,-0.1453176886,0.1037896052,0.0080720559,0.1306941956,0.0954823419,-0.0303623602,-0.0535186753,-0.0792052597,-0.2223518938,0.0409040637,-0.0312793553,0.1334782243,0.2560643852,-0.170611307,0.2422808111,-0.0323660299,0.2163337767,-0.2987174392,0.3679385483,0.1134490818,0.1150858998,0.1203411296,0.1817978024,0.102775082,-0.3937260509,0.2938078046,0.2180601805,-0.1342508197,-0.3611879647,0.2778706253,0.411811769,-0.042282436,0.0191788636,0.0044959118,-0.2609452009,-0.2788941264,0.0022908694,0.0535298139,0.284119606,0.0009955565,-0.1397132427,0.1420059502,-0.0057213553,-0.0119816279,0.0052724448,-0.3401517272,-0.0475044958,0.1347774118,0.2582471669,-0.0777915344,-0.2393870652,0.1086823642,-0.1181267053,-0.1963922083,-0.1292683184,-0.0849284455,-0.2074318975,-0.0080732163,-0.1872795969,0.0858818367,-0.3121996522,0.4465610087,-0.0449161455,0.0027646481,-0.4077846706,0.1059214324,0.3924972713,0.3049200475,0.1198190302,0.4659336507,0.2885671258,0.1070851907,0.2401887327,0.5957577229,0.3030447662,0.421821326,-0.187180236,0.0574703775,-0.0797938108,-0.3175366223,-0.0661245435,0.2808930874,0.1926131099,-0.0128190443,0.3018321991,0.1150097102,-0.0837434232,0.3055701852,0.1440033615,-0.065081872,-0.2222093344,0.2926779091,0.0533282571,-0.0894696489,0.0699352995,-0.1069631353,-0.5942695737,-0.1775361598,0.1187378392,0.1969788224,0.2502897084,-0.0112900212,0.0487038009,-0.1111362129,0.0428541675,0.1670473963,0.1388010532,-0.3474678099,0.2432246953,-0.4494113028,0.258580327,0.0451372042,-0.0504662134,-0.2393674701,0.1519651562,-0.0016839303,0.013859191,-0.1522593051,-0.1242899597,0.3138854802,0.0405595154,-0.6332930326,0.0363241732,0.2352402657,-0.0478335842,0.0201331154,-0.2898215353,-0.0292829499,0.0028303722,-0.1069519296,0.1977178752,0.1175835058,0.0818364918,0.0537293628,0.0638592169,0.2540250421,0.6821110845,-0.0234102793,0.1346951127,-0.0523871854,-0.2156037986,-0.0907000601,0.104758881,0.1677621603,0.2770222723,0.2091981918,-0.2570753694,0.1253728718,0.15254502,0.1906967461,-0.1393731982,0.0428805053,0.1240224019,0.084628351,-0.0611387864,0.0215719901,-0.0070201471,-0.0116981436,-0.0304475091,-0.2408078313,-0.5170319676,0.4177877605,-0.4951081276,-0.3498734534,-0.3034233153,0.1974147856,0.1040621325,-0.2009649575,-1.0071469545,-0.1534320712,0.4317228198,-0.1761640012,-0.0324237756,0.2053984404,-0.2931627631,-0.1114903241,-0.1606302261,0.030078765,0.0450900048,-0.1146132275,-0.1391947418,-0.4239300787]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/804","title":"Empty output\/answer in TriviaQA test set (both in 'kilt_tasks' and 'trivia_qa')","comments":"Yes: TriviaQA has a private test set for the leaderboard [here](https:\/\/competitions.codalab.org\/competitions\/17208)\r\n\r\nFor the KILT training and validation portions, you need to link the examples from the TriviaQA dataset as detailed here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md","body":"# The issue\r\n\r\nIt's all in the title, it appears to be fine on the train and validation sets.\r\n\r\nIs there some kind of mapping to do like for the questions (see https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md) ? \r\n\r\n# How to reproduce\r\n```py\r\nfrom datasets import load_dataset\r\nkilt_tasks = load_dataset(\"kilt_tasks\")\r\ntrivia_qa = load_dataset('trivia_qa', 'unfiltered.nocontext')\r\n# both in \"kilt_tasks\"\r\nIn [18]: any([output['answer'] for output in kilt_tasks['test_triviaqa']['output']]) \r\nOut[18]: False\r\n# and \"trivia_qa\"\r\nIn [13]: all([answer['value'] == '<unk>' for answer in trivia_qa['test']['answer']]) \r\nOut[13]: True\r\n# appears to be fine on the train and validation sets.\r\nIn [14]: all([answer['value'] == '<unk>' for answer in trivia_qa['train']['answer']]) \r\nOut[14]: False\r\n\r\nIn [15]: all([answer['value'] == '<unk>' for answer in trivia_qa['validation']['answer']]) \r\nOut[15]: False\r\n\r\nIn [16]: any([output['answer'] for output in kilt_tasks['train_triviaqa']['output']]) \r\nOut[16]: True\r\n\r\nIn [17]: any([output['answer'] for output in kilt_tasks['validation_triviaqa']['output']]) \r\nOut[17]: True\r\n\r\n```","comment_length":32,"text":"Empty output\/answer in TriviaQA test set (both in 'kilt_tasks' and 'trivia_qa')\n# The issue\r\n\r\nIt's all in the title, it appears to be fine on the train and validation sets.\r\n\r\nIs there some kind of mapping to do like for the questions (see https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md) ? \r\n\r\n# How to reproduce\r\n```py\r\nfrom datasets import load_dataset\r\nkilt_tasks = load_dataset(\"kilt_tasks\")\r\ntrivia_qa = load_dataset('trivia_qa', 'unfiltered.nocontext')\r\n# both in \"kilt_tasks\"\r\nIn [18]: any([output['answer'] for output in kilt_tasks['test_triviaqa']['output']]) \r\nOut[18]: False\r\n# and \"trivia_qa\"\r\nIn [13]: all([answer['value'] == '<unk>' for answer in trivia_qa['test']['answer']]) \r\nOut[13]: True\r\n# appears to be fine on the train and validation sets.\r\nIn [14]: all([answer['value'] == '<unk>' for answer in trivia_qa['train']['answer']]) \r\nOut[14]: False\r\n\r\nIn [15]: all([answer['value'] == '<unk>' for answer in trivia_qa['validation']['answer']]) \r\nOut[15]: False\r\n\r\nIn [16]: any([output['answer'] for output in kilt_tasks['train_triviaqa']['output']]) \r\nOut[16]: True\r\n\r\nIn [17]: any([output['answer'] for output in kilt_tasks['validation_triviaqa']['output']]) \r\nOut[17]: True\r\n\r\n```\nYes: TriviaQA has a private test set for the leaderboard [here](https:\/\/competitions.codalab.org\/competitions\/17208)\r\n\r\nFor the KILT training and validation portions, you need to link the examples from the TriviaQA dataset as detailed here:\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md","embeddings":[0.4346867502,-0.3452418447,-0.0812816396,0.1217559502,0.2600176632,-0.079768993,0.4260964692,0.4618622661,0.13298136,0.2688274384,0.0946825445,0.2613138556,-0.0670852587,0.5240859985,0.0683326572,0.2598775327,-0.1134255379,0.0808698758,-0.1446951777,-0.2219866514,-0.2116925418,0.0270461477,-0.2168559432,-0.1244957,-0.3129106164,0.0097973114,-0.0458583422,-0.0411906838,0.048802264,-0.3793427348,0.2050960362,0.0018789937,-0.4776950479,0.250802815,-0.0001265892,-0.0889239535,0.1993868202,-0.1674762517,-0.4454841018,-0.1441676319,-0.6783222556,0.0141277779,0.170403555,-0.2482234389,-0.3704464138,0.332406044,0.0045207287,-0.3541144133,0.3715062439,0.4303570986,0.0710942596,0.3091676831,-0.3334417641,-0.1123591363,-0.0177280717,0.078342855,-0.2280900925,0.0803658217,0.219762668,-0.1270766258,0.0533198975,0.330270946,0.1531341374,0.1341962814,-0.0423491672,0.1115850955,0.1451227814,-0.5243434906,0.2711947858,0.4891029,0.1355248094,-0.0259978324,-0.4715607762,-0.3702868521,-0.0004940171,-0.2080955058,0.2873293459,0.3626281917,-0.2731277049,-0.0013897181,-0.3779554665,0.0880932957,-0.1987210512,-0.0318931043,-0.6022466421,0.2828427255,-0.0419524498,0.056307897,-0.2972771823,0.0585988387,-0.1214954033,-0.2579809427,0.0553128049,0.2168537378,-0.5435726047,-0.0804850832,0.2416573912,-0.1061613038,0.021168476,-0.1563095897,0.0860012099,0.0318802781,-0.247733444,0.133972764,0.4812690318,0.4112280011,0.4215257168,0.1192411184,0.1123541594,0.0562162548,-0.1539959162,0.0744720474,0.0348182209,0.1464500576,-0.1980704218,0.0675947815,0.335577786,-0.4146649241,-0.5860321522,0.1421547383,-0.3326154053,0.0864446908,-0.1067222133,0.320563525,-0.2858104408,0.2536134124,-0.1336442977,0.4243001938,-0.2777159512,0.0795583054,-0.1509981602,-0.0046977056,-0.2762502432,0.1931227595,0.224704355,0.0346412919,0.1216206253,0.1096358076,0.0274162982,-0.0919776112,0.1531841308,-0.2339354455,0.4344162047,0.188072145,0.1408801973,0.1189044341,0.2922724783,-0.1410393864,0.0318254307,0.4630865157,-0.0387344584,-0.1531048566,0.2196510732,0.1176871657,-0.53619802,0.0612278245,-0.0697553232,0.3681988418,0.2204560041,0.0522886068,0.0639485046,-0.17597422,-0.0918827653,-0.1526447237,0.2948548198,0.1581320167,-0.4779413342,-0.107231468,0.2162587047,0.0987294242,-0.1795674562,0.0834397152,-0.0059622596,0.3908629119,-0.2375229746,0.0595924743,0.1495541036,-0.5808662772,-0.049199108,-0.2524282932,0.010969785,0.3100386262,-0.0182249602,-0.1082027107,0.1928690821,0.1797241122,0.0380761921,0.0307709333,-0.0385318212,-0.1282064915,-0.2770301104,-0.0048107761,0.2065397203,0.1226963475,0.2405065149,-0.0155215608,-0.1409450322,0.0033424345,0.1804189235,0.0150427595,0.1349632591,0.0477367826,0.415969938,-0.0334042497,-0.0810122117,0.1435109526,-0.2264648527,0.2795407772,-0.0904851481,0.4886126518,-0.0579417124,-0.2643123567,-0.2712050676,-0.100305073,-0.5140523911,-0.6296908259,0.0657849908,0.042335704,0.0234564636,0.0934630856,-0.178200841,0.0506216809,0.0379335284,-0.1656878889,-0.3817005455,0.2512590885,0.0596449114,0.0076799048,-0.3221111894,0.5083587766,-0.0573302358,0.0728075057,0.1012719348,-0.0902488902,0.0823971555,-0.0617132597,0.1081170887,0.3362760544,0.3323972225,-0.1298761666,-0.2818245292,0.3482183516,0.0935201496,-0.2479854822,-0.5444375277,0.5965662599,0.0071430248,0.4199711978,0.0516379662,0.0222856048,-0.0676893592,-0.2766861916,-0.218831867,-0.1478293985,-0.0463503338,-0.2262968719,0.1864347905,0.1002461314,-0.2956453562,0.2027985752,0.4188897908,0.0210408829,0.2085370421,-0.3053345382,0.0180465262,-0.0307684503,0.184289664,0.0231524054,0.4413944483,0.084580645,0.2386827767,0.0726953074,0.1604368389,-0.2435355335,0.152713567,-0.0230280645,0.1729786992,0.3898729682,0.3559095263,-0.1646558493,-0.0903521478,0.3359053135,0.0750357136,0.1385468841,-0.3268065453,-0.1666371226,-0.1502793282,-0.0419876017,-0.1752014607,0.0523343533,0.0248411261,-0.4100830853,0.3285831213,-0.0310880356,-0.2622547448,0.2447648793,-0.0283536986,0.1885126978,-0.0595245399,-0.3672983944,-0.0129037043,-0.5210822821,-0.2430393249,0.1383781731,-0.1649945974,-0.1913993061,0.1399034113,-0.1132650226,-0.5428164005,-0.2194762528,-0.429346025,0.1460483223,-0.2055821866,0.3748728931,0.0741121471,0.4747892022,-0.3547728956,-0.0322884209,0.5146630406,0.021866139,-0.3692892492,0.012095049,0.157763347,0.0929032192,-0.1327739358,-0.7486938834,-0.1053432226,-0.2937638462,0.266032517,-0.086060524,0.1985882968,0.3975977898,-0.1153228208,-0.039976459,-0.1941642463,-0.1932611018,-0.1947054118,0.0486537777,0.1662123948,-0.1380502284,-0.3828946948,0.0876780897,-0.2003528923,0.5565998554,-0.0043731234,-0.4171013534,-0.4424279928,0.0046635936,0.5091130733,0.2116530091,-0.1150768623,0.2026955187,0.0264127888,-0.0459441505,-0.1555680037,0.0131564932,0.5017895699,0.3555322886,0.107933417,0.0652409568,0.0247997716,0.3510317802,0.641597569,0.3786245286,0.3638949096,0.1289847493,-0.2493529618,0.3066883981,-0.3129513264,-0.3488161564,0.3050559461,0.1317903101,-0.0871136189,0.3041988015,0.0239400975,0.4075453281,-0.204497084,0.2693419158,-0.5056167245,-0.3185827434,0.049968563,-0.3135150969,0.1914495975,0.1279785335,0.2195844799,0.1310315579,-0.1225958616,0.0516673513,-0.0825516433,0.1423896849,-0.0420479812,-0.4505999684,-0.2251456529,-0.1829802096,0.1323272288,0.0155580482,0.1016553119,-0.1608117074,-0.2590690553,0.2262765616,0.1118173078,0.866363883,-0.0023163371,-0.0271386541,-0.0249185264,-0.1305039376,-0.3616461456,-0.0936799198,-0.3786865771,-0.1036000922,0.0740301237,0.2937024534,-0.2711283565,-0.1167269647,-0.0729561821,0.0201096795,-0.1635788828,-0.0270102229,-0.3407826424,-0.2246469408,-0.0129654035,0.1619568616,0.0700423643,0.4928734899,-0.3183727562,-0.19655478,-0.0594216362,0.0894155428,0.285448581,0.3582051992,0.2959146798,-0.0059017879,0.1324221939,-0.0149633018,-0.1460424811,-0.144663617,0.4854558706,-0.0010807263,0.0362753011,0.1864040345,-0.2770252824,0.3975395858,0.4037605524,0.1326678246,0.1354075372,-0.2506318092,-0.0152144097,-0.1475256234,0.1111450344,0.0727035627,-0.0618529879,0.0788977146,-0.4128292501,0.2229206264,-0.0405777805,-0.1696251482,0.204856351,0.0768689066,-0.2017410696,-0.035131745,0.0974616185,0.8919274807,-0.0662113577,0.4134542644,0.6620019078,0.2334069461,0.3648494184,-0.0069403644,0.2821791172,-0.3736187518,-0.112103723,0.0200377982,-0.131688863,-0.0268998444,0.292886436,-0.1461332291,0.6321951747,-0.0056277239,0.5025166869,-0.0048433137,-0.3010719717,-0.0134274689,0.2009545416,-0.5890649557,0.0264617447,0.0050038607,0.4218776524,-0.1418740898,-0.0885391161,-0.3659604788,-0.0504755862,-0.0132301655,0.275477916,-0.4245266616,-0.0454652496,0.3758049011,-0.3739206791,-0.2352229953,0.5581883192,0.009953212,0.1803457737,-0.173019886,0.2048060149,0.0844999552,0.3016399443,0.2449371219,0.2526378632,0.2085363269,-0.1569750458,-0.3818918765,0.2623743415,-0.2410897911,0.2431210577,-0.2989960909,0.4151664674,0.2272371054,-0.1952878982,-0.4141255021,-0.0060981219,0.2115725875,-0.1560892463,0.066116713,-0.1261325032,-0.2398512959,0.2514291406,0.0196974669,-0.2873941362,0.1271053255,0.2731330395,0.1748232841,0.0525534302,0.7209165692,-0.0428944081,0.0004571178,-0.2384641171,0.0822174847,0.1635748297,-0.0944215506,0.2792905271,0.1726958901,-0.136572659,-0.0052635809,0.5161393285,0.0768225938,0.0829376504,-0.0560750067,-0.3959750831,0.0580971017,0.2901723683,-0.0414154828,0.1876639426,-0.1468648612,0.0004632414,0.2005597651,-0.1800325513,-0.2512340546,-0.2729480565,-0.4118520617,0.4171562791,0.0375249721,0.0428510644,0.2883031666,-0.1083395779,0.0845078006,-0.3131258786,-0.3429970741,-0.2158810347,0.0343305767,0.1953560561,0.2434573472,0.1520544738,0.0318907499,0.0267938692,-0.0017874956,-0.2190286964,0.3636837006,0.2121641636,-0.0104553755,0.2798034847,-0.0637831539,0.3293096721,-0.2127484828,0.0035482249,0.039635852,-0.1352977008,0.2089324296,-0.1167586446,0.0492621623,-0.0300230552,0.1665134728,0.1337016821,-0.0454162396,-0.0679204091,-0.0298076682,-0.2352175415,0.0124886911,-0.0512313657,0.1661792845,0.2602639794,-0.2053518891,0.2154125422,-0.0571049675,0.1771767884,-0.2969951332,0.3653949499,-0.0021922654,0.1182781607,0.134694621,0.2747142017,0.1755045652,-0.4053771794,0.3537343144,0.1953548044,-0.1874177903,-0.5741314292,0.287094444,0.4074603319,-0.1071141362,0.0490605794,0.0404189453,-0.3518250883,-0.3844277561,0.036948476,-0.008321452,0.360829711,0.0497738197,-0.1657890677,0.20719558,0.0841390043,-0.0662977099,0.0219835974,-0.2879250944,-0.060148742,0.1251440495,0.2944073379,-0.1905690283,-0.2490133494,0.1038477346,-0.1933701038,-0.1430124193,-0.0769762769,-0.1141200364,-0.2166651338,-0.034094885,-0.1444530785,0.065853931,-0.3297702372,0.43579337,-0.0282859113,0.0185010843,-0.3719087839,0.0090901926,0.4641456306,0.3565843999,0.1119121909,0.5008341074,0.21571742,0.1341407895,0.2078406364,0.6149324775,0.2586732805,0.3419251144,-0.3233906031,0.1192589998,-0.0336957052,-0.2811513245,0.0290107224,0.2035481334,0.1608706266,-0.0742123351,0.3146443069,0.0817553997,-0.0466070697,0.2061839998,0.1415967643,-0.1039579213,-0.2096673548,0.3489553928,0.0770474225,-0.1407625824,0.1081671417,-0.0765817314,-0.549679935,-0.1222210377,0.0663465783,0.2608987391,0.2080575079,-0.0133956317,0.0482256263,-0.0840744227,0.1265189052,0.1357902288,0.2407106012,-0.3984416425,0.2194608897,-0.5065744519,0.2775183618,0.1063662097,-0.1580080241,-0.2228921354,0.1281230599,-0.0125848521,0.0807103962,-0.1649163365,-0.2625878751,0.3132312,-0.0272871926,-0.6367018819,-0.0326891281,0.2390697598,-0.0783295929,-0.0863595158,-0.2787926793,0.0069236588,-0.0296368301,-0.1167245582,0.2831625938,0.0622331426,0.0659655705,0.1093872339,0.0391211435,0.2266278416,0.5997713208,0.0350952819,0.1122694239,-0.0781710222,-0.0843577906,-0.0785217732,0.1004514024,0.0732875839,0.288697511,0.2508956492,-0.2273737937,0.0480622239,0.2219939679,0.2241440713,-0.0424578115,0.0655517504,0.0978976861,0.1245917454,-0.0345968157,0.0319908224,0.0805548355,-0.0692722052,0.0166463163,-0.2565056086,-0.4255900085,0.363728255,-0.5796006322,-0.3352664113,-0.2134286761,0.1866195649,0.0342832133,-0.2241690904,-0.9314985275,-0.1921284795,0.3569231033,-0.2097530216,-0.0426321812,0.145632863,-0.2674560249,-0.1038010269,-0.1568580568,0.1217676252,0.0308610182,-0.0544659719,-0.1367144287,-0.422246635]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/804","title":"Empty output\/answer in TriviaQA test set (both in 'kilt_tasks' and 'trivia_qa')","comments":"Oh ok, I guess I read the paper too fast \ud83d\ude05, thank you for your answer!","body":"# The issue\r\n\r\nIt's all in the title, it appears to be fine on the train and validation sets.\r\n\r\nIs there some kind of mapping to do like for the questions (see https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md) ? \r\n\r\n# How to reproduce\r\n```py\r\nfrom datasets import load_dataset\r\nkilt_tasks = load_dataset(\"kilt_tasks\")\r\ntrivia_qa = load_dataset('trivia_qa', 'unfiltered.nocontext')\r\n# both in \"kilt_tasks\"\r\nIn [18]: any([output['answer'] for output in kilt_tasks['test_triviaqa']['output']]) \r\nOut[18]: False\r\n# and \"trivia_qa\"\r\nIn [13]: all([answer['value'] == '<unk>' for answer in trivia_qa['test']['answer']]) \r\nOut[13]: True\r\n# appears to be fine on the train and validation sets.\r\nIn [14]: all([answer['value'] == '<unk>' for answer in trivia_qa['train']['answer']]) \r\nOut[14]: False\r\n\r\nIn [15]: all([answer['value'] == '<unk>' for answer in trivia_qa['validation']['answer']]) \r\nOut[15]: False\r\n\r\nIn [16]: any([output['answer'] for output in kilt_tasks['train_triviaqa']['output']]) \r\nOut[16]: True\r\n\r\nIn [17]: any([output['answer'] for output in kilt_tasks['validation_triviaqa']['output']]) \r\nOut[17]: True\r\n\r\n```","comment_length":16,"text":"Empty output\/answer in TriviaQA test set (both in 'kilt_tasks' and 'trivia_qa')\n# The issue\r\n\r\nIt's all in the title, it appears to be fine on the train and validation sets.\r\n\r\nIs there some kind of mapping to do like for the questions (see https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md) ? \r\n\r\n# How to reproduce\r\n```py\r\nfrom datasets import load_dataset\r\nkilt_tasks = load_dataset(\"kilt_tasks\")\r\ntrivia_qa = load_dataset('trivia_qa', 'unfiltered.nocontext')\r\n# both in \"kilt_tasks\"\r\nIn [18]: any([output['answer'] for output in kilt_tasks['test_triviaqa']['output']]) \r\nOut[18]: False\r\n# and \"trivia_qa\"\r\nIn [13]: all([answer['value'] == '<unk>' for answer in trivia_qa['test']['answer']]) \r\nOut[13]: True\r\n# appears to be fine on the train and validation sets.\r\nIn [14]: all([answer['value'] == '<unk>' for answer in trivia_qa['train']['answer']]) \r\nOut[14]: False\r\n\r\nIn [15]: all([answer['value'] == '<unk>' for answer in trivia_qa['validation']['answer']]) \r\nOut[15]: False\r\n\r\nIn [16]: any([output['answer'] for output in kilt_tasks['train_triviaqa']['output']]) \r\nOut[16]: True\r\n\r\nIn [17]: any([output['answer'] for output in kilt_tasks['validation_triviaqa']['output']]) \r\nOut[17]: True\r\n\r\n```\nOh ok, I guess I read the paper too fast \ud83d\ude05, thank you for your answer!","embeddings":[0.4681178033,-0.4883505404,-0.1169174835,0.0633333772,0.3016143143,-0.0618819781,0.5027799606,0.441729635,0.1443300396,0.2951969802,0.150508225,0.2478302419,-0.1058168337,0.4204440415,0.0343295373,0.204022795,-0.0182283055,0.134105131,-0.2242505699,-0.2463575304,-0.2494607121,-0.0017767396,-0.2310018092,-0.1340512782,-0.3148827255,0.0192815457,-0.018374715,-0.0441296548,-0.0066700438,-0.2820155323,0.2964689136,0.0416831449,-0.5087098479,0.2796152532,-0.0001253978,0.0176601093,0.165543437,-0.1627589613,-0.3021452427,-0.2690975666,-0.5443492532,0.003459824,0.2009383291,-0.2457512021,-0.3536373079,0.2595878839,-0.0016623527,-0.2429096997,0.3208558857,0.4414460063,0.0885331482,0.3402518332,-0.2983184457,-0.0976915359,0.1007717326,0.1736653894,-0.2127652615,0.1006111056,0.3349828124,-0.1839226484,0.066155158,0.4312537313,0.115748547,0.0765572712,-0.0095213763,0.0850534961,0.192847997,-0.5020107627,0.2259553075,0.4908399284,0.1194553599,0.0117059685,-0.461663574,-0.3616985679,-0.0677119046,-0.1615849137,0.2391583323,0.2563347518,-0.3226281404,0.0545236282,-0.3806765079,0.0091280751,-0.1806761771,-0.0156914108,-0.5306880474,0.3866957426,-0.0629000738,0.0666511431,-0.234006092,0.0058015864,-0.0944485217,-0.2278283834,0.106599234,0.1917997748,-0.4877149761,-0.0786596537,0.2468911409,-0.1749946177,0.0170607883,-0.1575658619,0.1409419477,-0.0101393433,-0.2379292548,0.1809385568,0.4561161399,0.371773392,0.3836920559,0.1170968637,0.1253914237,0.024882203,-0.1779976189,0.068300955,-0.0132526876,0.104945302,-0.173080191,0.0444740392,0.3815128803,-0.3759962618,-0.5460976362,0.1028343514,-0.3343901634,0.0862018541,-0.1280481815,0.4136289358,-0.2671588957,0.323705703,-0.1753737032,0.484808147,-0.2423684746,0.1622779816,-0.1694838703,0.0083823279,-0.2954889536,0.0338302106,0.1820193827,0.0345722064,0.1677876711,0.0791904926,-0.0682363361,-0.0843573734,0.1491902024,-0.29161641,0.5289235711,0.2174315751,0.1086589694,0.1371990144,0.3113140762,-0.1387363672,-0.0306898654,0.4569903016,-0.1073429361,-0.0885391235,0.1621474773,0.1351391673,-0.4702080488,0.0000834229,0.0173491593,0.3994839787,0.2739620209,0.0624445751,0.1275006235,-0.1487631649,-0.0352493897,-0.1899267733,0.205567047,0.158592999,-0.3715784848,-0.1733774841,0.3192159832,0.0479408093,-0.1288811713,0.0926913023,-0.0432998426,0.4183746576,-0.2419098616,0.0545514785,0.1526187062,-0.6342750192,-0.0166913364,-0.2027662247,-0.0080077397,0.3960766196,0.0161043759,-0.1464105844,0.1910933554,0.2044422776,0.2436323762,-0.0492489189,-0.0163895898,-0.101412721,-0.2839249671,-0.0278917216,0.2803300619,0.0384866111,0.2653223872,-0.0381516963,-0.1428515464,-0.0613704324,0.2356456667,-0.0408080593,0.0899868831,0.1494654715,0.4336160123,-0.0053794263,-0.0390727557,0.0198617131,-0.2733375728,0.1653865129,-0.0145297125,0.439550519,-0.0143588698,-0.2540389895,-0.2323454916,-0.1023680419,-0.4545465708,-0.5898591876,0.0748970881,0.014747561,-0.0367927887,0.0872119442,-0.1551977843,0.0207577292,0.0445901267,-0.1249775887,-0.3414300084,0.3159354031,0.058935713,0.028890267,-0.3574270606,0.3923297524,-0.0711159855,0.0870083347,0.1109094843,-0.0962955803,0.1215808392,-0.0367224552,0.0171239562,0.3487404585,0.3160446286,-0.1571756154,-0.3339691758,0.2780619562,0.1288728863,-0.2391919643,-0.5651096702,0.5706455112,-0.015483479,0.4227933586,0.1110909656,0.0158536006,0.0158130974,-0.3055428267,-0.2595428228,-0.1548806876,0.0180579703,-0.2967331111,0.1731819212,0.1353041381,-0.3907076418,0.1122884527,0.4177013934,0.0491052642,0.2202721685,-0.2979328632,-0.0175317377,-0.0396430343,0.1566833109,-0.0367524624,0.4276127815,0.1694587767,0.2256353348,0.0127891097,0.2090923935,-0.2430087328,0.1855034232,-0.0402710624,0.2182824463,0.4549237192,0.354709506,-0.1246863976,-0.1934615076,0.2973785102,0.0703052655,0.138393566,-0.3260527849,-0.2424041778,-0.1273448914,0.0182369594,-0.1301850528,0.0313257873,0.0962652564,-0.3991000056,0.3158450723,0.0024841891,-0.3320240676,0.2252082974,-0.1030021459,0.163310796,-0.0388344564,-0.3680047691,-0.0482127853,-0.4899754822,-0.2169424444,0.1613847911,-0.1952280253,-0.1368522346,0.1412716061,-0.1272917092,-0.531288743,-0.1544178277,-0.5428310037,0.1698828191,-0.222117722,0.3141136169,0.1305209547,0.4695051014,-0.3395379484,-0.0461444035,0.4725041687,0.0025181998,-0.4022099972,0.0161709189,0.1601449996,0.0788671598,-0.1604304463,-0.7739167809,-0.0642809197,-0.3372270763,0.286716193,-0.1638684422,0.1809030026,0.448207885,-0.0710506216,-0.0587882176,-0.2128799707,-0.1925424933,-0.1631088406,0.0021028754,0.1418824196,-0.0721007437,-0.3668020666,0.007778734,-0.2205589116,0.3654289842,0.093249701,-0.400913775,-0.410728097,-0.0315597616,0.4304032922,0.2287753969,-0.0975355133,0.2679589093,-0.0008596504,-0.0774438679,-0.2450527549,0.0018970593,0.5727344155,0.3348303735,0.0469684862,0.1178127527,-0.0271696839,0.3268133998,0.5415535569,0.3504087329,0.3235523701,0.1822449267,-0.2408680916,0.3834410012,-0.2624072433,-0.3972393274,0.2047518343,0.1588831544,-0.1285179257,0.2511807084,0.0158952698,0.3036919236,-0.2424140722,0.1569146812,-0.5188209414,-0.2767140269,0.1316638589,-0.3602773845,0.104478322,0.1489859521,0.1482095569,0.096497573,-0.0558830388,0.0600434244,0.0161809996,0.2040061802,-0.0765060708,-0.5741508603,-0.2322176546,-0.2483025342,0.1044028848,0.049725242,0.0934292674,-0.1396410763,-0.2572351992,0.2230919451,0.0743572116,0.98418957,-0.1124546453,-0.0223980099,-0.0163532253,-0.1371794492,-0.3577502668,-0.0575903431,-0.4521742165,-0.0635492802,0.1953746974,0.4829484522,-0.2719297707,-0.1840404719,-0.0610236786,-0.0780666694,-0.1377324462,-0.0551752038,-0.2825162709,-0.2471379936,0.0243515037,0.1792369932,0.1969057471,0.4291003942,-0.2604786456,-0.3233055472,-0.1244604439,0.1280684024,0.3246284425,0.3263876736,0.2666051388,-0.0662712157,0.0842259005,-0.0825690106,-0.0721119866,0.0257791542,0.5085005164,-0.1403199285,-0.0239045154,0.2063492984,-0.2590042651,0.4047759771,0.3651055098,0.1447870433,0.1355725378,-0.2059096694,-0.018225139,-0.1303575933,0.0782125965,0.1164329872,0.0254862458,-0.0361099206,-0.3983120024,0.1516757458,-0.0035752435,-0.1881574094,0.2331043035,0.0684020147,-0.2754635513,0.0854391307,0.067401737,0.9151231647,-0.0583753288,0.3912844658,0.6253470182,0.1934918016,0.3975296617,-0.0712790564,0.2521809936,-0.2924403846,-0.1937761903,0.0185146369,-0.1295040399,-0.0091074528,0.3428021073,-0.2154778987,0.5422403812,0.0139141409,0.5060142279,-0.0369393528,-0.3477807343,-0.0369482115,0.2454443276,-0.6066042781,0.0199462138,0.0903159231,0.4494953454,-0.1545343846,-0.10682071,-0.3858297169,-0.1560795158,-0.0228127223,0.2315672338,-0.4061106443,-0.0109606497,0.3604068756,-0.2347042859,-0.2855452895,0.4578832686,-0.0436796322,0.200664565,-0.2060175091,0.1875185072,0.0958188698,0.320232898,0.2116155922,0.2297783047,0.3145462573,-0.1325013489,-0.4006197751,0.1819072962,-0.2952146828,0.1937138885,-0.2160972506,0.3912729621,0.117652081,-0.2437204123,-0.4899375737,0.0495038852,0.2228689194,-0.2034307867,0.0622597039,-0.1620174795,-0.2052844316,0.2987907231,0.0618486665,-0.3066736162,0.1303577721,0.2504441142,0.2714741528,-0.0283011291,0.7059748173,-0.0800249279,-0.0637375116,-0.3099608719,0.0706003085,0.2637481093,-0.1664155424,0.3383428454,0.1393623501,-0.1203225553,0.0208787397,0.495787859,0.0764520913,0.1481766105,-0.1019191518,-0.434078455,0.0048647197,0.3871413469,-0.040783748,0.2559140921,-0.0730892941,0.0107484972,0.20003362,-0.171774745,-0.284694016,-0.2280094773,-0.3947721422,0.4053123593,0.020616509,0.0340003371,0.2809011936,-0.1246354431,0.1063427106,-0.3126686215,-0.3165102601,-0.2303075492,0.0233013853,0.1947916299,0.1834421307,0.1439168155,0.0407313667,0.0159107391,-0.0184927769,-0.2311230302,0.3753026128,0.3558829725,0.0073424689,0.2380220294,0.0011619339,0.3851644695,-0.1636802703,0.0180251598,0.0536325239,-0.1786247641,0.2189986259,-0.1234062761,0.0861577317,0.0103212381,0.1456641406,0.1310774684,-0.0173463672,-0.070716545,-0.0687456951,-0.2028094381,0.0087885531,-0.0472328104,0.1345523894,0.2589514554,-0.1851588488,0.2671405673,-0.0103208879,0.2198467851,-0.3239837885,0.3387541175,0.1191086248,0.0781588927,0.1221268177,0.1845982224,0.1412796676,-0.3834759295,0.3110426068,0.2137459666,-0.153291449,-0.3943413794,0.270372957,0.4016449451,-0.0494834073,0.0024952334,-0.0100786518,-0.2710435688,-0.2409231216,0.0042643473,0.0656297654,0.3041697741,0.0251283199,-0.1227378175,0.1586383432,0.0234342944,-0.0066123973,0.0212233197,-0.3439193368,-0.0287951156,0.1235759333,0.2682903707,-0.0385910384,-0.2265938371,0.1016585305,-0.0892059207,-0.1853413284,-0.1691816002,-0.0796572939,-0.2356610596,-0.016027512,-0.1868152022,0.0669320822,-0.3150311112,0.4789046347,-0.0453367382,0.0214859452,-0.3790132105,0.0813250616,0.3750403821,0.3184185922,0.089819096,0.5027284026,0.2552713752,0.0881296322,0.1871246696,0.6208726764,0.3228026032,0.4014385343,-0.1718603075,0.0328344181,-0.0612051412,-0.3069051206,-0.0683343709,0.2621795833,0.2224468589,-0.0030690499,0.3125353158,0.1099693477,-0.075420469,0.3012111485,0.1460101753,-0.0650471225,-0.1963528693,0.2436033338,0.0570653751,-0.0841096342,0.0970196798,-0.0905789509,-0.5747677088,-0.1879495829,0.0915049687,0.1847905666,0.271912396,-0.0292702205,0.0529377274,-0.0763235688,0.0537510403,0.1515025496,0.1362342834,-0.3360588253,0.2409199923,-0.4618490338,0.2489724904,0.0628399923,-0.0488863736,-0.2233596742,0.1597939879,0.0144211017,0.0276804678,-0.1524353772,-0.1517508328,0.3210305572,0.0558781661,-0.6173195839,0.0694772005,0.2226677984,-0.0316431783,0.0072421888,-0.3014624417,-0.0318908021,0.0041656755,-0.1023189798,0.1885470748,0.1236172244,0.0946243778,0.055021178,0.07967107,0.2517285943,0.653708756,-0.0398855209,0.0944277719,-0.0759158581,-0.185772717,-0.1136369109,0.0784753412,0.1745887995,0.2737561762,0.19055067,-0.2834953964,0.147964105,0.1163090467,0.1791380793,-0.1351221502,0.071369037,0.1045638844,0.1388771981,-0.0669206157,0.0154684829,0.0287789684,-0.0240659416,-0.0409313217,-0.2523094118,-0.5186450481,0.3893420398,-0.5014826059,-0.3424704969,-0.2832122147,0.1933733374,0.1038309261,-0.1890992075,-1.0374089479,-0.1832494736,0.3999889195,-0.1623470336,-0.0257816911,0.2184678912,-0.2688953876,-0.1232359186,-0.1669478863,0.0554085001,0.0288154203,-0.1088837981,-0.1726090759,-0.4612897933]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/801","title":"How to join two datasets?","comments":"Hi this is also my question. thanks ","body":"Hi,\r\n\r\nI'm wondering if it's possible to join two (preprocessed) datasets with the same number of rows but different labels? \r\n\r\nI'm currently trying to create paired sentences for BERT from `wikipedia\/'20200501.en`, and I couldn't figure out a way to create a paired sentence using `.map()` where the second sentence is **not** the next sentence (i.e., from a different article) of the first sentence.\r\n\r\nThanks!","comment_length":7,"text":"How to join two datasets?\nHi,\r\n\r\nI'm wondering if it's possible to join two (preprocessed) datasets with the same number of rows but different labels? \r\n\r\nI'm currently trying to create paired sentences for BERT from `wikipedia\/'20200501.en`, and I couldn't figure out a way to create a paired sentence using `.map()` where the second sentence is **not** the next sentence (i.e., from a different article) of the first sentence.\r\n\r\nThanks!\nHi this is also my question. thanks ","embeddings":[0.0086422656,-0.1031597331,-0.0017199666,0.1256665438,-0.0711243749,0.1530651003,-0.0795226172,0.1573186368,0.0155739533,-0.1239238977,-0.2840072513,0.2373536229,0.2954332829,0.027612282,0.0667384341,-0.1882848293,0.3172922432,0.1062992066,-0.1907269359,0.0525125563,0.0805284157,0.1282547265,-0.2607732415,-0.0547854267,-0.2916154563,0.1160952151,-0.1048387736,-0.1895434856,0.0183501914,-0.1205204129,0.2118845433,0.2474035025,0.0552277081,0.5249673128,-0.0001241543,-0.1600589305,0.1300558597,-0.2044917494,-0.026994599,-0.5992218852,-0.4386677146,-0.1244255006,-0.0684446245,-0.2696658075,-0.1819266677,-0.2206917107,-0.0206239726,-0.3450590968,0.3431601822,-0.0077251904,0.0235067103,-0.2668405175,0.3433602452,-0.237370953,0.2798164189,0.0688903332,0.2601788342,0.2782798111,0.3661361933,0.0107132811,0.368961513,0.0006436061,-0.3020179272,-0.1312525272,0.2876220345,0.0624770001,-0.2066230327,-0.5621252656,0.1306302547,0.5374888778,0.8537281752,-0.1552707702,-0.0969372988,-0.0561281629,0.2328011096,0.2921544015,-0.125739947,0.0982352197,0.1026913971,-0.0487832986,-0.0958057046,-0.3874238431,-0.0362189971,0.0918626711,-0.1444103122,0.3872371316,0.1295877844,0.114217937,-0.1108001098,-0.4039044976,0.0344295017,-0.1002297625,-0.2110011876,0.3856729567,-0.1243759766,-0.3320113122,-0.4771630168,-0.1038377658,0.3813834488,-0.268133074,-0.096886538,-0.1609976441,-0.2692092359,0.2105233967,0.3882365525,-0.1755322963,0.4047798216,0.2322459966,-0.1577465534,-0.2801693678,-0.7673329115,0.0171409436,0.057970874,-0.3421472311,0.0242968313,-0.1243067756,0.0199182406,-0.1801628172,0.1250840276,0.1007327884,-0.5118473172,-0.1183593422,-0.5240088701,-0.0114412205,0.0239154976,-0.1450465769,-0.0497147925,0.315967679,0.0395556204,-0.1686112732,0.0404199921,-0.0252320766,0.0633670539,0.1321119219,-0.3469753861,0.0445857123,0.1961617321,0.2331966162,0.1481040269,-0.1694045365,0.1197610348,-0.3020023704,0.137069881,-0.2697345316,0.2602796853,0.3120116591,0.0390945189,-0.2776152194,-0.0857516006,0.3248585463,-0.399923265,0.0864213854,-0.0653598607,0.0181378275,-0.0871320739,-0.2783280015,0.1598998606,0.3460610509,0.2326325476,-0.3940080702,-0.0587631725,0.0874176547,-0.4775606692,-0.1881870031,0.109285675,0.0678342357,-0.5827179551,-0.2335524559,0.2393262535,0.1420500875,-0.1750497967,0.3611309528,0.0280942861,0.471137315,0.0110191032,0.0962039679,0.2799146473,-0.0416910127,-0.2332488149,0.3025940061,-0.1320319921,-0.1147474274,-0.0706467256,0.3762509227,0.3382355869,-0.1347243339,0.3099463582,0.1458836198,-0.1441027671,-0.21162875,0.1479812562,-0.2224194854,0.128157407,-0.3557189405,-0.0910582393,-0.0409252569,0.1389116794,0.0660104901,0.187409088,-0.1573628634,0.17595011,0.3667401373,-0.0178749319,0.3519771099,-0.0946980268,-0.3118662238,-0.2776841819,-0.0857054666,-0.2087226212,0.1330672801,0.2850875854,-0.2510748208,-0.2248070091,-0.4016153812,0.3682670593,0.0408781432,-0.0449785553,-0.2483270466,-0.0694530681,-0.0517402366,-0.0536402278,-0.0219240282,0.2495816797,-0.1454919875,-0.2651919425,0.3668872714,0.1355704069,0.0374587886,-0.1235461608,0.3267448545,0.082300216,0.1811544746,0.1703365147,0.0001242725,-0.1772167981,0.3313055038,0.2264990658,-0.2807731032,-0.0298117772,-0.6336616874,-0.1286899298,-0.2222702056,-0.0479822569,-0.3042877316,-0.2024027407,0.1808590144,-0.0077022715,0.2465722114,0.0540199876,0.027798336,0.1328593194,-0.0343166925,-0.1086187139,-0.1904694289,0.0041406448,0.1025782451,0.1617056876,0.4100299776,-0.2182912081,0.0109193111,0.0501033776,0.0643859804,0.148682192,-0.3202102482,-0.4482902884,-0.0348510742,-0.2666668892,0.0605816729,0.197446391,0.2928033173,0.132996425,0.072188668,-0.0502494723,0.0031277109,-0.0175338835,0.0059488295,0.3809412718,0.031369675,0.191565305,0.0719103813,0.3916142285,0.2183511555,0.0273338035,-0.1882561743,-0.1368374676,-0.2739343643,0.0664582625,-0.2779691815,-0.3786148429,-0.1660847366,-0.2059377879,-0.0985172167,-0.0661770254,0.1446314156,-0.0018049927,0.2189115137,0.0251739416,0.0922363102,-0.1496050656,-0.0921243131,0.1309968084,-0.4827658534,0.2758358717,0.1367046237,0.1106621698,0.1821271032,0.4315045476,-0.2377767563,-0.2119956911,-0.4708075225,-0.4882440269,0.0271523632,-0.0832055658,0.2464647889,-0.1378098428,0.1196589693,-0.2433224618,0.0710356534,0.2020082623,0.2868791223,-0.0123796556,-0.3474816382,-0.1886772662,0.1141239777,0.2969013453,-0.3524129689,-0.1484709829,0.0310868844,0.1545841098,-0.0145354029,0.02892318,-0.5503239036,0.0835937411,-0.1618026942,-0.1571054757,0.0197177082,-0.0626593307,-0.4202101827,0.2303313464,-0.0485292152,-0.2263950258,0.180537641,-0.0165021271,-0.0026643169,-0.1699916124,0.0912156329,0.1795441955,-0.1564996243,0.017315222,0.2109203488,0.1189802215,0.1186756268,0.1498249322,0.0817559659,-0.1987539381,-0.1539820284,0.2602343857,0.0678912923,0.1008096039,0.0394634157,0.191930443,-0.3098877072,0.4737738669,0.6141434908,0.076026544,-0.1159236282,-0.1300909966,0.2858804166,-0.2061953843,-0.1062206551,-0.2308024019,0.2120374292,-0.0494063422,0.2222881317,0.1095358506,-0.2075289786,-0.228347376,0.0117431115,-0.3206907809,-0.0554457232,0.1207911745,-0.3535489142,0.2410795689,-0.0327361003,-0.5159078836,-0.2512241006,-0.2810396552,-0.1688420475,-0.2948459685,0.3148240149,-0.0221113469,-0.3921690285,-0.0365405306,-0.2824001312,0.253818512,0.40655002,-0.0717705116,-0.0816280395,-0.0252755899,-0.1406655312,0.2328858376,0.1659705937,-0.5191042423,-0.1185200289,0.1043717265,0.3830145895,-0.0816202164,0.3008957803,-0.1140013561,-0.1145123318,0.2535797656,0.29868415,-0.2875777781,0.1268176287,0.2329123169,0.3936496079,-0.206752345,-0.1619843692,0.2041006386,0.1742919534,-0.2514653206,-0.1059252098,0.1992509514,-0.0022914794,0.3379429579,-0.1705728471,0.0395038724,0.2164407223,0.504206419,0.1598282009,0.1012154445,0.056913808,0.2088782489,0.3993816078,-0.0731377527,0.1246046945,0.6548756957,-0.206765458,-0.3234260678,-0.2453838438,-0.1485227793,0.603535831,-0.0066761998,0.326483041,0.2336768359,0.0155491866,-0.1053068638,0.0016642222,-0.2192119807,0.103212446,0.1787758321,-0.7175309658,-0.4144426882,0.2319056392,0.3042742014,-0.364851594,-0.2006314248,0.0711929128,-0.2089423835,0.4880278409,0.0587127022,1.228703618,0.0238937158,0.1136939451,-0.3191665113,-0.1325744987,0.964812398,0.2026370764,-0.1596537083,-0.1462640464,-0.0730420351,-0.276109457,-0.2371081263,-0.1339363754,0.5694881082,-0.2063186914,0.4062193334,0.0868064538,0.1637608707,0.1207803562,-0.0000894385,0.5442727804,0.0194833409,-0.0250551403,-0.0180922449,0.121000737,-0.1127625257,-0.169895038,-0.045424033,-0.2727007866,0.0906821862,-0.316078186,0.0326010846,0.0298692528,-0.0137567949,0.0835134685,-0.1703158617,0.1221235096,0.7723956108,0.3151183724,0.223227486,-0.2482532561,0.1493998617,-0.0061447909,0.0168300085,0.144923836,-0.1038145125,0.2543903589,0.137360692,-0.3173004687,0.2153636664,0.1100396067,-0.0935541764,-0.4201115966,-0.0133239152,0.1026447043,-0.1006449908,-0.565748632,0.339728266,0.1132969856,-0.1969403476,-0.030613564,0.001103805,0.0554434136,0.1212077141,0.0847899243,-0.3526939154,0.0882991627,0.4132322371,0.3616730571,-0.0235290378,0.3309416175,0.2952297926,0.1101936921,-0.0820872337,0.0800176635,-0.0253093988,-0.4209328592,-0.169740811,0.1466586739,0.1951458752,-0.3111126423,0.0009082233,0.39122805,0.2568185329,0.1717341691,-0.1205009967,-0.3020161986,0.3099752069,0.161845088,0.1732452661,0.1090029851,0.0675505474,-0.0143959438,0.5571688414,-0.1639313996,0.1643266827,0.0790718123,0.2311948836,0.5408015847,0.3541598916,-0.357722193,-0.3111326098,0.0362372994,-0.1955042332,0.319306612,-0.1476972997,0.0221149772,0.1934092343,0.0916572064,0.0138143096,-0.0608714558,-0.241854623,0.2942898571,-0.3047936261,0.0784801915,0.2183600068,0.2359467596,-0.2006735355,0.3163810372,0.1476241201,-0.6093301177,0.3690555394,0.1104672402,0.1390142441,-0.3326762617,-0.2126074582,0.2531023324,-0.0996456295,0.2612755001,-0.0456405282,0.3477787375,0.024453314,0.0170433298,-0.0017798028,0.0083928695,0.2284443527,0.2729088962,-0.2444421202,0.0104264943,0.1001653075,0.1599237323,0.0806898177,-0.2722022831,-0.0695912018,0.5358014107,-0.077158682,-0.0319123603,0.0377199575,-0.1588039547,0.3696362972,-0.230673492,0.0490114614,0.0963440314,-0.1534030139,0.1248424649,0.3797160387,0.5193222165,-0.0463134907,0.038142994,0.1963178366,0.2385991961,0.271928072,0.2435814887,0.1333655417,0.5227660537,0.1349106133,0.6452955008,0.2325282991,0.2036983669,0.3884786665,0.2866891026,-0.2723762691,-0.0315466635,0.1465769857,0.172004357,-0.2927004397,0.1405534446,0.4561542273,0.0107864775,-0.2749221623,-0.0954526886,-0.3910925984,0.0870933011,-0.0577321798,-0.0810335875,-0.2759215832,0.2165053487,-0.0792778805,0.1875716597,-0.10843651,0.0230926797,-0.0766615868,0.0501497649,-0.2299422175,0.2903550863,-0.259270668,-0.0517148636,0.3107283115,0.6606180668,0.2608997822,0.1323453486,-0.0639328435,0.3687167168,0.0414149798,-0.1542628258,0.1616036147,-0.0336610265,-0.0599313378,-0.5259224176,0.2316687703,0.1365369558,-0.031260781,0.147347033,-0.0189385712,0.3329987526,-0.0503992811,-0.1576759815,-0.2462496907,-0.1383473128,0.0546116382,-0.0360702798,-0.3028747141,0.0585352406,0.0212423336,-0.1802733094,0.043890655,0.0455232747,0.0288159847,0.031202253,0.229914546,0.4334386289,0.1651583761,-0.2036474347,-0.2191979438,-0.5811525583,0.1105577648,-0.1987785101,0.05822438,-0.4460592866,-0.051602751,0.2955670357,0.1844227463,0.180706203,0.0409898311,-0.2941391766,0.1627306491,-0.149999544,-0.3005218208,0.1340182275,0.0533790365,-0.1266585737,0.0253766123,0.039991945,0.0776743293,-0.1486878097,0.0291925538,-0.1557798535,0.1156280115,-0.1833212674,0.0761142597,0.1654583216,0.7683653235,-0.3764029741,0.1086057425,0.5397925377,-0.3010936081,0.0065515181,0.6301190257,0.081260629,0.2191247046,-0.3869260848,0.1559415758,0.2556770742,-0.4890652597,-0.3831445873,0.2920721173,-0.6273818612,-0.1828433722,-0.0717921257,-0.0253555104,-0.0694948211,0.2531303465,-0.0198190287,0.2582797408,0.2272818387,-0.1414998621,0.1035178304,-0.1454102993,-0.2328959852,-0.2178673893,-0.0500617437,0.2873121202,-0.0637888312,-0.3062477112,0.0970285386,0.4636545181,0.1179197878,0.0729727596,0.1533372849,-0.0911429748,0.0488585941,-0.2329123765,0.1791358739,0.0205669496,-0.0396443419,-0.4858398736,-0.3726804256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/801","title":"How to join two datasets?","comments":"Hi ! Currently the only way to add new fields to a dataset is by using `.map` and picking items from the other dataset\r\n","body":"Hi,\r\n\r\nI'm wondering if it's possible to join two (preprocessed) datasets with the same number of rows but different labels? \r\n\r\nI'm currently trying to create paired sentences for BERT from `wikipedia\/'20200501.en`, and I couldn't figure out a way to create a paired sentence using `.map()` where the second sentence is **not** the next sentence (i.e., from a different article) of the first sentence.\r\n\r\nThanks!","comment_length":24,"text":"How to join two datasets?\nHi,\r\n\r\nI'm wondering if it's possible to join two (preprocessed) datasets with the same number of rows but different labels? \r\n\r\nI'm currently trying to create paired sentences for BERT from `wikipedia\/'20200501.en`, and I couldn't figure out a way to create a paired sentence using `.map()` where the second sentence is **not** the next sentence (i.e., from a different article) of the first sentence.\r\n\r\nThanks!\nHi ! Currently the only way to add new fields to a dataset is by using `.map` and picking items from the other dataset\r\n","embeddings":[-0.0807135701,-0.0231983922,-0.0087042591,0.0758321956,-0.0804641396,0.1938629001,-0.0964456275,0.1502997428,0.0917028189,-0.1372491866,-0.156401813,0.2632109523,0.2652062774,0.0553944521,0.0318246484,-0.1623478681,0.2753705978,0.1366879791,-0.1497536302,0.0182720907,0.0281484146,0.0739849955,-0.1760732085,-0.0964749157,-0.2948448658,0.0638359413,-0.1693757772,-0.2055224776,-0.007480918,-0.1788524538,0.1081667691,0.2946743965,0.0237622503,0.5012590885,-0.0001200251,-0.182182014,0.148232311,-0.1188671067,-0.0978056639,-0.5917367339,-0.4379109144,-0.1937709898,-0.0372159071,-0.2641343176,-0.1709364951,-0.2597176135,-0.0092760064,-0.3962280154,0.1879961193,-0.0077911085,0.0517811179,-0.2181003094,0.3178336918,-0.2899117172,0.368132025,0.0600768179,0.231391415,0.204635486,0.4667863846,0.0924607143,0.35362342,0.0318278708,-0.2515158951,-0.1778289527,0.2656972706,0.1065583304,-0.2725177407,-0.5039494634,0.1312267184,0.5253686905,0.8458815813,-0.1958146393,-0.2037546039,-0.1436982155,0.2615859509,0.1677378267,-0.1626797467,0.065555878,0.0571829155,-0.1085786298,-0.0172955748,-0.4783712029,-0.0146891614,0.1460682899,-0.0518698655,0.4198278785,0.1197788194,0.1123552471,-0.1201640666,-0.3536275029,-0.0021790173,-0.1028894931,-0.2314326614,0.3964161575,-0.1357270479,-0.3579379618,-0.4266589582,-0.207686156,0.2877579927,-0.2325144559,-0.109399341,-0.1312533915,-0.3093343377,0.2262607515,0.4135226309,-0.1765180826,0.4025959671,0.1679807901,-0.1497799158,-0.2919568717,-0.6388095617,-0.0956266075,0.0734440163,-0.2497820556,-0.0123184929,-0.1612832695,0.123229228,-0.0954735726,0.1156069934,0.1166566536,-0.3803674877,-0.1796399355,-0.4582118988,-0.0172274187,0.0473739319,-0.1791682541,-0.0947115719,0.3156735897,0.0625576377,-0.2051766515,0.0225611832,-0.0225598402,0.0753076598,0.1359440684,-0.3133123219,0.0678401217,0.2115236372,0.2421742678,0.1165313795,-0.0914426222,0.1404797584,-0.3214638233,0.0882012248,-0.2670930028,0.2662129402,0.3648284078,0.0332867876,-0.3661639392,-0.077468127,0.3192604482,-0.3424720168,0.0919076502,-0.1811608523,0.0503414199,-0.0347732008,-0.3690575063,0.1503404826,0.4734448195,0.2010741085,-0.379218936,-0.0435686931,0.1888428777,-0.4432769418,-0.1435356885,0.0995916724,0.0958330482,-0.610889852,-0.246776849,0.130536139,0.18384628,-0.2199200839,0.2735758126,0.0045823003,0.4709973633,0.0193588324,0.0979583412,0.2940160334,0.0065495647,-0.1389174908,0.238259837,-0.061020188,-0.1116313562,-0.0856147856,0.3941873908,0.3761968315,-0.1148844287,0.2615608871,0.1333845556,-0.205403015,-0.2404527366,0.1724509895,-0.2500698864,0.2003563643,-0.3064526021,-0.0209771879,-0.0451242886,0.2041331381,0.0940571874,0.1203064397,-0.2302704453,0.2388392687,0.361251533,0.0425797552,0.4038475156,-0.1544989944,-0.3849770129,-0.357668519,-0.0640172362,-0.1553385556,0.2091977,0.3712136447,-0.2315183133,-0.1918838769,-0.3699703515,0.3332959712,-0.0095565738,0.0069703101,-0.1931199729,-0.1151965261,-0.1268462539,-0.1134818792,-0.0729124695,0.2275792062,-0.1046024486,-0.2657743096,0.27956146,0.1391690969,-0.0043341462,-0.249317944,0.3853373528,0.08726307,0.1868847311,0.1661546379,0.0304406304,-0.0998971909,0.3716405332,0.249361068,-0.2330416441,0.0095545622,-0.5110837817,-0.1398496926,-0.2362093776,-0.0440366454,-0.265645206,-0.2391246706,0.1430874914,-0.0139346393,0.2417303473,0.1318977028,-0.0250237696,0.1806058735,-0.0407013409,-0.0761865079,-0.2563690543,-0.050618533,0.1348457336,0.1929733455,0.3609315157,-0.2296914309,0.0628190935,0.0342490375,0.0914989114,0.1899907589,-0.2439336926,-0.4683243334,0.0066163177,-0.2882340848,0.0960114896,0.1645246297,0.3466561139,0.1398829222,0.0730472878,-0.0421125665,-0.0236181002,0.0177689642,0.0443218946,0.3303335607,0.109584108,0.256780237,0.0742078945,0.3293254972,0.2420234382,0.1224824488,-0.1384059936,-0.1348122507,-0.2841987014,0.1067584902,-0.2484176904,-0.4215637147,-0.1875657886,-0.1845408529,-0.1230121776,-0.0149282655,0.1865074635,-0.0561988056,0.1403868496,0.0554290675,0.1559760273,-0.1527743936,-0.0191225801,0.1331458986,-0.5160618424,0.2592284381,0.1565908492,0.1139577702,0.1209732369,0.472194165,-0.1159626096,-0.2033684999,-0.5184364915,-0.5585831404,0.0389996357,-0.1104806066,0.2622460127,-0.1279797405,0.1481089741,-0.1493074,-0.0289613381,0.2102096677,0.2862325907,0.0190500263,-0.3370529115,-0.1956279725,0.0766984373,0.259927839,-0.4277597368,-0.180271104,0.0028607389,0.187552616,0.0680031925,0.0265829805,-0.4053646326,0.1125639677,-0.122062929,-0.1676352471,0.0574354641,-0.0919366777,-0.3873731196,0.2161987722,-0.0784542337,-0.2595170438,0.1928527057,-0.0159029327,-0.0558186658,-0.2051000446,0.1267775893,0.1307677478,-0.1578683704,0.1032423452,0.2249044776,0.1547768265,0.1833477914,0.1117331535,0.0685983002,-0.1756863147,-0.1481736153,0.268507719,0.0928248242,0.1118733361,-0.0606018007,0.2222197205,-0.3472218812,0.3986665905,0.5254830718,0.1192845628,-0.059492372,-0.2288450301,0.3587059081,-0.245824635,-0.105610311,-0.2335343659,0.1408099234,-0.1052911878,0.1787989587,0.0762337968,-0.1367232054,-0.1691931933,0.080647558,-0.2816328704,-0.0469959825,0.1670812964,-0.4001509845,0.2265855819,-0.0490962341,-0.489815861,-0.3174824119,-0.3232747018,-0.1970895231,-0.3505018651,0.2054466158,-0.0804451406,-0.4070578218,-0.0595487803,-0.2789439559,0.1943852007,0.3742375374,-0.0190009251,-0.0954905748,-0.0674528405,-0.1392037272,0.2143233865,0.1779632866,-0.542999506,-0.2740229964,0.0634253621,0.3556329012,-0.0907816738,0.2844603658,-0.1493005604,-0.1365783364,0.2212078422,0.332383275,-0.2749550641,0.1516233981,0.2588717043,0.3530679345,-0.2272440642,-0.0717946365,0.1638049036,0.1854703724,-0.3372766078,-0.1539022028,0.1404724121,-0.0018162654,0.31705603,-0.1068641618,0.101103425,0.2996177673,0.4610784054,0.1862187684,0.0585753173,0.12902385,0.196083501,0.3700777888,-0.0976310819,0.0814772323,0.6518191099,-0.2090458572,-0.3070388138,-0.18773368,-0.2284006476,0.6691730618,-0.0614032559,0.368329227,0.3365754485,0.0028094286,-0.0528609492,-0.1312807649,-0.1078722551,0.1989752501,0.1381972283,-0.615201056,-0.449105829,0.2315099537,0.1934190094,-0.3068053126,-0.1744422913,0.1185641065,-0.2160873264,0.584150672,0.0494881049,1.0757403374,0.0293671861,0.0874434263,-0.3162237108,-0.1883552074,1.0178573132,0.1305323988,-0.1236894429,-0.1478768885,-0.0850096494,-0.229445681,-0.1666280627,-0.0572976619,0.5215896964,-0.2304558456,0.3305691481,0.1110762879,0.2486634701,0.1022545323,0.0712074563,0.4821222723,-0.0573531538,0.0192611795,0.0158687904,0.1044017449,-0.2079253942,-0.1785885841,0.0272899028,-0.2807316184,0.0825845152,-0.4025096893,-0.0166642722,0.1252090931,-0.0403472409,0.0672277212,-0.179091543,0.1475476623,0.6222158074,0.3884709775,0.2047899365,-0.2520731986,0.1811258495,-0.0499880016,0.0590853915,0.1429446191,-0.1582109779,0.2772467434,0.1772700846,-0.3260656297,0.2746420801,0.0858253911,-0.1454246193,-0.4181193411,-0.0514045842,0.1851321459,-0.1382945329,-0.5660066605,0.4074316025,0.160322994,-0.3235922158,0.0023532382,-0.0015480134,0.0556906164,0.0581135638,0.1035229862,-0.3034124374,0.0461766198,0.3858907521,0.3112258911,0.0276320744,0.4255210757,0.3808818161,0.1432446241,-0.0935774222,0.0510323793,0.0641199276,-0.44381845,-0.2619391382,0.2240642905,0.1107028052,-0.2755554318,0.0275216736,0.294808954,0.2812787294,0.1131932363,-0.1540003717,-0.3026463687,0.298366487,0.1990268975,0.1886807084,0.1341676116,0.1064022928,-0.0667069107,0.6051511168,-0.1955351084,0.1030364558,0.0947640613,0.2416863889,0.5612391233,0.3211012781,-0.3037019372,-0.3529649377,0.0524257496,-0.2563051879,0.3680343628,-0.165379703,-0.0343191735,0.1800291836,0.1391880661,0.0141354911,-0.1297420114,-0.2268852592,0.305467099,-0.3553519547,0.1036901325,0.2047757059,0.1460906118,-0.1658950895,0.2882985473,0.1458716094,-0.6236057878,0.3157991171,0.1428540498,0.156602934,-0.2186437398,-0.1664490551,0.2113189548,-0.0294229593,0.223419413,0.0287347212,0.4243158102,-0.0088737598,0.0212812666,-0.0118985604,0.0641594753,0.2644204795,0.2463218421,-0.2564184964,0.039067056,0.0752099976,0.1435414404,0.1195016429,-0.3249657452,-0.0525959954,0.513418138,-0.1326962262,0.0561384968,0.0678662062,-0.1298228204,0.3916148245,-0.2596837878,0.0113476729,0.0583179146,-0.1029904932,0.2119308859,0.4004094005,0.4970461428,-0.0868292525,0.0304796137,0.1863889992,0.2096036077,0.2692551911,0.2416104376,0.1177023202,0.5011899471,0.1195439771,0.5809908509,0.1392070353,0.3784672618,0.3143641353,0.3798806071,-0.3143441975,-0.039888896,0.1034483835,0.2028982788,-0.257453531,0.1483082026,0.4723989666,-0.0298810843,-0.297696352,-0.0858098119,-0.3873427808,0.0948563889,-0.0732982904,-0.0769011006,-0.3153764904,0.1335423142,-0.1397841573,0.1501650661,-0.1131890342,0.052401185,-0.0905737877,-0.0022759505,-0.2257418782,0.3338592052,-0.1965188086,-0.0932018384,0.2818794847,0.6687037349,0.1901691407,0.1966962814,-0.0468821228,0.4440394342,0.0426716805,-0.1884844601,0.1407414526,-0.054325372,-0.0936097354,-0.441822499,0.2229226083,0.1742852777,-0.0512589142,0.1346774995,0.0311518274,0.3035240769,-0.0694840029,-0.1697521061,-0.0649950504,-0.1368065178,0.0962802991,-0.0222407077,-0.3277292848,-0.0021431819,0.1096536443,-0.2466522902,0.047279492,0.0412832163,0.0517095849,0.0994184539,0.2816911042,0.4313210547,0.0839763656,-0.2127912194,-0.194523111,-0.571377039,0.0890259519,-0.1803854257,0.0644965023,-0.40974617,-0.0178543068,0.3409053981,0.1188748255,0.2114411891,0.0123330038,-0.3273979127,0.0921842381,-0.1394601911,-0.3116663694,0.1199031249,0.0357649885,-0.0921876803,0.085191004,0.0246659126,0.1429485828,-0.1299191862,-0.0156781785,-0.1433270574,0.1077952161,-0.2914768755,0.0785342008,0.0836480632,0.7063450217,-0.3609155715,0.0404663831,0.458216995,-0.3141142428,-0.0788453445,0.5984266996,0.0694171563,0.248877123,-0.3487597108,0.1082893685,0.2200300246,-0.4537163675,-0.3922596872,0.1436914653,-0.6623858809,-0.0504209697,-0.0922140107,-0.0260310154,-0.0347718,0.1968842447,0.0055903746,0.3021912873,0.2433598042,-0.1516219974,0.0402150154,-0.1162698045,-0.2251847982,-0.2587262392,-0.0455798022,0.240624845,-0.0868317559,-0.3463819325,0.0724972188,0.4907080829,0.110297136,0.0547937602,0.1335089654,-0.0682993233,0.0065599098,-0.2171505839,0.2488212436,0.0190879349,0.0077922279,-0.4408718646,-0.3627175689]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/801","title":"How to join two datasets?","comments":"Closing this one. Feel free to re-open if you have other questions about this issue.\r\n\r\nAlso linking another discussion about joining datasets: #853 ","body":"Hi,\r\n\r\nI'm wondering if it's possible to join two (preprocessed) datasets with the same number of rows but different labels? \r\n\r\nI'm currently trying to create paired sentences for BERT from `wikipedia\/'20200501.en`, and I couldn't figure out a way to create a paired sentence using `.map()` where the second sentence is **not** the next sentence (i.e., from a different article) of the first sentence.\r\n\r\nThanks!","comment_length":23,"text":"How to join two datasets?\nHi,\r\n\r\nI'm wondering if it's possible to join two (preprocessed) datasets with the same number of rows but different labels? \r\n\r\nI'm currently trying to create paired sentences for BERT from `wikipedia\/'20200501.en`, and I couldn't figure out a way to create a paired sentence using `.map()` where the second sentence is **not** the next sentence (i.e., from a different article) of the first sentence.\r\n\r\nThanks!\nClosing this one. Feel free to re-open if you have other questions about this issue.\r\n\r\nAlso linking another discussion about joining datasets: #853 ","embeddings":[-0.0025697404,-0.0501239039,0.0077043655,0.0776305571,-0.0018986185,0.1729735434,-0.059212938,0.1528545618,-0.0238949601,-0.1077590883,-0.2796118557,0.2284302264,0.3064365983,0.0511664487,0.0457657985,-0.2121218741,0.2977069616,0.0878454968,-0.139589116,0.0677665919,0.0877424926,0.1052053496,-0.2417863607,-0.0796104893,-0.3017899394,0.1389433295,-0.1065389886,-0.178644225,0.0257563423,-0.0823389664,0.2509201765,0.2338497043,0.0578636974,0.5224131346,-0.0001189344,-0.1573140323,0.1734097898,-0.1755609959,-0.06744311,-0.6412494779,-0.4305659533,-0.1681635231,-0.0630625188,-0.2254205495,-0.1610951871,-0.19150047,-0.0141816447,-0.377686739,0.2850196362,-0.0464441441,0.0730513707,-0.233841747,0.3569043875,-0.2776043713,0.2534075677,0.1100591347,0.229191348,0.251378566,0.405372262,-0.0196343306,0.3723791242,0.0401941501,-0.2859000564,-0.1112813205,0.2673867345,0.0482694805,-0.2433438003,-0.5618752241,0.1242505312,0.5158067942,0.8751185536,-0.1637456268,-0.1189346835,-0.1058145016,0.2280834168,0.2984703481,-0.1458121538,0.1257112026,0.0739404559,-0.0743975937,-0.0713713616,-0.3805038035,-0.0199619513,0.0570362061,-0.1911927462,0.3875505924,0.1164398417,0.083529301,-0.1151774228,-0.3857506216,0.0374712497,-0.0749263093,-0.1794656664,0.3205149174,-0.1373662353,-0.3005454838,-0.4265521467,-0.0406180359,0.3847048581,-0.2180030644,-0.1096995026,-0.1719014347,-0.2832714021,0.1812517643,0.4421011806,-0.170942083,0.3890824914,0.2038502246,-0.1573475301,-0.2631590962,-0.7858472466,-0.014169612,0.0385817923,-0.3469096422,-0.0002780605,-0.1453927755,0.055726245,-0.1759075522,0.0601257272,0.1142937168,-0.5267914534,-0.1042295024,-0.4912877381,-0.0284900256,0.0544901006,-0.1384902149,-0.0492657535,0.320995152,-0.0293860994,-0.2145489156,0.0017576166,-0.0458912812,0.032955382,0.092620641,-0.2935263813,0.0123675494,0.1479326189,0.2374411821,0.1795552671,-0.1296134591,0.1128565297,-0.2888922989,0.1297527105,-0.2324273437,0.2570161223,0.3389841318,0.0836734325,-0.2809276581,-0.0974410996,0.3202989995,-0.3467897177,0.0968274325,-0.0751114711,0.0482253209,-0.1077681482,-0.3144002557,0.1393895,0.3638130128,0.1911332011,-0.3920109272,-0.0591251329,0.0474587008,-0.4272734225,-0.1695651859,0.1270514727,0.0835983381,-0.5973306298,-0.211365819,0.1870436817,0.1715711355,-0.1690764129,0.3093141019,-0.0003779458,0.4118531942,0.0478395224,0.1017804518,0.188334778,-0.0319394767,-0.2228376269,0.2945310175,-0.1456040591,-0.067813009,-0.1044202149,0.3420464396,0.3204126358,-0.1875416338,0.3151755631,0.1449416131,-0.1745819002,-0.2085203677,0.1155626923,-0.2270266414,0.1572457701,-0.3124389052,-0.0611692443,-0.1168108806,0.1118064374,0.0585076138,0.1921531111,-0.1500457078,0.1555631608,0.3324890137,0.0116774989,0.3420877457,-0.0901248977,-0.2835850418,-0.3240111172,-0.0644662678,-0.0928679183,0.1528823376,0.2960220575,-0.2420353591,-0.1924745142,-0.4010330141,0.3508935273,0.0098660961,0.014594241,-0.2698222399,-0.1220166758,-0.0657829419,-0.0629407614,-0.0144885601,0.2110172063,-0.1504108459,-0.3179164827,0.3872092366,0.117640771,0.0571581647,-0.1679102629,0.3586957157,0.048453249,0.2030505687,0.1642742008,-0.0272339061,-0.2241767794,0.3396662772,0.2753737271,-0.2568640709,-0.0250149872,-0.6368867159,-0.1241964921,-0.2141603678,-0.033792384,-0.3123409748,-0.2352536768,0.1731517315,-0.0345406681,0.2715982497,0.0634436905,0.0401524678,0.1113931984,-0.0591189116,-0.1764664203,-0.1673034132,0.0902206376,0.0907126218,0.1448710412,0.3844320774,-0.2436366677,0.0482861213,-0.0284506846,0.076481469,0.1372897625,-0.3264172077,-0.4592152834,0.0074926852,-0.2270474434,0.1214903742,0.2082408071,0.349242717,0.1527168155,0.0456990264,-0.0921820775,-0.0213659946,-0.0105895605,0.0681030154,0.3793623745,0.0219300017,0.2481969744,0.0916100144,0.3852550089,0.2168067694,0.0136740403,-0.2023299485,-0.1696556211,-0.2334572077,0.0689230785,-0.2786014676,-0.3739550412,-0.1919828653,-0.2463329881,-0.1471285522,-0.0810213462,0.1199394763,-0.0407525674,0.1887602806,-0.0296876468,0.1275856048,-0.2064576,-0.1042085066,0.1654075533,-0.4634034336,0.2584106326,0.1879547387,0.1876951754,0.177471444,0.4568811059,-0.2174549103,-0.1817057282,-0.4573141634,-0.4930145741,0.0420013703,-0.1547994167,0.2963698804,-0.1615109444,0.1329655498,-0.2619121969,0.0421557799,0.2144407928,0.3131137192,0.0053436663,-0.3617569804,-0.1295341402,0.15239802,0.2682896256,-0.3597653806,-0.1527059823,0.0125954207,0.1549457312,-0.0277927723,0.0147465561,-0.4901633263,0.0906729549,-0.1672899872,-0.1309010237,-0.0067184488,-0.101703167,-0.3829989731,0.2168268412,-0.0516363308,-0.2362247258,0.2038921565,0.0279137641,-0.0117243584,-0.2106576115,0.0760313943,0.1673894227,-0.154104799,0.0374530703,0.227654621,0.1397216916,0.092387341,0.1287388653,0.0564554408,-0.1926456392,-0.1596468687,0.2166396677,0.0616349988,0.1089248657,-0.0098423483,0.1810977608,-0.3064339459,0.4406492412,0.5961535573,0.0669471473,-0.1543173492,-0.0552444533,0.319121182,-0.2241158634,-0.1444145292,-0.2162664533,0.1355162561,-0.1104243323,0.2012257576,0.1469836831,-0.2011134923,-0.2267197222,0.0276117921,-0.2966864705,-0.0866997242,0.1483125091,-0.3616520464,0.2280011028,0.0028276788,-0.4828451276,-0.254280448,-0.2783206105,-0.1992045939,-0.2758313715,0.3278150558,-0.0452019051,-0.352881819,-0.0166930687,-0.3142630458,0.2406503707,0.366499722,-0.0843384266,-0.0907220244,-0.0604189783,-0.129215762,0.2606255412,0.1542933583,-0.5128048062,-0.163343206,0.1516357213,0.4062105119,-0.065043807,0.2887904942,-0.0917122364,-0.1130859405,0.2488005906,0.2509052157,-0.2583779395,0.1570632309,0.2447043806,0.4032779336,-0.1736409813,-0.1313845515,0.2057297379,0.165261358,-0.2362172455,-0.0962601602,0.1956248879,-0.0006142814,0.2740318179,-0.0693942606,0.0365953222,0.2171511352,0.5396988392,0.1209776551,0.0811526626,0.0803797618,0.2357332408,0.4530342817,-0.0320292115,0.137833789,0.670058012,-0.2145473659,-0.3292794526,-0.1934350282,-0.1311738342,0.6428803205,0.0062167235,0.3512897193,0.1989220232,0.018884873,-0.1209156364,-0.0490795486,-0.2452421486,0.0792133063,0.1984337866,-0.7353713512,-0.4121163785,0.2752217352,0.2985345721,-0.3370001316,-0.2035911381,0.1112074703,-0.2003127337,0.494710058,0.0612143725,1.1625715494,0.0410089195,0.1179920584,-0.3089979291,-0.1229391173,0.9703682661,0.1930099279,-0.1272209883,-0.0968471318,-0.1137187034,-0.2565720677,-0.224820599,-0.1375611871,0.5079173446,-0.1840531677,0.3734824061,0.0310727637,0.1875920296,0.1369579434,-0.0327431709,0.5116969347,0.0012901746,-0.0064742193,0.0229872707,0.115257591,-0.0902241245,-0.1849841774,-0.0297493227,-0.302593857,0.1189246327,-0.3847559988,0.0003966058,0.0650695115,-0.0481357649,0.085029833,-0.2355203629,0.1955161691,0.7198811173,0.356329143,0.1922262311,-0.284825623,0.1726632118,-0.0092221908,0.0049915863,0.1629317403,-0.1281560957,0.2780495584,0.1617682874,-0.2777023911,0.194997713,0.1021158323,-0.0899804682,-0.4744349122,0.0123073738,0.0980078653,-0.0945380703,-0.5388156176,0.3663734496,0.1767819226,-0.1856520027,0.0131999282,0.0001883262,0.074918732,0.0935518444,0.1123004556,-0.3210189342,0.0606168061,0.3731020093,0.3647237718,-0.0629596114,0.3424706161,0.3815141618,0.0855754837,-0.087284863,0.0814722329,0.0288084466,-0.3980549872,-0.1969986856,0.1057754382,0.238794744,-0.3082545102,-0.0100204041,0.3829267919,0.1897433102,0.1722963601,-0.1259191334,-0.2626465559,0.3369967341,0.1391752809,0.1633089632,0.0490511917,0.0862304717,-0.0101358844,0.5093473196,-0.2070423514,0.1580708921,0.0907078087,0.2006306499,0.5335078239,0.3277243376,-0.3454565406,-0.3669421375,0.0643301979,-0.2048447132,0.3108893931,-0.1922218055,0.0480891801,0.1692542881,0.101621829,0.0221291203,-0.0469608605,-0.2170206308,0.3040750027,-0.316380173,0.0354828313,0.2301063538,0.2214178592,-0.26301229,0.2981283665,0.1858418286,-0.56121701,0.3651367426,0.1390035152,0.1313632131,-0.312027961,-0.227317214,0.2362160236,-0.0268927887,0.2411732227,-0.0294415075,0.3321395814,0.0570386425,-0.0083749797,-0.0273229219,-0.0005379622,0.2442921847,0.3078631461,-0.2212412059,-0.0240019187,0.1011852771,0.1767583489,0.1289907396,-0.2226727903,-0.0394648761,0.5133666992,-0.0380398333,-0.0375019237,0.0438097268,-0.1290798038,0.3635984659,-0.2724665403,0.0844293386,0.123299323,-0.1309456825,0.1443681121,0.4139513075,0.5091969967,-0.0637902245,0.0432810299,0.1681563854,0.2342418581,0.263600111,0.2262664288,0.1249122471,0.5415756702,0.1831618398,0.6217491031,0.2145633847,0.2305883318,0.3615840077,0.2951465249,-0.3101603687,-0.073356472,0.1703614295,0.1269236654,-0.2907981277,0.1274164468,0.4434099197,0.0040870984,-0.2422980368,-0.0603941828,-0.4170541167,0.1044449136,-0.0596784204,-0.0944367722,-0.2965032458,0.2017251551,-0.1277420968,0.1328473538,-0.1001606733,0.0389208831,-0.081516169,0.0456891283,-0.2055385709,0.2178045362,-0.2609117925,-0.0936178416,0.2776680589,0.642724216,0.2933251262,0.1303417534,-0.1099643335,0.3280540407,0.0163042136,-0.1135228872,0.1416664124,-0.0146136852,-0.0806663707,-0.5470141172,0.2632035315,0.1730249077,-0.0409362391,0.1357989758,-0.0137350559,0.3255069256,-0.1209612042,-0.1709827781,-0.2564886808,-0.0963709652,0.080746226,-0.0569785982,-0.2953187227,0.0591468811,0.0585425757,-0.1512655169,0.0465597212,0.0381093286,0.0607062317,0.0433832221,0.1929327846,0.4572063684,0.1805943996,-0.2380101681,-0.2230480909,-0.6116508842,0.1156006753,-0.2424541265,0.0879886597,-0.4129204452,-0.0658173263,0.3155344129,0.1827477366,0.1582343131,0.0365645178,-0.3021129966,0.1524007618,-0.1534939408,-0.324220866,0.155217126,-0.0207843017,-0.1128108501,0.0763467699,0.0181961022,0.0454507247,-0.1144116297,0.0254975632,-0.1378238946,0.1756382436,-0.1870414019,0.0889947638,0.1502768993,0.7541152239,-0.3090142012,0.0606666319,0.4328764081,-0.3222761452,-0.0024946798,0.6328269839,0.0188936517,0.2234545946,-0.3817552924,0.1293927282,0.2589629889,-0.5065020323,-0.339641124,0.3446277678,-0.6829172373,-0.2021073997,-0.0883289278,-0.0181745421,-0.0917239785,0.226677835,-0.0303806476,0.270624876,0.2221622914,-0.1556902081,0.1093000025,-0.1322838664,-0.2360172719,-0.2123873085,-0.0463723242,0.2906597555,-0.0349209458,-0.3357986808,0.0541259684,0.4443776309,0.1232088581,0.0623674877,0.1291297525,-0.0715733841,0.0446424335,-0.2311865389,0.1862354875,-0.0285791829,-0.0046890429,-0.4483374357,-0.337302804]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/798","title":"Cannot load TREC dataset: ConnectionError","comments":"Hi ! Indeed there's an issue with those links.\r\nWe should probably use the target urls of the redirections instead","body":"## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. ","comment_length":20,"text":"Cannot load TREC dataset: ConnectionError\n## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. \nHi ! Indeed there's an issue with those links.\r\nWe should probably use the target urls of the redirections instead","embeddings":[-0.1691725552,0.0168359298,0.0626651496,0.3633992374,0.2918767631,-0.1796784997,0.2199582756,-0.0434447527,-0.1838140339,0.0629585609,-0.3238992393,0.1360173374,0.0722290277,0.1577587724,-0.0579137243,-0.0945270881,-0.1901099831,-0.2000556588,-0.0688836575,0.2535367906,-0.1328548044,0.4753988683,-0.2073137164,0.2046911418,-0.4112499952,0.0120672574,0.2174304873,0.4217193425,-0.2242174745,-0.6152897477,0.4672029912,0.0918213353,0.2583209872,0.630194068,-0.0001212188,0.0438775904,0.3747947514,-0.0809114501,-0.3682197928,-0.5027552843,-0.5078126788,-0.0790523142,0.2648388743,0.2292865217,-0.3196886182,0.4285567403,0.0791531503,-0.325060308,0.3382033706,0.6332944036,0.0820835829,-0.0768375695,0.0945120901,-0.1623590142,0.3016184568,-0.1488979459,-0.0706101805,0.4869538546,0.2155133337,-0.0386384055,0.0549374111,-0.0216429308,-0.0271049347,-0.0117052123,0.1798272282,-0.0742924362,-0.1101848111,-0.4340969324,0.4241297245,0.352265209,0.6402148008,-0.1733739376,-0.2272690535,0.1454129517,-0.081264168,-0.4302360415,0.3003183305,0.3217078149,-0.1221241057,0.1870257258,-0.4078291059,-0.2597643733,-0.3686292768,0.0043727914,-0.416408807,0.1270349473,-0.2663984001,0.131530717,0.2910034359,0.0152950874,0.3388897479,0.2237246931,-0.4470455348,-0.0075067496,-0.6929902434,0.1829680651,-0.1482103318,-0.0363469273,0.0934938565,0.3697319627,0.3174377084,0.0368965492,-0.2863053977,0.4065686166,0.0869918466,0.3421801031,0.0432393365,0.1872974336,0.3242341876,0.1837179959,-0.0132212332,-0.0930793434,-0.195083186,-0.2066059262,-0.1330894083,0.1809051484,0.0973416939,-0.2702093422,-0.391297996,0.1465577036,-0.4139411747,0.009193588,0.1225641072,0.4541309178,-0.2755913138,-0.0242432654,-0.0385249332,0.2208224833,-0.0965710506,-0.1626555324,-0.066012606,0.2261447608,-0.2326459587,0.1317702383,0.3381001651,-0.1813087612,0.1413071007,-0.164917931,-0.066844523,-0.2156009972,-0.0086505944,-0.2676545978,-0.1067192107,0.4307493269,0.0776933208,0.0562675633,0.1243787333,-0.3719575107,-0.0352643579,0.0133798691,-0.5714361072,-0.3518609107,0.0621165261,0.0705782101,-0.1204286665,0.0219029319,-0.3443544507,0.0012689885,-0.0043544434,-0.1800432205,-0.1748710871,-0.2704450488,-0.3334649503,-0.3149197698,0.3458762765,0.451456517,-0.5566045642,-0.0270752627,0.0968400612,0.0743105635,0.1674599946,0.3314805627,-0.1907550693,-0.0951525122,-0.2181226611,-0.1000668108,0.0956019834,-0.3112884164,-0.2985938787,-0.0033583376,-0.0194073506,-0.0483571663,-0.2369234562,-0.1034136266,0.013775479,0.0616033152,0.6465601921,0.2157903463,-0.128316775,-0.1761939228,-0.1502933949,-0.276214689,-0.2542467415,0.0869560763,0.1214700714,0.0059731621,0.2363967299,-0.4764141738,0.070228681,0.041881226,-0.0073060701,0.3011595905,0.3711313903,-0.0534163527,-0.1965114772,0.2541643977,-0.4512370229,0.3701867163,-0.0718612894,0.2864062488,-0.2534539998,0.1539354026,-0.2667068243,0.1449401677,-0.0538807735,-0.1338416189,-0.0591719225,0.0979989544,0.2250677347,0.2401248366,-0.1199853867,0.5677842498,0.135291785,0.1387476027,-0.2335021943,0.2586082816,0.0218197275,0.0688075498,0.1329162866,0.0106323007,0.2142108232,-0.2170554101,-0.0641218349,0.318633914,0.2173629254,0.0417961814,0.057461638,0.1577191204,0.2542841733,-0.3891916275,-0.0997398421,0.0472882427,-0.0086243441,-0.0252697486,-0.0405539423,0.2425730079,0.0648595691,0.4205703437,0.1400236785,0.0894204974,0.3215745389,0.0298587233,-0.0160045065,0.179474324,0.1751997471,0.2178916037,0.0230622217,-0.0838506073,-0.3602313399,0.1493657082,-0.1642055213,0.0495590158,-0.1129640341,0.1247549057,-0.1547979712,-0.0220802911,0.0700044334,0.3708671927,0.2201276124,0.0490480177,-0.2005204111,0.1874097735,0.0605272613,-0.2058147639,-0.0445009433,0.1096539572,0.3646667898,0.2655833662,0.426687777,-0.0487633571,-0.095108144,-0.3787639439,-0.0358973593,0.2614510655,-0.27078861,0.4008783698,-0.2434709668,-0.3584270775,0.0938047096,-0.3161964118,-0.3449476063,-0.2649385929,-0.1177102774,0.4375137389,0.1163083985,0.0623287521,-0.2111103535,0.3716548681,0.0254304223,-0.5300382972,-0.023688294,-0.2573983371,-0.3198367953,0.015924165,0.3333905041,-0.2213886827,0.0698618591,-0.2547039092,0.0297967102,-0.1255459189,-0.2030438036,0.161564365,0.1523073763,0.3011630476,-0.0669810548,0.2267765105,-0.2425041497,0.0940599963,0.2750999629,0.067961067,0.1023404077,0.048934143,-0.0400866605,0.0441079624,0.0249122102,-0.2555710673,-0.4800460041,-0.3582805991,0.1784227639,0.0196136292,0.3458847404,0.2960084677,0.0382493213,-0.0866623893,0.0641809925,-0.0769640952,-0.1533749849,-0.3865545392,0.1724119484,-0.2786794007,-0.2431341708,0.2509688139,0.0747392997,0.384760499,0.0763949454,-0.6094400883,-0.1564110816,-0.1264603585,0.1048995107,0.0258263517,-0.2825921476,0.3334425092,-0.0882083923,-0.0051477673,0.216064468,-0.4404521585,0.0323337875,0.3356822431,0.3913239837,0.2737467587,0.3707469404,-0.1709450483,0.4239256084,0.2026095539,0.0485433601,0.3296657801,-0.1572235525,0.3341207802,-0.0892179385,-0.2969244123,-0.2172842175,-0.2103471756,-0.3245612681,0.1617722362,0.2636647522,-0.0344154313,-0.4967108071,0.1173620149,-0.0801406428,-0.3276883364,0.0556484796,0.1679751873,0.1110880598,0.0038787234,-0.0378257744,0.2212286294,-0.2119385451,-0.3172354698,0.3629997075,0.0803808197,0.0725494176,-0.1304673105,-0.0161201619,-0.8267481923,0.3045907319,0.0043531074,0.3990243375,-0.3158361316,0.0614299923,0.0791887119,-0.1068061665,0.1192461029,-0.3070542216,-0.0818270966,0.3087825179,0.0600856058,-0.2109733224,-0.2092314512,0.1269949377,0.1855375469,0.4327666461,0.3602838516,0.013694901,0.0053407452,0.0741606504,0.1480065137,-0.0896288231,0.0230829157,-0.3215021193,-0.4804897308,-0.2376385331,-0.0714276284,-0.062724866,0.298281461,0.1305000633,0.0849203244,0.2911790609,-0.2723354399,0.2081023902,0.0922953635,0.0188673791,0.1057648584,0.1027646288,0.0991389379,0.0624909215,0.3354113996,0.9395033717,0.1009224579,-0.120416373,0.3544933796,0.3962078691,0.0735524446,0.2287358344,-0.051833827,0.1031669229,-0.0455486812,0.1310721934,0.0472777598,0.3104491532,-0.2412313968,-0.0123398695,-0.844632566,-0.1633535475,0.3940396309,-0.1157400385,-0.0963782668,-0.0885319561,0.191869244,-0.1223828048,0.0664271712,-0.0798335299,0.9389856458,-0.2157409042,-0.1139465198,0.0964188576,-0.147055313,0.4845153391,-0.5781700015,0.4780519307,-0.3271455765,-0.1596903652,-0.2196512967,-0.2689325809,-0.1141616106,-0.1754051447,-0.002866735,0.4294499457,-0.1375024319,0.1797110438,0.4103512764,0.0312496778,0.1668056101,0.0313799717,-0.1563229114,0.0742265582,-0.0684729144,0.3795146346,-0.0753371939,-0.0879741907,-0.2208290249,-0.0232577417,-0.3305442631,0.2558438778,-0.7046300769,0.0951708704,0.0003454578,-0.2027993649,0.1387290508,0.1491003036,0.2019768804,0.1502197832,-0.3103254139,0.2650389969,-0.0473067723,-0.1506277919,-0.009702215,-0.0158758238,0.010616634,-0.0578824542,-0.0800179094,0.4451651275,-0.0131923249,-0.4016885757,0.0497705564,0.0414262153,0.0156010203,-0.0672010556,0.1310195774,-0.0913955346,0.0334716663,0.0690084398,0.0271308459,0.1205178499,0.1539868563,-0.0196321197,0.2282925695,-0.259336561,-0.0157153793,0.3636079133,-0.0135959983,-0.0319505855,0.1860428751,0.4898221791,-0.0020252562,-0.0705311596,-0.1395792663,0.2126883119,-0.3972610533,0.1164028496,-0.0432341732,0.1496537328,0.0333801284,0.3472812772,0.3143411875,0.0317257643,0.3876162767,-0.5564649701,-0.2944401801,0.0544478521,-0.1690248102,0.184166044,0.1040887386,-0.1632938981,-0.2276393175,-0.0018326067,-0.1922600567,0.0723112226,-0.3587886393,-0.175903365,0.1726374626,-0.1161857471,0.5185421109,-0.1368280649,-0.0479452088,-0.0184052158,-0.3581270874,-0.0564916581,-0.0145608978,0.2017893344,0.1712418199,-0.4430913627,-0.1755434275,-0.077528283,0.0485689603,0.0543948784,0.1847438067,0.1984396875,-0.1803075075,-0.3293845952,-0.4175056219,0.0081544761,-0.3177674711,0.3675778508,0.2319299281,0.4606673419,0.2174215019,0.2435667217,-0.0169100836,-0.1336589158,-0.117658779,0.2129968703,-0.053510312,-0.2645329237,0.1965231746,-0.4327823818,0.0601596422,-0.2405550629,0.2324357629,0.0786642805,-0.2138750255,0.14287664,0.1881443113,0.0525172353,-0.0704044029,-0.0328343585,0.1413883418,-0.0272031706,0.0310485549,0.2250308245,0.386873275,-0.1784227788,0.0799675807,0.2459060699,0.5618061423,0.1223886907,0.1421130449,-0.1236973405,-0.1237056777,0.1257034093,0.4094533622,0.0089930175,0.0327677615,0.0211193711,-0.3650708497,-0.0349260531,-0.0059551336,0.0183139034,0.0332748108,-0.5767639279,0.1064377353,0.1411761791,-0.1795779765,-0.1717036068,-0.0894220397,0.4524857998,-0.2100297809,-0.0174033958,-0.3908969462,0.1880070865,-0.0242402554,-0.039714206,0.1182918698,-0.4415909052,-0.1305991709,0.2427270114,0.0069916546,-0.1593962014,-0.0046678944,0.0859766155,-0.2388294488,-0.3000748754,-0.0755934119,0.3069646955,-0.0426480062,-0.0936480537,0.4497712553,0.0273359865,-0.1706606746,0.1557034552,0.0153890671,0.62725842,0.2654240429,0.1823752671,0.1087691635,-0.1266422421,0.0071312743,0.1697154045,0.0883647799,0.1900411397,-0.332215935,0.1685412526,0.078648366,0.0258462951,0.2831476927,-0.092812404,0.3121075034,0.0294574928,0.0179795437,-0.1460418701,0.1118161753,-0.1592104733,0.0422839858,-0.31451267,0.178357631,0.3862515092,-0.140285328,0.2693823576,0.1543028802,0.008642192,-0.178006798,0.4599909782,0.3405545056,0.2323293835,-0.3487294912,-0.1899100095,-0.2473540008,-0.0055192844,-0.1115458161,0.2073093206,-0.0416926779,0.1502738595,-0.204635039,0.3032639623,-0.1217940003,0.3058781028,0.0848768055,0.0063988287,-0.2651689947,0.0663421229,0.0090942262,0.0498305373,-0.1537736654,-0.2624773383,0.0815571919,-0.3734006286,-0.0794994906,-0.1199191436,0.1511416584,0.0426995084,-0.2643913627,0.2991390824,0.2458395213,0.7336868048,0.2630758286,-0.0093748644,-0.1382427812,-0.3809232116,-0.0225659534,0.0093688183,0.3737497926,0.2838226855,0.0853671506,-0.2404976487,-0.385111481,0.4988894165,0.065247342,0.236622721,0.1650582105,-0.0864073858,-0.0182050411,0.0852971599,-0.0759951696,-0.1094401255,0.1904141009,0.1407316327,-0.1363411546,-0.3142247498,0.5675712824,-0.3454417884,0.1128251329,-0.1528615505,0.2606040537,0.3078807592,0.0205607451,-0.270526737,0.1239754856,0.2145230472,-0.1328070611,-0.4126018882,0.3355360031,-0.2914167941,-0.0648081899,-0.0611460693,0.1464544684,0.0038937069,-0.3008677363,-0.2765820324,-0.0266715754]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/798","title":"Cannot load TREC dataset: ConnectionError","comments":"Hi, the same issue here, could you tell me how to download it through datasets? thanks ","body":"## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. ","comment_length":16,"text":"Cannot load TREC dataset: ConnectionError\n## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. \nHi, the same issue here, could you tell me how to download it through datasets? thanks ","embeddings":[-0.1691725552,0.0168359298,0.0626651496,0.3633992374,0.2918767631,-0.1796784997,0.2199582756,-0.0434447527,-0.1838140339,0.0629585609,-0.3238992393,0.1360173374,0.0722290277,0.1577587724,-0.0579137243,-0.0945270881,-0.1901099831,-0.2000556588,-0.0688836575,0.2535367906,-0.1328548044,0.4753988683,-0.2073137164,0.2046911418,-0.4112499952,0.0120672574,0.2174304873,0.4217193425,-0.2242174745,-0.6152897477,0.4672029912,0.0918213353,0.2583209872,0.630194068,-0.0001212188,0.0438775904,0.3747947514,-0.0809114501,-0.3682197928,-0.5027552843,-0.5078126788,-0.0790523142,0.2648388743,0.2292865217,-0.3196886182,0.4285567403,0.0791531503,-0.325060308,0.3382033706,0.6332944036,0.0820835829,-0.0768375695,0.0945120901,-0.1623590142,0.3016184568,-0.1488979459,-0.0706101805,0.4869538546,0.2155133337,-0.0386384055,0.0549374111,-0.0216429308,-0.0271049347,-0.0117052123,0.1798272282,-0.0742924362,-0.1101848111,-0.4340969324,0.4241297245,0.352265209,0.6402148008,-0.1733739376,-0.2272690535,0.1454129517,-0.081264168,-0.4302360415,0.3003183305,0.3217078149,-0.1221241057,0.1870257258,-0.4078291059,-0.2597643733,-0.3686292768,0.0043727914,-0.416408807,0.1270349473,-0.2663984001,0.131530717,0.2910034359,0.0152950874,0.3388897479,0.2237246931,-0.4470455348,-0.0075067496,-0.6929902434,0.1829680651,-0.1482103318,-0.0363469273,0.0934938565,0.3697319627,0.3174377084,0.0368965492,-0.2863053977,0.4065686166,0.0869918466,0.3421801031,0.0432393365,0.1872974336,0.3242341876,0.1837179959,-0.0132212332,-0.0930793434,-0.195083186,-0.2066059262,-0.1330894083,0.1809051484,0.0973416939,-0.2702093422,-0.391297996,0.1465577036,-0.4139411747,0.009193588,0.1225641072,0.4541309178,-0.2755913138,-0.0242432654,-0.0385249332,0.2208224833,-0.0965710506,-0.1626555324,-0.066012606,0.2261447608,-0.2326459587,0.1317702383,0.3381001651,-0.1813087612,0.1413071007,-0.164917931,-0.066844523,-0.2156009972,-0.0086505944,-0.2676545978,-0.1067192107,0.4307493269,0.0776933208,0.0562675633,0.1243787333,-0.3719575107,-0.0352643579,0.0133798691,-0.5714361072,-0.3518609107,0.0621165261,0.0705782101,-0.1204286665,0.0219029319,-0.3443544507,0.0012689885,-0.0043544434,-0.1800432205,-0.1748710871,-0.2704450488,-0.3334649503,-0.3149197698,0.3458762765,0.451456517,-0.5566045642,-0.0270752627,0.0968400612,0.0743105635,0.1674599946,0.3314805627,-0.1907550693,-0.0951525122,-0.2181226611,-0.1000668108,0.0956019834,-0.3112884164,-0.2985938787,-0.0033583376,-0.0194073506,-0.0483571663,-0.2369234562,-0.1034136266,0.013775479,0.0616033152,0.6465601921,0.2157903463,-0.128316775,-0.1761939228,-0.1502933949,-0.276214689,-0.2542467415,0.0869560763,0.1214700714,0.0059731621,0.2363967299,-0.4764141738,0.070228681,0.041881226,-0.0073060701,0.3011595905,0.3711313903,-0.0534163527,-0.1965114772,0.2541643977,-0.4512370229,0.3701867163,-0.0718612894,0.2864062488,-0.2534539998,0.1539354026,-0.2667068243,0.1449401677,-0.0538807735,-0.1338416189,-0.0591719225,0.0979989544,0.2250677347,0.2401248366,-0.1199853867,0.5677842498,0.135291785,0.1387476027,-0.2335021943,0.2586082816,0.0218197275,0.0688075498,0.1329162866,0.0106323007,0.2142108232,-0.2170554101,-0.0641218349,0.318633914,0.2173629254,0.0417961814,0.057461638,0.1577191204,0.2542841733,-0.3891916275,-0.0997398421,0.0472882427,-0.0086243441,-0.0252697486,-0.0405539423,0.2425730079,0.0648595691,0.4205703437,0.1400236785,0.0894204974,0.3215745389,0.0298587233,-0.0160045065,0.179474324,0.1751997471,0.2178916037,0.0230622217,-0.0838506073,-0.3602313399,0.1493657082,-0.1642055213,0.0495590158,-0.1129640341,0.1247549057,-0.1547979712,-0.0220802911,0.0700044334,0.3708671927,0.2201276124,0.0490480177,-0.2005204111,0.1874097735,0.0605272613,-0.2058147639,-0.0445009433,0.1096539572,0.3646667898,0.2655833662,0.426687777,-0.0487633571,-0.095108144,-0.3787639439,-0.0358973593,0.2614510655,-0.27078861,0.4008783698,-0.2434709668,-0.3584270775,0.0938047096,-0.3161964118,-0.3449476063,-0.2649385929,-0.1177102774,0.4375137389,0.1163083985,0.0623287521,-0.2111103535,0.3716548681,0.0254304223,-0.5300382972,-0.023688294,-0.2573983371,-0.3198367953,0.015924165,0.3333905041,-0.2213886827,0.0698618591,-0.2547039092,0.0297967102,-0.1255459189,-0.2030438036,0.161564365,0.1523073763,0.3011630476,-0.0669810548,0.2267765105,-0.2425041497,0.0940599963,0.2750999629,0.067961067,0.1023404077,0.048934143,-0.0400866605,0.0441079624,0.0249122102,-0.2555710673,-0.4800460041,-0.3582805991,0.1784227639,0.0196136292,0.3458847404,0.2960084677,0.0382493213,-0.0866623893,0.0641809925,-0.0769640952,-0.1533749849,-0.3865545392,0.1724119484,-0.2786794007,-0.2431341708,0.2509688139,0.0747392997,0.384760499,0.0763949454,-0.6094400883,-0.1564110816,-0.1264603585,0.1048995107,0.0258263517,-0.2825921476,0.3334425092,-0.0882083923,-0.0051477673,0.216064468,-0.4404521585,0.0323337875,0.3356822431,0.3913239837,0.2737467587,0.3707469404,-0.1709450483,0.4239256084,0.2026095539,0.0485433601,0.3296657801,-0.1572235525,0.3341207802,-0.0892179385,-0.2969244123,-0.2172842175,-0.2103471756,-0.3245612681,0.1617722362,0.2636647522,-0.0344154313,-0.4967108071,0.1173620149,-0.0801406428,-0.3276883364,0.0556484796,0.1679751873,0.1110880598,0.0038787234,-0.0378257744,0.2212286294,-0.2119385451,-0.3172354698,0.3629997075,0.0803808197,0.0725494176,-0.1304673105,-0.0161201619,-0.8267481923,0.3045907319,0.0043531074,0.3990243375,-0.3158361316,0.0614299923,0.0791887119,-0.1068061665,0.1192461029,-0.3070542216,-0.0818270966,0.3087825179,0.0600856058,-0.2109733224,-0.2092314512,0.1269949377,0.1855375469,0.4327666461,0.3602838516,0.013694901,0.0053407452,0.0741606504,0.1480065137,-0.0896288231,0.0230829157,-0.3215021193,-0.4804897308,-0.2376385331,-0.0714276284,-0.062724866,0.298281461,0.1305000633,0.0849203244,0.2911790609,-0.2723354399,0.2081023902,0.0922953635,0.0188673791,0.1057648584,0.1027646288,0.0991389379,0.0624909215,0.3354113996,0.9395033717,0.1009224579,-0.120416373,0.3544933796,0.3962078691,0.0735524446,0.2287358344,-0.051833827,0.1031669229,-0.0455486812,0.1310721934,0.0472777598,0.3104491532,-0.2412313968,-0.0123398695,-0.844632566,-0.1633535475,0.3940396309,-0.1157400385,-0.0963782668,-0.0885319561,0.191869244,-0.1223828048,0.0664271712,-0.0798335299,0.9389856458,-0.2157409042,-0.1139465198,0.0964188576,-0.147055313,0.4845153391,-0.5781700015,0.4780519307,-0.3271455765,-0.1596903652,-0.2196512967,-0.2689325809,-0.1141616106,-0.1754051447,-0.002866735,0.4294499457,-0.1375024319,0.1797110438,0.4103512764,0.0312496778,0.1668056101,0.0313799717,-0.1563229114,0.0742265582,-0.0684729144,0.3795146346,-0.0753371939,-0.0879741907,-0.2208290249,-0.0232577417,-0.3305442631,0.2558438778,-0.7046300769,0.0951708704,0.0003454578,-0.2027993649,0.1387290508,0.1491003036,0.2019768804,0.1502197832,-0.3103254139,0.2650389969,-0.0473067723,-0.1506277919,-0.009702215,-0.0158758238,0.010616634,-0.0578824542,-0.0800179094,0.4451651275,-0.0131923249,-0.4016885757,0.0497705564,0.0414262153,0.0156010203,-0.0672010556,0.1310195774,-0.0913955346,0.0334716663,0.0690084398,0.0271308459,0.1205178499,0.1539868563,-0.0196321197,0.2282925695,-0.259336561,-0.0157153793,0.3636079133,-0.0135959983,-0.0319505855,0.1860428751,0.4898221791,-0.0020252562,-0.0705311596,-0.1395792663,0.2126883119,-0.3972610533,0.1164028496,-0.0432341732,0.1496537328,0.0333801284,0.3472812772,0.3143411875,0.0317257643,0.3876162767,-0.5564649701,-0.2944401801,0.0544478521,-0.1690248102,0.184166044,0.1040887386,-0.1632938981,-0.2276393175,-0.0018326067,-0.1922600567,0.0723112226,-0.3587886393,-0.175903365,0.1726374626,-0.1161857471,0.5185421109,-0.1368280649,-0.0479452088,-0.0184052158,-0.3581270874,-0.0564916581,-0.0145608978,0.2017893344,0.1712418199,-0.4430913627,-0.1755434275,-0.077528283,0.0485689603,0.0543948784,0.1847438067,0.1984396875,-0.1803075075,-0.3293845952,-0.4175056219,0.0081544761,-0.3177674711,0.3675778508,0.2319299281,0.4606673419,0.2174215019,0.2435667217,-0.0169100836,-0.1336589158,-0.117658779,0.2129968703,-0.053510312,-0.2645329237,0.1965231746,-0.4327823818,0.0601596422,-0.2405550629,0.2324357629,0.0786642805,-0.2138750255,0.14287664,0.1881443113,0.0525172353,-0.0704044029,-0.0328343585,0.1413883418,-0.0272031706,0.0310485549,0.2250308245,0.386873275,-0.1784227788,0.0799675807,0.2459060699,0.5618061423,0.1223886907,0.1421130449,-0.1236973405,-0.1237056777,0.1257034093,0.4094533622,0.0089930175,0.0327677615,0.0211193711,-0.3650708497,-0.0349260531,-0.0059551336,0.0183139034,0.0332748108,-0.5767639279,0.1064377353,0.1411761791,-0.1795779765,-0.1717036068,-0.0894220397,0.4524857998,-0.2100297809,-0.0174033958,-0.3908969462,0.1880070865,-0.0242402554,-0.039714206,0.1182918698,-0.4415909052,-0.1305991709,0.2427270114,0.0069916546,-0.1593962014,-0.0046678944,0.0859766155,-0.2388294488,-0.3000748754,-0.0755934119,0.3069646955,-0.0426480062,-0.0936480537,0.4497712553,0.0273359865,-0.1706606746,0.1557034552,0.0153890671,0.62725842,0.2654240429,0.1823752671,0.1087691635,-0.1266422421,0.0071312743,0.1697154045,0.0883647799,0.1900411397,-0.332215935,0.1685412526,0.078648366,0.0258462951,0.2831476927,-0.092812404,0.3121075034,0.0294574928,0.0179795437,-0.1460418701,0.1118161753,-0.1592104733,0.0422839858,-0.31451267,0.178357631,0.3862515092,-0.140285328,0.2693823576,0.1543028802,0.008642192,-0.178006798,0.4599909782,0.3405545056,0.2323293835,-0.3487294912,-0.1899100095,-0.2473540008,-0.0055192844,-0.1115458161,0.2073093206,-0.0416926779,0.1502738595,-0.204635039,0.3032639623,-0.1217940003,0.3058781028,0.0848768055,0.0063988287,-0.2651689947,0.0663421229,0.0090942262,0.0498305373,-0.1537736654,-0.2624773383,0.0815571919,-0.3734006286,-0.0794994906,-0.1199191436,0.1511416584,0.0426995084,-0.2643913627,0.2991390824,0.2458395213,0.7336868048,0.2630758286,-0.0093748644,-0.1382427812,-0.3809232116,-0.0225659534,0.0093688183,0.3737497926,0.2838226855,0.0853671506,-0.2404976487,-0.385111481,0.4988894165,0.065247342,0.236622721,0.1650582105,-0.0864073858,-0.0182050411,0.0852971599,-0.0759951696,-0.1094401255,0.1904141009,0.1407316327,-0.1363411546,-0.3142247498,0.5675712824,-0.3454417884,0.1128251329,-0.1528615505,0.2606040537,0.3078807592,0.0205607451,-0.270526737,0.1239754856,0.2145230472,-0.1328070611,-0.4126018882,0.3355360031,-0.2914167941,-0.0648081899,-0.0611460693,0.1464544684,0.0038937069,-0.3008677363,-0.2765820324,-0.0266715754]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/798","title":"Cannot load TREC dataset: ConnectionError","comments":"Same issue. ","body":"## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. ","comment_length":2,"text":"Cannot load TREC dataset: ConnectionError\n## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. \nSame issue. ","embeddings":[-0.1691725552,0.0168359298,0.0626651496,0.3633992374,0.2918767631,-0.1796784997,0.2199582756,-0.0434447527,-0.1838140339,0.0629585609,-0.3238992393,0.1360173374,0.0722290277,0.1577587724,-0.0579137243,-0.0945270881,-0.1901099831,-0.2000556588,-0.0688836575,0.2535367906,-0.1328548044,0.4753988683,-0.2073137164,0.2046911418,-0.4112499952,0.0120672574,0.2174304873,0.4217193425,-0.2242174745,-0.6152897477,0.4672029912,0.0918213353,0.2583209872,0.630194068,-0.0001212188,0.0438775904,0.3747947514,-0.0809114501,-0.3682197928,-0.5027552843,-0.5078126788,-0.0790523142,0.2648388743,0.2292865217,-0.3196886182,0.4285567403,0.0791531503,-0.325060308,0.3382033706,0.6332944036,0.0820835829,-0.0768375695,0.0945120901,-0.1623590142,0.3016184568,-0.1488979459,-0.0706101805,0.4869538546,0.2155133337,-0.0386384055,0.0549374111,-0.0216429308,-0.0271049347,-0.0117052123,0.1798272282,-0.0742924362,-0.1101848111,-0.4340969324,0.4241297245,0.352265209,0.6402148008,-0.1733739376,-0.2272690535,0.1454129517,-0.081264168,-0.4302360415,0.3003183305,0.3217078149,-0.1221241057,0.1870257258,-0.4078291059,-0.2597643733,-0.3686292768,0.0043727914,-0.416408807,0.1270349473,-0.2663984001,0.131530717,0.2910034359,0.0152950874,0.3388897479,0.2237246931,-0.4470455348,-0.0075067496,-0.6929902434,0.1829680651,-0.1482103318,-0.0363469273,0.0934938565,0.3697319627,0.3174377084,0.0368965492,-0.2863053977,0.4065686166,0.0869918466,0.3421801031,0.0432393365,0.1872974336,0.3242341876,0.1837179959,-0.0132212332,-0.0930793434,-0.195083186,-0.2066059262,-0.1330894083,0.1809051484,0.0973416939,-0.2702093422,-0.391297996,0.1465577036,-0.4139411747,0.009193588,0.1225641072,0.4541309178,-0.2755913138,-0.0242432654,-0.0385249332,0.2208224833,-0.0965710506,-0.1626555324,-0.066012606,0.2261447608,-0.2326459587,0.1317702383,0.3381001651,-0.1813087612,0.1413071007,-0.164917931,-0.066844523,-0.2156009972,-0.0086505944,-0.2676545978,-0.1067192107,0.4307493269,0.0776933208,0.0562675633,0.1243787333,-0.3719575107,-0.0352643579,0.0133798691,-0.5714361072,-0.3518609107,0.0621165261,0.0705782101,-0.1204286665,0.0219029319,-0.3443544507,0.0012689885,-0.0043544434,-0.1800432205,-0.1748710871,-0.2704450488,-0.3334649503,-0.3149197698,0.3458762765,0.451456517,-0.5566045642,-0.0270752627,0.0968400612,0.0743105635,0.1674599946,0.3314805627,-0.1907550693,-0.0951525122,-0.2181226611,-0.1000668108,0.0956019834,-0.3112884164,-0.2985938787,-0.0033583376,-0.0194073506,-0.0483571663,-0.2369234562,-0.1034136266,0.013775479,0.0616033152,0.6465601921,0.2157903463,-0.128316775,-0.1761939228,-0.1502933949,-0.276214689,-0.2542467415,0.0869560763,0.1214700714,0.0059731621,0.2363967299,-0.4764141738,0.070228681,0.041881226,-0.0073060701,0.3011595905,0.3711313903,-0.0534163527,-0.1965114772,0.2541643977,-0.4512370229,0.3701867163,-0.0718612894,0.2864062488,-0.2534539998,0.1539354026,-0.2667068243,0.1449401677,-0.0538807735,-0.1338416189,-0.0591719225,0.0979989544,0.2250677347,0.2401248366,-0.1199853867,0.5677842498,0.135291785,0.1387476027,-0.2335021943,0.2586082816,0.0218197275,0.0688075498,0.1329162866,0.0106323007,0.2142108232,-0.2170554101,-0.0641218349,0.318633914,0.2173629254,0.0417961814,0.057461638,0.1577191204,0.2542841733,-0.3891916275,-0.0997398421,0.0472882427,-0.0086243441,-0.0252697486,-0.0405539423,0.2425730079,0.0648595691,0.4205703437,0.1400236785,0.0894204974,0.3215745389,0.0298587233,-0.0160045065,0.179474324,0.1751997471,0.2178916037,0.0230622217,-0.0838506073,-0.3602313399,0.1493657082,-0.1642055213,0.0495590158,-0.1129640341,0.1247549057,-0.1547979712,-0.0220802911,0.0700044334,0.3708671927,0.2201276124,0.0490480177,-0.2005204111,0.1874097735,0.0605272613,-0.2058147639,-0.0445009433,0.1096539572,0.3646667898,0.2655833662,0.426687777,-0.0487633571,-0.095108144,-0.3787639439,-0.0358973593,0.2614510655,-0.27078861,0.4008783698,-0.2434709668,-0.3584270775,0.0938047096,-0.3161964118,-0.3449476063,-0.2649385929,-0.1177102774,0.4375137389,0.1163083985,0.0623287521,-0.2111103535,0.3716548681,0.0254304223,-0.5300382972,-0.023688294,-0.2573983371,-0.3198367953,0.015924165,0.3333905041,-0.2213886827,0.0698618591,-0.2547039092,0.0297967102,-0.1255459189,-0.2030438036,0.161564365,0.1523073763,0.3011630476,-0.0669810548,0.2267765105,-0.2425041497,0.0940599963,0.2750999629,0.067961067,0.1023404077,0.048934143,-0.0400866605,0.0441079624,0.0249122102,-0.2555710673,-0.4800460041,-0.3582805991,0.1784227639,0.0196136292,0.3458847404,0.2960084677,0.0382493213,-0.0866623893,0.0641809925,-0.0769640952,-0.1533749849,-0.3865545392,0.1724119484,-0.2786794007,-0.2431341708,0.2509688139,0.0747392997,0.384760499,0.0763949454,-0.6094400883,-0.1564110816,-0.1264603585,0.1048995107,0.0258263517,-0.2825921476,0.3334425092,-0.0882083923,-0.0051477673,0.216064468,-0.4404521585,0.0323337875,0.3356822431,0.3913239837,0.2737467587,0.3707469404,-0.1709450483,0.4239256084,0.2026095539,0.0485433601,0.3296657801,-0.1572235525,0.3341207802,-0.0892179385,-0.2969244123,-0.2172842175,-0.2103471756,-0.3245612681,0.1617722362,0.2636647522,-0.0344154313,-0.4967108071,0.1173620149,-0.0801406428,-0.3276883364,0.0556484796,0.1679751873,0.1110880598,0.0038787234,-0.0378257744,0.2212286294,-0.2119385451,-0.3172354698,0.3629997075,0.0803808197,0.0725494176,-0.1304673105,-0.0161201619,-0.8267481923,0.3045907319,0.0043531074,0.3990243375,-0.3158361316,0.0614299923,0.0791887119,-0.1068061665,0.1192461029,-0.3070542216,-0.0818270966,0.3087825179,0.0600856058,-0.2109733224,-0.2092314512,0.1269949377,0.1855375469,0.4327666461,0.3602838516,0.013694901,0.0053407452,0.0741606504,0.1480065137,-0.0896288231,0.0230829157,-0.3215021193,-0.4804897308,-0.2376385331,-0.0714276284,-0.062724866,0.298281461,0.1305000633,0.0849203244,0.2911790609,-0.2723354399,0.2081023902,0.0922953635,0.0188673791,0.1057648584,0.1027646288,0.0991389379,0.0624909215,0.3354113996,0.9395033717,0.1009224579,-0.120416373,0.3544933796,0.3962078691,0.0735524446,0.2287358344,-0.051833827,0.1031669229,-0.0455486812,0.1310721934,0.0472777598,0.3104491532,-0.2412313968,-0.0123398695,-0.844632566,-0.1633535475,0.3940396309,-0.1157400385,-0.0963782668,-0.0885319561,0.191869244,-0.1223828048,0.0664271712,-0.0798335299,0.9389856458,-0.2157409042,-0.1139465198,0.0964188576,-0.147055313,0.4845153391,-0.5781700015,0.4780519307,-0.3271455765,-0.1596903652,-0.2196512967,-0.2689325809,-0.1141616106,-0.1754051447,-0.002866735,0.4294499457,-0.1375024319,0.1797110438,0.4103512764,0.0312496778,0.1668056101,0.0313799717,-0.1563229114,0.0742265582,-0.0684729144,0.3795146346,-0.0753371939,-0.0879741907,-0.2208290249,-0.0232577417,-0.3305442631,0.2558438778,-0.7046300769,0.0951708704,0.0003454578,-0.2027993649,0.1387290508,0.1491003036,0.2019768804,0.1502197832,-0.3103254139,0.2650389969,-0.0473067723,-0.1506277919,-0.009702215,-0.0158758238,0.010616634,-0.0578824542,-0.0800179094,0.4451651275,-0.0131923249,-0.4016885757,0.0497705564,0.0414262153,0.0156010203,-0.0672010556,0.1310195774,-0.0913955346,0.0334716663,0.0690084398,0.0271308459,0.1205178499,0.1539868563,-0.0196321197,0.2282925695,-0.259336561,-0.0157153793,0.3636079133,-0.0135959983,-0.0319505855,0.1860428751,0.4898221791,-0.0020252562,-0.0705311596,-0.1395792663,0.2126883119,-0.3972610533,0.1164028496,-0.0432341732,0.1496537328,0.0333801284,0.3472812772,0.3143411875,0.0317257643,0.3876162767,-0.5564649701,-0.2944401801,0.0544478521,-0.1690248102,0.184166044,0.1040887386,-0.1632938981,-0.2276393175,-0.0018326067,-0.1922600567,0.0723112226,-0.3587886393,-0.175903365,0.1726374626,-0.1161857471,0.5185421109,-0.1368280649,-0.0479452088,-0.0184052158,-0.3581270874,-0.0564916581,-0.0145608978,0.2017893344,0.1712418199,-0.4430913627,-0.1755434275,-0.077528283,0.0485689603,0.0543948784,0.1847438067,0.1984396875,-0.1803075075,-0.3293845952,-0.4175056219,0.0081544761,-0.3177674711,0.3675778508,0.2319299281,0.4606673419,0.2174215019,0.2435667217,-0.0169100836,-0.1336589158,-0.117658779,0.2129968703,-0.053510312,-0.2645329237,0.1965231746,-0.4327823818,0.0601596422,-0.2405550629,0.2324357629,0.0786642805,-0.2138750255,0.14287664,0.1881443113,0.0525172353,-0.0704044029,-0.0328343585,0.1413883418,-0.0272031706,0.0310485549,0.2250308245,0.386873275,-0.1784227788,0.0799675807,0.2459060699,0.5618061423,0.1223886907,0.1421130449,-0.1236973405,-0.1237056777,0.1257034093,0.4094533622,0.0089930175,0.0327677615,0.0211193711,-0.3650708497,-0.0349260531,-0.0059551336,0.0183139034,0.0332748108,-0.5767639279,0.1064377353,0.1411761791,-0.1795779765,-0.1717036068,-0.0894220397,0.4524857998,-0.2100297809,-0.0174033958,-0.3908969462,0.1880070865,-0.0242402554,-0.039714206,0.1182918698,-0.4415909052,-0.1305991709,0.2427270114,0.0069916546,-0.1593962014,-0.0046678944,0.0859766155,-0.2388294488,-0.3000748754,-0.0755934119,0.3069646955,-0.0426480062,-0.0936480537,0.4497712553,0.0273359865,-0.1706606746,0.1557034552,0.0153890671,0.62725842,0.2654240429,0.1823752671,0.1087691635,-0.1266422421,0.0071312743,0.1697154045,0.0883647799,0.1900411397,-0.332215935,0.1685412526,0.078648366,0.0258462951,0.2831476927,-0.092812404,0.3121075034,0.0294574928,0.0179795437,-0.1460418701,0.1118161753,-0.1592104733,0.0422839858,-0.31451267,0.178357631,0.3862515092,-0.140285328,0.2693823576,0.1543028802,0.008642192,-0.178006798,0.4599909782,0.3405545056,0.2323293835,-0.3487294912,-0.1899100095,-0.2473540008,-0.0055192844,-0.1115458161,0.2073093206,-0.0416926779,0.1502738595,-0.204635039,0.3032639623,-0.1217940003,0.3058781028,0.0848768055,0.0063988287,-0.2651689947,0.0663421229,0.0090942262,0.0498305373,-0.1537736654,-0.2624773383,0.0815571919,-0.3734006286,-0.0794994906,-0.1199191436,0.1511416584,0.0426995084,-0.2643913627,0.2991390824,0.2458395213,0.7336868048,0.2630758286,-0.0093748644,-0.1382427812,-0.3809232116,-0.0225659534,0.0093688183,0.3737497926,0.2838226855,0.0853671506,-0.2404976487,-0.385111481,0.4988894165,0.065247342,0.236622721,0.1650582105,-0.0864073858,-0.0182050411,0.0852971599,-0.0759951696,-0.1094401255,0.1904141009,0.1407316327,-0.1363411546,-0.3142247498,0.5675712824,-0.3454417884,0.1128251329,-0.1528615505,0.2606040537,0.3078807592,0.0205607451,-0.270526737,0.1239754856,0.2145230472,-0.1328070611,-0.4126018882,0.3355360031,-0.2914167941,-0.0648081899,-0.0611460693,0.1464544684,0.0038937069,-0.3008677363,-0.2765820324,-0.0266715754]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/798","title":"Cannot load TREC dataset: ConnectionError","comments":"Actually it's already fixed on the master branch since #740 \r\nI'll do the 1.1.3 release soon","body":"## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. ","comment_length":16,"text":"Cannot load TREC dataset: ConnectionError\n## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. \nActually it's already fixed on the master branch since #740 \r\nI'll do the 1.1.3 release soon","embeddings":[-0.1691725552,0.0168359298,0.0626651496,0.3633992374,0.2918767631,-0.1796784997,0.2199582756,-0.0434447527,-0.1838140339,0.0629585609,-0.3238992393,0.1360173374,0.0722290277,0.1577587724,-0.0579137243,-0.0945270881,-0.1901099831,-0.2000556588,-0.0688836575,0.2535367906,-0.1328548044,0.4753988683,-0.2073137164,0.2046911418,-0.4112499952,0.0120672574,0.2174304873,0.4217193425,-0.2242174745,-0.6152897477,0.4672029912,0.0918213353,0.2583209872,0.630194068,-0.0001212188,0.0438775904,0.3747947514,-0.0809114501,-0.3682197928,-0.5027552843,-0.5078126788,-0.0790523142,0.2648388743,0.2292865217,-0.3196886182,0.4285567403,0.0791531503,-0.325060308,0.3382033706,0.6332944036,0.0820835829,-0.0768375695,0.0945120901,-0.1623590142,0.3016184568,-0.1488979459,-0.0706101805,0.4869538546,0.2155133337,-0.0386384055,0.0549374111,-0.0216429308,-0.0271049347,-0.0117052123,0.1798272282,-0.0742924362,-0.1101848111,-0.4340969324,0.4241297245,0.352265209,0.6402148008,-0.1733739376,-0.2272690535,0.1454129517,-0.081264168,-0.4302360415,0.3003183305,0.3217078149,-0.1221241057,0.1870257258,-0.4078291059,-0.2597643733,-0.3686292768,0.0043727914,-0.416408807,0.1270349473,-0.2663984001,0.131530717,0.2910034359,0.0152950874,0.3388897479,0.2237246931,-0.4470455348,-0.0075067496,-0.6929902434,0.1829680651,-0.1482103318,-0.0363469273,0.0934938565,0.3697319627,0.3174377084,0.0368965492,-0.2863053977,0.4065686166,0.0869918466,0.3421801031,0.0432393365,0.1872974336,0.3242341876,0.1837179959,-0.0132212332,-0.0930793434,-0.195083186,-0.2066059262,-0.1330894083,0.1809051484,0.0973416939,-0.2702093422,-0.391297996,0.1465577036,-0.4139411747,0.009193588,0.1225641072,0.4541309178,-0.2755913138,-0.0242432654,-0.0385249332,0.2208224833,-0.0965710506,-0.1626555324,-0.066012606,0.2261447608,-0.2326459587,0.1317702383,0.3381001651,-0.1813087612,0.1413071007,-0.164917931,-0.066844523,-0.2156009972,-0.0086505944,-0.2676545978,-0.1067192107,0.4307493269,0.0776933208,0.0562675633,0.1243787333,-0.3719575107,-0.0352643579,0.0133798691,-0.5714361072,-0.3518609107,0.0621165261,0.0705782101,-0.1204286665,0.0219029319,-0.3443544507,0.0012689885,-0.0043544434,-0.1800432205,-0.1748710871,-0.2704450488,-0.3334649503,-0.3149197698,0.3458762765,0.451456517,-0.5566045642,-0.0270752627,0.0968400612,0.0743105635,0.1674599946,0.3314805627,-0.1907550693,-0.0951525122,-0.2181226611,-0.1000668108,0.0956019834,-0.3112884164,-0.2985938787,-0.0033583376,-0.0194073506,-0.0483571663,-0.2369234562,-0.1034136266,0.013775479,0.0616033152,0.6465601921,0.2157903463,-0.128316775,-0.1761939228,-0.1502933949,-0.276214689,-0.2542467415,0.0869560763,0.1214700714,0.0059731621,0.2363967299,-0.4764141738,0.070228681,0.041881226,-0.0073060701,0.3011595905,0.3711313903,-0.0534163527,-0.1965114772,0.2541643977,-0.4512370229,0.3701867163,-0.0718612894,0.2864062488,-0.2534539998,0.1539354026,-0.2667068243,0.1449401677,-0.0538807735,-0.1338416189,-0.0591719225,0.0979989544,0.2250677347,0.2401248366,-0.1199853867,0.5677842498,0.135291785,0.1387476027,-0.2335021943,0.2586082816,0.0218197275,0.0688075498,0.1329162866,0.0106323007,0.2142108232,-0.2170554101,-0.0641218349,0.318633914,0.2173629254,0.0417961814,0.057461638,0.1577191204,0.2542841733,-0.3891916275,-0.0997398421,0.0472882427,-0.0086243441,-0.0252697486,-0.0405539423,0.2425730079,0.0648595691,0.4205703437,0.1400236785,0.0894204974,0.3215745389,0.0298587233,-0.0160045065,0.179474324,0.1751997471,0.2178916037,0.0230622217,-0.0838506073,-0.3602313399,0.1493657082,-0.1642055213,0.0495590158,-0.1129640341,0.1247549057,-0.1547979712,-0.0220802911,0.0700044334,0.3708671927,0.2201276124,0.0490480177,-0.2005204111,0.1874097735,0.0605272613,-0.2058147639,-0.0445009433,0.1096539572,0.3646667898,0.2655833662,0.426687777,-0.0487633571,-0.095108144,-0.3787639439,-0.0358973593,0.2614510655,-0.27078861,0.4008783698,-0.2434709668,-0.3584270775,0.0938047096,-0.3161964118,-0.3449476063,-0.2649385929,-0.1177102774,0.4375137389,0.1163083985,0.0623287521,-0.2111103535,0.3716548681,0.0254304223,-0.5300382972,-0.023688294,-0.2573983371,-0.3198367953,0.015924165,0.3333905041,-0.2213886827,0.0698618591,-0.2547039092,0.0297967102,-0.1255459189,-0.2030438036,0.161564365,0.1523073763,0.3011630476,-0.0669810548,0.2267765105,-0.2425041497,0.0940599963,0.2750999629,0.067961067,0.1023404077,0.048934143,-0.0400866605,0.0441079624,0.0249122102,-0.2555710673,-0.4800460041,-0.3582805991,0.1784227639,0.0196136292,0.3458847404,0.2960084677,0.0382493213,-0.0866623893,0.0641809925,-0.0769640952,-0.1533749849,-0.3865545392,0.1724119484,-0.2786794007,-0.2431341708,0.2509688139,0.0747392997,0.384760499,0.0763949454,-0.6094400883,-0.1564110816,-0.1264603585,0.1048995107,0.0258263517,-0.2825921476,0.3334425092,-0.0882083923,-0.0051477673,0.216064468,-0.4404521585,0.0323337875,0.3356822431,0.3913239837,0.2737467587,0.3707469404,-0.1709450483,0.4239256084,0.2026095539,0.0485433601,0.3296657801,-0.1572235525,0.3341207802,-0.0892179385,-0.2969244123,-0.2172842175,-0.2103471756,-0.3245612681,0.1617722362,0.2636647522,-0.0344154313,-0.4967108071,0.1173620149,-0.0801406428,-0.3276883364,0.0556484796,0.1679751873,0.1110880598,0.0038787234,-0.0378257744,0.2212286294,-0.2119385451,-0.3172354698,0.3629997075,0.0803808197,0.0725494176,-0.1304673105,-0.0161201619,-0.8267481923,0.3045907319,0.0043531074,0.3990243375,-0.3158361316,0.0614299923,0.0791887119,-0.1068061665,0.1192461029,-0.3070542216,-0.0818270966,0.3087825179,0.0600856058,-0.2109733224,-0.2092314512,0.1269949377,0.1855375469,0.4327666461,0.3602838516,0.013694901,0.0053407452,0.0741606504,0.1480065137,-0.0896288231,0.0230829157,-0.3215021193,-0.4804897308,-0.2376385331,-0.0714276284,-0.062724866,0.298281461,0.1305000633,0.0849203244,0.2911790609,-0.2723354399,0.2081023902,0.0922953635,0.0188673791,0.1057648584,0.1027646288,0.0991389379,0.0624909215,0.3354113996,0.9395033717,0.1009224579,-0.120416373,0.3544933796,0.3962078691,0.0735524446,0.2287358344,-0.051833827,0.1031669229,-0.0455486812,0.1310721934,0.0472777598,0.3104491532,-0.2412313968,-0.0123398695,-0.844632566,-0.1633535475,0.3940396309,-0.1157400385,-0.0963782668,-0.0885319561,0.191869244,-0.1223828048,0.0664271712,-0.0798335299,0.9389856458,-0.2157409042,-0.1139465198,0.0964188576,-0.147055313,0.4845153391,-0.5781700015,0.4780519307,-0.3271455765,-0.1596903652,-0.2196512967,-0.2689325809,-0.1141616106,-0.1754051447,-0.002866735,0.4294499457,-0.1375024319,0.1797110438,0.4103512764,0.0312496778,0.1668056101,0.0313799717,-0.1563229114,0.0742265582,-0.0684729144,0.3795146346,-0.0753371939,-0.0879741907,-0.2208290249,-0.0232577417,-0.3305442631,0.2558438778,-0.7046300769,0.0951708704,0.0003454578,-0.2027993649,0.1387290508,0.1491003036,0.2019768804,0.1502197832,-0.3103254139,0.2650389969,-0.0473067723,-0.1506277919,-0.009702215,-0.0158758238,0.010616634,-0.0578824542,-0.0800179094,0.4451651275,-0.0131923249,-0.4016885757,0.0497705564,0.0414262153,0.0156010203,-0.0672010556,0.1310195774,-0.0913955346,0.0334716663,0.0690084398,0.0271308459,0.1205178499,0.1539868563,-0.0196321197,0.2282925695,-0.259336561,-0.0157153793,0.3636079133,-0.0135959983,-0.0319505855,0.1860428751,0.4898221791,-0.0020252562,-0.0705311596,-0.1395792663,0.2126883119,-0.3972610533,0.1164028496,-0.0432341732,0.1496537328,0.0333801284,0.3472812772,0.3143411875,0.0317257643,0.3876162767,-0.5564649701,-0.2944401801,0.0544478521,-0.1690248102,0.184166044,0.1040887386,-0.1632938981,-0.2276393175,-0.0018326067,-0.1922600567,0.0723112226,-0.3587886393,-0.175903365,0.1726374626,-0.1161857471,0.5185421109,-0.1368280649,-0.0479452088,-0.0184052158,-0.3581270874,-0.0564916581,-0.0145608978,0.2017893344,0.1712418199,-0.4430913627,-0.1755434275,-0.077528283,0.0485689603,0.0543948784,0.1847438067,0.1984396875,-0.1803075075,-0.3293845952,-0.4175056219,0.0081544761,-0.3177674711,0.3675778508,0.2319299281,0.4606673419,0.2174215019,0.2435667217,-0.0169100836,-0.1336589158,-0.117658779,0.2129968703,-0.053510312,-0.2645329237,0.1965231746,-0.4327823818,0.0601596422,-0.2405550629,0.2324357629,0.0786642805,-0.2138750255,0.14287664,0.1881443113,0.0525172353,-0.0704044029,-0.0328343585,0.1413883418,-0.0272031706,0.0310485549,0.2250308245,0.386873275,-0.1784227788,0.0799675807,0.2459060699,0.5618061423,0.1223886907,0.1421130449,-0.1236973405,-0.1237056777,0.1257034093,0.4094533622,0.0089930175,0.0327677615,0.0211193711,-0.3650708497,-0.0349260531,-0.0059551336,0.0183139034,0.0332748108,-0.5767639279,0.1064377353,0.1411761791,-0.1795779765,-0.1717036068,-0.0894220397,0.4524857998,-0.2100297809,-0.0174033958,-0.3908969462,0.1880070865,-0.0242402554,-0.039714206,0.1182918698,-0.4415909052,-0.1305991709,0.2427270114,0.0069916546,-0.1593962014,-0.0046678944,0.0859766155,-0.2388294488,-0.3000748754,-0.0755934119,0.3069646955,-0.0426480062,-0.0936480537,0.4497712553,0.0273359865,-0.1706606746,0.1557034552,0.0153890671,0.62725842,0.2654240429,0.1823752671,0.1087691635,-0.1266422421,0.0071312743,0.1697154045,0.0883647799,0.1900411397,-0.332215935,0.1685412526,0.078648366,0.0258462951,0.2831476927,-0.092812404,0.3121075034,0.0294574928,0.0179795437,-0.1460418701,0.1118161753,-0.1592104733,0.0422839858,-0.31451267,0.178357631,0.3862515092,-0.140285328,0.2693823576,0.1543028802,0.008642192,-0.178006798,0.4599909782,0.3405545056,0.2323293835,-0.3487294912,-0.1899100095,-0.2473540008,-0.0055192844,-0.1115458161,0.2073093206,-0.0416926779,0.1502738595,-0.204635039,0.3032639623,-0.1217940003,0.3058781028,0.0848768055,0.0063988287,-0.2651689947,0.0663421229,0.0090942262,0.0498305373,-0.1537736654,-0.2624773383,0.0815571919,-0.3734006286,-0.0794994906,-0.1199191436,0.1511416584,0.0426995084,-0.2643913627,0.2991390824,0.2458395213,0.7336868048,0.2630758286,-0.0093748644,-0.1382427812,-0.3809232116,-0.0225659534,0.0093688183,0.3737497926,0.2838226855,0.0853671506,-0.2404976487,-0.385111481,0.4988894165,0.065247342,0.236622721,0.1650582105,-0.0864073858,-0.0182050411,0.0852971599,-0.0759951696,-0.1094401255,0.1904141009,0.1407316327,-0.1363411546,-0.3142247498,0.5675712824,-0.3454417884,0.1128251329,-0.1528615505,0.2606040537,0.3078807592,0.0205607451,-0.270526737,0.1239754856,0.2145230472,-0.1328070611,-0.4126018882,0.3355360031,-0.2914167941,-0.0648081899,-0.0611460693,0.1464544684,0.0038937069,-0.3008677363,-0.2765820324,-0.0266715754]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/798","title":"Cannot load TREC dataset: ConnectionError","comments":"Hi\nthanks, but I did tried to install from the pip install git+... and it does\nnot work for me,. thanks for the help. I have the same issue with wmt16,\n\"ro-en\"\nthanks.\nBest\nRabeeh\n\nOn Mon, Nov 16, 2020 at 10:29 AM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> Actually it's already fixed on the master branch since #740\n> <https:\/\/github.com\/huggingface\/datasets\/pull\/740>\n> I'll do the 1.1.3 release soon\n>\n> \u2014\n> You are receiving this because you commented.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/798#issuecomment-727854736>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ABP4ZCEUBJKPOCLABXCKMPDSQDWH3ANCNFSM4TJBUKSA>\n> .\n>\n","body":"## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. ","comment_length":98,"text":"Cannot load TREC dataset: ConnectionError\n## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. \nHi\nthanks, but I did tried to install from the pip install git+... and it does\nnot work for me,. thanks for the help. I have the same issue with wmt16,\n\"ro-en\"\nthanks.\nBest\nRabeeh\n\nOn Mon, Nov 16, 2020 at 10:29 AM Quentin Lhoest <notifications@github.com>\nwrote:\n\n> Actually it's already fixed on the master branch since #740\n> <https:\/\/github.com\/huggingface\/datasets\/pull\/740>\n> I'll do the 1.1.3 release soon\n>\n> \u2014\n> You are receiving this because you commented.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/datasets\/issues\/798#issuecomment-727854736>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ABP4ZCEUBJKPOCLABXCKMPDSQDWH3ANCNFSM4TJBUKSA>\n> .\n>\n","embeddings":[-0.1691725552,0.0168359298,0.0626651496,0.3633992374,0.2918767631,-0.1796784997,0.2199582756,-0.0434447527,-0.1838140339,0.0629585609,-0.3238992393,0.1360173374,0.0722290277,0.1577587724,-0.0579137243,-0.0945270881,-0.1901099831,-0.2000556588,-0.0688836575,0.2535367906,-0.1328548044,0.4753988683,-0.2073137164,0.2046911418,-0.4112499952,0.0120672574,0.2174304873,0.4217193425,-0.2242174745,-0.6152897477,0.4672029912,0.0918213353,0.2583209872,0.630194068,-0.0001212188,0.0438775904,0.3747947514,-0.0809114501,-0.3682197928,-0.5027552843,-0.5078126788,-0.0790523142,0.2648388743,0.2292865217,-0.3196886182,0.4285567403,0.0791531503,-0.325060308,0.3382033706,0.6332944036,0.0820835829,-0.0768375695,0.0945120901,-0.1623590142,0.3016184568,-0.1488979459,-0.0706101805,0.4869538546,0.2155133337,-0.0386384055,0.0549374111,-0.0216429308,-0.0271049347,-0.0117052123,0.1798272282,-0.0742924362,-0.1101848111,-0.4340969324,0.4241297245,0.352265209,0.6402148008,-0.1733739376,-0.2272690535,0.1454129517,-0.081264168,-0.4302360415,0.3003183305,0.3217078149,-0.1221241057,0.1870257258,-0.4078291059,-0.2597643733,-0.3686292768,0.0043727914,-0.416408807,0.1270349473,-0.2663984001,0.131530717,0.2910034359,0.0152950874,0.3388897479,0.2237246931,-0.4470455348,-0.0075067496,-0.6929902434,0.1829680651,-0.1482103318,-0.0363469273,0.0934938565,0.3697319627,0.3174377084,0.0368965492,-0.2863053977,0.4065686166,0.0869918466,0.3421801031,0.0432393365,0.1872974336,0.3242341876,0.1837179959,-0.0132212332,-0.0930793434,-0.195083186,-0.2066059262,-0.1330894083,0.1809051484,0.0973416939,-0.2702093422,-0.391297996,0.1465577036,-0.4139411747,0.009193588,0.1225641072,0.4541309178,-0.2755913138,-0.0242432654,-0.0385249332,0.2208224833,-0.0965710506,-0.1626555324,-0.066012606,0.2261447608,-0.2326459587,0.1317702383,0.3381001651,-0.1813087612,0.1413071007,-0.164917931,-0.066844523,-0.2156009972,-0.0086505944,-0.2676545978,-0.1067192107,0.4307493269,0.0776933208,0.0562675633,0.1243787333,-0.3719575107,-0.0352643579,0.0133798691,-0.5714361072,-0.3518609107,0.0621165261,0.0705782101,-0.1204286665,0.0219029319,-0.3443544507,0.0012689885,-0.0043544434,-0.1800432205,-0.1748710871,-0.2704450488,-0.3334649503,-0.3149197698,0.3458762765,0.451456517,-0.5566045642,-0.0270752627,0.0968400612,0.0743105635,0.1674599946,0.3314805627,-0.1907550693,-0.0951525122,-0.2181226611,-0.1000668108,0.0956019834,-0.3112884164,-0.2985938787,-0.0033583376,-0.0194073506,-0.0483571663,-0.2369234562,-0.1034136266,0.013775479,0.0616033152,0.6465601921,0.2157903463,-0.128316775,-0.1761939228,-0.1502933949,-0.276214689,-0.2542467415,0.0869560763,0.1214700714,0.0059731621,0.2363967299,-0.4764141738,0.070228681,0.041881226,-0.0073060701,0.3011595905,0.3711313903,-0.0534163527,-0.1965114772,0.2541643977,-0.4512370229,0.3701867163,-0.0718612894,0.2864062488,-0.2534539998,0.1539354026,-0.2667068243,0.1449401677,-0.0538807735,-0.1338416189,-0.0591719225,0.0979989544,0.2250677347,0.2401248366,-0.1199853867,0.5677842498,0.135291785,0.1387476027,-0.2335021943,0.2586082816,0.0218197275,0.0688075498,0.1329162866,0.0106323007,0.2142108232,-0.2170554101,-0.0641218349,0.318633914,0.2173629254,0.0417961814,0.057461638,0.1577191204,0.2542841733,-0.3891916275,-0.0997398421,0.0472882427,-0.0086243441,-0.0252697486,-0.0405539423,0.2425730079,0.0648595691,0.4205703437,0.1400236785,0.0894204974,0.3215745389,0.0298587233,-0.0160045065,0.179474324,0.1751997471,0.2178916037,0.0230622217,-0.0838506073,-0.3602313399,0.1493657082,-0.1642055213,0.0495590158,-0.1129640341,0.1247549057,-0.1547979712,-0.0220802911,0.0700044334,0.3708671927,0.2201276124,0.0490480177,-0.2005204111,0.1874097735,0.0605272613,-0.2058147639,-0.0445009433,0.1096539572,0.3646667898,0.2655833662,0.426687777,-0.0487633571,-0.095108144,-0.3787639439,-0.0358973593,0.2614510655,-0.27078861,0.4008783698,-0.2434709668,-0.3584270775,0.0938047096,-0.3161964118,-0.3449476063,-0.2649385929,-0.1177102774,0.4375137389,0.1163083985,0.0623287521,-0.2111103535,0.3716548681,0.0254304223,-0.5300382972,-0.023688294,-0.2573983371,-0.3198367953,0.015924165,0.3333905041,-0.2213886827,0.0698618591,-0.2547039092,0.0297967102,-0.1255459189,-0.2030438036,0.161564365,0.1523073763,0.3011630476,-0.0669810548,0.2267765105,-0.2425041497,0.0940599963,0.2750999629,0.067961067,0.1023404077,0.048934143,-0.0400866605,0.0441079624,0.0249122102,-0.2555710673,-0.4800460041,-0.3582805991,0.1784227639,0.0196136292,0.3458847404,0.2960084677,0.0382493213,-0.0866623893,0.0641809925,-0.0769640952,-0.1533749849,-0.3865545392,0.1724119484,-0.2786794007,-0.2431341708,0.2509688139,0.0747392997,0.384760499,0.0763949454,-0.6094400883,-0.1564110816,-0.1264603585,0.1048995107,0.0258263517,-0.2825921476,0.3334425092,-0.0882083923,-0.0051477673,0.216064468,-0.4404521585,0.0323337875,0.3356822431,0.3913239837,0.2737467587,0.3707469404,-0.1709450483,0.4239256084,0.2026095539,0.0485433601,0.3296657801,-0.1572235525,0.3341207802,-0.0892179385,-0.2969244123,-0.2172842175,-0.2103471756,-0.3245612681,0.1617722362,0.2636647522,-0.0344154313,-0.4967108071,0.1173620149,-0.0801406428,-0.3276883364,0.0556484796,0.1679751873,0.1110880598,0.0038787234,-0.0378257744,0.2212286294,-0.2119385451,-0.3172354698,0.3629997075,0.0803808197,0.0725494176,-0.1304673105,-0.0161201619,-0.8267481923,0.3045907319,0.0043531074,0.3990243375,-0.3158361316,0.0614299923,0.0791887119,-0.1068061665,0.1192461029,-0.3070542216,-0.0818270966,0.3087825179,0.0600856058,-0.2109733224,-0.2092314512,0.1269949377,0.1855375469,0.4327666461,0.3602838516,0.013694901,0.0053407452,0.0741606504,0.1480065137,-0.0896288231,0.0230829157,-0.3215021193,-0.4804897308,-0.2376385331,-0.0714276284,-0.062724866,0.298281461,0.1305000633,0.0849203244,0.2911790609,-0.2723354399,0.2081023902,0.0922953635,0.0188673791,0.1057648584,0.1027646288,0.0991389379,0.0624909215,0.3354113996,0.9395033717,0.1009224579,-0.120416373,0.3544933796,0.3962078691,0.0735524446,0.2287358344,-0.051833827,0.1031669229,-0.0455486812,0.1310721934,0.0472777598,0.3104491532,-0.2412313968,-0.0123398695,-0.844632566,-0.1633535475,0.3940396309,-0.1157400385,-0.0963782668,-0.0885319561,0.191869244,-0.1223828048,0.0664271712,-0.0798335299,0.9389856458,-0.2157409042,-0.1139465198,0.0964188576,-0.147055313,0.4845153391,-0.5781700015,0.4780519307,-0.3271455765,-0.1596903652,-0.2196512967,-0.2689325809,-0.1141616106,-0.1754051447,-0.002866735,0.4294499457,-0.1375024319,0.1797110438,0.4103512764,0.0312496778,0.1668056101,0.0313799717,-0.1563229114,0.0742265582,-0.0684729144,0.3795146346,-0.0753371939,-0.0879741907,-0.2208290249,-0.0232577417,-0.3305442631,0.2558438778,-0.7046300769,0.0951708704,0.0003454578,-0.2027993649,0.1387290508,0.1491003036,0.2019768804,0.1502197832,-0.3103254139,0.2650389969,-0.0473067723,-0.1506277919,-0.009702215,-0.0158758238,0.010616634,-0.0578824542,-0.0800179094,0.4451651275,-0.0131923249,-0.4016885757,0.0497705564,0.0414262153,0.0156010203,-0.0672010556,0.1310195774,-0.0913955346,0.0334716663,0.0690084398,0.0271308459,0.1205178499,0.1539868563,-0.0196321197,0.2282925695,-0.259336561,-0.0157153793,0.3636079133,-0.0135959983,-0.0319505855,0.1860428751,0.4898221791,-0.0020252562,-0.0705311596,-0.1395792663,0.2126883119,-0.3972610533,0.1164028496,-0.0432341732,0.1496537328,0.0333801284,0.3472812772,0.3143411875,0.0317257643,0.3876162767,-0.5564649701,-0.2944401801,0.0544478521,-0.1690248102,0.184166044,0.1040887386,-0.1632938981,-0.2276393175,-0.0018326067,-0.1922600567,0.0723112226,-0.3587886393,-0.175903365,0.1726374626,-0.1161857471,0.5185421109,-0.1368280649,-0.0479452088,-0.0184052158,-0.3581270874,-0.0564916581,-0.0145608978,0.2017893344,0.1712418199,-0.4430913627,-0.1755434275,-0.077528283,0.0485689603,0.0543948784,0.1847438067,0.1984396875,-0.1803075075,-0.3293845952,-0.4175056219,0.0081544761,-0.3177674711,0.3675778508,0.2319299281,0.4606673419,0.2174215019,0.2435667217,-0.0169100836,-0.1336589158,-0.117658779,0.2129968703,-0.053510312,-0.2645329237,0.1965231746,-0.4327823818,0.0601596422,-0.2405550629,0.2324357629,0.0786642805,-0.2138750255,0.14287664,0.1881443113,0.0525172353,-0.0704044029,-0.0328343585,0.1413883418,-0.0272031706,0.0310485549,0.2250308245,0.386873275,-0.1784227788,0.0799675807,0.2459060699,0.5618061423,0.1223886907,0.1421130449,-0.1236973405,-0.1237056777,0.1257034093,0.4094533622,0.0089930175,0.0327677615,0.0211193711,-0.3650708497,-0.0349260531,-0.0059551336,0.0183139034,0.0332748108,-0.5767639279,0.1064377353,0.1411761791,-0.1795779765,-0.1717036068,-0.0894220397,0.4524857998,-0.2100297809,-0.0174033958,-0.3908969462,0.1880070865,-0.0242402554,-0.039714206,0.1182918698,-0.4415909052,-0.1305991709,0.2427270114,0.0069916546,-0.1593962014,-0.0046678944,0.0859766155,-0.2388294488,-0.3000748754,-0.0755934119,0.3069646955,-0.0426480062,-0.0936480537,0.4497712553,0.0273359865,-0.1706606746,0.1557034552,0.0153890671,0.62725842,0.2654240429,0.1823752671,0.1087691635,-0.1266422421,0.0071312743,0.1697154045,0.0883647799,0.1900411397,-0.332215935,0.1685412526,0.078648366,0.0258462951,0.2831476927,-0.092812404,0.3121075034,0.0294574928,0.0179795437,-0.1460418701,0.1118161753,-0.1592104733,0.0422839858,-0.31451267,0.178357631,0.3862515092,-0.140285328,0.2693823576,0.1543028802,0.008642192,-0.178006798,0.4599909782,0.3405545056,0.2323293835,-0.3487294912,-0.1899100095,-0.2473540008,-0.0055192844,-0.1115458161,0.2073093206,-0.0416926779,0.1502738595,-0.204635039,0.3032639623,-0.1217940003,0.3058781028,0.0848768055,0.0063988287,-0.2651689947,0.0663421229,0.0090942262,0.0498305373,-0.1537736654,-0.2624773383,0.0815571919,-0.3734006286,-0.0794994906,-0.1199191436,0.1511416584,0.0426995084,-0.2643913627,0.2991390824,0.2458395213,0.7336868048,0.2630758286,-0.0093748644,-0.1382427812,-0.3809232116,-0.0225659534,0.0093688183,0.3737497926,0.2838226855,0.0853671506,-0.2404976487,-0.385111481,0.4988894165,0.065247342,0.236622721,0.1650582105,-0.0864073858,-0.0182050411,0.0852971599,-0.0759951696,-0.1094401255,0.1904141009,0.1407316327,-0.1363411546,-0.3142247498,0.5675712824,-0.3454417884,0.1128251329,-0.1528615505,0.2606040537,0.3078807592,0.0205607451,-0.270526737,0.1239754856,0.2145230472,-0.1328070611,-0.4126018882,0.3355360031,-0.2914167941,-0.0648081899,-0.0611460693,0.1464544684,0.0038937069,-0.3008677363,-0.2765820324,-0.0266715754]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/798","title":"Cannot load TREC dataset: ConnectionError","comments":"I just tested on google colab using\r\n```python\r\n!pip install git+https:\/\/github.com\/huggingface\/datasets.git\r\nfrom datasets import load_dataset\r\nload_dataset(\"trec\")\r\n```\r\nand it works.\r\nCan you detail how you got the issue even when using the latest version on master ?\r\n\r\nAlso about wmt we'll look into it, thanks for reporting !","body":"## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. ","comment_length":48,"text":"Cannot load TREC dataset: ConnectionError\n## Problem\r\nI cannot load \"trec\" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label')` returns <Response [302]>. \r\n* `requests.head('http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label', allow_redirects=True)` raises `requests.exceptions.TooManyRedirects: Exceeded 30 redirects.`\r\n* Opening `http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label' in a browser works, but opens a different address\r\n* Increasing max_redirects to 100 doesn't help\r\n\r\nAlso, while debugging I've seen that requesting 'https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/trec\/default\/1.1.0\/dataset_info.json' returns <Response [404]> before, but it doesn't raise any errors. Not sure if that's relevant.\r\n\r\n* datasets.__version__ == '1.1.2'\r\n* requests.__version__ == '2.24.0'\r\n\r\n## Error trace\r\n```\r\n>>> import datasets\r\n>>> datasets.__version__\r\n'1.1.2'\r\n>>> dataset = load_dataset(\"trec\", split=\"train\")\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/home\/przemyslaw\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/przemyslaw\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/trec\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7\/trec.py\", line 140, in _split_generators\r\n dl_files = dl_manager.download_and_extract(_URLs)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/download_manager.py\", line 179, in download\r\n num_proc=download_config.num_proc,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in map_nested\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/home\/przemyslaw\/.local\/lib\/python3.6\/site-packages\/datasets\/utils\/file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n```\r\n\r\nI would appreciate some suggestions here. \nI just tested on google colab using\r\n```python\r\n!pip install git+https:\/\/github.com\/huggingface\/datasets.git\r\nfrom datasets import load_dataset\r\nload_dataset(\"trec\")\r\n```\r\nand it works.\r\nCan you detail how you got the issue even when using the latest version on master ?\r\n\r\nAlso about wmt we'll look into it, thanks for reporting !","embeddings":[-0.1691725552,0.0168359298,0.0626651496,0.3633992374,0.2918767631,-0.1796784997,0.2199582756,-0.0434447527,-0.1838140339,0.0629585609,-0.3238992393,0.1360173374,0.0722290277,0.1577587724,-0.0579137243,-0.0945270881,-0.1901099831,-0.2000556588,-0.0688836575,0.2535367906,-0.1328548044,0.4753988683,-0.2073137164,0.2046911418,-0.4112499952,0.0120672574,0.2174304873,0.4217193425,-0.2242174745,-0.6152897477,0.4672029912,0.0918213353,0.2583209872,0.630194068,-0.0001212188,0.0438775904,0.3747947514,-0.0809114501,-0.3682197928,-0.5027552843,-0.5078126788,-0.0790523142,0.2648388743,0.2292865217,-0.3196886182,0.4285567403,0.0791531503,-0.325060308,0.3382033706,0.6332944036,0.0820835829,-0.0768375695,0.0945120901,-0.1623590142,0.3016184568,-0.1488979459,-0.0706101805,0.4869538546,0.2155133337,-0.0386384055,0.0549374111,-0.0216429308,-0.0271049347,-0.0117052123,0.1798272282,-0.0742924362,-0.1101848111,-0.4340969324,0.4241297245,0.352265209,0.6402148008,-0.1733739376,-0.2272690535,0.1454129517,-0.081264168,-0.4302360415,0.3003183305,0.3217078149,-0.1221241057,0.1870257258,-0.4078291059,-0.2597643733,-0.3686292768,0.0043727914,-0.416408807,0.1270349473,-0.2663984001,0.131530717,0.2910034359,0.0152950874,0.3388897479,0.2237246931,-0.4470455348,-0.0075067496,-0.6929902434,0.1829680651,-0.1482103318,-0.0363469273,0.0934938565,0.3697319627,0.3174377084,0.0368965492,-0.2863053977,0.4065686166,0.0869918466,0.3421801031,0.0432393365,0.1872974336,0.3242341876,0.1837179959,-0.0132212332,-0.0930793434,-0.195083186,-0.2066059262,-0.1330894083,0.1809051484,0.0973416939,-0.2702093422,-0.391297996,0.1465577036,-0.4139411747,0.009193588,0.1225641072,0.4541309178,-0.2755913138,-0.0242432654,-0.0385249332,0.2208224833,-0.0965710506,-0.1626555324,-0.066012606,0.2261447608,-0.2326459587,0.1317702383,0.3381001651,-0.1813087612,0.1413071007,-0.164917931,-0.066844523,-0.2156009972,-0.0086505944,-0.2676545978,-0.1067192107,0.4307493269,0.0776933208,0.0562675633,0.1243787333,-0.3719575107,-0.0352643579,0.0133798691,-0.5714361072,-0.3518609107,0.0621165261,0.0705782101,-0.1204286665,0.0219029319,-0.3443544507,0.0012689885,-0.0043544434,-0.1800432205,-0.1748710871,-0.2704450488,-0.3334649503,-0.3149197698,0.3458762765,0.451456517,-0.5566045642,-0.0270752627,0.0968400612,0.0743105635,0.1674599946,0.3314805627,-0.1907550693,-0.0951525122,-0.2181226611,-0.1000668108,0.0956019834,-0.3112884164,-0.2985938787,-0.0033583376,-0.0194073506,-0.0483571663,-0.2369234562,-0.1034136266,0.013775479,0.0616033152,0.6465601921,0.2157903463,-0.128316775,-0.1761939228,-0.1502933949,-0.276214689,-0.2542467415,0.0869560763,0.1214700714,0.0059731621,0.2363967299,-0.4764141738,0.070228681,0.041881226,-0.0073060701,0.3011595905,0.3711313903,-0.0534163527,-0.1965114772,0.2541643977,-0.4512370229,0.3701867163,-0.0718612894,0.2864062488,-0.2534539998,0.1539354026,-0.2667068243,0.1449401677,-0.0538807735,-0.1338416189,-0.0591719225,0.0979989544,0.2250677347,0.2401248366,-0.1199853867,0.5677842498,0.135291785,0.1387476027,-0.2335021943,0.2586082816,0.0218197275,0.0688075498,0.1329162866,0.0106323007,0.2142108232,-0.2170554101,-0.0641218349,0.318633914,0.2173629254,0.0417961814,0.057461638,0.1577191204,0.2542841733,-0.3891916275,-0.0997398421,0.0472882427,-0.0086243441,-0.0252697486,-0.0405539423,0.2425730079,0.0648595691,0.4205703437,0.1400236785,0.0894204974,0.3215745389,0.0298587233,-0.0160045065,0.179474324,0.1751997471,0.2178916037,0.0230622217,-0.0838506073,-0.3602313399,0.1493657082,-0.1642055213,0.0495590158,-0.1129640341,0.1247549057,-0.1547979712,-0.0220802911,0.0700044334,0.3708671927,0.2201276124,0.0490480177,-0.2005204111,0.1874097735,0.0605272613,-0.2058147639,-0.0445009433,0.1096539572,0.3646667898,0.2655833662,0.426687777,-0.0487633571,-0.095108144,-0.3787639439,-0.0358973593,0.2614510655,-0.27078861,0.4008783698,-0.2434709668,-0.3584270775,0.0938047096,-0.3161964118,-0.3449476063,-0.2649385929,-0.1177102774,0.4375137389,0.1163083985,0.0623287521,-0.2111103535,0.3716548681,0.0254304223,-0.5300382972,-0.023688294,-0.2573983371,-0.3198367953,0.015924165,0.3333905041,-0.2213886827,0.0698618591,-0.2547039092,0.0297967102,-0.1255459189,-0.2030438036,0.161564365,0.1523073763,0.3011630476,-0.0669810548,0.2267765105,-0.2425041497,0.0940599963,0.2750999629,0.067961067,0.1023404077,0.048934143,-0.0400866605,0.0441079624,0.0249122102,-0.2555710673,-0.4800460041,-0.3582805991,0.1784227639,0.0196136292,0.3458847404,0.2960084677,0.0382493213,-0.0866623893,0.0641809925,-0.0769640952,-0.1533749849,-0.3865545392,0.1724119484,-0.2786794007,-0.2431341708,0.2509688139,0.0747392997,0.384760499,0.0763949454,-0.6094400883,-0.1564110816,-0.1264603585,0.1048995107,0.0258263517,-0.2825921476,0.3334425092,-0.0882083923,-0.0051477673,0.216064468,-0.4404521585,0.0323337875,0.3356822431,0.3913239837,0.2737467587,0.3707469404,-0.1709450483,0.4239256084,0.2026095539,0.0485433601,0.3296657801,-0.1572235525,0.3341207802,-0.0892179385,-0.2969244123,-0.2172842175,-0.2103471756,-0.3245612681,0.1617722362,0.2636647522,-0.0344154313,-0.4967108071,0.1173620149,-0.0801406428,-0.3276883364,0.0556484796,0.1679751873,0.1110880598,0.0038787234,-0.0378257744,0.2212286294,-0.2119385451,-0.3172354698,0.3629997075,0.0803808197,0.0725494176,-0.1304673105,-0.0161201619,-0.8267481923,0.3045907319,0.0043531074,0.3990243375,-0.3158361316,0.0614299923,0.0791887119,-0.1068061665,0.1192461029,-0.3070542216,-0.0818270966,0.3087825179,0.0600856058,-0.2109733224,-0.2092314512,0.1269949377,0.1855375469,0.4327666461,0.3602838516,0.013694901,0.0053407452,0.0741606504,0.1480065137,-0.0896288231,0.0230829157,-0.3215021193,-0.4804897308,-0.2376385331,-0.0714276284,-0.062724866,0.298281461,0.1305000633,0.0849203244,0.2911790609,-0.2723354399,0.2081023902,0.0922953635,0.0188673791,0.1057648584,0.1027646288,0.0991389379,0.0624909215,0.3354113996,0.9395033717,0.1009224579,-0.120416373,0.3544933796,0.3962078691,0.0735524446,0.2287358344,-0.051833827,0.1031669229,-0.0455486812,0.1310721934,0.0472777598,0.3104491532,-0.2412313968,-0.0123398695,-0.844632566,-0.1633535475,0.3940396309,-0.1157400385,-0.0963782668,-0.0885319561,0.191869244,-0.1223828048,0.0664271712,-0.0798335299,0.9389856458,-0.2157409042,-0.1139465198,0.0964188576,-0.147055313,0.4845153391,-0.5781700015,0.4780519307,-0.3271455765,-0.1596903652,-0.2196512967,-0.2689325809,-0.1141616106,-0.1754051447,-0.002866735,0.4294499457,-0.1375024319,0.1797110438,0.4103512764,0.0312496778,0.1668056101,0.0313799717,-0.1563229114,0.0742265582,-0.0684729144,0.3795146346,-0.0753371939,-0.0879741907,-0.2208290249,-0.0232577417,-0.3305442631,0.2558438778,-0.7046300769,0.0951708704,0.0003454578,-0.2027993649,0.1387290508,0.1491003036,0.2019768804,0.1502197832,-0.3103254139,0.2650389969,-0.0473067723,-0.1506277919,-0.009702215,-0.0158758238,0.010616634,-0.0578824542,-0.0800179094,0.4451651275,-0.0131923249,-0.4016885757,0.0497705564,0.0414262153,0.0156010203,-0.0672010556,0.1310195774,-0.0913955346,0.0334716663,0.0690084398,0.0271308459,0.1205178499,0.1539868563,-0.0196321197,0.2282925695,-0.259336561,-0.0157153793,0.3636079133,-0.0135959983,-0.0319505855,0.1860428751,0.4898221791,-0.0020252562,-0.0705311596,-0.1395792663,0.2126883119,-0.3972610533,0.1164028496,-0.0432341732,0.1496537328,0.0333801284,0.3472812772,0.3143411875,0.0317257643,0.3876162767,-0.5564649701,-0.2944401801,0.0544478521,-0.1690248102,0.184166044,0.1040887386,-0.1632938981,-0.2276393175,-0.0018326067,-0.1922600567,0.0723112226,-0.3587886393,-0.175903365,0.1726374626,-0.1161857471,0.5185421109,-0.1368280649,-0.0479452088,-0.0184052158,-0.3581270874,-0.0564916581,-0.0145608978,0.2017893344,0.1712418199,-0.4430913627,-0.1755434275,-0.077528283,0.0485689603,0.0543948784,0.1847438067,0.1984396875,-0.1803075075,-0.3293845952,-0.4175056219,0.0081544761,-0.3177674711,0.3675778508,0.2319299281,0.4606673419,0.2174215019,0.2435667217,-0.0169100836,-0.1336589158,-0.117658779,0.2129968703,-0.053510312,-0.2645329237,0.1965231746,-0.4327823818,0.0601596422,-0.2405550629,0.2324357629,0.0786642805,-0.2138750255,0.14287664,0.1881443113,0.0525172353,-0.0704044029,-0.0328343585,0.1413883418,-0.0272031706,0.0310485549,0.2250308245,0.386873275,-0.1784227788,0.0799675807,0.2459060699,0.5618061423,0.1223886907,0.1421130449,-0.1236973405,-0.1237056777,0.1257034093,0.4094533622,0.0089930175,0.0327677615,0.0211193711,-0.3650708497,-0.0349260531,-0.0059551336,0.0183139034,0.0332748108,-0.5767639279,0.1064377353,0.1411761791,-0.1795779765,-0.1717036068,-0.0894220397,0.4524857998,-0.2100297809,-0.0174033958,-0.3908969462,0.1880070865,-0.0242402554,-0.039714206,0.1182918698,-0.4415909052,-0.1305991709,0.2427270114,0.0069916546,-0.1593962014,-0.0046678944,0.0859766155,-0.2388294488,-0.3000748754,-0.0755934119,0.3069646955,-0.0426480062,-0.0936480537,0.4497712553,0.0273359865,-0.1706606746,0.1557034552,0.0153890671,0.62725842,0.2654240429,0.1823752671,0.1087691635,-0.1266422421,0.0071312743,0.1697154045,0.0883647799,0.1900411397,-0.332215935,0.1685412526,0.078648366,0.0258462951,0.2831476927,-0.092812404,0.3121075034,0.0294574928,0.0179795437,-0.1460418701,0.1118161753,-0.1592104733,0.0422839858,-0.31451267,0.178357631,0.3862515092,-0.140285328,0.2693823576,0.1543028802,0.008642192,-0.178006798,0.4599909782,0.3405545056,0.2323293835,-0.3487294912,-0.1899100095,-0.2473540008,-0.0055192844,-0.1115458161,0.2073093206,-0.0416926779,0.1502738595,-0.204635039,0.3032639623,-0.1217940003,0.3058781028,0.0848768055,0.0063988287,-0.2651689947,0.0663421229,0.0090942262,0.0498305373,-0.1537736654,-0.2624773383,0.0815571919,-0.3734006286,-0.0794994906,-0.1199191436,0.1511416584,0.0426995084,-0.2643913627,0.2991390824,0.2458395213,0.7336868048,0.2630758286,-0.0093748644,-0.1382427812,-0.3809232116,-0.0225659534,0.0093688183,0.3737497926,0.2838226855,0.0853671506,-0.2404976487,-0.385111481,0.4988894165,0.065247342,0.236622721,0.1650582105,-0.0864073858,-0.0182050411,0.0852971599,-0.0759951696,-0.1094401255,0.1904141009,0.1407316327,-0.1363411546,-0.3142247498,0.5675712824,-0.3454417884,0.1128251329,-0.1528615505,0.2606040537,0.3078807592,0.0205607451,-0.270526737,0.1239754856,0.2145230472,-0.1328070611,-0.4126018882,0.3355360031,-0.2914167941,-0.0648081899,-0.0611460693,0.1464544684,0.0038937069,-0.3008677363,-0.2765820324,-0.0266715754]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/797","title":"Token classification labels are strings and we don't have the list of labels","comments":"Indeed. Pinging @stefan-it here if he want to give an expert opinion :)","body":"Not sure if this is an issue we want to fix or not, putting it here so it's not forgotten. Right now, in token classification datasets, the labels for NER, POS and the likes are typed as `Sequence` of `strings`, which is wrong in my opinion. These should be `Sequence` of `ClassLabel` or some types that gives easy access to the underlying labels.\r\n\r\nThe main problem for preprocessing those datasets is that the list of possible labels is not stored inside the `Dataset` object which makes converting the labels to IDs quite difficult (you either have to know the list of labels in advance or run a full pass through the dataset to get the list of labels, the `unique` method being useless with the type `Sequence[str]`).","comment_length":13,"text":"Token classification labels are strings and we don't have the list of labels\nNot sure if this is an issue we want to fix or not, putting it here so it's not forgotten. Right now, in token classification datasets, the labels for NER, POS and the likes are typed as `Sequence` of `strings`, which is wrong in my opinion. These should be `Sequence` of `ClassLabel` or some types that gives easy access to the underlying labels.\r\n\r\nThe main problem for preprocessing those datasets is that the list of possible labels is not stored inside the `Dataset` object which makes converting the labels to IDs quite difficult (you either have to know the list of labels in advance or run a full pass through the dataset to get the list of labels, the `unique` method being useless with the type `Sequence[str]`).\nIndeed. Pinging @stefan-it here if he want to give an expert opinion :)","embeddings":[0.0350031629,-0.0001607512,-0.0594194382,-0.0637514293,0.1976903826,-0.0139794052,0.3227498531,0.0851441175,-0.2420099229,0.2861992717,0.027930202,0.2858993113,-0.0211663228,0.2161299288,-0.0129341474,-0.0656497553,0.0784552693,0.0014890271,0.3787277043,-0.1776746213,-0.3578615487,0.1468468457,-0.2107651234,0.3935232162,-0.4533465803,-0.1961256117,-0.1360747963,-0.1901365072,-0.2056523114,-0.717294693,-0.0173183549,0.3722059131,0.2083736211,0.1923229843,-0.0000971424,-0.1899494827,0.2502624691,0.0374604948,-0.1334196627,-0.122222431,-0.1032280847,0.035943985,0.0585464686,-0.3775436878,-0.2080832422,-0.121500127,0.1692837775,0.0566639341,0.0629290193,0.2456317842,0.336435914,0.0005296491,0.0464358144,-0.1934509575,0.3015998602,0.0884997696,-0.1493266672,-0.240615055,0.3280116022,0.0654913113,-0.1279607117,0.217209518,0.1009202302,-0.2430006415,0.1507545859,0.1688295156,0.094254069,-0.3894981444,-0.0099224197,0.2055738419,0.3400144279,-0.1157231107,-0.3494036496,-0.0222643409,0.1492648721,-0.55862993,0.0675736889,0.0308769569,-0.0935627148,0.1358009428,0.0156937204,-0.0045716581,-0.1262572557,-0.0185899884,0.0877543017,0.4997266531,-0.0228032377,-0.027160218,-0.1904568374,-0.2228474319,0.16465725,0.1624568105,-0.0823743641,0.0737618133,-0.2438113987,-0.0868022889,-0.2363198549,0.1093855798,0.0477180965,0.1121126711,-0.0749481916,0.1055150256,0.0927462205,0.1348267943,-0.1117241383,0.395766139,0.5546521544,0.3922184706,0.0402105413,-0.2716320455,-0.2277437299,0.0391084552,-0.3517991006,0.2292408049,0.2031823099,0.016437659,-0.2477627844,-0.1257696897,-0.2415265143,0.1795412898,-0.1840477884,0.1864147484,0.2231725454,0.2613739967,-0.0266696494,-0.0424281582,0.007533059,-0.1255536079,0.0068766894,-0.3670130372,-0.1972128898,0.0084629767,-0.3366155326,0.0632828772,0.0512441136,0.1780784875,0.188061282,-0.1087742001,0.0617884584,0.0322531089,0.0635774881,0.0654324144,0.206646651,0.1297711134,-0.1021380126,0.1842687726,-0.0234312993,-0.3758253455,-0.1441647857,-0.1054843962,-0.3935013115,-0.132826522,-0.0113310553,0.3403909206,0.1344616562,-0.2970930934,-0.197969541,0.1881863624,-0.1231216118,-0.0341042057,-0.0664968863,-0.2689226866,-0.084322542,-0.101395689,0.0229330864,0.0369330198,-0.2277213186,-0.2570510805,0.0453710705,0.0199872926,0.4235149324,0.2583203018,0.0031816107,0.1109550893,-0.134654969,0.3529075086,0.2797967792,-0.3064860702,-0.0472906828,-0.0853219554,0.0627386644,-0.4125154316,0.0275139976,0.138301909,0.472097218,-0.1728260219,0.1596099585,0.0471032523,-0.0773183703,-0.0183031932,-0.1319992542,0.0647274852,0.0080448799,0.0919801891,0.0159502141,0.0428286567,-0.1423151791,0.1821323782,-0.0863258019,-0.1286321431,0.1470537633,0.1467291713,0.1927825361,0.1134335473,-0.0091354568,-0.0309921913,-0.2807364762,-0.0334076211,-0.3541069627,0.3298346698,0.2239772379,-0.2360616028,-0.2177780718,0.1501003653,-0.1359316558,-0.0540139601,0.3477763236,0.1591985077,0.0812196806,-0.0347421467,-0.0864059478,0.0008741764,-0.2795131803,0.1390859485,0.0881425366,-0.0992286652,-0.1602345407,0.0235174578,-0.0599102639,0.1723402888,0.1994087398,0.1601340175,-0.0373576172,0.2957814932,0.2755984664,-0.4649072587,0.1125853509,0.2467777729,0.1680390388,-0.2610549331,-0.414224565,0.0256004557,-0.1231020615,-0.003348663,-0.0160584785,0.270419836,0.2479393333,0.2177502215,-0.0145743629,0.1826540381,0.217083171,0.107642211,-0.2002450824,-0.159858942,0.0426862091,-0.0928429365,0.1601582617,0.1883615255,-0.4900699258,0.1938947141,0.3071890473,-0.0805614069,0.1081473008,0.0773414224,-0.099817872,0.0461558104,0.015513896,0.084458895,-0.2311723679,0.3433656394,-0.033774171,-0.1372735649,-0.2281221449,-0.1279218942,0.1324526668,0.2013268769,-0.0843024179,-0.2098189145,0.0737981498,-0.0397628956,-0.1248621643,-0.4825110435,0.0163275599,0.1923668236,-0.2007617205,0.048723951,-0.1972482204,-0.116203554,-0.0454851463,-0.2751347721,-0.2497377396,-0.3715057075,0.4299250841,-0.1366638392,-0.0957767889,0.3745796978,-0.0885076746,0.2739830911,0.0266593061,0.1468486935,-0.2292273492,-0.2800453603,-0.3183439672,0.2023639679,0.0767199844,0.0573961362,0.3979504108,-0.2423951328,0.192178458,-0.2986437082,-0.5859181285,0.217320472,-0.1486257911,0.0096909851,0.1501301676,-0.2273232937,-0.4433462322,-0.081498526,0.0971875489,-0.2823037505,-0.3094882369,0.0128482673,-0.2065055221,-0.2268408537,-0.1377842426,-0.4923896194,-0.2999449968,-0.127739802,0.0790113062,0.0238751825,0.1990058422,0.2407225072,-0.3758854866,0.04311217,-0.1707640588,0.0600910038,-0.2539617419,-0.3775146306,0.2286599874,-0.3630736172,-0.3764009476,0.0130171999,-0.1310437471,0.2553437054,-0.1377676576,0.1219348684,-0.0814180374,0.0544045158,-0.0126058664,0.3006524444,-0.0456376784,0.1174534634,0.1584851146,-0.2375655025,-0.0584896244,-0.0649566278,0.3345787525,-0.2344478369,0.2199874669,-0.1146834493,0.0500537567,-0.0140285455,0.4248697758,-0.0691383854,-0.258764118,0.0796913952,-0.0711845234,0.1409361362,-0.1137579754,-0.0783374757,-0.0410843976,0.2003465444,-0.0651737675,0.3507784009,-0.1385075897,-0.169291988,0.1241262108,0.5172884464,-0.2133411616,-0.2612123489,-0.0629026368,0.0954945236,0.2508298755,0.1019332707,-0.0373913608,-0.167249769,-0.2514786124,0.0446825735,0.0676524043,-0.0419918634,0.1585939229,-0.902545929,-0.0804271325,-0.2539156675,0.0811365619,0.29298684,0.2695068121,-0.070383437,0.0068732807,0.1629113406,0.1983059049,0.312925905,-0.1775045097,-0.1798855513,0.3331149817,-0.1368389577,0.0723095536,0.0321740694,-0.28112486,-0.3639214337,0.6006526351,0.0414530598,-0.2259044349,-0.2619999647,-0.0859967992,-0.2618111372,-0.2456054538,-0.2110067606,-0.0273071174,0.0845717862,-0.3046589196,0.1012179554,0.356225431,0.1125336662,-0.0867803693,-0.1078528911,-0.3242354095,0.0071839751,0.295410037,0.104176797,0.0336927585,0.1462771595,0.1312729567,0.0762681067,0.481806457,0.4538158774,0.0541323647,-0.0687093213,-0.3990893364,0.3069589138,-0.0735366046,0.2728540301,-0.0009970719,0.0457565077,0.1356070638,-0.3390246928,0.0966146067,-0.1931502819,0.4727392495,0.0362458825,-0.0298503302,-0.6357679963,-0.23104693,0.1302806884,0.1391617954,-0.2732615471,-0.3353086114,0.4254508018,-0.3225258291,0.4082541466,0.3264205456,0.8285477757,-0.2165777981,0.1639978886,0.0103779295,0.2324353158,0.6974747181,0.0035841707,-0.0251044184,-0.4769381583,0.020634437,0.0147459563,0.0212301817,-0.2591609359,0.3274283409,-0.1326738,0.4008939862,0.1156716123,0.1138783321,0.0812377632,0.1768794954,0.3932454288,-0.4638050795,-0.1236988977,0.3103972971,-0.0699491873,0.0394717529,-0.0411620997,0.0479836576,-0.0813491493,0.0207824465,-0.0454828478,-0.0221462809,-0.0086637102,-0.0787343755,0.216261372,0.0975541174,0.0470392145,0.1673613042,0.6818380952,0.2556539774,-0.1083639637,0.0573003031,-0.1246865541,-0.0476914793,-0.0938087627,-0.1128688082,0.4482224286,0.0470509604,0.0443302989,0.0971325263,0.0243537743,-0.073586762,-0.3714022338,-0.2500919998,0.484652102,-0.1897815019,0.4156513512,-0.2231244743,-0.1002846733,-0.2731494009,0.2529295087,-0.008078252,-0.067133233,0.0078922864,0.2857302725,0.0017045356,-0.0553830862,0.0018423004,0.0839053243,-0.2129363865,0.162195757,-0.044644177,-0.1614256352,-0.3116636574,0.2374245077,0.4482563734,-0.2493665367,0.1492907554,0.055360008,-0.2358972728,0.2110121995,0.5633289218,0.1453163773,0.0041717407,0.1349109262,-0.0351782292,-0.5713064671,0.1059853956,-0.2385664731,0.3235824108,0.1591055989,0.1993603408,-0.2249979079,-0.0104533108,-0.4885942936,-0.2514983714,-0.4939608574,0.2033024728,-0.0032749597,0.0052450318,0.0112457899,0.2404531687,0.2125934809,0.1743282378,0.0421953835,-0.3392443359,-0.286247164,0.0347928815,0.2760323882,0.2940199673,0.0031454079,0.2205540091,-0.0490705818,-0.187001586,0.047419142,0.1663818359,0.1112558991,0.279163003,0.2095111758,0.137746349,0.0837804675,0.1196728051,0.2841970623,0.0849598944,0.0563635156,0.0894140899,-0.2080334127,-0.1286296248,-0.0041142376,0.3540931046,0.1926195771,-0.0641748905,0.0181951746,-0.2109034359,0.1032622084,0.3507905602,-0.108611837,-0.124019593,0.009169491,-0.3190105557,0.3946101964,0.3906536996,-0.1114765853,-0.123971045,-0.078564167,0.1223995164,0.1569989771,0.3135912716,0.1439275742,0.0335731208,0.1115781441,-0.052386187,0.0757064596,-0.2655915022,-0.090999648,-0.107219182,0.0667005479,0.099166207,0.4958135188,-0.0042765155,0.0846451148,0.4054196477,0.1304116398,0.4477429986,0.2858557105,0.4066837132,0.3508921266,0.0672225729,-0.1436798126,0.2004576623,0.1806756407,0.1406303346,-0.1268150955,0.0696615651,0.1770754009,-0.3234837055,-0.107166931,0.2992774546,-0.2483650744,-0.3178417087,-0.5303170681,0.0101908455,0.0022008158,-0.2631137967,-0.1260016114,-0.1456173509,-0.1627977192,-0.0769602433,0.0715138689,-0.3848086298,-0.3624493778,0.204036653,0.1798235178,-0.1608282477,0.1660821885,-0.220979467,-0.1499871612,0.3793928623,0.1603470743,0.2623375952,0.1384985149,0.2793231606,-0.183774665,-0.1967742443,-0.1375142038,-0.0023552612,0.0836754963,0.1412356794,-0.2003842145,0.2575336993,0.1711284816,-0.1930637062,-0.1340712756,0.0104958769,0.1413371712,0.3460427225,0.4542584419,0.2282142788,-0.0471822061,-0.1883623451,0.1970095485,-0.4383941293,0.0962390006,0.0832787007,0.1558158398,0.1422765553,-0.2215391546,0.168950662,-0.0346190408,0.3042966127,0.2141024768,-0.1930206567,-0.1467295587,-0.1127393395,-0.3020224571,0.2415391654,0.1288704574,0.0971034169,0.0563373454,-0.0626509413,-0.0122432308,0.0118335867,-0.2484165132,0.0553066358,-0.1565613896,0.197577998,-0.3323502541,0.2531730533,0.1967790872,-0.0281871613,0.0860479847,-0.0213063769,-0.0030633721,-0.2175346017,0.0960169807,-0.0300132092,0.2845150232,-0.0224189833,0.1014717892,0.0939109772,0.2199846506,0.3727922142,0.0434934646,0.0393336974,-0.0560393706,-0.4385955334,-0.2883540988,0.1123439297,-0.0132215284,0.2901932895,0.1896495968,-0.120069325,-0.0578397661,-0.0809314996,0.1779896021,-0.072114937,-0.4625945985,0.4179562628,-0.1178072691,0.178084746,0.0219542086,0.1571649164,0.3349438012,0.1236732677,-0.1393987387,-0.6529505849,0.4633997381,-0.3174412549,-0.1197007298,-0.1023882776,0.0734986737,0.3855136037,0.0415465161,-0.2959670722,0.0067361025,0.360131979,-0.0762610212,-0.4154408276,-0.0022460716,0.0064337784,0.1619369835,-0.2334664762,0.2191654146,0.0959200189,0.0310722366,0.0834465027,-0.1430361867]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/797","title":"Token classification labels are strings and we don't have the list of labels","comments":"Related is https:\/\/github.com\/huggingface\/datasets\/pull\/636","body":"Not sure if this is an issue we want to fix or not, putting it here so it's not forgotten. Right now, in token classification datasets, the labels for NER, POS and the likes are typed as `Sequence` of `strings`, which is wrong in my opinion. These should be `Sequence` of `ClassLabel` or some types that gives easy access to the underlying labels.\r\n\r\nThe main problem for preprocessing those datasets is that the list of possible labels is not stored inside the `Dataset` object which makes converting the labels to IDs quite difficult (you either have to know the list of labels in advance or run a full pass through the dataset to get the list of labels, the `unique` method being useless with the type `Sequence[str]`).","comment_length":3,"text":"Token classification labels are strings and we don't have the list of labels\nNot sure if this is an issue we want to fix or not, putting it here so it's not forgotten. Right now, in token classification datasets, the labels for NER, POS and the likes are typed as `Sequence` of `strings`, which is wrong in my opinion. These should be `Sequence` of `ClassLabel` or some types that gives easy access to the underlying labels.\r\n\r\nThe main problem for preprocessing those datasets is that the list of possible labels is not stored inside the `Dataset` object which makes converting the labels to IDs quite difficult (you either have to know the list of labels in advance or run a full pass through the dataset to get the list of labels, the `unique` method being useless with the type `Sequence[str]`).\nRelated is https:\/\/github.com\/huggingface\/datasets\/pull\/636","embeddings":[0.0755310729,-0.0799544826,-0.0420272276,0.0135806985,0.2348721474,0.0576448627,0.288079828,0.0719854236,-0.1486560702,0.2549916208,-0.0774475113,0.2283372581,-0.0282218773,0.3258889019,0.1232515126,-0.1418097913,0.1076608747,0.0012035223,0.4015982747,-0.1471077502,-0.3520070314,0.2045272738,-0.2058588266,0.3849816322,-0.4847157598,-0.2040015906,-0.1309118867,-0.164705947,-0.2079819143,-0.7251663208,0.0566709861,0.3897967041,0.2277411371,0.1994657964,-0.000099972,-0.182755366,0.2364167422,0.0487466343,-0.1670109779,-0.1688988656,-0.0595357008,-0.0329083987,0.1024990231,-0.3272850215,-0.2307683975,-0.1570009291,0.1786465645,0.0497040898,0.1350328326,0.2011854798,0.296880722,0.1148574799,0.0867767856,-0.1718408465,0.2746346593,0.1877480298,-0.1748783141,-0.226794973,0.3306677341,0.0648146793,-0.1091674045,0.2237626165,0.0889683068,-0.2415721864,0.1583586931,0.2200432271,-0.0420874096,-0.3756918907,-0.0427009016,0.2831664979,0.3383015692,-0.1796706617,-0.4307259023,-0.1690694541,0.1180717573,-0.5898023248,0.09213119,0.0497876033,-0.0942507386,0.1646869779,-0.0469512045,-0.0044506127,-0.0895305723,-0.0484714285,0.1076706871,0.4045359492,-0.1058153734,-0.0514166653,-0.1222712323,-0.206692487,0.008106119,0.1406385899,-0.048430454,0.1325135827,-0.285680145,-0.1170440391,-0.2419831604,0.1599194556,0.1839010119,0.1588218212,-0.1587827802,0.0877850056,-0.0111068431,0.115377821,-0.1110702679,0.393943429,0.6009917259,0.3100237548,0.0906699896,-0.1890732795,-0.1126156151,0.0014402028,-0.3363901973,0.2961976826,0.1396992952,0.0309481509,-0.1684665829,-0.1394877434,-0.2112128139,0.163738966,-0.1387309581,0.1093179435,0.2631298602,0.3080347478,-0.0839489251,-0.035269253,0.0338571668,-0.1173383296,0.0335149728,-0.3808597326,-0.2181453407,-0.0275092665,-0.325158,0.1289238781,0.0705821812,0.072751388,0.2719308734,-0.0728559271,0.0454622023,0.0309191551,-0.0202546902,0.0631063059,0.1177740991,0.1517551541,-0.1070421934,0.1800435632,0.0202697571,-0.3578546047,-0.1692445874,-0.1879466921,-0.4213067889,-0.1164273396,-0.0128177358,0.2891223133,0.0611739047,-0.3009089828,-0.4125224054,0.2202513516,-0.1456914544,0.0406402014,-0.0448528789,-0.234956786,-0.1504900157,-0.0861688256,0.1077341512,0.0936342031,-0.2191357315,-0.3532776237,0.1459342986,-0.0539470799,0.3317972422,0.321477592,0.0433978364,0.1016717702,-0.152334407,0.392164588,0.2442564368,-0.2834664285,-0.0330986753,-0.1448730528,0.0959314555,-0.3251680732,0.0213491041,0.0795537457,0.4945345223,-0.1944049448,0.1663547158,0.0648788512,-0.0803058147,-0.0513417237,-0.0587053448,0.0324293971,-0.0291811693,0.1054271609,-0.0699723959,0.0480222851,-0.2027177513,0.1365336329,-0.0632969141,-0.1476316601,0.1257200092,0.1711212993,0.1820428669,0.2451600134,-0.0384036675,-0.0875356197,-0.409095943,0.0009884972,-0.2968892455,0.3175148964,0.1794406474,-0.3005050123,-0.1960091442,0.1506244987,-0.143931374,-0.056152083,0.2782714963,0.147999689,0.1324552298,0.0195310973,-0.1495277286,0.0243149139,-0.2915483415,0.2639170885,-0.0594957657,-0.0554041341,-0.0947801024,0.0293414649,-0.0311482474,0.2046000361,0.1610000879,0.0897464678,0.0021730834,0.2839975655,0.2379150987,-0.4330663979,0.099878937,0.2944245934,0.2381518632,-0.2559423149,-0.4095581472,-0.0246567782,-0.1180522889,-0.005613076,-0.0560318083,0.2782593966,0.2599344552,0.2534674406,-0.0217689052,0.1863395274,0.216273129,0.1439461708,-0.1945036948,-0.1044190302,0.0806882232,-0.115239948,0.2146605104,0.1840274334,-0.5060653687,0.181235984,0.3841530681,-0.1542703658,0.0879698098,0.1259580851,-0.1702739,0.0581509694,-0.0065899971,0.0138176037,-0.2446641028,0.3634675741,-0.0591625422,-0.108531706,-0.1908883899,-0.1340588331,0.1163067743,0.2142671347,-0.1215767488,-0.2503065467,0.0553270392,-0.009102039,-0.1628084183,-0.4696801007,-0.0253826827,0.1480930746,-0.2036803663,0.0377145745,-0.1941829324,-0.1590667218,-0.0824075565,-0.3069811165,-0.3127501011,-0.3622931838,0.3359507918,-0.1132248789,-0.0677312016,0.3541318774,-0.0723467767,0.2815901339,-0.0116785821,0.1275184453,-0.2823343575,-0.2379372269,-0.251925081,0.1519290358,0.0933847427,0.0253330786,0.4100186825,-0.2825551033,0.1890667677,-0.3143168986,-0.5732880831,0.2905284762,-0.25844872,0.0648541003,0.2199686021,-0.1457832754,-0.3888683319,-0.0333772302,0.1715263575,-0.2711287737,-0.3626662493,-0.0250701252,-0.1583240479,-0.1702765077,-0.0864467174,-0.433205843,-0.2901567817,-0.1232596189,0.2089541703,-0.0312268697,0.1811626554,0.2885760367,-0.4318687022,0.1222113669,-0.2701400518,0.098091796,-0.2786156535,-0.4308565855,0.184660241,-0.3317806125,-0.337066412,0.0120338723,-0.1116356775,0.1916644275,-0.1736760885,0.0711489916,-0.1671478599,-0.0187535249,0.0394979678,0.3404955268,0.0439289697,0.1520327926,0.0976218954,-0.1603118926,-0.0321746916,-0.1724944413,0.3113151193,-0.2413232177,0.1498761326,-0.0271666162,0.1294258386,-0.0356341563,0.5465072989,0.0167985409,-0.2188199461,0.1124776304,-0.0698381513,0.2078786492,-0.131040886,-0.1153970361,-0.0175258014,0.1962258369,0.0488493815,0.3266423643,-0.0686601475,-0.0202880241,0.1527094394,0.4638473094,-0.2224692851,-0.2553728819,-0.0573589616,0.1836829185,0.1976723522,0.0822704136,-0.0100333551,-0.1460075676,-0.2055372745,0.0624394342,0.1053210571,-0.0139163062,0.21535106,-0.9272253513,-0.1239781082,-0.3518595695,0.1149428636,0.2784931362,0.2117043287,-0.1005886123,0.0231982302,0.139922902,0.2506639063,0.3925659359,-0.172354266,-0.198786214,0.2679311633,-0.2405377924,0.0573203713,0.0797777325,-0.2286064476,-0.3710949421,0.5744315982,0.1174965128,-0.3089258373,-0.3376754522,-0.035222888,-0.1807746142,-0.2285365015,-0.2050445974,-0.0344224498,0.0398283489,-0.2755060792,0.1085913926,0.363856554,0.1341395676,-0.0492361076,-0.1017664894,-0.3207028508,-0.0139379753,0.3085620701,0.110900715,0.0721494928,0.1642024815,0.0960753337,0.198173821,0.4637289643,0.3716628551,0.1914014667,-0.1280635446,-0.5385272503,0.2372922003,-0.0191451814,0.3033801317,-0.0156916901,0.0658005476,0.1816542149,-0.2474085689,0.1058613658,-0.2882263064,0.4614713192,0.0643679574,0.0808532238,-0.614556849,-0.2713691294,0.1870171726,0.1723172963,-0.2398056239,-0.2446308881,0.5395313501,-0.3418426216,0.3141379654,0.2545683086,0.8012390733,-0.2716864049,0.1736133546,0.0838110149,0.2266589999,0.8018407226,0.0197082944,-0.0548003688,-0.4453334808,-0.0223484393,-0.0401118919,0.012304836,-0.286054194,0.2042441517,-0.0622473769,0.4891516566,0.1152650788,0.1473940015,0.0736764371,0.1538778096,0.3308555484,-0.5550276637,-0.2502360344,0.2629196048,-0.1049464718,0.1011684164,-0.0502717532,0.0625055879,-0.0382312611,0.0031321004,-0.0962834209,-0.0727041811,-0.0741582066,-0.1342280209,0.3775980175,0.0809680745,0.0744003952,0.1943087131,0.7423373461,0.2340450585,-0.1419470608,-0.0250721257,-0.2329995781,-0.0316801332,-0.1227996349,-0.1196628213,0.4821770489,0.0812706128,0.0539786331,0.0829158723,0.0519554839,-0.0574677922,-0.3214282393,-0.2873733044,0.3971169889,-0.2270532697,0.3813668489,-0.2406784594,-0.0081387814,-0.2755420506,0.2029157281,0.0803899094,-0.1075828671,-0.0885190815,0.3111152053,0.0174665432,-0.0844879001,-0.0107158292,0.0398881733,-0.2402342111,0.1945092976,0.0408592224,-0.1827233583,-0.2653930187,0.2161968648,0.5392181873,-0.3252310157,0.1503188908,0.0968730003,-0.2248763293,0.110020861,0.6035030484,0.1310160905,-0.0424070768,0.1361672878,-0.091932185,-0.5880686641,0.1746411771,-0.2361624986,0.3300422132,0.2454454005,0.1923038661,-0.2232650816,-0.0218772627,-0.4449545145,-0.2353122979,-0.5003683567,0.1826831251,0.098872751,0.0071916794,0.145377636,0.1626972407,0.1726816893,0.178195104,0.0183822773,-0.2790432274,-0.3350640833,0.0463402979,0.2418377399,0.2792968452,0.0048966268,0.2466958463,0.0067548114,-0.2719830275,0.1146459356,0.1886238158,0.1420633793,0.2818845212,0.1411853731,0.1273689717,0.0757931173,0.1339850873,0.299782455,0.0177531485,0.1172628403,0.0153931361,-0.2881683707,-0.0887881294,-0.0191126242,0.386275202,0.2473716736,-0.1032756791,0.0380102433,-0.2212549597,0.0356871784,0.3983194232,-0.0356273353,-0.1485816836,0.0230256245,-0.2126665413,0.4153338373,0.3393226862,-0.1160741001,-0.110945724,0.0520004034,0.1824388951,0.1357592195,0.347199738,0.1274812073,0.1097617671,0.114107877,-0.0938314125,0.1275753677,-0.2317717671,-0.0522967726,-0.1410776973,0.0761955157,0.1322786659,0.5697593689,0.1241173223,0.1137580723,0.3200816512,0.0547745042,0.4050596654,0.1499900073,0.4258761108,0.3700139821,-0.0102956118,-0.138897866,0.2034137994,0.1344485432,0.1304856241,-0.2309441715,0.1932895333,0.0831905529,-0.3695017099,-0.1058568135,0.3525800109,-0.2580944896,-0.3478722572,-0.4649516642,0.058193922,0.0739827603,-0.2685129941,-0.1080918536,-0.2631648481,-0.1158763915,-0.1411875784,0.0868222639,-0.4861607254,-0.3856500089,0.2645304501,0.1960576624,-0.0958373398,0.1850445122,-0.2260196209,-0.1638351232,0.3651375175,0.2383434772,0.2375177294,0.0303071141,0.3444776535,-0.1424370408,-0.2216662019,-0.0912244245,0.0039834892,0.102182813,0.0953884125,-0.1628293395,0.231757015,0.1330412477,-0.1487390995,-0.1543930173,-0.0843736306,0.1567488313,0.2958145738,0.5019689798,0.1202239618,-0.0479635037,-0.2133332342,0.176492393,-0.4314566255,0.1594846547,0.0941373035,0.1836383492,0.1259750128,-0.1876222193,0.1515058726,0.0010168105,0.2498112768,0.3580768704,-0.1188349873,-0.1300881207,-0.041894041,-0.3163099885,0.1451972276,0.1794715375,0.0322514437,0.059865091,-0.095348753,0.0339572504,0.056916479,-0.2780731618,0.0186931565,-0.0529055521,0.0569178611,-0.2804805636,0.1978539377,0.1653849781,-0.0442250445,0.1083645299,0.0616231225,0.0139041385,-0.1952131987,0.006646066,-0.0783292726,0.4227509499,-0.1425765008,-0.0378605537,0.009359709,0.2284357846,0.3607788682,0.0830439553,0.0435539931,-0.0505282842,-0.4318092763,-0.3073491454,0.0982900038,-0.0965508968,0.36786744,0.1136207655,-0.1626976281,-0.1074949503,-0.0215720199,0.2414415032,-0.1362692416,-0.4723663032,0.433100611,-0.167316258,0.0928559676,0.1339776665,0.1941726208,0.3070421815,0.1179795787,-0.0809316784,-0.6299244761,0.532125473,-0.3958278596,-0.0686670244,-0.0602631085,0.0539494865,0.4667930305,-0.0481405593,-0.4107825756,0.0318905488,0.342508018,-0.0756124854,-0.4658171833,0.0576854385,-0.0169345848,0.1860681623,-0.246646598,0.2820748687,0.1233507246,-0.0033511603,0.1426492631,-0.1557134688]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/797","title":"Token classification labels are strings and we don't have the list of labels","comments":"Should definitely be a ClassLabel \ud83d\udc4d ","body":"Not sure if this is an issue we want to fix or not, putting it here so it's not forgotten. Right now, in token classification datasets, the labels for NER, POS and the likes are typed as `Sequence` of `strings`, which is wrong in my opinion. These should be `Sequence` of `ClassLabel` or some types that gives easy access to the underlying labels.\r\n\r\nThe main problem for preprocessing those datasets is that the list of possible labels is not stored inside the `Dataset` object which makes converting the labels to IDs quite difficult (you either have to know the list of labels in advance or run a full pass through the dataset to get the list of labels, the `unique` method being useless with the type `Sequence[str]`).","comment_length":6,"text":"Token classification labels are strings and we don't have the list of labels\nNot sure if this is an issue we want to fix or not, putting it here so it's not forgotten. Right now, in token classification datasets, the labels for NER, POS and the likes are typed as `Sequence` of `strings`, which is wrong in my opinion. These should be `Sequence` of `ClassLabel` or some types that gives easy access to the underlying labels.\r\n\r\nThe main problem for preprocessing those datasets is that the list of possible labels is not stored inside the `Dataset` object which makes converting the labels to IDs quite difficult (you either have to know the list of labels in advance or run a full pass through the dataset to get the list of labels, the `unique` method being useless with the type `Sequence[str]`).\nShould definitely be a ClassLabel \ud83d\udc4d ","embeddings":[0.0768210217,0.0408481397,-0.0622679777,-0.0633265823,0.2008129209,0.031326782,0.3044342697,0.0615330637,-0.2412469834,0.2134516984,0.0375984684,0.2922468483,-0.0581920072,0.2134546787,0.0440989397,-0.0849336833,0.1225070134,-0.0203047227,0.4037330747,-0.1659709364,-0.4042012095,0.0990402624,-0.1822333336,0.418227762,-0.4765669405,-0.2484193444,-0.1079657823,-0.2100836039,-0.2112367004,-0.7866371274,0.01376897,0.4073163867,0.2567683458,0.2016524225,-0.000099815,-0.2282706499,0.2769232392,0.0468798727,-0.1004512459,-0.1356547922,-0.132893905,-0.0208617765,0.1267424226,-0.3612166643,-0.1561586708,-0.130320549,0.1956954598,0.0553812869,0.0488849618,0.2450037897,0.307174921,-0.1079552621,0.0439232625,-0.1192446128,0.3123594224,0.08579126,-0.1171496958,-0.2361769378,0.3827905655,0.0606327653,-0.1801498234,0.2091386467,0.048502259,-0.2505214214,0.2122522891,0.1730182022,0.0583947487,-0.4053652883,-0.0621318184,0.2142696679,0.3668856621,-0.1557866335,-0.345233053,-0.0470486246,0.1908561289,-0.5691116452,0.0959332883,0.0806351751,-0.080550015,0.1211731657,-0.032531891,-0.0456739031,-0.1141422614,0.0379827097,0.1086833403,0.4687159956,0.00419875,0.0096691726,-0.2317759097,-0.2129191458,0.2041487396,0.1588688195,-0.076425761,0.0995420441,-0.2871589661,-0.06811288,-0.291015625,0.1549748778,0.0579142161,0.1181846261,-0.1640138626,0.0706309527,0.0777925327,0.102330029,-0.0987206548,0.3891686797,0.63230896,0.4124375284,0.0062644761,-0.2831750512,-0.201220721,0.0388593152,-0.3928970695,0.2875709832,0.2268941998,0.0354381315,-0.2978462875,-0.1191999242,-0.2023610622,0.2386752814,-0.1887488663,0.151620701,0.2438747883,0.2963256836,-0.0335858688,-0.0048513906,0.0642039403,-0.1481015831,0.117570892,-0.4146531224,-0.1679750681,0.0013823258,-0.3165435195,0.1231608018,0.0364442319,0.1781609654,0.1901351362,-0.0979544818,0.0218374971,-0.0045778644,0.0278856605,0.0312349591,0.1512978822,0.101462692,-0.1582397819,0.2000208497,0.0218751971,-0.3404643536,-0.1182463393,-0.0583085977,-0.4470881522,-0.0702943206,0.0174259283,0.3134384453,0.1736944169,-0.298832953,-0.2355999053,0.2007579058,-0.1186724156,-0.000668532,0.0200697258,-0.2740599811,-0.0681719854,-0.1230470538,0.0008262526,-0.0003314416,-0.2502900958,-0.2497319132,0.0574743785,0.0549387597,0.4358788133,0.2037663311,0.0340983942,0.1649187803,-0.0843898281,0.3176147044,0.2941150665,-0.3207540214,-0.0480906665,-0.1095457748,0.0903788656,-0.4227242172,0.0299677327,0.1495949328,0.4886767864,-0.185723424,0.2216118127,0.0355025642,-0.1019278541,-0.0455460511,-0.1317717582,0.014778167,-0.0321079493,0.1599477828,0.0010358335,0.0144398678,-0.2233490348,0.1796093583,-0.0895706192,-0.0955076963,0.1153866053,0.1230525449,0.2066576183,0.1273615658,-0.1059815884,-0.0217767134,-0.3198322058,-0.0540475249,-0.4033003449,0.3149658144,0.2263285518,-0.2728545368,-0.1995934099,0.1020718217,-0.1571552008,-0.0320423245,0.3134826422,0.0858375654,0.005109462,0.0308738984,-0.0993152633,0.0031067741,-0.29141891,0.1979734004,0.1333473474,-0.1013983563,-0.1244109794,0.0086920271,-0.0881629884,0.1404920816,0.2090094686,0.1559676379,-0.0424544029,0.3259553015,0.3027758896,-0.5033314824,0.0531026907,0.2369454652,0.1832408756,-0.2334075123,-0.4235955477,-0.0028389262,-0.157462284,-0.0109184319,0.0127114449,0.2579411566,0.2276392728,0.2110081464,-0.1010567918,0.1889338642,0.140272066,0.131369859,-0.1896604896,-0.0998370424,0.058209829,-0.1037061438,0.1796892732,0.2372550517,-0.4404655695,0.170366466,0.4048959017,-0.108738713,0.1101436839,0.1259507686,-0.1069037393,0.0537482612,0.0068631065,0.1044772044,-0.2589973211,0.315366298,-0.1305079311,-0.1475556046,-0.1903862506,-0.106580846,0.1317346841,0.2322927266,-0.0266170576,-0.2668626904,0.0060945642,-0.0316680856,-0.0331959203,-0.532052815,0.0106634889,0.1944475025,-0.1925121695,0.0129320081,-0.1499054879,-0.1529020518,-0.0605578423,-0.3183797002,-0.2868320346,-0.4123721421,0.4081130326,-0.1618976146,-0.0490075387,0.3676025867,-0.098071076,0.2918481529,-0.0241700374,0.1050853282,-0.1929809898,-0.2683356702,-0.3229005039,0.1699775606,0.0977562889,0.0489696786,0.3840393722,-0.2414394915,0.2228828222,-0.2187254876,-0.5636425018,0.2270750254,-0.1606742442,0.0286350194,0.1670449674,-0.2327639461,-0.4268393517,-0.1218441725,0.0852383226,-0.225970313,-0.2283295542,0.0026585239,-0.2077169716,-0.1686622053,-0.0664309189,-0.4556927085,-0.3312296271,-0.1236740649,0.0408038571,-0.0233279187,0.1746591032,0.1868965775,-0.431640476,0.0529649593,-0.1822777241,0.1012332886,-0.2143922001,-0.4131071866,0.1690001488,-0.2728218734,-0.3291990161,-0.046279069,-0.1354906857,0.260530293,-0.142766878,0.1251312941,-0.0946493,0.0706814453,-0.0226576962,0.3241481483,-0.0844372511,0.0872837678,0.2019927502,-0.211509645,0.0054970416,-0.0688067302,0.3137127757,-0.2144615501,0.2282593995,-0.0420599021,0.0848075449,-0.0460321344,0.4352812767,0.0055042896,-0.2631635964,0.0455760993,0.0022457109,0.0918791145,-0.1107524112,-0.1034810692,-0.0519464351,0.2237147242,-0.0670660734,0.3357895613,-0.1570311934,-0.1210895404,0.1330244094,0.5455912948,-0.2789964676,-0.2406085134,-0.0106635382,0.054960195,0.2454065531,0.0478863195,-0.0303443335,-0.1355297118,-0.2610328496,0.0024895009,0.0764808133,-0.0506648533,0.1783785522,-0.8838063478,-0.0876517296,-0.2758919597,0.0650000498,0.2896383405,0.3287802637,-0.1050058156,0.0875159651,0.1378634274,0.243316263,0.2921641469,-0.1724601239,-0.2023004591,0.306627214,-0.0915133432,0.1048011258,0.0684384108,-0.2548244596,-0.3751781583,0.6587129235,-0.0013229237,-0.1980962604,-0.2892563641,-0.1003281474,-0.2542065978,-0.2552270889,-0.2299426198,-0.0103047295,0.1338199824,-0.2432802469,0.0954227,0.3469467759,0.0970515981,-0.0816390365,-0.1210769489,-0.2930516005,-0.0052206474,0.3091537058,0.0813534036,0.0382580161,0.1295605004,0.1090776175,0.0679771751,0.471378684,0.468429625,0.0446008667,-0.1178455353,-0.4347439408,0.2569064498,-0.1234912947,0.3224527836,-0.0421053804,0.0402706414,0.1581085026,-0.3506857157,0.0792169049,-0.2233100533,0.4037976861,0.0329969041,0.0056237117,-0.6572426558,-0.2905606627,0.1734958738,0.1785162836,-0.2949580252,-0.3039404154,0.4339734316,-0.3457606733,0.4166151285,0.3446819484,0.8402222395,-0.2765398622,0.153770715,-0.0398757793,0.2484613806,0.8161224127,-0.033860296,-0.04928305,-0.4499648511,0.0454568304,-0.0397962183,-0.0126445284,-0.272613287,0.3278288245,-0.1575098485,0.4755187333,0.0842606723,0.0694004893,0.116682291,0.1682195663,0.4235061109,-0.4813767076,-0.0906634778,0.2716469467,-0.1478792876,-0.008489741,-0.1036666706,0.0865017027,-0.0916432291,0.0664000064,-0.0650573969,-0.0761936978,-0.0084571233,-0.085259378,0.2099221945,0.1007264778,0.1054402441,0.1962734014,0.6826722622,0.2396541834,-0.102055192,0.0042603058,-0.143812716,0.0096500749,-0.0612915382,-0.0923682079,0.4756271541,0.0628627986,0.0184507929,0.0871885493,0.0366614833,-0.0873480737,-0.3624782264,-0.2737263441,0.5779613256,-0.1959302723,0.3835754991,-0.2282650471,-0.0542849116,-0.2488140762,0.2299529314,0.0389279798,-0.0340617299,-0.0036484501,0.305210948,0.0065665091,-0.0574012063,-0.0475485548,0.1108567268,-0.2413233817,0.1471494883,-0.0268196091,-0.162515372,-0.2937345207,0.3379104435,0.5152076483,-0.2600741088,0.073965773,0.0553442389,-0.2391115427,0.1730944216,0.5471160412,0.1998198628,0.033034578,0.119345665,-0.0346743427,-0.5989367366,0.1492888182,-0.2309644371,0.2959704697,0.1551129967,0.1831935942,-0.1980195493,0.0012208305,-0.4458531737,-0.265277952,-0.506985724,0.1666662544,-0.0069521014,0.0518447757,0.0069249896,0.2098841965,0.1706136465,0.1566649079,0.0548336059,-0.2948642969,-0.2774916291,0.047804907,0.2919935882,0.2778143585,-0.0212246273,0.2169543356,-0.0021873023,-0.1961897016,-0.0159720592,0.1201599613,0.130068332,0.3469142616,0.1631504148,0.1440748572,0.0214746501,0.1141786799,0.2603269815,0.036883954,0.0771664456,0.0935700983,-0.1737708151,-0.1207502708,-0.0746066943,0.3826414347,0.1736990213,-0.036709249,0.0033082264,-0.2357729077,0.1335970759,0.3667702377,-0.139902696,-0.1279017329,0.0227096584,-0.3365601003,0.4009516537,0.3626036644,-0.0783069134,-0.1216499209,-0.0831931904,0.1834907532,0.1262177825,0.3075411916,0.1038674489,0.0938933417,0.0761686862,-0.0806920007,0.1641827524,-0.2672724724,-0.0073191393,-0.1113125905,0.0753161684,0.1344532222,0.544870317,0.0293023679,0.0395039581,0.4018202126,0.1554852724,0.4802336693,0.2808579206,0.4061281383,0.3893161118,0.0870949849,-0.1251718849,0.2771378458,0.1205809563,0.110397011,-0.1324063987,0.0871644467,0.1545950174,-0.4034617543,-0.0741439238,0.2192277759,-0.2240506709,-0.2999415994,-0.5332089663,0.0337834321,0.0711610466,-0.2867443562,-0.1224692017,-0.1979743689,-0.1575935483,-0.0478855036,0.0660463348,-0.4325397313,-0.3930718303,0.2419251055,0.1953437179,-0.1422200948,0.1416897327,-0.2518833876,-0.1245210469,0.4135083258,0.1293587536,0.1772696525,0.1297988296,0.2966983914,-0.2020821273,-0.2102628648,-0.0950009152,0.0324160978,0.0505374111,0.153938368,-0.1952284575,0.2203783989,0.1667108983,-0.1458132863,-0.1190961376,-0.0071834344,0.1094721332,0.381819874,0.5128670335,0.2157718241,-0.0837436914,-0.1622065455,0.1934518218,-0.4342210293,0.1483648866,0.1103886813,0.1622074097,0.1297765672,-0.1996585578,0.1550732106,-0.0077289618,0.2510840297,0.2670448124,-0.1258800477,-0.16584225,-0.1097758934,-0.2516238093,0.241121918,0.1479324549,0.0337538123,0.1138040423,-0.1368687749,-0.0222299825,0.0437764078,-0.2637335956,0.0545471124,-0.1886152476,0.1083354652,-0.3189408183,0.2377624214,0.2266927958,0.0191873983,0.0753318146,0.0241258573,-0.0150903752,-0.1803647578,0.0547152385,-0.0724307746,0.3150385618,-0.0269466266,0.0773629919,0.0789454058,0.2060728371,0.3618319333,0.0883559585,0.0555193312,-0.003059987,-0.4147556126,-0.2860845923,0.0799134001,-0.0380091146,0.3430612683,0.1546089649,-0.1247401908,-0.0981063843,-0.0826285481,0.1476854384,-0.0809504613,-0.5174705982,0.4141790867,-0.0643085018,0.1444258243,0.0157763269,0.1383515149,0.3136616349,0.1641432494,-0.0957003459,-0.5930022597,0.4904518127,-0.3036496639,-0.0730917156,-0.0427083559,0.0568354912,0.3752419353,0.0315561667,-0.3241288066,-0.0314144567,0.3910897374,-0.0594875589,-0.5095586777,0.0164803267,-0.0071561174,0.1668933183,-0.239215225,0.2097654939,0.0473340414,0.0278132726,0.0539805554,-0.1533282548]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/796","title":"Seq2Seq Metrics QOL: Bleu, Rouge","comments":"Hi ! Thanks for letting us know your experience :) \r\nWe should at least improve the error messages indeed","body":"Putting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand:\r\n\r\n+ Bleu expects tokenization, can I just kwarg it like sacrebleu?\r\n+ different signatures, means that I would have had to add a lot of conditionals + pre and post processing: if I were going to replace the `calculate_rouge` and `calculate_bleu` functions here: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py#L61\r\n\r\n\r\n#### What I tried\r\n\r\n\r\nRouge experience:\r\n```python\r\n\r\nrouge = load_metric('rouge')\r\nrouge.add_batch(['hi im sam'], ['im daniel']) # fails\r\nrouge.add_batch(predictions=['hi im sam'], references=['im daniel']) # works\r\nrouge.compute() # huge messy output, but reasonable. Not worth integrating b\/c don't want to rewrite all the postprocessing.\r\n```\r\n\r\nBLEU experience:\r\n```python\r\nbleu = load_metric('bleu')\r\nbleu.add_batch(predictions=['hi im sam'], references=['im daniel'])\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n```\r\nAll of these raise `ValueError: Got a string but expected a list instead: 'im daniel'`\r\n\r\n#### Doc Typo\r\nThis says `dataset=load_metric(...)` which seems wrong, will cause `NameError`\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/98004483-ff0d0580-1dbd-11eb-9f35-6f35904611bb.png)\r\n\r\ncc @lhoestq, feel free to ignore.","comment_length":19,"text":"Seq2Seq Metrics QOL: Bleu, Rouge\nPutting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand:\r\n\r\n+ Bleu expects tokenization, can I just kwarg it like sacrebleu?\r\n+ different signatures, means that I would have had to add a lot of conditionals + pre and post processing: if I were going to replace the `calculate_rouge` and `calculate_bleu` functions here: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py#L61\r\n\r\n\r\n#### What I tried\r\n\r\n\r\nRouge experience:\r\n```python\r\n\r\nrouge = load_metric('rouge')\r\nrouge.add_batch(['hi im sam'], ['im daniel']) # fails\r\nrouge.add_batch(predictions=['hi im sam'], references=['im daniel']) # works\r\nrouge.compute() # huge messy output, but reasonable. Not worth integrating b\/c don't want to rewrite all the postprocessing.\r\n```\r\n\r\nBLEU experience:\r\n```python\r\nbleu = load_metric('bleu')\r\nbleu.add_batch(predictions=['hi im sam'], references=['im daniel'])\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n```\r\nAll of these raise `ValueError: Got a string but expected a list instead: 'im daniel'`\r\n\r\n#### Doc Typo\r\nThis says `dataset=load_metric(...)` which seems wrong, will cause `NameError`\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/98004483-ff0d0580-1dbd-11eb-9f35-6f35904611bb.png)\r\n\r\ncc @lhoestq, feel free to ignore.\nHi ! Thanks for letting us know your experience :) \r\nWe should at least improve the error messages indeed","embeddings":[0.1769534498,-0.0281658582,-0.0853118747,0.0595532954,0.1316637546,-0.2772905231,0.0935510099,0.1194606274,-0.0715432093,0.0205882788,-0.2373280972,0.0115100741,-0.1035928577,-0.0028772415,0.0880227238,-0.231206879,-0.0192949343,0.0564785786,0.022084957,-0.209247455,-0.2625801265,0.3527042568,-0.0217220634,0.2763134837,-0.2073199898,0.2331521213,-0.0639110953,0.2056541741,-0.2941870987,-0.1854021698,0.0096485056,0.3292575777,0.0186899304,0.2050482631,-0.0001137066,-0.3125211596,0.0184408836,-0.1469954848,-0.1234472394,-0.1045938656,-0.1109406278,-0.195944801,0.0969132185,-0.1325443089,-0.0355381072,0.1201137006,-0.318225652,-0.0693274662,0.5164248943,-0.0813493282,0.168647632,-0.0962301493,-0.1875781119,-0.1486988664,0.3641577363,-0.1070586815,0.0575782731,0.004555345,0.2955908477,-0.0346819237,-0.1117430776,0.2009179592,0.0118584456,-0.214055717,0.1996009201,0.0452660359,0.2193309963,-0.1476968676,-0.1681335866,0.2540251315,0.1309775263,-0.0490120389,-0.4720951617,0.0694950223,-0.0954873562,-0.7001909018,-0.0471167974,-0.2837664187,0.1515046805,-0.1515443474,-0.1255643219,0.0069697043,0.1390461624,-0.2734936476,-0.136483863,0.0243614838,-0.1411269158,-0.0774002075,0.1040128618,-0.1568869799,-0.6167755127,0.2696429491,-0.1904052645,0.0336680152,-0.5840451121,0.0813066885,0.0431840122,0.1194143519,-0.1405089945,0.4723142385,0.1637546718,0.0697114617,0.1093147695,0.0862753019,-0.3274796307,0.8337019682,0.087496534,0.0176879056,0.1399512142,-0.0052242139,0.1916358024,-0.0466981567,0.0865587071,-0.5624172091,-0.0675756931,0.3009655178,-0.2086916715,-0.4418406785,-0.4798848629,0.1379334778,-0.2506815195,-0.1557816118,0.3023255765,0.3863398731,-0.335867852,0.0597458482,0.40995875,0.3697021306,-0.2531629503,-0.1009762734,-0.1679190695,0.0602376014,-0.2565958202,0.225626573,0.0209779311,0.0801946297,0.3454347551,0.0971023589,0.3335015774,0.004242064,0.1434872895,0.0575459935,-0.2455943674,-0.1922012866,-0.1943170577,0.0088598132,0.0848355591,-0.2524261475,-0.0161993299,-0.382820487,-0.3244469762,-0.1584510505,0.0603077821,0.1470858306,-0.1753459871,0.1231860295,-0.0035436349,0.3194283247,-0.0393082947,-0.1132980362,0.005361788,0.1049269363,-0.2670657039,-0.0042275405,0.493555367,-0.0255884975,-0.3216415644,-0.3817739487,0.4146699309,0.1683200598,-0.0606814697,0.270088762,0.0352221616,0.2871845067,0.0276340041,-0.1804422289,0.3904258907,-0.5985485911,-0.0689604282,-0.0158594176,-0.0115918266,-0.2362428457,0.2013259977,-0.1834621727,0.3161441982,-0.1831455827,0.4822333455,0.1010600477,0.0653936863,-0.2801431715,-0.2663798928,-0.1388558894,0.1392518878,-0.0844184905,0.1085980386,0.1279085279,-0.2155474573,0.3492257893,0.0701876506,-0.2282511145,0.0856477097,0.0376856215,0.1824711561,0.0331669003,0.2077793777,0.3472885191,0.1417553872,0.1336536705,-0.697609067,0.1592135727,0.0037085325,0.0860563889,-0.3027497232,-0.2085744739,0.0227134824,-0.4090459943,0.1354869604,-0.2492710054,0.2036932856,0.1237623915,0.001924585,0.1244986206,0.2435036302,0.182470277,-0.2680424452,-0.1919852942,0.202580899,0.0015646351,0.0111755775,0.4812694788,0.2334780693,0.1712279469,0.04375894,0.2832264006,0.0035683014,0.1227550358,0.1778895706,0.2196926922,0.2059095204,0.0626964644,0.1532783359,-0.0062190397,0.1477894634,-0.2205672115,-0.0119060511,0.455041945,-0.4066660404,0.1981160939,-0.2130469829,-0.0134293316,-0.0060455082,0.027414592,-0.3535484672,0.0184353963,0.325185895,0.1578260958,0.4982276857,-0.2068050951,0.2832622528,-0.1557312608,0.0057616415,-0.1006078348,0.0947469696,-0.040629521,0.1436415911,-0.0996471047,-0.0770215541,-0.4216920137,0.0819059685,0.1051404104,-0.1054541767,0.0385688841,-0.0031587586,-0.09889777,-0.1584773511,-0.0599150099,-0.2916539907,0.037877731,0.2082577944,0.1982465535,0.2047153264,-0.3442966044,-0.2024778128,-0.0576463677,-0.3000858426,0.1449369639,-0.0344874784,0.539752841,0.2608511448,0.1027044579,-0.2336456478,-0.3465655744,0.3593018949,-0.039825689,0.0450821519,0.1441108286,0.0771829188,0.4756244421,0.038900096,-0.0621122345,-0.0366739109,-0.1377660632,0.0031252198,0.0776271671,-0.3016474545,0.1216313019,0.1167247444,-0.3648528755,0.2510655224,0.1978717893,-0.5468524098,0.124114424,-0.1195817515,0.5138995051,0.3033767939,-0.301456511,0.023169307,-0.0888207555,0.2371552736,-0.0323478281,-0.2406721413,0.3399661481,-0.0452200957,0.1214821339,-0.0695555136,-0.1393463463,0.0858451575,-0.2094722241,0.3407860696,-0.0766743347,-0.2421108335,-0.1812707037,-0.0949681625,0.2206552625,0.1995648444,0.4270272851,-0.1554479897,-0.2086057812,0.3116022646,-0.1577573866,-0.2221233696,-0.2294902205,-0.1876649112,0.491013974,-0.1489214301,-0.0511640161,-0.3940578103,0.1894480735,-0.1001680195,0.2220212966,-0.0307293721,0.1455084085,0.1064650342,-0.1404236406,-0.0989815742,-0.1612590551,0.191022709,0.1008730531,0.4178522527,-0.2088980824,0.339697659,-0.0117482059,0.6103149056,0.1115433425,0.2892510593,0.1768062711,0.3831200898,0.1211188436,-0.0213042125,-0.2075167298,0.0879762396,-0.0352794267,-0.2933488488,0.1319354326,0.0984315276,0.4414548874,-0.2185959965,0.3410234451,-0.210394904,-0.1881791055,0.1058161855,-0.2011750489,-0.098917596,-0.1430128664,0.0723319352,0.1112953871,-0.0060305069,0.241292432,0.1390670687,0.3017922044,0.0698097423,-0.2126363516,0.1514588743,-0.3834177852,0.4566945732,-0.2437226623,-0.2038207948,-0.2165651768,-0.1013507843,-0.1215602979,-0.0875823721,0.214755252,0.0374684706,-0.2363127768,0.1458826363,-0.5916590691,-0.4131349325,-0.0120864147,-0.2095865607,-0.1605082005,0.2359598577,0.0095172767,-0.4821446836,-0.0756690502,-0.2295518965,0.057655625,-0.3014695048,-0.0373364463,-0.2727277875,-0.1172621697,-0.0696871579,0.0924727693,-0.2291588783,0.2875803113,-0.2921278179,0.1601521075,-0.1564980149,-0.0742772743,0.3070153594,0.1465413868,0.1504179686,-0.3316683471,0.078501232,0.2081436068,-0.1293841451,-0.1273933649,0.1969533414,-0.2532327771,-0.2951292992,0.3716384172,0.0178300925,0.2221780568,0.2946822643,-0.2342992872,0.0944958106,-0.3147600591,0.3028075993,-0.3613408506,-0.0944890603,0.3689983189,0.1033305824,0.0125515759,-0.2299200296,0.0690530762,0.1653673947,-0.3130488694,-0.3780952692,-0.0152913975,-0.278544426,0.5295838714,0.2223323137,0.9652849436,0.0470921658,0.0763823763,0.1551992446,0.0962854624,0.4637106061,-0.5718903542,-0.0351011157,-0.3878887594,-0.0913151056,0.0291725341,0.0367921069,-0.1170589402,-0.0540103428,-0.393668443,0.4557964206,-0.2637159228,-0.2005107701,-0.1655346602,-0.0229667872,0.1302414685,-0.2507967055,-0.3668765426,0.1868800819,0.0889842808,0.0542574562,-0.2669872046,-0.0393872634,-0.2502802014,-0.0585819371,0.0858374536,0.0232327431,-0.0212349873,0.1083106548,-0.0477350764,-0.15406394,-0.1582447439,0.544236064,0.5773678422,0.217205748,-0.0208211653,0.0600922368,-0.1236229986,0.5404776335,0.0934721455,-0.2224397063,0.2294900715,-0.0243547689,-0.020823203,0.3062645495,0.0182671379,-0.0775452256,-0.1989146322,-0.1544184834,-0.325826019,-0.3220331967,0.4116707444,0.2232385427,0.2346820235,-0.0311754644,0.0928960443,0.452919215,-0.3160113394,0.1871128827,0.1644583941,-0.32837376,0.0176813807,0.15692918,-0.2035150975,-0.2034178525,0.1447074562,-0.1222935095,0.0662494823,-0.1050124392,-0.2246523052,0.3282240927,-0.6158259511,0.1495267302,-0.0137531999,-0.2014415711,0.1079682112,0.5085456371,0.3355732262,-0.1233519018,0.3740538061,-0.1671467423,-0.2427298427,0.4513656497,0.0616037957,0.1218226328,-0.2861672938,-0.3142947853,-0.3326703906,0.6031278968,-0.3359348774,-0.2359313071,-0.4539035559,0.1343765706,-0.362043649,-0.2385994941,0.2074179947,0.2969813049,0.009648269,0.1842312962,-0.3344164193,-0.1230726838,-0.4081398249,0.1003050879,-0.1193289906,0.2269994169,-0.0668595135,0.4527094662,0.2298652232,-0.2671133578,0.1697508395,0.1802743822,0.2975305319,0.2053432614,0.0191491116,0.2244738787,-0.0069990172,-0.0262908693,0.2954408526,0.1316293776,0.1685137898,0.0559650958,-0.2218063325,-0.1717917323,0.3374065459,0.1799139082,0.2362015098,-0.0936959535,0.1139821261,-0.0410220772,0.0842017978,-0.2749198079,0.3915095925,0.3128079474,-0.0195628181,-0.0796531066,-0.1372188032,0.2240026593,0.0244851746,0.3415912092,-0.149001345,0.2489686459,-0.0112866266,0.3523798287,0.1521824896,0.0147120096,0.4298763573,-0.0189230144,-0.0702096298,-0.2236607671,0.1153164878,-0.2348205894,-0.1815804243,0.3596546948,0.4222204387,0.0077236784,0.0152922552,-0.0210295375,-0.1447943747,0.2545048892,-0.3928196132,0.6502040625,0.2231901735,0.0503745377,-0.1527456492,0.2164507508,0.0517553799,-0.2472672611,-0.2434281409,0.4218269885,0.0581700839,-0.3403600454,-0.3058310449,0.3959472775,-0.0788398981,-0.4969257414,-0.0322642475,-0.3686673641,-0.2251703739,-0.1759877354,0.0558964014,0.0035126531,-0.1908208132,0.0639134273,0.1989696622,-0.276350528,-0.5920048356,0.1823782623,0.0905692875,-0.0429586619,-0.0077434652,0.3834717572,0.0954965428,0.5452119708,0.4397696853,0.2471140325,0.1698268354,-0.4162213206,0.0099131372,-0.0941653773,0.1179095358,-0.117886357,0.4269874692,0.2142495662,-0.3480429947,0.1749587208,0.1014097407,-0.1391798109,0.2263496071,-0.2509038746,0.2887138426,-0.1855385751,0.4289951324,0.1102497876,0.3845459521,-0.2034668922,-0.054612238,-0.4368281364,0.0242949538,-0.2412265539,-0.0238137748,0.1368947774,0.1685836762,0.0787073448,0.1651517153,0.0353789181,0.0816258863,0.1796215922,-0.2963189483,0.2632241845,-0.1453604996,-0.0498373993,0.1904272586,0.3427992165,-0.3479309976,0.2818306983,0.1530394703,0.2411677092,-0.1471676826,0.086702168,-0.1079351977,0.1036498621,-0.4891263545,0.2790877223,-0.0577862002,0.0819780901,0.0164200626,0.1622750759,-0.2522097826,-0.030885892,0.0889545083,-0.0395877846,-0.219846189,0.120538421,0.4369190633,0.38095963,0.4347174168,0.2746351361,0.0491110533,-0.0917204767,-0.0348362811,-0.2374585867,0.0091844862,-0.1451218724,-0.0991911963,0.1483568549,0.1272278279,-0.0981932357,-0.2859273255,0.7760738134,0.10803321,-0.5372213125,-0.452224493,0.1022330523,0.2105946392,-0.0174015574,-0.2600540221,0.3088754416,-0.0147051318,0.6009328365,-0.1941381544,-0.1132751405,0.4475610256,-0.4307902455,0.3061504066,0.1107731387,0.3764396012,0.4347320795,-0.0453522541,-0.1358400434,0.0445140786,0.0719929487,0.0879144147,-0.006922279,0.1729042679,0.2557842135,-0.084685564,-0.2360916138,-0.2595790923,0.0349611454,0.0786984488,0.0781525522,-0.2363346517]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/796","title":"Seq2Seq Metrics QOL: Bleu, Rouge","comments":"So what is the right way to add a batch to compute BLEU?","body":"Putting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand:\r\n\r\n+ Bleu expects tokenization, can I just kwarg it like sacrebleu?\r\n+ different signatures, means that I would have had to add a lot of conditionals + pre and post processing: if I were going to replace the `calculate_rouge` and `calculate_bleu` functions here: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py#L61\r\n\r\n\r\n#### What I tried\r\n\r\n\r\nRouge experience:\r\n```python\r\n\r\nrouge = load_metric('rouge')\r\nrouge.add_batch(['hi im sam'], ['im daniel']) # fails\r\nrouge.add_batch(predictions=['hi im sam'], references=['im daniel']) # works\r\nrouge.compute() # huge messy output, but reasonable. Not worth integrating b\/c don't want to rewrite all the postprocessing.\r\n```\r\n\r\nBLEU experience:\r\n```python\r\nbleu = load_metric('bleu')\r\nbleu.add_batch(predictions=['hi im sam'], references=['im daniel'])\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n```\r\nAll of these raise `ValueError: Got a string but expected a list instead: 'im daniel'`\r\n\r\n#### Doc Typo\r\nThis says `dataset=load_metric(...)` which seems wrong, will cause `NameError`\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/98004483-ff0d0580-1dbd-11eb-9f35-6f35904611bb.png)\r\n\r\ncc @lhoestq, feel free to ignore.","comment_length":13,"text":"Seq2Seq Metrics QOL: Bleu, Rouge\nPutting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand:\r\n\r\n+ Bleu expects tokenization, can I just kwarg it like sacrebleu?\r\n+ different signatures, means that I would have had to add a lot of conditionals + pre and post processing: if I were going to replace the `calculate_rouge` and `calculate_bleu` functions here: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py#L61\r\n\r\n\r\n#### What I tried\r\n\r\n\r\nRouge experience:\r\n```python\r\n\r\nrouge = load_metric('rouge')\r\nrouge.add_batch(['hi im sam'], ['im daniel']) # fails\r\nrouge.add_batch(predictions=['hi im sam'], references=['im daniel']) # works\r\nrouge.compute() # huge messy output, but reasonable. Not worth integrating b\/c don't want to rewrite all the postprocessing.\r\n```\r\n\r\nBLEU experience:\r\n```python\r\nbleu = load_metric('bleu')\r\nbleu.add_batch(predictions=['hi im sam'], references=['im daniel'])\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n```\r\nAll of these raise `ValueError: Got a string but expected a list instead: 'im daniel'`\r\n\r\n#### Doc Typo\r\nThis says `dataset=load_metric(...)` which seems wrong, will cause `NameError`\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/98004483-ff0d0580-1dbd-11eb-9f35-6f35904611bb.png)\r\n\r\ncc @lhoestq, feel free to ignore.\nSo what is the right way to add a batch to compute BLEU?","embeddings":[0.1769534498,-0.0281658582,-0.0853118747,0.0595532954,0.1316637546,-0.2772905231,0.0935510099,0.1194606274,-0.0715432093,0.0205882788,-0.2373280972,0.0115100741,-0.1035928577,-0.0028772415,0.0880227238,-0.231206879,-0.0192949343,0.0564785786,0.022084957,-0.209247455,-0.2625801265,0.3527042568,-0.0217220634,0.2763134837,-0.2073199898,0.2331521213,-0.0639110953,0.2056541741,-0.2941870987,-0.1854021698,0.0096485056,0.3292575777,0.0186899304,0.2050482631,-0.0001137066,-0.3125211596,0.0184408836,-0.1469954848,-0.1234472394,-0.1045938656,-0.1109406278,-0.195944801,0.0969132185,-0.1325443089,-0.0355381072,0.1201137006,-0.318225652,-0.0693274662,0.5164248943,-0.0813493282,0.168647632,-0.0962301493,-0.1875781119,-0.1486988664,0.3641577363,-0.1070586815,0.0575782731,0.004555345,0.2955908477,-0.0346819237,-0.1117430776,0.2009179592,0.0118584456,-0.214055717,0.1996009201,0.0452660359,0.2193309963,-0.1476968676,-0.1681335866,0.2540251315,0.1309775263,-0.0490120389,-0.4720951617,0.0694950223,-0.0954873562,-0.7001909018,-0.0471167974,-0.2837664187,0.1515046805,-0.1515443474,-0.1255643219,0.0069697043,0.1390461624,-0.2734936476,-0.136483863,0.0243614838,-0.1411269158,-0.0774002075,0.1040128618,-0.1568869799,-0.6167755127,0.2696429491,-0.1904052645,0.0336680152,-0.5840451121,0.0813066885,0.0431840122,0.1194143519,-0.1405089945,0.4723142385,0.1637546718,0.0697114617,0.1093147695,0.0862753019,-0.3274796307,0.8337019682,0.087496534,0.0176879056,0.1399512142,-0.0052242139,0.1916358024,-0.0466981567,0.0865587071,-0.5624172091,-0.0675756931,0.3009655178,-0.2086916715,-0.4418406785,-0.4798848629,0.1379334778,-0.2506815195,-0.1557816118,0.3023255765,0.3863398731,-0.335867852,0.0597458482,0.40995875,0.3697021306,-0.2531629503,-0.1009762734,-0.1679190695,0.0602376014,-0.2565958202,0.225626573,0.0209779311,0.0801946297,0.3454347551,0.0971023589,0.3335015774,0.004242064,0.1434872895,0.0575459935,-0.2455943674,-0.1922012866,-0.1943170577,0.0088598132,0.0848355591,-0.2524261475,-0.0161993299,-0.382820487,-0.3244469762,-0.1584510505,0.0603077821,0.1470858306,-0.1753459871,0.1231860295,-0.0035436349,0.3194283247,-0.0393082947,-0.1132980362,0.005361788,0.1049269363,-0.2670657039,-0.0042275405,0.493555367,-0.0255884975,-0.3216415644,-0.3817739487,0.4146699309,0.1683200598,-0.0606814697,0.270088762,0.0352221616,0.2871845067,0.0276340041,-0.1804422289,0.3904258907,-0.5985485911,-0.0689604282,-0.0158594176,-0.0115918266,-0.2362428457,0.2013259977,-0.1834621727,0.3161441982,-0.1831455827,0.4822333455,0.1010600477,0.0653936863,-0.2801431715,-0.2663798928,-0.1388558894,0.1392518878,-0.0844184905,0.1085980386,0.1279085279,-0.2155474573,0.3492257893,0.0701876506,-0.2282511145,0.0856477097,0.0376856215,0.1824711561,0.0331669003,0.2077793777,0.3472885191,0.1417553872,0.1336536705,-0.697609067,0.1592135727,0.0037085325,0.0860563889,-0.3027497232,-0.2085744739,0.0227134824,-0.4090459943,0.1354869604,-0.2492710054,0.2036932856,0.1237623915,0.001924585,0.1244986206,0.2435036302,0.182470277,-0.2680424452,-0.1919852942,0.202580899,0.0015646351,0.0111755775,0.4812694788,0.2334780693,0.1712279469,0.04375894,0.2832264006,0.0035683014,0.1227550358,0.1778895706,0.2196926922,0.2059095204,0.0626964644,0.1532783359,-0.0062190397,0.1477894634,-0.2205672115,-0.0119060511,0.455041945,-0.4066660404,0.1981160939,-0.2130469829,-0.0134293316,-0.0060455082,0.027414592,-0.3535484672,0.0184353963,0.325185895,0.1578260958,0.4982276857,-0.2068050951,0.2832622528,-0.1557312608,0.0057616415,-0.1006078348,0.0947469696,-0.040629521,0.1436415911,-0.0996471047,-0.0770215541,-0.4216920137,0.0819059685,0.1051404104,-0.1054541767,0.0385688841,-0.0031587586,-0.09889777,-0.1584773511,-0.0599150099,-0.2916539907,0.037877731,0.2082577944,0.1982465535,0.2047153264,-0.3442966044,-0.2024778128,-0.0576463677,-0.3000858426,0.1449369639,-0.0344874784,0.539752841,0.2608511448,0.1027044579,-0.2336456478,-0.3465655744,0.3593018949,-0.039825689,0.0450821519,0.1441108286,0.0771829188,0.4756244421,0.038900096,-0.0621122345,-0.0366739109,-0.1377660632,0.0031252198,0.0776271671,-0.3016474545,0.1216313019,0.1167247444,-0.3648528755,0.2510655224,0.1978717893,-0.5468524098,0.124114424,-0.1195817515,0.5138995051,0.3033767939,-0.301456511,0.023169307,-0.0888207555,0.2371552736,-0.0323478281,-0.2406721413,0.3399661481,-0.0452200957,0.1214821339,-0.0695555136,-0.1393463463,0.0858451575,-0.2094722241,0.3407860696,-0.0766743347,-0.2421108335,-0.1812707037,-0.0949681625,0.2206552625,0.1995648444,0.4270272851,-0.1554479897,-0.2086057812,0.3116022646,-0.1577573866,-0.2221233696,-0.2294902205,-0.1876649112,0.491013974,-0.1489214301,-0.0511640161,-0.3940578103,0.1894480735,-0.1001680195,0.2220212966,-0.0307293721,0.1455084085,0.1064650342,-0.1404236406,-0.0989815742,-0.1612590551,0.191022709,0.1008730531,0.4178522527,-0.2088980824,0.339697659,-0.0117482059,0.6103149056,0.1115433425,0.2892510593,0.1768062711,0.3831200898,0.1211188436,-0.0213042125,-0.2075167298,0.0879762396,-0.0352794267,-0.2933488488,0.1319354326,0.0984315276,0.4414548874,-0.2185959965,0.3410234451,-0.210394904,-0.1881791055,0.1058161855,-0.2011750489,-0.098917596,-0.1430128664,0.0723319352,0.1112953871,-0.0060305069,0.241292432,0.1390670687,0.3017922044,0.0698097423,-0.2126363516,0.1514588743,-0.3834177852,0.4566945732,-0.2437226623,-0.2038207948,-0.2165651768,-0.1013507843,-0.1215602979,-0.0875823721,0.214755252,0.0374684706,-0.2363127768,0.1458826363,-0.5916590691,-0.4131349325,-0.0120864147,-0.2095865607,-0.1605082005,0.2359598577,0.0095172767,-0.4821446836,-0.0756690502,-0.2295518965,0.057655625,-0.3014695048,-0.0373364463,-0.2727277875,-0.1172621697,-0.0696871579,0.0924727693,-0.2291588783,0.2875803113,-0.2921278179,0.1601521075,-0.1564980149,-0.0742772743,0.3070153594,0.1465413868,0.1504179686,-0.3316683471,0.078501232,0.2081436068,-0.1293841451,-0.1273933649,0.1969533414,-0.2532327771,-0.2951292992,0.3716384172,0.0178300925,0.2221780568,0.2946822643,-0.2342992872,0.0944958106,-0.3147600591,0.3028075993,-0.3613408506,-0.0944890603,0.3689983189,0.1033305824,0.0125515759,-0.2299200296,0.0690530762,0.1653673947,-0.3130488694,-0.3780952692,-0.0152913975,-0.278544426,0.5295838714,0.2223323137,0.9652849436,0.0470921658,0.0763823763,0.1551992446,0.0962854624,0.4637106061,-0.5718903542,-0.0351011157,-0.3878887594,-0.0913151056,0.0291725341,0.0367921069,-0.1170589402,-0.0540103428,-0.393668443,0.4557964206,-0.2637159228,-0.2005107701,-0.1655346602,-0.0229667872,0.1302414685,-0.2507967055,-0.3668765426,0.1868800819,0.0889842808,0.0542574562,-0.2669872046,-0.0393872634,-0.2502802014,-0.0585819371,0.0858374536,0.0232327431,-0.0212349873,0.1083106548,-0.0477350764,-0.15406394,-0.1582447439,0.544236064,0.5773678422,0.217205748,-0.0208211653,0.0600922368,-0.1236229986,0.5404776335,0.0934721455,-0.2224397063,0.2294900715,-0.0243547689,-0.020823203,0.3062645495,0.0182671379,-0.0775452256,-0.1989146322,-0.1544184834,-0.325826019,-0.3220331967,0.4116707444,0.2232385427,0.2346820235,-0.0311754644,0.0928960443,0.452919215,-0.3160113394,0.1871128827,0.1644583941,-0.32837376,0.0176813807,0.15692918,-0.2035150975,-0.2034178525,0.1447074562,-0.1222935095,0.0662494823,-0.1050124392,-0.2246523052,0.3282240927,-0.6158259511,0.1495267302,-0.0137531999,-0.2014415711,0.1079682112,0.5085456371,0.3355732262,-0.1233519018,0.3740538061,-0.1671467423,-0.2427298427,0.4513656497,0.0616037957,0.1218226328,-0.2861672938,-0.3142947853,-0.3326703906,0.6031278968,-0.3359348774,-0.2359313071,-0.4539035559,0.1343765706,-0.362043649,-0.2385994941,0.2074179947,0.2969813049,0.009648269,0.1842312962,-0.3344164193,-0.1230726838,-0.4081398249,0.1003050879,-0.1193289906,0.2269994169,-0.0668595135,0.4527094662,0.2298652232,-0.2671133578,0.1697508395,0.1802743822,0.2975305319,0.2053432614,0.0191491116,0.2244738787,-0.0069990172,-0.0262908693,0.2954408526,0.1316293776,0.1685137898,0.0559650958,-0.2218063325,-0.1717917323,0.3374065459,0.1799139082,0.2362015098,-0.0936959535,0.1139821261,-0.0410220772,0.0842017978,-0.2749198079,0.3915095925,0.3128079474,-0.0195628181,-0.0796531066,-0.1372188032,0.2240026593,0.0244851746,0.3415912092,-0.149001345,0.2489686459,-0.0112866266,0.3523798287,0.1521824896,0.0147120096,0.4298763573,-0.0189230144,-0.0702096298,-0.2236607671,0.1153164878,-0.2348205894,-0.1815804243,0.3596546948,0.4222204387,0.0077236784,0.0152922552,-0.0210295375,-0.1447943747,0.2545048892,-0.3928196132,0.6502040625,0.2231901735,0.0503745377,-0.1527456492,0.2164507508,0.0517553799,-0.2472672611,-0.2434281409,0.4218269885,0.0581700839,-0.3403600454,-0.3058310449,0.3959472775,-0.0788398981,-0.4969257414,-0.0322642475,-0.3686673641,-0.2251703739,-0.1759877354,0.0558964014,0.0035126531,-0.1908208132,0.0639134273,0.1989696622,-0.276350528,-0.5920048356,0.1823782623,0.0905692875,-0.0429586619,-0.0077434652,0.3834717572,0.0954965428,0.5452119708,0.4397696853,0.2471140325,0.1698268354,-0.4162213206,0.0099131372,-0.0941653773,0.1179095358,-0.117886357,0.4269874692,0.2142495662,-0.3480429947,0.1749587208,0.1014097407,-0.1391798109,0.2263496071,-0.2509038746,0.2887138426,-0.1855385751,0.4289951324,0.1102497876,0.3845459521,-0.2034668922,-0.054612238,-0.4368281364,0.0242949538,-0.2412265539,-0.0238137748,0.1368947774,0.1685836762,0.0787073448,0.1651517153,0.0353789181,0.0816258863,0.1796215922,-0.2963189483,0.2632241845,-0.1453604996,-0.0498373993,0.1904272586,0.3427992165,-0.3479309976,0.2818306983,0.1530394703,0.2411677092,-0.1471676826,0.086702168,-0.1079351977,0.1036498621,-0.4891263545,0.2790877223,-0.0577862002,0.0819780901,0.0164200626,0.1622750759,-0.2522097826,-0.030885892,0.0889545083,-0.0395877846,-0.219846189,0.120538421,0.4369190633,0.38095963,0.4347174168,0.2746351361,0.0491110533,-0.0917204767,-0.0348362811,-0.2374585867,0.0091844862,-0.1451218724,-0.0991911963,0.1483568549,0.1272278279,-0.0981932357,-0.2859273255,0.7760738134,0.10803321,-0.5372213125,-0.452224493,0.1022330523,0.2105946392,-0.0174015574,-0.2600540221,0.3088754416,-0.0147051318,0.6009328365,-0.1941381544,-0.1132751405,0.4475610256,-0.4307902455,0.3061504066,0.1107731387,0.3764396012,0.4347320795,-0.0453522541,-0.1358400434,0.0445140786,0.0719929487,0.0879144147,-0.006922279,0.1729042679,0.2557842135,-0.084685564,-0.2360916138,-0.2595790923,0.0349611454,0.0786984488,0.0781525522,-0.2363346517]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/796","title":"Seq2Seq Metrics QOL: Bleu, Rouge","comments":"prediction = [['Hey', 'how', 'are', 'you', '?']] \r\nreference=[['Hey', 'how', 'are', 'you', '?']]\r\nbleu.compute(predictions=prediction,references=reference)\r\n\r\nalso tried this kind of things lol\r\nI definitely need help too","body":"Putting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand:\r\n\r\n+ Bleu expects tokenization, can I just kwarg it like sacrebleu?\r\n+ different signatures, means that I would have had to add a lot of conditionals + pre and post processing: if I were going to replace the `calculate_rouge` and `calculate_bleu` functions here: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py#L61\r\n\r\n\r\n#### What I tried\r\n\r\n\r\nRouge experience:\r\n```python\r\n\r\nrouge = load_metric('rouge')\r\nrouge.add_batch(['hi im sam'], ['im daniel']) # fails\r\nrouge.add_batch(predictions=['hi im sam'], references=['im daniel']) # works\r\nrouge.compute() # huge messy output, but reasonable. Not worth integrating b\/c don't want to rewrite all the postprocessing.\r\n```\r\n\r\nBLEU experience:\r\n```python\r\nbleu = load_metric('bleu')\r\nbleu.add_batch(predictions=['hi im sam'], references=['im daniel'])\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n```\r\nAll of these raise `ValueError: Got a string but expected a list instead: 'im daniel'`\r\n\r\n#### Doc Typo\r\nThis says `dataset=load_metric(...)` which seems wrong, will cause `NameError`\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/98004483-ff0d0580-1dbd-11eb-9f35-6f35904611bb.png)\r\n\r\ncc @lhoestq, feel free to ignore.","comment_length":25,"text":"Seq2Seq Metrics QOL: Bleu, Rouge\nPutting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand:\r\n\r\n+ Bleu expects tokenization, can I just kwarg it like sacrebleu?\r\n+ different signatures, means that I would have had to add a lot of conditionals + pre and post processing: if I were going to replace the `calculate_rouge` and `calculate_bleu` functions here: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py#L61\r\n\r\n\r\n#### What I tried\r\n\r\n\r\nRouge experience:\r\n```python\r\n\r\nrouge = load_metric('rouge')\r\nrouge.add_batch(['hi im sam'], ['im daniel']) # fails\r\nrouge.add_batch(predictions=['hi im sam'], references=['im daniel']) # works\r\nrouge.compute() # huge messy output, but reasonable. Not worth integrating b\/c don't want to rewrite all the postprocessing.\r\n```\r\n\r\nBLEU experience:\r\n```python\r\nbleu = load_metric('bleu')\r\nbleu.add_batch(predictions=['hi im sam'], references=['im daniel'])\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n```\r\nAll of these raise `ValueError: Got a string but expected a list instead: 'im daniel'`\r\n\r\n#### Doc Typo\r\nThis says `dataset=load_metric(...)` which seems wrong, will cause `NameError`\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/98004483-ff0d0580-1dbd-11eb-9f35-6f35904611bb.png)\r\n\r\ncc @lhoestq, feel free to ignore.\nprediction = [['Hey', 'how', 'are', 'you', '?']] \r\nreference=[['Hey', 'how', 'are', 'you', '?']]\r\nbleu.compute(predictions=prediction,references=reference)\r\n\r\nalso tried this kind of things lol\r\nI definitely need help too","embeddings":[0.1769534498,-0.0281658582,-0.0853118747,0.0595532954,0.1316637546,-0.2772905231,0.0935510099,0.1194606274,-0.0715432093,0.0205882788,-0.2373280972,0.0115100741,-0.1035928577,-0.0028772415,0.0880227238,-0.231206879,-0.0192949343,0.0564785786,0.022084957,-0.209247455,-0.2625801265,0.3527042568,-0.0217220634,0.2763134837,-0.2073199898,0.2331521213,-0.0639110953,0.2056541741,-0.2941870987,-0.1854021698,0.0096485056,0.3292575777,0.0186899304,0.2050482631,-0.0001137066,-0.3125211596,0.0184408836,-0.1469954848,-0.1234472394,-0.1045938656,-0.1109406278,-0.195944801,0.0969132185,-0.1325443089,-0.0355381072,0.1201137006,-0.318225652,-0.0693274662,0.5164248943,-0.0813493282,0.168647632,-0.0962301493,-0.1875781119,-0.1486988664,0.3641577363,-0.1070586815,0.0575782731,0.004555345,0.2955908477,-0.0346819237,-0.1117430776,0.2009179592,0.0118584456,-0.214055717,0.1996009201,0.0452660359,0.2193309963,-0.1476968676,-0.1681335866,0.2540251315,0.1309775263,-0.0490120389,-0.4720951617,0.0694950223,-0.0954873562,-0.7001909018,-0.0471167974,-0.2837664187,0.1515046805,-0.1515443474,-0.1255643219,0.0069697043,0.1390461624,-0.2734936476,-0.136483863,0.0243614838,-0.1411269158,-0.0774002075,0.1040128618,-0.1568869799,-0.6167755127,0.2696429491,-0.1904052645,0.0336680152,-0.5840451121,0.0813066885,0.0431840122,0.1194143519,-0.1405089945,0.4723142385,0.1637546718,0.0697114617,0.1093147695,0.0862753019,-0.3274796307,0.8337019682,0.087496534,0.0176879056,0.1399512142,-0.0052242139,0.1916358024,-0.0466981567,0.0865587071,-0.5624172091,-0.0675756931,0.3009655178,-0.2086916715,-0.4418406785,-0.4798848629,0.1379334778,-0.2506815195,-0.1557816118,0.3023255765,0.3863398731,-0.335867852,0.0597458482,0.40995875,0.3697021306,-0.2531629503,-0.1009762734,-0.1679190695,0.0602376014,-0.2565958202,0.225626573,0.0209779311,0.0801946297,0.3454347551,0.0971023589,0.3335015774,0.004242064,0.1434872895,0.0575459935,-0.2455943674,-0.1922012866,-0.1943170577,0.0088598132,0.0848355591,-0.2524261475,-0.0161993299,-0.382820487,-0.3244469762,-0.1584510505,0.0603077821,0.1470858306,-0.1753459871,0.1231860295,-0.0035436349,0.3194283247,-0.0393082947,-0.1132980362,0.005361788,0.1049269363,-0.2670657039,-0.0042275405,0.493555367,-0.0255884975,-0.3216415644,-0.3817739487,0.4146699309,0.1683200598,-0.0606814697,0.270088762,0.0352221616,0.2871845067,0.0276340041,-0.1804422289,0.3904258907,-0.5985485911,-0.0689604282,-0.0158594176,-0.0115918266,-0.2362428457,0.2013259977,-0.1834621727,0.3161441982,-0.1831455827,0.4822333455,0.1010600477,0.0653936863,-0.2801431715,-0.2663798928,-0.1388558894,0.1392518878,-0.0844184905,0.1085980386,0.1279085279,-0.2155474573,0.3492257893,0.0701876506,-0.2282511145,0.0856477097,0.0376856215,0.1824711561,0.0331669003,0.2077793777,0.3472885191,0.1417553872,0.1336536705,-0.697609067,0.1592135727,0.0037085325,0.0860563889,-0.3027497232,-0.2085744739,0.0227134824,-0.4090459943,0.1354869604,-0.2492710054,0.2036932856,0.1237623915,0.001924585,0.1244986206,0.2435036302,0.182470277,-0.2680424452,-0.1919852942,0.202580899,0.0015646351,0.0111755775,0.4812694788,0.2334780693,0.1712279469,0.04375894,0.2832264006,0.0035683014,0.1227550358,0.1778895706,0.2196926922,0.2059095204,0.0626964644,0.1532783359,-0.0062190397,0.1477894634,-0.2205672115,-0.0119060511,0.455041945,-0.4066660404,0.1981160939,-0.2130469829,-0.0134293316,-0.0060455082,0.027414592,-0.3535484672,0.0184353963,0.325185895,0.1578260958,0.4982276857,-0.2068050951,0.2832622528,-0.1557312608,0.0057616415,-0.1006078348,0.0947469696,-0.040629521,0.1436415911,-0.0996471047,-0.0770215541,-0.4216920137,0.0819059685,0.1051404104,-0.1054541767,0.0385688841,-0.0031587586,-0.09889777,-0.1584773511,-0.0599150099,-0.2916539907,0.037877731,0.2082577944,0.1982465535,0.2047153264,-0.3442966044,-0.2024778128,-0.0576463677,-0.3000858426,0.1449369639,-0.0344874784,0.539752841,0.2608511448,0.1027044579,-0.2336456478,-0.3465655744,0.3593018949,-0.039825689,0.0450821519,0.1441108286,0.0771829188,0.4756244421,0.038900096,-0.0621122345,-0.0366739109,-0.1377660632,0.0031252198,0.0776271671,-0.3016474545,0.1216313019,0.1167247444,-0.3648528755,0.2510655224,0.1978717893,-0.5468524098,0.124114424,-0.1195817515,0.5138995051,0.3033767939,-0.301456511,0.023169307,-0.0888207555,0.2371552736,-0.0323478281,-0.2406721413,0.3399661481,-0.0452200957,0.1214821339,-0.0695555136,-0.1393463463,0.0858451575,-0.2094722241,0.3407860696,-0.0766743347,-0.2421108335,-0.1812707037,-0.0949681625,0.2206552625,0.1995648444,0.4270272851,-0.1554479897,-0.2086057812,0.3116022646,-0.1577573866,-0.2221233696,-0.2294902205,-0.1876649112,0.491013974,-0.1489214301,-0.0511640161,-0.3940578103,0.1894480735,-0.1001680195,0.2220212966,-0.0307293721,0.1455084085,0.1064650342,-0.1404236406,-0.0989815742,-0.1612590551,0.191022709,0.1008730531,0.4178522527,-0.2088980824,0.339697659,-0.0117482059,0.6103149056,0.1115433425,0.2892510593,0.1768062711,0.3831200898,0.1211188436,-0.0213042125,-0.2075167298,0.0879762396,-0.0352794267,-0.2933488488,0.1319354326,0.0984315276,0.4414548874,-0.2185959965,0.3410234451,-0.210394904,-0.1881791055,0.1058161855,-0.2011750489,-0.098917596,-0.1430128664,0.0723319352,0.1112953871,-0.0060305069,0.241292432,0.1390670687,0.3017922044,0.0698097423,-0.2126363516,0.1514588743,-0.3834177852,0.4566945732,-0.2437226623,-0.2038207948,-0.2165651768,-0.1013507843,-0.1215602979,-0.0875823721,0.214755252,0.0374684706,-0.2363127768,0.1458826363,-0.5916590691,-0.4131349325,-0.0120864147,-0.2095865607,-0.1605082005,0.2359598577,0.0095172767,-0.4821446836,-0.0756690502,-0.2295518965,0.057655625,-0.3014695048,-0.0373364463,-0.2727277875,-0.1172621697,-0.0696871579,0.0924727693,-0.2291588783,0.2875803113,-0.2921278179,0.1601521075,-0.1564980149,-0.0742772743,0.3070153594,0.1465413868,0.1504179686,-0.3316683471,0.078501232,0.2081436068,-0.1293841451,-0.1273933649,0.1969533414,-0.2532327771,-0.2951292992,0.3716384172,0.0178300925,0.2221780568,0.2946822643,-0.2342992872,0.0944958106,-0.3147600591,0.3028075993,-0.3613408506,-0.0944890603,0.3689983189,0.1033305824,0.0125515759,-0.2299200296,0.0690530762,0.1653673947,-0.3130488694,-0.3780952692,-0.0152913975,-0.278544426,0.5295838714,0.2223323137,0.9652849436,0.0470921658,0.0763823763,0.1551992446,0.0962854624,0.4637106061,-0.5718903542,-0.0351011157,-0.3878887594,-0.0913151056,0.0291725341,0.0367921069,-0.1170589402,-0.0540103428,-0.393668443,0.4557964206,-0.2637159228,-0.2005107701,-0.1655346602,-0.0229667872,0.1302414685,-0.2507967055,-0.3668765426,0.1868800819,0.0889842808,0.0542574562,-0.2669872046,-0.0393872634,-0.2502802014,-0.0585819371,0.0858374536,0.0232327431,-0.0212349873,0.1083106548,-0.0477350764,-0.15406394,-0.1582447439,0.544236064,0.5773678422,0.217205748,-0.0208211653,0.0600922368,-0.1236229986,0.5404776335,0.0934721455,-0.2224397063,0.2294900715,-0.0243547689,-0.020823203,0.3062645495,0.0182671379,-0.0775452256,-0.1989146322,-0.1544184834,-0.325826019,-0.3220331967,0.4116707444,0.2232385427,0.2346820235,-0.0311754644,0.0928960443,0.452919215,-0.3160113394,0.1871128827,0.1644583941,-0.32837376,0.0176813807,0.15692918,-0.2035150975,-0.2034178525,0.1447074562,-0.1222935095,0.0662494823,-0.1050124392,-0.2246523052,0.3282240927,-0.6158259511,0.1495267302,-0.0137531999,-0.2014415711,0.1079682112,0.5085456371,0.3355732262,-0.1233519018,0.3740538061,-0.1671467423,-0.2427298427,0.4513656497,0.0616037957,0.1218226328,-0.2861672938,-0.3142947853,-0.3326703906,0.6031278968,-0.3359348774,-0.2359313071,-0.4539035559,0.1343765706,-0.362043649,-0.2385994941,0.2074179947,0.2969813049,0.009648269,0.1842312962,-0.3344164193,-0.1230726838,-0.4081398249,0.1003050879,-0.1193289906,0.2269994169,-0.0668595135,0.4527094662,0.2298652232,-0.2671133578,0.1697508395,0.1802743822,0.2975305319,0.2053432614,0.0191491116,0.2244738787,-0.0069990172,-0.0262908693,0.2954408526,0.1316293776,0.1685137898,0.0559650958,-0.2218063325,-0.1717917323,0.3374065459,0.1799139082,0.2362015098,-0.0936959535,0.1139821261,-0.0410220772,0.0842017978,-0.2749198079,0.3915095925,0.3128079474,-0.0195628181,-0.0796531066,-0.1372188032,0.2240026593,0.0244851746,0.3415912092,-0.149001345,0.2489686459,-0.0112866266,0.3523798287,0.1521824896,0.0147120096,0.4298763573,-0.0189230144,-0.0702096298,-0.2236607671,0.1153164878,-0.2348205894,-0.1815804243,0.3596546948,0.4222204387,0.0077236784,0.0152922552,-0.0210295375,-0.1447943747,0.2545048892,-0.3928196132,0.6502040625,0.2231901735,0.0503745377,-0.1527456492,0.2164507508,0.0517553799,-0.2472672611,-0.2434281409,0.4218269885,0.0581700839,-0.3403600454,-0.3058310449,0.3959472775,-0.0788398981,-0.4969257414,-0.0322642475,-0.3686673641,-0.2251703739,-0.1759877354,0.0558964014,0.0035126531,-0.1908208132,0.0639134273,0.1989696622,-0.276350528,-0.5920048356,0.1823782623,0.0905692875,-0.0429586619,-0.0077434652,0.3834717572,0.0954965428,0.5452119708,0.4397696853,0.2471140325,0.1698268354,-0.4162213206,0.0099131372,-0.0941653773,0.1179095358,-0.117886357,0.4269874692,0.2142495662,-0.3480429947,0.1749587208,0.1014097407,-0.1391798109,0.2263496071,-0.2509038746,0.2887138426,-0.1855385751,0.4289951324,0.1102497876,0.3845459521,-0.2034668922,-0.054612238,-0.4368281364,0.0242949538,-0.2412265539,-0.0238137748,0.1368947774,0.1685836762,0.0787073448,0.1651517153,0.0353789181,0.0816258863,0.1796215922,-0.2963189483,0.2632241845,-0.1453604996,-0.0498373993,0.1904272586,0.3427992165,-0.3479309976,0.2818306983,0.1530394703,0.2411677092,-0.1471676826,0.086702168,-0.1079351977,0.1036498621,-0.4891263545,0.2790877223,-0.0577862002,0.0819780901,0.0164200626,0.1622750759,-0.2522097826,-0.030885892,0.0889545083,-0.0395877846,-0.219846189,0.120538421,0.4369190633,0.38095963,0.4347174168,0.2746351361,0.0491110533,-0.0917204767,-0.0348362811,-0.2374585867,0.0091844862,-0.1451218724,-0.0991911963,0.1483568549,0.1272278279,-0.0981932357,-0.2859273255,0.7760738134,0.10803321,-0.5372213125,-0.452224493,0.1022330523,0.2105946392,-0.0174015574,-0.2600540221,0.3088754416,-0.0147051318,0.6009328365,-0.1941381544,-0.1132751405,0.4475610256,-0.4307902455,0.3061504066,0.1107731387,0.3764396012,0.4347320795,-0.0453522541,-0.1358400434,0.0445140786,0.0719929487,0.0879144147,-0.006922279,0.1729042679,0.2557842135,-0.084685564,-0.2360916138,-0.2595790923,0.0349611454,0.0786984488,0.0781525522,-0.2363346517]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/796","title":"Seq2Seq Metrics QOL: Bleu, Rouge","comments":"Hi !\r\n\r\nAs described in the documentation for `bleu`:\r\n```\r\nArgs:\r\n predictions: list of translations to score.\r\n Each translation should be tokenized into a list of tokens.\r\n references: list of lists of references for each translation.\r\n Each reference should be tokenized into a list of tokens.\r\n```\r\n\r\nTherefore you can use this metric this way:\r\n```python\r\nfrom datasets import load_metric\r\n\r\npredictions = [\r\n [\"hello\", \"there\", \"general\", \"kenobi\"], # tokenized prediction of the first sample\r\n [\"foo\", \"bar\", \"foobar\"] # tokenized prediction of the second sample\r\n]\r\nreferences = [\r\n [[\"hello\", \"there\", \"general\", \"kenobi\"], [\"hello\", \"there\", \"!\"]], # tokenized references for the first sample (2 references)\r\n [[\"foo\", \"bar\", \"foobar\"]] # tokenized references for the second sample (1 reference)\r\n]\r\n\r\nbleu = load_metric(\"bleu\")\r\nbleu.compute(predictions=predictions, references=references)\r\n# Or you can also add batches before calling compute()\r\n# bleu.add_batch(predictions=predictions, references=references)\r\n# bleu.compute()\r\n```\r\n\r\nHope this helps :)","body":"Putting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand:\r\n\r\n+ Bleu expects tokenization, can I just kwarg it like sacrebleu?\r\n+ different signatures, means that I would have had to add a lot of conditionals + pre and post processing: if I were going to replace the `calculate_rouge` and `calculate_bleu` functions here: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py#L61\r\n\r\n\r\n#### What I tried\r\n\r\n\r\nRouge experience:\r\n```python\r\n\r\nrouge = load_metric('rouge')\r\nrouge.add_batch(['hi im sam'], ['im daniel']) # fails\r\nrouge.add_batch(predictions=['hi im sam'], references=['im daniel']) # works\r\nrouge.compute() # huge messy output, but reasonable. Not worth integrating b\/c don't want to rewrite all the postprocessing.\r\n```\r\n\r\nBLEU experience:\r\n```python\r\nbleu = load_metric('bleu')\r\nbleu.add_batch(predictions=['hi im sam'], references=['im daniel'])\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n```\r\nAll of these raise `ValueError: Got a string but expected a list instead: 'im daniel'`\r\n\r\n#### Doc Typo\r\nThis says `dataset=load_metric(...)` which seems wrong, will cause `NameError`\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/98004483-ff0d0580-1dbd-11eb-9f35-6f35904611bb.png)\r\n\r\ncc @lhoestq, feel free to ignore.","comment_length":142,"text":"Seq2Seq Metrics QOL: Bleu, Rouge\nPutting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand:\r\n\r\n+ Bleu expects tokenization, can I just kwarg it like sacrebleu?\r\n+ different signatures, means that I would have had to add a lot of conditionals + pre and post processing: if I were going to replace the `calculate_rouge` and `calculate_bleu` functions here: https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/utils.py#L61\r\n\r\n\r\n#### What I tried\r\n\r\n\r\nRouge experience:\r\n```python\r\n\r\nrouge = load_metric('rouge')\r\nrouge.add_batch(['hi im sam'], ['im daniel']) # fails\r\nrouge.add_batch(predictions=['hi im sam'], references=['im daniel']) # works\r\nrouge.compute() # huge messy output, but reasonable. Not worth integrating b\/c don't want to rewrite all the postprocessing.\r\n```\r\n\r\nBLEU experience:\r\n```python\r\nbleu = load_metric('bleu')\r\nbleu.add_batch(predictions=['hi im sam'], references=['im daniel'])\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n\r\nbleu.add_batch(predictions=[['hi im sam']], references=[['im daniel']])\r\n```\r\nAll of these raise `ValueError: Got a string but expected a list instead: 'im daniel'`\r\n\r\n#### Doc Typo\r\nThis says `dataset=load_metric(...)` which seems wrong, will cause `NameError`\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/98004483-ff0d0580-1dbd-11eb-9f35-6f35904611bb.png)\r\n\r\ncc @lhoestq, feel free to ignore.\nHi !\r\n\r\nAs described in the documentation for `bleu`:\r\n```\r\nArgs:\r\n predictions: list of translations to score.\r\n Each translation should be tokenized into a list of tokens.\r\n references: list of lists of references for each translation.\r\n Each reference should be tokenized into a list of tokens.\r\n```\r\n\r\nTherefore you can use this metric this way:\r\n```python\r\nfrom datasets import load_metric\r\n\r\npredictions = [\r\n [\"hello\", \"there\", \"general\", \"kenobi\"], # tokenized prediction of the first sample\r\n [\"foo\", \"bar\", \"foobar\"] # tokenized prediction of the second sample\r\n]\r\nreferences = [\r\n [[\"hello\", \"there\", \"general\", \"kenobi\"], [\"hello\", \"there\", \"!\"]], # tokenized references for the first sample (2 references)\r\n [[\"foo\", \"bar\", \"foobar\"]] # tokenized references for the second sample (1 reference)\r\n]\r\n\r\nbleu = load_metric(\"bleu\")\r\nbleu.compute(predictions=predictions, references=references)\r\n# Or you can also add batches before calling compute()\r\n# bleu.add_batch(predictions=predictions, references=references)\r\n# bleu.compute()\r\n```\r\n\r\nHope this helps :)","embeddings":[0.1769534498,-0.0281658582,-0.0853118747,0.0595532954,0.1316637546,-0.2772905231,0.0935510099,0.1194606274,-0.0715432093,0.0205882788,-0.2373280972,0.0115100741,-0.1035928577,-0.0028772415,0.0880227238,-0.231206879,-0.0192949343,0.0564785786,0.022084957,-0.209247455,-0.2625801265,0.3527042568,-0.0217220634,0.2763134837,-0.2073199898,0.2331521213,-0.0639110953,0.2056541741,-0.2941870987,-0.1854021698,0.0096485056,0.3292575777,0.0186899304,0.2050482631,-0.0001137066,-0.3125211596,0.0184408836,-0.1469954848,-0.1234472394,-0.1045938656,-0.1109406278,-0.195944801,0.0969132185,-0.1325443089,-0.0355381072,0.1201137006,-0.318225652,-0.0693274662,0.5164248943,-0.0813493282,0.168647632,-0.0962301493,-0.1875781119,-0.1486988664,0.3641577363,-0.1070586815,0.0575782731,0.004555345,0.2955908477,-0.0346819237,-0.1117430776,0.2009179592,0.0118584456,-0.214055717,0.1996009201,0.0452660359,0.2193309963,-0.1476968676,-0.1681335866,0.2540251315,0.1309775263,-0.0490120389,-0.4720951617,0.0694950223,-0.0954873562,-0.7001909018,-0.0471167974,-0.2837664187,0.1515046805,-0.1515443474,-0.1255643219,0.0069697043,0.1390461624,-0.2734936476,-0.136483863,0.0243614838,-0.1411269158,-0.0774002075,0.1040128618,-0.1568869799,-0.6167755127,0.2696429491,-0.1904052645,0.0336680152,-0.5840451121,0.0813066885,0.0431840122,0.1194143519,-0.1405089945,0.4723142385,0.1637546718,0.0697114617,0.1093147695,0.0862753019,-0.3274796307,0.8337019682,0.087496534,0.0176879056,0.1399512142,-0.0052242139,0.1916358024,-0.0466981567,0.0865587071,-0.5624172091,-0.0675756931,0.3009655178,-0.2086916715,-0.4418406785,-0.4798848629,0.1379334778,-0.2506815195,-0.1557816118,0.3023255765,0.3863398731,-0.335867852,0.0597458482,0.40995875,0.3697021306,-0.2531629503,-0.1009762734,-0.1679190695,0.0602376014,-0.2565958202,0.225626573,0.0209779311,0.0801946297,0.3454347551,0.0971023589,0.3335015774,0.004242064,0.1434872895,0.0575459935,-0.2455943674,-0.1922012866,-0.1943170577,0.0088598132,0.0848355591,-0.2524261475,-0.0161993299,-0.382820487,-0.3244469762,-0.1584510505,0.0603077821,0.1470858306,-0.1753459871,0.1231860295,-0.0035436349,0.3194283247,-0.0393082947,-0.1132980362,0.005361788,0.1049269363,-0.2670657039,-0.0042275405,0.493555367,-0.0255884975,-0.3216415644,-0.3817739487,0.4146699309,0.1683200598,-0.0606814697,0.270088762,0.0352221616,0.2871845067,0.0276340041,-0.1804422289,0.3904258907,-0.5985485911,-0.0689604282,-0.0158594176,-0.0115918266,-0.2362428457,0.2013259977,-0.1834621727,0.3161441982,-0.1831455827,0.4822333455,0.1010600477,0.0653936863,-0.2801431715,-0.2663798928,-0.1388558894,0.1392518878,-0.0844184905,0.1085980386,0.1279085279,-0.2155474573,0.3492257893,0.0701876506,-0.2282511145,0.0856477097,0.0376856215,0.1824711561,0.0331669003,0.2077793777,0.3472885191,0.1417553872,0.1336536705,-0.697609067,0.1592135727,0.0037085325,0.0860563889,-0.3027497232,-0.2085744739,0.0227134824,-0.4090459943,0.1354869604,-0.2492710054,0.2036932856,0.1237623915,0.001924585,0.1244986206,0.2435036302,0.182470277,-0.2680424452,-0.1919852942,0.202580899,0.0015646351,0.0111755775,0.4812694788,0.2334780693,0.1712279469,0.04375894,0.2832264006,0.0035683014,0.1227550358,0.1778895706,0.2196926922,0.2059095204,0.0626964644,0.1532783359,-0.0062190397,0.1477894634,-0.2205672115,-0.0119060511,0.455041945,-0.4066660404,0.1981160939,-0.2130469829,-0.0134293316,-0.0060455082,0.027414592,-0.3535484672,0.0184353963,0.325185895,0.1578260958,0.4982276857,-0.2068050951,0.2832622528,-0.1557312608,0.0057616415,-0.1006078348,0.0947469696,-0.040629521,0.1436415911,-0.0996471047,-0.0770215541,-0.4216920137,0.0819059685,0.1051404104,-0.1054541767,0.0385688841,-0.0031587586,-0.09889777,-0.1584773511,-0.0599150099,-0.2916539907,0.037877731,0.2082577944,0.1982465535,0.2047153264,-0.3442966044,-0.2024778128,-0.0576463677,-0.3000858426,0.1449369639,-0.0344874784,0.539752841,0.2608511448,0.1027044579,-0.2336456478,-0.3465655744,0.3593018949,-0.039825689,0.0450821519,0.1441108286,0.0771829188,0.4756244421,0.038900096,-0.0621122345,-0.0366739109,-0.1377660632,0.0031252198,0.0776271671,-0.3016474545,0.1216313019,0.1167247444,-0.3648528755,0.2510655224,0.1978717893,-0.5468524098,0.124114424,-0.1195817515,0.5138995051,0.3033767939,-0.301456511,0.023169307,-0.0888207555,0.2371552736,-0.0323478281,-0.2406721413,0.3399661481,-0.0452200957,0.1214821339,-0.0695555136,-0.1393463463,0.0858451575,-0.2094722241,0.3407860696,-0.0766743347,-0.2421108335,-0.1812707037,-0.0949681625,0.2206552625,0.1995648444,0.4270272851,-0.1554479897,-0.2086057812,0.3116022646,-0.1577573866,-0.2221233696,-0.2294902205,-0.1876649112,0.491013974,-0.1489214301,-0.0511640161,-0.3940578103,0.1894480735,-0.1001680195,0.2220212966,-0.0307293721,0.1455084085,0.1064650342,-0.1404236406,-0.0989815742,-0.1612590551,0.191022709,0.1008730531,0.4178522527,-0.2088980824,0.339697659,-0.0117482059,0.6103149056,0.1115433425,0.2892510593,0.1768062711,0.3831200898,0.1211188436,-0.0213042125,-0.2075167298,0.0879762396,-0.0352794267,-0.2933488488,0.1319354326,0.0984315276,0.4414548874,-0.2185959965,0.3410234451,-0.210394904,-0.1881791055,0.1058161855,-0.2011750489,-0.098917596,-0.1430128664,0.0723319352,0.1112953871,-0.0060305069,0.241292432,0.1390670687,0.3017922044,0.0698097423,-0.2126363516,0.1514588743,-0.3834177852,0.4566945732,-0.2437226623,-0.2038207948,-0.2165651768,-0.1013507843,-0.1215602979,-0.0875823721,0.214755252,0.0374684706,-0.2363127768,0.1458826363,-0.5916590691,-0.4131349325,-0.0120864147,-0.2095865607,-0.1605082005,0.2359598577,0.0095172767,-0.4821446836,-0.0756690502,-0.2295518965,0.057655625,-0.3014695048,-0.0373364463,-0.2727277875,-0.1172621697,-0.0696871579,0.0924727693,-0.2291588783,0.2875803113,-0.2921278179,0.1601521075,-0.1564980149,-0.0742772743,0.3070153594,0.1465413868,0.1504179686,-0.3316683471,0.078501232,0.2081436068,-0.1293841451,-0.1273933649,0.1969533414,-0.2532327771,-0.2951292992,0.3716384172,0.0178300925,0.2221780568,0.2946822643,-0.2342992872,0.0944958106,-0.3147600591,0.3028075993,-0.3613408506,-0.0944890603,0.3689983189,0.1033305824,0.0125515759,-0.2299200296,0.0690530762,0.1653673947,-0.3130488694,-0.3780952692,-0.0152913975,-0.278544426,0.5295838714,0.2223323137,0.9652849436,0.0470921658,0.0763823763,0.1551992446,0.0962854624,0.4637106061,-0.5718903542,-0.0351011157,-0.3878887594,-0.0913151056,0.0291725341,0.0367921069,-0.1170589402,-0.0540103428,-0.393668443,0.4557964206,-0.2637159228,-0.2005107701,-0.1655346602,-0.0229667872,0.1302414685,-0.2507967055,-0.3668765426,0.1868800819,0.0889842808,0.0542574562,-0.2669872046,-0.0393872634,-0.2502802014,-0.0585819371,0.0858374536,0.0232327431,-0.0212349873,0.1083106548,-0.0477350764,-0.15406394,-0.1582447439,0.544236064,0.5773678422,0.217205748,-0.0208211653,0.0600922368,-0.1236229986,0.5404776335,0.0934721455,-0.2224397063,0.2294900715,-0.0243547689,-0.020823203,0.3062645495,0.0182671379,-0.0775452256,-0.1989146322,-0.1544184834,-0.325826019,-0.3220331967,0.4116707444,0.2232385427,0.2346820235,-0.0311754644,0.0928960443,0.452919215,-0.3160113394,0.1871128827,0.1644583941,-0.32837376,0.0176813807,0.15692918,-0.2035150975,-0.2034178525,0.1447074562,-0.1222935095,0.0662494823,-0.1050124392,-0.2246523052,0.3282240927,-0.6158259511,0.1495267302,-0.0137531999,-0.2014415711,0.1079682112,0.5085456371,0.3355732262,-0.1233519018,0.3740538061,-0.1671467423,-0.2427298427,0.4513656497,0.0616037957,0.1218226328,-0.2861672938,-0.3142947853,-0.3326703906,0.6031278968,-0.3359348774,-0.2359313071,-0.4539035559,0.1343765706,-0.362043649,-0.2385994941,0.2074179947,0.2969813049,0.009648269,0.1842312962,-0.3344164193,-0.1230726838,-0.4081398249,0.1003050879,-0.1193289906,0.2269994169,-0.0668595135,0.4527094662,0.2298652232,-0.2671133578,0.1697508395,0.1802743822,0.2975305319,0.2053432614,0.0191491116,0.2244738787,-0.0069990172,-0.0262908693,0.2954408526,0.1316293776,0.1685137898,0.0559650958,-0.2218063325,-0.1717917323,0.3374065459,0.1799139082,0.2362015098,-0.0936959535,0.1139821261,-0.0410220772,0.0842017978,-0.2749198079,0.3915095925,0.3128079474,-0.0195628181,-0.0796531066,-0.1372188032,0.2240026593,0.0244851746,0.3415912092,-0.149001345,0.2489686459,-0.0112866266,0.3523798287,0.1521824896,0.0147120096,0.4298763573,-0.0189230144,-0.0702096298,-0.2236607671,0.1153164878,-0.2348205894,-0.1815804243,0.3596546948,0.4222204387,0.0077236784,0.0152922552,-0.0210295375,-0.1447943747,0.2545048892,-0.3928196132,0.6502040625,0.2231901735,0.0503745377,-0.1527456492,0.2164507508,0.0517553799,-0.2472672611,-0.2434281409,0.4218269885,0.0581700839,-0.3403600454,-0.3058310449,0.3959472775,-0.0788398981,-0.4969257414,-0.0322642475,-0.3686673641,-0.2251703739,-0.1759877354,0.0558964014,0.0035126531,-0.1908208132,0.0639134273,0.1989696622,-0.276350528,-0.5920048356,0.1823782623,0.0905692875,-0.0429586619,-0.0077434652,0.3834717572,0.0954965428,0.5452119708,0.4397696853,0.2471140325,0.1698268354,-0.4162213206,0.0099131372,-0.0941653773,0.1179095358,-0.117886357,0.4269874692,0.2142495662,-0.3480429947,0.1749587208,0.1014097407,-0.1391798109,0.2263496071,-0.2509038746,0.2887138426,-0.1855385751,0.4289951324,0.1102497876,0.3845459521,-0.2034668922,-0.054612238,-0.4368281364,0.0242949538,-0.2412265539,-0.0238137748,0.1368947774,0.1685836762,0.0787073448,0.1651517153,0.0353789181,0.0816258863,0.1796215922,-0.2963189483,0.2632241845,-0.1453604996,-0.0498373993,0.1904272586,0.3427992165,-0.3479309976,0.2818306983,0.1530394703,0.2411677092,-0.1471676826,0.086702168,-0.1079351977,0.1036498621,-0.4891263545,0.2790877223,-0.0577862002,0.0819780901,0.0164200626,0.1622750759,-0.2522097826,-0.030885892,0.0889545083,-0.0395877846,-0.219846189,0.120538421,0.4369190633,0.38095963,0.4347174168,0.2746351361,0.0491110533,-0.0917204767,-0.0348362811,-0.2374585867,0.0091844862,-0.1451218724,-0.0991911963,0.1483568549,0.1272278279,-0.0981932357,-0.2859273255,0.7760738134,0.10803321,-0.5372213125,-0.452224493,0.1022330523,0.2105946392,-0.0174015574,-0.2600540221,0.3088754416,-0.0147051318,0.6009328365,-0.1941381544,-0.1132751405,0.4475610256,-0.4307902455,0.3061504066,0.1107731387,0.3764396012,0.4347320795,-0.0453522541,-0.1358400434,0.0445140786,0.0719929487,0.0879144147,-0.006922279,0.1729042679,0.2557842135,-0.084685564,-0.2360916138,-0.2595790923,0.0349611454,0.0786984488,0.0781525522,-0.2363346517]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/795","title":"Descriptions of raw and processed versions of wikitext are inverted","comments":"Yes indeed ! Thanks for reporting","body":"Nothing of importance, but it looks like the descriptions of wikitext-n-v1 and wikitext-n-raw-v1 are inverted for both n=2 and n=103. I just verified by loading them and the `<unk>` tokens are present in the non-raw versions, which confirms that it's a mere inversion of the descriptions and not of the datasets themselves.\r\n\r\nAlso it would be nice if those descriptions appeared in the dataset explorer.\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/87bd0864845ea0a1dd7167918dc5f341bf807bd3\/datasets\/wikitext\/wikitext.py#L52","comment_length":6,"text":"Descriptions of raw and processed versions of wikitext are inverted\nNothing of importance, but it looks like the descriptions of wikitext-n-v1 and wikitext-n-raw-v1 are inverted for both n=2 and n=103. I just verified by loading them and the `<unk>` tokens are present in the non-raw versions, which confirms that it's a mere inversion of the descriptions and not of the datasets themselves.\r\n\r\nAlso it would be nice if those descriptions appeared in the dataset explorer.\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/87bd0864845ea0a1dd7167918dc5f341bf807bd3\/datasets\/wikitext\/wikitext.py#L52\nYes indeed ! Thanks for reporting","embeddings":[0.1271498054,-0.5342550874,-0.0764296576,0.199450016,-0.2635107934,-0.0975624472,-0.1517811716,0.1650318205,0.1415553838,0.0051441216,-0.1823011339,0.1891187876,0.2802779973,0.1241092682,-0.0588884316,-0.079759866,0.3380807936,0.4160712361,-0.395632267,-0.2604143322,-0.2653579712,0.4940167367,-0.430556953,0.4160549641,0.0485300347,0.0852274299,0.1297219992,-0.0773343369,-0.4234927893,-0.2389664948,0.0949080437,0.000567615,0.0724763945,0.1639053375,-0.0001165944,-0.0758342221,0.2539846897,0.0833365694,-0.1843284667,0.4150004387,-0.1760118604,-0.2924150229,0.1187356561,-0.2455894947,-0.0535319895,-0.532694757,0.041229941,-0.1595708132,0.2786550522,0.025496712,0.1125580817,0.3560987115,0.0245670341,0.124584876,-0.1345583051,0.286449492,-0.2153586745,-0.4236371517,0.0287407171,0.17908369,0.2989568114,0.150022313,0.348872602,-0.060028255,0.1075158119,0.0864711255,0.3863157928,-0.3547534049,0.1696989536,0.3579327762,0.2167419344,0.0660902113,-0.4084389806,-0.1008179188,-0.0188740809,-0.3260421753,0.2659242153,0.0923222527,0.0638873726,0.3853289485,-0.2993252873,0.2338102907,0.1161369234,-0.0574889407,-0.6261865497,0.6967131495,-0.0628164709,0.0385348164,-0.1525173187,-0.3424025476,0.1719247252,-0.2624451816,-0.16916053,0.203207925,-0.3918353617,-0.012089191,-0.1657560766,0.1438358873,-0.0514968298,-0.1135589927,-0.0636554509,-0.0434999876,-0.1558102965,-0.0059807636,0.1320894212,0.1838515699,0.39950791,0.070474349,-0.2665069103,-0.0628487691,0.1084995791,-0.0115851555,-0.0853319168,-0.2782616019,-0.450632751,-0.102912426,-0.0614860728,-0.0690714791,-0.4563675523,-0.0578318685,-0.0301945973,0.1524980515,-0.2179052681,-0.2854312956,0.0726698339,0.3041699231,0.0029112508,0.0572773069,-0.1242960244,-0.0931699574,-0.1911326498,0.0438715518,-0.1438820809,0.0170809254,-0.0749831647,-0.1570484638,0.1655365378,0.0371770598,0.2188750952,0.0404536203,-0.1470821202,0.0900025591,0.5967655778,0.3653899133,-0.1491587162,0.5452039838,0.252130419,-0.1572161019,-0.1102239937,0.2990644872,-0.4543692172,-0.1101096421,-0.0526351891,0.174958244,-0.480931133,-0.0376615115,0.0569536909,0.2095514238,-0.0696393326,0.0032996393,0.2747723162,-0.2568790913,-0.1510345042,0.1624949276,0.2256581634,-0.1824951321,-0.5066196322,0.0219913255,-0.3199557364,-0.3118613958,-0.0690267235,0.5556253791,-0.0442124829,0.2121711224,-0.3629625738,0.3563351333,0.2109129429,-0.3331426382,0.0435870104,-0.1216727048,-0.1114798039,0.517957747,-0.3326848745,-0.41662848,0.0664626285,-0.1495087743,-0.1471648216,0.2496144325,0.4921108186,0.0501330011,-0.2034593225,-0.2606161237,0.6235286593,-0.3352941573,-0.138620466,-0.1611353308,-0.3391688764,-0.2146282941,0.3015082479,-0.0331984013,0.0035036115,-0.0056270952,-0.1916204989,0.1425783038,0.2735396624,0.2149775922,0.2107700258,0.2240692079,-0.456635505,0.0504483134,0.0736033544,-0.1799944192,-0.1384661645,0.1102711707,-0.2431517243,-0.4826768339,0.2714870274,-0.0576591939,0.0205870867,0.1860450506,-0.060605552,-0.2786284089,-0.0278052874,-0.0795296803,0.0359196812,-0.2037046254,0.1678601503,0.3199970424,-0.0206683092,-0.0412133746,0.1360388696,0.1009020656,0.0320518725,0.0148577234,-0.0274772588,0.4518759847,-0.0738083795,0.2224618644,-0.0367959179,-0.4124333858,-0.1352995038,0.3876947463,0.1421967447,-0.096272938,-0.0590629056,0.3712679148,0.3757946193,0.208806932,-0.1190598682,-0.1204330474,-0.018557284,0.1902958155,-0.2560938895,-0.401103884,-0.199561879,0.2541176975,-0.159465,-0.2629769146,0.291434437,0.3229745626,0.3300278783,-0.0626238286,-0.1192412302,-0.0411781333,-0.3235078454,-0.2578913867,0.060818743,-0.1730234027,0.2401383072,0.0722115263,0.1969630718,0.1020618752,0.0918280408,-0.0117396712,0.4295480549,0.1783465147,0.081121549,0.2869003415,0.2331811041,-0.0005492548,-0.3484930694,0.0509897545,0.0939048231,0.0469579957,-0.2442384958,-0.1325813681,-0.4801343381,0.0466335267,-0.2232690752,-0.1018253714,-0.0979822874,-0.4937857091,0.0546128787,-0.2819820344,-0.3108118176,0.2638486326,0.0463599525,0.1353929788,-0.1387054026,0.5016967058,-0.1144294441,-0.0443309806,-0.1221578866,0.0753727555,-0.2566229701,0.0978889614,0.099580057,-0.3706761599,0.1072077081,-0.4202881753,-0.3816201389,0.389434129,-0.087020494,-0.3375490904,0.1074736118,-0.0183748361,-0.1846232414,-0.3452057838,0.3258253038,-0.3925983906,-0.231434375,-0.2954516709,0.1398272216,-0.0754718781,-0.2425052077,0.0917229876,-0.1082660034,0.0532614477,0.1795574576,-0.1388073564,0.2465940714,0.1362661123,-0.3173582554,-0.1223066077,-0.1020742804,-0.2190776169,-0.3070394993,-0.2235675156,0.3470560908,-0.4757857323,-0.1583953351,0.0339936502,0.1421860456,-0.0508774333,0.262332201,0.0668139607,0.2145922035,-0.1252363175,-0.0695116371,0.3426079452,0.1241785213,0.2455888242,-0.1016205326,-0.042086266,-0.2940187156,-0.0273365024,0.1810752898,-0.0172745101,0.1592989117,-0.2762622237,-0.1641188562,-0.0342820883,0.2968429327,0.3348824382,0.2511936724,-0.2591611147,-0.0684287623,0.1136457697,-0.1110109538,0.1933900267,0.0214911923,0.1982367337,0.0541499816,0.709515214,0.0155821834,-0.1922372282,0.192288816,0.3546833694,-0.4887339175,-0.1965767294,-0.0783907697,0.2225352973,0.2406316698,-0.0249815788,-0.0021354232,-0.3048395813,0.0270717088,0.2300369442,0.1821662188,0.0763163045,0.0120338658,-0.2723428607,-0.294344604,-0.4310628772,-0.0190300513,0.05198377,0.4435116649,0.2483714223,-0.1682101488,0.2704311609,-0.0863975883,0.3242615163,-0.3602322936,0.1099799424,0.3611762226,-0.1975456476,-0.3167290688,-0.3746773899,-0.5127691627,0.0129086161,0.1461500823,0.051324293,0.0587631203,-0.3442403972,0.0346415974,-0.0439486764,-0.3073993623,-0.2792111337,0.0384161323,-0.2097461373,0.2508639097,-0.5460854769,0.417945236,0.2924569845,-0.5816605091,0.0741631389,-0.450496316,0.4885100722,0.1526669711,0.1978829056,0.2892878056,-0.0151860891,-0.0662796125,0.1886483729,0.2112591714,-0.1427602321,0.2967686653,0.2501598299,-0.3005088866,0.0077911424,-0.1984394342,-0.0146401459,0.253821522,-0.1120650321,0.0978574976,-0.5057727695,0.0219944306,-0.2583948374,0.1004826203,0.6158753037,0.0327369943,-0.394331038,-0.1818728447,0.3970836997,0.2251659632,-0.3253792524,0.0335221961,0.6692587137,-0.3617210686,-0.1420273483,0.4021150172,0.6736541986,0.0866683125,0.0203670189,0.3263881803,-0.0576104186,0.7016404271,-0.0638740137,0.0329564475,-0.3559820354,-0.2928742766,-0.0499338247,0.1108135879,0.1582686305,0.0723484606,-0.0352906026,-0.0871669948,-0.0997976735,0.2022983879,-0.0124775264,0.0346984193,0.2200429142,0.039790906,-0.0844847336,0.0972257555,0.5972161889,0.3229548633,-0.1972982585,-0.098367773,-0.026406344,-0.0381320976,-0.3968482614,-0.0018452716,0.1411920488,0.4528174996,0.062469881,-0.3829787672,0.441439271,0.2217008471,0.5327445269,-0.0396521576,-0.2861419618,0.1373756826,0.1176266447,-0.108745344,0.0060188868,-0.0833506733,-0.2097361684,-0.2212498635,-0.1080787405,-0.1236005872,0.0194364116,0.2174092233,-0.0779929608,-0.0094664181,0.0420487486,-0.0601389371,-0.086048007,0.1980539858,0.0544234253,0.0464055315,0.0527766608,-0.0268777218,-0.1355150193,0.132006824,0.0620029233,-0.3948559165,0.1578318328,0.234104231,0.2054586709,-0.1008146703,0.6108059287,-0.0993383303,-0.2503284514,-0.1607430875,0.0475602336,0.1869795918,-0.3103298247,-0.0130598089,0.1432362497,0.0154911485,0.0826172605,0.7759912014,0.0203219354,0.136984989,-0.4763255119,-0.1916448027,0.2156343013,0.1299427301,-0.2361616343,0.1474556476,-0.0238647368,0.1906870902,0.1701062918,-0.1386550814,-0.2814676464,0.1525866985,0.2324339747,0.0353736691,-0.4836800694,0.1188027933,-0.1101586819,0.1139537916,-0.0178602189,0.0740650594,-0.1856975406,-0.1450542063,-0.097356759,0.203242749,0.0808133334,-0.1675107032,0.3927460611,0.1622995883,0.1847814769,-0.1923657507,0.3247318864,0.1988266706,0.0650609508,0.3917618692,0.2971048951,0.2304317802,-0.2909979224,0.0269532781,0.1634774208,0.159767434,-0.0654395372,0.1494700462,-0.0778533295,0.032700967,0.4458997846,0.3129549623,-0.1949962378,0.5398123264,0.3058560789,-0.0615587533,0.2408313453,-0.1055331528,0.1859016418,0.4257743955,-0.352668941,-0.3963012993,0.2062830627,0.185619086,-0.1882981211,-0.2197807133,0.1489608139,0.1153729632,-0.2010199428,-0.0186850317,0.3285317421,-0.4414663613,-0.0704066977,-0.2098180354,0.1991513371,-0.1171024144,0.2530889511,0.3169166148,0.0568382032,0.0110730371,0.0221466217,-0.1400333941,0.1131103113,-0.0560883619,0.0914068371,-0.0469134115,0.1737012863,0.2082313597,0.222836718,-0.0457989834,0.0974933207,0.1776451319,0.0367021672,0.5943415761,0.3548409939,-0.0387674645,0.0125612002,0.0324754752,-0.528689146,0.2551363707,0.0001917571,-0.325430423,-0.2161678225,0.0300792791,0.006844637,-0.1277940124,-0.3067353666,-0.0481860638,0.0981359929,-0.0278131422,0.0450821333,-0.0457807295,-0.1563924402,0.2021134347,0.6080766916,-0.1188907623,0.0829355195,0.2031081766,0.2416888326,0.2060552984,0.6607326865,0.3005817831,0.1668931246,-0.0574238449,0.0485440083,-0.2012662143,-0.0535872504,0.120954141,-0.002572024,-0.0511805825,-0.2191366255,0.3901647329,0.1010953262,-0.0642973632,-0.2973686159,0.0657148063,0.1644857079,-0.0735214278,-0.0915411487,0.0755387172,0.068785347,0.1257434636,0.0490064211,-0.1212530509,-0.2315224856,-0.0675667748,0.049959559,0.1386755705,0.0661870614,0.0979144201,-0.0878077224,0.6923757792,0.2479281574,-0.1096170619,0.0893633291,-0.5694807172,-0.1026733369,0.1465600282,-0.0919954181,-0.388058722,0.2200866044,-0.0132185658,0.0270650722,0.192589879,0.2108723968,0.3054949343,-0.4455823898,0.1929500103,-0.3293270767,0.0259463694,0.0819468275,-0.0949818343,-0.0666919127,0.0757904202,0.2474085242,-0.0921689495,0.0656652674,0.0502836369,0.6037338376,-0.1314944774,-0.4238614738,0.3227267563,0.6272678971,0.2015083283,-0.1722560525,0.0755418912,-0.2394635677,0.1105547845,-0.3852342367,0.5749335885,-0.1550790966,0.1682860404,0.0808781832,-0.196016416,0.3053569496,0.2663375735,0.2701197565,-0.0778565481,-0.3772153556,0.0930842981,0.0144936582,-0.2753415406,0.2626592517,0.4421533644,0.1009861901,-0.0835012719,-0.1074565277,-0.5508131981,0.357096225,-0.4575970471,-0.1022377312,-0.1616833955,0.1738405526,0.2853312492,0.1053195149,-0.3537100255,-0.1055703312,0.2114420831,-0.1065451279,0.0298289508,0.2314236313,-0.1309427023,-0.1069141254,-0.1179708317,0.4257190824,0.2235053182,-0.2746570706,-0.3533921838,-0.2724451125]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/794","title":"self.options cannot be converted to a Python object for pickling","comments":"Hi ! Thanks for reporting that's a bug on master indeed.\r\nWe'll fix that soon","body":"Hi,\r\n\r\nCurrently I am trying to load csv file with customized read_options. And the latest master seems broken if we pass the ReadOptions object.\r\n\r\nHere is a code snippet\r\n```python\r\nfrom datasets import load_dataset\r\nfrom pyarrow.csv import ReadOptions\r\nload_dataset(\"csv\", data_files=[\"out.csv\"], read_options=ReadOptions(block_size=16*1024*1024))\r\n```\r\nerror is `self.options cannot be converted to a Python object for pickling`\r\nWould you mind to take a look? Thanks!\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-ab83fec2ded4> in <module>\r\n----> 1 load_dataset(\"csv\", data_files=[\"out.csv\"], read_options=ReadOptions(block_size=16*1024*1024))\r\n\r\n\/tmp\/datasets\/src\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 602 hash=hash,\r\n 603 features=features,\r\n--> 604 **config_kwargs,\r\n 605 )\r\n 606 \r\n\r\n\/tmp\/datasets\/src\/datasets\/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)\r\n 162 name,\r\n 163 custom_features=features,\r\n--> 164 **config_kwargs,\r\n 165 )\r\n 166 \r\n\r\n\/tmp\/datasets\/src\/datasets\/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)\r\n 281 )\r\n 282 else:\r\n--> 283 suffix = Hasher.hash(config_kwargs_to_add_to_suffix)\r\n 284 \r\n 285 if builder_config.data_files is not None:\r\n\r\n\/tmp\/datasets\/src\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/tmp\/datasets\/src\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/tmp\/datasets\/src\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/tmp\/datasets\/src\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n~\/.local\/lib\/python3.6\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in dump(self, obj)\r\n 407 if self.proto >= 4:\r\n 408 self.framer.start_framing()\r\n--> 409 self.save(obj)\r\n 410 self.write(STOP)\r\n 411 self.framer.end_framing()\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n~\/.local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in _batch_setitems(self, items)\r\n 850 k, v = tmp[0]\r\n 851 save(k)\r\n--> 852 save(v)\r\n 853 write(SETITEM)\r\n 854 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 494 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 495 if reduce is not None:\r\n--> 496 rv = reduce(self.proto)\r\n 497 else:\r\n 498 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\n~\/.local\/lib\/python3.6\/site-packages\/pyarrow\/_csv.cpython-36m-x86_64-linux-gnu.so in pyarrow._csv.ReadOptions.__reduce_cython__()\r\n\r\nTypeError: self.options cannot be converted to a Python object for pickling\r\n```","comment_length":15,"text":"self.options cannot be converted to a Python object for pickling\nHi,\r\n\r\nCurrently I am trying to load csv file with customized read_options. And the latest master seems broken if we pass the ReadOptions object.\r\n\r\nHere is a code snippet\r\n```python\r\nfrom datasets import load_dataset\r\nfrom pyarrow.csv import ReadOptions\r\nload_dataset(\"csv\", data_files=[\"out.csv\"], read_options=ReadOptions(block_size=16*1024*1024))\r\n```\r\nerror is `self.options cannot be converted to a Python object for pickling`\r\nWould you mind to take a look? Thanks!\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-ab83fec2ded4> in <module>\r\n----> 1 load_dataset(\"csv\", data_files=[\"out.csv\"], read_options=ReadOptions(block_size=16*1024*1024))\r\n\r\n\/tmp\/datasets\/src\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 602 hash=hash,\r\n 603 features=features,\r\n--> 604 **config_kwargs,\r\n 605 )\r\n 606 \r\n\r\n\/tmp\/datasets\/src\/datasets\/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)\r\n 162 name,\r\n 163 custom_features=features,\r\n--> 164 **config_kwargs,\r\n 165 )\r\n 166 \r\n\r\n\/tmp\/datasets\/src\/datasets\/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)\r\n 281 )\r\n 282 else:\r\n--> 283 suffix = Hasher.hash(config_kwargs_to_add_to_suffix)\r\n 284 \r\n 285 if builder_config.data_files is not None:\r\n\r\n\/tmp\/datasets\/src\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/tmp\/datasets\/src\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/tmp\/datasets\/src\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/tmp\/datasets\/src\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n~\/.local\/lib\/python3.6\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in dump(self, obj)\r\n 407 if self.proto >= 4:\r\n 408 self.framer.start_framing()\r\n--> 409 self.save(obj)\r\n 410 self.write(STOP)\r\n 411 self.framer.end_framing()\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n~\/.local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in _batch_setitems(self, items)\r\n 850 k, v = tmp[0]\r\n 851 save(k)\r\n--> 852 save(v)\r\n 853 write(SETITEM)\r\n 854 # else tmp is empty, and we're done\r\n\r\n\/usr\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 494 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 495 if reduce is not None:\r\n--> 496 rv = reduce(self.proto)\r\n 497 else:\r\n 498 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\n~\/.local\/lib\/python3.6\/site-packages\/pyarrow\/_csv.cpython-36m-x86_64-linux-gnu.so in pyarrow._csv.ReadOptions.__reduce_cython__()\r\n\r\nTypeError: self.options cannot be converted to a Python object for pickling\r\n```\nHi ! Thanks for reporting that's a bug on master indeed.\r\nWe'll fix that soon","embeddings":[-0.1403232962,-0.2921461761,-0.0467801094,0.2873256505,0.2217250913,0.0459425151,0.1823574007,0.1356180608,0.380415231,0.2151599377,-0.3029938042,0.5059405565,-0.2015548497,0.1517243981,-0.0564416796,0.1482837796,-0.0205383543,0.1317485124,-0.3516437113,0.0892095715,-0.3891227245,0.0643444359,-0.1519772261,-0.1157856435,0.4620776772,0.0435499437,-0.1418200582,0.2592187822,-0.3916413188,-0.2674822509,0.1634338051,0.0020817125,-0.0158527344,0.5413383245,-0.0001268832,0.2973466814,0.1171191931,-0.2019997984,-0.4036060572,-0.4351240695,-0.0985570177,-0.2100194842,0.2869416773,-0.3753811121,0.1487613767,0.07623671,-0.0614841692,-0.0429830663,0.3782578111,0.393106252,0.1152820885,0.468098551,0.2499979138,-0.0597077757,0.3551157415,0.2313318551,-0.325042516,0.3276945055,0.1314764768,0.2220353484,0.0228245817,-0.1151580289,-0.2867771089,-0.0280095339,0.354917109,-0.0732259527,0.549372375,-0.0124467136,-0.1513556242,-0.1726771295,0.3438975215,-0.2142896801,-0.3042000532,-0.4991585612,0.0642954409,-0.2003352642,-0.0002789274,0.2812205851,-0.1512132734,0.2247692347,-0.1278672814,-0.2004902363,-0.2886084616,0.3308737874,0.0999206826,0.1578830481,-0.2153031379,0.1038201228,0.6068193913,-0.0363067649,0.3987533748,-0.0768780336,0.0969080627,-0.1331616342,-0.184429571,-0.1289448589,0.136929065,0.2637735307,0.0068291323,0.0780360773,0.20820494,-0.0412264839,0.0157355871,0.3209885955,0.4501567483,-0.0324495845,-0.0773923323,0.3748054504,-0.1822792739,0.1989266276,-0.1268820614,0.1121694073,-0.0127738668,-0.2905876637,-0.0195031762,0.1632480025,0.4059026539,-0.1769283861,-0.0812058896,-0.1245692894,-0.2991523743,-0.0586395115,0.1847747862,0.204525888,-0.3079758883,0.1132094488,0.1419731081,0.1423262358,-0.2071622014,-0.0708041638,-0.0194415767,0.1855652034,-0.1923321784,-0.0261504874,0.0566978455,-0.1108380631,0.150934279,0.1469876468,0.3401753306,-0.078221783,-0.0282949638,0.002263353,-0.1086893454,0.3448125124,0.0962360278,-0.1620787233,0.4541670978,-0.410831809,-0.2694441974,0.3895312846,-0.3179284036,-0.4788861275,-0.0373666883,0.0666102692,-0.2896010578,-0.0734687075,-0.3707227409,-0.1446176618,0.283295691,-0.1576367319,-0.0271708611,-0.527002275,0.0821114779,-0.3469915688,-0.0964786336,0.4230982363,-0.5324553251,-0.0283388924,-0.4654639661,-0.1476585418,0.1756512374,0.2236339748,-0.1544360816,-0.0434624255,-0.135651499,0.0262871683,0.1260533482,0.0045371098,-0.5446788669,0.2412202507,0.1035120264,0.0943921655,0.1984214932,0.0039723082,0.1808269471,-0.0223537292,0.1013056785,-0.0582606383,-0.0329425186,-0.1281404048,-0.0877419412,-0.2728547752,0.1795296967,0.095884636,-0.0569376536,0.2374151945,0.0127819739,-0.3489665389,0.1784833372,-0.2371652573,-0.0816265419,-0.0116732325,0.4998986423,0.3710561097,-0.1615672857,-0.2299890071,-0.29878214,0.0260611381,-0.0877176374,0.0040484481,-0.2226441801,-0.2757273018,-0.4293171763,0.2370640635,-0.14393498,0.2768559158,-0.0496272221,-0.0367647633,-0.3075680733,-0.2785373032,-0.0438070782,-0.0240632463,-0.4844091833,0.1769717038,-0.3434560895,0.1361306459,-0.126248166,-0.3152449131,0.0921068788,-0.0174325854,0.3951207995,-0.0854850635,-0.173475042,0.1607419401,0.1858005971,-0.1829994172,-0.4633665979,0.0537890606,0.0165376179,-0.0540396832,-0.0187198166,0.3865171373,0.1647677422,-0.1947767884,-0.4074974954,0.7201971412,-0.0754367113,0.2049194276,-0.0390443355,0.1780920923,-0.115152441,-0.0955008641,-0.0928062499,-0.2387395352,-0.0338814557,-0.0025257545,0.2434200943,0.2903364599,-0.3657972813,0.0967688337,0.4356248975,-0.0268993601,0.0458077937,0.3836815357,0.2033196688,0.0104284426,0.1009548232,-0.1325433254,0.2712198794,0.0269846451,-0.2470352054,0.1836658865,0.0419702567,-0.11051213,0.1859935522,-0.0646611452,0.3500154316,0.1089130715,-0.0072601554,-0.0823675469,-0.1555547714,0.0342395604,-0.0852991268,0.2701582015,-0.4575350583,-0.0381181873,-0.6042368412,0.1981678754,-0.0064929905,-0.111160934,0.039913848,-0.2437177747,-0.2502857745,0.7373925447,0.0520901904,0.0106200445,-0.2468041927,0.0209891889,0.1103015095,-0.2601396441,0.0727273747,-0.0769404322,-0.1483912617,-0.0533351749,0.4077910781,-0.3798011839,0.1699371338,0.3562939167,-0.2902002931,-0.0750094205,-0.0585024729,-0.075809136,-0.0513725169,0.4167123735,0.1559699178,0.1609045863,-0.0150280762,-0.0718843862,0.1736012101,0.3256579936,-0.0626796782,0.0496864393,-0.0082239136,-0.1608755738,-0.1188521609,-0.2092637271,-0.5312121511,-0.2071380913,0.2949053049,0.3157903552,0.1492503732,0.3000403345,0.4598853886,-0.0066648633,0.5319833755,-0.1212378368,-0.0442953669,-0.2850902677,0.421581924,-0.1854872704,-0.1398955733,-0.2216750085,0.1336575896,0.1232367456,0.0548660196,-0.458710283,-0.0139286285,-0.1405092925,0.2392766923,-0.2014530301,-0.0975550264,0.3397494555,0.2523632944,0.1016439497,-0.1011289209,-0.5078522563,-0.028156925,0.3000195324,0.1851739734,0.2082374692,0.2476593256,-0.0282058641,0.9208983779,-0.0583232567,0.0600039251,0.3136671782,-0.1577302963,0.0182341076,-0.063078545,-0.6761226654,-0.1354931742,-0.4294220805,-0.2412139922,0.00878252,-0.2116856724,-0.2122322917,-0.0444799885,-0.0478492416,-0.4161897898,-0.3223855495,0.1162997782,0.0957282484,0.1269042194,-0.1484262198,0.1263941377,-0.3233261406,-0.025581073,-0.0057591032,0.3014504611,0.225794211,-0.0972031578,-0.1361616254,-0.1614249945,0.2291592956,0.2387820035,-0.2420640588,0.1241814569,-0.029188022,0.1448299885,0.1322267354,-0.3131263256,0.6434317231,0.2779018581,0.1437000334,0.3891632557,-0.1371430457,-0.1485996544,-0.1650471538,-0.0195952337,0.5672869682,0.1455990374,0.267234385,-0.3615693152,-0.0271537397,0.2386968285,0.1969642192,-0.2361561507,-0.3092321157,-0.0124453492,-0.1784662157,-0.2769854963,0.0338649265,-0.061799258,0.1688997447,-0.2057445347,0.1151597425,-0.0848394334,-0.1369769275,0.3504689336,-0.0994277969,-0.1277621388,-0.0156317428,0.0453689098,-0.2773779929,0.0899235234,0.1809950173,0.7667171955,0.2956429422,-0.5679453015,-0.0701803938,0.0876726657,0.1976390034,0.1271583289,-0.0866677612,0.3688420951,0.2420348674,0.0048795962,0.2013764083,-0.1857448369,0.1051647663,-0.1985411048,-0.2558301389,-0.6757668257,0.6392724514,-0.1064045429,0.0310960338,0.166144684,0.2497899383,-0.0734207705,0.211444512,0.1999350339,1.0999052525,0.0792013407,0.0475049168,0.4570145309,-0.0982623324,0.0898875669,0.0361090824,0.0191743746,-0.4776610732,0.3264194131,-0.0615015775,-0.1392160505,0.4623233378,0.2080128193,-0.3344580233,0.019821655,-0.2506667376,0.0257412679,0.2320177108,-0.0880761743,-0.1678493023,-0.1920723617,-0.1912151873,0.0592902265,-0.1186949611,0.0443567485,0.1802477688,-0.1773608029,-0.07596118,0.0990659893,-0.3381187022,0.0368413888,-0.2570302188,0.2578777671,0.0349162892,-0.5870956779,0.6370249987,0.1324518472,-0.111379832,0.0796734542,-0.110530816,0.1400686651,-0.0610908791,-0.0900501534,-0.1417550594,-0.0304920916,0.2746500373,0.013836165,-0.4408680797,0.1968343407,0.0097983461,-0.462321192,-0.0467087328,0.0124052446,0.2069481164,-0.6034260392,-0.7554270029,-0.2798535228,0.0121662207,-0.1812632978,-0.0262234844,-0.0960913524,0.0895833522,0.2296590805,-0.1381182969,-0.1673850417,-0.0615621209,0.4501515329,0.1901147962,0.2384325266,0.5417335629,0.2448257655,-0.1001778096,-0.1580555141,0.2274994701,-0.4632049501,0.0083735818,0.3463308513,0.0054029208,0.2151445746,-0.129687801,0.0877702683,0.2980994582,0.0886145905,0.0100820409,-0.4404700994,-0.3490543365,-0.1371877789,0.1166813076,0.0332587771,0.0872122124,0.2731642425,-0.2003501654,0.1008615717,-0.1760636419,0.1046683043,-0.0256009083,0.0095371073,0.3400537968,0.0527973175,-0.4185222089,-0.1103717759,0.098297514,0.0790472999,0.031528689,-0.1203828529,-0.3677987158,0.1830463856,0.1528980434,0.1838090122,-0.0134256408,-0.2521539032,-0.1117928252,0.0659481958,-0.1034400836,0.2932052314,0.0764456242,0.3824566901,-0.0900858268,-0.0252230205,-0.1875738502,0.1897868067,-0.1555317342,0.1305842698,0.0460529737,0.1299112439,0.0092245871,0.0176482946,-0.3067448437,-0.085755676,0.0376608297,0.1388911307,0.5037251711,-0.137460053,0.1802222431,-0.2621749938,0.1588571668,0.3484945893,0.0299550574,0.1921784729,0.393868804,0.0090430686,-0.3098551631,0.1624947041,0.5765071511,-0.0910829157,0.0047536851,0.212090835,-0.1875551194,0.3533939123,-0.0597220622,0.0240116939,0.3657215536,0.2323854268,0.2779027522,0.4846344888,-0.0617348291,-0.0849440843,0.3900028467,0.1062316746,0.2844568193,0.4103330076,-0.2074468732,0.0886121541,-0.06528458,0.2622410059,0.152435109,-0.556197226,0.2674730122,0.0913131237,0.0684965625,0.3092001975,0.2027073801,0.4740094244,-0.4187022746,-0.0951303244,-0.1333922148,0.275503546,0.0405147485,-0.1237839833,-0.47387591,-0.2722549438,-0.2307634503,0.2269040495,-0.1256423146,-0.2556895614,0.1766289026,-0.0009519383,-0.0243386906,-0.4201588333,-0.046983894,0.0118778888,0.3110469878,-0.2171035856,-0.0809936374,-0.0076020304,0.1574141383,-0.2234152704,-0.1060443223,0.3947450817,0.3337611258,-0.3022177815,-0.1372738779,0.0427636392,-0.2144188732,0.1043699905,-0.1024690196,-0.0896472707,0.2614206076,0.1665321738,0.0634532496,-0.1728792191,0.3626419604,0.1425715834,-0.102348648,-0.5338348746,0.0254684202,0.2249034196,-0.3996298015,0.2391256988,-0.184943527,-0.2334575057,-0.2561180592,0.5679363012,0.3157876432,0.180770278,0.1301283985,0.034629032,0.1815340668,0.4306892157,0.2542312741,0.117229715,-0.2939957976,0.2649128139,-0.5290288925,-0.0423475355,-0.0448289104,-0.2420785874,0.078393355,-0.0058252877,0.2037858069,0.2901687026,-0.2694903612,-0.1146580875,0.4101172388,0.483368814,-0.0096787028,-0.207579121,-0.0436376184,0.0125791552,0.0700064301,-0.3191119134,-0.0225299597,0.1407341361,0.0311577134,-0.3804076016,0.0263317507,0.2624572814,-0.0698070675,0.3701711893,-0.0534555465,0.6384229064,-0.0913228691,0.0913003534,0.2077893466,0.0305033661,-0.1754881144,0.0555888563,0.0107528251,0.5786215663,-0.0796466321,-0.0193635114,-0.3208862245,0.4343104064,0.0038745257,-0.2540392578,-0.0166071597,0.2931112349,-0.0950205997,-0.0592393726,-0.1440723836,0.1043164805,-0.2025951445,0.3448414505,-0.3339146376,-0.0043052607,0.7086890936,-0.3476385474,-0.1421352029,0.2020212263,-0.1209824756,-0.2059451938,-0.0698617995,-0.2821038663,-0.0640388504,0.4296365678,-0.2452226132,-0.0145931132,0.2832253873,0.069512032,-0.0061394344,-0.1478779018,0.5812503099,0.1736564785,-0.2334060967,0.1215844601,-0.1397002786]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/792","title":"KILT dataset: empty string in triviaqa input field","comments":"Just found out about https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md\r\n(Not very clear in https:\/\/huggingface.co\/datasets\/kilt_tasks links to http:\/\/github.com\/huggingface\/datasets\/datasets\/kilt_tasks\/README.md which is dead, closing the issue though :))","body":"# What happened\r\nBoth train and test splits of the triviaqa dataset (part of the KILT benchmark) seem to have empty string in their input field (unlike the natural questions dataset, part of the same benchmark)\r\n\r\n# Versions\r\nKILT version is `1.0.0`\r\n`datasets` version is `1.1.2`\r\n[more here](https:\/\/gist.github.com\/PaulLerner\/3768c8d25f723edbac20d99b6a4056c1)\r\n\r\n# How to reproduce\r\n```py\r\nIn [1]: from datasets import load_dataset\r\nIn [4]: dataset = load_dataset(\"kilt_tasks\") \r\n# everything works fine, removed output for a better readibility\r\nDataset kilt_tasks downloaded and prepared to \/people\/lerner\/.cache\/huggingface\/datasets\/kilt_tasks\/all_tasks\/1.0.0\/821c4295a2c35db2847585918d9c47d7f028f1a26b78825d8e77cd3aeb2621a1. Subsequent calls will reuse this data.\r\n\r\n# empty string in triviaqa input field\r\nIn [36]: dataset['train_triviaqa'][0] \r\nOut[36]: \r\n{'id': 'dpql_5197',\r\n 'input': '',\r\n 'meta': {'left_context': '',\r\n 'mention': '',\r\n 'obj_surface': {'text': []},\r\n 'partial_evidence': {'end_paragraph_id': [],\r\n 'meta': [],\r\n 'section': [],\r\n 'start_paragraph_id': [],\r\n 'title': [],\r\n 'wikipedia_id': []},\r\n 'right_context': '',\r\n 'sub_surface': {'text': []},\r\n 'subj_aliases': {'text': []},\r\n 'template_questions': {'text': []}},\r\n 'output': {'answer': ['five \u00a3', '5 \u00a3', '\u00a35', 'five \u00a3'],\r\n 'meta': [],\r\n 'provenance': [{'bleu_score': [1.0],\r\n 'end_character': [248],\r\n 'end_paragraph_id': [30],\r\n 'meta': [],\r\n 'section': ['Section::::Question of legal tender.\\n'],\r\n 'start_character': [246],\r\n 'start_paragraph_id': [30],\r\n 'title': ['Banknotes of the pound sterling'],\r\n 'wikipedia_id': ['270680']}]}}\r\nIn [35]: dataset['train_triviaqa']['input'][:10] \r\nOut[35]: ['', '', '', '', '', '', '', '', '', '']\r\n# same with test set \r\nIn [37]: dataset['test_triviaqa']['input'][:10] \r\nOut[37]: ['', '', '', '', '', '', '', '', '', '']\r\n# works fine with natural questions\r\nIn [34]: dataset['train_nq']['input'][:10] \r\nOut[34]: \r\n['how i.met your mother who is the mother',\r\n 'who had the most wins in the nfl',\r\n 'who played mantis guardians of the galaxy 2',\r\n 'what channel is the premier league on in france',\r\n \"god's not dead a light in the darkness release date\",\r\n 'who is the current president of un general assembly',\r\n 'when do the eclipse supposed to take place',\r\n 'what is the name of the sea surrounding dubai',\r\n 'who holds the nba record for most points in a career',\r\n 'when did the new maze runner movie come out']\r\n```\r\n\r\nStay safe :)","comment_length":21,"text":"KILT dataset: empty string in triviaqa input field\n# What happened\r\nBoth train and test splits of the triviaqa dataset (part of the KILT benchmark) seem to have empty string in their input field (unlike the natural questions dataset, part of the same benchmark)\r\n\r\n# Versions\r\nKILT version is `1.0.0`\r\n`datasets` version is `1.1.2`\r\n[more here](https:\/\/gist.github.com\/PaulLerner\/3768c8d25f723edbac20d99b6a4056c1)\r\n\r\n# How to reproduce\r\n```py\r\nIn [1]: from datasets import load_dataset\r\nIn [4]: dataset = load_dataset(\"kilt_tasks\") \r\n# everything works fine, removed output for a better readibility\r\nDataset kilt_tasks downloaded and prepared to \/people\/lerner\/.cache\/huggingface\/datasets\/kilt_tasks\/all_tasks\/1.0.0\/821c4295a2c35db2847585918d9c47d7f028f1a26b78825d8e77cd3aeb2621a1. Subsequent calls will reuse this data.\r\n\r\n# empty string in triviaqa input field\r\nIn [36]: dataset['train_triviaqa'][0] \r\nOut[36]: \r\n{'id': 'dpql_5197',\r\n 'input': '',\r\n 'meta': {'left_context': '',\r\n 'mention': '',\r\n 'obj_surface': {'text': []},\r\n 'partial_evidence': {'end_paragraph_id': [],\r\n 'meta': [],\r\n 'section': [],\r\n 'start_paragraph_id': [],\r\n 'title': [],\r\n 'wikipedia_id': []},\r\n 'right_context': '',\r\n 'sub_surface': {'text': []},\r\n 'subj_aliases': {'text': []},\r\n 'template_questions': {'text': []}},\r\n 'output': {'answer': ['five \u00a3', '5 \u00a3', '\u00a35', 'five \u00a3'],\r\n 'meta': [],\r\n 'provenance': [{'bleu_score': [1.0],\r\n 'end_character': [248],\r\n 'end_paragraph_id': [30],\r\n 'meta': [],\r\n 'section': ['Section::::Question of legal tender.\\n'],\r\n 'start_character': [246],\r\n 'start_paragraph_id': [30],\r\n 'title': ['Banknotes of the pound sterling'],\r\n 'wikipedia_id': ['270680']}]}}\r\nIn [35]: dataset['train_triviaqa']['input'][:10] \r\nOut[35]: ['', '', '', '', '', '', '', '', '', '']\r\n# same with test set \r\nIn [37]: dataset['test_triviaqa']['input'][:10] \r\nOut[37]: ['', '', '', '', '', '', '', '', '', '']\r\n# works fine with natural questions\r\nIn [34]: dataset['train_nq']['input'][:10] \r\nOut[34]: \r\n['how i.met your mother who is the mother',\r\n 'who had the most wins in the nfl',\r\n 'who played mantis guardians of the galaxy 2',\r\n 'what channel is the premier league on in france',\r\n \"god's not dead a light in the darkness release date\",\r\n 'who is the current president of un general assembly',\r\n 'when do the eclipse supposed to take place',\r\n 'what is the name of the sea surrounding dubai',\r\n 'who holds the nba record for most points in a career',\r\n 'when did the new maze runner movie come out']\r\n```\r\n\r\nStay safe :)\nJust found out about https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/kilt_tasks\/README.md\r\n(Not very clear in https:\/\/huggingface.co\/datasets\/kilt_tasks links to http:\/\/github.com\/huggingface\/datasets\/datasets\/kilt_tasks\/README.md which is dead, closing the issue though :))","embeddings":[0.4154247642,-0.3093399704,-0.0309624653,0.0308451131,0.2169784904,0.0481552258,0.6097394824,0.4196387529,0.1841181815,0.1952489913,0.3654612899,0.5179553032,-0.0468075871,0.443081677,0.0452616364,0.0297371373,0.0262947138,0.0816094503,0.2449792027,-0.1349078119,-0.070260793,0.1185156778,-0.5049002171,-0.3920861781,-0.6020743251,0.0715131238,-0.1471185088,0.1056134403,0.1308469176,-0.3451221585,0.2047567815,-0.1966237426,-0.2979108095,0.234799698,-0.0001233155,-0.151816383,0.4405516088,-0.1911161393,-0.5569390655,-0.2748937905,-0.5155441761,0.0414867885,0.1126812473,-0.3192623258,-0.0520286188,0.3783489466,-0.0556893386,-0.5345303416,0.1037189364,0.3284580708,0.0828188136,0.2482213229,-0.427097708,-0.0155318491,0.0678356588,0.2325283736,-0.0963372067,-0.1887251586,-0.0307343956,0.0090678968,-0.0292269066,0.3749616146,0.0552502051,0.0512047894,-0.0871716738,0.1463882774,-0.1052402556,-0.5381202698,0.4375944436,0.4902975857,0.0022711277,-0.0752102211,-0.3428653777,-0.3614497185,0.074659653,0.0352610387,0.1127898246,0.4587815702,-0.1910824776,-0.128039673,-0.0515876785,0.0197217278,-0.0616242215,-0.0467875451,-0.557240665,0.3227185309,0.0275498293,-0.0352287889,-0.118647553,-0.079226315,0.0615168065,-0.3293149769,-0.2264239937,0.1363897473,-0.4327518642,-0.0133391628,0.0735009536,0.0046049915,0.0714485347,-0.0606742166,0.0952832028,0.1380965114,0.0809555203,0.0362243317,0.3264836371,0.2131464481,0.606731236,-0.2336505055,-0.1019396335,-0.0146458298,-0.289934814,0.0237061009,0.015417112,0.2261016667,-0.1452406347,-0.1047834158,0.5170952082,-0.1638511717,-0.432489872,0.1947111189,-0.3115721047,0.0000848518,-0.251018554,0.1180457547,-0.2242430896,0.4553079605,-0.2690626383,0.3509604037,-0.1322484165,-0.3637267947,-0.2435650229,0.0825486705,0.0769863799,0.0719891563,0.1621539891,-0.3181550205,0.247709319,0.1627962738,-0.0797332376,-0.1926609129,0.3062608838,-0.2796264589,0.164816618,0.2342649996,0.1821506172,0.1010709777,0.2721576095,-0.1779694855,0.1281515658,0.4900316596,-0.2172444761,-0.159130156,-0.1640295833,0.0902906805,-0.5352227688,-0.1692023426,0.0077250013,0.3015995026,-0.0443646833,-0.0157206971,-0.0389804244,-0.0784888789,-0.0181758441,-0.1367528588,0.3983205855,0.4051298201,-0.6290889382,-0.0194267798,-0.1245946959,0.4337553084,0.0284347795,-0.0846947059,0.002301543,0.4501265883,0.0450163893,0.4883404076,0.1594999433,-0.6424002051,0.1234371215,0.1109090075,0.0765333474,0.2729054689,0.0170958098,-0.152042836,0.2271609455,0.3121139407,0.1535473913,0.0757146105,-0.1106448695,-0.0357699879,-0.4687983692,-0.2415012717,0.3733985126,0.1570773721,0.0501448363,-0.1175259948,-0.0163034443,0.3112974465,0.1443440318,0.0159370303,0.1407600641,0.2728122473,0.4098913372,-0.1147982776,-0.0665833578,0.051839523,-0.4112521112,0.1828290373,0.02530648,0.4265533388,0.0540686511,-0.2753804326,-0.3356623948,-0.0422768667,-0.366150707,-0.5387892127,0.0638962165,-0.0977963582,0.0266561024,0.2791167796,-0.2297948003,0.0499987155,-0.068345502,0.0233631264,-0.4188793898,0.1438494474,0.1057638526,-0.0649687797,-0.3397308886,0.447250247,0.0299084373,-0.0277109426,0.0762488991,0.2349160463,0.3358787596,0.0155418394,0.3939433992,-0.1505122185,0.2231674641,-0.1009625569,-0.0389687121,0.2608843148,0.2673651278,-0.2981997728,-0.5522263646,0.4946863353,0.0474158712,0.1908845305,-0.1122226864,0.0049033449,-0.2117701322,-0.17569682,-0.1037322879,-0.0738084316,-0.0485363267,-0.4181039631,0.0915501639,0.0548916347,-0.2473305315,0.2331138253,0.3761444986,0.1827236712,0.1425498724,-0.1430834234,-0.273860991,-0.1377910078,0.1052863598,0.190239951,0.2923644483,0.1332271397,0.0850077569,-0.198642984,0.2707414925,-0.1987440884,0.2556269765,0.2737278938,0.0222738199,0.3003895581,0.2671950459,0.0109333806,-0.0619636998,0.2227361798,0.2534605563,0.1360461414,-0.1222360283,-0.0501957685,-0.2849498987,-0.1190199703,-0.0362408496,-0.0462100692,-0.114528209,-0.5389600396,-0.0615606979,-0.1473469436,-0.0517592132,0.0327147618,-0.1225998476,0.2000129819,0.1589460969,-0.4246274829,-0.2907564938,-0.4309128523,-0.2542892694,0.0144867217,0.0569267049,0.0548986532,-0.0078120911,0.0646720976,-0.1538101584,-0.1130127236,-0.5093662143,-0.0147382505,-0.3350986242,0.4982024729,0.0131052872,0.496679455,-0.3786387444,-0.0884267762,0.3220568001,0.0506732464,-0.2215313762,0.1099406332,0.1743040979,0.1307162195,-0.4416171908,-0.7334843874,-0.0226352513,-0.096458964,0.0566737242,0.0583065376,-0.1101087928,0.402231127,-0.0306141581,0.3108605444,0.0632256418,-0.3287021816,-0.1513582766,-0.1570688188,0.3093299866,-0.0767925158,-0.3370245099,0.2594738305,-0.2156599313,0.4304480553,-0.1713768244,-0.4541926086,-0.1187740266,-0.0140375998,0.4953697622,0.1077462509,-0.2205437124,0.1702052504,-0.0410149284,0.0568225347,-0.1034188047,0.0223851092,0.2761441171,0.2957688272,0.2035993785,-0.0116870571,0.125397265,0.2966403961,0.4599235654,0.0736202821,0.3123475909,0.1346438378,-0.3639746606,0.278357178,-0.2543643415,-0.0631811023,0.2531127036,0.0127647556,0.0002652777,0.3142951429,-0.0756889656,0.3516102433,-0.1441332251,0.2436687946,-0.4402600229,-0.2558328807,0.0311886612,-0.3067443669,0.1770628393,0.1730919629,0.3619164824,-0.0012299025,-0.2652309239,-0.0778275505,-0.2306248248,0.1061030924,-0.095485799,-0.2690649033,0.0621837862,-0.2401293069,0.1976777613,0.2481335402,0.2003753185,0.0504098721,-0.3557696939,0.1461398005,-0.0704362616,0.7759654522,-0.0856233016,-0.1360178888,0.1071123481,0.0604435205,-0.4118791521,-0.0626499578,-0.2942467034,-0.1385632604,0.2875736356,0.2528029978,-0.2654583454,-0.0095642461,0.1142154634,0.2417783141,-0.1778235584,0.1597910672,-0.3231333494,-0.2890191972,0.0330559239,-0.1219754666,-0.0633419901,0.5701518655,-0.1483330131,0.0315103456,-0.209376514,0.0593042783,0.3717397153,0.1712847203,0.5655801296,-0.1246213317,0.2135389149,-0.0758487582,-0.3037151992,-0.2119699121,0.3906045854,0.0980002359,0.0490783527,0.0780460611,-0.3931384981,0.4202157557,0.4430984259,0.1791418195,0.3714133799,-0.2220454067,-0.0331352428,-0.1860253364,0.1904967576,0.0833218098,-0.1519438177,-0.0232092272,-0.2001502216,0.2331856638,-0.1087859496,0.1211943999,0.1551268846,0.1107929721,-0.0085573532,0.2784830332,0.1261936724,0.8262000084,0.1877862215,0.3258031011,0.494194001,0.2213577777,0.2021895498,-0.2013834268,0.0570372567,-0.4586983025,-0.0392895415,-0.0270802192,0.0808954313,0.2617099583,0.0104308585,-0.2232201099,0.1786382347,-0.0291297194,0.3884908557,0.0265641082,0.0543493219,-0.1113885194,0.0684937909,-0.001660899,0.0688228533,-0.0286207329,0.1248941645,0.081386283,-0.1425544173,-0.2323156446,-0.052559454,-0.0902594924,0.2771009505,-0.1811248362,-0.0206347723,0.4377576411,-0.2984696329,-0.1590685397,0.6575038433,0.1915818304,0.1558430493,-0.1707692444,0.3402536213,0.142495811,0.0556431413,0.3146828115,0.1460884511,0.15738222,0.0270131715,-0.3823417127,0.3358509243,-0.373190999,0.230252102,-0.47028175,0.2857643962,0.1192923412,-0.2572232783,-0.5466195345,-0.0484772362,0.1750349551,-0.1081222221,0.060101863,-0.2077105194,-0.0742991343,0.1068056226,-0.0123084215,-0.3217718601,0.0258914139,0.155945614,-0.0144759566,0.1936582625,0.6203402877,-0.0312073547,0.0547928289,-0.2772302628,0.3882187605,-0.1792588532,-0.0877210125,0.1648567915,0.1916460693,-0.0352723673,0.0381989107,0.4818542004,-0.122028634,-0.0431361012,-0.1034762412,-0.2885365486,0.1736021936,0.2671673298,-0.0728942826,0.1629694998,0.128193751,-0.0150424251,0.1816092432,0.0585280545,-0.1931365877,0.1472180784,-0.2093851268,0.3508725762,-0.1339951605,-0.0659423918,0.5642385483,-0.1818608344,0.0584103018,0.0066463146,0.1339650899,-0.197018981,-0.2558175623,0.158515051,0.1788584739,0.2830893397,-0.0859070048,0.0186234917,-0.0393614806,-0.1434950829,0.1618809849,0.0761126429,-0.0985507295,0.3253028393,0.2054098845,0.4561514854,-0.4297527075,-0.0462753251,0.0225315168,-0.0728754699,0.1214918047,0.0392028615,0.0421901755,0.1383526325,0.0407990664,0.0577998273,-0.1012860388,-0.3686122298,0.0466551855,-0.4157793522,0.0677360892,-0.1016999185,-0.0981459245,0.2605660856,-0.3660267889,0.1143000126,0.0204199106,0.1428683847,-0.1888909936,0.32215482,-0.1179342046,0.1958060116,0.0714628324,0.2107002586,0.1286302507,-0.1363971233,0.2871667147,0.1175800934,0.0364160389,-0.6664677858,0.3755865097,0.3628862202,-0.0350854844,0.0309827272,-0.0771119669,-0.2520345151,-0.3919234276,0.4506194293,0.0629536882,0.482809633,0.1476824135,-0.2308730185,-0.1054599509,-0.0070993598,0.0878536105,0.1100846529,-0.2691179514,0.01393524,0.1670454443,0.5890690088,0.0149973314,-0.3044736385,-0.0055160681,-0.2003384978,-0.2105042487,-0.1267281473,-0.0836274773,-0.1083182022,-0.0754744038,-0.0436581969,0.2366993725,-0.1012167186,0.1387130916,0.2177638113,-0.0327056572,-0.2060092688,-0.1638248116,0.3368455172,0.3360430598,-0.0687319934,0.3295077682,0.1154530942,0.0655894503,0.345582217,0.2013691664,0.2794666886,0.1423922926,-0.4256040454,0.0391972922,-0.0918238834,-0.2768032253,0.2572437823,0.1200449392,0.1936910897,0.220084846,0.4399842918,0.0890710205,-0.073740609,-0.0668838918,0.3249744177,-0.13887465,-0.3395922184,0.3811064065,0.0413788445,-0.005341453,-0.067585066,-0.0224708281,-0.4967587888,-0.0474369787,-0.0220717993,0.3517432511,0.0326015204,-0.0114298845,0.0394004323,-0.0662378147,0.1697606444,0.1508942842,-0.0164638199,-0.2606024444,-0.0459897779,-0.7141709328,-0.0054459935,-0.058167167,-0.2517243922,0.0189056545,0.164165765,-0.1548176557,-0.0626190826,-0.3487443328,-0.2288343161,0.1530577689,-0.1076024026,-0.6533303857,-0.0671836808,-0.0007704482,-0.252815634,-0.0974926949,-0.2104124278,0.0520490408,0.0282566641,-0.1157071292,0.1471559852,-0.2626264393,0.0655457005,0.0791369751,-0.2817266881,-0.1671254933,0.4770954847,0.1151685044,0.0676666573,-0.1748857498,0.08076334,-0.0295648389,0.0588541888,0.1463176012,0.3296325803,0.1459769458,-0.1990508139,0.1463579983,0.4578507245,0.2763280869,-0.0117400689,-0.1037370339,0.1773374677,-0.0981269032,0.1541680247,0.1440638006,0.2231404781,-0.1246735752,0.2193232179,-0.1163196638,-0.253995955,0.266592741,-0.5438081026,-0.3536892235,-0.2601950467,0.1049405485,0.0536958836,-0.1772624701,-0.9156144261,-0.293112874,0.2973383665,-0.0863226876,0.008624373,0.1512214839,-0.0172133725,0.0925923064,-0.2654484212,-0.1597658247,0.0733579323,0.0531114899,0.0430421568,-0.2308625877]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/790","title":"Error running pip install -e \".[dev]\" on MacOS 10.13.6: faiss\/python does not exist","comments":"I saw that `faiss-cpu` 1.6.4.post2 was released recently to fix the installation on macos. It should work now","body":"I was following along with https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html#adding-tests-and-metadata-to-the-dataset when I ran into this error.\r\n\r\n```sh\r\ngit clone https:\/\/github.com\/huggingface\/datasets\r\ncd datasets\r\nvirtualenv venv -p python3 --system-site-packages\r\nsource venv\/bin\/activate\r\npip install -e \".[dev]\"\r\n```\r\n\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/59632\/97868518-72871800-1cd5-11eb-9cd2-37d4e9d20b39.png)\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/59632\/97868592-977b8b00-1cd5-11eb-8f3c-0c409616149c.png)\r\n\r\nPython 3.7.7\r\n","comment_length":18,"text":"Error running pip install -e \".[dev]\" on MacOS 10.13.6: faiss\/python does not exist\nI was following along with https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html#adding-tests-and-metadata-to-the-dataset when I ran into this error.\r\n\r\n```sh\r\ngit clone https:\/\/github.com\/huggingface\/datasets\r\ncd datasets\r\nvirtualenv venv -p python3 --system-site-packages\r\nsource venv\/bin\/activate\r\npip install -e \".[dev]\"\r\n```\r\n\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/59632\/97868518-72871800-1cd5-11eb-9cd2-37d4e9d20b39.png)\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/59632\/97868592-977b8b00-1cd5-11eb-8f3c-0c409616149c.png)\r\n\r\nPython 3.7.7\r\n\nI saw that `faiss-cpu` 1.6.4.post2 was released recently to fix the installation on macos. It should work now","embeddings":[-0.0769137219,-0.4744410813,-0.1149005964,-0.0313249752,0.0470356792,0.0747042671,-0.2394820005,0.2321131378,0.3420989513,0.0246700086,-0.0713810101,0.2375308126,0.0024278027,0.066258654,0.0126584889,0.0398926735,0.4198972881,0.2260307074,-0.0316481218,-0.2329014987,0.0014593964,0.3070571423,-0.0191779342,0.0570393763,-0.1384629309,0.0629520267,-0.2245894074,0.0982487574,-0.2784340382,-0.4324710667,0.1324747354,-0.1166830063,0.0231352206,0.4486960769,-0.0001248198,0.1516883075,0.2149834633,0.0974139795,-0.2460971624,-0.3716697395,-0.1054948047,-0.307315439,0.1879502684,-0.0491277799,0.15651232,-0.0496589951,0.0473566465,-0.1054647192,0.2997088134,0.1950040609,0.1107113287,0.459600836,0.3656337261,-0.1996756494,0.0875823796,0.1747200787,-0.2323464006,0.1902423799,0.2368948609,0.1777076125,0.5242687464,0.1032261401,-0.0240859278,-0.049831178,0.0052525774,-0.0410645306,0.2971595526,-0.2509606481,-0.1255887747,0.0883223712,-0.0781328827,-0.2779022753,-0.4838891327,-0.1123396754,0.1949889213,-0.1475775391,0.2691460848,0.2406691313,-0.2661559582,0.2806168795,-0.0845102519,-0.5820125937,-0.111558266,0.0773854405,0.1476298571,0.0403320938,-0.1895345747,0.0433260575,0.2459104508,0.0419017933,-0.3036502302,0.0213176515,0.2408441901,0.2564199567,-0.2195349485,-0.1184682548,-0.0794374645,-0.0824062452,0.0761013553,-0.2071825862,-0.6648669839,-0.3138446808,-0.2615450919,0.2230482548,-0.0323677324,-0.0902222246,-0.0856781825,-0.0056550601,-0.0009375344,0.023171328,0.2461711615,0.0166704543,-0.1541399956,-0.1352241337,-0.7007260323,-0.3548826277,0.2498062551,-0.1318226755,-0.4621340632,0.1604498178,0.0744873956,0.0655441955,0.0822720677,0.3852746189,0.0288051385,-0.105658479,0.1026321799,0.2561386824,-0.2377709746,0.1440687478,-0.0480729602,0.1899397224,-0.1215004027,-0.0111378739,0.121295549,-0.3222525418,0.3431439102,0.1137485206,0.0346564241,0.136552915,-0.2019346952,0.1111452505,-0.2067697644,0.5875923038,0.0539902113,0.1794100255,0.1431519389,0.1792557985,-0.1760357022,0.1573632658,-0.2410850674,-0.244509086,-0.2892553508,0.069831565,-0.2390052378,-0.1153463125,-0.0618102364,-0.2575477362,-0.089860037,0.1062235609,0.0646146536,-0.0425125323,0.1820692867,-0.1051482931,0.1427769065,0.0718968287,-0.1721661985,-0.2253894359,-0.251116693,-0.1698783338,-0.093329452,0.3640250564,0.2783349752,0.2200587541,-0.26883322,-0.1268996149,0.435151279,-0.5085304976,0.0122087738,0.2115213573,0.0271888059,-0.1193129569,0.1028826088,-0.1287745684,-0.4710634053,0.10203477,-0.0978003442,0.0465045869,-0.0181763005,-0.2232404053,-0.1655162275,-0.5149326921,-0.0163664259,0.3023715615,0.1532721221,-0.0726735517,-0.0383706912,-0.3819718063,0.1565912813,-0.0442993604,0.0273570158,0.4160861969,0.8438920379,0.5213233829,-0.0873947442,-0.366656661,0.2548884749,0.219048664,0.0821953192,0.0615975261,-0.4336166084,-0.1164684817,-0.1909641922,0.4214517474,0.0527328514,-0.1817916334,-0.0493554324,-0.0084572975,0.220620513,0.258086592,-0.3192430735,0.2147803754,-0.0341657884,0.3442053199,-0.4534943998,0.1650847197,-0.2749231458,-0.0954194516,-0.1468310356,0.3459373713,-0.001902262,-0.0620544888,0.0765880048,0.2894162536,0.0595697723,-0.2615385652,0.2932676971,-0.0617162138,0.0749857128,-0.2080413699,0.150472194,-0.0534184948,-0.0314483494,0.396376282,0.0543086417,0.0109212417,-0.026063066,0.1357116699,0.2305481583,0.2402281165,0.2736192644,0.0662929937,0.0941658393,-0.2988328636,0.3442799449,0.0353484787,0.3008334339,-0.1841428131,0.119016014,-0.0637628064,0.1355283558,-0.0006024914,0.0433985926,0.2118345201,-0.096296154,0.2498711348,0.0397552438,-0.3664267659,0.5468246341,0.0831656232,-0.2761651278,0.1839732379,-0.055418361,-0.048827488,0.2700510621,0.007651459,0.0988890007,-0.1048019901,-0.0599188283,-0.086297743,-0.2169666588,-0.4104287326,-0.0999905542,0.2072316557,-0.3612687588,-0.0186494756,-0.0393618122,0.076958321,-0.1209061742,-0.3665560782,-0.4147287011,-0.160359323,0.2663107216,0.0446308106,0.0994489416,0.2024787217,-0.0282269716,0.0350032188,0.0785415098,-0.1282932609,-0.2449283451,-0.073949948,0.0042509297,-0.0038533323,-0.1449314505,0.0087598739,0.3355324864,-0.1896279454,-0.1491673738,-0.3814901412,-0.3965193331,0.2551424205,-0.4297249019,0.4415021837,0.3835313022,0.0721150935,-0.0986621678,0.0896866322,0.2708003521,-0.3785279989,-0.3092986047,-0.055280529,0.0339288302,-0.2750465572,-0.1942581683,-0.2093297988,-0.3411789834,-0.095007807,0.3561174572,-0.0179157313,-0.0884842947,0.1934183836,0.3617049158,0.10972251,-0.0950982943,0.0694146231,-0.3011163175,-0.3425506949,0.0825262889,-0.2646844089,-0.2384745479,0.3331691325,0.1514116526,0.4445262253,0.0719862357,-0.3710216284,-0.2489838153,-0.2236432582,0.4790898561,-0.1239096299,0.6306265593,0.1899009347,0.008816855,0.0424109809,-0.1913038641,0.1098376364,-0.1345633417,-0.1592764854,0.2964021564,-0.1310562193,0.4414909184,-0.1539944857,0.8269540668,0.3067979217,0.0285531003,0.2114059031,-0.2127243578,0.6402563453,-0.0050484003,-0.2645453811,0.1621617675,0.3845385015,0.0281237271,0.0711053684,0.0809923559,0.3834081292,-0.0617692471,0.1307184994,-0.3502662182,0.0121707907,-0.1711339504,0.0333107971,0.4369103909,-0.1325215101,0.0167511124,-0.0921440199,-0.0246128682,0.0159830581,0.4872438312,0.2405374199,0.1135439649,-0.0113850571,-0.4339258075,-0.4145162106,0.1889892519,-0.1134177074,0.360956341,-0.1997369677,0.1450020969,0.2208952606,0.1292618662,0.7559735775,-0.0684147477,-0.2122815698,-0.0792637244,-0.0276479889,-0.6903051138,-0.1361582428,-0.1578816026,0.0670234188,0.4556030631,0.0213148985,-0.2783058286,-0.1888050586,0.3221387565,0.0266918391,0.0881487802,-0.3671644926,-0.4545159936,-0.4209932983,-0.3265373409,-0.0843287483,0.0937032849,-0.0363886468,-0.1579377651,0.2681912184,-0.0576956458,0.1077918634,0.1095705852,0.0996327102,0.0358028226,-0.2422961891,0.2098245323,0.1346709728,-0.0248493552,0.1933073848,0.570361197,-0.0988542885,-0.6263926029,-0.1072739214,0.1094896272,0.34236148,0.2046235055,-0.180457294,0.0336655788,0.156871587,0.2679293156,-0.0799073726,-0.0983624309,0.3739109337,0.2174788564,-0.099006623,-0.0926613808,0.2381321043,0.3819628358,0.1373059601,-0.1019655541,0.3287504017,-0.0422448292,0.2344636023,-0.1059828699,0.7789868712,-0.0867024064,0.1590816081,0.6274837255,-0.0765039846,0.8931114078,-0.3930598199,0.0722030699,-0.2330359966,0.0784893632,0.0208487082,-0.1331031919,0.4661638141,-0.0283789821,-0.1705223173,0.2575988173,0.3362355232,-0.12261457,0.1674825549,0.0626431108,-0.1270908266,-0.1052532792,-0.3380537331,0.1101873815,0.0485466011,0.3355702162,-0.1552785188,0.21168603,-0.3033243418,-0.1789358258,-0.3892088532,-0.036400985,0.0443954207,0.3537194431,0.4338510633,-0.0443854146,0.2167690545,-0.215242058,0.6817358732,0.1902540028,-0.3260304034,0.3821090758,-0.2835861444,-0.2325629592,0.0873810872,0.0371801183,0.0813595355,-0.1498856694,-0.411220789,0.0643606484,-0.114429675,-0.1316247433,-0.2042658776,-0.0707614943,-0.2570985556,-0.1815521568,-0.3102241457,-0.0528333336,0.2538765371,-0.013730553,0.0050241062,0.2653237283,-0.2492889017,0.2417332679,-0.2785974443,-0.0418083444,-0.2203862369,0.1574952006,0.1125861406,-0.3182970881,0.0629968122,0.0499428958,-0.1636691689,-0.0395620167,-0.3638559282,-0.1107881367,-0.2669747174,-0.0840692148,0.3426751792,-0.2099535912,-0.030302899,0.0186719336,0.1352364421,-0.3024706841,0.0215454642,-0.3745931983,-0.0151576782,0.1055134386,-0.1055936888,0.1561023444,0.2724726796,0.2867538929,-0.0114068221,-0.1035005823,-0.1404443681,0.0588219799,0.0497312397,-0.0730718747,0.2619276047,-0.028208388,0.2338467538,-0.252137661,0.0764388368,-0.1887474209,-0.1489500999,-0.0388767384,0.044917807,0.1710525602,0.089426294,0.0066138338,0.004243982,-0.0406659245,-0.277926296,-0.0815342143,0.1986554861,0.1787802279,-0.232716769,0.2087052017,-0.4240727425,-0.0751486197,0.0302082878,0.1574312449,0.0957506672,0.3788109124,0.2075852752,0.2072230875,-0.1240368709,0.1277692467,-0.1183503345,-0.0022443975,0.1205425188,0.1774629503,0.1549585611,-0.2892750204,-0.1221728995,-0.1425925195,0.3925551474,0.3052443266,-0.1155149415,0.173723191,0.4368033409,0.0998252332,-0.3585011363,0.0340152793,0.4373786151,0.0289703142,-0.0176879056,0.2994979322,0.2153308839,-0.0105129881,0.2599743009,0.2954028845,0.253575474,0.1891105473,0.1018955931,-0.11949604,0.1733826101,0.0737443715,-0.0307610314,0.1063571051,0.1197481453,0.5863202214,-0.1923638731,0.1603039801,-0.1387946904,0.1327591538,0.1539234817,-0.2321935743,0.1293734908,-0.2308664471,0.1097538248,-0.0612265654,-0.1860149652,0.2869600356,-0.2451802939,-0.2856755853,-0.2843365669,0.3253025413,-0.1350783855,-0.078796111,0.2950965166,0.0836127624,-0.0231719725,-0.0377520956,0.2906121016,-0.4265969992,0.1664597094,0.3014774919,-0.0952513143,-0.4780474305,-0.0592677258,0.4519571066,0.0746057704,0.0984166414,0.2571079731,-0.0159455184,-0.39975667,-0.1361056715,0.3689001203,0.7117063999,0.1954545975,0.1044975743,0.1515866071,-0.2034646273,0.1490750164,-0.0896825939,0.1261641383,-0.1106125191,0.3228428066,0.0242033601,0.0970540866,-0.020511115,-0.3307760656,-0.0061045238,0.2146985084,-0.2782035172,0.15845038,-0.2380261719,-0.0854861587,-0.0862473026,0.0479842052,-0.6899208426,0.1493898034,0.2461606264,0.1270478368,-0.1731130034,-0.321834147,-0.0119329253,-0.1000670865,0.1747179478,0.6672796607,0.3155692518,-0.2467534393,-0.1719846874,-0.3748016059,0.1280721277,-0.2415083796,-0.286724478,0.0783744082,0.0326481014,-0.0608052425,0.1358946413,0.3229253888,-0.0712390095,-0.0536537841,0.0441737808,-0.0927747786,-0.154350996,-0.1638700813,-0.2126151472,0.2131003439,0.0746029541,-0.0040110201,-0.1527809799,-0.1440476477,-0.1683674604,0.1801643819,-0.0409720838,-0.3412232399,0.5376930833,0.2197144032,0.2743596733,-0.0054457854,-0.3348965347,-0.2708339095,-0.2661760449,-0.1113564298,0.3895553052,-0.1045996323,0.0746271536,-0.3636821508,-0.3409645259,-0.3371858895,0.4985400736,0.2556684315,-0.4602464437,0.0196081586,0.3215573132,-0.32603392,0.4232559502,0.170706436,0.2509987354,-0.1147209257,0.5524257421,-0.212650463,-0.1802524328,0.38535133,-0.1864649355,-0.2123773396,-0.1088995263,0.2661765516,0.2009283006,-0.1819356531,-0.8109849691,0.0378586501,0.1709183007,0.2873196006,-0.1208509952,0.1078975201,-0.0183062889,-0.081739448,-0.0000585547,0.4948093891,-0.1948238015,-0.0330200344,0.4954161644,-0.097830981]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/790","title":"Error running pip install -e \".[dev]\" on MacOS 10.13.6: faiss\/python does not exist","comments":"Closing this one.\r\nFeel free to re-open if you still have issues","body":"I was following along with https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html#adding-tests-and-metadata-to-the-dataset when I ran into this error.\r\n\r\n```sh\r\ngit clone https:\/\/github.com\/huggingface\/datasets\r\ncd datasets\r\nvirtualenv venv -p python3 --system-site-packages\r\nsource venv\/bin\/activate\r\npip install -e \".[dev]\"\r\n```\r\n\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/59632\/97868518-72871800-1cd5-11eb-9cd2-37d4e9d20b39.png)\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/59632\/97868592-977b8b00-1cd5-11eb-8f3c-0c409616149c.png)\r\n\r\nPython 3.7.7\r\n","comment_length":12,"text":"Error running pip install -e \".[dev]\" on MacOS 10.13.6: faiss\/python does not exist\nI was following along with https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html#adding-tests-and-metadata-to-the-dataset when I ran into this error.\r\n\r\n```sh\r\ngit clone https:\/\/github.com\/huggingface\/datasets\r\ncd datasets\r\nvirtualenv venv -p python3 --system-site-packages\r\nsource venv\/bin\/activate\r\npip install -e \".[dev]\"\r\n```\r\n\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/59632\/97868518-72871800-1cd5-11eb-9cd2-37d4e9d20b39.png)\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/59632\/97868592-977b8b00-1cd5-11eb-8f3c-0c409616149c.png)\r\n\r\nPython 3.7.7\r\n\nClosing this one.\r\nFeel free to re-open if you still have issues","embeddings":[-0.0273194257,-0.5231038928,-0.096934855,-0.0106153516,0.0975628421,0.1341893673,-0.1737166494,0.1631080657,0.2741362751,-0.0022856402,-0.0708376169,0.2156621218,-0.1082775667,0.0985628888,0.0775272325,0.01747627,0.3742345273,0.2318359315,-0.0920170918,-0.2636029124,0.0145522403,0.2337813824,-0.0420286246,0.100073278,-0.1499058753,0.0378152989,-0.2194844931,0.1425873339,-0.353730768,-0.402471602,0.2347421199,-0.107522659,0.0215717796,0.3952128887,-0.0001235658,0.1032380909,0.2734370828,0.0662982985,-0.2936411202,-0.3245178759,-0.1458859742,-0.320397228,0.1754925102,-0.0178544708,0.1331555843,-0.104241699,0.0076296604,-0.1902157366,0.3414696455,0.2017251551,0.1084856763,0.4261991084,0.4068373144,-0.1632526219,0.2117305994,0.2050491124,-0.2012479156,0.3519266844,0.28031528,0.116853185,0.6213964224,0.2089429647,-0.0343484767,-0.1102250293,0.1948963404,0.03770772,0.3722572923,-0.2741408646,-0.1215145737,0.114058353,0.0176468231,-0.3055261672,-0.5101055503,-0.1120397076,0.2295579612,-0.2107520849,0.2574312389,0.2584105432,-0.233322531,0.2482834607,-0.1876980215,-0.5710384846,-0.1166625321,0.1707498431,0.0451320596,0.051870659,-0.208025679,0.0745922998,0.2429015785,-0.0027772777,-0.296561718,-0.018023748,0.1877973974,0.2639892995,-0.0956165344,-0.1575754732,-0.1221077368,-0.0881233588,0.1529884487,-0.1431347579,-0.6991064548,-0.3801017106,-0.2554598749,0.2517642677,0.0650157481,-0.0301941056,-0.1212369278,0.1054580435,0.0112979757,-0.0623411313,0.3388760388,-0.0208951272,-0.0926854759,-0.1136968061,-0.6635062099,-0.2941928804,0.3216288984,-0.0755484328,-0.4111911654,0.1367155463,-0.0189097226,0.043910142,0.0345933959,0.3214659095,0.039343033,-0.1889539212,0.0554462969,0.3249624074,-0.1292402744,0.2426787466,-0.0277330615,0.1997445673,-0.1755705923,-0.1165654212,0.1093894914,-0.3418809772,0.306987375,0.0969894752,-0.0439148545,0.1191236973,-0.2606989741,0.0845028982,-0.1990306973,0.5372544527,-0.0130999163,0.1334887296,0.1182977557,0.1031897217,-0.2123124301,0.1156303883,-0.2179847211,-0.2217271477,-0.3601792157,0.0784799382,-0.1801202595,-0.091880776,-0.0741581097,-0.2496301383,-0.1176223904,0.0571911708,0.1138581708,-0.049436558,0.2282342166,-0.0834384188,0.0972391367,0.1804355234,-0.1646166593,-0.2316526771,-0.1654298902,-0.1200642437,-0.1138377339,0.3416836262,0.1908354014,0.1514125317,-0.3073675931,-0.1426826119,0.452106297,-0.5518192053,0.0349034816,0.1649605632,-0.006434239,-0.1189611852,0.0722640827,-0.0798432752,-0.4267463088,0.0257764235,-0.093643941,0.1506675631,-0.0481708907,-0.2697546482,-0.1425447315,-0.4599078596,0.0446076058,0.251193285,0.1198393255,-0.1239597648,-0.0295924265,-0.4910314679,0.1126911491,-0.0767626539,0.0207868069,0.4012482762,0.8845362663,0.5233003497,-0.1491191089,-0.3564125001,0.1082523838,0.200244382,0.0694542676,0.0255157277,-0.4744249582,-0.1305096298,-0.2103652656,0.4095723033,-0.0286410488,-0.1983034015,-0.0555915013,0.0495223813,0.2186944783,0.3279632032,-0.3118236065,0.1200481281,-0.0527677201,0.3686856925,-0.4925777018,0.2077807784,-0.258531183,-0.0113746691,-0.1513752937,0.3697832227,0.0258268993,-0.0742040649,0.1240360141,0.2985503972,-0.0168909412,-0.2344589233,0.2954894304,-0.063585043,0.105995819,-0.3291001916,0.1728343517,-0.0440846495,-0.027087409,0.3885313272,-0.0870241225,0.0119226677,-0.0828513876,0.1655735523,0.2052076906,0.2633179128,0.2058845609,0.0387275107,0.0466612987,-0.3190142214,0.283908695,-0.071476303,0.2851903141,-0.1596424133,0.2092591524,-0.1704484075,0.1309322268,-0.0693897307,0.0578127317,0.2683857083,-0.138446942,0.2407321781,0.0198557395,-0.1955583543,0.5839757919,0.1234679073,-0.2197704017,0.184215188,-0.0242488384,-0.0115588382,0.316785574,0.031579528,0.049329266,-0.0476397835,-0.070288375,-0.1011442244,-0.1815799922,-0.3469423354,-0.1386110187,0.1769586056,-0.3346327543,0.0368443541,-0.0684498847,-0.0093014557,-0.144128114,-0.3997024596,-0.4619700313,-0.2450969666,0.2865228057,-0.0359962918,0.0701265261,0.1897765398,0.0828685164,-0.0325652286,-0.0011017963,-0.0551039316,-0.2743811309,-0.1513365656,-0.0009942169,-0.0115679214,-0.0873634815,0.0056128614,0.4084482789,-0.2074378729,-0.0839894935,-0.4446350634,-0.4527582824,0.2999016941,-0.4500616789,0.513659358,0.4248966575,0.103361249,-0.0722663552,0.0533548445,0.3192103803,-0.3214621842,-0.3069763184,-0.0360465981,0.0428655632,-0.261259377,-0.2100961953,-0.2778549194,-0.4181415439,-0.0831286609,0.3565271795,-0.0697444975,-0.1467641145,0.3928925395,0.3376251459,0.0937834457,-0.1352419555,0.0761475936,-0.2972796261,-0.2731989324,0.0789754838,-0.2602440119,-0.1548136026,0.3411334753,0.1218075007,0.3819686472,0.099374041,-0.3784020841,-0.3387643695,-0.1463201493,0.3839563429,-0.1521921307,0.6276428103,0.2074134648,0.0536574461,0.0403194875,-0.169843331,0.132310003,-0.1780489385,-0.1670698076,0.2844780684,-0.1240300387,0.4813351929,-0.1860351264,0.8145718575,0.3639457524,0.0571102388,0.1872041672,-0.2137907892,0.737574935,-0.0186792314,-0.3020662963,0.0426588506,0.4065285623,0.0794692859,0.0715157464,0.142518729,0.3645596504,-0.0749675184,0.0880777538,-0.2960062325,-0.0092515331,-0.1502878517,-0.0762051046,0.4354586601,-0.1531744748,-0.0069325119,-0.0607232824,-0.0486379601,-0.0859324336,0.4860022962,0.2855802774,0.12686418,-0.0806930438,-0.4618548453,-0.2992686927,0.1853430569,-0.0801085159,0.3153047264,-0.2221250087,0.1964725107,0.219823584,0.1587648243,0.7820807695,-0.1615132242,-0.2203539163,-0.1352173686,0.0504485443,-0.7206327915,-0.0869576335,-0.1559451818,0.0614969842,0.3271164894,0.0169412084,-0.3512327671,-0.1422236264,0.4955541492,0.0654763579,0.1221221685,-0.3252646923,-0.4231464863,-0.4348807335,-0.3758741915,-0.0820174292,0.0824617818,-0.10968595,-0.069688499,0.2016672343,-0.0391265489,0.0933047011,0.1115741283,0.0503666997,0.0509381145,-0.149692148,0.2095062435,0.1426092237,0.0676998943,0.2173340619,0.613203764,-0.0780175775,-0.5313085318,-0.241089195,0.116397053,0.4383121133,0.2323056161,-0.1609713137,-0.0153918071,0.0812415406,0.2736735642,-0.0718278363,-0.1623792797,0.4459710717,0.2025856674,-0.101945959,-0.0681900904,0.191023007,0.3595486879,0.1482123882,-0.0385371931,0.3714217246,-0.066004768,0.1398102343,-0.0845543668,0.9183308482,-0.0935589969,0.1743528694,0.5837950706,-0.1239816844,0.8204815984,-0.3482013941,0.0128890388,-0.1981996,0.1030794382,-0.0064696353,-0.1772777289,0.4659460187,-0.11061012,-0.1929731667,0.2418677062,0.2256616801,-0.0120400321,0.081339553,0.0460570194,-0.1741657555,-0.0673997626,-0.2912816107,0.1096171141,0.0444961637,0.3541144133,-0.1628872007,0.1512389481,-0.2796699107,-0.2512816191,-0.3562674224,-0.0571627654,0.0786622986,0.3377154768,0.431250155,-0.1597439349,0.1616156995,-0.1158198938,0.566421628,0.2179654539,-0.3136972785,0.3472543955,-0.3345133364,-0.211920321,0.0641935617,0.020036988,0.042236045,-0.1076134667,-0.4160010517,0.1206100732,-0.1102076173,-0.0664628819,-0.2851743996,-0.0654131547,-0.1878081709,-0.1924899817,-0.2579020262,-0.0478700511,0.188326776,-0.0146487765,-0.000699463,0.2223066986,-0.2333004475,0.2254861295,-0.229225859,-0.0020009559,-0.2171983719,0.1658300012,0.0962589234,-0.2224345952,0.0714422911,0.0163808577,-0.1461010724,-0.0677005202,-0.2771719992,-0.0531375259,-0.3326382339,-0.1151178926,0.3919978738,-0.1433207691,-0.0666090697,0.0503009483,0.1436928064,-0.2790313661,0.0690598786,-0.3599552512,0.0055800108,0.1604699045,-0.0455572642,0.2205876112,0.1734713465,0.2651356459,0.0494715385,-0.0774117783,-0.1629629433,0.073098287,0.0075453813,-0.0150976451,0.3665457666,-0.009751685,0.2340275496,-0.2675764263,0.0854779705,-0.1930811852,-0.1960017681,-0.0356489643,0.1032852307,0.1589487791,0.1105028614,-0.0412770882,0.0679556057,-0.0818392038,-0.2758215666,-0.10208866,0.1726560444,0.1530869901,-0.1743701249,0.1730977446,-0.4592134655,-0.0504361019,0.0588917211,0.073495239,0.0531229153,0.3864119947,0.1768254936,0.1051933169,-0.1224669442,0.0694595203,-0.2162146419,0.0482474081,0.1612968594,0.1768624038,0.2280511707,-0.3039438426,-0.0922592729,-0.1575178206,0.3909111023,0.2778919935,-0.2286200821,0.2800627947,0.4462081194,0.0800167397,-0.3118492067,-0.0229379125,0.353392601,0.0105322367,-0.011286688,0.2549135983,0.1318441629,-0.0136800036,0.3224307597,0.2732493579,0.3008185327,0.1432294101,0.1492088288,-0.0091541903,0.0522161573,-0.0192527659,-0.0818803534,0.1467209011,0.1594325602,0.5342829227,-0.2300500721,0.1766435653,-0.2516861558,0.1258686632,0.1754400581,-0.262452513,0.1094681844,-0.2088209391,0.2479609549,-0.0206602514,-0.1576407701,0.2855012715,-0.2710872591,-0.2905469537,-0.2819435298,0.3530016541,-0.081237942,-0.0171822514,0.249061957,0.0521929823,-0.0018140441,-0.0273732692,0.2175214887,-0.4868814349,0.0819065198,0.3186800182,-0.1122050881,-0.5041239858,0.0123204887,0.4065821171,0.0952528417,0.0509507805,0.2943376005,0.036145106,-0.2608292401,-0.1347305626,0.3485756516,0.6598058939,0.2011416405,0.112052545,0.1633978486,-0.2195015848,0.0933539644,-0.0048949248,0.1834415793,-0.0597007871,0.3164571226,0.0176756382,0.1095485911,-0.045827195,-0.2378726304,0.0015209866,0.2013236135,-0.2562315464,0.141290471,-0.2377260476,-0.0904776901,-0.1487169117,-0.000690642,-0.7581383586,0.1529798508,0.2633711696,0.2183815241,-0.1385465413,-0.2645559013,-0.0098914634,-0.0611316338,0.213260904,0.6859478951,0.2986184359,-0.2723548114,-0.1659692675,-0.3634494841,0.0656816289,-0.2906607091,-0.3080950677,0.093626596,-0.0413575098,0.0135971913,0.1747493893,0.235088259,-0.0100016817,-0.0604531057,-0.027522346,-0.0779140145,-0.2410973907,-0.0678266287,-0.1578721851,0.2311872095,0.0504185744,0.0008148697,-0.1927704811,-0.138252914,-0.1193757877,0.3012028635,-0.1198446602,-0.275883764,0.4817641079,0.2094062567,0.3409634829,0.0422541611,-0.3465901613,-0.2541124821,-0.2792686522,-0.0754595995,0.4428474903,-0.0490170084,0.1129928902,-0.4008405209,-0.4534961283,-0.3493567705,0.3797669411,0.2767670155,-0.2626463771,-0.0054045264,0.3104399145,-0.2675612867,0.3911505044,0.1185877472,0.228493616,-0.136116758,0.4821123481,-0.2284451723,-0.2020583153,0.4988089204,-0.0865284801,-0.1871720999,-0.0002494962,0.2483351976,0.2377104312,-0.1799276173,-0.8118523955,0.1222163588,0.1981204599,0.2975278199,-0.1140525118,0.1636185199,0.069153972,-0.0991344452,0.0584153049,0.4341754317,-0.1925430149,-0.0363544486,0.5292338729,-0.0853545889]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/786","title":"feat(dataset): multiprocessing _generate_examples","comments":"I agree that would be cool :)\r\nRight now the only distributed dataset builder is based on Apache Beam so you can use distributed processing frameworks like Dataflow, Spark, Flink etc. to build your dataset but it's not really well suited for single-worker parallel processing afaik","body":"forking this out of #741, this issue is only regarding multiprocessing\r\n\r\nI'd love if there was a dataset configuration parameter `workers`, where when it is `1` it behaves as it does right now, and when its `>1` maybe `_generate_examples` can also get the `pool` and return an iterable using the pool.\r\n\r\nIn my use case, I would instead of:\r\n```python\r\nfor datum in data:\r\n yield self.load_datum(datum)\r\n```\r\ndo:\r\n```python\r\nreturn pool.map(self.load_datum, data)\r\n```\r\n\r\nAs the dataset in question, as an example, has **only** 7000 rows, and takes 10 seconds to load each row on average, it takes almost 20 hours to load the entire dataset.\r\nIf this was a larger dataset (and many such datasets exist), it would take multiple days to complete.\r\n\r\nUsing multiprocessing, for example, 40 cores, could speed it up dramatically. For this dataset, hopefully to fully load in under an hour.","comment_length":46,"text":"feat(dataset): multiprocessing _generate_examples\nforking this out of #741, this issue is only regarding multiprocessing\r\n\r\nI'd love if there was a dataset configuration parameter `workers`, where when it is `1` it behaves as it does right now, and when its `>1` maybe `_generate_examples` can also get the `pool` and return an iterable using the pool.\r\n\r\nIn my use case, I would instead of:\r\n```python\r\nfor datum in data:\r\n yield self.load_datum(datum)\r\n```\r\ndo:\r\n```python\r\nreturn pool.map(self.load_datum, data)\r\n```\r\n\r\nAs the dataset in question, as an example, has **only** 7000 rows, and takes 10 seconds to load each row on average, it takes almost 20 hours to load the entire dataset.\r\nIf this was a larger dataset (and many such datasets exist), it would take multiple days to complete.\r\n\r\nUsing multiprocessing, for example, 40 cores, could speed it up dramatically. For this dataset, hopefully to fully load in under an hour.\nI agree that would be cool :)\r\nRight now the only distributed dataset builder is based on Apache Beam so you can use distributed processing frameworks like Dataflow, Spark, Flink etc. to build your dataset but it's not really well suited for single-worker parallel processing afaik","embeddings":[-0.5645692348,-0.1257423609,-0.1173854023,-0.166632995,-0.0713058785,-0.0848814175,0.3084898293,0.345071286,0.1016154811,0.4008847475,0.2586757839,0.1760060638,0.0299800057,0.1242403463,0.1066480726,0.0938062146,0.0195515174,0.1313365996,-0.2605703473,-0.0488939695,-0.3563632071,-0.0217418987,-0.0285441708,-0.1431766152,-0.3056107759,-0.1653455794,-0.0860954672,0.0822234824,0.0034817467,-0.2213504016,0.1820405722,0.5444226265,0.2789001465,0.412689507,-0.0001032015,-0.0390751399,0.2969443798,0.0052962429,-0.3610217869,-0.1338654608,-0.1356738806,-0.3160059154,0.0205591172,-0.2597048283,0.0211865883,-0.187975958,0.246158734,-0.4733569026,0.0596728586,-0.1250498146,0.2303161472,0.3476171196,-0.2190504968,-0.2033669204,-0.1540288478,-0.0212983526,-0.1601623893,0.1414245963,0.5995049477,-0.3429674506,-0.1796414107,0.1689000428,-0.0268125031,0.3214696348,0.258869797,0.0280346218,0.2928673923,-0.4736996889,-0.0322284326,0.2586959898,0.3223201632,-0.0653719604,-0.4411685467,-0.4137692153,0.1503238827,-0.1780725271,0.098107338,0.4271344543,-0.3625374436,0.1794993281,-0.3271012902,0.0309631135,-0.0444259606,-0.0228912085,0.0459823161,0.4582173228,0.1388121396,0.16112037,0.2576366365,0.0965296403,0.3622563779,-0.1368422061,0.3720157444,0.0939120203,-0.3785074353,-0.0503505878,0.1009605825,-0.176166445,0.2498613,-0.0211286955,0.2596452832,-0.0036759903,-0.0463215411,0.2976036072,0.3191949129,-0.3273790181,-0.1877823919,-0.0794773921,0.2441907078,-0.1838876605,-0.0110040419,0.0401750095,-0.0484521315,-0.2198261321,0.2778868973,0.0812505409,0.0187796224,0.162140891,-0.1537270546,-0.2023274153,0.0209043548,-0.0703686997,0.0725858733,0.0811427906,-0.0497736,0.5781404376,-0.1443649977,0.0651179478,-0.2028383017,-0.1945530623,-0.0828361958,0.0311295856,-0.2206898481,0.2142316848,0.3574272692,0.035415668,0.096330516,0.2090110183,0.2967506647,-0.0109740403,0.4788134396,-0.1122947857,0.1474232823,0.1567318588,0.0870099813,-0.1228972971,-0.1640131027,-0.1195179448,-0.2253357172,0.1451944113,-0.0699561089,-0.4257085323,-0.1593768001,0.2495116144,-0.055490911,0.1053343341,-0.1263075471,0.2614251673,0.1310819238,0.1803332418,-0.0915223584,-0.1575553715,-0.2608850896,-0.0602515563,0.325807035,0.2831625342,-0.1300890595,0.2780073881,0.0442510545,0.0856357887,0.1165908054,0.1238762811,-0.1102689132,0.2128217965,-0.1414506733,-0.052149646,0.2380713224,-0.1173013002,-0.0812401026,0.3136304617,-0.0579978116,0.0389036834,0.1617939472,0.1347354949,0.2970617115,-0.0906533971,0.242137447,0.4703310132,-0.3758898377,0.2193625569,-0.2116719484,-0.1738851219,-0.168848291,0.2470009327,0.1568025053,0.2567701936,-0.0740778744,-0.2490072846,0.485902667,-0.2836574018,0.2944054008,-0.0935088843,-0.0824324414,0.1030864865,0.1391343772,-0.4160619676,-0.2420691699,0.3760923147,0.1697995067,-0.0567646362,0.0959128216,-0.1960014552,0.0544757992,0.1322979927,-0.0876032412,0.1565964222,0.1157926917,0.1026672497,-0.0603532344,-0.441342175,-0.3751507998,0.1957874298,-0.0584253892,-0.1846154928,-0.1708138883,0.1369488835,0.2126252949,-0.2457502931,0.1578003317,0.0827546641,-0.0318877399,-0.1456852853,0.052638445,0.2507564723,-0.0995185301,0.3289547265,0.1844166666,0.3626998067,0.1138872877,0.1689618677,0.05482715,0.0197907388,0.0313610248,-0.0820285678,-0.1565284431,0.4870259166,-0.3659533858,0.3455126882,0.1757371128,0.1759542376,0.4313523173,0.1570576578,-0.1642201096,0.0836196616,0.085631378,-0.0001003791,0.33670187,-0.0109573444,-0.4514661729,0.3199449778,0.1571628004,-0.182391867,0.0331963301,0.0422897488,0.3300043344,-0.1101244017,0.1448792666,0.2291253358,0.522993803,0.2154854387,0.2187856734,-0.1615592539,-0.1944313496,-0.0688633695,0.1502071619,-0.0102946041,0.3209049404,0.0703947023,0.13579005,-0.2369163185,-0.3364010155,-0.3829324245,0.1167751253,0.0366262235,-0.0017936444,0.1108320504,-0.0108287456,0.1982709169,-0.2554686368,0.2144106328,-0.0470128655,-0.3205996752,-0.0406120308,0.5437672734,-0.09951704,0.1028357148,-0.0570215955,0.2325585037,-0.1785197109,0.1215940192,-0.1859091073,-0.0988349915,-0.0075362287,0.0907950774,-0.0857532024,0.0673558488,0.4577212632,-0.0036933012,-0.104328461,-0.423586607,0.1313679963,0.0695388839,0.0352151692,0.1806467772,0.2100236416,0.0348645896,0.1539574116,-0.2294790596,0.1865983903,-0.2671281695,0.1010825038,-0.1395892799,-0.1501678973,-0.0437807031,-0.078035444,-0.0934929401,-0.3203550279,-0.4222767651,0.1101503074,-0.0202856138,0.236502856,-0.2105718702,0.0797268078,-0.1382808834,0.3285043538,-0.0509297289,-0.1423629522,-0.6931467056,0.283370167,-0.2654668093,-0.1514147371,-0.1869240552,-0.1685637981,0.0771837831,0.4729092717,-0.1621153057,-0.0578048453,-0.2334820032,0.190585658,-0.1642941982,0.0998540297,0.236958757,0.1101685539,-0.0786666349,-0.191914022,-0.1630766541,0.0809303373,0.0563186631,-0.1001630276,0.1788741648,0.2567817569,0.0413926095,0.77250278,0.2194131911,0.1988105178,0.2402208596,0.0078342538,-0.0833366066,0.0221496802,-0.2844358087,0.0830958486,-0.2348846048,-0.0329468921,0.055926878,0.0616141446,-0.3387680948,-0.0208140817,0.0103650121,-0.2248638421,-0.2632091343,0.179087162,-0.3586717844,0.3584170341,0.0024205751,0.0339504182,-0.4918853045,0.1724939048,-0.1960074008,-0.0399304852,0.3679131866,-0.1098933816,-0.2278319299,0.0442344323,-0.6570861936,0.0778432861,0.0862706453,0.3042764664,0.3668858111,-0.0175182912,0.0984907448,-0.1723390371,0.7410872579,-0.2485455424,0.1069443822,0.0468892828,-0.4014096558,-0.1866678447,-0.129987821,-0.0900292695,0.5101513267,-0.002125812,0.1709709764,-0.2782422602,-0.3481648266,0.3125014901,0.0452574305,-0.146634981,-0.4707900882,-0.2848757207,-0.3556037545,-0.1581220031,0.1461620629,0.0710196272,0.2442399859,-0.1179488152,-0.2693722248,-0.1128313988,0.0451570079,0.1888669729,0.2391284853,-0.1960514486,-0.0799753889,0.3443269432,0.3194955587,-0.1129853278,-0.0440427251,0.2980947196,-0.2402561456,-0.1970484555,0.2273578346,-0.1342851371,0.3759778142,0.2793304622,0.0216843523,-0.0937470645,0.0857071877,0.2703036368,-0.4555911124,-0.1594095081,0.1810458153,0.0273322351,-0.2600704134,-0.764082253,0.3955158889,0.1333610713,-0.0282454491,0.5454919934,-0.4678730965,-0.0673905388,0.2504844964,0.1616753936,0.6716319919,-0.4381504059,-0.009694593,-0.0543324687,0.1736839861,0.2533314824,-0.4068738222,0.3735173345,-0.2473713607,-0.4962448478,0.0433148518,-0.0401770622,-0.040390674,0.5621011257,-0.1047965363,0.263728261,0.2577420175,-0.098618716,-0.1023866609,0.5208840966,-0.4047096372,-0.4902951121,-0.3534139395,0.161268115,0.1444072872,-0.1542768478,-0.2018207461,-0.0491420105,0.3302107453,0.0352741703,-0.276158005,-0.0581101738,0.0308567435,0.1420065016,0.172443524,-0.1520049572,0.0007065558,-0.0017293529,-0.0600595362,-0.0428900979,-0.043179892,0.0672828853,-0.0735359788,0.3222302794,-0.0456901267,-0.1448190063,0.1478889585,-0.1173668578,-0.3560072184,-0.0174214467,0.0659024492,-0.2980070412,-0.0524389446,0.0876125023,0.1900136024,0.0776729733,0.1739651114,0.220745638,-0.0774758756,-0.2634599805,0.1773031801,0.259216398,-0.1591294557,0.412106514,-0.1334676147,-0.1721586138,-0.0717407018,0.0189812183,0.1435147226,0.298317939,0.1975506693,0.2281439006,-0.1307230294,-0.2915332913,-0.0961618423,-0.0146206236,-0.2505441606,0.3830240667,-0.1070127636,-0.2170334309,0.127637729,0.0363725275,0.1422406882,-0.0898071378,-0.1092833653,-0.3186972141,0.0299315508,-0.0589667819,-0.0572342202,0.1204497516,0.1106843576,0.062713787,0.1337474883,0.1624000669,-0.422390908,-0.2437303066,-0.3316542804,0.2030829787,-0.0897187293,-0.0881297365,-0.306004405,0.15482907,0.1843590885,0.0427528471,-0.1568905413,-0.2756624222,-0.1421665698,0.1092785448,-0.2088005245,-0.0420717373,0.3286156952,-0.1136608347,-0.4773363471,-0.4502692521,0.0990395695,-0.0168695599,-0.101727657,-0.0424431898,0.4718869925,-0.0475666784,-0.1252267957,0.0859649405,0.0560657382,0.0705879852,-0.204803288,0.4086611271,0.3737766445,-0.1439300328,0.2186107188,0.0627008751,0.0857835561,0.0712894723,0.3711373806,0.0988220349,-0.021229744,0.1842219979,0.4321992397,0.3389217854,0.076869145,0.1310527027,0.0768427476,0.3061861992,-0.2509643734,-0.2406508327,0.1940493584,0.0534911864,0.3035959303,0.3634151816,0.0604066961,-0.0709597021,0.1178136095,-0.0013647528,0.1401929706,-0.5134748816,0.048839733,-0.0716380477,0.1603372246,0.0301354416,0.2439599484,-0.0188760851,0.0672632754,0.3109544814,0.0954861119,0.4005775154,0.2080335468,0.1054888442,0.170996353,-0.3073976338,0.0586872511,-0.0687562525,-0.1944054663,0.1913817227,0.0423526317,0.0321912989,-0.5813704133,0.102558516,-0.3715049326,0.0990831479,-0.0097251721,-0.1769031882,-0.0922112167,-0.2176011652,-0.0591438413,-0.053065192,-0.014271819,-0.0228079632,0.6877743602,-0.0163429584,0.3382546902,-0.3257788718,0.1143595502,-0.1425328553,0.3084833026,-0.3558644652,-0.0218264237,0.1639089584,-0.1802686602,-0.1380155981,0.4435689151,0.2227330357,0.5573070049,-0.164280057,0.2105742395,-0.2797246277,0.0043355292,-0.4294217825,0.1043737829,-0.0820267498,0.0729996786,0.3381593823,0.1928336173,-0.1412856132,-0.1535686255,0.1815540493,-0.1677209139,-0.4397653341,0.2338453233,0.2389797866,-0.1849841177,0.0335126966,0.2120634466,-0.1895170063,-0.0347717702,0.4634003043,-0.3911526203,0.0308991782,-0.0786007717,0.1328496784,-0.1313948035,0.6261084676,0.0333569124,0.1973296404,-0.2347290814,-0.0686433986,-0.4603084624,-0.0685930923,-0.3893324137,0.0762336478,0.191521585,0.4422732592,0.009624159,0.2547313571,0.0306570288,0.0915272459,-0.0673318654,0.4843102694,-0.2087249756,-0.3480143249,-0.040616855,0.1016362458,0.0375813581,-0.5921882987,0.2953978777,-0.1419312954,0.1797688752,-0.1234490275,-0.0392190106,0.4771953523,-0.1351536214,0.2389727235,0.1778978258,0.1398626119,-0.2325797528,-0.0292728674,-0.2865022719,0.0202240627,-0.2869169116,0.1610234231,0.0441069119,0.0686965734,-0.2110840529,0.1074419171,-0.3195213974,0.0339687839,-0.1048387215,0.1596552879,-0.1011455134,0.1625238806,-0.1785961986,0.1942610592,-0.0893850252,0.3384362757,-0.007113657,0.0096893329,-0.1954769343,-0.2716714144,0.1582383811,-0.4215095639,-0.5883421302,-0.5599812865,0.2254885435,-0.0359872803,0.1533369124,-0.1941949129,-0.0357685722,0.2993846536,-0.0229387973,-0.1977114379,0.2384685278,0.1119248644,-0.2339143157,-0.1713932157,0.1260728985,-0.0226450041,-0.4274653792,0.0151321255,-0.3583347797]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/784","title":"Issue with downloading Wikipedia data for low resource language","comments":"Hello, maybe you could ty to use another date for the wikipedia dump (see the available [dates](https:\/\/dumps.wikimedia.org\/jvwiki) here for `jv`) ?","body":"Hi, I tried to download Sundanese and Javanese wikipedia data with the following snippet\r\n```\r\njv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner')\r\nsu_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner')\r\n```\r\nAnd I get the following error for these two languages:\r\nJavanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nSundanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nI found from https:\/\/github.com\/huggingface\/datasets\/issues\/577#issuecomment-688435085 that for small languages, they are directly downloaded and parsed from the Wikipedia dump site, but both of `https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json` and `https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json` are no longer valid.\r\n\r\n Any suggestions on how to handle this issue? Thanks!","comment_length":21,"text":"Issue with downloading Wikipedia data for low resource language\nHi, I tried to download Sundanese and Javanese wikipedia data with the following snippet\r\n```\r\njv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner')\r\nsu_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner')\r\n```\r\nAnd I get the following error for these two languages:\r\nJavanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nSundanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nI found from https:\/\/github.com\/huggingface\/datasets\/issues\/577#issuecomment-688435085 that for small languages, they are directly downloaded and parsed from the Wikipedia dump site, but both of `https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json` and `https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json` are no longer valid.\r\n\r\n Any suggestions on how to handle this issue? Thanks!\nHello, maybe you could ty to use another date for the wikipedia dump (see the available [dates](https:\/\/dumps.wikimedia.org\/jvwiki) here for `jv`) ?","embeddings":[0.194715023,-0.0537347235,0.0254635084,0.4426942468,0.162628442,0.0445552096,0.0317906849,0.4150587618,0.069062382,-0.0053408244,-0.2461343408,-0.1570815593,0.3839263022,-0.2293255627,-0.1017463058,-0.5121183395,0.1875391304,0.0159037355,0.0008706302,-0.0452764966,-0.199597612,-0.052532658,-0.184902519,0.0329155363,-0.2519637942,-0.0589046702,-0.024784578,-0.0309203956,-0.3356662393,-0.2533815205,0.2615026236,0.1938856989,0.3040901124,0.3966013193,-0.0001232719,-0.0804195926,0.5516072512,-0.0958508849,-0.5192881227,0.1004576907,-0.2622410655,-0.4349927902,-0.2489447743,-0.4146053493,0.1061190516,-0.1654881686,0.2590542734,-0.2643605173,0.2706869245,-0.0570042245,0.1158644408,0.1668395847,0.0364764705,-0.232977882,0.4131913781,0.1898481995,0.0390349664,0.0292671751,-0.145152837,0.0636961833,0.0542547181,0.2319209576,-0.2538940609,-0.0225599408,0.1514424533,0.0082278987,-0.3298267126,-0.4701971412,0.442645818,0.4518272579,0.9045652747,-0.0191264153,-0.5640506148,0.0585735254,-0.123735778,0.0984341055,0.2193090022,0.0888863057,0.0514305569,0.0775278583,0.2321285009,-0.272044003,-0.165673539,0.5533049703,-0.0247975662,0.2829127014,-0.056256026,0.0581289418,-0.1414844692,-0.3711184859,-0.1817739159,-0.3521578312,-0.0746123418,0.4161820412,0.0065638893,-0.0056546964,-0.0606315918,0.0483588651,0.1835199296,-0.332510829,-0.0941729099,0.0016863931,-0.1492668688,0.0613805987,0.3583227694,-0.3411815464,0.1615508646,-0.0928154215,0.6022837162,0.2045957893,0.340293169,-0.1207564324,0.1436592191,-0.2561024427,-0.4732506871,-0.1315380484,0.3441410363,-0.3749166131,-0.0217502601,0.2255961299,-0.2683691084,-0.1382702738,-0.1619247496,0.2328361571,-0.1926217079,0.4905366302,0.2641061246,0.3150091469,-0.3032971919,-0.2604726851,-0.0365028605,0.2098365873,0.0697660893,0.1154753789,0.1212362573,0.1254822761,0.3096583486,0.1305269748,-0.4090053439,-0.1776621193,-0.1842097491,-0.3713512421,-0.3349807262,0.1222100183,0.1611678451,0.6101635695,-0.196900934,-0.0326289535,-0.186406672,0.0577010587,-0.3059124947,0.0471702963,0.0331944153,0.0651315227,-0.330011487,0.1294766366,-0.1013423875,0.2805507481,0.1138203889,-0.0870753527,-0.0420234352,0.1421462297,-0.5038461089,-0.2322125584,0.3121340871,0.5738974214,-0.2841928303,-0.21173051,-0.2771613002,-0.0527619421,0.3993547857,0.3285742104,-0.1766907871,0.5292114019,-0.1388204098,0.3341922164,0.3454712927,0.0033035409,-0.2640667558,0.3298973739,0.0933336318,-0.0103241792,0.1576756984,0.0565585457,0.3707054555,-0.2029201686,0.0256964229,0.4108987451,0.173025012,-0.0055995649,-0.4627687335,-0.1843359172,0.119093731,0.0089058978,0.4738454819,-0.1573347896,0.2517452538,0.8539651632,0.1662379205,-0.0764204487,0.33806476,0.4670497179,-0.1945731193,-0.0828919709,0.0862237215,0.0015233086,-0.0667467043,0.1740822941,-0.0128847109,0.0722475648,-0.0481515229,-0.0805421323,-0.4336926341,-0.1321930438,0.0781155005,0.2217119485,0.0627875924,0.016171528,0.1256843656,0.4031229019,0.3853701055,-0.286000967,0.0274523254,-0.0765146315,-0.2527941167,0.3991639316,-0.1151903644,0.1357635707,0.1157582551,-0.1102615297,0.4241468608,0.0913995504,-0.0832197741,0.1222256348,0.1489984095,0.2598078549,0.1998369843,-0.2001913786,0.4503698647,-0.10481105,0.3533116877,0.5322104096,0.3143365979,-0.1873884052,-0.0633390397,0.4248915315,-0.0930707902,-0.0543666966,-0.1247329563,0.0032045646,0.4307149053,-0.0187590159,0.2190228403,-0.1407553852,0.2445390075,0.308244884,-0.1421490759,0.1383502781,-0.4233990014,-0.1638499498,0.2858459353,-0.1843141019,0.2433815151,-0.0408219695,-0.2044894099,-0.0603389665,0.2206227481,-0.0591475442,0.1275752634,0.1027461439,0.2328379452,-0.2996247113,0.2365118414,-0.2004734576,0.1167781726,0.2346571535,0.2534106374,0.1489507109,0.0707730874,0.0559613891,-0.1311332434,0.3252004981,0.0034002636,0.3112410605,-0.140868634,-0.1817257404,-0.3317093253,-0.3425328135,-0.2265742719,0.0998420194,-0.3634946942,-0.3806305528,0.1673058271,0.129030928,-0.3707671165,0.038968008,-0.2124536484,-0.1553746164,0.0196424592,0.0361521468,-0.2018184215,-0.2261242121,-0.2062615454,-0.0452431478,0.4151237607,0.2590598464,0.1206606999,-0.1163295135,-0.4247234464,-0.502355814,-0.1835591346,0.0773844197,0.0999757648,0.1563127488,-0.0653547347,0.557911098,-0.0510461107,-0.0452511907,0.1225928068,-0.006120936,-0.0857530013,0.115576826,-0.0713458583,0.1899273247,-0.1455579251,-0.2589744627,-0.1714494675,-0.3056757152,-0.0707055703,-0.0572113432,0.1168038324,-0.2959667146,0.187506333,-0.0181983076,0.2588417828,0.222754389,-0.2554257214,-0.0984865651,0.5742362738,-0.0838482976,-0.4717204869,0.1295849532,-0.1484830528,0.0076137716,0.1879812181,-0.2797017992,0.1219319403,0.1180790365,0.1706920862,0.2602594495,0.0907301903,0.0554289296,-0.2978803813,0.1081660688,-0.052497521,0.2223838568,-0.1497191638,-0.1731573194,0.4568734765,0.3794937134,0.4606593847,0.004209389,0.7833336592,0.0899355188,0.2961522937,0.3146292567,0.4288235903,0.2446366549,0.1249081641,0.0123606371,0.0148987267,0.0630349219,-0.0058769104,0.3074193001,0.064747341,-0.2191385478,-0.32653597,-0.4428035617,-0.3569615781,-0.3261870444,-0.1451663226,0.1818222404,0.2052357346,0.038699653,0.388741225,-0.0583276339,-0.3096982241,0.2100349069,0.2091162652,0.2057548612,0.260368228,-0.2030285597,0.0200098846,-0.6251594424,0.3266809285,0.0376289934,0.4489038587,0.1211346239,-0.0426084921,-0.0064905761,-0.0114129558,0.1911596656,-0.2258455306,0.2486802042,-0.0369804464,0.0049878024,-0.3369857073,0.0306823812,-0.2206840813,0.3259129226,0.5136585832,0.1297414154,-0.6778854728,-0.2252326906,0.0436097607,0.2859038413,-0.0816388056,-0.0003025423,-0.123926267,-0.3683078885,-0.1982917488,-0.0900636092,-0.0220521763,0.1623125672,-0.1682663113,0.2582770884,-0.2203194648,0.3745952845,0.0511454009,0.0659215003,0.2008274049,0.015101417,0.1974331737,0.2343120128,-0.0045970664,0.2846351862,0.3751891851,0.1319918483,-0.287494719,-0.0421271548,-0.0891859084,-0.2077367157,0.2224480957,-0.0696755797,0.2212251127,0.2267780453,-0.1486577541,-0.0479206964,-0.0220695902,0.1886217147,-0.0827186331,-0.3992475569,-0.1423584968,0.4184012413,0.1938906461,-0.0800282583,-0.1462794244,0.1373908669,-0.2048898637,0.0422976799,0.1178171113,0.9017566442,-0.1378929913,0.109958455,-0.0084327003,0.0048325304,0.7497717142,-0.0360592827,0.0383355208,-0.3155221045,0.0912691504,-0.0437765457,0.34152174,0.1126816049,0.035423886,-0.1919229031,0.407677561,0.1949482411,0.2575367689,0.0163744837,0.4060837328,0.0624971353,-0.2345500141,-0.5463211536,0.0462047458,-0.3144664168,0.4104981422,-0.2353578508,-0.2359815687,-0.2067969888,-0.2671512067,-0.0566833653,0.1418094933,-0.0346803144,0.2981498241,-0.0713735968,-0.4402926564,0.1721672416,0.2931363583,0.116397053,0.2252401859,-0.2450677603,0.0552951358,-0.5325859785,-0.3806132972,-0.1697797626,0.1576576829,0.1166551337,-0.0257621184,-0.1954665333,0.0903369784,-0.1123016402,0.1175316721,-0.2212729901,0.0041604582,-0.3834601045,0.0532953143,-0.4173184037,-0.094732374,-0.1572713703,-0.2040213794,0.0688517764,0.1082766131,0.0193427689,-0.1349878758,0.0942509547,-0.0633989125,-0.0277463254,0.5898222327,0.2064320445,0.2121424228,0.4350481629,0.1797177494,-0.4563695788,-0.1493015438,-0.0309250765,-0.5784888864,-0.2278078496,-0.0356225818,-0.2284296751,0.1857256591,-0.2033202648,-0.164228797,0.0463215373,-0.3468266428,0.2272445261,-0.4120407701,-0.1529070139,0.4245547652,0.0086027943,0.0053232713,0.1962230504,-0.1225243509,0.0411097556,0.0502696075,-0.160721004,0.1442897767,-0.2905435264,0.2015252411,0.2590176761,0.0922574624,-0.0789045244,-0.0952034369,-0.0132442918,-0.1834618896,-0.054305993,-0.0117225973,-0.1304046363,0.1948337704,0.1145079732,-0.1623736173,0.2088941932,-0.3206487894,0.1475622654,-0.1380580068,-0.1336162984,0.2696509957,-0.029381793,0.1871213764,0.276697427,0.2042607367,-0.2710437477,0.3788249791,0.0792823806,0.4594862163,-0.0117100403,0.3862805665,0.4423726201,-0.0254813805,-0.2803555727,0.0266867615,-0.0359661989,-0.2174688727,0.3051567674,-0.1816093177,-0.0556699708,0.4315660298,0.1015961617,0.2603603303,-0.0156101221,0.0855963752,0.1956901848,0.0698102564,-0.3646604121,-0.1519405544,0.4377315342,-0.0203535836,-0.1106998622,0.2261767536,0.0932675377,-0.2287782282,0.0955598652,0.1222151667,-0.052952677,-0.0097084111,0.2100512981,0.1276113391,0.4284837246,0.1411617845,0.0625455454,-0.1187034324,0.0098214708,-0.0578228757,-0.0774039254,0.2355452627,0.0742842853,0.1460341811,-0.2558204532,-0.4492249191,-0.0968892351,0.3846179545,-0.2090798169,0.0910647959,-0.0918148085,0.1347708851,0.2871341705,-0.0784622431,-0.2468202263,0.2137849629,-0.1107968986,-0.1801237911,-0.3408865035,-0.3482739329,-0.1132453457,0.1399225444,0.0625814274,-0.3009446859,0.3215266764,0.2096991688,-0.1300755143,-0.5181475878,0.611189425,0.107377544,0.1182978675,-0.3415501416,-0.0192417223,-0.1715460718,-0.112022303,0.0544974133,0.3561193049,0.2328452617,0.198846966,-0.2760933936,-0.0859767646,0.2533798814,-0.0229251385,0.0836348459,0.0606668778,0.1258863509,0.0147506343,0.4033443034,0.0213888064,-0.0314346664,-0.2144175321,0.3113033772,-0.0779477954,-0.0563141219,-0.0922852159,-0.2281106412,-0.0808212087,-0.053799022,-0.1917764992,-0.4238056839,0.0264816396,-0.0189954266,0.1561598331,0.2094827443,-0.1527034342,-0.006661918,-0.2599538863,0.2386309057,0.3384189904,0.0721738413,-0.2710242271,-0.1018828377,-0.4817108214,-0.0005999006,-0.539791584,0.0505526215,-0.0656441078,0.3469389081,-0.0686803833,0.0769366845,0.1008952186,-0.1663620323,0.0166305527,0.1990389377,-0.3084483743,-0.1713161916,-0.1275111288,0.0138703305,-0.2212380916,-0.4771667421,0.282877028,-0.3365028203,-0.0790846944,-0.0288131814,-0.0130413249,0.0311747622,0.2563148141,0.2192016691,0.1050936282,0.1111669987,-0.1393067092,-0.0927668512,-0.2995740473,-0.1683185846,-0.0625778511,0.2928656936,-0.1225750297,0.475065887,-0.1518029422,0.1282854974,-0.1748405695,0.4684761763,-0.2808565795,0.1735428423,-0.1457725465,0.0948005244,-0.3616304696,0.0534701273,-0.0626029745,0.013030963,-0.2021164447,0.1032494903,-0.4907825291,-0.153317526,0.184914887,-0.2853300869,-0.1076995805,-0.2806595266,0.0994234383,0.2337026745,0.2223946005,-0.4161421061,0.1807072908,0.2665099502,-0.0033004391,-0.091663219,-0.078995645,-0.0880528092,-0.3081294894,-0.0442834236,-0.0599305257,0.0284393914,-0.4647625089,-0.0302373897,-0.2713389695]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/784","title":"Issue with downloading Wikipedia data for low resource language","comments":"@lhoestq\r\n\r\nI've tried `load_dataset('wikipedia', '20200501.zh', beam_runner='DirectRunner')` and got the same `FileNotFoundError` as @SamuelCahyawijaya.\r\n\r\nAlso, using another date (e.g. `load_dataset('wikipedia', '20201120.zh', beam_runner='DirectRunner')`) will give the following error message.\r\n\r\n```\r\nValueError: BuilderConfig 20201120.zh not found. Available: ['20200501.aa', '20200501.ab', '20200501.ace', '20200501.ady', '20200501.af', '20200501.ak', '20200501.als', '20200501.am', '20200501.an', '20200501.ang', '20200501.ar', '20200501.arc', '20200501.arz', '20200501.as', '20200501.ast', '20200501.atj', '20200501.av', '20200501.ay', '20200501.az', '20200501.azb', '20200501.ba', '20200501.bar', '20200501.bat-smg', '20200501.bcl', '20200501.be', '20200501.be-x-old', '20200501.bg', '20200501.bh', '20200501.bi', '20200501.bjn', '20200501.bm', '20200501.bn', '20200501.bo', '20200501.bpy', '20200501.br', '20200501.bs', '20200501.bug', '20200501.bxr', '20200501.ca', '20200501.cbk-zam', '20200501.cdo', '20200501.ce', '20200501.ceb', '20200501.ch', '20200501.cho', '20200501.chr', '20200501.chy', '20200501.ckb', '20200501.co', '20200501.cr', '20200501.crh', '20200501.cs', '20200501.csb', '20200501.cu', '20200501.cv', '20200501.cy', '20200501.da', '20200501.de', '20200501.din', '20200501.diq', '20200501.dsb', '20200501.dty', '20200501.dv', '20200501.dz', '20200501.ee', '20200501.el', '20200501.eml', '20200501.en', '20200501.eo', '20200501.es', '20200501.et', '20200501.eu', '20200501.ext', '20200501.fa', '20200501.ff', '20200501.fi', '20200501.fiu-vro', '20200501.fj', '20200501.fo', '20200501.fr', '20200501.frp', '20200501.frr', '20200501.fur', '20200501.fy', '20200501.ga', '20200501.gag', '20200501.gan', '20200501.gd', '20200501.gl', '20200501.glk', '20200501.gn', '20200501.gom', '20200501.gor', '20200501.got', '20200501.gu', '20200501.gv', '20200501.ha', '20200501.hak', '20200501.haw', '20200501.he', '20200501.hi', '20200501.hif', '20200501.ho', '20200501.hr', '20200501.hsb', '20200501.ht', '20200501.hu', '20200501.hy', '20200501.ia', '20200501.id', '20200501.ie', '20200501.ig', '20200501.ii', '20200501.ik', '20200501.ilo', '20200501.inh', '20200501.io', '20200501.is', '20200501.it', '20200501.iu', '20200501.ja', '20200501.jam', '20200501.jbo', '20200501.jv', '20200501.ka', '20200501.kaa', '20200501.kab', '20200501.kbd', '20200501.kbp', '20200501.kg', '20200501.ki', '20200501.kj', '20200501.kk', '20200501.kl', '20200501.km', '20200501.kn', '20200501.ko', '20200501.koi', '20200501.krc', '20200501.ks', '20200501.ksh', '20200501.ku', '20200501.kv', '20200501.kw', '20200501.ky', '20200501.la', '20200501.lad', '20200501.lb', '20200501.lbe', '20200501.lez', '20200501.lfn', '20200501.lg', '20200501.li', '20200501.lij', '20200501.lmo', '20200501.ln', '20200501.lo', '20200501.lrc', '20200501.lt', '20200501.ltg', '20200501.lv', '20200501.mai', '20200501.map-bms', '20200501.mdf', '20200501.mg', '20200501.mh', '20200501.mhr', '20200501.mi', '20200501.min', '20200501.mk', '20200501.ml', '20200501.mn', '20200501.mr', '20200501.mrj', '20200501.ms', '20200501.mt', '20200501.mus', '20200501.mwl', '20200501.my', '20200501.myv', '20200501.mzn', '20200501.na', '20200501.nah', '20200501.nap', '20200501.nds', '20200501.nds-nl', '20200501.ne', '20200501.new', '20200501.ng', '20200501.nl', '20200501.nn', '20200501.no', '20200501.nov', '20200501.nrm', '20200501.nso', '20200501.nv', '20200501.ny', '20200501.oc', '20200501.olo', '20200501.om', '20200501.or', '20200501.os', '20200501.pa', '20200501.pag', '20200501.pam', '20200501.pap', '20200501.pcd', '20200501.pdc', '20200501.pfl', '20200501.pi', '20200501.pih', '20200501.pl', '20200501.pms', '20200501.pnb', '20200501.pnt', '20200501.ps', '20200501.pt', '20200501.qu', '20200501.rm', '20200501.rmy', '20200501.rn', '20200501.ro', '20200501.roa-rup', '20200501.roa-tara', '20200501.ru', '20200501.rue', '20200501.rw', '20200501.sa', '20200501.sah', '20200501.sat', '20200501.sc', '20200501.scn', '20200501.sco', '20200501.sd', '20200501.se', '20200501.sg', '20200501.sh', '20200501.si', '20200501.simple', '20200501.sk', '20200501.sl', '20200501.sm', '20200501.sn', '20200501.so', '20200501.sq', '20200501.sr', '20200501.srn', '20200501.ss', '20200501.st', '20200501.stq', '20200501.su', '20200501.sv', '20200501.sw', '20200501.szl', '20200501.ta', '20200501.tcy', '20200501.te', '20200501.tet', '20200501.tg', '20200501.th', '20200501.ti', '20200501.tk', '20200501.tl', '20200501.tn', '20200501.to', '20200501.tpi', '20200501.tr', '20200501.ts', '20200501.tt', '20200501.tum', '20200501.tw', '20200501.ty', '20200501.tyv', '20200501.udm', '20200501.ug', '20200501.uk', '20200501.ur', '20200501.uz', '20200501.ve', '20200501.vec', '20200501.vep', '20200501.vi', '20200501.vls', '20200501.vo', '20200501.wa', '20200501.war', '20200501.wo', '20200501.wuu', '20200501.xal', '20200501.xh', '20200501.xmf', '20200501.yi', '20200501.yo', '20200501.za', '20200501.zea', '20200501.zh', '20200501.zh-classical', '20200501.zh-min-nan', '20200501.zh-yue', '20200501.zu']\r\n```\r\n\r\nI am pretty sure that `https:\/\/dumps.wikimedia.org\/enwiki\/20201120\/dumpstatus.json` exists.","body":"Hi, I tried to download Sundanese and Javanese wikipedia data with the following snippet\r\n```\r\njv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner')\r\nsu_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner')\r\n```\r\nAnd I get the following error for these two languages:\r\nJavanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nSundanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nI found from https:\/\/github.com\/huggingface\/datasets\/issues\/577#issuecomment-688435085 that for small languages, they are directly downloaded and parsed from the Wikipedia dump site, but both of `https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json` and `https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json` are no longer valid.\r\n\r\n Any suggestions on how to handle this issue? Thanks!","comment_length":342,"text":"Issue with downloading Wikipedia data for low resource language\nHi, I tried to download Sundanese and Javanese wikipedia data with the following snippet\r\n```\r\njv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner')\r\nsu_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner')\r\n```\r\nAnd I get the following error for these two languages:\r\nJavanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nSundanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nI found from https:\/\/github.com\/huggingface\/datasets\/issues\/577#issuecomment-688435085 that for small languages, they are directly downloaded and parsed from the Wikipedia dump site, but both of `https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json` and `https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json` are no longer valid.\r\n\r\n Any suggestions on how to handle this issue? Thanks!\n@lhoestq\r\n\r\nI've tried `load_dataset('wikipedia', '20200501.zh', beam_runner='DirectRunner')` and got the same `FileNotFoundError` as @SamuelCahyawijaya.\r\n\r\nAlso, using another date (e.g. `load_dataset('wikipedia', '20201120.zh', beam_runner='DirectRunner')`) will give the following error message.\r\n\r\n```\r\nValueError: BuilderConfig 20201120.zh not found. Available: ['20200501.aa', '20200501.ab', '20200501.ace', '20200501.ady', '20200501.af', '20200501.ak', '20200501.als', '20200501.am', '20200501.an', '20200501.ang', '20200501.ar', '20200501.arc', '20200501.arz', '20200501.as', '20200501.ast', '20200501.atj', '20200501.av', '20200501.ay', '20200501.az', '20200501.azb', '20200501.ba', '20200501.bar', '20200501.bat-smg', '20200501.bcl', '20200501.be', '20200501.be-x-old', '20200501.bg', '20200501.bh', '20200501.bi', '20200501.bjn', '20200501.bm', '20200501.bn', '20200501.bo', '20200501.bpy', '20200501.br', '20200501.bs', '20200501.bug', '20200501.bxr', '20200501.ca', '20200501.cbk-zam', '20200501.cdo', '20200501.ce', '20200501.ceb', '20200501.ch', '20200501.cho', '20200501.chr', '20200501.chy', '20200501.ckb', '20200501.co', '20200501.cr', '20200501.crh', '20200501.cs', '20200501.csb', '20200501.cu', '20200501.cv', '20200501.cy', '20200501.da', '20200501.de', '20200501.din', '20200501.diq', '20200501.dsb', '20200501.dty', '20200501.dv', '20200501.dz', '20200501.ee', '20200501.el', '20200501.eml', '20200501.en', '20200501.eo', '20200501.es', '20200501.et', '20200501.eu', '20200501.ext', '20200501.fa', '20200501.ff', '20200501.fi', '20200501.fiu-vro', '20200501.fj', '20200501.fo', '20200501.fr', '20200501.frp', '20200501.frr', '20200501.fur', '20200501.fy', '20200501.ga', '20200501.gag', '20200501.gan', '20200501.gd', '20200501.gl', '20200501.glk', '20200501.gn', '20200501.gom', '20200501.gor', '20200501.got', '20200501.gu', '20200501.gv', '20200501.ha', '20200501.hak', '20200501.haw', '20200501.he', '20200501.hi', '20200501.hif', '20200501.ho', '20200501.hr', '20200501.hsb', '20200501.ht', '20200501.hu', '20200501.hy', '20200501.ia', '20200501.id', '20200501.ie', '20200501.ig', '20200501.ii', '20200501.ik', '20200501.ilo', '20200501.inh', '20200501.io', '20200501.is', '20200501.it', '20200501.iu', '20200501.ja', '20200501.jam', '20200501.jbo', '20200501.jv', '20200501.ka', '20200501.kaa', '20200501.kab', '20200501.kbd', '20200501.kbp', '20200501.kg', '20200501.ki', '20200501.kj', '20200501.kk', '20200501.kl', '20200501.km', '20200501.kn', '20200501.ko', '20200501.koi', '20200501.krc', '20200501.ks', '20200501.ksh', '20200501.ku', '20200501.kv', '20200501.kw', '20200501.ky', '20200501.la', '20200501.lad', '20200501.lb', '20200501.lbe', '20200501.lez', '20200501.lfn', '20200501.lg', '20200501.li', '20200501.lij', '20200501.lmo', '20200501.ln', '20200501.lo', '20200501.lrc', '20200501.lt', '20200501.ltg', '20200501.lv', '20200501.mai', '20200501.map-bms', '20200501.mdf', '20200501.mg', '20200501.mh', '20200501.mhr', '20200501.mi', '20200501.min', '20200501.mk', '20200501.ml', '20200501.mn', '20200501.mr', '20200501.mrj', '20200501.ms', '20200501.mt', '20200501.mus', '20200501.mwl', '20200501.my', '20200501.myv', '20200501.mzn', '20200501.na', '20200501.nah', '20200501.nap', '20200501.nds', '20200501.nds-nl', '20200501.ne', '20200501.new', '20200501.ng', '20200501.nl', '20200501.nn', '20200501.no', '20200501.nov', '20200501.nrm', '20200501.nso', '20200501.nv', '20200501.ny', '20200501.oc', '20200501.olo', '20200501.om', '20200501.or', '20200501.os', '20200501.pa', '20200501.pag', '20200501.pam', '20200501.pap', '20200501.pcd', '20200501.pdc', '20200501.pfl', '20200501.pi', '20200501.pih', '20200501.pl', '20200501.pms', '20200501.pnb', '20200501.pnt', '20200501.ps', '20200501.pt', '20200501.qu', '20200501.rm', '20200501.rmy', '20200501.rn', '20200501.ro', '20200501.roa-rup', '20200501.roa-tara', '20200501.ru', '20200501.rue', '20200501.rw', '20200501.sa', '20200501.sah', '20200501.sat', '20200501.sc', '20200501.scn', '20200501.sco', '20200501.sd', '20200501.se', '20200501.sg', '20200501.sh', '20200501.si', '20200501.simple', '20200501.sk', '20200501.sl', '20200501.sm', '20200501.sn', '20200501.so', '20200501.sq', '20200501.sr', '20200501.srn', '20200501.ss', '20200501.st', '20200501.stq', '20200501.su', '20200501.sv', '20200501.sw', '20200501.szl', '20200501.ta', '20200501.tcy', '20200501.te', '20200501.tet', '20200501.tg', '20200501.th', '20200501.ti', '20200501.tk', '20200501.tl', '20200501.tn', '20200501.to', '20200501.tpi', '20200501.tr', '20200501.ts', '20200501.tt', '20200501.tum', '20200501.tw', '20200501.ty', '20200501.tyv', '20200501.udm', '20200501.ug', '20200501.uk', '20200501.ur', '20200501.uz', '20200501.ve', '20200501.vec', '20200501.vep', '20200501.vi', '20200501.vls', '20200501.vo', '20200501.wa', '20200501.war', '20200501.wo', '20200501.wuu', '20200501.xal', '20200501.xh', '20200501.xmf', '20200501.yi', '20200501.yo', '20200501.za', '20200501.zea', '20200501.zh', '20200501.zh-classical', '20200501.zh-min-nan', '20200501.zh-yue', '20200501.zu']\r\n```\r\n\r\nI am pretty sure that `https:\/\/dumps.wikimedia.org\/enwiki\/20201120\/dumpstatus.json` exists.","embeddings":[0.0533478856,0.0374849103,0.0597833283,0.4869290888,0.170536682,0.0233596079,0.093230091,0.4074784219,0.1670001149,0.1014821902,-0.180892691,-0.1391495764,0.4188603461,-0.1732718199,-0.0940446556,-0.495067656,0.1877741963,0.0399843417,0.0126653081,-0.0655117556,-0.2115678489,-0.0480744801,-0.1926138699,0.0627112836,-0.3057098985,-0.0056309984,0.0016467076,-0.0364647955,-0.3302689195,-0.2426016182,0.2610393167,0.1820007414,0.2705383003,0.414835304,-0.0001278781,-0.0497068129,0.5732793212,-0.1151508689,-0.5644016862,0.088967219,-0.2109943181,-0.4145148098,-0.1965319067,-0.3739251792,0.0392361134,-0.1627885997,0.234076649,-0.2813850939,0.2931888998,-0.0290631242,0.0679255798,0.1302415282,0.1000647396,-0.2170469314,0.3911059499,0.1138225943,0.0463740714,0.0177767929,-0.0791449174,0.0869459212,0.0281140544,0.215274632,-0.2443998754,-0.0413640589,0.211182043,-0.0207845569,-0.2214757651,-0.5394378901,0.4275811613,0.5065014362,0.9873847961,-0.0341156051,-0.5420722365,0.0208634101,-0.0884430483,0.1717970371,0.2517359555,0.0773003176,-0.0216554496,0.0064187408,0.1908419132,-0.3481397629,-0.1752084047,0.5972660184,-0.0458157919,0.3435457349,-0.0293708425,0.1244830489,-0.1325082183,-0.4068507254,-0.196453318,-0.4271188676,-0.0505387969,0.4434016645,0.0007426016,0.0057797455,-0.1001872495,0.0649934188,0.1602074206,-0.2895126045,-0.0883787498,0.0405808762,-0.1443123817,0.1057483703,0.2951692939,-0.2529043257,0.1639977545,-0.1666972041,0.6046065092,0.1877863258,0.3268435895,-0.0606869645,0.0451207832,-0.2781887949,-0.5877600908,-0.1193340272,0.2341959774,-0.3487767875,-0.0475725196,0.1888379306,-0.2759740651,-0.1863067895,-0.149003014,0.240567416,-0.2268441468,0.4528813958,0.3064103425,0.2869040668,-0.3019836545,-0.2889588773,-0.0044378703,0.2404556721,0.0577961206,0.1191031039,0.1557949781,0.0076080612,0.2810476422,0.1295275837,-0.3599660993,-0.1727759838,-0.2213955969,-0.2736275196,-0.328809917,0.1263093203,0.1419122815,0.5926673412,-0.1778511405,-0.1102340296,-0.1915977746,0.059934184,-0.3075916469,0.0290118828,-0.0388549939,0.0066995448,-0.3058397472,0.1448859721,-0.1023367941,0.2594457269,0.0931876302,-0.1298283637,-0.0810647085,0.1333623379,-0.4865556359,-0.2310166508,0.4018515944,0.6183560491,-0.3419445157,-0.2550344169,-0.2246883363,-0.0379407294,0.3226376474,0.2405833751,-0.1962711364,0.5838981271,-0.1156998426,0.1431093365,0.3990396261,-0.0626019314,-0.245179534,0.3416454196,0.1305207908,-0.0190299917,0.1855218709,0.0644040853,0.2784171104,-0.1964817494,0.0487367027,0.407776475,0.1818939,-0.0078275073,-0.3822375834,-0.2369911075,0.1355568171,0.074671343,0.4870291948,-0.0940967575,0.2486828268,0.8490917087,0.2026953995,-0.0856895447,0.3260680437,0.4646955431,-0.3252147436,-0.0939502642,0.0796766654,0.0160330478,-0.0819027573,0.2273687869,-0.0334083624,0.044444833,0.0221871193,-0.0124820368,-0.3875280917,-0.1059804931,0.0293927956,0.2333385944,0.0024630348,0.0668149292,0.1838377267,0.3588463366,0.3215928078,-0.1883118004,0.0001829717,-0.0873958021,-0.2711069286,0.4411354959,-0.0971421897,0.1178860739,0.1242606044,-0.0658245459,0.4569521844,0.037395373,-0.0878356472,0.1333163679,0.1280237138,0.2300779521,0.09697669,-0.1452700645,0.4124618769,-0.0279227216,0.3956124485,0.5580114722,0.2673858702,-0.1624314636,-0.0820200741,0.4217672944,-0.0235179048,0.0680852756,-0.0762696937,0.0021344603,0.3455212712,-0.0045456886,0.2148859799,-0.1222809702,0.2748683393,0.3064382374,-0.1189002544,0.1761002094,-0.4096235335,-0.140421316,0.3270944655,-0.1661039889,0.2058768868,-0.0424847528,-0.2158710212,-0.0264976043,0.250402689,-0.0422125235,0.0219621379,0.093374446,0.1414162815,-0.2690082192,0.185656935,-0.1909544021,0.117873475,0.2861928344,0.361367017,0.2144424319,0.0326790139,0.0631784126,-0.0983968005,0.270280093,0.0177615173,0.2583781481,-0.1833215505,-0.1620476097,-0.3506138921,-0.3741486073,-0.287103951,0.0854783803,-0.3898091018,-0.3666174412,0.1491402388,0.1553399563,-0.2938445508,0.132437095,-0.1568155736,-0.1100076362,-0.0155158909,-0.0630766451,-0.2139376104,-0.2459361553,-0.2690545022,-0.1106629074,0.4098207057,0.1954005808,0.0575194061,-0.1600296199,-0.384940505,-0.5117307901,-0.194742322,0.1738581508,0.0831749812,0.178712666,-0.0822415948,0.5658929348,0.0030056352,-0.0505431034,0.1183449998,0.0240438879,-0.0823360533,0.1118144244,-0.1079549119,0.1611085385,-0.1291732192,-0.2555248737,-0.1848215312,-0.354953289,-0.1326247156,0.0402967595,0.1444684267,-0.1802950799,0.1903788745,-0.0165350512,0.3194538653,0.2845262885,-0.1832271665,-0.141690895,0.5626681447,-0.0918639824,-0.4536193013,0.1298017651,-0.1563428491,-0.0148345353,0.1588099748,-0.3009743094,0.2276768535,0.1487456262,0.1069899201,0.248714596,0.0975349173,0.0849275887,-0.2770103514,0.1548021734,-0.0610632561,0.1394795179,-0.1391167939,-0.1906077415,0.4748793244,0.4391957521,0.4500229359,-0.0618055724,0.7416902184,0.1354653686,0.2463316023,0.3136912286,0.3680102527,0.2478992045,0.0193921328,0.0227028877,0.1015528291,0.127037257,0.04435708,0.3230541945,0.0436069481,-0.1844170839,-0.391053766,-0.4620245695,-0.2502823174,-0.3246948421,-0.1803740263,0.0865729302,0.2444776893,0.0344703496,0.4447145164,-0.1163271815,-0.3504458368,0.2178046405,0.1891652346,0.2316963524,0.2817645967,-0.1259294897,-0.0661058724,-0.5629535317,0.3282524943,0.0698234439,0.4110635221,0.0179468505,-0.0059716702,0.0319736339,-0.027553089,0.2141028196,-0.2173385918,0.2254904509,0.0406879708,-0.0304436497,-0.4630871713,-0.0115974536,-0.1469059139,0.3091903627,0.4478082359,0.1146976575,-0.693402946,-0.3136429191,0.0711742193,0.35815534,-0.1145717949,0.0238104817,-0.1017409712,-0.2344863713,-0.3066959381,-0.1252878457,-0.1091524363,0.22679618,-0.125094682,0.2264004499,-0.1811089367,0.3791970015,0.0265087951,0.0661102086,0.2037884295,0.0909418613,0.1747338325,0.2588312626,0.0616421103,0.2426662594,0.331539005,0.1963541508,-0.1387943029,0.0138742402,-0.1242533848,-0.1976689994,0.2619448304,-0.0786775053,0.1416076869,0.2371560037,-0.1181489006,-0.0922538415,-0.0754841268,0.1806946695,-0.1116924286,-0.4302593172,-0.1690327674,0.4821285605,0.184679538,-0.0995531604,-0.0832197964,0.1816089898,-0.1851888001,0.0870919377,0.1557226926,0.9597741365,-0.1339156926,0.139233917,-0.0139016472,0.0420069546,0.7529662848,-0.1289315671,0.0568466447,-0.3687592149,0.1460504532,-0.0803370103,0.3275386989,0.1196122319,0.034181647,-0.1696020961,0.4756477177,0.173636809,0.2811532915,-0.009662535,0.4018622041,0.0217164531,-0.2601259947,-0.4872604012,-0.007310058,-0.293237716,0.4567648768,-0.2152505368,-0.2132642865,-0.1895176321,-0.2625997066,-0.08690992,0.1909972727,-0.0871391371,0.3240563273,-0.0048374971,-0.4457882643,0.1779010743,0.3298443556,0.1616213471,0.1972560287,-0.2407045811,0.0811973587,-0.5297273397,-0.3864913881,-0.1677397788,0.1796525121,0.2071367353,-0.0624726899,-0.2309261858,0.1035086811,-0.1793391407,0.0926469937,-0.2239650935,-0.0617022067,-0.2331023365,0.1321616769,-0.3695795834,-0.0716166943,-0.1525004655,-0.202937156,0.0266377646,0.035256587,0.033963453,-0.1284422725,0.0965876877,-0.0923717469,0.0047385637,0.5840378404,0.1734890193,0.2268537432,0.4951579273,0.2195609957,-0.4367139935,-0.1338247061,-0.0482272208,-0.4838817418,-0.2081249356,-0.0898371413,-0.1886025816,0.1479988247,-0.1888006032,0.0197453648,-0.0024591563,-0.3703350127,0.1932285726,-0.4431479573,-0.1510879099,0.3806108534,-0.0277295001,0.0423925593,0.2279711068,-0.0980442092,0.0965980366,-0.0128570171,-0.1124013066,0.0797955841,-0.3406732082,0.1945511848,0.3010063171,0.1971975416,-0.1028865427,-0.0432373025,-0.0749794915,-0.2692519128,-0.0939339921,0.0452760458,-0.1220032051,0.2240490019,0.1475181431,-0.1261536926,0.2464759052,-0.3998472989,0.0873426571,-0.1955201775,-0.0513404869,0.1589803845,-0.0559088066,0.1787536591,0.1960092187,0.2117496729,-0.3149054646,0.3788196445,0.1158760339,0.5302122235,-0.0128993513,0.3653714657,0.3719911575,-0.0265790801,-0.36284253,0.1002208143,-0.0864365473,-0.2152409703,0.3673942387,-0.2350914776,-0.0242953729,0.3940367401,0.0838518441,0.215974912,-0.0255265255,0.0932177827,0.1395023614,-0.0043542832,-0.3624985814,-0.1738433093,0.5092694163,-0.0150032071,-0.0885135084,0.264208138,0.1587806493,-0.2973440588,0.1971719414,0.1173843443,-0.0105728591,-0.110145621,0.1757884473,0.1708860695,0.4562654197,0.1897693425,0.0589323826,-0.0892682597,0.0142996516,0.0067839795,-0.0889343098,0.2076757997,0.0953658149,0.1460134536,-0.2269896418,-0.4386489987,0.0295362771,0.3851924241,-0.1301937103,-0.0124926344,-0.1177861392,0.1243198961,0.2882288396,-0.0849068761,-0.3104221523,0.2391562015,-0.097833097,-0.1753344536,-0.4473358095,-0.3883261383,-0.0926381797,0.1555445343,0.1127262488,-0.2999770641,0.2941488624,0.2409376949,-0.1209258586,-0.5600200295,0.6237092614,0.1035370454,0.0964279845,-0.324909389,0.1133992523,-0.2321376354,-0.0669260323,0.0839915574,0.3659837246,0.2763702273,0.1966903657,-0.3062681556,-0.0782816112,0.2343600392,-0.0261938423,0.0934021473,0.0172870122,0.0709710419,-0.003841352,0.4331744909,-0.0164175741,0.0059205755,-0.2877863944,0.3169848025,-0.0339478776,-0.0657152981,-0.0700167343,-0.1899706423,-0.0451916978,-0.0729158521,-0.1918238401,-0.4208929241,0.0570517294,-0.0258307047,0.2027056813,0.1951415986,-0.1792075485,-0.0339696221,-0.2826877534,0.259136945,0.2360923141,0.0933158025,-0.2518785894,-0.1341605484,-0.4866870344,-0.0587203614,-0.4086301625,0.0708172992,-0.0541622117,0.2985700965,-0.042791903,0.054275468,0.1404776573,-0.205230549,0.0477352701,0.1422234327,-0.3190624118,-0.1981557161,-0.0381391086,0.0567876697,-0.2491871119,-0.4527922273,0.2786127031,-0.3910551667,-0.1276456863,0.0341988616,-0.0678333342,0.0969120637,0.1439946741,0.2436262071,0.019980032,0.1572601199,-0.1167752072,-0.0487134196,-0.2996519208,-0.1257106066,-0.0596666187,0.2554554939,-0.1438671052,0.5094142556,-0.2220048159,0.1459827274,-0.245387882,0.3832649887,-0.2661608756,0.202036351,-0.1202317551,0.0843218938,-0.4251119494,0.1710017622,-0.0190150235,0.0046252129,-0.1791450083,0.1041809842,-0.5216527581,-0.0987187698,0.231131658,-0.2941542268,-0.1555873305,-0.2660243511,0.0603289157,0.2114711851,0.1888337582,-0.3588400781,0.2254001498,0.2415476292,-0.0320400856,-0.1188902706,-0.1337679625,-0.1148092002,-0.2388495803,-0.0629512668,-0.1781378835,0.0128491204,-0.4606752694,-0.1186596304,-0.2578940988]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/784","title":"Issue with downloading Wikipedia data for low resource language","comments":"Thanks for reporting I created a PR to make the custom config work (language=\"zh\", date=\"20201120\").","body":"Hi, I tried to download Sundanese and Javanese wikipedia data with the following snippet\r\n```\r\njv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner')\r\nsu_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner')\r\n```\r\nAnd I get the following error for these two languages:\r\nJavanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nSundanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nI found from https:\/\/github.com\/huggingface\/datasets\/issues\/577#issuecomment-688435085 that for small languages, they are directly downloaded and parsed from the Wikipedia dump site, but both of `https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json` and `https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json` are no longer valid.\r\n\r\n Any suggestions on how to handle this issue? Thanks!","comment_length":15,"text":"Issue with downloading Wikipedia data for low resource language\nHi, I tried to download Sundanese and Javanese wikipedia data with the following snippet\r\n```\r\njv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner')\r\nsu_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner')\r\n```\r\nAnd I get the following error for these two languages:\r\nJavanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nSundanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nI found from https:\/\/github.com\/huggingface\/datasets\/issues\/577#issuecomment-688435085 that for small languages, they are directly downloaded and parsed from the Wikipedia dump site, but both of `https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json` and `https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json` are no longer valid.\r\n\r\n Any suggestions on how to handle this issue? Thanks!\nThanks for reporting I created a PR to make the custom config work (language=\"zh\", date=\"20201120\").","embeddings":[0.1526538581,-0.1061289981,0.0755999908,0.4940117896,0.1789224297,0.0055683325,0.0679008141,0.3676909208,0.1133994088,0.0056751757,-0.3016711175,-0.17044954,0.3786817789,-0.1107292846,-0.0672977492,-0.4383374751,0.1048311144,0.0070813135,0.0832636207,-0.0598160289,-0.1867440492,0.035805285,-0.1820746362,0.0552036837,-0.1861671209,0.0082284706,0.0775164366,0.0408576243,-0.2773841023,-0.2326842248,0.3428219557,0.2690679729,0.2624540627,0.3764817715,-0.0001280552,-0.0187113006,0.5662430525,-0.1834860295,-0.4274104834,0.0950750858,-0.1335455626,-0.3950597048,-0.2378305942,-0.3481878638,0.0490779541,-0.1722506583,0.192558825,-0.2668434381,0.2568952739,-0.0115212118,0.0862855241,0.130633384,-0.0527203716,-0.216627419,0.3638321757,0.2597155869,0.030824462,-0.0217191279,-0.1367985457,0.0430002138,0.0354902558,0.1803687513,-0.2397564501,-0.0631839707,0.172988832,0.0221187063,-0.3905936182,-0.4756036401,0.4972803891,0.580232203,0.9011561275,-0.0553264245,-0.4694180191,-0.001766557,-0.0843804404,0.1632567644,0.3252804875,0.1311999261,-0.0291333683,0.0777615383,0.1493345499,-0.3046934605,-0.2257986218,0.5692602992,-0.0370111987,0.3623142242,-0.022015145,0.080434911,-0.1243974864,-0.4102575481,-0.16132541,-0.3531893492,-0.1563144326,0.3891967833,0.0133657092,-0.0071604438,-0.0576971099,0.0094355568,0.2346815318,-0.312382102,-0.0551041886,0.0997328237,-0.1300583333,0.0554276407,0.3217167854,-0.3150476813,0.1741762906,-0.0797744393,0.6716029048,0.2427786589,0.3484408855,-0.0548691489,0.165194869,-0.2500068247,-0.6555879712,-0.2368738204,0.251689285,-0.3539804518,0.029356014,0.237207666,-0.2949358523,-0.1471674144,-0.1304592341,0.2185590416,-0.1626968235,0.4253706038,0.2704122961,0.2700175345,-0.3403922319,-0.1425594091,0.0183685794,0.199084416,0.0245516971,0.1895290464,0.2227441818,0.1596432924,0.3807292283,0.122325249,-0.2994661033,-0.1245596558,-0.2987483442,-0.3253648281,-0.356089592,0.1442196369,0.151744619,0.5440235138,-0.1746984869,-0.1287314892,-0.2352216393,0.0503557399,-0.297629863,-0.0180745199,-0.0041210465,-0.0009796439,-0.4188370109,0.0512231924,-0.0438098274,0.3929765821,0.0905742124,-0.1045627519,-0.0197997745,0.1071925387,-0.5476297736,-0.2153519094,0.3619441688,0.6867630482,-0.4186421931,-0.2224641144,-0.2167265415,-0.0864354968,0.4158600271,0.2858968973,-0.2329926938,0.5622168779,-0.1678168178,0.3130166531,0.3979900777,-0.0334052108,-0.2916879058,0.2905941606,0.1531894505,0.0726159364,0.2077383995,0.0213250685,0.3479525149,-0.1400873363,0.0649015009,0.4259873927,0.2378898859,-0.0425093956,-0.4124343097,-0.1760639399,0.0159434807,0.0525507033,0.5028327703,-0.1079766974,0.2653234899,0.7665739655,0.195831567,-0.0950566828,0.3388813436,0.4378651083,-0.1937485933,-0.0389747433,0.0712459311,0.0431113653,-0.0595117249,0.2998177409,-0.0650679618,0.1620691568,0.0010746504,0.0017704942,-0.4758811891,-0.1993391514,0.0123531595,0.0779741704,0.0188493114,0.0723759755,0.0817548484,0.3199214637,0.2967854142,-0.1952908784,-0.0325038359,-0.1245645508,-0.2483137697,0.3950290978,-0.0441408232,0.1025192961,0.0570982955,-0.084241651,0.3936720192,0.0966292322,-0.1037098691,0.1699816734,0.1470469087,0.2012401819,0.2138874382,-0.1969036907,0.4289682209,-0.1039944217,0.3915771842,0.472628355,0.262812376,-0.1848152429,-0.2010032237,0.4290519059,-0.0746011063,0.024366729,-0.1253179908,-0.0187888723,0.4589298069,-0.028729815,0.1646021158,-0.096668072,0.187578395,0.3348905146,-0.1116193235,0.1313783675,-0.5358742476,-0.1226341054,0.3183752,-0.1486278772,0.194128871,0.0456859767,-0.1726856232,-0.0190993603,0.2978651822,0.0274029709,0.2479238063,0.1079477295,0.1968321353,-0.1569529921,0.1614483595,-0.1754125655,0.1646887958,0.1803215146,0.2451460063,0.2604564726,0.0862060189,-0.0073769279,-0.181640476,0.2493368238,-0.0781705976,0.2403332442,-0.273628056,-0.1741499305,-0.3966381848,-0.3204063475,-0.2519058287,0.070106931,-0.4154380262,-0.3010443449,0.1070255339,0.0418841541,-0.2741851211,0.121706441,-0.1756345779,-0.1092248484,-0.0279711764,-0.2199746519,-0.1873394102,-0.160188511,-0.2678100169,-0.1040257663,0.3693732619,0.2846460044,0.0591353998,-0.0695960894,-0.4899002016,-0.4523572624,-0.20771873,0.1102852896,0.073571302,0.2013477534,0.0067345588,0.5252380967,-0.0593651608,0.0185865778,0.1463052928,0.0544725135,-0.0795837641,0.1357005686,-0.0678358674,0.1778654158,-0.1691139191,-0.3163207173,-0.1811513007,-0.3348080516,-0.0808606297,0.0410169624,0.1589318514,-0.2114672661,0.2881553471,0.0445523337,0.270445168,0.148271814,-0.2233910859,-0.1568760723,0.6194942594,-0.1082798094,-0.3799785078,0.1814606041,-0.0872298703,0.0460804887,0.2495883107,-0.3366976082,0.2148473263,0.0906118155,0.2380648553,0.1813330352,0.1697121412,0.1567647457,-0.2488587201,0.1324440837,-0.0680426061,0.0462311655,-0.1115966961,-0.1424637884,0.4910603166,0.4675517082,0.4277179241,-0.0231369827,0.7972062826,0.0853199959,0.3377407193,0.3522926867,0.3237594366,0.3205472529,0.0526934303,-0.0274001192,0.0346012823,0.1011268497,0.0354676656,0.3481436074,0.0660607591,-0.2073579133,-0.3829086125,-0.4077340662,-0.2719334066,-0.3348489106,-0.0900258049,0.2140910923,0.2052574456,-0.0314113945,0.4209453464,-0.0708308518,-0.2926141322,0.1936920136,0.221112445,0.0754745007,0.3169983923,-0.168396771,0.0019463063,-0.5844843388,0.2997379005,-0.0237161312,0.4526117444,0.0433058962,-0.009534358,0.0088541061,-0.0191756785,0.2007117569,-0.2794329524,0.1884330064,0.0152765838,-0.0070474125,-0.3988235295,-0.0351090617,-0.1534910798,0.2801508904,0.5265550017,0.2765311897,-0.6955715418,-0.2446588874,0.0546150431,0.2202473432,-0.0744250417,-0.0385931581,-0.1163214222,-0.4328581095,-0.2004733682,-0.1105900332,-0.0892352462,0.1428891867,-0.1833442897,0.2319131494,-0.1385722905,0.3117272258,0.0880182758,0.0756346881,0.1251627803,0.0624739416,0.1128011197,0.2267530411,-0.0117184557,0.2272232622,0.4371179938,0.1746421158,-0.261838764,-0.0674943104,-0.0863031894,-0.1876974255,0.3005430996,-0.017649848,0.1646880209,0.2321302742,-0.1561704874,0.0077963192,-0.0597303659,0.0763560236,-0.1446675211,-0.3538124859,-0.1924921572,0.4830202758,0.1428171545,-0.1032774076,-0.1775431484,0.2011538148,-0.2903551161,0.0309926234,0.1604841948,0.965144217,-0.0913610011,0.1023061723,-0.0350122936,-0.0705716163,0.7512097359,-0.104984872,-0.0075933365,-0.3222668171,0.2065496892,-0.0347509645,0.3142012358,0.2218555808,0.0131513383,-0.2440468222,0.4417722225,0.2002679408,0.1980752498,0.058526814,0.3157040179,0.0242270641,-0.3382952809,-0.583748281,-0.0024885198,-0.36233446,0.4211081862,-0.1653506756,-0.2722578049,-0.3313656449,-0.2512749732,-0.131403029,0.1604578346,-0.107055597,0.1632218957,0.0069442224,-0.4532437921,0.2660300434,0.3748112321,0.092020452,0.2868559957,-0.3253129423,0.0356325842,-0.4986883402,-0.3538831472,-0.1333377659,0.1337982714,0.0581639819,-0.0023210412,-0.3238852024,0.0269920528,-0.1235910207,0.1611974686,-0.2638441324,-0.0537394099,-0.2599516213,0.0516092591,-0.3734903038,-0.1075445116,-0.1260600388,-0.2077461183,0.0324530192,0.008986121,0.0371839702,-0.0908613577,0.0393213332,-0.0547889322,-0.0511371791,0.5814448595,0.1348665357,0.2558547556,0.4162301421,0.1885979921,-0.3900574148,-0.1254880726,0.0539543033,-0.5722735524,-0.1605397165,-0.1082498878,-0.1560444534,0.0224918816,-0.1950097382,-0.1159540117,0.035395544,-0.3885105252,0.2575335205,-0.4159332812,-0.1893941462,0.3224726915,-0.0519076586,0.0377475396,0.1257080138,-0.0930598825,0.1009743139,0.0170966499,-0.1186206639,0.1285017729,-0.3691442311,0.1367328465,0.2836230993,0.1846770495,-0.1155968606,-0.0738666058,-0.0809885338,-0.2218613327,-0.0824956298,0.0333905369,-0.1826776117,0.2178574651,0.1572387367,-0.1708374768,0.260830313,-0.3195835948,0.0899044275,-0.2070213258,-0.0810045302,0.2713994384,-0.0602881759,0.2347515523,0.1839458346,0.2207906246,-0.2862567008,0.38986516,0.1317741722,0.4377555251,0.0245286021,0.3439136744,0.3335205615,-0.0640837252,-0.2942876518,0.0883963555,-0.0271482151,-0.1182398871,0.3131261766,-0.1907463968,-0.0510718673,0.4062865674,0.0962452143,0.3031124175,-0.0038799893,0.0534748398,0.2515538931,0.0003137954,-0.3657249808,-0.1449488103,0.5537335277,0.01936635,-0.0792309195,0.3045827448,0.1624540687,-0.3283275366,0.1121639907,0.0645582154,-0.0207514986,-0.086420238,0.1543729603,0.1341918111,0.3315724134,0.1660039425,0.0867367759,-0.1338923573,0.0064770463,0.030562656,-0.0203278251,0.1603604108,0.0686553568,0.1406930238,-0.2823565304,-0.5433478355,0.0276963357,0.3001674712,-0.1916796416,0.0341291539,-0.0853605568,0.1436231732,0.2602636814,-0.0399068743,-0.2248319089,0.177001968,-0.0960346386,-0.2489120364,-0.3540971577,-0.4199373722,-0.0514397584,0.1501644105,0.0987486616,-0.348010689,0.3086063266,0.1745431572,-0.2039578259,-0.517226994,0.6209245324,0.0093129007,0.0703897998,-0.3211233914,0.0193198603,-0.193955183,-0.0917457119,0.0491575301,0.3262443841,0.2265163064,0.264282912,-0.252224803,-0.1073273718,0.2573589087,0.0472931489,0.0638954341,0.0337764025,0.1843176633,-0.0084542772,0.3986049891,-0.0208824445,-0.0029428622,-0.2704879344,0.3224714696,-0.0859106109,-0.1025524959,-0.1120009795,-0.2207192779,-0.0870827362,0.0235404149,-0.2096631825,-0.3902516067,0.0390547179,0.0122992918,0.2634193301,0.2793067396,-0.1467745155,-0.0257328488,-0.1814183146,0.3561479747,0.2147600353,0.0015996922,-0.2113437504,-0.013352884,-0.4859939516,-0.0285965037,-0.4546748102,-0.0159429852,-0.0595692061,0.3594653904,-0.0488422662,0.0640125498,0.0648742169,-0.2156126946,0.0658656955,0.1961434782,-0.2607964873,-0.2078612149,-0.0740692243,0.0722362772,-0.2272894382,-0.4673805833,0.2686919272,-0.3110842109,-0.0997855887,0.0278221536,-0.0226985775,0.1338787973,0.2216407508,0.2262118906,0.1189349517,0.1705466658,-0.1098874956,-0.0132623529,-0.351834625,-0.0679581091,-0.1064103842,0.2769089043,-0.1255407184,0.5174766779,-0.1731731147,0.078177914,-0.2346602529,0.4826259315,-0.178996712,0.2235510647,-0.042124901,0.1130044982,-0.3322497308,0.0773149356,-0.1158128604,0.0802660361,-0.2116837949,0.141291678,-0.5255597234,-0.1196477413,0.2435221076,-0.2628340125,-0.1852324158,-0.1589695513,0.0839172751,0.2192897201,0.2267581373,-0.4330992997,0.1814738363,0.2489357591,-0.0612569153,-0.1152703986,-0.0638967231,-0.1936850846,-0.3239337206,-0.056840919,-0.1383538991,0.025284905,-0.4177716672,-0.0991424024,-0.2453081608]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/784","title":"Issue with downloading Wikipedia data for low resource language","comments":"@lhoestq Thanks!","body":"Hi, I tried to download Sundanese and Javanese wikipedia data with the following snippet\r\n```\r\njv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner')\r\nsu_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner')\r\n```\r\nAnd I get the following error for these two languages:\r\nJavanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nSundanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nI found from https:\/\/github.com\/huggingface\/datasets\/issues\/577#issuecomment-688435085 that for small languages, they are directly downloaded and parsed from the Wikipedia dump site, but both of `https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json` and `https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json` are no longer valid.\r\n\r\n Any suggestions on how to handle this issue? Thanks!","comment_length":2,"text":"Issue with downloading Wikipedia data for low resource language\nHi, I tried to download Sundanese and Javanese wikipedia data with the following snippet\r\n```\r\njv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner')\r\nsu_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner')\r\n```\r\nAnd I get the following error for these two languages:\r\nJavanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nSundanese\r\n```\r\nFileNotFoundError: Couldn't find file at https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json\r\n```\r\n\r\nI found from https:\/\/github.com\/huggingface\/datasets\/issues\/577#issuecomment-688435085 that for small languages, they are directly downloaded and parsed from the Wikipedia dump site, but both of `https:\/\/dumps.wikimedia.org\/jvwiki\/20200501\/dumpstatus.json` and `https:\/\/dumps.wikimedia.org\/suwiki\/20200501\/dumpstatus.json` are no longer valid.\r\n\r\n Any suggestions on how to handle this issue? Thanks!\n@lhoestq Thanks!","embeddings":[0.1786196381,-0.0915542468,0.0355631709,0.5031728148,0.1778338701,0.0397492312,0.0245488081,0.4043541849,0.0918699726,-0.0169933047,-0.2727084458,-0.185557723,0.4074715376,-0.2015087903,-0.030964572,-0.4686701,0.216587618,0.0196259078,0.0117588453,-0.067821458,-0.2040485293,-0.0486265644,-0.1739368737,0.0612079613,-0.2426719069,-0.0486025289,0.0245044548,-0.0024733341,-0.2858707011,-0.2479097247,0.3051996529,0.1955074221,0.2533367276,0.3573896587,-0.0001251102,-0.0590585321,0.5865513086,-0.0844299272,-0.4797185361,0.0504574589,-0.2325666249,-0.4236339927,-0.2270681411,-0.3952985704,0.143907398,-0.1698812097,0.2828868926,-0.2438297868,0.2382211983,-0.0873441398,0.100811623,0.1669425666,-0.0109384358,-0.203741312,0.4232416153,0.2345332205,0.0728073642,0.0143171372,-0.1237182692,0.0602297857,0.0374926403,0.2517840266,-0.2883180082,-0.0316933133,0.1395505369,0.0108051859,-0.3933703899,-0.4915478826,0.4885276258,0.4985272586,0.8822617531,0.0038319835,-0.5526617169,0.0408159941,-0.108237952,0.1711731106,0.2507136762,0.1016828194,0.0178929567,0.0506415665,0.1936288327,-0.3127492964,-0.17426534,0.5356084108,-0.071602948,0.2669323981,-0.0510456301,0.067120634,-0.1324867159,-0.3753298223,-0.2184385806,-0.3497898877,-0.0890267566,0.3984516859,-0.0174780544,0.0412176028,-0.0235201046,0.0548783839,0.2482403219,-0.3261247873,-0.0886384472,0.0482970737,-0.1856922209,0.0788649544,0.3245296776,-0.3864936531,0.1867302805,-0.121783711,0.6492362618,0.1908603907,0.3468075693,-0.1166795269,0.1112030819,-0.217063874,-0.5272045135,-0.1972555667,0.3362259567,-0.3786851168,-0.0882673636,0.2484256178,-0.2982155979,-0.1388314366,-0.1205799058,0.2060297877,-0.1700335592,0.5032062531,0.2816097736,0.3396740556,-0.3097490072,-0.21336402,-0.002239431,0.2029214054,0.069549337,0.1337595433,0.143404305,0.1134837344,0.3551291227,0.0848193169,-0.363124162,-0.1624638587,-0.2073983997,-0.3688312471,-0.3252537251,0.1190088466,0.1564212292,0.5992799997,-0.1748053432,-0.0655795485,-0.1661454439,0.0451296568,-0.3330823481,0.0564587973,0.051783897,0.0392461568,-0.3329943717,0.1090476736,-0.1164016873,0.3271816671,0.0885259509,-0.0986814201,-0.0094955061,0.1647109985,-0.5120577812,-0.2402655333,0.3306535184,0.6452141404,-0.326087296,-0.2071326971,-0.2655140758,-0.07561028,0.3856321573,0.3427251279,-0.1973087192,0.5056831837,-0.137521714,0.3840234876,0.3763515353,-0.0123990737,-0.2839014828,0.32100752,0.1292559952,0.0141540924,0.1769782007,0.0481741317,0.3618089855,-0.1909774244,0.0554572977,0.4139858186,0.18120341,-0.0110287908,-0.4113020301,-0.1764989942,0.1268015206,0.0368722603,0.5070595741,-0.1893838495,0.2619358599,0.7989595532,0.1824272871,-0.1058205515,0.317281276,0.4312382638,-0.2146311402,-0.0489357971,0.0676003769,-0.0285136625,-0.0622393042,0.1756249815,-0.0413413085,0.0570103973,-0.0930704698,-0.046655558,-0.4039661586,-0.112831831,0.1085650846,0.2116232812,0.0497519188,0.0191762727,0.0695056319,0.388153255,0.3417139649,-0.2633938789,0.0318265073,-0.0645282865,-0.2306150943,0.4210470617,-0.0516635142,0.1122238412,0.1109053642,-0.0690091029,0.4112375677,0.0734462589,-0.0802801028,0.1754925698,0.1408019811,0.196492061,0.2494700104,-0.1921579391,0.4811608195,-0.1113340035,0.3379900157,0.5026731491,0.3292540312,-0.1666206867,-0.1174076349,0.3784888089,-0.0834079534,0.0025171721,-0.0888929814,0.0247497633,0.4182535112,0.0160185341,0.2166407257,-0.1116760001,0.2426496744,0.3060250282,-0.1103839427,0.1395713836,-0.4566942751,-0.1404311061,0.2824612856,-0.2183391303,0.2655531764,-0.0312675238,-0.2328888178,-0.0282630045,0.2565798163,-0.0503560603,0.1437120885,0.0924822092,0.2409334779,-0.3033424318,0.2487966865,-0.219381243,0.1258102059,0.2550686896,0.2549644113,0.1653358489,0.0808472931,-0.0121658817,-0.1813852191,0.3044628799,-0.0301462449,0.277102977,-0.1315004975,-0.1587382257,-0.3600617051,-0.3527292013,-0.2127041072,0.1200152785,-0.3987418711,-0.356364429,0.1294712126,0.0962580666,-0.3287450969,0.0252786148,-0.2165237665,-0.1850563884,-0.0124160033,0.0064579784,-0.1975738704,-0.2204752564,-0.2140103877,-0.0680581778,0.4079530537,0.2943391502,0.1117187142,-0.1248179674,-0.3859615922,-0.4942632318,-0.184283033,0.0787756369,0.0691676885,0.1322096735,-0.0315306075,0.558224082,-0.0929501206,-0.0101044485,0.1337634772,0.0035851283,-0.0801954046,0.147303775,-0.0301168337,0.1531237215,-0.1236448139,-0.2683154345,-0.1529656649,-0.2973510921,-0.0655479431,-0.0420922823,0.1355310678,-0.3054483235,0.1934994608,0.0462663881,0.2934267521,0.202178672,-0.2384553552,-0.0790672302,0.5974600315,-0.0729232877,-0.4496887922,0.1407751143,-0.0919638723,0.0392969437,0.221739918,-0.2928608358,0.1580236107,0.1104384214,0.1891470551,0.2667778432,0.147159785,0.0803797096,-0.2661951184,0.1100322232,-0.0619375408,0.1946413517,-0.1846102923,-0.2057250291,0.4600301981,0.3829562664,0.4377918541,0.0006268981,0.7386979461,0.0592851676,0.3264911473,0.3146104515,0.4192307889,0.2448082417,0.1449342072,-0.0070150159,0.0247489568,0.0311153904,-0.0028461833,0.3034727275,0.0790625066,-0.2064209878,-0.3367328644,-0.4319285452,-0.3675804734,-0.3299281895,-0.1495547742,0.1186969131,0.1929888576,0.0255287625,0.4056770802,-0.0485216714,-0.295245111,0.2263528556,0.2296291292,0.2215245515,0.2484011054,-0.1900672168,-0.0319219194,-0.5945740342,0.3525826931,0.0408853777,0.449030757,0.1072836518,-0.0300536193,0.0044802772,-0.0409317203,0.1834150404,-0.2203916162,0.1912971884,-0.0713709965,-0.0041033318,-0.3443957567,0.0163757149,-0.2284627259,0.3116476834,0.5036585927,0.1573521495,-0.7052986622,-0.2460361421,0.0263708029,0.2563281059,-0.0592275299,0.0044319774,-0.1619263291,-0.3942430317,-0.1885688603,-0.092487134,-0.0794457793,0.1261663139,-0.1796601862,0.231782034,-0.2112765312,0.3435726464,0.0781511217,0.08277677,0.1663426608,0.0334004052,0.119198516,0.240228042,0.049467504,0.2583320439,0.3908606768,0.1792383492,-0.2802014649,-0.0476858132,-0.080102846,-0.163898766,0.2009035349,-0.0933195427,0.1817849725,0.2336309701,-0.187992543,-0.0223181378,-0.0406159088,0.1872678548,-0.0808808208,-0.3738819063,-0.1514998972,0.4668899477,0.2011755407,-0.1046635583,-0.1993723661,0.1813402623,-0.2023518085,0.0129642813,0.1461385936,0.8764947653,-0.1438428611,0.1090643704,-0.0131250117,-0.0098207723,0.7917615175,-0.1033853367,0.0493870564,-0.3197218478,0.0981003791,-0.0539178997,0.3398922682,0.141756475,0.019194765,-0.1542919129,0.3997728825,0.2154833078,0.2348602712,0.0291629992,0.34571293,0.0518677495,-0.2633115947,-0.5727800131,0.0369480215,-0.3412213326,0.3903761208,-0.2308899462,-0.2268359959,-0.2393369079,-0.2476684302,-0.0719011128,0.1709100306,-0.043422658,0.2907345295,-0.0230044685,-0.4650378525,0.2174841613,0.2888849974,0.1022741348,0.2750402391,-0.2726037502,0.0301658809,-0.5550385714,-0.3479052782,-0.1218018681,0.1358603984,0.1164036095,0.0113018993,-0.2462144792,0.0957515463,-0.0859639868,0.1661217511,-0.2431435138,-0.0196005497,-0.3584475517,0.0617419668,-0.3674289286,-0.1077776626,-0.1462792009,-0.1871256232,0.0522834435,0.097918883,0.0049836063,-0.148899436,0.1068901047,-0.0341116786,-0.0073692598,0.5981314778,0.2128921449,0.1767122447,0.4064259231,0.1778462231,-0.4610136449,-0.134305492,-0.0315144435,-0.5632510185,-0.1674393117,-0.0751469955,-0.2273816168,0.1380964816,-0.2326830178,-0.2077345401,0.0343325026,-0.3724065423,0.2331857681,-0.4281468987,-0.1943157911,0.4392628074,0.0094840946,0.0037184977,0.1464810371,-0.1621364802,0.0680834278,0.0361681581,-0.1348018944,0.1292701364,-0.3369716406,0.1998314559,0.2901670039,0.1307560354,-0.0828574449,-0.0822687,-0.0351983644,-0.2036866844,-0.0694674253,0.0166335981,-0.1133049354,0.2017035484,0.1529972851,-0.1878193766,0.2513023615,-0.3150022626,0.1836509109,-0.1575126201,-0.076670073,0.2379492372,-0.0503966361,0.194876343,0.2114303112,0.2380336821,-0.2908056378,0.399616152,0.0756716877,0.4530146718,-0.0228668824,0.3233878016,0.4547655284,-0.0331994444,-0.2978628874,0.0705303028,0.0043080254,-0.1661849022,0.3223378956,-0.2162181735,-0.0425389297,0.4102816582,0.084397614,0.2964698374,0.0253358204,0.0926939622,0.1859788746,0.0455171019,-0.3490274251,-0.1479230374,0.4853527248,-0.0044414531,-0.122122623,0.2496337891,0.0801413208,-0.2782052159,0.0815776885,0.0763150454,-0.0641535595,-0.0163621381,0.1870658696,0.1299520731,0.4314402044,0.1541232169,0.0574352667,-0.1410813928,0.0016992873,-0.0270035472,-0.0721443295,0.2399929613,0.0442406833,0.1213699207,-0.2448453307,-0.4209448695,-0.1049141362,0.3403676152,-0.1921256185,0.0676372424,-0.1144096404,0.1404842436,0.2601520121,-0.1248865575,-0.2411623895,0.2110381871,-0.1283989549,-0.2009819001,-0.3725899458,-0.3641116321,-0.1140562072,0.1500466466,0.0618273728,-0.2987218797,0.3423910737,0.2166405618,-0.1334405541,-0.5170475245,0.5984264016,0.0947302431,0.0972291976,-0.3193147182,-0.0285699125,-0.1645079255,-0.0674931556,0.0357940495,0.3686744869,0.1922394633,0.2224947512,-0.2817058563,-0.1172855422,0.2308419645,0.0102715176,0.0948098525,0.0722701326,0.1131281033,0.0466273986,0.3882396817,0.0095031587,-0.0156154651,-0.1940419078,0.3086785376,-0.1422365904,-0.0702155679,-0.0948039889,-0.2593750954,-0.102469191,-0.0808661655,-0.2198457122,-0.4356157482,0.0398861356,-0.0046600685,0.1804938465,0.2062910348,-0.1517167687,-0.0106960787,-0.2149042785,0.2850955725,0.3328787386,0.0484216698,-0.2299475521,-0.0592596792,-0.4567591846,-0.0011113127,-0.4882083535,0.0174112599,-0.0775568262,0.3373667002,-0.086028561,0.0824351981,0.069781743,-0.1591700763,0.0252509024,0.1641402394,-0.2781136036,-0.1926007718,-0.1484652907,-0.0044724965,-0.2551518381,-0.446182549,0.2521808147,-0.3382547498,-0.0832721591,-0.0080539398,0.0014490138,0.0224726833,0.2075880319,0.239718318,0.1301614046,0.1202379316,-0.1176048592,-0.0828460976,-0.2929579616,-0.1603579968,-0.0565218553,0.3372421861,-0.1429798752,0.4741894901,-0.1184858605,0.1155937016,-0.1547490358,0.4774958789,-0.2393037677,0.1971368492,-0.1372748017,0.0874861404,-0.389781177,0.0357131921,-0.0828025788,0.0215747412,-0.2010059655,0.101582624,-0.5101776719,-0.1246637478,0.179445222,-0.3231382966,-0.1639368534,-0.2575347424,0.1458891034,0.2454683334,0.2190019488,-0.4140423238,0.153430149,0.2502042651,-0.0047992328,-0.1180201694,-0.0861852467,-0.1027927399,-0.3430725336,-0.0655502975,-0.0965878963,0.0135647096,-0.45523417,-0.0065944288,-0.262763381]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/778","title":"Unexpected behavior when loading cached csv file?","comments":"Hi ! Thanks for reporting.\r\nThe same issue was reported in #730 (but with the encodings instead of the delimiter). It was fixed by #770 .\r\nThe fix will be available in the next release :)","body":"I read a csv file from disk and forgot so specify the right delimiter. When i read the csv file again specifying the right delimiter it had no effect since it was using the cached dataset. I am not sure if this is unwanted behavior since i can always specify `download_mode=\"force_redownload\"`. But i think it would be nice if the information what `delimiter` or what `column_names` were used would influence the identifier of the cached dataset.\r\n\r\nSmall snippet to reproduce the behavior:\r\n```python\r\nimport datasets\r\n\r\nwith open(\"dummy_data.csv\", \"w\") as file:\r\n file.write(\"test,this;text\\n\")\r\n\r\nprint(datasets.load_dataset(\"csv\", data_files=\"dummy_data.csv\", split=\"train\").column_names)\r\n# [\"test\", \"this;text\"]\r\n\r\nprint(datasets.load_dataset(\"csv\", data_files=\"dummy_data.csv\", split=\"train\", delimiter=\";\").column_names)\r\n# still [\"test\", \"this;text\"]\r\n```\r\n\r\nBy the way, thanks a lot for this amazing library! :)","comment_length":36,"text":"Unexpected behavior when loading cached csv file?\nI read a csv file from disk and forgot so specify the right delimiter. When i read the csv file again specifying the right delimiter it had no effect since it was using the cached dataset. I am not sure if this is unwanted behavior since i can always specify `download_mode=\"force_redownload\"`. But i think it would be nice if the information what `delimiter` or what `column_names` were used would influence the identifier of the cached dataset.\r\n\r\nSmall snippet to reproduce the behavior:\r\n```python\r\nimport datasets\r\n\r\nwith open(\"dummy_data.csv\", \"w\") as file:\r\n file.write(\"test,this;text\\n\")\r\n\r\nprint(datasets.load_dataset(\"csv\", data_files=\"dummy_data.csv\", split=\"train\").column_names)\r\n# [\"test\", \"this;text\"]\r\n\r\nprint(datasets.load_dataset(\"csv\", data_files=\"dummy_data.csv\", split=\"train\", delimiter=\";\").column_names)\r\n# still [\"test\", \"this;text\"]\r\n```\r\n\r\nBy the way, thanks a lot for this amazing library! :)\nHi ! Thanks for reporting.\r\nThe same issue was reported in #730 (but with the encodings instead of the delimiter). It was fixed by #770 .\r\nThe fix will be available in the next release :)","embeddings":[0.1203237697,-0.1472451985,-0.1446630508,0.3973078132,0.0596904233,0.1563190073,0.7675094604,-0.1010695621,0.3270575106,0.0822042897,-0.0324487649,-0.0013563952,0.1088031158,-0.2761227489,-0.1162617058,0.1431807727,0.1974260807,0.1384890378,-0.0753118917,-0.0711717606,-0.2949311733,0.2251483202,-0.1458898336,-0.1837260425,-0.0212593339,0.3673403859,-0.0063925441,0.2704255283,0.1843010187,-0.3755665123,0.4452941418,0.204180479,0.2709940076,0.3557014763,-0.0001234261,-0.0847295299,0.0421750657,-0.2460636199,-0.1824997067,-0.2576674223,-0.3444088697,-0.3255288303,0.1659789383,-0.640501976,-0.0831279233,0.4193374217,0.0104722306,-0.3650924265,0.0418628417,0.289988786,0.1642397046,-0.2764177918,-0.4180568159,0.3066487312,0.2418942451,-0.1092139035,-0.2015373558,0.1077693254,-0.0016694436,0.2170140743,-0.0673587844,0.0988658071,-0.2695139349,0.3476351798,0.3233605027,0.2655831873,-0.1112889498,0.1209461093,0.2490369827,0.0957703143,0.5434871316,-0.0160900671,-0.1401772052,-0.3993289471,-0.1829320341,-0.9000109434,0.4474709034,0.1172697395,0.2949036956,0.2847469151,-0.1221222505,0.174939245,0.0017118829,-0.0270429906,-0.2510727048,0.0042830245,-0.3040367961,0.1181052923,-0.1376942694,0.0799042359,0.2939622998,-0.3672236502,-0.0892909467,0.0888362452,-0.2034035623,0.2997893393,-0.1713486165,0.052713871,-0.1950713694,0.051088877,0.1519305706,-0.0921810642,0.0524523109,0.0886828452,-0.0951909423,0.2707783282,0.134291783,0.1440228224,0.1321519762,0.0277069639,-0.36105299,-0.1131923497,-0.040470168,-0.0574660264,0.7582945228,-0.0240596309,-0.1336691678,-0.1446204334,-0.2497699261,0.1310025305,-0.0539615043,-0.1943701506,-0.310503155,0.1582721621,-0.0937401354,0.0074114646,-0.2572183013,0.1283313781,-0.1247400343,-0.1262661517,0.013780483,-0.2658877671,-0.2747731507,-0.0999717265,0.1676563174,-0.0354341939,0.1349232346,0.0876384303,-0.2041487396,-0.3977515399,-0.0401847735,-0.040093936,0.4941184521,0.1790121496,-0.0341455601,0.3948505223,0.2257198989,-0.7923141718,-0.1296625137,0.2836402655,-0.4482902586,-0.1099786162,0.0770924613,0.0487055443,-0.4173682034,-0.0573835187,-0.240827471,0.1323028803,0.4306553006,-0.2343614995,0.0749179125,-0.279586494,-0.3188855052,-0.4057266712,-0.1243101358,0.5804481506,-0.6152841449,-0.1354853958,-0.2932513654,0.0454964712,0.0796200335,0.2107527107,-0.0280726515,-0.3229223788,-0.4087453783,0.0544598401,0.1152574718,-0.1735103428,-0.5145955682,-0.0387589894,-0.1684631109,0.5907817483,0.3413653672,0.0757716,0.4065316021,-0.0589029714,0.1259220839,0.2793928385,0.2539744377,0.020392511,-0.1249643192,-0.062951766,0.3729768991,-0.0367324986,-0.0246832166,0.2701401114,0.3431218266,-0.4709174037,0.2977422178,0.1017794907,-0.1056778878,0.1167057529,-0.1657940447,0.1936220974,0.2169577777,-0.0856490135,-0.4177284539,0.1580454558,0.1372050792,-0.3520218134,-0.1161354929,-0.0502136871,-0.2620417178,-0.1570705622,-0.1790102124,-0.0666497722,0.0170435309,0.2542367578,0.022816889,-0.0022813804,0.0898884535,0.4797988236,0.006873731,0.1321927905,-0.2679718137,-0.2331027985,0.0207112674,-0.0352903008,-0.5559324026,-0.0199753065,0.2322299033,-0.1991242319,-0.2429983616,0.4130986035,0.212196663,0.3052728474,-0.0656284466,0.1739781946,0.02716472,0.2331711203,0.1174948886,-0.008733904,0.1551889032,-0.1189426407,0.1674257666,0.5678290725,0.0782879218,0.2709801793,-0.2086476386,-0.18897973,0.3382747769,-0.0512223542,0.0360137224,-0.3398999274,0.1569098234,0.0934806317,0.0617582798,0.2420657575,-0.1884282529,-0.0971823782,0.2958268225,-0.1709136963,-0.0693111792,-0.0697167441,-0.3157299459,0.0640438199,0.1098499596,0.4225114286,0.5408830643,0.0355893523,-0.0270284247,-0.1226385534,0.0877398625,-0.3477230966,0.2842814028,0.0567690432,-0.0782581866,0.2036445439,0.2016819417,-0.0129735712,-0.3612755537,0.2047887892,0.1962146461,-0.3082534075,-0.3972662985,0.2769383788,-0.4201709628,0.0169204082,-0.2480187863,-0.0532403365,-0.1965268105,-0.2177442312,-0.257578522,0.3369186819,0.1936834753,-0.1185887307,-0.3190780878,0.4337733686,-0.1746449322,-0.5059754252,0.1377483159,-0.0415904298,-0.2272444069,-0.051291991,-0.0573533028,-0.1890227199,-0.0249267872,-0.1329493076,0.0726283267,0.0894688666,-0.1508388519,-0.1577429622,0.1279167384,0.2775491476,-0.1862485558,0.0429017358,0.0478855073,-0.0051693525,-0.0678490251,0.2459173352,0.2164950073,-0.0260622688,0.1768840998,-0.0759266093,-0.0018829505,-0.2840174437,-0.3081313074,0.0247397479,0.0312394556,0.0404082946,0.1132464781,-0.001486378,-0.3966306746,-0.1957970709,0.1762916297,0.1216729656,-0.4113669097,-0.3755255044,0.4756684005,-0.2218316644,-0.0878486708,0.1451059282,-0.0337921306,-0.0669071376,-0.043367561,-0.3657758534,0.4310222864,-0.2626770437,0.2542522252,0.2346964031,-0.2392585725,0.2565965652,-0.0234662108,0.2168571353,-0.2194843441,-0.392771095,0.1954289824,0.1232841164,0.3100214899,0.1594948024,0.1339705437,-0.0279372744,0.7450904846,-0.1718225926,-0.0983114094,0.3850101531,0.0750990808,0.3412340581,-0.0181278549,-0.0712050423,-0.0378132388,-0.1265809834,-0.3021036386,0.2834189236,0.1066194177,-0.2146609873,-0.1131849959,0.1987177581,-0.248671338,-0.3150984347,-0.004506785,-0.1817226857,-0.0972391814,0.3563121259,0.0614511147,-0.2178488523,-0.2576109469,-0.2146377116,0.4987211525,0.2452061623,-0.0813571438,0.0362473726,-0.0781250969,-0.0385513008,0.2863278985,0.2001431733,0.1760803014,0.0247729793,-0.1757690012,0.1913099885,0.213643074,0.7930586934,-0.2811437845,0.1568856537,0.0941017419,0.099480696,0.0205852147,-0.294855535,0.0798056871,0.197943002,0.0210837424,0.2169817537,0.0903883129,0.1625833809,-0.0634288415,0.2885499299,-0.1760798991,-0.4301860631,-0.1813444644,0.2266565859,0.0245669112,-0.0818469003,-0.1082707793,0.1065232232,-0.325919807,0.0175329298,-0.1557329148,0.1471208781,0.0569858737,0.0696722865,0.1517105401,-0.0355041549,-0.0175293181,0.2442268282,0.1019753888,0.1202894673,0.4538984895,-0.2612040639,-0.3922269344,-0.0007036239,-0.1960945129,0.0274526514,0.1697650552,-0.1067796648,0.2334295064,-0.0790707022,-0.2604856491,-0.0035102335,-0.1814682484,0.1122139171,0.2038200498,-0.4512124956,-0.4058631957,0.3258876204,-0.0554784015,-0.2559295297,0.1496506631,-0.0949921384,-0.4765543342,0.3897123337,-0.2141076326,0.7907226086,-0.0782905594,0.2348376662,-0.2794219255,-0.3538901806,0.0862282217,0.0383168943,0.1260598749,-0.2394668311,-0.4958381355,-0.189215526,0.1001505256,0.3439461589,0.2582060695,0.1379610002,0.3105574548,-0.4034962058,0.4532230794,-0.0289509725,-0.1205717772,0.0407461226,-0.1886200607,-0.0197523963,0.0188004207,0.0391596667,0.2214196324,0.0735887885,0.1305435151,-0.033097975,0.0506797023,-0.2418760359,-0.1328680366,-0.0424514227,0.0774864331,-0.2033519745,-0.5004670024,0.1164987385,0.4811633825,0.095617719,0.069704935,0.0979126543,0.2905620933,0.4477483332,0.2044315636,0.0422960073,-0.1440746784,0.0852391869,0.2692612708,-0.0308079496,0.0312861092,-0.2342547774,-0.2706260383,0.0721679851,0.2193858474,-0.0028329997,-0.2397052497,-0.4396778047,-0.3141622841,0.0672016516,-0.0320407711,0.025117971,-0.1814769506,0.0804366022,0.1577833593,-0.2707761526,-0.6040502191,-0.252725482,0.4860739708,0.1338461936,-0.2265944481,0.49207744,-0.2527627945,-0.1719788015,0.0251070429,0.1729397178,-0.1921681762,-0.070651181,0.1088251323,0.0554745495,0.177981928,-0.0406764448,-0.0561645031,0.1289059371,0.18949458,-0.1891904324,-0.5128541589,-0.0139781069,-0.2773654759,-0.1763147712,-0.0321367607,0.333145082,-0.3526581824,-0.092777513,-0.1866702586,-0.147001341,0.2397243977,0.2396584004,0.0713043585,0.3087339997,0.2459814101,0.0709194988,-0.0696480274,0.0592723638,0.0721109957,0.0686555803,-0.1127389595,-0.3264077902,0.2004229128,-0.0409801565,0.1022209153,-0.2532328069,-0.3742201924,0.1008262262,-0.3272667527,0.1454121768,0.330365926,0.0903070495,0.1455649436,0.0393872857,0.1486801654,-0.3327032626,0.1666129678,-0.3360438347,0.1882571578,0.1342082173,0.1283867508,-0.3437955976,0.2336547077,-0.34372738,-0.0905324072,-0.2087256312,0.1693561375,0.3830248117,-0.3520566523,0.2139899135,0.4557955563,0.2927396595,0.2756817639,-0.1129899547,-0.2772304118,0.1314342171,0.0361262634,-0.028236011,-0.1241683885,0.1595762968,0.0667330772,-0.0311175454,0.3678384125,0.2150232047,0.0445857383,0.0652378276,-0.0151626607,0.7256931663,-0.2609648108,0.4738339782,0.4797536731,0.2778429389,0.3633262515,0.4889661074,0.018202154,-0.056457106,0.7593064308,0.1222186387,0.1837300062,0.5540966392,0.0704659969,-0.3536395431,-0.5032405257,0.2317384928,0.1406024843,-0.4301652014,0.562718153,0.3728712201,0.1879161298,-0.3548078537,0.01792996,-0.3499878645,-0.1077628285,-0.1346459538,-0.3442691565,0.1918267459,-0.1022472084,0.0610238239,0.1465801895,-0.2583983541,0.1731935143,0.2608609498,-0.003770995,-0.0425372981,-0.4601888955,0.0837598592,-0.1845098883,0.4517956674,-0.180287376,0.0952509493,-0.0390569381,-0.0451595746,0.4648664892,-0.312107414,0.3018802106,0.2521142066,0.3326146603,-0.2277377546,-0.0551614426,-0.0324273072,0.0382371359,0.3212358952,0.1287715435,0.0200460423,0.1461442858,0.0891899243,-0.0379310921,0.5240586996,0.4534051418,0.27709499,-0.5207794905,-0.3061748445,0.0269795842,0.0006313204,-0.2785097361,-0.0913929865,-0.0344317816,0.0943515822,0.2659895718,0.0418441296,0.3121053278,0.2237243801,0.0124191111,0.2612023354,0.1443605721,0.232830599,-0.0991451517,-0.2329228222,0.214176923,-0.7888926864,0.0532443076,0.2286850661,0.1859432608,-0.012466751,0.0698960498,-0.0053087543,0.0196038093,0.1303019077,0.0490174219,0.2602853179,0.4160325229,-0.0297498442,-0.3272883296,0.1327348351,-0.1052253097,-0.2089363635,-0.3332625628,0.1337274313,0.2565193474,-0.0738355741,-0.1368063688,-0.0006603462,0.150446713,0.345936507,0.3318856359,0.0089831678,0.4752737582,0.0264084991,0.1231704354,-0.1790117919,-0.0965864509,-0.2501386702,0.1303341687,0.0363448225,0.1360018551,-0.3375628591,0.0401218273,0.0102688447,0.3192097247,0.1589882225,-0.0755918995,-0.3314594328,-0.1442620307,-0.1662588716,0.0904366225,0.0469603166,0.2856378257,0.0346994326,0.4123387039,0.1802210957,-0.1800531894,0.1023079827,-0.194243446,-0.1762610823,0.1269196123,0.2522345483,0.0130312014,-0.0983864963,-0.2418574691,0.0179552678,0.5038869381,-0.2512527108,-0.1480120569,0.120695591,-0.0764407143,0.1261561066,-0.2026415169,0.4905095696,0.3065373302,-0.3316522539,-0.055414658,-0.2716313899]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/778","title":"Unexpected behavior when loading cached csv file?","comments":"Thanks for the prompt reply and terribly sorry for the spam! \r\nLooking forward to the new release! ","body":"I read a csv file from disk and forgot so specify the right delimiter. When i read the csv file again specifying the right delimiter it had no effect since it was using the cached dataset. I am not sure if this is unwanted behavior since i can always specify `download_mode=\"force_redownload\"`. But i think it would be nice if the information what `delimiter` or what `column_names` were used would influence the identifier of the cached dataset.\r\n\r\nSmall snippet to reproduce the behavior:\r\n```python\r\nimport datasets\r\n\r\nwith open(\"dummy_data.csv\", \"w\") as file:\r\n file.write(\"test,this;text\\n\")\r\n\r\nprint(datasets.load_dataset(\"csv\", data_files=\"dummy_data.csv\", split=\"train\").column_names)\r\n# [\"test\", \"this;text\"]\r\n\r\nprint(datasets.load_dataset(\"csv\", data_files=\"dummy_data.csv\", split=\"train\", delimiter=\";\").column_names)\r\n# still [\"test\", \"this;text\"]\r\n```\r\n\r\nBy the way, thanks a lot for this amazing library! :)","comment_length":17,"text":"Unexpected behavior when loading cached csv file?\nI read a csv file from disk and forgot so specify the right delimiter. When i read the csv file again specifying the right delimiter it had no effect since it was using the cached dataset. I am not sure if this is unwanted behavior since i can always specify `download_mode=\"force_redownload\"`. But i think it would be nice if the information what `delimiter` or what `column_names` were used would influence the identifier of the cached dataset.\r\n\r\nSmall snippet to reproduce the behavior:\r\n```python\r\nimport datasets\r\n\r\nwith open(\"dummy_data.csv\", \"w\") as file:\r\n file.write(\"test,this;text\\n\")\r\n\r\nprint(datasets.load_dataset(\"csv\", data_files=\"dummy_data.csv\", split=\"train\").column_names)\r\n# [\"test\", \"this;text\"]\r\n\r\nprint(datasets.load_dataset(\"csv\", data_files=\"dummy_data.csv\", split=\"train\", delimiter=\";\").column_names)\r\n# still [\"test\", \"this;text\"]\r\n```\r\n\r\nBy the way, thanks a lot for this amazing library! :)\nThanks for the prompt reply and terribly sorry for the spam! \r\nLooking forward to the new release! ","embeddings":[0.0635633022,-0.1376715899,-0.1611223966,0.3697888255,0.0240291655,0.16985856,0.7317376137,-0.0952790827,0.3456759751,0.0621810257,-0.0196884777,0.0263576768,0.0926185325,-0.25169909,-0.1203582361,0.1674454808,0.2184598148,0.151863724,-0.1047642007,-0.076608099,-0.2986094654,0.2335488945,-0.1557903588,-0.2052044719,0.0038210799,0.367806524,0.0094520831,0.2638163865,0.147474438,-0.3517835438,0.487827301,0.2849021256,0.2567985058,0.3535687625,-0.0001240682,-0.1025354043,0.0330841728,-0.2503727674,-0.167268157,-0.2247184962,-0.3010952771,-0.3242630959,0.1865069568,-0.6597775817,-0.0849225,0.4206769168,0.0141710974,-0.3834719658,-0.0065421867,0.2676975727,0.1516819149,-0.2033159584,-0.4238824546,0.2895323336,0.3160742819,-0.1204257905,-0.2209732831,0.1117423624,0.0116422446,0.2073915899,-0.0646739379,0.1226585433,-0.2582234442,0.3790101707,0.4034818411,0.2493329346,-0.156485185,0.1434735656,0.2393333167,0.1148699671,0.6052809358,-0.0586435758,-0.1364403218,-0.4068745673,-0.1744253933,-0.8670362234,0.4361382127,0.0539970174,0.3069908321,0.2467764318,-0.1264297962,0.1484864503,-0.0315542929,-0.0197894163,-0.2462922335,0.0236229915,-0.2951024473,0.1403372288,-0.1081191003,0.1042862758,0.2891933322,-0.3306134641,-0.1047821268,0.1154227331,-0.2247299403,0.2880317569,-0.1413272321,0.0636456683,-0.1719520539,0.1004812941,0.182649374,-0.0790001377,0.050377354,0.0490896516,-0.0287298504,0.2890386581,0.1269850582,0.1208611578,0.1639618129,-0.0074727358,-0.3551072478,-0.1157704964,-0.032382004,-0.0672994852,0.7882553339,-0.0502976552,-0.1280403733,-0.1559327394,-0.2330631167,0.1018707827,-0.0144270966,-0.2141044885,-0.259835273,0.17105712,-0.10397356,0.0490734465,-0.2729883194,0.0823000595,-0.1471873224,-0.0479465611,0.0177310221,-0.2720465362,-0.3006331921,-0.1074334681,0.1478467882,-0.0366420448,0.1687697917,0.0754716769,-0.2045571655,-0.3488115072,0.0092526432,-0.0401295163,0.49402228,0.2162912637,-0.0531949103,0.4039781392,0.1593113244,-0.7147265673,-0.151832521,0.2730327249,-0.4315810502,-0.1569108963,0.0763362348,0.0492023714,-0.3940647244,-0.0674636215,-0.2217472792,0.1151744425,0.4086594284,-0.2435841411,0.130919084,-0.2905111611,-0.2552412748,-0.4207433164,-0.1456499696,0.5577818155,-0.6230276823,-0.1625296026,-0.3136772215,-0.0159263145,0.0115611022,0.1959097087,-0.0412421227,-0.3160282969,-0.4081629515,0.0511095896,0.188867718,-0.2318172604,-0.5006058216,-0.0646831617,-0.1167153046,0.5130479932,0.3648191988,0.1294814795,0.3524497747,-0.0367375687,0.1297385246,0.2571272552,0.2665065229,-0.0025443092,-0.0771075562,-0.0675275251,0.343197763,-0.024657486,-0.0168968104,0.246047467,0.3091090918,-0.4324559867,0.2762365043,0.1057215557,-0.1308590025,0.0779493973,-0.1777837723,0.2233598083,0.2501087785,-0.1072759479,-0.4229783118,0.1568689942,0.153378427,-0.3806136549,-0.0688580573,-0.0861040875,-0.2696950138,-0.1501080394,-0.1360625327,-0.0737996027,-0.0055735814,0.2782970667,0.051058799,-0.0540139526,0.0189768802,0.4882960916,-0.0023063628,0.147507295,-0.3020290136,-0.2891509235,0.0296077114,-0.0281032287,-0.5489796996,-0.0158299468,0.2357132137,-0.2063465863,-0.1993586868,0.4474981725,0.2140793651,0.3029052019,-0.0488799065,0.1910493523,0.0259766392,0.300465107,0.1206995845,-0.0338539742,0.1411740482,-0.1210455447,0.1655182987,0.5786083937,0.0464110933,0.2709972262,-0.1822339743,-0.1857110709,0.3285136223,-0.0327109583,0.0395495631,-0.3351762593,0.1468869001,0.1117740422,0.087748155,0.2042727619,-0.1907063425,-0.103051506,0.2695429623,-0.1807536334,-0.0829336941,-0.1325068176,-0.3261303306,0.0357434861,0.1322520673,0.4256306291,0.5540503263,0.0465918854,-0.0194570757,-0.1137767956,0.0336298309,-0.3271061182,0.3168638647,0.0437981077,-0.1439072639,0.1917705834,0.174673304,-0.0726502687,-0.4310675263,0.196579814,0.2052826583,-0.3256497979,-0.3820693195,0.2823598683,-0.3616985679,-0.0112228831,-0.2117984146,-0.0639657676,-0.1837892979,-0.2088836282,-0.2568313181,0.3508309722,0.1859330237,-0.1418526769,-0.3276410699,0.4634444714,-0.1766599417,-0.4898566306,0.1285752356,-0.0669388995,-0.1974189281,-0.0571517088,-0.1172994226,-0.1658727378,0.0224384796,-0.1532191634,0.0854036137,0.0578615516,-0.1726199389,-0.1545440257,0.1547121257,0.2643894851,-0.0985086635,0.0352152139,0.113488026,0.0029837184,-0.1029512361,0.2151020169,0.2049877346,-0.0899734497,0.1729825884,-0.0393679142,-0.0015088801,-0.289640516,-0.2656746507,0.0336521864,0.0385937467,0.0936559215,0.1115277484,0.0407924019,-0.4411357045,-0.1720250249,0.1978138685,0.1351348907,-0.4718317688,-0.3725627959,0.4636910856,-0.2460991591,-0.0603994317,0.145340845,-0.0254069306,-0.0932394713,0.0253035892,-0.3467017412,0.3654436171,-0.2504650056,0.2698564827,0.2420420051,-0.2039552033,0.2793787122,-0.0469130278,0.2026349902,-0.2383543998,-0.3993454576,0.1840856075,0.1458318979,0.3319687843,0.1404758543,0.0835792422,-0.0853042305,0.7569001317,-0.2657800913,-0.1096806452,0.4194296002,0.0669876114,0.38353315,-0.0114389835,-0.0806446746,-0.0420293212,-0.1353840679,-0.3232308626,0.2571461797,0.1152186915,-0.2303030938,-0.1356868595,0.1844886094,-0.2355636656,-0.2845561206,-0.0371874757,-0.154018417,-0.099135749,0.3531802595,0.06558685,-0.2425338328,-0.2362831533,-0.1866888553,0.5406585336,0.2278038859,-0.0869030356,0.0188933313,-0.0698982775,-0.0342898183,0.2807681859,0.1706693769,0.1107893363,0.0488953032,-0.2249499857,0.207617864,0.1762588471,0.8500131965,-0.3017201424,0.1375033706,0.1269579083,0.1223954037,0.0215551537,-0.308884114,0.0773890913,0.2468484789,0.0460802428,0.2135677487,0.1007380709,0.1525954753,-0.040083278,0.2730457485,-0.1604162306,-0.433727175,-0.1859287918,0.1993876547,0.0514731593,-0.0831764042,-0.1386318207,0.0528631397,-0.2383345813,0.0204529651,-0.1870723665,0.0975981802,0.0552866198,0.1373878121,0.1846157163,-0.0090537379,-0.031852562,0.2525266409,0.1103543714,0.0802549198,0.4240502417,-0.2737776935,-0.3302909732,-0.0236262809,-0.1315259635,0.0048070941,0.1606246084,-0.1355074644,0.2559552193,-0.1172237992,-0.2612065971,-0.0024598048,-0.1361798644,0.135482356,0.1744659096,-0.4639264345,-0.4767591953,0.3936649859,-0.0294940751,-0.2998777628,0.1450924724,-0.0846484527,-0.4775096774,0.3497653008,-0.1961151659,0.7891477346,-0.0871274024,0.2582281828,-0.2649728358,-0.3750717938,0.1874131262,-0.0057634013,0.1413217634,-0.2242701352,-0.4079174697,-0.1849196851,0.097390458,0.3535410464,0.2116335183,0.1636779159,0.3410289586,-0.38504228,0.4457184672,-0.0862529501,-0.080700241,0.0339009464,-0.1628030092,-0.0581448972,0.0111838318,0.0680223107,0.2422316223,0.0425454341,0.1523145586,0.0202691611,0.0615143925,-0.2251264155,-0.2366585433,-0.0081434809,0.0796070471,-0.238348797,-0.4841563106,0.0456716083,0.464289546,0.1064681634,0.0661360398,0.0910858065,0.3285362422,0.414917618,0.1834980398,0.0273444876,-0.1332663596,0.12055742,0.2587367296,-0.0125687532,-0.0029493975,-0.2334611714,-0.3164705336,0.0615183227,0.2487823516,0.0067434064,-0.2406788319,-0.443022579,-0.2619326711,0.0514279231,-0.029436009,0.0135703553,-0.1599300951,0.1055688933,0.2052897066,-0.2813999057,-0.5869441032,-0.2842292786,0.4930380881,0.1405247748,-0.2589586973,0.4142863154,-0.2807942033,-0.1433475316,0.0440575369,0.1567829102,-0.1718550622,-0.0439561866,0.13659437,0.0878812745,0.1142444685,-0.0062844767,-0.0296535082,0.1100325957,0.1960732043,-0.214922294,-0.5005058646,-0.0566539392,-0.2965898812,-0.160222128,0.0069441018,0.3611193597,-0.3827982247,-0.0757133365,-0.2007721812,-0.1371497512,0.2301628292,0.2385781556,0.0901947618,0.3121227026,0.2519566417,0.0857986137,-0.0643089414,0.0509354547,0.063499026,0.0298831668,-0.1041643992,-0.353456378,0.2112171799,-0.0843500569,0.0853285044,-0.2347505987,-0.3963413537,0.123164162,-0.332413733,0.2065859884,0.3151215911,0.1087853834,0.1266508698,0.028576659,0.1004778221,-0.3587028682,0.173361972,-0.2975005805,0.1408244967,0.1457120031,0.1496670842,-0.3840323687,0.2196987569,-0.2868731022,-0.0941992104,-0.205914557,0.1303634793,0.3181651831,-0.2837614119,0.172904402,0.4548774064,0.3350809515,0.3565506935,-0.0920297727,-0.2812874019,0.097518377,0.0202412326,-0.0355742648,-0.168987453,0.0808817968,0.0242506787,0.0017579059,0.3426400721,0.2381062359,0.0890928209,0.0593754947,-0.0445889793,0.7067676187,-0.2251773924,0.4917102456,0.5061523914,0.274591893,0.354285866,0.5110496879,0.055295594,-0.0622365773,0.8159404993,0.1131004095,0.2301755846,0.5932115316,0.0600903146,-0.3401613533,-0.4667787552,0.2498339862,0.1751907617,-0.4075567126,0.5800361037,0.3924189806,0.1885271519,-0.3639142215,0.0339204669,-0.3551177979,-0.1239668354,-0.1373892576,-0.4192869365,0.2092298865,-0.11279165,0.1046060845,0.1503666639,-0.259322226,0.1763900518,0.2875089645,-0.0324245356,0.0066044531,-0.4523352385,0.1276136488,-0.2387907356,0.4864200354,-0.1691358238,0.0776891783,-0.0531561002,-0.0248335488,0.4844307899,-0.2887794077,0.3032095134,0.2217226624,0.3409115672,-0.2214237303,-0.0586367734,-0.0433511585,0.0088816341,0.2829450071,0.1242282093,-0.0098378118,0.1609075516,0.0716842487,-0.0147771984,0.5873548388,0.4030327201,0.24359864,-0.5450233817,-0.3061108589,0.0272208713,-0.0067701358,-0.2705374062,-0.0504450127,-0.0404726118,0.0560702235,0.2388051599,0.03503659,0.2924752533,0.1901610047,0.0105318474,0.2531072199,0.217340678,0.2002668977,-0.0803632662,-0.181676954,0.2082995772,-0.8221272826,0.0112593286,0.2440929562,0.171255365,-0.0040077255,0.0789775029,-0.0287959911,0.0089651123,0.1912526935,0.0268275328,0.2903029919,0.3698172569,-0.0227639899,-0.3277322352,0.0798026919,-0.1364519596,-0.1759259254,-0.3469602168,0.1330257505,0.2561320662,-0.107690528,-0.1383015066,0.0438510664,0.2025581449,0.354804486,0.3398697674,-0.030731447,0.4765971303,0.0161878392,0.1322488338,-0.1990061104,-0.0862092972,-0.2432594299,0.0814133659,0.0601316653,0.1108943373,-0.3156266212,0.0690521747,0.0587656461,0.3522517979,0.1713689417,-0.0699589327,-0.3251211345,-0.1448878497,-0.1875336468,0.0587783232,0.0332420357,0.3184863329,0.0365314633,0.3696567118,0.1352797002,-0.1923961192,0.1058984771,-0.2028210908,-0.1624308974,0.090148434,0.228380546,0.0470427908,-0.0738488138,-0.3213833869,0.0497563183,0.4919182956,-0.2325772196,-0.1498887837,0.1395071745,-0.0885125026,0.1234906986,-0.2168569118,0.4710744619,0.3272688091,-0.3144697547,-0.0552008674,-0.2917206287]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/773","title":"Adding CC-100: Monolingual Datasets from Web Crawl Data","comments":"cc @aconneau ;) ","body":"## Adding a Dataset\r\n- **Name:** CC-100: Monolingual Datasets from Web Crawl Data\r\n- **Description:** https:\/\/twitter.com\/alex_conneau\/status\/1321507120848625665\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1911.02116\r\n- **Data:** http:\/\/data.statmt.org\/cc-100\/\r\n- **Motivation:** A large scale multi-lingual language modeling dataset. Text is de-duplicated and filtered by how \"Wikipedia-like\" it is, hopefully helping avoid some of the worst parts of the common crawl.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":3,"text":"Adding CC-100: Monolingual Datasets from Web Crawl Data\n## Adding a Dataset\r\n- **Name:** CC-100: Monolingual Datasets from Web Crawl Data\r\n- **Description:** https:\/\/twitter.com\/alex_conneau\/status\/1321507120848625665\r\n- **Paper:** https:\/\/arxiv.org\/abs\/1911.02116\r\n- **Data:** http:\/\/data.statmt.org\/cc-100\/\r\n- **Motivation:** A large scale multi-lingual language modeling dataset. Text is de-duplicated and filtered by how \"Wikipedia-like\" it is, hopefully helping avoid some of the worst parts of the common crawl.\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\ncc @aconneau ;) ","embeddings":[0.019558955,-0.2458286881,-0.1407929659,0.0268324167,0.0210234877,-0.0069015389,0.0924171731,0.1932940036,0.2508240044,0.1388318241,-0.2297824919,0.0172861423,-0.0723891258,0.4199198782,0.3427250385,-0.0529735088,0.0437292159,-0.1144956797,-0.1201308817,-0.1344495118,-0.1145138219,0.0710338727,0.1586278826,-0.2492505163,-0.1892216206,-0.1295406967,-0.1623324454,0.0898537189,-0.2325121611,-0.2987708151,0.0484831035,0.2808976471,0.1874767542,0.3730974793,-0.0000978591,-0.1484959275,0.3082101643,-0.1089620814,-0.4055474997,-0.1219775826,-0.0863632485,-0.1656171083,-0.0293379053,-0.1493299901,-0.2100080401,0.1484013945,-0.1826635301,-0.3099412918,0.2331986278,0.0586921349,0.3411509395,0.0515444949,-0.1264812797,-0.2095034868,-0.134409979,0.0987801254,-0.2047049403,0.4086226821,0.4788091481,0.1310467571,-0.0368922986,0.4958826005,0.011832254,-0.2205280811,0.0328529663,-0.2485093027,-0.0889576674,-0.2924669981,0.4486075938,0.5274826884,0.3609340787,-0.1307183951,-0.0979034305,-0.0842869133,-0.0833579451,-0.3118236363,-0.0303972866,0.3020348251,0.01247379,0.3431238234,-0.1748789549,-0.3682477176,0.0526489466,0.1340465695,-0.1012530327,0.374397099,-0.0466002226,-0.1761053056,0.2330160588,-0.178849265,-0.1637152284,-0.1707095057,0.0841050893,0.0898214504,-0.2419683337,-0.1954806447,0.3108842373,0.2363536209,0.4654799998,-0.0100714555,0.2817516923,0.215299502,-0.2693049908,0.0531920604,0.0554700606,0.1024267375,0.0092416648,-0.1768706739,0.3070274293,-0.0813275725,0.142281577,-0.008009987,0.0949930623,0.1607337594,-0.1921677291,-0.0939933136,0.0002809213,0.0507376529,-0.1964688152,0.2133120149,-0.2160896212,-0.0881870985,-0.0176079683,0.2856505215,0.251060605,0.1129023433,-0.0612481982,0.1644828469,-0.2837659419,-0.3191107213,-0.2630117238,0.1083738133,-0.1493460983,0.0674037263,0.1550023407,-0.086929366,-0.1096210107,-0.0828677118,0.2581441402,-0.004740966,0.0328102075,0.1159557477,-0.0129310731,-0.0930903926,0.0682386756,-0.0111959781,0.0085113104,-0.3455114663,-0.2770397365,0.0578536317,-0.1529181153,-0.1296080798,-0.3322165012,0.305549264,-0.1291368157,-0.2227707356,0.025774708,0.6889924407,0.0633791238,0.2916725278,0.2131776214,0.2518178225,-0.4083694816,-0.0086570149,0.0158470478,0.454177916,-0.1194418743,-0.1639713943,-0.0817381367,0.2107457668,0.1430377215,0.2737182379,-0.1329913437,-0.0202466752,0.0414150432,0.2591582835,-0.061844606,-0.2123963535,-0.1696907282,0.162947759,-0.1814220548,0.0437634215,0.0653368309,-0.1051054001,0.0494387187,0.0169843957,0.0813686401,0.2537022829,0.0015189998,0.1831729263,-0.2716716528,-0.217121318,0.3542041779,0.0622784607,0.0345298126,-0.2712417841,0.1570064574,-0.3200773299,0.3340692818,-0.3511089087,0.2514114976,-0.0095913531,0.2183507532,0.0710518956,0.180694133,-0.198560074,-0.2129664272,0.1470086426,0.0911539942,0.5386317372,-0.1650795043,-0.3207189143,-0.2928385735,-0.0793835521,-0.2214478701,-0.2008801848,0.2797168791,0.0977504924,-0.0199926719,0.229910627,0.0095961876,0.1924323738,-0.1565918028,-0.172799781,-0.1295485944,0.2533018887,-0.1796920896,0.1942432672,0.0597393662,0.4072172344,0.0240919013,-0.0340761133,0.0178555176,0.070624359,-0.3835492134,0.3688845038,0.5324430466,0.0307071917,0.4753707349,-0.5629479289,0.156393379,0.0988296047,0.1895219684,-0.0763363242,-0.0671690255,0.1608071923,0.1406563669,0.0454203226,0.053161826,0.0549321659,0.4579496086,-0.0446522608,-0.1067942828,-0.1989095956,0.2756573558,0.1451415122,0.3557890058,0.010574081,-0.4160343707,0.269554764,0.2469158322,-0.1409265399,0.3581875265,0.1919935048,-0.134202227,-0.3277059197,0.1519975364,-0.0735122263,-0.1204088554,0.2630755305,0.2457788587,-0.1333987713,0.0599472336,-0.1932128072,0.1192267686,-0.0656176805,-0.0026207473,0.1292984784,0.2423623651,-0.0341373198,-0.6187634468,-0.1480258703,-0.1066490039,-0.1053244174,-0.0087514594,0.0502373911,-0.2998757064,-0.212210387,-0.2236091793,-0.2370548695,-0.3498994708,-0.222333923,0.1589159518,-0.2296414673,-0.2210365683,0.0437825471,-0.1670891941,0.3609310687,-0.0514609031,0.1903482676,-0.146911636,-0.4738811851,0.0377512611,0.2415796667,0.0865053684,-0.0856188312,0.4025804996,-0.1345062256,0.1637924016,-0.1383548528,-0.5799003243,0.2076095343,-0.1775819212,-0.1691960245,-0.0769415349,-0.0828277767,-0.282216996,-0.1716155112,0.0882818699,-0.0583525263,-0.0875802562,0.0583322532,-0.0080064936,0.0358752199,-0.157517314,-0.2355936766,-0.3475712538,-0.298620522,0.0782110766,0.0942434967,0.0408081189,0.2560202479,0.0209715646,0.1061018333,0.0605710894,0.1603265256,-0.4990579784,-0.061362341,0.137726754,-0.3708117604,-0.291683048,-0.2793731987,-0.2789724171,0.1874396354,-0.0278580245,-0.0745265782,0.236430645,-0.1619305611,0.0138486782,0.1831455529,-0.0020075177,0.211910665,0.0978290439,-0.2152866125,-0.0926851928,0.06590119,-0.1396343857,0.2063630074,0.2098786682,-0.0066532847,0.0015716791,0.2086409032,0.6755363345,0.0307241417,0.1668455601,0.3882454932,0.1991461962,0.307472527,-0.1415151507,-0.369156599,0.1912224293,-0.1212691218,0.0689029694,0.5518729091,0.2474448234,-0.4288788736,-0.2415324003,-0.2114653736,-0.3021586239,-0.2667426169,0.0003897541,0.1350618452,0.1538924128,0.1310902536,-0.1356113702,-0.1088466868,-0.2052305341,0.0118160285,0.172673136,-0.2635315955,-0.0350015052,-0.3038572371,-0.2049835622,-0.8291309476,0.3415456116,0.2582693994,-0.0929087922,0.0327879414,-0.2342918366,0.0040922365,0.0649207085,-0.053489428,-0.3294539452,-0.0238073375,0.0639211908,0.0027842592,-0.0717481971,0.2534622252,-0.2141213268,0.3189731538,0.0981223434,-0.3130387068,-0.2449313849,0.069877468,0.1640625894,0.0984380618,-0.0184433181,-0.0091657601,-0.1830419302,-0.4018121064,-0.3176680207,0.0975057036,0.0931517482,0.3267916441,-0.0621930435,-0.0985588059,0.2693282068,0.0469869636,0.2671407163,0.206699118,-0.1088641956,0.3983151913,-0.0626179501,-0.0069185765,0.0558112897,-0.0529253706,0.525780797,-0.2254241258,-0.1997720748,-0.0858204514,-0.3426140845,0.1478449553,0.3237369657,0.3239377439,0.121394828,0.3233345747,-0.0729576498,-0.3016679883,0.0063750041,-0.1696735173,0.2496765256,-0.5083060265,-0.2028451562,0.3726942241,-0.0756934136,-0.1164898425,0.2815102935,0.0863811523,-0.3699843884,0.2073044181,0.0482583977,1.0934170485,0.0378021002,0.0797071829,-0.0892126039,0.179354623,0.3126837313,-0.1548178643,-0.0278431699,0.033688996,0.0600886382,0.020133391,0.2553302944,0.0137846014,0.1084748954,-0.2370541394,0.069730632,0.1137361899,0.2313947976,0.0689331815,0.3527994752,-0.1567326337,-0.3920912147,0.0129908156,0.2877604961,-0.0142307216,0.1240747422,-0.1320296973,0.0112924473,-0.0635243878,-0.2321610451,-0.3157537878,-0.1300537586,-0.0645562485,0.1693513989,-0.0983776003,-0.5530154109,0.2707975805,-0.0645507202,0.2124875337,0.1603774279,-0.3756097555,0.1424481869,-0.0126968082,-0.0184209757,-0.04170461,-0.075096935,0.254973501,0.0145845162,-0.3632721007,0.0269495863,0.3738246262,-0.0649622455,-0.1652504057,0.1523855776,-0.1075834855,0.0164418593,0.1307674199,0.3177154064,-0.2229151726,-0.3201321363,0.2758306861,0.1101215705,-0.0453927293,-0.1118017361,0.3097019494,-0.0271703191,-0.1288572699,0.2009067088,0.4519582093,0.0114871599,-0.0581445433,0.3753163517,-0.335467726,-0.2953867316,-0.1445604563,-0.116127342,-0.0827377215,0.0511339493,0.0060389708,-0.2575618923,-0.3616196215,0.2852728963,0.015123005,0.1142554954,0.0772675052,-0.0641471297,-0.0852163807,0.2860579491,-0.1048208997,0.1316508502,0.1622338593,0.2614533007,0.1349839121,0.1853707135,-0.4973127842,-0.0355070084,-0.0929123834,0.0071822964,0.3457914293,-0.2495725006,0.0403999053,-0.1101836115,0.1768378317,0.0772113204,-0.1299849004,-0.3995573819,-0.1381450146,0.0644889921,0.0048898687,-0.1332053244,0.2155862153,0.0624623224,0.1555373073,-0.246427983,0.1654261202,0.1099605411,0.109224245,0.122427009,-0.1367253214,-0.0854183212,0.0307878349,0.0252786558,0.0170423966,0.3796521723,0.0624494813,-0.0711349994,0.2760121822,-0.026377961,0.131242305,0.3368937671,-0.0143773677,0.0406746604,0.198781386,0.1964373142,0.170100823,0.4569205046,0.3109584749,-0.1597716957,-0.264097333,-0.0032521321,0.1107328609,0.3617193699,-0.4293848872,-0.0890872851,0.2017124593,-0.2224202752,0.3496288359,0.2883780301,0.3803534508,-0.1070593819,0.0640478507,-0.0685043558,0.2551399469,-0.0465776399,0.1782621592,0.0389080606,0.1245658845,0.0571152195,0.2227455378,-0.007131665,0.2035651356,0.2585514188,-0.0822005197,0.135156408,0.1036763936,0.1982891709,0.1639811099,-0.125375703,0.4556225538,0.1993482262,-0.3189878464,0.2262587547,0.0185355078,0.6119368672,-0.0556796752,0.1829176694,-0.2286576927,0.3193634152,-0.0991486013,-0.0798503757,-0.1340055168,-0.2147182524,-0.2670286,-0.0964690894,-0.0047495407,-0.0982402936,0.2147307396,-0.0509726293,-0.0671592578,-0.3928829432,0.1925808787,-0.0888626501,0.220148012,-0.1688978523,0.1299061626,0.3411770165,-0.1344313473,-0.0542614721,0.2736620009,0.1365206987,-0.2446054518,-0.1911772043,0.1870629191,-0.1402547508,-0.1559617817,-0.131689027,0.0842768997,0.256539315,-0.0517839529,0.3244769275,0.2099992931,-0.2784639299,-0.1579649597,0.5813961625,-0.0687161461,-0.1405733079,-0.2861142159,-0.0357260667,-0.018026717,-0.3723568916,0.0691990182,-0.4322626889,0.1830631196,0.1670181006,-0.2073450983,0.0982797593,-0.4162221551,0.1447786242,0.0815115571,0.4760172069,0.2228267342,0.0055108769,0.0817984939,-0.2737545967,-0.3413495719,0.0883633271,-0.3253697455,0.0877840742,-0.3026158512,0.2402378172,0.2388957441,0.21069175,-0.0160635225,0.2080584913,0.0607749522,-0.1910277754,-0.1565054059,0.0583258159,-0.0950065702,0.1093187854,-0.0385372415,-0.1216002703,0.0949980989,0.0427342318,0.1630875766,-0.2648103535,-0.1893464923,-0.0671083927,0.1017249227,0.294367522,0.1419875771,0.3941188455,-0.2851956189,-0.1314961612,0.1162795275,-0.2172035277,0.0310829859,0.4171354771,0.1171302721,0.0828169584,-0.0998035893,-0.2595484853,0.243877396,0.4825592339,0.0999961272,0.0367981978,-0.4139083326,-0.0024925678,-0.1404554397,-0.2758323848,-0.0571981221,0.4641697109,-0.0945583358,-0.0226088054,0.0923150852,-0.6629714966,0.0574688166,-0.1480608881,-0.3260959387,-0.29876104,0.0992474407,0.2374415845,0.1201201603,-0.2811151147,0.1766601205,0.1057725623,0.2473517209,-0.0978381038,0.2894420028,-0.095385626,-0.3460719883,-0.1351392567,0.1749862432,0.2377533019,-0.1289899647,-0.0760029107,-0.3905571401]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/771","title":"Using `Dataset.map` with `n_proc>1` print multiple progress bars","comments":"Yes it allows to monitor the speed of each process. Currently each process takes care of one shard of the dataset.\r\n\r\nAt one point we can consider using streaming batches to a pool of processes instead of sharding the dataset in `num_proc` parts. At that point it will be easy to use only one progress bar","body":"When using `Dataset.map` with `n_proc > 1`, only one of the processes should print a progress bar (to make the output readable). Right now, `n_proc` progress bars are printed.","comment_length":56,"text":"Using `Dataset.map` with `n_proc>1` print multiple progress bars\nWhen using `Dataset.map` with `n_proc > 1`, only one of the processes should print a progress bar (to make the output readable). Right now, `n_proc` progress bars are printed.\nYes it allows to monitor the speed of each process. Currently each process takes care of one shard of the dataset.\r\n\r\nAt one point we can consider using streaming batches to a pool of processes instead of sharding the dataset in `num_proc` parts. At that point it will be easy to use only one progress bar","embeddings":[-0.4709307849,-0.3086602092,-0.2045882791,0.0282438751,-0.1614840925,-0.0598893911,0.427454412,0.1954978704,-0.1440223157,0.5090811253,0.0353115238,0.2185556293,0.1020263955,0.5341004729,0.0146906162,-0.0062402082,0.0625615567,0.1021450013,-0.0488366038,0.2574520707,-0.0931532532,-0.0605296418,0.0501902178,0.0376354083,-0.0566342883,0.1142689139,-0.0659439117,-0.2918230891,0.2581534684,-0.643638134,-0.0705317333,0.4315348268,0.1479462981,0.2352978736,-0.0001087357,0.0400767028,0.4632191658,0.1836577058,-0.1190845594,0.0478734896,-0.3992361724,-0.4790298939,0.1656673849,-0.5045090318,0.2479811162,-0.0170508046,0.3479945958,-0.2314990759,0.1535726041,-0.0839441568,0.2232875526,0.2877654433,-0.4592035413,0.0202864818,0.1309508532,-0.1544233114,-0.0351945572,-0.2405854315,0.4689825177,0.3584709764,-0.3671169579,0.3515626192,0.0340411253,0.1914627105,0.02362841,0.1282078028,0.6972887516,-0.435975641,0.0484607182,0.2073627263,0.1045902818,-0.2849717438,-0.0844797641,-0.3046000004,-0.1938551217,-0.2056360096,-0.4532286525,0.157513544,0.0647703558,0.0178036876,-0.6200436354,0.1119635031,-0.1470354646,-0.1840860397,0.1084482968,0.0067434371,0.0264973752,0.4649553299,-0.0114118997,0.4518721402,-0.0388087146,-0.2027369291,0.1169731095,-0.213753745,-0.2411209792,0.0346872769,0.1544521153,0.0547254793,0.028870007,0.0516919158,-0.0968007892,0.1749569476,0.191467151,0.2314073592,0.2765748501,-0.2618946731,0.4729348421,-0.1703924537,0.2630069852,-0.2552232742,0.0469894223,0.1240871251,0.2900397182,-0.0927148014,0.1685600281,0.2086966783,-0.2425430119,-0.1615764052,0.0094220107,0.0010094575,0.2904830873,-0.2892686427,0.0285682678,0.3051282763,0.1149003133,0.0218197573,-0.3511643708,-0.0229732394,-0.0496004559,-0.1967031807,-0.1722453535,-0.1654002219,-0.2894493639,0.2183250636,-0.2506631911,-0.0741724297,0.0814437866,-0.1543471664,-0.0386420004,0.2655332386,0.1757421196,-0.3396120667,0.3054901361,0.141982764,0.0801868066,0.200136885,0.0496960767,0.2786524892,-0.2928563058,0.2514970899,0.0229024682,-0.2703835666,0.107291624,0.157570824,-0.1535101384,0.2138035595,-0.4039472938,0.5481806993,0.0917498395,-0.302537024,-0.1266032755,-0.0488536544,-0.624352634,-0.2712551057,0.0405892804,0.1322113872,-0.2416712493,-0.0713708326,-0.0160461143,-0.4697975814,0.3007001281,0.1141307354,-0.2179165035,-0.146421805,-0.0874192938,0.3367351294,0.0742602274,-0.3036816418,-0.2639031112,0.3852697313,-0.0571713448,-0.2511284351,0.4734428227,0.2763561904,0.3704321086,-0.3024572134,-0.3997048438,-0.0985456258,-0.2268814892,0.1177967712,-0.0596595816,-0.0823585913,-0.0747476667,0.1807522476,0.0954396352,0.056310311,0.2851853371,-0.5802689195,0.5668189526,-0.012318668,0.3039407134,-0.0066162306,0.0904614776,-0.278301239,-0.1680067778,-0.0736055672,-0.1742736846,0.3511590958,-0.3102012277,-0.1334916651,0.1653239876,-0.1894379407,0.1740193367,-0.0404673479,0.1339490414,0.0983711779,0.0937825367,0.098795563,0.087910369,-0.2542567849,-0.0760834217,0.231139183,-0.4155819118,-0.4498784244,-0.1657538563,0.2748689055,0.1072590724,-0.1088303775,-0.0145297553,-0.1064905599,-0.1093677431,-0.0450528674,0.1098114774,0.1773904711,-0.0653681755,0.3082703352,0.0906936303,0.3370721638,0.1831723452,0.0336816236,-0.0647954121,-0.2131910771,-0.0464835167,-0.1356776506,-0.0577809699,0.1218629777,0.0721362382,0.287117213,0.1541026831,-0.1059589982,0.0830248222,0.100044094,-0.297229737,-0.2367681563,0.1199104711,0.439856261,0.2085003108,-0.1657771617,-0.0968090072,0.4711030722,0.6808502078,0.1080581024,-0.3228969276,0.0744972527,-0.0657468662,-0.0953093767,0.3842852712,0.2983469963,0.4065902829,0.4276604652,0.2452062368,-0.0813861862,-0.1607411653,-0.1940162927,-0.1250543296,-0.0150070386,0.0925907344,-0.072960712,-0.0337651856,-0.2909654379,-0.0688744634,-0.4779720604,0.0624379255,-0.0120142875,0.0565489158,-0.1840170622,-0.1028188393,-0.1075433865,-0.1546378136,0.0755565241,0.2769264877,-0.3682191372,0.0886848047,0.3004539907,-0.1737615615,0.1589633673,0.3338918388,-0.0382236466,-0.1383941472,0.4887757003,-0.3873391747,-0.3365318775,-0.0283084661,0.0608985387,0.0255718511,0.1921042353,0.4016205072,-0.1541722417,0.0286473557,-0.5659902096,0.1881171167,0.0100368569,-0.2315947711,0.0740462989,-0.1519623697,0.0504134148,-0.1466164738,0.1371094584,0.1928319931,-0.4140234292,-0.0366277359,-0.1570467949,-0.0845530629,-0.2458758056,-0.0915630013,0.0465402231,-0.5235555768,-0.3526059985,0.1706472486,-0.0176661536,0.3490067124,-0.3921487927,-0.2120516598,0.252512753,0.1036813855,-0.0600114278,-0.1542237997,-0.4738194942,-0.1403233409,-0.041454792,-0.2489225417,-0.0744524896,0.1832370162,-0.0684680194,0.1845680177,-0.3350621462,-0.1403236687,-0.3038953841,0.2524101436,0.2147917897,0.2422604561,0.3234031796,-0.0331825353,-0.1720406413,0.037523672,-0.1058023497,0.144033134,0.0208565891,0.0050096679,-0.097892791,0.1340559572,0.029406894,0.5478230119,0.4378448129,-0.3881170154,0.0392033942,-0.1871100217,0.0481577776,-0.0191749223,0.0773201957,0.1866517961,-0.159836337,-0.0833100006,0.2705334425,-0.1076511964,-0.1935266703,0.0353363641,0.2834667265,-0.0969982147,-0.1716634333,-0.0966486782,0.1232764795,0.2239535898,-0.046775803,0.0259045791,-0.0451237522,-0.0510886908,-0.3012230098,0.0600423589,0.3548401892,-0.1982379407,-0.1405563653,-0.1563342214,-0.7523559332,0.3945201039,0.1789794564,-0.2582481503,-0.0577145852,-0.0581943877,0.0475282259,-0.0013057127,0.4540849328,-0.2986177206,-0.1308717728,0.0290299524,-0.2905183434,-0.1905002892,-0.1094192043,-0.2077116817,0.321266681,0.1919551045,0.1171801016,-0.2532632649,-0.0952740461,0.039236784,0.0910213962,-0.1395304203,-0.3234478831,-0.3122164309,0.0280114636,-0.0952098295,0.1391064674,-0.2067492604,-0.1781121492,-0.3750582337,-0.4214895368,0.1731848866,0.0366111286,-0.0659501404,0.2938268781,0.0106505966,0.1366571784,0.1566676497,-0.0006709912,0.3739665747,0.2307821065,-0.0054620537,-0.0372016169,0.0880712494,0.337656498,-0.251706183,0.1992788464,0.2213506103,0.1350201815,0.2246278226,0.2069747299,0.3590278327,-0.1791724265,-0.0434530638,-0.0544576496,0.1770003587,-0.3458555341,-0.1547007114,0.3590994477,0.2664080262,-0.1463781148,0.3667649329,-0.4146936238,-0.0973809287,-0.0846942738,0.0067301169,0.6404374838,-0.6624414921,-0.001204261,-0.0647825673,-0.0951276422,0.2084502131,0.136594817,0.0811423138,-0.5132695436,-0.3465443254,0.0587400533,-0.0803439692,0.1771927923,0.3014800251,-0.4407893419,0.1053336486,0.3916264474,0.001831464,-0.290468961,0.106615223,0.1912131757,-0.0743321702,-0.2516307235,0.1422360539,0.0133031271,-0.386944294,-0.1404528767,0.1109565347,0.1591571569,0.057625223,-0.0779486895,-0.2499789298,0.2778437436,0.2395842075,-0.1950539201,-0.2763399482,-0.1448337287,0.2340297848,0.0931279957,-0.2261737734,-0.0485870652,-0.0528202504,0.1196554899,0.2626811564,-0.0149985766,-0.2381752729,-0.0223779958,-0.1648025066,-0.3957273066,0.4399856925,0.010797034,-0.5087302327,-0.0926297158,0.4703203738,0.2749843895,0.2117157429,0.2134143859,0.3843767941,-0.4483097196,-0.1238320246,0.0901229382,0.3654623926,-0.3199890852,0.6993054152,-0.1303531826,-0.3771055043,-0.1530619264,0.053110972,0.1262799352,0.2108528018,-0.2507434487,0.0740582198,-0.1265089959,-0.2178712785,-0.0359621793,-0.0132286381,0.061245136,0.1062274203,0.0704891607,0.176189959,0.0199555345,0.1330261379,-0.0340776816,0.2667894661,-0.2423925698,-0.1095247418,-0.4801993668,-0.0197232328,-0.1943641901,0.4233628213,0.0978430882,0.0701469332,0.0185034238,0.226143077,-0.3720760345,-0.2011400163,-0.3223800361,0.3809590638,0.0674004927,0.0509366579,-0.0165480096,0.0326886661,0.1133495048,0.2673209012,-0.5762633085,-0.2781189978,0.1331930906,0.127991721,-0.1458604485,0.039799355,-0.1244824603,0.2529825568,-0.0980201364,-0.1641075462,0.2600605488,0.2658551633,-0.087669082,0.0391850173,0.3612810671,-0.0274142884,-0.1187922806,0.0602248125,0.0727331638,0.2029426992,-0.2687400579,0.4438777268,-0.0692406744,-0.0369615741,0.2527549863,-0.0771589279,0.4333483577,0.3147988915,-0.1164795533,0.1425156593,0.1690786332,-0.1300194263,0.6643269658,0.2362796664,0.0557555296,-0.2299117595,0.1429085135,0.2428239584,-0.0022660075,0.0768529549,0.4340244234,-0.0005965632,0.4263560474,0.1969960183,0.2532644272,0.1968126893,0.2387292236,0.0222468842,0.090924643,-0.4251683354,-0.0452213734,0.0376616716,0.2356757075,0.10962338,0.2234322727,0.0046122214,0.023478929,0.1442659348,-0.0361306816,0.6295459867,0.0342194736,0.0963837728,-0.1208274439,0.101235047,-0.0883341208,0.1687146872,-0.4824109674,0.00890816,0.1505179256,0.3759705424,-0.1967581511,-0.0613402613,-0.1119130179,0.1956705302,0.0537944809,-0.1770872474,-0.180997625,-0.0394151285,0.0887821838,0.1929963082,0.1651941687,0.2455890179,0.6638801694,0.1000951827,0.2418382019,-0.1280755997,-0.0973651633,-0.0180218481,0.3237379789,-0.1522717476,-0.2086367756,-0.1455534995,0.0961072892,0.1927004904,0.0040111071,0.0950795636,0.0377642512,-0.0075850864,0.0366202481,0.0180714857,0.0439565144,-0.2682685554,-0.0291495826,0.2027613372,-0.1386905909,0.3927569091,0.164030835,-0.1708239317,0.4104486108,0.0363222361,0.0448455289,-0.2567974627,0.3423312902,0.0226628743,-0.2043817192,0.0877961665,-0.1102412269,-0.253087014,0.0432952419,0.4033915699,-0.5760534406,0.0838309154,-0.0725017786,0.1238424405,-0.2732954621,0.4418148696,0.0909436718,0.5844121575,-0.2890387475,-0.1479343474,-0.2420908958,0.0854978636,0.0594621412,0.0200902335,-0.1961842328,0.3414389491,-0.0361194201,0.4042432308,0.4936386943,-0.2218496203,-0.1770952195,0.3003432751,-0.2015597224,0.1698973775,-0.1292448193,-0.105120115,0.0681951791,-0.2566551268,0.2003932148,0.0981537849,0.1799392104,0.0307900067,0.0592922159,0.3669612408,0.0161051154,0.4975039661,0.0005671917,0.3057104051,-0.2027689368,0.1349598914,0.0104377316,0.0972570479,-0.0923429653,0.1135526672,0.0688799471,0.3002985418,-0.2168045342,0.2165971696,-0.0770048797,0.1311831623,0.2174637765,0.0848717764,0.3008787632,-0.1078954265,0.0197880659,0.0333395973,-0.1427423656,0.2304473817,0.2945032418,0.0949380994,-0.0765371099,0.0933339968,0.1654152274,-0.381416291,-0.2248799503,-0.2506607175,0.1779890209,-0.2140320092,-0.0164401289,-0.2841692865,-0.3223105669,0.257937789,-0.1390108466,-0.3697535694,0.1706743985,-0.0640493408,-0.354069531,0.0441492908,0.0142035475,0.1825512797,-0.2490623742,-0.3193565011,-0.4881700575]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/769","title":"How to choose proper download_mode in function load_dataset?","comments":"`download_mode=datasets.GenerateMode.FORCE_REDOWNLOAD` should work.\r\nThis makes me think we we should rename this to DownloadMode.FORCE_REDOWNLOAD. Currently that's confusing","body":"Hi, I am a beginner to datasets and I try to use datasets to load my csv file.\r\nmy csv file looks like this\r\n\r\n``` \r\ntext,label\r\n\"Effective but too-tepid biopic\",3\r\n\"If you sometimes like to go to the movies to have fun , Wasabi is a good place to start .\",4\r\n\"Emerges as something rare , an issue movie that 's so honest and keenly observed that it does n't feel like one .\",5\r\n```\r\n\r\nFirst I try to use this command to load my csv file . \r\n\r\n``` python\r\ndataset=load_dataset('csv', data_files=['sst_test.csv'])\r\n```\r\n\r\nIt seems good, but when i try to overwrite the convert_options to convert 'label' columns from int64 to float32 like this.\r\n\r\n``` python\r\nimport pyarrow as pa\r\nfrom pyarrow import csv\r\nread_options = csv.ReadOptions(block_size=1024*1024)\r\nparse_options = csv.ParseOptions()\r\nconvert_options = csv.ConvertOptions(column_types={'text': pa.string(), 'label': pa.float32()})\r\ndataset = load_dataset('csv', data_files=['sst_test.csv'], read_options=read_options,\r\n parse_options=parse_options, convert_options=convert_options)\r\n```\r\n\r\nIt keeps the same:\r\n\r\n```shell\r\nDataset(features: {'text': Value(dtype='string', id=None), 'label': Value(dtype='int64', id=None)}, num_rows: 2210)\r\n```\r\n\r\nI think this issue is caused by the parameter \"download_mode\" Default to REUSE_DATASET_IF_EXISTS because after I delete the cache_dir, it seems right.\r\n\r\nIs it a bug? How to choose proper download_mode to avoid this issue?\r\n","comment_length":17,"text":"How to choose proper download_mode in function load_dataset?\nHi, I am a beginner to datasets and I try to use datasets to load my csv file.\r\nmy csv file looks like this\r\n\r\n``` \r\ntext,label\r\n\"Effective but too-tepid biopic\",3\r\n\"If you sometimes like to go to the movies to have fun , Wasabi is a good place to start .\",4\r\n\"Emerges as something rare , an issue movie that 's so honest and keenly observed that it does n't feel like one .\",5\r\n```\r\n\r\nFirst I try to use this command to load my csv file . \r\n\r\n``` python\r\ndataset=load_dataset('csv', data_files=['sst_test.csv'])\r\n```\r\n\r\nIt seems good, but when i try to overwrite the convert_options to convert 'label' columns from int64 to float32 like this.\r\n\r\n``` python\r\nimport pyarrow as pa\r\nfrom pyarrow import csv\r\nread_options = csv.ReadOptions(block_size=1024*1024)\r\nparse_options = csv.ParseOptions()\r\nconvert_options = csv.ConvertOptions(column_types={'text': pa.string(), 'label': pa.float32()})\r\ndataset = load_dataset('csv', data_files=['sst_test.csv'], read_options=read_options,\r\n parse_options=parse_options, convert_options=convert_options)\r\n```\r\n\r\nIt keeps the same:\r\n\r\n```shell\r\nDataset(features: {'text': Value(dtype='string', id=None), 'label': Value(dtype='int64', id=None)}, num_rows: 2210)\r\n```\r\n\r\nI think this issue is caused by the parameter \"download_mode\" Default to REUSE_DATASET_IF_EXISTS because after I delete the cache_dir, it seems right.\r\n\r\nIs it a bug? How to choose proper download_mode to avoid this issue?\r\n\n`download_mode=datasets.GenerateMode.FORCE_REDOWNLOAD` should work.\r\nThis makes me think we we should rename this to DownloadMode.FORCE_REDOWNLOAD. Currently that's confusing","embeddings":[-0.2590157092,-0.1065792218,-0.0324916057,0.0735094324,0.3922304213,-0.0588000938,0.4731936157,0.0939230695,0.1750782579,-0.0883600116,-0.2146189511,0.1605177373,0.2840963006,0.1457755417,-0.0188392736,0.0348515958,0.1505480856,0.07911475,-0.2978679538,0.0446997993,-0.4603429437,-0.0685208291,-0.215477407,-0.1834250838,0.1294441372,0.1641882956,0.167755276,0.184518382,-0.150361836,-0.4337638021,0.4201339185,0.1227659062,0.4481950402,0.0825559869,-0.000126469,0.0161511004,0.2499263585,-0.360117048,-0.1838915199,-0.339320004,-0.1870489419,-0.2415881604,0.2389954478,-0.3605295718,0.0253108777,-0.1680576503,0.0001144358,-0.1210178584,0.1644955873,0.3842561841,0.1007685661,0.0781160295,-0.1188412756,0.1853100359,0.1774165183,0.1141077802,-0.0534457266,0.4895187616,0.3477607071,0.2589950264,0.0926229209,0.0057656784,-0.2732252777,0.1277932823,0.4341212213,0.0734032914,0.5844093561,-0.204193458,-0.0044305627,-0.0334150158,0.8179601431,-0.2546541393,-0.1575488448,-0.2335109711,0.0608678162,-0.3556732833,0.1805979908,0.0530941375,-0.0197090209,0.2729556859,-0.3150651157,-0.1112685651,-0.0100988643,0.3499962986,0.0204178467,0.0267175604,-0.1969670653,0.3487562239,0.297935456,0.0454990976,0.2575612068,-0.4173841774,0.2050903738,0.033781372,-0.337611258,0.1275116354,-0.152001664,0.2336505055,0.2249239981,0.2109157592,-0.1419799477,-0.1402306408,-0.0618305616,0.040321894,0.2665914595,0.068963252,0.1517966986,-0.3156386018,0.020187173,0.0933016464,0.201643914,0.078651011,-0.0195763633,-0.2522189617,0.5681581497,0.0203414224,0.1890556067,-0.1091510504,-0.295209825,-0.1585887969,-0.0716224611,-0.1131579056,-0.0582123809,0.3231161237,0.0163159464,0.1556693166,-0.1274563074,0.4184004068,-0.2559093833,-0.130095467,-0.0085907727,-0.185862869,-0.2180095166,-0.1199804693,0.320186317,-0.2929286063,0.2597480118,0.3904041052,-0.186355859,-0.1730380207,-0.0431318879,-0.0347855613,0.0328734703,0.4530647099,0.0829750448,0.2083443701,0.2178869098,-0.1803499609,-0.0289740078,0.5776111484,-0.3491699994,-0.4118051529,0.0366017297,0.0541905016,-0.2928566635,0.0450675376,-0.3806696534,0.1129915193,0.0654720515,-0.2287005633,0.0757200345,-0.3123002946,-0.002802257,-0.345122993,-0.0431127809,0.7118008137,-0.5010120869,-0.0086220037,-0.2906814814,-0.2806813121,0.3792184591,0.0094999373,-0.3852303028,-0.0082766498,-0.2139983177,-0.1566238254,0.4293205738,-0.252058059,-0.5417465568,0.2455400378,-0.0221253615,0.0183113739,0.2990805209,0.2141675949,0.158557266,-0.1784640551,-0.1892070621,0.2094290257,0.0773767531,0.1481962353,-0.0499617532,-0.2345117033,0.2615923285,0.0459154323,-0.2453910857,0.2021593153,0.0001852767,-0.2519760132,0.1741706878,0.0604267865,-0.1360568851,0.0143710244,0.105925262,0.1329714209,-0.1386200041,0.1003473923,-0.7810969353,-0.0219680034,0.2704707384,-0.2892319262,-0.2882654965,-0.1170225292,-0.4807579219,-0.1440787017,-0.2776653767,0.0687077045,-0.0466013327,0.2238108218,0.1725722104,-0.0634523258,-0.1081570163,0.2495522797,-0.2250650972,0.2756273448,-0.0094522983,0.2079646885,0.084070839,-0.0544580705,0.067093581,-0.3101555407,0.344012171,-0.2031056136,-0.1752134115,0.573730588,0.2345304936,0.2890976071,-0.5119717121,0.0518639535,0.1824987084,0.1577418596,-0.0079128919,-0.0924282596,0.2343434095,0.0416088216,-0.2178079039,0.1471984088,-0.2413241565,0.0972039402,-0.173224315,-0.1812309325,0.2972182035,0.0828680322,-0.1139889583,-0.1572600752,-0.148605302,0.1153389961,0.2650654316,0.1879075617,-0.1227846444,0.0679866821,0.365275234,-0.2784500122,0.1460534036,0.1251476258,-0.0094349841,-0.2080771327,0.3465999663,0.1289145499,0.6112674475,0.065418452,-0.0754428431,-0.0606172048,-0.1442533284,-0.2393881977,0.4105067849,-0.0799897015,0.3619145155,0.1431763917,-0.004822968,-0.3801917434,-0.4402510524,0.2052964717,0.3782948554,0.240487203,-0.4643081129,0.0005684635,-0.3054567873,-0.1668855697,-0.1141233817,0.0788459331,0.1046214625,-0.1976700425,-0.310896337,0.3138912022,0.0799465552,0.2038165629,-0.2323596627,-0.1495496333,0.0818993971,-0.7005093694,0.0314336754,-0.0007050142,-0.318924576,-0.1002830863,0.2216484547,-0.1986538917,0.1419191211,-0.3100055158,-0.0023085154,-0.3742949665,0.0458809547,-0.0204491206,0.0504450724,0.2778349519,0.0114183743,0.3690236807,0.0027449108,0.1271579266,0.2017466128,-0.0212167986,0.0861169845,0.1441334188,-0.1074910983,0.1517088711,-0.1264477372,-0.650013864,-0.2897948921,-0.1877052039,-0.1765018851,0.0215188917,-0.0308007076,0.1151905283,0.2868536413,-0.1826750636,0.1109595895,0.1347949505,-0.3872719407,-0.1322165877,0.4630843699,0.1229642555,-0.1389097422,-0.0985699594,-0.0331159942,-0.016652301,0.3244376481,-0.2626566291,-0.2504927516,-0.1454032809,0.3590501249,-0.0520164743,0.3368940651,0.2954981625,0.0277527533,0.1441572458,-0.0169327408,-0.2331857383,0.0601134934,0.2177973539,0.1577083319,0.3307515085,0.4238836467,-0.1738361865,0.4840948582,0.1531599015,-0.08593788,0.3802889287,-0.14450638,0.3086200058,0.1956124753,-0.1443714947,-0.0471145622,-0.1648572087,0.0203243289,0.2302609533,-0.1398406923,-0.1838411689,-0.1184366867,0.1428219825,-0.5527740121,-0.3069429398,0.1033210158,0.3018544912,0.1691403985,0.0940616354,0.2061573416,-0.1130303964,-0.0955078304,-0.0175239332,0.5914154053,0.160253942,0.192405358,-0.1053320616,0.1548231989,-0.1293418705,0.2942940295,0.1301005632,0.5032503605,0.2339003831,-0.0042753718,0.1287279129,-0.063338317,0.841311872,-0.158316344,0.1861680299,0.0334561691,0.3317713141,-0.2436377853,-0.1645714194,-0.1117376909,0.1501456946,0.0081763826,0.0972548276,-0.3879661262,-0.1228454039,-0.0531562977,0.4365920424,-0.0659547001,-0.4471813738,-0.2841564715,-0.4376521409,-0.4462884665,-0.0384515077,0.0179048739,0.0549997613,-0.2249456346,0.0436707921,0.1634768695,0.0042288899,0.038498044,-0.0271527506,0.1122301817,0.154384613,-0.01488287,0.0019931556,-0.3943810165,0.2264147252,0.5883174539,-0.2196050286,-0.3321849108,0.0661516339,0.0548972748,0.1920361817,0.1493545622,-0.2415282726,-0.076754868,-0.1030721664,-0.0792036504,-0.1488855481,0.0430940092,0.2652294338,-0.0753985792,-0.1187168136,-0.8903482556,0.5784325004,-0.0845570415,-0.150904119,0.0467396416,0.2526814342,-0.1398811042,-0.0810653046,0.0712349489,0.764534831,-0.2446301728,-0.0281876028,0.3306536376,-0.0313084722,0.5819005966,-0.2537856102,0.1236410663,-0.2166161239,0.1507577896,-0.0235891789,-0.2288700938,0.3793826699,0.1672848165,-0.116410628,0.4459980726,-0.1671122909,0.4117119908,0.0936771184,0.127107963,-0.3512286544,-0.2483047694,-0.2595976293,0.0272984263,-0.2073265314,0.1427218765,0.0314566717,0.1156668812,0.0584837645,0.1598692238,-0.0451962166,-0.1212367192,-0.2295788378,0.0017908625,-0.3966817856,-0.2720352709,0.1914281994,0.5301158428,-0.0140773281,-0.0348480828,-0.0856881142,0.0302889552,0.042707514,0.1711218208,-0.1391275823,-0.0649025142,0.310939163,0.0959453881,-0.2223520726,0.124022007,0.023286378,0.0939395577,-0.0355423018,-0.1313208491,0.2292875201,-0.4811432362,-0.5293102264,-0.0353401415,0.118579343,0.0136947809,-0.0309410822,-0.0769595876,0.1665363908,0.2045455128,-0.3246525228,-0.370827198,-0.2980579734,0.347121805,0.1109898835,0.1481265873,0.0971216038,0.0001112359,-0.1615226865,-0.0228978842,0.0979121178,-0.3864884079,-0.474314183,0.117806226,0.0866523981,0.2270768136,-0.1768538654,0.057498496,0.1341666877,-0.0762520805,0.0526522473,-0.2332753539,-0.2053671181,-0.0183455553,0.230430007,0.0131689608,0.3149641752,-0.1650414169,0.1117767915,-0.5403917432,-0.1521462351,0.0065380055,0.2150774747,0.1250231117,0.154485181,0.172017023,-0.0364370607,-0.1557382345,-0.0178513341,-0.0280965343,-0.2507571578,-0.0906890854,-0.207592532,0.2264264971,0.039130263,-0.1974167973,-0.0468600802,-0.4403027594,-0.0294078868,-0.1761541218,0.216253832,0.2070033699,-0.0021215626,0.1487759948,-0.1941693872,-0.1165112182,-0.4289857745,0.043186672,-0.2956339419,0.2552612424,0.2014316767,0.2439677715,0.0786950067,0.0370713919,-0.4545849562,0.040139772,0.1246692538,0.1746946424,0.1457859725,-0.214253217,0.2793748677,0.0259606317,0.2397788912,0.5284750462,-0.1379821301,-0.1658487767,0.1686346084,0.0423345342,-0.2324798107,-0.0726399943,-0.113152571,0.0683512837,0.0421705842,0.4152803719,0.1428386569,0.1111135036,-0.0356445834,-0.1007110775,0.5492886305,0.2058724463,0.0787739977,0.7013623714,0.1059087813,0.1620962918,0.3812274039,0.2146875113,0.0375310443,0.2751406729,-0.0656058341,0.0851575285,0.1346253008,0.1996565908,-0.2870312333,-0.5832771659,0.3192648888,0.4408172071,-0.5928213596,0.1920885593,0.3697037399,0.2482479513,-0.2218416482,-0.0297818929,-0.2140135467,0.1440788358,-0.1356633306,-0.3109328151,-0.0124060456,-0.38986516,-0.0671014488,0.3456325233,-0.371106267,-0.2835859358,0.1845790893,0.1933751702,-0.0784093514,-0.5394612551,0.0342948064,-0.0206877459,0.2379596233,-0.0495142192,0.3212913871,0.3100027442,-0.0349971503,0.1355138123,-0.0280303694,0.37149629,0.1949170232,0.1508665979,-0.2217585146,0.2618805766,0.0822885111,-0.0570817366,0.0875394344,0.1370287687,0.418859899,0.237957865,0.1006568223,-0.0418211445,0.0913760513,0.2075316608,-0.3471577764,-0.3320117891,0.0766511783,-0.0371056721,0.0312490948,0.0664570332,0.0859541893,-0.1704254895,-0.0716075152,0.3773832321,0.0223165993,0.1777162105,0.0167677458,0.0174005479,0.0680174008,0.1567424089,-0.0432049669,-0.0552344769,-0.3518215418,0.2130565047,-0.7748298645,-0.1737463325,-0.0107038468,-0.2574958205,0.1445352286,-0.0978599265,0.0943769887,0.4485420883,0.1237157509,-0.0018752248,0.4510881901,0.283392638,0.2019158751,-0.3725009263,-0.0247972645,0.0199973378,-0.2211813331,-0.5889338255,0.2781899273,-0.233443141,-0.0461061038,-0.398005873,0.0939549133,0.3640525043,-0.0353138074,0.5893185139,0.0643815622,0.4532062411,0.0070295595,0.0586422943,0.1992734671,0.2542237937,-0.0262673106,-0.0073240534,-0.2787487805,0.2373551577,-0.2713417411,0.2070479989,-0.2745248377,0.2398115844,0.2244433165,0.0214670897,-0.2905832231,-0.0523138344,-0.1326320171,-0.3739037216,0.0139941853,0.3103607893,-0.1122552678,0.4573250115,-0.1265623271,-0.1473312974,0.3881329298,-0.0494086035,-0.1769427806,0.0829060003,0.1278465837,0.3624182045,0.1884074211,-0.4318867028,0.3127398193,0.3188152611,-0.2238079906,-0.3838649988,0.4087830186,-0.2351440936,0.0964606851,-0.0378321819,0.4400514662,0.1302730143,-0.307821244,0.0186477471,-0.2998788655]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/769","title":"How to choose proper download_mode in function load_dataset?","comments":"Can we just use `features=...` in `load_dataset` for this @lhoestq?","body":"Hi, I am a beginner to datasets and I try to use datasets to load my csv file.\r\nmy csv file looks like this\r\n\r\n``` \r\ntext,label\r\n\"Effective but too-tepid biopic\",3\r\n\"If you sometimes like to go to the movies to have fun , Wasabi is a good place to start .\",4\r\n\"Emerges as something rare , an issue movie that 's so honest and keenly observed that it does n't feel like one .\",5\r\n```\r\n\r\nFirst I try to use this command to load my csv file . \r\n\r\n``` python\r\ndataset=load_dataset('csv', data_files=['sst_test.csv'])\r\n```\r\n\r\nIt seems good, but when i try to overwrite the convert_options to convert 'label' columns from int64 to float32 like this.\r\n\r\n``` python\r\nimport pyarrow as pa\r\nfrom pyarrow import csv\r\nread_options = csv.ReadOptions(block_size=1024*1024)\r\nparse_options = csv.ParseOptions()\r\nconvert_options = csv.ConvertOptions(column_types={'text': pa.string(), 'label': pa.float32()})\r\ndataset = load_dataset('csv', data_files=['sst_test.csv'], read_options=read_options,\r\n parse_options=parse_options, convert_options=convert_options)\r\n```\r\n\r\nIt keeps the same:\r\n\r\n```shell\r\nDataset(features: {'text': Value(dtype='string', id=None), 'label': Value(dtype='int64', id=None)}, num_rows: 2210)\r\n```\r\n\r\nI think this issue is caused by the parameter \"download_mode\" Default to REUSE_DATASET_IF_EXISTS because after I delete the cache_dir, it seems right.\r\n\r\nIs it a bug? How to choose proper download_mode to avoid this issue?\r\n","comment_length":10,"text":"How to choose proper download_mode in function load_dataset?\nHi, I am a beginner to datasets and I try to use datasets to load my csv file.\r\nmy csv file looks like this\r\n\r\n``` \r\ntext,label\r\n\"Effective but too-tepid biopic\",3\r\n\"If you sometimes like to go to the movies to have fun , Wasabi is a good place to start .\",4\r\n\"Emerges as something rare , an issue movie that 's so honest and keenly observed that it does n't feel like one .\",5\r\n```\r\n\r\nFirst I try to use this command to load my csv file . \r\n\r\n``` python\r\ndataset=load_dataset('csv', data_files=['sst_test.csv'])\r\n```\r\n\r\nIt seems good, but when i try to overwrite the convert_options to convert 'label' columns from int64 to float32 like this.\r\n\r\n``` python\r\nimport pyarrow as pa\r\nfrom pyarrow import csv\r\nread_options = csv.ReadOptions(block_size=1024*1024)\r\nparse_options = csv.ParseOptions()\r\nconvert_options = csv.ConvertOptions(column_types={'text': pa.string(), 'label': pa.float32()})\r\ndataset = load_dataset('csv', data_files=['sst_test.csv'], read_options=read_options,\r\n parse_options=parse_options, convert_options=convert_options)\r\n```\r\n\r\nIt keeps the same:\r\n\r\n```shell\r\nDataset(features: {'text': Value(dtype='string', id=None), 'label': Value(dtype='int64', id=None)}, num_rows: 2210)\r\n```\r\n\r\nI think this issue is caused by the parameter \"download_mode\" Default to REUSE_DATASET_IF_EXISTS because after I delete the cache_dir, it seems right.\r\n\r\nIs it a bug? How to choose proper download_mode to avoid this issue?\r\n\nCan we just use `features=...` in `load_dataset` for this @lhoestq?","embeddings":[-0.2981237769,-0.1768483669,-0.0481677987,0.1568156332,0.3728834391,-0.0615380704,0.4276604056,0.089145571,0.2138329744,-0.0597590953,-0.2047565281,0.1524998248,0.2695509493,0.1945245415,-0.02517014,-0.032211259,0.1418517232,0.0792977214,-0.3833942115,0.0501097962,-0.4520176351,-0.0457154848,-0.2436609119,-0.2203825563,0.156972751,0.0908670202,0.1405728161,0.1358856559,-0.1495460123,-0.3516951501,0.3829599619,0.1064492911,0.4512521029,-0.0100312587,-0.0001241387,0.0170138814,0.2087269872,-0.3701337576,-0.181666851,-0.3073402345,-0.2030085474,-0.2878098786,0.2948141694,-0.3818411827,0.0118732508,-0.1798610687,-0.0173795875,-0.0702103376,0.141678527,0.3876386583,0.1243161857,0.15115273,-0.1043918729,0.1965364069,0.2249485552,0.0705995783,-0.0739335492,0.4122777581,0.3178265691,0.2856720984,0.1484193355,0.0274688657,-0.2809508443,0.0964084342,0.4331076145,0.0875602961,0.5315608978,-0.2286047786,0.0053384779,-0.0347638838,0.7926638722,-0.1557966471,-0.135601908,-0.2464806288,0.0157369822,-0.3848022819,0.1987775415,0.0420798957,0.0117137535,0.2732231915,-0.3474273682,-0.0860700905,-0.0133885471,0.3546049595,-0.0003889874,-0.0385024771,-0.2043245882,0.3374574482,0.2525494993,0.039589256,0.1619923115,-0.449942261,0.1867036223,0.0429582223,-0.4047822356,0.1304172277,-0.1374565959,0.2664746642,0.236819014,0.2299542576,-0.1548228264,-0.0926622823,-0.0792515129,0.0665773004,0.3199547231,0.12878263,0.1043737084,-0.3092871904,-0.0257253144,0.0703999102,0.1916398853,0.049117025,-0.0227866825,-0.2348311245,0.540137887,0.0002596265,0.2006766945,-0.1229735687,-0.2641528547,-0.1420144588,-0.0945665836,-0.0383802354,-0.049783878,0.3799830675,-0.0062793135,0.1521746814,-0.1268564463,0.4177851677,-0.2722910941,-0.1340729892,-0.0317891128,-0.1556083411,-0.1270587593,-0.1750640571,0.3053696752,-0.2639714181,0.2644144893,0.3575712144,-0.1747554243,-0.1617553681,-0.0174312778,-0.0884962156,0.0090374248,0.4237791896,0.1181288064,0.2123081684,0.2219851166,-0.1075821444,-0.0576024316,0.5666725636,-0.3441151381,-0.4022223651,-0.0058282944,0.0820841715,-0.3421829641,0.0225692149,-0.3763908446,0.143909961,0.0255218279,-0.2425636947,0.0790130049,-0.3571929038,-0.0704407468,-0.3164943457,-0.0200986043,0.6734722853,-0.4964317977,-0.0264704004,-0.2076707631,-0.2569437921,0.3642568886,0.0711122379,-0.4089884162,-0.0016056004,-0.2066299915,-0.1053454876,0.4453241825,-0.2087462097,-0.5429396033,0.2525599301,0.0336443596,0.0838887766,0.3102212846,0.2460404485,0.2766160071,-0.1398203969,-0.2555863559,0.2856087983,0.0932242945,0.141048789,-0.1062248796,-0.1885549873,0.3347449899,0.032627929,-0.2766446173,0.2249306291,-0.0438881926,-0.2530863881,0.1702097654,0.0345000103,-0.1126075611,0.0266716983,0.1021530628,0.1962307245,-0.1578259021,0.0795820355,-0.740213573,-0.0470241606,0.275233537,-0.2690194845,-0.272133857,-0.1280592382,-0.4426000416,-0.1316079199,-0.2991558015,0.0841811076,-0.0243577063,0.1747184098,0.181345135,-0.0368158929,-0.0784598887,0.2794538736,-0.1729464978,0.2299374342,-0.0222674664,0.2342641801,0.1129128709,-0.0626010299,0.1545609683,-0.2694577873,0.3410236537,-0.2135236263,-0.1886177212,0.496961087,0.2019570023,0.2254667133,-0.5017622113,0.0642606765,0.1720906049,0.1831092685,-0.0046683187,-0.0824803784,0.2830440402,0.0475750342,-0.3002482951,0.2096813023,-0.1713535041,0.0963776931,-0.1203131527,-0.1745775044,0.3404302001,0.0862245262,-0.1101804972,-0.1306815892,-0.0998930112,0.0866135657,0.2328527421,0.2049894631,-0.2370623946,0.0971372724,0.3599206805,-0.2388669401,0.1664720625,0.1412457675,0.0024575717,-0.2067211121,0.3860456347,-0.0181118548,0.5374552608,0.0922593772,-0.0647424534,-0.0982477292,-0.0890899375,-0.2349399775,0.4215328991,-0.1088552773,0.3762981892,0.0928524658,0.0468557216,-0.3739598095,-0.3573624194,0.1660375148,0.3270439506,0.2457007617,-0.4814566076,0.0129325101,-0.3572595417,-0.222611472,-0.1084775552,0.0790281221,0.1679350287,-0.2359506488,-0.2646328807,0.3349597752,0.081688486,0.2029771507,-0.2452377975,-0.1585022211,0.1610485613,-0.7079679966,0.0811407417,0.0265648998,-0.3262808621,-0.0736054853,0.2181967646,-0.116008237,0.1361428052,-0.2623019218,-0.0183591954,-0.3739025295,0.0654756129,0.003710686,0.0381481759,0.2727887332,-0.0044191666,0.4206526279,-0.0180208627,0.161377579,0.2462988794,0.0313079096,0.1154044271,0.2124305069,-0.0913026631,0.1203143373,-0.1219239086,-0.665851295,-0.3462766707,-0.1764652282,-0.1428951472,0.0490285158,-0.0059698918,0.0681011155,0.3256371319,-0.2047025114,0.0919452161,0.1029180884,-0.3671438694,-0.101931192,0.5462025404,0.1081275269,-0.1844171584,-0.0829290152,0.0043348945,-0.0424770564,0.2701987624,-0.2931649685,-0.2834721804,-0.1338075548,0.4665611088,-0.071446456,0.2954917848,0.2789805233,-0.0357418433,0.1449510604,0.0093251504,-0.1726491302,0.0674054772,0.1677254587,0.120783031,0.2852635384,0.4253148139,-0.1940834224,0.4808486998,0.1087180749,-0.0481882207,0.3894237876,-0.1903661937,0.2911720276,0.1304110289,-0.1176428646,-0.0516301692,-0.1776134968,-0.0352940634,0.260171771,-0.152953431,-0.2147369385,-0.1080888435,0.1152466834,-0.5872038007,-0.3035957813,0.1666941941,0.309710443,0.186901778,0.0740365237,0.2190928161,-0.1001063734,-0.1389508694,-0.0384620242,0.5875084996,0.1796838492,0.1998489648,-0.1805736274,0.0758691654,-0.1356766522,0.3035323918,0.0808761045,0.5172296762,0.2157154083,-0.0056346687,0.1064935401,-0.0685854182,0.8779754639,-0.0827821121,0.2284183949,0.0228288081,0.252314508,-0.2822555006,-0.1667332947,-0.1551891118,0.1341359168,-0.020944519,0.109267652,-0.3843249083,-0.099893935,-0.0522498488,0.4547657967,-0.1064655781,-0.4445143938,-0.2785519063,-0.4291524589,-0.4333376884,-0.0731729567,0.0012979989,0.0780208558,-0.2645585537,0.0401663482,0.1101416647,0.0512307361,0.0196343511,-0.0447431393,0.1604816765,0.1033977419,0.0223812908,-0.0519924536,-0.3631554544,0.1739183962,0.5789352655,-0.1612998396,-0.3909615874,0.1187860072,0.0163608622,0.2010006905,0.1088360995,-0.2684977055,-0.024138039,-0.0524266884,-0.0666960999,-0.1502482444,-0.0128512643,0.266967833,-0.0155068263,-0.067289643,-0.8737095594,0.5976746678,-0.0449595489,-0.12616238,-0.0705083907,0.1871443689,-0.14240852,0.006019711,0.0193552561,0.7355379462,-0.1768292785,-0.0156320035,0.3854378462,0.0207958613,0.6126236916,-0.2803078294,0.1384081244,-0.2318548858,0.1305318773,-0.0500006527,-0.2114840448,0.3716138601,0.1093444899,-0.1584900618,0.4707829058,-0.2379138172,0.3962750435,0.0369724222,0.1265867501,-0.3134088814,-0.2045218498,-0.2969937921,0.0438266508,-0.1785491705,0.120608218,0.0427952409,0.09447781,0.0201106481,0.1392270774,-0.0122677526,-0.1219582185,-0.2379900813,-0.0195518602,-0.3047446012,-0.3051156104,0.173195377,0.5150883198,-0.0864030495,-0.0246979669,-0.096759446,-0.009255073,0.0315299705,0.1695071161,-0.1301307082,-0.0340194628,0.2499612421,0.0489736013,-0.2247208357,0.1546764672,0.0042671557,0.1110539064,-0.0113720903,-0.1688061804,0.1451446712,-0.5207368731,-0.5995803475,-0.0852722451,0.1448283195,0.0155990655,-0.0054140389,-0.0975520611,0.1484616548,0.2198123932,-0.2083281875,-0.348364234,-0.3166775703,0.3343304098,0.0832558647,0.2338026762,0.1310770512,0.0895513371,-0.1941696703,-0.0318968855,0.0975025073,-0.3501374424,-0.46191594,0.0273964051,0.0944738016,0.2779285312,-0.2709822357,0.179777056,0.1189967468,-0.1085133106,0.0530548356,-0.3003046811,-0.1264685094,0.0483862348,0.2746642232,0.0077370489,0.2899661362,-0.1609075963,0.0919786543,-0.4921782613,-0.1814489216,-0.0524534509,0.183160603,0.1061625034,0.1717377901,0.2398357093,-0.0158353187,-0.1442894936,0.0165037028,-0.0397973545,-0.2177537829,-0.1113153175,-0.1632699072,0.2119177133,-0.0050746379,-0.2085061371,-0.0870049819,-0.4270334542,-0.0240687151,-0.1532411873,0.2351087332,0.1838352829,-0.0224882364,0.1400490254,-0.3013098836,-0.1095990092,-0.4557982981,0.095651269,-0.2496722043,0.2622475624,0.1927668303,0.2125907987,0.0839632675,0.0352892987,-0.4501771033,0.0212385058,0.0787536353,0.151515767,0.1444978863,-0.2035013735,0.3315588534,0.0542064495,0.2407667637,0.4869584739,-0.1531594098,-0.1538933218,0.1533464938,0.0671751872,-0.2148039937,-0.056878794,-0.1200529337,0.0571374819,0.0483214334,0.4295982122,0.1810687333,0.1054699048,-0.015584168,-0.1012237296,0.5195700526,0.2290925682,0.0258818995,0.6048824787,0.0278030653,0.1371875852,0.3535607457,0.2151578963,0.026156297,0.2934732735,-0.0717798173,0.0508133918,0.2015752643,0.1975307912,-0.2920260429,-0.5941326618,0.2725558579,0.3836588562,-0.6025776863,0.1743893176,0.295768708,0.3177723587,-0.221743241,-0.0346707515,-0.2121822536,0.1675433517,-0.1605745405,-0.2510948479,-0.0632870495,-0.4354324937,-0.1161422506,0.3221443594,-0.3627356291,-0.2999818027,0.2042711377,0.1932821572,-0.0679655522,-0.5329004526,0.0914443433,-0.014601808,0.2269548774,-0.0683528483,0.3582848012,0.2929713726,-0.006374876,0.150937438,-0.0296657775,0.3619471192,0.1994712651,0.1471490264,-0.2643459439,0.2242703289,0.0610304512,-0.0474915542,0.0560037456,0.1521061212,0.4220897257,0.2838714421,0.1128289923,-0.0554460883,0.2023054361,0.1842376292,-0.35645473,-0.3805241883,0.0537989661,-0.0333851166,0.0335576572,0.0858370289,-0.0225181151,-0.1427862942,-0.0669066757,0.3543439507,0.0768149793,0.2226689607,0.0618858449,0.0231942087,0.0333583318,0.1920264512,-0.0697750822,-0.0040392629,-0.3594412506,0.2542813122,-0.7662941813,-0.1680250168,-0.078842029,-0.3173618317,0.1970583946,-0.0482757837,0.0766493678,0.4915007055,0.0903009474,-0.0531331562,0.4645199478,0.2546527982,0.1510413587,-0.3897033632,-0.0744935349,0.075032562,-0.2163581252,-0.5871942043,0.3272801042,-0.2979813814,-0.0050256848,-0.3962117434,0.0559455939,0.3025632501,-0.0665262938,0.5888697505,0.0941470861,0.4858923256,-0.004767953,0.1134746,0.2599978447,0.2403260618,-0.0254954025,-0.0187586453,-0.3002490103,0.2615805864,-0.2164957523,0.2196522951,-0.2548787892,0.2183210999,0.2235890925,-0.0175197665,-0.237051487,-0.0644004866,-0.1191227362,-0.3833273351,0.01581694,0.3213788271,-0.1315158904,0.3969665468,-0.1332647353,-0.116931662,0.3631123304,-0.0391675383,-0.1580031216,0.072420314,0.092719391,0.4168378413,0.1926236898,-0.4525889158,0.2807515562,0.3402057588,-0.1753683835,-0.3141161799,0.3955916464,-0.1597575992,0.0925184116,-0.0176204611,0.4644208252,0.1077457964,-0.2916650176,-0.0107596768,-0.3381458521]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/769","title":"How to choose proper download_mode in function load_dataset?","comments":"Indeed you should use `features` in this case. \r\n```python\r\nfeatures = Features({'text': Value('string'), 'label': Value('float32')})\r\ndataset = load_dataset('csv', data_files=['sst_test.csv'], features=features)\r\n```\r\nNote that because of an issue with the caching when you change the features (see #750 ) you still need to specify the `FORCE_REDOWNLOAD ` flag. I'm working on a fix for this one","body":"Hi, I am a beginner to datasets and I try to use datasets to load my csv file.\r\nmy csv file looks like this\r\n\r\n``` \r\ntext,label\r\n\"Effective but too-tepid biopic\",3\r\n\"If you sometimes like to go to the movies to have fun , Wasabi is a good place to start .\",4\r\n\"Emerges as something rare , an issue movie that 's so honest and keenly observed that it does n't feel like one .\",5\r\n```\r\n\r\nFirst I try to use this command to load my csv file . \r\n\r\n``` python\r\ndataset=load_dataset('csv', data_files=['sst_test.csv'])\r\n```\r\n\r\nIt seems good, but when i try to overwrite the convert_options to convert 'label' columns from int64 to float32 like this.\r\n\r\n``` python\r\nimport pyarrow as pa\r\nfrom pyarrow import csv\r\nread_options = csv.ReadOptions(block_size=1024*1024)\r\nparse_options = csv.ParseOptions()\r\nconvert_options = csv.ConvertOptions(column_types={'text': pa.string(), 'label': pa.float32()})\r\ndataset = load_dataset('csv', data_files=['sst_test.csv'], read_options=read_options,\r\n parse_options=parse_options, convert_options=convert_options)\r\n```\r\n\r\nIt keeps the same:\r\n\r\n```shell\r\nDataset(features: {'text': Value(dtype='string', id=None), 'label': Value(dtype='int64', id=None)}, num_rows: 2210)\r\n```\r\n\r\nI think this issue is caused by the parameter \"download_mode\" Default to REUSE_DATASET_IF_EXISTS because after I delete the cache_dir, it seems right.\r\n\r\nIs it a bug? How to choose proper download_mode to avoid this issue?\r\n","comment_length":55,"text":"How to choose proper download_mode in function load_dataset?\nHi, I am a beginner to datasets and I try to use datasets to load my csv file.\r\nmy csv file looks like this\r\n\r\n``` \r\ntext,label\r\n\"Effective but too-tepid biopic\",3\r\n\"If you sometimes like to go to the movies to have fun , Wasabi is a good place to start .\",4\r\n\"Emerges as something rare , an issue movie that 's so honest and keenly observed that it does n't feel like one .\",5\r\n```\r\n\r\nFirst I try to use this command to load my csv file . \r\n\r\n``` python\r\ndataset=load_dataset('csv', data_files=['sst_test.csv'])\r\n```\r\n\r\nIt seems good, but when i try to overwrite the convert_options to convert 'label' columns from int64 to float32 like this.\r\n\r\n``` python\r\nimport pyarrow as pa\r\nfrom pyarrow import csv\r\nread_options = csv.ReadOptions(block_size=1024*1024)\r\nparse_options = csv.ParseOptions()\r\nconvert_options = csv.ConvertOptions(column_types={'text': pa.string(), 'label': pa.float32()})\r\ndataset = load_dataset('csv', data_files=['sst_test.csv'], read_options=read_options,\r\n parse_options=parse_options, convert_options=convert_options)\r\n```\r\n\r\nIt keeps the same:\r\n\r\n```shell\r\nDataset(features: {'text': Value(dtype='string', id=None), 'label': Value(dtype='int64', id=None)}, num_rows: 2210)\r\n```\r\n\r\nI think this issue is caused by the parameter \"download_mode\" Default to REUSE_DATASET_IF_EXISTS because after I delete the cache_dir, it seems right.\r\n\r\nIs it a bug? How to choose proper download_mode to avoid this issue?\r\n\nIndeed you should use `features` in this case. \r\n```python\r\nfeatures = Features({'text': Value('string'), 'label': Value('float32')})\r\ndataset = load_dataset('csv', data_files=['sst_test.csv'], features=features)\r\n```\r\nNote that because of an issue with the caching when you change the features (see #750 ) you still need to specify the `FORCE_REDOWNLOAD ` flag. I'm working on a fix for this one","embeddings":[-0.2571399212,-0.2127034068,-0.0402400121,0.1630151272,0.404009968,-0.0686804056,0.4393961132,0.0929248035,0.2167056799,-0.0813829452,-0.2033498436,0.1510446519,0.2890435755,0.2188754976,-0.0475560352,-0.0434880331,0.1723636389,0.066375263,-0.3534504175,0.0588456541,-0.4380922616,-0.0215164162,-0.2259062082,-0.2040410191,0.1186358109,0.1169724539,0.1479007751,0.1589901596,-0.1337802112,-0.378020972,0.3889071941,0.0531567298,0.4369566739,0.025336422,-0.0001253529,0.0090917898,0.2151527107,-0.3684231043,-0.169039771,-0.2970288694,-0.1894733012,-0.2959619164,0.2958136201,-0.3904573321,0.0227865353,-0.1693538576,-0.0471357293,-0.0768366829,0.1460930109,0.3617572784,0.1143755913,0.1577693969,-0.0766454935,0.1977392882,0.2081690729,0.0930887982,-0.0868413672,0.3931901455,0.3226602077,0.2978898883,0.1851437241,0.04641008,-0.3264553249,0.0901592225,0.4344960451,0.0826564878,0.5438580513,-0.2351289988,0.0120473299,-0.0455434769,0.7830659151,-0.1429341286,-0.1257714629,-0.2601625621,-0.008001701,-0.3558511138,0.182985574,0.0512514822,0.0279595926,0.2612246573,-0.3293582797,-0.0617427714,-0.0040019108,0.3507049084,-0.0163131692,-0.0037298871,-0.1991258264,0.3501984477,0.2780058682,0.0251901429,0.1337103993,-0.4628354609,0.1859646291,0.0551593415,-0.3747671843,0.1197132617,-0.1101632714,0.2270287573,0.2134388089,0.2350176126,-0.1731294245,-0.105664365,-0.0790382028,0.0543873981,0.3305209875,0.1084478721,0.0923051238,-0.3073598742,-0.0159139428,0.0750562027,0.1939509362,0.0455879122,-0.0070974957,-0.2097703516,0.5614816546,-0.0100340722,0.2120887637,-0.1557591856,-0.2577976584,-0.1193052456,-0.1015517712,-0.0093439156,-0.0417833365,0.3487726152,-0.0212357547,0.167053774,-0.1175338626,0.4232870042,-0.2588958442,-0.1044211611,-0.0268054381,-0.1692228466,-0.1096673235,-0.1769726425,0.3057777286,-0.2919177115,0.2943980396,0.3564922214,-0.1896508485,-0.1925509274,-0.038214311,-0.0922378898,-0.009829415,0.428913027,0.1095343456,0.2218271047,0.2428340316,-0.1180534065,-0.0441173539,0.5599557757,-0.341029495,-0.4236623645,0.0073513207,0.0734670013,-0.356769681,0.01916481,-0.3555149436,0.1227714568,0.0327582397,-0.2373168468,0.0837816894,-0.3370423913,-0.0787452608,-0.3249077499,-0.0326400511,0.6747348309,-0.4708639979,-0.0143058971,-0.1925148964,-0.2515560389,0.3785536289,0.0774227604,-0.403101176,-0.0403574593,-0.1870816648,-0.0902742967,0.4304097593,-0.2198917121,-0.5396748185,0.2700550258,0.0350185335,0.0890857279,0.3254155517,0.257817328,0.2394130379,-0.155254975,-0.2509272695,0.3023552001,0.0893182904,0.1630808264,-0.1126749739,-0.1849411279,0.3365391195,0.0209904481,-0.2687707841,0.2023787498,-0.0638904497,-0.2422065884,0.1783636659,0.0243125968,-0.1316060573,0.036118824,0.1187472865,0.2286567688,-0.1529634446,0.0663990378,-0.6884865165,-0.0347267538,0.2795573771,-0.2823960781,-0.2883073688,-0.1402547956,-0.4521226883,-0.1210737079,-0.3010733724,0.081807673,-0.0375691205,0.1825788021,0.1832840294,-0.0370614603,-0.0729631856,0.2751716375,-0.1564928144,0.2462969124,-0.0133046694,0.2497580945,0.1108405069,-0.0725681111,0.147635147,-0.2571159005,0.3647033572,-0.2116820663,-0.1865238249,0.4816067517,0.2214027196,0.2470842153,-0.5144000053,0.0678613186,0.1991439015,0.1813839227,0.001518858,-0.0353651829,0.289742738,0.038843859,-0.2861204147,0.2031154782,-0.1808379889,0.102239877,-0.1256675273,-0.1688039005,0.3344340026,0.0932300016,-0.1076360494,-0.1208641008,-0.10855297,0.0665090233,0.1763236076,0.2263212502,-0.2570774555,0.1120337769,0.3664935827,-0.2465100586,0.1577565074,0.1405531615,-0.0184936877,-0.2118013203,0.366008848,-0.0334386677,0.5430467129,0.0791769922,-0.0656061396,-0.0924277008,-0.094740957,-0.2232439965,0.4322989285,-0.0901784897,0.3930632174,0.0928277522,0.0567188524,-0.3744744956,-0.3599195778,0.1701303273,0.315366745,0.2394780368,-0.4751332402,0.0356779806,-0.3608303368,-0.2496362776,-0.1142448485,0.0489516221,0.1525537372,-0.2344695181,-0.2881452143,0.377052635,0.0880210549,0.1681328416,-0.2675405145,-0.1980847567,0.1541350633,-0.7329838276,0.0845142081,0.0342933051,-0.3322189748,-0.0932079628,0.2178363651,-0.1219761744,0.1280557364,-0.2368908972,-0.0129104853,-0.379891634,0.0522337519,-0.0185541809,0.0284672398,0.2587381899,-0.006434219,0.4116538465,0.0028282807,0.1528276503,0.2606508136,0.0493853092,0.1308382899,0.2404662669,-0.0796580166,0.0912177414,-0.1080132946,-0.6392263174,-0.3208782375,-0.1806237996,-0.147621125,0.065221265,-0.0250942353,0.0646182224,0.3264109194,-0.1983589828,0.1241691262,0.1098171249,-0.3587359786,-0.0688697472,0.5512461662,0.133730486,-0.1682958454,-0.062290512,0.0134632699,-0.0581243597,0.261585027,-0.2785099149,-0.2853080928,-0.1439531446,0.4526166618,-0.0660182461,0.2976057231,0.284819752,-0.0180761702,0.1427216679,0.0206978545,-0.1765645593,0.0578630008,0.1887421608,0.1285846233,0.2652296126,0.4068438411,-0.1923152953,0.4521817267,0.1233739182,-0.0381090008,0.3927264214,-0.1760504395,0.2472157478,0.1235024631,-0.1327904016,-0.0638440102,-0.2023869902,-0.0143619012,0.2485182136,-0.1247897893,-0.2157243043,-0.1185932904,0.1108167171,-0.5693225861,-0.2923105955,0.1548262089,0.3429583907,0.2171761394,0.0642765537,0.2186961919,-0.1193911508,-0.125150457,-0.0268506892,0.6019415259,0.1677833349,0.1966409385,-0.1799769551,0.0977293476,-0.1179957613,0.2893590927,0.084662579,0.5040074587,0.2042645961,-0.0134956939,0.0957151055,-0.0743034035,0.862149179,-0.0778501555,0.2306222618,0.0169222094,0.2369264215,-0.3040497601,-0.1977231055,-0.1489269286,0.1329702735,-0.0191019103,0.1518365741,-0.3600399196,-0.1250207424,-0.043709226,0.4756930768,-0.1013354212,-0.4211811125,-0.2916677892,-0.4378746748,-0.4540203214,-0.055892691,0.0120012136,0.0790909976,-0.2337288111,0.0254931767,0.1000574976,0.0465209,-0.0006262445,-0.0486608259,0.159667328,0.0804436952,0.030793678,-0.0606101789,-0.389847219,0.136727497,0.5804674029,-0.123164393,-0.4300067425,0.0780539885,0.0023522174,0.1949391663,0.1240272075,-0.2761417627,-0.0230704416,-0.0561558791,-0.0612984449,-0.1533517838,-0.0169717446,0.313621223,-0.006554279,-0.0639152378,-0.8783692122,0.5908769965,-0.0470537692,-0.122818239,-0.0681608841,0.1936955154,-0.137100473,0.0068526408,0.0088330163,0.7403290868,-0.1477386951,-0.0234253816,0.4106375575,0.0165233277,0.6323752999,-0.2574595213,0.1452035308,-0.2199908346,0.1368507892,-0.0592441671,-0.2123483568,0.3720144033,0.0788652152,-0.1590979695,0.4722997844,-0.2580112517,0.3843183815,0.0349201448,0.1149370819,-0.2998351753,-0.2017106712,-0.2972393334,0.0437286496,-0.1808031052,0.1441550404,0.048972033,0.0881351531,0.0088872248,0.1423192024,-0.0402462706,-0.1051114649,-0.2145475447,0.00285812,-0.2828826606,-0.2913124859,0.1696634591,0.4940200746,-0.0729221702,-0.0279583,-0.0893944204,-0.0332272165,0.0269724168,0.1663254052,-0.1406584382,-0.0422090255,0.257943064,0.0400789492,-0.2443522662,0.1300709993,0.0104620242,0.108861044,0.0024220357,-0.1839755476,0.1504509747,-0.5237014294,-0.6343913674,-0.0505853854,0.130431667,0.0128201405,-0.0131848175,-0.1037010401,0.1341365576,0.207993418,-0.2031947523,-0.3364051878,-0.3015027046,0.3213744164,0.0739852712,0.2300237566,0.1035767943,0.092312403,-0.2006052583,-0.0208337642,0.0716893151,-0.3993091583,-0.4469237328,0.0379945599,0.0748232529,0.2494314313,-0.2763229012,0.1764595807,0.1284055561,-0.1219971254,0.04915094,-0.3013528585,-0.1138836667,0.0468746871,0.2838817835,-0.0032972251,0.2612095177,-0.1571601778,0.0932509676,-0.5119244456,-0.1655477732,-0.0171333645,0.1885731071,0.1145518124,0.1965781152,0.199713394,-0.0309405141,-0.1614879519,0.0120449467,-0.0239459686,-0.2194655985,-0.1023138463,-0.1890747994,0.2198649496,0.0047949557,-0.2370148748,-0.0781530216,-0.4396764338,-0.0063916463,-0.1508437246,0.2134187073,0.2114851326,-0.0225066822,0.1191334426,-0.344498992,-0.130096823,-0.4444688857,0.0951450467,-0.2700436711,0.2525141835,0.219247371,0.2146097422,0.1194797009,0.0346096605,-0.4590881467,0.0269877911,0.0676205605,0.1436444521,0.1754936874,-0.214136079,0.3252046704,0.070473671,0.2229964137,0.4845416546,-0.1388935745,-0.1331438571,0.1473707706,0.0626072586,-0.2442159504,-0.11622563,-0.0963729024,0.0584983341,0.0248968285,0.4365079701,0.2103257477,0.138575837,-0.0232630037,-0.0939512476,0.530348897,0.2382722348,0.053014461,0.5988234878,0.0372681841,0.1260793507,0.3304837942,0.2008469701,0.024920905,0.2706526816,-0.0549688675,0.0294166021,0.2067639977,0.1966654211,-0.2933321893,-0.568454802,0.2525375485,0.3639097512,-0.602901876,0.1919147968,0.3080990314,0.3266709745,-0.2500045598,-0.0425350741,-0.211919874,0.1807708144,-0.1514454484,-0.2573174536,-0.0307767652,-0.4313881993,-0.1392447352,0.3422462344,-0.3485726118,-0.2752635479,0.2215133905,0.1923908889,-0.0687969998,-0.523971498,0.0820436701,-0.0109852655,0.2564552724,-0.0744184852,0.3178033531,0.310449481,-0.0251560472,0.1266564727,-0.0302233957,0.3832311332,0.173589468,0.1232963055,-0.2349727601,0.204733178,0.0548499897,-0.0312468074,0.0443176106,0.1425590813,0.4424269199,0.2582496405,0.0986924469,-0.0493013449,0.190691486,0.1652114838,-0.3828429282,-0.3693989515,0.0812259763,-0.0412446968,0.0211110078,0.0820622295,-0.0193278622,-0.1319551468,-0.060937658,0.3863160908,0.103305243,0.2183603346,0.0937743261,0.0164575074,0.0454543456,0.177882731,-0.0822125822,-0.0131134018,-0.3509376347,0.2613555193,-0.7569337487,-0.1792022437,-0.0720736012,-0.3162717819,0.2269176245,-0.0680761859,0.0832482353,0.4920197725,0.1015997827,-0.0779432431,0.4903147221,0.2357798666,0.16114375,-0.3786865473,-0.0462721996,0.055692181,-0.2231165171,-0.5922364593,0.3191397488,-0.30664289,-0.0053920173,-0.3925586641,0.0455300286,0.2697087526,-0.0871839672,0.5793070793,0.0889639482,0.4939719737,0.0061497628,0.0801969841,0.2767911255,0.2586174011,-0.0210693479,-0.0076073674,-0.3024641871,0.293603301,-0.2321776152,0.2348771244,-0.2624464333,0.2284349054,0.2493194491,-0.0449304767,-0.2544177175,-0.0677946731,-0.1488250345,-0.3843573034,0.0296917353,0.3405500948,-0.1550570726,0.4002464712,-0.1127438769,-0.109699972,0.3859417737,-0.0091191726,-0.176210776,0.0742671043,0.0939318016,0.444763571,0.2256976217,-0.4652258158,0.2522892654,0.3160280287,-0.1756307483,-0.3445521593,0.3969544768,-0.1829556674,0.0570503958,0.0019052976,0.4485104978,0.0847407579,-0.3051199615,0.011831996,-0.330213517]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/768","title":"Add a `lazy_map` method to `Dataset` and `DatasetDict`","comments":"This is cool! I think some aspects to think about and decide in terms of API are:\r\n- do we allow several methods (chained i guess)\r\n- how do we inspect the currently set method(s)\r\n- how do we control\/reset them","body":"The library is great, but it would be even more awesome with a `lazy_map` method implemented on `Dataset` and `DatasetDict`. This would apply a function on a give item but when the item is requested. Two use cases:\r\n\r\n1. load image on the fly\r\n2. apply a random function and get different outputs at each epoch (like data augmentation or randomly masking a part of a sentence for BERT-like objectives).","comment_length":41,"text":"Add a `lazy_map` method to `Dataset` and `DatasetDict`\nThe library is great, but it would be even more awesome with a `lazy_map` method implemented on `Dataset` and `DatasetDict`. This would apply a function on a give item but when the item is requested. Two use cases:\r\n\r\n1. load image on the fly\r\n2. apply a random function and get different outputs at each epoch (like data augmentation or randomly masking a part of a sentence for BERT-like objectives).\nThis is cool! I think some aspects to think about and decide in terms of API are:\r\n- do we allow several methods (chained i guess)\r\n- how do we inspect the currently set method(s)\r\n- how do we control\/reset them","embeddings":[-0.0316452757,0.013758203,-0.2658601999,0.0116509115,-0.1261357516,-0.1206701994,0.0822107866,0.315966934,0.3856461048,0.1577916294,0.090818353,0.4605909884,-0.3150094151,-0.0637615249,-0.099035129,-0.047613658,-0.0590976812,0.1570822299,-0.1722947806,-0.1399807483,-0.1990464628,-0.1581519544,-0.0754315853,-0.1457812935,0.1748659164,-0.2872923017,-0.2392606586,-0.1879741102,-0.3999473751,-0.1362809092,-0.0379625224,0.6174533367,-0.0462141819,0.0646637157,-0.0001012224,-0.1686680168,0.231354326,0.0854558796,-0.0684590787,-0.0441985913,-0.253770709,-0.1252306551,0.2018482387,-0.5509762168,-0.0584174879,-0.2714449167,0.4295844734,-0.4214327335,0.5393214226,-0.0856241882,0.2302741557,-0.0530725755,-0.3106282651,0.0365993567,0.3807837367,0.1583934426,-0.0576467477,-0.0953391343,0.4525526166,0.1018311754,-0.124576956,0.2925328314,-0.1559335738,0.1263547093,0.4784334302,0.0078927781,-0.4412189722,-0.4262873232,-0.1540486515,0.0608588085,0.5324354172,-0.1861354113,-0.2275097817,-0.1035418361,-0.0565495193,-0.012565772,-0.2463449836,-0.0079961913,0.2070353627,0.2098948509,-0.2132597119,-0.1770933121,-0.2358946502,0.2579637766,0.2851657867,0.0231089815,0.0494544171,0.1578091681,-0.0619069152,-0.0088356752,0.1775438488,-0.1693768799,-0.0915952548,0.3831130266,0.0738331079,-0.2585502267,0.1556254625,0.0686063841,0.3009393215,0.3678823709,0.4818527997,0.2997098565,-0.5404179692,0.2487291992,0.3010678887,-0.1054353938,-0.139364332,0.049615223,0.3160534799,-0.0766110048,0.080386579,0.016533317,0.1527781785,0.1714900136,0.1464107037,-0.1503849924,0.0439560376,-0.0677957311,0.1476469636,-0.1441446543,0.0641759485,-0.1750659943,-0.1025168225,0.2155337483,-0.100213632,0.1461665928,0.1481036097,-0.072214365,0.3096322715,0.0257673897,-0.0828994662,-0.0079526603,0.049043566,0.1977220923,0.1436965615,0.1598265469,0.0956062973,0.1071432978,-0.2972553968,0.0631073713,0.7239733934,-0.2636158466,0.4300808609,-0.1151483357,-0.1545201987,-0.2465788722,-0.1759558916,0.2088591009,-0.4214297235,0.1569105983,-0.4271495938,-0.134318769,0.0897209793,0.2743909359,-0.0737845525,0.0397363529,-0.1317954808,0.6621151567,-0.0275940076,-0.1764548719,0.3585615456,0.1350755841,-0.2970649004,-0.2560357749,-0.1324668527,0.3818058968,0.0152168553,-0.1152827069,-0.077579841,0.0068317223,-0.2633752823,0.0767960772,-0.3004282713,0.2324619591,-0.103215605,0.0379180908,0.5653924942,-0.4251792133,-0.2009767592,0.1914840043,0.1012704521,0.0121787218,0.0566214807,0.4613977075,0.2702679932,-0.0736534819,0.115966782,0.280369699,-0.0191990491,-0.0063708974,0.0200869869,-0.58931458,0.0421126634,0.260127753,0.0152670154,0.0322850123,0.2342322171,-0.0375205651,0.1960207075,-0.1750436425,0.0870560184,-0.3264836967,0.265180409,-0.0300459117,-0.0183052067,-0.2506521642,-0.4685738087,0.0288641695,0.0278265513,-0.0176963694,0.0963466242,-0.285038799,0.38956514,-0.1420194656,-0.1039258838,0.0119635127,0.2112865746,-0.075606741,-0.076759398,-0.0496201254,-0.3771501184,-0.0185872242,0.0087939659,0.0086954506,-0.3124372661,0.0176002719,0.2360194474,0.0682337135,-0.1141873598,-0.0841505826,-0.2094907612,-0.0732255355,0.0805928856,0.1130607873,0.4026201665,0.1198883951,-0.008840817,0.3823271394,0.3317900002,-0.0653146133,0.3423094153,0.2478310615,0.148562029,-0.0943941399,-0.2361268103,0.1199825034,-0.2427388579,0.0815189257,-0.033425875,-0.1496855915,0.2441015393,-0.0081077423,-0.3279363215,-0.2521275282,-0.0721689016,0.2961583436,0.3833504915,-0.0039719576,-0.0957821086,0.0504410528,0.2785755098,-0.0491424352,0.0562419668,0.1528507024,0.3605255485,-0.0903854668,0.2327747792,0.0485902391,0.3236717582,0.254142195,-0.2183730006,-0.0732039511,0.2605610192,0.0174726043,0.0921619236,0.1772110015,0.1935932487,-0.2423489541,0.3541028202,0.1604021788,-0.1092864349,-0.6076213717,0.1243430227,-0.1385724097,0.2387004495,-0.0957700685,0.0831852555,-0.0339872204,0.0074487072,-0.149992317,0.065113835,0.0129220057,0.2524854541,-0.0130217271,-0.4675897956,0.0029508839,-0.2462662756,0.6607556939,-0.2335275114,-0.144695729,-0.2904372513,-0.1868735105,0.2194749415,0.1533401012,0.0763652697,0.0247766264,0.6802108884,-0.0467184447,-0.0210895725,-0.2449754477,-0.4454804361,0.1895516813,0.0219036676,0.0836613625,0.2553391457,-0.262978673,0.1285716295,-0.1577276736,-0.0089158826,-0.1806893498,0.0045980336,-0.4637442231,-0.0139714461,-0.0161142424,-0.3169600666,-0.0770591274,-0.3230470121,-0.5011575818,0.4366027117,0.1657423079,0.3440570831,-0.0191955362,0.0011946532,0.0038690502,0.2962954044,0.1945665777,-0.2044894397,-0.3378013968,0.1303236783,-0.3507660031,-0.2318703532,-0.2668490708,-0.1379507929,0.0826428235,0.3081731498,-0.4109921753,-0.5004564524,-0.1030673161,0.3629426956,0.1154215038,0.0040320368,0.1514191031,0.3908299208,-0.1510018408,-0.0657531843,-0.1858519912,0.3007418811,0.0845982805,-0.3038173616,0.4934718609,0.0059860731,-0.179616645,0.4737222195,-0.0612437017,-0.1227849498,0.6362372637,0.1666857898,0.1869206131,-0.1618528813,-0.0220694225,-0.1229344457,0.0153929172,-0.173243165,0.1689973474,-0.0143931909,-0.4332821667,-0.2108092457,-0.089545466,-0.3204256892,-0.1249476001,0.1773780584,-0.1927106529,-0.1760842651,-0.0012910326,0.0785413906,-0.3619856238,0.0555810183,0.3197860718,-0.0887707323,0.3706781566,-0.2170149386,-0.2395542264,0.0721357614,-0.2971974909,0.387866497,-0.0153105976,-0.1779416502,0.0673948377,-0.1837063432,-0.041526895,-0.212174952,0.3732042909,-0.1810091436,-0.287150383,-0.0875945315,-0.2525025904,0.0918015018,0.2313669473,-0.2248200625,-0.2691423297,-0.1314213574,0.5000820756,-0.3117642999,-0.030294776,0.0825473219,-0.0228257068,-0.1114180312,-0.2034701705,0.0398612991,-0.2650546432,-0.3558033705,-0.2392140329,-0.1007635966,-0.2114698738,0.117920436,-0.4718104005,-0.2868666649,0.2348798662,-0.0000960802,0.1809700429,0.089544341,-0.0455648787,0.1941203624,0.1056026965,-0.0448657386,0.0704802051,0.3506188095,-0.4467808306,-0.1341320574,0.0475538298,-0.0452931486,0.000335795,0.1973542273,0.2177892923,0.308333993,0.1552434415,0.1554035842,-0.4040930271,0.1636748314,0.2215841413,-0.0303742159,-0.2600880265,-0.8813298941,0.4342654347,0.193916887,-0.3777368963,0.5089743137,-0.2627505958,-0.3286872506,0.5772544742,0.408349812,0.8184028268,-0.0983371437,0.1716862172,0.0122325514,-0.183648631,0.7522255182,-0.0472678542,-0.0118853897,-0.220019266,-0.1830983758,-0.0496231429,-0.0183233414,0.0103892088,0.0384048335,-0.4284000397,0.0952544063,0.4016557634,0.0883122832,-0.2598718703,0.4192357659,-0.389760673,-0.3177863657,-0.3965379,0.2290991694,0.0577062704,-0.0568927899,-0.2013415992,-0.1468288451,-0.0216848813,-0.1450176537,0.1410877556,-0.0497241132,-0.0982011557,0.2889795303,-0.20261693,-0.0617842749,-0.1000945866,-0.0405115932,-0.1206941679,0.2051262558,-0.2008714825,0.0101791769,0.0540811308,0.1703216285,-0.0095123406,-0.2149825096,0.3916645944,-0.0903270021,-0.120635882,0.1506973952,0.0843684524,-0.3436558247,-0.3285520077,0.0956087112,0.2398023903,-0.0266485121,-0.3221815228,0.2734158337,-0.1842143387,0.0300228372,0.1777677983,0.1538802385,0.1435182989,0.341430366,0.0145319942,-0.0362866968,-0.11415869,0.0221975259,0.0997279957,-0.1149287,0.0972167104,-0.155438602,0.0116543081,-0.3061263263,-0.0144719193,0.3259465992,-0.227126196,0.1405778378,-0.0424365066,-0.2730095387,-0.1316498518,0.3220629096,0.0462200493,0.1609128416,-0.0436262563,-0.0265114643,-0.4319006503,0.426245749,0.2698981762,0.5220265985,0.0628468469,-0.3603258133,-0.072220929,0.2898747623,-0.4333048761,-0.043410562,-0.0606926084,0.113906078,-0.2334891707,0.1204824373,-0.2686123848,0.1077727377,0.1110161394,0.2047243267,-0.321447432,-0.274540782,-0.2825702429,0.1382749081,-0.0045825113,0.0800474361,-0.0675571933,-0.3578974009,-0.158934623,-0.2625755966,0.0855759829,-0.1021583304,-0.1191416085,-0.0682687014,0.2984782755,0.0095486408,-0.1344479769,0.3815545738,0.0871392116,0.1423792541,0.097519286,0.1498387307,-0.1974892169,-0.1771635711,0.3952691257,0.1929668188,0.2850932479,0.1449053586,0.1226111129,0.3286502063,0.1575086564,0.1026376933,0.350815922,0.3296618462,-0.1900773346,-0.0669352189,0.1558900923,0.326631695,-0.2356421053,-0.0083129788,0.0163212214,-0.0832165852,0.1398475021,0.1998024136,-0.1729813367,0.074086234,-0.1409592479,0.0248277839,0.3546094894,-0.2582054138,-0.0061140945,-0.1341189146,-0.1970200837,-0.0595916174,0.2059364319,0.257748127,0.2602255046,0.0806211457,-0.1177510768,0.4487949312,0.1916569173,0.1141955853,0.2455382198,0.0187024437,0.3160087466,0.4250283539,0.2525939643,0.5317397714,0.3590681255,-0.0213168189,0.1426806301,-0.1604798138,-0.0961520076,0.2260684669,0.1921571344,0.0371529125,-0.0375727117,-0.1719564348,0.0304731708,0.0237922799,0.0455731004,-0.2055731863,0.5976638794,0.2425042838,0.0016859535,-0.1537800282,0.0611675605,-0.050387647,0.3446070552,-0.2554200292,-0.0170028005,0.1320172548,-0.1450250149,0.2190375477,0.3608762324,-0.0550073162,0.3974847198,0.0111281173,0.3629963994,-0.0713224858,-0.082165204,-0.3574269414,0.2183413804,0.2614524066,-0.1214355156,0.2981262207,0.1965662986,-0.2094759345,0.0629977658,0.013949764,0.0031707948,-0.2451432496,0.2634298801,0.3895851076,-0.3707019687,0.1628740728,0.039695777,-0.448816359,0.0132013932,-0.0758393034,-0.3752388954,0.00783036,-0.0192508604,0.1329202801,-0.1248855814,0.2074125856,0.0773235783,0.4497225583,-0.136465326,-0.1437561214,-0.4140063822,0.2090253234,0.0366400145,-0.1769729257,-0.0558742099,-0.0564514138,0.1387464851,0.0441668071,-0.0253303535,-0.2873387933,0.0377660133,0.0703257993,-0.3291042745,0.0735041648,-0.01508711,0.1148948297,0.0628169477,-0.3707211018,0.0946092159,-0.266374737,0.0631345883,-0.1402848214,0.0729795843,0.1142408624,-0.1053731889,0.7482131124,0.2279096097,-0.0328709036,-0.2948721945,0.0191608835,0.0244486909,-0.044572562,-0.2085472792,0.0011635412,0.2717054486,0.0931544527,-0.0104420194,0.0511706881,-0.1470230073,0.1538031101,-0.0686118826,0.1370601505,-0.1405926347,0.0735397041,0.0568241924,-0.2435269207,-0.2090392411,0.372202903,0.1220259741,-0.0956752747,-0.0523088835,-0.032848686,0.3887146115,-0.5408634543,-0.0991328806,-0.1203035936,0.1036350951,-0.1390575171,-0.0125474529,-0.4450729191,-0.0697336048,0.2892309427,-0.1003464833,-0.1196117029,0.4591073096,0.2960068583,-0.1567553878,-0.2075937688,0.370503366,0.0094875423,-0.3057690263,0.0125032263,-0.3759892881]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/767","title":"Add option for named splits when using ds.train_test_split","comments":"Yes definitely we should give more flexibility to control the name of the splits outputted by `train_test_split`.\r\n\r\nRelated is the very interesting feedback from @bramvanroy on how we should improve this method: https:\/\/discuss.huggingface.co\/t\/how-to-split-main-dataset-into-train-dev-test-as-datasetdict\/1090\/5\r\n\r\nAnd in particular that it should advantageously be able to split in 3 splits as well instead of just 2 like we copied from sklearn.","body":"### Feature Request \ud83d\ude80 \r\n\r\nCan we add a way to name your splits when using the `.train_test_split` function?\r\n\r\nIn almost every use case I've come across, I have a `train` and a `test` split in my `DatasetDict`, and I want to create a `validation` split. Therefore, its kinda useless to get a `test` split back from `train_test_split`, as it'll just overwrite my real `test` split that I intended to keep.\r\n\r\n### Workaround\r\n\r\nthis is my hack for dealin with this, for now :slightly_smiling_face:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\u200b\r\n\u200b\r\nds = load_dataset('imdb')\r\nds['train'], ds['validation'] = ds['train'].train_test_split(.1).values()\r\n```\r\n","comment_length":58,"text":"Add option for named splits when using ds.train_test_split\n### Feature Request \ud83d\ude80 \r\n\r\nCan we add a way to name your splits when using the `.train_test_split` function?\r\n\r\nIn almost every use case I've come across, I have a `train` and a `test` split in my `DatasetDict`, and I want to create a `validation` split. Therefore, its kinda useless to get a `test` split back from `train_test_split`, as it'll just overwrite my real `test` split that I intended to keep.\r\n\r\n### Workaround\r\n\r\nthis is my hack for dealin with this, for now :slightly_smiling_face:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\u200b\r\n\u200b\r\nds = load_dataset('imdb')\r\nds['train'], ds['validation'] = ds['train'].train_test_split(.1).values()\r\n```\r\n\nYes definitely we should give more flexibility to control the name of the splits outputted by `train_test_split`.\r\n\r\nRelated is the very interesting feedback from @bramvanroy on how we should improve this method: https:\/\/discuss.huggingface.co\/t\/how-to-split-main-dataset-into-train-dev-test-as-datasetdict\/1090\/5\r\n\r\nAnd in particular that it should advantageously be able to split in 3 splits as well instead of just 2 like we copied from sklearn.","embeddings":[0.0725738257,-0.0852903724,-0.1180252507,-0.045865681,0.15466851,0.2035396695,0.6252444386,0.1981317103,0.1147508919,0.3940248191,0.0419959575,0.3692273498,-0.2435895652,0.2782408893,0.191498816,-0.4520206153,0.0777252316,-0.2225460112,0.2795640826,0.1441616565,-0.1701510102,0.0290893447,-0.10989476,0.2488474548,-0.4482490122,0.0521879084,-0.0278837271,-0.0407919213,0.0729061589,-0.4732066393,0.0694575384,0.5114654303,-0.0923280418,0.4597217441,-0.0001201279,-0.2272785604,0.3921885788,-0.1634059697,-0.3604239523,-0.2719935775,-0.2721001208,-0.0508755185,0.191066891,-0.3920917809,-0.0140885385,-0.0546336323,-0.0692299753,-0.2750483453,0.1217140108,0.2252164781,0.1275219619,-0.0359692238,-0.1207832396,-0.0023297553,0.0822661072,0.4533889294,-0.3355133832,0.1411752254,0.0616262667,-0.0589847378,0.2678494453,0.035397239,0.0078432718,-0.17598553,-0.0337530002,0.068655543,-0.1521918923,-0.2227631807,-0.0949150696,0.400662899,0.3223872483,-0.3291953206,-0.407809943,-0.4338805974,0.1580452025,-0.2891124785,-0.0043494841,-0.0610060431,0.00684661,0.2569349408,-0.1519077867,-0.0383322202,-0.1391979456,0.1504184306,0.167988494,0.2730539441,0.1757129133,0.1539538652,0.4608617723,0.0318741426,0.2671248317,-0.275901258,0.1429983675,0.0135914311,-0.1820885837,-0.2353997082,-0.3166776597,-0.4003146589,-0.0385823026,0.2578307092,-0.089904584,0.0029252549,-0.1185064763,-0.036200352,0.1213353276,0.2042189091,0.1922021061,0.7993481755,0.1940679997,0.1425664276,-0.0509872809,-0.0838256404,-0.0732704252,-0.2198163271,0.1319383532,0.11735183,-0.1480457783,0.0972706452,-0.2157282531,-0.1622211337,-0.3056786954,-0.160752967,0.2621247172,0.2394130081,0.0350664519,0.0869063213,-0.0795196444,0.0568442196,0.1468181014,-0.0548607074,-0.1057783291,0.0489816442,-0.252707839,0.1430670917,0.2551349103,-0.0202106182,0.0251753498,-0.1186656654,0.2427727133,-0.2523259223,-0.0712954998,-0.0101479227,0.3710731268,0.1764828712,-0.2565138936,-0.0895712972,-0.2148702592,-0.0413802266,-0.1269136667,0.0241174344,-0.183362633,-0.4484345317,0.0184313636,0.0609599091,-0.1824507713,0.0002761522,-0.0470030382,0.2932840586,0.2690389752,0.0702133179,0.1655514985,0.034612231,-0.1679075807,-0.0667347088,0.389613241,0.087421909,-0.3150944114,-0.1640779525,-0.337852329,0.0888542384,0.0349265188,0.003133609,-0.0557395108,0.2230695784,-0.0702297017,0.240098834,0.3350121081,-0.1714231819,0.0440316312,0.0458580256,0.0866497979,-0.0092960801,0.2888076603,-0.155959785,0.1066175625,-0.03297792,-0.1210659072,0.1750892252,-0.324523747,-0.197215274,0.0226151776,-0.0928212106,0.3992806375,0.1736496389,0.0248096939,-0.1548391134,-0.2240208536,0.2639464438,0.4356185496,-0.0077086515,0.0810170174,-0.2269656956,0.3232060671,0.4250838459,0.126909703,-0.43144539,-0.1730945557,0.1214078814,0.1779445559,-0.1334395111,-0.1241060197,-0.3705472946,-0.3022031486,-0.1747015566,-0.0957440734,0.002574404,0.0565087125,-0.1349787116,0.1169960871,-0.0912767351,-0.3370675743,0.0810627416,-0.4117369652,0.2173947245,-0.0233671255,0.3630812764,0.3540457487,-0.0059769913,-0.4311004579,0.4017511606,0.1534221172,-0.1506371498,-0.1249693185,0.4914858341,0.5371119976,-0.1535212845,-0.1335526854,-0.0277760606,0.0029967285,-0.1223408952,0.0540128276,0.2383948565,-0.1113675386,-0.1098863631,-0.1630325466,0.3751423955,-0.4445958734,0.3816928566,-0.1641297787,0.0651184618,-0.1421230435,-0.224709183,-0.4564577341,-0.3743859529,-0.284876734,-0.5980314016,0.1653639525,-0.3122050464,-0.2530174553,-0.2069655508,0.3411161005,-0.1814827323,0.0084395213,-0.1815954745,0.2201620042,0.187446177,-0.0904239565,0.4492094219,0.3208989501,0.1303464621,-0.0994146317,-0.1213775352,-0.0566756092,-0.0879041404,0.1856535524,0.3309950531,-0.0504077822,0.1269751042,-0.2277490497,-0.2841596305,0.0142253274,-0.1667642742,0.2953528166,-0.2170961648,-0.2728697956,-0.0041689798,-0.314860642,-0.1530627459,-0.344350189,-0.1824578047,-0.1817972213,-0.3793221414,0.0949562863,0.095782958,-0.3661894202,0.3272184134,-0.093679972,0.1565567106,-0.3529231548,-0.3120750487,0.0425678864,-0.19191809,0.0191946402,0.0373649485,-0.0636336803,0.0886320546,0.4473897219,0.0682875365,-0.2181128263,-0.2077765465,-0.218667239,-0.088274762,-0.1306657046,0.0344410539,0.3023154438,-0.057647936,0.2030570805,-0.3977272511,-0.018455863,0.1599649042,-0.0437739454,-0.3815746605,0.392447859,0.1830078065,-0.2294371575,-0.3533361852,-0.326146394,-0.1929230988,0.4223927259,-0.5255027413,-0.0241620224,0.0357887708,-0.2785822451,0.1127695441,0.1251918674,-0.1200393364,-0.1015919223,0.3104766309,0.0435890779,-0.2384750396,0.2072957605,0.1077677086,-0.2218754888,0.0217224658,-0.3567278981,0.0503735952,0.0290435776,0.0944005325,0.277043432,0.1614022702,-0.0828790963,0.0974330902,0.0858204812,0.1860565543,-0.0896666199,-0.2755860984,0.1496620476,0.2805896997,0.2331838012,0.2243523002,0.2292753458,-0.063043341,0.5140989423,0.20800744,0.0556261316,-0.104147993,-0.0233318843,-0.1820419431,0.00504759,-0.4781915545,-0.0282481294,-0.1414563209,-0.3043084145,0.0748206079,-0.1577545851,0.0144820465,-0.0594567433,0.1999998689,0.0677571893,-0.45758003,0.0303161144,-0.5357938409,0.2113292962,-0.165037185,0.0746634081,0.0475669131,-0.0800646394,-0.0441265181,0.1198736131,0.2014363408,-0.2043307275,-0.4229975939,-0.1509073973,-0.1587821841,0.1874196231,0.2792785168,0.0823590979,-0.0913566723,0.0013253056,0.0624870509,0.4954948127,0.7230034471,-0.2819217443,-0.2045078278,0.1420887858,-0.2660385668,0.0259140618,-0.2045836896,0.0634126961,-0.0938839614,0.0959156081,0.22840482,-0.2615970075,-0.2743298709,0.773730576,0.3482439518,-0.0437669493,-0.3818064332,0.040544115,0.1989288181,0.1722832024,0.1197547391,0.3067525625,0.1788460463,-0.263659507,0.1035736054,0.0802274719,-0.1090175882,0.3623279929,0.0943774953,0.2493738532,-0.0701748356,0.4353824258,0.2356794477,0.106368348,0.1392605603,0.2227809727,-0.6137983203,-0.3588292003,-0.0213640537,0.0092540337,0.7729195952,0.4099295735,0.004097166,0.0827449113,0.060240563,0.1202039942,-0.1783584952,0.2659524083,0.3220493197,-0.0594890825,-0.5040188432,-0.7637935877,0.3296658695,0.1832436621,-0.1185743213,0.3174264729,0.1097892597,-0.586191833,0.3563875556,-0.0307516456,0.8402527571,0.1252982914,-0.1269396842,-0.1279689372,-0.1924713403,0.2156065553,-0.1647609174,0.0591560826,-0.1754287183,-0.1701663136,0.0659290403,0.0700342879,0.1669836342,0.3497168124,-0.072399348,0.2874290049,-0.0611942448,0.315091759,-0.2372083664,0.3600487709,0.2801631987,-0.2505971789,0.0356524363,-0.0113717513,0.072773926,-0.2413556427,0.2442024946,-0.0484637469,-0.0896628052,0.1317906678,-0.317253679,-0.2060526609,-0.167503044,0.0700305626,-0.059014298,-0.1223873124,0.3028968871,0.2367867529,0.5312367678,0.3700748086,0.0490423143,0.2635010779,0.2412390411,0.201323241,-0.3235864639,-0.047324013,0.4290888011,0.1332023442,-0.3335013092,0.0428418405,0.0747987926,-0.2521353066,-0.3842042089,0.110477075,0.3876135051,-0.654373765,-0.0626044422,-0.1125048399,0.320140183,-0.2469780445,0.1250055283,0.0599188283,-0.4423554838,0.1899299175,-0.15077281,-0.1416680366,-0.2008636147,-0.0343137197,-0.1398941129,0.042077601,0.2589150965,-0.1256493777,-0.0968632475,-0.1337192357,0.2407983392,0.2073108405,-0.2533508539,0.3469287157,0.0773177817,-0.2552933693,-0.0869388133,0.0673063993,0.1135153323,0.2626499534,-0.3044111133,0.1077474654,-0.2245469987,0.1135282069,0.2546497881,0.5764343143,-0.2348312587,0.2845920622,-0.4285892844,-0.1029544547,-0.2131628692,-0.1262359768,0.0857158154,0.1994167566,-0.1226366758,0.0781815946,0.2479048222,-0.2391082495,0.1612109095,0.1463834494,0.1289463639,-0.1457268745,-0.3270038068,0.139277935,-0.0562939569,0.2425657958,-0.2842194736,-0.1260062903,-0.0555741377,-0.3837483227,0.1106065661,0.3372883797,0.1458010226,0.5945234299,0.3972748518,-0.1521611214,0.1135407314,0.0934755355,-0.1003343463,0.3577884734,0.0185507033,0.2664894164,-0.0210022703,0.0388522521,-0.2621868253,-0.0964431316,-0.0161900744,-0.0803805217,-0.0262340941,0.0726046339,-0.2164149731,-0.1017519906,0.374994725,0.6219152212,-0.0251165573,-0.1903734356,0.4380894303,0.1251278222,-0.0155330366,-0.005557769,0.2794888616,0.0093912845,0.0778253675,0.5012418032,0.0126589136,0.3346582353,-0.1331539899,-0.0422241427,0.1388972551,0.0729200095,0.616724968,0.603295505,0.0429683253,0.1051749364,0.0790356845,-0.0545432195,-0.0081061227,0.2645595074,0.1413599551,0.3694077134,0.0159688387,0.0699227527,0.6006399393,0.365057528,-0.1086037531,-0.2764751911,-0.6752753854,0.1378994286,0.003504687,0.4029050469,-0.5035100579,-0.4012943804,-0.1975800246,-0.0319979042,0.006027522,-0.1104460657,0.0063267364,-0.1201000735,-0.1787228584,-0.0517515168,0.1769791692,-0.2383187413,0.4209338129,-0.0876793563,0.2677019536,0.090659298,0.0492176488,0.198245883,0.5991569161,-0.3150802553,-0.3227382004,-0.0698992759,-0.084241651,0.223815918,-0.1887881011,-0.0286738258,-0.0362362042,-0.0037658226,-0.3235393763,-0.0474772863,-0.0433065593,0.2021854222,0.2306499332,-0.0169332772,-0.0884166211,0.3804459274,0.1166617051,-0.0080757625,-0.0487259291,0.1965478808,0.2034022808,-0.6443556547,0.5634001493,0.0641526431,-0.0380412079,-0.1440705061,-0.0020379887,0.0053900741,-0.3700372279,0.2464805096,-0.0466667563,0.0320865251,-0.0453154333,0.0182371158,-0.1982019544,0.4282315373,-0.0705242008,-0.1208032817,-0.255595088,-0.089640826,0.0743729323,-0.2915914655,-0.0037942806,-0.2765485942,0.1966213137,-0.2916831672,-0.0787368715,-0.0342497006,-0.0117221624,-0.029026581,0.1444286108,0.3947259486,-0.3441644311,-0.3122899234,-0.1738013029,0.1536170691,0.3368343115,-0.134168908,-0.2430787683,0.1341224313,-0.3172352612,0.0556537583,-0.1340082139,0.5019895434,0.5471751094,0.2083570212,0.0402881466,0.4481987357,-0.0253648832,0.1909373105,0.0970267057,-0.059652213,-0.2668220401,0.3416194022,0.0094074244,0.2608264089,0.0513363853,-0.1047208458,-0.1050292626,0.3171876371,0.1245828271,-0.217676267,-0.1637939066,-0.125607878,0.0263712276,-0.0855356157,0.2697676718,0.2940776646,-0.2826081812,0.1645583957,0.382966876,-0.2912107408,0.4963204861,0.0327722467,-0.3277797997,-0.1031862348,0.1062436551,0.2823024988,-0.063306652,-0.6950249076,-0.3996464014,0.2895876765,-0.061067313,0.0544713587,0.2236010581,0.2504685819,0.0640293956,0.0995943025,0.2653733194,0.0949116424,0.088641189,0.1532568187,-0.2677426338]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/766","title":"[GEM] add DART data-to-text generation dataset","comments":"Is this a duplicate of #924 ?","body":"## Adding a Dataset\r\n- **Name:** DART\r\n- **Description:** DART consists of 82,191 examples across different domains with each input being a semantic RDF triple set derived from data records in tables and the tree ontology of the schema, annotated with sentence descriptions that cover all facts in the triple set.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2007.02871v1\r\n- **Data:** https:\/\/github.com\/Yale-LILY\/dart\r\n- **Motivation:** the dataset will likely be included in the GEM benchmark\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":7,"text":"[GEM] add DART data-to-text generation dataset\n## Adding a Dataset\r\n- **Name:** DART\r\n- **Description:** DART consists of 82,191 examples across different domains with each input being a semantic RDF triple set derived from data records in tables and the tree ontology of the schema, annotated with sentence descriptions that cover all facts in the triple set.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2007.02871v1\r\n- **Data:** https:\/\/github.com\/Yale-LILY\/dart\r\n- **Motivation:** the dataset will likely be included in the GEM benchmark\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nIs this a duplicate of #924 ?","embeddings":[0.0988729298,-0.0980768874,-0.0275114868,0.350076586,0.0993140265,0.0469358005,0.258694768,0.3785449266,-0.0969167948,-0.2644746304,0.2073587626,0.5722734928,-0.1000443101,0.0192577243,0.5630921125,0.0530177578,0.1939743906,-0.2244779617,-0.3903519213,-0.0217747465,0.106942229,0.3720033169,0.1015331373,-0.0193789154,-0.635699451,0.2571799159,-0.1006249562,0.1250395328,-0.2920392454,-0.3775512576,0.006468832,-0.0010598919,-0.0840749517,0.4894186556,-0.0001187936,-0.1333637387,0.1667408794,-0.0384612046,-0.1481407881,-0.2852532864,-0.210058555,-0.1043136045,0.0862447917,-0.0063329018,0.0289835744,0.0061654039,-0.2527204454,-0.3121990561,0.394235462,0.1942659318,0.1569889188,0.033913631,0.0199142564,-0.2601304352,0.2901246548,0.2960472405,-0.1966936886,0.1644269228,-0.0477752015,0.3365482986,0.1044842303,0.2548925579,-0.0377891175,0.1732083559,0.2677750587,-0.000854259,-0.3398010731,-0.0540848263,0.0438177474,0.5913277268,0.8983151317,-0.363299787,-0.3293124437,0.1702233851,-0.0129242064,-0.1437496692,-0.0794106126,-0.0914276093,-0.0259866752,0.031276457,-0.2138018161,-0.397338599,0.0883114189,-0.3343637586,0.1520093232,0.2257260382,-0.1210035309,0.0903150067,0.0361606739,-0.0034982038,-0.0090872608,-0.1915135682,-0.1762194484,-0.075630419,-0.1305220723,-0.0690736845,-0.2278890759,-0.1339284778,0.3035479188,0.1804266721,0.2035829574,-0.0193707421,0.0211625714,0.1496049762,0.3074965477,-0.1848963648,0.044335831,0.0368999504,-0.2244940698,-0.3875078857,0.011465434,0.048581738,-0.2552141249,0.223172456,0.0925284848,-0.2992147207,0.1698149443,-0.1166061535,-0.1215433404,0.2470410615,0.0773092657,-0.2973740399,-0.3376011252,-0.05626636,0.4398546815,-0.1319330037,-0.0393760726,0.1651702225,0.0881936178,-0.3748404384,-0.2507176995,0.1357170343,0.0231937263,0.0788443759,-0.2898987234,-0.1385315657,0.2476454973,0.0958974361,-0.1219097674,-0.2695758641,0.22706604,0.2696163356,0.1988913864,0.0687931404,-0.1379242241,0.2127192914,0.3740468323,-0.2079595327,-0.067153424,0.3547562957,0.0194183607,-0.2442435175,-0.4820486605,0.1965944767,-0.1424455345,-0.3130765259,-0.4166918099,0.5955224633,0.1373622417,-0.2137447298,0.2350785881,0.044000525,-0.1670071334,-0.2260800153,-0.0268068239,0.3685036898,-0.4458666146,-0.1463292986,-0.3063276112,0.0838676989,0.1560159475,0.2987765074,-0.3719986975,0.3442012072,-0.1406643689,0.3169780672,0.7513971329,0.0258749649,-0.0864702389,0.3657265306,-0.4230511189,-0.0874798521,0.2633732855,0.1928401291,0.084758237,0.0241459869,-0.2245238125,-0.3597026467,-0.0218284577,-0.2212935537,-0.1365464032,-0.2139517516,0.2315149456,-0.1666816771,-0.1431424022,-0.0936246291,-0.0301008746,0.0738776028,0.1043303534,-0.0316350125,-0.0516329817,-0.0487388708,0.2785757184,-0.0191222988,0.0426942818,-0.2108176202,0.0187276099,0.0219255742,0.6184740067,0.0165491197,0.6390024424,-0.3688984215,-0.1806297749,0.1318056285,-0.1339609474,-0.2385409027,0.0373388119,0.1418485641,-0.1612809151,-0.110103786,-0.4346881509,-0.0428754203,-0.1420986056,-0.0032330235,-0.0170046706,-0.2313560992,-0.2018598765,0.1709397733,-0.1273108423,0.031884104,-0.1584165692,-0.0425694026,0.1322999448,0.2030131221,0.0358976349,-0.0289500598,0.4761678874,0.1439667195,-0.0547991022,0.0039592856,0.1740035266,-0.1044699326,0.0595470816,-0.0822093189,-0.2062324286,-0.0327326879,0.3180542886,-0.1099602357,-0.0616687499,0.0029061609,0.0848505571,0.0487625599,0.0295694079,-0.1483750045,0.1645673662,0.0748133287,0.1418149322,0.2813539207,0.0569862798,0.4797866344,0.2565688789,-0.2023940086,0.4014044702,0.0102691799,-0.1661016196,-0.4657965302,-0.4106000066,0.6037593484,0.4670322835,0.1278611571,0.2289361656,0.0802996755,0.175574556,-0.1084442884,0.3508048058,0.064515844,0.0467643887,0.2156829834,-0.1078708023,-0.0773476809,-0.515571475,0.0483757146,0.3778814077,0.0970778912,-0.1156402528,0.0187050067,-0.0567044653,0.5352275372,-0.5397740602,-0.1553143561,-0.2948694527,-0.2834212184,-0.0217515901,-0.1148894951,0.2102578878,0.0176412351,-0.1807465404,0.2472441643,-0.3219151497,0.06098295,-0.0526864901,-0.1935206205,-0.0059612552,0.1351085901,-0.0695126206,-0.0801018849,0.3668055534,0.0589908995,0.3221768737,-0.5971446633,-0.4483878613,0.1167039275,-0.3446514606,0.2959300876,0.2681049407,-0.054494869,-0.1231664196,-0.313144356,-0.0328082852,-0.1567122191,-0.3768523335,-0.3263103068,-0.098530218,-0.2274149656,0.0676250085,-0.4186515212,0.0890238583,-0.0937056988,0.1138701588,-0.1263243407,0.0072206124,-0.2170980722,-0.1147512794,0.2930325568,-0.1335349828,-0.1354702562,-0.5725265741,-0.3136395812,-0.0407283679,-0.1095159128,0.0471354313,0.152067408,-0.1426386982,0.5855349302,0.0204560608,-0.452355057,-0.0098248562,-0.0052763685,-0.0729337484,-0.1012635678,-0.1508395225,0.3285362124,0.1529829651,-0.2205057293,-0.1928496361,-0.1405508071,0.2592980266,-0.0368808582,0.3099412322,0.068221502,0.5151696801,-0.0812381059,0.8126270771,0.446205914,-0.0425883234,0.0581194349,-0.1701321006,0.3379823267,0.0084491409,0.0058836262,-0.1486030668,-0.1736524403,-0.1846275628,0.1969885379,-0.0290017538,-0.0187956747,0.2554332614,-0.0941857174,-0.2641350925,0.0410315953,-0.0594664775,0.029382756,0.1052393839,-0.0799755752,-0.1359203756,-0.0373152792,0.1904539615,-0.072464332,0.3401826918,0.3442643583,-0.2065854222,-0.4439557493,0.2071311176,-0.323297888,0.1979438365,0.2169160396,0.0022859129,-0.2155611068,-0.2410863936,-0.0058503207,-0.0704995021,0.4181734025,-0.3534934223,-0.235640794,-0.0143589946,-0.1153401136,-0.2249286473,0.0075795851,0.1707376242,0.2429126203,0.1947937757,-0.202701211,-0.1980549991,0.0689430535,0.656756103,-0.020881163,-0.1480041891,-0.0609047487,-0.168750152,-0.3228779137,-0.1812411249,-0.1236665547,-0.0071287681,-0.3138555586,0.0969131738,0.0061547467,0.060961321,0.1969413161,0.2031282634,-0.2151685506,0.291778326,0.1635988504,0.0514188111,0.1395184398,-0.2974291742,-0.0624705665,0.6524248123,0.0288486835,-0.2167010605,-0.155239135,-0.3309161663,0.3740257621,0.2339657247,0.0752095133,0.2793155313,0.0747181624,-0.3822289109,-0.090592362,-0.089272581,0.3950584829,-0.0094506638,-0.1291545182,-0.5858945251,0.420247972,0.1381185651,-0.1452927291,0.4729810655,-0.0222325753,0.0124130333,0.0172267184,-0.0551927686,0.9333055615,-0.045706328,0.401679337,0.0651424006,-0.0342301503,0.4074928164,0.092877686,-0.1259670705,-0.4495360553,0.0007196935,-0.1970116645,-0.1622864902,-0.0203026813,0.2087668329,-0.0797056258,-0.0084222658,-0.1075409278,0.6634231806,0.0715812519,0.1159664616,-0.0944952369,0.3804030418,0.3838323057,0.1368951648,-0.1408568621,-0.0738540441,0.0483471081,-0.0724556521,-0.2017084658,-0.2198576778,-0.5885322094,0.040042799,-0.0242282692,0.398454994,-0.2065500915,-0.4383299649,0.1919480413,0.3045876622,0.1704235524,-0.144043237,-0.1342657506,0.2318397164,-0.077362597,0.0611826032,-0.0260407832,-0.1831194609,-0.0425754599,0.1098757759,-0.2243956178,0.1009307727,0.0804208145,-0.257265985,-0.3312875032,0.0871459767,0.3353280425,-0.4275343716,-0.1303723603,0.612216115,-0.4032317698,-0.2729275823,0.072701335,-0.1018608883,-0.1470207721,-0.1621126235,0.1086275727,-0.3186969161,0.0368652157,0.3961189985,0.3225332499,0.0432607681,0.4178205431,-0.1108663902,-0.2124750018,-0.2080911547,0.0885689929,0.2260427028,-0.553683579,0.0248130877,-0.1825949103,0.1197067052,0.2363055646,0.2706229985,0.1691158116,0.2176901251,-0.5264430642,0.1929994226,-0.1805783212,0.0882899761,-0.3223178387,0.2953023911,-0.0627999231,-0.4011446536,0.0714578927,0.322284013,-0.238112092,0.1451880634,0.3327397108,0.0218558423,0.4325750172,-0.1136261672,0.2214685231,0.1204704419,0.028869329,0.1015625149,-0.1833716035,-0.1896610409,-0.0850067511,0.172476083,-0.0278012622,-0.1245050281,-0.1102598161,-0.0192324091,-0.0451807305,0.0168913528,0.2841817141,-0.0029871014,0.1366295666,0.0449630171,0.1005101874,-0.245605737,-0.2218289822,-0.1359584928,-0.0241825301,0.4902749062,0.2201190442,-0.1159670725,0.2581986487,-0.0951941162,-0.0594560578,0.21335271,0.4102405012,-0.0272050891,-0.0086146481,-0.260984987,-0.0383044966,0.1356027573,0.5035889745,0.2122598439,-0.2846780419,-0.1208239347,-0.2197920233,0.1193602532,-0.225446254,-0.4324531853,-0.0007738639,-0.0448882729,-0.1152649447,0.2119766027,0.2721165717,0.0091281468,-0.1500375867,-0.0857242271,0.4931603968,-0.5406800508,0.3032420874,-0.0727410391,-0.1149798259,-0.0729663074,0.3433061242,0.2397442609,0.2043849081,0.1425998956,0.2234502882,0.3507343233,0.2054551095,0.0776821151,0.3223765492,0.1043063849,0.346447438,0.4383260906,-0.0879150406,0.1910244972,0.3837686181,-0.0755568147,0.0592059083,0.2052500248,-0.4768886566,0.0083360337,-0.1016507521,-0.1476440877,0.4957711101,-0.1870937049,-0.1552146971,0.1818929166,-0.1414681077,0.0948634222,0.1785045713,0.0389696509,0.1882050484,-0.2064231783,0.1900179237,-0.0419975445,-0.0690920129,-0.422062546,0.0873901546,0.4368317723,-0.1876729131,0.2415230721,0.6833847165,0.5043447018,-0.0530195311,0.1666692346,0.0865822434,-0.3992860615,-0.017111972,0.0447194465,0.2123184353,-0.1126080155,-0.241735369,0.1980956644,0.1035953239,0.0349821337,0.0723369569,0.2705387771,-0.073227644,-0.3887533247,0.3480070233,-0.2196324468,0.0310821999,-0.2088199109,0.2679190338,-0.5332863927,-0.2253577709,0.7668530941,-0.0215279497,-0.1761582941,-0.0832003728,0.0638783947,-0.0050257407,0.3441522717,0.1115109101,0.149009943,-0.2609232664,-0.2122286111,-0.3162968159,0.0790937021,0.0060928524,0.1070630848,-0.0004046989,0.1981268078,0.1417787373,0.2186894715,0.2012457997,-0.2121328264,-0.3927355111,0.0651562288,-0.1557183266,0.0462429188,0.266189605,-0.0160137769,0.0191170666,-0.1658469439,0.1156618595,0.0858282447,-0.063562721,-0.2607768476,-0.0035680404,0.3212144077,-0.0579026006,0.125544101,-0.1547737271,0.6371241808,-0.0887152329,-0.2864796519,-0.0468890034,-0.1218662262,-0.2209957242,0.7509840727,0.1011247709,0.1625779122,-0.1541840881,0.1163823307,0.0634207651,-0.0837207213,-0.1105395705,0.0966355428,-0.3664129376,-0.2386519164,-0.1530487686,-0.1683168113,0.2043031156,0.4405966699,0.1496009976,0.310029,0.0364961363,-0.1461270452,0.5079833269,0.0140956482,0.1268916875,-0.3063159883,0.2120562196,0.3244034946,0.0509812087,-0.8261535168,-0.1225799471,0.5124255419,0.1247016937,0.0469575785,0.3988730311,-0.0929264501,-0.0542040356,-0.0171266068,0.5899211168,-0.261403352,-0.1750503331,-0.0936252624,-0.2796359658]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/766","title":"[GEM] add DART data-to-text generation dataset","comments":"Yup, closing! Haven't been keeping track of the solved issues during the sprint.","body":"## Adding a Dataset\r\n- **Name:** DART\r\n- **Description:** DART consists of 82,191 examples across different domains with each input being a semantic RDF triple set derived from data records in tables and the tree ontology of the schema, annotated with sentence descriptions that cover all facts in the triple set.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2007.02871v1\r\n- **Data:** https:\/\/github.com\/Yale-LILY\/dart\r\n- **Motivation:** the dataset will likely be included in the GEM benchmark\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n","comment_length":13,"text":"[GEM] add DART data-to-text generation dataset\n## Adding a Dataset\r\n- **Name:** DART\r\n- **Description:** DART consists of 82,191 examples across different domains with each input being a semantic RDF triple set derived from data records in tables and the tree ontology of the schema, annotated with sentence descriptions that cover all facts in the triple set.\r\n- **Paper:** https:\/\/arxiv.org\/abs\/2007.02871v1\r\n- **Data:** https:\/\/github.com\/Yale-LILY\/dart\r\n- **Motivation:** the dataset will likely be included in the GEM benchmark\r\n\r\nInstructions to add a new dataset can be found [here](https:\/\/huggingface.co\/docs\/datasets\/share_dataset.html).\r\n\nYup, closing! Haven't been keeping track of the solved issues during the sprint.","embeddings":[-0.1165715158,-0.1498909593,-0.1535683572,0.0439386107,0.260538578,-0.2065060586,0.185433805,0.389185071,-0.1367991269,-0.0169223361,0.3331602216,0.5668106675,-0.1821102947,0.1380637139,0.2359357327,-0.1724929959,0.1973419636,-0.1215189621,-0.5213478804,0.1080351546,0.133435756,0.2698468268,0.035192851,-0.1910237521,-0.6201092005,0.0308218896,0.0166552123,0.1336372495,-0.1805196553,-0.3659416437,-0.2929753065,0.0072043305,-0.163345933,0.3781469464,-0.0000997972,-0.1461132467,0.1192464978,-0.1305323839,-0.2418070138,-0.3008140922,0.059666127,-0.0532873571,-0.0201037638,0.074442789,-0.172281161,0.0258689914,-0.1052750722,-0.2220113724,0.3892162442,0.3362015486,0.3132357597,0.1191830337,0.0982175693,-0.2955075502,0.4428591132,0.1097980142,-0.3314697146,0.0464075394,0.2092490196,0.1897605211,-0.1730675548,0.3517737389,0.0577264763,-0.0246350076,0.2585192621,0.096790649,-0.1810339838,-0.2130011916,-0.1539095938,0.4925748408,0.6532606483,-0.2809268832,-0.2915416062,-0.0507242307,-0.1524730474,-0.1382891089,0.0392585397,0.0710814893,-0.2020069063,0.14568533,-0.2726161778,-0.3349556327,-0.0317715071,-0.2182060778,0.3974118829,0.3366064131,-0.1089653149,-0.2625451684,0.0517809279,-0.0541125014,-0.1883223802,-0.0172898341,-0.1796721518,-0.1492447704,-0.098371543,-0.1058865562,-0.1038891748,-0.2657783329,0.332143575,0.1423376501,0.0604174882,0.2241337895,-0.0226277541,0.0684215873,0.0461447649,-0.1365075558,0.1740273386,-0.1645440459,-0.0377706923,-0.1889099926,0.0872607827,0.002014569,-0.1512589604,0.3136811554,-0.0947684795,-0.2358093411,-0.1609538049,0.073098585,-0.0491508618,0.196730867,0.3594150245,0.1057792529,-0.2781334519,0.3874655664,0.1394106597,0.0388766974,0.0251988638,0.2252260596,0.2270750552,-0.4081973732,-0.1553746313,0.0101712514,-0.0715117678,0.0961304754,-0.0878260881,0.1084317639,0.2315607816,0.1326463819,-0.1647238284,-0.2060151696,-0.0428748988,0.0611430444,0.1585363299,-0.048825752,-0.1285116673,-0.1586810201,0.2517712414,-0.4023594856,-0.0447046533,0.1603981853,0.0406379402,-0.1097391248,-0.618694663,0.3249977231,-0.1270263642,-0.2856718898,-0.2304139733,0.7588210702,0.0950612277,-0.016381437,0.0818202049,0.1526955068,-0.1285683513,-0.0230773464,0.0994677693,0.3376885653,-0.1666916609,-0.2157339454,0.1123826131,0.1107927188,0.1499939412,0.0912837088,-0.1027064249,0.2809126675,-0.0017362954,0.2478442788,0.5600636005,0.1997710168,-0.0153468763,0.1967267394,-0.3695555627,-0.1073290259,0.2704252601,0.0445325449,0.1718811691,0.0760390908,-0.1367870718,0.1550159007,-0.0099579524,0.1075245962,-0.3213868737,-0.1592086852,-0.0996386558,-0.04123291,-0.0261284392,0.000569199,-0.0003524939,-0.0635776222,0.2092159539,-0.1702394485,0.2260420173,0.1303931028,0.1334885806,0.1521532685,-0.0049922438,-0.4654440284,0.0085852668,-0.0491103008,0.5721980333,0.2055872828,0.2405096143,-0.2698643208,-0.1150939316,0.1573669612,0.1232749447,-0.0815620571,0.2380186319,0.1218586415,-0.1380584836,0.0518741868,-0.3135104179,-0.1425583959,-0.2048976719,0.19434008,0.1022189781,-0.1246024221,-0.1029465348,-0.2458452731,-0.0338945165,0.0377398618,-0.4477855265,0.0633279309,0.0315435119,0.3392884433,-0.0248928685,-0.0626844689,0.364782095,0.226969704,-0.1220436618,-0.0210109223,0.1069286838,-0.0906449035,0.0591600873,0.0565424338,-0.3695496917,0.0648393407,0.3528890014,0.0935240388,0.0904949978,0.204093948,0.2107468545,-0.0346663482,0.0285149794,-0.2066206932,0.1092444286,-0.2869579494,-0.1754281372,-0.0214460548,-0.1569079608,0.3283194304,0.4449425936,0.0423674583,0.6247909069,0.1997539699,-0.1794262528,-0.3734712899,-0.2994446158,0.3275955319,0.03594492,0.3196677268,0.167091608,-0.127369836,0.1325827837,-0.1444288939,0.287813127,0.0856088027,0.2351730764,0.3613308668,0.0754039288,-0.0512938239,-0.3199890554,-0.1770583242,0.0329321809,0.0330944024,-0.1204688102,0.1235278398,-0.0210624505,0.2796016932,-0.1243761256,-0.1955609024,-0.0944597498,-0.1808004826,0.2294594198,-0.0707854182,-0.0705101117,0.0688278899,-0.0594636872,0.2100172937,-0.1810294241,0.1303524971,-0.1797599643,-0.4284047782,-0.0232715122,0.2465163022,0.0317817256,-0.0007636692,0.4297660291,-0.0354573689,0.1278185546,-0.1593680829,-0.338712126,0.216504544,-0.3257946372,0.3702278435,0.1482627094,-0.0978470817,-0.0699263811,-0.2621046901,0.1357106119,-0.4005232453,-0.2407838553,0.0019144421,-0.1137961447,-0.2800056636,-0.1241418645,-0.4636096358,0.1895116717,-0.4452641606,0.0701606274,-0.0541283898,0.0698396266,-0.0829307288,0.1410372555,0.2990581691,-0.2412221432,0.1465534568,-0.151132077,-0.1785356998,0.0938367322,-0.1940399408,-0.2100235373,0.0158148836,-0.2255463004,0.257963568,0.0688639581,-0.2542402148,0.0689195022,-0.1384599358,0.0063318228,-0.0671823025,-0.324074626,0.2562712729,0.2003087848,-0.374694854,-0.1641590595,0.0135842189,0.1992840022,-0.1384343952,0.2085693926,0.0825936869,0.4492636025,0.1741908938,0.6949334145,-0.0340887979,-0.08740215,0.1039057821,-0.1723948568,0.1874068081,-0.0903564394,0.0847252905,-0.1261732876,-0.120888561,0.0290806275,0.1871547103,0.0645216629,0.0428812504,0.0672931522,0.0860505924,-0.2956048846,-0.0799993426,0.1849321276,-0.1435480118,0.1035513356,-0.1719338447,-0.083643131,0.068886824,0.0412440225,0.0701918527,0.0995467529,0.2663429081,-0.1246131584,-0.3607559502,0.3844753504,-0.1209727153,0.0696874633,0.3459402621,-0.0560411923,-0.2657413483,-0.4430799782,-0.1476469636,-0.0703251734,0.2152743042,-0.1894786209,-0.2073733956,0.0174107924,-0.1135666147,-0.1343559325,0.2052021176,-0.0992312878,-0.0481420755,0.2607309222,0.1375450343,-0.2008799911,-0.0808950737,0.3520312607,-0.0959809422,-0.0379314087,0.0218762606,-0.1342662722,-0.0269196574,-0.447335273,0.087160714,0.1879337579,-0.1219309941,0.0064930422,-0.1569654197,0.1176359877,0.2505837679,0.0261436086,-0.1266283691,0.1648120284,-0.0943299979,0.0398372747,0.0060172635,-0.157458961,-0.2148853093,0.310328126,0.1955432892,-0.2678176165,-0.2317107469,-0.1118957922,0.5726761818,0.1424116939,0.1816025227,0.3518012166,0.037607953,0.0732232332,-0.117017813,0.0609685294,0.1746285856,-0.1292452514,-0.0090008127,-0.426925391,0.278249681,-0.1937415153,-0.0242010914,0.1105456203,-0.1043256149,-0.0201144423,0.3952841461,0.1623084694,0.9261808991,-0.0712033808,0.2905678153,0.0836390331,0.1936235875,0.3770510256,-0.1571937501,0.082583271,-0.4151521921,0.045041319,-0.0346356183,-0.0251046009,-0.1056252569,0.0080453837,-0.1988641769,-0.0149705959,-0.0573670603,0.4939119518,0.0615813397,0.0903760642,-0.2605692744,-0.2442173213,0.0665751174,0.3035630584,-0.1324015111,-0.4097515643,0.0535100102,-0.1759440452,-0.3317796588,0.0220654197,-0.4327026308,0.2419035733,0.0204395987,-0.0085094543,0.0090859383,-0.2123127878,0.0826989114,0.4755913019,0.0202569403,0.2477395087,-0.2034921795,0.1941388398,-0.2079037577,-0.0794658735,0.0871902257,-0.0047396924,0.1704165936,-0.070648551,-0.2956527472,0.106964387,0.0570235774,-0.188574627,-0.1003850624,-0.3214813471,0.5800759792,-0.1747535616,0.1666214019,0.3485658169,-0.3621776104,-0.3868846297,0.2830595374,0.1124810427,-0.0701518506,-0.0462790206,0.2306562364,-0.0792665109,-0.0607864596,0.2294187546,-0.0659821406,0.1556735039,0.1192581058,0.2455958128,-0.2787877023,-0.4312754273,-0.0473550186,0.0248816237,-0.4599800706,0.1271917373,-0.082826905,-0.261654377,0.0263818949,0.3298555315,0.2149818838,-0.1254110932,-0.4379013479,0.0500923172,-0.2335363179,0.2601082623,-0.2705912888,0.1863708794,-0.2008126527,-0.4787630141,-0.0759548768,0.2935267389,-0.4990960956,-0.3012329042,0.0399609208,0.0403070897,0.3633587658,-0.2924446166,0.1766660213,0.0247659162,0.2197746485,0.1383731514,-0.1241815761,-0.400608927,-0.2167790532,0.0662255436,0.0578460768,0.0544000193,-0.1711077243,0.1726521254,-0.1839154959,-0.3579123616,0.2174836099,-0.0860442743,0.0333500542,0.1074248552,0.0334367305,0.2011024505,0.0606637672,-0.2599824667,0.0162948221,0.2650108933,-0.0435847081,-0.238947317,0.1211241409,-0.1795321703,0.0950341001,0.384967804,0.4249205291,-0.0003660894,0.2454815209,-0.0997762159,-0.0034887083,0.2424364686,0.3354515433,0.0503730699,-0.1998635828,-0.0333393216,0.1245561764,0.4007697701,-0.0827215463,-0.2995687127,-0.3071708381,-0.1641791463,0.0919094905,0.24236238,0.1141037643,-0.1713225394,-0.2621734738,0.0596946403,0.2554591894,-0.4275546372,-0.1463738382,-0.0895744637,-0.059045542,0.0160665102,0.132401973,0.0604429543,0.4192325175,0.1842124313,0.4129287601,0.3860855401,0.1060342491,0.1783267856,0.1417056769,0.1189149693,0.2870290279,0.133878082,0.0473706722,0.2359395325,0.1557631791,0.0390746482,0.1192451641,0.0002890812,-0.4375512004,0.0721774474,-0.0565718636,-0.1220239475,0.1933514029,-0.1832424849,-0.3041761518,0.0049323374,0.0650648326,0.127777949,0.11847464,0.0523641072,0.3291581571,-0.2734803557,0.1003848165,0.0041728271,0.1913673729,-0.5118032694,0.1887443513,0.2488616407,-0.0596719496,0.0444657654,0.3537020385,0.3933276832,0.0903520435,-0.0900077,0.1201020181,-0.5065389872,-0.1811780035,-0.1216164976,0.421444118,0.2122554928,-0.0171915852,0.2497558147,0.2684254646,-0.2805714011,0.0739318505,0.0865992308,0.0205074158,0.128135249,0.1778277159,0.1867726445,-0.0919190571,-0.1867256463,0.0900227278,-0.4382068813,-0.3870444,0.5009947419,-0.1617409438,-0.0705900565,-0.1446658224,0.1687218547,0.229646787,0.3206554651,0.0383556932,0.1701887399,-0.0263254084,-0.1508025825,-0.1290848553,-0.1143188477,0.1711804867,0.0999951288,-0.0092502506,0.1686090082,0.3334142864,0.3495034277,-0.136463806,-0.2151177227,-0.292118609,0.1837705821,-0.1592653394,0.1831764281,0.1509350985,0.0642901212,0.1814620942,-0.1594055593,0.0072506047,0.0973292887,0.1515723467,-0.2297084779,0.0294701718,0.3255977631,-0.3726344705,0.1761813164,-0.1810970157,0.2968932688,-0.2599690259,-0.2761796117,-0.1311565191,-0.2272078991,-0.3021377027,0.3453723192,0.0340980403,0.3032814264,0.0586679056,0.1503698975,0.0986611173,-0.0599968061,0.010710855,0.1960448623,-0.2368569821,-0.196191743,-0.0654784068,-0.0247029513,-0.0092873452,0.270147264,0.0033766499,0.1404532343,-0.0202693082,-0.2575499713,0.3462090492,0.1513204575,-0.1443157792,-0.180433929,0.2275689095,0.5469304323,-0.0042525991,-0.56985116,-0.0887095928,0.2386969775,0.1852922291,-0.0721036643,0.3853772581,-0.0648666769,0.0329451896,-0.0698629022,0.5610318184,-0.3088130355,-0.0821788907,-0.2214787006,-0.2792749405]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/761","title":"Downloaded datasets are not usable offline","comments":"Yes currently you need an internet connection because the lib tries to check for the etag of the dataset script online to see if you don't have it locally already.\r\n\r\nIf we add a way to store the etag\/hash locally after the first download, it would allow users to first download the dataset with an internet connection, and still have it working without an internet connection.\r\n\r\nI'll let you know when we add this feature.","body":"I've been trying to use the IMDB dataset offline, but after downloading it and turning off the internet it still raises an error from the ```requests``` library trying to reach for the online dataset.\r\nIs this the intended behavior ?\r\n(Sorry, I wrote the the first version of this issue while still on nlp 0.3.0).","comment_length":75,"text":"Downloaded datasets are not usable offline\nI've been trying to use the IMDB dataset offline, but after downloading it and turning off the internet it still raises an error from the ```requests``` library trying to reach for the online dataset.\r\nIs this the intended behavior ?\r\n(Sorry, I wrote the the first version of this issue while still on nlp 0.3.0).\nYes currently you need an internet connection because the lib tries to check for the etag of the dataset script online to see if you don't have it locally already.\r\n\r\nIf we add a way to store the etag\/hash locally after the first download, it would allow users to first download the dataset with an internet connection, and still have it working without an internet connection.\r\n\r\nI'll let you know when we add this feature.","embeddings":[-0.2098324746,0.2670384049,-0.0017677535,0.1898869723,-0.0595396459,-0.0748802498,0.3903472424,-0.0172846429,0.1707941145,-0.0225018989,0.1237998679,0.0126209315,-0.0386806279,0.0756761879,0.0625319555,0.1871642321,-0.1557508409,0.2216936052,-0.1252215505,-0.0784502327,-0.1236611605,0.1148991734,-0.4275744855,-0.0501417667,-0.1870036274,-0.3171657324,-0.0034036417,-0.0439859852,-0.1552307904,-0.2724395096,0.4218712747,0.0848470479,0.2210721076,-0.1517106444,-0.0001220542,-0.2454779446,0.4161519706,-0.1439789981,-0.4349360764,-0.3407869637,-0.275238812,-0.130433321,0.2988806665,-0.5066426992,0.0441146605,0.1309671402,0.3384900391,-0.4647698104,-0.1303227693,0.1508988738,0.0914961025,0.1822045296,-0.0067803198,0.0909337997,0.4891792238,0.0798255131,-0.2148955464,0.1665281504,0.1152893752,0.0191016942,0.1782421768,0.0184437726,-0.1545058042,0.0909674913,0.3861879408,0.1950401664,-0.1535570621,-0.264256984,-0.0379818715,0.3704021275,0.5538470745,-0.1437790245,-0.2343484908,-0.3610733449,0.0763947517,-0.2982944846,0.1947353929,0.2515519261,-0.0451297611,0.1533563435,-0.3593324721,-0.4431596398,-0.1371422261,0.5251787901,0.0860984623,0.2136146724,-0.1137967482,0.1065247357,0.0008924592,0.0111899041,0.0522401333,-0.3040782809,0.1637428999,-0.095231548,0.2223513871,-0.0641377941,0.1375614554,0.4724583626,-0.1811282486,0.3028615415,0.1037672609,-0.0921731666,-0.1592959911,0.2099303901,0.5247608423,0.3245604634,0.3381833136,-0.2607774436,0.5273810625,0.2759962082,-0.0852581114,-0.1607044637,0.1783451736,0.1802045852,0.0469965003,-0.1126272827,0.283922106,-0.3313288689,-0.1025822237,-0.1161459088,-0.0363384522,0.0267836638,0.2100475878,0.0840593576,-0.2664492726,-0.1593923569,0.0865489244,0.2714380324,-0.182312429,-0.4581656456,0.0118206562,-0.2709528208,-0.1404256076,-0.0769108534,0.4374878705,0.0036356177,0.3053943515,-0.2534536123,-0.09133441,-0.0206222814,0.4746436775,-0.2036610395,0.3540556431,0.2140980214,0.3410397768,0.0610174984,-0.0455781445,-0.0827937499,-0.0744462088,0.2547931671,-0.0524521917,-0.4573020637,0.0692087784,0.0919585526,-0.4702423215,-0.3144265413,-0.2976517379,-0.0881111026,0.0970745608,-0.5006312132,0.0282156132,0.0839831457,-0.1873994172,0.0166355651,0.024963472,0.4703622162,-0.3674836159,-0.065719232,-0.4391421974,-0.1759605259,-0.0243706182,0.2671213746,-0.425421983,0.3243421316,-0.3204441965,0.1207296401,0.6827148199,-0.1885617673,-0.6827122569,0.3701702356,-0.0995323807,-0.1788527668,-0.0034115152,0.3794584572,0.4256163836,-0.0089578917,-0.4156363606,0.526925683,0.0929822326,0.0212123487,-0.2327948958,-0.1643290818,0.2605100572,-0.0170766078,0.0112660453,0.3139621913,0.3630164564,0.1670653373,0.1006167307,0.3052434623,0.3341993093,-0.1018294767,0.0742243305,0.3801196814,-0.1144180223,-0.0251530949,-0.496630609,0.0201001372,-0.1487763673,0.0987895653,-0.5461813211,-0.4241181016,-0.0341123566,-0.2377707958,-0.0445402674,0.2871575952,-0.0388877802,0.1060238928,0.0273684487,0.1294902116,-0.3869740367,-0.0304966494,-0.0898807198,-0.0513926893,-0.5284494758,0.2566904724,-0.1686318219,0.0119885337,-0.1151870266,-0.08566311,-0.0241342541,-0.1413842738,-0.1097180918,0.3671101928,-0.2565329671,0.0871693268,0.1891967803,0.2710218728,0.2855142653,-0.3220532537,0.4973919094,-0.0081949756,0.2154596895,-0.0381353423,-0.3263987005,0.4724673033,-0.0683582947,0.2161989808,0.1441254616,0.0053595509,0.3300654888,-0.1547808796,-0.251830101,-0.582813859,0.0283664428,0.2074073851,0.0082336636,-0.0725473017,-0.1747859716,0.1800655276,0.5578531027,0.2726838887,0.2287419587,0.0783198699,0.0125849415,-0.2767373323,0.0853907838,0.1471385062,0.318325758,0.1985955089,0.3152204454,0.1368162483,-0.0074230353,-0.4224086106,-0.0591056459,0.08879558,-0.0197892617,-0.0263547413,0.0225592107,-0.0377119742,-0.3606550097,-0.0614188686,0.2296983153,0.2251725644,-0.2382855415,-0.0716267601,-0.6653861403,-0.7250332236,0.0148601914,-0.0738799125,-0.4821614325,-0.2052155584,0.0278104208,0.4159468114,-0.1507524848,0.0593610033,-0.2272671759,0.5010195971,-0.3477635682,-0.2220639735,-0.0665618703,-0.3527732193,-0.2486335337,-0.0017851433,0.3669335544,0.1794337332,0.4034474194,-0.1142761111,-0.1231461391,-0.2343118489,0.0056437459,0.0564889796,0.3319588304,0.0723532587,-0.0066338349,-0.107201986,-0.1433563381,0.0088691209,0.1476003081,-0.3277746439,0.0985141546,-0.049735833,-0.0195473451,0.0176217649,-0.3477735519,-0.4379716218,-0.2635905147,-0.1216423288,0.1585334092,0.1391206235,0.1520963162,0.0390637219,-0.1248000041,0.2505188882,-0.1376264691,0.0951220021,-0.0327172242,-0.4872590005,0.1624234766,-0.3436885476,-0.2576625347,0.4373550415,0.0587734766,0.1389320642,0.2758118808,-0.6511327028,0.0503680073,0.0381959118,0.169741258,0.3652427793,0.0016047496,0.5040283203,-0.2468517274,0.0363385603,0.0136324381,0.0794102326,0.3829860091,0.1882002801,0.2508431673,0.2320654541,0.3819538355,0.1213085353,0.1898973733,-0.1730674356,0.1170406044,0.6550409198,0.0645186529,0.4792505801,-0.3082836568,0.0291578453,-0.1174154952,0.1220538914,-0.2058031261,0.1454465538,0.1693542451,-0.4001720846,-0.3892103136,-0.0038684614,-0.5113554597,-0.3278262019,-0.0678065047,-0.22614941,0.1589135677,0.364777863,0.0479072258,-0.0485213809,-0.3160042763,-0.2498843223,0.1579231769,0.2608879209,0.0561422519,-0.1440253407,0.1137703955,-0.1902497858,0.2225916088,-0.0877936929,0.5331962109,-0.0404457003,0.1176536605,0.3277540207,0.1139679104,0.2481108755,-0.344139725,0.1711178571,0.113004975,0.0209714845,-0.4358959198,0.1319620907,-0.1606543362,-0.0598653965,0.5046318769,0.0172553789,-0.0680067912,-0.009995738,-0.0377483033,-0.0325024314,-0.1555780619,-0.2539770901,-0.2266999632,-0.0264634416,-0.5389502048,-0.0416597687,-0.205903545,-0.0964665413,0.0205958374,0.0667409152,-0.0529207774,0.1670696437,-0.2818746567,-0.0671864972,-0.0584724769,0.1195121855,0.0424493439,0.0909697711,0.0887786001,0.1651577652,0.4254350662,-0.0219534561,0.0714309886,0.187551558,0.0345635749,-0.0017001497,0.2828837633,0.0066326899,-0.2516593039,0.2152040452,-0.0694362,0.0846594945,-0.0642759949,0.2071262002,-0.5104782581,0.1333306283,-0.4468948841,0.2959697843,-0.0045052893,-0.0339875259,0.0833550841,0.365950644,-0.127603516,0.2154053599,-0.0442438014,0.8948523998,-0.0578305684,0.2072690874,-0.1336588562,0.0899424031,0.7640866041,-0.1290883273,0.2141270041,-0.1367114335,-0.0039025981,-0.1508295238,-0.0863255709,0.3068038225,0.4187225401,-0.1014700532,0.2692237496,-0.1646228284,0.2183361948,-0.2512001991,0.332308948,-0.3463969529,-0.1656761318,-0.1386066377,0.0241280459,0.0648591146,0.346332252,-0.1581467688,-0.0619405024,0.0075495932,0.1318493187,-0.1311005503,0.2823046446,0.2811912596,0.1110347062,-0.3716670275,0.1312393695,-0.0454718284,0.2234465629,0.3521395326,0.0173894111,-0.3721261322,-0.1389572322,-0.0207041446,-0.1765758991,0.2591637373,0.1892563403,0.3392987549,-0.2422880381,-0.2942515314,0.0028848248,0.1207222193,-0.1234666258,0.4066619277,0.0089007579,-0.0970207453,-0.1011501774,-0.1687726825,0.2969223261,0.0419694781,-0.0036161665,0.0299585741,0.1280787885,0.1855444759,0.0686011538,-0.1168610305,-0.2107695937,-0.1453315318,0.5676361322,-0.0454499386,0.0304639265,0.144708097,0.0587553754,-0.1140615866,-0.1490970254,0.1853619963,-0.3498146832,-0.37835446,0.100266628,0.2122075111,0.4764332771,-0.0790759698,0.1829330176,-0.1413034797,0.2391072065,-0.2073705792,-0.3093141913,-0.2009140849,0.1761251986,0.2414706647,0.0257361289,0.0215906203,-0.1163560823,0.5522446036,-0.1803213209,-0.2020921111,0.2219247222,0.0396476015,0.3655430079,0.2100832462,0.0063195936,0.0308527406,-0.1672137529,-0.0977066755,-0.0158760268,-0.2581114173,-0.0951535478,0.0318226852,0.2842283547,0.2625710964,-0.5424134135,-0.3297531307,-0.640756011,-0.2728307247,0.0241430495,-0.0947715715,0.4041729271,0.0400393307,0.1986295283,0.3577870727,0.1495356113,-0.2055553347,0.1183987483,-0.1634669304,0.285066098,-0.0969314948,0.194829464,-0.0042722081,0.0401362889,-0.2721362114,-0.0828468055,-0.058149498,0.1644284576,0.5369650722,-0.2493012995,0.2751140594,0.1975863874,0.4544934928,0.259500742,0.1570683867,-0.2199498713,0.3323820233,0.0507632121,-0.2865042984,0.1005188078,0.3589014709,0.4310481846,-0.1354837865,0.1744440794,0.431504488,0.1162718236,-0.1053698435,-0.1573918164,0.3952949345,-0.2768719196,0.1150014549,0.0184405353,-0.265710175,-0.0092565073,-0.051172059,0.2096353024,0.2432672232,0.2710631192,-0.1154282689,0.0126862684,0.189111501,0.3074535728,-0.1063921824,-0.4238238931,-0.1150287166,0.1230169088,0.1771202385,0.0514835604,-0.1880192906,0.1199184209,-0.1349375993,0.1401914656,-0.1819642633,0.4822356105,-0.0093495604,0.0067445156,-0.4403430223,-0.2450457662,-0.0685571134,0.1366954446,-0.2170639038,0.0558843724,-0.1265665591,0.2268839628,-0.0778879076,-0.3708289564,0.3754977286,-0.0852470323,0.1979827881,-0.2669508755,0.1558912396,-0.0620208867,0.2027586699,0.2308393866,0.3019276261,0.1709937304,0.211199373,-0.0228935517,0.1909483224,0.0558776222,-0.0773330107,-0.3302193582,0.2316437662,0.6044890881,0.5322275162,0.2374630272,-0.043161273,-0.0579138212,0.4244399071,-0.0598479249,0.256264925,-0.074632898,0.4026714265,0.003264043,-0.1355200857,-0.1652109474,0.0201941803,-0.2610503435,-0.3235000074,0.2602367699,-0.3222095966,0.1295648664,-0.4470629692,-0.0216494221,-0.2157458216,0.5108923912,0.4062946141,0.3394335508,-0.1166099757,-0.2419133186,-0.7239391804,-0.0030534735,-0.0680556744,0.1875721067,-0.0414260328,0.4513482153,-0.1687252969,0.2156714648,-0.3415602148,0.1776656806,-0.0359168276,-0.0597198494,-0.0325577594,-0.0526741557,-0.3596999049,0.0883637965,-0.1332339942,-0.2659159303,0.0844654813,-0.381169796,-0.1640300602,0.124690555,-0.2449233532,-0.2417311817,0.3652911484,0.5970693827,0.1111268401,0.4610145688,0.0763032809,-0.0047109942,-0.2299119085,0.0996695906,-0.3166602552,0.0791570097,0.0090571567,0.0214247424,-0.2934352756,0.4062313437,-0.0983727947,0.2165817916,-0.0734515339,0.318205744,-0.1857285649,-0.3723779619,-0.2545109689,0.102100879,-0.0136554223,-0.1437838227,0.1199756041,0.0732648298,-0.4484457374,0.0452960618,0.5243942738,0.1416594684,0.2392503172,-0.3593874872,0.4845661521,0.393553406,-0.2534211576,-0.4270009696,-0.0509544499,0.4510918558,-0.1463803053,0.0994957685,0.037373662,-0.1538922191,0.1346619129,-0.0968453437,0.4045368433,-0.1079006121,-0.3252615929,-0.1765982658,-0.2115278691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Are you running the script on a machine with an internet connection ?","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":13,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nAre you running the script on a machine with an internet connection ?","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Yes , I can browse the url through Google Chrome.","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":10,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nYes , I can browse the url through Google Chrome.","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Does this HEAD request return 200 on your machine ?\r\n```python\r\nimport requests \r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\")\r\n```\r\n\r\nIf it returns 200, could you try again to load the dataset ?","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":28,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nDoes this HEAD request return 200 on your machine ?\r\n```python\r\nimport requests \r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\")\r\n```\r\n\r\nIf it returns 200, could you try again to load the dataset ?","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Thank you very much for your response.\r\nWhen I run \r\n``` \r\nimport requests \r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\")\r\n```\r\nIt returns 200.\r\n\r\nAnd I try again to load the dataset. I got the following errors again. \r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 475, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"C:\\Users\\666666\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\cnn_dailymail\\0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\\cnn_dailymail.py\", line 253, in _split_generators\r\n dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\download_manager.py\", line 175, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\py_utils.py\", line 224, in map_nested\r\n mapped = [\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/drive.google.com\/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ\r\n\r\nConnection error happened but the url was different.\r\n\r\nI add the following code.\r\n```\r\nrequests.head(\"https:\/\/drive.google.com\/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ\")\r\n```\r\nThis didn't return 200\r\nIt returned like this:\r\n\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connection.py\", line 159, in _new_conn\r\n conn = connection.create_connection(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\util\\connection.py\", line 84, in create_connection\r\n raise err\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\util\\connection.py\", line 74, in create_connection\r\n sock.connect(sa)\r\nTimeoutError: [WinError 10060] \r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connectionpool.py\", line 670, in urlopen\r\n httplib_response = self._make_request(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connectionpool.py\", line 381, in _make_request\r\n self._validate_conn(conn)\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connectionpool.py\", line 978, in _validate_conn\r\n conn.connect()\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connection.py\", line 309, in connect\r\n conn = self._new_conn()\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connection.py\", line 171, in _new_conn\r\n raise NewConnectionError(\r\nurllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x000001F6060618E0>: Failed to establish a new connection: [WinError 10060] ","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":272,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nThank you very much for your response.\r\nWhen I run \r\n``` \r\nimport requests \r\nrequests.head(\"https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\")\r\n```\r\nIt returns 200.\r\n\r\nAnd I try again to load the dataset. I got the following errors again. \r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 475, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\builder.py\", line 531, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"C:\\Users\\666666\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\cnn_dailymail\\0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602\\cnn_dailymail.py\", line 253, in _split_generators\r\n dl_paths = dl_manager.download_and_extract(_DL_URLS)\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\download_manager.py\", line 254, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\download_manager.py\", line 175, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\py_utils.py\", line 224, in map_nested\r\n mapped = [\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\py_utils.py\", line 225, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\py_utils.py\", line 163, in _single_map_nested\r\n return function(data_struct)\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 300, in cached_path\r\n output_path = get_from_cache(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\", line 475, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/drive.google.com\/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ\r\n\r\nConnection error happened but the url was different.\r\n\r\nI add the following code.\r\n```\r\nrequests.head(\"https:\/\/drive.google.com\/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ\")\r\n```\r\nThis didn't return 200\r\nIt returned like this:\r\n\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connection.py\", line 159, in _new_conn\r\n conn = connection.create_connection(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\util\\connection.py\", line 84, in create_connection\r\n raise err\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\util\\connection.py\", line 74, in create_connection\r\n sock.connect(sa)\r\nTimeoutError: [WinError 10060] \r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connectionpool.py\", line 670, in urlopen\r\n httplib_response = self._make_request(\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connectionpool.py\", line 381, in _make_request\r\n self._validate_conn(conn)\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connectionpool.py\", line 978, in _validate_conn\r\n conn.connect()\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connection.py\", line 309, in connect\r\n conn = self._new_conn()\r\n File \"C:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\urllib3\\connection.py\", line 171, in _new_conn\r\n raise NewConnectionError(\r\nurllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x000001F6060618E0>: Failed to establish a new connection: [WinError 10060] ","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Is google drive blocked on your network ?\r\nFor me \r\n```python\r\nrequests.head(\"https:\/\/drive.google.com\/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ\")\r\n```\r\nreturns 200","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":15,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nIs google drive blocked on your network ?\r\nFor me \r\n```python\r\nrequests.head(\"https:\/\/drive.google.com\/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ\")\r\n```\r\nreturns 200","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"I can browse the google drive through google chrome. It's weird. I can download the dataset through google drive manually.","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":20,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nI can browse the google drive through google chrome. It's weird. I can download the dataset through google drive manually.","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Could you try to update `requests` maybe ?\r\nIt works with 2.23.0 on my side","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":15,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nCould you try to update `requests` maybe ?\r\nIt works with 2.23.0 on my side","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"My ```requests``` is 2.24.0 . It still can't return 200.","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":10,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nMy ```requests``` is 2.24.0 . It still can't return 200.","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Is it possible I download the dataset manually from google drive and use it for further test ? How can I do this ? I want to reproduce the model in this link https:\/\/huggingface.co\/patrickvonplaten\/bert2bert-cnn_dailymail-fp16. But I can't download the dataset through load_dataset method . I have tried many times and the connection error always happens .\r\n","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":56,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nIs it possible I download the dataset manually from google drive and use it for further test ? How can I do this ? I want to reproduce the model in this link https:\/\/huggingface.co\/patrickvonplaten\/bert2bert-cnn_dailymail-fp16. But I can't download the dataset through load_dataset method . I have tried many times and the connection error always happens .\r\n","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"The head request should definitely work, not sure what's going on on your side.\r\nIf you find a way to make it work, please post it here since other users might encounter the same issue.\r\n\r\nIf you don't manage to fix it you can use `load_dataset` on google colab and then save it using `dataset.save_to_disk(\"path\/to\/dataset\")`.\r\nThen you can download the directory on your machine and do\r\n```python\r\nfrom datasets import load_from_disk\r\ndataset = load_from_disk(\"path\/to\/local\/dataset\")\r\n```","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":75,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nThe head request should definitely work, not sure what's going on on your side.\r\nIf you find a way to make it work, please post it here since other users might encounter the same issue.\r\n\r\nIf you don't manage to fix it you can use `load_dataset` on google colab and then save it using `dataset.save_to_disk(\"path\/to\/dataset\")`.\r\nThen you can download the directory on your machine and do\r\n```python\r\nfrom datasets import load_from_disk\r\ndataset = load_from_disk(\"path\/to\/local\/dataset\")\r\n```","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Hi\r\nI want to know if this problem has been solved because I encountered a similar issue. Thanks.\r\n`train_data = datasets.load_dataset(\"xsum\", `split=\"train\")`\r\n`ConnectionError:` Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/xsum\/xsum.py`","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":26,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nHi\r\nI want to know if this problem has been solved because I encountered a similar issue. Thanks.\r\n`train_data = datasets.load_dataset(\"xsum\", `split=\"train\")`\r\n`ConnectionError:` Couldn't reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.3\/datasets\/xsum\/xsum.py`","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Hi @smile0925 ! Do you have an internet connection ? Are you using some kind of proxy that may block the access to this file ?\r\n\r\nOtherwise you can try to update `datasets` since we introduced retries for http requests in the 1.2.0 version\r\n```\r\npip install --upgrade datasets\r\n```\r\nLet me know if that helps.","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":56,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nHi @smile0925 ! Do you have an internet connection ? Are you using some kind of proxy that may block the access to this file ?\r\n\r\nOtherwise you can try to update `datasets` since we introduced retries for http requests in the 1.2.0 version\r\n```\r\npip install --upgrade datasets\r\n```\r\nLet me know if that helps.","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Hi @lhoestq \r\nOh, may be you are right. I find that my server uses some kind of proxy that block the access to this file.\r\n![image](https:\/\/user-images.githubusercontent.com\/46243662\/106456211-2ca24180-64c8-11eb-831e-47e9b40e7da4.png)\r\n\r\n","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":26,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nHi @lhoestq \r\nOh, may be you are right. I find that my server uses some kind of proxy that block the access to this file.\r\n![image](https:\/\/user-images.githubusercontent.com\/46243662\/106456211-2ca24180-64c8-11eb-831e-47e9b40e7da4.png)\r\n\r\n","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"> Hi @lhoestq\r\n> Oh, may be you are right. I find that my server uses some kind of proxy that block the access to this file.\r\n> ![image](https:\/\/user-images.githubusercontent.com\/46243662\/106456211-2ca24180-64c8-11eb-831e-47e9b40e7da4.png)\r\n\r\nI have the same problem, have you solved it? Many thanks","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":40,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\n> Hi @lhoestq\r\n> Oh, may be you are right. I find that my server uses some kind of proxy that block the access to this file.\r\n> ![image](https:\/\/user-images.githubusercontent.com\/46243662\/106456211-2ca24180-64c8-11eb-831e-47e9b40e7da4.png)\r\n\r\nI have the same problem, have you solved it? Many thanks","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/759","title":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py","comments":"Hi @ZhengxiangShi \r\nYou can first try whether your network can access these files. I need to use VPN to access these files, so I download the files that cannot be accessed to the local in advance, and then use them in the code. Like this,\r\n`train_data = datasets.load_dataset(\"xsum.py\", split=\"train\")`","body":"Hey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?","comment_length":49,"text":"(Load dataset failure) ConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\nHey, I want to load the cnn-dailymail dataset for fine-tune.\r\nI write the code like this\r\nfrom datasets import load_dataset\r\n\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctrain\u201d)\r\n\r\nAnd I got the following errors.\r\n\r\nTraceback (most recent call last):\r\nFile \u201ctest.py\u201d, line 7, in\r\ntest_dataset = load_dataset(\u201ccnn_dailymail\u201d, \u201c3.0.0\u201d, split=\u201ctest\u201d)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 589, in load_dataset\r\nmodule_path, hash = prepare_module(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\load.py\u201d, line 268, in prepare_module\r\nlocal_path = cached_path(file_path, download_config=download_config)\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 300, in cached_path\r\noutput_path = get_from_cache(\r\nFile \u201cC:\\Users\\666666\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\datasets\\utils\\file_utils.py\u201d, line 475, in get_from_cache\r\nraise ConnectionError(\u201cCouldn\u2019t reach {}\u201d.format(url))\r\nConnectionError: Couldn\u2019t reach https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.1.2\/datasets\/cnn_dailymail\/cnn_dailymail.py\r\n\r\nHow can I fix this ?\nHi @ZhengxiangShi \r\nYou can first try whether your network can access these files. I need to use VPN to access these files, so I download the files that cannot be accessed to the local in advance, and then use them in the code. Like this,\r\n`train_data = datasets.load_dataset(\"xsum.py\", split=\"train\")`","embeddings":[-0.1726205349,-0.0845809281,-0.0973370373,0.1288862824,0.4222565293,0.2019445449,0.3426600695,0.0621942766,-0.0575587004,0.0043574101,-0.1423110366,0.0824962258,-0.0660601109,0.2489195168,0.1123090535,-0.1254215389,-0.0808455199,-0.1146305054,-0.2291170061,0.0067147403,-0.2098105252,0.1440933049,-0.2096494436,0.0432117768,-0.4213045835,-0.1652373374,0.086121805,0.3555387259,-0.0301731415,-0.2586031556,0.3710347414,0.0136734368,0.0507145114,0.4067729115,-0.0001109412,0.2821536362,0.4214935005,-0.0077436431,-0.4700695574,-0.4119871855,0.0286166817,-0.2524267137,0.203611657,-0.1215822548,-0.08969336,-0.0993359014,-0.0485146157,-0.2351501137,0.4430734515,0.4244684279,0.2697091103,0.2736960649,0.3110575974,-0.0948241577,-0.3082554638,-0.2242909074,0.0436809249,0.5262338519,-0.3041549921,-0.0550037548,0.2369493842,0.1852680892,-0.1102643609,0.0394290015,0.1027930453,0.0368396975,-0.0811182633,-0.3631623089,0.0167236421,0.0721938908,0.2434540987,-0.2260054797,-0.4658710957,-0.1638950258,0.0304690413,-0.352312237,0.2059648484,0.072493799,-0.1310377121,0.1562265158,-0.4081345499,-0.1662252545,-0.2331795692,0.4767994583,-0.0834709108,-0.0691660941,-0.1013452783,0.1568791717,0.2081463486,-0.0147652691,0.2456072122,0.0163964685,0.3304591179,0.2135430425,-0.4810218215,0.115841113,0.0163757335,0.1395893097,0.135859251,0.2361456752,0.0839775056,-0.0570870079,-0.1238575578,0.0851527154,0.0606347546,0.2453833222,0.0089363651,0.1387266517,0.1872473806,0.305845052,-0.0122886831,0.0348720662,-0.1371922344,-0.2628456056,-0.08476042,0.1266726404,0.4672960341,-0.0668150634,-0.1457813829,0.2286665291,-0.2497208267,-0.0773124471,0.1255286932,0.5078197718,-0.2459857464,0.1030792743,-0.0911056921,0.2443113178,-0.0984816253,0.0281397756,-0.2782628238,0.1583172977,-0.1469017863,0.1858578175,0.4209389389,-0.4283124804,0.3650208414,-0.0827948749,0.1328939646,-0.2651251256,0.0727618709,-0.0731363744,-0.2295966446,0.3315565586,-0.0128483977,0.1878604293,0.3255273104,0.219899714,-0.0086199995,-0.0932961255,-0.2572560012,-0.5359469652,0.2843953967,0.2346502244,-0.1325822622,-0.2377674282,-0.1039849222,-0.0701877475,-0.0855135694,0.0499906503,-0.2254374325,-0.0758890808,-0.0001020563,-0.1851737052,0.4579845369,0.4795647562,-0.3562659025,-0.131882593,-0.0889203772,-0.1993902177,0.1022753343,0.2505486906,-0.2102564871,0.267824322,-0.0300623551,-0.200193122,0.3244513571,-0.2567795813,-0.4303721189,0.2295979559,-0.1415840089,0.0772204101,-0.159794271,-0.0198278166,-0.230043903,0.0976685733,0.2224139124,0.4544759095,-0.2219419777,-0.0075087207,-0.0761365741,-0.2887578607,0.1971730888,0.3495088518,0.1905551255,0.0837119222,0.112290211,-0.1229663789,0.1389329135,0.0792715624,-0.1288710833,0.2222590446,-0.0730446875,0.0600969754,0.005793402,0.1931337863,-0.4456617534,0.2171620876,0.2575122118,0.2054394931,-0.1591732353,0.0718008503,-0.3314308226,0.064478755,-0.3276307583,-0.1017315835,0.1681315899,0.1100925952,0.1149900779,0.1487406194,-0.3477537334,0.5853907466,-0.2807269692,0.1544780582,-0.3642602861,0.3336731791,-0.137807101,0.0208101701,0.0532537736,0.1948694438,0.2270952165,-0.2124796659,-0.1230695471,0.3350651264,-0.0970880687,0.1396979541,0.1005691662,0.0206223,0.0887210742,-0.2732220292,-0.0075916527,0.1752086878,-0.0484225005,-0.0636551455,0.2221704721,0.2624588609,-0.0824369416,0.3350197375,0.1653448045,-0.0167164225,0.155774191,0.0822781697,-0.0725755841,-0.080437921,0.348846674,-0.282363534,0.4914526939,-0.1779355407,-0.1883532107,-0.214296937,0.1073015034,-0.2060754895,0.0367991664,0.1807588488,-0.1527466178,-0.1959124207,-0.1415822357,0.2285894006,0.3831121027,0.1266689599,0.0174037032,0.3330525756,-0.1124231219,-0.1255553067,0.2139613777,0.1642096043,0.253416568,0.125690341,-0.1063401401,0.0876753926,-0.3335894048,-0.2977259159,0.0457495488,0.2941001952,-0.3505651951,-0.0266053155,-0.2348111719,-0.1075837016,0.0410018079,-0.4958513677,-0.1831308901,-0.4478105903,-0.1630501598,-0.0245089773,0.10010802,0.1913428754,-0.1383036971,0.2042535245,0.2681008875,-0.1757077277,-0.1990249157,-0.1413165927,-0.1222066134,0.042547036,0.1909830719,-0.2497787029,0.2615085244,-0.1237755716,-0.1931108236,-0.1338968277,-0.0432524495,0.0448913872,-0.0807589814,0.4253887534,0.0415047854,0.5178230405,0.1474954188,-0.0527280755,0.3786589503,-0.0750784799,0.0525866598,0.0137069914,-0.0126736257,0.1713650823,0.1054245755,-0.4197510779,-0.5957531929,-0.3186041415,0.1332792342,-0.140416503,0.160090819,0.4751754701,0.1390486509,0.0728231594,-0.0277652182,0.3243475854,-0.295239687,-0.5784931779,0.3427504003,-0.426258415,-0.3999854922,0.0517781228,0.2337378412,0.1585310996,-0.136862874,-0.4911381602,-0.2912989259,-0.0210318696,-0.2082522511,-0.1408883333,0.0342517607,-0.0817868337,-0.1706445515,-0.0807222873,0.0528550036,-0.215396449,0.0207666811,-0.1897796839,0.1059555113,0.0540594012,0.5800537467,-0.2069241405,0.4104644656,0.2627967,-0.0999407768,0.430472672,-0.1017871127,0.298207432,-0.0337928012,-0.5136099458,-0.039039325,-0.0248874258,-0.0493591093,0.246701926,0.1183115691,-0.1750100702,-0.4023716748,-0.1823289394,-0.4465422034,-0.1094008759,-0.0297658965,0.1102360114,0.0683861896,0.0337716155,0.117400825,-0.1562561542,-0.0286270864,0.0233907197,0.3478070199,-0.0388782918,0.1180436164,0.0067395014,-0.0921745598,-0.4314179718,0.3346539736,-0.174461022,0.4283652902,-0.1977282017,0.0767524093,-0.0671717823,-0.0039037017,0.6058324575,-0.265165776,0.1158642173,0.0965194181,0.0104380129,-0.3042702973,-0.093921043,-0.0325787701,0.2487047911,0.1657653004,0.4472000897,-0.0785692483,-0.167560637,0.3270871043,0.2548521757,-0.139084354,-0.3214136064,-0.1426791102,-0.3935246766,-0.2281122506,-0.2242874205,-0.340064913,0.3046940565,-0.0341386721,-0.1692191362,0.0657262579,-0.0941395164,0.1747616827,0.0775278434,0.1345154792,0.0945736766,0.1138609573,0.3893180788,0.0514724366,0.1969855726,0.6876193881,-0.174729377,-0.2818039656,0.0031070828,0.2653590143,0.1266304404,0.2776075304,-0.2134698927,-0.1107277125,0.1917884648,0.1974568516,-0.0059067407,0.2809213698,0.3785067499,0.0910988823,-0.595351696,-0.5943185687,0.4862521589,-0.0949102864,0.1161979958,0.1835698783,-0.1201060936,-0.1307161301,0.1753077954,-0.187167719,0.7539357543,-0.0516213477,0.0220213942,0.2660264075,0.2764648199,0.4896038473,-0.3141959012,0.0171414465,-0.2213471234,-0.3019835651,-0.0965612605,-0.1128649339,0.1232143492,0.27900213,0.0411609039,0.4529607296,-0.1397987157,-0.0076717441,0.1203009412,0.2758707404,-0.0123658553,-0.0990976468,-0.6099660993,0.2349269688,-0.1152764484,0.2859089077,-0.1456052214,-0.0266552493,0.0189348031,-0.2011065483,-0.1252482831,0.1103071794,-0.5258552432,0.3352452815,-0.1103226244,-0.1921404153,-0.2348798215,0.1402760744,0.2428161502,0.0209096316,-0.1805033088,-0.016431246,-0.1787301004,-0.242212832,-0.1114929095,0.0066054724,0.3349578977,-0.1443490833,-0.2299135327,0.0964632556,-0.1250468493,-0.3004514277,0.1181728318,-0.0709109455,-0.1080978066,-0.1321864873,-0.2444404364,-0.1845482588,0.0862775296,-0.1746641695,0.136753127,0.068800576,-0.0828412473,0.1016885862,0.1494758874,-0.2983340621,-0.1130416021,0.3407580853,-0.2677858472,-0.2979260683,0.4692345262,0.3519727588,-0.1586272866,-0.1613220274,0.1176549494,0.2354236245,-0.275716573,0.0287928637,0.162001878,0.3008447587,-0.0093471128,0.3109228313,0.305639714,0.1008466929,0.0073339115,-0.7207001448,-0.3249890506,0.1946189851,0.1271667629,0.2861980796,0.1761256456,0.1147448793,0.0217655767,-0.1419669837,-0.3108200431,0.0563896373,-0.1930020154,-0.2246221751,0.2044872493,-0.196063742,0.505836308,0.0961747617,0.2353154421,-0.1862322241,-0.2238080949,-0.2522972822,0.012597179,0.1250667423,-0.0072047054,-0.1599043012,0.076685071,-0.098935917,-0.1071699783,-0.1015126035,0.190955326,0.1383958757,0.0687803105,-0.0787461624,0.1001247391,0.0301096532,-0.1907965988,0.0285083577,-0.025133878,0.3009046018,0.2265717834,0.0618353374,0.0241579488,-0.0533063971,-0.0746178776,-0.112177223,-0.0414070375,-0.0779635236,0.3595576584,-0.3970214128,-0.0275725778,0.2810701132,0.498627305,0.3485302329,-0.3341702521,0.1404315233,0.0982502177,0.2008360773,-0.2418051809,-0.1320366263,0.1866642535,-0.0318389013,0.0309770051,0.1233275607,-0.0581408106,0.0505561866,-0.232844457,0.0567093082,0.5290329456,0.0924609601,0.2249322832,0.1308487058,-0.018567048,0.066922076,0.0650012344,0.0360546485,0.1229094267,0.2920594513,-0.3495812416,0.1462301612,-0.0567597263,0.0684003085,-0.1147470176,-0.6289654374,0.3336820304,0.1689161211,-0.3191784322,0.0878436938,-0.244912371,0.4515689611,-0.2186651677,0.0517134108,-0.2908123434,0.3248186111,0.0535502285,-0.0619030893,0.1166536137,-0.2613930106,-0.1222891808,-0.0241959225,0.0592864268,-0.3378441334,0.2547748387,0.1873698682,-0.1702088118,-0.2505170107,-0.0514734276,0.2003883421,-0.0278912894,-0.1070360094,0.3322894871,0.3139648736,-0.1075972989,-0.0052599772,0.2078180909,0.5430828929,0.3638230264,0.0598817654,0.2568224967,0.1145391539,-0.1374525726,0.0676150694,0.0322945118,0.2886669934,0.0836922228,0.3812452257,0.2170442194,-0.1802070588,0.1671673656,-0.019858133,0.1490117908,-0.2210004777,0.3357275426,-0.0727915093,0.0364464335,-0.3704914451,0.06097712,-0.627651453,-0.0764110237,0.4593777955,-0.0112017142,0.0094913254,-0.0697316974,0.0808876082,-0.1090028882,0.7130913734,0.3730811477,0.4081925154,-0.4481304586,-0.0922093242,-0.5804846883,0.0778250098,-0.0238746293,0.1688785702,0.0909519568,-0.1451891065,0.0226883758,0.2757076919,0.2652715743,0.0849523693,0.1511732638,0.0965434685,-0.1795458943,-0.1153760627,0.0229089558,0.0601159073,-0.0376164168,-0.2377488017,0.0560075715,-0.5345261097,-0.0055663222,-0.004136832,-0.0563894808,-0.0108250827,-0.0782608688,0.3397467136,0.0779523328,0.7522112131,0.0061275442,-0.1257435828,-0.3982895613,-0.3299315274,-0.1280596256,0.1461878717,-0.0317727588,0.2351618707,-0.1075725034,-0.3375989199,-0.3766066432,0.1866864413,0.049685251,-0.2255174518,-0.1647088081,0.0524485111,0.078807205,0.0078812353,0.1963621676,0.0920482352,-0.034826234,0.3125106692,-0.1061565951,-0.3599557877,0.67988199,-0.2259570062,-0.0474053249,-0.1042583883,0.036589101,0.0912002549,0.0880400538,-0.5794773102,0.1991289258,0.3769950271,0.113466166,-0.4397311509,0.2614184916,-0.0718317032,0.0115397554,0.0349884219,0.1389212608,0.0321028791,-0.2739159763,-0.0249953382,-0.1567310691]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/758","title":"Process 0 very slow when using num_procs with map to tokenizer","comments":"Hi ! Thanks for reporting.\r\nIs the distribution of text length of your data evenly distributed across your dataset ? I mean, could it be because the examples in the first part of your dataset are slower to process ?\r\nAlso could how many CPUs can you use for multiprocessing ?\r\n```python\r\nimport multiprocessing\r\nprint(multiprocessing.cpu_count())\r\n```\r\nWhich tokenizer are you using ?","body":"<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n","comment_length":62,"text":"Process 0 very slow when using num_procs with map to tokenizer\n<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n\nHi ! Thanks for reporting.\r\nIs the distribution of text length of your data evenly distributed across your dataset ? I mean, could it be because the examples in the first part of your dataset are slower to process ?\r\nAlso could how many CPUs can you use for multiprocessing ?\r\n```python\r\nimport multiprocessing\r\nprint(multiprocessing.cpu_count())\r\n```\r\nWhich tokenizer are you using ?","embeddings":[-0.4953111708,-0.5315650105,-0.1695286036,0.2302866876,0.1256165057,-0.3297326267,0.3093039989,0.160647437,-0.3321200609,0.2106941193,0.340248853,0.2400155813,-0.103669934,-0.0406741872,-0.559312582,0.036117129,-0.0464157276,-0.0402569175,0.5563800335,-0.000482796,-0.162746653,0.1379326582,-0.3582848608,-0.0826882422,-0.616525352,-0.2573291063,0.000495375,0.0575366244,-0.0524373204,-0.2949142456,-0.6529923677,-0.0069849179,0.047483597,0.351573348,-0.0001153084,-0.1335371137,0.0726515353,0.2212062627,0.1563179195,0.0101909926,0.3781885207,-0.4317011833,-0.083963111,-0.3196580112,0.1087613851,0.0827158689,0.1891488433,-0.0064333165,0.3377622962,0.3044478893,0.2040831149,0.216209501,-0.1863328516,0.253736347,-0.2040131688,0.2132370919,0.0725447983,-0.1697115302,0.2893604934,0.288334012,-0.0408059396,0.2808979452,-0.2223011553,-0.0609901622,0.072473824,-0.0260831136,0.3007549345,-0.6823173761,0.3525350988,0.1850749701,0.0784423128,0.1566152424,-0.1315235496,-0.308989346,-0.0853446573,-0.5059006214,-0.086782448,0.254968971,-0.2516716123,-0.1617123634,-0.4387652576,0.2242856473,0.0581689961,-0.2230465561,-0.2645806074,0.263395071,0.1062755138,0.2124119103,0.0047158902,-0.1841313541,-0.2799612284,-0.2286385596,0.2434158474,0.1527602822,-0.1565137357,0.0398825444,-0.1100483984,-0.2451602668,-0.1466781795,0.0842854306,-0.39053455,0.4203822315,0.0537959374,0.0610826612,-0.0953726321,-0.1121988297,0.1583889127,0.1098181531,0.1482037902,-0.213392809,-0.2921487689,-0.0550900474,-0.0397290103,-0.2578710616,0.2200279087,0.0208390355,-0.2356411666,-0.0229297392,-0.2340057343,-0.084954828,0.0402273871,0.0965623111,0.2033511698,0.2396940291,0.0495671667,0.5489522219,-0.2180420011,-0.0619762577,-0.4051899016,-0.0309106987,-0.1467278302,-0.1142647862,-0.1594428569,0.0469062254,-0.1378861368,0.1525100321,0.0012935771,-0.019640211,0.319987148,-0.1742588282,0.0151706701,-0.4309885204,0.0693969131,0.0789844915,-0.0225591045,0.27238217,-0.0778323933,-0.1469168812,-0.4247120023,0.3230879307,-0.4733678997,-0.1793971062,0.0462478399,0.1082266793,0.0358785316,0.1713533551,-0.0786886439,0.1096759737,0.3484973609,0.1327200383,-0.0248959344,-0.2907633483,-0.4852850437,-0.2281451076,0.0602697358,0.1922586113,0.0579522923,0.0291652083,0.4639880359,0.1940548718,0.6476488709,0.5497222543,0.1571917832,0.077386342,-0.0358739272,0.5512923598,0.1163762137,-0.1133762598,-0.4488638043,0.2650886774,-0.2791212201,-0.3341399133,0.2636803687,0.1984430403,0.3769236505,-0.0152228978,0.2889907658,0.2263309211,0.0917104706,0.4347145259,-0.334543705,-0.0326939858,0.1791485846,0.1259745359,-0.0471317619,-0.3097909391,-0.0765231922,-0.2440915406,0.1196052805,-0.1575352401,0.0428710058,0.2962051928,0.0219667312,-0.1225077063,-0.0653337166,-0.0063165948,0.2425462604,0.2852153182,0.0202555489,0.3112221956,-0.1868072152,-0.0310237799,0.2466645241,0.1015818045,-0.1156051531,-0.1957660317,0.0660164505,-0.171933651,-0.0385443978,0.1197176725,0.1025743037,0.2863748968,-0.082142286,0.0805877373,0.2190673947,-0.032503657,-0.0987994447,-0.3095045388,-0.1771299094,0.0805088133,0.3629036248,0.0737580806,-0.0865936354,0.2838687897,0.3752439916,-0.0114852265,-0.3486469388,0.0197771061,0.4422858953,0.0011342396,0.3061334789,-0.1134737581,-0.1244747043,-0.0252764169,0.2055552304,0.437979579,0.1963785589,0.3408122361,0.2434750944,0.032631427,0.0240829214,0.174734652,0.250009656,0.0186405331,0.1492000222,0.4589809775,0.2119396329,0.1745252311,-0.1839994192,0.3083343208,0.434673667,-0.0011619458,-0.1568811834,0.3453220427,-0.132697612,0.0322504267,0.2280595452,-0.5205900073,0.1883367747,0.3040079176,0.0177533459,-0.0917438492,-0.1519123614,-0.0713321418,0.0493676364,0.1606242508,0.029036019,0.0950754955,0.2052949667,0.1799435467,-0.2392560989,-0.3200096786,0.0207468458,0.2429746389,-0.0013515828,0.0810669959,-0.1852992773,0.3129558265,-0.0443003066,0.0234781951,-0.2518494129,-0.1960191429,0.1010063738,0.1181092411,0.1571971774,-0.0073515391,0.0690242723,-0.0020889777,0.2708703578,0.046013914,-0.0574385971,0.0569550768,-0.245106414,-0.0337209851,0.372238487,0.0790437981,0.3720944822,-0.0797705799,-0.0757287741,-0.1691731662,-0.3268736303,0.0174459703,0.0113238823,-0.101103358,-0.0361476429,0.0515645146,-0.1560918093,-0.0931717604,0.2367040962,-0.3002384007,-0.2327294648,0.0292148218,-0.0520456508,-0.424818784,0.0050974297,0.2082571983,-0.4068971276,-0.0667257234,0.1968749911,0.305744797,0.4399700165,-0.1341744512,0.1866541207,0.0175136831,0.2654106617,0.0374066494,-0.0729926899,-0.0881271735,0.2479168475,-0.0408946835,-0.1777158976,-0.2335484624,0.0500518605,-0.0702619031,0.2776545286,-0.0354017057,-0.1698897183,-0.3668973446,0.4108220935,-0.2165420055,0.0544036925,0.40763098,0.1833177954,-0.1927209944,0.2562755346,0.0459657609,-0.0532862842,-0.2195073962,0.0115286326,-0.0571604408,0.2237249315,0.1973442137,0.5768964887,0.2344726324,-0.5633326173,0.200248912,-0.2848826051,-0.2870135009,0.0133943772,-0.0590872839,0.447293967,0.1093296334,0.0086578904,0.2384128869,-0.0138358278,-0.2939450443,0.0700018853,0.0657205507,-0.270067811,-0.1444712579,0.0597491264,0.0912714452,0.2420790642,-0.1904567033,0.0679848269,-0.3495050669,-0.193739295,-0.214173764,-0.0551056825,-0.1080439314,-0.4080316722,-0.5878000855,-0.0678426623,-0.7291850448,0.5164377689,0.3962760866,0.1549691558,0.1175733432,-0.0569428876,0.1957085133,-0.3049109578,0.6568159461,-0.0021272311,-0.1984634697,0.0401748307,-0.3029310703,-0.3522105813,0.0171920508,-0.0218962487,0.3239723444,0.3868622184,0.8028352857,-0.0870434046,-0.2673941851,-0.1531684101,-0.1316916794,0.0950850844,-0.1812621653,-0.2314995527,-0.2156763822,-0.3307567239,0.3650007844,0.0735835359,0.4053303599,-0.2350696027,-0.3139365017,-0.1081780046,-0.1134598255,0.3099853396,0.0232444331,0.2032972127,-0.1580931991,0.0333793759,0.0740530118,0.179284215,0.1955321282,-0.1693627238,0.1570598781,-0.1679613143,0.041220095,0.0416859761,0.1953939646,0.4357809424,0.0437920354,0.1081308126,-0.3416278064,0.7598041892,-0.0696404055,0.3424262702,0.1069184542,0.3957113326,-0.35973683,0.1138608009,-0.0476059653,0.3227454126,0.0472927541,0.0249578301,-0.5631777048,-0.0555258468,-0.0977513418,0.4544780254,1.0650622845,-0.2870717645,0.1448943168,-0.2320344895,0.0827210695,0.177181229,-0.4226152301,0.1026196107,-0.5460086465,-0.1650374085,0.0591854788,-0.1342520565,0.3077512383,0.1799541563,0.0870343074,0.1531486064,0.222472772,0.1601034552,-0.1453618556,0.4769963324,-0.0559897386,-0.3710184395,0.1172494292,0.138214767,0.0543054938,-0.1033849642,0.1092977077,0.024671793,-0.1572262198,-0.1643770337,-0.206037268,-0.1709191948,-0.410236001,0.3210148215,0.2067409009,0.0411215685,-0.027972322,0.3484246433,0.1181503385,0.2764460444,0.0397459306,0.2110109329,0.1361618936,0.0295989458,0.209814623,-0.3426702917,0.0637550727,-0.1972798705,-0.1635498106,-0.0958657861,0.05441406,0.0214747973,-0.2595753372,-0.0013827095,0.4683527052,0.1173835397,-0.0985035673,-0.026353009,-0.3687486351,-0.1589290947,0.0924814194,-0.0453759506,-0.2404420525,0.4379301369,-0.0298691858,-0.2830432951,0.025305694,0.2725923955,-0.0286679845,-0.1106834114,0.2649385333,-0.1724040508,-0.3740751445,-0.1781917512,0.3151311576,-0.296015203,0.0108429929,-0.0633540601,0.0155357141,-0.2265297174,0.0385525078,0.079021357,-0.1980660558,0.011885508,-0.047774449,-0.2924386859,-0.5503083467,-0.2980762124,-0.2731850445,0.0654857531,-0.0155580444,0.1982997954,-0.245895341,0.0351966992,-0.2807807028,0.0488861911,-0.246790424,0.2797229588,-0.168992877,-0.3845107257,-0.3008621633,0.0526493229,0.2130944431,0.5159611106,-0.1575538218,-0.2491378635,-0.0155146541,0.1241656765,0.1078753397,-0.3995264173,0.1141860709,0.1762262881,-0.2904109061,-0.2819703519,0.0141654871,0.0386116877,-0.1537787318,-0.2576296926,0.3884235919,0.216482237,0.0658447072,0.2312609851,-0.1386731565,0.310728997,0.1447477043,0.2480475307,0.1469513476,0.0634007752,0.0029561762,0.0393638127,0.5078852773,-0.1104288399,0.0072403196,-0.115770027,-0.1256223619,0.1650376916,0.0608487055,0.2264772207,-0.1681413203,-0.2976419628,0.3968755603,0.242995888,-0.194947511,-0.1240112036,0.3900765777,-0.17881459,0.0872255489,0.1557323188,0.2607519031,0.1712001413,-0.2070616633,0.0470718034,0.2873012424,-0.4762457311,-0.3216883242,-0.11560072,0.0869914293,0.0641735345,0.4179252386,0.0104686813,0.0260185841,0.3452897966,-0.0258643441,0.5836871862,-0.017006116,0.3134092391,-0.1671292335,-0.0566670485,-0.1695800424,0.2938793302,-0.0251538381,0.1789567918,-0.1078280509,0.176477775,-0.1166059226,-0.2088208497,0.028081568,0.5991235375,-0.3017578125,-0.244185105,-0.1127300709,-0.0059687314,-0.2675645649,0.0405530594,0.0563713312,0.0178295467,0.1266622394,0.3407023251,0.0284128487,-0.1761397868,-0.1805212796,-0.0431923866,0.279088974,-0.2988127768,0.1856257468,0.0152856484,0.286272943,0.0928542838,-0.1117143109,0.550331533,0.4608724117,-0.4589522779,-0.0628087446,-0.1084597781,0.1165972203,0.0476760045,0.2200512588,0.0298723206,0.2477375865,0.1128172651,0.0021529123,-0.2931623757,-0.3562104106,-0.0105454084,-0.0149187855,0.1450190097,0.3033882976,-0.0318259485,-0.3631107807,-0.1850340515,0.0589166731,-0.3338305354,0.0291337557,0.1870931685,-0.3277789056,0.1497628987,0.0080980351,0.0680177212,-0.1722186059,0.2430778444,0.1725537479,0.300183326,-0.217551887,-0.1257348955,-0.5043819547,0.0843531862,0.0822476447,-0.3333667815,-0.0831627473,0.121652931,0.1011618525,0.1263032854,0.0969700664,0.2900788486,0.0721934214,0.4266233146,-0.4099590778,0.6290709376,-0.1745357662,-0.1619635522,0.1438540369,-0.4980400503,0.4280370772,-0.0921039209,0.1016476527,-0.1212154254,0.252484858,-0.2613580823,0.2489531338,0.3376577497,0.1373838931,0.3791015148,-0.0641289353,0.0380618982,0.0848203748,0.4413324296,-0.2321867645,0.082650423,0.0833215043,0.1831194609,0.0546411946,0.1247580871,-0.1352440268,0.1376781911,0.0462845862,-0.4186735153,-0.1336521655,-0.0341651812,-0.0455417335,0.2542337179,-0.2590353191,0.2475591749,0.1833343208,-0.0444634072,-0.2448836118,-0.232711181,0.3662111461,-0.3689890206,-0.2968351245,-0.3429130614,0.1519065499,0.1632058024,0.3504226506,0.0330881625,-0.1743750125,0.3958963454,-0.1429535896,-0.507832706,0.3442977369,-0.0517102666,-0.3934430182,-0.1015944406,0.0064044832,-0.2152473629,0.0475870967,-0.1189566627,-0.2414662838]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/758","title":"Process 0 very slow when using num_procs with map to tokenizer","comments":"Using pre trained HF tokenizer. The result is the same with tokenizer multiprocessing off and on.\r\nI have (absolutely) no idea about the distribution, but since this issue occurs on all of my datasets(regardless of files), I don't think distribution is the problems.\r\n\r\nI can use up to 16 cores.","body":"<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n","comment_length":50,"text":"Process 0 very slow when using num_procs with map to tokenizer\n<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n\nUsing pre trained HF tokenizer. The result is the same with tokenizer multiprocessing off and on.\r\nI have (absolutely) no idea about the distribution, but since this issue occurs on all of my datasets(regardless of files), I don't think distribution is the problems.\r\n\r\nI can use up to 16 cores.","embeddings":[-0.4506952465,-0.3895408213,-0.0472824574,0.1629320234,0.1149342209,-0.2815493047,0.2656876743,0.2587329149,-0.3923633397,0.1888355911,0.0629082173,0.3718959093,-0.089070864,-0.2105110139,-0.5482542515,0.2360698283,0.0839328393,-0.0310935583,0.5445747375,-0.1181110516,-0.1754408777,0.177372694,-0.2817248106,0.0584396273,-0.5508942604,-0.1075035185,0.0991637409,0.2062431872,-0.0079598241,-0.3365160823,-0.6048396826,-0.0062392484,-0.0067063249,0.2380303293,-0.0001208894,-0.0261882041,-0.0501437411,0.3045714498,0.1773629338,0.1757730395,0.5087664127,-0.2677305043,-0.2065653354,-0.0814368874,0.071761854,-0.1732417494,0.1646704227,0.1108722761,0.3598265648,0.3113973737,0.1259145588,0.3464170992,-0.2695309818,0.1674762368,-0.3548649848,0.1909781843,0.0127337407,-0.1612080932,0.1026980132,0.1404281855,-0.0731380284,0.1114659607,-0.206951499,0.0112679061,0.1181672961,-0.0765307769,0.3340560496,-0.5317914486,0.3109235168,0.1555503756,-0.1304652095,0.2259662002,-0.0447759517,-0.2271244377,-0.1173180416,-0.4069034457,-0.0659299865,0.1536652893,-0.2233647555,-0.2492575049,-0.2870896161,0.2072083056,0.1220583245,-0.1853976846,-0.1608427912,0.1784252673,0.2193142921,0.1878379583,0.0917399302,-0.0794800594,-0.3464467525,-0.189996779,0.0932571292,0.1641352475,-0.2648397088,0.0863497406,0.0162886754,-0.2284547687,-0.1790191531,-0.0498304591,-0.4908621907,0.5438275933,-0.0379492939,0.1100388244,-0.0540822633,-0.0507318787,0.0711289942,0.107380487,0.3145228326,-0.0370388553,-0.2704118788,0.0606514923,-0.1888573617,-0.2157375216,0.1592000425,-0.1182655767,-0.4101752043,-0.0712799132,-0.3443175852,-0.0421151891,0.06168871,0.2465012819,0.1357420832,0.2784369588,0.1164710522,0.60796386,-0.2985524535,-0.1146471873,-0.4788651466,0.0308691394,-0.065696761,-0.0983158946,-0.1533274055,0.1412196904,-0.0580749735,0.0928089991,-0.0424256958,0.1374468058,0.2228023261,-0.0566168576,0.1432552338,-0.3872431517,0.1344203055,0.1969755739,-0.0905412212,0.3274353147,0.0075174361,-0.1662386656,-0.3650250435,0.1921859384,-0.5542647243,-0.1544521153,-0.1172215492,0.0300147496,-0.0256938469,0.3321104944,-0.1265891045,0.065926671,0.0601636171,-0.0067125461,-0.1519165933,-0.4198389351,-0.5100108385,-0.1309079677,0.0354381427,0.2284735143,0.0076896199,-0.0061620814,0.4147745073,0.1895893365,0.6329798698,0.640367806,0.0365301333,0.0456745885,-0.2062079906,0.4334468842,0.0078639546,-0.1366570741,-0.3977273107,0.2765939534,-0.2492134571,-0.2269993275,0.1242460683,0.083922565,0.2745398581,-0.0625268966,0.1646543741,0.0608774461,0.1374640614,0.3524214029,-0.3490796983,0.0038467364,-0.0185448434,0.0742785409,-0.0408974811,-0.1899458319,-0.0398687422,-0.2290841043,0.0277229939,-0.1095842272,0.0088469079,0.2146681696,0.1113792136,-0.1869167089,0.0424843207,0.1480467767,0.2814905941,0.3698045313,-0.2151997536,0.2599093616,-0.0825202018,-0.0297249015,0.2806521356,0.1298107505,-0.1157191023,-0.2477763742,-0.003191859,-0.1168284938,-0.2420949936,0.0901817456,0.0670359358,0.3922367692,0.1027478278,-0.0189812109,0.2864282429,-0.0789332911,-0.1034206524,-0.2746038735,-0.2179026753,-0.0040626163,0.3598762453,0.020688016,-0.0209823065,0.1538755745,0.297991693,-0.102121003,-0.2200790495,0.1073790118,0.3305101693,0.0649032518,0.1737664342,-0.132722348,-0.1296590418,-0.1472826004,0.2585277259,0.4864607751,0.233011663,0.2920912504,0.1635444164,-0.0365977846,0.0878182575,0.1875586957,0.4652343392,0.1040159687,0.1991215497,0.6450462341,0.1737261266,0.2268470228,-0.10093683,0.4220349193,0.3750983179,0.0383585021,-0.2353627384,0.3845766485,-0.1413104832,-0.0957228616,0.3067465425,-0.5021467209,0.2544321418,0.0890975595,-0.0221233293,-0.0148402629,-0.24618119,-0.0257795248,-0.0438657589,0.1368954033,0.1569816917,0.1807011962,0.1568026394,0.217647776,-0.3013871908,-0.320089221,0.1313546896,0.2890082002,-0.1799802184,0.0986567512,-0.2343864441,0.5450378656,0.0290588476,0.1898016632,-0.2035871744,-0.3212849796,0.1369773895,0.3257724047,0.2375264913,0.0229058005,-0.0110994456,0.0008424884,0.0958475396,0.012094615,-0.1002325043,0.0772534236,-0.4060472548,-0.1035230905,0.4133099914,-0.0200189874,0.1561088413,0.0215083845,-0.1393605322,-0.1392398626,-0.385617286,0.0294844713,0.0673709363,0.0429110453,-0.0571673959,-0.0194588322,-0.2499811947,-0.0217318404,0.2867519557,-0.3686539829,-0.1650758386,0.1257063746,-0.2890628278,-0.3887500465,-0.0040579909,0.3380643427,-0.2690812051,0.0169429854,0.076980032,0.0362953693,0.3447990119,-0.0609910861,-0.0147451404,0.0760455728,0.217759192,-0.1731932014,-0.04938481,-0.1005866528,0.2364498824,-0.0065876739,-0.1847718954,-0.1687864661,0.0533468314,-0.059252739,0.1908752024,-0.0932112858,-0.0675301999,-0.279248327,0.5511071086,-0.2200596333,-0.1358428746,0.3535012901,0.2189850658,-0.2078604996,0.1770895869,-0.0537383817,0.1026754677,-0.2086285204,-0.0796081871,-0.022159813,0.1777781993,0.2634583414,0.656558454,0.1694375575,-0.5546483397,0.4604192078,-0.1575149447,-0.1610368192,0.0329057723,-0.0412072316,0.3707009554,0.1070766002,0.0640258417,0.2123041451,0.0450257696,-0.3682759106,0.1785327941,0.1515749097,-0.083699666,-0.1060364544,0.1239336506,0.0961198807,0.1262589097,-0.1447354704,0.0562665761,-0.3013880551,-0.2197107375,-0.035473384,0.0004954087,0.0039479868,-0.421312511,-0.5623945594,-0.0808738768,-0.6648821831,0.3926235735,0.2515885532,0.3031941056,0.0463701375,0.045965936,0.2508892119,-0.3131645918,0.7168605924,0.0974350423,-0.0989008099,0.1996324658,-0.3248124421,-0.4253256023,-0.0893369615,0.108610861,0.3160227239,0.2753644586,0.9014762044,-0.0413431413,-0.4147046804,-0.2874907255,-0.1901812106,0.0496883653,-0.1471251845,-0.1574978232,-0.246357128,-0.1956475973,0.2643303573,0.0114489077,0.2582146227,-0.2203469723,-0.246658057,-0.0617920905,-0.0987265781,0.3484372199,-0.017023379,0.1559100747,0.0167181958,0.1187133193,0.1637178957,0.1730431169,0.238593623,-0.2919011116,0.2931388617,-0.1807395369,0.1343401819,0.1393645108,0.0168271568,0.4543557465,0.0459164903,0.1290720552,-0.4070775211,0.6302614212,0.0490674973,0.3861462176,0.1993939579,0.3371492922,-0.3774442971,0.3004015684,-0.0296039265,0.2843730152,-0.0492784753,0.1018453389,-0.4722979367,0.1241325215,0.0038173932,0.5101215839,0.9462080002,-0.3860379457,0.1506647468,-0.2742774785,-0.0172018632,0.1639907956,-0.4686093926,0.1010441557,-0.4557462633,-0.3033402562,0.0086000254,-0.0756167769,0.3385697305,0.1963633746,0.2635620236,0.1289667338,0.1450027525,0.154741928,-0.0322183184,0.3646079898,0.2569506168,-0.361651361,0.2099115103,0.0865081698,0.1023159251,-0.0712675899,0.1038280576,0.0186189134,-0.1536647528,0.0483248495,-0.1398189366,-0.0889380723,-0.3795030415,0.3293952048,0.2940430045,0.2540568709,-0.0129706496,0.2836950123,0.1130837575,0.0257879402,-0.0259230528,0.2413244396,0.2427598834,0.0285983756,0.2847898304,-0.4256589711,0.0792293176,-0.2082633078,-0.0765728876,-0.1313895434,-0.0191263761,-0.0395891815,-0.2956015468,0.048102323,0.483363986,0.3967644274,0.0377289429,-0.0651781112,-0.2590231001,-0.1406303644,0.0191147253,-0.07555978,-0.1697646081,0.3989981115,-0.1596686095,-0.2789801955,0.0549476072,0.3190858364,-0.1758250296,-0.0476929024,0.2904019654,-0.2413253933,-0.2966598272,-0.0782152042,0.2210619003,-0.1975937039,0.0224780794,-0.085141778,0.0315319933,-0.3125318587,0.1756534576,-0.0263852086,-0.1682738215,-0.1190613061,-0.1653357148,-0.171416074,-0.5735752583,-0.475664258,-0.4286080003,-0.0276448149,0.0256061871,0.0915373266,-0.2318774909,0.0014342181,-0.2034559995,-0.0040846285,-0.2192680836,0.2226076126,-0.3501394987,-0.2844921052,-0.178934142,-0.0394424424,0.1347631365,0.438252002,-0.1185480431,-0.1647270769,-0.1132146865,0.1871752441,0.0996448994,-0.4745827019,0.1656764448,0.2503249645,-0.4242202938,-0.2861304283,0.043825753,0.0710114166,-0.2203971595,-0.2785336375,0.48030141,0.0463778339,0.1537204981,0.2163337767,0.037379548,0.3629663587,0.1119824424,0.1945257485,0.0304203611,-0.057316225,-0.0880845115,0.0203557741,0.3967165351,-0.0861904472,0.0864861161,-0.1419195384,-0.1733751297,0.0906084999,0.066642575,0.1409444362,-0.0879263654,-0.4116289318,0.5136474371,0.0784526542,-0.1498871148,-0.2812044621,0.2699749172,-0.096310325,-0.032620132,0.024041906,0.421164304,0.0776925907,-0.0321288183,0.0739575997,0.2720718682,-0.5317792296,-0.2952333093,-0.1408230811,0.0701704249,0.1684836447,0.4127981961,-0.144829005,-0.0542106107,0.2911914885,0.0599383302,0.5728496313,0.1703593284,0.4189440906,-0.120441474,-0.2007472813,-0.2446850836,0.3875007629,-0.021161003,0.2297845334,-0.1253006458,0.1403176486,-0.1662497818,-0.4690258503,-0.0589975752,0.6085990071,-0.3707394302,-0.2414241135,-0.0472479649,0.0307616312,-0.1974385232,0.1610166579,-0.0112268981,0.0607002228,0.170657441,0.3910303414,-0.1183738336,-0.0886759385,-0.3024642467,0.008697453,0.3411857784,-0.1234590337,0.1405360252,-0.1475322247,0.124508813,0.0321461484,-0.1595974714,0.6055913568,0.5521280169,-0.3411705792,-0.1187715232,-0.1353086084,0.1968234777,0.0527879484,0.1682761759,-0.1112035066,0.1252186596,-0.0026613849,-0.1213411242,-0.224385649,-0.4974940419,-0.0765586346,0.1628838331,0.2405463755,0.304907769,-0.049275957,-0.231068626,-0.0280052442,0.1674408317,0.009861269,0.069675833,0.2245883495,-0.2460327148,0.1521101743,0.0025108496,0.0438818038,-0.1660663933,0.2478758246,0.0387677252,0.1136815995,-0.2455688864,-0.0506364405,-0.5876187086,0.1819279045,0.1153169274,-0.1236989498,0.0707050487,0.0245338827,0.019817777,-0.087288022,0.1267776042,0.2281825989,0.2938823104,0.4610562623,-0.418581754,0.6147739291,-0.2220789939,-0.1689703166,0.0884751454,-0.4843531847,0.5632546544,-0.0318483785,0.0487949997,0.0099400952,0.1435698867,-0.2666410506,0.3695188463,0.4443905652,0.1392315924,0.3444249332,-0.0397045799,0.0122154225,0.0898186639,0.5262340903,-0.2202477902,-0.1035792083,-0.0771511272,0.1054722816,0.0977565423,0.144134596,-0.1994735897,-0.0298770852,0.1257633865,-0.3371966183,-0.252749145,0.0993980318,-0.0942586586,0.3721275926,-0.1503159851,0.3240285516,0.2423544377,-0.0549165979,-0.3557980657,-0.1633712649,0.3820388317,-0.4133448303,-0.1988784522,-0.3727411926,0.2103743702,-0.0255500935,0.266774863,0.1961726993,-0.2702803016,0.2786149383,-0.0668820292,-0.5310514569,0.289827615,-0.2048015594,-0.2587738931,-0.1606618464,-0.0317392163,-0.1840328574,-0.0887728333,-0.0115601923,-0.0237787087]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/758","title":"Process 0 very slow when using num_procs with map to tokenizer","comments":"Ok weird, I don't manage to reproduce this issue on my side.\r\nDoes it happen even with `num_proc=2` for example ?\r\nAlso could you provide more details about your OS and the versions of tokenizers\/datasets\/multiprocess that you're using ?","body":"<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n","comment_length":39,"text":"Process 0 very slow when using num_procs with map to tokenizer\n<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n\nOk weird, I don't manage to reproduce this issue on my side.\r\nDoes it happen even with `num_proc=2` for example ?\r\nAlso could you provide more details about your OS and the versions of tokenizers\/datasets\/multiprocess that you're using ?","embeddings":[-0.4499242008,-0.3384662867,-0.0881828442,0.1553920507,0.0886590257,-0.303383559,0.2761160433,0.248826772,-0.4145074189,0.1561545581,0.2132177651,0.3615123332,-0.2245154679,-0.1062326878,-0.5425727963,0.2070079744,0.0648821518,0.0284322016,0.497579217,0.0339274369,-0.2033851445,0.1207496226,-0.373627156,0.0247500073,-0.579837501,-0.1454626024,0.1142511964,0.1221130416,-0.1154534519,-0.3299306929,-0.6470974088,-0.0766082034,-0.0318846926,0.3341849744,-0.0001155862,-0.0463386886,0.1090576127,0.3409170806,0.059888225,0.0577889681,0.3721919358,-0.356223613,-0.1680819839,-0.1338618249,0.065941833,-0.1426615119,0.1589431912,0.0322805718,0.3406347036,0.4026558697,0.1911497116,0.3179595768,-0.270878762,0.2543511689,-0.2134602964,0.1684640348,-0.0395519584,-0.2626494169,0.1379910856,0.2332983911,0.0755674392,0.171354115,-0.2536707819,-0.1796937585,0.0796079189,-0.0737387091,0.3969673514,-0.5955144763,0.3338231742,0.0688479915,-0.0498033278,0.2084262371,-0.1226272732,-0.2446288168,-0.1194128171,-0.3561387062,-0.0315729678,0.1177541912,-0.145931378,-0.1997264475,-0.4144126475,0.1733220965,0.0927784368,-0.0780107528,-0.2313115299,0.1342223734,0.1872867495,0.1875690073,0.0374248996,-0.2165006995,-0.2947055697,-0.2338628024,0.046458371,0.2533203065,-0.2309342772,0.0660237148,0.0152793219,-0.130366236,-0.2016302198,0.0667562559,-0.4654318392,0.4875847101,0.0327738598,0.1434691548,-0.1114958152,0.028573351,0.170927912,0.1275348812,0.3001178801,-0.2543088794,-0.2343515307,0.0001800376,-0.1543804854,-0.1821833402,0.2154282629,-0.0612580925,-0.2682190239,-0.0233409647,-0.3308350146,-0.0115478653,0.0873355493,0.2619652152,0.2996673286,0.2002889067,0.06225859,0.5214753747,-0.2556758523,-0.1215853095,-0.4594755769,0.1133482605,-0.0815635771,-0.1647175997,-0.1865520477,0.1378510892,-0.0613127537,0.177166447,-0.0054514324,0.0253582746,0.1104048416,0.0252898894,0.1823234558,-0.3223147392,0.1147368103,0.2022774369,-0.067110002,0.3035331666,-0.1029452011,-0.1617824137,-0.3697654605,0.2381415516,-0.5384536982,-0.1524564475,-0.1589490175,0.0934766233,-0.1279311031,0.2658382952,-0.2042573839,0.0647621751,0.0831181258,-0.0184914488,-0.1679270715,-0.3296148479,-0.5520940423,-0.1329564005,-0.0274989866,0.1706850529,0.0621937141,-0.0064918748,0.4061864018,0.1375060827,0.5677406788,0.476406157,0.080736883,-0.0067321165,-0.1896296293,0.4975835383,0.1461083293,-0.0847519487,-0.3920003176,0.2247670889,-0.343560338,-0.2591389716,0.1443513781,0.1934028417,0.2961143255,-0.0813638046,0.1891617924,0.1684509665,0.1706659049,0.4236152172,-0.3447260559,0.1150199845,0.0540287793,0.0510680936,-0.1712332666,-0.2077021897,-0.1022623032,-0.2073799968,-0.075518772,-0.1466234326,0.0673182905,0.2143203318,0.0944850072,-0.1933384538,0.0748752281,0.0299660843,0.1912508309,0.3252852559,-0.101033181,0.1861101836,-0.133623302,-0.0344640464,0.259673357,0.1267644167,-0.1248527467,-0.146664843,0.0586642399,-0.1872172356,-0.2034932226,0.1582451314,0.1260849386,0.3644276857,0.0801934525,-0.0583211333,0.2455107272,-0.0991178155,-0.1141911075,-0.2937938869,-0.1932035387,0.0026086592,0.3814733028,0.0125687579,-0.0252914242,0.2470211834,0.381334722,0.0205002334,-0.2337235808,0.1055138111,0.4118440449,0.0603031777,0.2026038766,-0.0950057581,-0.0862973481,-0.1775080264,0.2709073424,0.5127896667,0.2094573975,0.2903516293,0.1432604343,0.0554517657,0.0688226894,0.1647234708,0.242582202,0.0216359366,0.1735888273,0.6222238541,0.2241018862,0.2362848967,-0.0152292065,0.3324324191,0.3139469028,0.0017893401,-0.1696559787,0.3111461997,-0.184545666,-0.0576296039,0.2842711508,-0.4943203628,0.2353970706,0.2080538273,-0.0235128459,-0.0208482351,-0.199622348,-0.0781733841,0.0861452892,0.1754349172,0.1429576427,0.0968637988,0.2540300488,0.1578579992,-0.2692854404,-0.2531763315,0.0884561986,0.3354820609,-0.1019076928,0.1001666933,-0.168754667,0.401204437,0.0715172142,0.0944321156,-0.1741177738,-0.317132175,0.1304926872,0.2831198573,0.1585765928,0.1246608719,-0.0297836568,0.0023998972,0.19020091,0.0668637529,-0.082956627,0.0202425588,-0.3556158543,-0.0494644456,0.4389982224,-0.0162160955,0.2940421999,-0.009238421,0.0038734111,-0.2317297608,-0.455683887,0.0031739497,0.041392073,0.025513377,-0.0680909976,-0.062108703,-0.2358986437,0.0194190461,0.248058036,-0.3583425283,-0.2180262059,0.1097904369,-0.2262161374,-0.3725872338,-0.0686939433,0.3364560902,-0.2102839351,0.075750798,0.1448751837,0.0660199672,0.3742148876,-0.0085043246,0.1516018957,0.1291854978,0.2616471052,-0.1156107262,-0.0668545216,-0.0575702265,0.2586344779,-0.0262191463,-0.1741725057,-0.1098842695,0.0847538635,-0.111498557,0.2365138978,-0.0075679761,-0.205554381,-0.3503775299,0.4797764122,-0.1803678423,-0.156106472,0.3318628669,0.2220036238,-0.2312347293,0.1933972389,-0.0816725567,-0.053440515,-0.2416557372,-0.0583512411,-0.015322376,0.1631850004,0.2545320094,0.5217760205,0.1270966381,-0.5550823808,0.2751224041,-0.241236642,-0.1587435007,0.0387795419,-0.1154578179,0.3504100144,0.2130519897,0.0414600782,0.171665743,0.089203544,-0.4674322605,0.0982109681,0.0906485394,-0.1467721015,-0.1113884971,0.0176670197,-0.0280592721,0.0948858708,-0.1602466404,0.0541890301,-0.3582014441,-0.180077374,-0.1543655246,-0.0298825931,0.0212050565,-0.4966046512,-0.6432381868,-0.0315760672,-0.7037484646,0.4059981704,0.2911790907,0.2319975346,0.1531505734,-0.0340889096,0.2259286642,-0.2468258291,0.7638767362,0.0278597586,-0.1418037266,0.1673061699,-0.2136551589,-0.5546609759,-0.038373217,0.1784723103,0.392282784,0.3510373235,0.9344829321,0.0170381851,-0.3225594163,-0.290311873,-0.1401003599,0.0864271522,-0.1478821784,-0.2009090632,-0.1770835668,-0.1991243064,0.2782624662,-0.0200461056,0.2604209483,-0.1195328832,-0.3668772876,-0.1575031877,-0.1159372106,0.2841863334,-0.0570019782,0.2025188804,-0.0913921222,0.1810497344,0.1577021927,0.2029871345,0.202324599,-0.298358947,0.2804027796,-0.1743219942,0.1097655445,0.1640728563,0.0676613897,0.3864969611,0.0313137621,0.0984006748,-0.477943778,0.5437563658,0.0703656077,0.4488608241,0.1909615248,0.319955945,-0.3200171292,0.2705405056,-0.011883012,0.3681922257,-0.0130440928,0.0603403114,-0.4594018161,0.0020533013,-0.0132884299,0.4890936911,0.9400268793,-0.3060779274,0.2133759707,-0.1931506097,-0.0184859186,0.1113680229,-0.2814562917,0.0406491496,-0.4389478862,-0.1258583069,0.0415324531,-0.0911841989,0.2763584554,0.0752055496,0.152997002,0.0236573033,0.1620554328,0.2183236331,-0.0904111117,0.3808493316,0.1806976348,-0.3440164626,0.1924465001,0.149454847,0.17389144,-0.1192565337,0.1402198672,0.0078363754,-0.1601858288,0.0186570846,-0.1336497217,-0.1067335159,-0.341956228,0.3004209101,0.2267999053,0.1218818501,-0.160355106,0.3938630819,0.0186712649,0.1552830935,-0.0950716212,0.2904054821,0.1818507612,0.0560357161,0.1889547706,-0.4750975966,-0.0106146876,-0.203719303,0.0104182865,-0.0543441959,0.0256799012,0.0624063797,-0.2789143324,0.0811029449,0.4807518125,0.3518825173,-0.0614795014,-0.0926863253,-0.2849007249,-0.1550027579,0.0923083946,-0.0917548686,-0.1146125644,0.3365730941,-0.099429436,-0.3127016127,0.0610965155,0.3479284644,-0.1225402281,-0.1190595031,0.3216855526,-0.2689369321,-0.2803430259,-0.1252522767,0.2756992579,-0.2910202146,0.0451862551,-0.0733116269,0.1523389518,-0.2231270671,0.1914171427,-0.0689254701,-0.224012658,-0.0658596009,-0.1013902724,-0.113081038,-0.5481932759,-0.4265356958,-0.399356693,0.0397684127,-0.0814379975,0.0588096827,-0.2969693542,-0.1574765891,-0.2581764162,0.0441185944,-0.2127568573,0.2459952533,-0.221379891,-0.2583941817,-0.2013216615,-0.0869795382,0.1961783767,0.358694911,-0.1027748138,-0.225650385,-0.0482552238,0.1419333071,0.1489546299,-0.5315022469,0.1496793628,0.2843876183,-0.4086700976,-0.3168076575,0.0190238282,0.1879047602,-0.2728007138,-0.3082474768,0.3944231272,0.100109078,0.2289350629,0.1660902947,-0.0942246914,0.3499560654,0.0723193958,0.2149965167,0.0402824096,-0.0236298013,0.0783562735,0.0154484734,0.3958331943,-0.0450424403,0.0588575341,-0.1191261858,-0.1441314518,0.1718428284,-0.0324360393,0.1243379042,-0.1295388788,-0.3658833206,0.4708152115,0.1685203165,-0.0764752254,-0.214103058,0.3278603554,-0.1524691433,0.0193353668,0.0183176007,0.288877517,0.1424130052,-0.0052008708,0.0909814239,0.2107021809,-0.4735336304,-0.2761483192,-0.1802954823,0.0596443266,0.0376940258,0.4591518342,-0.1018936485,-0.0324468687,0.365596801,-0.0302576944,0.5474577546,0.1747730374,0.4053159356,-0.1785577238,-0.0343229584,-0.2934327722,0.3591295481,-0.0543820858,0.2230077684,-0.0498660877,0.1106936634,-0.1132999659,-0.3018490672,-0.0353639163,0.6377309561,-0.3508571684,-0.2562386394,-0.1169731617,0.0019785929,-0.1625101715,0.1244661137,-0.0368133634,0.071831286,0.0127833085,0.3527554572,-0.1232694834,-0.0989833623,-0.2599740624,0.0705935881,0.2637036145,-0.1159006506,0.1439370513,-0.2477955818,0.1988353878,0.0508370101,-0.0602228865,0.6009982824,0.6137855053,-0.4024623334,-0.0945194513,-0.057326261,0.065483734,0.0428201221,0.2975777388,-0.1064927503,0.044569131,0.0193122886,-0.0360825621,-0.2808651626,-0.4103232026,-0.153126806,-0.0051798979,0.215658322,0.2116217762,-0.1012041271,-0.2952637672,-0.1231216639,0.1760964096,-0.0192733202,-0.0587597154,0.2348760962,-0.2314607799,0.1348613501,-0.0089756967,0.074546434,-0.2175699025,0.2754802406,0.1463979036,0.229322806,-0.2204809487,-0.0933713317,-0.5578266978,0.152080223,0.0814378336,-0.1484388709,-0.0096609844,0.0127661284,-0.0402444787,0.0563468412,0.1488924474,0.3091361523,0.2066040784,0.316795975,-0.433567822,0.6727858782,-0.0681334883,-0.2339499444,0.0665888712,-0.5331949592,0.4914660752,-0.0902959779,0.115521796,-0.0538884588,0.2237475812,-0.371753484,0.4269129634,0.5173207521,0.1495974362,0.3735641837,-0.0100631667,0.0645540431,0.0330894627,0.4389162958,-0.2294113338,-0.1510961205,-0.0262137223,0.0054043112,0.0434774272,0.1580454558,-0.1662987173,0.0185094066,0.1677069813,-0.3099854589,-0.1790644675,0.0888064727,-0.0960524753,0.3011552393,-0.325137645,0.1892915368,0.2122380883,-0.0346654318,-0.2913627625,-0.1555498391,0.3699110448,-0.3290713429,-0.1448325813,-0.3039197028,0.1994423121,0.077651605,0.3248166442,0.1124932691,-0.2303510606,0.3438750505,-0.1254346669,-0.5179362893,0.3420324326,-0.1467755735,-0.2482198328,-0.1574875712,-0.0366525352,-0.1720851064,0.0217488073,-0.0030949716,-0.0453708172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/758","title":"Process 0 very slow when using num_procs with map to tokenizer","comments":"Yes, I can confirm it also happens with ```num_proc=2```.\r\n```\r\ntokenizers 0.9.2\r\ndatasets 1.1.2\r\nmultiprocess 0.70.10\r\n```\r\n```\r\nLinux nipa2020-0629 4.4.0-178-generic #208-Ubuntu SMP Sun Apr 5 23:45:10 UTC 2020 x86_64 x86_64 x86_64 GNU\/Linux\r\n```","body":"<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n","comment_length":34,"text":"Process 0 very slow when using num_procs with map to tokenizer\n<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n\nYes, I can confirm it also happens with ```num_proc=2```.\r\n```\r\ntokenizers 0.9.2\r\ndatasets 1.1.2\r\nmultiprocess 0.70.10\r\n```\r\n```\r\nLinux nipa2020-0629 4.4.0-178-generic #208-Ubuntu SMP Sun Apr 5 23:45:10 UTC 2020 x86_64 x86_64 x86_64 GNU\/Linux\r\n```","embeddings":[-0.4147121906,-0.2900132239,-0.140239194,0.1775934547,0.0856034979,-0.3098762035,0.3165899217,0.2695382237,-0.4015316069,0.1938359886,0.1718184054,0.4401447177,-0.1775085926,-0.1734138131,-0.4684062302,0.1972351074,0.0476887487,0.0190271009,0.4751037061,0.0531506315,-0.2164152861,0.2401550859,-0.3768131137,0.0269497838,-0.4960835874,-0.098242946,0.1989468336,0.1213030815,-0.1391700208,-0.3674746752,-0.6068891287,-0.1121927947,-0.056367632,0.2119784504,-0.0001163927,-0.0296951998,0.064612627,0.2730703056,0.1113588661,0.0883958414,0.3482666612,-0.3627751768,-0.2092968374,-0.1784971654,0.0755575001,-0.1697577238,0.1654167473,0.0946597606,0.3169885278,0.3889237344,0.1781579107,0.3117571175,-0.2135271281,0.2594029605,-0.2771312296,0.1019631773,-0.0687709749,-0.2140848786,0.0303773228,0.24990201,0.0425839722,0.2547703981,-0.2161941975,-0.1262565553,0.0159442481,-0.0336816646,0.3718964159,-0.5546597242,0.2848030031,0.0327237323,-0.1297376305,0.214637503,-0.0601828024,-0.1541355848,-0.188558042,-0.3445382118,-0.0453153625,0.1204939932,-0.2108682245,-0.2227787226,-0.3741155565,0.1841550767,0.0840206817,-0.070907712,-0.1827478707,0.1372275054,0.1799198687,0.2410597801,0.0588663109,-0.225803718,-0.2731664479,-0.2224009186,0.1319557726,0.2197855115,-0.2365812212,0.0364632607,0.0125438012,-0.1788612455,-0.2265859991,0.061087925,-0.3848721981,0.479562819,0.0413384214,0.2088337541,-0.2314832211,-0.0097556654,0.183086291,0.1602279395,0.2368336618,-0.1341395229,-0.172889486,0.0311940089,-0.1120954826,-0.1960592866,0.2113963664,0.013027939,-0.2358991206,0.0171142239,-0.3119621873,0.0140123181,0.090544939,0.199494049,0.2683678567,0.2375326306,0.1195797473,0.4252529144,-0.2917689085,-0.175411433,-0.4300312698,0.1039827317,-0.0970262513,-0.1099700704,-0.11935734,0.0670904741,-0.0476063192,0.21875453,0.0155169098,-0.044767268,0.153735131,0.0297058746,0.1537378579,-0.3857297301,0.1910876781,0.237996161,-0.0603851639,0.2763994336,-0.0695460215,-0.1646362692,-0.3798218966,0.2488522083,-0.4449457228,-0.0545314811,-0.0999007523,0.1063408852,-0.0132189747,0.2420077324,-0.0977023616,0.0103092529,0.1538992524,-0.1459278613,-0.0659778565,-0.2974435389,-0.5686413646,-0.1623958498,-0.0126863318,0.1474449039,0.0966638029,-0.0690117702,0.4687348902,0.1338403225,0.633321166,0.4759652317,0.0957222506,-0.0273743309,-0.1826714426,0.5199816823,0.2056587487,-0.0890081972,-0.4167572856,0.1860774606,-0.2954647243,-0.3388358653,0.1845507324,0.1925991625,0.2760441303,-0.0736087337,0.152774483,0.1710793078,0.2181038409,0.3887585998,-0.3401806653,0.0432673022,0.057945244,0.0052662301,-0.143991366,-0.2692907155,-0.0136551736,-0.1358415335,-0.1160059199,-0.108673729,0.0707351044,0.1403703988,0.0405127108,-0.2887232006,0.035770189,0.0396937765,0.1362640411,0.267501384,-0.20482786,0.1992261261,-0.1795548052,-0.0568981692,0.213009119,0.110573329,-0.0963906348,-0.2223664671,0.0520984493,-0.1409859061,-0.0806370601,0.127523005,0.1280655414,0.3340401947,-0.0087754494,-0.0142836347,0.2048281729,-0.0602168515,-0.1882988065,-0.274857074,-0.2319175303,0.0195146333,0.381421119,-0.0707417205,-0.0461662859,0.2635863423,0.2881624103,0.0682804883,-0.252323091,0.0121111022,0.3966016769,0.0402958468,0.1687045097,-0.0676355585,-0.0248788558,-0.12055365,0.2534329891,0.4735650718,0.1911372542,0.2915652394,0.2026228756,-0.0064282566,0.0646826103,0.1991950274,0.2518364489,-0.1087460592,0.0632869005,0.7362861633,0.1480686069,0.2410648912,0.014350378,0.3129008412,0.4191803634,0.043556422,-0.1877629012,0.3188773692,-0.1470636427,-0.0989779457,0.299633503,-0.4021481276,0.2188609242,0.211352855,0.0284018014,-0.0751190335,-0.1632568836,-0.1046004221,0.1323328763,0.1673816592,0.1391259134,0.1631523818,0.239171654,0.159887284,-0.362634182,-0.2156475931,0.1362676919,0.2745377719,-0.0113989646,0.0717987195,-0.1668399572,0.3841632605,0.1069370657,0.0126776444,-0.0979371518,-0.2654924393,0.1053292006,0.2229979634,0.1926709563,0.1570714861,-0.0243053958,-0.0396595411,0.2687875032,0.0549105778,-0.1527597904,0.1180010661,-0.2908624709,-0.0477984622,0.3717545867,-0.0466328934,0.3291921616,-0.0740024522,-0.04973628,-0.2217082977,-0.44871521,0.0687726289,0.0453884266,-0.0049517802,-0.0374320187,-0.0763192698,-0.2055060714,0.0315735191,0.2338883132,-0.3088588119,-0.2734475732,0.0143820895,-0.192863524,-0.2959949076,-0.0909073353,0.2817852199,-0.2543785274,-0.0103561897,0.0863191411,0.1400083601,0.4790601432,0.0132499197,0.0994248912,0.1158333048,0.2477497458,-0.064958401,-0.1200689673,-0.0427146442,0.2542879283,-0.0453790873,-0.1241725832,-0.1803464293,0.0653553382,-0.0297835805,0.1603381187,0.0569864884,-0.2868960798,-0.3649917841,0.5504561067,-0.158702746,-0.1019768193,0.4046473801,0.2637842,-0.2704053223,0.2251262516,-0.0688944533,-0.0361203142,-0.2399604619,-0.1305760443,0.0161142368,0.2189335376,0.2953400612,0.4784649611,0.1764711291,-0.6153280735,0.3364477158,-0.2446139455,-0.1234952807,0.0053260941,-0.1004122645,0.3539540172,0.2251975685,0.0864546224,0.1184328496,-0.000278366,-0.3927925229,0.1327160597,0.137947157,-0.1562550366,0.0196580905,0.0123996735,0.1647292674,0.1075727418,-0.1255678684,0.0337744094,-0.3275463581,-0.1538697034,-0.200270161,-0.0036017196,0.0448774137,-0.4627699554,-0.6620952487,0.0109444018,-0.7336006761,0.4476736784,0.3246118128,0.2894565165,0.0942433923,-0.0669183135,0.2772776186,-0.3210558593,0.7273993492,0.0402909704,-0.1105608791,0.1507585943,-0.1687192917,-0.5041905046,-0.0186392218,0.1345017552,0.3634614646,0.3612509966,0.9468919039,0.0104201185,-0.3450061977,-0.1934397519,-0.1803585142,0.0897505209,-0.1380214542,-0.2206707895,-0.1580165178,-0.3020619154,0.2238353044,-0.0461464226,0.2723909616,-0.011340972,-0.4036687315,-0.1229081899,-0.2363972217,0.2057241052,-0.0100027816,0.172812134,0.049763497,0.1275156438,0.0655401275,0.1480696946,0.1801037937,-0.3304218352,0.2216948867,-0.0839641988,0.101543583,0.2097090334,-0.0073933764,0.3432660699,-0.0332050845,0.0733464137,-0.4329693019,0.5720503926,0.0549249575,0.4549362063,0.1713409126,0.3446987271,-0.2337751091,0.2966985703,0.0297070779,0.3458833396,-0.0601859018,-0.0055456362,-0.4889617562,-0.0444744378,-0.0668482333,0.4607114494,0.891725421,-0.2397613674,0.2062516958,-0.1659155935,-0.0210574567,0.1621925831,-0.2714112103,0.0365187638,-0.3896453083,-0.1325187534,-0.0069339811,-0.1025191844,0.2989655137,0.0224773735,0.1291529238,0.0697840974,0.209883526,0.252622813,-0.0702297688,0.3553849161,0.1373200268,-0.3419867456,0.1720630527,0.1230013669,0.1367600411,-0.1139203012,0.1542340368,0.0394431278,-0.1398709267,-0.0649955422,-0.0191090107,-0.0732492059,-0.364613086,0.4183243215,0.1932606697,0.0933818221,-0.2395739704,0.4018901289,-0.0070009227,0.0851286575,-0.0032063187,0.2591147125,0.1997061074,0.0484700538,0.0990618244,-0.390566051,0.021752663,-0.2086933404,0.0682059005,-0.0247486997,-0.0074607227,0.0552500077,-0.2507555187,0.1249546632,0.5679706335,0.3063177466,-0.1537461132,-0.119184278,-0.4213725924,-0.2262269109,0.0785809457,-0.1657254547,-0.1487763226,0.2996742129,-0.0420696288,-0.3072770238,0.1031314433,0.3724085093,-0.0656656548,-0.0549712479,0.3066338897,-0.3504789174,-0.3216534257,-0.1695223898,0.2203718722,-0.3156825304,-0.0300312322,-0.0820053667,0.1802492291,-0.2183584869,0.1481049508,-0.0514574721,-0.2333099991,-0.003437781,-0.1631019562,-0.1400856078,-0.5985137224,-0.4524354041,-0.4788419008,0.0775932297,-0.0065823621,0.0323681086,-0.3689259291,-0.0931765214,-0.2567000687,0.0747388452,-0.2269094139,0.2328599989,-0.1825962812,-0.3429647684,-0.2174154222,-0.0178750772,0.2077482045,0.440209657,-0.1981148124,-0.2322364897,-0.0115626929,0.1236372069,0.0910103917,-0.5356752872,0.1012628153,0.3338397741,-0.3837572932,-0.2721379697,0.0243247859,-0.0006904139,-0.2361641973,-0.2305086851,0.3917294443,0.0685489699,0.2969121635,0.1387380213,-0.1238871664,0.3315956891,0.0436571389,0.2184188962,0.0057537323,-0.0165228639,-0.0097491471,-0.0041036578,0.4135787785,-0.1210057214,0.082050994,-0.0716067627,-0.2191260308,0.1075770557,0.0023459084,0.0992526338,-0.1673137844,-0.4050678015,0.3858732581,0.1673626602,-0.1224883422,-0.2120354623,0.3046740294,-0.150965929,0.0866419673,0.0125600826,0.2928382754,0.1308485121,0.0111535089,0.1331164837,0.1800602973,-0.4440456927,-0.2878634632,-0.1062030941,0.1063470021,0.1254697293,0.4623204768,-0.1677715629,-0.0226486456,0.3262932897,-0.1280936599,0.6531358361,0.1967159212,0.4684258103,-0.1613358706,-0.0570569336,-0.1884211749,0.3861746192,-0.0950250477,0.2422851622,0.0119455233,0.1504633278,-0.0132034617,-0.219508186,-0.0053426921,0.6526269913,-0.3278417587,-0.3698408306,-0.1431083083,-0.0239283089,-0.1746342778,0.1361032128,-0.0119876182,0.0564166009,0.0192647818,0.337708354,-0.1200220808,-0.0481933951,-0.2910823822,0.062882632,0.2389504462,-0.160695821,0.2387548685,-0.1345288157,0.2236472964,-0.0656880736,-0.0269113462,0.6454128027,0.526904583,-0.2857975662,-0.0641069934,-0.0504818,0.0743008628,-0.0067117172,0.2845498621,-0.1009718403,0.0349355713,-0.0482496731,-0.0318318531,-0.2941146195,-0.4370779991,-0.0570170805,-0.0528935716,0.2819961607,0.247100696,-0.0916362926,-0.2933861911,-0.1078817919,0.1764712334,-0.116552867,-0.0134157427,0.1211390421,-0.2631138563,0.1782803982,0.0049773785,0.0632242113,-0.2402898073,0.2864079475,0.0018532153,0.1694778949,-0.1832761765,-0.1148007587,-0.4960201085,0.107350193,0.2139244825,-0.1393530518,-0.1253384948,0.0636212081,-0.0505611412,0.0377206318,0.1999995261,0.2938379049,0.3271349967,0.3384621143,-0.4123345017,0.7688537836,-0.034173999,-0.2421052307,0.0951436758,-0.6103778481,0.4780231714,-0.1345040798,0.1196133569,-0.0700356141,0.2151155323,-0.3757578135,0.3962632716,0.5963897705,0.1870025545,0.3403976262,-0.1220012009,0.0530664474,0.0965485424,0.5001577139,-0.1907625198,-0.1072809994,0.0201639235,0.0784714296,0.0902176797,0.1068732813,-0.1552941948,0.0336105339,0.0775835589,-0.367154628,-0.1219521835,0.0768133402,-0.0977123976,0.239567548,-0.337230742,0.1748056859,0.2785285115,-0.0663789213,-0.3390512168,-0.2964253724,0.3573903441,-0.2921608388,-0.1247851551,-0.2662941515,0.1766947806,0.0284450594,0.2948837578,0.0678192899,-0.1883568168,0.3833021224,-0.1505487114,-0.4722623825,0.4093291163,-0.1440520734,-0.2408438474,-0.1465682983,-0.0105964681,-0.1326970458,-0.0035110484,-0.1019627377,-0.1442957669]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/758","title":"Process 0 very slow when using num_procs with map to tokenizer","comments":"I can't reproduce on my side unfortunately with the same versions.\r\n\r\nDo you have issues when doing multiprocessing with python ?\r\n```python\r\nfrom tqdm.auto import tqdm\r\nfrom multiprocess import Pool, RLock\r\n\r\ndef process_data(shard):\r\n # implement\r\n\r\nnum_proc = 8\r\nshards = [] # implement, this must be a list of size num_proc\r\n\r\nwith Pool(num_proc, initargs=(RLock(),), initializer=tqdm.set_lock) as pool:\r\n results = [pool.apply_async(process_data, shard=shard) for shard in shards]\r\n transformed_shards = [r.get() for r in results]\r\n```","body":"<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n","comment_length":73,"text":"Process 0 very slow when using num_procs with map to tokenizer\n<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n\nI can't reproduce on my side unfortunately with the same versions.\r\n\r\nDo you have issues when doing multiprocessing with python ?\r\n```python\r\nfrom tqdm.auto import tqdm\r\nfrom multiprocess import Pool, RLock\r\n\r\ndef process_data(shard):\r\n # implement\r\n\r\nnum_proc = 8\r\nshards = [] # implement, this must be a list of size num_proc\r\n\r\nwith Pool(num_proc, initargs=(RLock(),), initializer=tqdm.set_lock) as pool:\r\n results = [pool.apply_async(process_data, shard=shard) for shard in shards]\r\n transformed_shards = [r.get() for r in results]\r\n```","embeddings":[-0.3621354103,-0.5087301135,-0.1394249797,0.074978061,0.0402247943,-0.3560976386,0.3881988823,0.1418159902,-0.4626859426,0.2230350375,0.0311758332,0.3382255137,-0.2334183604,-0.087559782,-0.4083906412,0.1022370458,0.0076495558,-0.1594914347,0.2531303465,0.0326251537,-0.1771969795,0.2042715698,-0.3319115341,-0.0055946107,-0.6348926425,-0.1702251732,0.0586432293,0.2909309566,0.0616428033,-0.2190775126,-0.6123360395,0.0709186122,-0.1117318198,0.3896766603,-0.0001135372,-0.0468342192,0.0159822349,0.2007141858,0.2702916861,0.2192208916,0.3434351683,-0.3710708916,-0.1273700297,-0.2034926414,0.0797320679,0.069500722,0.0555393435,0.1417084634,0.4168218672,0.1280537099,0.2039487511,0.5180537105,-0.1577168554,0.3365240991,-0.2163354009,-0.1263161302,-0.117036894,-0.2264710963,0.3812011778,0.2920641601,-0.0488046855,0.4048925936,-0.2574481666,0.0048076492,-0.0755035281,-0.0256644879,0.3905305564,-0.6222294569,0.2957596183,0.0915026441,-0.1952246875,0.1449882686,-0.2645458281,-0.1054365262,-0.1821681559,-0.446723491,-0.0244202148,0.1261487007,-0.1433109194,-0.1415111274,-0.5341367722,0.0735078827,0.1497888565,-0.203086257,-0.1841876954,0.1338226795,0.1804544777,0.1738160998,-0.0526166447,-0.2131351382,-0.2869664133,-0.1519455761,0.2450894564,0.0610678084,-0.3582467139,-0.0217568148,0.0150993718,-0.159471184,-0.2735793293,0.0945431963,-0.3346866071,0.6400344372,0.1284670681,0.1383740455,-0.2279019356,-0.1701498777,0.0366427638,-0.0220989808,0.2077170461,-0.3187614083,-0.2897849977,0.0965854004,0.0173377823,-0.2958787084,0.2639737427,0.073942475,-0.191046387,0.0637174249,-0.17707856,-0.0365515612,0.2190357596,0.2024123073,0.1830807924,0.2976763844,0.1855953187,0.4982711375,-0.3031582236,0.0021319415,-0.4452203214,-0.1513814181,-0.1632452756,-0.1266250312,-0.0921153203,0.0348824263,-0.1104717851,0.2267440557,-0.1102699265,0.0132167479,0.4914779663,-0.0523906201,0.2614605129,-0.2932785153,0.1446408033,0.3133213222,-0.0084735081,0.1547652781,-0.0946155265,-0.2155823112,-0.2673734128,0.1971883774,-0.475502789,-0.1252941638,-0.0142284892,0.1151569262,0.0019123899,0.3380803466,-0.0697642937,0.0178970899,0.2934842408,0.2063705176,-0.1937802434,-0.3681376278,-0.4264445007,-0.2172588855,0.0611192584,0.1044013947,0.1582095623,0.0513375401,0.3336831331,0.127746895,0.4713522792,0.5424495935,0.1154186204,0.0417913795,-0.0256374981,0.4444888234,0.2932839096,-0.1252952367,-0.3540532589,0.2891217768,-0.2604304254,-0.2326419502,0.2232977003,0.206240043,0.3754859567,-0.0470183007,0.2187004685,0.1550389677,0.0404801555,0.4499570429,-0.3326165378,0.0179055501,-0.0288358591,0.0906401426,0.1248635128,-0.2718461156,-0.0265716035,-0.340380609,-0.0024562262,-0.0861337557,0.0460289195,0.1847891361,0.0244432371,-0.2166291773,-0.0092744613,-0.0815287903,0.057569433,0.3631345332,-0.1907765418,0.2156136483,-0.1590053886,0.0149135282,0.3050498366,0.101495944,-0.2248282284,-0.1282920092,0.1212932244,-0.0976703092,-0.0827974081,-0.0777845308,0.0200119726,0.2593530715,0.1543824375,-0.0542898774,0.2893970013,0.0681320652,-0.0803182423,-0.4340829849,-0.2544192672,0.0692352504,0.3470928371,-0.0687739551,0.0581307113,0.260525614,0.2913330793,0.0728142411,0.0302236136,0.1023276299,0.3324845135,0.0144097013,0.1362325251,-0.070233576,-0.1539847255,-0.1466698349,0.3840997219,0.5922679901,0.0784226432,0.2981658578,0.1087239757,0.1097748578,0.0894447118,0.0530198216,0.203907147,0.0658640489,0.1775876284,0.5067195296,-0.0242265761,0.0737212077,-0.1616682261,0.3329648972,0.3019838631,0.0647395104,-0.1266432703,0.1125895381,0.0120350504,0.0771588981,0.1308263093,-0.3062311709,0.2695827484,0.2243545949,-0.0358175673,-0.0356405452,-0.2221834064,-0.1087189838,0.0581647791,0.2047241926,0.0137569401,0.2202899903,0.2550688684,0.0539958254,-0.1373395175,-0.2441286892,0.1931726635,0.1583592892,-0.0990221649,0.1024988592,-0.1780764759,0.5981843472,0.0530727021,0.0544366315,-0.1519256681,-0.39184618,0.1261687726,0.2221514881,0.0655586496,0.1626021862,0.0847128332,-0.1941848695,0.1956421137,0.0266870055,-0.0524273999,0.0328526124,-0.1817349941,-0.055635374,0.4671281278,-0.0657507256,0.3467688859,-0.2140849233,-0.1339963526,-0.31273669,-0.3596593738,0.0138718383,-0.0464475006,0.1118478775,-0.084439449,-0.1321330518,-0.2101774663,-0.0616514646,0.3838855028,-0.24511756,-0.1824629903,0.1848247498,-0.1050725877,-0.4156938791,-0.0517984852,0.1632776558,-0.2007575929,0.0415546224,0.0667129979,0.0860705003,0.4626763463,-0.1805195212,0.0466812328,0.08330331,0.1142295077,-0.0243651867,-0.0521159321,-0.0890167505,0.2752712071,-0.0479844846,-0.2823193967,-0.0810268074,-0.0195017513,0.0616551191,0.2946303785,0.0788054541,-0.2685387433,-0.2180759609,0.419676125,-0.0750634745,-0.0371207371,0.2952258885,0.2299128771,-0.2754557133,0.2846280038,0.0524856336,-0.1074728891,-0.1358467638,-0.0922812894,0.0228650309,0.1707631648,0.2578331232,0.8335025311,0.1053365096,-0.3736720979,0.1181062013,-0.2024864405,-0.2785546184,0.0446266867,-0.136722073,0.2043661028,0.0339700095,0.0279805586,0.0777293146,-0.0069423961,-0.232961446,0.1226839349,0.2657170892,-0.2766141295,-0.1160089523,0.0890835226,-0.0251193382,0.2085808963,-0.059984345,-0.0496159978,-0.219357878,-0.1357395053,-0.1355384141,0.1397665739,-0.1594334841,-0.5324476957,-0.5082417727,0.0124831125,-0.5454754233,0.5083017945,0.2737244368,0.1928673685,0.1510091871,-0.1070455462,0.3085050881,-0.3585401475,0.5831413865,-0.0358357355,-0.1943229139,0.154962793,-0.5109229684,-0.3739168942,-0.0847528279,-0.1013148278,0.3901563883,0.4315742552,0.775106132,-0.0408927724,-0.3296204507,-0.3171955943,-0.1898137927,-0.0759424716,-0.0733939409,-0.3145386875,-0.2553100586,-0.3619530499,0.313570857,-0.0463148281,0.2629353702,-0.1495233029,-0.2197107375,-0.1490081251,-0.2139433026,0.335039854,0.0112666544,0.1755581647,-0.1104609594,0.1834638566,-0.0883748904,0.0649578199,-0.0225624815,-0.3216440976,0.2395224869,0.0170024429,0.1262518764,-0.0186622236,0.1967800558,0.4264132977,0.0258304458,0.1372722685,-0.3804469109,0.6753234863,0.0794816837,0.2024178654,0.2228013426,0.3210676908,-0.0504429564,0.3414316177,-0.0597264878,0.0975773633,-0.0394863822,0.0543489382,-0.4334502816,0.071754232,0.0213815644,0.4368113875,1.0347914696,-0.4358527362,0.1840893179,-0.1924218982,0.083478719,-0.1007280052,-0.3134061098,0.1369253695,-0.3445001543,-0.1754264235,0.0421191156,-0.164789021,0.2886514664,0.1133079752,0.0367744714,0.1739537567,0.0925626531,0.0989109352,-0.0623234101,0.3206027448,0.2186335772,-0.3813768923,0.1290139705,0.1123657227,0.1768194735,-0.2750999928,0.0771922022,-0.0579502471,-0.1002491415,0.0317640454,-0.108646743,-0.1455237567,-0.259329766,0.4093880355,0.3598007858,-0.0098444941,-0.1751457751,0.349930495,0.0123373726,0.0349983089,0.0297679175,0.2201571018,0.3177932501,0.0949144661,0.1105690897,-0.3497492373,0.0134541243,-0.2003436834,-0.0784498677,0.1306775957,0.070853956,-0.0052540111,-0.2097006589,0.2090209424,0.4415949285,0.3092518151,0.0101622967,0.0367167406,-0.3541873991,-0.1117398664,0.1057373509,0.0339356139,-0.2966079712,0.4594818056,-0.0079578329,-0.2622581422,0.0560866632,0.185264945,0.0278342217,0.0182014536,0.2954867184,-0.2277003974,-0.3295256197,-0.168914929,0.1978014857,-0.5355457664,0.0459224992,0.0405766554,0.0183423199,-0.2097234577,0.1152487025,-0.0047400841,-0.1750835925,-0.0084918141,-0.2897951901,-0.1205674931,-0.431625843,-0.2958341837,-0.3127335608,-0.0246874485,-0.1550574005,0.317630738,-0.2068932801,0.0822609365,-0.300450176,0.0303124618,-0.3680616617,0.2349937558,-0.0730733797,-0.4186654389,-0.2548908889,0.0349327549,0.2051485926,0.3331361115,-0.2838686705,-0.2526035607,-0.0174226444,0.118307583,0.0069771977,-0.3224290013,0.0711073577,0.3588574231,-0.2826042473,-0.1976970732,0.0328662284,0.0944510624,-0.2486629635,-0.2427313626,0.4648455679,-0.0264098383,0.0024900858,0.0809749663,-0.130436182,0.2661955655,-0.0161989927,0.0968660265,0.0903720856,-0.0383185819,-0.0171180107,0.0095335394,0.3124613166,-0.1557528526,0.1602586955,-0.1301520467,-0.1964568049,0.0054129488,0.0661672801,0.0643637255,-0.1989088506,-0.3912786543,0.3398142755,0.1709181815,-0.1451821476,-0.164874211,0.2732416689,-0.0731093287,0.1227294505,0.1467119753,0.2926249802,0.1291103959,0.0293509737,0.2141939551,0.0854707211,-0.6312884688,-0.3609910607,-0.3290002346,0.0278360788,0.163904801,0.4400305748,-0.1444057971,-0.1780833751,0.2677778602,-0.0944199488,0.6451376081,-0.0538246408,0.3378815055,-0.0794796273,-0.0599521399,-0.0833438709,0.1656211019,-0.0078396825,0.2094669342,-0.1221727356,0.1142318547,0.0292198677,-0.2197521776,-0.1352491379,0.7050784826,-0.3179017901,-0.3046842813,0.0875306204,0.0079261484,-0.154281646,0.1212150753,0.0719904676,0.2372307628,0.4211827219,0.3638751209,0.0921752453,-0.0196367595,-0.2525719702,-0.0456360616,0.192047894,-0.2006839663,0.219558984,0.0583878793,0.1918105036,0.0434352122,-0.0627755001,0.5542598963,0.4263674319,-0.4307270348,-0.0321963727,-0.0631471276,0.0459778756,0.0214263424,0.2826527357,-0.168749705,0.1781712472,0.0393702313,0.0198884271,-0.3380837142,-0.5042130947,-0.0376183167,0.0465958193,0.1441509873,0.3209769726,-0.0669127405,-0.3010896146,0.0428715348,0.1848435253,-0.305460602,-0.1848952025,0.3300402462,-0.3622610867,0.1010539979,0.0227022693,0.0802270994,-0.1056891009,0.3916597366,0.0542831905,0.2547254562,-0.2889440358,-0.0571130142,-0.572809577,0.1688037664,0.1001289487,-0.1853969246,0.0557678491,0.3687535524,-0.0032704556,0.2638692558,0.2638880312,-0.0217773952,0.0840343609,0.4047763646,-0.5053955317,0.6482889652,-0.1292989999,-0.160079062,0.1227157563,-0.4431537986,0.4912435114,0.0863193423,0.1505400985,-0.0308376215,0.1532614082,-0.1095430702,0.4072914124,0.2263593376,0.1058964655,0.393953681,-0.0990698859,-0.0053786617,0.1252945065,0.3513063192,-0.0730170012,-0.1500096768,-0.0505192243,0.0579580106,0.1224644631,0.1014477462,-0.1174239442,0.115297921,0.0493200868,-0.2682154179,-0.1101316065,0.0493659936,0.0971316993,0.2865616679,-0.2756014168,0.3383454978,0.1794983596,0.0900956839,-0.3465446532,-0.2756187022,0.418520242,-0.3566580415,-0.2835139334,-0.356605649,0.208751291,-0.147946775,0.3784295917,-0.045532126,-0.2098756731,0.2967396975,-0.0251551028,-0.4643786252,0.3824313581,-0.0555950291,-0.3912913203,-0.1442773044,-0.1995296776,-0.2414436489,-0.028886294,-0.2539622784,0.010824793]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/758","title":"Process 0 very slow when using num_procs with map to tokenizer","comments":"Nah, I'll just wait a few hours. Thank you for helping, though.","body":"<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n","comment_length":12,"text":"Process 0 very slow when using num_procs with map to tokenizer\n<img width=\"721\" alt=\"image\" src=\"https:\/\/user-images.githubusercontent.com\/17930170\/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png\">\r\nThe code I am using is\r\n```\r\n\r\n dataset = load_dataset(\"text\", data_files=[file_path], split='train')\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), num_proc=8)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n dataset.save_to_disk(file_path+'.arrow')\r\n```\r\n\nNah, I'll just wait a few hours. Thank you for helping, though.","embeddings":[-0.3868586421,-0.4354364276,-0.1802584976,0.13410151,0.1569930166,-0.2301176339,0.3296427429,0.2962342501,-0.4585930705,0.1594596058,0.1962346137,0.3626259863,-0.2120571285,-0.1056413129,-0.4329688251,0.2020953,0.0638447776,0.0465459228,0.6445625424,0.1183298007,-0.1853454858,0.0837816745,-0.3396959305,-0.0557353608,-0.5184189677,-0.2215343416,0.0767242759,0.0712397099,-0.1653123349,-0.3613607287,-0.6819348931,-0.0374706909,0.0291147064,0.2848610282,-0.0001170108,-0.0874471515,0.0966431871,0.2252869159,0.0773267075,0.1193114296,0.3593384922,-0.2990218997,-0.2077735662,-0.2494343519,0.0547069758,-0.1598816067,0.2517273128,-0.0244272649,0.4466660917,0.405870378,0.156382367,0.2713700235,-0.1250472069,0.3425252736,-0.2332341969,0.2126068324,0.0013102533,-0.2122468948,0.2445870191,0.2727806568,0.1339929849,0.2432834059,-0.2232780606,-0.1013321206,0.1405190378,-0.0319232382,0.4205235541,-0.6231470704,0.2275371403,0.0143446671,0.1001030132,0.2192400098,-0.0588691495,-0.1161324233,-0.1346788853,-0.5051073432,-0.0436096378,0.1709661633,-0.1615620404,-0.1913740486,-0.4218501747,0.0368584208,-0.0077557182,-0.0745765492,-0.0898743197,0.1579105109,0.2144381404,0.1967972368,0.0449102707,-0.1643884331,-0.2029230148,-0.1023996472,0.1511241347,0.2848470211,-0.1631530225,-0.0750436485,-0.1017178372,-0.1100398377,-0.1572086364,0.1390159279,-0.3806190491,0.4428103566,0.0814446062,0.1524949372,-0.2354751378,0.0275356844,0.2487822175,0.179919824,0.2428968996,-0.2003704607,-0.1573898345,0.0363128968,-0.1788246781,-0.2794988751,0.2280446291,0.0307574421,-0.2253383696,-0.0045749969,-0.1468830556,-0.0642471761,0.0989969969,0.2049956769,0.1935445219,0.2434437126,-0.0013603744,0.2977653444,-0.225454405,-0.1638171524,-0.3836987317,0.0930541456,-0.0750038326,-0.0325040035,-0.1291438043,0.072226502,-0.0952870548,0.1423259228,-0.0169265289,-0.149977088,0.1320409477,0.0054332367,0.0430204235,-0.323187381,0.2110825479,0.154749006,-0.1294487417,0.212220028,-0.076766558,-0.1230314523,-0.3588643372,0.2667497098,-0.4313960671,-0.1472436637,-0.2475175858,0.1116718054,-0.0784983858,0.1869377792,-0.1061788201,0.0527346991,0.1981505007,-0.1673814803,-0.1065982878,-0.2098736912,-0.5488907099,-0.1729204506,0.0702576935,0.1332519799,-0.0267391689,0.0297846254,0.4306248426,0.218362242,0.5822152495,0.5189096928,0.0849284157,-0.0407894552,-0.1276498288,0.3768569231,0.1648821682,-0.0467640571,-0.4396910965,0.2034142315,-0.3376708627,-0.4503232241,0.2453542501,0.2947375476,0.2543146014,-0.0176461134,0.2343882173,0.1872132272,0.119199574,0.3299919665,-0.3725756705,0.0205147862,0.1226871833,0.099348627,-0.1906185895,-0.2688972354,-0.0967785195,-0.1852292717,0.0004192661,-0.0831514597,0.0685142428,0.2069037706,0.1599231064,-0.2835086584,-0.0290791262,0.1657842696,0.1408671886,0.223802343,-0.0864791721,0.2587096691,-0.2587969303,-0.0815152377,0.1333564818,0.0703755841,-0.0844828039,-0.200100556,0.0113468245,-0.2404735684,-0.1034041345,0.107672736,0.036469847,0.3120546937,-0.0544656776,-0.0056454339,0.0526377633,-0.1108678058,-0.1789439917,-0.2310866117,-0.1503865123,-0.0611532032,0.3542821109,-0.1036711633,0.0174255688,0.26953879,0.2351294011,0.0855450407,-0.2370819002,0.070896931,0.4436505735,-0.0407675281,0.2335380316,-0.0791156143,-0.0519739874,-0.0708527267,0.0535034649,0.4686065018,0.178964138,0.2506849468,0.172337696,0.0238311775,0.0512158312,0.1808116585,0.2077569962,0.0567524433,0.2169198841,0.6340783834,0.1424715966,0.2417574674,0.0436554067,0.3440587819,0.3492493927,-0.0491053462,-0.2114381343,0.3780713379,-0.0720847249,-0.1003918573,0.2829619348,-0.5480772257,0.0918913558,0.2359430939,-0.035710901,-0.0622122921,-0.190929845,-0.0386777893,0.1346299052,0.1598608345,0.1788556129,0.0395130217,0.2177822292,0.179943338,-0.2889860868,-0.2397528887,0.1236830428,0.3608110845,-0.0653002188,0.0593392774,-0.1140489876,0.3863296211,0.0479478315,0.0983444303,-0.1607990414,-0.2956201434,0.1647772789,0.2017847598,0.2086825967,0.1556718349,0.0790115073,0.0150866769,0.3185814321,-0.1264426857,-0.189647451,0.0535034016,-0.1705431491,-0.0290024318,0.4241763055,-0.0880040526,0.3978841901,-0.0834888741,-0.0052670259,-0.2546927929,-0.4800395668,0.0158046838,0.0318632983,0.0244897418,-0.0779073238,0.0620434359,-0.3044963479,-0.0307395868,0.1447134167,-0.430989176,-0.2562635839,-0.0632642284,-0.179904744,-0.2704198956,-0.0367128626,0.2118083835,-0.2973480821,-0.031154817,0.1404358298,0.2213062942,0.4067439437,-0.0544368029,0.1965536475,0.128539294,0.3354866207,-0.1129290164,-0.0556798652,-0.2245120406,0.3730615377,0.0039263689,-0.1762271076,-0.1658287793,0.0794989914,-0.1046111807,0.1886092275,-0.0469145626,-0.2700690031,-0.3126637638,0.4584556818,-0.1833041757,-0.1591628045,0.3546036184,0.1594922096,-0.2280762047,0.2647324204,-0.0014942122,-0.0970419124,-0.1285646558,0.0746217817,-0.046603445,0.2935386896,0.2267158777,0.5777073503,0.0520409197,-0.657663703,0.2194339037,-0.2869417071,-0.1751941442,0.0028184918,-0.0693595856,0.3885862827,0.1792207658,0.0151448734,0.0738896206,0.0413428582,-0.3943099678,0.0386106186,0.1252323389,-0.19124116,-0.055244986,0.0178034212,0.108916983,0.0941894948,-0.1417355537,-0.0382167958,-0.3680287004,-0.2644228339,-0.1909839213,-0.0072536948,0.0963615924,-0.3810267746,-0.5885569453,-0.0436379984,-0.7761921883,0.5087891817,0.3386169076,0.1255489439,0.0932441652,-0.0340648927,0.2616506219,-0.4111613631,0.8351029754,0.0705500245,-0.0535059907,0.0675406232,-0.1544954926,-0.4474802017,0.1085085869,0.1985217631,0.2989944816,0.2902831435,0.9171606898,0.0049875579,-0.3062026799,-0.1606014222,-0.1485232711,-0.0169021301,-0.1029158309,-0.1912089288,-0.2857264578,-0.4631206095,0.2842850089,0.1612758636,0.2377052903,0.0326133035,-0.3594405651,-0.1032888889,-0.17431207,0.232149139,-0.0357375443,0.1585120857,0.0333978795,0.0893579721,0.1960792542,0.2048343122,0.1541989595,-0.3156464994,0.1993881315,-0.0802174434,0.0803235471,0.1185577661,0.1544083506,0.4262932241,0.061097946,0.0165943056,-0.3274839818,0.6634241939,0.0143608255,0.5038141608,0.1549650431,0.3396929204,-0.3493783176,0.1584633142,0.0589402504,0.3506843448,0.0275165252,0.1028949916,-0.4589796662,0.0642464086,0.0171915349,0.4936144352,1.1489685774,-0.3767484426,0.196278885,-0.2332410961,-0.0053219977,0.1026098356,-0.3771379888,0.063765496,-0.414324522,-0.1479073316,0.031679254,-0.110341534,0.2179987878,0.1088583544,0.1165269986,0.0691412389,0.0713806748,0.2985092998,-0.1181405112,0.4149611592,0.175529182,-0.302102834,0.0794324055,0.1030875295,0.0600263365,-0.1230729371,0.1000511721,-0.0038368537,-0.1578944623,-0.0559587143,-0.0001858436,-0.1246965453,-0.3861263096,0.3683027923,0.0359349884,0.044979278,-0.1801919788,0.4966416657,0.0014412233,0.1288654357,-0.0315248184,0.3320120871,0.2719569206,-0.0623940304,0.0846105516,-0.3868014514,0.1015374437,-0.2599090636,0.0735289603,0.0592579879,0.0054553463,-0.0700884685,-0.3401174843,-0.0373889171,0.5543574691,0.2927549481,-0.0568932332,-0.1591741145,-0.5194124579,-0.1436770409,0.0701248571,-0.0623824671,-0.1752188355,0.305308789,0.045717977,-0.2282149345,0.0573859885,0.378036797,-0.1269847453,-0.1234340221,0.333162576,-0.2295175046,-0.3170024157,-0.1727051288,0.2842459381,-0.27013731,-0.0160320271,-0.0374397188,0.1870573312,-0.1717520505,0.0937918201,-0.0035765795,-0.1896744967,-0.0809546188,-0.0347070061,-0.1684312969,-0.6950533986,-0.3513256907,-0.4066695273,0.1483950317,-0.0177134573,0.0937714353,-0.267639488,-0.0354391709,-0.2852414846,0.1015298516,-0.2572075427,0.2609927058,-0.2267630547,-0.3777024448,-0.191192165,-0.0201066118,0.2279696167,0.4707986712,-0.1789884567,-0.2302058637,0.0089994716,0.1635390967,0.0850325599,-0.3943318725,0.0949643701,0.1965423077,-0.4006542265,-0.1671240926,-0.1331324577,0.0216651186,-0.1444446445,-0.3322989941,0.4063672125,0.0892203599,0.3379632533,0.0661463365,-0.1011143848,0.3379051089,0.0604139306,0.1913604736,-0.0002302253,-0.043518573,0.1263835728,0.037028946,0.4706883132,-0.0563031323,0.1058802381,-0.0679779127,-0.197763294,0.0779391974,0.0280793216,0.1444304585,-0.209707737,-0.3047909141,0.4175078571,0.1545967311,-0.174370423,-0.1509354264,0.2933526933,-0.1973045319,0.0160925314,0.0380243398,0.2129042,0.2618134022,-0.0382114127,0.158889994,0.2943808436,-0.5616085529,-0.380471319,-0.065765284,-0.0071543641,-0.0098911524,0.5164598227,0.0028408321,0.0364554487,0.270165503,-0.133932814,0.575045228,0.102573812,0.5172138214,-0.1260009855,0.1103643477,-0.2094849944,0.4156570137,0.0166746937,0.2539340258,0.0285375938,0.2568211555,-0.0453151539,-0.1782808155,0.026085848,0.6844560504,-0.2174592018,-0.2757240236,-0.1599409878,-0.057441406,-0.1849276572,0.0889554992,0.0746182725,-0.075128004,-0.0206818283,0.3935567141,-0.0261386205,0.0300011225,-0.1854310781,-0.0162082836,0.3558835983,-0.1863145381,0.0884681493,-0.1706596166,0.1824912131,0.0028872697,-0.2100394964,0.5446403027,0.4241334498,-0.5148571134,0.066367574,-0.0466068871,0.0577872954,0.0682013929,0.2501453459,0.0156495497,0.0376494974,0.0011939917,-0.0207438469,-0.2810029089,-0.3404556513,-0.1160094738,0.0318790823,0.3446784317,0.1906518489,-0.0390959419,-0.4372619689,-0.1937607676,0.0385897607,-0.167571649,-0.0887572169,0.1056996211,-0.3172115684,0.187000066,0.0411653481,0.0620840639,-0.2634866834,0.313816756,0.0088809179,0.2916043699,-0.2433420718,-0.101151228,-0.5551652312,0.0671917796,0.3327208161,-0.232482329,-0.054489214,0.0037507466,0.0440671518,0.0728895664,0.139260307,0.2297839373,0.149304539,0.3782693148,-0.4107660353,0.7121807337,-0.0387569927,-0.0717252418,0.133231923,-0.4897900224,0.4451346993,-0.1202931032,0.1071482077,-0.1304897666,0.1382304281,-0.3424626291,0.4242952466,0.4949221313,0.0921372697,0.313100785,-0.0799106732,0.0774298534,0.1227345169,0.5301703811,-0.2131965905,-0.1305776387,0.1620765328,0.0397319905,0.1119386256,0.1467001587,-0.1729177982,0.0367608815,0.0516472533,-0.4973432124,-0.1654943228,0.0267188381,-0.0417396985,0.2200377136,-0.3419336975,0.2175743729,0.189138189,-0.1107802093,-0.3523395061,-0.2495200783,0.5046346784,-0.2942952812,-0.1918192357,-0.1261372566,0.1038018018,0.1934222132,0.4178584218,0.052925501,-0.1292170733,0.3099063039,-0.1265565008,-0.5185464025,0.4152126014,-0.049122408,-0.2529774606,-0.1285522431,-0.0443628691,-0.1448897123,0.1328309923,-0.2833017409,-0.1970816553]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/757","title":"CUDA out of memory","comments":"Could you provide more details ? What's the code you ran ?","body":"In your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n","comment_length":12,"text":"CUDA out of memory\nIn your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n\nCould you provide more details ? What's the code you ran ?","embeddings":[-0.3311731219,-0.0803365931,-0.1735431254,0.3186485171,0.6322613358,-0.0376044698,0.4316090345,0.2629073262,-0.0620163083,0.4164027572,0.1511951685,0.1299936473,-0.0573243462,0.2183705419,-0.1382766962,-0.1967546642,-0.123469092,0.0716228336,-0.1611089706,-0.0300138388,-0.0045555653,0.0638250187,-0.2565544844,-0.0716901496,-0.3902517855,-0.0811709762,0.1256887466,0.1167816669,0.1096124053,-0.1783622205,-0.0624942705,-0.0950220749,-0.0217838548,0.3943614364,-0.0000940633,0.0375876427,0.0659187138,-0.0837553144,-0.1677701622,-0.1017927006,0.0188903157,-0.0859654248,0.2191261649,-0.3291553855,-0.3077498674,0.2350233644,-0.0466809943,-0.1646973938,0.4626291394,0.4594749808,0.4260752201,0.0357029699,0.0110095181,-0.1526985317,0.0603744686,-0.0910968706,-0.0591243394,-0.0800056085,0.1262499839,0.0347637944,-0.1038746685,0.1986875236,0.075070411,0.0340476781,-0.0085610347,0.0314447358,0.0038584229,-0.1088101789,0.1604140848,0.0178463105,0.2432644963,-0.0856911764,0.1108799726,0.0369610339,-0.0283977687,-0.4129005969,0.2014254928,0.0642444491,-0.2818688154,0.0893573239,-0.159812659,-0.0741754919,-0.0900432169,-0.0643579885,-0.1651007831,0.0746425316,-0.1421723813,-0.0402286574,0.2027718723,0.0095603447,0.2094968855,0.2059761882,0.0802452639,-0.0305371024,-0.2038987428,-0.0203365572,-0.0418366976,-0.1675187349,0.0468229614,-0.1817803085,0.2306900769,0.1186237782,0.107156761,0.0460258275,0.1037259549,0.2572423816,-0.1822040677,-0.0675459504,0.0978305191,0.1991364509,-0.4336097836,-0.2073480487,-0.0667169243,-0.2976900935,-0.0259211827,0.0247074235,0.0468581617,-0.1699950844,-0.1295734644,0.1909290105,-0.1784307957,0.1336729676,0.0317543224,0.3226697147,-0.0821040422,0.1202526465,0.140561834,-0.0741335154,-0.0965460017,-0.1776317209,-0.3109577,-0.0051309331,-0.3132040501,-0.0426123552,0.1858206838,-0.0482715033,0.3879832029,-0.0939747915,0.0566688627,-0.3478482664,0.0732938126,-0.2285946012,-0.0315331519,0.0205796808,0.2646313906,0.0070514106,0.0456756428,-0.1445839107,0.2854070663,0.3102224767,-0.2231054157,-0.2050350159,-0.106887877,0.3904202282,0.2958566248,0.122736752,0.0036712729,0.0560515486,0.2754685879,0.1003884152,-0.0335981809,-0.0617270172,-0.1021131128,-0.127489537,0.2904540598,0.210787192,-0.3614667356,-0.0237172283,0.0554282963,-0.0509923957,0.2017987967,0.4653459191,-0.3429935873,0.1515907049,-0.122111395,0.0513894893,0.210186094,-0.2085656077,-0.2186412513,0.1398255974,-0.1913934648,-0.0475156009,-0.1828304827,-0.0689988434,0.3632155657,0.1496115625,0.2107840031,0.1580813527,-0.1014302075,0.0260269959,-0.3733839393,-0.1600344777,0.1295024455,0.1154429987,0.2628787756,0.0322844461,-0.1616015732,0.1611563116,0.2789415419,-0.1528163552,0.0447711684,0.2019090056,0.3102604151,0.1053462029,-0.1177274287,0.029875977,0.0967184231,-0.0082143033,0.1750383377,0.1497916132,-0.0680057406,-0.1150592566,-0.092666477,-0.16996786,-0.0603418797,-0.2448085845,0.4099098146,0.0178954992,-0.261898905,0.0578211024,-0.0230336692,0.2859028876,-0.5411349535,0.1115070134,0.1304994375,0.1472744048,-0.2702678442,-0.0930233598,0.0305206366,0.1799910516,0.1064453349,-0.1321322769,-0.1624475121,0.2998970151,0.0255118366,-0.138762489,-0.0194935724,-0.101259321,0.0573986247,-0.1405729502,-0.0524192862,-0.038931217,0.1393303871,-0.0597623996,0.0115033304,-0.0125965467,0.1099652797,0.0141418269,0.0939700156,-0.0541635752,0.2759802639,-0.1126097515,0.244639501,-0.1336678416,0.1223605573,0.2197733372,0.1669643521,0.1134319156,-0.391397059,-0.005893473,0.4713177383,0.2091960311,0.0273191407,0.2356284708,-0.037014246,-0.0060504,-0.1432538182,-0.08511962,0.01367353,0.2423564941,0.1329057664,0.1344209015,-0.0216995366,-0.1787326485,0.1482079625,0.0289075188,0.1611384749,-0.0737490505,0.1134065911,-0.1119518355,-0.3872449398,0.0183104612,0.1684061587,0.4123133421,-0.1514962614,-0.0794492066,-0.2176454514,-0.0354715548,-0.1448896825,-0.0356646068,0.0117704105,0.0004585455,0.1053671241,0.147795856,0.0525828861,0.3354626298,0.2482247055,0.2383232713,0.1011589542,0.2712031007,0.100518465,-0.2856894135,-0.1223889366,0.320677489,0.3138345778,0.0534393229,0.1381481141,-0.0111003257,0.0002681672,0.0383604877,-0.1562920511,-0.1231292561,-0.0358194858,0.2442009896,-0.1248871461,0.300786376,-0.0663089156,-0.3005660176,0.1731528491,0.0527940765,-0.1846019179,-0.2301633805,0.2285455167,-0.1417070925,-0.2493442148,-0.2954123914,-0.0717201233,-0.393280983,-0.0183455553,-0.0496885404,0.0932413861,-0.0590521879,0.3122150898,0.2935273945,0.1413881928,0.1075324863,-0.2150531709,-0.223020792,-0.0351428501,-0.2034439743,-0.2818828225,-0.0360513963,0.1461365819,0.5090152621,0.0321478993,-0.5990983248,0.1021220982,-0.0142735001,-0.0337694846,0.0007518526,-0.1885185987,0.3492803574,0.0987028033,-0.28537637,0.0510881282,-0.172121048,0.1802300364,-0.2217407227,0.0990715101,-0.2772093117,0.6215999126,0.0944464728,0.5560269952,-0.1098445952,-0.1806336492,0.3287634552,-0.1136434898,0.2123535722,-0.333091557,-0.3265771866,0.228421554,-0.1671674997,-0.1870498806,0.123694621,-0.1918957233,-0.3011939526,-0.1743255556,0.0858827457,-0.324555546,-0.3147713244,0.1646285951,0.0400211513,0.1808036268,0.0428792723,0.1655947566,-0.0299806595,0.014737675,0.0131666483,-0.0546873137,-0.0710295811,-0.2482457012,-0.1424703747,-0.0607185625,-0.2698639929,0.0998352319,-0.0432176255,0.1668251753,-0.2431155592,-0.0983385518,0.1031222194,-0.1852108389,0.6754658818,0.268932879,-0.2628563344,0.0951890424,-0.2138830721,0.0596922822,0.0439357609,-0.209960863,0.0793595165,0.1608300805,0.2077421695,-0.0738334954,-0.1111094654,0.1235438883,-0.0315761678,-0.1565324962,0.0985502899,-0.0375633016,-0.087832354,-0.3092781305,0.3059138656,-0.0078616226,0.2585193217,-0.1554540843,-0.3953620195,-0.1663855612,-0.2413648814,0.0360967852,0.0422013327,0.2953031957,0.050062947,0.1851315945,-0.202246815,0.2679366469,0.0784121826,0.0078508072,0.0263577197,0.1424129307,0.0376091935,-0.2152002752,0.2844597101,0.2412382513,0.0610372163,0.1335196793,-0.0490477309,0.3227410316,0.0185516775,0.4345700145,0.3531694412,0.2294142395,-0.26488325,-0.1870297045,0.1468691081,0.048871439,0.101167202,0.3652064204,-0.0096090594,-0.2562055886,0.0946315601,0.1298971623,0.7965639234,-0.0583988875,0.1323883384,0.1354368776,-0.1410504878,0.0653771311,-0.1011435837,0.1491664201,-0.1996470094,-0.2138191164,-0.0141981812,0.0952409208,0.116299741,0.0406131148,-0.2712939978,-0.1398149282,-0.2560619712,-0.0388413817,-0.1506883353,-0.0929624438,0.257117331,-0.2681669891,-0.264736712,0.3419808149,-0.0221495759,-0.150381729,-0.1287244558,-0.0178672895,0.2062958628,-0.0034799045,-0.1676839441,-0.0297257062,-0.4293828607,0.0951388776,-0.2170300335,-0.419472903,0.0476106144,0.0151321907,0.3110411167,0.2793773115,-0.0539255664,0.3366350234,0.067164816,-0.0357217006,-0.0005637978,0.2001040578,0.3037866354,-0.1471116096,-0.2554418445,0.3248833418,-0.0419527516,-0.2529807687,0.0814296305,-0.0240970217,0.060961619,-0.3186044395,-0.2821189761,0.2090885341,-0.1154341623,-0.2515000999,0.3540236056,-0.0740089715,-0.1892713904,0.0499914289,0.2412368506,-0.1499489993,-0.1098418832,0.2127569467,0.1642225385,0.0602424107,0.511734128,0.2239760458,-0.181303218,-0.3512002528,-0.2402063757,0.2829802334,-0.11261338,0.1271608919,0.036297854,-0.1076725647,0.1308922172,0.2884369791,-0.0033330806,0.1373281479,-0.1398479939,-0.0534761362,-0.6767534614,-0.1101318523,-0.0739743784,0.2218864262,0.0403752662,0.3503749073,0.110531494,0.0904591605,-0.4928095639,0.2385096252,-0.5095472336,0.3083234131,-0.1420601606,-0.215845868,0.0945696309,0.0085475547,0.3011172414,0.4089073241,-0.047635749,-0.4878978133,-0.0309447031,-0.0267841406,0.0770379752,-0.0916838795,0.0260925926,-0.0740045235,-0.2193322629,-0.1696593612,0.1821986288,0.0776316077,-0.0109238159,0.0380769968,0.0228185784,0.1196374595,0.0334693268,-0.1772734225,0.112271741,0.173694402,0.1841409206,-0.1732635945,0.0396502279,-0.0324111879,-0.2685368359,0.1097730324,0.0589577332,0.1305491328,0.0575968735,-0.2334944755,-0.185941115,0.1650736779,0.1561413556,0.1100264341,-0.1040899903,-0.0669035688,0.3573342264,0.4721309245,-0.3924895823,0.1112552285,-0.0753188282,-0.2440089434,0.1936084181,0.0490845814,-0.1518117636,-0.1343807578,0.026849078,0.2819062471,-0.0664621741,-0.3213450909,0.0485154986,-0.0251542777,-0.1563159078,-0.0869131312,0.0183550064,0.0984572396,0.2066317648,0.635245502,-0.0316832475,0.2937098145,-0.0434062704,0.0341534093,-0.0845698789,-0.0747278184,-0.1691175401,0.2129261196,-0.1215696409,0.0246775579,-0.19291915,0.301271677,-0.0110547673,-0.4234782159,-0.2920731604,0.1390277892,-0.4996100664,-0.0150819281,-0.0183331892,-0.0783632025,-0.1010973528,0.1056126505,0.1402803659,-0.0025299513,0.2537689507,0.0120885475,-0.2410276383,-0.0951509401,-0.0513837896,0.1307509392,0.0489582978,-0.1624327451,0.3208777905,0.0614394546,-0.1387870163,0.0105939107,0.1033188626,0.2492858171,0.3025606871,-0.0330305323,0.0185005125,0.0010637415,-0.0717216358,-0.1371000856,0.0713851601,0.3954438865,0.3473542035,0.1495565176,0.3094603717,-0.3370725513,0.2506276369,0.2308125496,0.1188576147,-0.3156321347,-0.0236234944,0.0616114102,-0.0871367306,-0.3836601675,-0.1020905524,-0.4021757543,0.1510528177,0.0539669581,-0.1820160449,0.0053993454,-0.0583642311,0.182753846,-0.0996073186,0.4807362556,0.0025471684,0.0004217636,-0.3862189054,-0.4053084254,-0.376365453,0.1137571558,-0.101693444,0.2785610557,-0.1809933186,0.1877205372,-0.0699564442,-0.1047500595,0.0416394137,-0.0206025112,0.034537822,-0.2018388212,-0.5220823288,0.0366242416,-0.3324203789,-0.0939854011,0.154002741,-0.067653805,0.2420821488,-0.2044366449,0.3008611798,0.1361327171,0.1057702824,-0.0364364646,0.2178577334,0.2591045499,-0.215166986,0.1001019478,0.0422790125,0.216383338,-0.0015072393,-0.1756122708,-0.0843199342,-0.0063720164,0.4266128242,0.2691765428,-0.0864145309,0.0804400668,-0.1269478798,0.2170118392,0.2193619609,-0.0886540413,0.0157117154,0.0605428852,0.2258880138,-0.0485352166,0.2050215602,0.39304474,-0.0044092364,-0.0495738946,-0.4330110848,-0.3016313314,0.5321094394,-0.0943858251,-0.3146249056,-0.1302324384,0.012318654,0.2421713024,-0.1696163565,-0.7161417603,-0.1776245087,0.1271902323,-0.1036652178,-0.3127447665,0.2600634396,-0.1679189056,0.1135847121,0.0241355523,0.2241726369,0.0308397394,-0.0453081019,0.091014646,-0.1937315315]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/757","title":"CUDA out of memory","comments":"```python\r\ntokenizer = FunnelTokenizer.from_pretrained('funnel-transformer\/small')\r\n\r\ndef tokenize(batch):\r\n return tokenizer(batch['text'], padding='max_length', truncation=True,max_length=512)\r\n\r\ndataset = load_dataset(\"bookcorpus\",split='train[:1000]').shuffle()\r\ndataset = dataset.map(tokenize, batched=True, batch_size=512)\r\n\r\n# dataset = LineByLineTextDataset(\r\n# tokenizer=tokenizer,\r\n# file_path=\".\/wiki1000.txt\",\r\n# block_size=128\r\n# )\r\n\r\ndata_collator = DataCollatorForLanguageModeling(\r\n tokenizer=tokenizer, mlm=True, mlm_probability=0.15\r\n)\r\n\r\nconfig=FunnelConfig(\r\n return_dict=True\r\n)\r\n\r\nmodel= FunnelForMaskedLM(config=config)\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/checkpoints\",\r\n overwrite_output_dir=True,\r\n do_train=True,\r\n num_train_epochs=1,\r\n per_device_train_batch_size=16,\r\n per_device_eval_batch_size=16,\r\n save_steps=10000,\r\n logging_dir='.\/ptlogs'\r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=dataset,\r\n)\r\ntrainer.train()\r\n```","body":"In your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n","comment_length":64,"text":"CUDA out of memory\nIn your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n\n```python\r\ntokenizer = FunnelTokenizer.from_pretrained('funnel-transformer\/small')\r\n\r\ndef tokenize(batch):\r\n return tokenizer(batch['text'], padding='max_length', truncation=True,max_length=512)\r\n\r\ndataset = load_dataset(\"bookcorpus\",split='train[:1000]').shuffle()\r\ndataset = dataset.map(tokenize, batched=True, batch_size=512)\r\n\r\n# dataset = LineByLineTextDataset(\r\n# tokenizer=tokenizer,\r\n# file_path=\".\/wiki1000.txt\",\r\n# block_size=128\r\n# )\r\n\r\ndata_collator = DataCollatorForLanguageModeling(\r\n tokenizer=tokenizer, mlm=True, mlm_probability=0.15\r\n)\r\n\r\nconfig=FunnelConfig(\r\n return_dict=True\r\n)\r\n\r\nmodel= FunnelForMaskedLM(config=config)\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/checkpoints\",\r\n overwrite_output_dir=True,\r\n do_train=True,\r\n num_train_epochs=1,\r\n per_device_train_batch_size=16,\r\n per_device_eval_batch_size=16,\r\n save_steps=10000,\r\n logging_dir='.\/ptlogs'\r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=dataset,\r\n)\r\ntrainer.train()\r\n```","embeddings":[-0.3100788295,-0.129237324,-0.1342045665,0.1175396442,0.4688266218,-0.144197613,0.2783528566,0.0998793095,-0.2433777303,0.3591911495,0.0829845816,-0.0730310678,-0.1192087159,0.3212790191,-0.1076370701,-0.1704480499,0.0092369532,-0.0881269574,0.0153183658,0.0442902222,0.0046651633,0.0154469348,-0.2489062846,-0.0319562852,-0.2206834406,-0.1371230483,0.098574169,-0.0503237098,-0.1877260506,-0.3946231604,0.0375735722,0.1146164834,0.2151951045,0.4167882502,-0.0001027996,0.100179024,0.0244873706,-0.1405445188,-0.420407176,-0.044092454,0.2414928675,-0.2798286676,0.0506280325,-0.2537201047,-0.217214182,0.2278189063,-0.137842834,0.1720422953,0.4584761858,0.3797530234,0.3435774744,0.1088275239,0.0585624166,-0.0294913892,0.0538175181,0.0755457133,-0.094014205,0.0066134301,0.1109111086,-0.1369544417,0.0238310844,0.3242061734,0.0113588162,-0.0089151151,-0.0367027931,-0.0341678485,0.2736227214,-0.2479079366,0.1470226943,-0.0891442969,0.1329367161,-0.303568095,-0.0373581164,-0.154106468,-0.0806474909,-0.4634771049,0.104373008,0.1010245979,-0.4107734263,0.0172619764,-0.060389284,-0.0050989529,-0.0138990106,0.0261216536,-0.0662517399,0.5225920081,-0.105364047,-0.0356148928,0.1708458811,-0.1355095953,0.4120981097,0.0853057355,-0.1276610792,0.0237403847,-0.2587937713,-0.1875814199,-0.0713279247,-0.1986297518,0.0711449012,-0.1124361977,0.1579910815,-0.0060249884,0.485934943,0.0080661094,0.0980062783,0.3584094346,-0.1202131212,0.3187831044,0.036064826,0.1946543455,-0.2176494598,-0.1244595945,-0.1859422475,-0.3380061984,0.0242839288,0.0381148756,-0.0462808795,0.0161911696,-0.1271799654,0.1684389859,-0.2535944283,0.273471266,0.1817775816,0.4570909142,-0.0964738876,0.2302170694,0.1769119948,0.039496012,-0.2093317509,-0.0818557143,-0.2425555587,0.1164505482,-0.3404504359,-0.1421803385,0.1027912721,0.2205390334,0.325173974,-0.1931248605,0.039561864,-0.3057804406,0.0956942514,-0.2383188754,0.005653997,0.120123364,0.3433279693,0.0877017304,0.0954148024,-0.2783067524,0.1181035638,0.2075963616,-0.2924019396,-0.2291511446,0.2030310035,0.3269710243,0.2827208042,0.1686135232,-0.3689680994,-0.0063974787,0.2053465843,0.1965654343,0.0562733859,-0.3184034526,-0.217391789,-0.1352854371,0.2781272531,0.3030323982,-0.1902775317,-0.2108455002,0.0214644112,-0.0315198824,0.3774225116,0.3584521413,-0.3117789626,0.458008647,-0.0675834194,-0.0518542342,0.5453791618,-0.5652977228,-0.2127517611,0.1543048024,-0.1909199059,0.0454589501,-0.2386426181,-0.2231659591,0.4829888046,0.0969781056,0.3733592331,0.2380772829,-0.0351438671,0.1896499544,-0.2864001989,0.0090552494,-0.0297914427,-0.0055249878,0.048713401,-0.0527741164,-0.2343461215,0.1280329823,0.2338199168,-0.0104771908,0.2677803338,0.3582985401,0.3073626459,0.116965957,-0.1215274259,0.1936755478,0.1673586965,-0.011587088,-0.0309378542,0.1239467412,-0.0786432847,-0.2058527172,-0.0455837846,0.082807377,-0.0156984944,-0.359300375,0.3384872079,0.0375348926,0.0266572926,0.019130975,0.12357115,0.2986924946,-0.3541118503,0.2222425044,0.2377438098,0.0852836743,-0.188553378,-0.1800913662,0.0524535179,0.1183927581,0.2418252528,0.0297301617,-0.09869048,0.2562616467,-0.0663084537,-0.0891714245,-0.093724899,-0.1130933166,0.0153399678,0.0903212279,-0.0647252128,-0.0206237342,0.1655088067,0.0199047942,0.351183176,0.2012838572,0.3272457421,0.1273166686,0.20463714,-0.0206221342,0.2332151383,-0.0857417285,0.1878331602,0.0592325293,0.2610836327,-0.0133153806,0.1250681132,0.0058278828,-0.2236133516,0.1084053963,0.2264301479,0.0882090405,-0.0231090486,0.2589934468,-0.2839830816,0.011871255,-0.0420016795,0.0582629777,0.0772991255,0.2125343382,0.1205851436,0.02303656,0.0700101703,-0.198204115,0.2579903603,0.0380883589,-0.0887699425,0.0096041206,0.1740602255,-0.1931758523,-0.3281930387,-0.1449996084,0.1208069474,0.4868705869,-0.1175011918,0.0966288596,-0.0272011105,-0.1130644977,-0.2430841476,-0.0940625593,0.1485604197,-0.0184286889,0.043626897,0.4213282168,-0.0010324048,0.3150789142,0.3121339977,0.2186200023,0.1445562392,-0.0231679399,0.0133008128,-0.1335752606,-0.2880644798,0.2566813529,0.2196992189,0.0427814312,0.1171905473,0.0628623143,-0.0104829054,-0.0288040359,-0.0449764356,-0.016239509,0.0228475034,0.1572996229,-0.165305838,0.2963692546,-0.2617641687,-0.4079509079,0.1808433533,-0.060446579,-0.2805642784,-0.1905061454,-0.0001221336,-0.2679699361,-0.2092573792,-0.216757074,-0.2778628469,-0.4542639256,-0.0103372186,0.0486837067,0.0854049921,-0.0741245449,0.2338189036,0.2066883296,0.2127250135,0.0140997944,-0.0412977263,-0.2158882469,0.0728689283,-0.2318145186,-0.1928956658,0.0099314637,-0.1537414342,0.3478007019,0.1561032534,-0.4876750708,0.3729537427,-0.0741802976,-0.271235466,-0.1000492945,-0.0658702105,0.3611463308,0.1052038521,-0.2213782966,0.1072303504,-0.1624223888,-0.0407600775,-0.2125340253,0.2935686707,-0.3540355861,0.6404054761,0.1446554661,0.8162136078,-0.0796792433,0.0347614996,0.1796438992,-0.0052892175,0.0840145051,-0.1534436494,-0.2790311277,0.2999058366,-0.0641298592,-0.1879597008,0.2040009797,-0.0071196677,-0.3185498118,-0.0618811585,0.0575089678,-0.2475688607,-0.3978370428,0.2112483382,0.150199607,0.2670522034,-0.0949291959,0.0659341067,-0.1520652771,-0.093332313,-0.1404843181,-0.0002467613,-0.239154771,-0.0435561277,-0.3146158755,-0.0192615539,-0.5928471684,0.0117929587,0.0448497497,0.2776107788,-0.0176498666,-0.0586582907,0.0439761095,-0.123512648,0.7291725874,0.1925560981,-0.0641700327,0.053502474,-0.2891407609,-0.1317944825,-0.07199765,-0.2500882745,-0.0896914899,0.2495447993,0.1464287639,-0.0683418736,-0.0660945475,-0.0443850122,-0.1496639252,-0.0953735411,0.1433293372,-0.0345514938,-0.3247775435,-0.3178597391,0.4709649086,0.2082360238,0.2264200449,-0.0420390405,-0.2081307918,-0.0624823384,-0.1810879558,0.1463857442,0.0344937928,0.2394464016,0.103066735,0.1047032624,-0.2255038023,0.1473952532,0.3487256765,0.0569934584,0.2948224247,-0.2132103741,0.0019884591,0.1858615726,0.2423778772,0.3444125354,0.1217527166,0.3838116229,-0.1805278659,0.3563895226,-0.0233826134,0.5706766844,0.1339718997,0.2940819263,-0.5356917977,-0.1441344917,-0.0733003914,-0.0229516141,0.1358525604,0.3842638433,-0.2950781882,-0.4676644802,0.2749304771,0.0158594567,0.8640151024,0.0037842533,0.2327529788,0.227300182,-0.1183483601,-0.0840130523,-0.470031172,0.1709436029,-0.1486387402,0.02234363,0.0731078982,-0.0175683461,0.1096459031,0.052021727,-0.2435223758,0.0195088591,-0.1968875229,0.0436733328,-0.1842474937,0.0784862936,0.152348727,-0.229408294,-0.1582888365,0.3006845117,0.0136540011,-0.0153820841,-0.1721973121,0.141670987,0.1015123203,-0.195558846,-0.1096817926,-0.0296464786,-0.6171456575,-0.0148989251,-0.0467374884,-0.4415018559,-0.0558963493,0.1309693456,0.2249736786,0.3476777673,-0.0277321786,0.4075202048,0.1552727073,-0.0709672123,0.0084357942,0.0882745087,0.213149026,-0.1463827789,-0.2049345225,0.2800983489,-0.0239803642,-0.2458442003,-0.1298459023,-0.0052154092,0.0893291384,-0.3025335371,-0.1445054561,0.1670951545,-0.283595413,-0.1976293027,0.262966305,-0.1794404238,-0.3451642394,-0.0516928732,0.2241318673,-0.1548503488,-0.1165103316,0.1556385905,0.1142624766,0.0887870342,0.4982994497,0.0722691864,-0.1852856129,-0.337631464,-0.0996500999,0.3330563903,0.2025566697,0.1904195696,0.1405813545,-0.1786846071,0.0843879208,0.2934961021,-0.0304424446,0.1074917763,-0.0709209815,-0.1166767031,-0.5866552591,-0.0473163016,-0.0656330734,0.0548392721,0.0482311249,0.6134432554,-0.0489161499,0.0141096581,-0.4224278629,0.1629427224,-0.4016977847,0.3556700051,-0.338485837,-0.2318286151,0.0710265636,0.1506133527,0.2422724813,0.3419395685,-0.1139192581,-0.4090699852,-0.1222184747,0.0248394143,0.1291023642,-0.139543578,0.1018724665,-0.0636721998,-0.1137402803,-0.2677372992,0.1700218171,0.2002907693,0.0035485604,-0.1849630773,0.002073423,0.2846563458,0.0880323946,-0.0798809081,0.073350355,-0.0980495363,0.2036529928,-0.0998681858,0.1082951352,0.0222571958,-0.3535291553,0.0179361049,0.0953891352,0.1707889587,-0.0110190315,-0.1538038254,-0.2709652781,0.058976274,0.0929035097,0.1874949336,-0.1579524428,-0.0051905862,0.602673173,0.3695806563,-0.2009856254,0.1199315786,-0.093357943,-0.2630318403,0.1861790717,0.1145235524,-0.2968789339,-0.1913561523,0.0389554948,0.26869753,-0.0513367839,-0.3590491116,0.1856439859,-0.1367128342,-0.1490940601,0.0012126168,0.0390715748,-0.0112079857,0.2023048103,0.4113551378,0.0242128763,0.2234200984,-0.0093104169,0.0759712383,-0.2070939839,-0.1292761713,-0.4814424217,0.1469446123,-0.1471955925,0.1422041506,-0.1822510511,0.0924170688,0.0629551113,-0.3733498752,-0.228403911,-0.0257234965,-0.499127537,-0.1693498194,-0.1390725821,0.093276538,-0.0430417284,0.1080241129,0.227123335,-0.2272856981,-0.0021611617,-0.0101463608,-0.2188068479,-0.0737400949,-0.0802897289,0.2493526787,0.0023890834,-0.241336748,0.1899193525,-0.1874287724,-0.1118924916,-0.16120556,0.044636108,0.3779862523,0.3604880869,-0.0842973739,0.0188738648,-0.1190733165,-0.0572737716,-0.2295381129,0.1885077357,0.5024799109,0.3665912151,0.0358611904,0.2906791568,-0.3719209433,0.0271870624,0.2588149309,-0.1794044077,-0.1993596256,0.1815936267,0.1272799671,0.0256972965,-0.323984772,0.010772408,-0.4952422678,0.1463080347,0.1886751205,-0.0784363896,0.1943069845,-0.107389614,0.1361709982,-0.1501721144,0.3536183536,0.0934941247,-0.0816853717,-0.2944380939,-0.2529091239,-0.2216197401,0.002709745,-0.2570445538,0.261566937,-0.2474881858,0.1016448364,-0.1807212383,0.0619049035,-0.2251426876,-0.181820035,-0.1386444271,-0.1433751136,-0.4553714395,0.0008999794,-0.333802551,0.2049721777,0.2946285903,-0.2715359032,0.1546801329,-0.2045429051,0.1988458782,0.200481087,0.122614339,0.0814024657,0.0494868793,0.2690639198,-0.1180238947,0.24882631,0.0066958433,0.234377116,-0.1272746325,-0.2208383977,-0.1644573957,-0.1328837574,0.2034689635,0.3435234725,-0.1512968093,-0.1147878394,-0.1901462972,0.0819561034,0.2690295577,-0.0918128639,0.1933837384,0.2084004134,0.2739050686,0.1507564485,0.0112295505,0.3407836854,-0.0216361005,0.0838732943,-0.364559412,-0.316702038,0.5697761178,-0.3233434558,-0.3560694456,0.0281231571,0.0793672055,0.3703938723,-0.103385523,-0.6004105806,0.0551448837,0.0919704288,0.0242193751,-0.4487530589,0.3624128997,-0.2732500732,0.134933874,0.0062522916,0.5045967102,-0.0294561759,-0.1434838474,0.1576818079,-0.2444412708]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/757","title":"CUDA out of memory","comments":"`RuntimeError: CUDA out of memory. Tried to allocate 954.00 MiB (GPU 0; 15.90 GiB total capacity; 14.35 GiB already allocated; 753.75 MiB free; 14.39 GiB reserved in total by PyTorch)\r\nException raised from malloc at \/pytorch\/c10\/cuda\/CUDACachingAllocator.cpp:272 (most recent call first):`\r\n\r\npart of error output","body":"In your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n","comment_length":44,"text":"CUDA out of memory\nIn your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n\n`RuntimeError: CUDA out of memory. Tried to allocate 954.00 MiB (GPU 0; 15.90 GiB total capacity; 14.35 GiB already allocated; 753.75 MiB free; 14.39 GiB reserved in total by PyTorch)\r\nException raised from malloc at \/pytorch\/c10\/cuda\/CUDACachingAllocator.cpp:272 (most recent call first):`\r\n\r\npart of error output","embeddings":[-0.2118009329,-0.1436042786,-0.1251238734,0.3756681681,0.4871861339,0.0221625883,0.2372636348,0.1879336387,-0.0591316633,0.3615151942,0.203222096,0.0899497569,-0.1103607565,0.0615564361,-0.0812174976,-0.2425248623,-0.1598561704,-0.0211354885,-0.2000565976,0.0185324159,-0.0547572672,0.0996781886,-0.2718253136,0.0186517034,-0.4758870304,-0.2184392959,0.0810014233,0.1155068204,0.1469694525,-0.2040057778,0.0754881054,-0.259085238,0.0466847345,0.5353175402,-0.0000942944,-0.0917870924,0.0012609041,-0.0493776798,-0.1668550968,-0.0935569406,0.2256605923,-0.0973295197,0.169995144,-0.3079277277,-0.2159549147,0.1849445999,-0.1372168511,-0.2383294702,0.2873457968,0.4498963654,0.3909004629,-0.0431755222,0.0861904696,-0.1659832597,-0.0875586346,-0.1732179672,-0.0146951824,0.0005477721,0.138158083,-0.0281157382,-0.003598714,0.1856382936,0.0475580357,0.0354331508,0.0543559715,0.0449393056,-0.0404999256,-0.216718182,0.0772738829,-0.0306700766,0.11783804,-0.2315409631,0.0958704725,-0.0387620963,0.0093952063,-0.413179785,0.1378812343,0.1859952062,-0.3922763169,0.0164076183,-0.2586274743,-0.0303190649,-0.0934135765,0.0631963611,-0.1956590414,0.0080889892,-0.1182058156,0.0632062629,0.2467140555,-0.0173122883,0.2066463828,0.091149509,0.0142326057,-0.0221963245,-0.282905221,-0.0232094135,-0.1080686375,-0.3290325403,0.0293364115,-0.1882707328,0.2276498377,0.00415518,0.1609884501,0.0526035205,0.1466838717,0.3722299337,-0.3857885301,0.0974769741,0.0332972854,0.0865174532,-0.3276966214,-0.2392992377,-0.0550302528,-0.3216640651,0.0693228245,-0.0124912187,0.0669356212,-0.132598713,-0.1062621623,0.2111896425,-0.2619428933,0.2171588987,0.1159809604,0.2579928935,-0.140743345,0.1431768537,0.1726024449,-0.0793246031,-0.1121673137,-0.0890850127,-0.3556940854,0.1650259197,-0.3696238399,-0.0468623042,0.1952004582,0.0122246621,0.3212688863,-0.1181160659,0.202634424,-0.4704040587,0.0320863761,-0.3434745371,0.0106649967,0.0647375137,0.1504801065,0.0602492169,0.0967096463,-0.0062753516,0.2173627913,0.3751479089,-0.2610060871,-0.3041603863,-0.0635333136,0.3546237648,0.2509842217,0.119456321,0.0315395556,0.0001186078,0.2730512023,0.0621277541,0.0361014679,-0.1540497541,-0.0545561686,-0.1836684197,0.3471758962,0.188801229,-0.2707559168,0.0562871993,0.219892785,0.0580805875,0.1811884195,0.4126347303,-0.3015066087,0.1460334212,-0.1990502477,-0.1652383357,0.2115194798,-0.1684204489,-0.2656309605,0.0336978994,-0.0929884985,-0.1253866553,-0.195714578,0.0292374343,0.299109906,0.223430112,0.2497170269,0.1406296939,-0.256942302,0.0725273117,-0.3540612459,-0.1874581575,0.1888196319,0.0551340654,0.3159446716,-0.0105230073,-0.2614680231,0.1066229418,0.1901454926,-0.0579619519,-0.0212716665,0.1260482669,0.3186985552,-0.0929767191,-0.0987953618,-0.0475212559,0.0558381975,0.0396544635,0.066314213,0.1411865354,-0.1820722669,0.004149775,-0.0556848384,-0.0952627361,0.0390261151,-0.2896197736,0.3484506011,0.057923045,-0.1184399426,-0.0052982415,-0.0020452042,0.3104230762,-0.4857845008,0.0727449581,0.0900419578,0.17615062,-0.2769615352,-0.2281685323,0.0018621,0.1827633232,0.1517068446,-0.0721077621,-0.1934701204,0.2518944144,-0.0209872313,-0.2037340254,-0.1180720702,-0.0767886937,0.1183430925,-0.0845875293,-0.0492652506,0.0054363911,0.1151401699,-0.03707885,0.1277406514,0.0058230055,0.1121657416,0.0126649579,0.0891129151,-0.0318649225,0.1537890434,-0.0675355941,0.3522224426,-0.036048077,0.0746686235,0.2333031744,0.2196899951,0.0484552234,-0.3491225243,-0.0526744314,0.3689388335,0.1791172028,-0.0128230657,0.2499065548,-0.1032726541,0.0485810079,-0.0381968692,-0.1489265114,-0.0114671318,0.2289577425,0.0813973173,0.1702391207,-0.0014853079,-0.1016810462,0.1454582661,0.1066214219,0.232187286,-0.119015418,0.0962334499,-0.034775693,-0.2055913657,-0.1354940832,0.2554323077,0.4957813025,-0.181113556,-0.0380417854,-0.1817548573,-0.1290147156,-0.1607402414,-0.0649124607,-0.0140387565,-0.0476050861,0.0061219838,0.2428815961,0.0315907039,0.3472577035,0.1770738065,0.2584714592,0.121574305,0.152034685,0.2021979392,-0.2745970488,-0.1216749623,0.2975054681,0.2598125637,-0.0124936262,0.1879578084,0.168282941,0.0326671712,0.1049642041,-0.0073334319,-0.1140931472,0.0201444961,0.1777545959,-0.2008757293,0.1997723132,-0.1199362651,-0.2861014605,0.2192390561,0.0083032688,-0.1159449965,-0.0851343721,0.1891164929,-0.1369484961,-0.1508421749,-0.2441036254,-0.113745153,-0.359675616,-0.0707698762,-0.0618628114,0.120352298,-0.1065524518,0.3134255409,0.382746309,0.0811476856,0.0531449839,-0.1908714473,-0.1974184364,-0.0866251662,-0.1350986958,-0.224124074,0.0129015716,0.1013907343,0.4695522785,0.015268201,-0.6003519297,0.0291695036,-0.0662457421,-0.0540731959,-0.0895547271,-0.1722446978,0.4207229614,0.0196295939,-0.2582046092,0.1793518513,-0.1991150528,0.060993474,-0.1229906529,0.1664449573,-0.245826304,0.615635097,0.2445935607,0.5803915858,0.0293949954,-0.1628429443,0.1924736649,-0.1467560381,0.14566423,-0.2136935145,-0.337397486,0.3285893202,-0.2230758071,-0.3722511232,-0.0171127543,-0.2398682833,-0.2466320992,-0.2296698689,0.0987703353,-0.179925859,-0.2937253416,0.1986032128,0.0552717,0.1808993965,-0.1025265083,0.0867984444,-0.0677145571,0.0408196189,-0.0348983444,-0.1857337505,-0.0566520356,-0.1185684428,-0.1725615561,-0.1408476382,-0.2354017496,0.1992799938,-0.0753855407,0.2314153761,-0.1360682845,-0.009563488,0.0642503351,-0.0935910642,0.5463860035,0.4751864672,-0.2689293027,0.1434053183,-0.3233285546,-0.0171127543,-0.0441883355,-0.1852470487,0.1693897098,0.2353733778,0.1539678425,-0.0898829922,-0.0436335839,0.077950336,-0.0135362027,-0.1640754044,0.0817308426,-0.0410620086,-0.166617021,-0.2393947542,0.3663626015,0.068989709,0.3136427999,0.0140892109,-0.2847580016,-0.0850646421,-0.3089558482,0.0584318228,-0.0208028667,0.3193124235,0.0104775419,0.1350441724,-0.2640307844,0.2954733372,0.138093248,0.0301283617,0.1311089993,0.231735006,0.1520993114,-0.0048916307,0.3407326937,0.157431066,0.002329044,0.0391832665,-0.0219140109,0.3902660906,-0.1159143075,0.333533138,0.2773242295,0.214095071,-0.358030498,-0.2283451557,0.1492249817,0.0847178325,0.1235629469,0.5383540392,-0.0797040313,-0.2866714895,0.0895337015,0.2135761231,0.8268972635,-0.1299597174,0.1022189111,0.2026884258,-0.165912956,0.2399533689,-0.0622143671,0.232943058,-0.214916423,-0.1393241286,0.0242631789,-0.0390883386,0.1224852353,-0.0461100042,-0.2585179806,-0.0151257124,-0.2253760248,-0.0566122159,-0.115311265,0.0038393377,0.2276581675,-0.2236818373,-0.0902879313,0.330152154,-0.1999340802,-0.1413722187,-0.0961847007,0.1016440168,0.2570006549,0.1042393669,-0.344731003,-0.0503953621,-0.5622708201,0.0710705742,-0.2664097846,-0.4682614207,0.0003383165,-0.0283034109,0.212062031,0.2580199838,0.0051964619,0.2802524269,-0.0005394026,-0.0636644736,0.003279377,0.2407024205,0.4479750395,-0.1849565506,-0.2873692513,0.5506269336,0.0019773252,-0.2876962721,0.1027586237,-0.0076731076,-0.0647220835,-0.3952232599,-0.3604306579,0.1480996758,-0.1487227529,-0.2114571035,0.3057723641,0.1042609364,-0.2607138455,0.0723363683,0.1013842449,-0.2181289643,-0.1287758201,0.155597195,0.2336168289,-0.0123213781,0.4842815697,0.2296358347,-0.1145039275,-0.2515179515,-0.2586404979,0.3390387595,-0.0731497481,0.1311484724,-0.0318765007,-0.0501877703,0.1253538579,0.3023049533,0.1005005017,0.2856651247,-0.1165959612,-0.0184924807,-0.707621336,-0.0653294548,-0.060025502,0.2186035514,0.0235087741,0.3910016418,0.056273412,0.0909727067,-0.4736708999,0.1770469844,-0.5639112592,0.2904718518,-0.2566624582,-0.2054287791,0.0416059755,0.0040291618,0.2464370131,0.4243367612,-0.0476634018,-0.4267258942,0.0216061883,-0.0419056937,0.116637744,-0.10296157,-0.0304950904,-0.2445813268,-0.1433236003,-0.1361955702,0.1577349454,0.1895408034,-0.047686521,0.1194986701,-0.1096070558,-0.069539316,0.0237499867,-0.1213802546,0.0675211996,0.1495750993,0.3131262958,-0.0734327435,0.1248154715,-0.0373871848,-0.4617183805,0.0833917186,-0.099327147,0.1193157509,0.0694814026,-0.3183685541,-0.2227094769,0.1084955037,0.1746730953,-0.0177029148,-0.1391636729,0.01642753,0.3587087095,0.4299487174,-0.2469257265,0.2003047764,-0.1443239897,-0.1612069905,0.1398003399,0.12167456,-0.1282646954,-0.1331651658,0.0532888584,0.2242798507,0.0956679583,-0.3333975077,0.0906544104,-0.0894696489,-0.106448181,-0.1372927129,0.122179687,0.1889846921,0.25471735,0.493703723,-0.1295205951,0.2689762115,0.0964153558,0.0695051253,-0.0760655403,-0.1202311441,-0.2427157611,0.1334981024,-0.189061895,0.0296563804,-0.3137652874,0.1912958175,-0.1022826135,-0.425188005,-0.1854676604,0.1470761746,-0.5079362988,0.136041075,0.0604731925,-0.1074742451,0.0550127104,0.1124255359,0.1630518585,-0.2354062796,0.2463978082,0.0221468247,-0.3643345535,0.0353302099,-0.0773712769,0.2141335309,0.0386217423,-0.1875312179,0.307616353,0.1620953232,-0.223695457,-0.1291543394,0.2275910974,0.3513897061,0.3235750496,-0.0998087972,0.1433091164,0.0217639934,-0.0817323476,-0.0720934868,0.0789101571,0.4305180013,0.2887809873,0.1629038453,0.3104507327,-0.2854378819,0.1994967014,0.2230198979,0.1595558375,-0.3771579862,0.1278299242,0.0198392533,-0.1377460808,-0.2737796605,-0.0245922413,-0.3497669101,0.1815815568,0.1511025727,-0.2821189165,-0.0140581103,0.0394457281,0.1760407537,-0.0886111185,0.5066301823,0.1213658527,0.0465274118,-0.4374958873,-0.3422831595,-0.283996731,0.1650443077,-0.1825189739,0.2564493716,-0.0810848102,0.1938932985,-0.0959538221,-0.1081768349,-0.0486624874,0.0746900886,-0.0253950115,-0.0910334662,-0.4974974096,-0.0855910927,-0.3642121851,-0.0255061705,0.18548356,0.0089821937,0.156437546,-0.2137868404,0.2555188537,0.148517251,0.1038045809,-0.0307189897,0.2767020464,0.3055878878,-0.1157960445,0.1302896142,0.2141691446,0.1882017851,-0.0287711397,-0.217326805,-0.0955248624,0.0016783852,0.3564402163,0.2714662552,-0.1065699533,-0.0464790389,-0.2048802674,0.2488592118,0.1599502265,-0.2749106288,0.1392546594,0.018373752,0.2895314991,0.0470945276,0.3179522753,0.4763450325,0.0271086395,0.0310970768,-0.3615962565,-0.2376406938,0.5043510795,-0.0995829776,-0.3100988567,-0.1498211026,0.0823993683,0.2493338287,-0.1075268909,-0.7183137536,-0.1950536966,0.1931870878,-0.0899090245,-0.3274354041,0.1808322668,-0.2372242957,0.0768355504,0.1003223285,0.3857816458,-0.0535829291,-0.10840334,0.0701345056,-0.2166676819]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/757","title":"CUDA out of memory","comments":"from funnel model to bert model : error still happened\r\n\r\nfrom your dataset to LineByLineTextDataset : error disapeared","body":"In your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n","comment_length":18,"text":"CUDA out of memory\nIn your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n\nfrom funnel model to bert model : error still happened\r\n\r\nfrom your dataset to LineByLineTextDataset : error disapeared","embeddings":[-0.2429937124,-0.0383692868,-0.0662143826,0.19402197,0.5733138323,-0.096849665,0.3483140171,0.0940258205,-0.0914314687,0.4046832919,0.0462234057,0.0135813393,-0.0812719017,0.3208122551,-0.1960563958,-0.3946526647,-0.0328095481,0.0102433218,-0.1957312971,-0.0121695027,-0.0123564731,-0.0033930612,-0.1592791975,0.0870993361,-0.1410438865,-0.03772806,0.002121479,0.0090349838,-0.1163757145,-0.2628358305,0.0889436752,-0.0642799363,0.1672817916,0.3281091452,-0.0001007679,0.1339539886,0.1042969301,-0.1116643324,-0.2635507286,-0.1729589105,-0.1789835095,-0.1399423927,0.2102268338,-0.2823681831,-0.1664161384,0.4420573413,-0.0516923405,0.03373098,0.5577896237,0.4128130376,0.335689038,0.1212375909,0.1510000527,-0.2648308277,0.2176375985,-0.0050718458,-0.0425598994,-0.040778853,0.1038939878,-0.0698297545,-0.0774987489,0.3561424315,0.1329948902,0.1485108882,0.1547268331,-0.0627807453,0.1188859418,-0.2141754925,0.1081306785,0.1221961603,0.1534587294,-0.1914740056,-0.0599850528,0.0532139763,0.0153939854,-0.3182945848,0.1467740089,-0.0166956671,-0.3548562527,0.0946364105,-0.1052352041,-0.2175528854,-0.1025776491,-0.1140932888,-0.1665934473,0.1490675509,-0.03719173,-0.0587535836,0.2903490067,-0.1254053712,0.3205078244,0.1537186056,0.0427101962,-0.0738143772,-0.2277662158,-0.1170416698,-0.0533464141,-0.1595803052,0.0571266636,-0.3830276728,0.2500422895,-0.1162567958,0.1959917843,0.035354618,0.3170464039,0.286280036,-0.3701386452,0.2391137481,0.0503853895,0.1746375561,-0.3613769114,-0.2272451073,-0.1112045571,-0.3267977834,-0.0800593197,-0.0327487327,0.0792790577,-0.1619876474,-0.2722689211,0.1725315154,-0.254881978,0.2640634179,0.2559896111,0.4726040959,0.0060154749,0.2089016885,0.132183224,0.0104280077,-0.0713887066,-0.0696887374,-0.2688013911,0.0624106675,-0.3850221336,-0.1537840962,0.1105828807,0.0035166475,0.2379930913,-0.1026343405,0.0520373024,-0.1890207976,0.0747663528,-0.3209008873,0.0013835065,0.1301466972,0.3464878798,0.0360809378,0.020009907,-0.0249867551,0.2250242531,0.2758370936,-0.4562953115,-0.3984077275,0.0950752571,0.2998204231,0.3816589415,0.0991173759,-0.1765723526,0.1599657834,0.1963942051,-0.049138803,-0.0305246413,-0.2006733567,-0.0990436524,-0.1039260626,0.3158841431,0.391621232,-0.4216264486,-0.1140593961,-0.1328940541,-0.0142196221,0.3896189034,0.5055039525,-0.4047174752,0.1558742225,-0.0543605424,0.0269637015,0.3355779648,-0.273036778,-0.2073615491,0.1021931693,0.0825998113,-0.0646452382,-0.2675142288,-0.2481892258,0.3915963769,0.0264495462,0.250297159,0.0818310231,-0.2580812573,0.1815284491,-0.2347031683,-0.0493548736,0.1143747196,-0.0836780816,0.2223152965,0.0171532445,-0.2288851738,0.3926403522,0.0965137258,-0.1182944849,0.1400403231,0.2874973416,0.337926209,0.0090108616,-0.0741014928,0.1534021348,-0.0264252778,-0.0433663316,-0.0966737047,0.0782344192,-0.0732004866,-0.1002274454,-0.0470795892,0.0761650726,0.0329090543,-0.2485426068,0.3355541229,0.0063603483,-0.245523721,-0.071964182,0.0238838587,0.0617828108,-0.4273528457,0.1283402741,0.253841579,0.1770500988,-0.3252508342,-0.0646981448,0.0731215477,0.1953135729,0.1215332672,0.0226774067,-0.1938195527,0.2192295194,-0.0949145108,-0.0056969658,0.1321394742,-0.2101243883,0.0761880875,-0.1818877459,-0.1388998777,-0.0490953922,0.2414444685,0.0396053456,0.0389331654,-0.0427992903,0.2721196413,0.1030704603,0.27469486,0.0665912852,0.3181394339,0.030663345,0.2861680984,-0.1076054946,0.1705531627,0.1305482984,0.3169270158,0.0225526094,-0.341042608,0.0226457827,0.4099620283,0.2349964529,0.0324318856,0.2992371917,-0.1192656383,-0.1215240508,-0.056027934,-0.0252685305,0.0644885451,0.1640000939,0.1957211643,0.0184660032,-0.0664573759,-0.1216934174,0.2433014661,0.094361268,0.0452732779,-0.1018354744,0.177628085,-0.1306377202,-0.2493300736,-0.0861566141,0.2323688567,0.4989029765,-0.2284733653,0.062990576,-0.0876248479,-0.0317607522,-0.1993570775,-0.345600605,-0.0594019629,-0.0510536805,0.1763631999,0.2323435545,-0.0868498236,0.3003810048,0.3504382968,0.291087091,0.0862325281,0.3323001564,-0.0694113746,-0.2761394382,-0.1076880619,0.2419588864,0.291364938,0.0424761139,0.0451027155,0.0642616153,0.06966912,0.0128549114,-0.1309442371,0.0154970288,0.0609394014,0.2333246171,-0.00944856,0.3017679155,-0.2326458097,-0.3526659906,0.2210343927,0.0122787775,-0.2054139525,-0.0669541359,0.0479299575,-0.1271324009,-0.1160670593,-0.2654438615,-0.1898758858,-0.4146497548,-0.1663347185,-0.1074047536,0.0727820992,-0.273742348,0.2912783921,0.3617987037,0.2805692554,-0.010073415,-0.0860195458,-0.1375756711,-0.1062793583,-0.2191572338,-0.2372312248,-0.0587483384,0.1296009719,0.3659161031,0.0628581122,-0.4608674943,0.2224972993,-0.0445925705,-0.3251991868,0.033326745,-0.1258550584,0.3331925571,0.1592571884,-0.2517319024,0.0018177774,-0.0933884457,0.0826418847,-0.2295024395,0.2506913543,-0.4444752038,0.6022287607,0.0526815727,0.8941935897,-0.1279267222,-0.0536749102,0.3798964918,0.0599210672,0.1554176509,-0.3653232157,-0.2811015844,0.2810940742,-0.1616919041,-0.2178434134,0.0793063939,-0.212458089,-0.3545593917,-0.2142681479,0.0506604202,-0.1614798903,-0.3749134541,0.1603234708,0.1664025187,0.2591389418,0.0793324038,0.0330849253,-0.0824036896,-0.0262369905,-0.0618656464,-0.0066177277,-0.0404127501,-0.1220665276,-0.0217582006,-0.2404593825,-0.525493741,0.2049487382,-0.036858283,0.126574263,-0.210607186,-0.0097335018,0.0357164815,-0.2114922702,0.606128633,0.2060275376,-0.2061778009,-0.045147147,-0.1212493405,-0.0043879407,0.0452690795,-0.2596739233,-0.0493762568,0.224839583,0.1728409827,-0.1016692966,-0.0797531754,0.0945252404,-0.1879914105,-0.1370821744,0.1349076331,-0.0498633124,-0.0842896625,-0.313347429,0.3950281143,0.07743036,0.1835727245,-0.0899285004,-0.2849116921,-0.1212120876,-0.1138361916,0.0061074584,0.0708377063,0.2786101699,0.0029742657,0.152318567,-0.2993784845,0.3828773797,0.5157994628,0.1190181226,-0.0874172673,-0.1902525425,0.1858585477,0.1249073669,0.1454904079,0.2861873209,0.125754118,0.1824540347,0.0727723911,0.2340397239,0.1402717084,0.463852495,0.2891491055,0.2993736267,-0.4249512553,-0.1256252676,0.0856915265,-0.0340138227,0.0835330412,0.2930764258,0.0490733087,-0.3022394478,0.324385941,0.0870444626,0.8639284372,-0.0792955607,0.1802265644,0.1918712258,-0.170998022,0.1523243338,-0.093704477,0.1080510989,-0.1515615284,-0.0322037004,-0.0175327305,0.0499060415,0.0376557112,-0.0029825277,-0.3940147758,0.0134876659,-0.2069327086,-0.1053629741,-0.2126814872,-0.0633785054,0.2877331376,-0.1858210266,-0.2617961764,0.2688200474,0.0012025917,-0.0332702622,-0.1494913697,-0.1048860028,0.3056617379,-0.144726783,-0.2445027381,0.0299047735,-0.5581154823,0.027568657,-0.1809765846,-0.5129506588,0.1047138348,-0.0081554996,0.297390759,0.1603412926,-0.1886995882,0.4027428925,0.1041323245,-0.0309403408,-0.058482863,0.2015679926,0.3738806248,-0.200485751,-0.209376663,0.3739481568,-0.1468416452,-0.2974852324,0.0071604145,-0.0817929208,0.2527705431,-0.3473982811,-0.2459544092,0.2132873237,-0.1670848429,-0.2638564408,0.2789062262,-0.1355575621,-0.183731541,-0.0756056458,0.0795594677,-0.1285527945,-0.094704777,0.0497826338,0.0582109205,0.2909926474,0.5716347694,0.1200144291,-0.1638103276,-0.2546852529,-0.2172007412,0.5439648628,-0.0571184382,0.2054128945,0.0880149528,-0.1192775965,0.0942577571,0.3280457556,-0.0314899646,0.1516571641,-0.0771693289,-0.0861444175,-0.6292020082,-0.0080254795,-0.1333273053,0.1587743461,0.0317225903,0.3995530009,0.1359015107,0.0740636364,-0.4038144946,0.1854001433,-0.4803616107,0.2574853599,-0.3123631179,-0.2211485356,0.1500566602,0.0183973853,0.1983171552,0.3915615678,-0.0878051743,-0.3881203532,-0.0516212955,0.0129222656,0.0536250249,-0.0444266163,-0.0242950972,-0.0771126822,-0.1646130681,-0.1381662637,0.1605730057,0.1419165134,-0.0070240092,-0.1002926007,0.0486867614,0.0814610049,0.0218538456,-0.0666425601,0.3228673041,-0.0471478552,0.0425882861,-0.0203962661,0.0444233678,-0.1182852462,-0.3260061443,0.0615037903,0.071949102,0.2459683269,-0.0139943473,-0.1344070435,-0.1172321364,0.0094733657,0.2008037418,0.0664601475,-0.0816580802,0.0826825947,0.5198850632,0.354341656,-0.3999358416,0.1367903203,-0.1783644855,-0.2301891297,0.0403737761,-0.061636705,-0.2664192319,0.0175928567,0.1138856113,0.2983406782,-0.2819608152,-0.2614426315,0.0145400483,0.0535264239,-0.2307027429,0.0569043383,0.0776648745,0.0807906166,0.2725352049,0.5964998603,-0.088912189,0.2089633346,-0.0151839983,0.0499990247,0.0881484151,-0.2211844325,-0.4366448224,0.1212514266,-0.1020080075,0.1862652004,-0.2001481205,0.2823794782,-0.0579351708,-0.5546742082,-0.2274595499,0.1214079261,-0.4557276666,-0.2373722047,-0.0775249973,-0.0684665516,0.1123583913,0.1091134548,0.0812029541,0.0438422933,0.2696995139,0.0330019668,-0.1880842745,-0.0353125967,-0.0883558169,0.3132316172,-0.0430991873,-0.1455016136,0.3132103682,-0.0304947384,-0.3657895327,-0.0081723845,0.1843430549,0.3374364376,0.3683411777,-0.0031501129,-0.0410465188,0.069115445,-0.130632177,-0.117127113,0.0389338881,0.6038047671,0.3613379598,0.1704288274,0.2067405134,-0.2698080838,0.2160950899,0.1742768586,0.2378054559,-0.2703260183,0.1327986866,0.0880890191,0.0074963816,-0.3903089762,0.0345643796,-0.4844126403,0.2789462805,0.105946444,-0.2162210345,0.007046367,-0.0284634363,0.1326556206,-0.0766890571,0.4783808887,0.140668869,-0.081765838,-0.4501723647,-0.2213836163,-0.3204562962,0.0801499486,-0.199378714,0.496186763,-0.1312596351,0.1748059392,-0.1713673323,-0.0546035171,0.1064721122,-0.1912437528,-0.2039954811,-0.0729133636,-0.5115637183,-0.0568443276,-0.4868826866,-0.0484146029,0.2784799337,-0.1454190314,0.2637949884,-0.0652624145,0.1779109091,0.2398308516,0.137320146,0.1342184991,0.2333138436,0.2523612976,-0.1193414479,0.2122927159,0.084531121,0.1872464418,-0.0177744683,-0.3345736265,-0.1654559076,-0.1604999453,0.3346556425,0.2071069628,-0.2151176482,-0.151447475,-0.2727840245,0.0322611183,0.169640407,-0.2407682091,-0.0566637255,0.0915061384,0.3267953694,-0.0081630256,0.1193371862,0.3521054387,-0.0016530509,-0.019166708,-0.4411460757,-0.2393814921,0.6260103583,-0.1650120765,-0.2551419437,-0.0921275243,-0.0154952444,0.436137706,-0.1941234469,-0.9336695075,-0.076051712,0.0727671981,0.0200701393,-0.438741684,0.0850871503,-0.1784355789,0.2450280488,-0.0705038309,0.4420758188,0.0784842819,-0.219212085,0.2160793096,-0.1709849089]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/757","title":"CUDA out of memory","comments":"notice i just loaded 1000 rows of data","body":"In your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n","comment_length":8,"text":"CUDA out of memory\nIn your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n\nnotice i just loaded 1000 rows of data","embeddings":[-0.3812152743,-0.0389198512,-0.1876501441,0.3138402104,0.60308218,-0.0071777315,0.415705353,0.2851212323,-0.0135605745,0.3824834228,0.1306770891,0.0962304473,-0.0210618414,0.2401799113,-0.0979630426,-0.1301741302,-0.0960947499,0.067211099,-0.1696237326,-0.009127005,-0.0049824486,0.0544702001,-0.2367642522,-0.0888629034,-0.4131459892,-0.0563066415,0.0885530189,0.1063596234,0.0912219808,-0.1915098876,-0.03468493,-0.072201632,0.009956562,0.4099596441,-0.0000930291,0.0642284676,0.0822571144,-0.1069153026,-0.2179498374,-0.0453921817,0.0519016087,-0.0985587388,0.2327310443,-0.3425115049,-0.3222614825,0.1839506179,-0.0925688893,-0.155229792,0.4169371426,0.4358083606,0.4207719862,-0.0068016695,0.0106122615,-0.1649034619,0.0886292532,-0.0599356554,-0.0249394011,-0.038050998,0.1654275209,0.0055128047,-0.1050002575,0.1898452044,0.0719715059,0.0555040501,0.049591694,0.0219080001,0.0157662705,-0.0785030499,0.1501797587,0.0198351145,0.1874594986,-0.0836565122,0.0776007101,0.0230194535,-0.0112400083,-0.4555846155,0.1876895875,0.099816747,-0.2839977443,0.0938002542,-0.1378044337,-0.076225698,-0.0837320015,-0.0711060464,-0.2140318453,0.0676838383,-0.1265358031,-0.049638059,0.2462563664,0.0179061648,0.215514347,0.1944238096,0.0709191263,-0.0338374078,-0.262978375,0.0166493654,-0.0389961042,-0.1950876862,0.0725326166,-0.1784534305,0.2672151029,0.1208723858,0.1570736766,0.0355736651,0.0661742762,0.2700311542,-0.2264651358,-0.0726892054,0.1083563715,0.2026180029,-0.4128457904,-0.1994316131,-0.0856676698,-0.2492730618,-0.0554843061,-0.0136892684,0.0047030221,-0.1533468813,-0.1447552294,0.215988934,-0.1839341223,0.1242910922,0.0215021838,0.3472186029,-0.0895308033,0.093908608,0.1507825851,-0.0732548758,-0.1090597808,-0.2107584924,-0.338303715,-0.0016872557,-0.2765048444,-0.0363508649,0.252854526,-0.065810211,0.3705449104,-0.1299691349,0.0602897368,-0.3532404304,0.0659679696,-0.2342962474,-0.0355752222,0.0809512287,0.2552649677,-0.0448413156,0.0494249463,-0.1252025664,0.2894486487,0.3651044667,-0.2126658857,-0.2157422751,-0.0996710435,0.3848805726,0.2827079594,0.1392045468,-0.0310675688,0.048223909,0.2516840398,0.1013729796,-0.0695960447,-0.0687178895,-0.0905827731,-0.1240739003,0.3312743604,0.2068857402,-0.3937552273,-0.0318776183,0.1077007353,-0.0436316021,0.1602514535,0.4636470377,-0.3375457823,0.1540868729,-0.1117823645,0.0359497592,0.1838454306,-0.1709900796,-0.2442660332,0.1554470062,-0.1522451788,-0.0724625289,-0.1648952663,-0.040329244,0.4156182706,0.1780366451,0.2875840068,0.1548374295,-0.0767079666,0.0052048075,-0.3618183434,-0.2044395357,0.1157821491,0.152536571,0.2196314335,0.0182796214,-0.175075531,0.0994481966,0.2839729786,-0.1476559043,0.0246441849,0.2063085586,0.2594321072,0.102494657,-0.099264361,0.0578925014,-0.0139874555,0.0029466471,0.1617276073,0.1466023624,-0.0912485793,-0.1259187162,-0.0808713213,-0.1296757758,-0.0401313864,-0.1913609058,0.4012335241,0.0231630635,-0.2436453849,0.0380667076,-0.0326103903,0.2772913277,-0.5564014912,0.116895698,0.0856815726,0.143217355,-0.2468101531,-0.137972936,0.0451400056,0.1475305706,0.0777290538,-0.1749167293,-0.1595963985,0.2797528505,0.0514730252,-0.1702575833,-0.0542087927,-0.0598187298,0.079489097,-0.0780127868,-0.0605498403,-0.024109846,0.1391024143,-0.0725652426,-0.0437277444,0.0198000576,0.0858965814,0.0416929312,0.0643806532,-0.1075470522,0.2989222109,-0.0637832731,0.2183286399,-0.1254119426,0.150870204,0.2486631721,0.148085475,0.1347068548,-0.362984091,-0.0534495115,0.3653757274,0.1575312763,0.0221728105,0.2209618837,-0.0477435067,-0.018417282,-0.103763625,-0.1129925847,0.0285339579,0.2524369061,0.1050950512,0.1041292101,-0.0058004549,-0.1915750504,0.1266401708,0.0279739909,0.1460324079,-0.0926327482,0.1141227037,-0.131258741,-0.3890062273,0.0243687611,0.1840142459,0.4346303642,-0.1706636697,-0.0885762647,-0.2157404572,-0.022049306,-0.0796040669,-0.0304986257,0.0756887272,0.0063563334,0.0699867159,0.1291413158,0.0442749485,0.3340530097,0.1918050796,0.2862678766,0.1312666684,0.2533352971,0.0983201116,-0.2767162323,-0.1055767164,0.3229141533,0.3100510836,0.0713494718,0.1366760284,0.0134438239,0.0413338467,0.0615663566,-0.1175929308,-0.1257170141,-0.0255134907,0.2186775804,-0.1618102342,0.3219445348,-0.0582852438,-0.2901395261,0.1826696545,0.0444106832,-0.1741296351,-0.2125734091,0.2222233713,-0.142817691,-0.2326691151,-0.2663710117,-0.0892449841,-0.409698993,-0.0233369824,0.0033621364,0.1049363762,-0.1140832305,0.3382145464,0.2784394622,0.1678650528,0.0769889653,-0.2047870904,-0.2406984717,0.0010084623,-0.2001372874,-0.2909089923,-0.0196567513,0.1453319937,0.4484384954,0.0565781705,-0.6337876916,0.1381919086,-0.0206948686,-0.0298275836,-0.0199877489,-0.2075379938,0.3398735821,0.0861832276,-0.280495882,0.0727095082,-0.1896730363,0.1456229687,-0.2381015271,0.058368329,-0.2997899354,0.6262995601,0.1008589864,0.5357186794,-0.0868577585,-0.2136739641,0.3018294573,-0.104256995,0.2297618389,-0.3354148269,-0.317358762,0.2339946181,-0.199220553,-0.1545128524,0.1516166031,-0.2063909471,-0.2986852229,-0.1931258738,0.098382093,-0.2361880392,-0.3058814108,0.2095521837,-0.0062322887,0.1644606143,0.0454294905,0.2024331391,-0.0491923057,-0.0051069967,0.0218538083,-0.1319111735,-0.0742247403,-0.2311108559,-0.1345995367,-0.0370503552,-0.2912639678,0.1280488074,-0.0431974344,0.156329155,-0.1944759637,-0.1157444268,0.1013467237,-0.1733676642,0.7220885754,0.2981640995,-0.2517338693,0.1072583795,-0.203968823,0.0144455107,0.0148126967,-0.2182471752,0.0294568017,0.1691404432,0.2344909608,-0.1143613905,-0.1062971801,0.1589983404,-0.0040469458,-0.1628544629,0.0408049524,-0.0760105699,-0.0720082149,-0.3172062635,0.3062619567,0.0000751532,0.2586673498,-0.1385029554,-0.3885134161,-0.1903767586,-0.2432740629,0.0305486284,0.0406208634,0.2845758796,0.0615795776,0.1720088124,-0.1825068295,0.2778347433,0.10861478,0.0385409184,0.0303482078,0.1055795401,0.0627170578,-0.1964756101,0.24676539,0.2457129806,0.0696830228,0.1003780365,-0.0196533855,0.2985500991,0.0040321071,0.476364255,0.3379434645,0.1779947132,-0.2616218925,-0.1978024095,0.1484490335,0.0290620644,0.1205294728,0.3407535851,0.0061643096,-0.221606195,0.1033465192,0.1269276142,0.7859761715,-0.0798925683,0.1563754529,0.1622484326,-0.1299407929,0.0706565008,-0.1290599257,0.1332830489,-0.1454270035,-0.2420355231,-0.0117923142,0.0841082484,0.0947091579,0.0661015585,-0.2309584916,-0.1309621781,-0.2417940199,-0.0347432941,-0.12609981,-0.0682323426,0.3135948777,-0.2964110374,-0.2406769842,0.3474310935,-0.0320165195,-0.1755944937,-0.1217935011,0.0284505859,0.2073505968,0.0153881069,-0.1561407745,-0.0159294568,-0.4113078117,0.1085858122,-0.1915371567,-0.4395812154,-0.029584961,-0.0236884151,0.2372672856,0.2627708316,-0.0402761549,0.3022933304,0.0585159697,-0.0558726601,-0.0063183415,0.1892680079,0.2918223143,-0.1375239789,-0.2573608756,0.3752954304,-0.0264450721,-0.2566692531,0.1116961241,0.0290731899,0.0259254463,-0.3345773816,-0.2897405326,0.1766740531,-0.1414851397,-0.2449982166,0.3526760936,-0.0272036418,-0.2025978267,0.0763795748,0.2673071325,-0.1722609699,-0.1517580301,0.2372366786,0.154964596,0.0203166436,0.4756414294,0.2392723411,-0.1857561767,-0.3338655829,-0.2238676399,0.326946795,-0.1067072973,0.1346709579,0.0286466479,-0.1105858833,0.1345710158,0.2904469073,0.020186564,0.1323320419,-0.1118649617,-0.055157356,-0.6730902195,-0.0890404433,-0.0282639787,0.2352817506,0.0963257551,0.3263259828,0.080704093,0.1305153817,-0.494967401,0.2333150953,-0.4960041046,0.2853696644,-0.172099039,-0.2201405019,0.0647670552,0.0208612774,0.3008740842,0.362732172,-0.0457558185,-0.492246598,-0.0302207228,-0.0348126106,0.0733001009,-0.0307317525,-0.0112101892,-0.1036137417,-0.2086821795,-0.1670752019,0.192026332,0.0240252409,-0.0383788235,-0.0162650179,0.0509443097,0.1126783788,0.0208045412,-0.1305144131,0.1121998653,0.1436704695,0.180826664,-0.1954687834,0.0351361819,-0.0273380373,-0.2746548355,0.122923106,0.0775791183,0.1198843941,0.1027949676,-0.2572697699,-0.1766558588,0.1449923664,0.1661091447,0.1802067161,-0.1044656411,-0.0656742826,0.3316280246,0.4633030593,-0.3924390972,0.1020054892,-0.0790088475,-0.1917509586,0.2278173864,0.0615895502,-0.1700830907,-0.1509520561,0.0428877957,0.2860615849,-0.022990806,-0.3346066773,-0.006291314,-0.0429664627,-0.150827229,-0.1082945168,0.0415449888,0.1582049429,0.2203144133,0.6154788136,0.0093526291,0.2319006771,-0.0099897841,0.0347447544,-0.1309582889,-0.1149120927,-0.1790274531,0.2740025222,-0.1531104743,0.0393011831,-0.172564581,0.2436451763,0.0130120134,-0.4212600887,-0.3078480363,0.1233169138,-0.4836005569,-0.0121978503,-0.0400583297,-0.0613486841,-0.0605766177,0.1017880067,0.1151562557,-0.0109383035,0.2107693851,0.0035055373,-0.2606017292,-0.0867086351,-0.0420812368,0.1316572279,0.0348119661,-0.1533124447,0.3249037862,0.0746885613,-0.1277173012,0.012509726,0.0688772053,0.2640544474,0.319170624,-0.0278167538,0.0290548373,-0.0105462084,-0.0819647238,-0.1663133502,0.0741641372,0.4301971793,0.3298614323,0.1651235968,0.3218916059,-0.339232415,0.2583580315,0.2204084396,0.1436823756,-0.3009209335,-0.0231049377,0.0529404692,-0.1410055161,-0.3837491572,-0.0982577056,-0.4071527421,0.192723617,0.0170793515,-0.2206182182,0.02226777,-0.0359903723,0.1851820499,-0.1399058551,0.5546776652,-0.0025411726,0.0117142042,-0.3824906349,-0.3812130988,-0.3882708848,0.1285359114,-0.1605913192,0.2828680575,-0.1538952142,0.2353523225,-0.0089147408,-0.0835232288,-0.0540946648,-0.0728151873,-0.0034517294,-0.1966046691,-0.5599756241,0.0591165461,-0.3399908245,-0.0430115201,0.1515894383,-0.0656553879,0.2719309032,-0.2163300365,0.2977227271,0.1515057236,0.0880366713,-0.0437324196,0.2607205808,0.2643214762,-0.2277108431,0.107152313,0.0530695766,0.2349273711,-0.0313944221,-0.1635961235,-0.0745228678,0.0268716309,0.4325712323,0.2168190777,-0.0892074481,0.0543145537,-0.0881393552,0.2173937857,0.2053120583,-0.0961533338,0.0329457149,0.0254562348,0.2312337309,-0.0308846515,0.2101952285,0.3477514088,-0.0113335513,-0.0824661851,-0.3962859213,-0.2467128038,0.5341277122,-0.0801385418,-0.3153985441,-0.1390158087,0.0159558114,0.2202746123,-0.1125640348,-0.6554005742,-0.1193427667,0.1337941885,-0.0970617384,-0.3419774473,0.280549854,-0.1716111451,0.0893601179,-0.0015595042,0.2308924794,0.0325436443,-0.0209655091,-0.0010068377,-0.2170399278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/757","title":"CUDA out of memory","comments":"the error happens when executing loss.backward()","body":"In your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n","comment_length":6,"text":"CUDA out of memory\nIn your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n\nthe error happens when executing loss.backward()","embeddings":[-0.31564641,-0.010816033,-0.1397833377,0.3557707667,0.5490416288,0.0549638979,0.3401089609,0.1870996058,-0.0963943452,0.3438007832,0.1469631195,0.0729068071,-0.0957498997,0.0438975431,-0.094872959,-0.2896981537,-0.1612949222,0.010533723,-0.1986415684,-0.0925278962,-0.0258407816,0.0907961801,-0.274245441,0.0410889201,-0.270825386,-0.1058467627,0.1235825494,0.2037989795,0.016788058,-0.2245287299,-0.0593528189,-0.2873102129,0.0817103386,0.3620395064,-0.0000956524,0.0390043594,0.0995051935,-0.1255464256,-0.1418096125,-0.100565128,-0.1708562225,-0.0161173753,0.1422035396,-0.4185650647,-0.1446307153,0.1617532969,-0.0037339693,-0.0918630287,0.543689847,0.4574703872,0.3895350397,0.0158246048,0.2225860953,-0.2225929052,0.1398513466,-0.1257176846,-0.0274266433,0.0095368987,-0.0084953895,0.0138388695,-0.0091856215,0.262196213,0.1054680496,0.0030366001,0.0379907079,0.0026064795,0.1660846919,-0.0116580036,0.036936149,0.0348594375,0.3217495978,-0.2452491969,0.0693678632,0.0606562272,0.0084986882,-0.6133979559,0.100301303,0.1643625945,-0.3250306845,0.1595817804,-0.2409178317,-0.0316499174,-0.1197501421,-0.0034601877,-0.1334127635,0.0322957262,-0.0746697411,-0.076267004,0.2240057588,0.1076579392,0.2029847801,0.2959357202,0.0931832045,-0.0205987953,-0.2095094621,-0.0106492294,-0.0395789146,-0.1530500203,0.1044434234,-0.1947086453,0.3281950355,0.0620489717,-0.0489292629,0.0538025983,0.1030460298,0.3072592616,-0.2109565288,-0.0528979264,0.0645964518,0.1584420949,-0.2452726811,-0.2367709428,-0.0178624038,-0.304317534,0.0184357632,0.1003000215,0.0601469912,-0.1953869164,-0.3125452697,0.1320943236,-0.1957952082,0.1682513505,0.0323540308,0.1749895662,-0.0270143785,0.1146645099,0.226195246,-0.109569937,-0.0308432225,-0.0953110754,-0.2784354389,0.060761366,-0.423903048,-0.0511757396,0.0720314234,-0.0128035285,0.3640543818,-0.1403627694,0.0336359665,-0.262110889,0.0609678216,-0.2395918667,-0.0202664603,-0.0023841709,0.0394101106,-0.0800413266,0.0665972978,-0.1894902587,0.3109566271,0.2445489615,-0.3046654761,-0.277236551,-0.0088917194,0.3739856482,0.4005457163,0.2230486125,-0.0023853795,0.117806606,0.1940815449,-0.005274673,-0.0337365717,-0.0483542196,-0.1662281603,-0.1880370677,0.3504390121,0.3241146505,-0.5481361747,0.0133346515,0.0154401772,-0.0617501996,0.1582100242,0.5219310522,-0.3247281015,0.1570933312,-0.1622375101,-0.0740421712,0.2730964124,-0.2193171531,-0.2357416749,-0.020480033,-0.1757066548,-0.1173518077,-0.289848417,-0.0273472685,0.3263846934,0.0920841545,0.2938703895,0.0572015084,-0.1999763995,-0.034501873,-0.3494429588,-0.0233857911,0.290617317,0.060131643,0.3228906095,0.035028901,-0.0900689736,0.1642953604,0.2367517501,-0.14053379,0.0437354408,0.1679202467,0.2024253011,0.0116784135,-0.1214956194,-0.0275338273,0.0616441667,-0.0243732482,0.2551512122,0.0681808591,0.0417948328,0.0115838926,-0.1060592011,-0.0535358302,0.0146047603,-0.0742205158,0.3958166242,-0.0102273626,-0.353319943,0.1310016066,-0.0018274301,0.2823489904,-0.5627521873,0.1541346461,0.0159069374,0.1574269682,-0.2948842645,-0.1026845276,-0.0559116639,0.1571476161,0.1613367796,-0.1261093915,-0.1331169605,0.2861905396,-0.0858489797,-0.1700906008,-0.0993663669,-0.0582234487,0.1344282478,-0.1735226363,-0.0278811269,0.0502294637,0.1693145782,-0.1400817782,0.1111925468,-0.0593359806,0.0779413283,-0.1153729782,0.0766295493,0.006839517,0.1426194608,-0.059907753,0.1946242005,-0.1595277637,0.1698842794,0.039113041,0.0677829832,0.049449496,-0.4741793573,-0.131354779,0.4039709568,0.1393110007,0.01631606,0.3615306616,-0.0060977261,0.0327668414,0.0398848094,0.0475286841,0.1633776724,0.1593559086,0.1184923872,0.1819821894,0.0011765185,-0.2295042276,0.0704383403,0.046451088,0.2797433734,-0.1410229504,0.0333830565,0.0118102701,-0.2733566165,-0.0080093658,0.1707322896,0.4336521626,-0.139855653,-0.077527903,-0.1878044754,-0.0633341447,-0.2235547751,-0.0802501217,0.1132016554,-0.0141208339,0.1262356192,0.0676785633,-0.0559571944,0.2761357129,0.1921551228,0.234899655,0.0548701957,0.3834463954,0.0915650278,-0.3123745918,-0.1033894122,0.3026879728,0.1713854074,0.0452892892,0.1424122006,-0.0977122039,-0.0990715176,0.0411855578,-0.1186421961,-0.0972837806,-0.0202032477,0.1600909829,0.0116336774,0.1763346344,-0.0588270091,-0.3047590554,0.2294220626,0.0239431225,-0.1025527492,-0.1793430448,0.2303777784,-0.102394022,-0.1935748458,-0.1727543473,-0.1791308671,-0.4166761339,0.0154942041,-0.0317229182,0.0819825381,0.0239304844,0.1869585514,0.2758398354,0.3195187449,0.1646284163,-0.106213741,-0.2144313306,-0.0745860338,-0.1913425624,-0.2148296386,0.0866775736,0.2380378544,0.5438649654,0.154640317,-0.5281574726,0.0767605752,0.0862149745,-0.0541018769,-0.0293214824,-0.1608521044,0.391349107,0.1321766227,-0.2789078653,0.1114006117,-0.1977013946,0.0852959454,-0.2984719872,0.2280881554,-0.2335647196,0.6457072496,0.1576704085,0.7361820936,-0.0915618986,-0.1484402418,0.2981447875,0.0119061563,0.1827775389,-0.2183460593,-0.3649846315,0.2807450891,-0.1663895696,-0.3238139749,-0.0718270466,-0.289144069,-0.4093184173,-0.2226622999,0.0512452498,-0.3414093256,-0.2993402183,0.0342677608,0.136349991,0.1291639656,0.0024344611,0.1703540087,-0.0199504178,0.0140283965,0.1675712168,-0.0102230674,-0.0789005458,-0.1923816949,-0.1871146411,-0.045384597,-0.3494098783,0.0854888707,-0.1480243057,0.1282951236,-0.2318241894,-0.0270462297,-0.0785633251,-0.1117923334,0.5153464675,0.3372752368,-0.2887536883,0.1103975773,-0.1565349698,0.0563909188,0.023870999,-0.1109937727,0.12002673,0.2666622996,0.0054967925,-0.0027044597,-0.0692887977,0.1958482116,-0.0101054367,-0.0994422361,0.118015863,0.0291029029,-0.0275227893,-0.239708215,0.2248152941,0.0660681054,0.1850986779,-0.086823456,-0.2632046938,-0.100916855,-0.1189215556,0.0143736424,-0.0590821579,0.2989198565,-0.01665828,0.0533709377,-0.2147438675,0.3247770667,0.0657959655,-0.0344170593,0.0777474493,0.073027052,-0.0038579721,0.0167620685,0.1754386425,0.2425888479,0.0303519946,0.0243863314,-0.1068340763,0.2358882427,-0.1074275449,0.3339174688,0.2666712105,0.1786070764,-0.2018076032,-0.1371130198,0.1199197993,-0.0392174236,0.0910743773,0.4367209673,0.0023016965,-0.3331176639,0.1756852567,0.2667809427,0.8569936752,-0.1103152633,0.1203153208,0.1161390394,-0.0829539374,0.1656933427,-0.1410201788,0.1628840715,-0.1218005866,-0.2915000618,-0.0379929543,0.1216773093,0.1069872677,-0.0142039899,-0.2216931432,0.0734576285,-0.0073341448,-0.1387856305,-0.1699548066,-0.1267966032,0.3544151783,-0.2670301795,-0.2409174293,0.3230738938,0.0066836779,-0.2649596334,-0.2212513089,0.0271321442,0.3020628095,0.0371316858,-0.2670526803,0.0036066135,-0.5152701735,0.0826417655,-0.2912952006,-0.5483094454,0.1515272409,0.1342442185,0.4394280314,0.2811528742,0.0009983673,0.2615604401,-0.070071727,-0.0450474247,-0.0944623426,0.1674088538,0.394910723,-0.1602264941,-0.2374518961,0.4268359244,-0.065512456,-0.2615479231,0.0590770058,0.0342499353,0.0880456716,-0.2322923392,-0.3248641789,0.1125048622,-0.1430911124,-0.2401550561,0.330442518,0.0610306412,-0.2552140355,0.0379422121,0.0119666578,-0.1680006981,-0.0777005553,0.2038887888,0.0765972957,0.2148515284,0.543744266,0.2571394742,-0.1373666078,-0.2277008891,-0.3516314626,0.4117773175,-0.0907146484,0.1059139073,-0.0254666321,-0.0102335066,0.0739599466,0.327133894,0.0346570648,0.1408510208,-0.0831332728,-0.1438246369,-0.6825150847,0.0239593964,-0.0168730319,0.2750003934,0.117429696,0.2217218578,0.095013231,0.041198343,-0.4652204216,0.260825932,-0.5161613226,0.2988553643,-0.097272411,-0.152783066,-0.0282910578,-0.0178145319,0.2504968047,0.5145121217,-0.0354213789,-0.409393698,0.0455256738,-0.0226108767,0.0949225724,-0.1141312197,0.0806816444,-0.1420778334,-0.1801255196,-0.1227668673,0.1926711202,0.1163949966,0.0362228043,0.1584739536,0.0229416508,0.0745089278,0.0962705314,-0.077943705,0.1664848328,0.247264877,0.1639335304,-0.1118822768,0.0339459702,-0.1149280965,-0.4327068627,0.1239068583,-0.0110356873,0.0911709666,0.1170830131,-0.3508817554,-0.2229702771,0.2337804288,0.2268226743,-0.0006321807,-0.1073241308,-0.0300570223,0.4318662286,0.4114925861,-0.3573097289,0.0648657158,-0.050633546,-0.1189505905,0.1111201271,-0.0209064018,-0.2630654871,-0.1489220262,-0.052836623,0.2039272636,-0.0565547198,-0.3707235456,0.1187603176,-0.2142623514,-0.1379618645,0.041190479,-0.126768589,0.1320974827,0.2081155181,0.5931918025,-0.0882763863,0.2787097991,-0.0174750146,0.084879145,-0.1862758696,-0.1985101849,-0.2002349347,0.2736057937,-0.1820455343,0.0968246683,-0.1139092669,0.1769698858,-0.0101856692,-0.488881737,-0.1082068756,0.1098813117,-0.4296532571,0.062652804,-0.0709951371,-0.1286485493,0.1134263352,0.0836458951,0.0880312771,0.0104047982,0.3454327285,-0.0342555456,-0.2058932334,-0.0991562158,-0.0513331816,0.2123697549,0.0287263058,-0.2416535765,0.3065276444,0.1152034998,-0.2015155107,-0.0503761545,0.1600894481,0.2213081867,0.3233834803,0.0893906429,0.0538634583,-0.0159290079,-0.1344053596,-0.0792480931,0.0809931755,0.4986602068,0.3484740257,0.1854818463,0.3115571141,-0.2827804983,0.2630265355,0.1610875279,0.1788541228,-0.4384532571,-0.1269968748,0.0034740679,-0.136859104,-0.4108421504,-0.0460124277,-0.3099844158,0.2053343505,0.0561056249,-0.2135705054,0.0498800911,-0.0297513735,0.18045111,-0.11830955,0.4148381054,0.1722552329,0.0025228038,-0.5248805285,-0.3804501295,-0.385817498,0.1498835534,-0.1522703469,0.337959826,-0.0966934934,0.1588047296,-0.0448608436,-0.0576546267,0.0255870316,0.0217512045,0.0638277084,-0.0827858001,-0.4914128184,-0.065576598,-0.3620012701,-0.227669403,0.1687453538,-0.0420047678,0.1639514714,-0.2050747126,0.2453229874,0.1151475757,0.245810613,-0.1358170509,0.2223539054,0.2783902287,-0.1632572114,-0.0324651338,0.053853713,0.1090050265,-0.0434122942,-0.160142377,-0.1857739538,-0.0983558372,0.3605250418,0.3383214176,-0.110515289,0.1474747062,-0.212376833,0.2961773574,0.2157161534,-0.2161800116,-0.0753719807,0.1618209332,0.2564951479,-0.109938845,0.324115932,0.4253447056,0.0341927074,0.0904267654,-0.3607902229,-0.3526273668,0.5530930758,-0.0758845657,-0.1794924289,-0.1046767309,-0.1088304594,0.3364746571,-0.2318263948,-0.7959830761,-0.1413161755,0.111687623,-0.033426661,-0.3792395592,0.1259617209,-0.1281611621,0.0510554388,0.0778992251,0.2904193401,0.0200988073,-0.1298357397,0.2656153738,-0.1605157256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/757","title":"CUDA out of memory","comments":"Since you're using a data collator you don't need to tokenizer the dataset using `map`. Could you try not to use `map` and only the data collator instead ? The data collator is supposed to pad to the longest sequence in each batch afaik, instead of padding to 512.\r\n\r\nAlso cc @sgugger ","body":"In your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n","comment_length":52,"text":"CUDA out of memory\nIn your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n\nSince you're using a data collator you don't need to tokenizer the dataset using `map`. Could you try not to use `map` and only the data collator instead ? The data collator is supposed to pad to the longest sequence in each batch afaik, instead of padding to 512.\r\n\r\nAlso cc @sgugger ","embeddings":[-0.3538847864,-0.0687038526,-0.1432429403,0.3057073951,0.5980893373,-0.0301802307,0.445364207,0.2957354188,-0.0621404573,0.4082165956,0.2172657698,0.0536231026,-0.0748180449,0.1507742554,-0.1334300786,-0.1775912493,-0.0540916361,0.0361537896,-0.0717044994,0.032488171,-0.03451683,0.0090633547,-0.2618727088,-0.0872588232,-0.4492588639,-0.0958235189,0.0632595867,-0.0474194549,0.0135206999,-0.251195997,-0.0501876138,-0.0768315494,0.1226606742,0.444525212,-0.0000950249,-0.0089978678,0.0488267429,-0.0793752372,-0.1593962461,0.003702936,0.0010660542,-0.1991862059,0.0680232123,-0.3073745668,-0.3009664416,0.2325589657,-0.0483179353,-0.1761007756,0.4668056369,0.3489148617,0.4058131576,0.0642159581,-0.0480915904,-0.1667485684,0.0934195817,-0.100873664,-0.0437048562,-0.0739792436,0.2798736393,-0.0510345511,-0.163937524,0.2283999324,0.0955837592,-0.0248313732,0.0088261869,0.0286433809,0.1250680387,-0.1874313205,0.1657485813,-0.0107836062,0.1392261237,-0.1024777889,0.0851274878,0.0155014759,-0.1246241853,-0.496588558,0.1401783526,0.0640604496,-0.3090767562,0.083506383,-0.1675762385,0.1029407158,-0.065088205,-0.0328303687,-0.2521700561,0.2495867759,-0.1194060594,-0.0043268823,0.1522545964,-0.1460786462,0.1584289968,0.0882797018,-0.0002291196,0.0289949588,-0.2778501511,-0.0494477637,-0.0506685749,-0.2739793956,0.1108096316,-0.1995375007,0.169232145,0.1295817941,0.2328849882,0.0386407822,0.1223049834,0.3336030543,-0.1237151623,0.0134235201,0.068894282,0.18664141,-0.3788158596,-0.1959584802,-0.1081423312,-0.3177534044,-0.0082560694,0.1122581065,-0.1409739107,-0.0693289116,-0.061546538,0.2887709737,-0.1936705858,0.1201417893,0.0923148245,0.3297210038,-0.0462105535,0.0614209846,0.0122292172,-0.0794339105,-0.0581315942,-0.1534135789,-0.2928471565,-0.0796786919,-0.4233686626,-0.0175390206,0.1741263568,0.1073105112,0.2736075819,-0.0849092528,0.0604409613,-0.3291586637,0.0684488043,-0.2856378555,0.1045873016,0.0019267949,0.2612719834,0.0498303324,0.0454895832,-0.1384896338,0.2103121728,0.2335585654,-0.1626400352,-0.197183907,-0.1050039753,0.3500034511,0.3678776026,0.2698632181,-0.0442567021,0.0073637129,0.2607892156,0.1084063426,0.0064527215,-0.1018997356,-0.126054883,-0.1532547772,0.2914360166,0.2016898245,-0.3154522777,-0.0260048825,0.1940804571,0.0753099024,0.2819879949,0.4822386205,-0.3072268963,0.1555389315,-0.1969341785,0.1619755179,0.2056476027,-0.2838630676,-0.2273934633,0.154495582,-0.2635684907,-0.0734121203,-0.1839417964,-0.1002919972,0.5489272475,0.08630649,0.3641426861,0.1814376712,-0.1500034928,0.0459711216,-0.3630645573,-0.122411564,0.1280618459,0.0573726892,0.179444626,-0.0891334787,-0.1700423062,-0.0015277917,0.1990641505,-0.1354993582,0.1784704626,0.2134157866,0.3221105635,0.0361040384,-0.1397631019,0.1699185073,0.0245382804,0.045541849,0.1107642874,0.1084509417,-0.1470888853,-0.1655827761,-0.0250591226,-0.1039782241,-0.0178107712,-0.204149276,0.3734978735,-0.058643423,-0.1204763502,0.0145024955,0.0387140289,0.1495333761,-0.4617997706,0.1477302909,0.1302772164,0.0800374895,-0.2617274821,-0.1296993047,0.0376938507,0.2047877014,0.1960292011,-0.1215371266,-0.1078026593,0.2921037674,0.1026141196,-0.2014149427,-0.0868217722,-0.0230327453,0.0566488877,0.0194748826,-0.1182487085,-0.0796515346,-0.0015650917,-0.0857052654,0.0124379974,0.1006002054,0.2694793642,0.095901601,0.0480761975,0.0103645446,0.3586985767,-0.1140572056,0.1640129983,-0.1487576067,0.0851368904,0.2405935824,0.1001321822,0.1569625884,-0.2406001389,0.0131232971,0.3567017019,0.1434330642,0.0106118564,0.2353611141,-0.1175424829,0.0115699163,-0.0241484921,-0.0806039274,-0.1070519313,0.2489943206,0.1102844328,0.0638679639,0.0392749682,-0.1660852134,0.1283660829,0.0325569436,0.0714564249,-0.1245191023,0.1260451972,-0.1098839715,-0.2454166412,-0.00226641,0.1472186446,0.4225439727,-0.1127472371,-0.0931511745,-0.1891930997,-0.1160900369,-0.0977402851,-0.0801920965,0.0199677069,-0.0049275979,0.0364934541,0.1595837623,0.0110066505,0.4099168181,0.2289893925,0.3758627176,0.1902043819,0.2441900074,0.0989734158,-0.2734331489,-0.1997942626,0.2843850255,0.3823623657,-0.0000639176,0.152418524,0.0594177246,0.081992209,-0.0315729119,-0.1147615984,-0.0479231961,-0.0653529838,0.1442676038,-0.094034709,0.3292437792,-0.2314816862,-0.2863684893,0.1959941983,0.0621236935,-0.1783341318,-0.2070908397,0.1329571158,-0.2569445968,-0.2753174901,-0.109054625,-0.0758211687,-0.3809902072,-0.0540908836,-0.0438662879,0.0942372605,-0.1224954948,0.2664973438,0.2440908849,0.0497474708,-0.1020129323,-0.174149394,-0.2536875606,0.0660755932,-0.2427299917,-0.2249824554,-0.0060495893,0.0853962824,0.5231486559,0.0444665179,-0.5343627334,0.1466070712,-0.0581690595,0.0312941819,0.0156113757,-0.1462978572,0.3653272688,0.0678780973,-0.2486971915,0.0843081176,-0.2163977027,0.1763117164,-0.2280780524,0.1031284928,-0.2819189131,0.5773177147,0.1775274575,0.6508575082,-0.0862182155,-0.2206742465,0.2745224535,-0.1354077309,0.1569600254,-0.2709427476,-0.2667551935,0.305085063,-0.1458317488,-0.1728738993,0.2063252777,-0.1714203507,-0.2792171538,-0.0277968962,0.1729624867,-0.2085362822,-0.392279923,0.2551884353,-0.0207874142,0.204292208,-0.0058908607,0.1315239221,-0.1478251219,-0.0500062369,-0.0046176706,-0.23295708,-0.0746494457,-0.2474722713,-0.2771756649,0.0280426648,-0.3883842528,0.15052028,0.0546479449,0.2861517072,-0.1500143111,-0.1118377447,0.0644527823,-0.0936582163,0.782335043,0.2042611539,-0.2611229718,0.0629386008,-0.159555465,-0.0211343504,-0.0055560856,-0.1316348463,0.1276701987,0.2425470352,0.3826509714,-0.1022981331,-0.1320970207,0.0025556388,-0.0479671471,-0.1053622141,0.034954872,-0.0143243466,-0.018823659,-0.3001775444,0.3913336098,0.1144846007,0.2752559483,-0.0381480344,-0.409035027,-0.1154563576,-0.2068901509,0.0979672149,0.039875742,0.3180324137,0.0902369618,0.1538381577,-0.1518669575,0.2439211011,0.1580741704,-0.1134222969,0.0096574435,0.0622029528,0.0988895595,-0.2057331055,0.3376143575,0.2864977121,0.064644739,0.1591658443,-0.0700829625,0.3968051374,-0.0992700309,0.395696938,0.3567792773,0.2163238972,-0.474974215,-0.1481388509,0.0088896854,0.0897904038,0.1118635982,0.4182065427,-0.0859785452,-0.323579073,0.2160569727,0.2096794993,0.8515176177,-0.1282588989,0.1934074759,0.0857783407,-0.0494305976,0.1198040396,-0.1908569038,0.1594129801,-0.2709292471,-0.2158919722,-0.0053512305,0.0670854598,0.073482655,0.1007448956,-0.2985439897,-0.11484126,-0.242041871,0.2059373856,-0.2007932663,-0.0606809594,0.2365565151,-0.4092287421,-0.2159014642,0.336620152,0.0250645876,-0.2348307222,-0.1664670557,0.0406505428,0.154088527,-0.0858798921,-0.1135170758,-0.0651403293,-0.5472652316,0.0305750053,-0.1110365316,-0.2965449095,-0.0454847887,0.0024608285,0.3210063279,0.3504115939,-0.037463773,0.319716841,0.1968779862,0.0088998927,0.0480723344,0.0419994295,0.244452104,-0.0612894706,-0.1383037418,0.3267183006,0.0412620604,-0.2219098955,-0.0039600306,0.0022162392,0.1148075312,-0.3608170152,-0.3072972596,0.1309196502,-0.0943896845,-0.2317421287,0.3246448338,-0.1180545092,-0.2070478201,0.0923194736,0.1686405241,-0.1416135579,-0.0677458495,0.156057924,0.0390854217,-0.0217384808,0.4916595817,0.2805047929,-0.1319648921,-0.3513337672,-0.1413226128,0.3831877112,-0.109190546,0.0950895026,0.0252765641,-0.1052054167,0.1053043976,0.3492059708,0.0430460609,0.0727331638,-0.0052797776,-0.0628239512,-0.6887542605,-0.1755339354,-0.074403055,0.2248651683,0.0487100519,0.3843528032,-0.0564285554,0.0756999031,-0.495967567,0.2193604857,-0.5408854485,0.3170781136,-0.1981321126,-0.1573366523,0.0791442171,0.0071351118,0.2550952435,0.4369693696,-0.1066215411,-0.4734306037,-0.0444895625,-0.0278031435,0.125888139,-0.1045935676,0.0343579836,-0.0124308271,-0.1867099702,-0.2628528178,0.1092704535,0.0853817314,-0.036403425,-0.0833310112,0.0957247317,0.1358408183,-0.0031876892,-0.1440229714,0.0927834809,0.2048113495,0.2161944956,-0.1022167057,0.0665897429,0.0020528608,-0.1708849072,0.1673737615,0.1186639518,0.1209047586,0.0366824269,-0.2211583406,-0.2700939178,0.2185980082,0.1463062614,0.0635443106,-0.1977745146,-0.1002414599,0.4072958231,0.4704892933,-0.239029184,0.1033436209,0.0320849344,-0.1943844557,0.1942332983,0.139599815,-0.092006579,-0.1526381373,0.0075538335,0.2622355521,0.0186193138,-0.3551523387,0.1178145111,0.0326393396,-0.0349846482,-0.0391604491,0.0502266735,0.1195853204,0.2119369507,0.5023724437,-0.0060723871,0.3215268254,0.0628102869,0.0257031657,-0.142088905,-0.2357930541,-0.1566660553,0.2672967613,-0.1019145399,0.038077645,-0.192658633,0.3080667257,-0.0372764245,-0.4703216851,-0.2891817689,0.1532073021,-0.518683672,0.0103962449,-0.1042399406,-0.0209073275,-0.0538057424,0.1113285869,0.1576839089,0.0044830237,0.143153891,0.0038083408,-0.1930522472,-0.0180151705,-0.1817862242,0.1559149623,0.0765524358,-0.186331287,0.3365648985,-0.0710554346,-0.1482813954,0.0015257401,0.1111113355,0.3467008471,0.4203843772,-0.0579400621,0.0301737413,-0.032778848,-0.021985,-0.2057778537,0.1491652131,0.3884341717,0.2333977371,0.1835659742,0.3134994209,-0.3399816155,0.1441179514,0.2320036739,0.0603805371,-0.1204472333,-0.0974949598,0.099840723,-0.1227229759,-0.3545719087,-0.0971661583,-0.3997303843,0.1740504652,0.1012748927,-0.2028765678,0.1588527113,-0.1019739285,0.1801307946,-0.1071234569,0.5011608601,0.138869524,-0.0256978292,-0.4887701273,-0.4422722459,-0.4006253779,0.1141264662,-0.2080056667,0.2612013817,-0.2109655589,0.1496301889,-0.0620724186,-0.1585180163,-0.0386680588,-0.0583045967,-0.1338530779,-0.1742406487,-0.5670872331,0.0746911988,-0.2965662777,-0.0312129688,0.1268420815,-0.1103565916,0.3521364331,-0.1709131151,0.2749555111,0.1290583909,0.1577444524,-0.0920079201,0.0661805645,0.3151902258,-0.1220371276,0.0438802801,0.0789468288,0.2435211688,-0.0885535777,-0.0638367012,-0.1045812145,-0.0615077429,0.3914674222,0.2720617354,-0.1657000333,-0.0123304212,-0.2234526277,0.1061282158,0.2691363692,-0.0650159717,0.077610895,0.1415107846,0.1761374027,-0.0450098664,0.2141918689,0.3716608584,0.0165967755,0.111627087,-0.378683418,-0.2809372842,0.4669593871,-0.3317545354,-0.3854004443,-0.0718252584,0.0286341403,0.1684820056,-0.1871622801,-0.5829002857,-0.1516562253,0.1221458837,-0.0802416578,-0.3954402506,0.2426403612,-0.1927638948,0.0949652269,-0.0366261527,0.1453107446,0.0766773224,-0.1085129753,0.0523131005,-0.2184519917]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/757","title":"CUDA out of memory","comments":"Closing this one.\r\nFeel free to re-open if you have other questions about this issue","body":"In your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n","comment_length":15,"text":"CUDA out of memory\nIn your dataset ,cuda run out of memory as long as the trainer begins:\r\nhowever, without changing any other element\/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.\r\n\nClosing this one.\r\nFeel free to re-open if you have other questions about this issue","embeddings":[-0.3400762975,-0.0855902806,-0.1650410593,0.3249937594,0.6060909033,-0.0403195024,0.4270146787,0.2553589344,-0.0715739131,0.4209960401,0.185912773,0.1221811175,-0.051040113,0.2017464489,-0.1587797403,-0.2027603239,-0.1131787002,0.0617581196,-0.1540765911,-0.0391461961,0.004438587,0.0698728189,-0.2483747602,-0.0458693467,-0.3964010775,-0.0884181857,0.1277511865,0.1092882678,0.0894006938,-0.1695463508,-0.0438509621,-0.1101237461,-0.0630380139,0.3814639747,-0.0000936632,0.0425125882,0.0949545801,-0.0490948893,-0.171335578,-0.1015473306,0.0156421289,-0.058138188,0.2235110253,-0.2925045192,-0.302036196,0.2294334322,-0.0336429067,-0.152110815,0.4355399311,0.4175440371,0.4308171868,0.0612173602,-0.0246755965,-0.1824290901,0.0813177228,-0.0950727612,-0.044300653,-0.0738885701,0.1362668723,0.0088141989,-0.1182514355,0.2346219271,0.0538604409,0.0354974382,0.0048042312,0.0351718813,0.0231665354,-0.0948890969,0.185355112,0.0251222104,0.237459287,-0.0889609456,0.1323986799,0.0277361516,-0.0305460487,-0.4217492342,0.2069141567,0.075270012,-0.277279526,0.0840834752,-0.1511641294,-0.0594984218,-0.0895057619,-0.0653821006,-0.1920795888,0.088625811,-0.1401712596,-0.0428105257,0.2009375095,0.0061591309,0.2041730136,0.1913493425,0.0787252709,-0.0597333983,-0.1920144111,0.0113050295,-0.0220776126,-0.201595068,0.0347665586,-0.1506214142,0.2334517986,0.124194704,0.1289770901,0.0709453076,0.1078717038,0.2626609206,-0.1746461391,-0.0604802892,0.1262073964,0.1932847649,-0.4087637365,-0.2120487094,-0.0505955368,-0.3032140136,-0.0418411829,-0.0001887688,0.0368638188,-0.1598519981,-0.1553711146,0.221146211,-0.1987977028,0.1236580461,0.0667260811,0.3126460612,-0.0716070235,0.1257661879,0.1721587628,-0.0680552647,-0.1040358469,-0.1551535428,-0.3256243169,-0.034456417,-0.3082820177,-0.0513355434,0.1770140678,-0.0193198081,0.3744767606,-0.0833299309,0.0473525748,-0.3399529159,0.080479078,-0.2133800834,-0.0329791419,0.0403410196,0.2297329456,0.0224875715,0.048159603,-0.1347981691,0.2834426165,0.3216530383,-0.190508008,-0.1698579639,-0.1307217181,0.3972017765,0.3022097349,0.0994320661,0.0257190038,0.0477277525,0.2762895525,0.1040519625,-0.0142621556,-0.0342193171,-0.1065646559,-0.1342776865,0.2871774733,0.2008584887,-0.3629264235,-0.0488588847,0.0638104826,-0.0576666892,0.1965082586,0.4805929959,-0.3300361633,0.1375161409,-0.1110421196,0.0668406337,0.1920344681,-0.2152825594,-0.2251886427,0.1402208209,-0.204540059,-0.0471759997,-0.1945177317,-0.0534457974,0.3400162756,0.1742246598,0.2338988781,0.1507943273,-0.094427608,0.018159261,-0.3922608495,-0.1705776006,0.1141636372,0.104731068,0.261485815,0.010008581,-0.1901880205,0.1491459161,0.2865661085,-0.1397072673,0.0368277729,0.2064145207,0.3562002778,0.0744057,-0.1390744448,0.0396519862,0.0761431158,-0.0021479535,0.1540036798,0.1316402853,-0.0711243972,-0.1158535928,-0.0787349939,-0.146183461,-0.0566723756,-0.2374371588,0.4076394439,0.0202540513,-0.3147120774,0.0752842203,-0.0158300325,0.2739010155,-0.5340737104,0.1113759279,0.1289527416,0.149172917,-0.2603077292,-0.0838240534,0.0179040954,0.1712759435,0.1057353988,-0.1367147863,-0.1590766162,0.308786571,0.0007887104,-0.1010002792,-0.044936534,-0.0787518248,0.086120382,-0.104829967,-0.0711894184,-0.0502930582,0.1177254319,-0.0768345445,-0.0051340368,-0.0040495098,0.1321269721,0.012302951,0.0702195317,-0.0538013838,0.2601024806,-0.1253153831,0.2270533442,-0.1411993206,0.1333103925,0.2103100717,0.1568932831,0.1228440851,-0.4185473323,0.0060981554,0.4697881937,0.2106494606,0.0404407084,0.2278088927,-0.0439954512,0.0128286462,-0.1277989596,-0.0969707668,0.0139140319,0.2327601016,0.1368488818,0.1502041668,-0.0410155617,-0.2047105283,0.1561100185,0.0246369503,0.1538496912,-0.0434695408,0.1207278967,-0.1167823374,-0.3814041018,0.0529665016,0.1630574912,0.3730143011,-0.1512226909,-0.0754969046,-0.2338861376,-0.0396199115,-0.1348814368,-0.0393435135,0.0138080074,-0.0041975197,0.1167159975,0.1439021975,0.02598482,0.3389230072,0.2365446389,0.2566678524,0.0947068334,0.2807426155,0.1150416955,-0.2863055766,-0.1230714843,0.3250212967,0.3048816919,0.0342250988,0.1474909931,-0.0012203158,0.0274741612,0.0652433783,-0.1668473631,-0.0912919715,-0.0187747423,0.2327405065,-0.1283296347,0.2993411124,-0.0789595842,-0.3095910847,0.1849641204,0.0784201846,-0.177643016,-0.240021497,0.2312419862,-0.1656937301,-0.2861095965,-0.2690458596,-0.0419202857,-0.3965493441,-0.0144898957,-0.0555506498,0.0768921301,-0.0456274077,0.289339453,0.2885732651,0.1438882202,0.1057327017,-0.2135178596,-0.215309009,-0.0512290001,-0.2097036391,-0.2905012667,-0.0253605265,0.1519947648,0.5169380307,0.0216586404,-0.6221033931,0.0969807506,-0.0212161243,-0.0241065007,0.0140420506,-0.2129510343,0.3379654288,0.0871150941,-0.2922165692,0.0325407162,-0.1825626642,0.1928290874,-0.2174768001,0.1137465537,-0.3010709286,0.6044348478,0.0972035378,0.5691809058,-0.1142264828,-0.1915694326,0.3047715425,-0.0899735391,0.223745957,-0.3156792521,-0.3081545532,0.2146705836,-0.157166779,-0.1879882962,0.1249954924,-0.1847010255,-0.3134367168,-0.1789241582,0.1089940146,-0.2831968963,-0.3224898279,0.1455546021,0.0237573758,0.1704107523,0.057867609,0.1971821189,-0.0294204336,0.0211309288,0.0224216357,-0.0686881244,-0.0821171254,-0.2822330892,-0.143777132,-0.0737466961,-0.2440574318,0.0912624449,-0.055985447,0.1626595259,-0.2480314672,-0.1143383458,0.1207504347,-0.1693537533,0.6883091331,0.2774794102,-0.2866244614,0.1017926782,-0.1897334456,0.0351728685,0.0313639939,-0.198009029,0.0537550747,0.1525080055,0.2275473773,-0.0653866008,-0.1112277806,0.1266249567,-0.0373797677,-0.1598826051,0.1104247198,-0.038830936,-0.056668058,-0.2945433259,0.29752931,-0.016834192,0.2550349236,-0.1716603637,-0.3714458346,-0.1436739415,-0.2192379385,0.0272341873,0.0471045859,0.3073254526,0.0409180969,0.1846050173,-0.202318877,0.2915947735,0.0864014179,-0.001966232,0.0360926911,0.1422807574,0.0331321955,-0.2069731951,0.2715002596,0.2690289915,0.0532907322,0.1192269102,-0.0357325561,0.3006443679,0.0093773929,0.4253540039,0.3723422587,0.2170529962,-0.2285766602,-0.1505293399,0.1462781131,0.0409523919,0.0873235017,0.3668553531,-0.0405679494,-0.224035874,0.0912836567,0.1146864519,0.8032432795,-0.0321332887,0.1395725012,0.1416281611,-0.1637245715,0.0539265871,-0.1167421117,0.1528065056,-0.1889503896,-0.2280789614,-0.0124106491,0.0979782194,0.115790002,0.0234029796,-0.2537070513,-0.1894619018,-0.2580034733,-0.0333606601,-0.1674800068,-0.1135003269,0.2589678764,-0.2609410584,-0.2125396878,0.336253494,-0.0204355381,-0.1284044385,-0.1462150514,-0.0240053479,0.1864775717,-0.01563753,-0.1840071231,-0.017907843,-0.4245589077,0.1093555391,-0.1972277015,-0.4166310132,0.0207650084,0.0037761203,0.2984772623,0.2632172108,-0.0586931109,0.3550698161,0.0994175449,-0.0299297981,-0.0066212779,0.2013110071,0.2945919335,-0.1369209439,-0.2663726211,0.3205899894,-0.0480781831,-0.1996354908,0.0502453633,0.0024982328,0.0410067216,-0.2958438098,-0.2911486328,0.2036809325,-0.0922603309,-0.2497835606,0.3618250489,-0.0576844439,-0.1583008319,0.0392347164,0.2182587087,-0.1461775452,-0.1182287484,0.2355463058,0.1425219476,0.0640296489,0.513656497,0.224274531,-0.2006270289,-0.3460781276,-0.2409085929,0.2673728168,-0.1038022265,0.1420608908,0.0205014925,-0.1499487162,0.1193839237,0.300819993,-0.0045858631,0.1239377186,-0.1286375672,-0.0400311612,-0.6460931897,-0.1034621373,-0.088998951,0.2422215194,0.0092601394,0.3586094081,0.1377948076,0.0665734187,-0.4973830581,0.220697403,-0.5212618113,0.2922661602,-0.1465101838,-0.2166813463,0.1004297584,-0.0035990949,0.2994687259,0.3896147907,-0.074753806,-0.4908471107,-0.0459445491,-0.0243124515,0.0860559493,-0.0832673609,0.0259789079,-0.0547310412,-0.1835623682,-0.1643497497,0.1968172342,0.1007408425,-0.0044261739,0.0331208706,-0.0149058402,0.1464602649,0.0368579775,-0.1768917739,0.1195131019,0.1588378102,0.1533807069,-0.2128538191,0.0376653336,-0.0427197553,-0.254727155,0.0933508873,0.0430233963,0.1398325115,0.0582809895,-0.2366780341,-0.1732652634,0.1693043411,0.1497258991,0.1178266704,-0.1118547469,-0.0784648135,0.3277317584,0.484431982,-0.3830759227,0.1091232449,-0.0806196854,-0.2138261497,0.1889745742,0.0428555161,-0.1388588995,-0.1439116299,0.0112759592,0.3013266325,-0.0613593049,-0.3362208307,0.044475425,-0.0171552915,-0.1670493484,-0.0838595331,0.0138555318,0.1262952536,0.2215817422,0.5965085626,-0.0110206893,0.2545818985,-0.0056214551,0.0307637826,-0.0948383659,-0.0897640139,-0.1684496105,0.206345588,-0.1387440562,0.0197989028,-0.1843643039,0.2826740444,-0.0005801157,-0.4360289574,-0.2782769501,0.1547601819,-0.5187679529,-0.0329871699,-0.0443767086,-0.0737793893,-0.083694756,0.0961439461,0.1515669078,0.0504986383,0.2457415015,0.0141567737,-0.2495627403,-0.0792225599,-0.0577606782,0.1140130311,0.0513658859,-0.1550548077,0.3201295733,0.076343812,-0.1493938267,0.0214519445,0.1081909463,0.2388635129,0.3021705449,-0.0504621267,0.0369069688,-0.0089009823,-0.0723595098,-0.1400568038,0.0680067241,0.4152925909,0.3369742036,0.160490796,0.3123735487,-0.34422791,0.2975425124,0.2102182508,0.1393764913,-0.3182028532,-0.043603763,0.0090456614,-0.1056754738,-0.391024977,-0.1055347845,-0.3937880993,0.1430419683,0.0768699273,-0.1664940715,0.0286381505,-0.0788837969,0.1855547279,-0.1037226021,0.4702087343,0.0057124691,-0.0278958734,-0.3736037612,-0.3904382288,-0.3781890571,0.1311614811,-0.1199381202,0.2864547968,-0.1748952419,0.1810536385,-0.0524846725,-0.1311141998,0.037479762,-0.0438918509,-0.0146777779,-0.2192430347,-0.5541431904,0.0495910347,-0.3265770972,-0.1130880937,0.1711770147,-0.0495973378,0.256437242,-0.1949551851,0.3151395023,0.1427249163,0.0903890282,-0.0648952723,0.2487023026,0.2711836398,-0.2061080337,0.0713036582,0.0298020244,0.1918613613,-0.0072645466,-0.1872489899,-0.0796122402,0.0138643757,0.4124436677,0.2714191079,-0.0919050723,0.0775919706,-0.1340719461,0.2105646282,0.2166771144,-0.1141231433,0.0151604693,0.0663868189,0.2164980471,-0.0347402059,0.2126675546,0.3974253833,-0.0095456196,-0.0446493626,-0.4061530232,-0.3422017694,0.5504412651,-0.0954808444,-0.2921683788,-0.1106133461,0.008527982,0.2239868194,-0.1508717984,-0.7015339732,-0.1838160455,0.1246204972,-0.1167333201,-0.3178701103,0.2025784403,-0.1550266743,0.0944320858,0.0175093208,0.2260047793,0.0509908944,-0.062849693,0.1083443537,-0.1644290835]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/752","title":"Clicking on a metric in the search page points to datasets page giving \"Missing dataset\" warning","comments":"Thanks for the report, can reproduce. Will fix","body":"Hi! Sorry if this isn't the right place to talk about the website, I just didn't exactly where to write this.\r\n\r\nSearching a metric in https:\/\/huggingface.co\/metrics gives the right results but clicking on a metric (E.g ROUGE) points to https:\/\/huggingface.co\/datasets\/rouge. Clicking on a metric without searching points to the right page.\r\n\r\nThanks for all the great work!","comment_length":8,"text":"Clicking on a metric in the search page points to datasets page giving \"Missing dataset\" warning\nHi! Sorry if this isn't the right place to talk about the website, I just didn't exactly where to write this.\r\n\r\nSearching a metric in https:\/\/huggingface.co\/metrics gives the right results but clicking on a metric (E.g ROUGE) points to https:\/\/huggingface.co\/datasets\/rouge. Clicking on a metric without searching points to the right page.\r\n\r\nThanks for all the great work!\nThanks for the report, can reproduce. Will fix","embeddings":[0.0138868168,-0.2590185106,-0.1347299814,0.1538688242,0.1985722184,0.0988123268,-0.2235769778,0.1556014419,0.2025790662,0.1700214148,-0.1306263506,-0.1742985398,0.0339337178,0.1640307605,0.0571638308,-0.0835662112,0.1699934453,-0.0082547618,-0.2080666125,-0.0952992737,-0.4587498903,0.0717848763,-0.2841034234,0.2455725819,-0.0775839165,-0.0866024122,0.1250821352,0.1686936617,-0.3200645745,-0.5558127761,0.1122842208,-0.2138780355,-0.0013868241,0.6711270213,-0.0001121588,-0.0697444752,0.3285639286,0.0842434168,-0.0327275097,-0.2659235001,-0.6243675947,-0.3157479763,0.1981473267,-0.0676825643,-0.1213435084,-0.1254370809,0.1018436924,-0.1882414818,-0.1046907827,0.4701806903,0.2347341329,0.3536976278,0.0323269777,-0.2488666773,-0.0622467175,0.0763213634,-0.1959494203,0.0957220271,0.2512545288,0.0655290559,-0.1133638695,0.4912514687,0.0638902783,-0.1059122086,0.4851844609,-0.0290803611,0.0253280234,-0.2186673582,0.2654206455,0.0973471701,0.4821181595,-0.2113437057,-0.1332303733,-0.143647939,-0.3382252157,-0.2315908819,0.304387778,0.1375869662,-0.0842677653,0.1569431573,-0.3331531286,-0.3567634523,-0.0829878002,0.1238143668,-0.2830052078,0.0921453014,-0.191466108,-0.1915433705,0.176922664,-0.1090717316,0.0180362258,0.0765223354,0.144546628,-0.0287802294,-0.4101329148,-0.1142918915,0.1382602602,0.302982986,0.2955027223,0.0316028334,0.0012607392,0.0925565809,-0.1566217542,-0.1034472436,0.3471851945,0.058796756,0.0542681441,0.0732906535,0.1920351237,0.1301513016,0.1607804447,-0.1340001822,0.5133313537,-0.4375089109,0.0579601116,-0.0705342516,0.272367686,-0.3061744571,-0.5156134367,0.1673241258,-0.125966832,-0.1280092299,0.0948597491,0.343421936,-0.1425359994,-0.2213729322,-0.0683271661,0.3447940946,-0.3484070301,-0.3326973915,-0.2930971086,0.1101564094,-0.2703250051,0.1143047959,0.1177247539,-0.1018946618,0.2040361762,0.0454400666,-0.0375711434,-0.0721766576,-0.1242717355,0.025850812,-0.1280767918,0.2041968852,0.1413629055,-0.0448762439,0.1241032481,0.0092285341,-0.0333393291,0.2294922471,-0.473228544,-0.0969714671,0.1021313369,0.2210151255,-0.2891647518,-0.2120722383,-0.0134912003,0.430652678,-0.2602975368,0.1719716787,-0.0706087127,0.1512968838,-0.3660476506,-0.0205889419,0.2257571369,0.2947001457,0.1470671147,-0.2842588127,-0.0640910193,-0.1619671285,-0.1329323351,0.1376949698,-0.0250112712,0.3163910508,-0.3049425781,0.1078906581,0.5167884231,-0.4385119081,-0.130096063,-0.0733902901,-0.3255558312,-0.2133613378,-0.2720314562,-0.1520033479,0.0006657221,0.1230984181,0.2300811559,-0.3523473442,0.0341193862,0.0435651541,-0.3213861585,-0.0132055627,0.2244569212,-0.1685528904,0.4157805443,0.2149371356,0.3697504997,-0.25347808,0.0077152187,0.057655286,-0.0770615116,0.3272648752,0.2152714431,-0.0878648683,0.0096301353,-0.1986258924,-0.1629843861,0.0412661023,-0.1565166563,0.0561219417,0.058598645,0.0796728805,-0.5484156013,-0.0530382283,-0.1839652061,-0.072971493,0.0743235797,-0.2555933297,-0.1613328457,0.3258731067,-0.1683218628,-0.042085737,-0.215356946,0.3617949486,0.0008099945,0.129815951,-0.0014142877,0.4188789129,0.0864574984,-0.0334956124,0.4393720925,0.2154636383,-0.0967995524,0.251155138,-0.0530488975,-0.0240522455,0.6581953168,0.3323626816,0.4566036463,-0.7646163106,0.2024465799,0.0395838991,-0.005007342,0.2245073915,-0.1753560752,0.0469006598,0.0223515518,0.252217114,-0.2134652883,0.2206035703,0.1634398848,-0.2020428926,0.1020906344,-0.1159022674,0.1398020536,-0.0851840004,0.2492752075,0.0660372823,-0.3684417605,-0.0726895407,0.1136137173,-0.2576752901,0.0933288336,0.4077852964,-0.2487524152,-0.1085800976,-0.1217024103,0.0161231011,0.3228662014,0.2857126296,0.2518514693,0.1494778693,-0.0738979802,-0.2204920948,0.087503843,0.2965078354,-0.381850183,0.1417827457,-0.0477359593,-0.1156477779,-0.1887676269,0.3320029974,0.0115100844,0.0464113392,-0.4192179143,-0.1865996718,-0.1603645682,-0.1719167829,-0.2277240902,0.1517448425,-0.5471202135,-0.4959767461,0.348659873,-0.0376952551,0.0352131426,0.2497101873,-0.3452328742,0.4408254027,0.051136639,0.1350287646,-0.0756923333,0.1524094939,-0.0919968933,0.1025347486,-0.0352046899,0.2418362647,0.1730639637,-0.3817080259,0.4543709159,-0.6356626153,-0.3222495914,0.2063115537,0.117913641,0.1645546705,-0.01169129,-0.0893508345,0.0092501193,0.3791496158,0.1565129459,0.0629256889,-0.2390480489,-0.1946367919,-0.0083144177,0.1837930828,0.1055591181,-0.213346377,-0.0975819379,-0.0944089219,-0.2051437944,-0.1617653519,-0.1362751722,0.2040036619,0.0420273654,0.0316251777,0.0632561594,0.1813753992,-0.4870892465,-0.3373370171,0.077069521,-0.0948093385,-0.2629598379,0.1894202828,0.141778633,0.066582948,0.1766996831,-0.542996943,-0.2068535388,-0.1684780419,-0.2284798175,0.0978550687,0.0305536855,0.0929218158,-0.1951426119,-0.1861207187,-0.2495574504,-0.2915457487,0.0088632619,-0.0878552571,0.5628396869,-0.1143353283,0.0270545911,0.1435004473,0.5649107695,0.2356552333,-0.1392944008,0.2578812242,0.0185435582,0.6561044455,0.0024774543,-0.2151874602,-0.1151194498,0.1803146452,0.0790854394,0.0743066818,0.3040091693,0.0523759909,-0.2171332538,-0.1585350633,-0.2101048231,-0.2344189584,-0.2728164792,0.0344149359,0.1841323972,0.345754832,0.2644370496,-0.123717621,0.0309429523,0.1255777627,0.242399931,0.2875377834,0.0246072821,-0.1096633077,-0.0449165739,-0.5212144852,0.3016580343,-0.0729278699,-0.0076042265,-0.2869568169,-0.2215294391,0.1890176833,0.2778476477,0.4952213168,-0.3938258886,0.0608816072,-0.1987864971,-0.0348906405,-0.1339586526,0.3143748045,-0.016142495,0.4549722672,-0.0127728265,0.4092468619,-0.0701501295,-0.4068935513,0.187811479,-0.0487689413,-0.0341018513,-0.0475005172,-0.2480164468,-0.1279268712,0.0413530283,0.2280550152,0.2087482959,-0.0270694662,0.2315495759,-0.1615378708,-0.0187651552,-0.1108200401,0.1051531211,-0.0419523977,0.3017169237,0.1175907254,0.1330995113,0.5209602118,0.3547775447,0.1849462539,0.4250838459,-0.099112533,-0.4459905028,0.1588593125,0.0841009915,0.1232668459,0.4121253192,0.2314698249,0.0252769943,0.1604672521,0.2962163091,-0.3389381468,0.1902681738,0.233384937,0.26177302,-0.3931213617,-0.0320051871,0.2159645408,0.0318484567,-0.253151685,0.4241885245,0.3781324029,-0.2632953525,-0.3974777162,0.0153957289,0.8766179681,0.0272058155,-0.2556518614,0.05060203,-0.1025109738,0.2426203936,-0.1184162125,0.0312974676,-0.1022053882,-0.3698687851,-0.0571034588,-0.0982570648,-0.0525173955,-0.1650258899,0.1924425811,0.5062782168,-0.1309405565,0.278110981,0.1631657928,0.0848827139,0.220494777,0.0068757953,0.0876217335,0.1725884676,0.0049387976,0.5381197333,-0.0339917466,-0.0510313921,0.0884594098,-0.1463930607,-0.147344172,-0.2146735191,0.0431149639,0.1001232266,0.0116914902,-0.183703959,0.3398163617,0.3559016883,0.4361775219,0.0292170346,-0.5305453539,0.5314522386,-0.3548872471,0.1610069275,-0.0287373755,-0.0326257087,-0.0348709151,0.009779769,-0.2539579272,0.0726194307,0.0279433783,-0.2203007191,-0.3428987265,-0.0455507115,-0.491561234,-0.3170394301,0.1464278549,0.060387753,0.2865716815,-0.1065720618,0.1475594938,0.064781636,0.2985531986,0.0589611754,0.4142215252,-0.3429290056,-0.1178766415,0.257171005,-0.0984261036,-0.3740284443,-0.0241161417,-0.0838117599,-0.2762058377,-0.2640267611,0.112261638,0.2420338839,-0.0438908786,0.0306530483,-0.193767041,0.1903484613,-0.0682811812,0.1630147845,0.1603297889,-0.3017082512,0.1072248593,-0.1216941178,0.0230742805,0.4630523324,-0.2493902445,0.0291298125,-0.3251560628,0.1815855652,0.39885813,-0.3054764569,-0.3343915343,0.1613800526,-0.1297575533,-0.1453749686,-0.0501280688,-0.0263983607,0.2151660174,-0.1293588728,-0.0346198976,-0.1737166643,-0.1804287434,-0.172140494,0.1392081678,0.0994416326,-0.0129045807,0.0999450982,-0.1333156675,-0.1412257552,0.0059944615,-0.0629047528,-0.1564326286,0.3826382756,0.0335984342,-0.1853412837,-0.2143182158,-0.1816588789,0.143053785,-0.1348234862,0.2602129579,0.1578761786,-0.0716822818,-0.1284870505,0.000095592,0.0415329002,-0.1102594361,0.33677122,0.2184362113,0.1030960381,-0.0407130793,0.0006751731,0.1385269165,0.3928380311,0.3524151444,-0.0167350899,-0.3143013418,0.0091413651,0.0789793283,0.2420702577,-0.3307174742,-0.1999592334,0.1457009017,0.0849538371,0.0688222125,-0.0044089006,0.1601569802,0.2547399998,0.4330819249,0.0481610037,0.4658231437,-0.0778833926,0.3188324571,-0.152422592,0.0746219978,0.0817858055,0.4875438213,0.0766537637,0.0018153784,0.6047698855,-0.2799504399,0.2457928509,0.0583588555,0.2884169519,-0.1010561809,-0.1215472594,-0.0148956776,0.150306195,-0.3355039358,-0.3681937158,-0.132318303,0.3470921814,-0.2068569958,-0.2244766504,-0.4540046751,-0.0195716042,-0.0378560424,0.1883051097,-0.26399979,-0.3991151154,0.1173588857,-0.0250790101,0.1665721983,-0.1909920126,0.1038947776,-0.3718858659,-0.1558714956,-0.3581365943,0.008100668,0.4263917208,0.1492312104,0.0154736759,-0.0535448268,-0.0502366349,-0.0687149987,0.3836861551,0.4492231309,0.1185536012,-0.0378953554,-0.0495653301,-0.1176095754,-0.106166631,-0.1822034866,0.0889492407,0.1060332656,0.0586003996,-0.4294810891,0.2577261925,0.0968336314,-0.1383806765,0.401961416,0.1180521399,0.5321746469,-0.2780930996,-0.1293099523,-0.2602621913,0.2574866712,-0.2856926918,-0.3864430487,-0.4004351795,0.1375137568,0.2005814761,-0.2451538444,0.1301847249,-0.190572679,0.0827250257,0.2242900878,0.2528814077,0.2072481662,0.1714077592,-0.3620870113,-0.1552054435,-0.5934864879,0.1075956821,-0.0794775933,0.2288858593,0.2855836749,0.0649324208,-0.2747042775,0.2795449793,0.4814769626,0.4453721941,0.232341826,-0.4066531956,-0.424007684,0.0271779206,-0.0387978107,0.3199112713,-0.0400011502,0.0899584889,0.220943585,-0.2710217237,-0.0378240868,0.1284059435,0.2410328239,-0.1397734582,0.3584252596,0.2700154185,0.0986824781,0.5671448708,-0.0092687747,0.1239897236,0.0181892086,-0.4957275689,-0.1984972358,0.1498145312,-0.0497997999,-0.1819164157,-0.0589374267,-0.3803367317,0.0191903058,0.4631288946,0.3843745887,-0.0380848795,-0.3145304322,0.3210884333,0.0990458503,-0.0566293299,0.0738087595,0.065096274,-0.016311273,0.3024105728,-0.0568258092,-0.3829844296,0.4628623128,-0.1051903665,0.0682889745,0.0045278464,0.2093756199,0.6467360854,-0.226966247,-0.497743845,-0.1021280736,0.2542308271,0.225624755,-0.1511065215,0.1385073513,-0.4572719038,-0.0190929901,0.0450753458,0.2390248477,0.2790387273,-0.3925626576,0.3229399323,-0.0046961461]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/752","title":"Clicking on a metric in the search page points to datasets page giving \"Missing dataset\" warning","comments":"Fixed now @ogabrielluiz ","body":"Hi! Sorry if this isn't the right place to talk about the website, I just didn't exactly where to write this.\r\n\r\nSearching a metric in https:\/\/huggingface.co\/metrics gives the right results but clicking on a metric (E.g ROUGE) points to https:\/\/huggingface.co\/datasets\/rouge. Clicking on a metric without searching points to the right page.\r\n\r\nThanks for all the great work!","comment_length":3,"text":"Clicking on a metric in the search page points to datasets page giving \"Missing dataset\" warning\nHi! Sorry if this isn't the right place to talk about the website, I just didn't exactly where to write this.\r\n\r\nSearching a metric in https:\/\/huggingface.co\/metrics gives the right results but clicking on a metric (E.g ROUGE) points to https:\/\/huggingface.co\/datasets\/rouge. Clicking on a metric without searching points to the right page.\r\n\r\nThanks for all the great work!\nFixed now @ogabrielluiz ","embeddings":[0.0023028697,-0.2306553572,-0.1354810148,0.1818616241,0.2275328487,0.0867888853,-0.2079419196,0.165363431,0.2031517625,0.1696141958,-0.1338950098,-0.1912556589,0.0348800272,0.1646399945,0.0446574539,-0.0706809163,0.1153621003,-0.0191359781,-0.1752306521,-0.0575763956,-0.4398697019,0.0914392918,-0.2647807896,0.2195227742,-0.0668927878,-0.0671380162,0.0965758711,0.1989919841,-0.3605392575,-0.5699805617,0.0917942971,-0.178452298,0.0244314522,0.6447627544,-0.0001094502,-0.050991755,0.3402236998,0.0670220181,-0.0576732457,-0.253524363,-0.5838222504,-0.3026655912,0.1710213125,-0.0628319979,-0.1704299003,-0.1314768046,0.1050626487,-0.1826975048,-0.0618696846,0.4960189164,0.2518434823,0.363691926,0.0319610313,-0.2908335924,-0.0623068511,0.0617693551,-0.2042548209,0.0801228881,0.2696799934,0.0941962525,-0.1237577125,0.4899204373,0.056628041,-0.1195067242,0.4883194268,-0.017767841,0.023276709,-0.189586401,0.2188418061,0.1154517159,0.472109884,-0.1763438135,-0.1078622341,-0.1747105271,-0.323205024,-0.2698458731,0.2927730083,0.0936331674,-0.0584217682,0.1300390363,-0.3686936498,-0.3476693034,-0.0801590532,0.1457185,-0.2750630081,0.1242110431,-0.1700812131,-0.1821606159,0.156921491,-0.1161096916,0.0043143509,0.1120286882,0.0779852793,-0.017196279,-0.3803426921,-0.1258677691,0.1613702029,0.2848857343,0.2939622402,0.0497444645,0.0195055529,0.0878601894,-0.155617699,-0.0952707231,0.3300313354,0.0799922869,0.1180992275,0.0543439165,0.2014047951,0.1601124108,0.1840344965,-0.1336996853,0.504566431,-0.4122618139,0.0319722816,-0.0455038212,0.2973310351,-0.2877683938,-0.4887247682,0.1386820227,-0.0908987373,-0.1207335591,0.1029199213,0.3420713842,-0.1414200217,-0.2397079617,-0.0565237328,0.3359407187,-0.3226137757,-0.3531973064,-0.317738831,0.0719576553,-0.2513972223,0.1021669284,0.1329724491,-0.1006291509,0.2405451387,0.032980904,-0.0241792519,-0.0739487037,-0.1438767016,0.0670374408,-0.123554416,0.2192181051,0.1385983229,-0.0621343553,0.1423881054,-0.0408366732,-0.0151597774,0.2219846249,-0.4897268713,-0.0887431651,0.0669954866,0.2507590652,-0.3006451726,-0.2070492357,-0.0504175499,0.4620119631,-0.2468333393,0.1387707144,-0.0540939867,0.159329921,-0.4060099721,-0.0042211385,0.2438665628,0.2932732105,0.1292285472,-0.2584050298,-0.0603350215,-0.161602065,-0.1436241567,0.1365978271,-0.0399352238,0.3282317817,-0.3203574121,0.1281061023,0.5187843442,-0.4534321129,-0.1000890657,-0.088338159,-0.3034232557,-0.1969704479,-0.3123200536,-0.1738588661,0.0200536549,0.0949194655,0.2261722982,-0.3596830368,-0.005082624,0.0433255173,-0.3273943365,-0.0253378935,0.2049774975,-0.1705673188,0.4226075709,0.2374582291,0.3301587403,-0.2527008355,0.0299650338,0.0706593096,-0.0791709423,0.3613068163,0.1958260983,-0.0616118871,0.0043172496,-0.204345867,-0.1808377951,0.0773058534,-0.1674150378,0.0912185982,0.0609391443,0.0700867251,-0.5627890825,-0.0772305354,-0.1667735875,-0.1009017602,0.0992917493,-0.2719451785,-0.1299392134,0.3075654209,-0.1438613832,-0.0408475138,-0.1891775429,0.3575999737,-0.0767064467,0.1071658731,-0.0135074761,0.3886426985,0.0864851251,-0.0374134555,0.4643843472,0.2118619084,-0.1166931391,0.2976281643,-0.0637838319,-0.0465701297,0.631880939,0.3136090934,0.4275024235,-0.7101761103,0.1736825854,0.0379729606,-0.0100220926,0.1846601665,-0.1775794774,0.0276283845,0.043380376,0.234590888,-0.2275471836,0.1989266276,0.1756252944,-0.1858505756,0.0711626485,-0.1342663467,0.1132211089,-0.0945369676,0.2001928985,0.0490385443,-0.3649002612,-0.0547548532,0.1446225196,-0.2434924394,0.0721579567,0.4042616189,-0.2328485847,-0.1042668149,-0.1208497435,-0.0099619282,0.2993220985,0.3014070392,0.2439046353,0.1666931957,-0.061787229,-0.2304273546,0.0946041793,0.27961573,-0.3212819397,0.1164422184,-0.0377987698,-0.1462145001,-0.1943930238,0.3090232909,-0.0196960587,0.0743089914,-0.434186548,-0.2063901871,-0.1760717183,-0.1746907383,-0.2031081915,0.1259508431,-0.5335823298,-0.5055249929,0.405862391,-0.0212045554,-0.0181558058,0.226176545,-0.3213864267,0.4223285317,0.0493235067,0.1179481745,-0.1025234312,0.0877846777,-0.0806450844,0.1173336953,-0.0396733619,0.2047053128,0.1800265461,-0.389813602,0.4231034815,-0.6081802845,-0.3277659416,0.1924491972,0.1024923772,0.1590022594,0.0030540912,-0.0280342866,-0.0167467259,0.3596761227,0.2014616132,0.0392616317,-0.2380069643,-0.1642080843,0.0407157838,0.1763277352,0.0692413524,-0.1884820461,-0.0871755481,-0.1194871217,-0.2222726196,-0.1375467777,-0.1288262159,0.2072986215,0.0730284527,0.045698639,0.0704394802,0.1472104043,-0.5000901818,-0.3758772314,0.0928888991,-0.113998346,-0.2680042982,0.1808676124,0.1667085141,0.0679923669,0.1402273923,-0.5287048221,-0.2205048352,-0.1306190342,-0.2091549486,0.1171096191,-0.0036215787,0.1019535437,-0.1935956329,-0.2114095837,-0.2377591729,-0.2698810399,0.0045319027,-0.1162334457,0.5565574765,-0.1421591938,0.0429807827,0.1278697848,0.5696756244,0.2332948595,-0.1159546152,0.2539792061,0.0192488059,0.6769933701,-0.0335285924,-0.1946286559,-0.0998400152,0.1951596588,0.0854586661,0.0784659162,0.2860322297,0.06798812,-0.2056694627,-0.1146797016,-0.2246868908,-0.2557660937,-0.2503631413,0.0029132816,0.1255344748,0.3458967805,0.2812554836,-0.1015462056,0.0268193968,0.1510512382,0.2511417568,0.2503772974,0.0754797384,-0.1089869291,-0.0044796756,-0.520232439,0.3248206675,-0.1001469269,-0.0265436545,-0.2850501239,-0.2246510684,0.1604328901,0.2622739077,0.5027706027,-0.3649117947,0.012957409,-0.1990551203,-0.0187044553,-0.1456762105,0.2665476501,0.0007430427,0.4129953682,0.0100469384,0.417440325,-0.0901834816,-0.3750605881,0.1587879807,-0.0598073527,-0.0635226518,-0.0554588549,-0.2541932464,-0.1780851036,0.0169726368,0.1952089667,0.2298783809,-0.0401988998,0.2167811245,-0.1523698866,-0.0112173241,-0.09734267,0.1173696518,-0.0302589759,0.3101910651,0.1290655434,0.1133010238,0.5366524458,0.3863175213,0.182556808,0.4396431446,-0.1022798195,-0.4324796498,0.1263654083,0.0807645619,0.0917962715,0.3825279772,0.2282810807,0.0098320758,0.1796298325,0.3414626122,-0.3486413062,0.2161120176,0.2158865184,0.2247284055,-0.3792811036,-0.0512785129,0.2140853405,0.0084243976,-0.228031829,0.4474707544,0.3726086915,-0.2609837353,-0.3765658736,0.0108837103,0.8559231162,-0.0005075571,-0.2761902213,0.0712725818,-0.1165760905,0.2547118664,-0.1197197512,0.0482202433,-0.1043665707,-0.3465489447,-0.0485959649,-0.0576840341,-0.0557905734,-0.1597695351,0.1670924574,0.5044968724,-0.1185804307,0.2639591098,0.139691934,0.0732899159,0.2079769224,0.0158501975,0.0777862743,0.204560414,-0.0151417032,0.5186557174,-0.0579088293,-0.0436114557,0.0813793391,-0.1458067298,-0.1594888866,-0.1949306726,0.033228416,0.1136221662,0.0198689085,-0.1685597003,0.2804581523,0.3684002459,0.4441899955,0.0482571758,-0.5383098125,0.5487598777,-0.3262110353,0.1341676116,-0.0619190596,-0.0485032834,-0.0415797234,0.0033510407,-0.2498645335,0.0799362212,0.0348032117,-0.2203081399,-0.3699318171,-0.0346913971,-0.5010609031,-0.3089067936,0.1500739902,0.0208260864,0.299061358,-0.1306672543,0.1625590175,0.0808873624,0.2913474143,0.0193266235,0.4137648046,-0.336830467,-0.1136792004,0.2434334159,-0.0914811864,-0.3866438568,-0.0126687288,-0.0684444681,-0.2787750065,-0.2640400231,0.1114029512,0.2581952512,-0.0393032879,0.0552805625,-0.1659417748,0.1620966345,-0.0451273322,0.1627172232,0.134771198,-0.3041498065,0.1274490803,-0.1414226741,-0.0255084429,0.4440074563,-0.229393959,0.0552743226,-0.2753170431,0.1799018085,0.3585790098,-0.2491911501,-0.3549217284,0.1406505704,-0.1469866782,-0.1394169778,-0.0343351588,-0.0323173627,0.2225940377,-0.1500373483,-0.0201509856,-0.1408224851,-0.1874470115,-0.1893225014,0.1419792771,0.0818267092,-0.0069420733,0.0750400648,-0.1218417659,-0.0928732455,0.040127743,-0.0823528022,-0.1524798125,0.4031577706,0.0560126901,-0.1782039851,-0.2115263939,-0.161742866,0.1801910847,-0.1581651866,0.2849455476,0.1782688051,-0.0564316772,-0.1523879766,-0.0342132747,0.0385251082,-0.057699047,0.3388547003,0.2247945517,0.1196242124,-0.0190725345,-0.008619885,0.1082208455,0.3977595866,0.3679237068,-0.0308472253,-0.325355798,-0.0081412205,0.0813816264,0.261613369,-0.3292872012,-0.1788344383,0.1698093861,0.1018564776,0.084951058,0.0138748009,0.1279476136,0.2910418212,0.3879586458,0.0666546524,0.4660071731,-0.0803254321,0.3218431175,-0.1814601421,0.0322220884,0.0252708402,0.4613326192,0.0716954842,0.0092472266,0.5923385024,-0.2822355032,0.2668382823,0.0189998075,0.2631135583,-0.0966499373,-0.1353506744,-0.0080306139,0.1296001673,-0.3205666542,-0.3387644589,-0.1115160808,0.3212869763,-0.2175656706,-0.2407742441,-0.4470663071,-0.0062094466,-0.0577268712,0.1718736589,-0.2729333937,-0.3960034847,0.1182654127,-0.0327976421,0.1930080801,-0.190587461,0.098375544,-0.3794196248,-0.1739891768,-0.3663709164,-0.0150904153,0.3879541755,0.1374263018,0.0233078152,-0.0341822654,-0.0254400652,-0.0956813768,0.3887946308,0.4251699746,0.13235116,-0.0252876356,-0.0022768874,-0.1055460423,-0.0846148059,-0.1804338694,0.0327158049,0.0998874307,0.0631109029,-0.4226819277,0.2813698053,0.1210618317,-0.1560002118,0.4273279309,0.1183545887,0.5431877375,-0.2176530063,-0.1033181176,-0.2633710802,0.2425056994,-0.2749055326,-0.3826209903,-0.3894530535,0.1006851569,0.1928274184,-0.2198851258,0.1340947598,-0.2175198346,0.0968967229,0.2427660823,0.2664282024,0.2357715815,0.1567135453,-0.3701629639,-0.137179628,-0.6221871972,0.1140299737,-0.1111420169,0.2323836982,0.2777174711,0.0850160792,-0.2349454314,0.2610255182,0.4957269728,0.4553940296,0.2559295893,-0.4363835156,-0.4250341654,0.0499435104,-0.0421162322,0.2826510668,0.019021349,0.0934680179,0.2072382867,-0.2875594795,0.0092391223,0.1324151754,0.2533423901,-0.1209777221,0.3452447653,0.2592053115,0.1172510087,0.5166373849,-0.033070717,0.1258856803,-0.0109232487,-0.4550628066,-0.2270402312,0.1489707381,-0.0394911245,-0.1314862072,-0.0524612553,-0.3916349411,-0.0254797731,0.4709387124,0.3663428128,-0.067717053,-0.3319285512,0.3367603123,0.0771918446,-0.0457301214,0.0846281275,0.0249752849,-0.0059604342,0.2916124463,-0.0523504615,-0.4012836814,0.5153747201,-0.105760783,0.0620930381,-0.0120069431,0.1945961416,0.6587357521,-0.2018281966,-0.5411317348,-0.0909795314,0.235832572,0.2214307487,-0.1303569376,0.1029795185,-0.4471551478,-0.0133594703,0.0273675695,0.2227484435,0.3146625161,-0.3989068568,0.3025538921,-0.0041214293]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/751","title":"Error loading ms_marco v2.1 using load_dataset()","comments":"There was a similar issue in #294 \r\nClearing the cache and download again the dataset did the job. Could you try to clear your cache and download the dataset again ?","body":"Code:\r\n`dataset = load_dataset('ms_marco', 'v2.1')`\r\n\r\nError:\r\n```\r\n`---------------------------------------------------------------------------\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-16-34378c057212> in <module>()\r\n 9 \r\n 10 # Downloading and loading a dataset\r\n---> 11 dataset = load_dataset('ms_marco', 'v2.1')\r\n\r\n10 frames\r\n\/usr\/lib\/python3.6\/json\/decoder.py in raw_decode(self, s, idx)\r\n 353 \"\"\"\r\n 354 try:\r\n--> 355 obj, end = self.scan_once(s, idx)\r\n 356 except StopIteration as err:\r\n 357 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 388988661 (char 388988660)\r\n`\r\n```","comment_length":31,"text":"Error loading ms_marco v2.1 using load_dataset()\nCode:\r\n`dataset = load_dataset('ms_marco', 'v2.1')`\r\n\r\nError:\r\n```\r\n`---------------------------------------------------------------------------\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-16-34378c057212> in <module>()\r\n 9 \r\n 10 # Downloading and loading a dataset\r\n---> 11 dataset = load_dataset('ms_marco', 'v2.1')\r\n\r\n10 frames\r\n\/usr\/lib\/python3.6\/json\/decoder.py in raw_decode(self, s, idx)\r\n 353 \"\"\"\r\n 354 try:\r\n--> 355 obj, end = self.scan_once(s, idx)\r\n 356 except StopIteration as err:\r\n 357 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 388988661 (char 388988660)\r\n`\r\n```\nThere was a similar issue in #294 \r\nClearing the cache and download again the dataset did the job. Could you try to clear your cache and download the dataset again ?","embeddings":[0.010368593,0.1111898348,-0.1000250876,0.3606889546,0.4273394346,0.051915016,0.2879604995,0.5654810071,0.0125486925,0.3217283189,-0.0122815529,0.5649032593,-0.0228346568,0.2471558452,-0.1524105668,-0.3327514231,-0.0393442959,0.3240633905,0.130652234,0.1089469045,-0.0823314264,0.2334581763,-0.1433686763,0.234962821,0.0887256414,0.1099612266,0.0444128998,0.1186437234,-0.3717558682,-0.6237437129,0.2395816296,-0.0147982743,0.3841393888,0.4122720063,-0.0001079264,0.2034292966,0.4274071157,-0.1497963965,-0.328352809,-0.6941396594,-0.5283315182,-0.0733867437,0.330268383,-0.0713751391,-0.0157223493,0.0518315434,-0.2022550702,-0.4683452547,0.7235790491,0.3449121118,0.2332075983,0.0984003097,0.2272278368,-0.0567691959,0.1948496848,0.1532230824,0.0622134432,0.2472892404,-0.1464816779,0.2154764235,0.2918561697,0.2464020252,-0.1134709418,0.3241589069,0.1573538631,0.1225416735,0.3648796678,-0.4494427443,0.3054109216,0.1438313425,0.6902774572,-0.1550350785,-0.2551623285,0.076748848,0.1294802725,-0.4748290479,0.3319576681,0.0643206164,-0.2432660609,0.1752771735,-0.0887585357,0.0134896645,-0.2744025886,0.0100894226,-0.1931982934,0.0089537222,-0.1390662938,0.1527302563,0.3419303894,-0.2617520988,0.0862084851,0.0477184989,-0.2760582864,0.1573705971,-0.193217203,-0.1091299579,-0.1569622159,-0.4858180881,-0.0533559658,0.0962756202,0.2548743784,0.0609164909,0.3814300299,0.0362790786,0.6026726365,0.0255676787,-0.1211960316,0.0345732681,0.1222294345,0.0011011522,0.0688749999,-0.080301553,-0.2731036544,0.0020022115,0.1748583615,0.1447976679,0.3112102151,-0.2800496817,-0.338255614,0.1514056474,-0.5848353505,-0.2422351837,-0.2254050672,0.3102692664,0.0200363994,0.243372798,0.0605809614,0.2738599777,0.0199087597,-0.345985949,-0.0432902537,-0.151647523,-0.034664046,-0.1455734074,0.2728931308,-0.0849440917,0.1361089945,0.0872814953,0.1812023669,-0.3393514752,0.3095952272,-0.2254335731,0.0995342582,0.322071135,0.2754564881,0.0525145791,0.1816451102,-0.1673370153,-0.0530319996,0.3141585588,-0.2938114405,-0.2666989565,-0.272241056,0.2071204484,0.0013436769,-0.1711638719,-0.2540102899,0.0457333252,-0.0178011265,-0.0603865162,-0.3778084219,-0.166707918,0.0880220011,-0.1780563593,0.3932457268,0.2161404043,-0.3279469013,0.0394064449,-0.0959152654,-0.1318071783,0.2154319882,0.1643887758,-0.3142090142,0.5403991342,-0.0388913788,-0.1691236645,0.0986142009,-0.3066726029,-0.2937805653,0.3694313169,0.0046840599,0.1591133624,0.0017894151,-0.2512532771,0.3732081652,0.1588671207,0.1945876777,0.0618088804,-0.1427410543,0.091400139,-0.0991631299,-0.2575182319,0.0906345546,0.3011377454,0.0775809586,0.0796924233,-0.0985772461,0.2593408823,0.1528244913,-0.1152997166,0.125785172,0.1634820998,-0.1097195745,0.163519904,0.0510325208,0.1401446462,-0.2065712661,0.0699192733,0.041585654,0.2281046808,-0.3960620165,0.0385790132,-0.3178732395,0.0104322629,-0.3651974499,0.1068749055,0.2487116605,0.1227117181,0.0558310673,0.1275044233,-0.3013746142,-0.0274712723,0.3916778266,0.1682389826,-0.0519139394,0.2938364148,-0.1912588775,-0.2941619158,0.1120572463,-0.0442084298,0.0378200486,-0.1330128312,-0.3254247904,0.237740919,0.1895778179,0.2034036815,-0.1667607874,-0.1456785649,0.1348054856,-0.3300461173,0.083804056,0.5397526622,0.1680045724,-0.0221341699,-0.1026687026,0.1981106699,-0.0855673701,0.1137717143,-0.0782797113,-0.0304033551,0.3267214596,0.0816080049,-0.175970912,-0.2072336674,0.262455821,-0.0087886006,0.0951394513,-0.1814563274,-0.2600827813,-0.2177453041,0.2487670183,-0.078285411,-0.0508634597,0.1320009679,-0.0949898735,0.0630510747,-0.2078442574,0.0244821422,0.2064062655,0.0302203465,-0.2265804112,0.1264622509,0.1852026284,-0.0127630159,0.2005794495,0.1897395104,0.3189218044,0.273775816,-0.0581553727,0.0075665601,-0.2234509587,-0.1390782744,-0.110349752,0.450409919,-0.2590261102,0.1004737243,-0.2818918526,-0.0750758201,-0.1346456259,-0.4714370966,-0.0931167752,-0.0985968709,-0.2331870943,0.1405063719,0.0831357315,0.0812237486,-0.102752991,0.070249185,0.1847392172,0.0335136652,-0.1212211922,-0.3000424504,-0.2222504616,0.0501412116,0.3008953631,-0.3179346919,-0.0289157107,-0.402251184,0.0792943239,0.2344600856,-0.2420565039,0.2829415202,-0.2433229536,0.1071463674,0.0976634547,0.3463149965,-0.0807698891,-0.1787151247,0.2980033457,-0.033628013,-0.0701223835,0.2351943105,-0.024764929,0.0044142492,-0.0345524028,-0.6003323793,-0.2259482443,-0.1773615479,0.084404625,0.0945687369,0.0720440298,0.2071624398,0.2518556118,0.1329909265,0.1408315301,0.2209936231,-0.2139878869,-0.4691381752,0.1890761405,-0.1856191903,-0.288592279,0.3380995095,0.1319410056,0.3926619589,-0.3039434254,-0.3875097334,0.311933279,0.1138057634,-0.1046722755,-0.4057462215,-0.3225399256,0.1243585423,0.1816706359,0.0121639222,-0.0981237888,-0.2578531802,0.2355495989,-0.0515550934,0.1400612742,-0.0608264059,0.8555181623,-0.0773985311,0.328720808,0.2119135261,0.1571446508,0.2867904603,-0.0260023903,0.0474580079,-0.0931168646,-0.1263145804,-0.0346654542,-0.1132034808,-0.1697463691,0.1050458252,-0.1065763086,-0.2495369911,-0.3681685627,0.1578814089,-0.187519595,-0.1722089052,0.1125244051,-0.068363376,0.2198690325,0.0455243178,0.2872373462,-0.1594828665,-0.0221999772,-0.0674551055,0.0986637026,0.2741688192,0.0888988301,-0.5317465067,-0.2367526442,-0.4396628439,0.1123363525,0.0749688223,0.4724262953,-0.1391905844,-0.0173861589,0.234149158,-0.2088274211,0.4418927133,-0.0526613258,0.0111641213,0.1830281019,0.0017400355,-0.2354272008,-0.0946633294,-0.3305482268,0.3097439408,0.1387432218,0.2339223921,-0.2818982303,-0.1609295011,0.3215728998,0.0054150871,-0.203893736,-0.1696620882,-0.5031234622,-0.3280936778,-0.5201476812,-0.1598611176,-0.1911194474,0.4015160203,0.1686213166,-0.031186929,0.1380336136,-0.0285759922,-0.0021822425,0.4222212434,0.270342648,0.0837394893,0.1865050048,0.3958322406,-0.1319777668,0.4243676364,0.5508012772,0.1134043336,-0.6399230957,0.051218763,-0.2962983549,0.0976137072,0.0014407089,-0.0816591159,-0.049380444,-0.1000895947,-0.0209872276,-0.4086267948,-0.0042957924,0.5303221941,-0.2322350144,-0.4098721743,-0.4350350499,0.1329607368,-0.1702174097,0.035307534,0.0269357618,0.0334991552,-0.214471519,0.3636551201,0.1855357289,0.7658941746,-0.0235149823,0.0016502406,0.306660682,-0.4282897711,0.0305051953,0.0277304538,0.2228553891,-0.2719286978,0.0700471476,-0.0694496185,-0.0195545442,0.2279199511,0.0355337411,-0.4261459708,-0.043672625,-0.0445879959,0.1533707529,0.071881786,0.2442642897,0.0587790422,-0.127570048,-0.299549073,0.1998344809,-0.4082095027,0.1917597353,-0.1199819669,-0.2695819438,-0.0086304098,-0.2286145538,-0.1836059242,0.0741830468,-0.1076028123,0.1494814008,-0.0748655275,-0.2504875362,-0.0508267693,-0.1125723496,0.0799794644,-0.0627262667,-0.1535194963,0.1907380521,-0.1253462881,-0.074596405,-0.0077030445,0.2273984998,0.0108062783,-0.0616296716,-0.3674403131,0.3495846987,0.0062730093,-0.3701888323,-0.0235795155,-0.0904286876,0.0961809233,-0.0933389217,0.000480224,-0.1861479878,0.0908050835,-0.0614479445,0.1880188286,0.0396146439,-0.1077995822,0.0597769059,-0.0855291337,-0.3417269886,-0.022280246,0.4954523146,0.0744508952,0.0424285121,0.672642827,0.1097135469,0.0233121216,-0.2028910965,0.4121150076,-0.1919962466,-0.1603666246,0.1369841695,-0.161854431,0.2158029079,-0.0246122405,0.3283049762,0.1497757435,-0.1560216844,0.1812472045,-0.398331821,-0.359377116,0.080186367,0.0996654406,0.089029707,0.2458856404,0.1987779588,0.0193018243,-0.1460973769,-0.3054213822,0.1012097225,-0.0211047698,0.0389342718,0.0735517666,-0.1596219242,0.1458532959,-0.0804152638,0.1671361476,0.1694602966,-0.1528197229,-0.1993517727,-0.0595716573,0.0871389136,-0.0703356192,-0.3041724563,-0.1647040099,-0.1262097806,-0.1425148845,-0.0565400608,0.0619089194,0.032271076,0.0487479307,0.0251143593,-0.08564464,0.0832659602,-0.0585484542,-0.0687211081,0.1066514179,0.4175856411,0.2003389895,0.2956329882,0.0155429477,-0.0098339058,-0.168074429,0.1084363461,0.1326770782,-0.0601212978,0.4899217486,-0.5844715238,0.309333086,0.1525930017,0.0641689152,0.3737840056,-0.3491540551,0.1438475102,-0.0257819984,0.2500727475,-0.0072489972,-0.0197770391,-0.0187709853,-0.0589524396,-0.2512377501,0.1292736828,0.0368148685,0.0643110946,0.0658229515,0.1855467707,0.5413076282,0.0205544345,0.1777003855,0.4449892044,-0.2651417553,0.0103178015,0.1070830524,0.0288880877,0.1508512497,0.5691823363,-0.2521656752,0.3656856716,0.0674657896,0.1891845167,-0.0689479783,-0.2857246697,-0.1911573261,0.1283740103,-0.1039629877,0.205653131,0.1605815291,-0.0411468297,-0.2754294872,-0.1645678431,-0.4454323351,-0.0433725901,-0.1581476778,-0.1651981324,-0.3046612442,-0.237954855,-0.2086984813,0.1694152802,-0.0964413956,0.0689792931,-0.1310560405,0.3181930482,-0.1337466091,-0.4816061258,-0.4851031005,0.1531974524,-0.0802619457,-0.227401942,0.4024288356,0.3622741997,-0.1449483484,0.1752591729,0.2373050898,0.45814991,0.4365479052,-0.3334041238,-0.005134589,-0.1475551724,-0.1388259679,0.0681254566,0.1469890326,0.1474833041,0.4646258652,0.1928111017,0.2086624801,-0.139930442,-0.0940288976,0.2734103799,0.2114677429,-0.2410974503,0.1022985876,-0.0833409727,0.269402802,0.011460864,-0.1178097203,-0.537373662,0.0016914231,0.1048632711,0.3085604012,-0.038359765,-0.0835507736,0.094158724,-0.0242478028,0.3664717972,-0.02242033,0.0643890947,-0.3836486042,-0.1501823813,-0.4316813648,0.1047254056,-0.1395379007,-0.0761609003,0.0795028061,0.0511154868,0.1058837101,0.355951339,0.1389287412,-0.0680269673,0.1070758998,-0.0319246426,-0.1835423112,-0.1068297997,0.2802385688,0.0560278669,-0.1562605798,-0.1600483358,0.3295812011,0.0734450296,0.0841552466,-0.352758497,-0.2275990099,-0.1116228774,-0.0061711241,0.5378038883,0.1319498867,0.4072888792,-0.0044389046,-0.2568791211,-0.1435836256,-0.1493096054,-0.0891230777,0.2212728858,0.198007375,0.4451334774,-0.2095823437,0.1915100217,-0.5001687407,0.0465377532,-0.1751574576,-0.1508637965,-0.5225541592,-0.0235568471,0.0313592739,0.1359848678,-0.1398651451,0.2121208608,0.1611679345,0.2831196785,-0.3213100433,0.0549394712,0.4070506096,-0.4105957448,-0.4109965563,-0.0811065212,-0.144598186,-0.3138813078,0.0372826159,-0.6443098783,0.2687261999,0.2179948241,-0.1473029703,-0.3369533122,0.2057896256,0.2618826628,0.0653921664,-0.0011065411,0.3464151621,0.1690855026,-0.1502877027,0.0528870896,-0.2060890943]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/751","title":"Error loading ms_marco v2.1 using load_dataset()","comments":"I was able to load the dataset successfully, I'm pretty sure it's just a cache issue that you have.\r\nLet me know if clearing your cache fixes the problem","body":"Code:\r\n`dataset = load_dataset('ms_marco', 'v2.1')`\r\n\r\nError:\r\n```\r\n`---------------------------------------------------------------------------\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-16-34378c057212> in <module>()\r\n 9 \r\n 10 # Downloading and loading a dataset\r\n---> 11 dataset = load_dataset('ms_marco', 'v2.1')\r\n\r\n10 frames\r\n\/usr\/lib\/python3.6\/json\/decoder.py in raw_decode(self, s, idx)\r\n 353 \"\"\"\r\n 354 try:\r\n--> 355 obj, end = self.scan_once(s, idx)\r\n 356 except StopIteration as err:\r\n 357 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 388988661 (char 388988660)\r\n`\r\n```","comment_length":29,"text":"Error loading ms_marco v2.1 using load_dataset()\nCode:\r\n`dataset = load_dataset('ms_marco', 'v2.1')`\r\n\r\nError:\r\n```\r\n`---------------------------------------------------------------------------\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-16-34378c057212> in <module>()\r\n 9 \r\n 10 # Downloading and loading a dataset\r\n---> 11 dataset = load_dataset('ms_marco', 'v2.1')\r\n\r\n10 frames\r\n\/usr\/lib\/python3.6\/json\/decoder.py in raw_decode(self, s, idx)\r\n 353 \"\"\"\r\n 354 try:\r\n--> 355 obj, end = self.scan_once(s, idx)\r\n 356 except StopIteration as err:\r\n 357 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 388988661 (char 388988660)\r\n`\r\n```\nI was able to load the dataset successfully, I'm pretty sure it's just a cache issue that you have.\r\nLet me know if clearing your cache fixes the problem","embeddings":[-0.0129359569,0.1030107811,-0.1228333712,0.3122574687,0.3983422816,0.0521115921,0.2968255281,0.5685597658,0.0205572154,0.3210778832,-0.0469787084,0.5231743455,0.0632516816,0.1900868565,-0.1655188352,-0.2574036717,-0.0236327332,0.3507929146,0.0692266896,0.0745202303,-0.0900942162,0.1957165897,-0.1244621351,0.2141236961,0.0591254495,0.0740995929,0.0389386192,0.1152041405,-0.3439752758,-0.6365219951,0.1919737756,-0.0079323296,0.3155696392,0.4043069184,-0.0001020064,0.1777641773,0.4454354048,-0.0689876229,-0.3125953078,-0.582008779,-0.5053097606,-0.117103003,0.3165388107,-0.1282913983,-0.1209806576,0.1134021059,-0.1850688607,-0.4382014573,0.6948314309,0.3881035447,0.3028866053,0.1351647079,0.2301937789,-0.1009389684,0.2550452352,0.1251532286,0.0138985058,0.2112888247,-0.1272655278,0.2033331394,0.2619906366,0.2876456976,-0.0855982825,0.3251772821,0.1399059892,0.1585076749,0.3314511776,-0.3463332355,0.2681953311,0.1036736369,0.6430297494,-0.1307166368,-0.2762633264,0.0868243799,0.079673253,-0.5477735996,0.3189710081,0.0929363742,-0.2067407519,0.1782433689,-0.0895706862,0.0375676341,-0.1832725704,0.0019572945,-0.2085376978,0.0816446841,-0.1336175948,0.0782065094,0.3002834916,-0.2424610406,0.0746838823,0.0812309012,-0.2699528635,0.1097850651,-0.1803840995,-0.0948034674,-0.0430226736,-0.4446944594,-0.1199472025,0.1145335287,0.2236438096,0.0987935066,0.3771330118,-0.0008453242,0.463940829,0.0501384921,-0.1559505463,0.0771147609,0.1845827252,0.0444036573,-0.0029853527,-0.10394191,-0.2111139148,0.0192598104,0.1602069736,0.1095812023,0.3138606846,-0.2204987258,-0.2839262187,0.0783530921,-0.4800007939,-0.1939749122,-0.1975454837,0.3207300603,-0.0632954314,0.2304754704,0.0567353852,0.2245039493,-0.0679012388,-0.2256979644,-0.1210092083,-0.131233871,-0.0852742717,-0.1201667413,0.2747131586,-0.0853872821,0.1471427232,0.0709505007,0.1519605517,-0.2827930152,0.3410668373,-0.2363017797,0.1135585755,0.3298206329,0.2770091295,0.0701734424,0.181936115,-0.1325307488,-0.096114397,0.2491235584,-0.2524724007,-0.2802442014,-0.1876838207,0.2855273485,0.0113282856,-0.1512570083,-0.2134167403,-0.0533782616,0.0100037782,-0.0807726085,-0.3593396842,-0.1714255065,0.0606306829,-0.2019232213,0.3621022999,0.2705246508,-0.2723309696,0.0733098388,-0.1354018748,-0.0922605991,0.1877469867,0.2017704844,-0.3014396727,0.4362956285,-0.0464412123,-0.1566222161,0.1402229518,-0.3202365041,-0.324088335,0.3028960526,-0.0329794586,0.0878275633,0.025868902,-0.1924054474,0.3647069633,0.1537065655,0.1992154568,0.0544421673,-0.0356813632,0.123538658,-0.1245329753,-0.2586768568,0.0181741659,0.2463353723,0.0644575283,0.0442390405,-0.1041809767,0.2037008852,0.1403645724,-0.1245170459,0.1137161776,0.2085631043,-0.1304852813,0.1713306755,0.0299089886,0.1364278197,-0.1399660259,0.1214684993,0.025929274,0.1900323033,-0.2726621628,0.0064861607,-0.3014554679,0.0316114724,-0.368709594,0.1013638079,0.3043015003,0.1964470446,0.1233296543,0.1316632479,-0.2749223411,-0.0110125663,0.4417962432,0.1070485935,-0.1180402935,0.261908859,-0.2515699863,-0.2915503979,0.1108597741,-0.0367101543,0.0580888987,-0.0801604316,-0.3036938608,0.1816572249,0.2036078274,0.2495737076,-0.1795324236,-0.1363855153,0.0791234225,-0.2600405812,0.1992523521,0.5228087306,0.120222576,-0.024972599,-0.0676704645,0.2387714833,-0.0229306649,0.1125961915,-0.0246520545,-0.0547818206,0.3964394331,0.0283731315,-0.1547789574,-0.2609422505,0.2471832335,0.0534547232,0.1135942638,-0.1151223704,-0.3128592968,-0.1558890343,0.3872981369,-0.036732804,0.0159019101,0.1358817816,-0.1428260207,0.0166943911,-0.2165794075,-0.0299253613,0.1951276213,0.088865526,-0.1606716514,0.0816550702,0.1995746195,-0.0875950977,0.2196178734,0.2029738724,0.2804163694,0.1726022512,0.035534434,-0.0301926322,-0.2262343466,-0.0232357215,-0.1176520064,0.3725837469,-0.2375440896,0.084952496,-0.3510678709,-0.0531920083,-0.1461604536,-0.5117496848,-0.0538696796,-0.1060151458,-0.1435576379,0.1756749004,0.0639039576,0.1328894496,-0.0974535942,0.0262348205,0.1760558635,-0.0128644807,-0.1240892857,-0.3575330675,-0.2978409827,0.1148166433,0.2390877455,-0.3049159646,0.0696581602,-0.4312526882,0.0471266545,0.185514003,-0.2506832182,0.257073313,-0.1656760126,0.0676144883,0.0843776315,0.2562079728,-0.0387876891,-0.1973778456,0.3516507447,-0.1273055673,-0.1333613694,0.2202691287,-0.0234848429,0.0086861402,-0.0370711125,-0.6399906278,-0.223705247,-0.2604730129,0.1010658741,0.1509131491,0.0356118344,0.1985674798,0.1864263117,0.1326588094,0.0938302279,0.2327876091,-0.2737099826,-0.4300875366,0.2375005633,-0.2297223955,-0.3239218891,0.3098138273,0.0468384735,0.3736924231,-0.3118451238,-0.4044915438,0.2070856541,0.0206048284,-0.0509722754,-0.3278871477,-0.2940309048,0.2029005736,0.1398703456,-0.0699764043,-0.0671648607,-0.2701518834,0.1653960049,-0.0801305249,0.1338672191,-0.1019447222,0.850292027,-0.1222684085,0.408911556,0.1821502596,0.1215497926,0.2841959894,-0.0286712442,0.0892631635,-0.1776071936,-0.1607208401,-0.0405220464,-0.0886679664,-0.1416831613,0.1179836392,-0.0766924322,-0.3244694471,-0.3877325058,0.1473693997,-0.1907785535,-0.2491629124,0.1173725948,-0.0695967749,0.2437423617,0.1065364107,0.3213146925,-0.1834319234,-0.078320533,-0.0409559831,0.0246639568,0.1784259081,0.0543770306,-0.5196168423,-0.218445465,-0.3971497416,0.0976874456,0.0944587439,0.3790026903,-0.1187818497,-0.133283034,0.2848420441,-0.1996551603,0.4548556805,0.015690906,0.0225901473,0.2308395356,-0.0406474322,-0.2477574646,-0.1033172607,-0.3119055927,0.261957556,0.2248476893,0.2912777662,-0.2751751542,-0.1617020369,0.3189952374,0.0325333886,-0.1961504668,-0.142571792,-0.5580509305,-0.2862574458,-0.5117334723,-0.1522890776,-0.1444848329,0.3319621384,0.1333819032,-0.0510137454,0.1158998758,-0.0506252646,-0.0743982717,0.4398439527,0.3376657665,0.0571114942,0.2137265503,0.3307417035,-0.0897982195,0.3472658992,0.55099231,0.0874864683,-0.564702332,0.025258109,-0.230829075,0.0376107618,0.0295324326,-0.0426539928,-0.0523692518,-0.0759932622,0.0193470865,-0.3702815473,0.044480361,0.4498756528,-0.2386075705,-0.3628216684,-0.4035473168,0.175988242,-0.148846224,0.0461867005,-0.026190266,-0.0236491002,-0.1656990945,0.3092259765,0.2021736801,0.7795621753,0.0041691386,0.0423513502,0.3334950209,-0.4208038151,0.0586380325,0.0567669198,0.2716711164,-0.3065963387,0.1043093652,-0.0283184089,-0.0435964689,0.1442019045,0.0286618918,-0.365976274,-0.0277143624,0.0079402514,0.205215767,0.0679553971,0.2323817313,0.0098969853,-0.1805931032,-0.3090296984,0.2660910487,-0.3243823946,0.1789020747,-0.0787966102,-0.2388206273,0.0093731536,-0.2156899273,-0.1426713616,0.1112605184,-0.0953651369,0.2175239176,-0.0990179554,-0.3454363048,-0.1483260244,-0.1479493678,0.065723218,0.012239296,-0.1350323558,0.1931801289,-0.0936063901,-0.0968425125,-0.0193292145,0.2205018103,0.0276891477,-0.1507608742,-0.3855765462,0.29763484,-0.0065851724,-0.3523638844,-0.0291637089,-0.0187392421,0.1215568036,-0.0503315143,-0.006662529,-0.1516212821,0.0931434482,-0.0834323317,0.2480464727,-0.0020353019,-0.0652900711,0.0928145498,-0.1238288432,-0.3077957928,-0.0492697135,0.4987157881,0.1204651371,0.1030878574,0.6825522184,0.1222132221,-0.0366998948,-0.2413410395,0.3117476404,-0.2639380693,-0.0966204107,0.1311968118,-0.135455966,0.1641073823,0.01884179,0.2660928071,0.1431497335,-0.183100909,0.1255340725,-0.3727391958,-0.3420338333,0.0853065178,0.0867971629,0.0365233645,0.3154531419,0.2218178064,0.0249408837,-0.1445487142,-0.3769084811,0.0968957543,-0.054606095,0.0498188548,0.0529910736,-0.2355288416,0.1326881498,-0.0485053062,0.2245550603,0.1256546229,-0.1869513243,-0.2962935269,-0.0257777479,0.05056566,-0.137912944,-0.2670506239,-0.1554026902,-0.1380641907,-0.1070197895,-0.094381094,0.0379409716,0.0420823731,0.0546769239,-0.0086765187,-0.1130879223,0.0625442564,0.0154548436,-0.0259000883,0.0415078625,0.3601216972,0.1232106462,0.2862483263,0.0048703416,-0.0090656318,-0.1393584609,0.0529398397,0.1061153784,-0.0625224784,0.4956172407,-0.5514340997,0.2296753377,0.2057649642,0.0518743768,0.3414046168,-0.2846624255,0.0875460729,-0.0317554697,0.3160981834,-0.069755204,-0.0543361306,-0.058146175,-0.1145483851,-0.1619733125,0.0740739852,0.0659963489,0.0969431475,0.0974294618,0.2289814353,0.5092587471,-0.0384553596,0.1326857358,0.3509934545,-0.2204767466,-0.025668446,0.155064106,0.0337026194,0.1524450034,0.5504530072,-0.1738504022,0.3638179004,0.0467276685,0.2301944643,-0.0908970907,-0.2728544772,-0.179276824,0.0543198846,-0.1145457402,0.1795697808,0.1726710051,-0.0131100696,-0.2915280163,-0.1167616397,-0.5116007924,-0.0274865627,-0.189869076,-0.1445574909,-0.2732620835,-0.1898443401,-0.1252173781,0.1707194,-0.0580857508,0.0919250697,-0.0969295055,0.3143523932,-0.1261971742,-0.4820244908,-0.3915014863,0.0809077173,-0.0292962305,-0.2216845751,0.4314951301,0.330870688,-0.1228761002,0.1327256858,0.2395052314,0.4572738707,0.3687526882,-0.2967999876,-0.0131786577,-0.1482017934,-0.2029774785,-0.0427321643,0.1434620023,0.2300140411,0.4531635642,0.1332665235,0.2887100875,-0.2089294791,-0.0166743565,0.1810160875,0.2182873487,-0.2215468735,0.1298509836,-0.0419780649,0.1686277837,-0.0074696573,-0.0885153264,-0.5434630513,0.0401474908,0.1592126191,0.2889085114,-0.0623682849,-0.1588570923,0.1310419589,-0.0189989209,0.3759171963,-0.0097481441,0.0368048437,-0.3365987539,-0.1299781501,-0.4362042546,0.1682818532,-0.1909894645,-0.0220063683,0.0380228721,0.119991295,0.1020132527,0.3278622627,0.1788889319,-0.1205592826,0.0284022186,-0.0729248673,-0.2259826809,-0.0825774819,0.2572518587,-0.0021914726,-0.1161152422,-0.2137537897,0.311785996,0.0019224281,0.1452404261,-0.3561689556,-0.2029805481,-0.0674703345,0.0280018337,0.5374520421,0.1623466164,0.3618908823,-0.0228523538,-0.2558176816,-0.1589379013,-0.2314827144,-0.1078368202,0.1982891709,0.1674370021,0.4172362685,-0.1999168545,0.1303411424,-0.4659354985,0.0804221407,-0.1563739032,-0.1586231291,-0.4202130139,-0.0610855594,0.0501153693,0.1049485281,-0.120700106,0.2867714167,0.1374479979,0.2191431522,-0.3230113983,-0.0361346006,0.3846173286,-0.3938947916,-0.3895446956,-0.0443839915,-0.0917497575,-0.2922390401,0.0420934521,-0.5770190954,0.3236659169,0.2290095836,-0.1083639413,-0.3024738729,0.2415040284,0.224123612,0.1163617894,-0.0124492366,0.3833830953,0.1507356465,-0.1480346322,0.0213339143,-0.159285143]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/751","title":"Error loading ms_marco v2.1 using load_dataset()","comments":"Yes, it indeed was a cache issue!\r\nThanks for reaching out!!","body":"Code:\r\n`dataset = load_dataset('ms_marco', 'v2.1')`\r\n\r\nError:\r\n```\r\n`---------------------------------------------------------------------------\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-16-34378c057212> in <module>()\r\n 9 \r\n 10 # Downloading and loading a dataset\r\n---> 11 dataset = load_dataset('ms_marco', 'v2.1')\r\n\r\n10 frames\r\n\/usr\/lib\/python3.6\/json\/decoder.py in raw_decode(self, s, idx)\r\n 353 \"\"\"\r\n 354 try:\r\n--> 355 obj, end = self.scan_once(s, idx)\r\n 356 except StopIteration as err:\r\n 357 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 388988661 (char 388988660)\r\n`\r\n```","comment_length":11,"text":"Error loading ms_marco v2.1 using load_dataset()\nCode:\r\n`dataset = load_dataset('ms_marco', 'v2.1')`\r\n\r\nError:\r\n```\r\n`---------------------------------------------------------------------------\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-16-34378c057212> in <module>()\r\n 9 \r\n 10 # Downloading and loading a dataset\r\n---> 11 dataset = load_dataset('ms_marco', 'v2.1')\r\n\r\n10 frames\r\n\/usr\/lib\/python3.6\/json\/decoder.py in raw_decode(self, s, idx)\r\n 353 \"\"\"\r\n 354 try:\r\n--> 355 obj, end = self.scan_once(s, idx)\r\n 356 except StopIteration as err:\r\n 357 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 388988661 (char 388988660)\r\n`\r\n```\nYes, it indeed was a cache issue!\r\nThanks for reaching out!!","embeddings":[0.0922261998,0.1404983997,-0.1030100286,0.3711931705,0.3660838306,0.0681342706,0.3117438853,0.6182887554,-0.005093202,0.2712129951,0.0038577318,0.4279622734,0.0134958746,0.0225968547,-0.0885598511,-0.2377021611,0.0372549109,0.2955968976,0.1482123882,0.0860564858,-0.0212920327,0.240081653,-0.1054312959,0.1442669183,-0.005001015,0.1451946944,-0.0091406694,0.0862226486,-0.2775143683,-0.6665686965,0.2482306808,-0.0437338017,0.266197741,0.4317498803,-0.0001092058,0.1491607875,0.4879757762,-0.0973927975,-0.3604195118,-0.4838114083,-0.5681415796,-0.0567342006,0.353218168,-0.1525125802,-0.046058882,0.249455899,-0.1985055357,-0.5668948889,0.7049098015,0.3587841094,0.2581380308,0.1135658696,0.1226179004,-0.0644857436,0.3891299963,0.0687951446,-0.0100787655,0.1548397988,-0.1450454146,0.1880626231,0.2793936133,0.43785882,-0.1095831543,0.3514956236,0.1488466114,0.1496458203,0.2632724345,-0.2365249246,0.2547244728,0.0620301478,0.7168841362,-0.1921516806,-0.3531574607,0.0220708828,-0.021079693,-0.6890460253,0.4284820855,0.0506971031,-0.1903843433,0.2155735344,-0.1666305065,0.0634919629,-0.1299615651,-0.0529771745,-0.1716541648,0.0229255483,-0.1064105257,0.0386242345,0.1862325966,-0.2250207663,0.1551077962,0.1308619827,-0.3201520443,0.1800563931,-0.187403664,-0.1108969525,-0.0265473034,-0.3927610219,-0.1147668883,0.1494496316,0.3283623457,0.0706718862,0.3636214435,-0.0788253322,0.4711206257,0.2670929134,-0.1930588782,-0.0677645579,0.1605575085,-0.0846868306,0.0597604066,-0.1239688247,-0.1636045873,0.0575158149,0.2948910594,0.1318600625,0.2824220657,-0.2775970399,-0.1897220612,0.1391332895,-0.4733852446,-0.2513734996,-0.2246418297,0.3202269375,-0.0938966274,0.2881301641,-0.0121462997,0.1712311953,-0.0681220219,-0.2659787536,-0.1516557932,-0.1500047296,-0.0745827258,-0.0456694961,0.3001405895,-0.0571401492,0.1055291817,-0.008961048,0.1569736153,-0.3325084746,0.3220511973,-0.300011307,0.2625266016,0.393765986,0.2043040842,0.1398236454,0.2237453461,-0.1096066087,-0.1394321471,0.2308930755,-0.2680746913,-0.3547726572,-0.1435337514,0.2186732888,-0.07531894,-0.1104376242,-0.1414732933,-0.0076211817,0.1929127872,-0.1232361272,-0.3134557307,-0.1129929274,-0.0216345508,-0.1990177035,0.3235304952,0.340814501,-0.2782672346,0.0473009199,-0.1354072988,-0.0050474494,0.1961886585,0.1943644285,-0.2169827223,0.4180597663,-0.0698104128,-0.2173119187,0.1187304109,-0.39428249,-0.4204770923,0.2443167716,0.0000649911,0.1725618243,0.129857406,-0.2158607841,0.3943699598,0.0809171796,0.2884091139,-0.0108072925,-0.0429666713,0.1894393861,-0.1191842705,-0.3005268276,0.0487483516,0.2368618846,0.1102496758,0.0818012059,-0.160544917,0.2635349333,0.0409293398,-0.044967737,0.1864383519,0.1943557858,-0.218996793,0.088749975,0.1048822328,0.1359405518,-0.1246708333,0.1461257637,0.0431543179,0.1191758811,-0.2609390914,-0.0338719562,-0.2359333336,0.0118640363,-0.3827471435,0.0352362506,0.2485632151,0.1627869159,0.2266714722,0.1403045207,-0.2825946808,0.1378679872,0.6244596839,0.0761905685,-0.1072211936,0.1550353318,-0.225903675,-0.3396036029,-0.033932969,-0.010075626,0.1136221886,-0.0250232387,-0.3470095098,0.1828759015,0.2052425146,0.2740232944,-0.2516199946,-0.1218564734,0.0876979828,-0.217611596,0.2567509115,0.5755069852,0.0731469691,-0.0924073011,-0.0734419674,0.1745611727,-0.007058281,0.0990115628,-0.0765491948,-0.1114835665,0.3235239983,0.0661790818,-0.0905491039,-0.3190706074,0.3075248003,0.0368275419,0.2031317651,-0.1671235412,-0.2782463431,-0.1103214473,0.3933092952,0.0117621161,0.0146434596,0.1398809552,-0.1584527791,-0.1118672043,-0.0973372459,-0.060440924,0.2371606529,0.0537914671,-0.1431425512,0.0943356603,0.2405339628,-0.0699254796,0.2096734941,0.1660362631,0.2259545624,0.1727723181,0.1283366531,0.0403316878,-0.2227562815,-0.0020625319,-0.1176391169,0.326607734,-0.2499751598,0.1368027627,-0.3926931024,-0.0491468385,-0.0632992312,-0.5974288583,-0.0916694552,-0.1806727052,-0.0839082748,0.2263780981,0.09662126,0.0744971558,-0.1054051965,0.1311153471,0.1691013128,-0.0401838571,-0.2067136765,-0.4465826154,-0.2395186126,0.0347604342,0.2109881639,-0.3973528445,0.0731356218,-0.5161833167,0.0260230862,0.2396287322,-0.2327644378,0.2877539396,-0.0986082032,-0.0078398883,0.0390363447,0.2067961246,-0.0482727177,-0.1511716396,0.2659276426,-0.0928811654,-0.1638563126,0.1813531071,-0.0219168011,0.1314731985,-0.0707536861,-0.5130618811,-0.2585636079,-0.1761192679,0.0810766891,0.1249150783,0.069428049,0.2712322772,0.1708330661,0.1232851744,0.1088149995,0.3150093853,-0.3630496562,-0.525581181,0.2225391865,-0.2196513265,-0.2559969425,0.2633799911,0.0752168,0.3404311836,-0.2539471686,-0.4406087995,0.1957118064,0.0412995145,-0.1075026616,-0.3485988975,-0.3980806172,0.4110273719,0.1676250696,-0.0281013679,-0.074427709,-0.2933170497,0.1424517781,-0.1122413501,0.1653866023,-0.0890772417,0.7981615663,-0.0466545783,0.5489046574,0.1633441448,0.2031069696,0.3505617976,0.0410172977,0.173749432,-0.2444557697,-0.0596782416,-0.0605661795,-0.1694197059,-0.2410618812,-0.0057518198,0.0406070575,-0.3838979602,-0.4311557114,0.2103930116,-0.1773812473,-0.2024858296,0.0691665038,-0.1699276567,0.1924155653,0.1797181964,0.3524456322,-0.3076170087,-0.0852140784,-0.0645869896,0.0399615243,0.2050726861,0.0701097474,-0.4801722467,-0.1741949171,-0.4788979292,0.0910582766,0.1686000377,0.3973281384,-0.1676757634,-0.186828196,0.2215010822,-0.2812476158,0.5397033691,-0.0543841124,0.100290142,0.2773097456,-0.0662925765,-0.1961205304,-0.10151086,-0.1574149281,0.269649744,0.2142236978,0.2479230165,-0.2078156322,-0.1436362863,0.28268677,0.0331509523,-0.1663973927,-0.1609747708,-0.4863089025,-0.1944684684,-0.4285179079,-0.1266063899,-0.1567766517,0.2426392883,0.1495864093,0.0103581995,0.0760535076,0.0850170031,-0.175706327,0.4018840492,0.3567194939,-0.0076818592,0.2022645175,0.4010499716,-0.0954559669,0.2502842844,0.5017067194,0.055976972,-0.4775360525,0.0833321363,-0.1465931535,-0.1254974604,0.1542491913,-0.0592983775,-0.1031178311,-0.1361022741,-0.0320603624,-0.3839403689,0.0091384137,0.3691645861,-0.2533956766,-0.3804501295,-0.4277611971,0.1571734399,-0.0802072659,0.0253686942,0.0096834302,-0.0842315182,-0.1895846277,0.3751416802,0.1809796691,0.9085156918,0.0180659648,0.1930073947,0.2775514126,-0.3772618175,0.0347600989,-0.0005454399,0.2941225171,-0.2400703728,0.0985015929,-0.0641937479,-0.079008244,0.0703691095,-0.0732016265,-0.3800194561,0.0524525344,-0.0195997283,0.3411562145,0.1107098758,0.113410376,-0.0437632129,-0.1667152494,-0.3010195792,0.2261270732,-0.293423146,0.1827531755,-0.1057145447,-0.2049712539,-0.0223844238,-0.2369997203,-0.2042770535,0.1603508592,-0.1083218157,0.2146257907,-0.1410563588,-0.4502209127,-0.3198000193,-0.1316740811,0.1280993074,-0.0865076184,-0.1674452126,0.2223321646,-0.0218252894,0.0025337453,-0.0333136357,0.180787757,0.0883386135,-0.1321383566,-0.2826611698,0.2551697195,-0.0912790522,-0.4154534638,0.0055319876,-0.0040789321,0.1123237386,0.0371109731,-0.0658264458,-0.1286054701,0.0806469396,-0.072613351,0.202172935,-0.0143237505,-0.0467925966,0.0940959975,-0.2040370107,-0.2591896951,-0.0525132902,0.6456134319,0.1243520603,0.0730129853,0.6635721326,0.0678832084,-0.070420213,-0.1653012335,0.2126576155,-0.2950674593,-0.1620849371,0.1453900039,-0.3414604068,0.1347388327,-0.0653721616,0.1814276874,0.1560182571,-0.0710345283,0.1464050114,-0.3323258758,-0.38290748,0.1038775519,-0.0066666636,0.0278706998,0.3528233767,0.1230962127,-0.034189336,-0.1593034863,-0.3074298203,0.1269252598,0.0018293491,0.014512511,0.0086287251,-0.247473225,0.2330595553,-0.0903321877,0.1361277103,0.1920154542,-0.2695337236,-0.2490183264,-0.0397111848,0.1018313244,-0.0809290558,-0.3352916241,-0.1351198107,-0.1073935553,-0.0131255724,-0.1068015397,0.053372331,0.1123632044,0.0825725868,-0.080611527,-0.0266562756,0.123228319,-0.045336578,0.0318044908,0.0856542811,0.4084722102,0.1012702435,0.319313705,-0.1268482059,-0.0018729954,-0.1540661305,0.0862619653,0.0769049525,-0.1578167379,0.5385918617,-0.5605436563,0.2418406308,0.2267641872,0.033647608,0.2734521925,-0.3241758347,0.0432275087,-0.0507400148,0.2437397391,0.0484457426,-0.0994377956,-0.0629498661,-0.0082139121,-0.2478473186,0.0845819563,0.1309743226,0.1246442348,0.2092943341,0.3035587966,0.7101327777,-0.1016732305,0.1105427593,0.2532702684,-0.184758395,0.0712591112,0.1596059501,0.0129830223,0.1596039385,0.5148218274,-0.1826081574,0.3546608388,0.1568040848,0.2028791457,-0.2365229279,-0.2974717319,-0.1872513592,0.1655883342,-0.0644547194,0.2564346194,0.1474445462,-0.105971247,-0.2371380329,-0.1161539555,-0.4655508101,-0.0138903772,-0.0928811133,-0.1782216132,-0.1196657866,-0.1731272191,-0.1022536978,0.1926274598,-0.1243963018,0.1496271938,-0.0116211884,0.3059940636,-0.1290429682,-0.4961288571,-0.3904127777,0.0743030608,0.0377120562,-0.2835611403,0.4901018441,0.3155002594,-0.157800585,0.0821411461,0.3121913373,0.5379475355,0.3586464226,-0.2830999792,-0.0239132624,-0.0909249783,-0.1720856428,-0.0585204773,0.1884824932,0.2515133619,0.3798468411,0.0632460788,0.233029902,-0.1391274929,0.0902912915,0.1856700927,0.2774809301,-0.2383589596,0.0739056468,-0.0867700428,0.1960682124,0.015399673,0.0313271061,-0.5615615845,0.1413570046,0.1531335413,0.3081690371,-0.0489655472,-0.0897348821,0.1000823528,0.0347835943,0.3627692759,0.0521545522,0.0377190784,-0.3249898553,-0.1218122691,-0.4196885228,0.2370986789,-0.1932418644,0.1044274196,0.0636865869,0.1273896992,0.020149339,0.3312429786,0.1877754778,-0.1198931038,0.0421651267,-0.1101079509,-0.2886916995,-0.0344747938,0.3270440698,-0.0148830917,-0.1215215176,-0.3076904714,0.3285369277,0.0932516381,0.1029370949,-0.382219553,-0.2826789618,-0.0679738522,0.1207631975,0.5093848705,0.1550472826,0.32603392,0.0435602404,-0.2738367319,-0.2430631816,-0.2345045209,-0.0666147396,0.0479207374,0.1029402092,0.3664226234,-0.3142722249,0.1125809252,-0.4420832992,0.2091475278,-0.2107106447,-0.1394338608,-0.4605120122,-0.0790195018,0.0433085561,0.1629967839,-0.080113776,0.2915146351,0.0630863681,0.3607110679,-0.38161394,-0.0008938375,0.3831396997,-0.4747906923,-0.3554658294,-0.039909482,-0.0804476291,-0.2855094373,-0.1191882044,-0.5399656296,0.3031439483,0.1982624084,-0.1223167777,-0.3863053024,0.2170281857,0.2792541385,0.1101970673,-0.0119604468,0.3740397692,0.1887703389,-0.1657519639,0.0004311023,-0.2231602073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"Amazing! ","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":1,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nAmazing! ","embeddings":[-0.2740081251,-0.1259905696,-0.2096738219,-0.1152828783,-0.085577935,-0.1539034396,0.2437988073,0.1135552227,0.0426797085,0.0694850013,-0.0425657816,0.032485012,-0.0064367075,0.2527284622,0.4808174372,-0.180927515,0.1096525341,0.0150428507,0.0459406711,-0.0692423135,-0.187655583,0.1984163076,0.1257516444,-0.0591848753,-0.1633331627,-0.3267315924,-0.2135618031,0.1300342381,-0.1628780067,-0.2406724542,-0.1953875124,0.1770967692,-0.2163759321,0.0903349891,-0.000095542,-0.1802453846,0.0767291412,-0.0259295069,0.156364128,0.1371474266,-0.255466789,-0.4294767082,-0.3013774157,-0.0871413127,-0.1461730301,0.0090143094,0.1226361096,-0.3801429272,0.020383589,-0.0230527781,0.3431610167,0.0187794976,0.0872897729,-0.15130952,0.3792746067,-0.1896388084,-0.0816313848,0.1452640742,0.730150938,0.0506490618,-0.1013333872,0.2667188644,-0.0063192951,-0.2246520072,-0.0970326215,-0.1064151004,-0.08913479,-0.1025960669,-0.0798228383,0.5630835891,0.1643103808,-0.2732553184,-0.2529917657,-0.0919973254,0.0310752746,0.0540094525,0.0291704219,0.2060076892,0.1084144637,-0.0369790606,0.2317048758,-0.5008109808,-0.1182519272,0.1380945742,0.0144306421,0.4552817941,-0.1266608834,-0.0103100557,0.0927088261,-0.0269558243,-0.1027282402,0.0731194094,-0.1232708916,0.0171594471,-0.1164916232,-0.0148156974,0.045888532,0.2191704363,0.1923842579,-0.1278252453,0.203797102,0.3216381371,-0.1149738654,0.2649789751,-0.1147010624,-0.09191861,0.0970849097,-0.3354712129,0.2913810909,-0.3138774931,0.063448295,0.1163212135,-0.0441066623,0.0267651286,-0.3767687082,-0.0464067943,-0.1368882805,-0.0077390033,-0.1720094234,0.0822664052,-0.0664546117,-0.0011254132,0.0399729945,0.1546137333,0.0336032063,-0.0230831709,-0.0958107635,0.2113515586,-0.0283250641,-0.3019951582,-0.108627528,0.4636098146,-0.2917343378,-0.0483258516,-0.052694086,0.038363222,-0.0514313057,0.3267011344,-0.0964877233,0.1418595016,0.3119728565,-0.0048563378,0.1545909494,-0.1286393106,0.0233768653,-0.078124322,-0.0522680245,-0.2360471487,-0.3313300312,-0.1563606262,-0.0009431837,0.0757037848,-0.5339612365,0.2975057662,-0.1099703759,-0.3572642207,0.331610322,0.6563366055,-0.1480449885,-0.1214380786,0.3455623388,0.1777919978,0.0048373407,-0.2739084661,-0.0403134823,0.1988286525,-0.6493272185,-0.2786084712,0.0554041862,-0.172479406,0.1005809605,0.0540490374,-0.0917001888,-0.1825314611,-0.1123199314,0.3574241698,0.2289472073,-0.2438004911,-0.161611408,0.0798761845,-0.2118618935,0.0384770297,0.2292647213,0.3234772384,0.0778752938,0.0334900916,-0.0441727415,0.5808441639,-0.1170698553,-0.0981553644,-0.1652316153,-0.2288519442,0.0343786813,0.2146012038,0.0455935635,-0.31438905,0.4132901132,-0.0090686493,0.0182234924,-0.223644048,0.0699590519,0.3148040473,0.2189383805,0.027360009,-0.163599506,-0.2597948313,-0.1803670228,-0.0380032919,-0.1145455763,0.2601157129,0.0754028931,-0.1735390872,0.0244643092,-0.1665843129,-0.134727478,-0.0887754858,0.3383388817,0.3188284039,0.0218499266,0.2631798983,-0.0888848528,0.3878929019,0.0809027553,-0.0605126247,-0.2505098283,0.0767152682,0.0102188094,-0.1386885047,-0.1115308478,0.3331746757,-0.2703098655,-0.2035097927,0.2567645013,0.365062207,-0.1833450198,0.3364019692,0.3505297601,0.3580788076,0.3171123862,-0.0504770838,0.362482667,-0.2111686021,0.0591547042,-0.0133995293,0.2184201926,0.4292857945,0.3146835566,-0.241075173,0.0976756513,0.1113431826,0.3226427436,-0.0924296156,-0.008012617,-0.2258381844,0.3219336271,0.0696492568,0.3773954213,0.0021968151,-0.4232115746,0.2499121428,0.6222330928,-0.1070857048,0.5763735771,-0.104069598,-0.240789324,-0.0825404227,0.0888807252,0.3571736515,-0.0675126985,0.3280631304,0.244677633,-0.1673430651,-0.1429964304,-0.2281855792,0.1743458509,-0.2535194457,-0.0953680798,0.1621779948,-0.0776845589,-0.2518578172,-0.714936614,-0.0898660272,-0.0041341111,-0.1146981493,0.1462798119,-0.1714839786,-0.1934831738,-0.3776310384,-0.2258145213,-0.3228199184,-0.3644911647,-0.2172412723,0.3006415367,-0.3342939913,-0.1229221672,0.187061891,0.0862646773,0.2967572808,-0.2713572681,-0.2081194967,-0.0943219289,-0.2525451183,-0.1215501428,0.3224652708,0.0602322258,0.0244648475,0.4462008178,0.0132798152,0.3975153565,-0.1881445795,-0.6204220653,0.1625687629,-0.1651383638,0.1159903407,0.1963235289,-0.1639989913,0.0745067671,-0.2138014138,0.1180421188,-0.2875752747,-0.0472740047,-0.0699187666,-0.210794881,-0.1325814575,-0.1493310928,-0.3276374042,-0.4085571468,-0.3770136833,0.2492589951,-0.1070765927,-0.0458701402,-0.0757050291,0.0352497473,0.3546482325,-0.1547538787,-0.0002752795,-0.2782554328,-0.3132137656,0.0699281916,-0.1905059814,-0.2392394841,-0.1472988725,-0.1387489587,0.2131655961,0.0296470299,-0.1900844574,-0.0641153678,-0.1778041869,0.2099309564,0.3099562526,0.2457467467,0.2115518749,0.0447180681,-0.2723623812,-0.2334306091,0.0261475574,0.1015957817,0.0797742233,-0.0036870013,-0.0134988995,0.1415644437,0.158748433,0.2324057966,-0.2745358944,0.0182910971,0.4850561619,0.081079185,0.422632426,-0.1021519825,-0.173604086,0.216070056,-0.0864928961,0.3775252104,0.1644100994,0.0628645048,-0.4215979874,-0.0522737391,0.0001925726,-0.1731462777,-0.0461320132,0.1778260171,-0.0901877806,0.0555359088,-0.0134691205,-0.4685731828,-0.0921198428,-0.0199963301,0.2332333177,-0.0348986946,-0.1317107081,-0.082853578,-0.4028142989,0.0798169225,-0.4257951975,0.3605086505,0.2831116915,0.0586501807,-0.350432694,-0.1132873446,-0.0264898241,-0.0478112362,0.2210818529,-0.6574328542,-0.3443217278,0.2359645218,-0.0483047664,-0.1397420913,0.2266708612,-0.5377310514,-0.082625255,0.3854729533,0.1418530196,-0.2202335,0.0216020104,0.1184179485,0.1489619166,0.0340641364,0.2528591752,0.0069987644,-0.2305505723,-0.2722635865,0.1300552487,-0.0698082745,0.0436404273,0.5713452101,-0.1412484795,-0.0916261822,-0.0719131008,0.1805793494,0.2043439448,-0.200764969,0.1915305257,-0.2172931135,-0.0401503108,0.0573535487,-0.1553671658,0.2480091155,-0.1575164795,0.0116306972,0.111874707,-0.3532651961,0.4243742824,0.0605330393,0.2963663042,0.3933086991,-0.1996760964,-0.1249825731,-0.2476720065,0.2191778123,0.0283820182,0.455771178,-0.3011696041,-0.2651994526,0.178234309,0.1518860608,-0.2155308425,0.4003295302,0.1531763673,-0.2826994956,0.0648949221,0.0946592689,0.9481253028,-0.1126192063,0.107780315,0.1033141911,-0.2483292818,0.1967993677,-0.373591423,0.0203580968,-0.2243371159,-0.2712901533,-0.1507394463,0.0772467777,0.2397308201,0.0705959499,-0.3249650598,-0.0798884183,0.1269480139,0.0125662051,-0.0471854061,0.2483869642,-0.1716228127,-0.2477518171,-0.1232510656,0.2998535037,0.0907097086,-0.1468375325,-0.1861998588,-0.1135761142,0.040957056,0.0537296124,-0.2909002304,0.0666615069,-0.1734419465,0.1988662332,-0.0448089018,-0.1493704915,0.0109906998,-0.1081607714,0.1532862037,0.0394170359,0.0183004756,0.2215730399,0.0564294383,-0.0435245223,0.231717214,-0.0877924189,0.4154912531,-0.1071188152,-0.1776314676,0.2820279896,0.0781980827,-0.264824003,-0.4172394574,0.1587783992,-0.0289308708,0.1190787256,-0.0302397329,0.3807233274,0.0576839894,-0.3492359519,0.2441909164,0.2583135962,-0.025098836,0.0626960024,-0.0504770391,-0.2062104046,-0.3528767824,0.0223980732,0.071230717,-0.0079647331,-0.1086183637,0.1534139961,-0.1555837095,-0.4069345295,0.174131155,0.5170544386,-0.2647245228,-0.0741832852,-0.0843076408,-0.5779993534,0.0241447687,0.0331948474,0.2779384851,0.0951461866,-0.0811665878,-0.2115022987,-0.0216264259,0.2034643888,0.1292107552,0.2548630834,-0.020177858,-0.1625723094,0.1064107046,0.2226611823,-0.4934061766,0.0124994125,-0.0677489191,0.0302970354,0.2122467607,-0.0968005955,0.1686117649,0.1711134613,0.2806229293,-0.0951856226,-0.1042414233,-0.3137906194,-0.2547805905,0.0337000526,-0.1900168955,0.1549427509,0.3085594773,-0.062331114,0.0875210837,-0.0756662861,0.3005617857,-0.0858374611,-0.112607345,0.0823357478,-0.0470874161,-0.0332129486,0.2576291561,0.0378558151,0.063905552,0.1002015173,0.1432376653,0.0218731668,0.1047030911,-0.1873530447,0.2834077477,0.3649134338,0.2924134135,-0.0324858837,-0.0490705557,0.1632808596,0.0707130209,0.1942114234,0.2369095981,-0.1011927351,-0.1162846386,0.0109600304,0.1202130988,0.3716497719,-0.2763198316,-0.3083119392,0.2703245878,-0.0963933319,0.4322249293,0.3338332474,0.3938131034,0.1376609802,0.289961189,-0.2563150525,0.1353016049,-0.0795604065,0.1719275564,0.1469134539,-0.1166106761,-0.1369358301,0.20497711,0.2133967876,0.1272309572,0.3471454084,0.0647107661,0.2154336274,0.4733229876,0.3091386259,0.1758236587,0.0805913433,0.24930951,0.2131290436,0.0876320675,0.3100575507,0.1658376008,0.3436900973,-0.1384807825,-0.0825333744,-0.0304508284,0.3033235073,-0.131828025,-0.1960529834,0.1562865674,-0.1928267926,-0.2984406054,-0.0162534341,0.103521511,-0.0311728399,0.2645922005,0.0259645283,0.3362686336,-0.1554452628,0.2306855619,0.0086102318,0.0243610442,-0.2491053939,0.3090200126,0.0112102926,-0.0687546879,-0.0799334347,0.3145861626,0.1846066117,-0.110539332,-0.1674206704,0.417663902,-0.2803074718,0.0274517946,-0.1538176537,0.1568334997,0.2929053009,0.3304571211,0.2943795323,0.2328952849,-0.2604297101,0.0188244488,0.1416285932,-0.1392474324,-0.011533075,-0.0909025744,-0.1464195997,0.0182702839,-0.4602284729,0.1762191951,-0.5362858176,-0.1234403029,0.3123670816,-0.2999267578,-0.0571811236,-0.0400994234,0.1510595381,0.0999739543,0.3928399384,0.1860428005,0.2275375724,0.1153294295,-0.4415796697,-0.4549120367,0.1648922414,0.0315582119,-0.2081720829,-0.3066650331,0.1670392305,0.4337587357,0.03623005,-0.0204863083,0.2946508527,-0.0263866596,-0.2263118774,-0.3502197862,0.2827513516,0.260978967,0.1626158208,0.0504076034,-0.0120312646,-0.130791381,-0.0318869576,0.2443352044,-0.2844920158,-0.2122510821,-0.1070654094,-0.1721068472,0.3592990935,-0.0011637401,0.2899380922,-0.3720490932,-0.1346681863,-0.1458679438,-0.2490303516,-0.1819832176,0.2771489918,0.1966401041,0.1921930462,-0.0104104746,-0.1158290282,0.0917952806,0.2683229148,0.0866814777,0.3507815301,-0.1442614347,0.2574337125,0.028982373,-0.1405309737,-0.00713311,0.4213782251,0.072576575,-0.0483817123,0.1396882385,-0.3324322402,-0.0917197391,0.0359552018,-0.1232469752,-0.0032113169,0.0258899778,0.1153074205,0.0484387241,-0.3237890303,0.0210070945,0.2122395635,0.214266032,-0.1582542062,0.1832866818,0.113574326,-0.2388171703,-0.0697368234,-0.0975616276,0.2726408243,-0.227192089,0.1129854023,-0.315903008]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"Small poll @thomwolf @yjernite @lhoestq @JetRunner @qiweizhen .\r\n\r\nAs stated in the XGLUE paper: https:\/\/arxiv.org\/pdf\/2004.01401.pdf , for each of the 11 down-stream tasks training data is only available in English, whereas development and test data is available in multiple different language *cf.* here: \r\n\r\n![Screenshot from 2020-11-04 15-02-17](https:\/\/user-images.githubusercontent.com\/23423619\/98120893-d7499a80-1eae-11eb-9d0b-57dfe5d4ee68.png)\r\n\r\nSo, I'd suggest to have exactly 11 \"language-independent\" configs: \"ner\", \"pos\", ... and give the sample in each dataset in the config a \"language\" label being one of \"ar\", \"bg\", .... => To me this makes more sense than making languaga specific config, *e.g.* \"ner-de\", ...especially because training data is only available in English. Do you guys agree? ","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":105,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nSmall poll @thomwolf @yjernite @lhoestq @JetRunner @qiweizhen .\r\n\r\nAs stated in the XGLUE paper: https:\/\/arxiv.org\/pdf\/2004.01401.pdf , for each of the 11 down-stream tasks training data is only available in English, whereas development and test data is available in multiple different language *cf.* here: \r\n\r\n![Screenshot from 2020-11-04 15-02-17](https:\/\/user-images.githubusercontent.com\/23423619\/98120893-d7499a80-1eae-11eb-9d0b-57dfe5d4ee68.png)\r\n\r\nSo, I'd suggest to have exactly 11 \"language-independent\" configs: \"ner\", \"pos\", ... and give the sample in each dataset in the config a \"language\" label being one of \"ar\", \"bg\", .... => To me this makes more sense than making languaga specific config, *e.g.* \"ner-de\", ...especially because training data is only available in English. Do you guys agree? ","embeddings":[-0.1428691596,0.0099278111,-0.0491098911,0.1698313951,-0.0771438181,-0.2269947827,0.3096124828,-0.0059922379,-0.110640347,0.026588764,-0.0933472589,-0.1092793867,0.0642141253,0.3641544282,0.2442416698,0.0054116757,-0.0701899603,-0.2135346085,0.1959691942,-0.1101954356,-0.1925240606,0.207330808,0.0093115317,0.0382775813,-0.1812143326,-0.3838912845,-0.1454201639,-0.0401203074,-0.2106133103,-0.1362018734,-0.0949112251,0.2998587787,-0.0450537838,-0.1357192844,-0.0001101672,-0.1455022544,0.0337634087,-0.2893490493,0.0777739212,0.0766476691,-0.5392189026,-0.4189271033,-0.2100260109,0.021882318,-0.1322230548,-0.0703945234,0.0398570597,-0.3876658082,0.0150852306,0.2009891719,0.124359332,-0.1307045519,-0.0916959047,0.0022764772,0.1724182069,-0.0599789917,-0.0712107569,0.2573898733,0.6603621244,0.1180264205,-0.189739868,0.4227789044,0.1062077731,-0.2274499983,-0.2512735128,-0.1123048291,0.0275756065,-0.5208718181,-0.0307946373,0.7121629119,0.094414711,-0.2192554176,-0.3764995933,-0.5085293651,0.2311372608,-0.1060143784,-0.0385447927,0.3316169381,0.1360908449,0.124642767,0.2813400626,-0.4833071828,-0.0645141527,0.1144296676,0.0272163115,0.3500621319,-0.0517745912,0.1554934829,0.0582310744,0.2205421031,0.0908468664,-0.1914832592,-0.202879414,-0.158775717,-0.273771733,0.0238598641,0.1089216247,0.1924334317,0.2820926905,0.0103116855,0.050496798,0.2658737898,0.0358598493,0.009202701,0.095834069,0.1696853787,0.3132936656,-0.0710081309,0.389337182,-0.2828632891,0.0175504722,0.0768345892,0.1171840504,-0.047292687,-0.5885556936,0.1953620017,0.0552476496,-0.1350183636,-0.1649702936,0.018880019,-0.2845041156,-0.2749780715,0.1250263453,0.0475592464,0.2360591143,0.2134106159,-0.2342136502,0.4277414083,-0.2416114956,-0.7489882708,-0.0678648427,0.1831589639,-0.2963540852,0.0628098398,-0.016167229,-0.0169058815,-0.0826979056,0.5410317183,-0.1015983969,-0.1230051219,0.3408746421,-0.2684747875,0.0758993179,-0.1212970167,-0.1414213628,0.0661510676,0.058392968,-0.2884202898,-0.3733742535,-0.047720205,-0.0855763108,-0.3151535094,0.0327207111,0.1235511377,-0.11830277,-0.1122066081,0.0821984485,0.9992453456,0.0727828741,0.1342613548,0.2505101264,-0.2038619369,0.0183526259,-0.3103568852,-0.1085827723,0.4024304152,-0.6962332129,-0.0611513592,0.2866933644,-0.1877843291,0.272908628,0.0335709155,-0.3144292533,0.0204332713,-0.1424882412,0.3152839541,0.3137671351,-0.1194128394,-0.0993398204,0.1415378153,0.0063684937,0.0358393751,0.0838430971,0.1323588639,0.233286038,0.120140709,-0.2324824035,0.5208756328,-0.1417699158,-0.164527595,-0.1718298942,-0.1081952006,0.1924680173,0.2677874863,-0.1291402876,-0.2747170329,0.1060365587,-0.2061803341,0.1472239792,-0.2873649299,0.1071479842,0.191968888,0.0442438945,-0.0178577863,-0.1327811778,0.3083823919,-0.5418758392,0.1860675365,-0.1032324582,0.5869051814,0.2990982831,-0.0234107412,0.4260779321,-0.3329880238,-0.3582531512,-0.4428260624,0.1009378508,0.2809553742,-0.0295414515,0.1378966123,-0.3276758492,0.7076433301,-0.3452907205,0.0027657019,-0.4479755461,-0.0915754959,0.0930149853,0.0227046236,0.0056790849,0.0761755779,-0.0997641608,-0.3386251926,-0.0859095305,0.1294830292,0.2044926137,0.2254512012,0.1732321829,0.2139706463,0.3390405774,0.1038615257,0.2929342985,-0.1664485335,-0.1128624454,-0.1833568662,0.4820525944,0.6298456788,0.2662581801,0.2163892984,-0.1045628786,-0.040334627,0.2494565994,0.0179534834,-0.4191426337,-0.0703290775,0.3541181982,-0.0364464857,0.6003254056,0.2047507316,-0.6425072551,0.302306056,0.482670784,0.0564567782,0.0506091528,-0.1181972846,0.1667409688,0.0330387391,0.2556284964,0.6512213945,0.2741759717,0.242681399,0.2308560461,-0.0964263007,-0.254090786,-0.2738215625,0.0713033527,-0.1925884783,0.0294661094,0.0707008019,-0.288125813,-0.2493146658,-0.3792347014,0.0157705266,0.0477979481,-0.0397264175,0.0076279645,-0.0560179465,-0.1628450751,-0.451035738,-0.3454927504,-0.1865978241,-0.6983678937,-0.1090421155,0.2895420492,-0.6256923079,-0.174624145,0.3286410868,-0.1245049089,0.1849988401,-0.6705849171,-0.3457694352,0.2005444169,-0.0656346232,-0.41620785,0.0572257787,0.3163971305,0.103096284,0.2832971215,-0.0476399511,-0.194565624,0.0146334805,-0.4240469635,0.2005152702,-0.1981097162,0.1131811365,-0.0054037529,-0.2507760525,-0.0808509216,-0.1278044283,0.1899142414,-0.0978041291,0.0012914964,-0.167727381,-0.0949386507,0.111540392,-0.0865347162,-0.7437154651,-0.541238606,-0.3228508234,0.1131119654,-0.2410251498,0.1418784112,0.2627950907,-0.0575279519,0.0206513703,-0.020281231,-0.0362934247,-0.2076792121,-0.2961928546,-0.037475992,-0.013556174,-0.1812023371,-0.0514452271,-0.0529960319,0.3328413367,0.3901850879,-0.1726878732,0.1659601182,-0.0591804609,-0.0047128331,0.1148092821,0.1560313702,0.2730443776,0.0309688933,0.0631555915,-0.0121305697,-0.0293845739,0.4256522357,0.2989344597,0.0689527169,0.2517804205,0.0168336686,-0.0664106831,0.2445134073,0.3209450245,0.349347353,0.373721391,0.0874529555,0.3018291295,-0.10758131,-0.2019774318,0.438558042,0.1080710515,-0.066428192,0.2356900424,0.0161511134,-0.3635125458,-0.2256160378,0.1721408814,0.042355828,-0.2732737362,0.3677722812,0.0257313047,0.1816153675,-0.1625988483,-0.2604582608,0.0885374174,0.2066217661,0.0591914058,-0.0404700451,-0.1929956377,0.295612812,-0.4144994617,0.0610372052,-0.3991529346,0.2309011072,0.037193168,0.2656215727,-0.1830397993,0.0845903754,0.1323578954,0.1686956286,0.1794894487,-0.7963297963,-0.5081321597,0.2580786347,-0.041272521,-0.1390909851,-0.0831522942,-0.2883306742,-0.1484971344,0.4331048429,0.0448704883,-0.4176843464,-0.2206394076,-0.158303991,0.3739377558,0.1010929644,0.2104705721,-0.1974391192,-0.1782929003,-0.1170188636,0.2660335898,0.1186481416,-0.118128553,0.304286927,0.077981852,0.0410200059,0.1341985166,0.1456559598,0.1758653075,-0.0906779766,0.088982664,-0.1730852574,0.0043387427,0.0771199092,0.1129135191,0.3774052858,-0.1973873228,-0.1878219992,0.0458315238,-0.1984352022,0.4541032612,0.3063156903,0.3032436371,0.0627226979,-0.1367657185,-0.1835803092,-0.1374144405,0.1951071173,0.0464615524,0.3057445586,-0.4204003513,-0.4959651232,0.006268159,0.2172677368,-0.1160677895,0.6183974743,0.1820365041,-0.3637962341,0.0239387807,0.1260073334,0.9849279523,0.0476905704,-0.0739722252,0.0827908739,-0.335500896,0.2396176755,-0.3381158113,0.2396823168,-0.3148294091,-0.2635342181,-0.0513521321,-0.0874922797,0.3884121478,0.2145522386,-0.1485758424,0.1712998599,0.2019474506,0.0042495709,-0.1619071215,0.4123133421,-0.251999408,-0.2226080447,0.0543112233,0.079454869,-0.0743552968,-0.1005910113,-0.0212406497,-0.0743886605,-0.0881816968,0.0843421295,-0.1392341405,0.0377072506,-0.5448747277,0.0035331319,-0.3147045076,-0.4152894318,0.1437132359,0.1752530187,0.0847182721,-0.2490741611,0.2143059075,0.0554812588,0.029830344,0.0860664099,0.2507684231,-0.3000149429,0.3130244315,-0.0505432598,-0.3360145986,-0.0765107572,-0.1037969887,-0.0291364938,-0.4604718685,-0.0021134969,-0.0890020132,-0.0638961494,0.0474110804,0.5011912584,0.2137591094,-0.1229204759,0.0853982791,0.1309883147,-0.0093306452,0.2678685188,-0.3505829871,-0.1870658547,-0.4241065979,-0.3487687707,0.1348614246,0.2235309929,-0.1319103986,0.0799195394,-0.0816700011,-0.1965603679,0.4011835754,0.5465692282,-0.0879475325,-0.2451459318,0.0401041694,-0.3437674642,-0.1185652986,-0.0462442897,0.0961434171,0.2671129405,0.2269534171,-0.1093569249,-0.1220848188,0.2045995593,0.056041494,0.1586218774,-0.1716845334,-0.2304513454,0.2959856391,0.2857683003,-0.2842213511,-0.1052112877,-0.0574468225,0.081491217,0.032275185,0.1359032094,-0.0090275388,0.2274613827,0.0398166925,-0.2499084026,-0.2717956007,-0.0514711887,-0.0815995708,0.1058994681,0.1697379351,0.1716057509,0.3742363453,-0.224957034,-0.0880228281,-0.0770933107,0.3085909486,0.1879910529,-0.0275255367,0.1648973525,-0.1111714542,0.0813963637,-0.0637393445,0.2170333415,0.1264251769,-0.0327552408,0.4839588702,0.0760216489,0.1110294759,-0.1519164145,0.0326545425,0.2393975705,-0.0714389607,0.0702328086,-0.4901865125,0.1066792905,0.258430779,0.3129652143,0.2199156582,0.3236777782,-0.1063199192,0.1365760565,0.2814666629,0.1494235098,-0.217226401,0.008129321,0.4324062467,-0.1141849384,0.1893190295,0.6332744956,0.5160235167,0.0638527796,0.3735551834,-0.3664918244,0.0195513535,-0.1580128819,0.2876561582,0.1989529878,-0.3435722291,-0.0083470196,0.4673802853,0.2608576715,0.0131729664,-0.0204512496,-0.3702901602,0.1684003323,0.4748399556,0.2019535154,0.0955993906,-0.1417230964,-0.0462937467,0.4105340838,-0.0688803494,-0.12183211,-0.0768221915,0.2128556967,-0.1870400608,-0.2848091424,0.2531356514,0.1109501347,-0.264632374,0.1139840707,-0.0329072736,-0.0848696604,-0.2566361427,0.1382213533,0.0882047042,-0.3035604656,0.3480518162,0.2333990335,0.1001546979,-0.1707016081,0.194010973,-0.2031847835,0.0243384652,-0.1241456494,0.4189016819,-0.19699049,0.0209347289,0.0248303805,0.3182596266,0.0532477237,-0.0070852917,0.0305489898,0.2666794658,0.0592819601,0.2132052928,-0.0238646101,0.2628642619,0.2646433413,0.1775674671,0.3684826493,-0.019525053,-0.043736048,-0.2275279611,0.1963344365,-0.2719421983,-0.088873595,0.0641222447,0.027378602,0.027481582,-0.1904760301,0.1537705958,-0.2445710301,0.0825679302,0.3190802336,-0.1334525198,-0.0575570986,0.0144639239,0.0764833465,-0.0535373352,0.0364789516,-0.1369408667,0.1068810746,-0.27738747,-0.3973070085,-0.5407938957,0.0473791771,-0.0731301159,-0.1428748816,-0.2532853186,0.1221587807,0.448046416,0.2421810478,-0.022599319,0.3097188473,0.2443051189,0.0017370189,-0.3667016625,0.2064484805,0.0931754038,0.3003849089,-0.0071130972,0.0821961612,-0.0121722873,0.2777653337,-0.0754387602,-0.0177313183,-0.0172352456,0.1883176416,-0.0081596533,0.318954885,0.1656632572,0.2753291428,-0.0120376227,-0.033523377,-0.2090339661,-0.0022414229,-0.4333558679,0.4266465604,-0.0440641679,0.1827373803,-0.2122016251,-0.1623466164,-0.2630758882,0.1672794968,0.0798287317,0.2675004005,-0.0244208686,0.3098958731,0.3590690494,-0.0127311787,0.0631133541,0.3329375386,0.2044288367,-0.128671214,-0.1421044916,-0.1900202483,-0.0214149933,0.351306051,0.0433107689,0.03501831,0.1487542689,0.4033871591,0.1723926663,-0.5031597018,0.0182689652,0.199925974,-0.1416805834,-0.0927898213,0.28622666,0.1462645233,-0.3139478862,-0.0572579987,-0.1617413908,0.0820256993,-0.1910757571,0.1697109193,-0.125371784]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"In this case we should have named splits, so config `ner` has splits `train`, `validation`, `test-en`, `test-ar`, `test-bg`, etc...\r\n\r\nThis is more in the spirit of the task afaiu, and will avoid making users do the filtering step themselves when testing different models or different configurations of the same model.","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":50,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nIn this case we should have named splits, so config `ner` has splits `train`, `validation`, `test-en`, `test-ar`, `test-bg`, etc...\r\n\r\nThis is more in the spirit of the task afaiu, and will avoid making users do the filtering step themselves when testing different models or different configurations of the same model.","embeddings":[-0.4026003182,-0.0958493128,-0.1003199592,-0.1407158375,-0.1486179531,-0.1347500682,0.2657963932,0.3144331574,0.0133637274,0.1866730452,-0.17175138,0.0889778659,-0.0634423941,0.341416955,0.3362111151,-0.0489127561,-0.0718723834,0.1033512279,0.2111007869,-0.0516765267,-0.068702355,0.2674790621,0.0762642696,0.1808439344,-0.1707201898,-0.396427691,-0.0069606001,0.1819518954,-0.0863823593,-0.3037915528,-0.0648695454,0.2361162007,-0.2647854984,0.2227278203,-0.0001021953,0.0186399277,0.0620974973,-0.1218779758,0.0950857401,-0.0929759815,-0.2374880761,-0.2317353487,-0.0698724464,-0.0005482199,-0.390712589,0.1390844733,-0.0086090108,-0.3298976719,0.326613456,0.3445443511,0.237161845,0.1790290177,-0.1479173452,-0.1607901305,0.225808233,-0.0567048341,-0.0942573696,0.1117315516,0.42716676,0.1156872138,-0.2336812317,0.4810343981,0.3760601282,-0.1239869222,-0.1967464983,-0.0698347092,0.116224058,-0.2355744094,-0.22294949,0.6634970307,-0.0900202766,-0.18455486,-0.404548645,-0.1773203313,-0.0388085395,-0.2991451621,0.1057845727,0.2563796341,-0.1572853923,0.0980824232,0.2095357627,-0.5304242373,-0.101264745,-0.0068214149,-0.1744519323,0.5518788695,0.0126464935,-0.0335026868,0.1698779166,0.0689379275,-0.1111647189,-0.0333246812,-0.1056858152,-0.1299832463,-0.3020146787,-0.1565746516,-0.0306329112,0.1455106884,0.1133949161,0.0860124677,0.1005492434,0.2182037681,0.1289458871,0.1670177728,-0.1857685447,0.0559552126,0.2164851725,0.019334862,0.2798151076,-0.166944474,-0.1872885674,0.0404679291,0.1680810004,-0.074464418,-0.4383301437,-0.0118650096,-0.1712932289,-0.1781162918,-0.3314332664,0.024797691,-0.0068879151,-0.1436261386,0.2092060745,0.2793003321,0.0224499423,-0.1478819102,-0.1934198886,0.1645053774,-0.2525926828,-0.3452326357,-0.1880884916,0.4269104302,-0.2951265275,-0.0494643971,0.0286358763,-0.0868245885,0.1208114997,0.0471362844,-0.1909853667,0.0883193612,0.362795651,-0.0257854871,0.0616963431,0.1848794222,0.0155417407,-0.108741723,0.0349667706,-0.1832040399,-0.3418172002,-0.2202503234,0.0020331701,-0.2498936206,-0.2068327516,0.2593634725,-0.1927213967,-0.1097695604,0.501901269,0.7477043271,-0.1138684452,-0.3097315729,0.0103494432,0.0722304881,-0.1551324427,-0.1781288683,0.142340228,0.2348195016,-0.7939331532,-0.4151142538,-0.1149807125,-0.37403813,0.2320399284,0.0306626987,-0.1746026278,-0.179068774,-0.067640312,0.5377990603,0.4789310992,-0.0326394998,-0.150624916,0.1968884468,-0.3211741149,-0.0246893503,0.3203324676,0.1018750742,0.0402509756,-0.0075318152,-0.1099249423,0.5292525887,-0.3676404953,-0.3368195891,-0.2437324375,0.0249286927,0.3988683224,0.1669460535,0.2497586906,-0.3429686427,0.1850258559,0.0071820016,0.0768774003,-0.2135211527,0.0529179722,0.1174665317,0.3316627443,-0.1416356713,-0.1905556768,-0.1798572838,-0.1615664065,0.0233864971,-0.4569210708,0.3356417418,0.2028000057,-0.1793420315,-0.0773236156,-0.2438413352,-0.2253038138,-0.256334573,0.2326025069,0.2952122986,0.1288315207,0.1328589767,-0.1517306715,0.348810941,-0.2325498164,0.1285135746,-0.3362828493,0.1396402419,-0.1186682731,-0.185119763,-0.1478704065,0.1360009462,-0.1948617548,-0.2928481996,0.0158039331,0.4303457737,0.1253077984,-0.0231997836,0.2271459699,0.0917504504,0.2911508083,-0.0336008631,0.117426835,-0.0836509094,-0.0812103897,0.0772666857,-0.0643262267,0.5886890292,0.2394322455,-0.1911157072,0.189277336,0.0202029254,0.3595643044,-0.1796154678,-0.1154732704,-0.2785170972,0.3284752965,-0.0679824799,0.1604867578,0.1114509106,-0.4790667593,-0.0094287936,0.4859450758,0.0590871796,0.3648809493,-0.0237304829,0.1297607273,-0.1083734483,0.0682213306,0.4315474927,0.1073834673,0.2848714292,0.2683021128,-0.0328877755,-0.377515614,-0.2526907027,0.1884729117,-0.1644174308,-0.0211052727,0.2314586639,-0.178697601,-0.4338044226,-0.4670210779,-0.1032855213,0.2745408714,-0.0487776957,-0.1409630477,-0.1502785385,-0.1997580081,-0.2132203281,-0.3502658308,-0.097134158,-0.3459997773,-0.1867788583,0.5182228684,-0.2924462557,-0.1760054976,0.3785411119,-0.1598130018,0.307507813,-0.2044347972,-0.0438871458,0.0040378701,-0.0985976011,-0.3283260763,0.2513146996,0.1908450872,0.2258790731,0.3649789095,-0.0743880644,0.0986498892,0.0716004074,-0.7308645844,0.1624788344,-0.1577544957,0.2988201678,0.1267984957,-0.1182792559,0.1131324768,-0.2868728042,0.2103506625,-0.240805313,-0.1970635355,0.0792607963,-0.040716324,0.0391069427,-0.2071117312,-0.8082507849,-0.4853231013,-0.3589679003,-0.1158563197,-0.1579067409,0.1148273051,0.1002138779,0.0731069371,0.1114850715,-0.2805955708,0.007613312,-0.151039198,-0.123407498,0.1163697094,-0.2014593482,-0.1243724003,-0.0786941051,-0.20791623,0.1581918746,0.0877095088,-0.1286592036,-0.0306887534,-0.0390449129,0.1243704483,0.4112747014,-0.0139590818,0.3051303625,-0.0276305657,-0.1860746443,-0.2051464021,-0.0044551315,0.3356563449,0.0322706066,0.1974271834,-0.120324634,0.3349353373,0.1339778304,0.3058947623,-0.0827450231,-0.0348390378,0.3734642267,0.0061591254,0.1780783236,-0.0371430963,-0.1299080104,0.222192198,0.0400747694,0.1338741034,0.1958666742,-0.0403218754,-0.3794880509,-0.0636169612,0.0508130901,0.0142224832,-0.2456770539,0.276737839,-0.0690668598,0.233041212,-0.0647961348,-0.295793891,0.0468138643,0.0790789127,0.0880564228,0.0216556359,-0.0771389157,-0.0655036867,-0.6089606881,0.0905625746,-0.0617359467,0.3322027028,0.2150345296,0.2362167239,-0.3545494974,-0.2137061656,0.0067057349,-0.0045267702,0.4245501161,-0.6387299299,-0.325799644,0.1251165271,-0.1791613847,-0.1754115671,-0.0034703899,-0.4442122579,-0.2253208458,0.4683688581,0.1449165195,-0.3436183631,-0.2765908241,0.1827060133,0.0970703661,-0.2005446255,0.2465616465,-0.1700361669,-0.0871610567,-0.2539561093,0.425757736,-0.1659005731,0.0622204877,0.2831777036,-0.1150996312,0.0646717772,-0.0397014283,0.3100999296,0.4098554254,-0.0249954015,0.0888762474,-0.1791358888,-0.071295619,0.1417743117,-0.1259770393,0.2017907947,-0.2762452364,-0.0261723269,0.1000604182,-0.2319994867,0.355178535,0.3125629127,0.128004089,0.3205566108,0.049450919,-0.0521539412,-0.0842237547,0.4497030079,0.0158807077,0.2580981553,-0.3180544972,-0.3337843716,-0.0661654025,0.2285652608,-0.2107949406,0.2800613642,0.2676205635,-0.2318145037,0.1874694228,0.1028197035,0.9224636555,0.0598705113,-0.1200500131,0.1358322501,-0.1968904585,0.1186690181,-0.3971567452,0.1130547598,-0.2436854392,-0.3141518533,-0.1351844668,0.0055969525,0.4562395513,0.1878727674,-0.2996572852,0.2148268074,0.1358685046,-0.0031979871,-0.0746267736,0.3480921984,-0.1192134321,-0.1294186413,-0.0360304266,0.1912573278,0.0044823051,-0.1313331723,0.066704303,-0.1764061004,0.0329360589,-0.0699042603,-0.1038202569,0.0413483121,-0.1775310934,0.1177050248,-0.0536804274,-0.1844181418,0.1092241183,-0.0690407157,0.1380472928,0.1580834836,0.1200252473,0.2941990197,-0.0762510523,0.147102043,0.1365595311,-0.1427118629,0.6267646551,-0.1289560199,-0.3069669008,0.2094914764,0.056266211,-0.1752725542,-0.1619961411,-0.0475414991,0.0762052089,-0.2443705797,-0.0056422804,0.2292590588,0.0134975119,-0.4274725318,0.2006150186,0.0614404008,-0.4205586016,0.1520059705,0.0173499659,-0.1653868258,-0.2778964937,-0.2556268871,0.1216323376,0.3193447888,0.044809889,0.0902811363,-0.0682641938,-0.368065834,0.0006735869,0.787630558,-0.3748192787,0.1325340122,-0.0687289461,-0.5916549563,0.0773957819,0.0315190665,0.3649032712,0.3213346899,-0.0512353443,-0.1956501305,-0.0381207503,0.0758216158,0.0719577447,0.4078860581,-0.080503583,0.0682991445,-0.0274340641,0.4570300579,-0.4596618116,-0.119225502,-0.3372139335,0.2345479429,-0.2614818215,0.0434475988,0.098760888,0.1737754196,0.2200475037,-0.0975527018,-0.2103373557,-0.2379668802,-0.160009101,0.0657836199,-0.1205589548,0.2012432516,0.2069998682,0.1868963838,0.0833775923,0.0131908152,0.3018627763,-0.2768757641,-0.020870965,0.1772527248,0.0810480416,-0.247232154,0.1071163863,-0.0407219119,-0.089287594,-0.0021475621,0.2132828236,0.1496021599,0.040396329,-0.19369708,-0.0163378473,0.2318391502,0.1679383069,-0.0342714489,-0.3017490804,0.0924228877,0.0133815054,0.0927657709,0.3654087484,0.1559432447,0.0438667126,0.2086666971,0.2636063397,0.341898948,-0.3398114741,-0.0437317006,0.5402928591,-0.1636401415,0.4729087949,0.3172200322,0.3986676633,0.4365920722,0.3079837859,-0.1262674481,-0.0659623742,-0.0754210651,0.258602798,0.410956949,-0.0899896696,0.0364705957,0.4287820756,-0.0308452193,0.0710817203,0.1886335462,0.0119573139,0.4845588207,0.1763234735,0.3700490296,0.1930938363,-0.1433769315,0.0301436763,0.1022964641,-0.0258101542,0.1754262447,0.0862090588,0.2275396287,-0.1140866652,-0.2702920735,-0.0859168842,0.2736179233,-0.1358747333,-0.167897135,-0.2852213085,-0.2245147079,-0.258462131,0.11921224,0.3195637167,-0.0634582266,0.1701209247,0.1603124738,0.1590656936,-0.0753977001,0.2455549836,-0.0336562172,-0.0297224279,-0.3305515945,0.3696353734,0.1434023976,-0.0854761899,0.0771523491,0.0752734616,0.1206653491,-0.2689931989,-0.0746615976,0.148532778,-0.0547177866,-0.0543552414,-0.1038949192,0.2398795187,0.2367775589,0.432272613,0.346018672,0.1975118518,-0.2211991847,0.0832738876,0.110762164,-0.1613859683,-0.1416667551,0.2433816642,0.1883159876,0.2390518486,-0.4731658399,0.0350847803,-0.4495849907,0.0903759003,0.3246405721,-0.2897536755,-0.0027567314,-0.0898313299,0.1125625148,0.1293251365,0.3142217994,-0.1294065118,0.0516743399,-0.1591435075,-0.2019968927,-0.4413145185,0.0547309816,-0.040384315,-0.1835627258,-0.2636707127,0.3045909405,0.2783789039,0.0368914194,-0.1825136393,0.1389006078,0.0346687622,0.0216567982,-0.3971924186,0.267428726,0.0383198,0.2621863186,0.0511260629,-0.0487593673,-0.030092109,-0.0213439856,0.0624181293,0.0825682878,-0.2733342648,0.3914412856,-0.0397450067,0.1981405467,0.0442067645,0.565777719,-0.1486567408,0.0834050626,-0.1370043904,-0.0643252432,-0.4202280045,0.1341945678,0.2424262762,0.512812078,0.1195674092,-0.1767282337,-0.1290865093,0.3237064779,0.1242190078,0.0261499062,0.0069240648,0.2358243614,0.2107907981,-0.1603044122,0.3413079977,0.5047712922,0.1459008753,-0.1016613394,0.0846766979,-0.3917711377,0.1512981802,0.0446778052,-0.0946692973,-0.1255572885,0.0433895737,0.3962927163,0.0148865981,-0.4677683115,-0.1340250969,0.1449148208,-0.03995746,-0.0277339667,0.2085355967,0.4323608279,-0.0117175402,-0.0592451766,-0.3107558489,0.1372046173,-0.2201346308,-0.0102339657,-0.2752775848]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"I see your point! \r\n\r\nI think this would be quite feasible to do and makes sense to me as well! In the paper results are reported per language, so it seems more natural to do it this way. \r\n\r\nGood for me @yjernite ! What do the others think? @lhoestq \r\n","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":49,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nI see your point! \r\n\r\nI think this would be quite feasible to do and makes sense to me as well! In the paper results are reported per language, so it seems more natural to do it this way. \r\n\r\nGood for me @yjernite ! What do the others think? @lhoestq \r\n","embeddings":[-0.0309355296,-0.0157449823,-0.1258900017,-0.1005204991,-0.1579117477,-0.1708758622,0.5267757177,0.0056056301,0.0238729157,0.0956136808,-0.0717838034,0.004015557,-0.0292193759,0.2641967833,0.4696397781,-0.0064959237,0.0179228708,-0.092742525,0.1055237725,-0.0311875083,-0.2255672514,0.0408765003,0.193488881,0.128103286,-0.1308085173,-0.2262523323,-0.1815516353,0.0034792195,-0.1671005636,-0.1984765381,-0.1351335496,0.1272173673,-0.2173869312,0.0498540252,-0.0001046478,-0.0820348337,0.1216271967,-0.0414647125,0.3431464434,0.0987389162,-0.2769545913,-0.5269677639,-0.3578154147,-0.2653224468,-0.0164584536,-0.1502900273,-0.0198334958,-0.5409762859,-0.2230930328,-0.0831348747,0.2834675908,0.0115481792,-0.0138974572,-0.0006390887,0.3840331733,-0.2163021564,-0.1017036885,0.3516216278,0.6787539124,0.199782446,-0.0622564666,0.5023947358,0.1719843894,-0.3639404774,-0.1554909796,0.1548494995,-0.0157572031,-0.1777889878,0.0044577923,0.4887833893,0.173227191,-0.2198902369,-0.5166098475,-0.151089713,-0.0256642811,0.1456084549,-0.1370281279,0.3663782179,0.3165973425,0.060445115,0.3006005883,-0.5922685266,-0.0474885926,0.0295654032,-0.0114819733,0.3938578069,-0.0192892551,0.2202584296,-0.0696594045,0.1565402448,0.0442250147,-0.0453594588,-0.204782337,-0.1475442201,-0.0427239612,-0.0027037682,0.0609702058,0.2228372246,0.2347803414,-0.2167901397,0.1660678089,0.1159691736,-0.1366979927,0.1365009844,0.0484070703,-0.0670579821,0.1184437573,-0.1217701584,0.1872103214,-0.4778737128,0.1259695441,-0.0056065572,0.1264991909,-0.016919516,-0.2524387836,0.0758058429,-0.1182870269,-0.1687624305,-0.1406301409,0.0159833841,-0.1670643687,-0.2575620711,-0.1345190555,0.1522021443,0.1174124181,0.0314531066,-0.0274043716,0.3192719519,-0.0030035877,-0.2601382136,-0.0182076767,0.4555564523,-0.3091601133,-0.0275301822,-0.1148491129,0.0074088625,-0.2264327556,0.1829274595,0.1247549951,-0.0296654589,0.293187052,-0.0992409214,0.1960194558,-0.2719397843,0.0791913196,0.1489207894,-0.0673978552,-0.1974616349,-0.3425369859,0.1498023272,-0.0699459091,-0.0092685549,-0.4532342255,0.1946405917,-0.0317016505,-0.3639160395,0.2336077541,0.6543378234,0.0268253498,-0.1571780443,0.3393564224,0.1094693989,-0.1671152562,-0.3618362248,-0.1114996001,0.1597103328,-0.5814822912,-0.203640148,0.1548270285,-0.109760277,0.2220728844,0.0983588547,-0.3220052421,0.0248717554,-0.1704873741,0.4339231551,0.2253325135,-0.3000765145,-0.1407328695,0.0266584624,-0.3058372736,-0.0017787977,0.3271745145,0.2802770734,0.0364675559,-0.0436035655,0.000573321,0.2896563709,-0.1517328173,0.0055794748,-0.2849932313,-0.2372041643,0.3821450472,0.1792827547,0.0955134854,-0.2297505885,0.4316015542,-0.0239207186,0.1088688225,-0.4153653085,0.1452231556,0.0073805819,0.3876383007,-0.1258910447,-0.2548124194,-0.1163973585,-0.189257443,0.1494081169,-0.1739997417,0.3878805339,0.2392870933,-0.3596981466,0.1090210229,-0.3047987223,-0.0997588709,-0.2800357342,0.2225359231,0.2846927643,0.0505989417,0.2357271463,-0.179836601,0.5148310065,-0.0650091097,-0.2156123966,-0.2844980955,0.2719512284,0.0615566373,0.0166402161,-0.3019689023,0.3540425003,-0.2673735023,-0.1714290977,0.224920854,0.2003877908,-0.2806517184,0.2007848471,0.3705322444,0.2134396136,0.3387433589,-0.0368191376,0.4451640844,-0.2990204394,-0.0064352131,-0.050741341,0.2629058361,0.5275496244,0.3378444314,-0.020178156,0.0635634065,-0.0839885473,0.1785039902,-0.066599004,-0.2512602508,-0.1478615701,0.2185904235,0.1833827049,0.3503826559,-0.0147361625,-0.5163173079,0.3014975786,0.6254872084,-0.0292419922,0.3895092905,-0.0920934603,0.019295929,0.0229834095,0.0512353405,0.3502872884,0.243791163,0.3249136209,0.2395406812,-0.122097522,-0.0230479985,-0.0875405967,-0.0093188984,-0.2209279537,-0.0916001499,0.1658771932,0.0760681331,-0.1549377292,-0.5849462152,0.0798439234,0.02613841,-0.322053045,0.048566442,-0.0801828578,0.061007157,-0.3757747412,-0.4289927185,-0.2512623966,-0.3670115471,-0.101163052,0.2450068295,-0.5446105003,-0.1156204268,0.0839172825,0.1357588172,0.2886963189,-0.4931294918,-0.0888581052,0.0717839226,0.0521639548,-0.1208332255,0.2150434107,0.0649465099,0.0613452233,0.6506130695,-0.0297608357,0.2131448984,-0.3529451191,-0.5960319638,0.0957663953,-0.2590044439,0.0248189643,0.102424629,-0.4434499741,0.0465821251,-0.1395166814,0.2609799802,-0.2049952,-0.0151072675,-0.2305301726,-0.095352456,0.0693260282,-0.1832613647,-0.4359954894,-0.6468153,-0.3434995413,0.2010104954,-0.0807763264,0.0785352811,-0.2243638486,-0.1936884075,0.203163743,-0.1049615964,-0.1506704241,-0.3018170893,-0.2455632687,0.0002138582,-0.1281707138,-0.2957010269,-0.1156471148,-0.0906981155,0.1820831895,0.1190915406,-0.1470335871,-0.0806309506,-0.0669408739,0.0293893516,0.2819399238,0.4289917946,0.2110100538,-0.0370901003,-0.151827246,-0.1413828284,-0.0387817211,0.0877371654,0.1865488589,-0.0784775093,0.1115677878,0.1964416504,0.2549844384,0.0981981382,-0.0094350213,0.0713188946,0.4470978677,-0.0311266575,0.3983821869,-0.2327969074,-0.3313070834,0.2287492305,-0.1843125373,0.1617482603,0.1289574057,0.0806174204,-0.4328382313,-0.1319629103,0.1240723878,-0.2155031115,-0.122989215,0.2715333998,0.1218841895,0.1368045211,0.0268952809,-0.5252269506,-0.0980471522,0.0657746345,0.1927143782,0.0876701176,0.1008327156,0.0204323381,-0.4529147148,0.1035572812,-0.6617617607,0.3235164881,0.3332972825,0.0845971778,-0.1681655943,-0.026971329,0.069974333,-0.0831562728,0.1652536988,-0.7830041647,-0.403886497,0.2012939006,-0.1501503885,-0.0906749964,0.0944576785,-0.4324661493,-0.0122224772,0.4249150753,0.1434320509,-0.4903404415,0.0758595243,0.1031939909,0.2933271229,-0.0366541892,0.1504915804,-0.1315781921,-0.2456531972,-0.2036443353,0.0841045082,0.089802593,0.0225345623,0.4347544909,-0.2493488044,-0.0478661321,-0.0738417432,0.1002015024,-0.0376438387,-0.2105487287,0.2348745316,-0.1014237329,-0.1369154453,0.1677592546,-0.2899704278,0.4333432317,-0.2037704587,0.1332355589,0.1908707917,-0.4832280874,0.497356832,0.2202970833,0.2189731449,0.2997460067,-0.1233195364,-0.1388925612,-0.1428907514,0.1032074317,0.1299720109,0.4290404916,-0.4069592655,-0.365621984,0.2014386058,0.2321173102,-0.0997955427,0.5052694082,0.3039492369,-0.4377226532,-0.0691275448,0.0089062927,1.0040178299,-0.2564153671,0.1959460676,0.0578914806,-0.1796460748,0.3725501597,-0.3016745746,-0.1616475433,-0.2634377778,-0.2750139832,-0.1262907386,-0.0533952117,0.3447171152,0.1411759555,-0.2564907968,0.0261163656,0.199685052,-0.0821727291,-0.0594161972,0.5094427466,-0.163037315,-0.2448430359,-0.071276553,0.1428415775,0.1211063936,-0.1961435974,-0.2274865657,-0.0276074912,-0.1567518115,-0.0975118801,-0.2363921404,-0.0394493379,-0.052648101,0.1502730101,-0.2101896703,-0.2341324538,0.2122021765,-0.223246783,0.1911126673,0.0326602608,0.1682790816,-0.0387043431,-0.0087649338,0.1832389534,0.4271745086,-0.2175198346,0.2066283375,-0.0606686436,-0.1605679691,0.2272233069,0.0343399905,-0.5253437161,-0.513217628,0.2669925094,-0.0754744411,-0.1264099777,-0.0614992231,0.6240118742,-0.02113254,-0.2172413319,0.1606660336,0.3921788335,-0.0677852109,0.1104492098,-0.1601873338,-0.2068914175,-0.3168252707,-0.1244885698,0.2557013333,0.096789971,-0.367056936,0.0462516025,-0.155927524,-0.1189372763,0.1399715841,0.6067633033,-0.3782593012,-0.288372606,-0.0039127832,-0.3069781661,-0.2110751122,0.0479427911,0.166558221,0.2156911045,-0.0879016221,-0.1916357279,0.0870058611,0.3477586508,0.1552591622,0.1503543407,0.0880595744,0.00334629,-0.040152099,0.106058307,-0.4068502784,-0.000190837,0.1143755168,0.122680746,0.1580331624,0.0464225896,0.1341831386,0.158901751,0.1971456558,0.0210437514,-0.05176102,-0.1954434812,0.0383411311,0.0593883768,0.0830747858,0.1936604232,0.343683362,-0.1540673226,0.0578893833,0.0714375302,0.3258045614,0.0293072183,-0.2294366211,0.041692324,-0.1714077145,-0.0834431276,0.1062697992,0.2818501592,0.0851332843,0.0294380561,0.0412163772,0.1196720377,0.1314347684,-0.0701946989,0.2505352497,0.2301358581,0.1264226586,0.3121958375,-0.1550155431,0.0587662868,0.3654409349,0.1273471862,0.4827502668,0.0336059108,-0.1689761281,0.0823084563,0.0709677786,0.313657552,-0.2416971028,-0.2967020273,0.2061612755,-0.2194766849,0.3902001083,0.5765987635,0.5690054893,0.2632627785,0.400457859,-0.4132812023,0.1162246764,-0.0859441459,0.0182973053,0.0456557572,-0.2124733478,0.0696043819,0.2831362486,0.1888386458,-0.0315789469,0.2012163401,0.161131829,0.2795372009,0.4283171594,0.2462067157,0.0271936376,0.1925108284,0.3381958902,0.1764977872,0.2595503628,0.1481246054,0.1729340255,0.2318243831,-0.1147492528,-0.0794658288,0.1311266869,0.1838276386,-0.0495331623,0.0149042243,0.2187445313,-0.0903742164,-0.242602542,0.0026589273,-0.1728191525,-0.1071863472,0.0830358639,0.0575165004,0.287306875,-0.0143342698,0.1595532149,-0.1490688175,0.1236341074,-0.2348009944,0.2202052325,-0.1538406014,0.1885551512,-0.0892922133,0.4514083564,0.0201955643,-0.0293596964,0.0048573669,0.3698182702,-0.2189735621,0.1569337547,-0.1249351278,0.133853972,0.2747658193,0.1748584062,0.3241960406,0.1354686916,-0.2330461293,0.0122821359,0.1976193935,-0.0397358537,-0.3034680188,0.0021590977,-0.2225866914,0.012468108,-0.2757618129,0.0758448243,-0.5223271251,-0.0902155638,0.223592937,-0.3513811231,-0.0257022511,-0.1148734614,0.1095340997,0.0350923575,0.1946718544,0.0465830564,0.4593777657,-0.1099379435,-0.3727731407,-0.4824965,0.2873205245,0.0362300314,-0.089895919,-0.3536243439,0.1818208843,0.4511235952,0.2547105253,0.0121647855,0.2115444094,-0.130023405,-0.1583548337,-0.2679615617,0.1299415827,0.2145318687,0.2390740067,0.0139873559,-0.0470988266,-0.0220199935,0.094101347,0.1614574492,-0.212352097,-0.139204964,0.0058388938,-0.0993597209,0.4858988225,0.0645322576,0.4056458175,-0.2137445956,-0.1554318368,-0.1162483618,-0.2418531775,-0.1959832907,0.4852727354,0.0653437376,0.2561366558,0.0638890341,-0.2525367737,-0.0699633956,0.1303884536,0.1136340946,0.35016945,-0.0011427741,0.3119924963,0.1250369251,-0.2902595699,0.0264784563,0.4481160939,0.1819459498,-0.0777805075,-0.0206536539,-0.2677201927,-0.1471849233,0.039859768,-0.06971246,0.0591698624,0.2188392282,0.1532175839,-0.1139371619,-0.4214032292,-0.075693503,0.3150254488,0.1775716096,0.0155050075,0.1421088874,0.0229909737,-0.3856188357,-0.0903417319,-0.0870592669,-0.0052371281,-0.1229972839,0.1273497343,-0.3738960028]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"I agree with Yacine on this!","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":6,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nI agree with Yacine on this!","embeddings":[-0.2618382275,-0.123177737,-0.1966786087,-0.1130408496,-0.0705870762,-0.1502931565,0.2662186921,0.1024881676,0.049780786,0.0746994615,-0.0425684713,0.0228724871,0.0352675132,0.2335611284,0.4506898224,-0.1696523428,0.0840329826,0.0307854917,0.0524904542,-0.0738410056,-0.21584858,0.2035902143,0.139892593,-0.0431322008,-0.1695101559,-0.318888694,-0.189951703,0.127466023,-0.146120742,-0.2392690331,-0.185634315,0.1550931484,-0.2172268778,0.0614663213,-0.0000953538,-0.1535126716,0.1057476774,-0.0402307622,0.1728555113,0.1327358484,-0.2640264928,-0.4511128366,-0.2658006251,-0.1031641737,-0.1453418285,0.0052080643,0.1220647395,-0.3816010952,0.0163565241,-0.0442703366,0.3555456996,0.042644646,0.0801235959,-0.160888344,0.3837733269,-0.1927790493,-0.0744989216,0.169593066,0.7076691389,0.0635426268,-0.0952736214,0.3055783212,0.0180030223,-0.209280476,-0.1201850176,-0.0888182446,-0.0507968143,-0.1288697273,-0.0659716651,0.5248299241,0.2002372593,-0.2432499677,-0.2598036826,-0.0923134908,0.0383914895,0.0588283725,0.0508521982,0.2048772424,0.1018913612,-0.0547104105,0.2638055682,-0.4965796173,-0.110403426,0.1435364634,0.0240762755,0.4227922261,-0.121576637,0.0113294534,0.0754692182,-0.0099703111,-0.0927945897,0.0704177171,-0.1156066507,-0.0181693323,-0.0965230837,0.010071571,0.0339906402,0.2233591527,0.1715191007,-0.1303382665,0.1880859584,0.3099137247,-0.1009657234,0.2755238414,-0.1013737172,-0.0969772637,0.0875863507,-0.3152606189,0.3033432961,-0.3120131195,0.0695121139,0.0858463049,-0.0146761695,0.0051078238,-0.326584816,-0.0282836575,-0.1211543158,-0.0209592152,-0.1886761039,0.0818557814,-0.070929721,0.0029591713,0.030375639,0.1763184667,0.0665554479,0.0044896528,-0.0676994026,0.213798061,-0.0642719641,-0.2925975025,-0.1070516333,0.4638045132,-0.2865490019,-0.0393219851,-0.0516330861,0.0296518821,-0.024669189,0.3144475818,-0.1159033403,0.1255046576,0.3007852137,-0.0192747358,0.1609501839,-0.1042611748,0.0251536481,-0.0596641563,-0.040239241,-0.2256619185,-0.3323200047,-0.144607693,-0.014218987,0.0597632863,-0.5089257956,0.2977975011,-0.0637629852,-0.3550580442,0.3392281234,0.6093209386,-0.1343419254,-0.1511291414,0.3318512142,0.1528866142,0.0057707126,-0.2723930776,-0.0577803403,0.2082204968,-0.6349985003,-0.2782033086,0.047687687,-0.1966089904,0.1323487908,0.0768143386,-0.1345886439,-0.2032061219,-0.1329520196,0.3717733026,0.2095955461,-0.2701429725,-0.1727863997,0.0872588456,-0.2041032165,0.061079897,0.2187633216,0.3435662091,0.0656014532,0.0287310928,-0.0377488919,0.5434561372,-0.09593077,-0.0715478584,-0.1809606552,-0.2019014359,0.0535563454,0.2144715339,0.0669218227,-0.3102498055,0.4042906761,0.0032153588,0.0190345384,-0.2211040407,0.0832316875,0.3163431883,0.2314508408,0.0008979499,-0.1725084186,-0.2485290319,-0.2069747746,-0.0411159545,-0.1203226671,0.2797703743,0.0994388461,-0.189672336,0.0166738462,-0.1728433818,-0.1407924443,-0.0969699621,0.34195292,0.3396552205,-0.0140738087,0.2250525802,-0.0734630376,0.4131920636,0.0579950511,-0.0676998124,-0.2143491656,0.0981365964,-0.0098228743,-0.1458452493,-0.1142768264,0.3254329264,-0.2678229809,-0.2128294408,0.2251909375,0.3847061694,-0.1422930807,0.3194935918,0.3387418091,0.3216494918,0.3257926702,-0.0356976874,0.3474074006,-0.2136949748,0.0753547326,-0.0129052075,0.244113043,0.4136523306,0.3465028703,-0.2504204214,0.1145524979,0.0955615714,0.3204693198,-0.1004021019,-0.0162196998,-0.2594081163,0.2842747867,0.0774397179,0.3983788788,0.0115043288,-0.4496573508,0.2280444205,0.6372754574,-0.1074441597,0.5722082853,-0.0827446356,-0.239277482,-0.0877125636,0.0917470753,0.3858900964,-0.0317183286,0.3275500238,0.253210634,-0.1788594574,-0.1514480561,-0.224912405,0.1766568422,-0.2296546996,-0.0940539911,0.1727136374,-0.0663125664,-0.2443848103,-0.7233072519,-0.0536221936,-0.0025880395,-0.1132152304,0.1130845845,-0.1879325062,-0.21476762,-0.3762944937,-0.2371354997,-0.2767520845,-0.3849237859,-0.2230731249,0.3052365184,-0.3404968679,-0.14573358,0.1877456009,0.1033339202,0.2711805701,-0.2790420949,-0.1893102378,-0.0971448347,-0.2272491902,-0.1440226883,0.3178740144,0.0562149063,0.0282727517,0.4346320629,0.0023422607,0.3996018767,-0.2386180609,-0.6179055572,0.1740061194,-0.1789790988,0.1134227589,0.2017412335,-0.1690248996,0.0743033811,-0.2025606036,0.1267107129,-0.2986336648,-0.0506571308,-0.0854031742,-0.2182105929,-0.130650267,-0.1694166064,-0.3412024081,-0.4239289463,-0.394972235,0.2163518965,-0.0966688991,-0.0409293212,-0.062113978,0.0050877267,0.321833998,-0.1429105401,0.032082729,-0.3033633232,-0.2933074236,0.0736350194,-0.1875010729,-0.2606122792,-0.1475309432,-0.1449529678,0.2272320539,0.0347452126,-0.2136043757,-0.0945152193,-0.1731255651,0.2020224333,0.3151617348,0.2346010506,0.2273909301,0.0478300489,-0.2951035798,-0.2257036567,0.022518998,0.1256612092,0.1002054587,-0.0171451122,-0.0410738438,0.139824152,0.1364652961,0.2040445358,-0.2930427194,0.0010839527,0.5024713278,0.0870520398,0.4506745934,-0.1411722302,-0.168377459,0.1922772825,-0.0851032808,0.3842844367,0.1940753609,0.0453380235,-0.4332190156,-0.0474616624,0.0027261807,-0.1839999855,-0.0691039339,0.1845856309,-0.0762820169,0.0736895204,-0.013014378,-0.4526517689,-0.1065232605,-0.0095329676,0.2192477435,-0.0248152632,-0.1171465963,-0.0941126198,-0.4033853114,0.0828110278,-0.4163180888,0.3747442365,0.277977407,0.0811795816,-0.3489916027,-0.1386424899,-0.0176035035,-0.0400353707,0.1840273291,-0.6582477093,-0.3768718243,0.2216307968,-0.0670080557,-0.1654496342,0.2300039232,-0.5266138315,-0.0506538674,0.3916855752,0.142287448,-0.2467406094,0.0106364014,0.1203776821,0.1604206115,0.0352125242,0.2456943691,0.0035948243,-0.2158908546,-0.2861940563,0.1211995333,-0.0732498169,0.0269424394,0.5596287251,-0.144780159,-0.1001693308,-0.0954545811,0.1795411706,0.2224706262,-0.178403452,0.1738958359,-0.2230199873,-0.0671028942,0.0868504792,-0.1394432038,0.2605708241,-0.1675680131,0.0170625523,0.1382520348,-0.3688492477,0.4051083624,0.0614149794,0.2773840725,0.379843384,-0.1939739883,-0.1538829654,-0.2364285439,0.214667201,0.0346330255,0.4595777392,-0.2819295824,-0.2186391205,0.1664097309,0.1457823962,-0.2202766687,0.366022259,0.1607603878,-0.2985445261,0.0733988732,0.0955035463,0.933837831,-0.1132375449,0.1155146435,0.1309912503,-0.2363582402,0.193717733,-0.34082672,0.0119084837,-0.2077218443,-0.283842057,-0.137303412,0.0940710828,0.2520132363,0.0839770064,-0.3144944608,-0.0678570122,0.1487810463,0.0024535912,-0.0633661374,0.2722357213,-0.1709335446,-0.2460264266,-0.1171183214,0.2920070291,0.0892297179,-0.1095148772,-0.1930678636,-0.0876056328,0.0609943941,0.0141086914,-0.2716012299,0.0521688834,-0.1726543009,0.1994530857,-0.0497342423,-0.1398762316,0.0067763007,-0.1169487089,0.1397252977,0.0282848459,0.0336578079,0.2369663417,0.0860798657,-0.031052297,0.232526511,-0.0763554722,0.4202235937,-0.1104880869,-0.1704436392,0.2629339397,0.0493143015,-0.2874183655,-0.4121020138,0.1683930904,-0.045314528,0.0926717296,-0.0516589545,0.3927904367,0.0349156223,-0.3446025252,0.2487957776,0.217421636,0.0012569248,0.0513979755,-0.0542113148,-0.2095478028,-0.3484693468,-0.0038436684,0.075491108,0.005682827,-0.0827022344,0.1142074838,-0.1739058197,-0.4119566977,0.1221974045,0.5029071569,-0.2650448978,-0.0603152588,-0.0746972263,-0.5638820529,0.0249707308,0.0322278664,0.2666935027,0.0849789307,-0.0962813869,-0.2266627848,-0.046968475,0.2115354091,0.131243214,0.2668715119,-0.0062968796,-0.138812229,0.1008431017,0.2236604691,-0.4980123937,0.0147320116,-0.0847297981,0.0517527945,0.2127681673,-0.0830934644,0.1858550459,0.1863524169,0.2837540805,-0.1115483195,-0.0968864262,-0.3121792674,-0.2252245545,0.0292495508,-0.1915279627,0.1571982056,0.3110649586,-0.0553085469,0.0855762586,-0.0759091452,0.3098083436,-0.1066405028,-0.1540370136,0.1089998409,-0.0430287197,-0.0377529636,0.2773827016,0.0577183478,0.0348580629,0.1125091836,0.1293311417,0.0148261748,0.0882244334,-0.1898311973,0.2626052797,0.3460908532,0.2754569054,0.0025742338,-0.0661720186,0.1631418616,0.0854619741,0.2141254395,0.2283053696,-0.1157781258,-0.1058685258,0.0004209341,0.1301726252,0.3859511912,-0.2879527807,-0.2697156668,0.2733829618,-0.120880872,0.4493782818,0.3400016129,0.4199036658,0.1637471467,0.3342348039,-0.2560541332,0.1339589506,-0.0895751789,0.1678557843,0.1405151039,-0.1229332015,-0.1152761877,0.2087747753,0.1885407716,0.1267953366,0.3226368129,0.0774015933,0.2301945686,0.4554693103,0.3126531839,0.1735657007,0.0702381432,0.2690598667,0.2314387858,0.1125183329,0.3201837242,0.1745916307,0.3403559625,-0.1504477859,-0.0803410634,-0.0499991626,0.3193646967,-0.148103103,-0.2050585151,0.1425908208,-0.1779529899,-0.2923854589,-0.0144203501,0.1012850106,-0.0165984947,0.2690847814,0.0395875461,0.3244552612,-0.1624946296,0.2201465517,-0.01475811,0.0293923393,-0.2435404211,0.3038011491,0.0021788806,-0.0573617108,-0.0445412807,0.3097155392,0.1879519075,-0.087484479,-0.1679289192,0.3861912489,-0.265776962,0.0323433988,-0.1779052913,0.1577568501,0.3044340014,0.3496803045,0.3144257069,0.2365701944,-0.2658677697,0.0024121033,0.139647603,-0.1504211277,-0.0193661246,-0.0867674202,-0.1257082969,-0.0129014598,-0.4462788403,0.1609750986,-0.5164832473,-0.1259454787,0.288387686,-0.3102853894,-0.0472939946,-0.0701579377,0.1513628662,0.0883361772,0.4046935439,0.1665095389,0.2397549152,0.0992886052,-0.4342949688,-0.4536548853,0.1775550097,0.0130660897,-0.2281029373,-0.3178385794,0.1631723195,0.4340861142,0.0132601513,0.0013953329,0.2829301059,-0.0448556654,-0.2597806156,-0.364072144,0.2854064405,0.2543532252,0.1260091364,0.064809911,-0.0410133079,-0.1498716325,-0.0042460365,0.2463144064,-0.2714927495,-0.2101431638,-0.0912606791,-0.1504568905,0.3735405803,0.0145685906,0.2956715822,-0.3853962719,-0.1448934376,-0.135174349,-0.2745050192,-0.1822858751,0.2835918963,0.1892194748,0.2176278234,-0.0008721882,-0.0889423266,0.0709464997,0.2585667968,0.0867869481,0.337067157,-0.1259837598,0.2799831331,0.0451552346,-0.1590208709,-0.0127900513,0.4234778881,0.0872911289,-0.0733054876,0.1100385636,-0.3415997922,-0.0910098404,0.0555362701,-0.1071021482,0.0195278171,0.0401372798,0.1069045886,0.0455627628,-0.352596879,0.0224057529,0.2243743986,0.2194463462,-0.1672592908,0.1753019691,0.1053165495,-0.2558670938,-0.0744999498,-0.0737636313,0.2589468062,-0.2257397771,0.1010412276,-0.3174892068]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"Okey actually not that easy to add things like `test-de` to `datasets` => this would be the first dataset to have this.\r\nSee: https:\/\/github.com\/huggingface\/datasets\/pull\/802","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":24,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nOkey actually not that easy to add things like `test-de` to `datasets` => this would be the first dataset to have this.\r\nSee: https:\/\/github.com\/huggingface\/datasets\/pull\/802","embeddings":[-0.1846334934,-0.2418597192,-0.1847652048,-0.1303508282,-0.0518171452,0.0069968733,0.2505635023,0.1113591269,0.1802298576,0.1449396908,-0.1167486832,0.1012669802,0.0485733636,0.3905768692,0.6209658384,-0.178280443,0.1267161667,0.0511756316,-0.0147035746,-0.0934559926,-0.1027881503,0.2167291492,0.1080801263,-0.0527150519,-0.1677287966,-0.2435363978,-0.28423962,0.1435454786,-0.0887115151,-0.3312167525,-0.0131604495,0.2882205844,-0.1768379807,0.1054641306,-0.0001009602,-0.1696676761,0.1100735739,-0.0895158276,-0.0204101969,-0.0245752875,-0.2445190102,-0.4427774549,-0.143860206,-0.0647223443,-0.1366013885,-0.0817899257,0.1576744914,-0.511418581,0.0948295146,0.1543592066,0.2733411491,0.2456284314,-0.0281457156,-0.235783875,0.3271295726,0.0656340495,-0.1248783842,0.2170629948,0.7011628747,0.1104023978,0.0426903777,0.2003490627,0.1661870927,-0.0251015443,-0.0942072794,0.0687786862,-0.1166618764,-0.1290900409,-0.0265672635,0.6269078255,0.1902135313,-0.4049749374,-0.4105316699,-0.1999892145,0.0986906812,-0.052923277,0.089781791,0.0726022571,0.1213970259,0.0318280049,0.1945581585,-0.4966183007,-0.0597505048,0.1288611144,0.0036636719,0.2533256412,-0.154023096,-0.0024558634,0.1759861559,0.0424560644,-0.1609525979,0.0287617818,-0.2005242556,0.1064115092,-0.096863158,-0.0444806516,0.0238624662,0.1938484758,0.4159869552,0.0361158326,0.0447767302,0.2233369946,-0.2803298533,0.1791000217,-0.1463030428,0.0760989711,0.2867692411,-0.4129454195,0.3636011779,-0.2374961525,0.2125659436,-0.0065588597,-0.0568500049,0.0141510293,-0.3175007999,-0.008249063,-0.0268741958,-0.050289616,-0.1470522881,0.0037651628,0.04865136,-0.1039883718,-0.0061707301,0.2225236148,0.0996550471,-0.1233040839,-0.1621115953,0.3138157427,-0.0012074781,-0.3761256635,-0.0911941528,0.3496199548,-0.269058466,0.1013754681,-0.0004975646,-0.1953832209,0.1036227047,0.1846352816,0.0611572266,0.0986920446,0.2261886001,0.0720733181,0.1173693091,-0.1074078232,-0.0736710206,0.0463803634,0.0098902006,-0.3629238904,-0.3809714913,-0.2105388939,0.0378625728,0.0682305396,-0.5171687603,0.2253204435,-0.1216332614,-0.3320922852,0.0610728674,0.7353149652,-0.2060721964,-0.1873333007,0.3742213845,0.1848040372,0.0606342368,-0.182090655,0.1231317297,0.3011280596,-0.7084853649,-0.3878898919,0.0161899347,-0.3116719425,0.015024215,-0.007212868,-0.0522164777,-0.1488976926,-0.2193853557,0.3566614985,0.3074155748,-0.2631284893,-0.1048190519,0.1527160853,-0.1363544017,0.0324215107,0.1493159831,0.2461178303,0.0076194662,0.0819620118,-0.228445515,0.4928712845,-0.1376854926,-0.0693279281,-0.0171897598,-0.1178915724,0.1163965687,0.2341865301,-0.017736556,-0.1847666949,0.3748797774,-0.0480547622,0.0143025937,-0.2178456634,0.0823105574,0.2338499725,0.2585249543,0.0419751368,-0.1527703702,-0.3222052455,-0.3203323483,0.117840603,-0.0965881869,0.23034814,0.1352093667,-0.2790805995,-0.0196157284,-0.0212945659,-0.205671683,-0.2783054709,0.1957814693,0.3253358305,0.1568973511,0.3068617284,-0.1636470705,0.2912967801,0.0270691346,0.0393407121,-0.382066071,0.1652247906,0.1009419039,-0.0487701856,-0.2471422702,0.2708751261,-0.2948076129,-0.3454168439,0.0475604273,0.4005276263,-0.0562770739,0.3210214078,0.4340528548,0.3132542372,0.3498570919,-0.0516707823,0.2166532725,-0.1477486938,0.0593076646,-0.0179117396,0.0332002603,0.3806247413,0.2867743075,-0.1806423515,0.0645806566,0.0515145473,0.3325469494,-0.0578965135,-0.0247553792,-0.2705189884,0.3214475513,-0.0682074577,0.462057054,-0.0891146958,-0.4519284666,-0.0041489908,0.5400702357,-0.2069958597,0.5007694364,-0.038568195,-0.2486328334,-0.1093044207,0.2232329845,0.2447239757,0.1789746583,0.2782121897,0.2959163487,-0.0094852336,-0.1759129912,-0.2030114532,0.1238840893,-0.2936118245,-0.2068038434,0.0600175858,0.0018413918,-0.2386073917,-0.6312516928,-0.1454912871,0.0645731837,-0.0627526641,0.0172268339,-0.1164710149,-0.2509187758,-0.3461498022,-0.2239780277,-0.2117842436,-0.4119899273,-0.2713960111,0.3582319021,-0.3860490918,-0.101405561,0.2287199944,0.1626435518,0.3069648445,-0.3231985271,-0.2430468202,-0.3107090294,-0.2273840308,-0.1320769638,0.2030258328,0.0638731644,0.0810738206,0.5975708365,-0.1493800133,0.3434199393,-0.3206497133,-0.6411496401,0.1712259054,-0.2336112112,0.3255996704,0.227258727,-0.1397531629,0.1338934749,-0.0364695564,0.1930484474,-0.373699069,-0.1994298249,-0.1343020648,-0.1690985262,0.0130697396,-0.1824685633,-0.4047191441,-0.4013569653,-0.3103552163,0.3401172757,-0.1355203837,-0.1033565253,0.0624947473,-0.1144509241,0.4457504153,-0.1651321352,0.0383789837,-0.3390419483,-0.3150513172,0.0420127362,-0.3130304515,-0.2071701586,-0.1308929175,-0.0832276568,0.1937596798,-0.104283832,-0.2162323743,-0.2951601148,-0.2080211788,0.304359287,0.2436639667,0.2603614032,0.2636737823,-0.0056840666,-0.1536726058,-0.1816168427,-0.1033957005,0.1106702387,0.1836391836,-0.1427585334,-0.0462077335,0.2588562071,0.0358119942,0.316762507,-0.1313972026,-0.0619766638,0.4659319222,-0.1548800617,0.4611420631,-0.219260931,-0.175665006,0.2025173455,-0.0464686453,0.3559702635,0.166242525,0.059358336,-0.19397524,-0.0081261769,0.0799493119,-0.0653869659,-0.0245025698,0.0498045534,-0.2300821692,-0.0210223068,-0.163060084,-0.3697709441,-0.1072552577,0.0006032686,0.2522925735,0.0814653188,-0.170551911,-0.0207876172,-0.5745233893,0.0311257001,-0.3425473571,0.4614945054,0.2403463721,0.0498220958,-0.3317689896,-0.0788035393,0.0167604741,-0.0146973152,0.3401902616,-0.7034711838,-0.3920474946,0.1245712265,-0.1734478921,-0.1693348736,0.1326391846,-0.4778984487,0.0108445091,0.2867746651,0.1882969439,-0.2953968644,-0.0790121406,0.1735622585,0.244952485,0.007372837,0.1449945867,-0.08718808,-0.2766906619,-0.3222582042,0.1501875967,-0.1172306836,0.0369660668,0.5093621016,-0.134950012,-0.1290971488,0.0091319811,0.1853175312,0.3124790788,-0.2047714293,0.2354229987,-0.1029847413,0.0458191633,0.0343781412,-0.1046586558,0.2496352196,-0.2602446079,-0.1087375134,0.0472425409,-0.2257299572,0.499892056,0.0635030642,0.2434426546,0.3960293233,-0.1816016883,-0.1033774987,-0.4619444907,0.4287548363,0.203068465,0.5127570033,-0.3543935716,-0.2824347317,0.1989815086,0.1494351327,-0.1407776177,0.5283777118,0.2236380428,-0.3452454805,0.0749753565,-0.0176576432,0.8527343869,-0.2429270744,0.1053046137,0.0347235352,-0.1981409639,0.3081974387,-0.3240333498,-0.0729882494,-0.2227023393,-0.3267223239,-0.1645401418,0.0058657336,0.372156173,0.0501755625,-0.3208467364,0.0604899637,0.2180280834,0.1049018428,-0.0844884962,0.3619780242,-0.2068517059,-0.2756367922,-0.1969658583,0.2479059696,0.0991240442,-0.0737509206,-0.0998443663,-0.0280158538,0.0922902152,-0.0161178112,-0.307030797,-0.0946009234,-0.2411384284,0.0730493665,0.0579347648,-0.1569432169,0.0994086936,0.0679132864,0.3513994217,0.1998198926,0.005828565,0.1908431202,-0.1399709433,0.1001738682,0.1871105731,-0.0880351141,0.3409434855,-0.0073426054,-0.2477324605,0.2806872725,0.0567599759,-0.1837937981,-0.2904188931,0.0760108158,-0.221619159,-0.040140368,-0.0003999865,0.2231275737,0.1862411797,-0.384999156,0.1687087417,0.2530966103,-0.2087425888,-0.0268264767,-0.0650622472,-0.2486869544,-0.363185674,-0.0296371114,-0.1757624,0.0355609357,-0.0354002938,0.1423131377,-0.062777698,-0.3454819918,0.2907411456,0.6144684553,-0.3988986909,-0.0256390665,-0.0218600277,-0.525449574,-0.1252254844,0.131118685,0.2892465889,0.065188013,-0.0017302133,-0.1653303355,-0.1337577403,0.0964198634,0.2026550621,0.4214718342,0.1294968724,-0.1154472753,0.0959627852,0.3398857415,-0.4340874255,-0.0085881762,-0.1618078053,0.0188085474,0.1833848506,0.0882824808,0.3379588425,0.0983434021,0.2130499333,-0.1175852045,-0.1208423078,-0.2233335227,-0.2579259276,0.0717110783,-0.2370135039,0.1891643554,0.3106756806,0.0677388012,0.1408637315,-0.2422078699,0.38416484,-0.0748245046,-0.1034875438,0.1999962628,0.08423841,-0.1229715571,0.2088065743,0.0343965851,0.0824556574,0.0929292589,0.1213834062,0.0487042889,0.0149398185,-0.1842785925,0.3117762506,0.4180962443,0.3014640808,-0.1611471623,-0.1428309381,0.1835541576,0.1057185978,0.2218572646,0.3776785731,-0.0637553483,-0.1747801006,0.2375528216,0.2852987051,0.2506440878,-0.2058854103,-0.2343479693,0.436740756,-0.0743765011,0.4397975206,0.3935502172,0.4036575854,0.3437797725,0.4054287076,-0.3494438231,0.0447571576,-0.0479858331,0.2133401185,0.1163633019,-0.1503828615,-0.0357876383,0.2522676885,0.2084745914,0.0992716625,0.2140417248,-0.0287269652,0.3963953555,0.1795141548,0.2958677113,0.2443053573,-0.0552623235,0.192082569,0.0173485875,0.0954189524,0.2742870152,0.1217975095,0.5185917616,-0.2648636401,-0.0779418573,-0.0724580586,0.2538952827,-0.0582274497,-0.2553240657,0.2280032784,-0.083093524,-0.2353249788,-0.0134180039,0.133975476,-0.2005642056,0.1956748962,-0.0318537354,0.3330758512,-0.2071227282,0.269357264,-0.069068104,-0.0569591112,-0.1648180783,0.3427710533,-0.0086142467,-0.0757123455,-0.0116378637,0.3626711965,0.1208774224,-0.2137694359,-0.065699473,0.434623301,-0.355791986,0.067497395,-0.0108750295,0.2708671987,0.0738674328,0.3266423345,0.3112015426,0.1924106181,-0.1710516065,-0.0774526894,0.0605593622,-0.202601105,-0.2428016365,0.112141721,-0.1197674051,0.0963719413,-0.4614645839,0.1329244524,-0.4521275163,-0.0432649702,0.3228740692,-0.2712717056,-0.0590008497,-0.0097448155,0.1391970217,0.0850889608,0.361779958,0.2722505033,0.3546929955,0.104063198,-0.367102623,-0.5847011805,0.0854197145,0.1345130652,-0.2198191285,-0.2222940326,0.0091503654,0.4471949339,0.0087295147,0.0666519627,0.1300784796,0.1190741137,-0.2491855919,-0.3235985637,0.1706228107,0.0405224562,0.1078919843,0.1124057993,0.0585383922,-0.1096152589,0.0941534713,0.0735964924,-0.2597961426,0.0547442064,-0.1511449218,-0.2828803658,0.4270501733,-0.0038169418,0.488386035,-0.3071035743,0.0392458439,-0.1248999611,-0.2984401882,-0.2123712003,0.3169874251,-0.0084684035,0.1640867591,-0.068251811,-0.3227053881,-0.0810526237,0.3033623993,0.1809957922,0.2069409788,-0.1580900699,0.1864404082,-0.0598232858,-0.2287287414,0.1990087181,0.3820851147,0.0345334411,0.0233497173,0.2187623978,-0.2297603786,0.0560141988,0.0270907525,-0.0925478563,0.0313281156,0.0093064699,0.2244930863,-0.0694353655,-0.4996795356,0.0741702095,0.307428658,0.196398735,-0.1526353657,0.297727108,0.0619612113,-0.1063736677,-0.099522464,0.0301187932,0.2739199698,-0.09409637,0.1999083757,-0.3941316009]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"IMO we should have one config per language. That's what we're doing for xnli, xtreme etc.\r\nHaving split names that depend on the language seems wrong. We should try to avoid split names that are not train\/val\/test.\r\nSorry for late response on this one","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":44,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nIMO we should have one config per language. That's what we're doing for xnli, xtreme etc.\r\nHaving split names that depend on the language seems wrong. We should try to avoid split names that are not train\/val\/test.\r\nSorry for late response on this one","embeddings":[-0.1462257057,-0.1761126667,-0.0905722529,0.0123518491,-0.0241257884,-0.1703824699,0.4644282162,0.2293003947,0.0064150211,0.1897848248,-0.0784216374,-0.0123210121,0.063715063,0.3173589706,0.3393554986,-0.147364676,0.064874202,-0.120767951,0.1857912093,-0.1410236955,-0.1970852613,0.164675653,0.141678527,0.2563765943,-0.3061516285,-0.2510169446,-0.1212384701,0.1231266856,-0.0533281155,-0.26364398,-0.1327727735,0.0994950905,-0.1483211666,0.1550781876,-0.0000984329,-0.1763023585,0.1570734084,-0.1662795395,0.3259849548,0.0369416624,-0.3540754318,-0.4925276935,-0.2166533321,-0.2867809832,-0.0726134628,0.0749670267,0.0203127936,-0.4325012267,0.0292520337,0.0513051972,0.2826530933,-0.0485953242,-0.1415353864,-0.0816421062,0.2933580875,-0.2213200927,-0.1784569621,0.002078576,0.5535991788,0.2362468839,-0.1021504775,0.3790180683,0.0996620655,-0.1741277128,-0.0846165493,0.0462375432,-0.0509669743,-0.3432504833,-0.0554944761,0.6024343967,0.0136226807,-0.2653130889,-0.1561050266,-0.2016578913,0.0942969546,-0.2133093178,0.3548102379,0.1433286816,0.1954391599,-0.0333734304,0.313854903,-0.4838635325,0.0441258475,0.0608955212,0.0301384553,0.6080756783,-0.0815607458,0.2101168036,0.0818187073,-0.0044063856,0.055607412,0.018105261,-0.1166885123,-0.1805522591,-0.278193444,0.0348439664,-0.133814171,0.0385017693,0.1513521671,-0.0915089622,-0.1049464494,0.1773038059,0.1033563018,0.1640123576,-0.0740814656,0.0739103556,0.3556385934,-0.1418625563,0.3384209275,-0.3893523216,0.0308436155,0.0682724193,0.0302113201,-0.1210897565,-0.3905890882,-0.0484874696,-0.2725441158,-0.0133919427,-0.0945795998,0.0500221699,-0.08578033,-0.202482149,0.1805745661,0.1018227637,0.0545383655,-0.0015527862,-0.2859128416,0.2533295751,-0.0977325439,-0.3352463841,-0.155273214,0.3696979582,-0.4024463296,0.0887354165,-0.0924175307,0.0822391734,0.0640990734,0.1878533512,-0.0102433059,-0.0425229743,0.2050230652,-0.053944096,-0.0014978915,-0.0343507156,0.0269424915,0.0131021254,-0.0330026187,-0.165032953,-0.4244991839,-0.215992257,-0.2676058412,-0.1515643597,-0.1938266903,0.2565615773,-0.0661879331,-0.1259263158,0.3019111156,0.6030246019,0.0658158287,-0.0055519342,0.2308276743,0.0461863801,0.0359373204,-0.3438422382,-0.0265953317,0.1573953778,-0.560788691,-0.2898280919,0.0050518517,-0.2177070975,0.2166292816,0.1000580415,-0.191809684,-0.2419541031,-0.1731529683,0.4972555935,0.4736790657,-0.240694955,0.0017588803,0.139537856,-0.2263592631,-0.2233425677,0.3514536917,0.1564726084,0.1404679567,-0.0067435224,-0.0991059467,0.3637696803,-0.2389213145,-0.1682415456,-0.179953143,-0.3024235964,0.0140931718,0.2203172147,-0.0066400929,-0.4272894263,0.1903276891,-0.0931604877,0.0925678834,-0.1461269557,0.1340751648,0.151235491,0.2007117122,0.0784001276,-0.1671226025,0.0491228178,-0.2886480093,0.0849829912,-0.1902270615,0.2473544925,0.3146641552,-0.1260802597,0.1242845282,-0.4120953381,-0.307710737,-0.2251723856,0.246767953,0.3807688951,-0.0058764932,0.0702522174,-0.3039516211,0.4216486514,-0.0748284236,0.0373808108,-0.3291956484,-0.006822926,-0.0320982449,-0.1438957751,-0.1064084172,0.2375395447,-0.1356835812,-0.3920332789,0.0525945835,0.4568064809,0.0867202207,0.0595481545,0.2669987082,0.1823760569,0.3151248991,0.0306577925,0.4122351706,-0.3238601089,-0.1963976473,-0.2101339102,0.2165523916,0.4376225471,0.1735754609,-0.0046036956,-0.097738795,0.0390911475,0.272116214,-0.1845210046,-0.1671703011,-0.2693996429,0.1353587359,0.033449512,0.4870150089,0.1121011898,-0.537476778,0.0372514836,0.7857860327,-0.0592366755,0.2184426934,-0.0784685388,0.1288003623,0.0427695811,0.1004865319,0.4957652092,0.1081290841,0.2882475555,0.0216908045,-0.1542692631,-0.2354768068,-0.2799210846,0.0978767425,-0.1289460361,0.0073645664,-0.0057402877,-0.254562676,-0.2525956929,-0.3949924111,-0.1950167716,0.1849539876,-0.1311198175,0.1120041385,-0.1308682561,-0.1418882012,-0.3241763711,-0.3846465647,-0.2120252252,-0.6176679134,-0.264865011,0.2558196783,-0.658339262,-0.1273485273,0.3772856891,0.0395399369,0.0717264414,-0.390676409,-0.3015213311,-0.0437556952,-0.0720791221,-0.4109976292,0.171017468,0.1405231208,0.0250371359,0.4922403395,-0.0793910399,0.1855514944,-0.074500598,-0.5724858046,0.2775664032,-0.4036213756,0.0977482796,0.1708719283,-0.3082919717,0.0878910348,-0.3493945897,0.3390375972,-0.1168491021,0.1387781352,-0.0910428986,0.0008447964,-0.1255896091,-0.0018690528,-0.6168110371,-0.5302234888,-0.3577502966,0.0649172217,-0.2152603716,0.0530571081,0.0140573578,-0.2009966522,0.2050771862,-0.2021887004,-0.0735040605,-0.1877882183,-0.3085239828,0.0390862003,-0.179997474,-0.1100341082,0.0209620614,0.0351652466,0.4360462725,-0.0134434495,-0.1864354312,0.0974280834,-0.1578869224,0.2900023758,0.4185093641,0.23428002,0.1832298338,0.1810414344,-0.1444858015,-0.0840473995,-0.126610592,0.2406166345,0.1379864216,0.2461365461,0.1285828501,0.188356325,-0.0293926969,0.256842047,-0.0068727173,0.105248943,0.4177756906,0.0217579138,0.1973075569,-0.2238929123,-0.2783102095,0.4182807803,-0.0201440472,0.1917502731,0.2493044436,0.0366868973,-0.3029431999,-0.2026093751,0.1793306172,0.0589087456,-0.2773952186,0.1101572439,0.0364853367,0.0889327973,-0.1228726581,-0.2811173499,-0.0412498489,0.0089712692,0.213023901,0.1249706596,-0.3043186665,0.0758680403,-0.5326246023,0.0725217462,-0.2875136435,0.2076584697,0.1831062585,0.1627009511,-0.1951339245,-0.1184329912,-0.0212416705,0.1403120458,0.2180921733,-0.613001883,-0.4700454772,0.2924254239,0.022241829,0.0317458399,0.0676582605,-0.3907157481,-0.174292326,0.6021929979,0.1423159391,-0.3299188912,-0.2401295751,0.2197658271,0.2776794732,-0.0095865298,0.0851046294,-0.1398838311,-0.1039840505,-0.1854957342,0.02123712,-0.1654780805,-0.035362348,0.4166640639,-0.1096868068,0.029072253,-0.3474702537,0.2417775542,0.2491857558,-0.1018166989,-0.0015162987,-0.099885948,-0.0325593986,0.1437212825,-0.1647284478,0.2211935818,-0.3730180264,-0.0413094871,0.1606752574,-0.4128875136,0.4341976047,0.214568451,0.1774011403,0.2229921967,-0.3140594661,0.0970984921,-0.0133316973,0.2715943754,0.1935470998,0.325214833,-0.4524241686,-0.4648650885,0.2338782996,0.274925977,-0.1879258007,0.5551062226,0.2615851164,-0.3070254922,0.0185072031,0.087161459,1.0130937099,-0.06652724,-0.0320996791,0.1240567714,-0.1468687505,0.2870386243,-0.3465816379,0.0044148993,-0.2121769637,-0.1168294698,-0.007645485,0.0637091324,0.4263187051,0.1828870028,-0.2988040447,0.2446153611,0.1877578199,-0.1389592737,-0.3435710967,0.4902936518,-0.212834537,-0.4288300574,-0.0554878712,0.196996823,0.0424996838,-0.1195895672,0.017214952,-0.0604139045,-0.0423927233,0.0951641947,-0.1126551777,-0.0568947569,-0.1916469187,0.1518830806,-0.19913432,-0.2050966918,0.1081305742,-0.1030364186,0.2312948853,0.0377985835,0.17686373,0.2509280145,0.166286841,-0.02907997,0.1491759419,-0.188683033,0.3108358383,-0.0743235797,-0.262655288,0.1429659724,0.0153478133,-0.0575440489,-0.4738851786,0.0995389074,0.0154341459,-0.1869218051,0.0646899864,0.3470923305,0.1042923629,-0.2016884983,0.221280694,0.1342332512,-0.1066631153,0.1801571101,-0.2058048397,-0.2673102617,-0.4408117235,-0.247443229,0.1526275575,0.0741651431,-0.169661507,0.1126614586,-0.0482331626,-0.247978881,0.3831702769,0.4610032439,-0.2728407383,-0.280883342,-0.0260930825,-0.6038403511,-0.0128137534,0.0397433899,0.3965600729,0.2652055621,0.1307495236,-0.0010352515,-0.1320715696,0.1347316802,0.1667663008,0.2131154537,-0.0842902958,0.0817787051,0.148028031,0.1975748986,-0.4271973073,-0.1203593537,-0.1311844289,0.0853154212,0.0615266897,-0.0055331904,0.2368793488,0.2738026679,0.2482172698,-0.2009928077,-0.190261364,-0.1815940738,-0.2109779567,0.0351683497,-0.0612023883,0.0691287443,0.1474352181,-0.1265222281,0.0697391331,-0.0203310661,0.3101062179,-0.0406604782,-0.0212694295,0.2212351263,-0.036828693,-0.2369242609,0.2370628417,0.1390256435,0.0758242011,0.2203001976,0.2095202506,0.1347356886,-0.0029808364,-0.2144669294,0.119520016,0.3670914471,0.0384648629,0.0667798892,-0.3034000993,0.0576504022,0.2281790376,0.2372751087,0.1905229837,0.3459758461,-0.087965332,-0.0574204251,0.3325531781,0.3427843451,-0.152900666,-0.2157159895,0.2266250104,0.1065103859,0.4202777445,0.5838722587,0.4350994229,0.2854172885,0.2471595556,-0.2668680251,0.198841244,-0.1636265218,0.1349105388,0.2016613036,-0.220858261,0.0675410554,0.502920568,0.1542478949,0.0893119946,0.3030202687,0.0422180817,0.353084743,0.4188139141,0.4055214822,0.1972192526,0.0351754129,0.0403027534,0.1591703445,-0.1290499121,0.109258838,0.018138906,0.3010971546,-0.2082658112,-0.2138486058,0.075454697,0.2279537022,-0.1240233555,-0.0842729807,0.0544374138,-0.1027564108,-0.3516123891,0.0516852103,0.0991103575,-0.0377660468,0.2453121841,0.2175559551,0.1678795069,-0.1135488376,0.2204888463,-0.0661543831,0.0834219158,-0.2337682545,0.3854663074,0.0023047265,-0.0819230378,-0.0665482059,0.1008752659,0.0661214218,-0.0585437603,0.0252409354,0.1825351268,0.0323137976,0.0240091216,-0.0823515803,0.1625451893,0.1854986995,0.402844578,0.3717899919,0.1848207861,-0.2366011739,-0.1963731945,0.1103989109,-0.133354947,-0.0876799747,0.1387682855,-0.0674078837,-0.0267455634,-0.3724149466,0.0236279983,-0.6408744454,0.0907235965,0.3472354114,-0.2498900741,-0.0363755114,-0.1081303582,0.1391492784,0.0205908995,0.2904621661,-0.205486089,0.1605260521,-0.1513404101,-0.3798654974,-0.3629666567,0.068690896,-0.0741145983,-0.2374187261,-0.3008517325,0.1499517858,0.2086910307,-0.0430909097,-0.0213474259,0.1685630381,0.1100888699,-0.0755361244,-0.3630594313,0.1835388839,0.0967719853,0.3167694211,0.1844221801,-0.0548717305,0.0051204516,0.139504388,0.0582029819,-0.2079132199,-0.1573339999,0.3688467741,-0.0546832271,0.3100132346,0.1430501044,0.3413970768,-0.195165202,-0.0823991001,-0.2316165417,-0.2218691409,-0.3323300779,0.2775380313,0.1269987822,0.1420237273,-0.0280279238,-0.0776402354,-0.277282536,0.4186033607,0.0982594863,0.3743474185,-0.0265424475,0.3107008636,0.2739525735,-0.0954766497,0.165767312,0.5336943865,0.1344620734,0.0001994505,0.0348067656,-0.2444978356,0.0066154478,0.3034052551,-0.0619243979,0.0352333449,0.1215721145,0.3006125391,0.1463197023,-0.4625615478,-0.0849028751,0.2468048632,0.0522881001,-0.3042321205,0.1125002578,0.220403403,-0.3317742646,0.0073492141,-0.1698811054,0.2085607201,-0.3166735172,0.2862680256,-0.0598953255]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"@lhoestq agreed on having one config per language, but we also need to be able to have different split names and people are going to want to use hyphens, so we should at the very least warn them why it's failing :) E.g. for ANLI with different stages of data (currently using underscores) or https:\/\/www.tau-nlp.org\/commonsenseqa with their train-sanity or dev-sanity splits","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":61,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \n@lhoestq agreed on having one config per language, but we also need to be able to have different split names and people are going to want to use hyphens, so we should at the very least warn them why it's failing :) E.g. for ANLI with different stages of data (currently using underscores) or https:\/\/www.tau-nlp.org\/commonsenseqa with their train-sanity or dev-sanity splits","embeddings":[-0.2000913322,-0.1522804648,-0.1069084406,-0.112353228,-0.0792811289,-0.2000014931,0.5592647791,0.1603520364,0.1297950149,0.1972673535,-0.0472673699,0.0549896657,0.0048433915,0.3074243367,0.300499469,-0.2359765619,0.0617931373,-0.0328123979,0.1070400625,-0.0768814161,-0.144819051,0.1187251955,0.1501040757,0.06426806,-0.2914257348,-0.3140029907,-0.2087100595,0.1164670661,-0.0417363569,-0.2866325676,-0.1831393391,0.2181471586,-0.1481360495,0.0460788421,-0.0000994877,-0.1674608588,0.1176850125,-0.1754718572,0.2268093526,-0.0554833598,-0.4276823997,-0.5850602388,-0.1877811253,-0.1727513075,-0.0764963478,0.0747532323,0.1615252197,-0.3223183751,0.1621859074,-0.016851211,0.2486219853,-0.035681095,0.0158526208,-0.0933979824,0.2921179235,-0.1258252263,-0.1625911295,0.1111362278,0.5344116688,0.0421693847,-0.1167415977,0.2462140024,0.0771799386,-0.1617516875,-0.2693609297,-0.0677690282,-0.069693692,-0.2593415976,-0.0977884457,0.6707048416,0.0018181055,-0.2552489638,-0.2341094613,-0.2518298328,0.1429242492,-0.1198723614,0.2260702699,0.2438061237,0.0440104827,0.00164947,0.3162364066,-0.490437448,-0.0035794664,0.2545977235,0.1498419195,0.4481433332,0.0083515672,0.0914842039,0.1814766526,-0.0291998815,-0.1529129446,0.1295476258,-0.0814182982,-0.1340005696,-0.3194148242,-0.0192977302,-0.0664459243,-0.0460366867,0.1971566081,-0.1193258241,-0.0815122426,0.2506907284,-0.1215141863,0.2826818824,-0.0358217359,0.0253942981,0.2486378103,-0.1062540635,0.353379041,-0.2628124654,0.0464551337,0.0820845738,-0.0214688182,-0.1186307743,-0.4098235965,0.0041323812,-0.2621773779,0.0790798292,-0.1390137225,-0.0721227899,-0.1061048582,-0.0694446415,0.0607370846,0.0999145284,0.1349299848,-0.0380974673,-0.269852519,0.3025579751,-0.10853789,-0.3179698288,-0.0852049142,0.340692848,-0.3904686868,0.2825713456,-0.0011110291,0.1071576998,0.0306002367,0.2661173344,-0.2616833746,-0.0092200674,0.2431921214,-0.025525311,0.1116447374,-0.0860612392,0.0445116684,-0.0805907995,-0.1204369143,-0.3179932535,-0.4474070668,-0.1331004947,-0.1393804103,-0.0474322326,-0.5369630456,0.2319704592,-0.1346972287,-0.2424329072,0.3213674426,0.702580452,0.1054731533,-0.1079343036,0.3025411665,0.0773232505,0.0856088176,-0.283187449,-0.085445933,0.2961887121,-0.583910346,-0.3671292663,-0.0423091203,-0.1428717077,0.1852105558,0.2240878642,-0.0939926431,-0.2408581227,-0.0905022025,0.5788308978,0.2949937582,-0.0800256431,-0.0437667929,0.0483726077,-0.2129266411,-0.2295107394,0.4102588296,0.2193477154,0.1522818506,-0.1028388068,-0.1026312709,0.5329892635,-0.1997340322,-0.19203116,-0.2687574029,-0.3714087307,-0.0669597909,0.2908557355,-0.1108097062,-0.4445999265,0.2499771267,-0.1283310354,0.0342377983,-0.2165666968,0.1971199811,0.1866378486,0.1733614653,0.1198027804,-0.1649485379,-0.1606724113,-0.2717201114,-0.0421764106,-0.1458329409,0.3109076321,0.1103108302,-0.2175168842,0.0918080211,-0.3954267502,-0.2288378477,-0.0556196719,0.2442547977,0.3357694745,-0.0485513173,0.2046450227,-0.2657516003,0.3758607805,0.0656333193,0.0459028296,-0.2742502987,0.1087707132,-0.0235380586,-0.1934509426,-0.041371569,0.4167004228,-0.1104963869,-0.273794353,0.1936376244,0.4694652259,-0.0133421598,0.2094618231,0.2545446455,0.1569453776,0.3164723217,0.0097726109,0.4155246317,-0.3414929509,-0.1877511889,-0.1871416122,0.1694924384,0.3568233848,0.2377646416,-0.0222448669,0.0605877154,0.0200414062,0.2959116697,-0.2497529536,-0.0891408026,-0.3473980129,0.214688614,-0.129070431,0.3781507313,0.0310075227,-0.5096861124,0.1320846528,0.7999991179,-0.039274741,0.5096131563,-0.074428916,-0.0220924933,-0.0267020129,0.092434369,0.3850877285,0.0828881487,0.3525207937,0.0537568927,-0.2341546863,-0.2267881632,-0.365375936,0.1503828019,-0.2305915207,-0.0933856666,0.1149935424,-0.1172457561,-0.2532336712,-0.4707606137,-0.1996025294,0.0749980062,-0.1893981099,0.1460272074,-0.1899925172,-0.1876575351,-0.4007490873,-0.3989224732,-0.3629322052,-0.5685099363,-0.260099262,0.3544929624,-0.5420058966,-0.1989032328,0.4133724272,0.0591998659,0.1492994428,-0.4386818111,-0.2791981697,-0.0675601438,-0.2322134823,-0.3018974662,0.1828038394,0.1581467837,-0.0470945872,0.4090433717,0.0335788988,0.0718934089,-0.0733387247,-0.5325052738,0.2297291905,-0.2416845262,0.0931292176,0.1921675056,-0.2887467444,0.1889669448,-0.4706890881,0.2868530452,-0.0119840391,0.0934711918,-0.0222060978,-0.0688297302,-0.0162252896,-0.1070671231,-0.4447098076,-0.4915530086,-0.4010550976,0.1596658826,-0.3207647502,-0.0119162686,0.0915309861,-0.0613051131,0.1787116677,-0.1809263676,-0.0398546308,-0.2107488513,-0.177683562,0.0296898,-0.2467573285,-0.0975076184,-0.0743637308,-0.0030583297,0.2528924644,0.0278940778,-0.1953131706,0.0266535431,-0.2030615658,0.1726285368,0.3333505094,0.2352800071,0.2445562035,0.1802091599,-0.2039491236,-0.0323086977,-0.1615832895,0.1644463092,0.0807995498,0.2233292311,0.0131587051,0.167428866,0.2213269621,0.2721554637,-0.0015126242,0.0654091761,0.4926971793,0.0336894207,0.2852627039,-0.1993743479,-0.152240023,0.359172374,-0.0646159574,0.1802013516,0.1841147691,0.0539193489,-0.4981171191,-0.127374962,0.094533138,-0.1506454796,-0.3248206079,0.1541626006,-0.040158201,0.1239834949,-0.0951660722,-0.3498984277,0.042620156,-0.0896913335,0.2541393638,0.1558100134,-0.2382898182,-0.040866293,-0.5034109354,-0.065452002,-0.2861918509,0.3564997017,0.2081937492,0.2657046616,-0.3406349719,-0.1893959194,-0.0595183223,0.1341715604,0.1475784928,-0.7052649856,-0.3108758628,0.2482823879,0.255171597,-0.0351362862,0.1081281453,-0.3380486071,-0.0579069592,0.4620375037,0.1651182026,-0.3659995496,-0.1376523674,0.2716058493,0.1894518733,0.0914007798,0.059588857,-0.0273278244,-0.1717239469,-0.1447698921,0.0563178658,-0.2224403769,-0.0263783522,0.4388493001,-0.1601878107,-0.014355557,-0.3033636212,0.2234648615,0.1760692298,-0.1793837994,0.0400891677,-0.0872563645,-0.0311553422,-0.0244033374,-0.0464481786,0.2221960425,-0.3324345946,-0.0447668284,0.1119777784,-0.3413735926,0.491805166,0.1527469605,0.3442548513,0.2925677001,-0.1617774367,0.058671765,-0.0644989014,0.2131449729,0.1382068396,0.3538115323,-0.504909873,-0.325312227,0.2326054871,0.347741276,-0.1057607681,0.4562176168,0.253054291,-0.3633546829,0.281695962,0.0046307687,0.9073777199,-0.1592378467,0.0663195476,-0.013173393,-0.2073615193,0.4077249467,-0.3071686029,0.012664903,-0.231626451,-0.0525071435,-0.0314998478,0.1162346303,0.3724484742,0.2617295086,-0.4031896293,0.128276825,0.1052311435,-0.0297354814,-0.2424730062,0.4963005483,-0.2735046744,-0.4457586408,-0.1557405889,0.1923010051,0.0387297533,-0.2478262186,0.0037269734,-0.1316088289,0.0787885934,0.1053312719,-0.3344317079,-0.0245852768,-0.1835359186,0.0838166624,-0.1833967566,-0.1816042364,0.2296942025,-0.1530426741,0.1499914378,-0.0009480078,0.1856348515,0.1846385747,0.0655901879,-0.0768916756,0.1543825418,-0.159883976,0.4245723188,-0.0769161656,-0.1944670081,0.1506235898,0.1532031,-0.0731140897,-0.5054073334,0.1016770005,0.0147130284,-0.1413274109,0.1182180047,0.292265296,0.0873077437,-0.2045971751,0.2007818371,0.1990916282,-0.0027951021,0.1231182441,-0.1729202121,-0.2294327766,-0.354834348,-0.2141833603,0.0716999844,0.0619225651,-0.151859358,0.0999381468,-0.0621285699,-0.3110248446,0.2243016511,0.3248922229,-0.3135169744,-0.2035947591,-0.0149541553,-0.5611590147,-0.0316385366,0.0489393733,0.2902243137,0.2276911736,0.0714248568,-0.2066080123,0.0279340968,0.2727835774,0.0022749586,0.3183094859,-0.0489145964,0.0769051313,0.0972466171,0.2931251824,-0.4255700409,-0.0918623134,-0.099408105,0.0639564842,-0.0310464092,-0.0595779158,0.2252209783,0.1836123765,0.2673657835,-0.1263179928,-0.112641789,-0.2048867643,-0.2327113897,0.044927638,-0.0455986932,0.2447638661,0.1332669109,-0.1732302457,0.1027107686,-0.0981097519,0.2340750545,-0.0067625279,-0.156259954,0.3492742777,0.0658076927,-0.030431943,0.2229417264,0.073059231,0.1538510174,0.1939463764,0.2320749015,0.148744002,-0.0079509253,-0.0711521879,0.2236993462,0.4041438401,0.2519884109,0.0371975638,-0.2712742686,0.0843023434,-0.0215310752,0.0744254962,0.2611341178,0.2805994749,-0.0766169205,-0.0450055413,0.325258106,0.3334692419,-0.1243688017,-0.163198635,0.3182978928,-0.0139754564,0.4204299152,0.483969152,0.4422833323,0.177260831,0.2632223666,-0.2364410907,0.0219297465,-0.1194806173,0.138064146,0.0977178365,-0.2348605543,-0.0554500297,0.3983246088,0.1041608751,0.1517788619,0.1233953759,0.1996935159,0.225780338,0.4948224127,0.3259187341,0.228032738,0.052729845,0.0314613245,0.1201630905,0.0122143971,0.16796422,0.0551325344,0.5647082329,-0.1505911648,-0.1659447104,0.1796604097,0.2376630008,-0.0135356458,-0.1544373631,0.073702462,-0.0206397139,-0.3714703023,-0.0858862847,0.099519968,-0.1106293201,0.3970496058,0.1965210438,0.2857390642,-0.1417672485,0.1813798398,0.1122639477,0.24675107,-0.2470305562,0.350484103,0.0195937175,-0.1192037836,-0.1085195467,0.0991262272,0.2237402201,-0.1180719733,-0.1607848704,0.20195961,-0.2517102957,-0.0079949638,-0.0457957685,0.2725595534,0.2854666114,0.2857907414,0.3367886245,0.1849639565,-0.2013082057,-0.1097560897,0.2182750106,-0.0986415893,-0.1151624843,0.1749023348,-0.0541750267,-0.1403588802,-0.2893491387,-0.0321437754,-0.5881889462,-0.1756471246,0.3687169552,-0.2487146556,-0.0288488269,-0.1787038296,0.1249965653,0.0950523689,0.328476131,0.053480804,0.2066310644,0.0587490872,-0.3791368902,-0.3151342571,0.0004419042,-0.1313317418,-0.2205827981,-0.3215102553,0.2591942847,0.2777638137,0.1195501313,-0.0105504533,0.3541172743,0.124132596,0.013410328,-0.3721802831,0.2482224256,0.0209269319,0.1585820168,0.1502375603,-0.1453205347,-0.0784850717,0.0602995157,0.0443588831,-0.2087066621,-0.1215166748,0.3126517832,-0.0396512821,0.4404267967,0.1888224781,0.3760468364,-0.2245993316,-0.1751354337,-0.323101759,-0.2360529155,-0.3889352977,0.2374414057,0.1850437373,0.2181522101,-0.117486693,-0.0293772146,-0.118984744,0.3965694308,0.0227113403,0.2944974005,-0.001514019,0.1962413192,0.2674978971,-0.1677458882,0.0743267536,0.5738527775,0.0561533719,0.0317840874,0.0877580643,-0.34475559,-0.0165468194,0.2008775175,-0.0890753567,0.1210014746,0.1768995374,0.2702555656,0.1854370236,-0.4542295039,-0.0464490168,0.2504996061,0.2181609124,-0.228809908,0.2579584718,0.1778253317,-0.244492203,0.0192547422,0.0247735437,0.2521750331,-0.3240350187,0.2067234814,-0.2375120223]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"Yes sure ! Could you open a separate issue for that ?","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":12,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nYes sure ! Could you open a separate issue for that ?","embeddings":[-0.2647427619,-0.0139869424,-0.1921773106,0.0359471515,-0.0434695147,-0.0982092544,0.3603390753,0.0627091974,-0.0184259731,-0.1154731289,-0.0389284641,0.0209468026,-0.08068268,0.2378351092,0.4504718482,-0.0975920483,0.0373907052,-0.0432208963,0.2783952355,-0.099371016,-0.2711967826,0.2142597735,0.1464569122,0.0897132978,-0.1819248199,-0.3932442963,-0.1387724876,0.2345739603,-0.2227648497,-0.2039751261,-0.0870670304,0.1873851269,-0.1738463193,0.0015961245,-0.0000983301,-0.1283178777,0.1833903342,-0.0398954302,0.226495713,0.1820923686,-0.3832429647,-0.4079212844,-0.3313592076,-0.1559972018,-0.1218965054,0.0686075315,-0.0658945963,-0.4664652348,0.0391333997,0.041204989,0.363858521,0.1037029549,0.1544457972,-0.0809745863,0.4940542877,-0.2180846035,-0.1248433143,0.2643489838,0.6296162605,0.1443551928,-0.0279787369,0.346641928,0.0517254323,-0.2266931385,-0.066929549,-0.0756984726,0.0440771133,-0.0943228751,-0.0027972888,0.5223833323,0.2083785534,-0.2428066581,-0.3809204698,-0.1125743017,0.0447276346,-0.0495979525,0.1553055346,0.213593334,0.0252314433,-0.0582756549,0.3420246243,-0.7054735422,-0.0316873826,0.0867259279,-0.0944582373,0.4752363265,-0.1749087125,0.0634466931,0.0298928954,-0.0816632584,-0.0291946866,0.0685072392,-0.2097952366,-0.0673230663,-0.1582744271,0.0078948177,0.1215531528,0.2376289517,0.1963365972,-0.1370898634,0.1802581251,0.2105797231,-0.1265626252,0.2100778818,0.1066820174,-0.0179853849,0.0359322093,-0.2398753464,0.3223237693,-0.3500244021,0.0417179279,0.0329738036,-0.0969661549,-0.1362719387,-0.3655448556,0.0945240036,0.0065116375,-0.0537303835,-0.1887385249,0.1021315008,-0.1061969846,-0.1278383136,0.0022158949,0.1328441352,0.1668203324,-0.0216534249,-0.0929848775,0.1935635209,-0.1013735682,-0.2571038306,-0.0922213718,0.4428953826,-0.2843577266,-0.0087096607,-0.1215822995,0.0566399172,-0.2255034894,0.277752459,-0.0973026678,0.1801305562,0.1993173212,-0.0652268976,0.0591287687,-0.0969118997,-0.081387803,0.0183145776,-0.0137241278,-0.2029397041,-0.3541913033,0.011640152,-0.050062146,0.0446609668,-0.4625561535,0.295707643,-0.0769589841,-0.3422270119,0.2619192302,0.5549826622,-0.0514017195,-0.1621511877,0.3513506949,0.1422885209,0.0377332754,-0.3050616086,-0.1562540233,0.128671512,-0.6564578414,-0.2426055074,0.0382716581,-0.2157424241,0.134234041,0.0552261956,-0.1644967049,-0.2113583833,-0.2053091973,0.4141138196,0.2762158811,-0.2975952923,-0.15790914,0.1058422029,-0.3106619716,0.0574993119,0.2120180279,0.3159177005,0.0528679229,0.0105208335,0.0817761719,0.4460647404,-0.0849808231,-0.1274641752,-0.3038905859,-0.1958521158,0.0674187094,0.153446883,0.0478106365,-0.3661046624,0.2526943684,0.0066063646,0.0063677407,-0.1728488505,0.08926972,0.358271122,0.3219415843,-0.0011018065,-0.1597639322,-0.196077615,-0.1436997652,-0.0463056155,-0.1390229613,0.2675254345,0.2119537294,-0.1923638433,0.0275642909,-0.2283999473,-0.1267476529,-0.1495580822,0.3025671542,0.3021502793,-0.0039269552,0.258497715,-0.2771050334,0.4925562441,-0.0299043022,-0.1175995916,-0.2775684893,0.083403334,-0.0804031342,-0.1337040067,-0.103669703,0.2650892437,-0.2006918043,-0.1945941299,0.116114229,0.2085275054,-0.253665328,0.2737390697,0.2924781442,0.3462963104,0.3957484365,-0.0487123653,0.2588745058,-0.2618387938,0.025994949,-0.0619549155,0.3367300928,0.4578630924,0.3043382466,-0.1681433916,0.1504900903,0.0590393692,0.3933729231,-0.1898751855,-0.0334755853,-0.1738599986,0.3060208857,0.1165392175,0.3803611398,-0.0879270732,-0.3736033142,0.1595176458,0.5955366492,-0.0505071059,0.4780003726,-0.0052015772,-0.2134760022,-0.0220599547,0.1378979534,0.3991895616,0.1495487094,0.3466322124,0.1789194643,-0.088893503,-0.0543040335,-0.2007930279,0.1330865026,-0.1944322884,-0.1086108387,0.1986415535,-0.0175326671,-0.2067572773,-0.6382713914,0.0002516316,0.0827830508,-0.0400782488,-0.081197761,-0.242780447,-0.0609628484,-0.3861940205,-0.1348689646,-0.2713386416,-0.4131394923,-0.1885077059,0.2982266247,-0.4564900994,-0.1141680032,0.2115776241,0.0227513183,0.2943324745,-0.3579414487,-0.079332754,-0.1385708898,-0.2441864163,-0.1943514347,0.3225551844,0.1500778794,0.0714581907,0.5224553943,0.0133263646,0.3942250907,-0.2903516889,-0.5398347378,0.2158973515,-0.1609514654,0.1236621067,0.2172199041,-0.1816847771,-0.0419921875,-0.2021117061,0.2714400291,-0.262989521,-0.1013208032,-0.2663555741,-0.1510933042,-0.0391175225,-0.1000778973,-0.4043988585,-0.4563241005,-0.4155798554,0.2238685787,-0.0921476409,-0.0210107099,0.0343694873,-0.1286590993,0.3319850564,-0.0553996004,-0.0963205621,-0.4068902135,-0.3033789098,-0.0465371422,-0.1489078999,-0.2380395383,-0.0363060981,-0.1591510922,0.2299891561,0.0416196994,-0.1625043303,-0.0799973458,-0.0699970573,0.1694485694,0.287904501,0.1856399477,0.1983991414,-0.0112313358,-0.2875878513,-0.1339599639,-0.0416770019,0.0701544881,0.1296934485,0.006973647,0.0583659336,0.1693323553,0.0338592045,0.2717657089,-0.2620794773,0.0577034429,0.5060428977,0.0381621346,0.3928982615,-0.0506087691,-0.2885011137,0.2060594857,-0.1435455084,0.179176122,0.1482581943,0.1132999063,-0.3378033042,-0.060639713,0.0498514026,-0.1404470801,-0.142699793,0.1719920337,-0.0505423285,0.1096465141,-0.0171030946,-0.5033966899,-0.1553092301,0.0136660971,0.2279440016,0.083078824,-0.0222410653,-0.1959115267,-0.3956011534,0.134155035,-0.4080187678,0.3272369802,0.2141761929,0.0806222633,-0.3271196187,-0.0128654987,0.0456922129,-0.1296967417,0.2366665453,-0.7440313697,-0.4256722629,0.4182428718,-0.0248652883,-0.1652431488,0.2018192112,-0.4702907503,0.0177544076,0.3857242763,0.1370377541,-0.268638283,0.0607724711,0.1033602282,0.0851282105,0.0260390323,0.1092958972,-0.0375114121,-0.2013277113,-0.3085419834,0.156859383,-0.0107471282,-0.0094216149,0.5359956026,-0.1166977659,-0.1306119859,-0.0897638723,0.2905734479,0.0213159919,-0.1299557984,0.1373336017,-0.2213906795,-0.0502160415,0.0604709648,-0.1738540381,0.3664509952,-0.1478516459,0.0183738209,0.1256835908,-0.35794276,0.3236688077,0.1791090667,0.1820282638,0.2951367795,-0.2078621835,-0.1534554362,-0.2343786806,0.1705844104,0.0656810254,0.408649534,-0.381194979,-0.2579296231,0.219300434,0.2772715688,-0.1837475747,0.4523904324,0.2555423379,-0.3076181114,-0.0509312972,0.0558194369,1.0330278873,-0.1289116293,0.0865811184,0.1105158702,-0.2335989028,0.2222301066,-0.3710967302,0.0019269186,-0.2933121026,-0.2001183778,-0.1299994588,0.0252405573,0.3633016348,0.1056003794,-0.3064724207,-0.1127450541,0.0087388372,-0.035265442,-0.1325241923,0.4052647352,-0.0950508863,-0.2374022901,-0.1807707995,0.285929054,0.1412178874,-0.0342609547,-0.2078039348,-0.1313460916,-0.0225606449,-0.0647179261,-0.3884841204,-0.0655406192,-0.2169139236,0.1796509176,-0.1496282071,-0.1979130358,0.0808593258,-0.172009334,0.1848994046,-0.1170371994,0.0280428566,0.2095607817,0.0949864835,-0.0617584512,0.2495844364,-0.050684806,0.351069808,-0.1661036462,-0.1310409606,0.3706371486,-0.0199719723,-0.3196701407,-0.5220580101,0.225324288,-0.0378404334,0.0157431867,-0.1464454085,0.3312796652,0.0697702095,-0.249219507,0.2445730418,0.2363467216,-0.0127484221,0.005425862,-0.0626262575,-0.1233884022,-0.3303997815,-0.0062632915,0.2248469293,0.0292284191,-0.0766762123,0.0810906813,-0.1856088489,-0.3380162716,0.1519450992,0.5602865815,-0.2667227685,-0.1041969582,0.096348621,-0.4846907556,-0.0298193675,0.0484404936,0.2404088974,0.164521113,-0.0475455485,-0.1873999834,-0.0530907139,0.169773519,0.0944154337,0.1669229418,-0.0474651381,-0.110462375,0.0914345384,0.192244634,-0.4879330397,0.0201933496,-0.1726436764,-0.0684843883,0.1564400047,0.0276219342,0.1491114944,0.0521919131,0.2759647369,-0.0797405168,-0.1004062891,-0.2928227186,-0.1073266119,0.0318703651,-0.0953899696,0.1151035205,0.4072877467,0.0127625465,0.0443904772,0.1297347695,0.2823311687,0.0241723228,-0.0984913781,-0.0193569604,-0.128188923,-0.0866441131,0.1371849477,0.1887793392,0.0558508821,0.2510929406,0.0936824307,0.0472313315,0.1170993522,-0.0751739144,0.1892903745,0.2810501158,0.244940877,0.0663313866,-0.1331507415,0.0131087769,0.1328531504,0.3380985558,0.2147030085,-0.0047740322,-0.1635367125,0.1110127941,0.106915012,0.3676954806,-0.3061505258,-0.2016221434,0.3028011918,-0.1867834032,0.3714001775,0.2973410487,0.4434659183,0.1925105304,0.3000945449,-0.2383126318,0.2664511502,-0.098376587,0.2026672512,0.1675505489,-0.2406470627,-0.0670785233,0.316575706,0.2543980777,0.1449220181,0.2586155236,0.0724117681,0.2174856365,0.5248108506,0.3870584369,0.1872797757,-0.0654622391,0.2178006321,0.2417145818,0.1484408677,0.2852225304,0.1362885833,0.3225629628,-0.0732868016,-0.0957979932,0.0021498592,0.361191988,-0.187585786,-0.0834339783,0.2951524556,-0.1643246114,-0.2590714395,-0.0288820714,-0.0335016847,0.0200015064,0.2205688953,-0.0063870396,0.2992775738,-0.0489805266,0.2292802334,-0.1513505429,-0.0416500829,-0.1875191182,0.3817808926,-0.0719969347,0.0830610022,-0.0250509493,0.3998768628,0.2404097766,0.0409953408,-0.1279479712,0.3074738979,-0.2693833411,0.0853135213,-0.1036826298,0.2024284452,0.3084992468,0.3361353874,0.2477551401,0.2231673151,-0.2754342556,0.0787269697,0.1235194728,-0.0638355389,-0.0827496201,-0.0276259761,-0.2714805305,0.0361698866,-0.4454158247,0.0819117874,-0.466108948,-0.0603413731,0.3135890067,-0.268529743,-0.0101583814,-0.0251538325,0.1527639329,-0.0212940406,0.2776848674,0.1215824708,0.2425265759,0.0009030537,-0.3308187127,-0.568457067,0.2323015779,-0.0763855726,-0.2178693265,-0.2524388134,0.1626952887,0.4503721297,0.1099235788,0.0069954609,0.2110242099,-0.1269143373,-0.2184176743,-0.4312402308,0.3489074111,0.2380550206,0.1228104383,0.0626394972,-0.0256107971,-0.0529160947,0.114094615,0.2004552633,-0.2561736107,-0.2614584267,-0.0136345625,-0.1440519094,0.4633581638,-0.0002401396,0.4071262777,-0.2810736001,-0.1339345425,-0.1302715391,-0.3298838139,-0.1636816561,0.3796445727,0.1289864033,0.1515926868,-0.0035783274,-0.1131781265,-0.022495918,0.1821066886,-0.0194283761,0.3837529123,-0.1485262066,0.276864022,0.1654167622,-0.1741964072,0.0130524561,0.3631302416,0.0621650852,-0.0504043326,0.0062767263,-0.3073987067,0.0113785844,0.0719135776,-0.1186818853,0.1092512161,0.1160031557,0.1518441141,0.0413482562,-0.4653996229,-0.0516140647,0.2968170941,0.2025944889,-0.0227437243,0.0869670957,0.1771949679,-0.2494412065,-0.076418072,-0.2795963883,0.1685655862,-0.1784936935,0.1591571122,-0.212465018]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"Really cool dataset \ud83d\udc4d btw. does Transformers support all 11 tasks \ud83e\udd14 would be awesome to have a xglue script (like the \"normal\" glue one)","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":25,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nReally cool dataset \ud83d\udc4d btw. does Transformers support all 11 tasks \ud83e\udd14 would be awesome to have a xglue script (like the \"normal\" glue one)","embeddings":[-0.3609884679,-0.2430483103,-0.1333388388,-0.1087900698,-0.0488198325,-0.1585718393,0.4023332894,0.1078701392,0.1723619401,0.0442129448,-0.0323640816,0.0167611782,-0.1374104619,0.3313932717,0.5176857114,-0.2384225577,0.1370011419,-0.0414835922,-0.2136694044,-0.1239042655,-0.2755305469,0.1441282779,-0.0130416006,0.0020683771,-0.3241946101,-0.3382702172,-0.2491800189,0.0200917218,0.0855365247,-0.0490060486,-0.1051613837,0.2372089475,-0.1000937745,0.1456027627,-0.0001029017,-0.2386933267,0.1524556726,-0.0855204463,0.1233778894,0.1465126723,-0.4335066676,-0.5227155685,-0.284322232,-0.126483649,-0.0534855463,0.1218110025,0.2125723958,-0.1292340308,0.2736060917,0.0016656595,0.283796519,0.2364493906,0.0420415588,-0.1583895683,0.3880887926,0.0325444713,-0.0888498947,-0.1677185297,0.8024855852,0.0244995225,-0.14155595,0.1799804419,-0.1484936625,-0.3167272508,-0.1134844422,-0.1171215773,-0.1555602401,-0.2585976422,-0.0800477117,0.5484515429,0.1892139465,-0.231271863,-0.3534981012,-0.1925369203,-0.1706315428,0.1413885057,-0.1134962738,0.2801836133,0.1149859503,-0.1105148345,0.2131234705,-0.4164280891,-0.0800762028,0.0479947254,0.0213420168,0.5528358221,-0.1332029998,0.080798246,0.2142960131,0.0414795391,0.038148731,-0.0028293354,-0.0836118534,0.118608281,-0.3118964434,-0.1651511639,0.0856062397,0.0546375215,0.1615312397,0.1043573245,0.1731624156,0.3329555094,-0.0175849013,0.247294113,0.0191554707,0.0790556371,-0.0230050124,-0.0928765535,0.2940820158,-0.4682030082,0.1120841727,0.0056076217,-0.130060941,0.0141564878,-0.3059464991,0.1412544101,-0.0782055631,0.0545207597,-0.0553626716,-0.0613579415,0.2052805424,-0.0639327466,-0.0996132717,0.132367,0.1529039443,0.0511071682,-0.1864060313,0.2986691594,0.0478519164,-0.2754711211,-0.0506898053,0.33980304,-0.1761929542,-0.085260354,0.0325038694,0.2464813441,-0.1237427145,0.2510396838,-0.1529009789,0.0991487503,0.4014096856,-0.0238538552,0.1298343241,0.1362227947,-0.1011758074,0.0525082201,-0.050817918,-0.1687996835,-0.366877526,-0.0154537763,0.0796928033,-0.0815573633,-0.3735295534,0.2293701023,-0.2696936429,-0.344999373,0.086140506,0.628996253,-0.0880313143,0.0049777371,0.2904132307,0.105007343,0.2124749422,-0.3345422447,-0.0929958671,0.3001294136,-0.4730134904,-0.3338137567,-0.0290499087,-0.2311615497,-0.0021570779,0.052013617,-0.0440511741,-0.0961744413,-0.0073740389,0.2651539445,0.288978368,-0.264759481,-0.207452938,0.2098602355,-0.193057999,-0.0273487438,0.2175819427,0.4717178345,0.2743147016,-0.126091823,-0.3171695471,0.6667872071,-0.1519976258,0.0343477428,-0.0801531523,-0.2497335672,0.1098253578,0.3647827804,0.2909832597,-0.2167263329,0.1439425796,0.0396679081,-0.043223124,-0.1433188021,0.141574949,0.2905838788,0.2136435956,-0.0768426359,-0.0356039442,-0.3173741698,-0.5034196973,-0.0827657282,-0.1377798319,0.2258930206,0.0800666064,-0.1218504906,0.1599663794,-0.0953992382,-0.1808453798,-0.2863193154,0.2205733657,0.1641145796,0.1397455037,-0.0252317581,-0.2686023414,0.4233843684,0.3195320666,0.0075033177,-0.3495750427,-0.0012606267,0.0281364378,-0.0337243862,-0.0981782749,0.1319551021,-0.0795587972,-0.4043676853,0.1795045882,0.2557623088,-0.1880486906,0.3382915258,0.2561551332,0.6219782233,0.3133595586,-0.01916508,0.3991294205,0.069849506,0.0090057161,-0.1501580775,0.0675852597,0.390750885,0.1409379095,0.0182825942,-0.0826657861,0.2094757855,0.2510527968,-0.1255118102,-0.1356682628,0.0638031587,0.3023630381,0.0520361476,0.3860325515,-0.0337325558,-0.1731179357,0.1216315255,0.6009011865,-0.1979758292,0.5357917547,-0.0756154731,-0.376876384,0.099514477,0.0751873255,0.4574151933,-0.0352169909,0.3183375895,0.1094111428,-0.0901598483,-0.0832100883,-0.1180773377,0.1242982745,-0.1158265397,-0.1036301926,0.0889113769,-0.094376348,-0.2065753192,-0.4662317932,-0.2011834979,0.0638101771,-0.0539061241,0.1487091184,-0.0733557269,0.1418774575,-0.2242625952,-0.3412419558,-0.2206401527,-0.3772154748,-0.2209184915,0.2375549078,-0.3334909976,-0.2636536658,0.2999638617,0.0928904191,0.3302301764,-0.3172257245,-0.258176446,-0.2700491846,-0.2920682728,-0.0477995761,0.2156932801,0.0847446918,-0.0897062272,0.3635438681,0.033396434,0.3812054098,-0.0919457227,-0.5689717531,0.1191645637,-0.1226817593,0.0685385242,0.1305507272,-0.1295525879,0.146707505,-0.2462765574,0.1192120165,-0.2974481583,0.0594247691,-0.0277623404,-0.1027413979,-0.1354483515,-0.1127604693,-0.2742823064,-0.4515147805,-0.399027735,0.4014700949,-0.1385790855,0.0809628367,0.0772029981,-0.0410939939,0.1876616925,-0.0107725887,-0.1963230968,-0.1856986582,-0.1812002808,0.2295367122,-0.1653044373,-0.2328504175,-0.3834370077,-0.3701706827,0.1409291625,0.1719186455,-0.0434372872,0.0923865736,-0.4001675546,-0.0342918821,0.3072669208,0.407863915,0.3317818046,0.0975070447,-0.115450874,-0.1843906492,0.0087157516,0.0510748439,0.2351146638,-0.1191026345,0.0116807008,0.1188951582,0.094940871,0.3151741624,-0.2041479349,0.0975989923,0.3520654142,0.0841611773,0.2209134698,-0.0726856291,-0.2704918087,0.1274237037,-0.1682581604,0.16810669,-0.0417464301,0.0772242397,-0.3364407122,-0.0496734455,-0.015291201,-0.202797085,-0.1033513322,0.3815230429,-0.0514091589,0.1208204255,-0.0068344139,-0.3145824373,-0.2032824159,-0.0024617203,0.317787528,-0.1241957247,-0.0532717146,-0.0749130845,-0.3606041074,0.1025639847,-0.4060278833,0.4150725901,0.0535133965,0.0381959155,-0.1957623065,-0.1038285643,-0.1164113283,-0.0720748305,0.5585858822,-0.7669788003,-0.2477972209,0.2671756446,-0.1752997041,-0.2825009227,0.0456872843,-0.6122175455,-0.1853534132,0.3356102705,0.2638915181,-0.2619715035,0.0222214796,0.135365963,0.2076642364,-0.0366966762,0.1109816283,-0.0739147514,-0.2088598013,-0.3884415627,0.1725718081,0.1043167412,-0.124515295,0.4162594378,-0.0996796861,-0.1448263526,-0.0257691257,0.1814978123,0.0621249191,-0.2512047589,0.0350438543,-0.235184148,-0.0061818529,-0.1770451367,-0.3659862578,0.1819806844,0.0489051379,-0.0261666514,0.150062561,-0.3189902306,0.5837523341,0.0618356392,0.4266435802,0.1943758428,-0.0578186065,0.0082548987,-0.4315266311,0.1220738664,0.0495351441,0.436293304,-0.5069682598,-0.2749755383,0.3288793862,0.2196706831,-0.2500652969,0.3241352737,0.0630880222,-0.1747760922,0.231112048,0.0590977594,1.0470569134,-0.2853870094,0.082585223,-0.0768945664,-0.3744665682,0.3689417541,-0.4711643159,0.1320772618,-0.4165920615,-0.4293307662,-0.1159039065,-0.1274816394,0.188461706,0.2491799444,-0.4533571005,0.1104439124,0.0337638408,0.2077458948,-0.0778501108,0.4188004434,-0.2352682352,-0.3907487988,0.1043958589,0.2456569225,0.0125923902,0.1032886654,-0.2052792907,-0.1338415444,0.1223039404,-0.0382560007,-0.4704657197,0.2153525054,-0.0674795657,0.2509670854,-0.0331185535,-0.1206813753,0.0749347806,-0.2031740695,0.1126596704,0.0963343084,0.0640146062,0.1682970673,-0.0994523987,0.0459487326,0.293320477,-0.0573473237,0.3924855292,-0.0306717847,-0.1983870566,0.1636905074,0.1191077903,-0.1707437038,-0.4024882913,0.0486913063,-0.060189154,0.0203221217,0.0897640511,0.3942229748,-0.0316863731,-0.2778942287,0.1909075677,0.1337256879,-0.0844059587,0.2934395969,-0.1149677262,-0.1069840044,-0.2799264193,-0.1366925985,0.1460370868,-0.0961657166,-0.0794297308,0.3056715727,-0.1349848956,-0.4338507652,0.2840723097,0.6293212175,-0.4850778878,0.0036916081,-0.3216128647,-0.6450653672,-0.1130215973,0.1339280903,0.2503407896,0.121266447,-0.1605266482,-0.1393659413,-0.2133230567,0.2515177131,0.2450024039,0.2780736983,-0.1337166131,-0.1324390322,0.2573319077,0.4048441947,-0.4375584424,-0.082991913,-0.1618523896,0.1344718039,0.3316450715,-0.1460802853,0.1819198281,0.1513404548,0.1655953974,-0.0386900343,-0.1681568474,-0.2187852263,-0.2697502375,0.07200744,-0.0865029693,0.1566186398,0.1791739762,0.037860699,0.0818518326,-0.1696382016,0.2937637866,-0.0619389452,-0.1026089266,0.2647128403,-0.0327847861,-0.0574530475,0.143862471,0.0093399649,0.0467736758,-0.010986479,0.1935272813,0.1348608434,0.3256726265,-0.1500690281,0.0536516793,0.3959354162,0.4930198491,-0.1005025133,-0.0145374453,0.1071007103,0.0501955412,0.2038672268,0.240139991,-0.1264400929,-0.0027318704,0.1049311683,0.1144299805,0.3260230422,-0.2255145013,-0.2132744193,0.295029372,-0.0916679353,0.3021838665,0.4559682608,0.2727414966,0.0748630166,0.4452632666,-0.2198544145,0.439669013,-0.2672578692,0.3010184169,0.1515907794,-0.1996720284,-0.1657268852,0.341111064,0.1333608031,0.0590924919,0.2153359652,0.0762529969,0.0807504803,0.4413011968,0.2158814222,0.3455506563,0.0495134257,0.0585016012,0.1124445349,0.2237920463,0.3284390867,0.1237655133,0.1969251186,-0.1285902262,0.0466523506,-0.0742896497,0.3286129534,-0.2025379092,-0.0439695716,0.4355279505,-0.0704428405,-0.2919859588,0.0813801438,0.1336533576,-0.203766942,0.4764049053,0.0273859184,0.3670365214,0.0936613157,0.3817861378,0.0440279059,0.0127913235,-0.3642971814,0.2106716931,0.0945617557,0.0458102226,-0.1554964334,0.3691646457,0.234443903,-0.0413458422,-0.3102605045,0.5758237243,-0.3054921925,0.0054962682,-0.0814949125,0.2868722081,0.1579225957,0.4094847739,0.1979366839,0.1584171802,-0.1892091185,-0.2466739118,0.0180139933,-0.2632712126,0.0546193197,0.1511254162,-0.1083961427,-0.2024659365,-0.3343569338,0.1445026845,-0.5961225033,-0.2081917077,0.2707670629,-0.3062333465,0.0033914598,-0.0259529129,0.116457589,0.0409544632,0.3355064988,0.1505192667,0.44129017,-0.1046013385,-0.243960008,-0.3743424118,0.0297326501,0.1027981266,-0.1452240795,-0.16904217,0.0720212832,0.3530841172,0.169112429,-0.0904693604,0.2801841199,-0.0896497816,-0.1212618276,-0.2620971501,0.4705163836,0.2503425181,0.2977389991,0.1526476592,-0.0619150661,-0.0834920034,0.1330938935,0.1720967442,-0.3086570203,-0.1317838579,-0.0311154649,-0.258413583,0.3762411475,0.057452675,0.4025652111,-0.2408711761,-0.0790463462,-0.2328089178,-0.2008935213,-0.256952554,0.1051389799,0.132986322,0.1505862176,-0.0286011435,-0.1116519198,-0.1817277074,-0.0419372171,-0.0280584227,0.5094855428,-0.1684677601,0.1310403794,-0.0654511303,-0.0967623442,0.032171011,0.2447135299,0.1230115145,-0.0456186272,0.0157964006,-0.2617710829,0.0546609834,0.1993365437,-0.4142240882,-0.1303807199,0.0549485907,0.0566857979,0.206425041,-0.3133169711,0.0007215161,0.1439418197,0.0925224647,-0.0868832394,0.1879674494,0.2319283932,-0.367206037,-0.0741343126,-0.1359804869,0.1137863398,-0.2101170123,-0.0523656309,-0.3238989711]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"Just to make sure this is what we want here. If we add one config per language, \r\n\r\nthis means that this dataset ends up with well over 100 different configs most of which will have the same `train` split. The train split is always in English. Also, I'm not sure whether it's better for the user to be honest. \r\n\r\nI think it could be quite confusing for the user to have\r\n\r\n```python\r\ntrain_dataset = load_dataset(\"xglue\", \"ner-de\", split=\"train\")\r\n```\r\n\r\nin English even though it's `ner-de`.\r\n\r\nTo be honest, I'd prefer:\r\n\r\n```python\r\ntrain_dataset = load_dataset(\"xglue\", \"ner\", split=\"train\")\r\ntest_dataset_de = load_dataset(\"xglue\", \"ner\", split=\"test-de\")\r\ntest_dataset_fr = load_dataset(\"xglue\", \"ner\", split=\"test-fr\")\r\n```\r\n\r\nhere","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":107,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nJust to make sure this is what we want here. If we add one config per language, \r\n\r\nthis means that this dataset ends up with well over 100 different configs most of which will have the same `train` split. The train split is always in English. Also, I'm not sure whether it's better for the user to be honest. \r\n\r\nI think it could be quite confusing for the user to have\r\n\r\n```python\r\ntrain_dataset = load_dataset(\"xglue\", \"ner-de\", split=\"train\")\r\n```\r\n\r\nin English even though it's `ner-de`.\r\n\r\nTo be honest, I'd prefer:\r\n\r\n```python\r\ntrain_dataset = load_dataset(\"xglue\", \"ner\", split=\"train\")\r\ntest_dataset_de = load_dataset(\"xglue\", \"ner\", split=\"test-de\")\r\ntest_dataset_fr = load_dataset(\"xglue\", \"ner\", split=\"test-fr\")\r\n```\r\n\r\nhere","embeddings":[-0.1910263747,0.05320004,-0.0776923746,0.0108434688,0.038727738,-0.0806754977,0.5347980261,0.0377284102,-0.0204801112,0.14164415,-0.1144325361,0.068052195,-0.18432872,0.2480089068,0.3943488598,-0.1348956078,0.0209186338,-0.1578741223,0.2138730437,-0.1265163422,-0.1170296818,0.1462667137,0.2255790681,0.0706288889,-0.3789371252,-0.2115485519,-0.1750750542,0.0153059065,0.0442247875,-0.2807618976,0.026356373,0.0578123778,-0.048457969,0.1628393084,-0.000103672,0.1411678642,-0.0540858395,-0.0771505013,0.1324319243,-0.1961729825,-0.2736229002,-0.4907403886,-0.0753002167,-0.1089987531,-0.1303656399,-0.0766442493,0.0178500637,-0.4681305885,0.1797653586,0.1816543788,0.2313515693,-0.098832339,-0.4710571766,-0.0546997786,0.154390797,-0.1809719205,-0.0108413668,0.2446625829,0.2024462968,0.1673575342,-0.2104752809,0.4398652017,0.1683361977,-0.2313731462,-0.1968753934,0.1649381816,-0.1484714895,-0.236178726,-0.1522667706,0.567723453,0.0808059573,-0.3228999972,-0.39285478,-0.2704529762,0.1999154389,-0.288887471,0.0518833436,0.211303696,0.2176040113,0.1751316488,0.2317298204,-0.6836288571,0.0540126339,0.1356295347,-0.0971826985,0.513705492,-0.0015301312,0.2439488322,0.0815680474,0.1402274966,0.020834947,-0.007843568,-0.0897864774,-0.1002551019,-0.240918234,0.106680952,-0.1642866284,-0.0150391851,0.0946135595,-0.0935695246,0.0727476105,0.1536908448,0.1089784577,0.2050895989,0.1776739359,0.1180366501,0.2366352975,0.0627688617,0.2887322307,-0.4710458517,-0.077805832,0.019506963,0.0746848956,-0.1912410408,-0.172714293,-0.0459464751,0.026984781,0.0663522854,-0.0959774479,0.1052939743,-0.2061987072,-0.2935826778,0.0990042239,0.144900769,0.1277307272,0.1413604468,-0.2030745447,0.3725821972,-0.2662522793,-0.2571800649,-0.2277455777,0.3409709036,-0.3748532534,0.0921020061,-0.202175647,-0.0808232799,-0.0476233251,0.3292251229,-0.1071465835,-0.1580495834,0.0915870816,-0.1337833554,0.0519525707,-0.1247290224,0.0125548588,0.158018738,0.0570185855,-0.3566270173,-0.4626468122,-0.0300364345,-0.2527826428,-0.2549269497,-0.2875261307,0.2056722492,-0.1694247574,-0.2321385145,0.2545635998,0.7391151786,0.1273803115,-0.1175596789,0.2827657163,-0.1243179813,-0.1663301587,-0.2764618695,-0.0702709928,0.2362784445,-0.5743151903,-0.4132932127,0.2253564447,-0.1976189017,0.2069160342,0.1966434866,-0.3473395705,0.0109948069,-0.1177361086,0.353420794,0.5970428586,-0.292347014,-0.1303872913,0.2141420841,-0.2635833323,0.0889391527,0.2357155979,0.2031340897,0.102727823,0.1982204914,0.0678457916,0.4630224407,-0.0053035812,-0.1573871821,-0.1402194053,-0.1902844608,0.3542566597,0.2323129326,-0.0381515399,-0.2863727212,0.1947164685,0.0899080932,0.2390928119,-0.14268291,0.0964538679,0.2986013293,0.0468662828,0.0179023575,-0.1953507513,-0.070842877,-0.3336547315,0.1470629275,0.03686722,0.3827440143,0.2917511761,-0.266276747,0.1779703349,-0.3421458006,-0.4258976281,-0.3459898531,0.1614493877,0.2230120301,-0.06804052,0.2329872996,-0.2800186276,0.5938700438,0.0390464291,0.1116575375,-0.3817915618,0.0530276559,0.0686305612,-0.1893719882,-0.2376475036,0.2393784821,-0.1718095094,-0.240515247,-0.0398931168,0.4583503008,0.2323985696,0.0488370694,-0.0718663484,0.1190843657,0.3836656511,0.1094573438,0.181332618,-0.2294235229,-0.1822823733,-0.1237252206,0.0827196315,0.5217264891,0.3244667053,0.1004315913,-0.1075853109,0.0357733592,0.1910133958,-0.0415985622,-0.3569164872,-0.2650184631,0.0556679294,0.0556831546,0.4347801805,0.1592078805,-0.5791935325,-0.071252428,0.7180764675,0.0097676571,0.169363454,-0.0693663284,0.0558392256,0.1011972576,0.1895572692,0.520197928,0.212292105,0.2389431447,0.2982468307,-0.0165144466,-0.1940762103,-0.1828314662,-0.105938226,-0.237099722,0.079501465,-0.0362919569,-0.2030076832,-0.2575028539,-0.408272028,-0.1509490013,0.1471932828,-0.0292324778,-0.0294269305,0.0245032161,-0.3005194366,-0.180154115,-0.5986668468,-0.301335603,-0.6115759015,-0.1811361611,0.2447134852,-0.6047827005,-0.0467013642,0.4229427278,-0.1425320357,-0.0406717062,-0.4698758125,-0.5025680065,0.2565771937,-0.042105563,-0.4064779878,0.0918783247,0.1579806954,-0.0327246822,0.2512928843,-0.1421437413,-0.0645048916,0.1113217846,-0.4008816779,0.1825397164,-0.206215471,0.1482295245,0.1005489975,-0.3347789347,-0.0572909713,-0.3021550775,0.3210506439,0.0150025142,0.145675987,-0.0646712929,0.0767859817,0.0475415513,-0.065370813,-0.720400095,-0.5405853987,-0.1982033402,0.0103970701,-0.1640195251,0.1104089469,0.2534588873,0.0890554488,0.0119740162,-0.1456250101,-0.004497014,-0.2257570624,-0.382006228,-0.0291856434,-0.109988898,-0.1568063796,-0.0304675456,-0.0716736317,0.2077922523,0.1767010987,-0.1378644258,0.0166729614,-0.19390513,0.2275470346,0.0353257358,0.4253292978,0.1598967463,0.154330641,-0.0326043293,-0.0435756557,-0.1585964561,0.3780442774,0.2033304721,0.0127778435,-0.0691448152,0.0634470582,0.0594621636,0.2326344848,0.1773548722,0.0641298592,0.2799993455,-0.3380756676,0.3977771997,-0.1409880519,-0.2067123204,0.218290031,-0.0323910825,0.0140596228,0.1528440863,0.0271884687,-0.3039860725,-0.0851356983,0.0778562278,-0.0166709181,-0.3773218691,0.3567853272,-0.153914839,0.0320723541,-0.1507051289,-0.1522399038,0.102948159,0.1817676872,0.0295737479,0.0969726667,-0.3498258293,0.183231771,-0.5369267464,-0.099141866,-0.2999083996,0.107970953,0.1305076778,0.2762563825,-0.1866474003,0.002990575,0.017507067,0.0150314895,0.3966665268,-0.4808499217,-0.4820994437,0.2004496604,0.0212572627,-0.1305934042,-0.0855679214,-0.3278746307,-0.0188421216,0.3498181999,0.0562957413,-0.3079891801,-0.15863581,0.0676218793,0.3820748925,-0.2028707564,0.0607276,-0.2034718692,-0.1728468835,-0.0992191285,0.197016269,0.0209773909,0.1204798892,0.2889884114,-0.122318536,0.0317437984,-0.0846944675,0.1988639832,0.1464780718,0.0917129219,0.1164064705,-0.1610087305,0.050102599,0.0676592737,-0.148519367,0.4402639568,-0.3087310493,-0.2286410779,0.078777723,-0.0829526111,0.4901841283,0.263143003,0.1338195354,0.0982677862,0.0177663378,-0.085103184,-0.0298895277,0.3046809435,0.2123527825,0.3412998915,-0.5329069495,-0.7967947721,0.2185469419,0.382243216,-0.0536779016,0.3410787284,0.1901209205,-0.4465326667,0.2073233724,-0.0830268636,0.9011126161,-0.1306343526,0.2235770822,0.1495044976,-0.1318377256,0.3479498327,-0.3285906017,0.0038764626,-0.2408125997,-0.0437447689,-0.0364985541,0.0516547151,0.4080277383,0.1710567325,-0.3253445029,0.219081223,-0.0632047579,-0.024665365,-0.2653277218,0.538980484,-0.3405443132,-0.3104391694,0.0117441993,0.1211121976,0.0164596755,-0.1240203157,-0.0320049971,-0.0346303955,-0.102709122,-0.0499443673,-0.2220515311,-0.0139930965,-0.3344768882,0.2656894326,-0.2952314615,-0.3578465879,0.0453363582,0.2026152611,-0.0023020764,-0.0105984192,0.265540421,0.0741099417,0.1182727143,0.0995843112,0.2259416133,-0.2287583351,0.4069571197,0.0496987812,-0.4066694081,0.1384824067,-0.0037560533,-0.0418353193,-0.2877307534,0.1658207476,-0.0140044428,-0.3451161087,-0.0867623165,0.5174047947,0.1695557833,-0.1511254907,0.1396108568,0.2260053158,-0.1892263442,0.1703297943,-0.1977103502,-0.3133745492,-0.3624938428,-0.1934811473,0.3652395904,0.0774339885,0.0183695909,-0.0373930149,-0.0392577387,-0.1547609568,0.3754968941,0.6007117033,-0.2000176758,-0.2600066662,0.0255490746,-0.1986386031,-0.1310257316,-0.0964418948,0.0995766595,0.2887752056,-0.0135679711,-0.136773631,-0.0293975752,0.18599841,0.2647145391,0.2356603146,0.0603537709,0.1310637146,0.0571040809,0.1626475304,-0.3480123281,-0.1127506793,0.0621661097,0.0632514656,-0.0794918537,0.1583418846,0.1201239526,0.0510213003,0.2173428088,-0.2318613082,-0.0895734653,-0.1527539641,-0.2958568037,0.0701785237,0.031105699,0.2595134676,0.2123280913,-0.0669621676,0.2546616495,-0.0671432465,0.4487274587,0.3342232108,-0.0203397069,0.3825972974,0.0965799317,0.0775317252,0.0216036271,0.0777426958,0.0062760166,0.0919873565,0.3465153873,0.213204056,-0.0094083818,-0.068460539,0.1386409253,0.1333218217,0.0467717797,0.1533385217,-0.561603725,-0.1983240843,0.2329377383,0.1103418022,0.307398051,0.519389987,-0.2277439982,0.0489994399,0.2958116233,0.2463622987,-0.1652408093,-0.1334269196,0.3222333491,-0.1974413544,0.3228979707,0.6079252958,0.3970000148,0.3502963185,0.178486526,-0.388710618,-0.0774658546,-0.0573844351,0.376455605,0.3210880458,-0.2772592306,-0.0390720591,0.3862944245,0.2131917626,0.0836509988,0.1898492575,-0.203017503,0.2447144836,0.4480059743,0.2559129298,-0.0196452588,-0.1803528517,0.1119464263,0.122697629,-0.0923089758,-0.0242624786,-0.0230912175,0.38338539,-0.1555438042,-0.1396637112,0.298561126,0.2837642729,-0.1940131932,-0.1535001248,0.1968219876,0.0511735566,-0.2921155691,0.1323249787,-0.0847921297,-0.382085979,0.101246424,0.2282993346,0.1256137937,-0.0172211584,0.1757413596,-0.0920663849,-0.0584927313,-0.1508609205,0.3755715191,-0.0813728198,-0.0365996435,-0.0309992097,0.0649983063,0.1281387806,-0.0776271001,0.1141756624,0.2989533246,0.0123910718,0.1952902675,0.0215530973,0.0813497528,0.2636418939,0.278467387,0.3846481442,0.1433396637,-0.1800047904,-0.1037963554,0.1969032735,-0.1185327619,-0.2046389729,0.4069387019,0.0168795921,0.148429513,-0.3007180393,0.1054492965,-0.5907669663,-0.1081681326,0.4824235737,-0.1324926317,-0.0243054274,-0.1197230294,0.1084380671,0.1769620776,0.1353194267,-0.1425591707,0.2858889997,-0.163303718,-0.0971363634,-0.5655861497,0.0163472835,-0.2357181162,-0.2222581506,-0.3270236254,0.1791228354,0.254714191,0.1469032615,-0.0250163134,0.3019028306,0.0407265127,-0.022203574,-0.2168000042,0.2168511748,0.0339496806,0.375287652,0.0588303395,-0.0458982214,-0.0710566267,0.1366904825,-0.0337916277,-0.1267012507,-0.0554821491,0.1835243106,-0.1502084583,0.4845892191,0.0721929371,0.3468746841,-0.0968670323,0.0449644737,-0.3203819692,-0.1027017832,-0.4841050804,0.3530173004,0.070481047,0.179307282,-0.1841542721,-0.1012338102,-0.268538028,0.3025504053,0.1291616261,0.0757252872,0.1334379762,0.2273640186,0.3005334735,-0.1105148643,0.0799842924,0.6495319605,0.1090995446,0.0263891872,-0.0370237902,-0.2057641149,0.0126171978,0.1199869961,-0.1545431614,0.1037767008,0.1777692884,0.3171560168,0.1134345829,-0.5646631718,0.0775776207,0.3290458322,-0.0124461427,-0.1623531431,0.1617171466,0.1916934699,-0.1885296404,-0.0747933984,-0.0379169583,0.1036526635,-0.2620896101,0.1151695028,-0.3432026505]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"Oh yes right I didn't notice the train set was always in english sorry.\r\nMoreover it seems that the way this dataset is used is to pick a pretrained multilingual model, fine-tune it on the english train set and then evaluate on each test set (one per language).\r\nSo to better fit the usual usage of this dataset, I agree that it's better to have one test split per language. \r\n\r\nSomething like your latest example patrick is fine imo :\r\n```python\r\ntrain_dataset = load_dataset(\"xglue\", \"ner\", split=\"train\")\r\ntest_dataset_de = load_dataset(\"xglue\", \"ner\", split=\"test.de\")\r\n```\r\n\r\nI just replace test-de with test.de since `-` is not allowed for split names (it has to follow the `\\w+` regex), and usually we specify the language after a point. ","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":122,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nOh yes right I didn't notice the train set was always in english sorry.\r\nMoreover it seems that the way this dataset is used is to pick a pretrained multilingual model, fine-tune it on the english train set and then evaluate on each test set (one per language).\r\nSo to better fit the usual usage of this dataset, I agree that it's better to have one test split per language. \r\n\r\nSomething like your latest example patrick is fine imo :\r\n```python\r\ntrain_dataset = load_dataset(\"xglue\", \"ner\", split=\"train\")\r\ntest_dataset_de = load_dataset(\"xglue\", \"ner\", split=\"test.de\")\r\n```\r\n\r\nI just replace test-de with test.de since `-` is not allowed for split names (it has to follow the `\\w+` regex), and usually we specify the language after a point. ","embeddings":[-0.1850294322,-0.013266677,-0.1486910731,0.0019631763,0.0076410761,0.0446558781,0.5710830688,0.2278438359,0.0795819983,-0.0067034229,-0.0968189761,0.0388207212,-0.1122408062,0.2615087628,0.3110351861,-0.238413766,-0.0684592649,-0.0565335415,0.1240939349,-0.1399791688,-0.204636693,0.1350339055,-0.0491297767,0.1263543814,-0.2888018489,-0.2710744441,-0.2809756994,0.1510722935,0.0982090011,-0.3324356377,-0.0145562431,0.0820208639,-0.1000953764,0.2394534498,-0.0000984545,0.0035837635,0.0116474032,-0.1343545318,0.093176946,-0.2468428165,-0.1667533666,-0.3409177065,-0.0839922577,-0.2314071357,-0.1988062412,-0.2100844234,-0.0071408399,-0.4129564464,0.2574446201,0.4221245944,0.2658775449,0.0925966501,-0.4435324073,-0.0577486791,0.0836897567,-0.2422899902,-0.014265514,0.3178781271,0.3424064517,0.0916497484,-0.2131611705,0.4722446501,0.2147878855,-0.2140097916,-0.374986887,0.2226093858,0.1621941179,-0.2566306591,-0.244612053,0.5190034509,0.0885888115,-0.1996836215,-0.4328460395,-0.3251274526,0.1437958181,-0.2628550231,-0.0695554391,0.1311028749,0.1654103547,0.1995814741,0.2035406828,-0.464897722,-0.0041480963,-0.027546702,-0.0512328148,0.4490268528,0.0837484375,0.233571291,0.0661156476,0.1411518157,0.07010144,-0.0458780155,-0.0986158177,0.0553029887,-0.307097286,0.0117034521,-0.1072815955,0.0474770516,0.0492759831,-0.0513091646,0.057756979,0.1498975605,-0.0426060557,0.0831398293,0.1448103786,0.198608771,0.26109308,-0.0025003587,0.2398752123,-0.2484660596,-0.1595713496,0.0447020717,-0.0625420436,-0.2378837317,-0.0981818214,0.1467842311,0.0020360735,-0.0192321278,-0.2591941953,0.021880392,-0.2672099471,-0.256757468,0.1375149637,0.3601800799,0.1010513231,0.1783555597,-0.1949617118,0.4000519812,-0.2575907409,-0.3261202574,-0.2360519767,0.4139589071,-0.3500255346,-0.1115129665,-0.1663131416,-0.0035025266,0.0561321378,0.2235698998,-0.069418028,-0.2711364329,0.2184349746,-0.1956034452,0.0430458859,-0.1566675305,0.1363712996,0.1480958611,0.0551049337,-0.2610817254,-0.2872411013,0.124961853,-0.2613191605,-0.2190708071,-0.1309768856,0.2636016309,-0.0556391291,-0.202593416,0.1537780315,0.6856271625,0.0679442063,-0.0829038024,0.16224733,-0.1698952615,-0.2368926406,-0.1498965323,0.0604185946,0.2283627987,-0.4998071492,-0.2230638266,0.1563819945,-0.1898705363,0.2398961186,0.1788352877,-0.3093422651,0.1431647539,-0.1819398701,0.3296810985,0.4337345064,-0.1968709677,0.0012615731,0.1648803949,-0.276986599,0.1439598352,0.0981913656,0.0963478014,-0.0398086384,0.0914687663,-0.1129784361,0.4214546084,-0.1313754618,-0.0832726508,-0.204592526,0.0518585443,0.4903110862,0.1938472092,0.0115592694,-0.2899575531,0.0232222509,0.2018413395,0.4168175757,-0.2848474383,-0.0102514243,0.1488902867,0.0356930196,-0.1913880706,-0.1631722152,-0.0502905585,-0.2708038986,0.2367835939,-0.0572082326,0.3583614826,0.2925870121,-0.1614951044,-0.0340897031,-0.2929207683,-0.3429955244,-0.3280101418,0.1961519569,0.1275319457,-0.0183659736,0.2233020663,-0.1144111753,0.4993126988,-0.2044032067,0.0379449353,-0.2293879986,0.0690778941,0.0574301146,-0.0924549028,-0.1822476238,0.2621295452,-0.0277300626,-0.2642833292,-0.0730083734,0.459644109,0.2617920339,0.0888954103,0.0312576741,-0.0496136919,0.2933147848,0.0038492025,0.080895789,-0.0492724217,-0.057169091,-0.0665819049,0.0750342757,0.6284967661,0.2595184445,0.0366196297,0.0453434885,0.021935882,0.1795022041,-0.1326722354,-0.158524394,-0.1459661126,0.1779647768,-0.0484450161,0.3755075336,0.0467371568,-0.5525548458,-0.0395340882,0.6682415605,-0.0861353725,0.221055612,-0.1629425436,0.0032062849,0.0128655219,0.0814107731,0.4090770483,0.19759655,0.1926131397,0.323312223,0.0363064818,-0.2348086983,-0.185053423,-0.0121339308,-0.1885347366,-0.0220245309,-0.0073001985,-0.0227289777,-0.3276414871,-0.3438843489,-0.0637880713,0.2475376576,0.0396653861,-0.1704257578,0.0497674569,-0.1639364511,-0.0924926549,-0.5131726265,-0.1697528064,-0.2990209162,-0.2755838633,0.2779988647,-0.4795269966,-0.1530558169,0.351680249,0.0142616201,-0.0065271133,-0.2841562629,-0.2699314058,0.2208058387,-0.0940740034,-0.3664023578,0.2050338387,0.2037620097,0.0812525451,0.3301102519,-0.2344191074,-0.169103384,0.0525085628,-0.3607347608,0.1120211631,-0.3015360236,0.2112489045,-0.0038545432,-0.2363384366,-0.0772204921,-0.0966504291,0.291098088,-0.2559049726,0.0668354556,0.0072893901,0.1401249468,0.0815488771,-0.1243102178,-0.8642572761,-0.4875789285,-0.1643183678,-0.0253249481,-0.1726877987,0.1272609383,0.1203700677,-0.0235025231,0.0834427997,-0.1305628568,-0.1321270168,-0.1328826249,-0.3648807704,0.0541831702,-0.1330412924,-0.2362875789,-0.0005247894,-0.1073926911,0.191878587,0.010263131,-0.0284333844,-0.1482353508,-0.0269304644,0.168405965,0.0729791,0.2140501887,0.1686771512,0.0360341519,-0.0599837862,-0.1326552778,0.1027214453,0.3575430512,0.2007726729,0.0179576743,-0.0859696493,0.1646140665,0.114430584,0.316139549,0.1050371155,0.0751701221,0.3473781347,-0.2949373424,0.0827238411,-0.0996648967,-0.2263886631,0.1534629166,-0.0825012252,-0.0072273514,0.2077981234,-0.0784923434,-0.27692312,-0.0320977606,0.1027442142,-0.1094154119,-0.3094548881,0.3692904711,-0.2520239353,0.1452081054,-0.0956943184,-0.0400843658,0.087116085,0.1184847951,0.070456095,-0.1126701981,-0.153316319,0.0615469366,-0.5243782997,0.0631783456,-0.4058930278,0.2378018498,0.2078359425,0.2764014602,-0.0740614086,-0.0548351072,0.0277633537,-0.0255988259,0.4136492014,-0.5377339721,-0.3000136018,0.1335586905,-0.1395608038,-0.1210924461,-0.0396731943,-0.4901552498,0.0567325726,0.3245157599,0.0268932451,-0.2364980727,-0.2350190282,0.0957181603,0.4035211802,-0.2324522138,0.1144091189,-0.2139937282,-0.2013907433,-0.0795557722,0.3560833037,0.1235738099,0.1735672951,0.0661977977,-0.0526301749,0.0362093709,0.0801758468,0.2874006927,0.2367445081,0.1805154383,0.0960932598,-0.01521949,0.0856458619,0.1028416753,-0.1949971765,0.5514800549,-0.279573977,-0.1800486594,0.0751633048,-0.2313220054,0.4920178056,0.3158755004,0.0637909025,0.0542763025,-0.14702259,-0.0298241675,-0.1273115724,0.4952500761,0.3691213429,0.3038398623,-0.5208013654,-0.7163482904,0.0763604939,0.1433677971,-0.0718169957,0.3379656672,0.0801023915,-0.4435579777,0.3029436767,-0.1379787177,0.829978168,-0.1547329426,0.1495676488,0.2511034906,-0.0103379814,0.1198968664,-0.2770326734,0.0504343212,-0.3902565837,-0.1816767454,-0.018542055,0.0266717002,0.3309811652,0.2546290457,-0.1719591171,0.3106066883,0.0629124939,-0.0823835954,-0.1602986008,0.5479097366,-0.1836716086,-0.1779917777,-0.0554749891,0.2073782682,0.1209597066,-0.2522977293,-0.0005774759,0.0440300517,-0.1167805642,-0.1211788729,-0.149024725,0.0681142807,-0.3145233095,0.1270879656,-0.3392326236,-0.4663173854,0.1525374651,0.2302163988,0.0238432661,0.0414802097,0.2004444152,0.2233334929,-0.0361074433,0.2259847075,0.286973983,-0.2161968797,0.4443508387,0.0769711062,-0.4147386253,0.1529234797,-0.0429296345,-0.155361712,-0.3987050653,0.1413903534,-0.0879075676,-0.3317220211,-0.2798876464,0.3710279465,0.1566068232,-0.3257326484,0.1937778592,0.1746720076,-0.356644094,0.1035791859,-0.2939981222,-0.340898037,-0.2279255688,-0.1467459649,0.2551461756,0.1478311121,0.1853164583,0.1771680564,0.0631781593,-0.2510706484,0.2935941219,0.7671792507,-0.360485822,-0.0944762751,-0.0430952199,-0.0761963502,-0.1275304556,0.0466260612,0.0012513705,0.292834878,-0.0060276873,-0.1465603113,0.0262541734,0.1060533524,0.3661714196,0.213556394,0.1051953658,0.2165450156,-0.0657735467,0.1873431206,-0.4249196053,-0.0218880437,0.033500392,0.1952533126,-0.1877145022,0.101508908,0.2462445647,0.0794786364,0.2716882825,-0.0982098579,-0.0227529276,-0.2768943012,-0.2326977551,0.0420945212,0.0030401289,0.146489948,0.1917452514,-0.1105339527,0.1963160932,-0.1995993555,0.3462591767,0.093392849,-0.0289440602,0.4050987661,0.1337253004,0.0200064182,-0.0226388704,-0.0154917976,-0.0756204873,0.0577673428,0.2735755742,0.2285783142,0.1661154479,-0.033849854,0.1037751138,0.0515045375,0.102488704,0.1802722365,-0.4070773125,-0.1074664816,0.2417317331,0.0146529665,0.4209896922,0.455082655,-0.2232889831,0.1202175468,0.1841107905,0.3284201026,-0.3032054603,-0.1017670482,0.3478955925,-0.1835583895,0.3683897257,0.5443088412,0.3609474301,0.3915966451,0.2173250914,-0.332057327,-0.0590574071,-0.0370326526,0.228353858,0.3521629572,-0.22525841,-0.0137925576,0.2947935462,0.0122066643,0.0204939488,0.3185677826,-0.1346903443,0.5100477338,0.2785696387,0.2657672167,-0.1102461889,-0.0381643511,-0.0429405458,0.0359503664,-0.0425849557,0.0030518244,-0.0134197166,0.3571196795,-0.2191537768,-0.2307671309,0.0600328818,0.1721811444,-0.1735138148,-0.1581004411,0.0456712767,-0.0769228712,-0.3853396177,0.1428916007,-0.0948412344,-0.2333560586,0.0164350793,0.1870516539,0.0885120854,-0.1807622761,0.0049232608,-0.0268227849,0.00056922,-0.263656348,0.2993964553,0.1494666189,0.0189232863,-0.0019945907,0.0059083225,0.0925205275,-0.1938612461,-0.1582998931,0.1911499202,-0.0999314263,0.126741305,0.0645078644,0.1244990751,0.1079674736,0.3817975819,0.4639883935,0.2002543211,-0.2372895628,0.0335981473,0.3672507703,-0.1782061905,-0.3492485881,0.3454866111,0.0347282737,0.1469871104,-0.4240697026,0.0471312888,-0.4729422629,0.0539718121,0.3605460227,-0.2008821368,0.008877662,-0.0449940078,0.1398203224,0.0305372402,0.1862389743,0.0164784379,0.4034821391,-0.3780132234,-0.2280184329,-0.6396083236,0.1061454788,-0.1306387633,-0.1318805367,-0.2288672179,0.1273167133,0.2996813655,0.076071687,0.0178166404,0.2471982837,-0.0661533475,0.0361259058,-0.3035609722,0.1115652993,-0.1736971587,0.2723175287,0.0318913907,-0.091309607,-0.0970785692,0.1003885418,0.011223725,-0.0569188558,-0.1481508613,0.2393166125,-0.0982648134,0.4576745629,0.0818586797,0.4475521445,-0.1374200583,0.1620275825,-0.2884498537,-0.1693184525,-0.3598200083,0.3531141579,0.0327526964,0.387263149,0.0410273001,-0.1317428797,-0.218370676,0.2081664205,0.2204309106,0.0191570949,-0.0410343073,0.2259750068,0.2056661397,-0.1205068007,0.1088378131,0.6713241935,0.0820435882,-0.0427994244,0.019226132,-0.1867138296,-0.0438632183,-0.0447645225,-0.2145180851,0.061785277,0.1554207355,0.2474339902,0.0241124984,-0.555437088,-0.0113707315,0.2986544073,0.0716427788,-0.1736570895,0.1647885144,0.3442055285,-0.0516922548,-0.0447650664,0.0026242773,0.1113646328,-0.2034366578,0.1111545563,-0.3900694549]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/749","title":"[XGLUE] Adding new dataset","comments":"Closing since XGLUE has been added in #802 , thanks patrick :) ","body":"XGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance ","comment_length":12,"text":"[XGLUE] Adding new dataset\nXGLUE is a multilingual GLUE like dataset propesed in this [paper](https:\/\/arxiv.org\/pdf\/2004.01401.pdf).\r\n\r\nI'm planning on adding the dataset to the library myself in a couple of weeks.\r\nAlso tagging @JetRunner @qiweizhen in case I need some guidance \nClosing since XGLUE has been added in #802 , thanks patrick :) ","embeddings":[-0.2742777169,-0.1562268734,-0.2091785818,-0.1225316375,-0.0959226638,-0.1439804584,0.2599635422,0.153185457,0.0344703831,0.0236391965,-0.0522928052,0.0668241754,0.0312706493,0.2067699283,0.4939941168,-0.163086161,0.1301592886,0.0034520761,0.0903411135,-0.0966654867,-0.2660258412,0.2607573271,0.1444862485,-0.0042417818,-0.1715925336,-0.2991217673,-0.1470070928,0.1057496592,-0.1567191035,-0.2873096764,-0.1490171552,0.1604385674,-0.2286649197,0.0137421554,-0.0000952988,-0.2044720501,0.1632692963,-0.0263256449,0.1437775344,0.1274793595,-0.2965531051,-0.473511219,-0.2323931158,-0.0420043953,-0.1184929535,0.0144469086,0.1320712417,-0.3555095792,-0.0362836532,-0.0418734774,0.3510540426,-0.0210884456,0.1043468416,-0.1704379767,0.3903431594,-0.230871588,-0.0915244073,0.1624965519,0.7590865493,0.0936698318,-0.0416146107,0.3046829998,-0.00105332,-0.2337667644,-0.0864954591,-0.0876996368,-0.0885687992,-0.1163651571,-0.0421283841,0.5368894339,0.2547674775,-0.2672837377,-0.2090119421,-0.0759900436,0.0242577214,0.0555257052,0.0751243606,0.1695384234,0.1265574843,-0.033608567,0.2392605841,-0.4888306856,-0.0894594789,0.1397513449,-0.0046262485,0.4441057742,-0.1634523422,-0.0035619664,0.1060658172,-0.0219564512,-0.0789262503,0.0640343651,-0.1478862613,0.0097090499,-0.1306984425,-0.0074032736,0.0957263559,0.2829773426,0.2035048604,-0.1293129921,0.1632052064,0.3025079072,-0.0952644497,0.2429935485,-0.1022223756,-0.0690801144,0.1271852404,-0.3906663358,0.2885267735,-0.3118975163,0.069825232,0.0863194242,-0.0212149173,0.0198808499,-0.3518382311,-0.0355433598,-0.0874540284,-0.0195089858,-0.1918211281,0.1291367263,-0.0789504051,-0.0441314317,0.0182418078,0.1366966218,0.0814800709,-0.0045614606,-0.0681345016,0.1595114321,-0.0233968999,-0.3239021301,-0.1112427637,0.4582774341,-0.2896256447,-0.0675472915,-0.0937699899,-0.0117569622,-0.0565662719,0.3399093449,-0.0969850793,0.1450445801,0.3158318996,-0.0096173966,0.1648840755,-0.0947501212,0.0147123048,-0.0287498534,-0.0303222518,-0.2079471797,-0.3481211662,-0.1435165703,-0.0459752306,0.0378921814,-0.5237320065,0.3018259704,-0.0602787957,-0.3669784069,0.3705608845,0.628477931,-0.1873027682,-0.145001784,0.4069532454,0.1466540992,0.0393727534,-0.2790876031,-0.0604388192,0.1980302185,-0.7063208222,-0.2835400403,-0.0141691221,-0.2221071571,0.132270053,0.0466780849,-0.1279182434,-0.1907623261,-0.1125500426,0.3500831425,0.2328341156,-0.2542153597,-0.2042989135,0.0883870497,-0.2166795135,0.0407686792,0.1817221344,0.3123669922,0.0125206895,0.0214362089,-0.0126557015,0.4998044372,-0.0969865024,-0.1383253336,-0.1928712279,-0.1960292757,0.0616425984,0.2080968022,0.0572019704,-0.3447773755,0.3928572536,0.0142166233,0.0035616613,-0.1504184306,0.0723526329,0.2602885067,0.2550289035,0.0174827762,-0.1518219262,-0.2284299582,-0.171663627,-0.0465569347,-0.1549233794,0.2416294366,0.0896044672,-0.1191537455,-0.0232251063,-0.1639386714,-0.1449377239,-0.1328471899,0.3246168196,0.3142593503,-0.0267926324,0.3098312914,-0.1112602502,0.4249493182,0.0398069322,-0.0925852582,-0.2444120646,0.0706681386,0.0340821519,-0.1215469688,-0.1632174551,0.3433803618,-0.2660542727,-0.2048622072,0.2076402456,0.3820528686,-0.170656547,0.2914150059,0.3637571931,0.3595967591,0.2947700918,-0.0527838729,0.3880995512,-0.2382720411,0.0704112276,-0.0284144338,0.2534967363,0.4290236533,0.3430476487,-0.2762015164,0.0739808008,0.1146147922,0.2889913917,-0.0745631233,0.020768743,-0.2264362425,0.307629019,0.077410154,0.4284240901,0.0180358198,-0.4193187356,0.2382009923,0.6861438155,-0.1019127965,0.5521813035,-0.0436834358,-0.2791469097,-0.1156053543,0.0578306764,0.3536409736,-0.0159910265,0.2982699573,0.2538118958,-0.1532248557,-0.1018530652,-0.2309988886,0.2162145525,-0.2101432681,-0.1110442132,0.1501610577,-0.0538076758,-0.2277784646,-0.6874163151,-0.0757365227,0.0182351675,-0.0968999341,0.1712910682,-0.2017069161,-0.2172479779,-0.3654176593,-0.2121680677,-0.3007868826,-0.3960062861,-0.2351064384,0.2575345635,-0.3937160671,-0.0567926168,0.1918391287,0.0398194864,0.3007309735,-0.303132683,-0.1705218256,-0.1445472389,-0.2165330201,-0.1467843652,0.3178697824,0.002797008,0.0089306338,0.4153338373,0.0300526209,0.4382993281,-0.2172783464,-0.6204845905,0.1703937203,-0.2214509994,0.0870820582,0.2307307422,-0.1992539167,0.0191566944,-0.2321950793,0.1240424141,-0.3179441392,-0.059223678,-0.1150261387,-0.2475600392,-0.1028009877,-0.1238960698,-0.3499466181,-0.3936875165,-0.3744566143,0.2842513323,-0.1569627225,-0.0362687558,-0.0057125646,-0.0788377523,0.3879067302,-0.2138548642,-0.0427505486,-0.3189327419,-0.3353681564,0.0284324382,-0.1614083052,-0.2121576667,-0.1112881899,-0.077175498,0.2667379379,-0.0319627002,-0.183155477,-0.0728240237,-0.1678088456,0.2342032492,0.329438746,0.2102174908,0.2111905664,0.0299131703,-0.2941529155,-0.2109241188,0.0089010494,0.032301072,0.1285784692,0.0096426765,-0.0202821922,0.1596352756,0.1161478311,0.2275880277,-0.2759818435,0.0567686521,0.4715641141,0.1356564611,0.4334883392,-0.197437048,-0.2036078721,0.2248235345,-0.0871945024,0.3798136711,0.2312250882,0.0606806353,-0.4111117423,-0.0607576333,-0.0088170553,-0.1732037365,-0.037573155,0.1463441849,-0.0517286696,0.0790261254,-0.0204739422,-0.4773836136,-0.0922211111,-0.0290654972,0.2082470208,0.0037320605,-0.1087006405,-0.0774764195,-0.4288495481,0.0735837892,-0.4345083237,0.3826474845,0.2757713795,0.0534247123,-0.4289885461,-0.0833275318,0.0096842982,0.0045841211,0.204345122,-0.6696708202,-0.3693343401,0.2843756378,-0.0253501795,-0.1088625342,0.2056323439,-0.5182785988,-0.0948737636,0.4616099596,0.0619569942,-0.1430101693,-0.0150112966,0.1007983908,0.1149680838,0.0920410529,0.2686449885,-0.0234302171,-0.2141582966,-0.252055347,0.1282011867,-0.0731537938,0.0328202993,0.5301657319,-0.1180454344,-0.0077520851,-0.0827431902,0.1971974224,0.1604093015,-0.1937441826,0.219630003,-0.2133018076,-0.0148815308,0.0405406244,-0.1574484557,0.2339265943,-0.123452872,-0.0116745243,0.1653179079,-0.3759150207,0.449303329,0.0674554855,0.232697919,0.3954828382,-0.2194926292,-0.1685137451,-0.214167878,0.1572878212,0.0742386729,0.4364535809,-0.304697454,-0.2440004647,0.1930669248,0.1594609618,-0.2432352304,0.4759348631,0.1507908553,-0.25961712,0.01861207,0.0877216533,0.9410984516,-0.0782822594,0.0978498459,0.0937493518,-0.2397655994,0.2125080377,-0.3397639394,0.0507284105,-0.2307864726,-0.2511447668,-0.1371396929,0.0488259383,0.2838183343,0.0590272136,-0.3113557398,-0.1352628171,0.0898457095,-0.022964647,-0.0496555939,0.2600930333,-0.1917844862,-0.2250609547,-0.1021987349,0.296916604,0.0922854841,-0.096296981,-0.2292804718,-0.0955704823,0.0142140416,0.0250678863,-0.3029597402,0.0615355,-0.0918174461,0.1942910999,-0.0479764864,-0.1653027683,0.0636471361,-0.1276819408,0.1905606538,-0.014483152,0.0277530029,0.2603907883,0.0452265739,-0.0018432272,0.2649957836,-0.0676777437,0.4289161861,-0.1257977188,-0.1602305025,0.3073972762,0.0614620335,-0.2659898102,-0.4786525071,0.2083249241,-0.0601964146,0.0518059842,-0.02062672,0.3895162642,0.0610155948,-0.3236421943,0.2439119667,0.2899209559,-0.0018567329,0.0419820808,-0.078847602,-0.2587674856,-0.3496501744,0.0361692384,0.0911347717,-0.0295489375,-0.0626850799,0.1648306996,-0.1603526473,-0.399741143,0.2003573924,0.5205903053,-0.3110843301,-0.0589933395,-0.0755918398,-0.5100243092,0.0624842569,0.0846669376,0.2255124003,0.0914389193,-0.0827186853,-0.1961147636,-0.0008067821,0.1594514251,0.0634786636,0.2461168617,-0.029408237,-0.1255350709,0.1702504456,0.1905012727,-0.4892847836,0.0260768365,-0.1046092585,0.0023274487,0.1547085494,-0.1131711751,0.1838531494,0.1778156012,0.2720205486,-0.1322544515,-0.1588822454,-0.2928080857,-0.2194390148,0.0335001573,-0.1818083227,0.1496213526,0.3129426837,-0.0526561402,0.0643696487,-0.0356205925,0.3001505733,-0.1327546686,-0.082308419,0.1247885376,-0.0526339822,-0.0787711218,0.2935212553,0.020669952,0.09071403,0.132244423,0.1352324188,-0.002423679,0.0949669331,-0.1704802513,0.2274750769,0.3677877486,0.3038661182,-0.0087136868,-0.0256131925,0.1653397977,0.0542828254,0.2000505477,0.1796058267,-0.1096459702,-0.1126142293,-0.0263937227,0.0965260714,0.3588613868,-0.2548570931,-0.3004674017,0.2909342051,-0.0199230611,0.3635517359,0.2576375604,0.4654313028,0.1626667231,0.3227239549,-0.2760549784,0.20226717,-0.1456116289,0.2158945501,0.1679212451,-0.0837132186,-0.103766568,0.272023052,0.2163497955,0.1066098437,0.3222955167,0.1085861996,0.2026608735,0.4585995972,0.3181387782,0.2081259936,0.0177060962,0.1841727495,0.1825535893,0.0019001334,0.3019014001,0.1556647271,0.3037400842,-0.1416059881,-0.0552943125,-0.0401775613,0.3042356372,-0.133985728,-0.1856438071,0.1882041842,-0.1813130677,-0.2850637436,-0.0271798745,0.0641444996,0.0165890343,0.2899083793,0.039817743,0.3159254491,-0.1422883719,0.1996173114,0.0006902194,0.0411421396,-0.2550051212,0.2672004998,-0.0083445217,-0.0420538634,-0.0598075129,0.3378546238,0.177268818,-0.0934022963,-0.1404700428,0.4026169777,-0.286077261,0.0405382365,-0.1464448422,0.2270843983,0.3201424778,0.3445664346,0.2554882467,0.2267757058,-0.2437592894,-0.0012400495,0.1416504532,-0.0962449461,-0.0055065746,-0.0710261464,-0.1543770432,-0.0052716406,-0.4544988871,0.2391119152,-0.5083774924,-0.089176856,0.3120138347,-0.2233425528,-0.0836262703,-0.0944956094,0.1502225548,0.067548655,0.3851423264,0.1739802659,0.2395537794,0.0991646647,-0.4898750186,-0.4001961946,0.1798395813,0.0053985142,-0.1453634351,-0.2671976388,0.1726998538,0.4052922726,-0.0511570573,0.0044821315,0.2739199102,-0.0479621403,-0.2896851003,-0.3363948166,0.2676069736,0.296299547,0.1526449174,0.0082904547,0.0605987385,-0.13210769,-0.0489855185,0.2248755842,-0.3410332501,-0.2622797489,-0.0828525499,-0.1693163365,0.4184142053,0.019289786,0.2519781291,-0.3697897494,-0.1481476426,-0.1701712608,-0.2969485819,-0.1570016146,0.2974040508,0.1577706784,0.1777445674,0.0313716009,-0.1149886698,0.0886008888,0.3226155043,0.009637773,0.4005437195,-0.1830954403,0.2164245993,-0.0060983705,-0.1322339326,-0.0078031607,0.3858168125,0.0863222033,-0.0409533754,0.0627282709,-0.3116734922,-0.0615143515,0.0648160279,-0.1616826057,-0.008993404,0.0640085936,0.0841846764,0.064127475,-0.3489833176,-0.0237735864,0.2662064433,0.233957231,-0.1916790009,0.1523661315,0.1183539927,-0.2164268792,-0.0549221337,-0.0883634314,0.2709396183,-0.176272139,0.1392744631,-0.2985340357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/744","title":"Dataset Explorer Doesn't Work for squad_es and squad_it","comments":"Oups wrong click.\r\nThis one is for you @srush","body":"https:\/\/huggingface.co\/nlp\/viewer\/?dataset=squad_es\r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=squad_it\r\n\r\nBoth pages show \"OSError: [Errno 28] No space left on device\".","comment_length":9,"text":"Dataset Explorer Doesn't Work for squad_es and squad_it\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=squad_es\r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=squad_it\r\n\r\nBoth pages show \"OSError: [Errno 28] No space left on device\".\nOups wrong click.\r\nThis one is for you @srush","embeddings":[-0.2234398425,-0.4402095675,-0.1517270803,0.145961985,0.1462641954,0.1553376466,0.1371591091,0.2519143224,-0.0682639778,0.2752149105,-0.4201240838,0.2000239938,0.0364610776,0.0686686039,0.0943653956,0.151311636,0.0965063944,0.1675711274,-0.0708209276,-0.0054966472,-0.1675249487,0.1495395601,-0.2742112577,0.0363539159,-0.0629873797,0.0800039694,-0.2244669795,0.0782328248,-0.384829849,-0.5345374346,0.2730305493,-0.0230431091,0.2929546535,0.5748831034,-0.0001086667,0.186245963,0.0248020571,-0.0072288234,-0.2695414424,0.0506284535,-0.0131876981,-0.1526875198,0.0983918831,-0.1606301218,-0.1878699511,-0.195244804,0.2092588544,-0.3585654497,0.2048866302,0.3193196654,0.2758903205,0.3985975087,0.1886552125,-0.3512822688,0.2656227946,0.2310827971,-0.1567801386,-0.0990175903,0.2990483344,0.2853501141,0.1349099129,0.1675653309,0.0955989435,-0.1476712525,0.161743179,0.1158927009,0.1390375793,-0.4237047732,0.3848167658,-0.1525680572,0.4322877824,-0.0804457963,-0.0980347618,-0.0941657498,-0.2108548433,0.2332536727,0.4124106169,0.224057138,0.0587718412,0.1121616215,-0.158686921,-0.0569637269,-0.0735386759,0.015049234,-0.2281894386,0.078337729,-0.1577430815,-0.0159548949,0.3848657906,-0.2780478001,0.1338586658,0.1794280708,-0.2060808837,0.0069116554,-0.458635211,-0.2569373846,0.3723630905,0.160320729,0.0147591894,0.0829183385,-0.085153684,-0.0866970718,0.3140895367,0.0477084778,0.3404707909,-0.0117321201,0.3494079113,0.3023525476,0.1197706088,0.0556391068,0.0254449695,-0.0043953038,-0.0374175087,-0.4061364532,-0.2090521753,-0.522371769,0.2554076314,-0.2348600179,-0.3668984771,0.0401463062,-0.0519686155,0.0029537093,0.1729019135,0.5089779496,0.1654711515,-0.1329497546,-0.1014919728,0.180378899,-0.2655635774,-0.0745124817,-0.192122221,0.1390963793,-0.0637816414,-0.0755947679,0.2702535987,-0.1928589344,0.1058584303,-0.1962919235,-0.3221698105,-0.0487365536,0.0425680429,-0.274464488,-0.0895947888,0.1391910613,0.4376441538,0.1323512942,-0.0040678321,-0.3398418725,-0.0370307341,0.2474331111,-0.075470902,-0.0792496577,-0.3246981204,0.2920432389,-0.1661214828,-0.174058184,0.0698922873,0.0990317091,-0.0096071251,-0.2658840716,0.0625528768,0.1980029047,-0.268330276,-0.0281459671,0.1698357761,0.1600189358,-0.5003048182,-0.1374456882,0.0072684069,-0.3086799681,-0.2967842221,0.2676844001,0.0219451264,0.1391532123,-0.4932158291,0.2635519803,0.4709008336,-0.4916559458,-0.3728571236,0.0045102071,-0.1832658201,-0.3734286129,0.1396850348,-0.0224992204,0.3653149307,-0.0873811767,-0.0205250233,0.0075755226,0.0223731156,0.1790546477,-0.2408419847,-0.1227968186,0.2672136128,0.1999234259,0.2312605232,-0.266988188,0.1281405091,0.1061307713,0.3399959207,0.0032302055,0.0475034229,0.2265247703,0.381298095,-0.1033391878,0.0824259371,-0.2284855247,-0.5290923119,0.1028946266,-0.15992257,-0.031792447,0.1933537424,-0.1414292455,-0.4058427513,-0.0825412422,-0.3567085266,-0.2678045034,0.1916503906,0.2298726588,0.2232365459,0.0549226031,-0.1956097335,-0.3706454933,0.0656058788,0.2588733435,-0.1514396966,0.4658104181,-0.3590956628,0.0304150525,-0.0777742639,-0.1701313108,0.1541738808,-0.1241408661,-0.0389397517,0.3524455726,0.1188392788,0.1033529267,0.5112420917,-0.0336495675,0.293697983,-0.4487921,0.1349865794,0.1567705572,0.0307809003,0.1064600796,-0.576908052,-0.1558910012,0.0331678167,0.2438534796,0.1301893741,0.0209527295,0.1251507998,-0.1615510434,0.1561150253,-0.2148759365,-0.1042641848,0.1952779591,0.5971779227,0.1425471306,-0.0596967116,0.2989946008,0.4813050032,0.2349683642,0.0187755004,0.2161989361,-0.2257878482,-0.2157331705,0.1341759861,-0.0498538092,0.1974401027,0.3116164207,0.1880741864,-0.0143284583,0.167436704,-0.1175291836,0.4005159736,-0.0520440266,-0.1146574542,0.1091348976,0.1970361918,-0.0750243291,-0.3947299719,0.0118937297,0.0384029038,0.1230225787,-0.2854169607,-0.3301422298,-0.1461731344,-0.2606966197,-0.0297925137,0.1997716427,-0.3709495962,-0.3142088056,0.3471026719,0.1373558044,0.0867677256,0.0912988409,-0.1635465473,0.3301775157,0.057735201,0.3134717047,0.0113002881,0.1272867024,-0.2209336162,0.0944232643,0.4051365852,0.0588024072,0.2761109471,-0.1189917475,0.161521405,-0.4382263422,-0.3704350591,0.3745687902,0.1348064393,0.0493189842,0.0235639531,0.1651585102,-0.089035742,-0.0871198922,0.2475448847,-0.128824532,-0.2427146882,-0.2209071368,0.1081131548,-0.1156411543,-0.1801084131,-0.4107953608,-0.209849298,-0.3672865033,0.1204875335,-0.0883965865,0.1230024621,0.2280883789,0.0245218482,0.2867732942,0.0007508479,0.0597681515,-0.279853791,-0.3397994637,0.1284044534,-0.3015019596,-0.2347818613,0.2843191922,0.4243242443,-0.1859542131,-0.1623532623,-0.4029372334,-0.199808225,-0.2867853642,-0.0241020247,0.0215538666,0.0440464318,0.4154180884,0.0182647966,-0.1682421416,-0.0366741717,-0.395341903,-0.0321983881,0.0103890831,0.3113749027,-0.3675692976,0.1509856284,0.1951705366,0.3447280824,0.2323287576,0.0033631669,0.3162208796,-0.2036316544,0.5018703938,-0.2453964502,-0.3656563163,0.5349478722,0.1695492268,-0.1146515608,0.5388630033,0.1985167414,-0.0217370279,-0.1946578175,0.1344207823,-0.1831215918,-0.263944,-0.1646105945,0.4319920242,-0.1885179132,-0.1141950488,0.1994416565,-0.1052016541,-0.3615525067,-0.095579721,0.1002745777,-0.1827435046,0.1476974338,-0.557857275,0.1474080682,-0.2226858586,0.4686960876,-0.0856671929,0.1411573887,-0.1766588092,-0.2724738419,0.3133682013,0.0638271943,0.7511894107,-0.1392703354,0.1164806336,-0.1657005697,0.0916325152,-0.4107932448,-0.0469416976,-0.0948059931,0.4150362909,0.2884638309,0.3086353242,-0.0252074003,-0.1560217589,0.2403050214,-0.1024344265,0.0072467183,-0.0058234702,-0.0915156752,-0.1118335798,-0.2791357934,-0.0077951192,-0.0631194934,0.0553143173,0.1716785431,-0.1518667042,-0.0044593774,-0.2452210039,0.0462501645,0.2754687667,0.2162992209,0.2289494872,0.2078976482,0.3748396635,0.4376555979,0.2863232791,0.4913018048,-0.1778350621,-0.2533530295,0.5660368204,-0.3280970156,0.2779302895,0.4320658743,0.1831551045,0.0061232699,-0.0182994436,0.3134420812,-0.069041118,0.090514861,0.3183722198,-0.0392315127,-0.241953373,0.155253619,-0.1632509977,0.1180526987,0.0516167805,-0.0403438918,0.4304373264,-0.314978987,0.0226399843,-0.2286599427,0.8593530655,-0.0836120844,-0.1480150819,0.0708762929,-0.1643661708,0.5332263112,-0.1868626475,0.0886869952,-0.2660962343,-0.3191604018,0.1530744731,0.0176716708,0.1885938197,0.1788944155,-0.3197067678,0.2799776793,-0.0644097477,0.3155790567,-0.0232860334,0.2297875583,0.0764765665,-0.1924987286,-0.3569566309,0.2586321235,-0.0121861957,0.1191288382,-0.0568133891,0.177657485,0.0890847668,-0.1162126288,-0.2176722586,0.0169369224,-0.0516844727,0.2933002412,-0.0391190499,-0.3088862598,0.1016941071,0.4017638564,0.0827997327,0.2805646062,-0.1748368293,0.2485937476,-0.1351228505,0.0581959598,-0.1038220674,0.0309919436,0.1862732917,-0.0684259236,-0.1956147701,-0.00832792,0.044547122,-0.1834152937,-0.1934018284,0.284071207,-0.1807427555,-0.2782163918,-0.1197003946,-0.432438761,-0.1412380785,-0.1099802926,0.1480585784,-0.1941190064,0.1928865463,0.0779497027,0.2961778343,-0.1612393111,0.0019791522,0.397240907,-0.1696939468,-0.2206758559,0.3314729333,0.033836212,-0.2281062752,-0.330145359,0.2094393373,0.0056532891,-0.1599404663,-0.0018850922,-0.0972246826,-0.0817659274,0.138718009,0.0896066427,0.1493631899,-0.1708790511,0.0733508319,-0.2378573865,-0.1257589608,0.253044188,0.0354278386,0.2537222207,-0.1475792527,0.0797393322,0.4233693779,-0.1488526016,-0.3815091848,0.1978015453,-0.3341688812,0.0540331863,-0.1728246361,-0.161648348,-0.0421965197,-0.3832290173,0.1153600886,-0.0231710561,-0.2624351978,-0.2596676946,0.0453222692,0.0850084051,0.1289906949,-0.0338047519,0.1080881134,-0.0258699171,-0.0051345108,-0.2253529131,-0.1026955843,0.1898593754,0.0746432915,-0.2765327692,-0.164630115,-0.277325213,0.0920858607,0.0368066058,0.3832513094,0.0400811099,0.0687186569,0.2369822115,-0.2245240659,-0.0636385381,0.1028056145,0.3906838,0.5452753901,0.1040454656,-0.1004010066,-0.2468933165,-0.07135088,0.3907612562,0.3345629871,0.4041971564,-0.1080931053,0.1183247417,0.2297127694,0.2484002411,-0.2580810189,0.0283648595,0.6684547663,-0.1455660909,0.0252239946,0.1914816648,-0.0421984047,0.1434655339,0.288384378,0.2390310317,0.4860727191,0.0212910119,0.2060350329,-0.1908002794,-0.2738453448,-0.0218849983,0.3374457061,-0.2252998948,0.1781250685,0.1996039897,-0.3639682829,0.1348520517,-0.0359093361,0.0714310333,0.4560913444,-0.1718347222,0.0009817743,0.0268809441,0.0290902443,-0.1799277812,0.0754978433,0.6324539781,-0.2814785242,-0.2391844094,-0.4220416248,0.1404184699,-0.1951296031,0.0952443406,-0.0253704023,-0.2549024224,-0.2626312673,0.0110958181,0.0079182778,-0.4651292264,0.0924999192,-0.141389966,-0.1604019552,-0.2068317384,-0.1000781879,0.1744306386,-0.1241189241,-0.1710980237,0.3448863626,-0.1370967329,-0.1177558526,0.270637989,0.4900676012,0.2644387484,0.0730273277,-0.0536922403,-0.039657928,-0.1066853404,-0.1015334427,-0.2389807552,-0.048848547,0.0084974207,-0.2065933645,0.4321264029,0.1840347946,-0.1868656278,-0.0998616144,0.2089326382,0.2730803788,-0.1900375038,-0.5773889422,-0.3892473876,0.0222205296,-0.0387369394,-0.4212571383,-0.5128836036,-0.1529961526,-0.0701648593,-0.1401499659,0.1646196246,-0.1110213175,0.1047925055,0.1442159116,0.2252937406,0.2122979909,-0.1825629175,0.253057152,-0.2342097461,-0.7275021076,0.0883990675,0.0454922318,-0.0574165657,0.0153756859,0.4763936996,-0.0353577957,0.0897859037,0.5816265345,0.0082141142,0.1955064535,0.0975737348,-0.3863015473,0.155879885,0.0966452062,0.0985280946,0.1335322559,-0.2835779786,0.0503087416,-0.2562494576,0.0159403402,-0.0097815935,0.1733886898,-0.201256156,-0.5459365845,0.3870345056,-0.1953514069,0.6339592934,-0.0232188199,0.054083135,-0.0203459337,-0.3139606416,-0.2908736169,0.2257361561,0.3106091917,0.1783697903,-0.2628771365,-0.281149298,0.0403330736,0.1866378337,0.1744310856,0.2610611618,-0.0880378038,0.0434223339,0.0529542714,0.0686003268,0.097774148,-0.0130191855,-0.1009733602,0.1262278855,-0.3603181243,-0.4724701941,0.3098034561,-0.0291241016,-0.0794356465,-0.1577093452,0.2791119218,0.3063009679,-0.1680700779,-0.345787257,-0.1435119808,0.1469212621,-0.0352621712,-0.0508615039,0.0133655826,-0.3224466443,-0.0262208674,-0.0661617219,0.2232442945,0.2602333426,-0.1318943948,0.1287337691,-0.0547551103]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Thank you !\r\nCould you provide a csv file that reproduces the error ?\r\nIt doesn't have to be one of your dataset. As long as it reproduces the error\r\nThat would help a lot !","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":36,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nThank you !\r\nCould you provide a csv file that reproduces the error ?\r\nIt doesn't have to be one of your dataset. As long as it reproduces the error\r\nThat would help a lot !","embeddings":[-0.0531706028,0.0708665848,-0.0034636373,0.3629573286,0.2447588295,0.2201897651,0.4253186285,0.3036738336,0.3109277785,0.0509861857,-0.0670640767,0.2354821861,-0.0957939476,-0.0218962226,-0.274692744,-0.1102110669,0.0493473858,0.0534440912,-0.3843253851,0.037524797,-0.3989782333,0.3356210291,-0.342168659,-0.0064872117,-0.1416352987,0.2397978306,-0.0553938709,0.3360723555,-0.1290960461,-0.3701969385,0.4902819991,-0.0949141309,0.5313909054,0.2242039144,-0.0001188278,0.0561855994,0.3267244995,-0.0927152708,-0.2536969185,-0.5014423728,-0.0495749563,-0.4797810912,0.3480935395,-0.0440137945,-0.0695052668,0.0655863881,-0.1130333692,-0.3628052771,0.3677662909,0.5194334984,0.1816963255,-0.2161453515,-0.0732249096,-0.0899274051,0.5836966634,0.0712949336,-0.1410909891,0.2741574347,0.0311340336,-0.2326374799,-0.0960092619,0.1110259891,-0.0724671558,0.2592616379,0.2572398484,0.2152855694,-0.1385260969,-0.1685495973,0.0878666192,0.1150503457,0.7295114398,-0.1114873886,-0.1594744027,-0.0453408323,-0.0405998491,-0.0944409966,0.3628718555,0.2874688208,0.0556214824,0.349548161,-0.0392498896,0.0313880295,-0.2870233655,0.1172613874,0.0713907778,-0.0096609592,-0.1667080075,0.3307066858,0.1022032648,-0.0038514612,0.2687622905,-0.0411292315,-0.1087320149,0.2209747434,-0.7359090447,0.3105952442,-0.145863384,-0.0289481953,-0.0242926143,0.1971954703,0.2464096844,-0.0232908204,-0.0816031247,0.3690752387,0.2595978379,0.0325677507,0.0150489165,0.2488965839,0.2425163686,-0.2127750814,0.0344926678,-0.1823337972,-0.3340726197,-0.5727440715,-0.0450330302,0.172230944,0.2509998083,-0.2274134457,-0.2134582698,0.1509696543,-0.265511781,0.0042609028,-0.1363606155,0.2583585382,0.1820407212,0.3217639625,0.102238223,0.3233686984,-0.0812660158,-0.4609397948,-0.0293140858,0.1431435496,-0.4125258923,-0.04699637,0.4920209944,-0.1724782735,0.1270473599,0.2477502078,-0.1354368776,-0.3477976024,-0.0136513198,-0.2814347446,0.0046409201,0.3041781187,0.0724007636,0.2251592278,0.3489328623,-0.202039063,-0.0728904754,0.3434776366,-0.4518966079,-0.2065818906,-0.1280949712,0.0978843123,-0.2282010019,0.0720365569,-0.339265883,0.071510911,0.0557156317,-0.1955105513,-0.1711066067,-0.319401443,-0.2678619921,-0.2312421948,0.0226150919,0.4551670849,-0.8168234825,0.2262132317,-0.0098615317,0.0500961393,0.0799537003,0.2039691955,-0.26779145,0.0344988219,-0.2549441755,0.1157383472,0.1402722597,-0.2161636353,-0.2937296331,0.1842556894,-0.0505548716,-0.0749084279,0.0845157504,0.0922507197,0.3441407084,-0.0177558176,-0.2049922645,0.1499433815,0.0094183013,0.0440510996,-0.2377330959,-0.1822585315,0.244654566,0.2318156809,0.0159404799,0.1241321713,0.2280937433,-0.6181377769,0.3143367767,-0.0560143255,0.2522957623,0.1251528263,0.2919839323,0.0680332705,0.331485182,0.0560596585,-0.5451503992,0.19358778,0.4534468949,-0.1884105951,-0.090702869,-0.0167362671,-0.4643526971,0.0068462631,-0.1820607185,-0.0329604782,-0.0002793333,0.0497451872,-0.0883560553,-0.1819704771,-0.115388982,0.4194991291,-0.2807845473,0.0929733589,-0.3300335109,0.280695945,0.1371117383,-0.0622213967,0.2043175697,-0.0787182376,0.315700233,-0.0505178608,-0.1529254764,0.2792815268,0.3547394872,-0.0294624101,0.1198978052,-0.1979084909,0.0191887449,-0.2967647016,-0.009896161,0.2180438787,0.3470375836,-0.0931992084,-0.484100908,0.324023366,-0.3189892173,0.2089032382,-0.2283426225,0.1850348562,0.2576622367,-0.0349268019,-0.0827434361,0.0283303466,0.0569879189,0.0253285822,0.2225339562,0.0946080312,-0.1744120866,-0.141436547,0.0628362969,0.181490019,-0.2388355881,0.2301218361,-0.1126146689,-0.0576637127,0.3364206254,0.2839767933,0.4159408808,0.0849344209,-0.2692225575,-0.0180044211,0.0707981363,-0.1863808781,0.0902066603,0.0085394103,0.0635168999,0.1984077096,0.2047156543,0.0949885398,-0.1371750087,-0.1268930137,0.1822054684,0.2504083812,-0.6004428864,-0.0170621369,-0.1977541596,0.3690707386,-0.231453076,-0.0469768457,0.1054718569,-0.2846023142,-0.2780268788,0.3058357239,-0.1332714558,0.0533150509,-0.2049994469,0.2518928945,0.1134851575,-0.1292214096,0.0387605093,-0.1857044101,-0.2578786016,0.0051694517,0.0914032385,-0.02278913,-0.113893643,0.1033109799,0.0897010043,-0.3063434064,-0.0327222049,0.1250150055,0.0054551079,0.0982463285,0.1273556799,-0.0856845826,-0.0653377175,-0.1550449729,0.2096203864,0.0910362974,0.0384264775,0.4184431136,0.2405577749,-0.1648762971,-0.1206735224,-0.6375058293,-0.2401450276,-0.2200728506,0.0447856039,-0.1669529229,-0.0543237068,0.0653824583,0.3040782213,-0.0523246713,0.2384745777,-0.1168534085,-0.1043981314,-0.2288609892,0.5683006048,-0.2417068183,-0.2770061493,0.2639984488,-0.0347369425,0.0478000417,0.3508776724,-0.396723628,0.2066193074,-0.0642232522,0.0257207323,-0.0835890323,0.0371143781,0.0664877445,-0.03237468,0.0934540704,-0.0514092781,-0.1807925254,0.2330056578,-0.0544732921,0.04972969,0.1682844162,0.556370616,-0.2356366813,0.6716985106,0.1087843254,0.1012935862,0.3297623992,-0.0753615946,0.2564425766,0.171028018,-0.224067837,0.0610778555,-0.1236017346,-0.3338024318,0.3212258816,0.0084113618,-0.1048502102,-0.0734925643,0.2458775342,-0.2612117827,-0.168322742,0.1560325772,-0.0741376504,0.0590328127,-0.0084796166,0.1152755618,0.0636246577,-0.2108647227,-0.0467193946,0.1300501078,0.1895849109,-0.0741187781,0.1287033856,-0.125347361,-0.2497012019,0.2363940924,-0.1328609586,0.4405017495,0.0873843133,-0.0750820413,0.0294266976,0.185069859,0.7332041264,-0.2643435001,0.0262117945,0.1199072748,-0.054726731,-0.1926729232,-0.3259961605,-0.082532838,0.221994251,-0.041279383,0.2150913179,-0.0131860226,0.0535505526,0.385982424,0.0705647245,-0.1538188607,-0.4156410992,-0.1248806119,-0.0565223284,-0.1333364695,-0.2428255379,0.1577137709,0.056347739,-0.2371139228,0.1061412245,-0.218386367,-0.2058781981,0.3085451424,0.1126157418,0.2635262311,0.0404759087,0.1960004568,0.2719372809,0.1261612475,0.2306050956,0.7737972736,-0.1709044129,-0.8922721744,0.2450987548,-0.5180836916,0.3393909931,-0.0342068598,-0.3213843107,-0.0762142986,-0.2587061226,-0.1450984031,-0.1039241552,-0.1216759458,0.0747581124,-0.0106530106,-0.1297347546,-0.4533146918,0.5562050939,-0.132987991,-0.1653013974,0.1423024684,-0.0152589437,-0.3003289104,0.4268864989,-0.0881930813,0.6560729742,-0.1391480565,-0.0989319012,0.1329803616,-0.2335326374,0.3357267976,-0.2730204761,-0.0558692142,-0.5120702386,-0.1469634026,-0.1590158939,-0.003742598,0.1034840345,0.5652615428,-0.2156032771,0.3151760697,-0.1441103816,0.4138491452,0.0545260124,0.0907121003,0.3058978021,-0.167015478,-0.1141053736,0.1325440705,-0.1335925162,-0.0413920842,0.0516827106,0.0316274539,-0.2306943536,-0.0953274518,-0.3436087966,0.3158481717,-0.0588376336,0.0750975981,0.1861771196,-0.4859144986,0.5535690188,0.4605587125,0.1939994246,-0.1114400774,-0.220122844,0.1092188731,0.0466070436,0.0503436066,-0.1632008106,-0.0240384992,-0.0615488328,0.2152404487,-0.2079122812,0.280238241,-0.1053205431,-0.082716547,0.2299666852,0.272598505,0.2134948373,-0.4861702919,-0.355322659,-0.2251552492,0.1214449927,-0.1187396944,0.0530749112,-0.0247510895,0.1066573709,-0.0446031094,0.0643731132,-0.3480309844,0.1839970946,0.2229261547,-0.2328323871,0.0587838292,0.4570632279,0.4909372628,-0.1183462441,-0.1546209902,0.0396810398,0.1785497367,-0.6025990248,0.1056115702,0.0670924932,0.2354789972,0.1761479974,0.1965438724,0.2040313631,-0.211319834,0.1216653883,-0.5683940053,-0.0207868945,-0.0678644925,-0.1500790715,0.0608933493,-0.0695432127,-0.4517227113,-0.0591822341,-0.0651101097,-0.1966659427,0.080567129,0.0609497726,0.0550402254,0.169142887,0.3050051928,-0.2626045346,-0.2842249274,0.017431343,0.0360207893,0.078960672,-0.1089611873,-0.1910560131,0.1684363186,-0.0031499341,0.1523512155,-0.2686831653,-0.1957891285,-0.2490988672,-0.1680260152,0.0847067684,0.3114747405,-0.0280915815,0.4006356001,0.033898335,-0.1250307858,-0.494327575,0.2662498951,-0.2059058994,0.1760911793,0.0802545026,0.1721066982,0.1556380391,0.0842218548,-0.1465845108,0.0710210577,-0.1401133537,-0.1272111088,0.2502036691,-0.3445920646,0.1869973987,0.2318910062,0.3359310031,0.3265926838,-0.1360160112,-0.0745562389,0.1204319447,0.1417509168,-0.3591190577,0.1969389915,0.0459274277,0.2983696759,-0.121851787,0.2741483748,-0.0423089117,-0.1998051107,0.0856372938,0.1599937677,0.6721033454,-0.0011855514,0.0585343502,0.2218623161,-0.0432721451,0.0969640538,0.1767922491,-0.0449934341,0.0933889523,0.4701834321,-0.1698187888,-0.0640111193,0.3982224464,0.2136934549,-0.1883458048,-0.5901343226,0.202379927,0.3499044478,-0.2514417768,-0.1630514264,0.1874289513,0.6417908669,0.0472944528,0.067786485,-0.3391375542,-0.0705886632,0.0864299834,-0.0597232319,-0.2506252527,-0.1946726888,-0.4349193573,0.2554750741,-0.2009301931,0.0267460532,0.501475811,0.1956419349,-0.0481147617,-0.5706117749,0.1003892049,0.2635083795,0.033516638,-0.317953378,0.0991414636,0.0033047139,-0.0355581418,0.3778116703,-0.138439998,0.388450563,0.450404942,0.0671169907,-0.3161944449,-0.1921661794,-0.0744566917,-0.0587588772,0.4470014572,-0.0163553488,0.2828291953,0.2275445163,0.1104676574,0.0034672467,0.3139467835,0.0639150515,0.1889017373,-0.406386286,-0.11627312,0.0189674906,-0.0651797727,-0.0975170434,-0.2211100012,-0.0924166739,-0.2157883346,0.4446016252,0.1799662411,0.2071263194,0.1574011892,0.0032119942,-0.1497551948,0.1894354671,0.4540915489,0.0465285704,-0.5287822485,-0.0883442238,-0.8030949235,-0.0099850828,-0.2449389994,-0.1308873743,0.0474374741,0.2087431848,0.307241559,0.2748607695,-0.02985036,0.1222134903,-0.0819613785,0.3319580853,-0.1741870642,-0.1797350496,-0.1086854711,0.100477308,-0.0626904443,-0.4160607755,0.0869849622,0.3398024738,-0.1194699481,-0.1836851835,-0.1013246998,0.0599796288,-0.2456318587,0.4114705324,-0.0058905259,0.4132962227,0.0335021615,0.2946876884,-0.4566014111,-0.2764179111,-0.1911329925,0.1752238572,-0.1380836815,0.4519537985,-0.3454532027,0.0347793885,-0.1817922145,-0.0487446897,0.0195972677,0.2384879738,-0.4080229998,0.2792759836,-0.1106728837,0.2292044759,-0.1760122031,-0.1346166283,-0.1204197705,0.2537302673,-0.1035771221,-0.2741242945,0.3016401827,-0.2517799139,-0.2887698412,0.0469480008,0.2282957137,0.1620085537,0.1029795185,-0.5200520158,-0.036144238,0.3270946145,-0.0772534683,-0.1145509332,0.1069713235,0.1176324859,0.0564750768,-0.1519694477,0.2231712341,0.309307307,-0.1856913269,-0.1730553359,-0.3540093005]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"I think another good example is the following:\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sts-dev.csv\"], delimiter=\"\\t\", column_names=[\"one\", \"two\", \"three\", \"four\", \"score\", \"sentence1\", \"sentence2\"], script_version=\"master\")`\r\n`\r\n\r\nDisplayed error `CSV parse error: Expected 7 columns, got 6` even tough I put 7 columns. First four columns from the csv don't have a name, so I've named them by default. The csv file is the .dev file from STSb benchmark dataset.\r\n\r\n","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":72,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nI think another good example is the following:\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sts-dev.csv\"], delimiter=\"\\t\", column_names=[\"one\", \"two\", \"three\", \"four\", \"score\", \"sentence1\", \"sentence2\"], script_version=\"master\")`\r\n`\r\n\r\nDisplayed error `CSV parse error: Expected 7 columns, got 6` even tough I put 7 columns. First four columns from the csv don't have a name, so I've named them by default. The csv file is the .dev file from STSb benchmark dataset.\r\n\r\n","embeddings":[-0.0901070535,0.0171127245,0.0167961176,0.3881013691,0.2925541103,0.2115879506,0.3847278059,0.229759112,0.2224226147,0.089858532,-0.1013280302,0.2565506101,-0.0943530872,0.0382322073,-0.2512097657,-0.1036800966,0.0104466416,0.0889976621,-0.2961676121,0.0908844918,-0.3902235329,0.3168067336,-0.3827799261,-0.0461566821,-0.1766609251,0.2577837706,0.0003086621,0.2659320533,-0.0958688855,-0.4407778382,0.5010362267,0.0035070458,0.5712614655,0.3400994837,-0.000120307,0.0167349447,0.3593454361,-0.1578391045,-0.2551825643,-0.4866746068,-0.0636008605,-0.4186861515,0.3337731659,-0.1273178309,-0.1325138211,0.0567032024,-0.0902742967,-0.5262620449,0.3351167738,0.4816686213,0.1698999405,-0.1670440137,-0.1962601095,0.012288156,0.6344578266,0.0383948833,-0.1703587919,0.2307640314,0.060028255,-0.0478319153,-0.077239953,0.0414571948,-0.0999131575,0.2801174223,0.2486495972,0.2107812464,-0.2227522135,-0.1397711188,0.1007597297,0.1845953166,0.7821591496,-0.1575741917,-0.160560146,-0.0145655898,-0.0244933534,-0.088375546,0.3267967999,0.3057576716,0.0497207083,0.3230619133,-0.1105097607,0.1278137118,-0.2835086882,0.0646023676,0.0731800869,-0.0285636242,-0.2363406867,0.3204714358,0.0431455448,-0.0127093485,0.3014170825,-0.128121689,-0.1516457647,0.1567753404,-0.7318518162,0.2994091809,-0.1890997142,-0.0831176937,-0.0139113748,0.1470643878,0.2610993981,0.0409128815,-0.0495255105,0.3694851696,0.1844192296,0.0623423234,0.074477002,0.2123364657,0.2606661618,-0.1923152953,-0.0245101713,-0.1213865876,-0.2915119529,-0.4715484083,-0.1039494053,0.1235364452,0.2158168554,-0.2365605086,-0.2136596739,0.1357294023,-0.2236691117,0.0182397123,-0.1153503507,0.287561208,0.0961308181,0.2370782644,0.0844724104,0.1946082264,-0.0593218282,-0.4193832576,-0.0320638455,0.0889588967,-0.396494329,-0.0089434134,0.4253871143,-0.2059131116,0.155105412,0.2090531737,-0.1807143092,-0.3763448596,-0.1051809862,-0.29985255,0.1227002218,0.2350853831,0.0722876862,0.22527197,0.1758594811,-0.2096450329,-0.0948163494,0.3743303716,-0.4765123725,-0.203762114,-0.1129608005,0.0870465264,-0.3702276945,0.042812489,-0.4717314839,0.0912087634,0.0114765568,-0.1476510614,-0.1846654564,-0.3796912134,-0.3459873796,-0.2003361583,0.0562690981,0.4900042713,-0.7549694777,0.2011175901,0.0685681775,0.0840682685,0.1757277995,0.1387699544,-0.2232251018,0.0998104513,-0.2925262749,0.0760307387,0.117212534,-0.1124075651,-0.3250758946,0.1897105426,-0.0210658405,-0.0349105038,0.1909705698,0.0766447634,0.4279769659,-0.0000178403,-0.2978307605,0.1610570997,0.0182893444,0.0978719965,-0.2641838193,-0.2243623137,0.21959728,0.2429786772,0.0428958684,0.0870601982,0.2906083465,-0.4773252904,0.2948067486,-0.0837217569,0.2005083561,0.1304012984,0.1938507706,-0.0031297801,0.3042298853,-0.0209543332,-0.5490700603,0.2013014704,0.5036683083,-0.1273956448,-0.054686565,-0.0433550999,-0.4341191947,-0.0346208662,-0.1757203788,0.0496488176,-0.0256375745,0.1051591933,-0.0783729777,-0.1835370809,-0.0745789483,0.3672889173,-0.3182172179,0.0592879392,-0.3288670182,0.3052717745,0.1356878728,-0.0979554504,0.2334704846,-0.0340172574,0.2635864615,-0.1565463692,-0.1594044268,0.2825955749,0.3972230554,-0.0089151813,0.0568526611,-0.1653582603,0.0449949019,-0.2886992395,-0.039644707,0.2083660811,0.3289196491,-0.052529674,-0.388286978,0.3406111598,-0.3962658346,0.2122683227,-0.185850665,0.0636317208,0.2840827405,-0.0627926365,-0.0625529289,0.0368210375,0.0639444813,0.0541221611,0.2319158167,0.1155366451,-0.1605666876,-0.0800897554,0.2399234921,0.1933390945,-0.2646180987,0.2249322534,-0.1317929924,-0.0654701069,0.3131316006,0.2746325135,0.3821977377,0.1115466356,-0.2823289633,-0.0077096047,0.1524128914,-0.2335685045,0.1316001713,0.103461735,0.0353699438,0.2109541148,0.1988956332,0.104850173,-0.1725974083,-0.1863415837,0.1725758761,0.2190569788,-0.5805425644,0.0048966892,-0.2296882421,0.3273848593,-0.2782441974,0.0585424565,0.1733689457,-0.2118965089,-0.3372685611,0.2768916786,-0.1294859648,0.0467706956,-0.1542281657,0.3068214059,0.1176594198,-0.2467561811,0.046924077,-0.2137753069,-0.3152797222,-0.0071447152,0.104252696,0.0340959989,-0.1043617055,-0.037348751,0.0876660049,-0.24026151,-0.0827936083,0.1639792025,0.0137606245,0.1385041326,0.1075599939,-0.1262121946,-0.0486987755,-0.15883407,0.1437249035,0.0194418002,0.0146210454,0.3827898204,0.2163419425,-0.209561646,-0.1090568677,-0.6926314831,-0.1795829237,-0.2248694301,0.025014244,-0.1732673943,-0.0428306684,0.047063034,0.3236681819,-0.0238800831,0.2033355534,-0.0535822399,-0.0562276058,-0.3528072536,0.5881902575,-0.2043498456,-0.3377241194,0.2926166058,-0.0908306092,0.0103339078,0.3109973967,-0.4887527227,0.2925721109,-0.055077184,0.0720270127,-0.0035766959,0.0297304932,0.0222400259,-0.0356193259,0.1201732904,-0.0546184741,-0.2449769825,0.197209239,-0.075742051,0.119669348,0.2698403597,0.434009552,-0.2318444848,0.6490834951,0.0943914577,0.006586032,0.2863881886,-0.0369606726,0.3239893317,0.1193069592,-0.1768079102,0.1326798797,-0.0976964459,-0.3513123393,0.445933193,0.0127536049,-0.0088594798,-0.0752964392,0.2621181607,-0.1663679779,-0.1729792356,0.1763002574,-0.105628565,0.0349246114,-0.0148693668,0.115628317,0.1018472612,-0.2759167254,-0.047250621,0.1879240572,0.1916073412,-0.0905418098,0.0025331709,-0.0383909605,-0.2599628866,0.2224005908,-0.032826338,0.4320131838,0.1223770902,-0.0044626454,0.0574928001,0.1865427941,0.6240664721,-0.2614255548,0.0962498114,0.1275530308,-0.0515436567,-0.2346757054,-0.2981118858,-0.1358577609,0.1346204728,0.0321769938,0.1876169145,-0.0290286057,0.0071552182,0.4268553853,0.1015386879,-0.1602927148,-0.4608369172,-0.1265007704,-0.0646243021,-0.0824340507,-0.2703128159,0.1493461877,0.1094130576,-0.2964922488,0.0023714667,-0.2398501784,-0.2355900705,0.2956104279,0.1571993232,0.2442665249,0.1096603498,0.2460027784,0.1996624768,0.1105183735,0.3350627422,0.7943231463,-0.2134048045,-0.8323941827,0.250285387,-0.4999430776,0.3353821933,-0.0302735902,-0.3484945595,-0.0081958482,-0.2992192209,-0.1562387198,-0.151406154,-0.0122491904,0.0357026085,0.0479065366,-0.0827963799,-0.4753679037,0.540831089,-0.1637014002,-0.1382284313,0.0493597575,-0.008981986,-0.3495826721,0.4944770932,-0.0811155066,0.6880194545,-0.1763131618,-0.1290588677,0.0714659169,-0.2028657198,0.3421262205,-0.253944546,-0.0582686141,-0.4942832589,-0.2201174647,-0.1425331682,-0.0417384878,0.1611692458,0.5277663469,-0.2060645074,0.3125900626,-0.0184136219,0.5204398036,0.1101227477,0.17576994,0.3855985105,-0.1565244794,-0.1727927327,0.0980241522,-0.1834193766,-0.1085250229,0.0740488842,-0.0011511861,-0.2079226077,-0.0559914671,-0.3236173987,0.2994101346,-0.0577524006,-0.0715499818,0.116463311,-0.518186152,0.6153252721,0.479262948,0.1595358402,-0.0059801228,-0.186324209,0.0537879877,-0.0308515243,0.0172794275,-0.1223404855,-0.0043034451,-0.0322226472,0.2211479992,-0.2293285877,0.2947227061,-0.0806328505,-0.1131349355,0.303347677,0.225277409,0.2555881739,-0.4614585638,-0.3070752621,-0.2253314704,0.1091273725,-0.1241648048,0.0422348566,-0.0958329588,0.071121715,0.0882018954,0.1338154227,-0.389028877,0.1239156425,0.2190839052,-0.1784569472,0.0967702419,0.4302880764,0.4298396707,-0.1339468062,-0.1351148039,0.1060404405,0.0850794539,-0.4980711639,0.1417179555,0.0060172998,0.2284625322,0.1437237114,0.2385239303,0.1999004483,-0.2316476405,0.1122665256,-0.563555181,-0.0219043791,-0.1168763191,-0.1212229729,0.1035309955,-0.0057635885,-0.3888217211,-0.1755323708,-0.0525572933,-0.1856177747,0.0855126306,0.0474717319,0.1009461731,0.2080987394,0.3193138242,-0.2697918117,-0.3296963871,-0.0113080572,0.0608374625,0.0761810541,-0.1239528209,-0.2231418937,0.191404134,0.0444764234,0.1103503257,-0.3429810703,-0.2133450657,-0.2508711815,-0.2187032402,0.13683559,0.3249601126,-0.0459598154,0.459479332,0.0526612177,-0.1193289235,-0.5032320619,0.318826437,-0.2290262729,0.3656277955,0.0728344917,0.2288548052,0.1421497911,0.1032780856,-0.1216418669,0.0269141868,-0.0953238234,-0.0933778659,0.3047415912,-0.3803228736,0.154953137,0.2595809698,0.3093921542,0.3506708741,-0.1301964074,-0.2125749737,0.0924840495,0.0970567539,-0.3641675115,0.1188809127,0.0667255968,0.3423583508,-0.0695171505,0.3417379558,0.0332781486,-0.2004050016,0.1132652238,0.1477335393,0.6941482425,0.1396491081,0.1238417849,0.1839379668,-0.0157241523,0.1573635489,0.2216382921,-0.005368331,0.0212602261,0.4586307704,-0.1159985065,0.0383616649,0.4242616296,0.2985851169,-0.1576149762,-0.5123638511,0.2923939526,0.3320300877,-0.1976994872,-0.1058691889,0.2124358863,0.5360161662,0.034632612,0.0418690965,-0.3084710836,-0.1861308068,0.0684134737,-0.0889787301,-0.1709145159,-0.1957310736,-0.3741849065,0.2757281065,-0.2003026009,-0.0128349531,0.4741038978,0.1880678385,-0.0518330932,-0.5019614697,0.1363547295,0.2327768803,0.0486846045,-0.3137757778,0.1842669547,-0.0501237027,0.0104461806,0.4037235677,-0.0495144613,0.3289115131,0.4287268817,0.054724738,-0.4291597903,-0.1533870399,-0.0846829712,-0.0640428737,0.4988794029,-0.0862210616,0.1686631441,0.2531365156,0.0904214159,0.0121704005,0.3275724649,0.1407345831,0.1382683069,-0.4157441258,-0.1013825908,0.0889673308,-0.0586135499,-0.092644237,-0.1818147451,-0.0369553082,-0.1980533898,0.4690285325,0.06864582,0.2260659039,0.1646199077,0.0022851576,-0.1956307739,0.2916514575,0.3827778697,0.0330598913,-0.5024465322,-0.0478829183,-0.7514445186,0.0344696157,-0.1055863127,-0.1439944357,0.0284269173,0.1896352321,0.2427166849,0.3166953623,-0.067066133,0.1263182759,-0.0666285232,0.3325902224,-0.2332709879,-0.2137629688,-0.0828296542,0.049770657,-0.1109112352,-0.4989263713,0.1240908131,0.2586054802,-0.1643968821,-0.1973007619,-0.0831733197,0.088095583,-0.332262665,0.4126713276,-0.0374362171,0.4819592834,-0.0488995165,0.2920432091,-0.3914601505,-0.2983885705,-0.1549179852,0.2218884677,-0.0646122247,0.4354818463,-0.3110987842,0.0208778866,-0.2157057226,-0.1098568216,0.0801395997,0.2470867783,-0.4290534854,0.2016562372,-0.190440774,0.2222689539,-0.1485283673,-0.1051933989,-0.1315124482,0.2425079942,-0.0621553138,-0.2259265184,0.2711745501,-0.2285078317,-0.2634802461,0.0477851182,0.2037994862,0.1680153459,0.0844740495,-0.5076405406,0.0372539051,0.4007314742,-0.1354267895,-0.068827942,0.1657315195,0.0443446934,0.1395900249,-0.1589803398,0.1779274493,0.2822124064,-0.1992393285,-0.2089348286,-0.3870800734]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi, seems I also can't read csv file. I was trying with a dummy csv with only three rows.\r\n\r\n```\r\ntext,label\r\nI hate google,negative\r\nI love Microsoft,positive\r\nI don't like you,negative\r\n```\r\nI was using the HuggingFace image in Paperspace Gradient (datasets==1.1.3). The following code doesn't work:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', script_version=\"master\", data_files=['test_data.csv'], delimiter=\",\")\r\n```\r\nIt outputs the following:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-3b6254ff4dd403e5 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/csv\/default-3b6254ff4dd403e5\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nDataset csv downloaded and prepared to \/root\/.cache\/huggingface\/datasets\/csv\/default-3b6254ff4dd403e5\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2. Subsequent calls will reuse this data.\r\n```\r\nBut `len(dataset)` gives `1` and I can't access rows with indexing `dataset[0]` (it gives `KeyError: 0`).\r\n\r\nHowever, loading from pandas dataframe is working.\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\ndf = pd.read_csv('test_data.csv')\r\ndataset = Dataset.from_pandas(df)\r\n```\r\n\r\n","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":141,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi, seems I also can't read csv file. I was trying with a dummy csv with only three rows.\r\n\r\n```\r\ntext,label\r\nI hate google,negative\r\nI love Microsoft,positive\r\nI don't like you,negative\r\n```\r\nI was using the HuggingFace image in Paperspace Gradient (datasets==1.1.3). The following code doesn't work:\r\n\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', script_version=\"master\", data_files=['test_data.csv'], delimiter=\",\")\r\n```\r\nIt outputs the following:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-3b6254ff4dd403e5 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/csv\/default-3b6254ff4dd403e5\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nDataset csv downloaded and prepared to \/root\/.cache\/huggingface\/datasets\/csv\/default-3b6254ff4dd403e5\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2. Subsequent calls will reuse this data.\r\n```\r\nBut `len(dataset)` gives `1` and I can't access rows with indexing `dataset[0]` (it gives `KeyError: 0`).\r\n\r\nHowever, loading from pandas dataframe is working.\r\n```\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\ndf = pd.read_csv('test_data.csv')\r\ndataset = Dataset.from_pandas(df)\r\n```\r\n\r\n","embeddings":[-0.0118629886,0.0808789656,-0.0193711855,0.4624471068,0.2891317606,0.1691015065,0.3954825103,0.2487924546,0.2549251318,0.1220230907,-0.1881923825,0.2273677438,-0.1354269832,-0.0153029179,-0.2497221678,-0.0935218707,0.0260975901,0.0599039607,-0.3732225597,-0.0090303589,-0.3582966924,0.3465511799,-0.3180826902,-0.0767824054,-0.100431487,0.1366493404,-0.0739043728,0.3516087532,-0.2565934062,-0.2810692787,0.4602858126,0.027672179,0.5446310043,0.3226149082,-0.0001188028,0.1157893538,0.257464081,-0.125779599,-0.1716525704,-0.6218040586,0.0017178034,-0.4847868681,0.2545427382,-0.1264301389,-0.1261842698,0.0441867076,-0.0974872708,-0.3808397949,0.4975135028,0.5468569994,0.1608204097,-0.2178964913,-0.002659607,0.0565558858,0.67342031,0.047714334,-0.1692203432,0.270789206,0.0260531921,-0.0865285397,-0.0880396217,0.1368653774,-0.0883378312,0.2905078828,0.2541857958,0.2660015523,-0.1604845822,-0.2315118462,0.0664864704,0.1838555038,0.6319200993,-0.0759829357,-0.1848375648,-0.0473118685,0.0015072487,-0.2142893523,0.3849784434,0.32181862,-0.0536062904,0.3348769844,-0.1593370736,0.0491260625,-0.4021936655,0.1563876867,0.0565393716,0.0167305879,-0.2711093724,0.2473977357,0.1620324701,0.0409820639,0.3063595295,0.0125093972,-0.0958893001,0.2550705969,-0.6877479553,0.3242422044,-0.1610052288,-0.0471008494,-0.0373417251,0.1841595173,0.3348793387,0.018604733,-0.100939855,0.3323724568,0.1696489453,0.0724511296,-0.0694890693,0.3101442754,0.3111258745,-0.1418602616,0.0889902264,-0.1957607716,-0.3813566566,-0.5510748625,-0.1370717734,0.153520748,0.2152677923,-0.191373378,-0.1348076761,0.0824970827,-0.29584378,0.0412005521,-0.1037042812,0.2584906816,0.1181269959,0.1929664612,0.1120182201,0.217828393,-0.0669630021,-0.3518299758,-0.0048816982,0.1507574767,-0.4125787318,-0.0420589931,0.4580542147,-0.1998315305,0.1430309713,0.1652851403,-0.0317209847,-0.3051397204,-0.027504025,-0.345038265,0.1042648628,0.3114556968,0.0919667482,0.1780766696,0.2598017156,-0.2742810249,-0.0650035962,0.2549707592,-0.4401567578,-0.1146177351,-0.0722140744,0.0712279826,-0.2610419691,0.1687274575,-0.5450232625,0.1131039113,-0.0171524789,-0.1490582228,-0.1941609979,-0.3184810877,-0.2939890623,-0.2533073127,0.0877587721,0.4105604887,-0.7388890386,0.1863694638,-0.0079962043,0.0084585734,0.1423375905,0.2571278811,-0.2166781574,-0.0076674805,-0.2400081605,0.0525184162,0.1994749755,-0.2259487063,-0.4637648165,0.1456199139,0.0125081893,-0.0459893011,0.1068215147,0.0247449912,0.4188020527,0.0953659117,-0.2324724197,0.2021092921,0.0218010601,0.0537071675,-0.2709424198,-0.2555734217,0.2727690637,0.2312987149,0.0567305461,0.0727248862,0.2554844618,-0.5654004812,0.3175981641,-0.0955061093,0.2537537217,0.103790231,0.305829823,0.0119233225,0.2585232258,0.0209247191,-0.5154770017,0.1781058162,0.3431998193,-0.1195503771,-0.1280687302,-0.0305286143,-0.4309688807,-0.1388527304,-0.2483625561,-0.1230339333,0.0097857034,0.0791332871,0.0067214277,-0.176499486,-0.1114899814,0.4863345623,-0.3433071971,0.1073693335,-0.3811258078,0.2562277913,0.1100255027,-0.0257717185,0.2472402304,-0.0620501488,0.2698598504,-0.1324793398,-0.1780457497,0.2074373513,0.348687917,-0.0601232275,0.0584408008,-0.1465489864,0.0486193746,-0.278768152,-0.0639291406,0.1557741612,0.3600311279,-0.0460514463,-0.5185007453,0.3144533634,-0.3182805181,0.1841373742,-0.1650368422,0.1002310738,0.2616048455,0.011042295,-0.1340304017,0.006500219,0.054611396,0.1028838083,0.3047707677,0.0061566592,-0.0917265415,-0.169329986,0.0866777003,0.1697233468,-0.2521165609,0.209491834,-0.078904435,-0.0527838208,0.3029036522,0.3002995253,0.3695094585,0.0668440759,-0.2557020485,0.0274065994,0.0987148434,-0.2053614408,0.0438876525,0.0450873449,0.1305779815,0.167106539,0.2208328098,0.1397261918,-0.1462558806,-0.1437519342,0.1935928166,0.2692375183,-0.5836182237,0.0046391208,-0.2520117164,0.2815494239,-0.177628994,0.0017936039,0.1016970277,-0.2104956359,-0.3504090011,0.2913831472,-0.0326991342,0.0169307422,-0.0212115459,0.3473856151,0.1048926115,-0.0763040558,-0.0240837857,-0.1735054851,-0.2351216227,0.0199389532,0.0443579592,0.0041387249,-0.0721165389,0.0316162929,0.0195567273,-0.2486414164,-0.1216644198,0.1549575031,-0.0072373631,0.18079032,0.0941220745,-0.0761066079,-0.1042506322,-0.0307540875,0.171576336,0.1048674658,-0.0070715062,0.2781068087,0.203579098,-0.1809631735,-0.0832260624,-0.5839006305,-0.3633822799,-0.2267879546,0.0676118359,-0.1146416292,-0.0052905944,0.0991690531,0.2561677396,0.0274770912,0.2174453735,-0.0561032481,-0.1361311823,-0.3465022147,0.5890855789,-0.2951643169,-0.3459739089,0.2518579364,-0.0580747612,0.0655628592,0.454408139,-0.4482577145,0.1498956829,-0.0319364406,0.0750386342,0.0224172082,0.0262204837,0.0370181464,-0.0896322504,0.1411664933,-0.0410196222,-0.1947057843,0.1235551015,-0.0017070151,0.1028957069,0.248731181,0.441820085,-0.1863634586,0.7564931512,0.1680112332,0.0771736503,0.303815335,-0.1413715333,0.3919916451,0.1962672919,-0.2521838546,0.1308351606,-0.0893941,-0.2790390551,0.2910518348,-0.0030916915,-0.1245509908,-0.1432153434,0.1890479773,-0.341802448,-0.1645838022,0.1209822148,0.0723368898,0.0835640654,0.0051700012,0.1673494726,0.1078059226,-0.2757094502,-0.1325341612,0.2759643793,0.1677877009,-0.0350736827,0.1664510369,-0.1171142608,-0.3932091296,0.2338520437,-0.0699125752,0.3405809104,0.1547214836,0.0269662254,0.0491143912,0.128609255,0.7292755842,-0.2867426872,-0.0499208383,0.1683304459,-0.1148460433,-0.185339734,-0.2699624896,-0.1347569972,0.1397525668,0.0422792584,0.2738830447,0.0090459259,0.0384510718,0.3809129894,0.0371800773,-0.1505873352,-0.3422771096,-0.1247256771,-0.1089220941,-0.1365215033,-0.2156278789,0.1030825302,0.1527048051,-0.26317963,0.0338280275,-0.2128648907,-0.1947259009,0.2898357809,0.2248447239,0.2315249145,0.1369865537,0.1335770935,0.3725126386,0.1532457173,0.2307439446,0.7831414938,-0.250382036,-0.8798404336,0.2520810068,-0.53959167,0.2746089399,0.0672599226,-0.3926186562,-0.0692605302,-0.1908975542,-0.1358876675,-0.1330513805,0.039461799,0.0727600753,0.0298984051,-0.1287193447,-0.5149635077,0.5257356763,-0.0998035297,-0.1812289804,0.0202281531,0.0141944038,-0.3694266975,0.4248115718,-0.055052951,0.7185191512,-0.1543842405,-0.0664678738,0.1947962344,-0.158122018,0.4200713634,-0.2776701152,0.0082029868,-0.3842919171,-0.1745027453,-0.1966463178,0.0137680648,0.1733606309,0.5210571885,-0.1451996416,0.316396296,0.0016985977,0.3645551801,0.0264602937,0.0155003387,0.3741771877,-0.1792826504,-0.1920854449,0.1107749492,-0.1384782791,0.0154473335,0.09687718,0.057058163,-0.3032521009,-0.0505689755,-0.4131678343,0.2384798527,-0.1725191176,0.0442087166,0.1149675846,-0.5480040908,0.5427275896,0.450619787,0.2410171032,-0.0006868581,-0.2381753474,0.0594833195,0.0089390036,-0.0455417782,-0.1769888997,-0.0232021119,-0.0791270509,0.1729516834,-0.2429803759,0.3244460821,-0.0582636669,-0.0837417841,0.220012933,0.2635119855,0.2422812581,-0.432959497,-0.2782201171,-0.2240790725,0.0929488093,-0.1038480997,0.044498153,-0.0798013583,0.0879833475,-0.003703441,0.1256099492,-0.3943379521,0.1475495249,0.2292748988,-0.2091286331,0.0742137134,0.3961915672,0.4418799877,-0.1595523357,-0.1338270009,0.0772403628,0.1888357252,-0.4828431606,0.1160870418,0.1250001937,0.2439948916,0.1087796688,0.2083628178,0.213423565,-0.2776620388,0.1464932412,-0.6198529005,-0.0254824143,-0.1095703393,-0.1404000819,0.0756146088,0.0497044213,-0.449672997,-0.1291389167,-0.0117754759,-0.1885747612,0.1399783641,0.04078063,0.0180434138,0.2311222702,0.2858001888,-0.2057445049,-0.3224655986,-0.0362218991,0.1214752793,0.0656232536,-0.1071390435,-0.2727225721,0.1728659868,-0.0013992608,0.0539030097,-0.2284443229,-0.1858399361,-0.2576567233,-0.1746505648,0.2525277138,0.2778947055,0.0411927812,0.4188427627,-0.0836181715,-0.0894363523,-0.4818773866,0.2702876031,-0.1971144974,0.2839874029,0.1542857438,0.2316162586,0.1593157649,0.1004863009,-0.1338308305,-0.0245897174,-0.0708419681,-0.1876980513,0.3153250515,-0.3347911835,0.2185911834,0.1790135056,0.3079961836,0.3168783188,-0.1215095967,-0.1023305655,0.1719919294,0.0902403668,-0.336694628,0.1967000365,0.0594761111,0.3061063886,-0.0220685676,0.3245767057,-0.0537061766,-0.0796434432,0.0239635371,0.1151189432,0.7046646476,-0.0164931845,0.0524813309,0.0548262522,-0.0494296029,0.1115370169,0.2270132899,0.0380735472,0.0507204011,0.5308746696,-0.1390244663,-0.0151982019,0.247492373,0.3807494938,-0.162783742,-0.6306781173,0.2385840565,0.3995527029,-0.3072297573,0.0190617312,0.1212392598,0.6014146805,0.034966398,0.1163831279,-0.2986671329,-0.0107396357,0.1315867901,-0.103157945,-0.2647881508,-0.2150602043,-0.4387287796,0.2846031189,-0.2169097364,-0.0843455195,0.352935791,0.1971827596,-0.0669354573,-0.5603028536,0.1278917342,0.2119740099,0.1103988066,-0.3089471459,0.1559385508,-0.0493138619,0.0194816682,0.2848858535,-0.0313387103,0.3540910184,0.4284122288,0.069250986,-0.3456064165,-0.1722152978,-0.1174993664,-0.1075449139,0.4433703423,0.0557371117,0.2020919919,0.2329978645,0.0930867717,-0.0080326591,0.3259983659,0.0701494887,0.2354776561,-0.3867930472,-0.1437149495,-0.0548563115,0.0474510565,-0.1470628679,-0.283655256,-0.1698649675,-0.0785293803,0.4186337888,0.1290511191,0.2892831266,0.164312005,0.0040268544,-0.1119125038,0.263640821,0.4633713067,0.1063430831,-0.4908936024,-0.0468468107,-0.7612836361,-0.0230253115,-0.2092337608,-0.1463592499,-0.0102712307,0.2479962409,0.2465246171,0.3274960518,-0.0813509971,0.1213954613,-0.0378901027,0.2988810539,-0.2078569233,-0.1493215263,-0.0753906444,0.0610629953,-0.0357330181,-0.4650059342,0.0714385435,0.3284134567,-0.1604241878,-0.1580476463,0.0090502249,0.0185574051,-0.2787310183,0.3543348312,0.0200481676,0.3844937086,-0.0162355397,0.2808375955,-0.4696329534,-0.2169648856,-0.1940386295,0.2134044617,-0.0509442948,0.4670439065,-0.3124223948,-0.1008952186,-0.2378869057,0.0001280919,0.088023521,0.137953952,-0.3464074135,0.265065223,-0.1035997272,0.202070564,-0.1530989259,-0.1017107368,-0.2005610466,0.1393135637,-0.0791473538,-0.261423856,0.2789075673,-0.3219949901,-0.2220238894,0.035374999,0.1209697947,0.1921243221,0.1258191168,-0.5225005746,0.0423808768,0.3424724042,-0.1380214244,-0.1590197831,0.1584722102,0.1098292768,0.0463948548,-0.2152625322,0.1998657882,0.3526127338,-0.2159950584,-0.2149200886,-0.3720293641]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"This is because load_dataset without `split=` returns a dictionary of split names (train\/validation\/test) to dataset.\r\nYou can do\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', script_version=\"master\", data_files=['test_data.csv'], delimiter=\",\")\r\nprint(dataset[\"train\"][0])\r\n```\r\n\r\nOr if you want to directly get the train split:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', script_version=\"master\", data_files=['test_data.csv'], delimiter=\",\", split=\"train\")\r\nprint(dataset[0])\r\n```\r\n","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":55,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nThis is because load_dataset without `split=` returns a dictionary of split names (train\/validation\/test) to dataset.\r\nYou can do\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', script_version=\"master\", data_files=['test_data.csv'], delimiter=\",\")\r\nprint(dataset[\"train\"][0])\r\n```\r\n\r\nOr if you want to directly get the train split:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', script_version=\"master\", data_files=['test_data.csv'], delimiter=\",\", split=\"train\")\r\nprint(dataset[0])\r\n```\r\n","embeddings":[-0.0920246169,0.0564349107,0.0039141956,0.372161001,0.2419901937,0.2112514675,0.5301389098,0.3948995173,0.2474182695,0.0687352568,-0.1450405717,0.2728796601,-0.1780427992,0.0853817388,-0.2621557415,-0.1945452243,0.0322136991,0.0497310571,-0.3552699387,0.0039880648,-0.3571982682,0.3788069189,-0.3598428667,-0.0790998042,-0.2235179991,0.1674229652,-0.1466455311,0.341891259,-0.0646044761,-0.3069737852,0.4642286599,-0.1027503386,0.4105321765,0.3379975855,-0.0001209265,0.091597721,0.2450603694,-0.1091812253,-0.2504522502,-0.6010637283,-0.0259371456,-0.3921331465,0.3510513306,-0.1411182433,-0.1290298402,0.0570398457,-0.1692730635,-0.3528496027,0.4674951434,0.5159787536,0.1531329453,-0.2059019059,-0.1971256584,-0.0124462983,0.5450271368,0.1082914993,-0.1457669139,0.288138032,-0.0139438249,-0.2046414018,-0.111815691,0.06374906,0.0106420731,0.2190205753,0.1316032112,0.2965677381,-0.2615711093,-0.2017144263,0.0166659206,0.194961369,0.570548296,-0.0731799155,-0.1748564541,-0.1613171101,-0.059986379,-0.2130065262,0.3124711812,0.3574937284,-0.0169693641,0.3381771445,-0.1206576899,0.0709901527,-0.2584077418,0.0994611308,-0.0562138148,0.0654799715,-0.1489541382,0.2812530398,0.2017697394,0.1015310362,0.2514773607,-0.0579952151,-0.0673245639,0.1580315381,-0.6834616661,0.2036809176,-0.2168770581,-0.1526712775,-0.054263711,0.2458858341,0.1760437191,-0.0476836078,-0.0708104372,0.3254227638,0.3645863235,0.0074674669,-0.0258980151,0.4044125676,0.2407534719,-0.1334234476,-0.0388788767,-0.1857776791,-0.2934213877,-0.5016217828,-0.086302571,0.2012825906,0.1611034721,-0.1534703076,-0.2344414443,0.0494668446,-0.3282530606,-0.0160041694,-0.1057873964,0.3072037101,0.0792054757,0.3233473003,0.0002639497,0.2762194574,-0.1499179155,-0.3500113487,-0.0380696207,0.1538398564,-0.4294225276,-0.0667094663,0.4375000298,-0.1740852147,0.1433625072,0.1928385347,-0.0566435307,-0.3830625713,-0.0729662925,-0.3045082986,0.127741605,0.2792576253,0.1017231271,0.1716285944,0.2939251661,-0.1996124834,-0.0871599913,0.2777005136,-0.4890449941,-0.2541227639,-0.0913462788,0.1075182632,-0.2549365759,0.1419746131,-0.3131116331,0.0468015783,0.0950297117,-0.1052417681,-0.1675722599,-0.3225544095,-0.3190791607,-0.2440020442,0.0466859564,0.3962135911,-0.7064715028,0.1609789878,-0.0436272025,0.0166498329,0.1765231192,0.2286920249,-0.2605005205,0.0781636834,-0.2938950658,0.1257948726,0.2624050379,-0.21647802,-0.3129369318,0.2795172334,-0.0455346778,-0.0195158757,0.1386790872,0.0048780697,0.4192972481,0.0327761024,-0.1222806796,0.2695268393,-0.0665015802,0.0513628274,-0.2043424994,-0.2208314836,0.230395481,0.2533374131,0.09017317,0.064632073,0.2520318627,-0.4523490071,0.4311442971,-0.0185215902,0.2009960264,0.0600047223,0.2045498639,0.0673958138,0.2995598316,-0.0448753238,-0.5501888394,0.1751102954,0.4230570197,-0.1836449653,-0.1712397188,-0.0816116408,-0.4569071531,-0.0527785569,-0.2632133961,-0.0353595838,-0.0295918472,-0.0453094617,-0.0110517554,-0.1825748086,-0.1791177988,0.4217712879,-0.261186868,0.1721967459,-0.3425503671,0.3586449921,0.1426158845,-0.0517234728,0.1552652717,0.0561741702,0.2390625179,-0.1527739614,-0.1483723372,0.3356363773,0.4766589999,-0.0618811175,-0.0393589959,-0.2567850351,0.1200383306,-0.2125388384,-0.0765184611,0.2014890015,0.266081214,-0.0892454609,-0.4791173935,0.3983052075,-0.3870799243,0.3109741211,-0.1254491508,0.1998694092,0.241813615,-0.1007643268,-0.1040568799,-0.0747825727,0.0224811137,-0.0448641144,0.2591088712,0.0762873888,-0.1619548649,-0.1972352862,0.3197233379,0.1151808053,-0.2512791157,0.1153136343,-0.0970705897,-0.0526020266,0.2553263307,0.388995707,0.4357036948,0.083102718,-0.2003983855,0.0272413809,0.1205655038,-0.2420069575,0.1677270532,0.0737355798,0.0803622305,0.2283989042,0.1371799409,0.0185588915,-0.1221626997,-0.1420411319,0.2180795521,0.1531413943,-0.5580414534,0.0672471225,-0.2492441237,0.3542293012,-0.3054797351,-0.088093169,0.1177534387,-0.3376854062,-0.2673149109,0.2258105576,-0.1555700451,0.0849401802,-0.1968522966,0.168894276,0.0460411496,-0.2257519066,0.127748847,-0.2152117789,-0.262229979,0.0265395846,0.0546971373,0.0048142457,-0.0604922995,0.0089467121,-0.0171862077,-0.1233489215,-0.0182160083,0.1867986768,-0.070273146,-0.0006798789,0.1853316277,-0.1570403576,-0.0293877814,-0.2066927999,0.1623373479,0.077324219,0.056724295,0.3653017879,0.293117404,-0.0645011738,-0.1795730293,-0.7377924919,-0.2892294824,-0.2201963365,0.0692026615,-0.1671292484,-0.0350531973,0.0936679617,0.2560247779,-0.0321816392,0.2291259617,-0.1719628274,-0.1385293454,-0.1858995855,0.5807957649,-0.2469512373,-0.2713826299,0.2195544392,-0.1315249503,0.007083829,0.3676760197,-0.4218454957,0.2262401134,-0.0697562769,0.0749832019,-0.0520796403,0.0243040752,0.0830063075,-0.0699172094,0.0978282839,-0.0387605727,-0.1052255258,0.2508596182,-0.0496933125,0.1034532785,0.1427065581,0.5326666832,-0.0993343964,0.8147948384,0.0513601415,0.059384767,0.1929327548,-0.1644670963,0.2508680522,0.1152941734,-0.2596244216,0.0838407576,-0.1691666692,-0.3220829964,0.3516044021,-0.0214848425,-0.152440846,-0.0998279303,0.2590567172,-0.2020308226,-0.2619891167,0.1752263159,-0.1171213761,0.029884927,-0.0312765054,0.2175438404,0.1629425585,-0.2070378661,-0.0227658134,0.1213583499,0.1313208938,-0.1712009311,0.0724048615,-0.1677563488,-0.2650647461,0.1700244546,0.0065876897,0.3186220825,0.1230911165,-0.1427007467,0.0781695992,0.2231677175,0.8126674294,-0.1891089231,-0.0012189214,0.1494552791,-0.075190194,-0.1324187368,-0.3159856796,-0.1486232728,0.2276510596,0.0883170813,0.2510455549,-0.0496397763,-0.010330664,0.4505234063,0.0838636681,-0.2074792236,-0.3661195636,-0.1562462747,-0.0125673898,-0.0931672603,-0.1308235973,0.1905769557,0.1685333848,-0.3185512424,-0.0054584728,-0.1650790721,-0.2574357092,0.3499359488,0.1709162891,0.2823196054,0.0279283077,0.2067068219,0.2287578434,0.1520322859,0.194015339,0.7980725765,-0.2726474404,-0.8757582903,0.2597289979,-0.5026088357,0.4005541205,0.0253318585,-0.3638646901,-0.1037124246,-0.1918034256,-0.0526365601,-0.1298295408,0.0165172759,0.0891316831,-0.0678077191,-0.2455695271,-0.5857750773,0.4648813605,-0.0820494965,-0.0724826381,0.0997197628,-0.0700680465,-0.3810084164,0.5269698501,-0.1301181018,0.6978643537,-0.094052285,-0.0197088774,0.1578088403,-0.2015551627,0.3961661756,-0.3102938235,-0.0145095233,-0.4537198246,-0.160011515,-0.1839677095,0.0067586116,0.0700714663,0.5914731026,-0.171021238,0.3970745206,-0.1107248217,0.4428042173,0.0041380678,0.1333013177,0.2929831743,-0.1906875372,-0.1368488222,0.0773172602,-0.0579028428,-0.0676557422,0.1529643983,-0.0103089483,-0.2295545042,-0.0298130494,-0.2959817648,0.3327460885,-0.0851014704,0.0975557491,0.0962044671,-0.518379271,0.5273531079,0.440700084,0.1441206187,-0.0339886621,-0.1290064007,0.134818092,0.0089673866,0.0739119574,-0.0905188993,-0.0146647627,0.0735804215,0.2807483375,-0.1852082908,0.245151639,-0.0977552608,-0.0829650536,0.3111159205,0.3133254647,0.2825339437,-0.586096108,-0.2721990347,-0.208545059,0.1540906876,-0.1515071392,0.0469689742,-0.118547976,0.0065012667,0.0510134958,0.0078975977,-0.3608900607,0.121984683,0.1970485598,-0.0657681525,0.112635985,0.4779486358,0.4211015701,-0.0839853436,-0.189010486,0.101125218,0.1984159052,-0.6184692979,0.1737714857,0.0567198955,0.2487824559,0.0968458503,0.1344331801,0.1138446704,-0.169321999,0.0602245368,-0.5078524947,-0.0251084976,-0.0420170203,-0.1369598657,0.059743017,0.0674264058,-0.2848598957,-0.1618444026,-0.0702818558,-0.1783937067,0.1401689351,0.076066263,0.0713703632,0.1139593646,0.222956717,-0.1100144833,-0.2800535262,0.0251350682,0.0276515055,0.1282836944,-0.1222079769,-0.2492934018,0.1729645729,-0.0203151517,0.1893836558,-0.3006106317,-0.1253087372,-0.2562193274,-0.1848993897,0.2022196054,0.2780593336,0.0276399851,0.4822484255,0.1122752279,-0.0452574901,-0.5738332868,0.2749340534,-0.2892035842,0.2634687126,0.1305773407,0.2388586849,0.1035437658,0.0980132893,-0.1666341722,-0.0499367043,-0.1110425219,-0.1527338326,0.2965564728,-0.3347997367,0.1624127775,0.1323401183,0.4000544548,0.4437752068,-0.1005393043,-0.1168725118,0.1531316638,0.124657616,-0.272716254,0.2014936358,0.1138086095,0.2503403723,-0.0640137643,0.3176334798,-0.0359446555,-0.0939216018,-0.0171212144,0.1545409113,0.6070341468,0.0804748237,0.2360871136,0.4467528462,0.011463454,0.1598700136,0.2003496587,-0.147692278,0.0838204846,0.5120176673,-0.1648634821,0.0128322085,0.3407443762,0.1974021047,-0.0812954605,-0.5081665516,0.1211583987,0.2955353856,-0.362416774,-0.0714387,0.0745174363,0.7061758637,-0.0843044147,0.0414132997,-0.3371932209,-0.0668341219,0.0725576878,-0.0509174727,-0.1879673749,-0.1644283831,-0.5028641224,0.3112888634,-0.1561876833,-0.0690337867,0.4760004878,0.2167230546,-0.0086945193,-0.433771193,0.0320503414,0.1697372347,0.1164811328,-0.3514192998,0.0458693504,-0.0698429719,0.0414159484,0.3675315678,-0.2498469651,0.3425768912,0.352757901,0.0092108101,-0.3907544315,-0.0755686089,-0.0503061563,-0.0214369297,0.4297500849,0.001763513,0.3196757734,0.2829744816,0.0923240557,-0.0092783188,0.2641902864,0.1017781273,0.1555731595,-0.4826744199,0.0886415318,0.075677909,-0.079569757,-0.1519210637,-0.2095644474,-0.1678600162,-0.2077058852,0.3991562724,0.0987178609,0.2374105453,0.1570792049,-0.0125420773,-0.1016106382,0.2418651432,0.3471136987,0.0301855989,-0.4722242355,-0.023413077,-0.6675115824,-0.1041138619,-0.1713967174,-0.1148777828,-0.0195009597,0.2386513948,0.2016133219,0.2796035409,-0.0930212438,0.1285197139,-0.0175936278,0.3582010567,-0.1670179963,-0.1782082915,-0.1963286251,0.0773874968,-0.0550286211,-0.4330329001,0.0000825173,0.2998559773,-0.1908219904,-0.0665819421,-0.1808104962,0.1633932889,-0.2126727104,0.2939324379,0.0215848722,0.4189729691,0.037959028,0.3084060848,-0.3747854531,-0.2793860435,-0.2077226639,0.2350733429,-0.0631015971,0.438677609,-0.3243896663,0.0658723786,-0.207908079,0.0394379832,0.0888457596,0.1680847853,-0.3443849385,0.1940704733,-0.0793132931,0.2522798777,0.0013850692,0.0767941251,-0.2081237733,0.1591545939,-0.0463752523,-0.3086296916,0.2655185163,-0.3227376044,-0.2894771099,0.0198449399,0.1921177804,0.2426323593,0.1170973331,-0.5216588378,-0.0667871609,0.3476248384,-0.1688327044,-0.1139031127,0.1149121672,0.1570246667,0.0978991464,-0.1428449601,0.201773718,0.3349767029,-0.3364254832,-0.0670104623,-0.3840115368]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Good point\r\n\r\nDesign question for us, though: should `load_dataset` when no split is specified and only one split is present in the dataset (common use case with CSV\/text\/JSON datasets) return a `Dataset` instead of a `DatsetDict`? I feel like it's often what the user is expecting. I break a bit the paradigm of a unique return type but since this library is designed for widespread DS people more than CS people usage I would tend to think that UX should take precedence over CS reasons. What do you think?","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":89,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nGood point\r\n\r\nDesign question for us, though: should `load_dataset` when no split is specified and only one split is present in the dataset (common use case with CSV\/text\/JSON datasets) return a `Dataset` instead of a `DatsetDict`? I feel like it's often what the user is expecting. I break a bit the paradigm of a unique return type but since this library is designed for widespread DS people more than CS people usage I would tend to think that UX should take precedence over CS reasons. What do you think?","embeddings":[-0.0537754819,0.1033459008,0.0190047286,0.3612280488,0.1807834059,-0.0042513018,0.5432186127,0.222499162,0.3372097313,-0.003618543,-0.0237615574,0.3069189489,-0.0317640789,0.118917942,-0.2025142312,-0.0749718025,-0.012215076,0.1207054928,-0.3683388829,-0.0174310263,-0.4309170544,0.292850107,-0.2213966399,-0.0585704185,-0.1151770204,0.1571623236,0.0017044473,0.3660930991,-0.1084838063,-0.257276535,0.6280568838,0.0230042394,0.4292282462,0.2245847285,-0.0001190717,0.0388116241,0.3117853403,-0.1329061836,-0.2897416055,-0.5877243876,-0.027259687,-0.3956184685,0.3377561569,-0.1040745974,-0.0616714507,0.172775656,-0.1652778685,-0.5165011287,0.3123306334,0.3730696142,0.1405084878,-0.2834521234,-0.2309973836,0.0034147301,0.5908178687,0.2067524493,-0.1971361637,0.2810762227,0.1669810265,-0.1015687287,-0.1598022133,0.083155334,-0.1043233722,0.1894381344,0.29018116,0.2623083591,-0.2809916139,-0.11749015,-0.0169992391,0.1899370998,0.8341629505,-0.0735759661,-0.2151966095,-0.2079991996,-0.0091483137,-0.1942547411,0.2959364951,0.224168852,-0.0256013237,0.2741404772,-0.039895419,0.0002723018,-0.2295822501,0.0643773004,-0.1193104833,0.0102746459,-0.1569630653,0.3779331744,0.0456255935,0.1549314559,0.3429839015,-0.1116522849,-0.1162602082,0.0835307017,-0.567054987,0.1575641781,0.0551394075,-0.1592389643,0.0174478833,0.3070752323,0.3632949889,0.0668959916,-0.17210491,0.2512286603,0.434494853,0.0817990974,0.0020746593,0.2576614916,0.3648816347,-0.2776294947,0.013783399,-0.1647897661,-0.1719340533,-0.2707709968,-0.0575214885,0.1610938907,0.1682668328,-0.1684322953,-0.2216254175,0.0732909441,-0.2644221187,-0.0547595546,-0.1519445032,0.2221849561,0.1522415429,0.5024954677,0.0156444684,0.1761675328,-0.0987721533,-0.3069808483,0.0024105534,0.0743475333,-0.4379191697,-0.0022643732,0.4775822163,-0.1238090396,0.0213834923,0.1910146922,-0.146714136,-0.2925012708,-0.0691041574,-0.3121488094,0.1865879744,0.2201448679,0.0137046082,0.2126200795,0.204693377,-0.2328411341,-0.2364430428,0.3585527539,-0.536218226,-0.3046457469,-0.0398351848,0.0985862017,-0.2729690969,0.2291877419,-0.3323957026,0.1283838004,0.0017088264,-0.1008917093,-0.0892250985,-0.3063386679,-0.2946177125,-0.33223176,-0.0385905392,0.5090972185,-0.8320771456,0.2264830768,-0.0860439837,0.1219893396,0.0956524983,0.344486624,-0.3025426567,0.096642442,-0.3494292498,0.0400712192,0.3277105987,-0.2119710743,-0.3828731477,0.2315531373,-0.1052042544,0.0784768239,0.164193958,0.1649427414,0.444454968,-0.043529626,-0.0862467512,0.2531380653,-0.1244701743,0.0795064718,-0.2437729388,-0.2483042628,0.2499010712,0.170079872,0.1038125083,0.0540584698,0.1944538951,-0.4752427936,0.358985275,-0.0931508318,0.1257806569,-0.1334753633,0.3010072112,0.0260031652,0.2912549078,-0.0078927856,-0.6777339578,0.1792105287,0.4068228602,-0.1115447357,-0.0382764116,-0.1820578724,-0.3484876454,-0.0155588202,-0.169530496,0.0278336145,-0.0358911566,0.0391564444,-0.1085351482,-0.1549387872,-0.2077755928,0.4025168419,-0.2128464878,0.0848495588,-0.3851655424,0.3566335142,0.2739728093,0.0097316541,0.1406469643,0.1305159926,0.2248567194,-0.1097731665,-0.0581612885,0.2158618718,0.357062906,-0.0946864858,-0.0116337584,-0.1943476945,0.0966617912,-0.1202944145,-0.0475928932,0.2353375703,0.3100091517,-0.165341869,-0.5367754698,0.4306187034,-0.4352806211,0.2406532615,-0.1465645581,0.0563365296,0.1764354557,-0.0112934625,-0.1437526196,-0.0274792295,-0.1477054954,0.0337617807,0.2478099763,0.1867637783,-0.1355113536,-0.0533684976,0.3618720472,0.1317256093,-0.1845165491,0.2107610554,0.0488578379,-0.237596035,0.3247870803,0.3543029428,0.6315694451,0.095159784,-0.1953668892,-0.1066858545,0.0825454518,-0.2646082938,0.1412470192,-0.0092707984,0.0298803374,0.2075575739,0.1890493482,0.0394458137,-0.1802781969,-0.1906566024,0.0667323321,0.0999250039,-0.5104933381,0.0682827085,-0.1229088902,0.2668240666,-0.2863724232,-0.0286341738,0.0715720952,-0.3165184855,-0.2467340976,0.1903038323,-0.1523376703,0.0006298273,-0.1428436786,0.3498927057,-0.0687937289,-0.2591992617,-0.0169770215,-0.1245573983,-0.271569401,0.0175173245,0.0291548818,0.0485817418,0.0630102456,0.0337894745,0.0646716133,-0.2758508325,-0.1107280478,0.2082014233,0.027093282,0.0583995543,0.3743036985,-0.1855602264,0.0611077026,-0.1489398926,0.1078897193,0.1560414881,0.1085718945,0.2716521621,0.1851983666,0.0015301723,-0.2390160114,-0.7112779617,-0.1844305545,-0.2410863042,0.1508251876,-0.1209753156,0.0023745727,-0.0167912655,0.1608787179,-0.1565036476,0.2552342117,-0.0619459786,-0.1892142594,-0.2331595719,0.4649218023,-0.1569713652,-0.2644332051,0.1217835993,-0.1825101972,0.1514082551,0.4018861651,-0.4829933643,0.1828479171,0.0471073501,0.1323951334,-0.0508755594,0.14494133,0.0866822302,0.0103522362,0.1132815108,-0.0133335805,-0.1951068789,0.2277120352,-0.0892839804,0.0242395401,0.2536148131,0.3969816864,-0.138894558,0.705930233,0.0036074312,0.015542306,0.2021128535,0.0485987701,0.4244733751,0.0127241453,-0.1967828274,0.096319139,-0.1831023842,-0.3161441088,0.3190335631,0.0838498101,-0.1625922769,-0.1532451808,0.2083591074,-0.1835206151,-0.3161172867,0.2805400789,-0.1011907309,-0.117179513,-0.0002713483,0.2491073757,0.1046134233,-0.1490245759,-0.0405976586,0.2144348621,0.2860654294,-0.1061940417,0.0506759882,-0.2143294364,-0.312066257,0.151555717,-0.0513514392,0.3003502786,0.2327192426,-0.1304375529,-0.0041040815,0.1650894582,0.6726489067,-0.2029227763,-0.0361734368,0.1090754569,-0.0360408835,0.0532151088,-0.1702297479,-0.0146458084,0.230233863,-0.0576607473,0.217935577,-0.1228799447,0.0076648942,0.3096750677,0.0963699743,-0.2221122235,-0.4531788826,-0.1049963161,-0.0549614504,-0.0579492636,-0.1739383936,0.0660612285,0.0278715435,-0.3355501592,-0.1227798462,-0.2196180522,-0.2645541728,0.3511791229,0.1359793246,0.1844701618,-0.029492978,0.1564248949,0.1462995112,0.1883157343,0.2318917215,0.7988601327,-0.2999086976,-0.7502194643,0.2594363987,-0.4703613818,0.3502209187,0.1668519825,-0.3451269865,-0.0672018528,-0.3084926307,-0.1468196809,-0.1964682937,-0.1515558511,0.0179830864,-0.069571577,-0.3025731742,-0.660019815,0.5407512188,-0.2247382253,-0.1652632058,0.1719543487,0.084384501,-0.4312987626,0.4897021949,0.0409175046,0.800549686,-0.1704283059,0.0720024779,0.1085456461,-0.2047685832,0.600874126,-0.276424408,-0.1295630038,-0.5226104259,-0.0656086206,-0.2052313983,-0.0517630018,0.0951955467,0.5655970573,-0.1754756421,0.4081784487,0.1344902515,0.4029865265,0.1213078052,0.2673028111,0.2354559898,-0.2553951442,-0.2133433223,0.0296826493,-0.118363589,-0.0844964534,0.0401427485,0.0512280427,-0.1784121096,-0.1373188347,-0.2695530057,0.1262074113,-0.0577276088,0.0150859775,0.0499325469,-0.6408004761,0.5324746966,0.3608645499,0.0921528861,-0.0202020034,-0.1526842266,0.0340645015,0.0920539051,0.2161552906,0.0190283861,-0.0949911252,0.0825951919,0.2941956818,-0.1607955992,0.2153991461,-0.141071856,-0.0050880411,0.196378991,0.3348126709,0.3393918276,-0.5301561356,-0.2820550799,-0.1227415949,0.2097257078,-0.0928211883,0.0123441694,0.1200500652,0.2153827101,0.0124088731,-0.0287131146,-0.3354755044,0.000258995,0.0358208939,-0.0057812487,0.0289145652,0.1569942087,0.2840822041,-0.1687953025,-0.1051959023,0.1372450441,0.1519786566,-0.6934360266,0.1855424494,0.1447969973,0.0985500589,0.0743139759,0.2256253362,0.1408829987,-0.1747946739,0.0635985807,-0.3946326673,0.0747546703,0.0736219883,-0.0600177236,0.0922558531,0.079018198,-0.4209208786,-0.1307505816,-0.1268936545,-0.1906572282,-0.0081958808,0.1432707161,0.0891461819,0.0176287889,0.309777379,-0.2794101834,-0.2250023335,0.0017567994,0.0087609012,0.0047906265,-0.0894954279,-0.1914918274,0.2107375562,0.0226567518,0.2284318209,-0.1968971491,-0.2100581229,-0.1015769988,-0.2319881916,0.2390558869,0.3404797912,-0.0775873587,0.3745772541,0.0658349246,-0.0109581826,-0.6506854892,0.4418027699,-0.3084153831,0.3093321919,0.0754541904,0.3467370272,-0.0001511069,0.1098976508,-0.0824043825,0.0565916188,-0.0377752967,0.0474769436,0.312366128,-0.2570691109,0.2241755575,0.17159006,0.3068363369,0.3681269288,-0.1031856537,-0.1775368601,0.0890467912,0.1088251993,-0.2263747156,0.2014872283,-0.0887983516,0.2183039039,0.0073704841,0.4874252677,-0.0458956361,-0.1511247456,0.0754916295,0.0471836105,0.7774942517,0.058211863,0.143811807,0.3218049109,-0.1171600893,0.2135417461,0.2073939741,-0.0632122159,0.0398133025,0.5263146758,-0.1259410083,0.0107296221,0.377314955,0.2663340867,-0.0828818306,-0.4553400874,0.2748794258,0.3771361113,-0.2825215757,-0.0846111998,0.2595617473,0.5131548643,-0.1591779739,0.1302594841,-0.1870142221,-0.2644057572,0.0163356159,-0.0597800165,-0.1930786967,-0.2350784987,-0.3509510458,0.2641786337,-0.3510676026,-0.0516301095,0.3634466529,0.1875218898,0.082823433,-0.4788799882,0.1263748556,-0.0500871353,0.0921473056,-0.288184911,0.1098938435,-0.0134696243,0.0782679245,0.3593219817,-0.2522948384,0.2834373116,0.4704683721,0.1390377134,-0.3918017149,-0.1036396772,-0.0271543004,-0.216751948,0.4326716661,-0.0285319034,0.2285542935,0.2073419094,0.1219134852,0.0160052404,0.3366029859,0.1267327666,0.0241187941,-0.5890341997,0.0372147821,0.0035395634,-0.1234009117,-0.0259874146,-0.0557220317,0.0087389369,-0.2356420308,0.3753245175,0.1757012755,0.1725266427,0.0632754192,0.015184978,-0.0169149451,0.095560208,0.2717058957,0.1832728237,-0.4492387474,0.0707439184,-0.7308097482,0.0228078272,-0.0885443464,-0.0567186959,-0.0234387498,0.2650984228,0.3051563203,0.2791956365,-0.1910183877,0.256103009,0.0036744319,0.2863952518,-0.0526523367,-0.3129012883,-0.2451200187,-0.0457745194,-0.1484946162,-0.5657708645,0.0250716619,0.4132972062,-0.1882381737,-0.2094072551,-0.0156296473,0.3124324381,-0.1002548262,0.3512981832,0.0730804056,0.323829025,-0.0007769538,0.1384551525,-0.3173959553,-0.231896922,-0.2114670873,0.2685054541,-0.0895991176,0.4095950425,-0.2896968126,0.121375449,-0.2343211174,0.0011415348,0.0855453089,0.2449109405,-0.5026375055,0.2409135848,-0.2224508077,0.1260558963,-0.1578632295,0.057847932,-0.1234941781,0.0754875988,-0.0371279307,-0.1940155029,0.2651410401,-0.1993717551,-0.205096513,0.0285416618,0.2436950356,0.1691824645,0.0668128207,-0.6058081388,-0.1650235802,0.376796633,-0.1661026329,-0.1423831582,-0.0075519374,0.0256169494,-0.0475918576,-0.225037694,0.2251909375,0.2840964496,-0.2430257499,0.048657354,-0.449652046]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"In this case the user expects to get only one dataset object instead of the dictionary of datasets since only one csv file was specified without any split specifications.\r\nI'm ok with returning the dataset object if no split specifications are given for text\/json\/csv\/pandas.\r\n\r\nFor the other datasets ton the other hand the user doesn't know in advance the splits so I would keep the dictionary by default. What do you think ?","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":73,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nIn this case the user expects to get only one dataset object instead of the dictionary of datasets since only one csv file was specified without any split specifications.\r\nI'm ok with returning the dataset object if no split specifications are given for text\/json\/csv\/pandas.\r\n\r\nFor the other datasets ton the other hand the user doesn't know in advance the splits so I would keep the dictionary by default. What do you think ?","embeddings":[-0.0446914434,0.0243465658,-0.0048847762,0.419436276,0.1463128179,0.1458506286,0.4632930756,0.290150255,0.3439593017,0.0506453626,-0.1350136995,0.2295309901,-0.1126452386,0.0743184164,-0.360443294,-0.0890146419,-0.0051947599,0.1351227313,-0.3674571514,-0.0321296565,-0.3761761785,0.3206699491,-0.2852636576,-0.0821710601,-0.1697315127,0.1815558225,-0.0694450289,0.2880131602,-0.1186519191,-0.3370015621,0.5611509681,-0.0190644655,0.4348040819,0.1747192889,-0.0001195433,0.0651759878,0.2265417129,-0.1385782957,-0.2591245472,-0.5547061563,-0.0865493342,-0.495644033,0.3576343954,-0.1113045886,-0.149463743,0.0304121636,-0.092438072,-0.4187635779,0.3667915463,0.4456266463,0.1584687233,-0.252972573,-0.1881751269,0.0030297169,0.5168049932,0.1237124205,-0.1239998788,0.2595596015,0.07364849,-0.1548053473,-0.1259088963,0.0728909448,-0.0548554771,0.2917454839,0.1838179678,0.2202951759,-0.2079448551,-0.2291031927,-0.0040888102,0.2079269141,0.7277753949,-0.0347440019,-0.2305931747,-0.1744337529,-0.0231071748,-0.1313573569,0.2728374004,0.3152931035,-0.0410876498,0.3006059229,-0.0826784894,-0.0083451057,-0.2871993184,0.1252717674,-0.0235241633,0.0793898851,-0.1653632522,0.3553797901,0.0618962757,0.0458485223,0.2907164991,-0.1185568422,-0.0849388465,0.1433923692,-0.6625823379,0.1903320253,-0.0949976817,-0.0448507927,-0.0221889708,0.1342255622,0.2266720384,0.0177782178,-0.1208943725,0.2445975244,0.4031658173,-0.0086480612,0.0469429009,0.2923464775,0.2768140733,-0.1933324337,0.0447108857,-0.1059078574,-0.2271493226,-0.4162047207,-0.1200387925,0.1544906497,0.2859624028,-0.2022923529,-0.2549926341,0.0759332776,-0.2841730714,0.0128482683,-0.1795963049,0.2437569499,0.0655156821,0.3408356309,0.0492644645,0.2859683633,-0.1149515957,-0.4505905211,-0.0125414655,0.14165245,-0.3554002047,-0.0124977939,0.4809399843,-0.1329941303,0.1060488895,0.2314292192,-0.1102559716,-0.3242432177,-0.0325178541,-0.3137765229,0.086547032,0.2456090599,0.1420630366,0.2350402325,0.3222496808,-0.2917710841,-0.1687238514,0.3128437698,-0.4632062316,-0.2403516769,-0.0398925282,0.0929584056,-0.3021011055,0.1972524822,-0.3214191496,0.1838947833,-0.0014772764,-0.1085455567,-0.2330001444,-0.3334562182,-0.3624798954,-0.2817545235,-0.0565882735,0.5261699557,-0.8061848283,0.2232605368,-0.0182995908,0.0256729871,0.0277712848,0.2435151339,-0.2859251797,0.0788338631,-0.3219588399,0.0521328188,0.1771896929,-0.2237006724,-0.3795720041,0.20960401,-0.0132211782,0.0899483413,0.2062482387,0.0760797858,0.4837306738,-0.008639507,-0.1743772477,0.1678628772,-0.0659872517,0.0446666256,-0.209931016,-0.2436727434,0.2614076138,0.1777191311,0.1183044463,0.1183997244,0.1923597604,-0.5197708011,0.3524926007,-0.049706392,0.2328709215,0.0597576723,0.2039584666,0.041094061,0.2808486521,0.0149374371,-0.6455548406,0.2398843914,0.4423056245,-0.1205580533,-0.0816649571,-0.1694549769,-0.399446249,-0.0598017201,-0.215688929,-0.0491334386,-0.002058561,0.0483992435,-0.0326792225,-0.1884748191,-0.103998892,0.4660894871,-0.2199749053,0.0929078087,-0.3543525636,0.3410103023,0.1808355004,0.0232977346,0.2312521487,-0.0563939437,0.2629020512,-0.1184227988,-0.0771762505,0.1964529604,0.3999143541,-0.0858183354,0.0429502577,-0.1645276994,0.0793244317,-0.2305038571,-0.0083328858,0.2045976818,0.3715140224,-0.1554052681,-0.5231380463,0.4438102543,-0.3133574724,0.2376755029,-0.1666315496,0.1455672979,0.2703783214,-0.0266004987,-0.1132093519,-0.0289892871,-0.1073944792,-0.0079468358,0.2660500109,0.149736166,-0.2108669132,-0.1419972479,0.2749673426,0.2318961471,-0.2266092896,0.1718943566,-0.0651103705,-0.1108972356,0.2969529331,0.3253509998,0.4923043847,0.1208926886,-0.258210659,-0.0712934509,0.1183274984,-0.2021027356,0.0870299265,-0.029579306,0.1463160068,0.2104919702,0.1246300042,0.0422474183,-0.0960944816,-0.1743089408,0.1943690032,0.1912709177,-0.5371576548,-0.0354342572,-0.2490179837,0.245905146,-0.2650808096,-0.0541447215,0.0794021413,-0.2963488102,-0.2862538397,0.2433870435,-0.1253311038,0.0655803606,-0.1074497625,0.2319077104,-0.0061182091,-0.1723755747,0.1206972823,-0.1466157883,-0.2509982586,0.0133886011,0.0352111459,-0.0364192165,-0.0454987101,0.0550033152,-0.060106609,-0.2914685905,0.0006237893,0.189397648,-0.0112397987,0.0695458502,0.2390502989,-0.0659901947,0.0432222076,-0.1368498504,0.1452718526,0.2191331387,0.0598854646,0.2933091223,0.2413597256,-0.0689229891,-0.1228289679,-0.6996182799,-0.2679125071,-0.1816014498,0.0728140697,-0.1040656343,-0.0151293864,-0.0078915935,0.2856502533,-0.1130621657,0.2407524437,-0.1150324047,-0.1243783459,-0.1990409195,0.5246908069,-0.1994052529,-0.323693186,0.212506488,-0.1198170409,0.0817697495,0.3567140698,-0.4152070582,0.2089580148,-0.0329122841,0.0867071897,-0.0653101578,-0.01460349,0.1049970016,-0.0258057155,0.1222163886,-0.0642417967,-0.1947111785,0.2693861723,-0.0943278521,0.0281988755,0.1903737932,0.4873292446,-0.1435077637,0.7951076627,0.0928443074,0.04721478,0.201102823,-0.0355097242,0.3330549598,0.1751451194,-0.2264029682,-0.0061438312,-0.1244723499,-0.3189521134,0.3500702977,-0.0268965531,-0.1173555329,-0.1424244046,0.1945772618,-0.235439539,-0.2284532934,0.1594709605,-0.1251471937,-0.0245742016,0.0497913025,0.1937774122,0.1017239019,-0.1457491368,-0.1344631314,0.1896149516,0.191930294,-0.044725962,0.1049398109,-0.1503963172,-0.2751072049,0.2131285071,-0.0694784597,0.3745324016,0.1689337343,-0.0444944128,0.0118110655,0.1957569867,0.7295382023,-0.2287737131,0.0030460856,0.1243919879,-0.0855473652,-0.1337897629,-0.3158795536,-0.0860540792,0.298548162,0.001284605,0.2565813661,-0.0799904093,-0.0319196172,0.3263701797,0.1161973849,-0.2338622659,-0.3317708969,-0.1265914142,0.0044454644,-0.1254691929,-0.1931747496,0.1252732873,0.0521134324,-0.3080663979,-0.0141283413,-0.214544937,-0.2291708738,0.3313292861,0.1550188512,0.2398004085,0.0414591841,0.2401252538,0.1598702967,0.1619763523,0.2585996091,0.8187352419,-0.2255117297,-0.9100395441,0.2532353997,-0.4812318683,0.3621435761,0.0518251061,-0.3327270448,-0.0682387128,-0.2888256609,-0.1223362163,-0.0978192315,-0.092101872,0.1060933396,-0.0652477741,-0.2792618573,-0.5379041433,0.5231564045,-0.1839359999,-0.1688441634,0.1658743322,0.0635139123,-0.4008098841,0.4619475007,-0.0779626444,0.6648093462,-0.143561393,-0.02632696,0.1368159205,-0.1935030818,0.5352106094,-0.2681345046,-0.0582276136,-0.4784962535,-0.1556538194,-0.188233912,0.0233326927,0.1336218417,0.6201779246,-0.2137886435,0.4137539864,-0.0305569544,0.3697021604,0.0277275052,0.2698028088,0.2819107771,-0.1970161945,-0.1146226972,0.0792973265,-0.1219548061,-0.0136859175,0.0405084714,0.0256664157,-0.2784499526,-0.1080828086,-0.3273004293,0.2571620047,-0.0309282057,0.0844721571,0.0456776954,-0.4729815722,0.5333228111,0.4806918204,0.0964348763,-0.0579727292,-0.2122698426,0.0419796556,-0.0075088814,0.070980683,-0.1156744286,-0.0343130901,0.0392010026,0.2451462895,-0.1931064278,0.2969657481,-0.1279441267,0.029141672,0.2356510013,0.3048042953,0.2445830107,-0.558244586,-0.3475918174,-0.1782844663,0.1838510782,-0.0923165679,0.0286441762,-0.0112061938,0.1078994051,0.0184130706,0.0635010526,-0.3889575899,0.0707364529,0.1621206105,-0.0464439653,0.1069690138,0.4289334416,0.3648098409,-0.1255620122,-0.1633298397,0.1097066402,0.0889374688,-0.5510398149,0.1483136415,0.0768055245,0.2905543745,0.0877697989,0.1595234871,0.1559222341,-0.239258498,0.1204695478,-0.5238209367,0.0897693634,-0.0340383761,-0.1151964515,0.0711519569,-0.040878363,-0.3904766142,-0.0813329816,-0.0195437409,-0.183399722,0.0560255162,0.0836377591,0.0804499835,0.1844597161,0.4139981866,-0.2509410083,-0.253223002,-0.0012747335,-0.0102723958,0.0814380422,-0.0961059481,-0.2050713599,0.1875602901,0.0344714224,0.2173551768,-0.2460440546,-0.2080379128,-0.1947423518,-0.1506970525,0.154033497,0.3942095637,-0.0514224246,0.4527197778,0.1114693284,-0.0255170781,-0.5381059051,0.3280771673,-0.2612595558,0.2569449842,0.103313081,0.2795568705,0.1250805706,0.0944782868,-0.0898012593,-0.0122042652,-0.1515176892,-0.1182865798,0.2566821873,-0.2914991379,0.2917625904,0.1998850703,0.3145687878,0.3723840415,-0.1177826896,-0.0988317207,0.109185189,0.1090015844,-0.3487354219,0.1822365224,0.1090171784,0.2419047356,-0.0584848858,0.414698869,-0.0443696007,-0.095819585,0.1168787628,0.131759122,0.6230407357,0.1043817103,0.1987516284,0.2564282417,-0.0296462085,0.1116351336,0.2250181437,-0.0835401192,0.0890388042,0.472912699,-0.2165098339,0.0019182889,0.2744267285,0.1816958785,-0.0412162095,-0.6070023775,0.2713732421,0.3079600334,-0.3180388212,-0.1184831932,0.1882697046,0.5466785431,-0.0783549622,0.098703295,-0.3214305341,-0.1059165522,0.0685794801,-0.0557230897,-0.1259046644,-0.2485820055,-0.3915130794,0.3009372354,-0.2011852711,-0.125779748,0.4019694924,0.1675672531,-0.0332874209,-0.4865242839,0.1772979647,0.1307816654,0.0838177726,-0.3009275794,0.0558152832,-0.0414289162,0.0241742451,0.3911566734,-0.1515443176,0.2833005488,0.4373911321,0.063422285,-0.3570860922,-0.1043281779,-0.0803956091,-0.1107578203,0.4107109308,-0.0791143328,0.2689143121,0.3173108399,0.097559616,0.0045518181,0.2404071838,0.1902999133,0.0589355044,-0.5318906307,-0.0652642325,0.0404917486,-0.0581807084,-0.0185877159,-0.1208131015,-0.0996511951,-0.2963516116,0.3969897032,0.1738691479,0.2589991689,0.1141659543,0.002013742,-0.1082005575,0.1776602268,0.2751132548,0.1055841669,-0.4967321455,-0.0113580646,-0.7635504007,-0.062765941,-0.2667761445,-0.2044683993,-0.0269498862,0.2673875391,0.2948272824,0.2880780697,-0.1117066368,0.2403885871,0.0136979297,0.3622200787,-0.1996468753,-0.2887046933,-0.227032125,0.0318626985,-0.0690125376,-0.4857380092,0.0564628653,0.3484521508,-0.1489176005,-0.1766722649,-0.0491102561,0.1913100332,-0.2589651346,0.3701530695,0.0659780502,0.4796224833,0.0222701002,0.299364239,-0.3624299765,-0.1713561416,-0.2586036026,0.2603024244,-0.0973287523,0.4215645194,-0.3740360141,0.076334171,-0.1894480139,0.0059450381,0.0698222965,0.2765709758,-0.3861725032,0.2962089777,-0.1280640215,0.2370443791,-0.1018579155,-0.0365716107,-0.1631512791,0.162623331,-0.1010828987,-0.1878380924,0.2851313353,-0.2237531394,-0.235858351,-0.0228021126,0.2156576663,0.175096333,0.0522414185,-0.5023126602,-0.0722332075,0.4346192777,-0.1690045297,-0.0793062299,0.0470205322,0.1012325212,0.0152681451,-0.1957579553,0.2328616679,0.3037622571,-0.2695995271,-0.1059476286,-0.3923304975]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Thanks for your quick response! I'm fine with specifying the split as @lhoestq suggested. My only concern is when I'm loading from python dict or pandas, the library returns a dataset instead of a dictionary of datasets when no split is specified. I know that they use a different function `Dataset.from_dict` or `Dataset.from_pandas` but the text\/csv files use `load_dataset()`. However, to the user, they do the same task and we probably expect them to have the same behavior.","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":78,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nThanks for your quick response! I'm fine with specifying the split as @lhoestq suggested. My only concern is when I'm loading from python dict or pandas, the library returns a dataset instead of a dictionary of datasets when no split is specified. I know that they use a different function `Dataset.from_dict` or `Dataset.from_pandas` but the text\/csv files use `load_dataset()`. However, to the user, they do the same task and we probably expect them to have the same behavior.","embeddings":[-0.0780608729,0.058403451,0.0233000908,0.4148484468,0.1361280382,0.1216799617,0.5085968375,0.2503667772,0.360342443,-0.0349156,-0.1385968179,0.3222544789,-0.0713082179,0.1254664958,-0.3528847694,-0.0622133799,0.0382989831,0.1323879957,-0.3376760483,-0.0414372906,-0.380692035,0.3519899845,-0.2903811336,-0.1016624868,-0.162970379,0.0910416842,-0.0846713334,0.3872539401,-0.0802054852,-0.3159049749,0.5106101632,-0.0687854514,0.3722463548,0.2686672807,-0.0001207905,0.0423916653,0.2599446177,-0.1325830966,-0.2995002866,-0.5992068052,-0.0885680392,-0.3759452403,0.4667295218,-0.1101180017,-0.0792847723,0.0684129149,-0.1559933275,-0.4584492743,0.4170620441,0.4606310725,0.1560240984,-0.2249399722,-0.1643290222,0.0096417991,0.4941339493,0.1596400887,-0.1423371434,0.2697026134,0.0866434574,-0.1821702719,-0.2313863486,0.0202852022,-0.1544834673,0.3192254305,0.2253212184,0.2923902273,-0.3081013262,-0.2810398042,0.0261405949,0.2212562412,0.7244586945,-0.0942640826,-0.2048068792,-0.2172994465,-0.0480781011,-0.0917173252,0.3131278753,0.3217790425,-0.0733008608,0.295707792,-0.0414612368,0.0229659714,-0.2213735133,0.1205664724,-0.1148103923,-0.0006338183,-0.1768136621,0.2903594375,0.1218063757,0.0948193893,0.2343305647,-0.1621436775,-0.0009302908,0.1288641691,-0.5839777589,0.2475134283,-0.0986645594,-0.0668170676,-0.1103039235,0.253992945,0.1501595229,0.0090181809,-0.1736856401,0.2963423133,0.4208086133,-0.0071674292,0.0559095033,0.3210427165,0.3060675263,-0.1665060818,-0.0051969648,-0.1115169525,-0.1817283481,-0.3427421451,-0.1860247105,0.1037283838,0.2094877064,-0.2004436702,-0.2784621716,0.0141659556,-0.1827183515,0.0419439301,-0.1135773659,0.2982382178,0.0146188112,0.3928487599,0.014573033,0.2610868216,-0.1982986629,-0.2762896419,-0.0069853971,0.1286406666,-0.3538252711,-0.039692279,0.4889247119,-0.1660848707,0.1336771995,0.2067694068,-0.0537003912,-0.2936573923,-0.1007575616,-0.3508053422,0.1512218863,0.2754834294,0.1813350618,0.2705023587,0.3625773489,-0.2818308473,-0.1320963353,0.2809002399,-0.4552856982,-0.1679245234,-0.0974194631,0.0909826159,-0.2902528048,0.1167384163,-0.318028897,0.1667846739,0.0186548587,-0.1930303872,-0.174176991,-0.2862383723,-0.3174207509,-0.3460519612,-0.0322397426,0.5567400455,-0.8378716111,0.1386426836,-0.0670072213,-0.0069621145,0.1202307865,0.2839854956,-0.2087622434,0.0152275898,-0.2989397347,0.1312758178,0.1732632667,-0.2612000406,-0.3863241971,0.2486075312,-0.0008814359,0.0901698098,0.1690715253,0.075541921,0.496361047,-0.0160967764,-0.0889544412,0.3020797074,-0.0491137654,0.080986008,-0.1848463416,-0.284966588,0.3172256649,0.1995102465,0.0977134109,0.0848607793,0.2303352207,-0.4987604022,0.3067665994,-0.0629990473,0.2209568769,0.0525365658,0.1889110357,0.0994128808,0.2802127302,-0.0316577293,-0.5756883025,0.1889436096,0.4474719167,-0.1121796891,-0.1767742634,-0.105791904,-0.4251611531,-0.0712024122,-0.2718628347,-0.0307784192,-0.0208103936,0.0431174897,-0.0757090971,-0.0927123725,-0.1473402977,0.4585489929,-0.2381284833,0.218170166,-0.3602612317,0.314499855,0.1941236556,-0.0531481057,0.1314400136,0.0408290178,0.1705841124,-0.0698477104,-0.1043954045,0.2648106813,0.4699584544,-0.0506553575,-0.0218371283,-0.2311656922,0.149664849,-0.1610616595,0.0174792912,0.1880163848,0.3294549584,-0.1443995982,-0.4745770097,0.3915098906,-0.3383184373,0.3178929985,-0.1259980351,0.1517553926,0.2254446447,0.0197415687,-0.0936718062,-0.1194756031,-0.1482660472,-0.032560315,0.3659872115,0.1129661649,-0.1503008306,-0.1129395887,0.2878534198,0.1690546423,-0.1196036488,0.1870760471,-0.1231057867,-0.1474548876,0.2803917825,0.367105186,0.4955613017,0.0771473199,-0.23485291,-0.0580582917,0.0573453903,-0.2526851892,0.1350136697,0.04401825,0.0865214318,0.1999137104,0.0943010822,-0.0033025509,-0.1571340114,-0.1080671921,0.1180733889,0.1504926682,-0.5525853038,0.0730808899,-0.285284102,0.173477456,-0.2393057048,-0.1132838354,0.0525890887,-0.3208289444,-0.2696237862,0.2311982065,-0.1116294712,0.0210419111,-0.1159698665,0.19407022,-0.0259667709,-0.2768673003,0.1353655607,-0.1660187393,-0.2796300948,0.02613068,0.0899908245,0.0405326672,-0.0314903818,0.0105322953,-0.0054161027,-0.1679026783,-0.0294126999,0.1906539053,-0.0258915573,0.048165638,0.2501961887,-0.1107524335,0.0286539961,-0.171365574,0.1270343959,0.1098005995,0.0811418444,0.2782082856,0.2357174605,-0.0067102071,-0.1312566251,-0.7236306071,-0.3136633337,-0.2459084988,0.0754915625,-0.0628424138,-0.0796345696,0.0326724313,0.3050904274,-0.0846501812,0.256850481,-0.024679441,-0.1409158856,-0.1120964587,0.5172942877,-0.201225847,-0.2881705761,0.2345358431,-0.1206822023,0.0465553366,0.3901838064,-0.4093537331,0.1691296101,-0.0497052707,0.0715131313,0.0637714043,0.0803195015,0.1051056162,-0.0090013454,0.1107044369,-0.0589840934,-0.1868160963,0.2101540864,-0.0723890662,0.0164383724,0.1945227534,0.4666732848,-0.180589661,0.7081639171,0.052876763,0.058684025,0.1935268193,-0.0913909078,0.3675380647,0.032750912,-0.2014241517,-0.0085765095,-0.1361856312,-0.307294488,0.4365998507,0.044011198,-0.1472160965,-0.1671951711,0.2117059976,-0.2908908725,-0.2611537278,0.1816856414,-0.1667779386,-0.0816012472,0.0153723154,0.2761571109,0.0979495645,-0.1838973761,-0.1667266935,0.1501668692,0.1727153212,-0.0977322385,0.1566798985,-0.106340684,-0.3628371358,0.2003892362,-0.0233827643,0.3507727683,0.1390687823,-0.1785807312,0.0689700916,0.2334896326,0.676369369,-0.2118424624,-0.1088769659,0.1362042725,-0.1670362502,-0.2020245045,-0.2269340158,-0.08950302,0.321803689,-0.0135834832,0.1687471867,-0.0981902406,-0.044085227,0.2892365754,0.1672764122,-0.1867423952,-0.2867829204,-0.117634058,-0.0463212579,-0.1826345623,-0.2411186993,0.1062955856,0.0728992075,-0.2928985357,-0.0041776635,-0.1973250508,-0.2476390898,0.3436926305,0.1330286115,0.2578410506,-0.0313516669,0.1394304782,0.2009285539,0.1762399822,0.2331240624,0.7674669623,-0.3200322688,-0.896895647,0.2190898061,-0.47677055,0.294875741,0.0699816272,-0.3103145361,-0.0204845164,-0.2447078079,-0.1114460602,-0.1208607033,-0.1385059208,0.1131333187,-0.0770476833,-0.2178081572,-0.6420865059,0.451900661,-0.0804931149,-0.138715446,0.1971892565,-0.0059807715,-0.3941088021,0.5317512751,-0.1025062948,0.6488485932,-0.1443080753,0.066354908,0.1926665455,-0.1060860232,0.5815523267,-0.252615571,-0.0314489901,-0.4836898148,-0.0993925259,-0.2113437504,0.0161525942,0.0745463371,0.5343914628,-0.1654260755,0.4403239191,-0.134607017,0.4269822836,0.0749558732,0.1848100871,0.2163040042,-0.2788418531,-0.1900909543,0.0659345835,-0.0644843578,-0.0235290881,0.0954507887,0.0706775934,-0.2394135892,-0.03834977,-0.3074463904,0.2404232919,0.0134579604,0.1497154832,0.0567385629,-0.5112692714,0.5276226401,0.4659340382,0.1288299263,-0.0771813616,-0.1715561599,0.0647411197,-0.0769019052,0.1065504476,-0.0476461202,-0.0652330592,0.1183977276,0.3097569644,-0.184985593,0.1965093017,-0.1329264641,-0.060037002,0.2879072428,0.3575039804,0.2925893962,-0.5540482998,-0.3186806738,-0.2082457989,0.1750919968,-0.0921975821,0.0287257414,-0.0640124828,0.1739258319,0.124816753,-0.0214965492,-0.3738910258,0.045145154,0.1663763672,-0.0290663596,0.0572426692,0.4186035097,0.3911679685,-0.169159919,-0.1707745194,0.0406671353,0.0574521832,-0.5586921573,0.272693783,0.0945262089,0.1553149074,0.0937950984,0.1295140684,0.0750710368,-0.1472866386,-0.0274174865,-0.5871501565,0.0391716324,0.0518657975,-0.1045517325,0.0467989407,0.0984390154,-0.4158429205,-0.0531222001,-0.1460790634,-0.1662377417,0.1227066591,0.0952958837,0.0282586943,0.2055833042,0.3031570613,-0.1256092489,-0.2308121473,-0.0158372149,0.0365633145,0.1334188282,-0.0922423229,-0.1767408401,0.200640291,-0.037426576,0.1355479211,-0.2335270643,-0.2175002843,-0.2205772996,-0.2437777221,0.2594327331,0.3854045272,-0.0624373257,0.3423619866,-0.007410048,-0.0218210872,-0.5484731793,0.3721247315,-0.333745569,0.264854461,0.1167161837,0.2449628562,0.0713142455,0.1001943201,-0.0735298097,0.040366821,-0.0935590938,-0.1060140654,0.3031860888,-0.330824852,0.2521577775,0.1782936901,0.3616377711,0.4359160364,-0.0412281565,-0.1274756342,0.1186124012,0.1057944149,-0.2864648104,0.1520410776,0.1158831716,0.1954275966,0.0093329567,0.4370470941,-0.0044519398,-0.1398126781,0.0296286251,0.0849419981,0.6649833322,0.1072682217,0.2528632581,0.3818801939,-0.0435524508,0.1784691364,0.1605795026,-0.1021741927,0.12071006,0.5896924138,-0.1346414089,0.0309120324,0.2568347752,0.1964881569,-0.0786211938,-0.6072866321,0.3335479498,0.2953578532,-0.2968009412,-0.0950201899,0.17487517,0.6289703846,-0.1853761524,0.1386339515,-0.3035647869,-0.072698392,0.0657653362,-0.0995672122,-0.1449110955,-0.2396784425,-0.3253216445,0.2576449513,-0.2221736908,-0.1446971744,0.3557284176,0.1453876495,0.0212243348,-0.5101605654,0.1769391,0.0611311682,0.0676239729,-0.2418274879,0.0452640876,-0.0599957444,0.1075006127,0.411095947,-0.1862286031,0.3582159579,0.4266976416,-0.0102404049,-0.4081254005,-0.073347427,-0.0697489902,-0.0991298333,0.4141913056,-0.0458859988,0.3552892506,0.3006964624,0.107278645,-0.0228898618,0.1493742466,0.1904636174,0.0595689267,-0.4607411623,0.0260301363,0.0315621831,-0.0554631911,-0.0809353292,-0.1332417876,-0.120658949,-0.2544751167,0.4193520844,0.0971190035,0.2352307886,0.1572280675,0.0010582089,-0.0490168892,0.1945968866,0.3154040277,0.1179358959,-0.3888403475,0.0148337325,-0.70088166,-0.0577346161,-0.1603074223,-0.2227324843,-0.0657904744,0.2667681277,0.3000711203,0.2722353637,-0.1075643748,0.2855184674,0.0407360122,0.2488112748,-0.1010121703,-0.2874626517,-0.2420125157,-0.0227797739,-0.1214359701,-0.5083060265,-0.010401777,0.3372935057,-0.156740889,-0.1524049491,-0.0398764424,0.1726846844,-0.3162718117,0.2824459672,0.0485674068,0.4625611901,0.0350711308,0.2286790758,-0.3366465271,-0.2433710992,-0.2299163491,0.3020690084,-0.0797186792,0.4605024457,-0.367775619,0.0734988302,-0.2131593227,0.0699645057,0.1490738839,0.2884459794,-0.4138551056,0.2748316824,-0.1599977165,0.2001490444,-0.0224916916,0.0812497884,-0.2107774317,0.1545372158,-0.0083400682,-0.2705018818,0.1735009998,-0.2811928093,-0.2791195214,-0.0397654474,0.1720833331,0.1727098078,0.093914859,-0.4474797249,-0.0440837033,0.342249006,-0.1715088338,-0.1493270248,0.0345320664,0.0146459034,0.044138018,-0.2010121942,0.1968457997,0.3159100413,-0.2689936161,-0.0672228113,-0.3839844763]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='.\/amazon_data\/Video_Games_5.csv', delimiter=\",\", split=['train', 'test'])\r\n```\r\nI was running the above line, but got this error.\r\n\r\n```ValueError: Unknown split \"test\". Should be one of ['train'].```\r\n\r\nThe data is amazon product data. I load the Video_Games_5.json.gz data into pandas and save it as csv file. and then load the csv file using the above code. I thought, ```split=['train', 'test']``` would split the data into train and test. did I misunderstood?\r\n\r\nThank you!\r\n\r\n","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":78,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='.\/amazon_data\/Video_Games_5.csv', delimiter=\",\", split=['train', 'test'])\r\n```\r\nI was running the above line, but got this error.\r\n\r\n```ValueError: Unknown split \"test\". Should be one of ['train'].```\r\n\r\nThe data is amazon product data. I load the Video_Games_5.json.gz data into pandas and save it as csv file. and then load the csv file using the above code. I thought, ```split=['train', 'test']``` would split the data into train and test. did I misunderstood?\r\n\r\nThank you!\r\n\r\n","embeddings":[0.0013099385,0.05171258,-0.019382393,0.4013643861,0.2351075709,0.2222292572,0.5522068739,0.3211097121,0.2396999747,-0.0363929719,-0.146097675,0.2243235707,-0.025056662,0.102229014,-0.3207229376,-0.1825466752,0.0453077927,0.0672884807,-0.2340873033,0.0617866963,-0.3590757251,0.3422645926,-0.2727225721,-0.0144880507,-0.2078830004,0.0124817826,-0.1053493991,0.296968013,-0.1483759135,-0.2648254633,0.4790059626,-0.1716454923,0.5383775234,0.3396247923,-0.0001182179,0.0470696352,0.2971403897,-0.1246482879,-0.2083082646,-0.5637422204,-0.056119509,-0.3337765634,0.3093161285,-0.1384035796,-0.0306093525,0.0026258738,-0.1647102088,-0.37167117,0.5169056058,0.5258500576,0.1700556278,-0.2575105131,-0.1054589003,0.010397505,0.5956737995,0.0796283782,-0.1261375397,0.2239545137,0.0111493096,-0.128290683,-0.1261236072,0.0465885103,-0.0409378149,0.3501431644,0.1306047738,0.2808638811,-0.2173999995,-0.2594374418,0.0357642174,0.1646072417,0.6517999172,-0.0907546654,-0.25573349,-0.0893641934,-0.0458670929,-0.1906019151,0.2356330603,0.4025241733,-0.0143940095,0.2629038095,-0.0302205607,0.0621254407,-0.2307951003,0.143907845,-0.0291519202,0.0988970771,-0.2419938594,0.3104786575,0.0748927519,0.0376325175,0.2845781446,-0.0803926289,-0.0853785947,0.1097130626,-0.73577106,0.1972242743,-0.2606303692,-0.1393675655,-0.1818331033,0.2301413119,0.1986734569,-0.048055131,-0.0711116493,0.3267121017,0.3619975746,0.1196252257,-0.0493249558,0.3038940132,0.2154174745,-0.1192011908,-0.0823735744,-0.189896971,-0.2964744866,-0.5364337564,-0.0690400451,0.1694035679,0.2778784931,-0.2056666315,-0.3103758693,0.0415622666,-0.3661009669,-0.0593541749,-0.0650034547,0.3313742578,0.0484936647,0.1800385416,-0.0199314784,0.3329245448,-0.1016789228,-0.2537936866,-0.0099349301,0.2129421234,-0.3322185278,-0.0475763343,0.4394691885,-0.1399374455,0.169890821,0.1438711435,-0.1016609594,-0.3589423001,-0.0439114124,-0.3201140761,0.0598089546,0.2911030948,0.0977208391,0.2263792157,0.2708276212,-0.2004960328,-0.1134681404,0.3200462759,-0.4528761506,-0.1949418187,-0.1236148328,0.0638868064,-0.2958910167,0.0516121313,-0.3975146711,0.0070914431,0.1372073293,-0.1690459847,-0.1107430756,-0.2598054111,-0.223449856,-0.2329428494,-0.0297738779,0.3591510355,-0.729249835,0.2099155039,-0.1305401772,0.139788568,0.204263553,0.2783480585,-0.2136737108,0.0838187039,-0.1957585216,0.1398439556,0.1922795922,-0.1948812455,-0.2584933043,0.3368588388,-0.0548021421,0.0118795587,0.14804326,0.0079602227,0.2661395669,0.0118092205,-0.2148744017,0.277939707,-0.1107130125,0.0532059409,-0.2178413868,-0.2958856225,0.393114984,0.2233066708,0.0809110627,0.0637842864,0.219349429,-0.4887855649,0.2911222577,-0.0074691433,0.213099435,0.1266507357,0.2399577498,-0.0118248109,0.3064186573,-0.0468057133,-0.5506250858,0.1211448312,0.2973080575,-0.0805335045,-0.2491194308,-0.0293331854,-0.4990498126,-0.1263436526,-0.1931822002,-0.1471881717,-0.0113766007,0.0028395574,-0.0695435628,-0.1811055839,-0.1231717244,0.4104182422,-0.1968202293,0.1574802846,-0.420506835,0.3484317362,0.0685198754,-0.046662081,0.232258305,-0.0470527746,0.317230612,-0.1737196296,-0.172609672,0.2992421985,0.3452042341,0.0514555573,0.0920722485,-0.2689474523,0.0843897611,-0.1345872581,-0.07657893,0.1975729764,0.3707106113,-0.090206176,-0.4967188537,0.3981715739,-0.373051554,0.3407868743,-0.1548341811,0.1214029342,0.2545317709,-0.0342797935,-0.0573217347,0.0154310679,-0.0674974099,-0.0414524674,0.1813625395,0.1204914972,-0.1412470639,-0.0947611406,0.2932521701,0.1242887378,-0.2171434313,0.1496353447,-0.0212595519,-0.0907215476,0.3234220147,0.4788161814,0.4989160895,0.0934970304,-0.2257995307,-0.0502079949,0.090594776,-0.2129285783,0.1495766789,0.1314753294,0.1568460763,0.1497189403,0.1046723127,0.0309833139,-0.0564096943,-0.153122142,0.2575722635,0.2865161896,-0.5545143485,0.0832569972,-0.2742894888,0.251234889,-0.2948322296,-0.1245409399,0.1124371514,-0.3302797377,-0.2978978753,0.2424979359,-0.1188882664,0.0766470656,-0.1887726784,0.1858246028,0.0945126414,-0.1874366701,0.0460106358,-0.2297289073,-0.1853150278,0.0440204144,0.1608080864,0.1147634089,-0.1593758166,0.0924839228,0.1120204851,-0.1131626442,-0.0734895021,0.1205578297,-0.0692943409,0.1066164449,0.0611006469,-0.0971011519,-0.1153244749,-0.3006359041,0.0956041068,0.0619500428,0.0524068102,0.3358862698,0.3010444641,-0.1343676746,-0.0637167245,-0.771294415,-0.292863667,-0.2254818976,0.1364150345,-0.0603985898,-0.0800451189,-0.009466528,0.2729968131,-0.046676036,0.3181810975,-0.0819869265,-0.1524343342,-0.3559525907,0.6282113194,-0.33587569,-0.2506924868,0.212349385,-0.0891900733,0.0134996045,0.3305257559,-0.3797052801,0.2495916933,-0.0787978917,-0.0502249151,-0.0006279607,0.0238592699,0.0380966552,-0.0071674823,0.081921041,-0.0103262225,-0.1273985505,0.2148552835,0.0483199209,0.1261241436,0.104697682,0.4444700181,-0.1454232782,0.832663238,0.0092135007,0.0208379421,0.2185450792,-0.1146018207,0.2767150104,0.1338882148,-0.2511597872,0.0126347225,-0.1921897978,-0.2479174137,0.3181246221,-0.0622342378,-0.1635575444,-0.0671132579,0.2253295779,-0.3498931229,-0.2616161406,0.219756335,-0.0992869139,-0.0568968803,0.0018099865,0.197286129,0.1112910137,-0.245378688,-0.1210482419,0.0973607302,0.1619684547,-0.0868247449,0.0986912772,-0.1675404161,-0.2355763316,0.2341689765,-0.0257508103,0.465012908,0.1191044897,-0.0924313739,0.0427705534,0.1901145726,0.790368259,-0.2772937119,-0.0669711009,0.0895091668,-0.0843078718,-0.1386624724,-0.2487661242,-0.1439347863,0.3455865681,-0.0632849634,0.2461963296,-0.0161964092,0.0098986914,0.4104843438,0.0764349997,-0.1893097758,-0.3667446971,-0.1203171685,-0.0003531738,-0.1793943644,-0.1840387434,0.2391402274,0.1934204847,-0.3377687335,0.0585261062,-0.2245917618,-0.1983197629,0.4003881216,0.1000645459,0.222487241,0.0537363254,0.1503234953,0.1658302099,0.0966288745,0.2035507709,0.7843945026,-0.2009359747,-0.8349447846,0.2318479717,-0.5099244714,0.3329105377,0.0696405619,-0.3419288993,-0.0750046149,-0.152355656,-0.0593255833,-0.064184159,-0.0652371272,0.1192682907,-0.0315260664,-0.170163691,-0.4961595237,0.5014415979,-0.1725474447,-0.099379614,0.165089488,0.0060842289,-0.3151542842,0.5069113374,-0.1264930218,0.7633357048,-0.0809842944,-0.0384552181,0.1238727644,-0.2518376708,0.4888052344,-0.3862514496,0.0171930566,-0.4449011087,-0.187705934,-0.1797460914,-0.0094155679,0.1393853426,0.5158739686,-0.1586592644,0.3798233569,-0.0642664805,0.3847050667,0.0100509506,0.1798697114,0.2399886698,-0.254288435,-0.1591579616,0.061796777,-0.1136795282,0.0402820185,0.1215542555,-0.0685013682,-0.3014987707,-0.0222902056,-0.258771956,0.278627038,0.0064092157,0.169696629,0.0670542717,-0.5253942013,0.4796038866,0.4157079756,0.1502281129,-0.1548792422,-0.1765185148,0.1454425752,-0.0027779422,0.0345427133,-0.1173274368,-0.0234220736,0.1226177365,0.2246334553,-0.2167378664,0.2141619176,-0.1105540246,-0.0573693737,0.2784081995,0.2830651402,0.1158407927,-0.5786654949,-0.4050552547,-0.1994121671,0.0355778039,-0.1358472407,0.0427845269,-0.095901899,-0.0073771188,0.0790164173,-0.0109543689,-0.379995048,0.0163821205,0.2304484099,-0.1041684598,0.1677046716,0.4894424975,0.3991492391,-0.120305635,-0.1565667093,0.1239945814,0.1943260729,-0.6159325838,0.1485491991,-0.0160779599,0.2707149088,0.1173539683,0.0935313553,0.1481492072,-0.1131062582,0.1228978187,-0.6571761966,-0.0065172301,-0.0163924806,-0.025537286,0.0661222339,-0.0053585488,-0.3471963704,-0.0063308096,-0.0185468178,-0.1774584502,0.2192814797,-0.016016826,0.0773725435,0.0875928029,0.2237473726,-0.1134931445,-0.3160484135,0.0059332154,0.0807985663,0.1289913654,-0.1185339466,-0.2344442606,0.1807435155,0.0513067357,0.1172635704,-0.2916288972,-0.171203047,-0.2519420981,-0.1157091036,0.122261025,0.3097827733,0.0763874203,0.4201918542,0.0632062554,-0.1346996278,-0.5766661167,0.2431892008,-0.2517924607,0.2563348114,0.200434804,0.2189345658,0.1526429355,0.132629931,-0.1045133621,-0.1199813262,-0.1409800351,-0.1232136264,0.3064601719,-0.3543687761,0.1658968925,0.1318966001,0.3639472723,0.4501647651,-0.0877354816,-0.0296064764,0.194535926,0.1060161293,-0.3109939098,0.2500671148,0.1015497372,0.2192715555,-0.0666508526,0.3654021323,-0.0586885735,-0.0059233694,0.0418658406,0.1571998149,0.6517170072,0.0037278084,0.2534441054,0.2700840831,-0.1033791676,0.171817854,0.2659032941,-0.039995037,0.0503829941,0.4718825817,-0.1745712608,0.0558809526,0.3098167479,0.2301667035,-0.0958890393,-0.5000249147,0.2013560981,0.3368729949,-0.226283446,-0.061237853,0.2030343562,0.75197649,-0.0741559789,0.0944383293,-0.2837046683,-0.0529755987,0.0688966885,-0.0506859608,-0.1667233109,-0.194307521,-0.4568819404,0.2175016254,-0.157866478,-0.1699221432,0.3505592942,0.2255699933,-0.0205493588,-0.4643119574,0.1008405536,0.0951001123,0.1084275916,-0.2965501249,0.0249687601,0.0136129353,-0.0371115766,0.3568257689,-0.2303853333,0.3943682611,0.3589632213,-0.0824880004,-0.2996542454,-0.0001913275,-0.1467819214,0.0811101571,0.4575303197,0.0188797582,0.2979622185,0.291082263,0.1223561168,-0.0376565717,0.3405008018,0.195913136,0.1682380289,-0.4209381342,0.0973645002,0.0194368046,-0.0772222131,-0.1342261434,-0.2494690865,-0.2247814834,-0.1644600928,0.28709203,-0.0139043229,0.228413105,0.1483282745,0.0054250853,-0.0634680092,0.2566858232,0.2689575255,0.0835417658,-0.5096370578,0.0921724141,-0.7402230501,-0.1185131222,-0.1397496611,-0.1350007951,0.0336732641,0.2456679195,0.2258338183,0.3275522292,-0.0969278738,0.1426140517,-0.0532682016,0.3893442452,-0.1827596873,-0.149500981,-0.2000493556,0.1063718647,-0.0652989298,-0.4260183275,0.0570442826,0.2859656513,-0.1811177135,-0.1203172207,-0.2147776186,0.1211224124,-0.2325747311,0.3196118772,-0.0330539756,0.4393061101,0.0305328183,0.3137936294,-0.4146061242,-0.316257447,-0.162063539,0.3063241839,-0.0047634831,0.4763437212,-0.3523975313,0.043399781,-0.321406424,-0.0328664593,-0.026316274,0.2140740007,-0.3108560145,0.2241715193,-0.1598862261,0.2404781729,0.0255052522,-0.0010709423,-0.2014998049,0.1945049465,-0.0734492317,-0.2631210983,0.2682917416,-0.2515657842,-0.3278561532,0.1213800088,0.1324686259,0.3412543237,0.0790841505,-0.49938187,0.0459823869,0.3171979487,-0.1684502065,-0.0226213541,0.045990523,0.104587011,0.1017698422,-0.1623759568,0.098918654,0.3107347488,-0.2570582628,-0.1267532706,-0.360827893]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi ! the `split` argument in `load_dataset` is used to select the splits you want among the available splits.\r\nHowever when loading a csv with a single file as you did, only a `train` split is available by default.\r\n\r\nIndeed since `data_files='.\/amazon_data\/Video_Games_5.csv'` is equivalent to `data_files={\"train\": '.\/amazon_data\/Video_Games_5.csv'}`, you can get a dataset with \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='.\/amazon_data\/Video_Games_5.csv', delimiter=\",\", split=\"train\")\r\n```\r\n\r\nAnd then to get both a train and test split you can do\r\n```python\r\ndataset = dataset.train_test_split()\r\nprint(dataset.keys())\r\n# ['train', 'test']\r\n```\r\n\r\n\r\nAlso note that a csv dataset may have several available splits if it is defined this way:\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files={\r\n \"train\": '.\/amazon_data\/Video_Games_5_train.csv',\r\n \"test\": '.\/amazon_data\/Video_Games_5_test.csv'\r\n})\r\nprint(dataset.keys())\r\n# ['train', 'test']\r\n```\r\n","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":123,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi ! the `split` argument in `load_dataset` is used to select the splits you want among the available splits.\r\nHowever when loading a csv with a single file as you did, only a `train` split is available by default.\r\n\r\nIndeed since `data_files='.\/amazon_data\/Video_Games_5.csv'` is equivalent to `data_files={\"train\": '.\/amazon_data\/Video_Games_5.csv'}`, you can get a dataset with \r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='.\/amazon_data\/Video_Games_5.csv', delimiter=\",\", split=\"train\")\r\n```\r\n\r\nAnd then to get both a train and test split you can do\r\n```python\r\ndataset = dataset.train_test_split()\r\nprint(dataset.keys())\r\n# ['train', 'test']\r\n```\r\n\r\n\r\nAlso note that a csv dataset may have several available splits if it is defined this way:\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files={\r\n \"train\": '.\/amazon_data\/Video_Games_5_train.csv',\r\n \"test\": '.\/amazon_data\/Video_Games_5_test.csv'\r\n})\r\nprint(dataset.keys())\r\n# ['train', 'test']\r\n```\r\n","embeddings":[-0.1236993298,0.0545275696,0.0046684053,0.3440138996,0.2238630652,0.2072241902,0.5557584763,0.3320873976,0.2292669863,0.0723564923,-0.1219850779,0.1893656254,-0.0988404006,0.0580993779,-0.3150721788,-0.194993943,0.0256498456,0.0631211326,-0.2891602218,0.0247312747,-0.2967126966,0.3495420814,-0.2854318023,-0.1041691303,-0.2474516779,0.1247463822,-0.1284908503,0.2887857258,-0.1067125648,-0.2828909755,0.5074794292,-0.0799292624,0.4618830681,0.3443642855,-0.0001203501,0.0970003307,0.2799856365,-0.091465354,-0.2787306607,-0.5799635053,-0.0071041742,-0.375300616,0.3846252859,-0.1235786229,-0.1427274942,0.0628602952,-0.0909830704,-0.3723733723,0.4926059544,0.502761364,0.1562837213,-0.2277879417,-0.2041125447,-0.051319059,0.6431521177,0.1048247665,-0.1537516266,0.270380944,0.0426355526,-0.1800077707,-0.1540324092,0.0384248085,0.0187865905,0.2721130848,0.1503915191,0.3210097253,-0.1926092505,-0.2060050964,0.0234964862,0.1783809215,0.5856808424,-0.0617553107,-0.1998890489,-0.1086915657,-0.0031236117,-0.2112372667,0.2918922901,0.3944924474,-0.0140921967,0.2673090696,-0.0840748101,0.0562876873,-0.2715720832,0.1967208087,-0.0411109254,0.0441895053,-0.160697028,0.3093575537,0.1828568578,0.1352512985,0.2616449594,-0.0672460869,-0.0468277894,0.1808257997,-0.7518927455,0.2322711051,-0.2511095703,-0.1547812074,-0.0936986431,0.2901218832,0.2221346498,-0.0328100696,-0.0704159141,0.3143113554,0.3604484797,0.0580574311,-0.0649092644,0.373855114,0.2827087641,-0.0966242552,-0.0495880581,-0.1580700725,-0.2829408944,-0.5059507489,-0.0812057331,0.1843688339,0.2114668787,-0.1583938897,-0.2342969477,0.0050778058,-0.3175958097,-0.0391786583,-0.0982536599,0.324622035,0.122901082,0.2582788169,0.0182131343,0.3555499017,-0.1680651605,-0.3251541555,-0.0390207842,0.163063094,-0.3775836825,-0.0187384021,0.527097404,-0.2026496828,0.1497711837,0.2378042489,-0.0695297271,-0.3647531271,-0.0183723271,-0.3357921243,0.060033977,0.2678999603,0.1478130966,0.2357051224,0.2901499271,-0.1314113438,-0.0991395116,0.3206919432,-0.4503082037,-0.2298634648,-0.0609409958,0.0792621821,-0.2566174567,0.1285786182,-0.3197812736,0.1215938702,0.1233536229,-0.1894128919,-0.1974068284,-0.2858693004,-0.3190152645,-0.2252106816,0.0287518781,0.4265123904,-0.7253031731,0.1600687057,-0.0735779107,0.0818117931,0.1762104481,0.2192542106,-0.3011444211,0.0285338406,-0.2588867843,0.1298175007,0.2128348202,-0.1452604532,-0.3053944111,0.3130154312,-0.0814771727,-0.0150277643,0.1646915376,0.0735130981,0.3808694184,0.0346264243,-0.1928317547,0.273281008,-0.0599818751,0.0648037419,-0.2179353982,-0.2626712918,0.268104732,0.2398951948,0.121989511,0.0886864811,0.2000192255,-0.4222188294,0.3684153557,0.0018666971,0.2421445996,0.0944420248,0.2172833234,0.0069560236,0.2594904304,-0.0468955226,-0.5880161524,0.175081268,0.3549012542,-0.1456279159,-0.1272691637,-0.0827832744,-0.4457899332,-0.1092281342,-0.2193202376,-0.0531642027,-0.0301740468,0.0546009205,-0.0905675292,-0.2170437425,-0.150492087,0.3589515984,-0.2253829837,0.1345571727,-0.347008884,0.3731137514,0.1203849763,-0.074222669,0.1797977686,-0.0364786349,0.3094663024,-0.1655763984,-0.1550948173,0.3368910849,0.4324690402,-0.0177731831,-0.0070517296,-0.2337273061,0.1350266635,-0.1509144008,-0.0988640636,0.2633002996,0.2995477617,-0.1329545379,-0.6124714613,0.3689530492,-0.34813115,0.3461934924,-0.1082850099,0.1398616582,0.254413873,-0.0694497824,-0.0920359567,-0.0502822772,-0.0246854499,-0.0507279113,0.2192940414,0.0786125585,-0.1820709705,-0.1413666159,0.1904687732,0.1454722136,-0.2642039955,0.1631590277,-0.0574991964,-0.0909903049,0.286041081,0.3959853947,0.4706317782,0.0796952099,-0.1567980647,-0.0353182107,0.074081555,-0.2415788472,0.1159654185,0.0701689273,0.0640580058,0.2292760313,0.2091118246,0.0643702894,-0.087200731,-0.1415751725,0.2427642196,0.1857290268,-0.5693569779,0.0414618142,-0.2665552199,0.3180040419,-0.3426658511,-0.056278944,0.1317123622,-0.3142736554,-0.2392712831,0.2791632712,-0.1528598368,0.0895879045,-0.1620828956,0.241129145,0.0410924926,-0.189879939,0.1409055889,-0.1960645169,-0.273843199,0.0282651931,0.1539470106,0.0585999936,-0.1277996004,0.1029053107,-0.0550785661,-0.1706884354,-0.0086257122,0.1619700938,-0.0035848648,0.0908206999,0.138238579,-0.1053299382,-0.0646063834,-0.2693961263,0.1504160911,0.0771211237,0.0710023791,0.314376086,0.2691509128,-0.0828716531,-0.179314509,-0.780608654,-0.2570376992,-0.2282651514,0.0682769269,-0.0943248793,-0.0680395365,-0.0079498822,0.2537840307,-0.0882224739,0.3137325346,-0.1078761145,-0.1132512689,-0.2240658998,0.5512241721,-0.2445445955,-0.3344505131,0.2084482312,-0.1100612208,-0.0290145148,0.3913619816,-0.4320285916,0.2069480866,-0.0562909469,-0.0276837908,-0.0692590624,0.0554977842,0.0666981637,-0.06051654,0.0955885202,-0.0226425305,-0.1251135468,0.2517670393,-0.020024037,0.1037029028,0.0690994859,0.4551238418,-0.1279479712,0.840536654,0.0934606194,0.0382797495,0.2161407471,-0.1279042959,0.2627877295,0.1161440983,-0.2052368075,0.0491353013,-0.2187660784,-0.2989220023,0.343603313,0.0062897736,-0.158817783,-0.0846071616,0.3055967391,-0.2164925635,-0.2865198255,0.2109033167,-0.1444711685,0.0212432612,-0.049886778,0.1768529266,0.1406903565,-0.1986887306,-0.0943417326,0.1105646119,0.1068759561,-0.1261726618,0.1068718582,-0.1246998683,-0.258471787,0.1572492421,-0.0449999683,0.3818120658,0.1139998436,-0.1269673109,0.0701359734,0.2051007897,0.8161818981,-0.1706873178,-0.0262791291,0.1365031749,-0.0943441987,-0.1557804644,-0.3316735923,-0.1315875202,0.2589629292,0.0289924331,0.2791790962,-0.0939700156,0.0061338344,0.4122930169,0.1237512603,-0.1991576999,-0.3709280491,-0.1176720187,0.0243075211,-0.1085612252,-0.1439118981,0.1700904071,0.1237185225,-0.3360510468,0.0344572589,-0.2056022435,-0.258038938,0.3210894465,0.1019874215,0.2867149711,0.067013219,0.2236173749,0.2309412807,0.0865521654,0.1521280259,0.8312818408,-0.2476365119,-0.8462259769,0.3043625355,-0.5130507946,0.3058377206,0.1006135419,-0.3390438259,-0.0740043074,-0.2073810399,-0.0908136517,-0.0643537119,0.0385148488,0.1052993461,-0.1000569984,-0.1768036932,-0.5368046165,0.5061169267,-0.1158609092,-0.1219228879,0.1163800508,-0.064086549,-0.2929754257,0.5268897414,-0.1695971936,0.6717287302,-0.1161372289,0.0086252922,0.1661095321,-0.1738390923,0.3820894957,-0.3971768916,0.0192866251,-0.4172923267,-0.1871000975,-0.1575158089,0.028835902,0.045035772,0.5768893361,-0.1867121309,0.3853290975,-0.1072077453,0.4090873897,0.0115924142,0.131070748,0.2775966823,-0.2057442218,-0.0936230719,0.0730208457,-0.0481260158,-0.0305720493,0.1257380992,0.0225633327,-0.2299945205,-0.0861866698,-0.3585097194,0.3015389144,-0.0580208041,0.1496921778,0.0550172552,-0.5366518497,0.4740409851,0.4240622818,0.0932305455,-0.1158969626,-0.1576935947,0.1158879325,0.0249201767,0.0712478682,-0.1567979008,-0.0490732305,0.0624352805,0.2436160743,-0.1971800923,0.2300535291,-0.1238343567,-0.0305629615,0.305573374,0.3118975759,0.2201452851,-0.5418587923,-0.3424956799,-0.1763364673,0.1146238446,-0.1732373685,0.0370578393,-0.0883756727,0.0022506043,0.0070836921,0.0366398133,-0.3459068537,0.0961277038,0.2780534029,-0.079132624,0.0692589134,0.4756678343,0.3913168311,-0.0722296238,-0.1721457988,0.0924541578,0.1939412653,-0.6047871709,0.1933580786,-0.0401426516,0.2394957989,0.1869920343,0.1891756803,0.154988721,-0.1268637478,0.0681020021,-0.5866593719,0.0394465327,-0.0464410856,-0.0775226876,0.1153251454,0.0230263397,-0.3197153509,-0.0997641459,0.0018915024,-0.1686231792,0.1739281416,0.0741657689,0.108770743,0.1004060954,0.2541715205,-0.1494038552,-0.2636879683,0.0088842344,0.0132841878,0.1262332201,-0.1043557748,-0.2518170476,0.1814361364,-0.0176955834,0.2021208256,-0.2842322588,-0.1261136383,-0.1908904761,-0.1310359985,0.1956932545,0.2580342293,-0.0112840366,0.4334688783,0.1149795428,-0.0632633865,-0.5831515193,0.2543992698,-0.2832888365,0.1787214875,0.2102704346,0.2335810959,0.1305256635,0.1091150492,-0.1699352562,-0.092206277,-0.0863703862,-0.0878939107,0.2572662234,-0.2967858315,0.1809120029,0.1053466946,0.3611459434,0.4378594756,-0.0838360786,-0.0477893353,0.113895677,0.1188237444,-0.3225206733,0.2063741386,0.09352424,0.2312634289,-0.0427020416,0.3317616284,0.0004151528,-0.064343594,0.0140073532,0.1694218814,0.5861902237,0.0251170881,0.2621178329,0.3344129622,-0.0588498376,0.1211710796,0.1560478359,-0.1573315412,0.0728304982,0.4479795992,-0.2174402326,0.0932001248,0.3141719699,0.1881577522,-0.1047521904,-0.568825841,0.1891770065,0.2869030833,-0.286314249,-0.1088717133,0.1301440895,0.7349344492,-0.1073281169,0.0295041744,-0.3249364197,-0.0867304355,0.0972531214,-0.0965511128,-0.1583584696,-0.2195175886,-0.484954983,0.28180632,-0.184979558,-0.0466711968,0.4361394346,0.2232358307,-0.0277029872,-0.4806584716,0.0966255143,0.1565505862,0.0846172497,-0.3090490997,0.0834893286,-0.021261964,-0.0038151813,0.3777878284,-0.2420455664,0.3354338706,0.3684580922,0.0073713278,-0.3682180345,-0.0619435087,-0.0630295053,-0.007812541,0.376955837,0.0059910733,0.3021041751,0.2983396947,0.090074189,-0.0261352528,0.3797567487,0.1009403244,0.1272027344,-0.5206719637,0.047071714,0.061259266,-0.1090081856,-0.0982129052,-0.2307288498,-0.1610397249,-0.1864679754,0.323150456,0.0533484891,0.2272536457,0.1613298357,-0.0059358771,-0.0697870702,0.2767516971,0.3064690232,0.0358427167,-0.5206972361,-0.0163222495,-0.7225837708,-0.0909704864,-0.2059160322,-0.1488364786,-0.000634077,0.2883149385,0.2138852477,0.3061715961,-0.0822249651,0.1076726541,-0.0376043022,0.3772444427,-0.2053090334,-0.1663640887,-0.1563258022,0.0986460522,-0.0578413233,-0.4393822551,-0.0255547818,0.298512131,-0.1839352399,-0.0591370203,-0.2375973016,0.1494272649,-0.2458091974,0.3341518939,-0.0345351025,0.4197685421,0.0257531013,0.3332746327,-0.4953846633,-0.2681282759,-0.2189337909,0.2006675899,-0.0507924855,0.4944992363,-0.3649024963,0.0594792888,-0.2512992024,0.0077369455,0.0468175747,0.1744482815,-0.2876694798,0.2198364139,-0.1487819403,0.2688424885,-0.0124190571,0.0497606359,-0.2028205395,0.1340834051,-0.0502321273,-0.2751541138,0.2319502831,-0.2538375556,-0.2683008313,-0.0170317609,0.1544953734,0.2229427695,0.1902010888,-0.5178751349,-0.0281958226,0.3256676793,-0.1713875532,-0.1206701174,0.1029580384,0.1581060141,0.0384632163,-0.1509172171,0.228666082,0.3510144055,-0.2745783329,-0.1515935659,-0.4364373088]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"> In this case the user expects to get only one dataset object instead of the dictionary of datasets since only one csv file was specified without any split specifications.\r\n> I'm ok with returning the dataset object if no split specifications are given for text\/json\/csv\/pandas.\r\n> \r\n> For the other datasets ton the other hand the user doesn't know in advance the splits so I would keep the dictionary by default. What do you think ?\r\n\r\nYes maybe this would be good. I think having to select 'train' from the resulting object why the user gave no split information is a confusing and not intuitive behavior.","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":107,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\n> In this case the user expects to get only one dataset object instead of the dictionary of datasets since only one csv file was specified without any split specifications.\r\n> I'm ok with returning the dataset object if no split specifications are given for text\/json\/csv\/pandas.\r\n> \r\n> For the other datasets ton the other hand the user doesn't know in advance the splits so I would keep the dictionary by default. What do you think ?\r\n\r\nYes maybe this would be good. I think having to select 'train' from the resulting object why the user gave no split information is a confusing and not intuitive behavior.","embeddings":[-0.0685335547,0.0198079459,0.0123501066,0.4163457453,0.1547634006,0.1310773194,0.5161136985,0.3348332942,0.3283668458,0.0745705143,-0.1174225211,0.2238826901,-0.1303579956,0.0772807971,-0.3574971259,-0.1409811527,-0.0161414035,0.1667791605,-0.3396335244,-0.0541298836,-0.3720598519,0.3022883534,-0.3071189523,-0.103884697,-0.2333146483,0.1547114998,-0.1243326813,0.2636172473,-0.1306831241,-0.3113133609,0.5264427662,-0.0348026864,0.4296829104,0.1796253175,-0.0001202436,0.0683241561,0.2098931521,-0.1570206881,-0.2832562327,-0.5928282738,-0.0948916599,-0.5000761151,0.371696502,-0.1824131757,-0.1818093508,0.0587633513,-0.055730477,-0.3692515492,0.4264984727,0.4404452145,0.1417728961,-0.2375429273,-0.2223368734,0.0053099506,0.5322136283,0.1271093339,-0.159385249,0.2628558576,0.0841457024,-0.1708380729,-0.146781832,0.0727015436,-0.0378269888,0.250505805,0.1527512074,0.2516296506,-0.2149782777,-0.2802566588,-0.0473094918,0.2274475098,0.6906309724,-0.0038556212,-0.2016994506,-0.2366113514,-0.0083019249,-0.1586325616,0.2352561206,0.3178546727,-0.0731082335,0.3024133444,-0.1214098781,0.0119989626,-0.2797518969,0.1332839131,-0.0117351543,0.1140175834,-0.159745574,0.3371417224,0.0914066955,0.1130064726,0.2736600935,-0.1419514567,-0.050943207,0.1469705999,-0.6654461026,0.1463200301,-0.0904512107,-0.0697755963,-0.0807317495,0.1399653256,0.2095887661,0.0222028363,-0.0940629467,0.2417105883,0.4279995859,0.0152157107,0.0089434115,0.2975203991,0.2508084476,-0.1638221741,-0.0027235139,-0.0982123688,-0.2451073229,-0.3989139199,-0.1590597928,0.1913882792,0.2666960657,-0.1663199514,-0.2446895391,0.0393525884,-0.3511972725,0.0108301658,-0.1631431282,0.2812631726,-0.0030980257,0.306346029,0.0556032732,0.2700705826,-0.1631046236,-0.4479640126,-0.0052000051,0.1509830803,-0.3910146654,-0.0152986366,0.460850805,-0.0783595219,0.0757180154,0.1967018694,-0.098565042,-0.3172414303,-0.0278438255,-0.3539455533,0.1186480522,0.2464787811,0.1359597892,0.2408735752,0.3258979023,-0.2979562879,-0.1760974228,0.3370364904,-0.510586381,-0.2835355997,-0.0214070193,0.0790357664,-0.3109275103,0.216049701,-0.2705379725,0.2318892628,0.0025535435,-0.0811033621,-0.2396937907,-0.3311726749,-0.3782068789,-0.3062495589,-0.0184656214,0.5380911231,-0.7937141061,0.2113918662,-0.0439789854,0.0304449033,0.0645339489,0.2373640984,-0.3146848977,0.145540446,-0.3004151285,0.0451257639,0.2184087038,-0.1572796106,-0.3521279097,0.2249472886,-0.0438504554,0.1430745125,0.1946258098,0.1160269603,0.4747645855,0.0054960218,-0.1496740729,0.1738155186,-0.103760235,0.0363456868,-0.2112027258,-0.2809379995,0.2493326515,0.1949699521,0.1628943235,0.1472988278,0.1863788366,-0.4255873561,0.3947057426,-0.0318465978,0.217958495,0.0551677607,0.2408977449,0.0201239865,0.2374704331,0.001313109,-0.6431271434,0.2445382327,0.4549826384,-0.058966089,-0.0715787113,-0.1860950738,-0.3736642301,-0.085830465,-0.2312476635,-0.0666555539,-0.0176788997,0.0032555866,-0.0478532016,-0.1757822484,-0.1272353232,0.477886498,-0.2450702637,0.1183171272,-0.3356822729,0.3488133848,0.1532374322,0.0226258691,0.1873706132,-0.0584682524,0.2424250394,-0.1430356652,-0.0515100509,0.2048112005,0.3868688047,-0.1422166824,-0.0124562914,-0.1445665807,0.0729097873,-0.2080978006,-0.0453578681,0.1602115184,0.3658777177,-0.1580575109,-0.5012956262,0.4670168757,-0.2943073213,0.263843447,-0.1589294821,0.1330959797,0.233153671,-0.0610940196,-0.1029713675,-0.0019982792,-0.1220468134,-0.006638811,0.2533269227,0.1580813676,-0.2474706471,-0.1264657676,0.3704720736,0.2474065572,-0.2348777205,0.1483876258,-0.0330001377,-0.0938444063,0.2924904823,0.342839241,0.4789253175,0.1217185855,-0.2401447594,-0.0546514429,0.1157442108,-0.1957455575,0.0661826879,-0.015107044,0.1794178486,0.2044955045,0.084171623,0.0573599301,-0.0746098384,-0.1888594031,0.2203994244,0.1704453975,-0.5285298228,-0.006785084,-0.2358623892,0.2502968013,-0.3065753281,-0.0481774397,0.0667919666,-0.2848746777,-0.2406479418,0.2247037143,-0.129499495,0.0832543299,-0.1042544395,0.2468931675,-0.0310731418,-0.1913963109,0.1393317878,-0.1290226877,-0.2795024216,0.0142542943,0.0443121567,-0.0355804302,-0.0780696049,0.034434963,-0.0694945976,-0.2210642248,-0.0055247238,0.2116847932,-0.0453424938,0.0798723474,0.1930060089,-0.0438242257,0.0246098042,-0.1626574993,0.1271110326,0.2256253064,0.0464623943,0.220476836,0.2706637383,-0.0505867973,-0.1775081009,-0.764139533,-0.2559477091,-0.1818013638,0.0870506093,-0.1146430597,-0.0085540516,-0.0022579252,0.3254054189,-0.0912772119,0.2516673505,-0.1356215328,-0.0944925025,-0.2183258682,0.5146271586,-0.1794226319,-0.3070169091,0.2129276395,-0.1601766795,0.0302102249,0.3554811478,-0.4325180054,0.225927785,-0.0241388604,0.0817532241,-0.0706900135,-0.0414355844,0.1300362349,-0.0349057168,0.1392338574,-0.0643521845,-0.1901704669,0.3100559413,-0.0989008397,0.098874338,0.1584572494,0.4907527566,-0.1198672578,0.8878324032,0.0415000953,0.0302721038,0.1754058301,-0.0502931029,0.3468176126,0.1235480085,-0.2270592302,0.0012114414,-0.1244316772,-0.3443960249,0.3532143831,-0.0356557481,-0.1530629694,-0.1682810336,0.2079945803,-0.2423226833,-0.2623269558,0.2239594758,-0.1331666112,-0.0297012143,0.0486553535,0.2154509425,0.1233167276,-0.1159010604,-0.171932891,0.2245167941,0.2053622305,-0.0583073683,0.0913787484,-0.1581435651,-0.2692704797,0.1932853311,-0.0391291305,0.3117917776,0.1488640308,-0.0657137483,0.0432025567,0.2029720545,0.7653227448,-0.228073597,-0.0244639702,0.1349873096,-0.1019808874,-0.0461988002,-0.3012251556,-0.0984894037,0.2842860222,0.0264640469,0.2631985545,-0.1154613942,-0.0627299398,0.3587510288,0.1263293624,-0.2580283284,-0.3287102878,-0.1299742907,0.0623990186,-0.1074779481,-0.163281858,0.1607128233,0.0724806339,-0.3396209776,-0.0296115056,-0.202733323,-0.1998030841,0.3576449156,0.1423184425,0.2183056325,0.0453952774,0.2475808561,0.135484755,0.188069582,0.2552326322,0.8236984015,-0.2230930477,-0.8741761446,0.277115643,-0.4476580024,0.3850242794,0.1348782629,-0.3043456078,-0.0478809439,-0.2633568645,-0.1017216444,-0.0708947778,-0.038050212,0.1352691501,-0.0860932097,-0.3213091791,-0.5509390831,0.4983870089,-0.2160193175,-0.152677983,0.1288544089,0.050186798,-0.3783160448,0.4786064625,-0.0916960984,0.7594624758,-0.1534635872,-0.0095579447,0.1201103851,-0.1898284405,0.5620492697,-0.3316332698,-0.0203491673,-0.473703295,-0.14791812,-0.1999812126,0.0326113142,0.1313994825,0.6535795331,-0.2025294453,0.4069699943,-0.0216778405,0.388076365,0.0278303586,0.2338407189,0.302188158,-0.1964590251,-0.1149774641,0.0562805347,-0.0653756335,0.0055887178,0.0434991643,-0.0228690039,-0.2929770648,-0.098129943,-0.2989468873,0.2829633951,-0.0461634994,0.0756249875,0.0363135971,-0.4664282203,0.52027601,0.4734298885,0.050010331,-0.0365422219,-0.2011891454,0.0731039569,-0.0223002657,0.0527495369,-0.1005424485,-0.0462411642,0.1261733323,0.2554222345,-0.2163132131,0.2950339615,-0.1905880272,0.0038558026,0.2529632449,0.3160943687,0.2802587152,-0.5410072207,-0.3653708994,-0.1273264736,0.1805908084,-0.0940900221,0.0280139856,-0.0596976541,0.1222131997,0.0425459445,0.051162947,-0.397280544,0.0528306663,0.1606616378,-0.0169754215,0.1448103189,0.4566460848,0.3148107529,-0.1113458648,-0.1686692238,0.1123681217,0.1194436029,-0.5705486536,0.1371758282,0.0857450664,0.2909768224,0.1026712507,0.1484766006,0.1521828175,-0.2443030775,0.0830208287,-0.4805288017,0.0864085704,-0.0174754839,-0.1169790104,0.0771264434,-0.0416627303,-0.2976436913,-0.0453313626,-0.0287650246,-0.1823899448,0.0598155148,0.0349877067,0.0892363787,0.158857584,0.4067537189,-0.2238025665,-0.2627977729,-0.0229016058,-0.0026906878,0.0822317228,-0.0843560845,-0.2125822902,0.1974669248,0.0379100032,0.2562304139,-0.2446068376,-0.1998939365,-0.2053586245,-0.1485075504,0.1421973556,0.3760248423,-0.047276888,0.4416263103,0.1379790902,0.0541282743,-0.5553631186,0.3322670758,-0.2706056833,0.2641060948,0.1230383813,0.2363655716,0.0792644769,0.1204858199,-0.0828914344,-0.0638344735,-0.1384110302,-0.1215425059,0.2717137933,-0.2314950973,0.3026963472,0.1328430325,0.3071193993,0.3615554869,-0.0719616637,-0.096796602,0.1469042152,0.0921405032,-0.3382647634,0.2435209006,0.1269054413,0.21628353,-0.045146212,0.4322556853,-0.0617987812,-0.0199855529,0.12371815,0.1230706349,0.5733159184,0.1306252033,0.2434351891,0.2835692763,-0.0492272824,0.1505297422,0.2348001897,-0.1263485551,0.0890480503,0.4723775685,-0.2439918518,0.0161353182,0.2719652057,0.1843554974,-0.0243693963,-0.5952978134,0.2227465957,0.2796722353,-0.3082632422,-0.1141301915,0.1221162453,0.542779088,-0.1207558438,0.0791807696,-0.3032690585,-0.0884521008,0.0508166179,-0.0339204855,-0.1227207854,-0.2430635244,-0.4158157706,0.2870648801,-0.1807208955,-0.1423523426,0.3956698179,0.1432180703,-0.0040157083,-0.4216868281,0.1565317512,0.1227784678,0.1232950911,-0.2992652953,0.0295118559,-0.1005281582,0.0502832346,0.3894432485,-0.1678595692,0.2446740717,0.4575298727,0.0600555129,-0.3788633645,-0.0480150804,-0.0691348538,-0.0939064249,0.3936230242,-0.0502735712,0.2647095621,0.3244104385,0.0758297071,-0.0198911913,0.292480737,0.1600425839,0.0725020245,-0.5550231934,-0.0429757163,0.0614253543,-0.0587637052,-0.0343561545,-0.1492106616,-0.1095791161,-0.2929442227,0.3705632985,0.132608071,0.2546764016,0.0892879665,-0.0064055175,-0.0946696028,0.1924179494,0.241294533,0.0998176858,-0.5010780692,0.008691689,-0.7816181779,-0.0986604691,-0.254185617,-0.1950033903,-0.0570685938,0.27555269,0.2792519331,0.2758787274,-0.1493148953,0.2054737061,0.0526693128,0.3565359712,-0.2402208149,-0.2829420269,-0.2615454495,0.0218832064,-0.0611356944,-0.4771634042,0.0334574133,0.3663001359,-0.1833497286,-0.1305167824,-0.0736192241,0.213641867,-0.1871750206,0.3538753092,0.0364601798,0.502486527,0.034453012,0.31734851,-0.3440791667,-0.1859062165,-0.2605512738,0.2560063899,-0.0559963174,0.4194791913,-0.3687512577,0.1282571703,-0.1930968612,-0.00863274,0.098813124,0.2830576599,-0.3722726703,0.2633211911,-0.1280460507,0.2847011089,-0.0594868138,0.0075211315,-0.1713931113,0.1335120499,-0.134759292,-0.1800507307,0.2812542915,-0.207953617,-0.2370752245,-0.0244700313,0.187266916,0.212369144,0.0568695925,-0.5106430054,-0.1206396744,0.4468553364,-0.1862429678,-0.0663676113,0.055176083,0.0801299065,0.0314652659,-0.1960247159,0.2442516685,0.2878347635,-0.2607948482,-0.091153115,-0.4131068587]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"> Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n> \r\n> `from datasets import load_dataset`\r\n> `dataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")`\r\n> \r\n> Displayed error:\r\n> `... ArrowInvalid: CSV parse error: Expected 2 columns, got 1`\r\n\r\nI'm also facing the same issue when trying to load from a csv file locally:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('csv', data_files='sample_data.csv')\r\n```\r\n\r\nError when executed from Google Colab:\r\n```python\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-34-79a8d4f65ed6> in <module>()\r\n 1 from nlp import load_dataset\r\n----> 2 dataset = load_dataset('csv', data_files='sample_data.csv')\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 547 # Download and prepare data\r\n 548 builder_instance.download_and_prepare(\r\n--> 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n 551 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 461 if not downloaded_from_gcs:\r\n 462 self._download_and_prepare(\r\n--> 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n 465 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 535 try:\r\n 536 # Prepare split will record examples associated to the split\r\n--> 537 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 538 except OSError:\r\n 539 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 863 \r\n 864 generator = self._generate_tables(**split_generator.gen_kwargs)\r\n--> 865 for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n 866 writer.write_table(table)\r\n 867 num_examples, num_bytes = writer.finalize()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 213 def __iter__(self, *args, **kwargs):\r\n 214 try:\r\n--> 215 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 216 # return super(tqdm...) will not catch exception\r\n 217 yield obj\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/tqdm\/std.py in __iter__(self)\r\n 1102 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1103 \r\n-> 1104 for obj in iterable:\r\n 1105 yield obj\r\n 1106 # Update and possibly print the progressbar.\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/datasets\/csv\/ede98314803c971fef04bcee45d660c62f3332e8a74491e0b876106f3d99bd9b\/csv.py in _generate_tables(self, files)\r\n 78 read_options=self.config.pa_read_options,\r\n 79 parse_options=self.config.pa_parse_options,\r\n---> 80 convert_options=self.config.convert_options,\r\n 81 )\r\n 82 yield i, pa_table\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: CSV parse error: Expected 1 columns, got 8\r\n```\r\n\r\nVersion:\r\n```\r\nnlp==0.4.0\r\n```","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":319,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\n> Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n> \r\n> `from datasets import load_dataset`\r\n> `dataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")`\r\n> \r\n> Displayed error:\r\n> `... ArrowInvalid: CSV parse error: Expected 2 columns, got 1`\r\n\r\nI'm also facing the same issue when trying to load from a csv file locally:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('csv', data_files='sample_data.csv')\r\n```\r\n\r\nError when executed from Google Colab:\r\n```python\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-34-79a8d4f65ed6> in <module>()\r\n 1 from nlp import load_dataset\r\n----> 2 dataset = load_dataset('csv', data_files='sample_data.csv')\r\n\r\n9 frames\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 547 # Download and prepare data\r\n 548 builder_instance.download_and_prepare(\r\n--> 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n 551 \r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 461 if not downloaded_from_gcs:\r\n 462 self._download_and_prepare(\r\n--> 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n 465 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 535 try:\r\n 536 # Prepare split will record examples associated to the split\r\n--> 537 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 538 except OSError:\r\n 539 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 863 \r\n 864 generator = self._generate_tables(**split_generator.gen_kwargs)\r\n--> 865 for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n 866 writer.write_table(table)\r\n 867 num_examples, num_bytes = writer.finalize()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 213 def __iter__(self, *args, **kwargs):\r\n 214 try:\r\n--> 215 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 216 # return super(tqdm...) will not catch exception\r\n 217 yield obj\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/tqdm\/std.py in __iter__(self)\r\n 1102 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1103 \r\n-> 1104 for obj in iterable:\r\n 1105 yield obj\r\n 1106 # Update and possibly print the progressbar.\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/nlp\/datasets\/csv\/ede98314803c971fef04bcee45d660c62f3332e8a74491e0b876106f3d99bd9b\/csv.py in _generate_tables(self, files)\r\n 78 read_options=self.config.pa_read_options,\r\n 79 parse_options=self.config.pa_parse_options,\r\n---> 80 convert_options=self.config.convert_options,\r\n 81 )\r\n 82 yield i, pa_table\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/_csv.pyx in pyarrow._csv.read_csv()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n\/usr\/local\/lib\/python3.7\/dist-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: CSV parse error: Expected 1 columns, got 8\r\n```\r\n\r\nVersion:\r\n```\r\nnlp==0.4.0\r\n```","embeddings":[-0.0300141871,0.1694913059,0.0235360079,0.2714547515,0.2753945589,0.1793165207,0.4045257568,0.27959463,0.2873067856,0.0150397848,-0.0485231131,0.3499826193,-0.1212732792,-0.068123661,-0.1520148665,-0.1536566764,-0.0122299837,0.1196543425,-0.4033187032,-0.0239689164,-0.3553748131,0.3309987187,-0.2975638509,0.0998900831,-0.0714402348,0.1982583553,-0.0852632672,0.3722810149,-0.0867927074,-0.3423321545,0.5356943607,-0.0946969464,0.5112503767,0.2374755144,-0.000117985,0.0406568088,0.3306674957,-0.1105561927,-0.2995376289,-0.6809139252,0.0257156584,-0.5392317772,0.4017808139,-0.1466551125,-0.0527872182,0.0457104594,-0.0339156613,-0.4215278625,0.3402782679,0.4766380191,0.1821974665,-0.0922983885,-0.1318363994,0.0157459285,0.4828670025,0.073752366,-0.1259507537,0.376285553,0.0207763575,-0.239270404,-0.1518488377,0.0073804525,-0.0787402168,0.255281359,0.2491716594,0.2138702422,-0.2403781712,-0.1153684705,0.01403144,0.1130977795,0.6585769057,-0.1782788783,-0.1866249293,-0.0933595672,0.0274395104,-0.2190636396,0.3825586736,0.411262095,0.009916761,0.3105875552,-0.1560350358,0.056958098,-0.2717018127,0.2379808724,0.1065417454,-0.0040133619,-0.1554982364,0.311968416,0.0586880073,0.0225891359,0.2661928236,-0.1402395517,-0.0138087049,0.2710196674,-0.6992922425,0.3220380843,-0.0580023006,-0.0070773927,-0.0701758564,0.143386364,0.3351231813,-0.0486308858,-0.0925272033,0.3708167374,0.2820994854,0.1138467863,-0.0387672856,0.2824321389,0.2630428374,-0.2030360848,0.0290505793,-0.1195268556,-0.278203696,-0.509021163,-0.1248255,0.224749893,0.2519869506,-0.2572544217,-0.1120413765,0.0733333454,-0.2260328829,0.0147897396,-0.0744889155,0.2470282465,0.0601703785,0.3175560236,0.0203008205,0.2700637579,-0.1704099178,-0.3315723538,-0.0171890836,0.1262030602,-0.3891074359,-0.0880584419,0.5497084856,-0.1239738017,0.1671730429,0.1720090955,-0.1229457706,-0.2391269058,-0.0686041415,-0.2752360702,0.048061531,0.2800689936,0.1143449247,0.1900819838,0.3180615008,-0.1938027889,-0.1355801076,0.2522106469,-0.385945946,-0.2324071527,-0.195453614,0.090378508,-0.321360141,0.0626311451,-0.3915347159,0.015699001,0.0457716361,-0.2120894641,-0.1808320135,-0.3092979789,-0.3017327785,-0.2416182309,0.0150502138,0.439994216,-0.6876297593,0.202816084,-0.1004167125,0.0604222976,0.1709206104,0.2772907615,-0.298163563,0.1015653536,-0.2844896615,0.0790061653,0.2385806441,-0.2647067308,-0.3350452483,0.359580934,-0.0717722252,-0.0879997015,0.085076347,0.0455226339,0.3238097727,0.0071840598,-0.1462555081,0.3527291119,-0.0047068587,0.1135263741,-0.238440603,-0.1901738048,0.2679091096,0.2576404512,-0.0618824847,0.1129786223,0.2957235277,-0.4787016213,0.3347389102,-0.1028870493,0.2226001024,0.1331427842,0.1572675556,0.0099994345,0.218157202,0.0542478301,-0.6148228049,0.1521190852,0.2315641493,-0.0791522712,-0.2202433795,-0.0017674118,-0.3930812478,-0.0305651203,-0.2295912206,0.0227669589,-0.0195139833,0.0806080624,0.0607920885,-0.1256914139,-0.0473370999,0.3860825002,-0.2677580118,0.0173374768,-0.4094599485,0.306309551,0.0902065188,-0.0382477455,0.2489789575,0.0863740221,0.2496588975,-0.1442183256,-0.2104988992,0.2916672528,0.2853559256,0.0471600108,0.0695529282,-0.2708469331,0.0942480564,-0.213366732,0.0322463699,0.2261844873,0.3391343653,-0.0313163921,-0.5053218007,0.397185266,-0.3178347647,0.2437009811,-0.1135058179,0.1233608425,0.345356822,-0.0892557874,-0.0802449137,-0.0036903089,0.1109050438,0.1402174681,0.2556707561,0.1141963229,-0.2472608835,-0.1746401638,0.1861750036,0.1400503814,-0.1960377693,0.213381201,0.0154024968,-0.1066111475,0.3125395775,0.2956349254,0.4571997523,0.087733224,-0.2595733404,0.0056093941,-0.0213326346,-0.2613018453,0.0891569406,-0.0073952861,0.0612976886,0.2624716759,0.2842783034,0.0976949036,-0.1761980355,-0.1439563781,0.1858266294,0.2496437132,-0.5318591595,0.045227129,-0.3609481156,0.1762495041,-0.2729723752,0.0102986777,0.0426914208,-0.2635087967,-0.3743283749,0.250363946,-0.1324359477,0.0816713944,-0.1343074292,0.2072207183,0.0467177778,-0.2297455966,0.1525468528,-0.2799641192,-0.3581955731,0.0062141931,0.1478461474,0.0675947666,-0.0600930229,0.0656328052,-0.0216071587,-0.2102790177,-0.0601661056,0.1272987425,0.0478489771,0.1065663546,0.1082932204,-0.1527350247,-0.1061812192,-0.1949448586,0.2128024846,0.0481399782,-0.032766521,0.3588487506,0.2019795477,-0.1589808315,-0.2023465782,-0.630129993,-0.401289463,-0.2943170667,0.1171190888,-0.1430850029,0.0297290646,0.1440817416,0.2027333081,0.0192646347,0.2147850841,-0.0837452039,-0.0446465909,-0.2041229755,0.5625194907,-0.3179401159,-0.4271067381,0.1390203238,-0.0806052983,0.0882797316,0.3525590897,-0.4951308072,0.1878040284,-0.1395660788,0.0535571463,0.0482543707,0.0520809889,0.0155839585,-0.0864479989,0.0881131589,-0.0037086506,-0.1890086681,0.1955624372,0.035877265,0.0729663149,0.2429578751,0.4747933149,-0.2480227947,0.5842331052,0.0097642895,0.0339329801,0.3762377799,-0.100591749,0.3321916759,0.1103129387,-0.2510380149,0.1159814671,-0.0786157772,-0.2959708869,0.3519629538,-0.0678921342,-0.1585746557,-0.1452058107,0.2150488496,-0.2901725769,-0.2035695016,0.2112045586,-0.0854677111,0.1272023618,0.0595780611,0.054494746,0.0178848952,-0.3537164927,-0.1255517304,0.1114378944,0.1716856956,-0.0553593114,0.0629697517,-0.196658209,-0.3097870052,0.2110659927,-0.0279056989,0.524882257,0.0256070308,-0.0179026425,0.0413185246,0.1429857016,0.5542986393,-0.1822309792,0.1075029746,0.1693184674,-0.0439377539,-0.2650451064,-0.2759802341,-0.0867696106,0.2865701616,0.0068582906,0.2773997188,-0.0021761395,-0.0606578737,0.3453439474,0.1594411433,-0.220526889,-0.4335669577,-0.1399963945,-0.0273316186,-0.1427411586,-0.2897568643,0.0066125784,0.2067679167,-0.2675772905,0.102711007,-0.1962901056,-0.1330083013,0.2451456189,0.1147702634,0.2343942523,0.1104077324,0.1611224562,0.1183592081,0.2257016599,0.0926594958,0.7872273326,-0.1599137038,-0.7599287033,0.2973036766,-0.4518717527,0.3619061112,-0.001382218,-0.2888433635,-0.0669871494,-0.2123679668,-0.1393371522,-0.1512573063,-0.1044284031,0.0254814308,-0.0071146358,-0.0671476051,-0.4919825494,0.556945622,-0.1400377452,-0.0984783024,0.0469172113,-0.1453376412,-0.3126710653,0.5209854841,-0.1284262687,0.5789877772,-0.1519059837,-0.0054508732,0.2000327855,-0.1395255625,0.4839040339,-0.2072645724,-0.0124839032,-0.4585461318,-0.0315013304,-0.152623266,-0.0345811918,0.1453834772,0.6091606021,-0.1539794952,0.3765082657,-0.013852085,0.322267592,0.0654192939,0.2620986104,0.3155660629,-0.2835987806,-0.3397400975,0.0922290608,-0.1965550333,0.0049896161,0.1000767052,-0.0706438348,-0.2413423061,-0.0535262451,-0.297670871,0.3831697106,0.0292883515,0.1549091935,0.1162338927,-0.5188087821,0.5603206158,0.440164417,0.1686895788,-0.0252272896,-0.1860709637,0.0463445261,-0.0480936766,0.0278106164,-0.1102224067,0.0982132182,-0.023759814,0.1585789919,-0.169118315,0.1771288663,-0.1403254718,-0.1148723215,0.3304219842,0.2738518417,0.1867150664,-0.3853925467,-0.3837366402,-0.2151712626,0.1129150689,-0.0398342237,0.0504122525,-0.0296674091,0.1127496436,-0.0301914085,0.1432684958,-0.3665798903,0.144713521,0.278531462,-0.1977973431,0.0565510131,0.4927229285,0.4739419222,-0.0741870254,-0.1500219405,-0.0280874223,0.0692967325,-0.5587103963,0.0750577971,0.1246412843,0.2721292675,0.0574427806,0.1689198613,0.1549792737,-0.1622064859,0.0998938084,-0.710355401,-0.0916901305,-0.0616350397,-0.1580159366,-0.0111920889,0.1162052378,-0.3119654357,-0.016569633,-0.0504602566,-0.1851052642,0.0479075164,-0.0239854772,0.0872046128,0.2771908641,0.2218031883,-0.226923421,-0.2990846336,0.0183526967,0.0005218192,0.0555062257,-0.130814448,-0.184629336,0.19107458,0.0090253623,0.0584627688,-0.3238874078,-0.2742693722,-0.1618220359,-0.14653413,0.0760336518,0.2272851467,-0.0637922734,0.4846363366,-0.0836571828,-0.0416973121,-0.4219721556,0.2852167487,-0.2585068643,0.2118567675,0.0728896037,0.249730438,0.1491447538,0.0933686793,-0.2618623972,-0.0494161546,-0.2759154141,-0.0645886809,0.2769887745,-0.375610739,0.2140119523,0.1697146297,0.352368325,0.3072981238,-0.1214207187,-0.1546291709,0.1908563673,0.0932114646,-0.3811018169,0.1589248925,0.0928864554,0.3418718576,-0.0889320597,0.1998246759,0.1019363403,-0.2631264925,-0.0558813363,0.13679488,0.6348989606,0.0320949592,0.0673783869,0.3244942427,0.0105597517,0.1128842756,0.1104467884,0.0207565129,0.0966800898,0.4950858355,-0.1185559705,-0.1133052185,0.3128224611,0.3223560452,-0.1348772496,-0.6856020689,0.2883699238,0.339371115,-0.1629763097,-0.0226084385,0.1849779636,0.6478062272,-0.0429786481,0.0582773462,-0.2627935112,0.0404914729,0.1200712919,-0.0829428211,-0.2660669982,-0.1276918352,-0.3850884438,0.3157810867,-0.2050881088,0.0343599021,0.4355418086,0.2526460886,-0.0578061044,-0.5292367339,0.2077122778,0.2296304703,0.0024830736,-0.2734231055,0.1571275741,-0.0257155653,0.0132471379,0.2618347406,-0.1036355644,0.3829801381,0.524379909,-0.0099792434,-0.2622888982,-0.1545640826,-0.0780448467,-0.1029748097,0.46793589,0.0346287042,0.2928476334,0.3257479668,0.1019153967,0.0164211411,0.3097040355,0.1844928414,0.1199439093,-0.2552144825,0.028373044,0.1363503486,-0.032304842,-0.1254562587,-0.1230392307,-0.1966532767,-0.1693419218,0.474527806,0.2143625915,0.2723410428,0.0460844785,0.0024913312,-0.157891959,0.3675529361,0.4769450128,0.1246797442,-0.4641588628,-0.0805023834,-0.8023317456,-0.0381992236,-0.1752462238,-0.1333916783,0.0795962662,0.2359370738,0.2038174272,0.2918545902,-0.2377463132,0.1682191342,-0.1148635074,0.2035628557,-0.2020422369,-0.184230268,-0.1682485044,0.0981217921,-0.1045111343,-0.4657826424,0.0697082803,0.2155240476,-0.1185950562,-0.1940643787,-0.059023682,0.0045161946,-0.1609342247,0.2894135714,0.0564139597,0.4611916244,0.0201220345,0.2057639956,-0.4807578027,-0.2439059317,-0.1527059376,0.28661713,-0.0924916789,0.4068799913,-0.3040652871,0.1196668819,-0.2718935609,0.0706932098,-0.0065751858,0.1731262356,-0.2251237482,0.173098594,-0.1398609877,0.1454409212,-0.1590025574,-0.0354489759,-0.1513582021,0.2548293769,-0.1024668217,-0.2797609568,0.2372237146,-0.3887586594,-0.2771699429,0.1228235886,0.1983723789,0.0480458699,0.0820777342,-0.4914503992,0.0580233969,0.3590140939,-0.1576883048,-0.0586697273,0.0771085247,0.013606838,0.0447380319,-0.1328226477,0.1817386001,0.2837622464,-0.2722812593,-0.1785669923,-0.3695840836]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi @kauvinlucas\r\n\r\nYou can use the latest versions of `datasets` to do this.\r\nTo do so, just `pip install datasets` instead of `nlp` (the library was renamed) and then\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='sample_data.csv')","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":38,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi @kauvinlucas\r\n\r\nYou can use the latest versions of `datasets` to do this.\r\nTo do so, just `pip install datasets` instead of `nlp` (the library was renamed) and then\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('csv', data_files='sample_data.csv')","embeddings":[-0.0319091491,0.0998910815,0.0016058994,0.3416559994,0.2253115177,0.1857520193,0.3781754375,0.2843493819,0.3076332808,-0.0046042749,-0.0736300051,0.3338021338,-0.1202119589,-0.036175251,-0.2130927294,-0.0587122701,0.0360811427,0.0771625936,-0.3771968484,0.0088428417,-0.3649479747,0.3057532012,-0.3254077137,-0.0012915325,-0.0824024007,0.2179437876,-0.0744925812,0.3262184858,-0.1025668606,-0.3898982704,0.4821476638,-0.080058597,0.4992660582,0.2201852053,-0.0001179857,0.0123486947,0.3326276839,-0.0737969577,-0.2836776376,-0.5951529145,0.0139686335,-0.5699070096,0.3534557819,-0.115568459,-0.0387592912,0.0617986023,-0.0708740726,-0.3653351963,0.3587920666,0.5203021765,0.1995920837,-0.2032059133,-0.0811686441,-0.0118461316,0.5204197764,0.0610235557,-0.1259580702,0.2708447278,0.0428973995,-0.2233492732,-0.0964197516,0.0951562971,-0.1126321852,0.2678712606,0.2468886822,0.2386394143,-0.1517352313,-0.1456529647,0.0518766902,0.1153152809,0.6635107398,-0.1355136186,-0.176729992,-0.1062554643,-0.0070097786,-0.1568265557,0.3422959745,0.3546644747,-0.0102011804,0.3242204189,-0.0250324197,0.0301253237,-0.2691831589,0.1886892468,0.1149083152,0.027992826,-0.1570668072,0.3057209253,0.114892371,0.0086925505,0.2700324059,-0.0869944915,-0.0580076687,0.2113421112,-0.6860401034,0.3148464859,-0.1057037562,-0.0131408172,-0.0355927013,0.2131419182,0.2888286412,-0.0168030355,-0.087785773,0.370107919,0.2730760276,0.0119341025,-0.0131305298,0.2724969983,0.2526120842,-0.2041951269,0.0607698746,-0.1565530896,-0.2978444397,-0.5129746199,-0.0728966221,0.1486257911,0.2486549616,-0.1994403005,-0.181474179,0.1242232621,-0.2517411113,-0.0170060508,-0.1177749485,0.260204196,0.0609210394,0.3025761843,0.0645253956,0.2838415802,-0.1195834279,-0.3835787475,-0.0249955859,0.1081288904,-0.384224385,-0.0725212768,0.5409681797,-0.1791391969,0.1588603854,0.2351213247,-0.1358587444,-0.3058106601,-0.0346773267,-0.2870726883,0.0495085083,0.2614042759,0.0752800852,0.2058995217,0.3096886575,-0.2199794948,-0.0928175375,0.3182603121,-0.4290420711,-0.1496896595,-0.1494023353,0.1122689098,-0.2766397595,0.019795645,-0.3394761384,0.0994089991,0.0399336889,-0.2738412917,-0.1835940629,-0.2900147736,-0.2794313133,-0.2576267421,0.0108211935,0.4110125899,-0.7598640919,0.2014146149,-0.0385610051,0.063506037,0.1321390569,0.2012669444,-0.2380695045,-0.0085036838,-0.2470524311,0.1170615256,0.2001416832,-0.2251778692,-0.3695258498,0.2039723098,-0.0277520828,-0.0621384047,0.1256328523,0.0767565146,0.3356224895,-0.0383712798,-0.1593500674,0.2200727016,0.020736536,0.0405360498,-0.2515647113,-0.2080706507,0.2807472646,0.2186114639,-0.017313784,0.0922422409,0.2300778776,-0.596401751,0.328918159,-0.0431947187,0.2586236,0.0781924874,0.2754150033,0.0668581799,0.276987046,0.0377102606,-0.5350408554,0.1619918793,0.4143816233,-0.1803888232,-0.1539355814,-0.0427342318,-0.3967543244,-0.0268209055,-0.1542127728,-0.0362929888,0.0212248582,0.0724720359,-0.0212220177,-0.16581586,-0.1091084406,0.3955930769,-0.3020714223,0.0671919063,-0.368129313,0.2836926281,0.1172193065,-0.0429604799,0.2385161221,0.0490527786,0.2982416749,-0.0532653332,-0.1485976726,0.2183762789,0.2961485684,-0.0237372331,0.0841621459,-0.2446893901,0.0538063347,-0.239152804,-0.0022596212,0.2839276791,0.3712147474,-0.0852704495,-0.5303786993,0.3305390477,-0.2939954698,0.2343984544,-0.1652853191,0.1570280343,0.30540061,-0.0682469532,-0.1116520762,0.0074358834,0.1065215394,0.0153380921,0.251903981,0.1161820367,-0.2326779962,-0.1420400143,0.105782263,0.1679900587,-0.2043682039,0.2531194687,-0.0756668299,-0.11570815,0.3229525387,0.2924413681,0.4490132928,0.1327143162,-0.2260429859,-0.0074027041,0.0159655362,-0.2307155132,0.0959510878,-0.0015275988,0.0104517154,0.2142614424,0.2492847443,0.1078535989,-0.1244596541,-0.0946541205,0.1780858785,0.2668474913,-0.5732170939,0.0274307393,-0.2653020322,0.275398016,-0.2149259597,-0.0302288085,0.1330321133,-0.2786577046,-0.3122953773,0.3066865206,-0.0924999565,0.0600423031,-0.1841714382,0.2245592475,0.0431744792,-0.1701749712,0.0359234363,-0.2332860082,-0.3239174187,0.0234433077,0.1193632409,0.0211772751,-0.009588981,0.1299151629,0.0245018937,-0.2708138525,-0.061675638,0.081739895,0.0176403709,0.100418292,0.1105277836,-0.1548156738,-0.0887109637,-0.1159626618,0.2029960454,0.0429197177,0.045604378,0.3967782259,0.2087472975,-0.149520278,-0.1410145611,-0.6391364336,-0.3571048975,-0.2650282681,0.0800263733,-0.1071311012,-0.0271581337,0.1139522865,0.277887851,-0.0299414881,0.19760032,-0.0752973184,-0.1388437152,-0.2092313468,0.58896029,-0.2381600291,-0.3440220654,0.2079728991,-0.077859439,0.0502680503,0.3534047604,-0.451241076,0.1194554567,-0.0669652373,0.0515395552,-0.0284980275,0.0880324692,0.0203698091,-0.0561100505,0.076953657,-0.0587273389,-0.1870900691,0.1889620125,-0.0395994671,0.0470475703,0.2158955038,0.4871893227,-0.17796655,0.5998911858,0.1483035535,0.0594503656,0.3233929276,-0.0991083905,0.319134295,0.1864452958,-0.2664167881,0.0995654166,-0.1026838869,-0.2561929822,0.3076239228,-0.0021679311,-0.1340125203,-0.0833596215,0.2881454825,-0.3194304705,-0.1619986445,0.1712899655,-0.0611390285,0.0499433316,0.0561219342,0.094966583,0.0194642935,-0.2995353937,-0.1109833866,0.0919160023,0.1619724333,-0.0799226686,0.0896489993,-0.1746819764,-0.3334073126,0.2311434895,-0.106622301,0.466958046,0.0766970217,-0.0314852111,0.063958399,0.179696098,0.6740596294,-0.2698614001,0.0136661455,0.1487313807,0.0062908228,-0.2670135796,-0.231546402,-0.0807920247,0.2888073623,0.0171443913,0.2355082333,-0.0577477552,0.0189796761,0.38669765,0.1330913156,-0.1693070978,-0.4248486161,-0.1115750819,-0.0922456086,-0.1378076971,-0.278554976,0.078789793,0.0951751918,-0.2379315943,0.0914721042,-0.2159522623,-0.1552424729,0.3093272448,0.1034007668,0.2509729564,0.0663090944,0.1835236996,0.2088621557,0.1656487435,0.1542244852,0.7836306095,-0.2060810775,-0.8606128693,0.1866773069,-0.4865391254,0.3519999981,-0.0379060842,-0.3236556053,-0.0384505019,-0.2536812723,-0.1381228864,-0.1436244398,-0.1135772541,0.0669509023,0.0202170163,-0.1511394382,-0.5407761931,0.584284842,-0.0988450348,-0.1432559639,0.1049453691,-0.083362177,-0.3322691023,0.4414898157,-0.1370904446,0.6065697074,-0.0918170586,-0.0183178894,0.1956824064,-0.1937958002,0.4673049152,-0.2404603809,-0.0483806767,-0.4810394049,-0.0766469687,-0.133880198,0.0181350149,0.1137555912,0.5482631922,-0.2174833715,0.371859163,-0.0522970334,0.3556436598,0.113758482,0.1765640527,0.2892500758,-0.2167294621,-0.2115511298,0.1327034235,-0.1762226671,-0.0496775582,0.0918722823,0.0414727554,-0.2431035936,-0.0586544387,-0.3483195901,0.2833224833,-0.0396465026,0.0894251317,0.1615220904,-0.5426152349,0.5322107673,0.4186404347,0.1603853703,-0.0692004338,-0.1993330866,0.0857054219,-0.0149807995,0.0396604948,-0.1163498014,0.0064520123,-0.0793641061,0.2109176368,-0.2009479403,0.263461262,-0.0781920478,-0.0890198052,0.2962186337,0.2987450063,0.2094630152,-0.4391548336,-0.4187197983,-0.2294868976,0.1130284145,-0.1019381136,0.0660437271,0.0144572975,0.0806369707,0.0152487559,0.084503822,-0.332015872,0.1660001278,0.2239613235,-0.2200846374,0.0675337985,0.4705270529,0.4601638019,-0.115125671,-0.1658548564,0.0118581979,0.0688948706,-0.562055409,0.0882047266,0.1442666054,0.2655460835,0.1164891049,0.1526083648,0.1590538919,-0.2141202986,0.0994581878,-0.6414242387,-0.0282890536,-0.0687486678,-0.1128343716,0.0191989411,0.0454149134,-0.4401465952,-0.0679347813,-0.0427924804,-0.2041550726,0.1093608588,0.0837499797,0.0677037537,0.2318925261,0.2830806077,-0.2703290284,-0.30738765,0.0274370145,0.0335030071,0.0787499174,-0.1534463018,-0.2020764947,0.1672978252,0.0078801084,0.1011727005,-0.2869077921,-0.2297091186,-0.2180497795,-0.1598300785,0.0565553866,0.2896663845,-0.0289052408,0.4706905782,-0.0412926562,-0.1117323786,-0.4391998649,0.2942407429,-0.2483570874,0.1980830282,0.0932521746,0.2220350057,0.1999329627,0.1284148842,-0.1848100275,0.0201671012,-0.1850952208,-0.128226161,0.2976302505,-0.3851639628,0.2280173451,0.2646118104,0.3228787184,0.3634847701,-0.1232189834,-0.0728207678,0.1645085067,0.1382203698,-0.4042392075,0.1750418097,0.1052817404,0.3410072029,-0.0651910752,0.2365866303,-0.0005588638,-0.2272938788,0.0137957381,0.1443219185,0.6307868958,-0.017970141,0.071678564,0.2516496778,0.0317493975,0.0858843178,0.1555887312,-0.0269959196,0.0680792332,0.4835270941,-0.1209281012,-0.0494444072,0.3966093063,0.2534804344,-0.2115227729,-0.6216854453,0.2509863675,0.3503720164,-0.2413451523,-0.0964915678,0.1905129701,0.7036423087,-0.0018420574,0.0893176347,-0.2806842029,-0.0770020038,0.1189150289,-0.1277201474,-0.2567797899,-0.1458923966,-0.3834769428,0.2625007033,-0.2112941295,-0.0272011086,0.4355693161,0.2089945525,-0.0739797801,-0.6115416884,0.1231908947,0.2633067071,0.0291670226,-0.2828443944,0.1036590338,-0.0319648758,0.0002865648,0.3149217367,-0.104939498,0.4309365451,0.4255015552,0.0442368202,-0.2811281085,-0.2040180117,-0.0756347999,-0.1283522844,0.494276911,0.0512205772,0.2805725038,0.2355047315,0.1105442643,-0.0171552729,0.3259221911,0.1368455738,0.1284111142,-0.4042657614,-0.0327764489,0.0683988929,-0.090345934,-0.1030429006,-0.2066888362,-0.1445633918,-0.1984061897,0.4890687466,0.2041657418,0.2407859862,0.1066623479,0.0175768696,-0.1467791945,0.2029802799,0.4359095395,0.0723897666,-0.4763014615,-0.0362267196,-0.8368656635,-0.0388572849,-0.2405024171,-0.1494382322,0.0304138493,0.2102050334,0.2651368678,0.3032228947,-0.1305641234,0.1635681987,-0.0295826327,0.2831856608,-0.1725388914,-0.1970801651,-0.1247223765,0.1123401374,-0.0759784505,-0.4329157472,0.0330928303,0.2475735694,-0.1090585664,-0.1960031688,-0.0595827289,0.060254503,-0.2196055353,0.3812231421,0.0339942724,0.4314745665,0.0173779465,0.295524776,-0.4949624836,-0.2224110216,-0.1963848174,0.2198836356,-0.1356467158,0.4301290512,-0.3860015869,0.0560084581,-0.1707784981,0.0034308806,0.0037432923,0.1853771806,-0.3079570532,0.2642029524,-0.1082593501,0.2179622948,-0.1841491014,-0.067771852,-0.1615976989,0.2248890251,-0.0729279742,-0.2904005647,0.2474670112,-0.3145464361,-0.2629497647,0.0616531186,0.2380640954,0.1233009174,0.1169598177,-0.5120447278,-0.0280583147,0.3570630252,-0.1396004409,-0.0915261582,0.1276018769,0.0424169675,0.0625734031,-0.1565251797,0.2157999426,0.3222095072,-0.2383171916,-0.1307475269,-0.3803531826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi \r\nI'm having a different problem with loading local csv. \r\n```Python\r\nfrom datasets import load_dataset \r\ndataset = load_dataset('csv', data_files='sample.csv') \r\n``` \r\n\r\ngives `ValueError: Specified named and prefix; you can only specify one.` error \r\n\r\nversions: \r\n- datasets: 1.1.3 \r\n- python: 3.9.6 \r\n- pyarrow: 2.0.0 ","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":42,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi \r\nI'm having a different problem with loading local csv. \r\n```Python\r\nfrom datasets import load_dataset \r\ndataset = load_dataset('csv', data_files='sample.csv') \r\n``` \r\n\r\ngives `ValueError: Specified named and prefix; you can only specify one.` error \r\n\r\nversions: \r\n- datasets: 1.1.3 \r\n- python: 3.9.6 \r\n- pyarrow: 2.0.0 ","embeddings":[-0.050101459,0.078890942,-0.0220736191,0.340208441,0.2207659781,0.1807745844,0.4585178792,0.2670554519,0.2990832925,0.1005545333,-0.1566747874,0.2878015637,-0.0605567135,-0.0458843485,-0.2373053581,-0.1309784204,0.0242509339,0.1026108935,-0.3309668601,-0.0062435013,-0.3678452671,0.3093583584,-0.2815257311,0.0530989505,-0.0545527861,0.2814908326,-0.1077686772,0.3274840415,-0.0620354079,-0.3914218247,0.5134364963,-0.060867846,0.5080314279,0.3084244728,-0.0001191223,0.0650712252,0.3402869403,-0.1058762297,-0.2548354566,-0.6348005533,-0.0315943509,-0.4776789248,0.407428503,-0.1194225177,-0.0851382241,-0.0584867708,-0.1120312735,-0.3718575835,0.3140225708,0.5228904486,0.1772251129,-0.1155418307,-0.1115062535,0.0169778336,0.5616163015,0.0575069636,-0.1525167078,0.4509942234,0.1394277513,-0.2184303999,-0.0677595586,-0.0209364668,-0.0959598944,0.3456400633,0.2188953161,0.271359086,-0.1611296088,-0.1651814878,0.062767975,0.1464377642,0.6752828956,-0.1511281431,-0.2087820768,-0.0510118008,-0.0382058881,-0.1732246578,0.3668098748,0.3682530522,-0.0369195677,0.3713665307,-0.0950149521,0.1026137918,-0.3096706569,0.1899678111,0.0927571505,-0.0103784641,-0.1794596016,0.3294480741,0.0513701737,0.008640456,0.3520628512,-0.1739129722,0.030434018,0.2446758151,-0.6652464271,0.3353160918,-0.1578666717,0.0001058316,-0.0663317069,0.1045316383,0.2552757263,-0.0045825816,-0.0678219125,0.3870537877,0.2508233786,0.0504334569,0.0290666539,0.2726481557,0.2801621258,-0.1642535776,0.0055883639,-0.13599433,-0.3793538511,-0.4904814959,-0.051405102,0.1927994192,0.2272313833,-0.2714297473,-0.1685214788,0.1140698716,-0.2511085272,-0.027104212,-0.127081722,0.2699306607,0.0280062798,0.289981693,-0.0452711247,0.2199497223,-0.090474017,-0.2733890414,-0.0209608693,0.1261294186,-0.3207287192,-0.0600758083,0.4831056595,-0.2676691413,0.1819711328,0.1681607068,-0.0794380307,-0.2476108819,-0.0827695802,-0.2674682438,0.0243966077,0.3265566826,0.0672450438,0.2302682847,0.3057488799,-0.2579580843,-0.110199362,0.3398165703,-0.432371676,-0.2835350633,-0.1473101974,0.0990742519,-0.2438331544,0.1004437879,-0.2957794368,-0.000056291,0.0635556579,-0.1275317669,-0.1993082017,-0.3081683517,-0.2920990288,-0.2420302629,0.0490508564,0.4407739937,-0.8225072622,0.1986010224,0.0250368826,0.027388351,0.1153561622,0.2424414903,-0.3161163032,0.004800037,-0.2935258448,0.0552282892,0.1316671222,-0.2720192075,-0.3150507212,0.2806831598,-0.0270450246,-0.0634991825,0.141702652,-0.0138663761,0.2569593489,0.0029977325,-0.1404381543,0.2176546454,-0.0065031582,0.0761716589,-0.259701997,-0.2113527954,0.207602948,0.2757164538,-0.0061859931,0.108214274,0.2752881348,-0.6618198752,0.3255362809,-0.0791740343,0.236444965,0.1412426382,0.3167905509,0.0424468964,0.313113004,0.0630818233,-0.5907900929,0.1977384835,0.2924916446,-0.1200514212,-0.2475693822,-0.0772442147,-0.4460267723,0.0104096755,-0.239118576,0.0370681062,-0.0059876153,0.0592396036,0.0421137661,-0.1589002311,-0.1259151995,0.3727186918,-0.2769035995,0.0706972629,-0.3248630166,0.2909721136,0.0696870834,-0.0526059791,0.1898505986,-0.0392167829,0.325453043,-0.1290652752,-0.1919662058,0.2637890279,0.3160148561,-0.0336085074,0.0590862334,-0.2230442315,0.077670902,-0.1986172646,-0.0506002419,0.2209782749,0.3760454059,-0.0419052914,-0.5225034952,0.3844500482,-0.3002507985,0.2447667867,-0.1516911387,0.0497826561,0.2744937241,-0.0778861195,-0.0841779634,0.0022267736,0.0433275625,0.0710103884,0.3040778935,0.167039156,-0.2275003493,-0.237296477,0.069299534,0.1240548417,-0.2722059786,0.2221032828,-0.0933093727,-0.0296445992,0.32342273,0.2225303799,0.450609237,0.0627956241,-0.2948082089,-0.0311818495,0.1055251732,-0.2698894441,0.0893970057,0.0152474185,0.008845645,0.2963629663,0.2396397144,0.0683367476,-0.1314299703,-0.1208921894,0.2028779238,0.2157005668,-0.5725854635,-0.0012655248,-0.2637578547,0.3105990291,-0.2766802609,-0.0802132711,0.1246065497,-0.223446086,-0.3464516997,0.3340693414,-0.1314715147,0.0806201622,-0.1163329557,0.181334421,0.055045642,-0.2825608253,0.0345876925,-0.1984506696,-0.3249219656,0.009058564,0.0852184221,-0.0181425288,-0.0775880441,0.0292050336,-0.0280321594,-0.2400929779,-0.0159601588,0.1261330843,0.0647608042,0.1917006224,0.0844557285,-0.0688858628,-0.1227831692,-0.1982715428,0.2833007574,0.0806418806,0.0428979993,0.2975507677,0.1895868182,-0.1723712981,-0.1275824606,-0.6349033713,-0.4294931591,-0.2392594665,0.1076039448,-0.118669793,-0.0019448254,0.0566577166,0.2694094777,-0.0629708469,0.2377326936,-0.1118587926,-0.0665362552,-0.2725560069,0.5867142081,-0.2760928869,-0.3623427451,0.1451467127,-0.1022788063,0.0377784446,0.2768310606,-0.3733101785,0.1132083982,-0.0661438033,0.1471009403,0.0610943921,0.0687354133,0.02736862,-0.0744934455,0.0918395296,-0.0271730609,-0.2052680552,0.1610162705,0.0298476685,-0.012604502,0.3149689734,0.4116819501,-0.2683644593,0.7119854093,0.0053101219,0.0461543575,0.3699708879,-0.1068791896,0.2614112198,0.1299588233,-0.2914157808,0.0598405264,-0.1390175819,-0.361502707,0.3220266402,-0.0177444685,-0.1171161309,-0.1575265378,0.1837616712,-0.3016541898,-0.1233405992,0.1480936408,0.0486130901,0.0678979978,0.0312702805,0.0231347643,0.0952631757,-0.2626835704,-0.0269714594,0.1845025122,0.1151131317,-0.0825677812,0.0579194799,-0.1454227418,-0.2668076754,0.2060009539,-0.0304789413,0.463822335,0.1204189584,-0.038091097,0.0751824081,0.139124617,0.6658430696,-0.1872533858,0.0402411558,0.1610817462,-0.0728144124,-0.2294280976,-0.3184125125,-0.0606936775,0.2017460763,-0.0700996667,0.3571449816,0.0216822121,-0.0156351943,0.353515476,0.1525818557,-0.195493862,-0.5041841269,-0.1489433497,-0.0295130517,-0.1700459421,-0.2663184404,-0.0026112925,0.169308871,-0.2428933382,0.0588111505,-0.1647517085,-0.1062650234,0.2855170667,0.1203782186,0.1944304854,0.0343126543,0.2272322178,0.1703966558,0.2108255178,0.2185293436,0.7849230766,-0.2690060735,-0.8792741299,0.2317835242,-0.4429601729,0.3828547597,-0.0565310307,-0.3647814989,-0.0973411873,-0.1691260785,-0.1317820698,-0.1059450507,-0.0430910066,0.020401841,-0.0467944331,-0.0806568787,-0.4824158549,0.5758624673,-0.1096653044,-0.1644360125,0.103181012,-0.0994907096,-0.2497218549,0.4641616642,-0.1504679769,0.5777961016,-0.1825271547,-0.0528247133,0.1385999918,-0.2188474387,0.2997281253,-0.180532679,-0.0591463484,-0.5199486613,-0.1374166757,-0.1423643827,0.0187158901,0.1822494417,0.5470429659,-0.11068663,0.398213923,-0.0904222876,0.484508872,0.0546464622,0.1585680246,0.364418596,-0.1809336543,-0.193136245,0.1023848131,-0.1136552095,0.0771946386,0.1125712544,0.0335527286,-0.2811995447,-0.0472303182,-0.2457051426,0.275659889,-0.0848595947,0.0824685544,0.1637711823,-0.5518720746,0.4269961417,0.4201009274,0.1565538347,-0.0616821647,-0.2027568072,0.04663001,0.0201693811,0.0526571497,-0.1770026535,0.0165129174,-0.073630169,0.2172013968,-0.1906293035,0.2250493318,-0.1424359977,-0.0726065859,0.2998256981,0.2638697624,0.3123708069,-0.4656921625,-0.3996197879,-0.3334167004,0.1910330951,-0.0796642527,0.0545595139,-0.0669752583,0.0864299536,-0.0383711755,0.0369112156,-0.3518146574,0.1033335254,0.2805387378,-0.1340022683,0.0681498274,0.443764627,0.4114876688,-0.0924189091,-0.1417542994,0.0123451203,0.1805324107,-0.5031499863,0.0629161522,0.164639473,0.2261803895,0.0708832368,0.2218476236,0.2093816251,-0.1926318556,0.1231275275,-0.5760722756,-0.1013640612,-0.1048061848,-0.0483660586,0.0492364131,0.0740978569,-0.3924596906,-0.0293158554,-0.0528812744,-0.1870634407,0.1170195788,0.0693912581,0.0727759674,0.2216253877,0.3139066398,-0.1698289961,-0.2607459128,0.0497959778,0.0246517602,0.0938417986,-0.1491534114,-0.1748726964,0.1690513641,0.0386072062,0.0781580508,-0.3387264907,-0.2455083281,-0.1488036215,-0.1550678015,0.1201864481,0.3304129243,-0.0106807938,0.4284214377,0.0258767139,-0.1009341106,-0.4691784978,0.3606740534,-0.2741923034,0.2413979322,0.1107829511,0.172846809,0.1472524852,0.0840285718,-0.131484434,-0.0908034518,-0.1636728644,-0.0913070142,0.2350666821,-0.3960172832,0.1661299914,0.2181178033,0.3637970686,0.3214376271,-0.1575125605,-0.0788168162,0.1981801987,0.1224990115,-0.3466894329,0.1106185615,0.065676257,0.2738384306,-0.0489147007,0.2328171432,0.1088710278,-0.2285522223,0.05112027,0.1101180166,0.697028935,0.1044492275,-0.016417671,0.294200778,0.0353478938,0.1553337127,0.1816571504,0.0114567084,0.0506067611,0.4985149801,-0.166196391,-0.1359436959,0.361710012,0.3011926413,-0.1136225238,-0.5898247957,0.3065499961,0.2861317098,-0.2570914626,-0.1384093314,0.2038695812,0.6618592739,0.0233908854,0.102483876,-0.232744351,-0.0765229836,0.129336074,-0.1519845277,-0.2526064515,-0.1717599928,-0.3143319786,0.2734127343,-0.1854747236,-0.0579471849,0.5163354874,0.1813938916,-0.0574016795,-0.5273650289,0.1054735705,0.2769794762,0.0459995046,-0.2584888041,0.1292857379,-0.007651221,-0.0329293609,0.2504449785,-0.1367973089,0.4047161341,0.4960860312,0.0311213546,-0.3256340623,-0.2005137056,-0.0616444871,-0.0888142958,0.4282944202,-0.077324681,0.2897395492,0.2834821641,0.1197699457,-0.0191068817,0.3488196731,0.1070305705,0.1097613275,-0.3893099725,-0.0922246277,0.0806775913,-0.0843076035,-0.0558481216,-0.196643889,-0.1175943613,-0.1686717868,0.5300334692,0.1841522008,0.3017850816,0.1599747092,0.0007537697,-0.1638134867,0.2719569504,0.4349094927,0.103567116,-0.492909193,-0.0291486885,-0.7631896734,0.0415348001,-0.1710262299,-0.0912178457,0.0520508103,0.1729028523,0.2211141437,0.3072668016,-0.2144627124,0.1830375344,-0.0511868,0.3276064098,-0.1933553517,-0.1775848567,-0.20316495,0.1117483452,-0.0412390791,-0.4847297966,0.1184746549,0.1818851084,-0.1158415303,-0.226164639,-0.0943899304,0.1627674997,-0.2500812113,0.4163210988,-0.0341711566,0.4087325633,0.0086558033,0.3287061155,-0.3305092454,-0.1907590926,-0.1695470363,0.2170059383,-0.1300562471,0.3343851268,-0.30078125,0.0276948083,-0.2088326663,0.0200406164,0.0203171689,0.2079422027,-0.1729598194,0.1353712827,-0.0705585331,0.1225851923,-0.1733812094,-0.0880132094,-0.1297494024,0.2034527808,-0.0837133229,-0.2797818184,0.3309130371,-0.3489595354,-0.2443291694,-0.0096810861,0.1662707776,0.0950971618,0.115621075,-0.4329107702,0.0243950933,0.4081395268,-0.1634295881,-0.1169580072,0.0899385065,0.068726629,0.0980335623,-0.1716623902,0.1761628985,0.3099908233,-0.2136202306,-0.2278588414,-0.3829801977]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Oh.. I figured it out. According to issue #[42387](https:\/\/github.com\/pandas-dev\/pandas\/issues\/42387) from pandas, this new version does not accept None for both parameters (which was being done by the repo I'm testing). Dowgrading Pandas==1.0.4 and Python==3.8 worked","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":35,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nOh.. I figured it out. According to issue #[42387](https:\/\/github.com\/pandas-dev\/pandas\/issues\/42387) from pandas, this new version does not accept None for both parameters (which was being done by the repo I'm testing). Dowgrading Pandas==1.0.4 and Python==3.8 worked","embeddings":[0.0241772942,0.12562038,0.0306745227,0.319699347,0.2818748057,0.1347274333,0.470521152,0.3493146896,0.2720313966,-0.0102589251,-0.1092102081,0.2939046621,-0.044851318,-0.0265678819,-0.3008044064,-0.1147533208,0.0091162156,0.0720298886,-0.4082455635,0.0407130383,-0.4287291169,0.284797132,-0.3110737801,-0.0556099638,-0.152755484,0.1461137533,-0.1490888745,0.2877849042,-0.1212894768,-0.399774164,0.4710716605,-0.0675764605,0.4576898217,0.2826908827,-0.0001190207,0.0470557921,0.3922930956,-0.1001775339,-0.2630170882,-0.5131826997,-0.0591822937,-0.4227257073,0.3733281195,-0.0852224305,-0.046688281,0.0004197524,-0.1162619591,-0.2997397184,0.3223545551,0.5077928901,0.197917968,-0.1602941602,-0.0620777011,-0.0502253436,0.4172463417,0.1050913036,-0.1397542655,0.2756552696,0.0337564647,-0.1532249302,-0.180978328,0.0272843409,-0.1014828458,0.2782543898,0.1798395813,0.2050615847,-0.0613027588,-0.2106291652,0.1289485842,0.2036833167,0.6754380465,-0.078778483,-0.2856331468,-0.1039790958,-0.0243786722,-0.1393211037,0.3564254642,0.2502124608,0.0016349813,0.3509472907,0.0232161768,0.0296683796,-0.2598004043,0.1540524513,0.0373894386,0.0158521775,-0.1935314983,0.2507592142,0.0340466313,-0.0823894143,0.3526718318,-0.1788177937,-0.0623942763,0.197615087,-0.6042397618,0.2810893059,-0.0836109594,-0.0234648902,-0.1294538528,0.2142701298,0.2488881201,-0.0499460101,-0.1401538253,0.2775976658,0.252376765,0.0532183312,-0.0278310105,0.2208503634,0.3117902875,-0.1449721903,0.1107845604,-0.1051988527,-0.222037822,-0.5018793344,-0.0677434355,0.0968574733,0.3531314433,-0.1722039282,-0.2210885882,0.1006415039,-0.2622009814,0.0220399573,-0.1620692015,0.2267710119,0.0502322465,0.4545515478,0.0701662973,0.2115022689,-0.0344944336,-0.3089322746,-0.0769338682,0.2031660825,-0.3034771979,-0.0684907511,0.4976694286,-0.2253912538,0.1549333036,0.3578968942,-0.1139468104,-0.2443407625,-0.1136397198,-0.3365449309,0.0902300701,0.3289268613,0.0469029807,0.2032093704,0.2896735072,-0.2509704232,0.0028574821,0.3258141279,-0.4054673314,-0.156913951,-0.0999016613,0.1198122278,-0.2537450194,0.0574827828,-0.3325140476,0.1230519116,-0.0078405384,-0.151597634,-0.1227682382,-0.2970442176,-0.2000421733,-0.2799708545,0.0212287959,0.5202496648,-0.7918003201,0.2299589813,0.0292756967,0.1033632308,0.0896141902,0.1875173002,-0.2415774614,-0.1062611863,-0.2405188531,0.0625411719,0.1165735126,-0.237686798,-0.3421730995,0.2998892069,0.0453347303,-0.0108899875,0.1983411163,-0.0019309281,0.2845632434,-0.04353882,-0.1288126856,0.126778692,-0.0160801522,0.0224632267,-0.2836357951,-0.2488529831,0.2401542813,0.1519769132,0.130537197,0.094353117,0.1428275257,-0.6783123612,0.3488560617,-0.0634345189,0.2517107129,0.1308165491,0.3389266133,0.0228330884,0.2688421905,0.0505939238,-0.5418136716,0.1698775589,0.5113579035,-0.1673701853,-0.1205927134,-0.0629021823,-0.3830545843,-0.0392312631,-0.099480696,-0.0518742502,0.017235551,0.0323562659,-0.0781595409,-0.1198081449,-0.1356744468,0.4858939946,-0.2923885882,0.0885697305,-0.258349508,0.256650269,0.1534156799,-0.1027728841,0.275652498,-0.0653968081,0.3416048288,-0.0444978252,-0.1651556939,0.2488504946,0.2802524865,-0.0365780033,0.0584948584,-0.2135780156,0.0895586982,-0.129989624,-0.0512957759,0.2908011973,0.3992747664,-0.0659131631,-0.4376945198,0.4224687815,-0.3204883039,0.2275464833,-0.0841154754,0.1403484344,0.3445750177,-0.0801549181,-0.152662009,0.0182648096,-0.0319590047,0.0388623103,0.3248752356,0.1877735257,-0.3184331059,-0.1537049115,0.0780854449,0.1375251859,-0.2164755464,0.2041727304,-0.0594577938,-0.1062221378,0.3829270005,0.3608201444,0.4717500806,0.1498554945,-0.2529484928,-0.0557890125,-0.01006338,-0.2269156724,0.1954914182,-0.0310393851,0.0481213443,0.1908816695,0.1868442446,0.0977427289,-0.1696167439,-0.0430323705,0.2352081537,0.2486244291,-0.6367043853,-0.0294911135,-0.2627077103,0.3036528826,-0.1164851189,-0.0980645567,0.1143851429,-0.3141474128,-0.3436886072,0.3097485006,-0.1371875107,0.0977562815,-0.3261144757,0.2158664763,0.0691104531,-0.1033549011,-0.0268367529,-0.1415639222,-0.2339593023,0.0365678035,0.0974037871,-0.0330904983,-0.0662379935,0.1117405221,0.0317247063,-0.288046211,-0.0658863038,0.0880052224,-0.0089438949,0.1357587576,0.1609303057,-0.0979804546,-0.1344276816,-0.1008068323,0.1499018073,0.0083998181,0.0724000111,0.2714380026,0.1862328798,-0.1533819735,-0.1734954864,-0.6161032915,-0.2489308417,-0.2709670365,0.0619678237,-0.1247730777,-0.058524888,0.135047093,0.365917027,0.021160759,0.3091208637,-0.0635993406,-0.1574359089,-0.2968179882,0.6772490144,-0.1667860597,-0.3600411713,0.2659082413,-0.0987677053,0.0851714388,0.343912065,-0.3236020803,0.0558186956,-0.117518723,0.0973820686,0.0033491177,0.1036384255,0.0825691447,-0.0110409744,0.0334409028,-0.0814219564,-0.1823910028,0.1710588038,-0.0009193884,-0.0323532522,0.2558740377,0.563256979,-0.1767581552,0.6726998091,0.0456445329,0.0922084004,0.3408700228,-0.0584755018,0.4342031777,0.2090545893,-0.2655888498,-0.0067741885,-0.0711784288,-0.181089431,0.2537304461,-0.0517102554,-0.0780986175,-0.1305233091,0.2020471692,-0.3098101318,-0.1612519026,0.1412205547,-0.0876443312,0.0256496929,0.050725203,0.1857020259,0.0174520761,-0.1868579239,-0.0818033442,0.1275239885,0.261594981,-0.1261100173,0.1983493865,-0.0240594838,-0.3371816874,0.2336102277,-0.0698892176,0.5109390616,0.1607391983,-0.1134153083,0.0573602468,0.1283015758,0.7509254813,-0.2501649559,-0.0441519469,0.187016353,-0.1603053808,-0.3192566633,-0.2150829732,-0.078167215,0.2765806615,-0.0009714197,0.2078567594,-0.0168184079,-0.0802662596,0.3738730252,0.042471081,-0.1355861574,-0.3771294653,-0.1334594935,-0.1177297384,-0.0732270777,-0.274385184,0.183987096,0.0483439527,-0.2974915802,0.1116578355,-0.3248671591,-0.2815507948,0.3783840537,0.0840915218,0.2123315483,0.1083485857,0.1812478453,0.1921921968,0.1203031167,0.2890277207,0.702907443,-0.2705733478,-0.8506482244,0.1913441122,-0.5073080063,0.3283641636,0.0872669294,-0.3635573685,-0.0722120926,-0.2784616053,-0.1500739604,-0.1070750058,-0.1296276897,0.1055824161,-0.043873854,-0.1721906364,-0.493545413,0.4711264074,-0.0656682029,-0.1855989248,0.1417224109,0.0517840907,-0.327922076,0.3642807603,-0.1267917901,0.5505488515,-0.0172249749,-0.0004293724,0.2155709267,-0.2124698907,0.3772599399,-0.0127177369,0.0168256816,-0.444534272,-0.1848396957,-0.1234467998,0.031282071,0.2225219905,0.4833863378,-0.1904896647,0.3246837556,-0.1199102476,0.4244746268,0.1669234335,0.1593374908,0.1721351147,-0.1484532803,-0.3074183166,0.1248734295,-0.0998324603,-0.1680870801,0.1148492619,-0.0113008572,-0.2571187615,-0.0778157488,-0.2717769146,0.2373145074,-0.0636966452,0.0961729214,0.1771979481,-0.4958979189,0.4706732929,0.4114856422,0.2012061775,-0.1171216443,-0.2064071,0.1093757376,0.0152817918,0.0764585882,-0.2028641105,-0.0343280621,0.0464251563,0.2529504597,-0.1738139093,0.3091202378,-0.1578205526,-0.0804691538,0.2528537214,0.2375118434,0.1309558898,-0.4937244952,-0.4120789766,-0.3213724494,0.1861267388,-0.1494590044,0.0768082365,-0.0576720089,0.1612201482,-0.0047300109,0.0211010501,-0.3172330558,0.104773432,0.342759341,-0.1670037955,0.1807068437,0.4932451546,0.3881204724,-0.1507330984,-0.1403361708,0.0555117279,0.1645432413,-0.5076981187,0.2135135084,0.1413078457,0.2084482908,0.1376141906,0.161295265,0.1666217446,-0.1460340172,0.091046609,-0.5139111876,-0.0772565752,-0.0654829666,-0.1267828792,0.0436272547,-0.0141349509,-0.4589305222,0.0014190257,-0.1196263283,-0.219667539,0.1575871855,0.1223779023,0.0131109953,0.2223804295,0.2580272853,-0.1582408845,-0.3068284988,0.0520422384,-0.0006026339,0.0987873077,-0.1344638169,-0.2471244186,0.1642010659,0.0322313756,0.0624058843,-0.2403770834,-0.2371185869,-0.280154109,-0.1346853673,0.0647224113,0.4185419381,-0.0397819728,0.4131694734,0.0035393555,-0.1441858113,-0.4067820013,0.2532990277,-0.2691311538,0.210089311,0.1308118403,0.198647514,0.2070214897,0.113695167,-0.0841311663,0.0257699415,-0.3353297412,-0.1486731917,0.256475091,-0.3860013187,0.1391390711,0.2921793163,0.250737071,0.3862535954,-0.1412204802,-0.0474280231,0.0925103873,0.1378982961,-0.380081445,0.1879015714,0.0773668215,0.2832441926,-0.0096325511,0.2677393556,-0.020547742,-0.2951770425,0.0983581915,0.1745265722,0.6494079232,-0.0390684158,0.1272902936,0.3346779346,0.0346802287,0.0982005596,0.2503215075,-0.0515310988,0.0501430891,0.4086123109,-0.1400688887,-0.0618080795,0.2793547511,0.2557901442,-0.2222185433,-0.5973098278,0.3044040203,0.2913835943,-0.3806108832,-0.1511500329,0.2046428919,0.697853446,-0.0141467014,0.1091274545,-0.2954770625,-0.0840990767,0.0314258039,-0.163073197,-0.2595496178,-0.2324003428,-0.3630485833,0.276663363,-0.1744512469,-0.0118660247,0.4851983786,0.1758655459,-0.0940839201,-0.5722904801,0.1402863115,0.1994098127,0.0795066506,-0.2412603348,0.0388635322,-0.0257792082,-0.0872339681,0.3151928782,-0.0733094513,0.4634872973,0.4264777005,0.0831427947,-0.3347374499,-0.0677053854,-0.1541987807,-0.0575158708,0.4183604419,0.0534323752,0.3305989802,0.1860104352,0.1443066746,-0.012256979,0.3317760229,0.0806537196,0.1279873699,-0.4893304706,0.0413609296,0.1082577407,-0.0676129758,-0.0256094895,-0.1773844212,-0.109637484,-0.2583309114,0.4641063511,0.1054635793,0.2055932432,0.0880961493,0.0172811728,-0.1367624402,0.1560656726,0.3946134448,-0.070170857,-0.4267624915,0.0598126426,-0.8037463427,-0.1125817597,-0.1431775093,-0.0794061199,-0.0164541956,0.2091906816,0.2341060191,0.1763246655,-0.0745594278,0.1277616024,0.0429232679,0.2066366524,-0.2172836363,-0.131273061,-0.1636935771,0.0422593355,-0.0842612013,-0.4247184098,0.0470632054,0.2354150265,-0.0874219537,-0.1702657044,-0.1204782501,0.0710048676,-0.1706643403,0.3737181425,0.0368576571,0.3752529621,0.0016463454,0.2292733043,-0.4571590424,-0.1647520512,-0.2008585036,0.2513170838,-0.1026316732,0.454159826,-0.3759488165,-0.0517935827,-0.160790965,0.0499484129,-0.0451475717,0.2026387751,-0.3483147025,0.3250577748,-0.1704608351,0.1564630121,-0.1707881093,-0.0988595262,-0.2129938751,0.2510595918,-0.0880642384,-0.3208688498,0.2898322046,-0.3992585242,-0.2384376228,-0.0725912899,0.2006807178,0.1523417532,0.1147531345,-0.4594575763,-0.0546600446,0.3439137042,-0.0973579139,-0.1922149956,0.0997206345,-0.0161417425,0.1078802273,-0.1731543839,0.1773280501,0.3391388953,-0.2554277182,-0.1896863431,-0.3545406163]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi, \r\nI got an `OSError: Cannot find data file. ` when I tried to use load_dataset with tsv files. I have checked the paths, and they are correct. \r\n\r\nversions\r\n- python: 3.7.9\r\n- datasets: 1.1.3\r\n- pyarrow: 2.0.0\r\n- transformers: 4.2.2\r\n\r\n~~~\r\ndata_files = {\"train\": \"train.tsv\", \"test\",: \"test.tsv\"}\r\ndatasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n~~~\r\n\r\nThe entire Error message is on below:\r\n\r\n```08\/14\/2021 16:55:44 - INFO - __main__ - load a local file for train: \/project\/media-framing\/transformer4\/data\/0\/val\/label1.tsv\r\n08\/14\/2021 16:55:44 - INFO - __main__ - load a local file for test: \/project\/media-framing\/transformer4\/data\/unlabel\/test.tsv\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-00a4200ae8507533 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/usr4\/cs542sp\/hey1\/.cache\/huggingface\/datasets\/csv\/default-00a4200ae8507533\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nTraceback (most recent call last):\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 592, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 944, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 307, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"run_glue.py\", line 484, in <module>\r\n main()\r\n File \"run_glue.py\", line 243, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 610, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 515, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 594, in _download_and_prepare\r\n raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\nOSError: Cannot find data file. ```","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":229,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi, \r\nI got an `OSError: Cannot find data file. ` when I tried to use load_dataset with tsv files. I have checked the paths, and they are correct. \r\n\r\nversions\r\n- python: 3.7.9\r\n- datasets: 1.1.3\r\n- pyarrow: 2.0.0\r\n- transformers: 4.2.2\r\n\r\n~~~\r\ndata_files = {\"train\": \"train.tsv\", \"test\",: \"test.tsv\"}\r\ndatasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n~~~\r\n\r\nThe entire Error message is on below:\r\n\r\n```08\/14\/2021 16:55:44 - INFO - __main__ - load a local file for train: \/project\/media-framing\/transformer4\/data\/0\/val\/label1.tsv\r\n08\/14\/2021 16:55:44 - INFO - __main__ - load a local file for test: \/project\/media-framing\/transformer4\/data\/unlabel\/test.tsv\r\nUsing custom data configuration default\r\nDownloading and preparing dataset csv\/default-00a4200ae8507533 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/usr4\/cs542sp\/hey1\/.cache\/huggingface\/datasets\/csv\/default-00a4200ae8507533\/0.0.0\/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2...\r\nTraceback (most recent call last):\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 592, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 944, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 307, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"run_glue.py\", line 484, in <module>\r\n main()\r\n File \"run_glue.py\", line 243, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 610, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 515, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 594, in _download_and_prepare\r\n raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\nOSError: Cannot find data file. ```","embeddings":[-0.1220383421,0.1353530139,-0.0049578166,0.4093729258,0.3437595367,0.1575909108,0.4911705256,0.3018618524,0.2409298271,0.0778130069,-0.1208111942,0.3113324344,-0.222933799,0.0161729231,-0.2666302919,-0.138849169,-0.0121987769,0.1486898512,-0.3797127008,0.0842762068,-0.2968249023,0.3350248337,-0.23477377,-0.0235737097,-0.2068452984,0.3015033901,0.0195733905,0.2683660388,-0.0678415745,-0.2882575989,0.4837555885,-0.1166169122,0.5780785084,0.5965259671,-0.0001198588,0.1980403811,0.2585474253,-0.1649317443,-0.2870138288,-0.6305498481,0.0434910543,-0.437597841,0.261670351,-0.1079069749,-0.1369617581,0.0029657511,-0.0563248098,-0.3104810119,0.3318168223,0.580542326,0.1549984962,-0.0283102971,-0.0582819022,-0.0010095276,0.6244791746,0.0946319997,-0.1630176157,0.3080536127,0.0226266012,-0.1955775917,-0.0984998941,0.0650285929,0.0058259796,0.1974367201,0.2176357061,0.2914288938,-0.173745811,-0.2069217116,-0.0035459534,0.1451856494,0.7800492048,-0.1723983586,-0.2525832355,-0.0038263958,-0.1281324029,-0.1057975665,0.3823305368,0.2369377762,-0.048913423,0.2923953235,-0.1913403571,0.0496631712,-0.3450913131,0.2236623913,-0.0214209072,-0.0197535418,-0.1582514346,0.2584479153,0.1375644654,0.0710336789,0.3356910646,-0.0396718346,0.0543107241,0.2524381876,-0.7086184025,0.1900539398,-0.1133599132,-0.1570455581,-0.1534983516,0.115812473,0.2388054878,-0.0914529413,-0.1756827533,0.3815530539,0.2372174114,0.1486689597,-0.129113555,0.2593872547,0.2610002458,-0.1225181893,-0.0553485043,-0.1889454573,-0.3713354766,-0.5662067533,-0.0335551873,0.1246437356,0.245372206,-0.1781110317,-0.1711992621,0.0048956862,-0.2986387014,-0.0057112207,-0.0554229245,0.2902826369,0.1472160816,0.3855616748,0.1423828453,0.1907602251,-0.1389487088,-0.2804300785,-0.041624818,0.2023255974,-0.3305784762,-0.0056452821,0.4941843748,-0.1304323524,0.1737867147,0.082842268,-0.091869615,-0.2563663125,-0.0123549318,-0.2786598206,-0.055250328,0.3757532537,0.0055698087,0.1856311411,0.2911657989,-0.2491188645,-0.0616967157,0.3790513277,-0.4626737833,-0.2957409322,-0.0280461628,0.0852449983,-0.3014573455,0.1072890311,-0.3230452538,-0.0427241176,0.1460918486,-0.2255543768,-0.1203534007,-0.3606698215,-0.235120222,-0.2219359726,0.00656028,0.4152827561,-0.7811157107,0.1281213909,-0.0042020045,0.0325595923,0.0963324159,0.288096875,-0.3107186258,0.1574350744,-0.2816325724,-0.0167383458,0.1991070509,-0.2672846615,-0.3795184493,0.1948359311,-0.1619416922,-0.0049851062,0.1852816343,0.1021047533,0.2606715858,-0.0783005506,-0.2450922728,0.1544238627,-0.0333498791,0.0658834651,-0.2152729034,-0.2210890204,0.2718627155,0.3141382635,0.0862169191,0.0300115757,0.1951025575,-0.3645020425,0.3027190566,-0.0952498019,0.1820617169,0.1559844464,0.3703453243,0.0269140936,0.3065308928,0.0473045781,-0.4875147641,0.1468679309,0.4081065357,-0.1211295053,-0.1660327464,0.0102656381,-0.3846558928,-0.0491423383,-0.2423173338,-0.090641655,-0.0145480921,0.0530857854,-0.0414341353,-0.1651814282,-0.1047927067,0.3845700026,-0.2736659646,0.0664697513,-0.3920545578,0.4043228924,0.0824075416,-0.055935666,0.0855056792,-0.0344316028,0.1692339033,-0.1672074199,-0.1799739301,0.2550437152,0.4084514976,-0.0330250636,0.0269570537,-0.0329795294,0.0791254118,-0.258315444,-0.1261023879,0.3344679177,0.3087880611,-0.0029713996,-0.4763065279,0.2504805326,-0.3923824131,0.181960091,-0.1012003496,0.0932749361,0.251886487,-0.119603321,-0.106306985,0.1031586528,0.0559722073,0.0228890888,0.2572350502,0.1939049363,-0.0633053109,-0.1562314034,0.1896046549,0.1293913126,-0.288569659,0.2658117414,-0.1528628021,0.0079371752,0.2042433619,0.3819603026,0.5386138558,0.0883824527,-0.293579489,-0.0668624863,0.1644527763,-0.2344348431,0.0807885826,0.0785160512,0.1907888055,0.3258116245,0.1997103393,0.1552238017,-0.0841264576,-0.3111198843,0.1669424623,0.2850769162,-0.5293880701,0.0387531556,-0.1866257936,0.338039577,-0.3329127729,-0.0377601497,0.0703953505,-0.2028545439,-0.3053887784,0.2585046589,-0.1360059828,0.0595632046,-0.1811581254,0.1324446499,0.1742648035,-0.2269573808,0.089508377,-0.1951800585,-0.220851317,-0.0164600331,0.0973833948,-0.1575862318,-0.1652231663,-0.0160056688,0.066575788,-0.2609189451,-0.1183249056,0.1628821194,0.0610100739,0.091319941,0.1020861417,0.0297820903,-0.0470242649,-0.2525389493,0.2062446028,-0.0025440597,-0.016047148,0.3312510848,0.1999717355,-0.0864830688,-0.1531871557,-0.7353489995,-0.3709748685,-0.3064213693,0.0816308483,-0.1460936219,-0.0306182746,0.2606905401,0.3235732019,-0.004563496,0.2491169572,-0.0899321437,-0.02238358,-0.28042382,0.6574794054,-0.2137465328,-0.2847048342,0.1995658726,-0.1170615554,0.0257779937,0.3571566641,-0.4444891214,0.1330472976,-0.0713608339,0.0335714556,-0.0613024235,0.0512672327,0.1103764027,-0.0150371455,0.1081173643,-0.0345507488,-0.0960643217,0.1769181639,0.0580694266,0.1803807318,0.2061363012,0.5240602493,-0.1935347915,0.752599895,-0.0069442852,-0.0061677769,0.2724443376,-0.0853926614,0.3250020146,0.1740597039,-0.2503384352,0.0761615038,-0.1681621224,-0.3343825936,0.3666778505,0.0272584725,-0.1200250611,-0.1239110976,0.1599295288,-0.2371294051,-0.1694990098,0.2616626322,0.0317293927,0.124811776,-0.0547839366,0.0938649625,0.0552779846,-0.2856650352,-0.0114622582,0.2788128853,0.245703131,-0.0685827807,0.1414246708,-0.0923432931,-0.2467011362,0.1020107269,-0.0651865005,0.4200946093,0.1193907112,-0.1087347269,-0.0435616821,0.1764675826,0.7794322371,-0.1432041824,0.0212296825,0.086968556,-0.1605240107,-0.2050446123,-0.3354282677,-0.1334927827,0.1190553382,-0.0339228399,0.3815132976,0.0613930933,0.0302610938,0.3845025599,0.1609823555,-0.256991297,-0.4833341837,-0.0821001977,-0.0507574566,-0.1799066812,-0.3221031427,0.0665794984,0.1676475853,-0.286970973,0.1186983809,-0.1045254543,-0.23453255,0.3253744543,0.0865179226,0.2651091516,0.0390847474,0.1910636872,0.2929503322,0.093632862,0.3215658069,0.7372086644,-0.1628676504,-0.9020167589,0.2740162611,-0.4344469905,0.3657584488,-0.0305296034,-0.3615438342,-0.1097800136,-0.2838808298,-0.0921163484,-0.0388168767,0.0345797092,0.030549245,-0.0115143536,-0.0776138827,-0.4349418879,0.4664115608,-0.123480536,-0.0214227065,0.0374093764,-0.1266365796,-0.3070187569,0.5303562284,-0.1694428921,0.6691302657,-0.1261789352,0.0582753234,0.1655172259,-0.0813237131,0.3352088332,-0.5055786967,0.048769135,-0.4333445132,-0.327960968,-0.1936734915,-0.0775998458,0.1067102924,0.6288593411,-0.2026550621,0.32326442,-0.0583984777,0.4683563113,0.1017164513,0.0682224184,0.4265602529,-0.1912725419,-0.1417884231,0.0763952583,-0.0988230258,0.1147894487,0.0959546119,-0.0519035496,-0.3455936313,-0.0138464533,-0.2423666269,0.2610686719,-0.1793634742,0.172583878,0.2288490534,-0.4907253683,0.4911367595,0.5247833133,0.0727556273,-0.0776696131,-0.2039031833,0.0855794773,0.0435255095,0.0373002291,-0.0899276584,-0.0227657277,0.0164169818,0.2192912102,-0.1826176941,0.2748560309,-0.1109125018,-0.184493646,0.3382415473,0.1942050308,0.0900050029,-0.4632531106,-0.3599356115,-0.2527321577,-0.0199256316,-0.0519251972,0.051641915,-0.0221761297,0.000784152,-0.0333643258,0.0866052881,-0.3998742998,0.1197555959,0.2619692385,-0.1787328422,0.0760382488,0.4830133617,0.4360223413,-0.1482409239,-0.1247603968,0.0865892097,0.149539426,-0.6645997763,0.0831816271,0.1039365605,0.2797349393,0.1543491185,0.1433707029,0.2814809978,-0.249594003,-0.0201755837,-0.6343808174,-0.1504680514,-0.0241676476,-0.0182663258,0.0828894153,-0.114701055,-0.3087159991,-0.0492930673,-0.0441081598,-0.183860153,0.0794505179,0.0156732854,0.078860648,0.2726916671,0.2159263194,-0.1990477294,-0.3160897791,0.0124449534,-0.0366629064,0.0632762536,-0.0981342494,-0.1809089035,0.1783587337,0.0285369772,0.0892830491,-0.2773477733,-0.1277479678,-0.1596653461,-0.1591578424,0.0903621614,0.2512242794,-0.0392178409,0.4274789095,0.0592345968,-0.1586241275,-0.5465249419,0.2915149331,-0.1882787198,0.328620404,0.2261325121,0.2161956578,0.1401555687,0.1189815924,-0.2003940046,-0.0395810753,-0.1273439676,-0.0579887964,0.257036984,-0.3816359341,0.1869713515,0.1779757142,0.3278340399,0.2979975045,-0.2113125473,-0.0658378229,0.1483927071,0.1105599254,-0.3501855135,0.1427778453,0.1776920259,0.2974821329,-0.0638365597,0.2781265974,-0.0291179009,-0.131716758,0.157521978,0.244769156,0.5773417354,0.1098440364,0.1476423442,0.3500671387,-0.0882378146,0.1736589968,0.1366865486,-0.0580663458,0.087748386,0.4587211609,-0.2164085954,-0.0720299706,0.1937769204,0.1926274747,-0.1403484643,-0.5329040289,0.0453971215,0.1736008227,-0.2526529729,-0.150647372,0.1242325008,0.5423734188,0.068657726,0.1494096369,-0.2550836504,-0.0606675223,0.052697815,-0.0530096851,-0.0698421001,-0.2602415085,-0.4337778389,0.2687378228,-0.1538171023,-0.0083177881,0.3635636568,0.2617110908,-0.1166488975,-0.3991003931,0.1071271971,0.2506076992,-0.0293979738,-0.2409007996,0.1203518361,-0.006111369,0.0315284915,0.2301202863,-0.1253584027,0.3768225014,0.4576132298,0.0243073534,-0.2782173157,-0.1779769659,-0.0580862425,-0.0733145773,0.4218493998,-0.0790282413,0.1666322351,0.2153225392,0.0950129032,-0.012566491,0.303709656,0.0253088083,0.0546890348,-0.3918221295,-0.0358554609,0.1125537232,-0.061074052,-0.0871239305,-0.1870794594,-0.1338915974,-0.1520318687,0.5085308552,0.1754084826,0.2575397789,0.1847787052,-0.0037028126,-0.1210379899,0.2251693755,0.4045380652,0.126837939,-0.5164458156,-0.0470704027,-0.8025571108,0.033315219,-0.1674451232,-0.0851123109,-0.0130050797,0.1274859309,0.2851581872,0.2845371366,0.0436287932,0.0738850683,-0.0118813626,0.2930821776,-0.1820256561,-0.2165475488,-0.1133944243,0.1580870897,-0.0652677938,-0.4554162621,-0.0155877322,0.2373419404,-0.0940121636,-0.2334360033,-0.0742748305,0.1010088846,-0.1628885716,0.3524152339,-0.1388456225,0.4168570638,-0.0341716185,0.282664746,-0.3049477339,-0.2810989618,-0.1505681574,0.165022701,-0.0602572933,0.4468724132,-0.3955821097,-0.0342179798,-0.2920238078,-0.0623516627,0.0651758909,0.1752175242,-0.2441074997,0.188214317,-0.1435171664,0.1499857455,-0.0294642095,-0.1104987785,-0.1713940203,0.1815409958,-0.119181864,-0.2850665748,0.324449122,-0.3510090709,-0.2308666557,0.0145493802,0.0453776531,0.0070409006,0.2257379293,-0.4288408756,0.0607375354,0.3522358537,-0.0931597203,-0.1766714156,0.0753208473,0.0886227861,0.1043147668,-0.1456499696,0.399692595,0.2324021459,-0.1665874124,-0.2668822706,-0.2740080655]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi ! It looks like the error stacktrace doesn't match with your code snippet.\r\n\r\nWhat error do you get when running this ?\r\n```\r\ndata_files = {\"train\": \"train.tsv\", \"test\",: \"test.tsv\"}\r\ndatasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n```\r\ncan you check that both tsv files are in the same folder as the current working directory of your shell ?","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":57,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi ! It looks like the error stacktrace doesn't match with your code snippet.\r\n\r\nWhat error do you get when running this ?\r\n```\r\ndata_files = {\"train\": \"train.tsv\", \"test\",: \"test.tsv\"}\r\ndatasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n```\r\ncan you check that both tsv files are in the same folder as the current working directory of your shell ?","embeddings":[-0.1632082611,0.1201281324,-0.0203899965,0.3711948693,0.2908929586,0.1906737834,0.3376256526,0.308806628,0.2558354139,0.1050071195,-0.1869195849,0.2054682523,-0.1593434066,0.061179474,-0.1956521869,-0.0624763109,0.0268127378,0.0686473697,-0.4645629525,0.0177216865,-0.3333188891,0.3470653892,-0.2780237794,-0.0542951822,-0.0767822489,0.2361236066,-0.0599401705,0.3367495239,-0.081656985,-0.3559603095,0.5194858909,-0.0635318682,0.560716033,0.5292677283,-0.0001137903,0.129445225,0.2825890183,-0.1467752755,-0.1785620749,-0.6393190026,0.0613104552,-0.4359839857,0.2645874918,-0.1104227826,-0.1284557283,0.0712635145,-0.0708982721,-0.4720506668,0.4504335225,0.5279279947,0.2148371339,-0.0864070952,-0.0118197463,-0.0221919268,0.7886179686,0.0594406687,-0.1394074708,0.2724424303,0.0125369942,-0.0810968876,-0.0302768908,0.0458028689,-0.0359600559,0.1779315323,0.1814219505,0.296615392,-0.1023824066,-0.1386378556,-0.0735867172,0.2100553215,0.7554205656,-0.1943683177,-0.1298263371,0.0654361546,-0.1378994882,-0.1567752808,0.406347245,0.2865529954,0.0064115208,0.3294068873,-0.1184835657,0.1312605739,-0.256188333,0.0950852111,0.098347187,-0.0493261069,-0.1786335856,0.2593528926,0.1806005985,0.0034185105,0.3713912666,-0.0517851748,-0.058962699,0.2235426456,-0.7684729695,0.3118459284,-0.0455594212,-0.1471838802,0.014965063,0.179177165,0.2139286548,-0.0393063612,-0.1528491527,0.3345857263,0.2363773733,0.1051969528,-0.1414199769,0.2154708654,0.2190301716,-0.1684966683,-0.102421768,-0.1876339912,-0.2559287846,-0.5735869408,-0.0645407438,0.2073688507,0.3200314641,-0.3114317954,-0.1733118594,0.0734636188,-0.2722924352,-0.0609482974,-0.109194845,0.2909167111,0.2314547151,0.2941367328,0.1858211011,0.1946026683,-0.1058751419,-0.2749941051,-0.0832266957,0.1312164217,-0.403624326,-0.1082926467,0.4074251354,-0.1435656697,0.0499262922,0.209153235,-0.044293914,-0.2685553432,-0.035612233,-0.3195300698,0.0450207554,0.2549913824,0.1352955699,0.1181483567,0.3289474249,-0.1746667027,-0.1117878854,0.3744624555,-0.4798728526,-0.1793120503,-0.0459833853,0.1310596317,-0.2370796353,0.0573173016,-0.289082557,0.0589945316,0.000939445,-0.1820271462,-0.2509436011,-0.3599371612,-0.2920160294,-0.2913584113,0.1218943521,0.4260993302,-0.8851076961,0.2679671943,-0.0419623069,-0.0201598052,0.1030402482,0.2525891662,-0.286568433,0.1934656799,-0.2951545119,0.0659349412,0.1927287132,-0.3063805699,-0.226077646,0.1231190339,-0.118889071,-0.0308844969,0.0815995261,-0.0159799103,0.2922246456,0.0229680575,-0.2450550199,0.1886671185,-0.0248834491,0.0727707818,-0.2346982211,-0.1997045279,0.2740665078,0.2147894949,-0.0159138534,0.0792820826,0.2235260457,-0.5767694712,0.3881706893,-0.1135374233,0.2037299871,0.1781434715,0.3660393059,0.1063058153,0.2167533636,0.0133437794,-0.4905884266,0.1591786593,0.3609518707,-0.1155055836,-0.0092602121,-0.0075382371,-0.450137645,-0.0408641472,-0.1863635927,0.0030228717,0.0847011283,0.0825764611,0.0114414012,-0.1034900472,-0.0824752748,0.3019010425,-0.1349658966,0.0457387194,-0.2683048546,0.3771105111,0.0310081206,-0.1467919499,0.2174808681,-0.0860394463,0.3234868646,-0.1837220341,-0.1866134852,0.3333313465,0.3146055341,-0.0872737169,0.0854247138,-0.159656167,-0.0496572256,-0.2291757762,-0.0167078115,0.2488171458,0.314687252,-0.1180462316,-0.4590246975,0.2613494396,-0.4094178975,0.206133008,-0.1435100138,0.0847721547,0.2101373971,-0.14181602,0.0210019778,0.0874420106,0.1191504151,0.0604729094,0.1988692731,0.005549375,-0.0911601782,-0.1056592017,0.119756341,0.1689518839,-0.2727974355,0.1789902747,-0.1137832776,-0.0162422266,0.242197454,0.357165277,0.4763866663,0.1523467749,-0.1742680818,0.0455711745,0.1098353192,-0.1504481435,0.0936007947,-0.0194585808,0.1371385455,0.2552471459,0.1668246537,0.0839905292,-0.1430705041,-0.2369464487,0.0455449112,0.2538726926,-0.5752357244,0.0083326316,-0.249729678,0.3108226061,-0.3310995996,-0.0051644901,0.1356990635,-0.2337109447,-0.3300172985,0.2485228926,-0.1445584297,0.0756820589,-0.1767652184,0.1835032552,0.1014495492,-0.1629744172,0.0509379804,-0.173088938,-0.3325943947,0.0628696084,0.1289216727,-0.0283848252,-0.061428573,0.0472806394,0.1133723781,-0.2062104195,-0.069977954,0.1670141667,-0.0763246343,0.0880631581,0.1179210618,-0.0030940298,-0.0499594063,-0.1978800446,0.246868819,-0.0530651137,-0.0031851677,0.3039884865,0.2936697304,-0.1829455346,-0.1477911621,-0.7467257977,-0.3022550941,-0.2975016534,0.0628438219,-0.0657057986,0.0176544636,0.1362267286,0.3105674088,0.0593861453,0.1433230191,-0.0918809026,-0.0467757061,-0.3576166034,0.5338152051,-0.2907546759,-0.2972563803,0.2983312905,-0.0351573192,0.0133842342,0.3558810651,-0.4415941238,0.1364328116,-0.0556880422,-0.0080279978,-0.06351161,0.1525565982,0.0533665977,0.019340897,0.0493189469,-0.0812957287,-0.2418559194,0.2346500009,-0.0742264614,0.1467011869,0.1442626268,0.5319949389,-0.2617813051,0.6968776584,0.0899335667,0.0690586641,0.3502433598,-0.1133134142,0.3590984643,0.1277568191,-0.2564528286,0.091887854,-0.0979034901,-0.3763469756,0.3155032992,0.006017589,-0.0746962503,-0.088450484,0.2623583078,-0.2288102359,-0.2408952862,0.1966506094,0.0378157422,0.1186003163,-0.1037179083,0.0290416218,0.0580466054,-0.2552058697,-0.0950018242,0.128601402,0.0879365802,-0.0659375712,-0.0196913797,0.0130823543,-0.1995950639,0.2741596401,-0.1319939494,0.3353087902,0.0291078202,-0.0947662964,-0.0097495858,0.1962525249,0.7513359189,-0.1963489652,0.0381776579,0.1308378577,-0.0254559182,-0.2028169185,-0.3558411598,-0.1370390058,0.1230921075,0.0332484208,0.1556124538,0.0182997975,0.0646016747,0.3633918464,0.1286439449,-0.1475206614,-0.4154252112,-0.1174383834,-0.1033015549,-0.114068605,-0.2285342813,0.1015130803,0.1351855248,-0.28753829,0.0572807901,-0.1101413518,-0.1874686033,0.2784743607,0.059882395,0.3630291224,0.0445536189,0.2284569889,0.2116134167,0.0818404183,0.2669923902,0.8654198647,-0.2275514156,-0.8078784347,0.1678776592,-0.465336591,0.2808261216,0.0099713504,-0.3242847323,-0.0899668708,-0.2835259736,-0.1283735037,-0.0162382908,0.0010448804,0.0133929299,0.0381392837,-0.0996670425,-0.4074403942,0.4686181843,-0.1068051234,-0.1038858145,-0.0314271301,-0.0486309677,-0.3610852957,0.5382452607,-0.1863855571,0.7635279894,-0.1556736976,-0.0716926008,0.1424181461,-0.2298340648,0.3213377893,-0.3747919202,-0.0774681419,-0.502373755,-0.2104349583,-0.1804514378,-0.0968683809,0.0697605535,0.597076118,-0.3314548135,0.3451715112,-0.1161097586,0.3273295164,-0.0137004917,0.1385611892,0.4446401,-0.1152485609,-0.0814957768,0.1851789057,-0.0905858949,-0.0137265194,0.0729703009,-0.0674788505,-0.2689510882,-0.0614796802,-0.353815645,0.3456997275,-0.2019076496,0.0655875877,0.1689490378,-0.5422196984,0.460337311,0.4039847553,0.1948414594,-0.0011996529,-0.1929984689,0.0927047879,0.0807235613,0.0506324247,-0.1227283329,-0.0151686519,-0.0294888746,0.1688002795,-0.2706636488,0.2851683497,-0.0561245792,-0.1420143992,0.2620805204,0.2446110249,0.1543709487,-0.5034883022,-0.288025707,-0.1574339718,0.11342448,-0.0749626681,0.0980294719,0.0046114763,0.0553123318,0.0296995956,0.1411348432,-0.2868427932,0.1739550084,0.1763403565,-0.2092242092,0.1472908556,0.4387862384,0.5524030328,-0.0883836523,-0.1231174991,0.1129119769,0.1212616563,-0.5274673104,0.1565487534,0.0258636568,0.3393357992,0.1452896148,0.236192137,0.2931863666,-0.1802963912,0.0669545904,-0.4703954458,-0.0330226831,-0.0938968435,-0.1051436141,0.1869398504,-0.0565323047,-0.2851337492,-0.038804397,-0.0609747581,-0.2626056373,0.0676803589,0.0063291625,0.0691457838,0.2368307859,0.1666108072,-0.1574672163,-0.2865997553,0.0556916483,-0.0622217357,0.0121746501,-0.1614169478,-0.1675748825,0.1348129809,-0.0042217169,0.096193701,-0.2614148259,-0.1657780111,-0.1452402472,-0.1733609438,0.0625734851,0.1997066885,-0.0398008078,0.3725084662,-0.025244059,-0.172689721,-0.4611665308,0.2609953582,-0.2078049183,0.1719646901,0.1307312548,0.2049126178,0.11233785,0.0092727663,-0.2228019983,0.0354624912,-0.0008793259,-0.0371561013,0.2153118998,-0.301107347,0.0429497622,0.2243585885,0.3837178051,0.2965293527,-0.1746602654,-0.0378762856,0.1227534637,0.1579858065,-0.3390794396,0.2375322282,0.1367159337,0.2987371087,-0.0851852447,0.2680241466,-0.1103682518,-0.0799849182,0.097489208,0.1993958354,0.6316300631,0.0561968759,0.1229816005,0.1048306823,-0.1335862875,0.1393507719,0.150758177,-0.0844926164,0.0812596604,0.4595391452,-0.1926772296,-0.1196798459,0.2415973842,0.063960284,-0.1346393675,-0.5399410725,0.1558781564,0.2517969608,-0.1703035831,-0.1522077769,0.1404367834,0.5916336179,0.0165722854,0.0147288786,-0.3151999116,-0.0298845507,0.0900036767,-0.0841346011,-0.0351762213,-0.145972237,-0.4134395719,0.17644912,-0.0717597604,0.0099335993,0.3621029854,0.2145014256,-0.0319225006,-0.4427134395,0.0857791379,0.2318394333,-0.02281278,-0.2336860448,0.1210372746,0.0231404118,0.0358833745,0.2976679802,-0.015728917,0.3811991811,0.4854708016,0.0620260872,-0.3394706249,-0.1351813525,-0.0548018776,-0.0553645715,0.3558157086,-0.1178854108,0.273103714,0.2957620323,0.1765792966,-0.0403585732,0.2760414481,0.0370879844,0.1097964868,-0.4307216406,-0.0895508006,0.0860620439,0.0078106895,-0.1641460508,-0.2182767391,-0.1917066872,-0.1558233052,0.4550687373,0.1204833239,0.2162842602,0.1460715532,0.0373648219,-0.1193289757,0.246837303,0.4382474124,0.1416926384,-0.574309051,-0.0275462605,-0.8004928827,0.0029473368,-0.2535035312,-0.1772418469,-0.0825795382,0.2530684769,0.2895002961,0.3318608701,0.1106532514,-0.0300684776,-0.1929352731,0.2943170965,-0.2472021133,-0.1607429981,-0.1506142467,-0.0059295562,-0.0098957289,-0.4133943021,0.0394508578,0.30848369,-0.0588939376,-0.1396917403,-0.1065005809,0.0850683376,-0.3280164599,0.4423456192,-0.0652806982,0.4719922543,-0.0089084692,0.2330144197,-0.4741684496,-0.3503286242,-0.2095749527,0.1834397465,-0.1208372787,0.4187626541,-0.3576852381,-0.0196065661,-0.2524614632,-0.0442239381,0.0910942033,0.1757109463,-0.2568310201,0.1950098723,-0.0321600586,0.1595012248,-0.076503858,-0.1335105747,-0.0936495587,0.3417082727,-0.1618132144,-0.3193213046,0.3301817477,-0.1463798583,-0.2415775359,0.0700480714,0.1511906385,0.0502488725,0.0939265117,-0.5559325814,0.0279939286,0.3319455683,-0.0682038665,-0.1687952131,0.1472918689,0.0379714519,0.0707561895,-0.1159297153,0.2154304683,0.2917870879,-0.1881298721,-0.2043753266,-0.2656930089]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi @lhoestq, Below is the entire error message after I move both tsv files to the same directory. It's the same with I got before.\r\n```\r\n\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/torch\/cuda\/__init__.py:52: UserWarning: CUDA initialization: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http:\/\/www.nvidia.com\/Download\/index.aspx (Triggered internally at \/pytorch\/c10\/cuda\/CUDAFunctions.cpp:100.)\r\n return torch._C._cuda_getDeviceCount() > 0\r\n08\/29\/2021 22:56:43 - WARNING - __main__ - Process rank: -1, device: cpu, n_gpu: 0distributed training: False, 16-bits training: False\r\n08\/29\/2021 22:56:43 - INFO - __main__ - Training\/evaluation parameters TrainingArguments(output_dir=\/projectnb\/media-framing\/pred_result\/label1\/, overwrite_output_dir=True, do_train=True, do_eval=False, do_predict=True, evaluation_strategy=EvaluationStrategy.NO, prediction_loss_only=False, per_device_train_batch_size=8, per_device_eval_batch_size=8, gradient_accumulation_steps=1, eval_accumulation_steps=None, learning_rate=5e-05, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=8.0, max_steps=-1, lr_scheduler_type=SchedulerType.LINEAR, warmup_steps=0, logging_dir=runs\/Aug29_22-56-43_scc1, logging_first_step=False, logging_steps=500, save_steps=3000, save_total_limit=None, no_cuda=False, seed=42, fp16=False, fp16_opt_level=O1, fp16_backend=auto, local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=500, dataloader_num_workers=0, past_index=-1, run_name=\/projectnb\/media-framing\/pred_result\/label1\/, disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=None, ignore_data_skip=False, sharded_ddp=False, deepspeed=None, label_smoothing_factor=0.0, adafactor=False, _n_gpu=0)\r\n08\/29\/2021 22:56:43 - INFO - __main__ - load a local file for train: \/project\/media-framing\/transformer4\/temp_train.tsv\r\n08\/29\/2021 22:56:43 - INFO - __main__ - load a local file for test: \/project\/media-framing\/transformer4\/temp_test.tsv\r\n08\/29\/2021 22:56:43 - WARNING - datasets.builder - Using custom data configuration default-df627c23ac0e98ec\r\nDownloading and preparing dataset csv\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/usr4\/cs542sp\/hey1\/.cache\/huggingface\/datasets\/csv\/default-df627c23ac0e98ec\/0.0.0\/9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff...\r\nTraceback (most recent call last):\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 693, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1166, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 428, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"run_glue.py\", line 487, in <module>\r\n main()\r\n File \"run_glue.py\", line 244, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 852, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 616, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 699, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file. \r\nOriginal error:\r\nerror closing file\r\n```","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":311,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi @lhoestq, Below is the entire error message after I move both tsv files to the same directory. It's the same with I got before.\r\n```\r\n\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/torch\/cuda\/__init__.py:52: UserWarning: CUDA initialization: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http:\/\/www.nvidia.com\/Download\/index.aspx (Triggered internally at \/pytorch\/c10\/cuda\/CUDAFunctions.cpp:100.)\r\n return torch._C._cuda_getDeviceCount() > 0\r\n08\/29\/2021 22:56:43 - WARNING - __main__ - Process rank: -1, device: cpu, n_gpu: 0distributed training: False, 16-bits training: False\r\n08\/29\/2021 22:56:43 - INFO - __main__ - Training\/evaluation parameters TrainingArguments(output_dir=\/projectnb\/media-framing\/pred_result\/label1\/, overwrite_output_dir=True, do_train=True, do_eval=False, do_predict=True, evaluation_strategy=EvaluationStrategy.NO, prediction_loss_only=False, per_device_train_batch_size=8, per_device_eval_batch_size=8, gradient_accumulation_steps=1, eval_accumulation_steps=None, learning_rate=5e-05, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=8.0, max_steps=-1, lr_scheduler_type=SchedulerType.LINEAR, warmup_steps=0, logging_dir=runs\/Aug29_22-56-43_scc1, logging_first_step=False, logging_steps=500, save_steps=3000, save_total_limit=None, no_cuda=False, seed=42, fp16=False, fp16_opt_level=O1, fp16_backend=auto, local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=500, dataloader_num_workers=0, past_index=-1, run_name=\/projectnb\/media-framing\/pred_result\/label1\/, disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=None, ignore_data_skip=False, sharded_ddp=False, deepspeed=None, label_smoothing_factor=0.0, adafactor=False, _n_gpu=0)\r\n08\/29\/2021 22:56:43 - INFO - __main__ - load a local file for train: \/project\/media-framing\/transformer4\/temp_train.tsv\r\n08\/29\/2021 22:56:43 - INFO - __main__ - load a local file for test: \/project\/media-framing\/transformer4\/temp_test.tsv\r\n08\/29\/2021 22:56:43 - WARNING - datasets.builder - Using custom data configuration default-df627c23ac0e98ec\r\nDownloading and preparing dataset csv\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/usr4\/cs542sp\/hey1\/.cache\/huggingface\/datasets\/csv\/default-df627c23ac0e98ec\/0.0.0\/9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff...\r\nTraceback (most recent call last):\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 693, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1166, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 428, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"run_glue.py\", line 487, in <module>\r\n main()\r\n File \"run_glue.py\", line 244, in main\r\n datasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 852, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 616, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 699, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file. \r\nOriginal error:\r\nerror closing file\r\n```","embeddings":[-0.0713824928,0.0363221169,-0.024686452,0.5199025869,0.3269370794,0.1928398013,0.44621858,0.3241172433,0.224331066,0.0428135805,-0.019542044,0.2665969729,-0.0372168943,0.0397648402,-0.2458034754,-0.1460843086,0.0505351909,0.0280245766,-0.3117779493,0.0664772168,-0.260904789,0.3204184771,-0.1558368057,-0.0361693315,-0.2530987263,0.119819276,0.0454959869,0.2727533579,0.0326381475,-0.3402159512,0.4851877093,-0.1499628276,0.533010304,0.4523147643,-0.0001191331,0.1187679619,0.2267004997,-0.1120318621,-0.2227353454,-0.6003681421,0.0691084787,-0.4721171856,0.3808331192,-0.1278465241,-0.169731915,0.068624936,-0.1462235749,-0.4836004674,0.3415564597,0.4694939256,0.1677398831,-0.2340410054,-0.0738900751,0.0044538765,0.4453818202,0.1310156137,-0.1876753718,0.2258012593,0.009724549,-0.1359496862,-0.0403806679,0.1108501628,-0.130706951,0.2922811806,0.2499799132,0.2974606156,-0.2419434488,-0.2736930251,0.0225309189,0.1213904694,0.6164861917,-0.1500841528,-0.1400368065,-0.0326985009,-0.0828981847,-0.1896711886,0.3366042972,0.3559082747,-0.1302392781,0.2394926399,-0.1136730015,0.1308840364,-0.2241668999,0.0715673938,0.049326621,-0.0001854788,-0.1302859634,0.3417437673,0.0915791616,0.0901505724,0.3912831545,-0.004260757,0.002390295,0.1033155695,-0.7571657896,0.3397616744,-0.0897243097,-0.17062895,-0.162389636,0.1441668272,0.3156573772,-0.0900196284,-0.0864366218,0.3288353682,0.0821671486,0.0484148823,-0.0442836992,0.2279284298,0.2494887561,-0.1326309741,0.0859453157,-0.188307777,-0.3141408861,-0.6196153164,-0.1500588059,0.2270557433,0.3016443551,-0.1968029141,-0.2878167033,0.0655969828,-0.2806815207,0.066669859,-0.1228434145,0.2648940682,0.0883032531,0.3031728566,0.1720811129,0.2169495225,-0.0404541604,-0.2523911595,-0.0274142977,0.1094366089,-0.3116878569,-0.0032872735,0.5225704908,-0.2512436807,0.1504862458,0.2206335515,-0.0773041472,-0.3231442273,0.0181409009,-0.3295076489,0.1302096099,0.2901154757,0.1380326301,0.3126874268,0.3299174607,-0.1744098216,-0.070775561,0.3949204981,-0.4245900214,-0.1792442054,-0.2038639933,0.0775322244,-0.0885883346,0.1009734198,-0.2871201932,0.0694979653,0.1445493251,-0.2312690765,-0.0790209547,-0.3212262392,-0.3405198455,-0.3129277527,0.1117309853,0.4026996195,-0.8818349242,0.2132673413,0.198320508,0.0556938276,0.0214947499,0.3038444221,-0.2842591107,-0.0352121368,-0.2930731475,-0.0345344506,0.0590712763,-0.288402468,-0.2493695915,0.1552080065,-0.0890525058,-0.0228530131,0.2429944128,0.1101588085,0.3556645215,-0.0237174034,-0.0928946137,0.1914544106,-0.0591227002,0.0131877027,-0.2130649835,-0.1901592016,0.3738617003,0.1716047227,-0.0081138918,0.0118381307,0.1692466736,-0.6463864446,0.3319563568,-0.1069131419,0.1779621542,0.0905336738,0.325034976,0.0173670687,0.2009603232,0.1837183982,-0.6111759543,0.1801491976,0.4992318749,-0.1896237284,-0.2479818463,-0.1043635309,-0.3395034671,-0.0565126985,-0.1028170288,-0.1103380173,-0.0110080214,0.1839656383,-0.1009614468,-0.0751741454,-0.040860422,0.4944821894,-0.4295209944,0.0704878941,-0.3189311028,0.3528833091,0.0992905796,-0.1022058576,0.141086325,-0.0712023675,0.27628842,-0.1578126103,-0.1376788318,0.2136023492,0.3686512113,-0.0230407659,-0.0386369973,-0.2215279043,0.1488551795,-0.1720895022,-0.0323602557,0.2876801789,0.358145684,-0.0749094114,-0.4160434902,0.3269958198,-0.4082435071,0.1859025806,-0.1869416833,0.0978934839,0.3414917588,0.0086155161,0.0448844656,0.0606090277,-0.0055956342,0.1934436113,0.1597741842,0.1372384876,-0.2749932408,-0.0690765604,0.1137360707,0.2358391583,-0.2924256027,0.1657490879,-0.1018801779,-0.0168215632,0.439864397,0.2854737341,0.4559783638,0.0421452485,-0.2504726052,-0.0080843512,0.0931468531,-0.2148771435,0.0924219713,-0.0473628677,0.0956739485,0.2528899312,0.196509093,0.1560171694,-0.0759197176,-0.1987670809,0.2110615969,0.2683174312,-0.6494820118,0.0274201259,-0.2144494951,0.2339817286,-0.2603882849,-0.104509525,0.103598237,-0.1728931814,-0.3434310555,0.344153285,-0.0421172939,0.1579322219,-0.0993181616,0.1854149103,0.0730421916,-0.1689814329,0.0334094986,-0.2104582936,-0.2221073508,-0.008325018,0.1756947488,0.0255544335,-0.0517407879,0.0312438663,0.075098142,-0.2419181615,-0.0653229058,0.0522311255,-0.0377469473,0.1385564357,0.035073515,-0.0259483196,-0.0233432185,-0.3175126016,0.1683665216,0.072531648,-0.0024671736,0.2256510705,0.2033078521,-0.2221161574,-0.1577876061,-0.6343611479,-0.3415049911,-0.2504074872,-0.0175146479,-0.0438977964,-0.034607254,-0.0959560424,0.3808245063,-0.0116766794,0.2445736825,-0.0465579517,-0.0434968397,-0.2414994985,0.5401979089,-0.1761843115,-0.3055940568,0.2881653011,0.0057944898,0.0812152103,0.401751399,-0.4659741223,0.1516272277,0.0125736389,0.1310024559,-0.0981328413,0.00196495,0.1370166242,0.0259656571,0.085618265,0.0137890568,-0.2664068043,0.1253763139,-0.0736640915,0.1062660664,0.1298653036,0.6638559699,-0.1733186543,0.6843754649,0.131282866,-0.1560296118,0.2848717868,-0.1524146497,0.3640497029,0.166402027,-0.3182913363,0.1269710064,-0.2186773717,-0.3227928579,0.204973653,-0.1381778121,-0.0686757043,-0.1539912224,0.1816746891,-0.3221245706,-0.1505452693,0.210945338,-0.1003916115,0.0678139552,-0.0305393338,0.1973719299,0.0237294864,-0.1864508688,-0.0467939787,0.2185323983,0.1680103391,-0.0804374218,0.1470590681,0.031991031,-0.2279854715,0.2945488691,0.0224822219,0.4111844599,0.0595987551,-0.0883613229,0.0089130001,0.1625760794,0.7426168919,-0.1256064326,-0.0731673017,0.0853601843,-0.2333904803,-0.1674278677,-0.2884995043,-0.0899905413,0.1868292242,-0.0027646853,0.198735103,0.0296388809,-0.0472526811,0.3281803727,0.1264507771,-0.187419191,-0.2498578727,-0.1188151762,-0.00936992,-0.2244079709,-0.2099975944,0.1062965989,0.0763743892,-0.3267103136,-0.0191834662,-0.271435529,-0.2140168399,0.2801027596,0.1397762597,0.3112785816,0.0763530731,0.2653453052,0.1252960861,0.2106350511,0.2983998656,0.6239320636,-0.1243404746,-0.7515357733,0.1691110581,-0.458096236,0.2946280539,0.0273762867,-0.3869454265,-0.0640055537,-0.3385323882,-0.109625034,-0.0703078285,-0.0726530403,0.1365453303,0.0585652813,-0.184651345,-0.4386244714,0.4338480532,-0.0518908612,-0.1056194976,0.2825895846,-0.0692652687,-0.3331142068,0.3501770198,-0.0277878996,0.6251299381,-0.0991299003,0.0102299983,0.1925514042,-0.2428119481,0.4717301726,-0.2204425782,0.0258232951,-0.4598237872,-0.2451589853,-0.1922226101,0.0322490111,0.0966751277,0.4480052292,-0.201062277,0.2890181839,-0.1025590301,0.3831981122,0.0078430483,0.0220265165,0.380310148,-0.1665107757,-0.1624476612,0.0953690484,-0.1526246518,-0.0245816857,0.0563908182,0.0197181422,-0.2976280451,0.0125805382,-0.2996997535,0.2973289788,-0.20800969,0.2216138542,0.0806187987,-0.7200571895,0.4817132354,0.3934561312,0.2646308839,-0.1215755418,-0.1750627458,0.0561405495,0.0133059574,0.011479076,-0.149382174,-0.0852063224,0.0047015557,0.1939673871,-0.1968918592,0.3111004233,-0.0344010554,-0.0646474436,0.3603192866,0.2636821568,0.0922409445,-0.4585042,-0.462254405,-0.1424634457,0.1399500221,-0.0867381096,0.0505421795,-0.091066055,0.0593239516,-0.0913203061,0.0969792679,-0.4294605255,0.0909238756,0.2542864382,-0.1923609227,0.0302958768,0.4609158635,0.4661313891,-0.1570168287,-0.1127240956,-0.0519315191,0.124244675,-0.6640465856,0.1180858314,0.0874280035,0.2713101506,0.1106669605,0.2245652974,0.2448486537,-0.1299523115,0.1159513667,-0.5455234051,-0.0697467551,-0.0773646608,-0.2316938192,0.0858305022,-0.0465646163,-0.3287047148,-0.0299688112,-0.0460659564,-0.1822274327,0.1479650736,0.0829787776,0.0912668407,0.2328791171,0.3219473362,-0.2838843763,-0.2637484968,0.029215537,0.0507224612,0.1216696724,-0.1127028167,-0.1922750324,0.1594114751,0.0283143949,0.1106836721,-0.2598891258,-0.2662132978,-0.2686952055,-0.1795853525,0.1353785992,0.3537660539,-0.0081257774,0.4036307633,-0.0431269258,-0.1033697128,-0.5479774475,0.2926706672,-0.2415863872,0.2237766087,0.199037984,0.1964119077,0.1309591085,0.203501597,-0.1996586472,0.0008341488,-0.1360931545,-0.0493976809,0.1639295816,-0.3254681826,0.1608218998,0.1159815714,0.2796325982,0.3563481867,-0.1452045441,-0.1158741638,0.114940986,0.11075259,-0.3894898295,0.249901101,0.0710891187,0.2626110613,-0.1218530312,0.3188999891,0.080301404,-0.1559212059,0.1330567747,0.2122031599,0.5601608157,-0.1304023862,0.1679929942,0.1382879913,0.0153289782,0.1011405438,0.131508261,-0.0248818789,0.0874198973,0.4638621211,-0.1963236183,0.0259222258,0.3378002346,0.2243233174,-0.1777549684,-0.5586881638,0.1542459577,0.1441560686,-0.3409478366,-0.243239224,0.2154238522,0.4956796467,0.0113115087,0.007662551,-0.3150693476,-0.0030331227,0.0038702802,-0.045924712,-0.1675787419,-0.189921245,-0.335691303,0.3205699921,-0.1349436939,0.020549573,0.4814255834,0.2324887514,-0.2511176765,-0.5126866102,0.1508887261,0.2508907914,0.0274047069,-0.1713472605,0.0852352306,0.0495510548,-0.0540526062,0.1698529124,-0.0192384757,0.4859765768,0.4486764073,0.0273987968,-0.3042316735,-0.0401586555,0.0159792118,-0.0440404937,0.4091783166,0.0131132836,0.3308348358,0.1833188832,0.1357971281,-0.0060800305,0.2983112037,0.1140477657,0.0704922602,-0.4106969237,-0.0540727749,-0.0409771092,-0.0714359581,-0.1586268991,-0.1491562575,-0.0871632174,-0.1485601962,0.519643724,0.1215119064,0.188619718,0.1503846645,0.0073909406,-0.1367951781,0.2055473328,0.3561823964,0.0819178,-0.5459756851,-0.0732731968,-0.8299985528,0.0774980932,-0.1281850487,-0.0972929373,-0.0878585503,0.2418970615,0.2935998142,0.2321042866,-0.0436902009,0.1798369288,0.0490001217,0.2618585825,-0.2553984821,-0.1562901288,-0.3128726184,-0.0214397404,-0.0712897032,-0.3751957715,0.0317862332,0.1568144858,-0.0899392217,-0.1936057061,-0.0788183212,0.074318476,-0.1119757593,0.4728642702,-0.0571855977,0.4539398849,0.0706689507,0.3540861905,-0.4080405235,-0.2868048251,-0.1724875718,0.409463048,0.0054507949,0.377499789,-0.3005765378,-0.0191425607,-0.2216658592,0.1338578314,0.0602834746,0.1935337633,-0.2525572181,0.2503116131,-0.1184998825,0.1764087975,0.0193541758,-0.0358276479,-0.1018786281,0.2717542052,-0.164885819,-0.2978326976,0.2681587636,-0.2065344453,-0.2151190341,-0.0239162818,0.0972871482,0.1572106928,0.1748520732,-0.5737085342,-0.0117292609,0.3259617984,-0.182708919,-0.1304383874,0.1300726533,0.000290238,0.05965719,-0.1849017888,0.2264322042,0.2564281523,-0.1877179444,-0.2046785355,-0.3282538652]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi !\r\nCan you try running this into a python shell directly ?\r\n\r\n```python\r\nimport os\r\nfrom datasets import load_dataset\r\n\r\ndata_files = {\"train\": \"train.tsv\", \"test\": \"test.tsv\"}\r\nassert all(os.path.isfile(data_file) for data_file in data_files.values()), \"Couln't find files\"\r\n\r\ndatasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\nprint(\"success !\")\r\n```\r\n\r\nThis way all the code from `run_glue.py` doesn't interfere with our tests :)","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":56,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi !\r\nCan you try running this into a python shell directly ?\r\n\r\n```python\r\nimport os\r\nfrom datasets import load_dataset\r\n\r\ndata_files = {\"train\": \"train.tsv\", \"test\": \"test.tsv\"}\r\nassert all(os.path.isfile(data_file) for data_file in data_files.values()), \"Couln't find files\"\r\n\r\ndatasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\nprint(\"success !\")\r\n```\r\n\r\nThis way all the code from `run_glue.py` doesn't interfere with our tests :)","embeddings":[-0.0062861182,0.0618332289,0.0004360501,0.3439050019,0.2524879575,0.1751339436,0.5055190325,0.32783705,0.3839304149,0.0075207599,-0.094731316,0.2648680508,-0.0998279154,-0.0093660085,-0.2245774418,-0.039557457,0.0509862155,0.0297241844,-0.3710769117,0.0725511983,-0.3521109223,0.2782054842,-0.3021057844,-0.0440652966,-0.1426056921,0.1825702786,-0.1116087809,0.3334528208,-0.099801138,-0.4000122845,0.4402317405,-0.0391456969,0.4721477032,0.3309651613,-0.0001206773,0.0893450528,0.3573922217,-0.0702851191,-0.2225422859,-0.5089859962,-0.0338167772,-0.4686772525,0.296729207,-0.1405784786,-0.1099878922,0.1247854009,-0.1471688002,-0.4223869145,0.3362388313,0.4916538596,0.1666339934,-0.0498269834,-0.0776987374,-0.0487935245,0.6013208628,0.0008950081,-0.107385233,0.3473229408,0.0976693556,-0.2549462914,-0.0749159977,0.0784208179,-0.0796423927,0.2506426871,0.1551815718,0.2411040068,-0.1426447928,-0.1929208934,0.0080289897,0.1800806522,0.5383203626,-0.1723931134,-0.2085453123,-0.0655025765,-0.0718722492,-0.0408947691,0.3501514196,0.3825588822,0.029311765,0.3788618147,0.013485969,0.0541821085,-0.1889536381,0.126449421,0.0808381736,-0.000396442,-0.1469158232,0.2984919548,0.1597528756,0.0262267981,0.1689083278,-0.1150744557,-0.0622056946,0.1477849483,-0.7149015069,0.3251987696,-0.1247542053,-0.0206366517,0.0187079646,0.2130948007,0.2478446364,-0.042283155,-0.0525682718,0.3610035777,0.2418078333,0.0896460861,-0.0427161828,0.2770175934,0.2427061945,-0.1332805008,-0.0287302025,-0.1192627922,-0.3003031015,-0.5950941443,-0.0516846143,0.1479632556,0.307525605,-0.1655176282,-0.2107231319,0.0764237642,-0.3221824467,0.0662977695,-0.1481997073,0.229479,0.1425035149,0.2334895581,0.0762687698,0.3107706904,-0.1320535988,-0.3083095551,-0.0419163592,0.2396093458,-0.379576385,-0.0501148328,0.4993746877,-0.2040912956,0.1153782755,0.1919148713,-0.1458258927,-0.3528422415,0.0152910464,-0.2772711217,0.0987497717,0.3070719838,0.0310322456,0.1829850376,0.3100918233,-0.3143515885,-0.0652119443,0.2924493551,-0.3839842379,-0.1958357692,-0.073478803,0.0947331712,-0.3603913784,0.0796126425,-0.3193351626,0.025692692,-0.008764972,-0.1769020706,-0.1282196492,-0.356578052,-0.2364251912,-0.2965958416,0.1110925674,0.4650505185,-0.7945942879,0.238329649,-0.0096505545,0.0674593449,0.0469612703,0.2029931396,-0.175192818,0.0009627892,-0.3269457817,0.1846571863,0.1312337667,-0.3535791039,-0.3629861176,0.2135502994,-0.1101615503,-0.0140276952,0.1743032336,0.0087105827,0.2934425771,-0.0421954803,-0.175726518,0.1522732228,0.008166681,0.0231556855,-0.3212611377,-0.2146085203,0.2824979722,0.1791924387,0.0050034234,0.0181187242,0.2591667473,-0.7052285075,0.3372290432,-0.0257217083,0.2324234694,0.0988066494,0.372282356,0.0113708097,0.301420778,-0.0037520607,-0.5378309488,0.2384977788,0.4775056541,-0.1331087351,-0.1889230758,-0.0681338608,-0.4203745723,0.0107433051,-0.2168693542,-0.1097213328,0.0102444394,0.0357291028,-0.0001164881,-0.1082693115,-0.0550721921,0.4337472022,-0.2398032397,0.1292284727,-0.3495222628,0.3247340322,0.1120966896,-0.059616562,0.1662975848,0.037718799,0.2580135763,-0.1628742814,-0.1263890266,0.3396655619,0.3961007595,-0.0519712567,0.160907954,-0.1690751165,0.0690016448,-0.2459465712,-0.0559245907,0.2450754493,0.3397670686,-0.0777557641,-0.4732547104,0.2924112678,-0.3288089037,0.3145468235,-0.1321841329,0.1685900688,0.2230485827,-0.0696802065,-0.1975671649,0.0641156211,0.0427139364,0.0154422,0.3271248341,0.1262814254,-0.1261104643,-0.1338541657,0.1003465354,0.1134780571,-0.1979324669,0.1449226886,-0.1365266144,-0.0266337264,0.3255553544,0.4348104894,0.4938296676,0.1152940765,-0.1612459123,-0.0280058421,0.0757928044,-0.1865972131,0.1406206936,0.0464963689,0.0275778249,0.1959311217,0.2476542741,0.0444588624,-0.0910918713,-0.1231984198,0.1277632713,0.185345605,-0.6318135262,-0.0364054255,-0.2342247665,0.3816649616,-0.2050330043,-0.0471458361,0.1352360547,-0.2819332182,-0.3157407641,0.203558296,-0.1404105872,0.0973091722,-0.2060460597,0.2179884166,0.1297790706,-0.2205867618,0.0024759686,-0.171497345,-0.2777157128,-0.0176119246,0.0986566097,0.0389547199,0.0045793718,0.1216914579,0.0442261361,-0.2933131158,-0.1359817088,0.1183634251,-0.0273984317,0.1899541467,0.2141971886,-0.1362731904,-0.0025947772,-0.1741008759,0.2184861451,0.0330374613,0.0418562219,0.4006383121,0.2771764994,-0.1021996513,-0.1422503591,-0.6783524752,-0.3884107769,-0.2795219123,0.1429558396,-0.1724932641,-0.1111443713,0.1216984391,0.3000505567,0.0373916067,0.1272048652,-0.1077364758,-0.117249079,-0.2421749532,0.5630997419,-0.2089006603,-0.3398690522,0.1851096898,-0.1722486913,0.0041179014,0.3509646058,-0.3908100128,0.018076485,-0.0999999195,0.0693459138,-0.0278682131,0.1542191803,0.0238924045,-0.047691077,0.1044527367,-0.0720873252,-0.1617171317,0.2585965693,-0.0316148922,-0.0129348328,0.2223136127,0.4758070409,-0.1225692928,0.5804255009,0.0619642809,0.050454434,0.2805772424,-0.1669808328,0.3939423263,0.1707340628,-0.2829434574,0.0277966596,-0.1335334331,-0.3184961975,0.2585748136,0.0010107184,-0.0635258257,-0.0499602221,0.3057394028,-0.288435638,-0.2198022902,0.1359740198,-0.0676380694,0.0328279473,0.0285982322,0.0905077457,0.0752010494,-0.2221941799,0.0259455927,0.1324168593,0.2070467174,-0.0871777982,0.1151671857,-0.1608940959,-0.2878498435,0.2011395693,-0.0998934135,0.4879913628,0.092275925,-0.1246773228,0.0289812628,0.1562558413,0.7973386049,-0.3307304382,0.0607476979,0.2085767239,-0.0579907,-0.2705810666,-0.3227037191,-0.197427839,0.1449974328,0.0405170247,0.3329616189,-0.0296818055,0.0314081013,0.4015727341,0.1198751479,-0.0922064483,-0.3959646821,-0.1509085447,-0.0523686558,-0.1299741864,-0.201807484,0.1642456204,0.063848868,-0.2626346946,0.0484170392,-0.2081122249,-0.2489888072,0.3064344823,0.0998005122,0.3315268457,0.001684419,0.2351656854,0.2444294393,0.1030283347,0.1286245137,0.735691309,-0.2316629589,-0.7495519519,0.1535028219,-0.4999154806,0.4155621827,-0.022335697,-0.3868151009,-0.0933660865,-0.32850945,-0.0750938654,-0.1487456411,0.0221094079,0.1300385445,0.0235475488,-0.1224919409,-0.3920351267,0.5094703436,-0.0529166088,-0.1391727328,0.148479566,0.0504317246,-0.3435452878,0.5170083046,-0.2264779806,0.6625568867,-0.1139025539,-0.0429214239,0.2207420617,-0.2347032279,0.4015940428,-0.2745751441,-0.0725832731,-0.4662559927,-0.2859759629,-0.1123683751,-0.0579464175,0.0895486251,0.5569059253,-0.2300378382,0.3319657147,-0.1128526926,0.3904385567,0.065334782,0.1853723824,0.3027922511,-0.1388522536,-0.1399407685,0.1091327742,-0.0370872207,-0.045306094,0.0928043723,0.0325299725,-0.2246846557,-0.0832145512,-0.3111889362,0.2645092309,-0.221382156,0.1078122705,0.1895973831,-0.457907021,0.47436288,0.3568017781,0.1146369055,-0.0029017217,-0.1852585524,0.1015009359,0.1217641458,0.0406750143,0.0068598208,-0.0186508074,-0.0073122266,0.2269879729,-0.2133965343,0.2065441012,-0.1337005943,-0.0998805463,0.2832755148,0.2662884295,0.1629767269,-0.463259995,-0.4328168631,-0.2219876647,0.0756355077,-0.0825303495,0.0529006496,-0.028985424,-0.0261390079,0.0518195406,0.0938666612,-0.4068253934,0.1623443365,0.2168688029,-0.3158495128,0.1522035897,0.4322910011,0.475780338,-0.0787295923,-0.1661250442,0.0207740758,0.1792140752,-0.533282578,0.168761149,0.0779091567,0.1887216717,0.0965940803,0.1122952551,0.2716129422,-0.2194173038,0.056315694,-0.5997289419,-0.0178624932,-0.0290189683,-0.0808221474,0.1214091033,-0.0437527634,-0.4180786908,-0.0769166499,-0.0492746793,-0.1964782029,0.1414736956,0.0534505807,0.038065277,0.3048774004,0.2614005804,-0.1836981326,-0.2774608135,0.0480910502,0.0333735198,0.0787975639,-0.1318198889,-0.1984200329,0.1645409316,-0.0724463761,0.2178934664,-0.1894136369,-0.1972860694,-0.1989601254,-0.183006227,0.1245347559,0.2621741891,-0.0282929502,0.3648383617,-0.0076064598,-0.1850651503,-0.4463832378,0.3087603748,-0.2491591722,0.1177139804,0.1212879121,0.21079202,0.1490400881,0.1265259087,-0.0876453146,0.0389842354,-0.1910273284,-0.0634854436,0.2676966488,-0.3426944613,0.1156221852,0.1754963547,0.4235918224,0.3090229332,-0.1356925219,-0.0606913827,0.0762435496,0.1580803692,-0.3811096847,0.206409052,0.1673161387,0.2456314713,0.0201372281,0.283411175,-0.0382842161,-0.1048923582,0.0237701628,0.1907588542,0.6733418107,0.0377098359,0.0342129655,0.2278372347,-0.0785998553,0.0914947838,0.1461608708,-0.168244496,0.0773076937,0.3710555136,-0.0827592984,0.003663878,0.3795471489,0.1875592023,-0.1309307218,-0.5275370479,0.2405227274,0.3179276288,-0.2321213782,-0.1387685984,0.1503387094,0.8276310563,-0.0742329359,0.0468161143,-0.25398314,-0.1251809448,0.1233496517,-0.0945746377,-0.1050643921,-0.1339813024,-0.4749995768,0.220361203,-0.1088465154,-0.0279281884,0.4934461117,0.2190087438,0.007231439,-0.4880752861,0.0196796544,0.2167332619,0.0583826341,-0.3225283921,0.1015523598,-0.0613080822,0.0048887702,0.2801894546,-0.1002024263,0.4725545049,0.383972764,-0.0325991996,-0.3448341787,-0.1978309751,-0.0683287159,-0.0755073056,0.4705097377,-0.0550898761,0.2542237341,0.2449971139,0.1167153865,-0.0182929281,0.2133943886,0.0777039602,0.0493583716,-0.4734903872,0.0162538234,0.0388623923,-0.124118425,-0.1102314219,-0.2345827818,-0.2053638399,-0.1926809698,0.5107021928,0.1163474992,0.2256508023,0.1695706099,0.0143696619,-0.1597373933,0.1876871139,0.4251697361,0.0898337439,-0.5339030623,-0.0086079799,-0.8671326041,-0.0264169183,-0.1721196622,-0.2194489986,-0.003831469,0.2222914398,0.3162292242,0.2648622096,-0.001159697,0.211091131,-0.0564790741,0.2417295128,-0.2078548372,-0.1545615345,-0.2132895142,0.0549823344,-0.019938875,-0.4609722197,0.0379728712,0.2296255678,-0.1171760336,-0.0772046521,-0.1085590273,0.1284982562,-0.2856185436,0.4325979948,-0.0586674996,0.4626667798,0.0414262339,0.3189246953,-0.4940350354,-0.2361551672,-0.1467922926,0.2007579505,-0.2002476007,0.3741764724,-0.2807653546,-0.0034886301,-0.20289056,-0.0559482165,0.0842913911,0.1892364472,-0.3115457296,0.2394579202,-0.1083200648,0.2191468179,-0.084726572,-0.1101704538,-0.2025339901,0.2691792548,-0.0543310232,-0.3667738736,0.3107223213,-0.3082571328,-0.2810049355,0.0863787383,0.2669145465,0.1490323842,0.1282244623,-0.4939648807,-0.0689735487,0.3465808928,-0.1181809008,-0.169647485,0.1538556814,0.0902242437,0.1440924555,-0.1082378775,0.139920488,0.2795184255,-0.2076472938,-0.1717350632,-0.3059180379]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi @lhoestq, \r\n\r\nBelow is what I got from terminal after I copied and run your code. I think the files themselves are good since there is no assertion error. \r\n\r\n```\r\nUsing custom data configuration default-df627c23ac0e98ec\r\nDownloading and preparing dataset csv\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/usr4\/cs542sp\/hey1\/.cache\/huggingface\/datasets\/csv\/default-df627c23ac0e98ec\/0.0.0\/9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff...\r\nTraceback (most recent call last):\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 693, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1166, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 428, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 7, in <module>\r\n datasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 852, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 616, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 699, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file. \r\nOriginal error:\r\nerror closing file\r\n```","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":160,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi @lhoestq, \r\n\r\nBelow is what I got from terminal after I copied and run your code. I think the files themselves are good since there is no assertion error. \r\n\r\n```\r\nUsing custom data configuration default-df627c23ac0e98ec\r\nDownloading and preparing dataset csv\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/usr4\/cs542sp\/hey1\/.cache\/huggingface\/datasets\/csv\/default-df627c23ac0e98ec\/0.0.0\/9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff...\r\nTraceback (most recent call last):\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 693, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1166, in _prepare_split\r\n num_examples, num_bytes = writer.finalize()\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/arrow_writer.py\", line 428, in finalize\r\n self.stream.close()\r\n File \"pyarrow\/io.pxi\", line 132, in pyarrow.lib.NativeFile.close\r\n File \"pyarrow\/error.pxi\", line 99, in pyarrow.lib.check_status\r\nOSError: error closing file\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"test.py\", line 7, in <module>\r\n datasets = load_dataset(\"csv\", data_files=data_files, delimiter=\"\\t\")\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 852, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 616, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/projectnb2\/media-framing\/env-trans4\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 699, in _download_and_prepare\r\n + str(e)\r\nOSError: Cannot find data file. \r\nOriginal error:\r\nerror closing file\r\n```","embeddings":[-0.0766204745,0.0920656696,-0.0017136495,0.4262596369,0.260334909,0.1757656038,0.385474503,0.2841121852,0.2640709579,0.0310262423,-0.1189905927,0.2788829803,-0.0557524003,0.0436067656,-0.3021000326,-0.0312552042,0.0004868807,0.0566506386,-0.3885477185,0.0410138145,-0.3662183285,0.3667454123,-0.3187191486,-0.0609998479,-0.1914399117,0.2263685465,-0.0368314795,0.3382658362,-0.1301845312,-0.3756915331,0.5021184683,-0.0521184951,0.5287908316,0.3145392835,-0.0001166436,0.0442261696,0.3172171116,-0.080038406,-0.2472346574,-0.4812021852,-0.0732618049,-0.4360754192,0.2785712481,-0.0932627395,-0.0746381879,0.0831260681,-0.1838671267,-0.4044435322,0.3599595428,0.4905599058,0.2041939795,-0.2266175747,-0.0533966236,-0.0375091098,0.4672858119,0.1351828426,-0.1646743715,0.2667809427,0.033427421,-0.1499333084,-0.0587946437,0.0766809732,-0.0680063516,0.2834093571,0.2143716961,0.2197063267,-0.1669077724,-0.1799285561,0.1194262952,0.1464961022,0.7298074365,-0.1525365412,-0.1637503803,-0.0346819237,-0.0223762766,-0.0578420088,0.3883640468,0.3147871494,0.0142923724,0.3722792864,-0.0827065781,0.0316018648,-0.263974905,0.0726018175,0.0781468526,-0.0138204424,-0.2187384963,0.3078764081,0.0906548798,0.0069222543,0.297133714,-0.1402279586,-0.0977815613,0.1826354414,-0.6836587191,0.2746154666,-0.1902481765,-0.008679051,-0.0350687504,0.1952123642,0.2596012354,-0.043297451,-0.1149725541,0.2992675304,0.1880345345,0.0414905027,0.0026777333,0.2720786631,0.2288470268,-0.1444371492,0.0412311144,-0.1318569779,-0.3197771311,-0.5077925324,-0.0904954299,0.1599707305,0.2539120913,-0.243753776,-0.227644816,0.1201614067,-0.2313852608,-0.0067400448,-0.1025092751,0.2435493618,0.1126939952,0.2905176282,0.0676930919,0.2179534733,-0.0906946287,-0.4313392341,-0.0451537371,0.1183533147,-0.3720602393,-0.046401687,0.5200387836,-0.2410795689,0.1423213035,0.2660549879,-0.1297727227,-0.3421305418,-0.0322260521,-0.3169589639,0.0239527002,0.2362485677,0.0389745571,0.2405245155,0.2866297066,-0.1781049669,-0.0325141177,0.3158904016,-0.443251431,-0.1592340767,-0.1350477934,0.1337426007,-0.27512604,0.05435827,-0.4043934047,0.1153675318,0.0411192514,-0.1912349463,-0.1712235659,-0.281470269,-0.3073675036,-0.2589139342,0.0480230786,0.4904466867,-0.8519433141,0.2261208892,0.0183286071,0.0025323441,0.0836948082,0.1709737182,-0.2485905588,0.0157358274,-0.2890161574,0.14831613,0.0495959707,-0.1857823282,-0.3911417127,0.2297905535,-0.0328218825,-0.0420922972,0.1136223748,0.0496373698,0.369251281,-0.071498096,-0.2215430439,0.176014185,0.0029329371,0.0403850414,-0.2676356137,-0.1843561232,0.1896206588,0.1752882749,0.007474679,0.1149878353,0.2790457308,-0.60149014,0.3060942888,-0.0593442731,0.2580801249,0.1458770484,0.2768695652,0.0048321439,0.3109204173,0.068410866,-0.5979942679,0.2334529459,0.4385319352,-0.2033963799,-0.1050681099,-0.0430174991,-0.4723057747,-0.0249618161,-0.2017359883,-0.0651664287,0.025429761,0.1114392206,-0.0386273377,-0.1682164371,-0.1144261211,0.4684047401,-0.3507110775,0.0618633665,-0.3040837348,0.3389953077,0.1605005413,-0.0358106717,0.2233442068,-0.0775785223,0.3259297907,-0.1029166803,-0.1839357167,0.2597768605,0.3249568343,0.0075665405,0.1412291378,-0.2065003663,0.0305013545,-0.2821558714,-0.0077941343,0.200310424,0.3678487539,-0.1124740839,-0.439716965,0.2634721398,-0.308039844,0.2148465365,-0.1841071844,0.0894398615,0.2627619505,-0.0551595129,-0.0341272652,-0.0108826719,0.0296714623,0.0668125376,0.2504251003,0.166002214,-0.1503114998,-0.1154987141,0.131458804,0.1737294793,-0.2715708017,0.25199458,-0.0939705297,-0.0774245486,0.3294499218,0.3344382048,0.4893704653,0.1043431237,-0.2505175471,0.0342538357,0.1167054474,-0.1767628938,0.0952144936,-0.0147623774,0.0734598115,0.2274675816,0.1784396619,0.1246469542,-0.1151839346,-0.1555411816,0.2122543454,0.2461774647,-0.5974342823,-0.0291215237,-0.2400234342,0.3107037246,-0.2130826265,-0.0169079825,0.1156351641,-0.2395446748,-0.3577504754,0.2877122462,-0.075320594,0.0539753176,-0.177641809,0.2452251613,0.0701454207,-0.1850723028,-0.0241964459,-0.1246553212,-0.2909143567,0.0290511791,0.1494376808,0.019947404,-0.0702546462,0.0626022816,0.0362962075,-0.3233754933,-0.0476093441,0.1449241936,0.0297123641,0.1076415479,0.1203675121,-0.1193226203,-0.046053648,-0.1017050594,0.1734133959,0.0696248338,0.0196209364,0.346560806,0.2277445346,-0.170257777,-0.1075497717,-0.6430516243,-0.2873028219,-0.2208728194,0.0694131702,-0.1359775513,-0.0641472712,0.0360071994,0.2975532413,-0.0357256941,0.2472857982,-0.1144540161,-0.1707125902,-0.2853313982,0.5567880273,-0.2298516631,-0.2987788022,0.2550551593,-0.0699987933,0.0933172256,0.3623686731,-0.4278604984,0.1757393926,-0.0856653899,0.0414425917,-0.0948990434,0.0472871326,0.0213367194,-0.0696596652,0.0943147987,-0.0596975833,-0.2065607756,0.2107337117,-0.0503792353,0.0644159466,0.1991274804,0.482239157,-0.1926071197,0.5956169367,0.1381940395,0.0244416744,0.3114423454,-0.07276728,0.3431052566,0.1681955904,-0.2420172542,0.1264448911,-0.0926905498,-0.2797218859,0.3001132905,-0.0175865646,-0.0444650128,-0.0809145868,0.2538172901,-0.262996465,-0.1329061538,0.1441645473,-0.0021719635,0.0858086944,0.0104690222,0.1078696921,0.0803341568,-0.21458821,-0.0315031223,0.1162984818,0.1495630741,-0.0665685162,0.1472606957,-0.0682260692,-0.2853095531,0.2672547698,-0.0902932957,0.3842556477,0.0616445057,0.0128766615,0.0238410253,0.2124439031,0.6938650608,-0.2765268385,-0.0001417281,0.1259997934,-0.0777212083,-0.1775041521,-0.2861438096,-0.0797559842,0.2022363096,0.0240980498,0.3000525832,-0.0484632365,0.0295048431,0.3985030353,0.1114721224,-0.1546878368,-0.4441296756,-0.0788959339,-0.08399304,-0.1326429844,-0.2752335966,0.1259173304,0.0835051239,-0.2511596084,0.0488567464,-0.2311005443,-0.2213755399,0.2830351591,0.136982739,0.2412112802,0.0732309222,0.2369686663,0.2524127662,0.1075080782,0.2686490417,0.7835130692,-0.2061071694,-0.8399780393,0.2030007839,-0.5181749463,0.2781227529,-0.0280248169,-0.3763557076,-0.0876169354,-0.2681033909,-0.1287165582,-0.1402122527,-0.0398665927,0.0613795407,0.0089078723,-0.1584059447,-0.4933409989,0.5363292694,-0.1677071452,-0.1483057737,0.1495650709,-0.0071604,-0.3145232499,0.3444467783,-0.1033034101,0.5854893923,-0.1115463153,-0.1564457268,0.161339581,-0.2082685828,0.3750273585,-0.227734983,-0.0961464942,-0.4784932137,-0.1491326988,-0.1390066594,0.0153579367,0.1141175255,0.5559030175,-0.2015703022,0.3405825496,-0.0971138477,0.3816972971,0.1246869266,0.1603046656,0.3174616992,-0.1933637708,-0.1356250793,0.1509141326,-0.1649430096,-0.0352075472,0.0691534877,0.0267955158,-0.2295257598,-0.0699788556,-0.3319748938,0.3233262002,-0.1663999259,0.0867159739,0.1780104488,-0.4880495369,0.5669565797,0.4021629393,0.2192201167,-0.0531578474,-0.1779717356,0.0928103626,0.017558191,0.0870793983,-0.1146431342,-0.0385068282,-0.0487275422,0.2040377706,-0.2052993476,0.2556633353,-0.0796705782,-0.0600396395,0.2950843871,0.2164248377,0.2323938161,-0.4881139398,-0.3482004404,-0.2358520776,0.0916323736,-0.0862164944,0.0711607784,-0.0459510237,0.0737519711,0.0044297692,0.090941757,-0.3700616956,0.1304409802,0.2394040674,-0.2664078474,0.0189806204,0.4548167586,0.4276998639,-0.1150632873,-0.1518424749,0.0279893205,0.1362561136,-0.5780556798,0.096986942,0.0927347317,0.2737703919,0.1611937135,0.182644248,0.2050187886,-0.1695383787,0.1299560368,-0.5699573755,-0.0315435231,-0.1393492669,-0.1653977185,0.0614223927,-0.0202136952,-0.3926450312,-0.0837385058,-0.0696602985,-0.2136393487,0.133776322,0.0512168296,0.0132510113,0.2452448756,0.2931753099,-0.2296188921,-0.2767742276,0.0442633778,0.1065781042,0.0634158626,-0.1454249322,-0.2136187851,0.1501998603,-0.0102408994,0.1067873314,-0.2959846854,-0.1934594065,-0.2556288242,-0.1705665886,0.1196297184,0.331541121,-0.0172575228,0.3716807365,-0.0607876815,-0.1611565053,-0.5020973682,0.2788507044,-0.1882903427,0.2069951743,0.0959108472,0.2364961952,0.1481495202,0.0996187553,-0.1412270069,0.0340405405,-0.1563390046,-0.1072929278,0.2648641169,-0.3902792037,0.2014696747,0.2545625865,0.3285291791,0.3011502028,-0.1477268934,-0.1272476465,0.1206850708,0.1623675674,-0.3560805619,0.180093348,0.1094101667,0.3291898072,-0.0979899392,0.3037487566,0.0115371151,-0.2040809095,0.0461662933,0.1682648659,0.666778326,0.03482721,0.0676321536,0.2113007456,-0.0096680988,0.1443352401,0.2122198343,-0.0167905204,0.0573296286,0.5108150244,-0.1701605171,-0.0370443836,0.3046994209,0.2604802251,-0.1606282145,-0.6539058685,0.2500951886,0.3208517134,-0.2636173368,-0.1137662232,0.1824982911,0.5810176134,0.0035529342,0.0682522133,-0.3186700046,-0.0973011479,0.0951301679,-0.0132612335,-0.213712886,-0.2044332623,-0.3498357236,0.2641787231,-0.1733150482,-0.0056747827,0.4497130513,0.1849378049,-0.1245447323,-0.5479856133,0.0882893279,0.2583621144,0.0428550318,-0.3176088929,0.1223269105,-0.0279606897,-0.0076812799,0.3330886364,-0.0986668393,0.3893023431,0.4318366051,0.1205390692,-0.3330965638,-0.1233512983,-0.0530246645,-0.0745155811,0.4672670662,-0.036854066,0.2521445453,0.2423177063,0.1311698705,-0.0224856287,0.2287944853,0.0818985552,0.2228803635,-0.4266517758,-0.1245938614,0.0320029855,-0.0391336977,-0.1038715616,-0.1802230626,-0.1029359922,-0.1934598833,0.4847235084,0.1438227892,0.22729671,0.1781719476,0.0232287813,-0.1932987124,0.1868279278,0.4103273451,-0.0159697421,-0.5346615314,-0.0820005834,-0.8077808619,-0.0004987333,-0.1560807824,-0.1756918877,0.0273725428,0.2069417387,0.2682349682,0.2591080666,-0.0207101535,0.1343044192,-0.0364686474,0.3154428899,-0.1966075301,-0.2066788375,-0.171055451,0.0741356239,-0.0421346314,-0.4293818176,0.0723792389,0.3088013828,-0.1039273292,-0.1886914968,-0.0647485033,0.0617913306,-0.2954472005,0.3805260956,0.0014711894,0.4309867918,-0.0023989477,0.3294302523,-0.4426386356,-0.2062971443,-0.1647175252,0.2472613603,-0.1406600177,0.4589703679,-0.3075934947,0.0491960049,-0.258679986,-0.0177992489,0.0659851134,0.2437577546,-0.374265939,0.2780484557,-0.1355160922,0.2285493165,-0.1251909584,-0.0648873895,-0.1331039369,0.2377364635,-0.0869754255,-0.276063323,0.3023509979,-0.2629050612,-0.2814046144,0.0674730167,0.2234089077,0.1626129299,0.0625678822,-0.4981215298,0.0108685168,0.3823477626,-0.1610226184,-0.1190343276,0.1228456125,0.0050548334,0.1080479622,-0.1312912107,0.1549797654,0.3163824379,-0.1977689862,-0.1361123621,-0.3402167261]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Hi, could this be a permission error ? I think it fails to close the arrow file that contains the data from your CSVs in the cache.\r\n\r\nBy default datasets are cached in `~\/.cache\/huggingface\/datasets`, could you check that you have the right permissions ?\r\nYou can also try to change the cache directory by passing `cache_dir=\"path\/to\/my\/cache\/dir\"` to `load_dataset`.","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":58,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nHi, could this be a permission error ? I think it fails to close the arrow file that contains the data from your CSVs in the cache.\r\n\r\nBy default datasets are cached in `~\/.cache\/huggingface\/datasets`, could you check that you have the right permissions ?\r\nYou can also try to change the cache directory by passing `cache_dir=\"path\/to\/my\/cache\/dir\"` to `load_dataset`.","embeddings":[-0.0871088356,0.1020701975,0.005998747,0.4314921498,0.2185793817,0.2184780687,0.4292903244,0.2389688194,0.3231713176,0.0248068217,-0.1687801331,0.2384760678,-0.1101290509,-0.0933418646,-0.2372374684,-0.0756902173,0.0261753611,0.0410047062,-0.3811805844,-0.0062905271,-0.4014434218,0.3093370497,-0.3083968759,-0.0182189606,-0.1635810286,0.1486214399,-0.0188286174,0.3629311025,-0.088870883,-0.3489400744,0.4900049269,-0.0543516316,0.5426036716,0.2987747788,-0.0001183373,0.0226844866,0.3176947236,-0.0851583183,-0.2275812477,-0.4275220931,-0.0256141461,-0.4985882044,0.3423484564,-0.06707187,-0.0731020123,0.0908294916,-0.0877925381,-0.4491149187,0.4408088923,0.5118373036,0.2004705667,-0.1390616298,-0.0094959829,0.0101438593,0.6154857874,-0.001861185,-0.1703196168,0.2800339758,0.0369507,-0.1588282287,-0.0883734897,0.1460739523,-0.0622680411,0.2507012188,0.31141904,0.1999147385,-0.2324631065,-0.2419393361,0.1065781191,0.0923586562,0.8474333882,-0.1498607844,-0.2260136008,-0.0473056883,-0.0552015752,-0.1341855377,0.3748573661,0.3546611965,0.0664210916,0.3369226456,-0.0629092678,-0.0407476351,-0.2813122272,0.1430380195,0.1430392563,-0.0449890979,-0.1894115806,0.2800143361,0.1055740714,-0.0342199914,0.3509489298,-0.0848069414,-0.1161021888,0.2580316663,-0.7147852182,0.3244916201,-0.1518708318,0.0887799039,-0.0345322937,0.2629768252,0.2805000842,0.022332605,-0.1302015036,0.3442869186,0.2256664336,0.1672879308,-0.0359066837,0.2805074751,0.3097871244,-0.2277148366,-0.0528800599,-0.1950272769,-0.2916230559,-0.5650448203,-0.0099357637,0.1524816453,0.2901365757,-0.2326605916,-0.136371091,0.124290213,-0.2214371413,0.0102816029,-0.090347901,0.3146436214,0.0904793143,0.3160606623,0.1083630994,0.2409510314,-0.1057104468,-0.3624810576,-0.0699226111,0.0720898286,-0.3807951212,0.038262289,0.4703016877,-0.1982618421,0.1263877898,0.1747371256,-0.0393530875,-0.2941902578,0.0493728518,-0.291264385,0.0694968998,0.3317597508,0.0307913981,0.2064109147,0.3324394524,-0.1748835742,-0.0883184075,0.2969932258,-0.4597531259,-0.2354933172,-0.0657947138,0.0806414336,-0.2515916526,0.132902354,-0.3708431125,-0.0102759385,0.1059380248,-0.1918373853,-0.1255963147,-0.2630356252,-0.2816921473,-0.3016434908,0.0190750826,0.5431545377,-0.7934693694,0.2218711078,-0.0641072839,0.1017690822,0.0040062084,0.2952950299,-0.2567481995,0.0294366553,-0.2967207432,0.0368544273,0.1628824174,-0.2294470072,-0.4436288476,0.1812134981,-0.0252471622,-0.0147933764,0.0866156742,0.0915315077,0.362511158,-0.0462300591,-0.2393852919,0.18926467,-0.0175592806,0.0613049679,-0.2599571645,-0.2040599734,0.224815473,0.1709872037,0.0435030572,0.1520606279,0.2611731589,-0.6375333071,0.2761252224,-0.0541094914,0.2460305691,0.1802657694,0.2484291196,0.0611739159,0.3364658654,0.0358324423,-0.596761167,0.2510197461,0.3427164853,-0.2168759406,-0.1212083325,-0.0457303263,-0.4000038207,0.0011727435,-0.2293895036,-0.0621970631,0.0090293186,0.0889609978,-0.0153363999,-0.1442162246,-0.0935701355,0.5005298257,-0.2158488333,0.081880495,-0.3989738524,0.300879091,0.1314293295,-0.0771669298,0.1612426043,-0.0463612601,0.3139698803,-0.0904841274,-0.2068674117,0.2543218434,0.3026152551,-0.0388666764,0.1233864725,-0.1583100259,0.0368795432,-0.3185792267,0.0433271006,0.1833157688,0.3288411796,-0.1012388244,-0.4540030956,0.2778144479,-0.3211692274,0.2337656766,-0.2180206031,0.156018585,0.2118747532,-0.0045894478,-0.0888199434,0.0230344869,0.0618801862,0.0452033319,0.3459661305,0.0847883001,-0.1154410541,-0.1102720797,0.1374965459,0.1963553876,-0.1975062042,0.2129110694,-0.1349883527,-0.0453927144,0.353230685,0.2573804855,0.4387530684,0.0898240507,-0.2252526581,-0.0238307621,0.1509096622,-0.152304545,0.0165314153,-0.0135874683,0.1102258638,0.16401577,0.1770403087,0.1246003509,-0.1302484274,-0.0883391201,0.1637058109,0.2491808832,-0.4925572872,0.0511966832,-0.1800087392,0.3748138547,-0.1793498993,-0.0850353241,0.0108298818,-0.2172825634,-0.31294626,0.3145366907,-0.0917653516,0.1111260876,-0.1890403479,0.2836256027,0.131011039,-0.1542268991,-0.0150205409,-0.1804295033,-0.2538110912,-0.0217656139,0.1527909189,-0.0590267442,-0.0866953731,0.0797144324,0.1314535141,-0.3408510387,-0.0016875594,0.0904464573,0.0303205382,0.1180033013,0.0525206178,-0.0170974117,-0.0492991284,-0.1359845102,0.1895927191,0.0173037294,0.0275386497,0.3591513038,0.2666297257,-0.1454355121,-0.0984572247,-0.6051182747,-0.2989300191,-0.3046963215,0.1061419174,-0.1025243104,-0.0014811994,0.116023019,0.2876195908,-0.0665228218,0.2434743196,-0.0484236106,-0.1217130795,-0.2950877845,0.5952745676,-0.2911517918,-0.3286848366,0.268263489,-0.0384926982,0.0919636637,0.3597333133,-0.4090351164,0.1694013774,-0.0824309289,0.0561728328,-0.0561686978,0.0685708374,0.0811326206,-0.0554329008,0.1393038779,-0.0295770783,-0.1968939304,0.1694541276,-0.0910888836,0.0490142629,0.2071580291,0.4857744575,-0.2566995025,0.7927243114,0.0675841272,0.1144968942,0.3349134922,0.0031341482,0.3701060116,0.1599618047,-0.2955479026,0.0845650658,-0.1591249257,-0.3198029697,0.3384991884,0.0584359095,-0.1529040188,-0.0706506893,0.2594616413,-0.2728158534,-0.2192439288,0.1771332622,-0.156610027,0.0803580061,0.0080378046,0.1146808714,-0.0200731382,-0.2414990366,-0.026980374,0.1884839833,0.1865026057,-0.0551850274,0.1528104693,-0.1228947043,-0.2886929214,0.2886113524,-0.1544601321,0.4078564346,0.0328982808,-0.0363260694,0.03973772,0.1954895258,0.776231885,-0.2733801603,0.0186978038,0.1182882115,-0.0143118752,-0.15080221,-0.3008757234,-0.001133894,0.2157413661,-0.0758567378,0.1894022971,-0.0392207578,0.0064115003,0.252094686,0.1005202681,-0.2171374708,-0.4065541029,-0.180907622,-0.0111552272,-0.1775229722,-0.2359721214,0.088709712,0.0679389983,-0.249018684,0.1165003926,-0.2494221032,-0.2153302282,0.2302294672,0.0987843052,0.3196050525,0.0893960968,0.1771992296,0.262976706,0.1359718889,0.2286501676,0.8212495446,-0.2555838227,-0.8745865226,0.2248188257,-0.4375938773,0.2542509139,-0.0137700578,-0.3433295488,-0.048304528,-0.2197701782,-0.1364994943,-0.0811645314,-0.0562362187,0.0102993418,0.0052309772,-0.0873702914,-0.5048304796,0.5631707311,-0.08336173,-0.1484945714,0.1437069327,-0.0225303229,-0.3080236912,0.4410199523,-0.1134528071,0.7396519184,-0.1897674054,-0.0513516031,0.0989359692,-0.2008097321,0.3925123215,-0.3037558198,-0.0298086535,-0.4993213117,-0.1921999305,-0.1562449187,-0.038052313,0.0910793245,0.54327631,-0.1915143877,0.3158293664,-0.1240997836,0.3808956444,0.0440304764,0.0923432037,0.2995007634,-0.2266685665,-0.1182795316,0.1387898475,-0.1794545949,0.0204701405,0.0378143527,-0.0112659847,-0.2284193337,-0.0456547961,-0.352573663,0.3131829798,-0.1434078217,0.0971926302,0.1712797582,-0.6031698585,0.5004423261,0.4744847417,0.2492394,-0.0955313444,-0.2055733353,0.0834538937,0.0743317902,-0.0045969253,-0.2426962107,0.0059339567,0.0072347801,0.1907503307,-0.1637479216,0.3280681968,-0.1499566436,-0.1462507546,0.2858631015,0.3220999539,0.2411754876,-0.4651108086,-0.3601719439,-0.2402686179,0.0512117557,-0.0718256831,0.0563277192,0.0646344423,0.0796990544,-0.0023343952,0.0640948117,-0.3267735541,0.1688649952,0.2198811918,-0.2082929611,0.0752046704,0.4720528126,0.4608855844,-0.1236692518,-0.1189092323,0.0018366389,0.1272981465,-0.6390781999,0.1294370741,0.1166711524,0.2417464554,0.1317799985,0.124538824,0.2230076343,-0.1942317486,0.073514156,-0.5760642886,-0.1028078645,-0.0645833462,-0.1964429766,0.0786467642,-0.057004869,-0.5029879212,-0.057699319,-0.017358562,-0.2017152756,0.0509648435,-0.0363410488,-0.0126664909,0.1817023605,0.2421429902,-0.2382351756,-0.2921729982,-0.0281904265,0.0323550664,0.0052473932,-0.1087043509,-0.1650231779,0.1757085025,-0.0072001955,0.1040344536,-0.307500869,-0.2177542597,-0.1669041961,-0.2070279568,0.1396328509,0.2555310428,-0.0076166671,0.3323541582,0.0271904208,-0.1070898399,-0.5441235304,0.2826118171,-0.2114539742,0.187242344,0.1003946438,0.1679219306,0.1192823797,0.0578045286,-0.1548666209,0.0199019685,-0.0952334926,-0.1539371908,0.3234480023,-0.3792797923,0.1397074312,0.2306100279,0.2930877507,0.2602367103,-0.1094966531,-0.0748061538,0.1451209188,0.1173848137,-0.3229917586,0.1120470613,0.0211968869,0.2629726529,-0.0840862468,0.2402922958,-0.0250097718,-0.1733133346,0.1008311138,0.2054375708,0.6525261998,0.0066219433,0.0822381973,0.1425482333,-0.0172165353,0.1489227265,0.297252208,0.0581633784,0.0477336869,0.439869076,-0.1683690548,-0.0418318026,0.3462232351,0.1906350255,-0.114179939,-0.6883136034,0.1853644848,0.272980988,-0.1937285811,-0.1051933765,0.1018466651,0.5970719457,0.0254736245,-0.0268965084,-0.2631659508,-0.0485790484,0.0614283457,-0.0357040428,-0.1355413944,-0.1496821642,-0.3093096614,0.2167368829,-0.1998447925,-0.081889227,0.4407967627,0.1543577015,-0.1003409773,-0.4828207195,0.1358698606,0.26647228,0.0216101296,-0.263672322,0.1589758992,-0.0735052526,-0.0148157664,0.2434874028,-0.1003579125,0.3400716484,0.4579682648,0.0386538729,-0.2136320025,-0.0977050886,-0.0766636059,-0.0986232534,0.4576211572,-0.0504556522,0.2011354715,0.1784655899,0.1070088819,-0.0207585469,0.3069820404,-0.0026891003,0.2105031312,-0.3866465688,-0.148170501,-0.0119448826,-0.0929755867,-0.1426808238,-0.1752101332,-0.1145598665,-0.1086821258,0.4643135369,0.1401850283,0.1937510371,0.0654954687,0.0033941839,-0.1118401662,0.2868217826,0.5028715134,0.1731095463,-0.5306534767,-0.0562082902,-0.8355731964,0.0631824583,-0.2315540016,-0.0948862061,0.0296853334,0.2148626,0.2593040168,0.2721314728,-0.0273133498,0.1377696693,-0.0158321839,0.2853050828,-0.2362405062,-0.1603053808,-0.1035052687,0.0983134136,-0.0479214191,-0.4799541831,0.0935249999,0.3068058193,-0.1077924892,-0.1020182669,-0.1223229393,0.0748175234,-0.1577813625,0.4501149952,-0.0165146664,0.3705914021,0.0817033872,0.2856374979,-0.5112701654,-0.2877731621,-0.1633494794,0.0769248754,-0.1115249917,0.4098854065,-0.3690055907,0.0069668032,-0.2801869512,0.0396508127,0.0769017264,0.2473265678,-0.274169296,0.226364553,-0.1124691889,0.2035575658,-0.0860854313,-0.1226061508,-0.1627916545,0.1960602552,-0.1498049945,-0.2431539148,0.3379566669,-0.2973978817,-0.2897456884,0.0942837596,0.2492740452,0.1518336684,-0.0067107845,-0.4973727465,0.0293828137,0.3351517618,-0.1269314289,-0.1486630887,0.0459613055,0.0588275529,0.1021991447,-0.164056465,0.3040836453,0.3110781908,-0.1586262882,-0.1608160287,-0.3319260478]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/743","title":"load_dataset for CSV files not working","comments":"Thank you!! @lhoestq\r\n\r\nFor some reason, I don't have the default path for datasets to cache, maybe because I work from a remote system. The issue solved after I pass the `cache_dir` argument to the function. Thank you very much!!","body":"Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you","comment_length":40,"text":"load_dataset for CSV files not working\nSimilar to #622, I've noticed there is a problem when trying to load a CSV file with datasets.\r\n\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv\", data_files=[\".\/sample_data.csv\"], delimiter=\"\\t\", column_names=[\"title\", \"text\"], script_version=\"master\")\r\n`\r\n\r\nDisplayed error:\r\n`\r\n...\r\nArrowInvalid: CSV parse error: Expected 2 columns, got 1\r\n`\r\n\r\nI should mention that when I've tried to read data from `https:\/\/github.com\/lhoestq\/transformers\/tree\/custom-dataset-in-rag-retriever\/examples\/rag\/test_data\/my_knowledge_dataset.csv` it worked without a problem. I've read that there might be some problems with \/r character, so I've removed them from the custom dataset, but the problem still remains.\r\n\r\nI've added a colab reproducing the bug, but unfortunately I cannot provide the dataset.\r\nhttps:\/\/colab.research.google.com\/drive\/1Qzu7sC-frZVeniiWOwzoCe_UHZsrlxu8?usp=sharing\r\n\r\nAre there any work around for it ?\r\nThank you\nThank you!! @lhoestq\r\n\r\nFor some reason, I don't have the default path for datasets to cache, maybe because I work from a remote system. The issue solved after I pass the `cache_dir` argument to the function. Thank you very much!!","embeddings":[-0.0859662145,0.0921201184,0.0054692053,0.3930594325,0.2793588638,0.1994730979,0.4239506125,0.2563423812,0.3121513426,0.00980769,-0.1000402272,0.2833045721,-0.0902011842,-0.0692626685,-0.2416357994,-0.0534538962,0.0487427153,0.0339308605,-0.4194591641,0.0217577126,-0.398001343,0.3447437286,-0.3417994976,-0.0670136884,-0.1728119552,0.2312749773,-0.0274074618,0.3173514903,-0.0774215236,-0.3483394682,0.49772048,-0.1169507429,0.500279367,0.3159280419,-0.0001189884,0.0335704125,0.3231101632,-0.1217699423,-0.2798947692,-0.5001267195,-0.0860216245,-0.4523081481,0.3036141992,-0.078889586,-0.0746764019,0.1003823653,-0.0871309862,-0.4603987932,0.3343280852,0.5259562135,0.1874039024,-0.2362770885,-0.1149116457,0.0203485657,0.5270505548,0.0821966901,-0.1715659946,0.2562561035,-0.0119706914,-0.1819378287,-0.0650214925,0.1478487253,-0.0790288895,0.2778861821,0.2819365561,0.2120058984,-0.2090995163,-0.1924725622,0.1150583476,0.0516182072,0.7780841589,-0.1751658916,-0.2040853649,-0.0703984648,-0.0710535273,-0.0905073807,0.3608498275,0.2796316147,0.0471044742,0.3526415825,-0.0739563033,0.0198418032,-0.2478520125,0.1096846163,0.1472314298,-0.0370750353,-0.1845748425,0.3111932576,0.0819834694,0.0278162509,0.3614430428,-0.1226573363,-0.0612086058,0.2232882828,-0.6913048029,0.3598780632,-0.1117809787,0.0476231165,-0.0488829464,0.2159273475,0.253226757,-0.0178384166,-0.0982983708,0.3449589312,0.2084871233,0.0942893401,0.03507635,0.249590531,0.2458130419,-0.240046218,-0.0242090952,-0.1735283583,-0.3206408918,-0.4924867153,0.0513884798,0.1816558838,0.211786285,-0.2026416957,-0.1311562359,0.1114847362,-0.1737065762,-0.0094566187,-0.0949874967,0.3192047179,0.0954763293,0.3276444077,0.0360649265,0.2327312231,-0.1163586378,-0.3847432733,-0.0511320345,0.1306393594,-0.3966401219,-0.0078006592,0.499743551,-0.2116759121,0.1611486226,0.2013804168,-0.1022118405,-0.3209100664,0.0098241894,-0.3409103155,0.1178882346,0.2810696959,0.0596529432,0.2322187573,0.310836643,-0.1735278219,-0.0792894661,0.3202610612,-0.4694957733,-0.2254518718,-0.053723976,0.0856733099,-0.2701482773,0.1136537343,-0.3645306528,0.0592239499,0.0930911601,-0.2261726111,-0.148012653,-0.2921352684,-0.3328604698,-0.2874190509,0.0227104947,0.527285099,-0.7632331848,0.2065825462,0.0069278814,0.0596400052,0.077546671,0.201359719,-0.2320704013,-0.0131589593,-0.3008173406,0.084708415,0.1511888057,-0.2414093018,-0.4044985175,0.1948494166,-0.0091790361,-0.0367273539,0.1076726317,0.072963208,0.3510636687,-0.0444780849,-0.1565225273,0.2063654363,0.0127947778,0.079222329,-0.2519009113,-0.1841476113,0.2559945285,0.1668462902,0.0265599005,0.1279062033,0.2115993351,-0.6256981492,0.2878684402,-0.026080206,0.2456631958,0.1336244941,0.240569219,0.020431878,0.3675477803,0.0579219759,-0.5920143127,0.2176810503,0.4573203027,-0.2367344201,-0.12829566,-0.0426198766,-0.4006431401,0.0049440595,-0.1802303642,-0.051015377,0.0053335573,0.1431531161,0.0037192472,-0.166486457,-0.1181761697,0.5300241709,-0.2569983602,0.0464286283,-0.3279723525,0.2669471502,0.1607712805,-0.0713458657,0.1533521861,-0.0794750452,0.3291767538,-0.0821014568,-0.2026355863,0.2568963766,0.3679184616,-0.0227024481,0.0851090699,-0.1917208433,-0.0070678876,-0.2358018756,-0.0132546611,0.2161183953,0.3630660474,-0.0771617219,-0.4703349769,0.282397002,-0.329693377,0.2206449211,-0.1862371415,0.1222787425,0.2683041692,-0.0067571732,-0.0762864873,0.0306818839,0.0430371426,-0.016078135,0.3278394043,0.1356023103,-0.1543986797,-0.0686571375,0.104084596,0.2219552994,-0.1962908655,0.2119081914,-0.1135942265,-0.1372815222,0.3695862293,0.2924639583,0.4771190882,0.1017706245,-0.2151629329,-0.03321575,0.1260908991,-0.1481108814,0.0860050619,-0.0315163843,0.0596535169,0.1926266849,0.2122889012,0.0930455029,-0.1251711696,-0.1155806929,0.17658481,0.2235135734,-0.5865730047,0.0359858647,-0.2217503041,0.3033283353,-0.1915452182,-0.0284086727,0.1247221753,-0.2356092483,-0.3277724385,0.3035749495,-0.0688414797,0.0889437273,-0.1954190433,0.2737675011,0.0884293616,-0.2449583858,-0.0298119988,-0.1432175487,-0.2710935771,-0.0190364961,0.0947937593,0.0005641778,-0.1103101224,0.0575547032,0.1049836054,-0.3393681943,0.0112823285,0.138859272,0.0470259152,0.1042666659,0.1056636572,-0.1229335219,-0.0585625321,-0.100685738,0.1599597633,0.0252639428,0.0772768706,0.3538436592,0.229650408,-0.1431204975,-0.1136703938,-0.6079529524,-0.3110757172,-0.2347203642,0.0422148593,-0.1346933544,-0.0265995637,0.0542938448,0.2813584208,-0.0934910253,0.2276720554,-0.0884653553,-0.1698265374,-0.3080235124,0.6029754877,-0.2578561306,-0.2866512835,0.2364552021,-0.0638770014,0.0657695159,0.3948900402,-0.4063292742,0.1408141255,-0.0391779132,0.0357070789,-0.0525728613,0.0657720342,0.0635105297,-0.0665663034,0.1175326779,-0.0782246068,-0.2430461198,0.2015887648,-0.0742934272,0.0763092563,0.2349738181,0.4599756002,-0.196335122,0.6576650143,0.1088631675,0.0870276839,0.3282864988,-0.0926542878,0.3686032593,0.1682643145,-0.2678069174,0.0895845443,-0.0932174847,-0.3670552671,0.269513756,0.0467762202,-0.123863019,-0.0919571966,0.2348202169,-0.2747565806,-0.1522409767,0.1995553076,-0.0784694254,0.0519669317,0.0399094783,0.1436175853,-0.0207252186,-0.2149112374,-0.0282757431,0.1274815798,0.2044937611,-0.0482011177,0.1675854921,-0.1045095399,-0.3080209792,0.2213120461,-0.0817261115,0.437636584,0.071454607,-0.0554930866,-0.0034884836,0.1947611272,0.7499451637,-0.336463362,0.0298958328,0.1303633898,-0.0646097213,-0.2020518929,-0.2559672892,0.0101779513,0.2573974431,-0.044601921,0.2114073932,0.0081580738,0.0392203555,0.3173394799,0.124707967,-0.1731449962,-0.4302400947,-0.1287029386,-0.0440912247,-0.1343362182,-0.260258913,0.0956653655,-0.0007538416,-0.1850678474,0.0882486552,-0.2512223423,-0.1873823851,0.2498172224,0.1124929935,0.2862683237,0.0475280806,0.197009027,0.2487416863,0.0807058066,0.2081256211,0.7596247792,-0.2756142914,-0.8639798164,0.2319833487,-0.4705553651,0.2806535959,-0.0335888378,-0.3992097676,-0.1155501232,-0.2685472071,-0.1504105777,-0.0894087255,-0.1064961851,0.0078194635,0.0154668875,-0.1265347749,-0.5553480387,0.5315487981,-0.0984754786,-0.1718450189,0.1925376654,-0.1100402623,-0.3113648891,0.4189883471,-0.1144802198,0.6271879077,-0.1241634935,-0.1032958105,0.1455404907,-0.1528799385,0.3744075596,-0.2721971273,-0.0416736603,-0.4573805034,-0.1809865385,-0.1555781662,-0.0239725634,0.0471607968,0.5235294104,-0.2072220296,0.3843708336,-0.1009207517,0.4271258712,0.0781405419,0.0950178578,0.2918179929,-0.2190098614,-0.1630765498,0.1391296983,-0.1176220551,-0.0087528676,0.0714028627,0.0635630563,-0.2258556932,-0.0736265332,-0.3024659157,0.3099745512,-0.0996361226,0.0925187618,0.1466941237,-0.51267308,0.4581801891,0.4478138387,0.1973773241,-0.102102913,-0.1980636865,0.0802133009,0.0769241527,0.0982713923,-0.2025686651,-0.0496243164,-0.0435052775,0.2136051953,-0.1751904637,0.242034182,-0.1271027774,-0.1109358519,0.3289182186,0.2879223824,0.2425295115,-0.4307258129,-0.387932092,-0.1784506142,0.1143456101,-0.0891140252,0.0522813164,-0.0043365946,0.1033893228,0.0306254122,0.0312224608,-0.3578824401,0.1549049169,0.2224382609,-0.2110670209,0.0424115621,0.4137513936,0.4314633906,-0.1372980624,-0.1447460353,-0.0101709692,0.1325487643,-0.5681897402,0.1376532018,0.0400586762,0.2680269778,0.136052981,0.1017231271,0.1819981635,-0.1829439551,0.063343063,-0.5674280524,-0.048464857,-0.0937195048,-0.1349463612,0.0383743159,0.0078521455,-0.5268148184,-0.0391837284,-0.0920737609,-0.1954650134,0.0826359913,0.0543640368,0.01612808,0.1905421913,0.3258012235,-0.2146797031,-0.2855106294,0.0000114315,0.0599250384,0.0498943999,-0.1171048805,-0.1825485826,0.1764264554,-0.0121126343,0.1497071981,-0.3138529956,-0.205768615,-0.2322503775,-0.2072654665,0.10746824,0.3564963937,-0.0696219951,0.3649552763,-0.0061729718,-0.1737711132,-0.4967603385,0.290572226,-0.2242492586,0.1979621649,0.0926667377,0.2064141035,0.1576426327,0.1024193019,-0.1475694478,0.0159455519,-0.1610230058,-0.0975365266,0.2784028351,-0.3515321016,0.1970977783,0.2572290301,0.2927462161,0.2969406545,-0.1435653716,-0.0857069194,0.1398878098,0.1210196763,-0.3374234736,0.1304138303,0.0534944199,0.3156667054,-0.0910222828,0.2972840071,0.0034365244,-0.2099749595,0.1134114638,0.1574269533,0.7327905297,-0.0232780278,0.0393834524,0.1834812313,0.0020634176,0.1581076533,0.2012578398,0.0261133611,0.0677802265,0.4985857904,-0.1393862218,-0.031464573,0.3940755725,0.2223598659,-0.2321441621,-0.6349562407,0.2484429479,0.3177166283,-0.2765003145,-0.1067677438,0.1735525876,0.5750449896,0.0018488275,0.0608277768,-0.2844551504,-0.0516806096,0.1050643399,-0.0605890043,-0.1403058171,-0.1845536381,-0.347091049,0.2614604831,-0.1911302805,-0.0177368429,0.4982432723,0.1881555766,-0.0476525463,-0.5127802491,0.1151892841,0.2723284066,0.0355487093,-0.294654876,0.126491487,-0.0687676147,-0.0099533666,0.2945648134,-0.1287981123,0.4101406038,0.4447221756,0.1125388071,-0.3156537712,-0.1382453889,-0.0626907051,-0.1432363391,0.4691237509,-0.0354215279,0.2465351224,0.2225378156,0.1133174151,-0.0038163341,0.3163536489,0.0765826702,0.1817938685,-0.3753982484,-0.1211301759,0.0400015898,-0.0224781726,-0.1208899617,-0.139284268,-0.1074324995,-0.1666510105,0.5061461329,0.1402078122,0.2272045165,0.1511776,0.0067096525,-0.1303758472,0.2235109359,0.4077506661,0.1012787297,-0.5023837686,-0.0518347025,-0.8332253098,0.003514016,-0.2273966819,-0.0952329412,0.0056153727,0.2180944234,0.2635284662,0.2751927376,-0.031958811,0.1715863794,-0.0168863814,0.3326143324,-0.2261124551,-0.1757160872,-0.1433279067,0.1256488413,-0.0311014075,-0.4723520279,0.0634715185,0.3111626506,-0.1095227078,-0.1654501259,-0.1251498163,0.0520647392,-0.246922031,0.4691313207,-0.0141462293,0.37859115,0.0237498544,0.33029598,-0.4599416256,-0.2367649525,-0.14237611,0.1348722428,-0.1412815154,0.3937490284,-0.4040461779,0.0120197264,-0.2420687973,-0.0377768539,0.0698111951,0.2664785087,-0.3444803357,0.2908079326,-0.1341441125,0.2600900233,-0.1846758425,-0.1040553227,-0.1662362516,0.2463909835,-0.0808947384,-0.2135210484,0.3060748875,-0.2560876012,-0.2865568995,-0.0037384627,0.2286492586,0.1515983343,0.043349009,-0.4878182411,-0.0173042063,0.3493456244,-0.1020397171,-0.1388151348,0.1015590578,0.0189562291,0.1301361918,-0.1563574225,0.2366661578,0.2966106534,-0.1945958287,-0.1509202123,-0.3429087996]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Thanks for reporting.\r\nIn theory since the dataset script is just made to yield examples to write them into an arrow file, it's not supposed to create memory issues.\r\n\r\nCould you please try to run this exact same loop in a separate script to see if it's not an issue with `PIL` ?\r\nYou can just copy paste what's inside `_generate_examples` and remove all the code for `datasets` (remove yield).\r\n\r\nIf the RAM usage stays low after 600 examples it means that it comes from some sort of memory leak in the library, or with pyarrow.","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":96,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nThanks for reporting.\r\nIn theory since the dataset script is just made to yield examples to write them into an arrow file, it's not supposed to create memory issues.\r\n\r\nCould you please try to run this exact same loop in a separate script to see if it's not an issue with `PIL` ?\r\nYou can just copy paste what's inside `_generate_examples` and remove all the code for `datasets` (remove yield).\r\n\r\nIf the RAM usage stays low after 600 examples it means that it comes from some sort of memory leak in the library, or with pyarrow.","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Here's an equivalent loading code:\r\n```python\r\nimages_path = \"PHOENIX-2014-T-release-v3\/PHOENIX-2014-T\/features\/fullFrame-210x260px\/train\"\r\n\r\nfor dir_path in tqdm(os.listdir(images_path)):\r\n frames_path = os.path.join(images_path, dir_path)\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n```\r\n\r\nThe process takes 0.3% of memory, even after 1000 examples on the small machine with 120GB RAM.\r\n\r\nI guess something in the datasets library doesn't release the reference to the objects I'm yielding, but no idea how to test for this","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":75,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nHere's an equivalent loading code:\r\n```python\r\nimages_path = \"PHOENIX-2014-T-release-v3\/PHOENIX-2014-T\/features\/fullFrame-210x260px\/train\"\r\n\r\nfor dir_path in tqdm(os.listdir(images_path)):\r\n frames_path = os.path.join(images_path, dir_path)\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n```\r\n\r\nThe process takes 0.3% of memory, even after 1000 examples on the small machine with 120GB RAM.\r\n\r\nI guess something in the datasets library doesn't release the reference to the objects I'm yielding, but no idea how to test for this","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"I've had similar issues with Arrow once. I'll investigate...\r\n\r\nFor now maybe we can simply use the images paths in the dataset you want to add. I don't expect to fix this memory issue until 1-2 weeks unfortunately. Then we can just update the dataset with the images. What do you think ?","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":53,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nI've had similar issues with Arrow once. I'll investigate...\r\n\r\nFor now maybe we can simply use the images paths in the dataset you want to add. I don't expect to fix this memory issue until 1-2 weeks unfortunately. Then we can just update the dataset with the images. What do you think ?","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"If it's just 1-2 weeks, I think it's best if we wait. I don't think it is very urgent to add it, and it will be much more useful with the images loaded rather than not (the images are low resolution, and thus papers using this dataset actually fit the entire video into memory anyway)\r\n\r\nI'll keep working on other datasets in the meanwhile :) ","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":65,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nIf it's just 1-2 weeks, I think it's best if we wait. I don't think it is very urgent to add it, and it will be much more useful with the images loaded rather than not (the images are low resolution, and thus papers using this dataset actually fit the entire video into memory anyway)\r\n\r\nI'll keep working on other datasets in the meanwhile :) ","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Ok found the issue. This is because the batch size used by the writer is set to 10 000 elements by default so it would load your full dataset in memory (the writer has a buffer that flushes only after each batch). Moreover to write in Apache Arrow we have to use python objects so what's stored inside the ArrowWriter's buffer is actually python integers (32 bits).\r\n\r\nLowering the batch size to 10 should do the job.\r\n\r\nI will add a flag to the DatasetBuilder class of dataset scripts, so that we can customize the batch size.","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":97,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nOk found the issue. This is because the batch size used by the writer is set to 10 000 elements by default so it would load your full dataset in memory (the writer has a buffer that flushes only after each batch). Moreover to write in Apache Arrow we have to use python objects so what's stored inside the ArrowWriter's buffer is actually python integers (32 bits).\r\n\r\nLowering the batch size to 10 should do the job.\r\n\r\nI will add a flag to the DatasetBuilder class of dataset scripts, so that we can customize the batch size.","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Thanks, that's awesome you managed to find the problem.\r\n\r\nAbout the 32 bits - really? there isn't a way to serialize the numpy array somehow? 32 bits would take 4 times the memory \/ disk space needed to store these videos.\r\n\r\nPlease let me know when the batch size is customizable and I'll try again!","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":55,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nThanks, that's awesome you managed to find the problem.\r\n\r\nAbout the 32 bits - really? there isn't a way to serialize the numpy array somehow? 32 bits would take 4 times the memory \/ disk space needed to store these videos.\r\n\r\nPlease let me know when the batch size is customizable and I'll try again!","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"The 32 bit integrers are only used in the writer's buffer because Arrow doesn't take numpy arrays correctly as input. On disk it's stored as uint8 in arrow format ;)","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":30,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nThe 32 bit integrers are only used in the writer's buffer because Arrow doesn't take numpy arrays correctly as input. On disk it's stored as uint8 in arrow format ;)","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"> I don't expect to fix this memory issue until 1-2 weeks unfortunately.\r\n\r\nHi @lhoestq \r\nnot to rush of course, but I was wondering if you have a new timeline so I know how to plan my work around this :) ","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":41,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\n> I don't expect to fix this memory issue until 1-2 weeks unfortunately.\r\n\r\nHi @lhoestq \r\nnot to rush of course, but I was wondering if you have a new timeline so I know how to plan my work around this :) ","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Hi ! Next week for sure :) ","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":7,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nHi ! Next week for sure :) ","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Alright it should be good now.\r\nYou just have to specify `_writer_batch_size = 10` for example as a class attribute of the dataset builder class.","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":25,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nAlright it should be good now.\r\nYou just have to specify `_writer_batch_size = 10` for example as a class attribute of the dataset builder class.","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"I added it, but still it consumes as much memory\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/722\/files#diff-2e0d865dd4a60dedd1861d6f8c5ed281ded71508467908e1e0b1dbe7d2d420b1R66\r\n\r\nDid I not do it correctly?","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":17,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nI added it, but still it consumes as much memory\r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/722\/files#diff-2e0d865dd4a60dedd1861d6f8c5ed281ded71508467908e1e0b1dbe7d2d420b1R66\r\n\r\nDid I not do it correctly?","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Yes you did it right.\r\nDid you rebase to include the changes of #828 ?\r\n\r\nEDIT: looks like you merged from master in the PR. Not sure why you still have an issue then, I will investigate","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":37,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nYes you did it right.\r\nDid you rebase to include the changes of #828 ?\r\n\r\nEDIT: looks like you merged from master in the PR. Not sure why you still have an issue then, I will investigate","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Hi @lhoestq, any update on this?\r\nPerhaps even a direction I could try myself?","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":14,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nHi @lhoestq, any update on this?\r\nPerhaps even a direction I could try myself?","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Sorry for the delay, I was busy with the dataset sprint and the incredible amount of contributions to the library ^^'\r\n\r\nWhat you can try to do to find what's wrong is check at which frequency the arrow writer writes all the examples from its in-memory buffer on disk. This happens [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L257-L258) in the code.\r\n\r\nThe idea is that `write_on_file` writes the examples every `writer_batch_size` examples and clear the buffer `self. current_rows`. As soon as `writer_batch_size` is small enough you shouldn't have memory issues in theory.\r\n\r\nLet me know if you have questions or if I can help.\r\n\r\nSince the dataset sprint is over and I will also be done with all the PRs soon I will be able to go back at it and take a look.","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":128,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nSorry for the delay, I was busy with the dataset sprint and the incredible amount of contributions to the library ^^'\r\n\r\nWhat you can try to do to find what's wrong is check at which frequency the arrow writer writes all the examples from its in-memory buffer on disk. This happens [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/datasets\/arrow_writer.py#L257-L258) in the code.\r\n\r\nThe idea is that `write_on_file` writes the examples every `writer_batch_size` examples and clear the buffer `self. current_rows`. As soon as `writer_batch_size` is small enough you shouldn't have memory issues in theory.\r\n\r\nLet me know if you have questions or if I can help.\r\n\r\nSince the dataset sprint is over and I will also be done with all the PRs soon I will be able to go back at it and take a look.","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Thanks. I gave it a try and no success. I'm not sure what's happening there","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":15,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nThanks. I gave it a try and no success. I'm not sure what's happening there","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"I had the same issue. It works for me by setting `DEFAULT_WRITER_BATCH_SIZE = 10` of my dataset builder class. (And not `_writer_batch_size` as previously mentioned). I guess this is because `_writer_batch_size` is overwritten in `__init__` (see [here](https:\/\/github.com\/huggingface\/datasets\/blob\/0e2563e5d5c2fc193ea27d7c24607bb35607f2d5\/src\/datasets\/builder.py#L934))","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":37,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nI had the same issue. It works for me by setting `DEFAULT_WRITER_BATCH_SIZE = 10` of my dataset builder class. (And not `_writer_batch_size` as previously mentioned). I guess this is because `_writer_batch_size` is overwritten in `__init__` (see [here](https:\/\/github.com\/huggingface\/datasets\/blob\/0e2563e5d5c2fc193ea27d7c24607bb35607f2d5\/src\/datasets\/builder.py#L934))","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Yes the class attribute you can change is `DEFAULT_WRITER_BATCH_SIZE`.\r\nOtherwise in `load_dataset` you can specify `writer_batch_size=`","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":16,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nYes the class attribute you can change is `DEFAULT_WRITER_BATCH_SIZE`.\r\nOtherwise in `load_dataset` you can specify `writer_batch_size=`","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Ok thanks for the tips. Maybe the documentation should be updated accordingly https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html.","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":13,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nOk thanks for the tips. Maybe the documentation should be updated accordingly https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html.","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/741","title":"Creating dataset consumes too much memory","comments":"Thanks for reporting this mistake in the docs.\r\nI just fixed it at https:\/\/github.com\/huggingface\/datasets\/commit\/85cf7ff920c90ca2e12bedca12b36d2a043c3da2","body":"Moving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n","comment_length":14,"text":"Creating dataset consumes too much memory\nMoving this issue from https:\/\/github.com\/huggingface\/datasets\/pull\/722 here, because it seems like a general issue.\r\n\r\nGiven the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400):\r\n```python\r\n def _generate_examples(self, base_path, split):\r\n \"\"\" Yields examples. \"\"\"\r\n\r\n filepath = os.path.join(base_path, \"annotations\", \"manual\", \"PHOENIX-2014-T.\" + split + \".corpus.csv\")\r\n images_path = os.path.join(base_path, \"features\", \"fullFrame-210x260px\", split)\r\n\r\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\r\n data = csv.DictReader(f, delimiter=\"|\", quoting=csv.QUOTE_NONE)\r\n for row in data:\r\n frames_path = os.path.join(images_path, row[\"video\"])[:-7]\r\n np_frames = []\r\n for frame_name in os.listdir(frames_path):\r\n frame_path = os.path.join(frames_path, frame_name)\r\n im = Image.open(frame_path)\r\n np_frames.append(np.asarray(im))\r\n im.close()\r\n\r\n yield row[\"name\"], {\"video\": np_frames}\r\n```\r\n\r\nThe dataset creation process goes out of memory on a machine with 500GB RAM.\r\nI was under the impression that the \"generator\" here is exactly for that, to avoid memory constraints.\r\n\r\n\r\nHowever, even if you want the entire dataset in memory, it would be in the worst case\r\n`260x210x3 x 400 max length x 7000 samples` in bytes (uint8) = 458.64 gigabytes\r\nSo I'm not sure why it's taking more than 500GB.\r\n\r\nAnd the dataset creation fails after 170 examples on a machine with 120gb RAM, and after 672 examples on a machine with 500GB RAM.\r\n\r\n\r\n---\r\n\r\n## Info that might help:\r\nIterating over examples is extremely slow.\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359590-3c666780-111d-11eb-9347-1f833ad982a9.png)\r\nIf I perform this iteration in my own, custom loop (Without saving to file), it runs at 8-9 examples\/sec\r\n\r\nAnd you can see at this state it is using 94% of the memory:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359606-7afc2200-111d-11eb-8c11-0afbdba1a6a3.png)\r\n\r\nAnd it is only using one CPU core, which is probably why it's so slow:\r\n![image](https:\/\/user-images.githubusercontent.com\/5757359\/96359630-a3841c00-111d-11eb-9ba0-7fd3cdf51d26.png)\r\n\nThanks for reporting this mistake in the docs.\r\nI just fixed it at https:\/\/github.com\/huggingface\/datasets\/commit\/85cf7ff920c90ca2e12bedca12b36d2a043c3da2","embeddings":[-0.2472248971,-0.0484526679,-0.0161761772,0.2615477145,0.1814279854,0.2023240179,0.1410058886,0.2836433053,-0.0936915278,0.2016249001,0.4972395897,0.1490257382,-0.234624356,0.0067335241,-0.0158069357,-0.3033439517,0.0866438746,0.0552453026,-0.1493512541,-0.0251512229,-0.4015744925,-0.0364088193,0.0979415551,-0.2667103112,-0.5073342323,-0.2018489987,-0.0731501654,0.0787201747,-0.1270601451,-0.2315939218,-0.1274705678,-0.0400048792,0.1177601293,0.3480851054,-0.0001107187,-0.0142071992,0.2528391778,0.056058187,-0.2610005736,0.2551356554,0.0024836487,-0.2779872119,-0.1727146357,-0.0488479808,0.2572742403,-0.300396353,0.1659740806,-0.3587361276,0.2852363586,0.3103289604,0.1906799227,0.2333978266,0.0225876123,-0.0645367801,0.0283546913,0.1364507228,-0.2712320983,-0.0043950318,-0.0322477482,-0.1391170472,-0.1249336302,0.3213855624,0.1251342595,0.0555912778,0.3767856956,-0.0258995909,-0.3605363071,-0.4826664329,0.1298619807,0.2105497569,0.4798392355,-0.2130855024,-0.1907441765,-0.3127378821,-0.0608719289,-0.2809078097,-0.0386913605,0.5055589676,-0.4797238708,-0.156840533,-0.4373593628,0.0112411054,-0.0739322454,-0.0789558962,0.2725889087,-0.4267039895,-0.0361702442,0.2112913728,0.4411447942,0.0069410964,0.1683401465,-0.1735475063,0.0405645072,-0.0908014327,-0.2532210648,-0.0523642302,0.2876224518,-0.1453632712,0.3291010559,-0.0412678234,0.024514053,-0.2715124786,0.1530065686,-0.0058941585,0.0967000052,0.2201670259,-0.4858937263,-0.1767417938,0.2278588116,0.0444847383,0.0899148658,-0.159083873,-0.0274225194,-0.1531633884,0.2210888267,-0.3335966468,0.1902000308,0.2805503607,-0.2625469565,0.2027172446,-0.2653440237,0.1029191092,-0.146130681,0.1159540489,0.0655315965,0.1141222045,-0.1104514003,0.1520260572,-0.2346293479,-0.2064009309,-0.2843567729,0.0219008811,-0.2518551946,0.201895386,0.133929655,0.2431201041,0.217063278,0.0080225421,0.0536194853,-0.0079490449,0.4056237936,-0.2521450818,-0.006455563,0.2825173438,0.1068824902,0.1090517938,-0.0290899146,0.1394088417,-0.0898926258,0.2382693589,0.0490416102,-0.4122443199,0.0999208838,0.1508389413,0.095170334,0.2297330797,-0.4434710443,0.2292865813,0.1947136968,0.3888747096,-0.054538507,-0.0666551888,-0.3439253271,-0.2636183798,0.1964847147,0.7139770985,-0.3620707393,0.2919789255,-0.1818291992,0.1198833138,0.1442041844,0.4414608181,-0.0528309084,0.4882627726,-0.2243760228,-0.2710960507,-0.0880410895,-0.094117634,-0.2132639587,0.2518543005,-0.1075511724,0.1572639942,0.3729579151,0.2033801377,-0.0334945098,-0.0882051438,0.0497728027,0.2308515012,-0.3586278558,0.124964945,-0.1097932085,-0.389754653,-0.0506447367,-0.001079825,0.1469896585,0.0434499905,-0.1591925323,0.0015705117,0.1117232442,-0.3314083517,0.1802747697,0.2826015949,0.1433033198,-0.138225913,-0.0929224715,0.0918074995,-0.2511996031,0.2500662208,0.2217899859,0.1579179615,-0.0372050181,-0.1576395929,0.1384074986,0.0796204284,-0.040574912,-0.1741780937,0.0987476259,0.0408518016,0.2424800396,-0.1990905702,0.0407792144,0.4570713639,-0.2844359577,0.036405541,-0.6800735593,-0.0023433422,-0.0597890802,-0.1311700791,-0.0204953104,0.0292914025,-0.0755665004,-0.0433139056,-0.0115973717,0.298815459,0.2784372568,0.0222370271,-0.0605900697,0.1867603958,0.1487032622,-0.1612704545,0.3168522716,0.0585198477,0.1633767188,-0.071909681,-0.2061856836,0.0668226629,-0.095822297,0.3077232838,0.2266746014,0.0122988485,0.0099700829,0.2215041071,0.1509661227,-0.050812643,0.0896444991,0.2920495272,0.1799527705,0.1877627522,-0.3006313741,0.0827777237,0.1254699379,-0.0749200955,-0.1055606827,0.2994289696,-0.1035914943,-0.2416673452,0.0810325369,0.2266719788,0.5017876625,0.1558399349,-0.1444334984,0.0478165336,0.0574757233,-0.2169308364,0.1586061567,0.1800119579,0.3289294541,-0.1385491341,0.0901937708,-0.0619816333,-0.4440478981,-0.0912801176,0.184844926,0.2858141065,-0.1865232587,0.2504453957,0.0615522042,-0.0788814276,-0.0309186876,0.4246840179,-0.0496711768,-0.2127269059,-0.1706473529,0.2725588679,-0.018957546,-0.0283497255,0.0066530071,0.261056304,0.2244727314,-0.02329541,0.0941695124,-0.0079309605,-0.0817143396,0.1298800558,0.1777150482,-0.2123888582,0.4530031681,-0.170789659,-0.0871114284,-0.2084838599,-0.0723169595,0.1136234626,-0.2280815989,0.5252376199,0.1035802439,0.252276659,-0.117324926,-0.1160969213,0.0819555968,-0.0663691983,-0.1613080353,0.0102339908,-0.0259694383,-0.114658013,-0.0101680011,-0.1553606391,0.0795831308,-0.3436994553,0.356094867,0.1022500247,0.1887512654,0.5033752918,0.4098785222,-0.0466410108,0.1624164879,0.1523554623,-0.1796501279,-0.3409413695,0.1021591201,-0.3160607517,-0.33717224,0.0466640629,-0.227466464,0.2909381688,0.4419275224,-0.5405673385,-0.2555706799,-0.512783587,0.178343758,-0.0151668554,0.2652284801,0.090331912,-0.0019471067,-0.1002620608,0.0065976339,-0.1557332873,-0.0046922737,0.0445123166,0.1287520081,0.2151253819,0.59130162,0.0606980622,0.7865701914,0.415009141,0.1420457959,0.3436701,-0.0614233911,0.4706937373,-0.2932085395,-0.1800679713,0.2319211662,-0.1787169725,-0.2574402094,-0.0831607431,0.1046686992,0.0142659806,0.1273812354,0.058672294,0.0460089445,-0.1819885671,0.266802758,-0.0274689402,0.3461062014,0.0467389002,0.0366080366,-0.1479356736,-0.016332021,-0.1037027016,-0.1109503433,0.4852962196,-0.044174578,-0.4239206612,0.1098548472,-0.6938717365,-0.1332408041,-0.0965696424,0.2234214991,-0.0490844324,0.0660194308,0.047589045,0.0453441367,0.9892038703,0.0007064728,-0.2966735065,-0.2548519373,-0.3207207918,-0.4081355631,0.1653787196,0.0722287595,0.1199957579,-0.1644738317,0.1693557203,-0.152690053,-0.203691408,0.2588399649,0.4127705097,-0.1430650353,0.0404216908,-0.2801088095,-0.5489321351,-0.2964856923,-0.1632171273,-0.0055411756,0.0484779589,0.2352022678,0.0255420431,-0.0487965867,-0.1557963192,0.2111781538,0.1084236428,0.2446286976,-0.0259844493,0.4257113636,0.1911683977,-0.3292507827,0.2183281332,0.324498266,-0.0656343922,-0.1088474393,-0.0289757513,0.1022888795,0.2475358695,0.4917110205,-0.2284799367,-0.3292703629,-0.1282335222,0.2469951063,-0.563716054,0.1934460849,0.481580615,-0.2082684487,-0.1923841238,-0.7430677414,0.3641576171,0.3175493181,0.1278441548,0.3630657196,0.0887512565,-0.1641530991,0.3955097795,0.2683007419,0.6815106273,-0.1718036085,0.4881143868,0.3426570296,0.2432327121,0.3742546141,-0.0830155015,0.421284765,-0.0590210147,-0.4138422906,0.0700451657,-0.1448521018,0.1890719831,0.1100216359,0.2438379377,0.3425058722,0.0781698599,0.0153281912,-0.1596763879,0.27529791,-0.3474652171,-0.4178962111,-0.0688188151,0.1925273538,0.0487389416,-0.186029315,-0.0940656811,0.1372859478,-0.0540784113,-0.1073366553,-0.2376541644,-0.1235265583,-0.2773991227,0.2645848095,-0.0288833939,-0.1680644304,-0.1311216056,0.0741543546,0.0400642529,0.1833519787,0.0875317529,0.122982122,0.0256561097,-0.0608532242,0.0861571431,-0.2096137106,0.3166009784,-0.061050687,0.0272021927,0.0231162794,-0.0450544804,0.0166374464,-0.2683945,-0.0405098908,-0.0224609002,-0.0948094875,-0.3335689604,-0.1030184627,-0.2790772021,-0.180031538,0.1073122695,0.1676468551,-0.0732505843,0.3933377564,-0.3213761151,-0.1825398654,-0.1805839539,-0.0805969387,0.2867523432,0.2247370332,0.3443740308,0.1867702156,-0.0413239561,-0.1183303967,0.2398533821,0.2784740031,-0.3960603774,0.3994882405,-0.0061355652,-0.0835081711,-0.1741186976,0.0222124234,-0.0347575769,-0.1690721959,-0.0011354813,0.0251391362,-0.43785429,-0.0114882337,-0.0359180495,0.309658587,-0.0809226409,-0.2384006679,0.0569273122,0.0623049438,-0.2509617209,-0.1806093007,-0.0545278415,0.1300976127,0.1445485055,-0.0649977624,0.2266220301,-0.0639647841,0.0592436232,0.0769756585,-0.2337152064,-0.1937136501,-0.2092723399,0.1047557294,0.1538061649,-0.5640975237,0.315808177,-0.2536366284,-0.2878131568,-0.3615008295,0.2732218504,0.1587805152,-0.0976024568,-0.2814570665,0.0315277614,-0.2384159714,0.1754502356,0.177296564,-0.0986446738,0.4965091646,0.2690390348,0.2676720917,0.2177262306,-0.1677125543,-0.3551409543,0.2323107123,-0.1757250577,-0.2370799631,0.1685317606,-0.145319283,-0.1803852171,0.1270939559,0.1900752634,0.3366607428,-0.1190790012,0.0566112176,0.0406586975,0.157879144,-0.0442557149,-0.1291687638,-0.1578691602,-0.0657628775,0.1783049554,0.4919783473,0.1858622134,0.0189377069,-0.4908386469,0.0165941529,0.2409807891,-0.0848554075,0.1650960594,0.2115793079,0.1125838161,0.0071789902,0.1997704804,0.2766707838,0.2983657718,0.459677577,0.0335596949,0.3375521898,0.1533709168,-0.0437303558,-0.0128277214,-0.3350018263,0.0359271057,0.2339578569,-0.1810237169,0.2591881454,-0.064316988,-0.1826624572,-0.0755789578,-0.2131424546,-0.2734743357,0.1559334695,-0.1004549041,-0.0916266143,0.1778089702,-0.059228424,-0.0837883204,0.4182516932,0.0563772656,-0.6564088464,0.185810551,-0.0279028937,-0.218339175,-0.0040361267,0.2878212035,0.0002884011,0.1268146932,-0.1856125295,-0.056896884,0.1019747183,-0.1687731892,-0.2335278243,0.4316358566,0.5210511684,0.2962177396,-0.0155671164,0.2253416181,0.1285903752,-0.0580609702,-0.2433924228,0.207829833,-0.0938507169,0.3470301032,0.0298524946,0.056428764,-0.119941242,-0.2313876599,0.4667378366,-0.0322324522,-0.5099734068,0.0713702291,-0.270957768,-0.0831899866,-0.0014889478,0.1556717008,-0.3743119538,0.014201696,0.6617762446,-0.3358905613,0.0529741794,-0.1549557745,0.0486678369,-0.4334238172,0.5622742176,0.2479630411,0.0245904196,-0.5293908715,-0.0578941889,-0.4145038426,-0.2513063252,-0.5728726387,0.1208465621,0.215716213,0.2678526938,-0.0817704052,0.1097253338,-0.0991437584,0.1220087558,-0.2385518849,0.286741823,-0.0255693365,-0.2052706033,-0.051268816,0.0605058856,-0.1328707188,-0.4785497189,0.2171453387,-0.3334968984,0.0321596712,-0.1802652776,0.0915698335,0.0828931853,0.1091512591,0.3665707409,0.3001115322,0.0304139275,-0.0899439231,-0.4216742814,-0.4540976882,-0.0191898588,-0.236771971,0.1167336777,0.128349036,0.3148860037,-0.254237175,0.1739491969,-0.2994903326,0.3420858085,0.2490573376,-0.2070638686,-0.0978223085,0.267593652,-0.3416646123,0.1588354558,0.4218654633,0.2471740544,-0.2160965353,0.0314047858,-0.2623040378,-0.3741120398,0.234436065,-0.39076069,-0.3366749585,-0.1116325557,0.0749210343,-0.0133863222,-0.194516629,-0.1902235746,0.1216385514,0.1056280807,-0.1063162237,-0.2567142844,0.2248544991,-0.032845322,-0.2498074025,0.0614074394,0.3020296395,-0.2057998031,-0.3960452676,0.4934438765,-0.1076571792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/737","title":"Trec Dataset Connection Error","comments":"Thanks for reporting.\r\nThat's because the download url has changed. The old url now redirects to the new one but we don't support redirection for downloads.\r\n\r\nI'm opening a PR to update the url","body":"**Datasets Version:**\r\n1.1.2\r\n\r\n**Python Version:**\r\n3.6\/3.7\r\n\r\n\r\n**Code:**\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"trec\")\r\n```\r\n\r\n**Expected behavior:**\r\nDownload Trec dataset and load Dataset object\r\n\r\n**Current Behavior:**\r\nGet a connection error saying it couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label (but the link doesn't seem broken)\r\n\r\n<details>\r\n <summary>Error Logs<\/summary>\r\n \r\n\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/root\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\n---------------------------------------------------------------------------\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-8-66bf1242096e> in <module>()\r\n----> 1 load_dataset(\"trec\")\r\n\r\n10 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n\r\n<\/details>","comment_length":34,"text":"Trec Dataset Connection Error\n**Datasets Version:**\r\n1.1.2\r\n\r\n**Python Version:**\r\n3.6\/3.7\r\n\r\n\r\n**Code:**\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"trec\")\r\n```\r\n\r\n**Expected behavior:**\r\nDownload Trec dataset and load Dataset object\r\n\r\n**Current Behavior:**\r\nGet a connection error saying it couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label (but the link doesn't seem broken)\r\n\r\n<details>\r\n <summary>Error Logs<\/summary>\r\n \r\n\r\nUsing custom data configuration default\r\nDownloading and preparing dataset trec\/default (download: 350.79 KiB, generated: 403.39 KiB, post-processed: Unknown size, total: 754.18 KiB) to \/root\/.cache\/huggingface\/datasets\/trec\/default\/1.1.0\/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7...\r\n---------------------------------------------------------------------------\r\nConnectionError Traceback (most recent call last)\r\n<ipython-input-8-66bf1242096e> in <module>()\r\n----> 1 load_dataset(\"trec\")\r\n\r\n10 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag)\r\n 473 elif response is not None and response.status_code == 404:\r\n 474 raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\n--> 475 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 476 \r\n 477 # Try a second time\r\n\r\nConnectionError: Couldn't reach http:\/\/cogcomp.org\/Data\/QA\/QC\/train_5500.label\r\n\r\n<\/details>\nThanks for reporting.\r\nThat's because the download url has changed. The old url now redirects to the new one but we don't support redirection for downloads.\r\n\r\nI'm opening a PR to update the url","embeddings":[-0.2436395139,0.1074598357,-0.014969049,0.1505164951,0.3800676465,-0.1288013607,0.2745419741,0.3322343528,-0.2046611756,0.1162951291,-0.1912941933,0.2581116557,0.072281599,-0.1658235937,-0.1021717414,0.0497933179,-0.1717509627,0.0062029338,-0.2832172811,0.0877527371,-0.0442226492,0.2602857649,-0.0055203196,0.2427942008,-0.5984306335,-0.1884526461,0.1804316342,0.4515568316,-0.2506002486,-0.5709514618,0.3888585269,0.0727966577,0.12875247,0.5149000287,-0.0001109371,0.056604173,0.2422578931,-0.0557584353,-0.5349962115,-0.4334150553,-0.5455427766,-0.0273083281,0.1403693557,-0.1366654336,-0.2754541934,0.4765296876,0.0609516352,-0.4809705913,0.231019482,0.6710780263,0.2249374539,0.2940627038,0.0680481568,-0.2650927603,0.2843118906,-0.1481320709,-0.1591581553,0.4663556218,0.1364138126,-0.0583354458,-0.0214009155,0.1513233185,0.0624634773,0.1721279323,0.1745289564,0.0565684251,0.1772675514,-0.3218351305,0.2633388937,0.2760293186,0.4948776364,-0.3989416659,-0.2013637871,0.0997701138,-0.0143683497,-0.658031702,0.2096267641,-0.0679786652,-0.1465303004,0.2543896139,-0.3509271145,0.0848670676,-0.3491250277,0.1558761001,-0.2747200131,-0.1051498055,-0.200848341,-0.0068778759,-0.1133388355,-0.0278355852,0.1749705821,0.0200294312,-0.2384963632,0.0650303811,-0.5116367936,0.0560323447,0.0803835317,-0.1478816718,-0.0281517096,0.3335969448,0.0427142084,0.038103085,-0.1636677086,0.2506561577,0.0588728264,0.2197348475,-0.1769096851,0.3202084899,0.2743516266,0.233279407,-0.1748762131,-0.1759629995,-0.1833797991,-0.0463008061,0.1591341943,0.1637717783,0.3319268525,-0.1655625403,-0.5211207867,0.2541392148,-0.1025256291,0.1115969643,0.0097520873,0.4247652888,-0.1067638621,0.0908646733,-0.1643857658,0.0531059578,-0.2054545432,-0.058853671,-0.1864492446,0.019791089,-0.2384144664,0.0471161604,0.3205462098,0.0270043239,0.1940556467,-0.0945624411,0.0846998394,-0.1832986772,0.1642670035,-0.324323386,0.0265539382,0.3521924019,-0.0864366814,0.1659188271,0.2515583634,-0.2422627509,-0.0546777807,-0.0281223208,-0.4812555611,-0.4549756646,-0.0710661337,0.2084080875,0.1548659056,-0.0179841239,0.1349477768,-0.2025086731,0.0613218844,-0.2018948942,-0.0231011249,-0.2384263724,-0.1747210324,-0.2204871029,0.3749778867,0.3168279529,-0.5557841659,0.0769451335,-0.0761044174,0.1106991619,0.1391872168,0.2035178989,-0.1795076132,0.149582848,-0.2587057054,-0.0297243446,0.3721457422,-0.430886507,-0.5118092895,0.0736469775,-0.1449172944,0.1010434702,-0.1864410043,0.163426578,0.2461973429,0.0745620281,0.3918343782,0.0501541793,-0.2675366104,-0.1172212064,-0.2557188869,-0.1203596294,-0.2273943573,-0.0058596572,0.0968563557,0.0362855792,0.2212335765,-0.153466776,0.2883461416,0.1403520256,0.0544888899,0.2114231437,0.5641490221,-0.0762912184,-0.2149109542,-0.0440055393,-0.2572885752,0.274630338,0.0434309617,0.1699548513,-0.3841554821,-0.0731172785,-0.5478708744,0.0116442656,-0.1969336122,-0.2449869663,0.1305416524,0.2496086359,0.1088530198,0.1462856084,-0.1167913079,0.4556647539,0.0960789695,0.104242757,0.0182092991,0.249475345,-0.0983027369,-0.0271586217,-0.1154528931,0.2011945695,0.1797367781,-0.1083229259,-0.2002046257,0.4768730998,0.2418258041,-0.0474010147,-0.094438456,0.1279828101,0.2421502769,-0.1477313191,0.0017794613,-0.0935001001,0.1546034962,0.0253841225,-0.0430585183,0.3091504574,-0.0131638292,0.1504990309,0.1926895827,-0.0952668861,0.2412174493,0.0416728891,-0.0225719586,0.0657677054,0.018250579,0.2073457837,0.1549005806,-0.1155516282,-0.2918310463,0.1244803667,0.342433244,0.0697862878,-0.0812748522,0.0232161861,-0.2145894468,0.0729293078,0.1710629165,0.2937608063,0.4095773995,0.1543966532,0.0596784391,0.123982057,-0.2317830026,-0.1342634559,-0.0668331683,0.0030505019,0.2552390993,0.3379247785,0.2496048361,0.0391560681,-0.0740778521,-0.3382083774,-0.0782374963,0.2021334469,-0.3116461337,0.2916006148,-0.280492276,-0.2236142904,-0.0499674231,-0.4288078547,-0.0789081827,-0.3910887241,-0.104875356,0.4562005103,0.0663824975,0.1978205442,-0.4925712347,0.0761058256,0.1892468929,-0.4692899883,0.0011927318,-0.1383705437,-0.314391762,0.1267779171,0.0829908624,-0.1739951372,0.251747787,-0.2204298377,0.0104240412,-0.2139101774,-0.2533704042,0.0877238587,0.1492300034,0.1897948533,0.0066279345,0.0844305828,-0.1221543923,-0.0734417811,0.3308883309,0.0803055614,0.1345987469,0.0250462424,-0.0844899788,-0.0403569527,-0.028842343,-0.5540796518,-0.4347090423,-0.4233095646,0.0265906602,0.0278892498,0.2145307213,0.3089239299,-0.0137016736,0.1361341476,-0.1069938466,-0.0539860278,-0.13827461,-0.4636726379,0.2503553629,-0.3218648434,-0.3025492132,0.1655755937,-0.0695523843,0.5321611762,0.0086490298,-0.4650232196,-0.2676859498,-0.0982105434,0.6345692277,0.0915184766,-0.2353428453,0.2610230744,-0.0646533519,-0.1607629508,0.0247910656,-0.1128008366,-0.0059721055,0.0834122896,0.3514324427,-0.0541349128,0.4497286975,-0.0431193672,0.5627840757,0.1353463531,-0.1065158844,0.4253454208,-0.3066239357,0.5101931095,-0.2793849409,-0.2797890902,-0.1413617134,-0.2360912561,-0.2632610202,0.0153983226,0.2132201046,0.0382813029,-0.3137512207,0.3839368522,-0.0350420438,-0.3948903978,-0.0420298465,-0.0412913822,0.2510329187,0.1905149668,0.1083178967,-0.0192351062,-0.1726915389,-0.243965745,0.3814210296,-0.0949059948,0.0308566391,-0.06078697,0.2049268335,-0.6307371259,0.1342715621,0.1163181737,0.439796567,-0.294167757,-0.1318408847,-0.0849412903,0.080196768,0.2074448317,-0.0995101333,0.0841597393,0.426617533,0.0836059377,-0.1135634333,-0.1509215385,0.0033232879,0.0818037689,0.6349381804,0.434804827,0.074315086,0.0479372405,0.2320050597,0.0532673858,0.0615116842,0.0781043246,-0.2655322552,-0.2171864808,-0.1533264667,-0.0631938353,-0.0232888348,0.3115465939,-0.0437206216,-0.1754530817,0.0472505465,-0.1432409137,0.1013275087,0.1371771842,0.0246305484,-0.0246402603,0.3024009764,0.0889823362,0.2391394973,0.2632663846,0.6313427091,-0.1889279336,-0.2625210881,0.1411161423,0.0982744619,0.0219226442,0.2090880573,-0.0440428965,-0.0713758543,-0.1043322012,0.1706364006,0.007620662,0.4123369157,0.1464786083,0.0978743583,-0.6797946095,-0.1825641394,0.0809005797,-0.1793085784,-0.0737979189,0.031826932,-0.0273168031,-0.1100971624,-0.1013319269,-0.3407725394,0.7502122521,0.1051542088,0.1461166143,0.2128794789,-0.1000243351,0.3703199327,-0.6265904307,0.2992111444,-0.3391892314,-0.1779634506,-0.1715645343,-0.064457491,0.042701982,-0.0872035399,-0.058789365,0.3925181925,-0.0232675113,0.1652839631,0.0695065409,-0.048204314,-0.0568276756,0.1105865687,-0.0691036284,0.2089674175,0.074148342,0.3698207438,-0.0456227958,-0.0529730171,-0.1782884002,-0.1797784567,-0.2008440346,0.1354667991,-0.6049464345,0.1428332478,-0.026967017,-0.2785516977,-0.080890812,-0.1312805861,-0.014492196,0.181736514,-0.1985465288,0.2753667533,-0.0028113436,0.0133361137,0.1627682894,-0.0252936315,0.0766094178,-0.0047154301,-0.1728844792,0.4410365522,-0.0212819371,-0.1607270539,0.2650497258,0.0032535377,0.4391771257,-0.2914651334,-0.0309472047,-0.1398818046,-0.0298233628,-0.1710261703,0.1774249971,-0.1339993179,0.1550931185,0.0815624893,0.0286250301,-0.3594526649,-0.0321366563,0.4235404432,-0.3132712543,0.0055291303,0.300578773,0.2023658007,-0.144022122,-0.1973550767,0.1408713162,0.1862573624,-0.5022172928,-0.0584826097,0.1394028664,0.1922177374,0.0379430838,0.2046674341,0.2547497153,-0.0284734685,0.0760834292,-0.3847139478,-0.3824962676,-0.0481147245,0.0224588979,0.2585418224,0.0243497156,0.0337019935,-0.0517693125,-0.0630752221,-0.3346201479,-0.0273290835,-0.2622383237,0.012638771,0.320886761,-0.0575916469,0.5082917213,-0.0331908129,0.1827071607,-0.0389721729,-0.3674542904,-0.2421674281,-0.0671875626,0.1398568153,0.1127960086,-0.5299235582,0.0158395804,-0.0475150235,0.0082087805,-0.0270215254,0.1269881576,0.1069453359,-0.1901586801,-0.0041024163,-0.4619935453,0.1204537079,-0.0626276284,0.1747385561,0.2353819162,0.4555337131,0.0081379106,0.2163468003,-0.1688103676,-0.064915888,-0.2859056294,0.0310637131,-0.0958550572,-0.2007585615,0.3629794121,-0.2602885962,0.0281877089,-0.0610699244,0.2967252135,0.0044664717,-0.1540704668,-0.1034802869,0.0677492991,0.3193880916,-0.1193158403,-0.045767419,-0.0623318665,0.0754420832,0.2138039917,-0.0005392199,0.0787804425,0.0369653367,0.0524556078,0.1924662441,0.4129530191,-0.2597540021,0.1540228724,0.2356329113,0.0374271721,0.1041542664,0.3256738186,-0.0771511272,0.0053670653,0.498523891,-0.3577448428,0.1061747223,0.1248714328,0.0072757564,-0.033147309,-0.3750703037,0.0520821698,-0.0397391953,-0.3141109943,-0.1003410742,-0.0710787401,0.2068194598,-0.1854576021,0.1388015449,-0.2510991395,0.0079707177,-0.1095806956,-0.0449367352,0.1683595628,-0.4044199884,-0.1095409393,-0.0832012445,-0.1020624638,0.0784510598,0.0795240551,0.1483847201,-0.0089100869,-0.4815813005,0.0227165725,0.1650723219,0.0516879968,-0.1169057637,0.4407913089,0.2734145522,-0.0331330001,0.0902165025,0.1800569296,0.5992760062,0.327675432,0.233602643,0.1383290142,-0.098086305,-0.112054266,0.1027596965,0.0545509756,0.1344367862,-0.226527676,0.189315334,0.2322294712,-0.1436278373,0.2335052043,-0.0963100791,0.0910944492,-0.1523352116,0.4199562371,-0.0260075647,-0.1321964264,-0.184330076,0.1535143703,-0.2354542166,0.2184886485,0.5031722188,0.0140184797,0.2912636101,-0.0805422664,0.0961987972,-0.2021582872,0.2815181911,0.2193568945,0.1679267883,-0.4231604338,-0.0806224346,-0.3168094754,0.1508148611,-0.0500686169,-0.2932548821,-0.1034139991,0.104993917,-0.2948221266,0.2415295243,0.2013257444,0.0316748098,0.1292237937,0.162303254,-0.2166964412,-0.0907827243,-0.2303544879,-0.0631657243,-0.0273146182,-0.5374889374,0.0025256944,-0.1326077133,0.0612035953,-0.1515268236,-0.0150005994,0.1695074588,0.0840091929,0.3490768075,0.1543771774,0.6344005466,0.1395266652,-0.0493563786,-0.1607923508,-0.3679970205,-0.0795508921,-0.018755978,0.2487581223,0.3183907568,0.1491256803,-0.4152117074,-0.378049165,0.5051019192,0.0849959701,0.0548370704,0.1152707115,0.0360174328,-0.1020010784,0.1671333909,-0.0482431538,0.0784009174,0.171269998,0.3160137832,-0.1110542789,-0.4060392082,0.4514369071,-0.3892627358,-0.0561610423,-0.2896740437,0.4707393646,0.1972310841,-0.1055885404,-0.373465687,0.158931151,0.4805704951,-0.1257908046,-0.3419944644,0.3100984693,-0.1594555974,0.1204138845,0.0147893606,0.6515817642,-0.0898189619,-0.2755630016,-0.156762287,-0.1716499627]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/735","title":"Throw error when an unexpected key is used in data_files","comments":"Thanks for reporting !\r\nWe'll add support for other keys","body":"I have found that only \"train\", \"validation\" and \"test\" are valid keys in the `data_files` argument. When you use any other ones, those attached files are silently ignored - leading to unexpected behaviour for the users.\r\n\r\nSo the following, unintuitively, returns only one key (namely `train`).\r\n\r\n```python\r\ndatasets = load_dataset(\"text\", data_files={\"train\": train_f, \"valid\": valid_f})\r\nprint(datasets.keys())\r\n# dict_keys(['train'])\r\n```\r\n\r\nwhereas using `validation` instead, does return the expected result:\r\n\r\n```python\r\ndatasets = load_dataset(\"text\", data_files={\"train\": train_f, \"validation\": valid_f})\r\nprint(datasets.keys())\r\n# dict_keys(['train', 'validation'])\r\n```\r\n\r\nI would like to see more freedom in which keys one can use, but if that is not possible at least an error should be thrown when using an unexpected key.","comment_length":10,"text":"Throw error when an unexpected key is used in data_files\nI have found that only \"train\", \"validation\" and \"test\" are valid keys in the `data_files` argument. When you use any other ones, those attached files are silently ignored - leading to unexpected behaviour for the users.\r\n\r\nSo the following, unintuitively, returns only one key (namely `train`).\r\n\r\n```python\r\ndatasets = load_dataset(\"text\", data_files={\"train\": train_f, \"valid\": valid_f})\r\nprint(datasets.keys())\r\n# dict_keys(['train'])\r\n```\r\n\r\nwhereas using `validation` instead, does return the expected result:\r\n\r\n```python\r\ndatasets = load_dataset(\"text\", data_files={\"train\": train_f, \"validation\": valid_f})\r\nprint(datasets.keys())\r\n# dict_keys(['train', 'validation'])\r\n```\r\n\r\nI would like to see more freedom in which keys one can use, but if that is not possible at least an error should be thrown when using an unexpected key.\nThanks for reporting !\r\nWe'll add support for other keys","embeddings":[0.316801697,-0.2654780149,-0.1059445664,0.159991473,0.1830630451,0.0241956171,0.1906820387,0.2526224554,0.3097397685,0.1597378701,0.2905041575,0.3881573975,-0.3354470432,0.1538857073,-0.288726747,0.1160684675,0.125699982,-0.1590540707,0.1830485165,-0.0739719048,-0.3281506598,0.1299612671,-0.1742626131,0.0519801937,-0.2565570772,-0.0325198509,-0.0652077645,-0.1249923557,-0.0559795089,-0.7033967376,0.2441999912,0.1866964698,-0.1685740352,0.3661860526,-0.0001206838,0.0227669124,0.2600605786,-0.0612923354,-0.4180273712,-0.284078449,-0.1266269088,-0.2503322959,0.2157292217,-0.2927863896,-0.0095770406,-0.5335063934,-0.1659465879,-0.2136441767,0.428491503,0.4755265415,0.1441039294,-0.0281475242,0.1074903011,0.103443943,0.1855832487,0.173317939,0.0523978397,0.1129865125,0.2005833089,-0.1863103062,0.1974927783,-0.015203868,-0.0855333209,-0.1166258156,0.4169049263,0.014463705,0.0431937128,-0.5853435397,-0.0673373938,0.4680506885,0.4733634889,-0.2158763111,-0.4072593153,-0.4962639809,0.0168514624,0.1437435001,0.1239900962,0.0262904894,-0.2754368186,0.1716492921,-0.0186458454,-0.1279143393,-0.3145546019,0.3645599484,0.1588759869,0.1667305082,0.1998009682,0.1751462072,0.4177148342,0.1041818261,-0.2417998314,-0.2961252928,-0.1687681079,0.0153912129,-0.1717357039,-0.2395186126,0.0096005592,-0.1021072492,0.3478270769,-0.1615573317,0.0814781561,0.0147552509,0.0688540414,0.1160583869,0.3295890391,0.0167197008,0.2521263957,0.27151829,0.4053871036,-0.1177208647,-0.0782537982,-0.1062863916,0.4558876753,-0.3214888275,0.2844702899,0.3004194498,0.2107571363,0.0294454712,-0.0727447644,0.2422875166,-0.1851319224,-0.2118499577,0.2068524063,0.3518288136,0.0572953559,0.1997139901,0.0050809565,0.4353574812,0.1753629297,-0.0635584444,-0.0750683323,0.2083379626,0.0066031022,-0.1846792549,0.2966028154,-0.195021376,0.072069563,0.4366956651,0.219838351,-0.2566190362,0.0217719562,-0.1647385806,0.3601348102,0.2705901861,-0.3926011622,0.1859975457,0.0535216779,-0.318626076,-0.1505866349,0.1927851439,-0.4798439443,-0.3855746984,-0.0199395679,0.148662895,-0.173707813,0.1109164804,-0.3096131384,0.2625356615,-0.1487076581,-0.0787951946,0.1654056013,-0.1909547448,0.0015476707,-0.2966027856,-0.0613504164,0.2132754028,-0.1867026389,0.0856343806,0.0245898403,-0.1329401731,-0.1904079914,0.0431639999,-0.5409638286,0.5332465172,-0.2760227621,0.0758310109,0.3136712909,-0.6196815372,-0.4739336371,0.2544255555,0.2400846183,0.0542289019,0.3313697278,0.0146613438,0.1378411204,0.0395525061,0.3327175677,0.1336993128,0.2205701172,-0.0022337136,-0.0966703147,-0.1880936921,0.3692169189,-0.0734605268,0.023219429,0.1141415462,0.2966578901,0.0823726878,0.2609554827,-0.3047966063,-0.3681036532,0.1661448479,0.2164793015,-0.043949116,0.0608692579,0.0930199251,-0.4039118886,0.3568021953,0.1617289037,-0.2016153783,-0.3329255283,-0.3264976144,-0.1857260913,0.0251686033,-0.4044780731,0.1097436175,0.1087307557,-0.2512439787,0.206527099,0.2972320914,-0.2340177298,0.1557913274,-0.0337294862,0.2250504047,-0.0219865367,-0.1630951464,-0.0199313499,0.2315615863,-0.1368559003,0.019049542,0.2613013983,-0.150744617,0.0013917639,0.4357238114,0.4243524671,-0.1707628965,-0.2887263596,0.2721306086,0.2272150367,0.0273606759,-0.2132174522,0.3163019419,0.1341622323,-0.0600632131,0.0794736743,0.6082780957,-0.1413636506,0.1784777492,-0.1875850856,-0.0560442582,-0.0752283707,-0.0145493615,-0.3586849272,-0.2051805556,0.1175777093,-0.2243866175,0.007542891,-0.1712309271,-0.103690818,-0.3479410112,0.4102439284,-0.0472923964,0.2244063765,0.1516033113,0.1955777258,0.1432718635,0.0237270296,0.4587876201,0.4947429001,0.1478353888,-0.1695305109,0.0823711157,-0.0985467806,-0.2065082788,0.0551128052,0.0971310809,-0.3650157452,0.3951963484,0.122982353,0.0379791595,-0.2471842319,-0.050597202,0.0124774827,0.1871052086,-0.3328945041,0.2572689354,-0.4460756481,-0.1765044183,0.0867044032,-0.1396218389,-0.0752752945,-0.3988204598,0.2380988002,-0.0556424409,-0.158140853,-0.1837353557,-0.2249031216,-0.052273158,-0.2386897951,-0.2873674631,-0.1737000644,-0.1168125123,-0.0246256199,-0.0382243991,0.2852021158,-0.222488746,0.357503742,0.0925312415,-0.2790093422,0.0549247675,-0.2404105067,-0.0759595558,-0.0169443581,0.1494110227,0.2026516497,0.2256762385,0.3076711893,-0.3808083832,0.1275920719,-0.0765158162,-0.1585383862,0.4020903409,0.1625909805,0.1447659582,-0.2372924238,-0.3876551688,-0.2504473627,-0.080073379,0.0585017875,-0.1418123096,0.0744012222,0.3585133255,0.3497640789,0.0151081467,-0.0751057491,0.3658713698,-0.0952299908,0.0383574888,0.4012955427,0.0997118577,-0.3444525599,0.0867446214,-0.1080586985,-0.2782904208,0.0616066717,-0.4604220688,-0.5260500908,-0.1220487133,0.2159826756,-0.1367788613,0.0903569907,-0.0118063251,0.1144048944,0.0603317097,0.0076171821,0.0693806484,-0.069802314,0.1368387938,0.1260422766,0.1830426306,0.24760665,0.2479145974,0.377846688,0.1189415082,0.0415818281,0.33198753,-0.1227000207,0.3132101893,-0.0703658685,-0.2405774891,-0.1832890362,0.0242551025,0.0448255017,-0.0764272958,-0.1696386188,0.3032894433,-0.045164939,0.0614811182,-0.2494579107,-0.3481260836,0.2374229133,-0.5032522678,-0.1085943282,-0.02013845,0.3095923066,0.1701372266,-0.0433075428,0.0347894654,0.2693816721,-0.0054581426,0.0079274168,-0.2242091745,-0.540913403,-0.4010589719,0.3015395403,-0.1871969402,0.3280073404,-0.0276482999,0.0671183243,0.1418225765,0.1057135016,0.7394579649,-0.3708121777,-0.2351822704,-0.1151553541,-0.1043522432,-0.176985234,-0.2543654144,-0.1803963929,0.2238178253,-0.3039506078,0.5197979808,-0.4486809969,-0.0724492222,-0.3375962973,0.4328943491,-0.1186673641,-0.2142701894,-0.2181779742,-0.2280169874,-0.6136746407,0.0906747952,0.2800222635,-0.0063888957,-0.3038771451,0.0755136162,-0.0514055379,0.0802263021,0.2476643026,0.2924230099,0.2296675742,-0.2909874916,-0.0418380015,0.1697078049,-0.1283289939,0.4376193583,0.3582804203,-0.103136383,-0.4273714721,-0.4494740069,-0.216669023,0.2137969881,0.1823346764,-0.1980460733,0.3466600478,-0.1286956072,-0.2928498685,-0.0560928881,0.0661616474,0.6232289672,-0.1975838989,-0.3550898135,-0.492277652,0.1085235849,0.0375762656,-0.3533420563,0.1429382563,0.1257270277,-0.4656327367,0.6197372675,-0.0357916951,0.8890610933,-0.0563272163,0.2924438119,-0.1200383082,-0.054693304,0.5364705324,-0.1870442331,-0.0003774582,-0.2019851059,-0.0878447965,-0.1401995718,-0.0576836914,0.2362201065,0.1754661947,-0.1895309091,0.3187419474,-0.3747239709,0.080651775,-0.38819471,0.4054502845,0.0213492662,0.1384967864,-0.2590313852,0.0468513332,-0.040299125,0.1357626915,-0.1278158873,-0.0638511702,-0.0407986008,-0.7564883232,-0.2518976033,-0.4478384256,-0.2224171758,0.1853726506,0.178618148,-0.4904553592,0.1751361787,0.9425093532,0.2816868722,0.2434365302,0.0368765816,-0.0752516687,-0.0491545759,0.1594684869,-0.3204378486,-0.2523343265,0.0967485532,0.0960732773,-0.35826841,0.2794502079,-0.0934548005,-0.4843669236,-0.1092201546,0.224901408,-0.1836192161,-0.3045082688,-0.1533350348,-0.3470221162,-0.0440563597,-0.0618835017,0.0789851099,0.2746475637,-0.1786049604,0.4952195287,-0.0355433002,-0.233111605,-0.0529856794,0.1000684127,-0.1401808858,-0.0239010639,0.5580211282,-0.2631718516,-0.0060901265,-0.2033471763,0.2132144868,0.0729472414,-0.0281335991,0.1582365334,0.2658636272,0.1028251946,-0.017732786,0.2799800336,0.2504447401,0.0614958741,0.082097508,-0.3686722517,-0.2333796769,0.223278597,0.1594523787,0.3823647201,-0.0945816264,-0.2262554467,-0.1755571961,0.0547526404,-0.2186206281,-0.0043829312,-0.0208428986,-0.1360767037,0.0259730034,0.197192207,-0.0331522264,-0.0582246482,0.0395488404,-0.0387759693,-0.0343009792,-0.1245293766,-0.2114706486,0.1554586887,0.127065286,0.2961866558,-0.2051555663,-0.4633936584,0.179374963,-0.1252330393,0.1445632428,0.3797843456,-0.0325997993,0.7761133313,0.1800706238,0.3035766184,-0.0005575059,0.0001976158,-0.1242279857,0.25644961,0.361599952,0.2067274153,0.0920269787,0.0749769658,-0.3671783209,0.2010153234,-0.1292878538,0.0401251875,0.1179736927,0.0312240366,0.1884763092,0.113078326,0.0931501463,0.5713256598,-0.4997255206,0.1459197104,0.285589993,0.0999429673,-0.1062477008,-0.5389022231,0.2994879186,0.1093986556,-0.0123101175,0.2852394283,0.2920354605,-0.2698893547,0.0249548964,-0.1093523577,0.0066219363,-0.0632150024,0.1603449285,0.802572608,-0.1439393014,-0.1153720096,0.198474139,0.0631730035,0.1533917189,0.4139851034,-0.2399118543,0.1607054323,0.1065083295,-0.1746343076,0.1504771113,-0.1964418292,0.5519580841,-0.0189920105,-0.1386803836,-0.0161232501,0.219086051,0.1060645655,-0.104502745,0.0166263785,-0.3090135753,-0.2974226773,-0.0313752033,-0.1498438418,-0.0491711199,-0.1686347574,-0.0217822324,-0.0622665361,-0.1330319494,-0.219183445,0.0366337486,0.1761519164,-0.0783657283,-0.074264586,0.3176168203,0.0668309107,0.4520940781,-0.1790332943,-0.2803371847,0.2196555138,0.0278560873,0.6177238226,-0.2754909992,0.4259009063,0.1500042379,-0.2373513132,0.1693226993,-0.2940306067,0.015603235,-0.0602523237,0.4385226071,0.1985034645,0.257815361,0.2947139144,0.0853877142,-0.0364014618,0.1011472717,-0.0332877636,0.2056467384,-0.3567962348,0.2917584777,0.1165286005,-0.3315205276,-0.0348504484,0.1958280057,-0.0505585037,-0.2955284417,0.1781949997,0.5037840605,0.2918821573,-0.1816748828,0.0336296149,0.1424205303,0.3772466183,0.2152693421,0.1113035083,-0.204020679,0.2999002337,-0.723395884,-0.1216728613,-0.2340506762,0.0723701939,0.1547695398,-0.1319907308,0.1139990613,0.1974890381,0.1747167856,-0.3036686182,-0.2114664167,0.3931582272,0.0087357992,-0.177488476,0.1569247246,0.0477939248,-0.0651287138,-0.0781766325,0.2374575287,-0.0448867381,0.0295173135,-0.2535606921,0.1244048551,0.0873044953,0.1525101662,0.3869750798,0.3887454271,0.564969182,-0.0199309066,0.0986781269,0.0476350822,0.0947995931,-0.0860051289,0.0813600942,-0.0006985722,0.1280229688,-0.2033862174,-0.253823638,-0.1101024076,0.1801722646,0.0578602664,-0.4248856902,-0.2771167159,0.1014536172,0.000554908,-0.0377875827,-0.0545937642,0.3721119463,-0.0957153961,-0.0594441965,-0.000119928,-0.2207419872,0.2309884131,-0.1772340536,-0.279220432,-0.012768928,0.2935386896,-0.1618194431,0.0712580159,-0.3407915533,-0.036695566,0.3377033472,-0.2515605986,-0.3088917136,0.2081037909,-0.0752706975,0.0838494003,-0.0034179001,0.0867788419,-0.0935960487,0.0006317665,0.0479247943,-0.3399060071]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/730","title":"Possible caching bug","comments":"Thanks for reporting. That's a bug indeed.\r\nApparently only the `data_files` parameter is taken into account right now in `DatasetBuilder._create_builder_config` but it should also be the case for `config_kwargs` (or at least the instantiated `builder_config`)","body":"The following code with `test1.txt` containing just \"\ud83e\udd17\ud83e\udd17\ud83e\udd17\":\r\n```\r\ndataset = datasets.load_dataset('text', data_files=['test1.txt'], split=\"train\", encoding=\"latin_1\")\r\nprint(dataset[0])\r\ndataset = datasets.load_dataset('text', data_files=['test1.txt'], split=\"train\", encoding=\"utf-8\")\r\nprint(dataset[0])\r\n``` \r\nproduces this output:\r\n```\r\nDownloading and preparing dataset text\/default-15600e4d83254059 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155...\r\nDataset text downloaded and prepared to \/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155. Subsequent calls will reuse this data.\r\n{'text': '\u00f0\\x9f\u00a4\\x97\u00f0\\x9f\u00a4\\x97\u00f0\\x9f\u00a4\\x97'}\r\nUsing custom data configuration default\r\nReusing dataset text (\/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155)\r\n{'text': '\u00f0\\x9f\u00a4\\x97\u00f0\\x9f\u00a4\\x97\u00f0\\x9f\u00a4\\x97'}\r\n```\r\nJust changing the order (and deleting the temp files):\r\n```\r\ndataset = datasets.load_dataset('text', data_files=['test1.txt'], split=\"train\", encoding=\"utf-8\")\r\nprint(dataset[0])\r\ndataset = datasets.load_dataset('text', data_files=['test1.txt'], split=\"train\", encoding=\"latin_1\")\r\nprint(dataset[0])\r\n```\r\nproduces this:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset text\/default-15600e4d83254059 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155...\r\nDataset text downloaded and prepared to \/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155. Subsequent calls will reuse this data.\r\n{'text': '\ud83e\udd17\ud83e\udd17\ud83e\udd17'}\r\nUsing custom data configuration default\r\nReusing dataset text (\/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155)\r\n{'text': '\ud83e\udd17\ud83e\udd17\ud83e\udd17'}\r\n```\r\n\r\nIs it intended that the cache path does not depend on the config entries?\r\n\r\ntested with datasets==1.1.2 and python==3.8.5","comment_length":35,"text":"Possible caching bug\nThe following code with `test1.txt` containing just \"\ud83e\udd17\ud83e\udd17\ud83e\udd17\":\r\n```\r\ndataset = datasets.load_dataset('text', data_files=['test1.txt'], split=\"train\", encoding=\"latin_1\")\r\nprint(dataset[0])\r\ndataset = datasets.load_dataset('text', data_files=['test1.txt'], split=\"train\", encoding=\"utf-8\")\r\nprint(dataset[0])\r\n``` \r\nproduces this output:\r\n```\r\nDownloading and preparing dataset text\/default-15600e4d83254059 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155...\r\nDataset text downloaded and prepared to \/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155. Subsequent calls will reuse this data.\r\n{'text': '\u00f0\\x9f\u00a4\\x97\u00f0\\x9f\u00a4\\x97\u00f0\\x9f\u00a4\\x97'}\r\nUsing custom data configuration default\r\nReusing dataset text (\/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155)\r\n{'text': '\u00f0\\x9f\u00a4\\x97\u00f0\\x9f\u00a4\\x97\u00f0\\x9f\u00a4\\x97'}\r\n```\r\nJust changing the order (and deleting the temp files):\r\n```\r\ndataset = datasets.load_dataset('text', data_files=['test1.txt'], split=\"train\", encoding=\"utf-8\")\r\nprint(dataset[0])\r\ndataset = datasets.load_dataset('text', data_files=['test1.txt'], split=\"train\", encoding=\"latin_1\")\r\nprint(dataset[0])\r\n```\r\nproduces this:\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset text\/default-15600e4d83254059 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155...\r\nDataset text downloaded and prepared to \/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155. Subsequent calls will reuse this data.\r\n{'text': '\ud83e\udd17\ud83e\udd17\ud83e\udd17'}\r\nUsing custom data configuration default\r\nReusing dataset text (\/home\/arne\/.cache\/huggingface\/datasets\/text\/default-15600e4d83254059\/0.0.0\/52cefbb2b82b015d4253f1aeb1e6ee5591124a6491e834acfe1751f765925155)\r\n{'text': '\ud83e\udd17\ud83e\udd17\ud83e\udd17'}\r\n```\r\n\r\nIs it intended that the cache path does not depend on the config entries?\r\n\r\ntested with datasets==1.1.2 and python==3.8.5\nThanks for reporting. That's a bug indeed.\r\nApparently only the `data_files` parameter is taken into account right now in `DatasetBuilder._create_builder_config` but it should also be the case for `config_kwargs` (or at least the instantiated `builder_config`)","embeddings":[0.0632612184,-0.1689583212,-0.1356959939,0.3066903055,0.3663588464,-0.0822662711,0.2109704763,0.1932276636,-0.0809329078,0.1459649205,-0.0522681624,0.0610482432,0.1775195897,-0.0559846647,0.1316590309,0.2184448987,0.0873197839,0.06633313,-0.1263160855,-0.210990116,-0.0102705723,0.2062479109,0.0173396915,-0.1451737732,-0.4371198714,0.1228281334,-0.0854631141,0.0826427341,0.0433101729,-0.2312449813,0.0753922686,0.1557570398,-0.1427129656,0.3933540881,-0.0001079512,-0.0463827141,0.0049218708,-0.0751267746,-0.4451858997,-0.02410217,0.0668388307,0.1159065664,-0.0537758544,-0.1792441458,-0.0638852715,0.089059785,0.0805219784,-0.6234307289,0.4211957753,0.3092533946,0.2928189635,0.0181103088,-0.3509517014,0.1025870889,0.2549996674,0.2605431676,-0.0601470061,0.0162583049,0.2008947283,0.1844874471,-0.1797304899,0.4905602038,0.0224634875,-0.0929092988,0.0126927979,0.3502127528,-0.1973470896,0.1078346819,0.4176999927,0.2047117203,0.4102971852,-0.4855338931,-0.2600766122,-0.3530350327,-0.1391108781,-0.3968032897,0.3521828651,-0.0104581323,-0.0381049551,0.2515204251,-0.2373825461,-0.0499013253,0.0433750972,-0.0378617831,-0.1471813321,0.0639895871,-0.2387994826,-0.1049143374,-0.0565128364,-0.0936133042,0.0106648346,-0.2465815395,-0.0519431904,0.1307703704,-0.0817520916,0.1496287733,-0.0354530141,0.0141138528,0.086954616,0.0944403037,0.2200161815,0.1656225473,-0.0772432536,-0.1178843901,-0.0753500983,0.4089231193,0.1265441924,-0.1236865297,0.1996533424,0.1021118313,-0.2863745391,-0.1559411585,0.0823156089,-0.066374436,0.3497540951,-0.2117480785,0.1754067391,-0.1227143034,-0.1732728779,0.1200615466,-0.013754081,-0.0166406445,-0.0045666327,0.2791540623,-0.2176062614,0.0804498643,-0.1096234173,0.0016868971,-0.2030216008,-0.2661352158,-0.4274788797,-0.1317585856,-0.1823942065,0.1311712116,0.1808477938,-0.0524299927,0.1408107132,0.1967484951,0.1277533025,-0.3219616711,-0.1136642098,-0.2299860269,0.2369440645,-0.0064800219,-0.2771321237,0.2032752931,0.116921939,-0.2279536873,-0.0955291018,0.0284312069,-0.2402467728,-0.1247919276,0.0569322817,0.2919853628,-0.2366798371,-0.1262016296,0.0572552234,0.0542957149,0.5505169034,0.0660656169,0.1900551021,0.013956055,-0.4436228573,-0.1973882914,0.0649240911,0.3066074252,-0.2653443515,-0.2186737359,0.0841267109,0.0026896682,0.221651718,0.3893524706,-0.1010949761,0.1029098183,-0.3480854332,0.3724368811,0.1862321049,-0.3808503151,-0.5444986224,0.2635792196,-0.0366775505,0.3307319283,0.219825834,0.0197251141,0.1979728937,0.0097494535,0.2746679783,0.1109910756,0.3599116802,-0.053945411,-0.2629933953,-0.1501543969,-0.0911056921,-0.0414852612,-0.1381851882,0.0672327429,0.0885716751,-0.1138967425,0.3438817859,-0.1182795167,0.2481808066,0.1298857033,0.2086013556,0.1743334979,0.1099112704,-0.0403509289,-0.3164944351,0.1880825013,-0.0852982849,-0.0295293257,-0.3783096373,-0.3485529125,-0.1945723444,-0.0583882965,-0.2409086525,-0.1717122793,0.2385733128,0.4372510016,0.1811357439,0.2541648746,-0.0017992576,0.4922895133,-0.2259601504,0.0430978574,-0.4107952118,-0.063932918,0.2076205909,-0.1066015288,-0.3256843686,0.229467988,-0.0379208438,-0.0506643765,-0.0245662201,0.3689850867,-0.0090350537,0.5451434851,-0.1396271586,0.0909882411,0.0636203513,0.0560450368,0.0607162118,0.1490697861,0.2253748775,-0.2123257369,-0.2007148415,0.4290677011,0.1339482516,0.0944108367,0.1008229703,-0.3489763141,0.1937699467,-0.1465731263,-0.0214689467,-0.4593221247,0.3228978813,-0.186393261,0.3557103276,0.3392952085,-0.4177323878,0.154691875,0.648634553,-0.0221168213,0.0970359817,0.1618766338,-0.0894946456,-0.1046929806,-0.2154047042,0.3116965592,0.4378974736,0.1979781389,0.3201467097,0.217566669,0.1970346272,-0.4783408642,0.2719818056,-0.2613622844,-0.0497345775,0.3116179109,0.2337706387,-0.1056016609,-0.5463246107,0.1784432381,-0.0284870882,0.0335313939,-0.2530533373,0.1925891489,-0.48399809,-0.3050448298,-0.4533791244,-0.0004835404,-0.190690279,-0.2545401752,0.0010062737,0.0512658879,0.0556201078,0.0102702938,-0.1824670732,0.1323817372,-0.1237922907,-0.376193136,-0.2686216831,-0.0481047779,-0.3546080589,0.1346622258,-0.0071911663,-0.1643130779,0.2114718705,-0.2651865184,-0.0801099315,-0.1270552427,-0.2446544021,0.1423727125,-0.0068997727,0.0158227626,0.0206633061,-0.1169913411,-0.0731749088,0.0459409766,0.2070356011,-0.1467127651,-0.2457351685,-0.1974944472,0.029874634,0.0841844156,-0.3097217977,-0.3685527146,-0.1084531918,-0.1123129427,0.308909893,-0.092616342,0.1283420622,0.3613407314,-0.1768619269,0.2316176295,-0.0665881783,0.2496745884,-0.5686767101,-0.4457151592,0.3216125071,-0.0838067159,-0.2544903457,-0.0816360563,0.0415059552,0.2657451332,0.2059588134,-0.7856476903,-0.1250162423,-0.3293462992,0.0665295124,-0.0561567843,-0.0514068902,0.1831353456,-0.1885332465,-0.1543068439,-0.1911274195,-0.0294206757,0.1126435772,-0.0549775735,0.1539233476,-0.0694048703,0.2732542157,0.3387182355,0.5176922679,0.2432592809,0.0198027082,0.2993902564,-0.1592840999,0.3542063534,-0.3163362145,-0.0658517927,0.0955877304,-0.2693164647,-0.0386303589,0.3650593162,0.1198536456,0.1819965839,0.0208005905,0.0422853008,-0.1887388676,-0.3113326728,0.2352000475,-0.2445181459,0.1443541795,0.1757642031,0.1281597018,-0.0332734995,-0.1915501207,-0.1471121907,0.000200557,0.0424690768,-0.0058078957,-0.3604525924,0.1492104828,-0.5625633001,0.2226690948,0.1985478103,0.174543187,-0.1717173457,-0.0993915275,0.2057587057,-0.0968841463,0.7381255627,0.0631980151,0.0537928417,0.0406153575,-0.1717147529,-0.3108314872,0.1234151125,-0.0936188549,0.0283297058,0.1906214952,0.4489213228,-0.1566934884,-0.0801721662,-0.1717058718,0.1107357293,-0.0656813756,-0.0861324668,-0.2430154532,-0.33938241,-0.172399044,0.0112697436,0.0917003453,0.1303760856,-0.1485615075,0.2206910104,-0.0087004332,0.0192539431,0.2384978533,-0.086740762,0.3990580738,-0.1014834344,0.0295701474,0.1886966079,0.1497644633,-0.029397957,0.5788064003,-0.0160102658,-0.2766880691,-0.2209342867,-0.2195673436,0.0078752395,0.5253742337,-0.1171140894,0.1944184452,0.2502798736,-0.0213026386,-0.2537555993,0.0922321901,0.2513122261,0.1302412301,-0.2400285155,-0.3940942287,0.3845086396,0.0931559727,-0.1280459911,0.0852604359,0.0336195305,-0.5783619285,0.012126294,-0.1870983094,1.0526423454,0.0697231144,0.5438438654,0.1644590646,-0.2343123555,0.2079813331,0.0299140941,0.0776039883,-0.1372855157,-0.1231848076,-0.0392164625,0.0493473448,0.1252042949,0.0445113406,-0.1628301293,0.1048292518,-0.132482633,0.4626202285,0.0255191345,0.0554909743,-0.3534593582,-0.2368060797,-0.1421909034,0.1792891026,-0.004024928,0.1795131713,-0.0311280414,0.0284819733,-0.1299783736,-0.1594067365,-0.0400604084,0.0552265011,-0.2622327805,0.0518419109,0.1309134513,-0.2550185323,-0.1852672994,0.3582657576,0.4886320829,0.058994893,-0.0226820111,0.181029439,0.1769503802,0.3690515757,0.1631838232,-0.07118617,0.1402853727,0.0347783454,-0.2269736081,-0.0868552998,0.0416178294,-0.2070045024,-0.1773606241,0.1021571979,-0.2035160661,-0.3819696307,-0.1024412736,-0.0708086044,0.0490292683,-0.1543621719,0.2164075971,0.0182434004,-0.147185564,0.1232392117,-0.2301244885,-0.3487478197,-0.2812531292,0.6182484627,0.3330359459,-0.1536255181,0.6093755364,0.0298628397,-0.2135214955,-0.233123973,-0.0615369491,-0.0391099863,-0.3748999238,0.1753538996,-0.165901348,-0.0390571654,-0.1358298361,0.0943607613,-0.0543193333,-0.004858824,-0.0295857415,-0.2917799354,-0.2042460889,-0.0120187737,0.2862655222,0.2789078653,0.2129031718,0.1974249631,-0.0714538991,0.0140298922,-0.4197110236,0.2962636054,0.1460199356,0.1726618558,0.0889052376,-0.1397820711,0.3646045625,-0.0083557302,0.2012527734,0.2086616457,-0.2705426216,-0.3591167331,-0.2536821067,0.0615098216,-0.1423659325,-0.0799691454,0.0010466283,0.0470733345,0.0717447102,-0.3589701355,0.3513580859,0.4166297317,0.0382954665,0.1437445134,-0.0534596704,0.3532390296,0.1702318937,-0.1074290127,-0.0596709996,0.175994128,0.0771526396,0.1461575478,-0.3433862329,0.0248257127,0.1210998446,-0.0221874565,0.2799771726,0.0071432041,0.085866563,-0.4575456083,-0.1887638122,0.1324160546,0.4153516591,0.3262939751,-0.3256803751,-0.3639090955,0.0287527274,0.2579559684,-0.0927454978,-0.2130314112,0.2842707038,0.0909185335,0.0438745432,0.3295021355,0.3479600549,-0.0955454558,0.0770696104,-0.0678816736,0.502114296,-0.5044654012,0.1415495425,0.1453690231,-0.0150588462,0.104600206,0.2185519636,0.4027286172,0.0882586166,0.7428874373,0.0694247782,0.2655401528,0.0344629623,0.3497874439,-0.3061888516,-0.2714808881,0.008783008,0.3392660022,-0.5183728337,0.3594899476,0.0399041437,0.2947771251,0.1023311466,0.0146622257,-0.1142988801,0.0610385463,-0.1927390993,-0.4710101485,0.1543368697,-0.1860476434,-0.0982344151,0.0535010546,-0.0423500426,0.0210374128,0.1267087758,-0.0657493398,-0.0144562367,-0.3596044183,-0.0651701018,-0.2634501457,0.2892220914,-0.2736432254,0.3618342578,0.251624167,-0.0244300608,0.2664103508,0.1843282729,0.4134000838,0.0964873135,0.1391860098,0.2343908697,-0.3441286087,0.0173973534,-0.2722978592,0.3577625453,0.3037012219,-0.0722733214,0.1832407415,0.227822572,-0.1841198653,0.1968555599,0.2459048033,0.2491842508,-0.4332534969,0.1303792149,-0.2235026807,-0.0221770573,-0.2555291951,0.0684460923,-0.401979655,0.0493896678,0.4003736675,-0.0126161426,0.3317835629,-0.0323130302,0.1192988753,0.0970787257,0.3910096288,0.246708557,0.0202430859,0.0852722526,-0.1726862341,-0.5865501165,0.2221587896,-0.0296891071,0.134458974,-0.1434549987,0.2045404613,-0.0341443382,-0.0542166792,-0.0914015993,-0.1793846935,-0.2699118555,-0.2181200981,-0.2253098935,-0.3098490536,-0.057099171,0.0176589433,0.1262981147,-0.2951638401,0.1995938867,-0.1299218386,0.171095401,-0.0734913722,0.1191050187,-0.1012372077,-0.0817152485,0.3345757723,0.2134465724,0.0621457286,-0.1052110791,-0.0699443445,-0.2133233249,-0.0632963851,-0.1850117743,0.397562474,0.138302207,0.240978241,-0.2387066185,-0.1000701413,-0.0154632032,0.3532556891,0.3000925779,-0.2251818329,0.0405854769,0.0742222518,0.0315963812,0.0677158907,-0.0424550287,0.7119355798,-0.1609653383,0.1956686378,-0.1226238906,-0.4357623458,0.4972657859,-0.4109683335,-0.3635086715,0.0999604687,0.1996232271,0.1974008381,-0.2245982289,-0.6479585767,0.2402899712,0.3244629204,-0.1465073675,-0.289839983,0.2578243613,-0.0659369752,-0.0560158901,-0.1371863931,0.4104946852,0.2625234723,-0.1719959676,0.1602532119,-0.4273609519]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/726","title":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset","comments":"Hi try, to provide more information please.\r\n\r\nExample code in a colab to reproduce the error, details on what you are trying to do and what you were expected and details on your environment (OS, PyPi packages version).","body":"Hi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.","comment_length":38,"text":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset\nHi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.\nHi try, to provide more information please.\r\n\r\nExample code in a colab to reproduce the error, details on what you are trying to do and what you were expected and details on your environment (OS, PyPi packages version).","embeddings":[-0.4920600653,0.0389875546,-0.0594882704,0.3662634194,0.2443764508,-0.0337870046,0.13654387,0.550494194,-0.00061072,-0.0602816343,-0.0666541532,-0.0310110506,0.0238088965,0.3648073375,-0.2299020737,0.2352990061,-0.1055581272,0.0833998173,-0.3995411098,-0.0401990227,-0.2967323363,0.1188192815,-0.1012961715,-0.1545280963,0.0459318496,0.0663644373,0.2705113292,0.1403153092,-0.2775344551,-0.2145338356,0.1424371004,-0.1216851696,-0.1503286809,0.324788034,-0.0001102053,0.2582596838,0.3656147122,-0.2235869169,-0.2715173662,-0.1560243964,-0.4222053289,-0.1340990663,-0.1875573993,-0.1528116018,0.0739320889,-0.2589229047,-0.1242828369,-0.4580081105,0.0629987568,0.2549592257,0.2594358921,0.5298959017,0.4289398491,0.1028628424,0.2806044221,-0.0724724978,0.0572922044,0.1174185649,0.1282486916,0.1596287191,0.0161918625,-0.0816258788,-0.1522659957,0.1627645791,0.1623219252,-0.3617706895,0.0090026567,-0.1375187486,0.1970816851,0.3654464483,0.7156856656,-0.1815347672,-0.3790054619,-0.0936338827,-0.1986480951,0.0984760225,0.3893739879,0.3778521121,-0.0309308544,0.0653221011,-0.5064879656,0.1339263022,0.1091979817,0.0512661226,0.0404527225,0.0759120733,0.0293370411,0.008689425,-0.1022983715,-0.1159981489,0.3256934285,-0.2656748891,-0.3963989913,0.137359947,-0.3631493449,0.0508111529,-0.1439990103,0.0500055067,0.366363883,0.6212907434,0.2462342829,0.0420111604,-0.308077246,0.0778114572,0.0899420232,0.3613916039,-0.0054691094,0.1346786171,0.1948350817,0.10071899,0.0852302238,0.0521176606,0.0023271991,-0.542937696,0.1448046714,0.0951064378,0.3429888189,-0.3038514555,-0.4549231231,0.047217641,0.1378372461,-0.1998882294,0.1437949091,-0.0076497574,-0.0868444666,0.0802833959,0.3436453938,0.0880122259,-0.1021173596,-0.2038995028,-0.2310939729,-0.2034747303,-0.0948625132,-0.0629801005,0.1900294423,-0.2003930807,0.6136524081,-0.0917051286,0.0911761746,-0.1107355282,0.0129413661,0.046453923,0.0409467779,0.321911335,-0.1849273741,0.0135516413,0.1199886575,0.1528467536,-0.0941035226,0.2584348321,-0.2343892604,-0.3227925301,-0.1861532629,0.219078213,-0.3183863163,0.0472868122,-0.1684523225,-0.3878966272,0.3156642318,-0.2622767985,-0.0602091402,-0.2444446683,-0.1759607047,-0.138668865,0.3337224126,0.2220809311,-0.1910729706,0.126049161,0.1682969481,-0.1456349194,0.0463749357,0.2319472134,-0.1313233376,0.0905488282,-0.2481278926,-0.0242459103,0.0881310552,-0.5387325287,-0.4547849596,-0.0014980707,0.2015110701,0.0489920415,0.1314415038,-0.1612880677,-0.0724657029,-0.2714241743,-0.2563689649,0.258287251,-0.034631554,0.389313817,-0.2438358366,-0.2965556383,0.0666808411,0.11800237,-0.0492598414,-0.0816775784,0.1693636179,-0.0095629962,0.205123961,0.0532166623,0.0325854793,0.108541891,0.1249753609,-0.1738582253,-0.0426240675,-0.0249788035,-0.1966191828,0.2551011145,-0.0707978681,0.1320103109,0.0627517551,0.178278774,-0.4489194155,-0.1798519492,-0.2049353272,0.0868078247,0.1547742933,0.1201993003,-0.1345562041,0.0315445885,0.097442545,0.1682491302,-0.1691212803,0.0556032844,-0.2289179713,0.2009186,-0.0244281776,-0.1356206685,0.4263866842,0.0483330823,-0.0907463729,0.0649090111,-0.4173122048,0.4893443882,-0.0621393695,0.0347024985,0.0946347788,0.4519070983,0.0079325,-0.3528214693,0.0029319238,0.0297744963,0.1392337382,-0.1301709712,0.2713800669,-0.0893520936,0.0362150632,-0.1035486832,-0.1582131386,-0.1385202408,0.0254937224,-0.0858455822,-0.0036497831,-0.1547219008,0.036334835,0.2815303206,0.2090259939,0.1824414283,0.3501009941,-0.2496142536,0.3917295635,0.0523996353,0.0041363831,0.1314524859,-0.0301868841,-0.0547485016,-0.123130776,0.2424137145,0.4796056449,0.0881364644,0.0710710213,0.1027916372,0.0828066617,-0.0666035786,0.2418119758,0.1984371692,0.3564124107,0.4244209528,-0.1783008277,-0.1144582033,-0.2286888808,0.1582078785,0.1266210526,0.296477139,-0.5625895262,-0.2332416624,-0.3110375106,0.1262899935,-0.0553843863,0.0316452757,-0.3143775761,-0.1802434623,-0.1689627618,0.1081855819,-0.0475393161,0.1951696426,-0.1458839178,0.0477331541,0.3103329241,0.0443977341,0.1516741216,-0.0028465451,-0.1123988777,0.0865353346,0.6746875644,-0.1140086427,0.3397882283,-0.477501303,0.1271569282,-0.3800797462,0.0426583849,-0.0863100588,-0.0893176869,0.2315170616,0.2202654928,0.2755811214,0.0788401291,-0.2032470554,0.0359021463,0.0523619428,-0.4162954092,0.0309536979,-0.0636543036,-0.1320037097,0.1025646105,-0.0423322469,0.0432340465,-0.3893762231,-0.0724421144,-0.0345688462,0.1005237103,0.0962934569,0.0817191377,-0.0508541204,0.0996042937,0.2866059244,-0.1770675182,-0.8880888224,0.3540101349,-0.1670295596,-0.3587040901,-0.0214136578,0.1422802508,0.2374749631,0.0288894922,-0.4529916644,-0.1852105707,-0.2191690356,-0.0274334569,0.0638383254,0.1969694644,0.1162543371,0.0691954195,-0.0823741183,-0.3447431326,-0.2273833901,0.1099680588,-0.2630299628,0.5279319286,-0.0101229204,0.062973924,0.0090144631,0.1534843743,0.411018759,-0.1777416617,0.424354881,0.2426645756,0.3268184364,-0.2149715424,-0.1142090335,0.0165318102,-0.1049812809,0.333722651,0.2043135762,0.0627743155,0.2587611377,-0.1610162407,-0.0275116041,-0.5019219518,-0.3119254112,-0.1261365712,-0.4771397412,0.1579095274,0.0015204475,-0.0457791425,-0.1766192168,-0.2818224728,0.1243380383,0.0594626069,-0.0118080024,0.0062621678,-0.1055888012,0.1124698818,-0.0435051918,0.3818668425,0.1593196094,0.4464143515,0.0325865448,-0.0394472592,-0.1186385751,0.2015811354,0.1771973073,-0.1391032636,0.3249413669,0.0442026407,0.1309691817,-0.1988399774,-0.2067622989,0.2606891096,-0.0236098375,0.3816577494,-0.1196590289,-0.4288542867,0.0298066325,0.2281961292,0.4090623558,-0.1562371254,-0.1394207478,-0.2068855464,-0.3080563843,-0.3402201235,-0.3694739044,0.1338023245,0.2898083329,-0.1530362666,-0.0328774005,-0.1857864559,-0.266718328,0.3404831588,-0.1804640293,0.4067646861,0.0107006691,0.3208973706,0.0213317052,0.2319337279,0.4421767592,0.7271000743,-0.1791524738,-0.21856004,-0.0536129028,-0.1541273147,-0.2087100595,0.0071359742,0.0306307096,-0.2557501495,0.0924119502,-0.0574301518,0.044365827,0.0608743764,0.1258152276,0.0699756593,-0.4000614882,-0.2998196185,0.2181069702,-0.156189397,0.0378430523,0.5064824224,0.0676232576,-0.0924678519,0.0669312477,0.131942898,0.6712408066,-0.1444398016,-0.1336941421,0.0441391766,-0.1740802377,-0.2378405482,0.0318626277,-0.2122766972,-0.3817217052,-0.5945901871,0.0930716395,-0.2704686224,0.0329011157,0.2246014923,0.0661379397,0.0567725748,-0.0167895537,-0.0726982355,0.0699928254,0.3150302172,-0.3102372885,0.0056242491,0.0674703941,0.1505780518,-0.2651734948,0.2959451079,-0.0667072609,-0.1344791949,-0.0533542447,-0.2241715193,-0.6046833396,0.063356787,-0.2882435322,0.2968223691,0.2508451641,0.0494315699,0.157456249,0.4011854529,0.4436883628,0.2695259452,-0.2344846129,0.172759518,0.3623528481,-0.0145823015,-0.0064136069,0.2117148489,0.4534879327,-0.1813885272,-0.2389175892,0.037739303,-0.0976290405,-0.4282446206,0.1780985296,-0.2890240252,0.1226878613,-0.4741533399,-0.0448160693,-0.1610464007,0.0495670922,-0.11707744,0.1266120225,0.3015568852,-0.0258204844,0.3938731849,0.2872494459,-0.4737055004,0.0818112344,0.4354539514,-0.2639756501,-0.0656893551,0.7143592834,-0.0224221386,-0.1156795174,-0.08455538,0.3356194794,0.1899365485,-0.4613088965,0.1759456396,-0.1370148063,0.1676396132,-0.0239781868,0.5160508752,0.4567060471,0.3805165291,0.3290770948,-0.5187920928,-0.1200366989,0.0344009176,-0.1044829786,0.3924307227,-0.1633544564,0.0690328032,0.0379482545,0.0105704777,-0.2830748558,0.1500228792,-0.3312316835,0.0517838746,0.1376151294,0.0577630289,0.091541335,-0.0988202617,0.1534478515,-0.1528942436,-0.3658686876,-0.2194008082,-0.0937628448,0.1155195087,0.0607097931,-0.3785829842,0.1993471235,-0.3264406323,-0.1617555618,-0.1271059513,0.0065265517,0.033089947,-0.036198806,0.027435381,0.2251442075,-0.0241950527,0.0211056527,0.056758821,-0.362577498,0.0497004502,0.0169090182,0.1100955606,0.0522976965,-0.1036025807,0.0623393506,0.0843107179,-0.0241844431,0.1517633647,0.3195523918,-0.2027384341,0.0648035333,0.482668072,0.306938976,0.02040576,-0.2247438431,-0.1236111969,0.1812740713,0.1958637685,-0.3893767595,-0.0456152521,0.0189761855,0.2228658348,-0.00635838,-0.1106350273,0.2290060967,-0.2677126527,-0.0688804686,0.0097189955,0.2057594508,0.2482970953,0.2930920422,0.5799326897,-0.175913319,-0.1093828455,-0.0568373799,-0.1746608615,0.0977744833,0.0135684013,-0.1358999163,0.4929817021,-0.323251158,0.2292830646,-0.2266659588,-0.7084224224,0.3919857144,-0.1421013623,0.0754047036,0.0047916854,0.2124478519,0.2448038906,-0.0497319512,0.1501278579,-0.3210782111,0.4191267788,0.0932898372,-0.0296805296,-0.2051957548,-0.2670288384,0.1096700281,0.0063231117,-0.1057886407,0.1221756935,-0.0712562352,0.1627322435,-0.302936852,-0.1216503754,0.0532986298,-0.1697726995,-0.0355583094,-0.2146684378,0.2357464135,0.2564383149,-0.1772019267,0.3424805999,0.7380926013,0.371566534,0.4830287099,0.1371567845,-0.1722052097,0.1276460886,0.127407223,-0.0393284895,0.2327709645,0.0634962171,-0.0034288804,0.3761826754,0.1473830491,-0.0341962501,-0.3039711714,-0.2243894935,-0.2451120466,0.0394476093,0.2884531021,0.0391374789,0.0263525322,-0.2462791055,0.1971482038,-0.0284639299,-0.0342451669,0.204590708,-0.0086461948,0.1436299533,-0.0167889558,0.095203504,-0.1158063412,0.4963770211,0.3752224743,-0.1486891657,-0.1958691478,-0.3705138862,-0.5975680351,0.2106411606,0.1432383657,-0.1041443199,0.1782567203,-0.0372597538,0.0617663376,-0.1067649201,0.492202729,0.0131060984,-0.196629703,0.0617452897,-0.244164601,-0.2179751843,-0.2281747907,-0.2018176019,-0.0431281328,-0.3614442348,0.1915898472,0.0650836453,0.1579857618,-0.1930200607,0.0880778953,0.0480715744,0.0848006159,0.1467909068,-0.0055310964,0.4740925133,0.053589765,-0.0633306205,-0.0972624496,-0.1911649406,-0.1607998013,0.5104889274,-0.0338879116,0.2085700631,-0.0188264493,-0.0699278265,-0.2319627553,0.3770512044,-0.1640319824,-0.1332943738,-0.4568928182,-0.2069936842,-0.0227444656,-0.2393719554,0.3742827773,-0.108578369,-0.2206009477,0.3657265007,0.0113386977,-0.3828713298,0.4573955536,0.0634475201,0.0366952717,-0.147557646,0.0569182262,-0.1345574558,-0.0479519702,-0.5954985023,0.2696992755,0.4791096151,0.0051987641,-0.0346901678,0.2399044782,-0.1095316857,0.2111077607,0.0491965562,0.1587296426,-0.03080176,-0.2375629991,-0.2700150609,-0.0389962234]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/726","title":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset","comments":"> Hi try, to provide more information please.\r\n> \r\n> Example code in a colab to reproduce the error, details on what you are trying to do and what you were expected and details on your environment (OS, PyPi packages version).\r\n\r\nI have update the description, sorry for the incomplete issue by mistake.","body":"Hi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.","comment_length":53,"text":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset\nHi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.\n> Hi try, to provide more information please.\r\n> \r\n> Example code in a colab to reproduce the error, details on what you are trying to do and what you were expected and details on your environment (OS, PyPi packages version).\r\n\r\nI have update the description, sorry for the incomplete issue by mistake.","embeddings":[-0.4920600653,0.0389875546,-0.0594882704,0.3662634194,0.2443764508,-0.0337870046,0.13654387,0.550494194,-0.00061072,-0.0602816343,-0.0666541532,-0.0310110506,0.0238088965,0.3648073375,-0.2299020737,0.2352990061,-0.1055581272,0.0833998173,-0.3995411098,-0.0401990227,-0.2967323363,0.1188192815,-0.1012961715,-0.1545280963,0.0459318496,0.0663644373,0.2705113292,0.1403153092,-0.2775344551,-0.2145338356,0.1424371004,-0.1216851696,-0.1503286809,0.324788034,-0.0001102053,0.2582596838,0.3656147122,-0.2235869169,-0.2715173662,-0.1560243964,-0.4222053289,-0.1340990663,-0.1875573993,-0.1528116018,0.0739320889,-0.2589229047,-0.1242828369,-0.4580081105,0.0629987568,0.2549592257,0.2594358921,0.5298959017,0.4289398491,0.1028628424,0.2806044221,-0.0724724978,0.0572922044,0.1174185649,0.1282486916,0.1596287191,0.0161918625,-0.0816258788,-0.1522659957,0.1627645791,0.1623219252,-0.3617706895,0.0090026567,-0.1375187486,0.1970816851,0.3654464483,0.7156856656,-0.1815347672,-0.3790054619,-0.0936338827,-0.1986480951,0.0984760225,0.3893739879,0.3778521121,-0.0309308544,0.0653221011,-0.5064879656,0.1339263022,0.1091979817,0.0512661226,0.0404527225,0.0759120733,0.0293370411,0.008689425,-0.1022983715,-0.1159981489,0.3256934285,-0.2656748891,-0.3963989913,0.137359947,-0.3631493449,0.0508111529,-0.1439990103,0.0500055067,0.366363883,0.6212907434,0.2462342829,0.0420111604,-0.308077246,0.0778114572,0.0899420232,0.3613916039,-0.0054691094,0.1346786171,0.1948350817,0.10071899,0.0852302238,0.0521176606,0.0023271991,-0.542937696,0.1448046714,0.0951064378,0.3429888189,-0.3038514555,-0.4549231231,0.047217641,0.1378372461,-0.1998882294,0.1437949091,-0.0076497574,-0.0868444666,0.0802833959,0.3436453938,0.0880122259,-0.1021173596,-0.2038995028,-0.2310939729,-0.2034747303,-0.0948625132,-0.0629801005,0.1900294423,-0.2003930807,0.6136524081,-0.0917051286,0.0911761746,-0.1107355282,0.0129413661,0.046453923,0.0409467779,0.321911335,-0.1849273741,0.0135516413,0.1199886575,0.1528467536,-0.0941035226,0.2584348321,-0.2343892604,-0.3227925301,-0.1861532629,0.219078213,-0.3183863163,0.0472868122,-0.1684523225,-0.3878966272,0.3156642318,-0.2622767985,-0.0602091402,-0.2444446683,-0.1759607047,-0.138668865,0.3337224126,0.2220809311,-0.1910729706,0.126049161,0.1682969481,-0.1456349194,0.0463749357,0.2319472134,-0.1313233376,0.0905488282,-0.2481278926,-0.0242459103,0.0881310552,-0.5387325287,-0.4547849596,-0.0014980707,0.2015110701,0.0489920415,0.1314415038,-0.1612880677,-0.0724657029,-0.2714241743,-0.2563689649,0.258287251,-0.034631554,0.389313817,-0.2438358366,-0.2965556383,0.0666808411,0.11800237,-0.0492598414,-0.0816775784,0.1693636179,-0.0095629962,0.205123961,0.0532166623,0.0325854793,0.108541891,0.1249753609,-0.1738582253,-0.0426240675,-0.0249788035,-0.1966191828,0.2551011145,-0.0707978681,0.1320103109,0.0627517551,0.178278774,-0.4489194155,-0.1798519492,-0.2049353272,0.0868078247,0.1547742933,0.1201993003,-0.1345562041,0.0315445885,0.097442545,0.1682491302,-0.1691212803,0.0556032844,-0.2289179713,0.2009186,-0.0244281776,-0.1356206685,0.4263866842,0.0483330823,-0.0907463729,0.0649090111,-0.4173122048,0.4893443882,-0.0621393695,0.0347024985,0.0946347788,0.4519070983,0.0079325,-0.3528214693,0.0029319238,0.0297744963,0.1392337382,-0.1301709712,0.2713800669,-0.0893520936,0.0362150632,-0.1035486832,-0.1582131386,-0.1385202408,0.0254937224,-0.0858455822,-0.0036497831,-0.1547219008,0.036334835,0.2815303206,0.2090259939,0.1824414283,0.3501009941,-0.2496142536,0.3917295635,0.0523996353,0.0041363831,0.1314524859,-0.0301868841,-0.0547485016,-0.123130776,0.2424137145,0.4796056449,0.0881364644,0.0710710213,0.1027916372,0.0828066617,-0.0666035786,0.2418119758,0.1984371692,0.3564124107,0.4244209528,-0.1783008277,-0.1144582033,-0.2286888808,0.1582078785,0.1266210526,0.296477139,-0.5625895262,-0.2332416624,-0.3110375106,0.1262899935,-0.0553843863,0.0316452757,-0.3143775761,-0.1802434623,-0.1689627618,0.1081855819,-0.0475393161,0.1951696426,-0.1458839178,0.0477331541,0.3103329241,0.0443977341,0.1516741216,-0.0028465451,-0.1123988777,0.0865353346,0.6746875644,-0.1140086427,0.3397882283,-0.477501303,0.1271569282,-0.3800797462,0.0426583849,-0.0863100588,-0.0893176869,0.2315170616,0.2202654928,0.2755811214,0.0788401291,-0.2032470554,0.0359021463,0.0523619428,-0.4162954092,0.0309536979,-0.0636543036,-0.1320037097,0.1025646105,-0.0423322469,0.0432340465,-0.3893762231,-0.0724421144,-0.0345688462,0.1005237103,0.0962934569,0.0817191377,-0.0508541204,0.0996042937,0.2866059244,-0.1770675182,-0.8880888224,0.3540101349,-0.1670295596,-0.3587040901,-0.0214136578,0.1422802508,0.2374749631,0.0288894922,-0.4529916644,-0.1852105707,-0.2191690356,-0.0274334569,0.0638383254,0.1969694644,0.1162543371,0.0691954195,-0.0823741183,-0.3447431326,-0.2273833901,0.1099680588,-0.2630299628,0.5279319286,-0.0101229204,0.062973924,0.0090144631,0.1534843743,0.411018759,-0.1777416617,0.424354881,0.2426645756,0.3268184364,-0.2149715424,-0.1142090335,0.0165318102,-0.1049812809,0.333722651,0.2043135762,0.0627743155,0.2587611377,-0.1610162407,-0.0275116041,-0.5019219518,-0.3119254112,-0.1261365712,-0.4771397412,0.1579095274,0.0015204475,-0.0457791425,-0.1766192168,-0.2818224728,0.1243380383,0.0594626069,-0.0118080024,0.0062621678,-0.1055888012,0.1124698818,-0.0435051918,0.3818668425,0.1593196094,0.4464143515,0.0325865448,-0.0394472592,-0.1186385751,0.2015811354,0.1771973073,-0.1391032636,0.3249413669,0.0442026407,0.1309691817,-0.1988399774,-0.2067622989,0.2606891096,-0.0236098375,0.3816577494,-0.1196590289,-0.4288542867,0.0298066325,0.2281961292,0.4090623558,-0.1562371254,-0.1394207478,-0.2068855464,-0.3080563843,-0.3402201235,-0.3694739044,0.1338023245,0.2898083329,-0.1530362666,-0.0328774005,-0.1857864559,-0.266718328,0.3404831588,-0.1804640293,0.4067646861,0.0107006691,0.3208973706,0.0213317052,0.2319337279,0.4421767592,0.7271000743,-0.1791524738,-0.21856004,-0.0536129028,-0.1541273147,-0.2087100595,0.0071359742,0.0306307096,-0.2557501495,0.0924119502,-0.0574301518,0.044365827,0.0608743764,0.1258152276,0.0699756593,-0.4000614882,-0.2998196185,0.2181069702,-0.156189397,0.0378430523,0.5064824224,0.0676232576,-0.0924678519,0.0669312477,0.131942898,0.6712408066,-0.1444398016,-0.1336941421,0.0441391766,-0.1740802377,-0.2378405482,0.0318626277,-0.2122766972,-0.3817217052,-0.5945901871,0.0930716395,-0.2704686224,0.0329011157,0.2246014923,0.0661379397,0.0567725748,-0.0167895537,-0.0726982355,0.0699928254,0.3150302172,-0.3102372885,0.0056242491,0.0674703941,0.1505780518,-0.2651734948,0.2959451079,-0.0667072609,-0.1344791949,-0.0533542447,-0.2241715193,-0.6046833396,0.063356787,-0.2882435322,0.2968223691,0.2508451641,0.0494315699,0.157456249,0.4011854529,0.4436883628,0.2695259452,-0.2344846129,0.172759518,0.3623528481,-0.0145823015,-0.0064136069,0.2117148489,0.4534879327,-0.1813885272,-0.2389175892,0.037739303,-0.0976290405,-0.4282446206,0.1780985296,-0.2890240252,0.1226878613,-0.4741533399,-0.0448160693,-0.1610464007,0.0495670922,-0.11707744,0.1266120225,0.3015568852,-0.0258204844,0.3938731849,0.2872494459,-0.4737055004,0.0818112344,0.4354539514,-0.2639756501,-0.0656893551,0.7143592834,-0.0224221386,-0.1156795174,-0.08455538,0.3356194794,0.1899365485,-0.4613088965,0.1759456396,-0.1370148063,0.1676396132,-0.0239781868,0.5160508752,0.4567060471,0.3805165291,0.3290770948,-0.5187920928,-0.1200366989,0.0344009176,-0.1044829786,0.3924307227,-0.1633544564,0.0690328032,0.0379482545,0.0105704777,-0.2830748558,0.1500228792,-0.3312316835,0.0517838746,0.1376151294,0.0577630289,0.091541335,-0.0988202617,0.1534478515,-0.1528942436,-0.3658686876,-0.2194008082,-0.0937628448,0.1155195087,0.0607097931,-0.3785829842,0.1993471235,-0.3264406323,-0.1617555618,-0.1271059513,0.0065265517,0.033089947,-0.036198806,0.027435381,0.2251442075,-0.0241950527,0.0211056527,0.056758821,-0.362577498,0.0497004502,0.0169090182,0.1100955606,0.0522976965,-0.1036025807,0.0623393506,0.0843107179,-0.0241844431,0.1517633647,0.3195523918,-0.2027384341,0.0648035333,0.482668072,0.306938976,0.02040576,-0.2247438431,-0.1236111969,0.1812740713,0.1958637685,-0.3893767595,-0.0456152521,0.0189761855,0.2228658348,-0.00635838,-0.1106350273,0.2290060967,-0.2677126527,-0.0688804686,0.0097189955,0.2057594508,0.2482970953,0.2930920422,0.5799326897,-0.175913319,-0.1093828455,-0.0568373799,-0.1746608615,0.0977744833,0.0135684013,-0.1358999163,0.4929817021,-0.323251158,0.2292830646,-0.2266659588,-0.7084224224,0.3919857144,-0.1421013623,0.0754047036,0.0047916854,0.2124478519,0.2448038906,-0.0497319512,0.1501278579,-0.3210782111,0.4191267788,0.0932898372,-0.0296805296,-0.2051957548,-0.2670288384,0.1096700281,0.0063231117,-0.1057886407,0.1221756935,-0.0712562352,0.1627322435,-0.302936852,-0.1216503754,0.0532986298,-0.1697726995,-0.0355583094,-0.2146684378,0.2357464135,0.2564383149,-0.1772019267,0.3424805999,0.7380926013,0.371566534,0.4830287099,0.1371567845,-0.1722052097,0.1276460886,0.127407223,-0.0393284895,0.2327709645,0.0634962171,-0.0034288804,0.3761826754,0.1473830491,-0.0341962501,-0.3039711714,-0.2243894935,-0.2451120466,0.0394476093,0.2884531021,0.0391374789,0.0263525322,-0.2462791055,0.1971482038,-0.0284639299,-0.0342451669,0.204590708,-0.0086461948,0.1436299533,-0.0167889558,0.095203504,-0.1158063412,0.4963770211,0.3752224743,-0.1486891657,-0.1958691478,-0.3705138862,-0.5975680351,0.2106411606,0.1432383657,-0.1041443199,0.1782567203,-0.0372597538,0.0617663376,-0.1067649201,0.492202729,0.0131060984,-0.196629703,0.0617452897,-0.244164601,-0.2179751843,-0.2281747907,-0.2018176019,-0.0431281328,-0.3614442348,0.1915898472,0.0650836453,0.1579857618,-0.1930200607,0.0880778953,0.0480715744,0.0848006159,0.1467909068,-0.0055310964,0.4740925133,0.053589765,-0.0633306205,-0.0972624496,-0.1911649406,-0.1607998013,0.5104889274,-0.0338879116,0.2085700631,-0.0188264493,-0.0699278265,-0.2319627553,0.3770512044,-0.1640319824,-0.1332943738,-0.4568928182,-0.2069936842,-0.0227444656,-0.2393719554,0.3742827773,-0.108578369,-0.2206009477,0.3657265007,0.0113386977,-0.3828713298,0.4573955536,0.0634475201,0.0366952717,-0.147557646,0.0569182262,-0.1345574558,-0.0479519702,-0.5954985023,0.2696992755,0.4791096151,0.0051987641,-0.0346901678,0.2399044782,-0.1095316857,0.2111077607,0.0491965562,0.1587296426,-0.03080176,-0.2375629991,-0.2700150609,-0.0389962234]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/726","title":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset","comments":"Hi, I have manually downloaded the compressed dataset `openwebtext.tar.xz' and use the following command to preprocess the examples:\r\n```\r\n>>> dataset = load_dataset('\/home\/admin\/workspace\/datasets\/datasets-master\/datasets-master\/datasets\/openwebtext', data_dir='\/home\/admin\/workspace\/datasets')\r\nUsing custom data configuration default\r\nDownloading and preparing dataset openwebtext\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/default\/0.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nDataset openwebtext downloaded and prepared to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/default\/0.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02. Subsequent calls will reuse this data.\r\n>>> len(dataset['train'])\r\n74571\r\n>>>\r\n```\r\nThe size of the pre-processed example file is only 354MB, however the processed bookcorpus dataset is 4.6g. Are there any problems?","body":"Hi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.","comment_length":87,"text":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset\nHi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.\nHi, I have manually downloaded the compressed dataset `openwebtext.tar.xz' and use the following command to preprocess the examples:\r\n```\r\n>>> dataset = load_dataset('\/home\/admin\/workspace\/datasets\/datasets-master\/datasets-master\/datasets\/openwebtext', data_dir='\/home\/admin\/workspace\/datasets')\r\nUsing custom data configuration default\r\nDownloading and preparing dataset openwebtext\/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/default\/0.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nDataset openwebtext downloaded and prepared to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/default\/0.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02. Subsequent calls will reuse this data.\r\n>>> len(dataset['train'])\r\n74571\r\n>>>\r\n```\r\nThe size of the pre-processed example file is only 354MB, however the processed bookcorpus dataset is 4.6g. Are there any problems?","embeddings":[-0.4920600653,0.0389875546,-0.0594882704,0.3662634194,0.2443764508,-0.0337870046,0.13654387,0.550494194,-0.00061072,-0.0602816343,-0.0666541532,-0.0310110506,0.0238088965,0.3648073375,-0.2299020737,0.2352990061,-0.1055581272,0.0833998173,-0.3995411098,-0.0401990227,-0.2967323363,0.1188192815,-0.1012961715,-0.1545280963,0.0459318496,0.0663644373,0.2705113292,0.1403153092,-0.2775344551,-0.2145338356,0.1424371004,-0.1216851696,-0.1503286809,0.324788034,-0.0001102053,0.2582596838,0.3656147122,-0.2235869169,-0.2715173662,-0.1560243964,-0.4222053289,-0.1340990663,-0.1875573993,-0.1528116018,0.0739320889,-0.2589229047,-0.1242828369,-0.4580081105,0.0629987568,0.2549592257,0.2594358921,0.5298959017,0.4289398491,0.1028628424,0.2806044221,-0.0724724978,0.0572922044,0.1174185649,0.1282486916,0.1596287191,0.0161918625,-0.0816258788,-0.1522659957,0.1627645791,0.1623219252,-0.3617706895,0.0090026567,-0.1375187486,0.1970816851,0.3654464483,0.7156856656,-0.1815347672,-0.3790054619,-0.0936338827,-0.1986480951,0.0984760225,0.3893739879,0.3778521121,-0.0309308544,0.0653221011,-0.5064879656,0.1339263022,0.1091979817,0.0512661226,0.0404527225,0.0759120733,0.0293370411,0.008689425,-0.1022983715,-0.1159981489,0.3256934285,-0.2656748891,-0.3963989913,0.137359947,-0.3631493449,0.0508111529,-0.1439990103,0.0500055067,0.366363883,0.6212907434,0.2462342829,0.0420111604,-0.308077246,0.0778114572,0.0899420232,0.3613916039,-0.0054691094,0.1346786171,0.1948350817,0.10071899,0.0852302238,0.0521176606,0.0023271991,-0.542937696,0.1448046714,0.0951064378,0.3429888189,-0.3038514555,-0.4549231231,0.047217641,0.1378372461,-0.1998882294,0.1437949091,-0.0076497574,-0.0868444666,0.0802833959,0.3436453938,0.0880122259,-0.1021173596,-0.2038995028,-0.2310939729,-0.2034747303,-0.0948625132,-0.0629801005,0.1900294423,-0.2003930807,0.6136524081,-0.0917051286,0.0911761746,-0.1107355282,0.0129413661,0.046453923,0.0409467779,0.321911335,-0.1849273741,0.0135516413,0.1199886575,0.1528467536,-0.0941035226,0.2584348321,-0.2343892604,-0.3227925301,-0.1861532629,0.219078213,-0.3183863163,0.0472868122,-0.1684523225,-0.3878966272,0.3156642318,-0.2622767985,-0.0602091402,-0.2444446683,-0.1759607047,-0.138668865,0.3337224126,0.2220809311,-0.1910729706,0.126049161,0.1682969481,-0.1456349194,0.0463749357,0.2319472134,-0.1313233376,0.0905488282,-0.2481278926,-0.0242459103,0.0881310552,-0.5387325287,-0.4547849596,-0.0014980707,0.2015110701,0.0489920415,0.1314415038,-0.1612880677,-0.0724657029,-0.2714241743,-0.2563689649,0.258287251,-0.034631554,0.389313817,-0.2438358366,-0.2965556383,0.0666808411,0.11800237,-0.0492598414,-0.0816775784,0.1693636179,-0.0095629962,0.205123961,0.0532166623,0.0325854793,0.108541891,0.1249753609,-0.1738582253,-0.0426240675,-0.0249788035,-0.1966191828,0.2551011145,-0.0707978681,0.1320103109,0.0627517551,0.178278774,-0.4489194155,-0.1798519492,-0.2049353272,0.0868078247,0.1547742933,0.1201993003,-0.1345562041,0.0315445885,0.097442545,0.1682491302,-0.1691212803,0.0556032844,-0.2289179713,0.2009186,-0.0244281776,-0.1356206685,0.4263866842,0.0483330823,-0.0907463729,0.0649090111,-0.4173122048,0.4893443882,-0.0621393695,0.0347024985,0.0946347788,0.4519070983,0.0079325,-0.3528214693,0.0029319238,0.0297744963,0.1392337382,-0.1301709712,0.2713800669,-0.0893520936,0.0362150632,-0.1035486832,-0.1582131386,-0.1385202408,0.0254937224,-0.0858455822,-0.0036497831,-0.1547219008,0.036334835,0.2815303206,0.2090259939,0.1824414283,0.3501009941,-0.2496142536,0.3917295635,0.0523996353,0.0041363831,0.1314524859,-0.0301868841,-0.0547485016,-0.123130776,0.2424137145,0.4796056449,0.0881364644,0.0710710213,0.1027916372,0.0828066617,-0.0666035786,0.2418119758,0.1984371692,0.3564124107,0.4244209528,-0.1783008277,-0.1144582033,-0.2286888808,0.1582078785,0.1266210526,0.296477139,-0.5625895262,-0.2332416624,-0.3110375106,0.1262899935,-0.0553843863,0.0316452757,-0.3143775761,-0.1802434623,-0.1689627618,0.1081855819,-0.0475393161,0.1951696426,-0.1458839178,0.0477331541,0.3103329241,0.0443977341,0.1516741216,-0.0028465451,-0.1123988777,0.0865353346,0.6746875644,-0.1140086427,0.3397882283,-0.477501303,0.1271569282,-0.3800797462,0.0426583849,-0.0863100588,-0.0893176869,0.2315170616,0.2202654928,0.2755811214,0.0788401291,-0.2032470554,0.0359021463,0.0523619428,-0.4162954092,0.0309536979,-0.0636543036,-0.1320037097,0.1025646105,-0.0423322469,0.0432340465,-0.3893762231,-0.0724421144,-0.0345688462,0.1005237103,0.0962934569,0.0817191377,-0.0508541204,0.0996042937,0.2866059244,-0.1770675182,-0.8880888224,0.3540101349,-0.1670295596,-0.3587040901,-0.0214136578,0.1422802508,0.2374749631,0.0288894922,-0.4529916644,-0.1852105707,-0.2191690356,-0.0274334569,0.0638383254,0.1969694644,0.1162543371,0.0691954195,-0.0823741183,-0.3447431326,-0.2273833901,0.1099680588,-0.2630299628,0.5279319286,-0.0101229204,0.062973924,0.0090144631,0.1534843743,0.411018759,-0.1777416617,0.424354881,0.2426645756,0.3268184364,-0.2149715424,-0.1142090335,0.0165318102,-0.1049812809,0.333722651,0.2043135762,0.0627743155,0.2587611377,-0.1610162407,-0.0275116041,-0.5019219518,-0.3119254112,-0.1261365712,-0.4771397412,0.1579095274,0.0015204475,-0.0457791425,-0.1766192168,-0.2818224728,0.1243380383,0.0594626069,-0.0118080024,0.0062621678,-0.1055888012,0.1124698818,-0.0435051918,0.3818668425,0.1593196094,0.4464143515,0.0325865448,-0.0394472592,-0.1186385751,0.2015811354,0.1771973073,-0.1391032636,0.3249413669,0.0442026407,0.1309691817,-0.1988399774,-0.2067622989,0.2606891096,-0.0236098375,0.3816577494,-0.1196590289,-0.4288542867,0.0298066325,0.2281961292,0.4090623558,-0.1562371254,-0.1394207478,-0.2068855464,-0.3080563843,-0.3402201235,-0.3694739044,0.1338023245,0.2898083329,-0.1530362666,-0.0328774005,-0.1857864559,-0.266718328,0.3404831588,-0.1804640293,0.4067646861,0.0107006691,0.3208973706,0.0213317052,0.2319337279,0.4421767592,0.7271000743,-0.1791524738,-0.21856004,-0.0536129028,-0.1541273147,-0.2087100595,0.0071359742,0.0306307096,-0.2557501495,0.0924119502,-0.0574301518,0.044365827,0.0608743764,0.1258152276,0.0699756593,-0.4000614882,-0.2998196185,0.2181069702,-0.156189397,0.0378430523,0.5064824224,0.0676232576,-0.0924678519,0.0669312477,0.131942898,0.6712408066,-0.1444398016,-0.1336941421,0.0441391766,-0.1740802377,-0.2378405482,0.0318626277,-0.2122766972,-0.3817217052,-0.5945901871,0.0930716395,-0.2704686224,0.0329011157,0.2246014923,0.0661379397,0.0567725748,-0.0167895537,-0.0726982355,0.0699928254,0.3150302172,-0.3102372885,0.0056242491,0.0674703941,0.1505780518,-0.2651734948,0.2959451079,-0.0667072609,-0.1344791949,-0.0533542447,-0.2241715193,-0.6046833396,0.063356787,-0.2882435322,0.2968223691,0.2508451641,0.0494315699,0.157456249,0.4011854529,0.4436883628,0.2695259452,-0.2344846129,0.172759518,0.3623528481,-0.0145823015,-0.0064136069,0.2117148489,0.4534879327,-0.1813885272,-0.2389175892,0.037739303,-0.0976290405,-0.4282446206,0.1780985296,-0.2890240252,0.1226878613,-0.4741533399,-0.0448160693,-0.1610464007,0.0495670922,-0.11707744,0.1266120225,0.3015568852,-0.0258204844,0.3938731849,0.2872494459,-0.4737055004,0.0818112344,0.4354539514,-0.2639756501,-0.0656893551,0.7143592834,-0.0224221386,-0.1156795174,-0.08455538,0.3356194794,0.1899365485,-0.4613088965,0.1759456396,-0.1370148063,0.1676396132,-0.0239781868,0.5160508752,0.4567060471,0.3805165291,0.3290770948,-0.5187920928,-0.1200366989,0.0344009176,-0.1044829786,0.3924307227,-0.1633544564,0.0690328032,0.0379482545,0.0105704777,-0.2830748558,0.1500228792,-0.3312316835,0.0517838746,0.1376151294,0.0577630289,0.091541335,-0.0988202617,0.1534478515,-0.1528942436,-0.3658686876,-0.2194008082,-0.0937628448,0.1155195087,0.0607097931,-0.3785829842,0.1993471235,-0.3264406323,-0.1617555618,-0.1271059513,0.0065265517,0.033089947,-0.036198806,0.027435381,0.2251442075,-0.0241950527,0.0211056527,0.056758821,-0.362577498,0.0497004502,0.0169090182,0.1100955606,0.0522976965,-0.1036025807,0.0623393506,0.0843107179,-0.0241844431,0.1517633647,0.3195523918,-0.2027384341,0.0648035333,0.482668072,0.306938976,0.02040576,-0.2247438431,-0.1236111969,0.1812740713,0.1958637685,-0.3893767595,-0.0456152521,0.0189761855,0.2228658348,-0.00635838,-0.1106350273,0.2290060967,-0.2677126527,-0.0688804686,0.0097189955,0.2057594508,0.2482970953,0.2930920422,0.5799326897,-0.175913319,-0.1093828455,-0.0568373799,-0.1746608615,0.0977744833,0.0135684013,-0.1358999163,0.4929817021,-0.323251158,0.2292830646,-0.2266659588,-0.7084224224,0.3919857144,-0.1421013623,0.0754047036,0.0047916854,0.2124478519,0.2448038906,-0.0497319512,0.1501278579,-0.3210782111,0.4191267788,0.0932898372,-0.0296805296,-0.2051957548,-0.2670288384,0.1096700281,0.0063231117,-0.1057886407,0.1221756935,-0.0712562352,0.1627322435,-0.302936852,-0.1216503754,0.0532986298,-0.1697726995,-0.0355583094,-0.2146684378,0.2357464135,0.2564383149,-0.1772019267,0.3424805999,0.7380926013,0.371566534,0.4830287099,0.1371567845,-0.1722052097,0.1276460886,0.127407223,-0.0393284895,0.2327709645,0.0634962171,-0.0034288804,0.3761826754,0.1473830491,-0.0341962501,-0.3039711714,-0.2243894935,-0.2451120466,0.0394476093,0.2884531021,0.0391374789,0.0263525322,-0.2462791055,0.1971482038,-0.0284639299,-0.0342451669,0.204590708,-0.0086461948,0.1436299533,-0.0167889558,0.095203504,-0.1158063412,0.4963770211,0.3752224743,-0.1486891657,-0.1958691478,-0.3705138862,-0.5975680351,0.2106411606,0.1432383657,-0.1041443199,0.1782567203,-0.0372597538,0.0617663376,-0.1067649201,0.492202729,0.0131060984,-0.196629703,0.0617452897,-0.244164601,-0.2179751843,-0.2281747907,-0.2018176019,-0.0431281328,-0.3614442348,0.1915898472,0.0650836453,0.1579857618,-0.1930200607,0.0880778953,0.0480715744,0.0848006159,0.1467909068,-0.0055310964,0.4740925133,0.053589765,-0.0633306205,-0.0972624496,-0.1911649406,-0.1607998013,0.5104889274,-0.0338879116,0.2085700631,-0.0188264493,-0.0699278265,-0.2319627553,0.3770512044,-0.1640319824,-0.1332943738,-0.4568928182,-0.2069936842,-0.0227444656,-0.2393719554,0.3742827773,-0.108578369,-0.2206009477,0.3657265007,0.0113386977,-0.3828713298,0.4573955536,0.0634475201,0.0366952717,-0.147557646,0.0569182262,-0.1345574558,-0.0479519702,-0.5954985023,0.2696992755,0.4791096151,0.0051987641,-0.0346901678,0.2399044782,-0.1095316857,0.2111077607,0.0491965562,0.1587296426,-0.03080176,-0.2375629991,-0.2700150609,-0.0389962234]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/726","title":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset","comments":"NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n\r\ni got this issue when i try to work on my own datasets kindly tell me, from where i can get checksums of train and dev file in my github repo","body":"Hi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.","comment_length":39,"text":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset\nHi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n\r\ni got this issue when i try to work on my own datasets kindly tell me, from where i can get checksums of train and dev file in my github repo","embeddings":[-0.4920600653,0.0389875546,-0.0594882704,0.3662634194,0.2443764508,-0.0337870046,0.13654387,0.550494194,-0.00061072,-0.0602816343,-0.0666541532,-0.0310110506,0.0238088965,0.3648073375,-0.2299020737,0.2352990061,-0.1055581272,0.0833998173,-0.3995411098,-0.0401990227,-0.2967323363,0.1188192815,-0.1012961715,-0.1545280963,0.0459318496,0.0663644373,0.2705113292,0.1403153092,-0.2775344551,-0.2145338356,0.1424371004,-0.1216851696,-0.1503286809,0.324788034,-0.0001102053,0.2582596838,0.3656147122,-0.2235869169,-0.2715173662,-0.1560243964,-0.4222053289,-0.1340990663,-0.1875573993,-0.1528116018,0.0739320889,-0.2589229047,-0.1242828369,-0.4580081105,0.0629987568,0.2549592257,0.2594358921,0.5298959017,0.4289398491,0.1028628424,0.2806044221,-0.0724724978,0.0572922044,0.1174185649,0.1282486916,0.1596287191,0.0161918625,-0.0816258788,-0.1522659957,0.1627645791,0.1623219252,-0.3617706895,0.0090026567,-0.1375187486,0.1970816851,0.3654464483,0.7156856656,-0.1815347672,-0.3790054619,-0.0936338827,-0.1986480951,0.0984760225,0.3893739879,0.3778521121,-0.0309308544,0.0653221011,-0.5064879656,0.1339263022,0.1091979817,0.0512661226,0.0404527225,0.0759120733,0.0293370411,0.008689425,-0.1022983715,-0.1159981489,0.3256934285,-0.2656748891,-0.3963989913,0.137359947,-0.3631493449,0.0508111529,-0.1439990103,0.0500055067,0.366363883,0.6212907434,0.2462342829,0.0420111604,-0.308077246,0.0778114572,0.0899420232,0.3613916039,-0.0054691094,0.1346786171,0.1948350817,0.10071899,0.0852302238,0.0521176606,0.0023271991,-0.542937696,0.1448046714,0.0951064378,0.3429888189,-0.3038514555,-0.4549231231,0.047217641,0.1378372461,-0.1998882294,0.1437949091,-0.0076497574,-0.0868444666,0.0802833959,0.3436453938,0.0880122259,-0.1021173596,-0.2038995028,-0.2310939729,-0.2034747303,-0.0948625132,-0.0629801005,0.1900294423,-0.2003930807,0.6136524081,-0.0917051286,0.0911761746,-0.1107355282,0.0129413661,0.046453923,0.0409467779,0.321911335,-0.1849273741,0.0135516413,0.1199886575,0.1528467536,-0.0941035226,0.2584348321,-0.2343892604,-0.3227925301,-0.1861532629,0.219078213,-0.3183863163,0.0472868122,-0.1684523225,-0.3878966272,0.3156642318,-0.2622767985,-0.0602091402,-0.2444446683,-0.1759607047,-0.138668865,0.3337224126,0.2220809311,-0.1910729706,0.126049161,0.1682969481,-0.1456349194,0.0463749357,0.2319472134,-0.1313233376,0.0905488282,-0.2481278926,-0.0242459103,0.0881310552,-0.5387325287,-0.4547849596,-0.0014980707,0.2015110701,0.0489920415,0.1314415038,-0.1612880677,-0.0724657029,-0.2714241743,-0.2563689649,0.258287251,-0.034631554,0.389313817,-0.2438358366,-0.2965556383,0.0666808411,0.11800237,-0.0492598414,-0.0816775784,0.1693636179,-0.0095629962,0.205123961,0.0532166623,0.0325854793,0.108541891,0.1249753609,-0.1738582253,-0.0426240675,-0.0249788035,-0.1966191828,0.2551011145,-0.0707978681,0.1320103109,0.0627517551,0.178278774,-0.4489194155,-0.1798519492,-0.2049353272,0.0868078247,0.1547742933,0.1201993003,-0.1345562041,0.0315445885,0.097442545,0.1682491302,-0.1691212803,0.0556032844,-0.2289179713,0.2009186,-0.0244281776,-0.1356206685,0.4263866842,0.0483330823,-0.0907463729,0.0649090111,-0.4173122048,0.4893443882,-0.0621393695,0.0347024985,0.0946347788,0.4519070983,0.0079325,-0.3528214693,0.0029319238,0.0297744963,0.1392337382,-0.1301709712,0.2713800669,-0.0893520936,0.0362150632,-0.1035486832,-0.1582131386,-0.1385202408,0.0254937224,-0.0858455822,-0.0036497831,-0.1547219008,0.036334835,0.2815303206,0.2090259939,0.1824414283,0.3501009941,-0.2496142536,0.3917295635,0.0523996353,0.0041363831,0.1314524859,-0.0301868841,-0.0547485016,-0.123130776,0.2424137145,0.4796056449,0.0881364644,0.0710710213,0.1027916372,0.0828066617,-0.0666035786,0.2418119758,0.1984371692,0.3564124107,0.4244209528,-0.1783008277,-0.1144582033,-0.2286888808,0.1582078785,0.1266210526,0.296477139,-0.5625895262,-0.2332416624,-0.3110375106,0.1262899935,-0.0553843863,0.0316452757,-0.3143775761,-0.1802434623,-0.1689627618,0.1081855819,-0.0475393161,0.1951696426,-0.1458839178,0.0477331541,0.3103329241,0.0443977341,0.1516741216,-0.0028465451,-0.1123988777,0.0865353346,0.6746875644,-0.1140086427,0.3397882283,-0.477501303,0.1271569282,-0.3800797462,0.0426583849,-0.0863100588,-0.0893176869,0.2315170616,0.2202654928,0.2755811214,0.0788401291,-0.2032470554,0.0359021463,0.0523619428,-0.4162954092,0.0309536979,-0.0636543036,-0.1320037097,0.1025646105,-0.0423322469,0.0432340465,-0.3893762231,-0.0724421144,-0.0345688462,0.1005237103,0.0962934569,0.0817191377,-0.0508541204,0.0996042937,0.2866059244,-0.1770675182,-0.8880888224,0.3540101349,-0.1670295596,-0.3587040901,-0.0214136578,0.1422802508,0.2374749631,0.0288894922,-0.4529916644,-0.1852105707,-0.2191690356,-0.0274334569,0.0638383254,0.1969694644,0.1162543371,0.0691954195,-0.0823741183,-0.3447431326,-0.2273833901,0.1099680588,-0.2630299628,0.5279319286,-0.0101229204,0.062973924,0.0090144631,0.1534843743,0.411018759,-0.1777416617,0.424354881,0.2426645756,0.3268184364,-0.2149715424,-0.1142090335,0.0165318102,-0.1049812809,0.333722651,0.2043135762,0.0627743155,0.2587611377,-0.1610162407,-0.0275116041,-0.5019219518,-0.3119254112,-0.1261365712,-0.4771397412,0.1579095274,0.0015204475,-0.0457791425,-0.1766192168,-0.2818224728,0.1243380383,0.0594626069,-0.0118080024,0.0062621678,-0.1055888012,0.1124698818,-0.0435051918,0.3818668425,0.1593196094,0.4464143515,0.0325865448,-0.0394472592,-0.1186385751,0.2015811354,0.1771973073,-0.1391032636,0.3249413669,0.0442026407,0.1309691817,-0.1988399774,-0.2067622989,0.2606891096,-0.0236098375,0.3816577494,-0.1196590289,-0.4288542867,0.0298066325,0.2281961292,0.4090623558,-0.1562371254,-0.1394207478,-0.2068855464,-0.3080563843,-0.3402201235,-0.3694739044,0.1338023245,0.2898083329,-0.1530362666,-0.0328774005,-0.1857864559,-0.266718328,0.3404831588,-0.1804640293,0.4067646861,0.0107006691,0.3208973706,0.0213317052,0.2319337279,0.4421767592,0.7271000743,-0.1791524738,-0.21856004,-0.0536129028,-0.1541273147,-0.2087100595,0.0071359742,0.0306307096,-0.2557501495,0.0924119502,-0.0574301518,0.044365827,0.0608743764,0.1258152276,0.0699756593,-0.4000614882,-0.2998196185,0.2181069702,-0.156189397,0.0378430523,0.5064824224,0.0676232576,-0.0924678519,0.0669312477,0.131942898,0.6712408066,-0.1444398016,-0.1336941421,0.0441391766,-0.1740802377,-0.2378405482,0.0318626277,-0.2122766972,-0.3817217052,-0.5945901871,0.0930716395,-0.2704686224,0.0329011157,0.2246014923,0.0661379397,0.0567725748,-0.0167895537,-0.0726982355,0.0699928254,0.3150302172,-0.3102372885,0.0056242491,0.0674703941,0.1505780518,-0.2651734948,0.2959451079,-0.0667072609,-0.1344791949,-0.0533542447,-0.2241715193,-0.6046833396,0.063356787,-0.2882435322,0.2968223691,0.2508451641,0.0494315699,0.157456249,0.4011854529,0.4436883628,0.2695259452,-0.2344846129,0.172759518,0.3623528481,-0.0145823015,-0.0064136069,0.2117148489,0.4534879327,-0.1813885272,-0.2389175892,0.037739303,-0.0976290405,-0.4282446206,0.1780985296,-0.2890240252,0.1226878613,-0.4741533399,-0.0448160693,-0.1610464007,0.0495670922,-0.11707744,0.1266120225,0.3015568852,-0.0258204844,0.3938731849,0.2872494459,-0.4737055004,0.0818112344,0.4354539514,-0.2639756501,-0.0656893551,0.7143592834,-0.0224221386,-0.1156795174,-0.08455538,0.3356194794,0.1899365485,-0.4613088965,0.1759456396,-0.1370148063,0.1676396132,-0.0239781868,0.5160508752,0.4567060471,0.3805165291,0.3290770948,-0.5187920928,-0.1200366989,0.0344009176,-0.1044829786,0.3924307227,-0.1633544564,0.0690328032,0.0379482545,0.0105704777,-0.2830748558,0.1500228792,-0.3312316835,0.0517838746,0.1376151294,0.0577630289,0.091541335,-0.0988202617,0.1534478515,-0.1528942436,-0.3658686876,-0.2194008082,-0.0937628448,0.1155195087,0.0607097931,-0.3785829842,0.1993471235,-0.3264406323,-0.1617555618,-0.1271059513,0.0065265517,0.033089947,-0.036198806,0.027435381,0.2251442075,-0.0241950527,0.0211056527,0.056758821,-0.362577498,0.0497004502,0.0169090182,0.1100955606,0.0522976965,-0.1036025807,0.0623393506,0.0843107179,-0.0241844431,0.1517633647,0.3195523918,-0.2027384341,0.0648035333,0.482668072,0.306938976,0.02040576,-0.2247438431,-0.1236111969,0.1812740713,0.1958637685,-0.3893767595,-0.0456152521,0.0189761855,0.2228658348,-0.00635838,-0.1106350273,0.2290060967,-0.2677126527,-0.0688804686,0.0097189955,0.2057594508,0.2482970953,0.2930920422,0.5799326897,-0.175913319,-0.1093828455,-0.0568373799,-0.1746608615,0.0977744833,0.0135684013,-0.1358999163,0.4929817021,-0.323251158,0.2292830646,-0.2266659588,-0.7084224224,0.3919857144,-0.1421013623,0.0754047036,0.0047916854,0.2124478519,0.2448038906,-0.0497319512,0.1501278579,-0.3210782111,0.4191267788,0.0932898372,-0.0296805296,-0.2051957548,-0.2670288384,0.1096700281,0.0063231117,-0.1057886407,0.1221756935,-0.0712562352,0.1627322435,-0.302936852,-0.1216503754,0.0532986298,-0.1697726995,-0.0355583094,-0.2146684378,0.2357464135,0.2564383149,-0.1772019267,0.3424805999,0.7380926013,0.371566534,0.4830287099,0.1371567845,-0.1722052097,0.1276460886,0.127407223,-0.0393284895,0.2327709645,0.0634962171,-0.0034288804,0.3761826754,0.1473830491,-0.0341962501,-0.3039711714,-0.2243894935,-0.2451120466,0.0394476093,0.2884531021,0.0391374789,0.0263525322,-0.2462791055,0.1971482038,-0.0284639299,-0.0342451669,0.204590708,-0.0086461948,0.1436299533,-0.0167889558,0.095203504,-0.1158063412,0.4963770211,0.3752224743,-0.1486891657,-0.1958691478,-0.3705138862,-0.5975680351,0.2106411606,0.1432383657,-0.1041443199,0.1782567203,-0.0372597538,0.0617663376,-0.1067649201,0.492202729,0.0131060984,-0.196629703,0.0617452897,-0.244164601,-0.2179751843,-0.2281747907,-0.2018176019,-0.0431281328,-0.3614442348,0.1915898472,0.0650836453,0.1579857618,-0.1930200607,0.0880778953,0.0480715744,0.0848006159,0.1467909068,-0.0055310964,0.4740925133,0.053589765,-0.0633306205,-0.0972624496,-0.1911649406,-0.1607998013,0.5104889274,-0.0338879116,0.2085700631,-0.0188264493,-0.0699278265,-0.2319627553,0.3770512044,-0.1640319824,-0.1332943738,-0.4568928182,-0.2069936842,-0.0227444656,-0.2393719554,0.3742827773,-0.108578369,-0.2206009477,0.3657265007,0.0113386977,-0.3828713298,0.4573955536,0.0634475201,0.0366952717,-0.147557646,0.0569182262,-0.1345574558,-0.0479519702,-0.5954985023,0.2696992755,0.4791096151,0.0051987641,-0.0346901678,0.2399044782,-0.1095316857,0.2111077607,0.0491965562,0.1587296426,-0.03080176,-0.2375629991,-0.2700150609,-0.0389962234]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/726","title":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset","comments":"Hi, I got the similar issue for xnli dataset while working on colab with python3.7. \r\n\r\n`nlp.load_dataset(path = 'xnli')`\r\n\r\nThe above command resulted in following issue : \r\n```\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']\r\n```\r\n\r\nAny idea how to fix this ?","body":"Hi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.","comment_length":44,"text":"\"Checksums didn't match for dataset source files\" error while loading openwebtext dataset\nHi,\r\nI have encountered this problem during loading the openwebtext dataset:\r\n```\r\n>>> dataset = load_dataset('openwebtext')\r\nDownloading and preparing dataset openwebtext\/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to \/home\/admin\/.cache\/huggingface\/datasets\/openwebtext\/plain_text\/1.0.0\/5c636399c7155da97c982d0d70ecdce30fbca66a4eb4fc768ad91f8331edac02...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 476, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 536, in _download_and_prepare\r\n self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n File \"\/home\/admin\/workspace\/anaconda3\/envs\/torch1.6-py3.7\/lib\/python3.7\/site-packages\/datasets\/utils\/info_utils.py\", line 39, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\ndatasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/zenodo.org\/record\/3834942\/files\/openwebtext.tar.xz']\r\n```\r\nI think this problem is caused because the released dataset has changed. Or I should download the dataset manually?\r\n\r\nSorry for release the unfinised issue by mistake.\nHi, I got the similar issue for xnli dataset while working on colab with python3.7. \r\n\r\n`nlp.load_dataset(path = 'xnli')`\r\n\r\nThe above command resulted in following issue : \r\n```\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']\r\n```\r\n\r\nAny idea how to fix this ?","embeddings":[-0.4920600653,0.0389875546,-0.0594882704,0.3662634194,0.2443764508,-0.0337870046,0.13654387,0.550494194,-0.00061072,-0.0602816343,-0.0666541532,-0.0310110506,0.0238088965,0.3648073375,-0.2299020737,0.2352990061,-0.1055581272,0.0833998173,-0.3995411098,-0.0401990227,-0.2967323363,0.1188192815,-0.1012961715,-0.1545280963,0.0459318496,0.0663644373,0.2705113292,0.1403153092,-0.2775344551,-0.2145338356,0.1424371004,-0.1216851696,-0.1503286809,0.324788034,-0.0001102053,0.2582596838,0.3656147122,-0.2235869169,-0.2715173662,-0.1560243964,-0.4222053289,-0.1340990663,-0.1875573993,-0.1528116018,0.0739320889,-0.2589229047,-0.1242828369,-0.4580081105,0.0629987568,0.2549592257,0.2594358921,0.5298959017,0.4289398491,0.1028628424,0.2806044221,-0.0724724978,0.0572922044,0.1174185649,0.1282486916,0.1596287191,0.0161918625,-0.0816258788,-0.1522659957,0.1627645791,0.1623219252,-0.3617706895,0.0090026567,-0.1375187486,0.1970816851,0.3654464483,0.7156856656,-0.1815347672,-0.3790054619,-0.0936338827,-0.1986480951,0.0984760225,0.3893739879,0.3778521121,-0.0309308544,0.0653221011,-0.5064879656,0.1339263022,0.1091979817,0.0512661226,0.0404527225,0.0759120733,0.0293370411,0.008689425,-0.1022983715,-0.1159981489,0.3256934285,-0.2656748891,-0.3963989913,0.137359947,-0.3631493449,0.0508111529,-0.1439990103,0.0500055067,0.366363883,0.6212907434,0.2462342829,0.0420111604,-0.308077246,0.0778114572,0.0899420232,0.3613916039,-0.0054691094,0.1346786171,0.1948350817,0.10071899,0.0852302238,0.0521176606,0.0023271991,-0.542937696,0.1448046714,0.0951064378,0.3429888189,-0.3038514555,-0.4549231231,0.047217641,0.1378372461,-0.1998882294,0.1437949091,-0.0076497574,-0.0868444666,0.0802833959,0.3436453938,0.0880122259,-0.1021173596,-0.2038995028,-0.2310939729,-0.2034747303,-0.0948625132,-0.0629801005,0.1900294423,-0.2003930807,0.6136524081,-0.0917051286,0.0911761746,-0.1107355282,0.0129413661,0.046453923,0.0409467779,0.321911335,-0.1849273741,0.0135516413,0.1199886575,0.1528467536,-0.0941035226,0.2584348321,-0.2343892604,-0.3227925301,-0.1861532629,0.219078213,-0.3183863163,0.0472868122,-0.1684523225,-0.3878966272,0.3156642318,-0.2622767985,-0.0602091402,-0.2444446683,-0.1759607047,-0.138668865,0.3337224126,0.2220809311,-0.1910729706,0.126049161,0.1682969481,-0.1456349194,0.0463749357,0.2319472134,-0.1313233376,0.0905488282,-0.2481278926,-0.0242459103,0.0881310552,-0.5387325287,-0.4547849596,-0.0014980707,0.2015110701,0.0489920415,0.1314415038,-0.1612880677,-0.0724657029,-0.2714241743,-0.2563689649,0.258287251,-0.034631554,0.389313817,-0.2438358366,-0.2965556383,0.0666808411,0.11800237,-0.0492598414,-0.0816775784,0.1693636179,-0.0095629962,0.205123961,0.0532166623,0.0325854793,0.108541891,0.1249753609,-0.1738582253,-0.0426240675,-0.0249788035,-0.1966191828,0.2551011145,-0.0707978681,0.1320103109,0.0627517551,0.178278774,-0.4489194155,-0.1798519492,-0.2049353272,0.0868078247,0.1547742933,0.1201993003,-0.1345562041,0.0315445885,0.097442545,0.1682491302,-0.1691212803,0.0556032844,-0.2289179713,0.2009186,-0.0244281776,-0.1356206685,0.4263866842,0.0483330823,-0.0907463729,0.0649090111,-0.4173122048,0.4893443882,-0.0621393695,0.0347024985,0.0946347788,0.4519070983,0.0079325,-0.3528214693,0.0029319238,0.0297744963,0.1392337382,-0.1301709712,0.2713800669,-0.0893520936,0.0362150632,-0.1035486832,-0.1582131386,-0.1385202408,0.0254937224,-0.0858455822,-0.0036497831,-0.1547219008,0.036334835,0.2815303206,0.2090259939,0.1824414283,0.3501009941,-0.2496142536,0.3917295635,0.0523996353,0.0041363831,0.1314524859,-0.0301868841,-0.0547485016,-0.123130776,0.2424137145,0.4796056449,0.0881364644,0.0710710213,0.1027916372,0.0828066617,-0.0666035786,0.2418119758,0.1984371692,0.3564124107,0.4244209528,-0.1783008277,-0.1144582033,-0.2286888808,0.1582078785,0.1266210526,0.296477139,-0.5625895262,-0.2332416624,-0.3110375106,0.1262899935,-0.0553843863,0.0316452757,-0.3143775761,-0.1802434623,-0.1689627618,0.1081855819,-0.0475393161,0.1951696426,-0.1458839178,0.0477331541,0.3103329241,0.0443977341,0.1516741216,-0.0028465451,-0.1123988777,0.0865353346,0.6746875644,-0.1140086427,0.3397882283,-0.477501303,0.1271569282,-0.3800797462,0.0426583849,-0.0863100588,-0.0893176869,0.2315170616,0.2202654928,0.2755811214,0.0788401291,-0.2032470554,0.0359021463,0.0523619428,-0.4162954092,0.0309536979,-0.0636543036,-0.1320037097,0.1025646105,-0.0423322469,0.0432340465,-0.3893762231,-0.0724421144,-0.0345688462,0.1005237103,0.0962934569,0.0817191377,-0.0508541204,0.0996042937,0.2866059244,-0.1770675182,-0.8880888224,0.3540101349,-0.1670295596,-0.3587040901,-0.0214136578,0.1422802508,0.2374749631,0.0288894922,-0.4529916644,-0.1852105707,-0.2191690356,-0.0274334569,0.0638383254,0.1969694644,0.1162543371,0.0691954195,-0.0823741183,-0.3447431326,-0.2273833901,0.1099680588,-0.2630299628,0.5279319286,-0.0101229204,0.062973924,0.0090144631,0.1534843743,0.411018759,-0.1777416617,0.424354881,0.2426645756,0.3268184364,-0.2149715424,-0.1142090335,0.0165318102,-0.1049812809,0.333722651,0.2043135762,0.0627743155,0.2587611377,-0.1610162407,-0.0275116041,-0.5019219518,-0.3119254112,-0.1261365712,-0.4771397412,0.1579095274,0.0015204475,-0.0457791425,-0.1766192168,-0.2818224728,0.1243380383,0.0594626069,-0.0118080024,0.0062621678,-0.1055888012,0.1124698818,-0.0435051918,0.3818668425,0.1593196094,0.4464143515,0.0325865448,-0.0394472592,-0.1186385751,0.2015811354,0.1771973073,-0.1391032636,0.3249413669,0.0442026407,0.1309691817,-0.1988399774,-0.2067622989,0.2606891096,-0.0236098375,0.3816577494,-0.1196590289,-0.4288542867,0.0298066325,0.2281961292,0.4090623558,-0.1562371254,-0.1394207478,-0.2068855464,-0.3080563843,-0.3402201235,-0.3694739044,0.1338023245,0.2898083329,-0.1530362666,-0.0328774005,-0.1857864559,-0.266718328,0.3404831588,-0.1804640293,0.4067646861,0.0107006691,0.3208973706,0.0213317052,0.2319337279,0.4421767592,0.7271000743,-0.1791524738,-0.21856004,-0.0536129028,-0.1541273147,-0.2087100595,0.0071359742,0.0306307096,-0.2557501495,0.0924119502,-0.0574301518,0.044365827,0.0608743764,0.1258152276,0.0699756593,-0.4000614882,-0.2998196185,0.2181069702,-0.156189397,0.0378430523,0.5064824224,0.0676232576,-0.0924678519,0.0669312477,0.131942898,0.6712408066,-0.1444398016,-0.1336941421,0.0441391766,-0.1740802377,-0.2378405482,0.0318626277,-0.2122766972,-0.3817217052,-0.5945901871,0.0930716395,-0.2704686224,0.0329011157,0.2246014923,0.0661379397,0.0567725748,-0.0167895537,-0.0726982355,0.0699928254,0.3150302172,-0.3102372885,0.0056242491,0.0674703941,0.1505780518,-0.2651734948,0.2959451079,-0.0667072609,-0.1344791949,-0.0533542447,-0.2241715193,-0.6046833396,0.063356787,-0.2882435322,0.2968223691,0.2508451641,0.0494315699,0.157456249,0.4011854529,0.4436883628,0.2695259452,-0.2344846129,0.172759518,0.3623528481,-0.0145823015,-0.0064136069,0.2117148489,0.4534879327,-0.1813885272,-0.2389175892,0.037739303,-0.0976290405,-0.4282446206,0.1780985296,-0.2890240252,0.1226878613,-0.4741533399,-0.0448160693,-0.1610464007,0.0495670922,-0.11707744,0.1266120225,0.3015568852,-0.0258204844,0.3938731849,0.2872494459,-0.4737055004,0.0818112344,0.4354539514,-0.2639756501,-0.0656893551,0.7143592834,-0.0224221386,-0.1156795174,-0.08455538,0.3356194794,0.1899365485,-0.4613088965,0.1759456396,-0.1370148063,0.1676396132,-0.0239781868,0.5160508752,0.4567060471,0.3805165291,0.3290770948,-0.5187920928,-0.1200366989,0.0344009176,-0.1044829786,0.3924307227,-0.1633544564,0.0690328032,0.0379482545,0.0105704777,-0.2830748558,0.1500228792,-0.3312316835,0.0517838746,0.1376151294,0.0577630289,0.091541335,-0.0988202617,0.1534478515,-0.1528942436,-0.3658686876,-0.2194008082,-0.0937628448,0.1155195087,0.0607097931,-0.3785829842,0.1993471235,-0.3264406323,-0.1617555618,-0.1271059513,0.0065265517,0.033089947,-0.036198806,0.027435381,0.2251442075,-0.0241950527,0.0211056527,0.056758821,-0.362577498,0.0497004502,0.0169090182,0.1100955606,0.0522976965,-0.1036025807,0.0623393506,0.0843107179,-0.0241844431,0.1517633647,0.3195523918,-0.2027384341,0.0648035333,0.482668072,0.306938976,0.02040576,-0.2247438431,-0.1236111969,0.1812740713,0.1958637685,-0.3893767595,-0.0456152521,0.0189761855,0.2228658348,-0.00635838,-0.1106350273,0.2290060967,-0.2677126527,-0.0688804686,0.0097189955,0.2057594508,0.2482970953,0.2930920422,0.5799326897,-0.175913319,-0.1093828455,-0.0568373799,-0.1746608615,0.0977744833,0.0135684013,-0.1358999163,0.4929817021,-0.323251158,0.2292830646,-0.2266659588,-0.7084224224,0.3919857144,-0.1421013623,0.0754047036,0.0047916854,0.2124478519,0.2448038906,-0.0497319512,0.1501278579,-0.3210782111,0.4191267788,0.0932898372,-0.0296805296,-0.2051957548,-0.2670288384,0.1096700281,0.0063231117,-0.1057886407,0.1221756935,-0.0712562352,0.1627322435,-0.302936852,-0.1216503754,0.0532986298,-0.1697726995,-0.0355583094,-0.2146684378,0.2357464135,0.2564383149,-0.1772019267,0.3424805999,0.7380926013,0.371566534,0.4830287099,0.1371567845,-0.1722052097,0.1276460886,0.127407223,-0.0393284895,0.2327709645,0.0634962171,-0.0034288804,0.3761826754,0.1473830491,-0.0341962501,-0.3039711714,-0.2243894935,-0.2451120466,0.0394476093,0.2884531021,0.0391374789,0.0263525322,-0.2462791055,0.1971482038,-0.0284639299,-0.0342451669,0.204590708,-0.0086461948,0.1436299533,-0.0167889558,0.095203504,-0.1158063412,0.4963770211,0.3752224743,-0.1486891657,-0.1958691478,-0.3705138862,-0.5975680351,0.2106411606,0.1432383657,-0.1041443199,0.1782567203,-0.0372597538,0.0617663376,-0.1067649201,0.492202729,0.0131060984,-0.196629703,0.0617452897,-0.244164601,-0.2179751843,-0.2281747907,-0.2018176019,-0.0431281328,-0.3614442348,0.1915898472,0.0650836453,0.1579857618,-0.1930200607,0.0880778953,0.0480715744,0.0848006159,0.1467909068,-0.0055310964,0.4740925133,0.053589765,-0.0633306205,-0.0972624496,-0.1911649406,-0.1607998013,0.5104889274,-0.0338879116,0.2085700631,-0.0188264493,-0.0699278265,-0.2319627553,0.3770512044,-0.1640319824,-0.1332943738,-0.4568928182,-0.2069936842,-0.0227444656,-0.2393719554,0.3742827773,-0.108578369,-0.2206009477,0.3657265007,0.0113386977,-0.3828713298,0.4573955536,0.0634475201,0.0366952717,-0.147557646,0.0569182262,-0.1345574558,-0.0479519702,-0.5954985023,0.2696992755,0.4791096151,0.0051987641,-0.0346901678,0.2399044782,-0.1095316857,0.2111077607,0.0491965562,0.1587296426,-0.03080176,-0.2375629991,-0.2700150609,-0.0389962234]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/724","title":"need to redirect \/nlp to \/datasets and remove outdated info","comments":"Should be fixed now: \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/35882\/95917301-040b0600-0d78-11eb-9655-c4ac0e788089.png)\r\n\r\nNot sure I understand what you mean by the second part?\r\n","body":"It looks like the website still has all the `nlp` data, e.g.: https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\n\r\nshould probably redirect to: https:\/\/huggingface.co\/datasets\/wikihow\r\n\r\nalso for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).","comment_length":16,"text":"need to redirect \/nlp to \/datasets and remove outdated info\nIt looks like the website still has all the `nlp` data, e.g.: https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\n\r\nshould probably redirect to: https:\/\/huggingface.co\/datasets\/wikihow\r\n\r\nalso for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).\nShould be fixed now: \r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/35882\/95917301-040b0600-0d78-11eb-9655-c4ac0e788089.png)\r\n\r\nNot sure I understand what you mean by the second part?\r\n","embeddings":[0.2948436141,0.110556297,-0.0633747578,0.1094637066,-0.0713283047,0.0353955403,-0.260299772,0.6290499568,-0.2917530835,-0.4541862309,-0.1458730847,0.2195103467,0.2947607338,0.0850115344,0.1424258053,-0.3171723187,0.0133651607,0.0606414154,-0.0077770688,-0.1089684814,-0.186075002,0.0806226283,-0.403195709,0.116498366,-0.0495447777,0.0418990068,-0.2173669338,0.3226778805,-0.0874300599,-0.42714867,0.1224372461,0.0153161846,-0.0209404994,0.1970620006,-0.0001060939,-0.3152425885,0.3930561543,0.1965439171,-0.3718169928,-0.0397977009,-0.2657049596,-0.5014023781,-0.0838163868,-0.1341035664,0.0758307502,-0.1319982708,0.0686012283,0.3240991533,0.4053322673,0.4329928458,0.2685536742,-0.2723818123,-0.1567654908,0.1352109313,0.0452874862,0.3373916447,0.1736284792,-0.1308058947,0.1029123887,-0.0435013175,-0.1377879679,0.6198434234,-0.1445967555,-0.272770226,0.0824484453,-0.0814060494,0.0386756659,-0.1081278846,-0.0297006462,0.1601059139,0.5080628991,-0.1142880544,-0.0372500196,-0.1996771693,0.002060547,-0.2207480073,0.2206597626,0.1426458508,-0.155316025,0.1439509392,-0.2525389194,-0.644289434,-0.2582044899,0.3472139537,0.0650897548,0.5357505679,0.0393282659,-0.050986398,0.0237035751,-0.0212874264,0.2273884416,0.001841528,-0.151028946,0.2054466754,0.2761593759,-0.1788726002,-0.0123925805,0.4630789459,0.0318617187,-0.0050338199,-0.005989654,0.1522022784,-0.2559991777,-0.060855072,0.1496139318,0.0887265354,0.1551512778,-0.0099968156,0.1412384659,0.2226877958,0.1227829456,0.1127246916,0.2170561105,-0.0220430028,-0.4140966833,0.0599268787,0.3093501925,-0.3077695072,-0.0351443216,0.010313062,-0.0311403442,-0.156090647,0.046028506,0.0781720728,-0.2404728979,0.2759421468,0.2424458712,0.0025035948,-0.2576318681,-0.1979256719,0.0392020196,0.1484769732,-0.5195070505,-0.0037097603,0.1793030053,0.3031880558,0.1963305771,0.2088835537,-0.4914647341,-0.0103206355,-0.1324130148,0.0879880488,0.3486702144,-0.0020263144,-0.1399039179,0.1751738936,-0.0648551062,-0.1758546531,-0.1601034403,0.1516988873,-0.3903276026,-0.1063865125,-0.1017534062,0.2205360234,0.0132216113,-0.0695660561,0.1046729833,0.522659421,-0.0152454171,-0.2866698802,0.0681395009,0.249386549,-0.0525919236,-0.160615474,0.0980929881,0.4052127004,0.0794249922,-0.0646125898,-0.0768415034,-0.0301984455,-0.0066010901,0.1782200783,-0.137875095,0.0530477688,-0.1463340968,-0.2217721194,0.6264216304,-0.3208308518,-0.0289240386,0.2638678253,-0.0541572273,0.1031716317,-0.2276762724,0.0109351939,-0.1501184702,-0.2054261863,-0.077296935,0.2869209647,0.1903690845,-0.0708118603,-0.5612444282,-0.1219867617,0.0364453308,-0.145921424,-0.118674621,0.2481559962,-0.128270492,0.1427189112,0.5751604438,0.1089756563,0.2362707108,0.0461571813,-0.0069270916,0.2162114084,-0.0232686903,0.1955585331,-0.3031808436,-0.2531629801,-0.2206091434,0.2218620479,0.3798179924,-0.2203883529,-0.2938579917,-0.3824447095,-0.1047745347,-0.4056072235,0.2608014643,0.0412185416,0.0370696932,0.1344967782,-0.2119551301,0.106041573,0.0213722773,0.1596153378,-0.281570524,0.2219356149,-0.2116109282,0.1130198613,0.0817909911,0.3167993128,0.3062698841,0.0317722075,0.0738056153,0.3021827042,-0.2413655072,0.2140286416,0.3557474315,0.0032243282,0.3906036615,-0.3758459985,0.13474828,0.2379235178,-0.2560378313,0.1230783686,-0.3258954287,0.014580911,0.1396802366,0.3352673948,0.0611242428,0.2751969993,0.0969419181,-0.4506769776,0.0199287087,0.0352055691,0.035097219,0.1807038337,-0.1705889702,0.0479946248,-0.2274552584,0.5184099078,-0.1327476501,-0.0902360827,-0.1630612463,0.4421414137,-0.1375371665,-0.4395441413,0.1946420968,0.1848457158,-0.2379846722,0.2428795099,0.1344380826,0.2076068968,0.0131123681,-0.1935088336,0.3632593453,0.1367352903,0.013725807,-0.1369606555,0.0938369855,0.2238072306,-0.1558355689,-0.1874810159,-0.1377896518,0.3766480684,-0.200019002,-0.0217866879,-0.2344863266,-0.8449617624,-0.266662091,-0.0810860023,-0.49774459,-0.4589166939,-0.0094369287,-0.019099053,-0.2085675299,0.1083395779,0.0919973403,0.2354967743,-0.2721153796,0.3169374168,-0.535997808,-0.3892775774,-0.1888776273,0.1811912209,0.1817516685,0.3827549219,0.1627145112,-0.0475809872,0.0194041189,-0.2012942135,-0.6216056347,-0.0174883623,0.135520786,-0.0172693375,0.0692636371,0.3452521265,-0.1315455288,-0.0921540037,0.1151114926,-0.0690847859,0.0220816936,-0.2542694807,0.0454222411,0.2809758782,-0.0525238067,-0.2910253704,-0.2359342873,-0.3109239042,0.021790389,0.1182686761,0.3605513871,0.4227543175,0.0719411299,-0.2329274565,-0.1364896595,0.1704229563,-0.3226844966,-0.2482672483,0.0414103344,-0.1580933034,-0.1972006261,-0.0252275281,0.064657025,0.0775933489,-0.049986288,-0.6025821567,-0.1269472986,-0.1313961595,-0.1286750287,0.0709887221,0.1048422158,0.283921659,0.2563126683,-0.3228650391,0.1037822217,-0.1825142652,-0.1681531817,0.007123149,0.4162358046,-0.0056421845,-0.0713063926,0.2042766809,0.2970851362,0.1327783018,0.2455477268,0.1495489776,0.2490274608,0.3893796802,0.1003492847,-0.2246986032,0.1389163136,-0.0300456882,0.1968904734,0.2607525289,0.2953700423,-0.0710903779,-0.3592978418,-0.2804770768,-0.5068995357,-0.3678104877,0.2198278159,-0.0039306195,0.1960823983,0.4365545809,-0.0586107522,-0.0607510805,-0.3621489406,0.0613125227,0.3835931122,0.0864200369,-0.0400692485,-0.5179490447,-0.2183751315,-0.3404259384,0.1040318757,-0.016234586,-0.0700975507,-0.1219053715,-0.0853976831,-0.0417905748,0.0335987918,0.0560586341,-0.2636699378,-0.1103122756,-0.068856962,0.3781457841,-0.2133826613,0.3772484958,0.0054370333,0.2163894922,0.4170415998,0.2548283339,-0.0886950493,-0.294691056,0.4732964039,-0.0484016575,-0.1588539183,0.1752280444,-0.0133169619,-0.2834280133,-0.2460772097,0.0070943567,-0.1668279767,-0.1399956495,-0.1437566131,0.1650793552,0.3778227866,0.06473355,0.1096120924,0.0086405575,0.1039742902,0.2948508561,-0.3788731992,0.0881222039,0.2247514427,0.1009909734,0.3582809567,-0.1854430437,0.0883201063,-0.4033206701,0.1375195533,0.5038176179,0.5648974776,0.4843834937,0.1057311743,-0.0381659493,0.0434280224,-0.2725447416,0.1219573915,-0.0363102145,-0.0668789148,-0.4212539196,-0.2740131617,0.4274557233,-0.2244532406,-0.3345918059,0.1839265972,0.3845708668,-0.3714579046,0.0986312404,-0.0541161969,1.303673625,0.1177560836,0.2643562853,0.2851367593,-0.3264497519,0.6118419766,0.135192275,0.180323422,-0.3115263283,0.2939246297,0.0758846924,0.1033245027,0.0355616063,-0.0169036649,-0.27623564,0.146912992,-0.0094638234,-0.1048598364,0.3148038089,0.5139609575,0.0888590291,-0.0101270871,-0.4954625666,0.2104873359,-0.3010275066,0.1579083502,-0.1387442052,-0.2165047377,-0.1414223164,-0.0079964856,0.1591984779,0.1593579948,0.1233633608,-0.0102320258,-0.1897466928,-0.406208992,-0.0905774981,0.1806754321,0.1331230551,0.2241073549,-0.2983700931,0.2756780982,-0.1352891475,-0.0084200064,0.1512207538,0.0875727907,0.0199186634,-0.0425280109,-0.2699320912,-0.0053209444,0.0387092605,-0.0304473024,-0.40817976,-0.2474005073,0.2023917288,-0.1728231758,0.1758574545,0.3450033367,-0.3697670102,-0.2266703248,0.166582942,0.1998155117,-0.0727843791,-0.1729570925,0.4413701594,-0.0301595107,-0.0509580299,0.1921456307,0.1212329492,-0.3229702115,0.3000794649,0.1150268912,-0.0765450895,-0.348790437,0.112345919,-0.0191056635,-0.2420697063,-0.0497961678,-0.2220897526,-0.0446446314,-0.0433059037,0.1412590295,-0.1310528368,0.1130351871,0.0411256962,-0.2262660712,-0.285675019,0.3485090137,-0.123034358,-0.0245348401,-0.1657323688,0.1705522835,0.2855715156,0.1165855005,-0.3988348246,0.2132160813,0.1558334231,0.3843374848,0.0738096908,-0.1693646759,-0.033110939,-0.1091758013,0.0152047435,-0.3327125311,-0.0191912986,-0.31116575,-0.1570253074,0.0958077535,0.1315570921,-0.4206082225,-0.1938262731,0.0048293248,0.3410654962,-0.0282674246,-0.3932342827,-0.0528078042,0.1041390523,0.3544598818,-0.0349465497,-0.1348582655,0.2051250637,0.1046093553,-0.0054755635,0.3132798672,0.0977767184,0.098578833,-0.1884911954,-0.1168107316,-0.0280481409,0.2046812177,0.0945469663,0.3071016967,0.1823250502,-0.0172695071,0.084300518,0.1988328993,0.2487662882,0.3345123529,-0.1385534704,-0.1020023674,0.2987512946,0.2516271472,-0.1525026709,0.306453675,0.4109356403,-0.2264778465,-0.140647158,0.04764653,0.0535164475,-0.1912403554,-0.0045951898,0.16637218,0.3280133307,-0.3724982142,0.1765822023,-0.0227067415,-0.1544036269,-0.1868982315,0.0649691448,-0.0025248469,-0.0139820492,0.0510520041,0.1612492502,0.07209865,0.2448982,0.1589886993,-0.0305771735,0.1322415322,-0.1021421999,0.2460074872,0.0474035442,0.2853547633,0.1320997179,0.3234423697,0.1637513787,-0.5597475767,0.1861124039,0.2872393727,-0.0840739459,-0.0687477961,-0.2393766344,-0.0950438306,0.0250345897,-0.1406839043,-0.0047860513,-0.3113849461,0.2902386487,0.005597502,-0.2713837624,-0.616717577,0.0540416799,0.2305917889,0.3701771796,0.130607143,-0.016153587,0.0870969221,0.1048497856,0.1810632795,0.3856883049,0.2037330419,0.089304097,-0.2723139524,0.1367505193,-0.0468956754,-0.1020632833,-0.0629247278,0.1047944129,0.4423394203,-0.101404503,0.0856407508,0.1887861639,-0.160644412,0.0905619115,0.2436363548,-0.0757840499,0.1476674974,0.1856596023,0.2298227996,-0.3325279653,-0.1175151616,-0.260014236,-0.37333408,-0.1117496714,0.2153915316,-0.1159693748,0.1428100318,-0.0462600812,0.1098516434,0.2666891813,0.3352409005,0.4374489188,0.0977304876,-0.2789232731,-0.4578016698,-0.5455568433,-0.1064728424,0.0231017098,-0.2141452581,-0.0622801781,0.083075285,-0.3161224425,0.1527830064,-0.1049753428,-0.1918206513,0.1794049442,-0.2479571998,-0.3061390519,0.3506569564,0.1376372129,0.4061353505,-0.0138249956,-0.0866429284,0.2708300352,-0.1054262742,0.0904941559,0.1379040182,0.1030693203,0.0266215727,0.2724760771,0.0227294937,0.3079683781,0.0580761842,-0.1209438667,-0.4421961904,-0.4232117236,-0.2431793809,-0.073003076,0.3129670322,0.1411991417,0.3724712729,-0.1217118129,-0.0141428774,-0.0595383905,0.2734529376,0.1081346869,0.2610917687,-0.2706222534,0.1001240611,0.3021621704,0.2522706687,-0.0820597112,0.2043442279,-0.0370109677,0.1082299501,-0.4340156615,-0.3835900426,0.4355790019,-0.2187874764,-0.1724579781,0.0330864117,0.3791277707,-0.0644902289,0.2866306305,-0.6331457496,-0.2555653155,0.2507276237,0.0525368303,-0.2529153824,0.2137627304,-0.368619144,-0.3414417803,0.0525199138,0.0384796746,0.2273092419,-0.4464933872,0.1576785892,-0.1497326344]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/724","title":"need to redirect \/nlp to \/datasets and remove outdated info","comments":"Thank you!\r\n\r\n> Not sure I understand what you mean by the second part?\r\n\r\nCompare the 2:\r\n* https:\/\/huggingface.co\/datasets\/wikihow\r\n* https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\nCan you see the difference? 2nd has formatting, 1st doesn't.\r\n","body":"It looks like the website still has all the `nlp` data, e.g.: https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\n\r\nshould probably redirect to: https:\/\/huggingface.co\/datasets\/wikihow\r\n\r\nalso for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).","comment_length":31,"text":"need to redirect \/nlp to \/datasets and remove outdated info\nIt looks like the website still has all the `nlp` data, e.g.: https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\n\r\nshould probably redirect to: https:\/\/huggingface.co\/datasets\/wikihow\r\n\r\nalso for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).\nThank you!\r\n\r\n> Not sure I understand what you mean by the second part?\r\n\r\nCompare the 2:\r\n* https:\/\/huggingface.co\/datasets\/wikihow\r\n* https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\nCan you see the difference? 2nd has formatting, 1st doesn't.\r\n","embeddings":[0.3347691298,-0.031311173,-0.0245432202,0.122392565,-0.1146651059,0.0453228764,-0.3248898685,0.6779083014,-0.4332506657,-0.5133858323,-0.1581162512,0.0308898725,0.2777537107,0.0370305218,0.0912264884,-0.3001837134,0.083900176,0.068306528,0.108894676,-0.0865245089,-0.185257107,0.076251559,-0.4062897265,0.112150088,-0.0114804832,0.1721997857,-0.2518543899,0.2978983521,-0.0522664674,-0.4027541578,0.2221582085,-0.0079392707,-0.0050460813,0.3382590413,-0.0001157864,-0.4116526842,0.4116846025,0.1879996806,-0.4584499002,-0.0305531826,-0.4210923314,-0.4628536999,-0.0959183425,-0.0723487809,0.0769471079,-0.0744698048,-0.0712039396,0.3125736713,0.344877243,0.4643400609,0.1784186959,-0.2903703153,-0.2416379601,0.1717674881,0.0189941525,0.420421809,0.2907690108,-0.1033793986,0.1266770363,0.0224710423,0.0177703183,0.6001993418,-0.2633239031,-0.2864018083,0.0878261626,-0.1737945527,0.01004091,0.018770678,0.0099457484,0.1783487201,0.5071709156,-0.1586164236,-0.0345927998,-0.1617970318,-0.0723986849,-0.3024497628,0.1925821304,0.0792358518,-0.1341700256,0.1910063475,-0.274197489,-0.6634989381,-0.2567833662,0.371929884,0.0737498626,0.5255467892,0.0017782624,-0.0485953577,-0.0253332872,0.0524695925,0.3299182653,-0.1322030276,-0.149890855,0.1076483727,0.3520326912,-0.2275372893,-0.1650887281,0.3888019919,-0.0739497617,0.0860262662,-0.0450351946,0.1166939363,-0.2130912244,-0.0696077719,0.0730750188,0.0922397524,0.2408150136,0.0387114286,0.0163716115,0.2204917669,0.0733710229,0.0563520864,0.2624738812,0.0330741331,-0.3904784322,0.1451823413,0.3424996138,-0.3545633256,-0.1334958076,-0.0107321003,-0.130911991,-0.2028384209,0.1030148193,0.0233705062,-0.2182503939,0.239734292,0.2594830394,0.0784377381,-0.2327984571,-0.2440481782,0.1151465923,0.1520035565,-0.4599367976,-0.0848177969,0.1431110054,0.2702641189,0.1501922607,0.2429584712,-0.5883119106,-0.0873502269,-0.1672285497,0.0308704153,0.4714208841,-0.021943152,-0.213327229,0.254794389,0.0051450529,-0.0993284062,-0.2255294174,0.2119438797,-0.3909611702,-0.1834077388,-0.0727772787,0.1527855247,0.0505205728,-0.0792206377,0.0594301336,0.5098864436,0.0243786406,-0.3396795392,0.1360300928,0.245713532,-0.0578637496,-0.1810048223,0.1157800257,0.5219936371,0.0678177252,-0.053227894,-0.1457304209,0.0465020239,0.1242175773,0.1956205517,-0.0921932757,0.0991828442,-0.0882086456,-0.3709585369,0.6637300849,-0.351118505,-0.0227153935,0.3351544142,-0.0081379581,0.1604323983,-0.2453326732,-0.0492001548,-0.1875029504,-0.2134551257,-0.0744737983,0.224197045,0.3232367337,-0.1579634398,-0.5133376122,-0.1274110973,0.0353278667,-0.2628821135,-0.1963531822,0.2379519343,-0.1808660626,0.0626831427,0.6130368114,0.2045864165,0.3268550038,0.0366968066,-0.0764690936,0.2783766091,-0.0655194446,0.164321959,-0.2536492646,-0.3068724275,-0.2187329978,0.0782774016,0.3773010671,-0.1764589995,-0.3194611371,-0.4541701078,-0.1285852939,-0.370475024,0.1609469652,0.0367840119,0.0669013262,0.0918622017,-0.2161007077,0.1195017323,0.122188963,0.1626618803,-0.1556651294,0.234363243,-0.1073224619,0.1185677797,0.0680757836,0.3675707281,0.3903980255,0.0310998894,0.033188317,0.3623469174,-0.2315732539,0.1050688922,0.3430273831,-0.0005020927,0.2701265514,-0.3319279253,0.1148291528,0.3078564703,-0.2995312512,0.1157678738,-0.2941272855,-0.0970844701,0.1080173403,0.4377196729,0.0486137755,0.2380438447,-0.0427469164,-0.4784863889,0.0134781944,0.012924836,0.0927122831,0.1475967914,-0.1852423102,0.1496761441,-0.1700379252,0.489030242,-0.2432168871,-0.2002646476,-0.1469666064,0.4490039349,-0.2449116707,-0.5029028058,0.1238067895,0.1460931897,-0.1699922383,0.1770005077,0.1778144538,0.1586057395,0.1773630828,-0.1300580055,0.391132921,0.1921301335,-0.1077755764,-0.1821141541,0.0611279532,0.2805323601,0.0116043091,-0.2150404304,-0.1216224656,0.3242534399,-0.1866368651,0.0014892109,-0.3264100552,-0.8657703996,-0.3520624638,-0.2004303336,-0.5172258019,-0.4104232788,-0.0949900299,-0.1143295988,-0.1342767477,0.0450775363,0.1265208274,0.2379600853,-0.2683819234,0.1645227522,-0.5306436419,-0.2966865897,-0.2325037271,0.0822637007,0.1692272127,0.4310949147,0.1023434475,-0.0135781635,0.1317180097,-0.2384172827,-0.5461174846,-0.0772489086,0.1511229873,-0.1104475632,0.0499580279,0.3271476328,-0.0515405722,-0.0389799103,0.1058775336,0.0922546014,0.0479799062,-0.1891593039,0.1006740928,0.3573777676,0.0506377891,-0.1992050409,-0.2320473641,-0.2316874117,-0.1065358222,0.0011562388,0.303037852,0.3275019526,0.0424165204,-0.2985092103,-0.1533736587,0.1980222911,-0.2720221877,-0.2821151614,-0.0865133554,-0.0520646535,-0.0480839573,-0.0406525321,0.1638768762,-0.0399586968,0.0063309935,-0.5437178612,0.0172786396,-0.1721352935,-0.2262423187,0.0743889436,0.0549087524,0.3278782964,0.2925352156,-0.2564878464,0.1380576938,-0.221404165,-0.2019814253,-0.0032891636,0.470210284,0.0355643556,-0.080441989,0.2707927525,0.224577263,0.1259835213,0.24359034,0.0709466189,0.3061742187,0.3696184158,0.0687897056,-0.2160182595,0.1123954654,-0.0352278799,0.2803629339,0.2662485242,0.3352378607,0.0372058749,-0.3397895694,-0.3865875304,-0.5043460727,-0.3230564296,0.2027923912,-0.1036222577,0.2499017566,0.4611448646,-0.0946934,-0.0032176673,-0.376596421,0.0748055801,0.4609953165,0.1067184731,0.0093266657,-0.5229618549,-0.2309429348,-0.2563528419,0.2096590698,-0.0110528087,-0.0789282098,-0.0681877285,-0.1151459292,-0.0564891994,0.1711410731,0.0056753843,-0.2897023261,-0.1365116984,-0.1385129541,0.4089796543,-0.2733708918,0.3602390587,0.0629552826,0.2372463644,0.4312343001,0.2234023362,-0.1197765172,-0.3254252672,0.5697222352,0.0645321682,-0.0614050664,0.0964406878,-0.0524778627,-0.2221926749,-0.1560297608,0.0855973586,-0.1968943328,-0.0881664231,-0.1269774288,0.2383783013,0.3612759709,0.082232587,0.1876226217,-0.1319362968,0.0705622658,0.1995438784,-0.458871603,0.1354636848,0.2541296482,0.2300762236,0.2352903634,-0.2480003089,0.0296633448,-0.4350307882,0.1664230525,0.5890612006,0.5690331459,0.5141723156,0.1899053305,-0.0087944074,-0.0551820621,-0.2320433259,0.0557170436,-0.1033248082,0.028269846,-0.6692448854,-0.2759590149,0.343464762,-0.201421693,-0.2996373773,0.1327136755,0.368013978,-0.3256421089,0.1358929873,-0.138660565,1.4525067806,0.0161544736,0.3074936271,0.2622662783,-0.3694678545,0.5940526128,0.1238385588,0.2053142637,-0.2839100659,0.2496680021,0.0036739632,0.0618798137,0.1228402331,0.0623341836,-0.2701880038,0.2022422552,-0.2004921138,0.0299870986,0.3993929029,0.4282920957,0.1693124771,0.1005339622,-0.5160793066,0.1123960912,-0.2810961008,0.1808936,-0.1239767522,-0.1440436244,-0.2226492614,0.0313794576,0.1877167672,0.0839281306,0.1421214193,-0.0183626954,-0.2806868553,-0.5062955022,-0.1181014702,0.1595148295,0.1678707153,0.3043162525,-0.2099966705,0.2386780083,-0.0549512058,0.0356582366,0.2162674218,-0.0535351336,-0.0558998287,0.0556209236,-0.2328428328,0.0376837626,0.0312075447,-0.1029536426,-0.3474051356,-0.1976026297,0.218541339,-0.2566686571,0.0584217086,0.396440953,-0.3623544276,-0.2125409544,0.0836696401,0.1809797138,0.0384856239,-0.1831214279,0.3429775238,-0.0922758281,-0.058297772,0.1865049005,0.1581071764,-0.3483942449,0.3420065045,0.03138832,-0.0707622245,-0.2912493348,0.1405386627,-0.0006707911,-0.1967129558,0.0280776024,-0.2834335268,0.0484337062,0.0363944583,0.0581098311,-0.2070178539,0.152587831,0.0304014515,-0.1802909523,-0.2243570834,0.2958810329,-0.0278944168,-0.0560011864,-0.1810217947,0.22586371,0.2381573617,0.0413483381,-0.2912551761,0.3446398079,0.2932645977,0.4035772383,0.0138337016,-0.1386965364,0.083667241,-0.1347232908,-0.0722849369,-0.4095422029,0.091105938,-0.2318191975,-0.135504052,0.1214023158,0.0832747668,-0.4337322414,-0.2626564801,-0.0199546646,0.4507327676,-0.0571439341,-0.4337203503,0.0790682212,0.136782214,0.3330495954,-0.0103830332,-0.2221971899,0.0730658248,0.207111448,-0.0344449952,0.2445626408,0.0574368872,0.1464105248,-0.1366048753,-0.0882486328,-0.063011758,0.2419251502,0.2073027045,0.248589918,0.1636388004,0.0078867013,0.0967051834,0.1874339581,0.2349242121,0.3742614686,-0.1672325879,-0.190737471,0.2022379637,0.1570818275,-0.0633992627,0.3117002249,0.3845291436,-0.2091356963,-0.195726037,-0.017918678,0.0369765162,-0.1840925962,-0.0103482688,0.1297141165,0.3523136973,-0.3151237071,0.1350959241,0.028424859,0.0555564985,-0.1573742777,0.1570022553,-0.02068929,-0.0604657792,-0.1080057248,0.2944135964,-0.00699904,0.4499303401,0.0754646137,-0.1183186471,0.0667596757,-0.1927372217,0.2933219969,0.0403670929,0.2244839519,0.1030922234,0.2842251062,0.196530953,-0.5416135192,0.3025431931,0.2099866122,-0.0410254374,-0.0964831263,-0.1445709616,-0.0314274542,0.051625669,-0.2259208709,-0.0453448929,-0.3614836335,0.3704491258,-0.0049913912,-0.2521316707,-0.5928734541,0.0533662066,0.2811320126,0.3872300982,0.1061195359,-0.1477273852,0.1504846215,0.1130765229,0.2815261483,0.2899907529,0.2494257241,0.0743724704,-0.2620927095,0.1713245809,-0.0873468891,-0.1128724664,-0.0194508675,0.1650666595,0.4752903879,-0.1073506922,0.0096322689,0.133734405,-0.0545267463,0.106719926,0.3185575604,0.0217739884,0.0777640715,0.1855484396,0.1987583488,-0.3515220881,-0.139888823,-0.1542218328,-0.3481053412,-0.0948709697,0.2665059865,-0.088078618,0.0736078322,0.0979898795,0.0551433489,0.3159165382,0.2274737507,0.5344837904,0.0322412103,-0.3407536745,-0.4845394492,-0.3706590235,-0.231229946,0.0498786345,-0.1062839702,-0.1174254194,0.047318019,-0.3516747952,0.0669400617,-0.1297477633,-0.2198439687,0.1422678679,-0.2810109556,-0.2178428173,0.3459418118,0.1754386127,0.4400460124,-0.05391277,-0.133134678,0.2523139417,-0.0499270931,0.0098541658,0.1244200319,0.1042019203,0.1302523315,0.284040153,-0.0573797636,0.2944975197,0.1350450367,-0.0791729838,-0.4958041012,-0.3676656485,-0.1456313133,0.0074781156,0.436678946,0.0757361278,0.2836892307,-0.1762489825,0.0180865265,0.0131636383,0.388620615,0.0656279027,0.243244946,-0.3522212505,0.0078461226,0.2614372671,0.2554189265,-0.0641343594,0.2637672424,-0.0437513925,0.2339209914,-0.5088478327,-0.24350743,0.4126277268,-0.109449707,-0.1396605968,-0.0169326589,0.4314415157,0.0141383819,0.4137528837,-0.6328085661,-0.3446973264,0.3429203033,0.107165128,-0.2154010534,0.2676656842,-0.4782138467,-0.23450993,0.1565807164,0.0931104794,0.211096406,-0.431619674,0.1794063449,-0.1144845039]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/724","title":"need to redirect \/nlp to \/datasets and remove outdated info","comments":"For context, those are two different pages (not an old vs new one), one is from the dataset viewer (you can browse data inside the datasets) while the other is just a basic reference page displayed some metadata about the dataset.\r\n\r\nFor the second one, we'll move to markdown parsing soon, so it'll be formatted better.","body":"It looks like the website still has all the `nlp` data, e.g.: https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\n\r\nshould probably redirect to: https:\/\/huggingface.co\/datasets\/wikihow\r\n\r\nalso for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).","comment_length":56,"text":"need to redirect \/nlp to \/datasets and remove outdated info\nIt looks like the website still has all the `nlp` data, e.g.: https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\n\r\nshould probably redirect to: https:\/\/huggingface.co\/datasets\/wikihow\r\n\r\nalso for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).\nFor context, those are two different pages (not an old vs new one), one is from the dataset viewer (you can browse data inside the datasets) while the other is just a basic reference page displayed some metadata about the dataset.\r\n\r\nFor the second one, we'll move to markdown parsing soon, so it'll be formatted better.","embeddings":[0.1384783983,0.0236901566,-0.0980000347,0.0289668012,-0.0519955009,0.0458978377,-0.1316044331,0.4961518943,-0.1375634372,-0.2072739899,-0.1407679468,0.3467579484,0.2411794066,0.0714614987,0.1461839974,-0.1521984339,-0.0214515217,0.1132967472,-0.0513905771,-0.1351284236,-0.1931079328,0.0566270165,-0.3132694066,0.1716675758,-0.0621245429,0.0175958984,-0.1403496563,0.2805624008,-0.1343233436,-0.4249670506,0.102028273,0.0087717818,-0.0224362407,0.1715869755,-0.0000975793,-0.1550674587,0.4177809656,0.2299157381,-0.4386451542,0.0474657714,-0.1592706442,-0.4129195213,0.0570607409,-0.2196607888,-0.0308993682,-0.2214241624,0.137592271,0.2655303478,0.2472458333,0.3672331572,0.3661076427,-0.019929383,-0.2039603144,0.0462491661,0.0675611421,0.2029665858,0.0472302809,-0.1386793703,0.1671683341,-0.1134862825,-0.2322293669,0.5095509887,-0.0722998306,-0.2618313134,0.0562584512,-0.0017119646,-0.0014855193,-0.070292905,0.046875637,0.1706978679,0.436910212,-0.0547419526,-0.0492673069,-0.2271544784,-0.0232388247,-0.1028957665,0.2928943634,0.1422806829,-0.1384128332,0.0905143023,-0.1495300829,-0.4678625762,-0.1713795066,0.3049802184,-0.0220149513,0.56427598,0.0080763763,-0.1040589735,0.074239254,-0.0737466514,0.1059833243,0.0995106027,-0.0752474815,0.0970747918,0.1928911507,-0.0765538886,0.2435375452,0.4252410531,0.0815108493,-0.0695508495,0.0161836036,0.1859911829,-0.198638469,-0.0451902598,0.1165728718,0.068919152,0.0976125598,-0.1040062606,0.2782568038,0.3188299239,0.077490069,0.1003954634,0.2027646154,-0.0784591734,-0.2845477164,-0.1333402097,0.1968569905,-0.2208771259,-0.0664087161,-0.0128305396,0.0212563667,-0.1261846721,0.0431220122,0.2340467721,-0.1883869618,0.1729913354,0.0916701928,0.0043306686,-0.3476365805,-0.2004378885,-0.1041787714,0.062396165,-0.5295287371,0.0361767597,0.2182329595,0.2483716607,0.222904548,0.2349675894,-0.4926007986,0.0033514709,-0.1259416193,0.2126781195,0.2112226337,0.0128597161,0.0014351723,0.2291309386,-0.0713291466,-0.2442645431,-0.0399351902,0.0254284684,-0.3115333617,-0.1752976775,-0.1647272855,0.3262220323,0.0564432628,-0.0685154498,0.18339324,0.397426784,-0.036995735,-0.2568757236,-0.0134542286,0.301879108,-0.170118615,-0.1396078616,0.0991517305,0.3828622699,0.0952413604,-0.1544090509,-0.0409387872,-0.0012377891,0.0293821972,0.173140496,-0.2121288776,0.0950792581,-0.218935281,-0.102346234,0.5547258854,-0.2279092371,-0.1078956276,0.2369597107,-0.1232387871,0.052433379,-0.1402726769,-0.0220217165,-0.1591455042,-0.0985369757,0.0036777037,0.1784611493,0.268003732,0.0771942139,-0.6268885136,-0.114068307,0.070186913,-0.0578800477,-0.1109447777,0.1748309284,-0.0566855557,0.1209659502,0.5310482383,-0.0093563842,0.1440632641,0.136564821,0.0343695357,0.0989690274,-0.052304022,0.0686732009,-0.2810053825,-0.1475200951,-0.2399805039,0.3287532926,0.326877594,-0.2241638899,-0.403958559,-0.2527482808,-0.1470331401,-0.4105485678,0.3645327091,0.1858108342,0.040363986,0.2435142398,-0.1546339989,0.1094995886,-0.0035772857,0.094842501,-0.3514446616,0.1881889552,-0.279169172,0.0763323531,0.1151818782,0.2385628968,0.2698487043,0.0877145231,0.0782882199,0.2478946298,-0.2400452644,0.2538331151,0.3343015313,0.0537943766,0.3653838336,-0.3246807754,0.0520896465,0.2069708407,-0.1764696091,0.1793634892,-0.3303443193,0.1332094222,0.1781615019,0.2474444807,0.1418089867,0.2600362003,0.2022663504,-0.4204209745,0.0664190575,-0.0762854293,0.0819931701,0.1620307565,-0.0552762561,0.0285124127,-0.2853654921,0.5206699967,0.0286518987,0.0097688157,-0.0239305515,0.3777434528,-0.1655337811,-0.3923074007,0.154936716,0.2252096236,-0.1950206906,0.2894262969,0.2376935482,0.2057400793,-0.1825958043,-0.2519817948,0.31046018,0.0524475761,0.0171946883,-0.0813839436,0.1677494943,0.1088386849,-0.3862136006,-0.064042598,-0.190724194,0.3152854145,-0.2420074344,-0.0095195165,-0.2421172112,-0.6865228415,-0.1989766806,-0.0486601219,-0.3761569262,-0.5674775839,0.1000581011,0.0890862197,-0.2643342018,0.096295163,0.0220667068,0.2123070061,-0.26298967,0.403229773,-0.4363816381,-0.3189541996,-0.2448745221,0.2402091324,0.1480138302,0.4297671318,0.177730605,-0.066489473,0.0116419494,-0.2758164108,-0.5960994363,0.0688938275,0.1598372906,0.065273501,0.0897851214,0.2754555345,-0.1754969805,-0.1561299413,0.1922322065,-0.1071819812,0.0250942484,-0.1592062861,-0.0489173345,0.1280827224,-0.0721236542,-0.4142029881,-0.130512476,-0.3529879153,0.0151670761,0.0577152446,0.3492697477,0.3905658126,-0.0092571238,-0.1279177517,-0.1278903782,0.2130807787,-0.3561644852,-0.1776023656,0.1717917323,-0.2524499893,-0.3447203338,-0.0921413898,0.0265414398,0.1429200619,-0.1665219814,-0.5557922125,-0.1072855741,-0.2250916362,-0.0492843576,0.0959306881,0.0670157894,0.2911284268,0.2124027312,-0.3828006685,0.0761632994,-0.2007709295,-0.1852182299,-0.0926346481,0.3297071457,-0.1473432481,-0.0259679593,0.1464907229,0.2578772306,0.1349194646,0.1953126341,0.2204305083,0.0854305848,0.419657737,0.011008135,-0.2900315523,0.0634929314,0.0090660583,0.1195359826,0.3961511552,0.2554289103,-0.1339019388,-0.3042543828,-0.217126444,-0.4790750444,-0.4104218483,0.1557454914,0.0932827815,0.1844753921,0.4510629773,0.0493236072,-0.1134250462,-0.3020300269,0.1069517061,0.3718749583,-0.0175994933,-0.0517834164,-0.4626865387,-0.1169179529,-0.4204596877,0.008606567,0.012529036,0.0068326085,-0.1453172266,-0.1797373146,0.0675043762,-0.0653715804,0.1453809142,-0.2783140838,0.0355925299,0.0482695401,0.3203530908,-0.2505602837,0.343912065,-0.0825131014,0.2027129978,0.4124555588,0.1730882525,-0.1125212759,-0.233201012,0.283608526,-0.088860698,-0.1713405997,0.1754533201,-0.1122347638,-0.2408093959,-0.170441106,-0.0391073488,-0.2217512131,-0.1524666846,-0.1458521038,0.0307615437,0.3162342608,0.0287967864,0.0636043772,0.2657086849,0.1218656674,0.2786855102,-0.3115952909,0.0927836522,0.3133794367,-0.0665598884,0.4313217998,-0.0712837875,0.1204112768,-0.2262969464,0.0282771308,0.3692047298,0.4006429017,0.3853724599,0.0272873435,0.0499906428,0.0483686738,-0.2750867605,0.0617832132,0.111887753,-0.0430064425,-0.3030899763,-0.1651149541,0.4166704118,-0.1593436748,-0.3416793048,0.1539679766,0.3954318762,-0.2817866206,0.1258599907,-0.1199480295,1.0915570259,0.0552989803,0.2253075391,0.3524339795,-0.2674306929,0.4936009049,0.1012987569,0.1801206321,-0.2920378745,0.1751583368,0.1270914078,0.1457635909,-0.055767078,-0.0058921366,-0.2550282478,0.1094239801,0.1402429193,-0.0785358623,0.2105832547,0.4886168838,0.0843971521,-0.0641260073,-0.5037642717,0.2985770702,-0.2262145579,0.1996663064,-0.1277562678,-0.1631535143,-0.0601420999,-0.0321708098,0.1165290624,0.1792828739,0.1863968521,-0.0179047547,-0.0378010832,-0.2672226131,-0.037340533,0.1575271189,0.1092490256,0.2163615674,-0.3734364808,0.2302187383,-0.1959143281,-0.069423683,0.0487480797,0.1498441696,0.0336107984,-0.0839719698,-0.288569957,-0.0285137836,0.0361263007,-0.0350860469,-0.3423162997,-0.0780352727,0.0723431036,-0.2086376846,0.1602956951,0.2669920623,-0.2990272343,-0.3812173605,0.2623146772,0.0443825014,0.029635597,-0.1978316605,0.5369786024,-0.0462450013,-0.1385474652,0.2511784732,0.0826558098,-0.1980268359,0.3518412709,0.1640083194,-0.1358989477,-0.4078323841,0.0390856788,0.0016323213,-0.162639007,-0.1120079756,-0.1261799484,-0.0972317383,0.0267891046,0.1738174856,-0.105613932,0.0543489829,-0.0123708155,-0.1686031371,-0.2082092613,0.3446324766,-0.056145817,-0.0612733066,-0.0452142693,0.1294675022,0.2444751859,0.0769977644,-0.4937223792,0.0811451823,0.0188795421,0.2040676475,0.0736447424,-0.2258643806,0.0162376463,-0.1037576944,0.1585815996,-0.3116462529,-0.0532302745,-0.3942674696,-0.1345126182,0.0667915493,0.1154428497,-0.262173146,-0.171346426,0.0479212105,0.1974187344,-0.0503354296,-0.2641848624,-0.0713342577,0.0700575858,0.2239848524,-0.084443517,-0.0575154237,0.1428698599,-0.0570463799,-0.0508438088,0.2087192386,0.0167040993,-0.0038872785,-0.1956360489,-0.1665590554,0.0156889129,0.1917095482,0.043832738,0.3789802194,0.1804862618,-0.1043237522,0.0294196997,0.239362359,0.1878098845,0.247782737,-0.0290873311,-0.1235619262,0.2525253296,0.3605397046,-0.2280332446,0.1633861661,0.3158961535,-0.2228287011,0.0871226192,0.0798252895,0.0638512447,-0.1516431868,0.0634769201,0.12425825,0.2361385077,-0.4074928463,0.0833906159,-0.0008195193,-0.1425863504,-0.1591534019,0.1008664146,-0.0154249799,-0.0259342249,0.1727671027,0.1672540307,0.1875387132,0.1990916431,0.1983813047,-0.0316238441,0.1032966599,-0.0272880532,0.1560550481,0.1149837375,0.2188353986,0.1606697738,0.3957930505,0.0339275263,-0.380235374,-0.0479916669,0.2377634645,-0.1753586084,-0.1134166867,-0.2992447019,-0.1522037387,-0.0028739804,-0.0327752084,-0.0263366681,-0.1942236274,0.1382886618,-0.1051362306,-0.2177474499,-0.5958793163,0.0416439399,0.1248269007,0.3702645004,0.1492662877,0.1023702398,-0.0352713875,0.0460775346,0.0458044894,0.4355767667,0.179875657,0.1616175622,-0.2159407884,0.0523510017,-0.1297667474,-0.1344488859,-0.1680035144,0.1029506624,0.3525752127,-0.0743375942,0.2070360482,0.2491971999,-0.2899200022,0.0684604198,0.1442240924,-0.0367556624,0.2184754461,0.073574312,0.0894304067,-0.2402599454,-0.1313077211,-0.2430719584,-0.3983203173,0.027714381,0.1920780391,-0.1561357379,0.1798248738,-0.2516317368,0.1540109515,0.3098346293,0.502558887,0.3681330979,0.116983898,-0.0317990407,-0.4122181535,-0.6755204201,0.0369235352,-0.0604937449,-0.2278496623,-0.0889483616,0.0863098577,-0.2890084684,0.0963240042,-0.0644585341,-0.0280502848,0.0477548763,-0.262989521,-0.3645362556,0.2613121569,0.1004491448,0.2882197499,-0.0359893106,-0.1095353961,0.3306038082,-0.1103644893,0.2111804038,0.1841337681,0.1408011019,-0.1183778644,0.2390456945,0.0826437697,0.266258508,0.0853166431,-0.2080555409,-0.4103218913,-0.4120026529,-0.2276323587,-0.0203007422,0.3156791627,0.1692177206,0.3763763905,0.0344467536,-0.0559514984,-0.0370041877,0.2447821796,0.2542082071,0.273785919,-0.2028926015,0.2134634852,0.310757935,0.1523429304,-0.0311639756,0.2368996739,-0.0129877618,0.0038760877,-0.3176868558,-0.518298924,0.3584730029,-0.2934133112,-0.1650796384,0.0446168147,0.3463246822,0.0199225321,0.1421867907,-0.520522058,-0.1700234264,0.1570398062,0.0337525234,-0.2412440926,0.1453214735,-0.3328873217,-0.3634594977,0.0072161318,0.1913143545,0.3161476552,-0.457780242,0.1836125553,-0.0948202163]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/724","title":"need to redirect \/nlp to \/datasets and remove outdated info","comments":"I understand. I was just flagging the lack of markup issue.","body":"It looks like the website still has all the `nlp` data, e.g.: https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\n\r\nshould probably redirect to: https:\/\/huggingface.co\/datasets\/wikihow\r\n\r\nalso for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).","comment_length":11,"text":"need to redirect \/nlp to \/datasets and remove outdated info\nIt looks like the website still has all the `nlp` data, e.g.: https:\/\/huggingface.co\/nlp\/viewer\/?dataset=wikihow&config=all\r\n\r\nshould probably redirect to: https:\/\/huggingface.co\/datasets\/wikihow\r\n\r\nalso for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).\nI understand. I was just flagging the lack of markup issue.","embeddings":[0.3121137917,0.049953334,0.0001663364,0.0672624707,-0.1151986718,0.0831950009,-0.1976590008,0.7124435902,-0.3074539006,-0.4837414026,0.0184820816,0.1823898405,0.2734232545,0.0811766684,0.0155482776,-0.224872008,0.0418451242,0.1194557846,0.0926132798,-0.1298215538,-0.1840707958,0.0760487914,-0.4259464145,0.0895498246,-0.057430461,0.1863916665,-0.274757266,0.2186981738,0.0128066568,-0.4521507621,0.233032912,-0.041215755,-0.0035548999,0.2869084179,-0.0001153144,-0.3819181323,0.4201750159,0.2734089494,-0.4040446579,0.0266998839,-0.3029521406,-0.5209726691,-0.0512088798,-0.0561881065,0.0821864158,-0.1064834669,0.0175983999,0.3641363382,0.3566517532,0.4198992848,0.1767500937,-0.2354136705,-0.2281059772,0.1842486858,-0.0145172467,0.4247660935,0.1794860065,-0.0424174219,0.218910709,-0.108294785,-0.0463079177,0.6112314463,-0.2842897475,-0.3035524189,0.0342958756,-0.1799987406,0.0585613474,0.0137152802,0.0608603656,0.2468251735,0.5079960227,-0.0884071589,-0.0573685095,-0.1837991625,0.0358981565,-0.2231395245,0.2622285187,0.1159586385,-0.1138907373,0.1685933024,-0.2343468666,-0.7096889615,-0.2176048309,0.3618764281,0.1301011592,0.5949155688,0.0067921905,-0.0207128804,-0.024813341,0.0375945792,0.28298226,-0.0150205018,-0.1389777064,0.1442379653,0.2614130378,-0.2010602802,-0.0916535482,0.3249067366,-0.043853078,0.0278132148,-0.0326323062,0.1668570638,-0.2085462362,-0.0112059526,0.1398984343,0.1188817695,0.1547500789,0.0274594519,0.1959964633,0.1880315393,0.1771432161,0.1031875908,0.1646202058,0.0255090818,-0.4745702744,0.0554477647,0.3781544566,-0.367338419,-0.1460703462,-0.0107235666,-0.1698699743,-0.1774009615,0.0622730926,-0.002443657,-0.1762068272,0.3352080584,0.2665275931,0.108097963,-0.2088692337,-0.2620197237,0.1379564106,0.1621140093,-0.5227422714,-0.0487293378,0.1706895083,0.2870421708,0.1318935752,0.1845576316,-0.5318283439,-0.0059547229,-0.1387128532,0.0999911278,0.3505681157,0.064644888,-0.2796598971,0.2416704744,0.0723626986,-0.1382050216,-0.227845192,0.2116680294,-0.4205908179,-0.2350843102,-0.1214417964,0.1213486791,0.0237509999,-0.0749187097,0.0474984273,0.5399441123,0.0321400501,-0.3299078345,0.0808482915,0.230441168,-0.2118103355,-0.1967500895,0.0096923979,0.5410260558,0.0182653069,-0.0333944708,-0.2096469849,0.1550637335,0.0964485481,0.1794069558,-0.0828172714,0.0267068818,-0.1406253576,-0.3276030719,0.6296696663,-0.4168357253,0.019913407,0.3359433115,-0.0560782701,0.0356473327,-0.1769518554,-0.0058480799,-0.2500549555,-0.2651512027,-0.1353917122,0.1659582704,0.329456985,-0.1538588554,-0.5116532445,-0.206457898,0.0390739739,-0.1328460276,-0.1730636954,0.2373200804,-0.2350547165,0.0705238581,0.5517298579,0.1773053557,0.2991080582,0.0618381314,-0.0289415605,0.2161879838,-0.0664126724,0.2408443242,-0.2139448673,-0.2351850718,-0.12299826,0.1791379303,0.4254705608,-0.1890263855,-0.3055891097,-0.3230370879,-0.1817670017,-0.4135717154,0.14646402,0.0512773618,0.0428871475,0.1019634083,-0.2781560719,0.08420863,-0.0519889966,0.1351807564,-0.2034198493,0.284430474,-0.103057988,0.1038426235,0.1682254672,0.3632629514,0.3530548811,0.0637160465,0.0755793154,0.2720287144,-0.2182272673,0.0641740263,0.3869410157,0.0215834398,0.3037807643,-0.4388135672,0.1363336146,0.3686822653,-0.3081907332,0.1256404668,-0.3620037735,-0.0851353407,0.1350598484,0.4353942573,-0.0133787869,0.3100491464,0.0063065025,-0.4687981904,0.0244808793,-0.0529614836,0.1359496713,0.2446075976,-0.133996278,0.1312762797,-0.1736081392,0.5197536349,-0.1244020015,-0.1267498583,-0.1058771312,0.4842295051,-0.1369006038,-0.4732503295,0.1806265265,0.1709880233,-0.1684914231,0.2007486075,0.1216370836,0.1624431759,0.0803192258,-0.1805135757,0.3031093478,0.1471854001,-0.1061947048,-0.2216950357,0.0899641365,0.2064427584,-0.1198119521,-0.195438087,-0.147754848,0.3691796064,-0.2084124982,-0.0082296906,-0.3075575233,-0.8154377937,-0.3433330357,-0.1136377826,-0.5743998289,-0.507727623,-0.0435523763,-0.0674857274,-0.1105821133,0.0973115787,0.0914751664,0.2796940506,-0.2902366519,0.2444553077,-0.5706679225,-0.3709126413,-0.3021612763,0.0668940544,0.116711989,0.3781304955,0.1288924962,0.0108151622,0.0981406942,-0.257445246,-0.6133751273,-0.0708967596,0.1427922994,-0.0794783235,0.1220057532,0.3299681544,-0.1069513559,-0.0468527265,0.1477011889,0.0200351402,0.049784027,-0.1970223039,0.0969124809,0.3820397258,0.0254282784,-0.1837107241,-0.1944239587,-0.3151046634,-0.0138591519,-0.0320823453,0.2687450349,0.3569940031,0.048788581,-0.2859113216,-0.1558191031,0.168817237,-0.3459967673,-0.219113335,-0.0060026441,-0.0334227867,-0.1243130714,0.0266373847,0.2025225759,0.0507321134,0.0060664518,-0.6127337217,-0.1377636492,-0.1659317017,-0.2174377888,0.0845919028,0.0340219364,0.310770601,0.2806150019,-0.2519896328,0.1439770013,-0.1950098872,-0.2122138143,-0.0581770502,0.4314448833,0.0834526792,0.0186503492,0.279836297,0.2041777521,0.1755305231,0.3253601193,0.1641829461,0.2870451212,0.4816306531,0.0762197152,-0.2647235692,0.0886185467,-0.0647676289,0.2359607965,0.2644204795,0.3525199294,0.0108002098,-0.3699546158,-0.365655899,-0.3869067132,-0.3215874135,0.1865542829,-0.0682817176,0.2619588077,0.3954570591,0.0270936079,-0.0196742825,-0.4043807387,0.0363866501,0.4727167785,0.1090523601,-0.0662123859,-0.5515373945,-0.2882637978,-0.3238580525,0.1774097681,-0.0130424118,0.0790669471,-0.196094498,-0.045199845,0.0029488616,0.0936395302,0.0771860853,-0.3526243269,-0.0739295557,-0.0307348128,0.3724617064,-0.290350765,0.3712398112,0.1018923894,0.2400370687,0.4315045476,0.124755688,-0.1151724532,-0.3203204572,0.6249171495,-0.0167506952,-0.1162151992,0.1279618144,-0.1502049565,-0.2667894363,-0.1926303506,-0.0562716462,-0.1294878572,-0.1645620614,-0.07874313,0.2433791459,0.3415629268,-0.0237792972,0.1277205497,-0.0806938782,0.0222130492,0.2495805174,-0.4481226802,0.1008876413,0.2697258294,0.185569033,0.2672497928,-0.1671937704,-0.0988892913,-0.3628971577,0.1002174839,0.6756293774,0.6172820926,0.4762189984,0.1955121458,-0.0506991856,-0.0535919853,-0.300742805,0.0931104794,-0.0075235907,-0.0439049564,-0.5921224356,-0.2635300159,0.414119482,-0.1439275742,-0.2659974098,0.1820229888,0.3974396288,-0.2978719473,0.1096360162,-0.0897750854,1.4291721582,-0.000865522,0.256526649,0.3096844554,-0.4007885158,0.6899108291,0.16811499,0.216217652,-0.3022377193,0.2702332735,0.0069261692,0.0764393583,0.1282144189,0.0751267448,-0.2815228999,0.2460896075,-0.1172614768,0.0463160053,0.4460447729,0.4472658336,0.0421958715,0.0401994921,-0.5951189399,0.096384339,-0.2864804566,0.2229105979,-0.1483656168,-0.1773940474,-0.1411143094,0.008632916,0.0850536898,0.1216822118,0.2743267715,0.0407964066,-0.2017478794,-0.3902743161,-0.0283115245,0.13207452,0.2264724225,0.2344796509,-0.2757345736,0.2643720806,-0.1536898613,-0.0960848406,0.2205400318,0.0294983424,-0.0248385798,-0.000995228,-0.2474957556,0.0554591157,0.0210520551,-0.0954895094,-0.4248725772,-0.1283671707,0.2023754567,-0.2268251479,0.1403612792,0.3582680821,-0.308185935,-0.2166725248,0.0972310752,0.1946189851,0.0321375467,-0.1957535893,0.3877584636,-0.095033668,-0.0632411391,0.2224540412,0.1618603319,-0.2750146985,0.4436224997,0.0874807984,-0.0679790378,-0.284730047,0.1291270256,0.0141269639,-0.2611299753,0.0008159549,-0.2003303468,-0.0111970576,0.0881563351,0.119385995,-0.2090937495,0.0770391375,0.0151648177,-0.2666352093,-0.2520269156,0.308339417,-0.0541184656,-0.0800754428,-0.2058586031,0.2816421986,0.2856034338,-0.021230096,-0.2877645493,0.19966124,0.1726448238,0.3165671527,0.1310645491,-0.1372187436,0.1183192283,-0.075678654,-0.0960454568,-0.4450204372,0.0421475023,-0.1942452043,-0.1171189472,0.1558960527,0.0893357471,-0.3943462372,-0.2206305414,-0.1115606278,0.3140355647,-0.0245635137,-0.5111380219,0.0885364339,0.088786684,0.2897472084,-0.1059973761,-0.1677989513,0.1546612084,0.1393998563,-0.0248186439,0.253153652,0.0223623682,0.0944724977,-0.0723539442,-0.0655811355,0.0062947245,0.1999638081,0.0854543671,0.1052829176,0.1806842089,-0.0357230976,0.1077946052,0.1527138948,0.2118401378,0.4078911543,-0.0903725699,-0.1380528361,0.2557862401,0.125784874,-0.1839438826,0.2373187542,0.4143259823,-0.1657807231,-0.2028378993,-0.0111092813,0.0744210333,-0.2060944438,0.0247758161,0.1541421115,0.357872963,-0.4810714722,0.0896400809,0.0917777345,-0.0135561423,-0.1553218812,0.1172784641,0.0603422709,-0.0076976838,-0.0849929675,0.2721902728,0.0660780519,0.392267108,0.1440452337,-0.169031173,0.0897945762,-0.2024345547,0.2951602638,0.1297168881,0.2689244449,0.1412351727,0.3790894449,0.1474922448,-0.4827577174,0.1624906063,0.2640354335,-0.0548146069,-0.0982024819,-0.238650471,0.0177368075,0.0016244646,-0.1285876483,0.0343153514,-0.2685404122,0.2576007247,-0.0373353101,-0.2693895698,-0.5754164457,-0.0225143377,0.2759726048,0.4431509078,0.0976536721,-0.1535121351,0.0038909777,0.1772876382,0.2929264605,0.3548230529,0.3087698221,0.018375393,-0.2700078487,0.0778377354,-0.1856582165,-0.0614573397,-0.0389045253,0.226038307,0.4249528944,-0.0111356303,0.011509805,0.0903569385,-0.0609310567,0.1445835382,0.2097390294,-0.0604004301,0.0478034019,0.1780730337,0.134366557,-0.3508805335,-0.1559800208,-0.1859836429,-0.2923302948,-0.0471232273,0.3014952242,-0.0555227064,0.1203449592,0.0191251542,0.062717922,0.3014355898,0.2425976247,0.5738956928,-0.0606247522,-0.2362863272,-0.4625973105,-0.4981448054,-0.1876816452,-0.0191421993,-0.1543782651,-0.0630073398,-0.0323370397,-0.236650154,0.1229447722,-0.18451114,-0.2078841031,0.1534264386,-0.314740628,-0.2530477643,0.3148359954,0.1685433537,0.4534058273,-0.0440449864,-0.1592959166,0.23332344,-0.027178118,-0.0161928218,0.1484400928,0.0836957321,0.0210778676,0.2964026034,-0.0546433665,0.2856386006,0.1004963368,-0.0624935515,-0.4700012803,-0.4627919793,-0.2253038138,-0.0050991313,0.5007927418,0.0615272187,0.2555485666,-0.2155905515,0.0858847275,-0.0254798513,0.3489435613,0.0724282786,0.3452991247,-0.3424378335,0.0822933391,0.2715480328,0.3163326085,-0.0896165073,0.2536320388,-0.0599368773,0.1819185913,-0.5088608265,-0.3227801323,0.4950957298,-0.2659641802,-0.195047766,0.0237622876,0.3641318977,-0.0573648065,0.3212889433,-0.6140242219,-0.3967353702,0.2766321898,0.1400351971,-0.2757867575,0.2085591257,-0.4300899804,-0.3208374083,0.1410806477,0.1570859402,0.1715867668,-0.4577231705,0.236540705,-0.0621435009]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/723","title":"Adding pseudo-labels to datasets","comments":"Nice ! :)\r\nIt's indeed the first time we have such contributions so we'll have to figure out the appropriate way to integrate them.\r\nCould you add details on what they could be used for ?\r\n","body":"I recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n","comment_length":36,"text":"Adding pseudo-labels to datasets\nI recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n\nNice ! :)\r\nIt's indeed the first time we have such contributions so we'll have to figure out the appropriate way to integrate them.\r\nCould you add details on what they could be used for ?\r\n","embeddings":[0.35087201,-0.1339654475,0.0640572533,0.0146746505,0.1172191575,0.0425939336,0.2484151274,-0.055692628,-0.1454755217,-0.0428759754,-0.227577284,0.0762309954,-0.1649908125,0.5450238585,0.4980666637,-0.1480551064,0.3547057807,-0.1910561919,0.1890376806,-0.1643102169,-0.3369725645,-0.2637389898,0.1954767555,-0.177405566,-0.3335675597,-0.1520867795,-0.188161701,0.0017013031,0.0816946998,-0.5430544615,0.0081283478,0.3411814272,0.2522218525,0.6251304746,-0.0001249106,-0.1294019818,-0.0920366049,-0.1204193681,-0.1105555296,0.190760687,-0.0204500537,-0.2631741762,-0.214708522,0.196944952,-0.09029378,-0.4064389765,-0.1335061342,0.0102537191,0.1772880256,-0.0072268839,0.0163613167,0.1366703957,0.2723228633,0.1299423873,0.5037116408,0.3426888287,-0.2014466077,0.3047901094,0.3386514485,0.3808413744,0.4996564984,0.4813934267,-0.0260894764,-0.1911639571,0.5817210078,0.0361270718,-0.0280091148,-0.2036755383,-0.3807860315,0.1567308158,0.4856583178,-0.2559436262,-0.6600388288,-0.4705187678,0.0994573832,-0.2843776643,-0.1884535402,-0.1961780787,0.4040839374,-0.0619082637,-0.0100463936,-0.2337230146,-0.0850889012,0.2030259669,0.1202186421,0.5472038984,-0.2007035315,0.2139021754,0.0618340038,-0.1964243501,-0.0501444563,-0.2513433397,0.1982958913,0.1462722868,0.1533986032,-0.1464244574,-0.0554305874,0.035822615,0.2769576609,0.306052506,-0.0804551244,-0.1082585827,-0.2618288696,-0.0918944031,-0.0831924751,0.1469036788,0.0197892208,0.2458151579,0.2622636557,0.0032124117,0.3767152131,-0.0831515342,-0.4829546213,0.5400187373,-0.0798437819,-0.184470728,0.0166599695,-0.1186084896,-0.0216396227,-0.0099645751,0.1242481619,-0.1875851005,0.078972511,-0.113677375,0.0045646601,0.073093608,-0.0961935148,0.3804770708,0.2330143303,-0.0041363514,-0.0311319679,-0.1038940996,-0.0761066303,0.3563047945,0.0114585357,-0.3125945926,0.2468480021,-0.2488931119,0.198173508,-0.2102049887,0.0347838104,0.1707411706,-0.0479497649,0.2335938662,-0.1842186451,0.0526870042,0.0468658768,0.0510809645,-0.2296123803,-0.0847219676,-0.0553180501,-0.1274133474,-0.3076143861,-0.007403973,-0.3845643401,-0.3234862685,-0.3623290658,0.3362720907,0.146713376,-0.014857064,0.0895832852,0.1612404734,-0.1204106584,-0.119428359,0.3341804743,0.1501198262,-0.3646391928,-0.2879036665,0.1363932937,0.027432194,0.1304156631,0.2103500217,0.0395956598,0.1638571173,0.2064767182,-0.0699356869,-0.118732132,-0.4139882326,-0.1851709634,-0.0824119002,-0.0383956283,-0.0672680587,-0.0235427301,0.447196722,-0.0290771779,-0.2683320642,-0.0834478214,0.0239267237,-0.0027650036,-0.0937337726,-0.0517457947,0.009491222,-0.1063969806,0.5546265244,-0.0790554658,0.0317616798,-0.0460544899,-0.1878689229,-0.1235040873,-0.0379969329,0.112446934,-0.0244041886,0.4308757782,0.4505261779,-0.0841156617,-0.3351357281,-0.4225297272,0.0643078163,0.3131957352,0.1518680602,0.2139699608,-0.3543727398,-0.2606374025,-0.1056091785,-0.1188807711,-0.0032252385,-0.1172124222,0.1667604148,0.21048522,0.0392617024,-0.351926595,-0.1949398965,0.0344020948,0.2529699504,-0.0170265827,0.1207967103,0.1968993843,0.0700706169,-0.218046084,0.5762590766,0.0122700259,-0.0774033144,0.2291522622,0.2382849604,-0.0247948989,-0.0827146396,0.1057941392,0.3093137443,0.4576770365,-0.1437068284,0.1279944032,-0.305983603,0.0114496276,-0.1778569967,-0.134126693,0.0533383228,0.1222603694,0.1460716277,-0.3488898277,-0.1220633239,-0.064772293,0.1816105545,-0.1966081113,0.0349039547,-0.0665336326,0.1509809941,0.4198540449,0.0485272035,-0.1534426212,0.3250514567,-0.017040005,-0.3731516302,0.1307826489,0.1720673442,-0.4888537526,0.2132398188,-0.0017561347,0.3827134967,-0.0193849653,0.0638613105,0.2843587399,0.0112366378,-0.0600267053,0.0600339547,-0.1200994998,0.313303262,0.0610037185,0.047300674,0.0482840054,0.1464210451,0.1569345295,0.0194783043,-0.0098285247,-0.0560709536,-0.255022794,-0.070810616,0.2456015944,-0.0945197418,-0.2045022398,-0.6011893749,-0.4497777224,-0.245485127,0.0487111993,-0.134419933,-0.1453191191,0.1909579039,0.1226615235,0.4898834825,-0.3512583077,-0.2254018039,0.1413585842,-0.1690039635,0.1628854871,-0.1073578298,0.1542331129,-0.1214861125,0.4159196317,-0.1388185173,0.1517219543,-0.1876494437,-0.5756072402,0.0636467338,-0.4807730317,0.1864727437,0.2330859452,-0.2272555083,0.3000864089,-0.2512019873,0.0596793853,0.1218449026,0.1086961329,-0.2782357037,-0.1743098348,0.0284928493,-0.3839595318,0.130328998,-0.4349294305,-0.1555609256,0.3866901696,0.1300320923,0.0219939239,0.2576771379,-0.273460269,0.1378861368,-0.058666721,0.008496304,-0.0351490192,-0.5451734662,0.4454016089,-0.0228166711,0.0330391526,0.1718705148,-0.1100521907,-0.0759961978,-0.0923953131,-0.1332699805,-0.3692465723,0.0426127203,-0.0050614271,0.35817945,0.3633756638,0.0498868078,0.0678379014,0.1727570444,-0.1507345587,-0.3482174575,0.1096350178,-0.0155550726,0.2649582624,-0.114940092,0.1408982128,-0.2226189524,0.627032876,0.176966846,-0.2345686555,0.1262998134,0.2979913652,0.2040144354,0.2684484422,0.0368236713,0.1379386485,0.0083837947,0.1878974885,0.1212026626,0.2570758462,0.3262632489,0.0405183174,0.1368399113,0.0201743618,0.0571698844,-0.0872630179,0.1436314136,0.2712720931,-0.2260143161,-0.3325639963,-0.447147876,0.0698933303,-0.1096526086,0.183258757,0.294257313,0.2618865967,-0.6611452699,-0.0527033694,-0.3960132897,0.1748719811,0.244799152,-0.3084754348,0.056871336,-0.1571287215,0.0546976067,0.0243412498,0.2687183917,-0.3838991225,-0.3637040555,-0.068895027,0.0773931071,-0.0373581648,0.0751038194,-0.0106498683,-0.3649801016,0.1830191761,0.3641944528,-0.5304680467,0.0452031419,0.309527427,0.0591489486,-0.0988950655,-0.1686684638,-0.2364957631,-0.465775013,-0.2433303297,-0.0507255085,0.14411138,-0.0580592342,0.1163626537,-0.0124951256,0.0377096236,0.095112212,0.3177097142,0.1828722805,-0.1146069691,0.5422511697,0.208555907,0.4471334219,-0.0509898737,0.5076368451,0.5242843628,-0.3771447837,-0.4337592721,-0.2176911831,-0.2883626819,0.5144400597,-0.0313558765,0.3144396544,0.1530524194,0.0846590996,0.1049327254,-0.5463404655,0.3300388157,0.0861147791,0.1523803025,-0.3281347752,-0.4354931116,0.6497976184,0.1358162016,-0.1339677274,-0.095717147,0.503556788,-0.1957800537,-0.0542914383,0.3126242757,0.9838216305,-0.2847186327,0.3361450732,0.0421895981,0.1361599714,1.0483944416,-0.0625781417,-0.2609398961,0.0118772537,-0.1134008542,-0.1125742421,-0.0819378719,-0.0747892857,-0.0406231955,-0.1220680773,0.1819446832,-0.1291684955,-0.0067709908,-0.0532925464,-0.1155592054,0.2831186056,-0.3984207213,-0.0422256961,0.0064976937,-0.2303169519,0.0134916231,-0.2043283731,-0.1657581925,-0.1504695266,0.2722949982,-0.0898268893,-0.1634229124,0.0383806303,-0.4214065373,0.0512669906,-0.0482460745,0.2185052186,-0.1262954921,0.7867920399,-0.0273045078,-0.0350665748,0.0809144303,-0.2605219781,0.0955459476,-0.1866031885,-0.2229931206,0.1329692751,0.1036499739,-0.312680006,0.0249662846,0.2777660787,-0.2456264943,-0.3948039114,-0.2721706331,0.1418700367,-0.1429704577,0.2579843402,0.5297660232,0.20099549,-0.1535038799,-0.0071223364,0.2283542305,-0.1628368795,-0.0830904394,-0.3562004864,0.1110170782,-0.2488144338,-0.1822171658,-0.093763046,-0.0923338756,-0.0918209031,0.1986861974,-0.0517796837,0.0627983436,-0.1350747943,0.0867515281,-0.3095857501,-0.1989090294,0.2536410987,-0.5064380765,-0.3839590549,0.31824404,0.4344847202,-0.3665504158,0.1841074824,0.0995577201,-0.5240021348,0.1594320536,-0.0566583909,0.2099518925,-0.0559373163,0.3442655206,-0.1018175781,0.3042809963,-0.152016148,-0.1078563705,-0.091113627,-0.0349740982,0.2185702026,0.0066314712,0.153787449,-0.0997627154,-0.1474537998,0.0516326725,-0.0515033714,0.0500356965,-0.3707665503,0.2346185446,0.3052381277,-0.4240824282,0.1333966255,0.0920967683,0.0870786309,-0.1338680089,0.3114676774,-0.1317249686,0.0855242983,0.3123394549,-0.1351475865,0.3421088755,-0.0184536278,0.2410942614,0.3581305444,0.247486487,0.2924283445,-0.0333468057,0.0091084233,-0.1073802561,-0.1162339821,0.2303807288,0.2635426819,-0.132940799,0.2961379886,0.0176959205,0.0926961601,0.3134860098,0.225747481,0.1083608419,-0.128477037,0.1317342073,0.2081294507,0.0750151351,-0.0244705398,-0.1884942353,0.0053492025,0.0648608133,0.1536205411,0.3775359094,0.2366767526,0.4687484205,-0.007442046,-0.0442513973,0.2752285302,0.158906728,0.0878041461,0.040505629,-0.03639552,0.1538603753,0.4635971487,0.4276756644,0.0430368073,-0.0315198973,0.1420859098,0.5564658642,-0.0167644825,-0.051404614,0.14983356,0.030840518,0.2447377741,0.2975254655,0.2078222781,0.0666140318,0.0108239595,0.0854119137,-0.3406426609,-0.2226180881,-0.0471166447,0.4238791764,0.1994228065,-0.3127707243,-0.2827529013,0.0637792274,-0.0132114924,-0.1547596008,0.0589314699,-0.7018326521,0.3001566231,-0.0447965041,0.159914583,-0.4526433051,-0.5470297933,0.3091581464,0.3365088701,0.0706115142,0.1328466088,0.4084155262,-0.148955062,0.1680801213,0.2040942013,0.1232496649,-0.2361437529,-0.0411452875,0.4352783263,-0.3590355217,0.1281316429,0.086117357,0.0697084069,-0.1634103507,0.0828655288,0.1794780195,0.007413296,0.009549398,-0.1601138413,-0.3326464593,0.0501829609,-0.2415525466,0.4797544181,-0.3484759331,-0.1995490342,-0.1394896954,0.3175758421,-0.3148475587,0.4054407477,0.4090920389,-0.0522536263,-0.0789522156,-0.1367018968,-0.0390763134,-0.0270060096,0.2926070094,0.3595173061,0.3388386369,-0.1954761893,0.017700471,-0.3608800173,0.0704909265,0.0115094595,-0.510710299,-0.0568262339,-0.2980892062,0.4002337456,-0.1214181259,-0.1359264404,-0.2043287456,0.0047957865,-0.324830085,-0.1442607641,0.0766164288,-0.3093498647,0.2042374611,0.1153706759,0.1524992585,-0.1460665017,0.3247851729,-0.1544343084,-0.0842668638,0.2609606981,-0.200767085,-0.0252762102,0.1296456903,-0.0191547982,0.486382395,0.1082788706,-0.2453238219,0.0161076561,-0.0890296772,-0.3318750262,-0.0126882186,-0.4134606123,0.2549862564,-0.2719509304,-0.0313540548,-0.2527205944,0.1063453257,0.1150481477,0.0308027286,-0.3354078531,0.4304694831,-0.2455373257,-0.2542997003,0.5712053776,0.2878658473,0.1374261826,0.0928031206,-0.0813299268,-0.0595384948,0.3417321444,-0.0514122769,-0.0964718014,0.0206324328,0.1394737512,0.3758439422,-0.1184827238,-0.8774783611,0.1891068369,0.2884035408,0.0653264523,-0.5017668605,0.2753763795,-0.0326778069,-0.4225319028,-0.0748912469,0.2685986161,-0.2749883831,-0.0157779455,0.0272682328,-0.1747651696]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/723","title":"Adding pseudo-labels to datasets","comments":"They can be used as training data for a smaller model.","body":"I recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n","comment_length":11,"text":"Adding pseudo-labels to datasets\nI recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n\nThey can be used as training data for a smaller model.","embeddings":[0.3236106336,-0.1705471873,0.0767482594,0.0553025417,0.1455969363,0.03575496,0.2866197526,-0.059739016,-0.0898068473,-0.0202813949,-0.2633213103,0.0311118513,-0.1358729452,0.5349795818,0.4721788764,-0.1100855172,0.2953128219,-0.2011565268,0.1507671028,-0.1810066551,-0.3599384427,-0.3136106431,0.2260216773,-0.1640241295,-0.3601994514,-0.1645295024,-0.1485246569,-0.0048646159,0.0626980066,-0.5612202883,-0.0211327486,0.3344987631,0.2650378942,0.5716098547,-0.0001252374,-0.1076325625,-0.1466301531,-0.1324191839,-0.1382960677,0.2175367326,-0.0470574498,-0.239052847,-0.2226939946,0.1851053834,-0.0898205489,-0.4023342729,-0.1135311946,0.0199588072,0.179651022,0.0606059656,0.0117403939,0.1274621487,0.2593846321,0.1219251826,0.4692736864,0.2954324484,-0.2046612799,0.3461514115,0.2935628891,0.4200083613,0.4833305776,0.4464986324,0.0034462938,-0.193264544,0.5569983721,0.0687073097,0.0339032821,-0.1904724985,-0.3122011125,0.186435461,0.4571416378,-0.2422807962,-0.6876101494,-0.4750324786,0.0990681723,-0.3165664077,-0.1687474996,-0.1687679738,0.4196640253,-0.0432869531,0.0009718654,-0.2683121264,-0.0523364842,0.1794773191,0.0941794962,0.5503262877,-0.191838339,0.2296448797,0.0576347597,-0.2194597423,-0.0889671519,-0.254592061,0.2057287246,0.1169036999,0.1560484022,-0.1756041348,-0.1183653027,-0.0010495095,0.268558383,0.2642661035,-0.0267506521,-0.1335746348,-0.2277931124,-0.1190177053,-0.0653408319,0.1547617614,0.0556248687,0.256757617,0.2520856559,0.00090321,0.3067933321,-0.0871245936,-0.5431601405,0.5461075902,-0.105010137,-0.1740448773,-0.045517318,-0.087097235,-0.0180455092,0.0294812266,0.142124027,-0.177833721,0.0912935734,-0.0558533072,0.0101742065,0.0628716573,-0.0915892571,0.3678977191,0.2050561309,-0.0117363445,-0.0286490414,-0.0926965401,-0.1068819612,0.3703257143,-0.01511175,-0.2746581435,0.2885118723,-0.2767260075,0.1808677316,-0.2599985003,0.028037658,0.16761522,-0.0208101869,0.2268609703,-0.1771686524,0.0860536918,0.0260488428,0.0599065013,-0.1998985708,-0.0881496146,-0.0627203509,-0.0934131891,-0.2837078571,0.000316258,-0.405438751,-0.2913951874,-0.3764716685,0.2953603864,0.1808551699,0.0011611893,0.1232507303,0.1125305668,-0.1404820085,-0.1208121479,0.2901604474,0.1359441578,-0.3464312553,-0.2773628533,0.2104981542,0.0327050313,0.0956838876,0.2340742648,0.0417409353,0.1593033671,0.2100940347,-0.0620449223,-0.1108028293,-0.3975424469,-0.2085312158,-0.100314796,-0.0812878013,-0.1018395126,-0.0257699117,0.4673004448,-0.013927469,-0.2657881677,-0.0761870965,-0.0532632582,0.0495560467,-0.0954939872,-0.0713133067,0.0138945291,-0.1166138127,0.4727445543,-0.0940469205,0.0543226004,-0.0853611976,-0.0997000337,-0.1125518158,-0.0341796726,0.0995260254,-0.0078767724,0.3924225569,0.4675444365,-0.0960373878,-0.3215367496,-0.3847274184,0.0788920894,0.2710045576,0.1160689518,0.2893601954,-0.3910988271,-0.2456501126,-0.1154258028,-0.1004146263,-0.0317078307,-0.1258582175,0.1458956599,0.2966170013,0.0099384412,-0.3283920288,-0.1993019879,-0.0424213484,0.2517051995,-0.0187256746,0.0951761901,0.2252668291,0.07680922,-0.2107329667,0.5853446126,-0.0017094128,-0.0702502429,0.2283571213,0.2384004593,-0.0035662532,-0.161528185,0.065992713,0.2846771777,0.4756277502,-0.1371748745,0.0974622071,-0.2794744968,0.0071344133,-0.1295837462,-0.1223206744,0.0352574326,0.1744290739,0.1044036821,-0.2804199159,-0.1535691321,-0.0706861019,0.1915058196,-0.1755067408,0.0208014976,-0.0537518412,0.1719561219,0.4328586459,0.0887179896,-0.1478509605,0.3163104355,0.0343166329,-0.3489584327,0.1492570639,0.1420131475,-0.511095345,0.2602009177,-0.0332821347,0.3716858923,0.0305472873,0.059053082,0.2921098173,0.0191535503,-0.056062419,0.0425069667,-0.1283709556,0.302752018,0.0522857159,0.0290875509,0.0287051015,0.0958486944,0.146037966,0.0429511592,-0.0104490034,-0.0619661547,-0.2876375318,-0.1020039394,0.2572000027,-0.0939967558,-0.2186250389,-0.5356428623,-0.4487345517,-0.2085801959,0.0651586875,-0.1343584657,-0.1437036544,0.1640323251,0.0797575563,0.4695492387,-0.3450471461,-0.1987034082,0.2104361802,-0.1664390266,0.1741382927,-0.1217311993,0.0913000256,-0.1822049618,0.3962112963,-0.1217460558,0.1060194448,-0.1886327565,-0.5276361704,0.0436970852,-0.4738060832,0.1451777965,0.239948377,-0.1563364416,0.2840083838,-0.2199819386,0.0735880211,0.0627435222,0.0822078958,-0.3243557215,-0.0924052522,0.0356562063,-0.3760551512,0.1180931032,-0.405501008,-0.1406731755,0.4104865789,0.1494239867,0.0048129051,0.2804119587,-0.2361316085,0.1205309927,-0.0468914323,0.0120421099,-0.0608828105,-0.5602765083,0.4077919126,-0.0525955744,0.0190129001,0.1539668739,-0.1636202335,-0.0806388482,-0.0600142442,-0.1388378888,-0.359738946,0.0399610177,-0.0283835325,0.3964413106,0.3382200897,0.0393798426,-0.0122915842,0.1701073796,-0.1517384648,-0.3298409283,0.1699948311,-0.0135946991,0.2757686079,-0.1529276818,0.1247687489,-0.2460490316,0.6183304191,0.1462423801,-0.2673268914,0.0889830217,0.2282872051,0.2022008151,0.238674432,0.0776953325,0.1591652334,0.0495318361,0.1625450402,0.1308613867,0.269672662,0.3977327645,0.0793766305,0.203293249,0.0485626906,0.0449625924,-0.0842107832,0.1603960544,0.2533026338,-0.2020492554,-0.2511749566,-0.4260875583,0.0768298954,-0.1592129767,0.1497004479,0.2452454716,0.2454757988,-0.6999825239,-0.0805011243,-0.4132456779,0.1729496121,0.2795114219,-0.2491959631,0.0677825212,-0.1459705085,0.0437211692,0.0090917628,0.279209435,-0.3869775832,-0.3823457956,-0.1146598831,0.0667323396,-0.0785387829,0.0902425721,0.023412643,-0.3819142878,0.2053471804,0.357078433,-0.5349148512,0.0449987315,0.2951855659,0.0593914092,-0.1521806866,-0.155501768,-0.2612583041,-0.4407454431,-0.2159960568,-0.0621569827,0.1316289455,-0.0885623395,0.0866670609,0.0210243557,0.0464734249,0.0607404932,0.3342549503,0.1943166405,-0.0947806537,0.5751731992,0.1907665282,0.44811517,0.0067705489,0.5023270845,0.5508075356,-0.4038015604,-0.4175478816,-0.2469518483,-0.2677246928,0.5048151612,-0.0427201837,0.3125376701,0.1756529659,0.090309687,0.0931700543,-0.5563325882,0.406506449,0.0986593887,0.1042241305,-0.3072217703,-0.4790277183,0.645817399,0.149361521,-0.0848046318,-0.0854144022,0.506028831,-0.1912818104,-0.1578022689,0.3315126896,0.9062218666,-0.3334765434,0.3655091822,0.0131009361,0.1562291086,1.0811457634,0.0033211063,-0.256093055,-0.0159779154,-0.1378635466,-0.0842902437,-0.0985352471,-0.0891081691,-0.0119750863,-0.0431570336,0.186240375,-0.1342516989,-0.0050168773,-0.0737391114,-0.052968841,0.242264241,-0.3497599661,-0.0392264724,-0.0075751999,-0.2592848837,0.00220139,-0.1769247502,-0.1001577601,-0.1548250169,0.2414825857,-0.1008635014,-0.1672659218,-0.0238416158,-0.4493339062,0.0524116158,-0.0501913503,0.1648599207,-0.113325119,0.7807733417,-0.0103021953,-0.0261253528,0.0547878295,-0.3196213245,0.0584369563,-0.1833813041,-0.1899094582,0.0945091397,0.1322172433,-0.2486992925,0.0619627275,0.2867776453,-0.2255601734,-0.4018236995,-0.2655332685,0.143485859,-0.1803411394,0.2619341314,0.528208673,0.1343051642,-0.165765956,-0.0139381681,0.2420060784,-0.1624602675,-0.1309479922,-0.3226356804,0.0755101591,-0.2600620389,-0.1361034513,-0.079211615,-0.0575200841,-0.06330733,0.1612929255,-0.0311610606,0.0492313951,-0.0930720046,0.1090695336,-0.3015504777,-0.1558669209,0.2778174281,-0.505035758,-0.41854617,0.3102189302,0.3964101374,-0.3695798814,0.2166962177,0.1266154647,-0.5569292307,0.1624207199,-0.0688453466,0.2214820236,-0.0179772433,0.385145247,-0.1129819974,0.2741782367,-0.1475791037,-0.0971751437,-0.130520612,-0.0183126703,0.1846455634,0.02937489,0.1764271408,-0.1242447793,-0.153591007,0.0654529855,-0.0220909119,0.0473726429,-0.3549407423,0.2310741395,0.3131256402,-0.4221338332,0.1323610693,0.0865930021,0.0741610229,-0.1343493164,0.3031973243,-0.0629230291,0.0735789463,0.3213457167,-0.1773859411,0.3383938074,-0.030060336,0.2580043674,0.3491809368,0.2711409628,0.3097110987,-0.0361437798,-0.0033085507,-0.1242608652,-0.1788883656,0.2156116068,0.2669431269,-0.1784222424,0.2639374137,-0.0269617662,0.0859639421,0.3056539297,0.2354824841,0.1830295324,-0.1293051392,0.1637925655,0.2418400347,0.0689755604,0.0014420775,-0.1497674584,0.0030115966,0.060351111,0.1754912883,0.4341250658,0.2106015831,0.4329251647,-0.0293114558,-0.063058421,0.2237009406,0.1672794819,0.0849507824,0.0849132836,-0.0175961517,0.1635657996,0.48567608,0.4265227318,0.0674312562,-0.0160254426,0.1899795979,0.6487360597,0.0044181072,-0.0524943508,0.1508476138,0.028226098,0.1753160208,0.3093416095,0.1775844097,0.0783883408,-0.0537376255,0.0474493392,-0.382368803,-0.2152680457,-0.0376274288,0.3847441971,0.169697091,-0.3462902009,-0.298066318,0.0780194327,0.0251304638,-0.1415096521,0.0557551011,-0.7503657937,0.285649091,-0.1025271937,0.1747113317,-0.4682204425,-0.5723503232,0.3234819174,0.3111141026,0.078252621,0.157945767,0.4032856822,-0.1206088588,0.2232599407,0.2052865922,0.1001311168,-0.2368672788,-0.004581355,0.3985375464,-0.3177670836,0.1261327416,0.0327679478,0.0902690515,-0.144632861,0.0540255122,0.2588135898,0.0035147762,0.0087617459,-0.1507186145,-0.2779210508,0.0636352003,-0.2348604202,0.4316190779,-0.3251055479,-0.1927683502,-0.1062208414,0.3080970347,-0.3092028797,0.4655345976,0.3847272694,-0.0573392808,-0.0606169403,-0.1608618051,-0.0399090722,-0.0662580431,0.3288370073,0.3810606897,0.3720329106,-0.2288176715,0.0549941026,-0.3086269498,0.0609697625,-0.0616751723,-0.5124160647,-0.091246143,-0.2879363894,0.3729308248,-0.0900705382,-0.1256012917,-0.1805207878,-0.023457434,-0.2921602726,-0.1426432133,0.0222477112,-0.312833339,0.2529564202,0.1500561386,0.1267486066,-0.1434247643,0.3120370507,-0.165549174,-0.0364412181,0.2860501409,-0.2266689539,-0.0232150629,0.1433687508,-0.0475377142,0.4864451885,0.0999650136,-0.2336501628,-0.0023351919,-0.0677524656,-0.3144228756,-0.0466906503,-0.3797927797,0.2594606578,-0.268569231,-0.0020491458,-0.241060406,0.0829983726,0.1157295257,0.0540380217,-0.29678002,0.4357704222,-0.2345369309,-0.2568191886,0.5996598005,0.3250459433,0.0987215787,0.0490599796,-0.0345505066,-0.0198894236,0.3759042621,-0.07449352,-0.0791880414,0.0058689695,0.1684659719,0.4147238433,-0.1418843269,-0.8466008306,0.2211653888,0.3185895383,0.0548537225,-0.5564775467,0.2264992595,-0.0804183856,-0.4305396676,-0.0225379188,0.3301892281,-0.2793427408,-0.0013932257,0.0352891125,-0.186635986]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/723","title":"Adding pseudo-labels to datasets","comments":"Sounds just like a regular dataset to me then, no?","body":"I recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n","comment_length":10,"text":"Adding pseudo-labels to datasets\nI recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n\nSounds just like a regular dataset to me then, no?","embeddings":[0.3098163605,-0.073686637,0.0788327679,0.0502193421,0.1483774483,0.064456068,0.2848760486,-0.0867986232,-0.176591754,-0.113085106,-0.2497952729,0.0274977945,-0.138913691,0.5297517776,0.450722903,-0.0800261945,0.3305429518,-0.1942687333,0.2078152448,-0.2288157642,-0.3326999247,-0.3138588369,0.2468191385,-0.1605732143,-0.3161626458,-0.2350748479,-0.1671558022,0.0179832354,0.0190474633,-0.5172463655,0.0243184548,0.3594777882,0.2320507169,0.6230777502,-0.0001245047,-0.0930988044,-0.1576466262,-0.1632835865,-0.1270368546,0.2477617413,-0.0783547908,-0.2267749161,-0.1635601521,0.1813987792,-0.1364328861,-0.3832185566,-0.0941722244,-0.0087911747,0.1753597707,0.0393101461,0.0270005185,0.1182208806,0.2331093699,0.1397381872,0.4637860358,0.3221139908,-0.2436993122,0.3084616661,0.2786679566,0.4386538565,0.4769410789,0.4095715582,-0.0336054899,-0.1056167036,0.6208615899,0.0606488958,-0.0060085803,-0.1803056598,-0.3337912261,0.2313068062,0.4991978109,-0.2710779011,-0.7025510669,-0.4659937918,0.0760054886,-0.3196178973,-0.1870369166,-0.1058692783,0.3414154649,-0.0471548662,0.003702641,-0.2444640547,-0.0227476377,0.1859458387,0.0180374496,0.5205722451,-0.1851842701,0.2339127511,-0.003882484,-0.2170413435,-0.059614256,-0.2445372343,0.1620963216,0.1288419068,0.1126239821,-0.1607144177,-0.0765358284,0.0234620739,0.3366958201,0.3094500005,-0.0428376757,-0.143932417,-0.3169184327,-0.1127314568,-0.0114139076,0.0899884552,-0.0004383897,0.2661900818,0.2482797652,-0.0281014033,0.2920183539,-0.0877215788,-0.509095192,0.546718955,-0.0522184223,-0.2023925334,0.0438271612,-0.0896684676,-0.0073592081,0.0333210863,0.113646321,-0.1795961261,0.043670319,-0.0328714512,0.0484431647,0.0181626901,-0.1198237017,0.4060238302,0.2424546629,0.0486480072,-0.0560292304,-0.1219252348,-0.0949172229,0.3752702177,-0.0604504719,-0.3154687583,0.2706238627,-0.2346781492,0.1446810067,-0.2384348363,0.069081746,0.1547600776,-0.0132015999,0.2396950424,-0.1269270927,0.1267541796,0.0188518353,0.0030103107,-0.23264952,-0.1019717008,-0.0473884791,-0.0878580362,-0.3120549321,0.0115569206,-0.4288146794,-0.3029751182,-0.3602036238,0.2782149017,0.1989504993,-0.0586553328,0.1180477589,0.172534585,-0.0766597465,-0.1462652832,0.250947386,0.1416517943,-0.3954949677,-0.2266838998,0.1373609006,0.0454571582,0.1233935282,0.2846758664,0.0137401195,0.1591172814,0.1665392518,-0.0563894473,-0.0785102397,-0.3177955747,-0.2149504274,-0.0107226148,-0.0421728343,-0.0987236276,-0.0162170008,0.4997740686,-0.0344073251,-0.2872255147,-0.0855292156,-0.0388564244,0.0156851951,-0.0605357103,-0.0187013112,0.0269505661,-0.1122945771,0.4192120135,-0.1120919362,0.0362510867,-0.0967117846,-0.1454378217,-0.0679651573,-0.0335562639,0.0626452714,0.0215310305,0.4627296329,0.4635120928,-0.1319904029,-0.3320657313,-0.4214040935,0.0558236055,0.3083326817,0.1856974661,0.2704796493,-0.4083002806,-0.2508350611,-0.1292894334,-0.1484947652,-0.0652147979,-0.1003694385,0.2113861889,0.1631660908,-0.0453920849,-0.3714341521,-0.1713396907,0.052831132,0.2452875078,-0.0778191239,0.1458304375,0.1956076324,0.0783619508,-0.1884623766,0.5431229472,-0.0114947222,-0.0809042752,0.2427551746,0.2535338402,0.0497574545,-0.1231769919,0.0735284686,0.3053961694,0.4409328997,-0.0558783039,0.095363833,-0.2860271037,0.0262372959,-0.1562604606,-0.1547542959,0.0229252838,0.1678315848,0.1422424018,-0.2536532283,-0.1215398833,0.0439311042,0.1698319316,-0.2101525068,0.0330372937,-0.065046981,0.2022144347,0.4685901701,0.0883104429,-0.1373460144,0.3082611859,0.0281754695,-0.3854891956,0.1695020348,0.1227527261,-0.5392128825,0.2260270864,0.0100205131,0.4092067778,0.0080515621,0.0682568923,0.3067458272,0.0199515019,-0.0524558574,0.0600802153,-0.1393863708,0.3049429655,0.0571600534,0.0458538309,0.0370499119,0.1020892262,0.1249625608,0.0351284929,-0.004118551,-0.0209866893,-0.2624418139,-0.0825800076,0.2878836989,-0.0429316983,-0.2284687757,-0.5607491732,-0.4760740399,-0.2232433259,0.023550598,-0.1197864711,-0.1296577752,0.1882886738,0.0621396415,0.4003103077,-0.2920649052,-0.2911193073,0.2338989228,-0.1990503967,0.1824574769,-0.1028177366,0.1352082938,-0.1243391335,0.4377942681,-0.142568931,0.1274449825,-0.2115978897,-0.5272506475,0.1104164124,-0.5111126304,0.1881586462,0.2846199572,-0.1162844002,0.2558780015,-0.2260627151,0.0801432803,0.1272312105,0.0782635733,-0.2512591481,-0.133881256,0.0368195139,-0.3204395175,0.1230423152,-0.4065716565,-0.1351296455,0.3523740172,0.1700830162,0.0051358044,0.3083994687,-0.2543621659,0.083268784,0.0155996196,-0.0244372804,-0.0969084576,-0.4898469448,0.4022898078,-0.0309677459,0.0306624789,0.1679752171,-0.1609854549,-0.0378445536,-0.0580840446,-0.11693912,-0.3661031425,0.001762769,-0.0091450037,0.393114388,0.3475563228,0.137969017,0.0126321595,0.1610682607,-0.1280919015,-0.3824179173,0.149972558,-0.0020157297,0.2357483506,-0.1752831638,0.1630832553,-0.2472160757,0.637722075,0.1977131814,-0.2878910005,0.0773420781,0.1941648275,0.2422816902,0.2006517351,0.0659928769,0.1496909112,0.0586045794,0.0593225621,0.1145875752,0.2662224472,0.3589520454,0.0818890631,0.2087306827,0.0325850695,0.010980133,-0.0578490384,0.1123771369,0.2750446796,-0.183592841,-0.2388730645,-0.4218675494,0.0581412539,-0.1841807812,0.2141856253,0.2380071729,0.19958736,-0.6758162975,-0.0205183346,-0.4138442278,0.1781001985,0.2200511545,-0.2674053311,0.090525426,-0.1286990941,0.0429533683,-0.0176688023,0.2548457682,-0.3990539908,-0.3311429024,-0.0952272341,0.1338106394,-0.1348559707,0.1069837064,-0.0147903189,-0.3166734874,0.1980721802,0.340883702,-0.5565126538,0.056778457,0.2829640508,0.0571998321,-0.1577194929,-0.1797414571,-0.3056754768,-0.5021149516,-0.1930465996,-0.1195303425,0.159616515,-0.1527983099,0.1589968204,0.0032353166,-0.01403559,0.0751755685,0.3526730239,0.2016186714,-0.065966785,0.5528895259,0.1689908504,0.3891877234,-0.0374332778,0.4574795067,0.6365597248,-0.3747137785,-0.3864565492,-0.2323791832,-0.2896787226,0.4548307657,-0.0600494929,0.3047736883,0.1004615277,0.1085141599,0.0226362366,-0.5715113878,0.3784736991,0.1348147541,0.1017560959,-0.3226232827,-0.5324237347,0.6639290452,0.1734095216,-0.1657990366,-0.0975202098,0.5208674669,-0.1833907068,-0.0676982626,0.3497229517,0.9251251817,-0.3402081132,0.3692504168,0.089386642,0.1624438465,1.0703035593,-0.0549558848,-0.2816874385,-0.0436497964,-0.1256705523,-0.1385435164,-0.1378173977,-0.1130425408,-0.0050831805,-0.0380305797,0.2086778432,-0.1540289074,0.0427259803,-0.0325813591,-0.0309839696,0.2736977339,-0.3206111491,-0.0463141762,0.0106112594,-0.2755618393,0.0060173557,-0.1870793402,-0.1040566415,-0.1724939793,0.2346333712,-0.1657537669,-0.1942425668,0.0065937792,-0.4786692858,0.0298513267,-0.0655192211,0.1125091314,-0.1437826008,0.764765203,-0.008483951,-0.0440957583,0.0978011861,-0.3074002564,0.0603915118,-0.1638819724,-0.2074524313,0.0873991698,0.1504178196,-0.2351899892,0.007990825,0.2446397096,-0.2272354066,-0.3556442261,-0.2518899143,0.1198010072,-0.2442423403,0.254879266,0.5359111428,0.1437408775,-0.1427090317,-0.0064275791,0.2152624577,-0.13534154,-0.1334627867,-0.3340030611,0.1401178688,-0.2746910453,-0.1642646044,-0.0705529675,-0.0748223588,-0.08573661,0.1579608321,0.0020001312,0.0321660675,-0.0521001741,0.1497536749,-0.2563918531,-0.1615019739,0.236702323,-0.5272604227,-0.3810845315,0.3116165698,0.4240096807,-0.350061655,0.1770195663,0.1268406808,-0.5051674247,0.1304293424,-0.0502328277,0.2248882055,-0.03009177,0.3470942676,-0.1093998328,0.2857561111,-0.1585695446,-0.1217036247,-0.1107746139,-0.0179893449,0.2011557519,0.0281385295,0.1726350486,-0.1554268599,-0.1318470836,0.0203533806,-0.0412121527,0.0225128513,-0.3279897869,0.2265195101,0.3207247257,-0.4298757613,0.1708727926,0.0904043764,0.0297566224,-0.1469672173,0.2935111821,-0.077819258,0.0759380907,0.3098110855,-0.1754194498,0.2969334424,-0.0640214607,0.1832231283,0.295230031,0.2794476449,0.3251145482,-0.0373063944,0.0460680127,-0.1093309447,-0.1085442156,0.2416185439,0.2581656277,-0.0821704417,0.2647285461,-0.0399412178,0.1047691703,0.3044362664,0.2879047692,0.2022820711,-0.1490662992,0.1566319019,0.1962457746,0.0966201052,-0.0308438428,-0.158583805,0.0253886729,0.0791772678,0.2121003121,0.424436152,0.1952596158,0.4297283888,-0.0370261744,-0.0806300864,0.3002631664,0.1621903181,0.0880272388,0.1017729864,-0.0581912771,0.1491179913,0.4947297275,0.4353736639,0.0675814524,0.064937517,0.1865001321,0.5670723319,0.0219700448,-0.105205521,0.201448068,0.0079528065,0.2051110119,0.3066345453,0.2205009609,0.0318184905,0.0051765819,0.0112079559,-0.3440894186,-0.2327881455,-0.0553414635,0.3427706659,0.1720449477,-0.2845652401,-0.3234486878,0.0794580728,0.0455606915,-0.1957406998,0.010358125,-0.7603212595,0.3236016929,-0.1173227206,0.2239730954,-0.4359399676,-0.4555698037,0.260270834,0.2396520972,0.0969375148,0.2058233619,0.4202291071,-0.1685076207,0.1724863499,0.2137111723,0.1304792464,-0.247311905,-0.0065970519,0.4755756259,-0.3449179232,0.1098707169,0.0376648083,0.0394646004,-0.2256655246,0.0678499192,0.2151242346,0.0179233216,0.020522248,-0.1157146692,-0.3120968342,0.0303748082,-0.2730508447,0.4743926525,-0.3677545488,-0.2162243873,-0.0857257098,0.2981347442,-0.3052148521,0.4317930341,0.428373456,-0.0770943463,-0.0451779179,-0.1449841559,-0.0278286412,-0.0611798428,0.3342255652,0.3694381416,0.3831316233,-0.1900883168,0.0611990169,-0.375887841,0.0878613889,-0.1073901057,-0.5902005434,-0.0892133564,-0.2459474951,0.3976393342,-0.0479489677,-0.1397959143,-0.179437235,-0.1012513638,-0.2905360758,-0.1318623126,-0.0567958504,-0.3264635205,0.2493855506,0.1107960939,0.0886786059,-0.132731095,0.3213802576,-0.1418015361,-0.0289692394,0.231187582,-0.1831809133,-0.0427688211,0.1266191602,0.0065209074,0.4872638583,0.07805641,-0.2360869199,-0.0023194493,0.0006455186,-0.3507651985,0.0020263558,-0.3973657191,0.2501219511,-0.2804778218,0.0006991252,-0.261313498,0.0633928999,0.1567376107,0.0762818679,-0.3622764349,0.4714178443,-0.2266539335,-0.2493025661,0.527005434,0.3202368319,0.0835427493,0.0520756431,-0.0162507929,-0.084172152,0.379681915,-0.0524729155,-0.0951329842,0.0781966299,0.1298217028,0.4085799754,-0.1211648211,-0.8176765442,0.2679797709,0.3405527771,0.0760081112,-0.5392848849,0.2038773149,-0.0499905087,-0.4629510641,-0.0578464903,0.3135511577,-0.2662079334,-0.0460031778,0.0242788512,-0.2036228776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/723","title":"Adding pseudo-labels to datasets","comments":"A new configuration for those datasets should do the job then.\r\nNote that until now datasets like xsum only had one configuration. It means that users didn't have to specify the configuration name when loading the dataset. If we add new configs, users that update the lib will have to update their code to specify the default\/standard configuration name (not the one with pseudo labels).","body":"I recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n","comment_length":65,"text":"Adding pseudo-labels to datasets\nI recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n\nA new configuration for those datasets should do the job then.\r\nNote that until now datasets like xsum only had one configuration. It means that users didn't have to specify the configuration name when loading the dataset. If we add new configs, users that update the lib will have to update their code to specify the default\/standard configuration name (not the one with pseudo labels).","embeddings":[0.2998081446,-0.1344268322,0.1064767763,-0.042103827,0.1416293383,0.0385911241,0.3353704512,-0.0611251034,-0.1213262454,0.0122373067,-0.2308751792,0.1057627127,-0.1279500425,0.5004864931,0.4318877757,-0.072646223,0.3291837275,-0.1554969102,0.2026746273,-0.154693082,-0.3575718999,-0.2737482488,0.2698325813,-0.1786535531,-0.320311904,-0.1530624032,-0.1499094367,-0.0244253874,0.1104237139,-0.5956424475,0.03784661,0.4156611264,0.2552714348,0.5836076736,-0.0001237342,-0.089877069,-0.1095676944,-0.1492246985,-0.1526645422,0.186216265,-0.0167718735,-0.2705730498,-0.103080228,0.1222240925,-0.1497649848,-0.3686211109,-0.0767853931,-0.0310839936,0.1511062682,-0.0925186798,0.0303302091,0.1451057494,0.1626512855,0.1272767633,0.3680979013,0.2810037732,-0.2382881194,0.3393152356,0.2810574472,0.4066390097,0.4596785903,0.495606482,-0.0231238827,-0.0686169863,0.5873855352,0.0751341805,0.0005662205,-0.1729782373,-0.3276819587,0.2396902293,0.4566649497,-0.3164331615,-0.6837278605,-0.5047792196,0.186757043,-0.3624064624,-0.0957857668,-0.2155633718,0.3606551588,-0.0346914567,0.0314153694,-0.2650181949,-0.0182420593,0.1895082146,0.0594545864,0.5780060887,-0.1692253798,0.2320201695,0.0101037323,-0.2667593062,-0.0918954983,-0.266084522,0.1848731637,0.0811003968,0.109753862,-0.1387264729,-0.1134848222,-0.0457396507,0.2438456714,0.2950548232,-0.0437582806,-0.0620150231,-0.2086888105,-0.1097071916,-0.1035718396,0.120954074,0.0874914825,0.2561557293,0.2847910523,0.0440343879,0.3656741679,-0.0779900178,-0.4416409433,0.5037554502,-0.1021510139,-0.1318588704,0.0779781863,-0.0652201772,0.0415383391,-0.0243347026,0.1696671098,-0.2140185833,0.0382408239,-0.0785938352,0.0153718702,0.0808275193,-0.0493975282,0.3345745802,0.1719156206,0.0234977677,-0.0198372286,-0.1612987965,-0.0988189727,0.364339143,0.0371232517,-0.3528376818,0.2718880177,-0.2289155722,0.1966086477,-0.2133773714,0.0106298393,0.2160994411,-0.0613005795,0.22638309,-0.2178615183,0.1044051498,0.0036458736,-0.0238881763,-0.2566460669,-0.1206147447,-0.0953300893,-0.0889450163,-0.324175477,0.0147484932,-0.4072446823,-0.3032717109,-0.3384978473,0.3656526208,0.1131786928,-0.0525872856,0.0730779171,0.1723829359,-0.0328396074,-0.134907037,0.2840846479,0.2005071789,-0.3355906308,-0.3381416798,0.1224144846,0.0566466339,0.078985855,0.1605279148,0.0321430862,0.0978841782,0.2206550688,-0.0889733806,-0.067524299,-0.382614553,-0.1448272616,-0.0675379038,-0.0896222368,-0.0237168763,0.0551170558,0.4480420351,-0.0472856201,-0.2814720869,-0.1423392445,-0.0717054978,0.0374272801,-0.0787703171,-0.0332069062,-0.0672871247,-0.144183591,0.4287656546,-0.0894117802,0.0541012548,-0.0772441998,-0.1134878173,-0.0747608393,-0.0645027384,0.1017478332,-0.0114808753,0.4220195115,0.4557925463,-0.0916468725,-0.2570293248,-0.4810330868,0.1187712476,0.3459650278,0.1877480745,0.2587982714,-0.3725942075,-0.2178406566,-0.18145293,-0.1556061953,-0.0693745464,-0.1133408993,0.2501719892,0.1705455929,-0.0530524775,-0.3825318217,-0.1964839101,0.0427181721,0.2877751589,-0.1020270437,0.0700213835,0.1953106821,0.0370659307,-0.2225669771,0.4910910726,0.0013808043,-0.1018630192,0.1985341907,0.3084602654,-0.0033228048,-0.0713012293,0.0395375341,0.3227552176,0.4791212678,-0.0398230702,0.153262645,-0.2845877707,-0.0183797833,-0.0789748728,-0.1487845182,0.0462142751,0.2128414512,0.1702634394,-0.3061312139,-0.1273018569,-0.0200702492,0.1623200178,-0.2193689942,-0.0518594421,-0.0748730451,0.1818132997,0.4083220065,0.0816129372,-0.2233512402,0.2831476629,0.0431273393,-0.3308486938,0.0959251076,0.1368115544,-0.4195856452,0.2700432837,0.0196660664,0.437004745,-0.0176327806,0.0807090253,0.2086809725,0.0106297815,-0.1043510288,0.0108252103,-0.1265753955,0.2933960557,0.1004739404,0.0300719496,0.0111526167,0.1413501799,0.1606981009,-0.0773590207,0.0576363541,-0.1008791625,-0.2513286769,-0.1020588428,0.1794400215,-0.0913679004,-0.3065832257,-0.5622990131,-0.5115970969,-0.2192707658,0.0635916889,-0.1512479633,-0.2140166014,0.3226282001,0.0542418063,0.3781419098,-0.3508310318,-0.2711749673,0.1680888981,-0.1181780621,0.1399009079,-0.123668462,0.0609191768,-0.2141324133,0.4188081026,-0.1850165576,0.1274848729,-0.2075192034,-0.5780903101,0.1039367914,-0.4823778272,0.2204030305,0.2675939798,-0.1833579689,0.3257483244,-0.2888326645,0.1066211313,0.1330201775,0.1489219964,-0.3089151382,-0.1425167471,0.0886036679,-0.320543617,0.0195567682,-0.3618177474,-0.1463156641,0.3072495162,0.1251452565,0.0840290487,0.3436688483,-0.24739483,0.0715096071,-0.0873192921,0.0665148124,-0.0324894264,-0.5351750255,0.3863258064,-0.0446912907,0.1066346914,0.1578733027,-0.1220841855,-0.0349110253,-0.0149162663,-0.0934661776,-0.4165671766,0.055048883,0.0668338463,0.3598328531,0.3506225348,0.0857239813,0.1384339929,0.1219525859,-0.1305479556,-0.4114057422,0.1381793916,-0.0073749246,0.2347899228,-0.0767000094,0.1047660783,-0.2347685695,0.6055535674,0.1378824413,-0.2848439813,0.1233670413,0.213933453,0.2773593366,0.2527821362,0.0802338943,0.1231472194,0.09794043,0.1768232435,0.1326763481,0.2246923447,0.330196172,0.0223206636,0.2020407021,0.0368253961,-0.0451365113,-0.061231479,0.0891546682,0.2640226185,-0.1869099587,-0.1962648928,-0.4515404999,0.0963482633,-0.1350032836,0.2445066869,0.1784919202,0.2270118594,-0.6722140312,-0.058122173,-0.373331815,0.1896591634,0.2063449472,-0.2074152827,0.0756509304,-0.1758773178,0.0184060428,0.0383834019,0.2505616844,-0.3790582716,-0.375082761,-0.0604158714,0.1172766685,-0.0674980432,0.047480341,0.0205092989,-0.3828158379,0.2052793354,0.3741309345,-0.5892170668,-0.065943636,0.3306910396,0.0927901864,-0.1313436925,-0.2216567248,-0.2396981269,-0.452411741,-0.2524763346,-0.1162350401,0.0258924458,-0.0764975771,0.12029998,0.0198930725,0.0247088317,0.0644343495,0.3508708477,0.2437824905,-0.0878818035,0.5968904495,0.1840090305,0.4270557761,0.0044480646,0.5245034695,0.5270329714,-0.4492975175,-0.4075215757,-0.2239102125,-0.2210036367,0.4601719677,0.0576396883,0.317401588,0.1746520251,0.1301114857,0.1193813607,-0.5778493285,0.3765330017,0.076203227,0.1317571998,-0.3012309074,-0.5492858291,0.67757833,0.1472238004,-0.1553729922,-0.0644810721,0.5425695181,-0.2256025374,-0.0472144745,0.3019182682,0.9438102245,-0.3285489976,0.2973953485,0.0565994903,0.129018113,1.0668774843,-0.042660933,-0.2235747278,0.0169675853,-0.0606753863,-0.0645463243,-0.0610054247,0.0100098271,-0.0399791561,-0.0866565481,0.2626317143,-0.1319755167,-0.004861223,-0.0524095967,-0.007408042,0.1156437621,-0.443428129,-0.0620699227,-0.0200740527,-0.3278385699,0.0460784584,-0.1779173166,-0.106475316,-0.1384462267,0.1808561236,-0.0782787576,-0.2005703151,0.0090066567,-0.4644095302,0.0459138379,-0.0538132899,0.1084043309,-0.1514496356,0.742577374,0.0360546336,-0.0078963386,0.060147617,-0.2602529228,0.0951503441,-0.2032085657,-0.2163182199,0.1532936096,0.1491838396,-0.303573072,-0.0210453533,0.2473726273,-0.162926659,-0.3291380703,-0.3351070583,0.18148534,-0.1905628145,0.3235607445,0.4867148697,0.2079122514,-0.176824376,-0.0023308499,0.1727014184,-0.1673434079,-0.0578976236,-0.3185966015,0.0766081586,-0.299900353,-0.180229798,-0.1048278362,-0.074519679,-0.0544158816,0.1132205278,-0.0107801287,0.0677029267,-0.1136699617,0.054384578,-0.3007849455,-0.2121462673,0.2832657993,-0.5372553468,-0.3484558761,0.3236203194,0.4322231412,-0.3114257455,0.1920561939,0.1686388701,-0.4940820634,0.1778880954,-0.0366944894,0.2591643035,-0.0096713714,0.4078370035,-0.123025395,0.2761672139,-0.17470254,-0.1363597214,-0.0695876405,0.0363732278,0.1963931322,0.0861264989,0.1853114218,-0.0782117918,-0.1013666615,-0.025417529,-0.0959804505,0.0536715388,-0.4087136388,0.2254897654,0.2022174001,-0.3918568492,0.0987538919,0.0750087947,0.0946207941,-0.1908841133,0.3108455241,-0.0324766003,0.0353285447,0.3759314716,-0.1677114218,0.2833200991,0.021812791,0.2113074213,0.319945842,0.2520157993,0.2788857818,-0.0036710238,-0.0880028978,-0.0903637707,-0.0697052032,0.2032179534,0.1966564059,-0.0789690018,0.1939894259,0.0197473411,0.0650335252,0.3351961672,0.2463516146,0.183452338,-0.1384530067,0.0547075197,0.2898483872,0.0893732831,-0.01775527,-0.1585976928,0.0265491046,0.0859427899,0.227418527,0.4988335371,0.2685062587,0.3956277668,-0.0129990522,-0.0677447096,0.2840413153,0.0716920346,0.1701555252,0.0837532803,-0.0594907589,0.1950876266,0.5000240803,0.4496957064,0.0386446305,-0.0388258174,0.1041432694,0.5346742272,-0.0056027924,-0.062078286,0.1263550222,-0.0043645725,0.2718457282,0.2628068924,0.1722565889,0.1215167716,0.0264301263,0.0776845515,-0.3849411607,-0.1851020008,-0.0095245428,0.3350860476,0.1930453032,-0.3616013527,-0.2882723808,0.089672491,0.0271635037,-0.1577624381,0.0481804758,-0.7170671225,0.3460785747,-0.051770255,0.2031672001,-0.5002550483,-0.4857996106,0.2336293161,0.3497897089,0.0259191729,0.1968863159,0.4087084532,-0.1984284073,0.2039638162,0.2398054153,0.1072587445,-0.2287327945,0.0653764606,0.4185298979,-0.2756542563,0.1485028863,0.0549340658,-0.0024057608,-0.1555739343,0.0931705534,0.2085333765,0.0101444302,-0.0063745938,-0.1862779409,-0.3286513686,0.0274994206,-0.2652023733,0.4785491526,-0.2938702404,-0.1582933515,-0.056515608,0.3390560746,-0.3580414355,0.4116491675,0.443757385,-0.0325084068,-0.1012703702,-0.1758369952,-0.0251276456,0.0027255537,0.2918702066,0.2755597532,0.2722046971,-0.2335116416,0.0860366002,-0.3426698744,0.0365585238,-0.0267138444,-0.5635901093,-0.1061020792,-0.3014020324,0.3944289386,-0.1160802469,-0.121190846,-0.2254061848,0.06929221,-0.2247915417,-0.1723793894,-0.0211837851,-0.2945216,0.2290796787,0.1749406159,0.1203293577,-0.1489344686,0.2933540344,-0.1696756929,-0.0576495789,0.2612989545,-0.0561523885,-0.0316381678,0.1851319969,0.0004077401,0.4343340397,0.0401111171,-0.3034462333,-0.0748706684,-0.082951583,-0.3774099052,-0.0144115156,-0.356215477,0.2779452503,-0.2749114335,0.0190288853,-0.3525388539,0.0970998853,0.0886838734,0.0865051821,-0.2903698385,0.4144820571,-0.1972637624,-0.2823800743,0.5451181531,0.3401219547,0.1585597843,0.1627668589,-0.0933095962,-0.123703666,0.4480282068,-0.0308442991,-0.0846610963,0.0062058331,0.1697437465,0.4019327164,-0.0590473078,-0.8440304399,0.2708303332,0.3358246088,0.058369305,-0.535623312,0.2374904454,-0.1027968451,-0.4529278278,-0.0601694994,0.3132772744,-0.2470305413,-0.0735243633,-0.0563191883,-0.1896903813]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/723","title":"Adding pseudo-labels to datasets","comments":"Could also be a `user-namespace` dataset maybe?","body":"I recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n","comment_length":7,"text":"Adding pseudo-labels to datasets\nI recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n\nCould also be a `user-namespace` dataset maybe?","embeddings":[0.3372250795,-0.050370343,0.0705883056,0.0136920055,0.1387504041,0.0619150139,0.3166058362,-0.0735883862,-0.1538384557,-0.0340620652,-0.2433054298,0.0398036949,-0.1421181411,0.5068695545,0.5139107108,-0.0928782076,0.3329214156,-0.161922425,0.1973849833,-0.17973046,-0.3573174179,-0.2903221548,0.2575227618,-0.1741076112,-0.3187001944,-0.1787993163,-0.1542833149,0.0150797796,0.0611836798,-0.5726603866,0.0432887599,0.374940455,0.2304802388,0.64429003,-0.0001240043,-0.0763929859,-0.1512939334,-0.1524621844,-0.1376603246,0.2009308338,-0.0528901331,-0.218797043,-0.1666469127,0.1626427323,-0.1171764508,-0.3696009517,-0.0746916831,0.0174398869,0.1938840151,0.0266593266,0.0249343142,0.1186638847,0.2309546918,0.1051767245,0.4362747669,0.3171413839,-0.2496043295,0.3025824726,0.2650249004,0.3800933063,0.4628627896,0.408708185,-0.0102578495,-0.1446854621,0.616460681,0.0513366647,-0.0366443843,-0.1837252527,-0.3468540609,0.2112408578,0.5033496618,-0.2692509294,-0.6825976372,-0.4553024471,0.0848091543,-0.2915610075,-0.1433606446,-0.1595896482,0.3417896628,-0.0638809726,0.0224082991,-0.2425260246,-0.0692846105,0.2046854049,0.0658252239,0.5435054898,-0.1693327129,0.2218388617,0.0300021507,-0.2402008176,-0.0781440362,-0.2168484926,0.1853793561,0.125719443,0.1152946427,-0.1544454396,-0.0584880039,-0.0068262354,0.312628597,0.3015111089,-0.043009419,-0.1620018631,-0.2468457967,-0.1121434122,-0.0771837384,0.1189124063,0.0210794862,0.2551678121,0.269862771,-0.0218350384,0.3048101664,-0.1139756814,-0.5342983603,0.5411620736,-0.0541201048,-0.2128379047,0.0036297694,-0.0625899732,0.0031487742,-0.0051162941,0.1455241889,-0.1451524943,0.0581165031,-0.0149584515,0.0637847856,0.0147829531,-0.0999502316,0.3897129595,0.2270878702,0.0200959817,-0.05121959,-0.134305492,-0.1395832598,0.3545896113,-0.0194652453,-0.2863017619,0.3018845916,-0.2576481998,0.1924594641,-0.275267154,0.0817670301,0.2055664808,-0.0544824935,0.2389345914,-0.1204430535,0.1105919778,0.0214736797,-0.0027890599,-0.2402633131,-0.1197118163,-0.0554947704,-0.0945935324,-0.3270657063,0.0155994464,-0.3937909007,-0.3156652451,-0.3782431185,0.2677690685,0.1736882031,-0.0418858565,0.1006088108,0.1814104468,-0.0852108076,-0.1596137434,0.2854138613,0.14059259,-0.3296193779,-0.2724534869,0.0815309584,0.0685857385,0.087481983,0.247994557,0.0392374694,0.1267900616,0.1781820506,-0.035612449,-0.0342132784,-0.3879116178,-0.2021701038,-0.0469164737,-0.06433779,-0.1276117265,0.027205281,0.4909307361,-0.0151986955,-0.2996301353,-0.086886175,-0.0073207086,0.0067967782,-0.0367102548,0.0159466714,0.0192606952,-0.1314559877,0.4342541695,-0.0709865913,0.0237194095,-0.0694636405,-0.1364616305,-0.0965490937,-0.0559106804,0.0847948417,0.0388861299,0.453989774,0.5106609464,-0.1366450638,-0.2999775112,-0.4262473583,0.0608281679,0.2992438376,0.1649462432,0.2493289858,-0.4019422233,-0.2741505802,-0.0979810432,-0.1292949021,-0.051951766,-0.1132292449,0.214834094,0.20876275,-0.0250972472,-0.3422092497,-0.1877381355,0.0577514134,0.2788983881,-0.0486825556,0.1046392992,0.1948835254,0.0439524241,-0.1548945457,0.5722846985,0.0037882533,-0.0727778226,0.2341006696,0.2575864196,0.0209753718,-0.1038708016,0.0966780186,0.3244815469,0.4285235703,-0.0999615937,0.1173951551,-0.2590292096,-0.0158545282,-0.1104129776,-0.1411903948,0.0248061232,0.1656399816,0.1345708668,-0.2868353724,-0.1047198176,-0.0061523984,0.1768272221,-0.1596791893,0.0104618138,-0.0507154316,0.1858514398,0.4787361324,0.1056185141,-0.1444310099,0.2968212664,0.0278852209,-0.3427369297,0.1812280715,0.119561404,-0.5345843434,0.2192011774,-0.0060151611,0.3959254026,-0.0321908854,0.0675944909,0.2912071049,-0.0239492841,-0.0405232422,0.0190820191,-0.1470579803,0.2811025083,0.0375399739,0.0213345606,0.071924679,0.0894377679,0.1300142407,0.0193517394,-0.0039990963,-0.0496643856,-0.250521034,-0.09520071,0.2610931695,-0.0846618935,-0.2297170907,-0.5908981562,-0.505007267,-0.2148566097,0.0485087521,-0.1013945043,-0.1506157219,0.2207663357,0.0781791359,0.3964470327,-0.3146063089,-0.2530259192,0.1790063083,-0.1775618941,0.2007614523,-0.1132066622,0.120511055,-0.177684322,0.4657227993,-0.1034091413,0.1453158259,-0.1856699288,-0.5438774824,0.0526022315,-0.5006883144,0.182919547,0.2707560658,-0.1279103756,0.2879445851,-0.2323251516,0.0673317313,0.1091383323,0.0826247111,-0.2611098588,-0.1224006414,0.0458420068,-0.3174271286,0.1527843326,-0.3833405972,-0.1743234396,0.3617884219,0.1312460899,0.012477221,0.2764031887,-0.2608338296,0.0804369599,-0.0426834673,0.0129134459,-0.0240160767,-0.5220912695,0.4109560847,-0.0681833774,0.0742995068,0.179184854,-0.1460767835,-0.039368961,-0.0817361847,-0.0980944782,-0.3686917722,0.0078179985,0.0263644643,0.39359501,0.3240304887,0.1298532188,0.0402110703,0.1510763168,-0.1397590339,-0.3809151649,0.1409541368,0.0039061294,0.2158198953,-0.1555594057,0.1129913032,-0.2410590649,0.6436130404,0.2247477472,-0.2596519589,0.1200383008,0.225541994,0.2154758424,0.2066877037,0.0681018978,0.1544284821,0.0993949696,0.132549718,0.1646129936,0.2598731816,0.3516198397,0.075247474,0.1566935778,0.0224758927,0.001733302,-0.0677464157,0.1254003048,0.2947210371,-0.2372250557,-0.2885355055,-0.4345577955,0.0502384566,-0.1725725383,0.2192239612,0.2361645699,0.2339865267,-0.6525697708,-0.0646853521,-0.418353945,0.1497697085,0.2101670802,-0.2585080564,0.0487791598,-0.1139919013,0.0413299836,0.0553585403,0.2756233513,-0.4068242013,-0.3215429187,-0.1338912398,0.1080361828,-0.0592176206,0.1334582865,0.0538672693,-0.335139513,0.1464511305,0.3255427182,-0.5488436222,0.0289644189,0.2874841392,0.0305426735,-0.1422052085,-0.1794624478,-0.2558195293,-0.4790070653,-0.2184931338,-0.0961850509,0.1289184391,-0.0987130329,0.206369102,0.0250991918,0.0195248574,0.0605213456,0.3410648406,0.2048428357,-0.0839761719,0.5489271283,0.1841719896,0.4057816565,-0.0367206037,0.498236239,0.543840766,-0.4102806747,-0.4458573163,-0.2406907529,-0.2364233732,0.5088031888,-0.0471299365,0.295406431,0.1613481641,0.1000375524,0.0301486086,-0.5421208143,0.3827722967,0.1044642478,0.101625666,-0.3163246512,-0.5153921843,0.6500331759,0.1372375041,-0.1403207034,-0.093703106,0.5290517807,-0.2122666985,-0.0464482084,0.3127535284,0.9245706201,-0.3420910537,0.3865251243,0.0498978533,0.1496538222,1.0509083271,-0.0660311729,-0.2529517412,-0.0452660322,-0.0859199762,-0.108365424,-0.0940411538,-0.1133008972,-0.0278171618,-0.0488783568,0.2143078446,-0.1284105778,0.0406639911,-0.057110846,-0.0287318863,0.2884884775,-0.3976852596,-0.0739053339,0.0057488354,-0.2724364996,-0.0052735382,-0.1983730942,-0.1336512566,-0.1448269486,0.2605913877,-0.1283146143,-0.1970808357,-0.0115776304,-0.4661141932,0.0442018434,-0.0668639317,0.1576806009,-0.0987012908,0.8020288944,0.0030566975,-0.0333352946,0.1023518741,-0.2983785868,0.0487954393,-0.2134904116,-0.1815410256,0.1033248752,0.1320537478,-0.2284690738,0.0096617462,0.286708951,-0.231173411,-0.3780369163,-0.2839672863,0.09970963,-0.2393221557,0.2914377451,0.491335094,0.1522365659,-0.1644828469,-0.0018144717,0.2500420511,-0.1749132127,-0.164575994,-0.3446571529,0.1069930866,-0.2558177114,-0.200831458,-0.1038671583,-0.0515512936,-0.091775693,0.1607387364,-0.0174699239,0.0351752713,-0.091680795,0.0929260999,-0.272079587,-0.1390751302,0.2453610897,-0.5394354463,-0.3860781789,0.2829026878,0.418913275,-0.3687256277,0.199231863,0.1497184485,-0.5050113797,0.147755295,-0.0755768791,0.2470769435,-0.044813659,0.3661976457,-0.1089798436,0.2876859307,-0.1661036015,-0.1402059197,-0.1243806854,0.0014686568,0.2083008289,0.0045568626,0.1921467781,-0.1206984892,-0.132948935,0.0091822073,-0.0551242717,0.0319452733,-0.3826414943,0.235648647,0.2814162374,-0.3803116083,0.1496724635,0.0697534084,0.0734398887,-0.1687825024,0.2959616184,-0.1004008129,0.0683975443,0.3203061819,-0.1592958272,0.2912622392,-0.0435626879,0.2320719361,0.3278766572,0.252913177,0.2915497422,-0.0429420955,-0.0000352938,-0.1267882138,-0.1028731316,0.2473246157,0.2575520277,-0.117952019,0.241146341,-0.0154884644,0.0684320629,0.3221442103,0.2460658848,0.1326191723,-0.1216163561,0.1707419753,0.2096450627,0.0880345181,0.0055891867,-0.177562654,0.0430709012,0.0518799573,0.2331981808,0.4577735364,0.1930025369,0.4258338511,-0.0130385021,-0.0845982432,0.26731655,0.1696623117,0.1071987301,0.091803506,-0.0400454178,0.1556315869,0.4965419173,0.4767507911,0.0760963857,0.0422004536,0.1736752987,0.5621284842,-0.010004608,-0.1365950257,0.1830108762,0.0466432832,0.1775755882,0.2730251849,0.2157990038,0.0626088828,-0.0165415723,0.0590946674,-0.3680163026,-0.2046263069,-0.0443443581,0.36524266,0.1625914127,-0.3252568245,-0.3301976621,0.1017808467,0.0079294657,-0.1792191267,0.0368532725,-0.7533146143,0.2983182371,-0.0801833645,0.2437644005,-0.4584320784,-0.4779801369,0.3081508577,0.2598594129,0.0663928762,0.2118909657,0.3727384508,-0.1756145358,0.170288384,0.1928735673,0.1293082535,-0.3071805835,-0.0201793518,0.4482208788,-0.3258323669,0.1118373573,0.0137011306,0.0178779438,-0.1996189803,0.1314798146,0.1867457628,0.0145564647,0.0168784242,-0.1589948088,-0.2898467183,0.0207005795,-0.2335769236,0.4583547115,-0.3620073199,-0.2303649336,-0.1014963984,0.3317963481,-0.3246144652,0.4316393733,0.4480347931,-0.0498042405,-0.0851531625,-0.1653256714,-0.0306280013,-0.0388955697,0.3118703663,0.3628661036,0.3550554812,-0.2159344256,0.006523395,-0.3346562982,0.077940464,-0.0795393735,-0.5314953923,-0.0749535263,-0.2790946364,0.386651665,-0.1143806279,-0.1173889562,-0.1609950513,-0.0456244759,-0.23464863,-0.136664018,-0.023665769,-0.3316567838,0.239864856,0.1465487331,0.1286137253,-0.1669041514,0.3373683095,-0.1495797634,-0.0518984869,0.2519812584,-0.1544785649,-0.0376677215,0.1252938211,-0.0341436565,0.4620826542,0.0743211657,-0.2768168449,0.0147172408,-0.0490716211,-0.3608732522,-0.0394783057,-0.3856681585,0.2619729638,-0.2971333861,0.014146246,-0.2721405923,0.0574108548,0.118015334,0.0544012897,-0.2993408442,0.4488971531,-0.2359231561,-0.2571635246,0.5835531354,0.3259752095,0.0826630592,0.056698624,-0.0444326587,-0.091830872,0.3495596647,-0.0792772025,-0.1233575121,0.0639469698,0.1544058323,0.3551385999,-0.1040099338,-0.834869206,0.2928942442,0.3128514886,0.0903129056,-0.5386824608,0.2407338917,-0.0605649725,-0.4271275699,-0.0267980304,0.3106340468,-0.2396766394,-0.0399483405,0.0327998102,-0.2008678764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/723","title":"Adding pseudo-labels to datasets","comments":"Oh yes why not. I'm more in favor of this actually since pseudo labels are things that users (not dataset authors in general) can compute by themselves and share with the community","body":"I recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n","comment_length":32,"text":"Adding pseudo-labels to datasets\nI recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n\nOh yes why not. I'm more in favor of this actually since pseudo labels are things that users (not dataset authors in general) can compute by themselves and share with the community","embeddings":[0.2814165056,-0.0320715606,0.079331696,-0.0085590035,0.0833111629,0.0303214397,0.3540932834,-0.1127623767,-0.1846269965,0.0258010179,-0.2245690823,0.0616265349,-0.1164304838,0.5070895553,0.4590643346,-0.0578463003,0.2990211248,-0.1764775962,0.2383633405,-0.1696520001,-0.3217107058,-0.3061875999,0.2590849102,-0.126566872,-0.3013052046,-0.1725135297,-0.1364627481,-0.0555804558,0.0722709969,-0.5509163141,0.0033327753,0.3938508332,0.2134947032,0.6353236437,-0.000124414,-0.1101553142,-0.135169819,-0.0913159624,-0.1328714192,0.1952459365,-0.0241945535,-0.2377730459,-0.208563298,0.1316625327,-0.1162411943,-0.3956457376,-0.1321433336,-0.0004315871,0.1173866987,-0.0123413848,0.0267378427,0.134383142,0.2572763562,0.1317085624,0.43737486,0.3451999724,-0.2409696281,0.3131239712,0.2690953314,0.4111149609,0.4979709685,0.4555452466,0.0043593589,-0.165012911,0.5459336638,0.0560602807,-0.0244434234,-0.1503382027,-0.324547708,0.1916165799,0.5204410553,-0.2745699584,-0.7477304339,-0.4786067009,0.133940801,-0.2930112183,-0.1681708395,-0.1566387713,0.3763653934,-0.0557193346,0.0327078737,-0.2535891831,-0.0172958877,0.153938219,0.0945208594,0.5230644941,-0.1751050651,0.2421139628,0.0016060428,-0.2349187881,-0.0595188066,-0.2889635265,0.1365657598,0.0885229856,0.137140736,-0.1648374051,-0.0790546387,-0.0250637978,0.2834421992,0.280524224,-0.0019304454,-0.1625097096,-0.2505151629,-0.1046128795,-0.0205005873,0.0994948,0.0174180306,0.3107479513,0.2868222892,-0.0311153252,0.3509349823,-0.1240788624,-0.4875078499,0.5173031092,-0.0417704731,-0.1651961058,0.0297067929,-0.1252249181,-0.0369141102,-0.004920797,0.1060510725,-0.1697644889,0.0614376478,-0.0727177113,0.0455628559,0.010078826,-0.1098928452,0.3468409181,0.2264734358,-0.0099695306,-0.0446697958,-0.1268308759,-0.0849921256,0.3702796996,-0.0208641235,-0.3191276193,0.2367655039,-0.2677616179,0.2227566689,-0.2555199564,0.039737545,0.1664511412,-0.07879439,0.207292363,-0.1710773706,0.1300288737,-0.0064152367,-0.0165117159,-0.2232581079,-0.0755658746,-0.0772995278,-0.0885614753,-0.3378321528,0.0163432322,-0.3694381416,-0.3178050816,-0.4057458043,0.267841965,0.1625999957,-0.0445063189,0.1314776838,0.1778279394,-0.0925155953,-0.1411683261,0.3071728945,0.175780043,-0.3151316047,-0.2468121499,0.0979709104,0.0708536282,0.106221281,0.273260951,0.0273605641,0.178281337,0.1910387725,-0.0352513865,-0.0270007849,-0.3939894736,-0.1756528318,-0.0302183814,-0.0872522369,-0.1030465066,0.0150569174,0.4775819778,-0.0071330732,-0.3312185109,-0.0948215723,-0.0311435889,0.0078072837,-0.0605014861,-0.0415951796,0.0101687349,-0.0982987508,0.4570453167,-0.1014263555,0.046385631,-0.0321486108,-0.1647463441,-0.0758874789,-0.0500925779,0.0503234379,-0.0062795901,0.4870735705,0.4675250947,-0.1388309002,-0.3249926269,-0.3900238574,0.0731933936,0.3471091092,0.1582392901,0.2516158819,-0.4307462573,-0.243163541,-0.1022994667,-0.1021443382,-0.0536688045,-0.1130784899,0.1869273782,0.2021273673,0.0301728025,-0.3294788599,-0.1759525388,0.019927023,0.2592219114,0.0052640559,0.0779906809,0.1858629435,0.0633951202,-0.1451585144,0.5716484785,0.003700197,-0.0492232479,0.2115737349,0.2514578402,-0.0341302641,-0.1268545389,0.0779056326,0.2899094224,0.4492066801,-0.1023445204,0.1202338785,-0.3016070724,-0.00628724,-0.1210778803,-0.1607472897,0.0325910933,0.1353400052,0.1079138592,-0.3179320395,-0.1142704189,-0.0362302959,0.2148618698,-0.202592209,0.0159474351,-0.0470697545,0.1973655075,0.4283904433,0.0691283941,-0.127035901,0.3050814569,0.0148116983,-0.3362833261,0.1448364258,0.1154859886,-0.4773359597,0.2646057308,0.0246778559,0.3967917562,-0.0382044762,0.0665167645,0.285852313,-0.0505741686,-0.0769792721,0.0505822301,-0.1785834283,0.2639607191,0.044479765,0.0173525773,0.0846812129,0.111520499,0.168547973,0.0215705577,0.0050871167,-0.0695074573,-0.2577661872,-0.0902420655,0.3064964712,-0.1113592014,-0.2288336903,-0.5478327274,-0.4770701826,-0.243519038,0.0457399376,-0.1163429692,-0.1515194029,0.2037445456,0.0828662962,0.3769288659,-0.3681286871,-0.2341170311,0.2060318589,-0.1579406559,0.1751887202,-0.1088225469,0.1013144106,-0.1739375591,0.4595066905,-0.0909392312,0.1576002836,-0.2395855486,-0.5610396862,0.0532135554,-0.4868705571,0.1262972653,0.2272780985,-0.2064217627,0.3010920286,-0.2319125235,0.1144014597,0.0976795629,0.10711357,-0.3073494136,-0.0991429389,0.0282772295,-0.3704604208,0.1602454334,-0.383256197,-0.1707057506,0.3619334698,0.1385250837,0.0145235276,0.2654289901,-0.2644994259,0.0720604211,-0.0012258035,0.0056854854,-0.0457010306,-0.5406401157,0.4097239077,-0.0630020723,0.0450029373,0.1832072735,-0.0863226354,-0.0787305981,-0.0366437808,-0.0980235487,-0.3702332377,0.0372954719,0.0601624995,0.3296085596,0.3650105596,0.0784116536,0.0801626667,0.1541107595,-0.1120469868,-0.3777053654,0.1402949989,-0.0060771666,0.2007026225,-0.1400783509,0.1237652749,-0.2093783766,0.6475243568,0.1835399717,-0.2741861641,0.0909754112,0.2186028212,0.2789510489,0.218400985,0.0629039034,0.1245019659,0.0851549059,0.11877729,0.0922010243,0.2761643827,0.3328823447,0.0363051295,0.1667709202,-0.0027102784,0.0036796988,-0.0742745847,0.1283902526,0.3042136729,-0.2279257327,-0.239344433,-0.4707076252,0.0233208854,-0.1196448579,0.2016664743,0.2855396569,0.2183596343,-0.6506949663,-0.0337944068,-0.454113394,0.1646956801,0.2548808753,-0.2515058517,0.0080291322,-0.0985730886,0.0416149981,0.0023288273,0.2471233904,-0.4359412789,-0.3356826901,-0.1177065,0.128551051,-0.0474906191,0.087363869,0.032630112,-0.3131889999,0.1383748949,0.3104916513,-0.5903473496,0.0822378024,0.3321760297,0.0320971571,-0.1716781259,-0.2003044933,-0.2402271628,-0.4535846412,-0.2054431587,-0.1358858645,0.1491918117,-0.0974565372,0.1312792003,0.022057103,-0.0093616806,0.1175542027,0.3391140997,0.1882612258,-0.1085172519,0.5585429668,0.2293509692,0.3981238902,-0.0017458304,0.4851604998,0.5506290197,-0.3786619008,-0.4305565059,-0.2744676173,-0.2977836728,0.498726368,-0.0239651259,0.301517725,0.1344216019,0.1156232283,0.0313125513,-0.57185781,0.3317236304,0.1132676899,0.1221410483,-0.3394647837,-0.5148450136,0.64348948,0.170930326,-0.1328039318,-0.0769741088,0.5751504302,-0.1831365526,-0.0781731457,0.3106164038,0.9312649369,-0.3546266854,0.340282619,0.0096715912,0.1206126958,1.1054617167,-0.0826207697,-0.269320637,-0.0225140285,-0.1110867634,-0.0988920704,-0.1021145955,-0.0899720415,-0.0190814938,-0.0169626791,0.2067678869,-0.1112793535,0.0422427766,-0.0314115472,-0.0288625266,0.2507326305,-0.3882371783,-0.0254844818,0.0088858837,-0.2618277669,0.0024439078,-0.20294635,-0.1366048455,-0.1641849726,0.252175808,-0.1696010977,-0.1654903144,0.0459349342,-0.4238361418,0.0301601179,-0.066127494,0.1879517436,-0.1320540756,0.8199995756,0.0269342139,-0.0102601107,0.0369616933,-0.3038391769,0.1092842296,-0.2030627877,-0.1565515697,0.0781785399,0.1217697188,-0.243327871,0.0456175171,0.2912818193,-0.2466258854,-0.3772779107,-0.3010877967,0.1230072901,-0.2192463726,0.3524191082,0.5262486339,0.164069593,-0.1386026144,-0.0056376434,0.2523898184,-0.1644957513,-0.1531691849,-0.3391691148,0.1415354609,-0.2554901838,-0.1679081917,-0.1020875424,-0.0440895148,-0.1231990084,0.1124008968,-0.038566865,0.0602386147,-0.1376800686,0.0562337525,-0.3003266752,-0.2097403258,0.2434345186,-0.5009253025,-0.4061090946,0.3031411767,0.4211236537,-0.377358824,0.1892689764,0.1726640165,-0.4455311,0.1084853262,-0.0434048213,0.2288219929,-0.0389686115,0.3678354621,-0.1221154928,0.2558901906,-0.1640507132,-0.1598451287,-0.0952679068,-0.0218968857,0.2124329656,0.0834646821,0.1798060685,-0.1412159801,-0.1252384931,0.0408810824,-0.027508121,0.048272904,-0.3353329599,0.2303613275,0.2928524911,-0.3914384544,0.1151466519,0.062715061,0.0759790018,-0.1434515566,0.3050139844,-0.0676116943,0.0672236308,0.2963534892,-0.1880676448,0.3651655316,-0.0393471234,0.2685129344,0.3243949413,0.3087015748,0.3029631674,-0.0125713609,0.0382776633,-0.0775370002,-0.0944027603,0.2028935701,0.2126232982,-0.0965493992,0.243411392,-0.0327301398,0.0925734341,0.3101596832,0.2636582553,0.1240705997,-0.1384516805,0.1463162154,0.234185338,0.092959024,0.0249175988,-0.1522690952,0.0213429462,0.0459068641,0.2269611359,0.4519796073,0.2267324924,0.423509568,-0.0457368754,-0.0722153187,0.270006299,0.1288207471,0.0932013988,0.0590199493,-0.0335601531,0.1987065524,0.467820406,0.4600561261,0.0148075335,0.0027085554,0.1728649586,0.5681412816,0.0498109348,-0.1248593628,0.1993463039,0.0748150423,0.2682193518,0.2298011184,0.2823825479,0.0614926368,0.0262116436,-0.0040476229,-0.400695771,-0.1534422636,-0.0729166418,0.3058268428,0.1598051339,-0.2744481266,-0.31610021,0.1162031218,0.0345210135,-0.170460999,0.0241412446,-0.7365186214,0.2727882862,-0.0846388787,0.2248656154,-0.4340877533,-0.4931125045,0.2739287019,0.3674144149,0.0638419539,0.1478434205,0.3270253539,-0.1447973102,0.1678352058,0.2406982034,0.1115865409,-0.26908198,0.0377585888,0.4323944449,-0.3454434276,0.1450866163,0.0528867431,0.0522493757,-0.2151185274,0.087914601,0.2273890227,0.0081383726,0.0221003108,-0.1368333101,-0.3013674021,0.051869262,-0.2918008566,0.473669827,-0.3938247263,-0.2142862827,-0.0800402313,0.3053113222,-0.2383293808,0.4440695047,0.4377977252,-0.0314485691,-0.1012361497,-0.1655845642,-0.0306710247,-0.0058881622,0.2613072097,0.3529650867,0.3276174366,-0.1994434744,0.0735999718,-0.3824089766,0.06465929,-0.0807911381,-0.5424757004,-0.1202826053,-0.273755163,0.3947317898,-0.0442484394,-0.1000176594,-0.1935767084,0.0120290248,-0.2517722845,-0.1323552877,-0.0638131499,-0.3120062649,0.1835526377,0.1365460306,0.1423668116,-0.1529538929,0.299376905,-0.1273600906,-0.0680193529,0.2645392716,-0.1696015,-0.0250620022,0.1452663988,-0.0240018163,0.4277547002,0.1341848522,-0.2914246917,0.0280221812,-0.0703518167,-0.3189187646,0.0110618444,-0.390827328,0.2293149382,-0.2439148724,-0.0141698513,-0.3038215041,0.0421396643,0.1075420603,0.064372547,-0.2987958491,0.4347042441,-0.1804732978,-0.2937621176,0.5569198132,0.2664344013,0.1236700714,0.053293854,-0.0540952981,-0.0849280357,0.3328028023,-0.1001133397,-0.0730896518,0.0278330538,0.1979803741,0.382681638,-0.1051414236,-0.838541925,0.2493816614,0.3333539069,0.0442685001,-0.4850102961,0.2213869095,-0.0739427954,-0.4330570698,-0.0398637578,0.2943261564,-0.3045844138,-0.0232384056,0.0507726185,-0.1518713534]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/723","title":"Adding pseudo-labels to datasets","comments":"![image](https:\/\/user-images.githubusercontent.com\/6045025\/96045248-b528a380-0e3f-11eb-9124-bd55afa031bb.png)\r\n\r\nI assume I should (for example) rename the xsum dir, change the URL, and put the modified dir somewhere in S3?","body":"I recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n","comment_length":22,"text":"Adding pseudo-labels to datasets\nI recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/96045248-b528a380-0e3f-11eb-9124-bd55afa031bb.png)\r\n\r\nI assume I should (for example) rename the xsum dir, change the URL, and put the modified dir somewhere in S3?","embeddings":[0.3824186623,-0.0590515547,0.0575784147,0.0074140299,0.1342783868,0.0131380288,0.2963197231,-0.1217540801,-0.2198833525,-0.0435279682,-0.2787754536,0.0804837942,-0.0839418247,0.5122793913,0.4707950652,-0.0267684255,0.3672170043,-0.1611523479,0.1646211743,-0.1702684611,-0.3577936292,-0.284829855,0.2717745602,-0.2073100656,-0.3152383268,-0.1203107312,-0.2245425135,0.0721220225,0.0589686148,-0.5473186374,-0.0266391244,0.4308969676,0.2439986616,0.6618462801,-0.0001245916,-0.1142086759,-0.156179741,-0.1502090394,-0.1823928207,0.2175573707,-0.0315193571,-0.1962168813,-0.206166178,0.1632058173,-0.1114674062,-0.324881196,-0.0985213965,0.0733948052,0.1952245086,0.0326190107,0.0204025283,0.1139785945,0.2299670279,0.1345407367,0.4793930948,0.3306820393,-0.1912674457,0.3493455648,0.2104496062,0.3990181386,0.494325161,0.4715925157,-0.0191405583,-0.1135742515,0.6187542677,0.042759899,-0.0338233523,-0.1695554852,-0.3249826729,0.121823445,0.5115076303,-0.2819126844,-0.731289506,-0.4637393355,0.1253407747,-0.3203245103,-0.1134286523,-0.2277140617,0.4337185323,-0.0508428477,-0.0439432859,-0.3035870194,-0.0564605072,0.2161291689,0.0429830402,0.5357024074,-0.1913152039,0.1910958886,-0.0058261519,-0.2352779508,-0.0667269155,-0.2758331895,0.1633462906,0.124404572,0.1628944129,-0.1206509471,-0.0619253442,-0.0050701685,0.2346730381,0.3252992332,0.0175860897,-0.1301683635,-0.2440437675,-0.1120035872,-0.0779500678,0.1155245751,-0.0518663377,0.289770931,0.2308941483,0.0532464907,0.3382211626,-0.12401364,-0.4614148438,0.4557311237,-0.1177974343,-0.1218679026,0.07338617,-0.0636416078,0.0258405544,-0.0516759418,0.1071726456,-0.1597567052,0.0295777787,-0.0548921265,-0.0067709219,0.0393005647,-0.0814270675,0.371589303,0.177459389,0.0856072679,-0.0314126387,-0.132732302,-0.0919188261,0.3518421054,-0.0154538229,-0.2775564194,0.2595612705,-0.2489346564,0.1895795465,-0.2381272912,0.0119780172,0.2015544623,-0.0093984995,0.2288381755,-0.1336815953,0.130899027,-0.0534304976,-0.0058628661,-0.2272228897,-0.1346719563,-0.077406317,-0.1108798161,-0.3310883939,0.0051318696,-0.4340772331,-0.3205754459,-0.3698263466,0.2506905496,0.1282762885,-0.0645630881,0.109578602,0.187093854,-0.0930071101,-0.1487048268,0.3043155968,0.1734105051,-0.3290264308,-0.2644236684,0.1126336604,0.0758917406,0.0661482587,0.2349952459,0.0647127107,0.1130079031,0.165698573,-0.0372064486,-0.054179702,-0.4142211378,-0.2015166879,-0.0602946915,-0.1206694692,-0.0934125409,0.0243772902,0.4403903186,-0.0290453583,-0.324778378,-0.1446695179,0.0162284821,0.0528157912,-0.053754624,-0.032078281,0.0505713373,-0.1820052713,0.4311705232,-0.1165945381,0.0583196804,-0.0444591381,-0.1440145969,-0.1113424227,-0.0915574953,0.1040965617,0.0736880377,0.500715971,0.5515341759,-0.1131003052,-0.2475200295,-0.4533876181,0.0532470718,0.315992862,0.1733866483,0.2216348052,-0.3751897216,-0.2680237293,-0.1474182904,-0.1475027502,-0.0535799675,-0.0952903405,0.24758479,0.2361048758,-0.019332435,-0.2998919487,-0.2160845101,0.1283097118,0.2848452628,-0.0539394021,0.1205325648,0.1492785066,0.020002516,-0.1491583139,0.5127786398,0.0104662608,-0.1040401086,0.2029866576,0.3039376736,-0.0074415477,-0.0276306737,0.1016785055,0.3263528347,0.4377632439,-0.0990099385,0.1382637471,-0.2529663742,-0.0132120606,-0.0791060477,-0.1793911606,0.0229728557,0.109579958,0.13348037,-0.2986077964,-0.152113378,0.0372588933,0.1393225491,-0.1947154403,0.0079036588,-0.1108312383,0.2330360413,0.3945525885,0.1087793857,-0.095689103,0.3215187192,-0.004888718,-0.3789289594,0.1060980111,0.1068026647,-0.4911374152,0.1963323653,0.0301518682,0.4219139516,-0.0347625203,0.0619035214,0.340346843,0.0125119602,-0.04759432,0.02531977,-0.127601251,0.3111085892,0.0907710344,0.082946822,0.1214504987,0.1007056013,0.1734576672,-0.0229759179,0.0254100412,-0.0625023395,-0.2944380045,-0.063615337,0.2560421824,-0.1195860878,-0.1983718425,-0.5820183754,-0.4939034879,-0.2003488243,0.0969549939,-0.0954896659,-0.2112884223,0.2024484128,0.1277593374,0.3231777549,-0.3230867088,-0.2451108992,0.1739901751,-0.1317127943,0.1823119223,-0.1230627373,0.1306600571,-0.1697481722,0.3952826858,-0.1792241335,0.1420127302,-0.2615374029,-0.5440930128,0.0700345561,-0.4652977884,0.1871996373,0.3011380732,-0.1650559604,0.2846772373,-0.2040569633,0.1087174714,0.1524674445,0.1083435714,-0.2670251131,-0.1588961631,0.0674753636,-0.366787225,0.1237903461,-0.2899271548,-0.1475366652,0.3569642007,0.1526903063,0.027100997,0.2999839187,-0.1780371964,0.0791995302,-0.0334858112,0.0238806754,-0.0537361987,-0.5200936198,0.4123340547,-0.0438693166,0.1034152433,0.2109449059,-0.1105611175,-0.0498708598,-0.0476945788,-0.0883577839,-0.4446310997,0.0703269467,0.0244686529,0.3464213908,0.391292125,0.1474480629,0.0580122955,0.1238531247,-0.1444465071,-0.3642102182,0.0896709338,0.0311482977,0.2948604822,-0.1509866714,0.165262714,-0.2183775902,0.6097123623,0.1429276019,-0.2653751671,0.1522255242,0.2022414953,0.2568854094,0.261213392,0.1293950826,0.1242536306,0.0688544661,0.0794270113,0.1312377155,0.2361262441,0.3546675444,0.0890387967,0.1559652239,-0.0222204681,-0.0110286521,-0.056861598,0.1160861924,0.2757038772,-0.2265879661,-0.3143399954,-0.443998307,0.0765801892,-0.1103463471,0.208367005,0.2990710139,0.2141655684,-0.6747987866,-0.0784450844,-0.3940447271,0.1974583119,0.236542657,-0.2909457386,0.0761525705,-0.1362183392,0.014246895,-0.0457082801,0.32473436,-0.4556795061,-0.3164601624,-0.0828949288,0.0773023218,-0.0874930769,0.0609924681,-0.0438268967,-0.3415932655,0.1914530098,0.3583162427,-0.5709721446,0.0013944801,0.2717257738,0.0476425104,-0.1293657124,-0.1451887786,-0.2763870955,-0.5452921987,-0.2844866216,-0.1423661709,0.1368618757,-0.1235774532,0.1678208411,-0.0092439232,0.0457215011,0.0674624443,0.308762759,0.1833769232,-0.0732656494,0.6347836852,0.1727158278,0.4002009034,-0.0571849421,0.512804389,0.556219399,-0.3854486942,-0.4082216322,-0.258903563,-0.1833526492,0.4894203842,0.015521598,0.342421174,0.1198452339,0.0873250067,0.1135363728,-0.5821657181,0.3661463857,0.1292794794,0.0931179896,-0.2194298953,-0.4939138889,0.6453342438,0.1312827468,-0.1917436421,-0.1537511349,0.5240007043,-0.2538593709,-0.0462197624,0.3815089166,0.9618303776,-0.2958473563,0.3563614488,0.094573237,0.1051079929,1.0278152227,-0.0388761088,-0.2415130138,-0.0164210238,-0.0441416204,-0.0776996762,-0.1024250239,-0.0843888894,-0.0279040299,-0.0509029962,0.2172041535,-0.1412744969,0.0122615863,-0.1010473967,-0.033538539,0.1910453737,-0.3727122247,-0.0745813325,0.0126226386,-0.2252281457,0.003348131,-0.1828036606,-0.1266426444,-0.1935335845,0.2440081835,-0.0582861528,-0.1784499884,-0.0051928661,-0.5162863135,-0.0061499719,-0.0745098442,0.1101946011,-0.1065224931,0.7671622038,0.0037602987,0.0082289735,0.0739077702,-0.2669685781,0.0935750902,-0.2059941888,-0.2251676023,0.0853210092,0.1157646775,-0.2600768507,-0.0054510841,0.2936967313,-0.2009813935,-0.3357983232,-0.3550329804,0.078013055,-0.2176427543,0.2874295712,0.4839892387,0.1382375211,-0.146063894,-0.0037413014,0.2006308287,-0.1607755423,-0.1167374551,-0.3657362163,0.1184464917,-0.2654045522,-0.1230231076,-0.1780189872,-0.0528781973,-0.0653270781,0.1566042304,-0.0373747759,0.0552239753,-0.1481808275,0.0123097338,-0.2691958249,-0.1362373382,0.236418739,-0.5443879366,-0.386346221,0.3006246984,0.4367991388,-0.325363636,0.1571236104,0.1752294898,-0.4280358553,0.1592925191,0.0112792142,0.2756153345,-0.0802958757,0.335321933,-0.0866628289,0.2324058414,-0.1703130603,-0.107602343,-0.1109550744,0.0298788566,0.153689906,0.0183866881,0.1671490222,-0.1511325389,-0.1201809794,0.0055343909,-0.0324260257,0.0397699401,-0.3695160151,0.2196667045,0.2820345461,-0.425211519,0.1378548443,0.0831332281,0.1048725247,-0.1739906371,0.3192070127,-0.0955537781,0.0653206632,0.3136176467,-0.182766065,0.2739145458,0.0023458009,0.1688380837,0.2984244227,0.2922674716,0.2921116948,0.0073129921,-0.0223777927,-0.1077290624,-0.1074771285,0.2464704216,0.2351890802,-0.0750162601,0.2435467541,-0.0104906112,0.0915156901,0.3255590498,0.278716594,0.1505318731,-0.1560919434,0.1145433709,0.2418032289,0.0898408741,0.0196207389,-0.1615661979,-0.0305625014,-0.0147936223,0.2058152258,0.4395158291,0.1985477656,0.4025268555,-0.0616395399,-0.0391606763,0.305508554,0.2193682194,0.0831356496,0.0658195317,-0.128072083,0.1707731038,0.5006363988,0.4155251682,0.0268865731,0.0178106725,0.1701321602,0.5361150503,-0.0182364564,-0.0289441142,0.1848016083,0.060424909,0.1947866082,0.2642328441,0.2715235651,0.1048758775,0.0036642575,0.0176010914,-0.3718640804,-0.2035439461,-0.039251335,0.4397819042,0.2133210897,-0.3237321973,-0.3560729027,0.0858793631,0.026256036,-0.1113549322,0.013047481,-0.7505530715,0.3144698441,-0.055699721,0.2290304452,-0.4288556278,-0.4792248011,0.2262845784,0.2891719043,0.0530890003,0.1781933904,0.4162328243,-0.1435064971,0.1604761332,0.1805380732,0.1668718457,-0.2905869484,0.0015827608,0.4147277176,-0.2863204479,0.1159629673,0.0697534904,0.0077965385,-0.2317918092,0.0784617066,0.1609253734,0.0187409557,-0.0161950905,-0.1664539874,-0.3410788178,0.0363428853,-0.1856468916,0.4932834804,-0.3427179456,-0.2011359781,-0.0916402414,0.3285875916,-0.3409661651,0.4163323343,0.4194240272,-0.0913837254,-0.0692442879,-0.1162862703,-0.0272595063,-0.0003643057,0.3093910217,0.352286011,0.3389692605,-0.2107869536,0.0307674464,-0.3650116622,0.0416252539,-0.0584665798,-0.5650427341,-0.1128221825,-0.2199030817,0.3593733609,-0.1182817817,-0.0938144252,-0.1921337098,-0.002083817,-0.1872657984,-0.1280307919,0.0157940518,-0.3481068909,0.2568414211,0.1659721434,0.0915383101,-0.1572402716,0.3233322501,-0.137570411,-0.0374385752,0.288926363,-0.1651370525,-0.0691806898,0.1549958289,-0.0092366571,0.4199464321,0.0389690176,-0.3318716586,-0.036680419,-0.0794460848,-0.3724073768,-0.0080791749,-0.4380820394,0.28690961,-0.2511079907,0.0664695054,-0.301242888,0.0728721693,0.1339550316,0.0613200404,-0.3008301258,0.4762836695,-0.2457075715,-0.2488268614,0.5414467454,0.2740878761,0.1216177866,0.0951972678,-0.1107436195,-0.1305067092,0.4118334949,-0.0960555896,-0.1009868383,0.0511977971,0.1962416917,0.3703525364,-0.1034028083,-0.8380081058,0.2865531743,0.3024728298,0.1257441342,-0.5482910275,0.2628130317,-0.0457666367,-0.4233816564,-0.0525522418,0.2995759845,-0.2467922419,-0.0660707131,-0.0452023745,-0.1770839244]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/723","title":"Adding pseudo-labels to datasets","comments":"You can use the `datasets-cli` to upload the folder with your version of xsum with the pseudo labels.\r\n\r\n```\r\ndatasets-cli upload_dataset path\/to\/xsum\r\n```","body":"I recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n","comment_length":23,"text":"Adding pseudo-labels to datasets\nI recently [uploaded pseudo-labels](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/seq2seq\/precomputed_pseudo_labels.md) for CNN\/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo.\r\nSince pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution.\r\nI read https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html, but it doesn't really cover this type of contribution.\r\n\r\nI could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https:\/\/github.com\/huggingface\/datasets\/blob\/5f4c6e830f603830117877b8990a0e65a2386aa6\/datasets\/xsum\/xsum.py\r\n\r\nWhat do you think @lhoestq ?\r\n\r\n\r\n\nYou can use the `datasets-cli` to upload the folder with your version of xsum with the pseudo labels.\r\n\r\n```\r\ndatasets-cli upload_dataset path\/to\/xsum\r\n```","embeddings":[0.3005567491,-0.0865521207,0.0789877996,0.0127475159,0.1379013658,0.042592112,0.2965248227,-0.0892710388,-0.1311770082,0.0045273653,-0.2615566254,0.1170213744,-0.101874873,0.5030103326,0.4647361934,-0.0501749516,0.3400839269,-0.1612035632,0.1209385246,-0.1879306287,-0.3377030492,-0.2674909234,0.2434321344,-0.1668701917,-0.3428444862,-0.1463990957,-0.1756635904,0.044133652,0.0913106054,-0.5311963558,0.0352854356,0.4109178782,0.2603910565,0.644187808,-0.0001249052,-0.0960011184,-0.1164314598,-0.1558603644,-0.1917866468,0.1872811019,-0.0406610146,-0.1984737664,-0.1967243254,0.1530397087,-0.1072224453,-0.374322921,-0.0886982307,-0.0048337351,0.1889227629,-0.0055733151,0.016215466,0.2126973718,0.2619461417,0.11458157,0.3966299593,0.312405467,-0.2199382186,0.329834342,0.2574898601,0.4000386298,0.4975061417,0.4654204845,0.0034594834,-0.1363222152,0.5766840577,0.0685753152,0.0213656109,-0.1638678908,-0.3478551209,0.1882938147,0.5000043511,-0.2947636247,-0.7588800788,-0.4253388643,0.1248204783,-0.2834514081,-0.1056061387,-0.2003812939,0.3972685337,-0.0533891357,-0.0065780366,-0.254371047,-0.084383741,0.1836412847,0.0637215823,0.5415856838,-0.1714508235,0.2019059211,0.0627435893,-0.2626826465,-0.0453341231,-0.2511155009,0.156747669,0.1438652277,0.1374804229,-0.1627710462,-0.1111432239,0.0124162966,0.2875876725,0.2554081678,-0.0147288051,-0.1271378547,-0.2798385322,-0.090706788,-0.0495130159,0.1478044987,0.0205173716,0.2564308047,0.2562672198,0.0137418807,0.3419875801,-0.1323289275,-0.5306587219,0.5164095163,-0.0874370411,-0.1615372896,0.0084345825,-0.1053058356,-0.0154332751,-0.0224043578,0.1680036783,-0.1824706197,0.0620614626,-0.0694893748,0.0375250652,0.071760796,-0.0921622962,0.3676926792,0.1835538894,-0.0092500374,-0.052627679,-0.0945297033,-0.087111637,0.3537300229,-0.0076326779,-0.3353779018,0.2623464465,-0.2569543421,0.2294719517,-0.1936970204,0.01232043,0.1991736293,-0.0730665103,0.2315552235,-0.1415062994,0.150870949,-0.0122410776,-0.0095902886,-0.2297956347,-0.1294585466,-0.0831670314,-0.0762726963,-0.3556428552,0.0113501186,-0.4253790379,-0.3329553902,-0.4054366052,0.2301377952,0.1485272944,-0.0096587623,0.0667015091,0.1540709883,-0.1053178608,-0.1127883121,0.314979285,0.1690998822,-0.4170022011,-0.2718285024,0.1106524616,0.0382295586,0.0504448488,0.2552046478,0.0666076764,0.1342077553,0.1700593382,-0.05008642,-0.0997086689,-0.4665155113,-0.2023460865,-0.0561346672,-0.0590186566,-0.0885888413,0.0227516871,0.4571012557,-0.015280202,-0.3111627996,-0.13169384,-0.0294852275,0.0568469949,-0.0687908083,-0.0553259291,-0.0011320733,-0.1273282021,0.4534283876,-0.1470853388,0.0185575504,-0.0733295158,-0.1068966165,-0.0800819919,-0.0740582123,0.0727354959,-0.0007414462,0.4443889558,0.4343406558,-0.1196610183,-0.3397579491,-0.4548968077,0.1086860523,0.2950422168,0.1677487046,0.240214631,-0.3612698019,-0.2660116851,-0.1327229589,-0.1280923039,-0.0307635013,-0.1144382805,0.1807572991,0.2044485956,-0.0083762901,-0.3424597979,-0.1521357596,0.0372333564,0.2448761463,-0.0409918167,0.1239593923,0.1826606095,0.0440870188,-0.1847420186,0.5197100043,0.0191852208,-0.1109805927,0.1982563585,0.3092309535,-0.0159006752,-0.0633957684,0.1213110313,0.2998844385,0.4853625298,-0.0897499472,0.1092524379,-0.2770066559,-0.0257104132,-0.0954571813,-0.1917815059,0.0236201603,0.186128661,0.1155374423,-0.2611899674,-0.1196405664,0.0220842585,0.2011290044,-0.1726160049,-0.0360716879,-0.0711005554,0.2465389818,0.4118812382,0.0672813281,-0.1039957628,0.3385364115,0.0379811786,-0.3380328119,0.1061590761,0.1020345539,-0.5140880346,0.2836023867,-0.0069670076,0.4016209841,-0.0132528748,0.0652671307,0.2741593719,-0.0282700229,-0.0877167657,0.0564490482,-0.1335628182,0.2707892358,0.0627320856,0.043914482,0.0381421298,0.119860895,0.1434423029,0.0160628892,0.0239307471,-0.045829013,-0.2845480144,-0.1033421978,0.2123590261,-0.1288749278,-0.2173863053,-0.5689035654,-0.4772747159,-0.2290362269,0.0598107427,-0.0901624858,-0.1786348373,0.2162619531,0.0645548552,0.3908275068,-0.2997992635,-0.2319872826,0.1835287809,-0.1345534623,0.1550382972,-0.119854793,0.1175880209,-0.1475571245,0.3982762694,-0.1288672388,0.1292864233,-0.238943696,-0.5371047258,0.0983880237,-0.4517771602,0.1908956766,0.2576667666,-0.1502099931,0.3111432791,-0.2268038839,0.1120988876,0.1109325811,0.0671593472,-0.2997234464,-0.1118497998,0.0367011577,-0.3677363098,0.0979162678,-0.3780786991,-0.1553580761,0.3291868567,0.1468504518,0.0413799137,0.3024629653,-0.2576722205,0.0737459138,-0.0114810914,-0.0000506818,-0.0530079231,-0.5784045458,0.4582704306,-0.0525679588,0.0255833659,0.1926742941,-0.1010329127,-0.0518296175,-0.0418860801,-0.0808490962,-0.4202512801,0.0469216332,0.0205524266,0.3903393149,0.3359698057,0.1172429323,0.0368375517,0.1378154904,-0.1517869532,-0.3558728099,0.1650114954,0.0210258663,0.2545069158,-0.1433811337,0.1291761696,-0.2391848415,0.5956507921,0.2005895078,-0.2929969132,0.1209338978,0.2484630793,0.2504304647,0.2333758771,0.0773001164,0.1127719432,0.0980243832,0.1541238129,0.1273020208,0.2796022892,0.3740213513,0.0768565461,0.1701912433,0.0721968487,0.0184610393,-0.0998083875,0.1475939155,0.2638201118,-0.2336713523,-0.2543124557,-0.4365407228,0.066539444,-0.1036041528,0.2030191869,0.2736344337,0.2449995428,-0.6622622609,-0.0344684459,-0.4082279205,0.1590770185,0.2376983464,-0.2505781054,0.0388490111,-0.1129466966,0.0092552658,-0.005698686,0.2935488522,-0.4202443361,-0.3173517883,-0.0715713203,0.0930697843,-0.1117160246,0.0709920004,0.0091690123,-0.3605170548,0.2394087017,0.3792348504,-0.5476180315,0.0103417654,0.276030302,0.0657602102,-0.1663586348,-0.2120378911,-0.2518391609,-0.4434755445,-0.2146153748,-0.1189867929,0.1250091344,-0.086353831,0.1271366179,0.0071246843,0.021521654,0.0315180235,0.3175496757,0.1892130226,-0.0314187631,0.5588920116,0.2225304842,0.4542620778,0.0002596597,0.4915840626,0.5206646919,-0.3872155547,-0.4062038958,-0.2454668581,-0.2910209,0.4526287019,-0.0365534462,0.3053741157,0.0981245041,0.134692505,0.0305018332,-0.5840031505,0.3780729771,0.1097161099,0.1045726612,-0.2726257145,-0.4577994347,0.657779336,0.1590689868,-0.1287040114,-0.0958611593,0.5477828979,-0.2254185826,-0.1127008274,0.3520124257,0.932847321,-0.336961776,0.3329705298,0.0396052562,0.1636604667,1.0672823191,-0.0586959012,-0.294801861,-0.0047523081,-0.1631879359,-0.1015256271,-0.0841587186,-0.108916305,-0.0217036102,-0.0522612073,0.2386965007,-0.1711103916,-0.0124164363,-0.0560391732,-0.0046274792,0.2183507532,-0.3940485418,-0.0156153357,0.0040183705,-0.2048399746,0.0047004353,-0.1885085851,-0.1357959509,-0.1353674233,0.2488988489,-0.1408220679,-0.1361058801,0.0379498675,-0.4584195018,0.0112074353,-0.0335458852,0.1570054144,-0.1123263985,0.8262199759,0.0003976144,-0.0335019529,0.0715708211,-0.3101950288,0.0936180204,-0.176508829,-0.188576743,0.1225454733,0.104551509,-0.2536045313,-0.0054941257,0.2725922465,-0.2227247506,-0.3585734069,-0.3212807775,0.1308987141,-0.1887420267,0.2948666215,0.4917615354,0.1491300166,-0.1357511431,-0.0047588344,0.2458877265,-0.1723059118,-0.1269857138,-0.3595585823,0.1188425645,-0.284638226,-0.1014964804,-0.1405745298,-0.0804911554,-0.0579291284,0.1359797418,-0.043510139,0.0516582131,-0.1861876249,0.0736512244,-0.2694046497,-0.1747682989,0.2689924836,-0.519067347,-0.4004072845,0.3233268559,0.4202766418,-0.3643452525,0.1572706252,0.1166163087,-0.5053479075,0.1381151825,-0.0381040312,0.2523898184,-0.057800319,0.3598192036,-0.0919567347,0.2455291152,-0.1634774953,-0.1431566477,-0.0721925274,-0.0301044919,0.2206993401,0.0580460653,0.2157352716,-0.1299769431,-0.1238181144,0.0111894542,-0.0468099229,0.0507178195,-0.4097671211,0.2229915112,0.3041748405,-0.4157530367,0.1692835689,0.0494493432,0.0994267315,-0.15156959,0.2793646753,-0.0760485083,0.0589885227,0.3352190554,-0.1418848485,0.3381769657,-0.0312912762,0.2499747276,0.3630746901,0.2893466651,0.2955308855,-0.0161010083,-0.033383742,-0.1171479896,-0.1440601647,0.2044927627,0.2352798581,-0.0927622765,0.2259894758,-0.0146825444,0.0624651909,0.3354095221,0.2851724923,0.1215860769,-0.1429920197,0.0899208263,0.2511619031,0.0925044492,0.0040504015,-0.2070921659,-0.0063689295,0.0486393794,0.1939466298,0.425496459,0.2137170136,0.3945171535,0.0331697948,-0.0591631718,0.2672289014,0.132034719,0.1106245369,0.0543665253,-0.0569675453,0.1894164234,0.4579184949,0.4162216187,0.0169174057,0.0289580468,0.1181012765,0.5321837664,0.0321259201,-0.0732519627,0.1708705276,0.0514870659,0.196275413,0.2941991091,0.2453474551,0.1055003703,0.0069781211,0.0940504149,-0.3743757606,-0.1930261403,-0.0829578266,0.3710646033,0.1964569241,-0.332023561,-0.3283770978,0.0903383866,0.046790693,-0.1340342909,0.0357138962,-0.7238030434,0.2991562486,-0.073135227,0.203206107,-0.4646299183,-0.5324650407,0.2731387019,0.3042718768,0.0662788376,0.2049303651,0.4383504689,-0.1478362381,0.1783230305,0.2021567523,0.1201846898,-0.2224933356,0.0403566547,0.4374471903,-0.300110817,0.1411277354,0.0549574196,0.0160526279,-0.2383214235,0.0554358587,0.200874269,0.0154812522,0.0041684192,-0.1883987784,-0.3138978481,0.0832742676,-0.225464642,0.4769443274,-0.3305907547,-0.159789443,-0.0889857337,0.3159866631,-0.3112288415,0.4488584399,0.3846322596,-0.0216629654,-0.0756221637,-0.1248618364,-0.0264319405,-0.0443770178,0.2965546548,0.3579028845,0.3324427903,-0.2424770743,0.026712453,-0.3693448603,0.0541414805,-0.0543589853,-0.5596653223,-0.1114332601,-0.2042160481,0.3585690856,-0.1240681335,-0.0400462858,-0.2223491073,-0.0023262317,-0.2603672147,-0.1678107083,0.0122163985,-0.3369867802,0.2500520349,0.1526575089,0.1316899806,-0.1519873142,0.3455469012,-0.1487063617,-0.0182658937,0.2930288315,-0.1870582104,-0.0279352758,0.1037204191,-0.0328949802,0.4764545858,0.082005769,-0.2617827654,-0.0165562034,-0.033640977,-0.3932731748,-0.0621455908,-0.3720927536,0.245525226,-0.2771069407,0.0273795277,-0.2961142957,0.1016198099,0.1016144305,0.001855601,-0.3415241539,0.470675528,-0.2211702615,-0.2726424336,0.5616235733,0.2881712317,0.1016227305,0.0880300552,-0.0739528388,-0.0888884813,0.3690780997,-0.1002475619,-0.0697909445,0.0462503843,0.1882796884,0.3721483052,-0.092684187,-0.836920619,0.269846499,0.3096073568,0.0760967061,-0.5116295218,0.2501280308,-0.0500108562,-0.3855341971,-0.0597649068,0.3162653446,-0.2229470015,-0.0423219353,0.0261518378,-0.1755764186]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"We only support http by default for downloading.\r\nIf you really need to use ftp, then feel free to use a library that allows to download through ftp in your dataset script (I see that you've started working on #722 , that's awesome !). The users will get a message to install the extra library when they load the dataset.\r\n\r\nTo make the download_manager work with a custom downloader, you can call `download_manager.download_custom` instead of `download_manager.download_and_extract`. The expected arguments are the following:\r\n```\r\nurl_or_urls: url or `list`\/`dict` of urls to download and extract. Each\r\n url is a `str`.\r\ncustom_download: Callable with signature (src_url: str, dst_path: str) -> Any\r\n as for example `tf.io.gfile.copy`, that lets you download from google storage\r\n```\r\n","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":120,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\nWe only support http by default for downloading.\r\nIf you really need to use ftp, then feel free to use a library that allows to download through ftp in your dataset script (I see that you've started working on #722 , that's awesome !). The users will get a message to install the extra library when they load the dataset.\r\n\r\nTo make the download_manager work with a custom downloader, you can call `download_manager.download_custom` instead of `download_manager.download_and_extract`. The expected arguments are the following:\r\n```\r\nurl_or_urls: url or `list`\/`dict` of urls to download and extract. Each\r\n url is a `str`.\r\ncustom_download: Callable with signature (src_url: str, dst_path: str) -> Any\r\n as for example `tf.io.gfile.copy`, that lets you download from google storage\r\n```\r\n","embeddings":[-0.2777805328,0.0965271667,-0.0294328295,0.2827260196,-0.0488570444,0.0626954138,-0.2193705291,0.4863711894,0.2671810389,0.0046167555,-0.0645394847,0.1767717898,-0.1338637918,0.0276905168,0.1833811402,-0.3053059578,-0.2551793158,0.1110800207,0.0832186565,0.1619187891,-0.3100489378,0.2241034508,0.1274141669,0.0247967672,0.1058143452,-0.0016975444,-0.0335105099,0.2009846121,-0.3600776196,-0.5725451708,0.159604311,0.199882552,0.5002796054,0.3012334704,-0.0001155282,-0.0273309927,0.2771560252,-0.0509039946,-0.1939808726,-0.4507175088,-0.5581098199,-0.3901635706,0.1760960817,-0.0818983316,0.1434405744,-0.3247947991,-0.065694958,0.1504253596,0.1654934138,0.4937476218,0.2138060778,0.2187065184,0.0942479074,-0.1992247105,0.3822670281,0.0258258972,-0.0190181285,0.2136030048,0.2988850772,-0.0795222893,-0.0127831465,-0.2919237018,-0.1655909419,0.2839197814,0.3653366864,0.169727549,0.102001138,-0.4725706279,-0.3483140767,0.2249996811,0.6980264783,-0.0426879153,-0.1774830669,-0.0982262641,-0.1084202006,-0.0485818088,0.1637655646,0.0252196565,-0.38479653,0.1930247247,0.134944737,-0.2261709273,-0.2653417289,0.3329018354,-0.1544879824,0.4816545844,0.0088161603,-0.1243843883,0.3409364223,0.2407482266,-0.1731476039,-0.1758915633,0.0956520438,0.0276539046,0.2391528636,-0.0896978527,0.1219808757,-0.3390654325,-0.045642782,0.3806220591,0.0886813626,-0.0634947419,-0.1267754883,0.2225989997,0.0764587149,0.2798477113,-0.1293370426,0.0132238781,0.4056569636,0.3186863065,0.2450589091,0.0259217806,-0.1233363152,-0.1067257673,-0.3545540869,0.0777444541,0.2355566323,-0.1232114583,-0.1206975356,-0.0501870885,-0.121394664,-0.2151160538,-0.066038765,0.1542132646,-0.1493148357,0.1724778563,0.0286382735,0.1730007231,0.1612342,-0.3008628488,0.0119847218,0.0084588053,-0.0474399365,-0.1504227221,0.217641905,0.1390340924,-0.0804110244,-0.1865072399,0.0279484205,0.1799260825,-0.0033731801,0.0043380656,0.129010573,0.3257303834,0.030877959,0.2170122266,0.0210642368,-0.2257748991,-0.2357034087,0.2544204593,-0.0345582105,-0.3946982324,-0.3626570702,0.15627186,-0.2903930247,-0.4475151896,-0.3203414083,0.1746923178,-0.3312831521,-0.2136016488,-0.1230311543,0.0842836201,-0.232864663,-0.195668593,0.1440956295,0.4552592039,-0.492256403,0.060594257,-0.3768833876,-0.3115023077,0.4496421218,0.0502946638,-0.124092795,0.3914755583,-0.3554272652,0.4173650742,0.8397292495,-0.4267210364,-0.3148742318,0.6333284974,-0.2937240303,0.1136790961,0.2245716155,-0.0514887236,0.4169696271,-0.285752058,-0.6332723498,0.5020344853,-0.0726740584,0.0090590129,-0.2345491052,-0.199662298,0.2574515343,0.0535952002,0.2425493449,0.3365590274,0.3952041268,-0.0429333262,0.3353668153,0.3670587838,0.2471549511,0.0626695752,0.2879798114,0.2796064615,-0.1349156648,-0.2456129342,-0.2629238665,0.1651069224,0.0862020552,0.169700712,-0.1856204569,0.1597983092,-0.2809397578,-0.0604899079,-0.2156412601,-0.0157647207,0.0051061544,0.1655586362,0.2290703654,0.0236069784,-0.1031404659,0.1136211827,0.2875776887,0.307384938,0.1115076169,0.1563326418,-0.1724861264,0.0395464599,0.0204862412,-0.0457603186,0.2629824579,-0.3486967683,-0.158597663,0.1335842758,-0.123903431,0.0464619994,0.3054710925,0.1961762905,0.2601018846,0.1200301051,0.0739792585,0.4652787745,0.2284341604,-0.0791872293,-0.0061101671,-0.0467151813,0.3204243481,0.2881264389,0.0336574428,0.1974166185,0.2212958783,-0.1061479598,0.0918356776,-0.1773681641,0.0986497253,-0.098504819,-0.179042086,-0.2212844491,-0.0927878693,0.1402365565,-0.0492130965,-0.0503070131,0.1286207139,0.3818585277,0.1158388779,-0.0050424086,0.0836059898,0.8543784022,0.3756823242,0.1835933626,0.306401521,0.1088297367,0.1027726009,-0.2426906973,0.2653562129,0.1198403314,-0.3955549896,0.3678884208,-0.0257276297,0.1617210805,0.03200287,-0.2243326455,0.1551394314,0.3608131409,-0.2279530466,-0.0266593248,-0.6431180835,-0.5907304883,-0.1313005239,-0.1926848292,-0.1272398382,-0.4519247413,-0.1167864427,0.3211928904,-0.1044618934,0.0826620832,-0.284847945,-0.0026747747,-0.2534915507,-0.3995335102,-0.2082609534,0.1707179397,-0.192368567,0.0894376114,0.3211081624,-0.0662640333,0.3892353177,-0.0055792588,0.1255531013,-0.5770410299,0.0153351827,0.2387178242,-0.1064808145,-0.0664174706,0.2190169543,0.0265400391,0.337253809,-0.0493886471,0.2074355632,-0.1061513349,-0.123633936,-0.1736947596,0.1422599256,-0.0082121305,-0.1790382564,-0.2249698639,-0.3690969944,-0.6310892105,0.1111525744,0.1660155654,0.1625747383,-0.0175437406,-0.2038410306,0.2674894035,-0.2053000182,-0.1185505614,0.0029458227,-0.1617115736,0.4682974517,-0.3175647259,-0.3787608147,0.2251977623,0.0455160663,-0.136271134,0.1256070286,-0.35362342,-0.2745190263,-0.0791602433,0.3072172701,-0.1262131631,0.2914076746,0.2416874766,-0.0288196709,-0.0462423675,0.0606233366,-0.0025995267,-0.0029891578,0.4793727994,0.3309399188,0.2338623703,0.3097515106,-0.0318109356,0.6106972694,-0.0915334746,0.2176847756,0.5523766875,-0.0916210413,0.1654037833,0.1803812534,-0.238364473,0.074733071,-0.1104736999,0.0982212871,0.2661069036,-0.072028406,-0.0428019911,-0.3615036011,-0.0274920184,-0.1111903116,-0.2581684589,0.0229249876,0.0385503136,0.0067449682,0.0458630957,-0.0043211258,-0.0009690495,-0.1712151617,0.1932059377,0.541592896,0.0281076897,0.136547491,0.3491770923,0.3713872135,-0.5879129171,0.2287244946,0.2666150033,0.0969248042,0.0114947101,-0.0824269503,0.1188708618,0.0548427813,0.6556655169,-0.1353308409,0.0986518487,0.0666243732,-0.2603299022,0.0821249858,-0.1452472657,-0.1661816984,-0.0788561329,-0.1544735283,-0.2195294648,-0.1011289805,-0.3283968568,0.1870178431,0.0621362254,-0.0938748047,-0.2454237491,0.03369461,-0.2624855638,-0.409260273,-0.2815756798,-0.0104149776,0.0559045188,0.2774830461,0.1103716567,0.1333966255,0.0068075182,0.1335189641,-0.1390004605,-0.0577397682,-0.0085129039,0.0141142067,0.0822463334,-0.0128306244,0.1226616725,0.3396207988,-0.1821528226,-0.3488292694,-0.0330330469,-0.0441380367,0.3317430913,0.2396522611,-0.1392124593,-0.0971872881,-0.2428288013,0.2821039557,-0.2954567075,0.1141914278,0.4472608268,-0.3421837389,-0.1252638102,-0.3733779788,0.416162014,-0.1171206534,-0.0647555441,0.1792529672,0.1077392325,-0.1346815526,-0.1354333758,-0.2318955064,0.8524115682,0.0752383471,0.2058204859,0.3293749392,-0.2937966585,0.4327649176,-0.3588069081,0.1133267879,-0.1409437358,0.1773870438,-0.2753989697,0.0096643493,-0.0011534905,0.1388697773,-0.1896541268,0.3798859417,0.0313320793,0.0757569522,0.230248943,0.4625850022,0.0513147414,-0.2804062366,-0.004331111,0.1891535968,-0.2375760823,0.4009000063,-0.1536157876,-0.3216964006,-0.0560976379,-0.1654237062,-0.3222967684,0.1248774827,-0.376567632,0.1581715345,-0.0971781686,-0.2147282064,0.1250221133,-0.0915197432,-0.4460642338,-0.2091538757,-0.1902590543,-0.1778602898,-0.2255990803,-0.0631957352,-0.0331118219,0.1158379242,0.2146817893,-0.2299425155,0.0233681872,-0.1044761464,-0.0207359493,-0.3708561659,0.3247354925,0.1927947998,0.072100319,0.0160151068,-0.2242674977,-0.080971241,-0.151015833,-0.1725269854,0.1021626443,0.0208582152,-0.3076504767,0.0750787854,-0.1099052653,-0.1160028428,-0.1093643382,0.316216439,-0.3857374489,-0.0382144526,0.1901695877,0.1664547473,-0.2660354376,-0.1714635491,0.0781780481,-0.3098061085,-0.2643665969,0.2415321767,-0.0749417245,-0.0800470859,0.064437978,0.507814467,0.2872799039,-0.1613528877,-0.1703559458,-0.413156271,-0.0904966742,-0.2093134373,-0.1304544061,-0.0088740587,-0.3474535346,0.1425922066,0.2795874774,-0.259778738,-0.2710268795,0.1126305908,0.0846962109,-0.1183348522,0.6158233881,0.1940929294,0.3525829613,0.1051703021,-0.018004084,0.0892750248,-0.3229301572,-0.0869442746,-0.2921779156,0.186701566,0.1633133739,-0.2184811682,0.1312788427,-0.1773484796,-0.2802200019,0.1531658173,-0.0902672708,0.1098241508,-0.0582629368,0.1060827598,0.2737483382,-0.2235983908,-0.1254300922,0.3725326061,0.156497851,0.6829890013,-0.1120122746,0.0962002352,-0.1482327282,0.014458809,0.0712997541,-0.0329744071,0.0731552541,-0.1383981556,0.5538968444,-0.0048992336,0.0972772911,0.4584711492,0.4045279324,0.38379246,0.2066975832,-0.017228771,0.4155067801,0.0779508874,-0.2180012912,-0.1995036751,0.21274288,0.2397638857,0.0316163078,0.2869144976,0.0391414799,-0.0666026026,-0.0769210234,-0.0946746245,0.410094291,0.1849676818,-0.0245686844,0.3254160881,-0.0883878246,0.26181373,0.1731977165,0.3311846554,0.2310250103,0.1246363744,-0.2521328628,0.0938148424,0.0803409964,-0.0041574547,0.0959158242,-0.4635386169,0.1409638673,-0.0577626713,0.1350248605,0.0403649285,0.1422886699,0.391353637,-0.3291977644,-0.0282009393,-0.18395105,0.3366115987,0.1261387318,0.2924504578,0.1812870055,-0.0276791099,-0.5020226836,0.024705179,0.0370319188,-0.2099049687,-0.0175288226,0.1742316335,0.1238446012,-0.1536006629,-0.2406488508,-0.3937408924,0.008739097,-0.387493223,-0.1205585003,-0.0282662604,-0.2779155374,0.0365630202,0.1705744267,0.3475547731,0.2376671582,-0.0724418908,-0.1863111854,-0.3535120487,0.1091940328,-0.1476799548,0.2919328511,0.1634677351,0.1949480772,0.3043048978,-0.0075095873,-0.1206732169,0.002449478,0.0473204814,-0.2435740381,-0.0634313673,0.8210984468,0.2703952491,0.0882317722,-0.1724933386,-0.1439775229,-0.3292527795,0.2324587405,0.21502994,-0.1479426771,0.1105021387,-0.1574006677,0.0451663695,0.0527735539,0.3405618072,0.4622702897,-0.0701694414,-0.1550057828,-0.0200468246,-0.9037473798,0.1082122475,0.0543433279,-0.1670973599,-0.080864042,0.0378517024,0.1100160033,0.4262301326,0.1255442053,0.2361948937,0.1714093536,-0.1668512672,-0.0630202219,-0.1099749729,-0.0954844132,0.1355502009,-0.0711727366,-0.1057145372,0.0740881786,-0.4191245139,-0.0196009167,-0.2305950522,-0.2496817261,-0.1324823648,-0.1511650383,0.5930138826,0.0655725226,0.2280219346,-0.0595616587,-0.0824397132,-0.1598233283,0.3103452325,-0.3625293076,0.0681564733,-0.1976426989,0.0475333333,-0.2573741674,0.4077554941,-0.7308899164,0.0816035271,-0.1295918077,0.3909314275,-0.4042830467,-0.0866027027,-0.3370788991,-0.0990028232,0.467001617,0.1077034324,-0.0946613476,0.1264566183,-0.4665576816,-0.2809146941,0.3255301118,-0.0217460878,-0.1913801581,-0.1035375893,0.0917188972,-0.0933693945,-0.035032548,-0.2697997689,-0.193620488,0.1508775949,-0.0902043059,-0.0572738051,-0.001144612,-0.2958563566,-0.1792342961,-0.0286260098,0.2602679729,0.0896153301,-0.13442415,-0.0527024828,0.050984066]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"Also maybe it coud be interesting to have a direct support of ftp inside the `datasets` library. Do you know any good libraries that we might consider adding as a (optional ?) dependency ?","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":34,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\nAlso maybe it coud be interesting to have a direct support of ftp inside the `datasets` library. Do you know any good libraries that we might consider adding as a (optional ?) dependency ?","embeddings":[-0.3274756074,0.0598696545,-0.0164928641,0.2534114122,-0.0806325227,0.0911129639,-0.3827839494,0.4749818444,0.2600127161,-0.012885185,0.0683740675,0.0932430774,-0.1218968257,0.061171934,0.2366616279,-0.2707274258,-0.1249640808,0.167685017,0.0703390911,0.1570510566,-0.3043116629,0.2058279514,0.1108845472,0.0007483974,0.0134709068,-0.0703274235,-0.0722677484,0.1834948659,-0.2807340622,-0.4544092715,0.1355403811,0.1415678263,0.4248403609,0.257706821,-0.0001219638,-0.0764313564,0.2710136175,-0.0649453849,-0.145114854,-0.4268159568,-0.5726474524,-0.4681797624,0.2224383056,-0.0011480418,0.1972098202,-0.3350039124,-0.0732626393,0.0971458182,0.0442159399,0.4594890773,0.1673590541,0.2992135882,0.0093542505,-0.1778156161,0.4375674129,0.0231842883,-0.0504501872,0.153066352,0.3597561121,-0.1422450393,0.0555440299,-0.3372543752,-0.1355669498,0.17609936,0.3862240016,0.1103014052,0.0883987769,-0.4388636351,-0.3785873353,0.1965651363,0.8066228032,-0.0113757458,-0.2615829706,-0.1113187373,-0.0873063803,0.0143470513,0.1853109151,0.0202660579,-0.3324308693,0.1831110269,0.1143505871,-0.3629404008,-0.252194494,0.4006340504,-0.1277102828,0.5490372181,0.1106914505,-0.1825642735,0.2825981975,0.2046848089,-0.1684248596,-0.1567012221,0.0916963443,-0.0014251804,0.2913459241,-0.1304873973,0.1080868021,-0.263481915,-0.0174486358,0.3301560879,0.0243162625,-0.0238242205,-0.2186668217,0.2281159759,0.0460556336,0.229918018,-0.1359343827,-0.0047339913,0.5143220425,0.1797383577,0.3002929091,0.0320505537,-0.1394660324,-0.0609621108,-0.4214181006,0.0173486751,0.1475913525,-0.1914305687,-0.1164726242,0.0196850393,-0.1824108958,-0.0891887099,-0.0083674528,0.1318999231,-0.1650015265,0.264490068,0.0364155918,0.1938411891,0.1065606028,-0.3565620482,0.0130211208,0.0204761699,-0.0704325736,-0.1268787384,0.190478757,0.1413714439,-0.0339485891,-0.1371247917,0.0703350455,0.2175328285,0.0164672732,0.0222992525,0.071576938,0.4026202857,-0.0590976477,0.1796619445,-0.0069614006,-0.1225644648,-0.2095035166,0.2328031212,-0.0161455404,-0.3932777047,-0.4267930686,0.0802333504,-0.3305881917,-0.4670033455,-0.275801599,0.1919849217,-0.4273515642,-0.3020412624,-0.125728175,0.1029542983,-0.1907224655,-0.2069382668,0.1016146988,0.3802587688,-0.4875880182,-0.0005755357,-0.3736157119,-0.2895914912,0.4132218957,0.1140806377,-0.1112021208,0.3359066546,-0.3094168603,0.3546554148,0.8370084167,-0.4016798437,-0.2758085132,0.6220941544,-0.2426191568,0.0267043393,0.1759946644,0.0300458558,0.2986840606,-0.2243283987,-0.6506483555,0.5972316265,-0.0663956851,0.0441403799,-0.1612779647,-0.2266774178,0.3191091418,0.1394774914,0.3316980004,0.2728152871,0.418646276,-0.0235842355,0.2792365849,0.3195464313,0.1983102113,0.1458649486,0.2886610627,0.3233473003,-0.1175427511,-0.3403542042,-0.2940297127,0.0807635486,0.1218628213,0.145616442,-0.1783807725,0.1935077459,-0.122542657,-0.0336224549,-0.0697396398,0.0909017399,-0.0722767636,0.181281209,0.1942894012,0.0414849557,-0.1086733341,0.12853688,0.2902016938,0.318875879,0.1650923342,0.2465964407,-0.1522357166,0.0011203651,0.1121921167,0.0614286289,0.2203390896,-0.3315290511,-0.1351704597,0.092469655,-0.1107333153,0.0936811343,0.4048472345,0.2875028849,0.3451217115,0.1375146806,0.0838565826,0.4250737131,0.1754143536,-0.0993347764,0.0376974717,-0.014303389,0.2977044582,0.217909053,0.0552522168,0.2402647734,0.1647406667,-0.0411328562,0.0723397285,-0.1136398315,0.2451276779,-0.0986617208,-0.157007724,-0.1735036671,-0.1777079999,0.0258666202,-0.1199384108,-0.0781918764,0.1950016618,0.3921463192,0.1654786468,-0.0244768504,0.2006291151,0.7358022928,0.3375622034,0.2228246927,0.2597005367,-0.004251441,0.0336218365,-0.2456354797,0.1866513938,0.1149216965,-0.4359814227,0.4135332406,-0.0206876155,0.1957435459,-0.0184401404,-0.1528042108,0.1500848979,0.3519281745,-0.1117857918,-0.0641986206,-0.6870807409,-0.6479420066,-0.0983839184,-0.1577105969,-0.2821214199,-0.4398795664,-0.1192172542,0.2665982544,-0.0242006164,0.1208501086,-0.2807169557,0.0926330313,-0.38007164,-0.4477750063,-0.267439127,0.1894757897,-0.1792530268,0.0328132771,0.344219923,-0.0277355686,0.4250334799,0.0987618342,0.1917737126,-0.6420844197,-0.0550661869,0.1851613373,-0.0226981044,-0.0905228183,0.1784199625,-0.0125826104,0.3820634782,-0.0953880697,0.1900233775,-0.1293125004,-0.11783766,-0.1879602373,0.08741723,-0.1212427691,-0.1198334172,-0.0994132608,-0.4218278229,-0.6161510348,0.077366069,0.1289636642,0.0838463902,0.0494505242,-0.2287164181,0.2352489531,-0.1928627938,-0.0224560052,0.0970234796,-0.0697934255,0.4639669359,-0.2383593023,-0.42216748,0.2020000815,0.0862178504,-0.1741467863,0.2055050582,-0.3952102065,-0.3099576831,-0.1374028027,0.3529423773,-0.1657400131,0.2932446301,0.2597908378,0.0002999851,0.0142573928,0.0154755171,-0.0873998627,-0.0309707597,0.4773543477,0.3319088817,0.1926017851,0.2289376259,-0.0619901903,0.5613712668,-0.0307525713,0.1898088902,0.5820297599,-0.0288637355,0.1707128584,0.2009478211,-0.2395555526,0.0938971043,-0.0521430336,0.1095566303,0.2972458601,-0.0014327572,-0.061559286,-0.4131017625,-0.0158431083,-0.0998506919,-0.1463207752,-0.0108167809,-0.0215873811,0.1161765531,0.0083655734,-0.0709476471,-0.0334242731,-0.1968568116,0.2670749426,0.5796684027,0.0098497933,0.1781524867,0.422267884,0.2526375949,-0.5315039158,0.195726037,0.3185509443,0.2280303091,-0.1253430545,-0.0067171887,0.0833459049,0.096614413,0.6473723054,-0.2057652324,0.0129504651,-0.0232142936,-0.2333059311,0.1315872669,-0.0692591369,-0.1822122633,-0.0757050216,-0.1233205721,-0.1422018856,-0.0206315629,-0.3912426829,0.1817764193,0.0007341158,-0.0052395198,-0.214448154,0.032490056,-0.1783387363,-0.3002638221,-0.2489537448,-0.05332857,-0.0423792116,0.387375772,0.0693248659,0.1101591885,0.0047998964,0.1454618871,-0.1333232075,-0.1317717135,-0.0213034358,-0.0248574913,0.0687914491,0.0237296969,0.166273728,0.3021893203,-0.2050030231,-0.3380509615,-0.044901114,0.0523116551,0.4842704833,0.2997088432,-0.0966040343,-0.0566530377,-0.3616673648,0.2137583941,-0.320751369,0.0823279023,0.4879281521,-0.2429908067,-0.1529644579,-0.260897994,0.4238541722,-0.013433665,-0.0632636398,0.2081471086,0.1353587359,-0.1122121811,-0.140491277,-0.1748726815,0.8917751908,0.0603307858,0.2355369925,0.3040161133,-0.2715129554,0.4350160658,-0.3444151878,0.1655257791,-0.1084212288,0.0808987021,-0.3537234664,-0.049186267,-0.0066245594,0.0630970001,-0.2247193456,0.4155174196,0.1061229408,0.0291219484,0.2251032889,0.5095053315,0.0706729367,-0.2419833243,-0.0523020215,0.1287378967,-0.2999565601,0.4243385196,-0.2171368003,-0.2767744064,-0.0605714768,-0.1455177218,-0.399713397,0.0774171054,-0.3578985631,0.1576014161,-0.0801990926,-0.1929141134,0.1894233972,-0.0851839632,-0.424215436,-0.1446505934,-0.2077785581,-0.1700901091,-0.3140527308,-0.0997422934,-0.0234303921,0.1824529171,0.2166316658,-0.1875996143,0.0805773214,-0.155688405,-0.048763115,-0.349639684,0.3244370818,0.1433216929,0.0986517742,0.0827324837,-0.2016393989,-0.0164320134,-0.1232625842,-0.1591062099,0.0567836203,0.0518706925,-0.1336106509,0.0316607244,-0.0367127508,-0.1021822169,-0.1498824805,0.3496854007,-0.4334790409,-0.0269989725,0.120353058,0.2029506564,-0.334446162,-0.1306482553,0.1898956746,-0.327611506,-0.1984561384,0.2118930519,-0.0969294012,-0.1923658997,-0.0317956097,0.5712575912,0.368262887,-0.2154659629,-0.2258192748,-0.4168668687,-0.0864836946,-0.0604599491,-0.1216491684,0.0431944467,-0.3707618117,0.0794191286,0.3085111678,-0.2533963025,-0.1913246661,0.0808724612,0.0539453663,-0.101163134,0.5919700265,0.2054125071,0.3187373579,0.0919776037,-0.0924517363,-0.0520374887,-0.3047579527,0.0018518439,-0.2380412072,0.2110367268,0.204767108,-0.1962421536,0.1818410903,-0.2734619081,-0.2555222213,0.1578818411,-0.1329609454,0.0836247206,-0.0757939816,0.1405834854,0.3700135052,-0.2467726916,-0.0867529884,0.4178487957,0.2024151981,0.6275689602,-0.1459679604,0.0796029717,-0.0440056249,-0.0145969754,0.0605165996,0.0258421116,0.0859874561,-0.2036667913,0.5231630206,0.035040047,0.0709355101,0.5666955709,0.4435304105,0.3741248548,0.239710182,0.0741062462,0.468072176,0.0236284919,-0.2743926942,-0.3609838486,0.1873316616,0.2658543587,-0.01225923,0.2377750427,0.0419111289,-0.0460754707,0.0110874297,-0.1038415208,0.2900448442,0.1910845041,-0.0134967677,0.284881711,-0.0172860548,0.1840373427,0.0829845518,0.3664553463,0.1889047027,0.0957409963,-0.175790295,0.0497316122,0.0919539407,-0.0033337916,0.1116234884,-0.4135985374,0.1131742671,-0.0520080812,0.278986156,-0.007718124,0.1429690272,0.487472564,-0.3883010149,-0.1104035452,-0.2031031847,0.294211328,0.0637609512,0.372656554,0.2644439936,0.0005417435,-0.5394998789,0.0317131244,0.0097181834,-0.1933443248,-0.0180448834,0.2489153892,0.2446534336,-0.1977079064,-0.192632392,-0.3577364683,-0.0176846739,-0.3825719357,-0.0630708858,-0.1717903614,-0.2582587004,0.0350048244,0.2416263968,0.3221454918,0.1953469813,-0.0602960885,-0.2482363731,-0.3852999806,0.1303257942,-0.074746117,0.3600233793,0.269394666,0.2696221471,0.2516016066,-0.0916647986,-0.0394632779,0.0659736991,-0.0240453687,-0.3908688724,-0.1267070621,0.862013638,0.2531178594,0.0156932343,-0.2131469697,-0.1203512847,-0.25330019,0.2296481282,0.1763811558,-0.170840472,0.0227278545,-0.1361280829,0.0022075477,0.0250170007,0.3607389331,0.5824887753,-0.0028288725,-0.1660872698,-0.1079355404,-0.802097559,0.1326081455,0.0481190979,-0.1215113476,0.0048921965,-0.1043617278,0.0488994233,0.3867866993,0.1092296764,0.2687337101,0.1751996875,-0.2665206194,-0.1097715497,-0.0943668261,-0.063622959,0.0400733277,-0.1191881523,-0.0213764533,0.1035271809,-0.3948248029,-0.0899235532,-0.2828958631,-0.2706706524,-0.2346845269,-0.0991405323,0.5682695508,0.0655271336,0.3026096821,-0.0307279211,-0.1387579292,-0.1106653735,0.1714862883,-0.3984921873,0.0577309802,-0.2491495162,-0.0179353524,-0.3361113667,0.4585196972,-0.6438986063,0.0134512698,-0.177330792,0.4412663877,-0.4494836628,-0.1079820842,-0.4080908895,-0.0742144808,0.4904831648,0.1567518115,-0.120558776,0.077031076,-0.4493555427,-0.1922594458,0.278919071,-0.0046790238,-0.0711623356,-0.111241214,0.0880340263,-0.0942263603,-0.1011068299,-0.258911103,-0.2988597155,0.1036855206,-0.0099991783,-0.0829782262,0.0138975568,-0.2336737663,-0.2074281871,-0.0179911871,0.3222567737,0.0870110542,-0.1465880275,0.0722438544,0.0400179327]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"Downloading an `ftp` file is as simple as:\r\n```python\r\nimport urllib \r\nurllib.urlretrieve('ftp:\/\/server\/path\/to\/file', 'file')\r\n```\r\n\r\nI believe this should be supported by the library, as its not using any dependency and is trivial amount of code.","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":35,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\nDownloading an `ftp` file is as simple as:\r\n```python\r\nimport urllib \r\nurllib.urlretrieve('ftp:\/\/server\/path\/to\/file', 'file')\r\n```\r\n\r\nI believe this should be supported by the library, as its not using any dependency and is trivial amount of code.","embeddings":[-0.2455563694,-0.0056827385,0.0114010274,0.2680145502,-0.0810396895,0.0563570522,-0.3472591639,0.4960227907,0.210584417,0.0559276417,0.0054219542,0.1040770933,-0.1624985635,-0.0003008154,0.2173724473,-0.3047148883,-0.1787896901,0.1610744596,0.1181761697,0.1564905196,-0.3281778991,0.2467977852,0.0930166245,0.043319881,0.036886517,-0.0102594495,-0.0181726627,0.231210947,-0.3351477981,-0.4214291871,0.1100811958,0.0631712005,0.4437259138,0.1806788743,-0.0001204356,-0.0327388421,0.2680413425,-0.1002686322,-0.104510501,-0.4727423191,-0.4627041221,-0.4231175184,0.2234623879,-0.0714606047,0.1594662964,-0.2432890236,-0.0140075972,0.2061714381,0.0663921311,0.4829704463,0.1702732891,0.3050340414,0.0770800337,-0.1257840693,0.4577258229,-0.0401502363,-0.0512701757,0.248623386,0.2742137909,-0.0884236693,0.0582275316,-0.3440451622,-0.1427604556,0.252089411,0.3116759062,0.1652508527,0.1236155704,-0.4026138186,-0.3681646883,0.1997761279,0.63075459,-0.0083609242,-0.2014279366,-0.0267409869,-0.1026719362,-0.0145607283,0.1862136722,0.0254227892,-0.3613951802,0.1787406057,0.1856765449,-0.3242506385,-0.2554449737,0.3719015419,-0.1366192997,0.5931152105,0.0658612102,-0.1004417092,0.2225957811,0.2331712842,-0.2562013268,-0.0763917342,0.1533754021,0.0574023239,0.342338115,-0.0806174129,0.0626318008,-0.32030195,0.0128945205,0.3063532114,0.1055897027,-0.0480534174,-0.1900885701,0.2983517647,0.0536059104,0.3081001043,-0.17290093,-0.0186953414,0.5094934106,0.2821047008,0.2624134719,0.0560330637,-0.1178547964,-0.1194995865,-0.462259084,0.054779388,0.2014470696,-0.1590370089,-0.1283294559,-0.0145571018,-0.2463677675,-0.0972510129,0.0239095557,0.1393690854,-0.2065571994,0.1906870902,0.1232686862,0.1359563023,0.0505211502,-0.2750120759,0.0420158543,0.0029452278,-0.0010131976,-0.155328095,0.1617364585,0.1152453199,-0.0066440976,-0.1665710807,0.0627628565,0.1509384364,0.0086926483,-0.0056089424,0.1514650434,0.3983240128,0.0141673498,0.1728252769,0.0121454336,-0.1598490328,-0.243770197,0.2962782979,-0.1052594557,-0.3810894787,-0.4174828231,0.0826406702,-0.2794828713,-0.4609342813,-0.2328967601,0.1762951463,-0.3809169233,-0.3419992924,-0.102248162,0.0753771663,-0.2319997251,-0.212921828,0.1630425602,0.3832987845,-0.3571052551,0.007798316,-0.2994357646,-0.298350215,0.5073300004,0.1147064492,-0.1210711449,0.3324902356,-0.3670815527,0.4251608253,0.8384565711,-0.3629837632,-0.2417307049,0.6288360357,-0.2879821956,0.0591514036,0.1531856209,-0.0149559369,0.2839210927,-0.2343828976,-0.5873812437,0.5538029671,-0.0242186021,0.0745490938,-0.2459794879,-0.185032025,0.3185222745,0.0530269071,0.2675327063,0.3087219894,0.4214221239,0.0938497633,0.3757402897,0.3142935336,0.2340296209,0.1025049537,0.3178471029,0.2957000732,-0.0954593495,-0.3132984042,-0.2004291862,0.0692326352,0.0264658313,0.194797948,-0.1536310762,0.1266146898,-0.1657294184,-0.0594669282,-0.1859846115,0.0433253907,-0.0508963764,0.1662818044,0.1799157411,0.1008666158,-0.0685695037,0.0021913652,0.2449868172,0.3067437112,0.1160969734,0.1882428378,-0.1808428913,0.0640043169,0.0556332879,0.0540285408,0.2409528196,-0.3728549182,-0.1152889803,0.1083641201,-0.0711680353,0.0535149314,0.3289978206,0.2119719535,0.3346286118,0.1550185978,0.0265759416,0.5274912715,0.2497471571,-0.0545664579,0.0958918482,0.0023235921,0.3467338383,0.2445716113,0.0871291831,0.2104832679,0.1962825507,-0.0432377532,0.0579768345,-0.1230782047,0.1969663203,-0.066023387,-0.2364279628,-0.2281231135,-0.1946508437,-0.0071621803,-0.0965443328,-0.0657776818,0.2163253576,0.4139409363,0.2313465625,-0.0124053061,0.1543519199,0.6730174422,0.2891684473,0.1565618217,0.3078869581,0.1071752459,0.0316325985,-0.253988862,0.2010782361,0.1231395602,-0.389958173,0.4704409838,0.0261115432,0.1698914617,-0.0346687436,-0.139299795,0.1403881162,0.3499347866,-0.1492049992,-0.0147421025,-0.6423718333,-0.6558085084,-0.1747983396,-0.1938220263,-0.2456446439,-0.4119878709,-0.1122543067,0.2524807751,-0.0747434944,0.0646307245,-0.1988042891,-0.002335784,-0.2858401537,-0.4750635922,-0.183858797,0.2437696606,-0.2298301756,0.0236968584,0.3109210432,-0.0177714881,0.3862735629,0.0264210552,0.1282990426,-0.6088418961,-0.1003479213,0.2114328593,-0.0697606131,-0.1035302505,0.1966163367,-0.0041562659,0.3545367718,-0.0289515182,0.2459471375,-0.1386429369,-0.1292089224,-0.1795073897,0.1364054531,-0.0362183787,-0.2145157307,-0.1819323599,-0.4030963778,-0.6090556979,0.1008850932,0.1400819123,0.1651463211,0.0233957283,-0.2871911228,0.2522779405,-0.1295570284,-0.0499394946,0.0600947663,-0.0067992806,0.4296649992,-0.3297210634,-0.4547201693,0.2211284041,0.0919658989,-0.158678636,0.1415458918,-0.3535096049,-0.3210872412,-0.0745454282,0.2627055645,-0.132836625,0.3481777012,0.2570135295,-0.0688443854,0.0271192361,0.0475400575,-0.0865596011,0.0000654736,0.514926672,0.3293808699,0.2759384811,0.2812847793,-0.0223778728,0.5285961628,-0.0103555033,0.2793345153,0.6546480656,-0.1272286177,0.1461856961,0.1091930568,-0.1454851478,0.1041269079,-0.101027526,0.0010329379,0.3067156374,0.0013138877,-0.0367011353,-0.438008517,-0.0723110661,-0.1485421956,-0.1934870631,-0.0064141853,0.0628310218,0.0955515504,0.062379837,-0.0814205706,0.0075304168,-0.1989013702,0.2500849068,0.649758935,0.0407849811,0.1961861104,0.4838578105,0.2168342024,-0.5869781971,0.224169746,0.3614320159,0.2293214053,-0.0843847841,-0.0729418546,0.1597599983,0.019655969,0.5835807323,-0.135507375,0.0800044462,-0.0479177572,-0.2808924913,0.1180665269,-0.084871456,-0.146149233,-0.0691081434,-0.0731122717,-0.1658737808,-0.0676342994,-0.34342134,0.1716650426,0.0335461609,-0.0832164437,-0.1453421116,-0.0816606283,-0.287314564,-0.3666789532,-0.3185653985,-0.0382613465,0.0556676909,0.4373838603,0.0862969011,0.1122577339,-0.0060758539,0.0873371363,-0.128738746,-0.0676563755,-0.0199927092,-0.0933449119,0.0826147795,0.0600007735,0.0529027693,0.3506924212,-0.2123598456,-0.294642657,0.0113191577,0.036757838,0.4046431184,0.276871264,-0.130267024,-0.0820613801,-0.26342085,0.2308738679,-0.3231474459,0.0556549989,0.4918873906,-0.2641445696,-0.1055584177,-0.2606639266,0.4776839614,-0.0560221598,-0.0490369536,0.1593175083,0.1289281547,-0.12479233,-0.1060231328,-0.2019780725,0.8354321122,0.0649856031,0.2609160542,0.3359478712,-0.2869058251,0.3655300736,-0.3152897656,0.1893507391,-0.1061523855,0.2169631869,-0.3154323101,-0.0082638394,-0.0213815644,-0.0181404687,-0.2020681947,0.3630113602,0.0261504482,0.0235218648,0.1912669539,0.5003851652,0.0390172265,-0.1895582229,-0.0844311938,0.1286360323,-0.254157424,0.4741061628,-0.1847448647,-0.3270057738,-0.0958316624,-0.1978807151,-0.3076323569,0.0921462178,-0.3734368086,0.1551284939,-0.0893386751,-0.1881128997,0.0538570285,-0.1138696298,-0.4920062721,-0.2124909312,-0.1942313313,-0.2015946954,-0.3381808102,-0.1031343564,-0.0485131368,0.2205061316,0.2001671493,-0.261387527,0.1094696298,-0.1464538276,-0.064246051,-0.4030819237,0.3618185818,0.2063464373,0.0949151292,0.1159401014,-0.206764549,-0.041754663,-0.1645783782,-0.1535397619,0.0540538132,0.0112247542,-0.2702167928,-0.029682029,-0.0247210022,-0.1064534113,-0.1103557572,0.3347574472,-0.3714750707,0.0169313159,0.1162068471,0.1637612432,-0.2936356366,-0.0866940469,0.1692095101,-0.3307695985,-0.205976814,0.2427511066,-0.0761063024,-0.1194311306,0.0063624992,0.702466011,0.3082853556,-0.1577439606,-0.3027270138,-0.4830198288,-0.073663801,-0.1577166468,-0.1569102854,-0.00997198,-0.3055683374,0.183736667,0.3220192194,-0.3346721232,-0.1891777217,0.078967087,0.0266300347,-0.122843869,0.5291396976,0.1935974658,0.3772794604,0.147927165,-0.0849591345,0.0858315453,-0.257689327,0.0042257663,-0.272908777,0.2002912015,0.2139132023,-0.1712884456,0.0827091262,-0.2489266396,-0.3137808144,0.1957261115,-0.0881544948,0.0901515335,-0.0784056708,0.1302753687,0.2613147795,-0.2927399576,-0.0622371063,0.3811442554,0.1899445802,0.6598369479,-0.1306340247,0.0564748049,-0.1597788483,0.0279124733,0.0813171118,-0.0285926294,0.0531030968,-0.2434137613,0.4898528755,0.0037947611,0.0948319137,0.4420147538,0.4499049783,0.3753207326,0.2510335445,0.0683608651,0.4525552988,0.0076003796,-0.2783553898,-0.252345562,0.1419274509,0.2004345655,-0.0196965523,0.2268148363,0.0370485708,-0.0154264262,-0.0489892513,-0.0836614892,0.2508754134,0.1625110507,-0.1040665209,0.2833546996,-0.0776140839,0.2967441976,0.0831736922,0.3155852854,0.2682251334,0.0576899871,-0.1999992728,0.0581847951,0.068782717,0.0213815458,0.1277673542,-0.3788020611,0.066799596,-0.0872437209,0.2311303765,0.0379736125,0.1320905089,0.5408688188,-0.36582008,-0.0806363449,-0.2153704464,0.343701154,0.0707582235,0.3460200727,0.2400043458,0.018942235,-0.527771771,0.0124398042,0.0408414155,-0.1840171069,-0.0811839253,0.2307440937,0.1991312355,-0.1430063248,-0.2762212455,-0.3814669251,-0.0030757422,-0.3562513888,-0.0800009295,-0.1303960234,-0.2521276474,-0.015186782,0.1799149662,0.314827323,0.2418599129,-0.090104416,-0.2086496949,-0.4039201736,0.1592064053,-0.1104486436,0.3354885876,0.3182710111,0.2607212365,0.3081428707,-0.0947030932,-0.0463319272,0.074378863,-0.0318073854,-0.2770717144,-0.0186240785,0.898235321,0.1729502529,0.1287475675,-0.227530688,-0.1510954946,-0.3705109954,0.2907259762,0.1069380566,-0.1794191748,0.0605520718,-0.1141295061,0.0029370894,-0.0107164523,0.2982536256,0.5337735415,-0.0062849927,-0.0811569691,-0.0855011344,-0.8763769269,0.1558041722,0.1167662069,-0.16924496,-0.0115709333,-0.0814387724,0.005458218,0.3687034845,0.0448691808,0.1964470297,0.2288792282,-0.2801899314,0.0005456043,-0.107941255,-0.1100930423,0.0365700573,-0.1058528423,-0.0856233984,0.0952273309,-0.4726800323,-0.0680822134,-0.3004283011,-0.2796398997,-0.2702931762,-0.0819357038,0.5554597974,0.1181739867,0.2426398098,-0.0563713238,-0.0953730196,-0.1059931666,0.2550345957,-0.3881648183,0.0448978394,-0.2764706016,-0.0022369861,-0.2648373842,0.473952204,-0.7042845488,0.0773477778,-0.1230427027,0.3312744498,-0.3687281311,-0.1133645773,-0.393311888,-0.1210933551,0.4776001871,0.2249873877,-0.104355745,0.0414955691,-0.478423506,-0.1934379488,0.2941083312,-0.0705824941,-0.0907491148,-0.0579391234,0.1131941527,-0.0954375118,-0.1294773668,-0.3328773677,-0.2636527717,0.1053806543,-0.0307676047,-0.1391648948,0.0020565402,-0.2911555171,-0.2257433385,-0.0343778357,0.2167581469,0.1243622378,-0.1454560012,0.0569386296,0.0046372982]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"I know its unorthodox, but I added `ftp` download support to `file_utils` in the same PR https:\/\/github.com\/huggingface\/datasets\/pull\/722\r\nSo its possible to understand the interaction of the download component with the ftp download ability","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":33,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\nI know its unorthodox, but I added `ftp` download support to `file_utils` in the same PR https:\/\/github.com\/huggingface\/datasets\/pull\/722\r\nSo its possible to understand the interaction of the download component with the ftp download ability","embeddings":[-0.2079318464,-0.0998929814,-0.0202530045,0.3010161221,0.0044642645,0.038478259,-0.3553130329,0.3622789085,0.1900351942,-0.0035547365,-0.0368139669,0.0217449348,0.0155007364,-0.0038925642,0.2229107469,-0.3022278845,-0.1582735032,0.0847345889,0.0404505357,0.1436005384,-0.3097785711,0.2319472432,0.1025383323,0.0332681201,-0.0504573658,-0.014419958,0.0126235299,0.216467768,-0.3532263339,-0.4517711997,0.1148509458,0.0617914759,0.3427506685,0.1893779635,-0.000120639,0.0197103098,0.3356402516,-0.0973493159,-0.2229550034,-0.4683343768,-0.5454901457,-0.411131382,0.1993215382,0.0049577951,0.1873623431,-0.2422083467,-0.0317680351,0.2030040175,0.0860505104,0.4672448039,0.1789339334,0.3690392673,0.0318196863,-0.1910261363,0.3959606588,0.0568572283,-0.0408926457,0.2149498016,0.3350120485,-0.0349815078,0.0446458422,-0.3629597723,-0.0632884279,0.2762863636,0.3633669019,0.2289074808,0.1841487586,-0.4370957613,-0.2981275022,0.1590768248,0.699360013,0.0777133629,-0.2732343376,-0.1167283952,-0.0868948326,0.0861626565,0.2511809468,-0.079638727,-0.3449171782,0.1491218954,0.1730009615,-0.2589472532,-0.170397833,0.3341841996,-0.197999388,0.5212568045,0.0061053308,-0.1265738457,0.2468764633,0.2629101872,-0.2788716853,-0.1374629885,0.1785530448,0.0504078344,0.3377340138,-0.0213782303,0.1397557408,-0.2761086524,-0.0083406651,0.3188734949,0.1086752042,-0.111659728,-0.2271296084,0.2703460753,0.0646889135,0.3599967659,-0.1743905246,-0.0469691567,0.3693583608,0.3174996674,0.2482461184,0.006688117,-0.0072048814,-0.1123003885,-0.4965158105,-0.0217914972,0.2988591492,-0.1590376794,-0.2438595444,-0.0120566264,-0.1646545678,-0.1358545572,-0.054800421,0.15742217,-0.1514691859,0.1966957748,-0.0269040912,0.2070281357,0.082897678,-0.2286737561,-0.0190481041,-0.0415509604,-0.0312383361,-0.2148872614,0.161039114,0.1214091107,0.0064423801,-0.1272520274,0.0971123502,0.0847668126,-0.041205138,0.0134572312,0.1249888688,0.4334996343,-0.0127799418,0.2451587021,0.0644340813,-0.2286069989,-0.2067031562,0.2412199974,-0.1036896259,-0.4144616127,-0.3034451008,0.1017002612,-0.3632735312,-0.3726772368,-0.1595049798,0.1136647165,-0.3486055434,-0.315356493,-0.086832732,0.11650341,-0.1531251967,-0.2190992236,0.1667493433,0.383581847,-0.4564204514,-0.047297094,-0.2854032815,-0.3417429924,0.506298244,0.1772540808,-0.1460350007,0.3434446156,-0.353361845,0.4306581914,0.8133015037,-0.3829745352,-0.3716633618,0.705005765,-0.4238267839,0.2211805135,0.2233175039,-0.0965155438,0.2633066177,-0.2493056655,-0.5300371051,0.4649730027,-0.0029226474,-0.0015167119,-0.2496085465,-0.1350483745,0.2841676474,-0.0408967622,0.2322322428,0.337898463,0.3964929879,0.051944647,0.3205003142,0.3770959973,0.30320099,0.0766608566,0.2393833399,0.2546314597,-0.0940833017,-0.2554593384,-0.2583701611,0.1295386851,-0.0381314419,0.166278705,-0.1916953474,0.1113921329,-0.2765848637,-0.0130323609,-0.1457217634,-0.0526321158,-0.0306963269,0.216698125,0.1879994869,0.0291851554,-0.0721330196,0.0595457926,0.2788746357,0.2985241711,0.1559013724,0.1172475964,-0.2246597409,0.0273785274,-0.0366569459,0.00244577,0.2337848097,-0.4035803378,-0.1247583181,0.1427869946,-0.1231002286,0.1376376897,0.2947228253,0.1800485849,0.3189501762,0.2410545051,-0.0431585833,0.4741290808,0.2841287851,-0.0888023451,0.073186405,-0.0395187065,0.2465720922,0.3004812002,0.0893413648,0.1786063462,0.2016614676,-0.0689590946,0.0651522204,-0.1590226442,0.2348520756,-0.1180081218,-0.0782875642,-0.1455991417,-0.1292728931,0.0702390149,-0.1378353238,-0.0682510808,0.1008945331,0.3550207019,0.1246390864,-0.056308914,0.1041949987,0.7412886024,0.3876671195,0.160544917,0.4062668085,0.0733299777,0.0268041044,-0.2988274097,0.2523026466,0.1610683352,-0.4967062473,0.5382865667,0.0496946275,0.2061890364,0.0334009193,-0.0575258844,0.2096959054,0.3144170344,-0.2221792191,-0.0278006662,-0.6374946833,-0.600946188,-0.1810956895,-0.1889698505,-0.2641664147,-0.4526263177,-0.1407168359,0.2260042727,-0.1081146449,0.0873988718,-0.218007341,0.0166965518,-0.3855964839,-0.3432736397,-0.2249444276,0.2820158601,-0.1479971409,0.0275822561,0.2968629599,0.0026488476,0.3359049261,0.0013950104,0.0883508176,-0.5908955932,-0.0373855606,0.2585214376,-0.0719510689,-0.0291324556,0.151628077,-0.0415065326,0.3573758006,-0.088770628,0.2683821619,-0.094729282,-0.1510519981,-0.1963443905,0.1135050207,-0.0697748587,-0.1697794497,-0.1321094483,-0.3559782505,-0.5992090702,0.2074672729,0.0885049328,0.1327043623,-0.0062963073,-0.3061015606,0.2599484622,-0.1653517187,-0.0363086089,-0.0331934504,-0.0868117213,0.3976346254,-0.3162903786,-0.4053558111,0.1422792077,0.1079493091,-0.1221241355,0.2139302045,-0.3075752556,-0.2922318578,-0.1043406725,0.263728112,-0.1654445529,0.3404821754,0.3213359714,-0.1051378548,-0.0344130695,-0.0144899851,0.0042766188,0.0218883287,0.5916693807,0.3955231011,0.144984588,0.2034229934,-0.013806208,0.6329036355,-0.0128569389,0.3049511313,0.6129640341,-0.0857623369,0.1866588295,0.114027299,-0.1501547843,0.1227124929,-0.1731449813,-0.0008509151,0.3054537773,-0.0267824493,0.0470192134,-0.3564161658,0.0124452878,-0.1616762429,-0.175230667,0.0791460425,0.0412768461,0.08387205,0.1225029081,-0.0530436039,-0.0046052267,-0.1046294123,0.220670253,0.588650763,0.114394024,0.1677813083,0.5067594647,0.2036758065,-0.5534311533,0.1838194579,0.3727284372,0.1865550131,-0.0981064364,-0.1176688299,0.2423761487,0.039303761,0.7532058954,-0.1371991336,0.0951686502,-0.0227054413,-0.2773636878,0.0982569829,-0.0846483633,-0.2140581459,-0.125043124,0.0207217783,-0.1770860404,-0.0846206546,-0.3458729386,0.0855676979,0.0121726021,-0.0344856828,-0.1708544791,-0.0390406549,-0.191938594,-0.3701970279,-0.2853296399,-0.0260415375,0.0063993684,0.4199696183,0.1443029791,0.0727236718,0.0505155772,0.1016796529,-0.1505739242,-0.0936838761,-0.0648625642,-0.0634383261,0.1325102597,0.0797572136,0.1074750051,0.3926929832,-0.1583863795,-0.2790734768,0.0104631456,0.018178273,0.3472702503,0.3987167478,-0.0676064566,-0.0752261132,-0.3433760703,0.2555634081,-0.3039541543,0.1452272087,0.4953189492,-0.2727416456,-0.072167553,-0.2200797796,0.4214267731,-0.1367313564,-0.1279631406,0.049063798,0.2962502837,-0.094549939,-0.1387538612,-0.2388079017,0.9049050808,0.1610273719,0.2585718632,0.287981987,-0.2937954962,0.4278529286,-0.3093656898,0.1490637809,-0.0305961818,0.0727648064,-0.331931442,-0.044294931,-0.0298077594,0.0314101093,-0.1724551022,0.3442900479,0.0497858711,0.1140828654,0.2147442251,0.4392111003,0.0458884165,-0.2011903822,0.0374383181,0.1256831139,-0.2235944122,0.5764859915,-0.1773506254,-0.2568201125,-0.1791772097,-0.1348347366,-0.340405792,0.0957404152,-0.3617017269,0.1418791711,-0.0273757037,-0.3045834601,0.0767923668,-0.1332893968,-0.4745095074,-0.305487752,-0.2130950689,-0.1318248212,-0.2448129356,-0.0724516585,-0.0509195141,0.1327765882,0.2552112043,-0.2429924458,0.0630495325,-0.154395476,-0.1461341083,-0.2880741954,0.3319582641,0.1827702671,-0.0111748707,0.0390755646,-0.2839114368,-0.0404372588,-0.1150360703,-0.1783135533,0.0523862354,-0.0619813241,-0.279843688,-0.0301723666,-0.0061628465,-0.1752230823,-0.1325216889,0.4075365067,-0.4034814835,-0.0355725549,0.1406916529,0.0237755477,-0.2962320149,-0.0895974413,0.0937840566,-0.331438005,-0.2941505611,0.172530666,-0.0792762563,-0.1536644548,-0.0614638105,0.6250385642,0.4542081356,-0.0486427173,-0.3735595942,-0.3518791199,-0.0759581402,-0.1955539286,-0.0323210023,-0.0543921888,-0.4404830337,0.2066711485,0.368796289,-0.1951183677,-0.2043240815,0.0681994408,0.0282806624,-0.0735630989,0.5844718218,0.1414512247,0.3461655974,0.0799479038,-0.0698176026,0.0713339746,-0.3271226287,-0.0190804712,-0.270571053,0.2029909194,0.2286142409,-0.206267342,0.1012567431,-0.1428261101,-0.1720265001,0.1476672292,-0.0604830906,0.0595105402,-0.0511135682,0.0235313233,0.3245332241,-0.1672866046,-0.1156062931,0.4272964299,0.2004687935,0.6626449823,-0.1361947358,-0.0056374166,-0.185244292,0.0673514903,0.0413904302,-0.004974606,0.0635146424,-0.1287768185,0.4958369732,-0.0079004699,0.130793035,0.3999163806,0.4293607175,0.3676395416,0.2517672777,-0.0209887475,0.4518892765,0.0067912363,-0.1872095615,-0.2510673404,0.1127900928,0.2692395151,0.0050619184,0.2372924834,0.1238364279,-0.0276753604,0.0061598318,-0.0874172971,0.3391385078,0.1967254579,-0.146126315,0.3032358289,-0.0212113019,0.3283675015,0.1042569354,0.2872180045,0.2324739546,0.0244450253,-0.2458744049,0.0306977928,-0.0026337786,0.1167892888,0.13131468,-0.3960454166,0.1197963506,-0.0447750501,0.1062580496,0.0339801013,0.1775739789,0.4260559082,-0.4361589849,-0.0607525483,-0.1934913397,0.3412932754,0.0521253422,0.2550957799,0.2244356126,-0.0461192429,-0.5083350539,0.0352218524,0.0010543631,-0.1477964371,0.0082437927,0.1526286453,0.1336377114,-0.1495599598,-0.2861795723,-0.4118020535,0.0226510279,-0.3924216628,-0.0291099995,-0.1426413953,-0.2209938616,0.0700139552,0.1762846261,0.3575519621,0.2120589167,-0.0406742431,-0.2111977935,-0.3224902451,0.1345594376,-0.1515351832,0.3563965857,0.2794629037,0.1905568093,0.3203724623,-0.0804375932,-0.1068699658,0.0160681661,-0.0057101515,-0.3274736702,-0.0986219794,0.8825447559,0.1085711345,0.0574115291,-0.1607702225,-0.1185572445,-0.3888339996,0.35741359,0.0243344158,-0.2475163639,0.0983301476,-0.0492089167,0.0190822762,-0.0445245877,0.3944630027,0.4977112412,-0.0882662758,-0.1012614816,-0.073006019,-0.955273509,0.2226015925,0.1122501567,-0.1343935132,-0.0110493982,-0.0381552652,0.0013503941,0.3666800261,0.0849564746,0.2992457151,0.138347894,-0.2450374514,-0.0024799155,-0.1338802874,-0.1508325487,0.0156509876,-0.146272555,0.0034540321,0.1538896412,-0.4318594337,-0.0710960031,-0.1886007339,-0.2279030681,-0.2627551556,-0.1107052267,0.5203107595,0.1448254287,0.2329773903,-0.0968302563,-0.1354693174,-0.1302615106,0.2230708748,-0.3745639622,0.1055758968,-0.3027060926,0.015027591,-0.2752325833,0.4236091673,-0.6369168758,0.1009457931,-0.0892453268,0.4044830799,-0.4081545174,-0.0905559957,-0.4338996112,-0.1059968919,0.5274696946,0.2321295291,-0.0608138479,0.0233165063,-0.4625007808,-0.2424898148,0.3503893316,0.0346450657,-0.1024282128,-0.0361547768,0.1269068867,-0.0896053687,-0.168469429,-0.3031354547,-0.3479736149,0.1281346381,-0.083448723,-0.1291950941,-0.0283843242,-0.3198001087,-0.2465521246,-0.0224162862,0.3546431065,0.1518904716,-0.1414553374,0.1038638353,-0.0098321522]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"Awesome ! I'll take a look :)","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":7,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\nAwesome ! I'll take a look :)","embeddings":[-0.2323101014,-0.0522186235,-0.007700738,0.270209223,-0.0739417449,0.0673958436,-0.3258600533,0.4886372387,0.2403178513,0.0261085015,0.0047331825,0.0491164774,-0.1206184328,0.0258979313,0.2157631963,-0.2895117998,-0.1317707151,0.0758426785,0.1449715197,0.1285266429,-0.3687725067,0.2526968122,0.0979527608,0.0030220232,-0.0357273705,-0.0106627848,-0.0539005361,0.1975093931,-0.3309296966,-0.4783369303,0.0846154094,0.0830737874,0.4349913597,0.2375996262,-0.0001199804,-0.0423170403,0.3016745746,-0.0614725426,-0.1504947394,-0.4398259223,-0.5248809457,-0.421546787,0.1821025014,-0.0120052956,0.1363058388,-0.2649262547,-0.0794526786,0.2011477351,0.0850372761,0.4670846462,0.1787825227,0.288256675,0.0442712642,-0.2111332864,0.4216047227,-0.0171500891,-0.0451693684,0.2298566103,0.2866631448,-0.1339090019,0.1286519766,-0.316221863,-0.1232239306,0.2418138087,0.2954734564,0.1833904684,0.1795920581,-0.4384172261,-0.3550662696,0.2028669715,0.6560674906,0.0796597749,-0.2338946015,-0.0384366997,-0.1062551737,-0.0435800403,0.1859125942,0.0106265312,-0.367403537,0.1872695982,0.1668387949,-0.3132669032,-0.2391114235,0.3601313829,-0.1921871305,0.5815842748,0.0598433167,-0.1615957469,0.2679896057,0.268009454,-0.2352554947,-0.1112708673,0.0955639854,0.0555746444,0.3084036112,-0.0907865912,0.0957749933,-0.3303886354,-0.0245710965,0.3253615499,0.0885893777,-0.1108830199,-0.1574549824,0.272888869,0.0597683117,0.3095929921,-0.1660335213,-0.0379955433,0.454120189,0.2523880601,0.2839609981,0.0283919573,-0.1289796829,-0.1289862394,-0.4581525028,0.0608980954,0.2114820629,-0.1608365476,-0.1390478611,0.0565187857,-0.1972845942,-0.1454484612,-0.0611432567,0.1528953612,-0.1504372954,0.1851055324,0.058769159,0.2179650962,0.0919819623,-0.3163460195,-0.0012078716,-0.0218626633,-0.0148911783,-0.1716385037,0.1608157754,0.2000778168,-0.0408578776,-0.1884183735,0.063421607,0.1054395959,-0.0369242281,0.0170955639,0.1618222743,0.4071896672,0.0059996899,0.1881854236,0.0600562543,-0.1991619617,-0.1849802732,0.3172505498,-0.0459095612,-0.4196074009,-0.4449992478,0.0985748619,-0.3218239546,-0.4770755768,-0.2233060151,0.1810412258,-0.3135973215,-0.2631994486,-0.1270371079,0.1419162303,-0.1918174177,-0.1787087768,0.1242135018,0.371696353,-0.4418006241,0.0315090008,-0.3087265491,-0.266253233,0.4835541844,0.1880432814,-0.1360826939,0.345410198,-0.3295161128,0.3978422582,0.8266927004,-0.3348728716,-0.265930146,0.6340323687,-0.3567996919,0.0749033988,0.2040862143,-0.067672357,0.3194113076,-0.2655497193,-0.5724760294,0.5304371715,-0.0179234836,0.0237028748,-0.2779535055,-0.1761804968,0.3120366633,0.0047696638,0.2916070521,0.3320195973,0.4019729793,0.001360253,0.3842509985,0.3598610163,0.2111338973,0.10855221,0.3266764283,0.2945000827,-0.1101353541,-0.2515132725,-0.2113886178,0.081030421,0.024237141,0.1963675618,-0.1477672458,0.1370948404,-0.206694901,-0.0301483348,-0.1581299007,0.0073475158,-0.0367704593,0.1264226139,0.1666442752,0.1145016253,-0.0772116035,-0.0012543467,0.3100516498,0.306964159,0.1786165982,0.1703889668,-0.2259678096,0.0450895652,0.0175617319,0.0087145483,0.2557226717,-0.3684763014,-0.1210912988,0.101949431,-0.1318722218,0.0522053167,0.3705914617,0.2386092991,0.3024059534,0.1381682754,0.007253177,0.4942449927,0.2126208395,-0.0774280727,0.0471975356,0.0311504044,0.3347287774,0.2859232128,0.0341412425,0.2607295215,0.1912377477,-0.0836199373,0.0984601974,-0.1145061031,0.2232210636,-0.0471828766,-0.2856550217,-0.2089126259,-0.1502177417,0.0780643895,-0.136467129,-0.084148638,0.1606751531,0.3647230864,0.191279158,-0.0010774946,0.1218683422,0.7607136369,0.2721985281,0.1874955297,0.2952882349,0.0765018016,0.0513352975,-0.2567276657,0.254937619,0.1485017687,-0.3966371417,0.4738033712,0.0204718839,0.1821963042,-0.0022478215,-0.0878891498,0.1793987304,0.3458877206,-0.1910804063,-0.0361176915,-0.6302676797,-0.6095279455,-0.1457116455,-0.1752958447,-0.3059278131,-0.4953168035,-0.0939492285,0.2876924872,-0.0820775628,0.080241099,-0.2544182837,0.0726488754,-0.332279563,-0.4052580595,-0.1888598204,0.1786257029,-0.174638465,0.0340075493,0.3451590836,-0.0719688535,0.3641274869,0.0156590734,0.1630017608,-0.5805225372,-0.0341098383,0.2214657813,-0.088445425,-0.0963504314,0.1044033468,0.0479490496,0.3728501499,-0.057602454,0.2575690746,-0.1228814349,-0.1119804382,-0.1512124538,0.126661405,-0.0717323646,-0.1539563239,-0.1748428941,-0.4068228602,-0.5905546546,0.1086899936,0.1271524876,0.1227463633,0.0624490678,-0.2385023385,0.2134639323,-0.1629313678,-0.1435995847,0.0121043306,-0.1037436053,0.4455562532,-0.3121083081,-0.4407377541,0.2149594575,0.1073974967,-0.2023888379,0.0721579418,-0.3543744385,-0.3051295578,-0.072785601,0.2624230683,-0.1276576817,0.2905939817,0.2897391617,-0.0863473937,-0.0168699156,0.0618931428,-0.0243935101,-0.0005722956,0.5442129374,0.4199533463,0.1826222241,0.2969355583,0.0163720157,0.6336371303,-0.0423585996,0.3230960369,0.6264436841,-0.0973656774,0.1565067619,0.1170973405,-0.2085177451,0.0857888535,-0.1211948767,0.0133069791,0.2947114706,0.0125734825,-0.056375023,-0.3769802749,-0.0203673579,-0.1402379423,-0.1981895268,0.0357752666,-0.0019030438,0.0420269296,0.1089105308,-0.0800641179,-0.0180377644,-0.2021165341,0.247672379,0.5596757531,0.0652188212,0.1657548547,0.4549008012,0.2708334625,-0.5822426677,0.2195927054,0.3171445727,0.2178468555,-0.0899870247,-0.1110158637,0.173497811,0.0662167966,0.7055230141,-0.1372469366,0.0714982003,-0.0071141096,-0.2771007419,0.1491194069,-0.1258572638,-0.1508076787,-0.0758457184,-0.0548788793,-0.1670057178,-0.0248848442,-0.3125512004,0.1889318824,-0.0073673027,-0.1046995074,-0.2031930536,-0.0625448897,-0.2073055655,-0.3747368157,-0.276414603,0.0084983818,0.0465852879,0.3519515395,0.1400182098,0.1495913118,0.0082349377,0.105021596,-0.1345347911,-0.0881252512,0.0081927013,-0.0791939273,0.1413256079,0.0445072763,0.0557330064,0.3736692071,-0.1556573957,-0.3091413975,-0.0282895304,-0.0110604959,0.4624546766,0.3269808888,-0.0888105854,-0.1383894831,-0.280093044,0.2719325721,-0.3127980232,0.0486420095,0.4944605827,-0.287388891,-0.1275331378,-0.2351669371,0.4512132108,-0.1115192175,-0.0671577007,0.1349679232,0.1733578295,-0.0740333349,-0.0685102046,-0.2375212759,0.9377808571,0.0580414496,0.256670177,0.2678645551,-0.2649666369,0.3609484434,-0.3413946331,0.1959850639,-0.1357772499,0.1650269926,-0.2951318026,-0.0098424247,-0.0482832566,0.0298023373,-0.1373028457,0.337772131,-0.0297982339,0.059063483,0.1983510554,0.4265333116,0.1065265909,-0.1865420938,0.016135335,0.1449547857,-0.2132726759,0.481584996,-0.2049054205,-0.3102780282,-0.1348468065,-0.190675959,-0.348968327,0.1109610349,-0.3399208486,0.1432354599,-0.0542324036,-0.2990573049,0.0343605652,-0.1033728048,-0.4695111811,-0.2201677412,-0.2046877891,-0.1978290826,-0.25750646,-0.085170649,-0.0384584367,0.1663539261,0.1779996306,-0.2365094423,0.0419561006,-0.1462488323,-0.0405577607,-0.3945172429,0.3203729689,0.18020989,0.0504260547,0.095167689,-0.1821344942,-0.0331469961,-0.1009063423,-0.1971072257,0.0710965097,0.0191533342,-0.2945888638,-0.0491449609,-0.0315621682,-0.1207503229,-0.1241130456,0.3860171735,-0.4125863016,-0.0052474104,0.1505108625,0.1877989918,-0.2914903164,-0.1343897581,0.2103460133,-0.3033867776,-0.2543422282,0.2223829329,-0.1115188375,-0.1667922139,-0.0146605903,0.6656320691,0.3933482468,-0.1540711522,-0.260773927,-0.4490229785,-0.0700142458,-0.1726222485,-0.164131999,-0.0202804897,-0.4149360359,0.1708543152,0.3341239691,-0.2818113863,-0.215427354,0.0690346286,0.0233124718,-0.1131454632,0.526088655,0.1449679136,0.3692310154,0.1002573892,-0.058022771,0.0132969879,-0.2751323283,-0.0357258692,-0.2690887153,0.203623578,0.2234343886,-0.1856695712,0.1319426447,-0.1784322709,-0.2557868958,0.2268766165,-0.1452896595,0.075754419,-0.0410268195,0.0910200328,0.3216641545,-0.2734777629,-0.0384337604,0.3752794266,0.2164659947,0.708438158,-0.1393766999,0.0330246277,-0.1075611562,0.039334394,0.0995219275,-0.0170519929,0.0235843156,-0.1738117039,0.5475404859,-0.0660149828,0.1212156639,0.4624367654,0.4549086094,0.3463176489,0.2295672745,0.0658955649,0.4528761804,0.0315681212,-0.2494986355,-0.2837159038,0.2023290694,0.2518798709,-0.0667679757,0.2430782467,0.0543562211,-0.0405488685,-0.0235288516,-0.0747297257,0.3235633075,0.1616298705,-0.0410937332,0.2495569736,-0.0756332353,0.265334785,0.1085726619,0.3157592416,0.207068637,0.0600905642,-0.1767178178,-0.0036719344,0.0733807907,0.0304513089,0.1207732707,-0.4290504754,0.0949747488,-0.1390285939,0.1841121167,0.0307648201,0.13647829,0.4525681734,-0.3737442493,-0.0856005475,-0.2555031478,0.3492827415,0.0573923215,0.320907861,0.2318922132,-0.0216507744,-0.4817880988,0.0300166886,0.0683305413,-0.2199382037,-0.0444778949,0.2219981998,0.1696274132,-0.1470555663,-0.2900291681,-0.4080080688,0.0116184643,-0.4206080735,-0.0846557245,-0.1482022107,-0.2454571873,0.0370136388,0.2253718227,0.3424113393,0.2294157296,-0.1090273857,-0.1780029982,-0.4201919734,0.1284442097,-0.1133411527,0.3553616107,0.2969467342,0.2547872365,0.2975407839,-0.0766934231,-0.0739527345,0.1028153673,0.0043639187,-0.2936388254,-0.0119841713,0.8866648674,0.1815786809,0.0499610491,-0.221670866,-0.1543268263,-0.3260795176,0.2771731019,0.1211373359,-0.2143840194,0.0922301188,-0.1071480885,0.0184428245,-0.0133842854,0.3756723106,0.5387010574,-0.0299662277,-0.1477918029,-0.0379768088,-0.86496526,0.1638650745,0.0563824028,-0.0826989487,-0.0218810327,-0.0166995507,0.0569063723,0.4327149391,0.0975401923,0.2717875242,0.1420949697,-0.2102885395,-0.073967129,-0.0802449286,-0.0888027102,0.09229628,-0.1218988895,-0.0145118879,0.1678975224,-0.4554167092,-0.0582717136,-0.2435230911,-0.2710791528,-0.2796961069,-0.0773483068,0.5414490104,0.1081371084,0.2367742956,0.0103609916,-0.1080522388,-0.1547997594,0.2393555045,-0.3836968541,0.073734194,-0.248927623,-0.0139210066,-0.2949145436,0.4077955782,-0.6546960473,0.0210588071,-0.179632619,0.3861242235,-0.4314543605,-0.1367823184,-0.3671122491,-0.1105694994,0.4848162532,0.1734354943,-0.094808422,0.0456906632,-0.4706010222,-0.2095149159,0.3010290563,-0.0681697056,-0.1370650083,-0.0538238771,0.1517536193,-0.1034578979,-0.1024919897,-0.2440225631,-0.2954359949,0.1046513915,-0.0429268479,-0.1079491004,-0.0133445384,-0.2723070383,-0.2435987145,-0.002500975,0.2506129146,0.0840245336,-0.1545374542,-0.0079908418,0.0190129392]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"@AmitMY Can you now download the Phoenix2014 Dataset?","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":8,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\n@AmitMY Can you now download the Phoenix2014 Dataset?","embeddings":[-0.2729876041,-0.121160455,-0.0146644348,0.3014095724,-0.1079125181,0.1247347742,-0.3906079233,0.5260216594,0.1725602448,0.0058719809,0.0328398049,0.0677741393,-0.1204458624,0.0785517544,0.1591174603,-0.2751715183,-0.1589929014,0.0370276794,0.1541847736,0.1182602644,-0.3644980788,0.2508248091,0.0949482545,0.0418288559,0.0232647136,-0.0667340308,-0.035162624,0.14474608,-0.2999583483,-0.4333535433,0.0909486115,0.0641556755,0.4221330881,0.2631332576,-0.0001202675,-0.0266765915,0.2670078278,-0.0622318164,-0.1010396108,-0.452952981,-0.5797411799,-0.4428085983,0.1601156592,-0.0147779332,0.1252076775,-0.4274076521,-0.0819816217,0.2454808056,0.1802282631,0.5245463848,0.1700854301,0.2693458796,0.1126718968,-0.2105449438,0.2768781185,-0.065274775,-0.0387466662,0.2291698307,0.2163773179,-0.0682078525,0.0850204229,-0.3280288875,-0.0515193753,0.2132132947,0.2415710539,0.1459519714,0.2097450197,-0.4479530156,-0.2789220214,0.1582863331,0.7371519804,0.0724490806,-0.248405695,0.0274231005,-0.0954511464,0.0514095947,0.1537825018,0.0298709441,-0.3552879989,0.1846993268,0.1256214529,-0.3136164248,-0.2932279408,0.3587517142,-0.2015419304,0.580247283,0.0805571675,-0.1313536316,0.2098293006,0.2340998948,-0.1249889061,-0.0796662346,0.1231010705,0.1114751324,0.2863289416,-0.1130683646,0.0320711806,-0.336925447,-0.0634365603,0.3482524753,0.119962208,-0.0519705452,-0.20236139,0.2075619251,0.0313657299,0.2873896658,-0.2012322247,-0.0750032514,0.483661443,0.2284152508,0.3003913164,0.0401046798,-0.1498629153,-0.1126357689,-0.516656518,0.0618682578,0.1997544914,-0.1582774371,-0.1259271801,0.0564944111,-0.2179156244,-0.1501775682,-0.1112014204,0.1822033077,-0.160253495,0.2125310451,0.1048063189,0.1673953086,0.1267129779,-0.3416509032,-0.0013488753,-0.0313572586,-0.0064080949,-0.1590428501,0.1458460093,0.2707543075,-0.0171359517,-0.1208417118,0.0763477236,0.1392547935,0.034758836,0.0082598673,0.1322934628,0.4340328276,0.0426393375,0.2170617431,0.0564584583,-0.0531987809,-0.1602071822,0.3079189956,-0.0338113159,-0.3681851625,-0.3663426936,0.0849163681,-0.2923173308,-0.4707967341,-0.2747641206,0.1351107955,-0.3350332677,-0.1954900175,-0.1703357399,0.1040570661,-0.1960649937,-0.2253264487,0.116665557,0.3613219559,-0.4485612512,0.0420237891,-0.3433261514,-0.2684932053,0.433547914,0.1675172001,-0.1673358083,0.4218984842,-0.3451114297,0.3251282275,0.7254835367,-0.41470927,-0.2290217429,0.6577929258,-0.388859719,0.0293821543,0.1746576577,-0.0850054547,0.265429318,-0.181802243,-0.600851953,0.4793032706,0.0001939554,0.0079911808,-0.2402609438,-0.1558217108,0.3215585947,-0.0265491437,0.4053131342,0.3193722665,0.3584243655,-0.0089872209,0.4057266414,0.400777638,0.1981233954,0.1791159958,0.3663581014,0.2983915806,-0.0701222196,-0.1530264914,-0.174318105,0.1051786914,-0.0197607372,0.1888488084,-0.1204703078,0.2139324099,-0.2334628105,-0.0353509709,-0.1449323446,0.0448279567,-0.0436263382,0.081652455,0.1361609995,0.0571517497,-0.0544677339,0.1394223273,0.260053426,0.3442430198,0.1357770711,0.1589937806,-0.241996184,0.1036544144,0.0355154835,-0.0525291823,0.2644432485,-0.3973349333,-0.123803407,0.1003826261,-0.1142973974,0.0461548716,0.4365489483,0.2337718308,0.3273056149,0.0735014975,0.018048523,0.6015809774,0.2127688378,-0.086383462,0.0803783089,0.0640379265,0.3520664871,0.2351408452,0.0877009407,0.2479202449,0.2110846788,-0.0850770026,0.1058674753,-0.0994204804,0.1831993759,-0.1273368001,-0.2642623484,-0.26141572,-0.1796244085,0.0396056511,-0.2451809198,-0.0819443241,0.1301120967,0.3201668262,0.1818225086,0.0112556573,0.1691032946,0.7405772209,0.2752283812,0.184969008,0.2032272816,0.0601123199,0.0223659631,-0.2468018979,0.2231581658,0.1680711955,-0.4660722613,0.4722954333,-0.0098764068,0.1571714729,0.0251320731,-0.1079090983,0.2505765259,0.3372763395,-0.1950346529,-0.114720434,-0.6178670526,-0.6137973666,-0.177519545,-0.1595990211,-0.2571884096,-0.5025901794,-0.1125973538,0.2553154528,-0.0221549645,0.0824857578,-0.3406073153,-0.0157515761,-0.2803775966,-0.3749525547,-0.2095185369,0.221472621,-0.1683911234,0.0393543653,0.2671877742,-0.0207419619,0.3431386352,0.0164155234,0.1853420883,-0.6218441129,-0.0607331209,0.2398566157,-0.1127152368,-0.141832307,0.0436262824,0.1516503692,0.3746292889,-0.0767496675,0.2472211868,-0.1594891548,-0.1488558948,-0.2075939476,0.0846021026,-0.1313866973,-0.1132614166,-0.2578608394,-0.3839243054,-0.6021553874,0.0862217322,0.1499544084,0.0871223137,0.1439118534,-0.2540988922,0.2184660435,-0.1183515638,-0.1890586168,0.0049049328,0.0216433238,0.4551009536,-0.2924210727,-0.4757568836,0.3094584346,0.0788547844,-0.1271689087,0.1186314449,-0.3245375454,-0.2510761619,-0.0579116121,0.299800545,-0.1370936781,0.2469025999,0.2431820482,-0.1063542739,-0.0103679579,0.0781777129,0.0360161029,0.0258312747,0.4987714589,0.4349073172,0.1815470755,0.2763704658,-0.0165461916,0.6880322099,-0.0589644611,0.319414109,0.5639925599,-0.1137170047,0.1197838187,0.1605960578,-0.2317630649,0.0441464484,-0.0905136541,0.0560472421,0.2956251502,-0.0184949562,0.012885619,-0.3436952233,-0.0201386027,-0.154248938,-0.1895780861,0.0176399052,0.082046926,0.135411039,0.1197650954,-0.0756633133,-0.0361142978,-0.2214507312,0.3127985299,0.5210315585,0.0571315438,0.1689489484,0.3742898703,0.2641096413,-0.5434142351,0.2335715592,0.2768685222,0.3034380674,-0.0891269967,-0.0596626624,0.1516725868,0.1023599729,0.7210991979,-0.1091488078,0.0362288915,0.0048252447,-0.3374864459,0.1149931252,-0.1588766426,-0.1626011282,-0.0974450931,-0.0808466747,-0.0527287647,0.0189823657,-0.3054806292,0.1479519457,-0.0104530677,-0.1740223467,-0.1326397657,-0.0576261729,-0.2535949349,-0.3287672102,-0.2671148181,-0.012045553,0.0572728626,0.3296545744,0.1172263175,0.1582537293,-0.0281034261,0.1203735173,-0.1252276301,-0.037925642,-0.0146760996,-0.0358159319,0.2308700234,-0.0811178461,0.0836933032,0.2914409935,-0.1222698018,-0.2990047336,-0.0553752705,-0.0119789355,0.4326098859,0.3277004659,-0.1416583806,-0.1336937696,-0.2784712911,0.2247039229,-0.2623413801,0.0621279404,0.4702568352,-0.2851668298,-0.1684988439,-0.3053057492,0.4727476239,-0.0633067861,-0.0962888151,0.1470700949,0.1922398955,-0.0726158842,-0.0283530653,-0.2226088643,0.9253940582,0.0006219145,0.239593327,0.3015554845,-0.1495206654,0.2313610911,-0.4104020894,0.1956088692,-0.1698304713,0.1093025133,-0.3239030242,-0.0360896699,0.0134825427,0.0606245846,-0.099322021,0.3914314508,0.0025228083,0.1010187343,0.2412925214,0.3989268839,0.1077072695,-0.127554521,0.0696762353,0.1434963644,-0.2096883655,0.4521377683,-0.2083146125,-0.3111769259,-0.213216424,-0.2053733021,-0.3493562043,0.0959601626,-0.3789912164,0.2317427248,-0.0494100302,-0.2501916885,0.0373126343,-0.0652401075,-0.4683618546,-0.2049657404,-0.2142729908,-0.1728343219,-0.2795264125,-0.1806669682,-0.0733659491,0.2548052371,0.1379241645,-0.2332793921,0.0295738988,-0.087411128,-0.0753938109,-0.3793374002,0.2414846271,0.22015661,-0.0151133388,0.0703068823,-0.2330009788,-0.0599167608,-0.1364499032,-0.174561426,0.0590272397,0.0446616448,-0.2648908496,-0.0488293171,0.0360850468,-0.110570468,-0.0944533199,0.3712663054,-0.3449903131,0.0334360674,0.163670212,0.1493191123,-0.2496489882,-0.1262749583,0.1904715747,-0.2932347357,-0.1871403605,0.2193350494,-0.119507283,-0.1595657319,-0.0067003621,0.647742331,0.395906508,-0.1678339392,-0.2777324021,-0.4629293084,-0.0938885212,-0.1257422566,-0.1827734262,-0.0359730721,-0.3634029031,0.1754759401,0.3819897175,-0.331168741,-0.1985999793,0.0524301976,0.0106878383,-0.120566912,0.5426334739,0.1465328336,0.3576272726,0.1086517796,-0.0682146549,-0.0058810813,-0.2707851827,-0.0155655555,-0.2745249569,0.2112624049,0.2419572175,-0.1581131965,0.1384169757,-0.179476738,-0.2593320906,0.2399389744,-0.18505615,0.1252718121,-0.0161957033,0.1062942222,0.3337551355,-0.301935792,-0.1277133524,0.4062066674,0.2887072265,0.6514325142,-0.1028107256,0.0723117515,-0.1247851178,0.0460376963,0.0447835289,-0.0392298847,0.0457028188,-0.2047600001,0.5358090401,-0.0350688472,0.1376501173,0.43967852,0.4230821729,0.3530572057,0.2129304409,0.1364121437,0.3983596563,-0.0010099364,-0.3090783358,-0.3052457869,0.214076966,0.2089213878,-0.0823074654,0.2206403315,0.1025342196,-0.0964228585,-0.0179158729,-0.0795714557,0.3365235925,0.1307368129,-0.0493457578,0.3066780865,-0.0629749745,0.2555106878,0.1610785425,0.3184822202,0.1977926493,0.0837193951,-0.1517321467,0.0074438178,0.0919873342,0.0101718754,0.1563773155,-0.5059672594,0.029490618,-0.1238052174,0.1603077501,0.0392039418,0.182775408,0.4297809601,-0.3041255772,-0.1120025665,-0.2705145478,0.3142858446,0.0414867736,0.3501625657,0.3118754327,-0.0145279374,-0.5376652479,0.0140974419,0.1034462824,-0.2846463621,-0.0433056913,0.2061949223,0.1360609829,-0.1247794479,-0.245677501,-0.3682178855,-0.0894447938,-0.4216371775,-0.0930768624,-0.2030991763,-0.2526587546,0.0094547467,0.1671324074,0.2918306589,0.1727928072,-0.048460532,-0.1233839467,-0.3529158235,0.1164634824,-0.074503459,0.3798488379,0.3188082874,0.2932070196,0.3110564053,-0.0722426176,-0.0546796583,0.0514513776,0.0415739082,-0.2189872116,-0.0162614472,0.9244276881,0.1550083011,-0.0076674791,-0.2141106874,-0.1626442373,-0.2887750864,0.3796088099,0.1407541037,-0.1880159378,0.0479983427,-0.0975645557,0.0115659321,-0.0329439864,0.3414574862,0.5806118846,0.0109108118,-0.1533287764,-0.0650638491,-0.8778639436,0.1630043387,0.0442373864,-0.0038644657,-0.0154343629,-0.0504602566,0.094907701,0.4220899343,0.1475211978,0.1593259126,0.1363637447,-0.1571579278,-0.0653819889,-0.0643485934,-0.0424108803,0.1424551755,-0.11601495,-0.0007690141,0.1629443169,-0.3651428223,-0.0491002165,-0.2216645032,-0.3413749337,-0.2257584333,-0.2178337872,0.5345090032,0.042726323,0.3024739921,0.0546501353,-0.0683681071,-0.2081798911,0.2137438655,-0.3635400534,0.0761857331,-0.1749850214,-0.0227399468,-0.2470638603,0.3912584186,-0.6319795847,0.0416816063,-0.1613832712,0.30707708,-0.4761056602,-0.0989890471,-0.4565140605,-0.1278617829,0.4466244578,0.190320909,-0.1071146727,0.0239447113,-0.4305598438,-0.2083818167,0.2949058712,-0.2144971341,-0.1440057009,-0.051021494,0.1228268966,-0.2033042014,-0.1162307039,-0.2057293057,-0.2645466328,0.1072627231,-0.0636272728,-0.0753584281,-0.0804609284,-0.2199666649,-0.2440465093,-0.0166832507,0.336972177,0.1028690785,-0.1627583355,0.000097684,0.085802488]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"@hoanganhpham1006 yes.\r\nSee pull request https:\/\/github.com\/huggingface\/datasets\/pull\/722 , it has a loader for this dataset, mostly ready.\r\nThere's one issue that delays it being merged - https:\/\/github.com\/huggingface\/datasets\/issues\/741 - regarding memory consumption.","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":30,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\n@hoanganhpham1006 yes.\r\nSee pull request https:\/\/github.com\/huggingface\/datasets\/pull\/722 , it has a loader for this dataset, mostly ready.\r\nThere's one issue that delays it being merged - https:\/\/github.com\/huggingface\/datasets\/issues\/741 - regarding memory consumption.","embeddings":[-0.207817927,-0.1362151951,0.0223399419,0.368414551,-0.0426028669,0.0941679403,-0.4160577655,0.4587387145,0.246598646,-0.0219785664,-0.104866907,0.0549300723,-0.1739564091,0.0406175293,0.2238930464,-0.2853588164,-0.109582372,0.0470968671,0.0764123052,0.1486932337,-0.3671719432,0.3155606389,0.1266240329,-0.0520574041,-0.0608695634,0.0682501197,-0.1144394055,0.2525817156,-0.3424279988,-0.4832611382,0.1144609973,0.1468002796,0.3998353779,0.2815961838,-0.0001227519,-0.0169627946,0.3013392985,-0.0854624659,-0.205602482,-0.4361374676,-0.4879086912,-0.4381598532,0.204580158,0.0155038713,0.1532450169,-0.2656704485,-0.0692233518,0.1311786622,0.1514955461,0.4167751074,0.148070544,0.3967960179,0.0261271615,-0.2171597779,0.3806223869,0.0019360805,-0.0103933439,0.3392916918,0.2438274771,-0.1254053265,0.0567795187,-0.2961394489,-0.0569786802,0.2207807004,0.3716031015,0.133394137,0.0937364474,-0.4013227522,-0.3160227835,0.1644095778,0.5963615179,-0.0194317419,-0.2410590053,-0.1115393192,-0.1419954151,-0.0060443142,0.2234905064,-0.0008247739,-0.4154690504,0.1532354802,0.1061660573,-0.3735143542,-0.2633183599,0.333652854,-0.1387932599,0.5295559168,0.0314591974,-0.1647163182,0.3249212801,0.2675722837,-0.302785486,-0.1141962782,0.1184461117,0.074788712,0.2175052762,-0.1066058502,0.0439071506,-0.3347759545,0.0518928729,0.3083164096,0.1205500737,-0.0901618153,-0.1433764398,0.2116677016,0.1183404401,0.3491191566,-0.1774645597,-0.1086019799,0.384899199,0.2491233498,0.3148803413,0.0080584977,-0.1638400704,-0.0860047117,-0.5304250121,-0.0754911304,0.1751984954,-0.159505263,-0.171575442,0.0380557142,-0.1565726846,-0.1459298581,-0.0113843987,0.2451413125,-0.1559437513,0.2249639034,0.0938650295,0.1684688926,0.0178118125,-0.2756283581,-0.0035254352,-0.0146278683,-0.0804342479,-0.1210728586,0.1708238721,0.113258563,0.0096824868,-0.2057599723,0.1116544381,0.115702495,-0.0857889056,0.021078255,0.1708335727,0.4054081738,-0.0006323688,0.1629956961,0.0858401209,-0.1381656677,-0.1729673594,0.2869307697,-0.0179809872,-0.4325714707,-0.3796412051,0.0643770173,-0.354464978,-0.4458071887,-0.2828682363,0.219019264,-0.311722517,-0.2606828809,-0.1746931523,0.1834027022,-0.268594861,-0.1940467954,0.1424164921,0.4873999357,-0.3776033223,-0.0931542367,-0.2818639278,-0.2958835661,0.4628069699,0.243124485,-0.1338122785,0.3007458448,-0.3882220685,0.3272667825,0.7001002431,-0.347008884,-0.2477435619,0.6614431143,-0.3661957979,0.1240254492,0.227305904,-0.0370961502,0.2169185728,-0.2122607529,-0.4664954841,0.4712201953,-0.0262596216,0.0468968265,-0.2975867987,-0.2159231901,0.303047061,0.0292146616,0.2969394624,0.2930310369,0.3287850022,0.0340177156,0.3883121014,0.3318313062,0.2062091827,0.1013031751,0.3124403358,0.2461340278,-0.1173017249,-0.2410794497,-0.3073039651,0.1516397446,-0.048524417,0.2113685459,-0.116776526,0.1303155869,-0.2362406254,-0.0487335362,-0.1344463974,-0.0297987927,-0.0940584615,0.150647074,0.2303199023,0.0588666275,-0.1300015002,0.1075513363,0.1942079663,0.3531340063,0.057989005,0.2031669766,-0.1665961146,0.0836378112,0.0449276343,0.0117370198,0.1984714717,-0.3616105914,-0.1048683971,0.1805507243,-0.1400326341,0.1049061343,0.3677375913,0.1984399706,0.3461498916,0.1714134663,0.0351369902,0.430300802,0.2232325971,-0.1132473797,0.0082803071,0.0375885889,0.2606927454,0.3046484888,0.1443363875,0.1958293468,0.1838881373,-0.0570907667,0.0616536327,-0.1389453709,0.2580167055,-0.0735475942,-0.150655508,-0.154685542,-0.1694375575,0.0279493257,-0.1301021725,-0.0157377254,0.0681838617,0.4152185023,0.0894243643,-0.025441831,0.1282014549,0.7226035595,0.377538681,0.2367203683,0.2860601842,0.1250470132,0.0694644451,-0.2483469546,0.2356323004,0.1740339249,-0.4175153375,0.4896562397,0.0088774413,0.1388650239,-0.0630019829,-0.0210741013,0.2118410021,0.3253067136,-0.2296219617,-0.1021391749,-0.5599485636,-0.6452051401,-0.2360999286,-0.1311146617,-0.3341293931,-0.5082860589,-0.1180110127,0.2917005718,-0.0588609204,0.0850161538,-0.2439475209,0.1025520191,-0.3336437047,-0.4053463042,-0.2126259208,0.1854860038,-0.112926513,0.0057555819,0.384722054,-0.091664359,0.4335967898,-0.022649806,0.1536938548,-0.5495795608,-0.1093697846,0.2470646948,-0.0960324779,-0.1344262809,0.087641567,0.0862911567,0.3288471997,-0.076649785,0.286080569,-0.1717059612,-0.1637559533,-0.1800730526,0.1500083655,-0.0488474518,-0.154743284,-0.210155949,-0.3709218502,-0.6213084459,0.1715507805,0.1312792152,0.106965594,0.1341509819,-0.2162243426,0.2566540837,-0.1929634959,-0.1720329076,-0.0204367358,-0.1622503847,0.3975332081,-0.2929755151,-0.4353757799,0.1951984614,0.1112825125,-0.0704709291,0.1163327768,-0.3959720135,-0.242349878,-0.0722930208,0.2193657607,-0.1033337042,0.3142015636,0.3255229592,-0.0736394301,-0.0287296567,0.0735493973,-0.0482596606,0.0058852532,0.4612645209,0.397813201,0.2045585215,0.3464250863,0.0365738049,0.7163600326,0.047580786,0.3232117891,0.6261866093,-0.1124065295,0.2358408719,0.1207485348,-0.2355271578,0.0711995214,-0.1708737761,0.0622378439,0.3565026522,0.0827783048,0.0082883537,-0.4065494835,-0.0926927552,-0.1046226919,-0.1986020207,0.1033839732,0.0329179093,0.0569192842,0.0991131216,-0.1080402434,-0.0872318819,-0.2185586244,0.2813105285,0.5460278988,0.0595107786,0.1376953274,0.3997965455,0.3199513853,-0.6954449415,0.1886536628,0.2648611963,0.1873261482,-0.0747522637,-0.1366336197,0.1631584466,0.0525815524,0.722086966,0.0064717624,-0.0054477584,-0.049223993,-0.3526240587,0.031804055,-0.0241399873,-0.1140728891,-0.0388214327,0.0115564121,-0.0579373017,-0.1526669264,-0.3361531794,0.1889985502,0.0135110058,-0.1281348318,-0.2438368499,-0.0805294216,-0.1624545306,-0.4324138463,-0.2699868381,-0.0022198183,0.050934054,0.352484405,0.1653837413,0.1585516036,-0.0528848693,0.1838971674,-0.1535257995,-0.0394321047,0.0589002855,-0.0021917231,0.2227850705,0.0150483996,0.130135715,0.5010687709,-0.1754173487,-0.327858299,-0.0153752249,0.0642706007,0.473022759,0.3958739042,-0.065512605,-0.1401820779,-0.2006431073,0.2754945755,-0.3726553619,0.1158778444,0.4451705515,-0.2512331903,-0.1821686476,-0.2667693198,0.4942343831,-0.06494046,-0.0335146189,0.1643356234,0.2488303185,-0.1114163548,-0.0914459154,-0.2442488968,1.0181758404,-0.031690862,0.3275598884,0.3483323455,-0.2674277425,0.3755925596,-0.3395925164,0.1744762212,-0.1544803083,0.1659665108,-0.2527278662,-0.0472251885,-0.0009159505,0.0243480615,-0.1145456955,0.2757887244,0.025885107,0.1319821179,0.1885783225,0.542791605,0.0451934449,-0.2683275044,-0.0698959753,0.1125915274,-0.1917246282,0.4287308753,-0.2052816451,-0.3015991449,-0.1515591294,-0.1510379016,-0.3071917892,0.0640914068,-0.3433289826,0.1522391737,-0.0954722986,-0.2829515636,0.0741137341,-0.0723655298,-0.4508009255,-0.2348684371,-0.2485983223,-0.1661808491,-0.3236630559,-0.1033271924,-0.0118674999,0.1541040391,0.1638712883,-0.236692816,0.0903784335,-0.139910385,-0.0686791614,-0.428665638,0.2736342251,0.1849133372,0.0192768127,-0.005036545,-0.2380008101,-0.0062390254,-0.1350820214,-0.1940034628,0.0324501768,0.0239186361,-0.2946556211,-0.0810283646,0.029623149,-0.1089015454,-0.1592523307,0.446151644,-0.3408080041,-0.054403089,0.2034599781,0.2049850971,-0.2875416577,-0.0615017191,0.1565030813,-0.2197719961,-0.3023819625,0.2461261898,-0.0545959026,-0.147565335,-0.0591267794,0.6080919504,0.4208750129,-0.157550469,-0.1964900643,-0.4159407318,-0.1934881806,-0.1089417264,-0.1465702504,0.0608006865,-0.34273085,0.1878957748,0.2939175069,-0.1560539901,-0.1955595762,0.080340445,0.0700087547,-0.1715662479,0.5449560881,0.1415683031,0.417999804,0.0502819754,-0.0912822261,0.0684879348,-0.2794997394,0.0045609609,-0.2792131603,0.2280917019,0.1795910001,-0.2227763683,0.1254166216,-0.1593245417,-0.2294790149,0.1573985815,-0.065103054,0.1375335604,-0.1063269526,0.0330738761,0.3706350625,-0.2928020954,-0.0464773402,0.446018368,0.2914896905,0.7225209475,-0.1139562055,0.0192740094,-0.1755818576,0.0347231179,0.1067769974,0.0230561644,0.1457345635,-0.2059277445,0.5104584694,-0.0693237409,0.0486077033,0.4199607372,0.435328275,0.4048272967,0.2504307032,0.0979634598,0.410967499,-0.0126726553,-0.2710828185,-0.2341128141,0.2145987898,0.1695533097,-0.0466604903,0.2740049958,0.0503489412,-0.0358981006,0.0169585273,-0.0415601581,0.3557260334,0.2153108716,-0.0440153442,0.1700400263,-0.0099720731,0.2208690196,0.1354673654,0.3721471429,0.2672179639,0.0126920529,-0.2078489065,0.0160804447,-0.0214947406,0.1013779715,0.122909151,-0.4758791625,0.0133363213,-0.0129726678,0.1317626238,0.0926605016,0.0999051854,0.6003855467,-0.3106774688,-0.1090619341,-0.2531542182,0.3721357584,0.0499454588,0.3044594228,0.2552295029,-0.059763886,-0.4850014746,0.0903068781,0.0666397363,-0.3369698524,0.0156723466,0.1888315678,0.1627209932,-0.1406687647,-0.2628439367,-0.2625999749,-0.0054836627,-0.4511940181,-0.0340810195,-0.0800017118,-0.2324447483,0.0859891549,0.2378678173,0.3981639743,0.1851015985,-0.0101090819,-0.160721302,-0.3526368439,0.1537134349,-0.1102500781,0.3452934623,0.2785937488,0.2199971974,0.2887620032,-0.0998423249,-0.0763660595,-0.0040914463,-0.0015205255,-0.2752718627,-0.1056312844,0.8151876926,0.1196410954,0.0299633928,-0.1736686677,-0.1735024005,-0.3805889785,0.3163445592,0.070997417,-0.2252328545,0.0813223347,-0.0717931986,0.00109989,0.0225625895,0.4185544252,0.5415891409,-0.0539325848,-0.197893396,-0.1344046742,-0.8942683935,0.1354991198,0.0762996152,-0.0512326397,-0.0535026453,-0.0450740904,0.0491696708,0.4020185769,0.04172929,0.2032446861,0.16186966,-0.2408735901,-0.0749645531,-0.0302265212,-0.1067859903,0.0741190612,-0.0451446809,0.0063168351,0.153602913,-0.409619838,-0.1287774742,-0.2356541306,-0.170801118,-0.256341964,-0.1435922831,0.5043106079,0.0902054459,0.298143357,0.0171854496,-0.177183941,-0.1429008245,0.1860742867,-0.3783836067,0.0786260739,-0.2074833661,0.0568382666,-0.3484973013,0.2905078232,-0.6230998039,0.0363786668,-0.1540139318,0.3330832422,-0.4119401574,-0.0878354162,-0.4065058231,-0.0347218961,0.4750182927,0.3073501885,-0.0940460786,0.0729617476,-0.4795749485,-0.1952843517,0.3577583432,-0.1159282774,-0.0916161835,-0.0322151817,0.1063020304,-0.1141506284,-0.1664708704,-0.3694661856,-0.2317149192,0.1074244753,-0.0059917849,-0.065624319,0.0099365059,-0.29361099,-0.2170827091,-0.0365346484,0.2938864827,0.0873769298,-0.187648192,0.0494083948,0.0033309122]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"The problem which I have now is that this dataset seems does not allow to download? Can you share it with me pls","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":23,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\nThe problem which I have now is that this dataset seems does not allow to download? Can you share it with me pls","embeddings":[-0.2195911556,-0.1120689511,-0.0246193595,0.3834899366,-0.0526028275,0.1177402362,-0.3828471899,0.4139447808,0.2322627753,0.0223058201,-0.0140955709,0.0171899106,-0.1022191644,0.0319089331,0.2101533264,-0.2765799165,-0.2073134631,0.0704654902,0.1246696562,0.0944500268,-0.3448491991,0.2155543268,0.0538635552,-0.0150338225,-0.002244324,-0.0836158395,-0.0123813748,0.1466105282,-0.4219306409,-0.4405503869,0.0988955647,0.0674532875,0.4428101778,0.2283713818,-0.0001201383,-0.0678524375,0.2262694538,-0.1244195849,-0.1119161248,-0.4976867139,-0.6040620804,-0.4267119765,0.167746067,-0.0655230507,0.1008527204,-0.2254551649,-0.0777876675,0.1828066707,0.1271830946,0.5547147989,0.1909677982,0.2840135694,0.0148847355,-0.1946382225,0.3984506726,-0.0345649943,-0.0255294144,0.2603332996,0.21910505,-0.0397336334,0.1848991066,-0.3369786739,-0.0732683539,0.2586836517,0.2793957889,0.133928299,0.1869349927,-0.5500770211,-0.2737672031,0.1698972583,0.7219539285,0.0584876686,-0.2320967019,0.0399339609,-0.0457105376,-0.0285104048,0.1964208037,0.0886953101,-0.4111626148,0.2304250598,0.1812592298,-0.2941674888,-0.2398548722,0.3223291337,-0.2248306572,0.5259870887,-0.0105329342,-0.1534703523,0.1897880733,0.2882221341,-0.1858109981,-0.1448328793,0.1183310673,0.1023713499,0.3044013083,-0.0678838044,0.0628744289,-0.2980289757,-0.0205402896,0.3111588359,0.1126253605,-0.1349289864,-0.1364929974,0.227102384,0.0349793658,0.2347702235,-0.1673434228,-0.0229942743,0.4096495211,0.2589091957,0.2554894686,-0.0043367213,-0.1848020852,-0.0966101885,-0.4551785886,0.0259516574,0.2356131971,-0.1696947515,-0.190508455,0.0792635083,-0.2015102059,-0.1778980196,-0.0338972397,0.2074545771,-0.1466837823,0.1321310848,0.0629355684,0.2153571844,0.1690015048,-0.3285966814,-0.0077877478,0.0101578021,-0.0384752452,-0.1748359054,0.1545878202,0.2050777227,-0.0165672414,-0.1299769878,0.0299464501,0.0974588618,0.021439692,-0.0664843172,0.2248134017,0.4637010992,0.0701338872,0.2412337959,0.0547637045,-0.1556604058,-0.1665830016,0.3274118602,-0.0776167214,-0.3467562497,-0.3879059255,0.1058626175,-0.3641716838,-0.4292062223,-0.218113035,0.1445102841,-0.3195294738,-0.2302254885,-0.0655010268,0.1692455411,-0.1796378195,-0.2509208322,0.1736604422,0.3953446746,-0.4349624813,0.0290004183,-0.3200612962,-0.3286038637,0.5275195241,0.1898436397,-0.1224430203,0.3176210225,-0.3467461765,0.4102366865,0.8917118311,-0.3346193433,-0.3704231083,0.624605298,-0.3008839786,0.0145958727,0.2337301373,-0.0524067543,0.3178453147,-0.2185449153,-0.5774485469,0.539021492,-0.0026435682,-0.0013851926,-0.2533585727,-0.1590520591,0.3335995972,-0.0122704068,0.3183053434,0.3677804768,0.3756348789,0.0217861347,0.3828482032,0.3941229582,0.2284238338,0.141210556,0.3604916632,0.2412090451,-0.0930865705,-0.2692519128,-0.1937125921,0.0958233774,-0.0300673023,0.2068408728,-0.1162710264,0.1539364159,-0.267205894,-0.0368832499,-0.2007843405,-0.0366823301,-0.0223183408,0.1322454214,0.1591792703,0.0649762377,-0.1109997332,0.0250935741,0.244552508,0.2543138564,0.1832391471,0.2468275428,-0.2353002429,0.0370694473,-0.0002029525,-0.0345541649,0.2976437211,-0.3468154967,-0.1460515261,0.0344424136,-0.1146792397,0.0226303823,0.3501186073,0.2540035248,0.2828947902,0.0655724183,0.0102714244,0.4430164099,0.2658160925,-0.0549040698,0.0487808958,-0.0302391443,0.2986435592,0.218571946,0.0507325009,0.2505043149,0.2344325036,-0.0882805213,0.0875615254,-0.1297133118,0.2024669945,-0.0549488068,-0.1953208447,-0.1999290735,-0.1733127981,0.0314071402,-0.1219292954,-0.0436027795,0.2009671479,0.3435566723,0.1428687423,-0.0367139727,0.1076911986,0.7519909739,0.2806704938,0.1639990062,0.3033875227,0.0694541931,0.0829764307,-0.2507296503,0.2020807266,0.1029977277,-0.3249638677,0.521641016,0.0256121531,0.192514345,0.0141275153,-0.0935264826,0.2667743564,0.3767357171,-0.1761631072,-0.090033412,-0.5962128639,-0.6407173276,-0.1288952976,-0.1327850521,-0.3110176921,-0.4138788879,-0.1716987044,0.2639714777,0.0401951931,0.0294970591,-0.2321714312,0.0237820316,-0.2966636121,-0.3578520417,-0.2559229434,0.193116799,-0.1150306612,0.0433514565,0.320225805,0.0060025943,0.3827809691,-0.0271581486,0.11095386,-0.6365141273,0.0093950564,0.2352179587,-0.0255296826,-0.0233493689,0.128038317,0.1464680582,0.3535769582,0.0205632225,0.2245682627,-0.0904154778,-0.1382988393,-0.2259079516,0.055093877,-0.0535068735,-0.1188602298,-0.1805927753,-0.4567272067,-0.5308476686,0.0875444487,0.1722155809,0.1441112012,0.0104325507,-0.2622845173,0.2387209684,-0.1256548464,-0.1229226217,-0.0297259446,-0.1118735224,0.4354454875,-0.3324555755,-0.5053565502,0.2403377295,0.1019918621,-0.1038701534,0.0684414655,-0.3555796146,-0.2570869625,-0.124401249,0.3066698909,-0.1409659386,0.2711115479,0.3191989362,-0.1667193323,0.0051958007,-0.0055935071,-0.0693926588,0.0037146129,0.5728480816,0.4226831496,0.1594146192,0.3276003003,0.0024073636,0.6448269486,-0.0095672924,0.2822215557,0.6291103363,-0.1618081182,0.1219543144,0.1170671433,-0.2406397015,0.0876679868,-0.0754482448,0.0392876044,0.2569905221,-0.031655632,-0.0045874198,-0.3665598035,-0.1046306193,-0.2244952917,-0.1580657959,-0.0212368947,0.0689395592,0.1123700589,0.0793810636,-0.1203675196,-0.0176578425,-0.2099653184,0.1976162195,0.6311357617,0.072408475,0.164438352,0.4267177582,0.2370337844,-0.5568380952,0.1474598199,0.2749114633,0.2561472654,-0.1357026547,-0.0030323255,0.2488021255,0.0404339246,0.7217104435,-0.2233360708,0.1260593534,0.0106760291,-0.2298774868,0.1163207218,-0.1070605963,-0.1702568084,-0.0390463173,-0.0372091755,-0.1960916519,-0.030076623,-0.3179146349,0.1791249961,0.0211202037,-0.1266051233,-0.2111361027,-0.0882482678,-0.2241411805,-0.3866355419,-0.2749633789,0.0203403439,0.0635187551,0.3686639965,0.170425415,0.1055454016,0.0657222345,0.1088792533,-0.1017641276,-0.1081665382,0.050608743,-0.0923402607,0.0870002583,0.0051074708,0.0940076262,0.4323658943,-0.1201166511,-0.2625788748,0.071622923,-0.007045683,0.4005003572,0.3010057211,-0.1169489697,-0.1035785303,-0.2422213107,0.1936787963,-0.2980011702,0.0387144759,0.5218784213,-0.2922292054,-0.158819139,-0.2169635445,0.4570279121,-0.0837512985,-0.073540628,0.102676183,0.188549459,-0.0562191345,-0.1339755356,-0.1844018698,0.9474808574,0.0006262517,0.2272841632,0.2905671895,-0.2102317363,0.3450115025,-0.31956774,0.1814590245,-0.1474466473,0.1320079714,-0.312477082,-0.0764249861,0.0163861625,0.0375518054,-0.1446471661,0.3108085096,-0.0256562978,0.0877423733,0.2116396278,0.4602811038,0.1035515592,-0.1053005457,-0.0241237916,0.1478614211,-0.1752704829,0.5175879002,-0.1765725911,-0.3218096793,-0.1366093159,-0.1697395444,-0.3718522787,0.0758538768,-0.3833491206,0.0824168026,-0.106381841,-0.3292470276,0.1229932532,-0.0723284632,-0.4242785871,-0.2412085235,-0.2548487782,-0.1742917895,-0.2951797843,-0.1632646918,-0.0610752553,0.224673599,0.181134671,-0.280856818,-0.0130471122,-0.0724508986,-0.0246469621,-0.3059975505,0.3301275671,0.2087536305,0.0457782298,0.0393141955,-0.2539334297,-0.0312265567,-0.1195662171,-0.1648338288,0.078991577,0.0304570291,-0.2807351947,-0.0941655561,0.0118661756,-0.1039208174,-0.1088645011,0.4157003462,-0.3446667194,0.0068461578,0.169771269,0.1644347459,-0.248095274,-0.1656563133,0.2473639697,-0.2960809767,-0.2661187053,0.1913012862,-0.1005120948,-0.0942674801,-0.0505361035,0.6551602483,0.4011157155,-0.1281387955,-0.2591781616,-0.4842569828,-0.0643111914,-0.2011118382,-0.2008961886,-0.0664837211,-0.4469634295,0.1706749797,0.4031454921,-0.2508141994,-0.2210259885,0.1231693849,-0.0719735995,-0.1279245317,0.5456643105,0.1578342617,0.3221291006,0.0788176805,-0.0606031083,-0.0121935718,-0.2856484354,-0.029918557,-0.2306776941,0.2108377665,0.2161486149,-0.1658072472,0.1258391291,-0.2379730195,-0.3372804821,0.2775915265,-0.1696918756,0.070010744,0.0119971847,0.153385371,0.2888015807,-0.2897823751,-0.1860709488,0.3590416908,0.1816685945,0.6558068991,-0.1496990025,0.0859209001,-0.0532652661,0.0603953674,-0.0206981637,-0.0771495253,0.006219557,-0.2031152099,0.5135411024,-0.0538603961,0.1767507792,0.4479880333,0.3833984435,0.3180074692,0.2119265199,0.1303649843,0.4260506928,0.0462132059,-0.3384754956,-0.2188012153,0.1932468563,0.2005938888,-0.0252602566,0.2288807631,0.0827228203,-0.0342293158,-0.0437557735,-0.0720423535,0.3187459707,0.1803257316,-0.053669557,0.269136399,-0.0442194752,0.2605239451,0.1503638327,0.3784708977,0.1672927588,0.0411241874,-0.2037897706,-0.0412801355,0.0644503385,0.0052709971,0.1637534946,-0.451017648,0.0703416243,-0.1305047274,0.1743481457,-0.0327422209,0.0706915259,0.415880233,-0.3137712479,-0.0369261242,-0.25046736,0.3388393819,0.0419832505,0.3652963042,0.1648036838,-0.0635407045,-0.4671598077,-0.0256664921,0.0354944915,-0.2077417821,-0.0060373512,0.2209276706,0.1604884714,-0.1041954085,-0.2530535758,-0.3030154705,-0.0372196771,-0.4389596283,-0.0574247539,-0.1826453507,-0.2376430929,0.0271142144,0.2644408345,0.3302514255,0.2279307097,-0.0377206169,-0.1494458765,-0.3516336977,0.069665648,-0.1207115054,0.3769831955,0.3315568268,0.2289936692,0.3107689321,-0.061569795,-0.0512908585,0.0913041979,0.015280595,-0.2757429183,-0.0464224182,0.8493897319,0.1211000681,0.0498543642,-0.1883588284,-0.1459112614,-0.3266359568,0.2783893645,0.1192675382,-0.2139117718,0.1260946393,-0.1155523807,0.0180723164,-0.0688824952,0.3841584027,0.5230175853,0.0476042889,-0.1605906636,-0.0681992695,-0.896712482,0.1346840113,0.0837276354,-0.179769069,0.0002072031,0.02018103,-0.0139931524,0.4384086728,0.1066239998,0.2803474069,0.1367417723,-0.1346459687,-0.0751738325,-0.13756378,-0.1080867574,0.1513875276,-0.1366295218,0.0082220649,0.1951931268,-0.4555903673,-0.0540177524,-0.28187567,-0.2733702362,-0.2601463795,-0.040161632,0.51650846,0.1068888903,0.2701793909,-0.0284752008,-0.0465431772,-0.1167982072,0.2739857137,-0.3469812572,0.0959106982,-0.1564934701,-0.0202717464,-0.231190607,0.3776462674,-0.634260416,-0.0379633531,-0.1673160642,0.3864785433,-0.4003718495,-0.137267679,-0.3108042777,-0.0862047821,0.4823465347,0.1522886306,-0.1266354322,0.0065469146,-0.469194293,-0.1850824058,0.3231465816,-0.1022177115,-0.0873273015,-0.0631031096,0.20399113,-0.1114904806,-0.1419046223,-0.2607578039,-0.2664494812,0.1900374144,-0.0605474524,-0.1300034225,0.0141061675,-0.2847542465,-0.1756198704,-0.0043705283,0.2746061683,0.0575128011,-0.1825055331,-0.0254795235,0.0080391364]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"The dataset loader is not yet ready, because of that issue.\r\nIf you want to just download the dataset the old-fashioned way, just go to: https:\/\/www-i6.informatik.rwth-aachen.de\/ftp\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz (the ftp link is now broken, and its available over https)","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":37,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\nThe dataset loader is not yet ready, because of that issue.\r\nIf you want to just download the dataset the old-fashioned way, just go to: https:\/\/www-i6.informatik.rwth-aachen.de\/ftp\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz (the ftp link is now broken, and its available over https)","embeddings":[-0.2821508646,-0.0215412136,-0.0087811993,0.2353300899,-0.0750115365,0.0881375074,-0.2813771665,0.4609623253,0.2376454324,0.021310145,0.0197128095,0.0793237612,-0.1120400652,0.0288054869,0.2074834853,-0.2595874369,-0.1745924205,0.0500782542,0.0860737637,0.1392570436,-0.3367642462,0.2255442142,0.093302615,0.0308642723,0.0250631385,-0.0268508308,-0.0471630022,0.2105636001,-0.3284516335,-0.4584114254,0.1427454054,0.0775095671,0.4306091368,0.2409657538,-0.0001195454,0.0194303878,0.3105322719,-0.0597625934,-0.1921466589,-0.4733149707,-0.5679838061,-0.4453574717,0.1962510943,-0.0225957353,0.1427641064,-0.3135254979,-0.0587531067,0.1872797608,0.1010844186,0.5034633875,0.1821435839,0.2758154571,0.0917264596,-0.1938150674,0.330616802,-0.0564118177,-0.0530907661,0.2648175359,0.2582599819,-0.1079685166,0.0960217267,-0.3503322005,-0.0996441245,0.2647417486,0.2908240557,0.1488873512,0.1582138538,-0.4356733561,-0.3246853948,0.2026762217,0.7000156045,0.0620673262,-0.2714015245,-0.0443456843,-0.0864064172,-0.0254671145,0.1995577365,0.0327339061,-0.3621903062,0.1935665607,0.1648830622,-0.3147110343,-0.2591263354,0.3636896312,-0.1913252026,0.5927678943,0.0355499908,-0.1313801706,0.2586606145,0.2421822101,-0.136722967,-0.1332246512,0.1057113633,0.070227176,0.2541283369,-0.0678744242,0.0353642888,-0.3570435047,-0.0415239222,0.3222969174,0.1360559165,-0.0897942558,-0.17428267,0.2492266446,0.0741091147,0.3009946048,-0.1796060205,-0.0477748848,0.4765815437,0.2672050595,0.326230973,0.0576015227,-0.1351092011,-0.1535451263,-0.4611136913,0.0714939982,0.214524135,-0.1443417519,-0.1893423796,0.0324055701,-0.1543548256,-0.1804030836,-0.0766993761,0.1669009775,-0.1812235713,0.2234870046,0.0402884074,0.1987622976,0.0476837568,-0.3713335395,0.0091267638,-0.0419542864,-0.0121170115,-0.179774493,0.1604952365,0.189798817,-0.0275741052,-0.1526712626,-0.0065940763,0.1284850985,-0.0272581037,-0.014349116,0.1219643056,0.4336204827,0.0577198267,0.2097714841,0.0633033961,-0.1816210151,-0.1494152844,0.3578703105,-0.1170712635,-0.421894908,-0.4223869741,0.0920756534,-0.3440444171,-0.4904932976,-0.3125094175,0.1632703543,-0.3456213176,-0.2232789397,-0.1560643315,0.1160735041,-0.225792706,-0.1811288744,0.135245055,0.3909862936,-0.4929123223,0.0246767756,-0.3338528275,-0.2905285954,0.4726332724,0.1250493228,-0.1509480625,0.3816462457,-0.37542817,0.3525737524,0.8150752187,-0.3507859707,-0.3017644584,0.6629589796,-0.3365699947,0.0677952841,0.164497152,-0.051824335,0.3105320036,-0.2412600815,-0.5735133886,0.558311522,-0.0212691035,0.0276219603,-0.2257877886,-0.1839017421,0.3385206461,0.0228705481,0.2872436345,0.3535701334,0.381776005,-0.0145400017,0.3758553267,0.3502211869,0.1977120787,0.1511360556,0.2737631202,0.2791815996,-0.1216490045,-0.1910120696,-0.3182948828,0.1219421923,0.0116811609,0.1984686106,-0.1835459471,0.2057807297,-0.2299393415,-0.0293844622,-0.1918578893,-0.0055028312,-0.0523436628,0.1304647923,0.1405320466,0.0675436109,-0.0822006017,0.0480408706,0.3051608801,0.3396253884,0.1168485433,0.2010666132,-0.2154401094,0.0572634675,0.0254720505,-0.0420174003,0.2511591613,-0.3942451179,-0.1590341926,0.1220752448,-0.0913945138,0.0655208975,0.3380198181,0.2291091084,0.3188152313,0.1212419569,0.0016420583,0.4901714921,0.1867331266,-0.062870793,0.048834227,0.0486768298,0.2936161458,0.292632252,0.0477472395,0.239598617,0.2429680079,-0.0954048187,0.0971740559,-0.0827437937,0.1909264624,-0.0649102256,-0.2271276116,-0.215315327,-0.1315472424,0.0321932882,-0.1921300739,-0.0864678919,0.1772977114,0.3546396494,0.1660159528,-0.0113036698,0.1358819306,0.7480919957,0.3015371561,0.1720658541,0.2561052442,0.0626670718,0.0420245714,-0.2632793784,0.2398954779,0.1317509264,-0.3243546188,0.4762621224,-0.00190047,0.1587664187,-0.0111282719,-0.085818626,0.2294829488,0.3678600788,-0.2246294022,-0.0351526439,-0.6205755472,-0.5866054296,-0.1442284584,-0.1357115805,-0.2736100852,-0.5019877553,-0.1304797381,0.2951658666,-0.0787130445,0.0650274754,-0.3146880567,0.02385398,-0.2574695647,-0.376431644,-0.1914370954,0.2212374359,-0.1721339226,0.0412393324,0.3339522779,-0.0831134692,0.3308106363,0.0350064225,0.1513211876,-0.6007010341,-0.0285297129,0.2332389355,-0.0630536303,-0.0624087676,0.1022353992,0.1170671955,0.3741820157,-0.0691222548,0.2739335299,-0.1158630326,-0.1350189,-0.1439898759,0.092330955,-0.0261781756,-0.1369401067,-0.2131223232,-0.4117681384,-0.5828337669,0.0274788514,0.1414267421,0.1413069218,0.0885490924,-0.1958159655,0.1616010666,-0.0886616111,-0.155405432,0.0077403332,-0.1080392748,0.4505591393,-0.3225883245,-0.4660054147,0.2123737782,0.0972791016,-0.1651213914,0.0707212165,-0.4019480944,-0.3196632564,-0.0820868537,0.2933103144,-0.1356044412,0.2699176371,0.2686161399,-0.1009520516,-0.0188567005,0.0767827481,-0.0001750693,-0.0275560711,0.4820994139,0.4234930873,0.2095014453,0.3233124912,0.0368335731,0.6444626451,-0.0771290287,0.2494202256,0.6027050018,-0.1445818543,0.1560005993,0.1535471529,-0.1989412159,0.0645653382,-0.1352654397,0.1036542431,0.2738640606,-0.0293843169,-0.0050092638,-0.3752741516,-0.0089604296,-0.1581338048,-0.2018981129,0.0422366969,-0.0317868628,0.0560783409,0.1093128249,-0.063311927,-0.0176115967,-0.2139255106,0.2782017589,0.5194920301,0.0708413348,0.1476035714,0.4403631091,0.3020991087,-0.601334095,0.2155667245,0.2751151919,0.3023559153,-0.05192836,-0.0765480399,0.1519107223,0.04416316,0.7506873012,-0.1240007058,0.1207204983,-0.0285310019,-0.253030479,0.080157429,-0.1567016095,-0.1289250255,-0.0478217453,-0.0667254627,-0.1551598758,-0.050171379,-0.2962909043,0.1810438782,0.01441056,-0.1790700406,-0.1830759794,-0.0868665501,-0.2324106991,-0.3717514277,-0.2859154344,-0.0525193587,0.0626253411,0.3296981454,0.1340361089,0.1256370544,-0.0214328188,0.0917509943,-0.1359567791,-0.1142072529,0.0439662077,-0.0712708309,0.2089931667,0.023187805,0.0333664864,0.4457804859,-0.1423975825,-0.3427382708,0.0388872512,0.0148114953,0.4080195427,0.2724869847,-0.0987553671,-0.1731152087,-0.2164417505,0.2574450672,-0.3312039971,0.0613607988,0.4953374267,-0.267308563,-0.1414252222,-0.2788743675,0.4383151829,-0.1001902223,-0.0560497008,0.125070706,0.1384723037,-0.0760494471,-0.0614440478,-0.2324933112,0.8629674911,0.0144217648,0.2370169908,0.3038595021,-0.175624162,0.3585059941,-0.3500007689,0.1629862338,-0.1360680312,0.0968439952,-0.2991203368,-0.0423142873,0.0123619186,0.0892767534,-0.0915816799,0.3780939579,0.0022058457,0.0685440078,0.2407547981,0.4632816911,0.1174289659,-0.1882355064,0.0045812335,0.1454018503,-0.2447636276,0.4712463021,-0.2008114904,-0.2791591287,-0.1171380356,-0.2049705088,-0.3058803082,0.1529554129,-0.3597407043,0.1835206449,-0.0428470708,-0.2563566267,0.0438552536,-0.0704385042,-0.457016021,-0.2150658965,-0.2023921609,-0.1428188682,-0.2738178074,-0.0793440342,-0.0613126792,0.1927106082,0.1764434576,-0.2176543772,0.0413536541,-0.1109787896,-0.0507863089,-0.4298627973,0.3740624189,0.183421582,0.0304617211,0.0513000488,-0.1973140538,-0.069912836,-0.0936939642,-0.1903689355,0.0598284043,0.0476390384,-0.2772141993,-0.0420044325,0.0035592841,-0.131146118,-0.0879939124,0.4120772779,-0.3942357004,-0.0091985418,0.2196848243,0.1942667365,-0.2773800492,-0.1044059619,0.1771951467,-0.2804226875,-0.2604638338,0.2247442007,-0.129341051,-0.1274029911,0.014879494,0.6381238699,0.365467906,-0.1188959554,-0.2743760645,-0.5013254285,-0.0295176953,-0.1777423471,-0.1380681545,-0.028397873,-0.334826529,0.1493633091,0.3305063248,-0.2770786285,-0.2031548768,0.0706726313,0.0392596088,-0.1099808142,0.5276737809,0.1723448485,0.374548614,0.0889013931,-0.0562382452,-0.02578789,-0.2458661199,-0.0296868794,-0.278414458,0.2010084838,0.2245386988,-0.2131719738,0.085435614,-0.2243053317,-0.2534430027,0.250028193,-0.1331517547,0.0895387456,-0.0641103089,0.0805288777,0.3064457476,-0.2585799098,-0.0882368237,0.3885104358,0.2041369379,0.7370729446,-0.1020248607,0.0850819945,-0.1089493409,0.0646433309,0.0211400408,-0.0322586074,0.0539428182,-0.2059779018,0.5440858603,-0.1116633937,0.1611338705,0.4218918979,0.4294244349,0.3363713324,0.1875032932,0.0661525056,0.422524929,0.0326179489,-0.2813044488,-0.2828013897,0.1432813555,0.2398876846,-0.0510544553,0.2163697779,0.0733692721,-0.0466286764,-0.0501256548,-0.0949195921,0.339237541,0.2038549632,-0.0515505634,0.2960546315,-0.0968080983,0.235894084,0.1667422801,0.2681959569,0.1888979375,0.0904361233,-0.1493038088,-0.0187355839,0.1248281896,0.0542375818,0.0799058452,-0.4931425452,0.1297039837,-0.1220384091,0.1883015037,0.0023201369,0.1750240326,0.432664156,-0.3352252841,-0.0759861022,-0.2435451299,0.3419324458,0.0644242093,0.3038132787,0.1719255298,-0.0095828306,-0.4720115066,0.0489578359,0.0523403063,-0.285513252,-0.0627946407,0.2283147424,0.1387498975,-0.1609402448,-0.2634184361,-0.3478029072,-0.0558286346,-0.388550818,-0.0338365026,-0.1174168736,-0.2494976819,0.0382826254,0.2297584713,0.3301566243,0.2160333395,-0.059186209,-0.1435101628,-0.3559989035,0.1270053983,-0.1315833628,0.350456804,0.2799951434,0.2201452851,0.332847178,-0.0612771399,-0.0679303929,0.0908693895,0.0506149121,-0.2412766367,0.017053213,0.8381709456,0.1983498931,0.0631714091,-0.2176034003,-0.1121020988,-0.3152917922,0.2876056731,0.1484920532,-0.264836967,0.0957766101,-0.1258524209,0.0186044704,-0.0145046376,0.3802063465,0.5779767036,0.0022769845,-0.1475172043,-0.0319148377,-0.8814552426,0.1708936989,0.0187444612,-0.0922380835,-0.0153410472,-0.0211312883,0.0578299277,0.4521678388,0.0934954286,0.2524431646,0.160710007,-0.1544790566,-0.0674720928,-0.112852484,-0.0911042169,0.1765305251,-0.1561423093,-0.0853340626,0.1565014124,-0.4184089899,-0.0696822628,-0.2324753553,-0.2883923948,-0.253156364,-0.1012190953,0.5299973488,0.1074942723,0.2558782399,0.0311715286,-0.0750325322,-0.1839953959,0.2173179686,-0.3720805347,0.101464957,-0.1989294291,-0.0357299931,-0.2310822457,0.4101987779,-0.6611530185,0.0198226981,-0.1762946844,0.3600581884,-0.4730418622,-0.0884845555,-0.392757982,-0.1201888397,0.4417955279,0.1423179507,-0.0831570551,0.0282110162,-0.4381508231,-0.181188792,0.2574197948,-0.1110700294,-0.0902615935,-0.0728062764,0.1308198571,-0.1639696211,-0.0821132064,-0.2027184367,-0.231645301,0.1264264286,-0.0672128573,-0.0879971161,-0.0319006853,-0.2705280185,-0.1920363307,-0.0220188126,0.2878769934,0.0903607681,-0.1675287187,-0.0393887572,0.0255844202]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/721","title":"feat(dl_manager): add support for ftp downloads","comments":"Got it, thank you so much!","body":"I am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n","comment_length":6,"text":"feat(dl_manager): add support for ftp downloads\nI am working on a new dataset (#302) and encounter a problem downloading it.\r\n\r\n```python\r\n# This is the official download link from https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/\r\n_URL = \"ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz\"\r\n\r\ndl_manager.download_and_extract(_URL)\r\n```\r\n\r\nI get an error:\r\n\r\n> ValueError: unable to parse ftp:\/\/wasserstoff.informatik.rwth-aachen.de\/pub\/rwth-phoenix\/2016\/phoenix-2014-T.v3.tar.gz as a URL or as a local path\r\n\r\nI checked, and indeed you don't consider `ftp` as a remote file.\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/4c2af707a6955cf4b45f83ac67990395327c5725\/src\/datasets\/utils\/file_utils.py#L188\r\n\r\nAdding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.\r\n\r\n\r\n\r\n\nGot it, thank you so much!","embeddings":[-0.2126119137,-0.105668053,-0.0007811597,0.2975113988,-0.058780171,0.0699706078,-0.358628273,0.4665720761,0.2261405289,0.0345444754,0.0063683246,0.0323727652,-0.1088812128,0.0282213446,0.215288505,-0.3030371964,-0.1441714317,0.0891770869,0.1512079537,0.1197713092,-0.368216753,0.2432816029,0.0791300535,0.027631145,-0.035612762,-0.0311522558,-0.0498432964,0.1876486838,-0.3593316674,-0.4674740136,0.0837802738,0.062321268,0.4426888824,0.25443542,-0.0001212206,-0.0487639122,0.2991600335,-0.077933915,-0.1218912825,-0.482131362,-0.5479144454,-0.4215336442,0.1877740771,-0.0204157382,0.1280240417,-0.2801418602,-0.0812771395,0.2213316858,0.09118817,0.4994577169,0.1692496985,0.2896858752,0.0430356897,-0.219632715,0.4001064599,-0.0340692326,-0.0168172959,0.2405412346,0.277759105,-0.1213869154,0.1511707902,-0.325006038,-0.0999746323,0.245314911,0.2981637418,0.1833368689,0.188765645,-0.4589036405,-0.3279529512,0.1772979647,0.7004752755,0.0756181404,-0.2252479941,0.0019740108,-0.0881897882,-0.0389715247,0.1910102665,0.0173417348,-0.3815852404,0.1804523766,0.1664944291,-0.3180598021,-0.2576699257,0.3676729798,-0.1973506212,0.5571352839,0.0487504266,-0.1615852714,0.2658912838,0.2502712011,-0.2255968451,-0.1307333708,0.0936120152,0.0788671821,0.3097814023,-0.0936625078,0.0573463365,-0.3340337276,-0.0409815647,0.3245583177,0.0777752921,-0.1289359033,-0.1524286717,0.2710548639,0.0640627742,0.3102921247,-0.1451157928,-0.0148711288,0.4457752705,0.2382022291,0.2698770761,0.002382847,-0.131486252,-0.126184687,-0.4463693798,0.0386458784,0.2090035379,-0.1795550883,-0.1392163336,0.046216201,-0.1991677433,-0.156618461,-0.0373204835,0.1776865572,-0.1590933353,0.154797703,0.0666841343,0.2309334874,0.0879055783,-0.3005750477,-0.0041290587,-0.0183556043,-0.0278895255,-0.1718954295,0.1558862478,0.1921874881,-0.022678189,-0.1647319943,0.0480327271,0.0884923562,-0.0350799151,-0.0008676248,0.1571017355,0.4103775024,0.0308127385,0.2045414299,0.0508439019,-0.1825755388,-0.1705961227,0.3346003592,-0.0609676614,-0.417440474,-0.4218883812,0.0934561044,-0.310644716,-0.4798686206,-0.2458954751,0.1683319211,-0.32980299,-0.2513088882,-0.1468007863,0.1491383165,-0.1827988774,-0.1983227134,0.1209572032,0.3734580874,-0.4144250154,0.0291019827,-0.2889033258,-0.2861227989,0.504930377,0.1850250512,-0.1464401484,0.3502469063,-0.3416886032,0.4026052654,0.8359761238,-0.3352044821,-0.2715401351,0.648907721,-0.3387178779,0.0491234325,0.2037837803,-0.058865767,0.3201461434,-0.2384623587,-0.5967105031,0.5312612653,-0.0203742739,0.0281430893,-0.2488248497,-0.1428492069,0.3201168478,-0.0131673012,0.304184407,0.3515080512,0.39999488,0.003863445,0.3596712053,0.3598415852,0.2074237466,0.1116639525,0.3310908973,0.3007498085,-0.1056143865,-0.2577478886,-0.1999596208,0.0642469823,0.0120887952,0.2021362334,-0.1310111135,0.1451321691,-0.2414437532,-0.0366884209,-0.1761261225,0.0116253495,-0.0474337824,0.1167031974,0.1692971438,0.0934341773,-0.0810307562,0.0085509159,0.2862268984,0.3123243749,0.185193494,0.1991097331,-0.2432799637,0.0332196765,0.0235696603,-0.0079385238,0.2777887583,-0.3598569036,-0.1432510316,0.0852712169,-0.1224464551,0.0532171503,0.3762054145,0.2251874357,0.3104636073,0.1063867062,0.016662756,0.4917182922,0.2315866947,-0.0926904231,0.0679637492,0.0093036462,0.3512553871,0.2727631629,0.038854029,0.2436721772,0.1929914951,-0.0836679116,0.1350914836,-0.117683962,0.2175098956,-0.0382055864,-0.2735078037,-0.2215679288,-0.1565509439,0.0363968089,-0.1558702588,-0.0762391016,0.1572926193,0.3674932718,0.1893507391,-0.0055451808,0.0975138098,0.7340248227,0.2701306343,0.1755817384,0.2864920795,0.07965222,0.0605679266,-0.2452714294,0.2392493635,0.1280343831,-0.3814903796,0.4782448113,0.0112830848,0.1587582231,0.0200834647,-0.0762270167,0.2052078247,0.3778917193,-0.1911681443,-0.0497907326,-0.630124867,-0.6431543231,-0.1614997536,-0.1864856035,-0.3114741743,-0.4579983354,-0.1281510293,0.2783065736,-0.0310349483,0.0748625621,-0.2218396217,0.0461155586,-0.3055746257,-0.3691912591,-0.1946413368,0.1907774806,-0.1681050658,0.0221063029,0.3318359256,-0.0448213629,0.3709984124,0.0055191102,0.1643063128,-0.5872776508,-0.0316426232,0.2095719576,-0.0730698407,-0.0809546039,0.1068395153,0.0773827508,0.3795509934,-0.0368276387,0.2527633905,-0.1218668148,-0.126101613,-0.1605987549,0.1077131629,-0.074286066,-0.1270583123,-0.1851478219,-0.4269123375,-0.5641823411,0.0863819569,0.145265013,0.1356523186,0.0469334833,-0.2455386966,0.2061327249,-0.1473110467,-0.1365184784,0.008670249,-0.094322443,0.4327356815,-0.3193574548,-0.4523018599,0.2318858206,0.1106598452,-0.1800525635,0.0566388741,-0.3397907913,-0.3155540228,-0.0894830674,0.2680430412,-0.1362195164,0.3121235371,0.2732063234,-0.1192001104,-0.0038765224,0.0484087914,-0.0436860472,0.0211573299,0.5425739884,0.4273420274,0.1969109327,0.297011435,-0.0133165326,0.6181204915,-0.0228054337,0.2966670394,0.6242889762,-0.1203598678,0.1402164251,0.133283481,-0.2157981694,0.0744546577,-0.0880875587,0.0341362245,0.2886872888,-0.0039806669,-0.0240627527,-0.3874899149,-0.0473271608,-0.1590750515,-0.1604474336,0.0085544707,0.0273359083,0.065593794,0.0925231501,-0.09565413,-0.0283769704,-0.1959549487,0.2384900749,0.5998197198,0.0679303929,0.1679787934,0.4535048008,0.2414030135,-0.5688303113,0.212627247,0.3315867186,0.2319153845,-0.0963530317,-0.0619769804,0.1871579438,0.0768441558,0.713553369,-0.1558488607,0.0703887269,-0.0246325359,-0.2775731683,0.1602158844,-0.1093901917,-0.1580657214,-0.0580378175,-0.049335707,-0.1699758321,-0.0273812655,-0.3250612319,0.1878829747,-0.004773817,-0.1145173684,-0.222067222,-0.0641481504,-0.2081581205,-0.3624149263,-0.2637877166,-0.00736026,0.0727424771,0.3580117226,0.1383801997,0.1419562697,0.0219335463,0.1053147316,-0.1206187233,-0.0798171759,-0.001489668,-0.0871277004,0.1243532375,0.0263615698,0.0822223052,0.3834381104,-0.1606231332,-0.3037316799,-0.0248164702,-0.0114737628,0.4582042396,0.2911813259,-0.0992695242,-0.1137853637,-0.2919534445,0.25123474,-0.3035507798,0.0390977748,0.5152456164,-0.2670708597,-0.1256285161,-0.234898001,0.4583648443,-0.1041586995,-0.0633342415,0.129591614,0.175859347,-0.0752712488,-0.090374209,-0.2545063496,0.924384892,0.0561261848,0.2265282571,0.2778679132,-0.2586701214,0.3327999413,-0.3221211135,0.1914760768,-0.1493446827,0.1535163075,-0.2988134325,-0.040913716,-0.0225866325,0.0349114984,-0.1446696073,0.3509344459,-0.029135555,0.0620302446,0.200875774,0.4563136697,0.1177383959,-0.1599223316,0.0087761311,0.1314154118,-0.2177159935,0.5015893579,-0.2063933313,-0.3046430051,-0.1339556724,-0.1986084729,-0.3338581324,0.106393002,-0.3586746156,0.1266753674,-0.0573138259,-0.3070451021,0.0710494965,-0.0925111324,-0.4661950171,-0.2269196212,-0.1952413917,-0.2027527839,-0.2538070977,-0.1161324605,-0.0751287267,0.2028677016,0.164948374,-0.2543311417,0.0212622304,-0.122402072,-0.0377865955,-0.3619621098,0.3261345327,0.1754250377,0.0337467566,0.0668708161,-0.2135999799,-0.031743966,-0.1207416207,-0.1793361008,0.0618916452,0.0303149428,-0.3192237318,-0.0401467048,-0.0070212348,-0.1228300855,-0.1137941852,0.3765180707,-0.3748160601,-0.0277742259,0.1581548452,0.1798696369,-0.289537549,-0.1254481971,0.193814069,-0.3133497536,-0.2441553026,0.238464728,-0.097234346,-0.138985768,0.001052083,0.6697244048,0.3858258128,-0.1358945668,-0.2478831708,-0.4633538127,-0.0759450868,-0.1868367493,-0.1755925268,-0.0185471885,-0.4202850163,0.191336602,0.3578615189,-0.2595577836,-0.20212017,0.0966386646,0.0229102019,-0.0845932439,0.5350301862,0.138502568,0.3574820757,0.1101758108,-0.0685657114,0.0076221153,-0.2728740573,-0.0187282972,-0.2584894598,0.2065690905,0.2220849544,-0.1939870268,0.1105964109,-0.1941284835,-0.2564472556,0.2486084402,-0.1435788572,0.0808290094,-0.03667349,0.1119670123,0.3359970152,-0.3032180369,-0.0806062147,0.3831391335,0.2155951113,0.7001659274,-0.1524683982,0.0238346551,-0.1086545289,0.0309887733,0.0564430095,-0.0352484882,0.0351461507,-0.1925798655,0.5361096263,-0.0765660927,0.1506664604,0.4634658694,0.448949039,0.3358605206,0.2186780274,0.0899601877,0.4598769844,0.025373362,-0.2702007294,-0.2755449116,0.2061547041,0.2384947687,-0.0652312115,0.2407082915,0.0626269579,-0.0273518208,-0.0325812027,-0.0781954303,0.3028678894,0.1784471273,-0.0628427789,0.2563156188,-0.0595512316,0.2510738075,0.133586213,0.3439708352,0.1981960386,0.0493551306,-0.1852570474,-0.0075311554,0.0687543303,0.0293905288,0.1221798882,-0.4461005032,0.0730282664,-0.1373590082,0.1878450066,0.0037990818,0.1161297336,0.4550787508,-0.3649117053,-0.0877450705,-0.2433637977,0.354462564,0.0400511287,0.3484389782,0.2293685228,-0.0204163454,-0.4830627143,0.012564647,0.0583398156,-0.2269357741,-0.064882122,0.2309322655,0.1682733297,-0.1308041513,-0.2869111896,-0.3740094304,-0.0084817978,-0.4153676927,-0.0640107915,-0.177776292,-0.2294123173,0.0465086885,0.2366785556,0.3457094431,0.242806226,-0.0939176008,-0.1676982343,-0.3937638402,0.1159779057,-0.0783596784,0.3565541208,0.3063958883,0.2554956079,0.2877852619,-0.0882408842,-0.0577697791,0.0996753648,0.0149557963,-0.2940506041,-0.0048696999,0.8790072203,0.1883275062,0.0485853702,-0.2287493646,-0.1502418965,-0.336753577,0.2911872864,0.1217537373,-0.2262770832,0.0933099762,-0.1079271659,0.0099698426,-0.0320791788,0.3683573902,0.5511055589,0.0066435882,-0.1698871255,-0.0683180541,-0.8546507955,0.1642346233,0.0614067279,-0.1075595915,-0.0152006596,-0.0272314679,0.005673869,0.4103737473,0.1009956151,0.274208039,0.1527303308,-0.1822310835,-0.078166455,-0.0830342099,-0.1028614864,0.1033862084,-0.12315052,0.0130612515,0.1495684534,-0.4573408365,-0.0651810095,-0.246527493,-0.2577905357,-0.2650389671,-0.0629730597,0.5231481194,0.0979339927,0.278883338,0.0115835285,-0.0737574995,-0.1028137654,0.234753862,-0.3906285763,0.0799727887,-0.2214599699,-0.012579971,-0.2705534101,0.4269626141,-0.6717286706,0.0106301336,-0.1676397622,0.3749326169,-0.4148010612,-0.1392906755,-0.3579791486,-0.119463779,0.5101983547,0.1830921322,-0.0954664871,0.0520747975,-0.4716263413,-0.1973584592,0.3010956049,-0.0599812977,-0.1154141724,-0.0559537262,0.1560561657,-0.1166034266,-0.1140902489,-0.2734312117,-0.278450042,0.1303945035,-0.0540446676,-0.0990918055,0.0006548799,-0.2916392684,-0.2278538793,0.0075549511,0.2673057318,0.0839251205,-0.1534629017,-0.0154921487,0.0266717728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/720","title":"OSError: Cannot find data file when not using the dummy dataset in RAG","comments":"Same issue here. I will be digging further, but it looks like the [script](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py#L132) is attempting to open a file that is not downloaded yet. \r\n\r\n```\r\n99dcbca09109e58502e6b9271d4d3f3791b43f61f3161a76b25d2775ab1a4498.lock\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n~\/anaconda3\/envs\/eqa\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 446 try:\r\n--> 447 return pickle.load(fid, **pickle_kwargs)\r\n 448 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n~\/src\/datasets\/src\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 559 \r\n--> 560 if verify_infos:\r\n 561 verify_splits(self.info.splits, split_dict)\r\n\r\n~\/src\/datasets\/src\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 847 writer.write(example)\r\n--> 848 finally:\r\n 849 num_examples, num_bytes = writer.finalize()\r\n\r\n~\/anaconda3\/envs\/eqa\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 227 try:\r\n--> 228 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 229 # return super(tqdm...) will not catch exception\r\n\r\n~\/anaconda3\/envs\/eqa\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1132 try:\r\n-> 1133 for obj in iterable:\r\n 1134 yield obj\r\n\r\n\/hdd\/rag\/cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2\/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n~\/anaconda3\/envs\/eqa\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 449 raise IOError(\r\n--> 450 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 451 \r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='\/hdd\/rag\/downloads\/99dcbca09109e58502e6b9271d4d3f3791b43f61f3161a76b25d2775ab1a4498'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-8-24351ff8ce44> in <module>\r\n 4 retriever = RagRetriever.from_pretrained(\"facebook\/rag-sequence-nq\", \r\n 5 index_name=\"exact\",\r\n----> 6 use_dummy_dataset=False)\r\n\r\n~\/src\/transformers\/src\/transformers\/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 321 generator_tokenizer = rag_tokenizer.generator\r\n 322 return cls(\r\n--> 323 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 324 )\r\n 325 \r\n\r\n~\/src\/transformers\/src\/transformers\/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 310 self.config = config\r\n 311 if self._init_retrieval:\r\n--> 312 self.init_retrieval()\r\n 313 \r\n 314 @classmethod\r\n\r\n~\/src\/transformers\/src\/transformers\/retrieval_rag.py in init_retrieval(self)\r\n 338 \r\n 339 logger.info(\"initializing retrieval\")\r\n--> 340 self.index.init_index()\r\n 341 \r\n 342 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n~\/src\/transformers\/src\/transformers\/retrieval_rag.py in init_index(self)\r\n 248 split=self.dataset_split,\r\n 249 index_name=self.index_name,\r\n--> 250 dummy=self.use_dummy_dataset,\r\n 251 )\r\n 252 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n~\/src\/datasets\/src\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 615 builder_instance.download_and_prepare(\r\n 616 download_config=download_config,\r\n--> 617 download_mode=download_mode,\r\n 618 ignore_verifications=ignore_verifications,\r\n 619 )\r\n\r\n~\/src\/datasets\/src\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 481 # Sync info\r\n 482 self.info.dataset_size = sum(split.num_bytes for split in self.info.splits.values())\r\n--> 483 self.info.download_checksums = dl_manager.get_recorded_sizes_checksums()\r\n 484 self.info.size_in_bytes = self.info.dataset_size + self.info.download_size\r\n 485 # Save info\r\n\r\n~\/src\/datasets\/src\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 560 if verify_infos:\r\n 561 verify_splits(self.info.splits, split_dict)\r\n--> 562 \r\n 563 # Update the info object with the splits.\r\n 564 self.info.splits = split_dict\r\n\r\nOSError: Cannot find data file.\r\n```\r\n\r\nThank you.","body":"## Environment info\r\n\r\n transformers version: 3.3.1\r\n Platform: Linux-4.19\r\n Python version: 3.7.7\r\n PyTorch version (GPU?): 1.6.0\r\n Tensorflow version (GPU?): No\r\n Using GPU in script?: Yes\r\n Using distributed or parallel set-up in script?: No\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behaviour:\r\n```\r\nimport os\r\nos.environ['HF_DATASETS_CACHE'] = '\/workspace\/notebooks\/POCs\/cache'\r\n\r\nfrom transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n\r\ntokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\nretriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False) \r\n```\r\n\r\nPlese note that I'm using the whole dataset: **use_dummy_dataset=False**\r\nAfter around 4 hours (downloading and some other things) this is returned:\r\n\r\n```\r\nDownloading and preparing dataset wiki_dpr\/psgs_w100.nq.exact (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/workspace\/notebooks\/POCs\/cache\/wiki_dpr\/psgs_w100.nq.exact\/0.0.0\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2...\r\n\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 459 try:\r\n--> 460 return pickle.load(fid, **pickle_kwargs)\r\n 461 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 552 # Prepare split will record examples associated to the split\r\n--> 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 840 for key, record in utils.tqdm(\r\n--> 841 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 842 ):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 217 try:\r\n--> 218 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 219 # return super(tqdm...) will not catch exception\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2\/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 462 raise IOError(\r\n--> 463 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 464 finally:\r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='\/workspace\/notebooks\/POCs\/cache\/downloads\/f34d5f091294259b4ca90e813631e69a6ded660d71b6cbedf89ddba50df94448'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-10-f28df370ac47> in <module>\r\n 1 # ln -s \/workspace\/notebooks\/POCs\/cache \/root\/.cache\/huggingface\/datasets\r\n----> 2 retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 307 generator_tokenizer = rag_tokenizer.generator\r\n 308 return cls(\r\n--> 309 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 310 )\r\n 311 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 298 self.config = config\r\n 299 if self._init_retrieval:\r\n--> 300 self.init_retrieval()\r\n 301 \r\n 302 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_retrieval(self)\r\n 324 \r\n 325 logger.info(\"initializing retrieval\")\r\n--> 326 self.index.init_index()\r\n 327 \r\n 328 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_index(self)\r\n 238 split=self.dataset_split,\r\n 239 index_name=self.index_name,\r\n--> 240 dummy=self.use_dummy_dataset,\r\n 241 )\r\n 242 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 474 if not downloaded_from_gcs:\r\n 475 self._download_and_prepare(\r\n--> 476 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 477 )\r\n 478 # Sync info\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n--> 555 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n 556 \r\n 557 if verify_infos:\r\n\r\nOSError: Cannot find data file. \r\n\r\n```\r\n\r\nThanks \r\n","comment_length":387,"text":"OSError: Cannot find data file when not using the dummy dataset in RAG\n## Environment info\r\n\r\n transformers version: 3.3.1\r\n Platform: Linux-4.19\r\n Python version: 3.7.7\r\n PyTorch version (GPU?): 1.6.0\r\n Tensorflow version (GPU?): No\r\n Using GPU in script?: Yes\r\n Using distributed or parallel set-up in script?: No\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behaviour:\r\n```\r\nimport os\r\nos.environ['HF_DATASETS_CACHE'] = '\/workspace\/notebooks\/POCs\/cache'\r\n\r\nfrom transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n\r\ntokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\nretriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False) \r\n```\r\n\r\nPlese note that I'm using the whole dataset: **use_dummy_dataset=False**\r\nAfter around 4 hours (downloading and some other things) this is returned:\r\n\r\n```\r\nDownloading and preparing dataset wiki_dpr\/psgs_w100.nq.exact (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/workspace\/notebooks\/POCs\/cache\/wiki_dpr\/psgs_w100.nq.exact\/0.0.0\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2...\r\n\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 459 try:\r\n--> 460 return pickle.load(fid, **pickle_kwargs)\r\n 461 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 552 # Prepare split will record examples associated to the split\r\n--> 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 840 for key, record in utils.tqdm(\r\n--> 841 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 842 ):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 217 try:\r\n--> 218 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 219 # return super(tqdm...) will not catch exception\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2\/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 462 raise IOError(\r\n--> 463 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 464 finally:\r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='\/workspace\/notebooks\/POCs\/cache\/downloads\/f34d5f091294259b4ca90e813631e69a6ded660d71b6cbedf89ddba50df94448'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-10-f28df370ac47> in <module>\r\n 1 # ln -s \/workspace\/notebooks\/POCs\/cache \/root\/.cache\/huggingface\/datasets\r\n----> 2 retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 307 generator_tokenizer = rag_tokenizer.generator\r\n 308 return cls(\r\n--> 309 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 310 )\r\n 311 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 298 self.config = config\r\n 299 if self._init_retrieval:\r\n--> 300 self.init_retrieval()\r\n 301 \r\n 302 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_retrieval(self)\r\n 324 \r\n 325 logger.info(\"initializing retrieval\")\r\n--> 326 self.index.init_index()\r\n 327 \r\n 328 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_index(self)\r\n 238 split=self.dataset_split,\r\n 239 index_name=self.index_name,\r\n--> 240 dummy=self.use_dummy_dataset,\r\n 241 )\r\n 242 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 474 if not downloaded_from_gcs:\r\n 475 self._download_and_prepare(\r\n--> 476 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 477 )\r\n 478 # Sync info\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n--> 555 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n 556 \r\n 557 if verify_infos:\r\n\r\nOSError: Cannot find data file. \r\n\r\n```\r\n\r\nThanks \r\n\nSame issue here. I will be digging further, but it looks like the [script](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wiki_dpr\/wiki_dpr.py#L132) is attempting to open a file that is not downloaded yet. \r\n\r\n```\r\n99dcbca09109e58502e6b9271d4d3f3791b43f61f3161a76b25d2775ab1a4498.lock\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n~\/anaconda3\/envs\/eqa\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 446 try:\r\n--> 447 return pickle.load(fid, **pickle_kwargs)\r\n 448 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n~\/src\/datasets\/src\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 559 \r\n--> 560 if verify_infos:\r\n 561 verify_splits(self.info.splits, split_dict)\r\n\r\n~\/src\/datasets\/src\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 847 writer.write(example)\r\n--> 848 finally:\r\n 849 num_examples, num_bytes = writer.finalize()\r\n\r\n~\/anaconda3\/envs\/eqa\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 227 try:\r\n--> 228 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 229 # return super(tqdm...) will not catch exception\r\n\r\n~\/anaconda3\/envs\/eqa\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1132 try:\r\n-> 1133 for obj in iterable:\r\n 1134 yield obj\r\n\r\n\/hdd\/rag\/cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2\/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n~\/anaconda3\/envs\/eqa\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 449 raise IOError(\r\n--> 450 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 451 \r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='\/hdd\/rag\/downloads\/99dcbca09109e58502e6b9271d4d3f3791b43f61f3161a76b25d2775ab1a4498'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-8-24351ff8ce44> in <module>\r\n 4 retriever = RagRetriever.from_pretrained(\"facebook\/rag-sequence-nq\", \r\n 5 index_name=\"exact\",\r\n----> 6 use_dummy_dataset=False)\r\n\r\n~\/src\/transformers\/src\/transformers\/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 321 generator_tokenizer = rag_tokenizer.generator\r\n 322 return cls(\r\n--> 323 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 324 )\r\n 325 \r\n\r\n~\/src\/transformers\/src\/transformers\/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 310 self.config = config\r\n 311 if self._init_retrieval:\r\n--> 312 self.init_retrieval()\r\n 313 \r\n 314 @classmethod\r\n\r\n~\/src\/transformers\/src\/transformers\/retrieval_rag.py in init_retrieval(self)\r\n 338 \r\n 339 logger.info(\"initializing retrieval\")\r\n--> 340 self.index.init_index()\r\n 341 \r\n 342 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n~\/src\/transformers\/src\/transformers\/retrieval_rag.py in init_index(self)\r\n 248 split=self.dataset_split,\r\n 249 index_name=self.index_name,\r\n--> 250 dummy=self.use_dummy_dataset,\r\n 251 )\r\n 252 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n~\/src\/datasets\/src\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 615 builder_instance.download_and_prepare(\r\n 616 download_config=download_config,\r\n--> 617 download_mode=download_mode,\r\n 618 ignore_verifications=ignore_verifications,\r\n 619 )\r\n\r\n~\/src\/datasets\/src\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 481 # Sync info\r\n 482 self.info.dataset_size = sum(split.num_bytes for split in self.info.splits.values())\r\n--> 483 self.info.download_checksums = dl_manager.get_recorded_sizes_checksums()\r\n 484 self.info.size_in_bytes = self.info.dataset_size + self.info.download_size\r\n 485 # Save info\r\n\r\n~\/src\/datasets\/src\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 560 if verify_infos:\r\n 561 verify_splits(self.info.splits, split_dict)\r\n--> 562 \r\n 563 # Update the info object with the splits.\r\n 564 self.info.splits = split_dict\r\n\r\nOSError: Cannot find data file.\r\n```\r\n\r\nThank you.","embeddings":[-0.2258688062,-0.0487169065,0.0123973489,0.1052551121,0.3706667423,-0.0698539317,0.3323160708,0.2692305744,0.0155129815,0.325091511,-0.13942644,0.217653051,-0.0637350604,-0.3219818473,-0.0156813674,0.0387514979,0.0088648982,0.176944688,-0.2071304321,-0.2274689972,-0.2748771906,0.0712658539,-0.2206947207,-0.0241290927,-0.6810176969,0.1146390289,-0.05781601,0.3083193898,-0.1954237521,-0.251630187,0.202607438,-0.3225424588,0.2263191193,0.7035835981,-0.0001225423,0.2990083098,0.1429332048,-0.2735655606,-0.2608669698,-0.278689146,0.4701996744,-0.0022001471,-0.1336365491,-0.2349150628,-0.3879041076,-0.0179219656,0.1504863203,-0.2608039379,0.1146472842,0.4453887343,0.1201401874,0.5135034323,-0.0789261982,-0.1666063219,0.4416405857,-0.1321118772,-0.2456288487,0.2880708873,0.1355750412,-0.3839831054,-0.2244360745,0.3043579757,0.0357949734,0.214489162,0.0499641933,0.1509962678,-0.331099242,-0.1621896625,0.2034894675,0.1592778862,0.4668168724,-0.1024601981,-0.4101263881,0.1078529805,-0.1492986828,-0.01189222,0.3080828786,0.250741303,-0.2307934165,0.3093445301,-0.1880170703,0.1411207914,-0.1838113368,0.3203998804,-0.1091684997,0.5142392516,-0.0597856715,0.1333447248,0.0515007935,0.266982913,-0.1275482178,-0.0532716103,0.2568569183,0.1107259393,-0.4334582984,-0.0851476789,-0.141758278,-0.2281454057,0.0090549998,0.283882916,0.119040519,-0.0872798711,-0.050617218,0.0559016839,-0.0747926906,0.0031038059,-0.4398024082,0.3263707459,0.1369172037,0.0633962378,-0.0617391281,-0.2556205094,0.0124686137,-0.0752046928,-0.1377577782,-0.0830673054,0.2952482104,-0.1419144869,-0.3792672455,-0.1376513541,-0.2718304992,0.044915434,-0.0649191588,0.5335163474,-0.1762444675,0.0942436904,0.2233512551,0.2114507407,-0.3598036766,-0.2135165185,-0.189420566,0.1914025992,-0.2589720488,-0.0695585534,0.336157769,0.1515605152,0.5469791293,-0.0422237702,-0.0395792834,0.0757380351,0.5255907774,-0.3452492356,0.1852508485,0.3671411276,-0.0957375243,0.1022145227,0.3197237849,0.0107704932,-0.0048441472,0.1602138281,-0.134794578,-0.4402830601,-0.0159310251,0.123059392,-0.096447885,-0.0022720867,0.5216624737,-0.3452539146,0.0150453327,-0.0186987352,0.1125785187,-0.0878424272,-0.3588058054,-0.177587077,0.1657052636,0.3699544072,-0.4192665219,-0.0607249737,0.0974297151,0.1354825199,0.272669822,0.407297045,-0.2612980008,0.5780154467,-0.3673042655,-0.0979130566,0.2546078265,-0.3039738536,-0.4831534326,-0.0294232927,-0.2176701427,0.3169541955,0.3235011697,0.1196951345,0.0952945575,0.1186122298,0.0603864826,0.2023348361,-0.0112885842,-0.1898554116,-0.3436117768,-0.36027354,0.3072978556,0.3618117571,0.3364816904,0.0394932702,-0.0466328859,0.2523000836,0.2554249763,-0.1765722781,0.0135172177,0.3977082074,0.2658633292,0.0652284846,0.1295245886,-0.2036168128,-0.0927456841,0.2144518346,0.0100941388,0.1594747603,0.0780224055,-0.2786820233,-0.2695260644,-0.1588254571,-0.2466272712,-0.4860778153,0.0125515806,0.3241280615,0.2520541549,-0.0928548723,-0.2521234155,0.4377958477,-0.1952871829,0.2539069653,-0.3626701534,0.3693114519,-0.1394768953,-0.1955277026,-0.1041209176,0.0695498437,0.0412254259,-0.1709701717,0.0008806916,0.3422067165,0.2573977411,-0.3763899207,0.2662567496,0.057855133,0.0825240463,-0.119228363,-0.0103514362,0.4316679239,0.228631705,0.0311591905,-0.1992195249,0.2342782319,-0.0837530121,-0.0150448382,-0.0632191002,-0.0380300283,0.2054506838,-0.2455834001,0.0543307029,0.0540948063,0.2960953116,0.0623819977,0.5005347133,-0.086363323,-0.1532837451,-0.0471563451,0.4012603462,-0.0439240448,0.0086467201,0.1421054751,-0.1760937274,0.100363411,-0.0817503482,-0.1644160599,0.589443326,0.1907703578,0.0153245684,0.0057594245,0.0197617151,-0.1352092624,-0.0579379462,0.0057319207,0.1119224355,0.2939564884,0.010454841,0.0019570047,-0.204659909,-0.2126511633,0.1033594385,0.2841060758,-0.086158067,0.0140030636,-0.229308635,0.2530737817,-0.1107315645,0.0085567124,-0.2427500635,-0.1266786009,0.0411275737,0.2089068741,-0.0423332155,0.3041032553,0.092355594,0.0221105833,0.2015033662,-0.195492208,-0.1071512103,-0.1214265823,-0.2365635782,-0.0093308585,-0.0095360149,-0.075437203,0.2127888501,-0.2755419016,-0.0830216929,-0.5109938383,-0.2276017517,0.0538613498,0.2870907784,0.2047031373,0.1335327327,0.2145299315,-0.062615633,-0.0509736314,0.2801601589,-0.2143454701,-0.1684692353,0.0525324866,-0.0789887682,0.2073695064,0.0410677455,-0.6410495043,-0.3373051286,-0.2971862853,-0.1248568967,-0.0847932026,-0.0033860109,-0.1192247272,0.1342246085,0.2047805488,0.0360465646,-0.0473384112,-0.123831518,-0.2134780288,0.5039732456,-0.0896632299,-0.2254339159,0.1115513816,-0.2123592645,0.0627879202,0.2687756121,-0.5052886009,-0.5236566663,-0.091148667,0.3369216025,-0.1908483356,0.089464426,0.4921669066,-0.3213092685,-0.0286873244,-0.0757975057,-0.048251491,0.1265075505,-0.1775063425,0.4617244005,-0.0131942481,0.2616951168,0.2069703788,1.1618537903,0.221200183,-0.0264962409,0.0756130442,-0.0601709187,0.3890325725,-0.2206432372,-0.1952376217,0.169987455,-0.0849358365,-0.0233410988,0.1195211336,-0.0171637014,0.131513387,-0.0434012078,-0.0431164354,-0.3271123767,-0.0630915686,0.3613573313,0.2374397814,0.1733063161,0.1608201563,0.21682778,-0.1940921843,-0.1063198075,0.194404766,0.4131210148,0.0717765316,0.1459725946,-0.2727347612,0.0053165574,-0.2924439907,0.1266802698,-0.1389422715,0.2734206617,-0.0478715971,-0.0166708715,0.0786937103,-0.0301470272,0.8610990644,-0.2157857567,0.0723627359,0.1461741179,-0.4310403466,-0.1535596102,-0.1024157703,-0.2645180821,-0.1194293648,0.6156060696,0.5470615625,-0.1647453606,-0.0651736259,0.401129216,0.0672709718,-0.0415488146,-0.0908466503,-0.3198335469,-0.3359068334,-0.2896226048,-0.0756401643,-0.0588875487,0.1726746112,0.0725102648,0.2630057335,-0.3187335432,0.1581834853,-0.0134523716,-0.1048918962,0.2868012786,-0.3563987911,0.233405143,0.2810381353,0.3141556978,0.1240825579,0.4758909941,-0.1620752215,-0.5533198118,0.2848446667,-0.0089414632,0.1100370586,0.3649648428,-0.0149248876,0.0206317194,-0.0600723773,0.3059141338,-0.2079595476,0.1888481379,0.1273121983,0.1805795282,-0.0055888225,-0.2922878563,0.2343427986,-0.0100594731,0.121710591,0.2594493032,-0.1489346325,-0.037145853,0.8069669008,-0.0851551369,1.0018033981,-0.0164092015,0.3242408335,0.2122611552,-0.1894021481,0.3186568022,-0.8604207635,0.3705362976,-0.3098131418,-0.2558270991,-0.2118623704,-0.2688034177,0.272595346,0.5007386208,-0.2377667129,0.3705540895,0.1126124337,0.5121015906,0.1188098565,0.3404255509,0.1157986969,-0.0725710988,0.1873907596,0.0555037782,-0.0894562155,0.5973527431,0.0034105976,-0.0756136701,-0.3768138289,-0.0559503622,-0.0864164159,0.1534355581,-0.5980362892,0.1214436367,0.4480594993,-0.3019877374,-0.1662076116,0.2210670263,0.4297845066,0.2084310651,-0.1015106961,0.3052919805,0.0094819376,-0.1383112669,0.2651391625,-0.1400008053,0.1618786156,-0.0763958842,-0.4398621917,0.2476784587,-0.3299744129,-0.3752295375,0.2602829337,-0.0489757359,0.3252377808,-0.3127461374,-0.2047451735,-0.0382046252,-0.1589909345,-0.1972965747,0.0725952163,-0.0142965252,-0.214747861,-0.0079059144,-0.1116086021,-0.4335786402,-0.2141374052,0.3808218539,0.0006370423,0.081953615,0.5034264922,-0.1717196256,-0.2764569521,-0.1857311279,0.0238212924,0.2257454544,-0.4661904275,-0.1291016191,0.0142129529,-0.0213841144,0.077913858,-0.3143901825,0.3677604198,-0.0758163631,-0.5247256756,-0.4727155268,-0.2337951809,-0.0284098126,0.2426598817,0.1366354525,0.0898085684,0.1172851548,0.1010095552,0.0829965845,-0.2611963451,0.1274620146,-0.3284144104,0.2241388261,-0.0823822469,-0.432854116,0.1134232283,-0.0398013927,0.0124952691,0.1324867904,-0.2328983843,-0.1520704329,-0.0522504523,0.1510518938,0.0293514002,-0.1678698659,-0.0627266392,-0.0445908681,-0.0316798128,-0.1187212691,0.1657162905,0.2946432531,0.1132879704,0.0098577179,-0.0731213689,-0.1555406004,-0.1787892282,-0.0207018834,-0.1610881239,0.2682994008,0.2411250174,0.1016138569,-0.0367552526,0.0535285659,-0.0402750336,-0.0307252686,-0.0901167691,-0.0944979787,0.3644737899,-0.4622706175,-0.0739855468,-0.1404031664,0.1395464242,-0.0114119621,-0.346552521,-0.1271581799,0.3380667269,0.1625091136,-0.4333016574,-0.0775827616,0.2363430411,0.0022680191,0.237414211,0.2966210544,0.0799449608,0.0015987612,0.2902956903,0.093101956,0.3097213209,-0.2249634415,-0.0230655484,-0.0278911963,0.048248969,0.1535485834,0.1837675571,0.047395207,0.0814307779,0.4353090525,0.1530235261,0.2819969356,-0.0381257124,0.1201000512,-0.1897055507,-0.4449408352,-0.2815672159,0.3054930568,-0.235494256,-0.0182448924,-0.3626365364,0.055864159,-0.1567477286,-0.1506158262,-0.0402768403,-0.1189904884,-0.2089492232,0.1465533823,-0.0952102095,-0.0301971864,-0.1931044161,0.102811411,0.1171641648,-0.1189524606,0.0187289137,0.0301830098,-0.2823504508,0.1210343465,-0.1132590547,0.0880358443,-0.0462888256,-0.2088495493,0.1870757788,0.2697921693,0.0533148795,0.0685409531,0.3685548902,0.4437269866,0.2311636508,-0.2230366319,0.192424655,-0.0318639688,-0.0763217807,-0.030568568,0.159598127,-0.1536246389,-0.0688809678,0.3970540166,0.0480838045,-0.189527005,-0.3199321032,0.0274739191,0.0225167349,-0.4600149393,0.2075293213,-0.0441925414,-0.0788600147,0.0065660863,-0.0245958772,-0.5514102578,0.127814576,0.3803643584,0.0252658669,0.2749297619,0.1281455904,0.0078003039,-0.2787654996,0.0311802514,0.4185286164,-0.0350203849,-0.5950027704,-0.3033539355,-0.4331485927,0.0840362385,-0.569270432,-0.292607069,0.001013095,0.248501271,0.0891625583,0.1719410121,0.3557214141,-0.105475843,0.0784786418,0.4126326144,-0.4025351107,-0.1663885862,-0.3254982531,0.2052407563,0.0800544843,-0.7052715421,0.0650318116,0.3884374797,0.0101613589,0.1391600668,-0.0993559957,0.3541438282,0.0327954255,0.1194630712,-0.1563095003,0.2681676447,0.0471640974,-0.0312050562,0.1960101873,0.0786783695,-0.1172625795,0.0075382129,0.0786147416,0.5020301342,-0.158609584,-0.160935536,-0.277139008,0.2690277398,0.1713728905,-0.1698029339,0.124161534,0.2329676896,-0.0899593756,0.227147907,0.2715693414,0.3370776772,-0.0686973631,0.0453815237,-0.3402681351,-0.0994826332,0.583381772,-0.3087985814,-0.2146814317,-0.1653167605,0.116828151,-0.053248167,0.2518343329,-0.4316256344,-0.0399834663,0.3314810097,0.064550586,-0.1202179641,-0.0008699126,0.0577102937,0.0541665554,-0.0688194558,0.2568713427,-0.0468433127,-0.2186792493,-0.1020397916,-0.2438035309]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/720","title":"OSError: Cannot find data file when not using the dummy dataset in RAG","comments":"An update on my end. This seems like a transient issue. Reran the script from scratch overnight with no errors. ","body":"## Environment info\r\n\r\n transformers version: 3.3.1\r\n Platform: Linux-4.19\r\n Python version: 3.7.7\r\n PyTorch version (GPU?): 1.6.0\r\n Tensorflow version (GPU?): No\r\n Using GPU in script?: Yes\r\n Using distributed or parallel set-up in script?: No\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behaviour:\r\n```\r\nimport os\r\nos.environ['HF_DATASETS_CACHE'] = '\/workspace\/notebooks\/POCs\/cache'\r\n\r\nfrom transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n\r\ntokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\nretriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False) \r\n```\r\n\r\nPlese note that I'm using the whole dataset: **use_dummy_dataset=False**\r\nAfter around 4 hours (downloading and some other things) this is returned:\r\n\r\n```\r\nDownloading and preparing dataset wiki_dpr\/psgs_w100.nq.exact (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/workspace\/notebooks\/POCs\/cache\/wiki_dpr\/psgs_w100.nq.exact\/0.0.0\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2...\r\n\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 459 try:\r\n--> 460 return pickle.load(fid, **pickle_kwargs)\r\n 461 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 552 # Prepare split will record examples associated to the split\r\n--> 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 840 for key, record in utils.tqdm(\r\n--> 841 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 842 ):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 217 try:\r\n--> 218 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 219 # return super(tqdm...) will not catch exception\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2\/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 462 raise IOError(\r\n--> 463 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 464 finally:\r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='\/workspace\/notebooks\/POCs\/cache\/downloads\/f34d5f091294259b4ca90e813631e69a6ded660d71b6cbedf89ddba50df94448'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-10-f28df370ac47> in <module>\r\n 1 # ln -s \/workspace\/notebooks\/POCs\/cache \/root\/.cache\/huggingface\/datasets\r\n----> 2 retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 307 generator_tokenizer = rag_tokenizer.generator\r\n 308 return cls(\r\n--> 309 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 310 )\r\n 311 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 298 self.config = config\r\n 299 if self._init_retrieval:\r\n--> 300 self.init_retrieval()\r\n 301 \r\n 302 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_retrieval(self)\r\n 324 \r\n 325 logger.info(\"initializing retrieval\")\r\n--> 326 self.index.init_index()\r\n 327 \r\n 328 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_index(self)\r\n 238 split=self.dataset_split,\r\n 239 index_name=self.index_name,\r\n--> 240 dummy=self.use_dummy_dataset,\r\n 241 )\r\n 242 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 474 if not downloaded_from_gcs:\r\n 475 self._download_and_prepare(\r\n--> 476 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 477 )\r\n 478 # Sync info\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n--> 555 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n 556 \r\n 557 if verify_infos:\r\n\r\nOSError: Cannot find data file. \r\n\r\n```\r\n\r\nThanks \r\n","comment_length":20,"text":"OSError: Cannot find data file when not using the dummy dataset in RAG\n## Environment info\r\n\r\n transformers version: 3.3.1\r\n Platform: Linux-4.19\r\n Python version: 3.7.7\r\n PyTorch version (GPU?): 1.6.0\r\n Tensorflow version (GPU?): No\r\n Using GPU in script?: Yes\r\n Using distributed or parallel set-up in script?: No\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behaviour:\r\n```\r\nimport os\r\nos.environ['HF_DATASETS_CACHE'] = '\/workspace\/notebooks\/POCs\/cache'\r\n\r\nfrom transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n\r\ntokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\nretriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False) \r\n```\r\n\r\nPlese note that I'm using the whole dataset: **use_dummy_dataset=False**\r\nAfter around 4 hours (downloading and some other things) this is returned:\r\n\r\n```\r\nDownloading and preparing dataset wiki_dpr\/psgs_w100.nq.exact (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/workspace\/notebooks\/POCs\/cache\/wiki_dpr\/psgs_w100.nq.exact\/0.0.0\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2...\r\n\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 459 try:\r\n--> 460 return pickle.load(fid, **pickle_kwargs)\r\n 461 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 552 # Prepare split will record examples associated to the split\r\n--> 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 840 for key, record in utils.tqdm(\r\n--> 841 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 842 ):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 217 try:\r\n--> 218 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 219 # return super(tqdm...) will not catch exception\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2\/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 462 raise IOError(\r\n--> 463 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 464 finally:\r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='\/workspace\/notebooks\/POCs\/cache\/downloads\/f34d5f091294259b4ca90e813631e69a6ded660d71b6cbedf89ddba50df94448'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-10-f28df370ac47> in <module>\r\n 1 # ln -s \/workspace\/notebooks\/POCs\/cache \/root\/.cache\/huggingface\/datasets\r\n----> 2 retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 307 generator_tokenizer = rag_tokenizer.generator\r\n 308 return cls(\r\n--> 309 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 310 )\r\n 311 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 298 self.config = config\r\n 299 if self._init_retrieval:\r\n--> 300 self.init_retrieval()\r\n 301 \r\n 302 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_retrieval(self)\r\n 324 \r\n 325 logger.info(\"initializing retrieval\")\r\n--> 326 self.index.init_index()\r\n 327 \r\n 328 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_index(self)\r\n 238 split=self.dataset_split,\r\n 239 index_name=self.index_name,\r\n--> 240 dummy=self.use_dummy_dataset,\r\n 241 )\r\n 242 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 474 if not downloaded_from_gcs:\r\n 475 self._download_and_prepare(\r\n--> 476 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 477 )\r\n 478 # Sync info\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n--> 555 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n 556 \r\n 557 if verify_infos:\r\n\r\nOSError: Cannot find data file. \r\n\r\n```\r\n\r\nThanks \r\n\nAn update on my end. This seems like a transient issue. Reran the script from scratch overnight with no errors. ","embeddings":[-0.2258688062,-0.0487169065,0.0123973489,0.1052551121,0.3706667423,-0.0698539317,0.3323160708,0.2692305744,0.0155129815,0.325091511,-0.13942644,0.217653051,-0.0637350604,-0.3219818473,-0.0156813674,0.0387514979,0.0088648982,0.176944688,-0.2071304321,-0.2274689972,-0.2748771906,0.0712658539,-0.2206947207,-0.0241290927,-0.6810176969,0.1146390289,-0.05781601,0.3083193898,-0.1954237521,-0.251630187,0.202607438,-0.3225424588,0.2263191193,0.7035835981,-0.0001225423,0.2990083098,0.1429332048,-0.2735655606,-0.2608669698,-0.278689146,0.4701996744,-0.0022001471,-0.1336365491,-0.2349150628,-0.3879041076,-0.0179219656,0.1504863203,-0.2608039379,0.1146472842,0.4453887343,0.1201401874,0.5135034323,-0.0789261982,-0.1666063219,0.4416405857,-0.1321118772,-0.2456288487,0.2880708873,0.1355750412,-0.3839831054,-0.2244360745,0.3043579757,0.0357949734,0.214489162,0.0499641933,0.1509962678,-0.331099242,-0.1621896625,0.2034894675,0.1592778862,0.4668168724,-0.1024601981,-0.4101263881,0.1078529805,-0.1492986828,-0.01189222,0.3080828786,0.250741303,-0.2307934165,0.3093445301,-0.1880170703,0.1411207914,-0.1838113368,0.3203998804,-0.1091684997,0.5142392516,-0.0597856715,0.1333447248,0.0515007935,0.266982913,-0.1275482178,-0.0532716103,0.2568569183,0.1107259393,-0.4334582984,-0.0851476789,-0.141758278,-0.2281454057,0.0090549998,0.283882916,0.119040519,-0.0872798711,-0.050617218,0.0559016839,-0.0747926906,0.0031038059,-0.4398024082,0.3263707459,0.1369172037,0.0633962378,-0.0617391281,-0.2556205094,0.0124686137,-0.0752046928,-0.1377577782,-0.0830673054,0.2952482104,-0.1419144869,-0.3792672455,-0.1376513541,-0.2718304992,0.044915434,-0.0649191588,0.5335163474,-0.1762444675,0.0942436904,0.2233512551,0.2114507407,-0.3598036766,-0.2135165185,-0.189420566,0.1914025992,-0.2589720488,-0.0695585534,0.336157769,0.1515605152,0.5469791293,-0.0422237702,-0.0395792834,0.0757380351,0.5255907774,-0.3452492356,0.1852508485,0.3671411276,-0.0957375243,0.1022145227,0.3197237849,0.0107704932,-0.0048441472,0.1602138281,-0.134794578,-0.4402830601,-0.0159310251,0.123059392,-0.096447885,-0.0022720867,0.5216624737,-0.3452539146,0.0150453327,-0.0186987352,0.1125785187,-0.0878424272,-0.3588058054,-0.177587077,0.1657052636,0.3699544072,-0.4192665219,-0.0607249737,0.0974297151,0.1354825199,0.272669822,0.407297045,-0.2612980008,0.5780154467,-0.3673042655,-0.0979130566,0.2546078265,-0.3039738536,-0.4831534326,-0.0294232927,-0.2176701427,0.3169541955,0.3235011697,0.1196951345,0.0952945575,0.1186122298,0.0603864826,0.2023348361,-0.0112885842,-0.1898554116,-0.3436117768,-0.36027354,0.3072978556,0.3618117571,0.3364816904,0.0394932702,-0.0466328859,0.2523000836,0.2554249763,-0.1765722781,0.0135172177,0.3977082074,0.2658633292,0.0652284846,0.1295245886,-0.2036168128,-0.0927456841,0.2144518346,0.0100941388,0.1594747603,0.0780224055,-0.2786820233,-0.2695260644,-0.1588254571,-0.2466272712,-0.4860778153,0.0125515806,0.3241280615,0.2520541549,-0.0928548723,-0.2521234155,0.4377958477,-0.1952871829,0.2539069653,-0.3626701534,0.3693114519,-0.1394768953,-0.1955277026,-0.1041209176,0.0695498437,0.0412254259,-0.1709701717,0.0008806916,0.3422067165,0.2573977411,-0.3763899207,0.2662567496,0.057855133,0.0825240463,-0.119228363,-0.0103514362,0.4316679239,0.228631705,0.0311591905,-0.1992195249,0.2342782319,-0.0837530121,-0.0150448382,-0.0632191002,-0.0380300283,0.2054506838,-0.2455834001,0.0543307029,0.0540948063,0.2960953116,0.0623819977,0.5005347133,-0.086363323,-0.1532837451,-0.0471563451,0.4012603462,-0.0439240448,0.0086467201,0.1421054751,-0.1760937274,0.100363411,-0.0817503482,-0.1644160599,0.589443326,0.1907703578,0.0153245684,0.0057594245,0.0197617151,-0.1352092624,-0.0579379462,0.0057319207,0.1119224355,0.2939564884,0.010454841,0.0019570047,-0.204659909,-0.2126511633,0.1033594385,0.2841060758,-0.086158067,0.0140030636,-0.229308635,0.2530737817,-0.1107315645,0.0085567124,-0.2427500635,-0.1266786009,0.0411275737,0.2089068741,-0.0423332155,0.3041032553,0.092355594,0.0221105833,0.2015033662,-0.195492208,-0.1071512103,-0.1214265823,-0.2365635782,-0.0093308585,-0.0095360149,-0.075437203,0.2127888501,-0.2755419016,-0.0830216929,-0.5109938383,-0.2276017517,0.0538613498,0.2870907784,0.2047031373,0.1335327327,0.2145299315,-0.062615633,-0.0509736314,0.2801601589,-0.2143454701,-0.1684692353,0.0525324866,-0.0789887682,0.2073695064,0.0410677455,-0.6410495043,-0.3373051286,-0.2971862853,-0.1248568967,-0.0847932026,-0.0033860109,-0.1192247272,0.1342246085,0.2047805488,0.0360465646,-0.0473384112,-0.123831518,-0.2134780288,0.5039732456,-0.0896632299,-0.2254339159,0.1115513816,-0.2123592645,0.0627879202,0.2687756121,-0.5052886009,-0.5236566663,-0.091148667,0.3369216025,-0.1908483356,0.089464426,0.4921669066,-0.3213092685,-0.0286873244,-0.0757975057,-0.048251491,0.1265075505,-0.1775063425,0.4617244005,-0.0131942481,0.2616951168,0.2069703788,1.1618537903,0.221200183,-0.0264962409,0.0756130442,-0.0601709187,0.3890325725,-0.2206432372,-0.1952376217,0.169987455,-0.0849358365,-0.0233410988,0.1195211336,-0.0171637014,0.131513387,-0.0434012078,-0.0431164354,-0.3271123767,-0.0630915686,0.3613573313,0.2374397814,0.1733063161,0.1608201563,0.21682778,-0.1940921843,-0.1063198075,0.194404766,0.4131210148,0.0717765316,0.1459725946,-0.2727347612,0.0053165574,-0.2924439907,0.1266802698,-0.1389422715,0.2734206617,-0.0478715971,-0.0166708715,0.0786937103,-0.0301470272,0.8610990644,-0.2157857567,0.0723627359,0.1461741179,-0.4310403466,-0.1535596102,-0.1024157703,-0.2645180821,-0.1194293648,0.6156060696,0.5470615625,-0.1647453606,-0.0651736259,0.401129216,0.0672709718,-0.0415488146,-0.0908466503,-0.3198335469,-0.3359068334,-0.2896226048,-0.0756401643,-0.0588875487,0.1726746112,0.0725102648,0.2630057335,-0.3187335432,0.1581834853,-0.0134523716,-0.1048918962,0.2868012786,-0.3563987911,0.233405143,0.2810381353,0.3141556978,0.1240825579,0.4758909941,-0.1620752215,-0.5533198118,0.2848446667,-0.0089414632,0.1100370586,0.3649648428,-0.0149248876,0.0206317194,-0.0600723773,0.3059141338,-0.2079595476,0.1888481379,0.1273121983,0.1805795282,-0.0055888225,-0.2922878563,0.2343427986,-0.0100594731,0.121710591,0.2594493032,-0.1489346325,-0.037145853,0.8069669008,-0.0851551369,1.0018033981,-0.0164092015,0.3242408335,0.2122611552,-0.1894021481,0.3186568022,-0.8604207635,0.3705362976,-0.3098131418,-0.2558270991,-0.2118623704,-0.2688034177,0.272595346,0.5007386208,-0.2377667129,0.3705540895,0.1126124337,0.5121015906,0.1188098565,0.3404255509,0.1157986969,-0.0725710988,0.1873907596,0.0555037782,-0.0894562155,0.5973527431,0.0034105976,-0.0756136701,-0.3768138289,-0.0559503622,-0.0864164159,0.1534355581,-0.5980362892,0.1214436367,0.4480594993,-0.3019877374,-0.1662076116,0.2210670263,0.4297845066,0.2084310651,-0.1015106961,0.3052919805,0.0094819376,-0.1383112669,0.2651391625,-0.1400008053,0.1618786156,-0.0763958842,-0.4398621917,0.2476784587,-0.3299744129,-0.3752295375,0.2602829337,-0.0489757359,0.3252377808,-0.3127461374,-0.2047451735,-0.0382046252,-0.1589909345,-0.1972965747,0.0725952163,-0.0142965252,-0.214747861,-0.0079059144,-0.1116086021,-0.4335786402,-0.2141374052,0.3808218539,0.0006370423,0.081953615,0.5034264922,-0.1717196256,-0.2764569521,-0.1857311279,0.0238212924,0.2257454544,-0.4661904275,-0.1291016191,0.0142129529,-0.0213841144,0.077913858,-0.3143901825,0.3677604198,-0.0758163631,-0.5247256756,-0.4727155268,-0.2337951809,-0.0284098126,0.2426598817,0.1366354525,0.0898085684,0.1172851548,0.1010095552,0.0829965845,-0.2611963451,0.1274620146,-0.3284144104,0.2241388261,-0.0823822469,-0.432854116,0.1134232283,-0.0398013927,0.0124952691,0.1324867904,-0.2328983843,-0.1520704329,-0.0522504523,0.1510518938,0.0293514002,-0.1678698659,-0.0627266392,-0.0445908681,-0.0316798128,-0.1187212691,0.1657162905,0.2946432531,0.1132879704,0.0098577179,-0.0731213689,-0.1555406004,-0.1787892282,-0.0207018834,-0.1610881239,0.2682994008,0.2411250174,0.1016138569,-0.0367552526,0.0535285659,-0.0402750336,-0.0307252686,-0.0901167691,-0.0944979787,0.3644737899,-0.4622706175,-0.0739855468,-0.1404031664,0.1395464242,-0.0114119621,-0.346552521,-0.1271581799,0.3380667269,0.1625091136,-0.4333016574,-0.0775827616,0.2363430411,0.0022680191,0.237414211,0.2966210544,0.0799449608,0.0015987612,0.2902956903,0.093101956,0.3097213209,-0.2249634415,-0.0230655484,-0.0278911963,0.048248969,0.1535485834,0.1837675571,0.047395207,0.0814307779,0.4353090525,0.1530235261,0.2819969356,-0.0381257124,0.1201000512,-0.1897055507,-0.4449408352,-0.2815672159,0.3054930568,-0.235494256,-0.0182448924,-0.3626365364,0.055864159,-0.1567477286,-0.1506158262,-0.0402768403,-0.1189904884,-0.2089492232,0.1465533823,-0.0952102095,-0.0301971864,-0.1931044161,0.102811411,0.1171641648,-0.1189524606,0.0187289137,0.0301830098,-0.2823504508,0.1210343465,-0.1132590547,0.0880358443,-0.0462888256,-0.2088495493,0.1870757788,0.2697921693,0.0533148795,0.0685409531,0.3685548902,0.4437269866,0.2311636508,-0.2230366319,0.192424655,-0.0318639688,-0.0763217807,-0.030568568,0.159598127,-0.1536246389,-0.0688809678,0.3970540166,0.0480838045,-0.189527005,-0.3199321032,0.0274739191,0.0225167349,-0.4600149393,0.2075293213,-0.0441925414,-0.0788600147,0.0065660863,-0.0245958772,-0.5514102578,0.127814576,0.3803643584,0.0252658669,0.2749297619,0.1281455904,0.0078003039,-0.2787654996,0.0311802514,0.4185286164,-0.0350203849,-0.5950027704,-0.3033539355,-0.4331485927,0.0840362385,-0.569270432,-0.292607069,0.001013095,0.248501271,0.0891625583,0.1719410121,0.3557214141,-0.105475843,0.0784786418,0.4126326144,-0.4025351107,-0.1663885862,-0.3254982531,0.2052407563,0.0800544843,-0.7052715421,0.0650318116,0.3884374797,0.0101613589,0.1391600668,-0.0993559957,0.3541438282,0.0327954255,0.1194630712,-0.1563095003,0.2681676447,0.0471640974,-0.0312050562,0.1960101873,0.0786783695,-0.1172625795,0.0075382129,0.0786147416,0.5020301342,-0.158609584,-0.160935536,-0.277139008,0.2690277398,0.1713728905,-0.1698029339,0.124161534,0.2329676896,-0.0899593756,0.227147907,0.2715693414,0.3370776772,-0.0686973631,0.0453815237,-0.3402681351,-0.0994826332,0.583381772,-0.3087985814,-0.2146814317,-0.1653167605,0.116828151,-0.053248167,0.2518343329,-0.4316256344,-0.0399834663,0.3314810097,0.064550586,-0.1202179641,-0.0008699126,0.0577102937,0.0541665554,-0.0688194558,0.2568713427,-0.0468433127,-0.2186792493,-0.1020397916,-0.2438035309]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/720","title":"OSError: Cannot find data file when not using the dummy dataset in RAG","comments":"Closing this one. Feel free to re-open if you have other questions about this issue","body":"## Environment info\r\n\r\n transformers version: 3.3.1\r\n Platform: Linux-4.19\r\n Python version: 3.7.7\r\n PyTorch version (GPU?): 1.6.0\r\n Tensorflow version (GPU?): No\r\n Using GPU in script?: Yes\r\n Using distributed or parallel set-up in script?: No\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behaviour:\r\n```\r\nimport os\r\nos.environ['HF_DATASETS_CACHE'] = '\/workspace\/notebooks\/POCs\/cache'\r\n\r\nfrom transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n\r\ntokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\nretriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False) \r\n```\r\n\r\nPlese note that I'm using the whole dataset: **use_dummy_dataset=False**\r\nAfter around 4 hours (downloading and some other things) this is returned:\r\n\r\n```\r\nDownloading and preparing dataset wiki_dpr\/psgs_w100.nq.exact (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/workspace\/notebooks\/POCs\/cache\/wiki_dpr\/psgs_w100.nq.exact\/0.0.0\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2...\r\n\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 459 try:\r\n--> 460 return pickle.load(fid, **pickle_kwargs)\r\n 461 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 552 # Prepare split will record examples associated to the split\r\n--> 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 840 for key, record in utils.tqdm(\r\n--> 841 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 842 ):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 217 try:\r\n--> 218 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 219 # return super(tqdm...) will not catch exception\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2\/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 462 raise IOError(\r\n--> 463 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 464 finally:\r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='\/workspace\/notebooks\/POCs\/cache\/downloads\/f34d5f091294259b4ca90e813631e69a6ded660d71b6cbedf89ddba50df94448'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-10-f28df370ac47> in <module>\r\n 1 # ln -s \/workspace\/notebooks\/POCs\/cache \/root\/.cache\/huggingface\/datasets\r\n----> 2 retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 307 generator_tokenizer = rag_tokenizer.generator\r\n 308 return cls(\r\n--> 309 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 310 )\r\n 311 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 298 self.config = config\r\n 299 if self._init_retrieval:\r\n--> 300 self.init_retrieval()\r\n 301 \r\n 302 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_retrieval(self)\r\n 324 \r\n 325 logger.info(\"initializing retrieval\")\r\n--> 326 self.index.init_index()\r\n 327 \r\n 328 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_index(self)\r\n 238 split=self.dataset_split,\r\n 239 index_name=self.index_name,\r\n--> 240 dummy=self.use_dummy_dataset,\r\n 241 )\r\n 242 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 474 if not downloaded_from_gcs:\r\n 475 self._download_and_prepare(\r\n--> 476 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 477 )\r\n 478 # Sync info\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n--> 555 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n 556 \r\n 557 if verify_infos:\r\n\r\nOSError: Cannot find data file. \r\n\r\n```\r\n\r\nThanks \r\n","comment_length":15,"text":"OSError: Cannot find data file when not using the dummy dataset in RAG\n## Environment info\r\n\r\n transformers version: 3.3.1\r\n Platform: Linux-4.19\r\n Python version: 3.7.7\r\n PyTorch version (GPU?): 1.6.0\r\n Tensorflow version (GPU?): No\r\n Using GPU in script?: Yes\r\n Using distributed or parallel set-up in script?: No\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behaviour:\r\n```\r\nimport os\r\nos.environ['HF_DATASETS_CACHE'] = '\/workspace\/notebooks\/POCs\/cache'\r\n\r\nfrom transformers import RagTokenizer, RagRetriever, RagTokenForGeneration\r\n\r\ntokenizer = RagTokenizer.from_pretrained(\"facebook\/rag-token-nq\")\r\nretriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False) \r\n```\r\n\r\nPlese note that I'm using the whole dataset: **use_dummy_dataset=False**\r\nAfter around 4 hours (downloading and some other things) this is returned:\r\n\r\n```\r\nDownloading and preparing dataset wiki_dpr\/psgs_w100.nq.exact (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/workspace\/notebooks\/POCs\/cache\/wiki_dpr\/psgs_w100.nq.exact\/0.0.0\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2...\r\n\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 459 try:\r\n--> 460 return pickle.load(fid, **pickle_kwargs)\r\n 461 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 552 # Prepare split will record examples associated to the split\r\n--> 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _prepare_split(self, split_generator)\r\n 840 for key, record in utils.tqdm(\r\n--> 841 generator, unit=\" examples\", total=split_info.num_examples, leave=False, disable=not_verbose\r\n 842 ):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/notebook.py in __iter__(self, *args, **kwargs)\r\n 217 try:\r\n--> 218 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 219 # return super(tqdm...) will not catch exception\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wiki_dpr\/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2\/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/numpy\/lib\/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 462 raise IOError(\r\n--> 463 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 464 finally:\r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='\/workspace\/notebooks\/POCs\/cache\/downloads\/f34d5f091294259b4ca90e813631e69a6ded660d71b6cbedf89ddba50df94448'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-10-f28df370ac47> in <module>\r\n 1 # ln -s \/workspace\/notebooks\/POCs\/cache \/root\/.cache\/huggingface\/datasets\r\n----> 2 retriever = RagRetriever.from_pretrained(\"facebook\/rag-token-nq\", index_name=\"exact\", use_dummy_dataset=False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 307 generator_tokenizer = rag_tokenizer.generator\r\n 308 return cls(\r\n--> 309 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 310 )\r\n 311 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 298 self.config = config\r\n 299 if self._init_retrieval:\r\n--> 300 self.init_retrieval()\r\n 301 \r\n 302 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_retrieval(self)\r\n 324 \r\n 325 logger.info(\"initializing retrieval\")\r\n--> 326 self.index.init_index()\r\n 327 \r\n 328 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/transformers\/retrieval_rag.py in init_index(self)\r\n 238 split=self.dataset_split,\r\n 239 index_name=self.index_name,\r\n--> 240 dummy=self.use_dummy_dataset,\r\n 241 )\r\n 242 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 474 if not downloaded_from_gcs:\r\n 475 self._download_and_prepare(\r\n--> 476 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 477 )\r\n 478 # Sync info\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 553 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 554 except OSError:\r\n--> 555 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n 556 \r\n 557 if verify_infos:\r\n\r\nOSError: Cannot find data file. \r\n\r\n```\r\n\r\nThanks \r\n\nClosing this one. Feel free to re-open if you have other questions about this issue","embeddings":[-0.2258688062,-0.0487169065,0.0123973489,0.1052551121,0.3706667423,-0.0698539317,0.3323160708,0.2692305744,0.0155129815,0.325091511,-0.13942644,0.217653051,-0.0637350604,-0.3219818473,-0.0156813674,0.0387514979,0.0088648982,0.176944688,-0.2071304321,-0.2274689972,-0.2748771906,0.0712658539,-0.2206947207,-0.0241290927,-0.6810176969,0.1146390289,-0.05781601,0.3083193898,-0.1954237521,-0.251630187,0.202607438,-0.3225424588,0.2263191193,0.7035835981,-0.0001225423,0.2990083098,0.1429332048,-0.2735655606,-0.2608669698,-0.278689146,0.4701996744,-0.0022001471,-0.1336365491,-0.2349150628,-0.3879041076,-0.0179219656,0.1504863203,-0.2608039379,0.1146472842,0.4453887343,0.1201401874,0.5135034323,-0.0789261982,-0.1666063219,0.4416405857,-0.1321118772,-0.2456288487,0.2880708873,0.1355750412,-0.3839831054,-0.2244360745,0.3043579757,0.0357949734,0.214489162,0.0499641933,0.1509962678,-0.331099242,-0.1621896625,0.2034894675,0.1592778862,0.4668168724,-0.1024601981,-0.4101263881,0.1078529805,-0.1492986828,-0.01189222,0.3080828786,0.250741303,-0.2307934165,0.3093445301,-0.1880170703,0.1411207914,-0.1838113368,0.3203998804,-0.1091684997,0.5142392516,-0.0597856715,0.1333447248,0.0515007935,0.266982913,-0.1275482178,-0.0532716103,0.2568569183,0.1107259393,-0.4334582984,-0.0851476789,-0.141758278,-0.2281454057,0.0090549998,0.283882916,0.119040519,-0.0872798711,-0.050617218,0.0559016839,-0.0747926906,0.0031038059,-0.4398024082,0.3263707459,0.1369172037,0.0633962378,-0.0617391281,-0.2556205094,0.0124686137,-0.0752046928,-0.1377577782,-0.0830673054,0.2952482104,-0.1419144869,-0.3792672455,-0.1376513541,-0.2718304992,0.044915434,-0.0649191588,0.5335163474,-0.1762444675,0.0942436904,0.2233512551,0.2114507407,-0.3598036766,-0.2135165185,-0.189420566,0.1914025992,-0.2589720488,-0.0695585534,0.336157769,0.1515605152,0.5469791293,-0.0422237702,-0.0395792834,0.0757380351,0.5255907774,-0.3452492356,0.1852508485,0.3671411276,-0.0957375243,0.1022145227,0.3197237849,0.0107704932,-0.0048441472,0.1602138281,-0.134794578,-0.4402830601,-0.0159310251,0.123059392,-0.096447885,-0.0022720867,0.5216624737,-0.3452539146,0.0150453327,-0.0186987352,0.1125785187,-0.0878424272,-0.3588058054,-0.177587077,0.1657052636,0.3699544072,-0.4192665219,-0.0607249737,0.0974297151,0.1354825199,0.272669822,0.407297045,-0.2612980008,0.5780154467,-0.3673042655,-0.0979130566,0.2546078265,-0.3039738536,-0.4831534326,-0.0294232927,-0.2176701427,0.3169541955,0.3235011697,0.1196951345,0.0952945575,0.1186122298,0.0603864826,0.2023348361,-0.0112885842,-0.1898554116,-0.3436117768,-0.36027354,0.3072978556,0.3618117571,0.3364816904,0.0394932702,-0.0466328859,0.2523000836,0.2554249763,-0.1765722781,0.0135172177,0.3977082074,0.2658633292,0.0652284846,0.1295245886,-0.2036168128,-0.0927456841,0.2144518346,0.0100941388,0.1594747603,0.0780224055,-0.2786820233,-0.2695260644,-0.1588254571,-0.2466272712,-0.4860778153,0.0125515806,0.3241280615,0.2520541549,-0.0928548723,-0.2521234155,0.4377958477,-0.1952871829,0.2539069653,-0.3626701534,0.3693114519,-0.1394768953,-0.1955277026,-0.1041209176,0.0695498437,0.0412254259,-0.1709701717,0.0008806916,0.3422067165,0.2573977411,-0.3763899207,0.2662567496,0.057855133,0.0825240463,-0.119228363,-0.0103514362,0.4316679239,0.228631705,0.0311591905,-0.1992195249,0.2342782319,-0.0837530121,-0.0150448382,-0.0632191002,-0.0380300283,0.2054506838,-0.2455834001,0.0543307029,0.0540948063,0.2960953116,0.0623819977,0.5005347133,-0.086363323,-0.1532837451,-0.0471563451,0.4012603462,-0.0439240448,0.0086467201,0.1421054751,-0.1760937274,0.100363411,-0.0817503482,-0.1644160599,0.589443326,0.1907703578,0.0153245684,0.0057594245,0.0197617151,-0.1352092624,-0.0579379462,0.0057319207,0.1119224355,0.2939564884,0.010454841,0.0019570047,-0.204659909,-0.2126511633,0.1033594385,0.2841060758,-0.086158067,0.0140030636,-0.229308635,0.2530737817,-0.1107315645,0.0085567124,-0.2427500635,-0.1266786009,0.0411275737,0.2089068741,-0.0423332155,0.3041032553,0.092355594,0.0221105833,0.2015033662,-0.195492208,-0.1071512103,-0.1214265823,-0.2365635782,-0.0093308585,-0.0095360149,-0.075437203,0.2127888501,-0.2755419016,-0.0830216929,-0.5109938383,-0.2276017517,0.0538613498,0.2870907784,0.2047031373,0.1335327327,0.2145299315,-0.062615633,-0.0509736314,0.2801601589,-0.2143454701,-0.1684692353,0.0525324866,-0.0789887682,0.2073695064,0.0410677455,-0.6410495043,-0.3373051286,-0.2971862853,-0.1248568967,-0.0847932026,-0.0033860109,-0.1192247272,0.1342246085,0.2047805488,0.0360465646,-0.0473384112,-0.123831518,-0.2134780288,0.5039732456,-0.0896632299,-0.2254339159,0.1115513816,-0.2123592645,0.0627879202,0.2687756121,-0.5052886009,-0.5236566663,-0.091148667,0.3369216025,-0.1908483356,0.089464426,0.4921669066,-0.3213092685,-0.0286873244,-0.0757975057,-0.048251491,0.1265075505,-0.1775063425,0.4617244005,-0.0131942481,0.2616951168,0.2069703788,1.1618537903,0.221200183,-0.0264962409,0.0756130442,-0.0601709187,0.3890325725,-0.2206432372,-0.1952376217,0.169987455,-0.0849358365,-0.0233410988,0.1195211336,-0.0171637014,0.131513387,-0.0434012078,-0.0431164354,-0.3271123767,-0.0630915686,0.3613573313,0.2374397814,0.1733063161,0.1608201563,0.21682778,-0.1940921843,-0.1063198075,0.194404766,0.4131210148,0.0717765316,0.1459725946,-0.2727347612,0.0053165574,-0.2924439907,0.1266802698,-0.1389422715,0.2734206617,-0.0478715971,-0.0166708715,0.0786937103,-0.0301470272,0.8610990644,-0.2157857567,0.0723627359,0.1461741179,-0.4310403466,-0.1535596102,-0.1024157703,-0.2645180821,-0.1194293648,0.6156060696,0.5470615625,-0.1647453606,-0.0651736259,0.401129216,0.0672709718,-0.0415488146,-0.0908466503,-0.3198335469,-0.3359068334,-0.2896226048,-0.0756401643,-0.0588875487,0.1726746112,0.0725102648,0.2630057335,-0.3187335432,0.1581834853,-0.0134523716,-0.1048918962,0.2868012786,-0.3563987911,0.233405143,0.2810381353,0.3141556978,0.1240825579,0.4758909941,-0.1620752215,-0.5533198118,0.2848446667,-0.0089414632,0.1100370586,0.3649648428,-0.0149248876,0.0206317194,-0.0600723773,0.3059141338,-0.2079595476,0.1888481379,0.1273121983,0.1805795282,-0.0055888225,-0.2922878563,0.2343427986,-0.0100594731,0.121710591,0.2594493032,-0.1489346325,-0.037145853,0.8069669008,-0.0851551369,1.0018033981,-0.0164092015,0.3242408335,0.2122611552,-0.1894021481,0.3186568022,-0.8604207635,0.3705362976,-0.3098131418,-0.2558270991,-0.2118623704,-0.2688034177,0.272595346,0.5007386208,-0.2377667129,0.3705540895,0.1126124337,0.5121015906,0.1188098565,0.3404255509,0.1157986969,-0.0725710988,0.1873907596,0.0555037782,-0.0894562155,0.5973527431,0.0034105976,-0.0756136701,-0.3768138289,-0.0559503622,-0.0864164159,0.1534355581,-0.5980362892,0.1214436367,0.4480594993,-0.3019877374,-0.1662076116,0.2210670263,0.4297845066,0.2084310651,-0.1015106961,0.3052919805,0.0094819376,-0.1383112669,0.2651391625,-0.1400008053,0.1618786156,-0.0763958842,-0.4398621917,0.2476784587,-0.3299744129,-0.3752295375,0.2602829337,-0.0489757359,0.3252377808,-0.3127461374,-0.2047451735,-0.0382046252,-0.1589909345,-0.1972965747,0.0725952163,-0.0142965252,-0.214747861,-0.0079059144,-0.1116086021,-0.4335786402,-0.2141374052,0.3808218539,0.0006370423,0.081953615,0.5034264922,-0.1717196256,-0.2764569521,-0.1857311279,0.0238212924,0.2257454544,-0.4661904275,-0.1291016191,0.0142129529,-0.0213841144,0.077913858,-0.3143901825,0.3677604198,-0.0758163631,-0.5247256756,-0.4727155268,-0.2337951809,-0.0284098126,0.2426598817,0.1366354525,0.0898085684,0.1172851548,0.1010095552,0.0829965845,-0.2611963451,0.1274620146,-0.3284144104,0.2241388261,-0.0823822469,-0.432854116,0.1134232283,-0.0398013927,0.0124952691,0.1324867904,-0.2328983843,-0.1520704329,-0.0522504523,0.1510518938,0.0293514002,-0.1678698659,-0.0627266392,-0.0445908681,-0.0316798128,-0.1187212691,0.1657162905,0.2946432531,0.1132879704,0.0098577179,-0.0731213689,-0.1555406004,-0.1787892282,-0.0207018834,-0.1610881239,0.2682994008,0.2411250174,0.1016138569,-0.0367552526,0.0535285659,-0.0402750336,-0.0307252686,-0.0901167691,-0.0944979787,0.3644737899,-0.4622706175,-0.0739855468,-0.1404031664,0.1395464242,-0.0114119621,-0.346552521,-0.1271581799,0.3380667269,0.1625091136,-0.4333016574,-0.0775827616,0.2363430411,0.0022680191,0.237414211,0.2966210544,0.0799449608,0.0015987612,0.2902956903,0.093101956,0.3097213209,-0.2249634415,-0.0230655484,-0.0278911963,0.048248969,0.1535485834,0.1837675571,0.047395207,0.0814307779,0.4353090525,0.1530235261,0.2819969356,-0.0381257124,0.1201000512,-0.1897055507,-0.4449408352,-0.2815672159,0.3054930568,-0.235494256,-0.0182448924,-0.3626365364,0.055864159,-0.1567477286,-0.1506158262,-0.0402768403,-0.1189904884,-0.2089492232,0.1465533823,-0.0952102095,-0.0301971864,-0.1931044161,0.102811411,0.1171641648,-0.1189524606,0.0187289137,0.0301830098,-0.2823504508,0.1210343465,-0.1132590547,0.0880358443,-0.0462888256,-0.2088495493,0.1870757788,0.2697921693,0.0533148795,0.0685409531,0.3685548902,0.4437269866,0.2311636508,-0.2230366319,0.192424655,-0.0318639688,-0.0763217807,-0.030568568,0.159598127,-0.1536246389,-0.0688809678,0.3970540166,0.0480838045,-0.189527005,-0.3199321032,0.0274739191,0.0225167349,-0.4600149393,0.2075293213,-0.0441925414,-0.0788600147,0.0065660863,-0.0245958772,-0.5514102578,0.127814576,0.3803643584,0.0252658669,0.2749297619,0.1281455904,0.0078003039,-0.2787654996,0.0311802514,0.4185286164,-0.0350203849,-0.5950027704,-0.3033539355,-0.4331485927,0.0840362385,-0.569270432,-0.292607069,0.001013095,0.248501271,0.0891625583,0.1719410121,0.3557214141,-0.105475843,0.0784786418,0.4126326144,-0.4025351107,-0.1663885862,-0.3254982531,0.2052407563,0.0800544843,-0.7052715421,0.0650318116,0.3884374797,0.0101613589,0.1391600668,-0.0993559957,0.3541438282,0.0327954255,0.1194630712,-0.1563095003,0.2681676447,0.0471640974,-0.0312050562,0.1960101873,0.0786783695,-0.1172625795,0.0075382129,0.0786147416,0.5020301342,-0.158609584,-0.160935536,-0.277139008,0.2690277398,0.1713728905,-0.1698029339,0.124161534,0.2329676896,-0.0899593756,0.227147907,0.2715693414,0.3370776772,-0.0686973631,0.0453815237,-0.3402681351,-0.0994826332,0.583381772,-0.3087985814,-0.2146814317,-0.1653167605,0.116828151,-0.053248167,0.2518343329,-0.4316256344,-0.0399834663,0.3314810097,0.064550586,-0.1202179641,-0.0008699126,0.0577102937,0.0541665554,-0.0688194558,0.2568713427,-0.0468433127,-0.2186792493,-0.1020397916,-0.2438035309]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/712","title":"Error in the notebooks\/Overview.ipynb notebook","comments":"Do this:\r\n``` python\r\nsquad_dataset = list_datasets(with_details=True)[datasets.index('squad')]\r\npprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n```","body":"Hi,\r\n\r\nI got the following error in **cell number 3** while exploring the **Overview.ipynb** notebook in google colab. I used the [link ](https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb) provided in the main README file to open it in colab. \r\n\r\n```python\r\n# You can access various attributes of the datasets before downloading them\r\nsquad_dataset = list_datasets()[datasets.index('squad')]\r\n\r\npprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n```\r\n\r\nError message\r\n```\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-8dc805c4949c> in <module>()\r\n 2 squad_dataset = list_datasets()[datasets.index('squad')]\r\n 3 \r\n ----> 4 pprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n \r\nAttributeError: 'str' object has no attribute '__dict__'\r\n```\r\n\r\nThe object `squad_dataset` is a `str` not a `dataclass` .","comment_length":15,"text":"Error in the notebooks\/Overview.ipynb notebook\nHi,\r\n\r\nI got the following error in **cell number 3** while exploring the **Overview.ipynb** notebook in google colab. I used the [link ](https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb) provided in the main README file to open it in colab. \r\n\r\n```python\r\n# You can access various attributes of the datasets before downloading them\r\nsquad_dataset = list_datasets()[datasets.index('squad')]\r\n\r\npprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n```\r\n\r\nError message\r\n```\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-8dc805c4949c> in <module>()\r\n 2 squad_dataset = list_datasets()[datasets.index('squad')]\r\n 3 \r\n ----> 4 pprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n \r\nAttributeError: 'str' object has no attribute '__dict__'\r\n```\r\n\r\nThe object `squad_dataset` is a `str` not a `dataclass` .\nDo this:\r\n``` python\r\nsquad_dataset = list_datasets(with_details=True)[datasets.index('squad')]\r\npprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n```","embeddings":[-0.2680433691,-0.1624178737,0.0057080346,0.2799274027,0.1056714728,0.2901156545,0.4072164595,0.1146837324,-0.2850091755,0.2991931736,-0.2500123978,0.0650306493,-0.1404008865,0.3564788699,-0.1135938242,-0.2361594588,-0.0553790517,0.1900392175,0.1701914072,0.1000705957,-0.3340093791,0.3672005534,-0.3910942376,0.1765102893,-0.1706778109,-0.1613949537,-0.2402285337,-0.008894545,-0.2815186977,-0.3519656956,0.4908081293,0.1682859808,0.2882230282,0.1846988946,-0.0001186387,-0.1437991112,0.2313589156,0.0259393789,-0.4177237749,-0.2631405592,-0.3337314725,0.1955717355,0.4131062031,-0.4264855087,0.0406544209,0.1536474377,0.2174620926,0.20104976,0.0396129452,0.1690451354,0.1968215555,0.1096186712,0.1222046837,-0.2075660229,0.2651115358,0.0014240976,-0.1841227114,0.5570950508,0.1321611702,-0.0184438452,0.084775582,0.3262999356,-0.1288239211,0.3469691575,0.3461412787,0.1949061602,-0.106752485,-0.4708770514,0.1931686103,0.0890839994,0.3395051062,-0.2843323946,-0.2140650451,-0.0499199964,0.2775289118,-0.4328393936,-0.0025660386,0.1433969885,-0.0905105993,0.1429048479,-0.0587709323,-0.0315349028,-0.201241076,0.1499438286,-0.1740650535,0.0952292904,-0.2355967611,0.0439464077,0.1605493724,-0.0812851191,0.1328320354,0.084228605,-0.0634953231,0.0995640084,-0.1119460389,-0.1084662676,0.1217441335,0.2231748104,-0.0211918298,-0.0111756008,0.0971410796,-0.0971522182,0.2428537458,0.2357889265,0.0750366226,0.0634773746,0.3124435842,0.2323675007,0.4430889189,0.4049856663,-0.1138893515,-0.1448300779,-0.2141363621,0.0769702718,-0.0758639574,-0.0896096379,0.5129563808,-0.1647848189,-0.4963900149,0.1015359387,-0.4635719657,0.0636398345,0.054602135,0.2858685553,-0.152328521,-0.1873095632,0.3515741229,-0.0834875554,-0.3648814261,-0.0802448839,-0.0735398456,0.283171773,-0.2353118509,0.1417765468,0.0598339662,0.0439876765,0.1056997925,-0.1330519021,0.3888125718,-0.0231856741,0.0917182937,-0.2827881575,-0.1918691695,0.2702095509,0.0696490407,-0.0634803101,0.1260766238,-0.4340178967,-0.0948853716,0.1770343035,-0.2880666554,-0.4469555318,-0.2491353303,0.104183279,-0.3000849187,0.1708068252,-0.7285445929,0.0945369452,-0.1119123995,-0.1812943071,-0.1398629248,-0.3296223283,-0.2246630937,-0.054172609,0.1804208755,0.3580517769,-0.0352263935,-0.2884978354,-0.0337338857,-0.0603192598,-0.049087733,0.0678471252,-0.1399157941,0.2216641158,-0.1868937463,-0.0091469241,-0.1042655706,-0.084740378,-0.5290681124,-0.0980416089,0.0235238522,-0.1184472069,-0.0355644636,0.1707216054,0.428248316,-0.0456840098,0.2019578218,0.1759639978,0.0051796269,0.0013563985,-0.1519694924,0.0248671714,-0.109906286,0.212497741,0.0735141188,0.1547418535,-0.0791203827,0.465932548,0.0108987642,-0.1637555063,-0.1997836232,0.1005726755,0.4894437492,-0.0820248649,0.0474484526,-0.1341384202,-0.5307047963,-0.0178963169,-0.072453633,0.3959489763,-0.139690429,-0.305118829,-0.113137424,0.0237655193,-0.2772391737,-0.1911684126,0.0858863145,-0.000569904,-0.2817528248,0.0758074448,-0.0012496948,0.3564808071,-0.4302850962,0.2019212395,-0.2638512254,0.401825577,-0.286195159,-0.0436715148,-0.0822878703,0.1484449506,-0.1577308774,-0.0911605507,-0.1315301061,0.0958798081,0.2473683059,-0.1269755065,0.355309546,0.3423580229,0.280144304,-0.4387547374,-0.043503277,0.0559467077,0.1610783786,-0.0229150932,0.0214926358,0.5706778169,0.1465102434,0.1678668708,0.0685209632,0.2316018641,0.0849063024,0.0138206286,-0.2052835524,-0.0483931527,0.1654042751,0.3209697008,0.1427537203,-0.0484020151,-0.5993828177,0.0734367743,0.1319135129,-0.0261154138,0.0088003967,0.2466322929,-0.2354722321,0.0221447088,0.3418874741,-0.0977244526,0.2555797398,0.0470569022,-0.196243301,0.3231500089,0.1189130098,0.0405603871,0.0883290172,0.0982785523,0.041908551,-0.2259611189,0.1929256469,0.2676983774,-0.1872539669,-0.4457127452,-0.1428239793,0.4448599219,-0.2659401894,0.3601466715,0.102452226,-0.4525662959,0.1166955754,-0.4259652793,-0.0240261871,-0.3914908171,-0.1269020587,0.5747728944,0.1097899526,0.1173803508,0.2175751179,0.3346266747,0.1861429065,0.0591771938,0.1993421763,-0.2130044699,-0.146005407,0.0366370492,0.3140764832,0.050988324,0.1050368398,-0.6140634418,-0.0197883192,-0.0385726206,-0.4488918185,0.268299073,-0.1640116423,0.2577382922,0.1871744245,0.2099899948,-0.7529500723,-0.22980313,0.3059401512,-0.0376649126,-0.2324609458,0.1190868244,-0.1549127698,-0.1829179972,0.0899532661,-0.2882977724,-0.4302333593,-0.1131294668,0.1976495832,-0.0859065354,0.2230022848,0.3915525377,0.0277271774,0.2863467634,0.1266853809,-0.0855448991,-0.0821151733,-0.1945269257,0.2196407467,-0.2291409969,-0.3775385022,0.0392293073,-0.1950063556,-0.0072278301,-0.1822704226,-0.5103940964,-0.3671375811,-0.122978963,0.313549459,-0.0497242771,-0.0816242769,0.1563565731,-0.0511809252,0.0318393372,0.1398670077,-0.3180902898,-0.1953433901,0.1001820043,0.3249399364,0.0017015204,0.224160403,-0.1860172749,0.5411051512,0.1028810516,-0.0359829254,0.0695795193,-0.0453116484,0.1569643617,-0.2114745975,-0.5914644599,0.1970109493,-0.1381000727,-0.3979639709,0.2297605425,-0.1665318757,-0.0795570761,-0.2628512084,0.1618176848,-0.1922082007,-0.3569773734,0.0852901191,0.1166200712,0.1483758539,-0.0892480165,-0.0057441341,-0.1952414066,-0.0886950418,-0.2610430419,0.5911771059,-0.091869399,-0.0214480534,0.1323955655,-0.5692622662,-0.4948889911,0.4468664229,0.1703414917,0.1428690255,-0.0721523464,0.224242568,0.2004028559,0.1086162254,0.0683004558,0.2909052372,-0.1129506826,0.2354920357,-0.4176886082,-0.0911087692,-0.2939256728,-0.4655536711,0.529384017,0.3720712364,0.1062611565,-0.1365952641,-0.0097650383,0.573844552,-0.1389477104,-0.0916603133,0.0966088176,-0.2874094546,-0.3922017217,-0.1987729371,0.0892160684,0.1517078429,0.2342650145,0.0489914156,0.2229000181,0.1575094014,0.0435913615,0.2172969282,0.3824620843,0.2098367959,0.1929333359,0.0575739928,-0.1584874392,-0.349472791,0.7139824033,0.4645614922,0.144123286,-0.4969211221,0.31544891,-0.1728610843,0.1697321981,0.0845355764,-0.226408869,-0.0118468376,-0.2608957589,0.2563427687,-0.0606618971,0.0684362501,0.3001720011,0.1774365008,-0.2723436356,-0.3452796638,0.4646578431,-0.0106101455,-0.0182178281,0.2829066217,0.0248022769,-0.0549458042,0.2821848392,0.0408880971,0.9024171233,0.0645119846,0.0284949727,0.1929081678,0.2728843093,0.3773007393,0.3931041062,0.3450742066,-0.3494737744,-0.2869944274,-0.1521319747,-0.180066973,0.0884026811,-0.216887787,-0.3517301977,0.0275155194,-0.0837276727,0.1566710621,0.0940631405,0.0500561893,-0.0159842931,0.0590103827,-0.3776443899,0.1115469858,-0.375056237,0.4265988767,-0.1083527207,0.1254327148,-0.3670600653,-0.0781564415,-0.5884280801,-0.0665231496,-0.130300045,0.2520198226,0.0622887239,-0.5302674174,0.1898329705,0.2397199273,0.0647180453,0.1100267991,-0.235809654,-0.2128587365,-0.2796963155,-0.2514392138,-0.1393909454,0.0161634963,0.2972866893,-0.1131438762,-0.1678958684,-0.0154138105,0.2299166769,-0.2607729435,0.1760595441,0.1149929017,0.200120613,-0.2405137569,-0.0061042709,-0.1636625081,-0.316161871,-0.229700163,0.0579216592,-0.0258397292,-0.199735716,0.1484617442,0.2773488462,-0.2335744798,0.0986726359,0.2935664654,0.1577779949,0.1107102185,0.4761755466,0.2053770125,-0.1986375004,-0.1115731299,0.2342332751,-0.1365652084,-0.1222183183,0.3233883679,0.0085981702,0.2407441884,-0.0790722668,0.5385704041,0.2694066167,0.2081065476,0.0015244349,-0.5426689982,-0.2820931971,0.3901743889,-0.1453262866,0.1676611602,0.1595192701,0.3155119717,0.2815088928,-0.2674945891,-0.2563422322,0.1806341112,-0.437087059,-0.0640645251,0.1904750913,-0.0554087684,0.0867994353,-0.1483875066,0.0073136482,0.0208700448,-0.0912166312,-0.1655988097,-0.0240269396,0.1615605056,0.0460752696,-0.0756019205,-0.051860068,0.0440785214,-0.0717514083,-0.0727590919,0.1863159388,0.3165102601,0.1193936169,0.2586029768,0.1162759364,0.0453382097,0.0807558522,0.1910562664,0.185588479,0.2228157073,0.1261791736,0.1831989735,0.0197480768,0.1389545798,0.2146682888,0.0720496252,-0.2869362235,-0.1504399478,0.1660373509,-0.3575297594,0.2767143548,-0.1285526603,0.0660817027,0.3950617015,-0.2022429258,0.0270218458,-0.004575788,0.0574726164,-0.2570908666,0.1495775729,0.4041148424,0.1624238491,-0.0664008558,0.3276006877,0.1113180965,0.3538653851,-0.0464223437,0.0474667028,0.2803732157,0.3593430817,0.4133222997,0.0183337703,-0.0806365237,0.3399491906,0.3725320101,0.1064107418,0.227258116,0.0667687878,-0.2028881311,0.0916994736,-0.1639574319,0.0988614857,0.4115161598,-0.3746778965,0.0359906778,0.0409786701,0.1612733603,-0.0843925327,-0.1017038748,0.6201440692,-0.4127743244,-0.0247708429,-0.0767311081,0.1286787242,-0.3219184279,0.0986487493,-0.2481276095,0.0152739016,-0.0480610505,-0.0361101069,-0.1940017641,-0.1827042401,0.5503336191,0.0065804413,-0.3892282844,-0.112789169,0.0053987871,0.2927092612,0.035471078,-0.0405700728,0.0932959542,0.0513207391,0.3335129619,-0.0237994809,0.8110225201,0.3740229905,0.2293677032,-0.2217999995,-0.1573830247,-0.0455614962,-0.0863502249,0.1110345349,-0.1982842237,0.1920984387,-0.0301301386,0.1099635065,0.098899357,-0.0875383541,0.4234429598,0.0647073835,-0.0278456062,0.0268988758,-0.1066753492,-0.1649829149,-0.0893630385,-0.1092354953,-0.0006791845,-0.2936024666,0.1458192617,0.1783524901,-0.0438204929,-0.1287208647,0.1657616198,0.010623008,0.0148266228,0.4214701355,0.6968566775,0.3276043832,-0.3690921366,-0.3808443248,-0.4658767581,0.2936625183,-0.0831911489,-0.0707500204,0.3097544014,0.0940103233,-0.1100137606,0.4130517244,0.2218168229,-0.0342633389,-0.2744422257,-0.0159663595,-0.4642702043,-0.2310316563,0.2206012309,-0.177477479,-0.121523656,-0.0392395109,0.0641866848,-0.1259896159,-0.0884254724,-0.2508864701,-0.0538548678,-0.1454885602,0.0109998835,0.2066849768,0.1236989126,0.5438165069,0.0670169517,-0.1376764774,0.3238222301,-0.3819396496,-0.04842728,0.367575109,0.1307536364,0.4540177882,0.0304153673,-0.3018222153,-0.1593368649,0.3684819043,0.1545033455,-0.1052933559,-0.5301445723,0.2924960256,-0.2869573832,0.1213328093,-0.1795976013,0.2327983528,-0.1241348907,0.2184194922,-0.4095369577,-0.5055328608,0.6645439863,-0.3631699681,-0.1893636137,-0.009533057,0.081398204,-0.3022824526,-0.1839391142,-0.5669856668,0.0801141709,0.1354445815,-0.1657749116,-0.4114905298,0.2612744868,-0.114575237,0.1225590035,-0.051665578,0.6925320625,-0.0167422816,-0.1436629146,-0.2550248504,-0.2705749869]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/712","title":"Error in the notebooks\/Overview.ipynb notebook","comments":"Thanks! This worked. I have created a PR to fix this in the notebook. ","body":"Hi,\r\n\r\nI got the following error in **cell number 3** while exploring the **Overview.ipynb** notebook in google colab. I used the [link ](https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb) provided in the main README file to open it in colab. \r\n\r\n```python\r\n# You can access various attributes of the datasets before downloading them\r\nsquad_dataset = list_datasets()[datasets.index('squad')]\r\n\r\npprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n```\r\n\r\nError message\r\n```\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-8dc805c4949c> in <module>()\r\n 2 squad_dataset = list_datasets()[datasets.index('squad')]\r\n 3 \r\n ----> 4 pprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n \r\nAttributeError: 'str' object has no attribute '__dict__'\r\n```\r\n\r\nThe object `squad_dataset` is a `str` not a `dataclass` .","comment_length":14,"text":"Error in the notebooks\/Overview.ipynb notebook\nHi,\r\n\r\nI got the following error in **cell number 3** while exploring the **Overview.ipynb** notebook in google colab. I used the [link ](https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb) provided in the main README file to open it in colab. \r\n\r\n```python\r\n# You can access various attributes of the datasets before downloading them\r\nsquad_dataset = list_datasets()[datasets.index('squad')]\r\n\r\npprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n```\r\n\r\nError message\r\n```\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-8dc805c4949c> in <module>()\r\n 2 squad_dataset = list_datasets()[datasets.index('squad')]\r\n 3 \r\n ----> 4 pprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n \r\nAttributeError: 'str' object has no attribute '__dict__'\r\n```\r\n\r\nThe object `squad_dataset` is a `str` not a `dataclass` .\nThanks! This worked. I have created a PR to fix this in the notebook. ","embeddings":[-0.270930469,-0.1667739153,0.0211873837,0.2973203063,0.1410933733,0.2470905185,0.3973854184,0.1097450629,-0.2881629467,0.3136608303,-0.2585085332,0.0930699185,-0.1493968219,0.4009942114,-0.0883977786,-0.2678328753,-0.0487496741,0.1776246876,0.1273774058,0.1111325249,-0.3454554677,0.4073553085,-0.3989681602,0.2037108392,-0.15288423,-0.1347142756,-0.2426262945,-0.0368740149,-0.2974142432,-0.3404968977,0.5058689713,0.1504085064,0.3084569275,0.1716338545,-0.0001185117,-0.1617465615,0.2194536179,0.0204981621,-0.4183129072,-0.2656316161,-0.2946822345,0.2121896595,0.434034735,-0.3997350633,0.0107030412,0.1181016266,0.1772234291,0.2562707663,0.0087522455,0.1836417019,0.2012922168,0.0767680407,0.1377631724,-0.1891687959,0.2763320208,0.0083338227,-0.1860652715,0.5503295064,0.1448304802,-0.0061604409,0.0790594742,0.3227718174,-0.1269728392,0.3200143576,0.3800161481,0.2021083981,-0.1059345528,-0.4415276945,0.1806607842,0.0910915062,0.3644813597,-0.2890426517,-0.1792014241,-0.046033904,0.2806011736,-0.424282521,0.0049188817,0.1416762769,-0.1011838391,0.1620668769,-0.0623333752,0.0090483623,-0.2079133391,0.1519640982,-0.1807147861,0.0621673279,-0.2343833148,0.0561950654,0.1653573364,-0.0860943794,0.1261062324,0.0844906941,-0.085326001,0.0727773905,-0.1169367731,-0.1083694324,0.1056106538,0.2446455806,-0.0286494158,-0.0063754595,0.1252856404,-0.0734497458,0.2627391517,0.2319093198,0.0585801713,0.1010458916,0.3260593414,0.2221671194,0.43383798,0.3870790899,-0.1393183619,-0.1483535171,-0.1849612594,0.0776706412,-0.0636769533,-0.0976769924,0.4729758501,-0.1677965075,-0.4722085893,0.0763178393,-0.4486215413,0.0683887675,0.0495142117,0.3009141386,-0.1404068172,-0.1751619726,0.3757564425,-0.053410016,-0.3574758172,-0.0860870183,-0.057840243,0.275573194,-0.257599473,0.154520601,0.0538943373,0.0846694037,0.094255656,-0.1075522155,0.3705225289,-0.0418016762,0.0951529071,-0.2657516301,-0.1925259084,0.2624136806,0.0562211126,-0.0872476175,0.1263485551,-0.4123106301,-0.0957977027,0.2137513012,-0.2628417015,-0.4416675866,-0.2556983232,0.1024007052,-0.2830074728,0.1690697074,-0.7398453355,0.1132219806,-0.1152189076,-0.1714206636,-0.144604966,-0.3594257236,-0.2101719826,-0.0476947054,0.1662577391,0.3064883351,-0.0253558774,-0.3012582064,-0.0395722836,-0.0651265606,-0.0251944959,0.0590853766,-0.1258510798,0.2229011655,-0.1902214736,-0.0404840484,-0.1007223949,-0.0579410642,-0.5573680997,-0.1149858758,0.0168263558,-0.1207929552,-0.0177816041,0.1846337914,0.4527177513,-0.0442931913,0.2009233087,0.1655567139,0.0361241437,0.016640855,-0.1560817659,0.0522623844,-0.1451889575,0.1938721836,0.0665834546,0.1774530411,-0.0857876241,0.4238735437,0.0167625304,-0.1737090498,-0.191334933,0.0910271034,0.4840039015,-0.0848592669,0.0749718472,-0.1302783489,-0.5194602609,-0.0149873737,-0.0771147534,0.3989536464,-0.1061048433,-0.3118350208,-0.1082766876,0.028894823,-0.268199563,-0.2229119688,0.0792467296,-0.0152423698,-0.2789445221,0.0818037912,0.0211563166,0.3589418828,-0.4287518859,0.1999123842,-0.2265112251,0.380209595,-0.275723815,-0.0553507023,-0.0871134251,0.1282591224,-0.1438953131,-0.0731389448,-0.1668476313,0.0817962512,0.2856699824,-0.1570422649,0.3240010142,0.3493815362,0.2778704464,-0.4468308091,-0.0326773673,0.0533654988,0.1557855159,-0.048127111,0.029815238,0.6075223088,0.153249979,0.1698352844,0.0582216829,0.2513449192,0.100116156,0.0018508786,-0.1944767833,-0.0474026911,0.1448477507,0.2917172313,0.1379937828,-0.0470897518,-0.6045613885,0.1073669121,0.133743614,0.0019791855,0.0019884477,0.2708523571,-0.2778955996,0.0235560704,0.3681385517,-0.1046715304,0.2718116939,0.0501519404,-0.221997261,0.3388575613,0.1493779272,0.012199969,0.1157260314,0.0963022709,-0.0018015968,-0.2401884794,0.1908581108,0.253501296,-0.1765674651,-0.3803108037,-0.1501056999,0.4352173507,-0.2706015706,0.36913234,0.0850111991,-0.4967216551,0.0958659723,-0.4260323942,0.0140681872,-0.3825870752,-0.1229045838,0.5801084042,0.1315604299,0.1268530786,0.2733921409,0.3275909126,0.1646072268,0.0684550852,0.1850305647,-0.1790891737,-0.1883226931,0.0352635011,0.30385077,0.0639656857,0.1011504754,-0.6036487818,-0.0127085848,-0.0683145747,-0.450547576,0.2691711187,-0.1578437537,0.253667146,0.1643698364,0.1689088345,-0.7538238168,-0.1972087026,0.2999323308,-0.0564218462,-0.2320850044,0.1623414606,-0.1534205824,-0.1599992365,0.1000580415,-0.2786118686,-0.438177377,-0.0924907923,0.2107614279,-0.0821261778,0.2223611623,0.339510411,0.0101453317,0.255471915,0.1340217441,-0.1075526401,-0.0720693395,-0.151478529,0.2061123401,-0.2204024345,-0.3829008937,0.0423497632,-0.1764658093,-0.0046123005,-0.1765857041,-0.4933642149,-0.3251665533,-0.1232387573,0.3248824179,-0.0101135056,-0.1082609296,0.1791148484,-0.0313624516,0.043699313,0.1430506259,-0.3167512119,-0.1727762371,0.0781061277,0.3243124485,0.0214569643,0.2312369347,-0.1931480914,0.5219928026,0.1132162809,-0.0329450034,0.0710314438,-0.0321002044,0.1423357427,-0.2465922385,-0.6022719145,0.1946360767,-0.1608224511,-0.3940148354,0.262686044,-0.161374256,-0.0890003443,-0.2681726813,0.1659751981,-0.1808676571,-0.3569934368,0.112445198,0.1158666238,0.1267744452,-0.095348306,-0.0039674076,-0.1777112484,-0.129415229,-0.2598792613,0.5796514153,-0.1171306819,-0.0102184946,0.1248733774,-0.5551859736,-0.5046379566,0.4659616947,0.1363296509,0.1327106953,-0.0625401288,0.2162555754,0.2164950967,0.1371911019,0.0545272976,0.2948690355,-0.1343552321,0.2173094451,-0.4540901482,-0.0705878437,-0.2973164022,-0.4497696757,0.5143113732,0.3962590992,0.1072687358,-0.1270727217,-0.0021027126,0.5705778003,-0.154373616,-0.1032789797,0.0685631335,-0.2651600242,-0.3707567751,-0.1871915758,0.071819514,0.1491259336,0.2224368453,0.0234897323,0.1801182032,0.1484892666,0.0292511955,0.181476891,0.372476548,0.223598659,0.1375333816,0.0807725489,-0.1862015277,-0.3375489116,0.7604062557,0.4432624578,0.1400884986,-0.516569078,0.3147984743,-0.2135024369,0.1624967307,0.0894594342,-0.2334130704,0.005905374,-0.2915703356,0.2339954376,-0.0725532621,0.043491926,0.2837259173,0.2039369345,-0.2519575357,-0.364665091,0.44737342,-0.009719884,-0.0332737006,0.2692568302,-0.0129842712,-0.0844917819,0.3016189933,0.0411354005,0.913112998,0.0439756624,0.0118644182,0.1632259488,0.2658332586,0.3516104817,0.4249352217,0.3413000405,-0.3987880349,-0.2872536778,-0.15217641,-0.1688665003,0.0934722126,-0.2255028039,-0.382260859,0.0277762823,-0.0755121261,0.1592479497,0.1124096662,0.0421201363,-0.0256585237,0.0501437783,-0.3361033499,0.1063807085,-0.3889293969,0.4352809489,-0.0967839882,0.1258522719,-0.3661421835,-0.0694668368,-0.6155385971,-0.0753550306,-0.1396891773,0.2223044187,0.0657755956,-0.5587177277,0.1665498912,0.2718721628,0.0604533963,0.1166683584,-0.2194302827,-0.2018918544,-0.2605550885,-0.2280198932,-0.1176263317,0.0079856897,0.2915400565,-0.1155446768,-0.1756666154,-0.0531185977,0.2619985342,-0.2806656361,0.1660973132,0.0781501532,0.2317319065,-0.2402873784,-0.0139818098,-0.2124376893,-0.3221185803,-0.2666913867,0.0596001819,-0.0498164035,-0.1972203702,0.1640742421,0.2764459252,-0.2160393596,0.1018853709,0.2794586718,0.1628874838,0.1101850495,0.4423402846,0.2286207974,-0.1892180592,-0.1094829291,0.2742224038,-0.1263520122,-0.0969656706,0.3325920999,0.0192970652,0.2447302788,-0.0536528751,0.5748010874,0.2682849169,0.2037184238,-0.0190459639,-0.5321005583,-0.2849088907,0.3553669155,-0.1385358274,0.1389624625,0.1529977471,0.3002109826,0.2885152698,-0.2477992922,-0.2559144497,0.1302964389,-0.4166849554,-0.0837080553,0.1977601349,-0.0389571674,0.0527720153,-0.1217178702,-0.0080897631,0.047534693,-0.0718201026,-0.16222848,0.0007717088,0.1630601883,0.0565468557,-0.0455884412,-0.0837821513,0.0418861136,-0.0764463022,-0.0940982997,0.1915303022,0.31864959,0.1008209512,0.2645284832,0.1437436342,0.0332962982,0.0867701843,0.2080909014,0.1631020904,0.2157804966,0.1040193364,0.1661801785,0.0382035598,0.1274047196,0.2375237048,0.061385788,-0.3057027459,-0.1105992198,0.1711522192,-0.340857774,0.2600957751,-0.1495318264,0.0375181064,0.392272681,-0.2082165033,0.0097793536,0.0104601355,0.0488296859,-0.2452409267,0.1422697604,0.4010728896,0.1789648533,-0.0743131638,0.344742626,0.1297760606,0.3132273257,-0.0571474284,0.0418851189,0.2915983796,0.3508552611,0.3779059052,0.0177083947,-0.0528208651,0.3602612615,0.4090751708,0.130143851,0.2151070237,0.0706430972,-0.1877745092,0.0540827401,-0.1746215373,0.0918059647,0.4187891781,-0.3737809062,0.0323903002,0.0304902904,0.1338222623,-0.0984511822,-0.0764590278,0.5935812593,-0.395186007,-0.0461310074,-0.0859396383,0.1298427284,-0.3327770531,0.1137521341,-0.2831107676,0.0192352533,-0.0256266575,-0.0366127901,-0.2145611346,-0.1634477079,0.5673809052,0.0111440336,-0.3946866989,-0.1227046326,-0.0411480479,0.2837224603,0.0160519835,-0.0276303962,0.0825928077,0.0034854589,0.38932845,-0.012874661,0.7882614136,0.3325275183,0.2531560361,-0.2101394385,-0.1845037639,-0.0457822941,-0.0812522992,0.1033669785,-0.1798466444,0.2010289431,-0.0251585338,0.0883645937,0.0815819651,-0.0792253315,0.4392504692,0.0644623712,-0.0095170001,0.0313556679,-0.0821871832,-0.1517181396,-0.0982890651,-0.1130388454,0.0014051137,-0.2937871516,0.1572235078,0.1767118424,-0.0298029725,-0.1228393167,0.1625349671,0.0115085272,0.0426121019,0.4082304835,0.667522788,0.3122265041,-0.3599291742,-0.395676434,-0.4652534425,0.2966089845,-0.0774477646,-0.0532648601,0.3142609,0.0503375418,-0.0916436017,0.4129706919,0.2264231443,-0.0059880107,-0.2697605491,0.0003434708,-0.4813435972,-0.2437752783,0.2439744323,-0.1807579398,-0.1419511586,-0.0112962024,0.0783867612,-0.1063896716,-0.0845295787,-0.264700979,-0.0342514925,-0.1343003064,0.0113397967,0.2143099755,0.1299092323,0.5411167741,0.0662499666,-0.0942865387,0.3903933764,-0.3790248036,-0.0596658066,0.3528016806,0.112555854,0.448194325,0.0488496646,-0.3138788342,-0.1702610552,0.3521558046,0.1613577157,-0.1250025332,-0.5275723934,0.291346997,-0.2855599821,0.1263653487,-0.1786051691,0.2519372702,-0.1180200651,0.21032767,-0.4187256694,-0.5308037996,0.6453858614,-0.3496780694,-0.187312454,-0.0021525451,0.0727198645,-0.343190968,-0.1957636923,-0.5364913344,0.07067018,0.1289410293,-0.1637672186,-0.3913559616,0.271876663,-0.1155463532,0.1596179456,-0.0733094066,0.7147974372,-0.0013397041,-0.1138385907,-0.2748626173,-0.2732389569]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/709","title":"How to use similarity settings other then \"BM25\" in Elasticsearch index ?","comments":"Datasets does not use elasticsearch API to define custom similarity. If you want to use a custom similarity, the best would be to run a curl request directly to your elasticsearch instance (see sample hereafter, directly from ES documentation), then you should be able to use `my_similarity` in your configuration passed to datasets\r\n\r\n```\r\ncurl -X PUT \"localhost:9200\/index?pretty\" -H 'Content-Type: application\/json' -d'\r\n{\r\n \"settings\": {\r\n \"index\": {\r\n \"similarity\": {\r\n \"my_similarity\": {\r\n \"type\": \"DFR\",\r\n \"basic_model\": \"g\",\r\n \"after_effect\": \"l\",\r\n \"normalization\": \"h2\",\r\n \"normalization.h2.c\": \"3.0\"\r\n }\r\n }\r\n }\r\n }\r\n}\r\n'\r\n\r\n```","body":"**QUESTION : How should we use other similarity algorithms supported by Elasticsearch other than \"BM25\" ?**\r\n**ES Reference**\r\nhttps:\/\/www.elastic.co\/guide\/en\/elasticsearch\/reference\/current\/index-modules-similarity.html\r\n**HF doc reference:**\r\nhttps:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html\r\n\r\n**context :**\r\n========\r\n\r\nI used the latest Elasticsearch server version 7.9.2\r\nWhen I set DFR which is one of the other similarity algorithms supported by elasticsearch in the mapping, I get an error\r\n\r\nFor example DFR that I had tried in the first instance in mappings as below.,\r\n`\"mappings\": {\"properties\": {\"text\": {\"type\": \"text\", \"analyzer\": \"standard\", \"similarity\": \"DFR\"}}},`\r\n\r\nI get the following error \r\nRequestError: RequestError(400, 'mapper_parsing_exception', 'Unknown Similarity type [DFR] for field [text]')\r\n\r\nThe other thing as another option I had tried was to declare \"similarity\": \"my_similarity\" within settings and then assigning \"my_similarity\" inside the mappings as below \r\n\r\n`es_config = {\r\n \"settings\": {\r\n \"number_of_shards\": 1,\r\n **\"similarity\": \"my_similarity\"**: {\r\n \"type\": \"DFR\",\r\n \"basic_model\": \"g\",\r\n \"after_effect\": \"l\",\r\n \"normalization\": \"h2\",\r\n \"normalization.h2.c\": \"3.0\"\r\n } ,\r\n \"analysis\": {\"analyzer\": {\"stop_standard\": {\"type\": \"standard\", \" stopwords\": \"_english_\"}}},\r\n \r\n },\r\n \"mappings\": {\"properties\": {\"text\": {\"type\": \"text\", \"analyzer\": \"standard\", \"similarity\": \"my_similarity\"}}},\r\n }`\r\n\r\nFor this , I got the following error\r\nRequestError: RequestError(400, 'illegal_argument_exception', 'unknown setting [index.similarity] please check that any required plugins are installed, or check the breaking changes documentation for removed settings')\r\n\r\n","comment_length":88,"text":"How to use similarity settings other then \"BM25\" in Elasticsearch index ?\n**QUESTION : How should we use other similarity algorithms supported by Elasticsearch other than \"BM25\" ?**\r\n**ES Reference**\r\nhttps:\/\/www.elastic.co\/guide\/en\/elasticsearch\/reference\/current\/index-modules-similarity.html\r\n**HF doc reference:**\r\nhttps:\/\/huggingface.co\/docs\/datasets\/faiss_and_ea.html\r\n\r\n**context :**\r\n========\r\n\r\nI used the latest Elasticsearch server version 7.9.2\r\nWhen I set DFR which is one of the other similarity algorithms supported by elasticsearch in the mapping, I get an error\r\n\r\nFor example DFR that I had tried in the first instance in mappings as below.,\r\n`\"mappings\": {\"properties\": {\"text\": {\"type\": \"text\", \"analyzer\": \"standard\", \"similarity\": \"DFR\"}}},`\r\n\r\nI get the following error \r\nRequestError: RequestError(400, 'mapper_parsing_exception', 'Unknown Similarity type [DFR] for field [text]')\r\n\r\nThe other thing as another option I had tried was to declare \"similarity\": \"my_similarity\" within settings and then assigning \"my_similarity\" inside the mappings as below \r\n\r\n`es_config = {\r\n \"settings\": {\r\n \"number_of_shards\": 1,\r\n **\"similarity\": \"my_similarity\"**: {\r\n \"type\": \"DFR\",\r\n \"basic_model\": \"g\",\r\n \"after_effect\": \"l\",\r\n \"normalization\": \"h2\",\r\n \"normalization.h2.c\": \"3.0\"\r\n } ,\r\n \"analysis\": {\"analyzer\": {\"stop_standard\": {\"type\": \"standard\", \" stopwords\": \"_english_\"}}},\r\n \r\n },\r\n \"mappings\": {\"properties\": {\"text\": {\"type\": \"text\", \"analyzer\": \"standard\", \"similarity\": \"my_similarity\"}}},\r\n }`\r\n\r\nFor this , I got the following error\r\nRequestError: RequestError(400, 'illegal_argument_exception', 'unknown setting [index.similarity] please check that any required plugins are installed, or check the breaking changes documentation for removed settings')\r\n\r\n\nDatasets does not use elasticsearch API to define custom similarity. If you want to use a custom similarity, the best would be to run a curl request directly to your elasticsearch instance (see sample hereafter, directly from ES documentation), then you should be able to use `my_similarity` in your configuration passed to datasets\r\n\r\n```\r\ncurl -X PUT \"localhost:9200\/index?pretty\" -H 'Content-Type: application\/json' -d'\r\n{\r\n \"settings\": {\r\n \"index\": {\r\n \"similarity\": {\r\n \"my_similarity\": {\r\n \"type\": \"DFR\",\r\n \"basic_model\": \"g\",\r\n \"after_effect\": \"l\",\r\n \"normalization\": \"h2\",\r\n \"normalization.h2.c\": \"3.0\"\r\n }\r\n }\r\n }\r\n }\r\n}\r\n'\r\n\r\n```","embeddings":[-0.1099796966,-0.6802694798,-0.0450149104,-0.0399373472,-0.1662632376,0.0516545996,-0.1838181019,0.2227956951,0.445256263,0.1382717043,-0.3413906097,-0.1468974054,0.0635201782,-0.2089974731,-0.3592832088,-0.0928407833,0.0289196949,0.0444094092,0.104161717,0.1085229665,-0.1743733436,0.1254539192,-0.000062598,-0.0005490784,-0.1480967999,0.187689364,0.0392793305,-0.2065735906,0.1528002173,-0.3148995042,0.1453121305,-0.2109477669,0.1577213705,0.044019714,-0.0001227613,0.0690059215,0.369327873,-0.0321302786,0.4254214466,-0.4718561471,-0.0608858466,0.216519013,0.1834836155,-0.177677691,-0.2151821107,-0.3694826066,0.0328343771,-0.4300846457,0.2202268839,0.3700291514,0.1151460186,-0.1253225058,0.178447336,-0.0453588031,0.1369128078,0.0639811084,0.0044003711,-0.2802915871,0.3826874793,0.4209572971,0.3343570828,-0.0454553254,-0.2023601979,0.3465203643,0.3322952092,0.1597774178,0.3609420955,-0.2127847522,0.2631377578,0.1443555504,0.1755000949,0.1586993188,0.1121972054,0.0569028407,-0.1709520221,-0.0436668098,0.0706234202,-0.2081927657,0.147807315,0.0993624851,-0.1951645613,-0.1788298786,-0.1105849594,0.2574418783,-0.0947720185,0.1925845742,-0.0943944752,-0.060371127,0.256309092,-0.4657059312,-0.1320724636,0.0108752036,0.1735929549,0.1292192787,-0.0163315777,-0.2779355049,-0.3406863213,-0.1888903081,0.3027116358,-0.1236504167,-0.075883396,0.1966890097,0.1552668363,0.2615856826,-0.3191659153,0.2049432248,0.066407159,0.4087871909,0.0873084664,-0.2203004509,-0.1675471514,0.0207611546,0.4790687561,-0.0699595362,0.0530895814,-0.4394777119,-0.2394437939,-0.4781885147,0.1970800012,-0.1506071687,-0.2655495703,-0.2916416824,0.1298919767,0.1182668284,0.0235036481,-0.4387791157,0.0350998938,0.086308904,0.0781803951,-0.0199933443,-0.0445248969,-0.0094926823,0.0087711113,0.3264604509,0.3060792685,-0.4353673458,0.5557736754,0.118113786,0.0581613109,0.3656066954,0.2970832288,0.0120175947,0.3168718219,-0.0040030503,-0.2208718807,0.0614858866,0.181742534,-0.0633637607,-0.3740391731,-0.0667839721,-0.5649577379,-0.2988860309,-0.2933887541,0.1480261087,-0.3074215949,-0.2548411787,-0.2832580209,0.2734139264,-0.2297459394,-0.1462838948,0.0989792868,-0.0884415507,-0.2237434983,0.0443157107,0.0850663036,-0.1622759551,-0.0286308713,-0.309638828,-0.0529246293,0.4652359486,0.1320650578,0.1934707612,-0.1054061055,-0.6684598923,0.0378974527,0.5693466663,0.5097745657,0.0055426504,-0.1204319596,0.1143238619,0.2543828189,-0.0709597841,0.2039704621,-0.1388183385,0.200619638,0.1326102912,0.1450300664,0.2731208801,0.2092545331,-0.2835346162,-0.3051760793,-0.2141604125,-0.0925754905,0.0189207196,-0.0333237126,0.0647436008,0.3815340698,-0.1603909731,-0.3229284883,-0.1779347062,-0.1087432355,0.0910640359,0.177333802,0.2237422019,0.2579643428,-0.1902689487,0.259560734,0.176634118,-0.3559484184,0.0818277076,0.2356043905,0.1936783195,-0.0839837566,-0.3139104247,-0.0714864954,-0.076721184,0.0441464931,-0.219810918,-0.1627634317,-0.0379004888,-0.1630102098,-0.255543828,-0.0120575782,-0.0245708581,-0.175784722,0.5014539361,-0.2597070634,0.032013379,-0.0591114424,-0.1223389208,0.2278618217,-0.0852190033,-0.1969066113,0.0806867406,-0.1826835275,0.3506126702,0.3075753152,0.7557079792,0.0849625394,-0.4385830462,0.3527544439,-0.008834363,-0.2499325126,0.0982048661,0.0114404131,-0.038631659,0.2910941541,0.3620179892,0.0191407558,-0.2755416036,-0.0913378075,-0.0194800962,0.241254732,-0.3768110871,0.2612398267,-0.3031679094,-0.1589258015,0.1838589758,0.1624084562,-0.161703825,0.0292062797,0.1390403211,0.2721504271,0.2021351606,0.022936523,-0.1996165365,-0.0208406579,-0.1506257057,0.2454142272,0.1411077678,-0.074645862,0.1552448422,-0.1430294216,-0.1715416163,0.3471221626,0.1294501126,-0.028844336,-0.0878201574,0.0389396362,0.0593524985,0.0525161289,-0.387117058,-0.0891268551,-0.102802895,-0.1088057458,0.1004484147,-0.5472633839,0.1111528501,-0.1829553545,0.4247772992,-0.2955296934,-0.036486432,0.0299583022,0.179770276,-0.0549660586,0.1705628783,-0.2438575327,-0.1432878375,-0.1753666252,-0.1681888252,-0.2786498368,0.2655235529,-0.2386914045,-0.0093539348,0.0260451511,0.3414477408,0.0146750361,0.0368778296,0.0348646119,-0.3051835001,-0.6715813875,0.1528566778,0.3799234927,0.2668421566,0.2583189309,-0.3530448675,-0.4748586714,-0.2644218206,0.3256278634,-0.282907933,0.1872529984,-0.0278585479,-0.3894019127,0.3343360126,0.3319964409,-0.294306308,-0.095977217,-0.2260111719,-0.1961698383,0.0739624947,0.1944969147,0.3149732649,-0.1686628014,0.0810831338,0.0829168931,0.3201363981,-0.1368821263,0.1280291229,0.5260477662,-0.1801885217,0.2626565099,0.0549673811,-0.0480560251,0.2496026605,0.394746244,-0.0880166888,-0.0321652628,-0.005675585,-0.417722702,0.0833182558,0.1446457058,-0.1467420906,0.0077060866,-0.085375607,0.1881073862,0.1078551039,0.0606455915,0.1781775355,0.3725463152,0.3807065189,0.1436383873,-0.0966634154,0.8414549232,-0.0144099146,0.0652125701,0.3936395347,0.1621803492,-0.2119904459,-0.1389683634,-0.1302311569,-0.218854174,0.3948901892,0.1391048878,0.1219386384,-0.012267516,-0.3806421459,-0.229127869,-0.4974321723,0.0169697665,0.1065993086,0.0358183682,0.2787443995,0.4080377221,-0.2741197348,-0.0735107437,0.0552526601,-0.3691664338,-0.1137173921,0.059699472,-0.2399495244,0.1703723073,-0.0230796877,-0.5380219817,-0.32050246,0.7414457202,0.0002173933,0.0987046286,-0.2876690626,0.1439871639,0.0832361504,0.3491668105,-0.0098141544,-0.2938861549,-0.5016282201,-0.3293353617,-0.4276208282,-0.3831458688,-0.1308456063,0.2402856946,-0.2232843339,-0.3592094779,0.3704482317,-0.0129618598,-0.1446431577,-0.0215710979,-0.2039801925,0.0451560318,-0.2999050915,0.0072280951,0.1388060898,-0.1072053909,-0.3052333295,0.0139533449,0.2452327609,0.1477886885,-0.0708150715,-0.067677699,-0.0388912372,0.2269310355,0.0728657246,0.0179732274,0.0661873668,0.0671498403,0.5369636416,0.2987402678,0.2468257844,-0.1220910475,0.0192238465,0.3615914881,-0.2456898093,0.1638257802,0.043209672,-0.2378498167,0.0613403171,0.3586055338,0.1583438069,-0.0166843627,0.1164001375,-0.3012368381,0.1999047548,0.2137165964,0.0041518244,-0.0577542894,0.67900002,0.3671955764,-0.0127790477,-0.0531664602,0.3668000102,-0.3741472661,0.3981563449,0.3430175781,0.3471108973,0.4064126611,-0.2098452747,-0.3018680811,0.1915673316,0.3319590986,-0.0766078308,0.1850986332,-0.182146281,-0.3758681417,-0.0414293222,-0.0942478329,0.3397330046,0.0652392805,-0.1956473738,0.4940738082,0.0631527826,-0.2016892731,-0.1147125661,0.3365665674,-0.0548303574,-0.0579649284,0.2160901427,0.0381141603,-0.1906488389,0.3520360887,0.334575206,0.1518594623,-0.1074482426,-0.0434568487,-0.0971069932,-0.1127913222,-0.0231549162,0.226640597,0.180139482,0.2071553916,0.779913187,-0.2633679509,0.3296678662,0.0258928761,-0.405015111,-0.0541692264,0.0409363285,-0.1658571362,-0.036997091,-0.0493085757,0.1772247106,-0.1595717072,-0.4186327457,0.1097158715,0.4122183025,-0.135384053,-0.6603272557,-0.4940025806,0.0222968403,0.185439229,-0.1221523657,-0.1672212332,-0.1612369865,0.237484619,0.0848484263,-0.2467111498,-0.1824184805,0.5444489717,0.2052455544,-0.174591884,-0.0993676558,0.5531528592,0.0559505783,-0.2092758864,0.2950088978,-0.2328965515,-0.2230932266,0.0890229642,-0.1984615773,0.4163675308,0.1171630993,-0.2509320378,-0.0408289135,-0.0475679263,0.1526893526,0.5386033058,0.4139128029,0.5027273297,-0.2389631271,0.1192825958,-0.0697953925,0.3097170591,0.1394025236,0.3073012829,-0.5134685636,-0.2494951189,-0.033646863,-0.0060045645,-0.1531669199,0.3232014775,0.0497691929,-0.1040933877,-0.1565631777,-0.0293832887,-0.0254745372,0.03224314,-0.0558070689,0.0541609563,-0.3288837671,0.0023267155,-0.0626846924,0.1406768113,0.0628935844,-0.2919207811,-0.2410839796,-0.422299087,-0.1617659032,-0.0557123236,-0.2245861888,0.3070334494,0.1734170169,-0.0448971093,-0.079356119,-0.5070895553,0.0743330568,0.0686031356,0.0944649875,0.0851660371,0.0325710699,-0.0201976225,-0.0556060374,-0.2246446311,-0.0915173367,0.1086319089,-0.2324387133,0.0110346032,0.1500189006,0.0877879485,-0.3217265904,0.7859103084,0.0040244483,0.0206365343,-0.0271261688,-0.2612791657,0.1330943853,0.1213181391,0.0456235111,-0.1868679076,0.4898707271,0.3075526655,0.0129378643,0.2696974576,0.5250288248,0.048668392,0.2212863415,0.005495551,0.3301386535,-0.3393138349,0.1181462556,-0.14236103,0.0951446965,-0.0422689989,0.2914741933,0.2757108212,0.4842646718,0.0848611593,-0.3516989946,0.2595898509,-0.2052451819,0.0762458295,0.7275319099,0.0867138654,0.3168497384,0.3144491613,0.2281218767,0.4571487904,0.051415056,0.2946278751,0.0566789955,-0.1594220549,0.0405904241,0.0062791905,0.119250752,-0.0850611851,0.0231246781,-0.3159181774,-0.4254609048,-0.1276066303,0.0082545439,-0.0437546149,-0.1752752364,0.0180729833,-0.1200730801,0.2199278325,0.0000328688,0.1514882743,0.5412801504,-0.0908547863,0.0245082323,-0.1568830609,-0.1649010181,0.4298416078,0.2256816924,-0.1486890018,0.1915638596,-0.1007814631,-0.0809649751,0.0556520671,-0.1301408559,0.3380600512,0.0261419304,0.1664376855,-0.6764463186,-0.0454053171,-0.0414452292,-0.0645261183,-0.2009960264,-0.3495067954,0.3897270858,-0.1165069044,-0.2713944912,0.1432805806,0.1515005678,0.3517342806,-0.112020582,-0.3581876457,-0.0862292498,0.0518590696,-0.080984205,-0.142534107,0.2655230761,0.0237308219,-0.2251168489,0.0217373427,0.1432458311,-0.0875028148,-0.2112697065,0.0929914862,-0.1047799736,0.1971467882,0.0840773955,0.4553633034,0.5093330145,0.0920788869,-0.0255699307,-0.2781482339,0.3214883506,-0.1484740078,-0.3624585569,-0.0977649838,0.1022541225,-0.1342019439,-0.2379291356,-0.1270836294,0.0090426365,0.3054428101,0.277174294,-0.283210516,-0.1311529428,-0.6396153569,0.1917271912,0.0363001004,0.0382275842,-0.0924960822,0.0709762052,0.1456525475,0.0485923588,0.285474956,0.4260379672,-0.1719560623,0.0081042647,0.2560974658,0.0248180386,0.1707031131,-0.050307557,-0.5536447763,-0.2471792102,0.1923092157,-0.1681049764,-0.2826262116,-0.1764518023,-0.2089059204,-0.0843344331,-0.2383701056,-0.1878714859,0.1503479481,0.1285378188,0.0168372039,-0.1015243456,-0.1317765266,0.4407918453,-0.021504987,0.1676945239,-0.1945523471,0.054125458,0.6103241444,0.1389135569,-0.1990364492,-0.3519845009,0.3735135198,0.0313636474,-0.0052880151,-0.024831973,0.3395010829,-0.0619784072,-0.2084068805,-0.1932700425,0.1773454845,0.0633519217,-0.3400342166,0.0889995918]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"Facing a similar issue here. My model using SQuAD dataset takes about 1h to process with in memory data and more than 2h with datasets directly.","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":26,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nFacing a similar issue here. My model using SQuAD dataset takes about 1h to process with in memory data and more than 2h with datasets directly.","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"And if you use in-memory-data with datasets with `load_dataset(..., keep_in_memory=True)`?","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":10,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nAnd if you use in-memory-data with datasets with `load_dataset(..., keep_in_memory=True)`?","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"Thanks for the tip @thomwolf ! I did not see that flag in the docs. I'll try with that.","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":19,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nThanks for the tip @thomwolf ! I did not see that flag in the docs. I'll try with that.","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"We should add it indeed and also maybe a specific section with all the tips for maximal speed. What do you think @lhoestq @SBrandeis @yjernite ?","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":26,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nWe should add it indeed and also maybe a specific section with all the tips for maximal speed. What do you think @lhoestq @SBrandeis @yjernite ?","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"By default the datasets loaded with `load_dataset` live on disk.\r\nIt's possible to load them in memory by using some transforms like `.map(..., keep_in_memory=True)`.\r\n\r\nSmall correction to @thomwolf 's comment above: currently we don't have the `keep_in_memory` parameter for `load_dataset` AFAIK but it would be nice to add it indeed :)","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":51,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nBy default the datasets loaded with `load_dataset` live on disk.\r\nIt's possible to load them in memory by using some transforms like `.map(..., keep_in_memory=True)`.\r\n\r\nSmall correction to @thomwolf 's comment above: currently we don't have the `keep_in_memory` parameter for `load_dataset` AFAIK but it would be nice to add it indeed :)","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"Yes indeed we should add it!","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":6,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nYes indeed we should add it!","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"Great! Thanks a lot.\r\n\r\nI did a test using `map(..., keep_in_memory=True)` and also a test using in-memory only data.\r\n\r\n```python\r\nfeatures = dataset.map(tokenize, batched=True, remove_columns=dataset['train'].column_names)\r\nfeatures.set_format(type='torch', columns=['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\nfeatures_in_memory = dataset.map(tokenize, batched=True, keep_in_memory=True, remove_columns=dataset['train'].column_names)\r\nfeatures_in_memory.set_format(type='torch', columns=['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\nin_memory = [features['train'][i] for i in range(len(features['train']))]\r\n```\r\n\r\nFor using the features without any tweak, I got **1min17s** for copying the entire DataLoader to CUDA:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(features['train'], batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nFor using the features mapped with `keep_in_memory=True`, I also got **1min17s** for copying the entire DataLoader to CUDA:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(features_in_memory['train'], batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nAnd for the case using every element in memory, converted from the original dataset, I got **12.5s**:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(in_memory, batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nTaking a closer look in my SQuAD code, using a profiler, I see a lot of calls to `posix read` api. It seems that it is really reliying on disk, which results in a very high train time.","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":170,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nGreat! Thanks a lot.\r\n\r\nI did a test using `map(..., keep_in_memory=True)` and also a test using in-memory only data.\r\n\r\n```python\r\nfeatures = dataset.map(tokenize, batched=True, remove_columns=dataset['train'].column_names)\r\nfeatures.set_format(type='torch', columns=['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\nfeatures_in_memory = dataset.map(tokenize, batched=True, keep_in_memory=True, remove_columns=dataset['train'].column_names)\r\nfeatures_in_memory.set_format(type='torch', columns=['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\nin_memory = [features['train'][i] for i in range(len(features['train']))]\r\n```\r\n\r\nFor using the features without any tweak, I got **1min17s** for copying the entire DataLoader to CUDA:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(features['train'], batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nFor using the features mapped with `keep_in_memory=True`, I also got **1min17s** for copying the entire DataLoader to CUDA:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(features_in_memory['train'], batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nAnd for the case using every element in memory, converted from the original dataset, I got **12.5s**:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(in_memory, batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nTaking a closer look in my SQuAD code, using a profiler, I see a lot of calls to `posix read` api. It seems that it is really reliying on disk, which results in a very high train time.","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"I am having the same issue here. When loading from memory I can get the GPU up to 70% util but when loading after mapping I can only get 40%.\r\n\r\nIn disk:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='\/home\/ad\/Desktop\/bookcorpus', split='train[:20%]')\r\nbook_corpus = book_corpus.map(encode, batched=True, num_proc=20, load_from_cache_file=True, batch_size=2500)\r\nbook_corpus.set_format(type='torch', columns=['text', \"input_ids\", \"attention_mask\", \"token_type_ids\"])\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/mobile_bert_big\",\r\n overwrite_output_dir=True,\r\n num_train_epochs=1,\r\n per_device_train_batch_size=32,\r\n per_device_eval_batch_size=16,\r\n save_steps=50,\r\n save_total_limit=2,\r\n logging_first_step=True,\r\n warmup_steps=100,\r\n logging_steps=50,\r\n eval_steps=100,\r\n no_cuda=False,\r\n gradient_accumulation_steps=16,\r\n fp16=True)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=book_corpus,\r\n tokenizer=tokenizer)\r\n```\r\n\r\nIn disk I can only get 0,17 it\/s:\r\n`[ 13\/28907 01:03 < 46:03:27, 0.17 it\/s, Epoch 0.00\/1] `\r\n\r\nIf I load it with torch.utils.data.Dataset()\r\n```\r\nclass BCorpusDataset(torch.utils.data.Dataset):\r\n def __init__(self, encodings):\r\n self.encodings = encodings\r\n\r\n def __getitem__(self, idx):\r\n item = [torch.tensor(val[idx]) for key, val in self.encodings.items()][0]\r\n return item\r\n\r\n def __len__(self):\r\n length = [len(val) for key, val in self.encodings.items()][0]\r\n return length\r\n\r\n**book_corpus = book_corpus.select([i for i in range(16*2000)])** # filtering to not have 20% of BC in memory...\r\nbook_corpus = book_corpus(book_corpus)\r\n```\r\nI can get:\r\n` [ 5\/62 00:09 < 03:03, 0.31 it\/s, Epoch 0.06\/1]`\r\n\r\nBut obviously I can not get BookCorpus in memory xD\r\n\r\nEDIT: it is something weird. If i load in disk 1% of bookcorpus:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='\/home\/ad\/Desktop\/bookcorpus', split='train[:1%]')\r\n```\r\n\r\nI can get 0.28 it\/s, (the same that in memory) but if I load 20% of bookcorpus:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='\/home\/ad\/Desktop\/bookcorpus', split='train[:20%]')\r\n```\r\nI get again 0.17 it\/s. \r\n\r\nI am missing something? I think it is something related to size, and not disk or in-memory.","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":247,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nI am having the same issue here. When loading from memory I can get the GPU up to 70% util but when loading after mapping I can only get 40%.\r\n\r\nIn disk:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='\/home\/ad\/Desktop\/bookcorpus', split='train[:20%]')\r\nbook_corpus = book_corpus.map(encode, batched=True, num_proc=20, load_from_cache_file=True, batch_size=2500)\r\nbook_corpus.set_format(type='torch', columns=['text', \"input_ids\", \"attention_mask\", \"token_type_ids\"])\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/mobile_bert_big\",\r\n overwrite_output_dir=True,\r\n num_train_epochs=1,\r\n per_device_train_batch_size=32,\r\n per_device_eval_batch_size=16,\r\n save_steps=50,\r\n save_total_limit=2,\r\n logging_first_step=True,\r\n warmup_steps=100,\r\n logging_steps=50,\r\n eval_steps=100,\r\n no_cuda=False,\r\n gradient_accumulation_steps=16,\r\n fp16=True)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=book_corpus,\r\n tokenizer=tokenizer)\r\n```\r\n\r\nIn disk I can only get 0,17 it\/s:\r\n`[ 13\/28907 01:03 < 46:03:27, 0.17 it\/s, Epoch 0.00\/1] `\r\n\r\nIf I load it with torch.utils.data.Dataset()\r\n```\r\nclass BCorpusDataset(torch.utils.data.Dataset):\r\n def __init__(self, encodings):\r\n self.encodings = encodings\r\n\r\n def __getitem__(self, idx):\r\n item = [torch.tensor(val[idx]) for key, val in self.encodings.items()][0]\r\n return item\r\n\r\n def __len__(self):\r\n length = [len(val) for key, val in self.encodings.items()][0]\r\n return length\r\n\r\n**book_corpus = book_corpus.select([i for i in range(16*2000)])** # filtering to not have 20% of BC in memory...\r\nbook_corpus = book_corpus(book_corpus)\r\n```\r\nI can get:\r\n` [ 5\/62 00:09 < 03:03, 0.31 it\/s, Epoch 0.06\/1]`\r\n\r\nBut obviously I can not get BookCorpus in memory xD\r\n\r\nEDIT: it is something weird. If i load in disk 1% of bookcorpus:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='\/home\/ad\/Desktop\/bookcorpus', split='train[:1%]')\r\n```\r\n\r\nI can get 0.28 it\/s, (the same that in memory) but if I load 20% of bookcorpus:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='\/home\/ad\/Desktop\/bookcorpus', split='train[:20%]')\r\n```\r\nI get again 0.17 it\/s. \r\n\r\nI am missing something? I think it is something related to size, and not disk or in-memory.","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"There is a way to increase the batches read from memory? or multiprocessed it? I think that one of two or it is reading with just 1 core o it is reading very small chunks from disk and left my GPU at 0 between batches","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":45,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nThere is a way to increase the batches read from memory? or multiprocessed it? I think that one of two or it is reading with just 1 core o it is reading very small chunks from disk and left my GPU at 0 between batches","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/708","title":"Datasets performance slow? - 6.4x slower than in memory dataset","comments":"My fault! I had not seen the `dataloader_num_workers` in `TrainingArguments` ! Now I can parallelize and go fast! Sorry, and thanks.","body":"I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n","comment_length":21,"text":"Datasets performance slow? - 6.4x slower than in memory dataset\nI've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset.\r\n\r\nNow, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower.\r\n\r\nFor example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33.\r\n\r\nIs this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss.\r\n\r\nFor reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU.\r\n\r\nI can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower.\r\n\r\nWhat am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance?\r\n\r\nAt 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice?\r\n\r\nIn any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test.\r\n\r\n``` py\r\nimport sys\r\nfrom datasets import load_dataset\r\nfrom transformers import DataCollatorWithPadding, BertTokenizerFast\r\nfrom torch.utils.data import DataLoader\r\nfrom tqdm import tqdm\r\n\r\nif __name__ == '__main__':\r\n tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\n collate_fn = DataCollatorWithPadding(tokenizer, padding=True)\r\n\r\n ds = load_dataset('yelp_polarity')\r\n\r\n def do_tokenize(x):\r\n return tokenizer(x['text'], truncation=True)\r\n\r\n ds = ds.map(do_tokenize, batched=True)\r\n ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\n if len(sys.argv) == 2 and sys.argv[1] == 'memory':\r\n # copy to memory - probably a faster way to do this - but demonstrates the point\r\n # approximately 530 batches per second - 17500 batches in 0:33\r\n print('using memory')\r\n _ds = [data for data in tqdm(ds['train'])]\r\n else:\r\n # approximately 83 batches per second - 17500 batches in 3:31\r\n print('using datasets')\r\n _ds = ds['train']\r\n\r\n dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4)\r\n\r\n for data in tqdm(dl):\r\n for k, v in data.items():\r\n data[k] = v.to('cuda')\r\n```\r\n\r\nFor reference, my conda environment is [here](https:\/\/gist.github.com\/05b6101518ff70ed42a858b302a0405d)\r\n\r\nOnce again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints.\r\n\r\nThanks for all your great work.\r\n\nMy fault! I had not seen the `dataloader_num_workers` in `TrainingArguments` ! Now I can parallelize and go fast! Sorry, and thanks.","embeddings":[-0.380666554,-0.0424089059,-0.0314095877,0.4377960861,0.1005082354,0.1263741553,0.0136552053,0.3071820736,0.0914101228,-0.0730988383,-0.0662421435,0.3011278808,-0.0139888329,-0.2742891014,-0.0800373852,0.0623485856,0.2097818255,0.0289249532,-0.2366964072,-0.1054769978,-0.2640983164,-0.2973698974,-0.0698852018,-0.2284685522,-0.3223609924,-0.1434253901,0.0818204358,-0.0005959411,-0.3450991511,-0.403732717,0.3832427859,0.0766954646,0.0686847046,0.5503901243,-0.0001125017,-0.0448605046,0.2153707594,0.1636437178,-0.1740950942,0.3116467595,-0.0444426648,-0.2415705472,-0.0513389893,-0.1154529974,-0.0132120242,-0.1297846586,-0.0774296299,-0.7160379887,0.0181460567,0.1087614447,0.1370815635,0.319951117,-0.3872780502,0.2173005193,0.1421412528,0.3198285103,-0.4389561415,-0.103455998,0.3924718201,-0.1402355283,-0.0548132025,0.2295684218,-0.182692647,0.2396481335,0.4582098126,-0.0244686864,-0.1892895699,-0.2351143211,0.0883351192,0.4672665894,0.1863689572,-0.1389111727,-0.6342912316,-0.3808341622,-0.2253736407,-0.1158320382,-0.0066045467,0.2012721598,-0.1965836883,0.1110184193,-0.5031599998,0.0621067099,0.0210670028,-0.0106447944,-0.1593937129,0.3560089469,0.108682014,0.2212632298,0.3045376241,0.1173553243,0.395116508,-0.3474166691,-0.0426575504,0.0569677539,-0.6727506518,0.0735261813,0.1322349459,0.1327432245,0.2916539013,0.0192395188,0.1375171244,0.2773087919,-0.0180944502,-0.1378564835,0.2188938707,0.132766515,-0.3069685102,0.4844334424,0.3089217246,-0.2571935356,0.1131687313,0.2347921282,-0.0429064818,-0.1230405346,0.0808739588,-0.2456587255,0.0628576428,-0.2480963022,-0.3668070436,-0.1015461758,-0.128623262,-0.0355535038,-0.0565834977,0.1639846861,-0.2934108377,0.2284448445,-0.1210528836,0.0039493861,-0.2586508393,0.0483607389,-0.0533004217,-0.0860240832,-0.3140702248,0.3556382954,0.2916277051,-0.0728289783,0.1457336843,-0.0316702574,0.0242215451,0.2429368049,0.3603523374,-0.38301754,0.2523589134,0.1450635195,-0.2260443121,0.5183621645,0.1015642062,0.4579477608,-0.2209920734,0.3493163586,-0.0600941256,-0.2040511966,0.2610084414,0.1124157831,-0.264804244,-0.1167250723,-0.4379530251,0.2817272842,0.1080345884,-0.0582337268,-0.0100165736,-0.3681587577,-0.1956059635,-0.1200666726,0.3318262398,0.3998510838,-0.5381987095,0.1778304726,0.060519293,0.211497277,0.2493430674,0.6614067554,-0.3634275496,0.2380728871,-0.1285901368,-0.2693578005,-0.1169592217,-0.1960741431,-0.5835575461,0.150880143,0.0044011883,0.137756452,0.103662096,0.4255777299,0.3806003928,0.1373888403,0.1205449998,0.2378032357,0.0506333634,0.1219747737,-0.3522131741,-0.266743809,0.5058071017,0.2659254074,-0.018459741,-0.1692710668,-0.1304339319,-0.2034452409,0.171515435,-0.1723211259,0.1123998016,0.2761603296,-0.0154160084,-0.1323151141,0.1253483295,0.0962624475,-0.2414122373,0.4494385421,0.0475798473,0.0823373869,-0.0722553283,-0.0690130517,0.2340127826,0.1237764359,-0.2757113576,-0.0913971737,-0.020273624,-0.1243479028,0.0506436378,-0.2392936796,-0.1698194146,0.3928050995,-0.2049032599,-0.0397737734,-0.2208188623,0.0434946306,0.0289860237,-0.0810197294,-0.0130326152,-0.1096339822,-0.1200399175,-0.0273582377,-0.1362688243,0.1346836239,0.2414696813,0.260176301,0.1974770576,0.3538561463,0.1384093165,-0.3115014434,0.3684659004,0.282561928,0.1614504755,-0.1878308803,-0.3582589626,0.5080374479,0.0408374928,0.6212487817,0.131858483,-0.2736033797,0.0354713872,0.161772728,-0.152790606,0.361810714,0.2422434837,0.183134377,0.5993636847,0.3990736306,-0.2270187736,0.1293455511,0.3130761981,0.0727399737,-0.1557809412,-0.0534361303,-0.2871873081,-0.2821799815,0.2410202771,-0.0649421513,0.5271029472,0.2901991308,0.1248301491,-0.1665140986,-0.0623956509,-0.088564761,0.0004647589,0.2956546247,0.2021181136,0.0236077383,0.1321617812,-0.0330443047,-0.2369342744,-0.1259484142,-0.013615571,0.0920877904,-0.1836669147,0.1966559887,-0.048241172,0.0431790762,-0.1185244098,0.0486888587,-0.1458606869,-0.2508894205,-0.07871975,-0.1703781933,0.1408747733,0.0997162461,-0.0557448044,0.297817409,-0.0532311983,-0.3437313437,-0.1330957562,-0.1097652093,-0.2056207806,0.0347893834,0.0790492743,0.201562956,0.3894380927,0.2333966941,0.0630549118,-0.2507633567,-0.1281210929,-0.081134595,-0.1727131903,0.2470872104,-0.3129540384,0.1722493172,0.0450521857,0.0761780441,0.0374998301,-0.3253332973,0.0282385517,-0.1393776536,-0.1856903732,-0.095603399,0.094167456,-0.283816576,-0.2353224307,-0.231665045,0.1122174487,0.1856211126,-0.1038918123,-0.0777875558,0.2100544274,0.1400318295,-0.0230241437,-0.0323164165,0.0010908976,-0.2395755649,0.5963212252,0.1307837963,-0.2874099314,-0.1484392285,0.0325048938,0.0926051661,0.4593246579,-0.4363919497,0.3887852132,-0.4268547595,0.0659126043,-0.0283237677,0.1497395784,0.3875692785,-0.04899057,-0.0553492531,-0.0899968445,-0.2898074687,-0.2084778547,-0.2657877803,0.1750806719,0.0703375936,0.2047269493,0.1540494412,0.7702897191,0.2816254497,-0.1282120496,-0.0700765774,-0.0199760105,0.3221504092,-0.3042885363,-0.1336632073,0.0767839104,-0.2323928922,-0.3620031178,0.1022298336,0.1592204869,-0.2228553742,0.0571660362,0.4334772527,0.2629042268,0.0344460271,0.1643804461,-0.1444606781,0.1150298938,0.1291987896,0.2734664679,-0.2965278924,-0.2377397567,0.1282240599,-0.1157930046,0.1971655339,-0.0538895167,-0.5846302509,0.0750427768,-0.7312512994,0.1715986729,-0.0042754202,0.2625332177,0.1212930679,0.0990571305,0.2460765243,0.0343444459,0.4295500517,0.0074468786,-0.1342289299,-0.0209551565,-0.262619704,-0.5181395411,0.001941319,-0.1180637106,0.0979928449,-0.0180772208,0.1409915537,-0.144948706,-0.2084661424,0.0128051573,0.2898257077,-0.1102436483,-0.2574890852,-0.3075466752,-0.2508772016,0.0031966071,-0.1106914133,0.0640596449,-0.2158922702,-0.0053233677,-0.2174378783,-0.1411990672,0.1710705906,0.287007153,-0.0199488755,0.2277724892,0.0420350321,0.2784542441,0.3091863096,-0.0779345483,0.31096825,0.6148449183,0.0961860493,-0.0449313186,-0.057990361,-0.0302234162,0.1740664989,0.3999004066,-0.1160176396,-0.1174309105,-0.0470448099,-0.0459505059,-0.4416711032,0.0930080861,0.360679239,0.0545088053,-0.4937981665,-0.6961229444,0.4829239249,0.3221649528,-0.0029609448,0.5213718414,0.1529991031,-0.1188371181,0.2289226204,0.3288167715,0.7686120272,-0.4081908762,0.342733115,0.0339299589,-0.0006147145,0.2862083614,-0.1507107466,0.0477475896,-0.1387883574,-0.0451609902,0.0209274869,-0.3210028112,0.2163613886,0.0421519428,0.2567621768,0.2072264403,0.1430619508,-0.0285382587,0.009281992,0.3683234751,0.1415445656,-0.3528417349,0.0806298405,0.2060843408,-0.0636474565,0.0266835541,0.1052983031,0.0771686435,-0.5997769833,0.2118579149,-0.2221498489,-0.0983413085,-0.2747580707,0.2297734916,-0.2467642426,-0.1948339492,0.1204785332,0.1752462685,-0.0273540523,-0.0050550401,-0.2925045192,0.2589013875,-0.1889818609,0.1734658927,0.46968472,-0.4210250676,0.1938281804,0.0822748691,-0.0500955544,0.042528823,-0.270781666,-0.3727757931,-0.1190479621,-0.0976812318,0.046276927,-0.0969666466,-0.177412957,0.1778092235,-0.1157280877,-0.0738806948,0.1192866489,0.2847138047,0.1483693868,0.5707704425,0.1340293884,-0.4016315937,-0.1551916599,0.2970864177,0.0244420748,0.0889516696,0.3403579295,0.1648631394,-0.2617582381,-0.0780200064,0.3460661769,0.5033900142,-0.1370906681,-0.0944340825,-0.3355536163,0.1455243677,0.1652888656,0.101435855,0.1463968903,-0.4325959086,-0.2803824842,-0.1083587259,-0.2876849771,0.1452467144,-0.1482289284,0.2979682088,-0.1901215762,-0.3174223304,-0.1114949062,0.018394433,-0.2807396948,-0.2505933046,-0.0107442504,0.1090027615,-0.101121746,0.2202881724,0.0244037174,-0.1119341254,0.0007436512,-0.0798370391,-0.1273417771,-0.1610721946,0.0595525987,0.1735744923,0.0525193512,-0.4033669531,-0.0549083687,-0.3336770833,-0.2233922631,-0.302418679,0.0491736084,0.3302561641,-0.1085040122,0.0007786985,0.0641422272,-0.0916875601,-0.1112686545,0.2955054045,-0.1259222031,0.1686892956,0.0309227761,0.1445612162,0.4665194452,-0.2060361505,-0.1343386769,0.1394993365,0.0919986293,0.2192005068,0.3119925261,-0.3754245639,0.3008852303,0.1782549918,0.5255870819,0.577677846,-0.0864237919,-0.0482995547,-0.1358982921,0.1196112111,-0.0082848733,-0.2492899597,0.0623386651,-0.111069724,0.1503143162,0.2359203696,0.3171467185,0.3776896596,-0.0135363294,-0.1543546915,0.6034693718,-0.4535905123,-0.0637547076,-0.1066441908,0.1659359187,0.1051363274,0.4445793033,0.1751129925,0.1501991898,0.5105736852,0.0687420592,0.6162849069,0.3639474511,0.313824445,-0.1088632941,-0.1895516366,-0.0064244275,0.083041288,-0.2204790413,-0.0311428737,0.0744839087,0.0459100455,-0.3543307185,-0.314951092,0.0743726939,-0.2537949979,-0.0074903257,-0.048367139,-0.0298703108,-0.1297776848,0.0728950724,0.2107991725,-0.2031932473,-0.1960491538,0.3447195292,-0.0779259726,-0.0568211861,-0.1621299088,0.280852139,-0.0214888789,0.1728643924,-0.1264068484,-0.0823844671,-0.284964323,0.0103563713,-0.1397747844,0.3787524104,0.2705047429,0.2324059308,-0.1640647352,0.050375063,-0.148900643,-0.1381408721,0.0803578123,0.3532447517,-0.2355175614,-0.0428107865,0.1823548824,0.1039419696,-0.1256211847,-0.0923540518,-0.188364476,0.0659575909,-0.4343673289,0.1180912927,-0.0840086266,-0.2931295931,0.0669182017,0.07898397,-0.2430418432,-0.1107975617,0.4821847081,-0.4651018381,-0.0238466635,-0.112780042,0.0683293864,-0.2132829726,0.7433564663,0.3071635365,0.2977477312,-0.388754487,0.1410839856,-0.629714489,-0.0627761632,-0.3574149609,0.0241962019,0.1204939187,0.2420873344,0.0829267129,0.3555292785,0.1255458742,0.0015364517,-0.0328144133,-0.0863277614,-0.2309202999,-0.0251039956,-0.2001771629,0.2183193266,-0.1794754118,-0.334594816,0.3847670853,-0.106375955,-0.043455869,-0.1651989669,-0.0484439619,0.3652563393,0.1291300356,0.3027887046,0.1771481782,0.4205040336,0.2366502583,-0.0797343925,-0.3657915592,-0.053249687,-0.0327494107,0.0968487412,-0.2428236902,0.2796160579,-0.1563116163,-0.4385993481,-0.1572759748,0.070731923,-0.0875135511,-0.0967460126,-0.3099508882,0.0481159538,-0.0935274437,0.1731007099,0.0528877564,0.1320955902,0.0943317115,-0.1697658598,-0.3104428649,0.0754998624,0.0954924896,-0.2117907852,-0.1013581157,-0.0900925323,-0.0787789971,0.4433886111,0.1187652647,-0.5254154205,-0.0171316732,0.321495533,-0.0486117974,-0.102469489,0.1972341388,-0.0890345648,-0.0165294632,-0.1392822713,0.2070880532,-0.1946892738,-0.1024894044,-0.0100645898,0.0016356963]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/707","title":"Requirements should specify pyarrow<1","comments":"Hello @mathcass I would want to work on this issue. May I do the same? ","body":"I was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.","comment_length":15,"text":"Requirements should specify pyarrow<1\nI was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.\nHello @mathcass I would want to work on this issue. May I do the same? ","embeddings":[-0.2184681743,-0.1796475202,0.0095730992,0.2583691478,0.0810354128,-0.0046880511,-0.0524621457,0.1359867752,-0.0720611736,0.0817059278,-0.0015765247,0.165083915,-0.0410678424,0.0125884796,0.1982615143,-0.3465024233,0.2946472168,0.7113897204,-0.206025809,-0.0302419327,-0.0526288897,0.2589272261,-0.3135199249,0.241330415,-0.3759223223,0.199465245,0.0763885826,-0.021983711,-0.2523011565,-0.6445119381,0.525924325,0.3301421106,0.0510368086,0.4057582319,-0.0001344269,-0.0024110288,0.1571382284,-0.0464288332,-0.0123291146,-0.0229075402,-0.1473000199,-0.1561318934,0.4686639309,-0.0483778007,-0.1823444963,-0.0561741367,0.0182243977,0.1818112135,0.3393440545,0.0799945369,-0.0025348903,0.2847137153,0.6692750454,0.0813626945,0.6278495193,-0.104434289,-0.1627849042,0.2372824401,0.2553147078,-0.2777830362,0.1320671141,-0.1496046633,0.1420923322,0.1130004153,0.182820186,-0.0948167369,0.4513402283,-0.2707454562,-0.1126469448,0.338280946,0.3595126569,-0.1958723515,-0.4813543856,-0.2269831598,-0.0875378698,0.0196941327,0.2253726274,-0.2525044382,-0.351623863,0.3974847198,0.000126842,-0.3094378114,-0.2692816257,0.1413810253,-0.4481065273,0.6317117214,0.1931491792,0.1571956724,0.1491387337,-0.0099027362,-0.2063032687,0.34385553,0.0197915006,0.0820774585,-0.2628011405,-0.1552556008,0.1938420534,0.3674228787,0.2781217992,0.0625823289,-0.2063448131,0.0357606597,0.1680972129,0.1465825289,0.1606892496,0.0482313149,-0.1425948292,0.0638123229,0.2782729864,-0.0032687068,0.2045881003,0.1701836735,-0.1130155772,-0.3809115589,0.0659815818,-0.2222530246,0.3967120945,-0.2652838528,-0.3305368423,-0.1373946667,-0.3141295314,0.055398114,-0.063442938,0.0595487021,0.0525634512,0.4498504102,-0.0011431596,0.1504696012,-0.3192583621,-0.1740972549,0.001476059,-0.1121686846,0.0886964127,0.1441914886,0.3981486559,-0.0086736875,0.2444756329,0.2111054659,0.1229913384,0.3438071609,-0.0262345001,0.0587381124,-0.4241413176,0.5613344908,-0.4919880033,0.1927098185,-0.0267107282,0.1162308455,-0.26808393,0.2310970277,-0.2706459165,0.0316975005,-0.1708506644,-0.0038416185,-0.4511280954,-0.0888761953,-0.2467582226,0.1062645912,-0.0633806288,-0.4063470662,0.0196325257,-0.2907786667,-0.0282245632,-0.0126736648,0.2648762167,0.1395405978,-0.4285311699,-0.0921003371,0.1573464423,-0.6298226714,-0.1456084698,0.2305918783,-0.2470219433,0.3063081503,-0.1489901394,0.3098466992,0.2117111236,-0.2791602612,-0.5069134831,-0.1852958649,-0.2865203023,0.3666157126,0.048749689,-0.1778726429,0.0859937221,-0.0412257649,-0.022593163,0.0162717421,0.1593060791,0.1239512861,-0.2605391741,-0.4413793981,0.6475383043,0.1815765798,0.0953393281,0.0171399526,-0.1749277264,-0.0594116151,0.196628198,-0.1953442097,0.063839592,0.0558572151,0.4046800435,-0.123991847,0.0776911005,-0.2948793173,-0.5367867351,0.2503334582,-0.0789953321,0.1200971007,-0.1541086137,-0.0576478355,-0.0936772004,-0.117729798,-0.2218975425,-0.0450864509,-0.0002401062,-0.3510248959,0.1932462901,0.0494407266,-0.2208816856,0.0793564543,0.4042799473,0.146463275,0.0654478222,0.3477944732,-0.1267597973,-0.2591908276,0.2620239556,0.1021635085,0.1128299162,-0.0408804826,-0.0343778431,0.2145361304,0.2017305046,0.3799296916,-0.2047355473,0.4031554759,0.1886650324,-0.5318733454,-0.0813160017,0.428008616,0.0507431291,-0.1015331671,0.1925344467,0.1949117333,0.3216634393,0.0487086996,0.0860464945,-0.1437521428,0.0229295623,0.0788786486,-0.0770915598,-0.0544076115,0.0746237859,0.1505306214,0.1111836582,0.0289035439,-0.0151340198,0.1760713309,0.2172666788,-0.0504407994,0.3995642364,0.0873619094,-0.2445646673,-0.1183868647,0.1589450091,-0.5414170027,0.2736468911,0.1807460934,-0.1555959731,0.0263441838,-0.2064407915,-0.0906103402,0.1290439516,0.0129333222,0.2398475558,0.2827976048,0.118326433,-0.01598108,-0.3418098092,-0.2424255311,-0.0591057688,0.3333070874,-0.571005702,-0.3344192207,-0.1035021916,0.3565844893,0.2867717445,-0.0556394383,-0.3220829368,-0.1297465861,0.1878872067,0.4278853834,0.2341695726,0.1362022609,-0.2086892426,0.1023266613,0.0394862704,-0.3329018056,-0.4262370765,-0.1749710441,-0.4419396222,-0.0323528759,-0.0733021423,0.0494342633,0.1888718456,-0.0716860294,0.0273340661,-0.1472794712,-0.5166342854,-0.0126185743,-0.1964084208,0.2887322903,0.2825522423,0.4836349785,-0.0143905263,0.0071936115,0.2166141868,-0.3632283509,-0.2659369409,-0.0521604903,-0.2391353995,0.1267824024,-0.1768006235,-0.0699720755,0.1442874968,-0.0736284256,0.4621376991,0.0718505681,0.0055371872,-0.0567086488,0.2460021079,0.2129061967,-0.1323195547,0.0499868467,-0.067451112,-0.1870760173,0.4790800512,0.0279959142,-0.237732172,0.0326807015,0.0117164962,0.3401373625,0.2471800596,-0.2556170523,-0.1977005005,-0.0914533734,0.5086722374,0.065155454,0.2349864841,0.6176438332,0.1453297585,0.1375848651,-0.1374483407,-0.2975311875,0.0366049223,0.1328589618,0.0410355963,-0.062410336,0.0190239139,-0.0689202771,0.7220748067,0.1805410832,0.2497692704,0.1880050749,-0.1340093613,-0.038727589,-0.1497257054,0.0109773427,0.0612105504,0.2387140989,0.0626510531,0.362392664,-0.0082371663,-0.1929070204,-0.3028367162,-0.0788288265,-0.2791213095,-0.3678583205,-0.1491655856,0.2377404273,0.2466145307,-0.0884538516,0.1118366346,-0.1570489407,-0.2174735814,0.2932991385,0.1461163312,0.0940377191,0.1437790245,-0.1467586905,-0.3636945486,-0.313872546,0.0422386825,0.1464237273,0.2950319052,0.0062028873,0.001568514,0.2513278723,-0.2766465843,0.5854098201,-0.2075577676,0.0745045468,-0.013466279,0.0262271781,-0.5049881339,-0.2446169853,-0.1738785654,-0.2472160608,-0.0167782847,0.0896965861,-0.7596051097,-0.1595819145,0.0578873716,0.0162847973,0.060979601,-0.0304881167,-0.3312805295,-0.3436859846,-0.1295042634,-0.2457185239,0.0617394149,0.1002186984,-0.1446244568,0.138832584,-0.1400859356,-0.1135425568,0.1472508013,0.3280717432,-0.0061052055,-0.1756786257,-0.1201879308,0.3453245759,-0.3474427462,0.2772834897,0.3458767235,0.3807895482,-0.4025198519,0.0481721014,0.2226974815,0.3583159149,0.4074164927,-0.1179825068,0.1971431226,-0.1694965512,0.0359963924,-0.2643255889,-0.1354401857,0.5726701021,-0.2287177593,-0.445019424,0.081261985,0.552116394,0.1940074712,-0.0467723422,-0.0632883683,0.7096910477,-0.3362965882,-0.1792881042,0.1273531467,1.1406264305,-0.0248490311,0.2473261058,0.6765403748,-0.1588457823,0.5147961974,-0.1974467933,0.0909665227,-0.4388770461,0.1908202618,-0.215523228,-0.1201725081,0.5187138319,-0.060135182,-0.3835266232,0.1679504216,-0.0410193391,0.390275985,0.3309307992,0.4093617201,0.050019376,-0.0208778437,-0.0926342085,-0.045070298,-0.1533531696,0.4164012074,-0.16749385,-0.1173002496,-0.1032201797,-0.3516890109,-0.5647407174,-0.1512971967,-0.632887125,0.1708581895,0.1716879308,-0.1631986946,0.4662188888,-0.065022774,0.060000971,-0.0614450276,-0.6299888492,0.2026435286,-0.2727764845,-0.316409409,0.1050971746,0.0251444057,0.0979503393,-0.222682029,-0.4676271081,0.1323000789,-0.1306631416,-0.2055814415,-0.3713487089,-0.2895797193,-0.1128320917,-0.1328438818,0.2229925543,-0.1476543695,0.2213087529,-0.1424404234,-0.0324849337,-0.0495837219,-0.0568765737,0.0584738292,0.1219949573,0.0095223775,-0.1087786555,0.3497531712,-0.0164230447,-0.0317866579,0.4289358258,0.0457732007,-0.1335276216,-0.2631809711,-0.1348035187,0.15174371,-0.2094664127,-0.0927212983,0.4336149096,-0.0520086251,-0.0601006746,0.6476336718,0.2194587737,-0.0705861524,0.0888518766,-0.205693543,-0.0781486258,0.4570547044,0.0452597775,0.3360444009,-0.1726716608,0.3661982417,0.1156333089,0.0359348133,-0.1197629347,0.1660009176,0.3005580902,-0.0142167704,-0.0259292908,-0.0334366746,0.1244254485,0.0516490117,0.0103749717,0.0135359457,-0.1695408076,-0.0254522208,-0.0384886526,0.2444959283,0.1455010176,-0.1281895638,0.2163442075,0.0136029217,-0.0190166775,-0.0348856114,0.2454285622,0.1642632931,-0.0955564827,0.2209249139,0.2699700296,0.0770268813,-0.3596716225,-0.0208439045,-0.0620745644,0.1586059481,0.1254786104,0.1768544912,-0.1476745754,-0.0063407351,0.0949037299,0.5591252446,0.0180631205,0.0748907328,0.1377249658,-0.2974795103,-0.0877950788,-0.0668647811,0.1378324032,0.4681286216,-0.0245772507,0.0170821957,0.2390627116,0.0323481411,-0.1606028527,0.0219473485,0.7301490903,0.0066128159,0.0462656915,0.2955257297,0.1630705297,-0.2300028205,-0.2191772461,0.0415242314,0.1825381517,0.1701244265,0.143195048,0.4622653425,0.0121410443,-0.038077835,0.2080612034,0.0445591882,0.111924544,0.1219315603,-0.3308547735,-0.0615758337,-0.2602821887,0.3340809941,0.3091212809,0.0904528573,-0.273843199,0.0813075155,0.13753739,-0.2009998262,0.0836105719,0.4977889359,0.2663789093,0.000754586,-0.4683029652,0.1404999942,-0.063182801,-0.2132144421,-0.0124405241,-0.2669169605,-0.0257676318,-0.0448554158,-0.1177963614,-0.4823380411,0.450829953,0.1795624495,-0.0793716684,-0.4652190804,0.2614364326,0.0961107761,0.4365414679,0.0740831196,0.2808980644,0.0957234949,-0.3562628627,-0.1094975546,0.5503528118,0.3795802891,0.304433614,-0.4581252933,-0.149341464,-0.1852260679,-0.1116145775,-0.0626784265,0.1420795768,0.112323612,0.1942065656,-0.0610600151,-0.0494807325,0.025495695,-0.3664503992,-0.1379855275,-0.0661098659,-0.4894159436,0.2052889615,-0.2217478007,-0.0763396248,0.1193129718,-0.0436979532,-0.0506600589,-0.125938952,0.0089358445,0.2037197053,-0.0018360092,-0.0624133646,-0.0365494005,0.066512987,0.4709437191,0.268700242,0.2825923264,-0.2857880592,-0.1075992137,-0.2465941906,0.1521293223,-0.279561162,-0.06989941,0.1277179271,0.1753626317,-0.2369573414,0.3721760213,0.1520057768,0.2086332142,0.0452933908,0.2363936156,-0.1257058829,0.2006753385,-0.0794207379,-0.1585906744,-0.1003259271,-0.3147715032,0.3072324395,-0.4304306507,-0.146254003,-0.1349830627,0.4565312266,0.0745008439,-0.3596764505,0.3986965418,0.4116817713,0.5397037864,0.1616663188,-0.0280598886,-0.4173544943,-0.178258568,-0.3377934098,0.1736617684,-0.1985481828,0.3198767006,-0.0661583766,-0.2826152146,-0.0794190615,0.2281568795,0.1725917011,0.2167717069,-0.1380596459,0.1227037236,-0.0266372506,0.111544013,-0.2197209597,0.0003308874,-0.0041569164,-0.173173219,-0.4804472327,-0.5023652911,0.3418385088,-0.6672945619,-0.2077832371,-0.2846734226,0.2944103777,0.1157280579,-0.2131050676,-0.5761139393,0.0057171378,0.2340031266,0.1797155589,-0.4021957815,0.3607995808,0.1584106386,-0.1529832482,-0.0651513711,0.3677757978,-0.1255606264,0.1176461354,-0.1512745023,-0.1483117491]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/707","title":"Requirements should specify pyarrow<1","comments":"@punitaojha, certainly. Feel free to work on this. Let me know if you need any help or clarity.","body":"I was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.","comment_length":18,"text":"Requirements should specify pyarrow<1\nI was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.\n@punitaojha, certainly. Feel free to work on this. Let me know if you need any help or clarity.","embeddings":[-0.2744113207,-0.2090404034,0.0152528426,0.18003501,0.0473458432,-0.0619748719,0.0390898697,0.1851368546,-0.0446801148,0.0569026507,-0.0508330688,0.2923365533,-0.0050392509,0.0731776655,0.1184344813,-0.3647958636,0.2140396386,0.5875732899,-0.3834571838,-0.0399198346,-0.0928857177,0.26306355,-0.29579404,0.2726375163,-0.3719050586,0.1225018725,0.0235204715,-0.0084750494,-0.1621870846,-0.5378573537,0.4328965843,0.2639689445,0.028079696,0.5115487576,-0.0001171745,0.0794191658,0.1408243179,-0.0046915193,-0.0439403914,-0.0354698151,-0.0304914769,-0.0833085403,0.3720040917,-0.0927432105,-0.2928206027,-0.0792524815,0.0061559151,0.2068326324,0.4049660861,0.1370559633,0.1354582161,0.41404742,0.6308282018,-0.0483633019,0.5168268681,-0.2037336528,-0.2146046609,0.1246313602,0.2182046473,-0.2637349665,0.0977444574,-0.0559179597,0.2176226079,0.0844787881,0.1007053256,0.0064795171,0.2837662995,-0.2852359116,-0.0767906383,0.3960488141,0.3805250525,-0.2011321783,-0.3604101837,-0.1703670174,-0.2085843235,0.0597833134,0.2075708658,-0.232400164,-0.3061923981,0.3499825299,-0.0237948038,-0.1572733521,-0.2329319715,0.0398114808,-0.3779924214,0.61867553,0.1304641664,0.2004017979,0.0787140951,-0.138892293,-0.2523773015,0.2585268319,-0.0129787577,0.0636072531,-0.293458432,-0.1428923458,0.2488572448,0.2419497669,0.2727940083,0.0403146632,-0.1411073655,0.1671097875,0.1981676668,0.2720076442,0.1446544677,0.0335154794,-0.2245220691,0.1350048631,0.3313992918,-0.0000973628,0.1322936565,0.0535429455,-0.0543981716,-0.2799725235,0.0378361493,-0.3460942805,0.2894438207,-0.1931288242,-0.3421807885,-0.1763553172,-0.1166223884,0.1466711164,-0.0318934657,0.1282465607,0.0388669521,0.3341369033,0.0885849744,0.1577054411,-0.2806058228,-0.0087910155,-0.132013455,-0.1841318309,-0.0091545731,0.0849368274,0.3447553217,0.050514672,0.332337141,0.1975707263,0.068075031,0.4564490318,-0.0531207435,0.1364161074,-0.3700836301,0.4508806169,-0.3323380351,0.1976619512,-0.0581384934,0.0628103763,-0.1915748864,0.1300496459,-0.1899691522,0.0482477732,-0.1915132552,0.1574584693,-0.3573656976,-0.082664758,-0.1527121961,0.0702023059,-0.0288152564,-0.4366858304,-0.0195416939,-0.2502210736,-0.0132398773,-0.0616674386,0.2646416128,0.0713782385,-0.362818718,-0.0274630133,0.1427609324,-0.6227787733,-0.1789803058,0.2393549085,-0.2394668162,0.2606437206,-0.1663723737,0.3105495572,0.1664200425,-0.269310981,-0.4029172361,-0.1308529675,-0.2516518235,0.4186939895,-0.0464100912,-0.1123274341,0.0676421225,-0.0331610255,-0.0478268117,0.1581586301,0.1312135458,0.2168271989,-0.3007755578,-0.3674801886,0.5387315154,0.2180080563,0.0704710558,-0.0844723061,-0.1506546885,-0.07358706,0.2024605721,-0.2686364949,0.133695662,0.0354774483,0.4232333302,-0.0971348882,0.2044667453,-0.2488170713,-0.5351415277,0.319075048,-0.094259806,0.2063255608,-0.1566704512,0.038478896,-0.0378696322,-0.0542144403,-0.3329945803,-0.2113429904,0.1806448847,-0.2434691936,0.1964030266,0.1206182241,-0.1091237515,0.0806275234,0.449213028,0.1215978563,0.0134600187,0.3498786092,-0.1768245101,-0.3050769269,0.2239729464,0.0812370852,0.0774515942,-0.0824306607,-0.0425241143,0.286680609,0.164312169,0.3654063642,-0.1780032665,0.3176199496,0.1306525767,-0.5406620502,-0.0758885443,0.4269067347,0.1035571247,-0.1261876523,0.1132316515,0.1425604224,0.2797724009,0.0193208996,0.1519389451,-0.1222834364,0.1331613213,0.0406503417,0.0729066283,-0.0568339936,0.0058672535,0.1114464849,0.1371488273,0.0175449196,-0.0516746566,0.2303841114,0.3417685032,0.0505564809,0.3291549683,0.1389820129,-0.2897283733,-0.1328887045,0.0805216581,-0.5257543921,0.2791210711,0.2369832247,-0.0444273166,0.1000379026,-0.2512210011,-0.1261416823,0.1744200289,-0.0804376379,0.178245008,0.3231774271,0.0959307104,-0.0637741536,-0.3817213476,-0.1507439464,-0.0748539418,0.3306754231,-0.3862912357,-0.2413176,-0.1049693823,0.3216769695,0.272595346,-0.0808423758,-0.3064685464,-0.2201129049,0.2603960037,0.4772414863,0.1165540069,0.1628492922,-0.0737927631,0.0380958728,0.139837876,-0.2964511812,-0.3593039513,-0.2747438252,-0.4682711661,0.1212390661,0.0316688716,0.0594272241,0.2674309015,0.0404028371,0.0869702622,-0.2797509134,-0.5713727474,0.0258007515,-0.2757348418,0.2965800464,0.2656935453,0.3712022603,0.0146423532,-0.1520946175,0.334280163,-0.5331454873,-0.2797227502,0.0561480001,-0.1597094983,-0.0188374985,-0.227671802,-0.200541243,0.0545630939,-0.2094620019,0.487305969,0.0893598571,-0.0777641311,-0.0578465946,0.2512842417,0.1818445325,-0.100018099,0.0056195739,-0.1132392287,-0.2412524372,0.4282823205,-0.1403423101,-0.37278229,-0.0131195094,0.048239544,0.3445105553,0.2113744169,-0.2584074438,-0.2144919336,-0.160062328,0.4973121881,0.1709284335,0.2897213995,0.5415073633,0.0972692147,-0.0025504425,-0.2357560247,-0.2569276989,0.2003496736,0.0552067384,0.0855824277,-0.1022009477,0.0219613146,-0.0871281177,0.7050462961,0.210354358,0.1172728911,0.2652158141,-0.1464131773,0.0713077039,-0.2871912718,-0.0204011016,0.1816247851,0.1265805662,0.1141961813,0.4659206867,0.0153718991,-0.2948928773,-0.3402727246,-0.0141839506,-0.265638411,-0.3895333409,-0.1034848243,0.359426707,0.1929802299,-0.0434745513,-0.0052686129,-0.1688804179,-0.1890220344,0.2966147959,-0.0397454314,-0.0240895562,0.0552966595,-0.1579433084,-0.2891508937,-0.3336388469,0.0818578079,0.1664174795,0.1847863048,-0.0680765286,-0.1340449899,0.2387772352,-0.2572526038,0.5482107997,-0.2503589988,0.0644779205,0.074098289,0.0079346877,-0.5168680549,-0.2064503878,-0.2619241774,-0.2034538984,-0.0051719351,0.1305753142,-0.8042139411,-0.1416983902,-0.0203215536,0.0543859378,-0.0029820025,-0.0186938159,-0.3273305893,-0.330019325,-0.1926887631,-0.2540754974,0.029383501,0.1346323192,-0.2136129439,0.0616567023,-0.137757197,-0.0156033468,-0.0004649303,0.5012168288,0.0365820117,-0.2493593693,0.0479821935,0.296244055,-0.2872927785,0.2205220908,0.4225324392,0.3253390789,-0.3725339472,0.0434294194,0.091584377,0.2149431258,0.2550437748,-0.1127832532,0.1612326503,-0.2287414521,0.1712809801,-0.2005300671,-0.0550193675,0.6458345056,-0.1297022402,-0.4264853597,0.1230620518,0.4306225777,0.0931048095,-0.085847944,-0.0152061256,0.5795417428,-0.412197113,0.0432370938,0.060630966,0.9315183759,-0.0630197376,0.2635147274,0.6802076101,-0.182333529,0.4589168429,-0.1581913531,0.0483356267,-0.5087715387,0.1453892291,-0.0921903178,0.020042751,0.3349588215,-0.0943754092,-0.4093668461,0.1114815921,0.0546160005,0.3981406689,0.2363925129,0.3883899152,0.0551471487,-0.0701506287,-0.0819075033,0.1282799244,-0.1015640348,0.2616382241,-0.1411422193,-0.2780426145,-0.1079298332,-0.4004366398,-0.437412858,0.0159220137,-0.5444221497,0.2426148206,0.1901870072,-0.1010000557,0.4206794202,-0.0795523673,0.0983965397,-0.0371325798,-0.5875917077,0.2600964308,-0.3174569905,-0.3028061092,-0.0156446546,0.0216641985,0.1476682574,-0.2335633039,-0.4823552668,0.1443567276,-0.0612180643,-0.2304868102,-0.2473804951,-0.2755086422,-0.1288022697,-0.1659006029,0.1807358414,-0.0787779689,0.0377132855,-0.1690511405,0.0951925591,-0.0664826334,-0.1401763558,0.1735385954,0.1389178187,0.0415222645,-0.0753181875,0.3403992355,-0.1505477577,-0.0002171361,0.4557004273,0.184414193,-0.2143869996,-0.3254329264,-0.1513898522,0.1860679388,-0.2213612348,-0.0901408121,0.4567629993,-0.0418890081,0.1108160913,0.6615354419,0.13069354,-0.1404129416,0.0508983433,-0.1539063007,-0.0674634948,0.3901177049,0.0278619304,0.3385826945,-0.0731707215,0.365308702,0.2052224874,0.0990205258,-0.30164814,0.1949870139,0.1255561709,0.0547828227,-0.0497926287,-0.2189925909,0.1979608834,0.1060333773,0.1444692761,-0.0088695893,-0.2713912427,-0.1890643537,-0.027004784,0.134412542,0.2101166248,-0.204305172,0.1009770557,-0.0373051502,-0.0319802314,-0.0614558272,0.3890509903,0.0531606004,-0.1427123398,0.1534888744,0.1570970118,0.0543242581,-0.2639515996,-0.0870099366,-0.094180271,0.0004089592,0.1391077191,0.098004967,-0.2191276699,-0.1056191921,0.2948779464,0.504645884,0.0544735417,0.0643531159,0.2324198782,-0.3727887273,-0.1605534703,-0.1102430746,0.284666419,0.3755245805,-0.0397622138,-0.0546569154,0.284601897,0.1878604293,-0.2341835052,-0.0106671713,0.6830043197,0.0071560801,0.155257374,0.1619551033,0.1528609544,-0.3002581,-0.2156451792,0.0755801871,0.2629773021,0.0707815886,0.0582783744,0.4259404242,-0.0978896022,-0.0580639541,0.1368257999,0.1183037385,0.1697496772,0.2068119794,-0.3776710927,0.0475411639,-0.2825198174,0.2500368059,0.3207114041,0.1081866622,-0.2281583697,-0.038232699,0.0943588614,-0.2090092599,0.010834083,0.4563935995,0.1838445067,0.0454131141,-0.490598917,0.2197510898,-0.1127001718,-0.1174176261,-0.0129564442,-0.2486073375,-0.0332058035,-0.1253835261,-0.1384198815,-0.3741065264,0.4117640257,0.1140366569,0.0056991726,-0.4141677618,0.1806014627,0.0854028761,0.3866275251,0.1436970532,0.2506822348,0.2383488417,-0.3593237996,-0.155341804,0.5158734322,0.3609436154,0.2361157984,-0.4983881116,0.0330157168,-0.1927713901,-0.0972192436,-0.073848322,0.2962005138,0.0832715929,0.1003081277,0.0081479587,0.0812422261,-0.1380749792,-0.3515096307,-0.1028728783,-0.0035800484,-0.2982715368,0.0544766523,-0.2025982291,-0.0291763376,0.0609574877,-0.1285893321,-0.1703608781,0.0027335589,0.0466495827,0.1201156229,0.0524003133,-0.1782757193,0.0606786087,0.0881635025,0.5512146354,0.2267614454,0.2592985034,-0.1696711481,-0.1509065926,-0.3126692772,0.2088225484,-0.2903214991,-0.053837724,-0.006735621,0.2106719166,-0.1541049629,0.341060251,0.1830129623,0.2890583575,-0.0141778048,0.2211007327,-0.2086694092,0.1971787065,-0.1781776696,-0.164937228,-0.0768914968,-0.2805628777,0.2442214042,-0.3426177204,0.0537522696,-0.1285856962,0.4614972472,0.0093119219,-0.3884457946,0.3599073291,0.4141126275,0.4446305037,0.0833821446,0.0008541755,-0.4352279305,-0.2522577345,-0.3233967125,0.0065197409,-0.1300686002,0.349300772,0.0371886864,-0.3272906244,-0.0716946572,0.1934989542,0.1795130074,0.1985123307,-0.0171341728,0.0927244425,0.0163966343,0.0711495206,-0.2019703984,0.1262801886,-0.0323904566,-0.2043989748,-0.3565768898,-0.6132531762,0.3638980091,-0.6347136497,-0.2741698921,-0.2015027404,0.2138897032,0.1161636561,-0.2915629148,-0.509082973,0.1441642344,0.1522792876,0.1124377549,-0.3741119802,0.3102248013,0.0858423561,-0.1579924822,-0.0411101356,0.3362417221,0.0036743693,0.0551079139,-0.084028095,-0.1470239609]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/707","title":"Requirements should specify pyarrow<1","comments":"Hello @mathcass \r\n1. I did fork the repository and clone the same on my local system. \r\n\r\n2. Then learnt about how we can publish our package on pypi.org. Also, found some instructions on same in setup.py documentation.\r\n\r\n3. Then I Perplexity document link that you shared above. I created a colab link from there keep both tensorflow and pytorch means a mixed option and tried to run it in colab but I encountered no errors at a point where you mentioned. Can you help me to figure out the issue. \r\n\r\n4.Here is the link of the colab file with my saved responses. \r\nhttps:\/\/colab.research.google.com\/drive\/1hfYz8Ira39FnREbxgwa_goZWpOojp2NH?usp=sharing","body":"I was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.","comment_length":103,"text":"Requirements should specify pyarrow<1\nI was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.\nHello @mathcass \r\n1. I did fork the repository and clone the same on my local system. \r\n\r\n2. Then learnt about how we can publish our package on pypi.org. Also, found some instructions on same in setup.py documentation.\r\n\r\n3. Then I Perplexity document link that you shared above. I created a colab link from there keep both tensorflow and pytorch means a mixed option and tried to run it in colab but I encountered no errors at a point where you mentioned. Can you help me to figure out the issue. \r\n\r\n4.Here is the link of the colab file with my saved responses. \r\nhttps:\/\/colab.research.google.com\/drive\/1hfYz8Ira39FnREbxgwa_goZWpOojp2NH?usp=sharing","embeddings":[-0.2086724043,-0.2492389828,0.0104706734,0.143163085,-0.1351191252,-0.1367644966,0.0601092875,0.1777841598,-0.0722263679,0.1876842231,-0.1349501312,0.3921408951,-0.0105912182,0.2025151551,0.0202153549,-0.3189196289,0.1558381766,0.6317852139,-0.2972459197,-0.0332883708,-0.0148538966,0.2716951072,-0.3856919408,0.1468177438,-0.4066579044,-0.0718952566,0.0602015518,-0.1913313568,-0.1789083779,-0.4104488492,0.5019386411,0.2933356464,0.0380189568,0.5386695266,-0.0001246494,0.0441618152,0.0715481415,-0.058221627,-0.0730772763,-0.4739520848,-0.0124219982,0.0288729072,0.5552842021,-0.1167393997,-0.2710622251,0.1298567802,0.0246236678,0.3614777923,0.4647699296,0.224004969,0.024686344,0.3220967054,0.519674778,0.027610248,0.4959204793,-0.1979997456,-0.2250536829,0.4007034898,0.2845157087,-0.293859601,0.2136534899,0.0722620934,0.1143926382,0.2241774201,0.0857390538,-0.0296407621,0.306420207,-0.543872416,-0.0512404032,0.4704498947,0.1956389546,-0.2339904308,-0.2855395377,-0.1750824153,-0.076056689,-0.0668525696,0.3151905835,-0.19068636,-0.4019597471,0.2506096959,0.0359963775,-0.0399612859,-0.3839387298,0.1969906092,-0.532189548,0.7457957268,0.0781329572,0.140368104,0.1287548691,-0.0064596632,0.0153422831,0.1976995319,0.2052425593,-0.0080036595,-0.2740123272,-0.1842935085,0.0807474703,-0.0599373505,0.186905086,-0.1586042643,-0.1703172773,0.1618908644,0.0550829098,0.3274672925,0.0876166373,0.1712814569,-0.225545615,0.1574649513,0.3345220089,0.10782291,0.0957444534,0.1650237739,-0.0801129118,-0.2739412785,-0.2060223222,-0.3268119991,0.3248876929,-0.1852011532,-0.543025136,-0.2471684366,-0.3868255019,0.1440582424,0.007572154,0.193520993,-0.168186903,0.2706876099,0.1613992006,0.1995216012,-0.2905425429,0.1262453645,-0.0097410362,-0.0157327522,0.0473034531,0.1668775827,0.3879338205,0.0598513223,0.3442662656,0.2795542777,0.2312610149,0.1923202425,0.1490726173,-0.0685155839,-0.1666152626,0.5420348644,-0.238415882,0.1536210626,-0.0455889925,0.1128447428,-0.1617376059,0.2223211378,-0.3807045221,0.0063013029,-0.2397688776,0.0668680519,-0.4700929821,-0.0116190957,-0.2089211196,0.2450740188,-0.0518171676,-0.2779945731,-0.0258288421,-0.4117038846,-0.2063217014,-0.0955633819,0.3335030675,-0.1586707532,-0.4914388955,-0.0222056601,0.1562986374,-0.343288064,-0.0639649332,0.2183389217,-0.3571609557,0.356990993,-0.1360036582,0.2632779777,0.1253128499,-0.2123099864,-0.4445071518,-0.2307714373,-0.1657947898,0.4377110302,-0.0319378376,-0.1881282777,0.2705567181,-0.0057018003,0.3001312315,0.0910482481,0.0968549624,0.1326059699,-0.2460104674,-0.4140643775,0.6402183771,0.1166730896,0.3057157993,-0.1149227023,-0.2733577788,-0.076900214,0.2295586914,-0.3223408163,-0.0021763155,-0.026989311,0.4426197708,0.0258706734,0.0911421254,-0.3212692738,-0.5001301765,0.2173959762,-0.0801016539,0.3355025649,-0.1914384365,0.060521286,-0.0343133807,-0.1909963489,-0.3320446014,-0.2931030691,0.1080883965,-0.3093075454,0.0226568878,0.1018400863,-0.1950880587,0.1308250874,0.2655107379,0.0207538102,0.2158547193,0.4626187682,-0.1991202086,-0.4572329521,0.2032588124,0.1807994395,0.0625622943,-0.0591128655,-0.1452132314,0.093349576,-0.0193592459,0.1822324395,-0.2986144125,0.2605123222,0.2605971694,-0.4232666194,-0.1955486089,0.3503348827,-0.0008110038,-0.0930695385,0.2096849829,0.3274608552,0.2745113671,0.017079927,0.0372881182,-0.2270409912,0.1220123917,-0.0016372741,-0.0083608758,0.0001987793,0.0653400049,0.1265785098,-0.0049333228,0.080294393,-0.1801214367,0.2274035811,0.4201957285,0.0835636333,0.4133821428,0.055138547,-0.2376071066,-0.1505482495,0.1739173383,-0.3802369833,0.2892012,0.1841975749,-0.0777000561,0.0568747893,-0.3004885316,-0.1140455753,0.1123887077,-0.0253327359,0.4185265005,0.2763240635,0.1527326256,-0.024946237,-0.2231233716,-0.1890087128,0.0537125058,0.2534185648,-0.3964975178,-0.245354414,-0.1942079663,0.1345811188,0.1920168251,0.0723449737,-0.1970225722,-0.048667416,0.2232683599,0.653278172,0.2153997421,0.2832365334,0.0743162706,-0.051443357,0.0335617065,-0.3396537304,-0.2768448293,-0.2075626552,-0.3671346009,0.0441854112,0.0132677574,0.133673802,0.230551362,-0.0909765139,-0.1738574952,-0.1247097477,-0.5882374644,0.1316970289,-0.3359529972,0.3475390971,0.0953313559,0.3303355873,-0.2024286836,-0.2611160278,0.2653265297,-0.3258785307,-0.2533763051,-0.0841299966,-0.2031656206,0.1257281452,-0.1894627512,-0.3544028997,-0.0187996104,-0.1409433037,0.2841362059,-0.1004943103,0.0101524582,0.0029494928,0.3139958978,0.2860015035,0.0444627069,0.0526336655,-0.1595649868,-0.1258476228,0.4896460772,-0.1127681807,-0.3425046802,0.1377508044,-0.0884387046,0.4012726247,0.2340296209,-0.3173576295,-0.3688831925,-0.0117837349,0.4804005921,0.1285279393,0.0694005191,0.552575469,0.0312730111,0.1009884104,-0.1931358874,-0.2890688479,0.2264042348,0.1350500584,0.1334231049,-0.0029779437,-0.0167331435,0.0700285733,0.771420598,0.0803975835,0.0786430463,0.1744908541,-0.0797611251,-0.0188491642,-0.1822079569,-0.0707815811,0.1368534416,0.0644049793,0.1029367149,0.3807842433,-0.1876033247,-0.3487024605,-0.3500249982,-0.0201874524,-0.3264290988,-0.4082984328,0.1243103966,0.2795991302,0.2563576698,-0.0651810095,-0.0377918929,-0.0926645622,-0.165588364,0.215112105,0.2523818612,-0.1616763324,0.0830574781,-0.0622322932,-0.4791888297,-0.2220380902,0.0489147976,0.2555783391,0.2051359266,-0.0589679293,0.1463351101,0.2981480956,-0.111692898,0.5202952027,-0.0870405585,0.0110761067,0.0811137259,-0.172903046,-0.2840286791,-0.3062788844,-0.4232672751,-0.2646532059,0.1585467011,0.1654624492,-0.7676321864,-0.2575203776,0.0505747758,-0.0228640605,0.0322505087,-0.0421996899,-0.3416551352,-0.4767364264,-0.2516466081,-0.2090094239,0.1732095182,0.1290228069,-0.3761099875,-0.0231266059,-0.0450554341,0.0272252522,0.0778521672,0.4401805103,0.0053054122,-0.0469987765,-0.0451438166,0.2295737714,-0.2422747612,0.4745727777,0.1285251677,0.324803561,-0.4303437471,0.2039560974,0.2204692215,0.1308780462,0.3809553087,-0.1939184815,0.2553448975,-0.2127743214,0.0657228827,0.0090952683,0.1710775942,0.6335740685,0.0473977327,-0.3903391361,-0.0285553243,0.3548615277,0.1598041952,-0.1670046747,-0.0544485822,0.3755366504,-0.3867868185,-0.0233918577,0.1086357161,1.0120888948,-0.0974006653,0.2016103864,0.7103577852,-0.1251997799,0.6407565475,-0.1507136375,0.2180497944,-0.3029672503,0.1637694538,-0.1251390427,-0.1464380473,0.3246546388,-0.2493910044,-0.517329216,0.1654734761,0.1835632473,0.2719470859,0.3307575285,0.3664639294,0.0562734976,-0.1726071537,-0.0661861151,-0.0214585904,-0.1906188577,0.3654985428,-0.143700853,-0.2710828483,-0.1689605266,-0.1311110854,-0.3711029291,-0.0510370322,-0.86431247,0.2675008178,0.1508107632,-0.2401708812,0.4472086728,-0.0309454184,0.1376144141,-0.1097687334,-0.4689815342,0.1472121477,-0.2493483424,-0.3123396933,0.0910324678,-0.141910553,0.3792173266,-0.3129012287,-0.5409114361,0.1033018604,-0.0511757955,-0.1094794646,-0.1680267006,-0.2288939357,-0.1445110887,-0.0526683442,0.2425125986,0.0594205409,0.1441214681,-0.203457579,0.0338383615,-0.0299207754,-0.1953060925,0.1536721885,0.1596405357,-0.1387038678,-0.0109998519,0.2203334868,0.0495722666,0.0212057643,0.5640969276,0.079010427,-0.133562848,-0.1652043462,-0.2041968405,-0.0878229588,-0.0198419187,-0.0097453343,0.4078778028,-0.2062466592,-0.087968275,0.8415583372,0.2351234853,0.0826877952,-0.0706818998,-0.1525163054,-0.0312063172,0.3379786015,0.0881566033,0.364815563,-0.0028082824,0.5681377649,0.2385709137,-0.0400059633,-0.1985764354,0.1549594402,0.0803827047,0.1271061748,-0.2191509902,-0.1455724239,0.1834778488,0.1521439403,-0.0072622537,0.0748372227,-0.2029429674,-0.092824012,-0.0556371957,0.1708943099,0.1209280491,-0.1932411939,-0.0316785164,-0.1339410096,0.0819301978,-0.0132941464,0.3547129631,0.1251430959,-0.1496167183,0.0882559195,-0.0083853109,0.1833599359,-0.2099686265,0.0715834498,-0.1528007239,0.0815197751,0.1265183985,0.0473816954,-0.1234634295,-0.0083737811,0.2713010013,0.2433030009,-0.175644666,0.1151928231,0.0139481807,-0.3107477129,-0.0969542116,-0.2689747512,0.1533014774,0.3499805629,0.0143595003,-0.0810011923,0.3281117976,0.0702682808,-0.1180961654,0.2799508274,0.5855948329,-0.0165175349,0.1711314172,0.2889331877,0.1592019349,-0.0347109623,-0.1282282621,0.094213523,0.117757991,0.1074803844,0.1329014003,0.0949653536,-0.1843547374,0.0916816294,0.1803431958,0.0330013521,0.135751307,0.0499206409,-0.1537524462,0.0347510949,-0.2302161604,0.2335726917,0.4710671306,0.25595209,-0.2555856109,0.0093812412,-0.0461414494,-0.174395442,0.0023055139,0.3529137075,0.0215682015,-0.1229898185,-0.3077789247,0.1919380873,-0.1925984919,-0.0194737297,-0.1789566278,-0.2740046084,0.0337448977,-0.1235163733,0.0066535859,-0.3109351099,0.3535003662,0.2466618121,-0.1854855716,-0.2436199635,0.1156520247,0.1982557178,0.3388854563,0.1959418207,0.3653029799,0.0804524645,-0.2732245624,-0.2526738942,0.5344126821,0.3100623488,0.3571822643,-0.5270896554,-0.1186390743,0.050837405,-0.038234394,-0.0993915647,0.0742177069,0.3719755709,0.1159041151,-0.0343117006,0.0166897047,-0.1020942554,-0.1851174384,-0.162114352,-0.1160400659,-0.3301354051,0.0211435705,-0.1357685328,0.0005404799,0.1968808621,-0.1748357713,-0.1484432518,0.1207373291,0.1013769805,-0.0851651654,0.0744904205,-0.012759028,0.0261575337,0.1295501739,0.4061409831,0.0721219182,0.3320486844,-0.2823962569,-0.0656898692,-0.2463539839,0.2152016908,-0.1876375973,0.0049150465,-0.0860483423,0.2188675255,-0.1591217369,0.3309309185,0.234250471,0.0933299661,0.2342053652,0.1184995696,-0.3123105466,0.1131899059,-0.1605767757,-0.112708509,0.0392215811,-0.2111784518,0.1725055873,-0.4397734106,-0.0324785672,-0.0812579244,0.4167866409,0.1804928184,-0.2006976455,0.3395538628,0.3661825359,0.6058911681,0.0794832259,0.0877500698,-0.0646250919,-0.1996700913,-0.3480488658,0.0865994245,-0.064177461,0.3849518001,-0.0420470387,-0.3016376793,-0.0782693326,0.2787020802,0.1799233258,-0.1153789163,-0.0422519334,0.0655879974,0.0778086036,0.1526705623,-0.1873805374,0.2286456078,-0.0583111718,-0.1775380224,-0.4938701987,-0.5086950064,0.3863790333,-0.7380798459,-0.3214219213,-0.2718997002,0.1684070528,0.1280069351,-0.0242552664,-0.5557008386,-0.0519464016,0.1206484437,0.250685066,-0.4771203101,0.269864887,0.1755547076,-0.0765638873,-0.0590749197,0.3788481951,-0.0047910665,-0.1537258625,-0.2274892032,-0.195642516]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/707","title":"Requirements should specify pyarrow<1","comments":"Also, please share some links which made you conclude that pyarrow < 1 would help. ","body":"I was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.","comment_length":15,"text":"Requirements should specify pyarrow<1\nI was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.\nAlso, please share some links which made you conclude that pyarrow < 1 would help. ","embeddings":[-0.2884189188,-0.2130024731,0.0232637059,0.1959736198,0.0543849804,-0.0585194491,0.0271022562,0.1870937645,-0.0278004091,0.0787156224,-0.0430811495,0.2951213717,0.0129551571,0.0651427433,0.1308615655,-0.3642075062,0.1980435252,0.5893019438,-0.4127082825,-0.0204928108,-0.1149846911,0.2640613019,-0.2953101099,0.2697647512,-0.3863803446,0.1321671605,0.032500688,0.0084769623,-0.105433166,-0.5279193521,0.4393040836,0.2499842793,0.0158766452,0.4985886514,-0.0001162542,0.0902105793,0.1582967043,-0.018678382,-0.0388238803,-0.0482270196,-0.033626575,-0.0796304643,0.3679337204,-0.0938403755,-0.3176647127,-0.0656953529,0.0177631304,0.2201688588,0.4021618366,0.134517014,0.1432773024,0.4291445017,0.6221467257,-0.0559801906,0.5306423903,-0.2237151116,-0.2100241631,0.1086437181,0.2228767574,-0.269215703,0.0951325595,-0.050817553,0.2044183165,0.0820160061,0.0697146282,0.0101763979,0.3071729541,-0.2699528039,-0.0762089267,0.3755886257,0.3744158745,-0.1932648569,-0.3323806226,-0.1581032276,-0.2210713327,0.0853025466,0.2259534448,-0.2543055713,-0.2877891362,0.3496618569,-0.0230650511,-0.1415728629,-0.217463091,0.0307581015,-0.3705138564,0.6154306531,0.1274533421,0.2102266252,0.0850322321,-0.1379047632,-0.2763998806,0.2418058068,-0.0273374803,0.0434532575,-0.3027988672,-0.1337102652,0.2557691336,0.2169889808,0.2609139383,0.0583833009,-0.1163472906,0.1746545732,0.2093643844,0.2908481658,0.1288990378,0.0310670398,-0.2532361448,0.1519726068,0.3379288018,0.0021519579,0.0923068449,0.0469338,-0.0562966801,-0.2438757122,0.0505803674,-0.3447391093,0.276422441,-0.1954369843,-0.3501516283,-0.1822246015,-0.0832382143,0.162123993,-0.013500371,0.1416802555,0.0479840972,0.3418589234,0.0987584442,0.1481140107,-0.2741329372,0.0235562902,-0.1509594768,-0.2161423117,0.0096637746,0.0791789517,0.3454882205,0.0691154301,0.3587607741,0.1929480731,0.0396846421,0.4338293672,-0.0368303284,0.1316811144,-0.3620532155,0.4557719827,-0.3318060935,0.1814910024,-0.0390531309,0.0716069788,-0.1901975721,0.1185606867,-0.1962600052,0.0315162726,-0.1900782883,0.1686164588,-0.3494861126,-0.0849615633,-0.1358319819,0.0537204221,-0.0379853062,-0.4488982558,-0.0353503563,-0.2522087991,-0.0285261925,-0.0763801411,0.2768583298,0.0834865198,-0.3063120246,-0.0223295875,0.1181501448,-0.6360762715,-0.141303882,0.2398983389,-0.2291075885,0.2597793341,-0.1677664965,0.2862717211,0.1563287526,-0.2725051939,-0.4217221439,-0.1164945364,-0.2695347071,0.4236685038,-0.0534157977,-0.090876244,0.0705735385,-0.0361696072,-0.0721378848,0.1831407845,0.1319004893,0.2432777435,-0.2878929079,-0.3644507825,0.5293803215,0.2395403832,0.0658148378,-0.0783991963,-0.1571026444,-0.0383710712,0.1621546894,-0.2566770911,0.1221959144,0.0505723581,0.4054592252,-0.0996585637,0.2301550359,-0.2327405661,-0.5437594056,0.3214948773,-0.0972875357,0.1614063382,-0.1610365957,0.0357198529,-0.0397969298,-0.0607667454,-0.3407133222,-0.2189716697,0.1890426129,-0.230576694,0.2103099674,0.1103476807,-0.0933237523,0.0938846618,0.4587007165,0.1336646527,0.0319539271,0.3562092483,-0.1763494611,-0.3174824417,0.2404371202,0.0770044401,0.0751899183,-0.1000207514,-0.0663075447,0.3032898307,0.1804076284,0.37339136,-0.1613334268,0.3018469214,0.1253358722,-0.5215623975,-0.1006548926,0.4425125122,0.1099250168,-0.1337175071,0.1248511523,0.1103714034,0.2693481743,0.0308032781,0.1520830244,-0.1175029799,0.1235675067,0.0351256318,0.1010116711,-0.0817971602,0.0243122615,0.0937995315,0.1685349494,0.0213546865,-0.0607167594,0.2339461446,0.363301456,0.0720332786,0.3373264074,0.1363614053,-0.2839112282,-0.1312531084,0.0659226477,-0.5286070108,0.263975054,0.2446134388,-0.0513978526,0.09380766,-0.2473659217,-0.1190732419,0.1849914789,-0.09476199,0.1906589717,0.3164351583,0.0882510841,-0.0662374198,-0.3864838779,-0.1912595928,-0.0982206017,0.3551743329,-0.3472343385,-0.2545511723,-0.0955661312,0.3142175376,0.2870850265,-0.0719521716,-0.2914169431,-0.2105284035,0.2494694591,0.5105282664,0.0870300084,0.1683196723,-0.0505847372,0.0320817679,0.1617701203,-0.2861552536,-0.3745260537,-0.2782730162,-0.4979173839,0.1265777349,0.0173711926,0.0569923334,0.2521235943,0.017532682,0.1189049631,-0.2761261761,-0.5717075467,0.0195273105,-0.2541222274,0.2961116135,0.267426759,0.3653386235,-0.0024966912,-0.1739783585,0.3474290073,-0.5429822803,-0.2628821433,0.0644827858,-0.1493652016,-0.0377887003,-0.2412446439,-0.1955478191,0.0689712763,-0.2338234186,0.4771602452,0.0582995154,-0.077684626,-0.0930535495,0.2367125154,0.1903440952,-0.1197477952,0.0196016673,-0.1057230979,-0.2424111366,0.4232416749,-0.1305646002,-0.3693697453,-0.0595454127,0.0262689702,0.3354455829,0.2058672458,-0.2777559161,-0.2156028152,-0.2062187642,0.4504065514,0.1560463607,0.3061792552,0.534140408,0.1000198349,-0.0175255295,-0.2492809594,-0.2447465211,0.2127530277,0.0567573197,0.0985297859,-0.1080293506,0.0211731363,-0.0744062364,0.6856952906,0.2347756475,0.1267780662,0.2585782111,-0.1518139839,0.0764069706,-0.2973735631,-0.0133580053,0.1752147675,0.1543961316,0.115083091,0.4884473383,0.0112211248,-0.3030709624,-0.3405341506,-0.0163158774,-0.2733669877,-0.4006022513,-0.1156450585,0.35567379,0.214151606,-0.0556888543,-0.0006395359,-0.1483120322,-0.1827561706,0.2967506349,-0.0072427699,-0.0256605428,0.057108067,-0.1521092355,-0.3151553273,-0.328956455,0.087934643,0.165817067,0.2093942016,-0.0940652862,-0.1421719044,0.2307176441,-0.252114892,0.5581433773,-0.2723191977,0.0665509403,0.0655351579,-0.0005386861,-0.5107936263,-0.1990357488,-0.2789282203,-0.1900371462,0.0027415238,0.1434878856,-0.8148322701,-0.1501693577,-0.0170536768,0.0582493283,-0.0053717834,-0.0284521207,-0.3249716163,-0.3300275505,-0.2102109492,-0.2348714173,0.0176281445,0.1313700676,-0.2182593048,0.0490581095,-0.1593446136,-0.0190639719,-0.003210223,0.5185790658,0.0622739159,-0.2540459037,0.0881812721,0.2997968793,-0.2922690213,0.2169052064,0.4214388132,0.3327272534,-0.372635901,0.0617817827,0.0982978344,0.214127332,0.2341071814,-0.1033746228,0.1758540273,-0.2350974232,0.1757507324,-0.2090771496,-0.0389076546,0.6191834807,-0.1155053675,-0.428322047,0.1293959022,0.433989197,0.1009066626,-0.0819729045,-0.020034181,0.550339818,-0.4129359424,0.059869729,0.0418654867,0.9076311588,-0.0629235432,0.2472717762,0.6524153948,-0.1866419315,0.43964535,-0.1180082858,0.0496772788,-0.5093026757,0.1383062899,-0.0762239993,0.0113007491,0.3197404444,-0.0736124665,-0.4296858609,0.1145518795,0.0697153881,0.39747715,0.2144890428,0.3926328719,0.0419497415,-0.073696278,-0.053067524,0.1322509646,-0.1223123074,0.2743061483,-0.1285284907,-0.279614687,-0.0910420045,-0.3897846043,-0.4185862243,0.0127870925,-0.533516705,0.2520986497,0.1780143231,-0.0908802003,0.418019563,-0.0899131373,0.0837225094,0.0134004848,-0.5699629784,0.2557992637,-0.3203342259,-0.2919153571,-0.0274421535,0.0375049897,0.1340354979,-0.220176205,-0.4835812151,0.1174221337,-0.0550971143,-0.2445966601,-0.2249414921,-0.2855810225,-0.1274466068,-0.157169655,0.1741639823,-0.0563730896,0.0269269738,-0.1763037145,0.1012103781,-0.0728914291,-0.1337454021,0.1904887408,0.162714377,0.0285729319,-0.0645867437,0.3040250838,-0.1398829669,0.0092501258,0.4439619184,0.1605631113,-0.2012172639,-0.335215807,-0.1562720984,0.1725357473,-0.1964999139,-0.0683013499,0.4442242682,-0.0354581885,0.1382122785,0.6535784602,0.1216332018,-0.1547073573,0.0448694117,-0.1794140637,-0.075460583,0.4141306281,0.0167075396,0.3614344597,-0.0614759475,0.3435841203,0.1603622586,0.1033100709,-0.3085553944,0.1960342973,0.0800771415,0.0753422976,-0.0514179654,-0.2330048233,0.2034725845,0.1268011183,0.1486518383,-0.0089972652,-0.2830576301,-0.198303625,-0.0317386575,0.1284049004,0.2010845691,-0.1915730387,0.0715099052,-0.0326537155,-0.0255067516,-0.0739990622,0.4183784723,0.0116327098,-0.1399015039,0.1661022156,0.121614337,0.060937345,-0.2573490441,-0.0761313364,-0.1215017363,-0.018378906,0.1265917271,0.0972920284,-0.2283302248,-0.1099855304,0.3120474219,0.4752480686,0.0638492778,0.0714169592,0.2174027562,-0.3540939689,-0.1691927761,-0.0960473716,0.2554406226,0.3597443104,-0.0306040794,-0.0581423938,0.2677448094,0.1974910647,-0.2309544235,-0.012707348,0.6733983755,-0.0153030939,0.1685168594,0.1452654451,0.1580358297,-0.2852303982,-0.2165612876,0.0746904537,0.2761015296,0.0822916999,0.0622879192,0.3978444934,-0.0982525423,-0.0740163624,0.1238050014,0.128155604,0.1585317105,0.2188188583,-0.3485071361,0.0422690175,-0.2654130459,0.2607809305,0.3450279534,0.123630546,-0.2138407677,-0.0470369235,0.0708537847,-0.179735139,0.0121128894,0.4386491179,0.1415000856,0.0579469241,-0.4869778454,0.1763892174,-0.1399082839,-0.1317978054,-0.001620808,-0.2269510627,-0.0348796025,-0.1186161563,-0.1176229194,-0.3630170226,0.3812019825,0.1194360554,0.0343099125,-0.415478766,0.1770950705,0.1217279136,0.4034839571,0.1384263784,0.2604581714,0.2337625027,-0.3612535894,-0.1436404586,0.5182735324,0.3549620807,0.2138041854,-0.4979650676,0.0287269577,-0.1955886483,-0.1093650833,-0.0893893465,0.3242752254,0.0754805431,0.0665356368,0.0065159001,0.0878771394,-0.1414429843,-0.3864957988,-0.1149380356,-0.0126780886,-0.2656838298,0.0635332316,-0.2135778517,-0.0108951116,0.0580915548,-0.1321555078,-0.1663777828,-0.0273609869,0.0539070442,0.0987845585,0.0450497232,-0.1749381125,0.0666020811,0.0867039189,0.5437386036,0.2416298836,0.2390600443,-0.149918735,-0.1806225479,-0.3015088141,0.2141608298,-0.2693429589,-0.0514820367,0.0098815579,0.2190559953,-0.16715765,0.3232294619,0.1952762455,0.2796862125,0.00488265,0.2342150211,-0.2166932225,0.2024385333,-0.1751505733,-0.1685825437,-0.0731466487,-0.3066116273,0.2222513109,-0.3252540827,0.065374881,-0.1551135778,0.4660894573,-0.0019075178,-0.4149541557,0.3522657454,0.4237250388,0.4327597916,0.0837661847,0.0076497099,-0.4300166965,-0.2715637088,-0.297950536,-0.0146119939,-0.1282048076,0.359266907,0.0515940376,-0.3163743615,-0.0666829944,0.1658657938,0.1844336689,0.1959703118,-0.0079905521,0.0761536062,0.0065748035,0.0738817826,-0.1974889487,0.1553751528,-0.0027347486,-0.2165616155,-0.3409823179,-0.6350567341,0.352094233,-0.6040747762,-0.2789495885,-0.1949500889,0.2212252319,0.0911131203,-0.2894885242,-0.5017029047,0.1414109915,0.1455617696,0.0892093182,-0.3809329271,0.3301685154,0.0929168314,-0.1507183015,-0.0393585637,0.3450451493,-0.000523483,0.0334852971,-0.0949892774,-0.1373833567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/707","title":"Requirements should specify pyarrow<1","comments":"Access granted for the colab link. ","body":"I was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.","comment_length":6,"text":"Requirements should specify pyarrow<1\nI was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.\nAccess granted for the colab link. ","embeddings":[-0.2640045285,-0.1825000942,0.0247028172,0.2080772668,0.0334078148,-0.0568866469,0.018309813,0.1705020666,-0.051373437,0.0628898516,-0.0578348935,0.3004088998,0.0103776539,0.0493106544,0.1372278035,-0.3382230401,0.1985091269,0.6012260318,-0.3903082013,-0.0117722517,-0.1071513295,0.277674973,-0.3115366399,0.2668267488,-0.3911313415,0.1409055591,0.0372027531,-0.0074684555,-0.1455481201,-0.5624930263,0.4644871056,0.2674184144,0.0232731737,0.5105575323,-0.000118152,0.0626951307,0.1689137667,0.0015901406,-0.0691752508,-0.0360992365,-0.0351523384,-0.0729539543,0.4277050197,-0.0911680833,-0.2840470374,-0.0597744584,0.0152262375,0.2307688743,0.3799954057,0.1311753839,0.1249586046,0.4022147059,0.6234029531,-0.0591734648,0.55938375,-0.21055758,-0.2087494731,0.1733294874,0.2411623895,-0.2995124757,0.1012863591,-0.0731993616,0.1831062734,0.0921853185,0.0832972601,-0.0062339511,0.3094590604,-0.2671729624,-0.0847477168,0.3785930574,0.3645021021,-0.2019049078,-0.3395178616,-0.1651227176,-0.1809519976,0.0457897186,0.2400628179,-0.264613241,-0.3053819239,0.3500322402,-0.0254025757,-0.1578205973,-0.2061026841,0.0427111462,-0.3993184268,0.6330034733,0.1217722744,0.189049691,0.1084371209,-0.1207811981,-0.2246674448,0.2483856082,-0.0336769745,0.0602966994,-0.280421555,-0.1262994856,0.2551509142,0.2685022652,0.2674371302,0.0109174969,-0.1362269819,0.1159326658,0.1953293234,0.2715395093,0.1356712729,0.0448755436,-0.2389433235,0.1318921894,0.3234206438,0.0240907781,0.0976330712,0.0692867711,-0.0618452691,-0.2690609992,0.043004401,-0.3400786519,0.310482353,-0.1881036311,-0.3721840978,-0.1729421169,-0.1333751529,0.1471593976,-0.015003616,0.1085261032,0.0445757098,0.3093115389,0.0646194667,0.1430523843,-0.2932569981,0.0013063619,-0.1354033649,-0.1897899508,0.0189166274,0.1017965898,0.3523018658,0.0441264212,0.3299775422,0.2019640207,0.0673893094,0.4334993958,-0.068460159,0.1157963052,-0.3825947344,0.4894653261,-0.3330786228,0.186836049,-0.0436479375,0.0435026363,-0.1896668226,0.1388448179,-0.2169218361,0.0475114509,-0.1925468594,0.1535531133,-0.3837744892,-0.1018717363,-0.1582276076,0.0733761042,-0.0264303815,-0.4709495604,-0.0271199625,-0.2830387056,-0.0348394848,-0.0807878077,0.2809008956,0.1106805354,-0.3640484512,-0.0134712309,0.1015510932,-0.613037467,-0.1257239133,0.2249627113,-0.2526943088,0.2793261707,-0.1900654733,0.3094345927,0.1549070179,-0.2540619969,-0.4199714363,-0.1402296126,-0.2621178329,0.4392996132,-0.0526423119,-0.1308819801,0.0571928695,-0.04758396,-0.0124159791,0.144418329,0.1292423308,0.2089393288,-0.2873423398,-0.3793187439,0.550583303,0.1957227588,0.0593589284,-0.0628846288,-0.1617864966,-0.0708619878,0.1840371937,-0.2281886339,0.0989016891,0.0178445484,0.4258736372,-0.1256642789,0.1921008974,-0.2277698815,-0.5535191298,0.3316273391,-0.0752472281,0.1770821214,-0.184125036,0.0507137291,-0.0509962067,-0.0359207578,-0.3395514786,-0.2181845158,0.1747591496,-0.2470089346,0.2021829337,0.131853953,-0.0847525597,0.0917461216,0.4679502249,0.1318423152,0.0473705828,0.3581856191,-0.1776650697,-0.3521677256,0.2274091691,0.0816311166,0.084101595,-0.0874470249,-0.0819165334,0.2887557745,0.2040927559,0.3674218357,-0.1579416096,0.2927808464,0.1251823604,-0.5347675681,-0.0921271518,0.4410662949,0.107898809,-0.1125908867,0.138616547,0.1178340241,0.2655041516,0.0160754081,0.141372174,-0.1135568842,0.1182827502,0.0596063957,0.0869314,-0.0635824651,0.0249249674,0.1095331237,0.1459946334,0.0369430333,-0.0381435417,0.2296832502,0.3299396932,0.0545891225,0.3232450783,0.1171465293,-0.2885356247,-0.146664694,0.0940232202,-0.546282351,0.3037647605,0.2316075861,-0.061521668,0.0919838995,-0.2672388256,-0.1340173334,0.16674456,-0.075330317,0.2131572515,0.301574856,0.1094832048,-0.057512749,-0.3743309975,-0.2043806762,-0.0831646472,0.3292681277,-0.3665182292,-0.2274840474,-0.112080045,0.3007388413,0.2766219676,-0.0859723613,-0.2964093387,-0.2233691663,0.2391152531,0.5081591606,0.1429130584,0.1486463994,-0.0795440003,0.0362967737,0.1617348492,-0.3212462068,-0.3873093128,-0.2732061148,-0.4676873088,0.1054064184,0.007660069,0.0391306728,0.2460561693,0.0237638447,0.1016461849,-0.2425812185,-0.5656111836,0.026364021,-0.2614628077,0.3213639855,0.2565511763,0.3441059887,-0.0135897789,-0.1424618512,0.3339657187,-0.5064530969,-0.2627179325,0.0347051769,-0.1838896573,-0.013372574,-0.224862203,-0.1690051407,0.0707008019,-0.1985435933,0.4534217119,0.06592796,-0.0857119039,-0.0507137515,0.2667607665,0.2038111985,-0.1204631105,-0.0051770848,-0.1060460731,-0.2177686244,0.4272252619,-0.1125899404,-0.3516944945,-0.0267899428,0.0174157489,0.3518814445,0.2199183553,-0.2742263377,-0.2319765985,-0.1878056526,0.4793027937,0.1513108611,0.2765623927,0.5465868115,0.0883967951,0.0069768322,-0.2390901595,-0.2716271579,0.1582140177,0.067517817,0.0896215439,-0.1138579547,-0.0063391314,-0.0854398534,0.6846992373,0.2212386131,0.1325291395,0.2649469376,-0.1608120501,0.0886531547,-0.2707132995,-0.0260698069,0.1666799933,0.1526383162,0.0678483248,0.4624449313,-0.0095654745,-0.3078714609,-0.333961457,-0.0137895914,-0.2605930269,-0.3953533769,-0.132596612,0.3443104625,0.1936661303,-0.0320131555,0.0173314866,-0.1554399878,-0.1932237744,0.2747856379,0.0024021023,-0.0154053923,0.0693521798,-0.1165623516,-0.3573284149,-0.3245784938,0.0749714747,0.1739663035,0.2043685764,-0.0804925412,-0.1200086102,0.241010651,-0.2460516542,0.5684053302,-0.25218606,0.065993838,0.0878360868,0.0231783707,-0.5199002624,-0.2485121936,-0.2673258483,-0.2003259808,0.0019616394,0.1321568936,-0.7699299455,-0.1349441111,0.0160106029,0.0429690443,0.0254197195,-0.0277173501,-0.3403947055,-0.3557534814,-0.171530664,-0.2400254905,0.0476204008,0.1283057928,-0.2292140275,0.0626842454,-0.1202869564,-0.0202972088,0.0068394593,0.515656054,0.0356265455,-0.2251554281,0.0457167476,0.2897121608,-0.3025925457,0.2273629606,0.4176037908,0.3635075092,-0.394479692,0.0600364543,0.1315339506,0.2427432388,0.2666661441,-0.1338459551,0.155461207,-0.2488211393,0.144179821,-0.181955725,-0.0170387756,0.6404119134,-0.1524086893,-0.4390966594,0.1479772329,0.4384315312,0.1381115615,-0.0688195005,-0.0335017368,0.5509485602,-0.3969899416,0.0266494807,0.0471438915,0.9068719745,-0.069555454,0.2456902266,0.7057409883,-0.1641305685,0.4846913815,-0.1276680529,0.0798361301,-0.4921944439,0.1543032676,-0.0874926597,-0.0180054083,0.345639348,-0.1010526419,-0.4325716794,0.1215936542,0.0453118011,0.4206387699,0.2381878644,0.3708129525,0.0645726323,-0.0582233295,-0.0536208525,0.1187885255,-0.110498853,0.2907907069,-0.1490032524,-0.2534666955,-0.1082738712,-0.3817815781,-0.4406842887,0.0247161109,-0.5484663248,0.2193097919,0.1956246942,-0.1236229613,0.4255849123,-0.0919874087,0.0700304061,-0.0218029041,-0.5774720311,0.2410622239,-0.3283961713,-0.3249269724,-0.0168522689,0.0160340406,0.1424678117,-0.2004353404,-0.4658229351,0.1338423043,-0.0540548712,-0.2076292485,-0.2341814637,-0.2612690628,-0.1499400735,-0.1801768839,0.2061419487,-0.0947233364,0.0852411911,-0.1790660471,0.0874984339,-0.0511961058,-0.120016329,0.1713012904,0.1432052255,0.0203042869,-0.0751672536,0.3200232387,-0.1242592335,0.0015585679,0.4473478198,0.1689130217,-0.1917541325,-0.3090502322,-0.1596631408,0.1361554861,-0.1876279861,-0.0640119389,0.4336289167,-0.0374432728,0.1203792989,0.6748108864,0.1317801178,-0.1503217369,0.054472208,-0.1782490611,-0.0511736572,0.4023901522,0.001954735,0.3464175463,-0.0611957014,0.3917293251,0.1604878604,0.0819919929,-0.2824753225,0.1933896095,0.1088349447,0.0439873114,-0.0597863384,-0.2192470133,0.190731287,0.127321586,0.1434732676,-0.0032835873,-0.2474149913,-0.1800688207,-0.0254747961,0.1384269446,0.1964617968,-0.2059121579,0.0826283991,-0.0493857227,-0.0090508256,-0.0824080855,0.3874429166,0.0158422347,-0.1408599466,0.1461486816,0.1456985921,0.0501231141,-0.2568610609,-0.0556547269,-0.0972851366,0.0148074506,0.1389526427,0.1097923964,-0.2210910171,-0.0896321014,0.2877051532,0.4870175719,0.0243346207,0.0843614563,0.1900818944,-0.3717831373,-0.1567863822,-0.1273949593,0.2305163741,0.3896375,-0.0364933647,-0.065403834,0.2632423639,0.1756502688,-0.2014955133,-0.0217096228,0.7180562019,0.0413499735,0.1442599744,0.1641548872,0.1518385857,-0.256200552,-0.2277313769,0.0712770745,0.2802804112,0.1349544525,0.0521216914,0.4194869399,-0.0784267411,-0.0558062606,0.1510203779,0.1136972234,0.1680547893,0.2300292104,-0.3444737494,0.0036181635,-0.2733975053,0.2572287917,0.3534219265,0.0960295871,-0.2179039419,-0.0434323102,0.080022499,-0.218395263,0.0045721699,0.4608478546,0.1482567489,0.0641990155,-0.5053674579,0.1826054007,-0.1403050572,-0.1281055063,0.0065520243,-0.2422787547,-0.0179849677,-0.1201921105,-0.1376276612,-0.3610452116,0.4276951849,0.1090376973,-0.007999341,-0.4191276729,0.1947673708,0.1165578067,0.3725156188,0.142149806,0.2442077547,0.2000763565,-0.3698288202,-0.1661584079,0.51447016,0.3726584315,0.2405040413,-0.4919936657,-0.0274051558,-0.1933962107,-0.1078368723,-0.1031218991,0.2831028402,0.0954339206,0.1105467156,-0.0022192658,0.0780437365,-0.1196467727,-0.3614462614,-0.1206302419,-0.0309411082,-0.2924908996,0.0651671067,-0.2311267257,-0.0109412754,0.0729457662,-0.0941872895,-0.1422733516,-0.0063423938,0.0627510324,0.1302629113,0.0435019694,-0.1565409601,0.0517775416,0.088344574,0.5389368534,0.2632124126,0.2757868469,-0.1789373606,-0.1609123051,-0.2972084284,0.2488731444,-0.2938908041,-0.0626605675,0.0041398206,0.2197458744,-0.19699727,0.3427710533,0.1688167602,0.2962501347,0.0238844212,0.2132087201,-0.2030909061,0.1639484018,-0.177874893,-0.2063555121,-0.0891958848,-0.2804979384,0.2431616932,-0.3883217573,0.0470606685,-0.1458314061,0.4335891306,0.0176427942,-0.3936564922,0.3618168533,0.4307121634,0.4717588127,0.1147362143,0.0199389998,-0.449865371,-0.2684952915,-0.3030336201,0.0024801204,-0.1619077027,0.3303642571,0.0322031192,-0.2877375185,-0.0704602227,0.2043200582,0.1888189167,0.2193284631,-0.0171949808,0.090707399,-0.0158310421,0.0904022008,-0.1888982654,0.1323809773,-0.0070127528,-0.1748738587,-0.3533162773,-0.6101939678,0.3555683792,-0.6441082954,-0.2690891922,-0.2278824002,0.2118414938,0.0992330164,-0.2858141661,-0.4798710942,0.1209515557,0.1649071723,0.1052860543,-0.4051917195,0.3210611939,0.1087444052,-0.1447425038,-0.0397307724,0.357611835,-0.0105961571,0.0434550084,-0.0862424672,-0.1204055622]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/707","title":"Requirements should specify pyarrow<1","comments":"Thanks for looking at this @punitaojha and thanks for sharing the notebook. \r\n\r\nI just tried to reproduce this on my own (based on the environment where I had this issue) and I can't reproduce it somehow. If I run into this again, I'll include some steps to reproduce it. I'll close this as invalid. \r\n\r\nThanks again. ","body":"I was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.","comment_length":56,"text":"Requirements should specify pyarrow<1\nI was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.\nThanks for looking at this @punitaojha and thanks for sharing the notebook. \r\n\r\nI just tried to reproduce this on my own (based on the environment where I had this issue) and I can't reproduce it somehow. If I run into this again, I'll include some steps to reproduce it. I'll close this as invalid. \r\n\r\nThanks again. ","embeddings":[-0.2600724101,-0.2294466645,0.044546254,0.2166786045,0.0749733075,-0.0639310703,-0.0022501727,0.1691178232,-0.0127054844,0.0657434016,0.0208967216,0.2901545167,-0.0119958557,0.0165852979,0.1564203203,-0.3912960589,0.2354253829,0.5927878022,-0.3565097153,-0.0511188917,-0.0974970087,0.2818730175,-0.3155817389,0.2457960546,-0.4022749364,0.1330782473,0.0570968278,-0.0212265309,-0.1788376272,-0.5676388741,0.460424453,0.1918254197,-0.0006512939,0.4826521873,-0.0001226647,0.0562133119,0.1434753388,-0.0056539029,-0.0374005176,-0.0778044462,-0.0551672168,-0.1225150526,0.4323880672,-0.0663315579,-0.2351198643,-0.061791569,0.0275407284,0.206842199,0.3814501464,0.1206037775,0.0818247721,0.3831529319,0.6507722139,-0.0118111502,0.5365049839,-0.1792535931,-0.2274965346,0.1885087341,0.2283306569,-0.2860308588,0.0947201476,-0.1229192317,0.1636893749,0.1266192943,0.1389255226,-0.0354292691,0.3300071359,-0.3048428893,-0.0674957037,0.4033079743,0.3782411516,-0.1925992668,-0.3459409773,-0.1335900426,-0.1589212865,0.0724678636,0.2532921433,-0.1948823631,-0.2869555652,0.3625525832,0.0005612909,-0.1138784215,-0.2268296629,0.1020552516,-0.4096747041,0.5994551778,0.1086408868,0.2016237974,0.0570896231,-0.1125908643,-0.2647919357,0.2614644468,-0.0001149078,0.0883189365,-0.3127442002,-0.059696693,0.2298751026,0.2597002983,0.2728823423,0.0418971144,-0.1688748449,0.103894867,0.2539030313,0.2548690736,0.1296690404,0.0837187991,-0.2245155126,0.1010372341,0.2753295302,0.0285481345,0.1286762804,0.1025176048,-0.0497893393,-0.2957265377,0.0719251782,-0.3837428689,0.3628817797,-0.2149203271,-0.3580614328,-0.1517550796,-0.1998100132,0.1401495785,-0.0616697855,0.1240726262,0.0389086828,0.3743239343,0.067445986,0.1658492237,-0.2863715589,-0.01235676,-0.0845561847,-0.1550300866,-0.0043634567,0.0966148898,0.3515407741,0.0227618832,0.3478555381,0.2659069896,0.0562620573,0.3823370039,-0.0564059503,0.1206506863,-0.3797757924,0.5211027861,-0.3667622507,0.2458288372,-0.0599478856,0.0836729333,-0.216229111,0.1611863524,-0.2448174208,0.0681528971,-0.2135563344,0.1039168462,-0.3787121773,-0.0746729821,-0.1572631299,0.0571043417,-0.0352057032,-0.408365041,0.0018554332,-0.3095980287,-0.0211989656,-0.0501910709,0.2657848597,0.0732726231,-0.4110647142,-0.0553536266,0.1464295983,-0.6278679967,-0.0937169343,0.2602693141,-0.2143107951,0.2615125775,-0.171292156,0.3319961429,0.125491783,-0.2687300146,-0.4613823295,-0.097947903,-0.2837933898,0.4334025085,0.0229181219,-0.1663631499,0.0577699468,0.0078358976,-0.0059621758,0.126890704,0.180409357,0.2102524042,-0.2970997393,-0.3904989362,0.5659132004,0.2077561617,0.077319935,-0.0564301051,-0.122668229,-0.0641804934,0.1925192922,-0.2520045638,0.1266434491,0.0647221133,0.4038166404,-0.1244884953,0.1586154401,-0.2392110825,-0.5577841401,0.3368268311,-0.1081323028,0.1521909237,-0.1528612077,0.0712798536,-0.093885377,-0.0258187391,-0.3526058793,-0.2056111097,0.1142313704,-0.2597649097,0.1468533725,0.1433563679,-0.0995418504,0.0853913948,0.4340462983,0.1081717461,0.0264091697,0.3328432739,-0.1912792921,-0.3221775293,0.2326421142,0.0861702338,0.1178943291,-0.0763146281,-0.0634006858,0.2662022114,0.1697747111,0.3685865402,-0.2167717069,0.3443760574,0.1450622827,-0.5561097264,-0.0559157953,0.4561325312,0.1285793483,-0.1256193668,0.1880304813,0.13425605,0.3463529944,0.0614618249,0.1090695262,-0.1019291952,0.0950394198,0.0584769882,0.0756847337,-0.0650201663,0.0488045961,0.1172883436,0.1151427627,0.0534863845,-0.0210108329,0.1688484401,0.33660537,0.0282275639,0.3812072575,0.1301885396,-0.300298661,-0.1400309354,0.1055176184,-0.543846488,0.2666546106,0.1710639149,-0.1121409088,0.0792613775,-0.2599853277,-0.1251826286,0.1742731333,-0.0578921512,0.2054306269,0.3278053701,0.0803237706,-0.0520259328,-0.3668719828,-0.1731809378,-0.0534409061,0.3192165494,-0.448913008,-0.2553728223,-0.1153847054,0.4096748233,0.2679757476,-0.0832600445,-0.3234757483,-0.15157184,0.1907196343,0.487103194,0.1732062846,0.1628544331,-0.0860742256,0.0134216864,0.1226308197,-0.3300580978,-0.3696283698,-0.2419108152,-0.4619723558,0.0492066704,-0.0058045071,0.1037280783,0.1937630028,0.0053327945,0.0835207999,-0.2515048087,-0.5143217444,0.0052005085,-0.290040195,0.3554943204,0.2452578694,0.4037764966,-0.0252845567,-0.1328048855,0.3036792576,-0.4741431475,-0.2775180936,0.0788417161,-0.2127462029,0.003696393,-0.2109628916,-0.1804171801,0.1119275019,-0.1781096458,0.4739205241,0.0716430992,-0.0863906965,-0.0142506277,0.2749122679,0.2249675542,-0.1203625798,0.0545447916,-0.1089293882,-0.2190484703,0.4475554228,-0.0897278339,-0.3637834191,-0.0060731168,0.0331213847,0.3714805245,0.2010559291,-0.3030137718,-0.1785420924,-0.1841654629,0.5038659573,0.1345975399,0.269472152,0.5878297687,0.0813112706,0.0553361773,-0.2243198901,-0.2839870751,0.1419700235,0.0696283057,0.0827935934,-0.0956305712,0.0275062863,-0.0847217739,0.7192692161,0.201434657,0.168365255,0.2772160769,-0.1349285096,0.0249315538,-0.2609725296,-0.0178890284,0.161109671,0.1278343797,0.1197075322,0.4493446052,-0.0099443197,-0.2933170497,-0.3085644245,-0.0194654632,-0.2585767508,-0.3600152135,-0.1014119461,0.3348831236,0.2495855987,-0.0422803536,0.0532908589,-0.1134441122,-0.1968870014,0.3039842844,0.0223848745,-0.0011477702,0.0949145257,-0.1835049689,-0.3080985248,-0.294521749,0.0470187142,0.1664323509,0.2763676047,-0.0939330831,-0.069214642,0.2698101699,-0.2304870486,0.5348501205,-0.2213624567,0.1032870933,0.0726944655,0.0049912091,-0.5632783175,-0.2405838072,-0.2651919425,-0.2444620728,-0.02852742,0.0840953365,-0.7947930694,-0.161608398,-0.0031478058,0.0745918378,0.0348151326,-0.0079380311,-0.3810628951,-0.3311390281,-0.1961113662,-0.2764700949,0.0251088291,0.1427241713,-0.1972413659,0.1291232407,-0.1272191554,-0.036442861,-0.0063953972,0.4384379983,0.0167681184,-0.1937925518,0.0190713704,0.2959504724,-0.293839097,0.216850996,0.3549230695,0.3580702841,-0.3900272548,0.0701681301,0.1502629817,0.2579312325,0.2847651839,-0.1211136729,0.1475898474,-0.2307736427,0.0978849977,-0.1897483915,-0.1056676209,0.7092196941,-0.1743064225,-0.4266014993,0.1528984606,0.4809176028,0.1264927238,-0.0636676252,-0.0694280043,0.6567612886,-0.3557112515,0.0084249293,0.0415333398,0.9474662542,-0.0492686927,0.2523489296,0.7000189424,-0.1713727564,0.4334260225,-0.1260406226,0.0683534369,-0.4909998775,0.1541275084,-0.1144988164,-0.0247976426,0.397518754,-0.0940966383,-0.3963697851,0.1129442826,-0.0252527799,0.4111185372,0.2764179707,0.3371881545,0.0277610514,-0.0784793347,-0.0978203192,0.0642561913,-0.1611084044,0.3095940053,-0.1548897624,-0.2225808203,-0.0984712467,-0.3774934411,-0.4782420695,0.0154625149,-0.5383328199,0.2571948767,0.2164738327,-0.1365374029,0.4455132782,-0.0758886784,0.0518844724,-0.0593140721,-0.6055544615,0.2674524784,-0.3270075321,-0.3188637495,0.0530103445,0.0074724946,0.1546264142,-0.224529162,-0.4871273935,0.1202137098,-0.1397122294,-0.1552681625,-0.2389632314,-0.2483064681,-0.1534742266,-0.185462907,0.167465508,-0.1240853444,0.1325313449,-0.1715200841,0.0447930694,-0.0898996964,-0.1219869629,0.1516408473,0.1526881903,-0.0129656661,-0.0802376568,0.3831326962,-0.1722148359,0.0095832553,0.4769024253,0.1530897468,-0.1831881553,-0.2931614816,-0.1196411327,0.1579947621,-0.248840183,-0.1189776361,0.4333809912,-0.0437363684,0.0681127161,0.7004838586,0.1763393581,-0.1456474066,0.0540268645,-0.2041210383,-0.0548990108,0.3778792918,0.0031667338,0.31283921,-0.0913680345,0.3753881156,0.1995435804,0.0651790202,-0.2247831672,0.2147883475,0.1590881944,0.0638614744,-0.0293738078,-0.1718463451,0.190786168,0.1145640165,0.1144405752,-0.0069732391,-0.2571875453,-0.1317492127,-0.0618795976,0.1686325073,0.2130324394,-0.1596172899,0.119885385,-0.0422661602,-0.0043332102,-0.0849364176,0.3771917224,0.0965899229,-0.1567594707,0.1390298456,0.1304193586,0.0886541232,-0.3128328025,-0.0523621626,-0.1344356686,0.0260586888,0.1303717494,0.0753604472,-0.2008762956,-0.0683857799,0.2166530192,0.4980845451,0.0106346682,0.0860229135,0.2288598716,-0.4094147086,-0.1212662756,-0.1192341,0.1960599124,0.3971195817,-0.0784616992,-0.0645659491,0.21815373,0.142302528,-0.1961407065,0.0011727202,0.6975147724,0.031433966,0.1126797944,0.1705732197,0.182759285,-0.3354976773,-0.2131242603,0.0825447366,0.2384019047,0.0979657546,0.0413377546,0.4126865566,-0.0530695952,-0.0369048715,0.1738544703,0.0716801509,0.1724842042,0.1892310232,-0.3467921019,0.0061219162,-0.2893975973,0.3153369725,0.3084582686,0.0791492239,-0.2358136326,-0.0338427573,0.0548740067,-0.1931436956,0.0286297388,0.4667815566,0.2011779845,0.0236374401,-0.5152425766,0.1780370027,-0.1269298345,-0.1181456223,-0.0037992788,-0.2237559855,-0.0253534634,-0.1171616018,-0.1390086412,-0.3470347822,0.4516313374,0.1168604493,-0.0679067597,-0.4596051276,0.1974963695,0.1343003809,0.4040614963,0.1194643825,0.2511403859,0.1893767267,-0.3837008476,-0.1953844726,0.5415599346,0.3914606273,0.2434290797,-0.4737422168,-0.0138548613,-0.1702418625,-0.101205647,-0.0652632788,0.2659266889,0.1034448072,0.1449329108,0.000227833,0.0285556875,-0.0796172768,-0.3607688844,-0.0961584002,0.0007537874,-0.3436560333,0.0800774693,-0.2495083809,-0.0392743871,0.0575807877,-0.0598285049,-0.1600278914,0.0024758796,0.0466144346,0.1515882313,0.0417585,-0.1684443802,0.0292738341,0.0942955986,0.5413830876,0.2419524193,0.2027942538,-0.2083523422,-0.1428866982,-0.2838066816,0.2257681042,-0.2848079801,-0.0250259824,0.0288371015,0.2524174154,-0.2243113071,0.3560341001,0.1580049992,0.297983855,-0.0117505537,0.2616168857,-0.2010524869,0.1591654867,-0.1653672457,-0.1854951233,-0.0884958655,-0.2935526073,0.2803270519,-0.4329345822,0.0006533143,-0.1449590176,0.4242351055,0.0093396129,-0.3736906648,0.3833175004,0.4528143108,0.476791352,0.1089384407,-0.0122996271,-0.4657028913,-0.2472915798,-0.2630336881,0.0687008277,-0.1872000992,0.3646776676,0.0250796564,-0.3029357493,-0.0547212996,0.238977015,0.1853815317,0.2155537456,-0.072176449,0.1199946553,-0.0422298014,0.1143095195,-0.2109838426,0.0943513736,-0.0189971495,-0.1899803728,-0.3945068419,-0.6121429801,0.3251288831,-0.6871471405,-0.2863022387,-0.2247801721,0.2356779128,0.0586220399,-0.26669994,-0.4676551521,0.0777356625,0.1692678183,0.135789305,-0.3994818032,0.275660336,0.0832402185,-0.1162421852,-0.0143997008,0.3591381609,-0.0228888132,0.0447546244,-0.062857002,-0.1172901988]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/707","title":"Requirements should specify pyarrow<1","comments":"I am sorry for hijacking this closed issue, but I believe I was able to reproduce this very issue. Strangely enough, it also turned out that running `pip install \"pyarrow<1\" --upgrade` did indeed fix the issue (PyArrow was installed in version `0.14.1` in my case).\r\n\r\nPlease see the Colab below:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/15QQS3xWjlKW2aK0J74eEcRFuhXUddUST\r\n\r\nThanks!","body":"I was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.","comment_length":52,"text":"Requirements should specify pyarrow<1\nI was looking at the docs on [Perplexity](https:\/\/huggingface.co\/transformers\/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error,\r\n\r\n```\r\nmodule 'pyarrow' has no attribute 'PyExtensionType'\r\n```\r\nI traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. \r\n\r\nhttps:\/\/github.com\/huggingface\/datasets\/blob\/e86a2a8f869b91654e782c9133d810bb82783200\/setup.py#L68\r\n\r\nDowngrading by installing `pip install \"pyarrow<1\"` resolved the issue.\nI am sorry for hijacking this closed issue, but I believe I was able to reproduce this very issue. Strangely enough, it also turned out that running `pip install \"pyarrow<1\" --upgrade` did indeed fix the issue (PyArrow was installed in version `0.14.1` in my case).\r\n\r\nPlease see the Colab below:\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/15QQS3xWjlKW2aK0J74eEcRFuhXUddUST\r\n\r\nThanks!","embeddings":[-0.2761943936,-0.1401243955,0.0347475186,0.1603855789,0.048561275,-0.0427087359,0.0266228598,0.198436603,-0.0582807027,0.0600505993,0.0015630063,0.2838506997,-0.0044425442,0.0376977585,0.1520177722,-0.3687959611,0.2303959429,0.5788680911,-0.3444227278,-0.022595562,-0.1056371629,0.2532047927,-0.3011270761,0.2348771393,-0.3952791691,0.1669677496,0.0467276238,-0.0211397596,-0.1742847264,-0.5824115276,0.4924414754,0.252129674,0.0277054142,0.456689775,-0.0001201362,0.0909883007,0.1616440862,-0.0273129307,-0.0626678616,-0.0653416291,-0.0663691536,-0.1003902555,0.4112484753,-0.0659435689,-0.258389771,-0.0569337532,0.0501447804,0.1991277039,0.3331044316,0.1226729155,0.1010975912,0.3785631061,0.6543204784,-0.0274976362,0.5464001298,-0.1852430105,-0.2104420215,0.1501597464,0.2215284705,-0.2979403734,0.0968258232,-0.0911414325,0.1513130516,0.1252644062,0.129845202,-0.0355437733,0.3226665258,-0.30716905,-0.0943656042,0.3945658505,0.3904876411,-0.1932222843,-0.3653137684,-0.1660096794,-0.178064093,0.0701816529,0.2186295837,-0.2005637884,-0.2938903272,0.3531870842,0.0228029191,-0.1070260853,-0.2230653763,0.0897029489,-0.4169590473,0.6115364432,0.1379712522,0.2251069099,0.0881054774,-0.1168655008,-0.2416556031,0.2601368427,-0.0086097121,0.0636655018,-0.3093869686,-0.0776964575,0.2582891285,0.2512825727,0.2540597916,0.0727146715,-0.1410643756,0.1450536549,0.2471518666,0.2898202538,0.1449887007,0.0582231358,-0.2342856973,0.1094237864,0.3053756654,-0.0057278415,0.1311250925,0.0875021145,-0.0547012985,-0.2781450152,0.0926290825,-0.3468330503,0.323202163,-0.1806135625,-0.3758249581,-0.1754792035,-0.1947010756,0.1470202655,-0.0443434045,0.1393225342,0.0568133518,0.380368948,0.0832966343,0.1703187823,-0.2854255736,-0.0176735651,-0.097957924,-0.1771339178,0.0434861667,0.058685489,0.3848065436,0.0592871085,0.319157362,0.2018921822,0.0387462042,0.4158566296,-0.0089400476,0.1414913386,-0.3730438054,0.5116958618,-0.3510389328,0.2050296664,-0.0354541093,0.0312404651,-0.2431348413,0.1497157961,-0.2207836062,0.0337291025,-0.2022808641,0.134383589,-0.4136125743,-0.1025794372,-0.1556591094,0.0494392999,-0.0739996061,-0.4285871983,-0.0487005785,-0.3369171321,-0.0042493721,-0.0796426833,0.2585652471,0.0653273463,-0.414026022,-0.0424384996,0.0710194558,-0.6321167946,-0.1191628501,0.2066752166,-0.2231913805,0.2573313117,-0.1579539329,0.3253785372,0.1259886324,-0.281404078,-0.4366946518,-0.1255472153,-0.2708389759,0.4315644503,-0.0321910977,-0.1144108772,0.0957523212,0.0036066687,-0.0575910397,0.1472731382,0.1389974207,0.2289622873,-0.2556209564,-0.3824260831,0.5881329179,0.2423675805,0.0550807193,-0.0834098235,-0.149246335,-0.036235854,0.1523858458,-0.2435113639,0.1465931386,0.0390842929,0.4210374057,-0.0975347385,0.1797184944,-0.2006237507,-0.5595870018,0.323648721,-0.0932345092,0.1747029126,-0.1643192321,0.0497593209,-0.0593132153,-0.0135820657,-0.3554161489,-0.1940595657,0.1503903866,-0.2420448512,0.1908438802,0.1120263934,-0.0961319357,0.0760718584,0.4792812765,0.1153134257,0.0438901447,0.3394732475,-0.1945205033,-0.3024532497,0.2515620589,0.0911321864,0.111483492,-0.0667078868,-0.0520862713,0.2847637236,0.1765264273,0.3445160985,-0.1805079281,0.3716143966,0.1172270104,-0.5627410412,-0.081143111,0.4557859302,0.1287183315,-0.1143580675,0.1628384888,0.1477029622,0.3197205961,0.0652770549,0.1070592999,-0.0887629241,0.1299800873,0.0553453006,0.0559812747,-0.0689043254,-0.002261241,0.131443128,0.1101041585,0.0435634404,-0.0192339346,0.2190407664,0.3266852498,0.0325320363,0.3579244316,0.1216076463,-0.3350586593,-0.1262318641,0.1116462648,-0.507922411,0.2845572531,0.2100426704,-0.0693620592,0.0497773439,-0.2698375583,-0.1088167876,0.1571920663,-0.0635257214,0.2181335837,0.2919237018,0.1006836444,-0.0767323151,-0.3659601808,-0.2055685371,-0.077467829,0.340467751,-0.4137086868,-0.2654786408,-0.138037473,0.381837517,0.2513831556,-0.1123967245,-0.3306830227,-0.2146785855,0.2353895605,0.494522512,0.1171966121,0.1509263068,-0.1158697009,-0.0013261178,0.1205762476,-0.3492171764,-0.3686798513,-0.2633909881,-0.4776780605,0.0790006891,-0.0042073377,0.0857134536,0.2334501892,0.0129469978,0.0899726152,-0.2458569407,-0.5565886497,-0.0041595125,-0.2848057449,0.349306345,0.2681626379,0.394120127,-0.0177511424,-0.1600650102,0.3386487961,-0.4612849057,-0.2450103909,0.0826767311,-0.1957324147,0.009552585,-0.2275137454,-0.1870778799,0.0804662779,-0.1823906004,0.4641319215,0.074658528,-0.0986429751,-0.0391061939,0.2710894644,0.1867854744,-0.0900549516,0.0501333103,-0.1244107485,-0.2010659873,0.4448416829,-0.1024022028,-0.346373111,0.0103552565,0.0043825815,0.3437042832,0.2179702669,-0.2875938714,-0.1910873652,-0.1717970222,0.4668476284,0.174009189,0.2660770416,0.5662255883,0.116081953,0.0288577732,-0.2397851348,-0.2698367238,0.1398894191,0.0620218031,0.1233609915,-0.1146316975,-0.020297626,-0.0600939095,0.7015287876,0.2194091976,0.143311128,0.245865941,-0.1431449801,0.0402624793,-0.2635708153,-0.0018192854,0.1594168693,0.1043254882,0.1012833714,0.4308771193,-0.0101439431,-0.3055635095,-0.3132544458,0.0514972545,-0.274436444,-0.3873589039,-0.0785480589,0.3216457665,0.2337643057,-0.0422740318,0.0207918212,-0.147366792,-0.1888929904,0.263607502,0.048506245,-0.0017894536,0.1032019556,-0.1735251993,-0.3677696586,-0.3169654012,0.0823120326,0.1721159816,0.2570376098,-0.0665053874,-0.1105390191,0.2313485593,-0.2317669243,0.5387541652,-0.2662319839,0.0727584288,0.0955241323,0.0072519239,-0.538071692,-0.2452507466,-0.2579601705,-0.2703019679,-0.0064096581,0.0428120755,-0.8047869802,-0.1370545924,0.0025426375,0.0652824417,0.0224762987,0.0021824623,-0.3476671278,-0.3513223827,-0.2089853436,-0.3042515516,0.0158094689,0.162087664,-0.2289929092,0.0841265619,-0.1313247085,-0.0271121021,0.0254173987,0.467074275,0.0367717594,-0.1930948496,0.0135715557,0.2992069423,-0.3078332543,0.2499613464,0.3812302649,0.3898441494,-0.3924266398,0.0751300156,0.1768058985,0.2352270335,0.2457041591,-0.1128676981,0.1488084346,-0.2493561059,0.1089376807,-0.2314786166,-0.0435917862,0.6460393667,-0.1837123781,-0.4247295856,0.1305992752,0.46757406,0.1244366914,-0.0668817908,-0.0846195221,0.5950852633,-0.3703900576,0.0306381602,0.0692139044,0.8994724751,-0.0803756565,0.2363677025,0.6777761579,-0.1762773544,0.4314590991,-0.131967023,0.0896903872,-0.5020123124,0.1679986119,-0.1159755737,-0.0205783378,0.3543125391,-0.0853118822,-0.4229280651,0.1149851009,0.0423141941,0.4126974046,0.2536417842,0.3395940661,0.0222538412,-0.0871961489,-0.0327684283,0.1034799293,-0.146859169,0.32390064,-0.1414451599,-0.206780225,-0.1191481799,-0.392141521,-0.4757883251,-0.0012190758,-0.5666529536,0.2706696093,0.1667610407,-0.1066929474,0.4281308055,-0.0930189192,0.0278596915,-0.024059739,-0.5778464675,0.2720609903,-0.2703409493,-0.3054293394,0.0382399,0.0015378628,0.1709767431,-0.2011300325,-0.4782135189,0.1206074506,-0.1202965379,-0.1980054826,-0.2413238138,-0.2334801108,-0.1403298825,-0.165648669,0.204869926,-0.088970013,0.1008691266,-0.164320603,0.0706041604,-0.0845047086,-0.1225200742,0.185841307,0.1371440589,0.0163509175,-0.0627229586,0.3350047171,-0.1525661051,0.0187820718,0.4783485234,0.1441951245,-0.1743831038,-0.3062952757,-0.1273806095,0.1637389362,-0.194612354,-0.0876766294,0.4245507717,-0.0416405462,0.0999379456,0.6526227593,0.1358080506,-0.1568565071,0.029003216,-0.2108036727,-0.0722336918,0.4104577005,0.0001245645,0.3364676237,-0.1000608876,0.3942663074,0.1756132692,0.0507485792,-0.2581934929,0.1766735017,0.1273927838,0.0723767504,-0.0281810295,-0.1699816883,0.1859834939,0.1228221208,0.1235018596,-0.0236156229,-0.2789948285,-0.1525100917,-0.0532410853,0.1535165161,0.2072379142,-0.1573552191,0.1092613563,-0.0439141393,-0.0047672233,-0.0857668743,0.3785856068,0.0532588884,-0.1712069511,0.1584713906,0.1245287359,0.0749544948,-0.2570697963,-0.0818345845,-0.1065371409,0.0422274545,0.1521182656,0.1105312109,-0.2425280064,-0.06965819,0.2371832132,0.502761066,-0.0046943277,0.0697078258,0.2416599393,-0.3728254735,-0.1401909143,-0.1293221116,0.2363274544,0.4021983147,-0.0301402416,-0.0339136198,0.2486865073,0.1598043293,-0.1791618317,0.0249481164,0.7095746994,0.0255396459,0.144178614,0.1672735214,0.1907720119,-0.2652100325,-0.2495258749,0.0724356249,0.2639098763,0.105128631,0.0782728642,0.4552574158,-0.0888005421,-0.0570246428,0.1600898951,0.0752012879,0.1843453944,0.2008951455,-0.3341520429,0.0326524079,-0.2809025943,0.279322058,0.3199020028,0.1067882925,-0.2530281842,-0.0313566737,0.0981715322,-0.1619111001,0.0430844277,0.4668079913,0.1513280123,0.063598156,-0.4788576365,0.1876887828,-0.1072025374,-0.11174348,-0.046061568,-0.2203003466,-0.0647186488,-0.1144329384,-0.1065220833,-0.3939550817,0.4086517096,0.124942638,-0.0485959612,-0.4404159188,0.193289727,0.112508595,0.3823840916,0.0976767913,0.1988639235,0.2222632766,-0.3799968362,-0.1416541487,0.5052872896,0.4059638381,0.2238578498,-0.5180214047,-0.0117542129,-0.2111394703,-0.0981388465,-0.1031869277,0.2561290562,0.0835584179,0.1160430685,0.0260429978,0.0627270341,-0.0898388922,-0.4097585976,-0.1104363427,0.0061263693,-0.3029188216,0.0790456533,-0.2354569882,-0.0238410328,0.0639061853,-0.0892145634,-0.1156930402,-0.0091181565,0.0319977254,0.1436076015,0.0451708697,-0.1636793613,0.0443829075,0.0875869766,0.5330723524,0.2499726117,0.2016882747,-0.1850212365,-0.1509507596,-0.2981627285,0.1833940595,-0.300362438,-0.0558608621,0.0372362919,0.2560040653,-0.1863509864,0.3407575786,0.1460999697,0.2600632012,0.0267326254,0.2587114573,-0.1819263548,0.1599327177,-0.1776016206,-0.184375152,-0.0919502899,-0.3208380342,0.2229168862,-0.3668428957,0.0208827686,-0.1403071582,0.4087250829,0.0492090695,-0.4200240672,0.4051322341,0.4248364568,0.4987818301,0.0859793723,0.0233909953,-0.462423414,-0.2273351699,-0.3084006608,0.0380230956,-0.1979131252,0.4011147022,0.0511554442,-0.2752503455,-0.0668677092,0.206860885,0.1771453172,0.1820498556,-0.07849022,0.1207780764,-0.0507498011,0.1081958264,-0.218326211,0.0959962085,-0.0026972804,-0.1745442748,-0.3816213608,-0.6334139109,0.3339706361,-0.6409389377,-0.2875405252,-0.1929623932,0.1954268962,0.042068284,-0.2452954203,-0.4759141505,0.1125751436,0.158261776,0.1307853162,-0.3932452798,0.2494207919,0.1017882526,-0.1549134254,-0.0295701623,0.3505800664,-0.0256106276,0.0519909374,-0.0929412916,-0.1061921269]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/705","title":"TypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'","comments":"Hi !\r\nThanks for reporting :) \r\nIndeed this is an issue on the `datasets` side.\r\nI'm creating a PR","body":"## Environment info\r\n<!-- You can run the command `transformers-cli env` and copy-and-paste its output below.\r\n Don't forget to fill out the missing fields in that output! -->\r\n \r\n- `transformers` version: 3.3.1 (installed from master)\r\n- `datasets` version: 1.0.2 (installed as a dependency from transformers)\r\n- Platform: Linux-4.15.0-118-generic-x86_64-with-debian-stretch-sid\r\n- Python version: 3.7.9\r\n\r\nI'm testing my own text classification dataset using [this example](https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow) from transformers. The dataset is split into train \/ dev \/ test, and in csv format, containing just a text and a label columns, using comma as sep. Here's a sample:\r\n```\r\ntext,label\r\n\"Registra-se a presen\u00e7a do acad\u00eamico <name> . <REL_SEP> Ao me deparar com a descri\u00e7\u00e3o de dois autores no polo ativo da a\u00e7\u00e3o junto ao PJe , margem esquerda foi informado pela procuradora do reclamante que se trata de uma reclama\u00e7\u00e3o trabalhista individual . <REL_SEP> Diante disso , face a aus\u00eancia injustificada do autor <name> , determina-se o ARQUIVAMENTO do presente processo , com rela\u00e7\u00e3o a este , nos termos do [[ art . 844 da CLT ]] . <REL_SEP> CUSTAS AUTOR - DISPENSADO <REL_SEP> Custas pelo autor no importe de R $326,82 , calculadas sobre R $16.341,03 , dispensadas na forma da lei , em virtude da concess\u00e3o dos benef\u00edcios da Justi\u00e7a Gratuita , ora deferida . <REL_SEP> Cientes os presentes . <REL_SEP> Audi\u00eancia encerrada \u00e0s 8h42min . <REL_SEP> <name> <REL_SEP> Ju\u00edza do Trabalho <REL_SEP> Ata redigida por << <name> >> , Secret\u00e1rio de Audi\u00eancia .\",NO_RELATION\r\n```\r\n\r\nHowever, @Santosh-Gupta reported in #7351 that he had the exact same problem using the ChemProt dataset. His colab notebook is referenced in the following section.\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behavior:\r\n\r\n1. Created a new conda environment using conda env -n transformers python=3.7\r\n2. Cloned transformers master, `cd` into it and installed using pip install --editable . -r examples\/requirements.txt \r\n3. Installed tensorflow with `pip install tensorflow`\r\n3. Ran `run_tf_text_classification.py` with the following parameters:\r\n\r\n```\r\n--train_file <DATASET_PATH>\/train.csv \\\r\n--dev_file <DATASET_PATH>\/dev.csv \\ \r\n--test_file <DATASET_PATH>\/test.csv \\\r\n--label_column_id 1 \\\r\n--model_name_or_path neuralmind\/bert-base-portuguese-cased \\\r\n--output_dir <OUTPUT_PATH> \\\r\n--num_train_epochs 4 \\\r\n--per_device_train_batch_size 4 \\\r\n--per_device_eval_batch_size 4 \\\r\n--do_train \\\r\n--do_eval \\\r\n--do_predict \\\r\n--logging_steps 1000 \\\r\n--evaluate_during_training \\\r\n--save_steps 1000 \\\r\n--overwrite_output_dir \\\r\n--overwrite_cache\r\n```\r\n\r\nI have also copied [@Santosh-Gupta 's colab notebook](https:\/\/colab.research.google.com\/drive\/11APei6GjphCZbH5wD9yVlfGvpIkh8pwr?usp=sharing) as a reference.\r\n\r\n<!-- If you have code snippets, error messages, stack traces please provide them here as well.\r\n Important! Use code tags to correctly format your code. See https:\/\/help.github.com\/en\/github\/writing-on-github\/creating-and-highlighting-code-blocks#syntax-highlighting\r\n Do not use screenshots, as they are hard to read and (more importantly) don't allow others to copy-and-paste your code.-->\r\n\r\nHere is the stack trace:\r\n\r\n```\r\n2020-10-02 07:33:41.622011: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n\/media\/discoD\/repositorios\/transformers_pedro\/src\/transformers\/training_args.py:333: FutureWarning: The `evaluate_during_training` argument is deprecated in favor of `evaluation_strategy` (which has more options)\r\n FutureWarning,\r\n2020-10-02 07:33:43.471648: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcuda.so.1\r\n2020-10-02 07:33:43.471791: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.472664: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1716] Found device 0 with properties: \r\npciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1\r\ncoreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB\/s\r\n2020-10-02 07:33:43.472684: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n2020-10-02 07:33:43.472765: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10\r\n2020-10-02 07:33:43.472809: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10\r\n2020-10-02 07:33:43.472848: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10\r\n2020-10-02 07:33:43.474209: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10\r\n2020-10-02 07:33:43.474276: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10\r\n2020-10-02 07:33:43.561219: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7\r\n2020-10-02 07:33:43.561397: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.562345: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.563219: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1858] Adding visible gpu devices: 0\r\n2020-10-02 07:33:43.563595: I tensorflow\/core\/platform\/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations: AVX2 FMA\r\nTo enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.\r\n2020-10-02 07:33:43.570091: I tensorflow\/core\/platform\/profile_utils\/cpu_utils.cc:104] CPU Frequency: 3591830000 Hz\r\n2020-10-02 07:33:43.570494: I tensorflow\/compiler\/xla\/service\/service.cc:168] XLA service 0x560842432400 initialized for platform Host (this does not guarantee that XLA will be used). Devices:\r\n2020-10-02 07:33:43.570511: I tensorflow\/compiler\/xla\/service\/service.cc:176] StreamExecutor device (0): Host, Default Version\r\n2020-10-02 07:33:43.570702: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.571599: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1716] Found device 0 with properties: \r\npciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1\r\ncoreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB\/s\r\n2020-10-02 07:33:43.571633: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n2020-10-02 07:33:43.571645: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10\r\n2020-10-02 07:33:43.571654: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10\r\n2020-10-02 07:33:43.571664: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10\r\n2020-10-02 07:33:43.571691: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10\r\n2020-10-02 07:33:43.571704: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10\r\n2020-10-02 07:33:43.571718: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7\r\n2020-10-02 07:33:43.571770: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.572641: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.573475: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1858] Adding visible gpu devices: 0\r\n2020-10-02 07:33:47.139227: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1257] Device interconnect StreamExecutor with strength 1 edge matrix:\r\n2020-10-02 07:33:47.139265: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1263] 0 \r\n2020-10-02 07:33:47.139272: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1276] 0: N \r\n2020-10-02 07:33:47.140323: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.141248: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.142085: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.142854: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1402] Created TensorFlow device (\/job:localhost\/replica:0\/task:0\/device:GPU:0 with 5371 MB memory) -> physical GPU (device: 0, name: GeForce GTX 1070, pci bus id: 0000:01:00.0, compute capability: 6.1)\r\n2020-10-02 07:33:47.146317: I tensorflow\/compiler\/xla\/service\/service.cc:168] XLA service 0x5608b95dc5c0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:\r\n2020-10-02 07:33:47.146336: I tensorflow\/compiler\/xla\/service\/service.cc:176] StreamExecutor device (0): GeForce GTX 1070, Compute Capability 6.1\r\n10\/02\/2020 07:33:47 - INFO - __main__ - n_replicas: 1, distributed training: False, 16-bits training: False\r\n10\/02\/2020 07:33:47 - INFO - __main__ - Training\/evaluation parameters TFTrainingArguments(output_dir='\/media\/discoD\/models\/datalawyer\/pedidos\/transformers_tf', overwrite_output_dir=True, do_train=True, do_eval=True, do_predict=True, evaluate_during_training=True, evaluation_strategy=<EvaluationStrategy.STEPS: 'steps'>, prediction_loss_only=False, per_device_train_batch_size=4, per_device_eval_batch_size=4, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=1, learning_rate=5e-05, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=4.0, max_steps=-1, warmup_steps=0, logging_dir='runs\/Oct02_07-33-43_user-XPS-8700', logging_first_step=False, logging_steps=1000, save_steps=1000, save_total_limit=None, no_cuda=False, seed=42, fp16=False, fp16_opt_level='O1', local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=1000, dataloader_num_workers=0, past_index=-1, run_name='\/media\/discoD\/models\/datalawyer\/pedidos\/transformers_tf', disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=False, tpu_name=None, xla=False)\r\n10\/02\/2020 07:33:53 - INFO - filelock - Lock 140407857405776 acquired on \/home\/user\/.cache\/huggingface\/datasets\/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock\r\n10\/02\/2020 07:33:53 - INFO - filelock - Lock 140407857405776 released on \/home\/user\/.cache\/huggingface\/datasets\/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"run_tf_text_classification.py\", line 283, in <module>\r\n main()\r\n File \"run_tf_text_classification.py\", line 222, in main\r\n max_seq_length=data_args.max_seq_length,\r\n File \"run_tf_text_classification.py\", line 43, in get_tfds\r\n ds = datasets.load_dataset(\"csv\", data_files=files)\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 604, in load_dataset\r\n **config_kwargs,\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 158, in __init__\r\n **config_kwargs,\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 269, in _create_builder_config\r\n for key in sorted(data_files.keys()):\r\nTypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'\r\n```\r\n\r\n## Expected behavior\r\n\r\nShould be able to run the text-classification example as described in [https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow](https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow)\r\n\r\nOriginally opened this issue at transformers' repository: [https:\/\/github.com\/huggingface\/transformers\/issues\/7535](https:\/\/github.com\/huggingface\/transformers\/issues\/7535). @jplu instructed me to open here, since according to [this](https:\/\/github.com\/huggingface\/transformers\/issues\/7535#issuecomment-702778885) evidence, the problem is from datasets.\r\n\r\nThanks!","comment_length":19,"text":"TypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'\n## Environment info\r\n<!-- You can run the command `transformers-cli env` and copy-and-paste its output below.\r\n Don't forget to fill out the missing fields in that output! -->\r\n \r\n- `transformers` version: 3.3.1 (installed from master)\r\n- `datasets` version: 1.0.2 (installed as a dependency from transformers)\r\n- Platform: Linux-4.15.0-118-generic-x86_64-with-debian-stretch-sid\r\n- Python version: 3.7.9\r\n\r\nI'm testing my own text classification dataset using [this example](https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow) from transformers. The dataset is split into train \/ dev \/ test, and in csv format, containing just a text and a label columns, using comma as sep. Here's a sample:\r\n```\r\ntext,label\r\n\"Registra-se a presen\u00e7a do acad\u00eamico <name> . <REL_SEP> Ao me deparar com a descri\u00e7\u00e3o de dois autores no polo ativo da a\u00e7\u00e3o junto ao PJe , margem esquerda foi informado pela procuradora do reclamante que se trata de uma reclama\u00e7\u00e3o trabalhista individual . <REL_SEP> Diante disso , face a aus\u00eancia injustificada do autor <name> , determina-se o ARQUIVAMENTO do presente processo , com rela\u00e7\u00e3o a este , nos termos do [[ art . 844 da CLT ]] . <REL_SEP> CUSTAS AUTOR - DISPENSADO <REL_SEP> Custas pelo autor no importe de R $326,82 , calculadas sobre R $16.341,03 , dispensadas na forma da lei , em virtude da concess\u00e3o dos benef\u00edcios da Justi\u00e7a Gratuita , ora deferida . <REL_SEP> Cientes os presentes . <REL_SEP> Audi\u00eancia encerrada \u00e0s 8h42min . <REL_SEP> <name> <REL_SEP> Ju\u00edza do Trabalho <REL_SEP> Ata redigida por << <name> >> , Secret\u00e1rio de Audi\u00eancia .\",NO_RELATION\r\n```\r\n\r\nHowever, @Santosh-Gupta reported in #7351 that he had the exact same problem using the ChemProt dataset. His colab notebook is referenced in the following section.\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behavior:\r\n\r\n1. Created a new conda environment using conda env -n transformers python=3.7\r\n2. Cloned transformers master, `cd` into it and installed using pip install --editable . -r examples\/requirements.txt \r\n3. Installed tensorflow with `pip install tensorflow`\r\n3. Ran `run_tf_text_classification.py` with the following parameters:\r\n\r\n```\r\n--train_file <DATASET_PATH>\/train.csv \\\r\n--dev_file <DATASET_PATH>\/dev.csv \\ \r\n--test_file <DATASET_PATH>\/test.csv \\\r\n--label_column_id 1 \\\r\n--model_name_or_path neuralmind\/bert-base-portuguese-cased \\\r\n--output_dir <OUTPUT_PATH> \\\r\n--num_train_epochs 4 \\\r\n--per_device_train_batch_size 4 \\\r\n--per_device_eval_batch_size 4 \\\r\n--do_train \\\r\n--do_eval \\\r\n--do_predict \\\r\n--logging_steps 1000 \\\r\n--evaluate_during_training \\\r\n--save_steps 1000 \\\r\n--overwrite_output_dir \\\r\n--overwrite_cache\r\n```\r\n\r\nI have also copied [@Santosh-Gupta 's colab notebook](https:\/\/colab.research.google.com\/drive\/11APei6GjphCZbH5wD9yVlfGvpIkh8pwr?usp=sharing) as a reference.\r\n\r\n<!-- If you have code snippets, error messages, stack traces please provide them here as well.\r\n Important! Use code tags to correctly format your code. See https:\/\/help.github.com\/en\/github\/writing-on-github\/creating-and-highlighting-code-blocks#syntax-highlighting\r\n Do not use screenshots, as they are hard to read and (more importantly) don't allow others to copy-and-paste your code.-->\r\n\r\nHere is the stack trace:\r\n\r\n```\r\n2020-10-02 07:33:41.622011: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n\/media\/discoD\/repositorios\/transformers_pedro\/src\/transformers\/training_args.py:333: FutureWarning: The `evaluate_during_training` argument is deprecated in favor of `evaluation_strategy` (which has more options)\r\n FutureWarning,\r\n2020-10-02 07:33:43.471648: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcuda.so.1\r\n2020-10-02 07:33:43.471791: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.472664: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1716] Found device 0 with properties: \r\npciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1\r\ncoreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB\/s\r\n2020-10-02 07:33:43.472684: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n2020-10-02 07:33:43.472765: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10\r\n2020-10-02 07:33:43.472809: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10\r\n2020-10-02 07:33:43.472848: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10\r\n2020-10-02 07:33:43.474209: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10\r\n2020-10-02 07:33:43.474276: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10\r\n2020-10-02 07:33:43.561219: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7\r\n2020-10-02 07:33:43.561397: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.562345: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.563219: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1858] Adding visible gpu devices: 0\r\n2020-10-02 07:33:43.563595: I tensorflow\/core\/platform\/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations: AVX2 FMA\r\nTo enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.\r\n2020-10-02 07:33:43.570091: I tensorflow\/core\/platform\/profile_utils\/cpu_utils.cc:104] CPU Frequency: 3591830000 Hz\r\n2020-10-02 07:33:43.570494: I tensorflow\/compiler\/xla\/service\/service.cc:168] XLA service 0x560842432400 initialized for platform Host (this does not guarantee that XLA will be used). Devices:\r\n2020-10-02 07:33:43.570511: I tensorflow\/compiler\/xla\/service\/service.cc:176] StreamExecutor device (0): Host, Default Version\r\n2020-10-02 07:33:43.570702: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.571599: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1716] Found device 0 with properties: \r\npciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1\r\ncoreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB\/s\r\n2020-10-02 07:33:43.571633: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n2020-10-02 07:33:43.571645: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10\r\n2020-10-02 07:33:43.571654: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10\r\n2020-10-02 07:33:43.571664: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10\r\n2020-10-02 07:33:43.571691: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10\r\n2020-10-02 07:33:43.571704: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10\r\n2020-10-02 07:33:43.571718: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7\r\n2020-10-02 07:33:43.571770: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.572641: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.573475: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1858] Adding visible gpu devices: 0\r\n2020-10-02 07:33:47.139227: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1257] Device interconnect StreamExecutor with strength 1 edge matrix:\r\n2020-10-02 07:33:47.139265: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1263] 0 \r\n2020-10-02 07:33:47.139272: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1276] 0: N \r\n2020-10-02 07:33:47.140323: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.141248: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.142085: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.142854: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1402] Created TensorFlow device (\/job:localhost\/replica:0\/task:0\/device:GPU:0 with 5371 MB memory) -> physical GPU (device: 0, name: GeForce GTX 1070, pci bus id: 0000:01:00.0, compute capability: 6.1)\r\n2020-10-02 07:33:47.146317: I tensorflow\/compiler\/xla\/service\/service.cc:168] XLA service 0x5608b95dc5c0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:\r\n2020-10-02 07:33:47.146336: I tensorflow\/compiler\/xla\/service\/service.cc:176] StreamExecutor device (0): GeForce GTX 1070, Compute Capability 6.1\r\n10\/02\/2020 07:33:47 - INFO - __main__ - n_replicas: 1, distributed training: False, 16-bits training: False\r\n10\/02\/2020 07:33:47 - INFO - __main__ - Training\/evaluation parameters TFTrainingArguments(output_dir='\/media\/discoD\/models\/datalawyer\/pedidos\/transformers_tf', overwrite_output_dir=True, do_train=True, do_eval=True, do_predict=True, evaluate_during_training=True, evaluation_strategy=<EvaluationStrategy.STEPS: 'steps'>, prediction_loss_only=False, per_device_train_batch_size=4, per_device_eval_batch_size=4, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=1, learning_rate=5e-05, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=4.0, max_steps=-1, warmup_steps=0, logging_dir='runs\/Oct02_07-33-43_user-XPS-8700', logging_first_step=False, logging_steps=1000, save_steps=1000, save_total_limit=None, no_cuda=False, seed=42, fp16=False, fp16_opt_level='O1', local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=1000, dataloader_num_workers=0, past_index=-1, run_name='\/media\/discoD\/models\/datalawyer\/pedidos\/transformers_tf', disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=False, tpu_name=None, xla=False)\r\n10\/02\/2020 07:33:53 - INFO - filelock - Lock 140407857405776 acquired on \/home\/user\/.cache\/huggingface\/datasets\/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock\r\n10\/02\/2020 07:33:53 - INFO - filelock - Lock 140407857405776 released on \/home\/user\/.cache\/huggingface\/datasets\/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"run_tf_text_classification.py\", line 283, in <module>\r\n main()\r\n File \"run_tf_text_classification.py\", line 222, in main\r\n max_seq_length=data_args.max_seq_length,\r\n File \"run_tf_text_classification.py\", line 43, in get_tfds\r\n ds = datasets.load_dataset(\"csv\", data_files=files)\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 604, in load_dataset\r\n **config_kwargs,\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 158, in __init__\r\n **config_kwargs,\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 269, in _create_builder_config\r\n for key in sorted(data_files.keys()):\r\nTypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'\r\n```\r\n\r\n## Expected behavior\r\n\r\nShould be able to run the text-classification example as described in [https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow](https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow)\r\n\r\nOriginally opened this issue at transformers' repository: [https:\/\/github.com\/huggingface\/transformers\/issues\/7535](https:\/\/github.com\/huggingface\/transformers\/issues\/7535). @jplu instructed me to open here, since according to [this](https:\/\/github.com\/huggingface\/transformers\/issues\/7535#issuecomment-702778885) evidence, the problem is from datasets.\r\n\r\nThanks!\nHi !\r\nThanks for reporting :) \r\nIndeed this is an issue on the `datasets` side.\r\nI'm creating a PR","embeddings":[-0.2919273078,-0.6131456494,-0.0538775101,0.0724286065,0.5358543992,0.0512854382,0.5790147781,0.301427871,0.288526386,0.1793083698,-0.0349086858,0.1695188433,-0.0423114263,-0.1191879734,-0.1670343578,-0.2730587125,-0.1069535911,0.1412786692,-0.4372419417,-0.0667978898,0.0061711208,0.1115164608,-0.3358640075,0.0874110833,-0.3546169102,-0.1470186561,0.1937073618,0.0334143899,0.1056338474,-0.3845270872,0.366317004,-0.0270779245,0.1710627675,0.491507113,-0.0001042778,0.081310913,0.4950879514,-0.138119936,-0.0701588094,-0.343044579,-0.4419636428,0.0307981856,0.0160502605,-0.3915827572,-0.2448375076,-0.0110362982,-0.1189130321,-0.0461345613,0.6161757112,0.278801918,0.2397651076,0.2381774187,-0.1034501866,-0.3372361064,-0.1007003114,0.3406432867,-0.1360172033,-0.1092183441,0.1256229579,-0.2894225121,-0.0251126885,0.2308775187,0.0032396459,-0.089165844,0.3362298608,0.126687631,-0.1719463766,0.0904838964,0.01274959,0.2484526038,0.3977201581,-0.191258952,-0.201420784,-0.1912941635,-0.4447488189,-0.3844889998,0.0586855598,0.0405776054,0.0833992809,0.2616128325,-0.0061038891,0.4328720272,-0.2141648531,-0.1010648534,0.0205578785,0.5832354426,-0.1706664115,0.0971066877,0.0810640901,-0.0337578095,0.1662235856,-0.1250212342,0.2540244162,0.0342846215,-0.2468905896,-0.012379894,-0.0421408713,-0.4438541234,-0.214487657,-0.1216932535,0.2730001509,-0.0499446429,0.0718224198,0.2485999912,0.1498123109,0.2958030701,0.0911678746,0.537527442,0.2819210589,0.3849364817,-0.0427861772,-0.1365961283,-0.142474398,-0.158597067,-0.0144823557,0.4171228111,0.0576643199,-0.1039039567,-0.30254969,-0.2212817669,0.0309037361,-0.0660278946,-0.0246453453,0.4157520235,-0.0416241586,0.3547725976,0.2065986395,0.0611954629,-0.0000107179,-0.1164428368,-0.1507170498,0.1743426323,-0.3866131902,-0.2293417454,-0.0172503423,0.2606917918,0.0060736635,-0.1308667958,-0.083957687,-0.0662648901,-0.0675454959,0.0104777887,0.038608443,0.2141883671,-0.1814568043,0.1716786176,0.1413978785,-0.142892912,-0.0168804023,0.2545795739,-0.0376725495,-0.2689933479,0.0494651496,0.1983073056,0.0924026594,0.0822201595,0.0646613985,0.1997675002,0.2888107002,-0.0023917381,0.0389889143,0.0640591905,-0.1613370031,0.1434037834,0.0981706604,0.3937546015,-0.7670513988,-0.2362775803,-0.0231548995,0.0841976553,-0.0441100746,0.4377836287,-0.3081965744,0.2548838556,0.1809004545,0.2804787159,0.1315836161,-0.4054885507,-0.2366541028,-0.0511300489,0.0623557344,-0.1371805072,0.1976591647,-0.1599786133,-0.0393890887,-0.2148528397,-0.2731992602,0.1258901358,-0.0372625664,0.0372501723,-0.3978560269,-0.0586486682,0.1387393624,0.0166069958,0.0908488706,-0.0245048665,-0.0067748767,0.5229411125,0.1247486845,-0.3379059136,0.1744470745,0.2543297112,0.0626068562,0.2155818939,0.018898882,-0.2434887141,-0.2023972422,0.1049191728,-0.0260952972,0.3419173658,-0.2349476367,-0.1734545231,-0.4128475785,0.0400041379,0.0303186756,-0.1581599265,0.1817014217,0.0163617879,0.0562101454,-0.1548790336,0.0504410565,0.0850315765,-0.0169340409,0.1152144745,-0.3652016521,0.205051899,-0.2516756356,-0.0821671262,-0.126055941,0.208872363,0.2595393658,-0.2008835673,-0.1568623483,0.3399887383,-0.2882199287,0.1538384408,-0.3089927435,-0.0751828775,0.2699451447,-0.4763702154,0.0024343291,0.1310064346,0.0902314931,-0.0006721493,-0.4511058927,0.1370696127,0.0112843513,0.1648834497,0.1284329742,0.0007869658,0.1388457268,-0.1832198799,-0.0573632009,-0.1072910875,-0.0001398468,0.1755349934,-0.2161804736,0.0637440309,-0.1466740072,-0.1216330752,0.4897304475,-0.0830696449,0.2121409774,-0.0066706208,-0.2014056146,0.1740379184,0.0147640789,0.2376895696,0.1019209698,0.1063609719,-0.1818083525,0.1594488174,-0.1297113597,-0.1348221153,0.1366052777,0.1401495934,0.1842842102,0.047402259,-0.0023738095,0.0637102649,0.0001989906,-0.3237716556,0.0035916145,0.0375333168,-0.2615197003,0.1092323586,-0.2121478766,-0.1581652462,-0.2769756615,0.241360575,0.1923743635,-0.0711507946,-0.0779944956,0.0264965203,-0.2088043094,0.3075107038,0.1213724911,0.2280253023,0.2589328587,-0.1669288725,-0.1522791386,-0.3805861771,-0.1148757711,0.1299654692,0.1442116201,0.1163031459,-0.0103829578,0.0811195746,0.0132961869,0.4569865167,-0.5159244537,0.0750539526,-0.314931035,-0.2156138271,-0.0008078428,-0.0266429707,0.0513779521,-0.1849368513,0.3361887038,-0.1817682087,0.0587969385,-0.0685050115,0.1174616069,-0.2828928232,-0.1832380444,-0.6926159859,-0.2156877816,-0.4069219828,0.2902898192,-0.1129137129,-0.1000652313,0.4519687295,-0.04145924,-0.1367301047,0.2323179543,-0.0375305265,0.0842413232,0.2739644647,0.3995569944,-0.1479191929,-0.3361201286,-0.037801262,-0.1635225713,0.0331597701,-0.2620622218,-0.1347199827,-0.0583056845,-0.2712892592,-0.1474146694,0.1512570232,-0.1404368728,0.6117979288,0.1744675785,-0.1556137651,-0.1580144018,-0.4322287738,0.2149780095,0.0620988421,0.0518928878,-0.0564923882,0.5008730292,0.1867362559,0.6495117545,0.0560511127,-0.1962005943,-0.0095487041,0.1530277133,0.0647748932,-0.0284662768,-0.2616336942,0.0940554515,-0.1156798303,-0.1356900036,0.1677749753,-0.1896721721,-0.1575728357,-0.1722387522,0.1542970687,-0.0772994012,-0.0894062594,0.1257669032,0.1191693991,0.2297199517,0.1420507431,-0.0787234679,-0.10656555,-0.1022997648,0.096433714,-0.0871651992,-0.1488742679,-0.0652406886,-0.3501496017,0.2422767282,-0.1469796896,0.2825861275,0.2371597588,0.1074166596,-0.2071240842,-0.1536873728,-0.2497842014,0.014652282,0.1408586353,-0.3110143244,0.1866850853,0.1122649312,-0.0809816793,-0.1901675463,-0.2099327296,-0.3738399744,-0.298717618,0.453838259,0.2226490676,-0.185783416,-0.1639241427,0.2885839641,-0.1199201718,-0.0421009362,-0.066820167,-0.0768557042,0.0182460137,-0.2515257597,-0.0107603529,0.5364021659,0.2025124282,-0.2095840871,-0.0444211029,0.0943264514,-0.2034685016,0.0882901251,0.2573613226,0.1361691803,-0.1558400393,-0.0760446042,0.1248814166,0.2837224007,-0.3310799897,0.5674952865,-0.0255851038,-0.4515457153,0.0420375802,-0.0200555008,0.1364636868,0.1572538167,-0.0916863009,-0.0461933538,0.1043341309,0.0756744668,-0.3240720034,0.1180027127,0.1891417652,0.3833045363,-0.4595752358,-0.2022822946,0.1838428825,0.0292732008,-0.0668438151,-0.0761145428,-0.1605715901,-0.4441316724,0.3648698926,0.0461951494,0.8255496025,-0.1720346361,-0.0274047237,-0.2060128003,0.1214447916,0.5401180983,-0.4126571417,0.2696087658,-0.2367256135,-0.3283510208,-0.1095912084,-0.0516983978,0.3606872559,0.4803798199,-0.3971676826,0.2573061287,0.103219524,0.5092794299,0.0648841634,0.136406064,0.4459854364,-0.3245070875,-0.2694648802,0.1508009285,0.0748465508,0.0509600416,0.0084982747,-0.2146799117,-0.2909614444,-0.1904337853,-0.318582505,0.3504792452,-0.2043990791,0.0957831964,0.2979163826,0.0655586198,0.4973452091,0.3372198641,0.3578670621,0.2304675281,-0.2153190523,0.2350308448,-0.4642540514,-0.0610615797,0.0088658445,0.251606375,0.227899462,0.0168934893,-0.1576444358,0.1514526308,-0.0659205243,-0.239291802,-0.0809610933,0.0206167698,0.2474818826,-0.5286822319,0.0103748515,0.1242398694,-0.263369292,-0.2806583047,0.2036949992,-0.3791154921,-0.1899057478,0.1543290168,0.1162807718,-0.2372897416,0.0348868333,0.3466538191,0.007645539,0.2065514922,0.2748655975,0.1189157292,-0.1234726608,-0.2976979017,-0.1585630625,0.2176708281,-0.5927738547,0.0982700586,-0.2696886063,-0.4663753808,0.121622853,0.4874913394,0.3218313754,0.1377170533,-0.1893444061,-0.4025084376,-0.2644128799,0.0103745395,0.1245822832,0.2435663193,0.1476768851,0.6597263813,0.0173819643,0.0904994458,-0.3986144662,-0.2800250053,-0.1478807628,0.295852989,0.3108174205,-0.3957287371,0.3502090871,-0.0701561272,0.1842743754,0.4157313108,-0.2173074782,-0.2901489437,-0.0977177992,0.0736077875,0.1292785853,-0.2479224503,-0.2928442061,0.0705482587,0.0738156065,-0.3057259917,0.060341455,0.0323538966,-0.0002617335,0.3716009855,-0.2391617298,0.1032218561,-0.1942325234,-0.1802102327,-0.1308957338,0.2095948011,0.0563998781,0.0817958862,0.2029102743,-0.0294597894,-0.4282607436,0.0897560492,0.2320590913,0.260022521,0.3225409389,-0.3676284552,-0.0552807488,0.1726189703,0.5626053214,0.0809191167,-0.1066219062,0.0329805873,-0.0473479591,0.2685658038,-0.3096589446,0.2288379371,-0.0930913463,0.067760922,0.2696631253,0.2911012471,0.2975101173,-0.0638441667,0.4036651254,0.2677652836,0.5184584856,-0.4800507724,0.3569121361,0.5745406747,0.1697940677,0.1794567853,0.3768188655,-0.0295521896,0.3553950787,0.5121091604,0.2054487318,0.1787361652,0.0035348476,0.1895809472,0.0774413496,-0.15311791,0.0910538435,0.085694313,-0.1319064498,0.0842484906,-0.2042768598,0.3030064106,-0.3348991275,-0.0643551946,-0.137806654,0.0549418107,-0.2147033364,-0.2030396312,0.1460205913,0.0916059837,-0.0967579708,-0.1491556168,0.1459876448,0.2719106674,0.3311924934,-0.2415809184,0.011369125,-0.186536476,-0.1872024983,-0.022918554,0.3771433234,-0.2421434671,0.3255573809,0.1919579655,-0.2189380825,-0.0256017949,0.3311004043,0.3879540563,0.346213311,-0.0973591655,-0.0188501198,-0.1468135267,-0.2428473681,0.1603179723,0.290438056,0.1453945041,0.1796865314,0.2034573555,0.1824687421,-0.2012612075,0.2441810966,-0.045337636,0.0992114991,0.0013043385,0.0638233572,0.0748212859,-0.0474862158,-0.0048640999,0.0407558642,-0.3722771108,-0.0838894472,0.364733696,-0.2776278853,0.4735959768,-0.0817337409,0.1044070572,0.0357249528,0.3190372884,-0.0509206988,-0.2098562717,-0.220578298,0.0366850682,-0.246158421,-0.0787129104,0.0871472582,-0.2164678574,-0.2482390255,0.0417579189,0.1564619392,0.0311662164,-0.1432110369,-0.315692544,0.010035662,-0.0800714567,-0.1900778562,-0.1123893782,0.0135460198,0.0193593539,0.2289910316,-0.2942387164,-0.0722259209,0.4261686206,0.0742083043,-0.4140973091,-0.0033793557,-0.0250940695,0.2368596941,0.1539556831,0.097318992,0.569098711,-0.2223181427,0.1948460042,-0.077134043,-0.3015601039,-0.1696515381,0.2426602244,0.3290059865,0.3649009764,-0.1513296813,0.1434555501,-0.2218807787,-0.1535740495,0.0497781672,-0.2138116062,-0.2299725264,-0.0495054238,-0.2060346454,-0.0614992492,0.1562587619,0.2371859103,0.0235220343,-0.0085253129,-0.181430459,-0.5764254928,0.4325292706,-0.3074938059,-0.4217503965,0.119653061,-0.0462134816,0.2443118989,0.0112836268,-0.4401860535,0.0656878874,0.2044553459,-0.0260742437,-0.2218872756,0.058609765,0.1953352988,-0.0228522941,-0.0969520509,0.2508562207,0.1856728643,-0.079318516,-0.0641221404,-0.1950844824]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/705","title":"TypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'","comments":"Thanks @lhoestq !","body":"## Environment info\r\n<!-- You can run the command `transformers-cli env` and copy-and-paste its output below.\r\n Don't forget to fill out the missing fields in that output! -->\r\n \r\n- `transformers` version: 3.3.1 (installed from master)\r\n- `datasets` version: 1.0.2 (installed as a dependency from transformers)\r\n- Platform: Linux-4.15.0-118-generic-x86_64-with-debian-stretch-sid\r\n- Python version: 3.7.9\r\n\r\nI'm testing my own text classification dataset using [this example](https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow) from transformers. The dataset is split into train \/ dev \/ test, and in csv format, containing just a text and a label columns, using comma as sep. Here's a sample:\r\n```\r\ntext,label\r\n\"Registra-se a presen\u00e7a do acad\u00eamico <name> . <REL_SEP> Ao me deparar com a descri\u00e7\u00e3o de dois autores no polo ativo da a\u00e7\u00e3o junto ao PJe , margem esquerda foi informado pela procuradora do reclamante que se trata de uma reclama\u00e7\u00e3o trabalhista individual . <REL_SEP> Diante disso , face a aus\u00eancia injustificada do autor <name> , determina-se o ARQUIVAMENTO do presente processo , com rela\u00e7\u00e3o a este , nos termos do [[ art . 844 da CLT ]] . <REL_SEP> CUSTAS AUTOR - DISPENSADO <REL_SEP> Custas pelo autor no importe de R $326,82 , calculadas sobre R $16.341,03 , dispensadas na forma da lei , em virtude da concess\u00e3o dos benef\u00edcios da Justi\u00e7a Gratuita , ora deferida . <REL_SEP> Cientes os presentes . <REL_SEP> Audi\u00eancia encerrada \u00e0s 8h42min . <REL_SEP> <name> <REL_SEP> Ju\u00edza do Trabalho <REL_SEP> Ata redigida por << <name> >> , Secret\u00e1rio de Audi\u00eancia .\",NO_RELATION\r\n```\r\n\r\nHowever, @Santosh-Gupta reported in #7351 that he had the exact same problem using the ChemProt dataset. His colab notebook is referenced in the following section.\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behavior:\r\n\r\n1. Created a new conda environment using conda env -n transformers python=3.7\r\n2. Cloned transformers master, `cd` into it and installed using pip install --editable . -r examples\/requirements.txt \r\n3. Installed tensorflow with `pip install tensorflow`\r\n3. Ran `run_tf_text_classification.py` with the following parameters:\r\n\r\n```\r\n--train_file <DATASET_PATH>\/train.csv \\\r\n--dev_file <DATASET_PATH>\/dev.csv \\ \r\n--test_file <DATASET_PATH>\/test.csv \\\r\n--label_column_id 1 \\\r\n--model_name_or_path neuralmind\/bert-base-portuguese-cased \\\r\n--output_dir <OUTPUT_PATH> \\\r\n--num_train_epochs 4 \\\r\n--per_device_train_batch_size 4 \\\r\n--per_device_eval_batch_size 4 \\\r\n--do_train \\\r\n--do_eval \\\r\n--do_predict \\\r\n--logging_steps 1000 \\\r\n--evaluate_during_training \\\r\n--save_steps 1000 \\\r\n--overwrite_output_dir \\\r\n--overwrite_cache\r\n```\r\n\r\nI have also copied [@Santosh-Gupta 's colab notebook](https:\/\/colab.research.google.com\/drive\/11APei6GjphCZbH5wD9yVlfGvpIkh8pwr?usp=sharing) as a reference.\r\n\r\n<!-- If you have code snippets, error messages, stack traces please provide them here as well.\r\n Important! Use code tags to correctly format your code. See https:\/\/help.github.com\/en\/github\/writing-on-github\/creating-and-highlighting-code-blocks#syntax-highlighting\r\n Do not use screenshots, as they are hard to read and (more importantly) don't allow others to copy-and-paste your code.-->\r\n\r\nHere is the stack trace:\r\n\r\n```\r\n2020-10-02 07:33:41.622011: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n\/media\/discoD\/repositorios\/transformers_pedro\/src\/transformers\/training_args.py:333: FutureWarning: The `evaluate_during_training` argument is deprecated in favor of `evaluation_strategy` (which has more options)\r\n FutureWarning,\r\n2020-10-02 07:33:43.471648: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcuda.so.1\r\n2020-10-02 07:33:43.471791: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.472664: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1716] Found device 0 with properties: \r\npciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1\r\ncoreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB\/s\r\n2020-10-02 07:33:43.472684: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n2020-10-02 07:33:43.472765: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10\r\n2020-10-02 07:33:43.472809: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10\r\n2020-10-02 07:33:43.472848: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10\r\n2020-10-02 07:33:43.474209: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10\r\n2020-10-02 07:33:43.474276: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10\r\n2020-10-02 07:33:43.561219: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7\r\n2020-10-02 07:33:43.561397: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.562345: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.563219: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1858] Adding visible gpu devices: 0\r\n2020-10-02 07:33:43.563595: I tensorflow\/core\/platform\/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations: AVX2 FMA\r\nTo enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.\r\n2020-10-02 07:33:43.570091: I tensorflow\/core\/platform\/profile_utils\/cpu_utils.cc:104] CPU Frequency: 3591830000 Hz\r\n2020-10-02 07:33:43.570494: I tensorflow\/compiler\/xla\/service\/service.cc:168] XLA service 0x560842432400 initialized for platform Host (this does not guarantee that XLA will be used). Devices:\r\n2020-10-02 07:33:43.570511: I tensorflow\/compiler\/xla\/service\/service.cc:176] StreamExecutor device (0): Host, Default Version\r\n2020-10-02 07:33:43.570702: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.571599: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1716] Found device 0 with properties: \r\npciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1\r\ncoreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB\/s\r\n2020-10-02 07:33:43.571633: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n2020-10-02 07:33:43.571645: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10\r\n2020-10-02 07:33:43.571654: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10\r\n2020-10-02 07:33:43.571664: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10\r\n2020-10-02 07:33:43.571691: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10\r\n2020-10-02 07:33:43.571704: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10\r\n2020-10-02 07:33:43.571718: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7\r\n2020-10-02 07:33:43.571770: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.572641: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.573475: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1858] Adding visible gpu devices: 0\r\n2020-10-02 07:33:47.139227: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1257] Device interconnect StreamExecutor with strength 1 edge matrix:\r\n2020-10-02 07:33:47.139265: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1263] 0 \r\n2020-10-02 07:33:47.139272: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1276] 0: N \r\n2020-10-02 07:33:47.140323: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.141248: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.142085: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.142854: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1402] Created TensorFlow device (\/job:localhost\/replica:0\/task:0\/device:GPU:0 with 5371 MB memory) -> physical GPU (device: 0, name: GeForce GTX 1070, pci bus id: 0000:01:00.0, compute capability: 6.1)\r\n2020-10-02 07:33:47.146317: I tensorflow\/compiler\/xla\/service\/service.cc:168] XLA service 0x5608b95dc5c0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:\r\n2020-10-02 07:33:47.146336: I tensorflow\/compiler\/xla\/service\/service.cc:176] StreamExecutor device (0): GeForce GTX 1070, Compute Capability 6.1\r\n10\/02\/2020 07:33:47 - INFO - __main__ - n_replicas: 1, distributed training: False, 16-bits training: False\r\n10\/02\/2020 07:33:47 - INFO - __main__ - Training\/evaluation parameters TFTrainingArguments(output_dir='\/media\/discoD\/models\/datalawyer\/pedidos\/transformers_tf', overwrite_output_dir=True, do_train=True, do_eval=True, do_predict=True, evaluate_during_training=True, evaluation_strategy=<EvaluationStrategy.STEPS: 'steps'>, prediction_loss_only=False, per_device_train_batch_size=4, per_device_eval_batch_size=4, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=1, learning_rate=5e-05, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=4.0, max_steps=-1, warmup_steps=0, logging_dir='runs\/Oct02_07-33-43_user-XPS-8700', logging_first_step=False, logging_steps=1000, save_steps=1000, save_total_limit=None, no_cuda=False, seed=42, fp16=False, fp16_opt_level='O1', local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=1000, dataloader_num_workers=0, past_index=-1, run_name='\/media\/discoD\/models\/datalawyer\/pedidos\/transformers_tf', disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=False, tpu_name=None, xla=False)\r\n10\/02\/2020 07:33:53 - INFO - filelock - Lock 140407857405776 acquired on \/home\/user\/.cache\/huggingface\/datasets\/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock\r\n10\/02\/2020 07:33:53 - INFO - filelock - Lock 140407857405776 released on \/home\/user\/.cache\/huggingface\/datasets\/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"run_tf_text_classification.py\", line 283, in <module>\r\n main()\r\n File \"run_tf_text_classification.py\", line 222, in main\r\n max_seq_length=data_args.max_seq_length,\r\n File \"run_tf_text_classification.py\", line 43, in get_tfds\r\n ds = datasets.load_dataset(\"csv\", data_files=files)\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 604, in load_dataset\r\n **config_kwargs,\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 158, in __init__\r\n **config_kwargs,\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 269, in _create_builder_config\r\n for key in sorted(data_files.keys()):\r\nTypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'\r\n```\r\n\r\n## Expected behavior\r\n\r\nShould be able to run the text-classification example as described in [https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow](https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow)\r\n\r\nOriginally opened this issue at transformers' repository: [https:\/\/github.com\/huggingface\/transformers\/issues\/7535](https:\/\/github.com\/huggingface\/transformers\/issues\/7535). @jplu instructed me to open here, since according to [this](https:\/\/github.com\/huggingface\/transformers\/issues\/7535#issuecomment-702778885) evidence, the problem is from datasets.\r\n\r\nThanks!","comment_length":3,"text":"TypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'\n## Environment info\r\n<!-- You can run the command `transformers-cli env` and copy-and-paste its output below.\r\n Don't forget to fill out the missing fields in that output! -->\r\n \r\n- `transformers` version: 3.3.1 (installed from master)\r\n- `datasets` version: 1.0.2 (installed as a dependency from transformers)\r\n- Platform: Linux-4.15.0-118-generic-x86_64-with-debian-stretch-sid\r\n- Python version: 3.7.9\r\n\r\nI'm testing my own text classification dataset using [this example](https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow) from transformers. The dataset is split into train \/ dev \/ test, and in csv format, containing just a text and a label columns, using comma as sep. Here's a sample:\r\n```\r\ntext,label\r\n\"Registra-se a presen\u00e7a do acad\u00eamico <name> . <REL_SEP> Ao me deparar com a descri\u00e7\u00e3o de dois autores no polo ativo da a\u00e7\u00e3o junto ao PJe , margem esquerda foi informado pela procuradora do reclamante que se trata de uma reclama\u00e7\u00e3o trabalhista individual . <REL_SEP> Diante disso , face a aus\u00eancia injustificada do autor <name> , determina-se o ARQUIVAMENTO do presente processo , com rela\u00e7\u00e3o a este , nos termos do [[ art . 844 da CLT ]] . <REL_SEP> CUSTAS AUTOR - DISPENSADO <REL_SEP> Custas pelo autor no importe de R $326,82 , calculadas sobre R $16.341,03 , dispensadas na forma da lei , em virtude da concess\u00e3o dos benef\u00edcios da Justi\u00e7a Gratuita , ora deferida . <REL_SEP> Cientes os presentes . <REL_SEP> Audi\u00eancia encerrada \u00e0s 8h42min . <REL_SEP> <name> <REL_SEP> Ju\u00edza do Trabalho <REL_SEP> Ata redigida por << <name> >> , Secret\u00e1rio de Audi\u00eancia .\",NO_RELATION\r\n```\r\n\r\nHowever, @Santosh-Gupta reported in #7351 that he had the exact same problem using the ChemProt dataset. His colab notebook is referenced in the following section.\r\n\r\n## To reproduce\r\n\r\nSteps to reproduce the behavior:\r\n\r\n1. Created a new conda environment using conda env -n transformers python=3.7\r\n2. Cloned transformers master, `cd` into it and installed using pip install --editable . -r examples\/requirements.txt \r\n3. Installed tensorflow with `pip install tensorflow`\r\n3. Ran `run_tf_text_classification.py` with the following parameters:\r\n\r\n```\r\n--train_file <DATASET_PATH>\/train.csv \\\r\n--dev_file <DATASET_PATH>\/dev.csv \\ \r\n--test_file <DATASET_PATH>\/test.csv \\\r\n--label_column_id 1 \\\r\n--model_name_or_path neuralmind\/bert-base-portuguese-cased \\\r\n--output_dir <OUTPUT_PATH> \\\r\n--num_train_epochs 4 \\\r\n--per_device_train_batch_size 4 \\\r\n--per_device_eval_batch_size 4 \\\r\n--do_train \\\r\n--do_eval \\\r\n--do_predict \\\r\n--logging_steps 1000 \\\r\n--evaluate_during_training \\\r\n--save_steps 1000 \\\r\n--overwrite_output_dir \\\r\n--overwrite_cache\r\n```\r\n\r\nI have also copied [@Santosh-Gupta 's colab notebook](https:\/\/colab.research.google.com\/drive\/11APei6GjphCZbH5wD9yVlfGvpIkh8pwr?usp=sharing) as a reference.\r\n\r\n<!-- If you have code snippets, error messages, stack traces please provide them here as well.\r\n Important! Use code tags to correctly format your code. See https:\/\/help.github.com\/en\/github\/writing-on-github\/creating-and-highlighting-code-blocks#syntax-highlighting\r\n Do not use screenshots, as they are hard to read and (more importantly) don't allow others to copy-and-paste your code.-->\r\n\r\nHere is the stack trace:\r\n\r\n```\r\n2020-10-02 07:33:41.622011: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n\/media\/discoD\/repositorios\/transformers_pedro\/src\/transformers\/training_args.py:333: FutureWarning: The `evaluate_during_training` argument is deprecated in favor of `evaluation_strategy` (which has more options)\r\n FutureWarning,\r\n2020-10-02 07:33:43.471648: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcuda.so.1\r\n2020-10-02 07:33:43.471791: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.472664: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1716] Found device 0 with properties: \r\npciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1\r\ncoreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB\/s\r\n2020-10-02 07:33:43.472684: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n2020-10-02 07:33:43.472765: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10\r\n2020-10-02 07:33:43.472809: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10\r\n2020-10-02 07:33:43.472848: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10\r\n2020-10-02 07:33:43.474209: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10\r\n2020-10-02 07:33:43.474276: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10\r\n2020-10-02 07:33:43.561219: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7\r\n2020-10-02 07:33:43.561397: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.562345: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.563219: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1858] Adding visible gpu devices: 0\r\n2020-10-02 07:33:43.563595: I tensorflow\/core\/platform\/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations: AVX2 FMA\r\nTo enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.\r\n2020-10-02 07:33:43.570091: I tensorflow\/core\/platform\/profile_utils\/cpu_utils.cc:104] CPU Frequency: 3591830000 Hz\r\n2020-10-02 07:33:43.570494: I tensorflow\/compiler\/xla\/service\/service.cc:168] XLA service 0x560842432400 initialized for platform Host (this does not guarantee that XLA will be used). Devices:\r\n2020-10-02 07:33:43.570511: I tensorflow\/compiler\/xla\/service\/service.cc:176] StreamExecutor device (0): Host, Default Version\r\n2020-10-02 07:33:43.570702: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.571599: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1716] Found device 0 with properties: \r\npciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1\r\ncoreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB\/s\r\n2020-10-02 07:33:43.571633: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n2020-10-02 07:33:43.571645: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10\r\n2020-10-02 07:33:43.571654: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10\r\n2020-10-02 07:33:43.571664: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10\r\n2020-10-02 07:33:43.571691: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10\r\n2020-10-02 07:33:43.571704: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10\r\n2020-10-02 07:33:43.571718: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7\r\n2020-10-02 07:33:43.571770: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.572641: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:43.573475: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1858] Adding visible gpu devices: 0\r\n2020-10-02 07:33:47.139227: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1257] Device interconnect StreamExecutor with strength 1 edge matrix:\r\n2020-10-02 07:33:47.139265: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1263] 0 \r\n2020-10-02 07:33:47.139272: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1276] 0: N \r\n2020-10-02 07:33:47.140323: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.141248: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.142085: I tensorflow\/stream_executor\/cuda\/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero\r\n2020-10-02 07:33:47.142854: I tensorflow\/core\/common_runtime\/gpu\/gpu_device.cc:1402] Created TensorFlow device (\/job:localhost\/replica:0\/task:0\/device:GPU:0 with 5371 MB memory) -> physical GPU (device: 0, name: GeForce GTX 1070, pci bus id: 0000:01:00.0, compute capability: 6.1)\r\n2020-10-02 07:33:47.146317: I tensorflow\/compiler\/xla\/service\/service.cc:168] XLA service 0x5608b95dc5c0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:\r\n2020-10-02 07:33:47.146336: I tensorflow\/compiler\/xla\/service\/service.cc:176] StreamExecutor device (0): GeForce GTX 1070, Compute Capability 6.1\r\n10\/02\/2020 07:33:47 - INFO - __main__ - n_replicas: 1, distributed training: False, 16-bits training: False\r\n10\/02\/2020 07:33:47 - INFO - __main__ - Training\/evaluation parameters TFTrainingArguments(output_dir='\/media\/discoD\/models\/datalawyer\/pedidos\/transformers_tf', overwrite_output_dir=True, do_train=True, do_eval=True, do_predict=True, evaluate_during_training=True, evaluation_strategy=<EvaluationStrategy.STEPS: 'steps'>, prediction_loss_only=False, per_device_train_batch_size=4, per_device_eval_batch_size=4, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=1, learning_rate=5e-05, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=4.0, max_steps=-1, warmup_steps=0, logging_dir='runs\/Oct02_07-33-43_user-XPS-8700', logging_first_step=False, logging_steps=1000, save_steps=1000, save_total_limit=None, no_cuda=False, seed=42, fp16=False, fp16_opt_level='O1', local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=1000, dataloader_num_workers=0, past_index=-1, run_name='\/media\/discoD\/models\/datalawyer\/pedidos\/transformers_tf', disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=False, tpu_name=None, xla=False)\r\n10\/02\/2020 07:33:53 - INFO - filelock - Lock 140407857405776 acquired on \/home\/user\/.cache\/huggingface\/datasets\/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock\r\n10\/02\/2020 07:33:53 - INFO - filelock - Lock 140407857405776 released on \/home\/user\/.cache\/huggingface\/datasets\/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"run_tf_text_classification.py\", line 283, in <module>\r\n main()\r\n File \"run_tf_text_classification.py\", line 222, in main\r\n max_seq_length=data_args.max_seq_length,\r\n File \"run_tf_text_classification.py\", line 43, in get_tfds\r\n ds = datasets.load_dataset(\"csv\", data_files=files)\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 604, in load_dataset\r\n **config_kwargs,\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 158, in __init__\r\n **config_kwargs,\r\n File \"\/media\/discoD\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 269, in _create_builder_config\r\n for key in sorted(data_files.keys()):\r\nTypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'\r\n```\r\n\r\n## Expected behavior\r\n\r\nShould be able to run the text-classification example as described in [https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow](https:\/\/github.com\/huggingface\/transformers\/tree\/master\/examples\/text-classification#run-generic-text-classification-script-in-tensorflow)\r\n\r\nOriginally opened this issue at transformers' repository: [https:\/\/github.com\/huggingface\/transformers\/issues\/7535](https:\/\/github.com\/huggingface\/transformers\/issues\/7535). @jplu instructed me to open here, since according to [this](https:\/\/github.com\/huggingface\/transformers\/issues\/7535#issuecomment-702778885) evidence, the problem is from datasets.\r\n\r\nThanks!\nThanks @lhoestq !","embeddings":[-0.2919273078,-0.6131456494,-0.0538775101,0.0724286065,0.5358543992,0.0512854382,0.5790147781,0.301427871,0.288526386,0.1793083698,-0.0349086858,0.1695188433,-0.0423114263,-0.1191879734,-0.1670343578,-0.2730587125,-0.1069535911,0.1412786692,-0.4372419417,-0.0667978898,0.0061711208,0.1115164608,-0.3358640075,0.0874110833,-0.3546169102,-0.1470186561,0.1937073618,0.0334143899,0.1056338474,-0.3845270872,0.366317004,-0.0270779245,0.1710627675,0.491507113,-0.0001042778,0.081310913,0.4950879514,-0.138119936,-0.0701588094,-0.343044579,-0.4419636428,0.0307981856,0.0160502605,-0.3915827572,-0.2448375076,-0.0110362982,-0.1189130321,-0.0461345613,0.6161757112,0.278801918,0.2397651076,0.2381774187,-0.1034501866,-0.3372361064,-0.1007003114,0.3406432867,-0.1360172033,-0.1092183441,0.1256229579,-0.2894225121,-0.0251126885,0.2308775187,0.0032396459,-0.089165844,0.3362298608,0.126687631,-0.1719463766,0.0904838964,0.01274959,0.2484526038,0.3977201581,-0.191258952,-0.201420784,-0.1912941635,-0.4447488189,-0.3844889998,0.0586855598,0.0405776054,0.0833992809,0.2616128325,-0.0061038891,0.4328720272,-0.2141648531,-0.1010648534,0.0205578785,0.5832354426,-0.1706664115,0.0971066877,0.0810640901,-0.0337578095,0.1662235856,-0.1250212342,0.2540244162,0.0342846215,-0.2468905896,-0.012379894,-0.0421408713,-0.4438541234,-0.214487657,-0.1216932535,0.2730001509,-0.0499446429,0.0718224198,0.2485999912,0.1498123109,0.2958030701,0.0911678746,0.537527442,0.2819210589,0.3849364817,-0.0427861772,-0.1365961283,-0.142474398,-0.158597067,-0.0144823557,0.4171228111,0.0576643199,-0.1039039567,-0.30254969,-0.2212817669,0.0309037361,-0.0660278946,-0.0246453453,0.4157520235,-0.0416241586,0.3547725976,0.2065986395,0.0611954629,-0.0000107179,-0.1164428368,-0.1507170498,0.1743426323,-0.3866131902,-0.2293417454,-0.0172503423,0.2606917918,0.0060736635,-0.1308667958,-0.083957687,-0.0662648901,-0.0675454959,0.0104777887,0.038608443,0.2141883671,-0.1814568043,0.1716786176,0.1413978785,-0.142892912,-0.0168804023,0.2545795739,-0.0376725495,-0.2689933479,0.0494651496,0.1983073056,0.0924026594,0.0822201595,0.0646613985,0.1997675002,0.2888107002,-0.0023917381,0.0389889143,0.0640591905,-0.1613370031,0.1434037834,0.0981706604,0.3937546015,-0.7670513988,-0.2362775803,-0.0231548995,0.0841976553,-0.0441100746,0.4377836287,-0.3081965744,0.2548838556,0.1809004545,0.2804787159,0.1315836161,-0.4054885507,-0.2366541028,-0.0511300489,0.0623557344,-0.1371805072,0.1976591647,-0.1599786133,-0.0393890887,-0.2148528397,-0.2731992602,0.1258901358,-0.0372625664,0.0372501723,-0.3978560269,-0.0586486682,0.1387393624,0.0166069958,0.0908488706,-0.0245048665,-0.0067748767,0.5229411125,0.1247486845,-0.3379059136,0.1744470745,0.2543297112,0.0626068562,0.2155818939,0.018898882,-0.2434887141,-0.2023972422,0.1049191728,-0.0260952972,0.3419173658,-0.2349476367,-0.1734545231,-0.4128475785,0.0400041379,0.0303186756,-0.1581599265,0.1817014217,0.0163617879,0.0562101454,-0.1548790336,0.0504410565,0.0850315765,-0.0169340409,0.1152144745,-0.3652016521,0.205051899,-0.2516756356,-0.0821671262,-0.126055941,0.208872363,0.2595393658,-0.2008835673,-0.1568623483,0.3399887383,-0.2882199287,0.1538384408,-0.3089927435,-0.0751828775,0.2699451447,-0.4763702154,0.0024343291,0.1310064346,0.0902314931,-0.0006721493,-0.4511058927,0.1370696127,0.0112843513,0.1648834497,0.1284329742,0.0007869658,0.1388457268,-0.1832198799,-0.0573632009,-0.1072910875,-0.0001398468,0.1755349934,-0.2161804736,0.0637440309,-0.1466740072,-0.1216330752,0.4897304475,-0.0830696449,0.2121409774,-0.0066706208,-0.2014056146,0.1740379184,0.0147640789,0.2376895696,0.1019209698,0.1063609719,-0.1818083525,0.1594488174,-0.1297113597,-0.1348221153,0.1366052777,0.1401495934,0.1842842102,0.047402259,-0.0023738095,0.0637102649,0.0001989906,-0.3237716556,0.0035916145,0.0375333168,-0.2615197003,0.1092323586,-0.2121478766,-0.1581652462,-0.2769756615,0.241360575,0.1923743635,-0.0711507946,-0.0779944956,0.0264965203,-0.2088043094,0.3075107038,0.1213724911,0.2280253023,0.2589328587,-0.1669288725,-0.1522791386,-0.3805861771,-0.1148757711,0.1299654692,0.1442116201,0.1163031459,-0.0103829578,0.0811195746,0.0132961869,0.4569865167,-0.5159244537,0.0750539526,-0.314931035,-0.2156138271,-0.0008078428,-0.0266429707,0.0513779521,-0.1849368513,0.3361887038,-0.1817682087,0.0587969385,-0.0685050115,0.1174616069,-0.2828928232,-0.1832380444,-0.6926159859,-0.2156877816,-0.4069219828,0.2902898192,-0.1129137129,-0.1000652313,0.4519687295,-0.04145924,-0.1367301047,0.2323179543,-0.0375305265,0.0842413232,0.2739644647,0.3995569944,-0.1479191929,-0.3361201286,-0.037801262,-0.1635225713,0.0331597701,-0.2620622218,-0.1347199827,-0.0583056845,-0.2712892592,-0.1474146694,0.1512570232,-0.1404368728,0.6117979288,0.1744675785,-0.1556137651,-0.1580144018,-0.4322287738,0.2149780095,0.0620988421,0.0518928878,-0.0564923882,0.5008730292,0.1867362559,0.6495117545,0.0560511127,-0.1962005943,-0.0095487041,0.1530277133,0.0647748932,-0.0284662768,-0.2616336942,0.0940554515,-0.1156798303,-0.1356900036,0.1677749753,-0.1896721721,-0.1575728357,-0.1722387522,0.1542970687,-0.0772994012,-0.0894062594,0.1257669032,0.1191693991,0.2297199517,0.1420507431,-0.0787234679,-0.10656555,-0.1022997648,0.096433714,-0.0871651992,-0.1488742679,-0.0652406886,-0.3501496017,0.2422767282,-0.1469796896,0.2825861275,0.2371597588,0.1074166596,-0.2071240842,-0.1536873728,-0.2497842014,0.014652282,0.1408586353,-0.3110143244,0.1866850853,0.1122649312,-0.0809816793,-0.1901675463,-0.2099327296,-0.3738399744,-0.298717618,0.453838259,0.2226490676,-0.185783416,-0.1639241427,0.2885839641,-0.1199201718,-0.0421009362,-0.066820167,-0.0768557042,0.0182460137,-0.2515257597,-0.0107603529,0.5364021659,0.2025124282,-0.2095840871,-0.0444211029,0.0943264514,-0.2034685016,0.0882901251,0.2573613226,0.1361691803,-0.1558400393,-0.0760446042,0.1248814166,0.2837224007,-0.3310799897,0.5674952865,-0.0255851038,-0.4515457153,0.0420375802,-0.0200555008,0.1364636868,0.1572538167,-0.0916863009,-0.0461933538,0.1043341309,0.0756744668,-0.3240720034,0.1180027127,0.1891417652,0.3833045363,-0.4595752358,-0.2022822946,0.1838428825,0.0292732008,-0.0668438151,-0.0761145428,-0.1605715901,-0.4441316724,0.3648698926,0.0461951494,0.8255496025,-0.1720346361,-0.0274047237,-0.2060128003,0.1214447916,0.5401180983,-0.4126571417,0.2696087658,-0.2367256135,-0.3283510208,-0.1095912084,-0.0516983978,0.3606872559,0.4803798199,-0.3971676826,0.2573061287,0.103219524,0.5092794299,0.0648841634,0.136406064,0.4459854364,-0.3245070875,-0.2694648802,0.1508009285,0.0748465508,0.0509600416,0.0084982747,-0.2146799117,-0.2909614444,-0.1904337853,-0.318582505,0.3504792452,-0.2043990791,0.0957831964,0.2979163826,0.0655586198,0.4973452091,0.3372198641,0.3578670621,0.2304675281,-0.2153190523,0.2350308448,-0.4642540514,-0.0610615797,0.0088658445,0.251606375,0.227899462,0.0168934893,-0.1576444358,0.1514526308,-0.0659205243,-0.239291802,-0.0809610933,0.0206167698,0.2474818826,-0.5286822319,0.0103748515,0.1242398694,-0.263369292,-0.2806583047,0.2036949992,-0.3791154921,-0.1899057478,0.1543290168,0.1162807718,-0.2372897416,0.0348868333,0.3466538191,0.007645539,0.2065514922,0.2748655975,0.1189157292,-0.1234726608,-0.2976979017,-0.1585630625,0.2176708281,-0.5927738547,0.0982700586,-0.2696886063,-0.4663753808,0.121622853,0.4874913394,0.3218313754,0.1377170533,-0.1893444061,-0.4025084376,-0.2644128799,0.0103745395,0.1245822832,0.2435663193,0.1476768851,0.6597263813,0.0173819643,0.0904994458,-0.3986144662,-0.2800250053,-0.1478807628,0.295852989,0.3108174205,-0.3957287371,0.3502090871,-0.0701561272,0.1842743754,0.4157313108,-0.2173074782,-0.2901489437,-0.0977177992,0.0736077875,0.1292785853,-0.2479224503,-0.2928442061,0.0705482587,0.0738156065,-0.3057259917,0.060341455,0.0323538966,-0.0002617335,0.3716009855,-0.2391617298,0.1032218561,-0.1942325234,-0.1802102327,-0.1308957338,0.2095948011,0.0563998781,0.0817958862,0.2029102743,-0.0294597894,-0.4282607436,0.0897560492,0.2320590913,0.260022521,0.3225409389,-0.3676284552,-0.0552807488,0.1726189703,0.5626053214,0.0809191167,-0.1066219062,0.0329805873,-0.0473479591,0.2685658038,-0.3096589446,0.2288379371,-0.0930913463,0.067760922,0.2696631253,0.2911012471,0.2975101173,-0.0638441667,0.4036651254,0.2677652836,0.5184584856,-0.4800507724,0.3569121361,0.5745406747,0.1697940677,0.1794567853,0.3768188655,-0.0295521896,0.3553950787,0.5121091604,0.2054487318,0.1787361652,0.0035348476,0.1895809472,0.0774413496,-0.15311791,0.0910538435,0.085694313,-0.1319064498,0.0842484906,-0.2042768598,0.3030064106,-0.3348991275,-0.0643551946,-0.137806654,0.0549418107,-0.2147033364,-0.2030396312,0.1460205913,0.0916059837,-0.0967579708,-0.1491556168,0.1459876448,0.2719106674,0.3311924934,-0.2415809184,0.011369125,-0.186536476,-0.1872024983,-0.022918554,0.3771433234,-0.2421434671,0.3255573809,0.1919579655,-0.2189380825,-0.0256017949,0.3311004043,0.3879540563,0.346213311,-0.0973591655,-0.0188501198,-0.1468135267,-0.2428473681,0.1603179723,0.290438056,0.1453945041,0.1796865314,0.2034573555,0.1824687421,-0.2012612075,0.2441810966,-0.045337636,0.0992114991,0.0013043385,0.0638233572,0.0748212859,-0.0474862158,-0.0048640999,0.0407558642,-0.3722771108,-0.0838894472,0.364733696,-0.2776278853,0.4735959768,-0.0817337409,0.1044070572,0.0357249528,0.3190372884,-0.0509206988,-0.2098562717,-0.220578298,0.0366850682,-0.246158421,-0.0787129104,0.0871472582,-0.2164678574,-0.2482390255,0.0417579189,0.1564619392,0.0311662164,-0.1432110369,-0.315692544,0.010035662,-0.0800714567,-0.1900778562,-0.1123893782,0.0135460198,0.0193593539,0.2289910316,-0.2942387164,-0.0722259209,0.4261686206,0.0742083043,-0.4140973091,-0.0033793557,-0.0250940695,0.2368596941,0.1539556831,0.097318992,0.569098711,-0.2223181427,0.1948460042,-0.077134043,-0.3015601039,-0.1696515381,0.2426602244,0.3290059865,0.3649009764,-0.1513296813,0.1434555501,-0.2218807787,-0.1535740495,0.0497781672,-0.2138116062,-0.2299725264,-0.0495054238,-0.2060346454,-0.0614992492,0.1562587619,0.2371859103,0.0235220343,-0.0085253129,-0.181430459,-0.5764254928,0.4325292706,-0.3074938059,-0.4217503965,0.119653061,-0.0462134816,0.2443118989,0.0112836268,-0.4401860535,0.0656878874,0.2044553459,-0.0260742437,-0.2218872756,0.058609765,0.1953352988,-0.0228522941,-0.0969520509,0.2508562207,0.1856728643,-0.079318516,-0.0641221404,-0.1950844824]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/699","title":"XNLI dataset is not loading ","comments":"also i tried below code to solve checksum error \r\n`datasets-cli test .\/datasets\/xnli --save_infos --all_configs`\r\n\r\nand it shows \r\n\r\n```\r\n2020-10-02 07:06:16.588760: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nTraceback (most recent call last):\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 268, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 474, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/.\/datasets\/xnli\/xnli.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 279, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 474, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/.\/datasets\/xnli\/xnli.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/opt\/conda\/bin\/datasets-cli\", line 36, in <module>\r\n service.run()\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/commands\/test.py\", line 76, in run\r\n module_path, hash = prepare_module(path)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 283, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at .\/datasets\/xnli\/xnli.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/.\/datasets\/xnli\/xnli.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/.\/datasets\/xnli\/xnli.py\r\n```\r\n\r\n","body":"`dataset = datasets.load_dataset(path='xnli')`\r\n\r\nshowing below error \r\n```\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']\r\n```\r\n\r\nI think URL is now changed to \"https:\/\/cims.nyu.edu\/~sbowman\/xnli\/XNLI-MT-1.0.zip\"","comment_length":170,"text":"XNLI dataset is not loading \n`dataset = datasets.load_dataset(path='xnli')`\r\n\r\nshowing below error \r\n```\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']\r\n```\r\n\r\nI think URL is now changed to \"https:\/\/cims.nyu.edu\/~sbowman\/xnli\/XNLI-MT-1.0.zip\"\nalso i tried below code to solve checksum error \r\n`datasets-cli test .\/datasets\/xnli --save_infos --all_configs`\r\n\r\nand it shows \r\n\r\n```\r\n2020-10-02 07:06:16.588760: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nTraceback (most recent call last):\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 268, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 474, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/.\/datasets\/xnli\/xnli.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 279, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/file_utils.py\", line 474, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/.\/datasets\/xnli\/xnli.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/opt\/conda\/bin\/datasets-cli\", line 36, in <module>\r\n service.run()\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/commands\/test.py\", line 76, in run\r\n module_path, hash = prepare_module(path)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 283, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at .\/datasets\/xnli\/xnli.py, or remotely at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.2\/datasets\/.\/datasets\/xnli\/xnli.py or https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/datasets\/datasets\/.\/datasets\/xnli\/xnli.py\r\n```\r\n\r\n","embeddings":[-0.2810523212,0.140100345,-0.0911717787,0.1260449886,0.2266674191,-0.2409968972,0.3072917163,0.4281842411,0.0896520987,-0.0614517219,-0.0190948769,0.4737130105,0.1864162832,0.1176810041,0.1495385915,0.219538033,0.0383550376,0.094383806,-0.0668891147,-0.1008866653,-0.2314456105,0.0685328618,-0.2965807021,-0.1154323369,-0.0837840065,0.1186984032,0.215154022,0.2002559602,-0.0664989576,-0.2786804438,0.3651867807,-0.0079631768,0.263148725,0.2762711346,-0.0001126595,0.2332450598,0.3754191995,-0.1614189744,-0.4101258516,-0.0952022448,-0.4924982488,-0.1245010495,-0.1306751519,-0.2602872849,0.1164859161,0.061470978,-0.0578304157,-0.1681259871,-0.0545159765,0.4935926199,0.1621233076,0.5488842726,0.2709943354,0.2150414437,-0.0674156547,-0.3017497957,0.0076497304,0.4067320824,0.3300507665,0.014924285,0.055706773,0.1214362085,-0.0206639655,0.000889206,0.158074826,-0.1624352485,0.2153967768,-0.0836901218,0.0030177215,0.2048472762,0.397978425,-0.1201249063,-0.4151076376,0.1187436283,0.0617506504,-0.1697445065,0.4212925434,0.1787170917,-0.3158423603,-0.1220314577,-0.3779917061,-0.0248488933,-0.175771445,0.3431745172,-0.0073401993,0.4367713928,-0.0140386047,0.0259262361,0.0205369312,-0.0988183767,0.2914394438,-0.0606681034,0.0233033914,0.1861115694,-0.6986988187,0.0651394948,-0.0342235416,-0.0879542679,0.1597488374,0.395940125,0.3262369931,0.0806987062,-0.2925149202,0.1523523331,-0.1719446629,0.387106508,0.1406884938,0.0790466443,0.2752794623,0.2361992598,-0.0131360972,0.0545628145,-0.300019145,-0.1971398443,-0.0815355331,0.0388958752,0.1401141584,-0.2044802755,-0.2629959583,0.0759377778,-0.0784603208,-0.138380006,0.4358338416,0.4694562554,-0.320874244,0.2498867661,0.0098449374,0.0181956422,-0.0383259542,-0.1316920221,-0.1493628621,0.071669206,-0.2305678427,-0.0721811801,0.2829676867,-0.3151513636,0.4981717169,-0.1555375308,-0.1146094799,-0.3088402748,0.0614730939,0.0018759954,-0.0404581651,0.2428904325,-0.0249700118,-0.033209119,0.072953172,-0.0030836209,-0.0557516925,0.1683819145,-0.3589117229,-0.2988475561,-0.0545134991,0.2073375136,-0.1995881349,-0.1916325092,-0.0024796955,-0.4883229434,0.0660807788,-0.0477588437,-0.1237207875,-0.2153945267,-0.0667610615,-0.209039852,0.5407055616,0.1293569803,-0.3490072489,0.0516496077,-0.1369031072,-0.0557792149,0.0964061096,0.2605536282,-0.1590194851,0.3728075624,-0.3403566778,-0.1178918779,0.5776913762,-0.476293087,-0.4399244487,0.2114343941,-0.1018751934,0.0079044383,-0.1255787462,0.2523783445,-0.1356548816,-0.0352004357,0.0137475841,0.2980307937,0.1415857971,0.1215579435,-0.2224131823,-0.2473589331,0.1558611095,0.3051367104,0.0188241675,0.1345297247,-0.0449265689,0.0668575466,0.3322269022,-0.0833622739,-0.0464372523,0.221496582,0.4732953012,-0.0393512286,0.0370533764,-0.0048491447,-0.4131316543,0.3650059998,-0.0015582185,-0.0114370799,-0.0650833026,0.1789069325,-0.3976843357,-0.024558926,-0.3283422887,-0.1062684953,0.0848114043,0.2809491754,0.1377623379,0.0063051474,-0.3725928068,0.281760335,-0.8748972416,0.045638036,-0.4280187786,0.4332666099,0.0865833908,-0.0375143439,0.0658052489,0.2386031151,-0.1012093499,-0.2264588177,-0.3001839519,0.3219572604,-0.0343309045,-0.0100150499,0.2217055708,0.1605622619,0.0023727182,-0.3698914945,0.0630576536,0.1883188784,0.0396234393,-0.0004497514,-0.057606101,0.2553064227,-0.1893199533,-0.0278984662,0.166231975,0.0318163037,0.4790482521,-0.3187073171,-0.0347367786,-0.1404661089,0.2331364304,0.30008775,0.604013741,0.1275435239,-0.0661315992,-0.2273055166,0.3051785827,0.0821915567,-0.1016039029,0.1684033722,0.0007089683,0.0061235605,0.0291468389,0.306537658,0.3231015801,-0.0140058435,-0.1969237477,0.0177570265,-0.1565663815,-0.1717172265,0.0702453777,-0.1741173416,0.2569725215,0.2882356048,-0.0568803772,-0.0213246997,-0.2173702568,-0.3831168413,0.2945067883,0.3578618765,-0.3981689513,-0.2213178128,-0.226242587,-0.4142237008,-0.3695698977,-0.1048752293,-0.2279290706,-0.2551909685,-0.2312048525,0.3115474582,0.183365792,0.2592550814,-0.1243881136,0.1447745562,0.2544369698,-0.4136667252,0.0239859987,-0.0903487653,-0.1083464026,0.0917411596,0.5014837384,0.1665120721,0.1969742775,-0.2534525394,0.0452200584,-0.5949463248,-0.2944060862,-0.0161856711,0.0371351354,0.0740418658,-0.074483037,0.3668269813,0.1976650357,-0.3344924748,0.3055459857,0.0922012329,-0.1450958401,0.1104357317,-0.1262268722,-0.160679549,0.0007932255,-0.3663280904,-0.1312322766,-0.2736053765,-0.1452491134,-0.1186463833,0.24539873,0.2374483794,0.0201194752,0.1560905874,0.10875085,0.0869797915,0.0980706513,-0.5159553885,0.3297222853,0.0439850315,-0.406933099,0.0055010873,-0.0666834489,0.5552334785,0.2053047568,-0.4230065048,-0.0054273717,-0.1825606525,0.2202462852,0.2052992284,-0.1584539115,0.0227844305,-0.2259042114,-0.0106735984,-0.1840025187,-0.0045016324,0.2184030861,0.1114478484,0.6592665315,0.3477965593,0.2220592499,-0.0564368404,0.11664965,0.5052437782,-0.3023961186,0.2608957589,-0.0986781493,0.1178187355,0.0728700235,-0.2802758217,0.1006157994,0.0264769923,-0.070226714,-0.0979865342,-0.0083872667,0.13620314,-0.1991450042,0.1025647596,-0.2736385763,-0.2417384833,0.0471471027,-0.360360533,0.3171441555,-0.0406246446,-0.0160366595,-0.1175242066,-0.2542997599,0.0467978865,0.5146548748,-0.0103428178,-0.1609942913,-0.4704329073,-0.168560341,-0.1489563286,0.0452453494,0.0624945015,0.7762842178,-0.0286614597,0.027804926,-0.0215605814,-0.0729917809,0.3670107424,-0.2659906745,0.3630941212,0.1105107814,0.0825680345,-0.1649070531,-0.0846161172,0.0318948738,-0.354118228,0.6948394179,0.3220119178,0.1106785461,-0.0690888613,0.0601812936,0.2269338071,-0.2035052776,-0.2965390086,-0.3927070498,-0.2024308145,-0.3153542578,-0.2363517731,-0.4139214158,0.2197584063,-0.0074460115,-0.205004245,0.0748780295,-0.267609179,0.13763915,0.1148417518,0.1625653356,0.1734159291,0.1625549942,-0.2402063757,0.2352827489,0.1725134701,0.318329215,-0.1475915313,-0.2885357141,-0.003833594,-0.13595137,0.0959357396,-0.0350906663,-0.153297022,-0.0839427263,0.0986499265,0.0651265606,0.0621594749,0.2465730906,0.1049913317,0.230098933,-0.3477571905,-0.3568046689,0.2147634178,0.0604007952,0.1153791174,0.7290050983,-0.2933385372,-0.1288176924,-0.0421995297,0.1484813243,0.5102241635,-0.0432354063,-0.2476459742,0.3940462172,0.1369661689,0.2234213054,-0.1048373058,0.0619109981,-0.3449874222,-0.253799051,-0.0249620788,-0.3489011824,0.0853083134,-0.010192492,0.025941005,0.2195088565,0.0459111221,0.2277482748,0.3119932711,0.5255157948,-0.0172784142,-0.1265279353,-0.3445289433,0.0857932046,-0.0783056542,0.3728111386,-0.0303588081,-0.0005271425,-0.1853592247,0.0276624281,-0.2067966461,0.4837397933,0.0697026402,0.4298190176,0.0664979443,-0.0753155649,-0.079552874,0.5727033615,0.5105199218,0.0839171484,-0.223528862,0.2138012201,0.1702681333,0.1728147268,0.0185814071,-0.0168045573,0.3583163619,-0.260016799,-0.2901149392,0.1926204264,-0.104221262,-0.4215745032,0.0394274555,-0.0934735611,-0.1304834634,-0.0870987549,0.0148835331,0.0587170459,-0.1404230446,-0.0161365476,0.1427765489,0.1371825337,0.0019288848,0.1217521727,0.1317557544,-0.349781692,0.0614108406,0.2858654559,-0.0064142803,-0.0014055388,0.5304970741,-0.0700174719,0.1519463956,-0.2194432765,-0.0852406994,0.0494390391,-0.221242249,-0.1986135542,-0.0840748325,0.1187090352,0.0725146458,0.4365159571,0.2535585761,0.303075552,-0.0325133689,-0.6876363158,-0.3423904479,0.0568886101,-0.1715882868,0.1885706931,-0.3884999156,0.3960308135,0.0748401359,-0.1774811298,-0.2731969357,-0.0326643288,-0.4017135799,-0.0267323758,0.0280214716,-0.0176643934,0.1672491133,-0.0709349364,0.1682512611,-0.1040004119,-0.2033444941,-0.1754015535,-0.0935202166,0.1171932518,0.1630305052,-0.2208752334,-0.2047118694,-0.259968698,-0.0981271639,-0.0686718896,-0.128551811,-0.1265994906,-0.0201878361,0.1087884307,-0.0913866609,0.0690638945,-0.1085224301,0.0635514483,-0.3368259966,0.084159717,0.0912641361,0.183292076,-0.0921068043,0.080088608,-0.4293678999,-0.0950773209,-0.1050019115,-0.1158129051,0.2806726098,-0.1528832465,0.0681754574,-0.0840163082,0.1499198824,0.3600474,-0.3264266253,-0.0737648681,0.2524220049,0.2111978382,-0.2323976904,0.0079622502,0.0975191817,-0.0128283892,-0.0809166133,-0.0202721227,0.1736266762,-0.3428405523,0.383328855,-0.0789623559,0.286729455,-0.2502049506,0.136477977,0.271020174,0.0146723399,0.0450892746,0.0056294403,-0.0391099453,-0.2255387753,0.195848003,-0.085773766,0.2376942486,-0.1642831862,0.5021861792,-0.104235217,-0.3124946654,-0.0067890813,0.1271679252,-0.4421218634,-0.0838351622,0.0046896874,0.352208674,-0.1192335933,0.2068729252,-0.2747467756,0.1378641874,-0.1530071497,0.1626491249,-0.6818800569,-0.2202686369,0.1333305836,0.0085037798,0.1705940366,-0.0937691778,-0.2688525319,0.3143114448,-0.3887181878,-0.3878394663,0.0650330335,0.3368549645,-0.0120461415,-0.2356077433,0.3604293764,0.1550465673,-0.0725589097,0.0559229851,0.0326236635,0.279287219,0.5495951176,0.5028504133,0.141716063,0.2905434966,0.0051555685,0.0096761556,0.2727641761,0.2030745894,0.1011421978,0.3449248075,0.2064351141,-0.0650844201,-0.1087738723,-0.2298966795,0.2433483899,0.0245732795,0.5131148696,0.1457114518,0.1192423329,-0.2304068208,0.0868743435,-0.2940407991,0.0426469855,0.3270275593,0.0896346048,0.123476319,0.0342260487,0.0581901483,-0.3124855161,0.365192771,0.3919512033,0.1219502985,-0.2810330987,-0.7514004707,-0.785620153,0.081270963,0.1031554937,-0.2206645161,0.1371860802,0.1202625185,-0.2577443719,-0.1677394211,0.0545695461,-0.2219514549,0.052715268,-0.0193315893,-0.3758400083,-0.0948539525,-0.3110665381,0.1669647694,-0.2070860118,-0.2815288305,0.0894439965,0.0347800478,0.0689683929,0.0055830562,0.0771625116,0.1766042709,0.2179130167,0.2082778662,-0.0733696893,0.6868266463,0.0379902609,-0.1654130071,-0.2968116403,-0.0329283401,-0.0112932781,0.1104076877,0.1009301022,0.2436001301,0.1682621986,0.2485682666,-0.4946317375,0.4898337722,0.0473821834,-0.5125205517,-0.1670876443,-0.0022595946,0.0885946974,0.0051781288,0.2704469562,0.1382778138,0.0334279276,0.2477683574,-0.1295214295,-0.1330000162,0.3874620199,-0.1668900251,0.0412997082,-0.0317719467,0.1995337158,-0.3064464331,0.1330668181,-0.5193888545,0.0498286635,0.4189535677,-0.0191901661,-0.0867763832,0.1279964149,0.1789079159,0.3869445026,-0.076261051,-0.106206134,0.0338066481,-0.2182243317,-0.0525420122,-0.1444694251]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/699","title":"XNLI dataset is not loading ","comments":"Hi !\r\nYes the download url changed.\r\nIt's updated on the master branch. I'm doing a release today to fix that :)","body":"`dataset = datasets.load_dataset(path='xnli')`\r\n\r\nshowing below error \r\n```\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']\r\n```\r\n\r\nI think URL is now changed to \"https:\/\/cims.nyu.edu\/~sbowman\/xnli\/XNLI-MT-1.0.zip\"","comment_length":22,"text":"XNLI dataset is not loading \n`dataset = datasets.load_dataset(path='xnli')`\r\n\r\nshowing below error \r\n```\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']\r\n```\r\n\r\nI think URL is now changed to \"https:\/\/cims.nyu.edu\/~sbowman\/xnli\/XNLI-MT-1.0.zip\"\nHi !\r\nYes the download url changed.\r\nIt's updated on the master branch. I'm doing a release today to fix that :)","embeddings":[-0.1545367837,0.328220129,-0.0674623325,0.1016497761,0.1184456944,-0.0849034786,0.1139141247,0.3517860472,-0.0838156715,-0.1345048845,-0.0737449005,0.351093322,0.2048691064,0.0506398119,0.1383061111,0.3167421222,0.1006802842,0.0527608953,-0.1457019895,-0.1142641678,-0.3562244773,0.2969622612,-0.3298872411,0.0699843317,-0.161618337,0.1915494651,-0.0262833908,0.2155114561,-0.2241702974,-0.5219001174,0.46504426,-0.0289612003,0.315335989,0.0534079447,-0.0001107943,0.013320378,0.4856797755,-0.0701202005,-0.4200716019,0.0991579592,-0.5813625455,-0.368876636,-0.0407271609,-0.2189436555,0.1747738719,-0.0294561759,0.0603701323,0.0087949587,-0.055560302,0.4349479377,0.2022579461,0.3315931261,0.2125155479,0.1041497588,-0.1241293699,-0.1358566433,0.2027838379,0.4185233712,0.2098928541,0.0184144136,0.0820090547,0.0695741698,-0.1658983529,-0.1311559528,0.0142947109,-0.0650999174,0.1832603663,-0.0951679498,-0.0232734568,0.1045117155,0.4157608747,-0.0431302972,-0.414258033,0.0711860582,0.1402248889,-0.1816168427,0.2897528112,0.1656956822,-0.3068470657,-0.0380344763,-0.3423021138,-0.2752628624,-0.0842999369,0.3529605865,0.005246751,0.5990450978,-0.0411307588,0.05516655,0.0090629691,-0.080049701,0.2823019326,-0.0525373705,-0.1047084928,0.162681073,-0.5495138764,0.0248671025,-0.005983158,0.1834041476,0.2655077577,0.4027745724,0.0387204066,-0.1163020805,-0.2263501734,-0.0094019566,-0.0039071543,0.300869137,0.1394992024,0.1594572216,0.2239145637,0.1344506443,0.0951100215,0.0431366004,-0.1842562854,-0.1783154607,-0.0636249706,0.0801098347,0.1682343632,-0.2972934246,-0.1405249536,-0.0837188363,-0.0799831524,-0.1666853875,0.1680999696,0.2521229684,-0.222952202,0.3372999728,0.0686314031,0.111228697,0.0157314986,-0.2440187335,-0.1695141792,-0.1212179884,-0.2854925692,-0.0604430065,0.2744226158,-0.2481134385,0.463131845,-0.0921317413,-0.1814789921,-0.1822262406,-0.1476952881,0.1697592288,-0.0256858245,0.1520397216,-0.0257621091,0.0911136121,0.1403442919,-0.0117276981,-0.1172379181,0.2947634459,-0.3946950734,-0.2646890283,-0.0457637869,0.1853075027,-0.1605921239,-0.3530971408,0.0199411251,-0.390781939,-0.0066498616,-0.1254160404,-0.067285046,-0.2273852974,0.0255347565,-0.2347317487,0.4058351517,0.3372837305,-0.1800175905,-0.050451234,-0.2473857701,-0.1954921484,0.0923826844,0.2312322855,-0.153695479,0.1757413,-0.4672147334,-0.0179841816,0.6156256199,-0.5054792762,-0.4822386801,0.3668369651,-0.1643920392,-0.093921721,-0.0435516275,0.2854287922,-0.0775715262,0.0263480246,0.0611944646,0.351580292,0.1483877003,0.1347034127,-0.2264336348,-0.3120892644,0.077529259,0.2374334335,-0.1020624042,0.2401536405,-0.0833321661,-0.0964250043,0.6564130783,-0.0033448022,-0.039024163,0.1771044135,0.3496188223,0.0461644456,-0.0561558343,0.1321394145,-0.2743854821,0.2317928523,-0.0811063424,0.0774697065,-0.136914283,0.1522574127,-0.4671771824,-0.0133984918,-0.3502943516,-0.2138036638,0.0977900699,0.3238503039,0.2103271782,0.1386871189,-0.1924282759,0.4510427415,-0.514996171,0.0892236009,-0.4599666297,0.3872051835,0.051192265,0.026924463,0.1402985007,0.222624138,0.0464308262,-0.0905627906,-0.2875969112,0.4172315001,-0.075165838,-0.1176199242,0.3687200546,-0.0515867397,0.0299848393,-0.4615370333,0.2841538191,0.4072540998,-0.0159509163,0.0830567777,0.040477924,0.1762773097,-0.2168430835,-0.030602742,0.0199314468,0.2044681162,0.449331969,-0.2942862809,-0.1360215843,-0.1536149085,0.1517793685,0.3045043945,0.5157181025,0.0725361332,-0.0740447342,-0.138126269,0.3581745625,-0.0692819282,-0.1495406777,0.3571566343,-0.1663306355,-0.0016493857,-0.0482868962,0.5162200332,0.3303979039,0.0800945088,-0.167221114,0.0157247148,-0.0763310641,-0.2316537499,0.1646535546,-0.1718080044,0.1088634431,0.1868586987,-0.069393374,-0.1027749553,-0.2119053304,-0.2347465605,0.0774969235,0.3297291696,-0.3639899194,-0.2274917513,-0.2603933513,-0.6383991241,-0.486920625,-0.1401940286,-0.2508092523,-0.4705287814,-0.2021902204,0.0587797128,0.2062979788,0.3153224885,-0.3141490817,0.0116963238,0.0457248539,-0.3295633495,0.1589161307,-0.1446891129,-0.2713507712,0.1072728038,0.3671357036,0.1439752281,0.100624457,-0.4280472994,0.2426960021,-0.630448103,-0.2325656563,-0.0655870587,-0.0149605162,0.1261070669,0.0270945895,0.2817869186,0.2949373126,-0.1686242223,0.2453976125,-0.0050110687,-0.1012967229,0.1765973568,-0.0970524773,-0.1183386296,0.1169453189,-0.4073301852,-0.2895722091,-0.2114052922,-0.0537784956,-0.0920130759,0.1696119606,0.1500200331,0.1081298739,0.0406973921,-0.0460923538,-0.0529094115,0.0463920347,-0.514942348,0.346663177,0.0684273541,-0.3327194452,-0.027035756,0.0320866592,0.3410561383,-0.0058479393,-0.5295609236,0.0738169402,-0.1447530538,0.3154476583,0.2459883392,-0.0960092843,0.1120342389,-0.1378752589,-0.0642815679,-0.0147826318,-0.1439587176,0.0396769084,0.0106241666,0.6999570727,0.2855210006,0.2176012397,-0.1492902339,0.0831896216,0.5553334951,-0.0592046306,0.2632457614,-0.0163616259,0.1333655417,0.1375619173,-0.3414487839,0.0171845183,-0.0367083661,-0.1097874194,-0.0237469561,0.1147009656,0.0748707578,-0.1925242096,0.0020041561,-0.2976507246,-0.2230404466,-0.0241424162,-0.2135828882,0.282389909,0.0843419954,-0.0323964991,-0.1558046937,-0.2728619277,-0.0168794263,0.4304164052,0.0510933399,-0.0669245273,-0.4378406107,-0.1503817737,-0.2656535208,0.1857946366,0.0449981391,0.7549546957,0.0679737478,-0.0616988242,0.0239417031,-0.036279168,0.4664522409,-0.193282038,0.2973357737,0.1407938302,0.3209072649,-0.233184576,-0.0348829627,0.0646288246,-0.1848784238,0.7279908657,0.2987177372,0.0721627027,-0.1937947571,0.0732216313,0.2993072569,-0.257579565,-0.2874571681,-0.4475958943,-0.2088444978,-0.2936892807,-0.1837622523,-0.3898381591,0.1613376886,-0.1003872752,-0.0517880507,0.1099185869,-0.2723607123,-0.0352507345,0.0367544666,0.2607632577,0.1131936535,0.0107165892,-0.1164238006,0.2308255732,-0.0326765776,0.5089076161,-0.1746757925,-0.245076403,-0.0136891045,-0.1131558567,0.1418546587,0.0185816772,-0.0477691218,-0.1855541915,0.1721397042,-0.0042621116,0.0221608374,-0.0433779061,0.2097509503,0.0897536948,-0.3601856232,-0.282235533,0.4302022755,0.1393752098,-0.0578979999,0.5784233809,-0.117174387,-0.0944814011,-0.0784033537,0.0654617846,0.6112681031,0.0496003032,-0.0541409962,0.3504298031,-0.1887284815,0.3375973403,0.1458977312,0.0347135179,-0.4030075967,-0.0592873171,0.0196089186,-0.2003259659,0.0101816738,0.0027694099,-0.0356467776,0.2889032066,-0.084027417,-0.0155217834,0.3707637787,0.501766324,0.0225259773,0.0111113703,-0.3719960153,0.1639682353,-0.0784762278,0.3671861887,-0.0763884783,-0.0838334337,-0.2551733851,0.0709594637,-0.2950749695,0.5404978395,0.2757504284,0.3389620185,0.0648258403,-0.12801449,-0.2496617585,0.3878884315,0.4305009544,0.0638912693,-0.275644362,0.2003508061,0.0786878243,0.1224249154,0.089209415,0.1224875376,0.2299338281,-0.1404502094,-0.4328583777,0.2280066609,-0.0215889737,-0.5032142997,-0.0519705229,-0.1553738415,-0.2494817674,-0.2342972755,-0.0343453735,0.0702423602,0.0576719195,-0.1293161958,0.1366470009,0.2121319324,0.0934360325,0.0752327442,0.1593122929,-0.2969798148,0.062420994,0.3193431795,-0.0685607791,0.0689793229,0.4445504844,0.0333652236,0.156944409,-0.1756383926,0.0826241896,-0.1344573051,-0.3435791135,-0.2444276214,-0.0872385129,0.3326417208,0.088528268,0.2776812911,0.2276193649,0.3206563294,-0.0000836218,-0.6334676147,-0.303922087,0.0985505283,-0.2064141035,0.0633975416,-0.3180437982,0.3124531209,0.2417284697,-0.2426433712,-0.2966819406,0.0722850189,-0.3123780489,-0.071740374,0.2395819575,0.0292003918,0.090396814,-0.0425592884,0.2007019818,-0.2257982343,-0.1257001609,-0.1602731794,-0.0246867519,0.0996063054,0.1776626557,-0.202981934,-0.1229582429,-0.2508124113,-0.0240061302,0.0717611164,-0.3325246871,-0.1658546478,-0.0307434276,0.2758225203,-0.1615842432,-0.0562870093,-0.038375672,0.1534134299,-0.1991807818,0.0390923098,0.155216977,0.281017065,-0.0904579461,0.0632269904,-0.477396369,0.0386107527,-0.1450752169,0.0427148193,0.2743893862,-0.2584930956,0.1577838361,0.0820706114,0.1168674305,0.4848008454,-0.1882724315,-0.1262040436,0.2777551711,0.2049924582,-0.2251925915,-0.0539958961,0.0980131552,0.0857037455,-0.2504562438,-0.0555801392,0.0586953908,-0.3978050053,0.4871254563,-0.0234777648,0.3694750667,-0.1294136494,0.1393547058,0.5224536061,0.0935406759,0.0208097082,0.1553806067,0.0035149187,-0.1617502272,0.250354588,-0.1632585973,0.1540075392,-0.172718361,0.4835169911,-0.1974907368,-0.172699064,0.0231973976,0.0449124724,-0.3199961782,-0.2160704285,0.0066733509,0.399944514,-0.1598161608,0.118490167,-0.2759808004,0.2097931057,-0.1217805892,-0.0070004165,-0.5526407957,-0.2350109071,0.1237560809,0.0368723013,-0.0431374311,-0.2234437168,-0.0480027609,0.1611389816,-0.3857373893,-0.757891953,0.0713072494,0.2273156494,-0.0725756139,-0.1632796675,0.2439629138,0.1988022178,0.0104247611,0.2831303179,0.1787039638,0.4121367335,0.5009096861,0.3841798902,0.0317535214,0.1298063695,-0.0685225949,0.0512293689,0.2722649872,0.2894987762,0.0480667278,0.3093041778,0.2313951254,-0.0603740327,0.0422833301,-0.1409118325,0.2453172505,0.072713308,0.6251955032,0.0614112616,0.0537467711,-0.2286472619,0.1315890104,-0.3768253326,0.1797703803,0.3087476492,0.0949520916,0.0438651443,-0.0172966961,0.0809464976,-0.2652841806,0.3914026022,0.5102474093,0.1521365345,-0.2963318825,-0.7047803402,-0.9020798206,0.0867724419,0.0547012687,-0.3410697579,0.046408549,0.0831419826,-0.2063589841,-0.0568964519,0.0139277093,0.0702682883,0.0295171235,-0.0250442587,-0.266559273,-0.0999606624,-0.104674466,0.2980384529,-0.2714779377,-0.1906729341,0.0879943147,-0.1157109812,0.0625946224,0.0176295061,-0.09320198,0.1104453504,0.3106443286,0.2195031941,0.1678663492,0.5355522633,0.0561113246,-0.2521293461,-0.4081022441,-0.1263269037,-0.098552011,0.1180329695,-0.0269521829,0.2019404471,0.0512727536,0.3044298589,-0.423625201,0.3305320144,0.0209780596,-0.3489254713,-0.2254323661,0.0212250445,0.1123480946,-0.1054457799,0.1905827224,0.1687425822,-0.0324886143,0.2855879664,-0.0658100843,-0.0228894018,0.3058545589,-0.0856062844,0.0935542434,-0.0168858655,0.1959046125,-0.3091352284,0.0666457638,-0.4109690785,0.146302253,0.4854401946,0.1082174703,-0.0762619823,0.223399058,0.0738197416,0.2739032507,-0.0148150204,-0.0121203931,0.1058727205,-0.3055033982,0.0107610868,-0.1301611066]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/699","title":"XNLI dataset is not loading ","comments":"the issue is fixed with latest release \r\n\r\n","body":"`dataset = datasets.load_dataset(path='xnli')`\r\n\r\nshowing below error \r\n```\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']\r\n```\r\n\r\nI think URL is now changed to \"https:\/\/cims.nyu.edu\/~sbowman\/xnli\/XNLI-MT-1.0.zip\"","comment_length":7,"text":"XNLI dataset is not loading \n`dataset = datasets.load_dataset(path='xnli')`\r\n\r\nshowing below error \r\n```\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']\r\n```\r\n\r\nI think URL is now changed to \"https:\/\/cims.nyu.edu\/~sbowman\/xnli\/XNLI-MT-1.0.zip\"\nthe issue is fixed with latest release \r\n\r\n","embeddings":[-0.1652062684,0.3078104556,-0.0719777122,0.1075766161,0.1345753968,-0.1903338283,0.1358870119,0.4159418046,-0.1426640898,-0.0951321945,-0.027760895,0.3722311854,0.1915638894,0.1019455791,0.0401312448,0.2788109183,0.038748242,0.1021191999,-0.1693783849,-0.0896563083,-0.4575359523,0.2308999151,-0.3699463308,0.044152692,-0.1650039554,0.1481255591,0.0880573541,0.2627589107,-0.1791169345,-0.5378155708,0.3541086018,-0.0477205925,0.2957901657,0.0565933324,-0.000109311,0.0805833563,0.5326402783,-0.0763499886,-0.4562686384,0.0316981301,-0.572711885,-0.3006376028,0.0188016053,-0.2262880951,0.1292701066,0.0173521992,0.0579939522,-0.0115622459,-0.0175031573,0.4165300727,0.2193570137,0.3823958039,0.1813428551,0.1187320054,-0.067767702,-0.1881858259,0.1244590878,0.3551398218,0.2661266625,-0.0105029028,0.0116481036,0.151140973,-0.0867596641,-0.1276048869,-0.0243225805,-0.147426188,0.2215246111,-0.0494712554,-0.0064158477,0.1203873977,0.332117945,-0.057444118,-0.4823557138,0.0820210204,0.0933560207,-0.1518306136,0.3315947056,0.1539075375,-0.3006024063,-0.0139094694,-0.3421601057,-0.1181052178,-0.1095728278,0.3462433815,0.0827415064,0.6632393003,-0.0087956861,0.0616594255,0.0214301031,-0.1097972244,0.3039150238,-0.0053871335,-0.0648238733,0.1792555004,-0.6147842407,0.0730719641,0.0505644642,0.1485415548,0.2771735191,0.4269566238,0.039081458,0.0163364392,-0.1285128295,0.043814607,-0.0745271072,0.3263207972,0.1951640844,0.1056501493,0.351210475,0.1847042143,0.0196165517,0.0024958097,-0.1620179713,-0.1966552436,-0.050215669,0.0619280376,0.1723054796,-0.2476417273,-0.2267495394,-0.0233323649,-0.073556073,-0.161790818,0.2516665161,0.3732225299,-0.1771217287,0.3023325205,0.0766479746,0.1219753996,0.0093009463,-0.1587177962,-0.1752922088,-0.1532102376,-0.2567434907,-0.063166216,0.2626055479,-0.2718550563,0.5153488517,-0.1368000209,-0.1612903625,-0.2008756995,-0.0837921649,0.077554211,-0.0055592544,0.133189261,0.0145602124,-0.0025966573,0.0718518272,-0.0072221076,-0.0910158232,0.2304249108,-0.3537057638,-0.2913222313,-0.0509357303,0.208982408,-0.1071547642,-0.2180344611,0.130789727,-0.4018658698,-0.0421145782,-0.0757014006,-0.11832425,-0.1761257201,0.0333496034,-0.1801979095,0.3947097063,0.2784049213,-0.0463645235,-0.082118839,-0.1354683936,-0.1878089756,0.0980778784,0.2005037516,-0.1321160793,0.1954370886,-0.4121010303,-0.0469686985,0.6427592635,-0.5244405866,-0.4740378857,0.3630495965,-0.0504577756,-0.0461297967,0.0254281219,0.2336148918,-0.1047653407,0.0642274469,0.0717120469,0.413803488,0.2458074838,0.1430757344,-0.2784319818,-0.3335769475,0.0373401232,0.2772501707,-0.0443936177,0.1746162623,-0.0215840545,-0.1162311286,0.5730970502,-0.0943751857,-0.0419180766,0.1305863857,0.3484081924,-0.0619368814,0.0078675747,0.0965677351,-0.2852457166,0.2265436202,-0.1295363605,0.0729999468,-0.131630972,0.1046619341,-0.4063738286,0.0326095596,-0.407994926,-0.1606801301,0.1448810101,0.3520865738,0.2110624611,0.1173777357,-0.1824099869,0.3575408757,-0.5728180408,0.0282797851,-0.4020316899,0.4245337546,0.0173414201,0.0038795327,0.1808323562,0.3252361417,0.0632347092,-0.1204888597,-0.2998119593,0.3453311324,0.0123365428,-0.1181395501,0.3638269603,-0.0567530133,-0.0296642426,-0.4274431467,0.2044515908,0.4161517322,0.0528737903,0.1057602614,0.0360035188,0.2661903799,-0.1614881903,-0.0124495039,0.1129470319,0.1203531176,0.4757416248,-0.3935986161,-0.0793851689,-0.2250277251,0.2007994652,0.3113394082,0.5478700399,0.0875683427,-0.167408064,-0.2204756737,0.3531446159,-0.0779190361,-0.079595834,0.3465532362,-0.1249015182,-0.0348315462,-0.055852633,0.484578371,0.2584535182,0.0815692544,-0.2308607846,0.0092068631,-0.0677508935,-0.214541778,0.2015321702,-0.1609097123,0.1754761487,0.2978350818,-0.0158965811,-0.0721065402,-0.2716808915,-0.3521552682,0.0770771131,0.3228310347,-0.3203577101,-0.2519187927,-0.1970744282,-0.6368752122,-0.4342008233,-0.110192351,-0.2041967511,-0.430598706,-0.1968216151,0.0759712681,0.1833988875,0.3953720927,-0.2950808406,0.1188163906,0.1504509449,-0.3178398013,0.1384288222,-0.1689960361,-0.2411791235,0.1053718701,0.3873987198,0.1350743324,0.1163083836,-0.4196836352,0.2214501351,-0.5555966496,-0.3054719865,-0.0203776415,0.054652337,0.1636812091,0.0304634906,0.2054142356,0.3402251601,-0.21178554,0.3098994195,-0.0164126866,-0.1042103618,0.167415902,-0.1008238718,-0.2035283893,0.0999489948,-0.3374554217,-0.2482861876,-0.2715961933,-0.134393692,-0.0716122538,0.2444032133,0.1916561276,0.0547086075,0.0592896715,0.0078518977,0.0136315217,0.0331883542,-0.434822619,0.3333008587,0.0581389479,-0.3860395253,-0.1018688083,-0.0353708453,0.383344233,0.0462521873,-0.4977454245,-0.030824339,-0.141589731,0.3415923119,0.2060551047,-0.0379793905,0.0579963177,-0.0789121389,-0.0884511024,-0.1016976312,-0.182782501,0.0661804229,-0.0196486916,0.6019392014,0.3503460884,0.23009637,-0.1074162498,0.0796942562,0.5629200339,-0.0856010541,0.3347431123,-0.0019795205,0.1370503008,0.0689804256,-0.2874595225,0.0412171558,0.0626139715,-0.0421500653,0.0016463434,0.1221648082,0.0798313171,-0.1781720072,-0.0184903499,-0.3015888929,-0.1898855716,-0.0631138533,-0.2164432108,0.3323167861,0.0930786878,0.007118993,-0.0782951787,-0.2623859942,-0.0094239833,0.3894610703,-0.0051748627,-0.1068841517,-0.4288728237,-0.1599216014,-0.3064721823,0.1798345447,0.1183662191,0.7898818851,0.0738975927,-0.101059638,0.0343399644,-0.0585338809,0.4581282735,-0.2872256339,0.3431702256,0.1105672941,0.2729529142,-0.2318642735,0.0647885874,0.1350179911,-0.1900219321,0.6813909411,0.4474297762,0.091034703,-0.1750723869,0.0587175526,0.2367841303,-0.2266165018,-0.2797169983,-0.4169642627,-0.1498186886,-0.3446804583,-0.2061668038,-0.4376671314,0.216854319,-0.0346146524,-0.1012668759,0.0125066005,-0.2486381531,-0.0531426221,0.1785583794,0.2914031744,0.0578597635,0.0792711899,-0.1697585583,0.2542459071,0.0267245695,0.4863105118,-0.1803604811,-0.2499850243,-0.0342615955,-0.1264213622,0.1709937006,0.0310659185,-0.0440691262,-0.163388595,0.0666763708,0.0405356549,0.0362628661,-0.0218937136,0.1529640257,0.1128033996,-0.3904330134,-0.1736035943,0.3502788842,0.054247722,-0.0271127447,0.4912928641,-0.0716575384,-0.1758960336,0.0144083444,0.0689746812,0.548458755,0.052184023,-0.0682374388,0.370490402,-0.1389340013,0.2149277925,0.2748998702,0.0398769304,-0.4101783931,-0.0543519147,0.0307185519,-0.2006665319,0.0302793421,-0.0076992204,-0.0463145524,0.2387344092,-0.0014613754,-0.0923180431,0.2938639224,0.5232002139,0.0253506191,-0.0398741849,-0.4175279438,0.1520083845,-0.0384372883,0.2745706141,-0.0583990999,-0.0724219531,-0.1936935484,-0.0283851959,-0.1732878536,0.5370412469,0.2373113036,0.3634009361,0.1421034038,-0.155518055,-0.251210928,0.3198786974,0.3755476773,0.0394073166,-0.2372034341,0.2062772661,0.0661774948,0.136397168,0.014370596,0.1800992787,0.2572447956,-0.1853321344,-0.4124382734,0.1643786877,-0.0900145173,-0.5163417459,0.0218146667,-0.1436648518,-0.2524609268,-0.1844024211,-0.0428488888,0.0188157912,-0.0491008461,-0.1930849403,0.1615733355,0.1543835104,0.0127429627,0.0813210905,0.2204843313,-0.3417276144,0.0500040613,0.2873558104,-0.0586036816,0.2185343355,0.4380055666,-0.0632230565,0.1107288823,-0.2146839797,0.1038361639,-0.098858349,-0.3196708858,-0.1674591303,-0.0207371768,0.1994189173,0.0823150724,0.383962363,0.1196190417,0.3336258531,0.029496152,-0.7458069921,-0.2727313638,0.0904877558,-0.2581298947,0.0962680355,-0.3080120683,0.2975505888,0.1578523815,-0.2161156982,-0.3137465119,0.0557735637,-0.3832300007,-0.0221263263,0.1816824377,-0.0469267666,0.1276578307,0.0132900672,0.2055689692,-0.2040094584,-0.1707394123,-0.1885140538,-0.0990260616,0.0908809602,0.2313120514,-0.2426413149,-0.2289054543,-0.1766489297,-0.019755993,-0.021458976,-0.2810612917,-0.1368478835,-0.0625501499,0.2884157002,-0.1646315008,-0.0780040771,-0.0060479767,0.1191944033,-0.2427588552,0.0181697886,0.1281769574,0.2090318352,-0.1759236306,0.086850442,-0.4298305809,-0.0034004366,-0.0677994564,0.0299839359,0.2535915077,-0.2149144262,0.07771153,0.1600028276,0.1125654504,0.4434487224,-0.2593226731,-0.1283624768,0.2621056139,0.224020049,-0.316437006,-0.093203783,-0.0248246491,-0.0275305156,-0.1356708258,-0.0546812154,0.1269209981,-0.3840124905,0.4839520454,-0.1164689288,0.3322085738,-0.1578206569,0.1145084426,0.3909533322,0.1161315739,0.0097341947,0.1117612496,0.0248170178,-0.1158675402,0.2192110419,-0.2134054154,0.2132608742,-0.1810670495,0.4345529079,-0.2127161324,-0.2137928456,0.0688529462,-0.0006934394,-0.3444710374,-0.2063694894,-0.0091895163,0.3396487832,-0.1394157112,0.1647350341,-0.3039878309,0.1474207342,-0.1442257017,-0.0836316943,-0.5544732809,-0.2305179387,0.1536713988,0.0022344063,0.0133923758,-0.2009896487,-0.0501650535,0.1978259087,-0.3533246219,-0.6947817206,0.0998798385,0.3292515874,-0.0616523065,-0.2170449197,0.3470833302,0.1661848575,0.0569744147,0.2067393214,0.1869245619,0.4446860552,0.519215107,0.3454692662,0.0137780802,0.0318365507,-0.054658249,-0.0442877673,0.3197089732,0.277659595,0.0808024257,0.3412194252,0.2260429859,-0.0718624145,0.0240201969,-0.093373321,0.231741339,0.0468514599,0.552989006,0.0451156124,0.0906263515,-0.1790807098,0.0865867063,-0.382640034,0.1628147364,0.2970855534,0.0860101208,0.1280750334,-0.0488247424,0.0915899277,-0.3019304872,0.4267116487,0.4726709127,0.0578404367,-0.2677156329,-0.6913499236,-0.8452142477,0.0982558504,0.0744490698,-0.3610561192,0.0699459612,0.0575078689,-0.3055672646,-0.1219892576,0.0002658217,0.037761528,0.0732541606,-0.1401530206,-0.3526489735,-0.0274445117,-0.0775761157,0.2200601399,-0.269798398,-0.3462952077,0.0953406096,-0.1271423548,0.1135565862,0.0286789406,-0.0613721758,0.0526691973,0.341129303,0.2710794508,0.1565523893,0.5291492939,-0.05415361,-0.1534776986,-0.4015060663,-0.1735394299,0.0137491738,0.061041899,0.0429229699,0.2143645436,0.0536125898,0.2884124815,-0.3946288526,0.3613432944,-0.0351518132,-0.368375957,-0.0824803188,0.0617433228,0.092276901,-0.0690653995,0.1938026249,0.1601412594,-0.0056817066,0.2154659629,-0.1285752356,-0.1044401824,0.2839360535,-0.1597376913,0.0158734228,-0.0604281574,0.1922597885,-0.3329337239,0.0406481512,-0.4265522361,0.1731559932,0.4620506465,0.1421972513,-0.0997513533,0.2333021462,0.0355379656,0.2758001089,-0.0163486935,-0.0943498835,0.1151054204,-0.2784427404,0.0145692993,-0.1046153978]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/690","title":"XNLI dataset: NonMatchingChecksumError","comments":"Thanks for reporting.\r\nThe data file must have been updated by the host.\r\nI'll update the checksum with the new one.","body":"Hi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!","comment_length":21,"text":"XNLI dataset: NonMatchingChecksumError\nHi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!\nThanks for reporting.\r\nThe data file must have been updated by the host.\r\nI'll update the checksum with the new one.","embeddings":[-0.2586137354,0.2562108636,0.0357761383,0.1724495739,-0.0067088124,-0.0366752595,0.0808679238,0.4067072868,0.175513044,0.2336782813,-0.2135459036,0.3052044809,-0.0187664814,-0.0720469877,-0.2150234878,0.5297763944,0.0938087702,0.2160157412,0.0079976609,0.0222377647,-0.2151079029,0.1584314406,-0.2035468668,-0.3193615377,-0.2458490431,0.1520858407,0.1421747804,-0.0535499007,-0.2479522675,-0.2221305817,0.6901740432,0.1887718588,0.258135736,0.0462866314,-0.0001240973,0.0080530262,0.2406711131,-0.1988253295,-0.4517176449,-0.1169145778,-0.8176319003,-0.3818692267,0.0475631207,-0.2078416198,0.1094296798,0.4795917571,0.2667313814,0.1372768134,0.1055045128,0.2807789743,0.111543946,0.420984,0.1347889304,0.143436417,0.2464423776,-0.0521562845,-0.1362707615,0.4895199239,0.4230527282,0.320353359,0.0393180661,0.1135347262,-0.3681002855,0.3269022107,-0.0359629355,-0.0250639338,0.0205162838,-0.2464183569,0.2830494046,0.4693900645,0.2261938304,-0.2822745442,-0.2415455878,-0.0743846595,0.0434021018,-0.2809100449,0.2798109949,0.08360897,-0.3455017507,-0.1063984632,-0.3278510571,-0.0058591957,-0.0017170587,0.1073263437,0.1493442059,0.5423236489,-0.0477058478,-0.0237834807,0.0697466657,-0.1552949101,0.628156364,-0.2680093348,-0.1252770722,0.1798190773,-0.4900614619,-0.0731538981,-0.2064715475,0.5996738672,0.4066444039,0.2380990088,0.2245099545,0.0259249248,-0.3553993702,0.1254947037,-0.2012140751,0.2479915321,0.099644132,0.0318743624,0.2320785224,0.3483143151,-0.0973198339,0.1221057847,-0.2033595145,0.0350357629,0.2290040255,0.2580448091,0.3275788426,-0.5910425782,-0.3378924727,0.0049089938,-0.2603462338,-0.1746095568,0.1753950864,0.1753493994,-0.1699453145,0.1092463359,-0.1756642014,-0.1288086921,-0.0445124619,-0.3918737769,-0.0676398054,-0.0543278456,-0.0956366882,0.1412098557,0.3571409285,-0.2953674197,0.3991790414,-0.2256494612,0.1817737669,0.0179107245,0.057620544,-0.1116041392,-0.0308814608,0.2910017371,-0.0173147395,0.0595719181,-0.0529284775,-0.140993461,-0.1972710192,0.2363901138,-0.3122947812,-0.2167454213,0.0996265188,0.0400579087,-0.5102123618,-0.3753417432,0.0736734644,-0.465560317,0.1015564576,-0.3904392719,-0.0073960521,-0.3887443542,-0.0695823655,-0.1876173466,0.1607968509,0.3164153397,-0.2147251368,0.0343073457,-0.1672418565,0.0257880948,0.4009526074,0.3156618774,0.0544878095,-0.1665255874,-0.2950071394,-0.2161164731,0.3075037897,-0.2204373926,-0.7255297899,0.1216126904,0.0719884932,0.3229758441,0.1239928231,0.2812179029,0.211918965,-0.0167510137,-0.0355769433,0.0872988254,-0.0745189488,-0.0421463996,-0.2804175019,-0.3618970811,0.1204608902,0.141869247,0.1601865441,0.1856274009,-0.0315918252,-0.0567067042,0.4973293841,-0.1714521796,0.0485826321,-0.0227915589,0.6201295853,-0.1340786219,-0.1849189103,-0.0258624982,-0.3418603837,0.3134120703,0.1858926266,0.0887587965,-0.1150415316,0.0676701441,-0.2243721485,-0.1555407196,-0.2370499223,0.0643190593,-0.024208568,0.3354491889,0.2417681813,0.1224176735,-0.0075558894,0.4944452047,-0.3668858409,0.092463173,-0.386790961,0.3697480559,0.0600519702,-0.0740127936,0.0644229129,0.092517592,-0.08177232,-0.0969615579,-0.2909850478,0.1332344562,0.1913929582,-0.2294040024,0.0671633855,0.2875173688,0.0111955991,-0.1594452411,0.1004255265,0.5777589083,0.1238051131,0.0808635429,-0.0785013288,0.4863092303,-0.1833893359,-0.0225315895,-0.1332876831,0.1746011525,0.3575408161,-0.2149170637,-0.1466772556,-0.1746447533,0.3238839507,0.0543514192,0.1948827505,0.111102879,-0.2262816578,0.0082300305,0.3536303937,0.0136982296,-0.1000851393,0.1387248486,-0.0055579841,-0.0261584017,0.2044367939,0.6320056915,0.380002141,0.0317563005,-0.0440924838,0.0113440827,-0.1384483576,-0.2532644272,-0.0992301106,-0.1576065868,0.2100726664,0.2351075262,0.1551225632,-0.1404782832,-0.3958390355,-0.4012742043,-0.0733323544,0.4360639751,-0.2105490565,-0.1275674552,-0.316608876,-0.5170448422,-0.4883892238,-0.2116082311,-0.2422738224,-0.34609285,-0.1825844944,0.3951963484,0.2581498921,0.2322383672,-0.3695223927,0.0123253996,-0.0546117499,-0.3932121396,0.1058228016,-0.155129239,-0.2283204198,-0.0100653842,0.3551742136,0.0392791517,0.253167659,-0.4118210077,0.1813425124,-0.4413237274,-0.1746168733,0.0173400808,-0.1531530917,-0.041174721,0.0783495978,0.2060880214,-0.2134456187,-0.3065015078,0.0666823983,-0.072919488,-0.1403520554,0.1195373833,-0.2479639947,-0.079216063,0.1542877406,0.0388552472,-0.2466491163,-0.0903133154,-0.1552611142,0.0761152282,0.220612824,0.2278158516,-0.0476889312,0.0238214415,0.0273994859,0.0180009548,0.0116397804,-0.6506959796,0.4857341945,0.1974257678,-0.2535623312,0.1648096591,-0.1050902382,0.2708635032,0.2326164544,-0.4453786314,-0.1136724725,-0.2259542197,0.3107627332,0.4475627542,-0.2022423893,0.0723847002,-0.0949512944,0.1288635284,-0.0753970221,-0.3394464552,0.0084449099,0.0520526543,0.7311505675,0.1499949992,0.1265174747,-0.1364470571,0.2031770051,0.417519927,-0.1546080858,0.2104763687,0.0111747365,0.1802267283,0.0195404235,-0.358356595,0.1440388411,0.0228398163,-0.3536070883,-0.0504716001,-0.0518312193,-0.001400759,-0.2670251429,-0.0106402002,-0.129398182,-0.3412097394,-0.0794882029,-0.0831220224,0.2536192536,0.2116161138,0.1231754869,-0.2357359976,-0.6318883896,-0.1708341241,0.4474797249,-0.1235526055,-0.0586419515,0.0527076796,-0.1895701885,-0.3307130933,0.4474558532,0.2980200052,0.6073165536,0.1178284287,0.0389348008,0.0599629506,-0.0782375932,0.0700902641,-0.0893066823,0.2266210467,0.0762593225,0.0211505592,-0.1093896106,-0.1656817645,-0.16470173,-0.1745555103,0.565957725,0.0911581367,-0.0176951047,0.1189156473,0.2135758251,0.4204087853,-0.2359035611,-0.0628216788,-0.3771163225,-0.2072806358,-0.3647631407,-0.118063733,-0.4754371643,0.0615997203,-0.0455348082,0.029505346,0.1479495168,-0.0657039657,0.3446292281,0.2249271125,0.1544180363,0.4531782568,0.2331416011,-0.1674919128,0.1083003283,0.2188342512,0.5179356337,-0.1935895383,-0.3603799343,-0.1526427567,-0.1453351378,-0.1083899066,0.1207606345,0.018649267,-0.0829491988,0.1656395495,-0.1359833032,0.2692396045,0.0904407874,-0.0642674267,0.1900078207,-0.5104939938,-0.1850135177,0.1662018746,0.1732836813,-0.0686925203,0.3116128147,-0.1918081939,0.0097127147,0.038213525,0.3153886795,0.8688945174,-0.0218935404,-0.0354281962,0.2181995958,-0.1299487203,0.5052806735,0.2657540739,0.210214451,-0.3559130728,-0.3961741328,-0.1350573897,-0.1732967049,0.0568499155,-0.1710415035,-0.2292763293,0.3527696133,0.2268236578,0.1746855229,0.2256747633,-0.0252701994,-0.2586165667,-0.0162424278,-0.2293198258,0.1181695387,-0.3430683315,0.4754451215,-0.1209284067,-0.104993321,-0.2857403159,0.067202054,-0.3832433522,0.3143416941,0.0966804326,0.1732212603,0.2541341186,-0.2059416026,-0.2779699564,0.0363748968,0.4871706665,0.0131751504,-0.3220879436,-0.0232944153,0.327339232,0.0417026952,0.0133996718,-0.1327213943,0.2953264713,-0.0555990078,-0.1177687496,0.1068492457,0.1666829884,-0.4556831121,0.1250381768,-0.1773844659,-0.1821448952,-0.0818685517,-0.0712002069,0.0713325143,0.1577724814,-0.1774205118,0.0476816036,0.2079446018,0.0523539484,0.2391225845,0.1268899143,-0.3064699173,-0.1135168523,0.4045442343,0.1057962254,-0.0367847681,0.3865575492,-0.1316715032,-0.0423777588,-0.0913608894,0.1321873963,-0.3543288112,-0.4095630944,-0.0604042374,-0.2701902986,0.4153294861,0.0231582262,0.6284530759,0.4644913077,0.1783502996,0.0964619517,-0.5210321546,-0.3888100684,0.0315414667,0.0129365046,0.2870261073,-0.3758367598,0.2927758098,0.1397345215,-0.3371580243,-0.168060571,0.0850643218,-0.3066517711,-0.0509789214,0.0953966752,0.1309387684,0.0691095516,0.0625506267,0.0046246853,0.0459244661,-0.2008612305,-0.0415531471,-0.102226831,0.1907496303,0.2005019933,-0.1846124381,-0.2284703851,-0.1442883015,-0.1481518745,-0.1284848601,-0.1365038902,0.1499313712,-0.1924204081,0.1312949359,-0.0278377943,-0.0140909441,0.0250854343,0.1129842103,-0.1522169709,0.1837059557,0.2570561767,0.2861245275,0.0253343806,0.0861968845,0.1437249929,0.0571357086,-0.1019802839,-0.2410803437,0.3060297072,-0.2665883005,0.1472246498,0.1099907681,0.0491780303,0.4875948429,0.0205439795,-0.2028518915,0.352579236,0.0845165476,-0.3221232891,0.0614541359,0.5308297276,0.2454988062,-0.0874180794,0.2134353667,0.0358163863,-0.3072910309,0.6014170051,-0.0430191047,0.1474882662,0.013823363,-0.0004376838,0.4038957655,0.0626997203,0.1253695935,0.0189434588,-0.1403688192,0.0287218075,0.0626489446,-0.231616661,0.3849839568,-0.1694199741,0.2810196579,-0.1107492447,-0.432471633,0.233934477,-0.0175396428,-0.2075413316,-0.0420889929,0.2931669652,0.3398036361,-0.2007785887,0.1734028161,-0.5587031245,0.1731672138,-0.1941285282,0.0291710775,-0.2529180348,-0.3007824123,-0.0168841872,0.3131954074,0.039506346,0.1465025842,0.1700263917,0.2076429725,-0.3036561012,-0.4746927321,-0.0965430066,0.2057171166,-0.0628147721,-0.1299209893,0.1693889499,0.0274246857,-0.0219151657,0.0453211926,0.2925057709,0.3776884079,0.4168750644,0.2413641661,-0.0295942649,0.0176223908,0.0726837441,0.0233120266,0.4255527556,0.29023844,0.0710030049,0.3429037035,0.0622064359,-0.0108724814,0.2022940814,-0.3271122873,-0.1813562214,0.0961442515,0.3088456392,0.185976848,0.3183511496,-0.2038242817,0.0416993313,-0.3885011971,0.1510501057,0.0275715608,-0.0624467842,0.1061573997,0.1031533703,0.0027692255,-0.369287312,0.2995936871,0.4130438268,0.1857761592,-0.2702026963,-0.5286294818,-0.7571751475,0.2116253078,-0.110047102,-0.1597783864,-0.2165713012,-0.0024012816,-0.1215699241,0.0625088513,-0.0136890002,-0.0481505729,-0.0637750179,-0.0609185584,-0.2312758118,-0.2271118611,0.1864972115,-0.2454633415,-0.1708978564,-0.2581579387,0.2013615817,-0.1007344723,-0.0210813638,0.0165490359,-0.2557745874,-0.0119885169,0.2333965451,0.1644077748,0.2768930495,0.6607109308,0.0985971764,-0.1049929112,-0.2200141251,-0.0525943711,0.0062342635,0.1635087729,0.1268311739,0.4762103558,0.0781691819,0.374635756,-0.271370858,0.4960834682,0.177851975,-0.3425251544,-0.2108488679,-0.0190948509,-0.3737196922,0.1639876217,-0.06081919,0.1279097348,0.017360609,0.3779901862,-0.235192582,0.1121449247,0.5374975204,-0.1511538178,-0.080838114,-0.3420774639,0.1271229982,-0.1766867936,-0.3017485738,-0.5635141134,0.1713236719,0.2985345721,-0.0366921537,-0.2237526327,0.2323482335,-0.0314048305,0.2705062628,-0.0579745695,0.2570025623,0.1416118443,-0.3666126132,-0.1271904856,-0.0590840131]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/690","title":"XNLI dataset: NonMatchingChecksumError","comments":"Well actually it looks like the link isn't working anymore :(","body":"Hi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!","comment_length":11,"text":"XNLI dataset: NonMatchingChecksumError\nHi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!\nWell actually it looks like the link isn't working anymore :(","embeddings":[-0.255310744,0.2392414361,0.0085211527,0.1634876728,0.0067739035,-0.0439035408,0.1372827739,0.3633373976,0.171156019,0.2286223024,-0.1727592498,0.3366044164,-0.0426545069,-0.0695162863,-0.163505435,0.4813378751,0.0465704687,0.233136937,0.0313721076,0.0874956325,-0.2718009949,0.2433460653,-0.2769919038,-0.2101065218,-0.2057008594,0.2202561796,0.1495936215,-0.0134021537,-0.1404132098,-0.2130139917,0.6650242209,0.1713921875,0.1935191602,0.0153912483,-0.0001241965,0.0300010871,0.3443054557,-0.1430316865,-0.5110990405,-0.2383280098,-0.9576476216,-0.2858887911,0.0788412988,-0.2258774638,0.13251701,0.5757232904,0.3302832544,0.1163121536,0.0017030794,0.327390492,0.1154778749,0.3802640438,0.1039595306,0.1111304164,0.3024130464,-0.0210408699,-0.1393845379,0.4701579809,0.4452569187,0.3324682117,0.1383518875,0.1801856011,-0.2893587351,0.3599727452,-0.0358764529,-0.0043312144,0.0399777889,-0.2448804379,0.3192991614,0.4592480659,0.2527225316,-0.2184615284,-0.2251195461,-0.1094611138,-0.0291664042,-0.2589724064,0.3085613847,0.0491242409,-0.4267017245,-0.0589586273,-0.3267292082,-0.051597137,-0.0306037944,0.1487843841,0.1163902357,0.5532471538,-0.0624029078,-0.0311506577,0.2217631638,-0.1669586897,0.6072987318,-0.2093271911,-0.0865448415,0.1287019253,-0.3354246914,-0.1029662639,-0.1649646461,0.6590434313,0.4060464799,0.2011806518,0.2180965096,0.0843380019,-0.2829890847,0.1547482908,-0.1918350905,0.2123289853,-0.0024171688,0.0310449246,0.3131835461,0.413022697,-0.098218061,0.1040841267,-0.2532060444,-0.0069389716,0.120624207,0.201683268,0.3460606337,-0.5813544393,-0.4805167615,-0.028539801,-0.3434022665,-0.1493207663,0.2136414647,0.2512396872,-0.2500895262,0.0916138142,-0.1839031279,-0.1607674658,-0.0903996527,-0.3228450119,-0.0528683923,-0.0008747963,-0.1846466959,0.1285142601,0.3199151158,-0.1887675375,0.3772034049,-0.2923530042,0.2073148787,-0.0472774692,0.0560475625,-0.1307127923,-0.1271960288,0.2881113589,0.0012395101,0.047682859,-0.0320687741,-0.1563788652,-0.175281629,0.2566285133,-0.355835855,-0.2280403823,0.1525427252,0.0405656956,-0.5855308771,-0.3886873126,0.0549940765,-0.4413233995,0.0434583426,-0.3909614682,-0.0410101861,-0.3778672516,-0.0869192556,-0.1570095718,0.1773321331,0.3334729373,-0.1910743862,-0.0496541746,-0.2098191082,0.0813598558,0.4352088869,0.2786945701,-0.0631316602,-0.0942101628,-0.3327966332,-0.1720232815,0.4384049475,-0.2183571607,-0.6873060465,0.232118845,0.0170926675,0.3388860226,0.0787715167,0.2691991925,0.1568959653,-0.009257393,-0.006289314,0.0932135507,-0.0263806675,-0.0337954536,-0.2823987901,-0.4234699309,0.1412936449,0.1014310345,0.2260431945,0.1875007749,0.0004545489,0.0009353977,0.5099486709,-0.1901926845,0.0348421298,-0.0265240222,0.583820641,-0.132076323,-0.1593952626,-0.0692844689,-0.3015118837,0.2537908554,0.1956477314,0.084881857,-0.1827760339,0.0334301405,-0.3075610697,-0.1212127134,-0.2691375613,0.0481208488,-0.0398521088,0.3143370748,0.2146538645,0.0825666711,-0.054383073,0.4295917451,-0.2378655672,0.1057004184,-0.3177689016,0.4189566672,0.0903876424,-0.0132352449,0.0044877622,0.1666118205,-0.0513178296,-0.0697367191,-0.2662812471,0.1956916749,0.19978185,-0.2107697278,0.1168404967,0.2540125251,0.0767762959,-0.2551115453,0.0840194449,0.6038797498,0.1429791451,0.0997914225,-0.0833297074,0.4522894025,-0.154230237,0.0467138328,-0.0043509752,0.1836584955,0.3160943985,-0.2437842637,-0.1882025748,-0.1297625899,0.3522753417,0.0088688703,0.1782954633,0.1195556521,-0.2827438414,0.0746538714,0.2864148617,-0.0105606131,-0.0486978292,0.11467278,-0.0397472791,-0.0458030626,0.1884431839,0.6610341072,0.419017911,0.0447397865,-0.0262146667,-0.0454915389,-0.2191208303,-0.3118264079,-0.0609550849,-0.1028338075,0.2148247212,0.1664218307,0.1326277405,-0.1728834361,-0.3709217012,-0.4593958557,-0.076135315,0.3870403171,-0.2864486277,-0.0547578968,-0.2972511947,-0.5700561404,-0.4208690226,-0.2278121114,-0.1808784455,-0.4096906483,-0.1674777865,0.4215687513,0.2923945487,0.2115745693,-0.4384241104,0.0700787306,-0.1171517149,-0.4150042832,0.1029814482,-0.1305647045,-0.2152257115,0.0221943632,0.3165490627,0.0453118049,0.2080613375,-0.4454585612,0.1805232763,-0.4627043307,-0.1436391622,0.0466488265,-0.0403421856,-0.0938906893,0.1401262134,0.1327039897,-0.2960180342,-0.3179702461,0.1250856817,-0.0126346992,-0.0766010582,0.0581146106,-0.1703249365,-0.0415658467,0.1137043759,0.1048019677,-0.2751274705,-0.0801452249,-0.110281609,0.067289032,0.2233922929,0.2933540046,0.0222055744,0.0908023342,0.1297766566,-0.0346496291,-0.0408190526,-0.537040472,0.5248768926,0.1440218538,-0.2934426367,0.1971141547,-0.0994708389,0.2816541195,0.2665379047,-0.492307514,-0.0525426567,-0.1957925409,0.2604302764,0.4470374286,-0.1659066528,0.1023683846,-0.0626803786,0.1052886024,-0.0871440247,-0.2887802124,0.0141040394,-0.0328743979,0.8322196007,0.16430372,0.1732864976,-0.0846164897,0.2370248437,0.3089038432,-0.1095726714,0.1927365065,-0.0685611367,0.2631010413,0.0845108479,-0.3629761934,0.1237300262,0.0215564426,-0.400944978,-0.0619616248,0.0108650215,-0.0468338393,-0.3123826683,0.0720107555,-0.0961806029,-0.3898316622,-0.122696802,-0.1184759364,0.1612131894,0.2262888104,0.1595988125,-0.2336450964,-0.6007033587,-0.1205367148,0.4900304675,-0.0762343034,-0.1148472056,0.0439735539,-0.2405601144,-0.3251512945,0.4079020023,0.3307505548,0.5425952673,0.060883671,0.0197302755,0.0806670561,-0.1190860942,0.1719765514,-0.081223011,0.1852892935,0.0631923303,0.0568976589,-0.1652707607,-0.1066547334,-0.0800202563,-0.084153004,0.5791826844,0.0523072071,0.0119999824,0.1365017444,0.3013047874,0.3168735802,-0.243986696,-0.0789486095,-0.3897927701,-0.2632984817,-0.3350715637,-0.1861327738,-0.4157499969,0.1151658073,-0.0033972808,-0.0390392169,0.1234101355,-0.153450653,0.374909997,0.2465710938,0.1699961424,0.4374325871,0.138075158,-0.1276906878,0.0963174775,0.2485939413,0.4512087107,-0.2059950083,-0.3352600932,-0.0543575026,-0.0971271545,-0.1294324249,0.1508696675,-0.0390126891,-0.1483182311,0.1782423407,-0.1310148239,0.3097861707,0.2154785991,-0.0746808574,0.21454449,-0.5271897316,-0.223398,0.1697745472,0.168580398,-0.0907169133,0.3737095892,-0.1777195781,-0.0577123836,-0.0817419216,0.2467073053,0.8652727604,0.0340811908,0.0265818611,0.2244563997,-0.0645353124,0.5286777616,0.2552682459,0.2164635807,-0.2638687193,-0.2813645303,-0.1403950155,-0.2654835284,0.0218896773,-0.1841464639,-0.1950537115,0.3138683736,0.2048721164,0.2350251675,0.2923361659,-0.0678566173,-0.2367922962,-0.0023196712,-0.2189317048,0.1229743734,-0.3019004762,0.5433884859,-0.1382317841,-0.0866944268,-0.3204324841,0.0957007408,-0.3543671668,0.3735297918,-0.0369929932,0.2108268589,0.1992673278,-0.2351453155,-0.2929539084,0.0551099814,0.4296872318,0.0156552903,-0.3775629401,-0.002322407,0.2431229651,0.0153148063,-0.0058364999,-0.1281497926,0.2512226105,-0.077915661,-0.0937398821,0.1966138184,0.1873093694,-0.5267367363,0.1964974105,-0.1897227913,-0.2113306671,-0.0434712917,-0.1286285669,0.0980362371,0.151769951,-0.2232264727,0.0336726904,0.2163868845,0.0842665285,0.175550729,0.1365674436,-0.3069829941,-0.1173517779,0.4108388424,0.020483518,-0.1196061298,0.2680166066,-0.1812877357,-0.1173817813,-0.0703620762,0.1895038486,-0.3482924402,-0.3630986214,0.0115461005,-0.359129101,0.3595589399,-0.0845240802,0.633433342,0.4420131445,0.2548853457,0.119230181,-0.5736899972,-0.3946498036,0.0735856295,0.0262836628,0.2805681229,-0.4309362769,0.2529241145,0.206659928,-0.4923039973,-0.1585398912,0.1004528999,-0.313103348,-0.1686593443,0.1005461887,0.1543332785,0.2331245095,0.0552469231,-0.027317645,0.0585817769,-0.1575045437,-0.0232928675,-0.1145900264,0.1947147846,0.216373831,-0.2415506393,-0.2482116818,-0.2061857879,-0.0886033401,-0.0780795589,-0.1512363702,0.1631639302,-0.185152784,0.0159288459,-0.0071725245,0.0016378372,0.0148268146,0.1262827069,-0.0983736068,0.3528677821,0.3310589194,0.2575093508,0.0463281572,0.096727632,0.1049781516,0.0362714082,-0.0587519929,-0.3040106595,0.3055242002,-0.3151383996,0.1117161438,0.0998762622,0.0645807311,0.4636318088,-0.0058921091,-0.1869293749,0.3600109518,0.0739226788,-0.2547706068,0.0994829684,0.4736136198,0.2084587365,-0.0927349851,0.2440030426,-0.0198348407,-0.2561646104,0.5993666649,-0.0337541923,0.2367288619,0.0651679859,-0.0064962995,0.4156466424,-0.0183228068,0.2096235305,0.0408801213,-0.0386385508,0.0911413655,0.0527115315,-0.3368493617,0.3105288744,-0.1109210402,0.3148570061,-0.1311902404,-0.2744791508,0.1205670238,0.0087699611,-0.1982615441,-0.0979734957,0.2826307714,0.3082252741,-0.2427444458,0.1804671437,-0.4344666302,0.1950188279,-0.1541193873,-0.0523243211,-0.3147389591,-0.3634771407,-0.0619343631,0.2429175675,0.0398818813,0.0609067157,0.1592306793,0.1471021473,-0.1875926256,-0.4107559323,-0.0372618921,0.2140091956,-0.0378823765,-0.1662287414,0.1053919718,-0.0538168997,0.0309428908,0.0425254479,0.2354684472,0.3828666508,0.368296504,0.1695964485,-0.1808357835,-0.0332278796,0.0626356676,-0.0205639545,0.3779393733,0.3738494813,0.0881867632,0.3433549702,0.0519359894,-0.0165604912,0.3092062771,-0.2958473861,-0.1162963584,0.0710616186,0.3299096525,0.0922407955,0.3065173626,-0.1619854271,0.0194590334,-0.3695020378,0.1287274361,-0.0074387803,-0.148357898,0.098257862,0.1055385992,-0.001338098,-0.2666970789,0.2667632997,0.4087016881,0.2720141709,-0.3113087118,-0.5257842541,-0.7715191841,0.2454327047,-0.1790910661,-0.1963342279,-0.2004187852,-0.0155204646,-0.2246599495,0.0646358579,-0.0255499277,0.0824032128,0.0000370311,-0.133960709,-0.1492927969,-0.2211563736,0.1960512251,-0.1825088859,-0.1785595268,-0.2523192763,0.1705125272,-0.1418397427,-0.0570616834,-0.0162363276,-0.2760982811,-0.1398755908,0.2364538014,0.1566121876,0.2343970835,0.6360192299,0.1286464781,-0.0918278992,-0.1987361461,-0.1054351702,0.0819835216,0.14933227,0.1141709387,0.3489145041,0.1132276505,0.3968115151,-0.3097335696,0.4353665709,0.2154790312,-0.3155360818,-0.1793326586,-0.0432311334,-0.3783478737,0.1850687712,-0.0316382609,0.1115725636,-0.0868502855,0.350279659,-0.180418551,0.0959245414,0.5883145332,-0.2307970226,-0.1297885627,-0.3142844141,0.2153085619,-0.1349747628,-0.3337156177,-0.5340923667,0.1075514778,0.3082754612,0.043045152,-0.2346515656,0.1928321421,-0.0446773358,0.2590782046,-0.0563054979,0.2505431473,0.1826297492,-0.3517955542,-0.129405424,-0.1041338742]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/690","title":"XNLI dataset: NonMatchingChecksumError","comments":"The new link is https:\/\/cims.nyu.edu\/~sbowman\/xnli\/XNLI-1.0.zip\r\nI'll update the dataset script","body":"Hi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!","comment_length":10,"text":"XNLI dataset: NonMatchingChecksumError\nHi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!\nThe new link is https:\/\/cims.nyu.edu\/~sbowman\/xnli\/XNLI-1.0.zip\r\nI'll update the dataset script","embeddings":[-0.3160946071,0.2545181513,0.0077860816,0.1722998023,0.0245504249,-0.0274721552,0.0829105601,0.3733117878,0.1383772641,0.2381430119,-0.160176158,0.3349710703,-0.0523419976,-0.0138887251,-0.2367373407,0.4848173857,0.0543490835,0.2295569777,-0.0119152199,0.04875594,-0.2844641209,0.2078164667,-0.2566102743,-0.3055530488,-0.2413582951,0.1408674866,0.1714137197,-0.0477383696,-0.2128823549,-0.2108860761,0.6704602838,0.1974633485,0.2794670761,0.0191521607,-0.0001240097,-0.003953591,0.2910263538,-0.1783125103,-0.4625137448,-0.2218521386,-0.8915368319,-0.3380475938,0.0831184015,-0.2290149331,0.1075151563,0.5386987925,0.3375046551,0.0708507225,0.0276175588,0.2684116364,0.1171906888,0.375667423,0.1043157056,0.1244531944,0.2789602876,-0.0545767508,-0.1612804085,0.3962781727,0.4237611294,0.2930906415,0.0608726069,0.1234650314,-0.3069435656,0.368691802,-0.0298285484,-0.0053281193,0.0213466566,-0.2856452167,0.3026607037,0.4251622856,0.2880077362,-0.2810034752,-0.2300743163,-0.0863995403,0.026348222,-0.2540665865,0.2726040483,0.1079102904,-0.3731226325,-0.1152871847,-0.3757872581,0.0140273152,-0.0409932882,0.1373051703,0.1632848382,0.5331887603,-0.0488443337,-0.0279885624,0.1708679348,-0.1853105277,0.6401736736,-0.2738181055,-0.1104228795,0.1589082927,-0.4452862442,-0.0866896659,-0.2309544384,0.6810389757,0.432706207,0.2793370187,0.1991811246,0.0499440096,-0.2715555727,0.1212544665,-0.1957128644,0.2773586214,0.0934230313,0.0368506685,0.2612954378,0.3615629077,-0.0771685392,0.1345850527,-0.2460973561,0.0053000604,0.2411555201,0.2529916167,0.3403802216,-0.5845649838,-0.419871062,-0.0255148448,-0.2608847618,-0.1788255423,0.1848389804,0.2305761129,-0.2177217752,0.1147004738,-0.1794048995,-0.1120341271,-0.047775086,-0.3324575126,-0.071427837,-0.0286183041,-0.1149005145,0.1308418512,0.360815376,-0.2371824384,0.4096808732,-0.2569949031,0.1652428657,0.0368049964,0.1026336104,-0.1463655531,-0.1097729653,0.2847666442,0.0376389064,0.0673509091,-0.0382908843,-0.1384540647,-0.181839332,0.2530845702,-0.2947968245,-0.179777503,0.1321018636,0.0559484586,-0.5604270697,-0.4045595229,0.0056302636,-0.4317547679,0.0922728628,-0.3287758231,-0.035398595,-0.3808316886,-0.1152077243,-0.171797514,0.1331047416,0.3164741099,-0.2183650434,0.0033139964,-0.2181382328,0.0305009354,0.4607720673,0.2656942606,0.0109801153,-0.1280210167,-0.3144918084,-0.1830315441,0.3567604125,-0.2277176231,-0.7030076981,0.2038350701,0.0312499572,0.3323637247,0.1455083936,0.2839791477,0.1917799115,-0.035418544,-0.0459539071,0.1118181869,-0.1084238514,0.0025983064,-0.2682961226,-0.3987186849,0.1553615779,0.1496720761,0.193248719,0.2040158659,-0.0066342568,-0.0685168579,0.4672687054,-0.151114136,0.0307037123,-0.0210363958,0.5686207414,-0.1449807882,-0.1780773848,-0.0558709279,-0.4229870439,0.26247859,0.1851583421,0.0916152969,-0.1629800051,-0.0058566099,-0.2717241645,-0.1367194355,-0.2690127194,0.0862980038,-0.0278629586,0.328510344,0.2669617236,0.0678903088,-0.0200144071,0.4973035157,-0.2618830204,0.1180475578,-0.3793559074,0.3724340498,0.0714881644,-0.0346243382,0.062581785,0.1065299734,-0.0516875722,-0.1024950966,-0.2842988968,0.1670198888,0.215520978,-0.2492003441,0.1226647645,0.3015940189,0.0181532018,-0.1910425127,0.0906570926,0.5642969608,0.1486576945,0.0649174899,-0.0842781514,0.4598152637,-0.2173133194,0.0348867252,-0.0706120953,0.1615808755,0.3541851938,-0.2312299758,-0.1465158463,-0.1614536792,0.3485026658,-0.0042236922,0.17922768,0.130396083,-0.26084885,0.0330655873,0.3517879844,0.0112041198,-0.0761374012,0.1135889515,-0.0039238408,-0.0190148763,0.187965095,0.669013679,0.4091190398,0.0576087795,-0.0814408436,-0.0368625,-0.1685521752,-0.2883860469,-0.0803477019,-0.1197787002,0.2366069704,0.217600286,0.1347666383,-0.1866415441,-0.3393860459,-0.4620058239,-0.0754526481,0.4363051653,-0.2271452099,-0.1441894323,-0.31031394,-0.5587186217,-0.4521065056,-0.1745581925,-0.1719480008,-0.3676106036,-0.2170446217,0.4340645373,0.2697534263,0.2102877945,-0.4119772315,0.0043868935,-0.0679293647,-0.3930529952,0.0950845852,-0.1406474411,-0.1859848797,0.0142272357,0.3564397395,0.0460537821,0.2404073626,-0.4504795969,0.1937728822,-0.4548211694,-0.0943770111,0.0120562594,-0.0839616135,-0.059875533,0.1309684366,0.174190715,-0.2559552491,-0.3329817951,0.0792548284,-0.0278346892,-0.1136782467,0.0864965767,-0.2105521709,-0.0980559811,0.1322626919,0.0410617031,-0.2752128541,-0.1168626994,-0.1300715208,0.1038184017,0.2140036225,0.250980854,-0.0373196639,0.0194646567,0.1284779757,0.0097576585,-0.0104718208,-0.5803815722,0.5372937918,0.2047131211,-0.2719651759,0.1865013838,-0.1506414264,0.2709552646,0.2707563341,-0.4572410583,-0.0807002708,-0.2461180538,0.3194866478,0.4392978847,-0.1445919871,0.0847446993,-0.0737023577,0.1387615055,-0.0755705684,-0.3086307943,0.0157176964,-0.0358265825,0.7910445333,0.160236761,0.1724600196,-0.1426649839,0.2079091817,0.3563118279,-0.1387093961,0.197384432,-0.0231454819,0.1869047582,0.0767353326,-0.3725516498,0.1055987701,0.0439445935,-0.3328412473,-0.0409908332,-0.0515513644,-0.0116770416,-0.2747036219,0.058316581,-0.126404956,-0.3247192502,-0.1167409942,-0.0829986408,0.1847261339,0.2347031832,0.1476030201,-0.2101095021,-0.6048804522,-0.1609661877,0.4566787481,-0.0959737971,-0.0884267986,0.0352201387,-0.2085704952,-0.3613708913,0.4269649982,0.3321536183,0.567489624,0.0968966931,0.0603144206,0.0602723807,-0.0894341245,0.1560380906,-0.1272510886,0.175715521,0.0525124073,-0.0084798206,-0.115796946,-0.1243382543,-0.1250339001,-0.1224976778,0.5582368374,0.0578195862,-0.0169157907,0.120940797,0.2647824883,0.3590869904,-0.2323239148,-0.1060407981,-0.3691002131,-0.240138486,-0.3707461059,-0.1803979874,-0.4470874667,0.118883431,-0.0520292781,-0.0415108912,0.087096639,-0.1462549716,0.3776928186,0.2574685514,0.1807541251,0.4366160929,0.2415048331,-0.1516434848,0.0830794722,0.2586716413,0.4875330627,-0.1805298328,-0.3844186962,-0.0826585516,-0.1259481609,-0.1228092834,0.0439571068,-0.0385271423,-0.1200430393,0.1281431913,-0.1144559681,0.2808301151,0.175107792,-0.0885214061,0.2047661394,-0.5165196657,-0.2286902666,0.1477170289,0.2041740716,-0.0640764609,0.2843591869,-0.2465116233,-0.0394900367,-0.0328395441,0.2776961625,0.7721500993,-0.0258909445,-0.0483707562,0.1798790544,-0.0615091845,0.5034281611,0.2231923342,0.2116127908,-0.3534955084,-0.3576406837,-0.1324877441,-0.2179043889,0.0445159487,-0.1324955225,-0.2310803235,0.3744522035,0.2239195704,0.1750621349,0.2270997167,-0.0000897513,-0.2538197041,-0.0371022746,-0.1641140133,0.1223141029,-0.3624246716,0.5054822564,-0.1451924294,-0.0746110901,-0.2625241876,0.0771306977,-0.3477314413,0.3486246467,-0.0258437581,0.1962925792,0.2156538516,-0.1985754073,-0.236065194,0.082139194,0.4985495806,0.0347594693,-0.3212915957,-0.0386599563,0.3170934021,0.0081187682,-0.0099782581,-0.1119951308,0.2667431831,-0.0426660627,-0.0751726553,0.0764969364,0.183497414,-0.4562922418,0.1531820744,-0.2107769251,-0.1691779196,-0.08042299,-0.0946721062,0.0830966011,0.1585878581,-0.1855007708,0.0471146181,0.2266539633,0.0351050496,0.2503043711,0.1518213302,-0.3103519976,-0.1358781159,0.3983116746,0.092706427,-0.0821238905,0.3218188584,-0.157571435,-0.0912478864,-0.0922826156,0.1568600535,-0.2979563475,-0.3764877617,0.0118838772,-0.2807072103,0.3591167033,-0.0141971149,0.6621084809,0.437707752,0.211810112,0.1338943243,-0.5541878343,-0.4126679301,0.0333890133,0.0304437056,0.3028182983,-0.3684923947,0.2307261825,0.1426840723,-0.3950164616,-0.1623651832,0.0403645486,-0.312350899,-0.0979024693,0.1356265396,0.1179734096,0.1462132782,0.0753952116,0.010548356,0.095137924,-0.1698838621,-0.0424373038,-0.0877029076,0.1844587922,0.1810245365,-0.2313551307,-0.2567962408,-0.1628440171,-0.1254494637,-0.1294579953,-0.1340705901,0.1638919413,-0.1637013108,0.0876208171,-0.0254555345,-0.0207980573,0.0191400945,0.1048887819,-0.1677226275,0.2461791635,0.275352478,0.2777412832,0.016829377,0.0786655098,0.137841031,-0.0340982825,-0.0623640493,-0.234870404,0.2922562361,-0.3091657162,0.1134811714,0.1831169128,0.0337855406,0.4374336898,0.0241530258,-0.2217780799,0.3685865998,0.0927828476,-0.293862462,0.0926533118,0.4783470929,0.2246861458,-0.0713844374,0.2703650594,-0.024427386,-0.2754060328,0.6229617596,-0.046276968,0.2520076036,0.0623073019,0.0238296334,0.3764099777,0.0159344077,0.1333604902,0.0274039656,-0.0542654321,0.0372009277,0.0339823514,-0.3000513911,0.3647603095,-0.1667413265,0.2961978316,-0.1277741492,-0.3592664301,0.1847795695,0.0331941247,-0.215343833,-0.0497018993,0.3189953268,0.2995985746,-0.2180698216,0.1897909045,-0.4841630161,0.1491916329,-0.1803987324,0.0183361303,-0.2926737964,-0.2941568792,-0.0242084302,0.2867627442,0.0501426496,0.0923137218,0.1550454199,0.1771600246,-0.2429914773,-0.4285787046,-0.0038833374,0.2105401009,-0.0519685037,-0.1507023722,0.1538318098,-0.020317968,0.0097761238,0.0208441149,0.2510379553,0.3779487014,0.4135080278,0.2020197362,-0.1285819858,0.0079084886,0.05331669,0.0135493763,0.4109458625,0.2806444466,0.1080696955,0.3543703556,0.0611788966,-0.015105227,0.2467937171,-0.2891025543,-0.1900946945,0.1118410751,0.2855423987,0.1654729992,0.3381035924,-0.1878156215,0.0556278639,-0.3720616996,0.0943116993,-0.0087661408,-0.1257174611,0.1105828211,0.0765798837,0.0075810812,-0.3380372226,0.2609250247,0.4220189154,0.2350218445,-0.3310851753,-0.5152939558,-0.7489028573,0.2124381959,-0.118589893,-0.193276912,-0.227858007,0.0206195358,-0.1967155635,0.0710373521,-0.0318758152,0.0320886634,0.0229606666,-0.065158613,-0.1825962514,-0.2111281753,0.2068996876,-0.247140795,-0.1370172501,-0.2896077037,0.1604126394,-0.1052887589,-0.0233660135,-0.0427276678,-0.212931484,-0.078984879,0.206130743,0.1292150766,0.2298737466,0.6657446623,0.0821982548,-0.0127790477,-0.1778599769,-0.0746037513,0.0293471944,0.1487686187,0.1586524993,0.4367358088,0.1104788482,0.4121868312,-0.2985737622,0.4166417122,0.1854349226,-0.3115433753,-0.1750054061,-0.0060119438,-0.3983434737,0.1600766033,-0.0643367395,0.0901764259,0.0066874796,0.3766325116,-0.1819073111,0.064168632,0.5656695366,-0.1315070987,-0.1157854646,-0.3266015053,0.133413136,-0.1509822756,-0.3287578523,-0.5074476004,0.1606133878,0.3070746958,-0.0123147164,-0.1938316822,0.2428957522,-0.0409609377,0.2672875524,-0.0545955598,0.2144455016,0.1643067002,-0.3376363814,-0.1326330304,-0.0852813497]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/690","title":"XNLI dataset: NonMatchingChecksumError","comments":"I'll do a release in the next few days to make the fix available for everyone.\r\nIn the meantime you can load `xnli` with\r\n```\r\nxnli = load_dataset('xnli', script_version=\"master\")\r\n```\r\nThis will use the latest version of the xnli script (available on master branch), instead of the old one.","body":"Hi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!","comment_length":49,"text":"XNLI dataset: NonMatchingChecksumError\nHi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!\nI'll do a release in the next few days to make the fix available for everyone.\r\nIn the meantime you can load `xnli` with\r\n```\r\nxnli = load_dataset('xnli', script_version=\"master\")\r\n```\r\nThis will use the latest version of the xnli script (available on master branch), instead of the old one.","embeddings":[-0.3440380991,0.2931839824,-0.0001978273,0.1477116346,0.0428991877,-0.0164932776,0.0669286624,0.4335777164,0.1969800442,0.2615415752,-0.1929385513,0.3978194892,-0.0380580127,-0.0033608482,-0.2366971821,0.4933423996,0.0317254066,0.2110858262,-0.0416039452,0.0739370659,-0.282717973,0.1787003577,-0.1861887276,-0.3178001344,-0.2123045772,0.1219339147,0.1847492605,0.0075170174,-0.2586039007,-0.2615925968,0.6109445691,0.2217766941,0.2797488868,0.0617806725,-0.0001173359,0.0237155594,0.2804249227,-0.1893090457,-0.4183906615,-0.19196105,-0.801671505,-0.347548455,0.0605280735,-0.2402590811,0.0747940168,0.5231491923,0.2952467203,0.0442489609,0.079988949,0.2395127267,0.1787977368,0.4278893173,0.0840491802,0.0535929874,0.2339523435,-0.0877392218,-0.1639369875,0.4085545838,0.3534652591,0.2922393084,0.0033407037,0.1321728528,-0.2840929329,0.3042035699,-0.0183510184,-0.0000574095,0.0221796464,-0.2624893785,0.2590136826,0.4535431266,0.2009491324,-0.333573699,-0.2044590116,-0.0561783388,0.0789384097,-0.3034098446,0.2755576968,0.0850338116,-0.3896925747,-0.1392334849,-0.3486262858,0.0117965499,-0.0150103439,0.124863416,0.1671703905,0.5360557437,-0.0082516186,-0.0296936072,0.1839886904,-0.1859424859,0.5659554005,-0.162644729,-0.1520641744,0.142889142,-0.441308558,-0.0617953017,-0.1626289785,0.5207799077,0.4231540263,0.2548365295,0.1699787825,0.0411418341,-0.293474406,0.1523209065,-0.1858281642,0.2886753082,0.1287662834,0.0313783176,0.2930379212,0.3644462228,-0.0386991017,0.1364924312,-0.1992698461,-0.0421607532,0.2076930553,0.2905877233,0.2817013562,-0.5635091662,-0.3809591234,-0.0410360508,-0.2130841613,-0.1139888614,0.1676994264,0.2496082187,-0.1615517139,0.1476880163,-0.1371200681,-0.119239673,-0.0691370144,-0.2893770933,-0.095134832,0.0040906724,-0.1191838756,0.1341282129,0.3768643141,-0.2947879136,0.4392274916,-0.2404561937,0.1950042546,0.1250956059,0.0708589554,-0.091416128,-0.0768904313,0.3040601909,0.0895576328,0.0418243445,-0.0311249979,-0.1528535485,-0.210223034,0.2301728725,-0.20999448,-0.2321097702,0.0797881857,0.1087787971,-0.541649878,-0.3774971366,0.0189006906,-0.4489604235,0.1088399366,-0.3180778921,-0.0343200341,-0.3917699456,-0.1067016497,-0.2412903607,0.1894364059,0.3107250035,-0.2163632065,0.0360517763,-0.2468756735,-0.0264514033,0.3716176152,0.2849496603,0.0138590569,-0.1486024857,-0.27336362,-0.1364937723,0.288885653,-0.2784935236,-0.6970908642,0.1289833635,0.0430604964,0.3183749914,0.1046937704,0.2983999848,0.2461851388,-0.016169142,-0.0438211299,0.1752974987,-0.0970218331,-0.006965484,-0.2635580599,-0.3880696893,0.1158207282,0.1368900388,0.1502616554,0.1209083945,-0.0502936281,-0.0395643339,0.4690094888,-0.1905785948,0.0268350784,-0.052196715,0.6033701897,-0.1217801943,-0.202829048,-0.0414078906,-0.4090159833,0.3111777902,0.1476166844,0.1218377203,-0.1686831713,-0.0094153909,-0.2805188298,-0.1390558332,-0.2671610117,0.0837762728,0.0549876131,0.3394343257,0.258228451,0.0887850821,-0.0038406069,0.4365166426,-0.3119868338,0.0726900771,-0.436504513,0.3094474375,0.0062044933,-0.0633645803,0.067551516,0.1108780801,-0.1225920767,-0.1070815399,-0.2557765841,0.2078650296,0.196037814,-0.2338299453,0.0494602732,0.3397025466,0.005655495,-0.1557008624,0.0675097406,0.6223809123,0.0914903134,0.1029840633,-0.1324334443,0.5266985297,-0.1924358904,0.020005757,-0.0674268678,0.1960758567,0.4001260698,-0.1682979465,-0.1617601812,-0.2088333517,0.2940329313,0.0468100272,0.193383038,0.0589754432,-0.2583680749,0.0743772462,0.3327547908,0.007445015,-0.0336877182,0.0930739045,0.0040551829,0.011394334,0.1987383515,0.5975401998,0.3432101011,0.0955268964,-0.0616085008,-0.0286781564,-0.1906537712,-0.2502810657,-0.0580552183,-0.1220419258,0.2735538185,0.1527737528,0.131263122,-0.183420077,-0.3542267978,-0.397900492,-0.0945035443,0.4105632007,-0.2074863017,-0.1465307623,-0.3155780137,-0.5586172938,-0.447663337,-0.2198947519,-0.1585790515,-0.3854695261,-0.1467304826,0.4163065255,0.1678808182,0.2521423995,-0.375821352,0.0352790765,-0.0468054414,-0.3844075799,0.1503746063,-0.2083703578,-0.1908456534,0.0462133773,0.3689054847,0.0867272094,0.2980211079,-0.403192848,0.1827478856,-0.4075948596,-0.1981237084,0.002381505,-0.1348774433,-0.0719987825,0.1256294996,0.1721436232,-0.2058403343,-0.3575612605,0.080321677,-0.1493177116,-0.1881509125,0.1096832603,-0.2746915221,-0.1528237909,0.0912578404,-0.0726803318,-0.2556606233,-0.2009692341,-0.097683385,0.1144784465,0.2240271866,0.2649918199,-0.0191736817,0.0423420072,0.1130246073,0.0400638469,-0.0267869346,-0.6075460911,0.4921409786,0.094733119,-0.2937291265,0.1848194152,-0.1470679343,0.2552671134,0.2104963809,-0.4421551228,-0.1724352241,-0.2389871627,0.37339136,0.4332044423,-0.1727671921,0.0812958628,-0.0512844585,0.0780458078,-0.0847655237,-0.2738974392,0.0359925143,-0.0335578248,0.7024351358,0.1258622259,0.1708853096,-0.0635146201,0.2098976374,0.3908780217,-0.1221815571,0.1051872522,-0.0278415307,0.235170573,-0.0090694129,-0.3265691698,0.1183073148,-0.016558988,-0.3186417222,-0.0053003621,-0.0556301102,-0.0392621867,-0.2463173121,0.1020684391,-0.0881523117,-0.3926936686,-0.0291871335,-0.079379499,0.191177994,0.2054201514,0.1937786937,-0.2194745094,-0.5721917748,-0.0995531306,0.3870615065,-0.1190756187,-0.0579636768,0.0472556204,-0.1828984022,-0.4157749712,0.4237547815,0.3293733001,0.5484646559,0.136691615,-0.011610291,0.0345033035,-0.0985600725,0.0837162063,-0.0597552918,0.1749057919,0.1334548146,-0.0519234464,-0.1844398975,-0.1501597166,-0.1745018363,-0.1552254856,0.5905579329,0.0591124035,-0.0776579678,0.1408092082,0.2333712429,0.3763696551,-0.2462835461,-0.0807745829,-0.3972806931,-0.2627044022,-0.4146534503,-0.1544626653,-0.3829424679,0.1329501718,-0.1005192399,0.0448333994,0.1235402673,-0.1154031232,0.3189052045,0.3315773904,0.2282118052,0.4145651162,0.2811504304,-0.1363643259,0.0512514636,0.2107328176,0.5573121309,-0.1722424626,-0.403103292,-0.1164150387,-0.1506759375,-0.1288350821,0.0716042593,-0.0065048593,-0.1395279467,0.1229689941,-0.1664734781,0.2401521951,0.1917052716,-0.1167017072,0.1203276441,-0.4312667549,-0.2116885185,0.2073089331,0.2092949301,-0.0363912806,0.2715171278,-0.2344303429,-0.0248650424,0.0846186429,0.3442436755,0.8187870979,-0.0619056039,-0.0139195733,0.2896169722,-0.0557262152,0.4866748452,0.2484872788,0.2380783558,-0.4096163213,-0.396299392,-0.0390985385,-0.1366098523,-0.0193005018,-0.1677212566,-0.254160881,0.3362489641,0.1775852591,0.1237634942,0.2067863792,0.0409847759,-0.2272512913,-0.126827836,-0.1875162274,0.1844399869,-0.2838767767,0.4083459675,-0.1444783062,-0.0671343431,-0.2317021489,0.0616197921,-0.3937956393,0.3232375681,0.0329275876,0.254294306,0.1914368421,-0.1826344281,-0.2856239378,0.0671776012,0.4448699355,0.0388735197,-0.3048975468,-0.004949281,0.3574894965,-0.0184867457,0.0408534855,-0.0647566915,0.3254228532,-0.0900863856,-0.0873252153,0.0833622217,0.1733257174,-0.4621197581,0.1082622781,-0.1714569032,-0.2083178461,-0.0360819809,-0.0317527242,0.066907458,0.1646899879,-0.1567268223,0.1015163064,0.2169137895,-0.0008439182,0.2573365867,0.1036614105,-0.319843024,-0.1011667326,0.3443153501,0.0908376575,-0.0055110827,0.3339008391,-0.070347406,-0.0624952577,-0.1247319877,0.1068125665,-0.2791675925,-0.3632577956,-0.0075325612,-0.2389723957,0.3001516759,0.0185424201,0.6877378821,0.4473281205,0.1496277899,0.0859907717,-0.5489456654,-0.4318609238,0.0741936192,0.0474209376,0.309594512,-0.3513423204,0.2906151712,0.044617448,-0.3251021206,-0.2508600354,-0.0226757042,-0.3561298847,-0.0524393395,0.1040027812,0.0975899547,0.135990113,0.0881349221,0.0636298805,0.0629810393,-0.2140293121,-0.1170706749,-0.0861676261,0.1459875852,0.1237000898,-0.2211586982,-0.1613106281,-0.1197179854,-0.1620863378,-0.1573929489,-0.0734207928,0.1817919463,-0.2101595253,0.1287515014,-0.0647247508,0.001096078,0.0696657896,0.1311083138,-0.1214490235,0.2450139225,0.2138001621,0.3322269619,-0.0241162591,0.0739712566,0.1956086159,0.004383415,-0.0255718902,-0.1465316117,0.2387058139,-0.2423232496,0.1400311589,0.1473293006,0.0703817829,0.4781847,0.0161632672,-0.1533745527,0.3481903076,0.1562802047,-0.3181652725,0.0861119479,0.3944348395,0.2042919099,-0.0174992401,0.2422493696,0.0090040769,-0.2437971532,0.5908532143,-0.0485161878,0.184991315,-0.0333949365,0.0413183868,0.3584519923,-0.0264852885,0.0386489145,-0.022557579,-0.1022627205,0.0775159448,0.0320197418,-0.2205406427,0.3828904331,-0.1937123239,0.2351669669,-0.0764031038,-0.3828199804,0.1927022189,0.0023949854,-0.1918043792,-0.0293089431,0.3255550861,0.3256494999,-0.2465953827,0.1597205251,-0.5443404913,0.13140136,-0.1738945693,-0.0069844224,-0.2848836184,-0.2271642238,-0.026050562,0.3117634058,0.0437478013,0.0755051449,0.1772057712,0.1944651604,-0.3063092232,-0.4392056465,-0.0449708775,0.2036381662,-0.0761502609,-0.135276258,0.1637635678,0.0486328341,-0.0134405931,-0.0013745048,0.2801886797,0.4108684063,0.4273441732,0.1495102644,-0.0487959944,-0.0148140127,0.0695647225,-0.0541239455,0.3933201432,0.2981081605,0.1234130934,0.3966402113,0.1295002252,-0.0697939992,0.1862992644,-0.2708442211,-0.1804251224,0.2033927143,0.3322363794,0.1945767105,0.3319145739,-0.1734601259,0.0504391603,-0.3767872751,0.0688006729,-0.0245942101,-0.0797668919,0.1014868617,0.075271219,0.0506347381,-0.3402002156,0.3276660442,0.3530296087,0.2047849298,-0.2460642904,-0.4818518162,-0.7534542084,0.2511005998,-0.1440832615,-0.1838150769,-0.2389981598,0.1205141023,-0.1504046023,0.1064958572,-0.0394608788,-0.031222865,-0.0206561182,-0.082969375,-0.2197119743,-0.2052154243,0.2064459175,-0.3215513825,-0.1116758585,-0.3138567507,0.1072224528,-0.1370761544,0.049774766,-0.0054869624,-0.2207742631,0.0075335382,0.2135256827,0.1802084148,0.2631951869,0.6508324742,0.0306208581,-0.1217552125,-0.2570914626,-0.1069287509,-0.0540275946,0.1279273629,0.168135196,0.4546613991,0.0822134987,0.3446868658,-0.3246959448,0.4766316712,0.1532453448,-0.3483620286,-0.2064530998,0.0085847452,-0.4279065132,0.1680554003,-0.0878248513,0.1111263633,-0.0251438636,0.3450973928,-0.2159132808,0.0450780578,0.5796919465,-0.1642714739,-0.0891054124,-0.3459124863,0.1120372489,-0.1620555818,-0.2647881806,-0.537189424,0.1969026476,0.2010464966,-0.0259510688,-0.2404850721,0.2116144449,-0.0444227494,0.2430715561,-0.0714146793,0.255710423,0.1401874572,-0.3658737838,-0.1365923882,-0.0890742764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/690","title":"XNLI dataset: NonMatchingChecksumError","comments":"That's awesome! Thanks a lot!","body":"Hi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!","comment_length":5,"text":"XNLI dataset: NonMatchingChecksumError\nHi,\r\nI tried to download \"xnli\" dataset in colab using \r\n`xnli = load_dataset(path='xnli')`\r\nbut got 'NonMatchingChecksumError' error\r\n\r\n`NonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-27-a87bedc82eeb> in <module>()\r\n----> 1 xnli = load_dataset(path='xnli')\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip']`\r\n\r\nThe same code worked well several days ago in colab but stopped working now. Thanks!\nThat's awesome! Thanks a lot!","embeddings":[-0.2819172144,0.2050142288,0.020113837,0.1967777759,0.0314993858,-0.0379749388,0.0950002298,0.3737634718,0.1586368382,0.2620050013,-0.1807173043,0.3254222274,-0.0744880959,-0.0367187671,-0.261670053,0.4829288423,0.0819851756,0.2462064773,0.0238743816,0.0569811761,-0.2583662271,0.1832848042,-0.235320732,-0.2836458087,-0.2500883937,0.1526790261,0.1717824787,-0.0539441518,-0.2348185927,-0.1783496588,0.683586657,0.1818370968,0.2733955681,0.0273053329,-0.0001248214,-0.0015254073,0.27713871,-0.1859742999,-0.4356919527,-0.1733842194,-0.8994476795,-0.3125433326,0.1000511125,-0.2207698822,0.1262109131,0.5519015193,0.3006771803,0.0774263814,0.0458501689,0.2314354479,0.1109215841,0.3881515265,0.0722048283,0.1265586764,0.2893660367,-0.0304410718,-0.1867221296,0.4033304155,0.4341873825,0.289860934,0.0874743536,0.1355348378,-0.3305095136,0.349709779,-0.0299160406,0.0063521559,0.0597864911,-0.2803854346,0.2838531137,0.4646708667,0.2570837438,-0.2853725553,-0.2109027207,-0.1135406196,0.0141513003,-0.2952498794,0.271613121,0.10427396,-0.3947315216,-0.126157254,-0.3881402314,0.0198688824,-0.0356611423,0.1344495416,0.1497301906,0.5504695177,-0.0427105278,-0.0172089469,0.152722314,-0.1757955104,0.6034109592,-0.2551377714,-0.1213727891,0.1728352457,-0.4698452055,-0.1092989817,-0.2321287394,0.659997046,0.4288351536,0.2676359415,0.2135632634,0.0420205742,-0.2813257277,0.1358925253,-0.1788654178,0.2811609209,0.0769755691,0.0494103208,0.2594471276,0.3046509624,-0.0887630284,0.1223223805,-0.2402966768,0.0487517007,0.2361861765,0.228083849,0.3581655622,-0.6151965261,-0.4077045619,-0.0175277069,-0.3078190386,-0.1488969326,0.2004669011,0.2228905857,-0.178687498,0.1112050414,-0.1888392419,-0.0959366336,-0.0434855707,-0.3230869472,-0.0632919595,-0.0409811251,-0.1402484179,0.1328617185,0.3503641188,-0.2307439595,0.4137473404,-0.2526093125,0.2016782016,0.0497365594,0.1017175764,-0.1510148644,-0.1092969328,0.279389292,0.0114606339,0.0469383672,-0.030188527,-0.1352948844,-0.2071782351,0.2490159571,-0.2570019066,-0.204978168,0.1035054699,0.0488035791,-0.5717355609,-0.4043061137,0.0282818675,-0.4414221346,0.1302997619,-0.3304274082,0.0079753287,-0.3951217532,-0.0843497366,-0.1734269708,0.1188671514,0.2891893983,-0.2118510306,-0.0066778446,-0.2215166688,0.0308208484,0.4953330457,0.3073114157,0.0254527945,-0.1474280059,-0.2925641239,-0.1817196608,0.3495540023,-0.2221652567,-0.6940304041,0.1845265925,0.0541046336,0.3548475504,0.1475637555,0.2943892479,0.1833803654,-0.0151305832,-0.0434627719,0.1026348919,-0.1208757833,-0.0101118768,-0.2606995106,-0.394066751,0.1631025672,0.1464027613,0.1930862069,0.1979783177,-0.018915901,-0.0294226818,0.4759738147,-0.1794171333,0.0241415948,-0.0655769482,0.6058382988,-0.1317158341,-0.1949792653,-0.0788427666,-0.3825056255,0.2945936322,0.1762135178,0.0867754221,-0.1250323355,0.0169063844,-0.22107324,-0.1302383393,-0.2804357409,0.0917461291,-0.042932298,0.3247237504,0.2662253976,0.0635759458,-0.0221443661,0.4768369496,-0.3175384104,0.1133968309,-0.364677161,0.3330651522,0.092230089,-0.0635774583,-0.0159393754,0.1159998029,-0.1098002419,-0.0678048059,-0.2860008776,0.1424648762,0.2039167732,-0.257297039,0.0812238827,0.3196516037,0.0108549073,-0.1800953746,0.098941423,0.5628541708,0.1422111094,0.0428454988,-0.0697422475,0.4678102136,-0.1707297266,0.0288460758,-0.0934389532,0.1493381113,0.3307910264,-0.2131583095,-0.1675635576,-0.1583297998,0.3806863725,0.0082950871,0.1767238975,0.1274272352,-0.297685802,0.0035841707,0.363689959,0.0265910029,-0.0485007316,0.1050585583,0.0134815658,0.0074122082,0.1936909258,0.6029521823,0.4207071364,0.0526914448,-0.0816665143,-0.0295806807,-0.1756836772,-0.2641055882,-0.0856960639,-0.1251631826,0.2225050926,0.2045379132,0.1434812248,-0.1744694859,-0.3789008856,-0.452999711,-0.085193947,0.431185931,-0.2166561782,-0.1307768524,-0.3028436303,-0.560225606,-0.4269486666,-0.2269605547,-0.1904080063,-0.3415316641,-0.2066312879,0.4175875485,0.2848624587,0.2056483477,-0.4289693832,0.0284334663,-0.0769986734,-0.4025599658,0.1074434519,-0.1687632501,-0.1888879836,0.0038370495,0.3540403545,0.0251270458,0.266795963,-0.4175130725,0.1723256707,-0.4076798558,-0.1350240707,0.0188128017,-0.108944945,-0.0911804736,0.1147025377,0.165010646,-0.2605905831,-0.3402802348,0.0865534693,-0.0366633385,-0.1509548575,0.1089596823,-0.2129782885,-0.0930652767,0.119558312,0.0871195272,-0.2788080871,-0.1130304039,-0.0946236551,0.1037109271,0.2096355408,0.2764999866,-0.0497860722,0.0562750027,0.0707229227,0.0053598294,-0.0209025797,-0.6090612411,0.5074281096,0.2171162665,-0.242058903,0.1817860603,-0.1126176789,0.2447627932,0.2946217358,-0.4478031993,-0.0700772703,-0.2626539767,0.3046044111,0.4580621719,-0.170944795,0.0855474472,-0.0884440988,0.1539672613,-0.0720202848,-0.3018549979,0.0240998659,0.0061714374,0.807857275,0.143890664,0.1760789305,-0.1466365755,0.2224985957,0.3546407223,-0.1305572987,0.2199125886,-0.0091819353,0.2038232386,0.0232406463,-0.3824899197,0.1121021956,0.0233587027,-0.361235559,-0.0599076487,-0.0405819267,-0.0460137613,-0.2724870741,0.0866593048,-0.0796980411,-0.3104072511,-0.0960920602,-0.1148920059,0.1741357744,0.2099877298,0.1504523903,-0.2111333907,-0.6193137169,-0.1932988018,0.4313744605,-0.1132007316,-0.0674277022,0.0240884628,-0.2055055499,-0.3486331403,0.4259659946,0.3353696167,0.57256037,0.0841257647,0.0515261143,0.0777399987,-0.0633421391,0.1146890447,-0.0857225731,0.1472020745,0.0615910292,-0.043126978,-0.1213084683,-0.1672414392,-0.143555522,-0.1369461417,0.5619136095,0.0811771154,-0.0064347149,0.142294839,0.2485618293,0.3619640172,-0.2316131443,-0.1120342687,-0.4073703885,-0.2235094011,-0.3646560311,-0.1365823597,-0.4404166937,0.0997952968,-0.0324799567,-0.0235743281,0.0961427093,-0.1213060766,0.3884248137,0.2627144456,0.1594759822,0.4073407054,0.2169599086,-0.1510777622,0.0900661573,0.2250166833,0.4917720258,-0.1683453619,-0.410967797,-0.0848407969,-0.1315216422,-0.0782832056,0.0832860619,-0.0152734192,-0.0769759119,0.1381513029,-0.1266430914,0.2824898958,0.1624497622,-0.0763054565,0.2159348726,-0.5175997019,-0.2168612927,0.1592998505,0.2026869208,-0.0583221093,0.3092672229,-0.2321166694,-0.0502904579,-0.0040987711,0.2826381028,0.843593359,-0.0212918408,-0.0186656099,0.1947042942,-0.0569824241,0.5236573219,0.2369910777,0.2200023383,-0.3398425579,-0.3638399541,-0.1351340413,-0.2310622334,0.0615488328,-0.1584023833,-0.2274930179,0.3509737849,0.1765418798,0.1906090528,0.2440531999,-0.0522845313,-0.2763481736,-0.0428297371,-0.2079283297,0.1186391935,-0.3366180956,0.4956990778,-0.1565288156,-0.064451918,-0.295995146,0.0787355006,-0.387858063,0.3220736682,0.0333812386,0.1946880519,0.2161872387,-0.2063944638,-0.2550320029,0.0782535449,0.4921063781,0.0072505316,-0.3277010322,-0.0715426952,0.3375067115,0.0087395599,0.0085216211,-0.129801631,0.2674111128,-0.0491223596,-0.0792591646,0.1303117573,0.1846353412,-0.4929713011,0.0933643505,-0.1945062429,-0.1652070284,-0.0646852031,-0.1081096157,0.0674279258,0.1680608839,-0.1755590886,0.0383589715,0.2552807331,0.0062106028,0.2716672122,0.1261206269,-0.3180354834,-0.1384302825,0.3755318522,0.1262700111,-0.0458085388,0.3302237391,-0.1699012667,-0.0786243379,-0.0773760825,0.1589231789,-0.2853263915,-0.3685470819,-0.003196653,-0.2948693633,0.3357711434,-0.0232990086,0.6605290174,0.4636396766,0.1747219563,0.1247405037,-0.541418016,-0.4280076623,0.0475120954,0.015636798,0.3181543648,-0.3863716125,0.2644638121,0.112308979,-0.3842808604,-0.1530707926,0.0220469832,-0.3532738686,-0.0839845836,0.1285564005,0.1037898511,0.119595103,0.0607418679,-0.0019193024,0.1233513355,-0.1819916964,-0.0402813591,-0.0929000825,0.1983344406,0.2173982561,-0.2106017321,-0.2237478644,-0.1485863179,-0.1218200922,-0.1191623807,-0.1202880293,0.1875154823,-0.1658313721,0.0912135392,-0.0161976796,0.0044004717,0.0190551635,0.1144554988,-0.1512770206,0.2192663103,0.2834704518,0.2664760351,0.0335621797,0.0787959471,0.1616857499,-0.0094194114,-0.0570490621,-0.2619839609,0.2811817825,-0.3038229048,0.1540289819,0.1424184144,0.0411778949,0.4485198855,0.029904468,-0.2073407918,0.3816370368,0.0803909823,-0.2912295759,0.0865014568,0.5048671365,0.2316154838,-0.0880133733,0.2624451816,-0.0341858119,-0.2621433139,0.598043263,-0.0686085597,0.2076094896,0.0526830852,0.0039694696,0.3820592761,0.0013126496,0.1300400496,0.0267407522,-0.0465889499,0.0500793196,0.0377709158,-0.3071244955,0.3832449019,-0.1867759526,0.2938640118,-0.0980819315,-0.3809989393,0.1678915769,0.0337479189,-0.2091995329,-0.0273648351,0.3020499051,0.3033664227,-0.2376255691,0.1738168299,-0.4978576899,0.1653204113,-0.2128407806,0.0150312725,-0.2766670287,-0.2974121869,0.0005558443,0.3077133298,0.044029098,0.1072626486,0.1599031538,0.17561768,-0.245573163,-0.4360531867,-0.0118185971,0.2139164656,-0.0622008257,-0.1757478416,0.1522284001,-0.0357555076,-0.0042406586,0.006509095,0.2702673078,0.3721913099,0.4484756589,0.1966051459,-0.0833345577,-0.0004710455,0.072952047,0.0203074384,0.4219423234,0.328112483,0.1199931055,0.3458139598,0.0470918119,-0.0159148742,0.231736064,-0.3053882122,-0.1808155477,0.0903594792,0.2880550921,0.1603920162,0.343225807,-0.175434798,0.0745824501,-0.3838873804,0.1117257774,-0.0103745153,-0.0846130401,0.1026891097,0.1028400809,-0.0042359135,-0.3401915431,0.268737793,0.4247135818,0.2476124614,-0.3216385245,-0.4812085927,-0.7580248713,0.2354024798,-0.1540753096,-0.1620262563,-0.2018033713,0.0319935605,-0.1501042545,0.0833065435,-0.0219451096,0.03021558,-0.0156061621,-0.0845616385,-0.1804195642,-0.1993200928,0.2254624218,-0.2760331929,-0.1444019526,-0.3023146093,0.1853452474,-0.1060482115,-0.0287240893,-0.0378244705,-0.2126424909,-0.0635991767,0.2427360266,0.1121840626,0.2488216162,0.6962276697,0.1176026165,-0.0530965477,-0.169786334,-0.0831357986,0.014936327,0.1258328259,0.1309275627,0.4253900051,0.0744149089,0.4211103022,-0.293369472,0.4874224067,0.1984412372,-0.351908356,-0.1890533268,-0.029189134,-0.416288197,0.1731599271,-0.0598062463,0.1060002819,-0.0056720939,0.3709750175,-0.1891993135,0.1152638346,0.5879713297,-0.1420108378,-0.1362783313,-0.3244630396,0.1158192828,-0.1114782691,-0.3294302225,-0.5316795707,0.1349311173,0.3023136854,-0.0331192464,-0.1879462451,0.2291242331,-0.0168488622,0.2463797629,-0.0619100817,0.2094120234,0.1433493793,-0.3186507821,-0.1013223976,-0.0981205702]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/687","title":"`ArrowInvalid` occurs while running `Dataset.map()` function","comments":"Hi !\r\n\r\nThis is because `encode` expects one single text as input (str), or one tokenized text (List[str]).\r\nI believe that you actually wanted to use `encode_batch` which expects a batch of texts.\r\nHowever this method is only available for our \"fast\" tokenizers (ex: BertTokenizerFast).\r\nBertJapanese is not one of them unfortunately and I don't think it will be added for now (see https:\/\/github.com\/huggingface\/transformers\/pull\/7141)...\r\ncc @thomwolf for confirmation.\r\n\r\nTherefore what I'd suggest for now is disable batching and process one text at a time using `encode`.\r\nNote that you can make it faster by using multiprocessing:\r\n\r\n```python\r\nnum_proc = None # Specify here the number of processes if you want to use multiprocessing. ex: num_proc = 4\r\nencoded = train_ds.map(\r\n lambda example: {'tokens': t.encode(example['title'], max_length=1000)}, num_proc=num_proc\r\n)\r\n```\r\n","body":"It seems to fail to process the final batch. This [colab](https:\/\/colab.research.google.com\/drive\/1_byLZRHwGP13PHMkJWo62Wp50S_Z2HMD?usp=sharing) can reproduce the error.\r\n\r\nCode:\r\n\r\n```python\r\n# train_ds = Dataset(features: {\r\n# 'title': Value(dtype='string', id=None), \r\n# 'score': Value(dtype='float64', id=None)\r\n# }, num_rows: 99999)\r\n\r\n# suggested in #665 \r\nclass PicklableTokenizer(BertJapaneseTokenizer):\r\n def __getstate__(self):\r\n state = dict(self.__dict__)\r\n state['do_lower_case'] = self.word_tokenizer.do_lower_case\r\n state['never_split'] = self.word_tokenizer.never_split\r\n del state['word_tokenizer']\r\n return state\r\n \r\n def __setstate(self):\r\n do_lower_case = state.pop('do_lower_case')\r\n never_split = state.pop('never_split')\r\n self.__dict__ = state\r\n self.word_tokenizer = MecabTokenizer(\r\n do_lower_case=do_lower_case, never_split=never_split\r\n )\r\n\r\nt = PicklableTokenizer.from_pretrained('bert-base-japanese-whole-word-masking')\r\n\r\nencoded = train_ds.map(\r\n lambda examples: {'tokens': t.encode(examples['title'], max_length=1000)}, batched=True, batch_size=1000\r\n)\r\n```\r\n\r\nError Message:\r\n\r\n```\r\n 99% 99\/100 [00:22<00:00, 39.07ba\/s]\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<timed exec> in <module>\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1496 if update_data:\r\n 1497 batch = cast_to_python_objects(batch)\r\n-> 1498 writer.write_batch(batch)\r\n 1499 if update_data:\r\n 1500 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 271 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 272 typed_sequence_examples[col] = typed_sequence\r\n--> 273 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 274 self.write_table(pa_table)\r\n 275 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_arrays()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.validate()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Column 4 named tokens expected length 999 but got length 1000\r\n```\r\n","comment_length":128,"text":"`ArrowInvalid` occurs while running `Dataset.map()` function\nIt seems to fail to process the final batch. This [colab](https:\/\/colab.research.google.com\/drive\/1_byLZRHwGP13PHMkJWo62Wp50S_Z2HMD?usp=sharing) can reproduce the error.\r\n\r\nCode:\r\n\r\n```python\r\n# train_ds = Dataset(features: {\r\n# 'title': Value(dtype='string', id=None), \r\n# 'score': Value(dtype='float64', id=None)\r\n# }, num_rows: 99999)\r\n\r\n# suggested in #665 \r\nclass PicklableTokenizer(BertJapaneseTokenizer):\r\n def __getstate__(self):\r\n state = dict(self.__dict__)\r\n state['do_lower_case'] = self.word_tokenizer.do_lower_case\r\n state['never_split'] = self.word_tokenizer.never_split\r\n del state['word_tokenizer']\r\n return state\r\n \r\n def __setstate(self):\r\n do_lower_case = state.pop('do_lower_case')\r\n never_split = state.pop('never_split')\r\n self.__dict__ = state\r\n self.word_tokenizer = MecabTokenizer(\r\n do_lower_case=do_lower_case, never_split=never_split\r\n )\r\n\r\nt = PicklableTokenizer.from_pretrained('bert-base-japanese-whole-word-masking')\r\n\r\nencoded = train_ds.map(\r\n lambda examples: {'tokens': t.encode(examples['title'], max_length=1000)}, batched=True, batch_size=1000\r\n)\r\n```\r\n\r\nError Message:\r\n\r\n```\r\n 99% 99\/100 [00:22<00:00, 39.07ba\/s]\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<timed exec> in <module>\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1496 if update_data:\r\n 1497 batch = cast_to_python_objects(batch)\r\n-> 1498 writer.write_batch(batch)\r\n 1499 if update_data:\r\n 1500 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 271 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 272 typed_sequence_examples[col] = typed_sequence\r\n--> 273 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 274 self.write_table(pa_table)\r\n 275 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_arrays()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.validate()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Column 4 named tokens expected length 999 but got length 1000\r\n```\r\n\nHi !\r\n\r\nThis is because `encode` expects one single text as input (str), or one tokenized text (List[str]).\r\nI believe that you actually wanted to use `encode_batch` which expects a batch of texts.\r\nHowever this method is only available for our \"fast\" tokenizers (ex: BertTokenizerFast).\r\nBertJapanese is not one of them unfortunately and I don't think it will be added for now (see https:\/\/github.com\/huggingface\/transformers\/pull\/7141)...\r\ncc @thomwolf for confirmation.\r\n\r\nTherefore what I'd suggest for now is disable batching and process one text at a time using `encode`.\r\nNote that you can make it faster by using multiprocessing:\r\n\r\n```python\r\nnum_proc = None # Specify here the number of processes if you want to use multiprocessing. ex: num_proc = 4\r\nencoded = train_ds.map(\r\n lambda example: {'tokens': t.encode(example['title'], max_length=1000)}, num_proc=num_proc\r\n)\r\n```\r\n","embeddings":[-0.3730494082,0.0209990088,-0.1026319936,-0.0166299157,0.1107027829,0.1678620726,0.1268472821,0.3861759603,-0.1962419301,0.116063267,0.2094027549,0.6037945747,-0.1167867631,0.0404954627,-0.3405689895,0.0122450693,-0.0138801886,0.2132770419,0.0759285465,-0.2047145665,-0.4590324759,0.0273315106,-0.3769527078,0.2137223929,-0.2793795764,-0.46280545,0.2793683112,-0.1693802327,-0.2238800675,-0.74639678,0.2083986551,-0.3749470711,-0.0207395498,0.3278741539,-0.0001172259,0.042348668,0.4025927782,0.0688235834,-0.2501145005,-0.0585927963,-0.443045795,0.0925626904,-0.0866485387,-0.1329717785,0.2778750062,-0.0361048207,0.0314678699,-0.4573630393,0.3919546306,0.2821494341,0.1263703704,0.2222543806,0.1442392766,0.0323247947,0.1279150993,-0.1200693548,-0.0969361216,0.1047375202,0.0181517769,-0.2857314944,-0.0512987599,0.5154019594,0.180965513,0.0922654271,-0.14175497,-0.1290374547,0.1971558928,-0.4126629233,0.3031370938,-0.0948614031,0.0813262165,-0.1926683784,-0.4155083001,-0.0138592795,0.1312633455,-0.1198019758,0.0628945976,0.0746485814,-0.0928725153,0.0599402264,-0.0519259274,0.167006433,-0.2132056057,0.0801583529,0.1788284779,0.4855768085,0.0136159519,0.4012263417,0.2158066779,-0.1161201075,0.0043541109,0.0970336497,0.0199738536,0.241735518,-0.2374131531,-0.127425164,-0.264008224,-0.1885884553,0.1048256159,-0.335390538,0.1042770669,0.0072844056,0.1545805335,0.1884318292,0.3311057091,0.0799308792,0.115690656,0.6522192955,-0.1187430173,-0.4918249249,-0.0285963994,0.0985013545,-0.0701802,-0.4311812818,0.2728593647,0.3366879225,0.1505464315,0.1117760018,-0.3864218295,0.1871385127,-0.589379549,-0.0437153503,-0.0479695611,0.1943127215,0.1245247126,0.0586616062,-0.0686166659,-0.0534041934,-0.0046589328,-0.1363479495,-0.0411160998,0.108408913,-0.4678287506,-0.0251717102,0.0634236187,-0.0948988721,0.3233495355,-0.0266724546,0.1230678931,0.0514647625,0.314424336,-0.2968288362,0.444064647,0.0961867496,-0.0872931406,0.3249668777,0.1295829117,0.1602937281,-0.0586597286,0.128712222,-0.1173376068,-0.230542466,0.0751926452,0.1038942412,0.3227936029,-0.0253573768,-0.1583411545,-0.0692773163,0.4308267832,-0.0557872914,0.178976655,-0.3901212513,0.108979933,-0.2312533557,-0.0915208161,0.2688697875,-0.586361587,0.1150185391,0.0958043039,0.2284136713,0.2394196093,0.2511422038,-0.3237181306,0.2681290507,-0.0672967136,0.7344178557,0.0913244188,-0.4536603689,-0.5696929097,0.1651970297,-0.0714170337,-0.2872891724,-0.1635802388,-0.0753729269,0.1052654237,-0.0703787953,0.2810492516,0.1559283584,-0.206625998,0.0811991617,-0.0850825608,-0.1805368215,0.1134609282,0.0650475174,0.1270639449,-0.2152681202,0.0416606329,-0.2121016532,-0.0126661211,-0.0910451859,0.3163724542,0.1728592515,0.1968010515,-0.0793101043,0.1716614217,0.0586773306,-0.0976552442,0.1258585751,0.0379586145,-0.1189674586,-0.4610619247,-0.2379849702,-0.3678559661,0.3181095123,-0.4086672664,-0.022437958,0.1247887909,0.154564321,-0.0500557572,-0.1316203624,-0.2792712748,-0.2015069723,0.0223132372,-0.0179371852,-0.1443542987,-0.0236081909,-0.1345303059,-0.3064616323,0.1208036989,-0.0018816206,0.1851107925,-0.0172275007,-0.2361747921,0.4019815028,0.1110517904,-0.1824735701,0.2651119232,-0.2703411281,0.0067328028,-0.2617128789,-0.1663633138,0.1191149056,0.2205321938,0.1281047016,0.2508363426,0.3558339179,0.1522091031,0.1357205212,-0.1720412374,0.0719726831,0.1325083375,0.1398831606,-0.1958693117,-0.2773173451,-0.2591813207,-0.1319920123,0.5122656226,-0.1623483896,-0.1164862514,0.0374570452,0.3888116777,0.0573440157,0.0788703561,-0.0464808904,-0.2116911262,0.1952369362,0.1356423497,0.2061406076,0.0987049192,0.0820548683,0.0950869471,-0.0831235722,0.0089299865,0.0431402698,0.0941936523,0.1881881356,0.3178276718,0.2141555399,-0.0749313682,0.0543545187,-0.2489284128,-0.1475625336,0.1688107401,0.2606538236,-0.1852554679,0.1725290269,-0.1887763441,-0.0659476593,-0.0683829561,-0.4471708238,0.170828104,-0.2091741413,-0.0630615354,-0.0351297706,-0.1834743172,0.4110240638,0.0006338063,0.1164892316,0.4737392068,-0.0154695604,-0.0733822435,-0.276661396,-0.1933734715,-0.0152658643,0.228355974,-0.1102727801,0.296617955,0.220810011,0.0844046921,-0.2958572507,-0.0957247838,0.2139255553,-0.1807052791,-0.0076032318,0.2341925651,0.0471360534,-0.5811719894,-0.3872248232,0.1693429947,-0.0757252946,-0.2313343287,-0.15086025,-0.002469372,-0.2634731829,0.0694960281,-0.3197098076,-0.2123733908,-0.2388973385,0.1862350702,-0.2556763291,-0.0316172689,0.0654448792,0.2022592425,0.3322208226,-0.081564784,-0.3609610796,-0.1151162907,0.0165958162,0.3085134029,-0.2723379433,0.0018396794,0.029960243,-0.1951613277,0.2619602382,0.2316198945,-0.1621741951,0.1872944087,-0.1391092986,0.2304623574,-0.3229131997,0.2915933132,0.3173620701,0.2290569842,-0.1079699919,-0.2228882611,0.0057419599,0.1587736607,-0.1930841953,0.0635004044,0.0404598899,0.4070858359,0.0578363314,0.9904161692,0.3290010095,-0.184348613,0.1601355821,0.0439409912,-0.130933702,-0.1031865254,-0.2093008757,-0.1285343915,-0.2524230182,-0.0792861208,0.1366002709,-0.1981544793,-0.3025023937,0.249754861,0.0105883637,-0.5632714033,-0.2375912964,0.254394114,-0.2797160745,0.2588542998,-0.0024122249,0.0555759594,-0.2642837763,0.2071430534,0.0718878284,-0.3598581851,0.3196207583,-0.1441478431,0.0894451216,-0.2419476509,-0.3778175116,0.3687334657,0.3592782915,0.6092718244,0.0377681069,-0.007376573,-0.0294802096,0.0970075056,0.4181975126,-0.3694489598,0.0814596862,-0.0399212688,-0.1231451482,-0.0497128032,0.0469382554,-0.2737925053,0.5196662545,0.3642606437,0.2670809925,-0.2750947773,0.0026986506,0.2009459138,-0.1017457247,0.0590404645,-0.0111810556,-0.1706310064,-0.0688263997,-0.558647275,0.1786676347,0.4432715178,-0.1226824522,0.2515909672,-0.1211954206,-0.255768925,-0.197385028,-0.0335434005,0.1136311516,0.3274050057,0.2703148425,0.0766888112,0.1624634713,0.2545147836,0.4407124817,0.2038445622,-0.1080116034,-0.4675176442,0.1281770617,-0.1398986429,0.2996688187,0.3086300492,-0.1780779213,0.0501075946,0.0881881416,0.0509525388,0.2329247296,0.242156893,0.2835567892,0.2226179987,-0.439637363,-0.2767536938,0.4049759805,-0.0138686616,-0.0912982821,0.6421588063,0.3474164903,-0.1849973798,0.3692859411,0.4013648927,0.7862712145,0.2532509863,-0.0124251796,0.1248406246,0.2170794308,-0.1023470685,0.4299082756,0.1834250242,-0.3477331698,0.2617433667,0.0012240701,-0.0906033665,0.0341689512,0.1903268099,-0.2757633626,-0.0677879825,-0.121113047,0.0702949092,0.0555842109,-0.0016870816,-0.0833590031,-0.331281811,0.1922793984,0.0611589663,0.0844454318,-0.0350100398,-0.1506827027,-0.2564366162,-0.1501162499,-0.1496674269,-0.1197075769,0.0060363072,-0.6829081774,0.565030992,-0.0459312908,-0.0839272514,0.15096955,-0.0842965841,0.3340730071,0.025580382,-0.0353463814,0.1205937564,0.4888465405,-0.1497036666,-0.1564611495,-0.2073270082,0.309296757,-0.0101799294,-0.1845765561,0.122286737,-0.1697846502,-0.3743094802,-0.0325552709,0.0764172077,0.1754269153,-0.3503093719,-0.2272080928,-0.0709934682,-0.3393777609,-0.1550418884,0.0502095185,-0.133615315,-0.1731118113,0.2058069855,0.0522395112,-0.3172770441,0.2279915959,0.1361300498,-0.0892230868,0.0557560474,0.8029521108,-0.1772937626,-0.2020333558,-0.2253221571,0.1978305429,0.5487523675,-0.3811806142,0.1656246632,-0.3322678506,0.0752159134,0.1803031117,0.3500501215,0.1772226393,0.1831656992,-0.1500357985,-0.3336010575,-0.3557195663,0.1322591305,-0.1198241189,0.3273670077,0.0387873054,0.3118831515,-0.0322746821,0.0812250674,-0.2580687702,-0.1572238952,-0.2238389403,0.2041134387,0.0935177729,0.1801964492,-0.0656198114,0.1206872463,0.0327497423,0.5000814795,-0.1537916362,-0.1777279824,0.0156333335,0.1212539002,0.1292287707,-0.0560172237,-0.0715844631,-0.1047161445,-0.284964174,-0.0446433835,-0.1142784581,-0.1097290367,0.0111781834,0.2695165873,0.1076420397,0.019040782,0.0980497599,-0.0959703028,-0.0032661033,0.0494811386,-0.0821532682,0.3108957112,0.1330127716,-0.002252745,-0.048183836,-0.1015686765,0.0886421874,0.1926507354,0.0492651127,-0.5152286887,-0.0905362442,0.0867754072,0.1024260223,0.2516393065,-0.3033187985,-0.2979983389,0.3304128647,0.1884914339,-0.2688320875,-0.0114826718,0.1021456942,-0.0675078928,0.1170439348,0.0695642084,-0.1452982724,0.0703056976,-0.2506695092,0.1509166956,-0.0121543929,-0.2579898834,0.1355048418,0.488517344,-0.0827949271,0.0725362301,0.2948015034,0.2546038032,0.4049094617,0.5791914463,-0.1654969454,0.6184579134,0.2191292942,0.0881407633,-0.0633769631,-0.1177311465,0.0770792887,0.2516070604,-0.0624556988,0.2434403002,0.4535259008,0.4902361929,-0.2501936257,-0.4952516556,0.1045922488,0.0324253552,-0.3045398593,-0.0541481525,-0.6603556871,-0.1197683439,-0.0567469932,0.015529843,-0.1565536112,0.3162733316,0.421939522,0.1472989768,-0.0421621874,-0.1396858543,-0.2276960313,0.2237729877,0.2358185649,-0.3784120977,0.0090177692,-0.0332620479,0.0109482845,-0.2814303935,0.4175459743,0.4623645544,0.3912786245,-0.3112112284,-0.0724628866,-0.0561839677,-0.0738908947,0.0694638118,0.1608723253,0.339402169,-0.0586858653,0.3028180301,0.0987344831,-0.1526950151,-0.0308848955,0.1996582001,0.222965464,0.2359169424,-0.286049813,-0.0518292673,-0.1972729117,-0.0275575332,0.0090576569,-0.3988699019,0.315263778,0.3137662411,-0.1130291,0.068776913,-0.1158980802,0.0592510998,-0.0700498596,0.5608128309,0.4262844324,-0.0452025943,-0.399895221,-0.1887740642,-0.3950439692,0.1455415934,-0.2550313771,-0.0440195054,0.0581868701,0.0753254965,0.1059387848,0.3091356158,0.3110461235,-0.0556932874,-0.2100166529,0.4253170788,-0.1567313969,-0.0507003553,0.0598544367,-0.0091930907,0.1691812128,-0.4654403031,0.0733946413,0.1306339204,0.0990783125,-0.1603719294,-0.4318068326,0.0909193009,-0.1860193759,0.5836700201,0.2212095559,0.1505903155,-0.2123440951,-0.0772735104,-0.1093366593,-0.2141530812,-0.1267417669,0.0200576838,0.3289568126,0.5786898732,-0.0753559023,0.0539231524,-0.2971648872,-0.1712229997,0.0016312442,-0.0309432037,-0.2057603151,0.1793767661,-0.1247466281,0.0558500625,-0.2652968466,0.2664737105,0.0658193007,0.1168129072,-0.3982465863,-0.347958982,0.5152161121,-0.6306791306,-0.676350534,0.0664059892,-0.1369708627,0.3331304491,-0.0322293751,-0.4938650131,-0.0091210017,0.1045906097,0.0493355431,-0.4475872815,-0.0207463726,0.0198748354,0.1007239595,0.0060973959,0.6639660597,0.1340132207,-0.1051306948,-0.0541317388,-0.3198482692]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/687","title":"`ArrowInvalid` occurs while running `Dataset.map()` function","comments":"Thank you very much for the kind and precise suggestion!\r\nI'm looking forward to seeing BertJapaneseTokenizer built into the \"fast\" tokenizers.\r\n\r\nI tried `map` with multiprocessing as follows, and it worked!\r\n\r\n```python\r\n# There was a Pickle problem if I use `lambda` for multiprocessing\r\ndef encode(examples):\r\n return {'tokens': t.encode(examples['title'], max_length=1000)}\r\n\r\nnum_proc = 8\r\nencoded = train_ds.map(encode, num_proc=num_proc)\r\n```\r\n\r\nThank you again!","body":"It seems to fail to process the final batch. This [colab](https:\/\/colab.research.google.com\/drive\/1_byLZRHwGP13PHMkJWo62Wp50S_Z2HMD?usp=sharing) can reproduce the error.\r\n\r\nCode:\r\n\r\n```python\r\n# train_ds = Dataset(features: {\r\n# 'title': Value(dtype='string', id=None), \r\n# 'score': Value(dtype='float64', id=None)\r\n# }, num_rows: 99999)\r\n\r\n# suggested in #665 \r\nclass PicklableTokenizer(BertJapaneseTokenizer):\r\n def __getstate__(self):\r\n state = dict(self.__dict__)\r\n state['do_lower_case'] = self.word_tokenizer.do_lower_case\r\n state['never_split'] = self.word_tokenizer.never_split\r\n del state['word_tokenizer']\r\n return state\r\n \r\n def __setstate(self):\r\n do_lower_case = state.pop('do_lower_case')\r\n never_split = state.pop('never_split')\r\n self.__dict__ = state\r\n self.word_tokenizer = MecabTokenizer(\r\n do_lower_case=do_lower_case, never_split=never_split\r\n )\r\n\r\nt = PicklableTokenizer.from_pretrained('bert-base-japanese-whole-word-masking')\r\n\r\nencoded = train_ds.map(\r\n lambda examples: {'tokens': t.encode(examples['title'], max_length=1000)}, batched=True, batch_size=1000\r\n)\r\n```\r\n\r\nError Message:\r\n\r\n```\r\n 99% 99\/100 [00:22<00:00, 39.07ba\/s]\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<timed exec> in <module>\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1496 if update_data:\r\n 1497 batch = cast_to_python_objects(batch)\r\n-> 1498 writer.write_batch(batch)\r\n 1499 if update_data:\r\n 1500 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 271 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 272 typed_sequence_examples[col] = typed_sequence\r\n--> 273 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 274 self.write_table(pa_table)\r\n 275 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_arrays()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.validate()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Column 4 named tokens expected length 999 but got length 1000\r\n```\r\n","comment_length":61,"text":"`ArrowInvalid` occurs while running `Dataset.map()` function\nIt seems to fail to process the final batch. This [colab](https:\/\/colab.research.google.com\/drive\/1_byLZRHwGP13PHMkJWo62Wp50S_Z2HMD?usp=sharing) can reproduce the error.\r\n\r\nCode:\r\n\r\n```python\r\n# train_ds = Dataset(features: {\r\n# 'title': Value(dtype='string', id=None), \r\n# 'score': Value(dtype='float64', id=None)\r\n# }, num_rows: 99999)\r\n\r\n# suggested in #665 \r\nclass PicklableTokenizer(BertJapaneseTokenizer):\r\n def __getstate__(self):\r\n state = dict(self.__dict__)\r\n state['do_lower_case'] = self.word_tokenizer.do_lower_case\r\n state['never_split'] = self.word_tokenizer.never_split\r\n del state['word_tokenizer']\r\n return state\r\n \r\n def __setstate(self):\r\n do_lower_case = state.pop('do_lower_case')\r\n never_split = state.pop('never_split')\r\n self.__dict__ = state\r\n self.word_tokenizer = MecabTokenizer(\r\n do_lower_case=do_lower_case, never_split=never_split\r\n )\r\n\r\nt = PicklableTokenizer.from_pretrained('bert-base-japanese-whole-word-masking')\r\n\r\nencoded = train_ds.map(\r\n lambda examples: {'tokens': t.encode(examples['title'], max_length=1000)}, batched=True, batch_size=1000\r\n)\r\n```\r\n\r\nError Message:\r\n\r\n```\r\n 99% 99\/100 [00:22<00:00, 39.07ba\/s]\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<timed exec> in <module>\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data)\r\n 1496 if update_data:\r\n 1497 batch = cast_to_python_objects(batch)\r\n-> 1498 writer.write_batch(batch)\r\n 1499 if update_data:\r\n 1500 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 271 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type)\r\n 272 typed_sequence_examples[col] = typed_sequence\r\n--> 273 pa_table = pa.Table.from_pydict(typed_sequence_examples)\r\n 274 self.write_table(pa_table)\r\n 275 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pydict()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_arrays()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.validate()\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Column 4 named tokens expected length 999 but got length 1000\r\n```\r\n\nThank you very much for the kind and precise suggestion!\r\nI'm looking forward to seeing BertJapaneseTokenizer built into the \"fast\" tokenizers.\r\n\r\nI tried `map` with multiprocessing as follows, and it worked!\r\n\r\n```python\r\n# There was a Pickle problem if I use `lambda` for multiprocessing\r\ndef encode(examples):\r\n return {'tokens': t.encode(examples['title'], max_length=1000)}\r\n\r\nnum_proc = 8\r\nencoded = train_ds.map(encode, num_proc=num_proc)\r\n```\r\n\r\nThank you again!","embeddings":[-0.3730494082,0.0209990088,-0.1026319936,-0.0166299157,0.1107027829,0.1678620726,0.1268472821,0.3861759603,-0.1962419301,0.116063267,0.2094027549,0.6037945747,-0.1167867631,0.0404954627,-0.3405689895,0.0122450693,-0.0138801886,0.2132770419,0.0759285465,-0.2047145665,-0.4590324759,0.0273315106,-0.3769527078,0.2137223929,-0.2793795764,-0.46280545,0.2793683112,-0.1693802327,-0.2238800675,-0.74639678,0.2083986551,-0.3749470711,-0.0207395498,0.3278741539,-0.0001172259,0.042348668,0.4025927782,0.0688235834,-0.2501145005,-0.0585927963,-0.443045795,0.0925626904,-0.0866485387,-0.1329717785,0.2778750062,-0.0361048207,0.0314678699,-0.4573630393,0.3919546306,0.2821494341,0.1263703704,0.2222543806,0.1442392766,0.0323247947,0.1279150993,-0.1200693548,-0.0969361216,0.1047375202,0.0181517769,-0.2857314944,-0.0512987599,0.5154019594,0.180965513,0.0922654271,-0.14175497,-0.1290374547,0.1971558928,-0.4126629233,0.3031370938,-0.0948614031,0.0813262165,-0.1926683784,-0.4155083001,-0.0138592795,0.1312633455,-0.1198019758,0.0628945976,0.0746485814,-0.0928725153,0.0599402264,-0.0519259274,0.167006433,-0.2132056057,0.0801583529,0.1788284779,0.4855768085,0.0136159519,0.4012263417,0.2158066779,-0.1161201075,0.0043541109,0.0970336497,0.0199738536,0.241735518,-0.2374131531,-0.127425164,-0.264008224,-0.1885884553,0.1048256159,-0.335390538,0.1042770669,0.0072844056,0.1545805335,0.1884318292,0.3311057091,0.0799308792,0.115690656,0.6522192955,-0.1187430173,-0.4918249249,-0.0285963994,0.0985013545,-0.0701802,-0.4311812818,0.2728593647,0.3366879225,0.1505464315,0.1117760018,-0.3864218295,0.1871385127,-0.589379549,-0.0437153503,-0.0479695611,0.1943127215,0.1245247126,0.0586616062,-0.0686166659,-0.0534041934,-0.0046589328,-0.1363479495,-0.0411160998,0.108408913,-0.4678287506,-0.0251717102,0.0634236187,-0.0948988721,0.3233495355,-0.0266724546,0.1230678931,0.0514647625,0.314424336,-0.2968288362,0.444064647,0.0961867496,-0.0872931406,0.3249668777,0.1295829117,0.1602937281,-0.0586597286,0.128712222,-0.1173376068,-0.230542466,0.0751926452,0.1038942412,0.3227936029,-0.0253573768,-0.1583411545,-0.0692773163,0.4308267832,-0.0557872914,0.178976655,-0.3901212513,0.108979933,-0.2312533557,-0.0915208161,0.2688697875,-0.586361587,0.1150185391,0.0958043039,0.2284136713,0.2394196093,0.2511422038,-0.3237181306,0.2681290507,-0.0672967136,0.7344178557,0.0913244188,-0.4536603689,-0.5696929097,0.1651970297,-0.0714170337,-0.2872891724,-0.1635802388,-0.0753729269,0.1052654237,-0.0703787953,0.2810492516,0.1559283584,-0.206625998,0.0811991617,-0.0850825608,-0.1805368215,0.1134609282,0.0650475174,0.1270639449,-0.2152681202,0.0416606329,-0.2121016532,-0.0126661211,-0.0910451859,0.3163724542,0.1728592515,0.1968010515,-0.0793101043,0.1716614217,0.0586773306,-0.0976552442,0.1258585751,0.0379586145,-0.1189674586,-0.4610619247,-0.2379849702,-0.3678559661,0.3181095123,-0.4086672664,-0.022437958,0.1247887909,0.154564321,-0.0500557572,-0.1316203624,-0.2792712748,-0.2015069723,0.0223132372,-0.0179371852,-0.1443542987,-0.0236081909,-0.1345303059,-0.3064616323,0.1208036989,-0.0018816206,0.1851107925,-0.0172275007,-0.2361747921,0.4019815028,0.1110517904,-0.1824735701,0.2651119232,-0.2703411281,0.0067328028,-0.2617128789,-0.1663633138,0.1191149056,0.2205321938,0.1281047016,0.2508363426,0.3558339179,0.1522091031,0.1357205212,-0.1720412374,0.0719726831,0.1325083375,0.1398831606,-0.1958693117,-0.2773173451,-0.2591813207,-0.1319920123,0.5122656226,-0.1623483896,-0.1164862514,0.0374570452,0.3888116777,0.0573440157,0.0788703561,-0.0464808904,-0.2116911262,0.1952369362,0.1356423497,0.2061406076,0.0987049192,0.0820548683,0.0950869471,-0.0831235722,0.0089299865,0.0431402698,0.0941936523,0.1881881356,0.3178276718,0.2141555399,-0.0749313682,0.0543545187,-0.2489284128,-0.1475625336,0.1688107401,0.2606538236,-0.1852554679,0.1725290269,-0.1887763441,-0.0659476593,-0.0683829561,-0.4471708238,0.170828104,-0.2091741413,-0.0630615354,-0.0351297706,-0.1834743172,0.4110240638,0.0006338063,0.1164892316,0.4737392068,-0.0154695604,-0.0733822435,-0.276661396,-0.1933734715,-0.0152658643,0.228355974,-0.1102727801,0.296617955,0.220810011,0.0844046921,-0.2958572507,-0.0957247838,0.2139255553,-0.1807052791,-0.0076032318,0.2341925651,0.0471360534,-0.5811719894,-0.3872248232,0.1693429947,-0.0757252946,-0.2313343287,-0.15086025,-0.002469372,-0.2634731829,0.0694960281,-0.3197098076,-0.2123733908,-0.2388973385,0.1862350702,-0.2556763291,-0.0316172689,0.0654448792,0.2022592425,0.3322208226,-0.081564784,-0.3609610796,-0.1151162907,0.0165958162,0.3085134029,-0.2723379433,0.0018396794,0.029960243,-0.1951613277,0.2619602382,0.2316198945,-0.1621741951,0.1872944087,-0.1391092986,0.2304623574,-0.3229131997,0.2915933132,0.3173620701,0.2290569842,-0.1079699919,-0.2228882611,0.0057419599,0.1587736607,-0.1930841953,0.0635004044,0.0404598899,0.4070858359,0.0578363314,0.9904161692,0.3290010095,-0.184348613,0.1601355821,0.0439409912,-0.130933702,-0.1031865254,-0.2093008757,-0.1285343915,-0.2524230182,-0.0792861208,0.1366002709,-0.1981544793,-0.3025023937,0.249754861,0.0105883637,-0.5632714033,-0.2375912964,0.254394114,-0.2797160745,0.2588542998,-0.0024122249,0.0555759594,-0.2642837763,0.2071430534,0.0718878284,-0.3598581851,0.3196207583,-0.1441478431,0.0894451216,-0.2419476509,-0.3778175116,0.3687334657,0.3592782915,0.6092718244,0.0377681069,-0.007376573,-0.0294802096,0.0970075056,0.4181975126,-0.3694489598,0.0814596862,-0.0399212688,-0.1231451482,-0.0497128032,0.0469382554,-0.2737925053,0.5196662545,0.3642606437,0.2670809925,-0.2750947773,0.0026986506,0.2009459138,-0.1017457247,0.0590404645,-0.0111810556,-0.1706310064,-0.0688263997,-0.558647275,0.1786676347,0.4432715178,-0.1226824522,0.2515909672,-0.1211954206,-0.255768925,-0.197385028,-0.0335434005,0.1136311516,0.3274050057,0.2703148425,0.0766888112,0.1624634713,0.2545147836,0.4407124817,0.2038445622,-0.1080116034,-0.4675176442,0.1281770617,-0.1398986429,0.2996688187,0.3086300492,-0.1780779213,0.0501075946,0.0881881416,0.0509525388,0.2329247296,0.242156893,0.2835567892,0.2226179987,-0.439637363,-0.2767536938,0.4049759805,-0.0138686616,-0.0912982821,0.6421588063,0.3474164903,-0.1849973798,0.3692859411,0.4013648927,0.7862712145,0.2532509863,-0.0124251796,0.1248406246,0.2170794308,-0.1023470685,0.4299082756,0.1834250242,-0.3477331698,0.2617433667,0.0012240701,-0.0906033665,0.0341689512,0.1903268099,-0.2757633626,-0.0677879825,-0.121113047,0.0702949092,0.0555842109,-0.0016870816,-0.0833590031,-0.331281811,0.1922793984,0.0611589663,0.0844454318,-0.0350100398,-0.1506827027,-0.2564366162,-0.1501162499,-0.1496674269,-0.1197075769,0.0060363072,-0.6829081774,0.565030992,-0.0459312908,-0.0839272514,0.15096955,-0.0842965841,0.3340730071,0.025580382,-0.0353463814,0.1205937564,0.4888465405,-0.1497036666,-0.1564611495,-0.2073270082,0.309296757,-0.0101799294,-0.1845765561,0.122286737,-0.1697846502,-0.3743094802,-0.0325552709,0.0764172077,0.1754269153,-0.3503093719,-0.2272080928,-0.0709934682,-0.3393777609,-0.1550418884,0.0502095185,-0.133615315,-0.1731118113,0.2058069855,0.0522395112,-0.3172770441,0.2279915959,0.1361300498,-0.0892230868,0.0557560474,0.8029521108,-0.1772937626,-0.2020333558,-0.2253221571,0.1978305429,0.5487523675,-0.3811806142,0.1656246632,-0.3322678506,0.0752159134,0.1803031117,0.3500501215,0.1772226393,0.1831656992,-0.1500357985,-0.3336010575,-0.3557195663,0.1322591305,-0.1198241189,0.3273670077,0.0387873054,0.3118831515,-0.0322746821,0.0812250674,-0.2580687702,-0.1572238952,-0.2238389403,0.2041134387,0.0935177729,0.1801964492,-0.0656198114,0.1206872463,0.0327497423,0.5000814795,-0.1537916362,-0.1777279824,0.0156333335,0.1212539002,0.1292287707,-0.0560172237,-0.0715844631,-0.1047161445,-0.284964174,-0.0446433835,-0.1142784581,-0.1097290367,0.0111781834,0.2695165873,0.1076420397,0.019040782,0.0980497599,-0.0959703028,-0.0032661033,0.0494811386,-0.0821532682,0.3108957112,0.1330127716,-0.002252745,-0.048183836,-0.1015686765,0.0886421874,0.1926507354,0.0492651127,-0.5152286887,-0.0905362442,0.0867754072,0.1024260223,0.2516393065,-0.3033187985,-0.2979983389,0.3304128647,0.1884914339,-0.2688320875,-0.0114826718,0.1021456942,-0.0675078928,0.1170439348,0.0695642084,-0.1452982724,0.0703056976,-0.2506695092,0.1509166956,-0.0121543929,-0.2579898834,0.1355048418,0.488517344,-0.0827949271,0.0725362301,0.2948015034,0.2546038032,0.4049094617,0.5791914463,-0.1654969454,0.6184579134,0.2191292942,0.0881407633,-0.0633769631,-0.1177311465,0.0770792887,0.2516070604,-0.0624556988,0.2434403002,0.4535259008,0.4902361929,-0.2501936257,-0.4952516556,0.1045922488,0.0324253552,-0.3045398593,-0.0541481525,-0.6603556871,-0.1197683439,-0.0567469932,0.015529843,-0.1565536112,0.3162733316,0.421939522,0.1472989768,-0.0421621874,-0.1396858543,-0.2276960313,0.2237729877,0.2358185649,-0.3784120977,0.0090177692,-0.0332620479,0.0109482845,-0.2814303935,0.4175459743,0.4623645544,0.3912786245,-0.3112112284,-0.0724628866,-0.0561839677,-0.0738908947,0.0694638118,0.1608723253,0.339402169,-0.0586858653,0.3028180301,0.0987344831,-0.1526950151,-0.0308848955,0.1996582001,0.222965464,0.2359169424,-0.286049813,-0.0518292673,-0.1972729117,-0.0275575332,0.0090576569,-0.3988699019,0.315263778,0.3137662411,-0.1130291,0.068776913,-0.1158980802,0.0592510998,-0.0700498596,0.5608128309,0.4262844324,-0.0452025943,-0.399895221,-0.1887740642,-0.3950439692,0.1455415934,-0.2550313771,-0.0440195054,0.0581868701,0.0753254965,0.1059387848,0.3091356158,0.3110461235,-0.0556932874,-0.2100166529,0.4253170788,-0.1567313969,-0.0507003553,0.0598544367,-0.0091930907,0.1691812128,-0.4654403031,0.0733946413,0.1306339204,0.0990783125,-0.1603719294,-0.4318068326,0.0909193009,-0.1860193759,0.5836700201,0.2212095559,0.1505903155,-0.2123440951,-0.0772735104,-0.1093366593,-0.2141530812,-0.1267417669,0.0200576838,0.3289568126,0.5786898732,-0.0753559023,0.0539231524,-0.2971648872,-0.1712229997,0.0016312442,-0.0309432037,-0.2057603151,0.1793767661,-0.1247466281,0.0558500625,-0.2652968466,0.2664737105,0.0658193007,0.1168129072,-0.3982465863,-0.347958982,0.5152161121,-0.6306791306,-0.676350534,0.0664059892,-0.1369708627,0.3331304491,-0.0322293751,-0.4938650131,-0.0091210017,0.1045906097,0.0493355431,-0.4475872815,-0.0207463726,0.0198748354,0.1007239595,0.0060973959,0.6639660597,0.1340132207,-0.1051306948,-0.0541317388,-0.3198482692]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/686","title":"Dataset browser url is still https:\/\/huggingface.co\/nlp\/viewer\/","comments":"Yes! might do it with @srush one of these days. Hopefully it won't break too many links (we can always redirect from old url to new)","body":"Might be worth updating to https:\/\/huggingface.co\/datasets\/viewer\/","comment_length":26,"text":"Dataset browser url is still https:\/\/huggingface.co\/nlp\/viewer\/\nMight be worth updating to https:\/\/huggingface.co\/datasets\/viewer\/\nYes! might do it with @srush one of these days. Hopefully it won't break too many links (we can always redirect from old url to new)","embeddings":[-0.1529248506,0.2032147646,-0.0998434573,-0.1728723794,0.1061696857,0.0467374735,0.179426223,0.2390701473,-0.0500670075,-0.0185796227,-0.1565577686,0.30587098,0.1280383468,0.1724736094,0.3105870485,0.0242812056,0.0012544101,-0.0140798306,-0.1639329493,-0.0225632489,-0.1863491833,0.0512106866,-0.195139125,0.180641219,-0.0065673641,0.0746079534,-0.0707541332,0.2305047661,-0.177075997,-0.3989592195,0.1384267807,0.1680058539,-0.0488178357,0.2282091677,-0.0000960609,0.0672143474,0.4233936071,0.2476758063,-0.3446639478,-0.0000749466,-0.3106382787,-0.065295957,0.1053099558,-0.1298185736,-0.1422041506,-0.2468431741,0.3551162481,-0.065297775,0.0736937225,0.2408897728,0.3633951843,0.3327011764,-0.1858961284,-0.089773424,-0.0542072952,0.1208049208,-0.0696513131,0.0434561446,-0.0182185322,0.066160433,-0.239288643,0.561126709,0.2315685153,-0.0279473159,0.200971365,-0.0143021494,-0.353766799,0.0535263084,0.079043135,0.055047933,0.6028001904,-0.1018182784,-0.1742539704,-0.0268567204,-0.0725952238,0.0548230186,0.2638870776,0.1444999427,0.0604988337,0.2288692594,-0.2419080734,-0.5008474588,-0.1607080102,0.1459707916,-0.131855011,0.3666667938,-0.0955795869,-0.0348553769,0.1961886585,-0.1912985295,0.2546566129,0.120418258,-0.0303971823,-0.263392508,0.1133593619,-0.0686037317,0.2350059301,0.4213660955,0.0431765169,0.2189738303,0.154790014,0.1485686898,-0.1884457916,-0.0044194642,0.0471914522,-0.0387113467,0.0920071825,-0.0088646691,0.6043276191,0.2491928041,0.2259243429,-0.0179779641,0.1733002812,-0.3365879655,-0.3934387267,-0.3392239511,0.2704303861,-0.1058420688,-0.162116915,-0.1167059839,0.1140188202,-0.0489502884,-0.0150499037,0.3982775509,-0.1377293169,-0.0157084912,-0.0406846069,-0.1462123841,-0.2074071765,-0.2186151147,-0.3124217391,0.0294859931,-0.3069172204,-0.0521591716,0.2305074185,-0.0377479866,0.1992869228,-0.1417885274,-0.1505475044,0.0472174175,-0.2635045052,0.2486747354,0.0366090983,0.1801086217,0.0523627028,0.044847589,0.0209172051,-0.2001548409,0.0460464954,0.0541520789,-0.2276464999,-0.1882440597,-0.1267290413,0.2991304398,-0.1316419691,-0.3244582117,0.0465979055,0.206476897,-0.252592802,-0.1256885082,-0.088467285,0.2324637026,-0.1503419429,0.0034821709,0.0761694983,0.2025158554,-0.1588964164,-0.2690931261,0.0305513591,-0.2208424509,-0.2299480885,0.2701799273,-0.2201052308,-0.2789732814,-0.3101256192,0.2633773386,0.6101366282,-0.169667244,-0.3256669641,0.1078999639,-0.2597180009,-0.2637858093,-0.087349087,0.1183230802,-0.0149701741,-0.0187621675,-0.014495492,0.1476344615,0.144870162,0.219928056,-0.2966607809,-0.2687580287,0.0122952219,0.085551165,0.0404122174,-0.0115381861,0.1467533261,-0.1777264178,0.4457988739,-0.0647552237,-0.0068596895,0.1231624037,0.3145541847,0.0974895507,-0.0806366727,-0.0708973333,-0.265624702,-0.1158179119,0.0696807653,0.1581618041,0.017750835,-0.2546255291,-0.3468809724,0.0013946585,-0.1944440603,-0.164983511,0.2775953412,0.2337370664,-0.2706082761,0.2563655674,-0.242731601,0.0653673932,0.1290623993,0.2343828082,-0.35971573,0.3342649937,-0.12441127,0.1992957294,0.1714464724,-0.0682586581,0.2153659165,-0.0765552297,0.0236231182,0.3552123308,-0.1429526359,0.333948791,0.4800492227,-0.0572706535,0.3015576303,-0.4770204425,0.1557276249,0.1190635711,-0.160752207,0.3650277555,-0.2998684645,0.0101009971,-0.068044588,0.0613684207,0.1825455427,0.2595223188,0.3625398278,-0.3248929977,-0.2032797784,-0.2347911149,-0.0175856706,0.1589594632,0.1128976122,0.0513959117,-0.4989529848,0.3118694127,0.1154963747,-0.0290463679,-0.0476505943,0.3867293298,-0.2414716333,-0.3074489832,0.2407457083,0.280131489,-0.0437580869,0.3403137922,0.2554681003,0.193293646,-0.0012780592,-0.3980717957,0.1768649668,0.1721522063,0.0027356723,0.0088600554,0.173740834,-0.2945383489,-0.6536637545,-0.0058761858,-0.0075584813,0.0107460693,-0.33941558,0.0236079656,-0.4454859197,-0.6335576773,-0.2669549882,-0.1445803046,-0.3213297427,-0.5371222496,0.1742296368,0.1724796742,-0.3501804471,0.3709135652,-0.3496403992,0.336697042,-0.1685070992,0.2599729002,-0.2313676625,-0.1633770317,-0.2284460366,0.2568376064,0.0770873949,0.1749694496,0.3014180958,-0.0565680042,0.1974834055,-0.6858537197,-0.4707405865,0.1699635386,0.2959243953,0.1817481965,0.1927588433,0.1188537478,0.0080202166,-0.0642928481,0.2318264842,0.0027228142,0.035466224,-0.2211149782,-0.0204486586,0.132950291,0.0118381921,-0.3134480715,-0.1686445475,-0.2678632736,0.0422301255,-0.18909958,0.1295805275,0.1280422658,-0.0340230949,-0.1728270352,-0.1392841786,-0.0964625105,-0.2624019086,-0.3715264201,0.2499987483,-0.4449140728,-0.306435883,0.0055771689,0.1480789334,0.1420973241,-0.0593496151,-0.506249547,-0.2804401815,-0.1307622045,0.0351853259,0.2475958169,0.0029961944,0.1048962176,0.1557350308,-0.3758841455,0.0317999981,-0.3364630044,-0.0537689514,-0.094900772,0.2806492746,0.008397799,0.2198595107,0.2289399207,0.3426800966,0.1634785384,0.0199132524,0.2377622873,-0.1535305381,0.4485100508,0.1821885854,-0.1965880394,0.2710147798,0.0751455203,0.0043195123,0.3426808417,0.268902421,0.098488085,-0.2077539414,-0.1613915116,-0.2719857097,-0.4872804582,-0.0833558068,0.0242726356,0.119744882,0.3425944448,0.0877243504,-0.2497603893,-0.2247463912,0.1520168334,0.3736483753,0.0169874672,-0.0495062321,-0.2696383297,0.2313541174,-0.6547798514,0.2687750757,-0.1115647182,0.1053990424,0.0420822874,-0.1473432779,0.1444115937,-0.0495910235,0.6202213168,-0.0586729646,0.1136013195,-0.0650485232,0.0678817853,-0.2452939302,0.2540537417,0.1758591682,0.1474565566,0.178911984,0.1102554053,-0.1125405058,-0.2522718608,0.3403680623,-0.2710326314,-0.1940625161,-0.1016262844,-0.021062132,-0.2907532752,0.0051707579,-0.0885715261,-0.2042384297,-0.0454114079,0.0387418009,-0.0686731264,0.0498037301,-0.3015016913,-0.0142160775,0.1484458447,0.1374404281,0.2187379748,-0.1437191814,0.3085556924,0.430329591,0.0874608532,0.434345603,-0.1887726933,-0.3583957553,0.0380587168,0.1178704873,0.2680726945,0.4299917221,0.0566061884,0.0496392511,0.3776905835,0.2128890753,-0.1849479824,0.1865593493,0.1830231994,0.0021558416,-0.2337555736,-0.2346233428,0.3446406424,-0.0576439835,-0.1991614848,0.0754687339,0.4092161953,-0.1977351904,-0.2255460769,-0.040585164,0.7620612383,-0.0155285951,0.1551640779,0.3432326317,-0.3480367064,0.4521665871,-0.039280802,0.2377868593,-0.308750242,-0.2042775899,-0.0663167611,0.0424938612,0.0276053809,-0.1109649315,0.1298296005,0.3003414869,0.2962694466,0.1463958323,0.2190443873,0.3614355624,0.0448447652,-0.1108948737,-0.3656885624,0.2949449122,-0.2370052636,0.2156790495,-0.0922102854,-0.125214681,0.1617491394,-0.0149423489,0.0792639852,0.2442039847,0.0020674558,0.0953832045,-0.057151705,-0.2528472245,-0.0088850139,0.2713748217,0.1386504918,0.1052312702,-0.2749758959,0.5181615949,-0.2161436528,0.0573766306,-0.0757090598,0.1377534568,0.0910498798,-0.1569630653,-0.1286019385,-0.0081182374,0.2216854393,-0.2824501693,-0.0727411583,-0.0528901406,0.0015451212,-0.2261842638,0.1129765436,0.0444749445,0.0497734733,-0.3487950563,0.2110318393,-0.0957791656,0.2648498118,-0.0577540807,0.3419021666,-0.0532673225,-0.1077435538,0.4212606847,-0.1554348767,-0.1962436289,0.1823097765,0.1184654981,-0.2897750437,-0.2650914192,-0.0043525123,-0.0290477872,-0.133482486,0.1474136263,-0.0611419529,0.0675660446,-0.0918387845,0.0226206761,-0.0831926838,0.1102432981,-0.1281268448,-0.0382154658,-0.1458683908,0.2807451785,-0.0058235163,0.121396631,0.0242964774,0.3524737954,0.1620064229,-0.3383163512,-0.4731090665,0.1348265409,0.0254057478,-0.2022316009,-0.1570260823,-0.0140348617,0.1987930089,-0.0477130413,0.2098970413,-0.213122949,-0.2175655663,-0.3029687703,-0.1685350239,0.0437042899,0.0954749137,-0.1264076978,-0.0680841506,-0.0920082927,-0.0225775409,0.0399917923,-0.0653442442,0.2169628739,0.0521011725,-0.0644420087,-0.1221574917,-0.3451391459,0.2292209417,0.1107683703,0.1380632371,0.3652173877,-0.0281618573,0.0966598764,-0.2874962091,-0.1066641733,0.0659698695,0.1543349624,0.1804783493,0.2186163515,0.017165361,-0.2897581458,-0.0631611571,0.2186051756,0.4709537923,0.1383360773,-0.1754731983,-0.1330549419,0.2779676914,0.3232357502,-0.3519645035,0.039263349,0.1405555755,-0.0650251061,0.1487740725,-0.1574735641,-0.0791427121,0.1033759639,0.0206994489,0.1116338968,0.2700257003,-0.1199916527,0.1802068502,0.008527698,-0.3254857957,0.0766150057,0.3500171304,0.2305956483,0.1211257055,0.3390010297,-0.0367269553,0.1108646169,0.0562501997,0.2756705284,0.1108814031,-0.0648860559,-0.1411033422,0.0755087286,0.0908265039,0.0157677233,0.186929509,0.497964561,-0.0143686095,-0.1983138621,-0.172334075,0.4549842179,-0.0198510773,-0.1479644477,-0.2496731281,-0.1424888074,-0.043998953,0.2142240703,-0.2048420608,-0.2597120106,-0.0717703253,-0.1286168545,-0.0478472672,-0.5942425728,0.045665767,-0.0341341496,0.2094418406,0.1238220111,0.1417789012,0.0351218581,0.0110922959,0.2486814559,0.3786734939,0.0604696609,-0.1654723883,0.2508488297,-0.1877603531,-0.0965400636,-0.1948823482,-0.213869825,-0.0835697576,0.3052229881,0.0201803576,0.4068004787,0.3149119318,-0.2492127568,0.3233197629,-0.0130927423,0.2501013875,-0.0206806585,0.1084684655,-0.0117508173,-0.1663473994,-0.2375216335,-0.3136681318,-0.3691691756,0.0267784502,0.1721661538,-0.2847143412,0.0880477354,-0.3673914671,0.1497754306,0.2927182317,0.5062050819,0.1465207189,0.0858292207,0.0920263529,-0.515545249,-0.6560676694,0.1436702758,0.0882201716,-0.0828440487,-0.0681077614,-0.0966338441,-0.0902022794,-0.0261958148,0.0687410235,0.1095627323,-0.0556523912,-0.1158264279,-0.186962381,0.162489295,-0.0758871064,0.3305950165,0.0751511976,-0.084387213,0.2293513566,-0.1216669455,0.1155182272,0.0726577044,0.3220477104,-0.2915939391,-0.32259354,0.3025619686,0.0703656152,0.1032533199,-0.0950523317,-0.1500782818,-0.3113046885,-0.4981371462,-0.3935403526,0.1178113893,0.1152288318,0.2336045206,0.1006616503,0.0067488742,0.0358720124,0.2105554193,0.1440186948,0.3561436236,-0.1577115953,0.15701361,0.2726441026,-0.0135753723,0.0609438866,0.1928554177,-0.091266118,-0.0285569932,-0.1198187396,-0.5693752766,0.5659230947,-0.3538654447,-0.0813192576,-0.0747722238,0.2972428799,0.2303742766,-0.0636409596,-0.2993016839,0.0576485023,0.136019066,0.0436314046,0.0388640948,0.2311002761,-0.2023027986,-0.0671898127,-0.1442138702,0.3890945613,0.4058899283,-0.3612388074,-0.0093424637,0.0128212478]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/686","title":"Dataset browser url is still https:\/\/huggingface.co\/nlp\/viewer\/","comments":"This was fixed but forgot to close the issue. cc @lhoestq @yjernite \r\n\r\nThanks @jarednielsen!","body":"Might be worth updating to https:\/\/huggingface.co\/datasets\/viewer\/","comment_length":14,"text":"Dataset browser url is still https:\/\/huggingface.co\/nlp\/viewer\/\nMight be worth updating to https:\/\/huggingface.co\/datasets\/viewer\/\nThis was fixed but forgot to close the issue. cc @lhoestq @yjernite \r\n\r\nThanks @jarednielsen!","embeddings":[-0.1974660605,0.0273436904,-0.0784601718,-0.1032570228,0.181561783,-0.0338328853,0.2075119317,0.2500345409,-0.0712756068,0.104487963,-0.1758001,0.2083590627,0.1768986583,0.2290751189,0.1969033778,0.0059935362,-0.1094174832,0.0742834732,-0.1585035175,-0.0686138496,-0.1156583428,0.153195262,-0.2245044857,0.1394579709,-0.0785325542,0.0543240309,-0.0041663684,0.2298039794,-0.2259248197,-0.2940465212,0.1507101357,-0.0173418783,-0.1685232818,0.2094027549,-0.0000959101,0.1282985806,0.3727542758,0.1688464731,-0.2621804476,-0.0823831111,-0.2555282414,-0.0074984962,0.1870837808,-0.0158358105,-0.1943366826,-0.2524350584,0.3059961796,-0.0592416339,0.1616151333,0.2116993815,0.401886791,0.4169558287,-0.1241875216,-0.1239035949,-0.0596560538,0.0358830504,-0.0895474106,0.0421341322,0.0613159128,-0.0207254384,-0.2333392054,0.4512352645,0.2613551319,-0.0077268556,0.1107012182,-0.0712212697,-0.2978827357,0.0363094732,0.1715471596,-0.0333992429,0.5183126926,0.0327016935,-0.0918525308,-0.0442182757,-0.0863645747,0.0427262969,0.2930809557,0.2634138167,0.1003871635,0.1855985969,-0.2968401015,-0.2738392353,-0.1774041206,0.133595258,-0.1769248098,0.2563844919,-0.1441618651,-0.0435543656,0.1497895122,-0.1700845063,0.1568141729,0.1429754049,-0.0701353401,-0.1276304126,0.0310982559,0.0248226915,0.3437185585,0.4302775562,0.0548315011,0.1577781439,0.204503715,0.1993731558,-0.1671537161,0.0326014198,0.0654028654,0.0515740328,0.1125673428,-0.091020748,0.5377908945,0.1836037636,0.1450007558,-0.0687400401,0.1833945811,-0.3058356047,-0.3549398482,-0.3171661496,0.2748822272,-0.2041047513,-0.1318510771,-0.0670159683,0.0757016242,0.020995805,-0.0373437889,0.4482566714,-0.154953748,-0.1480985284,0.0407034792,-0.0087935682,-0.2628766596,-0.2265728861,-0.3911529481,-0.0244881697,-0.2453728914,-0.1372548342,0.3169980049,-0.0720319077,0.3692349792,-0.0142215369,-0.1194595322,-0.0168606844,-0.1930647045,0.2169735879,0.0279376619,0.2399534285,0.1572362036,0.0713254064,0.1082638428,-0.2228159159,0.180398494,-0.0512345061,-0.1800405383,-0.1564188302,-0.2101799399,0.3378449976,-0.0805729032,-0.20712623,0.056504447,0.1396676153,-0.1483092606,-0.0317349993,-0.1122195274,0.1763474196,-0.2131728679,-0.010816358,0.1252395213,0.1734618247,-0.0859159008,-0.2365061343,0.0957480967,-0.2298832834,-0.0901157707,0.2802179456,-0.251814276,-0.1694702655,-0.3509841859,0.3048020899,0.3762409687,-0.1676816791,-0.4185831547,0.1490959525,-0.2779150307,-0.2218320817,-0.0990791172,0.0020068439,0.0110543352,0.0777365118,0.0809618086,0.1376070678,0.1807423383,0.2409165204,-0.3028908968,-0.2009304762,-0.0469012633,0.0132408654,0.0301344991,0.0239060465,0.1576497704,-0.1619078815,0.4284245968,-0.0154245347,0.0119457478,0.2059180737,0.2590494752,0.0803477243,-0.0480143875,-0.065082781,-0.2406261414,-0.0328416042,-0.0519442633,0.2378103882,-0.0058331727,-0.156351313,-0.345670253,-0.021460833,-0.2990147471,-0.2557522058,0.3288921118,0.1464810371,-0.1810379326,0.2520053387,-0.0942678899,0.1360912472,0.1611603945,0.1647181958,-0.3535585403,0.2865383923,-0.1542076319,0.0727541968,0.2222292274,-0.0629097298,0.2361103594,-0.0257163011,-0.0894754753,0.3488487601,-0.1196320429,0.3738693297,0.4263543487,-0.0244428143,0.2084365636,-0.4521690905,0.0737857819,0.0894138515,-0.0494929552,0.2450299859,-0.2935980856,0.096330002,0.0569775514,0.0592544563,0.2151220292,0.1758551002,0.3895345628,-0.2599368989,-0.1002948359,-0.2663059831,0.1481195837,0.147216484,0.1474496573,0.0232875831,-0.5029736757,0.2252423465,0.148873955,0.1364866942,-0.0624672398,0.2977840006,-0.321867913,-0.2559401989,0.2251434773,0.158508271,-0.006903985,0.3093116283,0.2602862418,0.1855838299,-0.1012605354,-0.4158044159,0.1550372541,0.0775817111,-0.0206567887,0.1006702706,0.2406623363,-0.3285664916,-0.8028116226,0.1236182824,-0.0710889846,0.0160447117,-0.3454656899,-0.0351603031,-0.488429755,-0.5637759566,-0.1451428831,-0.1193673611,-0.3447198868,-0.5111808181,0.209605068,0.243634522,-0.223038733,0.3255797029,-0.2251421809,0.3194464743,-0.0139513062,0.2616069913,-0.1516726017,-0.1704344898,-0.2728944123,0.263074249,-0.0522389896,0.1690321267,0.3369331956,-0.1240859851,0.1483501345,-0.5392875671,-0.4979757965,0.1466765553,0.272112757,0.2357880473,0.1745633185,0.2735047042,-0.0065587829,0.0095047951,0.29240641,-0.0310326889,-0.1246151254,-0.1360731572,0.0024792412,0.0782557577,-0.1521027386,-0.2140236646,0.0028284674,-0.3118531704,0.1058293879,-0.1396290213,0.0617649592,0.1695275158,0.0230551455,-0.0627425164,-0.1943280399,-0.010065536,-0.3440771699,-0.324375391,0.2471306324,-0.4713352323,-0.4756580293,0.0069068782,0.2107207179,0.1817521155,-0.0875607952,-0.4407213926,-0.2456109524,-0.1737812608,0.0558804497,0.2376805097,0.0277899988,0.1869580448,0.0396152958,-0.3498468995,-0.1598076969,-0.3615305424,0.0500346199,-0.1396215111,0.2461443394,-0.1019043401,0.2139475048,0.1814713478,0.3139735758,0.1883889884,-0.0042371345,0.3592143655,-0.2011668384,0.4598397315,0.0230813231,-0.2174991369,0.1868271232,0.0777249187,0.1025524065,0.3146655858,0.287132889,0.0138680376,-0.1832826734,-0.1223619655,-0.2982822955,-0.387178123,-0.1333837807,0.120830819,0.1195249557,0.4041796923,0.0985241234,-0.2048654854,-0.2593629062,0.1359426379,0.2581130862,-0.0252667442,-0.071333833,-0.2248267084,0.2221825123,-0.637063086,0.2313666195,-0.126742363,0.093814835,-0.1146973372,-0.1444635987,0.2424203753,-0.0305138323,0.6903861165,-0.0589891598,0.1687088907,-0.0796724632,-0.0389771201,-0.2865443826,0.2066481709,0.0473707169,0.216582194,0.2075992972,0.2117219716,-0.0798943266,-0.1644723415,0.1647363454,-0.2278536409,-0.1649111658,-0.1120238155,-0.0951729044,-0.1904585958,-0.0772356763,-0.106568858,-0.0804794282,-0.0203400925,0.0121811209,-0.0971115306,-0.0566086471,-0.221495837,-0.0685617402,0.167021066,0.1722956747,0.1537458003,-0.0183098,0.2694332004,0.4162510335,0.0434971079,0.4411606789,-0.0618641414,-0.3070908785,0.0651720837,0.0260470826,0.207272023,0.3419608474,0.0080391085,-0.0124965785,0.2996070385,0.2353086919,-0.1479694992,0.2208895683,0.2598156035,-0.0002156106,-0.0682832152,-0.1415542811,0.3250381052,-0.0959904119,-0.1430975497,0.1209715679,0.3451144993,-0.1537528485,-0.1891282946,-0.1212928519,0.7177481055,0.070479624,0.0990132466,0.3580032289,-0.2905935943,0.34465307,0.0518434234,0.2566057146,-0.3744661808,-0.2412984818,-0.0324432179,-0.0137491301,-0.0178163573,-0.1288865805,0.1598843187,0.1130844504,0.2826024294,0.138179794,0.1533910036,0.2530760169,0.0002257908,-0.0751468986,-0.3149949014,0.3230692148,-0.1355199516,0.2502849102,-0.0897034779,-0.1414242089,0.0822522864,-0.1100846827,0.007494403,0.3172851503,-0.0048320498,0.22215949,-0.0119919349,-0.216378659,0.0186017249,0.2947873771,0.1496807337,0.1280510128,-0.3559486866,0.5357000232,-0.1166999564,0.0166328531,-0.1068717763,0.233327046,0.0663204491,-0.2210595012,-0.1496175379,-0.0689432099,0.1177704185,-0.1777958572,-0.0917475075,-0.0056935158,-0.0917626172,-0.2471944094,0.0940962806,-0.017563235,0.1372779459,-0.3961023688,0.2428833991,-0.0468099676,0.2616125047,-0.1094832048,0.3743314147,-0.0868424773,-0.133857429,0.5016450882,-0.1606318504,-0.1081101298,0.2218483835,0.1497847736,-0.3185872436,-0.2778306603,-0.094901897,-0.0188943073,-0.2196323425,0.1873591095,-0.0137456339,0.0496612228,-0.0338049717,0.0999610424,-0.1409194618,0.0170827117,-0.07321468,-0.1446515471,-0.0928106382,0.1306103468,-0.1439695209,0.0942751393,0.0816676021,0.3466886282,0.2421336919,-0.2156917304,-0.4751226306,0.1386797726,-0.0766266435,-0.1673935056,-0.0832478404,-0.106690973,0.1977979094,-0.039603617,0.2327354997,-0.0665189922,-0.3025290072,-0.3437770009,-0.1553218365,0.0249250066,0.1081616729,-0.1740917265,-0.0455245264,0.0145715149,-0.1190724447,0.0031785676,-0.012049499,0.1558796167,0.0164955128,-0.0937587172,-0.0943586677,-0.2766507268,0.301554054,-0.0431516431,0.1119098961,0.2310573459,-0.0277925376,-0.1102184653,-0.3132819533,-0.0867305174,0.0566490702,0.0813306123,0.080271855,0.2652200162,0.0929904133,-0.2890507281,-0.027740255,0.149767682,0.4162068367,0.0300273802,-0.1694955975,-0.0977861881,0.2208977044,0.3564222157,-0.4050797224,-0.0020332905,0.2132008076,-0.0612955615,0.2013937384,-0.1754217148,-0.0248048734,0.1014504284,-0.0394261591,0.1695090681,0.2606306374,-0.1044750288,0.0647063777,-0.0401303358,-0.3225340545,0.0469818898,0.3188965619,0.2050915807,0.0931429192,0.3705171943,0.0382943936,0.0341243334,-0.0880159885,0.1793792546,0.0921193287,-0.1680691689,-0.1290023178,0.0845367089,-0.0011249684,0.0131163141,0.1349214911,0.4328964949,-0.1351351887,-0.1299032718,-0.2418042868,0.4619630277,-0.1506508589,-0.1110696867,-0.2960782051,-0.1463870406,-0.0114867138,0.2243330032,-0.1389539987,-0.09758389,-0.044652611,-0.1895443499,-0.129193306,-0.604567945,0.0254800022,-0.0207140483,0.185785383,0.0685232803,0.2162698209,-0.0314482301,0.0223948136,0.173094824,0.3657982349,0.1764479727,-0.0333605781,0.2814880908,-0.1389557272,-0.0623414144,-0.2161011845,-0.2258041352,0.0258338787,0.3186942041,0.0795325711,0.4054733217,0.299836874,-0.2988408506,0.2461194396,-0.0305343233,0.3594302833,0.0234926939,0.0148500325,-0.1371868253,-0.1350031793,-0.2627707124,-0.3503129482,-0.3947630227,0.0951644406,0.1055137962,-0.1969668418,0.1259727031,-0.4530036449,0.1553248018,0.2195655704,0.567620635,0.2256785929,0.0764167458,0.1508321315,-0.5539203882,-0.6505864859,0.1814620793,0.044435747,-0.0702971444,-0.1433730572,-0.0330340862,-0.0848247632,-0.0788952336,0.1709430665,0.1131158397,-0.0090197055,-0.2403028458,-0.2492431253,0.1814006716,-0.0292069782,0.2143139392,0.1185635105,-0.0940899253,0.2352159619,-0.1139642149,0.217145294,0.0486917272,0.3396807909,-0.4147003889,-0.3355849087,0.2693334222,0.1099150702,0.1559266597,-0.1001498997,-0.0957681388,-0.2673412561,-0.421908021,-0.2950202227,0.1550862044,0.131259948,0.2896251678,0.130605191,-0.0490837768,0.0515473969,0.2216097265,0.1836282909,0.2618759274,-0.1320513338,0.1721829176,0.1324615479,0.0265303385,0.0231165383,0.1753365993,-0.0885616094,-0.048725348,-0.1355973631,-0.6341204047,0.48963359,-0.3542748392,-0.1450948119,0.010684683,0.2051821351,0.1848870963,-0.1796301901,-0.3368443847,0.0978056639,0.1487629861,0.0152387973,0.0124241114,0.247321263,-0.2656150758,-0.053168498,-0.0950677171,0.4360678792,0.4730352163,-0.3662415445,0.1386486739,-0.003336502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/678","title":"The download instructions for c4 datasets are not contained in the error message","comments":"Good catch !\r\nIndeed the `@property` is missing.\r\n\r\nFeel free to open a PR :)","body":"The manual download instructions are not clear \r\n```The dataset c4 with config en requires manual data. \r\n Please follow the manual download instructions: <bound method C4.manual_download_instructions of <datasets_modules.datasets.c4.830b0c218bd41fed439812c8dd19dbd4767d2a3faa385eb695cf8666c982b1b3.c4.C4 object at 0x7ff8c5969760>>. \r\n Manual data can be loaded with `datasets.load_dataset(c4, data_dir='<path\/to\/manual\/data>')\r\n```\r\n\r\nEither `@property` could be added to C4.manual_download_instrcutions (or make it a real property), or the manual_download_instructions function needs to be called I think.\r\n\r\nLet me know if you want a PR for this, but I'm not sure which possible fix is the correct one.","comment_length":15,"text":"The download instructions for c4 datasets are not contained in the error message\nThe manual download instructions are not clear \r\n```The dataset c4 with config en requires manual data. \r\n Please follow the manual download instructions: <bound method C4.manual_download_instructions of <datasets_modules.datasets.c4.830b0c218bd41fed439812c8dd19dbd4767d2a3faa385eb695cf8666c982b1b3.c4.C4 object at 0x7ff8c5969760>>. \r\n Manual data can be loaded with `datasets.load_dataset(c4, data_dir='<path\/to\/manual\/data>')\r\n```\r\n\r\nEither `@property` could be added to C4.manual_download_instrcutions (or make it a real property), or the manual_download_instructions function needs to be called I think.\r\n\r\nLet me know if you want a PR for this, but I'm not sure which possible fix is the correct one.\nGood catch !\r\nIndeed the `@property` is missing.\r\n\r\nFeel free to open a PR :)","embeddings":[-0.2952925265,-0.1274301261,-0.1135854423,0.0708296672,0.2947362363,0.0913632959,0.1110735014,0.2096318156,-0.0788623765,0.1567240804,0.1122687683,0.2098814249,-0.0879549459,0.3875023425,-0.0222318582,0.0357285887,-0.2305764854,0.1732322574,-0.1882077754,0.0362125225,-0.0883256346,0.1371610016,-0.1378411651,0.0431257188,-0.1025673822,-0.0522560962,0.0911326855,0.2248452753,-0.131732896,-0.4425009787,0.3559399545,0.143772006,0.3937465847,0.0451206081,-0.0000938641,0.1411955655,0.2274561077,-0.1976596266,-0.2515526116,0.1758781224,-0.2253013998,-0.0594678447,0.0868790299,-0.280577749,-0.1739056855,0.0538876951,0.0915005803,-0.1386648566,0.0949869081,0.3721596003,0.4034843445,0.1437874585,0.191760093,-0.3148843348,0.0110757537,-0.1719944924,-0.0858436003,0.1718982905,0.1063720733,0.2503721714,0.2745094597,0.2165531814,0.189129293,0.0806846842,0.2857280672,-0.0794256181,0.0880433694,-0.2238693833,0.149920553,0.057664644,0.8168542385,-0.1879398823,-0.1010043919,0.1536811739,0.3333650827,-0.1642507166,0.1092226654,0.1731310189,-0.0236431099,0.1634045094,-0.2989314795,-0.293723166,-0.1688118428,-0.0074193319,-0.0459048897,-0.0008645593,-0.1846629679,-0.0471602231,0.0488130227,-0.0409464911,0.1524113864,0.0963839889,-0.2392666042,0.1344270408,0.0781139806,-0.1756483316,0.0404920131,0.3301278949,0.0258282945,0.0065668249,0.0842241049,0.0686276257,-0.0528978854,0.1270597577,0.0847075805,-0.1314075887,0.1438766122,0.2669115365,0.2410773337,0.1134388223,0.0030000112,-0.0759418905,-0.167848736,-0.2088225335,-0.1654671133,0.1303621829,0.0738515481,-0.2450197637,-0.3865617514,0.0805436373,-0.0979762077,0.0877509266,-0.1605391502,0.3545568585,-0.1133193821,-0.135311529,0.1627790034,0.0752347857,-0.0501861386,-0.2587626874,-0.3132553101,0.0565595068,-0.2274860591,-0.096283935,0.2555839717,0.0069883224,0.348084718,-0.0533080287,0.0385149494,0.0977297872,0.0625968874,0.0795601457,-0.3223217428,0.2906909585,0.0997691303,0.0222854856,0.1109223366,-0.1382524073,-0.0602918938,-0.0050689289,-0.0841321051,-0.4475284219,-0.0457093492,0.4004327953,0.0117915599,-0.1345737129,-0.1688542366,-0.0649744347,-0.0321873091,-0.1209722906,-0.1212418228,-0.2032362372,0.1658848226,-0.224857986,0.1895277798,0.4165790081,-0.4708591104,0.0804208145,0.0241042245,-0.286244601,0.1587120295,-0.0538099706,-0.3435450792,0.2462607175,-0.2540868521,0.0129749961,0.0195657499,-0.1926914454,-0.2476391792,0.2590120733,-0.090809077,-0.2682569027,-0.1800188869,0.0476541407,0.0699188113,0.0596556887,0.0657015517,0.1099572629,-0.0253978781,0.0016468918,-0.207278192,-0.2175521106,0.04963829,0.0392934419,0.0816996843,0.0089751305,0.2493124157,-0.0673993304,0.1978165209,-0.1398575157,0.0594883971,0.3982246816,0.1247814521,0.1764263958,-0.2655292153,0.1074199006,-0.4236545563,0.138012737,0.0316912271,0.222186625,0.1536938399,-0.0614108741,-0.2698266208,-0.2186504006,-0.2104483992,-0.32004866,0.3388337791,0.2055804878,-0.2076474726,-0.0500928573,0.068965435,0.4813523293,-0.3558864594,-0.0145559032,0.0101027628,0.1900322288,-0.3723829687,-0.0109958723,0.1352766752,-0.1734652519,0.1876429319,-0.0766421705,-0.1780222654,0.3868669569,-0.0223936867,0.0703889132,0.1814189106,0.0880919471,0.0806779638,-0.3901180029,0.0723966062,0.1072691083,0.1716867834,-0.0301120747,-0.1045943052,0.1489655375,0.026138721,-0.1028372794,0.1922296882,0.0867937058,0.3510356545,-0.0531119555,-0.0670326501,-0.1744185686,-0.0929188058,0.1219271645,0.1601719409,-0.153531149,-0.4608395994,-0.0183329508,0.2677019238,0.2692942023,-0.048124209,-0.0184236169,-0.1127979755,-0.0396463759,0.2409429401,0.3904123008,0.4621816278,0.3583667278,-0.0902841315,0.1028370559,-0.0989274457,-0.1665301919,0.2769759893,-0.2360740453,0.0766933784,-0.0100442581,-0.1796915382,0.0740877166,-0.2662764788,-0.2713891864,0.0707096756,0.1121157408,-0.3719191253,-0.2583868206,-0.2335111201,0.038343899,-0.0546552204,-0.2035372257,0.0084853387,-0.2680447996,0.0237966795,-0.0856864974,-0.2056819797,0.1266689897,-0.4074525237,0.0943043977,-0.0145385638,-0.0337636136,0.0593939386,-0.0157890823,-0.3294896483,0.2444536835,0.0478815027,0.1672565192,0.3354514539,-0.2728863955,0.0769898221,-0.3132178783,-0.3293321729,0.1319431812,-0.0139026092,0.1548405737,0.2868923247,0.3454277217,-0.0200399291,-0.1642941087,0.1777035445,-0.0676172972,-0.0803072751,-0.1428566426,-0.0234906282,-0.2711684108,-0.254339397,-0.6514555216,-0.4080271721,-0.387583375,-0.0587242842,-0.0408107378,0.0094668381,0.4306851625,0.2378659695,0.1363740861,0.1561000496,-0.0140108569,-0.3094950318,-0.3086686134,0.2638241649,-0.435118258,-0.2814389765,0.2458922863,0.1639101058,0.2736424506,0.1124274433,-0.3028967977,0.165701896,-0.1467757821,0.2527734935,-0.1490262449,0.0568699352,0.4291028976,0.0490832515,-0.192098245,-0.2636856437,-0.2061094046,0.2330669314,0.0097463652,0.3035892844,-0.3539081812,0.0888256207,-0.082161665,0.4178382158,-0.1425989866,-0.2592869103,0.3099949658,-0.0748697966,0.5479115844,-0.0035315435,-0.0974276364,0.2020566612,-0.0357910022,-0.0505725145,0.3431040347,-0.1990187317,-0.2512921095,-0.2705321908,0.0943482295,-0.2098183334,-0.137264356,-0.1280497015,0.1379722506,0.2824385166,0.1460064948,0.1212734878,-0.0238222089,0.121232748,0.0317086093,0.2491375953,0.0066711917,0.1008310392,-0.2829979658,-0.1761524528,-0.129693374,0.2367837727,-0.0202393625,0.2252919525,-0.1107786968,0.0262964256,-0.109706372,-0.1041380167,0.1723397225,-0.3338513374,-0.005550134,0.1862294674,0.2565667927,-0.1348627508,-0.2305299938,-0.2140298486,0.2606616318,0.359077096,0.1068173721,-0.1634493917,-0.0062134438,0.0997924805,-0.0412271619,-0.1040333286,-0.0564130135,-0.0513513424,-0.1530238539,-0.3319344223,0.0477227084,-0.0051995604,0.021796748,-0.2011698335,-0.1668344587,0.0718733445,0.1020668894,0.0059622061,0.1362481713,0.3738857508,-0.001772466,0.2168974876,0.0941227004,0.0984304994,0.5368898511,0.44797346,-0.1147343367,0.0696465522,0.0959386155,-0.2784751356,0.1975340992,0.1984547377,-0.0647660345,-0.2408042103,0.0585557409,0.1606145501,0.0237346217,0.2814680934,0.2125218809,-0.1988380104,-0.1813612431,-0.3780083358,0.2156289816,-0.2403035462,0.0017788182,0.1725716144,0.2052414566,-0.3346079886,-0.0224948116,0.2057650983,0.7688063979,0.2329611629,-0.0182664599,0.0582374409,-0.2121347785,0.083017759,-0.2897780836,0.0095688207,-0.2246858031,-0.0756369382,-0.0318175219,-0.0437695757,0.1898258775,0.265894413,-0.3571045101,0.1258708984,-0.057526987,0.1043520868,0.0555396639,0.1110893786,-0.0446675047,-0.0359606706,-0.1016631126,0.3267787099,-0.0888665989,0.2303604782,-0.2770774364,-0.2785871923,-0.0524114184,-0.272292614,-0.2474131137,0.1907442957,-0.2851499617,0.0868133679,-0.0042576417,0.1123535708,0.3332913518,0.0493767411,0.2300221175,0.3571946919,-0.3989551067,0.2011287212,0.0438202992,-0.2820204496,-0.1219434738,0.1810658872,0.2008391023,-0.2149366438,-0.2505923808,0.0652051717,-0.1483434588,-0.1720426083,0.029820662,-0.1384863853,0.058410652,-0.2664520741,-0.1122318879,0.1875180602,-0.147525087,-0.15597637,0.3168507218,-0.0526291914,-0.0727544948,0.2589355707,0.3189015388,-0.1281951368,-0.1235658005,0.259731859,-0.1791674942,0.0032544814,0.4107590318,-0.0871477947,-0.2052238733,-0.4251074791,-0.0747680143,-0.0888200104,-0.2961089015,-0.0503999107,0.0601064898,0.0171492882,0.1987743378,0.2797046304,0.164357841,-0.1016547978,0.184419021,-0.3283883035,-0.4239746034,0.0620470829,-0.2974878848,0.3756355047,-0.0996582285,0.4513445199,0.3099474013,-0.0258254167,-0.5571076274,0.058957275,-0.2986157537,-0.0191736128,0.2772828937,0.0253537986,0.0496468917,-0.0510816313,0.2824494839,0.149531886,-0.2774735093,-0.3205429316,-0.0104805939,0.0632289425,-0.0112866713,0.0427569151,0.1938339025,-0.1259935945,-0.3876332045,0.137242347,0.0289683361,0.0868571773,0.0087913908,0.2562069893,0.2374175787,0.1578209251,-0.1241822466,0.0148789734,0.2006451488,0.335167408,-0.0871025398,0.1256008893,-0.2146410793,-0.1034682244,0.1116388813,0.0699548349,0.0974765196,0.1989020258,0.108093977,-0.3085347116,0.1577941775,0.1264507174,0.0533504523,0.1908809543,-0.1451033801,0.07526353,0.1876571923,0.461645335,-0.4281205535,0.0538554266,-0.2618232965,-0.0148525313,-0.0256375428,-0.1163142398,0.2491564453,-0.1593383253,0.1663857996,0.2423924655,0.2042932659,-0.1009972617,-0.0006021188,0.2360047102,-0.2113386542,0.0030071605,0.1253673881,0.0289865546,0.114989765,0.4669352472,-0.079380855,-0.1202526465,0.1179667413,0.0541600287,0.0719345957,-0.5321770906,0.0264738873,0.218842417,0.0727425069,0.037773259,0.0748429596,0.2073807567,-0.099852927,0.0064115035,-0.2664357126,-0.0000012437,-0.2774181068,0.1265488118,-0.1524086446,-0.2897831202,-0.1268594414,0.0805559456,-0.0509897321,-0.0630458891,0.2491436303,-0.0408362746,-0.0483604632,-0.4156670868,0.0566806346,0.0661977381,-0.0227176566,-0.3296530545,0.2532843053,0.106669724,-0.1933308244,0.1415152699,0.1459695101,0.4041513801,0.3569764197,0.0852710232,0.1064156368,0.1866198927,-0.0933036134,-0.0474734418,0.0947546735,0.3691647947,0.0019585867,0.3078721464,0.3121574521,-0.2528561056,0.1171253845,-0.0325172395,0.1404514462,-0.039862778,0.1727115512,0.1886358708,-0.167920351,-0.0991174281,-0.0001495858,-0.3141236603,0.0415555462,0.2325172275,0.0819706023,0.1070414856,-0.173423931,0.2024026364,-0.0781998709,0.6107167602,0.1800044775,0.2917609215,-0.3144370019,-0.3652966917,-0.5705239177,0.0635299906,-0.1339280456,-0.0468859859,0.003095936,0.1522558033,0.2721349299,-0.0483174101,0.2243812978,0.243867889,0.2713009417,0.1631766856,-0.2638369203,-0.1335407346,0.1034183353,-0.0346105397,0.1267110109,-0.3644719422,-0.1800686419,-0.0160902422,0.2515571117,-0.310885489,0.0980365053,0.1693420112,0.1797309518,0.5669074059,-0.0797521621,0.6776628494,-0.2189651132,-0.1290173829,-0.1694681495,-0.0989269912,-0.2127552927,-0.1834476441,0.3084415793,0.1379079372,-0.0276605226,0.0041789636,-0.4089870751,0.1579998732,0.2589384913,0.3046479523,-0.265781343,0.1670839787,-0.1956088543,-0.0727322102,-0.0270441752,0.0656324849,-0.0450523645,-0.1307168007,-0.356697619,-0.4053038359,0.6265100241,-0.0702631474,-0.1206315383,0.0032990007,0.1856639534,0.011107306,-0.1001905575,-0.3287264109,0.2859677374,0.3074204922,-0.1295980513,-0.2068364769,0.2071796209,0.0252585225,-0.0200975612,-0.0380547419,0.5409080982,0.1175973564,-0.0519334674,-0.3357287943,-0.3130464256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/678","title":"The download instructions for c4 datasets are not contained in the error message","comments":"Also not that C4 is a dataset that needs an Apache Beam runtime to be generated.\r\nFor example Dataflow, Spark, Flink etc.\r\n\r\nUsually we generate the dataset on our side once and for all, but we haven't done it for C4 yet.\r\nMore info about beam datasets [here](https:\/\/huggingface.co\/docs\/datasets\/beam_dataset.html)\r\n\r\nLet me know if you have any questions","body":"The manual download instructions are not clear \r\n```The dataset c4 with config en requires manual data. \r\n Please follow the manual download instructions: <bound method C4.manual_download_instructions of <datasets_modules.datasets.c4.830b0c218bd41fed439812c8dd19dbd4767d2a3faa385eb695cf8666c982b1b3.c4.C4 object at 0x7ff8c5969760>>. \r\n Manual data can be loaded with `datasets.load_dataset(c4, data_dir='<path\/to\/manual\/data>')\r\n```\r\n\r\nEither `@property` could be added to C4.manual_download_instrcutions (or make it a real property), or the manual_download_instructions function needs to be called I think.\r\n\r\nLet me know if you want a PR for this, but I'm not sure which possible fix is the correct one.","comment_length":56,"text":"The download instructions for c4 datasets are not contained in the error message\nThe manual download instructions are not clear \r\n```The dataset c4 with config en requires manual data. \r\n Please follow the manual download instructions: <bound method C4.manual_download_instructions of <datasets_modules.datasets.c4.830b0c218bd41fed439812c8dd19dbd4767d2a3faa385eb695cf8666c982b1b3.c4.C4 object at 0x7ff8c5969760>>. \r\n Manual data can be loaded with `datasets.load_dataset(c4, data_dir='<path\/to\/manual\/data>')\r\n```\r\n\r\nEither `@property` could be added to C4.manual_download_instrcutions (or make it a real property), or the manual_download_instructions function needs to be called I think.\r\n\r\nLet me know if you want a PR for this, but I'm not sure which possible fix is the correct one.\nAlso not that C4 is a dataset that needs an Apache Beam runtime to be generated.\r\nFor example Dataflow, Spark, Flink etc.\r\n\r\nUsually we generate the dataset on our side once and for all, but we haven't done it for C4 yet.\r\nMore info about beam datasets [here](https:\/\/huggingface.co\/docs\/datasets\/beam_dataset.html)\r\n\r\nLet me know if you have any questions","embeddings":[-0.1576072276,-0.2520465553,-0.0443442911,0.2419738322,0.3052386045,-0.0173060261,0.0462005883,0.1740138978,-0.0926164016,0.1525098532,0.011208185,-0.0311050825,-0.1280990988,0.579795897,-0.0943659544,-0.2361066043,-0.1385877132,0.1183631122,-0.2195278853,-0.0132202459,-0.0061009866,0.2214009762,-0.1956916749,-0.0220425092,-0.2092388719,-0.1372852623,0.0151895769,0.1767092049,-0.25517115,-0.3560912907,0.3604578972,0.1128410026,0.2981440425,0.1601436138,-0.0000945143,0.0746880248,0.2518620491,-0.302344203,-0.284550488,0.1466871798,-0.0307056513,-0.0956450775,0.0670619756,-0.1968445182,-0.2997147441,-0.1451850832,0.1192829385,-0.1014602035,0.3640060127,0.2599665523,0.372258395,0.4407114983,0.195772022,-0.2333814949,0.0953537077,-0.1508592516,-0.1420059353,0.1691168994,-0.0406183712,0.2747166157,0.0004316543,0.2728057206,0.1778380871,0.0580458343,0.4070051908,-0.2139707059,-0.1956558675,-0.2863411903,0.1214683652,0.0440186225,0.7800216675,-0.1754940301,-0.0318887569,0.0684108809,0.3036718965,-0.0913942531,0.2903752327,0.2816475034,-0.1135262698,0.1321048588,-0.2517784536,-0.1577833593,-0.2102700472,-0.0275425036,0.1293176264,-0.2587459981,-0.1821366996,0.1505683213,0.1448352039,-0.0655170009,-0.0370307676,0.0024909708,-0.2538126707,0.1177524254,-0.2260058373,-0.1340099126,-0.1367894709,0.5243659616,0.1480009854,0.0493390597,0.0552521087,0.0395027064,-0.0958476365,0.0246077385,-0.0875779167,0.1341896057,0.1416077167,-0.0273241121,0.2056231797,0.0090816459,0.1363287568,-0.1324751228,-0.0923999473,-0.3220089972,-0.4273110032,0.1215066835,-0.1025412306,-0.1251258552,-0.2567414343,-0.0216875579,-0.1581317782,0.1170150787,-0.1226818711,0.502887249,0.0319798253,0.0357804447,0.3110114634,0.1301048398,-0.1264162809,-0.4252971411,-0.1609552205,0.0885234997,-0.3406014442,0.1032325551,0.2031144798,-0.0858152285,0.3210023642,-0.0105482349,0.203918308,0.2336816043,0.043708276,0.1382245421,-0.3299477994,0.2939603627,0.0892609581,0.1966176778,0.0503919981,0.0992412493,0.0147278104,-0.242895171,-0.1383904517,-0.38889727,-0.0293127373,0.2899577916,-0.0359732173,0.0604546554,-0.3929697275,-0.1538803875,-0.0779550523,0.0085782697,-0.018533932,-0.1841296703,0.1792027056,-0.3099900186,0.1034423634,0.5316656232,-0.193946287,0.0594609007,0.107191287,-0.3290980458,0.0939717069,0.1096454486,-0.1983128786,0.1528692544,-0.3218123615,-0.0717164576,0.0001487405,-0.2997231185,-0.0671718419,0.2491941899,0.0763702989,-0.1028418913,-0.0841958225,0.0457970314,0.0344794653,0.0471420623,-0.1555431336,0.0964351818,0.0459185466,0.0953812897,-0.2365958691,-0.2664739788,-0.0504299514,-0.0170552824,0.0486117378,0.0009503769,0.1752432585,-0.0223728772,0.2334792614,-0.1215633005,0.2053255886,0.3341152668,0.0897393376,0.0726901889,-0.0986366346,0.270585686,-0.4880263209,0.1984877437,-0.023317311,0.2936452329,0.2374486923,-0.0921918899,-0.2881412506,-0.1170640588,-0.1954953671,-0.392844677,0.233316198,0.0638065636,0.0160966571,-0.1424281895,-0.055114679,0.6949421167,-0.4194718599,0.136083737,-0.1755590439,0.2874920666,-0.2992298901,-0.0015135614,0.2668257058,-0.1448068172,0.2442115694,-0.1558156163,-0.1182781532,0.5188832283,-0.0700506046,0.1831248552,-0.0156595241,0.102499187,0.1684033424,-0.4067897499,0.2226454169,0.0969818383,0.0564252883,0.0465730317,0.1368783414,0.027752284,0.0609656163,-0.0385046788,0.1375764608,0.2114509344,0.3035189509,0.0903565884,-0.0515104011,-0.008439864,0.0460292585,0.0381696224,0.3691146672,-0.1547346562,-0.2686443925,-0.0204406343,0.372458607,0.1492622942,-0.2134382725,-0.0039702882,-0.2015951276,0.018254986,0.4242030084,0.0845953524,0.1815245748,0.3259015977,-0.2572725415,0.1394944936,0.0787117854,-0.1675015539,0.2197905928,-0.2067821026,0.173573032,-0.076530695,-0.0619713515,0.0288041923,-0.2854051292,-0.1263670325,-0.0495843105,0.0670863912,-0.4333347082,-0.1874492019,-0.07151068,-0.133476302,-0.2219575495,0.1057161689,-0.2684016824,-0.2493634522,0.0531169921,0.0519354008,-0.2146594524,0.2211509943,-0.0773784444,0.2021472454,0.1870241612,0.0604863875,-0.2790496349,0.0561805442,-0.3365766704,0.1974057108,0.0763333738,0.1776463687,0.3503883481,-0.2926580012,0.1538708955,-0.4253621697,-0.2524706423,0.2706113458,-0.1242211014,0.0341452211,0.2647260725,0.4457345605,-0.0405968353,-0.1978025287,0.1048382148,-0.2580126226,-0.2056209147,-0.0564710908,-0.0314116552,-0.1052621827,-0.2328078747,-0.4738240838,-0.2691175044,-0.5012937784,0.061664734,0.0238951892,-0.0179634243,0.3495784998,0.1636542827,0.0834817737,0.1862342507,-0.132588014,-0.1996641308,-0.25761801,0.2411109805,-0.4244994521,-0.4918941855,0.2326402962,0.0886932164,0.1810022593,0.1364176273,-0.2883037627,0.1077371091,-0.1272397339,0.2177413702,0.0037632859,0.0890457481,0.4727257192,-0.105982244,-0.0791258886,-0.2712836862,-0.0139193693,0.1867629588,-0.1283687949,0.4082902074,-0.141553849,0.092208609,0.0083458927,0.6587253213,0.0684503093,-0.2738857269,0.3950152695,-0.024522122,0.5483447909,-0.1196088716,-0.0483414382,0.2783370912,-0.1303411275,0.1182845756,0.256187886,-0.0717047974,0.0530204959,-0.2930820584,0.1332418174,-0.3371211588,-0.203250587,-0.0752009228,0.133770436,0.2385990769,0.1224342957,0.2099100053,0.0756298676,-0.0600336157,0.2200852633,0.3244325519,0.0945592895,0.0862483755,-0.097725898,-0.0006392779,-0.4687542915,0.4585257471,-0.0882852226,0.0121188555,-0.0354761444,-0.048440598,0.0560801476,-0.0770029053,0.3993708491,-0.3509088457,-0.0245684441,0.0671608374,0.3152973652,-0.3512361944,-0.1357654184,-0.147463873,0.0266345497,0.3536953032,0.1129958481,-0.3503396511,-0.1448244601,0.050570149,0.0257965643,-0.1164405048,-0.0779691786,-0.1261963397,-0.0568869226,-0.3180392385,-0.0524927452,-0.0564878471,0.064365007,-0.191808477,-0.0424596332,0.068495445,-0.0226399973,0.027367672,0.1267912984,0.3164551258,0.0824907497,0.1086835638,0.1601500064,0.1868722588,0.3519448042,0.2604799867,-0.0664425343,0.1868506819,0.1660544425,-0.0529631265,0.0243413951,0.2211947441,-0.0637338385,-0.0417155884,-0.1308413446,0.197663486,-0.155297935,0.3797974885,0.1449739039,-0.1973758638,-0.2799603939,-0.3411352336,0.3049995005,-0.3228604198,-0.0193392336,0.2203385234,0.5354681611,-0.2219850272,0.175055638,0.1773684025,0.805611968,-0.1524404585,-0.0300409831,0.2842763066,-0.1223610491,0.2576764822,-0.6670987606,0.0283860657,-0.2446983755,0.0566469058,0.0189890582,-0.0293506328,0.0844151005,0.1341787577,-0.2023187131,0.205209583,0.1579805911,0.2810953259,-0.0964582264,0.2312203199,-0.2328831553,-0.1702729762,-0.230320096,0.3361826837,-0.1028246209,0.3644041717,-0.3169815838,-0.2264071554,0.0476952232,-0.4318677187,-0.1580609083,0.1578135192,-0.3295030892,0.050670553,-0.0793561414,-0.0533052161,0.3071305156,0.1460714042,0.3061790168,0.2972402275,-0.383091867,0.225058347,-0.0498413593,-0.3608684838,-0.1509022862,0.1439300179,0.1543418616,-0.2163374573,-0.1585421562,0.1210782602,-0.1123431846,-0.2633713484,0.138936311,-0.225103125,0.0310812052,-0.1937533319,0.108003214,0.1789447814,-0.0672992393,-0.0808589831,0.2495893985,-0.2174934298,-0.087718904,-0.0223432947,0.2772181034,-0.0715375021,-0.1196030527,0.2357800156,-0.0965786949,-0.189482227,0.2412616163,0.091824159,-0.2193150669,-0.3616685271,0.0277366918,0.1239842996,-0.4699998796,-0.0984592065,0.0730648711,0.0260902643,0.044149626,0.4836724997,0.1310720742,-0.0179124586,0.2964156568,-0.1941359341,-0.5111047029,0.172272414,-0.3063954413,0.2224185467,0.0716702342,0.4083387554,0.2780316174,0.0212524422,-0.4761144519,-0.0957311094,-0.2024464309,0.0086193178,0.4760988057,-0.0535302311,0.2613423467,-0.0498861,0.1953081638,0.2604642808,-0.4230805337,-0.2319354713,0.0407515615,0.0788130686,-0.018850578,-0.0326265804,0.1591200382,-0.1057995558,-0.3744731843,-0.051499065,0.2088698149,0.1218388528,-0.0660383552,0.02446872,0.1511496156,0.187641412,-0.2258436531,0.0167261921,0.2588253617,0.319647342,-0.0280859172,-0.0400208384,-0.1756953448,-0.1103271842,0.1341460347,0.159148261,0.1411132216,0.1615344286,0.1578661501,-0.1192470342,-0.0192980096,-0.0792006701,0.2799156308,-0.1003041491,-0.0936973691,0.0063015902,0.258610785,0.3450060785,-0.3602778018,0.2975356579,-0.1334537417,0.0064386572,-0.0688623637,-0.0125061208,0.2499472052,-0.152866289,0.1491673887,0.2357269228,0.1898102313,0.0348894112,0.0292287841,0.0096283173,-0.0778544545,-0.0010325771,0.2097520828,-0.1130981296,0.1647657156,0.3772323132,-0.3160652816,0.0949926376,0.0621370897,0.0615479238,0.1109166518,-0.5008833408,-0.0245065764,0.1988550127,0.072304979,0.0479236171,-0.015957661,0.1939796805,0.0329560675,-0.3138833046,-0.4437049329,0.115845412,-0.1720166355,0.2750708461,-0.1679453999,-0.2745361924,-0.1335872114,0.1127260774,-0.0335199609,-0.2113982439,0.3847887814,0.0726080239,-0.0404763259,-0.2797919214,0.0317030363,0.0953296125,0.1308010817,-0.1873326451,0.2822522819,0.0710875243,-0.0393346623,0.0987020135,0.0260619204,0.317925334,0.2258337587,0.1370507926,0.1507961005,0.2091993839,0.0792203844,0.0103420224,0.086575143,0.1785832793,0.1475903541,0.3340989649,0.2623826861,-0.2869746983,-0.2721345425,-0.0941689461,0.2695444524,-0.35337767,-0.0039260546,-0.0860684738,-0.0313970409,-0.2981731296,0.0170973931,-0.4201547801,0.2572529614,0.0354243182,-0.0297196638,0.0713373795,-0.2428658903,0.1755140424,-0.1715015173,0.5958704948,0.1973862946,0.3026838899,-0.3092738092,-0.5232947469,-0.5052442551,0.0167879127,-0.0965643898,0.1665878147,-0.2559703887,0.1326684654,0.3628245592,0.0504737608,0.165999949,0.2906830609,0.2242631763,0.0743630007,-0.19414635,-0.071864754,-0.085461244,-0.1783625484,0.0745044202,-0.3372360468,-0.0929619297,0.0976881459,0.1946698874,-0.087953411,0.0600923114,0.2378426492,-0.284307152,0.5308744311,0.0084486287,0.5093175173,-0.0833964124,-0.1735006422,-0.2166521251,-0.2197478116,-0.0536232591,-0.237412706,0.0840961561,0.2714071572,0.0634320378,0.020726176,-0.503624022,0.3893963993,0.2477635443,0.3329304755,-0.2883208096,0.1675183326,-0.4077230394,-0.0201235637,-0.0112996697,0.101473406,-0.0199635774,-0.1913530529,-0.3272975087,-0.3006255925,0.5939635634,-0.5361892581,-0.0412924662,0.1515975893,0.2181707174,0.0904022008,-0.1211094856,-0.490555197,0.3521925211,0.1682988852,-0.0789178759,-0.1132170036,0.0657814443,-0.1576025486,-0.1189124137,-0.0759415105,0.3231970072,0.0979943275,-0.1713924408,-0.1286943257,-0.3984100223]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/676","title":"train_test_split returns empty dataset item","comments":"The problem still exists after removing the cache files.","body":"I try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty.\r\nThe codes:\r\n```\r\nyelp_data = datasets.load_from_disk('\/home\/ssd4\/huanglianzhe\/test_yelp')\r\n print(yelp_data[0])\r\n yelp_data = yelp_data.train_test_split(test_size=0.1)\r\n print(yelp_data)\r\n print(yelp_data['test'])\r\n print(yelp_data['test'][0])\r\n```\r\nThe outputs:\r\n```\r\n{'stars': 2.0, 'text': 'xxxx'}\r\nLoading cached split indices for dataset at \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-f9b22d8b9d5a7346.arrow and \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-4aa26fa4005059d1.arrow\r\nDatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)})\r\nDataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)\r\n{} # yelp_data['test'][0] is empty\r\n```","comment_length":9,"text":"train_test_split returns empty dataset item\nI try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty.\r\nThe codes:\r\n```\r\nyelp_data = datasets.load_from_disk('\/home\/ssd4\/huanglianzhe\/test_yelp')\r\n print(yelp_data[0])\r\n yelp_data = yelp_data.train_test_split(test_size=0.1)\r\n print(yelp_data)\r\n print(yelp_data['test'])\r\n print(yelp_data['test'][0])\r\n```\r\nThe outputs:\r\n```\r\n{'stars': 2.0, 'text': 'xxxx'}\r\nLoading cached split indices for dataset at \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-f9b22d8b9d5a7346.arrow and \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-4aa26fa4005059d1.arrow\r\nDatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)})\r\nDataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)\r\n{} # yelp_data['test'][0] is empty\r\n```\nThe problem still exists after removing the cache files.","embeddings":[-0.0136992559,-0.0298607461,-0.0888853148,0.3190387785,0.0190193374,0.3202376664,0.6021837592,0.4843112528,0.0905475616,0.0323204175,-0.0479131229,0.3081917465,-0.0566511452,0.0226379037,-0.048620604,-0.1337524951,-0.0076132338,0.3531485796,-0.0649880096,-0.1747340113,-0.1587130427,0.3243024051,-0.3151785433,-0.1851881444,-0.6670129299,-0.1959621161,-0.1485328227,0.0855437964,0.0238804091,-0.2852377295,0.4701847434,-0.1868356019,-0.1838475168,0.4389150739,-0.0001218947,0.2150428444,0.3642240167,-0.1148698106,-0.3536468148,-0.3756703138,-0.2658014596,-0.0999344215,0.2463333905,-0.2923759222,-0.1791367084,-0.3091073334,-0.3419113159,-0.5840457082,0.2359025925,0.3876549006,0.0513744801,0.0667883828,-0.116522491,-0.0540321209,0.3155407906,0.0369479731,-0.18412368,-0.2976082563,-0.1456727237,0.1644529551,0.1135949194,0.2889157534,0.0030854177,0.3384099305,-0.0266885124,0.192962423,-0.3500203788,-0.2590218484,0.236080572,0.2974523306,0.4226731062,-0.0276838932,-0.2365240157,-0.399590075,-0.0722277761,-0.3523159027,-0.0162757281,0.3600808084,0.0586258583,0.1915573329,-0.2560532987,-0.084402509,0.0160805471,0.0308360681,-0.232720986,0.5517379045,-0.0849775076,0.1775852889,0.154498145,0.2532979846,0.2454075962,-0.2248506248,-0.4156620502,0.0392152257,-0.4421793222,-0.1239895448,-0.0464895442,-0.1729997098,-0.0222148132,0.3425300419,0.1302364916,0.1256473511,-0.0212422442,0.0309303142,0.3599259257,0.3450015187,0.3145315647,0.5157473683,-0.1340911537,-0.0338124186,-0.2550422549,-0.1506865025,0.1562656164,0.0097790817,0.2196011245,-0.0395882875,0.1482719183,-0.0999606773,-0.2335683554,-0.0586475655,-0.3129959702,-0.149448961,0.0169360917,0.305033952,-0.0400337204,0.2198151797,-0.2828273177,0.2975720763,-0.139419198,-0.1364942491,-0.3409182131,0.0618148148,-0.0945326984,0.0015835721,0.3188843727,0.040157605,0.2988629937,-0.1462258846,-0.1401026845,-0.3392833769,0.3868964314,-0.3686065972,0.5032181144,0.3611723185,0.2450574785,0.2846489251,0.231769383,0.1634005606,-0.0403060541,0.337567091,-0.4073759913,-0.4581854939,0.1203239411,0.1071453914,-0.2793502808,0.1082999408,-0.1189456284,0.1031373739,0.3248194754,0.1345795989,0.091539301,-0.0434517413,-0.2772683203,-0.0889875367,0.2878130376,0.3444910049,-0.3921041191,-0.2018454671,-0.153083384,0.0590633564,0.0008789673,0.0779583454,-0.2070706189,0.2231879681,-0.3185917437,0.4633011222,0.4069260061,-0.1604932547,-0.3121798933,0.3719954789,-0.0052978508,0.0760053843,0.1348972172,0.0438948162,0.3947341442,0.2482829541,0.1477374136,0.1104831547,-0.1766646355,-0.2216054499,-0.1792588234,-0.1585918516,0.2288827449,0.1085337922,0.3979535103,0.0039851228,-0.0031826964,-0.0271819942,0.3104227483,0.0627038702,0.4102676809,0.3866920471,0.2406358272,0.1347073466,0.1598587483,-0.1691802591,-0.3582789004,0.2762641013,-0.1068533584,-0.1939646751,-0.170280382,-0.2097129375,-0.2480654269,-0.1084934771,-0.4937535226,-0.4181150496,-0.0053421678,0.0033979125,0.2890127897,0.0997730568,-0.2307969183,0.433611393,0.0265368205,0.2539686561,-0.1601684093,0.4588728845,0.1726094484,-0.1285167336,-0.3474343717,0.2893313169,0.1544469446,-0.2652579248,-0.1285501122,0.5229845643,0.1035675257,-0.0113746794,0.0057018143,-0.376186043,0.2769083977,-0.2356022596,0.004431244,0.1440576464,0.2102435976,-0.3076463342,-0.2030430585,0.4296516478,-0.234010458,0.4533729255,-0.0887819901,-0.0778889284,0.0869930983,-0.1984008253,0.0062057539,-0.252808392,-0.2671325803,-0.6313278675,0.4923230112,0.0421935134,-0.0809992626,0.1292485297,0.4800535142,0.1851798296,0.0490794741,-0.2963755429,-0.3799560964,-0.1307857931,0.0842701495,0.4853183925,0.5090169907,0.1987067461,0.4343140423,-0.2804032862,0.3053404689,-0.0969614312,0.2565339804,0.1113508344,0.2373649776,0.2491977364,-0.1361386478,-0.2146163881,-0.0883356705,0.1807025522,0.2340266705,0.0465632714,-0.1054376364,-0.1647658348,-0.3451281786,0.0007487613,-0.115665473,-0.2631690204,0.124257721,-0.4139946997,-0.1120493636,-0.129886955,0.0327498615,0.1377154738,-0.3459695578,-0.0460975915,0.2083253562,-0.1354780793,0.0428335443,-0.1137658134,-0.2015039772,0.0452771559,-0.2031094283,0.1034515649,-0.0160450712,0.0251933187,-0.1039461344,-0.1275982559,-0.0244784541,0.1733095348,-0.194246009,0.0365195423,0.045218166,0.1945366263,-0.0528563373,-0.0148094352,0.280952394,-0.049959261,-0.1307082474,0.0769945234,0.3277001679,0.1118408963,-0.4114836752,-0.5755783319,-0.1387521774,-0.1709298044,-0.127192989,-0.2229975164,-0.0377138555,-0.1934124529,0.0636086166,0.2387853116,0.1956992298,-0.381749928,-0.0796484351,-0.2661925256,0.5177537203,-0.2472695112,-0.1176090389,0.2494963259,-0.0936437696,0.1229230836,0.1205546185,-0.3497059941,0.2411096245,-0.207686767,0.2200057656,-0.0879789963,-0.1112608537,0.3703546822,-0.0325954221,-0.0200879946,-0.1400561333,-0.0493237972,0.2399649918,-0.1393930018,0.1800015718,-0.0418913662,0.2658566535,0.1897232831,0.8674818873,0.2328895777,0.0641227737,0.2545216084,-0.0346755907,0.3965483904,-0.2540817261,-0.2954222262,-0.0559012331,-0.0106537016,0.0043434431,0.2586315274,0.0059628244,-0.1888487488,-0.0591839068,0.1204615831,-0.3784198761,-0.2521870434,0.1118902043,-0.4925549626,0.1797518432,0.0905633047,0.2424609512,0.0790520757,-0.0261223409,-0.1524905264,0.0428648144,0.0930471718,0.0458773077,-0.4535360634,0.0042072856,-0.1950328946,0.1562372148,0.1405526549,0.1325220913,-0.016599359,-0.1549002081,0.0522066317,0.1071720049,0.907374382,-0.0301296879,0.0777048618,0.1425790936,-0.0036621979,0.1084909961,-0.1134132445,-0.123933576,0.4580434859,0.0612970628,0.1274152845,-0.2643380463,-0.0817533433,0.1501991898,0.0163423736,-0.129562676,0.0688163936,-0.1034420356,-0.0157243945,0.1850292832,0.0511429384,0.0645855144,0.2875162959,-0.2549770772,-0.4118429422,-0.2756813467,-0.4323099852,0.2257648706,-0.0940019488,0.6159130931,-0.165939793,0.2583933771,0.1138402447,0.0406177565,0.192991972,0.6257175803,-0.1958746314,-0.0670321882,0.1330210716,-0.4234817326,0.287256062,0.2791524827,-0.0975593776,0.1533161998,-0.1368684471,0.2121735364,0.1275228411,0.0586957708,-0.0826639608,-0.0537142605,-0.3288598359,-0.4959793091,0.0962387621,-0.0174683835,0.1985425353,0.2156388313,0.1023041382,-0.3705566823,0.3236138523,-0.3525702953,0.5845611691,0.1328759938,0.0851713717,0.0052091577,0.022955047,0.2684156299,-0.2451326698,0.2696959376,-0.1519471705,-0.0079324236,-0.1675604731,0.0058324123,0.129671365,0.4917007983,-0.1713621318,0.3351157308,-0.025430413,0.1402037889,-0.2097478062,0.0497784652,-0.0451711342,0.0010414986,0.1309424192,0.0418781266,-0.1400932074,0.1597982347,0.0681055412,-0.162135452,-0.2018470615,-0.1972043514,-0.0762648657,0.1086761579,-0.2591590285,0.3027068377,0.0807093233,-0.5935022235,-0.1768077612,0.1073797122,0.1040076986,0.0083487062,0.0298663359,0.3249391615,0.1016326621,0.00560035,-0.2331026345,0.0005327447,0.3225613832,0.137635693,-0.3769953251,0.0565445349,-0.5206513405,-0.3250971735,0.1465594023,0.3361848295,0.1775530279,-0.379842639,-0.4140000045,0.1946683973,0.0241738427,-0.1589615643,0.0461593792,0.0570100099,-0.3003157377,0.0351202302,0.0802178308,-0.4035818577,-0.1339070797,0.1038346216,0.210653767,0.1934316903,0.6813169718,-0.092593506,0.0945421606,-0.3001667261,0.1401201487,0.41185987,-0.2145386785,0.1946433038,-0.0499970913,0.1372864395,0.1457802951,-0.2230545133,-0.0110794269,0.4891737401,-0.2789618969,-0.3485830724,0.100769408,0.2095288783,-0.0201068465,0.4321010709,-0.0022062024,0.0575315431,0.0900983065,0.1692116559,-0.1688221097,0.3351387084,-0.1360630095,0.2530459762,-0.1570479721,-0.0456002764,0.2340331525,-0.1502685547,0.0003166376,0.0113981171,0.029603269,-0.1699844748,-0.0861199871,0.1566186249,-0.0588031523,0.180138886,-0.3464486301,-0.0389647596,-0.1619588584,-0.2065539658,0.2667501867,0.0144080045,0.0047060759,0.3027579486,0.5212694407,0.2162364721,-0.6195165515,0.1073318347,-0.0346402787,0.0817127526,-0.090405196,0.1297353506,0.1371285617,0.2646201849,-0.0776598006,-0.2009537369,-0.000033287,-0.326233685,0.3473887444,-0.3503061235,0.1630559564,-0.0317440964,0.3556675911,0.1046485379,0.0164327025,0.0993399471,-0.1005933508,0.1407768279,-0.0554904267,0.1748842001,0.0780899674,0.2436935604,0.1394677907,0.3184380829,0.2244095504,0.1559344679,-0.0156039465,0.2505425215,0.4774449468,-0.2229612768,0.498902142,0.3220537305,-0.0221667811,-0.0198533516,0.2059470117,-0.0831189752,0.0121137761,0.5560984015,0.0743142292,0.7793147564,0.311268419,0.1426578313,-0.066162169,-0.3953317702,-0.3098015189,0.0892859623,-0.3012505174,0.0493967272,-0.0110642342,0.6285704374,-0.4465096295,-0.3996686637,0.0379872024,-0.114848882,-0.2282067835,-0.0171772838,0.1587102115,-0.2879336178,-0.3298096657,-0.1775783747,0.3250267506,-0.1781408191,0.1572667509,-0.0398936532,-0.1690577269,-0.1408738792,0.1225883737,0.3199462295,0.0896509215,-0.1565380841,0.1336556375,0.082248807,-0.2207893133,0.2148956656,-0.0013862908,0.1916716695,0.0935179517,-0.2610373795,0.0486506745,0.2884696722,-0.3952535689,0.0046666334,0.1285031885,0.0023132879,-0.4220147729,0.1197435781,0.1231696904,-0.1481700093,0.2146699876,0.1635013968,0.4149589539,-0.3951157629,0.3899232149,-0.15072079,-0.3004842699,-0.1343429387,-0.150453046,-0.4707731307,-0.0725038573,0.2104393542,-0.2107331902,-0.0327815041,-0.0505678207,0.0014293782,-0.1745319664,0.17268686,0.1217111796,-0.1986365318,-0.2783783078,-0.1783223003,-0.5634427071,-0.0564445108,-0.2300333679,0.0582507886,-0.0388421044,0.436499238,-0.0455295779,-0.1148196906,0.2183071673,0.0158811864,0.3312583864,0.0577022545,-0.5139803886,-0.1032845154,-0.2207211554,0.0349747874,0.0166928954,-0.3294147253,-0.0837239176,0.1530560702,-0.1765542477,0.1609555334,-0.3053895533,0.3250873685,0.0807680562,0.0971418992,0.0623806119,0.2973851562,0.1774174124,0.2921417356,-0.1090765893,-0.1457717419,-0.0174437091,0.1880932599,0.1636358202,0.522834599,-0.2134700269,-0.0539803952,-0.0435173959,0.0357248597,0.11596369,-0.0554625317,-0.0488640405,0.1981749684,-0.3685344756,0.1995724589,0.4306891561,0.4390742481,-0.0992893353,-0.2021472603,-0.0715781823,-0.2667915225,0.0581404753,-0.2168218642,-0.2403342575,-0.1212190539,0.1190091595,0.5129519105,-0.1881358176,-0.6468869448,-0.2834579051,0.3674525619,-0.2636160254,0.001171627,-0.0739596561,0.1404255033,0.0992939174,-0.0843359903,0.2014192045,0.1445780843,-0.0160793662,0.08163587,-0.3269009888]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/676","title":"train_test_split returns empty dataset item","comments":"Can you reproduce this example in a Colab so we can investigate? (or give more information on your software\/hardware config)","body":"I try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty.\r\nThe codes:\r\n```\r\nyelp_data = datasets.load_from_disk('\/home\/ssd4\/huanglianzhe\/test_yelp')\r\n print(yelp_data[0])\r\n yelp_data = yelp_data.train_test_split(test_size=0.1)\r\n print(yelp_data)\r\n print(yelp_data['test'])\r\n print(yelp_data['test'][0])\r\n```\r\nThe outputs:\r\n```\r\n{'stars': 2.0, 'text': 'xxxx'}\r\nLoading cached split indices for dataset at \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-f9b22d8b9d5a7346.arrow and \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-4aa26fa4005059d1.arrow\r\nDatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)})\r\nDataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)\r\n{} # yelp_data['test'][0] is empty\r\n```","comment_length":20,"text":"train_test_split returns empty dataset item\nI try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty.\r\nThe codes:\r\n```\r\nyelp_data = datasets.load_from_disk('\/home\/ssd4\/huanglianzhe\/test_yelp')\r\n print(yelp_data[0])\r\n yelp_data = yelp_data.train_test_split(test_size=0.1)\r\n print(yelp_data)\r\n print(yelp_data['test'])\r\n print(yelp_data['test'][0])\r\n```\r\nThe outputs:\r\n```\r\n{'stars': 2.0, 'text': 'xxxx'}\r\nLoading cached split indices for dataset at \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-f9b22d8b9d5a7346.arrow and \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-4aa26fa4005059d1.arrow\r\nDatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)})\r\nDataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)\r\n{} # yelp_data['test'][0] is empty\r\n```\nCan you reproduce this example in a Colab so we can investigate? (or give more information on your software\/hardware config)","embeddings":[-0.1150352061,-0.0503816865,-0.0381936133,0.375646472,-0.0080879638,0.2310482264,0.6476925015,0.2886788249,-0.0222678706,0.1684438586,-0.0831825659,0.4956986606,-0.1554624736,0.2236279696,-0.1473039687,-0.0612270273,0.0395237468,0.416975975,-0.0837054104,-0.1733156145,-0.1575848758,0.3264073133,-0.5189428329,-0.2928180695,-0.7594946027,-0.2361606956,-0.3129857183,-0.0154720666,-0.1107865497,-0.0165034104,0.5567771196,-0.0819952339,-0.1336539239,0.5849176645,-0.0001244457,0.0981333107,0.3251812458,-0.1256597042,-0.4603216946,-0.4501077235,-0.5075224638,-0.0494799428,0.3426325321,-0.2649562359,-0.2269916832,-0.0226692744,-0.2063624561,-0.2614695132,0.2259005308,0.2926188409,0.0635271966,0.0414758995,-0.2472750545,-0.1242459789,0.2525101304,0.1699924916,-0.3104469478,-0.1098317429,-0.084317632,0.1739827245,0.1159843951,0.2465375662,0.0693301037,0.3401452601,-0.09772975,0.2747443318,-0.442955941,-0.3045037389,0.3036516905,0.2749741077,0.3398500383,-0.0917443037,-0.1515464038,-0.3225497007,0.0415977091,-0.4011600912,0.0713355243,0.3876688778,-0.090194419,0.1396924406,-0.2474467754,0.0599142238,0.0849869326,0.0673935786,-0.34357512,0.6945251822,-0.1937275529,0.2074731737,0.0974512175,0.2137624174,0.5437459946,-0.2621796131,-0.2669441402,0.0290783402,-0.3628301322,-0.1718165725,-0.1199132726,0.088610962,-0.0517791547,0.1024868041,0.196312204,0.0715745986,-0.0790911168,0.1310102344,0.295804441,0.274138689,0.3077515066,0.411904484,-0.0986466035,0.0760244876,-0.3278721273,-0.0480283536,-0.001570637,0.0881480873,0.0448615327,-0.0580453575,0.1930097491,-0.1583079696,-0.3444063663,-0.1840529591,-0.5183025002,-0.0567977987,0.1866762191,0.3126021326,-0.0572762713,-0.0559262596,-0.3544777334,0.1348287761,-0.0955052078,-0.0661771148,-0.2749396265,0.0778156891,-0.2299034297,0.1631309986,0.2406240255,0.0799832791,0.3387073874,-0.1525916606,-0.0916989893,-0.2025912255,0.2280801088,-0.5900961161,0.4591190815,0.3485629261,0.3222778141,0.1913000494,0.0980790704,-0.0664385781,-0.1165150926,0.4029569626,-0.2699317932,-0.3050501645,0.1691551358,0.0849727765,-0.3663966358,0.0794186145,-0.2584615946,0.2686917484,0.2266364843,0.0961481705,0.1163988858,-0.0441446565,-0.4369967878,-0.0590528809,0.3124624789,0.2948292494,-0.408418119,-0.1554810703,-0.1902661473,0.2188366354,0.2140793651,0.0398626626,-0.2395052761,0.1319135129,-0.3153765798,0.3399737179,0.3605463207,-0.0871430263,-0.2835251689,0.2771761417,0.0215708781,0.051359456,0.0451857708,0.0018947049,0.4075783491,0.1792588532,0.328889519,0.1144289449,-0.2773238122,-0.1597403884,-0.2146873772,-0.2300063521,0.1433477104,0.0812680051,0.3348293602,0.019803999,-0.0664328039,-0.0552028492,0.3579739034,0.0629952848,0.2749548852,0.2020434588,0.3893932104,-0.1076148227,0.0098543344,-0.1398951411,-0.2623062134,0.1802488565,0.1984196007,0.0323021151,-0.0946407765,-0.2139585763,-0.276968807,-0.0646537617,-0.3802941442,-0.3407127857,-0.0074102082,0.0401033908,0.1436058581,0.0353596397,-0.1622667164,0.5844069719,0.0103582954,0.2315689772,-0.0459007397,0.5272762775,0.1391587108,-0.1883159131,-0.2542985082,0.2448890358,0.1367236972,-0.2267313153,-0.0985794738,0.3551600277,0.1909105629,-0.1077017635,-0.0189544912,-0.3879105747,0.3679305017,-0.1755170822,-0.0699600354,0.0392054655,0.3265375495,-0.2248434722,-0.1471385509,0.5041164756,-0.2933380604,0.3285697699,-0.0480019301,-0.0602198541,0.0336071253,-0.1547931284,-0.0144889625,-0.2078600675,-0.3048747778,-0.4725946784,0.3259237409,0.1150533184,-0.2302742004,0.3069022894,0.4501276314,0.187895909,0.0188092645,-0.2738700807,-0.3083892763,-0.0857903063,0.3047046661,0.5613517165,0.5491407514,0.1559779942,0.3771387935,-0.1911429465,0.2594592571,-0.1260149032,0.2184940875,0.0867770016,0.4505963624,0.1112086698,0.0040066582,-0.1120923236,-0.0821984336,-0.0164337568,0.2886066437,0.0569797084,-0.0253688525,-0.0290943962,-0.3493950665,-0.1848792881,-0.229042232,-0.3024784923,0.2572642863,-0.3693206906,-0.1748346984,0.0454847775,0.0536181666,0.0645638555,-0.1782749146,-0.0004121346,0.1094598174,-0.0240367036,0.0801911056,-0.0892360955,-0.2991433442,0.0594417863,-0.1319158077,0.2729621828,-0.0352631845,0.0179516766,-0.1311779171,0.0478454605,-0.0884177163,0.3439834416,-0.3098200858,0.2339693457,0.0029196867,0.0047655958,-0.3975249529,-0.1547760516,0.1998274773,0.0130147431,-0.155743286,-0.138789475,0.2697178125,0.014482283,-0.3401608169,-0.3421728611,-0.1796095669,-0.1125278249,-0.1947131455,-0.2104953229,0.0175590981,-0.027923286,0.06466645,0.2902249098,0.3815029562,-0.4983840883,-0.0954140946,-0.204627201,0.4170759618,-0.2486552298,-0.140739724,0.2069021612,-0.2243056148,0.1000042334,0.1204108894,-0.3114476502,0.0275154077,-0.1803904623,0.098407127,-0.0162202697,-0.2345980853,0.2713419199,-0.0120163467,0.0414469205,-0.1543293446,-0.2120657265,0.0987327918,-0.0192148313,0.3380070329,-0.0995228514,0.0485847108,0.0872996375,0.7695263624,0.1302616596,-0.0590402596,0.185683161,-0.0803642944,0.294434607,-0.2148572505,-0.3306841254,0.085235022,0.0504689589,-0.2484665066,0.3751797676,0.0260967873,-0.2959826291,-0.066269137,0.0787969157,-0.3488972783,-0.4114096463,0.084487386,-0.3040919006,0.0734924451,0.17384395,0.2460621148,0.1376636028,-0.0574313626,-0.3457048237,0.0935205594,-0.047715053,-0.0272557456,-0.1181514934,-0.1120559946,-0.2812520564,0.1049769521,0.2453666478,0.1161459759,0.0674711764,-0.1029446051,0.0977059975,0.0931061357,0.8047711849,0.0951175541,0.0255576987,0.151276812,-0.0543951802,0.0289500859,-0.1543101072,-0.2661253214,0.5942175388,0.0351537652,0.107962139,-0.1630769372,-0.0757580101,0.2782518864,-0.0735875219,-0.0777219236,0.045203682,-0.0972709432,-0.2069830894,0.2816327214,0.0695414692,0.0513364933,0.2702434659,-0.2625232637,-0.4213937819,-0.2210748792,-0.322194159,0.3461486995,0.1722915918,0.4645672441,0.0360242091,0.2999585271,-0.1496406198,0.0685373098,0.3058344126,0.5060020089,-0.1083004549,-0.0132051185,0.1994806677,-0.4382297099,0.3078832924,0.3851027787,-0.0773703977,0.1544531882,-0.3961910605,0.108464025,0.3041463494,0.2651836574,-0.1541537344,0.0501260422,-0.4672730565,-0.3593154252,0.1445117444,0.0673304126,0.0478441268,0.1550257355,0.166083023,-0.3851067126,0.402581811,-0.1388577074,0.6618785262,0.1113434434,0.0216269195,0.1709940732,0.0542530268,0.3996334374,-0.0652922019,0.3572191596,-0.1694017202,-0.233551085,-0.1278268993,0.0270858835,0.0771726221,0.2690830529,-0.3201324344,0.2286082059,0.0635437295,0.1925323308,-0.0743437558,-0.2257066369,-0.0816244483,-0.0338220298,-0.1214183792,0.0666676164,-0.1046582311,0.1696460992,-0.0640848577,-0.1659599841,-0.3470738828,-0.1202433556,-0.1196520403,0.0415382758,-0.4431511164,0.1055006236,0.1274347305,-0.6294221878,-0.0395515412,0.1533181071,0.1393236816,0.0197388176,0.0174572952,0.2915143669,0.0716116205,0.0380920395,-0.2708716393,-0.2328441888,0.3416423798,0.1064351425,-0.2811036706,-0.1011908203,-0.3789856434,-0.1036455557,0.3149293065,0.4081469178,0.0745374784,-0.3411479294,-0.309253633,0.2713319957,0.1163845062,-0.2194038481,0.0284717027,-0.093044199,-0.3396018147,0.1892219037,0.213169992,-0.3500841856,-0.1038265675,0.0506324433,0.2158877701,0.3880850673,0.5702491999,-0.1257463694,0.1827216148,-0.2740946412,0.2483842224,0.3024659455,-0.121639125,0.2211587727,-0.2184768617,0.0917791948,0.2944124341,0.2036750615,0.1276053339,0.4597963095,-0.3011159599,-0.3193522394,0.0755893141,0.1595554352,-0.0148545252,0.3178521395,-0.0485178344,0.2364337146,0.1221807078,0.0492864214,-0.1707865596,0.336376816,-0.2865241468,0.291434288,-0.1016721502,-0.0084250681,0.323661685,-0.1891307384,-0.0468913019,0.1872662455,0.0975887477,-0.1881628036,-0.1224057823,0.1336726695,0.015345227,0.1390743703,-0.3899894953,0.0957617164,-0.1807764769,-0.2337371856,0.3308992088,0.0481334813,0.0454878807,0.1050186828,0.4617639184,0.3260038495,-0.5690649152,0.1453598291,0.0805115104,0.2416459918,0.028403841,0.1527653784,0.1809976548,0.2765832841,0.2391925454,-0.2424150407,-0.0734812319,-0.3484893441,0.2455792278,-0.2719382048,0.226150617,-0.1522747725,0.2825799286,0.2128136009,0.1177982613,-0.0052538929,-0.0358106792,0.0897980258,-0.0477615222,0.3134739995,0.1652405709,0.2337402701,0.1872961223,0.5071595907,0.1327540427,0.2762697637,-0.1346289366,0.1326930523,0.3872712553,0.1447565705,0.4235880673,0.2531102002,-0.2026654333,0.2513694763,0.1573185325,-0.2246248871,0.0482556596,0.4973720312,-0.0109178396,0.8306031823,0.2314440608,0.1009561494,0.1960763335,-0.3582451642,-0.0870556235,0.1386700124,-0.1664347649,-0.0359754637,0.0435004607,0.6147770882,-0.5687756538,-0.3220711052,0.033976756,-0.0765148774,-0.2327789962,-0.0246977471,0.1492420435,-0.2099410295,-0.415474683,-0.0766517296,0.2104127854,-0.2003394514,0.1773407906,-0.040449027,-0.1598809361,-0.0455557704,0.0901641101,0.4565052688,0.0363219082,-0.098339662,0.2844130993,-0.1642322093,-0.1471322924,-0.006801554,0.0563077889,0.1897610724,0.0989151299,-0.280459404,-0.0977751315,0.2884265184,-0.3211739361,-0.0576598123,0.000604596,0.0949134678,-0.2591908872,0.2111064643,0.1038462371,-0.1199933961,0.3641736507,0.2107376754,0.178730607,-0.3844685853,0.2586522698,-0.1900773495,-0.1395036727,-0.1266788095,-0.1292676032,-0.4150033593,0.0217049494,0.1124092788,-0.3707098365,0.0645275638,0.1396985948,-0.0088709369,-0.1747846603,0.3111534417,0.1171784997,-0.0375645943,-0.2863926888,-0.2873168886,-0.4279896617,-0.0448138751,-0.1681403518,-0.0244649258,-0.2829699516,0.4049312174,-0.0274438746,-0.0090823257,-0.0340911672,0.0970258713,0.3957843184,0.031071553,-0.5145777464,-0.2315471768,-0.2865189612,-0.2351807952,0.0116838347,-0.3804796338,-0.099413082,0.019452244,-0.2052187026,0.0983179212,-0.4067668319,0.1684648842,-0.039140787,0.129312098,0.1172951907,0.4498714209,0.1242432594,0.3546867967,-0.0034433296,-0.1645329595,-0.059148863,0.1902570426,0.2555554509,0.4428932369,-0.1055395156,0.2147505581,0.113697283,0.0909379199,0.2821521759,0.0163339935,-0.0582649894,0.1255048364,-0.5242887735,0.1475864202,0.3340400457,0.428783536,-0.21595864,-0.1925842613,-0.0772522315,-0.2364570498,0.1647450775,-0.2881684303,-0.2600948215,-0.304441452,0.0644336045,0.5333266854,-0.3095366061,-0.6090126634,-0.3760145009,0.2566226423,-0.3416646719,-0.0502407253,0.068371661,0.2078633159,0.0642896071,-0.1890226305,0.2928658724,0.1276476234,-0.1594105214,-0.0744992718,-0.2763881385]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/676","title":"train_test_split returns empty dataset item","comments":"Thanks for reporting.\r\nI just found the issue, I'm creating a PR","body":"I try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty.\r\nThe codes:\r\n```\r\nyelp_data = datasets.load_from_disk('\/home\/ssd4\/huanglianzhe\/test_yelp')\r\n print(yelp_data[0])\r\n yelp_data = yelp_data.train_test_split(test_size=0.1)\r\n print(yelp_data)\r\n print(yelp_data['test'])\r\n print(yelp_data['test'][0])\r\n```\r\nThe outputs:\r\n```\r\n{'stars': 2.0, 'text': 'xxxx'}\r\nLoading cached split indices for dataset at \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-f9b22d8b9d5a7346.arrow and \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-4aa26fa4005059d1.arrow\r\nDatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)})\r\nDataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)\r\n{} # yelp_data['test'][0] is empty\r\n```","comment_length":12,"text":"train_test_split returns empty dataset item\nI try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty.\r\nThe codes:\r\n```\r\nyelp_data = datasets.load_from_disk('\/home\/ssd4\/huanglianzhe\/test_yelp')\r\n print(yelp_data[0])\r\n yelp_data = yelp_data.train_test_split(test_size=0.1)\r\n print(yelp_data)\r\n print(yelp_data['test'])\r\n print(yelp_data['test'][0])\r\n```\r\nThe outputs:\r\n```\r\n{'stars': 2.0, 'text': 'xxxx'}\r\nLoading cached split indices for dataset at \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-f9b22d8b9d5a7346.arrow and \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-4aa26fa4005059d1.arrow\r\nDatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)})\r\nDataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)\r\n{} # yelp_data['test'][0] is empty\r\n```\nThanks for reporting.\r\nI just found the issue, I'm creating a PR","embeddings":[-0.0692258254,-0.1399071515,-0.0832422823,0.3288369477,0.1622399539,0.2811196446,0.5008799434,0.4487760365,0.0246849339,0.0437641256,0.0638897568,0.3331188858,-0.0612816736,0.1503981948,0.0031257027,-0.3133085668,0.014100763,0.3501460254,0.1334746182,-0.1657955199,-0.1645467728,0.3009945452,-0.3867303431,-0.0283491742,-0.6012365222,-0.2343055159,-0.1984614432,0.0304441769,-0.0407279357,-0.2345176786,0.4407738149,-0.1730792522,-0.0664960817,0.4659323394,-0.000122333,0.1851662993,0.3183399439,-0.1087606102,-0.2321871519,-0.4960864484,-0.2220596671,-0.0540375896,0.169494465,-0.1669001877,-0.2523914874,-0.5214830637,-0.3494725227,-0.3989439309,0.1032987386,0.4921701849,0.0618354566,0.1277791858,-0.0308778044,-0.126215294,0.2270111591,0.0731073618,-0.1628945917,-0.3062564135,-0.1305507869,0.0858042315,0.2098292708,0.1276037246,0.0727683157,0.3254510462,-0.035023652,0.2741033733,-0.2721252739,-0.2937547565,0.1546309292,0.2236322463,0.4064407051,-0.0446998626,-0.2039678544,-0.2823201716,-0.0188810378,-0.3768885136,-0.0126348287,0.4752480686,-0.0541098639,0.1777257472,-0.3229609728,-0.0062980871,-0.0366358459,0.1047730222,-0.2832176387,0.5424917936,-0.0420952514,0.2676186562,0.2208230793,0.2334234565,0.2378802299,-0.2615387738,-0.3574378788,-0.0335276909,-0.3384468257,-0.1822820008,-0.0514249392,-0.0835057124,-0.0333305076,0.4234345853,0.0241253488,0.0817929357,-0.0078448113,0.0299990736,0.5028821826,0.2192660868,0.250667423,0.7321005464,-0.197307393,-0.0857820958,-0.221535936,-0.1081528142,0.1725011468,0.0278904103,0.1677374244,-0.0898575112,0.175998643,-0.1116527319,-0.3258713186,-0.0096074129,-0.4230352938,-0.1525691748,0.0236974005,0.3949540854,0.0243018195,0.2217854857,-0.2038404197,0.2689420283,-0.1859201342,-0.1087822989,-0.3139494359,0.1262977868,-0.1566160768,-0.0336585119,0.1874148101,0.1495432854,0.2486540824,-0.1927955151,-0.1435847878,-0.3561177254,0.3846796155,-0.3879866302,0.45037058,0.3888926804,0.259203434,0.1444913745,0.1872133315,0.0973837003,-0.0630506277,0.4184350669,-0.2409835011,-0.4681658149,0.1210204512,0.1241473481,-0.249706611,0.0749129131,-0.1207142323,0.1754167676,0.2396708727,0.2106685042,0.1013673842,-0.0829644501,-0.2997838855,-0.0554357991,0.3250390887,0.298917979,-0.435490489,-0.1938529313,-0.0269413814,-0.0377576165,0.0987228975,0.0639264435,-0.1908593029,0.3027178943,-0.2560653985,0.3515174985,0.403164506,-0.1523319781,-0.1991901696,0.3608879745,-0.1322743744,-0.0881163031,0.0899762362,0.0333884805,0.3286120892,0.2344780713,0.1426410079,0.1282824427,-0.3315196931,-0.1611126065,-0.1317579001,-0.0540887788,0.2540296912,0.1392902732,0.3386989534,0.009810593,-0.0568286516,-0.0019291023,0.3218839467,0.1079667509,0.3927653432,0.3693246841,0.3182047307,0.1337313205,0.0898148715,-0.2464287877,-0.1984262913,0.1360871047,-0.0220111925,-0.1006483436,-0.0659577399,-0.1627495736,-0.3451218605,-0.0321597084,-0.4366373718,-0.4379944801,-0.0311503597,0.0099573871,0.2998206615,0.0316401124,-0.2314134389,0.3132321537,-0.0151020605,0.2709501088,-0.1372412145,0.5579671264,0.0858738646,-0.1602363884,-0.2550652921,0.3363881707,0.1351814568,-0.1897073686,-0.0282705445,0.5393128395,0.1561227739,-0.0963500366,0.1016719937,-0.4441440105,0.3311428726,-0.3701614439,-0.0459192879,0.1064216942,0.2560572028,-0.2861233056,-0.3384467959,0.4985240698,-0.2237933874,0.3937275708,-0.0617673174,-0.0050035943,0.008790548,-0.1795272529,0.0197951812,-0.179376632,-0.2345152199,-0.5615051985,0.3890978098,0.017576009,-0.2142573446,0.1536695957,0.4248917699,0.1010513902,-0.054269433,-0.2678136826,-0.3070979416,-0.0779673606,0.1100468785,0.4688039422,0.5742069483,0.2218060046,0.3895666897,-0.2460367382,0.2372018546,-0.0767608359,0.2861794531,0.12975353,0.0885100663,0.102936469,-0.0894576833,-0.2887299061,-0.0545921512,0.1589853913,0.2574138641,0.1586489826,-0.1498198807,-0.1529001743,-0.2970952392,0.0205893181,-0.1837679297,-0.299725771,0.218993783,-0.3766548634,-0.1327089518,-0.1720713526,0.0609136075,0.1534673572,-0.3233666718,-0.1307486147,0.2058083117,-0.095775038,0.1071398631,-0.1162371263,-0.302728802,0.0548043177,-0.2462932616,0.2195638716,0.1494247764,0.1213808805,-0.1647442281,-0.172728315,-0.0510458872,0.2122229487,-0.3291912675,0.1005400717,0.0924106836,0.1489076763,-0.0589384362,-0.0096183391,0.3382163346,-0.0752585456,-0.1832846552,0.0561846159,0.3177065551,0.039433863,-0.3621302843,-0.6693060994,-0.1701436341,-0.1311104894,-0.042758707,-0.1126700789,-0.1030751765,-0.3102213144,0.1905998141,0.1736620367,0.1487074345,-0.4946018457,-0.0123268915,-0.1887061298,0.3734691739,-0.1857345253,-0.1374296844,0.2316340804,-0.0453420952,0.0085654929,0.0681638345,-0.3258306086,0.0948413759,-0.1826951355,0.3482220471,-0.008392876,-0.1020498201,0.3345980942,-0.0186947081,-0.0149772884,-0.0903544128,-0.2031243742,0.182360962,-0.1545674503,0.2948225439,-0.0614191145,0.2654571533,0.0838236436,0.8389443755,0.3068560362,-0.0551252551,0.1542288661,-0.1708200276,0.3282755017,-0.2170514911,-0.3026107848,0.0331442505,0.0241007488,0.028293604,0.2113737613,0.0341551788,-0.1356151104,-0.0695623755,0.1628718525,-0.3378141522,-0.2322252393,0.0513067693,-0.4491755068,0.1951909959,-0.0111551574,0.2712552547,0.0580530316,-0.036712788,-0.2497382462,0.0165349767,0.0583529398,0.0073867259,-0.5049040914,0.0354797691,-0.1723843366,0.1477490664,0.0739840344,0.1601296067,-0.0775010586,-0.1007753387,0.1228605881,0.089914158,0.9712352753,-0.0609134473,0.0226266049,0.0474038683,0.0161762312,0.1253774762,-0.0026444367,-0.2329001725,0.4291203618,0.046927277,0.1052201167,-0.293818593,-0.0979690701,0.2789357901,-0.0779123977,-0.1038213447,-0.0369339585,-0.1101406142,-0.1111246422,0.1850880831,0.1682392806,0.0842991173,0.2928035855,-0.3033387661,-0.4022770524,-0.1171652377,-0.3910799325,0.300688535,-0.0950870439,0.587339282,-0.2661634386,0.3136860132,0.0757007524,-0.0035671208,0.3057568073,0.6267791986,-0.1922587454,-0.2090499103,0.129686445,-0.4727914035,0.4119720459,0.2536488175,-0.0598122925,0.153866604,-0.1273107678,0.2574986517,0.1033495292,0.0629096329,-0.0226616319,-0.0303056929,-0.3368639648,-0.5548412204,0.0236936864,0.0443819463,0.2274028659,0.1133423895,0.1477009505,-0.2912487388,0.3620987236,-0.3070876002,0.6621349454,0.1249903291,0.0384457409,0.0616829731,-0.1210616678,0.2905127108,-0.3129933476,0.2031474113,-0.2081340849,0.0379224978,-0.1686794609,-0.0400554352,0.1357192844,0.4900937378,-0.2692544758,0.3050555885,0.0130924657,0.0566475429,-0.1229907721,0.0454400852,-0.0610059686,0.0536985025,0.0273267012,0.0447062291,-0.1512764245,0.1466617733,0.0537761822,-0.1675367653,-0.2348791063,-0.2073855996,-0.1974988431,0.0746520013,-0.2080063671,0.1435916871,0.2049618661,-0.5883692503,-0.1354315579,0.2373399436,0.0900980085,-0.0269070119,-0.0377723165,0.3466602564,0.1001263037,-0.0251955688,-0.2198298573,-0.1089737937,0.3403795362,0.151378125,-0.4699990749,0.1359284222,-0.4741043746,-0.2202642709,-0.0289817303,0.3751183748,0.2001461983,-0.4143981338,-0.2675417364,0.1446793973,-0.0040655551,-0.2112818956,0.0441679172,0.039437931,-0.2804901898,0.0369543806,0.1986828148,-0.3770993352,-0.1434695572,-0.0015123916,0.2128536105,0.2328515351,0.6224246025,0.0193146057,0.1335740834,-0.355597645,0.1771530658,0.5034751892,-0.2378745526,0.2084354907,0.0083694402,0.1636676341,0.2857747674,-0.1548123062,-0.0278204829,0.3669712245,-0.2558273971,-0.3016428351,0.066472806,0.1214822531,-0.0239174627,0.4458616376,-0.0283885095,0.2968219519,0.1283537447,0.163887009,-0.1756293476,0.2410324812,-0.2868854702,0.2920682132,-0.1304365844,-0.0426908992,0.1848874986,-0.2096262872,-0.0039816643,0.0540040135,0.0402254276,-0.1483696997,-0.0591092072,0.1577388197,-0.0509914607,0.1809583157,-0.3827188611,0.0384443216,-0.2085055262,-0.1790226251,0.201378718,0.0246607512,0.0272086337,0.2389764488,0.5471405983,0.2556939423,-0.4877431393,0.0229049325,-0.0330386125,0.1655492634,0.0368326493,0.0993533209,0.192557767,0.2495333254,-0.0916492343,-0.1875119209,0.1401937902,-0.2722086608,0.3440042436,-0.3154484332,0.1098888367,-0.1637266427,0.4070223272,0.2370828688,-0.031045625,0.2151905894,-0.072019957,0.1305464804,-0.0705041364,0.2705824375,0.0778609812,0.284802258,0.1415928751,0.377294302,0.1088183597,0.2488646358,-0.1791144907,0.2729105949,0.4832733274,-0.2954136729,0.3878009021,0.3491734564,-0.0616737194,0.0226347912,0.2674302757,-0.0999648497,-0.0488859303,0.5379616022,0.1225563139,0.8046615124,0.3010440171,0.1920514405,0.0261310562,-0.3077985644,-0.250263989,0.0372233465,-0.2184349298,-0.0479722507,0.0906971395,0.6547471285,-0.4528113604,-0.3387594521,0.0375295691,-0.0983041674,-0.2058597654,-0.0376347005,0.0330090337,-0.2489822954,-0.3534676135,-0.1199418604,0.3764518201,-0.3002145588,0.1956713051,-0.0710040554,-0.1644324809,-0.2800527811,0.1135727689,0.297208786,0.0377394557,-0.1406639814,0.178888768,0.0109710442,-0.2542884052,0.272274375,-0.0720950142,0.2705946863,0.0833100677,-0.2751749456,0.021903675,0.2562428713,-0.3994549811,0.0114197573,0.1589533985,0.0220184792,-0.2706472576,0.1641083956,0.1273721159,-0.1094768792,0.2264485955,0.2214173675,0.4050032198,-0.4769397676,0.5263082981,-0.1397278607,-0.3470490873,-0.1880185157,-0.1941388398,-0.425721854,0.0217535384,0.1537575722,-0.1686933786,-0.0264133196,-0.0552439913,-0.0017204976,-0.1079888716,0.1301419735,0.1521756947,-0.1436682343,-0.361202687,-0.1339243054,-0.4773192406,-0.055799067,-0.2874367237,-0.0131239705,-0.1389248669,0.4015713036,0.0337339006,-0.0629594252,0.1022224948,0.1271187216,0.245279938,0.1763147265,-0.466230154,-0.2396556586,-0.2299015373,0.1189728752,0.0449386314,-0.2557972074,-0.123530753,0.2178478241,-0.2132015377,0.0463367067,-0.3076741397,0.2911501229,0.0874228179,0.0523960143,0.092025958,0.4203707874,0.1802183837,0.3620956242,-0.1223116592,-0.1324259341,-0.1293004453,0.2738141119,0.140817821,0.3951690495,-0.1212823093,-0.0960315913,-0.041967582,-0.0921561122,0.1321140528,-0.0122561287,-0.0141508551,0.1424739957,-0.330191195,0.0737133995,0.3887591958,0.3652699888,-0.089974843,-0.2397662848,-0.0408004858,-0.2522728741,0.1089337319,-0.1479311138,-0.23006998,-0.084854275,0.016916737,0.4508520365,-0.2538329363,-0.7680340409,-0.3387804627,0.3814491928,-0.358556062,-0.0557784811,-0.0562797189,0.1166052446,0.0956540406,-0.0851709694,0.27356866,0.1231876388,-0.0523716584,0.1689541936,-0.2802062929]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/676","title":"train_test_split returns empty dataset item","comments":"We'll do a release pretty soon to include the fix :)\r\nIn the meantime you can install the lib from source if you want to ","body":"I try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty.\r\nThe codes:\r\n```\r\nyelp_data = datasets.load_from_disk('\/home\/ssd4\/huanglianzhe\/test_yelp')\r\n print(yelp_data[0])\r\n yelp_data = yelp_data.train_test_split(test_size=0.1)\r\n print(yelp_data)\r\n print(yelp_data['test'])\r\n print(yelp_data['test'][0])\r\n```\r\nThe outputs:\r\n```\r\n{'stars': 2.0, 'text': 'xxxx'}\r\nLoading cached split indices for dataset at \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-f9b22d8b9d5a7346.arrow and \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-4aa26fa4005059d1.arrow\r\nDatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)})\r\nDataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)\r\n{} # yelp_data['test'][0] is empty\r\n```","comment_length":25,"text":"train_test_split returns empty dataset item\nI try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty.\r\nThe codes:\r\n```\r\nyelp_data = datasets.load_from_disk('\/home\/ssd4\/huanglianzhe\/test_yelp')\r\n print(yelp_data[0])\r\n yelp_data = yelp_data.train_test_split(test_size=0.1)\r\n print(yelp_data)\r\n print(yelp_data['test'])\r\n print(yelp_data['test'][0])\r\n```\r\nThe outputs:\r\n```\r\n{'stars': 2.0, 'text': 'xxxx'}\r\nLoading cached split indices for dataset at \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-f9b22d8b9d5a7346.arrow and \/home\/ssd4\/huanglianzhe\/test_yelp\/cache-4aa26fa4005059d1.arrow\r\nDatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)})\r\nDataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)\r\n{} # yelp_data['test'][0] is empty\r\n```\nWe'll do a release pretty soon to include the fix :)\r\nIn the meantime you can install the lib from source if you want to ","embeddings":[-0.1176419482,-0.0009298192,-0.1334935278,0.2837670147,0.0316409729,0.2487456352,0.5152361393,0.5261705518,0.0605890118,0.0824789405,0.0223485641,0.3803651631,-0.1783174723,0.1184177399,-0.0266839936,-0.1455709487,0.0281710345,0.3949723542,-0.0627409294,-0.1058102325,-0.1067902669,0.3772622943,-0.3872468472,-0.1233003289,-0.5268978477,-0.2172542512,-0.1449890882,0.0612099282,-0.114821814,-0.2209564894,0.5441961288,-0.1443571299,-0.1798071414,0.4662792683,-0.0001174493,0.1740997732,0.3488704562,-0.0959639996,-0.2131626606,-0.5335092545,-0.1603131443,-0.1066494063,0.259299159,-0.208511889,-0.2019205838,-0.3470432162,-0.3203684092,-0.5306897759,0.1431894153,0.3292081952,0.0906265676,0.0888477862,0.0527684987,-0.0330180675,0.4277412295,0.0449317768,-0.2140247673,-0.2842196524,-0.0964196846,0.1749909073,0.1409329027,0.249644056,0.0374673866,0.1979111582,0.063509129,0.2312937677,-0.3535765409,-0.2968400717,0.0674433857,0.3545723557,0.3835054338,-0.1285975277,-0.2091629803,-0.2360120267,-0.0195489172,-0.2849743664,0.0304181445,0.3383017778,-0.0515981652,0.1450700462,-0.1933145225,-0.1244631037,-0.0433207415,0.1351359636,-0.1743124723,0.6467148662,-0.1317276061,0.2141457647,0.2832026184,0.2190051973,0.1526591927,-0.1540828794,-0.3844638169,0.0220124591,-0.3630778491,-0.1830717176,-0.0904447287,-0.2417935431,-0.0404366367,0.4139630198,0.0901596174,0.154797554,-0.0677702874,0.051500693,0.3399357796,0.2884173095,0.3791084886,0.5538810492,-0.0362380408,-0.022138996,-0.0798987299,-0.0783459321,0.0799446329,-0.1314550042,-0.0016993288,0.0321706422,0.1306239218,-0.1279747784,-0.2455851436,-0.1167563349,-0.3425415754,-0.1276844889,0.1007220522,0.3022542298,-0.0558883809,0.2449736893,-0.2102753669,0.1757251769,-0.1284891814,-0.0782680884,-0.2893948257,0.2224498093,-0.1635094434,-0.0577653572,0.2249544561,0.0333971716,0.3561101258,-0.1195592508,-0.029349409,-0.1698011905,0.33678931,-0.3161755502,0.3781973422,0.3954267204,0.1683808863,0.2080931067,0.1540404856,0.2103333175,-0.1806223541,0.3590171039,-0.2874221504,-0.4169856906,0.0133517589,0.1523927897,-0.2529547215,0.0737892911,-0.1068571284,0.1531223208,0.2022406012,0.0466021188,0.1070308983,-0.0253007133,-0.1190391406,-0.1371486634,0.4137606025,0.1770732701,-0.4947104454,-0.1679688394,-0.2387651056,0.0077859187,0.0688475072,0.0316146389,-0.2810123861,0.1752462238,-0.1840929538,0.5340542197,0.4903393686,-0.1969677359,-0.2827195227,0.405551523,0.0146204652,-0.0382343493,0.0570961796,0.1137902513,0.3004012704,0.2838887274,-0.0063301325,0.2280911505,-0.1854338497,-0.2758221924,-0.1109347567,-0.2520990968,0.285246551,0.0868118331,0.4078781009,-0.1718384027,-0.074890241,0.0638745427,0.3752987385,0.0428891145,0.3869078159,0.3386223316,0.3746705353,0.1301590353,0.0755097419,-0.2861389816,-0.2770721614,0.1796784699,-0.0227101799,-0.1048072278,-0.1208274141,-0.2206861228,-0.2114479244,-0.0344585665,-0.442486763,-0.4250469804,0.0233588386,-0.0592853539,0.1825018376,-0.0041589229,-0.3373215199,0.2488869131,0.0244200528,0.3148395121,-0.2052736878,0.4214242101,0.0370678529,-0.1337924451,-0.296159476,0.2590880692,0.0629136339,-0.3750789165,-0.0266384166,0.631146431,0.0772382542,-0.0642370135,-0.052646663,-0.4810121953,0.3369107544,-0.2955011427,-0.0480407886,0.2148432732,0.2424692065,-0.2903488576,-0.1927265972,0.4654704034,-0.2564730942,0.4291006923,-0.056961488,0.0397228561,-0.02869091,-0.1467879266,-0.0515800901,-0.2208297998,-0.2495584339,-0.7266792059,0.3887437582,0.0322172083,-0.09846697,0.0736336112,0.4730113149,0.1401787102,0.0240965132,-0.2417277843,-0.2793951333,-0.1014119387,0.1261884272,0.4395976067,0.4393751323,0.1826652437,0.330413729,-0.293017447,0.2054164857,-0.112560153,0.2705596387,0.1046366841,0.2017994672,0.1577464938,-0.2934381366,-0.1967669278,-0.0721429586,0.0529737584,0.2197586894,0.0666935965,-0.0715603903,-0.151461646,-0.3876054883,0.0715205222,-0.1323045492,-0.4492847025,0.0836419687,-0.3369779885,-0.1412802041,-0.1519938111,-0.050725393,0.1981332898,-0.2802027464,-0.1647758186,0.1501642764,-0.1726747751,0.0782484561,-0.0823540613,-0.2641968131,0.1076270193,-0.10392344,0.16381827,0.0216651894,0.0750716552,-0.1681145877,-0.0679952949,-0.1577569395,0.1224758998,-0.2915972471,0.0457189493,0.2256562859,0.0956452414,-0.0532497875,-0.0517577119,0.2858192921,-0.1603218168,-0.1362540275,0.0089375339,0.2589613795,0.01628525,-0.4951088428,-0.6609779596,-0.1393256783,-0.21591717,-0.0639085546,-0.1677109003,0.0056940648,-0.1805396825,0.1855225116,0.3342963755,0.2362160832,-0.3171778321,0.0476164445,-0.1519611776,0.4656967819,-0.2278466523,-0.1176769435,0.264672786,-0.0819380805,0.1152106151,0.0929076895,-0.4088464081,0.2679079771,-0.1481287181,0.1757383049,-0.0095773283,-0.0769798532,0.2821514904,-0.0093663502,-0.0344192088,-0.2035983205,-0.0834062696,0.2235188037,-0.1755424738,0.2848979235,0.0168077704,0.2470324486,0.158352837,0.7831040025,0.1292257905,0.0327878557,0.2175579518,0.0052596913,0.3514150977,-0.1918498427,-0.2853290141,-0.0229540728,-0.0158400536,0.0131615028,0.2241443992,0.0619855486,-0.197503984,-0.0958135203,0.1189195067,-0.272631079,-0.22777839,0.1294079274,-0.4907452464,0.0027032555,-0.0313345902,0.2232808024,0.1610920727,-0.0105031244,-0.0947625339,0.0539047793,-0.0218640361,0.0488254204,-0.4401236176,-0.072063297,-0.1526599973,0.1704812795,0.0650078282,0.1534888297,-0.0169308148,-0.1147254631,0.1063382477,-0.0172792077,0.8683209419,0.0326389782,-0.0398173109,0.1215237901,0.0219677035,-0.0544051416,-0.0775596499,-0.1911672056,0.4157630801,0.0789694488,0.118790105,-0.3246720433,-0.0966298878,0.3375362754,-0.0949487388,-0.0662097633,0.0525599942,-0.0984244421,-0.1475110799,0.1238877252,-0.037711814,0.0867922306,0.3334338069,-0.1576208025,-0.3991945982,-0.1103441641,-0.4536967874,0.3376243412,-0.0919183642,0.6031657457,-0.2223276943,0.2215555161,0.0338773057,0.0620377921,0.163119629,0.6068787575,-0.2022864968,-0.0529225133,0.126764372,-0.4122452736,0.3593854308,0.2898715436,-0.122519955,0.2348027527,-0.1102875546,0.1477235407,0.2332483083,0.1892338097,-0.0167318061,-0.0478886962,-0.2788283825,-0.6171287298,0.1384761482,0.0598717928,0.1884882003,0.3024021387,0.2127149403,-0.313711226,0.3486145139,-0.2725932002,0.5921763778,0.2279917449,0.0438746624,0.0887681171,-0.0557851307,0.2563143373,-0.2995692492,0.2728048563,-0.1660836637,0.1047546118,-0.1470031291,0.0285783224,0.1856596023,0.4128020108,-0.2262186557,0.2714931071,-0.0224590432,0.1870783418,-0.2246891856,0.0352174193,-0.1360756308,-0.0008034711,0.1077120975,0.1070405915,-0.1398872137,0.1070436686,0.0506391041,-0.1895365268,-0.2601568103,-0.2190781534,-0.1861041933,0.1215514168,-0.2075137943,0.3387215734,0.0884593874,-0.5442290902,-0.0861151814,0.0376271121,0.1340201795,0.0693047866,0.041676715,0.3390548825,0.0296718795,-0.0141510786,-0.2532439232,0.0125055863,0.3496006429,0.0754997581,-0.391066283,0.0725511089,-0.4722877443,-0.347245574,0.1875555664,0.3271017373,0.2945356369,-0.2646553218,-0.3012766242,0.2109841555,0.077639468,-0.2029530853,0.0860263258,0.0515347421,-0.3947243094,0.1002955735,0.0873533413,-0.3391228318,-0.154499352,-0.0792507902,0.2447081059,0.2802988291,0.5937029123,0.0568265691,0.1011832878,-0.3393918872,0.1853217036,0.549046278,-0.2614412606,0.2128617465,-0.0905142054,-0.0102982447,0.1903695911,-0.0247823391,0.115178898,0.511980176,-0.316973567,-0.3421567678,0.0863105655,0.2374942005,-0.0332246162,0.542303741,-0.1008196101,0.1273095608,0.0738786459,0.1382212937,-0.1998396814,0.2956891358,-0.1303074807,0.2745501697,-0.151112929,-0.0046649412,0.301702261,-0.1407115161,0.02484967,0.0927410945,0.054685019,-0.1928481311,-0.1149365678,0.145131439,-0.1180891022,0.1305853277,-0.320877701,-0.0045679919,-0.2357295156,-0.179053247,0.275293082,-0.0488189198,0.0758124441,0.3523583114,0.5218029618,0.1811956018,-0.4933082163,0.1370322555,0.0181099847,0.1743115485,-0.1067427099,0.1228647456,0.1876126975,0.1868083626,0.0154261487,-0.1866953522,-0.0376754701,-0.2877111733,0.293887645,-0.2934230566,0.1822538972,-0.0972167924,0.4907480776,0.2497202009,0.0945034102,0.1780165583,-0.0638180375,0.1687703878,-0.0699116513,0.2082051933,0.0441996865,0.2500332892,0.0892053097,0.2784087956,0.1207119673,0.2616220117,-0.105404444,0.2577394545,0.3951508999,-0.139470458,0.5480863452,0.3780357838,-0.2128003389,0.0298310816,0.1767873466,-0.1068792343,0.0498505719,0.5641690493,-0.0350950025,0.7496700287,0.3097232282,0.1178717092,0.0220395513,-0.2546595931,-0.2642491162,0.0667234734,-0.1393632293,-0.0128742261,0.1237914786,0.7728577256,-0.5248380899,-0.4285672903,0.0550763272,-0.1308665276,-0.2159280181,-0.0293585006,0.0361305289,-0.1824024916,-0.3617464602,-0.1615842134,0.3612845838,-0.2226101458,0.218494907,-0.0202757027,-0.1748887599,-0.2373883575,0.1977123022,0.2362217009,-0.0220266674,-0.1897076815,0.1611776501,0.0687884167,-0.2228153795,0.1481445134,-0.0129223084,0.175176248,0.1693328023,-0.2550384998,-0.0734548494,0.1955959648,-0.3825569451,-0.0258781034,0.1070540324,0.0610248037,-0.249984026,0.1637921184,0.1491637677,-0.1430628598,0.2668722272,0.1677126288,0.4939667284,-0.4116888046,0.5075230598,-0.2429688424,-0.314692229,-0.1424963474,-0.1737228632,-0.5078582764,-0.1696650833,0.1771786362,-0.2113063633,-0.0921531245,-0.1402797103,0.0238409173,-0.0835376829,0.2316112965,0.0506577455,-0.1956673115,-0.2434002161,-0.1726279259,-0.5613914132,-0.048939798,-0.1307417303,0.0438064151,-0.111396946,0.4335915744,0.0071659149,-0.0571629293,0.247338593,0.0503176786,0.2724482417,-0.0084218523,-0.449344337,-0.0424110554,-0.1891454309,0.0121546872,0.0634559616,-0.3176318109,-0.1126365438,-0.0041291257,-0.1844093651,0.0583116524,-0.3946798742,0.3809708357,0.0556883253,0.1241075248,0.0109634595,0.3038922846,0.136105597,0.2473127395,-0.141997844,-0.1855760366,-0.0968350321,0.1947021782,0.1502645761,0.5261270404,-0.1752500683,-0.0080487188,-0.1147482023,0.0516407229,0.1026537642,-0.0092810523,-0.0756682903,0.1626409739,-0.4066336155,0.0679248646,0.3518167138,0.3614502251,-0.1492454112,-0.2877289951,-0.0849026144,-0.3186247349,0.1644227952,-0.2044497579,-0.2382300198,-0.1468767077,-0.0104743484,0.5484179854,-0.2199757397,-0.7831657529,-0.2817843854,0.2548730671,-0.1522158831,0.005790384,-0.079554148,0.1520287842,0.1231076196,-0.1109395176,0.1992077827,0.1201090217,-0.0776762739,0.1051460356,-0.389803797]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/675","title":"Add custom dataset to NLP?","comments":"Yes you can have a look here: https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#csv-files","body":"Is it possible to add a custom dataset such as a .csv to the NLP library?\r\n\r\nThanks.","comment_length":8,"text":"Add custom dataset to NLP?\nIs it possible to add a custom dataset such as a .csv to the NLP library?\r\n\r\nThanks.\nYes you can have a look here: https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#csv-files","embeddings":[0.0775588676,-0.0566500463,-0.1256115586,0.0260554589,-0.0703751743,0.2424899489,0.1162142679,-0.0917605385,0.281412214,-0.2676742673,0.0334090628,0.1616699249,-0.3363847136,0.6732513309,0.7522121072,-0.1350857317,-0.1240647063,0.429327637,0.0964669734,-0.0507278629,-0.0633848757,0.1361282915,0.0185158979,-0.1752955317,-0.0362362042,-0.1084434092,-0.2250680178,0.1384377182,-0.0664304644,-0.4487399161,0.261779964,-0.0204324573,0.2707085907,0.0869756415,-0.0001107963,-0.2628327012,-0.0211409479,-0.1515139043,-0.2518488765,-0.3269862533,-0.0239437297,-0.3652845621,0.0739788935,-0.1221370995,-0.1202088669,0.0363523588,-0.0042682937,0.0977276638,0.1700936109,0.2824110687,0.2567759752,0.1346395165,-0.1142457649,0.0041352059,0.1351333559,0.3959808648,-0.2559067905,0.2333760411,0.1768920571,-0.313834101,0.06913957,-0.1994894147,0.2164935917,-0.0720204487,0.5661614537,0.0607152916,-0.2870441675,0.0192016158,-0.2516106069,0.3968992233,0.5484801531,-0.0743985325,-0.1490182877,-0.3659101427,0.2712460756,-0.5170595646,-0.0035988465,0.445671618,-0.111013025,0.0421171486,-0.3270383775,-0.4381486773,-0.1884178668,0.2032312006,0.3590254784,0.1338698417,-0.1294729263,-0.0401557013,0.1879836917,-0.0811279565,-0.09756203,0.1182313859,-0.2314918041,0.3199084401,-0.0541983955,-0.4730792642,-0.0862495154,0.3110139072,0.1749601662,-0.0408643559,0.3802590668,-0.2313612252,-0.3682580888,-0.0016704083,0.4806701839,0.0633576587,0.0755041838,-0.2430075407,0.0819760114,-0.0251925439,-0.0476045124,0.0104643665,0.1027243808,0.142405659,-0.3298875988,-0.1552078575,-0.0700912327,-0.1713297963,0.0736687705,-0.2487524897,0.0490061827,-0.0450114608,-0.1421192735,-0.0301948208,0.0613118932,-0.2197824717,-0.0140245641,0.1297189444,0.0152756879,-0.2782378495,0.1125229299,0.1565082967,-0.2152211666,0.3599964082,0.2683266401,0.0264663566,0.2556294501,-0.0729989782,0.0015696712,0.0029313997,0.0089485468,0.0765186101,0.0353913866,-0.2195768207,-0.0413792655,-0.0388729125,-0.0976040959,-0.1963849813,-0.3492519557,-0.062978588,-0.1767267138,-0.1172613129,-0.3750561774,0.0853108913,-0.2939235568,-0.5153973699,-0.2834098041,0.610648632,-0.0300168414,-0.1298743039,0.0102911154,0.0031157089,-0.4993396997,-0.1047919914,-0.16257824,0.0090880888,-0.4061802626,-0.1649597734,-0.084653914,0.1289866567,-0.2680904865,0.3705161512,-0.2440252453,0.204222098,-0.2543714643,0.2297997326,0.4180313945,0.0849524662,-0.0105822384,0.1717072427,-0.0617400929,-0.0050123199,-0.061243102,0.4353130162,0.3030475676,-0.0541681461,-0.1115086004,0.5360097885,0.0100331726,-0.2378062308,-0.1025184914,-0.2391218692,0.4851387441,0.0223476011,-0.3711519539,0.1333415508,0.1073758528,-0.1674319953,0.1551389396,-0.1726332307,-0.0067964019,0.2376814634,0.3435646594,0.3218199313,-0.0223606229,-0.2113587558,-0.2037809491,-0.0839954689,-0.1104697138,0.1744533628,0.5439611077,-0.335442096,-0.17798917,-0.2638439536,0.1026517898,-0.3476859331,0.0413110033,0.0131868841,0.1162269935,-0.0541780032,-0.3343946636,0.1194322109,-0.1976954937,-0.0311139878,-0.3163869679,0.0357133299,0.3566322327,-0.026943095,0.1618965417,0.6866432428,-0.060641706,-0.0514127724,0.1286400408,-0.0129940435,-0.2734517753,0.2273615301,0.5741897225,0.1562145948,0.1398566663,-0.3005142212,0.0632262379,0.1164543629,0.0206078161,-0.0390703753,-0.304241538,0.3030946255,0.0014648254,0.0829678103,-0.0151977921,0.1326185018,0.3196250498,-0.1753004789,0.0044814222,0.0157005824,0.2691348493,0.0117504699,0.3069846928,0.0041259355,-0.5030314326,0.076573424,0.2852231562,0.0675170496,-0.0066327835,0.2697894573,-0.2515368462,-0.2003432214,0.1985690296,0.0334774517,0.4873532057,0.2200846076,-0.1198289096,0.1315952092,-0.1980983615,-0.1662986428,0.0128969252,-0.2820401788,-0.152466327,-0.1945736557,0.3033026457,0.0442486182,-0.3530921042,0.0303130001,-0.285916388,-0.0100170551,-0.17841281,-0.0066163773,-0.0846634805,-0.5216109753,-0.2716724575,-0.1133153737,-0.3984526992,0.0219037719,0.0281850453,-0.1018163338,0.1754820347,-0.1379193515,0.0991933197,0.7299356461,-0.4048038125,0.0334048532,-0.0054349843,-0.2608744204,-0.1277058572,0.1739484519,0.2084726542,0.230064705,0.5157439709,0.1039663702,0.0057491674,-0.2571940422,-0.364250809,-0.1075957119,-0.0123113589,0.3751443923,0.2003322691,0.0418238267,0.0513869077,-0.0798423439,-0.0329398178,-0.0609388873,-0.0434795879,-0.2752324641,-0.1312696934,0.0458953679,-0.1963222623,-0.2151353061,-0.4797278643,-0.1989177167,0.6407945156,0.3357377946,-0.0393862799,0.0741489083,-0.2446672022,0.0360871479,-0.3030263484,-0.2150272727,-0.0911983177,-0.1636633873,0.2925286889,-0.3907861114,-0.3553226292,0.1752773225,-0.285839051,0.2602013052,-0.2456716448,-0.5037968159,0.0058849705,-0.1441662461,0.0880592763,0.0153788477,0.001830921,0.1095384434,-0.1986367404,-0.0044243163,0.0537678115,-0.4234225154,0.1213832274,0.5514384508,0.3530923128,0.2252676636,0.072194308,-0.1570291668,0.5539880991,0.0432723947,-0.30643785,0.4635758698,-0.1624929756,0.320936501,-0.0395596586,-0.276031673,0.1705406308,0.0543980859,0.0410336293,0.3240988553,0.3654115498,-0.094702512,-0.0257425252,-0.0887339786,-0.0963069052,0.0672080591,0.0932609588,-0.0140218027,0.1564693153,-0.0416413993,-0.4380434752,-0.4497656524,-0.5354079008,-0.1823123991,0.3283074796,0.259997189,0.0954031274,-0.4211108685,-0.1547131985,-0.3439140916,0.1173314378,0.076766707,0.1072150245,-0.3053105772,0.2260614485,0.0668340549,0.3516839445,0.2493275553,-0.1067824513,-0.3824917376,0.0858204663,-0.0241574757,-0.0523822866,0.3113621175,0.080947347,0.2358848155,0.3028995395,0.3901837766,0.143347308,0.0508843996,0.298489064,0.2404459268,-0.284276396,-0.5547896028,0.1437440217,0.0037382445,0.0159027297,-0.1056059003,-0.2010005116,0.0463373251,0.1468924731,-0.191443935,0.1141133755,-0.0850100294,0.3719226122,-0.0277430061,-0.1457688063,0.3619446754,-0.0384067707,-0.140002057,0.1322872043,-0.3999531567,0.3720891774,-0.1413231343,-0.2353494167,-0.3332659006,-0.2779295743,0.7403286695,0.030234633,0.2221503556,0.5406273007,-0.1428799778,-0.1326203942,-0.3662570119,0.0392726175,0.0690493435,0.0836616307,-0.1957935989,-0.3911427259,0.5545285344,-0.1625733376,0.1009237394,0.1001631469,-0.0284444802,-0.3380544782,0.0474170446,0.0415390767,0.8102702498,-0.1718854308,0.4492128491,-0.0904891491,-0.1277540475,0.9267758727,-0.087061286,-0.0211338196,-0.3327535093,0.0929961503,-0.0922831595,-0.0448918268,0.1702288985,0.3052452803,0.0037798453,0.3782992363,0.3005853295,0.1206910685,-0.0635171235,0.4277610183,0.1763828993,-0.3847832084,-0.5538643599,0.1244023442,-0.0672208294,0.0045066448,-0.1738230288,-0.3498322368,-0.1250611842,0.0033066338,-0.3231876194,-0.2310388982,0.1421574056,-0.3760822117,0.3682158291,-0.4656081796,0.4713041782,0.2949881256,0.2567002773,-0.0506609641,-0.4115624428,-0.1037210152,-0.5036485195,-0.0069893305,0.0046939803,0.3530621827,0.006788834,-0.0751514882,-0.0582916476,0.1092181131,0.1230894104,-0.5109474063,-0.1076117828,0.1918686479,-0.0315937325,-0.2812797725,-0.3143726885,0.1285663098,0.1052999869,-0.069382675,0.1535828263,0.2200227678,0.0316355936,-0.0429257527,0.3324000239,-0.0003322211,-0.1222936735,0.019113183,-0.077919662,-0.12871705,0.3715503812,0.0906048119,0.030723216,-0.1900788993,0.1290351897,-0.0589607134,-0.2155820727,-0.092363216,0.4586572349,0.1093415618,-0.236688301,0.265126884,0.0807306319,0.0592210889,-0.0861199051,-0.2646733522,-0.3277939558,-0.0397505797,-0.2711414397,0.0826283023,0.1692409515,0.3065602779,0.1543675065,0.2509973049,-0.2773097754,0.0783152282,-0.0416826196,0.1116833612,0.6355682611,0.2288804054,-0.3173076212,0.0704065487,0.0197478868,-0.0988360122,-0.0634483621,-0.2268832922,-0.2224021256,0.1278135031,0.2310259938,-0.1348657161,0.046206031,-0.0069157314,-0.0160481036,-0.1626290679,-0.2796970606,-0.196663186,-0.0559004247,0.4375052154,-0.1880384386,0.0189923663,-0.1027008146,0.2936935723,-0.0953935608,0.3794404566,0.1414951235,-0.017849287,-0.0247769803,0.2432526052,-0.1801389307,0.076406084,-0.0830067173,-0.0357462652,0.0364451818,0.1747532189,0.2613742948,0.1623649597,0.2013068497,0.1422980726,-0.147576347,0.3566319644,0.5317546129,0.1040558815,-0.2674814463,0.2147462517,0.5399845243,0.1690997928,-0.0608614273,0.2528953552,0.3857486248,0.0968497619,0.0965160131,-0.3320907354,0.1057588458,-0.1771353483,-0.0023715617,-0.0157553405,-0.0952633023,-0.0505031943,0.0934331343,0.1950291991,0.2439582795,0.3640795648,0.1654857844,0.00809051,0.1443014443,0.1010067388,0.1842868626,-0.1806146801,0.1497448832,0.2120132297,0.5875739455,0.1764431149,0.2001717389,0.8229294419,-0.2091724277,-0.10860461,-0.091086328,0.344653964,0.0514752045,0.0099935131,-0.0024968269,0.1135105118,0.0259347856,0.0785143077,-0.1623707712,-0.2813535035,0.0468161777,-0.2133954018,0.0927779004,-0.3370256722,0.2226427346,0.1689711213,-0.0391505398,-0.0085231625,0.1482908875,-0.3518048525,0.2194603682,0.20916605,0.3167611063,-0.0420827158,0.1668914855,0.1987276077,0.1854719967,-0.3132882714,0.1532227844,0.0353523605,0.2433640659,0.0817299858,0.09628997,0.156498298,0.1070316881,-0.0389220864,0.3910911381,0.106501244,0.0443232134,-0.523757875,-0.0560768619,0.141556859,-0.2613912821,-0.4353482127,0.1091605499,-0.0786768943,0.0219052322,0.1268840134,0.1486736238,0.0874332711,-0.0391653292,0.0800713077,-0.1687254906,0.287006557,0.4402600527,0.3506292403,-0.0162521638,-0.2136728764,-0.6035621166,-0.0822869614,-0.0482378714,-0.3018891513,-0.2500346899,-0.0923250243,0.0275791027,0.1333278418,-0.1367410272,0.0499149337,0.1495786756,-0.1008412912,-0.1232595742,-0.1770251542,-0.1860551685,0.24054645,0.019936271,0.2670100033,-0.017300969,0.1601596177,-0.054902263,-0.2000137269,0.2097456604,-0.1976076961,0.2843711674,0.1540641338,-0.0917212591,0.392747283,-0.0093359528,0.0528407544,-0.1029452756,-0.1924402565,-0.1474255025,0.1173743978,0.0482888073,0.1915504634,-0.4278292358,0.2386998534,0.0424327292,0.2137159109,-0.1585284919,0.0355500355,-0.1404360533,0.1836489737,-0.1764484942,-0.1017247438,0.1943928748,0.2418884784,-0.1971387863,-0.192367658,0.1130754575,0.2830001414,0.1534094214,-0.3416302204,-0.1167089194,0.3161066175,0.0672189295,0.0863930732,-0.3088625371,-0.5332424045,-0.2029330432,0.4037673771,0.0293057449,0.3578562737,0.4019823968,-0.3012072444,-0.0618310124,-0.049261909,0.1737072617,0.1436092108,0.1735672355,0.0366602689,-0.2972665727]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/675","title":"Add custom dataset to NLP?","comments":"No activity, closing","body":"Is it possible to add a custom dataset such as a .csv to the NLP library?\r\n\r\nThanks.","comment_length":3,"text":"Add custom dataset to NLP?\nIs it possible to add a custom dataset such as a .csv to the NLP library?\r\n\r\nThanks.\nNo activity, closing","embeddings":[-0.0246852487,0.2721365988,-0.1693947464,-0.0960473865,-0.140953213,0.1985376179,0.164521724,-0.0333243981,0.122163035,-0.2603679597,0.0634898692,0.3089755177,-0.3101531863,0.3809611797,0.5698917508,0.010903623,-0.1581757963,0.4338408113,0.1714897156,-0.0899741799,-0.1088153347,0.0890194923,-0.0500819422,-0.2326126248,0.012981669,-0.1840332001,-0.3109160364,0.0216665585,0.0503891744,-0.464855969,0.1695680618,0.0495735332,0.0829865485,0.0443414226,-0.0001039244,-0.3218532205,0.0944960192,-0.0797037706,-0.3124259412,-0.2543268502,-0.1262194961,-0.3405673206,0.0744002163,-0.2056007385,-0.1044490412,0.0506827384,0.0459873453,-0.0611810498,0.056697201,0.31162709,0.3188592196,0.037134774,-0.0133301038,-0.1320730895,0.1730589569,0.193999052,-0.2667230666,0.0830561444,0.2120901048,-0.2655239701,0.0584843792,-0.2540983558,0.1990999281,-0.1035532504,0.3580239713,-0.0985715911,-0.2730415761,0.0850539505,-0.0572604872,0.3085813224,0.6531471014,-0.0452801995,-0.0890895128,-0.1547316015,0.3483633101,-0.5059443116,-0.0154973213,0.4471132457,-0.1188446432,-0.0521467887,-0.1563750654,-0.3904550076,-0.1722864062,0.1726415157,0.302403897,0.2724383473,-0.0278637465,-0.0570722744,0.0692740977,-0.1693551391,0.106343627,0.1583901793,-0.1631380618,0.1945649534,-0.0970731229,-0.5344083309,0.066948168,0.146524176,-0.0113134999,-0.0421860404,0.3797008395,-0.0531337596,-0.1696348041,-0.0040113172,0.4268395007,-0.0091086626,0.0658400804,-0.2488344014,0.0445798784,-0.2266687304,-0.2248475701,0.0954133794,0.1112821773,0.1264733523,-0.2264544815,-0.084489651,-0.0175539628,-0.1493349224,0.0124622751,-0.2063044161,0.0008548694,-0.0315816067,-0.202364713,0.0515983663,-0.0153309125,-0.1953684241,-0.1278804541,0.0548482686,-0.0342016816,-0.2542394996,0.0031585095,0.2893096805,-0.2754404247,0.1385729164,0.3936661482,0.2977691591,0.2072262168,-0.2248698026,-0.1811930537,0.0975299627,0.1424744874,-0.0072974064,0.1315401047,-0.2434982657,0.0070423437,-0.0326196142,-0.2594222128,-0.1769958436,-0.2777086198,-0.0070215706,-0.10118258,-0.0484111384,-0.4054218829,0.1948550642,-0.2956242263,-0.5886354446,0.0481817201,0.4805569351,-0.0042753611,-0.1434326023,-0.1506635547,0.0043633864,-0.4138018489,-0.1670317352,-0.2586372197,-0.0384746827,-0.4280094206,0.0247867033,-0.2856617868,0.1836133599,-0.1178864017,0.250850141,-0.1946612149,0.1438765675,-0.1640172303,0.2355490029,0.4564310908,0.1322707981,-0.1237746552,0.1252546012,-0.0999490842,-0.2862580717,0.0169130452,0.4900789857,0.2143241018,-0.1137451828,-0.0285949111,0.445974946,-0.0736425146,-0.216039151,-0.1384481043,-0.2348207384,0.502715826,0.1157157794,-0.1813714355,0.0821674839,0.2012870163,0.0774092227,0.1484279633,-0.0924983844,0.061406564,0.2432142645,0.3862870336,0.1460861564,-0.0674331784,-0.2225354016,-0.0373285301,-0.1804953218,-0.1361526251,0.3128543794,0.5657021999,-0.1476721764,-0.1946200728,-0.2427160144,0.1383232772,-0.3551922441,0.1833327115,0.1296024621,0.0096900146,-0.0395015627,-0.3007885814,-0.0085225133,-0.1401360333,-0.2271460742,-0.1467251927,0.0086712129,0.3027338982,-0.1357878149,0.2279568464,0.6698915958,-0.0552393943,0.0496963114,0.1091049686,-0.0279300213,-0.2216759324,0.2507086396,0.6161807775,0.0877701864,0.0648754239,-0.3078438938,0.030718267,0.2034991086,-0.0377116017,-0.0052570412,-0.4222348034,0.2371030897,0.0161598213,-0.0279775131,0.0909236446,0.0873376578,0.3488438725,-0.3161099255,0.0227293875,-0.0056161731,0.1427417845,0.0796404481,0.2110400051,-0.0164655391,-0.4958158135,0.1931754947,0.3322006166,0.1325723082,0.2262517661,0.1601351798,-0.1934612244,-0.2535790503,0.2574088871,0.2357343435,0.5376749635,0.3102097511,0.0346343741,0.0830808431,-0.2494507283,-0.2400306314,0.0538378246,-0.2221221924,-0.1143376455,-0.0966756865,0.3948643208,0.0811165422,-0.3476854563,-0.0454007946,-0.1863285899,0.0198741611,-0.0391601063,-0.1605142057,-0.223210305,-0.4384857118,-0.1360625625,-0.0015994468,-0.1757171601,-0.0856172293,0.1906778961,-0.0440264754,0.0481534712,-0.1306576729,-0.1208184808,0.6042627692,-0.3373201787,0.0238803867,0.0689422712,-0.3376608789,-0.1856709421,0.2890725732,0.1607080251,0.2856096327,0.5189067125,0.2577670515,0.0518450886,-0.1724942327,-0.3947350681,-0.0966190621,0.1642819792,0.2501286268,0.1050464436,-0.0546616763,-0.019397635,-0.1043699533,-0.0856619999,-0.0631399527,0.0141719589,-0.2887066007,-0.1155504361,-0.036719501,-0.2269544005,-0.443038404,-0.4617425203,-0.2956878543,0.4416509867,0.3239281178,-0.0050152615,0.1830353141,-0.2560373843,0.0588296764,-0.148912102,-0.2795785964,-0.1184377745,0.0411307774,0.3216764331,-0.4376955926,-0.3990927637,0.2586284876,-0.3845588863,0.4693359435,-0.1753715426,-0.4421000183,0.2945881784,-0.0543801077,-0.0164005272,0.0484241508,-0.0872547999,0.1458687484,-0.1231255978,-0.1614573449,-0.005731015,-0.3422439396,0.1494919956,0.4729136229,0.3996224403,0.0915660858,-0.1059177965,-0.1125054732,0.4529488981,-0.1334289014,-0.3053820133,0.406352818,-0.0700435266,0.2264881879,-0.022921348,-0.2507359982,0.1698897332,0.1192452684,0.0366997682,0.3236517608,0.271286875,-0.381570667,-0.0433127098,-0.0455437787,-0.1203955635,-0.0140453167,0.1255347282,-0.059151154,0.2015746385,0.048150517,-0.3874826133,-0.4675212502,-0.5137468576,-0.2667708695,0.0938187912,0.0827669203,-0.0941109359,-0.4561622441,-0.1450974196,-0.289929837,0.0400604792,-0.0211185869,0.2730748653,-0.3617060781,0.1949138492,-0.0107112387,0.2560217679,0.08529526,-0.2469847947,-0.2434651107,0.3072559834,0.1894251704,-0.0675998479,0.285972178,-0.0283592343,0.2705319524,0.3359294236,0.2513705492,0.2004014552,0.1837732941,0.3231347799,0.0242119711,-0.1474047899,-0.5249981284,0.2831519842,-0.0602713041,0.0679545999,-0.0586248785,-0.1988760382,0.0662800521,0.1008312106,-0.2645625174,0.0892810076,-0.114855513,0.3978985548,-0.1081315055,-0.1931193769,0.4888555408,0.0629684702,-0.4390733838,0.1477188915,-0.4123088419,0.1791461706,0.069370158,0.0163788032,-0.0169642288,-0.4071082473,0.6774002314,0.0294782836,0.2511779964,0.5810739994,-0.2724207938,-0.1723959148,-0.1244460791,0.0957418159,-0.0275955535,0.0127651803,-0.0872960687,-0.3797627985,0.3709666133,-0.2156876922,0.0236095749,0.1017266586,-0.2537083626,-0.3308169246,0.0757204145,0.1183183044,0.7347449064,-0.0095282365,0.3428400755,-0.0398861505,-0.1669938713,0.7713696957,-0.0878919587,0.0484156087,-0.3166424334,0.1967035383,0.0196030997,-0.0538259409,0.1984357983,0.4225087166,-0.0164726786,0.2747951448,0.3508663177,0.0083968844,0.0328202285,0.4729342163,0.1031668931,-0.3084146082,-0.5083747506,0.210156098,-0.0664832219,-0.0785811543,-0.0992129222,-0.358060509,-0.1783571094,0.1260504574,-0.2862391174,-0.0684378743,0.218135193,-0.2530501783,0.153743118,-0.3726985753,0.4663243294,0.1521077901,0.2903245389,0.0907128602,-0.3656150699,0.042472288,-0.3776226342,-0.0618493669,0.1614119709,0.3197495043,0.0973491669,-0.1948238164,-0.1415038258,0.1542900652,0.0712516159,-0.4752052426,-0.1446366459,0.2077693939,0.1553182453,-0.2464947402,-0.2605574429,0.1782981157,-0.0248682536,-0.0618023947,0.2312707454,0.1949729621,0.1150978729,0.0881908759,0.2863702476,0.0993767157,-0.1318785101,0.0424041785,0.1077272147,-0.0620020144,0.5071240664,0.0657510161,0.0882968456,-0.3349186182,0.1031305566,-0.2139761746,-0.0973308831,-0.0753853992,0.2809960544,-0.0052765845,-0.0356880985,0.1731676459,-0.0351127088,0.2356858402,-0.1704820395,-0.27320382,-0.2981154323,-0.0742990002,-0.2371469885,0.101073049,0.084624961,0.2633987069,0.2473420054,0.1135562062,-0.381282419,0.0116891656,-0.0965795293,0.1257996559,0.4738414586,0.1071376279,-0.4320290089,0.0634417385,0.1081693619,-0.1510082036,-0.0274590403,-0.3134797215,-0.1060293168,0.0551722534,0.1837221086,-0.0258557256,-0.0507009737,-0.1543504596,-0.0744200274,-0.0142359193,-0.4059553146,-0.2841967344,-0.093995899,0.4376886487,-0.2048701346,0.032410264,0.0243681073,0.2916100323,-0.0729093552,0.2626363337,0.0713959262,0.1383384317,0.137764588,0.2071292847,-0.2504392266,-0.045756001,-0.2730670273,-0.0639777407,0.0671597794,0.0730416849,0.1266888529,0.1274456382,-0.0045316881,0.0401717015,-0.0482170656,0.2606568635,0.4675655663,0.2337225527,-0.2975882888,0.246946618,0.5079210401,0.1500812769,0.0788907483,0.1880045086,0.3647298217,-0.0440945178,0.1197644323,-0.1913508475,0.0924788341,-0.3511106372,-0.0011780204,0.0120781902,-0.1965670437,-0.1065239161,0.0716196224,0.0790385306,0.17761904,0.4563633502,0.3603435755,0.2475412637,0.3561000824,0.0542956367,0.1751889735,-0.056054689,0.2001927644,0.1959909946,0.4949415326,0.1797489971,0.2184648514,0.6061689258,-0.1177618504,-0.0019219283,-0.0335490368,0.235518828,0.0500024483,0.0860325024,-0.1540162414,0.0254514348,0.0498644449,-0.0251439475,-0.1926139444,-0.190315485,-0.0778069869,-0.2634140253,0.036120899,-0.1345700324,0.350997448,0.1511772275,-0.0953017846,-0.0403566286,0.1372587085,-0.373010397,0.26259166,0.0877016336,0.3519905508,-0.057782758,0.1842782795,0.0077355127,0.1066294685,-0.1099711061,-0.0457656458,-0.0961577669,0.3193364143,0.2096555978,0.1690178365,0.1813977957,0.1997787356,-0.1114991233,0.4685297012,0.1964374781,-0.0034208,-0.455044508,-0.0579553135,0.2478536665,-0.4083847404,-0.2654701471,0.1538335234,-0.0491784401,-0.1813121587,0.1186719239,0.0813573003,0.0917286947,-0.1443740726,0.1474887282,-0.2473029494,0.3097587526,0.2068371624,0.3228884041,-0.0265527219,-0.298115015,-0.5479794145,-0.0429098047,0.0376353525,-0.2904359102,-0.2704144716,0.0650611669,0.0457327105,0.0926338062,-0.1580069065,0.1130649224,0.1292542219,-0.0296698771,-0.1693439633,-0.0161890164,-0.0055368962,0.2955280244,0.0059941267,0.1810116917,-0.0577668287,0.1430981755,0.0590551458,-0.1916663051,-0.0074161929,-0.0690815523,0.3769579828,0.1758186817,-0.11085549,0.5135433078,-0.0434629358,0.0193489231,-0.221256271,-0.2334871143,0.0284574423,0.0951418951,0.1655909866,0.2264915258,-0.2504951656,0.3891206682,0.2319185138,0.1416527033,-0.3056280613,0.1270947903,0.0209905766,0.2256880701,-0.1602775604,0.1102835387,0.0793493167,0.041438289,-0.1691362709,-0.2112443447,0.0222097412,0.1683037132,0.0099759484,-0.3777600527,-0.2323386967,0.1417543143,0.1396994442,-0.1582944542,-0.1827895194,-0.2834591269,-0.2367718518,0.4009599686,-0.0117526045,0.360245049,0.340446353,-0.3079028726,-0.015717851,0.0013454129,0.1487043202,0.0809452236,0.1337190121,-0.1450617611,-0.3186116815]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/674","title":"load_dataset() won't download in Windows","comments":"I have the same issue. Tried to download a few of them and not a single one is downloaded successfully.\r\n\r\nThis is the output:\r\n```\r\n>>> dataset = load_dataset('blended_skill_talk', split='train')\r\nUsing custom data configuration default <-- This step never ends\r\n```","body":"I don't know if this is just me or Windows. Maybe other Windows users can chime in if they don't have this problem. I've been trying to get some of the tutorials working on Windows, but when I use the load_dataset() function, it just stalls and the script keeps running indefinitely without downloading anything. I've waited upwards of 18 hours to download the 'multi-news' dataset (which isn't very big), and still nothing. I've tried running it through different IDE's and the command line, but it had the same behavior. I've also tried it with all virus and malware protection turned off. I've made sure python and all IDE's are exceptions to the firewall and all the requisite permissions are enabled.\r\n\r\nAdditionally, I checked to see if other packages could download content such as an nltk corpus, and they could. I've also run the same script using Ubuntu and it downloaded fine (and quickly). When I copied the downloaded datasets from my Ubuntu drive to my Windows .cache folder it worked fine by reusing the already-downloaded dataset, but it's cumbersome to do that for every dataset I want to try in my Windows environment.\r\n\r\nCould this be a bug, or is there something I'm doing wrong or not thinking of?\r\n\r\nThanks.","comment_length":41,"text":"load_dataset() won't download in Windows\nI don't know if this is just me or Windows. Maybe other Windows users can chime in if they don't have this problem. I've been trying to get some of the tutorials working on Windows, but when I use the load_dataset() function, it just stalls and the script keeps running indefinitely without downloading anything. I've waited upwards of 18 hours to download the 'multi-news' dataset (which isn't very big), and still nothing. I've tried running it through different IDE's and the command line, but it had the same behavior. I've also tried it with all virus and malware protection turned off. I've made sure python and all IDE's are exceptions to the firewall and all the requisite permissions are enabled.\r\n\r\nAdditionally, I checked to see if other packages could download content such as an nltk corpus, and they could. I've also run the same script using Ubuntu and it downloaded fine (and quickly). When I copied the downloaded datasets from my Ubuntu drive to my Windows .cache folder it worked fine by reusing the already-downloaded dataset, but it's cumbersome to do that for every dataset I want to try in my Windows environment.\r\n\r\nCould this be a bug, or is there something I'm doing wrong or not thinking of?\r\n\r\nThanks.\nI have the same issue. Tried to download a few of them and not a single one is downloaded successfully.\r\n\r\nThis is the output:\r\n```\r\n>>> dataset = load_dataset('blended_skill_talk', split='train')\r\nUsing custom data configuration default <-- This step never ends\r\n```","embeddings":[-0.4348374605,0.2638886869,-0.0968250707,0.3132835925,0.2672564983,0.3075503707,0.3540579677,0.1106310487,0.4425596595,-0.0165286697,0.1196889281,-0.1064612344,-0.0671281815,0.2139161676,0.1405703574,-0.1002630293,0.018564146,-0.0331001058,0.0595737882,-0.1480202377,-0.2890079319,0.1217577606,-0.4512639344,-0.1942238808,-0.1455354691,0.1002141535,-0.1550091505,0.0938178152,-0.128532812,-0.2223109603,0.4647398591,0.1681256443,0.2233542353,0.5532972813,-0.0001173147,-0.1689022481,0.4234569371,-0.0814600661,-0.2416763306,-0.2221424282,0.0686217472,-0.2787030935,0.0384051986,-0.043393191,0.0327521302,0.3367926478,0.0414203741,-0.2054827213,-0.0061557875,0.3664603829,0.1425715834,0.0846243128,0.0911352858,0.0108170155,0.0476456061,0.1078762114,-0.1996514201,0.4818440974,0.1704382598,-0.4610636234,0.2528343797,-0.0523954928,-0.2002899647,0.12531811,0.0094174054,0.0281769503,-0.0543065406,-0.5504904985,0.1858800203,0.3046257794,0.6023476124,0.0177001376,-0.1578363329,-0.1204510257,0.2528926432,-0.0260488745,0.2037484646,0.5220000148,-0.4007895291,-0.025911171,-0.3000251949,-0.1521255225,-0.0877072215,0.3059648573,-0.0330718718,0.0428854339,0.0164353736,0.2268963754,0.2488449812,0.0721496046,0.1917523444,-0.1281947941,0.0682789981,0.027903039,-0.1809323728,0.32985425,-0.0504387878,0.2989700735,-0.0418128595,0.1580298245,-0.0601914823,0.1740888655,0.3321507573,0.0459657125,0.2691091597,-0.099441953,0.2820907235,-0.0252827834,0.3245888948,0.121104911,-0.082859233,0.0667497888,-0.501216948,-0.194568336,0.052375149,-0.0869862512,0.1955155581,-0.3377199173,-0.2771073282,0.0688086748,0.1445810199,-0.1822151542,-0.0500968657,0.2808072567,-0.1014573798,0.4788323939,0.1540380716,0.2452641428,-0.2062827498,-0.1190844253,0.1272698939,-0.0692254156,-0.3084921837,0.1779513508,0.7094912529,-0.0529173873,0.0663962662,-0.1264293194,0.1003017649,-0.0624417365,0.0906467587,-0.2585457861,-0.0227721445,0.1866897345,0.1599061638,0.4259392023,0.0041597588,-0.035244897,-0.1602686942,0.2484574318,-0.1696225107,-0.0448931418,0.2376054525,0.0982200801,-0.1863003373,-0.2393067777,-0.1790407002,0.1990797669,-0.0354634225,-0.0408191569,-0.0609379299,-0.2394403219,-0.2222975194,-0.2471050322,0.2509646714,0.6065970063,-0.4733546078,0.1334774196,-0.0128212413,-0.0194936916,-0.0189592876,0.1573492736,-0.2292907238,0.2850937545,-0.3361421227,-0.0334928222,0.0889084116,-0.2796662152,-0.3336099982,0.4279620349,-0.0389764234,0.1260471195,-0.0503798984,0.1954921484,0.1609428674,-0.0208935887,0.1018467024,0.4022238553,-0.0085760411,-0.1034732461,-0.1004038155,-0.1858472526,0.3791280687,0.415197134,-0.0585369058,0.1164777875,-0.0023768656,0.1477486342,0.2085295767,0.2583529651,0.1554789841,0.3921535909,-0.100433819,0.194047302,0.0573909357,-0.0040819296,-0.5138206482,0.2298600376,0.3814336658,-0.1514694244,0.1052555069,-0.0277960431,-0.2260430306,0.0215200167,-0.0843482167,-0.2146155089,-0.0222895797,0.0955006182,0.3066940308,0.0342207812,-0.0121396556,0.4903284907,-0.1665464789,0.0066171535,-0.2134784907,0.2366082519,0.0909736678,-0.2106549144,0.2085710019,-0.0699847564,0.1091529354,-0.0718648434,-0.1570317298,0.2500483394,-0.0906808674,0.0076134386,0.0512502566,0.0115301656,0.0814980119,-0.3572994471,0.205064401,0.392457515,0.2962708473,-0.3245432973,-0.2591126859,0.0580980293,-0.2606731355,0.33670789,-0.005331513,0.0617840737,-0.0374517217,-0.0032223789,-0.1617287099,0.2174919397,0.5736041069,-0.104907535,0.0363553129,-0.1144829094,-0.1393821537,0.1269520819,0.4384441376,-0.1317360848,-0.0689161569,0.1883349121,-0.2489185929,0.1089598984,-0.0910281092,-0.0266572293,0.6324250102,0.1077284664,0.3119974136,0.0097935377,-0.1067500561,-0.3897905052,0.0751158372,0.0054400098,-0.12851955,0.2242888361,-0.2907730937,-0.0353042223,-0.1567242891,-0.0604768358,0.2551260889,0.1556317061,-0.254458487,0.1504667997,-0.1579031944,0.0311158001,-0.1553646028,-0.15486525,-0.2051059902,-0.1602920145,-0.3846489787,0.4122015834,0.2328325957,-0.1524040848,-0.5157061219,0.1887485832,0.1120762229,-0.3764174581,-0.0860777199,-0.0969321728,-0.3495421708,-0.0422155,0.368955493,0.1660794914,0.1740034223,-0.2884266973,-0.0160381626,-0.1641294658,0.1684151888,-0.1408004761,-0.0490464754,0.212425679,-0.1679721773,0.5251682997,-0.0327252001,-0.0148316082,0.163998723,-0.3974747658,0.1488021761,0.1389051676,0.273614943,-0.1533205211,-0.0900415853,-0.3636281192,-0.2869464159,-0.3258349299,0.1719224453,-0.0516751409,-0.0136036305,-0.0170345493,0.1706237793,0.0451981612,0.0707797706,0.0161171295,-0.1132758409,-0.1930204034,0.675116241,-0.1407323778,-0.6298113465,0.4133820236,0.1727474481,-0.112663798,0.2265709341,-0.5824322701,0.4229878485,-0.1837514937,0.1028695479,0.1750198752,0.1629975885,0.0262151174,-0.4262453318,0.1226002872,-0.009976618,-0.335233748,-0.1860522628,0.0036444147,0.3259528577,0.1315894127,0.1866594106,-0.1926443875,0.2649832964,0.1222916096,0.1034580246,0.116020605,-0.0910001919,0.4476820827,0.0732973143,-0.3743332922,0.3017482758,-0.4511118531,0.0208771862,0.243429482,0.0960293114,-0.3561885059,-0.4275498092,0.2047991902,-0.0676822215,-0.252379328,0.3047461212,-0.2845537961,0.0341998972,0.0155000128,0.3321449161,0.2220671475,-0.3462550938,0.1259207875,0.3952535689,-0.138132304,0.1101623848,-0.3709639609,-0.1167592183,-0.3521858752,0.0027283423,-0.2510640621,0.5077358484,-0.2720744908,-0.0153364455,0.0953399837,-0.0960849449,0.6394124627,-0.2933663428,-0.0067243311,0.2766592205,0.2527191937,-0.4101142585,-0.2863557041,-0.3217037022,0.1813432723,-0.0756404772,-0.0983816087,-0.0939827934,-0.0081345299,-0.0319498181,0.3258681297,-0.0518057831,-0.1561313719,-0.3686889112,-0.412083298,-0.2425460964,-0.0851676911,0.0589065216,-0.0259562545,-0.4836912751,0.0609271191,0.0045133317,0.126207605,0.1818000227,-0.2490932196,0.158599481,0.2035549581,0.33368963,0.2094846219,0.1814172268,-0.0510995872,0.6126553416,0.0319113359,-0.0074441815,0.3563105166,-0.3362620175,0.1000961214,0.4206353128,-0.0016221546,-0.297308296,0.1030741408,0.0874824524,-0.3481106758,0.1405719966,0.3611352444,-0.0728051141,-0.3170986176,-0.5456185937,0.2885629237,-0.0102033112,0.0821493268,0.2355408669,-0.2802584767,-0.0697857738,-0.1636259705,-0.2875608206,0.791492939,-0.1276796162,0.3157043159,-0.0315801091,-0.0601148382,0.0311626215,-0.1487036347,0.0579585806,0.0718751252,0.0076463153,-0.1388118267,-0.111723572,0.1739655882,0.4494954348,-0.2361567467,0.3529339731,-0.0425163768,0.168843627,-0.0821297541,0.25844118,-0.2362321764,-0.3395774662,-0.0656648725,0.0950022638,0.111419484,0.5089247227,-0.195538789,0.062819846,-0.0094823958,-0.035400182,-0.4129705429,0.4149441123,-0.1061964706,-0.0348649099,-0.3620473742,-0.2052252591,0.08474534,0.1291707903,0.1205319688,0.0469692685,-0.3560387194,0.2466623038,-0.0813157931,-0.2004201561,0.0379847772,-0.217981413,0.0581302829,0.0361326821,-0.3643952906,0.1239510998,-0.1406413764,-0.0554502346,-0.2276135534,0.0271032304,0.1607457399,-0.153775081,-0.1951226145,-0.0130888419,0.0929647833,0.0349462628,0.0534504317,0.201261878,0.0458813943,0.4033931196,-0.0145655945,-0.1353239864,-0.150363639,0.5252747536,-0.2151469886,-0.4387831688,0.3159073293,0.5385573506,-0.2298115492,-0.297206521,0.0839205757,0.0740603358,-0.2920134366,-0.1127802879,0.1028122008,0.3448622823,0.086360082,-0.0468950383,-0.0064448407,-0.4304711819,0.1015973762,-0.5667569637,0.0215958972,0.066563718,0.3942690492,0.0972845256,-0.1676266342,-0.4079504907,0.625030756,-0.1190573052,-0.190765366,0.088801533,-0.0049997973,-0.0484072343,0.4071675539,0.1431884617,0.1961261481,-0.0266036093,-0.0463298187,-0.0587076657,0.1866006851,-0.0785975233,-0.0118761193,0.1934809387,-0.1028184593,0.0462876335,0.098958835,-0.6630380154,-0.2647500038,0.0398306884,0.2257114947,0.1203350052,-0.1630231589,0.096638307,-0.1720335037,0.3454067707,-0.199025318,0.233397752,-0.3064649105,0.6293855309,0.0703851208,0.0528666638,0.3849901855,0.1681175232,-0.3313408494,-0.0132239955,-0.0591454022,-0.0936452895,0.3078832924,-0.4258176386,0.3199900985,-0.1857276857,0.2074411362,0.8582122326,-0.2658539712,0.0478883423,0.3950186074,0.0771475285,-0.064494811,0.0234608781,0.0666038841,-0.1402131319,-0.0744289756,0.1523786187,0.0200661216,-0.3098987341,-0.2634477019,0.0498912111,0.4344172478,0.2207412124,-0.1367764175,0.4721736014,0.1196809486,0.0527886525,-0.0661485642,0.2855157852,-0.1107043028,0.406011045,0.0934903622,0.2014708072,-0.0194685161,0.4325453639,-0.2531368732,-0.2909341455,-0.0486959144,0.0934745073,0.0326189622,0.2133301049,-0.2366552353,0.3340227008,-0.456348598,-0.2456885576,-0.453268975,-0.0144100301,0.1548987925,-0.0124606164,-0.1058923081,-0.1668462157,-0.1983261555,0.1738777906,0.1210285276,-0.3409792781,0.4674229324,0.1777604073,0.0178955346,-0.4753952622,0.0249211583,-0.0788091794,-0.0989109725,-0.1948540956,0.1853650808,-0.1579914391,0.0507142879,0.1232788712,0.0551744923,0.2349176705,0.3727568388,-0.2836601734,0.3019384444,0.0694615245,-0.058920078,-0.0072927806,0.5686403513,0.1995872259,0.0661858469,0.2364199609,0.0451256484,-0.0677763,-0.0304130185,0.0695494413,-0.1441613883,-0.2572297752,0.0617365427,-0.080319345,0.1083729193,-0.2424018234,-0.1758417636,-0.4169380963,-0.3438284993,0.3497318923,-0.1343279332,0.090949975,-0.2353840768,0.0050638542,0.0340485871,0.6027309299,0.2792074382,0.5272394419,-0.114135839,-0.1882113516,-0.9585389495,0.0234731007,-0.2382193804,-0.2091089636,0.0192716792,0.222942546,0.0303672738,0.4572855234,0.0115599316,0.0761600733,0.1781709492,0.382165283,-0.3112042546,-0.1928497702,0.0062751281,0.4389103651,-0.182726264,-0.4216966033,0.2759920359,-0.1390033811,-0.1636611521,-0.0513791963,-0.1799184084,0.0934566259,0.139980346,0.4547464252,-0.1862626076,0.5130954981,0.1956937313,-0.0569440685,-0.4508399069,0.0985100567,0.16231668,0.1784574836,-0.0569322705,-0.2573368847,-0.4159771502,0.1761966646,-0.0633940026,-0.0654212311,0.0987293571,0.1002532244,-0.1380513757,-0.2078807354,0.080041334,0.2679149806,0.1112214699,0.2325889319,-0.246371001,-0.1765277684,-0.2108065039,0.0051754462,0.2323575467,-0.2249078453,-0.338441819,-0.1001186818,0.1650857925,-0.3213374913,0.2227676362,-0.3573559225,0.1331174076,0.1265269369,-0.0716709718,-0.300235182,0.3131253123,-0.3208013177,0.017322544,-0.0236268714,0.2866987288,-0.0939519554,-0.1967639923,0.0633408129,-0.1457565278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/674","title":"load_dataset() won't download in Windows","comments":"This was fixed in #644 \r\nI'll do a new release soon :)\r\n\r\nIn the meantime you can run it by installing from source","body":"I don't know if this is just me or Windows. Maybe other Windows users can chime in if they don't have this problem. I've been trying to get some of the tutorials working on Windows, but when I use the load_dataset() function, it just stalls and the script keeps running indefinitely without downloading anything. I've waited upwards of 18 hours to download the 'multi-news' dataset (which isn't very big), and still nothing. I've tried running it through different IDE's and the command line, but it had the same behavior. I've also tried it with all virus and malware protection turned off. I've made sure python and all IDE's are exceptions to the firewall and all the requisite permissions are enabled.\r\n\r\nAdditionally, I checked to see if other packages could download content such as an nltk corpus, and they could. I've also run the same script using Ubuntu and it downloaded fine (and quickly). When I copied the downloaded datasets from my Ubuntu drive to my Windows .cache folder it worked fine by reusing the already-downloaded dataset, but it's cumbersome to do that for every dataset I want to try in my Windows environment.\r\n\r\nCould this be a bug, or is there something I'm doing wrong or not thinking of?\r\n\r\nThanks.","comment_length":23,"text":"load_dataset() won't download in Windows\nI don't know if this is just me or Windows. Maybe other Windows users can chime in if they don't have this problem. I've been trying to get some of the tutorials working on Windows, but when I use the load_dataset() function, it just stalls and the script keeps running indefinitely without downloading anything. I've waited upwards of 18 hours to download the 'multi-news' dataset (which isn't very big), and still nothing. I've tried running it through different IDE's and the command line, but it had the same behavior. I've also tried it with all virus and malware protection turned off. I've made sure python and all IDE's are exceptions to the firewall and all the requisite permissions are enabled.\r\n\r\nAdditionally, I checked to see if other packages could download content such as an nltk corpus, and they could. I've also run the same script using Ubuntu and it downloaded fine (and quickly). When I copied the downloaded datasets from my Ubuntu drive to my Windows .cache folder it worked fine by reusing the already-downloaded dataset, but it's cumbersome to do that for every dataset I want to try in my Windows environment.\r\n\r\nCould this be a bug, or is there something I'm doing wrong or not thinking of?\r\n\r\nThanks.\nThis was fixed in #644 \r\nI'll do a new release soon :)\r\n\r\nIn the meantime you can run it by installing from source","embeddings":[-0.5447326899,0.2937762439,-0.0705709532,0.2974264026,0.1420515478,0.3211807311,0.2740902603,0.2108792812,0.3762368858,-0.0303392913,0.1460882127,-0.0534878336,-0.0536378734,0.0993813351,0.1741901487,-0.0783414617,0.0386146717,0.0273939483,0.0398455709,-0.1172637492,-0.3460811079,0.1343625337,-0.5268455148,-0.2473067939,-0.1183369607,0.1657858789,-0.1818952411,0.1549066007,-0.1827932149,-0.3171298206,0.4274270236,0.2730977535,0.2871667445,0.5432711244,-0.000117681,-0.2694989741,0.4458218813,-0.0840959698,-0.0909612104,-0.1485743672,0.0208318233,-0.3754148185,0.0824159458,-0.0847013071,0.0044254111,0.2943906188,0.1066002473,-0.2247380465,-0.071463801,0.3034383655,0.1490247101,0.0583825186,0.1866412014,-0.0182447415,0.0309291799,0.0822879225,-0.2140993178,0.372680366,0.1696050018,-0.3838160336,0.1925896406,0.0155720823,-0.2315173447,0.1429763734,0.0123975426,-0.0477098152,-0.0509737581,-0.4643266499,0.1777079999,0.2681756616,0.7157570124,-0.0258965492,-0.1523783356,0.0638917908,0.170056805,0.0090480354,0.2117129415,0.446498096,-0.300486654,-0.0090608783,-0.3201352954,-0.2021882087,-0.1552502662,0.3006733656,-0.0191625115,0.172332719,0.0151738683,0.2417906523,0.2533405721,0.1294081807,0.2495853901,-0.0842911527,-0.0199931953,0.0875165612,-0.12029019,0.3168123662,-0.0019659237,0.4192122817,-0.0137599166,0.1223873347,-0.0840518922,0.2061972469,0.3144582212,0.0479619205,0.3432303667,-0.1232668385,0.2831478417,0.0193244293,0.4794267714,0.1123719737,-0.0280559398,0.0329987071,-0.456946075,-0.1710484773,0.0427136756,-0.0770588592,0.3007377386,-0.4118879735,-0.1713655889,0.0112365186,0.1327832043,-0.2327603996,-0.0835277289,0.2520165741,-0.0977337956,0.5410680771,0.1398295611,0.2110765874,-0.2146802843,-0.0960877091,0.133765012,-0.0282489993,-0.3755545914,0.16073668,0.6784365177,-0.0189376548,0.0583459921,-0.1191413552,0.0624016635,0.0006960102,0.0645575151,-0.2367898375,-0.0150662046,0.229765743,0.1399356425,0.4387683868,-0.0286705364,-0.0200705826,-0.227859512,0.2863512039,0.0123344762,-0.050815139,0.1796677262,0.0697398484,-0.1441726089,-0.1259819716,-0.1264906079,0.2690012455,-0.0426498652,-0.1282819211,0.0170346964,-0.2095018625,-0.2395153344,-0.271723181,0.2266691923,0.6464570165,-0.5141645074,0.2695254087,-0.094172813,-0.0030316773,0.013725725,0.0930571184,-0.1872507781,0.2323407531,-0.2925850749,-0.0279402398,0.0329585001,-0.3341713846,-0.329556495,0.4202957451,0.0340788141,0.1563066989,-0.0726221055,0.2671049237,0.0862022564,-0.0714237317,-0.0056537604,0.4295674264,-0.0183062293,-0.0602431111,-0.1239621341,-0.1605127305,0.3207066059,0.3485834002,-0.0397550315,0.1262582093,0.0998108014,0.1640673578,0.2192295045,0.2019403726,0.0795274377,0.3887034357,-0.0824792609,0.1266227216,0.1167503893,0.0072803437,-0.600615561,0.1996982545,0.4054840207,-0.0834777579,0.1673303992,-0.039528802,-0.2067569643,-0.0293626785,-0.0387593359,-0.2925137877,-0.0106431833,0.162091434,0.304017365,0.0198231582,-0.0307580996,0.4781758785,-0.0493634641,0.0320866592,-0.2876859605,0.128419295,0.0026968576,-0.1307896525,0.1691425294,-0.1523955911,0.1291792244,-0.1172845364,-0.1496613175,0.3383615315,-0.0526154935,0.0253006145,0.0424159765,0.0579257645,0.0808429718,-0.4428676367,0.2842034698,0.4864927232,0.2960579097,-0.3706301153,-0.3011002541,0.0170023069,-0.3206239641,0.2708461881,-0.003683669,0.0178484581,0.0608161725,-0.0296878014,-0.1301566809,0.1781577319,0.5836650133,-0.1008612365,0.0228968449,-0.1700228006,-0.1543064564,0.1075336561,0.3203768134,-0.0916451812,-0.0287028234,0.1544530839,-0.2725630701,0.0527827442,-0.0886603445,-0.0396871753,0.6740580201,0.120720759,0.3030719459,0.0141986422,-0.0475734882,-0.3879134357,0.1116904318,-0.0549832769,-0.1497775167,0.165714249,-0.3047574759,-0.1206816137,-0.2102823108,-0.0174453817,0.2316928804,0.1771310419,-0.200468719,0.094275333,-0.044428613,0.0238137599,-0.0655736849,-0.2384589016,-0.1516899019,-0.2049007267,-0.3307043016,0.4893770516,0.2152124643,-0.1464357078,-0.5216550827,0.1931720674,0.1326215714,-0.2481984943,-0.196603179,-0.1243068054,-0.3068206608,-0.0133875497,0.3423084915,0.1893699765,0.150108099,-0.2650111318,0.0399425961,-0.2486810982,0.0475106165,-0.1428280473,-0.0254679006,0.2613433897,-0.0454117768,0.4417714477,-0.0693355203,-0.0001030402,0.1407703459,-0.4553450346,0.1400993913,0.1691418141,0.2193464339,-0.2320087701,-0.079426378,-0.3101460338,-0.2250716686,-0.3428766131,0.2097816914,-0.0245217495,0.0428835936,-0.0143600795,0.1111810133,0.0196625702,0.1156873703,-0.0526175275,-0.1598743945,-0.1181115136,0.6609854102,-0.2714054883,-0.6282078028,0.3548407257,0.2915496826,-0.1052732915,0.1916489005,-0.583519578,0.3744540215,-0.1928445399,0.0926938877,0.1994612068,0.1941565126,-0.0166794863,-0.3557727933,0.1009251177,-0.0099809552,-0.1868429035,-0.2314061821,-0.0458582118,0.2779678106,0.1946287751,0.105609335,-0.2316230685,0.2021383494,0.1312455088,0.1842386574,0.2269838452,-0.0310881399,0.5086892843,0.0304643679,-0.3904331028,0.2541800439,-0.4080343843,-0.0001525798,0.2978044152,0.1291253567,-0.4621703625,-0.4852896631,0.1718789488,-0.0858805999,-0.2213990688,0.2845761478,-0.2263183296,-0.0195100214,0.0193995498,0.258694917,0.2411656678,-0.4452512264,0.1517035961,0.4347318113,-0.139911294,0.0723235607,-0.3505471647,-0.0789590478,-0.4334245026,0.1168992147,-0.237097308,0.4480698407,-0.1931823045,-0.0272666961,0.1092576981,-0.1246502325,0.6732804179,-0.344823271,0.0033152201,0.3158023059,0.1881671846,-0.3902427852,-0.2304425985,-0.3901026547,0.3355697691,-0.0902066752,-0.1591770351,-0.1144886538,0.0050000763,0.0183633361,0.2352613211,-0.0144216781,-0.1628565341,-0.3745222688,-0.4569240808,-0.2227561623,-0.0896032006,0.0780539066,-0.0599671081,-0.428075999,0.1249174029,-0.0234540794,0.1070839092,0.1048137322,-0.1557910144,0.1851313412,0.1956189573,0.2743834555,0.2562501729,0.1056020781,-0.0424185656,0.544921577,-0.0249647051,0.0179828387,0.3539716005,-0.3520484567,0.0853556246,0.3696365952,0.0084874304,-0.345415175,0.0923867151,0.0583309457,-0.3786527812,-0.0176139046,0.3195759058,-0.0532699078,-0.2735890448,-0.5353895426,0.3098929822,-0.0057077445,0.0565166511,0.3623389006,-0.2526048124,-0.0515085943,-0.1430336833,-0.2395103425,0.8468360305,-0.107557863,0.2447051257,-0.0618476234,-0.0157158598,0.0893332809,-0.1936181784,0.0676236823,-0.0049970816,-0.0329640135,-0.1138351038,-0.115787372,0.2238118351,0.4172495902,-0.4091871083,0.29440853,-0.0592645667,0.1347719878,-0.1224899814,0.3068839908,-0.2343120426,-0.2817795575,-0.100464128,0.0891601071,0.1238320619,0.4508275092,-0.2824506164,0.1202135459,0.0690917075,-0.1100990847,-0.4286577404,0.4342184365,-0.0060545923,0.0589021854,-0.4166520834,-0.0702740252,0.0664740726,0.1753120273,0.1721327752,0.0660057366,-0.3955233097,0.2719853818,-0.0355883464,-0.3001207113,0.0164686423,-0.1969319433,0.040742036,0.0036632807,-0.3454319537,0.1249590665,-0.1769378781,-0.2024333179,-0.1881968826,0.0765614957,0.186364755,-0.1148935035,-0.1953827888,0.0253343359,0.0366907008,-0.0113332178,0.058860667,0.2570262253,-0.0104694869,0.4543800056,0.0297168028,-0.1582191736,-0.1085017323,0.5431082845,-0.0927428007,-0.4721795619,0.1921751648,0.714869082,-0.2609756887,-0.2856498957,0.0789523795,0.1166231856,-0.289657563,-0.081263423,0.1003530771,0.3360960186,0.1872027814,0.0143998573,-0.0788416713,-0.3600306511,0.0363114439,-0.569488287,0.0812425092,0.0486535579,0.3408637643,0.0979807749,-0.1303733438,-0.5350725055,0.6288610101,0.066524595,-0.2064329684,0.0589357316,-0.0528288521,-0.000053618,0.3957705796,0.1604871899,0.2460856438,-0.0190905854,-0.0477106869,-0.0280646794,0.1173542589,-0.0809960812,-0.0294634998,0.2213721126,-0.1012445837,0.0198016111,0.1108257696,-0.6536310911,-0.3515220582,0.0445244499,0.1947991699,0.1392392665,-0.1913036555,0.0843778327,-0.0571952872,0.3019420803,-0.1776511222,0.2432241589,-0.2800223827,0.6548907161,0.0084427223,0.11064291,0.2920010984,0.0861852318,-0.2077659518,-0.0038366199,0.0258450322,-0.0390645228,0.2755316496,-0.2831008136,0.2741784155,-0.1997779459,0.2218065113,0.8365336061,-0.2370648682,0.0382379144,0.3122197986,0.0683991387,-0.1233076081,-0.058957316,0.0357364304,-0.1718779206,-0.0692727938,0.1240697801,-0.0218727235,-0.3659677804,-0.2176698744,0.0453307182,0.5127395391,0.1894974858,-0.190373227,0.4263847768,-0.0427493118,0.0131010152,-0.0716622844,0.3250524104,-0.1172903553,0.4445208013,0.1096294224,0.1706625968,0.0574068464,0.408953011,-0.2246468365,-0.2308945954,0.0619929135,0.1573178023,0.0614791997,0.1932944506,-0.1953667998,0.2428781092,-0.4373216927,-0.1304504126,-0.4729946256,0.0679324791,0.2066802233,-0.0204456449,-0.1333401948,-0.1568819433,-0.1622099429,0.2003723234,0.0817897916,-0.2086405605,0.5919040442,0.210847497,0.0559961721,-0.4479515851,0.1479143202,-0.1466023922,-0.0857848227,-0.2470907718,0.1983364373,-0.0722391754,0.1332909167,0.0773827583,0.1716842353,0.2039041966,0.4424135089,-0.1939238012,0.3094861209,-0.0019640026,-0.1057752967,-0.0110080522,0.5946021676,0.2005302906,0.032815326,0.1848658323,0.0196748059,-0.0471918397,0.0039445143,0.0807811394,-0.118106842,-0.1192593649,0.0060735303,-0.0527142398,0.0527915508,-0.1306189597,-0.1308090985,-0.4362170696,-0.3978845477,0.337580502,-0.2368493229,0.0197787602,-0.2708039284,0.0072684954,0.0866931155,0.6320139766,0.338993907,0.5349494815,-0.1189103499,-0.2785025537,-1.0058071613,0.1198851615,-0.199184075,-0.1983908415,0.024627544,0.2282790244,0.0540424846,0.522810936,0.0532014631,0.0760001764,0.2053549588,0.3606359661,-0.2714611888,-0.0808678418,0.1189026982,0.4150774181,-0.2541849613,-0.4474763572,0.3048176765,-0.1501553506,-0.1415580064,-0.1338585019,-0.2137736827,0.1502432674,0.1927309483,0.4347640872,-0.1761736721,0.4640434682,0.0954776257,-0.13095285,-0.4059804976,0.0920999497,0.1357538998,0.1186206415,-0.0427123979,-0.2561986148,-0.4338244498,0.0624667779,-0.050388854,-0.0436750986,0.0386263542,0.1462690979,-0.1939917803,-0.2305743247,-0.0851170719,0.2610270083,0.0376045257,0.2082821727,-0.2089046836,-0.1773203164,-0.2015266567,-0.0190663487,0.2685916424,-0.3023798466,-0.3177041113,-0.1464695781,0.0872825757,-0.4101589322,0.1301526874,-0.4316550195,0.1105723158,0.0562147088,0.0241490733,-0.290605098,0.2589587867,-0.3456102014,-0.0299872588,-0.0750534311,0.2575957775,-0.0224151332,-0.2148822546,-0.0522237755,-0.1408999264]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/674","title":"load_dataset() won't download in Windows","comments":"Closing since version 1.1.0 got released with Windows support :) \r\nLet me know if it works for you now","body":"I don't know if this is just me or Windows. Maybe other Windows users can chime in if they don't have this problem. I've been trying to get some of the tutorials working on Windows, but when I use the load_dataset() function, it just stalls and the script keeps running indefinitely without downloading anything. I've waited upwards of 18 hours to download the 'multi-news' dataset (which isn't very big), and still nothing. I've tried running it through different IDE's and the command line, but it had the same behavior. I've also tried it with all virus and malware protection turned off. I've made sure python and all IDE's are exceptions to the firewall and all the requisite permissions are enabled.\r\n\r\nAdditionally, I checked to see if other packages could download content such as an nltk corpus, and they could. I've also run the same script using Ubuntu and it downloaded fine (and quickly). When I copied the downloaded datasets from my Ubuntu drive to my Windows .cache folder it worked fine by reusing the already-downloaded dataset, but it's cumbersome to do that for every dataset I want to try in my Windows environment.\r\n\r\nCould this be a bug, or is there something I'm doing wrong or not thinking of?\r\n\r\nThanks.","comment_length":19,"text":"load_dataset() won't download in Windows\nI don't know if this is just me or Windows. Maybe other Windows users can chime in if they don't have this problem. I've been trying to get some of the tutorials working on Windows, but when I use the load_dataset() function, it just stalls and the script keeps running indefinitely without downloading anything. I've waited upwards of 18 hours to download the 'multi-news' dataset (which isn't very big), and still nothing. I've tried running it through different IDE's and the command line, but it had the same behavior. I've also tried it with all virus and malware protection turned off. I've made sure python and all IDE's are exceptions to the firewall and all the requisite permissions are enabled.\r\n\r\nAdditionally, I checked to see if other packages could download content such as an nltk corpus, and they could. I've also run the same script using Ubuntu and it downloaded fine (and quickly). When I copied the downloaded datasets from my Ubuntu drive to my Windows .cache folder it worked fine by reusing the already-downloaded dataset, but it's cumbersome to do that for every dataset I want to try in my Windows environment.\r\n\r\nCould this be a bug, or is there something I'm doing wrong or not thinking of?\r\n\r\nThanks.\nClosing since version 1.1.0 got released with Windows support :) \r\nLet me know if it works for you now","embeddings":[-0.5312045217,0.2285069823,-0.0575985946,0.2666499317,0.1350667328,0.3239186406,0.2561958432,0.219893232,0.3572812676,0.0436172225,0.1882215887,-0.0625395849,-0.0545900762,0.1764082909,0.1357093602,-0.0816593394,0.0941099972,0.0226964857,0.0226596929,-0.1013214365,-0.3583259881,0.1709131896,-0.4910069406,-0.1853563637,-0.120413743,0.2017234713,-0.2180992812,0.1222668961,-0.1995783895,-0.3122927547,0.4654698968,0.3261946738,0.3328743577,0.4956374168,-0.0001170868,-0.231574297,0.483954668,-0.0690701902,-0.1345821917,-0.2716261744,-0.0077642691,-0.3620634675,0.0949622169,-0.0017414198,0.010947383,0.3178222775,0.1540232897,-0.2169940323,-0.1187456921,0.3016267717,0.1500115842,0.0650988743,0.2076624781,-0.0231751185,0.1106366217,0.099093087,-0.237621218,0.3654014468,0.1448057294,-0.3507416844,0.1638025194,0.0025700624,-0.1906076223,0.0756300464,0.0991751701,-0.0257794689,0.0206769928,-0.4977041483,0.1921181828,0.2145073712,0.7776694894,-0.0320849679,-0.1946044117,0.0488454401,0.1430145055,-0.0328599066,0.2404405475,0.4651777446,-0.312772125,0.0017895943,-0.3549053073,-0.1745453179,-0.2147067785,0.3462227881,-0.0036783491,0.1802301556,0.0305233449,0.2561663985,0.2806198299,0.1594635397,0.2893838584,-0.1119091064,-0.0438714437,0.1055945083,-0.1552415937,0.2997744977,0.0266171321,0.4183351099,0.017508287,0.1570257992,-0.0587561205,0.1947076768,0.3071296215,0.0666199327,0.4180442095,-0.0607035682,0.3588195443,-0.0342604555,0.4944933653,0.096069999,0.0118606184,0.0071075396,-0.4333915412,-0.1531762481,0.0900343508,-0.0854356065,0.2950430512,-0.4243557751,-0.1556747854,0.0356016494,0.1440680921,-0.2066551149,-0.120803602,0.2567278147,-0.0902356654,0.5926252604,0.1480210871,0.1692719907,-0.2379569411,-0.0793555975,0.1535152048,-0.0292146113,-0.3896257877,0.1484386176,0.6411631107,-0.0656778291,0.0258341134,-0.1347650439,0.0625348687,-0.0405454151,0.0859267861,-0.2406888306,-0.0449584574,0.2409293652,0.1386255175,0.4186852872,-0.0617769621,0.0067275236,-0.2271531224,0.2992707789,-0.0781947151,-0.0815833956,0.1264632493,0.0773238465,-0.1531830132,-0.1732040197,-0.1545853019,0.2747504711,-0.1034760997,-0.1443240941,-0.0104165263,-0.1774776131,-0.3123803735,-0.2741759121,0.1735449582,0.6089103222,-0.5470990539,0.2760280669,-0.0341996253,-0.0876644105,-0.0104920547,0.0894189328,-0.141623497,0.1639377326,-0.2924817801,-0.0215298124,0.025942836,-0.3787801862,-0.3475463688,0.4641360044,0.0749896616,0.0911938697,-0.0299300402,0.2377229929,0.1555584818,-0.0865240097,-0.0461787768,0.4049776196,-0.0507781655,-0.0866378769,-0.148676917,-0.1572996527,0.2671688795,0.3516069949,-0.0558429807,0.1907018572,0.0927738845,0.078379117,0.2400555313,0.2020245939,0.14721632,0.40376845,-0.1247166693,0.0438928157,0.1558066905,-0.0253990274,-0.6035884023,0.206732735,0.3425562978,-0.031738732,0.1866748035,-0.0706361905,-0.2475808263,-0.0217133034,-0.0718118697,-0.2302504629,-0.0182543155,0.1467156708,0.2857332826,-0.0163864009,-0.0573725775,0.4816448092,0.0001618297,0.0678999722,-0.3044669628,0.1455879807,-0.0020915668,-0.1764754057,0.1640513688,-0.1513475329,0.120415017,-0.1093366817,-0.1514501274,0.3896580935,-0.0368123241,0.0581553727,0.049190674,0.0184193924,0.0687952787,-0.3777018487,0.2146212012,0.4776518941,0.2862357497,-0.3557849526,-0.3085817993,0.1224041283,-0.292832762,0.26884076,0.0279136244,0.0429542735,0.0413202234,-0.0010225258,-0.1258667111,0.2151589394,0.5775262117,-0.0929219127,-0.0020058907,-0.1643341333,-0.1167262644,0.074873291,0.2431920171,-0.1832413077,-0.0991430879,0.1818745881,-0.2656547129,0.046527721,-0.0512061678,-0.0279067066,0.6937488914,0.1468280107,0.2767846286,0.0139100784,-0.0489187054,-0.3464194536,0.1142464578,-0.0273889545,-0.1994453818,0.1210508719,-0.2485818416,-0.1819965839,-0.2715475261,-0.0218851939,0.1723171473,0.2161433101,-0.2237141579,0.1064540446,-0.0458935462,0.0451414101,-0.0418594889,-0.2032460868,-0.2282697707,-0.2365659475,-0.3288748264,0.5409021974,0.2697987854,-0.1737596095,-0.4869428575,0.2375602275,0.0965301841,-0.1562859416,-0.2288695276,-0.1317171007,-0.3606881201,-0.0115164006,0.3313680887,0.1251847148,0.1712014824,-0.2329801917,0.0551182292,-0.1988323033,0.1272965819,-0.1261771172,-0.0057974411,0.2244524509,-0.0427630991,0.4186484814,-0.1009568349,0.0202344488,0.11982961,-0.4642215967,0.0796953812,0.2274574935,0.1727819443,-0.2189692557,-0.1373376548,-0.2433601916,-0.2774139047,-0.2721554935,0.2006094456,0.0106294835,0.0735635236,0.0224569906,0.114382498,0.0053695044,0.0750945508,-0.0262677576,-0.1641979814,-0.2334053814,0.6322652102,-0.2217449248,-0.6599826813,0.4075287282,0.2974890172,-0.1475100964,0.2126586735,-0.5720703602,0.3467951119,-0.1019633785,0.0610632151,0.2333485782,0.1665354222,-0.0139730619,-0.3323441446,0.1042610928,-0.0081534088,-0.2222926915,-0.2352988571,-0.0378954262,0.2994037271,0.1876711547,0.1272700578,-0.1850459278,0.2182715684,0.1114651412,0.2113908231,0.2606621981,-0.0783335567,0.4497875571,-0.0028928609,-0.3595738709,0.2079339027,-0.4530533254,-0.0218362082,0.2410742193,0.1938201785,-0.4492456317,-0.5416457057,0.1022526696,-0.0325710885,-0.2288877666,0.2101327181,-0.228131935,0.0144009683,-0.0066878526,0.3081879318,0.2055810541,-0.4835149944,0.1495600641,0.511305213,-0.1142199859,0.0346211083,-0.2572854459,-0.0452077761,-0.5120877028,0.1543219239,-0.2412569225,0.4682085812,-0.1624612212,0.0226237848,0.1476805061,-0.1611587703,0.7127903104,-0.3348449469,0.0090464046,0.2911954522,0.1778964847,-0.3477024138,-0.2521779239,-0.3208561838,0.2707757056,-0.1333317757,-0.1066116095,-0.0513587482,-0.0276868381,0.0194046721,0.2085131705,-0.0282630604,-0.2030299455,-0.387596786,-0.4489168227,-0.2161347866,-0.1169514135,0.0080400333,-0.0554372035,-0.338124305,0.0704431906,-0.0389574505,0.0448572785,0.1545578092,-0.1704314947,0.1566195488,0.117770642,0.2518834174,0.2605077624,0.1264923811,-0.0444797911,0.5352252126,-0.0068990136,-0.0584064163,0.3456351161,-0.3245076537,0.065953739,0.3794381618,-0.0373905897,-0.3221268058,0.0372584127,0.0385672301,-0.3754252791,-0.0805498958,0.3639701903,-0.1234171316,-0.2981170714,-0.5118228197,0.3101017773,-0.0069003208,0.0436954871,0.3944942951,-0.2841474712,-0.0619641468,-0.0923063681,-0.289914608,0.9123569727,-0.1042078361,0.2789795697,-0.0560506508,-0.0926121026,0.060232386,-0.0597327016,0.0701088831,0.011450584,-0.010658822,-0.1747796685,-0.0912473053,0.2246440351,0.4207729399,-0.4020048678,0.3291707039,-0.0444841422,0.1301694661,-0.1456130445,0.3398071826,-0.2560743093,-0.2890045643,-0.1004848778,0.1076414436,0.1147084087,0.4825769365,-0.2655574381,0.0962700024,0.1247294992,-0.0826989561,-0.4117218256,0.3731667399,0.0219276864,-0.0088622514,-0.3657298088,-0.0865783691,0.0371257775,0.0833033174,0.1346684247,0.0007063883,-0.4090565741,0.2347535044,-0.0154679213,-0.2087961882,0.0245861094,-0.2035185248,0.0525614917,-0.0225745458,-0.3193159699,0.135324955,-0.1560494602,-0.1914230585,-0.1911019534,0.0524752215,0.1565386653,-0.0776957721,-0.1630139351,-0.036458984,-0.0055671064,-0.0188114382,0.0700897053,0.2706729472,0.0310924184,0.3902575672,0.0121833282,-0.1944542825,-0.1106987521,0.605768621,-0.0806575492,-0.5021287799,0.1530230641,0.6260691285,-0.2508814931,-0.243290782,0.1853442043,0.1611519307,-0.2851474881,-0.0198673699,0.1127226278,0.3741264045,0.1398125142,0.0936654285,-0.0518892482,-0.3265524507,0.0399390124,-0.5623198748,0.0030020759,0.0282329246,0.3559066951,0.1410147101,-0.1761784852,-0.5053316951,0.6522714496,0.0118451025,-0.2051388621,0.0404381528,0.043873854,-0.0516297221,0.3406750858,0.1937030107,0.216009289,-0.075050205,-0.0469053797,-0.0835297927,0.0763612241,-0.0669475645,-0.0165356249,0.2285728306,-0.0713757575,0.0152987186,0.1097325459,-0.665522635,-0.3301795125,0.0519565716,0.2147805542,0.1841876358,-0.2401821911,0.0629862025,-0.0378192812,0.3294989765,-0.1409052163,0.2073756009,-0.2496888489,0.678794086,0.0323360823,0.1153361425,0.2936725616,0.0756280497,-0.1755987406,-0.0027927915,-0.0256994013,-0.0552208796,0.2679033875,-0.2902222276,0.2437649518,-0.1873649806,0.2755455673,0.8399642706,-0.2070834637,0.0871811658,0.2832946777,0.0746318325,-0.0726210698,-0.0405225828,0.0131938355,-0.1571924239,-0.0853196681,0.0677443892,-0.0451035053,-0.3668614626,-0.21510382,0.0573219433,0.4865668118,0.1241235211,-0.1937589049,0.4410764873,0.0168469213,0.0754283369,-0.0500852577,0.2841109633,-0.1606819332,0.4408056736,0.1419351846,0.1854076087,0.098233223,0.4767306447,-0.2801034749,-0.2568105161,0.0381127335,0.1330263615,0.0329111889,0.1737668365,-0.1635855138,0.2226181924,-0.4554187953,-0.1005294845,-0.4397477508,0.0793714449,0.2263457477,-0.0563460812,-0.2158171088,-0.1652947664,-0.1753156781,0.1932718903,0.0722954497,-0.2186772078,0.5698959231,0.2334739566,0.0850056112,-0.3904230297,0.1136045307,-0.141228646,-0.0061675673,-0.2205160856,0.1858049184,-0.1524084657,0.2062789947,0.1133086532,0.1342744678,0.2519593835,0.528739512,-0.2413778156,0.2972762883,-0.0324133374,-0.1212810948,0.0077468278,0.5865789056,0.2404150814,0.1050585881,0.2001648694,0.0082779834,-0.0164264739,0.0515775084,0.0666478798,-0.1798188835,-0.1858170927,0.0692103878,-0.0355626084,0.036070969,-0.1252231151,-0.1398774832,-0.3972561657,-0.4243099391,0.3482150733,-0.2158009261,0.0149442796,-0.2794138491,0.0117860967,0.1118227243,0.5671476126,0.3322288692,0.5297795534,-0.1620533168,-0.230309993,-0.9565055966,0.1080500185,-0.1685977578,-0.1755849719,0.009884051,0.1923179179,0.0025185714,0.484390825,0.0235465225,0.0809898898,0.2134415656,0.3761447966,-0.2119746357,-0.0689434931,0.1132201552,0.3849098086,-0.2799790204,-0.4363290668,0.320438087,-0.1441283524,-0.1724737138,-0.0517528914,-0.1804921776,0.1560871601,0.1969275028,0.4207725823,-0.2091135532,0.4729420543,0.1104512662,-0.0546742119,-0.444727391,0.0897131711,0.0714227632,0.0718474016,-0.0723685771,-0.293153882,-0.3677377999,0.0291811619,-0.0544561334,-0.0275196731,-0.00033495,0.1716920584,-0.1966290325,-0.1751498133,-0.0856798813,0.2230069637,0.0221847557,0.1691586375,-0.1965220273,-0.153722927,-0.2684342265,-0.0431594998,0.2580578923,-0.2556962073,-0.3342415988,-0.1974039674,0.0821198896,-0.4035399556,0.1280595213,-0.4431243241,0.0934146121,0.0588153154,0.0057147183,-0.2654506266,0.2720973194,-0.2873011231,0.0370805338,-0.0591975152,0.2502172887,0.0090837153,-0.2215782702,-0.0548279211,-0.1950526088]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/673","title":"blog_authorship_corpus crashed","comments":"Thanks for reporting !\r\nWe'll free some memory","body":"This is just to report that When I pick blog_authorship_corpus in \r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=blog_authorship_corpus\r\nI get this:\r\n![image](https:\/\/user-images.githubusercontent.com\/7553188\/94349542-4364f300-0013-11eb-897d-b25660a449f0.png)\r\n\r\n","comment_length":8,"text":"blog_authorship_corpus crashed\nThis is just to report that When I pick blog_authorship_corpus in \r\nhttps:\/\/huggingface.co\/nlp\/viewer\/?dataset=blog_authorship_corpus\r\nI get this:\r\n![image](https:\/\/user-images.githubusercontent.com\/7553188\/94349542-4364f300-0013-11eb-897d-b25660a449f0.png)\r\n\r\n\nThanks for reporting !\r\nWe'll free some memory","embeddings":[-0.2163716406,0.2747993469,-0.1138018072,0.1898003966,0.146038726,0.0928721502,0.1720205843,0.1664657742,-0.1611467898,0.1571003199,-0.1133945137,-0.021676885,-0.1140432358,-0.1900718659,0.2385389656,-0.1146043986,0.0731984079,0.072356306,0.1017724797,0.0425746627,-0.0311759412,0.414680481,-0.1758517176,0.3988641798,0.0710927621,-0.0187141672,0.0456173569,0.2987168133,-0.3048721254,-0.2131317854,0.1047912017,-0.1060279682,-0.2227795422,0.1742460728,-0.0000998856,0.1172761023,0.4273268878,0.057323385,0.1372665316,-0.1416655481,0.1624460667,0.4034135938,-0.3906848431,-0.0303638261,0.095895119,-0.1135362163,0.3303670585,0.2484228313,0.0547101572,0.269975543,0.3805444241,0.0716325492,0.2540104687,-0.1128359213,0.1174256951,0.1137823313,-0.0484427884,0.1886665821,0.3746606708,-0.4860912561,0.0057068537,0.2744910419,-0.4425405264,-0.3512631953,0.0297152996,-0.0352445319,0.4075711668,-0.2089362592,0.3604542315,-0.0021250218,0.3157071769,0.1729424,0.3534378409,0.1994096935,0.2074600458,0.1549449563,0.2427862436,0.44027403,-0.3790329993,0.0430675,-0.1910772175,-0.0003574741,-0.1774763018,0.131278798,-0.1635001302,0.1062504873,-0.1483945251,0.17932491,0.1258559525,-0.0744511336,-0.1421850324,0.2133093774,-0.3459277749,0.0568486638,0.238921538,-0.031445954,0.1119413972,0.131578818,0.1883015037,0.1484288722,0.1014417037,0.030962931,0.3124092519,-0.0445341095,0.0850539431,-0.1316303015,-0.1139149368,0.2925547659,0.2066818774,0.0995167643,-0.0289487541,0.0814646259,0.056315843,-0.3565326333,-0.0049208519,0.0185365602,0.2699261904,-0.3455457389,-0.1154548153,0.0020947042,-0.1149151325,0.1820112914,-0.1271858364,0.5459301472,0.0536981411,-0.3293138742,0.1995938122,-0.0989635661,-0.1754956245,0.0096144108,-0.2455152124,0.0689297542,-0.0709898472,-0.0577284358,-0.0129124448,-0.4866568744,0.2495700568,-0.2782512009,0.2225845456,-0.3736403286,-0.0321246982,0.0354044139,0.0390856974,0.2968863249,-0.040260531,0.0073204986,0.025980806,-0.2214180082,-0.0156187667,-0.1187322214,-0.052832868,-0.4582737088,-0.1066641062,0.317006886,0.1600879729,-0.0140685979,-0.2878476083,0.1761578023,0.1307210773,0.0281492323,0.1392430663,-0.0033061663,-0.0725626498,-0.0208655912,-0.2250477225,0.2027204931,0.5055447221,-0.1421189159,0.284996748,0.0273071099,0.1254907995,-0.0109554958,-0.0573767312,0.1208899021,-0.2566081583,0.1346314251,0.0496563651,-0.3645309806,-0.117607668,-0.1531822532,-0.0455647632,-0.2274973691,-0.0526654832,-0.1433995515,0.1265978664,-0.4041801393,0.039721109,0.265882045,-0.1656779051,0.2386905849,-0.1009172276,-0.1558930576,-0.4006796479,0.1794584692,0.2606923282,-0.1484595686,0.2191464156,0.1936527193,0.1779512167,-0.1048373878,-0.1563020796,0.3646437228,-0.0283769555,-0.059642639,0.1045782641,0.0209348258,0.1133623719,0.0232321601,-0.0034976506,0.2783702314,0.0336939879,-0.095100224,-0.2760006785,0.0084948037,-0.0683562309,-0.1174954474,0.2607244551,0.0489604808,0.121272251,0.1861996204,-0.0599385016,0.0705476701,-0.3065277338,0.1190777123,-0.4907872081,0.0102182124,-0.2365529984,-0.2175215483,0.1544156373,0.0396885239,0.0799481347,-0.0527355075,0.0441531651,0.214770779,-0.7718183994,0.6506346464,-0.2539434135,-0.2737226486,0.1144889593,-0.6372538805,0.0004768134,-0.1097349226,0.0530254878,-0.2531743348,0.3928960562,-0.2119711637,-0.2148910612,-0.0337876044,-0.2117688805,0.2093642354,-0.0484340899,-0.2040019631,-0.037006069,-0.0455847085,0.4374957979,-0.1888944805,-0.4467028677,-0.2337962091,-0.3670018017,0.1857180595,0.2263275385,0.0725216642,0.1982420236,0.0556353629,-0.2048998475,0.1349256039,0.0351873077,-0.2587709129,0.2274333835,0.1091094241,0.1460367292,0.0680399761,-0.1317982972,-0.2587855756,0.5229434967,-0.1520927399,-0.0110620093,0.0306083281,0.2120207995,-0.0224969182,-0.5222176909,-0.0682234988,-0.1896177381,0.1195337698,-0.2316753566,-0.0390838422,-0.0302366093,-0.3653581738,0.2025453895,-0.0928438753,-0.0632834211,-0.076214008,0.2007237077,0.1787656695,-0.0375158004,0.0801548138,-0.0647923574,0.0700777099,-0.1833695918,0.5596791506,-0.2307097167,-0.0811082199,-0.2272028178,0.2060884982,0.2813688517,0.0382261388,0.2702437937,-0.103636764,-0.159312427,-0.2530929446,-0.3532364368,0.1014156193,-0.2390562445,-0.0733566061,0.1105378494,-0.0089041498,0.1484123915,-0.0193033423,0.0443686768,-0.1109773889,-0.3371574283,0.1738187671,0.0675216839,-0.0498753637,-0.183854863,0.1482604593,-0.1039499789,-0.2680473626,-0.0163095053,0.3017629087,0.1923937201,0.5250303745,0.0501661077,-0.0450664163,-0.1013776213,0.1904317886,-0.1738932282,-0.2707777023,-0.0193061642,-0.1489182413,-0.2960508764,-0.056727048,0.2712175846,0.0962744802,-0.2069989741,-0.6416386962,-0.1404098868,0.1288578063,-0.3816541731,-0.2078122497,-0.2474769354,0.1066152528,0.0309765544,-0.1957159191,0.2484492213,0.0768619776,-0.3203097284,-0.0228676535,0.1434488297,-0.1919718981,0.1064746901,0.0898859128,0.212917462,0.289996475,0.1720285118,0.1229227632,0.2468916625,0.1122581288,0.1162184998,-0.3091490865,0.5376442075,0.2538833022,0.0101129245,0.1342960298,0.1659036577,-0.1720529199,-0.154303804,-0.0007666118,-0.0709234253,-0.1224840134,0.093194373,0.0511602908,0.2124935985,-0.0443679318,-0.0659728423,-0.0818255767,-0.4461684227,0.2434774786,0.2005154192,-0.1107377335,-0.1275615394,-0.1696360707,0.3538482487,-0.7147620916,0.2759867609,-0.1562836766,-0.3036090434,-0.2649732232,0.2164523304,0.1240662411,-0.0703997463,0.3655041456,-0.2957212329,0.1908190846,0.2249024808,0.2819967568,-0.2597190142,0.0543479137,0.0367526971,0.4002539515,0.1397712976,0.2654617727,0.0262576044,0.1603334397,0.7020551562,-0.3217024505,-0.0891580433,-0.2188675404,-0.2493347079,-0.6438820958,0.1396394819,-0.0964808464,0.1745369136,0.009765923,0.0776319802,-0.2375160009,-0.1150625944,0.1204407811,0.0055798041,0.0739158243,-0.2053568363,-0.124871999,0.0376516506,-0.0473947152,-0.2069470435,0.0647765547,0.4308888018,0.121658206,-0.1135965362,-0.0474245101,0.0069956649,0.0345569886,0.389826417,0.0618968531,0.0127427196,0.0538301803,0.1377517581,-0.2983374894,-0.0832145214,0.1174977794,0.2116073817,-0.2302121669,-0.0441946685,0.3613269925,0.0553919487,-0.0996580496,0.1367396414,0.2086623162,-0.3961894214,0.1978868246,0.3166866004,0.9348781109,-0.1730396003,-0.134530738,-0.1766636819,-0.0376240052,0.382614255,-0.2042710632,0.2307477742,-0.1567114592,-0.0332039669,-0.0057803141,0.0557404421,0.1318207383,0.0326012969,-0.3525812924,-0.0119297542,0.3589737415,-0.0846713781,-0.1945872009,-0.1302062124,-0.3903216422,-0.1139891669,0.041737102,0.334177494,-0.2529545128,0.3164101839,-0.1191905215,0.1432706863,0.2935294807,0.0012377886,-0.5941622257,0.0666013137,-0.0300665293,-0.1395323277,0.4642055929,-0.2971438169,0.0706324205,0.1469232738,0.3992584348,-0.1498314738,0.1222780123,0.3739778399,-0.0354070812,-0.2205979973,-0.1211307198,0.1438591331,0.2649789453,-0.2014033347,-0.1341184527,0.403660655,0.1297254413,-0.2723623216,-0.3101667464,0.2061789632,-0.0897306502,0.2639447749,-0.0803362206,-0.012724393,0.0640442073,-0.0659422055,0.2213934511,0.0954504162,-0.0676598027,0.008147791,0.0892056152,-0.1765791774,0.2336664945,0.1825710684,0.0509348884,-0.1100532264,0.3958141208,0.3168579638,0.0363428742,-0.2088816464,-0.3589531779,-0.2563587725,-0.570510745,0.1261011511,-0.1381688565,-0.1262281388,0.2883445621,0.2633412182,0.0450145118,-0.0034419456,0.0417127423,-0.1899895072,-0.1214694455,-0.0370001122,-0.0334635451,-0.0341444574,-0.0617727377,-0.2161563337,-0.0440424196,0.0741852373,-0.4033314288,0.1847274154,-0.6634292006,0.0313054547,0.2421299219,0.0150886718,-0.1473194659,-0.1146333367,0.2116666734,0.4544166327,-0.111913532,-0.3365604281,-0.0327993445,0.0712163746,0.0604901873,-0.1068481207,-0.0819725394,-0.0614707433,0.1022454724,0.1483332366,0.2581620812,0.1286513805,0.1758640707,-0.2511729002,-0.4495416582,0.0717393607,0.3152889907,0.1469216496,-0.2241434157,0.2973846793,0.0401065089,-0.0758299902,0.0396534167,-0.1650167704,0.0204518791,0.2515373528,-0.3084841073,-0.084433727,0.5704000592,-0.0509946793,0.0379385762,0.0655620247,-0.0046626413,-0.0906579345,-0.3549615145,-0.0836425051,0.0924331993,0.3575937152,-0.0471189991,0.0192111656,-0.3026478291,0.312543124,-0.1287861168,0.0899212882,-0.1505315155,-0.1860652715,-0.0102050006,0.3068531156,0.4792306125,0.0303375889,-0.2070835084,-0.3460631371,-0.2849990427,-0.1498333216,0.2761592567,0.055676274,0.2739592493,0.0166111812,0.2179562598,0.2686451972,0.0278055221,0.3807986081,0.2029194981,-0.1844659448,0.1160868332,0.2407724261,-0.0053118696,0.4453006685,0.2224207073,0.192563504,-0.1722176224,-0.1838991791,-0.1505749822,0.1503908038,-0.0702551603,-0.0153436186,-0.4126037359,-0.0889842436,0.0874098092,0.2676990926,-0.1783732623,-0.0823500678,-0.0860712752,-0.0036437903,-0.3709817529,-0.4301372468,-0.0592041835,-0.0456796177,0.2653727233,0.0033413984,0.1840413511,0.2010908723,0.0111454502,0.1212541312,0.7572025657,0.3237584233,-0.1371999532,0.0595690086,0.1622218192,-0.0160473194,-0.0728575066,-0.1081149504,0.2844188809,0.4511514008,0.2431195229,-0.2382523119,0.244560197,-0.2395471185,-0.059790507,0.2348552048,0.5183149576,0.1451927572,-0.07414186,-0.4078825414,-0.267444253,-0.2777153254,0.1133237407,-0.3661178648,0.06467098,0.3448517025,-0.1750409305,0.1092706621,-0.250388056,0.1351570338,0.3811691403,0.2979952097,0.5235909224,-0.2599137723,0.0699838102,-0.3498252332,-0.0843215957,0.4450393021,-0.1876507401,0.0138829174,-0.01301804,-0.0102747604,0.2374962866,0.0252101365,0.1346125752,0.2876126766,0.092925176,0.3923954666,-0.3585024774,0.0830140114,0.2827360928,-0.1307966411,-0.0595165826,-0.0892921835,0.2283245623,-0.1874260902,0.2315941006,-0.1037282646,-0.0949992314,-0.2184896022,0.160442993,0.1742966622,-0.0775799304,-0.3268547058,-0.0526948608,-0.158288613,-0.1259012371,-0.3241021037,0.0225546323,0.1428435147,0.0003803703,0.2320614755,-0.1088771448,0.0918641835,-0.0014859459,0.0593311563,-0.0400166847,0.0610369965,-0.361582458,0.0518786907,-0.0145066585,-0.1085105985,0.1482366771,0.3998509943,0.143372938,-0.3150579333,-0.2990431488,-0.3841008246,0.2884457111,-0.3053405881,-0.3412559032,0.266107589,0.1187349111,0.0307515245,0.047075443,-0.745223999,0.0758084953,-0.0047413483,-0.2310057133,-0.0042881612,0.041546464,-0.0451794975,-0.0793226063,-0.0031296401,-0.2316308767,0.1657425016,-0.0500728711,0.173033163,-0.1965619773]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"We should try to regenerate the data using the official script.\r\nBut iirc that's what we used in the first place, so not sure why it didn't match in the first place.\r\n\r\nI'll let you know when the dataset is updated","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":41,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nWe should try to regenerate the data using the official script.\r\nBut iirc that's what we used in the first place, so not sure why it didn't match in the first place.\r\n\r\nI'll let you know when the dataset is updated","embeddings":[-0.0942947716,-0.3681170642,-0.1456752867,0.4842418432,0.3399179876,-0.0058333594,0.1115036383,-0.0186547488,0.2138756216,0.2600372732,-0.2300761342,0.2279759496,0.1195970625,0.4290395677,0.1250692308,0.1158343852,-0.0148444707,-0.0975626111,-0.3769937158,-0.2929213047,0.002830687,0.0524279438,-0.0749752671,-0.1406486034,-0.6725170612,0.0625345707,-0.1325014979,0.2614853978,-0.0806931481,-0.2141222507,0.2379073203,-0.0045654103,0.0516931489,0.5184449553,-0.0001057436,0.0076493816,0.1227379441,-0.0369254015,-0.1100252047,0.2392051667,-0.1737133414,-0.016915502,-0.0144083155,-0.3685780764,-0.0910531506,0.0295114294,-0.1441585869,-0.3795345724,0.2213725299,0.0326645672,0.2159141451,0.1681645364,0.1285500824,-0.0342706703,-0.0895384103,-0.0963415354,0.1405212134,0.2543079555,0.3807451129,0.2769884169,0.0853179842,0.1232881248,0.082179673,-0.1980160624,0.263024509,0.0695981309,0.3349301815,-0.1842196137,-0.3047405779,0.2811724842,0.8441703916,-0.0606472977,-0.2785620987,-0.003650225,0.0082516884,-0.1532494724,0.1460047811,0.0381583273,0.3942213655,-0.0236479882,-0.2479045838,0.4130246937,-0.2279997021,0.03008141,-0.1365813911,-0.0956770256,-0.038230814,0.0148485973,-0.2771068513,-0.0362515785,0.0802994892,-0.065788582,0.1476403475,0.1785173863,-0.288202405,-0.110301733,-0.0465075001,-0.0006226418,0.2270852923,0.1416063607,0.2942690551,0.0741879493,-0.3836719692,0.021458935,0.1987365484,0.2798708081,0.4257889688,-0.5590232015,0.0055819405,-0.0897394493,-0.1473904252,0.0754272565,-0.3453876972,-0.0143103991,0.0497555025,0.0657482818,-0.1653847247,-0.3561195135,-0.2374671698,0.278487891,0.0389001407,-0.1122002751,0.1571841091,-0.1443994194,0.0373629965,0.4274276495,0.0311978739,0.012314504,-0.5540838242,-0.302151978,-0.3512794375,0.0989583805,-0.0861772224,-0.015905166,0.0856985524,-0.1169761196,0.3941986561,-0.0891078785,0.2661682367,-0.0721295625,0.0343523063,0.1208451912,-0.3228808343,0.2343474478,0.0164745729,-0.011415123,-0.0378508307,0.0296648461,0.0094077559,0.0093311518,-0.2003113925,-0.2132353187,-0.0303956829,0.3145390451,-0.3848197162,-0.273540616,-0.178314805,-0.0391144827,0.1934795827,0.173362419,-0.2078889012,-0.17496714,0.1106308252,-0.1387371123,0.1916002929,-0.0392939076,-0.6223720908,0.0300172586,-0.0219870601,-0.278963834,0.0815735385,0.0211856607,-0.0069519021,0.279992342,0.0989442766,0.1685311943,-0.540019989,-0.4732564092,-0.1522229463,-0.4094890356,-0.1714006215,-0.0841521844,0.11836721,0.1125238314,0.0399888903,-0.0885651335,0.1391783953,0.0093496162,0.1051331684,-0.1891813427,-0.4143067002,-0.0582363009,-0.0200296994,0.1597853452,-0.0693599358,-0.2733851075,-0.0541004576,0.3515397906,0.2060988098,-0.0352558829,-0.095438078,0.0557038374,0.295684427,-0.1416187286,0.1109532937,-0.2052732706,-0.0060442802,0.3039955795,0.0715231597,0.1947310865,0.7513715029,-0.0283604506,-0.4254818857,-0.2496138215,0.0078303767,-0.16561234,0.2040338069,-0.1758873612,-0.0211424734,-0.2244169861,-0.1832626313,0.15129821,-0.5768190622,0.0455921739,-0.5188596249,0.3624452055,0.0184895527,-0.0043067127,0.0763605237,0.3057864904,-0.0273873582,-0.2645697296,-0.024875164,0.5631830096,0.1721718311,0.0439504161,0.1727522314,0.1609469503,0.3768123686,-0.1657699347,0.1663338691,-0.2367454916,-0.0379575007,-0.1960035712,-0.336510241,0.2413695455,-0.1057439148,-0.0122434944,-0.0910829753,0.0083981901,0.0046784626,-0.169797644,-0.0002173596,-0.0025254723,0.1883715838,0.1760815382,-0.2403234988,0.0694510266,0.0188233722,0.3829338849,-0.3535941839,-0.1978472024,-0.1869016737,-0.3545961976,-0.2246403247,0.1534084231,0.0963025466,-0.0267744623,0.4845177233,0.2857273519,0.0391670689,-0.019909434,-0.0704265311,-0.1554504186,0.0199633427,0.0216203444,-0.0475656278,-0.028564848,-0.043054577,0.0126293814,-0.1905241311,0.1189523265,0.1140751392,0.1595893502,-0.4636821449,-0.3495992124,-0.254878968,-0.0802589953,0.1686788946,0.1879429966,0.1514121443,-0.2807497978,0.14353396,0.0700476319,-0.026946513,0.0233165417,-0.3429911435,0.2449801415,0.0334233902,0.1110631153,0.0909133628,-0.1800023615,-0.1640715301,0.2022572458,-0.0004145729,0.3297639489,0.3800825179,-0.1204203293,-0.0575456694,0.1134605482,-0.4225591421,0.0218226258,-0.1813897341,0.284768641,-0.1475832313,0.0460562706,0.0708929226,-0.004852992,0.2478780597,-0.0120960735,-0.2618584037,-0.0627104938,0.1765855104,-0.2563492954,-0.0823205709,-0.5738776922,-0.1079507396,-0.059426643,-0.0769386441,0.4750975072,0.183270216,0.0703493506,-0.1054376885,0.1666536629,0.2087131739,-0.187780425,-0.3936911225,-0.8022010326,0.3011883199,-0.1008124501,-0.533334434,0.0156143578,-0.0288891774,0.3304855525,0.099997282,-0.3712185025,-0.3593392074,0.0408106893,-0.1371110529,0.4858476222,0.2386983931,0.1938264817,-0.1422718912,-0.1716768146,-0.2504088581,-0.0610037223,0.2543101311,0.3238006532,0.4589898288,-0.3080552816,-0.3199735582,-0.2250135243,0.0377587937,0.2135538608,-0.1853589118,0.0062891641,-0.0813466161,-0.0969929695,0.0534609221,-0.1199839264,0.6289497018,-0.1266483068,-0.068536967,0.1195929274,0.1906864047,0.2550229728,-0.0025167516,0.0918791816,-0.2440904975,-0.1512140036,-0.2004415691,-0.0462499596,0.0248736907,-0.0037870368,-0.0744263455,-0.1132846773,-0.104645595,-0.079851374,-0.0138376746,0.1383136511,-0.0518495925,-1.0071560144,0.2406782359,0.0479068495,0.3997605145,-0.2963574231,0.0098663876,0.0599974766,-0.1204705685,0.0912829563,0.2207179964,0.1551355273,-0.2316032946,-0.2294846624,0.3070248663,-0.0623219125,-0.2843466103,-0.0945147648,-0.1433060169,0.0610818714,0.2233985364,0.0505927801,-0.2597231567,-0.0079930546,0.1925090104,0.3008553088,-0.3191562891,-0.1693003923,-0.070815295,-0.0546886735,0.0299625024,-0.1493470222,-0.1430878788,0.2711488903,-0.2524779141,-0.1067786515,-0.1950679272,-0.1309178323,0.5445533395,0.033055149,0.463547498,0.0222121254,0.3310345411,0.3319663405,0.1000608504,0.2157465965,0.4837758839,-0.1562196016,0.2593025267,0.1151090264,-0.3584939539,0.376213491,0.1389822215,0.1031427383,-0.4917184711,-0.2125292718,0.0972299948,-0.1845370233,0.2611611784,0.0735679269,0.1788426787,-0.0851311088,-0.2825027704,0.5260211825,0.1357911527,-0.0083499635,0.3781108856,0.3599123955,-0.1380317658,-0.1896424294,0.5227787495,0.8718810081,0.0974758863,-0.1980103552,0.282738924,0.3868581951,0.332857579,-0.4695435166,0.0674933791,-0.2265428156,-0.314727962,0.0406505913,-0.1031452045,0.1411352158,0.2470818609,0.1490348577,0.0992128626,-0.1775530279,-0.2722309232,-0.1862836927,0.1960592717,-0.2520698905,-0.1280095279,0.3076058626,0.2597720027,0.0755261183,0.1924241334,-0.0095113497,0.0551902726,-0.3189354837,0.0040239627,-0.087775141,-0.0496271737,-0.0976950154,0.0431804769,0.047340855,-0.3594697118,-0.0802837461,0.3594199121,0.2364555001,-0.0355737992,-0.0733912066,0.1690243781,-0.3461524546,0.082434915,0.3339087069,-0.0104907062,0.3831386566,-0.1655253321,-0.2113124877,-0.0618125424,0.0983557999,0.238509655,-0.3904206753,-0.0386755094,-0.0042906716,-0.4929553568,-0.0564682074,0.1873193681,-0.061716672,-0.0932327509,0.189946726,0.2423671037,-0.0621952154,0.4939071834,-0.1194913313,0.0269078575,-0.1081115007,0.2109414786,0.4708757401,-0.1838504076,0.2785853744,0.0169368703,0.0392064564,-0.1465882063,0.0336620361,0.4933127165,-0.1134642661,-0.0812809393,-0.2681727409,-0.2203373164,-0.0074136523,0.225982964,0.2839486599,-0.00280051,0.0831825882,-0.3508226573,-0.4850940406,0.3159910738,0.1556036621,0.4166255593,-0.2392245829,0.2694608271,-0.0715032518,0.0385216214,-0.3680201769,0.0264106132,-0.0815889984,0.1011909321,0.0922044814,-0.0736327618,0.2503101826,-0.3807269931,0.2025808245,0.0263749864,-0.1334950775,-0.2435706705,-0.0969257429,0.0897025988,0.151542604,-0.0670351833,0.2224101275,-0.0237273369,-0.0340131335,-0.0770204514,0.0594956093,-0.1047699004,0.0720104054,0.3274497092,0.052566018,0.1143008396,-0.0189366229,0.1280949265,0.1096407026,0.2754950821,0.1357036233,-0.2312706858,0.0451415665,-0.205426231,-0.3339149654,0.0152442567,-0.052652996,0.1637735218,0.2915730178,-0.0418316498,0.1261018068,0.1449692547,0.3656479418,0.4139702022,-0.1543257087,0.1058540046,-0.0190595444,0.3068211079,-0.3131523728,-0.0106047513,0.0555676483,0.3534428775,-0.0921317488,0.0831032842,0.1575088352,-0.0645070076,-0.1458211094,0.159537971,0.2425075471,-0.0942249596,0.113700375,0.1915373206,-0.1950541586,-0.0373639055,0.1334225535,-0.0793756172,-0.0682396144,0.2332977802,0.0459243469,0.5049701333,0.190611884,0.1884461939,-0.0002468259,0.0380434133,-0.054433316,0.3526477516,0.1662671715,0.1667621583,0.0506561473,0.5731660128,-0.17005606,-0.0390254259,-0.3009103537,-0.0013903392,0.0906204879,0.1567979753,-0.2475579083,-0.1205097809,0.079577148,-0.1994359344,-0.1233735308,-0.2324358672,0.1365020424,-0.0118772406,-0.3977286518,-0.7796933055,-0.1341340244,0.1816382855,0.2501069605,-0.1011051759,0.1526043713,0.5754841566,-0.0296983309,0.0908437446,0.5537942052,0.14276658,0.1652055085,0.0234525409,0.0788345709,0.2297348827,-0.0456622951,0.1259018332,0.1113648713,-0.0031882927,-0.1627569348,0.2816799581,0.2152389288,-0.1006895453,-0.1492403299,0.0540590957,0.0037768953,-0.1000726745,0.0883461535,0.0580536313,-0.1477189064,-0.1118036881,-0.2121132761,-0.3597826958,-0.0281045511,-0.1232801527,0.1838154048,0.1735625863,0.0116041042,0.1354414523,-0.3553147316,0.2740556598,0.0691395774,-0.1476635337,-0.2951166332,-0.1201750338,-0.6131010652,0.0267107747,-0.0536688752,-0.5478016734,0.1653174013,0.3015881777,0.1290454119,0.034483321,0.0680996627,-0.4246810675,-0.1527599245,0.2639064789,-0.5903574824,0.0653875768,-0.2221127152,0.157260552,-0.0014695271,0.2976365387,0.232339561,0.2318881154,0.0173672363,0.1468279362,0.1510249823,0.1870525181,0.230774343,0.027423989,-0.0097883437,0.4499776661,0.1563043743,0.2933291793,-0.3235432208,-0.224089995,-0.203477487,0.3270482421,0.1905537695,0.1206229478,-0.0125326049,0.0145754768,0.0647433177,0.2950315177,0.0132933594,-0.429828763,-0.5134843588,-0.0048792008,0.0143990917,0.0360000357,0.2718863785,0.5303260684,-0.0038960101,0.1137369648,-0.2014437169,-0.3758945167,0.4186391234,-0.1064850986,0.1832005233,-0.0508390442,0.2534042895,0.4186257422,-0.0405696705,-0.3476036489,0.0056091705,0.420242697,-0.2656207383,-0.1188366115,0.2509295642,0.3692400753,-0.1447825134,-0.0115942638,-0.0424356237,-0.0008025188,0.0327931494,0.1694105566,-0.0311146583]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"Thanks, looking forward to hearing your update on this thread. \r\n\r\nThis is a blocking issue for us; would appreciate any progress on this front. We can also help with the fix, if you deem it appropriately. ","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":36,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nThanks, looking forward to hearing your update on this thread. \r\n\r\nThis is a blocking issue for us; would appreciate any progress on this front. We can also help with the fix, if you deem it appropriately. ","embeddings":[-0.0870945305,-0.376845479,-0.1521641165,0.4688611627,0.3402540982,-0.0078841466,0.1615420133,-0.0307646543,0.247697413,0.2832652032,-0.2237005979,0.2499283105,0.1018444821,0.4539228082,0.0879233852,0.1007004231,-0.0352866352,-0.1102674901,-0.3657335341,-0.2921965122,0.0072301151,0.0780227259,-0.067230843,-0.145237416,-0.6695821881,-0.0038879933,-0.1209945753,0.2803735733,-0.104296118,-0.214604184,0.2357613444,0.038694676,0.0067209424,0.5801949501,-0.0001046301,0.0152195944,0.1473651379,-0.0074839341,-0.137478441,0.2280700654,-0.1725152433,-0.0180964712,0.0208941996,-0.3256729841,-0.1103190482,0.0440427363,-0.1437742859,-0.3669281304,0.2572000921,-0.0139501514,0.2280746102,0.1909845173,0.1549105197,-0.0496396571,-0.0950625315,-0.117370382,0.0789831281,0.2292114943,0.4040428102,0.2403361648,0.0380713977,0.1594725549,0.0983691886,-0.1966067106,0.280148685,0.0289579406,0.3426020741,-0.2274544388,-0.2999531627,0.2939437628,0.847538352,-0.0466870405,-0.2914670408,-0.0303705409,0.0087754317,-0.1420633048,0.1820064187,0.0597105511,0.3434993029,-0.0311717447,-0.2439133376,0.4149996936,-0.2541270852,0.0437538289,-0.1570800394,-0.1155486852,-0.0435795598,-0.0082371747,-0.2389372438,-0.0282255746,0.0495098643,-0.0512786731,0.1646589935,0.1823989898,-0.3299439847,-0.1318350881,0.0134280017,-0.0122940484,0.1998295635,0.158960551,0.2649376988,0.101357922,-0.3621605337,0.0698371008,0.2414426655,0.2587036788,0.424094826,-0.5478389859,0.0477879122,-0.0696193799,-0.1134538278,0.0816056654,-0.3491240144,-0.0449133292,-0.0354393683,0.0332461707,-0.1748821288,-0.3821943104,-0.2380211502,0.258543402,0.0388883092,-0.0881350785,0.1911207139,-0.0995258391,0.0611412935,0.3915018439,0.0341392756,0.0414290391,-0.5715643167,-0.2726369202,-0.350409776,0.0863761976,-0.0972441807,-0.0213427376,0.1139399931,-0.1042929515,0.3827908635,-0.0802483857,0.2764771283,-0.0013614991,0.0811330006,0.1100352705,-0.352294445,0.2302160561,0.0268044937,-0.012643843,-0.0359075926,0.0272926949,-0.0133335879,-0.0298455264,-0.1658156812,-0.2073694766,-0.0207642987,0.3233802617,-0.3328650892,-0.2567097247,-0.1558976173,0.0178248957,0.1853138208,0.1802814007,-0.2122165859,-0.1361185759,0.0601905137,-0.1134830117,0.175613597,-0.0649841204,-0.589594841,0.0038927314,-0.0142197274,-0.2840304971,0.034830153,0.0177170858,-0.0126620792,0.2226736248,0.1310677528,0.1747597307,-0.5439710021,-0.4694362879,-0.1393547952,-0.358481586,-0.2200030088,-0.1051341593,0.1227104589,0.078452684,0.0269541945,-0.1187636331,0.1633792222,0.0666821823,0.1071553156,-0.1604182273,-0.4459184706,-0.0681828856,-0.032296706,0.2091264427,-0.0600988008,-0.3040694594,-0.0659138411,0.3356044888,0.1652618349,-0.0310724322,-0.0783647597,0.0805356055,0.3151931763,-0.1335745156,0.105501011,-0.2058998495,-0.0157993473,0.3022418916,0.0759724379,0.2600725591,0.7314265966,-0.0348708071,-0.3778423071,-0.2253963351,0.0159748029,-0.1700487286,0.2046742439,-0.2249566913,-0.0247681905,-0.203340441,-0.1770487577,0.1475644708,-0.5767453313,0.0628687665,-0.4978930354,0.3528924584,0.0002690142,-0.0101232976,0.0724387914,0.3005820215,-0.0322305411,-0.2561036348,-0.0397770181,0.5321791172,0.1301266104,0.0485317819,0.204424113,0.1736721247,0.3856614232,-0.1463344693,0.1554400921,-0.2496361881,-0.0562156551,-0.2027371228,-0.3685398698,0.2604891658,-0.1404676884,0.0063304231,-0.0556863397,0.0379640423,0.0084258281,-0.1628029644,0.0039468156,0.024797935,0.2130032182,0.1512256265,-0.2448852956,0.0545095764,0.0214266181,0.39623034,-0.3434987962,-0.1455949098,-0.207307294,-0.3295723498,-0.1979645193,0.1728601009,0.1191325784,-0.0176168717,0.4575639665,0.3126805127,0.069046028,-0.0542527623,-0.0904250219,-0.1600152403,0.0215936992,0.0304689948,-0.0730815455,-0.0313940831,-0.0046664374,-0.0337766148,-0.1768553853,0.0974946544,0.1154663488,0.1296596676,-0.4367828071,-0.3674219251,-0.2122879028,-0.0988145471,0.2028766274,0.1948630363,0.1293786168,-0.2920304835,0.1910259873,0.1212311536,-0.0716174617,0.0392532833,-0.3408944309,0.2852011919,0.0318596847,0.1425779164,0.0811291039,-0.1916878223,-0.1584848911,0.209480539,0.034960445,0.3384016156,0.3835898042,-0.0523936264,-0.0822839215,0.151828438,-0.4507327378,0.021873204,-0.1935034692,0.2933571637,-0.1625696868,0.0462265015,0.0689677969,-0.0419375561,0.2706375122,-0.0292679723,-0.2540873587,-0.087960273,0.2068856359,-0.2462812215,-0.1089710444,-0.5483402014,-0.097533077,-0.1158169881,-0.0183370374,0.443531841,0.189378798,0.0688930154,-0.1006862372,0.1724812686,0.2226365805,-0.1803920269,-0.3716886342,-0.8080675602,0.2895078361,-0.1316983402,-0.5295535922,0.0163203701,-0.0122795478,0.3129897118,0.0978410318,-0.3896394074,-0.3701252937,0.0081394035,-0.0939100534,0.4827438295,0.1899993569,0.2229121029,-0.1478611678,-0.1982781589,-0.2646466196,-0.0628993437,0.2635391653,0.2962849438,0.4528547525,-0.3367644846,-0.2986631691,-0.1789289117,0.0699736848,0.2015919089,-0.1326644123,0.0119043579,-0.0227626879,-0.0965105891,0.0338495709,-0.1430670023,0.610381186,-0.1313683093,-0.0790322348,0.1713482589,0.2282406241,0.2096860558,-0.0391468704,0.0601386763,-0.1790927798,-0.158929348,-0.1995088607,-0.0256968532,0.0269991755,-0.0228315648,-0.0663128272,-0.1054532304,-0.0996057615,-0.0844941586,-0.0198744554,0.1182788834,-0.05592677,-1.0049803257,0.2591084242,0.0375464484,0.3526399136,-0.3005052805,-0.0288194586,-0.0260787513,-0.1135907546,0.1070967466,0.2032565325,0.1990321726,-0.2164061815,-0.2058831453,0.3090201318,-0.0924059227,-0.2754475176,-0.0688432306,-0.1381198615,0.0797670409,0.2177277803,0.0466142297,-0.2844489217,-0.0444503687,0.1784988493,0.2122952342,-0.2845430374,-0.1637993306,-0.0753103644,-0.0201927274,0.0189944487,-0.091253005,-0.1106497124,0.2644147575,-0.2887608409,-0.1394648999,-0.1535237581,-0.1281749457,0.5450181961,0.0543302558,0.4168041945,0.0122888964,0.3488400578,0.357932657,0.0919462368,0.241031453,0.4887815118,-0.1336218417,0.2437861562,0.144253552,-0.4045612812,0.3732375205,0.2094112188,0.1001695991,-0.4752340317,-0.2483513355,0.0873067304,-0.1790094078,0.2362406254,0.0920023173,0.188735038,-0.0409707837,-0.2470601201,0.5007680058,0.1151901931,-0.0015584949,0.4271888435,0.3791874051,-0.1314914674,-0.1562880129,0.508336246,0.8966671228,0.082828708,-0.1813944429,0.2619684339,0.3698830903,0.3061384261,-0.4772644937,0.0974359661,-0.1830943972,-0.3640808165,0.054701671,-0.1012043282,0.1320711374,0.2592510879,0.1441241205,0.0532430857,-0.161688149,-0.2868407965,-0.1962385029,0.195965454,-0.2668993771,-0.1652957797,0.2729056478,0.2745414376,0.1123940721,0.1550142318,-0.0224235766,0.0194456782,-0.3178414106,0.0096410019,-0.076784119,-0.0440924577,-0.0591485016,0.0513107963,0.0463696085,-0.3314214647,-0.026158629,0.3168106973,0.230423674,-0.0533089191,-0.1130508408,0.1902426481,-0.3336274326,0.0785722807,0.3283728063,-0.0032496853,0.4269936085,-0.1823468804,-0.187206313,-0.060552489,0.0920939445,0.2167872041,-0.3755587339,-0.0423262306,-0.0067093302,-0.4976553917,-0.0271981433,0.1701218635,-0.0803688169,-0.1118330061,0.2077974379,0.2082478553,-0.0794618651,0.4497679174,-0.0659509897,0.0382523127,-0.0877890885,0.2386271507,0.4173071086,-0.1835372448,0.3220867217,0.0525721759,-0.0148553262,-0.1745233536,-0.0034505867,0.482532084,-0.1126062274,-0.0369279832,-0.2350037247,-0.2543023527,-0.009773476,0.2487630248,0.2622337043,-0.0032235384,0.08630725,-0.3580668867,-0.4583057463,0.3008619845,0.0951448157,0.4245128036,-0.2772362828,0.2854068279,-0.0411140174,0.0664204583,-0.3865275383,0.0025517812,-0.1480125189,0.0805461407,0.0871705115,-0.0949058682,0.2682088614,-0.3732394576,0.1924354732,-0.010902429,-0.1877463162,-0.243808344,-0.089805007,0.0889479369,0.1451299787,-0.0835331306,0.2687237263,-0.0355799012,-0.0553222932,-0.0533559173,0.0692585334,-0.1181078404,0.0437644906,0.3051767647,0.0440095477,0.1408572346,-0.0136882886,0.1271900982,0.1204276681,0.2415245324,0.1501477808,-0.2719298303,0.0273837391,-0.2203309089,-0.2931700349,0.0165194422,-0.0626758188,0.1528554559,0.2784412205,-0.0615113229,0.1116776988,0.1647971421,0.3731776178,0.3808892071,-0.1324039549,0.1207651645,-0.0105665624,0.3303654492,-0.2878557444,0.0007567775,0.0790150538,0.3087944388,-0.0843267068,0.0466641225,0.1687319875,-0.0542566329,-0.1187323332,0.1909874231,0.2143381685,-0.1191221699,0.1166087463,0.1912384927,-0.2486340255,-0.0254880171,0.1562619954,-0.070154652,-0.0570088662,0.2388652861,0.030884698,0.4838153124,0.1535273343,0.1727787107,0.0172187015,0.0146198701,-0.0558257587,0.3371435702,0.1983417124,0.1350704432,0.0260551404,0.6015375853,-0.1991492361,-0.0517543219,-0.3056697845,0.003507318,0.0700092241,0.1575064808,-0.2648793459,-0.1183076948,0.0507003702,-0.1897433549,-0.1171694547,-0.2110818475,0.1835687459,-0.0126266917,-0.3362026215,-0.8137513995,-0.1469639391,0.1683219522,0.298080653,-0.1029447913,0.1609456539,0.5675946474,-0.0264649466,0.0921069756,0.5185127854,0.1119827405,0.157761097,-0.0201044995,0.0527727641,0.2108608335,-0.0534414165,0.104845956,0.1572995335,0.0018041636,-0.1253673285,0.2945418358,0.2084714323,-0.1260446459,-0.1433202624,0.0451989286,0.0007639987,-0.0977523848,0.0937253162,0.0713279098,-0.1590177566,-0.098114863,-0.256118238,-0.3406968415,-0.0280125868,-0.1255313903,0.1458764076,0.1647303551,-0.0061447849,0.1409906447,-0.3291229606,0.3090844452,0.054998517,-0.132885322,-0.2961922288,-0.1193375587,-0.5893236399,0.0456676036,-0.0611139946,-0.5289897323,0.1521455497,0.3244359195,0.1560740173,0.0205802023,0.033994507,-0.4579609632,-0.1055276617,0.265213728,-0.6133133173,0.1041930392,-0.1988890469,0.1499170214,0.0008537704,0.2935955226,0.2201204449,0.2370463908,0.0215038657,0.1590389311,0.1622433066,0.141945228,0.2237860113,-0.0166986659,-0.0228509195,0.4889416397,0.1733771116,0.2755285203,-0.333951354,-0.2324575931,-0.2272508293,0.2942470312,0.2213608176,0.15554519,-0.0377789363,0.0029990275,0.0663448125,0.3365394175,0.0028025659,-0.414755404,-0.4890245497,0.009886642,0.0536733717,0.0490888543,0.2479671538,0.510253191,0.0132308695,0.0540295616,-0.2399694175,-0.4003309309,0.4349494874,-0.1649881601,0.1486083418,-0.0401165001,0.2681303024,0.4486649036,-0.0784517154,-0.3514889777,-0.0192534961,0.3600772321,-0.2413725108,-0.1331915408,0.2997153401,0.3858632147,-0.1430396289,-0.0351726934,-0.0849711075,0.0144364266,0.0524505042,0.1392469108,-0.0340140462]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"I just started the generation on my side, I'll let you know how it goes :) ","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":16,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nI just started the generation on my side, I'll let you know how it goes :) ","embeddings":[-0.0991990268,-0.3942707181,-0.1512490511,0.4992507696,0.3250366747,0.0082152458,0.1899509579,-0.0334032997,0.219748348,0.2975789905,-0.1856802404,0.2214599699,0.0975367576,0.4401781857,0.1296778321,0.1508605331,-0.0050972672,-0.0683598295,-0.3730553985,-0.3198929727,-0.0501614362,0.0540433154,-0.0487868935,-0.1231350377,-0.6788250804,0.0470918268,-0.1302306056,0.2782609463,-0.1247740686,-0.1770287752,0.2357266694,0.0572866686,0.0347149819,0.5405398011,-0.0001053364,-0.0207368452,0.1016165614,-0.0306879506,-0.0849425122,0.2506739795,-0.1658851504,-0.011623526,-0.0331669152,-0.3924215138,-0.079961136,0.0515269712,-0.1315433681,-0.382024616,0.225756377,-0.0086584343,0.2169919014,0.1548500061,0.1296726614,-0.0564978309,-0.0447452292,-0.0705575719,0.080738835,0.2239102125,0.360861212,0.2562358975,0.072766073,0.1726240963,0.0762178451,-0.1872816086,0.3275933862,0.063896291,0.3388044238,-0.2582989633,-0.2919934988,0.298853159,0.836256206,-0.0605643615,-0.2856417,-0.013232572,-0.0155020151,-0.136931628,0.1729129851,0.077917628,0.3255273998,-0.0402636714,-0.2604097128,0.3819594979,-0.2595346868,0.0025319546,-0.0883048549,-0.1223710105,-0.0531104468,-0.0039930139,-0.2595849931,-0.040273767,0.1087621152,-0.0885707214,0.1688392013,0.2064000368,-0.2878933549,-0.117923595,-0.0223937053,-0.0198950078,0.2344019413,0.1514639854,0.2566113472,0.0228564851,-0.3486999571,0.0581923351,0.1865371913,0.2489240468,0.4371138215,-0.5546713471,-0.0060968283,-0.1186807603,-0.122537896,0.0798360705,-0.3836108148,-0.0266290996,0.0392577127,0.0167447049,-0.1728738546,-0.3478006124,-0.2349817306,0.2712554634,0.0184829328,-0.092144914,0.1589310616,-0.1319564283,0.0435963795,0.4083958268,0.027719494,0.0327499919,-0.5587655902,-0.3067183793,-0.3500446379,0.1029693186,-0.069306992,0.0127624469,0.0917775482,-0.0723092258,0.3852658272,-0.0754475221,0.2438709438,-0.0177495796,0.065463461,0.1145852432,-0.3395024836,0.2190766335,0.0418366194,-0.0216601267,-0.0725513548,0.0082334392,0.0082001016,-0.0185361803,-0.1398711205,-0.2128385156,0.0050693084,0.3254516423,-0.3557157516,-0.265319109,-0.1458191574,0.0076628579,0.232460022,0.21268861,-0.2090700269,-0.1102747396,0.0803072155,-0.1324114501,0.1687544286,-0.0175945815,-0.6140215993,0.0293739308,-0.0249337126,-0.2568977773,0.087326169,-0.0196028613,0.015684722,0.260053426,0.1147686467,0.1682211608,-0.5175309181,-0.4441342652,-0.1044271365,-0.3349779844,-0.1746029854,-0.0914480165,0.1713097095,0.1171154305,0.0173606016,-0.1336980462,0.1548244655,0.0379910432,0.0789757222,-0.1654697955,-0.4269625545,-0.0553447381,-0.0285568126,0.1714439094,-0.0690064058,-0.2467221171,-0.106026724,0.3275652826,0.1769461036,-0.0419157557,-0.076879479,0.0513749942,0.3236053586,-0.1554244459,0.0769401491,-0.1586396545,-0.0301631559,0.3010194898,0.0683938935,0.2581225932,0.7335653901,-0.0615018159,-0.394277215,-0.2083212584,0.0027543746,-0.1839934736,0.1929906458,-0.1989011019,0.0005777174,-0.2506177425,-0.1730110496,0.1371758729,-0.5779194236,0.0343147554,-0.5610533357,0.3250957727,0.0066309562,0.0315568559,0.0762805715,0.2840644717,-0.033601556,-0.2483559549,-0.0288121868,0.5453804731,0.1616093069,0.0535059385,0.183155477,0.1931942254,0.3812390268,-0.1889412403,0.1518518478,-0.2646035552,-0.0439031646,-0.2053335458,-0.3492260277,0.2360476553,-0.114371784,0.0102000618,-0.0759773999,0.0012272017,0.0084400401,-0.1355068386,0.034424711,0.0099607781,0.1653895378,0.1845075935,-0.2131851912,0.0737973228,0.0260260757,0.4362431765,-0.3370135427,-0.155963406,-0.2162287682,-0.3423806131,-0.1967810988,0.1316416562,0.0940619558,0.0136564104,0.5156713128,0.2999622226,0.0722164661,-0.0422906913,-0.0838764086,-0.1740971059,0.0369105637,0.0046928958,-0.0680332556,-0.0423626006,-0.0661313161,-0.0061190771,-0.1768151075,0.1134700626,0.1010457724,0.1873127073,-0.4435324669,-0.3401427865,-0.14087677,-0.0707163811,0.2071112692,0.1828158945,0.1401575208,-0.2619085014,0.1876969934,0.0738056526,-0.0470513105,-0.0165371243,-0.3119051158,0.284686178,0.0427173972,0.1291547269,0.065996632,-0.1681566089,-0.1526060104,0.2071588188,0.0039650048,0.3634145558,0.4104003906,-0.1047771573,-0.0356321149,0.1301190555,-0.4129679501,0.0156611577,-0.2075364739,0.3164816499,-0.131006062,0.0450776815,0.073369123,-0.0389072113,0.2737810612,-0.0187504105,-0.2886246443,-0.1020701155,0.158085525,-0.265237987,-0.1001155898,-0.5698409081,-0.0769093186,-0.0694271773,-0.0577172562,0.4787977338,0.2264364511,0.08526586,-0.0846292898,0.1478224844,0.2793621123,-0.1870379001,-0.4015790522,-0.8443461657,0.3063878119,-0.1574644893,-0.5241183639,0.0214163158,-0.0598470867,0.3482739925,0.1301286519,-0.3654982448,-0.325777024,-0.0152124651,-0.1198371574,0.5110429525,0.2357797772,0.1952349544,-0.1350649446,-0.1865757406,-0.2790879905,-0.0319170728,0.2907136977,0.2476892918,0.4593497515,-0.2880563736,-0.2950657308,-0.25883618,0.0697335228,0.2178661674,-0.1610463113,-0.0074187131,-0.054726202,-0.0970285535,0.0554268956,-0.1517688632,0.6240478754,-0.0967046246,-0.085415408,0.1625877321,0.2218758911,0.2558006942,-0.0541107208,0.1269096285,-0.2075278908,-0.1293875426,-0.2394885719,-0.047531154,0.0372425877,-0.0244951975,-0.0653473213,-0.1325956732,-0.0933662578,-0.1098584458,-0.0593821071,0.1681641787,-0.0372900739,-0.9834538102,0.2603168786,0.0349398293,0.3408542573,-0.2668491602,-0.0271405336,0.0263139326,-0.1147090793,0.0955428034,0.2108888775,0.1977613717,-0.2447779924,-0.2083047032,0.2636697292,-0.0613511279,-0.2832501531,-0.0488575213,-0.1174369827,0.0494408272,0.1923976094,0.0555345155,-0.307526499,-0.0462106988,0.1702316552,0.2238605171,-0.3208246827,-0.1889399886,-0.0829787329,-0.0303321742,0.0447138213,-0.1673035473,-0.1453817487,0.3055654168,-0.3080767691,-0.1369552612,-0.1649604589,-0.0949142352,0.5748013854,0.0809492692,0.4154922664,-0.0184975788,0.3374986053,0.3749863803,0.0860504806,0.2072853893,0.486405164,-0.1634875834,0.252674818,0.1326466948,-0.3736195564,0.3611027896,0.1413076222,0.0928632393,-0.5118597746,-0.2403585613,0.0719364956,-0.1828967482,0.247979477,0.1052110642,0.1559087932,-0.0717962757,-0.3228155673,0.5426487923,0.1231088266,0.0081223436,0.3860912621,0.3564715087,-0.1340126097,-0.2051940113,0.5239417553,0.8795143366,0.0875157043,-0.1918264478,0.2991400361,0.4096212089,0.3452936113,-0.4674289823,0.0430298671,-0.235551253,-0.3408919871,0.0505557172,-0.1015463471,0.1307526529,0.3114386797,0.1271600574,0.0346694216,-0.1633205712,-0.2763163745,-0.1857145876,0.1975483745,-0.2770605385,-0.1748844534,0.2784390152,0.274924159,0.1019848362,0.166597724,-0.0270531196,0.0397352763,-0.3167531192,-0.0052995356,-0.0553643145,-0.0299164616,-0.0715082586,0.0308821537,0.0501126423,-0.3300329745,-0.0463315696,0.3366264105,0.2604948878,-0.0603243187,-0.0944990963,0.1569067985,-0.3263677657,0.0590722784,0.3139480054,-0.0156512763,0.382368058,-0.1759953946,-0.196354121,-0.0452414639,0.0787417963,0.2471217513,-0.381867826,-0.0753502548,-0.0234805886,-0.4934106171,-0.0453873686,0.1884387583,-0.0848584548,-0.1015375629,0.1960271895,0.2374441177,-0.1112999544,0.4571844935,-0.1239637882,0.0385814123,-0.1148345992,0.1939299405,0.459815532,-0.1795016378,0.2613755763,0.0343137346,0.0038184691,-0.1796964109,0.0693927184,0.4981959462,-0.1235932857,-0.0410488471,-0.2215934396,-0.2235259265,0.0073039411,0.2483253032,0.2811186016,0.0048839808,0.1370187998,-0.34417364,-0.4759086668,0.2609249055,0.1104766428,0.4316190779,-0.2707903683,0.2353208959,-0.0358550586,0.0875998512,-0.3749723136,-0.0241381619,-0.0902673304,0.0737866908,0.0886549205,-0.0670768544,0.2347098589,-0.3692100048,0.1837601662,-0.0140696503,-0.179276824,-0.238886103,-0.1359959841,0.0926149338,0.1513487995,-0.0998277292,0.2642742693,0.0025740797,-0.0953658968,-0.0564984567,0.0855174214,-0.1130073071,0.036393173,0.3185964823,0.044399377,0.1525774896,-0.0902124643,0.1271760017,0.1196649224,0.285410136,0.1434613615,-0.2454919815,0.0355580412,-0.2511769831,-0.295055002,0.0342085026,-0.0594218783,0.1259425581,0.2847483158,-0.0941464677,0.1497996449,0.1781172752,0.3441836536,0.3828387856,-0.1077492088,0.1276454329,-0.0130426576,0.324131906,-0.3050273657,-0.0148575837,0.04321279,0.2958384156,-0.0976335108,0.0905088559,0.1272356659,-0.0715451092,-0.1671731621,0.1673764884,0.193797037,-0.0980521962,0.121783562,0.2365946621,-0.2271397114,0.0157359038,0.1643600017,-0.042667225,-0.0649332702,0.2640988827,0.027695151,0.486045599,0.1873307228,0.1617645174,0.0328324325,0.0120532848,-0.084395729,0.3501393497,0.1519107074,0.1584688723,0.0076012844,0.5896815658,-0.2219512314,-0.0162749533,-0.3264860511,-0.0000326935,0.101241976,0.157463789,-0.2557415068,-0.1325260699,0.0417880602,-0.2113118023,-0.1075144112,-0.2485521734,0.1659447998,-0.0226367582,-0.3308962584,-0.7679842114,-0.1013107821,0.1610024124,0.2555581629,-0.0982127413,0.1442522556,0.5694270134,-0.0578342974,0.0739403889,0.5254844427,0.1098723561,0.2007332444,0.0134931104,0.0366667658,0.2117245197,-0.0628803521,0.0946505442,0.1388875395,-0.0252648313,-0.1828072667,0.2713635564,0.2025996149,-0.1187585071,-0.1726142019,0.1031561866,0.0065848432,-0.0567352474,0.0905831233,0.0345116146,-0.1706051677,-0.0913104191,-0.1935312301,-0.3727800548,-0.0227537472,-0.0996093452,0.1810112,0.1765191704,-0.0164286103,0.1349050999,-0.3623852432,0.3160422444,0.0677891523,-0.1362581998,-0.3066979945,-0.1303228289,-0.5872722864,0.0199033376,-0.0568403006,-0.5123307109,0.1407916993,0.3428821564,0.1273254156,0.0376042649,0.0000411544,-0.4160141647,-0.1572107375,0.3020840883,-0.5865224004,0.0664891601,-0.2207361311,0.165390715,-0.0105116675,0.2680391073,0.2200222015,0.2306521982,0.0275038928,0.1574284881,0.2008642554,0.1931498498,0.2320534885,-0.0202820599,-0.0036087416,0.4715546966,0.1600963175,0.2669734359,-0.2912588418,-0.2093981355,-0.2012654096,0.3093336225,0.2263792157,0.1445259005,-0.0006394225,0.0483767837,0.0573552735,0.32667014,0.0598765239,-0.4564238787,-0.4984981418,0.0336639173,0.0071401433,0.0170115046,0.2858602703,0.4912389219,-0.0170576759,0.0980672613,-0.2187522054,-0.3829304874,0.4389589429,-0.1473555714,0.1580481529,-0.0535886995,0.2750998437,0.4618854821,-0.0758124217,-0.3238136768,0.0202648696,0.4070436954,-0.2786367238,-0.1159614325,0.2650029957,0.379435122,-0.1333647668,-0.0222673602,-0.0395553075,-0.0174359288,0.0412032343,0.1854811758,-0.0360981263]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"Hmm after a first run I'm still missing 136668\/226711 urls.\r\nI'll relaunch it tomorrow to try to get the remaining ones.","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":21,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nHmm after a first run I'm still missing 136668\/226711 urls.\r\nI'll relaunch it tomorrow to try to get the remaining ones.","embeddings":[-0.0192339066,-0.2787488699,-0.1315880716,0.4514377117,0.3294107616,0.0132473074,0.0834711641,-0.0324111395,0.1632886678,0.25323686,-0.214021951,0.181240797,0.1041179374,0.3433215916,0.1806179285,0.1001857221,-0.0398060568,-0.1415234357,-0.4157710373,-0.3789446652,-0.1317090839,0.1342010051,-0.1089954525,-0.1242683679,-0.611969471,0.0783976391,-0.2388474196,0.3378801346,-0.0854114518,-0.2222675532,0.2241359353,0.0637293831,0.089367874,0.5341081619,-0.0001078297,-0.0220556576,0.1424761713,-0.0170413572,-0.1624374986,0.2538862228,-0.231166482,0.0268421471,0.0482006781,-0.405474782,-0.0263640303,0.1173033118,-0.0946482047,-0.3096888065,0.158078745,0.0959425792,0.1845942587,0.1571917236,0.0893572941,-0.0256532505,0.0051862001,-0.0828036368,0.1446168125,0.2197121382,0.3912556469,0.3032293618,0.0680401847,0.1804811209,0.0612244383,-0.2153697014,0.2798307538,0.098128058,0.2757803202,-0.2437672019,-0.3017998338,0.2616388202,0.797362864,-0.0047415853,-0.2618876398,-0.0269741826,-0.0765680969,-0.1153860837,0.2018673569,0.0557033569,0.3726731539,0.0041704988,-0.3282065392,0.3529005647,-0.2558207512,0.1323099285,-0.1288596541,-0.0443034545,0.0117962733,0.0079704309,-0.2504276931,0.0427734144,0.1230294555,-0.1049915478,0.2351700515,0.2270646989,-0.2158429772,-0.1437883079,0.060984429,0.0052842968,0.247093156,0.2200070769,0.2636894882,0.0973217487,-0.4094589651,0.0266528986,0.2186471969,0.2032875419,0.3677200377,-0.5091800094,0.0934059843,-0.0424155034,-0.0867900103,0.0701019987,-0.3361897171,0.0212913696,-0.0937169343,0.088907592,-0.1943087429,-0.3586162627,-0.1624151319,0.2117015272,0.0353169329,-0.136645779,0.2173826247,-0.1654897928,-0.0143268658,0.4042074382,0.0237145703,0.0212501306,-0.539381206,-0.2602357864,-0.3475570977,0.118217729,-0.0596303716,-0.0100992732,0.125677377,-0.0402502157,0.4135729074,-0.1710354835,0.2352254093,0.0067364755,0.0788006037,0.1520160735,-0.3228069246,0.2311338782,0.088969335,0.0422685146,-0.0792255104,-0.0212146528,-0.0260818787,-0.0076279873,-0.2080206126,-0.2832221091,0.0483163781,0.305248946,-0.3953118324,-0.2976082265,-0.1643511057,-0.0514992774,0.0617100298,0.2127865404,-0.2347178608,-0.0692887083,0.1486749947,-0.123633042,0.2656905055,0.0006675372,-0.4929637015,-0.0078990664,0.0381508395,-0.244260177,0.1114944592,0.0165498052,-0.0232431199,0.2643283308,0.0746211782,0.1598430425,-0.4744036496,-0.4691445529,-0.0426469147,-0.3227778673,-0.1889185607,-0.1445510685,0.0905963853,0.097147204,0.0018750384,-0.1155165285,0.2047923505,0.0755230337,0.0930427089,-0.2053320259,-0.4654321373,-0.1000907123,-0.0738273486,0.3007145822,0.0293438658,-0.299972564,-0.1004413813,0.3271587789,0.2232566029,-0.0668678582,-0.0878526047,0.1660561413,0.2788275182,-0.082377702,0.1109275371,-0.186829254,0.04142588,0.310914129,0.0750924349,0.2592254579,0.6795742512,-0.0216838457,-0.3754805624,-0.2413866818,0.0568287633,-0.1928550452,0.1584153324,-0.153461501,-0.0133115696,-0.1118040532,-0.2205297947,0.0573064238,-0.574752748,0.1063458994,-0.5254189968,0.3792724609,-0.0178775862,0.0787570924,0.1651778072,0.325288862,0.0451645851,-0.3062184453,0.0017507138,0.5623015761,0.160725683,0.0131055173,0.3588773906,0.1996974498,0.4656715095,-0.2428459674,0.0950104818,-0.1435525864,-0.0314964838,-0.2036189735,-0.339433074,0.2086922675,-0.1423111558,0.0569089204,0.0482508242,0.0840907693,0.0055028205,-0.1831810772,-0.0478080325,0.0529193655,0.2395460308,0.1336751729,-0.3054099381,0.0981754065,0.0219700914,0.3735604882,-0.3904986084,-0.2474121898,-0.2403516769,-0.2415396869,-0.2217556387,0.1088662669,0.0645262524,0.0066637751,0.4553330541,0.29851982,0.1484998614,-0.0245345216,-0.1464135349,-0.1921243221,0.0765454248,0.1265975982,-0.1374517083,-0.1376807392,-0.0914110988,0.0538808331,-0.2065616995,-0.0373752303,0.1147436649,0.102457881,-0.4339114428,-0.3744902909,-0.246969372,-0.1158339679,0.2495810986,0.1147520468,0.1498488188,-0.3861026466,0.1743736565,-0.0147843529,-0.1148667932,0.0133326417,-0.4554790258,0.2821137011,0.0333305188,0.0956916511,0.0830486789,-0.2291259021,-0.1364935786,0.1868573427,-0.0149911372,0.3306301236,0.3095939159,-0.1426821798,-0.0014282425,0.0614826344,-0.4463159442,0.0334600657,-0.1677661836,0.2153141201,-0.0946199223,0.0360548645,0.0441393554,-0.0399484448,0.2810697556,0.0453815982,-0.132837519,-0.0603456944,0.1951060146,-0.3132751286,-0.0596587546,-0.5780410767,-0.077090241,-0.0302528366,-0.0871214718,0.4910757542,0.2203983665,0.0963189751,-0.1681658626,0.1959906667,0.2443377525,-0.2565478683,-0.3770159483,-0.853300333,0.3729169071,-0.1395105571,-0.5446469188,0.0230577756,-0.0514396951,0.2916252613,0.1466273665,-0.467515558,-0.3342877328,0.0168569125,-0.1670075804,0.4653496146,0.2231473774,0.210283488,-0.11818441,-0.206025824,-0.249483481,-0.0421271659,0.2474309951,0.3107559383,0.555118978,-0.3252278566,-0.3398382068,-0.2012266815,0.1445459276,0.1646935195,-0.0298970174,0.0222609509,-0.0806642696,-0.1048713624,0.1389025599,-0.0809603781,0.6621094346,-0.1253650188,-0.069608964,0.2355695367,0.2911699712,0.2255482972,-0.1471757442,0.0858623981,-0.2458414435,-0.2381782681,-0.1877159178,-0.0227721855,0.0519184805,0.0170335099,-0.1006943807,-0.1145726815,-0.0720120147,0.0026852661,-0.0267572533,0.1960008889,-0.0618934594,-1.0400999784,0.2407350242,0.0374271013,0.3676604629,-0.3005236983,-0.015107411,0.0171065684,-0.1203267127,0.0688890517,0.2533524632,0.198268503,-0.2970383167,-0.1915179342,0.3072464168,-0.0602127202,-0.3816342652,-0.0156605504,-0.1553147435,0.0423862785,0.2700268328,0.0465191603,-0.2352750897,0.0022845657,0.2012717575,0.2129918635,-0.3095557392,-0.084908925,-0.0463386811,-0.2008480132,0.1059705019,-0.1783401072,-0.1248123869,0.2661638856,-0.2717261016,-0.1976008415,-0.140485093,-0.1948443502,0.550868988,0.02519246,0.4383126199,0.009295051,0.3014675379,0.4104407132,0.095239304,0.1888519526,0.4933249354,-0.1635041833,0.2424921244,0.1802616715,-0.4166297615,0.4104345441,0.2077487111,0.0506311208,-0.5396622419,-0.1687221229,0.127602607,-0.1933010668,0.3086474836,0.0555601269,0.1529191881,-0.1129278168,-0.2701956332,0.5335492492,0.0959653333,-0.0635619238,0.4199517667,0.4356451035,-0.1722257584,-0.1856307834,0.5252233148,0.9401891828,0.1180116981,-0.2010308504,0.285990119,0.3628589809,0.3200500906,-0.4389863312,0.1076745763,-0.0909889191,-0.3176470399,-0.0173184946,-0.0978896543,0.1144261807,0.2762749791,0.1894149929,0.1274252385,-0.1048102826,-0.2083971947,-0.0775121525,0.1711336374,-0.2112062275,-0.1323250681,0.3128091395,0.2388980091,0.1537725925,0.1969724745,-0.018821476,0.0476464219,-0.3277888,-0.0218333714,-0.0232866351,-0.0518541448,-0.1385186911,0.1136183366,-0.0270622559,-0.3384088874,-0.1717925817,0.3910569847,0.252484709,-0.0504942276,-0.1955041587,0.2480486333,-0.4263959229,0.0497898757,0.3418867588,0.0250158608,0.3480612934,-0.2133067548,-0.1571349204,-0.0502761826,0.0463956073,0.1945438981,-0.3856779933,-0.0647617728,-0.0856954306,-0.3978682756,-0.038868662,0.1745574921,-0.1564902514,-0.1440476775,0.1651884466,0.2185399532,-0.0178682618,0.4426909983,-0.0747731403,0.0288320314,-0.147003144,0.2139876038,0.4390409589,-0.2144436538,0.2931471169,0.0334881693,0.0004923187,-0.1198278815,0.0970496163,0.3956159353,-0.0324407592,-0.0934054926,-0.3001123369,-0.1183771268,-0.0597369447,0.2215801626,0.2761352658,-0.0285184029,0.0946879312,-0.3888754249,-0.4292516708,0.3315663636,0.1317848563,0.3345543444,-0.3417191207,0.2434822172,0.0145246675,-0.0421474837,-0.3486151993,0.0767510906,-0.1979299933,0.0903016552,0.0428444669,-0.050334353,0.3161005676,-0.3747347593,0.1644619852,-0.0852763727,-0.0834217742,-0.1887988448,-0.1612736732,0.1184464544,0.1784987003,-0.0808966607,0.2584476471,0.0195212457,-0.0407920219,-0.0206228625,-0.0118816961,-0.1061227024,0.0719916299,0.3074945807,0.0319890641,0.0527551025,-0.0597160608,0.1610738635,0.1112453192,0.3838041425,0.2100997269,-0.2839373946,0.0337956585,-0.2956763506,-0.2647460401,0.0088826669,-0.0692964941,0.0919447765,0.329475075,-0.0432856828,0.1233955175,0.2155062705,0.3226027787,0.4006254077,-0.2067875266,0.1559053361,-0.0439031795,0.2910838425,-0.261402756,-0.0116661433,0.0309898965,0.3084950745,-0.1409591883,0.0164672751,0.1240912527,-0.1446179003,-0.1146210805,0.1645523757,0.3177484274,-0.0181691367,0.1336696595,0.1239914075,-0.2424307913,-0.0135059096,0.059787754,-0.0122248679,-0.0660734996,0.1807798296,0.0019393035,0.5119621158,0.1870928407,0.169747293,-0.0036409881,-0.0299677327,-0.1239290088,0.2858310044,0.1655499041,0.1494066268,-0.064628467,0.5828592777,-0.2291433662,0.022733232,-0.2644196451,0.0174932387,0.1737523228,0.1246513575,-0.2637862861,-0.160568893,0.0370359272,-0.2200585753,-0.0629189685,-0.3337857127,0.1164907888,0.0125795668,-0.3229826391,-0.8925774097,-0.0908259153,0.1098231524,0.2754613459,-0.1272934973,0.1047695875,0.592668891,-0.0426868945,0.1159245595,0.4232537746,0.1104506403,0.1399971992,-0.1032320932,0.0757572204,0.2145930976,-0.0386865325,0.1354574263,0.1552173346,0.0579458475,-0.1702955663,0.3381465077,0.1881193668,-0.0985921845,-0.1025126427,0.0549401641,0.0699450523,-0.061226882,0.167812407,0.0603993461,-0.1503455341,-0.0869882479,-0.2606004775,-0.4143602848,-0.0776021779,-0.146246314,0.1356302798,0.1808057427,0.0280126277,0.1186301038,-0.3475383818,0.3643344641,0.1019619927,-0.1094222516,-0.3326221108,-0.2414314151,-0.6039551497,-0.0895560607,-0.0354717709,-0.5677009225,0.1905382872,0.2993082702,0.1537584066,0.0626469702,-0.042607449,-0.3574108481,-0.1242509484,0.1969060302,-0.6065823436,0.0870379731,-0.2483508289,0.271667093,-0.0730124563,0.3079241812,0.260155648,0.1919995546,-0.0033499773,0.1946383417,0.1530910134,0.1256481111,0.2274767458,0.0218566954,0.0238594543,0.4491069317,0.1651579142,0.2428272069,-0.2997594178,-0.2766379714,-0.0712353587,0.2449838817,0.1859591305,0.0427292846,-0.051421646,-0.0374948643,0.0562916584,0.3808864057,0.0428000614,-0.4350447655,-0.4997071028,-0.0145089785,0.0461161211,0.0758599266,0.3166548908,0.4712298214,-0.0947266743,0.0466996394,-0.1350196153,-0.3913336396,0.4939974546,-0.1814223379,0.1789795607,-0.0617736131,0.2855496407,0.4516842663,0.0753909945,-0.3038817048,-0.0554281399,0.4447309077,-0.2247909456,-0.1983687878,0.2394731343,0.3810158074,-0.1366303861,-0.0138139119,-0.0970000178,0.0049943854,0.006736889,0.1402167231,-0.0109383091]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"Update: I'm missing 36\/226711 urls but I haven't managed to download them yet","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":13,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nUpdate: I'm missing 36\/226711 urls but I haven't managed to download them yet","embeddings":[-0.0786484852,-0.3587717414,-0.1469128728,0.5101448894,0.3007078171,0.0413288213,0.1355312914,-0.0255487356,0.2116054893,0.2750748396,-0.2208661586,0.2018501163,0.1088096276,0.4054099023,0.1536063254,0.1125398725,-0.0107668405,-0.1048300341,-0.3727461696,-0.3230278194,-0.049412109,0.0881356895,-0.0959930941,-0.1305685043,-0.6366696954,0.0285530593,-0.163732931,0.2738044262,-0.1016616225,-0.2203664929,0.2498830706,0.0326056741,0.0439908579,0.5512446165,-0.0001064957,-0.0102737444,0.1216546446,-0.0262580961,-0.1309241205,0.2452793717,-0.2122080326,-0.0146167409,0.0220567882,-0.404873997,-0.0453346595,0.0481619909,-0.1241179556,-0.3442298174,0.1885783076,0.0384366512,0.2111816406,0.180425778,0.1202543154,-0.0671535358,-0.044425115,-0.1090220809,0.1268018186,0.2432686687,0.3858607411,0.2823593915,0.069826223,0.1650565267,0.0780529529,-0.1847888827,0.2842644155,0.0657752976,0.329177469,-0.2377765775,-0.2870981395,0.2864576876,0.8440484405,-0.0433222391,-0.2617526948,-0.0078279618,-0.0303761605,-0.124601379,0.1758044362,0.0809500739,0.3665628433,-0.0253524575,-0.2833823562,0.4096958935,-0.2574537396,0.0649301261,-0.1452460587,-0.081078276,-0.0485232212,0.0085586868,-0.2525517941,-0.0261823591,0.0863249376,-0.1084119827,0.2058113217,0.2092580795,-0.2674272358,-0.1104480624,-0.01029466,0.0013369605,0.2318953425,0.1648035645,0.2714709342,0.0452505797,-0.3984404206,0.0517967641,0.2074215114,0.2505196929,0.4076028764,-0.5350685716,0.0076874075,-0.0777196288,-0.1013191119,0.0812103897,-0.3757743835,0.002343684,-0.0263920128,0.0352237299,-0.1724919975,-0.3691794872,-0.2232318074,0.2634103298,0.031715963,-0.0976115093,0.1960591376,-0.1392249316,0.0292131547,0.3857043087,0.0119161187,0.0180005468,-0.5559822917,-0.3101855218,-0.3592703342,0.1109673902,-0.0530648641,-0.0222002454,0.110852845,-0.0720158666,0.4004827738,-0.1166237369,0.2740643024,-0.0031851865,0.0487857908,0.1095036045,-0.3340585828,0.2445800006,0.0671321973,0.0294872578,-0.0446272083,0.0057581961,-0.0325834975,0.0040111979,-0.1917028576,-0.2194197327,0.0227856133,0.3184735179,-0.3825630844,-0.2866083086,-0.1575307846,-0.0335176103,0.1435992271,0.1886528432,-0.2067073137,-0.1319628656,0.1127003059,-0.1455586702,0.2047200501,-0.00688195,-0.5829085112,0.0044352496,0.0092913974,-0.2898313999,0.0828917325,-0.0154359462,-0.0030139687,0.2542344332,0.1097736657,0.1992102265,-0.5189850926,-0.4834819138,-0.1372635216,-0.3550201952,-0.1833998114,-0.1033037603,0.1189661399,0.1262574941,0.0301867966,-0.1101572961,0.1655335575,0.0401767269,0.1082341969,-0.2075418085,-0.414519906,-0.0661348253,-0.0340366773,0.2221994847,-0.0620478392,-0.278073281,-0.0760408267,0.3273139,0.1863703132,-0.0055327709,-0.0973921493,0.0755881518,0.287073195,-0.1300795525,0.0824393407,-0.1762814373,-0.0173754077,0.3110890687,0.0811619237,0.223300755,0.7291049361,-0.0453017019,-0.385738939,-0.2154037058,-0.0097820675,-0.1862677485,0.1873774976,-0.1925709844,-0.0187897701,-0.1713922769,-0.191484943,0.1445645541,-0.5781942606,0.0466269031,-0.513125658,0.3279879689,0.0050349208,0.0378013626,0.1020331904,0.2791719139,0.0113373967,-0.2690429688,-0.0266941506,0.5668866634,0.1636475325,0.0422566831,0.2496344298,0.1973968893,0.3936899304,-0.2030072957,0.1371941268,-0.2146210968,-0.029627271,-0.2047377825,-0.3810374439,0.2312919348,-0.1225019991,0.0286590066,-0.0267326795,0.0361669958,0.0095526502,-0.1421406567,-0.0170786064,0.0299908239,0.2098237723,0.1546527594,-0.2356938571,0.0730327219,0.011369139,0.3710994422,-0.3674442768,-0.1867903769,-0.2211580127,-0.3204190135,-0.227793023,0.1493093371,0.1138046309,-0.0039152526,0.4859624207,0.2993046045,0.1127215698,-0.0250022616,-0.0873093084,-0.1738382429,0.0267921332,0.070074752,-0.1132489443,-0.0825649649,-0.0779607892,0.0170043614,-0.16715464,0.0658125132,0.1123546734,0.1347985864,-0.4500104785,-0.3438916206,-0.2276449054,-0.1115065292,0.2226544172,0.1502758712,0.1307689697,-0.3118938208,0.1746344864,0.0323131718,-0.0515429527,0.0069793495,-0.4058719873,0.3129604757,0.0428381339,0.120900631,0.1041473448,-0.168645516,-0.1676724553,0.1991018653,0.0027728805,0.3461155891,0.3852150738,-0.1130608171,-0.0200322587,0.1037744284,-0.4112503231,0.0277434606,-0.1778269708,0.2392630875,-0.1379300803,0.0326035023,0.0625598505,-0.0049605598,0.2779373527,-0.0122228935,-0.2213647068,-0.0858392939,0.1937889159,-0.2835340202,-0.0695760548,-0.5575791001,-0.12401627,-0.0489480272,-0.0760456994,0.4957128167,0.215826273,0.0828830972,-0.1113867834,0.1919533312,0.2487667352,-0.2108538747,-0.3730554879,-0.8316481113,0.3392534554,-0.1333644986,-0.5556632876,0.0347582735,-0.0107243555,0.3196974397,0.1372258067,-0.4156912863,-0.3285232484,0.0028609077,-0.1332719177,0.4988700747,0.2466121465,0.1999249309,-0.1665162444,-0.1875765175,-0.2685646415,-0.0677228794,0.2646456957,0.278077364,0.4762164056,-0.3166675866,-0.3312654793,-0.2116470337,0.0864077434,0.2232727706,-0.1222820058,-0.0018558812,-0.0742551312,-0.0947363451,0.0872071162,-0.1291690916,0.6249518991,-0.1004929766,-0.0743596703,0.1873518527,0.2371968925,0.2381603718,-0.0801334977,0.0954903737,-0.2258431166,-0.141831696,-0.2125950754,-0.0409519374,0.0509542823,-0.0420044176,-0.0933640748,-0.1038029268,-0.096297279,-0.0796592385,-0.0060387277,0.1317088157,-0.0344006494,-1.0077539682,0.2426573038,0.0604868084,0.371894449,-0.3021514118,0.0058259312,0.0018673902,-0.1264990121,0.1128279567,0.2424506694,0.1748816222,-0.2255022377,-0.1925474554,0.3095665574,-0.0791294277,-0.3255848587,-0.06665124,-0.1542109698,0.0815795287,0.1985381842,0.0263148751,-0.2524892986,-0.010719683,0.1903713048,0.2273431569,-0.3153850436,-0.1681496203,-0.0732905641,-0.0716855153,0.0586164556,-0.1715297401,-0.1118609011,0.2604054213,-0.2799581587,-0.1209540442,-0.1740711182,-0.1244765148,0.5657262206,0.0522190072,0.4223523438,0.0191579722,0.3164678812,0.3983605206,0.0959929302,0.2176501304,0.4899378419,-0.1678944379,0.2445158064,0.1520571262,-0.3752425611,0.3867096603,0.1771408767,0.0790656358,-0.5246110559,-0.2063470185,0.0706287026,-0.1910707951,0.2839501798,0.0786079317,0.1656914651,-0.0985878929,-0.2857264578,0.5754490495,0.1364170164,-0.0153934471,0.3863967955,0.3857402205,-0.1207363904,-0.2186936885,0.5087192655,0.9099233747,0.0933356211,-0.1867701113,0.2689580023,0.3895844817,0.314519763,-0.4774867594,0.0561192222,-0.1552691609,-0.3308633566,0.0102706812,-0.1089504063,0.1223558262,0.2782101035,0.1521281749,0.0763726085,-0.1585591882,-0.2480708808,-0.1482344419,0.181555137,-0.2294644415,-0.1465154141,0.2912207842,0.2644282579,0.12810646,0.2056466639,-0.0387386717,0.0456349067,-0.3216534853,-0.0078412713,-0.0484310351,-0.0600882061,-0.0976320878,0.0689852089,-0.0059334016,-0.3436070383,-0.08534275,0.359436661,0.2132128775,-0.062549606,-0.1257444024,0.1786783487,-0.3669171929,0.0576272681,0.3287370205,0.0124526629,0.3648482859,-0.1980662197,-0.1608164012,-0.0313753486,0.094133094,0.228657797,-0.3836551607,-0.0294618346,-0.0412901528,-0.4523265958,-0.0695839822,0.1678125262,-0.0893606842,-0.09791594,0.1863251626,0.2292987555,-0.0718699843,0.449798733,-0.1016354859,0.037131615,-0.117601268,0.2099590898,0.429729104,-0.2049558461,0.2754957378,0.0360344909,0.0052345917,-0.1443431228,0.0589408539,0.4662175179,-0.0782095864,-0.0909994915,-0.2957358956,-0.1659356654,-0.0604274347,0.2391818166,0.2923488319,-0.0224096533,0.1084384173,-0.3646565676,-0.4771287739,0.2965849042,0.1080270335,0.4138523936,-0.2822362781,0.2306647599,-0.0184349027,0.0322790407,-0.3637257814,0.0349334441,-0.1426316053,0.0633205995,0.0953675434,-0.0285140388,0.257401377,-0.3777071536,0.1761278212,-0.0493734404,-0.1377490163,-0.2226376086,-0.1133773699,0.103344284,0.1638176739,-0.0885369033,0.2341868579,-0.0245163236,-0.0553335026,-0.0356042273,0.0422471873,-0.105882965,0.0537427329,0.3159060478,0.0220756754,0.1330342144,-0.0847093686,0.1604008228,0.1310178488,0.3074553013,0.1683982015,-0.282037437,0.0260938909,-0.2338073403,-0.2848701775,0.0179096684,-0.0505883321,0.1148030758,0.2884682119,-0.0608447716,0.1907160729,0.1914209872,0.333340168,0.3816255331,-0.186850518,0.1370875984,-0.0227796603,0.3095036745,-0.2883691192,-0.0066335006,0.0627779737,0.3301757872,-0.1142457128,0.0569347218,0.1551648825,-0.0585037656,-0.1676236391,0.165860787,0.2570085526,-0.0311393607,0.1332113296,0.2040168345,-0.225135982,0.0041279909,0.1211696193,-0.0084303021,-0.0823717862,0.2305449396,0.0010614166,0.488392055,0.2151825577,0.1520121098,0.0299616866,0.0104308948,-0.1124079004,0.3381588161,0.1962863803,0.1544807106,-0.009430266,0.5764229298,-0.2117131203,-0.0099261589,-0.294464767,-0.0149887651,0.1193002537,0.1711391807,-0.2637554705,-0.1399070472,0.0453714691,-0.2038899362,-0.1080833152,-0.2699944079,0.1528010964,-0.0210159384,-0.3585354388,-0.8289134502,-0.1219780445,0.1738721877,0.2557282746,-0.1027105153,0.1203282401,0.5782190561,-0.0316473991,0.0955784172,0.5063878894,0.1323528588,0.1674837172,-0.0139366658,0.0357911959,0.2206596881,-0.0467847399,0.1264574826,0.1159114242,-0.000790038,-0.1671576351,0.2911850214,0.2048635334,-0.0981252864,-0.1249330416,0.0552311726,0.0274031367,-0.0900457725,0.1057320312,0.0168692004,-0.1648440361,-0.0819955096,-0.2306069434,-0.3721615672,-0.0468022861,-0.106701538,0.1789139062,0.1843640208,0.029317895,0.1284211427,-0.3652576804,0.3222817481,0.0909566954,-0.1236387342,-0.3167787492,-0.1710978448,-0.622776866,0.001430015,-0.0624746121,-0.5641044974,0.182269156,0.3159278035,0.1453770548,0.0527520925,0.031178223,-0.4149900973,-0.1453268677,0.2601168454,-0.6003611088,0.0534222163,-0.2321398854,0.2024497092,-0.0308518149,0.3180320561,0.2247194499,0.2201413959,0.0092765186,0.1568650305,0.185688436,0.1527652144,0.2310855985,0.0140155908,-0.020797465,0.4964556992,0.1912759542,0.2829366922,-0.2954441309,-0.1949322969,-0.1533316374,0.2975698709,0.1618734896,0.0609212369,-0.0355313718,0.0185935888,0.0723980442,0.3197269738,0.0364174992,-0.4527270794,-0.4951818585,-0.0151288239,0.0105010988,0.0400301255,0.303955853,0.5021279454,-0.0397007726,0.070709914,-0.1943258941,-0.3799972534,0.4460830688,-0.1446646154,0.1760379672,-0.0411515199,0.2653198838,0.4457305074,0.0169606414,-0.3346783817,-0.0056860596,0.4271606803,-0.2768956125,-0.1563839465,0.2719531059,0.3581559956,-0.1532853991,-0.0111295078,-0.0489597619,-0.012674341,0.0277566165,0.1528594345,-0.0404255167]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"Thanks! That sounds like a reasonable number! ","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":7,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nThanks! That sounds like a reasonable number! ","embeddings":[-0.1167578548,-0.4001875222,-0.1571099013,0.4909679592,0.3018277287,0.013298342,0.163067922,-0.0296909232,0.2293610275,0.2793543041,-0.219976902,0.2242952138,0.0920818001,0.425362438,0.118394509,0.1186279356,-0.0246461015,-0.1124243364,-0.3718460798,-0.2964376807,-0.0102031454,0.0654809102,-0.0789518282,-0.1322781444,-0.6428670883,-0.0013896989,-0.1362710595,0.2756492794,-0.1101892963,-0.1952979416,0.2403063327,0.0274113193,0.026389366,0.5462388396,-0.0001048368,0.0079390034,0.1067845747,-0.0254856385,-0.0968101621,0.260483712,-0.1833849698,-0.0307041891,-0.0009756866,-0.3822710216,-0.0826701522,0.0305082388,-0.1292806864,-0.3704583049,0.210828498,0.0230599567,0.2252570391,0.1732560247,0.1191313565,-0.0563634001,-0.0619362555,-0.1114320159,0.0909814611,0.23005943,0.3792605698,0.268030405,0.055041302,0.1818307936,0.0943560302,-0.2004666179,0.2908363044,0.0586025119,0.3298102617,-0.2149673402,-0.2978039384,0.2906467915,0.846260488,-0.030787224,-0.2686520517,-0.0306402296,-0.0212893244,-0.116286777,0.1561630219,0.0627702177,0.3555042446,-0.0477295071,-0.2468536198,0.4402652979,-0.226155892,0.0323073231,-0.1501324773,-0.0886049643,-0.062579982,0.01011765,-0.2362840176,-0.0510150418,0.0395439006,-0.0557919256,0.1713529527,0.1960915923,-0.3144487441,-0.1210187078,-0.0071856519,-0.0066858823,0.2326755673,0.1547258645,0.284388572,0.0368980505,-0.3853200376,0.0440469086,0.2232537866,0.2619153559,0.4053671658,-0.542455852,0.0013042449,-0.1044105142,-0.1109073013,0.0888874307,-0.3676219583,-0.0336851701,0.0091276756,-0.0239210594,-0.1455921084,-0.358456552,-0.232501775,0.2630746663,0.043420583,-0.0868002251,0.1700859666,-0.1166303232,0.0645963922,0.3793299198,0.0107838195,0.0437036119,-0.5489529371,-0.29221946,-0.3659675419,0.1173683852,-0.0682866126,-0.049507305,0.0918191597,-0.084259361,0.3836414516,-0.0886952057,0.2683841586,-0.0054676891,0.0769955218,0.098635301,-0.3332027495,0.2502072155,0.0531304032,0.000992257,-0.0326997563,0.0259366352,-0.0316033848,-0.026419213,-0.177808553,-0.2069142312,-0.0038863667,0.3261167705,-0.346576333,-0.273378402,-0.1719036996,0.0104818149,0.2030287385,0.1909862161,-0.1857059598,-0.1430527568,0.0689049512,-0.1321503073,0.1654997021,-0.0317264348,-0.5920577645,0.0319517776,0.0010234282,-0.2804661393,0.0692665577,0.0082712946,-0.0113249803,0.2472577244,0.1227510199,0.2057515681,-0.5087666512,-0.459145844,-0.1515549123,-0.3398270607,-0.1702478081,-0.1046170667,0.1429920346,0.1181333959,0.0195717849,-0.118281804,0.1685679853,0.0090522431,0.1221875623,-0.1737662256,-0.414943397,-0.0527438633,-0.0158749372,0.179792732,-0.0738685578,-0.3038567901,-0.0656073689,0.3422565162,0.1683950573,-0.0126107428,-0.0951703861,0.0540763512,0.2909930944,-0.1476549357,0.0794594213,-0.1941950768,0.0117893815,0.2972472012,0.0848153681,0.2337102592,0.7663412094,-0.0594773442,-0.3712383211,-0.2146568596,0.0006984443,-0.1874502599,0.2085249722,-0.2039660215,-0.014380794,-0.2161973566,-0.176518783,0.1511866748,-0.5576575398,0.0449179746,-0.5135161281,0.3054336011,0.0061251633,0.0216020141,0.0741378739,0.2626399696,-0.0369377919,-0.255161196,-0.0247753765,0.5458697081,0.1655876189,0.0235725772,0.2062017471,0.1933157444,0.3725995421,-0.164852649,0.1693312526,-0.244994849,-0.0462410226,-0.199475497,-0.3847044706,0.2421635389,-0.1056579053,0.0044654496,-0.0581552945,0.0273528844,0.0247651432,-0.1216316447,-0.0045091137,0.0273328871,0.1975600719,0.1597202718,-0.2162243277,0.0699165836,0.0156712625,0.3822619021,-0.3547385633,-0.1575454026,-0.1972106546,-0.3429452777,-0.2131985873,0.144252196,0.1099446937,-0.0173711255,0.5087643266,0.3122445643,0.0941863582,-0.0427583195,-0.1050608233,-0.1540257782,0.0064834291,0.0082663503,-0.0926611647,-0.048492901,-0.0438462384,-0.0016202327,-0.1926585883,0.0865472928,0.0965841115,0.1556443572,-0.4353733063,-0.3208337724,-0.1820989996,-0.092274949,0.1990901828,0.1944504678,0.127909556,-0.2799808681,0.1811150759,0.0528462119,-0.0400772728,0.0030225539,-0.3754591644,0.2995740771,0.042012047,0.1309506297,0.1244572699,-0.148749277,-0.171540156,0.2150459886,-0.0004270148,0.3419852555,0.3903032243,-0.0834798887,-0.0520527735,0.1247759983,-0.4192527831,0.0264567006,-0.1982966661,0.2562554777,-0.1238085628,0.0252714455,0.0750488564,-0.0109322285,0.2550654709,-0.0205222927,-0.2533429861,-0.0779409483,0.179470554,-0.2616745234,-0.0861553028,-0.5513278842,-0.1161064953,-0.0730515197,-0.0500715338,0.4884958267,0.1923639327,0.0913800746,-0.0933103114,0.1820322573,0.2397394776,-0.1836738437,-0.395264864,-0.7948393822,0.3041617274,-0.1171998978,-0.5338881016,0.026934443,-0.013064825,0.3292294443,0.1235873103,-0.3862064481,-0.3546859622,-0.0108040133,-0.1215000525,0.4833233058,0.2306261063,0.2224827111,-0.1804314703,-0.1956697255,-0.2644916177,-0.0566357896,0.2744171619,0.2775076032,0.4465969503,-0.3462668657,-0.3104788959,-0.2078265995,0.0511083156,0.233960256,-0.1498834193,0.0072134696,-0.0908195227,-0.0832794681,0.0434856787,-0.1303593218,0.6441615224,-0.1133824363,-0.0957322866,0.175354287,0.2157751322,0.2123878002,-0.0392580628,0.0730665401,-0.186233148,-0.1297777891,-0.1901371032,-0.0386423618,0.035843838,-0.0321568735,-0.0708934218,-0.1092753485,-0.1004438773,-0.1170008034,-0.0235792249,0.0930491164,-0.0286203437,-0.9927644134,0.2834077775,0.0438489392,0.3744670153,-0.3069837689,-0.0244734716,0.0068758726,-0.1291987151,0.1007369235,0.216666162,0.1885490119,-0.2201446891,-0.2002765536,0.2910711467,-0.0728572756,-0.300691992,-0.0757034048,-0.1526917666,0.0635092631,0.1998827457,0.0485054068,-0.267373085,-0.0117929187,0.2010687441,0.2163881212,-0.2986862659,-0.1580869406,-0.1005786061,-0.0461511724,0.0424675383,-0.1135312021,-0.1088564247,0.262062192,-0.2509604096,-0.1214638874,-0.1932528168,-0.1253427714,0.5454359055,0.0842804387,0.4125460386,0.0136615317,0.3276793361,0.348559916,0.1193937212,0.2117035687,0.5151635408,-0.1582102776,0.2467729598,0.1528934836,-0.3703597486,0.3883920908,0.1600954533,0.0839541331,-0.5188049674,-0.2274840623,0.0976376757,-0.1875761598,0.28106305,0.1153960377,0.1807526499,-0.075808607,-0.273689419,0.5317975283,0.13478145,-0.0028038512,0.3887014687,0.3704864085,-0.1336335838,-0.1923530549,0.5106663108,0.8990369439,0.0852601007,-0.1712842733,0.2725879252,0.397557795,0.3095962703,-0.4836807549,0.036449302,-0.1590983421,-0.3576008379,0.0437212549,-0.0892145708,0.1362828612,0.2720804513,0.1255271435,0.0590533093,-0.1530911028,-0.2679172754,-0.1915449649,0.1898313612,-0.2505795956,-0.1456047446,0.2827417552,0.2813921571,0.1055807397,0.1667972952,-0.0216614511,0.0538210198,-0.303925842,-0.0311253741,-0.0676174685,-0.0700528771,-0.0736553222,0.0463301055,0.0251998398,-0.3351991177,-0.0501870364,0.3363511264,0.2094851583,-0.0566519573,-0.100293234,0.156329453,-0.3482697904,0.0693837553,0.3146952391,-0.0027864452,0.3909422457,-0.1872898489,-0.1653948873,-0.0453364663,0.0984301865,0.2270621657,-0.3649527729,-0.0224130228,-0.0260261092,-0.5039666295,-0.0580411702,0.1854397207,-0.0847444236,-0.0845553428,0.2045910507,0.233733356,-0.0858128592,0.4452918768,-0.0976786762,0.0355004892,-0.1041898131,0.187541455,0.4382680357,-0.1820160151,0.2492120862,0.0063072038,0.0026788628,-0.1731395274,0.0372463726,0.49716416,-0.0897324458,-0.0653891414,-0.2602000237,-0.2154193074,-0.0321359187,0.245149374,0.2936851382,-0.0091089671,0.0951631442,-0.3545927107,-0.4549278021,0.2874722779,0.1043622866,0.4339017272,-0.269402355,0.2468311936,-0.0189342108,0.0732780918,-0.3803097904,-0.0051692575,-0.1200603992,0.0700716078,0.0809973851,-0.0699600726,0.2496467382,-0.3698225319,0.1934288591,-0.0115727549,-0.176400438,-0.2484067529,-0.1109169424,0.0881371722,0.1474966705,-0.0854537785,0.2627860904,-0.0250546969,-0.0720464066,-0.0719164684,0.0614328235,-0.1105754152,0.038131807,0.3096704185,0.0328780748,0.1332888305,-0.0532880016,0.1303756833,0.1261075884,0.242901206,0.1528848261,-0.2583137751,0.0404695496,-0.2405670136,-0.2951574028,0.0375955738,-0.0504559241,0.1465235502,0.2464679182,-0.0606100149,0.162951529,0.1662142575,0.3631498218,0.3863627315,-0.1512149423,0.1257970929,-0.0312004834,0.3302267194,-0.3243279457,-0.0202362668,0.0537020005,0.3276393414,-0.0782749876,0.0847102329,0.1624625474,-0.0260774661,-0.1673744768,0.1638192087,0.2182337642,-0.0851941332,0.1157339811,0.2151357979,-0.2492312342,-0.0128525831,0.1511539668,-0.0423803627,-0.0497897305,0.2698439658,0.0019788891,0.4952317476,0.187650159,0.1526925415,0.023977587,0.0282913633,-0.0691978037,0.348556459,0.1677264124,0.1530773938,0.0067916089,0.5699669123,-0.2129735202,-0.0128126321,-0.3186050355,-0.0199443735,0.0871669576,0.1523126066,-0.2450139523,-0.1156350821,0.0283762347,-0.1901276708,-0.1219640672,-0.2420213372,0.1666982919,-0.040146105,-0.3384954333,-0.7947514653,-0.1040157005,0.1587586403,0.2609121203,-0.1119638085,0.1587485224,0.5575824976,-0.0338778012,0.0941806734,0.5214620233,0.1418151408,0.1646910161,0.0142595489,0.0594945401,0.210086599,-0.0418510325,0.0721568242,0.0930826813,-0.0110545447,-0.143513456,0.2899384499,0.21162799,-0.1224962547,-0.1329342723,0.0507921241,0.0029789368,-0.1045157015,0.0848342255,0.030338645,-0.1579922587,-0.0689306259,-0.2018348724,-0.3602647483,-0.0174138844,-0.107451193,0.1938542873,0.1938017011,0.0137112364,0.1378953457,-0.3443205059,0.3184050918,0.0334680416,-0.1200440228,-0.2853075564,-0.128584221,-0.6340577602,0.0287928041,-0.0901555568,-0.5695977211,0.1581997871,0.3445473015,0.1636137217,0.0567654893,0.0524599217,-0.4342028201,-0.1503813416,0.2699612081,-0.5878266096,0.0524052195,-0.2097049803,0.1733077615,0.0048277127,0.2872542739,0.211448878,0.2206296176,0.0283691455,0.1516091824,0.1730168164,0.1553550363,0.2253242582,-0.0068826908,-0.0263559036,0.4970830083,0.1835382879,0.293407917,-0.3201337159,-0.1837273538,-0.2189203799,0.2920490503,0.2131930441,0.113897264,-0.0332902521,0.0113716247,0.0973601565,0.3136068285,0.0587047711,-0.4447043538,-0.4806538224,-0.0028026733,0.0574308746,0.0261105262,0.2621491551,0.5163059235,-0.0178603493,0.0546479933,-0.2004900426,-0.4012658298,0.4203479886,-0.1406367719,0.1721735001,-0.0506800972,0.2680188715,0.4456558228,-0.0457123779,-0.3527114391,0.0023798437,0.4105445445,-0.2611356378,-0.1351694018,0.2818577588,0.3704001606,-0.1813572496,-0.0144857122,-0.0505706482,-0.0281199496,0.035483405,0.1653698981,-0.0595856756]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"So I managed to download them all but when parsing only 226,181\/226,711 worked.\r\nNot sure if it's worth digging and debugging parsing at this point :\/ ","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":26,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nSo I managed to download them all but when parsing only 226,181\/226,711 worked.\r\nNot sure if it's worth digging and debugging parsing at this point :\/ ","embeddings":[-0.0844028369,-0.3709565699,-0.1518211812,0.5515391827,0.3594364822,0.040523842,0.0683168098,0.0596995018,0.1347435266,0.2831323445,-0.2639553547,0.2220126092,0.1665135622,0.3795992136,-0.0097541586,0.0870642513,-0.0458379053,-0.049531322,-0.3420444429,-0.2743534446,-0.1129426956,0.1672509909,-0.064018324,-0.093108803,-0.61611408,0.0304498505,-0.1506063342,0.2275840342,-0.1146603897,-0.1896904856,0.2454412878,-0.0064762058,0.0730718002,0.5845306516,-0.0001072108,0.0421477407,0.1385874599,-0.0378101729,-0.101608105,0.1958065629,-0.1344313771,0.0074811657,0.0375896581,-0.4290549159,-0.103844434,0.0078649484,-0.1288392097,-0.4139157236,0.2797066569,0.0508435443,0.2060117722,0.130140841,0.1220481396,-0.0610720254,-0.041498825,-0.1653328687,0.1135953739,0.1681452245,0.4286612868,0.3330434859,0.1132159159,0.1378391832,0.0627353042,-0.0928761587,0.2489815354,0.0891236067,0.3963843286,-0.2542513609,-0.2236058712,0.3075287342,0.7909542322,-0.0128415506,-0.2719562948,-0.0801350027,-0.1137403622,-0.1378805339,0.1684876829,0.0955293179,0.3050046265,0.0173783489,-0.2742906511,0.4346568584,-0.2719221711,0.0431274623,-0.1886090636,-0.019397527,-0.0990475565,0.0135308318,-0.2441749126,-0.0699344426,0.0598460473,-0.1378884912,0.1849856824,0.1969777942,-0.2687837183,-0.086529322,0.0225008763,-0.0889225975,0.2075470537,0.1495759934,0.2924564183,0.1740622371,-0.388445437,0.0360541567,0.2837073803,0.1982509941,0.4640651047,-0.5234386921,-0.0598212816,-0.094658576,-0.0922634676,0.089162305,-0.3523961306,0.0325111672,-0.174333334,0.0682125837,-0.1539725959,-0.3955872953,-0.2191723436,0.2432529926,0.0437817648,-0.0528692268,0.1824024916,-0.0723197684,0.0794285312,0.4509501755,0.0449281484,0.0883205011,-0.5913876891,-0.3623243272,-0.3410969079,0.076807633,-0.0388297327,-0.1517431885,0.0976497754,-0.0679078251,0.3594029248,-0.1681587547,0.2341181934,-0.0889768749,0.0521356091,0.0420723148,-0.2788230479,0.1807665974,0.1091001108,0.0562919229,-0.0498099849,0.0215040538,-0.0359664336,0.0135747548,-0.2149073929,-0.2208453268,0.0673824251,0.3082099557,-0.4003606141,-0.1919499785,-0.2521123588,0.0484697744,0.1929271817,0.1856541634,-0.2762892544,-0.1987271458,0.0695115179,-0.1438952535,0.2400277853,-0.0621405207,-0.6286695004,0.0207068976,-0.0551142246,-0.3237284422,0.1594667137,0.0425659791,0.0001080091,0.1796017587,0.0852619633,0.3123659194,-0.5448957086,-0.440623492,-0.1038651839,-0.3021672964,-0.1983390003,-0.0888617858,0.0656147674,0.05777191,0.1162542179,-0.1171016917,0.1266627312,0.090306364,0.1143721789,-0.1681381464,-0.3830769658,-0.0592863485,-0.0361176282,0.2105553895,-0.0382765308,-0.3569951653,-0.0710630268,0.3697214425,0.2348645627,-0.0128279943,-0.0519180447,0.0767108947,0.2566022277,-0.1620862782,0.1281024367,-0.2024354786,-0.0674671382,0.3197570443,0.148252219,0.1308018714,0.6472723484,0.0174675658,-0.3737598658,-0.2013436705,-0.0036841901,-0.156388551,0.179809466,-0.1540352106,0.016963657,-0.1929002553,-0.1964883357,0.1467991769,-0.5844185948,0.0877542347,-0.4837872684,0.307949394,0.035646379,-0.0554785617,0.1046856493,0.232011199,0.0160835478,-0.3351312578,-0.0028699031,0.5101035237,0.219106093,0.0154445935,0.0722845867,0.0785818174,0.36207515,-0.1975265145,0.0968506336,-0.2044920027,-0.0145210037,-0.2000112683,-0.4076152742,0.2605000734,-0.1173585057,0.0784933865,-0.02649693,0.0118698301,0.0656035244,-0.1671250612,0.0782168284,-0.0068109394,0.2217555791,0.1363307089,-0.2356708795,0.029454425,-0.0234066956,0.4073214829,-0.3067409694,-0.1834520996,-0.2272915393,-0.3764779568,-0.2248347551,0.2087489665,0.089134939,0.006139819,0.3895544708,0.2478598803,0.1647430658,-0.0026287346,-0.0421475358,-0.1760358959,0.0573837459,0.1193651929,-0.0931036249,-0.0215958096,-0.0650559962,0.0102848252,-0.1624887586,0.0591240749,0.149917677,0.1706698984,-0.4749034941,-0.3363263309,-0.2788443565,-0.1544815302,0.2290477604,0.2144044787,0.1654610783,-0.3635771871,0.1188497543,0.0655469149,-0.09844964,-0.0696898848,-0.4235265255,0.2534594536,0.0416911952,0.0493309051,0.0934920013,-0.1450417638,-0.2133835852,0.2005327642,0.1096160188,0.2957605124,0.2793506682,-0.0827667862,-0.0675938949,0.1530393213,-0.4200555682,0.0808792785,-0.2090593427,0.264739871,-0.1809299141,0.0637805089,0.0758882314,-0.0583418794,0.2622323036,-0.0175454151,-0.3150865138,-0.0072279945,0.2789834142,-0.2849284112,-0.1204456612,-0.620398581,-0.0950751305,-0.1197582334,0.0135916509,0.4494410157,0.2404864132,0.0030261728,-0.123798579,0.1671096385,0.1330026537,-0.198048383,-0.3257893026,-0.7631091475,0.3649157286,-0.1054657325,-0.5235021114,0.0694446117,0.0365557112,0.331969589,0.2015591711,-0.3550731838,-0.263184011,0.0283332523,-0.138043493,0.5030138493,0.1878084689,0.1311599612,-0.1825519353,-0.1283177286,-0.2406112701,-0.0633855909,0.2212834358,0.2867417037,0.4415383339,-0.2504809797,-0.2394237369,-0.1297158152,0.106730327,0.2224221677,-0.1233519316,0.0197561644,-0.0243240446,-0.1388284415,0.0865022615,-0.0701704472,0.6146479845,-0.100210458,-0.076683253,0.2349850237,0.201297909,0.1896943897,-0.1080243215,0.0543919913,-0.3126692176,-0.1980424821,-0.2103840262,0.02073914,-0.0154817728,-0.0241566692,0.0033168735,-0.0880531222,-0.0707055852,-0.1224520579,-0.094479993,0.0862987414,0.0019934974,-0.9377420545,0.2583681047,0.0919099152,0.4281425476,-0.3496419489,0.027894374,0.0480786376,-0.1031292379,0.162587449,0.1365582198,0.1698757559,-0.2361756563,-0.1420076191,0.3406646252,-0.1795443892,-0.325953126,-0.0398755446,-0.2304544896,0.0668514445,0.2693084478,0.0082560731,-0.3681924641,-0.0841492862,0.1560114622,0.1928171217,-0.2522937655,-0.093455255,0.0209813304,-0.0953884423,0.0589124411,-0.1013499051,-0.1001864523,0.2969883978,-0.201406911,-0.109172903,-0.2192253768,-0.1571836472,0.6077881455,0.0983486697,0.4593627155,-0.0179452784,0.348497957,0.3587201536,0.0554524139,0.3094827831,0.5613517165,-0.0473900773,0.140685156,0.2311219722,-0.4103594422,0.3843428195,0.2076773643,0.050920397,-0.5400857329,-0.2392208576,0.123028554,-0.123043187,0.3074818552,0.0821115226,0.1137974784,-0.1260647625,-0.2406028807,0.5399982333,0.2011412531,-0.0039305678,0.3453181088,0.363301754,-0.1422877014,-0.1741364002,0.4850198925,0.8846732974,0.0581081919,-0.2450192869,0.233622998,0.3626748919,0.3922798932,-0.5971927047,0.0725345388,-0.1595449448,-0.2600105107,0.0047875619,-0.1000371352,0.1541109234,0.3621688485,0.2189938873,0.1191185862,-0.1705763638,-0.2454868853,-0.0999174044,0.1773552448,-0.313826412,-0.1143446192,0.2780292034,0.2623187304,0.0470491499,0.1821988523,0.0251904987,-0.0102490922,-0.3843561411,-0.0390356071,-0.0858823061,0.0581313707,-0.1295070499,0.0087874271,0.0130772488,-0.3557404578,-0.0024173253,0.3001090288,0.1371404231,-0.0598600805,-0.138663128,0.1516336203,-0.3667663932,0.0080353161,0.3414291143,-0.054977309,0.3626557589,-0.1882983595,-0.0685659051,-0.0483330265,0.0838891342,0.2143613398,-0.3564110398,0.0453653522,-0.0632297099,-0.4032186866,-0.1070721895,0.1976603866,-0.0609891005,-0.071148783,0.1763242632,0.2075746506,-0.1406908035,0.4731565118,-0.1147413701,-0.0000974808,-0.0891012326,0.1888044477,0.4392287731,-0.234362945,0.3456516862,0.1064261049,-0.0194111317,-0.1558786333,-0.0648856536,0.4879930913,-0.0728849024,-0.0533645377,-0.3302249908,-0.1810845584,0.0378356874,0.2344367802,0.291099757,-0.0568905659,0.0859243572,-0.3747248352,-0.3915478587,0.3048351109,0.0615768507,0.3998572826,-0.2707557082,0.3009163439,-0.0394116566,0.013466794,-0.3510236442,0.0814448968,-0.1134228632,0.0863941982,0.0758703351,-0.0173688475,0.1995367408,-0.376380533,0.1671513319,0.0994853005,-0.1877503097,-0.2005375326,-0.1248668805,0.107813172,0.1627760977,-0.09144786,0.231532678,-0.0609393381,-0.0687168539,-0.0724650994,0.0416360572,-0.0996675789,0.1037641019,0.270760417,0.1801616251,0.1299978942,-0.0851482302,0.1632080376,0.1201812401,0.3497532904,0.1439631879,-0.2291558385,0.0017922587,-0.1992864311,-0.2425180376,-0.0323252566,-0.0972900912,0.0550502427,0.3261317611,-0.0826273113,0.1812698692,0.2118543237,0.3575928807,0.5210460424,-0.1173308119,0.0779539496,-0.0002382825,0.296744138,-0.2638825476,-0.0054880148,0.1626421213,0.433052808,-0.1679995209,0.1031333283,0.1227403954,-0.0439868346,-0.1741834283,0.1689157635,0.3321439922,-0.0323141441,0.1798407435,0.23881194,-0.1941303462,-0.0192902423,0.0813130438,0.0125190895,-0.0040598628,0.2605923414,-0.0456901528,0.4583982527,0.2378911376,0.1593589336,-0.0153165655,-0.0878006071,-0.139555648,0.3989134729,0.1636969298,0.1770862937,-0.0047665755,0.5884441733,-0.1515662074,-0.0095632067,-0.3527895808,0.1206694841,0.0668219477,0.1578653753,-0.2908957899,-0.1604319066,-0.0153111452,-0.1504224986,-0.1391030699,-0.2644779682,0.1907551587,-0.0136043578,-0.3546863496,-0.8215947151,-0.1265826225,0.1065370366,0.2467359155,-0.1747162193,0.1872714013,0.5541864038,-0.073822692,0.1061168164,0.4661080837,0.1588304043,0.1761285365,-0.0514695197,-0.0143454773,0.1804935485,-0.0300070178,0.0893063322,0.1932425648,-0.0220597703,-0.1006876081,0.3865216374,0.2039574981,-0.1132391691,-0.1893343329,0.0490954518,0.053751301,-0.0886314884,0.0892836154,-0.0158442762,-0.0779377148,-0.0738952011,-0.3469389677,-0.3450101018,-0.0823853686,-0.1266109347,0.208872512,0.1970622092,0.0725054219,0.1151135787,-0.3228029609,0.21750696,0.1203015521,-0.1338187754,-0.3367739618,-0.164731741,-0.5921193361,0.0133012943,-0.1393102407,-0.6057727337,0.1874084324,0.3719215691,0.2124894708,0.0715408102,0.0168451965,-0.3277927637,-0.1417146623,0.3222787976,-0.5912342072,0.0750558004,-0.2451960742,0.1103659645,0.0302405246,0.2611947358,0.2501969039,0.2228233069,-0.0242264587,0.1672596931,0.1731894016,0.1581403017,0.152683273,0.0076706093,-0.0108488267,0.4436746836,0.2360303104,0.3360919356,-0.3565450907,-0.1521240175,-0.1615939438,0.3721128404,0.1744452864,0.1416357607,-0.0451191589,0.057692565,0.0519835018,0.1859652102,-0.04373594,-0.3975057006,-0.4537165761,0.0076166894,-0.0223280489,0.073916465,0.2933510244,0.4988903403,-0.0320864581,0.0997441933,-0.2534347773,-0.4610159695,0.3716871142,-0.1335282624,0.0866397619,-0.0287303403,0.2500470281,0.4638034403,0.0116081676,-0.3417442441,-0.0264737662,0.4352745712,-0.3208561242,-0.1928695887,0.1822361052,0.3333805799,-0.25839293,-0.0364943556,-0.0554951243,0.0673059449,-0.0355634838,0.1169176996,-0.0384216458]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"Maybe @sshleifer can help, I think he's already played with xsum at one point","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":14,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nMaybe @sshleifer can help, I think he's already played with xsum at one point","embeddings":[-0.0879578963,-0.396235764,-0.1512721032,0.4915593863,0.3209068775,0.0120451516,0.1624783427,-0.0427685864,0.2166582197,0.2898221612,-0.2147962004,0.2517166436,0.103626281,0.4202885926,0.1204897091,0.125404194,-0.0164337549,-0.1060302332,-0.3534614444,-0.2994217277,-0.0101279486,0.0283189267,-0.0802705288,-0.1221043691,-0.6537907124,0.0015694995,-0.1277013719,0.2662664652,-0.0881038308,-0.2012134939,0.2491302937,0.0376654342,0.0282536857,0.556127429,-0.0001051683,-0.0071448125,0.1157903969,-0.0194812324,-0.0969398394,0.2441393137,-0.1864143163,-0.0163197871,0.0023606177,-0.4071934223,-0.0652088448,0.0490176566,-0.1269536912,-0.380408287,0.1947459728,-0.0033409526,0.2228730768,0.1702204943,0.1146195605,-0.0590576902,-0.0621842556,-0.1268646568,0.1010313109,0.2408597618,0.4033352435,0.2760911286,0.0722918212,0.1609850824,0.0703886077,-0.1998636723,0.2949674726,0.0657098591,0.3629882932,-0.193588689,-0.3117902875,0.271083504,0.8513344526,-0.0556539744,-0.2659498155,-0.0196695961,-0.0056978106,-0.1185164675,0.1655905694,0.0524143875,0.3655984402,-0.0346242227,-0.2396816462,0.4139131308,-0.2364279777,0.0443150662,-0.1244149804,-0.0856129602,-0.060050074,0.0134541215,-0.2395402193,-0.0409811623,0.0794396698,-0.0769191831,0.1826951653,0.1939542145,-0.2870345116,-0.1179374084,-0.0096426765,-0.0110937683,0.2150541693,0.1620862484,0.2834548652,0.0533788241,-0.3910829723,0.0552248731,0.2172700167,0.2657987177,0.4391618669,-0.5600765944,-0.0029499056,-0.106554307,-0.1215914711,0.0781917647,-0.3692628741,-0.0187186599,0.036253307,0.0168236978,-0.1521459669,-0.3608098626,-0.2249288261,0.2604300976,0.04552285,-0.0934022218,0.1966024637,-0.1213601083,0.0612661317,0.3942882419,0.0164197981,0.0329819284,-0.5537250638,-0.2839861512,-0.3537358344,0.1003869772,-0.0691327974,-0.0366998613,0.0944465771,-0.0812312588,0.3652492762,-0.0868920684,0.2678923011,-0.0066959984,0.0629577637,0.0928169116,-0.3408721387,0.2151789516,0.0383110009,0.0063792029,-0.0426913798,0.0111013418,-0.027415704,-0.0179207586,-0.1872780621,-0.1851917654,0.0151613439,0.325517863,-0.3460855782,-0.2690333426,-0.161677748,-0.0087987445,0.1791944206,0.1880225986,-0.1928846687,-0.133011505,0.0870937631,-0.1345073432,0.1623156816,-0.0509063303,-0.6085608602,0.0266496222,-0.0046081282,-0.2839947641,0.0577487685,-0.026933806,0.0000773518,0.2332571596,0.1214223728,0.1972533762,-0.5442844629,-0.4912055433,-0.1400835216,-0.3617620766,-0.1922598034,-0.0811311901,0.1312484443,0.1334284544,0.0044523734,-0.136386618,0.1758005172,0.0197602436,0.1149768159,-0.1973262131,-0.4245883524,-0.0459229946,-0.0254693106,0.1801759303,-0.0914816186,-0.2810633183,-0.0774027184,0.3514750004,0.1620838642,0.0037065318,-0.1059166938,0.0487542897,0.3123574555,-0.1451956779,0.0596055277,-0.2097690701,0.0057595568,0.2924492955,0.0920229703,0.2207249701,0.7576835155,-0.0583194233,-0.3749126792,-0.2107856423,-0.0011269161,-0.1633185446,0.2083177567,-0.197532177,-0.0243700705,-0.2116959393,-0.1774238795,0.1484070122,-0.5750912428,0.025980968,-0.4988195896,0.3014942408,0.011675355,0.0185693987,0.0415678658,0.2654131353,-0.0239366777,-0.2514491677,-0.0292652212,0.5703068376,0.1532436013,0.0237853173,0.2056856155,0.1927263886,0.3658851981,-0.1835869253,0.1456028074,-0.2482055873,-0.038825389,-0.1990296245,-0.3670314252,0.2275106311,-0.118374981,0.0121715292,-0.0710635707,0.0193024613,-0.0006857692,-0.1493818909,0.0113451406,0.0229897313,0.2028848529,0.1569024622,-0.2308676541,0.0646063983,0.0234613325,0.3836091757,-0.3727560639,-0.1598208249,-0.1802128702,-0.3582154214,-0.2148168981,0.1773784012,0.0991191193,-0.024069868,0.5084937215,0.3088600039,0.0807201341,-0.0417384729,-0.0892681777,-0.157313183,0.0142729785,0.0309174135,-0.1035773605,-0.0674735755,-0.0520167761,0.0037670285,-0.1331317276,0.1071076766,0.0987349525,0.1275331825,-0.4448856413,-0.3217203021,-0.2153685391,-0.1003635302,0.2128058821,0.1797181964,0.1537620276,-0.2801796496,0.16531156,0.057784006,-0.0484678149,0.0007240544,-0.3688282669,0.2979203165,0.0336221121,0.1137390286,0.0958118886,-0.1672858894,-0.1677383631,0.2043028474,0.0174683668,0.3386851251,0.3923749328,-0.0690915883,-0.0475043952,0.1368810833,-0.4199871123,0.0024094528,-0.1821399778,0.2876983285,-0.1547488868,0.0275087636,0.0741576254,-0.0217308179,0.2792061269,-0.024725033,-0.2462483943,-0.078604646,0.2014632076,-0.2688926756,-0.0860637203,-0.5443139672,-0.1212021485,-0.0790679976,-0.0547844879,0.4834779501,0.1991639882,0.0833586901,-0.1001398936,0.1911192238,0.251350224,-0.1760679185,-0.3761127591,-0.8053275943,0.3044070899,-0.1095170006,-0.5191166401,0.0300406236,-0.0273934659,0.3158865869,0.1306952685,-0.3709679246,-0.3450851738,0.0210608002,-0.1151625589,0.5133089423,0.22258991,0.1972033978,-0.1594251245,-0.1902523935,-0.2668702006,-0.0549484082,0.2752634287,0.2630960941,0.4530967772,-0.3220618367,-0.3342580199,-0.2175059915,0.0309932306,0.1877263188,-0.1511951983,0.0063681789,-0.0567324683,-0.102419965,0.0687124729,-0.1574212015,0.6126860976,-0.1005778536,-0.0859250948,0.150268957,0.2107169777,0.2229471654,-0.0477593876,0.0789719224,-0.1929571033,-0.134261921,-0.2051027268,-0.0543191433,0.0140580721,-0.0478526577,-0.0873532444,-0.1153435633,-0.090327166,-0.1104431748,-0.029831836,0.093039386,-0.0506407842,-1.0036779642,0.2833550274,0.0810382888,0.3722694218,-0.298047483,-0.0142641366,0.0122691048,-0.129674837,0.1106400117,0.2237814218,0.1755075455,-0.2141748965,-0.2116676271,0.3101729751,-0.0897351503,-0.3016416132,-0.0709492266,-0.146525532,0.0616944395,0.1914943904,0.038481541,-0.2584944665,-0.017409103,0.1863176972,0.2164259106,-0.300124526,-0.1658331156,-0.0737743452,-0.0248043146,0.0529182516,-0.1421644986,-0.1358912289,0.2707303762,-0.2613867223,-0.1138783395,-0.2111790031,-0.1083305478,0.5695816278,0.0409032218,0.411825031,-0.0140208369,0.3185135722,0.3607706428,0.1090598926,0.2171308547,0.4593220353,-0.15282543,0.2595822215,0.1420412064,-0.3771121204,0.3924203217,0.1489531547,0.0912804231,-0.4859552681,-0.2531524003,0.0776302516,-0.2056192607,0.2799872458,0.0921995938,0.194611758,-0.0778493434,-0.2772734761,0.5553429127,0.1270525753,-0.0143861929,0.4020263255,0.358953774,-0.141095221,-0.2001344413,0.5123339295,0.8936166167,0.0840885788,-0.1871736348,0.2681292892,0.3989028335,0.313541919,-0.4994595647,0.0347513892,-0.1576270014,-0.3551149666,0.0458439775,-0.0986993983,0.1329078376,0.2684042454,0.1199142039,0.0446141511,-0.1939385831,-0.2926597893,-0.1814146489,0.1697808653,-0.2594244778,-0.1431226581,0.3027238846,0.2797034979,0.1132410467,0.1533476561,-0.0178567357,0.0480956212,-0.3187372088,-0.0028283589,-0.055689726,-0.050293304,-0.0758076832,0.0480574109,-0.0008791484,-0.3251882195,-0.048046831,0.3322744071,0.2167807072,-0.0534215979,-0.0846696496,0.160999909,-0.3442172408,0.0681178495,0.3242946565,-0.0060112639,0.4086794257,-0.1703080386,-0.1804210395,-0.0153502608,0.1137395203,0.2143018395,-0.3885603845,-0.0180562399,-0.0055848705,-0.4725532234,-0.0741913021,0.1810488999,-0.0713575929,-0.0899611339,0.2024392784,0.2232955694,-0.0838009343,0.4466342926,-0.1171104461,0.0356777087,-0.0997894108,0.2028720677,0.4329761565,-0.1946015209,0.2617235482,0.005433687,0.0032989436,-0.1672134101,0.0255579446,0.4860497713,-0.0751136094,-0.0690284595,-0.2605270445,-0.2147228867,-0.0338639282,0.2190820873,0.2872092426,-0.0122303348,0.088808313,-0.3458296359,-0.4819651842,0.2854718566,0.1356697679,0.4279277027,-0.268433243,0.2468790412,-0.0480657183,0.0747461542,-0.3780378699,-0.0029303164,-0.1230372787,0.0814786702,0.0933952332,-0.0422297716,0.2464932203,-0.3719470799,0.2005640864,-0.0135506121,-0.1404107362,-0.2477356642,-0.1142132431,0.0917968825,0.1512233317,-0.0857264251,0.2550349832,-0.0327793397,-0.0284641944,-0.0656810477,0.0506962538,-0.1049196869,0.0468475372,0.3307105303,0.0517331734,0.159046784,-0.0714107528,0.1555794328,0.127342239,0.2622174025,0.1479554176,-0.276673764,0.0164503101,-0.2314392477,-0.308029145,0.0147160478,-0.0413128473,0.1328901798,0.2649381161,-0.052156359,0.1773873568,0.1764705479,0.3283917904,0.370691061,-0.1610096693,0.1119323,-0.0156154083,0.3274745047,-0.2878553271,-0.0030336035,0.0507418886,0.328287065,-0.0897791833,0.0664328635,0.1396362484,-0.033914391,-0.172856614,0.1628388911,0.2124100924,-0.0721174181,0.1325653344,0.2194254398,-0.215861097,-0.0179969277,0.1354779452,-0.0346864164,-0.0599680059,0.2524558008,0.0012235348,0.5017405748,0.1732399613,0.1563586593,0.0170002691,0.0399356037,-0.0766131878,0.3757353127,0.1781032234,0.1634464562,0.0103168953,0.5730484724,-0.2033393085,-0.0163600408,-0.2891273499,-0.0069278753,0.1030387431,0.1672000587,-0.2514303327,-0.1146501526,0.0474308953,-0.2095384002,-0.1093070731,-0.2430689186,0.1624627411,-0.0133168949,-0.3490655124,-0.7971129417,-0.0968284011,0.1765956134,0.2645561099,-0.0989554226,0.1194581613,0.5688374639,-0.012276507,0.0959236473,0.510873735,0.1168439984,0.1878489554,0.0088196518,0.0478271581,0.2163966298,-0.0476574637,0.0938705727,0.1153004989,-0.0242871679,-0.1276434958,0.2739405036,0.2140211165,-0.1223961264,-0.1516688019,0.0691220984,0.0125294952,-0.0978337899,0.0810987204,0.044432871,-0.165815413,-0.0655834302,-0.2105146497,-0.3483026028,-0.010542552,-0.0954700559,0.1822632253,0.1857572347,0.0227298122,0.1367100626,-0.3435218334,0.2940482199,0.0502388552,-0.1212358698,-0.3042699695,-0.1231847405,-0.6220884919,0.0256204307,-0.0430268347,-0.5687163472,0.160253644,0.3183881044,0.154721573,0.04080097,0.0377040207,-0.4410664141,-0.1403719187,0.2798912525,-0.5994938612,0.0579673536,-0.2111626416,0.1578747779,0.0108535429,0.2902826667,0.1936143339,0.2438032627,0.0254209172,0.154586643,0.1822303236,0.1727087647,0.2426517159,-0.0007079316,-0.0502603538,0.5023124814,0.1890145391,0.3064455688,-0.28627491,-0.1896206737,-0.2009622455,0.2739821672,0.1934179217,0.0795575604,-0.0406807773,0.0368497185,0.0848916844,0.3296248019,0.027294891,-0.463974148,-0.4913818836,-0.0244045351,0.0427307859,0.0343273804,0.2714102864,0.5099771023,-0.0100327656,0.1016071364,-0.2187136263,-0.3751087785,0.4201334715,-0.1226819158,0.1824613363,-0.0290633906,0.2659626305,0.4343281686,-0.0306755211,-0.3477798104,-0.0143892094,0.4245873392,-0.2670196891,-0.128981337,0.2694119513,0.3743841052,-0.1542746425,-0.0170514025,-0.0524208434,-0.0187273193,0.0629983172,0.1779447943,-0.0380061045]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"Thanks @lhoestq\r\nIt would be great to improve coverage, but IDs are the really crucial part for us. We'd really appreciate an update to the dataset with IDs either way!","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":30,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nThanks @lhoestq\r\nIt would be great to improve coverage, but IDs are the really crucial part for us. We'd really appreciate an update to the dataset with IDs either way!","embeddings":[-0.120538637,-0.3434901834,-0.1572831124,0.4010956287,0.3150455356,0.0539410487,0.1680675596,0.0182598215,0.2290308475,0.2051194459,-0.1785497963,0.2475751191,0.1144534126,0.4436739683,0.0983844176,0.1101119295,-0.0065614008,-0.1371245682,-0.3460201621,-0.2986064553,-0.043396119,0.0572821386,-0.1505293101,-0.1744760871,-0.6682879925,0.014502733,-0.1977975965,0.2540426254,-0.10829743,-0.2513684332,0.3075421453,0.0355967507,0.0455216244,0.4934388399,-0.0001066689,-0.0335242599,0.1318748742,-0.0461245961,-0.0730438009,0.2865335941,-0.1540056467,-0.0908150151,0.0087959999,-0.3714568019,-0.0635807887,0.0710184649,-0.1505089402,-0.3604391813,0.1319084018,0.0261039324,0.205384165,0.1816887707,0.1772074401,0.0034406546,-0.0710885674,-0.129971832,0.0547026582,0.2795854807,0.4867466986,0.2233678848,0.0364004858,0.2235723585,0.0667103305,-0.235170275,0.2473031282,0.0893050358,0.3610637188,-0.1546256393,-0.3263548911,0.2234194279,0.8148685098,-0.0138418199,-0.3631150424,-0.0914176404,-0.0141353784,-0.1905049831,0.1663066298,0.0231584311,0.4013923407,-0.0192244574,-0.319467634,0.3984606266,-0.261143297,0.0211198628,-0.1350149065,-0.1003202647,-0.0169938281,-0.0170955136,-0.2322692424,-0.0396247543,0.0964685157,-0.0244267955,0.1179354787,0.1987861246,-0.3089933693,-0.2048525214,-0.0104514612,0.0081872335,0.2604072988,0.1771929264,0.2447296977,0.0415656567,-0.4094988108,0.0169901028,0.2245205492,0.231425643,0.4467464983,-0.4870941639,0.0470451675,-0.0431569442,-0.126160875,0.0585928001,-0.3492003083,0.0123141892,0.0359336436,0.1200719625,-0.1307220161,-0.3616788089,-0.1987831742,0.2999648154,0.0037913849,-0.1758429557,0.1398107111,-0.1497582048,0.0295124333,0.4021392465,0.016287338,0.0602587834,-0.5490458012,-0.3304890096,-0.3623380363,0.0852195844,-0.0705257729,-0.0179139096,0.1263846755,-0.11324431,0.3872211576,-0.1766903996,0.248853758,0.0368418023,0.0897519737,0.1391469389,-0.3680388927,0.2284670472,0.0240021516,0.0156878158,-0.0285932366,0.0518154539,0.0010552821,-0.0023563453,-0.1921722442,-0.2217102498,-0.0001565067,0.2905130386,-0.3537013829,-0.3277324438,-0.0906434059,0.0307502002,0.1729968488,0.1653863788,-0.1721050441,-0.0541453846,0.0885859132,-0.103844218,0.1517871767,0.013269145,-0.5510629416,-0.0082750237,-0.0016376682,-0.2535509169,0.1331641376,-0.0280404631,0.0421098731,0.2410802394,0.1172041595,0.134856075,-0.5617380738,-0.4743942916,-0.1683333814,-0.3735470474,-0.1620114595,-0.1173287705,0.1920337379,0.06438528,0.0327112228,-0.1309922934,0.1321861595,-0.0206760354,0.0525833443,-0.1722007841,-0.4544654191,-0.0696137249,-0.0106009115,0.2401328832,-0.0399660468,-0.3308128417,-0.1268031448,0.3707153201,0.1579273045,0.0076082805,-0.1216362491,0.0300809592,0.3287639916,-0.1603640318,0.0529127829,-0.1468402296,-0.0415217914,0.3014721274,0.1175609529,0.2241449058,0.7303125262,-0.0659259111,-0.3103890121,-0.1799812019,0.019187035,-0.2510041893,0.188114807,-0.1872991025,-0.0368539989,-0.2384052575,-0.1892733723,0.1357982457,-0.6147508621,0.0889656171,-0.4941736758,0.3555936515,-0.0201430395,-0.0115462411,0.0635659918,0.2999169827,-0.0188042,-0.3205814064,-0.06464421,0.5364034772,0.1481600106,0.0586812049,0.2148335725,0.2414090782,0.4149382114,-0.1923774481,0.1212522313,-0.2626796067,-0.0870194286,-0.1991874278,-0.4233083427,0.2120577544,-0.1298015267,0.025815906,-0.1078194827,0.0357511714,-0.0327578261,-0.1688144803,0.0306376815,0.0417337976,0.1496233791,0.1417599767,-0.2297071666,0.0775303245,0.0226610247,0.4194988608,-0.3977731466,-0.1804336607,-0.2100398242,-0.3189482689,-0.1661608368,0.1576022953,0.1537567973,-0.0587063059,0.493157208,0.321742624,0.0985389426,-0.0338733457,-0.039565403,-0.1556171179,0.0235522371,0.0013054968,-0.1736928523,-0.0394245237,-0.0639208779,0.0033622223,-0.1511525512,0.0975713953,0.1108916327,0.1865143627,-0.4474341571,-0.4137154818,-0.2139122039,-0.1277546734,0.2385014743,0.1950375736,0.1759382337,-0.2985452712,0.1582861543,0.0456121601,-0.0374015272,0.0498322174,-0.4003778398,0.2830519378,0.0254464466,0.0708897263,0.0738012269,-0.16967839,-0.1166358516,0.2136633098,0.0037979127,0.2504509091,0.3692833781,-0.0844295844,-0.0666720793,0.1225024089,-0.4055014551,-0.0075853136,-0.1713564694,0.2512726784,-0.1238033548,0.0624052063,0.0091360817,-0.0322459452,0.2944410741,-0.0420788638,-0.2515262067,-0.1448777616,0.180883944,-0.1780132502,-0.0485051982,-0.5882418156,-0.1118778363,-0.1036614701,-0.0437036492,0.4338700771,0.1562176794,0.0717754588,-0.1139168367,0.1505248696,0.2319957614,-0.2218727618,-0.331430763,-0.8559192419,0.3598410487,-0.0937612653,-0.4732737541,-0.0068540871,-0.0225818809,0.3321779966,0.1283280402,-0.3856081665,-0.3623091578,0.0941484645,-0.0910067335,0.5071302652,0.227072686,0.2142199576,-0.1695542037,-0.1697413921,-0.2687189877,-0.070729807,0.2278573811,0.265399009,0.4700402319,-0.3075751364,-0.330424428,-0.2363013029,0.0882734731,0.2966341078,-0.1167368591,0.0679814517,-0.0080630872,-0.1070004106,0.0694277585,-0.2100117654,0.6536984444,-0.0930678472,-0.0585269332,0.1649472713,0.2104713023,0.1872383058,-0.0612776615,-0.0215398129,-0.2185076177,-0.156147182,-0.1904433966,-0.0317160562,0.1063784584,-0.0354219303,-0.0511396714,-0.1161283553,-0.0987932682,-0.0448368527,-0.0475972146,0.1295496225,-0.0689517781,-0.9873172045,0.1482983381,0.0614228174,0.3500617146,-0.2704577744,0.0611164384,0.0373717695,-0.1009262204,0.0586141907,0.2390669286,0.1709834188,-0.3181259036,-0.2277029157,0.3203563988,-0.0780299678,-0.2289000303,-0.0751320645,-0.1517715752,0.1124701053,0.2031222582,0.0463269167,-0.2426430136,-0.0057296157,0.2446615249,0.1740336865,-0.2819228768,-0.115053013,-0.0262235738,0.0369151458,0.0816228613,-0.0664374605,-0.1242515147,0.23765257,-0.2308771014,-0.1715973765,-0.1750877053,-0.1097575054,0.4990906119,0.025351543,0.4236938357,0.0014485696,0.3811964095,0.3656773865,-0.0023787925,0.2452310175,0.5255581737,-0.237906158,0.2452268898,0.1621163189,-0.3930387795,0.4691255689,0.1861153394,0.0855190977,-0.5343071818,-0.3154668808,0.0659409612,-0.2242779285,0.2662936747,0.096879594,0.1935724616,0.0061810631,-0.2685910463,0.4813047945,0.1461718678,-0.0028585738,0.4337332845,0.3625036776,-0.100150615,-0.1965574026,0.5348649025,0.8573752642,0.1341347992,-0.2769600749,0.1873970181,0.3989755809,0.2924266756,-0.4698548019,0.0309969652,-0.1797637045,-0.3730557263,0.0669314042,-0.0661152601,0.1716619581,0.2435258478,0.1326761693,0.142465204,-0.1018322855,-0.3105866611,-0.1415357739,0.1586584002,-0.2393762767,-0.1737303734,0.3241587579,0.2493679076,0.1057081074,0.1377291083,-0.0025496155,0.0902035087,-0.2908142507,-0.0636130348,-0.1376073211,-0.0866234303,-0.1024518535,-0.0253317691,0.100755319,-0.3295689821,-0.0205372293,0.3093982041,0.2168633938,-0.0141019188,-0.1341948211,0.1402646601,-0.3714327812,0.0706169158,0.3393009603,0.0430370383,0.3660301268,-0.1421201378,-0.179025799,-0.0798792168,0.0444958135,0.259280026,-0.3610317707,-0.0524776764,0.0012445579,-0.4801269472,0.0059419572,0.1389621496,-0.0478829555,-0.1669546366,0.1891179681,0.2385030985,-0.0703892633,0.4668538272,-0.0985428989,0.0819873586,-0.1123839095,0.2597110569,0.4520571828,-0.1689188182,0.2715690136,0.0058891862,0.0190446917,-0.1723523736,0.1103761569,0.5367072821,-0.0884712636,-0.083217144,-0.2841585279,-0.2118476629,0.0026132111,0.2482561171,0.2747527659,0.0115590068,0.1169230565,-0.3189619184,-0.4721277654,0.3399023712,0.1061609834,0.3993187249,-0.2627232671,0.2220871001,-0.0375858061,0.084335573,-0.3584275842,-0.0522868782,-0.1350082904,0.0268154703,0.1082239896,-0.048263602,0.2857343554,-0.3554716706,0.1944626719,0.0014775008,-0.1028859541,-0.2374936342,-0.0796812102,0.1035741046,0.1681618989,-0.0202260539,0.2944042981,0.0079547437,-0.0590882339,-0.0811659545,-0.0051046326,-0.1927944422,0.0886522681,0.3592905104,0.093133986,0.0844173953,-0.0601135194,0.1306255162,0.1125586331,0.2832222283,0.0832648128,-0.2470523566,0.0801814049,-0.2404566109,-0.1764965355,0.0024411068,-0.0544298477,0.1487930864,0.2819154859,0.0178397503,0.1254841238,0.2038832307,0.3088871837,0.3877467513,-0.1485746652,0.1263663173,0.017546799,0.2953405082,-0.2960583568,-0.0695298165,0.0591543317,0.3725500107,-0.0748170689,0.0944050848,0.1876342148,-0.1102064699,-0.1191943586,0.1681518704,0.2572802901,-0.0466940776,0.0978733152,0.2053415179,-0.1981027126,-0.0566766597,0.1275028735,-0.0845119506,0.0223543681,0.1790510267,0.0296545345,0.4828330874,0.168474853,0.1510419697,-0.039497029,0.0128700165,-0.0632435307,0.3031077683,0.1901881248,0.1405856907,0.0364697427,0.6264362335,-0.1472074986,-0.0713050216,-0.2482667118,-0.0444766842,0.1188216135,0.1468964517,-0.2875571549,-0.0665120408,0.018064836,-0.2042826265,-0.1263274103,-0.2352712154,0.2165547758,-0.0306694396,-0.3188955784,-0.817333281,-0.1121054068,0.1354611665,0.3281665146,-0.1114771366,0.1377736628,0.6112169027,-0.0511416718,0.0817146748,0.5912322402,0.1228112876,0.0653845072,0.0319043249,0.0496211722,0.2083977312,-0.074464187,0.144826591,0.1422780305,-0.0415838584,-0.1758447737,0.2718534768,0.1791799664,-0.0965657905,-0.1126216501,0.0980194286,-0.0643076822,-0.05607602,0.0713275,0.0975251272,-0.1953419,-0.0374278352,-0.2533317208,-0.3634219468,-0.0417784564,-0.1204866916,0.1425012797,0.1355956048,0.0111645376,0.1273574978,-0.2924350202,0.3528837562,0.0714260042,-0.0985987633,-0.2925545275,-0.1884861887,-0.5612860918,0.0191088747,-0.0751693621,-0.5208989382,0.2102709711,0.2816257477,0.1408775896,0.0216734316,0.0330645964,-0.406450808,-0.063203752,0.315318495,-0.6724225283,0.0180293694,-0.1777895689,0.1920741796,0.0134409899,0.3005068302,0.27896294,0.2261444777,-0.0230141487,0.1866990924,0.1346617192,0.1517716944,0.2687849104,0.020696098,-0.0172373615,0.4619522393,0.1609878093,0.3414975405,-0.3036714196,-0.2003555745,-0.1873812526,0.2807472348,0.1854951978,0.1262613535,-0.0563758463,-0.0309700146,0.1362665743,0.3118083179,0.027343424,-0.4815067351,-0.542869091,0.0589117818,0.0760929585,0.0016455313,0.1991196722,0.4701716304,-0.0038901241,0.0714812055,-0.2085636258,-0.3970816731,0.4352813363,-0.1364441365,0.2403235137,-0.0660243928,0.2361035943,0.4462292492,0.0233610664,-0.358189553,-0.0511543043,0.4260799587,-0.1957479119,-0.1757148206,0.2586028576,0.4284217954,-0.1749580055,-0.0401437916,-0.0537576377,0.0108095445,0.0529570915,0.2350607365,-0.0215719901]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"I gave up at an even earlier point. The dataset I use has 204,017 train examples.","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":16,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nI gave up at an even earlier point. The dataset I use has 204,017 train examples.","embeddings":[-0.0992734805,-0.3724809289,-0.1397751719,0.4924610853,0.3237828314,-0.0019074179,0.1788346171,-0.0305928998,0.2268309295,0.2765001953,-0.1957730949,0.2362507582,0.0817245916,0.4162519574,0.134375602,0.0998578593,-0.0211556833,-0.1195153072,-0.3618007898,-0.2939094305,-0.0009093486,0.0598329566,-0.0843998268,-0.1295554042,-0.6914079785,0.0213110577,-0.145983398,0.2635234892,-0.078791894,-0.1954722255,0.2472316027,0.0306974407,0.0586973578,0.5738731623,-0.0001047336,-0.0023851253,0.1076081246,-0.0218012314,-0.1017149165,0.2196226865,-0.1683497578,-0.0203225575,0.0056747426,-0.3873440325,-0.086523287,0.0476220846,-0.1387609392,-0.3502879143,0.2238690108,0.0132080894,0.2210143656,0.1479527652,0.1104970574,-0.0569568239,-0.052052319,-0.1075223237,0.0981687829,0.2489619553,0.4054052234,0.2546094656,0.0566349998,0.1681368798,0.0725834966,-0.2038870156,0.2872775495,0.0664794445,0.3404063582,-0.1934158653,-0.3072768152,0.3304914236,0.8445416093,-0.0371893533,-0.2612978816,-0.0371643417,-0.0114017399,-0.1463241428,0.1685736775,0.0669240579,0.3521563709,-0.0419837348,-0.2714678347,0.4366954565,-0.2350869328,0.0346358381,-0.1616154909,-0.0856527761,-0.0703878626,0.0162692089,-0.2013381124,-0.0584483407,0.0550670661,-0.0501671173,0.1705985367,0.1830284148,-0.2987988293,-0.1156960353,-0.0200141575,-0.0286237504,0.2357100993,0.153275907,0.2867733538,0.0642203465,-0.3826186061,0.0466699935,0.2231809944,0.2809687853,0.4171878397,-0.5452025533,-0.0015246755,-0.1041456237,-0.125152573,0.0936611891,-0.3935850263,-0.0319752246,0.0116169304,-0.0072853626,-0.1651062816,-0.3474945128,-0.2218122184,0.2655235827,0.0236572176,-0.0896358564,0.1897134483,-0.1298575997,0.0461675562,0.3801033795,-0.0006398535,0.0334395356,-0.5421857238,-0.2864270508,-0.3575430512,0.11921525,-0.081427969,-0.0300749056,0.1057599857,-0.0802363828,0.3754305542,-0.095330961,0.2607867122,-0.0121264681,0.0667468458,0.084838517,-0.323320657,0.2186362892,0.0322239771,0.0163808335,-0.0498163067,0.0311084781,-0.0335647538,-0.0112234047,-0.1946563572,-0.2256243229,0.0123133399,0.31813097,-0.36181283,-0.2805413306,-0.1812169105,0.0062777014,0.1754130274,0.2128992081,-0.1940576583,-0.1520864367,0.0713577494,-0.1175183728,0.1709081084,-0.0380265675,-0.5679003596,0.0302610546,0.0239037387,-0.2643796802,0.0763849914,0.0145943323,-0.0121541023,0.24252294,0.1205772683,0.1891646534,-0.5386627913,-0.4765385389,-0.1282004863,-0.3288970292,-0.1726729572,-0.0970215127,0.1299283504,0.1182871237,0.0263427068,-0.1328428239,0.1855413616,0.018991163,0.1131550893,-0.1729653478,-0.424064815,-0.0728575289,-0.0183058996,0.2005013376,-0.0861096606,-0.3178793192,-0.1180911064,0.3464241922,0.1670276523,0.006657199,-0.1042168513,0.07769081,0.3016307354,-0.1446907818,0.0841599628,-0.197345838,0.0233378988,0.3096409738,0.1285746694,0.226447314,0.7804632187,-0.0542837866,-0.3663805723,-0.2252172679,0.0094813388,-0.1826499999,0.2018443644,-0.2278423756,-0.0043877778,-0.2264474332,-0.1856216341,0.1292771101,-0.5900850892,0.0397550464,-0.4913723767,0.302693069,0.0295926854,0.0148401009,0.0524943061,0.2697274387,-0.0397591256,-0.2674591839,-0.0272119734,0.552048862,0.1712252945,0.0097715557,0.181179136,0.195578903,0.3629962206,-0.1806837469,0.1403544247,-0.2309368253,-0.0687372983,-0.1801776886,-0.3973367512,0.2244195938,-0.1129113436,0.0193621535,-0.0542983226,0.014410967,0.0128692957,-0.1489400864,-0.0007944622,0.0223872196,0.2171737999,0.1394062042,-0.2246800065,0.0960168019,0.0005261555,0.3793882728,-0.3613011539,-0.1710775644,-0.1926082671,-0.333950758,-0.2287157327,0.1707284153,0.1128912941,-0.0203486923,0.4611949325,0.3074295223,0.0879984796,-0.0402851105,-0.0910277963,-0.1547590792,0.0054164343,0.0353737622,-0.0759334639,-0.0705040246,-0.052472055,0.0217551719,-0.1583797336,0.0778034925,0.0994941145,0.1443371922,-0.4329530001,-0.3254625201,-0.2261655927,-0.1014579609,0.1962684095,0.1690274775,0.1492390186,-0.2707683742,0.1685469896,0.0528135151,-0.0432525761,0.012200119,-0.3721601069,0.299272269,0.0463093296,0.1064097807,0.1071182042,-0.1652509272,-0.1766562909,0.2086236775,0.0322136655,0.3394019604,0.3760530055,-0.0735379383,-0.0521858484,0.1309556365,-0.4230750501,0.0316206813,-0.2039082646,0.2718333304,-0.1448121071,0.0241173245,0.0496199355,-0.0104073929,0.2507301569,-0.0399122164,-0.2387535721,-0.0692075416,0.1907081604,-0.2634196579,-0.0955595896,-0.5472244024,-0.0945072621,-0.0876357481,-0.0245300122,0.482038945,0.2032702863,0.0930781588,-0.0947001055,0.1994303316,0.2466465533,-0.1848699003,-0.375430584,-0.8140149117,0.3188571036,-0.1111662909,-0.5230090022,0.0298817232,-0.034150552,0.2833774984,0.1337906867,-0.391861558,-0.3439759314,0.0009986291,-0.1173225865,0.5061702728,0.2171716243,0.2065120339,-0.1710245907,-0.1827392727,-0.2552517354,-0.0513568632,0.3030754924,0.2974095047,0.4628184438,-0.3247472644,-0.3170234263,-0.2085956484,0.0620310865,0.2097071707,-0.1498320401,-0.0147466455,-0.0671081915,-0.1116598397,0.0421481803,-0.115177542,0.6210405231,-0.1083307266,-0.0778184235,0.1630093455,0.2271415144,0.2326095998,-0.0459358022,0.0751870126,-0.1875701249,-0.1476857811,-0.170491159,-0.046982687,0.0304857232,-0.0582415275,-0.0764017701,-0.1003339961,-0.1045027375,-0.1077798307,-0.0384337716,0.1093312129,-0.0412414558,-0.9979836941,0.2905904949,0.0495607518,0.3575127125,-0.2970937788,-0.0282555595,0.0079540107,-0.1286726296,0.0868094116,0.2151583433,0.2042945176,-0.2144073993,-0.2301448882,0.3166199327,-0.0985527933,-0.31382972,-0.08796186,-0.1631443948,0.0783091635,0.2113242596,0.0470894948,-0.2636475265,-0.0264390316,0.2030963898,0.2186726034,-0.2879211307,-0.1855091155,-0.0728501156,-0.039471589,0.0451778844,-0.1197460592,-0.1093575358,0.2601594627,-0.2708394527,-0.1437163502,-0.1985710561,-0.1354959458,0.5682415366,0.0463847145,0.4107861519,0.0025732103,0.3345633745,0.3583497703,0.1150297299,0.2347040176,0.4786400199,-0.1482518911,0.2536826432,0.1591559798,-0.3659279644,0.3642497361,0.1660316139,0.0849331915,-0.4914815128,-0.2385421842,0.089146547,-0.1952740848,0.2906859815,0.0943458527,0.1904981881,-0.0780878887,-0.2748828828,0.5450213552,0.1387102753,-0.0097044231,0.3804354072,0.3659901321,-0.1164856628,-0.1873203665,0.5151320696,0.8932386637,0.0769281313,-0.1938199699,0.2794930935,0.4090279937,0.3263081908,-0.4818765521,0.0356591605,-0.1621779799,-0.3530299067,0.0403982401,-0.0972943529,0.1294897199,0.2668583989,0.1436258852,0.0591248833,-0.1650301516,-0.2733377814,-0.1946350485,0.1724725664,-0.2729213834,-0.1535346806,0.2773355544,0.2721074224,0.1305437982,0.1499999017,-0.0093121212,0.0358893201,-0.3253676295,0.0026430956,-0.0561356544,-0.067471981,-0.0896209553,0.0504179262,0.049583111,-0.3257537484,-0.0444530174,0.3328700066,0.2143797129,-0.0634539649,-0.1007064432,0.1711914241,-0.3458360136,0.0560581833,0.3358947635,-0.0251430608,0.4243965447,-0.1699172854,-0.1809005141,-0.0374361575,0.1033854708,0.2311223,-0.3712255955,-0.0264382847,-0.0079286927,-0.4698669612,-0.0746537223,0.1806062609,-0.0982644632,-0.083018288,0.2021785229,0.2307329476,-0.0614757985,0.4499377012,-0.1099568605,0.0509465411,-0.1047707498,0.190282926,0.4379962087,-0.1938695014,0.2532499731,0.0169406645,0.0171228871,-0.1588537246,0.0500547588,0.5058929324,-0.0729455128,-0.0720376968,-0.2528293133,-0.1984861493,-0.0053443909,0.2442036569,0.2831345201,-0.0258462187,0.1003974453,-0.3302686512,-0.4769794643,0.2921728492,0.1310619116,0.4174727499,-0.2587912083,0.2389146835,-0.0494063422,0.0539341047,-0.3782365322,0.0177129116,-0.1212550178,0.0698163509,0.0807663649,-0.0584794953,0.2591819763,-0.3779758215,0.1792130023,-0.0013165118,-0.1657775193,-0.2515692711,-0.117622152,0.0890248865,0.1660424173,-0.1049299836,0.2745767832,-0.0240850523,-0.0576299094,-0.0724285692,0.0604632236,-0.1173611879,0.0302801393,0.3011862338,0.0174231268,0.1451926827,-0.0611411929,0.1379962862,0.1187238246,0.2595630288,0.1873345226,-0.2682830691,0.0385159291,-0.2294558436,-0.2966991365,0.0164701287,-0.0487540774,0.1081340611,0.2654646337,-0.0660140291,0.1556817144,0.1725357473,0.3515242934,0.3776398599,-0.1848825365,0.1109502763,0.005314528,0.3281937242,-0.2923992872,-0.0011996442,0.0496316962,0.3295218349,-0.0788833275,0.0898390338,0.1686362922,-0.0530002229,-0.1599928439,0.1708453298,0.2100818306,-0.0832964778,0.1048952192,0.2073547244,-0.245584324,-0.0101327598,0.1504190117,-0.0316479616,-0.0314684808,0.2588604987,0.0131108882,0.511300385,0.1838934273,0.1635545492,0.0170183387,-0.0001414821,-0.0695835054,0.3481970429,0.1504073292,0.1592499018,-0.026223952,0.5760987997,-0.2066618651,-0.000573052,-0.3108677864,-0.0289764032,0.0962472484,0.1491611749,-0.2483234555,-0.1039681658,0.0481243059,-0.1860793084,-0.1290516406,-0.2516666055,0.1477458924,-0.0205544345,-0.3541607857,-0.8141177893,-0.1085595116,0.1517539471,0.276369065,-0.0995304063,0.1424702257,0.5436642766,-0.0344088823,0.0800300166,0.5004680753,0.1134971306,0.1973873526,0.0002368018,0.0506470129,0.2266128808,-0.0268162377,0.1014484391,0.1067324951,0.0004926063,-0.1504102796,0.29747805,0.1995944679,-0.1207931861,-0.1327406764,0.0375786237,0.0263077337,-0.1072421148,0.0461627096,0.0525134951,-0.154801324,-0.0805145577,-0.2089970261,-0.3566538692,-0.0255055875,-0.0997858569,0.2013784647,0.1901117414,0.0169145316,0.1408592761,-0.3599381447,0.3116765916,0.0382624157,-0.1093354374,-0.2739180923,-0.1286856532,-0.6168878078,0.0133419735,-0.0849928781,-0.5377328992,0.169963792,0.3220706284,0.1508049965,0.0472331308,0.0081662973,-0.4305217862,-0.1222295389,0.2721708715,-0.5976177454,0.0793443099,-0.2225140482,0.1740790308,0.0195114799,0.2972213924,0.2198368013,0.2128149718,0.0154558569,0.155493781,0.1872501075,0.1560057253,0.2486762851,0.0088381441,-0.040969789,0.4796516001,0.1974692792,0.3138945401,-0.3052363098,-0.1834245175,-0.1923083961,0.2781386673,0.205017522,0.0816336125,-0.0333200507,0.021579124,0.0972591415,0.3132694364,0.0403320268,-0.4563267529,-0.4822052121,-0.0240228418,0.0480608121,0.0503490306,0.2764172554,0.5173118114,-0.0225047227,0.0461858101,-0.2039900124,-0.3806153238,0.4077625871,-0.1462410092,0.1850555986,-0.0652313679,0.2599229515,0.4546602666,-0.0201598238,-0.3533847332,-0.0126582393,0.4080742896,-0.2572019398,-0.1338899881,0.2706454396,0.3691138327,-0.1420966685,-0.0221814904,-0.0474336706,-0.0362769775,0.056527853,0.1706421226,-0.0557023026]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"@lhoestq @sshleifer like @jbragg said earlier, the main issue for us is that the current XSUM dataset (in your package) does not have IDs suggested by the original dataset ([here is the file](https:\/\/raw.githubusercontent.com\/EdinburghNLP\/XSum\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json).) Would appreciate if you update the XSUM dataset to include the instance IDs. \r\n\r\nThe missing instances is also a problem, but likely not worth pursuing given its relatively small scale. ","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":63,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\n@lhoestq @sshleifer like @jbragg said earlier, the main issue for us is that the current XSUM dataset (in your package) does not have IDs suggested by the original dataset ([here is the file](https:\/\/raw.githubusercontent.com\/EdinburghNLP\/XSum\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json).) Would appreciate if you update the XSUM dataset to include the instance IDs. \r\n\r\nThe missing instances is also a problem, but likely not worth pursuing given its relatively small scale. ","embeddings":[-0.1664320678,-0.386536926,-0.1475521773,0.4103152156,0.3375102282,-0.0053339102,0.1342155039,0.042721767,0.2139728367,0.2906926572,-0.2072107047,0.3043749928,0.0520256944,0.3954081237,0.1108666435,0.1289572567,-0.0369182862,-0.0796159953,-0.3687422872,-0.2781106234,0.0103583038,0.0866705626,-0.0639782995,-0.1064027995,-0.625708282,-0.0011952977,-0.0703786388,0.2886465192,-0.124231711,-0.2318082452,0.2674055994,0.0330968015,0.0323780961,0.5344508886,-0.0001019957,0.0556662455,0.1717787236,-0.0185712148,-0.1103478968,0.2229582369,-0.1202735603,-0.0083083082,-0.0214206614,-0.3645848334,-0.1546657234,0.0537902489,-0.1565698832,-0.3595399261,0.2612186968,0.0356135406,0.26186198,0.1737311631,0.1423487514,-0.1017601192,-0.0694580376,-0.2182796597,0.0572399423,0.2420994192,0.3895654082,0.2819084823,0.0757031664,0.1821228862,0.0750418454,-0.1728700101,0.2609300017,0.081781514,0.3629446626,-0.1643868983,-0.3031727076,0.279101491,0.7783159614,-0.04961982,-0.2809360921,0.0017530457,0.0423495062,-0.1805801839,0.1583729386,-0.0204358641,0.3124708533,0.016694488,-0.1730660498,0.3848587275,-0.22889328,0.032479357,-0.1435603499,-0.0440803878,-0.0700866953,-0.0146878241,-0.1825108826,-0.0477214046,0.008945467,0.0036318516,0.1416556388,0.1554021984,-0.3144340217,-0.1411520988,0.020316394,-0.0635992885,0.192292273,0.1255430877,0.2808846235,0.1205654144,-0.3568879366,0.0459334739,0.1722762883,0.2889683843,0.3624550998,-0.5636849999,0.0376869291,-0.0363246426,-0.0748789757,0.0920707807,-0.3312948644,-0.038878683,-0.0027432879,0.0162025187,-0.1309365332,-0.3646146953,-0.3093115687,0.2398579568,0.0613963045,-0.0323374085,0.1363776326,-0.0783258975,0.0352188647,0.3873426318,0.0540400445,0.0547506846,-0.5485547781,-0.251409173,-0.3607020676,0.1081757694,-0.0969993472,-0.0304282699,0.0917695239,-0.0993936211,0.3530322313,-0.0829657242,0.267860502,-0.0113612311,0.0562427416,0.1415773034,-0.3755877316,0.1932751387,0.0244414788,-0.008043726,-0.0267244298,0.0531736165,-0.0133817336,-0.0639151111,-0.2145490348,-0.2001978755,-0.00622351,0.3539358377,-0.3119359314,-0.2647695839,-0.137862578,-0.0376398712,0.136985749,0.1669009477,-0.2109556049,-0.1167255715,0.0583696254,-0.1073395312,0.165079847,-0.0286728367,-0.5455512404,0.0090003312,-0.0160474032,-0.2775516212,0.0649718866,-0.0520804413,-0.0131399231,0.2424101382,0.134917438,0.1709459126,-0.4772084653,-0.5170894265,-0.1987747848,-0.2960907817,-0.2253487557,-0.1023816913,0.1153729334,0.0857764632,-0.0046228361,-0.0884748846,0.1872546077,0.0400169455,0.1452699751,-0.1639282256,-0.4526293874,-0.0969946831,-0.0505807512,0.2002424598,-0.040960636,-0.3017985821,-0.1011529565,0.3752578199,0.1631080955,-0.0287367702,-0.0725138187,0.1020362899,0.3146270812,-0.159934178,0.0529153645,-0.2208855003,0.0057637407,0.2957918942,0.0420864224,0.2648531199,0.6834337115,-0.0381439887,-0.3916221857,-0.216774255,-0.0180663634,-0.197099641,0.245301187,-0.1623886377,-0.021076763,-0.1782017499,-0.1711999327,0.1778254509,-0.5446184874,0.0424658135,-0.4682898223,0.3180691004,-0.0250312779,-0.0367774516,0.090384841,0.2675651312,-0.0208951328,-0.2781307101,-0.0240456127,0.5908074975,0.1188962087,-0.013860072,0.1396992952,0.1338040382,0.3510762155,-0.1848499179,0.1054525897,-0.2123266608,-0.0840017647,-0.137252152,-0.3525788486,0.2411815077,-0.1131720617,-0.0369871967,-0.0518106036,0.0354195163,0.0609064922,-0.1979057193,0.0302652847,-0.0250801761,0.1439371705,0.1197232604,-0.2250252068,0.0177500676,0.0416907743,0.3525958061,-0.2550519705,-0.12358623,-0.1564978212,-0.3450434804,-0.1798163503,0.1364571601,0.1099741906,0.0011714649,0.4829594791,0.3248569965,0.0352756567,-0.0094200652,-0.142456755,-0.1150498316,0.0122484034,-0.0189624596,-0.0676124394,-0.0167820025,-0.0689265877,0.0121765211,-0.1636830121,0.0999635532,0.1242448166,0.1916476935,-0.3735172749,-0.3346714377,-0.2764015496,-0.0854921639,0.1280031353,0.1754529327,0.1941688806,-0.2772435844,0.1896548271,0.0737726316,-0.0582990274,0.063777253,-0.3311454654,0.2439831644,0.0624697097,0.046151828,0.0960207656,-0.1329865754,-0.1988453865,0.2283081114,0.025777122,0.342841208,0.3706849217,-0.0861969665,-0.0631765872,0.0952487737,-0.4376233816,0.057785362,-0.1971290559,0.260050714,-0.1256955862,0.0024522298,0.1020699814,-0.0584809408,0.2918134034,-0.1022218466,-0.2722827792,-0.0577913821,0.1697316617,-0.2628473341,-0.1111091748,-0.6046498418,-0.0556150153,-0.1341214478,-0.1060399115,0.4435876906,0.1438321322,0.0852440298,-0.0556117259,0.2129310966,0.2396076471,-0.1701513082,-0.3796664178,-0.7306688428,0.2577499747,-0.1202305108,-0.5162162781,0.005153941,0.0044584493,0.3787733912,0.1294912845,-0.3869502842,-0.3610758185,0.0458289459,-0.0663785711,0.4467763901,0.1637556702,0.298425287,-0.1350425929,-0.2306153625,-0.273147285,-0.058093302,0.2619219422,0.2238494158,0.4402632415,-0.3096805513,-0.2505705059,-0.1828395575,0.0408152863,0.1465577334,-0.1679843217,0.0533489957,-0.0699994937,-0.0172503311,0.014959679,-0.1258848608,0.5826581717,-0.0808413923,-0.0290865265,0.1526952535,0.1758556068,0.1942481697,-0.0500300899,0.0634546056,-0.1752151251,-0.1591147929,-0.2025383562,-0.0330818184,0.0478348471,-0.0329142511,-0.022509234,-0.0822558254,-0.0675200075,-0.054332491,-0.0146009792,0.0958459303,-0.0414926298,-0.981575191,0.2990341485,0.0618139952,0.372712791,-0.2710430324,0.029136641,-0.0083192615,-0.1483341306,0.1056241244,0.1202371493,0.1292061508,-0.2489224523,-0.1767417192,0.3190035522,-0.083074294,-0.3016256392,-0.0921347514,-0.1500041038,0.089154087,0.239774242,0.0828034058,-0.3069484532,-0.0860883445,0.2430820018,0.1935516,-0.2661921084,-0.1325029284,-0.1314567775,-0.0764248148,-0.0045662094,-0.1026513949,-0.1512290984,0.2996350527,-0.2088680863,-0.1363309473,-0.1865468472,-0.1697449535,0.5243932605,0.0616226494,0.438652426,-0.0094692186,0.3387262523,0.3222386539,0.1136418581,0.2265524864,0.4560769796,-0.1674663723,0.2600370646,0.1371494979,-0.3577572703,0.3021462262,0.161236465,0.0334418155,-0.4339480102,-0.2414716631,0.065794453,-0.1872550994,0.2881970108,0.1533886641,0.1901845634,-0.0713487491,-0.233377099,0.5094255805,0.1026638001,-0.0210987013,0.3991836607,0.3343642652,-0.1494823843,-0.1722918749,0.5163660645,0.861309588,0.0960143507,-0.1809175313,0.2801845372,0.3966981173,0.306343317,-0.4772422016,0.0945080817,-0.1617206186,-0.3101737499,0.0645947903,-0.038032189,0.1542025357,0.2285586596,0.0645368695,0.0379699841,-0.1021009013,-0.2273483425,-0.1729261726,0.2186796218,-0.2502020299,-0.1699630767,0.2527996004,0.3129732907,0.0823612511,0.1384132355,-0.0397849493,-0.0000196773,-0.2951774597,-0.031917695,-0.0516387373,-0.0022180323,-0.0219538007,0.0386337861,0.043175254,-0.3069759905,-0.0473717712,0.2716632783,0.229662016,0.011652288,-0.1372314245,0.1732466519,-0.3802205622,0.0350510292,0.3145688474,-0.014671308,0.4454981983,-0.222269401,-0.2039477229,-0.0054138419,0.1337212026,0.1284480691,-0.3104580641,-0.0512507893,0.0256211609,-0.4915601015,-0.0540778264,0.234854728,-0.0531951375,-0.114726238,0.240081951,0.1850808263,-0.1270151734,0.4229767323,-0.0621418543,0.0092251133,-0.0947503373,0.2264069021,0.4063761234,-0.1673384607,0.2598178685,-0.0642514378,0.0166761559,-0.2114754617,-0.0516393408,0.5009376407,-0.1143606603,-0.0955844596,-0.2136882395,-0.2379795313,0.0407908373,0.2407123148,0.2750689089,0.0099793579,0.0775905177,-0.3289133608,-0.4556131959,0.2503007054,0.1304807514,0.4150586724,-0.2764041722,0.3125865459,-0.0520960391,0.0643858165,-0.4254912734,-0.0106896963,-0.1455169171,0.0274053719,0.1003261581,-0.0841947123,0.2783965468,-0.3419241905,0.2316702455,0.0478705131,-0.1745813489,-0.2762839794,-0.1365553588,0.0704332814,0.1491743624,-0.0776643753,0.2476326376,-0.0257330481,-0.0924856439,-0.050654415,0.093866162,-0.1393672973,0.0368533581,0.2981911898,0.0663713962,0.1586546302,-0.0354916006,0.0781941116,0.1130144224,0.2729159594,0.1305636615,-0.2734082043,0.0325730667,-0.2120262831,-0.3130917847,0.0232549496,-0.0222607423,0.1726319641,0.2782164514,-0.1222954467,0.1062289625,0.1534734517,0.3541401625,0.3513805568,-0.1526019275,0.0475028865,0.0252227616,0.3683805764,-0.3287710845,-0.0222936627,0.0255031046,0.2614387274,-0.0551192909,0.0587954186,0.1816557944,-0.0232636556,-0.0433841981,0.2002995461,0.2268705815,-0.1005388126,0.0965425,0.2139596641,-0.2077321708,-0.0051936368,0.1913173348,-0.0972520038,-0.0754692852,0.2856114209,0.0165811237,0.5100924969,0.1186038777,0.1476446986,-0.0365236364,0.0320656896,-0.0653874278,0.3482097983,0.1848556101,0.15584369,-0.0111770891,0.5806465149,-0.2293238491,-0.0498247445,-0.3147738278,0.0028497095,0.0439126715,0.1555723846,-0.2418771684,-0.0944522768,0.0503273457,-0.1549907625,-0.1050301343,-0.1727746576,0.1736911833,-0.0314557925,-0.3885213435,-0.7518583536,-0.1490351707,0.1357879192,0.2569057345,-0.0935985595,0.2078541517,0.5303714871,-0.0068027638,0.0690200329,0.4900555015,0.1336439103,0.1916834861,0.0421600863,0.0146905705,0.1988672614,-0.0280664414,0.0569916852,0.1483778656,0.0388246961,-0.1054349542,0.2769365311,0.246049881,-0.180809617,-0.0960547775,0.0311530735,0.0110993981,-0.0720743984,0.1268103719,0.124697268,-0.1188281327,-0.1453616023,-0.1919303536,-0.3672336042,0.0163552258,-0.1210338846,0.1684666276,0.1589729637,-0.0406360812,0.1584186107,-0.2721065581,0.3382392526,0.0010287641,-0.1129211411,-0.2934088707,-0.1318314075,-0.6782485843,0.074425891,-0.0800588876,-0.5033019781,0.1502721161,0.2766647637,0.1483566314,0.0735668316,0.0705633983,-0.458313942,-0.119702518,0.2027541846,-0.6247524023,0.0986149833,-0.1591926664,0.1226178855,0.026594596,0.1965227723,0.1745742708,0.2322185785,0.0766950771,0.135770902,0.1327916384,0.0952915922,0.1927194446,0.0488427505,-0.0298955273,0.4581679404,0.1391884983,0.2371512949,-0.3483576477,-0.1973422766,-0.2617532909,0.2313994318,0.2493049204,0.1476470828,0.0219679642,0.0007610826,-0.0060909968,0.3671992719,0.0588128865,-0.4401395321,-0.4696762264,0.0203338508,0.0525895096,0.0278285481,0.2831587493,0.5465506911,0.0105194077,0.0863931775,-0.3044322729,-0.3901852965,0.426586926,-0.1244323775,0.128526181,-0.0073514543,0.2603491545,0.4360617995,-0.0897243917,-0.3726535439,0.0235138163,0.3749047518,-0.1784073114,-0.1523081362,0.3035390973,0.3506132662,-0.1332809627,0.0069049643,-0.0433245599,0.0085623553,0.0376285017,0.1732897311,-0.0825776011]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":">So I managed to download them all but when parsing only 226,181\/226,711 worked.\r\n\r\n@lhoestq any chance we could update the HF-hosted dataset with the IDs in your new version? Happy to help if there's something I can do.","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":38,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\n>So I managed to download them all but when parsing only 226,181\/226,711 worked.\r\n\r\n@lhoestq any chance we could update the HF-hosted dataset with the IDs in your new version? Happy to help if there's something I can do.","embeddings":[-0.1108686551,-0.2911382914,-0.1378347427,0.4550742209,0.2817803621,0.07282985,0.0544087291,0.0913588926,0.2821643353,0.2655704319,-0.4389287233,0.0682012886,0.2014297694,0.3437097073,0.0978380814,0.098360911,0.0280660875,-0.0688482076,-0.39984712,-0.3520366251,-0.0486405119,0.0790460929,-0.0701801255,-0.1254000813,-0.6150904298,-0.0004532764,-0.2522834837,0.3219082654,-0.0684697777,-0.2613791823,0.2724297941,0.1768980622,0.1588336527,0.585562706,-0.0001079536,-0.0021305003,0.0701949671,-0.0362244174,-0.2027471662,0.2457122058,-0.1254195124,-0.0502531268,-0.104641296,-0.3473931253,-0.041289553,-0.0858429298,-0.1789296269,-0.2968728542,0.2022092491,0.0117187882,0.190333128,0.1982564479,0.1489178538,-0.0657885075,0.0004690879,-0.0772708431,0.1758524776,0.2949224412,0.384034425,0.2338938713,0.0522160083,0.2036968321,0.0793882012,-0.1030988246,0.2630741894,0.0071052993,0.2479682416,-0.1099337339,-0.2626760304,0.3442577422,0.7662885189,-0.008736901,-0.3143163025,-0.0585712045,-0.0513017103,-0.1753983647,0.1823643744,0.0166409016,0.3686720431,-0.0167589299,-0.1219330877,0.2837216258,-0.2416824698,0.083961755,-0.1171373576,-0.0086127352,-0.040566802,0.0076605235,-0.2974540889,-0.0240416266,0.1238147095,-0.136688605,0.1175005734,0.1794935018,-0.3197864294,-0.0899894759,-0.0294952542,-0.0761650801,0.2779528499,0.0087842755,0.1777738482,0.1494703144,-0.4885767996,-0.0050449204,0.2457178682,0.1824131161,0.4060859382,-0.5367921591,0.0424551889,0.0243291669,-0.0892695263,0.0768095329,-0.3691336513,0.0522828586,-0.1361779422,0.0634151101,-0.2079424411,-0.4489310086,-0.243271023,0.2925207913,0.0687269196,-0.11138805,0.10921707,-0.0368537605,0.0220870804,0.4510500431,0.027541237,0.060727492,-0.590329349,-0.4203078747,-0.3149056733,0.1223097816,0.0103718368,-0.0122706341,0.0663291216,-0.1219097227,0.3088372648,-0.1253540367,0.3376320004,-0.0142568406,-0.0368471257,0.0412789471,-0.3466891348,0.2455322295,0.0649793968,0.0139358658,-0.0470767021,0.0491428934,-0.0200307593,-0.0273537301,-0.2576954961,-0.1876059324,-0.0546092503,0.2807949781,-0.2417281121,-0.2601537704,-0.2303065211,0.0229586214,0.0973036513,0.1958481967,-0.2307623178,-0.1498784423,0.102684319,-0.0829291791,0.1686289608,0.0426320396,-0.6364934444,0.0590564907,0.0792019665,-0.245447576,0.0443412624,0.0417248681,-0.0496388674,0.2293722034,0.0753990412,0.1855361462,-0.5582084656,-0.5333790779,-0.1740589887,-0.3230131865,-0.0884199515,-0.0928148702,0.0452460162,0.0335601307,0.0858178064,-0.1337661892,0.0952780396,0.002880926,0.1338361651,-0.2259705216,-0.4646393657,-0.0772188157,-0.1532885283,0.220384568,0.0640577897,-0.2609004378,-0.0373654775,0.4446001947,0.3015833199,-0.0276407544,-0.0262201652,0.1332650185,0.2527402043,-0.0071262028,0.1067680493,-0.1595184207,-0.0529889576,0.2997504473,0.0269468036,0.1723457575,0.7534607649,-0.0607872084,-0.4119504392,-0.2014401853,0.0471296646,-0.1689955145,0.1534011662,-0.1586530507,-0.0525268354,-0.2474772483,-0.1893184781,0.29528144,-0.5699025393,0.0937045962,-0.4114120007,0.4359882176,0.0315888599,0.0111059481,0.1427347958,0.2196141183,0.0349663422,-0.3012127578,-0.1030707881,0.5682691336,0.0972987935,-0.0176486019,0.2086599767,0.1073379517,0.4287082553,-0.1494135261,0.2242486328,-0.2733376026,-0.0340476111,-0.1739271581,-0.3439344168,0.2497535497,-0.0589425489,0.0061067156,-0.0693114549,-0.0251988061,-0.0190039799,-0.1324999183,0.1328186244,-0.0106653543,0.1425020546,0.139212504,-0.2239284813,-0.0485858284,-0.0573377982,0.4266648591,-0.270881027,-0.1722155958,-0.2802174687,-0.2724072933,-0.2439087033,0.1403642148,0.1972509176,-0.1188449189,0.3779392242,0.2638700604,0.0522032343,0.0659703985,-0.0586499833,-0.1196560189,-0.0230520237,0.0211806502,-0.0362943113,-0.0213891231,-0.0011307135,0.0382127762,-0.2089093775,0.0571318232,0.1290062219,0.1677093357,-0.440556258,-0.3460095227,-0.2939581573,-0.1069727689,0.2392285019,0.2096288949,0.0524936654,-0.3409094512,0.1800817102,0.1349406391,-0.0414221548,-0.0222297367,-0.4305756688,0.2545246184,-0.0061297151,0.061091505,0.068220824,-0.0865661651,-0.2389592081,0.1626758426,0.0419396162,0.3007407188,0.3395791352,-0.0517323613,-0.0833902434,0.0332344994,-0.3909251988,0.0608562939,-0.1610302031,0.2292688042,-0.1898724735,0.0489438623,0.0256061349,0.0482384637,0.2835533321,-0.005784139,-0.1941837966,-0.0817633942,0.1162676215,-0.1925490052,0.0099135665,-0.5236006975,-0.1250429451,-0.1238108501,-0.1371329278,0.4118534029,0.2634223998,0.0554481708,-0.1676878631,0.1139132008,0.1504206359,-0.3045866191,-0.3980051875,-0.8720009923,0.3317911029,-0.1492508352,-0.5336837769,0.0280001499,-0.0066597317,0.3115903437,0.0910401195,-0.4845391512,-0.3018910885,0.0788630769,-0.0095073795,0.4274983406,0.1915900409,0.149491176,-0.2446286678,-0.1636426449,-0.2711959779,-0.0776796713,0.2578827441,0.3769550323,0.4586264491,-0.2341265082,-0.269828856,-0.1781576127,0.142458722,0.182148248,-0.0459706411,0.1569336951,0.0226600543,-0.0574579164,-0.000839628,-0.1019757912,0.6396290064,-0.1003964767,0.0041129384,0.1743919104,0.1856537312,0.291965872,-0.0159218553,-0.0278750435,-0.2039085329,-0.1663276255,-0.182744503,-0.0467879474,0.0416119285,0.0592499562,-0.080940403,-0.15938434,-0.1267841905,0.0207483489,-0.025505567,0.154053241,0.0580775067,-0.9112799168,0.3003269136,0.0890228674,0.5070568919,-0.3329073489,0.0954491869,0.0104669174,-0.0329248682,0.1351575255,0.1240379289,0.1658036113,-0.2133814245,-0.190477863,0.3586421609,-0.1311806291,-0.2136925608,-0.1286634654,-0.144305408,-0.0218649916,0.1743837744,0.0759186745,-0.3373077214,-0.1821484268,0.2263559252,0.1824091822,-0.3305494189,-0.1166630834,0.004144954,-0.1211457402,0.0137829781,-0.148611635,-0.1618931144,0.2823693156,-0.1857763529,-0.0846294984,-0.1780437976,-0.0911253467,0.5624822974,0.0079932036,0.3686125875,0.1240235716,0.3375982046,0.3508354127,0.1489098221,0.3623153567,0.5414234996,-0.1191594228,0.2617897987,0.1720621288,-0.3114438653,0.3362179101,0.2280520052,0.1045742258,-0.4190906286,-0.1923220456,0.1610761583,-0.2195259631,0.2987404168,0.0780677572,0.1364769191,-0.1735477597,-0.2240865976,0.5829634666,0.1190803498,0.0010877317,0.3726539612,0.499840498,-0.0982333496,-0.1306690574,0.5110326409,0.8817679286,0.0613003336,-0.2270235866,0.2880907357,0.2239576429,0.2927213013,-0.6754902601,0.0215901844,-0.1995118707,-0.3058543503,-0.0150679331,-0.0354091041,0.2169420421,0.2565379143,0.261311233,0.1693217754,-0.1373966038,-0.2695969343,-0.1536491364,0.2181302607,-0.3133374155,-0.1865031272,0.2252576202,0.2843100429,0.0592004806,0.2716241181,-0.0343548581,0.0325586051,-0.3010591567,-0.0350145474,-0.07575275,-0.0488687232,-0.0545129143,-0.0395404212,0.0539728105,-0.28198874,-0.0009389883,0.1233882084,0.2557562888,-0.0555337705,-0.0932698995,0.1878999472,-0.3696552515,0.0409896038,0.2919091284,0.0105398186,0.4175280333,-0.2100784928,-0.2193887532,-0.0190252904,0.0579995625,0.1530617476,-0.3923738599,-0.0931582525,-0.0376596749,-0.4144991934,0.0088060228,0.1827116162,-0.0068094004,-0.1136687696,0.170990929,0.150960356,-0.0611261167,0.4131099284,-0.0612229519,0.0377646498,-0.1916013509,0.2504793108,0.4290554225,-0.1598007828,0.3576131165,0.0856814533,-0.0061305007,-0.1509611458,-0.0412523635,0.5110541582,-0.097839646,-0.1124766991,-0.2973004878,-0.2116977423,-0.0785575211,0.2177188993,0.3156519532,-0.1096267328,0.1273468286,-0.3649717569,-0.4957455397,0.3224774301,0.0346124172,0.3332662284,-0.2135457397,0.3155978918,-0.0896363854,0.0993160978,-0.3504059911,0.0299434718,-0.0594397858,0.0497793891,-0.0519680381,0.0121137286,0.2587278187,-0.342249006,0.1581291854,-0.0003127404,-0.2006729543,-0.2028681487,-0.1906140894,0.1195234656,0.1737854481,-0.0475798622,0.2446395457,-0.0849996284,-0.0749365464,-0.049530793,0.0895788521,-0.0978588536,0.0650573745,0.3306813538,0.1396196634,0.0582726188,-0.0491175801,0.2019159049,0.3131592274,0.3580690622,0.0482144915,-0.242749095,-0.0168036968,-0.225228548,-0.3681827784,0.0697309971,-0.0707810074,0.0378898866,0.3394378722,-0.0883171037,0.2087966055,0.1925110668,0.3460713327,0.4225379527,-0.092950888,0.0592902973,0.0492136478,0.2705640197,-0.3063865602,0.0095024779,0.0382755212,0.3211839795,-0.1325514317,0.0830756947,0.3004551828,-0.0230054893,-0.0921783596,0.174843207,0.2036075145,-0.0020415646,0.0513606444,0.2435948998,-0.1754585505,0.0579288416,0.1570551842,-0.0379493274,-0.0833267048,0.228458181,0.0337203778,0.4363600016,0.2827654779,0.1618902683,-0.021522196,-0.1516973525,-0.1129184365,0.3693391681,0.2641839385,0.1251041591,-0.050791353,0.5699244738,-0.1266839802,-0.175521791,-0.2809105814,0.0863989964,0.102239646,0.1649278253,-0.2020112425,-0.1113703921,-0.0892658755,-0.1540822387,-0.1447122544,-0.3010968268,0.2157072425,-0.0265373867,-0.3665010333,-0.9039991498,-0.1263139397,0.1512472183,0.2594661713,-0.0811887905,0.2124182582,0.5510979891,-0.0092023266,0.1270493269,0.5412903428,0.0754256994,0.1766181439,-0.0222536251,-0.0107162101,0.2173730284,0.0039793965,0.2070180923,0.0969842002,-0.0436598621,-0.1632847339,0.3286494315,0.168301329,-0.0872084051,-0.160065949,0.0027589421,0.0158847291,-0.084942922,0.1660191566,0.0624355562,-0.0548020191,-0.0661267564,-0.3546641469,-0.3029305637,-0.0409936942,-0.0520972051,0.1444078386,0.2146542817,0.0307001192,0.1190431789,-0.4020766914,0.2650516033,0.112296477,-0.158799842,-0.311935395,-0.1485044211,-0.6052778363,-0.0056093377,-0.1523032337,-0.5811649561,0.2310092896,0.2647857666,0.2139344662,-0.0295127481,0.1296831369,-0.520601511,-0.0689320862,0.3080603778,-0.6280505657,0.1261766106,-0.2924522758,0.2437844127,-0.0257223658,0.2229900956,0.3024655879,0.3627904356,-0.0518991016,0.1664219201,0.1808859259,0.2209711373,0.0157558359,0.0665927753,-0.0450077802,0.4081068933,0.1880107075,0.2256182134,-0.3655063808,-0.2210374177,-0.2046412081,0.270203948,0.1040105894,0.199499771,0.03194106,-0.0354032926,0.0717182681,0.3342589736,-0.0194073785,-0.3952447176,-0.5124575496,0.100084886,-0.0262607876,0.0636978671,0.3593018651,0.5226396322,0.0657191575,0.123302646,-0.3297067583,-0.3902284801,0.4470281601,-0.0993783996,0.0889891237,-0.01128338,0.2847456634,0.4946372807,-0.1168806255,-0.3762240112,0.0116236703,0.424405396,-0.2406956404,-0.1203697398,0.2497921884,0.3464855552,-0.1600552797,-0.0128241125,0.0151669523,-0.0125525268,-0.0326250643,0.1349397451,0.024730131]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/672","title":"Questions about XSUM ","comments":"Well I couldn't parse what I downloaded.\r\nUnfortunately I think I won't be able to take a look at it this week.\r\nI can try to send you what I got if you want to give it a shot @jbragg \r\nOtherwise feel free to re-run the xsum download script, maybe you'll be luckier than me","body":"Hi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n","comment_length":55,"text":"Questions about XSUM \nHi there \u270b \r\n\r\nI'm looking into your `xsum` dataset and I have several questions on that. \r\nSo here is how I loaded the data: \r\n```\r\n>>> data = datasets.load_dataset('xsum', version='1.0.1')\r\n>>> data['train']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017)\r\n>>> data['test']\r\nDataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333)\r\n```\r\n\r\nThe first issue is, the instance counts don\u2019t match what I see on [the dataset's website](https:\/\/github.com\/EdinburghNLP\/XSum\/tree\/master\/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set)\r\n```\r\n \u2026 training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set.\r\n```\r\nAny thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https:\/\/github.com\/huggingface\/datasets\/pull\/289 (reviewed by @patrickvonplaten) \r\n\r\nAnother issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https:\/\/github.com\/EdinburghNLP\/XSum\/blob\/master\/XSum-Dataset\/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. \r\n\r\nCC @jbragg \r\n\r\n\nWell I couldn't parse what I downloaded.\r\nUnfortunately I think I won't be able to take a look at it this week.\r\nI can try to send you what I got if you want to give it a shot @jbragg \r\nOtherwise feel free to re-run the xsum download script, maybe you'll be luckier than me","embeddings":[-0.0726336688,-0.4036209285,-0.1623655409,0.5180847049,0.3325610161,0.0285643209,0.1227809936,-0.0330050364,0.2053789347,0.2613717318,-0.2468281239,0.2338201851,0.08773496,0.4263722897,0.1107674837,0.0942095369,-0.0165946502,-0.1213084236,-0.3660371602,-0.309211731,-0.0072388113,0.0955581591,-0.0635270998,-0.1014966741,-0.6309723854,-0.010049222,-0.1623434275,0.2903314233,-0.127583608,-0.2089564949,0.2380593866,0.015602544,0.0412374176,0.5506979227,-0.0001053002,-0.0060283383,0.1273684055,-0.0236889776,-0.0839767903,0.2564916909,-0.1792499125,-0.0246550217,0.0088917082,-0.4071148038,-0.0485546254,0.0278062299,-0.1476608664,-0.3758966327,0.2245383859,0.0081966771,0.2218513936,0.1583904177,0.102225408,-0.0368435942,-0.0639064908,-0.1222761944,0.1160336584,0.2276520431,0.4027490616,0.3057997227,0.0729457438,0.1370126009,0.0607239902,-0.1835125089,0.2962157726,0.0647668242,0.3481562734,-0.2427538484,-0.3049997091,0.3073093593,0.8674853444,-0.0344167873,-0.2513758838,-0.0144357095,-0.0167685356,-0.125263676,0.1652304381,0.08433415,0.3414100111,-0.0318596847,-0.2515144646,0.4186394215,-0.2792619467,0.0277597848,-0.18175973,-0.0714393929,-0.0967195705,0.0065125376,-0.2285345793,-0.0505385026,0.0692124069,-0.1019143388,0.1559629291,0.2080772966,-0.2770183086,-0.1285909116,-0.0161555801,0.0022373509,0.2066086084,0.1455343515,0.2617549002,0.063609913,-0.4012528956,0.028139418,0.2450592816,0.2434323132,0.4585989714,-0.5399466157,-0.0215715785,-0.104981795,-0.0611165352,0.0925377235,-0.3793508708,-0.0102072256,-0.0399882905,0.0393612459,-0.1692170203,-0.3824184239,-0.2278565317,0.2536819875,0.0415282957,-0.1091380641,0.1842543185,-0.1190410331,0.0361397453,0.3936980069,0.0177476201,0.0284670573,-0.5414434075,-0.3072095513,-0.3471832573,0.1222479269,-0.0597676337,-0.0438298061,0.0897295177,-0.0929655731,0.3672470152,-0.1047382131,0.2594784498,-0.0157939512,0.0495465323,0.0980077982,-0.3226242363,0.21496512,0.0601884574,0.001447715,-0.0674227029,0.0415629968,-0.0277159326,0.0168904755,-0.1902839094,-0.2040955424,-0.0045190994,0.3222165406,-0.3959135711,-0.2772588134,-0.190050751,0.0054745013,0.1448358148,0.2303758115,-0.1911048293,-0.1483587325,0.0973166674,-0.1408893615,0.1688584089,-0.0693916529,-0.6325304508,0.008455053,-0.0131408861,-0.3519762754,0.0666960701,-0.0030393882,0.0073680533,0.2608046234,0.1243237108,0.2177847475,-0.5206571221,-0.4857227802,-0.1147995442,-0.342882365,-0.1811395437,-0.1095322073,0.1012083739,0.1172326431,0.0287327711,-0.1006724313,0.1512282938,0.0304637682,0.1303051859,-0.2050456256,-0.4009779692,-0.0654452816,-0.0390540399,0.1878517121,-0.075581044,-0.3016083539,-0.0906671733,0.3413063884,0.2055914253,0.0135360723,-0.1134441122,0.0760913491,0.2956166565,-0.1378966272,0.0739980638,-0.1905764043,-0.0058532055,0.3091693521,0.1006015167,0.2124834657,0.7265692949,-0.0317329057,-0.392546922,-0.2147996426,0.0045553711,-0.1560645103,0.201537624,-0.2022472322,-0.0298810154,-0.2052899301,-0.1852127463,0.1675911546,-0.5990707278,0.0399206355,-0.4966748953,0.3045557141,0.0330085382,0.0258521549,0.0667311251,0.2611116469,-0.0271318741,-0.2616860569,-0.0266548637,0.5314072371,0.1615615189,0.0253946465,0.2021260262,0.178230539,0.36476776,-0.1731300801,0.1725413501,-0.2544903457,-0.0349752009,-0.1862795204,-0.4021555483,0.2673924565,-0.1242140606,0.0038795343,-0.067067638,0.0252663959,0.0050549749,-0.1537599415,0.0060087997,0.0304666776,0.2154404968,0.1709766239,-0.2440113276,0.0506175831,0.0227390714,0.393366456,-0.3404999971,-0.2269741297,-0.2323799133,-0.3499913812,-0.2318678945,0.190097928,0.1133133918,0.0082026757,0.4606724083,0.2917123735,0.0924161077,-0.0096252123,-0.0577662997,-0.144268617,0.0375588201,0.0757214352,-0.0849122852,-0.0752281845,-0.0750228241,0.0251599979,-0.145025298,0.0872807056,0.0913452208,0.1519302875,-0.4459275901,-0.3298214078,-0.2245246619,-0.1108564734,0.2284854352,0.1782716215,0.1340694427,-0.3090158999,0.1391239166,0.0357809365,-0.0282403082,-0.029433988,-0.3591500521,0.3002124131,0.0597737134,0.1229233593,0.1016185582,-0.1588439047,-0.167549029,0.2136395574,0.0532518104,0.3412386179,0.3781739175,-0.094384864,-0.0506140962,0.135437876,-0.4069740474,0.0473472811,-0.2130524218,0.2740540206,-0.1605814695,0.0470891409,0.049077455,-0.0007053016,0.2382680178,-0.0045107426,-0.2650096118,-0.0866273046,0.2043105513,-0.2747786641,-0.0735613704,-0.5671487451,-0.1443151832,-0.0556054525,-0.0253853537,0.4845307767,0.2149478942,0.0677579567,-0.132711947,0.2286450118,0.2270271629,-0.1904092878,-0.369502455,-0.8211238384,0.3219831884,-0.1093275174,-0.527531743,0.044315327,0.0050817188,0.3266919255,0.098662585,-0.3984340131,-0.3188937306,0.0212341603,-0.1178080663,0.4910205305,0.2024039477,0.1742156893,-0.1624936908,-0.169287324,-0.2453816682,-0.0643381551,0.2330094725,0.2933532,0.4438322484,-0.3135499656,-0.3200974464,-0.2105186284,0.0678214356,0.1921275854,-0.1336700916,-0.0430319346,-0.0406823121,-0.1121809408,0.1019943058,-0.118985422,0.6109253168,-0.1167218983,-0.0699785948,0.174207136,0.2060505897,0.2247810066,-0.0549101084,0.0715915635,-0.2151780427,-0.1094949618,-0.1889107972,-0.0313990936,-0.0109470468,-0.061919868,-0.0724670514,-0.1073607653,-0.1065226421,-0.102711238,-0.0129423952,0.0882943273,-0.0355589911,-0.9891731739,0.2564759552,0.063881278,0.3727460504,-0.3178672791,-0.0409980677,0.0148173859,-0.1117413417,0.130612433,0.1981731504,0.1457034647,-0.2057919353,-0.206325829,0.3384164274,-0.0943584368,-0.2961410284,-0.0803385079,-0.2180741727,0.0845301822,0.2185201943,0.0147242798,-0.2438407093,-0.0472686999,0.1853288561,0.2361969352,-0.3090729415,-0.1679389626,-0.058499217,-0.041337546,0.0609288216,-0.1247933358,-0.1112174392,0.2661823928,-0.2729489803,-0.1262558848,-0.16690588,-0.0967431068,0.5697981715,0.0501593351,0.4110444784,0.0066462229,0.3036608398,0.364882946,0.081345953,0.2145932764,0.4937660992,-0.1207815334,0.2570339441,0.1627689004,-0.4090260267,0.3568691611,0.181795612,0.0687776431,-0.5050532222,-0.2423398942,0.0817163438,-0.1969736814,0.2669080794,0.1065090969,0.1764202565,-0.0745199472,-0.2677347064,0.5765458941,0.153043285,-0.0216932409,0.3967195749,0.389215827,-0.1153717786,-0.2252329886,0.5142549872,0.9024685621,0.0755639747,-0.2098793685,0.2915862501,0.3819598854,0.336669445,-0.510951221,0.0388089903,-0.1483866125,-0.3224580884,0.0181175917,-0.0931540877,0.1443810761,0.2909357548,0.1441840231,0.0548482053,-0.1882929355,-0.2920118272,-0.1559663415,0.1759005189,-0.2523081303,-0.1252844632,0.280035764,0.2756405473,0.093914561,0.1585783511,-0.0221128333,0.0242550429,-0.3450347483,0.0137958545,-0.0757805854,-0.0314136781,-0.0659127086,0.0366405919,0.0075079296,-0.3132867217,-0.0257364362,0.3328459263,0.207234323,-0.0685253069,-0.1072039679,0.1550856084,-0.3515682817,0.0511082634,0.3374179304,-0.011840119,0.40498963,-0.2082604319,-0.1850056052,-0.031670399,0.0945555791,0.2242094576,-0.3591555953,-0.0039965753,0.0034982651,-0.4560909867,-0.067976363,0.18362014,-0.0690047443,-0.059200611,0.1960066855,0.2452936321,-0.0893047974,0.4629482031,-0.0940147266,0.0361172929,-0.0933116525,0.1915056854,0.447684139,-0.2230202407,0.2656006813,0.0410862751,-0.0035448782,-0.1695640981,0.0310574118,0.4695282876,-0.0643552989,-0.0931396335,-0.2682438493,-0.1900697649,-0.0251848679,0.2308848947,0.2963249385,-0.0065970118,0.0869633332,-0.3571560681,-0.4566163719,0.297208488,0.1031292379,0.4142374098,-0.2592293918,0.2355193347,-0.0219561476,0.0483899824,-0.375313729,0.0327717215,-0.1171204001,0.0454247817,0.1052995175,-0.0350338295,0.2321791947,-0.3884834349,0.1845381409,-0.0055950675,-0.1730097681,-0.2342030555,-0.1040064096,0.0975868925,0.1608510017,-0.0949795544,0.2690057456,-0.0395556241,-0.0355684049,-0.0280995201,0.0543981716,-0.1058267727,0.0968757868,0.3189457655,0.0706667826,0.1440327466,-0.0952517465,0.1414896846,0.1393973827,0.2810757458,0.1553093195,-0.2444820851,0.0287769251,-0.2200663835,-0.2927653193,0.0283166729,-0.0672241524,0.138917312,0.2825185359,-0.0517785065,0.1805142313,0.1599926353,0.3281486332,0.4037590623,-0.1820907593,0.1095534712,-0.0228266194,0.3217216432,-0.2869966626,-0.010161994,0.0545899682,0.3586607277,-0.1319585592,0.0748682246,0.1458770037,-0.0303300228,-0.1269547045,0.1584050953,0.2493472248,-0.0657369792,0.1471383572,0.2147453129,-0.2435144484,-0.0072115585,0.1375520974,-0.0283729956,-0.0748247579,0.2515573204,0.0104173068,0.4689598978,0.1743613333,0.1666101068,0.0173720457,0.000958267,-0.0749312118,0.3695009649,0.1639107615,0.1488026679,0.0254857484,0.588555932,-0.1843401492,-0.0380440019,-0.3076461256,0.0172274951,0.1117455438,0.1733884066,-0.2574014962,-0.1250401437,0.0255654827,-0.1796915829,-0.1197370216,-0.2582989037,0.1674255282,-0.0204450879,-0.3602648377,-0.7811257243,-0.124684535,0.1695304513,0.2597125173,-0.1041670293,0.1274808049,0.565345645,-0.0026227613,0.0757152662,0.5093676448,0.1179083958,0.1717958599,-0.0091006756,0.0523236282,0.2157287747,-0.0404576622,0.1399723887,0.1200163811,-0.0173581094,-0.1546125561,0.290456295,0.2103876621,-0.1007841825,-0.1195348799,0.0577823706,0.0160558969,-0.0858533457,0.0848625898,0.0565755032,-0.1321597844,-0.0987649038,-0.2414764762,-0.3680669069,-0.020138137,-0.1222256571,0.1972981244,0.1806603968,0.0501590483,0.1360905617,-0.3473694921,0.2889632285,0.0706477538,-0.1028144658,-0.2903155982,-0.121972017,-0.6011988521,0.0086711077,-0.0545973517,-0.5773779154,0.1665005088,0.3022072315,0.1711469591,0.059595339,0.045181524,-0.4302569628,-0.1458767802,0.2319048941,-0.6059362292,0.0783951059,-0.1940302551,0.1729171127,-0.0103041446,0.317861259,0.2231746316,0.2317785621,0.0007138273,0.1533498764,0.1520465165,0.156412825,0.2293556184,-0.0070346249,-0.0431755893,0.4805377424,0.1899569184,0.2845982313,-0.2929407954,-0.1684207022,-0.1762913913,0.3193457127,0.2029719502,0.0850899592,-0.0320546515,0.0369670689,0.0861716717,0.302007556,0.0019283418,-0.4383201301,-0.5029208064,-0.00462332,0.0276861936,0.0620464608,0.2641741931,0.482603699,-0.020442225,0.085680522,-0.235221535,-0.3571697474,0.4272537529,-0.1245932803,0.1778803766,-0.0091726473,0.253338486,0.4557929933,-0.0013344435,-0.352602452,-0.0086939652,0.4174803197,-0.2817873359,-0.1516844034,0.2618074417,0.3729294837,-0.1710085124,-0.0333488919,-0.0474460013,-0.0077627301,0.0264335368,0.1544996649,-0.0589766502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/669","title":"How to skip a example when running dataset.map","comments":"Hi @xixiaoyao,\r\nDepending on what you want to do you can:\r\n- use a first step of `filter` to filter out the invalid examples: https:\/\/huggingface.co\/docs\/datasets\/processing.html#filtering-rows-select-and-filter\r\n- or directly detect the invalid examples inside the callable used with `map` and return them unchanged or even remove them at the same time if you are using `map` in batched mode. Here is an example where we use `map` in batched mode to add new rows on the fly but you can also use it to remove examples on the fly (that's what `filter` actually do under-the-hood): https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset","body":"in processing func, I process examples and detect some invalid examples, which I did not want it to be added into train dataset. However I did not find how to skip this recognized invalid example when doing dataset.map. ","comment_length":95,"text":"How to skip a example when running dataset.map\nin processing func, I process examples and detect some invalid examples, which I did not want it to be added into train dataset. However I did not find how to skip this recognized invalid example when doing dataset.map. \nHi @xixiaoyao,\r\nDepending on what you want to do you can:\r\n- use a first step of `filter` to filter out the invalid examples: https:\/\/huggingface.co\/docs\/datasets\/processing.html#filtering-rows-select-and-filter\r\n- or directly detect the invalid examples inside the callable used with `map` and return them unchanged or even remove them at the same time if you are using `map` in batched mode. Here is an example where we use `map` in batched mode to add new rows on the fly but you can also use it to remove examples on the fly (that's what `filter` actually do under-the-hood): https:\/\/huggingface.co\/docs\/datasets\/processing.html#augmenting-the-dataset","embeddings":[-0.3242636323,-0.2246204466,0.031353876,0.0002581653,0.0896153748,0.3332220614,0.0300699323,0.1055706441,0.1422973573,0.2520426214,0.6323099732,0.4480103552,-0.2600845993,0.4908131957,0.0828728154,-0.0856683031,-0.0504853204,0.0724819228,0.1984967291,0.1779389232,-0.3641427457,-0.0961724371,-0.49770239,0.0496413969,-0.3158500195,-0.2935382724,-0.0769769698,0.140801385,-0.0265658572,-0.2878863811,0.0960213989,-0.0466132201,0.0982665867,0.504729867,-0.0001258959,0.1196550578,0.3082424402,-0.0945355967,0.1854922622,-0.3107355237,-0.3292454779,0.1670228541,-0.1726574451,-0.2398432195,-0.0762515888,-0.0612976849,0.0840196982,-0.0997303799,0.4658529162,0.0540934764,0.0529715009,0.2320326865,-0.3186951876,0.0486841574,-0.1116908491,0.409217,0.0898966044,0.2354432493,0.1143718734,-0.0532132871,0.0675040632,0.231179595,-0.1761514395,-0.0097621782,0.1850004047,-0.1228409261,0.1621088386,-0.3764846027,0.0962039158,0.4493292272,0.22457169,-0.126456663,-0.2365824878,-0.3658840358,-0.0463603511,-0.2754533589,-0.016118804,-0.1472744197,-0.015307745,0.4170869887,-0.6674239635,-0.1911608577,-0.1037272587,0.061202269,0.2733073235,-0.179596588,-0.0818134025,0.2964392006,0.329721123,0.1335158646,-0.2808516026,-0.0747990385,0.3179302514,0.3472799659,0.0003408075,-0.3308746219,0.0062255142,0.197625652,0.0249322187,-0.325789988,0.0555800684,0.1102719903,0.1363181025,0.5397353172,0.0191038679,-0.2823091745,0.0233992338,0.3064346015,0.2565121949,0.1218458638,-0.0133249564,0.1166886836,-0.2119525075,-0.1256945282,-0.2269021273,0.2246917337,0.0821375549,-0.1516428441,-0.1834163517,-0.002640435,-0.4451543987,-0.0977713987,0.1742839068,0.491938293,0.0618027374,-0.2413509339,-0.3540638685,0.1922273487,-0.0949275643,-0.1513017118,-0.0967249796,0.0498009995,-0.1144316345,0.1790896356,0.159690544,-0.0792608336,0.4232955277,0.068015404,-0.2709647715,0.1603313386,0.689712286,-0.1699372232,0.1788068265,0.3412056863,-0.0605335385,-0.057755243,0.2295355052,0.3670894206,-0.0047913282,0.35683617,-0.3024683893,-0.4309857786,0.3180296123,-0.0142080281,-0.0673328266,0.2039246112,-0.5382909179,0.346475184,0.2190382332,-0.4069853425,-0.0032906057,0.1302217692,-0.5322643518,-0.1984508038,0.1176695004,0.1650928259,-0.3112281263,-0.2914988697,-0.2650072873,0.0359472446,0.1033580974,-0.0010721558,-0.3380072117,0.480039537,-0.3783349693,0.1835429221,0.2497921884,-0.2121672928,0.0564303584,0.0568875559,-0.1763331294,0.0893616974,-0.3198416233,-0.1334590614,0.3183126152,-0.1414415985,-0.045973625,0.1065726951,-0.3901832104,-0.089728564,0.0181747284,-0.0168863405,0.5136342645,-0.2013853341,-0.2183645666,0.1387318671,-0.2506180406,-0.3810956776,0.2124197483,-0.138144061,-0.103997089,0.1810707599,0.3804059625,-0.3077709079,0.1553096473,-0.3345043361,-0.0102507612,0.1763610542,0.0401231162,-0.3803938627,0.2575185299,-0.5252299905,-0.1265999824,-0.1873489618,-0.0143144531,-0.0288681462,-0.0561559647,-0.3807054758,0.097371012,-0.1782945395,-0.3164161742,0.0432511419,-0.2925288081,0.3105837703,-0.1860637665,-0.025866719,0.2752487659,-0.0649267435,0.0075345528,-0.0372882746,0.2593930662,-0.4081422985,0.0443684459,0.3251231015,0.4812693298,0.1775559485,-0.0790168941,0.0932360664,0.1937626302,-0.3268751204,-0.2090979517,0.0282353032,-0.0245819092,0.0280555412,-0.2407762855,0.3548715115,0.1204803661,0.125122726,0.1026158258,0.1530056596,0.0357429534,-0.1511326283,0.0913808271,-0.1911781132,0.0256277956,0.058819484,0.1974120438,0.1561806053,-0.1957836747,-0.1475310922,0.4238241911,0.0040207244,-0.1040827483,0.2210841328,0.0002531631,-0.0520118102,0.5704846382,0.0326652527,0.0618723854,0.1499921381,0.0290911198,0.0220050458,-0.3287629783,0.2473180741,0.0675081685,0.1988274008,0.4481800199,-0.335295558,0.00959526,0.1830748469,-0.1451720893,-0.057996545,0.3105287254,0.0239243004,-0.3376735747,0.1957659423,-0.2525347173,-0.1397378743,0.1195069104,0.0803660005,-0.1787724644,-0.1134715974,0.0619538501,0.4838314056,-0.2872804701,0.2964291573,0.0118209347,0.12412972,-0.1833330095,-0.2632946372,-0.0218229517,0.1122780442,-0.2322319001,0.061067149,0.1845856011,0.4620588124,0.5137461424,-0.0903162137,-0.1922500432,0.0471119173,-0.5107018948,0.1798775494,-0.0647185072,0.2261238694,0.2627539635,0.2396679968,-0.5291053057,0.0256425068,-0.1005804241,-0.2393880635,-0.1765174717,-0.1198151186,0.025914276,0.3449905217,-0.2298613489,-0.2000429034,-0.0995236337,-0.0450547412,0.1051680222,-0.2208750397,0.1166531295,-0.1009434164,0.0337652154,0.219216004,-0.0252638608,-0.3246285617,-0.0130740255,-0.3057721257,-0.0206060037,-0.2947101891,0.048565127,0.2048910707,-0.3035925031,0.0695242584,0.3210887015,-0.174884811,-0.4121911526,-0.1730241477,0.0444512405,-0.1805033684,-0.3683572114,0.5422813892,0.0257188454,0.073286593,-0.310159713,-0.3047010899,0.2872392237,0.2302464396,0.0790903941,0.4050255418,0.586245358,0.1559972316,0.6632351279,0.3587724268,-0.0044967737,0.2946740687,-0.0173923522,0.1108123139,0.0867998078,0.2357931286,0.0060353619,0.0331217349,-0.3669233918,-0.0593147725,0.0174054652,-0.0560547821,0.1778513342,-0.1333879977,-0.3015755713,-0.249792248,-0.0237090494,-0.1485795677,0.4746419191,0.1500706971,0.0345524251,-0.2720657587,0.0499043278,0.1825569719,0.1444784701,0.8970304132,0.1220846847,-0.3651359677,-0.200261876,-0.3813802302,0.2161007077,0.2633254528,0.030687619,-0.1829452366,0.1107100174,-0.1326038241,0.0719926506,0.8909837008,-0.4783500433,0.0679053962,-0.0793978274,-0.0425561592,-0.225123316,0.1874660552,-0.0673202872,0.6205625534,0.3868528605,0.4097590446,-0.0332399793,-0.1684534997,0.2244223207,0.4525845349,0.003934267,-0.1690651923,-0.1096534654,-0.058669854,-0.6159629226,0.2032065243,0.633030653,-0.2209208906,-0.2785764337,-0.3541158736,0.0799758211,0.0908182934,0.2393754274,0.3062521517,0.2124631405,-0.1452292502,0.1064713523,0.4780971408,0.0251858085,0.2462866753,0.5393096209,-0.0041547855,-0.6362016797,-0.1552197784,-0.0606455542,0.4769300222,-0.0617463402,0.0176224522,0.1434142888,0.2348604202,0.0230429545,-0.0582349598,0.3424856067,0.1758184433,0.045240093,-0.4140235782,-0.050090231,0.07909289,0.2493262738,-0.1851271093,0.6250059009,-0.1639325619,-0.1471605748,0.3089538515,-0.0112283397,0.6379778385,0.0749557242,-0.2474688143,0.0789313018,0.0372072421,0.193800658,0.0915261507,-0.02088085,-0.0204039905,-0.3294067383,-0.1144296899,-0.1585920751,0.3285665512,0.5887830257,-0.2945955098,0.2747734487,-0.1022062153,0.2315357178,-0.0078366911,-0.0948732868,0.3148842156,-0.114477925,-0.2315273285,-0.0201951973,0.3014897406,-0.0872593746,0.0357923955,-0.1560101509,0.2789089084,-0.4633378088,-0.1052746028,-0.2805894613,-0.6895771623,-0.1791909486,0.1165178418,0.275136888,0.2830939591,0.318808347,0.1778908074,0.0506364666,-0.2302640378,0.0631311759,0.0429806598,0.2077715248,-0.326872617,-0.1661536992,0.2756370902,0.0375220105,-0.1649466604,0.3764308095,-0.2958546281,-0.1479957849,0.2756742835,-0.2241286486,0.048444014,0.0225730352,-0.154954046,-0.1398323178,0.0068104211,-0.2099109888,-0.0316440463,-0.0245200582,-0.0333436243,0.0926187113,-0.1169913784,-0.1770906448,-0.0702324659,0.0840760767,-0.2639919221,-0.1625937968,0.1618475914,-0.0930477232,-0.0219867043,-0.1291128844,-0.0113199167,0.2623629272,-0.1325615197,0.0652209297,0.3482711613,0.0519440211,0.1709774733,-0.054350052,-0.0329397619,-0.073522307,-0.2989520431,-0.2066054493,-0.5582759976,0.1039377004,0.0480968989,0.1627387255,-0.0097172428,0.4578390718,-0.0320746303,-0.0337234139,-0.149313733,0.1816614121,-0.1403558254,0.1874862611,0.1428201348,0.5469575524,-0.0235602986,-0.1670020819,-0.1679525375,0.3657305241,-0.0339384452,-0.0519330651,-0.1060867012,0.1871075034,-0.0148711931,-0.1825888604,0.2769848704,0.1012560204,-0.0096931485,-0.1023050025,0.2399615943,-0.095684737,0.0613521747,0.1268480569,0.0009975276,0.0823125914,-0.1285839975,-0.1799128354,-0.1966050118,0.0378272161,-0.1578034461,0.0638717934,0.1090375334,-0.0243338794,0.2529958189,-0.419449836,0.1464189887,-0.2020234317,-0.1831784397,0.0295828078,0.0717012361,0.1250089258,0.0034409596,-0.036871925,-0.1992926896,0.1470134556,0.4785037041,0.1193073243,-0.3371192813,0.0830580667,0.6689921021,-0.0708642527,-0.018205585,0.1826017797,0.0799477324,0.2994882762,-0.0443321094,0.1571815759,0.0131447371,-0.2660858035,0.4661112428,0.2899324894,-0.454511404,0.1229568347,-0.0141612971,0.1543002278,0.385658294,0.101633966,0.1284423918,0.4472121894,-0.0749505609,0.1344840825,0.2135424316,-0.725826323,-0.0063069402,0.4943154752,-0.0596663728,0.1720003337,0.2485917956,0.0685112029,-0.0997904465,-0.4781296551,-0.2441489697,0.023225937,-0.0691381767,0.084631905,-0.003051874,0.1252713352,-0.0031254089,-0.0998128206,0.2085620165,0.0370079055,0.5780152678,-0.1765360534,-0.1605189294,-0.278567493,-0.2988343835,-0.0486001261,0.19282417,0.145066753,-0.0741175637,-0.0620713569,-0.1806123853,0.0076460396,0.1089215502,0.0184160918,-0.1273633093,-0.1469281018,0.1265466064,-0.2374815345,0.125973925,-0.0354826711,0.3049452007,-0.2280288488,-0.1189911515,0.218514815,-0.0575866699,0.0158192739,-0.0139555465,-0.0117213912,0.3811348081,-0.3684956729,0.2288073748,-0.1111717373,0.0909270421,0.0603187606,0.011090707,0.0346961506,0.0386897027,0.4817609191,-0.4550876915,0.0002898185,0.0687901229,0.0042935866,-0.3503245711,0.0151158748,0.3317013979,0.1513875872,-0.3204599619,-0.1301911175,-0.4943649173,-0.0342449397,-0.3479117155,0.1900551319,0.3387843668,-0.31171,-0.1235571802,0.1725456566,0.1584008187,-0.161484316,-0.0366931371,0.2663267255,-0.1398996115,-0.249784559,-0.3644639552,0.1490165889,0.1726312488,-0.458494246,0.2356077582,-0.1520160288,-0.1531417519,-0.2025348544,0.0590690821,0.1058135033,-0.0214644596,0.3100109994,0.3010406792,0.242040053,-0.2407736033,0.2433021963,0.338647157,0.2236692905,-0.2236421704,0.1351250261,-0.1482826322,0.3675678372,-0.2840276957,0.0636148676,-0.0427500941,-0.1796211153,0.3713649511,0.0970496386,-0.3833548129,0.2957875431,-0.2796857953,0.1975485235,0.1138487458,0.0459220819,-0.0012813259,0.0344123766,-0.0164142381,-0.0105190603,0.583653748,-0.6836977601,-0.0483500138,-0.353382647,0.1826316565,0.1267497241,-0.4262392819,-0.6577043533,-0.3809611499,0.3768140972,0.0052215438,0.1006298587,0.3586829603,-0.0389041528,-0.1048538685,0.0579774268,0.3994214237,-0.0799558461,-0.0925761089,0.3459765017,-0.1557781547]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/669","title":"How to skip a example when running dataset.map","comments":"Closing this one.\r\nFeel free to re-open if you have other questions","body":"in processing func, I process examples and detect some invalid examples, which I did not want it to be added into train dataset. However I did not find how to skip this recognized invalid example when doing dataset.map. ","comment_length":12,"text":"How to skip a example when running dataset.map\nin processing func, I process examples and detect some invalid examples, which I did not want it to be added into train dataset. However I did not find how to skip this recognized invalid example when doing dataset.map. \nClosing this one.\r\nFeel free to re-open if you have other questions","embeddings":[-0.3116004169,-0.275143981,-0.044677373,0.0600629747,0.0858501866,0.2209694535,0.1170850918,0.0814867765,-0.0010126617,0.2320596129,0.7053828835,0.5303474665,-0.1657244712,0.2367559224,-0.0817117542,-0.1146364436,-0.0652721226,0.0616959669,0.2764957845,0.0338925421,-0.3509443104,-0.1140678823,-0.5746421218,0.097663939,-0.3303995132,-0.2800964713,-0.1267341077,0.2793287635,-0.0066571953,-0.3406547904,0.071965605,-0.1316940635,0.1424662173,0.4724344909,-0.000114269,-0.0090453243,0.2979634404,-0.0036793526,0.2733316123,-0.1840907782,-0.4169169068,0.2425941974,-0.2081405073,-0.3448901474,-0.0329260565,0.0175015908,0.178944096,-0.1509838998,0.4465960562,0.1765137315,0.1833696067,0.0438028798,-0.3004302084,0.0126147699,-0.0886531174,0.1837870181,0.1450846642,0.2166594565,0.1131799072,-0.0327155292,0.0332633182,0.1301446855,-0.1642156094,-0.148995325,0.0455880016,-0.219285205,0.2935702503,-0.3491054177,0.0852647126,0.2879982889,0.4023421705,-0.0522066094,0.0264740791,-0.0289817341,-0.1769023538,-0.2983714938,-0.0744488239,-0.1096068323,-0.0568666533,0.4329603612,-0.4413017929,-0.0922038332,-0.2592450976,0.0205906462,0.1587096155,-0.0384874307,-0.0532226972,0.2891343534,0.245820567,0.101094082,-0.1814101934,-0.1674471498,0.3781923056,0.1650206,0.0721630678,-0.4246366918,-0.0196412131,0.2232523561,0.0214418881,-0.2295525521,0.0268699415,0.0767953545,0.2791249156,0.5526400208,0.0059377886,-0.3542367816,-0.0444413684,0.2830202281,0.1372119188,-0.1490476429,-0.3205339909,0.1058036834,-0.3154727221,-0.2231092602,-0.0937445164,0.3639958799,0.0691171885,-0.2046124488,-0.1678655148,0.0204505492,-0.4849840105,-0.0616907068,0.1237588674,0.5533001423,0.0795432851,-0.13781932,-0.4026146531,0.0906885043,-0.0957671851,0.0863246992,-0.1235156581,0.1572559178,-0.1371674538,0.0494029298,0.0384660177,0.115911223,0.3035206199,0.0752757117,-0.3341506422,0.0895846188,0.6386831999,-0.2071260959,0.2160035968,0.2927184701,-0.1976246387,-0.0327954851,0.1785866767,0.4218839407,0.0036151337,0.5489127636,-0.3189277947,-0.3612024486,0.2330603004,0.1342833191,-0.1025677174,0.0370960608,-0.2820402384,0.1949703991,0.1580609381,-0.4047278464,0.0325804166,0.0445603095,-0.4094491005,-0.3968361616,-0.0321188383,0.0144897075,-0.3238740861,-0.0893682167,-0.3215657175,-0.0038260794,0.2544269562,-0.0053212829,-0.3519068658,0.4528498054,-0.4007963538,0.2714186311,0.349776715,-0.1893214434,0.009005826,0.1048287377,-0.2397967577,0.0848695114,-0.2678946257,-0.0999000818,0.0889650285,-0.2230821252,-0.1116780415,0.0074695121,-0.3792745173,-0.1720227599,-0.1059191003,-0.0101919491,0.7131124735,-0.1120503917,-0.0277937055,0.0759813115,-0.2024205774,-0.3447369039,0.2662272453,-0.1067549661,-0.2917796969,0.1463820487,0.4644774199,-0.4844546914,0.0561735705,-0.2964679301,0.1593330204,0.106228739,0.1102653369,-0.4690727293,0.3297270536,-0.4108855426,-0.163550511,-0.2216679901,0.1060729474,0.0723680332,0.1111005321,-0.2268821746,-0.0914246961,-0.1599628031,-0.2522238493,0.0974985957,-0.322347492,0.133223176,-0.019745009,-0.1716745198,0.0896144658,0.0863844231,-0.0747556239,-0.0985119045,0.2816791534,-0.3747140467,-0.0223763268,0.3735152483,0.5748207569,0.1624298394,-0.0706471577,0.0167278256,0.0794183835,-0.4032210708,-0.226887539,0.0248622969,-0.0716826692,0.0122008203,-0.2149142772,0.2462469935,0.1251076907,-0.0997257456,0.1627145708,0.0915305838,-0.0575253628,-0.306545347,0.1374838948,-0.0466962457,-0.0128422147,0.1992975026,0.1697935313,0.3546488285,-0.3145249784,-0.1168277264,0.5818113089,0.0139426347,-0.0912211835,0.2397665381,0.1171409935,0.03141094,0.5463516712,0.0562620051,0.2080324292,0.2348751426,0.1132725701,-0.0337234586,-0.2824081481,0.0871570781,0.1310851723,0.1556521803,0.5318427682,-0.2807664275,-0.0465164185,0.0725394189,-0.1272139549,-0.1315254718,0.3731694221,-0.0324681215,-0.2325488627,0.1655535698,-0.3040997684,-0.0613241233,0.0088989818,0.1290162504,-0.0578388348,-0.1112251207,0.1335957497,0.3207428157,-0.3179260194,0.2465324253,-0.0823405236,0.1503309011,-0.3041701615,-0.2173756957,0.0361299105,0.0493815616,-0.3074530363,0.1849097311,0.289822489,0.5032613873,0.4655097127,-0.022707535,-0.1006134152,0.1561005563,-0.4140819013,0.1365730911,-0.0293401759,0.1013196483,0.1015605927,0.2501784265,-0.6526506543,-0.0416160338,-0.1458315104,-0.1304936707,-0.0653425083,-0.1766010374,0.1687204391,0.1805300415,-0.1470572799,-0.4450690746,-0.1630945057,-0.1517759264,0.1123489961,-0.2219880223,0.2011135966,-0.0156771708,0.0428924002,0.2565562129,0.140794605,-0.3045994341,-0.0701673254,-0.3241927922,-0.0231326874,-0.2812818587,-0.0218514036,0.2121603191,-0.1461578757,0.1757969707,0.213076815,-0.2187995166,-0.3734529316,-0.0538807362,0.0605674647,-0.2117323577,-0.4013731778,0.4800862968,-0.0669076294,-0.0940288752,-0.356867522,-0.198976323,0.2725277841,0.156558156,0.2206348926,0.4197090268,0.4813550115,0.1184545159,0.759568274,0.4253959358,-0.0929639488,0.2549205124,-0.0280270353,-0.0921181515,0.1535868496,0.1980287284,0.0742754117,0.1008088216,-0.4482856393,0.0131158018,-0.0337807983,-0.1913268864,0.1160980016,-0.187563926,-0.2651685476,-0.2659244537,0.0169933029,-0.2806333899,0.4536403418,0.0817143843,-0.037755575,-0.3367842436,0.1838424206,0.2118087411,0.2109882534,0.7742440701,-0.0561313406,-0.4023346305,-0.1609561145,-0.2483397871,0.2167953402,0.2354632765,0.0238073878,-0.3214406967,0.0786977038,-0.132645905,0.0987576172,0.8940936327,-0.5955334306,0.0356054567,0.0652713031,0.0525507182,-0.0777351484,0.2334711254,-0.0537482239,0.6335684657,0.4227157533,0.301617831,0.1182494387,-0.2128224969,0.2583592832,0.2404031008,0.1046002209,-0.3142590523,-0.0312948525,-0.0432624444,-0.4051774144,0.2089471817,0.6485642195,-0.1963675618,-0.3819374442,-0.4394861162,0.0167536009,0.0017953898,0.2448793799,0.1892276406,0.1247112006,-0.1313396394,-0.0356980897,0.3018968999,0.1840536892,0.2614485621,0.3742477,0.1013040468,-0.4087122083,0.0061317016,-0.0797286704,0.5711025,0.0561901145,-0.1295699477,0.0406014621,0.1058165878,0.0244886763,-0.0407822765,0.2157162577,0.1700292677,0.0337740853,-0.3944417238,0.0509468541,0.0765936449,0.2261774838,-0.2815113664,0.5985379219,-0.4092262387,-0.0722212791,0.3379733562,0.0224004202,0.9034163952,0.0776852295,-0.3682831526,-0.122042194,-0.0128247337,0.0576925687,-0.0057572839,0.0021011152,-0.1196448356,-0.3289815784,-0.0611648113,-0.2163529098,0.3921758533,0.5727797747,-0.4000933468,0.066054225,-0.0367698781,0.319837451,0.0592120439,0.0691557899,0.2633821368,0.0478400104,-0.1181192547,0.0917188376,0.3535741866,-0.1800764501,-0.1631826162,-0.3346775472,0.1155678034,-0.3778954446,-0.0658099428,-0.1488494426,-0.6693279743,-0.027516963,-0.0493760929,0.2114768028,0.3682818413,0.426556617,0.2401660681,0.0601790138,-0.2309576422,0.1091192439,0.1093316972,0.204513222,-0.2705055773,-0.1617751867,0.2649917901,-0.020248862,-0.206182003,0.3760117888,-0.3116810024,-0.1890715212,0.2195124775,-0.2735117078,0.2840162516,-0.0273963064,-0.1715146899,-0.1398837268,-0.1841932684,-0.1463617831,0.121413298,-0.046665702,0.0636741221,0.2380730063,-0.0784925371,-0.1415310651,-0.0215498917,-0.0296832621,-0.067753233,-0.1018759459,0.1749740541,-0.0800902098,-0.0590972081,-0.2780593634,0.1010715663,0.0788937435,-0.021280352,0.0270720441,0.2048002183,0.1014457271,0.1911581904,-0.0180974752,-0.0049989712,0.0364446975,-0.4128344953,-0.1649052799,-0.5675908327,0.0696656257,-0.1487079412,0.3412941396,-0.1999352574,0.4308121502,0.0118782287,-0.1546693891,-0.3177521825,0.1883583963,-0.2911718786,0.1883616298,0.0021720477,0.4214789569,-0.2230183482,-0.2103959173,0.0113074435,0.4316955805,-0.0786424428,-0.1806723177,0.0721601024,0.1513021886,0.0218739472,-0.2524221539,0.228352502,0.0019864896,-0.0641413257,0.0814041197,0.0703032017,-0.2032799572,0.0290407389,0.2021375448,-0.0647261739,0.0148992296,-0.1282149851,-0.2480137199,-0.2126791477,0.141280815,-0.1847322881,-0.0093686162,0.1224418506,-0.0897384807,0.3123546839,-0.411776185,0.191290766,-0.1053033695,-0.2069860846,0.0964993313,0.0078680571,0.0423127078,-0.1220804751,-0.0796078667,-0.3112097085,0.2107894421,0.4466127157,0.2593601346,-0.2108139098,0.1834502816,0.6765562892,-0.0488072,-0.0583763011,0.1524925232,0.1595039517,0.3135410547,0.0694776848,0.1787582636,0.0871458128,-0.243190825,0.448300153,0.3453752398,-0.393204391,0.0850325674,-0.0567521751,0.2108356804,0.4469623864,0.1580300331,0.2145018876,0.3821800947,0.12946257,0.2104356438,0.1743154675,-0.5929300189,0.0659130067,0.4681944549,-0.2252958268,0.1540966481,0.3110831976,-0.0381588154,-0.0576830395,-0.4060981572,-0.227451086,-0.0349437855,-0.1657201201,0.0719667822,-0.0110761393,0.0105613088,-0.0690591335,-0.0965049118,0.0804078951,0.1080088094,0.5347546935,-0.0809473619,-0.1814576983,-0.0469844341,-0.1143631265,0.0040171803,0.146187827,0.0402435176,-0.0882473215,-0.0615627579,-0.1049113646,-0.0961183533,0.0694926679,0.1273569763,-0.0234412905,-0.2570350468,0.1598544121,0.0016512793,0.0557812564,-0.0062837121,0.30498752,-0.0789516866,-0.0157587212,0.196573928,0.0825002268,-0.063638255,0.0937113687,0.0783941522,0.3235256672,-0.4185680151,0.1278359294,-0.0264912006,0.0112397727,0.0632530153,-0.054210376,-0.0834181011,-0.1994279474,0.4450578094,-0.4438588917,0.1173106059,0.123884812,0.0765091479,-0.4493886828,0.1153320298,0.219560042,0.2876440883,-0.4472844005,-0.2774794698,-0.4475502968,0.0620142817,-0.2411218435,0.115953505,0.3168976307,-0.3243355453,-0.0696633086,0.0942135975,0.1178605855,-0.0571494624,-0.1065032259,0.4366444051,-0.1842484027,-0.0976533145,-0.2280449569,0.0904600024,0.0703790635,-0.3840627968,0.3422659338,-0.1254540235,0.0079767369,-0.1493956745,-0.1038270369,0.2283150554,0.0131886546,0.4199818969,0.2108830363,0.2916242182,-0.2451508492,0.3391368389,0.3155142665,0.1847156286,-0.0772146732,0.1414700001,0.0187563412,0.2832843363,-0.1715433449,0.1939759701,0.02531063,-0.3500695229,0.1939036548,0.1967999339,-0.2280448675,0.1515455544,-0.1228141561,0.2668715715,0.0897774249,0.0543835834,-0.0872373208,0.0276299827,-0.0993484259,-0.0247383378,0.5152356625,-0.4615027606,-0.1840762347,-0.2531863451,0.3410769403,0.0608647577,-0.1754593402,-0.5684073567,-0.4742026031,0.4363250732,-0.0186499283,-0.0709869638,0.3063427806,0.0312821381,-0.1013344675,0.1141058356,0.2170111388,-0.2094145566,-0.0246781316,0.2360495776,-0.2284669429]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/667","title":"Loss not decrease with Datasets and Transformers","comments":"And I tested it on T5ForConditionalGeneration, that works no problem.","body":"HI,\r\n\r\nThe following script is used to fine-tune a BertForSequenceClassification model on SST2.\r\n\r\nThe script is adapted from [this colab](https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb) that presents an example of fine-tuning BertForQuestionAnswering using squad dataset. In that colab, loss works fine. When I adapt it to SST2, the loss fails to decrease as it should. I attach the adapted script below and appreciate anyone pointing out what I miss?\r\n\r\n```python\r\nimport torch\r\nfrom datasets import load_dataset\r\nfrom transformers import BertForSequenceClassification\r\nfrom transformers import BertTokenizerFast\r\n# Load our training dataset and tokenizer\r\ndataset = load_dataset(\"glue\", 'sst2')\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\ndel dataset[\"test\"] # let's remove it in this demo\r\n\r\n# Tokenize our training dataset\r\ndef convert_to_features(example_batch):\r\n encodings = tokenizer(example_batch[\"sentence\"])\r\n encodings.update({\"labels\": example_batch[\"label\"]})\r\n return encodings\r\n\r\nencoded_dataset = dataset.map(convert_to_features, batched=True)\r\n# Format our dataset to outputs torch.Tensor to train a pytorch model\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'labels']\r\nencoded_dataset.set_format(type='torch', columns=columns)\r\n\r\n# Instantiate a PyTorch Dataloader around our dataset\r\n# Let's do dynamic batching (pad on the fly with our own collate_fn)\r\ndef collate_fn(examples):\r\n return tokenizer.pad(examples, return_tensors='pt')\r\n\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset['train'], collate_fn=collate_fn, batch_size=8)\r\n# Now let's train our model\r\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\r\n# Let's load a pretrained Bert model and a simple optimizer\r\nmodel = BertForSequenceClassification.from_pretrained('bert-base-cased', return_dict=True)\r\noptimizer = torch.optim.Adam(model.parameters(), lr=1e-5)\r\nmodel.train().to(device)\r\nfor i, batch in enumerate(dataloader):\r\n batch.to(device)\r\n outputs = model(**batch)\r\n loss = outputs.loss\r\n loss.backward()\r\n optimizer.step()\r\n model.zero_grad()\r\n print(f'Step {i} - loss: {loss:.3}')\r\n\r\n\r\n```\r\nIn case needed.\r\n\r\n- datasets == 1.0.2\r\n- transformers == 3.2.0","comment_length":10,"text":"Loss not decrease with Datasets and Transformers\nHI,\r\n\r\nThe following script is used to fine-tune a BertForSequenceClassification model on SST2.\r\n\r\nThe script is adapted from [this colab](https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb) that presents an example of fine-tuning BertForQuestionAnswering using squad dataset. In that colab, loss works fine. When I adapt it to SST2, the loss fails to decrease as it should. I attach the adapted script below and appreciate anyone pointing out what I miss?\r\n\r\n```python\r\nimport torch\r\nfrom datasets import load_dataset\r\nfrom transformers import BertForSequenceClassification\r\nfrom transformers import BertTokenizerFast\r\n# Load our training dataset and tokenizer\r\ndataset = load_dataset(\"glue\", 'sst2')\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\ndel dataset[\"test\"] # let's remove it in this demo\r\n\r\n# Tokenize our training dataset\r\ndef convert_to_features(example_batch):\r\n encodings = tokenizer(example_batch[\"sentence\"])\r\n encodings.update({\"labels\": example_batch[\"label\"]})\r\n return encodings\r\n\r\nencoded_dataset = dataset.map(convert_to_features, batched=True)\r\n# Format our dataset to outputs torch.Tensor to train a pytorch model\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'labels']\r\nencoded_dataset.set_format(type='torch', columns=columns)\r\n\r\n# Instantiate a PyTorch Dataloader around our dataset\r\n# Let's do dynamic batching (pad on the fly with our own collate_fn)\r\ndef collate_fn(examples):\r\n return tokenizer.pad(examples, return_tensors='pt')\r\n\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset['train'], collate_fn=collate_fn, batch_size=8)\r\n# Now let's train our model\r\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\r\n# Let's load a pretrained Bert model and a simple optimizer\r\nmodel = BertForSequenceClassification.from_pretrained('bert-base-cased', return_dict=True)\r\noptimizer = torch.optim.Adam(model.parameters(), lr=1e-5)\r\nmodel.train().to(device)\r\nfor i, batch in enumerate(dataloader):\r\n batch.to(device)\r\n outputs = model(**batch)\r\n loss = outputs.loss\r\n loss.backward()\r\n optimizer.step()\r\n model.zero_grad()\r\n print(f'Step {i} - loss: {loss:.3}')\r\n\r\n\r\n```\r\nIn case needed.\r\n\r\n- datasets == 1.0.2\r\n- transformers == 3.2.0\nAnd I tested it on T5ForConditionalGeneration, that works no problem.","embeddings":[0.0143272402,-0.0178403128,0.0766257048,0.2709603906,0.1909327656,-0.2114582062,0.3216443956,0.1833657622,-0.3186629415,0.1440708935,-0.1105745062,0.2251771837,0.0432985947,-0.3711665571,-0.3412014246,-0.5152077675,0.092767708,0.1171909422,-0.4170707762,-0.3085789084,0.2463035882,0.148686111,-0.5372670889,-0.2332571298,-0.4022983611,0.2082428634,-0.0268376134,0.08499185,-0.1229775921,-0.1508571804,0.4350038171,0.2246644944,0.6234452128,0.5011128783,-0.0001253695,0.0167352278,0.0884683058,-0.1851132661,-0.0964059457,0.1887830347,0.0095063001,-0.0720067918,0.0785151944,-0.2393475473,-0.2013340443,0.5764876008,-0.0589542091,-0.1579336375,0.5676104426,0.0713826045,0.0511577688,0.0994233415,-0.3086315989,0.0978987366,-0.0312880613,0.1299947351,-0.2217846066,-0.1588303745,0.142397061,-0.2430086881,-0.2492354661,0.244248569,-0.0914825723,0.1282944679,0.4990986884,0.2129201293,-0.3659015596,-0.2715988755,0.2734735608,0.0496383347,0.3472603858,-0.4256628454,-0.4216351211,-0.1504353583,0.1279190332,0.0961998925,0.0465043746,-0.1955449581,0.0177183952,0.0964226946,-0.1567200869,-0.2207983583,-0.4685472846,0.0264698416,0.091796726,0.4098714292,-0.0062107272,-0.1192714646,-0.1143856943,-0.3491779566,-0.1651540697,0.0337856859,0.1241988763,0.1486433446,-0.5537847281,-0.1846330315,-0.1117318422,-0.1745158732,-0.027535744,-0.3231819272,0.0780724362,0.101626128,0.0821538866,-0.1564653516,0.1735979915,0.3648107648,0.0943130329,0.4102594852,0.1208897606,0.1123721153,-0.2615919411,0.0032639012,-0.1972545534,-0.1704994887,0.3708522618,0.0576061383,0.3201502264,-0.192386955,-0.4567261934,-0.0554664135,-0.3279343843,-0.2216428518,0.0916512385,0.2830949724,-0.1445198804,0.2163469046,-0.1008516848,0.2693802416,-0.3416719139,0.0706471205,-0.0472738519,-0.0374250114,-0.5051249862,0.0723484084,0.3726485968,0.217414096,0.2313282341,0.0451216027,0.121455878,0.25222525,-0.0618577823,-0.0985316113,0.3288626969,-0.1315164566,-0.2866594195,0.0891968161,-0.2206839323,0.0552942418,0.0595232844,0.1287122816,-0.3794058859,0.2103689462,-0.0756791532,-0.0523015335,-0.0269694477,-0.0535664596,-0.1378819942,0.3341746628,0.1426265985,-0.1533270776,0.1410789341,-0.6094964743,-0.3426013589,-0.0242808796,0.5197733641,0.5062796474,-0.4620134532,-0.3242632151,0.2939985991,0.2669239342,0.5352307558,0.565192163,-0.1079585701,0.0349414162,0.0557246432,0.2512503266,0.1858744025,-0.2093686908,-0.7334695458,-0.0830294788,-0.0441935062,0.0869677886,0.2217092514,-0.1311743557,0.4703834951,-0.2120390534,-0.1244111434,-0.2151397169,-0.1037929729,0.1159090176,-0.2668857276,0.1395716518,0.4136345685,0.0328764208,0.1747836322,-0.0320379063,-0.1260010004,0.6209384799,0.2226442546,-0.1818389893,0.2899461389,0.0643967465,0.1717638671,-0.2174842805,0.2140316069,0.0806370005,-0.3901189268,0.2001280487,-0.1324909627,0.1447695047,0.014918928,0.0626084656,0.0884095952,-0.2703787982,-0.11363592,-0.1285360157,0.0465300828,0.0782191381,-0.1111471504,-0.4716425538,-0.183214739,-0.119209975,0.169731304,0.0719178542,-0.0869846269,0.3454029858,0.02591029,-0.4336415529,0.0098216785,0.1193503886,0.3469047248,-0.0527056307,-0.1413051337,0.19168441,-0.0627031326,0.1314253807,0.2388147414,0.3168720007,0.4315247536,-0.3996690214,0.277092427,0.4996348917,-0.0656077415,0.1053966805,-0.2424380183,-0.0199280195,0.010152882,0.1996157169,-0.0707076713,-0.2372710705,-0.172502771,-0.1236380488,-0.3080772758,0.1097653583,-0.3041394651,-0.1481756121,0.2156849056,0.06784226,-0.1514199674,0.1748232841,0.3401047885,0.0692100078,0.0954236612,0.2409140021,-0.4496682882,-0.2235210687,0.1370737255,-0.016748609,0.4678155184,-0.0258350112,0.1285744309,-0.1111882403,0.2357772738,-0.1046387479,-0.0667033121,0.301696986,-0.1883014292,0.1050804108,0.3168638051,0.3701637387,-0.0432829969,-0.0838378072,-0.0384021848,0.173660785,-0.2928595245,0.3218758106,-0.2582605183,0.06121286,-0.2138753384,-0.0527619421,0.1186532378,0.1473302543,-0.062692143,-0.0735690594,-0.047246471,0.2227206081,0.0347405151,0.3342996836,0.3281894624,-0.218313694,-0.0666736886,-0.0572618097,-0.0280947909,0.0324906334,-0.04106858,0.0276342891,-0.2657498717,-0.0736551881,-0.1680519432,-0.2551649809,-0.3171067834,0.070417881,-0.1802714169,-0.1311718225,0.2053000927,-0.0421734005,-0.3728771806,-0.2056106925,-0.0110588847,-0.3334241807,0.0961776599,0.1580109149,-0.1058861613,-0.2090536803,0.0334984735,-0.3507479429,0.1322992742,0.1283809543,-0.1099347919,-0.1493009031,-0.1089204922,0.3558456004,-0.0532984845,-0.0475715585,0.029614469,0.2345341742,0.1703339964,-0.3580476344,0.4893873334,-0.0219131485,-0.224332422,-0.1758491248,0.0223402623,0.1377399266,-0.0775631294,0.0666945949,-0.2440136224,-0.2023453563,0.1336150765,-0.1854813993,0.3551799953,0.3544382453,0.1523952782,0.110573411,-0.1908690035,-0.4509736896,0.3485206068,0.3832609057,0.4323945642,-0.4216374457,0.1823228896,0.0361409448,0.9876065254,0.1043779552,-0.335876137,0.2449721843,0.1232649684,-0.0912315398,0.040585354,-0.361766547,0.5099081993,-0.0015320354,-0.0853111371,0.3287678063,0.0312036648,-0.2091062069,0.1018045247,-0.2216373682,-0.1103068367,-0.367012471,0.0729520917,0.1993713081,0.2293976694,-0.0368091129,0.3337689936,-0.0824843347,-0.1033090726,0.3512706459,0.0184064731,0.4582268,-0.1077620238,0.1760492623,-0.1072182655,-0.6214439869,0.1384925842,-0.0775877237,0.3409398794,0.0871914327,-0.204132542,0.0871126875,0.2505132258,0.2127575874,0.2498399019,0.038217023,-0.0485703163,-0.098330684,-0.0534547642,-0.1513855755,-0.2904684246,0.0043621697,0.4999237955,-0.0523273572,-0.1654864997,-0.4805053771,0.2167249769,-0.1697978079,0.0787682086,0.0343557857,-0.2242307961,0.0256754383,0.116729416,0.0166058,0.2248081416,0.1711089611,0.0104533378,0.1089624017,0.1017682403,-0.3740820289,0.3025341332,0.363335222,0.2641854882,0.2990294993,0.1077369452,0.1964434385,0.1423639506,0.2262047976,0.1970766187,-0.0208579265,-0.6519159675,0.363923192,0.2096342891,0.1794057339,0.1377776414,-0.0091089094,-0.0136547517,-0.2079299986,0.2204765379,-0.0087606674,0.1700918525,0.4786550701,-0.004627212,-0.367867738,-0.0532867685,-0.0504152589,-0.2648195624,-0.1895358711,-0.5693151951,0.0968562588,-0.3616544902,0.1745099425,0.3774795532,0.838336587,0.1199332252,0.0641972199,0.0058644917,0.0753912255,0.4150811732,0.0422243476,0.1986550838,-0.1886021048,-0.1785337627,0.0570546202,0.0519798659,0.0577930138,0.0543221347,-0.5023739338,0.5196650028,0.0769692138,0.1709811687,-0.053226877,0.1329180747,0.3378891349,-0.2948259115,-0.2281223685,0.0177722629,-0.0332323089,0.2830355763,-0.0332401991,-0.3221073747,-0.0351764448,-0.1473837197,-0.1400850415,0.0983341783,-0.5406102538,-0.1973623633,0.2189487964,-0.4878490865,0.6094714999,0.3612021208,0.7112462521,0.0389166251,-0.1341971755,0.183658585,-0.186841324,0.2206083387,0.1224508435,-0.1901767403,0.1150003299,0.0256076939,-0.0704022422,-0.0071214689,0.2948194742,0.0246202163,-0.2678570747,-0.0589653403,0.5732697845,-0.3536780477,0.1465212852,-0.0981545821,-0.1048581824,-0.173830092,0.0196666308,0.2717900574,-0.1337009668,0.4722105563,0.0465045832,-0.2123426348,0.150770694,0.5336145759,-0.0612938702,0.0350701958,0.6399751306,-0.1759637892,-0.0593745634,-0.0515857302,-0.0456874929,0.0495316312,-0.3411386013,-0.1074682102,0.1047032177,-0.0610815473,0.2422991097,0.3580413461,-0.1100450754,-0.032171078,-0.3433737457,0.1485666037,-0.366133213,-0.0861077085,0.1647069603,0.2435697466,-0.065301165,0.1125687733,0.0855050385,0.0734752342,-0.1801758409,-0.1088634431,0.1182270423,0.2003917545,0.0096008237,-0.2287469655,-0.0130355228,-0.2042816132,-0.1301714182,0.0387509689,0.2149035335,-0.0377595238,-0.1776063442,0.1479622573,0.0929816887,-0.064505212,-0.5151131749,0.1278841048,0.1166149378,-0.2510861754,0.3547608256,0.1306277364,0.2909607887,0.4648828208,0.3471973538,-0.3416910768,0.0412849672,-0.1069640145,0.2297144681,0.2724525332,0.3789080083,0.0959770605,0.2298777401,-0.1649802178,-0.042753011,0.2534694672,0.3290734887,0.1927930117,0.145529747,-0.3761540651,0.011019744,-0.062625967,-0.0039872234,0.0892027095,0.0288749635,-0.3305234313,0.1826047301,0.125868395,-0.0624787807,-0.056751553,0.0470640659,0.0025412571,-0.0099788541,0.0793805793,-0.3615135252,0.1841420829,0.1266905367,0.274528712,0.3068202138,0.2341707349,0.2034870833,0.0538994893,-0.019959705,0.1516923457,0.0307156537,-0.3423266709,-0.0957271904,0.0079999426,0.0819253549,0.7272573113,-0.2339152992,0.2487430125,-0.0474247448,0.1425618976,-0.2965925038,0.0486089401,0.0628920048,-0.0674518943,-0.1082302108,0.3336431384,0.0007175052,-0.5508161783,-0.0054127891,0.223246038,-0.0829267278,-0.3166797161,-0.0854546726,-0.1897695959,0.3159942031,-0.1717508584,-0.331094861,0.2240343094,0.0462157093,0.043256022,-0.1881859452,0.0043717236,-0.3820967972,0.2529363334,0.1457211226,0.1615885645,0.3201903701,-0.0255887676,0.1071969867,-0.1764329076,0.5512974858,0.2408434451,0.3611789644,0.0062454245,-0.2476912141,0.1683620811,-0.0984279662,0.0391284972,0.0988013521,-0.0983741879,-0.3236695528,0.0431243964,-0.0247212276,-0.0108729219,0.1131593958,-0.032062728,-0.071500212,-0.3820296228,0.2489697933,-0.0715474486,0.0952049121,-0.2893677652,-0.1540643126,-0.1538222879,0.0290071573,0.0377821773,-0.2452782094,0.0288736541,-0.2547991276,-0.0192459822,-0.0545470491,0.554214716,0.0983995497,0.0764159039,-0.5310490727,0.1737270206,-0.3281563818,0.123385109,0.0902653039,0.0631550327,-0.0918572843,0.1053631976,0.0074244281,0.2421836853,0.370909512,-0.2622203231,0.0256211516,0.235778302,-0.2690033913,0.2432828844,-0.2464671135,0.081178382,0.1025464237,-0.2114072591,0.1775928438,0.0955260321,-0.0788858458,0.0357864387,0.1249653921,-0.1148098409,-0.0979764313,0.4438372552,0.2022013813,0.2865839899,-0.1549033374,-0.0948730633,-0.1554531306,-0.2424680293,-0.2454557866,-0.5271041393,-0.0060199285,0.413356632,-0.0477304682,-0.0831045061,-0.1928850114,0.1409992278,0.2417922467,-0.0685426518,-0.3989034295,0.1612616777,-0.0422265828,0.1907717735,-0.4427527487,0.0727856234,-0.0308055207,0.1064899042,-0.5212318897,-0.4688406587,0.6822171211,-0.6715806723,-0.2621474564,-0.4065579474,0.2129824311,-0.112140663,-0.154456526,-0.7287846208,0.074896045,0.0446135402,0.2008208632,-0.319816649,0.1755495518,0.163738519,0.2320660204,0.0097400155,0.5042396188,0.1847340614,-0.039670147,0.081794709,-0.1719207764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/667","title":"Loss not decrease with Datasets and Transformers","comments":"Hi did you manage to fix your issue ?\r\n\r\nIf so feel free to share your fix and close this thread","body":"HI,\r\n\r\nThe following script is used to fine-tune a BertForSequenceClassification model on SST2.\r\n\r\nThe script is adapted from [this colab](https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb) that presents an example of fine-tuning BertForQuestionAnswering using squad dataset. In that colab, loss works fine. When I adapt it to SST2, the loss fails to decrease as it should. I attach the adapted script below and appreciate anyone pointing out what I miss?\r\n\r\n```python\r\nimport torch\r\nfrom datasets import load_dataset\r\nfrom transformers import BertForSequenceClassification\r\nfrom transformers import BertTokenizerFast\r\n# Load our training dataset and tokenizer\r\ndataset = load_dataset(\"glue\", 'sst2')\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\ndel dataset[\"test\"] # let's remove it in this demo\r\n\r\n# Tokenize our training dataset\r\ndef convert_to_features(example_batch):\r\n encodings = tokenizer(example_batch[\"sentence\"])\r\n encodings.update({\"labels\": example_batch[\"label\"]})\r\n return encodings\r\n\r\nencoded_dataset = dataset.map(convert_to_features, batched=True)\r\n# Format our dataset to outputs torch.Tensor to train a pytorch model\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'labels']\r\nencoded_dataset.set_format(type='torch', columns=columns)\r\n\r\n# Instantiate a PyTorch Dataloader around our dataset\r\n# Let's do dynamic batching (pad on the fly with our own collate_fn)\r\ndef collate_fn(examples):\r\n return tokenizer.pad(examples, return_tensors='pt')\r\n\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset['train'], collate_fn=collate_fn, batch_size=8)\r\n# Now let's train our model\r\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\r\n# Let's load a pretrained Bert model and a simple optimizer\r\nmodel = BertForSequenceClassification.from_pretrained('bert-base-cased', return_dict=True)\r\noptimizer = torch.optim.Adam(model.parameters(), lr=1e-5)\r\nmodel.train().to(device)\r\nfor i, batch in enumerate(dataloader):\r\n batch.to(device)\r\n outputs = model(**batch)\r\n loss = outputs.loss\r\n loss.backward()\r\n optimizer.step()\r\n model.zero_grad()\r\n print(f'Step {i} - loss: {loss:.3}')\r\n\r\n\r\n```\r\nIn case needed.\r\n\r\n- datasets == 1.0.2\r\n- transformers == 3.2.0","comment_length":21,"text":"Loss not decrease with Datasets and Transformers\nHI,\r\n\r\nThe following script is used to fine-tune a BertForSequenceClassification model on SST2.\r\n\r\nThe script is adapted from [this colab](https:\/\/colab.research.google.com\/github\/huggingface\/datasets\/blob\/master\/notebooks\/Overview.ipynb) that presents an example of fine-tuning BertForQuestionAnswering using squad dataset. In that colab, loss works fine. When I adapt it to SST2, the loss fails to decrease as it should. I attach the adapted script below and appreciate anyone pointing out what I miss?\r\n\r\n```python\r\nimport torch\r\nfrom datasets import load_dataset\r\nfrom transformers import BertForSequenceClassification\r\nfrom transformers import BertTokenizerFast\r\n# Load our training dataset and tokenizer\r\ndataset = load_dataset(\"glue\", 'sst2')\r\ntokenizer = BertTokenizerFast.from_pretrained('bert-base-cased')\r\ndel dataset[\"test\"] # let's remove it in this demo\r\n\r\n# Tokenize our training dataset\r\ndef convert_to_features(example_batch):\r\n encodings = tokenizer(example_batch[\"sentence\"])\r\n encodings.update({\"labels\": example_batch[\"label\"]})\r\n return encodings\r\n\r\nencoded_dataset = dataset.map(convert_to_features, batched=True)\r\n# Format our dataset to outputs torch.Tensor to train a pytorch model\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'labels']\r\nencoded_dataset.set_format(type='torch', columns=columns)\r\n\r\n# Instantiate a PyTorch Dataloader around our dataset\r\n# Let's do dynamic batching (pad on the fly with our own collate_fn)\r\ndef collate_fn(examples):\r\n return tokenizer.pad(examples, return_tensors='pt')\r\n\r\ndataloader = torch.utils.data.DataLoader(encoded_dataset['train'], collate_fn=collate_fn, batch_size=8)\r\n# Now let's train our model\r\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\r\n# Let's load a pretrained Bert model and a simple optimizer\r\nmodel = BertForSequenceClassification.from_pretrained('bert-base-cased', return_dict=True)\r\noptimizer = torch.optim.Adam(model.parameters(), lr=1e-5)\r\nmodel.train().to(device)\r\nfor i, batch in enumerate(dataloader):\r\n batch.to(device)\r\n outputs = model(**batch)\r\n loss = outputs.loss\r\n loss.backward()\r\n optimizer.step()\r\n model.zero_grad()\r\n print(f'Step {i} - loss: {loss:.3}')\r\n\r\n\r\n```\r\nIn case needed.\r\n\r\n- datasets == 1.0.2\r\n- transformers == 3.2.0\nHi did you manage to fix your issue ?\r\n\r\nIf so feel free to share your fix and close this thread","embeddings":[0.0143272402,-0.0178403128,0.0766257048,0.2709603906,0.1909327656,-0.2114582062,0.3216443956,0.1833657622,-0.3186629415,0.1440708935,-0.1105745062,0.2251771837,0.0432985947,-0.3711665571,-0.3412014246,-0.5152077675,0.092767708,0.1171909422,-0.4170707762,-0.3085789084,0.2463035882,0.148686111,-0.5372670889,-0.2332571298,-0.4022983611,0.2082428634,-0.0268376134,0.08499185,-0.1229775921,-0.1508571804,0.4350038171,0.2246644944,0.6234452128,0.5011128783,-0.0001253695,0.0167352278,0.0884683058,-0.1851132661,-0.0964059457,0.1887830347,0.0095063001,-0.0720067918,0.0785151944,-0.2393475473,-0.2013340443,0.5764876008,-0.0589542091,-0.1579336375,0.5676104426,0.0713826045,0.0511577688,0.0994233415,-0.3086315989,0.0978987366,-0.0312880613,0.1299947351,-0.2217846066,-0.1588303745,0.142397061,-0.2430086881,-0.2492354661,0.244248569,-0.0914825723,0.1282944679,0.4990986884,0.2129201293,-0.3659015596,-0.2715988755,0.2734735608,0.0496383347,0.3472603858,-0.4256628454,-0.4216351211,-0.1504353583,0.1279190332,0.0961998925,0.0465043746,-0.1955449581,0.0177183952,0.0964226946,-0.1567200869,-0.2207983583,-0.4685472846,0.0264698416,0.091796726,0.4098714292,-0.0062107272,-0.1192714646,-0.1143856943,-0.3491779566,-0.1651540697,0.0337856859,0.1241988763,0.1486433446,-0.5537847281,-0.1846330315,-0.1117318422,-0.1745158732,-0.027535744,-0.3231819272,0.0780724362,0.101626128,0.0821538866,-0.1564653516,0.1735979915,0.3648107648,0.0943130329,0.4102594852,0.1208897606,0.1123721153,-0.2615919411,0.0032639012,-0.1972545534,-0.1704994887,0.3708522618,0.0576061383,0.3201502264,-0.192386955,-0.4567261934,-0.0554664135,-0.3279343843,-0.2216428518,0.0916512385,0.2830949724,-0.1445198804,0.2163469046,-0.1008516848,0.2693802416,-0.3416719139,0.0706471205,-0.0472738519,-0.0374250114,-0.5051249862,0.0723484084,0.3726485968,0.217414096,0.2313282341,0.0451216027,0.121455878,0.25222525,-0.0618577823,-0.0985316113,0.3288626969,-0.1315164566,-0.2866594195,0.0891968161,-0.2206839323,0.0552942418,0.0595232844,0.1287122816,-0.3794058859,0.2103689462,-0.0756791532,-0.0523015335,-0.0269694477,-0.0535664596,-0.1378819942,0.3341746628,0.1426265985,-0.1533270776,0.1410789341,-0.6094964743,-0.3426013589,-0.0242808796,0.5197733641,0.5062796474,-0.4620134532,-0.3242632151,0.2939985991,0.2669239342,0.5352307558,0.565192163,-0.1079585701,0.0349414162,0.0557246432,0.2512503266,0.1858744025,-0.2093686908,-0.7334695458,-0.0830294788,-0.0441935062,0.0869677886,0.2217092514,-0.1311743557,0.4703834951,-0.2120390534,-0.1244111434,-0.2151397169,-0.1037929729,0.1159090176,-0.2668857276,0.1395716518,0.4136345685,0.0328764208,0.1747836322,-0.0320379063,-0.1260010004,0.6209384799,0.2226442546,-0.1818389893,0.2899461389,0.0643967465,0.1717638671,-0.2174842805,0.2140316069,0.0806370005,-0.3901189268,0.2001280487,-0.1324909627,0.1447695047,0.014918928,0.0626084656,0.0884095952,-0.2703787982,-0.11363592,-0.1285360157,0.0465300828,0.0782191381,-0.1111471504,-0.4716425538,-0.183214739,-0.119209975,0.169731304,0.0719178542,-0.0869846269,0.3454029858,0.02591029,-0.4336415529,0.0098216785,0.1193503886,0.3469047248,-0.0527056307,-0.1413051337,0.19168441,-0.0627031326,0.1314253807,0.2388147414,0.3168720007,0.4315247536,-0.3996690214,0.277092427,0.4996348917,-0.0656077415,0.1053966805,-0.2424380183,-0.0199280195,0.010152882,0.1996157169,-0.0707076713,-0.2372710705,-0.172502771,-0.1236380488,-0.3080772758,0.1097653583,-0.3041394651,-0.1481756121,0.2156849056,0.06784226,-0.1514199674,0.1748232841,0.3401047885,0.0692100078,0.0954236612,0.2409140021,-0.4496682882,-0.2235210687,0.1370737255,-0.016748609,0.4678155184,-0.0258350112,0.1285744309,-0.1111882403,0.2357772738,-0.1046387479,-0.0667033121,0.301696986,-0.1883014292,0.1050804108,0.3168638051,0.3701637387,-0.0432829969,-0.0838378072,-0.0384021848,0.173660785,-0.2928595245,0.3218758106,-0.2582605183,0.06121286,-0.2138753384,-0.0527619421,0.1186532378,0.1473302543,-0.062692143,-0.0735690594,-0.047246471,0.2227206081,0.0347405151,0.3342996836,0.3281894624,-0.218313694,-0.0666736886,-0.0572618097,-0.0280947909,0.0324906334,-0.04106858,0.0276342891,-0.2657498717,-0.0736551881,-0.1680519432,-0.2551649809,-0.3171067834,0.070417881,-0.1802714169,-0.1311718225,0.2053000927,-0.0421734005,-0.3728771806,-0.2056106925,-0.0110588847,-0.3334241807,0.0961776599,0.1580109149,-0.1058861613,-0.2090536803,0.0334984735,-0.3507479429,0.1322992742,0.1283809543,-0.1099347919,-0.1493009031,-0.1089204922,0.3558456004,-0.0532984845,-0.0475715585,0.029614469,0.2345341742,0.1703339964,-0.3580476344,0.4893873334,-0.0219131485,-0.224332422,-0.1758491248,0.0223402623,0.1377399266,-0.0775631294,0.0666945949,-0.2440136224,-0.2023453563,0.1336150765,-0.1854813993,0.3551799953,0.3544382453,0.1523952782,0.110573411,-0.1908690035,-0.4509736896,0.3485206068,0.3832609057,0.4323945642,-0.4216374457,0.1823228896,0.0361409448,0.9876065254,0.1043779552,-0.335876137,0.2449721843,0.1232649684,-0.0912315398,0.040585354,-0.361766547,0.5099081993,-0.0015320354,-0.0853111371,0.3287678063,0.0312036648,-0.2091062069,0.1018045247,-0.2216373682,-0.1103068367,-0.367012471,0.0729520917,0.1993713081,0.2293976694,-0.0368091129,0.3337689936,-0.0824843347,-0.1033090726,0.3512706459,0.0184064731,0.4582268,-0.1077620238,0.1760492623,-0.1072182655,-0.6214439869,0.1384925842,-0.0775877237,0.3409398794,0.0871914327,-0.204132542,0.0871126875,0.2505132258,0.2127575874,0.2498399019,0.038217023,-0.0485703163,-0.098330684,-0.0534547642,-0.1513855755,-0.2904684246,0.0043621697,0.4999237955,-0.0523273572,-0.1654864997,-0.4805053771,0.2167249769,-0.1697978079,0.0787682086,0.0343557857,-0.2242307961,0.0256754383,0.116729416,0.0166058,0.2248081416,0.1711089611,0.0104533378,0.1089624017,0.1017682403,-0.3740820289,0.3025341332,0.363335222,0.2641854882,0.2990294993,0.1077369452,0.1964434385,0.1423639506,0.2262047976,0.1970766187,-0.0208579265,-0.6519159675,0.363923192,0.2096342891,0.1794057339,0.1377776414,-0.0091089094,-0.0136547517,-0.2079299986,0.2204765379,-0.0087606674,0.1700918525,0.4786550701,-0.004627212,-0.367867738,-0.0532867685,-0.0504152589,-0.2648195624,-0.1895358711,-0.5693151951,0.0968562588,-0.3616544902,0.1745099425,0.3774795532,0.838336587,0.1199332252,0.0641972199,0.0058644917,0.0753912255,0.4150811732,0.0422243476,0.1986550838,-0.1886021048,-0.1785337627,0.0570546202,0.0519798659,0.0577930138,0.0543221347,-0.5023739338,0.5196650028,0.0769692138,0.1709811687,-0.053226877,0.1329180747,0.3378891349,-0.2948259115,-0.2281223685,0.0177722629,-0.0332323089,0.2830355763,-0.0332401991,-0.3221073747,-0.0351764448,-0.1473837197,-0.1400850415,0.0983341783,-0.5406102538,-0.1973623633,0.2189487964,-0.4878490865,0.6094714999,0.3612021208,0.7112462521,0.0389166251,-0.1341971755,0.183658585,-0.186841324,0.2206083387,0.1224508435,-0.1901767403,0.1150003299,0.0256076939,-0.0704022422,-0.0071214689,0.2948194742,0.0246202163,-0.2678570747,-0.0589653403,0.5732697845,-0.3536780477,0.1465212852,-0.0981545821,-0.1048581824,-0.173830092,0.0196666308,0.2717900574,-0.1337009668,0.4722105563,0.0465045832,-0.2123426348,0.150770694,0.5336145759,-0.0612938702,0.0350701958,0.6399751306,-0.1759637892,-0.0593745634,-0.0515857302,-0.0456874929,0.0495316312,-0.3411386013,-0.1074682102,0.1047032177,-0.0610815473,0.2422991097,0.3580413461,-0.1100450754,-0.032171078,-0.3433737457,0.1485666037,-0.366133213,-0.0861077085,0.1647069603,0.2435697466,-0.065301165,0.1125687733,0.0855050385,0.0734752342,-0.1801758409,-0.1088634431,0.1182270423,0.2003917545,0.0096008237,-0.2287469655,-0.0130355228,-0.2042816132,-0.1301714182,0.0387509689,0.2149035335,-0.0377595238,-0.1776063442,0.1479622573,0.0929816887,-0.064505212,-0.5151131749,0.1278841048,0.1166149378,-0.2510861754,0.3547608256,0.1306277364,0.2909607887,0.4648828208,0.3471973538,-0.3416910768,0.0412849672,-0.1069640145,0.2297144681,0.2724525332,0.3789080083,0.0959770605,0.2298777401,-0.1649802178,-0.042753011,0.2534694672,0.3290734887,0.1927930117,0.145529747,-0.3761540651,0.011019744,-0.062625967,-0.0039872234,0.0892027095,0.0288749635,-0.3305234313,0.1826047301,0.125868395,-0.0624787807,-0.056751553,0.0470640659,0.0025412571,-0.0099788541,0.0793805793,-0.3615135252,0.1841420829,0.1266905367,0.274528712,0.3068202138,0.2341707349,0.2034870833,0.0538994893,-0.019959705,0.1516923457,0.0307156537,-0.3423266709,-0.0957271904,0.0079999426,0.0819253549,0.7272573113,-0.2339152992,0.2487430125,-0.0474247448,0.1425618976,-0.2965925038,0.0486089401,0.0628920048,-0.0674518943,-0.1082302108,0.3336431384,0.0007175052,-0.5508161783,-0.0054127891,0.223246038,-0.0829267278,-0.3166797161,-0.0854546726,-0.1897695959,0.3159942031,-0.1717508584,-0.331094861,0.2240343094,0.0462157093,0.043256022,-0.1881859452,0.0043717236,-0.3820967972,0.2529363334,0.1457211226,0.1615885645,0.3201903701,-0.0255887676,0.1071969867,-0.1764329076,0.5512974858,0.2408434451,0.3611789644,0.0062454245,-0.2476912141,0.1683620811,-0.0984279662,0.0391284972,0.0988013521,-0.0983741879,-0.3236695528,0.0431243964,-0.0247212276,-0.0108729219,0.1131593958,-0.032062728,-0.071500212,-0.3820296228,0.2489697933,-0.0715474486,0.0952049121,-0.2893677652,-0.1540643126,-0.1538222879,0.0290071573,0.0377821773,-0.2452782094,0.0288736541,-0.2547991276,-0.0192459822,-0.0545470491,0.554214716,0.0983995497,0.0764159039,-0.5310490727,0.1737270206,-0.3281563818,0.123385109,0.0902653039,0.0631550327,-0.0918572843,0.1053631976,0.0074244281,0.2421836853,0.370909512,-0.2622203231,0.0256211516,0.235778302,-0.2690033913,0.2432828844,-0.2464671135,0.081178382,0.1025464237,-0.2114072591,0.1775928438,0.0955260321,-0.0788858458,0.0357864387,0.1249653921,-0.1148098409,-0.0979764313,0.4438372552,0.2022013813,0.2865839899,-0.1549033374,-0.0948730633,-0.1554531306,-0.2424680293,-0.2454557866,-0.5271041393,-0.0060199285,0.413356632,-0.0477304682,-0.0831045061,-0.1928850114,0.1409992278,0.2417922467,-0.0685426518,-0.3989034295,0.1612616777,-0.0422265828,0.1907717735,-0.4427527487,0.0727856234,-0.0308055207,0.1064899042,-0.5212318897,-0.4688406587,0.6822171211,-0.6715806723,-0.2621474564,-0.4065579474,0.2129824311,-0.112140663,-0.154456526,-0.7287846208,0.074896045,0.0446135402,0.2008208632,-0.319816649,0.1755495518,0.163738519,0.2320660204,0.0097400155,0.5042396188,0.1847340614,-0.039670147,0.081794709,-0.1719207764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/666","title":"Does both 'bookcorpus' and 'wikipedia' belong to the same datasets which Google used for pretraining BERT?","comments":"No they are other similar copies but they are not provided by the official Bert models authors.","body":"","comment_length":17,"text":"Does both 'bookcorpus' and 'wikipedia' belong to the same datasets which Google used for pretraining BERT?\n\nNo they are other similar copies but they are not provided by the official Bert models authors.","embeddings":[0.1324191689,-0.133234635,-0.0841302946,0.4101325572,-0.0619927756,0.0686285198,0.4990635812,0.0508505479,0.0587276779,-0.1818362176,-0.5121811628,0.0105857374,0.1604425162,0.2634949088,0.3741164505,-0.062693581,0.2953804731,0.0685094073,-0.1061497778,-0.4837571979,-0.0717290044,0.2138784379,0.0865202546,0.2500123978,0.1909606457,0.2310380787,-0.2693659067,-0.0304537974,-0.0282236747,0.0310507566,0.3777728379,0.3160231113,0.2182165533,0.4636310935,-0.0001177422,-0.0648744255,-0.0596583672,-0.0256551411,-0.0905918926,-0.0317620933,-0.5987688303,0.4022060633,0.0205680188,-0.2447787374,-0.2479808629,0.1904209703,0.0798728392,0.0083656013,0.1955294758,0.096196942,0.1572679281,0.0176671911,0.1153461337,-0.0210660174,0.4745382667,-0.0441604853,0.0986416191,0.2570824623,0.2947520018,-0.007567313,0.1934588701,0.2982078195,0.0136623448,0.1792534292,0.3279303014,0.2027475834,0.0356970541,-0.228901118,0.1980826408,0.6075033545,0.5808110237,-0.1521254778,-0.2595596015,0.3440983891,0.1413084418,0.3872638941,-0.0057436214,0.3685973585,0.1439121813,0.060142111,-0.3490225077,-0.4425522089,-0.2131686807,0.0481958874,-0.4004027843,0.1453387141,-0.0261951387,-0.0515784025,0.1161745638,-0.0598331168,-0.1769890487,-0.4608443677,0.1334209293,0.0316440314,0.2590478063,-0.4295482039,-0.4468905926,0.4916749299,0.0198773798,-0.2651945353,0.1151548177,-0.2458391339,-0.4626718462,-0.0033402438,0.5650666356,-0.1606397033,-0.0972068533,0.4177864194,0.0059132604,0.0520161316,-0.4645185471,0.1407152563,-0.022649074,0.1596690118,-0.0140337851,-0.496568799,-0.0201715361,-0.46300596,-0.2084348649,0.3940341175,-0.2711075246,-0.3099861145,-0.0238772482,0.221752882,-0.2402871698,0.0059972312,-0.0714803264,0.0521042496,-0.3137901425,-0.0417638756,-0.1370812207,0.1919546574,-0.1478152424,0.0813206807,0.1714694798,-0.1659353524,0.2277069837,0.0033455594,-0.1874548644,0.1649449617,0.1627699435,-0.2344545126,0.2292414308,0.0988225415,0.1444210112,0.3086495101,-0.2100583613,0.0957047045,-0.2624313235,0.2983239293,-0.2653523982,0.1203858927,0.0544190705,0.1465215832,-0.0216831155,-0.4595587254,-0.1376622766,0.3717737198,0.2018166035,-0.3964059949,-0.0214513801,-0.0741342455,0.0003057899,-0.0498957746,0.1014835611,0.1136588976,-0.4338351786,-0.0991016999,-0.0695913061,0.1668549776,0.0413709059,0.4769982398,-0.2406729162,-0.0016950829,0.1298739165,0.2944982648,-0.0079249712,0.1406478584,-0.2155560255,-0.0801782757,0.6388680339,-0.0573148802,-0.2702763379,0.0628421158,-0.121490337,0.0706074089,-0.0278564934,0.1578110456,-0.0948843509,-0.0128700202,-0.0184743311,-0.3641971946,0.2479721308,0.2232214957,0.2096707523,-0.202590391,-0.1239563301,0.3146232069,0.0930859819,-0.20497033,-0.1000239849,0.3123936951,0.2197786123,0.135308668,0.3516609073,-0.1904551238,-0.0783343837,-0.103320241,-0.4503236711,0.1138187423,0.4208244681,-0.2247956991,-0.3842028081,-0.3467467427,0.1069406494,-0.269567281,0.0403116345,-0.1109490916,-0.1928634495,0.1469338685,-0.1691978276,-0.3153448999,-0.1299065202,-0.0747417286,-0.2905255556,0.4140481353,-0.3119814992,0.2114654034,-0.0152958324,0.2411210537,-0.2127108276,0.1486355513,0.1532109082,-0.0725458041,0.1858248562,0.2266049385,0.4529054761,0.2760826647,0.5355110765,-0.5498377085,-0.0086452104,0.2895117104,0.1784520298,-0.0589447953,-0.5726770163,0.0789673626,0.4249208272,0.0028086833,0.281206578,-0.0891249552,-0.1369204968,-0.0390577242,0.0601683818,-0.1927031428,-0.022814218,0.2331892848,0.2672102153,0.2973287106,-0.295699805,0.1688204557,0.1861185879,-0.2630072534,0.0444294363,0.1328230053,-0.5474489927,-0.2302994579,0.0332790539,-0.1503331363,-0.1289111525,0.0344535038,0.1915319562,-0.1533609182,-0.1680428088,-0.1252952367,0.2708812058,-0.021253448,-0.3044984341,-0.2566545904,-0.1274342239,-0.02233194,-0.1389683336,0.2556281388,-0.0996973217,0.0149442051,-0.3301988244,0.2602413595,0.0299835727,-0.0868941471,-0.1960293204,-0.1635618955,-0.2488681674,-0.0509138107,-0.0062400112,0.0555705018,0.0800185576,-0.0115499785,-0.1427585185,0.4850380421,-0.1470862776,0.1867519021,-0.0164172091,-0.2713660896,-0.2431888282,0.210370034,0.2645954788,-0.1789331585,-0.2447142154,-0.3091816604,-0.3024846315,-0.3940913081,-0.3749517798,0.1970193535,-0.2094998211,0.0385531001,0.0277205259,0.1748259515,-0.5439237356,-0.1631395221,-0.0401764102,-0.121094957,-0.0790127516,-0.2031615824,-0.0244428143,-0.0962444246,0.1075747758,-0.4731317163,-0.1166654378,0.4031413794,0.0157228038,0.3176196814,-0.027778782,0.1408171207,-0.0598806925,-0.0406191014,0.2494788319,0.129090786,-0.4798995554,-0.2105068713,0.4358055592,-0.028408153,-0.1621991247,-0.2142312527,-0.2048006803,0.1530332118,0.0487895906,-0.2156378329,-0.0616773404,-0.3050428629,-0.2642215788,0.3011384904,0.3834311366,-0.1515817791,-0.060688708,-0.0312170144,-0.1840981245,-0.0187146869,0.2367314547,0.2664717436,0.2112924904,-0.2286883146,-0.4692656398,-0.5121158957,1.0407934189,0.1825529635,-0.1018013358,0.17184636,0.2220474333,0.1048513427,0.0664469674,0.1045498848,0.0572123155,0.1135688052,-0.0143871205,0.5116860271,0.132976532,-0.4170606434,0.0342959948,0.1691185534,-0.6020396352,-0.195427835,-0.1465610713,0.2970230579,0.2466924638,0.2368284464,-0.0028760282,0.0660989434,-0.4088563621,-0.0887969509,0.0617677979,0.050660897,-0.1207098216,0.214133054,-0.1384859085,-0.6867517233,0.1933727413,-0.0709495693,-0.2052294761,0.0560388304,0.0194496345,0.086858362,-0.0800292864,0.1252575368,-0.3246702254,-0.1553137153,-0.0970320255,-0.1469941139,0.0126970103,-0.0250421818,-0.5403260589,0.007091267,0.1818078607,0.1770075709,-0.3593627512,0.158603102,0.3823160231,-0.0662709624,-0.1340385675,0.039491795,0.0597206131,-0.0261142775,0.2145725042,-0.1636144221,-0.058931265,0.0720094517,-0.2767025232,0.0830321088,0.1091548875,0.2225001603,0.5866476297,0.2598196864,0.3088317215,0.1266040206,0.1405755132,-0.0172692928,0.124607943,0.1883587688,0.4441550076,-0.3815987706,0.0418763608,-0.0205792449,-0.2442580014,-0.0208680891,0.2543655038,0.1880321652,-0.0247265752,0.576032877,-0.3531186879,-0.0513542555,0.0268813297,0.2596484721,0.0805870295,-0.4422298074,-0.7111808062,0.4129859209,0.3456959128,-0.2478869557,-0.0011171037,0.4646760225,-0.2808056474,0.12825571,0.3057814837,0.9812768102,-0.1408487558,0.5032827258,0.0965092629,0.2022590339,0.2007422596,-0.1934661418,-0.1131105572,-0.3900813758,-0.0107201375,-0.1948066503,-0.0556962267,-0.3312034011,0.2056029737,-0.1097878218,0.344586432,-0.1208686158,0.1145380288,-0.1543718129,0.0669138059,0.24833332,0.2096652836,0.2211130261,0.1095281392,0.1115722507,0.246091947,-0.0531759635,0.0593530461,-0.1016327366,0.0581501573,-0.4849038124,0.0853021443,0.17887564,0.2424292266,-0.0005605644,-0.5334535837,0.1449189186,0.2944183946,0.5593866706,-0.0638762265,-0.4661138356,0.4058112502,-0.0868351385,0.1213345602,-0.0691807047,-0.0592409186,0.0723409578,0.032682661,-0.0503385104,0.1156681851,0.0916895568,-0.1074122936,-0.5722077489,-0.0525913537,-0.0917850807,-0.2888646424,-0.000238916,0.1918547451,-0.1695802808,0.0064626769,0.0615779124,-0.1173144057,0.1855731308,0.2290791422,0.2940686643,-0.3287419081,-0.2035214454,0.3612873256,0.6086620688,0.2955796719,0.535658896,-0.1068026125,-0.1889112592,-0.2103467882,-0.2012575418,0.3934397697,-0.1953916997,0.0757810771,-0.1093032956,0.0502360091,0.0399315879,0.4836317003,0.1875104755,0.2573179603,0.017072523,0.0654167905,0.0673966631,0.0407558084,-0.1606556624,0.5022093654,0.2771162391,-0.2831906974,0.1389853507,-0.0427671261,-0.2440882325,0.296482563,-0.0983619764,0.029095782,-0.0750984848,-0.1111132354,-0.1471775919,-0.2076717466,0.0225398671,-0.4305154085,-0.1581306309,-0.205891192,-0.1212102845,0.139726758,-0.0677736029,-0.2945946753,-0.2339213341,-0.0598240495,-0.0185737386,-0.0001521255,0.0129810106,0.1715544909,0.3940028548,0.1722122282,-0.0029349944,-0.1716357768,-0.319110781,0.1670131534,0.1354023963,0.1584018022,0.2739040554,0.0659335256,0.2369225472,0.0057691154,0.0120736482,0.3034847081,0.2247575521,0.1233666167,0.0962817669,0.0573436394,0.0691860318,0.2848485708,0.1961987317,0.3843570948,-0.2311223,-0.0829437301,0.1619483382,0.1932964325,-0.1112757623,-0.0962242037,0.1826933622,-0.2088822573,-0.2195140868,0.1951330602,0.2600964308,0.27834481,0.1306708902,0.0764359385,-0.2148827612,0.047183156,0.4676146805,0.2438165992,0.225559786,0.125457868,0.2277617455,0.1582165062,0.155978784,0.1614185274,0.097919032,0.1106009856,-0.1034631878,-0.027125001,0.5461992025,0.0388218686,0.067728959,0.0852848068,0.042105414,0.2842883468,0.2741174996,0.403929323,-0.293607533,-0.2621202171,-0.5696946979,0.0139657753,-0.0656386241,-0.355717808,0.0519288145,-0.4053050876,0.2063619196,0.1411308497,-0.3217982054,-0.2162514329,0.1946169585,0.0757135376,-0.1093758345,-0.4116977453,0.1830304861,0.1711625159,0.1422017664,0.0944696218,0.1711367667,-0.1907295585,-0.1423849165,0.0793437511,0.6095259786,-0.0323604085,0.0049416167,-0.2050826252,-0.3360153139,0.3389038444,-0.0195932239,0.17508021,-0.1984649897,0.2157602906,-0.4320357144,0.2925240099,0.0897909999,0.0395124666,0.0613232069,0.1295151561,0.2644067109,-0.25579831,-0.0652650595,-0.129413262,0.1327611059,-0.0701980442,-0.0012829812,-0.3846642673,0.4658086598,0.057936307,0.2285820395,0.1208448559,0.1112714037,0.025290858,0.1268810183,0.3632686734,0.3803136945,0.0688281506,-0.3417423069,-0.1238423958,-0.1760856062,0.0878648609,-0.2083670944,0.2201406509,-0.0782327652,0.3059858978,-0.1753035933,0.0076720938,0.2532886565,-0.1376580149,-0.3357023001,0.2708671391,-0.3088422716,-0.0135835549,-0.1889760345,0.0675630122,-0.1638001204,0.1567009091,0.2238539755,-0.1125178412,-0.0717945546,-0.0859613121,-0.1239657477,0.2219494134,0.2634161115,0.1766878814,-0.0960950181,0.2713316083,-0.0753867328,-0.1762156337,0.1869928837,-0.3589598835,-0.0309926346,0.0352221094,-0.2463647574,0.0154422643,-0.1843353361,-0.0716844499,0.1203957722,0.2608821094,0.1746184677,0.1953514665,-0.2793253958,-0.3151467144,-0.0628902987,-0.2449338287,0.0448936522,0.4290236831,-0.1279314905,-0.0388094448,0.035948582,-0.2259385437,0.5050830841,-0.1065701619,-0.0702053308,-0.4230026901,-0.244433865,0.0838429183,0.1866805404,-0.7867591381,0.139200747,0.2918980122,-0.2339582592,-0.170379445,-0.0696728826,0.242632255,0.1770141721,-0.2896568775,0.2271283418,-0.0063795894,-0.2858683169,-0.1549101919,-0.2253444642]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/665","title":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects","comments":"Hi !\r\nIt works on my side with both the LongFormerTokenizer and the LongFormerTokenizerFast.\r\n\r\nWhich version of transformers\/datasets are you using ?","body":"I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n","comment_length":22,"text":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects\nI load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n\nHi !\r\nIt works on my side with both the LongFormerTokenizer and the LongFormerTokenizerFast.\r\n\r\nWhich version of transformers\/datasets are you using ?","embeddings":[-0.2387796193,-0.2978491187,-0.019680649,0.2330349237,0.4450323582,-0.1729919016,0.2851519585,0.1309388727,-0.2760912478,0.1103135347,-0.0697863474,0.4891909063,-0.0087260343,-0.046733927,-0.145709604,-0.111473918,0.0535604171,0.207853511,0.0181137323,-0.0561074987,-0.2787823677,0.2074424773,-0.4731775522,0.1585230231,-0.2825988233,-0.2674281299,0.1064047962,-0.1985800564,-0.2693119049,-0.3823218942,-0.1436450779,-0.073365584,-0.0288482774,0.5633369088,-0.0001192604,0.0994183272,0.108954519,-0.145972386,0.0052242242,-0.0495663248,-0.0748076513,-0.3035497069,-0.205873996,-0.4582723975,-0.1556460112,-0.1438439935,0.0111553641,-0.3324888647,0.5145235062,0.2914209962,0.139978081,0.4932267964,0.0206352938,0.0532785431,0.015327774,0.0298241358,-0.1314535141,-0.2608632743,0.1487917304,0.2165469229,-0.1302717179,0.3254893422,-0.1613031477,-0.2627820969,0.0994700417,0.094937481,0.1041127741,-0.4675594568,0.3230481744,-0.002158023,0.2189722508,-0.020810321,-0.2567602098,-0.2679677308,-0.0794852078,0.0625912696,-0.0412523523,-0.2022152543,0.0776130408,0.0512863845,-0.3526940644,-0.1723531634,-0.1798734814,0.108233422,0.0383866839,0.3674840629,-0.2666110098,0.2930521965,0.0967984498,-0.2501663864,-0.1138216779,-0.0991195515,0.2264450341,0.4768724144,-0.3883422315,-0.3008864224,-0.0943802446,-0.409999162,0.3477460146,-0.0384070538,-0.2276796252,0.2554124892,-0.0489559509,0.1893791109,0.0231759958,0.3520210385,0.2632215321,0.3699644506,0.0116726784,-0.1440164447,0.0652595013,-0.04887354,-0.0822229534,-0.2215761691,-0.0569127798,0.2582862377,0.3595126569,-0.0186132845,-0.046238631,0.0278620608,-0.2262803167,0.0427433066,0.1734637767,0.4068900049,0.1080137938,0.2318411618,-0.3183892965,0.0902408659,-0.124280192,0.0253062062,-0.1654862761,0.214023307,-0.0066213412,-0.1393794417,-0.1643196195,0.0698126107,0.2013424933,-0.026837565,0.0632021129,-0.2202773243,0.061538171,-0.1790308952,0.2758253217,-0.0943112522,-0.0502242185,0.1737568676,0.2081712335,-0.40948084,-0.2060563862,0.0352333188,-0.34115237,0.01733163,0.1311888844,0.1185405329,0.0562376231,0.0476904213,-0.3749628663,0.2920556068,0.4146037698,0.0075913849,0.0322678164,-0.1267381161,-0.2016420662,-0.111127086,0.2182486653,0.1899360716,-0.1373286545,-0.2079769075,0.2503727973,-0.002062849,0.1535669416,0.4049597979,-0.0530274957,0.2562201917,-0.1467852145,0.559774518,0.4704823494,-0.6255075336,-0.5309312344,0.1032184213,-0.4468844831,-0.1779434234,-0.035346359,0.0416382365,0.5613888502,0.1241554692,-0.113206543,0.3392548561,0.072295852,0.2676001787,-0.1563941687,-0.1415297985,0.4092992246,0.0926817507,0.21709764,0.116488941,-0.1682532728,0.1971376091,-0.1022469327,-0.1236947551,0.0914442316,0.1902052313,0.1654276252,-0.2108140886,0.0806120411,-0.1629784405,-0.5545789599,0.0651864186,-0.3476859927,0.2234149575,-0.2890501916,-0.194772318,-0.2887653708,0.1541353166,-0.4519188404,-0.0524003021,0.083847262,-0.0306146815,0.1889378428,-0.0136042386,0.0192883443,-0.0761402398,0.1458211094,0.186094299,-0.2552472055,0.2062030882,-0.2569340169,0.021148419,-0.05973031,0.09906286,0.408500433,-0.1759835035,-0.2551455796,0.2755496502,0.1720520854,-0.0848676786,-0.0500598922,-0.0350354798,0.2658101916,-0.2096956968,-0.0644069538,0.1785884053,0.166718334,-0.0287600234,-0.0929355696,0.4431660473,0.224703446,0.3178621531,-0.0191053115,0.1625137478,0.2412854582,-0.10274712,-0.0464949757,-0.1355438232,-0.1764924526,0.0628632903,0.2340398133,-0.0736531839,0.0656425655,-0.0590749644,0.8479928374,-0.0683462247,0.0793827847,0.204474166,-0.116351068,0.0497056618,-0.0788842216,-0.4018104076,0.0417372212,0.0392285287,-0.1734893918,-0.1066208854,0.2546639144,0.1459451169,0.041525092,0.3150502145,0.1146959588,0.2255442888,0.1599868983,0.1129512191,-0.1720335782,-0.1941431463,0.1629729271,0.1452347934,-0.1468617916,-0.0244321302,-0.0460567623,0.0444771908,0.0104709063,-0.0486729778,-0.195032984,-0.2504672706,0.0337895602,-0.0237582624,0.0957349986,0.4495555162,0.3668554127,0.3240292668,0.427821368,-0.1099201962,-0.170441404,-0.1969794482,-0.0876939595,-0.0677347481,0.1707262248,-0.0053565041,-0.0357478224,0.0103845401,0.1024920642,-0.456127733,-0.4291423857,0.2537454963,-0.178284511,0.1860092133,-0.0141354213,0.1210128963,-0.3429493308,-0.2921084464,0.2118502706,-0.2194129676,-0.3013744652,0.0706816763,-0.0610883981,-0.3209867477,0.0511774272,-0.2405867577,-0.3765725493,-0.3376917541,0.363016367,-0.008248332,0.1135794222,0.1541582793,-0.0030258943,0.2885194421,-0.0200715531,-0.0359078087,-0.2440013885,0.0543059148,0.4122084975,-0.1826187968,-0.2826860547,-0.3915768564,-0.1759163439,0.0525789037,-0.1293791682,-0.1020923853,-0.2275468558,-0.2711945474,0.2238460481,-0.1608631462,0.0458878055,0.558042407,-0.0492363572,0.0907655656,0.0115672005,-0.1609315425,0.2720036209,0.1358186603,0.0812894851,-0.0107793091,0.5259879827,0.1519594789,0.8704359531,0.1582149416,-0.2194377482,0.1683554202,-0.1485036314,-0.010637085,-0.0340552628,-0.3892847896,0.1844976842,-0.0443505682,0.1226070747,0.1395881474,-0.2094922066,-0.1556345224,0.1454913467,-0.0947229341,-0.2187247276,-0.2757362723,-0.0225727931,0.0979045033,0.0983442664,0.0477664843,0.1866315752,-0.4435558915,0.0679604933,0.1192273498,0.0044029737,0.0480467714,-0.0468376577,-1.0162568092,-0.0913212076,-0.1702794135,0.3654927015,0.2294975519,0.4982431829,-0.0179926045,-0.0469948612,0.0515056588,-0.2121096551,0.8495112658,0.2505568564,-0.0532124676,0.0796275288,0.0189491622,-0.1545509845,0.0577008165,-0.1632412374,0.6255755424,0.4945641458,0.7852733731,-0.4374043345,-0.0893706232,0.0312036462,-0.0599599853,-0.1476593465,0.0810301974,-0.2087632865,-0.180411458,-0.5762064457,0.2585665584,0.2810131907,0.1928339154,0.152831018,-0.2231671363,0.0690152347,-0.2840030491,-0.0985406041,0.2312042862,0.2143349946,0.1858991086,-0.0022402669,0.145538196,-0.1561143696,-0.2068816274,0.189942807,0.25589481,-0.6231859922,0.2523162365,0.1651500016,0.560379684,0.1582307816,-0.1150253788,0.2946467102,-0.0105134277,0.479396224,-0.0390628837,0.4788197279,0.3626244366,0.2778424919,-0.301658839,-0.1384537518,0.3859476149,-0.0154792443,0.0721439794,0.0153059829,0.2642338276,-0.1523793787,0.578543961,0.1050156429,1.0689468384,-0.0962891653,-0.091162324,0.1268847138,0.2655840218,0.5600728393,-0.2914608121,0.2052387893,-0.5174925327,-0.0904992297,0.0042558508,-0.1398295611,0.1350088418,0.1887042522,-0.4300575256,0.1389254481,0.0992789939,0.4275937676,0.0399220213,0.4134412706,-0.0181035865,-0.4771104455,-0.1259956062,0.0662658066,-0.2162788957,0.2179785967,0.0931357071,0.0167735219,-0.2299908102,-0.3334072232,-0.0685394555,0.0877202675,-0.3980352879,0.6731894612,0.342954129,-0.1945347339,0.2956734002,0.2294498533,0.1547928452,0.2516962588,-0.0637095571,0.1682659537,0.1497499645,-0.0029832826,0.1993835419,-0.0682335496,0.1190654561,-0.0103186648,-0.1015117094,0.1754909605,-0.0124609089,-0.2762164176,-0.2717876732,-0.1240058616,0.2903387249,-0.3395356238,-0.2814121246,-0.0395575985,-0.2850346565,-0.2035131902,0.0220345296,-0.0706589893,-0.3747996092,0.2263817191,0.2376686484,-0.1426926404,0.1695863307,0.3712310195,-0.0255008675,-0.0888265967,0.5584918261,0.0715670735,-0.1946096569,-0.2024845928,0.1518792063,-0.033778619,-0.2271941155,0.1091578379,-0.1162661538,-0.2636434436,-0.2287956625,0.3435637355,-0.0256999768,0.0402624384,-0.2143465132,-0.4193456769,-0.5076628923,0.0499494001,-0.1312001497,0.1811650097,0.4385974407,0.4080056846,-0.2568692267,0.3760201633,-0.2495917976,-0.0800243616,-0.5987675786,0.2662300766,0.0188432653,-0.5459757447,-0.0275539905,0.0216406416,0.0264433939,0.2940588593,-0.1200792193,-0.1588124633,-0.0702605769,0.1605631262,0.2280812263,-0.1013725996,-0.2458603829,0.1184105352,-0.1792993993,-0.1568111479,0.1025064737,0.1319525987,0.0121633019,0.1235533729,-0.0908354744,-0.1170054078,-0.0513545312,-0.0865232125,0.0492475443,-0.1551797241,0.200568229,0.3887774944,-0.1900877953,-0.0090764407,-0.2909185886,0.1795438677,0.6930254102,0.0096776029,0.3933944106,-0.3145877421,-0.1508094966,0.229211092,-0.2073029727,0.1850663126,-0.3238908648,-0.095886223,0.2286647111,0.1294852048,-0.2605695128,-0.2062440068,0.5028803945,-0.3981368542,0.1572274119,0.1296645403,-0.0983623341,0.2678847909,0.1832061708,0.1603047252,0.7458079457,0.0117864432,0.1765505672,-0.0622489601,-0.092001304,0.0551159568,0.512357831,0.1351608634,0.2308288515,0.1101420224,-0.2032733262,0.3773396611,-0.4285535812,0.2233328372,0.3698289692,-0.1252685338,-0.193761304,0.208747223,-0.1403710991,0.2644395828,-0.04349453,0.3522603214,-0.1093692854,-0.1740659326,-0.1687214673,0.4904870093,-0.3190308213,-0.0490976721,-0.0212323908,-0.0759386569,-0.1959419698,-0.2856764197,0.0508428477,-0.0616605505,0.1097760573,0.0157869291,-0.0804274678,0.2032717317,-0.1478792727,0.1775085181,0.2425409406,-0.3291356266,0.2424013913,0.1096198335,0.2629410625,-0.2478345931,0.1149145141,0.3023824096,0.2686603367,-0.4128156602,0.1197531819,-0.061437685,-0.1552257091,0.1539587826,0.1354496777,-0.1843046695,-0.2192908376,0.2301807553,0.0778790861,-0.2095594406,-0.1722642481,0.1902220994,0.1092769131,0.1233894005,-0.0497298315,-0.0805690289,-0.2053932846,-0.3169712722,-0.1771351695,-0.3990464211,-0.0486690924,0.0820412487,-0.2334050238,0.2352662832,-0.3069973588,0.0806782916,0.0823595896,0.3644183874,0.3155186474,0.0889510736,-0.3957800269,-0.0233593192,-0.4049155414,0.2132003903,0.3097579479,-0.1938925534,0.085277535,0.1949707717,0.060536474,0.3144331872,-0.1063973233,0.0570466518,-0.1318134964,-0.0067865741,-0.2199010402,0.134248659,-0.165356636,0.1062286198,0.2053378969,-0.3196312189,0.3005925119,0.2335089296,-0.0160384122,-0.2037670761,-0.0731834322,-0.2135284543,-0.3241593242,-0.0482806824,0.2011152506,0.2985326052,-0.2826665938,0.180134356,0.2097664326,-0.0438413583,-0.2875527143,-0.193121776,0.1564357877,0.4927534163,-0.0002099175,0.1263695359,-0.3023884594,-0.1757035553,0.0875676423,-0.2813342214,-0.242300421,0.3333650529,-0.205700621,0.1753344238,-0.0147263901,0.2186222672,0.0708544925,0.2834080458,-0.2453328669,-0.4549573958,0.6601228714,-0.4289312661,-0.4800792336,-0.0562647842,-0.0659726113,0.1373787075,-0.2057521641,-0.2559036016,-0.1504520178,0.2913421392,0.084562473,-0.2006697506,0.0518865958,0.1554601938,-0.0358313434,-0.1501895785,0.3721503019,0.1291872263,0.0628954768,0.1038767546,-0.2139432728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/665","title":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects","comments":"transformers and datasets are both the latest","body":"I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n","comment_length":7,"text":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects\nI load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n\ntransformers and datasets are both the latest","embeddings":[-0.2387796193,-0.2978491187,-0.019680649,0.2330349237,0.4450323582,-0.1729919016,0.2851519585,0.1309388727,-0.2760912478,0.1103135347,-0.0697863474,0.4891909063,-0.0087260343,-0.046733927,-0.145709604,-0.111473918,0.0535604171,0.207853511,0.0181137323,-0.0561074987,-0.2787823677,0.2074424773,-0.4731775522,0.1585230231,-0.2825988233,-0.2674281299,0.1064047962,-0.1985800564,-0.2693119049,-0.3823218942,-0.1436450779,-0.073365584,-0.0288482774,0.5633369088,-0.0001192604,0.0994183272,0.108954519,-0.145972386,0.0052242242,-0.0495663248,-0.0748076513,-0.3035497069,-0.205873996,-0.4582723975,-0.1556460112,-0.1438439935,0.0111553641,-0.3324888647,0.5145235062,0.2914209962,0.139978081,0.4932267964,0.0206352938,0.0532785431,0.015327774,0.0298241358,-0.1314535141,-0.2608632743,0.1487917304,0.2165469229,-0.1302717179,0.3254893422,-0.1613031477,-0.2627820969,0.0994700417,0.094937481,0.1041127741,-0.4675594568,0.3230481744,-0.002158023,0.2189722508,-0.020810321,-0.2567602098,-0.2679677308,-0.0794852078,0.0625912696,-0.0412523523,-0.2022152543,0.0776130408,0.0512863845,-0.3526940644,-0.1723531634,-0.1798734814,0.108233422,0.0383866839,0.3674840629,-0.2666110098,0.2930521965,0.0967984498,-0.2501663864,-0.1138216779,-0.0991195515,0.2264450341,0.4768724144,-0.3883422315,-0.3008864224,-0.0943802446,-0.409999162,0.3477460146,-0.0384070538,-0.2276796252,0.2554124892,-0.0489559509,0.1893791109,0.0231759958,0.3520210385,0.2632215321,0.3699644506,0.0116726784,-0.1440164447,0.0652595013,-0.04887354,-0.0822229534,-0.2215761691,-0.0569127798,0.2582862377,0.3595126569,-0.0186132845,-0.046238631,0.0278620608,-0.2262803167,0.0427433066,0.1734637767,0.4068900049,0.1080137938,0.2318411618,-0.3183892965,0.0902408659,-0.124280192,0.0253062062,-0.1654862761,0.214023307,-0.0066213412,-0.1393794417,-0.1643196195,0.0698126107,0.2013424933,-0.026837565,0.0632021129,-0.2202773243,0.061538171,-0.1790308952,0.2758253217,-0.0943112522,-0.0502242185,0.1737568676,0.2081712335,-0.40948084,-0.2060563862,0.0352333188,-0.34115237,0.01733163,0.1311888844,0.1185405329,0.0562376231,0.0476904213,-0.3749628663,0.2920556068,0.4146037698,0.0075913849,0.0322678164,-0.1267381161,-0.2016420662,-0.111127086,0.2182486653,0.1899360716,-0.1373286545,-0.2079769075,0.2503727973,-0.002062849,0.1535669416,0.4049597979,-0.0530274957,0.2562201917,-0.1467852145,0.559774518,0.4704823494,-0.6255075336,-0.5309312344,0.1032184213,-0.4468844831,-0.1779434234,-0.035346359,0.0416382365,0.5613888502,0.1241554692,-0.113206543,0.3392548561,0.072295852,0.2676001787,-0.1563941687,-0.1415297985,0.4092992246,0.0926817507,0.21709764,0.116488941,-0.1682532728,0.1971376091,-0.1022469327,-0.1236947551,0.0914442316,0.1902052313,0.1654276252,-0.2108140886,0.0806120411,-0.1629784405,-0.5545789599,0.0651864186,-0.3476859927,0.2234149575,-0.2890501916,-0.194772318,-0.2887653708,0.1541353166,-0.4519188404,-0.0524003021,0.083847262,-0.0306146815,0.1889378428,-0.0136042386,0.0192883443,-0.0761402398,0.1458211094,0.186094299,-0.2552472055,0.2062030882,-0.2569340169,0.021148419,-0.05973031,0.09906286,0.408500433,-0.1759835035,-0.2551455796,0.2755496502,0.1720520854,-0.0848676786,-0.0500598922,-0.0350354798,0.2658101916,-0.2096956968,-0.0644069538,0.1785884053,0.166718334,-0.0287600234,-0.0929355696,0.4431660473,0.224703446,0.3178621531,-0.0191053115,0.1625137478,0.2412854582,-0.10274712,-0.0464949757,-0.1355438232,-0.1764924526,0.0628632903,0.2340398133,-0.0736531839,0.0656425655,-0.0590749644,0.8479928374,-0.0683462247,0.0793827847,0.204474166,-0.116351068,0.0497056618,-0.0788842216,-0.4018104076,0.0417372212,0.0392285287,-0.1734893918,-0.1066208854,0.2546639144,0.1459451169,0.041525092,0.3150502145,0.1146959588,0.2255442888,0.1599868983,0.1129512191,-0.1720335782,-0.1941431463,0.1629729271,0.1452347934,-0.1468617916,-0.0244321302,-0.0460567623,0.0444771908,0.0104709063,-0.0486729778,-0.195032984,-0.2504672706,0.0337895602,-0.0237582624,0.0957349986,0.4495555162,0.3668554127,0.3240292668,0.427821368,-0.1099201962,-0.170441404,-0.1969794482,-0.0876939595,-0.0677347481,0.1707262248,-0.0053565041,-0.0357478224,0.0103845401,0.1024920642,-0.456127733,-0.4291423857,0.2537454963,-0.178284511,0.1860092133,-0.0141354213,0.1210128963,-0.3429493308,-0.2921084464,0.2118502706,-0.2194129676,-0.3013744652,0.0706816763,-0.0610883981,-0.3209867477,0.0511774272,-0.2405867577,-0.3765725493,-0.3376917541,0.363016367,-0.008248332,0.1135794222,0.1541582793,-0.0030258943,0.2885194421,-0.0200715531,-0.0359078087,-0.2440013885,0.0543059148,0.4122084975,-0.1826187968,-0.2826860547,-0.3915768564,-0.1759163439,0.0525789037,-0.1293791682,-0.1020923853,-0.2275468558,-0.2711945474,0.2238460481,-0.1608631462,0.0458878055,0.558042407,-0.0492363572,0.0907655656,0.0115672005,-0.1609315425,0.2720036209,0.1358186603,0.0812894851,-0.0107793091,0.5259879827,0.1519594789,0.8704359531,0.1582149416,-0.2194377482,0.1683554202,-0.1485036314,-0.010637085,-0.0340552628,-0.3892847896,0.1844976842,-0.0443505682,0.1226070747,0.1395881474,-0.2094922066,-0.1556345224,0.1454913467,-0.0947229341,-0.2187247276,-0.2757362723,-0.0225727931,0.0979045033,0.0983442664,0.0477664843,0.1866315752,-0.4435558915,0.0679604933,0.1192273498,0.0044029737,0.0480467714,-0.0468376577,-1.0162568092,-0.0913212076,-0.1702794135,0.3654927015,0.2294975519,0.4982431829,-0.0179926045,-0.0469948612,0.0515056588,-0.2121096551,0.8495112658,0.2505568564,-0.0532124676,0.0796275288,0.0189491622,-0.1545509845,0.0577008165,-0.1632412374,0.6255755424,0.4945641458,0.7852733731,-0.4374043345,-0.0893706232,0.0312036462,-0.0599599853,-0.1476593465,0.0810301974,-0.2087632865,-0.180411458,-0.5762064457,0.2585665584,0.2810131907,0.1928339154,0.152831018,-0.2231671363,0.0690152347,-0.2840030491,-0.0985406041,0.2312042862,0.2143349946,0.1858991086,-0.0022402669,0.145538196,-0.1561143696,-0.2068816274,0.189942807,0.25589481,-0.6231859922,0.2523162365,0.1651500016,0.560379684,0.1582307816,-0.1150253788,0.2946467102,-0.0105134277,0.479396224,-0.0390628837,0.4788197279,0.3626244366,0.2778424919,-0.301658839,-0.1384537518,0.3859476149,-0.0154792443,0.0721439794,0.0153059829,0.2642338276,-0.1523793787,0.578543961,0.1050156429,1.0689468384,-0.0962891653,-0.091162324,0.1268847138,0.2655840218,0.5600728393,-0.2914608121,0.2052387893,-0.5174925327,-0.0904992297,0.0042558508,-0.1398295611,0.1350088418,0.1887042522,-0.4300575256,0.1389254481,0.0992789939,0.4275937676,0.0399220213,0.4134412706,-0.0181035865,-0.4771104455,-0.1259956062,0.0662658066,-0.2162788957,0.2179785967,0.0931357071,0.0167735219,-0.2299908102,-0.3334072232,-0.0685394555,0.0877202675,-0.3980352879,0.6731894612,0.342954129,-0.1945347339,0.2956734002,0.2294498533,0.1547928452,0.2516962588,-0.0637095571,0.1682659537,0.1497499645,-0.0029832826,0.1993835419,-0.0682335496,0.1190654561,-0.0103186648,-0.1015117094,0.1754909605,-0.0124609089,-0.2762164176,-0.2717876732,-0.1240058616,0.2903387249,-0.3395356238,-0.2814121246,-0.0395575985,-0.2850346565,-0.2035131902,0.0220345296,-0.0706589893,-0.3747996092,0.2263817191,0.2376686484,-0.1426926404,0.1695863307,0.3712310195,-0.0255008675,-0.0888265967,0.5584918261,0.0715670735,-0.1946096569,-0.2024845928,0.1518792063,-0.033778619,-0.2271941155,0.1091578379,-0.1162661538,-0.2636434436,-0.2287956625,0.3435637355,-0.0256999768,0.0402624384,-0.2143465132,-0.4193456769,-0.5076628923,0.0499494001,-0.1312001497,0.1811650097,0.4385974407,0.4080056846,-0.2568692267,0.3760201633,-0.2495917976,-0.0800243616,-0.5987675786,0.2662300766,0.0188432653,-0.5459757447,-0.0275539905,0.0216406416,0.0264433939,0.2940588593,-0.1200792193,-0.1588124633,-0.0702605769,0.1605631262,0.2280812263,-0.1013725996,-0.2458603829,0.1184105352,-0.1792993993,-0.1568111479,0.1025064737,0.1319525987,0.0121633019,0.1235533729,-0.0908354744,-0.1170054078,-0.0513545312,-0.0865232125,0.0492475443,-0.1551797241,0.200568229,0.3887774944,-0.1900877953,-0.0090764407,-0.2909185886,0.1795438677,0.6930254102,0.0096776029,0.3933944106,-0.3145877421,-0.1508094966,0.229211092,-0.2073029727,0.1850663126,-0.3238908648,-0.095886223,0.2286647111,0.1294852048,-0.2605695128,-0.2062440068,0.5028803945,-0.3981368542,0.1572274119,0.1296645403,-0.0983623341,0.2678847909,0.1832061708,0.1603047252,0.7458079457,0.0117864432,0.1765505672,-0.0622489601,-0.092001304,0.0551159568,0.512357831,0.1351608634,0.2308288515,0.1101420224,-0.2032733262,0.3773396611,-0.4285535812,0.2233328372,0.3698289692,-0.1252685338,-0.193761304,0.208747223,-0.1403710991,0.2644395828,-0.04349453,0.3522603214,-0.1093692854,-0.1740659326,-0.1687214673,0.4904870093,-0.3190308213,-0.0490976721,-0.0212323908,-0.0759386569,-0.1959419698,-0.2856764197,0.0508428477,-0.0616605505,0.1097760573,0.0157869291,-0.0804274678,0.2032717317,-0.1478792727,0.1775085181,0.2425409406,-0.3291356266,0.2424013913,0.1096198335,0.2629410625,-0.2478345931,0.1149145141,0.3023824096,0.2686603367,-0.4128156602,0.1197531819,-0.061437685,-0.1552257091,0.1539587826,0.1354496777,-0.1843046695,-0.2192908376,0.2301807553,0.0778790861,-0.2095594406,-0.1722642481,0.1902220994,0.1092769131,0.1233894005,-0.0497298315,-0.0805690289,-0.2053932846,-0.3169712722,-0.1771351695,-0.3990464211,-0.0486690924,0.0820412487,-0.2334050238,0.2352662832,-0.3069973588,0.0806782916,0.0823595896,0.3644183874,0.3155186474,0.0889510736,-0.3957800269,-0.0233593192,-0.4049155414,0.2132003903,0.3097579479,-0.1938925534,0.085277535,0.1949707717,0.060536474,0.3144331872,-0.1063973233,0.0570466518,-0.1318134964,-0.0067865741,-0.2199010402,0.134248659,-0.165356636,0.1062286198,0.2053378969,-0.3196312189,0.3005925119,0.2335089296,-0.0160384122,-0.2037670761,-0.0731834322,-0.2135284543,-0.3241593242,-0.0482806824,0.2011152506,0.2985326052,-0.2826665938,0.180134356,0.2097664326,-0.0438413583,-0.2875527143,-0.193121776,0.1564357877,0.4927534163,-0.0002099175,0.1263695359,-0.3023884594,-0.1757035553,0.0875676423,-0.2813342214,-0.242300421,0.3333650529,-0.205700621,0.1753344238,-0.0147263901,0.2186222672,0.0708544925,0.2834080458,-0.2453328669,-0.4549573958,0.6601228714,-0.4289312661,-0.4800792336,-0.0562647842,-0.0659726113,0.1373787075,-0.2057521641,-0.2559036016,-0.1504520178,0.2913421392,0.084562473,-0.2006697506,0.0518865958,0.1554601938,-0.0358313434,-0.1501895785,0.3721503019,0.1291872263,0.0628954768,0.1038767546,-0.2139432728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/665","title":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects","comments":"Then I guess you need to give us more informations on your setup (OS, python, GPU, etc) or a Google Colab reproducing the error for us to be able to debug this error.","body":"I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n","comment_length":33,"text":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects\nI load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n\nThen I guess you need to give us more informations on your setup (OS, python, GPU, etc) or a Google Colab reproducing the error for us to be able to debug this error.","embeddings":[-0.2387796193,-0.2978491187,-0.019680649,0.2330349237,0.4450323582,-0.1729919016,0.2851519585,0.1309388727,-0.2760912478,0.1103135347,-0.0697863474,0.4891909063,-0.0087260343,-0.046733927,-0.145709604,-0.111473918,0.0535604171,0.207853511,0.0181137323,-0.0561074987,-0.2787823677,0.2074424773,-0.4731775522,0.1585230231,-0.2825988233,-0.2674281299,0.1064047962,-0.1985800564,-0.2693119049,-0.3823218942,-0.1436450779,-0.073365584,-0.0288482774,0.5633369088,-0.0001192604,0.0994183272,0.108954519,-0.145972386,0.0052242242,-0.0495663248,-0.0748076513,-0.3035497069,-0.205873996,-0.4582723975,-0.1556460112,-0.1438439935,0.0111553641,-0.3324888647,0.5145235062,0.2914209962,0.139978081,0.4932267964,0.0206352938,0.0532785431,0.015327774,0.0298241358,-0.1314535141,-0.2608632743,0.1487917304,0.2165469229,-0.1302717179,0.3254893422,-0.1613031477,-0.2627820969,0.0994700417,0.094937481,0.1041127741,-0.4675594568,0.3230481744,-0.002158023,0.2189722508,-0.020810321,-0.2567602098,-0.2679677308,-0.0794852078,0.0625912696,-0.0412523523,-0.2022152543,0.0776130408,0.0512863845,-0.3526940644,-0.1723531634,-0.1798734814,0.108233422,0.0383866839,0.3674840629,-0.2666110098,0.2930521965,0.0967984498,-0.2501663864,-0.1138216779,-0.0991195515,0.2264450341,0.4768724144,-0.3883422315,-0.3008864224,-0.0943802446,-0.409999162,0.3477460146,-0.0384070538,-0.2276796252,0.2554124892,-0.0489559509,0.1893791109,0.0231759958,0.3520210385,0.2632215321,0.3699644506,0.0116726784,-0.1440164447,0.0652595013,-0.04887354,-0.0822229534,-0.2215761691,-0.0569127798,0.2582862377,0.3595126569,-0.0186132845,-0.046238631,0.0278620608,-0.2262803167,0.0427433066,0.1734637767,0.4068900049,0.1080137938,0.2318411618,-0.3183892965,0.0902408659,-0.124280192,0.0253062062,-0.1654862761,0.214023307,-0.0066213412,-0.1393794417,-0.1643196195,0.0698126107,0.2013424933,-0.026837565,0.0632021129,-0.2202773243,0.061538171,-0.1790308952,0.2758253217,-0.0943112522,-0.0502242185,0.1737568676,0.2081712335,-0.40948084,-0.2060563862,0.0352333188,-0.34115237,0.01733163,0.1311888844,0.1185405329,0.0562376231,0.0476904213,-0.3749628663,0.2920556068,0.4146037698,0.0075913849,0.0322678164,-0.1267381161,-0.2016420662,-0.111127086,0.2182486653,0.1899360716,-0.1373286545,-0.2079769075,0.2503727973,-0.002062849,0.1535669416,0.4049597979,-0.0530274957,0.2562201917,-0.1467852145,0.559774518,0.4704823494,-0.6255075336,-0.5309312344,0.1032184213,-0.4468844831,-0.1779434234,-0.035346359,0.0416382365,0.5613888502,0.1241554692,-0.113206543,0.3392548561,0.072295852,0.2676001787,-0.1563941687,-0.1415297985,0.4092992246,0.0926817507,0.21709764,0.116488941,-0.1682532728,0.1971376091,-0.1022469327,-0.1236947551,0.0914442316,0.1902052313,0.1654276252,-0.2108140886,0.0806120411,-0.1629784405,-0.5545789599,0.0651864186,-0.3476859927,0.2234149575,-0.2890501916,-0.194772318,-0.2887653708,0.1541353166,-0.4519188404,-0.0524003021,0.083847262,-0.0306146815,0.1889378428,-0.0136042386,0.0192883443,-0.0761402398,0.1458211094,0.186094299,-0.2552472055,0.2062030882,-0.2569340169,0.021148419,-0.05973031,0.09906286,0.408500433,-0.1759835035,-0.2551455796,0.2755496502,0.1720520854,-0.0848676786,-0.0500598922,-0.0350354798,0.2658101916,-0.2096956968,-0.0644069538,0.1785884053,0.166718334,-0.0287600234,-0.0929355696,0.4431660473,0.224703446,0.3178621531,-0.0191053115,0.1625137478,0.2412854582,-0.10274712,-0.0464949757,-0.1355438232,-0.1764924526,0.0628632903,0.2340398133,-0.0736531839,0.0656425655,-0.0590749644,0.8479928374,-0.0683462247,0.0793827847,0.204474166,-0.116351068,0.0497056618,-0.0788842216,-0.4018104076,0.0417372212,0.0392285287,-0.1734893918,-0.1066208854,0.2546639144,0.1459451169,0.041525092,0.3150502145,0.1146959588,0.2255442888,0.1599868983,0.1129512191,-0.1720335782,-0.1941431463,0.1629729271,0.1452347934,-0.1468617916,-0.0244321302,-0.0460567623,0.0444771908,0.0104709063,-0.0486729778,-0.195032984,-0.2504672706,0.0337895602,-0.0237582624,0.0957349986,0.4495555162,0.3668554127,0.3240292668,0.427821368,-0.1099201962,-0.170441404,-0.1969794482,-0.0876939595,-0.0677347481,0.1707262248,-0.0053565041,-0.0357478224,0.0103845401,0.1024920642,-0.456127733,-0.4291423857,0.2537454963,-0.178284511,0.1860092133,-0.0141354213,0.1210128963,-0.3429493308,-0.2921084464,0.2118502706,-0.2194129676,-0.3013744652,0.0706816763,-0.0610883981,-0.3209867477,0.0511774272,-0.2405867577,-0.3765725493,-0.3376917541,0.363016367,-0.008248332,0.1135794222,0.1541582793,-0.0030258943,0.2885194421,-0.0200715531,-0.0359078087,-0.2440013885,0.0543059148,0.4122084975,-0.1826187968,-0.2826860547,-0.3915768564,-0.1759163439,0.0525789037,-0.1293791682,-0.1020923853,-0.2275468558,-0.2711945474,0.2238460481,-0.1608631462,0.0458878055,0.558042407,-0.0492363572,0.0907655656,0.0115672005,-0.1609315425,0.2720036209,0.1358186603,0.0812894851,-0.0107793091,0.5259879827,0.1519594789,0.8704359531,0.1582149416,-0.2194377482,0.1683554202,-0.1485036314,-0.010637085,-0.0340552628,-0.3892847896,0.1844976842,-0.0443505682,0.1226070747,0.1395881474,-0.2094922066,-0.1556345224,0.1454913467,-0.0947229341,-0.2187247276,-0.2757362723,-0.0225727931,0.0979045033,0.0983442664,0.0477664843,0.1866315752,-0.4435558915,0.0679604933,0.1192273498,0.0044029737,0.0480467714,-0.0468376577,-1.0162568092,-0.0913212076,-0.1702794135,0.3654927015,0.2294975519,0.4982431829,-0.0179926045,-0.0469948612,0.0515056588,-0.2121096551,0.8495112658,0.2505568564,-0.0532124676,0.0796275288,0.0189491622,-0.1545509845,0.0577008165,-0.1632412374,0.6255755424,0.4945641458,0.7852733731,-0.4374043345,-0.0893706232,0.0312036462,-0.0599599853,-0.1476593465,0.0810301974,-0.2087632865,-0.180411458,-0.5762064457,0.2585665584,0.2810131907,0.1928339154,0.152831018,-0.2231671363,0.0690152347,-0.2840030491,-0.0985406041,0.2312042862,0.2143349946,0.1858991086,-0.0022402669,0.145538196,-0.1561143696,-0.2068816274,0.189942807,0.25589481,-0.6231859922,0.2523162365,0.1651500016,0.560379684,0.1582307816,-0.1150253788,0.2946467102,-0.0105134277,0.479396224,-0.0390628837,0.4788197279,0.3626244366,0.2778424919,-0.301658839,-0.1384537518,0.3859476149,-0.0154792443,0.0721439794,0.0153059829,0.2642338276,-0.1523793787,0.578543961,0.1050156429,1.0689468384,-0.0962891653,-0.091162324,0.1268847138,0.2655840218,0.5600728393,-0.2914608121,0.2052387893,-0.5174925327,-0.0904992297,0.0042558508,-0.1398295611,0.1350088418,0.1887042522,-0.4300575256,0.1389254481,0.0992789939,0.4275937676,0.0399220213,0.4134412706,-0.0181035865,-0.4771104455,-0.1259956062,0.0662658066,-0.2162788957,0.2179785967,0.0931357071,0.0167735219,-0.2299908102,-0.3334072232,-0.0685394555,0.0877202675,-0.3980352879,0.6731894612,0.342954129,-0.1945347339,0.2956734002,0.2294498533,0.1547928452,0.2516962588,-0.0637095571,0.1682659537,0.1497499645,-0.0029832826,0.1993835419,-0.0682335496,0.1190654561,-0.0103186648,-0.1015117094,0.1754909605,-0.0124609089,-0.2762164176,-0.2717876732,-0.1240058616,0.2903387249,-0.3395356238,-0.2814121246,-0.0395575985,-0.2850346565,-0.2035131902,0.0220345296,-0.0706589893,-0.3747996092,0.2263817191,0.2376686484,-0.1426926404,0.1695863307,0.3712310195,-0.0255008675,-0.0888265967,0.5584918261,0.0715670735,-0.1946096569,-0.2024845928,0.1518792063,-0.033778619,-0.2271941155,0.1091578379,-0.1162661538,-0.2636434436,-0.2287956625,0.3435637355,-0.0256999768,0.0402624384,-0.2143465132,-0.4193456769,-0.5076628923,0.0499494001,-0.1312001497,0.1811650097,0.4385974407,0.4080056846,-0.2568692267,0.3760201633,-0.2495917976,-0.0800243616,-0.5987675786,0.2662300766,0.0188432653,-0.5459757447,-0.0275539905,0.0216406416,0.0264433939,0.2940588593,-0.1200792193,-0.1588124633,-0.0702605769,0.1605631262,0.2280812263,-0.1013725996,-0.2458603829,0.1184105352,-0.1792993993,-0.1568111479,0.1025064737,0.1319525987,0.0121633019,0.1235533729,-0.0908354744,-0.1170054078,-0.0513545312,-0.0865232125,0.0492475443,-0.1551797241,0.200568229,0.3887774944,-0.1900877953,-0.0090764407,-0.2909185886,0.1795438677,0.6930254102,0.0096776029,0.3933944106,-0.3145877421,-0.1508094966,0.229211092,-0.2073029727,0.1850663126,-0.3238908648,-0.095886223,0.2286647111,0.1294852048,-0.2605695128,-0.2062440068,0.5028803945,-0.3981368542,0.1572274119,0.1296645403,-0.0983623341,0.2678847909,0.1832061708,0.1603047252,0.7458079457,0.0117864432,0.1765505672,-0.0622489601,-0.092001304,0.0551159568,0.512357831,0.1351608634,0.2308288515,0.1101420224,-0.2032733262,0.3773396611,-0.4285535812,0.2233328372,0.3698289692,-0.1252685338,-0.193761304,0.208747223,-0.1403710991,0.2644395828,-0.04349453,0.3522603214,-0.1093692854,-0.1740659326,-0.1687214673,0.4904870093,-0.3190308213,-0.0490976721,-0.0212323908,-0.0759386569,-0.1959419698,-0.2856764197,0.0508428477,-0.0616605505,0.1097760573,0.0157869291,-0.0804274678,0.2032717317,-0.1478792727,0.1775085181,0.2425409406,-0.3291356266,0.2424013913,0.1096198335,0.2629410625,-0.2478345931,0.1149145141,0.3023824096,0.2686603367,-0.4128156602,0.1197531819,-0.061437685,-0.1552257091,0.1539587826,0.1354496777,-0.1843046695,-0.2192908376,0.2301807553,0.0778790861,-0.2095594406,-0.1722642481,0.1902220994,0.1092769131,0.1233894005,-0.0497298315,-0.0805690289,-0.2053932846,-0.3169712722,-0.1771351695,-0.3990464211,-0.0486690924,0.0820412487,-0.2334050238,0.2352662832,-0.3069973588,0.0806782916,0.0823595896,0.3644183874,0.3155186474,0.0889510736,-0.3957800269,-0.0233593192,-0.4049155414,0.2132003903,0.3097579479,-0.1938925534,0.085277535,0.1949707717,0.060536474,0.3144331872,-0.1063973233,0.0570466518,-0.1318134964,-0.0067865741,-0.2199010402,0.134248659,-0.165356636,0.1062286198,0.2053378969,-0.3196312189,0.3005925119,0.2335089296,-0.0160384122,-0.2037670761,-0.0731834322,-0.2135284543,-0.3241593242,-0.0482806824,0.2011152506,0.2985326052,-0.2826665938,0.180134356,0.2097664326,-0.0438413583,-0.2875527143,-0.193121776,0.1564357877,0.4927534163,-0.0002099175,0.1263695359,-0.3023884594,-0.1757035553,0.0875676423,-0.2813342214,-0.242300421,0.3333650529,-0.205700621,0.1753344238,-0.0147263901,0.2186222672,0.0708544925,0.2834080458,-0.2453328669,-0.4549573958,0.6601228714,-0.4289312661,-0.4800792336,-0.0562647842,-0.0659726113,0.1373787075,-0.2057521641,-0.2559036016,-0.1504520178,0.2913421392,0.084562473,-0.2006697506,0.0518865958,0.1554601938,-0.0358313434,-0.1501895785,0.3721503019,0.1291872263,0.0628954768,0.1038767546,-0.2139432728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/665","title":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects","comments":"And your version of `dill` if possible :)","body":"I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n","comment_length":8,"text":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects\nI load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n\nAnd your version of `dill` if possible :)","embeddings":[-0.2387796193,-0.2978491187,-0.019680649,0.2330349237,0.4450323582,-0.1729919016,0.2851519585,0.1309388727,-0.2760912478,0.1103135347,-0.0697863474,0.4891909063,-0.0087260343,-0.046733927,-0.145709604,-0.111473918,0.0535604171,0.207853511,0.0181137323,-0.0561074987,-0.2787823677,0.2074424773,-0.4731775522,0.1585230231,-0.2825988233,-0.2674281299,0.1064047962,-0.1985800564,-0.2693119049,-0.3823218942,-0.1436450779,-0.073365584,-0.0288482774,0.5633369088,-0.0001192604,0.0994183272,0.108954519,-0.145972386,0.0052242242,-0.0495663248,-0.0748076513,-0.3035497069,-0.205873996,-0.4582723975,-0.1556460112,-0.1438439935,0.0111553641,-0.3324888647,0.5145235062,0.2914209962,0.139978081,0.4932267964,0.0206352938,0.0532785431,0.015327774,0.0298241358,-0.1314535141,-0.2608632743,0.1487917304,0.2165469229,-0.1302717179,0.3254893422,-0.1613031477,-0.2627820969,0.0994700417,0.094937481,0.1041127741,-0.4675594568,0.3230481744,-0.002158023,0.2189722508,-0.020810321,-0.2567602098,-0.2679677308,-0.0794852078,0.0625912696,-0.0412523523,-0.2022152543,0.0776130408,0.0512863845,-0.3526940644,-0.1723531634,-0.1798734814,0.108233422,0.0383866839,0.3674840629,-0.2666110098,0.2930521965,0.0967984498,-0.2501663864,-0.1138216779,-0.0991195515,0.2264450341,0.4768724144,-0.3883422315,-0.3008864224,-0.0943802446,-0.409999162,0.3477460146,-0.0384070538,-0.2276796252,0.2554124892,-0.0489559509,0.1893791109,0.0231759958,0.3520210385,0.2632215321,0.3699644506,0.0116726784,-0.1440164447,0.0652595013,-0.04887354,-0.0822229534,-0.2215761691,-0.0569127798,0.2582862377,0.3595126569,-0.0186132845,-0.046238631,0.0278620608,-0.2262803167,0.0427433066,0.1734637767,0.4068900049,0.1080137938,0.2318411618,-0.3183892965,0.0902408659,-0.124280192,0.0253062062,-0.1654862761,0.214023307,-0.0066213412,-0.1393794417,-0.1643196195,0.0698126107,0.2013424933,-0.026837565,0.0632021129,-0.2202773243,0.061538171,-0.1790308952,0.2758253217,-0.0943112522,-0.0502242185,0.1737568676,0.2081712335,-0.40948084,-0.2060563862,0.0352333188,-0.34115237,0.01733163,0.1311888844,0.1185405329,0.0562376231,0.0476904213,-0.3749628663,0.2920556068,0.4146037698,0.0075913849,0.0322678164,-0.1267381161,-0.2016420662,-0.111127086,0.2182486653,0.1899360716,-0.1373286545,-0.2079769075,0.2503727973,-0.002062849,0.1535669416,0.4049597979,-0.0530274957,0.2562201917,-0.1467852145,0.559774518,0.4704823494,-0.6255075336,-0.5309312344,0.1032184213,-0.4468844831,-0.1779434234,-0.035346359,0.0416382365,0.5613888502,0.1241554692,-0.113206543,0.3392548561,0.072295852,0.2676001787,-0.1563941687,-0.1415297985,0.4092992246,0.0926817507,0.21709764,0.116488941,-0.1682532728,0.1971376091,-0.1022469327,-0.1236947551,0.0914442316,0.1902052313,0.1654276252,-0.2108140886,0.0806120411,-0.1629784405,-0.5545789599,0.0651864186,-0.3476859927,0.2234149575,-0.2890501916,-0.194772318,-0.2887653708,0.1541353166,-0.4519188404,-0.0524003021,0.083847262,-0.0306146815,0.1889378428,-0.0136042386,0.0192883443,-0.0761402398,0.1458211094,0.186094299,-0.2552472055,0.2062030882,-0.2569340169,0.021148419,-0.05973031,0.09906286,0.408500433,-0.1759835035,-0.2551455796,0.2755496502,0.1720520854,-0.0848676786,-0.0500598922,-0.0350354798,0.2658101916,-0.2096956968,-0.0644069538,0.1785884053,0.166718334,-0.0287600234,-0.0929355696,0.4431660473,0.224703446,0.3178621531,-0.0191053115,0.1625137478,0.2412854582,-0.10274712,-0.0464949757,-0.1355438232,-0.1764924526,0.0628632903,0.2340398133,-0.0736531839,0.0656425655,-0.0590749644,0.8479928374,-0.0683462247,0.0793827847,0.204474166,-0.116351068,0.0497056618,-0.0788842216,-0.4018104076,0.0417372212,0.0392285287,-0.1734893918,-0.1066208854,0.2546639144,0.1459451169,0.041525092,0.3150502145,0.1146959588,0.2255442888,0.1599868983,0.1129512191,-0.1720335782,-0.1941431463,0.1629729271,0.1452347934,-0.1468617916,-0.0244321302,-0.0460567623,0.0444771908,0.0104709063,-0.0486729778,-0.195032984,-0.2504672706,0.0337895602,-0.0237582624,0.0957349986,0.4495555162,0.3668554127,0.3240292668,0.427821368,-0.1099201962,-0.170441404,-0.1969794482,-0.0876939595,-0.0677347481,0.1707262248,-0.0053565041,-0.0357478224,0.0103845401,0.1024920642,-0.456127733,-0.4291423857,0.2537454963,-0.178284511,0.1860092133,-0.0141354213,0.1210128963,-0.3429493308,-0.2921084464,0.2118502706,-0.2194129676,-0.3013744652,0.0706816763,-0.0610883981,-0.3209867477,0.0511774272,-0.2405867577,-0.3765725493,-0.3376917541,0.363016367,-0.008248332,0.1135794222,0.1541582793,-0.0030258943,0.2885194421,-0.0200715531,-0.0359078087,-0.2440013885,0.0543059148,0.4122084975,-0.1826187968,-0.2826860547,-0.3915768564,-0.1759163439,0.0525789037,-0.1293791682,-0.1020923853,-0.2275468558,-0.2711945474,0.2238460481,-0.1608631462,0.0458878055,0.558042407,-0.0492363572,0.0907655656,0.0115672005,-0.1609315425,0.2720036209,0.1358186603,0.0812894851,-0.0107793091,0.5259879827,0.1519594789,0.8704359531,0.1582149416,-0.2194377482,0.1683554202,-0.1485036314,-0.010637085,-0.0340552628,-0.3892847896,0.1844976842,-0.0443505682,0.1226070747,0.1395881474,-0.2094922066,-0.1556345224,0.1454913467,-0.0947229341,-0.2187247276,-0.2757362723,-0.0225727931,0.0979045033,0.0983442664,0.0477664843,0.1866315752,-0.4435558915,0.0679604933,0.1192273498,0.0044029737,0.0480467714,-0.0468376577,-1.0162568092,-0.0913212076,-0.1702794135,0.3654927015,0.2294975519,0.4982431829,-0.0179926045,-0.0469948612,0.0515056588,-0.2121096551,0.8495112658,0.2505568564,-0.0532124676,0.0796275288,0.0189491622,-0.1545509845,0.0577008165,-0.1632412374,0.6255755424,0.4945641458,0.7852733731,-0.4374043345,-0.0893706232,0.0312036462,-0.0599599853,-0.1476593465,0.0810301974,-0.2087632865,-0.180411458,-0.5762064457,0.2585665584,0.2810131907,0.1928339154,0.152831018,-0.2231671363,0.0690152347,-0.2840030491,-0.0985406041,0.2312042862,0.2143349946,0.1858991086,-0.0022402669,0.145538196,-0.1561143696,-0.2068816274,0.189942807,0.25589481,-0.6231859922,0.2523162365,0.1651500016,0.560379684,0.1582307816,-0.1150253788,0.2946467102,-0.0105134277,0.479396224,-0.0390628837,0.4788197279,0.3626244366,0.2778424919,-0.301658839,-0.1384537518,0.3859476149,-0.0154792443,0.0721439794,0.0153059829,0.2642338276,-0.1523793787,0.578543961,0.1050156429,1.0689468384,-0.0962891653,-0.091162324,0.1268847138,0.2655840218,0.5600728393,-0.2914608121,0.2052387893,-0.5174925327,-0.0904992297,0.0042558508,-0.1398295611,0.1350088418,0.1887042522,-0.4300575256,0.1389254481,0.0992789939,0.4275937676,0.0399220213,0.4134412706,-0.0181035865,-0.4771104455,-0.1259956062,0.0662658066,-0.2162788957,0.2179785967,0.0931357071,0.0167735219,-0.2299908102,-0.3334072232,-0.0685394555,0.0877202675,-0.3980352879,0.6731894612,0.342954129,-0.1945347339,0.2956734002,0.2294498533,0.1547928452,0.2516962588,-0.0637095571,0.1682659537,0.1497499645,-0.0029832826,0.1993835419,-0.0682335496,0.1190654561,-0.0103186648,-0.1015117094,0.1754909605,-0.0124609089,-0.2762164176,-0.2717876732,-0.1240058616,0.2903387249,-0.3395356238,-0.2814121246,-0.0395575985,-0.2850346565,-0.2035131902,0.0220345296,-0.0706589893,-0.3747996092,0.2263817191,0.2376686484,-0.1426926404,0.1695863307,0.3712310195,-0.0255008675,-0.0888265967,0.5584918261,0.0715670735,-0.1946096569,-0.2024845928,0.1518792063,-0.033778619,-0.2271941155,0.1091578379,-0.1162661538,-0.2636434436,-0.2287956625,0.3435637355,-0.0256999768,0.0402624384,-0.2143465132,-0.4193456769,-0.5076628923,0.0499494001,-0.1312001497,0.1811650097,0.4385974407,0.4080056846,-0.2568692267,0.3760201633,-0.2495917976,-0.0800243616,-0.5987675786,0.2662300766,0.0188432653,-0.5459757447,-0.0275539905,0.0216406416,0.0264433939,0.2940588593,-0.1200792193,-0.1588124633,-0.0702605769,0.1605631262,0.2280812263,-0.1013725996,-0.2458603829,0.1184105352,-0.1792993993,-0.1568111479,0.1025064737,0.1319525987,0.0121633019,0.1235533729,-0.0908354744,-0.1170054078,-0.0513545312,-0.0865232125,0.0492475443,-0.1551797241,0.200568229,0.3887774944,-0.1900877953,-0.0090764407,-0.2909185886,0.1795438677,0.6930254102,0.0096776029,0.3933944106,-0.3145877421,-0.1508094966,0.229211092,-0.2073029727,0.1850663126,-0.3238908648,-0.095886223,0.2286647111,0.1294852048,-0.2605695128,-0.2062440068,0.5028803945,-0.3981368542,0.1572274119,0.1296645403,-0.0983623341,0.2678847909,0.1832061708,0.1603047252,0.7458079457,0.0117864432,0.1765505672,-0.0622489601,-0.092001304,0.0551159568,0.512357831,0.1351608634,0.2308288515,0.1101420224,-0.2032733262,0.3773396611,-0.4285535812,0.2233328372,0.3698289692,-0.1252685338,-0.193761304,0.208747223,-0.1403710991,0.2644395828,-0.04349453,0.3522603214,-0.1093692854,-0.1740659326,-0.1687214673,0.4904870093,-0.3190308213,-0.0490976721,-0.0212323908,-0.0759386569,-0.1959419698,-0.2856764197,0.0508428477,-0.0616605505,0.1097760573,0.0157869291,-0.0804274678,0.2032717317,-0.1478792727,0.1775085181,0.2425409406,-0.3291356266,0.2424013913,0.1096198335,0.2629410625,-0.2478345931,0.1149145141,0.3023824096,0.2686603367,-0.4128156602,0.1197531819,-0.061437685,-0.1552257091,0.1539587826,0.1354496777,-0.1843046695,-0.2192908376,0.2301807553,0.0778790861,-0.2095594406,-0.1722642481,0.1902220994,0.1092769131,0.1233894005,-0.0497298315,-0.0805690289,-0.2053932846,-0.3169712722,-0.1771351695,-0.3990464211,-0.0486690924,0.0820412487,-0.2334050238,0.2352662832,-0.3069973588,0.0806782916,0.0823595896,0.3644183874,0.3155186474,0.0889510736,-0.3957800269,-0.0233593192,-0.4049155414,0.2132003903,0.3097579479,-0.1938925534,0.085277535,0.1949707717,0.060536474,0.3144331872,-0.1063973233,0.0570466518,-0.1318134964,-0.0067865741,-0.2199010402,0.134248659,-0.165356636,0.1062286198,0.2053378969,-0.3196312189,0.3005925119,0.2335089296,-0.0160384122,-0.2037670761,-0.0731834322,-0.2135284543,-0.3241593242,-0.0482806824,0.2011152506,0.2985326052,-0.2826665938,0.180134356,0.2097664326,-0.0438413583,-0.2875527143,-0.193121776,0.1564357877,0.4927534163,-0.0002099175,0.1263695359,-0.3023884594,-0.1757035553,0.0875676423,-0.2813342214,-0.242300421,0.3333650529,-0.205700621,0.1753344238,-0.0147263901,0.2186222672,0.0708544925,0.2834080458,-0.2453328669,-0.4549573958,0.6601228714,-0.4289312661,-0.4800792336,-0.0562647842,-0.0659726113,0.1373787075,-0.2057521641,-0.2559036016,-0.1504520178,0.2913421392,0.084562473,-0.2006697506,0.0518865958,0.1554601938,-0.0358313434,-0.1501895785,0.3721503019,0.1291872263,0.0628954768,0.1038767546,-0.2139432728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/665","title":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects","comments":"I have the same issue with `transformers\/BertJapaneseTokenizer`.\r\n\r\n\r\n\r\n```python\r\n# train_ds = Dataset(features: {\r\n# 'title': Value(dtype='string', id=None), \r\n# 'score': Value(dtype='float64', id=None)\r\n# }, num_rows: 99999)\r\n\r\nt = BertJapaneseTokenizer.from_pretrained('bert-base-japanese-whole-word-masking')\r\nencoded = train_ds.map(lambda examples: {'tokens': t.encode(examples['title'])}, batched=True)\r\n```\r\n\r\n<details><summary>Error Message<\/summary>\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-35-2b7d66b291c1> in <module>\r\n 2 \r\n 3 encoded = train_ds.map(lambda examples:\r\n----> 4 {'tokens': t.encode(examples['title'])}, batched=True)\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in dump(self, obj)\r\n 407 if self.proto >= 4:\r\n 408 self.framer.start_framing()\r\n--> 409 self.save(obj)\r\n 410 self.write(STOP)\r\n 411 self.framer.end_framing()\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 608 else:\r\n 609 save(func)\r\n--> 610 save(args)\r\n 611 write(REDUCE)\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_tuple(self, obj)\r\n 749 write(MARK)\r\n 750 for element in obj:\r\n--> 751 save(element)\r\n 752 \r\n 753 if id(obj) in memo:\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in _batch_setitems(self, items)\r\n 850 k, v = tmp[0]\r\n 851 save(k)\r\n--> 852 save(v)\r\n 853 write(SETITEM)\r\n 854 # else tmp is empty, and we're done\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 519 \r\n 520 # Save the reduce() output and finally memoize the object\r\n--> 521 self.save_reduce(obj=obj, *rv)\r\n 522 \r\n 523 def persistent_id(self, obj):\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 632 \r\n 633 if state is not None:\r\n--> 634 save(state)\r\n 635 write(BUILD)\r\n 636 \r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in _batch_setitems(self, items)\r\n 845 for k, v in tmp:\r\n 846 save(k)\r\n--> 847 save(v)\r\n 848 write(SETITEMS)\r\n 849 elif n:\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 519 \r\n 520 # Save the reduce() output and finally memoize the object\r\n--> 521 self.save_reduce(obj=obj, *rv)\r\n 522 \r\n 523 def persistent_id(self, obj):\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 632 \r\n 633 if state is not None:\r\n--> 634 save(state)\r\n 635 write(BUILD)\r\n 636 \r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in _batch_setitems(self, items)\r\n 845 for k, v in tmp:\r\n 846 save(k)\r\n--> 847 save(v)\r\n 848 write(SETITEMS)\r\n 849 elif n:\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 494 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 495 if reduce is not None:\r\n--> 496 rv = reduce(self.proto)\r\n 497 else:\r\n 498 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tagger objects\r\n```\r\n\r\n<\/details>\r\n\r\ntrainsformers: 2.10.0\r\ndatasets: 1.0.2\r\ndill: 0.3.2\r\npython: 3.6.8\r\n\r\nOS: ubuntu 16.04 (Docker Image) on [Deep Learning VM](https:\/\/console.cloud.google.com\/marketplace\/details\/click-to-deploy-images\/deeplearning) (GCP)\r\nGPU: Tesla P100 (CUDA 10)\r\n","body":"I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n","comment_length":861,"text":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects\nI load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n\nI have the same issue with `transformers\/BertJapaneseTokenizer`.\r\n\r\n\r\n\r\n```python\r\n# train_ds = Dataset(features: {\r\n# 'title': Value(dtype='string', id=None), \r\n# 'score': Value(dtype='float64', id=None)\r\n# }, num_rows: 99999)\r\n\r\nt = BertJapaneseTokenizer.from_pretrained('bert-base-japanese-whole-word-masking')\r\nencoded = train_ds.map(lambda examples: {'tokens': t.encode(examples['title'])}, batched=True)\r\n```\r\n\r\n<details><summary>Error Message<\/summary>\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-35-2b7d66b291c1> in <module>\r\n 2 \r\n 3 encoded = train_ds.map(lambda examples:\r\n----> 4 {'tokens': t.encode(examples['title'])}, batched=True)\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in dump(self, obj)\r\n 407 if self.proto >= 4:\r\n 408 self.framer.start_framing()\r\n--> 409 self.save(obj)\r\n 410 self.write(STOP)\r\n 411 self.framer.end_framing()\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 608 else:\r\n 609 save(func)\r\n--> 610 save(args)\r\n 611 write(REDUCE)\r\n 612 \r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_tuple(self, obj)\r\n 749 write(MARK)\r\n 750 for element in obj:\r\n--> 751 save(element)\r\n 752 \r\n 753 if id(obj) in memo:\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in _batch_setitems(self, items)\r\n 850 k, v = tmp[0]\r\n 851 save(k)\r\n--> 852 save(v)\r\n 853 write(SETITEM)\r\n 854 # else tmp is empty, and we're done\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 519 \r\n 520 # Save the reduce() output and finally memoize the object\r\n--> 521 self.save_reduce(obj=obj, *rv)\r\n 522 \r\n 523 def persistent_id(self, obj):\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 632 \r\n 633 if state is not None:\r\n--> 634 save(state)\r\n 635 write(BUILD)\r\n 636 \r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in _batch_setitems(self, items)\r\n 845 for k, v in tmp:\r\n 846 save(k)\r\n--> 847 save(v)\r\n 848 write(SETITEMS)\r\n 849 elif n:\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 519 \r\n 520 # Save the reduce() output and finally memoize the object\r\n--> 521 self.save_reduce(obj=obj, *rv)\r\n 522 \r\n 523 def persistent_id(self, obj):\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 632 \r\n 633 if state is not None:\r\n--> 634 save(state)\r\n 635 write(BUILD)\r\n 636 \r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n\/usr\/local\/lib\/python3.6\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in _batch_setitems(self, items)\r\n 845 for k, v in tmp:\r\n 846 save(k)\r\n--> 847 save(v)\r\n 848 write(SETITEMS)\r\n 849 elif n:\r\n\r\n\/usr\/local\/lib\/python3.6\/pickle.py in save(self, obj, save_persistent_id)\r\n 494 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 495 if reduce is not None:\r\n--> 496 rv = reduce(self.proto)\r\n 497 else:\r\n 498 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tagger objects\r\n```\r\n\r\n<\/details>\r\n\r\ntrainsformers: 2.10.0\r\ndatasets: 1.0.2\r\ndill: 0.3.2\r\npython: 3.6.8\r\n\r\nOS: ubuntu 16.04 (Docker Image) on [Deep Learning VM](https:\/\/console.cloud.google.com\/marketplace\/details\/click-to-deploy-images\/deeplearning) (GCP)\r\nGPU: Tesla P100 (CUDA 10)\r\n","embeddings":[-0.2387796193,-0.2978491187,-0.019680649,0.2330349237,0.4450323582,-0.1729919016,0.2851519585,0.1309388727,-0.2760912478,0.1103135347,-0.0697863474,0.4891909063,-0.0087260343,-0.046733927,-0.145709604,-0.111473918,0.0535604171,0.207853511,0.0181137323,-0.0561074987,-0.2787823677,0.2074424773,-0.4731775522,0.1585230231,-0.2825988233,-0.2674281299,0.1064047962,-0.1985800564,-0.2693119049,-0.3823218942,-0.1436450779,-0.073365584,-0.0288482774,0.5633369088,-0.0001192604,0.0994183272,0.108954519,-0.145972386,0.0052242242,-0.0495663248,-0.0748076513,-0.3035497069,-0.205873996,-0.4582723975,-0.1556460112,-0.1438439935,0.0111553641,-0.3324888647,0.5145235062,0.2914209962,0.139978081,0.4932267964,0.0206352938,0.0532785431,0.015327774,0.0298241358,-0.1314535141,-0.2608632743,0.1487917304,0.2165469229,-0.1302717179,0.3254893422,-0.1613031477,-0.2627820969,0.0994700417,0.094937481,0.1041127741,-0.4675594568,0.3230481744,-0.002158023,0.2189722508,-0.020810321,-0.2567602098,-0.2679677308,-0.0794852078,0.0625912696,-0.0412523523,-0.2022152543,0.0776130408,0.0512863845,-0.3526940644,-0.1723531634,-0.1798734814,0.108233422,0.0383866839,0.3674840629,-0.2666110098,0.2930521965,0.0967984498,-0.2501663864,-0.1138216779,-0.0991195515,0.2264450341,0.4768724144,-0.3883422315,-0.3008864224,-0.0943802446,-0.409999162,0.3477460146,-0.0384070538,-0.2276796252,0.2554124892,-0.0489559509,0.1893791109,0.0231759958,0.3520210385,0.2632215321,0.3699644506,0.0116726784,-0.1440164447,0.0652595013,-0.04887354,-0.0822229534,-0.2215761691,-0.0569127798,0.2582862377,0.3595126569,-0.0186132845,-0.046238631,0.0278620608,-0.2262803167,0.0427433066,0.1734637767,0.4068900049,0.1080137938,0.2318411618,-0.3183892965,0.0902408659,-0.124280192,0.0253062062,-0.1654862761,0.214023307,-0.0066213412,-0.1393794417,-0.1643196195,0.0698126107,0.2013424933,-0.026837565,0.0632021129,-0.2202773243,0.061538171,-0.1790308952,0.2758253217,-0.0943112522,-0.0502242185,0.1737568676,0.2081712335,-0.40948084,-0.2060563862,0.0352333188,-0.34115237,0.01733163,0.1311888844,0.1185405329,0.0562376231,0.0476904213,-0.3749628663,0.2920556068,0.4146037698,0.0075913849,0.0322678164,-0.1267381161,-0.2016420662,-0.111127086,0.2182486653,0.1899360716,-0.1373286545,-0.2079769075,0.2503727973,-0.002062849,0.1535669416,0.4049597979,-0.0530274957,0.2562201917,-0.1467852145,0.559774518,0.4704823494,-0.6255075336,-0.5309312344,0.1032184213,-0.4468844831,-0.1779434234,-0.035346359,0.0416382365,0.5613888502,0.1241554692,-0.113206543,0.3392548561,0.072295852,0.2676001787,-0.1563941687,-0.1415297985,0.4092992246,0.0926817507,0.21709764,0.116488941,-0.1682532728,0.1971376091,-0.1022469327,-0.1236947551,0.0914442316,0.1902052313,0.1654276252,-0.2108140886,0.0806120411,-0.1629784405,-0.5545789599,0.0651864186,-0.3476859927,0.2234149575,-0.2890501916,-0.194772318,-0.2887653708,0.1541353166,-0.4519188404,-0.0524003021,0.083847262,-0.0306146815,0.1889378428,-0.0136042386,0.0192883443,-0.0761402398,0.1458211094,0.186094299,-0.2552472055,0.2062030882,-0.2569340169,0.021148419,-0.05973031,0.09906286,0.408500433,-0.1759835035,-0.2551455796,0.2755496502,0.1720520854,-0.0848676786,-0.0500598922,-0.0350354798,0.2658101916,-0.2096956968,-0.0644069538,0.1785884053,0.166718334,-0.0287600234,-0.0929355696,0.4431660473,0.224703446,0.3178621531,-0.0191053115,0.1625137478,0.2412854582,-0.10274712,-0.0464949757,-0.1355438232,-0.1764924526,0.0628632903,0.2340398133,-0.0736531839,0.0656425655,-0.0590749644,0.8479928374,-0.0683462247,0.0793827847,0.204474166,-0.116351068,0.0497056618,-0.0788842216,-0.4018104076,0.0417372212,0.0392285287,-0.1734893918,-0.1066208854,0.2546639144,0.1459451169,0.041525092,0.3150502145,0.1146959588,0.2255442888,0.1599868983,0.1129512191,-0.1720335782,-0.1941431463,0.1629729271,0.1452347934,-0.1468617916,-0.0244321302,-0.0460567623,0.0444771908,0.0104709063,-0.0486729778,-0.195032984,-0.2504672706,0.0337895602,-0.0237582624,0.0957349986,0.4495555162,0.3668554127,0.3240292668,0.427821368,-0.1099201962,-0.170441404,-0.1969794482,-0.0876939595,-0.0677347481,0.1707262248,-0.0053565041,-0.0357478224,0.0103845401,0.1024920642,-0.456127733,-0.4291423857,0.2537454963,-0.178284511,0.1860092133,-0.0141354213,0.1210128963,-0.3429493308,-0.2921084464,0.2118502706,-0.2194129676,-0.3013744652,0.0706816763,-0.0610883981,-0.3209867477,0.0511774272,-0.2405867577,-0.3765725493,-0.3376917541,0.363016367,-0.008248332,0.1135794222,0.1541582793,-0.0030258943,0.2885194421,-0.0200715531,-0.0359078087,-0.2440013885,0.0543059148,0.4122084975,-0.1826187968,-0.2826860547,-0.3915768564,-0.1759163439,0.0525789037,-0.1293791682,-0.1020923853,-0.2275468558,-0.2711945474,0.2238460481,-0.1608631462,0.0458878055,0.558042407,-0.0492363572,0.0907655656,0.0115672005,-0.1609315425,0.2720036209,0.1358186603,0.0812894851,-0.0107793091,0.5259879827,0.1519594789,0.8704359531,0.1582149416,-0.2194377482,0.1683554202,-0.1485036314,-0.010637085,-0.0340552628,-0.3892847896,0.1844976842,-0.0443505682,0.1226070747,0.1395881474,-0.2094922066,-0.1556345224,0.1454913467,-0.0947229341,-0.2187247276,-0.2757362723,-0.0225727931,0.0979045033,0.0983442664,0.0477664843,0.1866315752,-0.4435558915,0.0679604933,0.1192273498,0.0044029737,0.0480467714,-0.0468376577,-1.0162568092,-0.0913212076,-0.1702794135,0.3654927015,0.2294975519,0.4982431829,-0.0179926045,-0.0469948612,0.0515056588,-0.2121096551,0.8495112658,0.2505568564,-0.0532124676,0.0796275288,0.0189491622,-0.1545509845,0.0577008165,-0.1632412374,0.6255755424,0.4945641458,0.7852733731,-0.4374043345,-0.0893706232,0.0312036462,-0.0599599853,-0.1476593465,0.0810301974,-0.2087632865,-0.180411458,-0.5762064457,0.2585665584,0.2810131907,0.1928339154,0.152831018,-0.2231671363,0.0690152347,-0.2840030491,-0.0985406041,0.2312042862,0.2143349946,0.1858991086,-0.0022402669,0.145538196,-0.1561143696,-0.2068816274,0.189942807,0.25589481,-0.6231859922,0.2523162365,0.1651500016,0.560379684,0.1582307816,-0.1150253788,0.2946467102,-0.0105134277,0.479396224,-0.0390628837,0.4788197279,0.3626244366,0.2778424919,-0.301658839,-0.1384537518,0.3859476149,-0.0154792443,0.0721439794,0.0153059829,0.2642338276,-0.1523793787,0.578543961,0.1050156429,1.0689468384,-0.0962891653,-0.091162324,0.1268847138,0.2655840218,0.5600728393,-0.2914608121,0.2052387893,-0.5174925327,-0.0904992297,0.0042558508,-0.1398295611,0.1350088418,0.1887042522,-0.4300575256,0.1389254481,0.0992789939,0.4275937676,0.0399220213,0.4134412706,-0.0181035865,-0.4771104455,-0.1259956062,0.0662658066,-0.2162788957,0.2179785967,0.0931357071,0.0167735219,-0.2299908102,-0.3334072232,-0.0685394555,0.0877202675,-0.3980352879,0.6731894612,0.342954129,-0.1945347339,0.2956734002,0.2294498533,0.1547928452,0.2516962588,-0.0637095571,0.1682659537,0.1497499645,-0.0029832826,0.1993835419,-0.0682335496,0.1190654561,-0.0103186648,-0.1015117094,0.1754909605,-0.0124609089,-0.2762164176,-0.2717876732,-0.1240058616,0.2903387249,-0.3395356238,-0.2814121246,-0.0395575985,-0.2850346565,-0.2035131902,0.0220345296,-0.0706589893,-0.3747996092,0.2263817191,0.2376686484,-0.1426926404,0.1695863307,0.3712310195,-0.0255008675,-0.0888265967,0.5584918261,0.0715670735,-0.1946096569,-0.2024845928,0.1518792063,-0.033778619,-0.2271941155,0.1091578379,-0.1162661538,-0.2636434436,-0.2287956625,0.3435637355,-0.0256999768,0.0402624384,-0.2143465132,-0.4193456769,-0.5076628923,0.0499494001,-0.1312001497,0.1811650097,0.4385974407,0.4080056846,-0.2568692267,0.3760201633,-0.2495917976,-0.0800243616,-0.5987675786,0.2662300766,0.0188432653,-0.5459757447,-0.0275539905,0.0216406416,0.0264433939,0.2940588593,-0.1200792193,-0.1588124633,-0.0702605769,0.1605631262,0.2280812263,-0.1013725996,-0.2458603829,0.1184105352,-0.1792993993,-0.1568111479,0.1025064737,0.1319525987,0.0121633019,0.1235533729,-0.0908354744,-0.1170054078,-0.0513545312,-0.0865232125,0.0492475443,-0.1551797241,0.200568229,0.3887774944,-0.1900877953,-0.0090764407,-0.2909185886,0.1795438677,0.6930254102,0.0096776029,0.3933944106,-0.3145877421,-0.1508094966,0.229211092,-0.2073029727,0.1850663126,-0.3238908648,-0.095886223,0.2286647111,0.1294852048,-0.2605695128,-0.2062440068,0.5028803945,-0.3981368542,0.1572274119,0.1296645403,-0.0983623341,0.2678847909,0.1832061708,0.1603047252,0.7458079457,0.0117864432,0.1765505672,-0.0622489601,-0.092001304,0.0551159568,0.512357831,0.1351608634,0.2308288515,0.1101420224,-0.2032733262,0.3773396611,-0.4285535812,0.2233328372,0.3698289692,-0.1252685338,-0.193761304,0.208747223,-0.1403710991,0.2644395828,-0.04349453,0.3522603214,-0.1093692854,-0.1740659326,-0.1687214673,0.4904870093,-0.3190308213,-0.0490976721,-0.0212323908,-0.0759386569,-0.1959419698,-0.2856764197,0.0508428477,-0.0616605505,0.1097760573,0.0157869291,-0.0804274678,0.2032717317,-0.1478792727,0.1775085181,0.2425409406,-0.3291356266,0.2424013913,0.1096198335,0.2629410625,-0.2478345931,0.1149145141,0.3023824096,0.2686603367,-0.4128156602,0.1197531819,-0.061437685,-0.1552257091,0.1539587826,0.1354496777,-0.1843046695,-0.2192908376,0.2301807553,0.0778790861,-0.2095594406,-0.1722642481,0.1902220994,0.1092769131,0.1233894005,-0.0497298315,-0.0805690289,-0.2053932846,-0.3169712722,-0.1771351695,-0.3990464211,-0.0486690924,0.0820412487,-0.2334050238,0.2352662832,-0.3069973588,0.0806782916,0.0823595896,0.3644183874,0.3155186474,0.0889510736,-0.3957800269,-0.0233593192,-0.4049155414,0.2132003903,0.3097579479,-0.1938925534,0.085277535,0.1949707717,0.060536474,0.3144331872,-0.1063973233,0.0570466518,-0.1318134964,-0.0067865741,-0.2199010402,0.134248659,-0.165356636,0.1062286198,0.2053378969,-0.3196312189,0.3005925119,0.2335089296,-0.0160384122,-0.2037670761,-0.0731834322,-0.2135284543,-0.3241593242,-0.0482806824,0.2011152506,0.2985326052,-0.2826665938,0.180134356,0.2097664326,-0.0438413583,-0.2875527143,-0.193121776,0.1564357877,0.4927534163,-0.0002099175,0.1263695359,-0.3023884594,-0.1757035553,0.0875676423,-0.2813342214,-0.242300421,0.3333650529,-0.205700621,0.1753344238,-0.0147263901,0.2186222672,0.0708544925,0.2834080458,-0.2453328669,-0.4549573958,0.6601228714,-0.4289312661,-0.4800792336,-0.0562647842,-0.0659726113,0.1373787075,-0.2057521641,-0.2559036016,-0.1504520178,0.2913421392,0.084562473,-0.2006697506,0.0518865958,0.1554601938,-0.0358313434,-0.1501895785,0.3721503019,0.1291872263,0.0628954768,0.1038767546,-0.2139432728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/665","title":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects","comments":"> I have the same issue with `transformers\/BertJapaneseTokenizer`.\r\n\r\nIt looks like it this tokenizer is not supported unfortunately.\r\nThis is because `t.word_tokenizer.mecab` is a `fugashi.fugashi.GenericTagger` which is not compatible with pickle nor dill.\r\n\r\nWe need objects passes to `map` to be picklable for our caching system to work properly.\r\nHere it crashes because the caching system is not able to pickle the GenericTagger.\r\n\r\n\\> Maybe you can create an issue on [fugashi](https:\/\/github.com\/polm\/fugashi\/issues) 's repo and ask to make `fugashi.fugashi.GenericTagger` compatible with pickle ?\r\n\r\nWhat you can do in the meantime is use a picklable wrapper of the tokenizer:\r\n\r\n\r\n```python\r\nfrom transformers import BertJapaneseTokenizer, MecabTokenizer\r\n\r\nclass PicklableTokenizer(BertJapaneseTokenizer):\r\n\r\n def __getstate__(self):\r\n state = dict(self.__dict__)\r\n state[\"do_lower_case\"] = self.word_tokenizer.do_lower_case\r\n state[\"never_split\"] = self.word_tokenizer.never_split \r\n del state[\"word_tokenizer\"]\r\n return state\r\n\r\n def __setstate__(self, state):\r\n do_lower_case = state.pop(\"do_lower_case\")\r\n never_split = state.pop(\"never_split\")\r\n self.__dict__ = state\r\n self.word_tokenizer = MecabTokenizer(\r\n do_lower_case=do_lower_case, never_split=never_split)\r\n )\r\n\r\nt = PicklableTokenizer.from_pretrained(\"cl-tohoku\/bert-base-japanese-whole-word-masking\")\r\nencoded = train_ds.map(lambda examples: {'tokens': t.encode(examples['title'])}, batched=True) # it works\r\n```","body":"I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n","comment_length":153,"text":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects\nI load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n\n> I have the same issue with `transformers\/BertJapaneseTokenizer`.\r\n\r\nIt looks like it this tokenizer is not supported unfortunately.\r\nThis is because `t.word_tokenizer.mecab` is a `fugashi.fugashi.GenericTagger` which is not compatible with pickle nor dill.\r\n\r\nWe need objects passes to `map` to be picklable for our caching system to work properly.\r\nHere it crashes because the caching system is not able to pickle the GenericTagger.\r\n\r\n\\> Maybe you can create an issue on [fugashi](https:\/\/github.com\/polm\/fugashi\/issues) 's repo and ask to make `fugashi.fugashi.GenericTagger` compatible with pickle ?\r\n\r\nWhat you can do in the meantime is use a picklable wrapper of the tokenizer:\r\n\r\n\r\n```python\r\nfrom transformers import BertJapaneseTokenizer, MecabTokenizer\r\n\r\nclass PicklableTokenizer(BertJapaneseTokenizer):\r\n\r\n def __getstate__(self):\r\n state = dict(self.__dict__)\r\n state[\"do_lower_case\"] = self.word_tokenizer.do_lower_case\r\n state[\"never_split\"] = self.word_tokenizer.never_split \r\n del state[\"word_tokenizer\"]\r\n return state\r\n\r\n def __setstate__(self, state):\r\n do_lower_case = state.pop(\"do_lower_case\")\r\n never_split = state.pop(\"never_split\")\r\n self.__dict__ = state\r\n self.word_tokenizer = MecabTokenizer(\r\n do_lower_case=do_lower_case, never_split=never_split)\r\n )\r\n\r\nt = PicklableTokenizer.from_pretrained(\"cl-tohoku\/bert-base-japanese-whole-word-masking\")\r\nencoded = train_ds.map(lambda examples: {'tokens': t.encode(examples['title'])}, batched=True) # it works\r\n```","embeddings":[-0.2387796193,-0.2978491187,-0.019680649,0.2330349237,0.4450323582,-0.1729919016,0.2851519585,0.1309388727,-0.2760912478,0.1103135347,-0.0697863474,0.4891909063,-0.0087260343,-0.046733927,-0.145709604,-0.111473918,0.0535604171,0.207853511,0.0181137323,-0.0561074987,-0.2787823677,0.2074424773,-0.4731775522,0.1585230231,-0.2825988233,-0.2674281299,0.1064047962,-0.1985800564,-0.2693119049,-0.3823218942,-0.1436450779,-0.073365584,-0.0288482774,0.5633369088,-0.0001192604,0.0994183272,0.108954519,-0.145972386,0.0052242242,-0.0495663248,-0.0748076513,-0.3035497069,-0.205873996,-0.4582723975,-0.1556460112,-0.1438439935,0.0111553641,-0.3324888647,0.5145235062,0.2914209962,0.139978081,0.4932267964,0.0206352938,0.0532785431,0.015327774,0.0298241358,-0.1314535141,-0.2608632743,0.1487917304,0.2165469229,-0.1302717179,0.3254893422,-0.1613031477,-0.2627820969,0.0994700417,0.094937481,0.1041127741,-0.4675594568,0.3230481744,-0.002158023,0.2189722508,-0.020810321,-0.2567602098,-0.2679677308,-0.0794852078,0.0625912696,-0.0412523523,-0.2022152543,0.0776130408,0.0512863845,-0.3526940644,-0.1723531634,-0.1798734814,0.108233422,0.0383866839,0.3674840629,-0.2666110098,0.2930521965,0.0967984498,-0.2501663864,-0.1138216779,-0.0991195515,0.2264450341,0.4768724144,-0.3883422315,-0.3008864224,-0.0943802446,-0.409999162,0.3477460146,-0.0384070538,-0.2276796252,0.2554124892,-0.0489559509,0.1893791109,0.0231759958,0.3520210385,0.2632215321,0.3699644506,0.0116726784,-0.1440164447,0.0652595013,-0.04887354,-0.0822229534,-0.2215761691,-0.0569127798,0.2582862377,0.3595126569,-0.0186132845,-0.046238631,0.0278620608,-0.2262803167,0.0427433066,0.1734637767,0.4068900049,0.1080137938,0.2318411618,-0.3183892965,0.0902408659,-0.124280192,0.0253062062,-0.1654862761,0.214023307,-0.0066213412,-0.1393794417,-0.1643196195,0.0698126107,0.2013424933,-0.026837565,0.0632021129,-0.2202773243,0.061538171,-0.1790308952,0.2758253217,-0.0943112522,-0.0502242185,0.1737568676,0.2081712335,-0.40948084,-0.2060563862,0.0352333188,-0.34115237,0.01733163,0.1311888844,0.1185405329,0.0562376231,0.0476904213,-0.3749628663,0.2920556068,0.4146037698,0.0075913849,0.0322678164,-0.1267381161,-0.2016420662,-0.111127086,0.2182486653,0.1899360716,-0.1373286545,-0.2079769075,0.2503727973,-0.002062849,0.1535669416,0.4049597979,-0.0530274957,0.2562201917,-0.1467852145,0.559774518,0.4704823494,-0.6255075336,-0.5309312344,0.1032184213,-0.4468844831,-0.1779434234,-0.035346359,0.0416382365,0.5613888502,0.1241554692,-0.113206543,0.3392548561,0.072295852,0.2676001787,-0.1563941687,-0.1415297985,0.4092992246,0.0926817507,0.21709764,0.116488941,-0.1682532728,0.1971376091,-0.1022469327,-0.1236947551,0.0914442316,0.1902052313,0.1654276252,-0.2108140886,0.0806120411,-0.1629784405,-0.5545789599,0.0651864186,-0.3476859927,0.2234149575,-0.2890501916,-0.194772318,-0.2887653708,0.1541353166,-0.4519188404,-0.0524003021,0.083847262,-0.0306146815,0.1889378428,-0.0136042386,0.0192883443,-0.0761402398,0.1458211094,0.186094299,-0.2552472055,0.2062030882,-0.2569340169,0.021148419,-0.05973031,0.09906286,0.408500433,-0.1759835035,-0.2551455796,0.2755496502,0.1720520854,-0.0848676786,-0.0500598922,-0.0350354798,0.2658101916,-0.2096956968,-0.0644069538,0.1785884053,0.166718334,-0.0287600234,-0.0929355696,0.4431660473,0.224703446,0.3178621531,-0.0191053115,0.1625137478,0.2412854582,-0.10274712,-0.0464949757,-0.1355438232,-0.1764924526,0.0628632903,0.2340398133,-0.0736531839,0.0656425655,-0.0590749644,0.8479928374,-0.0683462247,0.0793827847,0.204474166,-0.116351068,0.0497056618,-0.0788842216,-0.4018104076,0.0417372212,0.0392285287,-0.1734893918,-0.1066208854,0.2546639144,0.1459451169,0.041525092,0.3150502145,0.1146959588,0.2255442888,0.1599868983,0.1129512191,-0.1720335782,-0.1941431463,0.1629729271,0.1452347934,-0.1468617916,-0.0244321302,-0.0460567623,0.0444771908,0.0104709063,-0.0486729778,-0.195032984,-0.2504672706,0.0337895602,-0.0237582624,0.0957349986,0.4495555162,0.3668554127,0.3240292668,0.427821368,-0.1099201962,-0.170441404,-0.1969794482,-0.0876939595,-0.0677347481,0.1707262248,-0.0053565041,-0.0357478224,0.0103845401,0.1024920642,-0.456127733,-0.4291423857,0.2537454963,-0.178284511,0.1860092133,-0.0141354213,0.1210128963,-0.3429493308,-0.2921084464,0.2118502706,-0.2194129676,-0.3013744652,0.0706816763,-0.0610883981,-0.3209867477,0.0511774272,-0.2405867577,-0.3765725493,-0.3376917541,0.363016367,-0.008248332,0.1135794222,0.1541582793,-0.0030258943,0.2885194421,-0.0200715531,-0.0359078087,-0.2440013885,0.0543059148,0.4122084975,-0.1826187968,-0.2826860547,-0.3915768564,-0.1759163439,0.0525789037,-0.1293791682,-0.1020923853,-0.2275468558,-0.2711945474,0.2238460481,-0.1608631462,0.0458878055,0.558042407,-0.0492363572,0.0907655656,0.0115672005,-0.1609315425,0.2720036209,0.1358186603,0.0812894851,-0.0107793091,0.5259879827,0.1519594789,0.8704359531,0.1582149416,-0.2194377482,0.1683554202,-0.1485036314,-0.010637085,-0.0340552628,-0.3892847896,0.1844976842,-0.0443505682,0.1226070747,0.1395881474,-0.2094922066,-0.1556345224,0.1454913467,-0.0947229341,-0.2187247276,-0.2757362723,-0.0225727931,0.0979045033,0.0983442664,0.0477664843,0.1866315752,-0.4435558915,0.0679604933,0.1192273498,0.0044029737,0.0480467714,-0.0468376577,-1.0162568092,-0.0913212076,-0.1702794135,0.3654927015,0.2294975519,0.4982431829,-0.0179926045,-0.0469948612,0.0515056588,-0.2121096551,0.8495112658,0.2505568564,-0.0532124676,0.0796275288,0.0189491622,-0.1545509845,0.0577008165,-0.1632412374,0.6255755424,0.4945641458,0.7852733731,-0.4374043345,-0.0893706232,0.0312036462,-0.0599599853,-0.1476593465,0.0810301974,-0.2087632865,-0.180411458,-0.5762064457,0.2585665584,0.2810131907,0.1928339154,0.152831018,-0.2231671363,0.0690152347,-0.2840030491,-0.0985406041,0.2312042862,0.2143349946,0.1858991086,-0.0022402669,0.145538196,-0.1561143696,-0.2068816274,0.189942807,0.25589481,-0.6231859922,0.2523162365,0.1651500016,0.560379684,0.1582307816,-0.1150253788,0.2946467102,-0.0105134277,0.479396224,-0.0390628837,0.4788197279,0.3626244366,0.2778424919,-0.301658839,-0.1384537518,0.3859476149,-0.0154792443,0.0721439794,0.0153059829,0.2642338276,-0.1523793787,0.578543961,0.1050156429,1.0689468384,-0.0962891653,-0.091162324,0.1268847138,0.2655840218,0.5600728393,-0.2914608121,0.2052387893,-0.5174925327,-0.0904992297,0.0042558508,-0.1398295611,0.1350088418,0.1887042522,-0.4300575256,0.1389254481,0.0992789939,0.4275937676,0.0399220213,0.4134412706,-0.0181035865,-0.4771104455,-0.1259956062,0.0662658066,-0.2162788957,0.2179785967,0.0931357071,0.0167735219,-0.2299908102,-0.3334072232,-0.0685394555,0.0877202675,-0.3980352879,0.6731894612,0.342954129,-0.1945347339,0.2956734002,0.2294498533,0.1547928452,0.2516962588,-0.0637095571,0.1682659537,0.1497499645,-0.0029832826,0.1993835419,-0.0682335496,0.1190654561,-0.0103186648,-0.1015117094,0.1754909605,-0.0124609089,-0.2762164176,-0.2717876732,-0.1240058616,0.2903387249,-0.3395356238,-0.2814121246,-0.0395575985,-0.2850346565,-0.2035131902,0.0220345296,-0.0706589893,-0.3747996092,0.2263817191,0.2376686484,-0.1426926404,0.1695863307,0.3712310195,-0.0255008675,-0.0888265967,0.5584918261,0.0715670735,-0.1946096569,-0.2024845928,0.1518792063,-0.033778619,-0.2271941155,0.1091578379,-0.1162661538,-0.2636434436,-0.2287956625,0.3435637355,-0.0256999768,0.0402624384,-0.2143465132,-0.4193456769,-0.5076628923,0.0499494001,-0.1312001497,0.1811650097,0.4385974407,0.4080056846,-0.2568692267,0.3760201633,-0.2495917976,-0.0800243616,-0.5987675786,0.2662300766,0.0188432653,-0.5459757447,-0.0275539905,0.0216406416,0.0264433939,0.2940588593,-0.1200792193,-0.1588124633,-0.0702605769,0.1605631262,0.2280812263,-0.1013725996,-0.2458603829,0.1184105352,-0.1792993993,-0.1568111479,0.1025064737,0.1319525987,0.0121633019,0.1235533729,-0.0908354744,-0.1170054078,-0.0513545312,-0.0865232125,0.0492475443,-0.1551797241,0.200568229,0.3887774944,-0.1900877953,-0.0090764407,-0.2909185886,0.1795438677,0.6930254102,0.0096776029,0.3933944106,-0.3145877421,-0.1508094966,0.229211092,-0.2073029727,0.1850663126,-0.3238908648,-0.095886223,0.2286647111,0.1294852048,-0.2605695128,-0.2062440068,0.5028803945,-0.3981368542,0.1572274119,0.1296645403,-0.0983623341,0.2678847909,0.1832061708,0.1603047252,0.7458079457,0.0117864432,0.1765505672,-0.0622489601,-0.092001304,0.0551159568,0.512357831,0.1351608634,0.2308288515,0.1101420224,-0.2032733262,0.3773396611,-0.4285535812,0.2233328372,0.3698289692,-0.1252685338,-0.193761304,0.208747223,-0.1403710991,0.2644395828,-0.04349453,0.3522603214,-0.1093692854,-0.1740659326,-0.1687214673,0.4904870093,-0.3190308213,-0.0490976721,-0.0212323908,-0.0759386569,-0.1959419698,-0.2856764197,0.0508428477,-0.0616605505,0.1097760573,0.0157869291,-0.0804274678,0.2032717317,-0.1478792727,0.1775085181,0.2425409406,-0.3291356266,0.2424013913,0.1096198335,0.2629410625,-0.2478345931,0.1149145141,0.3023824096,0.2686603367,-0.4128156602,0.1197531819,-0.061437685,-0.1552257091,0.1539587826,0.1354496777,-0.1843046695,-0.2192908376,0.2301807553,0.0778790861,-0.2095594406,-0.1722642481,0.1902220994,0.1092769131,0.1233894005,-0.0497298315,-0.0805690289,-0.2053932846,-0.3169712722,-0.1771351695,-0.3990464211,-0.0486690924,0.0820412487,-0.2334050238,0.2352662832,-0.3069973588,0.0806782916,0.0823595896,0.3644183874,0.3155186474,0.0889510736,-0.3957800269,-0.0233593192,-0.4049155414,0.2132003903,0.3097579479,-0.1938925534,0.085277535,0.1949707717,0.060536474,0.3144331872,-0.1063973233,0.0570466518,-0.1318134964,-0.0067865741,-0.2199010402,0.134248659,-0.165356636,0.1062286198,0.2053378969,-0.3196312189,0.3005925119,0.2335089296,-0.0160384122,-0.2037670761,-0.0731834322,-0.2135284543,-0.3241593242,-0.0482806824,0.2011152506,0.2985326052,-0.2826665938,0.180134356,0.2097664326,-0.0438413583,-0.2875527143,-0.193121776,0.1564357877,0.4927534163,-0.0002099175,0.1263695359,-0.3023884594,-0.1757035553,0.0875676423,-0.2813342214,-0.242300421,0.3333650529,-0.205700621,0.1753344238,-0.0147263901,0.2186222672,0.0708544925,0.2834080458,-0.2453328669,-0.4549573958,0.6601228714,-0.4289312661,-0.4800792336,-0.0562647842,-0.0659726113,0.1373787075,-0.2057521641,-0.2559036016,-0.1504520178,0.2913421392,0.084562473,-0.2006697506,0.0518865958,0.1554601938,-0.0358313434,-0.1501895785,0.3721503019,0.1291872263,0.0628954768,0.1038767546,-0.2139432728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/665","title":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects","comments":"We can also update the `BertJapaneseTokenizer` in `transformers` as you just shown @lhoestq to make it compatible with pickle. It will be faster than asking on fugashi 's repo and good for the other users of `transformers` as well.\r\n\r\nI'm currently working on `transformers` I'll include it in the https:\/\/github.com\/huggingface\/transformers\/pull\/7141 PR and the next release of `transformers`.","body":"I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n","comment_length":57,"text":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects\nI load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n\nWe can also update the `BertJapaneseTokenizer` in `transformers` as you just shown @lhoestq to make it compatible with pickle. It will be faster than asking on fugashi 's repo and good for the other users of `transformers` as well.\r\n\r\nI'm currently working on `transformers` I'll include it in the https:\/\/github.com\/huggingface\/transformers\/pull\/7141 PR and the next release of `transformers`.","embeddings":[-0.2387796193,-0.2978491187,-0.019680649,0.2330349237,0.4450323582,-0.1729919016,0.2851519585,0.1309388727,-0.2760912478,0.1103135347,-0.0697863474,0.4891909063,-0.0087260343,-0.046733927,-0.145709604,-0.111473918,0.0535604171,0.207853511,0.0181137323,-0.0561074987,-0.2787823677,0.2074424773,-0.4731775522,0.1585230231,-0.2825988233,-0.2674281299,0.1064047962,-0.1985800564,-0.2693119049,-0.3823218942,-0.1436450779,-0.073365584,-0.0288482774,0.5633369088,-0.0001192604,0.0994183272,0.108954519,-0.145972386,0.0052242242,-0.0495663248,-0.0748076513,-0.3035497069,-0.205873996,-0.4582723975,-0.1556460112,-0.1438439935,0.0111553641,-0.3324888647,0.5145235062,0.2914209962,0.139978081,0.4932267964,0.0206352938,0.0532785431,0.015327774,0.0298241358,-0.1314535141,-0.2608632743,0.1487917304,0.2165469229,-0.1302717179,0.3254893422,-0.1613031477,-0.2627820969,0.0994700417,0.094937481,0.1041127741,-0.4675594568,0.3230481744,-0.002158023,0.2189722508,-0.020810321,-0.2567602098,-0.2679677308,-0.0794852078,0.0625912696,-0.0412523523,-0.2022152543,0.0776130408,0.0512863845,-0.3526940644,-0.1723531634,-0.1798734814,0.108233422,0.0383866839,0.3674840629,-0.2666110098,0.2930521965,0.0967984498,-0.2501663864,-0.1138216779,-0.0991195515,0.2264450341,0.4768724144,-0.3883422315,-0.3008864224,-0.0943802446,-0.409999162,0.3477460146,-0.0384070538,-0.2276796252,0.2554124892,-0.0489559509,0.1893791109,0.0231759958,0.3520210385,0.2632215321,0.3699644506,0.0116726784,-0.1440164447,0.0652595013,-0.04887354,-0.0822229534,-0.2215761691,-0.0569127798,0.2582862377,0.3595126569,-0.0186132845,-0.046238631,0.0278620608,-0.2262803167,0.0427433066,0.1734637767,0.4068900049,0.1080137938,0.2318411618,-0.3183892965,0.0902408659,-0.124280192,0.0253062062,-0.1654862761,0.214023307,-0.0066213412,-0.1393794417,-0.1643196195,0.0698126107,0.2013424933,-0.026837565,0.0632021129,-0.2202773243,0.061538171,-0.1790308952,0.2758253217,-0.0943112522,-0.0502242185,0.1737568676,0.2081712335,-0.40948084,-0.2060563862,0.0352333188,-0.34115237,0.01733163,0.1311888844,0.1185405329,0.0562376231,0.0476904213,-0.3749628663,0.2920556068,0.4146037698,0.0075913849,0.0322678164,-0.1267381161,-0.2016420662,-0.111127086,0.2182486653,0.1899360716,-0.1373286545,-0.2079769075,0.2503727973,-0.002062849,0.1535669416,0.4049597979,-0.0530274957,0.2562201917,-0.1467852145,0.559774518,0.4704823494,-0.6255075336,-0.5309312344,0.1032184213,-0.4468844831,-0.1779434234,-0.035346359,0.0416382365,0.5613888502,0.1241554692,-0.113206543,0.3392548561,0.072295852,0.2676001787,-0.1563941687,-0.1415297985,0.4092992246,0.0926817507,0.21709764,0.116488941,-0.1682532728,0.1971376091,-0.1022469327,-0.1236947551,0.0914442316,0.1902052313,0.1654276252,-0.2108140886,0.0806120411,-0.1629784405,-0.5545789599,0.0651864186,-0.3476859927,0.2234149575,-0.2890501916,-0.194772318,-0.2887653708,0.1541353166,-0.4519188404,-0.0524003021,0.083847262,-0.0306146815,0.1889378428,-0.0136042386,0.0192883443,-0.0761402398,0.1458211094,0.186094299,-0.2552472055,0.2062030882,-0.2569340169,0.021148419,-0.05973031,0.09906286,0.408500433,-0.1759835035,-0.2551455796,0.2755496502,0.1720520854,-0.0848676786,-0.0500598922,-0.0350354798,0.2658101916,-0.2096956968,-0.0644069538,0.1785884053,0.166718334,-0.0287600234,-0.0929355696,0.4431660473,0.224703446,0.3178621531,-0.0191053115,0.1625137478,0.2412854582,-0.10274712,-0.0464949757,-0.1355438232,-0.1764924526,0.0628632903,0.2340398133,-0.0736531839,0.0656425655,-0.0590749644,0.8479928374,-0.0683462247,0.0793827847,0.204474166,-0.116351068,0.0497056618,-0.0788842216,-0.4018104076,0.0417372212,0.0392285287,-0.1734893918,-0.1066208854,0.2546639144,0.1459451169,0.041525092,0.3150502145,0.1146959588,0.2255442888,0.1599868983,0.1129512191,-0.1720335782,-0.1941431463,0.1629729271,0.1452347934,-0.1468617916,-0.0244321302,-0.0460567623,0.0444771908,0.0104709063,-0.0486729778,-0.195032984,-0.2504672706,0.0337895602,-0.0237582624,0.0957349986,0.4495555162,0.3668554127,0.3240292668,0.427821368,-0.1099201962,-0.170441404,-0.1969794482,-0.0876939595,-0.0677347481,0.1707262248,-0.0053565041,-0.0357478224,0.0103845401,0.1024920642,-0.456127733,-0.4291423857,0.2537454963,-0.178284511,0.1860092133,-0.0141354213,0.1210128963,-0.3429493308,-0.2921084464,0.2118502706,-0.2194129676,-0.3013744652,0.0706816763,-0.0610883981,-0.3209867477,0.0511774272,-0.2405867577,-0.3765725493,-0.3376917541,0.363016367,-0.008248332,0.1135794222,0.1541582793,-0.0030258943,0.2885194421,-0.0200715531,-0.0359078087,-0.2440013885,0.0543059148,0.4122084975,-0.1826187968,-0.2826860547,-0.3915768564,-0.1759163439,0.0525789037,-0.1293791682,-0.1020923853,-0.2275468558,-0.2711945474,0.2238460481,-0.1608631462,0.0458878055,0.558042407,-0.0492363572,0.0907655656,0.0115672005,-0.1609315425,0.2720036209,0.1358186603,0.0812894851,-0.0107793091,0.5259879827,0.1519594789,0.8704359531,0.1582149416,-0.2194377482,0.1683554202,-0.1485036314,-0.010637085,-0.0340552628,-0.3892847896,0.1844976842,-0.0443505682,0.1226070747,0.1395881474,-0.2094922066,-0.1556345224,0.1454913467,-0.0947229341,-0.2187247276,-0.2757362723,-0.0225727931,0.0979045033,0.0983442664,0.0477664843,0.1866315752,-0.4435558915,0.0679604933,0.1192273498,0.0044029737,0.0480467714,-0.0468376577,-1.0162568092,-0.0913212076,-0.1702794135,0.3654927015,0.2294975519,0.4982431829,-0.0179926045,-0.0469948612,0.0515056588,-0.2121096551,0.8495112658,0.2505568564,-0.0532124676,0.0796275288,0.0189491622,-0.1545509845,0.0577008165,-0.1632412374,0.6255755424,0.4945641458,0.7852733731,-0.4374043345,-0.0893706232,0.0312036462,-0.0599599853,-0.1476593465,0.0810301974,-0.2087632865,-0.180411458,-0.5762064457,0.2585665584,0.2810131907,0.1928339154,0.152831018,-0.2231671363,0.0690152347,-0.2840030491,-0.0985406041,0.2312042862,0.2143349946,0.1858991086,-0.0022402669,0.145538196,-0.1561143696,-0.2068816274,0.189942807,0.25589481,-0.6231859922,0.2523162365,0.1651500016,0.560379684,0.1582307816,-0.1150253788,0.2946467102,-0.0105134277,0.479396224,-0.0390628837,0.4788197279,0.3626244366,0.2778424919,-0.301658839,-0.1384537518,0.3859476149,-0.0154792443,0.0721439794,0.0153059829,0.2642338276,-0.1523793787,0.578543961,0.1050156429,1.0689468384,-0.0962891653,-0.091162324,0.1268847138,0.2655840218,0.5600728393,-0.2914608121,0.2052387893,-0.5174925327,-0.0904992297,0.0042558508,-0.1398295611,0.1350088418,0.1887042522,-0.4300575256,0.1389254481,0.0992789939,0.4275937676,0.0399220213,0.4134412706,-0.0181035865,-0.4771104455,-0.1259956062,0.0662658066,-0.2162788957,0.2179785967,0.0931357071,0.0167735219,-0.2299908102,-0.3334072232,-0.0685394555,0.0877202675,-0.3980352879,0.6731894612,0.342954129,-0.1945347339,0.2956734002,0.2294498533,0.1547928452,0.2516962588,-0.0637095571,0.1682659537,0.1497499645,-0.0029832826,0.1993835419,-0.0682335496,0.1190654561,-0.0103186648,-0.1015117094,0.1754909605,-0.0124609089,-0.2762164176,-0.2717876732,-0.1240058616,0.2903387249,-0.3395356238,-0.2814121246,-0.0395575985,-0.2850346565,-0.2035131902,0.0220345296,-0.0706589893,-0.3747996092,0.2263817191,0.2376686484,-0.1426926404,0.1695863307,0.3712310195,-0.0255008675,-0.0888265967,0.5584918261,0.0715670735,-0.1946096569,-0.2024845928,0.1518792063,-0.033778619,-0.2271941155,0.1091578379,-0.1162661538,-0.2636434436,-0.2287956625,0.3435637355,-0.0256999768,0.0402624384,-0.2143465132,-0.4193456769,-0.5076628923,0.0499494001,-0.1312001497,0.1811650097,0.4385974407,0.4080056846,-0.2568692267,0.3760201633,-0.2495917976,-0.0800243616,-0.5987675786,0.2662300766,0.0188432653,-0.5459757447,-0.0275539905,0.0216406416,0.0264433939,0.2940588593,-0.1200792193,-0.1588124633,-0.0702605769,0.1605631262,0.2280812263,-0.1013725996,-0.2458603829,0.1184105352,-0.1792993993,-0.1568111479,0.1025064737,0.1319525987,0.0121633019,0.1235533729,-0.0908354744,-0.1170054078,-0.0513545312,-0.0865232125,0.0492475443,-0.1551797241,0.200568229,0.3887774944,-0.1900877953,-0.0090764407,-0.2909185886,0.1795438677,0.6930254102,0.0096776029,0.3933944106,-0.3145877421,-0.1508094966,0.229211092,-0.2073029727,0.1850663126,-0.3238908648,-0.095886223,0.2286647111,0.1294852048,-0.2605695128,-0.2062440068,0.5028803945,-0.3981368542,0.1572274119,0.1296645403,-0.0983623341,0.2678847909,0.1832061708,0.1603047252,0.7458079457,0.0117864432,0.1765505672,-0.0622489601,-0.092001304,0.0551159568,0.512357831,0.1351608634,0.2308288515,0.1101420224,-0.2032733262,0.3773396611,-0.4285535812,0.2233328372,0.3698289692,-0.1252685338,-0.193761304,0.208747223,-0.1403710991,0.2644395828,-0.04349453,0.3522603214,-0.1093692854,-0.1740659326,-0.1687214673,0.4904870093,-0.3190308213,-0.0490976721,-0.0212323908,-0.0759386569,-0.1959419698,-0.2856764197,0.0508428477,-0.0616605505,0.1097760573,0.0157869291,-0.0804274678,0.2032717317,-0.1478792727,0.1775085181,0.2425409406,-0.3291356266,0.2424013913,0.1096198335,0.2629410625,-0.2478345931,0.1149145141,0.3023824096,0.2686603367,-0.4128156602,0.1197531819,-0.061437685,-0.1552257091,0.1539587826,0.1354496777,-0.1843046695,-0.2192908376,0.2301807553,0.0778790861,-0.2095594406,-0.1722642481,0.1902220994,0.1092769131,0.1233894005,-0.0497298315,-0.0805690289,-0.2053932846,-0.3169712722,-0.1771351695,-0.3990464211,-0.0486690924,0.0820412487,-0.2334050238,0.2352662832,-0.3069973588,0.0806782916,0.0823595896,0.3644183874,0.3155186474,0.0889510736,-0.3957800269,-0.0233593192,-0.4049155414,0.2132003903,0.3097579479,-0.1938925534,0.085277535,0.1949707717,0.060536474,0.3144331872,-0.1063973233,0.0570466518,-0.1318134964,-0.0067865741,-0.2199010402,0.134248659,-0.165356636,0.1062286198,0.2053378969,-0.3196312189,0.3005925119,0.2335089296,-0.0160384122,-0.2037670761,-0.0731834322,-0.2135284543,-0.3241593242,-0.0482806824,0.2011152506,0.2985326052,-0.2826665938,0.180134356,0.2097664326,-0.0438413583,-0.2875527143,-0.193121776,0.1564357877,0.4927534163,-0.0002099175,0.1263695359,-0.3023884594,-0.1757035553,0.0875676423,-0.2813342214,-0.242300421,0.3333650529,-0.205700621,0.1753344238,-0.0147263901,0.2186222672,0.0708544925,0.2834080458,-0.2453328669,-0.4549573958,0.6601228714,-0.4289312661,-0.4800792336,-0.0562647842,-0.0659726113,0.1373787075,-0.2057521641,-0.2559036016,-0.1504520178,0.2913421392,0.084562473,-0.2006697506,0.0518865958,0.1554601938,-0.0358313434,-0.1501895785,0.3721503019,0.1291872263,0.0628954768,0.1038767546,-0.2139432728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/665","title":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects","comments":"Thank you for the rapid and polite response!\r\n\r\n@lhoestq Thanks for the suggestion! I've passed the pickle phase, but another `ArrowInvalid` problem occored. I created another issue #687 .\r\n\r\n@thomwolf Wow, really fast work. I'm looking forward to the next release \ud83e\udd17","body":"I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n","comment_length":42,"text":"runing dataset.map, it raises TypeError: can't pickle Tokenizer objects\nI load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`.\r\n\r\n```\r\ndef convert_to_features(example):\r\n # Tokenize contexts and questions (as pairs of inputs)\r\n input_pairs = [example['question'], example['context']]\r\n encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512)\r\n context_encodings = tokenizer.encode_plus(example['context'])\r\n \r\n\r\n # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes.\r\n # this will give us the position of answer span in the context text\r\n start_idx, end_idx = get_correct_alignement(example['context'], example['answers'])\r\n start_positions_context = context_encodings.char_to_token(start_idx)\r\n end_positions_context = context_encodings.char_to_token(end_idx-1)\r\n\r\n # here we will compute the start and end position of the answer in the whole example\r\n # as the example is encoded like this <s> question<\/s><\/s> context<\/s>\r\n # and we know the postion of the answer in the context\r\n # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens)\r\n # this will give us the position of the answer span in whole example \r\n sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id)\r\n start_positions = start_positions_context + sep_idx + 1\r\n end_positions = end_positions_context + sep_idx + 1\r\n\r\n if end_positions > 512:\r\n start_positions, end_positions = 0, 0\r\n\r\n encodings.update({'start_positions': start_positions,\r\n 'end_positions': end_positions,\r\n 'attention_mask': encodings['attention_mask']})\r\n return encodings\r\n```\r\n\r\nThen I run `dataset.map(convert_to_features)`, it raise\r\n```\r\nIn [59]: a.map(convert_to_features) \r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-59-c453b508761d> in <module>\r\n----> 1 a.map(convert_to_features)\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/utils\/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in dump(self, obj)\r\n 435 if self.proto >= 4:\r\n 436 self.framer.start_framing()\r\n--> 437 self.save(obj)\r\n 438 self.write(STOP)\r\n 439 self.framer.end_framing()\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 636 else:\r\n 637 save(func)\r\n--> 638 save(args)\r\n 639 write(REDUCE)\r\n 640 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_tuple(self, obj)\r\n 787 write(MARK)\r\n 788 for element in obj:\r\n--> 789 save(element)\r\n 790 \r\n 791 if id(obj) in memo:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 547 \r\n 548 # Save the reduce() output and finally memoize the object\r\n--> 549 self.save_reduce(obj=obj, *rv)\r\n 550 \r\n 551 def persistent_id(self, obj):\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 660 \r\n 661 if state is not None:\r\n--> 662 save(state)\r\n 663 write(BUILD)\r\n 664 \r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 502 f = self.dispatch.get(t)\r\n 503 if f is not None:\r\n--> 504 f(self, obj) # Call unbound method with explicit self\r\n 505 return\r\n 506 \r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/dill\/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save_dict(self, obj)\r\n 857 \r\n 858 self.memoize(obj)\r\n--> 859 self._batch_setitems(obj.items())\r\n 860 \r\n 861 dispatch[dict] = save_dict\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in _batch_setitems(self, items)\r\n 883 for k, v in tmp:\r\n 884 save(k)\r\n--> 885 save(v)\r\n 886 write(SETITEMS)\r\n 887 elif n:\r\n\r\n\/opt\/conda\/lib\/python3.7\/pickle.py in save(self, obj, save_persistent_id)\r\n 522 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 523 if reduce is not None:\r\n--> 524 rv = reduce(self.proto)\r\n 525 else:\r\n 526 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tokenizer objects\r\n```\r\n\r\n\nThank you for the rapid and polite response!\r\n\r\n@lhoestq Thanks for the suggestion! I've passed the pickle phase, but another `ArrowInvalid` problem occored. I created another issue #687 .\r\n\r\n@thomwolf Wow, really fast work. I'm looking forward to the next release \ud83e\udd17","embeddings":[-0.2387796193,-0.2978491187,-0.019680649,0.2330349237,0.4450323582,-0.1729919016,0.2851519585,0.1309388727,-0.2760912478,0.1103135347,-0.0697863474,0.4891909063,-0.0087260343,-0.046733927,-0.145709604,-0.111473918,0.0535604171,0.207853511,0.0181137323,-0.0561074987,-0.2787823677,0.2074424773,-0.4731775522,0.1585230231,-0.2825988233,-0.2674281299,0.1064047962,-0.1985800564,-0.2693119049,-0.3823218942,-0.1436450779,-0.073365584,-0.0288482774,0.5633369088,-0.0001192604,0.0994183272,0.108954519,-0.145972386,0.0052242242,-0.0495663248,-0.0748076513,-0.3035497069,-0.205873996,-0.4582723975,-0.1556460112,-0.1438439935,0.0111553641,-0.3324888647,0.5145235062,0.2914209962,0.139978081,0.4932267964,0.0206352938,0.0532785431,0.015327774,0.0298241358,-0.1314535141,-0.2608632743,0.1487917304,0.2165469229,-0.1302717179,0.3254893422,-0.1613031477,-0.2627820969,0.0994700417,0.094937481,0.1041127741,-0.4675594568,0.3230481744,-0.002158023,0.2189722508,-0.020810321,-0.2567602098,-0.2679677308,-0.0794852078,0.0625912696,-0.0412523523,-0.2022152543,0.0776130408,0.0512863845,-0.3526940644,-0.1723531634,-0.1798734814,0.108233422,0.0383866839,0.3674840629,-0.2666110098,0.2930521965,0.0967984498,-0.2501663864,-0.1138216779,-0.0991195515,0.2264450341,0.4768724144,-0.3883422315,-0.3008864224,-0.0943802446,-0.409999162,0.3477460146,-0.0384070538,-0.2276796252,0.2554124892,-0.0489559509,0.1893791109,0.0231759958,0.3520210385,0.2632215321,0.3699644506,0.0116726784,-0.1440164447,0.0652595013,-0.04887354,-0.0822229534,-0.2215761691,-0.0569127798,0.2582862377,0.3595126569,-0.0186132845,-0.046238631,0.0278620608,-0.2262803167,0.0427433066,0.1734637767,0.4068900049,0.1080137938,0.2318411618,-0.3183892965,0.0902408659,-0.124280192,0.0253062062,-0.1654862761,0.214023307,-0.0066213412,-0.1393794417,-0.1643196195,0.0698126107,0.2013424933,-0.026837565,0.0632021129,-0.2202773243,0.061538171,-0.1790308952,0.2758253217,-0.0943112522,-0.0502242185,0.1737568676,0.2081712335,-0.40948084,-0.2060563862,0.0352333188,-0.34115237,0.01733163,0.1311888844,0.1185405329,0.0562376231,0.0476904213,-0.3749628663,0.2920556068,0.4146037698,0.0075913849,0.0322678164,-0.1267381161,-0.2016420662,-0.111127086,0.2182486653,0.1899360716,-0.1373286545,-0.2079769075,0.2503727973,-0.002062849,0.1535669416,0.4049597979,-0.0530274957,0.2562201917,-0.1467852145,0.559774518,0.4704823494,-0.6255075336,-0.5309312344,0.1032184213,-0.4468844831,-0.1779434234,-0.035346359,0.0416382365,0.5613888502,0.1241554692,-0.113206543,0.3392548561,0.072295852,0.2676001787,-0.1563941687,-0.1415297985,0.4092992246,0.0926817507,0.21709764,0.116488941,-0.1682532728,0.1971376091,-0.1022469327,-0.1236947551,0.0914442316,0.1902052313,0.1654276252,-0.2108140886,0.0806120411,-0.1629784405,-0.5545789599,0.0651864186,-0.3476859927,0.2234149575,-0.2890501916,-0.194772318,-0.2887653708,0.1541353166,-0.4519188404,-0.0524003021,0.083847262,-0.0306146815,0.1889378428,-0.0136042386,0.0192883443,-0.0761402398,0.1458211094,0.186094299,-0.2552472055,0.2062030882,-0.2569340169,0.021148419,-0.05973031,0.09906286,0.408500433,-0.1759835035,-0.2551455796,0.2755496502,0.1720520854,-0.0848676786,-0.0500598922,-0.0350354798,0.2658101916,-0.2096956968,-0.0644069538,0.1785884053,0.166718334,-0.0287600234,-0.0929355696,0.4431660473,0.224703446,0.3178621531,-0.0191053115,0.1625137478,0.2412854582,-0.10274712,-0.0464949757,-0.1355438232,-0.1764924526,0.0628632903,0.2340398133,-0.0736531839,0.0656425655,-0.0590749644,0.8479928374,-0.0683462247,0.0793827847,0.204474166,-0.116351068,0.0497056618,-0.0788842216,-0.4018104076,0.0417372212,0.0392285287,-0.1734893918,-0.1066208854,0.2546639144,0.1459451169,0.041525092,0.3150502145,0.1146959588,0.2255442888,0.1599868983,0.1129512191,-0.1720335782,-0.1941431463,0.1629729271,0.1452347934,-0.1468617916,-0.0244321302,-0.0460567623,0.0444771908,0.0104709063,-0.0486729778,-0.195032984,-0.2504672706,0.0337895602,-0.0237582624,0.0957349986,0.4495555162,0.3668554127,0.3240292668,0.427821368,-0.1099201962,-0.170441404,-0.1969794482,-0.0876939595,-0.0677347481,0.1707262248,-0.0053565041,-0.0357478224,0.0103845401,0.1024920642,-0.456127733,-0.4291423857,0.2537454963,-0.178284511,0.1860092133,-0.0141354213,0.1210128963,-0.3429493308,-0.2921084464,0.2118502706,-0.2194129676,-0.3013744652,0.0706816763,-0.0610883981,-0.3209867477,0.0511774272,-0.2405867577,-0.3765725493,-0.3376917541,0.363016367,-0.008248332,0.1135794222,0.1541582793,-0.0030258943,0.2885194421,-0.0200715531,-0.0359078087,-0.2440013885,0.0543059148,0.4122084975,-0.1826187968,-0.2826860547,-0.3915768564,-0.1759163439,0.0525789037,-0.1293791682,-0.1020923853,-0.2275468558,-0.2711945474,0.2238460481,-0.1608631462,0.0458878055,0.558042407,-0.0492363572,0.0907655656,0.0115672005,-0.1609315425,0.2720036209,0.1358186603,0.0812894851,-0.0107793091,0.5259879827,0.1519594789,0.8704359531,0.1582149416,-0.2194377482,0.1683554202,-0.1485036314,-0.010637085,-0.0340552628,-0.3892847896,0.1844976842,-0.0443505682,0.1226070747,0.1395881474,-0.2094922066,-0.1556345224,0.1454913467,-0.0947229341,-0.2187247276,-0.2757362723,-0.0225727931,0.0979045033,0.0983442664,0.0477664843,0.1866315752,-0.4435558915,0.0679604933,0.1192273498,0.0044029737,0.0480467714,-0.0468376577,-1.0162568092,-0.0913212076,-0.1702794135,0.3654927015,0.2294975519,0.4982431829,-0.0179926045,-0.0469948612,0.0515056588,-0.2121096551,0.8495112658,0.2505568564,-0.0532124676,0.0796275288,0.0189491622,-0.1545509845,0.0577008165,-0.1632412374,0.6255755424,0.4945641458,0.7852733731,-0.4374043345,-0.0893706232,0.0312036462,-0.0599599853,-0.1476593465,0.0810301974,-0.2087632865,-0.180411458,-0.5762064457,0.2585665584,0.2810131907,0.1928339154,0.152831018,-0.2231671363,0.0690152347,-0.2840030491,-0.0985406041,0.2312042862,0.2143349946,0.1858991086,-0.0022402669,0.145538196,-0.1561143696,-0.2068816274,0.189942807,0.25589481,-0.6231859922,0.2523162365,0.1651500016,0.560379684,0.1582307816,-0.1150253788,0.2946467102,-0.0105134277,0.479396224,-0.0390628837,0.4788197279,0.3626244366,0.2778424919,-0.301658839,-0.1384537518,0.3859476149,-0.0154792443,0.0721439794,0.0153059829,0.2642338276,-0.1523793787,0.578543961,0.1050156429,1.0689468384,-0.0962891653,-0.091162324,0.1268847138,0.2655840218,0.5600728393,-0.2914608121,0.2052387893,-0.5174925327,-0.0904992297,0.0042558508,-0.1398295611,0.1350088418,0.1887042522,-0.4300575256,0.1389254481,0.0992789939,0.4275937676,0.0399220213,0.4134412706,-0.0181035865,-0.4771104455,-0.1259956062,0.0662658066,-0.2162788957,0.2179785967,0.0931357071,0.0167735219,-0.2299908102,-0.3334072232,-0.0685394555,0.0877202675,-0.3980352879,0.6731894612,0.342954129,-0.1945347339,0.2956734002,0.2294498533,0.1547928452,0.2516962588,-0.0637095571,0.1682659537,0.1497499645,-0.0029832826,0.1993835419,-0.0682335496,0.1190654561,-0.0103186648,-0.1015117094,0.1754909605,-0.0124609089,-0.2762164176,-0.2717876732,-0.1240058616,0.2903387249,-0.3395356238,-0.2814121246,-0.0395575985,-0.2850346565,-0.2035131902,0.0220345296,-0.0706589893,-0.3747996092,0.2263817191,0.2376686484,-0.1426926404,0.1695863307,0.3712310195,-0.0255008675,-0.0888265967,0.5584918261,0.0715670735,-0.1946096569,-0.2024845928,0.1518792063,-0.033778619,-0.2271941155,0.1091578379,-0.1162661538,-0.2636434436,-0.2287956625,0.3435637355,-0.0256999768,0.0402624384,-0.2143465132,-0.4193456769,-0.5076628923,0.0499494001,-0.1312001497,0.1811650097,0.4385974407,0.4080056846,-0.2568692267,0.3760201633,-0.2495917976,-0.0800243616,-0.5987675786,0.2662300766,0.0188432653,-0.5459757447,-0.0275539905,0.0216406416,0.0264433939,0.2940588593,-0.1200792193,-0.1588124633,-0.0702605769,0.1605631262,0.2280812263,-0.1013725996,-0.2458603829,0.1184105352,-0.1792993993,-0.1568111479,0.1025064737,0.1319525987,0.0121633019,0.1235533729,-0.0908354744,-0.1170054078,-0.0513545312,-0.0865232125,0.0492475443,-0.1551797241,0.200568229,0.3887774944,-0.1900877953,-0.0090764407,-0.2909185886,0.1795438677,0.6930254102,0.0096776029,0.3933944106,-0.3145877421,-0.1508094966,0.229211092,-0.2073029727,0.1850663126,-0.3238908648,-0.095886223,0.2286647111,0.1294852048,-0.2605695128,-0.2062440068,0.5028803945,-0.3981368542,0.1572274119,0.1296645403,-0.0983623341,0.2678847909,0.1832061708,0.1603047252,0.7458079457,0.0117864432,0.1765505672,-0.0622489601,-0.092001304,0.0551159568,0.512357831,0.1351608634,0.2308288515,0.1101420224,-0.2032733262,0.3773396611,-0.4285535812,0.2233328372,0.3698289692,-0.1252685338,-0.193761304,0.208747223,-0.1403710991,0.2644395828,-0.04349453,0.3522603214,-0.1093692854,-0.1740659326,-0.1687214673,0.4904870093,-0.3190308213,-0.0490976721,-0.0212323908,-0.0759386569,-0.1959419698,-0.2856764197,0.0508428477,-0.0616605505,0.1097760573,0.0157869291,-0.0804274678,0.2032717317,-0.1478792727,0.1775085181,0.2425409406,-0.3291356266,0.2424013913,0.1096198335,0.2629410625,-0.2478345931,0.1149145141,0.3023824096,0.2686603367,-0.4128156602,0.1197531819,-0.061437685,-0.1552257091,0.1539587826,0.1354496777,-0.1843046695,-0.2192908376,0.2301807553,0.0778790861,-0.2095594406,-0.1722642481,0.1902220994,0.1092769131,0.1233894005,-0.0497298315,-0.0805690289,-0.2053932846,-0.3169712722,-0.1771351695,-0.3990464211,-0.0486690924,0.0820412487,-0.2334050238,0.2352662832,-0.3069973588,0.0806782916,0.0823595896,0.3644183874,0.3155186474,0.0889510736,-0.3957800269,-0.0233593192,-0.4049155414,0.2132003903,0.3097579479,-0.1938925534,0.085277535,0.1949707717,0.060536474,0.3144331872,-0.1063973233,0.0570466518,-0.1318134964,-0.0067865741,-0.2199010402,0.134248659,-0.165356636,0.1062286198,0.2053378969,-0.3196312189,0.3005925119,0.2335089296,-0.0160384122,-0.2037670761,-0.0731834322,-0.2135284543,-0.3241593242,-0.0482806824,0.2011152506,0.2985326052,-0.2826665938,0.180134356,0.2097664326,-0.0438413583,-0.2875527143,-0.193121776,0.1564357877,0.4927534163,-0.0002099175,0.1263695359,-0.3023884594,-0.1757035553,0.0875676423,-0.2813342214,-0.242300421,0.3333650529,-0.205700621,0.1753344238,-0.0147263901,0.2186222672,0.0708544925,0.2834080458,-0.2453328669,-0.4549573958,0.6601228714,-0.4289312661,-0.4800792336,-0.0562647842,-0.0659726113,0.1373787075,-0.2057521641,-0.2559036016,-0.1504520178,0.2913421392,0.084562473,-0.2006697506,0.0518865958,0.1554601938,-0.0358313434,-0.1501895785,0.3721503019,0.1291872263,0.0628954768,0.1038767546,-0.2139432728]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/664","title":"load_dataset from local squad.py, raise error: TypeError: 'NoneType' object is not callable ","comments":"Hi !\r\nThanks for reporting.\r\nIt looks like no object inherits from `datasets.GeneratorBasedBuilder` (or more generally from `datasets.DatasetBuilder`) in your script.\r\n\r\nCould you check that there exist at least one dataset builder class ?","body":"\r\nversion: 1.0.2\r\n\r\n```\r\ntrain_dataset = datasets.load_dataset('squad') \r\n```\r\n\r\nThe above code can works. However, when I download the squad.py from your server, and saved as `my_squad.py` to local. I run followings raise errors.\r\n```\r\ntrain_dataset = datasets.load_dataset('.\/my_squad.py') \r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-25a84b4d1581> in <module>\r\n----> 1 train_dataset = nlp.load_dataset('.\/my_squad.py')\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 602 hash=hash,\r\n 603 features=features,\r\n--> 604 **config_kwargs,\r\n 605 )\r\n 606 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n","comment_length":34,"text":"load_dataset from local squad.py, raise error: TypeError: 'NoneType' object is not callable \n\r\nversion: 1.0.2\r\n\r\n```\r\ntrain_dataset = datasets.load_dataset('squad') \r\n```\r\n\r\nThe above code can works. However, when I download the squad.py from your server, and saved as `my_squad.py` to local. I run followings raise errors.\r\n```\r\ntrain_dataset = datasets.load_dataset('.\/my_squad.py') \r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-25a84b4d1581> in <module>\r\n----> 1 train_dataset = nlp.load_dataset('.\/my_squad.py')\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 602 hash=hash,\r\n 603 features=features,\r\n--> 604 **config_kwargs,\r\n 605 )\r\n 606 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\nHi !\r\nThanks for reporting.\r\nIt looks like no object inherits from `datasets.GeneratorBasedBuilder` (or more generally from `datasets.DatasetBuilder`) in your script.\r\n\r\nCould you check that there exist at least one dataset builder class ?","embeddings":[-0.2243587077,0.3296630979,0.111471124,0.1580804884,0.3371319771,-0.0254657716,0.6201825738,0.3964118361,-0.1371878237,0.0546794124,-0.0827249959,0.4112722576,-0.124582462,-0.0242635924,0.3396508992,0.0455027483,-0.0610549562,0.2632742226,-0.1969885081,-0.2398484796,-0.2595587373,-0.0362527221,-0.0532293469,0.3426586688,-0.3587331474,0.0167298,-0.0222245716,0.4130075872,-0.2073467672,-0.4058050811,0.4260213673,-0.0520701259,0.4543029666,0.4609367549,-0.000108605,0.0964231789,0.2037233412,-0.0567755476,-0.4298074245,-0.4939341247,0.0174136348,-0.2600156665,0.348030448,-0.4434491098,-0.044265382,0.1277416945,0.1641693115,-0.2078907192,0.4198295176,0.5076459646,0.2114357054,0.4164859951,-0.0478948094,-0.1072460487,0.0606455281,0.2325323373,-0.1264866292,0.4050805569,0.0112330215,-0.2007900476,0.2467612177,-0.0516305193,-0.2405682504,-0.0204578638,0.2836372554,0.1646960825,-0.00388834,-0.3396708667,0.0379902907,0.0254600774,0.3043539226,-0.5623261929,-0.4327322841,0.0522474162,0.2386764288,-0.0903794169,0.0619195513,0.12697649,-0.1391509771,0.1650296748,-0.2995199859,-0.0174529459,-0.34699139,0.2185474634,0.1378310174,0.0882255137,-0.0734710768,0.2105123252,-0.0946086869,0.1174466833,0.1658370942,-0.0670502037,0.2600558996,0.2247153521,-0.1657711715,-0.0563905463,0.1484064311,-0.2879837751,0.0835403949,0.1376603991,0.1994940639,0.0354084782,0.1475709975,0.3584354818,0.3046308756,-0.0103832595,0.4460417926,0.3905521333,0.2092555761,0.0204773471,-0.3075550199,0.0816470981,-0.4033690989,-0.1090946198,0.2758958042,0.076598607,0.6148070693,-0.1758502126,-0.0884732679,-0.0214995984,-0.0159814488,-0.0150975827,0.1789544821,0.3566183746,-0.0046703769,0.1785069406,0.0226463974,0.3013427854,-0.2289461493,-0.3886495233,-0.1576166451,-0.0191612206,-0.1196386144,0.1119067594,0.3236369491,0.0200300962,0.0620407388,-0.0539848804,0.0096883215,0.0857435167,0.3394778669,-0.2605609894,-0.0883493051,0.0893337578,0.1972521245,0.0455686972,0.3239510953,-0.481225729,-0.1657757461,0.0935725421,-0.1421400309,-0.415040493,-0.2359469682,0.1811512709,-0.1541479975,-0.0848482996,-0.0737669617,-0.1494629234,0.222211197,-0.3436722457,-0.0893649682,-0.2118363082,-0.0979256555,-0.3135224283,0.1799425036,0.7318371534,-0.5566922426,-0.1748506129,0.0376518182,-0.1776361465,0.2152796388,0.0091927666,-0.396612227,0.2239100337,-0.0369548351,-0.1163446084,0.8239561319,-0.5646014214,-0.2108430266,0.3148209155,-0.2390918136,-0.1988459229,-0.0661059618,-0.0253550485,0.3400918543,0.092112653,0.2875827849,0.5355546474,-0.0294604786,0.0208227746,-0.1860280484,-0.2029984593,0.0848430544,0.2191347331,0.0231419913,0.3303081989,0.1459700316,0.0864074007,0.0052345647,-0.1767736226,0.1181473657,0.0941681117,0.3387049139,0.0741478726,-0.2490807027,-0.2523498535,-0.6033933163,0.0904372483,-0.3337163031,0.2756749392,-0.2055827975,-0.1493386626,-0.2154192328,0.0664722323,-0.3703392148,0.1264156103,0.1411835849,0.1770204306,0.0798652619,-0.042924922,-0.3529272079,0.0876473486,-0.2050779909,0.2082749009,-0.1766453534,0.2906638384,-0.1883773059,-0.2833380699,0.089140147,0.2023648024,0.2694806457,-0.1818110049,-0.1411184072,0.3598248661,-0.2107282728,-0.0693585575,0.1634406894,-0.2966667116,0.0126889218,-0.0751981437,0.2857248485,-0.0647990257,0.1065602377,-0.0678152293,-0.1150792912,0.5167067647,-0.0031732633,0.1721068323,-0.0526649728,0.0205951426,0.1081378162,0.0752766132,-0.2094512284,-0.1581074595,-0.2273749858,0.2697479427,0.4829780459,0.1188587025,-0.1470673233,0.0532565042,0.5740777254,-0.05465075,0.1372898668,0.0851620138,-0.1313451231,-0.0412323698,0.0708402544,0.1479733884,0.5725581646,0.0996330902,-0.0794533044,-0.0268103015,-0.1564649791,-0.0062203798,-0.033436507,-0.1205940321,0.2183308899,0.0642101169,0.0933713689,0.0356456973,-0.2885571122,-0.1144710928,0.1267339289,0.2433501333,-0.1978192329,0.2767357826,-0.2465389669,0.0061044395,-0.4481367171,-0.0640273765,-0.3408609629,-0.019282598,-0.201014474,0.1224423125,0.1562079191,0.1803924888,-0.1137546077,0.0052955281,0.0849541128,-0.5639901161,0.191577971,-0.07982371,-0.3821257055,-0.0223751832,0.4319239557,0.079372488,0.2485446036,-0.2006807923,-0.0224453732,-0.0097836722,-0.0044241687,-0.0129084401,0.1107249632,0.2616941631,0.0467463396,0.2120769173,-0.1695703864,-0.2411050797,0.5361255407,-0.3284081817,0.0225232262,0.2159548551,-0.1201800406,-0.0991033167,-0.2032280713,-0.6506759524,-0.5834841132,-0.3092330694,0.0982242003,0.3362255692,0.2940899134,0.2489456236,0.3768067658,0.4234935045,0.0208355319,0.107160382,-0.0720120817,-0.3683303893,0.3266993463,-0.3098600805,-0.3858877718,0.0721755922,-0.0773554444,0.1593592018,-0.1642022133,-0.143696934,-0.1241336241,0.1434112042,0.2297299951,-0.0108086048,0.1534465849,0.3795170486,0.0624459572,0.0130371302,-0.0232359711,-0.2466184199,0.130396232,0.2644398212,0.1292032599,0.1152427644,0.4664043784,-0.3106529713,0.6398789883,-0.0654288009,-0.1792743653,0.3221142292,-0.1978886276,0.3280735314,-0.265311271,-0.5119051933,0.0864262208,-0.0068482873,-0.1117152944,0.1300498843,-0.0943719596,-0.2312443405,-0.3441791534,0.0010108976,-0.3042766452,-0.287599802,0.1291430146,0.1015196368,0.0178156514,0.1656626016,0.1310942322,-0.4643912613,0.0817354843,-0.2946573794,0.1311013997,0.0188825782,0.1163309962,-0.6968171,-0.0040908651,-0.2606073916,0.2750989795,-0.0018119972,0.3443766832,-0.0007360847,-0.1504275203,0.0790584981,0.0071524805,0.3598929644,0.04435388,-0.0332388766,0.3494811356,0.0083162514,-0.366186589,-0.0690007657,0.0578530505,0.4126611054,0.1779303551,0.305872947,-0.1788778454,-0.1519776732,0.2920213938,0.1171981841,-0.1820330024,-0.1239636838,-0.4005764723,-0.1464546025,-0.4580982924,-0.1369232088,-0.3179372847,0.3135766089,-0.1764062047,-0.0169396698,0.010126112,0.2256152928,0.0984548628,0.2285173535,0.1719058156,0.0103987576,0.0234331805,0.1817039251,0.1397234648,0.0885703191,0.4274907112,-0.0154263955,-0.4276290238,0.0968842506,0.0247510932,0.2670660913,0.2610846162,0.050939545,-0.3438411951,0.1145096123,-0.0504516214,0.1614839882,0.0166757628,0.2412511259,-0.2432658821,-0.1654126793,-0.6092640758,0.2995340228,-0.0174322613,0.1257799119,0.0740782917,-0.1040706486,-0.0699960664,0.2641106546,-0.0333867185,0.9230502248,-0.2437045127,0.2671840191,0.4023739398,0.0270804707,0.5112031698,-0.1930642724,0.0954387784,-0.5287599564,0.0611551255,0.0787504837,-0.1406329125,0.2448664755,0.4530040324,-0.2674677968,0.2707574964,-0.1633042246,0.2369352281,0.0229802802,0.350240618,0.0826204345,-0.2968753278,-0.5334136486,0.1607750207,-0.152611196,0.2689841688,-0.1225842237,-0.3375274241,-0.1786660254,-0.1476680189,0.0995045379,0.2700579166,-0.1974030733,0.2189332247,0.0926455557,-0.1975752562,-0.0251707733,0.3961940706,0.1224067658,-0.0203800686,-0.1569622606,-0.1656003594,-0.0339815244,0.0331332348,-0.1739097387,0.0883125365,0.3278300166,-0.0933510885,-0.3661977351,0.1251247376,0.0381672345,-0.2050872445,0.2090356946,0.074556835,0.3146267831,-0.3893392384,-0.3871375024,-0.0512853898,0.0112018408,-0.1596219987,0.1664564312,-0.0108032301,-0.1299458295,0.0197042264,0.1222798601,-0.1492580324,0.0454932749,0.6396827698,0.0037159666,0.1097421572,0.5209497213,0.0998622924,0.074539572,-0.3004195392,0.3296760023,0.0122953337,-0.3325354457,-0.0051599317,0.2763139904,-0.0711304322,0.0493511669,0.0253269132,0.0048106676,0.1556652337,-0.0613143593,-0.1963359416,-0.392173171,0.0833666921,0.2229947001,0.0238742586,0.3970584869,0.3349558413,0.2966407239,-0.0667460635,-0.3174312413,-0.0635913759,-0.2534838319,0.0853929073,0.1011064276,-0.2618179619,0.0404368415,-0.003761916,0.1213666648,-0.0390645303,-0.0343276411,-0.2343989462,-0.0242191982,0.1676315367,0.1767394394,-0.0129675157,-0.1407909393,-0.1967523247,-0.2141488194,-0.065574266,-0.0679332018,0.1429409385,-0.0243410598,0.1971213371,-0.0413305983,0.2479348928,-0.0501688495,-0.0038267607,-0.0880311877,0.1236707717,-0.0530491844,0.2134479582,0.0203920733,-0.0633007586,-0.4011785984,-0.0378918312,0.0690681711,0.3248786926,0.2766212225,-0.496781826,-0.0730513036,0.1525527686,0.0463832915,0.4550617337,-0.190095216,0.1661151052,0.0314916149,0.1723899394,-0.235892877,-0.0859521925,0.4179473519,-0.3348882496,-0.0733614042,0.1853558421,-0.2392935455,-0.0198515169,-0.0200151075,0.0348276496,0.198942259,0.0168717988,0.2267225534,0.2360977978,-0.1316925138,0.0538280681,0.2121463418,0.2222722471,0.1212321222,0.3404949903,-0.3878471851,-0.0408475175,-0.3319381475,-0.0604662672,0.2438606173,-0.3466596007,0.0407510102,0.0672533959,0.0857072622,-0.0763254017,-0.2831714749,0.7181065083,-0.4838989079,-0.2441076487,-0.1867139935,0.133215487,-0.1122146696,0.0941124186,-0.4224556684,-0.0312602445,-0.0904556513,-0.090287298,-0.1380403489,-0.2224529237,0.1279004812,-0.0256494787,-0.1128323972,-0.1216526106,0.3253854513,0.1182349324,-0.2297561318,-0.0603047535,0.1626446396,0.0888856128,0.2507382631,-0.0889428258,0.095975332,0.3186889291,0.5235214233,0.0713191479,0.1520190984,0.065719232,-0.2197308987,-0.0342746004,-0.1044728532,0.0762779787,0.066329971,0.1891943216,0.1664166749,-0.1354343593,-0.127647832,0.0595247857,-0.0899126157,-0.0937430933,0.5399539471,-0.0734166354,-0.2679059207,-0.3223481476,0.120383434,-0.5080347657,0.0191759933,0.4161300361,0.3160795271,0.1190875396,-0.4740586579,0.0708643645,-0.0395558104,0.333486259,0.0198187903,0.0365329757,-0.2406849563,0.0904707313,-0.6870467663,0.2286404669,0.0082808826,0.1038643047,0.0660710707,0.3264712393,-0.1442703158,0.154947266,-0.207023859,-0.0874887705,-0.0201650746,0.1666975915,-0.2390473485,-0.2215585411,-0.3781501353,0.156494081,0.0617071949,-0.5121263266,0.0867967755,-0.2959515452,0.0290353335,-0.113111876,-0.1626592875,0.0351395383,-0.0939826816,0.4827966392,0.056031093,0.4043293595,0.1090317667,-0.1398846954,-0.07183595,-0.1703635752,-0.1583561897,0.3901117742,0.0352962613,0.3115004599,-0.0136261471,0.0669585764,-0.4052658379,0.2364003211,0.023013575,-0.1279178709,-0.0627652332,-0.1101918072,-0.1321988106,0.021562608,-0.0501381494,0.0795209929,-0.0191897284,0.2761864364,-0.3440571129,-0.0467146486,0.5400159359,-0.3772190809,-0.2842176855,-0.0952101648,0.3152951896,0.0061883796,-0.2457031459,-0.4609938264,0.1890916526,0.3097381592,-0.1667713821,-0.132629469,0.2897094488,-0.1685261726,0.1667710394,-0.1672700197,0.1841707528,-0.2169226706,0.006686274,0.0432691984,-0.1639606655]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/664","title":"load_dataset from local squad.py, raise error: TypeError: 'NoneType' object is not callable ","comments":"Hi @xixiaoyao did you manage to fix your issue ?","body":"\r\nversion: 1.0.2\r\n\r\n```\r\ntrain_dataset = datasets.load_dataset('squad') \r\n```\r\n\r\nThe above code can works. However, when I download the squad.py from your server, and saved as `my_squad.py` to local. I run followings raise errors.\r\n```\r\ntrain_dataset = datasets.load_dataset('.\/my_squad.py') \r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-25a84b4d1581> in <module>\r\n----> 1 train_dataset = nlp.load_dataset('.\/my_squad.py')\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 602 hash=hash,\r\n 603 features=features,\r\n--> 604 **config_kwargs,\r\n 605 )\r\n 606 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n","comment_length":10,"text":"load_dataset from local squad.py, raise error: TypeError: 'NoneType' object is not callable \n\r\nversion: 1.0.2\r\n\r\n```\r\ntrain_dataset = datasets.load_dataset('squad') \r\n```\r\n\r\nThe above code can works. However, when I download the squad.py from your server, and saved as `my_squad.py` to local. I run followings raise errors.\r\n```\r\ntrain_dataset = datasets.load_dataset('.\/my_squad.py') \r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-25a84b4d1581> in <module>\r\n----> 1 train_dataset = nlp.load_dataset('.\/my_squad.py')\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 602 hash=hash,\r\n 603 features=features,\r\n--> 604 **config_kwargs,\r\n 605 )\r\n 606 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\nHi @xixiaoyao did you manage to fix your issue ?","embeddings":[-0.0806185678,0.1176947579,0.0936616138,0.1588600874,0.3462714851,-0.0670320392,0.6041768789,0.3358294666,-0.1248953193,0.0278394595,-0.2387042642,0.4593687654,-0.0431698225,0.0077984366,0.2703522444,0.0201432016,-0.0901881531,0.2277538925,-0.1694865823,-0.2526986301,-0.2112970054,-0.086733982,-0.0967853889,0.3335448802,-0.3112369776,-0.1149061173,0.0190455131,0.3911852837,-0.122852914,-0.4314467311,0.4329204559,-0.1783745289,0.4354561269,0.4475769997,-0.000109348,0.1347309202,0.2311551273,-0.0777502432,-0.3977268934,-0.4779224694,-0.0558113866,-0.3129806817,0.4285405874,-0.4781847,-0.045282349,0.1453462541,0.0662642345,-0.2110323906,0.4976109564,0.5424274206,0.2103238851,0.491556406,-0.0830716789,-0.003671326,-0.0572602563,0.2062102854,-0.104882054,0.4251977503,-0.0675395802,-0.1556490064,0.286590457,-0.0383777395,-0.2218462378,-0.0107526677,0.1099499688,0.167563796,-0.0481997095,-0.3082281947,-0.0631497279,0.0014089027,0.2842186689,-0.4410152733,-0.4498500526,0.0503110625,0.2245669514,-0.1160942689,0.0855031684,0.1648088843,-0.0964554176,0.16060853,-0.2778938413,0.0494249314,-0.4305875599,0.269882679,0.1254404336,0.1310254186,-0.0436843745,0.2251705378,-0.1010334194,-0.0010944031,0.0856136903,-0.0640680641,0.2610620856,0.2828260362,-0.264559567,-0.1258904487,0.0208198968,-0.3314683735,0.0598817058,0.0440425538,0.2072066069,0.0643177256,0.1499866247,0.3807236254,0.3383734524,0.0530914776,0.4326679409,0.4140981436,0.254673332,0.0017180602,-0.2825131714,0.1016207412,-0.3732424676,-0.0204711016,0.1788279563,0.1019770578,0.5856095552,-0.2395004183,-0.1211257875,-0.0704000518,-0.052647043,0.0586843304,0.1986872852,0.2880812883,-0.0668461174,0.239775598,-0.0392161608,0.2709589601,-0.2719444931,-0.2944000959,-0.1162896082,-0.0202553496,-0.0757444799,0.0067593073,0.3047409654,0.0561952293,0.1117953211,-0.1125728562,0.0573193915,0.0361447223,0.2413858324,-0.3809649646,-0.1654928625,0.09285862,0.1998030096,0.0231265929,0.2800510228,-0.432163775,-0.1370766759,0.1014862955,-0.2477547228,-0.3766765594,-0.2730045915,0.1696526557,-0.1402013302,-0.1059202552,-0.025572395,-0.1946488768,0.2403365821,-0.2442179471,-0.0786004066,-0.1440846771,-0.1062929183,-0.2792617679,0.1275245398,0.6207863092,-0.4600565434,-0.1982714832,0.1088604182,-0.1848274469,0.2485819906,0.1012562364,-0.418309778,0.0973555744,0.0143492222,-0.0216259919,0.8349382877,-0.6732326746,-0.1963537782,0.3341652155,-0.3005271554,-0.1989126503,-0.0649010688,-0.0301371291,0.3758134544,0.1744355112,0.3727161586,0.6203901172,0.0326524377,0.0605263114,-0.268273443,-0.2184137106,0.0924154967,0.2240645289,-0.1035332307,0.305265069,0.258212328,0.0828729272,0.0356149785,-0.106291838,0.1102630123,0.0912532583,0.3867819011,0.1118131876,-0.2563160956,-0.1717478633,-0.5835316777,0.0703019649,-0.3740475774,0.3255114853,-0.2283221483,-0.165526554,-0.2743495703,-0.0077176387,-0.3477526307,0.1206470579,0.1305411458,0.1245081425,0.0745709538,0.1229885891,-0.2272370011,0.1845588684,-0.2240456194,0.212050721,-0.0563777871,0.2727299333,-0.1456245929,-0.2545230687,0.0751030743,0.192131564,0.2849406898,-0.1464078724,-0.2194321901,0.3381348252,-0.206066072,-0.0009007004,0.1736175567,-0.3849555552,0.1198664531,-0.088232778,0.211914137,-0.0204797704,0.1932997853,-0.0539380983,-0.1972520053,0.5450546741,0.0485697538,0.1374409944,-0.0682695433,-0.0187731553,0.1565325111,0.006862754,-0.2012736499,-0.1413447112,-0.2284275442,0.291485846,0.4233047962,0.0889551416,-0.1922463477,-0.1028838456,0.5874528885,-0.0022090157,0.1739878356,0.1498894542,-0.0687555075,-0.0268207006,0.1462869346,0.0710906535,0.5493784547,0.0975247398,-0.1152088791,-0.0204675868,-0.1288919449,0.0424742848,-0.0505417772,-0.1020898446,0.2875005007,0.0869314745,0.2397188097,-0.0181771088,-0.2553233802,-0.0320869312,0.1443401277,0.1585050374,-0.2210049778,0.2422586977,-0.281442076,-0.0893638879,-0.4535036683,-0.0300342608,-0.2935995758,0.0342804566,-0.2545829713,0.1180513576,0.1826432943,0.2490683645,-0.1946993023,-0.0162605569,0.1578800976,-0.5903301239,0.2964701056,-0.0197495986,-0.3813823164,-0.0230048206,0.3116194904,0.0575108416,0.2682855129,-0.1460652798,-0.0132466983,0.1193935871,0.0021420561,-0.0170681719,0.1818531156,0.261857748,-0.030691145,0.2657105923,-0.2190044224,-0.1725303382,0.5689585805,-0.2519004643,-0.0174036454,0.1087444201,-0.0937737748,0.0173760075,-0.2318332046,-0.6992468834,-0.6980066895,-0.3047943413,0.1478176862,0.2612267435,0.3123020232,0.2625707984,0.3045865893,0.460046351,-0.0200454686,0.0534782,-0.0453846306,-0.3136417568,0.3503160179,-0.2246292681,-0.3993594646,0.0677465275,-0.0174565893,0.0770074129,-0.134784475,-0.1639134735,-0.2062868327,0.1853364408,0.2672241926,-0.0822441429,0.0767815486,0.387201488,-0.0445964076,-0.0097692497,-0.0073603126,-0.2847049236,0.170734778,0.4015545249,0.1056399941,0.1255369931,0.4983547926,-0.1862341911,0.578397572,-0.0431978405,-0.1666303724,0.4039031267,-0.2085183412,0.1925643831,-0.2425951958,-0.4942960143,0.0960042626,-0.0078721056,-0.0431358926,0.1025289148,0.0013587869,-0.3603470027,-0.380141288,-0.1051583886,-0.3175726235,-0.2001556307,0.091403082,0.1860390604,0.0610087104,0.1480666995,0.039507512,-0.3310848176,-0.0232052412,-0.3033179045,0.1503264606,-0.0712689608,0.118620716,-0.8075171709,-0.0347948149,-0.1771851629,0.2771111727,0.0736542642,0.5131003261,-0.0552493297,-0.1190557629,0.0800627619,0.0442752913,0.3668909669,0.1586073637,-0.0348953716,0.3019957542,0.0263124388,-0.3892480135,-0.0006490431,0.0541016534,0.3697139621,0.1986767948,0.3945772946,-0.1663077027,-0.1035165936,0.1637847275,0.1087458432,-0.2328306139,-0.0936463177,-0.4278397858,-0.1181194335,-0.4635082483,-0.0377570167,-0.3359110355,0.2779995799,-0.1382509619,-0.0565969124,0.0515508205,0.2382313609,0.0339482687,0.1919505149,0.2053743154,0.0742502138,0.0032635846,0.1998532712,0.30115062,0.0235205665,0.3999909461,-0.0477525704,-0.3821667731,0.0328110717,0.0016871331,0.2976400852,0.2137624025,0.0477171503,-0.2274903208,0.0949631184,-0.0699609742,0.1666991264,0.0556224771,0.2104816586,-0.1812689453,-0.18331258,-0.5043637156,0.2545885444,-0.0445340388,0.097319141,0.0550148562,-0.1399977505,-0.1191406697,0.146799475,-0.0497168712,0.8953505754,-0.1864856929,0.2850016654,0.3895182014,-0.0481099933,0.5141850114,-0.0951653048,0.0977699757,-0.4156101942,0.2281057239,0.0586963743,-0.1465146095,0.2884069979,0.350684166,-0.2613078058,0.2742542624,-0.2266138047,0.1515625268,-0.0252920631,0.3689464629,0.2185619175,-0.3281519115,-0.596914053,0.1413387656,-0.2063062936,0.3003757596,-0.0901718885,-0.323051542,-0.2090037763,-0.1738661826,0.1828527302,0.2167179435,-0.1582095474,0.3018050492,-0.0049733236,-0.2463545501,-0.0841776803,0.3690027893,0.0208141748,-0.0521266684,-0.1479037255,-0.1882855147,0.0082124863,0.0487767234,-0.1506446153,0.0867507905,0.2128316462,-0.132731989,-0.4057173729,0.0624443553,0.0018144074,-0.212469995,0.1854114681,0.1405507624,0.3311612904,-0.3898367286,-0.4179916382,-0.1274420172,0.0712320358,-0.1393997222,0.1663014144,-0.0669062138,-0.0822308064,0.0076170727,0.2154862732,-0.1628956199,0.035863094,0.7231646776,0.0771195665,0.1255917549,0.5478281975,0.0430746265,0.0229579974,-0.284576118,0.4109072387,-0.1736898422,-0.2617224455,-0.0896436423,0.2568278611,-0.0743386894,-0.1257574856,0.0197097249,-0.0066912165,0.1398971528,0.0171969123,-0.2814970613,-0.4606661797,0.0895444453,0.1518680006,-0.0463186055,0.3797681928,0.420409143,0.3136231899,-0.1507326066,-0.2973442972,-0.0376431197,-0.3545924425,0.1447605193,0.1872859746,-0.2655496597,0.1369186342,0.0131834084,0.1493111253,0.0142670106,0.0265288744,-0.2314185351,-0.0650616959,0.1685629785,0.1047612503,-0.0766703784,-0.1733069569,-0.2671778202,-0.1777391434,0.0268649515,-0.1359387934,0.2319553196,-0.0360289626,0.2090586275,-0.103821747,0.275585413,0.0073030363,0.0758191347,-0.090870209,0.0256056935,-0.0488323718,0.1699072868,-0.0169348326,-0.0354032964,-0.4379576147,-0.1629447341,0.1216039732,0.3514449596,0.126103431,-0.4660779834,-0.1009855717,0.1057440639,0.0112706609,0.3717314303,-0.2233286202,0.1322943866,0.1064859852,0.1892614514,-0.2336390764,-0.1364200264,0.4439766705,-0.3704772592,-0.0799399242,0.140760839,-0.1987404376,0.066711776,0.0115559846,0.014415754,0.1749157161,0.0325722918,0.1833895743,0.1717113107,-0.1723045558,-0.0000910484,0.1286611557,0.1533839256,0.0848495662,0.2761272788,-0.3626051247,-0.0845367312,-0.2304720134,0.0316271596,0.1489379406,-0.3315999508,-0.0294113848,0.0186128691,0.0063710571,-0.021259347,-0.3019683659,0.7093449831,-0.4126901031,-0.2751137614,-0.127336815,0.109604381,-0.1604924202,0.0530812927,-0.3379632235,-0.0487438291,-0.0281584747,-0.0928766802,-0.1112693325,-0.1651954353,0.1398440301,0.0154252714,-0.1135646254,-0.1175372973,0.2047892809,0.2637485564,-0.1909432858,-0.0767310709,0.208198294,0.088266544,0.2496931404,-0.1305596977,0.0532316677,0.222833544,0.521212399,0.063247703,0.1084575579,0.1316713393,-0.1691597104,0.0381295048,-0.0792236775,0.1533258557,0.090287827,0.2072567791,0.1491101086,-0.1579980254,-0.0027270273,-0.0251505803,-0.1294742823,-0.142940104,0.5220686197,-0.065657191,-0.2317989469,-0.2707214355,0.1003431305,-0.5107808113,0.1482011825,0.4236058295,0.3288525641,0.1789383441,-0.4088073075,0.0661946312,-0.1410931647,0.377091974,-0.0934086666,0.0702102929,-0.1731377095,0.1539433151,-0.629229784,0.2495919913,-0.0292295273,0.0585418865,0.0405526459,0.2506961823,-0.1657961756,0.1661194712,-0.2065325677,-0.1086625606,0.0597292855,0.01259364,-0.2691973746,-0.187914297,-0.412630856,0.2236327529,0.1080814078,-0.5367693901,0.1327732056,-0.2291116863,0.0236638226,-0.1307989657,-0.169260636,-0.0314497836,0.0166047812,0.4433967769,0.0680539235,0.4574157894,0.1329732537,-0.1242775917,0.0595808104,-0.2444785535,-0.1108285934,0.3523015678,0.0732226074,0.287474066,-0.0444812924,0.0941297188,-0.3750967681,0.2254879773,-0.0027123413,-0.2421277463,0.0724023283,-0.1566695422,0.0006536315,-0.0763999149,-0.0931938663,0.1676049829,-0.0118708434,0.1729257554,-0.2581351697,-0.0266547054,0.5730899572,-0.4074673951,-0.2664195299,0.0212471057,0.3113992214,0.0176101159,-0.1853493303,-0.4607589543,0.1733926684,0.422688067,-0.1984933913,-0.1587991714,0.2864319086,-0.15618065,0.1307228357,-0.1358645111,0.1618546396,-0.1815832555,0.0046745623,0.0531527922,-0.1681726575]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/664","title":"load_dataset from local squad.py, raise error: TypeError: 'NoneType' object is not callable ","comments":"No activity, closing","body":"\r\nversion: 1.0.2\r\n\r\n```\r\ntrain_dataset = datasets.load_dataset('squad') \r\n```\r\n\r\nThe above code can works. However, when I download the squad.py from your server, and saved as `my_squad.py` to local. I run followings raise errors.\r\n```\r\ntrain_dataset = datasets.load_dataset('.\/my_squad.py') \r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-25a84b4d1581> in <module>\r\n----> 1 train_dataset = nlp.load_dataset('.\/my_squad.py')\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 602 hash=hash,\r\n 603 features=features,\r\n--> 604 **config_kwargs,\r\n 605 )\r\n 606 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n","comment_length":3,"text":"load_dataset from local squad.py, raise error: TypeError: 'NoneType' object is not callable \n\r\nversion: 1.0.2\r\n\r\n```\r\ntrain_dataset = datasets.load_dataset('squad') \r\n```\r\n\r\nThe above code can works. However, when I download the squad.py from your server, and saved as `my_squad.py` to local. I run followings raise errors.\r\n```\r\ntrain_dataset = datasets.load_dataset('.\/my_squad.py') \r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-25a84b4d1581> in <module>\r\n----> 1 train_dataset = nlp.load_dataset('.\/my_squad.py')\r\n\r\n\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 602 hash=hash,\r\n 603 features=features,\r\n--> 604 **config_kwargs,\r\n 605 )\r\n 606 \r\n\r\nTypeError: 'NoneType' object is not callable\r\n\nNo activity, closing","embeddings":[-0.066853784,0.0995030776,0.1286152154,0.1915656477,0.3325941563,-0.0487718955,0.6078260541,0.3216490746,-0.119287774,0.0478187092,-0.280388087,0.4274514318,-0.0573104247,0.0090635065,0.2343303412,0.0552773289,-0.0925886631,0.2136370987,-0.1214011982,-0.2714107335,-0.2697933614,-0.1037588939,-0.1063532978,0.3538266122,-0.3264617026,-0.1408044398,0.0196173787,0.42522645,-0.1174345538,-0.4143831432,0.424957335,-0.1676717997,0.4216913283,0.5132268071,-0.0001139371,0.1463842243,0.2594830692,-0.0537129454,-0.4158355892,-0.4642551839,-0.0249716118,-0.3368889689,0.4634298682,-0.4728784859,0.0357730612,0.1150925457,0.0788596347,-0.2133156806,0.4966112077,0.5514549613,0.1628592163,0.5375926495,-0.1094281301,0.0036652882,-0.0061952183,0.2733566761,-0.1244300306,0.4419011176,-0.0918779001,-0.1209315658,0.3270950913,-0.0799607411,-0.3061288893,-0.0329557694,0.1039705053,0.1689156741,-0.0495414659,-0.2777223289,-0.0081042415,0.0298658907,0.3103852272,-0.4151328802,-0.470343262,0.0371361822,0.191393435,-0.1044640169,0.0886788368,0.1706243604,-0.0743790045,0.1269944459,-0.2704857588,0.0765067562,-0.443479836,0.2536218762,0.1204907075,0.0998069495,-0.0491500609,0.2796372175,-0.1100179106,0.0165897403,0.1362159997,-0.111905694,0.2805261016,0.2867783606,-0.2474631965,-0.1053392291,0.0440336764,-0.3510856628,0.0503246561,0.0629415289,0.2130720615,0.0559444651,0.1396157444,0.4413034916,0.3271239996,0.0319412798,0.4496972561,0.3560720384,0.2692458034,-0.0259782486,-0.272654444,0.1220520809,-0.395328939,-0.0129764481,0.1951884627,0.0580949262,0.6444179416,-0.2534476519,-0.150013268,-0.0669307038,-0.0439964533,0.0555428006,0.1993235052,0.3033623397,-0.0779292658,0.2778894901,-0.0488807596,0.2646924853,-0.2939487696,-0.2867552638,-0.0930978581,-0.0465864614,-0.0574656762,0.0022237529,0.2899436653,0.0666521639,0.1015728191,-0.1533010453,0.0258625746,0.0284583811,0.2810722888,-0.4333745539,-0.1918549538,0.1407143921,0.1733035892,0.0355648249,0.2735150456,-0.4303772151,-0.1802366823,0.1264350563,-0.2432999313,-0.3901211023,-0.2810791433,0.1189031526,-0.1761102378,-0.119481802,-0.006648845,-0.2275927514,0.2389212996,-0.276360333,-0.0436668247,-0.1234723851,-0.0995882526,-0.3110636473,0.1127736941,0.718396008,-0.4603476524,-0.2119906247,0.0904013366,-0.1960785985,0.2584474385,0.0928622782,-0.3903159201,0.1151931435,0.0520894825,-0.0007080875,0.811224997,-0.6676915288,-0.1694393009,0.3581959903,-0.315184176,-0.2110363692,-0.0054747416,-0.0152611053,0.3347876966,0.1592498571,0.3711083829,0.6172578335,0.0295632537,0.0031173006,-0.2797783017,-0.2125980407,0.1260974705,0.2045026571,-0.0937167257,0.3147791326,0.3006416559,0.1213500649,-0.0389764309,-0.0730056837,0.1091554314,0.0271897037,0.4145427644,0.1237015799,-0.248243928,-0.1293156445,-0.6454059482,0.0846832171,-0.375215441,0.2967248261,-0.258151114,-0.1548545659,-0.2887417078,0.0254825447,-0.3430185616,0.1271225512,0.06427899,0.1650989503,0.053825967,0.1504105926,-0.2629725039,0.1401890814,-0.2705843449,0.226466924,-0.062795423,0.2615560591,-0.1090272442,-0.2577475905,-0.0147603247,0.1795382947,0.284961164,-0.146119073,-0.196494028,0.360973388,-0.1773881763,0.0523111746,0.1840339452,-0.3474236727,0.1192064509,-0.0372097865,0.2119618952,-0.0304623954,0.2122053951,-0.0540033542,-0.2087412179,0.5414131284,0.0354429409,0.2026906163,-0.0661788881,-0.0425245538,0.1195259467,0.0654075742,-0.2190113068,-0.1190921441,-0.2415505946,0.3161606789,0.462803185,0.1073014364,-0.1935411394,-0.1122130379,0.6106926203,-0.0589346103,0.2120309919,0.1559081078,-0.0751652867,-0.0572661795,0.2034239918,0.0585477538,0.5946799517,0.0656235516,-0.0674266741,-0.0158669371,-0.1330406219,0.0234597046,-0.0714743435,-0.0842890888,0.2936616838,0.1127628982,0.2223315835,-0.0037637472,-0.2657375634,-0.0621631332,0.1738451123,0.1842938662,-0.2377867997,0.2660960257,-0.2328731567,-0.0324031748,-0.462254405,0.0202576891,-0.2873675823,0.0177828837,-0.2952341139,0.1183724478,0.1650591046,0.1757132411,-0.2049600184,0.0190063696,0.1435684711,-0.6276679039,0.3255217671,0.0257114172,-0.3941057622,-0.0754065886,0.3209695816,0.04952804,0.2327014953,-0.1909303069,0.0105135022,0.1496471316,0.0047554253,-0.0046879095,0.2071150541,0.2160441577,-0.0063092131,0.2988104224,-0.2155127227,-0.1612794399,0.5661477447,-0.2876827121,-0.0342747904,0.1200489104,-0.0858461782,0.0726819411,-0.2706859112,-0.6789560318,-0.6777729392,-0.3114809096,0.2211694568,0.2406886369,0.3299731314,0.2866431773,0.3170989454,0.522442162,-0.0389064327,0.0809607059,-0.0219324846,-0.3142594397,0.351614207,-0.1871354431,-0.3561450839,0.1048744619,0.0210171174,0.0734873861,-0.0927556828,-0.1783774346,-0.2092662156,0.2102803737,0.2793536186,-0.0847087577,0.0731317028,0.409053266,-0.0145620946,0.0160883311,0.0125554446,-0.259999305,0.1366355568,0.4293095767,0.0518830195,0.1496693939,0.5200810432,-0.1869354546,0.5494720936,-0.0787070617,-0.1199544668,0.4224067926,-0.1818191856,0.2367590815,-0.2789506614,-0.4486786723,0.0931968987,0.0147099672,-0.1023569554,0.1252041608,0.0216381345,-0.3858487308,-0.3874108493,-0.1790390015,-0.3293198943,-0.1790106297,0.1009235382,0.2121476531,0.0574390739,0.1103114188,0.0696144328,-0.3129349649,0.0142771322,-0.2809855342,0.1780682504,-0.0355025269,0.1166437492,-0.8335525393,-0.0654066652,-0.1913765669,0.3020884693,0.036245048,0.5033860803,-0.082884185,-0.1583122313,0.0831573755,0.0445342623,0.3660436273,0.1793668866,-0.040174149,0.2974200547,0.0112736654,-0.4368326366,0.0043818457,0.0828426406,0.3547903299,0.1516488194,0.4076970518,-0.1823346913,-0.0708573312,0.1351133138,0.1384563595,-0.1945375055,-0.0562654287,-0.4482095242,-0.1823129207,-0.4763062,-0.0604771301,-0.3666487336,0.2709864676,-0.1451276541,-0.0593801253,0.011825107,0.2626023293,0.0522994138,0.1798921376,0.2120964527,0.0974095464,-0.0503193922,0.2229062319,0.2937876284,0.0030339332,0.4069620371,-0.0569769181,-0.3857686818,0.1001188755,0.0212153625,0.3089294136,0.2464347333,0.0440070666,-0.1845219433,0.1058196202,-0.0736738518,0.1682342291,0.0820863247,0.2412233651,-0.2016720474,-0.2027075887,-0.5433805585,0.2995260358,0.0016014788,0.112208806,0.0265312921,-0.1434326619,-0.0843043476,0.1438033581,-0.0501989946,0.9387040734,-0.2195483297,0.3159674704,0.4154958725,-0.0507548526,0.53798455,-0.1305721998,0.0771708339,-0.3324588239,0.248733446,0.0370570831,-0.1339110136,0.3338241279,0.3557857573,-0.2361179441,0.2623251081,-0.3056833148,0.1505704671,-0.0170386713,0.3333994448,0.2057210654,-0.3265504539,-0.5863435268,0.0847637877,-0.1969878227,0.3405351043,-0.0848734602,-0.2771324217,-0.2060414851,-0.1427658051,0.2262138575,0.1640796512,-0.1144768,0.2695013285,-0.0438066162,-0.2391508669,-0.1499842554,0.3654344976,0.0284860842,-0.075340502,-0.1199490502,-0.2409520149,0.0154249556,0.0340274312,-0.1692675799,0.0659590811,0.2128218114,-0.1161368191,-0.4060456753,0.0840343535,-0.0499502979,-0.2243577838,0.2568919063,0.1704580039,0.3350925148,-0.3792398572,-0.4349287748,-0.1255061328,0.0131867863,-0.1286685318,0.1185982525,-0.0501477197,-0.0853723735,-0.0247935299,0.1693380326,-0.2139330953,0.0121024176,0.7102419138,0.1227919608,0.1544817686,0.5271733403,0.056463711,0.028743552,-0.2516370714,0.4438868761,-0.1369747072,-0.2708518207,-0.0593430586,0.2328728884,-0.1062365621,-0.1532294154,0.0407936461,-0.0171394106,0.1637875736,-0.0237108953,-0.2799549401,-0.468398869,0.1292628646,0.2325588465,-0.0667242184,0.4054641426,0.3889571428,0.3437001705,-0.2130056471,-0.2405761629,-0.0601892397,-0.3482694924,0.1309663653,0.172085166,-0.2474369109,0.1798458993,0.0028930821,0.1025955603,-0.0119859399,0.0367739573,-0.1751099378,-0.0881105959,0.2154385,0.1041826233,-0.0396176837,-0.1611076891,-0.3057140708,-0.1736509949,0.0280911289,-0.1229889244,0.2543410063,-0.0789391547,0.1646820903,-0.1305028498,0.3442007899,-0.0312621482,0.1098884717,-0.1165250763,0.029705584,-0.0476818867,0.1354719698,-0.0101428879,-0.0301107503,-0.441590488,-0.1531834602,0.1211426407,0.2888450623,0.1289031953,-0.4736795723,-0.1102373526,0.0219162554,-0.0133010764,0.4136664867,-0.1731439084,0.0854178295,0.1194047183,0.124378629,-0.1822035611,-0.1135712937,0.4112038314,-0.382388413,-0.0698821992,0.1833585054,-0.2258871645,0.0673448145,-0.045454219,0.0504871197,0.1605644375,0.0334319063,0.1910150647,0.2033470422,-0.1767155379,0.0457966626,0.1557273865,0.1545379758,0.0752951428,0.2142463475,-0.327826649,-0.0623557828,-0.2296922505,0.057631202,0.1410638541,-0.3757316172,-0.0096031493,0.0703074485,0.0231056903,-0.0003415405,-0.2999442518,0.7356564403,-0.458088547,-0.2731758356,-0.0973641649,0.0991460904,-0.1358681023,0.0294245854,-0.3400792181,-0.053207498,-0.0367585942,-0.1075888947,-0.1050188839,-0.1947561204,0.1762378663,-0.0051207263,-0.1128530279,-0.0669324026,0.2054972053,0.2628647089,-0.1830122769,-0.0687152892,0.2042144537,0.1168224663,0.2395143509,-0.1342570633,0.058032658,0.2101529092,0.542621851,0.0283027589,0.1276351511,0.1404471695,-0.1811082661,0.0321750194,-0.0843549669,0.1311424226,0.113245748,0.1960847825,0.1002524868,-0.1349531561,0.0331801698,-0.0628243983,-0.1592794359,-0.1622516811,0.5783290267,-0.0896846429,-0.2541061938,-0.2238706946,0.1200513691,-0.5580485463,0.1264717877,0.4161293507,0.3060230911,0.1581511497,-0.371545285,0.026137108,-0.1622131467,0.3633931577,-0.0751319528,0.106511429,-0.159140721,0.1936148554,-0.6345376968,0.2495403886,-0.0525348596,0.0574738085,0.058162339,0.2537301481,-0.1865325719,0.1635277867,-0.2506873012,-0.0869673342,0.0821591616,-0.0081197489,-0.245050475,-0.1919616461,-0.4443419576,0.2117500901,0.0746628642,-0.5580812693,0.1694173813,-0.2686158419,-0.0267571043,-0.1174270287,-0.1710704565,-0.017356012,0.009701659,0.4901241064,0.0611307696,0.4263611436,0.1578920931,-0.1695724577,0.069610782,-0.2175941765,-0.0986016989,0.3248420656,0.0359757841,0.3015696406,-0.0707656294,0.1725644767,-0.3129721284,0.2349045724,-0.0181418061,-0.2249428779,0.093167156,-0.1749879867,-0.0540570579,-0.1124810874,-0.0966031924,0.1235048622,0.0363298021,0.1012006626,-0.2760375738,0.0279370919,0.5663385987,-0.429733336,-0.2301602215,0.0087847812,0.3087074161,0.0074773529,-0.1724757403,-0.4952301085,0.1691524386,0.4206260741,-0.198483631,-0.191821456,0.2607026994,-0.1705672294,0.075719662,-0.1466424167,0.1670562178,-0.2056396157,0.0100432178,0.0688930526,-0.1993165016]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/657","title":"Squad Metric Description & Feature Mismatch","comments":"Thanks for reporting !\r\nThere indeed a mismatch between the features and the kwargs description\r\n\r\nI believe `answer_start` was added to match the squad dataset format for consistency, even though it is not used in the metric computation. I think I'd rather keep it this way, so that you can just give `references=squad[\"answers\"]` to `.compute()`.\r\nMaybe we can just fix the description then.","body":"The [description](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/squad\/squad.py#L39) doesn't mention `answer_start` in squad. However the `datasets.features` require [it](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/squad\/squad.py#L68). It's also not used in the evaluation.","comment_length":63,"text":"Squad Metric Description & Feature Mismatch\nThe [description](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/squad\/squad.py#L39) doesn't mention `answer_start` in squad. However the `datasets.features` require [it](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/squad\/squad.py#L68). It's also not used in the evaluation.\nThanks for reporting !\r\nThere indeed a mismatch between the features and the kwargs description\r\n\r\nI believe `answer_start` was added to match the squad dataset format for consistency, even though it is not used in the metric computation. I think I'd rather keep it this way, so that you can just give `references=squad[\"answers\"]` to `.compute()`.\r\nMaybe we can just fix the description then.","embeddings":[0.0384348109,-0.1900426894,-0.0537597947,-0.0725321695,0.4192170799,-0.0426003039,0.1088642105,0.0794378519,-0.2159082294,0.1062574461,-0.1819749475,0.4140356183,0.3362865746,-0.0604616292,0.1003780365,0.156886667,0.0473284982,0.083982572,-0.0124868695,-0.139941901,-0.1415313631,0.2984051406,-0.2252353281,0.2715794444,-0.2736495733,-0.0178927816,-0.0506354868,0.1622512788,-0.3746500313,-0.5675835013,0.2677565515,0.0306367315,-0.3266791105,0.2287681401,-0.0001090158,-0.1445931643,0.232779637,-0.1377821714,-0.1771668792,-0.0342071764,0.1800467074,-0.3626684248,0.1918864399,-0.4109576643,-0.0487220921,0.114833653,0.0063604061,-0.0542373285,0.3847958148,0.1295463592,0.2297418267,0.4827030599,0.0115810884,-0.2939109802,-0.06691356,-0.035664212,-0.1096247807,0.2075034827,0.2950330675,-0.182152763,-0.2618212402,0.2340934575,0.0406408608,0.0445102379,0.4658947289,0.0488792248,0.2508768439,-0.1295684427,0.0629555508,0.0557779558,0.2274373919,-0.2071215808,-0.3798899055,-0.1025848687,0.0531466864,-0.0919421762,0.0872523338,0.082811147,-0.1797547489,-0.0123234494,-0.3148842156,0.101958096,-0.1891083568,0.0358035155,-0.0675041601,0.3396831453,-0.1549963802,0.0361524001,0.2265035808,-0.0704884753,-0.133175984,0.0341340005,-0.0735610873,0.2844384611,-0.4198206961,-0.1560457647,0.2228899747,-0.0200363155,0.6190392375,-0.0421511903,0.1646747142,0.0248638578,0.3776242137,0.177577734,0.0793167651,0.499111861,0.5429851413,-0.1149800494,-0.0014036827,0.1581891924,-0.1400866956,0.0006069957,0.0942751169,-0.14679268,0.1362455189,0.0379651599,0.1092147082,-0.1411068738,-0.1879776269,0.1140722781,0.0556677803,-0.0816742182,0.0181318857,0.3469876647,-0.0054648593,-0.0605896153,-0.1710804999,0.2534984052,-0.0183081664,0.1575538665,-0.3258033991,0.1219163984,-0.2628063262,0.0448266417,-0.1285425723,-0.1219590232,0.4747461081,-0.0501228869,0.7316465974,-0.301017642,-0.0586592481,0.0636631474,-0.0390798114,-0.0274202675,-0.0383260511,0.066543363,0.3291544616,-0.3137712181,-0.0523793772,-0.0128110182,-0.2679873407,-0.2070085108,0.0971626937,0.216930747,-0.5295918584,0.0841749236,0.01919806,0.322234869,0.0022982452,-0.0381796472,-0.0678529441,0.0785545856,-0.4157375097,-0.2373695523,0.4349510074,0.1600910276,-0.0744292885,-0.0936143175,0.2863320708,-0.0914693549,-0.2294351906,0.0623862445,-0.0687920302,0.0911994725,-0.1189325377,0.2205138654,0.4820013046,-0.6493625641,-0.3048389852,0.0117189279,-0.1494525969,-0.1385005713,-0.2431694567,-0.0276350416,0.2740222812,0.1774703264,0.0658344254,0.0588019751,0.2066640109,-0.0599471405,-0.3792648017,-0.1584869474,-0.0096754711,-0.049499087,-0.0848823562,0.1371218711,0.0600272268,0.1734498143,0.0599574819,0.0350070633,-0.1196739227,0.0897828713,0.1923914105,-0.1864150614,0.1418641061,-0.2624903619,-0.325923115,0.2314723432,-0.4219742715,0.1186333448,0.1511097252,-0.3517677188,-0.6282202005,0.2035232484,-0.1491964459,-0.2065085024,0.2031186521,-0.3312682807,0.2791960239,0.0093606655,-0.1268214881,-0.1456931084,-0.2963276505,0.0748766959,0.0622849949,0.0990387723,-0.2384701222,0.0805219635,0.1940943003,0.2733443975,0.1319565922,0.1275603324,-0.125713408,0.4458444118,0.0097677987,0.1940330416,0.2939254344,0.0445197187,-0.0508879311,0.007126587,-0.0308100171,0.2197391391,0.0883625671,-0.2323543727,-0.3145397902,0.4246505797,-0.1224974021,0.1658186466,0.0343078412,0.1127615124,0.1513227373,-0.0237647537,-0.2545131743,-0.0314414017,0.0066199945,-0.0501973704,0.054263439,-0.0734683499,-0.0802657232,-0.0812714621,0.6966173649,0.0119204083,0.0642427728,0.1774988472,-0.0329265259,-0.2631054223,-0.1926166117,-0.4333093762,0.3488348722,0.1907119304,0.0782111883,0.0230336208,-0.043449495,-0.0238941554,0.0534150153,0.1939957142,-0.0103237517,0.3542571664,0.3989413083,-0.1800171137,-0.0387113355,0.1886887103,-0.2210490704,0.0798667595,-0.2387335896,-0.00154825,0.240759939,0.0406457782,0.0088517489,-0.1270717084,-0.4600836337,-0.4150730371,0.3634660542,0.0263142101,0.216481477,0.3412558436,0.2149056941,0.1921670884,0.1436503232,0.0168056451,0.1201283485,-0.4616580904,0.1071901992,0.0702019855,-0.162426278,0.1004727408,0.2189997733,-0.3725537062,-0.0299972836,-0.1779542416,-0.4762518406,0.1616972685,-0.0275497325,0.3890663385,0.1912371367,0.1360046715,-0.497957021,-0.107298784,0.4114948213,-0.3360354006,-0.3585233688,-0.1036209464,0.0542120263,-0.1929950267,-0.2370444238,-0.3711310923,0.0113394381,-0.4087290764,0.387157023,0.0803839564,0.0601451471,0.2898188233,-0.0015505245,0.3929220736,-0.1249706298,-0.1458929032,-0.4327925444,-0.3826179504,0.3666389585,-0.0849316046,-0.4884542227,0.0189987086,-0.1149359792,-0.0452607162,-0.3103398383,-0.0352042839,-1.136441946,0.170079723,0.127178207,-0.1813671589,0.1192192808,0.2368253022,-0.560041666,-0.0788351595,-0.3416191638,-0.4113590419,0.3146167994,0.283890456,0.1413577199,-0.2506248951,0.2161833942,0.1242948472,0.8888487816,0.2331058532,0.1609022915,-0.2426312268,-0.1912288368,0.1344807446,0.0041368171,-0.3907559216,0.1829271019,-0.0979354605,0.0240481906,0.3376947641,0.2204973996,-0.0359367244,-0.0884128585,0.2805607319,-0.1713957042,-0.251763314,-0.2128238529,0.2326666266,0.1044014469,0.1094346493,0.2401740104,-0.3419875205,-0.1232593656,-0.0082461387,0.2320182472,0.065129295,0.1779607683,-0.7929934859,-0.123344928,-0.1499417126,0.221396178,0.2740212977,-0.0438226014,-0.0035351485,-0.0784568638,0.1278564781,-0.0885965154,0.6530706882,0.1421282142,-0.0944209844,0.1690146774,0.1755911112,-0.4683640301,-0.1306082308,-0.0884223357,0.33653754,0.3961407244,0.3704069257,-0.4528734684,-0.1798210889,-0.082412675,-0.0248382762,-0.1235277131,0.0070578456,-0.6551508904,-0.2761806548,-0.1653268337,0.2495297045,0.1240828484,0.0326754898,0.1633638591,0.0280109551,0.3029095829,0.1078598127,0.1145451069,0.4966824353,0.3356104493,0.022576008,-0.1246724799,0.2346404493,-0.0850382,-0.3230128884,0.3395520747,0.0269842912,-0.1159277931,0.0191253033,-0.3266032934,0.2890707254,0.2416296005,-0.1814799607,0.0906046107,-0.3929245174,0.0673670247,-0.1850719601,0.0427736416,0.4216829538,0.0768567622,-0.039862372,-0.1143276766,0.4105682969,-0.2401659936,-0.1583056599,-0.0344501659,0.3706847429,0.1414440274,0.0601721928,-0.0102416575,1.0200059414,0.0521302633,-0.1580085009,0.2711504102,-0.0472821966,0.519864738,-0.1478833258,0.2608551681,-0.4259823263,-0.1347872466,0.0839140937,-0.2199781835,-0.0904907286,0.1006634533,-0.2206875086,0.1749460101,-0.1338852346,0.4090332687,-0.0596338958,0.2469776422,-0.0361241065,-0.073927097,-0.4289419651,0.2308985293,-0.0863795504,0.1210720688,-0.050541956,-0.0839664042,-0.2879820168,-0.1469539553,-0.1935782135,0.0809503943,0.0188850742,0.2164885253,0.3956883252,-0.2521114647,0.0806011036,0.4948326349,0.3612062931,-0.0179025829,-0.1604380608,0.1334342062,-0.325479269,0.09686324,0.0188934971,0.2583013475,0.0252778679,-0.3979592025,-0.1800468415,0.0508729294,0.0217197556,-0.1788694263,-0.3010591865,0.1531126648,0.0394399315,-0.2973150909,-0.2082680613,-0.0005502225,-0.0494779423,-0.258351773,0.1320417374,0.0672353059,-0.2313183397,0.3131327629,0.3865985274,-0.1113000661,0.2371282876,0.4063925445,-0.2009040117,-0.1178988591,0.1609756351,-0.1262585968,-0.2319331616,-0.2166242003,-0.0883549824,-0.2817947865,-0.2436751872,0.099719651,0.11000067,-0.1503900737,-0.1758230925,0.2263848633,0.2988093495,0.1472547203,0.0732385367,-0.3525341749,-0.2241643667,0.2353284806,0.0177501701,-0.1402800977,0.012005276,0.3174556792,0.0925339684,0.3731617928,-0.3981429636,0.0657028481,-0.3477116823,0.1524640024,0.0164833181,-0.30829072,0.341386348,-0.0658024028,0.0670218021,0.0657467246,-0.4848766029,-0.225449726,-0.0560643002,0.1289645433,-0.1247254759,0.2390374392,0.2399483025,0.177708447,-0.1101103649,-0.1238408089,0.1057478264,0.2011776268,0.0197200235,0.1939899623,-0.2998627126,0.3818893433,0.1315320879,-0.327737838,0.2166209221,0.0656026676,-0.1260645241,0.0701941773,-0.1483680755,-0.0619577058,0.1529890448,0.2921739817,0.1241738126,0.1555727124,0.1947733462,0.0618586391,0.038716469,0.0105805807,0.140981108,0.1572409272,-0.200759992,0.1395862401,-0.2652696073,0.2051884681,-0.3029921353,-0.1528884023,0.4850533307,-0.0681184083,0.2212644815,0.1986716986,0.2453091145,0.0994070843,0.21999152,0.1079205796,0.599126637,-0.0610687956,0.0599963702,-0.0868286863,0.0118273823,0.1556928605,0.3291961551,0.0128815854,0.246457994,0.014765583,-0.0369104929,0.3591687679,-0.0478717834,0.0404678844,0.1642526239,0.2609469593,-0.2687528431,0.0692028105,0.1256689876,0.1583086401,-0.0700464994,0.7059479356,-0.2635207772,-0.049814716,-0.2375159562,-0.1269139647,-0.0177107584,-0.0924670249,-0.100991644,-0.2524999976,0.0155231543,-0.2578632236,0.2252056152,-0.1401660889,0.234828338,-0.2815510631,0.1279933304,0.1161624491,0.0049282219,0.3126760125,0.3076825738,-0.0500043668,0.0465313271,0.1621740162,0.3137163818,0.2272781581,0.4471285045,0.3903549612,0.0512046516,-0.0292133093,0.083307147,-0.092486985,-0.0691982806,0.0647646487,0.1232277602,-0.0407954939,-0.560495019,0.3493500054,0.1857195348,-0.2499478459,0.0837995708,-0.0217214487,0.0114817098,-0.0812999904,0.1417374015,-0.250515759,-0.0043504336,-0.1979366988,0.0736898482,-0.1613987982,0.0595474653,-0.160268873,0.2696827948,0.1501699686,-0.2571901679,0.1407111138,0.0442430042,0.1846884191,0.2414643317,-0.0327478871,-0.0230107177,0.0277725123,-0.7008520961,0.2437223345,0.2913865745,0.1446157843,0.0640144795,0.2949914038,-0.1753889471,0.2501959503,0.0487765782,-0.2412872165,-0.0748223513,-0.1962623596,-0.4760009348,0.0213080477,-0.2091042548,-0.0578983389,0.1915591806,0.0349221118,-0.0243190285,-0.2226160914,0.0521260649,0.0932007656,0.0291387588,-0.0105784088,0.0017508153,-0.1912248731,0.1408903152,0.4110670984,0.0085812137,0.0723993331,0.1618169099,-0.3288664222,-0.1204130575,0.1013039201,-0.0152519476,0.2814262807,0.02575415,-0.4013804495,-0.053398937,0.4772090316,0.2339962125,-0.3452475369,-0.4014283717,0.0463571213,0.0237442907,0.0774244219,-0.0262260418,0.4119680226,-0.0833834261,0.2936124206,-0.1765293628,-0.4193325639,0.7072589993,-0.3268920481,-0.3418447673,0.0193426479,0.3970513344,0.3459706008,-0.1593000144,-0.6697818637,-0.137917757,0.2949357629,0.0028172303,-0.1122922972,0.1081515774,-0.4581031799,-0.0393696837,-0.0313256197,-0.2585921586,0.116931513,-0.1140028089,0.1401878297,-0.2812470794]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/657","title":"Squad Metric Description & Feature Mismatch","comments":"But then providing the `answer_start` becomes mandatory since the format of the features is checked against the one provided in the squad [file](https:\/\/github.com\/huggingface\/datasets\/pull\/658\/files).","body":"The [description](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/squad\/squad.py#L39) doesn't mention `answer_start` in squad. However the `datasets.features` require [it](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/squad\/squad.py#L68). It's also not used in the evaluation.","comment_length":23,"text":"Squad Metric Description & Feature Mismatch\nThe [description](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/squad\/squad.py#L39) doesn't mention `answer_start` in squad. However the `datasets.features` require [it](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/squad\/squad.py#L68). It's also not used in the evaluation.\nBut then providing the `answer_start` becomes mandatory since the format of the features is checked against the one provided in the squad [file](https:\/\/github.com\/huggingface\/datasets\/pull\/658\/files).","embeddings":[0.0714836419,-0.3686691225,-0.0856218711,-0.0963039696,0.409524858,-0.104312025,0.1253436506,0.0395658389,-0.2808919847,0.0161719397,-0.0155542269,0.3508732915,0.2142232358,0.0877162814,-0.1212031543,0.1233918071,0.0228273273,0.1146427095,-0.1461589038,-0.0721072927,-0.1057656407,0.2958207428,-0.3101403713,0.1393026263,-0.3326881826,0.2032144964,-0.0322561525,0.2053763419,-0.5319287181,-0.4879083633,0.3031818569,0.0732051581,-0.2346524149,0.2549627125,-0.0001087786,-0.1570355594,0.1928605288,-0.1502050757,-0.0956225917,-0.1865075678,0.1002465189,-0.301348418,0.208040446,-0.2634674609,-0.1206465289,0.1266855448,-0.0258798562,0.015812831,0.5856692195,0.1773659736,0.2096451968,0.527971983,0.0051359362,-0.2664542496,-0.0079287859,0.0832580253,-0.1241676286,0.2411233634,0.3226965666,-0.009124618,-0.2592707872,0.215482235,0.1822725385,0.090565145,0.415504694,0.0111930734,0.1710323691,-0.1892596632,-0.1101630777,0.0799341425,0.2163025886,-0.1620045155,-0.2368020415,-0.1409005076,0.0533003248,-0.1408354789,0.2025523186,0.0599073395,-0.0294725504,0.0576386154,-0.2525348961,0.0616584718,-0.1925629675,-0.056944292,-0.0384190679,0.1845898181,-0.2179058045,-0.0223872345,0.0588668995,-0.077031076,-0.3285889626,0.0275729075,-0.0039779129,0.1555655897,-0.346353054,-0.2450008839,0.1401151717,-0.012090506,0.5897557139,0.130713135,0.0278300848,-0.0755563453,0.1916167736,0.0966398194,0.0683344454,0.359431833,0.5389780402,-0.260147512,-0.0470285937,0.0979135334,-0.0602265336,0.0362291634,0.2225055546,-0.1853069216,0.1102717891,-0.0105525507,0.1973386556,-0.1863773465,-0.3243863881,0.2362409085,0.0331388563,-0.1849229038,0.1610157937,0.3308874965,0.0394728146,-0.1160728559,-0.1364458352,0.3753709197,-0.0615498088,0.1788348407,-0.2362853438,0.0175781846,-0.1785049289,0.0176158082,0.0770150349,-0.094592005,0.468426168,0.1037268713,0.6205449104,-0.2139698863,0.0900896639,0.0907742456,-0.0086815041,-0.0588397644,-0.0381924137,0.0323758982,0.3155770302,-0.1487368196,-0.0383848026,-0.0121804858,-0.3050554395,-0.1685030907,0.0228985567,0.247195974,-0.4179111123,0.1742908955,0.1866101474,0.1592618972,-0.1578108668,-0.0198884755,0.0972554982,-0.0071580592,-0.2175530791,-0.0871073008,0.4565599561,0.3123025298,-0.0138828969,-0.1896150857,0.4936756194,-0.2422699481,-0.3501276076,0.1747062504,-0.087278381,-0.0625979453,-0.2254831791,0.1653340608,0.4319165349,-0.6663405299,-0.121143572,0.1177966073,-0.4087069035,-0.1348310262,-0.1633995771,-0.1139434278,0.2157614082,0.041427508,0.0647868514,0.1040396988,0.0628902018,-0.0814871639,-0.3227227032,-0.096813947,-0.1105398908,-0.0340452082,-0.2024563849,0.0050684786,-0.0367087238,0.2662734389,-0.0300772581,0.1435508728,-0.2213812768,0.0760081783,0.2579780221,-0.2862824798,0.0779082999,-0.1838898659,-0.3849327564,0.0795555711,-0.3601306379,0.0163337737,0.1635149717,-0.2715017796,-0.542047739,0.0469520278,-0.1306706965,-0.1301482469,0.196409449,-0.2922311127,0.1908030063,-0.0649100021,-0.2068103254,0.0555554591,-0.5343649387,0.2205049247,0.0747225508,-0.039069593,-0.2212112099,0.1489003301,0.1544100344,0.1852428764,0.1026900783,-0.0409526005,-0.1115171164,0.5497416854,-0.0968206823,0.1595625728,0.2671706676,-0.0458367988,0.0607396588,-0.1218105704,-0.1407346427,0.0995274484,-0.0703592896,-0.1978718042,-0.3766762614,0.4630094767,-0.1700356603,0.1172300279,0.1548947841,0.0790314749,0.19189924,-0.1135168746,-0.4225904047,-0.2107368112,0.1589574367,0.0613230653,0.0597392358,0.067355983,-0.2234184444,-0.0676215887,0.7663306594,-0.0128328009,0.0169440471,0.1607962549,0.0620675348,-0.1778208166,-0.1424493641,-0.4926059842,0.5218940973,0.2119178772,0.0149998432,0.0737033784,-0.1531002074,-0.192088604,0.0915806592,0.0498369373,-0.0947701484,0.2592765689,0.2015848011,-0.1314311922,-0.0671300739,0.0983822271,-0.1724919826,-0.1129594594,-0.2569566965,-0.1258196682,0.1275725663,0.0317176543,-0.1661224812,-0.1506333798,-0.3098291159,-0.2745752037,0.3487704396,-0.0223367456,0.2935876548,0.3687449694,0.1583545059,0.336248219,0.0544574857,-0.1106502786,0.1403378695,-0.2051765025,0.1116326004,0.0981991366,-0.1121825799,0.0899347961,0.2285868078,-0.5091598034,-0.0139453905,-0.0505290627,-0.4765910506,0.2807789445,-0.0454615019,0.3085437715,0.0962103009,0.0719201341,-0.3590722382,0.0183290895,0.3224570453,-0.3235698938,-0.3216747344,-0.1864978522,-0.0132516613,-0.2689188123,-0.3647123873,-0.5461171269,0.0952373967,-0.2989224494,0.3766980469,0.1400758773,-0.0273447372,0.2844640315,-0.0148345474,0.3455062509,-0.2718372643,0.006417166,-0.3936613798,-0.4697285295,0.226873219,-0.1384315342,-0.4657115042,0.0588319972,0.0330042243,-0.0363267474,-0.2294948697,-0.0529052913,-1.0586352348,0.0498611033,0.1250767112,-0.2215347439,0.1816776693,0.3187162578,-0.5645013452,-0.0541788153,-0.3160336912,-0.3421313465,0.3988611102,0.2140415609,0.2097068578,-0.3139363527,0.3115652204,0.1362070441,0.9513806105,0.2537688911,0.1607136726,-0.0673191771,-0.0600464046,0.3613639176,-0.1113016158,-0.278830111,0.301622659,0.0773936585,0.1918253452,0.3122570515,0.1924505532,0.2049937248,-0.0494212918,0.3052556217,-0.1406046301,-0.2224221975,-0.2549998164,0.2897501886,0.0553185232,0.103733547,0.1859121472,-0.2607045472,0.0168894045,-0.0293790456,0.4318302572,0.248383224,0.1019391045,-0.8079233766,-0.1121246442,-0.0766044259,0.1463543773,0.2055116892,0.085245505,-0.0143188676,-0.2051854134,0.0176267661,-0.059009254,0.5863503814,0.3503597379,-0.1195380613,-0.0343176201,0.2263382673,-0.5820803642,-0.1642438024,-0.1585285664,0.4097059369,0.5023412704,0.5473333597,-0.4737728834,-0.3406572938,0.1070994139,-0.1551385075,-0.0475146696,-0.0328378044,-0.5244740248,-0.2369216532,-0.2704086006,0.2653136253,0.1774002761,-0.0157449599,0.2089426816,0.0026247315,0.31666556,-0.0421448015,0.197839722,0.4141670763,0.1020237878,0.0150055997,-0.0443547294,0.0389795005,-0.0707699582,-0.3313654661,0.3410165608,-0.0112076979,-0.0931079835,0.1719153076,-0.329269588,0.2612441182,0.4037445188,-0.1517472565,0.1503641754,-0.4624970853,0.143188864,-0.211234659,0.117527768,0.4971188307,0.0586170107,-0.0626974404,-0.0065459763,0.2420920879,-0.2178587914,-0.0609112568,-0.0005251365,0.4124797583,0.1833920032,0.0914430097,-0.0982117802,0.9600061774,0.125190571,-0.0917282328,0.2422216684,0.0418736115,0.4939926565,-0.0268487316,0.2482564896,-0.2694194019,-0.2852006555,0.1116491705,-0.2703735828,0.044056233,0.0544328727,-0.211407423,0.1932344139,-0.0683503523,0.5172485709,-0.0599550679,0.3799267411,-0.1693510115,-0.0310522206,-0.3296203315,0.2161275744,0.0824195072,0.0719730482,-0.0860538408,-0.0578269251,-0.1177077517,-0.3063986003,-0.0524224751,0.0447315574,-0.0459589697,0.021889884,0.4349378347,-0.2040510923,0.1554914713,0.3100825846,0.3358572423,0.0439823791,-0.0902956054,0.186278075,-0.1875779182,0.0377072319,0.0011450431,0.2060986608,0.1023438647,-0.270734787,0.0024800703,-0.0309493616,0.082914792,-0.1804997921,-0.2589581907,0.0774727315,-0.0274888612,-0.3178336918,-0.3131406903,-0.0012585934,-0.1030943915,-0.1117482558,0.1407167315,0.0107708266,-0.1047642753,0.2426192164,0.3311521709,-0.1736882031,-0.0474306084,0.266350925,-0.3080544472,-0.0539530292,0.1893165857,-0.1640928835,-0.0731058121,-0.2250332534,-0.0445178449,-0.2047300339,-0.2611697018,-0.0439944342,0.2591428161,-0.1860081702,-0.0765228719,0.0950375423,0.3709025681,0.1623034626,0.2081296295,-0.3234916031,-0.2744864225,0.1854543984,-0.106858775,-0.2031576633,0.1021007895,0.416310817,0.0986238793,0.2295036763,-0.4194539487,0.1195935458,-0.1490286142,0.0883522108,0.0453021042,-0.3655311465,0.3733586669,-0.0366696566,0.0840919018,0.0656248182,-0.4903034866,-0.2547087967,-0.0756020322,0.1120662838,-0.0457424745,0.029030459,0.0841076672,0.0154872816,-0.1726397872,-0.0816719756,0.0525547601,0.2083191425,0.1003898308,0.0571951419,-0.3976977766,0.1715798378,0.0665492788,-0.3400273621,0.2161442339,0.0923185274,-0.0706029907,0.0571207479,-0.3034576178,-0.0367362164,-0.0810490251,0.3800121248,0.156839475,0.1931458563,-0.0197589379,0.0780095309,0.0301586483,-0.0885217488,0.1816163361,0.1655511558,-0.2564044595,0.115065068,-0.2157732844,0.2028285116,-0.1534432322,-0.0049390062,0.4422463179,-0.0443846285,0.2240913361,0.3737488687,0.2172664851,0.1900696903,0.3056747615,0.275015533,0.4591741264,0.0920072719,0.0804986581,-0.0482119732,0.0765820295,0.1516328603,0.501111865,-0.0244693141,0.3242197633,-0.1924531013,-0.0954265669,0.3372444212,-0.2135297507,0.1509516835,0.1670290828,0.1110204682,-0.2499295473,0.1986942589,0.1473441422,0.1880239695,-0.0462240539,0.8048347831,-0.3650283813,-0.1945146769,-0.2721890807,-0.1492254436,-0.0865748525,-0.1328081787,-0.0712952837,-0.2399506122,-0.0168199167,-0.1528507322,0.0741856247,-0.0917199999,0.0913095251,-0.1889369786,-0.0162144117,0.0621758774,0.0030816193,0.3148223162,0.1481962055,-0.0151353041,0.1107647717,0.0245119706,0.2311199307,0.3774188459,0.4738669097,0.2855056226,0.1190149412,0.1324759424,0.0000619727,0.0421579555,-0.1417860687,0.058939606,0.072218284,-0.0158619005,-0.4801882207,0.2754590511,0.2455084771,-0.2900890708,0.1766071469,0.0518721528,-0.0513360202,-0.2626931369,0.0709407628,-0.1839474738,0.1674204618,-0.228130728,0.0863311514,-0.3151623309,0.012538854,-0.1503469944,0.2780932784,0.1792273372,-0.2272629589,0.1218722388,0.0491319187,0.0213146601,0.2931469679,-0.0375427231,0.0089207729,0.0167212728,-0.7170854807,0.1781845689,0.3404882252,0.3211581409,0.1111536771,0.3643413782,-0.1321587563,0.321606189,0.1501121223,-0.2261168957,-0.0753632262,-0.2230733186,-0.34501791,0.0889578983,-0.0635394529,0.0358619541,0.1450353414,0.0524478294,0.0716812089,-0.1267293543,0.041466251,0.0683657378,0.1159667075,-0.0059020333,-0.1477762014,-0.0798495412,0.1761415452,0.3649193048,-0.0657626688,-0.0132515375,0.2049595416,-0.1980928332,-0.1940237135,-0.0136133432,-0.0612450354,0.3370971084,-0.0515946113,-0.4326334298,-0.0818319172,0.4637615383,0.1893284321,-0.4613316059,-0.3907948434,0.2044506967,-0.0979915857,0.0411250181,-0.100576438,0.5699029565,-0.0965008959,0.363768965,-0.2342429906,-0.3605833948,0.5602978468,-0.2598425746,-0.1011114568,-0.0214640144,0.413811177,0.3865252137,-0.2456570417,-0.6112876534,-0.1607122123,0.1788600236,-0.0031611568,-0.2295849919,0.1471381485,-0.3713880777,-0.0177422352,-0.0656003952,-0.0327629633,0.2191711813,-0.2374016196,0.1443990171,-0.1633359492]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/651","title":"Problem with JSON dataset format","comments":"Currently the `json` dataset doesn't support this format unfortunately.\r\nHowever you could load it with\r\n```python\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\n\r\ndf = pd.read_json(\"path_to_local.json\", orient=\"index\")\r\ndataset = Dataset.from_pandas(df)\r\n```","body":"I have a local json dataset with the following form.\r\n\r\n{\r\n 'id01234': {'key1': value1, 'key2': value2, 'key3': value3},\r\n 'id01235': {'key1': value1, 'key2': value2, 'key3': value3},\r\n .\r\n .\r\n .\r\n 'id09999': {'key1': value1, 'key2': value2, 'key3': value3}\r\n}\r\nNote that instead of a list of records it's basically a dictionary of key value pairs with the keys being the record_ids and the values being the corresponding record.\r\n\r\nReading this with json:\r\n\r\n```\r\ndata = datasets.load('json', data_files='path_to_local.json')\r\n```\r\nThrows an error and asks me to chose a field. What's the right way to handle this?","comment_length":32,"text":"Problem with JSON dataset format\nI have a local json dataset with the following form.\r\n\r\n{\r\n 'id01234': {'key1': value1, 'key2': value2, 'key3': value3},\r\n 'id01235': {'key1': value1, 'key2': value2, 'key3': value3},\r\n .\r\n .\r\n .\r\n 'id09999': {'key1': value1, 'key2': value2, 'key3': value3}\r\n}\r\nNote that instead of a list of records it's basically a dictionary of key value pairs with the keys being the record_ids and the values being the corresponding record.\r\n\r\nReading this with json:\r\n\r\n```\r\ndata = datasets.load('json', data_files='path_to_local.json')\r\n```\r\nThrows an error and asks me to chose a field. What's the right way to handle this?\nCurrently the `json` dataset doesn't support this format unfortunately.\r\nHowever you could load it with\r\n```python\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\n\r\ndf = pd.read_json(\"path_to_local.json\", orient=\"index\")\r\ndataset = Dataset.from_pandas(df)\r\n```","embeddings":[0.1458693445,0.1199862286,-0.0658117831,0.380238533,-0.099493064,0.2573181391,0.2378643304,0.4517599046,0.4588271081,-0.0621231049,0.1368803829,0.4856531322,-0.1455082148,0.2548018694,-0.3202168941,-0.1480671614,0.0985162109,0.2186025679,0.2703855634,-0.0470258854,-0.439702332,-0.1153432727,0.0489592962,0.0190584902,0.0482812189,-0.2671149373,-0.0260776114,0.0717845112,-0.1620740592,-0.4581755102,0.156412676,0.1291371733,0.0828994438,0.1049765572,-0.0001129885,0.0956471115,0.1844623536,-0.0450863615,-0.1903440058,-0.299005568,-0.7772290111,-0.185928598,0.3369084001,-0.2603326738,-0.2133954018,-0.8189460039,-0.1771575511,-0.548494637,0.3485087454,0.0538749993,0.1756142378,-0.0457043499,-0.1172699705,0.2295890301,-0.1242401972,0.4465412796,0.0052379719,0.2138409317,0.1540415138,0.2234113812,0.1710988283,-0.1115013063,-0.2714059353,-0.1085489169,0.3714460135,-0.0568574965,-0.0460285805,0.0683584064,0.0254373662,0.2053817064,0.730425477,-0.0766048282,-0.455851227,-0.20778808,-0.070964992,-0.0512549803,0.0490764119,0.1400657892,-0.031733308,0.4423697889,-0.1603931785,-0.1668311059,-0.0914005786,0.4621090591,-0.1848821789,-0.0156272762,-0.0242636725,0.1310406029,-0.0488182642,-0.3260755539,0.0286141243,-0.4399697781,0.1288776398,0.0635579601,-0.0337626673,-0.2168125808,-0.0416974947,-0.3340826035,0.2119557559,-0.0481729284,-0.0074766404,0.1113776192,-0.2421927154,0.3040264249,0.6694840193,-0.2420202792,0.4449410439,0.0218048561,0.0466984026,-0.1306554377,0.2354110777,-0.0629480258,-0.1822791696,0.029254932,-0.2261861563,-0.2180085033,0.094256714,-0.1267885864,-0.1877722442,0.463996768,-0.4986902475,0.038636975,0.0729411542,0.164200291,-0.0591540337,0.4581668079,0.0739175826,0.309589386,0.2918245196,0.2605212927,-0.092348285,0.0517236143,0.298828721,-0.200180918,0.0685544983,-0.0650402978,0.2729404271,0.1133925244,-0.0976343974,-0.1352136433,0.0604379103,0.0339622423,0.03365485,0.3438401818,0.115554899,0.2728944123,0.058720082,-0.7610890269,-0.2815734446,0.3070136011,-0.1231466979,-0.1196510568,-0.2131774276,0.1140166596,-0.214539364,-0.1683910191,-0.5135980844,0.1099394187,-0.0350384787,0.0967568904,-0.1146052927,0.0682745203,-0.0962640792,-0.3119336367,-0.0715676174,0.2158742994,-0.7670142055,0.0624736212,0.2604255378,0.1439981759,-0.0009404705,0.3863136768,-0.3122444153,0.1130622253,0.0466580428,0.2099815607,0.346383214,-0.4240632057,-0.1853321046,0.6589516401,0.131164059,0.0572888851,0.1835733652,0.0641262904,0.0861064792,-0.1249698699,0.4913157225,0.2782319784,-0.1301450729,0.13153328,-0.0413275771,-0.2937499583,0.3844843507,0.1433086097,-0.2987167835,0.0400857553,-0.0638527125,0.1849567145,0.1030333042,-0.1027433276,-0.0346187055,0.2544205189,0.1330920458,-0.2291435748,-0.0917216688,0.1678028256,-0.4931346178,0.0583169647,0.0407988913,-0.2706646919,-0.1732385457,-0.22928828,-0.2658457458,-0.0448416956,-0.3415186107,0.3753657937,0.1029696763,-0.1386786848,-0.0613617785,0.0832327902,-0.1890549809,-0.1287673414,0.1051743925,0.0485650003,-0.3560819626,0.2340255082,0.0891059116,0.1402282566,0.1026291475,0.0952601954,0.1381038129,-0.0816348866,0.049051594,0.2223201394,0.2931486964,-0.0688418895,0.0757608488,0.1965526193,0.2696464062,0.2145307511,-0.0805483013,0.0517453253,0.2669302523,-0.1208436042,-0.4998886585,0.5624603629,-0.0740086138,0.2754159272,-0.2417865694,0.1344882548,0.5070407391,0.1711891741,-0.2448195219,-0.2652659714,-0.1872988194,0.2555713952,0.0550000519,0.1753744334,-0.5730938911,0.0807779431,0.3986865878,-0.1162142083,0.1785604507,0.0774062574,0.0183163974,0.1102359593,0.0166593287,0.3469281793,0.1840494573,0.1019805148,-0.1722500473,-0.1593765467,0.1704727858,-0.1262270063,-0.030276861,0.0629003644,0.1021463275,0.1732898206,0.1294019371,-0.2646361589,0.0097442307,0.1082995757,0.1709943116,0.3211760223,-0.1843876839,0.2181835771,-0.5527148843,-0.1690990031,-0.1915680319,-0.3349094391,0.0057535684,-0.2660440803,0.0143419644,0.0908641964,-0.1945691407,0.0133133223,0.0123492973,0.0773505047,-0.0093088774,-0.4200900793,-0.0577931404,-0.170786798,-0.1814787835,0.0558716878,0.4747773409,0.0712883994,0.1124226898,-0.1644038558,-0.1423163116,-0.0538366847,0.1102759913,0.1040438935,-0.0833402127,0.1146071479,0.1896594167,0.5317562819,-0.3402480483,-0.1636222452,0.2713463604,0.3887449801,-0.1310292035,0.0392323136,0.139456138,0.059592884,0.0794297606,-0.3075087965,-0.1711047143,-0.2167973965,0.435690552,0.1593658328,0.1710596681,-0.0084817931,0.1128979996,-0.0164313708,0.1447687745,-0.1445475072,-0.0614725128,-0.3365549147,0.3744822741,-0.1530004442,-0.3797767758,0.0752773434,-0.0251518954,-0.0309548583,-0.2401718646,-0.099135004,-0.0014947819,0.042244602,0.0657009631,0.016220659,-0.1312144399,0.1923988461,0.1064992845,0.0139339566,-0.0919715241,-0.1125551462,0.2691612244,0.2732815742,0.1071154699,0.3044460416,0.5737470388,0.0091620311,0.1929805726,0.2185966223,-0.0632901192,0.082307376,-0.0631020144,-0.0089376373,-0.081286788,0.0083392365,-0.2582142651,0.0239418838,-0.12837556,0.0636104867,0.0780435205,-0.1122457385,-0.2132260501,0.1502926201,-0.4659946263,-0.1925676018,0.1423623115,-0.2337083817,-0.0703254193,-0.2169299871,-0.0009835879,0.1018207744,0.0360618234,-0.1283742785,0.2931230366,0.1016212404,0.0939056724,-0.4172830284,-0.3083844185,-0.4276412427,0.136715889,0.1104796156,0.7288771868,-0.0195554178,0.1968232244,0.0636686236,-0.0242178757,0.5950882435,-0.1059881896,0.0960163772,0.1525070965,0.1006457508,-0.4992520511,0.2310392112,-0.0852854177,0.3208945096,-0.2690153718,0.4516743124,-0.4052474797,-0.3435758948,-0.1920021325,0.2402451932,-0.2700727582,-0.2101640552,-0.1921625584,-0.2964921594,-0.2380502075,-0.2589570284,0.2959805131,0.1589806527,0.33596766,-0.0691056028,-0.1125484556,-0.0040718154,0.1293795705,0.0557941087,0.1993985027,0.2729334831,0.1131541952,0.1793407053,0.1726690978,0.4247415662,0.54626894,0.3334120512,-0.1455210596,-0.0143334875,-0.2502033412,0.2467488199,-0.0559655204,-0.0126835583,-0.0202426091,0.1780313104,-0.2735628784,-0.1463044286,-0.1517327577,0.3680625856,-0.3114070892,-0.3521307707,-0.5186553001,0.6033038497,0.0841253698,-0.0123819178,0.025399467,0.2774887085,-0.1227986738,0.4810540974,0.2179086953,0.5684335828,-0.3487242162,0.0267522503,0.1977459937,-0.5146607757,0.4830617011,-0.2183378041,-0.0967940241,-0.3696780205,0.3934253454,-0.1548226923,-0.0373506211,0.372744143,0.4814714491,-0.0547152981,0.1361578107,-0.3032663465,-0.0946575254,-0.1093672961,0.2447793782,0.1032146737,-0.1591505557,-0.4589945078,0.0816945285,-0.2037877887,-0.2845903635,0.0229372792,-0.1684264392,-0.3422836065,-0.1301282644,-0.1290245652,-0.217422992,0.3331001103,0.0794570819,0.2641437054,-0.2753999233,0.1338727176,0.254866004,0.0644343272,0.1171610951,-0.1934216022,-0.2033862025,-0.3793017268,0.0770608187,-0.0499614589,0.0118617984,0.118482545,0.098120667,-0.0415067859,-0.1009148881,-0.1732546091,-0.1489336938,-0.1010818556,0.2071549147,-0.1293741465,-0.3121788502,-0.56266886,-0.0551258624,0.0729702339,0.0901851729,0.1077717692,0.0797572285,0.0778418258,0.0646306947,-0.0084436396,0.0016054547,0.1139384657,0.3233715296,0.066838108,-0.0398823507,0.3139288127,0.1302960813,-0.1649362892,-0.1316999942,0.5236610174,0.3322329521,0.2576482892,-0.0298174508,-0.1930639595,-0.0104917483,-0.2179306,0.3178390861,-0.0223678034,-0.1482021362,0.2405258715,-0.4940349162,-0.2180171162,0.2946761549,0.2369864583,0.0139890658,0.1584945172,0.0604558848,-0.2530004978,-0.2274926305,-0.2128077745,-0.0762625486,-0.3122350872,0.2695815861,-0.1362293959,0.2458582222,0.0352968015,-0.2442626506,0.1339185238,-0.1740272343,0.0646461025,-0.2034875154,-0.0977171883,0.114831917,0.3019968867,-0.142211318,-0.0085985987,-0.2923331261,-0.2843970954,0.039489802,-0.1387791038,0.2076571435,-0.0695353746,0.4448586106,0.0516576953,0.019893622,0.0777343586,-0.0676850528,0.0560830235,0.2144360989,-0.0532594919,0.0691271946,0.1086296812,-0.0738356337,-0.2708690763,0.1666848511,-0.1808829457,0.1374463588,0.2203943878,-0.4979730546,0.41076684,0.4757048488,0.39567101,0.4644890726,-0.2230915427,0.2462891936,0.2214041203,0.1531096995,-0.0204890817,-0.50728935,0.3848720789,0.1289762259,-0.0782069415,0.2625131011,0.0682198703,-0.0413618051,-0.353328675,-0.2404543459,0.2358684838,-0.0373976007,-0.1141340509,0.5161632895,-0.0539601296,-0.2124410421,-0.0689396411,0.5151497126,0.2435200214,0.20899266,0.2079000175,-0.0601607077,0.4138967395,0.0516661257,-0.0301798619,-0.4153177142,0.4144587517,-0.0049183224,0.0307821613,-0.0177142303,0.4281779826,0.153610006,0.0701021403,-0.001690089,0.2020605654,-0.0245910157,-0.050845623,-0.2886847556,-0.3768408298,-0.1583996266,-0.0580376945,-0.1849884391,-0.2045609951,-0.3123615384,-0.1415838152,0.0404950194,0.2243338972,-0.3365134001,0.4711937606,-0.0136737144,0.1332265735,-0.2788590193,0.059794005,0.1122533455,-0.1988800019,0.2645430267,-0.0535119213,-0.0188860577,0.1589214206,-0.1320626587,0.0369451717,-0.5998958349,-0.1819780022,-0.0855077729,0.055605907,-0.0905480236,-0.1600075662,0.2966666222,0.1264388859,0.0164629538,0.3209455013,0.1543247253,0.0716397241,-0.0946691781,-0.0717239454,-0.1753923744,-0.1442168653,0.2563312054,-0.1602598578,-0.1590497643,-0.4251284003,0.1613835394,0.379804194,0.3913544714,0.1206353381,0.0776892975,-0.1784189343,0.3760291636,0.0717430636,0.0428183898,0.0304697324,0.1283907145,-0.207628414,-0.1148080081,-0.0716458559,-0.2363211066,0.4204461277,0.1062995046,0.2889698148,0.3643233478,-0.2331486642,0.4509801567,-0.3636855185,0.1910832971,0.0698704571,-0.4637273848,0.1111583263,0.1946075559,-0.0551870093,-0.083493866,0.2223051041,0.1503245085,0.0707844049,-0.3576923609,-0.0745876059,-0.2778406739,-0.1664799601,-0.1417878419,0.2965258956,0.0731141493,0.1139508188,0.0423507281,0.1826490909,-0.0085291015,-0.1535207927,0.4577557743,-0.232989803,0.4298778474,-0.2028204501,0.1843889803,-0.1410553753,-0.2135638595,-0.3715863526,-0.1534892917,-0.3980414271,0.2372504026,-0.1922960579,0.0056450949,-0.1154149771,0.2940388918,-0.1114013121,0.3417697549,-0.0218548495,-0.0153428456,0.2648360431,-0.3608396351,-0.1684058011,0.4840176105,-0.1143330261,0.0133906677,-0.1234349534,-0.0647939444,-0.0121954773,0.3006780446,-0.1669054478,-0.1197579801,-0.2389780283,-0.0009335708,-0.1093443483,-0.095691368,0.0829408169,-0.077634953,0.0185358394,-0.1587726027,-0.2314792871]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/651","title":"Problem with JSON dataset format","comments":"or you can make a custom dataset script as explained in doc here: https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html","body":"I have a local json dataset with the following form.\r\n\r\n{\r\n 'id01234': {'key1': value1, 'key2': value2, 'key3': value3},\r\n 'id01235': {'key1': value1, 'key2': value2, 'key3': value3},\r\n .\r\n .\r\n .\r\n 'id09999': {'key1': value1, 'key2': value2, 'key3': value3}\r\n}\r\nNote that instead of a list of records it's basically a dictionary of key value pairs with the keys being the record_ids and the values being the corresponding record.\r\n\r\nReading this with json:\r\n\r\n```\r\ndata = datasets.load('json', data_files='path_to_local.json')\r\n```\r\nThrows an error and asks me to chose a field. What's the right way to handle this?","comment_length":14,"text":"Problem with JSON dataset format\nI have a local json dataset with the following form.\r\n\r\n{\r\n 'id01234': {'key1': value1, 'key2': value2, 'key3': value3},\r\n 'id01235': {'key1': value1, 'key2': value2, 'key3': value3},\r\n .\r\n .\r\n .\r\n 'id09999': {'key1': value1, 'key2': value2, 'key3': value3}\r\n}\r\nNote that instead of a list of records it's basically a dictionary of key value pairs with the keys being the record_ids and the values being the corresponding record.\r\n\r\nReading this with json:\r\n\r\n```\r\ndata = datasets.load('json', data_files='path_to_local.json')\r\n```\r\nThrows an error and asks me to chose a field. What's the right way to handle this?\nor you can make a custom dataset script as explained in doc here: https:\/\/huggingface.co\/docs\/datasets\/add_dataset.html","embeddings":[0.1518978626,-0.0548713207,-0.0390631855,0.3815723062,-0.0431255847,0.2024196684,0.1602234393,0.2896431088,0.4436892569,0.0094779637,0.1342986375,0.3551860154,-0.2072727978,0.6003725529,-0.0831734389,-0.0579490475,0.1013266519,0.2328083366,0.1420461684,-0.0527270846,-0.395789355,0.0413110331,0.1581416875,0.0273906644,-0.0831316859,-0.1592040807,0.0789325982,0.1683625877,-0.1300393492,-0.4681346118,0.2148664743,0.1719652712,0.044767037,0.1334680021,-0.0001166466,0.0707662255,0.0919193327,-0.1179516166,-0.2690297067,-0.2829463184,-0.5637429953,-0.1780745387,0.3497334123,-0.2107811868,-0.2849601209,-0.7782123089,-0.1523974389,-0.4623758495,0.3354526758,0.0153777543,0.1571149826,0.1514682621,-0.0361434966,0.1626404226,-0.0200143345,0.652559638,-0.0668515936,0.2499583811,0.1642249376,0.2087757438,0.1236287355,-0.0067370082,-0.1066929847,-0.0750198662,0.5859818459,-0.0021051902,-0.256711483,0.0410431363,0.0158182234,0.3569475412,0.6222531199,-0.1870975494,-0.4084343314,-0.3341347575,-0.0829559863,-0.0947225764,0.1387061328,0.1777618974,-0.0051165614,0.4336256683,-0.3924814761,-0.2523345351,-0.0882712677,0.4107298255,0.0611397699,-0.0854333341,-0.1614599675,0.171757862,-0.0244887304,-0.2981067002,-0.2625541687,-0.4201046228,-0.034731362,-0.0047964584,-0.0733902678,-0.2878811359,-0.0630249605,-0.0454298817,0.4191361964,0.0468793772,0.0098545076,0.1410066783,-0.2734613419,0.1858308762,0.581612587,-0.1513639539,0.5947214365,-0.1334569752,0.1148177981,-0.0854144394,0.3587773144,-0.0369499028,-0.0074621919,0.0684220567,-0.3294129372,-0.2848255932,0.2201754302,-0.1641698778,-0.1078500599,0.4152380824,-0.2399584353,-0.0069708475,0.0520168766,0.2279072404,0.0070127277,0.2049752921,0.0375783257,0.2838076353,0.1509691626,0.1200970411,-0.0707179084,0.0218297038,0.311787039,0.0792425647,0.0439490229,-0.1809312552,0.4068748057,0.0214450192,-0.0643968359,-0.1230903268,-0.0460349992,0.0633512586,-0.0017345719,0.312379986,0.1519995332,0.2424383312,0.0549517125,-0.678863883,-0.2698859274,0.1283806413,-0.1238373369,-0.2885349989,-0.3096182644,0.0676742047,-0.336779803,-0.1962144375,-0.6254069805,0.2873277366,-0.0409296826,0.0317674614,-0.1228295043,0.1495005339,-0.3847355843,-0.3159891665,0.0342093892,0.4128384888,-0.6779385805,-0.0911154747,0.3226251006,-0.0568094924,-0.1318160444,0.365331918,-0.2939430773,0.2710194588,-0.087153703,0.1742832512,0.2444878668,-0.3256330788,-0.1440685838,0.6049585938,0.0289545227,0.0433966629,0.1686531901,0.0429909714,0.2124140263,-0.1310293972,0.4256637692,0.2614240944,-0.166025117,0.2054008245,0.0146942204,-0.3101083636,0.2307115644,0.1687070429,-0.4174053967,0.0358448066,0.0014335349,0.060240455,0.0878544971,-0.1614346355,0.0399083234,0.2464359701,0.225405708,-0.1384202242,-0.0148088466,0.1598281562,-0.8101267219,0.0894191191,0.0783330649,-0.1664144099,-0.1812603027,-0.251011014,-0.3407265544,-0.0264951326,-0.4412739277,0.195633024,0.0154511798,-0.025784865,0.0986443907,0.0230900571,-0.325723201,-0.1233536974,0.0486966036,0.1169337481,-0.5369209647,0.2333204448,0.194428429,0.0930254459,0.0751229674,0.137484923,0.0864106044,-0.2928431928,0.0864409879,0.2784948051,0.1890009195,-0.046988111,0.1903012097,0.2496979982,0.3467512131,0.1710522473,-0.0518033206,-0.0379513018,0.2165665478,-0.1478450447,-0.5269379616,0.5192542672,-0.0552628636,0.264398694,-0.3408648968,0.0983269289,0.4080820084,0.1785749793,-0.1861749738,-0.3059126735,-0.1800610721,0.1127015203,0.1632190645,0.0957221463,-0.4921213388,0.1594941616,0.5060635209,-0.1285547912,0.1044536382,0.1386920065,-0.056483753,0.1158549339,0.066189155,0.226153031,0.2699397504,0.1013890803,-0.2764717638,-0.0098673627,0.2332642823,-0.1635611057,-0.0168333631,0.0024980288,0.1029372513,0.1562787741,0.1946595758,-0.3284356296,-0.2016680688,0.1106605232,0.0623644702,0.2869332135,-0.2881720662,0.1150691882,-0.4400781691,-0.08745832,-0.2905260921,-0.254399389,-0.1714123935,-0.2230552286,0.0624001175,0.1358801574,-0.2535161078,-0.0218591988,0.2236781567,0.2010428458,-0.1078619212,-0.3723759055,-0.034044724,-0.1794715226,-0.1971369982,0.0057908869,0.441494137,-0.0181756597,0.2125764191,-0.3895281255,-0.1640284508,-0.2532625496,-0.0148642045,0.2355837077,-0.208352074,0.2300704122,0.2730090618,0.5455176234,-0.2114352882,-0.119498305,0.3884008229,0.3597330749,-0.2754244506,-0.0018676311,0.1320940703,0.0273656473,-0.0405271575,-0.332485944,-0.1743959039,-0.1648930311,0.546626687,0.0953944102,0.2684221268,-0.004737678,0.0919450074,0.0448512472,-0.1535470337,-0.1258562952,-0.1687311828,-0.4480712414,0.3488395512,-0.2413919121,-0.4294018447,0.1270628273,0.0357438773,-0.1360818893,-0.1900430918,-0.2047750354,-0.1808870435,0.043838162,0.143874377,0.0893156677,0.0208295994,0.3039059937,0.0842272341,0.0914083272,-0.1388596594,-0.2913130522,0.3426032662,0.261149466,0.1384408772,0.4060223699,0.6066058278,-0.0978474393,0.4318779707,0.2884668708,-0.0937830061,0.1609048545,-0.2751137614,0.1728776544,-0.0827634782,-0.0400857329,-0.1221234724,0.0663218275,-0.037628673,0.1682564914,0.1545103788,0.1476061195,-0.1952261925,0.1665136367,-0.4270516336,-0.2269026041,0.1370889544,-0.0021487572,-0.113665022,-0.1738328189,0.0098373583,0.0854336843,-0.0183087774,-0.1246229857,0.3706202209,0.167834267,0.1359750032,-0.5260899067,-0.2937138379,-0.5092406273,0.2776504755,0.0881556123,0.4594088793,-0.1146253124,0.1679283231,0.0838297606,0.0405243114,0.640681684,-0.1144799665,0.0357197002,-0.0914666504,-0.0944429934,-0.41669029,0.2163359672,0.0516773947,0.3282266855,-0.2141707838,0.5839707255,-0.5648413897,-0.4015636742,-0.0710465237,0.3303088844,-0.269297272,-0.2612278163,-0.1670202464,-0.3323805332,-0.3227077425,-0.2339201719,0.1251894832,0.1047254577,0.4464989603,-0.1163482815,0.0183416046,0.0449933559,0.0618399903,0.0449503958,0.2252071649,0.321330905,0.1240455061,0.3207392395,0.227886349,0.3861667216,0.6526228786,0.2381009459,-0.2334096283,-0.0148193659,-0.2096652687,0.3591895103,0.0716511309,0.0603582636,-0.0366692916,0.2573589087,-0.1422799975,-0.2241446525,-0.0346381217,0.2939943075,-0.2217294723,-0.3120138049,-0.4215351641,0.550645411,0.0819648951,0.0327649899,0.0181380901,0.438739568,-0.1620524973,0.3633080125,0.0149005279,0.5603979826,-0.3232106566,0.1063621566,0.2490279078,-0.4401795864,0.5534979105,-0.2797751725,-0.1974375099,-0.3419859707,0.3856818974,-0.1243120804,0.0222455841,0.3222679794,0.4229176939,0.0421257354,0.3495541811,-0.1405308247,0.0872501954,-0.2799370289,0.2551692724,0.0733219981,-0.2849473059,-0.4113448262,0.048883494,-0.284309417,-0.0435148776,0.0196592677,-0.188136518,-0.2444109321,-0.2672119141,-0.1139871031,-0.3135154545,0.1463708729,-0.1654650569,0.4151217043,-0.342243731,0.1641557962,0.2139154822,0.0958294272,0.0627259389,-0.284589231,-0.2553290427,-0.4199611545,0.1910274774,-0.1573657542,-0.0031786053,0.0675688088,0.0563662425,-0.1259816587,-0.1581781358,-0.1744192094,-0.147257179,-0.0822729021,0.0698694512,-0.2488603294,-0.3981580734,-0.4769451916,-0.0215016063,0.1572747976,0.0162103679,0.0623093471,0.1105716452,0.0930569768,0.013990121,0.1181188524,-0.0135530187,-0.0145945419,0.239553988,-0.106237188,-0.0447768085,0.2771878541,0.135288775,-0.0890241489,-0.0753622949,0.4444221854,0.4881884456,-0.0249143895,-0.0751604885,0.0390469693,-0.0287949964,-0.2141461819,0.426104039,0.0013095307,-0.1385851353,0.3630593419,-0.3883480132,-0.2910981178,0.3310472369,0.1526870877,0.0154981194,0.241190657,0.1343667656,-0.219185859,-0.1378929615,-0.2043963671,-0.1107109487,-0.1616651416,0.2030669749,0.0781574771,0.2802634537,0.189112708,-0.2581409514,0.0605769232,-0.1548637748,-0.0973527879,-0.1357849538,-0.1014540642,0.1591212749,0.2880918086,-0.159547314,-0.0996565148,-0.1496339887,-0.2277366221,-0.1170319021,0.0862470344,0.147896558,-0.036588084,0.318341285,0.0030590843,-0.0298570804,0.174221009,-0.0531225577,0.1787035167,0.2418582439,-0.0456478409,-0.0105849942,-0.1219931766,-0.1332558393,-0.0937801376,0.2638160586,0.0208998341,0.1135772243,0.1504711509,-0.4002380967,0.2392122895,0.5172943473,0.4587652385,0.3202010691,-0.22708942,0.1895919144,0.2228290588,0.0881423131,0.0277133193,-0.3655210435,0.4620092213,0.1475369334,-0.0479373559,0.3617207408,0.1514883935,0.0293574464,-0.2263712883,-0.2468200475,0.2360581458,-0.0020853581,-0.1432703733,0.2252085805,-0.10429281,-0.1289491206,0.0973900408,0.5097773671,0.2750142217,0.1287380755,0.0740316957,-0.0416187644,0.322987318,0.1246961951,0.1121880263,-0.5670818686,0.4407199919,0.0491373576,0.1332109272,-0.1298980266,0.389261961,0.3787291646,0.0270041674,-0.0782783777,0.0964568704,0.1244344488,0.0028751928,-0.2567520738,-0.3306165636,-0.091974996,-0.0198532492,-0.1177545413,-0.2587150633,-0.4576646686,-0.0438329913,-0.0662464127,0.1974647194,-0.3758418858,0.3885214925,0.0168707799,0.1566326171,-0.2038819939,0.1276708096,0.1544592828,-0.1893723309,0.2457656115,0.1402352601,0.0057493583,0.060281381,0.073764205,0.1214854419,-0.6196122766,-0.0415551774,-0.0259144474,0.0537305735,-0.2193846703,-0.2029133439,0.2871014476,0.0858973935,0.0252085309,0.1978961676,0.1073758677,0.1039506868,-0.1429748386,-0.2085123956,-0.2601112425,-0.1312231421,0.0672216043,-0.1454973221,-0.1673840731,-0.2738924325,0.2244312912,0.4102648497,0.3850536644,-0.0022571885,0.0498230755,-0.1614064872,0.3891780972,0.1098943427,-0.0231985208,0.0131301647,0.0589828864,-0.2698659599,-0.1748870313,0.0328287818,-0.1842172891,0.3299356699,0.0620039888,0.3498218656,0.3438363671,-0.2147798091,0.3234391212,-0.2594560087,0.0972567797,-0.0042576739,-0.4084130824,-0.03778309,0.1725681573,0.1015457734,-0.0029407649,0.1888600588,0.1150394082,-0.0378741063,-0.365365088,0.1548347175,-0.376906395,-0.3396052718,-0.1647881418,0.203949973,0.1296923161,0.0894177258,0.0538284443,0.167042464,-0.0350980572,-0.2098595798,0.4702451825,-0.267716676,0.4096673429,-0.147983104,0.1108629778,-0.2158671468,-0.0744786486,-0.2486751676,-0.1243249774,-0.2869073153,0.3227340281,-0.25279966,-0.1163362116,0.0139288167,0.3475815058,-0.1150641069,0.3030520082,-0.0177892763,-0.0474525429,0.4104952812,-0.2142120898,-0.0504966266,0.5103651285,-0.0688811541,0.1381971985,-0.3734433055,-0.2129079551,0.1266828328,0.2149323374,-0.1662498266,-0.0317089558,-0.0992148966,-0.1648225337,-0.1734967083,-0.0215530824,0.2306793332,-0.0025103267,-0.0075110225,-0.0919171497,-0.2163067907]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/650","title":"dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`","comments":"Hi :) \r\nIn your dummy data zip file you can just have `subset000.xz` as directories instead of compressed files.\r\nLet me know if it helps","body":"Hi, I recently want to add a dataset whose source data is like this\r\n```\r\nopenwebtext.tar.xz\r\n |__ openwebtext\r\n |__subset000.xz\r\n | |__ ....txt\r\n | |__ ....txt\r\n | ...\r\n |__ subset001.xz\r\n |\r\n ....\r\n```\r\nSo I wrote `openwebtext.py` like this\r\n```\r\n def _split_generators(self, dl_manager):\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n owt_dir = os.path.join(dl_dir, 'openwebtext')\r\n subset_xzs = [\r\n os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock\r\n ]\r\n ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75))\r\n nested_txt_files = [ \r\n [ \r\n os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt')\r\n ] for ex_dir in ex_dirs\r\n ]\r\n txt_files = chain(*nested_txt_files)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN, gen_kwargs={\"txt_files\": txt_files}\r\n ),\r\n ]\r\n```\r\nAll went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me.\r\n\r\nHow should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?","comment_length":25,"text":"dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`\nHi, I recently want to add a dataset whose source data is like this\r\n```\r\nopenwebtext.tar.xz\r\n |__ openwebtext\r\n |__subset000.xz\r\n | |__ ....txt\r\n | |__ ....txt\r\n | ...\r\n |__ subset001.xz\r\n |\r\n ....\r\n```\r\nSo I wrote `openwebtext.py` like this\r\n```\r\n def _split_generators(self, dl_manager):\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n owt_dir = os.path.join(dl_dir, 'openwebtext')\r\n subset_xzs = [\r\n os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock\r\n ]\r\n ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75))\r\n nested_txt_files = [ \r\n [ \r\n os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt')\r\n ] for ex_dir in ex_dirs\r\n ]\r\n txt_files = chain(*nested_txt_files)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN, gen_kwargs={\"txt_files\": txt_files}\r\n ),\r\n ]\r\n```\r\nAll went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me.\r\n\r\nHow should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?\nHi :) \r\nIn your dummy data zip file you can just have `subset000.xz` as directories instead of compressed files.\r\nLet me know if it helps","embeddings":[-0.2775548995,0.0534974262,0.0134969996,0.3913993239,-0.072360985,0.1356922984,0.4522341788,0.3349655271,-0.0165996812,0.0843322054,-0.0575094037,0.2236393094,-0.1632839441,-0.0482924581,0.0641588718,-0.1906499267,-0.1669515967,0.2437199652,0.0134650618,0.0311771519,-0.2944734991,0.0815392062,-0.2834986746,-0.0913858116,-0.1371546984,0.0181267131,-0.2095643431,0.307261467,-0.2457742095,-0.2825932205,0.0010185741,0.0183561444,0.0423617736,0.5455899835,-0.0001227712,0.1388189197,0.3071190715,-0.2575384676,-0.3333342373,-0.0337697677,-0.2507253587,-0.2779996395,-0.1756441742,0.021302674,-0.0840833411,-0.2268642932,-0.1854513139,-0.2622463405,0.3548209369,0.2717152238,0.0923322141,0.61311692,-0.0415804572,-0.1280228794,0.2287501246,-0.0264097862,-0.1879369766,0.2233155817,-0.0529861674,-0.0236796904,0.0873368829,-0.1019545943,-0.0011850322,0.4950457215,-0.3201796412,0.119338505,0.1507472992,-0.3189719617,-0.0463720262,0.550545156,0.4802380502,-0.1391587257,-0.3204444051,-0.2336205691,-0.2743136585,-0.1307140291,0.0747650787,0.2180380672,-0.2292438745,0.3225835264,-0.0293132477,-0.0288202334,0.1385269314,0.2705383897,-0.2816978097,0.3059706092,-0.123613447,0.0502374358,0.4046254754,0.3228848279,0.2671747506,-0.2975229323,-0.3519619703,-0.0660928413,-0.1584336311,-0.2081910372,-0.3658329248,-0.4488224387,0.1594887227,0.3530174494,-0.0114581212,0.1824113429,0.0564967133,0.153296262,0.0790388882,0.4907517433,0.2176588327,0.2714265287,0.0527708866,0.1392944902,0.00167685,0.04832891,0.0824967474,-0.3559404314,-0.2401300967,0.1764753908,0.321375221,-0.1328305751,-0.3999723792,-0.1878486127,-0.2729802728,-0.2551985681,-0.003033394,0.3480630219,-0.0552738905,0.152787447,0.12764512,0.472751826,-0.049808234,-0.5964107513,-0.1689494103,0.0743230879,-0.2225880921,0.113905549,0.1182279512,0.0219427533,0.4373009205,0.1158564389,-0.1179253533,-0.0122450339,0.6474174857,-0.0120400097,0.465523392,0.148574084,0.1298144609,-0.068881914,0.1100025848,-0.1474854946,-0.0640193745,0.0963684469,0.0561538599,-0.1661674827,0.0018030383,0.048893597,-0.4305842221,0.2151851058,-0.0022326279,0.1078094617,-0.0936745405,-0.0127065079,-0.0310162008,-0.046401605,-0.0932798758,0.1658296287,0.3068811595,0.4303734601,-0.2990069985,0.1886532307,-0.4357553422,0.105780296,0.5174133778,-0.0811935589,-0.0792279467,0.3894143403,-0.281427592,0.1036903337,0.4281542301,-0.1070776582,-0.1339181811,0.3495474756,-0.0856407434,0.2467711121,0.3035255373,-0.3556995988,-0.1740887761,-0.0409197249,-0.4010946453,0.1216159016,-0.0223300084,-0.0752302706,-0.3159443736,-0.1216393262,0.3932132125,0.0853224248,0.0043444955,0.1314850748,-0.0922577605,0.1189817488,0.2170329243,-0.0133684017,0.0638615638,-0.0933699682,-0.3415872753,-0.3053729534,-0.06134158,0.1441963464,-0.2157241702,0.3868087828,-0.3383391201,0.3813674152,0.3305300772,-0.0956297293,-0.5658615828,-0.0990950167,-0.2402652353,-0.3324826658,0.0340706035,0.3017258644,0.1029655412,0.0747674704,-0.0000563727,-0.2724884152,0.3430287838,0.1322463006,-0.3578069806,0.1702773571,0.0935114697,-0.1067070961,-0.0051657679,0.0438547842,-0.1398245096,-0.4835775197,-0.1919870526,0.2014565617,0.226877287,-0.1514827162,-0.1123245284,-0.2281041443,0.016749477,-0.148781836,0.0856785253,0.4087572396,0.05439239,-0.1391613334,0.1577101201,0.151158154,0.3081716597,-0.0488703065,-0.2156967819,0.0614776537,0.1661095172,-0.3463017642,0.214206472,-0.2726497054,0.0913260654,0.2350987345,0.3804805279,-0.0418399572,0.28460145,0.0348784402,0.285027802,-0.0671256632,0.1439881325,-0.0548982769,-0.3179729879,-0.1246314719,0.2577799857,0.2502178848,0.5530324578,-0.0051692496,0.3902249932,0.1067654788,0.1798737496,-0.1399505287,0.3181187212,0.0894173011,-0.1737308651,0.2202003449,-0.296266228,-0.3673281074,-0.1835876405,0.2700137794,0.3687210679,0.3929932714,-0.3055078983,-0.109008193,-0.418829143,0.1396477818,0.0166623853,-0.2076057792,-0.0101963971,-0.1257264167,-0.0310548041,0.0821425617,-0.231161356,0.0737062767,0.1541530341,0.1014865637,-0.0827399418,-0.0176653601,-0.2265048623,-0.0992558524,-0.3669633567,0.0497560315,0.3964762092,0.0628838018,0.1707433015,-0.256339252,-0.3311189711,-0.0444091037,-0.0038230647,0.283393085,0.0153302336,0.3275274932,0.2695204318,0.0554335229,0.3602997065,-0.1851060838,-0.074597761,-0.1926950365,-0.1971707642,0.1344675422,0.1516901702,0.1327989846,-0.2324692756,-0.5946473479,-0.150921464,-0.2586509287,0.0104379002,0.0821240395,-0.0298236683,0.28400442,0.0449707247,0.0342343412,0.221667245,-0.264914155,-0.2609222531,-0.3637113869,0.289039731,-0.23346138,-0.2668030262,0.0168261491,-0.5399339795,0.1551875323,-0.2124397755,-0.0537144877,-0.0245191772,0.1239686012,-0.1419387758,0.1030271351,0.2029168457,0.1878895611,-0.0453676954,0.0392214358,-0.1372805536,0.1033983901,0.2979875803,0.0868111998,0.5367175937,-0.1488341242,0.1738517433,0.0305474233,0.298599273,0.0357530974,0.0153352041,0.2445664853,-0.1797609329,0.292242676,-0.1095676348,-0.2095204443,-0.3042358756,0.0125172604,0.0374833681,0.2866483927,-0.1137525737,-0.0525308512,-0.060764201,0.1878269017,0.0676003024,-0.2674080133,0.0879296511,-0.2774010003,-0.1997950226,0.036231719,0.2855166793,-0.0677638203,-0.2022217214,0.091160126,-0.1012437418,0.1115730703,-0.2520765364,-0.1802863032,0.3822595775,-0.5033355355,0.4058063924,0.0177056361,0.286085397,0.3129207492,-0.2308182418,0.0782692283,0.0477271602,0.4159892201,-0.3579122126,0.1582535654,0.4472882748,-0.2505179942,-0.0422279611,-0.1235632375,-0.2599175274,-0.1890912652,0.0263638161,-0.3584081829,-0.367305398,-0.234705776,0.1790145636,0.1722272038,-0.1315891147,-0.0946047753,-0.1954205185,-0.1408470273,-0.173950091,-0.257831037,-0.0831954926,0.05112423,-0.2538347542,-0.1126225069,-0.0213235579,-0.2422351837,0.3344983757,0.3226644099,0.1653524339,-0.0121187614,0.0048921495,0.2595237792,0.032112211,-0.0002837362,0.5156345963,-0.1827042401,-0.2526942492,0.1532875896,-0.1646174192,0.3801353276,0.1221781224,-0.0386986807,0.0844475478,-0.4061064422,0.0917846486,-0.3141177595,0.5976012349,0.3187380135,-0.1618884355,0.1839360744,-0.2865276337,0.4400342107,-0.1198556721,-0.0586536154,0.5181100368,-0.4390115738,-0.2725733221,0.2292658538,-0.1000218242,0.6527184844,0.1551799476,-0.0103552602,0.2116364539,0.2181542963,0.1894761324,-0.2540006638,0.0620612241,-0.2933145463,-0.2902559936,0.1150470451,-0.1187709197,0.2748994231,0.3439331353,-0.371096611,0.2314968705,-0.1662519872,0.1161675975,0.2451562285,0.467181772,-0.2161771506,0.0160147771,-0.0778770745,0.0520516336,0.3815431297,0.2001960129,-0.0275135282,-0.1653925627,-0.2098923177,-0.1588151157,0.0254393648,0.2343832701,-0.4178391695,-0.0726502389,-0.0403762981,-0.2204381078,0.1376759559,0.1729962379,0.2418773621,0.0537331626,-0.1271219254,-0.1456125379,-0.2578451633,-0.0458920449,0.194044292,0.067223452,0.1478221864,0.0373468921,-0.374296248,0.2848320901,-0.1732952595,-0.2097570896,0.3150957227,-0.1556826532,0.1648638994,-0.3636010289,0.0174216125,-0.0640936717,-0.136324659,-0.078526333,0.0533220693,0.0521458909,-0.7080106735,0.1413471997,-0.2713920176,-0.3244532049,-0.0393841602,0.1066874638,-0.2228568494,0.4759735167,0.4173581004,-0.1163219512,0.3777332306,-0.2346686572,0.3289363086,0.3370268047,-0.1697399318,0.6116288304,-0.0555214994,-0.3803268671,0.1090204269,0.3577328324,0.2878513932,-0.0637150183,-0.2488834113,-0.1406081617,-0.1500300318,-0.1582457572,0.1125159264,0.3955524266,-0.1776655316,0.295866847,-0.1747024357,0.3727815747,-0.1949801147,-0.0024365336,-0.1881541312,0.1233800352,-0.0857628807,-0.182947129,0.2429677695,-0.0049513602,0.0499992259,-0.0530919172,-0.2602922916,-0.1787487417,-0.294421643,0.1744043082,0.0657970309,-0.238659665,0.1459914297,0.0540410914,-0.0258019716,-0.1207975522,0.1762764901,-0.3385201991,-0.0109814387,0.2525725663,0.3427561522,-0.0067282249,-0.6286964417,0.0099321706,-0.2735544443,0.2913064659,-0.219351992,0.0491062328,0.1606857926,0.032897003,0.1278441101,-0.2077875882,0.1320329756,-0.1334701329,0.5864819288,0.1192625463,0.0063474188,-0.0403591022,0.220347777,0.3805131018,0.0998040661,0.0783208087,0.1349651664,0.0863007009,-0.3046171963,0.0316624343,0.2106906176,0.0489396565,0.3257641494,0.274530679,0.2787441909,0.0171891041,-0.0304751955,-0.1940560639,0.2114888281,0.2699235976,0.1913394183,0.7015006542,-0.4936882555,0.3337839544,0.1052521989,-0.2653430104,0.2542865574,-0.4790070057,0.1929245144,0.6058324575,-0.3864374161,0.0741798952,-0.080145359,-0.211253792,0.2038942575,0.1999296099,0.2975566387,0.2877901793,0.026033625,0.3619494736,-0.5504577756,-0.0771566704,-0.1918422431,0.2061724663,0.2919972241,0.0384994112,-0.0975590497,0.0093709342,-0.4708305895,-0.2300649881,0.2454039603,-0.2955703735,-0.2263980508,-0.0519431494,-0.0112385619,0.3497267962,-0.2697206438,0.0553600155,-0.0817313492,-0.316609472,0.2097853869,0.1095300689,-0.0809019953,0.1963225454,0.4950662553,0.2401908338,0.0226626657,-0.4391456842,-0.3661148846,0.1244544089,0.0449609943,-0.1386365443,0.3397325277,-0.3198176324,-0.2074073851,0.2640452087,0.0414099731,-0.0806637257,-0.439886868,0.2591479123,-0.4032660127,-0.1769732833,0.6636248231,0.1579113901,0.0095406082,0.0294660125,-0.2400483936,-0.0903023556,0.0417932086,0.1923097223,-0.4423007965,0.1097011864,0.0098779434,0.0258049797,-0.2754499018,0.4726877213,-0.0548032969,-0.0301341657,-0.2771562636,-0.1637395322,-0.573884666,0.2634527385,-0.0527890585,-0.4865395427,0.0810311958,-0.1199837923,0.3776105642,0.1762557328,-0.1693374813,0.3486984968,0.2776536942,0.2912560999,-0.3625783324,-0.2438791841,-0.2989762425,0.0307736583,-0.0500393957,-0.4768276513,0.0896086022,-0.0004992259,-0.1357264072,0.1987256855,0.0193388946,0.3844162226,0.0003760862,0.2715222836,0.0879437923,0.4524197876,-0.1005812734,0.1148442626,-0.0384095423,0.2583157718,-0.1768371612,-0.0881766677,0.2198528945,0.2592366934,0.0071110297,0.2904507518,-0.3201829493,0.4042015374,0.3185971081,0.3075398207,-0.556835115,0.1301245689,-0.2209291756,-0.0340216942,0.6548748612,0.2844884694,-0.3331443667,0.117379941,-0.0490211435,0.0082143201,0.3080021441,-0.1347652823,-0.1568375677,-0.2240474224,0.0681478232,-0.0966738537,-0.1177390516,-0.145793587,-0.1565946043,0.1873218864,-0.0510618985,0.1215432063,0.1973781735,0.514441371,-0.0037093125,0.0087381527,0.287556529,0.2362604737,-0.0319847018,0.0270491075,-0.1153386012]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/650","title":"dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`","comments":"Thanks for your comment @lhoestq ,\r\nJust for confirmation, changing dummy data like this won't make dummy test test the functionality to extract `subsetxxx.xz` but actually kind of circumvent it. But since we will test the real data so it is ok ?","body":"Hi, I recently want to add a dataset whose source data is like this\r\n```\r\nopenwebtext.tar.xz\r\n |__ openwebtext\r\n |__subset000.xz\r\n | |__ ....txt\r\n | |__ ....txt\r\n | ...\r\n |__ subset001.xz\r\n |\r\n ....\r\n```\r\nSo I wrote `openwebtext.py` like this\r\n```\r\n def _split_generators(self, dl_manager):\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n owt_dir = os.path.join(dl_dir, 'openwebtext')\r\n subset_xzs = [\r\n os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock\r\n ]\r\n ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75))\r\n nested_txt_files = [ \r\n [ \r\n os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt')\r\n ] for ex_dir in ex_dirs\r\n ]\r\n txt_files = chain(*nested_txt_files)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN, gen_kwargs={\"txt_files\": txt_files}\r\n ),\r\n ]\r\n```\r\nAll went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me.\r\n\r\nHow should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?","comment_length":43,"text":"dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`\nHi, I recently want to add a dataset whose source data is like this\r\n```\r\nopenwebtext.tar.xz\r\n |__ openwebtext\r\n |__subset000.xz\r\n | |__ ....txt\r\n | |__ ....txt\r\n | ...\r\n |__ subset001.xz\r\n |\r\n ....\r\n```\r\nSo I wrote `openwebtext.py` like this\r\n```\r\n def _split_generators(self, dl_manager):\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n owt_dir = os.path.join(dl_dir, 'openwebtext')\r\n subset_xzs = [\r\n os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock\r\n ]\r\n ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75))\r\n nested_txt_files = [ \r\n [ \r\n os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt')\r\n ] for ex_dir in ex_dirs\r\n ]\r\n txt_files = chain(*nested_txt_files)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN, gen_kwargs={\"txt_files\": txt_files}\r\n ),\r\n ]\r\n```\r\nAll went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me.\r\n\r\nHow should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?\nThanks for your comment @lhoestq ,\r\nJust for confirmation, changing dummy data like this won't make dummy test test the functionality to extract `subsetxxx.xz` but actually kind of circumvent it. But since we will test the real data so it is ok ?","embeddings":[-0.2775548995,0.0534974262,0.0134969996,0.3913993239,-0.072360985,0.1356922984,0.4522341788,0.3349655271,-0.0165996812,0.0843322054,-0.0575094037,0.2236393094,-0.1632839441,-0.0482924581,0.0641588718,-0.1906499267,-0.1669515967,0.2437199652,0.0134650618,0.0311771519,-0.2944734991,0.0815392062,-0.2834986746,-0.0913858116,-0.1371546984,0.0181267131,-0.2095643431,0.307261467,-0.2457742095,-0.2825932205,0.0010185741,0.0183561444,0.0423617736,0.5455899835,-0.0001227712,0.1388189197,0.3071190715,-0.2575384676,-0.3333342373,-0.0337697677,-0.2507253587,-0.2779996395,-0.1756441742,0.021302674,-0.0840833411,-0.2268642932,-0.1854513139,-0.2622463405,0.3548209369,0.2717152238,0.0923322141,0.61311692,-0.0415804572,-0.1280228794,0.2287501246,-0.0264097862,-0.1879369766,0.2233155817,-0.0529861674,-0.0236796904,0.0873368829,-0.1019545943,-0.0011850322,0.4950457215,-0.3201796412,0.119338505,0.1507472992,-0.3189719617,-0.0463720262,0.550545156,0.4802380502,-0.1391587257,-0.3204444051,-0.2336205691,-0.2743136585,-0.1307140291,0.0747650787,0.2180380672,-0.2292438745,0.3225835264,-0.0293132477,-0.0288202334,0.1385269314,0.2705383897,-0.2816978097,0.3059706092,-0.123613447,0.0502374358,0.4046254754,0.3228848279,0.2671747506,-0.2975229323,-0.3519619703,-0.0660928413,-0.1584336311,-0.2081910372,-0.3658329248,-0.4488224387,0.1594887227,0.3530174494,-0.0114581212,0.1824113429,0.0564967133,0.153296262,0.0790388882,0.4907517433,0.2176588327,0.2714265287,0.0527708866,0.1392944902,0.00167685,0.04832891,0.0824967474,-0.3559404314,-0.2401300967,0.1764753908,0.321375221,-0.1328305751,-0.3999723792,-0.1878486127,-0.2729802728,-0.2551985681,-0.003033394,0.3480630219,-0.0552738905,0.152787447,0.12764512,0.472751826,-0.049808234,-0.5964107513,-0.1689494103,0.0743230879,-0.2225880921,0.113905549,0.1182279512,0.0219427533,0.4373009205,0.1158564389,-0.1179253533,-0.0122450339,0.6474174857,-0.0120400097,0.465523392,0.148574084,0.1298144609,-0.068881914,0.1100025848,-0.1474854946,-0.0640193745,0.0963684469,0.0561538599,-0.1661674827,0.0018030383,0.048893597,-0.4305842221,0.2151851058,-0.0022326279,0.1078094617,-0.0936745405,-0.0127065079,-0.0310162008,-0.046401605,-0.0932798758,0.1658296287,0.3068811595,0.4303734601,-0.2990069985,0.1886532307,-0.4357553422,0.105780296,0.5174133778,-0.0811935589,-0.0792279467,0.3894143403,-0.281427592,0.1036903337,0.4281542301,-0.1070776582,-0.1339181811,0.3495474756,-0.0856407434,0.2467711121,0.3035255373,-0.3556995988,-0.1740887761,-0.0409197249,-0.4010946453,0.1216159016,-0.0223300084,-0.0752302706,-0.3159443736,-0.1216393262,0.3932132125,0.0853224248,0.0043444955,0.1314850748,-0.0922577605,0.1189817488,0.2170329243,-0.0133684017,0.0638615638,-0.0933699682,-0.3415872753,-0.3053729534,-0.06134158,0.1441963464,-0.2157241702,0.3868087828,-0.3383391201,0.3813674152,0.3305300772,-0.0956297293,-0.5658615828,-0.0990950167,-0.2402652353,-0.3324826658,0.0340706035,0.3017258644,0.1029655412,0.0747674704,-0.0000563727,-0.2724884152,0.3430287838,0.1322463006,-0.3578069806,0.1702773571,0.0935114697,-0.1067070961,-0.0051657679,0.0438547842,-0.1398245096,-0.4835775197,-0.1919870526,0.2014565617,0.226877287,-0.1514827162,-0.1123245284,-0.2281041443,0.016749477,-0.148781836,0.0856785253,0.4087572396,0.05439239,-0.1391613334,0.1577101201,0.151158154,0.3081716597,-0.0488703065,-0.2156967819,0.0614776537,0.1661095172,-0.3463017642,0.214206472,-0.2726497054,0.0913260654,0.2350987345,0.3804805279,-0.0418399572,0.28460145,0.0348784402,0.285027802,-0.0671256632,0.1439881325,-0.0548982769,-0.3179729879,-0.1246314719,0.2577799857,0.2502178848,0.5530324578,-0.0051692496,0.3902249932,0.1067654788,0.1798737496,-0.1399505287,0.3181187212,0.0894173011,-0.1737308651,0.2202003449,-0.296266228,-0.3673281074,-0.1835876405,0.2700137794,0.3687210679,0.3929932714,-0.3055078983,-0.109008193,-0.418829143,0.1396477818,0.0166623853,-0.2076057792,-0.0101963971,-0.1257264167,-0.0310548041,0.0821425617,-0.231161356,0.0737062767,0.1541530341,0.1014865637,-0.0827399418,-0.0176653601,-0.2265048623,-0.0992558524,-0.3669633567,0.0497560315,0.3964762092,0.0628838018,0.1707433015,-0.256339252,-0.3311189711,-0.0444091037,-0.0038230647,0.283393085,0.0153302336,0.3275274932,0.2695204318,0.0554335229,0.3602997065,-0.1851060838,-0.074597761,-0.1926950365,-0.1971707642,0.1344675422,0.1516901702,0.1327989846,-0.2324692756,-0.5946473479,-0.150921464,-0.2586509287,0.0104379002,0.0821240395,-0.0298236683,0.28400442,0.0449707247,0.0342343412,0.221667245,-0.264914155,-0.2609222531,-0.3637113869,0.289039731,-0.23346138,-0.2668030262,0.0168261491,-0.5399339795,0.1551875323,-0.2124397755,-0.0537144877,-0.0245191772,0.1239686012,-0.1419387758,0.1030271351,0.2029168457,0.1878895611,-0.0453676954,0.0392214358,-0.1372805536,0.1033983901,0.2979875803,0.0868111998,0.5367175937,-0.1488341242,0.1738517433,0.0305474233,0.298599273,0.0357530974,0.0153352041,0.2445664853,-0.1797609329,0.292242676,-0.1095676348,-0.2095204443,-0.3042358756,0.0125172604,0.0374833681,0.2866483927,-0.1137525737,-0.0525308512,-0.060764201,0.1878269017,0.0676003024,-0.2674080133,0.0879296511,-0.2774010003,-0.1997950226,0.036231719,0.2855166793,-0.0677638203,-0.2022217214,0.091160126,-0.1012437418,0.1115730703,-0.2520765364,-0.1802863032,0.3822595775,-0.5033355355,0.4058063924,0.0177056361,0.286085397,0.3129207492,-0.2308182418,0.0782692283,0.0477271602,0.4159892201,-0.3579122126,0.1582535654,0.4472882748,-0.2505179942,-0.0422279611,-0.1235632375,-0.2599175274,-0.1890912652,0.0263638161,-0.3584081829,-0.367305398,-0.234705776,0.1790145636,0.1722272038,-0.1315891147,-0.0946047753,-0.1954205185,-0.1408470273,-0.173950091,-0.257831037,-0.0831954926,0.05112423,-0.2538347542,-0.1126225069,-0.0213235579,-0.2422351837,0.3344983757,0.3226644099,0.1653524339,-0.0121187614,0.0048921495,0.2595237792,0.032112211,-0.0002837362,0.5156345963,-0.1827042401,-0.2526942492,0.1532875896,-0.1646174192,0.3801353276,0.1221781224,-0.0386986807,0.0844475478,-0.4061064422,0.0917846486,-0.3141177595,0.5976012349,0.3187380135,-0.1618884355,0.1839360744,-0.2865276337,0.4400342107,-0.1198556721,-0.0586536154,0.5181100368,-0.4390115738,-0.2725733221,0.2292658538,-0.1000218242,0.6527184844,0.1551799476,-0.0103552602,0.2116364539,0.2181542963,0.1894761324,-0.2540006638,0.0620612241,-0.2933145463,-0.2902559936,0.1150470451,-0.1187709197,0.2748994231,0.3439331353,-0.371096611,0.2314968705,-0.1662519872,0.1161675975,0.2451562285,0.467181772,-0.2161771506,0.0160147771,-0.0778770745,0.0520516336,0.3815431297,0.2001960129,-0.0275135282,-0.1653925627,-0.2098923177,-0.1588151157,0.0254393648,0.2343832701,-0.4178391695,-0.0726502389,-0.0403762981,-0.2204381078,0.1376759559,0.1729962379,0.2418773621,0.0537331626,-0.1271219254,-0.1456125379,-0.2578451633,-0.0458920449,0.194044292,0.067223452,0.1478221864,0.0373468921,-0.374296248,0.2848320901,-0.1732952595,-0.2097570896,0.3150957227,-0.1556826532,0.1648638994,-0.3636010289,0.0174216125,-0.0640936717,-0.136324659,-0.078526333,0.0533220693,0.0521458909,-0.7080106735,0.1413471997,-0.2713920176,-0.3244532049,-0.0393841602,0.1066874638,-0.2228568494,0.4759735167,0.4173581004,-0.1163219512,0.3777332306,-0.2346686572,0.3289363086,0.3370268047,-0.1697399318,0.6116288304,-0.0555214994,-0.3803268671,0.1090204269,0.3577328324,0.2878513932,-0.0637150183,-0.2488834113,-0.1406081617,-0.1500300318,-0.1582457572,0.1125159264,0.3955524266,-0.1776655316,0.295866847,-0.1747024357,0.3727815747,-0.1949801147,-0.0024365336,-0.1881541312,0.1233800352,-0.0857628807,-0.182947129,0.2429677695,-0.0049513602,0.0499992259,-0.0530919172,-0.2602922916,-0.1787487417,-0.294421643,0.1744043082,0.0657970309,-0.238659665,0.1459914297,0.0540410914,-0.0258019716,-0.1207975522,0.1762764901,-0.3385201991,-0.0109814387,0.2525725663,0.3427561522,-0.0067282249,-0.6286964417,0.0099321706,-0.2735544443,0.2913064659,-0.219351992,0.0491062328,0.1606857926,0.032897003,0.1278441101,-0.2077875882,0.1320329756,-0.1334701329,0.5864819288,0.1192625463,0.0063474188,-0.0403591022,0.220347777,0.3805131018,0.0998040661,0.0783208087,0.1349651664,0.0863007009,-0.3046171963,0.0316624343,0.2106906176,0.0489396565,0.3257641494,0.274530679,0.2787441909,0.0171891041,-0.0304751955,-0.1940560639,0.2114888281,0.2699235976,0.1913394183,0.7015006542,-0.4936882555,0.3337839544,0.1052521989,-0.2653430104,0.2542865574,-0.4790070057,0.1929245144,0.6058324575,-0.3864374161,0.0741798952,-0.080145359,-0.211253792,0.2038942575,0.1999296099,0.2975566387,0.2877901793,0.026033625,0.3619494736,-0.5504577756,-0.0771566704,-0.1918422431,0.2061724663,0.2919972241,0.0384994112,-0.0975590497,0.0093709342,-0.4708305895,-0.2300649881,0.2454039603,-0.2955703735,-0.2263980508,-0.0519431494,-0.0112385619,0.3497267962,-0.2697206438,0.0553600155,-0.0817313492,-0.316609472,0.2097853869,0.1095300689,-0.0809019953,0.1963225454,0.4950662553,0.2401908338,0.0226626657,-0.4391456842,-0.3661148846,0.1244544089,0.0449609943,-0.1386365443,0.3397325277,-0.3198176324,-0.2074073851,0.2640452087,0.0414099731,-0.0806637257,-0.439886868,0.2591479123,-0.4032660127,-0.1769732833,0.6636248231,0.1579113901,0.0095406082,0.0294660125,-0.2400483936,-0.0903023556,0.0417932086,0.1923097223,-0.4423007965,0.1097011864,0.0098779434,0.0258049797,-0.2754499018,0.4726877213,-0.0548032969,-0.0301341657,-0.2771562636,-0.1637395322,-0.573884666,0.2634527385,-0.0527890585,-0.4865395427,0.0810311958,-0.1199837923,0.3776105642,0.1762557328,-0.1693374813,0.3486984968,0.2776536942,0.2912560999,-0.3625783324,-0.2438791841,-0.2989762425,0.0307736583,-0.0500393957,-0.4768276513,0.0896086022,-0.0004992259,-0.1357264072,0.1987256855,0.0193388946,0.3844162226,0.0003760862,0.2715222836,0.0879437923,0.4524197876,-0.1005812734,0.1148442626,-0.0384095423,0.2583157718,-0.1768371612,-0.0881766677,0.2198528945,0.2592366934,0.0071110297,0.2904507518,-0.3201829493,0.4042015374,0.3185971081,0.3075398207,-0.556835115,0.1301245689,-0.2209291756,-0.0340216942,0.6548748612,0.2844884694,-0.3331443667,0.117379941,-0.0490211435,0.0082143201,0.3080021441,-0.1347652823,-0.1568375677,-0.2240474224,0.0681478232,-0.0966738537,-0.1177390516,-0.145793587,-0.1565946043,0.1873218864,-0.0510618985,0.1215432063,0.1973781735,0.514441371,-0.0037093125,0.0087381527,0.287556529,0.2362604737,-0.0319847018,0.0270491075,-0.1153386012]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/650","title":"dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`","comments":"Yes it's fine for now. We plan to add a job for slow tests.\r\nAnd at one point we'll also do another pass on the dummy data handling and consider extracting files.","body":"Hi, I recently want to add a dataset whose source data is like this\r\n```\r\nopenwebtext.tar.xz\r\n |__ openwebtext\r\n |__subset000.xz\r\n | |__ ....txt\r\n | |__ ....txt\r\n | ...\r\n |__ subset001.xz\r\n |\r\n ....\r\n```\r\nSo I wrote `openwebtext.py` like this\r\n```\r\n def _split_generators(self, dl_manager):\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n owt_dir = os.path.join(dl_dir, 'openwebtext')\r\n subset_xzs = [\r\n os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock\r\n ]\r\n ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75))\r\n nested_txt_files = [ \r\n [ \r\n os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt')\r\n ] for ex_dir in ex_dirs\r\n ]\r\n txt_files = chain(*nested_txt_files)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN, gen_kwargs={\"txt_files\": txt_files}\r\n ),\r\n ]\r\n```\r\nAll went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me.\r\n\r\nHow should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?","comment_length":32,"text":"dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`\nHi, I recently want to add a dataset whose source data is like this\r\n```\r\nopenwebtext.tar.xz\r\n |__ openwebtext\r\n |__subset000.xz\r\n | |__ ....txt\r\n | |__ ....txt\r\n | ...\r\n |__ subset001.xz\r\n |\r\n ....\r\n```\r\nSo I wrote `openwebtext.py` like this\r\n```\r\n def _split_generators(self, dl_manager):\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n owt_dir = os.path.join(dl_dir, 'openwebtext')\r\n subset_xzs = [\r\n os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock\r\n ]\r\n ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75))\r\n nested_txt_files = [ \r\n [ \r\n os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt')\r\n ] for ex_dir in ex_dirs\r\n ]\r\n txt_files = chain(*nested_txt_files)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN, gen_kwargs={\"txt_files\": txt_files}\r\n ),\r\n ]\r\n```\r\nAll went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me.\r\n\r\nHow should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?\nYes it's fine for now. We plan to add a job for slow tests.\r\nAnd at one point we'll also do another pass on the dummy data handling and consider extracting files.","embeddings":[-0.2775548995,0.0534974262,0.0134969996,0.3913993239,-0.072360985,0.1356922984,0.4522341788,0.3349655271,-0.0165996812,0.0843322054,-0.0575094037,0.2236393094,-0.1632839441,-0.0482924581,0.0641588718,-0.1906499267,-0.1669515967,0.2437199652,0.0134650618,0.0311771519,-0.2944734991,0.0815392062,-0.2834986746,-0.0913858116,-0.1371546984,0.0181267131,-0.2095643431,0.307261467,-0.2457742095,-0.2825932205,0.0010185741,0.0183561444,0.0423617736,0.5455899835,-0.0001227712,0.1388189197,0.3071190715,-0.2575384676,-0.3333342373,-0.0337697677,-0.2507253587,-0.2779996395,-0.1756441742,0.021302674,-0.0840833411,-0.2268642932,-0.1854513139,-0.2622463405,0.3548209369,0.2717152238,0.0923322141,0.61311692,-0.0415804572,-0.1280228794,0.2287501246,-0.0264097862,-0.1879369766,0.2233155817,-0.0529861674,-0.0236796904,0.0873368829,-0.1019545943,-0.0011850322,0.4950457215,-0.3201796412,0.119338505,0.1507472992,-0.3189719617,-0.0463720262,0.550545156,0.4802380502,-0.1391587257,-0.3204444051,-0.2336205691,-0.2743136585,-0.1307140291,0.0747650787,0.2180380672,-0.2292438745,0.3225835264,-0.0293132477,-0.0288202334,0.1385269314,0.2705383897,-0.2816978097,0.3059706092,-0.123613447,0.0502374358,0.4046254754,0.3228848279,0.2671747506,-0.2975229323,-0.3519619703,-0.0660928413,-0.1584336311,-0.2081910372,-0.3658329248,-0.4488224387,0.1594887227,0.3530174494,-0.0114581212,0.1824113429,0.0564967133,0.153296262,0.0790388882,0.4907517433,0.2176588327,0.2714265287,0.0527708866,0.1392944902,0.00167685,0.04832891,0.0824967474,-0.3559404314,-0.2401300967,0.1764753908,0.321375221,-0.1328305751,-0.3999723792,-0.1878486127,-0.2729802728,-0.2551985681,-0.003033394,0.3480630219,-0.0552738905,0.152787447,0.12764512,0.472751826,-0.049808234,-0.5964107513,-0.1689494103,0.0743230879,-0.2225880921,0.113905549,0.1182279512,0.0219427533,0.4373009205,0.1158564389,-0.1179253533,-0.0122450339,0.6474174857,-0.0120400097,0.465523392,0.148574084,0.1298144609,-0.068881914,0.1100025848,-0.1474854946,-0.0640193745,0.0963684469,0.0561538599,-0.1661674827,0.0018030383,0.048893597,-0.4305842221,0.2151851058,-0.0022326279,0.1078094617,-0.0936745405,-0.0127065079,-0.0310162008,-0.046401605,-0.0932798758,0.1658296287,0.3068811595,0.4303734601,-0.2990069985,0.1886532307,-0.4357553422,0.105780296,0.5174133778,-0.0811935589,-0.0792279467,0.3894143403,-0.281427592,0.1036903337,0.4281542301,-0.1070776582,-0.1339181811,0.3495474756,-0.0856407434,0.2467711121,0.3035255373,-0.3556995988,-0.1740887761,-0.0409197249,-0.4010946453,0.1216159016,-0.0223300084,-0.0752302706,-0.3159443736,-0.1216393262,0.3932132125,0.0853224248,0.0043444955,0.1314850748,-0.0922577605,0.1189817488,0.2170329243,-0.0133684017,0.0638615638,-0.0933699682,-0.3415872753,-0.3053729534,-0.06134158,0.1441963464,-0.2157241702,0.3868087828,-0.3383391201,0.3813674152,0.3305300772,-0.0956297293,-0.5658615828,-0.0990950167,-0.2402652353,-0.3324826658,0.0340706035,0.3017258644,0.1029655412,0.0747674704,-0.0000563727,-0.2724884152,0.3430287838,0.1322463006,-0.3578069806,0.1702773571,0.0935114697,-0.1067070961,-0.0051657679,0.0438547842,-0.1398245096,-0.4835775197,-0.1919870526,0.2014565617,0.226877287,-0.1514827162,-0.1123245284,-0.2281041443,0.016749477,-0.148781836,0.0856785253,0.4087572396,0.05439239,-0.1391613334,0.1577101201,0.151158154,0.3081716597,-0.0488703065,-0.2156967819,0.0614776537,0.1661095172,-0.3463017642,0.214206472,-0.2726497054,0.0913260654,0.2350987345,0.3804805279,-0.0418399572,0.28460145,0.0348784402,0.285027802,-0.0671256632,0.1439881325,-0.0548982769,-0.3179729879,-0.1246314719,0.2577799857,0.2502178848,0.5530324578,-0.0051692496,0.3902249932,0.1067654788,0.1798737496,-0.1399505287,0.3181187212,0.0894173011,-0.1737308651,0.2202003449,-0.296266228,-0.3673281074,-0.1835876405,0.2700137794,0.3687210679,0.3929932714,-0.3055078983,-0.109008193,-0.418829143,0.1396477818,0.0166623853,-0.2076057792,-0.0101963971,-0.1257264167,-0.0310548041,0.0821425617,-0.231161356,0.0737062767,0.1541530341,0.1014865637,-0.0827399418,-0.0176653601,-0.2265048623,-0.0992558524,-0.3669633567,0.0497560315,0.3964762092,0.0628838018,0.1707433015,-0.256339252,-0.3311189711,-0.0444091037,-0.0038230647,0.283393085,0.0153302336,0.3275274932,0.2695204318,0.0554335229,0.3602997065,-0.1851060838,-0.074597761,-0.1926950365,-0.1971707642,0.1344675422,0.1516901702,0.1327989846,-0.2324692756,-0.5946473479,-0.150921464,-0.2586509287,0.0104379002,0.0821240395,-0.0298236683,0.28400442,0.0449707247,0.0342343412,0.221667245,-0.264914155,-0.2609222531,-0.3637113869,0.289039731,-0.23346138,-0.2668030262,0.0168261491,-0.5399339795,0.1551875323,-0.2124397755,-0.0537144877,-0.0245191772,0.1239686012,-0.1419387758,0.1030271351,0.2029168457,0.1878895611,-0.0453676954,0.0392214358,-0.1372805536,0.1033983901,0.2979875803,0.0868111998,0.5367175937,-0.1488341242,0.1738517433,0.0305474233,0.298599273,0.0357530974,0.0153352041,0.2445664853,-0.1797609329,0.292242676,-0.1095676348,-0.2095204443,-0.3042358756,0.0125172604,0.0374833681,0.2866483927,-0.1137525737,-0.0525308512,-0.060764201,0.1878269017,0.0676003024,-0.2674080133,0.0879296511,-0.2774010003,-0.1997950226,0.036231719,0.2855166793,-0.0677638203,-0.2022217214,0.091160126,-0.1012437418,0.1115730703,-0.2520765364,-0.1802863032,0.3822595775,-0.5033355355,0.4058063924,0.0177056361,0.286085397,0.3129207492,-0.2308182418,0.0782692283,0.0477271602,0.4159892201,-0.3579122126,0.1582535654,0.4472882748,-0.2505179942,-0.0422279611,-0.1235632375,-0.2599175274,-0.1890912652,0.0263638161,-0.3584081829,-0.367305398,-0.234705776,0.1790145636,0.1722272038,-0.1315891147,-0.0946047753,-0.1954205185,-0.1408470273,-0.173950091,-0.257831037,-0.0831954926,0.05112423,-0.2538347542,-0.1126225069,-0.0213235579,-0.2422351837,0.3344983757,0.3226644099,0.1653524339,-0.0121187614,0.0048921495,0.2595237792,0.032112211,-0.0002837362,0.5156345963,-0.1827042401,-0.2526942492,0.1532875896,-0.1646174192,0.3801353276,0.1221781224,-0.0386986807,0.0844475478,-0.4061064422,0.0917846486,-0.3141177595,0.5976012349,0.3187380135,-0.1618884355,0.1839360744,-0.2865276337,0.4400342107,-0.1198556721,-0.0586536154,0.5181100368,-0.4390115738,-0.2725733221,0.2292658538,-0.1000218242,0.6527184844,0.1551799476,-0.0103552602,0.2116364539,0.2181542963,0.1894761324,-0.2540006638,0.0620612241,-0.2933145463,-0.2902559936,0.1150470451,-0.1187709197,0.2748994231,0.3439331353,-0.371096611,0.2314968705,-0.1662519872,0.1161675975,0.2451562285,0.467181772,-0.2161771506,0.0160147771,-0.0778770745,0.0520516336,0.3815431297,0.2001960129,-0.0275135282,-0.1653925627,-0.2098923177,-0.1588151157,0.0254393648,0.2343832701,-0.4178391695,-0.0726502389,-0.0403762981,-0.2204381078,0.1376759559,0.1729962379,0.2418773621,0.0537331626,-0.1271219254,-0.1456125379,-0.2578451633,-0.0458920449,0.194044292,0.067223452,0.1478221864,0.0373468921,-0.374296248,0.2848320901,-0.1732952595,-0.2097570896,0.3150957227,-0.1556826532,0.1648638994,-0.3636010289,0.0174216125,-0.0640936717,-0.136324659,-0.078526333,0.0533220693,0.0521458909,-0.7080106735,0.1413471997,-0.2713920176,-0.3244532049,-0.0393841602,0.1066874638,-0.2228568494,0.4759735167,0.4173581004,-0.1163219512,0.3777332306,-0.2346686572,0.3289363086,0.3370268047,-0.1697399318,0.6116288304,-0.0555214994,-0.3803268671,0.1090204269,0.3577328324,0.2878513932,-0.0637150183,-0.2488834113,-0.1406081617,-0.1500300318,-0.1582457572,0.1125159264,0.3955524266,-0.1776655316,0.295866847,-0.1747024357,0.3727815747,-0.1949801147,-0.0024365336,-0.1881541312,0.1233800352,-0.0857628807,-0.182947129,0.2429677695,-0.0049513602,0.0499992259,-0.0530919172,-0.2602922916,-0.1787487417,-0.294421643,0.1744043082,0.0657970309,-0.238659665,0.1459914297,0.0540410914,-0.0258019716,-0.1207975522,0.1762764901,-0.3385201991,-0.0109814387,0.2525725663,0.3427561522,-0.0067282249,-0.6286964417,0.0099321706,-0.2735544443,0.2913064659,-0.219351992,0.0491062328,0.1606857926,0.032897003,0.1278441101,-0.2077875882,0.1320329756,-0.1334701329,0.5864819288,0.1192625463,0.0063474188,-0.0403591022,0.220347777,0.3805131018,0.0998040661,0.0783208087,0.1349651664,0.0863007009,-0.3046171963,0.0316624343,0.2106906176,0.0489396565,0.3257641494,0.274530679,0.2787441909,0.0171891041,-0.0304751955,-0.1940560639,0.2114888281,0.2699235976,0.1913394183,0.7015006542,-0.4936882555,0.3337839544,0.1052521989,-0.2653430104,0.2542865574,-0.4790070057,0.1929245144,0.6058324575,-0.3864374161,0.0741798952,-0.080145359,-0.211253792,0.2038942575,0.1999296099,0.2975566387,0.2877901793,0.026033625,0.3619494736,-0.5504577756,-0.0771566704,-0.1918422431,0.2061724663,0.2919972241,0.0384994112,-0.0975590497,0.0093709342,-0.4708305895,-0.2300649881,0.2454039603,-0.2955703735,-0.2263980508,-0.0519431494,-0.0112385619,0.3497267962,-0.2697206438,0.0553600155,-0.0817313492,-0.316609472,0.2097853869,0.1095300689,-0.0809019953,0.1963225454,0.4950662553,0.2401908338,0.0226626657,-0.4391456842,-0.3661148846,0.1244544089,0.0449609943,-0.1386365443,0.3397325277,-0.3198176324,-0.2074073851,0.2640452087,0.0414099731,-0.0806637257,-0.439886868,0.2591479123,-0.4032660127,-0.1769732833,0.6636248231,0.1579113901,0.0095406082,0.0294660125,-0.2400483936,-0.0903023556,0.0417932086,0.1923097223,-0.4423007965,0.1097011864,0.0098779434,0.0258049797,-0.2754499018,0.4726877213,-0.0548032969,-0.0301341657,-0.2771562636,-0.1637395322,-0.573884666,0.2634527385,-0.0527890585,-0.4865395427,0.0810311958,-0.1199837923,0.3776105642,0.1762557328,-0.1693374813,0.3486984968,0.2776536942,0.2912560999,-0.3625783324,-0.2438791841,-0.2989762425,0.0307736583,-0.0500393957,-0.4768276513,0.0896086022,-0.0004992259,-0.1357264072,0.1987256855,0.0193388946,0.3844162226,0.0003760862,0.2715222836,0.0879437923,0.4524197876,-0.1005812734,0.1148442626,-0.0384095423,0.2583157718,-0.1768371612,-0.0881766677,0.2198528945,0.2592366934,0.0071110297,0.2904507518,-0.3201829493,0.4042015374,0.3185971081,0.3075398207,-0.556835115,0.1301245689,-0.2209291756,-0.0340216942,0.6548748612,0.2844884694,-0.3331443667,0.117379941,-0.0490211435,0.0082143201,0.3080021441,-0.1347652823,-0.1568375677,-0.2240474224,0.0681478232,-0.0966738537,-0.1177390516,-0.145793587,-0.1565946043,0.1873218864,-0.0510618985,0.1215432063,0.1973781735,0.514441371,-0.0037093125,0.0087381527,0.287556529,0.2362604737,-0.0319847018,0.0270491075,-0.1153386012]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/650","title":"dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`","comments":"Thanks for the confirmation.\r\nAlso the suggestion works. Thank you.","body":"Hi, I recently want to add a dataset whose source data is like this\r\n```\r\nopenwebtext.tar.xz\r\n |__ openwebtext\r\n |__subset000.xz\r\n | |__ ....txt\r\n | |__ ....txt\r\n | ...\r\n |__ subset001.xz\r\n |\r\n ....\r\n```\r\nSo I wrote `openwebtext.py` like this\r\n```\r\n def _split_generators(self, dl_manager):\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n owt_dir = os.path.join(dl_dir, 'openwebtext')\r\n subset_xzs = [\r\n os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock\r\n ]\r\n ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75))\r\n nested_txt_files = [ \r\n [ \r\n os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt')\r\n ] for ex_dir in ex_dirs\r\n ]\r\n txt_files = chain(*nested_txt_files)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN, gen_kwargs={\"txt_files\": txt_files}\r\n ),\r\n ]\r\n```\r\nAll went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me.\r\n\r\nHow should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?","comment_length":10,"text":"dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`\nHi, I recently want to add a dataset whose source data is like this\r\n```\r\nopenwebtext.tar.xz\r\n |__ openwebtext\r\n |__subset000.xz\r\n | |__ ....txt\r\n | |__ ....txt\r\n | ...\r\n |__ subset001.xz\r\n |\r\n ....\r\n```\r\nSo I wrote `openwebtext.py` like this\r\n```\r\n def _split_generators(self, dl_manager):\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n owt_dir = os.path.join(dl_dir, 'openwebtext')\r\n subset_xzs = [\r\n os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock\r\n ]\r\n ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75))\r\n nested_txt_files = [ \r\n [ \r\n os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt')\r\n ] for ex_dir in ex_dirs\r\n ]\r\n txt_files = chain(*nested_txt_files)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN, gen_kwargs={\"txt_files\": txt_files}\r\n ),\r\n ]\r\n```\r\nAll went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me.\r\n\r\nHow should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?\nThanks for the confirmation.\r\nAlso the suggestion works. Thank you.","embeddings":[-0.2775548995,0.0534974262,0.0134969996,0.3913993239,-0.072360985,0.1356922984,0.4522341788,0.3349655271,-0.0165996812,0.0843322054,-0.0575094037,0.2236393094,-0.1632839441,-0.0482924581,0.0641588718,-0.1906499267,-0.1669515967,0.2437199652,0.0134650618,0.0311771519,-0.2944734991,0.0815392062,-0.2834986746,-0.0913858116,-0.1371546984,0.0181267131,-0.2095643431,0.307261467,-0.2457742095,-0.2825932205,0.0010185741,0.0183561444,0.0423617736,0.5455899835,-0.0001227712,0.1388189197,0.3071190715,-0.2575384676,-0.3333342373,-0.0337697677,-0.2507253587,-0.2779996395,-0.1756441742,0.021302674,-0.0840833411,-0.2268642932,-0.1854513139,-0.2622463405,0.3548209369,0.2717152238,0.0923322141,0.61311692,-0.0415804572,-0.1280228794,0.2287501246,-0.0264097862,-0.1879369766,0.2233155817,-0.0529861674,-0.0236796904,0.0873368829,-0.1019545943,-0.0011850322,0.4950457215,-0.3201796412,0.119338505,0.1507472992,-0.3189719617,-0.0463720262,0.550545156,0.4802380502,-0.1391587257,-0.3204444051,-0.2336205691,-0.2743136585,-0.1307140291,0.0747650787,0.2180380672,-0.2292438745,0.3225835264,-0.0293132477,-0.0288202334,0.1385269314,0.2705383897,-0.2816978097,0.3059706092,-0.123613447,0.0502374358,0.4046254754,0.3228848279,0.2671747506,-0.2975229323,-0.3519619703,-0.0660928413,-0.1584336311,-0.2081910372,-0.3658329248,-0.4488224387,0.1594887227,0.3530174494,-0.0114581212,0.1824113429,0.0564967133,0.153296262,0.0790388882,0.4907517433,0.2176588327,0.2714265287,0.0527708866,0.1392944902,0.00167685,0.04832891,0.0824967474,-0.3559404314,-0.2401300967,0.1764753908,0.321375221,-0.1328305751,-0.3999723792,-0.1878486127,-0.2729802728,-0.2551985681,-0.003033394,0.3480630219,-0.0552738905,0.152787447,0.12764512,0.472751826,-0.049808234,-0.5964107513,-0.1689494103,0.0743230879,-0.2225880921,0.113905549,0.1182279512,0.0219427533,0.4373009205,0.1158564389,-0.1179253533,-0.0122450339,0.6474174857,-0.0120400097,0.465523392,0.148574084,0.1298144609,-0.068881914,0.1100025848,-0.1474854946,-0.0640193745,0.0963684469,0.0561538599,-0.1661674827,0.0018030383,0.048893597,-0.4305842221,0.2151851058,-0.0022326279,0.1078094617,-0.0936745405,-0.0127065079,-0.0310162008,-0.046401605,-0.0932798758,0.1658296287,0.3068811595,0.4303734601,-0.2990069985,0.1886532307,-0.4357553422,0.105780296,0.5174133778,-0.0811935589,-0.0792279467,0.3894143403,-0.281427592,0.1036903337,0.4281542301,-0.1070776582,-0.1339181811,0.3495474756,-0.0856407434,0.2467711121,0.3035255373,-0.3556995988,-0.1740887761,-0.0409197249,-0.4010946453,0.1216159016,-0.0223300084,-0.0752302706,-0.3159443736,-0.1216393262,0.3932132125,0.0853224248,0.0043444955,0.1314850748,-0.0922577605,0.1189817488,0.2170329243,-0.0133684017,0.0638615638,-0.0933699682,-0.3415872753,-0.3053729534,-0.06134158,0.1441963464,-0.2157241702,0.3868087828,-0.3383391201,0.3813674152,0.3305300772,-0.0956297293,-0.5658615828,-0.0990950167,-0.2402652353,-0.3324826658,0.0340706035,0.3017258644,0.1029655412,0.0747674704,-0.0000563727,-0.2724884152,0.3430287838,0.1322463006,-0.3578069806,0.1702773571,0.0935114697,-0.1067070961,-0.0051657679,0.0438547842,-0.1398245096,-0.4835775197,-0.1919870526,0.2014565617,0.226877287,-0.1514827162,-0.1123245284,-0.2281041443,0.016749477,-0.148781836,0.0856785253,0.4087572396,0.05439239,-0.1391613334,0.1577101201,0.151158154,0.3081716597,-0.0488703065,-0.2156967819,0.0614776537,0.1661095172,-0.3463017642,0.214206472,-0.2726497054,0.0913260654,0.2350987345,0.3804805279,-0.0418399572,0.28460145,0.0348784402,0.285027802,-0.0671256632,0.1439881325,-0.0548982769,-0.3179729879,-0.1246314719,0.2577799857,0.2502178848,0.5530324578,-0.0051692496,0.3902249932,0.1067654788,0.1798737496,-0.1399505287,0.3181187212,0.0894173011,-0.1737308651,0.2202003449,-0.296266228,-0.3673281074,-0.1835876405,0.2700137794,0.3687210679,0.3929932714,-0.3055078983,-0.109008193,-0.418829143,0.1396477818,0.0166623853,-0.2076057792,-0.0101963971,-0.1257264167,-0.0310548041,0.0821425617,-0.231161356,0.0737062767,0.1541530341,0.1014865637,-0.0827399418,-0.0176653601,-0.2265048623,-0.0992558524,-0.3669633567,0.0497560315,0.3964762092,0.0628838018,0.1707433015,-0.256339252,-0.3311189711,-0.0444091037,-0.0038230647,0.283393085,0.0153302336,0.3275274932,0.2695204318,0.0554335229,0.3602997065,-0.1851060838,-0.074597761,-0.1926950365,-0.1971707642,0.1344675422,0.1516901702,0.1327989846,-0.2324692756,-0.5946473479,-0.150921464,-0.2586509287,0.0104379002,0.0821240395,-0.0298236683,0.28400442,0.0449707247,0.0342343412,0.221667245,-0.264914155,-0.2609222531,-0.3637113869,0.289039731,-0.23346138,-0.2668030262,0.0168261491,-0.5399339795,0.1551875323,-0.2124397755,-0.0537144877,-0.0245191772,0.1239686012,-0.1419387758,0.1030271351,0.2029168457,0.1878895611,-0.0453676954,0.0392214358,-0.1372805536,0.1033983901,0.2979875803,0.0868111998,0.5367175937,-0.1488341242,0.1738517433,0.0305474233,0.298599273,0.0357530974,0.0153352041,0.2445664853,-0.1797609329,0.292242676,-0.1095676348,-0.2095204443,-0.3042358756,0.0125172604,0.0374833681,0.2866483927,-0.1137525737,-0.0525308512,-0.060764201,0.1878269017,0.0676003024,-0.2674080133,0.0879296511,-0.2774010003,-0.1997950226,0.036231719,0.2855166793,-0.0677638203,-0.2022217214,0.091160126,-0.1012437418,0.1115730703,-0.2520765364,-0.1802863032,0.3822595775,-0.5033355355,0.4058063924,0.0177056361,0.286085397,0.3129207492,-0.2308182418,0.0782692283,0.0477271602,0.4159892201,-0.3579122126,0.1582535654,0.4472882748,-0.2505179942,-0.0422279611,-0.1235632375,-0.2599175274,-0.1890912652,0.0263638161,-0.3584081829,-0.367305398,-0.234705776,0.1790145636,0.1722272038,-0.1315891147,-0.0946047753,-0.1954205185,-0.1408470273,-0.173950091,-0.257831037,-0.0831954926,0.05112423,-0.2538347542,-0.1126225069,-0.0213235579,-0.2422351837,0.3344983757,0.3226644099,0.1653524339,-0.0121187614,0.0048921495,0.2595237792,0.032112211,-0.0002837362,0.5156345963,-0.1827042401,-0.2526942492,0.1532875896,-0.1646174192,0.3801353276,0.1221781224,-0.0386986807,0.0844475478,-0.4061064422,0.0917846486,-0.3141177595,0.5976012349,0.3187380135,-0.1618884355,0.1839360744,-0.2865276337,0.4400342107,-0.1198556721,-0.0586536154,0.5181100368,-0.4390115738,-0.2725733221,0.2292658538,-0.1000218242,0.6527184844,0.1551799476,-0.0103552602,0.2116364539,0.2181542963,0.1894761324,-0.2540006638,0.0620612241,-0.2933145463,-0.2902559936,0.1150470451,-0.1187709197,0.2748994231,0.3439331353,-0.371096611,0.2314968705,-0.1662519872,0.1161675975,0.2451562285,0.467181772,-0.2161771506,0.0160147771,-0.0778770745,0.0520516336,0.3815431297,0.2001960129,-0.0275135282,-0.1653925627,-0.2098923177,-0.1588151157,0.0254393648,0.2343832701,-0.4178391695,-0.0726502389,-0.0403762981,-0.2204381078,0.1376759559,0.1729962379,0.2418773621,0.0537331626,-0.1271219254,-0.1456125379,-0.2578451633,-0.0458920449,0.194044292,0.067223452,0.1478221864,0.0373468921,-0.374296248,0.2848320901,-0.1732952595,-0.2097570896,0.3150957227,-0.1556826532,0.1648638994,-0.3636010289,0.0174216125,-0.0640936717,-0.136324659,-0.078526333,0.0533220693,0.0521458909,-0.7080106735,0.1413471997,-0.2713920176,-0.3244532049,-0.0393841602,0.1066874638,-0.2228568494,0.4759735167,0.4173581004,-0.1163219512,0.3777332306,-0.2346686572,0.3289363086,0.3370268047,-0.1697399318,0.6116288304,-0.0555214994,-0.3803268671,0.1090204269,0.3577328324,0.2878513932,-0.0637150183,-0.2488834113,-0.1406081617,-0.1500300318,-0.1582457572,0.1125159264,0.3955524266,-0.1776655316,0.295866847,-0.1747024357,0.3727815747,-0.1949801147,-0.0024365336,-0.1881541312,0.1233800352,-0.0857628807,-0.182947129,0.2429677695,-0.0049513602,0.0499992259,-0.0530919172,-0.2602922916,-0.1787487417,-0.294421643,0.1744043082,0.0657970309,-0.238659665,0.1459914297,0.0540410914,-0.0258019716,-0.1207975522,0.1762764901,-0.3385201991,-0.0109814387,0.2525725663,0.3427561522,-0.0067282249,-0.6286964417,0.0099321706,-0.2735544443,0.2913064659,-0.219351992,0.0491062328,0.1606857926,0.032897003,0.1278441101,-0.2077875882,0.1320329756,-0.1334701329,0.5864819288,0.1192625463,0.0063474188,-0.0403591022,0.220347777,0.3805131018,0.0998040661,0.0783208087,0.1349651664,0.0863007009,-0.3046171963,0.0316624343,0.2106906176,0.0489396565,0.3257641494,0.274530679,0.2787441909,0.0171891041,-0.0304751955,-0.1940560639,0.2114888281,0.2699235976,0.1913394183,0.7015006542,-0.4936882555,0.3337839544,0.1052521989,-0.2653430104,0.2542865574,-0.4790070057,0.1929245144,0.6058324575,-0.3864374161,0.0741798952,-0.080145359,-0.211253792,0.2038942575,0.1999296099,0.2975566387,0.2877901793,0.026033625,0.3619494736,-0.5504577756,-0.0771566704,-0.1918422431,0.2061724663,0.2919972241,0.0384994112,-0.0975590497,0.0093709342,-0.4708305895,-0.2300649881,0.2454039603,-0.2955703735,-0.2263980508,-0.0519431494,-0.0112385619,0.3497267962,-0.2697206438,0.0553600155,-0.0817313492,-0.316609472,0.2097853869,0.1095300689,-0.0809019953,0.1963225454,0.4950662553,0.2401908338,0.0226626657,-0.4391456842,-0.3661148846,0.1244544089,0.0449609943,-0.1386365443,0.3397325277,-0.3198176324,-0.2074073851,0.2640452087,0.0414099731,-0.0806637257,-0.439886868,0.2591479123,-0.4032660127,-0.1769732833,0.6636248231,0.1579113901,0.0095406082,0.0294660125,-0.2400483936,-0.0903023556,0.0417932086,0.1923097223,-0.4423007965,0.1097011864,0.0098779434,0.0258049797,-0.2754499018,0.4726877213,-0.0548032969,-0.0301341657,-0.2771562636,-0.1637395322,-0.573884666,0.2634527385,-0.0527890585,-0.4865395427,0.0810311958,-0.1199837923,0.3776105642,0.1762557328,-0.1693374813,0.3486984968,0.2776536942,0.2912560999,-0.3625783324,-0.2438791841,-0.2989762425,0.0307736583,-0.0500393957,-0.4768276513,0.0896086022,-0.0004992259,-0.1357264072,0.1987256855,0.0193388946,0.3844162226,0.0003760862,0.2715222836,0.0879437923,0.4524197876,-0.1005812734,0.1148442626,-0.0384095423,0.2583157718,-0.1768371612,-0.0881766677,0.2198528945,0.2592366934,0.0071110297,0.2904507518,-0.3201829493,0.4042015374,0.3185971081,0.3075398207,-0.556835115,0.1301245689,-0.2209291756,-0.0340216942,0.6548748612,0.2844884694,-0.3331443667,0.117379941,-0.0490211435,0.0082143201,0.3080021441,-0.1347652823,-0.1568375677,-0.2240474224,0.0681478232,-0.0966738537,-0.1177390516,-0.145793587,-0.1565946043,0.1873218864,-0.0510618985,0.1215432063,0.1973781735,0.514441371,-0.0037093125,0.0087381527,0.287556529,0.2362604737,-0.0319847018,0.0270491075,-0.1153386012]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/649","title":"Inconsistent behavior in map","comments":"Thanks for reporting !\r\n\r\nThis issue must have appeared when we refactored type inference in `nlp`\r\nBy default the library tries to keep the same feature types when applying `map` but apparently it has troubles with nested structures. I'll try to fix that next week","body":"I'm observing inconsistent behavior when applying .map(). This happens specifically when I'm incrementally adding onto a feature that is a nested dictionary. Here's a simple example that reproduces the problem.\r\n\r\n```python\r\nimport datasets\r\n\r\n# Dataset with a single feature called 'field' consisting of two examples\r\ndataset = datasets.Dataset.from_dict({'field': ['a', 'b']})\r\nprint(dataset[0])\r\n# outputs\r\n{'field': 'a'}\r\n\r\n# Map this dataset to create another feature called 'otherfield', which is a dictionary containing a key called 'capital'\r\ndataset = dataset.map(lambda example: {'otherfield': {'capital': example['field'].capitalize()}})\r\nprint(dataset[0])\r\n# output is okay\r\n{'field': 'a', 'otherfield': {'capital': 'A'}}\r\n\r\n# Now I want to map again to modify 'otherfield', by adding another key called 'append_x' to the dictionary under 'otherfield'\r\nprint(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x'}})[0])\r\n# printing out the first example after applying the map shows that the new key 'append_x' doesn't get added\r\n# it also messes up the value stored at 'capital'\r\n{'field': 'a', 'otherfield': {'capital': None}}\r\n\r\n# Instead, I try to do the same thing by using a different mapped fn\r\nprint(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['otherfield']['capital']}})[0])\r\n# this preserves the value under capital, but still no 'append_x'\r\n{'field': 'a', 'otherfield': {'capital': 'A'}}\r\n\r\n# Instead, I try to pass 'otherfield' to remove_columns\r\nprint(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['otherfield']['capital']}}, remove_columns=['otherfield'])[0])\r\n# this still doesn't fix the problem\r\n{'field': 'a', 'otherfield': {'capital': 'A'}}\r\n\r\n# Alternately, here's what happens if I just directly map both 'capital' and 'append_x' on a fresh dataset.\r\n\r\n# Recreate the dataset\r\ndataset = datasets.Dataset.from_dict({'field': ['a', 'b']})\r\n# Now map the entire 'otherfield' dict directly, instead of incrementally as before\r\nprint(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['field'].capitalize()}})[0])\r\n# This looks good!\r\n{'field': 'a', 'otherfield': {'append_x': 'ax', 'capital': 'A'}}\r\n```\r\n\r\nThis might be a new issue, because I didn't see this behavior in the `nlp` library. \r\n\r\nAny help is appreciated!","comment_length":45,"text":"Inconsistent behavior in map\nI'm observing inconsistent behavior when applying .map(). This happens specifically when I'm incrementally adding onto a feature that is a nested dictionary. Here's a simple example that reproduces the problem.\r\n\r\n```python\r\nimport datasets\r\n\r\n# Dataset with a single feature called 'field' consisting of two examples\r\ndataset = datasets.Dataset.from_dict({'field': ['a', 'b']})\r\nprint(dataset[0])\r\n# outputs\r\n{'field': 'a'}\r\n\r\n# Map this dataset to create another feature called 'otherfield', which is a dictionary containing a key called 'capital'\r\ndataset = dataset.map(lambda example: {'otherfield': {'capital': example['field'].capitalize()}})\r\nprint(dataset[0])\r\n# output is okay\r\n{'field': 'a', 'otherfield': {'capital': 'A'}}\r\n\r\n# Now I want to map again to modify 'otherfield', by adding another key called 'append_x' to the dictionary under 'otherfield'\r\nprint(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x'}})[0])\r\n# printing out the first example after applying the map shows that the new key 'append_x' doesn't get added\r\n# it also messes up the value stored at 'capital'\r\n{'field': 'a', 'otherfield': {'capital': None}}\r\n\r\n# Instead, I try to do the same thing by using a different mapped fn\r\nprint(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['otherfield']['capital']}})[0])\r\n# this preserves the value under capital, but still no 'append_x'\r\n{'field': 'a', 'otherfield': {'capital': 'A'}}\r\n\r\n# Instead, I try to pass 'otherfield' to remove_columns\r\nprint(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['otherfield']['capital']}}, remove_columns=['otherfield'])[0])\r\n# this still doesn't fix the problem\r\n{'field': 'a', 'otherfield': {'capital': 'A'}}\r\n\r\n# Alternately, here's what happens if I just directly map both 'capital' and 'append_x' on a fresh dataset.\r\n\r\n# Recreate the dataset\r\ndataset = datasets.Dataset.from_dict({'field': ['a', 'b']})\r\n# Now map the entire 'otherfield' dict directly, instead of incrementally as before\r\nprint(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['field'].capitalize()}})[0])\r\n# This looks good!\r\n{'field': 'a', 'otherfield': {'append_x': 'ax', 'capital': 'A'}}\r\n```\r\n\r\nThis might be a new issue, because I didn't see this behavior in the `nlp` library. \r\n\r\nAny help is appreciated!\nThanks for reporting !\r\n\r\nThis issue must have appeared when we refactored type inference in `nlp`\r\nBy default the library tries to keep the same feature types when applying `map` but apparently it has troubles with nested structures. I'll try to fix that next week","embeddings":[0.3283730149,-0.292429328,-0.0719692931,0.0814116001,-0.0720765442,-0.2064360827,0.063771382,0.0195533317,0.2041487694,-0.0053224387,0.2920315266,0.5868220925,0.209287867,0.1178402677,-0.3162102103,0.1631347388,0.2691355646,-0.0755070373,0.0736097693,-0.1366803944,-0.250808388,0.0666691512,-0.5501969457,-0.1045745909,-0.0512326472,-0.2376290709,-0.2272645533,-0.2153911293,0.0933098421,-0.0640515685,-0.0802594051,0.1609658152,-0.3127307892,0.3500859141,-0.000115136,-0.1009186655,0.2150342166,-0.0772834495,0.0711237267,0.0254450012,-0.0651642308,-0.1851980239,-0.02156372,-0.2740437388,0.1911568046,-0.0858946815,-0.2054758519,-0.2903950214,0.0998265892,-0.2147040516,0.2083425671,-0.0882747099,0.0907837152,0.2036317587,0.0287717804,0.1171306372,-0.0643493831,0.0056961533,0.0296099745,-0.2457847744,0.1012372077,0.0323891975,-0.1886446178,-0.1628932059,0.3863431811,0.1219843328,0.324593991,-0.3239738941,0.0626328588,-0.019207634,0.0291196872,0.0251824632,-0.3432812095,-0.2539941072,-0.3001532257,-0.0827328116,0.227551654,-0.3730776906,-0.1039254665,-0.1914056838,-0.1468643993,-0.1279849857,0.2984395623,0.5585368872,-0.3571726084,0.0947527736,-0.1084648892,0.2734062374,-0.1603785902,-0.1644419581,-0.2618698478,-0.4508469999,0.085714303,0.3864740431,0.0043787709,0.0150338933,0.3914421797,-0.1868936867,-0.0469609238,-0.4379443824,0.1130760089,0.2154432684,-0.3255035281,0.2683792114,0.3306059539,0.1537620127,0.4373146594,0.2231998295,0.1474878341,-0.0318467394,-0.3212687969,0.0861442313,0.608695507,0.0624115653,0.2395029366,-0.2713067234,0.4256381691,0.079845883,0.0098994179,0.1574280113,-0.5427674055,-0.0663734451,0.0712640584,-0.1117040962,0.2715506852,-0.0866183415,-0.0785830468,0.3137060702,0.0207541306,0.5313529968,-0.1912368387,-0.0386036038,-0.2077517062,0.0864646509,0.0024236182,0.1304289103,0.2730062604,0.3178071976,-0.139219597,-0.2177396864,-0.3970099986,-0.0170148592,0.3400411308,-0.1534978598,-0.3614324033,0.368181169,-0.0495872162,-0.4164009988,0.0381338149,0.2708564699,-0.0403651558,0.1602687985,0.3031720817,0.1658837795,-0.1487569958,-0.0807443038,-0.1751647443,0.3520653546,0.4351246059,-0.2993870676,0.1443832368,0.03544081,-0.4979199469,-0.4050822854,0.0301575642,0.5204486847,-0.1916330755,-0.0488709137,-0.0841391534,0.3315013647,0.0023969505,0.242001608,0.0136620728,0.1531416774,-0.1249736398,0.2145013362,0.2045967281,-0.1615843773,-0.4443526566,0.2402662188,-0.0463324562,0.4343183339,-0.3960137069,-0.0249242336,0.1610620469,-0.1516000777,0.1488508582,-0.0462927967,0.1757988036,-0.0066658645,-0.2617482841,0.3233524561,0.2869602442,-0.3480004668,-0.2542611361,0.4002697766,0.260843873,-0.0804875866,0.2576496303,0.1320992708,0.2061035484,-0.0267862137,0.105247885,0.0331108123,-0.113696672,0.0844125077,-0.4560530186,0.3330092728,-0.0763356537,-0.1763486862,-0.1707538664,0.0100766737,0.0096046682,0.1199487522,-0.0798503235,-0.0691365898,0.1536289603,-0.0008641088,-0.2326338589,-0.0664435104,0.1253821552,-0.1909986287,0.233099848,0.0806063712,-0.1279177666,0.0695782751,0.1376043409,-0.1389885843,-0.4655871093,0.1393829286,0.4000686407,-0.1673889458,-0.2414455414,0.364187628,0.0692586452,-0.0023018715,-0.228089124,0.0989651978,0.332587719,0.3299348652,0.035865251,0.1566070318,0.1441521198,-0.2623310089,-0.0911436453,0.1735057682,0.3850866556,0.3000432849,-0.3314894736,-0.0568488799,-0.168634966,0.0851567611,-0.0463741496,-0.6482141614,-0.3556046486,0.1406761557,-0.0350368656,-0.1433358788,0.192409128,0.4853856266,0.1540116221,-0.1167074144,0.0462063625,-0.1036287546,-0.0340313539,0.0007395066,-0.1649100035,0.2468136698,0.5029404163,-0.0317557603,0.2145577073,0.0306693222,0.3216382563,-0.2117250413,0.0213167649,0.1173986942,0.0161563568,0.1465897858,0.5601219535,0.1424714178,-0.1684336811,0.2632064521,0.1871362031,0.0386451036,0.0624045432,0.0907623842,-0.2531965077,0.1338758916,0.0510713346,-0.316115737,-0.098414816,-0.5975273252,-0.0520730019,0.3707389235,-0.2355754673,0.0839985386,0.3379156291,-0.3251425028,0.0241299178,-0.2906601429,0.0372631438,-0.2106794268,0.0101511134,-0.0008530918,-0.1192711815,-0.2419505864,0.0747872069,0.4518421292,-0.242514655,-0.344588697,-0.7150273323,-0.0899501368,-0.3523359895,0.1481159776,-0.0713120624,-0.0721300542,-0.0156705193,-0.092732273,0.2402779013,-0.2040441781,-0.1975307316,0.0924527049,0.1870163977,-0.3144246936,-0.1214853376,-0.0873897523,0.0895959884,-0.2379463762,0.2975449562,-0.068632029,0.0802902728,0.1324363351,-0.1091013625,-0.2828168869,0.174284026,-0.0787646249,-0.174080804,0.3903700113,0.228212133,-0.1205741987,0.1720118374,-0.0620093793,-0.1167679951,-0.1341580003,0.1855477393,-0.1187653467,-0.2799407542,-0.1594984978,0.4333631098,0.3595969975,0.0029952493,0.2989759445,0.154724732,-0.0664265156,-0.1760396808,-0.4334548116,0.1408878267,0.5860908628,0.0139514804,0.1631966531,0.2730633616,0.0733999982,0.3057008684,0.3173773289,-0.0980089009,0.2232822925,-0.1188824922,-0.0478850156,-0.2906945944,-0.1633788198,-0.2218584567,-0.2506219149,0.0141529832,0.0295924079,0.0194457062,-0.4428293705,0.1400326043,-0.1187648922,-0.2647045255,-0.2848890722,-0.051630348,-0.2999198437,0.284106195,0.1367472112,0.2810207605,-0.155203253,-0.1385188699,-0.0789591968,-0.3428583741,0.2689992785,-0.2150084376,-0.297165215,-0.121071808,-0.3153410256,0.388584286,0.0652255639,0.2098755091,0.1250799149,-0.4220082462,0.0919418558,0.1380857527,0.4020265639,-0.4266528785,-0.0200011134,-0.1246158034,-0.3372523189,-0.5130752325,-0.1785174161,0.0059911869,0.807505548,0.1143296286,0.4953757823,-0.0577767864,0.0176593568,-0.1742608696,0.0347505398,0.027113555,-0.1188025102,0.0878935158,-0.1159714609,-0.2429102212,-0.2227448672,0.0133712497,0.0186490174,0.3174106479,-0.3497681618,0.1555934846,0.1434915811,-0.1378663629,-0.0431593806,0.2221333981,0.2328152508,0.1089468747,0.0408273712,0.0711774305,-0.3110002875,0.3243189752,-0.0059174616,0.1358445436,-0.0720113665,-0.2679109275,-0.0119421575,0.2709792852,0.1727703512,0.2241785973,-0.2227839977,-0.0291614421,-0.022587413,-0.2912722528,0.5955764055,0.0222176779,0.0258124173,-0.1589710414,0.422452718,-0.0279713254,-0.4867108464,0.3390016854,-0.0815821961,-0.2324020863,0.5947387815,-0.1880053282,0.5043503642,0.1774413884,-0.0333691053,0.2652246356,0.0960999727,-0.0781436339,0.100058198,0.2075048238,-0.3149503171,-0.173515141,-0.0446480848,0.1884694993,-0.1958250254,0.189678967,-0.1084643379,0.0508311726,-0.1657967269,0.6276268959,0.0688498244,-0.310153842,0.3981218934,-0.0345554054,-0.1007142067,0.0702549815,0.4636613131,0.3073572218,0.0203547776,-0.0549092405,-0.1906632185,-0.3429856896,-0.0222574528,0.0342615657,-0.1484508067,0.2699088752,0.0888038948,0.0082464358,0.0272732805,0.5211567283,0.0885119736,0.1108673066,0.178979367,-0.1381908059,0.5807952881,0.2329128832,0.1320389062,-0.2373419702,-0.2910971642,0.143728599,-0.280442059,0.0871349946,-0.255281508,-0.0709308013,-0.7718071938,0.3505956531,-0.2758606374,-0.2989434004,-0.0883858353,0.0231949054,-0.2038773596,-0.1183102205,0.0394661203,-0.1617492735,-0.0789023563,0.3045651615,-0.1264169663,-0.19723849,0.1007835716,0.5037269592,0.0695550889,0.4625643492,0.2889648974,-0.2222478092,-0.0142580522,-0.0470950566,-0.1965955049,-0.2833275795,0.2451336831,0.0803102851,-0.1415915638,0.0006988117,0.1558644772,0.3267731965,-0.3146661818,0.0501044542,-0.2087600678,-0.2365890294,-0.0785741284,0.3313898146,0.1773798168,0.077193141,-0.1244438738,-0.0876157656,-0.0371808261,0.5158184767,-0.249216035,0.0297319572,0.5445590615,0.3482589424,0.5488528609,0.1669693738,0.0330428854,-0.1139920652,0.1379629523,0.0576443821,-0.0070280912,-0.1969985366,0.0927834362,0.0975269601,0.246723026,0.1431359351,-0.0202176999,-0.0817831382,0.2235704362,-0.3962993622,0.3992108107,0.0477069207,-0.1095539033,0.2654606104,-0.1771471947,0.3552578688,0.2055754364,0.2373932898,-0.1122223809,-0.124263294,0.216254279,0.0146351205,0.0578589514,-0.0392930731,-0.0661937594,-0.0794739947,-0.376236558,0.1798868626,0.3319211304,-0.2201945633,0.0316512585,0.2090612054,0.0847323164,0.1028605103,-0.196068123,-0.199349761,-0.4081462026,0.2083772868,-0.1368820965,-0.2583907247,0.7140581608,0.1178330407,0.4833857417,0.5350748301,0.1387886703,-0.3639434278,-0.2383733988,-0.033843495,-0.1895392537,-0.2641141117,0.1569013298,0.5044384599,-0.093521148,0.0170585904,0.2735711336,0.1556527168,-0.0954524726,0.4887837172,0.5087665915,0.0641679242,0.0838842392,-0.1006654724,0.1373464912,-0.1960140765,0.3739913404,0.2221969813,0.1100254953,0.1483699679,0.2638943493,0.0017185556,0.1322786957,-0.3337880671,0.0190555062,0.2291946858,-0.211275354,-0.3193449974,0.0417507105,-0.3345662653,-0.0603923015,-0.2493552864,-0.1686457098,-0.1024260595,0.5230631828,-0.1845660508,-0.0359461308,-0.1700662822,0.0130208153,-0.2495191544,0.5381356478,-0.0387337096,0.2034341991,-0.0627981648,0.0051704259,-0.0611837208,-0.1318743229,0.1170631051,0.2796711326,-0.3479600549,0.4438444376,0.0131661156,-0.3360470533,0.1206346452,0.2105064839,-0.204814449,0.0788402483,0.1018049046,0.1002450436,-0.1397599727,0.1486285031,0.3657957911,0.434281081,-0.074540332,0.221506089,0.0665691346,-0.071638301,0.0137890205,-0.2227043957,0.1665644646,-0.5212206244,0.0166326631,-0.1729622036,0.2235323936,0.0654628649,0.0667767599,0.0536085851,0.2340756506,0.1249891222,-0.1575349569,-0.2167906314,0.0029990005,-0.4744950235,0.1615250856,-0.1510285884,-0.0472112074,-0.1579040885,0.0311761666,0.3251591027,-0.0296739638,0.1237818822,-0.4717045724,-0.26311782,0.339415133,-0.0446448363,-0.1144512072,-0.1804471314,-0.2015485615,0.1241681054,-0.3140755892,0.1728306413,-0.1143609658,0.1543126404,-0.0288260728,0.0082055004,-0.3365116119,-0.2416059822,0.5207398534,0.3949013054,-0.0324204378,-0.2991228402,-0.290127337,-0.1878260374,0.0111459875,-0.2121374756,0.3095389605,-0.0964078456,0.3693115711,0.0061808852,0.0086327661,0.0333503969,0.1296662539,0.1743185371,-0.3438021839,-0.0990054011,0.0399337895,-0.2183191925,0.1883189827,-0.017157238,0.5380637646,-0.1453256458,0.3856753707,0.0955808088,-0.3196715117,0.3257689178,-0.3355548084,-0.1595036089,0.1842433363,0.0158362351,0.0343552902,-0.2107642591,0.0301295538,-0.4007390141,0.2965137959,-0.3499027193,-0.1761531234,0.0013840265,-0.2839898467,-0.3851125836,-0.057937365,0.1256389767,0.0840851143,-0.1349222809,0.1699385494,-0.1088757217]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/648","title":"offset overflow when multiprocessing batched map on large datasets.","comments":"This should be fixed with #645 ","body":"It only happened when \"multiprocessing\" + \"batched\" + \"large dataset\" at the same time.\r\n\r\n```\r\ndef bprocess(examples):\r\n examples['len'] = []\r\n for text in examples['text']:\r\n examples['len'].append(len(text))\r\n return examples\r\nwiki.map(brpocess, batched=True, num_proc=8)\r\n```\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/multiprocessing\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 153, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 1486, in _map_single\r\n batch = self[i : i + batch_size]\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 1071, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 972, in _getitem\r\n data_subset = self._data.take(indices_array)\r\n File \"pyarrow\/table.pxi\", line 1145, in pyarrow.lib.Table.take\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/pyarrow\/compute.py\", line 268, in take\r\n return call_function('take', [data, indices], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nArrowInvalid Traceback (most recent call last)\r\n in \r\n 30 owt = datasets.load_dataset('\/home\/yisiang\/datasets\/datasets\/openwebtext\/openwebtext.py', cache_dir='.\/datasets')['train']\r\n 31 print('load\/create data from OpenWebText Corpus for ELECTRA')\r\n---> 32 e_owt = ELECTRAProcessor(owt, apply_cleaning=False).map(cache_file_name=f\"electra_owt_{c.max_length}.arrow\")\r\n 33 dsets.append(e_owt)\r\n 34 \r\n\r\n~\/Reexamine_Attention\/electra_pytorch\/_utils\/utils.py in map(self, **kwargs)\r\n 126 writer_batch_size=10**4,\r\n 127 num_proc=num_proc,\r\n--> 128 **kwargs\r\n 129 )\r\n 130 \r\n\r\n~\/hugdatafast\/hugdatafast\/transform.py in my_map(self, *args, **kwargs)\r\n 21 if not cache_file_name.endswith('.arrow'): cache_file_name += '.arrow'\r\n 22 if '\/' not in cache_file_name: cache_file_name = os.path.join(self.cache_directory(), cache_file_name)\r\n---> 23 return self.map(*args, cache_file_name=cache_file_name, **kwargs)\r\n 24 \r\n 25 @patch\r\n\r\n~\/datasets\/src\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1285 logger.info(\"Spawning {} processes\".format(num_proc))\r\n 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard]\r\n-> 1287 transformed_shards = [r.get() for r in results]\r\n 1288 logger.info(\"Concatenating {} shards from multiprocessing\".format(num_proc))\r\n 1289 result = concatenate_datasets(transformed_shards)\r\n\r\n~\/datasets\/src\/datasets\/arrow_dataset.py in (.0)\r\n 1285 logger.info(\"Spawning {} processes\".format(num_proc))\r\n 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard]\r\n-> 1287 transformed_shards = [r.get() for r in results]\r\n 1288 logger.info(\"Concatenating {} shards from multiprocessing\".format(num_proc))\r\n 1289 result = concatenate_datasets(transformed_shards)\r\n\r\n~\/miniconda3\/envs\/ml\/lib\/python3.7\/multiprocessing\/pool.py in get(self, timeout)\r\n 655 return self._value\r\n 656 else:\r\n--> 657 raise self._value\r\n 658 \r\n 659 def _set(self, i, obj):\r\n\r\nArrowInvalid: offset overflow while concatenating arrays\r\n```","comment_length":6,"text":"offset overflow when multiprocessing batched map on large datasets.\nIt only happened when \"multiprocessing\" + \"batched\" + \"large dataset\" at the same time.\r\n\r\n```\r\ndef bprocess(examples):\r\n examples['len'] = []\r\n for text in examples['text']:\r\n examples['len'].append(len(text))\r\n return examples\r\nwiki.map(brpocess, batched=True, num_proc=8)\r\n```\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/multiprocessing\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 153, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 1486, in _map_single\r\n batch = self[i : i + batch_size]\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 1071, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 972, in _getitem\r\n data_subset = self._data.take(indices_array)\r\n File \"pyarrow\/table.pxi\", line 1145, in pyarrow.lib.Table.take\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/pyarrow\/compute.py\", line 268, in take\r\n return call_function('take', [data, indices], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nArrowInvalid Traceback (most recent call last)\r\n in \r\n 30 owt = datasets.load_dataset('\/home\/yisiang\/datasets\/datasets\/openwebtext\/openwebtext.py', cache_dir='.\/datasets')['train']\r\n 31 print('load\/create data from OpenWebText Corpus for ELECTRA')\r\n---> 32 e_owt = ELECTRAProcessor(owt, apply_cleaning=False).map(cache_file_name=f\"electra_owt_{c.max_length}.arrow\")\r\n 33 dsets.append(e_owt)\r\n 34 \r\n\r\n~\/Reexamine_Attention\/electra_pytorch\/_utils\/utils.py in map(self, **kwargs)\r\n 126 writer_batch_size=10**4,\r\n 127 num_proc=num_proc,\r\n--> 128 **kwargs\r\n 129 )\r\n 130 \r\n\r\n~\/hugdatafast\/hugdatafast\/transform.py in my_map(self, *args, **kwargs)\r\n 21 if not cache_file_name.endswith('.arrow'): cache_file_name += '.arrow'\r\n 22 if '\/' not in cache_file_name: cache_file_name = os.path.join(self.cache_directory(), cache_file_name)\r\n---> 23 return self.map(*args, cache_file_name=cache_file_name, **kwargs)\r\n 24 \r\n 25 @patch\r\n\r\n~\/datasets\/src\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1285 logger.info(\"Spawning {} processes\".format(num_proc))\r\n 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard]\r\n-> 1287 transformed_shards = [r.get() for r in results]\r\n 1288 logger.info(\"Concatenating {} shards from multiprocessing\".format(num_proc))\r\n 1289 result = concatenate_datasets(transformed_shards)\r\n\r\n~\/datasets\/src\/datasets\/arrow_dataset.py in (.0)\r\n 1285 logger.info(\"Spawning {} processes\".format(num_proc))\r\n 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard]\r\n-> 1287 transformed_shards = [r.get() for r in results]\r\n 1288 logger.info(\"Concatenating {} shards from multiprocessing\".format(num_proc))\r\n 1289 result = concatenate_datasets(transformed_shards)\r\n\r\n~\/miniconda3\/envs\/ml\/lib\/python3.7\/multiprocessing\/pool.py in get(self, timeout)\r\n 655 return self._value\r\n 656 else:\r\n--> 657 raise self._value\r\n 658 \r\n 659 def _set(self, i, obj):\r\n\r\nArrowInvalid: offset overflow while concatenating arrays\r\n```\nThis should be fixed with #645 ","embeddings":[-0.4724798799,-0.5470885038,-0.0690055341,0.2029455453,-0.0922680721,0.0032601107,0.431145668,0.1213298067,0.0042691873,0.3337112367,0.2105832994,-0.0030280177,0.0025017336,0.0706486031,-0.2007900327,0.0564653911,0.1238672137,-0.1653056592,-0.1999184489,0.0210992452,-0.4086596668,0.0652992353,-0.2608809173,-0.0614106096,-0.0878032967,-0.1395558715,0.0227157604,-0.0221663248,-0.0390631594,-0.0816951841,0.1919379383,-0.0817265511,-0.112472944,0.3537240326,-0.0001111824,-0.0109583959,0.1242788509,0.0284189507,0.2034844458,0.1440587193,-0.0371055529,-0.1708365679,-0.2350585908,-0.3127315938,0.2882950306,0.0973996967,-0.1417520791,-0.2249159366,0.262106061,0.1953805238,0.2192112356,0.0814040527,-0.015035429,0.1142898053,-0.0161750019,-0.3034172058,0.0330700912,-0.2385724783,0.2039914578,-0.1968372017,-0.302244097,0.1477174759,-0.1354442239,0.1338905096,-0.1166878045,0.0154709024,0.343078047,-0.3428082764,0.2343281358,0.0401816517,-0.2495435923,0.0194753911,-0.2856802344,-0.4185943007,-0.2120327055,-0.2452063709,0.1186968908,0.3503252566,-0.1741090566,-0.1184589639,-0.7064749002,0.1826454848,-0.0528968424,0.0070800316,-0.3116375506,0.6008684039,-0.0026980031,0.2060533762,0.1709334999,0.1001436785,0.0688418746,-0.0227788482,-0.0647912994,0.1665035337,-0.1708871424,0.1595836431,0.4488881528,0.058397118,0.3597428501,-0.1830900162,-0.1618460417,-0.0563312434,0.2343739718,0.0594697259,0.3856415749,0.0190943126,-0.2259369195,0.0672159642,0.0536056608,0.074197188,-0.1498518586,-0.0234795306,0.1413926631,-0.4096342027,0.1917526573,0.225269258,0.0386171266,-0.0636864379,-0.0608621798,0.1051027328,0.1684311628,0.1251915991,0.0406736508,0.1396966577,0.0842215717,0.3683578372,-0.0960457772,0.1865689456,0.046105843,0.0762494057,-0.164629072,-0.2376046777,-0.1502536684,0.02442687,-0.0891724303,0.3796938062,0.2486175895,0.1886924207,0.2254287601,-0.1198965088,-0.0877317041,-0.2744494677,0.6435857415,0.1651964039,0.3063707352,0.2903899848,0.1194322109,0.2625783384,-0.1908123344,0.2612568438,-0.4289084077,-0.325552702,-0.0994048268,0.2534756958,-0.0046243309,0.3032531738,-0.3036207855,-0.0070459866,0.5798020959,-0.0259820241,0.2367868125,-0.2012345642,-0.3595503867,-0.071069032,0.0848903954,0.2256240398,-0.1289972067,0.2306214869,0.0653450936,0.2446785271,0.3458192945,0.3322739303,-0.273925513,0.2895392179,-0.2685263157,-0.0099180695,-0.0526836701,-0.4311921,-0.4203554988,0.1995908469,-0.1446069628,0.1888927966,0.067600131,-0.325679332,0.4037618041,-0.0403162017,0.4177243114,0.486638248,-0.3132050335,0.4952070117,-0.2428607047,-0.0260037761,-0.338896066,-0.0933043957,0.0249444526,-0.3816543818,-0.0841173306,-0.0041723531,0.3307575583,-0.0684263259,0.2112171799,0.4032656848,-0.0910114422,-0.0005261881,-0.0976596624,-0.1767158657,0.1018023938,0.1037141532,-0.2379125655,-0.0061032944,-0.1494359225,-0.1178122908,0.1189868525,0.2392622083,-0.465526104,-0.0865261182,0.2350709587,-0.1330577731,0.2071367949,-0.1156043336,0.3647233248,0.0955672786,-0.0473960899,-0.1623814106,-0.0440338589,-0.1695727557,-0.1972923279,-0.256564945,-0.1027143374,0.277923882,0.1670152098,-0.0678443238,-0.1385055035,0.3817984164,0.0144724855,-0.1126673967,0.0447344035,-0.0673192292,0.0487594195,-0.0025495754,0.0887556002,-0.1664895415,-0.1406800449,-0.0731712207,-0.0806001648,0.4928908944,-0.0429759696,0.450468123,0.1013248265,-0.0087925028,0.3333130181,0.0147212017,0.0952108949,-0.108680509,0.1052466109,0.1580382138,-0.1132940501,-0.1801329553,0.0783963725,0.2769286036,0.640108645,0.1786997169,-0.1603690535,-0.0665042922,-0.316055119,-0.1899700761,0.1805679351,0.0096732127,0.2360683829,0.3596695662,0.1982408315,-0.1609237045,0.0647445619,-0.0289550051,0.2104812413,0.1276272237,0.3061177135,0.1302338988,0.5117682219,-0.0963642895,-0.0896140933,-0.061671,0.0016787803,0.0674070418,-0.1021869108,0.2256777734,0.0637503341,0.0345311277,-0.342966795,-0.2171401083,-0.1741497815,-0.6312443614,-0.0058589131,0.2816341817,-0.1140101403,0.3577982783,0.1590273827,0.0000863484,0.3099007905,-0.3061830997,0.1523242146,-0.2501434088,-0.1402078122,0.0344589762,0.2034917474,-0.4103622437,0.2497137338,0.2418713868,-0.0073505361,-0.1619130671,-0.1759203672,-0.1180872172,0.0211110394,0.1095370501,0.0799917728,0.0196468625,0.0081992215,-0.358101666,0.1158828139,-0.2884114385,-0.155771181,0.0917602256,-0.1746366918,-0.1932019889,-0.2032632232,0.0032963841,-0.0095923301,-0.3650891185,0.1923043281,0.0071315547,0.1093452275,-0.0129435258,0.0516555198,-0.1818473786,-0.0121176485,-0.2586720884,-0.2050638199,-0.3385683298,0.2415130436,-0.240699172,-0.1149851307,-0.1148272604,0.0164141431,0.3596509099,0.2422824502,-0.2953150868,-0.0852859169,-0.0846701711,0.5255175829,0.0984388217,0.123869054,0.2237486243,0.1018691361,-0.1512203515,-0.127075538,-0.0096969903,0.1674998403,-0.1645011157,0.0745369568,-0.0171891805,0.1135187075,0.1911734641,0.8231935501,0.3125421107,0.0323668718,0.0461599529,-0.052046705,-0.0774144232,-0.0887126774,-0.465357244,0.155716002,0.0458407663,0.0824740678,0.1671870202,-0.0260663796,-0.3475735486,0.1483848393,-0.0090955412,-0.3566475213,-0.2947732508,0.3027471602,-0.109523721,0.1417776048,-0.0641486421,-0.1482805163,-0.496786803,-0.155815348,-0.1496390402,-0.3124412894,0.002707453,-0.0434226803,0.2983742356,-0.1023147926,-0.539770484,0.2341855168,-0.0011001605,0.1348960102,0.220674932,0.0508665182,-0.0949329287,-0.2084873766,0.8502280712,-0.2923719883,0.0277640894,-0.1798008233,0.0371931717,-0.4001259208,-0.1800974011,-0.2615940869,0.6538347006,0.691722393,0.4760460556,-0.1589214355,0.0430680811,0.3724712133,0.0677732229,-0.1501819342,-0.0856579468,-0.1202186197,-0.2358566374,-0.0233640037,0.3128212094,0.1812190115,0.3029714823,-0.0154260341,-0.2447441667,-0.0116983159,0.280726552,-0.2219143212,0.1512068212,0.0883365422,-0.1519748122,0.1290318668,-0.2227023095,0.2646117806,0.3021807969,0.1920861155,-0.2047645003,-0.2822438776,-0.1276487559,-0.2989731431,0.2527008057,0.214343071,0.0695905909,0.0509049296,0.1981263012,0.5551554561,-0.1061234549,0.0026973302,0.0856717229,0.3002865016,-0.1544084102,-0.2417069077,0.083387658,0.217290163,0.0350439325,0.2689328492,-0.2343525887,-0.416765511,0.1570390165,0.1068356484,0.5609591603,-0.090178743,-0.0360853747,-0.0071783587,0.1886250675,0.1050648317,-0.1770541072,0.2332184017,-0.1435902119,0.1433256716,0.156323567,-0.0471651666,0.0099960323,0.0531223454,-0.2045133859,0.2653582394,0.1411941946,-0.2298077047,-0.367251873,-0.0316803716,0.2411786616,-0.3824780285,-0.2204782218,0.1157284379,0.2500836253,-0.2820232511,-0.0830067247,-0.008922047,0.0639514625,-0.3150451481,-0.4909215569,0.0008943704,-0.4151181579,0.1651987284,-0.2484632879,-0.1469291896,-0.0685079843,0.1301172227,0.1298281401,0.0829190686,-0.0241292026,-0.049858924,0.0876322165,-0.1707328707,-0.0902273655,-0.0541367233,0.1428706199,-0.0124208014,-0.4450484216,-0.0248871818,-0.007269552,-0.0792046115,0.1262294799,0.367962122,0.1421956867,-0.14552176,-0.0489516668,-0.0944611728,-0.1117519587,-0.2680504322,0.0995053202,-0.1385320872,-0.4535485804,0.3498203754,-0.3295318782,-0.3506829441,-0.1023664623,0.3144615293,0.0909861401,0.0328479297,0.2585543096,0.2493997514,-0.1750436425,-0.1751388311,-0.0500649922,-0.0394997112,-0.2380501628,-0.0654175505,-0.0486120917,0.0547090732,0.1518817395,-0.0020762542,0.1433032453,0.1593454033,-0.3186149299,-0.285115093,-0.1217942685,-0.1914694607,-0.3388714194,-0.1731633395,0.440436244,0.1156077236,0.1084085181,0.2029795051,-0.3412675858,0.0495141074,-0.278188765,0.2500203848,-0.1471015811,-0.1701536626,0.1120382249,0.0142813195,0.1152472794,0.0972592086,-0.0491243675,-0.311406076,-0.1150640175,0.0383484252,0.0270475671,-0.2261347622,0.1507647783,0.1011740863,-0.0518595167,-0.3778333664,0.3535485268,-0.007179637,0.0689388812,0.0358417593,0.2494354546,0.0135658802,-0.090134427,-0.0148738995,0.0916951969,0.1114624292,-0.0135054663,0.512996614,0.1663272828,-0.1833403707,0.1983282268,-0.0952807665,0.1587287635,0.1796399951,0.2281201333,-0.2174176574,-0.2915329933,0.448148489,0.2858531475,0.1493940353,-0.1553090513,-0.2026843578,-0.0309864935,0.2425784916,-0.1084846631,-0.0957228094,0.3049129546,-0.1626545191,0.3035263121,0.4112124443,0.1383127421,-0.4035736918,0.1061665788,0.1432548314,0.3482195735,-0.3618722558,0.0650743023,0.2460177392,0.0900607184,0.1868119091,0.3765761554,0.1712119728,0.4348614216,0.5098477602,0.0847004279,0.3632257581,-0.0262439065,-0.4669933617,0.0102180159,-0.4753915966,0.0066250828,-0.0887213871,-0.08602532,0.3487884402,0.1319974065,0.1310573816,-0.0957048759,-0.1422550678,-0.241482541,0.2645242214,-0.3801378608,-0.1171856821,-0.1314582676,-0.0642014593,0.2291437685,0.0430430397,0.1383484453,0.2140901834,0.7544375062,-0.0522832423,-0.0296985153,-0.1664370447,-0.0638668463,-0.0322883353,0.1268276572,-0.2434642464,0.3029705286,0.2656693161,0.0155309532,-0.3197233975,0.3531139791,0.6688971519,0.3504451811,-0.0052596037,0.1054035276,0.1625346392,-0.163266331,-0.5884806514,0.0424423143,0.0139992479,0.3201362193,0.2330350429,0.2186632603,-0.2325776964,0.2933358848,0.3576749861,0.001839252,0.2867113054,-0.2394672483,0.0288397577,-0.3740426004,0.0247320458,0.0591021441,-0.1677625179,0.1108387858,0.273778528,-0.0377502441,0.3452721834,-0.0825715959,0.0966709778,0.0151069053,0.5231214762,0.5577062964,0.1013848782,-0.2210204154,-0.1186115891,-0.2194136977,0.3260878026,-0.4987770021,0.016561538,0.0156782158,0.2855914235,-0.1411339343,0.0241953731,0.4553402662,-0.3068294525,0.007509788,0.4098079801,-0.6726940274,-0.1202630773,0.2867000699,-0.3018577695,0.2044043988,-0.347600311,0.2213242501,0.1684019119,0.1794318855,-0.0864701048,0.3014255464,-0.0357852839,-0.0147195021,0.1236578748,0.503100574,0.1960096508,-0.2836230993,0.0266222022,-0.2694171369,0.1464063376,-0.3676693738,0.1431274861,0.0292756706,0.1391620338,-0.2626158595,-0.5490094423,-0.0643850043,-0.0190883167,0.3117630184,-0.0982327834,0.0152621511,0.1640101522,0.0423423275,-0.0885275081,-0.025079364,0.2250818461,-0.0628835484,0.1279391944,-0.1627910435,-0.1108391136,-0.0364003181,-0.2888792455,-0.3803110123,-0.2436331213,0.2614288628,-0.3879218102,-0.0652119741,-0.4406741858,0.0524187647,0.2756070495,0.1199200898,-0.4911474288,0.203205213,-0.2661342919,-0.0749607533,-0.044761885,-0.0224792548,0.0614942983,-0.5596327782,0.1373844147,-0.0810561255]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/648","title":"offset overflow when multiprocessing batched map on large datasets.","comments":"Feel free to re-open if it still occurs","body":"It only happened when \"multiprocessing\" + \"batched\" + \"large dataset\" at the same time.\r\n\r\n```\r\ndef bprocess(examples):\r\n examples['len'] = []\r\n for text in examples['text']:\r\n examples['len'].append(len(text))\r\n return examples\r\nwiki.map(brpocess, batched=True, num_proc=8)\r\n```\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/multiprocessing\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 153, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 1486, in _map_single\r\n batch = self[i : i + batch_size]\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 1071, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 972, in _getitem\r\n data_subset = self._data.take(indices_array)\r\n File \"pyarrow\/table.pxi\", line 1145, in pyarrow.lib.Table.take\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/pyarrow\/compute.py\", line 268, in take\r\n return call_function('take', [data, indices], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nArrowInvalid Traceback (most recent call last)\r\n in \r\n 30 owt = datasets.load_dataset('\/home\/yisiang\/datasets\/datasets\/openwebtext\/openwebtext.py', cache_dir='.\/datasets')['train']\r\n 31 print('load\/create data from OpenWebText Corpus for ELECTRA')\r\n---> 32 e_owt = ELECTRAProcessor(owt, apply_cleaning=False).map(cache_file_name=f\"electra_owt_{c.max_length}.arrow\")\r\n 33 dsets.append(e_owt)\r\n 34 \r\n\r\n~\/Reexamine_Attention\/electra_pytorch\/_utils\/utils.py in map(self, **kwargs)\r\n 126 writer_batch_size=10**4,\r\n 127 num_proc=num_proc,\r\n--> 128 **kwargs\r\n 129 )\r\n 130 \r\n\r\n~\/hugdatafast\/hugdatafast\/transform.py in my_map(self, *args, **kwargs)\r\n 21 if not cache_file_name.endswith('.arrow'): cache_file_name += '.arrow'\r\n 22 if '\/' not in cache_file_name: cache_file_name = os.path.join(self.cache_directory(), cache_file_name)\r\n---> 23 return self.map(*args, cache_file_name=cache_file_name, **kwargs)\r\n 24 \r\n 25 @patch\r\n\r\n~\/datasets\/src\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1285 logger.info(\"Spawning {} processes\".format(num_proc))\r\n 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard]\r\n-> 1287 transformed_shards = [r.get() for r in results]\r\n 1288 logger.info(\"Concatenating {} shards from multiprocessing\".format(num_proc))\r\n 1289 result = concatenate_datasets(transformed_shards)\r\n\r\n~\/datasets\/src\/datasets\/arrow_dataset.py in (.0)\r\n 1285 logger.info(\"Spawning {} processes\".format(num_proc))\r\n 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard]\r\n-> 1287 transformed_shards = [r.get() for r in results]\r\n 1288 logger.info(\"Concatenating {} shards from multiprocessing\".format(num_proc))\r\n 1289 result = concatenate_datasets(transformed_shards)\r\n\r\n~\/miniconda3\/envs\/ml\/lib\/python3.7\/multiprocessing\/pool.py in get(self, timeout)\r\n 655 return self._value\r\n 656 else:\r\n--> 657 raise self._value\r\n 658 \r\n 659 def _set(self, i, obj):\r\n\r\nArrowInvalid: offset overflow while concatenating arrays\r\n```","comment_length":8,"text":"offset overflow when multiprocessing batched map on large datasets.\nIt only happened when \"multiprocessing\" + \"batched\" + \"large dataset\" at the same time.\r\n\r\n```\r\ndef bprocess(examples):\r\n examples['len'] = []\r\n for text in examples['text']:\r\n examples['len'].append(len(text))\r\n return examples\r\nwiki.map(brpocess, batched=True, num_proc=8)\r\n```\r\n```\r\n---------------------------------------------------------------------------\r\nRemoteTraceback Traceback (most recent call last)\r\nRemoteTraceback: \r\n\"\"\"\r\nTraceback (most recent call last):\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/multiprocessing\/pool.py\", line 121, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 153, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 1486, in _map_single\r\n batch = self[i : i + batch_size]\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 1071, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/yisiang\/datasets\/src\/datasets\/arrow_dataset.py\", line 972, in _getitem\r\n data_subset = self._data.take(indices_array)\r\n File \"pyarrow\/table.pxi\", line 1145, in pyarrow.lib.Table.take\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/pyarrow\/compute.py\", line 268, in take\r\n return call_function('take', [data, indices], options)\r\n File \"pyarrow\/_compute.pyx\", line 298, in pyarrow._compute.call_function\r\n File \"pyarrow\/_compute.pyx\", line 192, in pyarrow._compute.Function.call\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays\r\n\"\"\"\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nArrowInvalid Traceback (most recent call last)\r\n in \r\n 30 owt = datasets.load_dataset('\/home\/yisiang\/datasets\/datasets\/openwebtext\/openwebtext.py', cache_dir='.\/datasets')['train']\r\n 31 print('load\/create data from OpenWebText Corpus for ELECTRA')\r\n---> 32 e_owt = ELECTRAProcessor(owt, apply_cleaning=False).map(cache_file_name=f\"electra_owt_{c.max_length}.arrow\")\r\n 33 dsets.append(e_owt)\r\n 34 \r\n\r\n~\/Reexamine_Attention\/electra_pytorch\/_utils\/utils.py in map(self, **kwargs)\r\n 126 writer_batch_size=10**4,\r\n 127 num_proc=num_proc,\r\n--> 128 **kwargs\r\n 129 )\r\n 130 \r\n\r\n~\/hugdatafast\/hugdatafast\/transform.py in my_map(self, *args, **kwargs)\r\n 21 if not cache_file_name.endswith('.arrow'): cache_file_name += '.arrow'\r\n 22 if '\/' not in cache_file_name: cache_file_name = os.path.join(self.cache_directory(), cache_file_name)\r\n---> 23 return self.map(*args, cache_file_name=cache_file_name, **kwargs)\r\n 24 \r\n 25 @patch\r\n\r\n~\/datasets\/src\/datasets\/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1285 logger.info(\"Spawning {} processes\".format(num_proc))\r\n 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard]\r\n-> 1287 transformed_shards = [r.get() for r in results]\r\n 1288 logger.info(\"Concatenating {} shards from multiprocessing\".format(num_proc))\r\n 1289 result = concatenate_datasets(transformed_shards)\r\n\r\n~\/datasets\/src\/datasets\/arrow_dataset.py in (.0)\r\n 1285 logger.info(\"Spawning {} processes\".format(num_proc))\r\n 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard]\r\n-> 1287 transformed_shards = [r.get() for r in results]\r\n 1288 logger.info(\"Concatenating {} shards from multiprocessing\".format(num_proc))\r\n 1289 result = concatenate_datasets(transformed_shards)\r\n\r\n~\/miniconda3\/envs\/ml\/lib\/python3.7\/multiprocessing\/pool.py in get(self, timeout)\r\n 655 return self._value\r\n 656 else:\r\n--> 657 raise self._value\r\n 658 \r\n 659 def _set(self, i, obj):\r\n\r\nArrowInvalid: offset overflow while concatenating arrays\r\n```\nFeel free to re-open if it still occurs","embeddings":[-0.4724798799,-0.5470885038,-0.0690055341,0.2029455453,-0.0922680721,0.0032601107,0.431145668,0.1213298067,0.0042691873,0.3337112367,0.2105832994,-0.0030280177,0.0025017336,0.0706486031,-0.2007900327,0.0564653911,0.1238672137,-0.1653056592,-0.1999184489,0.0210992452,-0.4086596668,0.0652992353,-0.2608809173,-0.0614106096,-0.0878032967,-0.1395558715,0.0227157604,-0.0221663248,-0.0390631594,-0.0816951841,0.1919379383,-0.0817265511,-0.112472944,0.3537240326,-0.0001111824,-0.0109583959,0.1242788509,0.0284189507,0.2034844458,0.1440587193,-0.0371055529,-0.1708365679,-0.2350585908,-0.3127315938,0.2882950306,0.0973996967,-0.1417520791,-0.2249159366,0.262106061,0.1953805238,0.2192112356,0.0814040527,-0.015035429,0.1142898053,-0.0161750019,-0.3034172058,0.0330700912,-0.2385724783,0.2039914578,-0.1968372017,-0.302244097,0.1477174759,-0.1354442239,0.1338905096,-0.1166878045,0.0154709024,0.343078047,-0.3428082764,0.2343281358,0.0401816517,-0.2495435923,0.0194753911,-0.2856802344,-0.4185943007,-0.2120327055,-0.2452063709,0.1186968908,0.3503252566,-0.1741090566,-0.1184589639,-0.7064749002,0.1826454848,-0.0528968424,0.0070800316,-0.3116375506,0.6008684039,-0.0026980031,0.2060533762,0.1709334999,0.1001436785,0.0688418746,-0.0227788482,-0.0647912994,0.1665035337,-0.1708871424,0.1595836431,0.4488881528,0.058397118,0.3597428501,-0.1830900162,-0.1618460417,-0.0563312434,0.2343739718,0.0594697259,0.3856415749,0.0190943126,-0.2259369195,0.0672159642,0.0536056608,0.074197188,-0.1498518586,-0.0234795306,0.1413926631,-0.4096342027,0.1917526573,0.225269258,0.0386171266,-0.0636864379,-0.0608621798,0.1051027328,0.1684311628,0.1251915991,0.0406736508,0.1396966577,0.0842215717,0.3683578372,-0.0960457772,0.1865689456,0.046105843,0.0762494057,-0.164629072,-0.2376046777,-0.1502536684,0.02442687,-0.0891724303,0.3796938062,0.2486175895,0.1886924207,0.2254287601,-0.1198965088,-0.0877317041,-0.2744494677,0.6435857415,0.1651964039,0.3063707352,0.2903899848,0.1194322109,0.2625783384,-0.1908123344,0.2612568438,-0.4289084077,-0.325552702,-0.0994048268,0.2534756958,-0.0046243309,0.3032531738,-0.3036207855,-0.0070459866,0.5798020959,-0.0259820241,0.2367868125,-0.2012345642,-0.3595503867,-0.071069032,0.0848903954,0.2256240398,-0.1289972067,0.2306214869,0.0653450936,0.2446785271,0.3458192945,0.3322739303,-0.273925513,0.2895392179,-0.2685263157,-0.0099180695,-0.0526836701,-0.4311921,-0.4203554988,0.1995908469,-0.1446069628,0.1888927966,0.067600131,-0.325679332,0.4037618041,-0.0403162017,0.4177243114,0.486638248,-0.3132050335,0.4952070117,-0.2428607047,-0.0260037761,-0.338896066,-0.0933043957,0.0249444526,-0.3816543818,-0.0841173306,-0.0041723531,0.3307575583,-0.0684263259,0.2112171799,0.4032656848,-0.0910114422,-0.0005261881,-0.0976596624,-0.1767158657,0.1018023938,0.1037141532,-0.2379125655,-0.0061032944,-0.1494359225,-0.1178122908,0.1189868525,0.2392622083,-0.465526104,-0.0865261182,0.2350709587,-0.1330577731,0.2071367949,-0.1156043336,0.3647233248,0.0955672786,-0.0473960899,-0.1623814106,-0.0440338589,-0.1695727557,-0.1972923279,-0.256564945,-0.1027143374,0.277923882,0.1670152098,-0.0678443238,-0.1385055035,0.3817984164,0.0144724855,-0.1126673967,0.0447344035,-0.0673192292,0.0487594195,-0.0025495754,0.0887556002,-0.1664895415,-0.1406800449,-0.0731712207,-0.0806001648,0.4928908944,-0.0429759696,0.450468123,0.1013248265,-0.0087925028,0.3333130181,0.0147212017,0.0952108949,-0.108680509,0.1052466109,0.1580382138,-0.1132940501,-0.1801329553,0.0783963725,0.2769286036,0.640108645,0.1786997169,-0.1603690535,-0.0665042922,-0.316055119,-0.1899700761,0.1805679351,0.0096732127,0.2360683829,0.3596695662,0.1982408315,-0.1609237045,0.0647445619,-0.0289550051,0.2104812413,0.1276272237,0.3061177135,0.1302338988,0.5117682219,-0.0963642895,-0.0896140933,-0.061671,0.0016787803,0.0674070418,-0.1021869108,0.2256777734,0.0637503341,0.0345311277,-0.342966795,-0.2171401083,-0.1741497815,-0.6312443614,-0.0058589131,0.2816341817,-0.1140101403,0.3577982783,0.1590273827,0.0000863484,0.3099007905,-0.3061830997,0.1523242146,-0.2501434088,-0.1402078122,0.0344589762,0.2034917474,-0.4103622437,0.2497137338,0.2418713868,-0.0073505361,-0.1619130671,-0.1759203672,-0.1180872172,0.0211110394,0.1095370501,0.0799917728,0.0196468625,0.0081992215,-0.358101666,0.1158828139,-0.2884114385,-0.155771181,0.0917602256,-0.1746366918,-0.1932019889,-0.2032632232,0.0032963841,-0.0095923301,-0.3650891185,0.1923043281,0.0071315547,0.1093452275,-0.0129435258,0.0516555198,-0.1818473786,-0.0121176485,-0.2586720884,-0.2050638199,-0.3385683298,0.2415130436,-0.240699172,-0.1149851307,-0.1148272604,0.0164141431,0.3596509099,0.2422824502,-0.2953150868,-0.0852859169,-0.0846701711,0.5255175829,0.0984388217,0.123869054,0.2237486243,0.1018691361,-0.1512203515,-0.127075538,-0.0096969903,0.1674998403,-0.1645011157,0.0745369568,-0.0171891805,0.1135187075,0.1911734641,0.8231935501,0.3125421107,0.0323668718,0.0461599529,-0.052046705,-0.0774144232,-0.0887126774,-0.465357244,0.155716002,0.0458407663,0.0824740678,0.1671870202,-0.0260663796,-0.3475735486,0.1483848393,-0.0090955412,-0.3566475213,-0.2947732508,0.3027471602,-0.109523721,0.1417776048,-0.0641486421,-0.1482805163,-0.496786803,-0.155815348,-0.1496390402,-0.3124412894,0.002707453,-0.0434226803,0.2983742356,-0.1023147926,-0.539770484,0.2341855168,-0.0011001605,0.1348960102,0.220674932,0.0508665182,-0.0949329287,-0.2084873766,0.8502280712,-0.2923719883,0.0277640894,-0.1798008233,0.0371931717,-0.4001259208,-0.1800974011,-0.2615940869,0.6538347006,0.691722393,0.4760460556,-0.1589214355,0.0430680811,0.3724712133,0.0677732229,-0.1501819342,-0.0856579468,-0.1202186197,-0.2358566374,-0.0233640037,0.3128212094,0.1812190115,0.3029714823,-0.0154260341,-0.2447441667,-0.0116983159,0.280726552,-0.2219143212,0.1512068212,0.0883365422,-0.1519748122,0.1290318668,-0.2227023095,0.2646117806,0.3021807969,0.1920861155,-0.2047645003,-0.2822438776,-0.1276487559,-0.2989731431,0.2527008057,0.214343071,0.0695905909,0.0509049296,0.1981263012,0.5551554561,-0.1061234549,0.0026973302,0.0856717229,0.3002865016,-0.1544084102,-0.2417069077,0.083387658,0.217290163,0.0350439325,0.2689328492,-0.2343525887,-0.416765511,0.1570390165,0.1068356484,0.5609591603,-0.090178743,-0.0360853747,-0.0071783587,0.1886250675,0.1050648317,-0.1770541072,0.2332184017,-0.1435902119,0.1433256716,0.156323567,-0.0471651666,0.0099960323,0.0531223454,-0.2045133859,0.2653582394,0.1411941946,-0.2298077047,-0.367251873,-0.0316803716,0.2411786616,-0.3824780285,-0.2204782218,0.1157284379,0.2500836253,-0.2820232511,-0.0830067247,-0.008922047,0.0639514625,-0.3150451481,-0.4909215569,0.0008943704,-0.4151181579,0.1651987284,-0.2484632879,-0.1469291896,-0.0685079843,0.1301172227,0.1298281401,0.0829190686,-0.0241292026,-0.049858924,0.0876322165,-0.1707328707,-0.0902273655,-0.0541367233,0.1428706199,-0.0124208014,-0.4450484216,-0.0248871818,-0.007269552,-0.0792046115,0.1262294799,0.367962122,0.1421956867,-0.14552176,-0.0489516668,-0.0944611728,-0.1117519587,-0.2680504322,0.0995053202,-0.1385320872,-0.4535485804,0.3498203754,-0.3295318782,-0.3506829441,-0.1023664623,0.3144615293,0.0909861401,0.0328479297,0.2585543096,0.2493997514,-0.1750436425,-0.1751388311,-0.0500649922,-0.0394997112,-0.2380501628,-0.0654175505,-0.0486120917,0.0547090732,0.1518817395,-0.0020762542,0.1433032453,0.1593454033,-0.3186149299,-0.285115093,-0.1217942685,-0.1914694607,-0.3388714194,-0.1731633395,0.440436244,0.1156077236,0.1084085181,0.2029795051,-0.3412675858,0.0495141074,-0.278188765,0.2500203848,-0.1471015811,-0.1701536626,0.1120382249,0.0142813195,0.1152472794,0.0972592086,-0.0491243675,-0.311406076,-0.1150640175,0.0383484252,0.0270475671,-0.2261347622,0.1507647783,0.1011740863,-0.0518595167,-0.3778333664,0.3535485268,-0.007179637,0.0689388812,0.0358417593,0.2494354546,0.0135658802,-0.090134427,-0.0148738995,0.0916951969,0.1114624292,-0.0135054663,0.512996614,0.1663272828,-0.1833403707,0.1983282268,-0.0952807665,0.1587287635,0.1796399951,0.2281201333,-0.2174176574,-0.2915329933,0.448148489,0.2858531475,0.1493940353,-0.1553090513,-0.2026843578,-0.0309864935,0.2425784916,-0.1084846631,-0.0957228094,0.3049129546,-0.1626545191,0.3035263121,0.4112124443,0.1383127421,-0.4035736918,0.1061665788,0.1432548314,0.3482195735,-0.3618722558,0.0650743023,0.2460177392,0.0900607184,0.1868119091,0.3765761554,0.1712119728,0.4348614216,0.5098477602,0.0847004279,0.3632257581,-0.0262439065,-0.4669933617,0.0102180159,-0.4753915966,0.0066250828,-0.0887213871,-0.08602532,0.3487884402,0.1319974065,0.1310573816,-0.0957048759,-0.1422550678,-0.241482541,0.2645242214,-0.3801378608,-0.1171856821,-0.1314582676,-0.0642014593,0.2291437685,0.0430430397,0.1383484453,0.2140901834,0.7544375062,-0.0522832423,-0.0296985153,-0.1664370447,-0.0638668463,-0.0322883353,0.1268276572,-0.2434642464,0.3029705286,0.2656693161,0.0155309532,-0.3197233975,0.3531139791,0.6688971519,0.3504451811,-0.0052596037,0.1054035276,0.1625346392,-0.163266331,-0.5884806514,0.0424423143,0.0139992479,0.3201362193,0.2330350429,0.2186632603,-0.2325776964,0.2933358848,0.3576749861,0.001839252,0.2867113054,-0.2394672483,0.0288397577,-0.3740426004,0.0247320458,0.0591021441,-0.1677625179,0.1108387858,0.273778528,-0.0377502441,0.3452721834,-0.0825715959,0.0966709778,0.0151069053,0.5231214762,0.5577062964,0.1013848782,-0.2210204154,-0.1186115891,-0.2194136977,0.3260878026,-0.4987770021,0.016561538,0.0156782158,0.2855914235,-0.1411339343,0.0241953731,0.4553402662,-0.3068294525,0.007509788,0.4098079801,-0.6726940274,-0.1202630773,0.2867000699,-0.3018577695,0.2044043988,-0.347600311,0.2213242501,0.1684019119,0.1794318855,-0.0864701048,0.3014255464,-0.0357852839,-0.0147195021,0.1236578748,0.503100574,0.1960096508,-0.2836230993,0.0266222022,-0.2694171369,0.1464063376,-0.3676693738,0.1431274861,0.0292756706,0.1391620338,-0.2626158595,-0.5490094423,-0.0643850043,-0.0190883167,0.3117630184,-0.0982327834,0.0152621511,0.1640101522,0.0423423275,-0.0885275081,-0.025079364,0.2250818461,-0.0628835484,0.1279391944,-0.1627910435,-0.1108391136,-0.0364003181,-0.2888792455,-0.3803110123,-0.2436331213,0.2614288628,-0.3879218102,-0.0652119741,-0.4406741858,0.0524187647,0.2756070495,0.1199200898,-0.4911474288,0.203205213,-0.2661342919,-0.0749607533,-0.044761885,-0.0224792548,0.0614942983,-0.5596327782,0.1373844147,-0.0810561255]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/647","title":"Cannot download dataset_info.json","comments":"Thanks for reporting !\r\nWe should add support for servers without internet connection indeed\r\nI'll do that early next week","body":"I am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/text\/default-53ee3045f07ba8ca\/0.0.0\/dataset_info.json\r\n```\r\n\r\nI tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0\r\nTensorFlow version 2.3.0\r\ndatasets version: 1.0.1 \r\n","comment_length":20,"text":"Cannot download dataset_info.json\nI am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/text\/default-53ee3045f07ba8ca\/0.0.0\/dataset_info.json\r\n```\r\n\r\nI tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0\r\nTensorFlow version 2.3.0\r\ndatasets version: 1.0.1 \r\n\nThanks for reporting !\r\nWe should add support for servers without internet connection indeed\r\nI'll do that early next week","embeddings":[-0.2581685483,0.0243737306,-0.0594887361,0.2031503767,0.0788523629,0.1288594604,0.0959735289,0.2360109985,0.1991672516,0.0795687139,0.1454199255,0.2383177727,0.2438558787,0.1896512508,0.1779324859,-0.1044934914,-0.1757180691,0.0602590367,0.0163063873,0.1639957577,-0.2036467642,0.1125223488,-0.0277882684,0.0930874273,-0.1486055255,-0.329529494,0.1726283729,0.1674591303,-0.1995363235,-0.0480165631,0.3676278591,0.1111577004,0.0231310483,0.1800840497,-0.0001172042,0.2482485324,0.3314515352,-0.0468823798,-0.3855315447,-0.6151015162,-0.2775319219,-0.3104656935,0.1326969713,-0.2921256423,0.0545545891,-0.0274387151,0.3667286336,-0.2455900609,0.3066195846,0.3828884363,0.081887275,0.1975744069,0.1668199748,-0.0151023362,-0.079669036,-0.156904757,0.1030239537,0.2234270424,0.0778491721,0.0222973935,0.1356130689,0.2242699414,-0.1415863484,0.1933998466,0.3185082078,0.0770921484,-0.2284037173,-0.2921458483,0.37315467,0.435272038,0.7340177894,-0.258764267,-0.3504030406,-0.0941050202,0.0685781986,0.0049100188,0.4181962609,0.3318796158,-0.2674552202,0.053244397,-0.3854344189,-0.4756149352,-0.5133515596,0.2596521974,-0.043795459,-0.0062728175,-0.0868521333,0.0650681183,0.1307830065,0.0529989488,0.1102165133,-0.0586893819,0.0596241057,0.2255720496,-0.1393045783,0.0642871931,-0.0231556147,-0.3781893849,0.0988889113,0.237310186,0.1910169274,0.0885317624,-0.2979581356,0.2700930536,0.2438108176,-0.019225426,-0.0753265992,0.0507073924,0.3943025768,0.1891427487,0.2085787952,-0.1946888864,-0.0591314472,-0.2078119069,-0.1013944671,0.0001048631,0.2709664404,-0.0075751944,-0.2217639685,0.1743744761,-0.2492113262,0.0577885956,-0.0308442898,0.2868508399,-0.4215427637,0.1188231632,0.3688768148,0.0901182219,-0.0124301109,-0.2255284041,-0.0405955687,0.0507175401,0.1518693566,0.1059540287,0.202420935,-0.0358267985,0.3656516075,-0.0974892154,-0.0094232261,-0.1456918865,0.3619759381,-0.0676196739,-0.3675275445,0.4784257114,0.3598552644,0.0498910509,-0.0304135587,0.0404353812,-0.0042618262,0.1028195098,-0.5772803426,-0.4673503637,-0.1304803342,0.0880941376,-0.1737635732,0.014178697,-0.3551717103,-0.0608476661,-0.2671119869,-0.279756099,-0.1031838655,0.0695324615,-0.167597428,-0.2041427195,0.3580983579,0.3052932918,-0.4322125018,-0.0583983511,-0.0573551022,-0.2180733234,0.0616178811,0.0804652497,-0.2448321134,0.2876029611,-0.1839329302,0.0781300366,0.4764019549,-0.3475132287,-0.7220580578,0.5315420628,-0.2805051804,-0.065564014,0.0445309468,0.209490478,0.1288461387,0.2588056922,0.4079203606,0.4810492992,0.0635359958,-0.0915062875,0.0023360976,-0.3649243712,0.0945882797,0.3108660579,0.0989390314,0.2225619704,0.1619277745,-0.0530178435,0.2041986287,0.0445404164,0.0700532272,0.4481004179,-0.0124171944,0.0885310099,-0.1026969478,-0.1130253449,-0.6764413714,0.1161552668,0.000277022,-0.0813723281,-0.6097717285,-0.0802460089,-0.2205836326,-0.0389364474,0.0883240774,0.2622841597,0.0493855998,0.049601417,0.1081951037,0.1324227154,-0.1000124738,0.4082681239,-0.3110571802,0.1138603836,-0.3475219905,0.1198774353,0.1842234731,0.1942978352,0.1091840714,-0.1518645138,0.0917049199,-0.0541608594,-0.1490963995,0.4382841885,-0.2289565802,0.4687093794,-0.0645764321,0.321454525,0.079349108,-0.0825936943,-0.0710100234,0.0418956243,0.1006268784,-0.0280569661,-0.288522929,0.3034878671,0.197091952,0.2505658567,0.0842370316,0.2263906747,0.2398903221,0.0106029343,-0.1017640233,0.3073266745,0.1692443341,0.1132228673,0.2009826452,-0.1750233918,-0.5216665268,0.1266985089,0.375092864,-0.1219119057,0.1262109131,0.1015705168,-0.1663372666,-0.0890585035,0.1078620851,0.1592134982,0.1848902404,-0.0357064307,-0.0391115434,0.1988347322,-0.0235857163,-0.243300423,-0.0912044868,0.0482785739,0.2365596443,0.0827148408,0.0179685745,-0.0010185915,-0.0160230845,-0.2249884307,0.1684878618,0.2603356838,-0.2508514822,0.2482616603,-0.2147570997,-0.1431507021,0.112368755,0.2177840173,-0.2780341208,-0.1969718337,-0.1126687303,0.4512029886,-0.0530626029,-0.2184859067,-0.2282786816,0.3127138615,0.1134808213,-0.5957485437,-0.0996805057,-0.1528220028,-0.0897533819,0.0087622888,0.1686343253,0.144774437,0.2064512968,-0.1780374944,0.0105940411,-0.4024783671,0.109743543,0.056042172,0.1726340353,0.1398850381,0.1356797367,0.6456850767,-0.1096475273,0.0676192641,0.1128703505,-0.0828174055,-0.1381504536,-0.0323999189,-0.1608317643,0.1464507878,-0.0544036254,-0.5343643427,-0.4285780787,-0.2995085716,0.4056402743,-0.0244592633,0.279620856,0.0601893328,0.0293879211,0.2110872567,0.3687650859,0.1299320012,-0.0944866166,-0.6934444308,0.3646911085,-0.1951303035,-0.3631881475,0.2532722354,0.064514786,0.159446761,0.4275206923,-0.6492444277,-0.1355833262,-0.0504805855,0.2179579884,-0.2815863192,-0.0338860489,0.4939148724,-0.202109769,0.070963569,-0.0619993955,-0.0312898643,0.0147357816,0.1017191708,0.2120196521,0.4684890807,0.538824141,0.1260890961,0.6983602047,-0.1538118869,0.0850010961,-0.0224630684,-0.3815024197,0.2022925913,0.1190600768,-0.122126326,-0.1462517977,-0.0883389637,-0.1300351471,-0.2253789753,-0.0755703896,0.1261390597,-0.3344646692,-0.1559286118,-0.4494193494,-0.281136781,0.1039604619,-0.0513807684,0.2389106452,0.2231043875,0.2089971602,-0.4047694206,-0.2367788255,-0.004999449,0.3802207708,0.1709281057,0.2029417753,-0.1041326821,-0.1379258484,-0.4995948076,0.431199193,0.0101294694,0.4071515501,-0.2748411,0.1849469692,0.0598673522,-0.0574491322,0.4644336402,-0.308701992,0.322222352,0.0798899755,-0.0434507616,-0.2367433906,0.0476072393,-0.1010697335,0.3014736772,0.4305861592,0.4523818493,-0.1853969097,-0.1183361039,-0.083656244,-0.041239731,-0.2490082532,-0.0373986736,-0.1405754387,-0.5846740603,-0.2606703043,-0.1465644687,0.0063717482,0.3492481709,0.0999024957,0.206900537,-0.058507096,0.2324883938,-0.0146720503,0.0018995011,-0.0240557194,0.1483072788,-0.1793149412,0.4764508307,0.2076442987,0.2092899531,0.6061328053,0.3182346523,-0.3097948134,0.0760813281,0.0074530016,0.0707541332,0.1588330716,-0.1001965478,-0.0475747325,0.5560957193,0.1700891256,-0.2265905291,0.216372475,0.2004491687,-0.0449656621,-0.2611289024,-0.5889946222,0.5028380156,0.1020913422,-0.0888957232,0.3096010685,-0.0867455155,-0.0259941202,-0.0179137401,-0.2588960528,0.7463462353,-0.0078486213,0.102185607,0.2200823128,0.1290268153,0.7392458916,-0.8184087873,0.4260359406,0.0502768047,-0.2318521738,-0.2800000608,-0.1687984318,0.042793978,0.2465156913,0.135396719,0.3537701368,0.0492912307,0.139837265,0.2809302211,0.327611208,-0.3123631477,-0.1970204413,-0.1295176297,0.0276071168,-0.2567071021,0.5734645724,-0.1012451649,-0.3177197576,-0.0105634443,-0.2015152127,-0.2205113918,0.2966547012,-0.3537477851,0.1829039603,-0.0251926575,0.0123269791,-0.1197537482,0.1188768819,0.0307789892,0.2270126641,-0.4064325094,0.0962214172,-0.3807075322,-0.2729485035,0.0600299537,0.1930928379,0.143187359,-0.1488229781,-0.3436147571,0.3103336096,-0.2093212754,-0.5062012672,0.3441852331,-0.1510679424,-0.3313428164,0.0829869211,-0.1414981633,-0.1066786498,-0.0420585275,0.013935349,0.0564835444,0.045998767,0.0307925064,-0.0586123578,0.1070291847,-0.1867025942,0.0720015019,0.6140071154,-0.0818191022,-0.1576109678,0.4818258584,0.1507883668,-0.0686292127,-0.147622928,0.0328451507,-0.0315195024,0.133966282,0.0476742238,-0.0774004459,0.4039573073,-0.3512095213,-0.0243556276,0.1211342514,0.0978179798,0.0288174879,-0.8456358314,-0.2271311581,0.2395487428,0.0572064035,0.0895522013,0.2577364743,-0.0146432556,0.1545117348,-0.1004174948,-0.1987087578,0.1689089537,-0.2931167185,-0.2407388389,0.443769753,0.2284310162,0.5434686542,-0.0682906434,-0.0336170197,-0.0724586621,-0.1338688135,-0.0819213018,-0.0315986834,0.1933619976,0.0343477428,-0.1537769437,-0.040189147,-0.5816941261,-0.238419205,0.2793464661,0.2583547533,0.1257843971,-0.0541532747,0.089436233,-0.3515660763,0.4427090585,-0.4098797739,0.2000609785,0.0118042221,0.6186724901,-0.2902218997,0.1899168789,0.0933985785,0.0799570233,-0.3963117301,-0.1322663873,0.0644161105,-0.1562742889,0.4741695821,-0.3039165139,0.2632939517,-0.1518127024,0.1123898327,0.1734662503,-0.2352962643,0.2265852839,0.1939248443,0.1053431109,-0.0775514022,-0.0710766688,0.3126709163,0.1430843771,0.0582971238,-0.0590570718,0.0815955848,-0.0269315932,-0.0752381384,-0.0132576488,0.3911983967,-0.219027251,0.1376026571,0.2432943285,-0.0124951685,0.0229650214,-0.0299673248,0.3261116743,0.0983023196,0.2900646627,-0.0794334188,-0.1628965437,0.0581233762,0.1574460715,-0.1831638366,-0.5663151741,-0.2402624041,0.1602426767,-0.3456601501,0.0673189312,-0.322435528,0.1991874278,-0.2051934153,0.0599440448,-0.405063957,0.2330885679,0.0006373553,-0.079073824,-0.4354059398,-0.2118585706,0.0559327155,0.2156287432,0.3409397006,-0.283557117,0.0105213383,0.0513613001,-0.0588661358,-0.087222673,0.036454238,0.1461372823,0.0961385146,-0.2760255039,-0.0853729844,-0.1579366475,-0.0975599512,0.1521982253,0.2372897714,0.2586217821,0.1005764678,-0.0636844933,0.3023335338,0.0530363135,-0.1611156315,0.0035622476,0.0596504509,0.4132117033,-0.4068720937,0.224431932,0.1159294993,-0.0663938448,0.4338053167,-0.0910516381,0.4080499113,-0.1714952141,0.1969529837,-0.1435177475,-0.0548874922,-0.2223404944,-0.1979023963,-0.4185056388,-0.3149419725,0.3104333878,0.0740805864,0.3062132597,-0.0984376818,0.0051385919,-0.2273651809,0.5094874501,0.3295617402,0.3142699003,-0.0078130048,-0.0333033279,-0.4118765295,-0.0295041539,-0.2248632908,-0.2211930305,0.1552860886,0.2018475682,-0.1512493044,0.1393018216,-0.1522100568,-0.073768504,-0.2099644542,0.1152875796,-0.1031565741,-0.306838423,-0.1324062049,0.1277769953,-0.0125604942,-0.3340646029,0.0885634497,-0.239780575,-0.0091690188,-0.1310245991,0.2173588872,-0.1317515671,0.101334393,0.4135287702,0.0147082638,0.4208114147,0.0513823889,-0.1141483784,-0.2717923522,-0.123175621,-0.2075553238,0.3100585938,0.1311533749,0.2213362902,-0.0996847749,-0.0406367332,-0.3142709434,0.1450714916,-0.1366638243,-0.0818043277,-0.160169825,-0.1002270728,-0.3212528527,0.2325598598,0.0840072334,-0.0241560955,-0.0681619793,-0.060340371,-0.2181842625,-0.0449179336,0.3933365941,-0.4936444163,-0.1012937352,-0.0289594103,0.185354203,-0.0953275114,0.1094550267,-0.2316055596,0.256243974,0.3378905058,-0.1461079121,-0.1566697657,-0.0595967621,-0.1753625572,-0.023731526,-0.0522246137,0.3484819829,-0.0584433042,-0.3829176128,-0.2968147993,-0.2815043926]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/647","title":"Cannot download dataset_info.json","comments":"Thanks, @lhoestq !\r\nPlease let me know when it is available. ","body":"I am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/text\/default-53ee3045f07ba8ca\/0.0.0\/dataset_info.json\r\n```\r\n\r\nI tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0\r\nTensorFlow version 2.3.0\r\ndatasets version: 1.0.1 \r\n","comment_length":11,"text":"Cannot download dataset_info.json\nI am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/text\/default-53ee3045f07ba8ca\/0.0.0\/dataset_info.json\r\n```\r\n\r\nI tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0\r\nTensorFlow version 2.3.0\r\ndatasets version: 1.0.1 \r\n\nThanks, @lhoestq !\r\nPlease let me know when it is available. ","embeddings":[-0.2836295366,-0.0356300995,-0.0750728026,0.2556798756,0.0749811083,0.1890455335,0.0522874184,0.2726956904,0.1538558006,0.0423615798,0.0565663129,0.2412359565,0.2468585223,0.1747367084,0.1874744445,-0.101119481,-0.208222717,-0.0152274324,0.0645620823,0.157960549,-0.1781320274,0.1768838614,-0.0382678732,0.1482433081,-0.1226924881,-0.300986588,0.1681018323,0.1378388256,-0.2036334872,-0.0502884462,0.3798254728,0.161773473,0.0586842746,0.1903471798,-0.0001162467,0.1944275051,0.3638163209,-0.0457046367,-0.381113559,-0.5967131257,-0.2494823933,-0.2973385155,0.1100192666,-0.2545582652,0.0731036887,-0.0997057259,0.3743366003,-0.2272342294,0.3153138459,0.3551135957,0.0908752605,0.1888761073,0.2049290538,-0.0277236011,-0.1030633748,-0.158737883,0.1508115232,0.2408153415,0.0700130686,0.0560199954,0.1360577941,0.2326438427,-0.121210061,0.2288177013,0.3147958219,0.0585186034,-0.2370508611,-0.258538276,0.4050700665,0.3839431107,0.7906491756,-0.2954023778,-0.3635589778,-0.0948560387,0.1129910648,-0.0083139101,0.4411802888,0.3437858224,-0.2761365771,0.0892968029,-0.3824490309,-0.4619383216,-0.5123433471,0.2484106272,-0.0504990779,-0.0829124004,-0.1134352759,0.0857232511,0.1864816546,-0.025770247,0.1857078373,-0.0318393968,0.0397249907,0.2290727347,-0.1832458675,0.0347999707,-0.0326584838,-0.3842658997,0.1213439479,0.1418446898,0.1805281639,0.086837627,-0.2933966517,0.291313678,0.2753976285,-0.0066163563,-0.0215257704,0.058890231,0.389688611,0.1760313064,0.2150290161,-0.1284320503,-0.1696780026,-0.2375366688,-0.081137307,-0.0169924889,0.2089046687,-0.0523862317,-0.293810159,0.1504876465,-0.2393262237,0.0182672217,0.0342030227,0.3381666243,-0.4359728694,0.0925228968,0.3580030203,0.0560295098,-0.0136811147,-0.1921963841,-0.0709482804,0.1126947775,0.0802725852,0.0851062462,0.2117019147,-0.0864717513,0.4037905633,-0.1432365775,-0.0240208115,-0.1649747491,0.3284343183,-0.0151170734,-0.4092355967,0.4660094082,0.3460184336,0.0941016674,-0.0922363251,0.0900407284,-0.0434597731,0.0956879705,-0.6135832071,-0.4312179983,-0.1211588234,0.0901190341,-0.1764487028,0.0328171328,-0.3972243369,-0.007526828,-0.2347230017,-0.2192370743,-0.1600467265,0.0038607509,-0.1636095345,-0.2463885248,0.3575221002,0.257029146,-0.4023265839,-0.0696031228,-0.0474545881,-0.1977138817,0.0395566635,0.1288488656,-0.2320835739,0.2186368555,-0.1934359968,0.0731985867,0.4759847522,-0.3694679737,-0.7223531604,0.472122699,-0.2550041378,-0.0980597287,0.0223274976,0.1902827173,0.069220975,0.2175524235,0.4344026446,0.5116984248,0.0490753092,-0.0971170813,0.0530255809,-0.3458020687,0.1375584006,0.2808712125,0.0005440955,0.1744597703,0.1421422064,0.0225911941,0.2160161585,0.051793851,0.0195316002,0.4409712851,-0.0089791566,0.074004285,-0.0913217217,-0.1263404638,-0.6730142832,0.1114275455,-0.0014205093,-0.225384891,-0.5628833771,-0.0616316088,-0.234390378,0.0206484739,0.0844853893,0.2007820308,0.0297161415,0.0952307954,0.179576695,0.1317225546,-0.0530734695,0.4545894265,-0.3153584301,0.1088761091,-0.3654281497,0.101332441,0.1997534186,0.1597425789,0.1662970632,-0.1006294563,0.0965397805,-0.0413076244,-0.1586725563,0.4500661492,-0.1536281258,0.4840446115,-0.0905861109,0.3049650788,0.0688972548,-0.0710703954,-0.0880602896,0.035084337,0.0187067911,0.0282035321,-0.3311638832,0.2993258834,0.2129252702,0.1538819671,0.0779213682,0.2333689183,0.2474469841,0.0309353732,-0.0864943415,0.3248325884,0.2136930674,0.1130780578,0.1974008977,-0.170709461,-0.5117841959,0.1430127919,0.4671474695,-0.1291960031,0.11435166,0.1440051794,-0.1887921989,-0.1393389553,0.1351850778,0.1753605604,0.1681917757,-0.0402290709,-0.0020954241,0.1290010065,-0.0566335805,-0.2142564207,-0.1114633903,0.0454943031,0.2592899799,0.0516564287,0.0680971891,0.0314497054,0.0106703481,-0.1637408286,0.1722943187,0.286379993,-0.2216579765,0.2577764988,-0.2242257893,-0.1017722115,0.068988651,0.2004743814,-0.2133779079,-0.1950456202,-0.1385336369,0.4860543013,0.0094758365,-0.2081688792,-0.1270845681,0.2987563312,0.1749956161,-0.5075445175,-0.0682964325,-0.0743432119,-0.1001739651,0.0019296136,0.1753407568,0.1013273597,0.2025064528,-0.1633908153,0.0673845932,-0.4128347337,0.0508188084,0.0573845245,0.1074214876,0.1554656774,0.1224716604,0.6201350093,-0.1053347141,-0.0022546549,0.0797243118,-0.1145559549,-0.1910119802,-0.0092532076,-0.1616915911,0.0794622153,0.0008136305,-0.4796889424,-0.4233120084,-0.3284586072,0.4593297243,-0.0458109193,0.2781172097,0.1093575582,-0.0119664716,0.2369737625,0.3646008074,0.1337388754,-0.1329399645,-0.6716684103,0.3766463995,-0.2624776661,-0.3970847428,0.1999997646,0.0810077935,0.1948112994,0.3747974336,-0.6048536301,-0.1300581098,-0.0520394631,0.1969448924,-0.2438790053,-0.0964714512,0.4335421324,-0.2252730876,0.0764248297,-0.0648086891,-0.1107157841,-0.0388915055,0.1485441923,0.2989765704,0.4727897048,0.489569813,0.1448770463,0.6489873528,-0.0566234961,-0.0086390935,-0.1140116602,-0.3621212244,0.126843363,0.167518273,-0.1699923575,-0.1445105374,-0.1061833203,-0.0862261504,-0.2185520977,-0.041770637,0.0887539834,-0.3020243943,-0.1280708462,-0.4389785528,-0.2546083927,0.0641142055,-0.0268537421,0.2325666547,0.1957739741,0.1866418123,-0.3734816313,-0.2427474111,0.0483100712,0.4246258736,0.1675929427,0.214969933,-0.1105742976,-0.1610172242,-0.5150237679,0.4450336397,-0.0464642532,0.3830299079,-0.2712254524,0.1927068084,0.104081437,-0.0749188289,0.5322337151,-0.2471116483,0.3469021618,0.019832164,-0.0296864454,-0.2515310943,0.0337256752,-0.0819581002,0.3308089375,0.4097300172,0.4240723252,-0.2173622102,-0.1461220831,-0.0376570895,-0.073057428,-0.2392347157,-0.05474158,-0.0902106762,-0.598169446,-0.2315344959,-0.2181072831,-0.0023780342,0.366640687,0.1303758174,0.1817200482,-0.0092580877,0.1933148056,0.0292195864,0.0117569901,-0.0070874486,0.2523843646,-0.1190272197,0.5460115075,0.1633838117,0.2022974938,0.6163425446,0.3051150441,-0.3728390634,0.1302846372,0.0481985211,0.0467651077,0.1003395692,-0.1750537306,-0.0257714428,0.4817495048,0.1339652538,-0.2277855575,0.2495918423,0.2406752408,0.0557207279,-0.2854783833,-0.6628177166,0.5089302659,0.079274945,-0.0980500802,0.3745701015,-0.1732532382,-0.0335143469,-0.0820529014,-0.2349797785,0.7159511447,-0.0937155485,0.0847724676,0.2268239707,0.0900217369,0.7099677324,-0.7943526506,0.4071442485,0.0305419415,-0.1623077393,-0.24813658,-0.1892915219,0.007821192,0.1695759445,0.1172313169,0.3078171611,0.1036105976,0.090541862,0.2692047656,0.419803977,-0.2725930214,-0.2063281536,-0.1654496193,0.0340438262,-0.3061686754,0.6217873096,-0.0530733094,-0.2730752528,0.0024951643,-0.1880281121,-0.2684150636,0.2410769612,-0.3396399915,0.25862813,-0.0061616781,-0.0796877965,-0.0739648864,0.1670385748,0.0283032674,0.2480962276,-0.403498739,0.1187100857,-0.37584126,-0.2939452529,0.0162975546,0.1462459713,0.1617604792,-0.1191487312,-0.2621125579,0.3097277284,-0.1070335135,-0.4919310808,0.2495809495,-0.145366028,-0.3048224449,0.0582894944,-0.1670315564,-0.1560491771,-0.1022773534,0.0433412679,0.0640314817,0.1126664728,0.0135199968,-0.0947747454,0.1139880642,-0.2039031237,0.0832241997,0.6080782413,-0.0589920841,-0.2333693951,0.5051184893,0.2048563361,-0.0648998618,-0.1277886182,-0.0117533039,0.0278691966,0.11033573,0.0088298647,-0.1190105155,0.3937375844,-0.3307348788,-0.0311474819,0.1413563043,0.023364041,0.0483698994,-0.8531724811,-0.253418237,0.2518245578,-0.008506285,0.114335537,0.2585679591,-0.0052972967,0.1121954024,-0.0558224171,-0.1969401389,0.1902649105,-0.2862555087,-0.2668044567,0.4553706646,0.2033553123,0.4987522066,-0.0187186413,-0.0267250929,-0.0563605279,-0.1670300066,-0.0663020685,-0.0981195197,0.1742230505,-0.0383925922,-0.1077263728,-0.0227834061,-0.595041573,-0.2270495743,0.2654409707,0.2656193376,0.1110570729,-0.0221053325,0.080348976,-0.3226752877,0.3380586803,-0.3448231816,0.2309819311,0.0411728993,0.5679775476,-0.2364568561,0.2201995254,0.0658860803,0.083639048,-0.4579710066,-0.1075477079,0.0463718548,-0.1774784774,0.4183572531,-0.2823946178,0.2595091164,-0.1754348278,0.1445869654,0.167454049,-0.2776069641,0.259172678,0.202020362,0.1054273322,-0.0678030923,-0.062700592,0.2681320906,0.059528403,-0.0339077525,-0.0177710969,0.0653823018,0.0741181597,-0.015342271,0.0344374329,0.4408204257,-0.2159710526,0.1089849994,0.2610510588,-0.0354504213,-0.0055050696,0.1022592187,0.37000525,0.0777910277,0.2805823684,-0.0342408307,-0.1504220814,0.0459386855,0.2147318274,-0.1329783797,-0.5768351555,-0.2581652701,0.162785545,-0.3690959811,0.1113963351,-0.2881442308,0.2386015803,-0.1748694479,0.0703686103,-0.3694103062,0.2531827986,-0.0194758009,-0.1326214969,-0.3300653398,-0.2338944077,0.0543714464,0.2013640404,0.3146101534,-0.3420751989,0.0198477209,0.0372558795,-0.0719688907,-0.1286867559,-0.0005833097,0.1860326827,0.0834481344,-0.298102051,-0.00710834,-0.2141523361,-0.0928767622,0.1437911093,0.1907740682,0.2920002341,0.0961105525,-0.0870069265,0.2963104248,0.004520582,-0.1576397866,0.0260365643,0.1132984459,0.3959763944,-0.4396766126,0.2458663583,0.1248684004,-0.0603563562,0.3807980418,-0.0588941202,0.4138514996,-0.1763534248,0.1871843487,-0.1356431246,-0.0542172492,-0.2197551429,-0.1501857638,-0.4087274671,-0.3268427551,0.3152669668,0.0977832675,0.2617763877,-0.0886970237,0.0133153098,-0.2075452954,0.4826298952,0.3682804108,0.3439180255,-0.0290631671,-0.0546230897,-0.4547855258,0.0061435769,-0.1834209412,-0.2308604121,0.1736797988,0.1868555844,-0.1689534783,0.1015363634,-0.0954168066,-0.0731583536,-0.1501572877,0.109169513,-0.1188473105,-0.3026703,-0.0785481185,0.1185615063,-0.0019124342,-0.2639161944,0.1165345982,-0.248997137,-0.0356290862,-0.2192046493,0.2607619464,-0.0442758687,0.0649944469,0.3897913992,-0.0012830885,0.4097664654,0.0638396591,-0.1327059269,-0.3079530001,-0.1985226125,-0.1465571076,0.2794715166,0.0859336779,0.2151842564,-0.1025955155,-0.1101848185,-0.355617255,0.1370177567,-0.1560878605,-0.1494034231,-0.1247703731,-0.0464178361,-0.3513237238,0.2281548381,0.1246922016,0.049195718,-0.0562466979,-0.0561646782,-0.1920246482,-0.0293024089,0.4452559352,-0.5346526504,-0.1444177032,0.0485171266,0.1739955544,-0.1337813288,0.0934993029,-0.2657158077,0.2842579782,0.3479267955,-0.1825049073,-0.1964315623,-0.0136570893,-0.1498102844,0.0140707949,-0.0336971879,0.3685227036,-0.0287478194,-0.4111033082,-0.2668015361,-0.2685930431]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/647","title":"Cannot download dataset_info.json","comments":"Right now the recommended way is to create the dataset on a server with internet connection and then to save it and copy the serialized dataset to the server without internet connection.","body":"I am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/text\/default-53ee3045f07ba8ca\/0.0.0\/dataset_info.json\r\n```\r\n\r\nI tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0\r\nTensorFlow version 2.3.0\r\ndatasets version: 1.0.1 \r\n","comment_length":32,"text":"Cannot download dataset_info.json\nI am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/text\/default-53ee3045f07ba8ca\/0.0.0\/dataset_info.json\r\n```\r\n\r\nI tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0\r\nTensorFlow version 2.3.0\r\ndatasets version: 1.0.1 \r\n\nRight now the recommended way is to create the dataset on a server with internet connection and then to save it and copy the serialized dataset to the server without internet connection.","embeddings":[-0.2739142179,0.0575308576,-0.0347832106,0.1870901138,0.0935543105,0.1803250462,0.0954198912,0.2747616768,0.08295542,0.0868135244,0.1172922999,0.2410640866,0.2068442702,0.1971997917,0.2280214131,-0.089716047,-0.1748178154,0.0688972697,0.0379331484,0.1446818113,-0.1514797807,0.0732105523,0.0222487431,0.0637850165,-0.1027823091,-0.3270795047,0.1324490309,0.1972536445,-0.2260443866,-0.0828825682,0.3538629115,0.1435671002,0.0654656366,0.190826878,-0.0001164499,0.2554253638,0.2761071026,-0.1264930665,-0.3891871274,-0.5971481204,-0.2190193981,-0.3448017538,0.1283259988,-0.3310894966,0.0493180268,-0.0733997971,0.3594780564,-0.1955469996,0.3597826064,0.3856014907,0.0774665177,0.2254529595,0.1947092861,0.0195436943,-0.1284128875,-0.121993199,0.0888299346,0.3027686775,0.032998383,0.1223273948,0.1463487148,0.1968403608,-0.1225891337,0.2005220503,0.3200191855,0.0434264839,-0.2740691602,-0.3115540445,0.375538975,0.4086201191,0.7346755862,-0.3074289858,-0.3879797459,-0.087109454,0.0589406826,-0.0454806946,0.4212038517,0.3348724842,-0.2379805744,0.1375946403,-0.3929382861,-0.5090522766,-0.5490764976,0.2441609502,-0.0238861237,-0.0689402819,-0.0839277133,0.0655507743,0.0700825527,0.0319329128,0.1131575331,-0.110623464,0.1374396235,0.2582965195,-0.0843182877,0.0147208814,-0.0684517026,-0.4558711052,0.099631235,0.1897504628,0.1806796342,0.0876994133,-0.228460148,0.2355820239,0.1796387732,0.0048505976,-0.0918789655,0.0451234542,0.3965954185,0.1481320113,0.2023462057,-0.1901379824,-0.1607414931,-0.119329311,-0.0338572413,0.0178025067,0.2646448016,-0.0304432567,-0.2177006155,0.2142812908,-0.2792238295,0.0577114783,-0.0448974743,0.3205521703,-0.4517366588,0.0873989239,0.364179343,0.0601880662,-0.041593913,-0.1907902509,-0.0322496891,0.0793243572,0.2298873663,0.0989318192,0.1862661541,-0.0769819319,0.405151397,-0.1058288068,-0.0253475569,-0.1501196474,0.3815007508,-0.0347792208,-0.3565560281,0.4856557548,0.4002937078,0.0500051454,-0.0698555037,0.0087387599,-0.0139075695,0.1168025061,-0.5142011642,-0.464381367,-0.1429093331,0.0872387588,-0.2037421316,0.0277113337,-0.3442267776,-0.0477546416,-0.1650036126,-0.2439986169,-0.1405543685,0.0832347944,-0.1811585277,-0.2239716351,0.4070423543,0.2606799006,-0.4162976444,-0.0523281135,0.009146465,-0.1822409332,0.0957825705,0.0717509538,-0.2450943291,0.3522921503,-0.1390307844,0.0617719106,0.522126615,-0.3030909896,-0.718416214,0.4686596394,-0.2963498235,-0.0514527969,0.0149099668,0.1851375699,0.1693569869,0.1772417724,0.3453245163,0.5167121291,0.0464096777,-0.0720027685,-0.0083472217,-0.3728986084,0.1220355928,0.2592833638,0.0820800662,0.1653037965,0.1632928848,0.0451843999,0.1699097902,0.000917975,0.0579500385,0.464712888,-0.002785559,0.1030352339,-0.1065410674,-0.07079871,-0.7039730549,0.1187509149,-0.0212403256,-0.1210229173,-0.6178978682,-0.123069182,-0.1917593479,-0.0210739579,0.0644884929,0.2469469309,0.0336726531,0.0527782328,0.1737811267,0.1497736871,-0.0895224214,0.4002639055,-0.2856488228,0.128625825,-0.4003409743,0.1032397971,0.1249527037,0.1720953733,0.1340138763,-0.1628952175,0.1166370958,-0.1240883172,-0.1739653945,0.4517517388,-0.2501598299,0.5106280446,-0.1213387698,0.3544617891,0.1358034164,-0.0870265067,-0.0439147204,0.0381285138,0.0842310116,0.0218504556,-0.3699810505,0.3021430671,0.1896648705,0.1218514442,0.1134383604,0.2273875475,0.3066497147,0.0234624948,-0.1310729384,0.2813239992,0.1746363491,0.1272785366,0.2188441008,-0.1274064779,-0.5014268756,0.2014043927,0.4261487722,-0.1354446858,0.1467635483,0.1135954484,-0.1700537205,-0.1370069534,0.0919980928,0.164185226,0.1402626783,-0.065551661,0.0136207957,0.1298469603,-0.044372391,-0.2295364141,-0.1062689498,0.0263654515,0.2623517811,0.0965017453,0.0491826795,0.033413738,0.0066689355,-0.1386484355,0.1981180757,0.2735874653,-0.1948744655,0.2650151849,-0.2267470658,-0.0929268822,0.0412768908,0.1424242407,-0.2560817599,-0.162516728,-0.092506133,0.4689167738,-0.0575848408,-0.1909465194,-0.1897982061,0.2515119612,0.1366597563,-0.526131928,-0.0057425424,-0.0959622636,-0.0684396923,-0.01858804,0.1396002322,0.1354275793,0.2400066406,-0.1796417683,0.0289248973,-0.3744145632,0.0997284129,0.0843668059,0.1751839221,0.1168126389,0.0857619345,0.6367673278,-0.1055241749,0.0176779404,0.0998934954,-0.0474891588,-0.169460088,0.0396545902,-0.1627120972,0.0653252602,-0.0126701249,-0.5224087238,-0.5092214942,-0.2645646036,0.3670660853,-0.0126496078,0.3041170537,0.1039887592,-0.0237409864,0.1928017437,0.3947021663,0.146002695,-0.0952269286,-0.7382696867,0.4059027731,-0.2604086995,-0.4382019937,0.2343431562,0.0621497184,0.1385695189,0.4293504059,-0.6090550423,-0.135950461,-0.0391379707,0.2375534773,-0.2261968553,-0.0720166564,0.5114514828,-0.1925547868,0.0823357478,-0.1040965319,0.0212002713,0.0836036801,0.12908943,0.2441371828,0.4184113443,0.5404879451,0.1681102514,0.7009679675,-0.056370154,0.0604445674,-0.0432019979,-0.3582232893,0.1943772286,0.1316646039,-0.1063024253,-0.1105115563,-0.1030639261,-0.1522717774,-0.1618840098,-0.081700094,0.1351223439,-0.2991721928,-0.1520617753,-0.4132375419,-0.3150022924,0.1286229193,-0.0678838938,0.2963247597,0.2067395747,0.215734303,-0.4287413955,-0.2588943243,0.0063701714,0.3762167394,0.178472206,0.2254025638,-0.132431671,-0.1656575054,-0.6117656231,0.3463323414,0.0103155123,0.4689979553,-0.2301760167,0.1798457056,0.0835047886,-0.0941344798,0.4290551841,-0.2495357841,0.3105312884,0.0183769893,-0.0356859043,-0.2880850136,0.0349950828,-0.0521402806,0.3610965908,0.3792183101,0.4429045618,-0.2210147083,-0.1046000794,-0.0795647055,-0.0167640392,-0.3497422934,-0.0143492566,-0.1387260258,-0.6285732985,-0.282969147,-0.187085703,-0.0445169136,0.3387078643,0.1342914701,0.2128896862,-0.0565803312,0.2776511014,-0.0254855081,0.0109867789,-0.0339394435,0.1958834082,-0.1098547652,0.5311949849,0.2137074023,0.2450786382,0.5885626674,0.3323698938,-0.3327031434,0.0199716575,0.0245345905,0.0532647185,0.1240947992,-0.1331605166,-0.0789608583,0.5337429643,0.1537056118,-0.3034186959,0.2127070725,0.1396434307,-0.0571989715,-0.298949182,-0.5814868808,0.4783387184,0.0775320455,-0.0566985868,0.2993632257,-0.2010769099,-0.0492857024,-0.0046018991,-0.1940877587,0.7162194848,-0.0931388736,0.1033353955,0.2150269896,0.1401149184,0.6555057764,-0.7900179625,0.3992990553,0.0381456539,-0.2679973245,-0.2783514559,-0.1457652003,0.025438942,0.1649257392,0.0266110823,0.3068061173,0.072044,0.1196134239,0.2397854626,0.4148024023,-0.289367646,-0.2671433985,-0.1212488711,0.0269073006,-0.2770169675,0.5673683882,-0.0577684753,-0.312671721,-0.0194542743,-0.196210295,-0.1964070052,0.2838349938,-0.3512867093,0.2460328043,-0.0581764579,-0.0332472138,-0.165585041,0.1115577221,0.0988958105,0.2406300008,-0.3796506524,0.0563735776,-0.3700048625,-0.2870515287,0.0400023647,0.1308580637,0.181256935,-0.1244323552,-0.2355837971,0.2666382194,-0.1747039557,-0.4450394213,0.3805356026,-0.2422423959,-0.2784985602,0.0918512568,-0.1555117518,-0.1224189475,-0.0879325271,-0.0163863357,0.0501872599,0.1084877625,-0.0505949333,-0.0403738953,0.0692366511,-0.153414607,0.0809173733,0.5816642642,-0.0806180164,-0.1587269455,0.5003042221,0.1704463214,-0.068115741,-0.1195202991,0.1162563413,-0.0412760079,0.1467309594,0.0195541717,-0.1293830574,0.4212910831,-0.3932151198,0.0289678387,0.1204623431,0.0393049568,-0.0330121703,-0.813549161,-0.2863284647,0.2287543565,0.1580988467,0.1293041259,0.2596993744,0.0100224772,0.1542126983,-0.0297784768,-0.19677113,0.131198138,-0.3208210468,-0.1879943758,0.4375839531,0.1676709503,0.4972367585,-0.0736194402,-0.0389232412,-0.0709862262,-0.1190066561,-0.0706461594,-0.0970102325,0.1862729341,-0.0162255,-0.1704066992,-0.0438649766,-0.5901021361,-0.231622979,0.2620551884,0.2117795348,0.1315574199,-0.0495784804,0.0545419827,-0.3019165397,0.3776362836,-0.3937428594,0.1543301493,0.0022256721,0.6276221275,-0.2808520496,0.2070873529,0.0500357598,0.0412114076,-0.4263059795,-0.1283075213,0.0890379101,-0.1582613438,0.4299826622,-0.2920497358,0.2547509074,-0.1186930314,0.1042036265,0.1905020475,-0.250333786,0.2916423082,0.23375763,0.0911973044,-0.0436655469,-0.0714254677,0.2956956327,0.009698703,0.0150977829,-0.0356242843,0.0838096589,0.0553011931,-0.1473531574,-0.0251035523,0.4249934256,-0.1997364014,0.1090438664,0.2411954552,-0.0350366868,0.0313412771,-0.0165633615,0.3380119205,0.1041948721,0.2461734861,-0.0894294754,-0.1074882075,0.1684742421,0.1592974365,-0.1315558851,-0.5190196633,-0.2859692872,0.0930766091,-0.3006415367,0.1135270521,-0.2937904894,0.1715636551,-0.2090359926,-0.0076380926,-0.398213923,0.2975952625,0.0084656877,-0.1158959121,-0.4368093312,-0.1715792418,0.0607246868,0.2427889556,0.3312906027,-0.3223013282,-0.0130050667,0.0944707319,-0.0486205332,-0.0815675929,0.0354829952,0.1492992342,0.0825720951,-0.2803864181,0.0063785836,-0.1056698412,-0.1089556292,0.0429225825,0.204076916,0.2475282252,0.0892019272,-0.0607504658,0.3859770596,0.0831939578,-0.1735335588,0.0214274153,0.0603294559,0.3365596533,-0.3860048056,0.2362745404,0.0912705064,-0.0563467033,0.3863759935,-0.0467832014,0.3695307672,-0.1001305729,0.2297031134,-0.0592059605,-0.0601835251,-0.1932042241,-0.1512310654,-0.4297840893,-0.2587772608,0.3552361727,0.0758617967,0.3142484128,-0.0911463797,0.0132115968,-0.2163451165,0.5425177217,0.3017623127,0.3233076036,-0.0104507916,-0.0313139893,-0.4100427926,-0.0654306561,-0.2089732438,-0.2313196212,0.1802463084,0.2493040711,-0.1689026654,0.1777102351,-0.15274252,-0.0287997462,-0.1564905196,0.1588185728,-0.073037304,-0.2842103839,-0.0704203993,0.129932791,-0.0345264487,-0.3565250337,0.1670374125,-0.2185204476,-0.0108180819,-0.1399709582,0.2111242712,-0.1460442543,0.0307210758,0.421356082,0.0243799854,0.349337548,0.0227135438,-0.131988883,-0.2136928737,-0.0371702313,-0.2274544537,0.300016731,0.123816587,0.2504255772,-0.1134877726,-0.0519046336,-0.3240870237,0.0799015686,-0.1384281665,-0.1868674308,-0.164100647,-0.1094836146,-0.357294172,0.2280038148,0.1295659989,0.046415735,-0.0875973403,-0.0353558548,-0.217574209,-0.0534098446,0.4117560983,-0.5382970572,-0.0688973293,-0.0045623141,0.1256928295,-0.128225401,0.1439084858,-0.2383605242,0.3045871854,0.3289352655,-0.2071260065,-0.1813850552,-0.0144907748,-0.1164342612,-0.0206493828,-0.0535508506,0.4048131704,-0.0768163651,-0.3848380744,-0.3207512796,-0.2601368427]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/647","title":"Cannot download dataset_info.json","comments":"#652 should allow you to load text\/json\/csv\/pandas datasets without an internet connection **IF** you've the dataset script locally.\r\n\r\nExample: \r\nIf you have `datasets\/text\/text.py` locally, then you can do `load_dataset(\".\/datasets\/text\", data_files=...)`","body":"I am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/text\/default-53ee3045f07ba8ca\/0.0.0\/dataset_info.json\r\n```\r\n\r\nI tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0\r\nTensorFlow version 2.3.0\r\ndatasets version: 1.0.1 \r\n","comment_length":30,"text":"Cannot download dataset_info.json\nI am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/text\/default-53ee3045f07ba8ca\/0.0.0\/dataset_info.json\r\n```\r\n\r\nI tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0\r\nTensorFlow version 2.3.0\r\ndatasets version: 1.0.1 \r\n\n#652 should allow you to load text\/json\/csv\/pandas datasets without an internet connection **IF** you've the dataset script locally.\r\n\r\nExample: \r\nIf you have `datasets\/text\/text.py` locally, then you can do `load_dataset(\".\/datasets\/text\", data_files=...)`","embeddings":[-0.2791209519,0.0578073859,-0.0465700962,0.1835988313,0.1229124293,0.1868670136,0.1547661722,0.2711952031,0.2081411928,0.0419166461,0.0489858687,0.2596798241,0.282694906,0.1925579756,0.1847233623,-0.044526998,-0.2054240704,0.0323137604,0.0241263341,0.1765891016,-0.2130242437,0.0461439006,-0.0344448686,0.1338437498,-0.1162246987,-0.326908946,0.1606549472,0.2022776902,-0.1810084879,-0.0399473011,0.4320088029,0.1607673913,0.058231581,0.1993090361,-0.0001162807,0.2081344724,0.3396979272,-0.0523446836,-0.3972235024,-0.6450169683,-0.2502361536,-0.2693962753,0.1729401201,-0.3017248809,0.0687525049,-0.1249875054,0.3304026425,-0.2659765482,0.2997096479,0.3808624148,0.0811682716,0.186101228,0.1585743576,-0.0045692651,-0.1447033733,-0.1378964782,0.1531009376,0.284047991,0.042373497,-0.0152341686,0.1086691618,0.2042810172,-0.1773227304,0.2587242424,0.3738180995,0.1103390381,-0.2407457829,-0.2773667574,0.4009684324,0.4214972556,0.7367280126,-0.3251824677,-0.3985488713,-0.1495109797,0.0299512371,-0.0038786351,0.3894180059,0.314329654,-0.2539661825,0.0717129856,-0.3475775123,-0.4419797063,-0.5005505681,0.2443239838,-0.0905222371,-0.0712040961,-0.0764266029,0.0694526061,0.1398933977,0.0175054949,0.2089918256,-0.0750415623,0.101654999,0.2641214132,-0.1614484191,0.0680155754,0.0132725742,-0.4202747643,0.0799991563,0.1321251541,0.1859790236,0.1021032482,-0.3240417838,0.2630485594,0.2888851464,0.0031036921,-0.0972453952,0.0537399054,0.411030978,0.2029902637,0.2062124461,-0.1693377793,-0.1028858051,-0.2158842832,-0.0904720947,-0.0342144184,0.2820253372,-0.0254259165,-0.2750402093,0.1628293246,-0.243377164,0.0810471177,0.0057741208,0.3083935976,-0.4633273184,0.1482780725,0.3730780482,0.0677468106,0.0011441385,-0.1491360962,-0.0526331402,0.1320006102,0.1611264646,0.0686709955,0.1769047827,-0.0410545953,0.352696389,-0.1015117466,-0.0355790481,-0.1430633068,0.3078434169,-0.0415491983,-0.3538811803,0.4735928178,0.3906008899,0.0865467116,-0.0025990999,-0.0301783979,-0.0229926612,0.0905554965,-0.5049895644,-0.4248255193,-0.1023380533,0.0907278657,-0.1651343852,0.0188833959,-0.3795021772,-0.054054182,-0.238785997,-0.2132173032,-0.1138018072,0.0657531619,-0.1576316804,-0.1937026978,0.3814212382,0.3426974118,-0.4553879797,-0.042890057,-0.02567669,-0.1867495775,0.0314610787,0.1171082929,-0.2610658407,0.3278618753,-0.1416548938,-0.0006124565,0.4650782049,-0.3743421137,-0.6786803603,0.5116195679,-0.2511117756,-0.0373778529,0.0120726023,0.1940940917,0.0640429854,0.2246436626,0.4172989428,0.5540279746,0.0352414064,-0.0761468932,0.033500459,-0.3742308617,0.1206460968,0.3107518256,0.1238316372,0.1601830423,0.1756913215,-0.0048195873,0.1874802709,0.0122599164,0.0361302681,0.4173101187,-0.0478650033,0.1298313141,-0.0888287127,-0.0869750679,-0.6877627969,0.1259184331,0.0302223749,-0.1560220718,-0.650418222,-0.1148975864,-0.2287505418,-0.0189219993,0.0801192299,0.2297023684,0.0302912556,0.067814514,0.1056918204,0.1507726759,-0.1127461493,0.4634071589,-0.2621696889,0.154511109,-0.364269346,0.0545704216,0.1842841208,0.2045586407,0.1370400786,-0.1432601362,0.0727346539,-0.0753005892,-0.1526921839,0.4997187257,-0.2178864777,0.4904348552,-0.0908864215,0.3165008426,0.1006958261,-0.01965666,-0.0989884287,0.0419103503,0.1325565577,0.00491085,-0.3661691546,0.3218409419,0.1772815138,0.2363182902,0.1218028069,0.2513240576,0.271055162,0.0193479937,-0.0697235316,0.2834273279,0.1404714733,0.0977991149,0.2253127992,-0.1177348197,-0.5117952228,0.1311011016,0.3453787863,-0.1234256998,0.1201108918,0.1088846698,-0.1616204083,-0.1135591492,0.0792017356,0.1813419908,0.1146394908,-0.051586818,-0.0541436784,0.1559963077,-0.0488835797,-0.2223096043,-0.1094931215,0.0705574155,0.1992183477,0.0120597025,0.0031578953,0.023165239,0.0083967345,-0.1761657149,0.1933675557,0.2756211758,-0.2250331193,0.2806057036,-0.242924273,-0.1875534654,0.127223894,0.1716041863,-0.2033437341,-0.1666338444,-0.1529945582,0.4840420485,-0.0454374515,-0.1549713463,-0.1914562136,0.2822308838,0.1641656458,-0.5675438046,-0.0469845086,-0.1555359215,-0.056237042,0.0154019883,0.2038508505,0.1700980365,0.2169905603,-0.1641974896,0.063286148,-0.436412096,0.1209371984,0.0438387468,0.1322410703,0.1461335272,0.1247555315,0.6109936833,-0.1397796273,0.029906109,0.0867849812,-0.0766616538,-0.0999118015,-0.0275352616,-0.157481581,0.0970095545,-0.0283454917,-0.5136215687,-0.4184240997,-0.3151581585,0.4051456153,0.0110967066,0.2497459352,0.142914623,0.0136674857,0.2127899528,0.3999161124,0.1146334857,-0.0925625637,-0.6567987204,0.400585264,-0.2118275911,-0.3886095881,0.1795176864,0.0410543606,0.1734778881,0.3741798699,-0.6029786468,-0.149218455,-0.0957719013,0.2385056466,-0.2502659261,-0.0587994121,0.4799557924,-0.2010459453,0.0503729433,-0.0519972555,-0.0412348285,0.0203936864,0.1076026559,0.2154725343,0.4581021965,0.5635311604,0.1050902009,0.6342528462,-0.1871810406,0.0657904893,-0.0232320502,-0.3957525194,0.2789325118,0.1227038577,-0.1484974176,-0.2132077515,-0.0865143239,-0.1262926608,-0.179120481,-0.0912338197,0.1277408749,-0.3572419286,-0.1581609845,-0.431129694,-0.2972108126,0.0731826276,-0.1085886285,0.1910106838,0.225886032,0.2070735991,-0.3853976727,-0.2513483465,0.0082503436,0.369232893,0.1811159998,0.186881125,-0.0452396832,-0.1204561368,-0.5152062774,0.3726807237,-0.0426244363,0.4191334844,-0.1932641566,0.1551182568,0.0765385628,-0.063243717,0.479293555,-0.2443970144,0.3249922395,0.077328898,-0.0583404452,-0.3281806409,0.0437759981,-0.0996175557,0.3088564873,0.3605071604,0.4001760781,-0.1641087085,-0.1108548939,-0.0523646884,-0.0363094322,-0.2664946914,-0.0389655456,-0.1516127586,-0.6053225994,-0.2809658349,-0.1779871881,0.0015701983,0.3210443556,0.1302578002,0.2168483138,-0.066672571,0.2029850483,0.0007949625,0.0263554491,-0.0381246246,0.2138100415,-0.1303320825,0.4847907126,0.1835351586,0.2229404896,0.5773658156,0.3251748681,-0.3378044665,0.023107497,0.015353119,0.0541646667,0.1097825542,-0.1483043283,-0.0399408676,0.5120699406,0.1543312073,-0.216136381,0.2679843605,0.1858838797,-0.0193492919,-0.255417347,-0.6606379151,0.5072332621,0.0906059667,-0.1004103795,0.3094779253,-0.1373150796,-0.0237892028,-0.0360246189,-0.2209880054,0.6713657975,-0.1245176792,0.1313057542,0.2163774073,0.1202802137,0.7099757195,-0.7732665539,0.3676516414,0.0492251106,-0.2292073518,-0.2622096241,-0.1553094089,0.0091339154,0.1447046399,0.1403291374,0.370215416,0.0631041452,0.1699427813,0.3019890785,0.3609046042,-0.312366873,-0.239106819,-0.129696995,0.0286568515,-0.2375215292,0.5534261465,-0.0718666837,-0.2736501694,-0.0073236474,-0.2225481719,-0.1852612048,0.262391001,-0.3086826801,0.2724312246,-0.0676866025,-0.0077404436,-0.1633070856,0.195807606,0.0608031936,0.2248882949,-0.3993468881,0.0888334438,-0.3730616271,-0.2940766811,0.0131200142,0.1559785753,0.2039615363,-0.0827014372,-0.2740246952,0.3077615201,-0.1745196283,-0.4703130424,0.3294248283,-0.1210385486,-0.2856192589,0.1029757932,-0.1716604084,-0.1696085632,-0.0616298541,0.0056703063,0.0552746467,0.0283126794,0.0480154678,-0.0401021168,0.0483588167,-0.2244568765,0.0567752309,0.6204462051,-0.1133008897,-0.1472988278,0.4869612455,0.2025518417,-0.0932652801,-0.128963232,0.0157462172,0.0206466317,0.1816702634,0.0618594773,-0.0970289931,0.4065269232,-0.3604156375,-0.0296382029,0.1149355918,0.0493562035,-0.0129001867,-0.8330102563,-0.2787037492,0.2436224073,0.0842235684,0.1087343469,0.2688502073,-0.0022197959,0.1604645848,-0.0900823921,-0.1942878664,0.1731632799,-0.29850927,-0.22021541,0.4320726991,0.2234663516,0.524835825,-0.0811828375,-0.0177935362,-0.0946031287,-0.0888049304,-0.0914296806,-0.0799653307,0.1858024895,0.001156977,-0.1877684295,-0.010869965,-0.6070572734,-0.2640092969,0.2268239707,0.2717204988,0.1981514096,-0.034951251,0.0191734489,-0.3432267606,0.4107033908,-0.3544176817,0.1874034852,-0.0310113262,0.5695282221,-0.2640039921,0.1736275554,0.0520727485,0.0772723481,-0.3644012511,-0.1432739496,0.0622182637,-0.1929645836,0.427082032,-0.2802501917,0.2403900921,-0.1060439572,0.1559551805,0.2138102353,-0.2523761094,0.2427348197,0.2068300843,0.0972651243,-0.0943275616,-0.0249108505,0.2158678025,0.0390221253,0.0503954068,-0.0125889387,0.0665477067,-0.0099800313,-0.1072607338,-0.0145940464,0.449859947,-0.185998857,0.1092173755,0.3178959787,-0.0188896954,-0.0018891189,0.0333694145,0.3920710683,0.1267469227,0.293499887,-0.0956425071,-0.2091741562,0.0239960253,0.2001218647,-0.1951174289,-0.5417727232,-0.2135819495,0.1336314529,-0.3624091744,0.0338804685,-0.3084082901,0.1986405551,-0.2256320715,0.091682978,-0.3848973811,0.2415825427,0.0161779467,-0.1153538004,-0.3337486982,-0.2156986147,0.0538560376,0.219604373,0.3604961932,-0.3386316895,-0.065882884,0.0333778299,-0.0416155159,-0.1041029692,0.0878843144,0.1423382908,0.0879145265,-0.2790158391,-0.0923997611,-0.1470281333,-0.0800864995,0.116598852,0.2271205634,0.2707594335,0.1289187372,-0.0918840468,0.3254320621,0.0407519154,-0.1868710965,0.0429278314,0.0326357074,0.3506328464,-0.3650422394,0.2370020598,0.1269571483,-0.0832222104,0.4458383322,-0.0731193423,0.4121448398,-0.1663271785,0.1976698637,-0.0692232922,-0.0354491249,-0.2082854956,-0.1467546225,-0.4075517058,-0.3283407092,0.3176425695,0.0635408461,0.3068359196,-0.0944892168,0.0135338195,-0.1860031784,0.5233602524,0.3315802813,0.3229912221,0.0180413201,0.0121835861,-0.4556885362,-0.0568073764,-0.2268435508,-0.2344715595,0.1527452469,0.1690326929,-0.1858552247,0.1104456335,-0.1380706877,-0.0008358348,-0.2150798291,0.0705234185,-0.0605563037,-0.3487340212,-0.1644011885,0.1297166646,-0.0251808465,-0.3342964053,0.0895451382,-0.2687938213,-0.0057376628,-0.1350550354,0.2368740141,-0.1141494811,0.0302369893,0.3860865235,-0.0399918258,0.3857759535,0.0656510592,-0.087156795,-0.2826167643,-0.1070456579,-0.2059087455,0.3248488307,0.1207988486,0.2053455114,-0.119281657,-0.1100053489,-0.309741199,0.1902949214,-0.1096412912,-0.1507735997,-0.1237106472,-0.099408038,-0.307834506,0.2365946323,0.069810845,-0.0148069104,-0.0909005627,0.0097412439,-0.1888253093,-0.0998867899,0.4315690696,-0.4962154031,-0.1133599803,-0.0368740372,0.1417098045,-0.1003408283,0.1042066664,-0.2002293617,0.3202190995,0.3165217936,-0.147985056,-0.1897581667,-0.05628407,-0.177348271,0.0039493884,-0.070331879,0.3125739098,-0.0593609288,-0.3655907512,-0.3102480769,-0.255245924]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"Thanks for reporting !\r\nIt uses a temporary file to write the data.\r\nHowever it looks like the temporary file is not placed in the right directory during the processing","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":30,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nThanks for reporting !\r\nIt uses a temporary file to write the data.\r\nHowever it looks like the temporary file is not placed in the right directory during the processing","embeddings":[-0.0562820248,0.175503239,-0.048730582,0.3908851743,-0.0356466025,-0.0572809353,0.2760777771,-0.0244471896,0.0408122353,0.1573939621,0.0506472811,0.1766866893,0.0519220531,0.3638361692,-0.035657201,0.3218174875,0.2768937051,-0.0544235073,0.0543463118,-0.0435270667,-0.3072208166,0.3765655458,-0.2001904845,-0.0186393708,-0.5099134445,-0.1708419919,-0.2043422461,-0.0366436392,0.0198707394,0.2087092996,0.1927695423,0.0984714553,0.1747972965,0.6240900159,-0.0001294347,-0.1149767116,0.2048154324,-0.266631186,-0.1961358786,-0.1022202075,-0.3119006157,-0.004488559,-0.1497934759,-0.0720397234,-0.1685159355,0.058167927,0.3384491801,-0.5294010639,-0.0324837863,0.197088182,0.0550545305,-0.4087739289,-0.4787145257,0.3768620193,0.1074029505,0.2737840116,-0.127657488,0.2617730498,0.0274086669,-0.1451185495,0.0063617732,0.2977354228,0.0007320868,-0.0749415234,0.281091392,0.3585237861,-0.2941537499,-0.3332058489,0.5126181841,-0.4663319886,0.3986636102,-0.3778884113,-0.2051143348,-0.2074549645,-0.3296646476,-0.2811291814,0.5710793734,0.1723903865,-0.0678301156,0.0381803103,-0.3426438272,-0.2352154851,0.0801221058,0.1073578671,-0.1712417603,0.2584484816,-0.2692752779,0.0078239897,-0.0813757405,0.2511205375,0.6566110849,-0.3971767128,-0.0177655574,0.3113218546,-0.017163083,0.0418484882,-0.2528593838,0.5198611021,-0.2589477003,-0.1941890419,-0.140186429,-0.137127474,-0.1453160644,0.1114479005,-0.0100332508,0.3893760443,0.0142912092,0.1784577966,0.1091897413,0.0854991004,-0.4649645686,-0.1024239138,-0.1018914804,-0.280590117,0.0215530582,0.1529057473,-0.0357462354,-0.2422925383,-0.1104080006,-0.0570120141,-0.3516086042,0.0353413634,0.0664604306,0.2854492366,0.0223693084,-0.1636580378,-0.2073559016,0.2409924716,-0.1883873343,0.3741033077,-0.0681627691,0.0487460122,-0.3213572502,0.4333797097,0.0406439193,-0.2225215286,0.1916906536,0.0157996304,-0.0768359974,-0.3220772743,-0.0095080128,-0.4610790014,0.3809150755,0.2202900946,0.1426358521,0.4319464564,0.001960421,-0.2612242103,-0.1253498644,0.2638800442,-0.5655605197,-0.2583985031,0.2529354692,-0.0420585498,-0.2712312639,-0.0508865528,-0.611823678,0.0973582864,0.4751488864,-0.2496961355,0.0291339941,-0.0071691517,-0.5547562838,-0.2493054271,0.0841405764,0.2138258815,-0.3784083426,0.2292414606,-0.2632399797,0.4557931721,0.4782080352,0.4059207439,0.0980356708,0.3541256785,-0.4703489542,-0.018397823,0.1202925667,-0.3309086561,-0.7670695186,0.2393060923,-0.0553215742,0.0022589867,0.0424810797,0.0430202186,0.1667083353,-0.1728424579,0.1321417242,0.146696642,0.1153548807,0.2249985486,-0.2722533047,-0.1135415733,-0.0317636095,-0.140868187,0.0147976708,0.1906705499,0.0313851163,-0.3111228645,0.1627004296,-0.0954517499,0.1095285341,0.2840329111,0.198729232,-0.1217923686,-0.0088152941,0.2195258588,-0.1978287846,0.078565374,-0.0435933098,-0.1874962896,-0.4925785065,-0.1061089486,-0.0936830565,-0.1442343742,-0.1505865753,-0.2586653531,-0.070561938,0.3078176081,0.1895607859,0.0718136579,-0.0172853749,0.5587958097,0.0425693914,-0.1547439992,-0.0659867153,0.2057202011,-0.110419631,-0.2670923173,-0.1782824546,-0.1684375703,0.2095979601,-0.2671532929,-0.1758679748,0.1820199937,0.2802887857,0.4550450146,0.0352442339,0.2760804296,0.1084619835,0.1304475814,0.2053912729,-0.0924353227,0.0293364003,0.1651969701,-0.0806705952,0.3651889861,-0.1599464864,0.0635770783,-0.0642158464,-0.3164082468,-0.0544167124,0.1412292272,0.2415256202,-0.1335027814,0.1456729174,0.406855911,0.38108024,0.1703892201,0.1708752066,0.3384847045,0.5877079964,0.0830793008,-0.0072799805,0.0868361369,-0.1417501271,-0.2576186359,0.208980903,0.4739125669,0.4611819983,-0.1000749618,0.2996844947,0.1259815693,0.0905264243,-0.1617136598,0.207000196,0.0486471578,0.1792274863,-0.0102858143,0.2194525152,-0.0296279509,-0.0411489606,0.2106044888,0.0995521247,0.2070693374,-0.0999977291,0.5765224695,-0.1749670357,-0.2962671518,-0.251406163,-0.2246739566,0.126494363,-0.1912495643,-0.0676634088,0.2575970292,-0.0080911685,-0.0047091036,0.3018825948,0.1664896458,-0.0958427787,-0.5050302148,-0.1803772151,0.1535266638,-0.2005008459,-0.113820903,0.3516035974,-0.0197491478,0.0268060695,-0.2237724513,0.0200731717,-0.3809759319,-0.0776564702,0.2111560702,-0.1135221496,0.2927731574,-0.2416089624,0.0524750315,-0.3985488117,-0.2574800551,-0.1008351073,-0.0618555062,-0.0682767481,-0.4157515466,-0.0122670354,-0.0417399481,-0.1293983757,0.0482095219,-0.2307545692,-0.0240253564,0.0910839885,-0.0031617447,0.3184533119,-0.0141437128,-0.0393549576,-0.2473150641,0.3702619076,-0.3497532308,-0.3336966038,-0.6516954899,0.2838873565,-0.1492913365,-0.0440371893,-0.0416815057,-0.0779947564,0.0555622615,0.4272268116,-0.42908746,-0.3207131922,-0.1345354468,-0.0098096468,0.2001427561,-0.0815653354,0.3003765941,-0.004548057,0.2103828043,-0.0307016075,-0.3322169483,0.0428318307,0.3315358758,0.3256826103,-0.0178597346,0.1884341687,0.0062960996,0.6751106381,0.2125789672,0.0176498033,0.3855674863,-0.062898986,0.2614315152,-0.2494227588,-0.0922999755,0.227580741,0.0313537978,-0.6610519886,0.3991059363,0.0960739702,-0.4041826427,0.1855948865,-0.2265985757,-0.327567637,-0.2346793413,0.2420727313,0.0593656637,0.2746419013,0.0666518956,0.1323849261,-0.2673313916,-0.2672293484,0.0602858886,0.1197602376,0.2946386039,0.1816901416,0.3472676873,-0.5023311377,-0.3641363084,0.3330274224,0.390514046,0.1117807552,-0.061358057,0.3069008589,0.1638711989,-0.1416215599,0.6055469513,-0.3414019644,-0.1618326753,-0.1456077695,0.0838119909,-0.0051114787,0.0111391721,0.0911115929,0.428199321,0.0127998609,0.502135694,0.3324736655,0.0061751353,-0.1754256338,0.2356707752,-0.3256070912,-0.0917089731,-0.262586385,-0.5155370831,-0.2017365545,-0.0243561864,0.1427342445,-0.0709052235,0.1209731102,0.0633415282,0.1503104568,-0.0181341935,0.0420119539,0.1489115059,0.4182381332,0.1263737381,-0.1437668204,0.1408954412,0.140935123,0.3221425116,0.5147647262,0.1704253554,0.21716021,0.2110309601,-0.2364504188,0.3073489368,0.4408431351,-0.1777643412,-0.1690578163,-0.169563964,-0.0828565136,-0.3682968318,0.3819870055,0.163952291,0.3501552045,-0.378600806,-0.4777137041,0.4054703414,0.1832027435,-0.2680819035,0.3681361675,-0.0032617189,-0.5314047933,0.4476582706,0.1192126796,1.0677214861,-0.0280556865,0.2187955379,0.0248621367,0.0335933603,0.026398072,-0.4069087505,0.2578330934,-0.2166917324,-0.6217984557,-0.1070586517,-0.2104317993,-0.1467541605,0.0568302386,-0.219652757,0.450304091,0.0789023787,-0.0052414099,-0.0949504524,-0.1046709344,-0.1614505947,-0.1285874993,-0.157095179,0.0151919387,0.0346292183,0.4675065875,-0.1815328747,0.1736757755,-0.2376923263,0.0873099044,-0.1408262849,-0.0248799324,-0.503633976,-0.0072958753,-0.3956756592,-0.278301388,-0.1324392557,0.3288989067,0.5609943867,0.2409276515,0.0507681035,-0.0050175819,0.134610191,0.2099206299,-0.0000260093,-0.4559622407,0.0346977152,0.0874912664,0.1809628606,-0.311373502,0.0901878998,-0.2999633551,-0.0884054974,0.0175642166,-0.0760427788,0.0428938046,-0.2532404661,0.019929802,-0.1287475079,0.0843431428,0.0125715369,0.0075321011,0.2728303671,0.4122650623,-0.5614525676,-0.3694346845,-0.2027308941,0.1790422648,0.2824056149,0.0935889855,0.524256289,-0.2457650155,-0.1774655432,0.0112958467,0.0560787916,-0.3277435601,0.2327900678,0.1281796992,-0.4396623075,-0.1521253586,-0.1758618355,0.073295176,0.2788392007,-0.0159517638,-0.4446371794,-0.3718070686,-0.4002510905,0.0330342166,0.0493193157,0.0435315557,0.0281921104,0.256239593,0.1701101363,-0.2793209851,-0.1557126492,0.16068694,-0.0612228885,0.1131735742,0.0303951781,0.0429069251,0.4526516795,-0.1833742708,-0.1235075742,0.1809015572,0.0278482549,-0.07766895,-0.0461231098,0.1795363426,0.1671635211,-0.0533492826,-0.024050653,-0.2914877832,0.1014394686,-0.4307421744,0.2056397051,0.2780205607,0.1040159538,-0.1492013782,0.1424132288,0.0012865819,-0.0319461972,-0.0352662727,-0.1286031902,0.3755893409,0.1459937543,0.2723150551,0.1754768044,0.0813812315,-0.0299717393,-0.0383945443,0.1442059129,0.3954287171,0.129550308,-0.2500985265,0.0136541389,-0.0697722882,0.0310945436,0.124249056,-0.1378899664,-0.3759160042,0.3196120858,-0.0090324283,0.13037543,-0.0355183631,0.050374452,-0.1860329956,-0.0418298244,0.5113438964,0.2504440546,0.1900663823,-0.0409482941,0.0315542556,0.4879702032,-0.210905835,0.0090667522,-0.2042765915,0.033489529,0.5536411405,0.3623776436,0.0902799144,-0.0718246251,0.5481159687,0.1639937609,0.1840903163,0.4036704898,-0.0341610797,-0.0197869111,-0.1628611535,-0.0979248956,0.2346383482,-0.1800171733,0.3569232523,0.2207547873,0.0212633833,-0.4289576709,-0.0185194165,-0.23010993,0.2464829236,-0.1843438596,-0.2571744621,0.2252781689,-0.1393987089,0.1097766608,0.3992193937,-0.1051693559,-0.0902865008,0.0895912647,0.1580913961,-0.1947741956,0.3890988827,0.0247050375,0.1284039617,0.2157687545,-0.0189616345,0.6912629604,-0.4017899632,0.1211556122,-0.0657231212,0.1271677017,0.2669976056,0.3885891438,-0.3067551553,-0.1319033355,0.1379080415,0.07229238,-0.0580658466,0.2541581094,-0.0454964079,0.0189016126,0.0825254023,0.0403179452,-0.0241406206,0.192021355,0.1807078719,0.1906842887,0.2427741438,0.0923807025,-0.1546779126,0.0720106736,-0.1111399457,0.0404135473,-0.3237197697,0.1414380521,0.1943964213,-0.2693227232,0.3378166258,0.2822555006,-0.0039548404,0.1737370938,0.4144847095,0.4470077753,0.3810233772,-0.3458941579,-0.2810313106,-0.3214671612,0.3492796123,-0.0205258373,0.1844316274,-0.5769681931,0.0726156533,0.039659448,-0.0608090386,0.011717651,0.0538607202,-0.1230439618,0.1567473859,-0.2668755352,-0.1203639135,-0.1544596553,-0.0252306089,-0.1251647174,-0.4647896886,0.2174205929,-0.2235407233,0.0293815639,0.0444851257,0.0315538533,-0.2476914674,-0.0166683346,0.4364027679,0.1388326436,0.1036730036,-0.2208717018,-0.0176571049,-0.2244780362,-0.0005778063,-0.3665600419,-0.042550344,-0.0079671973,0.3166204989,-0.3149808347,0.0972123966,-0.3729237616,0.0406824425,0.2318439186,0.1112584844,0.0731414482,-0.0877920315,-0.2125243694,0.072840564,0.1720267087,0.5756815672,0.0119075784,0.2167342901,-0.2610247433,0.004854783,0.2358810902,-0.3160810769,-0.4929459393,0.3831428289,0.0332278311,0.1937096119,-0.0659841672,-0.2832104862,0.054328654,0.1061108708,-0.2017673105,-0.1621267051,-0.005602608,0.0952657759,0.1294673681,-0.0604173951,0.3633388579,0.1185597852,-0.3589939177,-0.0725693628,-0.2732173502]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"Well actually I just tested and the temporary file is placed in the same directory, so it should work as expected.\r\nWhich version of `datasets` are you using ?","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":29,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nWell actually I just tested and the temporary file is placed in the same directory, so it should work as expected.\r\nWhich version of `datasets` are you using ?","embeddings":[-0.0980259776,0.188093707,-0.0280870888,0.389010936,-0.015298578,-0.0114677977,0.3272336423,-0.0096284868,0.0338673592,0.2286566049,0.0054276907,0.2275298089,0.0270035882,0.3596076667,-0.0818914026,0.3205984533,0.2760051787,-0.0400124863,0.064423047,-0.0552731417,-0.3491486609,0.3673860729,-0.250793457,-0.0342624858,-0.5128515363,-0.1368786991,-0.179927215,-0.0654983521,-0.05333478,0.2139814794,0.2417699695,0.0878142864,0.2495012879,0.661549449,-0.0001283521,-0.1254281253,0.2665183842,-0.2071329653,-0.2330143303,-0.1373612583,-0.3181043863,-0.014190034,-0.1339170635,-0.067687653,-0.1839046776,0.0559142567,0.272248596,-0.5808059573,-0.025083039,0.2409744561,0.0725319088,-0.4288330972,-0.4749288559,0.3053100109,0.0975685343,0.2478017658,-0.142292276,0.24306871,0.0100712012,-0.1486692727,-0.0139403753,0.3300213516,-0.0270178802,-0.0138578033,0.2550145984,0.3406344652,-0.3194044828,-0.3421723545,0.5276225209,-0.4035201669,0.4292229116,-0.3404000401,-0.1767730862,-0.2474320978,-0.3338007927,-0.2652026117,0.5508043766,0.1417493224,-0.0569063835,0.0546889231,-0.4403166473,-0.193094179,0.1015329212,0.1382195354,-0.238239482,0.2111843079,-0.2950587869,0.0276474431,-0.011653332,0.1779495776,0.6766315103,-0.3875518441,-0.0103655914,0.2630077302,-0.0764711723,0.0587412193,-0.1897370219,0.5381625891,-0.2099022716,-0.1787819415,-0.2083129883,-0.1328071654,-0.1667217165,0.0946199894,0.0385345221,0.4278603494,-0.0002032616,0.2039845288,0.1246508956,0.1027066931,-0.520960331,-0.0831503347,-0.0767244846,-0.2948054373,0.0310732555,0.0969406813,0.0182932839,-0.2827686071,-0.1402125508,-0.037873257,-0.3314204514,0.0076373308,0.1045960113,0.2721991539,0.0197714288,-0.1099094823,-0.2434050143,0.2486744374,-0.2118350863,0.3456625342,-0.087007165,0.0086983796,-0.3563786149,0.4384669363,0.0830166936,-0.2516986728,0.1514321864,0.0271900576,-0.0714058727,-0.2899838686,-0.0525118522,-0.4461802244,0.3620024025,0.1732686162,0.1398664713,0.4446723163,0.0133799035,-0.2829953432,-0.1326734275,0.2426331043,-0.5642427802,-0.2653877735,0.1985358,-0.0222923867,-0.2773581147,-0.0693465993,-0.6552339792,0.1527505368,0.4416986704,-0.2127338499,0.0085248305,-0.0149793057,-0.5274883509,-0.2238634527,0.0653401166,0.2004465312,-0.3652799428,0.2058353871,-0.3317601681,0.4557800591,0.4756256938,0.3935822248,0.060095001,0.2928664982,-0.4934003055,-0.0278436095,0.1094955131,-0.3094343543,-0.8372094035,0.2986419797,-0.0138159357,-0.0036940803,-0.0160198156,-0.0175278224,0.1869792342,-0.1316637397,0.1638331711,0.1282265484,0.1112386957,0.1921883523,-0.2752439678,-0.1077737361,-0.0107337227,-0.1389642209,0.0377673954,0.2096470594,0.041476462,-0.3141260445,0.1381217241,-0.0697862729,0.111001417,0.2728106678,0.2326658368,-0.1298245341,0.0115753245,0.1633971334,-0.2243786305,0.1474627405,0.0282337032,-0.1464349031,-0.4751341343,-0.1171608195,-0.1194390059,-0.1630004793,-0.1702288687,-0.251778841,-0.0769413188,0.3072750866,0.1577135324,0.0284263119,-0.0407034345,0.525129199,0.0626684353,-0.1593219489,-0.0775787085,0.2420241088,-0.1074183509,-0.2392097563,-0.148795411,-0.195563063,0.2205365747,-0.2613439858,-0.2108390331,0.1778932065,0.30595994,0.4031629562,0.0000764468,0.2722767293,0.1150537804,0.1092067435,0.169486776,-0.0730654225,0.0028634975,0.1589085311,-0.0870772079,0.2959052324,-0.1469192654,0.1218437925,-0.1049630567,-0.2768554688,-0.0485453308,0.1285164803,0.2272479832,-0.1897710115,0.0760212541,0.3430959284,0.3831253946,0.1982024014,0.2105258852,0.3193440139,0.6449167132,0.0745252147,-0.0350593664,0.130790621,-0.1365963966,-0.2520191967,0.2219441235,0.5235064626,0.4602493048,-0.0905334949,0.3149169981,0.1629967391,0.057884112,-0.1506624818,0.1834234148,0.0628222823,0.2291888595,-0.0392771289,0.2224123329,-0.0436756536,-0.0493959263,0.2005114406,0.0795630887,0.2584432662,-0.12045075,0.5334354043,-0.2883337736,-0.2838529646,-0.2435619235,-0.2431814522,0.1216463298,-0.2136332393,-0.0976688191,0.283724159,0.0229831152,0.0302868504,0.232142508,0.1779783815,-0.1119576395,-0.4812890887,-0.1549041867,0.1786360294,-0.2874040008,-0.1078283116,0.3575817943,-0.0645390972,-0.0030333495,-0.216862753,0.0499561653,-0.3820486665,-0.1052474976,0.2430865169,-0.146206215,0.2984904051,-0.2173310965,0.0096436944,-0.4151874185,-0.2589137852,-0.0426567122,-0.0731205493,-0.0663106441,-0.3781530261,-0.0269038547,-0.0367619842,-0.0858888701,0.0030577269,-0.1907083094,-0.0507298149,0.0393040031,-0.0032559084,0.2771962881,0.0263257567,-0.0282737389,-0.2060232162,0.3607292175,-0.3355197906,-0.3527882993,-0.6440094709,0.3036522269,-0.1445791572,-0.0496169664,-0.0091056125,-0.080620043,0.0811304525,0.4302971959,-0.4716597199,-0.3017313778,-0.1893893182,0.0136477137,0.2360906452,-0.0786211044,0.3027957976,0.0290112924,0.1964473873,-0.0391066931,-0.4018261731,0.012068999,0.3081807792,0.3579263091,0.0200459044,0.2512149811,0.054270599,0.6362167001,0.3114107549,-0.0254185032,0.4196685851,-0.0525017269,0.2791052759,-0.2759063244,-0.1334952414,0.2206906229,0.0085314624,-0.6468423009,0.4143176973,0.1350239664,-0.4364432395,0.1293365061,-0.2338132858,-0.309745878,-0.2208641171,0.2010632902,0.0150641752,0.2418791056,0.093167007,0.1408700645,-0.2238601893,-0.2583940923,0.0733997449,0.082902424,0.2794163525,0.1986617148,0.2841056883,-0.4492784142,-0.3418408036,0.3656334877,0.3487120569,0.1174183711,-0.0517198108,0.2863925099,0.1670887023,-0.1253135651,0.6017553806,-0.3549860716,-0.1576304138,-0.1396556348,0.0860990882,-0.0286171418,-0.0127259409,0.0868601501,0.4143778086,-0.0153196054,0.5273712277,0.3371331394,-0.0092130015,-0.1358054578,0.2135941386,-0.2908641398,-0.1111325696,-0.2496884167,-0.5072106719,-0.203581959,-0.030515084,0.1430048496,-0.0360442661,0.0577823296,-0.0157498047,0.131757617,-0.0788709521,0.0373287648,0.1687452644,0.3987447023,0.1434883773,-0.0929076746,0.1817075908,0.1506507993,0.3403621614,0.551104784,0.1500349194,0.1495835632,0.1981316954,-0.175190419,0.2572018802,0.3737203777,-0.1885133535,-0.2240612507,-0.1297322959,-0.0401520357,-0.3493865132,0.3809072077,0.1846684813,0.3165331781,-0.4214714766,-0.5009724498,0.4355785847,0.2205658704,-0.288097918,0.3781585097,0.0354241766,-0.5233605504,0.3892891705,0.1044136062,1.0422190428,-0.0292002298,0.1440987438,0.0531243011,0.0830091611,0.0313164182,-0.3644894361,0.2190930992,-0.1676378995,-0.6126423478,-0.0993559435,-0.170227021,-0.149457112,0.1020817906,-0.1965593398,0.4737482667,0.0464320853,0.0202112198,-0.1003165245,-0.0861563012,-0.1475774944,-0.1369802356,-0.2015279233,0.0424224623,-0.0138794929,0.4463690519,-0.1584409773,0.2061465085,-0.176995188,0.0774843916,-0.1612950861,-0.0300463755,-0.4129764438,-0.010057644,-0.3488592803,-0.277705729,-0.1216033697,0.3321136236,0.5052713752,0.2408725321,0.0065475721,0.0173622817,0.1277460605,0.2272510678,-0.0037387521,-0.444921881,0.0543819182,0.0711086169,0.1271183193,-0.3398972452,0.1443745047,-0.3299237788,-0.0805931315,0.0498799942,-0.0724172592,0.0122179389,-0.2223536372,0.0186888892,-0.0997203588,0.0818676427,0.018769823,-0.0277278069,0.2760848105,0.4491718113,-0.5062206388,-0.3761131763,-0.2003278732,0.2312143594,0.2537479401,0.0779747963,0.4709458053,-0.2033420354,-0.1989326626,0.0163193271,0.0493043251,-0.2708588243,0.2654402256,0.1176896915,-0.4325996935,-0.0760934427,-0.1049759686,0.0623806082,0.2794388831,-0.0690011084,-0.3804732859,-0.3362635076,-0.3996806145,0.0158156119,0.0075659486,0.0767894983,0.0372360572,0.1834812611,0.1978897005,-0.2522154152,-0.163239181,0.1962479502,-0.0291680396,0.0893339217,0.0308363438,0.0201762803,0.3836660683,-0.1811818928,-0.1201845482,0.1872044802,0.0288530625,-0.0975005478,-0.0426416136,0.1753437519,0.2000725865,-0.0873260498,-0.0261494014,-0.3177047074,0.0501990728,-0.4659591615,0.239288792,0.329480499,0.1140275449,-0.1467586756,0.160172984,-0.0901507363,0.0041615721,-0.0279722866,-0.1430335939,0.37413764,0.1804346889,0.2766548395,0.2307610959,0.0583929606,-0.0458598882,-0.035467755,0.1661788225,0.3798239827,0.1416240484,-0.2876999974,0.0398342051,-0.0435176305,0.081806004,0.1289951503,-0.1184119284,-0.4044561386,0.3410645425,0.0122533813,0.0788664445,-0.0586057082,0.0762959495,-0.1434560418,-0.0362136029,0.4677840471,0.2167766094,0.1726538539,-0.0198746826,0.0120632062,0.5274105668,-0.2263321429,0.0627222881,-0.1645873487,0.0231253486,0.5131480098,0.3814440966,0.1008141115,-0.0556069575,0.5689255595,0.127940163,0.1613864303,0.3975925744,-0.0264822021,0.0311635584,-0.2019389123,-0.0550160781,0.2214682549,-0.2355242819,0.3087995052,0.2096163929,0.0180903599,-0.4070118964,0.0104143992,-0.2349979877,0.1744944453,-0.153394714,-0.2313525379,0.2195528448,-0.1457259059,0.0759932846,0.4029558301,-0.1542879939,-0.1516586691,0.1393973678,0.1362307817,-0.2383453101,0.3084787726,0.0470379144,0.1173304021,0.2149900049,0.0212370567,0.7285473943,-0.3807320297,0.1181311086,-0.0495728217,0.1587444544,0.2729339898,0.3947113454,-0.2817905545,-0.1670736223,0.1221453175,0.061251197,-0.1060800403,0.2734852135,-0.0934416503,-0.0007922596,0.1241964325,0.0586615838,-0.0224676859,0.1964729577,0.134885177,0.2219512016,0.1697974801,0.1599490196,-0.1646047533,0.1446328312,-0.127521351,0.0413758978,-0.3454082906,0.1061098799,0.2801471353,-0.2671582401,0.3705637157,0.249825567,0.0027985349,0.1581301391,0.4029342234,0.4333430827,0.378410399,-0.3179951608,-0.2564052343,-0.3567007482,0.3485019505,-0.0395744182,0.2108885348,-0.5317530036,0.0837219134,0.009192938,-0.0288419873,0.0147677558,0.0835062787,-0.1280094534,0.194400683,-0.2590571344,-0.1269488037,-0.1365110874,0.0002452329,-0.1157820672,-0.4538482428,0.2380292714,-0.2168747187,0.0133956363,0.0219860077,0.0698784664,-0.2033717036,-0.0798680931,0.4682743847,0.2058281153,0.1373899579,-0.2023871988,0.0295157358,-0.3125998676,0.0013135684,-0.3460721672,-0.0226150136,-0.0177886318,0.3184143901,-0.3162148893,0.0922103971,-0.3713299036,0.0409188643,0.2394522876,0.1391164511,0.0661613867,-0.1014295965,-0.2200340927,0.0427592173,0.1457410753,0.5836619139,0.0304108001,0.2649640143,-0.2504309714,-0.0683840588,0.27648893,-0.3251815438,-0.5561156273,0.3612318039,0.0431550071,0.2361361235,-0.1428439021,-0.2722610533,0.0014823355,0.1511985511,-0.2202392668,-0.1296270192,0.0169416815,0.1076123491,0.1338991821,-0.0762368664,0.3681737781,0.1175735369,-0.34080562,-0.0224709138,-0.2614038587]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"`datasets-1.0.1`\r\nHere you can reproduce it here:\r\nhttps:\/\/colab.research.google.com\/drive\/1O0KcepTFsmpkBbrbLLMq42iwTKmQh8d5?usp=sharing\r\n","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":8,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\n`datasets-1.0.1`\r\nHere you can reproduce it here:\r\nhttps:\/\/colab.research.google.com\/drive\/1O0KcepTFsmpkBbrbLLMq42iwTKmQh8d5?usp=sharing\r\n","embeddings":[-0.0859269947,0.2105143666,-0.0259094182,0.4568368793,-0.0565692522,-0.0358938612,0.3175824285,0.0217307676,-0.0032252534,0.2064236104,-0.0123604294,0.2843514681,0.0427397005,0.4461206794,-0.0785819218,0.3241481781,0.2608175278,-0.0851598084,0.051477205,-0.0819793195,-0.3634168506,0.3861784339,-0.2248566598,-0.0192068089,-0.4734945595,-0.1032655686,-0.1497808695,-0.0416010767,-0.0594645478,0.2512085438,0.2503131628,0.0222314633,0.1794941127,0.6230234504,-0.0001284027,-0.1527742147,0.2034631968,-0.2297937572,-0.1955145299,-0.1140614823,-0.3389844596,0.0067239651,-0.0946708396,-0.0515125506,-0.209353283,0.145079568,0.2260410935,-0.5299069881,-0.0417477526,0.2581877708,0.0734880194,-0.438844651,-0.5047367811,0.3579004407,0.0671738759,0.2321267426,-0.1419566721,0.2834370434,0.0089863343,-0.1036424041,-0.0897616968,0.3229796886,-0.0508886948,-0.0091199512,0.2351137996,0.3252070844,-0.3294987082,-0.3114669919,0.5256301761,-0.4396483004,0.3353979886,-0.3274334371,-0.129487589,-0.2781305611,-0.3388575017,-0.3400581777,0.5355405807,0.1484738886,-0.0914822593,0.0689342991,-0.5111332536,-0.192500025,0.0778810978,0.0828469843,-0.2387162596,0.2999355495,-0.2760161459,0.0438477434,-0.0594080798,0.232314378,0.6684551835,-0.3926111162,0.0079123592,0.2085452527,-0.1130156294,0.0812150165,-0.1791343093,0.5704645514,-0.2104736716,-0.116953589,-0.2371394783,-0.1151392832,-0.1319716275,0.0896110907,0.0623460002,0.4495264888,-0.039935492,0.12878564,0.0753179863,0.0523317456,-0.5047028065,-0.0671607479,-0.0052911141,-0.2701866329,-0.0231546778,0.1327801347,-0.0336586833,-0.2817098796,-0.1495543569,-0.0432285964,-0.3456268013,0.0049989782,0.1195754632,0.2804924548,-0.0294180736,-0.1438725889,-0.2478039563,0.2346019447,-0.2164994925,0.3360671401,-0.0980529487,0.0096572191,-0.3533217311,0.463411957,0.0868544728,-0.2491939366,0.1437579691,0.0534504242,0.0120552927,-0.3182353377,-0.0554747581,-0.4173054993,0.3912698925,0.1545526683,0.1395917833,0.4230699241,0.01480613,-0.2285065949,-0.126036644,0.2177429348,-0.6053737998,-0.2480508387,0.1878372282,-0.017435465,-0.2685184479,-0.0699270889,-0.7038681507,0.1256981492,0.44171381,-0.18169339,-0.0001484236,-0.0358746238,-0.5169997215,-0.2336221635,0.0431912206,0.1529549956,-0.246355623,0.1710940003,-0.3305144012,0.5024914145,0.4864896238,0.3643982112,0.0411440134,0.246225372,-0.4963784516,-0.0254079364,0.071256794,-0.2785967588,-0.8266237974,0.3081037104,-0.0659767538,0.0089602629,-0.0225794502,-0.0386433192,0.2430036515,-0.1191023663,0.1943075806,0.146034807,0.1001713723,0.2107914388,-0.2540136576,-0.1479192972,-0.0308596771,-0.1766590029,0.0265350696,0.1936493218,0.0084530497,-0.3602747917,0.1458863616,-0.1321407557,0.1419627368,0.2384614944,0.243561089,-0.1596367061,-0.0031814098,0.1491408795,-0.2015686035,0.1593592912,0.0123271365,-0.1429904848,-0.5075867772,-0.0958065763,-0.1437901258,-0.1605003029,-0.1582639813,-0.2425342649,-0.0730035901,0.3165225387,0.106304884,0.0429452993,-0.0018431501,0.544718504,0.0838538036,-0.1620202363,-0.0596462414,0.2211902738,-0.1001619548,-0.2044643462,-0.1339729279,-0.1613882482,0.2472152263,-0.2365799993,-0.1905618161,0.1351072192,0.2361162454,0.4098289311,0.0015276857,0.2659107745,0.1049176231,0.0370874591,0.1842401773,-0.1047295704,-0.0314380974,0.1521859616,-0.0238724686,0.3346822262,-0.0980673805,0.1824603379,-0.0998011082,-0.2679816186,-0.0487923883,0.0778399482,0.272684902,-0.1619687974,0.0715112612,0.356780231,0.3740712702,0.2630158961,0.232255891,0.389228642,0.6078461409,0.0888204351,-0.0508456677,0.159095034,-0.1056971401,-0.2744719684,0.251042515,0.5018309355,0.4359130859,-0.109201096,0.2821084261,0.2095497698,0.0545706637,-0.1535334289,0.167074427,0.0697159469,0.2890127599,-0.017175002,0.2407371551,-0.0748685002,-0.0760969073,0.1666597575,0.0646106526,0.2317204773,-0.1293025464,0.5872750282,-0.3095234632,-0.2358986884,-0.2617283762,-0.3020540774,0.0744952485,-0.1905364096,-0.1237929463,0.3116358817,0.0411534943,0.0211631991,0.1923929155,0.2086050659,-0.1401247233,-0.432533145,-0.1313448995,0.1782785654,-0.2766745985,-0.1063859835,0.3385944366,-0.0532284863,0.0392203033,-0.1961381584,0.0434998237,-0.3656837046,-0.1418697387,0.2891896069,-0.1498378068,0.2874441743,-0.2322180569,-0.0268261824,-0.4468902349,-0.3565884233,-0.0927110091,-0.0000652044,-0.07298702,-0.4162923098,-0.0445621833,-0.033024922,-0.091918014,-0.0162865408,-0.1772962213,0.0048065642,0.0518919304,0.0014230608,0.294783771,-0.017781375,-0.0976191387,-0.2031913549,0.3764555454,-0.3099393547,-0.3719546497,-0.6327257156,0.2206784934,-0.1267177314,-0.0225250777,-0.0204467997,-0.0975492895,0.1260404587,0.3829345107,-0.4446614981,-0.2510305345,-0.1438884288,-0.0346541069,0.235169068,-0.116855748,0.2705733478,0.067327626,0.2098213583,-0.0291326679,-0.4357700348,0.0313161686,0.3048927486,0.3743615746,0.038527295,0.234416008,0.0510657765,0.6114255786,0.3091315329,-0.0226589534,0.4200598598,-0.0201362967,0.2290918827,-0.2524683177,-0.1324075311,0.2753468156,0.0232418533,-0.6568062901,0.4026684165,0.1236351654,-0.4404456913,0.1690159589,-0.2651309967,-0.3195312619,-0.2356149703,0.2237465531,0.0042809783,0.2263861448,0.0786603987,0.1232620403,-0.2254478037,-0.247725144,0.0637918487,0.0631776825,0.3021700382,0.1423139125,0.2853578627,-0.4864205718,-0.4033597708,0.3236558735,0.3794366717,0.0588944666,-0.0530069396,0.3046779037,0.1892239004,-0.1179858148,0.566806674,-0.3238274753,-0.1455767602,-0.165387094,0.0059846439,0.0071730176,-0.0027149678,0.0746291056,0.3955438137,-0.0238385797,0.4805594087,0.3246349096,0.0136521375,-0.0872606337,0.2133105844,-0.3349282146,-0.091222994,-0.1889190376,-0.5353719592,-0.1726032495,-0.0063267085,0.1511034518,-0.0123558743,0.046090249,-0.0064211427,0.1243515015,-0.0638395026,0.070753172,0.1734400094,0.4018012285,0.1729514897,-0.1075058803,0.1030358598,0.195174098,0.3815447092,0.5483594537,0.1643566936,0.1702649891,0.1976582557,-0.1668158919,0.2878287137,0.3344887495,-0.1925362051,-0.2355697751,-0.1161171943,-0.0494773313,-0.3378822803,0.3552585542,0.1480659992,0.3424273729,-0.4263763428,-0.4359541237,0.4566090405,0.1675349176,-0.3115753531,0.322681725,0.0678275675,-0.5420393348,0.4283147752,0.1543030888,1.0714001656,0.0033852037,0.14519611,0.0866712853,0.0854897872,-0.0246168021,-0.2929265201,0.2214425057,-0.173602581,-0.6409619451,-0.1054845974,-0.1736429781,-0.2099029273,0.0851637051,-0.1701610088,0.4203943312,0.0096939718,-0.0478363223,-0.1131186113,-0.1083497331,-0.0367388055,-0.132301867,-0.241411373,0.0416498594,-0.0608612783,0.4171026945,-0.1427449733,0.1785560846,-0.1685338765,0.0603563637,-0.1926646084,-0.0408711396,-0.3692883551,-0.0084927091,-0.3379705548,-0.2744612992,-0.0857492983,0.3298099637,0.4892736077,0.2108639032,0.0188965239,0.0051540616,0.1076816469,0.2748112977,0.0608406886,-0.4424714744,0.0144797135,0.0698906407,0.1044319049,-0.3181170225,0.1722688675,-0.3851829171,-0.0413322486,0.0806162804,-0.1244269609,0.0321558528,-0.1884502172,0.0097278189,-0.0740791783,0.0788282678,0.0203549433,-0.0319448821,0.282531172,0.4680353105,-0.4245634377,-0.4446209967,-0.2189762592,0.2427562773,0.3147049546,0.0773994699,0.4589720964,-0.1995372474,-0.1982268542,0.0224276986,0.0013363865,-0.2332913429,0.308386445,0.0870355591,-0.443601042,-0.064044565,-0.1276910752,0.0669273585,0.3167684078,0.006273441,-0.4122129679,-0.3120735288,-0.3458242118,0.0214615166,-0.0250640288,0.0159958452,0.033298634,0.1582906544,0.1738102287,-0.2094104588,-0.1639065892,0.176369518,-0.0739777386,0.0882997587,0.0552646592,0.0004676502,0.4113275409,-0.1642009616,-0.1075327545,0.1547516435,0.0445888788,-0.0944813043,-0.0453159884,0.1633691639,0.18625395,-0.0726311356,-0.0596916936,-0.2644861937,0.052268628,-0.4614967406,0.2731383145,0.3179630637,0.1334919035,-0.1683542728,0.1732744575,-0.141904369,-0.031213345,-0.0715209544,-0.1177848652,0.373078078,0.1653662771,0.2389952391,0.2501053214,0.0526352674,0.0117879491,-0.0419419408,0.1651650965,0.4098049104,0.138307482,-0.1994009018,0.0107082669,-0.0231417306,0.0838551596,0.1013283804,-0.0645256415,-0.4306384325,0.3228392005,0.0107785156,0.1113804132,-0.0769105405,0.1001283526,-0.1786119938,-0.042925138,0.4991569817,0.170720607,0.1455554366,0.0306864344,-0.0139356581,0.5526452661,-0.1769462973,0.0706767067,-0.1810891777,0.0004285282,0.5248662829,0.4290678203,0.0899589583,-0.0224515777,0.4933343828,0.140476957,0.097215496,0.3413825929,-0.0275382977,0.0665415302,-0.184803471,-0.0779664591,0.1536336988,-0.2290946841,0.2851954103,0.2491312474,-0.0172442459,-0.3904985785,-0.0501439944,-0.2299059331,0.2316945642,-0.2025544941,-0.2344719023,0.2113189995,-0.1458106935,0.1026369929,0.4035284817,-0.1777096093,-0.2079856247,0.192676425,0.1462937891,-0.268319428,0.3340502977,-0.0106526911,0.1251562685,0.1554382592,0.0702879503,0.7328562737,-0.4035183191,0.1121561602,-0.0930718035,0.1437781155,0.2503648102,0.4016755819,-0.2421788424,-0.2210508287,0.1754017621,0.0839866176,-0.1298199594,0.2136808485,-0.1136312336,0.0500252284,0.0996176079,0.0619367547,-0.0187563803,0.1787928641,0.1444265246,0.2375247478,0.2469804585,0.1737931371,-0.1421518475,0.1531006992,-0.1378596574,0.050131157,-0.3564988375,0.1108948961,0.2581638694,-0.2746676505,0.3892762065,0.3208779097,0.0001131504,0.2006645203,0.4254365265,0.3787307441,0.3124679327,-0.3084920943,-0.2201283276,-0.3098179698,0.3806098402,-0.0124031864,0.1874862015,-0.5203776956,0.0690181926,-0.0001826966,-0.051118575,0.0126315765,0.1607403904,-0.1287435144,0.1385003328,-0.302302748,-0.0964251086,-0.0995888263,-0.0191938318,-0.1124661341,-0.397503525,0.3286007941,-0.1192636639,0.0364226103,0.0316480808,0.0661184639,-0.2701520026,-0.1643407643,0.4762184024,0.2165219188,0.1547783911,-0.2523604929,0.033941593,-0.2433580607,0.0129390676,-0.3136281669,0.0196630489,-0.0127223935,0.3252117336,-0.2378225923,0.0975258201,-0.3865667284,0.0917084366,0.280357033,0.055234082,0.0762823075,-0.0562098734,-0.2146538943,0.0456930324,0.0900432393,0.5633693933,0.0695287883,0.274197489,-0.2228440046,-0.0457546376,0.2715085745,-0.3289845586,-0.5495134592,0.3473874331,0.0763195828,0.2181525081,-0.1747681648,-0.2520253658,-0.0071502095,0.1328004301,-0.2379005253,-0.1211836562,0.0403009765,0.0476477146,0.1299712807,-0.057671681,0.3296086788,0.1126061231,-0.3868745267,-0.0414467044,-0.2493579835]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"It looks like a pyarrow issue with google colab.\r\nFor some reason this code increases the disk usage of google colab while it actually writes into google drive:\r\n\r\n```python\r\nimport pyarrow as pa\r\n\r\nstream = pa.OSFile(\"\/content\/drive\/My Drive\/path\/to\/file.arrow\", \"wb\")\r\nwriter = pa.RecordBatchStreamWriter(stream, schema=pa.schema({\"text\": pa.string()}))\r\nwriter.write_table(pa.Table.from_pydict({\"text\": [\"a\"*511 + \"\\n\"] * ((1 << 30) \/\/ 512)})) # 1GiB\r\nwriter.close()\r\nstream.close()\r\n```\r\n\r\nMoreover if I `rm` the file on google drive, it frees disk space on google colab.","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":74,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nIt looks like a pyarrow issue with google colab.\r\nFor some reason this code increases the disk usage of google colab while it actually writes into google drive:\r\n\r\n```python\r\nimport pyarrow as pa\r\n\r\nstream = pa.OSFile(\"\/content\/drive\/My Drive\/path\/to\/file.arrow\", \"wb\")\r\nwriter = pa.RecordBatchStreamWriter(stream, schema=pa.schema({\"text\": pa.string()}))\r\nwriter.write_table(pa.Table.from_pydict({\"text\": [\"a\"*511 + \"\\n\"] * ((1 << 30) \/\/ 512)})) # 1GiB\r\nwriter.close()\r\nstream.close()\r\n```\r\n\r\nMoreover if I `rm` the file on google drive, it frees disk space on google colab.","embeddings":[0.0214205999,0.2508093715,-0.0044861692,0.4046226144,-0.0593296103,-0.0976660177,0.3005953729,0.0051357779,-0.2152584493,0.1271965504,-0.0451286696,0.3386108279,0.1068906263,0.2934122384,0.0605559871,0.223590672,0.2913432419,-0.0080436692,0.1109899879,-0.0101503739,-0.2467477173,0.3662295043,-0.2008794248,-0.0313707292,-0.4776650369,-0.068877548,-0.0705260411,-0.1316779405,-0.032523375,0.1094124839,0.1837549657,0.0785785317,0.150113225,0.5543874502,-0.0001274843,-0.2014526725,0.2319876403,-0.1646427512,-0.1672947705,0.0952369049,-0.0660738647,-0.0158476364,0.0897296444,-0.0740431994,-0.0477135181,0.0367304012,0.2621205449,-0.1535121351,-0.0225636475,0.1620317549,0.0549754538,-0.3191696107,-0.3850808144,0.4330038726,0.3903363049,0.2227914929,-0.1758393198,0.5310472846,0.076160714,-0.1830207855,-0.3298393786,0.2640752494,-0.1264401525,0.110678345,0.3018000126,0.4598481655,-0.2333257943,-0.1505207419,0.4259735644,-0.4988749921,0.3272492886,-0.5334558487,-0.002896925,-0.3041200638,-0.2905580401,-0.360919714,0.514906764,0.2798948586,-0.2408690751,0.0373578258,-0.389555037,-0.1254132688,0.0084921056,0.0503370427,-0.1800255924,0.1711009294,-0.122367993,0.0516222566,0.0287987366,0.3116212785,0.7678453326,-0.3648163378,0.0661570206,0.1391303986,-0.0734800249,0.1814907938,-0.1290129423,0.585472405,-0.1312903464,-0.1340878308,-0.1266849637,-0.1530451328,0.0074453205,0.1737578511,0.1315623969,0.3484873772,-0.2132810354,0.0425846986,0.0428300127,0.0529006533,-0.3369436264,-0.0782317445,0.0316968448,-0.2644616663,0.1953683048,-0.0778724998,0.0231226273,-0.2491862327,-0.1689238101,-0.2116445005,-0.3935553432,-0.0601108558,-0.0610288493,0.2024491578,-0.0228124522,-0.2642136514,-0.2638733089,0.2894004583,-0.2768119872,0.357504189,-0.0012719117,0.0600019097,-0.2810415626,0.5400337577,0.1548738033,-0.179901123,0.0203176383,0.1410446763,-0.0431125313,-0.2354217023,-0.0890352577,-0.4710323215,0.2895697653,0.191419825,0.1892023832,0.3132681549,-0.1785362363,-0.2473843545,-0.1625051349,0.4494559765,-0.5062214732,-0.3123239577,0.1799287349,-0.0252877772,-0.2993509471,-0.0670899823,-0.7675716877,0.1321822405,0.447485894,-0.2971973121,0.0309301317,-0.096560739,-0.4285757542,-0.3032741249,0.0149896126,0.1452413499,-0.359757185,0.0603813492,-0.1958555281,0.4015370607,0.4847250283,0.4746412337,-0.0242268052,0.3547291756,-0.4302553833,0.0213741045,-0.0106692435,-0.1567398459,-0.8236618638,0.1317311972,-0.132392332,0.0695866272,0.1081514135,0.0055570155,0.158888191,-0.0654986799,0.2246363759,0.2069316357,0.0737835467,0.2635890245,-0.3758739531,-0.1832395643,-0.0158579573,-0.0594401397,-0.0387478173,0.020846229,0.0070683332,-0.4193621874,0.2358665317,-0.1587572396,0.1791306287,0.2395592928,0.3115640879,-0.2237321734,0.0086980397,0.2210537195,-0.259278357,0.0565233119,-0.0845233351,-0.1852927953,-0.6225892305,-0.1273642927,-0.1777253449,-0.0660640448,-0.1020753458,-0.192406714,-0.0981779769,0.2356915325,0.1794316024,0.0773903951,0.1373386532,0.3301828504,0.1836962551,-0.1190631017,0.0269475933,0.3751268089,-0.1567516774,-0.3961032331,-0.0702992603,-0.1427293718,0.2049590796,-0.159670487,-0.0452094115,0.0729753524,0.1840246469,0.4450157285,-0.0777895376,0.2858203351,0.1316982657,0.0216108579,0.1732712388,-0.0894047916,0.01382716,0.0375896022,-0.036116533,0.3025759757,-0.0142473662,0.1336650997,-0.1301425695,-0.3485519886,-0.1377633065,0.0490126163,0.3428657651,0.085674867,0.0891671255,0.5186516047,0.4237043262,0.2823845744,0.1714913249,0.3646704853,0.6328815222,-0.0050880318,-0.0541194044,0.257504791,-0.2285197824,-0.3635298908,0.271355778,0.2931195199,0.4082559943,-0.013962456,0.1979498714,0.1831991225,0.1056673825,-0.2301922888,0.2001953125,0.1337825209,0.2639008164,-0.0150433686,0.3801979721,-0.1243999153,-0.1518711299,0.1923526824,0.0603036918,0.1992247552,-0.209303081,0.5275095701,-0.1733888984,-0.2081394345,-0.2764720917,-0.2005175054,0.0451708846,-0.1497702003,-0.0626137704,0.4318827093,0.0306840912,-0.0486451127,0.3401047587,0.1422446668,-0.162133202,-0.5259186625,-0.1383045167,-0.0220133737,-0.3286528587,-0.0774949044,0.4027002752,-0.0441007689,0.0771534592,-0.1384210736,0.0884591714,-0.3343403339,-0.1214695647,0.2740837634,-0.1896559149,0.3587255776,-0.3514130414,-0.1115452647,-0.4793860316,-0.3280753493,-0.1017479524,-0.0617399886,-0.114281103,-0.3009617925,-0.0578712933,0.0006670832,-0.0771662444,0.0741540119,-0.1301366836,0.020014409,0.1932437271,-0.0068307747,0.3806487024,-0.0050653028,-0.0432363115,-0.1388840973,0.3479846418,-0.3142756224,-0.2478320003,-0.4356377721,0.2898030579,-0.0098420447,-0.055216305,-0.0459030308,-0.1516476423,-0.042331405,0.4133680463,-0.5115941763,-0.3616776168,-0.1239698231,0.2218833715,0.1521575153,-0.110673748,0.2670279145,0.1321386248,0.1791777015,0.0847617015,-0.4711603224,-0.1082860529,0.3099479377,0.2349284887,0.0253391843,0.0066287001,0.0264174417,0.6789851189,-0.0198745523,0.0968693197,0.5025283694,-0.0739436969,0.2042860091,-0.1957439482,-0.0140864467,0.1454436928,-0.1160622314,-0.6394329071,0.3694887459,0.0848572478,-0.4844223857,0.2561872602,-0.2750031948,-0.2847714424,-0.3084435761,0.2834481597,0.2377592623,0.1873607188,0.0623418875,0.037746951,-0.1657356471,-0.3295600712,0.0454921164,-0.0543028302,0.2175507694,0.0898288339,0.3859018981,-0.4134770632,-0.4909831583,0.2205979377,0.3826884627,-0.1792142987,0.0343789868,0.2228515297,0.3623178005,-0.2491298914,0.558006525,-0.0694044754,-0.1339288205,-0.1462883651,0.1664867997,-0.1333398074,-0.0457114428,-0.0675492957,0.2604496777,-0.0163353886,0.3377733529,0.1728349328,0.0868892297,-0.0843029395,0.3391499817,-0.283531189,-0.0867726058,-0.1964409798,-0.7121967673,-0.0780667365,0.0588036031,0.1275708973,0.0980687216,0.0418757908,0.1731925607,0.2353619635,-0.0124085359,0.0576721802,0.1508389264,0.5057429075,0.0489221066,-0.1249409318,-0.006402371,0.0524299294,0.4683277905,0.487996608,0.3116280437,0.2212163657,0.1519512832,-0.1782701313,0.239008531,0.2694661021,-0.1086313128,-0.106631577,-0.2813803852,-0.0630207285,-0.2542047501,0.193458423,0.1144245565,0.2429378778,-0.2523652613,-0.212758258,0.5412887931,0.1812486351,-0.3203702569,0.165717721,0.0798251256,-0.5004062653,0.4107103646,0.1969547123,1.0303806067,-0.0471589305,0.3019300699,0.3087443113,-0.1668759882,-0.1174321696,-0.2488428801,0.3039296865,-0.2209102958,-0.5396696329,-0.0110264402,-0.2475277036,-0.1905698627,-0.0188333765,-0.2668173611,0.3062217534,-0.0438305326,0.0509454422,-0.0009036392,-0.1211186796,0.0458542779,-0.1328752041,-0.2634564936,-0.0213887505,-0.0880284086,0.2598496974,-0.1332017928,0.1381903887,-0.1858516783,0.0161058661,-0.168046236,-0.0160227139,-0.4394675195,-0.0721307024,-0.2420898676,-0.3970457613,-0.1454384625,0.282337755,0.2300014049,-0.0147827696,0.0704175159,0.002057482,0.1275438368,0.2864452302,0.0846182182,-0.4260691404,-0.144591555,0.0226753168,0.0339435041,-0.2728813589,0.1001212075,-0.3597154617,-0.1363663375,0.2026909441,-0.1626833975,0.0294711217,-0.0940956101,-0.1133505851,-0.0550969504,0.0660780519,0.0095067918,-0.0463662893,0.2639182806,0.4637506604,-0.4630711973,-0.3630080223,-0.1878152341,0.2531951666,0.4450027049,0.2145126611,0.5582187772,-0.2028559595,-0.2048253566,0.0576998144,-0.1367766261,-0.223781988,0.120476298,0.1326523125,-0.3732170463,0.0511844121,-0.1787276566,0.0684328228,0.2830272615,-0.0346837677,-0.4491626322,-0.1557337642,-0.3818812668,0.0471600741,0.1669677198,-0.0874844342,0.1780773252,0.1550747901,0.2216195166,-0.0953017846,-0.1456442475,0.2007080615,0.0055035301,0.2432675809,0.0172682367,-0.1123381257,0.4713480771,-0.1918706894,-0.1179881692,0.1168155223,0.1549628228,-0.1169964671,-0.0153459013,0.1827224046,0.1044028848,-0.1024020612,-0.1249281988,-0.1001676992,0.2199304998,-0.4058310986,0.3253194988,0.2486814708,0.1182093695,-0.2841389477,0.2652656436,-0.0952177793,-0.1122512817,-0.1452152133,-0.1768474728,0.3615041971,0.2094568014,0.2111814618,0.2580582798,0.0975412428,0.0806852579,-0.0012959536,0.2764171064,0.3049297035,0.0272615254,-0.2269961387,-0.167079255,-0.1834324598,0.0302474722,0.1225222796,-0.0035720614,-0.495803833,0.3037512302,0.0041007693,0.1544913799,-0.1203544289,0.1704882234,-0.257729739,-0.1299338937,0.4528477192,0.2115076035,0.1399543732,-0.0298390817,-0.0397718586,0.5303693414,0.0392148122,-0.1085442379,-0.1534610689,0.0309161097,0.4009896815,0.4879708886,0.1087321788,0.0028848737,0.2444178611,0.0172583163,0.0839073434,0.2110325396,-0.0905224234,-0.038835723,-0.1549651027,-0.1460747421,0.2240422219,-0.0110138031,0.1774436384,0.2570498288,0.1577571034,-0.1078217328,-0.0595412366,-0.1887115687,0.2701652646,-0.2121588141,-0.2904267609,0.2382966578,-0.1484885067,0.0976770669,0.4762371182,-0.2112547904,-0.2921563983,0.3226988018,0.2024369985,-0.2835077941,0.2390041798,0.1534790099,0.1942007393,0.1675261855,0.0098429406,0.5851386786,-0.4067664146,0.034578681,-0.189821586,0.1489820927,0.4157559574,0.3311626911,-0.2545661628,-0.2914428413,0.0924551636,0.127885133,-0.13094455,0.1330518872,-0.0176011585,0.2327791154,0.0003343276,0.0493427962,-0.0455655344,0.168859154,0.2412270159,0.1468367577,0.1848755479,-0.0148403468,-0.1873381436,-0.0714157596,-0.0255148001,-0.0618173741,-0.2494096607,0.1596643925,0.2407691777,-0.2312214971,0.301276803,0.310469985,-0.0102103343,0.1933280677,0.5336735845,0.3685654402,0.3064676821,-0.3856616616,-0.1242970005,-0.2901498675,0.377302438,-0.2015384138,0.1256184727,-0.3924018741,0.1516054273,-0.1680917889,0.0277754311,-0.0537368841,0.3108820915,-0.1991760284,0.0368190296,-0.3342728019,-0.0064273896,-0.0134857688,-0.1722969115,-0.1404057294,-0.4480997026,0.3833130002,-0.2276224494,0.0577225611,0.0032165102,0.0626593083,-0.1717342138,0.0512426049,0.5267401934,0.1537847519,0.1620921195,-0.1921782047,0.0589524172,-0.2873102725,-0.0165700112,-0.3212079406,0.0060098758,-0.0835126713,0.2568944395,-0.2164873779,0.075267911,-0.3169200718,0.3819307983,0.2212302238,0.0362061523,0.371244818,-0.1140624955,-0.1697267145,0.0451945849,-0.11739669,0.346899569,0.0570163243,0.235634312,-0.4045066833,-0.0970254093,0.2255552858,-0.435323,-0.4823940992,0.4828671515,0.1074682176,0.299225688,-0.2077957839,-0.2479835749,0.0159881711,0.1334010363,-0.1762092561,-0.2149590999,0.141854912,-0.0739011765,0.0904602408,-0.0241559781,0.2754199207,0.0608072206,-0.4763140678,-0.1921880096,-0.300022155]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"It looks like replacing `pa.OSFile` by `open` fixes it, I'm going to open a PR","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":15,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nIt looks like replacing `pa.OSFile` by `open` fixes it, I'm going to open a PR","embeddings":[-0.1422462761,0.1986719072,-0.0383415967,0.4084185958,-0.000056138,-0.059415888,0.2414347976,-0.0347257592,-0.0797591358,0.2280797362,-0.0223413855,0.2935992777,0.0998644978,0.4326467514,-0.085689418,0.2728121579,0.2725432813,-0.0640387535,0.1511678547,-0.019179171,-0.4347856939,0.3687989712,-0.2093541473,0.0260250121,-0.413898617,-0.1674830317,-0.1695943773,0.0356652103,-0.0862059221,0.1600921154,0.1018451452,0.0457544327,0.2067367733,0.6060429811,-0.0001277732,-0.1291867942,0.1854427308,-0.2101528198,-0.1699630767,-0.0673597008,-0.2280113548,0.0577390939,-0.1482831836,-0.0175693128,-0.1724234372,0.0106286779,0.3100663722,-0.48338449,-0.046771545,0.307231456,0.0535877571,-0.4076635838,-0.4075746834,0.3484997749,0.0977480263,0.2199754864,-0.1175883785,0.2327809632,0.0364818424,-0.1251986325,-0.1038886011,0.2282625437,-0.1174328253,0.0184736028,0.2055832148,0.3473006785,-0.2769013047,-0.2822091877,0.5518472195,-0.4264921248,0.3108395934,-0.3668774366,-0.1756084859,-0.2863199115,-0.3604894876,-0.4540383816,0.5563179255,0.2172890753,-0.1113750488,0.1160843149,-0.4088926315,-0.2206466645,0.0850178152,0.0851949826,-0.2766344845,0.3347867131,-0.2142973691,-0.0023257812,0.0831820592,0.2392783165,0.7119539976,-0.3475367129,-0.0151814837,0.1099161953,-0.04188703,0.0897475705,-0.158896789,0.5635985136,-0.2339966744,-0.0859678835,-0.3095452487,-0.1033949777,-0.216552645,0.0832411945,0.0350993685,0.3935157061,-0.1172950342,0.2156011909,0.0433319211,0.0907309502,-0.5223142505,-0.0509425998,-0.0023959228,-0.3334251046,-0.0524709299,0.111481607,0.0042617847,-0.293517977,-0.184518218,-0.0441469066,-0.3691815734,-0.0071516032,0.0517111942,0.3119595349,-0.0356405936,-0.0535083376,-0.1621025503,0.2497843951,-0.3047916889,0.3533391356,-0.0553682223,0.0202600863,-0.3277935088,0.5002159476,0.0343521275,-0.2283872515,0.1167821661,0.0170202702,-0.0154798497,-0.3201924562,-0.0323863477,-0.3897449672,0.3616607189,0.2101402879,0.1411185116,0.2772832215,0.0086917644,-0.2384637892,-0.0605964251,0.3146103024,-0.5840646029,-0.3119014204,0.1744987071,-0.0079071708,-0.2813263535,-0.1145861298,-0.5771445036,0.0865116715,0.399928987,-0.2811204791,-0.0266458374,-0.0705791712,-0.5173348188,-0.2211597115,0.0847556964,0.1141372845,-0.2487306297,0.2344523966,-0.2765614092,0.456158638,0.517649591,0.3933976591,0.0849183723,0.2573323548,-0.4998368323,-0.111710988,0.0785356238,-0.2734696567,-0.8139237761,0.2209342718,-0.0632194653,-0.0456420146,0.079459928,-0.1072182804,0.1865637153,-0.070211798,0.1823238581,0.1172748432,0.0643767342,0.2058708519,-0.2898390591,-0.1529694051,-0.0741849616,-0.0766907409,0.0320389383,0.1734116822,0.0119448705,-0.3798746169,0.2088817209,-0.1318611652,0.0997910127,0.2371345609,0.2858581543,-0.1157668382,0.0022213622,0.1314415932,-0.0968367383,0.0819292143,-0.0178206153,-0.1834215671,-0.5371497273,-0.1023863181,-0.1001651436,-0.1319344789,-0.1256255209,-0.2693501711,-0.1046035141,0.3369643986,0.1086574271,0.0557807982,-0.0425043292,0.4625955522,0.0752147883,-0.1314198077,-0.0676162094,0.2037629932,-0.1267394274,-0.207560569,-0.16725263,-0.1013979316,0.1836040318,-0.2338981926,-0.1515571922,0.188140288,0.3672493398,0.3668425083,0.051936999,0.3207263052,0.1042421907,0.0526043549,0.1189633608,-0.1081740186,-0.0683423653,0.1871767193,-0.1744893938,0.3758814931,-0.0536635816,0.1333827078,-0.063624166,-0.2601850927,-0.0748906955,0.085040085,0.2456500381,-0.08840774,0.1104600057,0.3789637387,0.4048433006,0.2822603881,0.1820171624,0.3573617935,0.5984858871,0.0325714573,-0.0868402869,0.1448584944,-0.0651772618,-0.2888077199,0.2270188928,0.4577192068,0.4406841397,-0.0414961874,0.232239753,0.2452438623,0.0631716028,-0.15523085,0.1744717956,0.1052293628,0.2572918832,-0.0047500581,0.2450516522,-0.0267407466,-0.0734094903,0.1420776397,0.0373859666,0.2701141238,-0.1314072907,0.5008135438,-0.2842007577,-0.1946406811,-0.2102927864,-0.1599930674,0.0943995863,-0.151170522,-0.1013871655,0.3433909118,0.0006402897,0.061652381,0.2011966407,0.2052982748,-0.1504097581,-0.4438265264,-0.0852102861,0.1282130182,-0.2280768752,-0.1032098532,0.3800701201,-0.0858365595,0.0378005691,-0.2371272892,-0.000076067,-0.4548173845,-0.0977833569,0.2738337219,-0.0867065713,0.2565067708,-0.254783988,-0.0352032371,-0.3800007403,-0.3792601526,-0.0911662877,-0.0368356705,-0.1166639999,-0.4200726449,-0.0774458796,-0.060431011,-0.078549929,-0.0877402872,-0.1520898044,-0.0974306837,0.1071541458,0.0327853002,0.2696706355,-0.0008746487,-0.0596933477,-0.2264128625,0.4060647488,-0.3458912373,-0.3902064562,-0.5710648894,0.1770489514,-0.1157949492,-0.0636668801,-0.0761626288,-0.0867176652,0.0481450297,0.3855960965,-0.4914422631,-0.3674526215,-0.1664719582,0.0240985844,0.1457050294,-0.1004504189,0.3305413723,0.0796436071,0.1804222018,0.0177433994,-0.4733370841,-0.0063891308,0.360966742,0.3901810348,0.0165281612,0.2349520177,-0.0318283513,0.6451180577,0.2406005263,-0.0005716826,0.4065792859,-0.1019583866,0.3029135466,-0.2793043852,-0.0734123364,0.2883846164,0.0304166283,-0.713496089,0.4362022877,0.1024607047,-0.4593780041,0.1621853411,-0.1576177627,-0.3234657347,-0.2360782325,0.2377761304,0.0102404514,0.2164751291,0.0608397387,0.1850255877,-0.2700043321,-0.1922589242,0.0680879205,0.0637985989,0.3217544854,0.1330997944,0.2887643576,-0.4332284331,-0.3722419143,0.3504219353,0.3609678447,0.0211262126,-0.0838944912,0.2518072426,0.2384886444,-0.1416171193,0.6589196324,-0.3321660459,-0.1907575279,-0.0690461099,0.095364362,0.0616944395,-0.0138185965,0.0208641887,0.3668631315,-0.0545157492,0.5082979202,0.2708223164,-0.0454920232,-0.1174647212,0.2194956541,-0.2853343189,-0.1476837695,-0.2132110447,-0.5835343003,-0.224505648,0.0180052072,0.1479576826,0.0136877028,0.0196397267,0.0940876678,0.1315902472,-0.0234412923,0.0514240712,0.1370977759,0.4452194571,0.1481254399,-0.1291105151,0.1241603643,0.2570526004,0.4552251995,0.5559539199,0.1767675579,0.1278216839,0.2849138975,-0.1611864269,0.2446209788,0.3518234193,-0.1343605816,-0.2747298777,-0.0385375209,0.0131936138,-0.3020339012,0.3487410247,0.142703101,0.3552173674,-0.4277957976,-0.4335630536,0.3987047672,0.1834510714,-0.3302587271,0.2861592174,0.058050964,-0.3915999532,0.5959499478,0.1556306928,1.1384776831,-0.0769858211,0.1892806441,0.1240559369,0.058353316,-0.0031137324,-0.4171907902,0.2900340259,-0.1768390834,-0.6259695888,-0.0584304631,-0.1869707108,-0.2189318836,0.0536282621,-0.2121810615,0.3463557363,0.1185621023,-0.1608695388,-0.0414642356,-0.0930934995,0.0032225584,-0.1204354763,-0.2353170663,0.0242516063,-0.0191810504,0.3901158273,-0.1369218081,0.182342127,-0.1488928199,0.1228972822,-0.2331905067,-0.031494271,-0.4409198463,-0.0328537486,-0.2758102119,-0.2513308227,-0.1786197424,0.4671134055,0.4909790754,0.1350540817,0.0193969868,0.0931423083,0.1353033632,0.2252166122,0.0942476541,-0.4611389935,0.0457771122,0.0526818037,0.0953730419,-0.2841707766,0.1204470471,-0.4322350323,-0.0339371487,0.0710949674,-0.0546355471,0.0055161817,-0.2037636042,-0.0434729531,-0.0900370777,0.0880717114,0.0218005143,-0.0121136922,0.2980711162,0.4234382212,-0.4189558923,-0.476545006,-0.230147168,0.2171956301,0.2243037522,0.0727388114,0.4416924417,-0.2073661536,-0.1893511266,0.0259845182,0.0225380082,-0.3053992987,0.368937403,0.0585906766,-0.3887958825,-0.1417413354,-0.0893315077,-0.0042166379,0.3380320668,-0.0040881261,-0.4160790145,-0.3144710362,-0.4086825848,0.0163096078,-0.0656821877,-0.0002444711,0.0658371821,0.1624255776,0.1061075479,-0.2829619348,-0.1668298393,0.1825982481,-0.1080674976,0.084744297,0.0251169093,-0.0752402693,0.3745762706,-0.1935940683,-0.0880492106,0.1043545306,-0.0225732941,-0.0909555703,-0.0425710082,0.1779299825,0.1841085702,-0.076868996,-0.0246869065,-0.2101166099,0.0590713359,-0.4563265443,0.2043543309,0.3002198935,0.1062506065,-0.2785976827,0.1827292293,-0.0827053338,-0.0563060939,-0.1001119837,-0.1271758676,0.383801192,0.1521463394,0.2458575815,0.1665582657,0.0431690849,0.0943585113,-0.0719715133,0.2739419937,0.4122424126,0.1204453409,-0.1605670452,-0.0337779596,-0.0360322744,0.1504793167,0.1403886527,-0.0612643547,-0.347091794,0.372877717,0.0275116414,0.1063742787,-0.0336705521,0.1889612526,-0.1593415141,-0.0325366296,0.4742387235,0.1866052598,0.1821029782,0.0804035962,0.0053271209,0.5819516182,-0.2198649347,0.0382110775,-0.1925040632,0.0337808877,0.5114142299,0.4418691695,0.1380953789,0.0030082935,0.530390501,0.1042935401,0.114467904,0.3726651371,0.0205049813,0.0345111154,-0.2604325116,-0.0694282353,0.0904657096,-0.2208849043,0.2243537009,0.246094346,-0.0237118695,-0.4161904752,-0.0826132745,-0.2048965693,0.2362450212,-0.1921284497,-0.2881277204,0.2334345281,-0.1257381737,0.1312112957,0.4411702156,-0.1343640089,-0.1849569678,0.1753748953,0.1119790748,-0.2571334839,0.327853173,-0.0471952409,0.0567835048,0.2232379168,0.0393877327,0.7306486368,-0.4694347382,0.0573474281,-0.0416257493,0.1692617685,0.343141675,0.3862788677,-0.3166682422,-0.1813006103,0.068496272,0.123988688,-0.1402437538,0.1842201054,-0.0822338015,0.0880178064,0.0789825395,0.0700456128,-0.0376378186,0.1894360334,0.0890753716,0.2039782703,0.2179751098,0.0654241145,-0.0860233977,0.1174592301,-0.1037731171,0.0394556299,-0.3149829805,0.1539651752,0.1531787962,-0.2533536255,0.4138925672,0.4065106213,0.0066892924,0.2526888549,0.3371219039,0.4522140622,0.3687004149,-0.2678707242,-0.1686691046,-0.3361262679,0.4362926781,-0.0028024761,0.1401071846,-0.4850183725,0.0318362936,0.0218570791,-0.0110365348,-0.0315673277,0.2673967779,-0.090834409,0.1765503883,-0.3612294197,-0.0543560833,0.005726241,-0.012753183,-0.1100380644,-0.4061972499,0.3412047923,-0.1115328074,0.0313159972,0.0874759927,0.0890932605,-0.269874692,-0.0917785838,0.4562672377,0.2106792629,0.1405868679,-0.2840675712,0.0296387598,-0.3148221374,-0.0061899303,-0.3753836751,-0.0092228288,-0.0226784758,0.2780998051,-0.2761690021,-0.0283657182,-0.3534768224,0.1527042538,0.2461850494,0.0644579157,0.1038305238,-0.0187361035,-0.1375332922,0.043624796,0.117470853,0.5533705354,0.0471012071,0.2553424239,-0.2453683764,-0.048362527,0.2990801632,-0.3552477956,-0.453869611,0.4186009169,0.051539354,0.2637205422,-0.1466786861,-0.22936517,-0.0075373068,0.084145464,-0.2654533684,-0.2002840638,-0.0499896556,-0.0440648645,0.0867281333,-0.0719552115,0.3976902664,0.0980238914,-0.4515044391,-0.054930266,-0.1866414249]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"Ok. Thank you so much!","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":5,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nOk. Thank you so much!","embeddings":[-0.060456261,0.1673737019,-0.0294772144,0.4371073842,-0.0683027133,-0.0368534848,0.317402333,0.0144377705,-0.0077836411,0.203405425,0.0204397291,0.2663175166,0.0607032105,0.4282112122,-0.0756157339,0.3190076649,0.2755341232,-0.0783572495,0.0503138751,-0.0926943719,-0.3528110981,0.3793799579,-0.2349521965,-0.0225024093,-0.4681852162,-0.1239464059,-0.1558191627,-0.0623899437,-0.0554476902,0.2577373683,0.2080550641,0.0255805217,0.1912323087,0.6428728104,-0.000128333,-0.167806074,0.2171703726,-0.2450951934,-0.1822337657,-0.09825176,-0.3136050999,0.0155748865,-0.1237585023,-0.0666165054,-0.2132172734,0.1130454019,0.2252760828,-0.5501648188,-0.0095465602,0.2673203647,0.0686934292,-0.4483325481,-0.4974781871,0.3798108399,0.046102982,0.2365247011,-0.1136518046,0.2773808837,-0.0112319579,-0.123082146,-0.0741559789,0.322047323,-0.0628916994,0.0154413376,0.2451543212,0.3284511268,-0.3148850501,-0.3024646044,0.525370121,-0.4499563873,0.3710609674,-0.3346081078,-0.1613344699,-0.2619041204,-0.3450641334,-0.346899569,0.5163353682,0.1443961263,-0.0990760401,0.0689156353,-0.5074322224,-0.2204099894,0.0699379146,0.0985092446,-0.2336286157,0.2928539515,-0.2735627592,0.0345008858,-0.0605346337,0.2344711572,0.6727767587,-0.3931878805,0.0138168707,0.2265203893,-0.0922441259,0.0641538873,-0.1834435463,0.5556010008,-0.2196488976,-0.1153276414,-0.2517289519,-0.1300277561,-0.1297519207,0.0963978246,0.052443888,0.4614912271,-0.035674125,0.1438517272,0.0700728744,0.0331230164,-0.5346032381,-0.098897174,-0.0233360119,-0.2653363943,-0.020941468,0.1335074306,-0.0389438495,-0.3010451794,-0.1093675643,-0.0452947281,-0.3292690814,-0.0301808957,0.1019481197,0.2871227264,-0.0410419852,-0.117310293,-0.2480662763,0.2609939277,-0.2055173814,0.3430996835,-0.0979407504,0.035608314,-0.3575517535,0.4629116654,0.070852913,-0.2481508106,0.1313719153,0.0166792627,-0.0092599085,-0.3316574693,-0.0551571026,-0.4371404052,0.3936647773,0.1404282004,0.1467010975,0.4134150743,0.0199359972,-0.2232182771,-0.1305555254,0.2400447875,-0.6035102606,-0.2666142583,0.1969403923,-0.0207188167,-0.269929111,-0.0823049173,-0.7065373659,0.1168561727,0.4764822125,-0.1968692094,-0.0012489705,-0.0145497508,-0.4899199307,-0.2396948636,0.0598827451,0.1181028262,-0.2469644397,0.1786674112,-0.3512824476,0.5111804008,0.5124985576,0.3832536638,0.0617158227,0.2783671021,-0.4794674218,-0.032861203,0.1103192568,-0.3036438227,-0.8256471753,0.321090132,-0.0373037718,-0.0114192348,-0.0120509602,-0.0546894744,0.2163040489,-0.1069970876,0.1838378161,0.1495525688,0.0930471271,0.1950432211,-0.2551704645,-0.1260472834,-0.0277917329,-0.1594712138,0.0319906957,0.2113371789,-0.0062828353,-0.3772325516,0.153066799,-0.1193877682,0.1384143531,0.2503773272,0.2501627803,-0.1377369165,-0.0087174037,0.1499860287,-0.1640667766,0.1378437579,0.0113901552,-0.1516707689,-0.5148812532,-0.066216372,-0.1218434498,-0.1658355147,-0.1372597218,-0.2341240793,-0.0784800351,0.3036627173,0.129999131,0.0466882139,-0.0341684185,0.5397333503,0.1032712832,-0.1655304432,-0.0304866303,0.2149823308,-0.1159545854,-0.2266227603,-0.1394385695,-0.159174189,0.256300956,-0.2202827185,-0.188075304,0.1363699734,0.2421599627,0.4023713768,0.0067340243,0.2738153636,0.0878415331,0.043854259,0.188293159,-0.0844165087,-0.0225113835,0.1623814255,-0.0375132822,0.3175189197,-0.0864669904,0.1687324345,-0.0932536498,-0.2730309665,-0.0628979728,0.0997424871,0.259542495,-0.151327461,0.1096719727,0.3349812031,0.383395493,0.248559311,0.2272684127,0.3417497873,0.6254071593,0.0774663761,-0.0242596325,0.1572722942,-0.1064121053,-0.2654297054,0.2465580553,0.473089546,0.4310381711,-0.1037662625,0.2897805572,0.1822990775,0.0891337022,-0.1331520528,0.1576814353,0.0572858155,0.2966574728,-0.019210849,0.2161849439,-0.0646691173,-0.0395767428,0.1827790588,0.0736656189,0.2401298732,-0.1200137436,0.5599375963,-0.2864474356,-0.2519131899,-0.2447170615,-0.2965691984,0.0666856319,-0.1773428917,-0.1172654778,0.3156601787,0.0425121002,0.038491074,0.2064196318,0.2040866911,-0.1411139369,-0.4638777077,-0.1500238925,0.1620169878,-0.2479168475,-0.1076444909,0.3529206514,-0.0566851683,0.0444393642,-0.1800248027,0.0654126108,-0.3621139824,-0.1017075032,0.2641431987,-0.1533628106,0.284496516,-0.2675742507,0.0230425373,-0.4210949838,-0.3491891921,-0.0831714272,-0.0317879952,-0.0842708349,-0.4205414653,-0.0307926089,-0.0372937769,-0.0755586252,0.0040332926,-0.188416034,-0.0132106189,0.0746422186,0.0064740879,0.2865063548,-0.0455820411,-0.1108236462,-0.2374458462,0.3854686916,-0.3026976585,-0.3842828274,-0.6341551542,0.2302432358,-0.1270264387,-0.0381801389,-0.0270582512,-0.0756087899,0.1181234494,0.3679939508,-0.453256011,-0.2758109272,-0.1491301358,-0.0410162508,0.2226082832,-0.1009749323,0.2913258076,0.0599741675,0.2083091438,-0.0328717716,-0.4329028428,0.0133674555,0.3180939853,0.3647685647,0.0532612391,0.2366448343,0.0465546064,0.6147227883,0.3157847524,0.0006196761,0.4315518439,-0.0236376077,0.2186195552,-0.2482051849,-0.141903311,0.2577672899,0.0356424637,-0.6475072503,0.3963169456,0.1195697039,-0.4500663877,0.1519375145,-0.2642638981,-0.3512252867,-0.2197337449,0.2121065259,-0.0032318651,0.2208560705,0.0800814629,0.1301181316,-0.2455980033,-0.2300494909,0.0878931135,0.0677069277,0.315030992,0.1584913731,0.2638950944,-0.5013681054,-0.3820397556,0.3400842547,0.3651652336,0.050528504,-0.0695242435,0.3062641025,0.1912072301,-0.1309281141,0.5714796782,-0.3397242129,-0.1749141663,-0.1701399684,0.0151945883,0.0311949067,0.001749487,0.0987669304,0.3884434104,-0.0513521619,0.4664028287,0.3263767958,-0.001717097,-0.1186120957,0.2090861946,-0.3313829303,-0.1450414211,-0.2094061077,-0.5329959393,-0.2008071542,-0.0029002514,0.1550676972,-0.0022595467,0.0805982128,0.0246182792,0.1240513325,-0.0578742623,0.0681305528,0.1499736011,0.4108996689,0.1605228335,-0.1281080246,0.1227666512,0.2180579305,0.3557561636,0.5134746432,0.1696090698,0.1748543382,0.1786006391,-0.1705738902,0.314306736,0.3289672434,-0.1720479429,-0.2153863311,-0.0842613727,-0.0419908352,-0.3513582051,0.3441999257,0.1442429721,0.3688139021,-0.4380836189,-0.4464225173,0.4601237774,0.1763325632,-0.2980312705,0.3435010314,0.0611300021,-0.512902379,0.4203824699,0.1444526166,1.1332173347,-0.0103692682,0.1552142352,0.0624827929,0.0950395167,-0.0427036695,-0.3295324147,0.2294168472,-0.1786450595,-0.6377150416,-0.1085321456,-0.1734284908,-0.2028935552,0.094775632,-0.1671697348,0.4362716973,0.0200463589,-0.0590527132,-0.0944550931,-0.0961731374,-0.0241844859,-0.1448550373,-0.2626713812,0.0357833765,-0.0397621877,0.4351899922,-0.1547851115,0.1855671704,-0.1714091301,0.0708350018,-0.1776671112,-0.0454940125,-0.3660924733,-0.0013639994,-0.3501954973,-0.2830081582,-0.1316222847,0.355617553,0.5147147179,0.2170017958,0.0166213978,0.0069479863,0.1162648946,0.2386836112,0.041979,-0.4373522103,0.0316480733,0.0378604978,0.0861163884,-0.3191848695,0.1702525467,-0.4062229991,-0.0293424092,0.0704087466,-0.1039343253,0.0386746265,-0.2192459404,0.023081962,-0.0806664303,0.1014169678,0.0251288116,0.0017692794,0.2629027367,0.4885014892,-0.434504807,-0.4152848125,-0.2243080586,0.2205363512,0.2989364564,0.068042241,0.4763426483,-0.192338869,-0.2190201879,0.0197377615,-0.0063820793,-0.2387237996,0.320954442,0.0656407177,-0.4401195347,-0.0714353025,-0.139489606,0.0455860086,0.3217843473,0.0227565784,-0.4103934467,-0.3189332187,-0.3720911443,0.0392180495,-0.0419935361,0.0076335133,0.0399075821,0.1591164023,0.1812925786,-0.200971067,-0.1711984873,0.1661234945,-0.0983568728,0.0911052153,0.0697882548,-0.023231566,0.4152715504,-0.1669968516,-0.0962416306,0.1486771703,0.0509158671,-0.0961405188,-0.0322142132,0.1683685035,0.2035573572,-0.072530821,-0.0320647806,-0.267170459,0.0556815267,-0.4596445262,0.221500203,0.3056057394,0.1351202428,-0.1597289145,0.1809068471,-0.1374412179,-0.071501337,-0.0658150464,-0.1228518635,0.363599509,0.1543242633,0.2317284793,0.2557786405,0.0614614412,-0.0047437898,-0.0602129884,0.2081991732,0.4198661447,0.1450264454,-0.2116784006,0.0292615071,-0.0277635418,0.0973519683,0.087252669,-0.087716952,-0.3895960748,0.3541232646,0.0179501884,0.1026289091,-0.0706161857,0.0873728767,-0.1946308613,-0.0451385118,0.5012277365,0.1294465214,0.1949753761,0.0250597969,-0.0093801608,0.5429433584,-0.2118059993,0.0418144502,-0.2268702239,0.0226060841,0.5083121061,0.4226641953,0.1123533398,-0.0454070233,0.4814372659,0.1322821826,0.0714679733,0.3586614132,-0.0433652252,0.0476317629,-0.1730791181,-0.0792840645,0.1631108522,-0.212156415,0.2862605155,0.2180590332,-0.0393928699,-0.3749186993,-0.0490138531,-0.2270479798,0.2377126366,-0.1845874041,-0.2329938859,0.2098865807,-0.1266938597,0.1135118082,0.3974912167,-0.1563858539,-0.2002331018,0.1650028378,0.1509024352,-0.2310722917,0.3549776971,-0.0034902964,0.1214134321,0.1821807176,0.0641114861,0.7182885408,-0.4093879461,0.121518448,-0.08843752,0.1298679858,0.2600771487,0.3952263594,-0.2697913945,-0.1796068847,0.152626574,0.0636613518,-0.1021967456,0.2406070381,-0.0837342963,0.0630977377,0.09434174,0.0667086244,-0.0216612723,0.1940475702,0.1418946534,0.2101719081,0.2374109328,0.1834969223,-0.134131968,0.135247916,-0.1609971076,0.044352524,-0.3645541668,0.1145360917,0.2537453473,-0.2655097246,0.3736873567,0.3293414414,-0.0005731989,0.195027262,0.417281121,0.3857181668,0.3806241155,-0.313904494,-0.216436401,-0.3051887453,0.369838357,-0.0038656152,0.188307628,-0.5266878605,0.066120103,0.0174712818,-0.0392613076,0.0130020911,0.1622861177,-0.1348773539,0.1616898924,-0.3015483916,-0.0942891687,-0.0909155905,0.0008063353,-0.1081473529,-0.3892262578,0.3206036389,-0.1331941932,0.041663371,0.0418236405,0.0725513026,-0.2574551404,-0.1132662594,0.4502836764,0.2056599408,0.1887113899,-0.2536135614,0.0303307604,-0.2292929888,0.0247610603,-0.3324453235,0.0260847472,-0.0189600457,0.3045559525,-0.2715017796,0.0932578519,-0.3927755058,0.1030114815,0.2600845397,0.0324523374,0.0816636011,-0.0760463476,-0.1772250831,0.0258838758,0.116041109,0.5565143824,0.0619826503,0.2580122948,-0.2433096617,-0.015569156,0.2863122225,-0.3389566243,-0.548126936,0.3631696999,0.0727461502,0.2338905782,-0.1632906944,-0.2613700926,-0.011047408,0.1338590831,-0.2170872539,-0.1264722943,0.0372462757,0.0467441529,0.1285656989,-0.0585524365,0.3239069581,0.0935545489,-0.3760496378,-0.0574709438,-0.2336916924]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"Actually I did more tests it doesn't >.<\r\nI'll let you know if I find a way to fix that","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":20,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nActually I did more tests it doesn't >.<\r\nI'll let you know if I find a way to fix that","embeddings":[-0.1001971662,0.1869404614,-0.0199802201,0.458135426,-0.0414948054,-0.0328615978,0.310751617,0.0137844495,0.0069233319,0.1962369829,0.0152610112,0.2812228799,0.0530100502,0.4453213215,-0.0918111876,0.3288263083,0.2770382762,-0.0716737285,0.0422292575,-0.0833311453,-0.3645289838,0.394906044,-0.2226699889,-0.0297363307,-0.4760729373,-0.1265688539,-0.1334368289,-0.0599123687,-0.0275350381,0.2656660974,0.2077789158,0.0172537975,0.1959412247,0.6396953464,-0.0001290814,-0.1540848464,0.213269189,-0.2452561557,-0.1593503058,-0.114685297,-0.2909755111,0.0105281863,-0.1092145965,-0.0579332449,-0.2169555575,0.1224940047,0.2166627944,-0.5325363278,-0.010078162,0.2681661546,0.0640946105,-0.47394979,-0.480668664,0.3833518326,0.0587677807,0.243763119,-0.1412620544,0.2848828435,-0.0153787006,-0.1163645834,-0.0842502192,0.2958207428,-0.0695914254,0.0128157008,0.2305926979,0.3311593533,-0.304877311,-0.3038527668,0.533562541,-0.4408625066,0.3466918766,-0.3541082442,-0.1448860615,-0.265127331,-0.3225090802,-0.3541128337,0.5375027061,0.1383071244,-0.1107280925,0.0715741888,-0.4810945094,-0.2326908559,0.0656590536,0.0965537429,-0.1904105842,0.31480214,-0.2614330351,0.0345509984,-0.0387202539,0.2343807966,0.6895991564,-0.3677733541,-0.0195311718,0.1927597374,-0.1213608384,0.0573783107,-0.2128070295,0.5516891479,-0.2121066153,-0.1210532933,-0.2416576743,-0.1189972684,-0.1252040416,0.086635828,0.0504911877,0.445263654,-0.0332958363,0.1330261528,0.0818043128,0.0395884551,-0.5268156528,-0.0845654309,-0.027151756,-0.2695402801,-0.0042475048,0.1287916005,-0.0369938202,-0.290201515,-0.0993199348,-0.0460208394,-0.3127284348,-0.013800025,0.0954971015,0.3146099746,-0.0434423462,-0.1035689488,-0.2572100461,0.2454315275,-0.1969503015,0.3418700993,-0.0803738534,0.0207407363,-0.3455398679,0.4818636477,0.0642976835,-0.2464165837,0.1298306137,0.0363991372,-0.0103645306,-0.3284159601,-0.0583500937,-0.4485407472,0.3813793063,0.1436963975,0.1453280002,0.4247841239,0.0019370598,-0.2195819765,-0.1311643571,0.2365234196,-0.6200785637,-0.2533975542,0.1688311994,-0.0321038254,-0.2495457083,-0.067345947,-0.6999961734,0.1224301383,0.4770286679,-0.205056563,-0.0080381455,-0.0043334458,-0.4935464859,-0.2447816879,0.0572936721,0.1244076341,-0.234338522,0.1792040914,-0.3634156585,0.4979801476,0.513926506,0.3871518075,0.075289458,0.2675049007,-0.4960411191,-0.041340895,0.1093340293,-0.2899128497,-0.8469179869,0.318788588,-0.0423517115,-0.0128071886,0.0056163082,-0.0255613904,0.2197386175,-0.0894946158,0.1782181114,0.1592388153,0.0948842615,0.2119773477,-0.2583329678,-0.1464414895,-0.0520190336,-0.1498499811,0.0301358029,0.2176420838,-0.0006994176,-0.3706175983,0.1509187371,-0.1175173745,0.1293628663,0.2562039196,0.2577470541,-0.1585953236,0.0000068447,0.1345375478,-0.2021965235,0.1305844635,0.0093786167,-0.1663395464,-0.5060367584,-0.0681093782,-0.1266943216,-0.164160803,-0.1428063065,-0.2228372693,-0.087469615,0.3029132783,0.1267005801,0.0550165363,-0.0214776266,0.5375711322,0.1039922982,-0.1637997031,-0.0310336314,0.2271191031,-0.1092197001,-0.2241535634,-0.1367475837,-0.1725661457,0.2570614219,-0.2123559117,-0.2030741274,0.1432766765,0.2454458475,0.4036563337,-0.0317771584,0.2653979063,0.0925490707,0.0413970873,0.1671624035,-0.0784718096,-0.0313067324,0.1556563675,-0.0009804078,0.2876313329,-0.0989504233,0.185148105,-0.0919763967,-0.2702074647,-0.0442136116,0.0820063055,0.2814543247,-0.1440525204,0.0922967345,0.3347589076,0.3947918713,0.2348679602,0.2349326611,0.3371480405,0.6447883844,0.119115226,-0.0171494558,0.1671895385,-0.1317229271,-0.2494183779,0.2548364401,0.4952813685,0.4230060875,-0.1029384285,0.2637450397,0.2036377937,0.0837236345,-0.1512748897,0.1672002077,0.0601979904,0.3180698752,0.0076511712,0.2109301984,-0.064953573,-0.0552661307,0.191541478,0.0658398122,0.2555198073,-0.1284782141,0.5481223464,-0.3072597086,-0.2564249933,-0.2413409054,-0.2760952413,0.0781546682,-0.1602677107,-0.1182289571,0.3369998336,0.0332643539,0.050681293,0.2371831685,0.1976035684,-0.1381628215,-0.4697015882,-0.1437356323,0.1527374685,-0.2454112619,-0.1133157238,0.3678135872,-0.0569793619,0.0241775718,-0.1859462559,0.0227216389,-0.3783481717,-0.0986185744,0.2697833478,-0.1195578128,0.2959055901,-0.2688421607,0.0188592281,-0.4218043387,-0.3589331508,-0.0800452754,-0.0346795283,-0.0787686631,-0.4026960135,-0.0332612507,-0.0269002784,-0.0811652839,0.0032078775,-0.1645305008,-0.0334004648,0.0611832626,0.0083255731,0.3051475286,-0.0450579301,-0.0811832994,-0.2303551286,0.3723268211,-0.3175048828,-0.3663457334,-0.622543931,0.2493044585,-0.1268671304,-0.036184106,-0.0311372411,-0.0876350924,0.1213364005,0.3739616275,-0.4566191435,-0.2531041205,-0.1389434189,-0.0384321883,0.2074797302,-0.0992779285,0.2927907407,0.0628481433,0.221025914,-0.0134768542,-0.4279940426,0.0216266587,0.3045812845,0.382042408,0.0778238848,0.2560872138,0.0397081226,0.6177136898,0.2947866321,-0.00883637,0.4568141401,-0.0352963097,0.2197225392,-0.228672266,-0.1513240635,0.2661589682,0.0503708832,-0.6271564364,0.3973252475,0.1080932617,-0.4347504377,0.1437236518,-0.2814536989,-0.3260056376,-0.2251602113,0.2219090164,-0.0011853445,0.2085199654,0.0663938001,0.1401647329,-0.2165212482,-0.24714683,0.1060164049,0.0449221842,0.2869082093,0.1469974816,0.2787696719,-0.4967870712,-0.3790457547,0.3464757502,0.3606938422,0.0563128181,-0.0647534579,0.3000008464,0.1742336154,-0.1240522116,0.5816720128,-0.3319545984,-0.1878365874,-0.1880345643,0.0243842918,0.0193963815,0.004804926,0.1159655452,0.3986047208,-0.0556098074,0.4732910097,0.3217411637,0.0159071125,-0.1249708086,0.1998544782,-0.3449236751,-0.1482804567,-0.1974350661,-0.5270221233,-0.2029923052,-0.004955105,0.1355162263,-0.0128361946,0.0748146921,-0.0141313812,0.1147589684,-0.0878750086,0.067526482,0.1530642062,0.4317626357,0.1444835812,-0.1085926294,0.093873471,0.2191275507,0.381564647,0.5146508217,0.1817273945,0.1613634378,0.1798891276,-0.1523645669,0.304582119,0.319531858,-0.1802613288,-0.2109510899,-0.0914971083,-0.0381050855,-0.3330701292,0.3399729729,0.1309188455,0.3543566763,-0.4325093627,-0.4281255603,0.4614940882,0.1713366061,-0.2966611981,0.3499470353,0.0447990969,-0.5196498036,0.4166848958,0.1613184363,1.119238019,0.0064206356,0.1492101848,0.0563516691,0.0863898695,-0.0224033836,-0.3167365193,0.2254672348,-0.1897275746,-0.6439607739,-0.0953606293,-0.1686271727,-0.2084507197,0.0929963514,-0.1833430827,0.4281603396,0.0131868944,-0.0684454739,-0.1066494957,-0.0882831439,-0.0291197728,-0.1644827574,-0.2493917346,0.0234037917,-0.0669641793,0.4125227928,-0.1377209574,0.1827721447,-0.1545264274,0.0587958954,-0.1939359158,-0.052264791,-0.3953526318,-0.0228390116,-0.3558655083,-0.2806157768,-0.1223724335,0.3432165384,0.5044690967,0.2127478868,0.010919855,0.0062493379,0.1235873252,0.2426975071,0.033219438,-0.4295626581,0.0050996407,0.0530467145,0.1005418599,-0.3044876754,0.1799134165,-0.4050152004,-0.028911829,0.0571527034,-0.1012034342,0.0512935296,-0.2298829705,0.0082561187,-0.0810013637,0.0929944739,0.0212133247,-0.0150381438,0.2631601095,0.4860735536,-0.4386587739,-0.4045324922,-0.2367364168,0.2274438292,0.3079238236,0.0721366853,0.4588257372,-0.1814826876,-0.211797595,0.0217545051,0.0081974398,-0.2481900752,0.3098764718,0.0689787939,-0.4246731997,-0.0832345635,-0.1431318372,0.0346490033,0.3159293234,0.0340315141,-0.4057687223,-0.2975547612,-0.390552938,0.0329502225,-0.053982228,0.0018422508,0.0533907339,0.1469669789,0.1759284735,-0.1808769256,-0.1588265896,0.1416149437,-0.0857368708,0.0707463846,0.0553784706,0.0067400048,0.3965462744,-0.1677688211,-0.1059673354,0.1616748422,0.063921541,-0.0879734978,-0.0432714634,0.1717988551,0.197226271,-0.0620708056,-0.0526820309,-0.2885946631,0.0479905047,-0.4657149613,0.2235802859,0.3048145473,0.1255899966,-0.1487080902,0.2037121207,-0.143911317,-0.0796404183,-0.0604648292,-0.1301433295,0.3830965459,0.1570235044,0.2263223976,0.2681724429,0.0695646629,-0.0093633467,-0.07343629,0.2125530839,0.3936017454,0.1422576457,-0.2025923729,0.0021381248,-0.0084064705,0.0643014684,0.0922193229,-0.0775453597,-0.4096875787,0.3800757229,0.0100467009,0.0969910473,-0.0744948983,0.0850033388,-0.2100767493,-0.056318637,0.5034117699,0.1240140796,0.1653403044,0.0301887952,0.0092838965,0.5287909508,-0.1984931529,0.0644593313,-0.2298065424,0.0299994331,0.5227378607,0.4158552587,0.0957127362,-0.0363947339,0.4889543951,0.1238824949,0.0548814163,0.3511891961,-0.039125029,0.048187241,-0.1636125743,-0.0937612131,0.1594876945,-0.2239948958,0.2865586281,0.231538415,-0.0386955477,-0.361273855,-0.0633785427,-0.2279272527,0.2280159444,-0.1960697919,-0.2139559388,0.211028412,-0.1182354391,0.1159966663,0.3922670782,-0.1613477319,-0.2112931013,0.1820067316,0.1608033776,-0.2490396202,0.345497638,-0.0114987856,0.1392014623,0.1760272384,0.0654562339,0.7366924882,-0.4153044224,0.1058186218,-0.1174083948,0.1387729198,0.2489478588,0.4229891896,-0.2619306147,-0.1962722093,0.1766308844,0.0808217153,-0.1100916862,0.2239323109,-0.0668054894,0.0656225979,0.0744922385,0.056740988,-0.0181809161,0.1978738755,0.1304050982,0.2323053479,0.2414747477,0.1912721545,-0.1268613487,0.1268394887,-0.15308173,0.0317869969,-0.3571462035,0.1267790347,0.2589235306,-0.2709531188,0.367713362,0.3034544885,-0.0048087393,0.1969352812,0.4202010334,0.3629051447,0.3622791767,-0.3203476369,-0.2182141542,-0.3224577308,0.3802945316,-0.0063217077,0.202030316,-0.5194978118,0.0457501076,0.0079019405,-0.0216594841,0.0384819508,0.1716012955,-0.1097048372,0.1593043655,-0.3157357574,-0.0987094939,-0.1103778109,-0.0025156776,-0.1039050668,-0.4072436392,0.3230719566,-0.1130490527,0.027075002,0.0397796407,0.0766298026,-0.2620250583,-0.1165412813,0.458986491,0.2134212554,0.1491034627,-0.2523421943,0.0602663718,-0.2260858268,0.0038760677,-0.327319324,0.0119494563,0.0091763418,0.3091532588,-0.2791394293,0.0806863755,-0.4037919343,0.0749665946,0.2575176954,0.0394628383,0.0847368985,-0.0773898736,-0.1713284701,0.0262787193,0.099920325,0.5492195487,0.0676569864,0.2419797629,-0.2468031794,-0.0222038142,0.2681179941,-0.3252448738,-0.5462810993,0.3673077822,0.0814202279,0.2363242507,-0.1652340442,-0.240495041,-0.0269384068,0.1208177507,-0.2052073628,-0.1055415049,0.0260220673,0.0342704281,0.1519045681,-0.0629816726,0.3229028583,0.1310681552,-0.3947972059,-0.0453302823,-0.2368305027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"Actually I also have the issue when writing a regular text file\r\n\r\n```python\r\nf = open(\"\/content\/drive\/My Drive\/path\/to\/file\", \"w\")\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) \/\/ 512)) # 1GiB\r\nf.close()\r\n```\r\n\r\nIs that supposed to happen ?","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":37,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nActually I also have the issue when writing a regular text file\r\n\r\n```python\r\nf = open(\"\/content\/drive\/My Drive\/path\/to\/file\", \"w\")\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) \/\/ 512)) # 1GiB\r\nf.close()\r\n```\r\n\r\nIs that supposed to happen ?","embeddings":[-0.0809623152,0.150933817,-0.004716095,0.4641732872,-0.0074457242,-0.0837897584,0.3782770634,-0.0160653237,-0.0122853341,0.1514238566,0.0223207176,0.2040352374,0.1237223744,0.459492594,-0.0932221487,0.3096233606,0.2900456488,-0.0745373294,0.0555239506,-0.1031743512,-0.3542973101,0.3774950802,-0.2202057987,-0.0087551316,-0.4959014654,-0.1943451613,-0.1198794842,0.0271139368,-0.007465004,0.2972819805,0.1181405634,0.0325630791,0.1704257727,0.6366829276,-0.0001297842,-0.1508616656,0.2038340122,-0.2284534872,-0.1701294482,-0.1342807263,-0.2650021017,0.0300459415,-0.0569734611,-0.0505863316,-0.1384722888,0.1548907012,0.2752764821,-0.536054492,0.0094062947,0.2767137289,0.0580062494,-0.4531153738,-0.4936726391,0.3889502585,0.1365727186,0.2704545856,-0.1813366562,0.2528256774,-0.0109171104,-0.1303181499,-0.0711399987,0.2785567641,-0.0806724727,0.0198998265,0.2697373927,0.4275933504,-0.3152773082,-0.2621915042,0.5415553451,-0.4134799838,0.3608884215,-0.3902038336,-0.2110204846,-0.2957029343,-0.3344982564,-0.3674358428,0.5965530872,0.1802713871,-0.1347963959,0.0444769673,-0.473382622,-0.2119874805,0.0828591883,0.1022464931,-0.1576583236,0.2304851413,-0.2503764927,0.0505878739,-0.0516202301,0.269718498,0.68134588,-0.4417134225,-0.0237598475,0.203751564,-0.0917869508,0.0940780118,-0.2643795013,0.4885321259,-0.2613624632,-0.1318342984,-0.1807596684,-0.1316577792,-0.1469574571,0.1340814084,0.1013710946,0.3952644467,-0.0123560624,0.1182581708,0.1236552745,0.0157679636,-0.5297678709,-0.1027585492,-0.0362670198,-0.2775636017,0.0065120058,0.0582706071,-0.0306735374,-0.2944576442,-0.1228455454,-0.0581407771,-0.3343105614,0.0305987541,0.0391496532,0.2700135708,-0.0227348469,-0.1497119069,-0.2075528204,0.2387869209,-0.2323745787,0.3985632062,-0.0660966337,0.0024269284,-0.3374920189,0.4626072943,0.0347473547,-0.2426494062,0.1173257008,0.0656713098,-0.0227639284,-0.3494029939,-0.0363068655,-0.4917875826,0.3772360682,0.1531202048,0.1754184663,0.4545671642,-0.016256839,-0.2907529473,-0.0823221579,0.2838455737,-0.5681078434,-0.2272555232,0.1737813503,-0.0337877572,-0.2504401207,-0.0980441272,-0.6474565864,0.0480876006,0.5046129227,-0.2406375706,0.0103453426,-0.0212473385,-0.5231416821,-0.2621290088,0.0020737362,0.1776815802,-0.2439186871,0.2162410021,-0.2958571315,0.5004327297,0.5286247134,0.4728852808,0.091679275,0.2983110249,-0.5009072423,-0.0403814986,0.1156427637,-0.2453132719,-0.7745263577,0.33667171,-0.0865652412,0.020292284,0.0238598604,0.0039782687,0.2072510272,-0.1113534346,0.2035693377,0.1161359698,0.1152517051,0.191468209,-0.2601095736,-0.0995403826,-0.0675879121,-0.1769413054,-0.0160327274,0.1607621759,0.0332443044,-0.3790466189,0.134554252,-0.1305041462,0.1587210149,0.2919359803,0.2254760414,-0.1097867042,0.0023407498,0.1582442671,-0.1636751592,0.0663116649,0.0064320127,-0.1549727917,-0.4688406885,-0.0954162627,-0.0691837147,-0.1704114974,-0.1157703474,-0.2090335041,-0.0990911499,0.3020282984,0.0527632795,0.1034016609,-0.0371871442,0.5598995686,0.0584912188,-0.1376343369,-0.0100797955,0.1996802539,-0.0974129736,-0.2990058362,-0.1839173734,-0.1400810778,0.2100965381,-0.1822098196,-0.2029643208,0.1439807266,0.2620843649,0.4544187784,-0.0678557977,0.2846869826,0.1313074827,0.1178170815,0.1736066788,-0.0596175753,0.0129096676,0.0975575,-0.0596396327,0.2911247015,-0.0939188749,0.1649128646,-0.0630543455,-0.3115476668,-0.1043291241,0.0928654373,0.2769778371,-0.0767954364,0.113763608,0.3560959697,0.3650048077,0.2202543914,0.2212857902,0.2814454138,0.6594775915,0.0796381086,0.0169117618,0.1316612363,-0.1434538066,-0.2267142385,0.252293855,0.4938217103,0.4509129226,-0.1102535054,0.2818646729,0.2308767438,0.0572349876,-0.192055136,0.2052621245,0.0635759383,0.2396275252,0.0193324275,0.1999723911,-0.0383975133,-0.1153916568,0.2090270668,0.0695498139,0.1952136755,-0.169008702,0.5489720702,-0.2330795825,-0.2392109632,-0.3021422625,-0.1950262636,0.0795688555,-0.1342739314,-0.1480572671,0.3275838196,0.0119456863,-0.0300209504,0.2582430243,0.2365769446,-0.1091117188,-0.4411351085,-0.1046401113,0.1154154837,-0.2608535588,-0.1139017045,0.3723790646,-0.0132912602,0.0179705378,-0.2058743536,0.0274478327,-0.3316384256,-0.0950437561,0.2160342634,-0.1245895699,0.3008134067,-0.2806513011,0.0052453275,-0.432366699,-0.323248148,-0.111687772,-0.0163679365,-0.1328595877,-0.4609757364,0.0380086079,-0.0903612152,-0.1600344181,-0.0169567633,-0.1087301373,-0.0706206784,0.1383949816,-0.0190440193,0.3141641319,0.0401264206,-0.1042646319,-0.2069066465,0.3973642886,-0.2842165232,-0.3432782888,-0.5802721381,0.2480850369,-0.1291804761,-0.076583758,0.0040053441,-0.0649519265,0.0509701334,0.4050847292,-0.4348990917,-0.2244686335,-0.1609023958,-0.0584499575,0.196693182,-0.0321757309,0.2875188887,0.0443622544,0.2087786347,-0.0488303602,-0.4330039024,0.1043199301,0.273560226,0.3252649903,0.041188091,0.2671677768,0.0133692743,0.5328205824,0.3042558134,0.0319025181,0.430485934,-0.0257738996,0.2479047179,-0.2646461129,-0.12208049,0.2533507049,-0.0026005267,-0.6874219775,0.3849371076,0.1331448704,-0.3752195835,0.1488452256,-0.2473292053,-0.2869769931,-0.2395713776,0.2274457216,0.0160554051,0.2355138361,0.0311372392,0.1894521564,-0.1962172389,-0.259650141,0.0710704178,0.1155289859,0.2832701206,0.0787959769,0.2494221032,-0.4221401513,-0.385180831,0.3280620575,0.4100993574,0.0784635022,-0.07906349,0.2768157721,0.2013736516,-0.1284291744,0.5718345642,-0.3225114346,-0.2128907889,-0.1575901955,0.0055418154,-0.0180004686,0.0042312806,0.0476930663,0.4170287549,-0.0004009898,0.4902386963,0.3160839379,0.0031770337,-0.1257741898,0.154505074,-0.3040198684,-0.1654364169,-0.1865906119,-0.5046249032,-0.1751075685,-0.0743584037,0.134340927,0.0022162534,-0.0032348435,-0.0233532861,0.1278809309,-0.05325187,0.094535403,0.0910330713,0.4122695923,0.1079916805,-0.1451215148,0.1175670028,0.2242972106,0.3180162311,0.5354546905,0.1898303628,0.1655128598,0.1603081524,-0.2400102466,0.2840308547,0.3559728563,-0.1273951381,-0.1553617865,-0.1032551304,-0.1012606993,-0.3344506025,0.3314042985,0.1584890187,0.3463461697,-0.3968758285,-0.4340343177,0.4349760413,0.1569945216,-0.2913616598,0.3219600916,0.0002822299,-0.5629954934,0.4082662761,0.1468815804,1.1026697159,-0.025093155,0.2338027358,0.0235595964,-0.0046498058,-0.0545314029,-0.3294460475,0.2526038885,-0.2281484306,-0.6414881349,-0.0915871635,-0.2005324215,-0.1712313145,0.0673605576,-0.1693917662,0.3570498526,0.0608160309,-0.0344433971,-0.0861869827,-0.0958181396,-0.1250860691,-0.1322516054,-0.2325712293,0.0000314218,-0.0144964429,0.4019239545,-0.1263515949,0.1470892727,-0.1871940047,0.0651531741,-0.2133183479,-0.0509143472,-0.4018420875,-0.0048369798,-0.3165283799,-0.3073404431,-0.0720216036,0.3442337215,0.5794885755,0.2448792607,0.0094383694,-0.0311703291,0.0687985942,0.2260994017,0.0237996317,-0.4319209158,-0.0044454094,0.0783524886,0.0978335813,-0.3306046128,0.1395055056,-0.3778975904,-0.0554662347,0.0662566647,-0.112194702,-0.0004935425,-0.2233911753,-0.0294295624,-0.0849098936,0.073418498,0.0106681949,-0.002215408,0.2830318213,0.4514401257,-0.4418176413,-0.4067653418,-0.23479487,0.258393079,0.2941422462,0.0805194229,0.505718112,-0.1924352497,-0.1908166558,0.0337104276,0.0553311855,-0.2769656181,0.3419545293,0.126182124,-0.4228163064,-0.1077380553,-0.1517355889,0.1283304095,0.2605071068,0.0372666419,-0.4263397753,-0.3310598135,-0.3691376448,0.0153495222,-0.051179029,0.0413538031,0.0808200762,0.1399013698,0.2029252797,-0.2439485937,-0.1427310854,0.1914767474,-0.0776702389,0.1080352888,0.1253316104,-0.0303895827,0.3905185759,-0.1557266861,-0.1167275086,0.1853381842,0.0732200295,-0.0771495625,-0.0613897331,0.1893729866,0.2049024403,-0.1101849154,-0.0476794951,-0.2540995777,0.0772399679,-0.4438061118,0.271540612,0.3231217563,0.0687290654,-0.1394196898,0.146487847,-0.0420566797,-0.1052661017,-0.0374791585,-0.1940354705,0.40677917,0.1452098042,0.1737913191,0.2661369443,0.0864082649,0.0027305395,-0.0581720918,0.2556654513,0.3912896514,0.1391523778,-0.2817211151,0.0091890134,-0.0093167415,0.0877390802,0.111807555,-0.1098465696,-0.3855486214,0.3469546735,0.0061654258,0.1675992608,-0.0860014707,0.0664690584,-0.2295808941,-0.0495780185,0.4888734519,0.1787821651,0.098316662,-0.0034018164,0.0161501206,0.4908652902,-0.1743362248,0.030837955,-0.2170121521,0.0416654497,0.4911317825,0.3654713929,0.0658307672,-0.0110052908,0.5228962898,0.177635923,0.0692608953,0.3879833817,0.014594987,0.0412263125,-0.244433865,-0.0556926653,0.1413013637,-0.2105817497,0.3062294722,0.2023257762,-0.0969418958,-0.4165747762,-0.0124940723,-0.2229884565,0.1945794821,-0.2766155899,-0.2702246606,0.223063454,-0.1256966144,0.1343888789,0.4146278501,-0.1829536855,-0.1224098876,0.1584744453,0.1388182491,-0.1914562434,0.3510169983,0.0748239756,0.1025699899,0.2082949132,0.0153051447,0.7186048031,-0.3996401131,0.1384305656,-0.1246139482,0.2011361718,0.2932036817,0.4530896842,-0.2599588633,-0.1737221628,0.1562871486,0.07342536,-0.0617403053,0.2570306063,-0.0532394275,0.1493946463,0.1064760163,0.0463066623,-0.0061288839,0.2349720746,0.1404094547,0.1613081694,0.153607741,0.1668096483,-0.1433655769,0.100396648,-0.1489121467,0.0240978263,-0.3396231234,0.0856763422,0.311399281,-0.2772238851,0.3688077629,0.2808133364,-0.0135309389,0.2127440721,0.4522306025,0.4078075886,0.3765894175,-0.3265292943,-0.1923316568,-0.251575917,0.3914064765,-0.0224455632,0.1984594017,-0.5478371978,0.042085018,-0.0205491744,-0.043378599,0.0044847215,0.1747304201,-0.1757806689,0.1482004821,-0.3036428988,-0.1035675183,-0.0763936937,-0.1082216725,-0.109682411,-0.4672412574,0.3492947221,-0.1440930814,0.0327432193,0.072681278,0.1453408003,-0.270029217,-0.1587573737,0.4588566422,0.2459577322,0.1691599786,-0.2327439934,0.005966817,-0.2120885104,-0.0001607053,-0.3116273284,0.0455616117,0.0111173149,0.3400745392,-0.2860616148,0.0865051001,-0.3210944533,0.1406545788,0.2385006547,0.0232114047,0.138678506,-0.0741586089,-0.2107339501,-0.0015974406,0.1609108001,0.5694400668,0.0424069986,0.1978941411,-0.2249976546,-0.0082821427,0.2339946032,-0.3523562849,-0.5315519571,0.3948780298,0.0739600435,0.2498631179,-0.1458754092,-0.3200004399,-0.0253370758,0.1369528919,-0.2490008771,-0.1409202218,-0.0034686869,0.0467934385,0.0967401862,-0.0225264449,0.3537873924,0.103175059,-0.4071530998,-0.0464940816,-0.2741344273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"The code you wrote should write a 1GB file in the Google Drive folder. Doesn't it? ","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":16,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nThe code you wrote should write a 1GB file in the Google Drive folder. Doesn't it? ","embeddings":[-0.0446973145,0.1795119941,-0.0474339947,0.406788528,-0.03534659,-0.0158605259,0.3132928014,0.0269284528,-0.003272105,0.2124627382,0.0527834035,0.233256489,0.0173061248,0.4683630168,-0.0888160914,0.3117387295,0.2533087134,-0.0614044629,0.0652199239,-0.0793858841,-0.3149404228,0.3846552968,-0.2003743798,-0.0362803452,-0.5003238916,-0.1209213734,-0.1952656358,-0.0283287391,-0.0948133543,0.249033466,0.169958815,0.0328491405,0.211313501,0.6244872808,-0.0001260697,-0.1791576445,0.2552550435,-0.2370764464,-0.187532559,-0.0942383483,-0.2913306952,-0.012065053,-0.1296514273,-0.0519676097,-0.1738840491,0.1368010193,0.2410814911,-0.5128407478,0.0125841079,0.2712931037,0.0960835293,-0.4741631448,-0.4019347429,0.3541018963,0.1001484245,0.2302198708,-0.0978730246,0.297321558,-0.0064145019,-0.12767151,-0.066675894,0.3040250838,-0.0224509761,-0.0017922681,0.2682514489,0.347874552,-0.3101857305,-0.3176042438,0.5124367476,-0.4732308984,0.345089972,-0.3447913527,-0.1621991247,-0.2686657012,-0.324944526,-0.3767522275,0.5275464654,0.1636802703,-0.1465069801,0.0504785217,-0.4966207445,-0.2640627921,0.0583114102,0.0859933272,-0.2157359421,0.3035001457,-0.2949536741,0.0281175449,-0.0502060093,0.2880083919,0.640224278,-0.3850523531,0.0206760466,0.1959230602,-0.0750690401,0.0538779162,-0.1898901463,0.5359105468,-0.2053767145,-0.0806944072,-0.2508230805,-0.1914845258,-0.095758073,0.0858409926,0.0577560142,0.4821523726,-0.0581905358,0.1287788302,0.055816073,0.0115779629,-0.4982734323,-0.0858849362,-0.061555557,-0.261061728,-0.0232868064,0.1011676043,-0.0176237337,-0.2813464999,-0.1566877365,-0.0024420226,-0.3255879879,-0.0047461581,0.0736411959,0.2545736134,-0.0352051742,-0.1727347225,-0.2061343938,0.2438772917,-0.2096227556,0.3257510066,-0.0973587632,0.0834292099,-0.3553804457,0.4614129066,0.0643578842,-0.2175325602,0.1618288904,-0.030333966,0.0486511551,-0.3234556317,-0.0199802183,-0.4359334409,0.3467865586,0.1586113274,0.1547554582,0.4147903025,0.0052948184,-0.3118020594,-0.1208236739,0.2784828842,-0.6148865819,-0.272470504,0.1674014777,0.004311407,-0.2542262971,-0.0784917325,-0.7056843042,0.1296862811,0.5126646161,-0.2033641785,0.0143642547,-0.029666286,-0.4853770733,-0.2789666951,0.0931211337,0.1397349536,-0.2764246762,0.2111438662,-0.3240930438,0.5097716451,0.5368673801,0.4086741507,0.0867168233,0.3118774295,-0.4806716144,-0.0023118255,0.1090578288,-0.2603101134,-0.7887850404,0.329482317,-0.0921199694,-0.0341041163,0.0108741419,-0.0524566062,0.249814868,-0.1281843632,0.2223532647,0.1184075922,0.0447631516,0.1605670303,-0.2890416384,-0.1415537,-0.0388858616,-0.1678702086,0.0084934952,0.1557980925,0.0203556363,-0.348626554,0.1816390902,-0.0968573987,0.193291679,0.2729940116,0.2430145293,-0.1512931287,-0.0257409383,0.1923045814,-0.1705409437,0.1176257133,0.0051346989,-0.1445207894,-0.4764550924,-0.1168647781,-0.1284796149,-0.1662072241,-0.1146271005,-0.2664223015,-0.047771126,0.2823030651,0.1235270649,0.0425949916,-0.0230107047,0.5242789984,0.1113477126,-0.176496461,-0.0565921441,0.2039708793,-0.1602296531,-0.2727033198,-0.1193593964,-0.1635743678,0.1950165331,-0.2037994415,-0.1764294803,0.1463667005,0.2399626672,0.3614382744,0.0290911347,0.2707649171,0.0838491172,0.0333187021,0.2098007351,-0.0554021485,-0.0060469657,0.1278685033,-0.114665553,0.2786545455,-0.0739006028,0.1376663595,-0.0880085453,-0.2931886017,-0.1379478276,0.1056081653,0.2040584236,-0.1038915738,0.1000170857,0.3206191957,0.3659143746,0.2343898863,0.2518416047,0.3290374279,0.6469297409,0.0343542583,-0.032018818,0.161945492,-0.1168718413,-0.3037322164,0.2609194219,0.4575291574,0.451004833,-0.0714942887,0.3330643177,0.17955257,0.0596499443,-0.1460272074,0.1589972377,0.0625516623,0.2426036596,-0.0450443178,0.2130969763,-0.0463405028,-0.0528769568,0.1504987925,0.0499276519,0.2370592356,-0.1375339031,0.5223639607,-0.2170304656,-0.2122853398,-0.2391374409,-0.2586258054,0.0896019265,-0.1758670658,-0.1378313452,0.2729135752,0.0500506721,-0.0241097026,0.2173446119,0.2124043554,-0.096877791,-0.4409404993,-0.0934197679,0.159826085,-0.2728134692,-0.0783884004,0.3454499841,-0.0601839311,0.0677398294,-0.1669664085,0.0556154773,-0.313252598,-0.1032665819,0.23997356,-0.1826256663,0.2825631797,-0.2589117885,0.0043534967,-0.4489739239,-0.3210093081,-0.1142052189,-0.0281106029,-0.1064752936,-0.4497766495,0.0043178764,-0.0534782149,-0.1324045062,-0.0062674847,-0.166413188,-0.0508487001,0.1070085913,0.0462805666,0.282414794,-0.0013550417,-0.0263722353,-0.1866543889,0.4291410446,-0.2935033441,-0.3800825179,-0.6091430783,0.2402899265,-0.1338403672,-0.0644774362,0.0137088262,-0.0646499917,0.1100766361,0.3217844665,-0.4183186293,-0.2801478207,-0.1830424815,-0.0565819591,0.1905384958,-0.0270395763,0.2956296802,0.0346344337,0.1806824952,-0.0232596565,-0.4232107699,-0.0011002694,0.3015029728,0.3348653018,-0.0046832152,0.2609412372,0.0532179289,0.5737909675,0.3420699239,0.0703262761,0.411179781,-0.0185390096,0.283570677,-0.2689017057,-0.0825405568,0.3358344734,-0.0045605418,-0.6965679526,0.3890519142,0.1380956769,-0.4020674825,0.1778074652,-0.2560751736,-0.2944072187,-0.1955456287,0.2278750539,0.0192515794,0.1546322256,0.0519250147,0.1103791818,-0.2017570287,-0.2173904181,0.0626599938,0.0561019145,0.2871218622,0.1901305765,0.3035385013,-0.482637614,-0.4505161047,0.3193284273,0.3583850265,0.0268737841,-0.028257383,0.2993693948,0.2221170515,-0.1431097239,0.5819005966,-0.3227649331,-0.2359462529,-0.1722202748,0.0360447802,0.018400915,0.0002365534,0.0674702376,0.4076424539,-0.0278029572,0.4173498452,0.3009566665,0.0232451186,-0.076606445,0.2138883024,-0.295899123,-0.1170052812,-0.2367077023,-0.5745474696,-0.2460456043,-0.0253193807,0.2116997838,0.0146663524,0.0607152842,-0.0048600915,0.1458210796,-0.0785794333,0.0907557309,0.1310393363,0.421870321,0.1577144563,-0.1334286332,0.0870031416,0.2162813544,0.3260956407,0.5480580926,0.1665854007,0.2173841,0.2050190419,-0.2190312892,0.327057153,0.3367375433,-0.1730904877,-0.2543376982,-0.081273824,0.0071180211,-0.3672324121,0.3636330366,0.1331376433,0.3152448237,-0.4551058114,-0.4510666132,0.440053612,0.1708369553,-0.2595471144,0.3553853929,0.0819831118,-0.5158582926,0.4137573838,0.1381800473,1.1279289722,-0.0231651459,0.2007926255,0.0686906725,0.0818093717,0.0057590418,-0.3252645731,0.2268906832,-0.149158299,-0.6130780578,-0.0723866001,-0.1585841924,-0.2013926208,0.0647542998,-0.2076790333,0.4101580977,0.0273536071,-0.0350207537,-0.0974772274,-0.0545840673,-0.0675400421,-0.1100804433,-0.157621339,0.0595994443,0.0056406953,0.4071371853,-0.1616102755,0.1993333697,-0.192860961,0.0326416381,-0.1974648237,-0.0513985418,-0.3769133985,-0.0154413665,-0.3699803948,-0.2988741398,-0.1157844439,0.3191693425,0.508995831,0.2731177211,0.0094284546,-0.0298534613,0.0898994282,0.2153796107,0.0090973796,-0.4728895128,0.012827632,0.029775301,0.1223492101,-0.2711934149,0.1495263129,-0.3859468699,-0.0631948113,0.1162364706,-0.1167571843,0.0370574445,-0.2118514925,-0.018298123,-0.1191811562,0.0690049306,0.0394492224,0.0435188934,0.2638119459,0.4571983218,-0.4170373678,-0.43055439,-0.2661255896,0.1776324809,0.3450031579,0.0716152266,0.4960001409,-0.208111316,-0.1910765171,0.0185073297,0.0000651494,-0.1955664009,0.3140620589,0.1060143933,-0.437708497,-0.0703066438,-0.1496840119,0.0698467419,0.3320937753,0.043769028,-0.3825742304,-0.2999253869,-0.4178439379,0.0382336937,-0.0361525044,0.0407303981,0.0147996768,0.136758849,0.2235816419,-0.193591252,-0.1928515583,0.1412618011,-0.0920391604,0.0669262484,0.0898918435,-0.006983683,0.4290527105,-0.1722289026,-0.0915494114,0.1663729697,0.0423534103,-0.1221255288,-0.0323462151,0.1566826999,0.2242708206,-0.0360697173,-0.0263478849,-0.2317689359,0.0401276946,-0.4575180113,0.2492562681,0.3022359014,0.1360566616,-0.2050191015,0.1941814423,-0.1127571762,-0.0335191935,-0.1197799221,-0.1145361289,0.3726640046,0.1349273622,0.2175253481,0.3037362099,0.0779422149,0.0291693565,-0.021077564,0.1902549714,0.4145394862,0.1018226892,-0.2399726361,0.0245175492,-0.0547637604,0.1027783006,0.1373428702,-0.0889267176,-0.3767094016,0.3508292139,0.0365209803,0.150793314,-0.0523608103,0.0547360703,-0.2197598666,-0.010061319,0.5260446072,0.1290273815,0.1877489835,-0.0440783016,0.020691786,0.537511766,-0.1673854738,-0.0024449239,-0.2392362356,0.0114913676,0.495944947,0.4118017256,0.0894749463,-0.0147218127,0.4912991226,0.1121211275,0.1070731133,0.3679322898,-0.0412073061,-0.012451062,-0.1857863814,-0.0598006845,0.1553999484,-0.1773815006,0.3018843234,0.2154381871,-0.0786966532,-0.3884577155,-0.0414212644,-0.2103201896,0.2080653459,-0.2081899941,-0.2179648131,0.2021914572,-0.1233753189,0.0614822805,0.4115557075,-0.1338239014,-0.1926843673,0.1497224122,0.0943595544,-0.2575376332,0.358805418,0.0322567374,0.087473534,0.1497797966,0.0274599381,0.7081152797,-0.4393172264,0.1149682999,-0.1275032461,0.1237793639,0.3178445101,0.4450235367,-0.2584784329,-0.1782467961,0.1497448832,0.0477909259,-0.1281424016,0.2253487259,-0.0409542136,0.0705203637,0.0697555766,0.0933058932,-0.0532026477,0.1945133507,0.1485283822,0.1457559913,0.2058077604,0.1708898246,-0.1727917939,0.1114588305,-0.1520684958,0.0414842479,-0.413954258,0.1071758419,0.2566117048,-0.2570850849,0.3453409672,0.2927150726,0.0105912331,0.1742859483,0.4269962907,0.4116210639,0.3664236963,-0.3534593284,-0.1867630333,-0.3090298772,0.3833360076,-0.0509146228,0.1405675411,-0.557947278,0.1094648838,-0.01199436,-0.046165809,-0.0060193953,0.1920229048,-0.1385643929,0.127083987,-0.2760286629,-0.1136429533,-0.0776509121,-0.0166891739,-0.0771117955,-0.4317476749,0.3348557949,-0.1419657469,0.0563002191,0.0217559189,0.096253112,-0.2522623241,-0.1563680172,0.4937503338,0.196736142,0.1923656464,-0.2603970766,0.000313119,-0.2698327303,0.098094888,-0.3428788185,0.0407129303,-0.0193716008,0.3388945758,-0.2815997601,0.0439135619,-0.3746964037,0.142073229,0.272471875,0.0337093659,0.1308254749,-0.0283605233,-0.2041523904,0.0594848953,0.1460769475,0.5029762387,0.0188888106,0.2355625927,-0.2690115571,-0.0395546891,0.2869609892,-0.3156882226,-0.5447868109,0.3355136812,0.0593464114,0.2711542845,-0.1600670069,-0.2820851803,0.0071517345,0.1244511232,-0.2202489525,-0.2036885619,0.0575490817,0.0194392428,0.1309789866,-0.0469449684,0.3310011327,0.0756135061,-0.3983117342,-0.0506492034,-0.2439450622]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"Yes it does, but the disk usage of google colab also increases by 1GB","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":14,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nYes it does, but the disk usage of google colab also increases by 1GB","embeddings":[-0.0579073243,0.1755220741,-0.044099763,0.4300816059,-0.0668439344,-0.0637377724,0.3105531037,-0.0151979169,0.0240935273,0.2048574388,-0.010906619,0.2981021404,0.024510745,0.4510393143,-0.0815978423,0.3266577721,0.2895388305,-0.1219265386,0.0352942944,-0.0866471082,-0.3507324755,0.4203332663,-0.233308509,-0.0814036429,-0.4888087809,-0.1078378931,-0.1276648492,-0.0955816358,-0.0324027538,0.2782038152,0.2286666036,0.0673102513,0.1566490084,0.6136825681,-0.0001276366,-0.1912479401,0.2082684338,-0.2142641097,-0.2040421814,-0.024232842,-0.2932702303,-0.028782241,-0.1060835719,-0.0617066659,-0.2049604654,0.1848263144,0.2094961256,-0.5256448388,-0.0178202521,0.2265815437,0.0681264624,-0.451680094,-0.5178921819,0.3490350544,0.0739812478,0.2234264165,-0.123977907,0.3315561116,-0.0154122682,-0.0842097625,-0.0951622948,0.3389725685,-0.0338269025,0.0157551058,0.2457561493,0.3415032327,-0.3221705854,-0.2806941569,0.5541262627,-0.4589180946,0.3567906916,-0.3349708021,-0.113792479,-0.3185136914,-0.3346970081,-0.3372330666,0.4843696952,0.1430778205,-0.1223365813,0.0643809065,-0.4892942607,-0.2339167446,0.0875386596,0.0506548956,-0.2401971817,0.2836458087,-0.2810623646,0.0217880197,-0.0594644994,0.2542212605,0.6544744968,-0.4119967222,0.0439070165,0.1637822241,-0.1068452224,0.0887926742,-0.1488082856,0.5583968163,-0.2165682763,-0.0939267427,-0.2615531683,-0.1097969636,-0.1050681993,0.110513024,0.0768997371,0.4400638938,-0.0759487897,0.125138998,0.0874458104,0.0261254553,-0.4825038016,-0.0934643969,0.0079720151,-0.2383855432,-0.0147595908,0.0811904594,-0.0277060457,-0.273077935,-0.1797699183,-0.0673600882,-0.29285267,-0.0187694412,0.0903148204,0.2778077126,-0.0463823676,-0.1041567549,-0.2465628833,0.2361903936,-0.2052471191,0.3447312117,-0.0882335231,0.010194079,-0.3250743449,0.4676035643,0.0815076083,-0.1799305379,0.1369340718,0.0422765687,0.0162927806,-0.2870463431,-0.0484832898,-0.4461915791,0.3656993508,0.1606285721,0.1375470161,0.3792952001,-0.0333984345,-0.2071691453,-0.1437879354,0.2496202439,-0.593831718,-0.254353255,0.2465388328,-0.0125353774,-0.2679803073,-0.0868449435,-0.7585638762,0.1522955298,0.4517327845,-0.202866748,-0.0214568097,-0.0434169844,-0.5144102573,-0.2331976593,0.0507898629,0.1395336837,-0.2214716971,0.1223362088,-0.3185424507,0.5079388022,0.5029506087,0.3948971033,0.0441798307,0.2825146616,-0.4761690795,-0.101576373,0.0546777844,-0.2911744118,-0.841163516,0.3037153482,-0.0567925647,0.0289074313,0.0197403021,-0.0219827071,0.1992881149,-0.1270340383,0.212266773,0.1625933647,0.0784864798,0.202511251,-0.2769399881,-0.1132743508,-0.0516907237,-0.1584375948,0.0463781953,0.1729450822,-0.0307471603,-0.3551956415,0.1693412364,-0.1109711528,0.1673916727,0.2421610653,0.2585195601,-0.1907779872,-0.0059154532,0.1661838442,-0.1667352468,0.1368945837,0.0102147181,-0.169147715,-0.5017287731,-0.1106358394,-0.1419491023,-0.1667768955,-0.1258208305,-0.2495489866,-0.0776531622,0.3122538328,0.1102217659,0.0156331938,-0.0011551894,0.5425358415,0.1114796326,-0.1763303578,-0.0159379859,0.2110683918,-0.1047189087,-0.1904285699,-0.1278766394,-0.1625626534,0.1867081672,-0.2280746251,-0.1446204334,0.0840243027,0.2548827827,0.4031203389,0.0223483369,0.257363081,0.1185857132,0.0660278946,0.1724972129,-0.0716333389,-0.0060776463,0.1346318573,-0.0680791438,0.2990705967,-0.0349998474,0.2065732628,-0.0985882655,-0.2913219929,-0.0527620614,0.06850373,0.2535994351,-0.1347607523,0.048276294,0.3635376394,0.399183929,0.2412126362,0.2315180302,0.3760573864,0.6175861955,0.0719399303,-0.0448696204,0.1943275183,-0.1331629306,-0.2966814339,0.2808429897,0.4768432975,0.4437509477,-0.084150143,0.318764329,0.1956067085,0.1062672287,-0.1657197475,0.170476377,0.0983121842,0.2657560706,-0.0049090376,0.2535420656,-0.0740612075,-0.0619144514,0.1687425673,0.0725579709,0.2071376145,-0.1305875033,0.5599091649,-0.2604081631,-0.2264876217,-0.2308239788,-0.2595694959,0.1325733215,-0.1989035606,-0.1546435207,0.3401542604,0.0226194207,0.0221345183,0.2268986851,0.2259625345,-0.1661250889,-0.4816790223,-0.117530793,0.146519348,-0.2646997571,-0.0800121501,0.3650301397,-0.0927654728,0.0887697116,-0.2135483474,0.067064248,-0.349732101,-0.11027468,0.2804393768,-0.1367677301,0.2790209651,-0.270807445,-0.0429415926,-0.4671076238,-0.2825538516,-0.1021936089,-0.0480850898,-0.0738576651,-0.4542439282,-0.0479856767,-0.0377782173,-0.0944714546,0.0057612923,-0.1711715162,-0.0020914401,0.0862128511,0.0237182584,0.3053760827,-0.0227343906,-0.0851875991,-0.2118315846,0.4396362603,-0.2831005156,-0.3987605572,-0.6017699242,0.1919758767,-0.0778582096,-0.0095466077,0.0061638849,-0.1056253463,0.1140042245,0.3935220838,-0.4325725138,-0.3030545115,-0.1556061357,0.0082541248,0.2246519178,-0.0641730204,0.2548262775,0.0814787298,0.1776910573,0.0015555398,-0.4455821216,-0.0200946257,0.2903625965,0.3632730544,0.0725536048,0.1969728321,0.0883628204,0.5838382244,0.2748143673,0.0427748375,0.4392297566,-0.0165761709,0.2697638571,-0.2366790473,-0.0926487967,0.2606828213,-0.009138193,-0.6388463378,0.3955721557,0.1468448937,-0.4218611419,0.1807880104,-0.2996293902,-0.305449754,-0.2231102437,0.2079669386,0.0375818238,0.2017766386,0.1075874269,0.1026197821,-0.2052800804,-0.2061718404,0.0772583485,0.0263891667,0.2952230573,0.1126998663,0.300424397,-0.4600166678,-0.4727796912,0.3076700866,0.3312336504,-0.0046752244,-0.037831638,0.2897976637,0.2132300138,-0.1476933509,0.5365321636,-0.3036015332,-0.1884712875,-0.1855019182,0.0013333644,-0.0077728843,0.0182861537,0.0395706259,0.3578896523,-0.0207867995,0.480646342,0.3120597899,0.0111730695,-0.1406567097,0.2696786821,-0.3342027068,-0.099679485,-0.2003918141,-0.6102374792,-0.208085686,0.0200633053,0.169949472,0.0136519084,0.1035135463,0.0281665064,0.1347721815,-0.0831257403,0.1259093136,0.1308348775,0.4536644518,0.1778806448,-0.1218984276,0.0582237504,0.1859662831,0.3675150573,0.5435396433,0.1489729285,0.2177361101,0.1850319058,-0.1524128467,0.2796268761,0.3332005441,-0.1861942858,-0.2289145142,-0.1627449244,-0.0404613912,-0.3313867748,0.3530606925,0.1173422933,0.3663960397,-0.4136309028,-0.4675089121,0.4466532171,0.2076259702,-0.338332653,0.2557046115,0.043730434,-0.5293267369,0.4215499759,0.1456007808,1.0473254919,0.0426898561,0.2036497444,0.0965250731,0.0423168018,-0.0396209247,-0.2900185287,0.2439421117,-0.1236196607,-0.6138933301,-0.0731990784,-0.1632012129,-0.2039301842,0.0696572438,-0.1578204632,0.4205283523,-0.0104257101,-0.0520723127,-0.0802365765,-0.0771027058,-0.01264193,-0.1358057857,-0.2375906706,0.0445882604,-0.0264932662,0.4088147581,-0.148512274,0.198130548,-0.1909011453,0.0557517186,-0.1523527056,-0.0380493738,-0.3426911831,-0.0042580678,-0.377460897,-0.2897384465,-0.1256650388,0.3257494867,0.474547267,0.1950808167,0.058074493,0.0167100187,0.0692284554,0.3094980717,0.0531396046,-0.4401885867,-0.0238161497,0.0324506909,0.1295283586,-0.3118658662,0.1636593938,-0.3916964233,-0.0708140507,0.1401041001,-0.1520666182,0.0486260615,-0.2261635959,0.0217770804,-0.1023206711,0.0963026509,0.0293044392,-0.0283266604,0.2613084018,0.5555589199,-0.4439566731,-0.4408777058,-0.2350784093,0.220039472,0.4006043375,0.1288367063,0.4692187607,-0.2755566835,-0.1959399879,0.0241919272,-0.0225867238,-0.226487264,0.3542645872,0.1070017591,-0.3871389627,-0.075612925,-0.152565375,0.001446342,0.3075831234,0.0256323256,-0.42087394,-0.2553508282,-0.3862680197,0.0234519187,0.0055377265,0.0077259559,0.0538008735,0.1463302821,0.1664783508,-0.1886976808,-0.1730728894,0.1588777006,-0.0415762626,0.0751744583,0.0663030446,-0.0095909405,0.4403245747,-0.1853152364,-0.1019598544,0.1266384274,0.0790098235,-0.112429969,-0.0168648567,0.1620710343,0.1791749001,-0.0589378439,-0.0627319962,-0.2030644715,0.0900871679,-0.4784107208,0.285243392,0.320758462,0.1350723654,-0.1904230565,0.1463860869,-0.1307566315,-0.0276272763,-0.0870440453,-0.1117655039,0.3455938995,0.1495471895,0.2315557152,0.3076937795,0.0697843954,0.0247148927,-0.0569558144,0.2307574302,0.366196245,0.098943904,-0.1866419166,0.0089026149,-0.0011618992,0.0620395541,0.116598703,-0.0455860272,-0.4300135672,0.3299910128,0.0201106053,0.1107413396,-0.0755204707,0.1027726009,-0.2004493624,-0.036615707,0.4840990901,0.189201802,0.1737752259,0.0230729207,-0.0117563745,0.5495623946,-0.1672741473,0.0605415814,-0.2448732853,0.0349009484,0.4988068044,0.4440789819,0.1197556183,-0.0108455708,0.4655616879,0.0947520509,0.0923563465,0.3152707815,-0.0526203178,0.0130182831,-0.1780070812,-0.0699592084,0.2038135231,-0.2209984958,0.276899308,0.2397387922,-0.0437286124,-0.3805903494,-0.0681986511,-0.211634472,0.2561579347,-0.2080422938,-0.2677923143,0.2450731844,-0.1310350001,0.1154426113,0.3954636753,-0.152320385,-0.2363842875,0.1956712753,0.1105599031,-0.2628990412,0.3691568375,-0.0119679971,0.1278661042,0.1712807864,0.0565041155,0.6875359416,-0.429317683,0.1306069493,-0.1128613502,0.1138771474,0.2791368663,0.3800466657,-0.2391382456,-0.2316983789,0.1513356119,0.0737593994,-0.1730723083,0.2255158126,-0.1018444598,0.0541970991,0.064258188,0.0636381954,-0.0352028646,0.2013346106,0.1513822228,0.2036570311,0.2066788226,0.1617737263,-0.1621638834,0.1142380312,-0.1058931947,0.0295409225,-0.349802494,0.0642443374,0.256606698,-0.3019021153,0.3766327798,0.3106136322,0.0010767694,0.1744494587,0.4297995269,0.3183529973,0.3082840145,-0.2989073694,-0.1687381119,-0.3139298558,0.3438679576,-0.0555843487,0.1898053885,-0.5433957577,0.0775294453,-0.0162996929,-0.0358938947,-0.0433548093,0.1638634503,-0.0986057147,0.0433638319,-0.2918907106,-0.0786823705,-0.1114868745,-0.0289388113,-0.0642483458,-0.4254477322,0.2891938686,-0.1395121366,0.0489671715,0.0474089533,0.0899207592,-0.2535679042,-0.1088721082,0.4920366704,0.2435350716,0.1973460913,-0.2810575962,0.016813945,-0.210606724,0.0240417644,-0.350753516,-0.0038205348,-0.01627234,0.2768189907,-0.2446689308,0.1167316735,-0.3384176195,0.167252332,0.2957621217,0.020621052,0.1403200477,-0.063522324,-0.2019758224,0.0330910496,0.0235777181,0.549865365,0.05847634,0.2296753824,-0.2739355862,-0.0549951009,0.2869323194,-0.3075362444,-0.5420182943,0.3407602012,0.083243683,0.2365081608,-0.1823928356,-0.2051012516,-0.0077944994,0.1253355145,-0.2199570984,-0.1504567713,0.0765758231,0.0095224027,0.1259295046,-0.0499590412,0.3359440565,0.0775671676,-0.4369404018,-0.0845040008,-0.238574639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"I could check it and as you say as I write to te Drive disk the colab disk also increases...","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":20,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nI could check it and as you say as I write to te Drive disk the colab disk also increases...","embeddings":[-0.0909002125,0.0898183137,-0.0347917937,0.4908721149,-0.0633708313,-0.0270626657,0.2676787972,0.0049588634,0.0242158957,0.227923736,-0.0148068471,0.2577182651,0.0650781915,0.4686147571,-0.0855880305,0.3258917928,0.2614867985,-0.0908618197,0.0236124657,-0.0889639705,-0.3442029357,0.3814185262,-0.1878898144,-0.0883715078,-0.4606517553,-0.1489057094,-0.1486406177,-0.0555905029,-0.0616812035,0.2551704347,0.1518823057,0.0476347059,0.2141139209,0.658629775,-0.0001266875,-0.2107377946,0.185264498,-0.2339996248,-0.1972978413,-0.0303513687,-0.2913546264,-0.0428781323,-0.0804462582,-0.093984209,-0.1805127263,0.1525765955,0.1875965446,-0.4624527693,0.0369793512,0.2215129435,0.0815562904,-0.4584001601,-0.4723300636,0.3777891099,0.0973054171,0.2221828401,-0.1482014358,0.3222098351,-0.0102290772,-0.0661063343,-0.1228185073,0.2910982072,-0.0457291193,0.0100274635,0.2436986715,0.3522854447,-0.2692084014,-0.338673532,0.6088119149,-0.4549605548,0.3356356919,-0.3624158204,-0.1185456812,-0.2765686512,-0.2856233716,-0.3757875264,0.5314132571,0.1389435977,-0.147815153,0.0686528757,-0.4924881756,-0.2548331022,0.0755867362,0.0571191795,-0.2038838714,0.2674259841,-0.2745117843,0.0091094803,-0.0179538541,0.2676405907,0.7121443152,-0.3867537677,-0.0097836275,0.2255641222,-0.1349747628,0.043014735,-0.176111728,0.5634051561,-0.2399387956,-0.1454800367,-0.2318115383,-0.1449790001,-0.1336895227,0.1113656312,0.0579791106,0.4367699325,-0.0656073838,0.1153593883,0.073388584,0.0440182611,-0.4660323262,-0.0788127035,-0.0530332364,-0.2441565245,0.0195791069,0.0858408213,-0.0243465174,-0.2801483572,-0.1593984067,-0.0670680925,-0.2925985157,0.0194888283,0.0709376782,0.2782577276,-0.0084352372,-0.116373077,-0.2690790892,0.2244533896,-0.1827167273,0.3975844681,-0.0862434953,0.0280139372,-0.3113510311,0.4865098298,0.0744594336,-0.2300519347,0.1022623703,0.0751506388,-0.0056708157,-0.332251966,-0.1181941479,-0.4248935282,0.3353984654,0.1824312806,0.1418440938,0.3487936258,-0.0752956569,-0.2226856798,-0.1101973206,0.2607309222,-0.6085553169,-0.2279908061,0.1420246065,-0.0007225169,-0.2446970195,-0.0833464563,-0.7278817296,0.0973593891,0.452542454,-0.1691073924,0.0063451272,-0.0134537118,-0.4763551354,-0.2047984153,0.0397766382,0.1366304606,-0.20871526,0.1826031953,-0.315536648,0.532546401,0.4800925553,0.4041325748,0.082662724,0.2509357333,-0.5047847629,-0.1302544922,0.068114236,-0.2428762466,-0.8206087947,0.2948431373,-0.0977444127,0.0086052353,0.0828308985,-0.0241562799,0.2168331146,-0.0997235477,0.1954146773,0.1250831038,0.073805742,0.2005304247,-0.3050268292,-0.1327726692,-0.1084555015,-0.1719167829,-0.002899226,0.1939235181,0.0233739186,-0.360453397,0.1965491772,-0.107761234,0.1663274169,0.2490561455,0.2758973837,-0.2232208848,-0.0158251952,0.1475936174,-0.1878914386,0.0984159708,0.030998541,-0.1838445216,-0.4914022386,-0.1147519946,-0.1692594439,-0.1446654797,-0.1153861657,-0.2338029295,-0.0712446496,0.3111445904,0.097951889,0.003481393,-0.0143393446,0.5263148546,0.0860157162,-0.1585880965,-0.0184127595,0.2007098794,-0.0831227675,-0.2301699519,-0.1330531687,-0.1328961551,0.1908378154,-0.2405849844,-0.1907077432,0.0976371467,0.265999496,0.4339229763,0.0383322537,0.2717318237,0.1190792099,0.0998934954,0.2144974321,-0.0961880237,-0.0499377586,0.1104994267,-0.0246756133,0.2426779568,-0.0756841451,0.1617136002,-0.103887789,-0.3116764426,-0.0743965879,0.0543084219,0.315312922,-0.1267091185,0.0170020703,0.3911203444,0.4046420455,0.2553063929,0.2483256757,0.3756074607,0.6352421641,0.0773170888,-0.0819763839,0.2210304886,-0.1640895009,-0.3080351055,0.2351090163,0.5566420555,0.419631511,-0.0789889097,0.3083798885,0.224425599,0.0988956094,-0.1601508856,0.1938725412,0.05116418,0.2598308027,0.0632752404,0.2444454283,-0.0309985504,-0.0644106045,0.1818398237,0.0606858693,0.2382087857,-0.1505656689,0.5132129192,-0.2193128765,-0.2198534012,-0.2537248731,-0.2095603496,0.1171369106,-0.1727666706,-0.1311654598,0.3562390506,0.0576346107,-0.0108536836,0.2667405605,0.2275639474,-0.1232322454,-0.4242453277,-0.1358309388,0.1687889248,-0.2426113337,-0.0781314746,0.3596924543,-0.0710113049,0.086566709,-0.1987028569,0.0729769692,-0.3626462221,-0.1350426078,0.2516106367,-0.1367730647,0.3073520064,-0.2948588133,0.006540122,-0.4329467118,-0.3261017501,-0.1517671794,-0.0480665863,-0.1187816486,-0.4421813488,-0.0419284552,-0.0632458776,-0.1138281301,0.0201725028,-0.1713443249,-0.0622775033,0.0999836847,0.0376602933,0.3234501183,0.0060572359,-0.0844256729,-0.1918919086,0.4517862797,-0.2782467008,-0.3682191968,-0.6094025373,0.2131614536,-0.0835438222,0.0003339312,0.0212867539,-0.1330575198,0.1271721274,0.3549331725,-0.4617934227,-0.3288876116,-0.1489284188,0.0233004652,0.190487355,-0.0547214076,0.2663661838,0.0645634159,0.1716391146,0.0243139062,-0.4264402092,-0.0048502618,0.3264302015,0.393153429,0.0035664367,0.2174666524,0.0104868496,0.5704649687,0.2698986828,-0.0205930695,0.4525624514,-0.0104251318,0.2274507731,-0.2605685294,-0.1322252005,0.227698043,0.020328328,-0.6987563372,0.4155741036,0.0959088281,-0.4083205163,0.1828237176,-0.3095015585,-0.2439012676,-0.2038915753,0.2121493816,0.0300389715,0.2055129111,0.0212329328,0.1008075029,-0.1913893074,-0.2594244778,0.055140622,0.0401586704,0.2412414253,0.1229618639,0.3073009849,-0.4604547322,-0.4348344803,0.3341538906,0.3555985391,-0.0023162856,-0.0438427292,0.2870714068,0.245826453,-0.1482657045,0.5649098754,-0.3025717437,-0.1659479141,-0.1720978618,0.0289180186,0.0175499469,0.0298031885,0.0448257104,0.3730847538,0.0043943361,0.5102786422,0.3682122231,0.0015424712,-0.0915673077,0.2732630074,-0.332229048,-0.1729174256,-0.1844392866,-0.5728099346,-0.176168859,0.0130243106,0.1412500292,0.0300491694,0.0221447404,-0.0319683999,0.1181981191,-0.1086964533,0.0976105854,0.1747451723,0.442756325,0.2253116965,-0.0914560407,0.0405405425,0.228124693,0.3766056597,0.5109022856,0.1520317793,0.2301543355,0.1836990267,-0.2098180503,0.2359918058,0.3207588792,-0.1442794204,-0.1745629609,-0.1821292341,-0.0112457052,-0.3184019625,0.3036257029,0.1171122193,0.3674888313,-0.4066756964,-0.4075476825,0.4499432147,0.134549886,-0.321631968,0.2389633805,-0.0022428238,-0.5306398869,0.4041656256,0.1716742665,1.1117690802,0.0401867367,0.1765854955,0.0392428115,0.0319475867,-0.0652521178,-0.3078401387,0.1893234402,-0.1908807307,-0.6514492631,-0.0732112974,-0.1613129973,-0.2037958205,0.0769555345,-0.2160165012,0.3886811137,-0.0033976953,-0.0716832355,-0.1122514084,-0.0528870858,-0.0460921638,-0.1305936277,-0.2803539038,0.037753541,-0.0601431243,0.3965791166,-0.1411873698,0.1791497171,-0.1214805096,0.0727491081,-0.1955318749,-0.0794792995,-0.3821721971,-0.0182949267,-0.367030561,-0.3077130318,-0.1050221175,0.3306053877,0.5086337328,0.2187628895,0.030099554,-0.0138404872,0.0932259187,0.2404084951,0.0285436418,-0.419696629,-0.0287540387,0.0454719029,0.0815917552,-0.2918027639,0.1761635542,-0.3630618751,-0.0291384216,0.098105073,-0.1210697815,0.0322280154,-0.240129292,0.0087007005,-0.1003763005,0.1044993624,0.0404587947,-0.050364539,0.2496072948,0.5033813119,-0.4237061441,-0.4663747251,-0.2362973392,0.2391015738,0.3680252135,0.101144664,0.4734413922,-0.2448921651,-0.181454584,0.0002700577,0.0250362493,-0.2598956525,0.342908144,0.0776615664,-0.3460297883,-0.087356858,-0.1469554156,0.0331371762,0.3445212245,0.0494187176,-0.4454934001,-0.2655265927,-0.4145762622,-0.000524574,-0.0301435925,0.0728258416,0.0513022356,0.1241998225,0.1792030334,-0.1231451184,-0.184248358,0.1778455824,-0.0687739179,0.1079052836,0.1092496663,-0.0445649438,0.4041939378,-0.1749344468,-0.0894473642,0.1353371143,0.0723468661,-0.118826896,-0.0061147586,0.1551075429,0.1744977236,-0.0285286214,-0.1091034412,-0.2261294723,0.083310917,-0.4550179541,0.3012824059,0.2682916224,0.1576905847,-0.1673208624,0.1907922328,-0.1116960347,-0.0656144097,-0.0736653134,-0.1129292548,0.3603917956,0.1441701949,0.2214203328,0.2751781344,0.0632538721,0.0111396099,-0.0444081575,0.2156650424,0.3725348711,0.0908573121,-0.1623178422,0.0083142398,0.0030542675,0.0255278703,0.1160376072,-0.0616766512,-0.4699501097,0.3420881331,0.0489591397,0.1168971658,-0.0590605289,0.1115255952,-0.2292788476,-0.0432575867,0.516887784,0.1539980024,0.1316804439,0.0440823399,0.0327839069,0.5230233669,-0.1982861161,0.0246882029,-0.2192222774,0.018057147,0.48650226,0.4259889722,0.0942511186,-0.0261397194,0.4793216884,0.0896813348,0.0908982381,0.3561840653,-0.0500971675,0.0549432002,-0.1865668446,-0.0388187617,0.1921585053,-0.2290362865,0.2488803715,0.2621865571,-0.0514820218,-0.3562253416,-0.0759625509,-0.2727500498,0.2419852614,-0.2199640721,-0.2384081483,0.2570852637,-0.1364896148,0.1088026166,0.4127176106,-0.1660005152,-0.1683591604,0.196452871,0.1456786394,-0.2558956742,0.3493218422,0.0103301266,0.1349759698,0.1556816995,0.0422728471,0.7475678325,-0.4142771959,0.140889734,-0.1750688851,0.1564682722,0.2523055077,0.4177719355,-0.2435997277,-0.2142863423,0.2119776309,0.0734439269,-0.131004408,0.1941266209,-0.0987379551,0.0808150396,0.0563206114,0.0752585456,-0.0412349142,0.2130665928,0.1265200078,0.1920007467,0.2290849984,0.1382924914,-0.1069689617,0.0859759822,-0.1110323146,0.013911115,-0.3589212,0.0974492878,0.2541071773,-0.3553035855,0.3786211312,0.2719302773,0.0068043564,0.1342209876,0.4763183594,0.3950312436,0.302531004,-0.2944328487,-0.1898594499,-0.2639845908,0.3594502509,-0.0134936757,0.1397278756,-0.5812434554,0.0616459586,-0.0707525462,-0.0522551276,0.003525943,0.1889203936,-0.1334938258,0.1061547846,-0.3326286077,-0.0784465894,-0.0884293392,-0.0156106176,-0.0666600466,-0.4335342646,0.3351227641,-0.1234756857,0.048489619,0.0505676642,0.16132918,-0.231479466,-0.1227805763,0.4660206735,0.2035220861,0.1742148101,-0.2906888425,0.0410778485,-0.196138069,-0.0451880619,-0.3456781507,0.0302924067,0.0224216562,0.3042700291,-0.19644849,0.054057993,-0.3024550676,0.1461640894,0.2656728327,0.0278430544,0.1670297682,-0.0894475952,-0.1717293113,0.0250690449,0.0971106812,0.5377408266,0.0824893862,0.2462516278,-0.1996489167,-0.0883886591,0.2424054444,-0.2692519426,-0.5248529911,0.3899616599,0.0918866768,0.2048326284,-0.2126502395,-0.2469800413,0.0134123154,0.1386600584,-0.2424541861,-0.1393662542,0.1419805437,0.0024185416,0.1437824965,-0.0412591472,0.3298479617,0.0980605334,-0.4087776244,-0.0612628125,-0.256520927]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/643","title":"Caching processed dataset at wrong folder","comments":"To reproduce it: \r\n```bash\r\n!df -h | grep sda1\r\n```\r\n```python\r\nf = open(\"\/content\/drive\/My Drive\/test_to_remove.txt\", \"w\")\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) \/\/ 512)) # 1GiB\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) \/\/ 512)) # 1GiB\r\nf.close()\r\n```\r\n```bash\r\n!ls -lh \/content\/drive\/My\\ Drive\/test_to_remove.txt\r\n\r\n!df -h | grep sda1\r\n\r\n!rm -rf \/content\/drive\/My\\ Drive\/test_to_remove.txt\r\n\r\n```\r\n[Colab](https:\/\/colab.research.google.com\/drive\/1D0UiweCYQwwWZ65EEhuqqbaDDbhJYXfm?usp=sharing)\r\n\r\n\r\n","body":"Hi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```","comment_length":56,"text":"Caching processed dataset at wrong folder\nHi guys, I run this on my Colab (PRO):\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='\/content\/corpus.txt', cache_dir='\/content\/drive\/My Drive', split='train')\r\n\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\ndataset = dataset.map(encode, batched=True)\r\n```\r\nThe file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it.\r\nThe dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs.\r\n\r\nWhat gets me crazy, it prints it is processing\/encoding the dataset in the right folder:\r\n```\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nCaching processed dataset at \/content\/drive\/My Drive\/text\/default-ad3e69d6242ee916\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/cache-b16341780a59747d.arrow\r\n```\nTo reproduce it: \r\n```bash\r\n!df -h | grep sda1\r\n```\r\n```python\r\nf = open(\"\/content\/drive\/My Drive\/test_to_remove.txt\", \"w\")\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) \/\/ 512)) # 1GiB\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) \/\/ 512)) # 1GiB\r\nf.close()\r\n```\r\n```bash\r\n!ls -lh \/content\/drive\/My\\ Drive\/test_to_remove.txt\r\n\r\n!df -h | grep sda1\r\n\r\n!rm -rf \/content\/drive\/My\\ Drive\/test_to_remove.txt\r\n\r\n```\r\n[Colab](https:\/\/colab.research.google.com\/drive\/1D0UiweCYQwwWZ65EEhuqqbaDDbhJYXfm?usp=sharing)\r\n\r\n\r\n","embeddings":[-0.0841225088,0.1945452243,-0.016326312,0.4593132436,-0.0439055972,-0.0728673413,0.3859433532,0.030553896,0.0337211378,0.2072662413,-0.0354525633,0.2843731046,0.047132723,0.4058027267,-0.1138568521,0.3554214239,0.2777641714,-0.125281468,-0.0011570895,-0.1184547693,-0.3421107829,0.381975174,-0.2049852908,-0.0256493818,-0.4567474127,-0.0956494436,-0.1473441124,-0.0011716173,-0.0159747936,0.2834328711,0.1546794027,0.034230385,0.2064488083,0.6668285131,-0.0001267947,-0.1738701314,0.2114759386,-0.2006000876,-0.2727822661,-0.1019956917,-0.3044102192,0.0360213965,-0.1111781448,-0.0811382532,-0.1309183091,0.1151060164,0.1956802756,-0.5906016827,-0.0217093918,0.2646551728,0.0812927112,-0.449357599,-0.5469542742,0.3519179225,0.1121328771,0.2696870863,-0.12555632,0.2454707474,-0.0296040047,-0.1314675957,-0.0899721906,0.3205049932,-0.1109801382,-0.0286945589,0.2242933512,0.3244673014,-0.3142950833,-0.2437792122,0.5524448752,-0.4051456451,0.3221742511,-0.3460015655,-0.1634961665,-0.3033885658,-0.3498371542,-0.3369373977,0.5142972469,0.1225768402,-0.0838442072,0.0870277807,-0.5243647099,-0.1893934309,0.0964756086,0.0881543607,-0.2207656205,0.2885017395,-0.2636702061,0.0324776769,-0.0383785665,0.2351333648,0.6513156295,-0.4771052599,0.0129428105,0.2287504226,-0.119115591,0.1084964201,-0.1420404315,0.4823613763,-0.2183911353,-0.1262344122,-0.2273806334,-0.1387718171,-0.1228205562,0.0968808532,0.0549007989,0.4719401598,-0.0667963997,0.1334535331,0.0674367249,0.0957627892,-0.495754689,-0.0539302304,0.0051828241,-0.2332229614,-0.0355139002,0.1311545521,-0.0165151991,-0.2762652636,-0.2102953345,0.0187313482,-0.370452255,-0.0185113885,0.0934006423,0.2619767785,-0.0247534588,-0.1263897717,-0.2220385373,0.2480699867,-0.2516383529,0.4041092098,-0.1224166825,0.0474488661,-0.3677769005,0.422128737,0.0711667091,-0.2111281604,0.1136209071,0.0618330427,-0.0047403676,-0.3111233115,-0.0762209743,-0.4326264262,0.4141172469,0.1112445965,0.1141543314,0.4161489904,0.0264522098,-0.267082572,-0.0823545903,0.2368801385,-0.5531480312,-0.2045004368,0.1599131525,0.0041203895,-0.2979134917,-0.0669916645,-0.69560498,0.1260885149,0.4034116864,-0.1562398821,0.0185713284,-0.0514682606,-0.4866917431,-0.2335226983,0.0505956858,0.1709021777,-0.2770163417,0.1444514245,-0.3489136398,0.5182402134,0.545022428,0.4078437984,0.0712292492,0.2802247107,-0.5115201473,-0.0195448175,0.0752299428,-0.257096976,-0.7950388193,0.3377696276,-0.0835522711,0.0778091103,-0.0157460086,-0.0586844683,0.1973324269,-0.1049470231,0.1535241008,0.1360723525,0.1003889292,0.1800670177,-0.276997,-0.1262809485,-0.0409163721,-0.173557803,0.0368149169,0.1725338548,0.0476657264,-0.3246938884,0.1804226339,-0.1166001111,0.1493276656,0.2664092481,0.2396847457,-0.1233286113,-0.0224798303,0.1364758909,-0.2220253795,0.15326491,0.0340711325,-0.1538006663,-0.5125191212,-0.1051220745,-0.1252607852,-0.1472769678,-0.1138537824,-0.2297271639,-0.0633671805,0.3790797293,0.0941518098,0.0542890616,-0.0053921985,0.5049632788,0.0877435058,-0.1718437672,-0.0558091924,0.1888891757,-0.1316520125,-0.2515571117,-0.1695169955,-0.1328744888,0.2208893299,-0.2341687232,-0.1833904386,0.1453886926,0.2613753974,0.423281163,-0.0350139476,0.2826690972,0.0854922608,0.1074394956,0.179387182,-0.0734495372,-0.0230706017,0.1655795872,-0.1058244705,0.2817675471,-0.11350099,0.1841135472,-0.1023149267,-0.3243633211,-0.0864485204,0.0631371513,0.2896333039,-0.1316164136,0.0799007788,0.3997071981,0.3467099965,0.2356557399,0.2422431111,0.3499772251,0.6368630528,0.0411725007,-0.0480863377,0.1404183656,-0.0809065253,-0.282887429,0.275200814,0.5394611359,0.4638066292,-0.0988675952,0.3149523735,0.1691281497,0.0862854645,-0.1600850523,0.2028650641,0.0676125661,0.2505049706,0.0337446369,0.2440981418,-0.0421402641,-0.0904779509,0.1334930211,0.04070637,0.223536551,-0.1144714281,0.5990411043,-0.3186650574,-0.2106164545,-0.2102443427,-0.2406889647,0.1269744188,-0.2022586316,-0.1322252899,0.3160975277,-0.0469736718,0.0264324937,0.2152574062,0.193661496,-0.1426065564,-0.4581068158,-0.1588827223,0.1610168815,-0.2578992248,-0.0891753063,0.3661741614,-0.0499390773,0.0422776751,-0.1529658139,0.0076157907,-0.3925890923,-0.1384721994,0.2612594664,-0.1823771447,0.3225194514,-0.2217480987,-0.0241664164,-0.439527452,-0.3129885793,-0.0940911397,-0.0256117694,-0.1019076034,-0.4108661711,-0.0089527834,-0.0493570417,-0.1378780156,0.0002188164,-0.0952606946,-0.0734974742,0.0395640321,-0.014223271,0.2749411762,0.0180198904,-0.0792432204,-0.1839408427,0.4457667768,-0.2575637102,-0.3838847578,-0.5558648109,0.1845136434,-0.0844427496,-0.019376928,-0.0445440747,-0.0801561922,0.1160566211,0.4078014493,-0.4832977653,-0.2445380539,-0.1550945789,-0.0202741269,0.1770356596,-0.0848400071,0.3010421097,0.0698694736,0.1874433309,-0.0501757152,-0.4178805053,0.034913525,0.2792589366,0.3262370825,-0.0075919875,0.2600141764,0.0915996209,0.5603269339,0.2719271779,-0.0357675925,0.4003481269,-0.0069067357,0.2754007578,-0.2521513402,-0.1157734394,0.2647081316,0.0552051701,-0.701398015,0.4209537804,0.0577467009,-0.4382657409,0.1772968471,-0.2116811424,-0.3041628003,-0.2247139215,0.2229962945,0.0012426039,0.1784544289,0.0605887473,0.1398739964,-0.1659089029,-0.2183031142,0.0834837556,0.0565795973,0.3373071551,0.1339708865,0.262817055,-0.4482693672,-0.3884060383,0.3494025469,0.3611344993,0.0470245965,-0.0848465785,0.2404615134,0.1869641244,-0.1444204897,0.5466490388,-0.3347469568,-0.1636122018,-0.129971087,0.0169887301,-0.0141098192,0.0017947638,0.0124021098,0.3665879369,-0.0325376056,0.5276474357,0.2805065811,-0.0422104336,-0.1070474908,0.2349510193,-0.284683913,-0.1501442492,-0.1825914979,-0.5665206313,-0.2081159949,-0.0184086487,0.1475915909,0.0265532583,0.0363263674,-0.0005759763,0.128592819,-0.0907087103,0.1047620028,0.1362764686,0.394087702,0.1735524535,-0.1035952419,0.0645233244,0.2168632001,0.3731351197,0.5316536427,0.1416012943,0.1814557165,0.214033857,-0.1720275283,0.2364907861,0.3640687168,-0.1516596675,-0.2099580914,-0.1321107,-0.0220669061,-0.3369803727,0.3455791473,0.1578528732,0.3325903416,-0.3623315692,-0.4115195572,0.4117583036,0.1483868957,-0.3303751647,0.3158074915,0.0331263468,-0.5511074066,0.4613457918,0.1280946881,1.1126844883,-0.002038056,0.1784241349,0.1135697588,0.0957011282,-0.0194324423,-0.2755327523,0.2005995065,-0.1734856069,-0.5922012925,-0.0799103603,-0.1618945748,-0.2040151656,0.1059159264,-0.1329108477,0.4099910557,-0.0166436974,-0.0567093268,-0.0946829244,-0.0610731654,-0.0468452759,-0.0961307511,-0.2189695537,0.0503207631,-0.0369420089,0.3914657831,-0.149508059,0.216874525,-0.1971985102,0.0422704741,-0.1725222766,-0.0299818218,-0.3739893734,0.0167011749,-0.3672956824,-0.3203129768,-0.1461954117,0.3489964008,0.4700143933,0.2422034144,0.0681796148,0.0135831824,0.045553945,0.2881101072,0.0449518263,-0.47294119,0.0219208039,0.0892639533,0.0966314375,-0.3060150743,0.1714949906,-0.3764615953,-0.0309768301,0.0909719616,-0.1157447919,0.0323731005,-0.2353328019,-0.0012709469,-0.1241158471,0.0911293179,0.0381635539,-0.024962455,0.2713142633,0.4538406134,-0.4913174808,-0.4338661432,-0.2429767102,0.2196909487,0.276155293,0.0923028365,0.471473217,-0.196315676,-0.2017212659,-0.0035782973,-0.0369497798,-0.2587949932,0.3299712837,0.1093433425,-0.468128264,-0.1054975986,-0.126248613,0.0081286645,0.281358242,0.0266783368,-0.4296484292,-0.3005276918,-0.3171999156,0.0161023438,-0.0597175807,0.0772564262,0.0695804507,0.1193826124,0.0892352536,-0.1950303912,-0.1829491258,0.2048536241,-0.0947705954,0.1227939799,0.0786335096,-0.017576322,0.4281571507,-0.193968609,-0.0813812017,0.143147558,0.0738637224,-0.1201180816,-0.0492002741,0.1548921913,0.1974501163,-0.0815239325,-0.0458107144,-0.2472177744,0.0709017739,-0.4901148081,0.2950856984,0.334808439,0.1295578629,-0.1613297611,0.2049960643,-0.1187464148,-0.0410110392,-0.0701126456,-0.1410942525,0.3600339293,0.1319341362,0.2307246178,0.3176661134,0.0463447273,-0.0278106239,-0.0652954355,0.2071718872,0.4547816217,0.1323383749,-0.1851794124,-0.0328354239,-0.0284981634,0.1077326685,0.0593158342,-0.0855814293,-0.4570811391,0.3199622631,0.0424062461,0.1196211427,-0.0851221234,0.0694155619,-0.203630656,-0.0116877025,0.5102900267,0.1288096309,0.1602648199,0.0473561883,-0.0059922612,0.5263209939,-0.1773720533,0.0566704869,-0.2004928589,-0.0103064328,0.507406652,0.3961098492,0.0522768609,-0.0095699104,0.4942304492,0.1224838868,0.0646164119,0.3955700099,-0.0130432015,0.0467665419,-0.1954602897,-0.0617316365,0.1664903313,-0.2371868938,0.283293128,0.187167421,-0.0691949576,-0.3907464147,-0.017704051,-0.1824212521,0.2373395413,-0.2330201715,-0.2296693325,0.1860862821,-0.1380093694,0.0792410672,0.3893232346,-0.1683446914,-0.1806377023,0.155827716,0.1771237999,-0.2397205383,0.3626711667,-0.0289096348,0.1323879659,0.1518109441,0.0500288904,0.7001646161,-0.3557700813,0.1334788501,-0.1025939882,0.1700500995,0.286038965,0.3981005549,-0.2447240502,-0.1887209415,0.1763647199,0.0670010895,-0.1224941909,0.2717932463,-0.1491892934,0.0327157974,0.097770609,0.0750924945,-0.0500117987,0.18273063,0.1619618088,0.1637403816,0.2166753411,0.1712021381,-0.1385445595,0.140000239,-0.1463423222,0.0411113389,-0.3651643991,0.0879890844,0.2932605743,-0.2734499574,0.4256699085,0.2981343269,0.0132941371,0.1791388988,0.4029976428,0.3924373686,0.3411541283,-0.321372956,-0.1617229134,-0.2840147913,0.3639715314,-0.0266983099,0.1606378555,-0.5455216169,0.0793506429,-0.0012819576,-0.0612959526,0.0257043056,0.1679891646,-0.1000321358,0.1487789601,-0.3114846647,-0.11306867,-0.0751886219,-0.0874413028,-0.1050553843,-0.4513826668,0.3197152019,-0.0907447189,0.0485186726,0.0477204211,0.1198868528,-0.2438071072,-0.111851804,0.4397116303,0.2618428767,0.168341428,-0.293587923,-0.025957346,-0.2678938508,0.0190906841,-0.31063658,0.0170399696,-0.0132976202,0.3414580822,-0.2391727418,0.0636947304,-0.366689086,0.1739198714,0.2756896913,0.0119208954,0.0677406937,-0.0571765639,-0.2674063444,0.0780813769,0.0970485508,0.5984996557,0.094274886,0.2917085886,-0.2487270087,-0.0865626112,0.3026677072,-0.3510763347,-0.5694525242,0.3708775342,0.0836350396,0.2353769988,-0.164602235,-0.272603184,0.0098311165,0.1228054836,-0.24636361,-0.1148190275,0.038476415,0.0152509697,0.1341851056,-0.0320450515,0.290769279,0.1469573975,-0.4093950093,0.0017055458,-0.2389717102]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/638","title":"GLUE\/QQP dataset: NonMatchingChecksumError","comments":"Hi ! Sure I'll take a look","body":"Hi @lhoestq , I know you are busy and there are also other important issues. But if this is easy to be fixed, I am shamelessly wondering if you can give me some help , so I can evaluate my models and restart with my developing cycle asap. \ud83d\ude1a\r\n\r\ndatasets version: editable install of master at 9\/17\r\n\r\n`datasets.load_dataset('glue','qqp', cache_dir='.\/datasets')`\r\n\r\n```\r\nDownloading and preparing dataset glue\/qqp (download: 57.73 MiB, generated: 107.02 MiB, post-processed: Unknown size, total: 164.75 MiB) to .\/datasets\/glue\/qqp\/1.0.0\/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4...\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in \r\n----> 1 datasets.load_dataset('glue','qqp', cache_dir='.\/datasets')\r\n\r\n~\/datasets\/src\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n~\/datasets\/src\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 467 if not downloaded_from_gcs:\r\n 468 self._download_and_prepare(\r\n--> 469 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 470 )\r\n 471 # Sync info\r\n\r\n~\/datasets\/src\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 527 if verify_infos:\r\n 528 verify_checksums(\r\n--> 529 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 530 )\r\n 531 \r\n\r\n~\/datasets\/src\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/dl.fbaipublicfiles.com\/glue\/data\/QQP-clean.zip']\r\n```","comment_length":7,"text":"GLUE\/QQP dataset: NonMatchingChecksumError\nHi @lhoestq , I know you are busy and there are also other important issues. But if this is easy to be fixed, I am shamelessly wondering if you can give me some help , so I can evaluate my models and restart with my developing cycle asap. \ud83d\ude1a\r\n\r\ndatasets version: editable install of master at 9\/17\r\n\r\n`datasets.load_dataset('glue','qqp', cache_dir='.\/datasets')`\r\n\r\n```\r\nDownloading and preparing dataset glue\/qqp (download: 57.73 MiB, generated: 107.02 MiB, post-processed: Unknown size, total: 164.75 MiB) to .\/datasets\/glue\/qqp\/1.0.0\/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4...\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n in \r\n----> 1 datasets.load_dataset('glue','qqp', cache_dir='.\/datasets')\r\n\r\n~\/datasets\/src\/datasets\/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 609 download_config=download_config,\r\n 610 download_mode=download_mode,\r\n--> 611 ignore_verifications=ignore_verifications,\r\n 612 )\r\n 613 \r\n\r\n~\/datasets\/src\/datasets\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 467 if not downloaded_from_gcs:\r\n 468 self._download_and_prepare(\r\n--> 469 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 470 )\r\n 471 # Sync info\r\n\r\n~\/datasets\/src\/datasets\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 527 if verify_infos:\r\n 528 verify_checksums(\r\n--> 529 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 530 )\r\n 531 \r\n\r\n~\/datasets\/src\/datasets\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 37 if len(bad_urls) > 0:\r\n 38 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 40 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 41 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['https:\/\/dl.fbaipublicfiles.com\/glue\/data\/QQP-clean.zip']\r\n```\nHi ! Sure I'll take a look","embeddings":[-0.3650631905,0.1962519735,-0.1235057488,0.0305407811,0.2150208056,-0.0330680311,-0.0687254071,0.4810118675,0.3090820611,0.1088016108,-0.2082169354,0.2490570843,0.0065834755,0.1479397863,-0.1978857666,0.2178172916,-0.1402305663,0.2429261506,-0.3468612432,0.0008772914,-0.2568644583,0.311511606,0.1454691589,-0.037117064,-0.0508371294,-0.0996373892,0.0225186907,0.0506936982,-0.1567400843,-0.1524573714,-0.0441206656,0.3554534018,-0.0527788065,0.1845129728,-0.0001048251,0.1206158549,0.2548413873,0.0099243093,-0.0700256601,0.0630101413,-0.6860588193,-0.4205325544,-0.0632946491,-0.0574780144,-0.1856556684,0.2971064448,0.014063336,0.0273699332,0.1056426838,0.1788785607,0.3663617969,0.4100304246,0.239244774,-0.0993599445,0.2890217006,-0.4846872091,0.0123000937,0.1321248412,0.2771624625,-0.2537031174,-0.2803534865,0.3557164967,-0.0638507679,0.1636538506,0.0280269496,-0.2150920779,0.3926641047,-0.3394522071,0.1215399653,0.3807606697,0.02845346,-0.3346045911,-0.1956349015,-0.1409159154,-0.152468726,-0.1387741417,0.1974964142,0.0815301538,-0.0096422387,0.009191554,-0.1941277236,0.0041787275,0.1872148812,-0.0287017599,0.1020861641,0.0277870893,0.009692179,-0.051424019,0.215486154,-0.1123123169,-0.0434386395,-0.2070212513,-0.1726285666,-0.2186170816,-0.5142478943,-0.1079793423,0.2796370685,0.3580858111,0.2040474713,0.2291440219,0.0207327642,0.1663113981,0.2158976942,0.1835996807,0.0717440099,0.3722954988,0.1897196025,-0.073409304,0.2134970278,0.2280592024,-0.269130826,0.1371052265,0.0912295133,-0.4280964434,0.2254415303,0.2580365241,0.0552637167,-0.2747471035,-0.2795647383,0.0197346359,0.045313023,-0.0639995039,-0.0465988852,0.4440578818,-0.2446081936,0.0562118441,-0.1668226421,0.0749490634,-0.2304637581,-0.376978457,-0.3119105101,-0.0639677271,-0.1396963,-0.0564260781,-0.0109345075,-0.2114254087,0.2194265723,-0.0692747831,0.0254345536,0.164484784,0.3677846789,-0.0609386712,-0.3187937737,0.4472129643,0.0252656601,0.1060391292,0.267300576,-0.1315524578,-0.1723053157,-0.0209263097,-0.070948191,-0.3960473835,-0.1307325959,0.3236110508,-0.2518887222,0.0305112842,0.1980349272,0.0180907659,0.3554721475,-0.2470597178,-0.0271833893,-0.1717544496,-0.0267702322,-0.3376527429,0.1118895859,0.128261283,-0.3563407063,-0.2266373038,-0.0205671936,-0.2394116372,0.1602957547,0.2766918242,0.0560979582,-0.0107363174,-0.2444953471,0.1683665812,0.1911902428,-0.1239527538,-0.6465619206,0.2447999418,-0.1370737851,0.1640499085,0.0881038159,-0.0142337652,0.1690144241,-0.0482003167,0.1858963966,0.0407693908,-0.0626278371,0.0194061939,-0.6014880538,-0.3445346653,0.2864218056,0.0696169212,0.0313448384,-0.0533347614,0.0483854227,-0.5890367031,0.1011041105,0.132873252,0.0534566902,0.2539786398,0.4030315876,-0.027172545,0.0397618636,-0.1177890673,-0.4048752487,0.3663899899,-0.3589467108,0.1689655036,0.0159482341,-0.0726779327,-0.1731645018,-0.1119106635,-0.1966955066,-0.1022052169,0.3123805821,0.0628054515,0.2177791595,-0.1044224799,0.0620517433,0.3803686202,0.026446281,0.0683532655,-0.498826921,0.1355920285,-0.2726540864,-0.0513453484,0.0612431467,0.1347525269,0.2113454342,-0.0627762899,-0.0227532983,0.5018537045,-0.0248763654,0.0237965118,0.0813682228,0.2808023393,-0.0775197074,-0.0222982857,-0.0654516593,0.1483798623,-0.0292976163,-0.0440152884,0.0631078333,0.3130654991,0.1534221172,0.4348816574,0.0270173755,0.1149048805,0.1049804688,-0.2004866898,-0.1465618461,-0.0535230115,0.3702305257,-0.0439681932,0.134275496,0.1561015099,0.1170144379,0.2556263804,0.2443335205,0.1038600281,0.257183075,-0.0588877052,-0.1836112887,-0.1258016676,0.0997289717,0.6300245523,0.3025740385,0.3073824346,0.01022931,-0.0542722233,-0.207089901,-0.1009188592,-0.0439228751,-0.0168120638,-0.0330733955,0.5266809464,0.3020893335,-0.2886449695,-0.2688984573,0.0763361976,0.0295359474,0.2115181535,-0.2110704482,-0.2006745189,-0.3368225992,0.070351243,0.1805512607,-0.2219465673,-0.3362565637,-0.4178234637,0.2449556589,-0.0055316589,-0.0849546269,0.4378315508,-0.076662086,-0.0288986918,-0.012571726,0.0281864088,-0.052345559,-0.1373202503,-0.0644090623,0.1317575127,-0.0055571236,-0.0205284506,0.3705486655,-0.1160964146,-0.0258435216,-0.2446191311,-0.3858230114,0.0517475866,0.0198435485,0.3773944378,0.4440660775,-0.0079118656,0.2512354553,-0.2213972956,0.4821390212,-0.0028145351,-0.2190364897,0.113586992,-0.1628128141,0.1468470097,-0.0902449191,-0.1874162108,-0.2316574007,-0.480322063,0.1482337415,-0.0148826558,0.1861845106,0.0865285471,0.0196707118,0.0170828942,-0.1916924268,0.1590493917,-0.2550878227,-0.3782135546,0.3728185892,-0.0872483775,-0.2741655707,0.0041132523,-0.3776219487,0.3588182926,-0.0538312979,-0.462300092,-0.4257757962,-0.361734122,0.140639931,0.2683550715,0.0066567026,0.2981892824,0.1599345952,-0.2407853156,-0.1765527725,-0.3038472533,0.2713657618,0.0797149017,0.3143028021,-0.2269009799,0.5414258838,-0.0390907638,0.5156348348,0.0573358797,0.2094944417,0.0351418704,0.0341601782,0.3793818951,-0.0664960593,-0.2141671777,0.0291889105,-0.0501799099,-0.028013017,0.071977824,0.0732845813,-0.0091875428,-0.0043602916,0.1662132293,-0.153477177,-0.2475631535,-0.0288516153,0.1173237115,0.1750668883,0.2975485027,0.0093084369,-0.1155454069,0.0356166475,0.2283277363,0.2288161218,-0.107960932,-0.272786051,-0.2606944442,-0.0201772973,-0.3037365079,0.3041863143,-0.069105342,0.1419113576,-0.1335133314,-0.114509508,-0.0167324003,-0.1531307399,0.6965787411,-0.4467490017,-0.0609181449,0.1860602647,-0.1845618337,-0.0849947557,-0.1380934417,-0.0670222118,-0.2307635844,0.1222133487,0.1436393559,-0.453040719,-0.0438402295,-0.0368505567,0.0355673097,-0.2282343358,-0.1748611629,-0.3163822889,-0.0098126084,-0.4208613932,-0.0508561656,-0.2226459086,0.1481028795,-0.0402867943,0.0342668071,0.021723045,-0.1523262113,0.2644501925,-0.0189654082,0.3451315165,-0.0580216832,0.4014099538,-0.0626601875,-0.0268991124,-0.0290011689,0.301527679,-0.2882125378,-0.1243282929,0.1067915559,-0.0029437989,0.0920218974,0.1087520346,-0.0189656056,0.0220061149,0.0785718933,0.2270199507,0.1134936735,0.0051786327,0.2286676019,0.0824933872,-0.1337325871,-0.0964887366,0.0041714171,-0.0560275428,-0.1275842637,-0.0208262857,0.0287565198,0.0006236063,0.1627625227,0.0146937901,0.8571218252,0.0224097241,-0.0242386591,0.1520280838,-0.4575130939,0.2288752645,-0.3107740283,0.0918884277,-0.4647422433,-0.5296952724,0.0109643154,-0.1254846156,0.0854337513,0.152937457,-0.2449940145,0.3977669179,-0.0673462451,-0.0074013174,0.0317809731,-0.0654386282,-0.0647530779,-0.0302895121,0.0824032351,0.3020190001,-0.2132348865,0.1333882064,-0.0592920594,-0.1437421143,0.3398318589,-0.0032236583,-0.0896132886,0.0885343403,-0.4293051362,0.0808462948,0.3521946669,0.0510565154,-0.3528407812,-0.0996696576,-0.0380380563,0.0926424861,-0.083297804,0.1071196124,0.6844993234,0.0468169749,0.1393172741,0.0377240889,0.1486459225,-0.0614378639,-0.0625254214,0.0177074429,-0.1595554203,-0.2094860226,-0.0611281916,-0.0150545845,0.2643513381,-0.2782352269,-0.1018016264,-0.0361815616,0.0900871381,-0.3350185454,0.2356123626,0.1903344095,-0.1853002906,0.2091421783,-0.0617509484,-0.2167958319,-0.2251700163,0.3962036967,0.02600771,0.0560546257,0.3702644408,0.030110402,-0.3334591091,-0.2968786061,-0.1077755541,0.1844270378,-0.3190284967,0.2102849185,-0.1081856117,-0.2645507455,0.0568307601,0.3402237892,0.2678715885,0.0562516637,0.0027437813,-0.1926036626,-0.2547009587,0.0753074959,0.1983680725,0.2389294952,-0.2917750776,-0.1506741345,-0.0729497671,0.3109094799,-0.4823000133,-0.1003986672,-0.3458444476,-0.0065361052,0.0761783347,-0.1883275956,0.1488784999,0.1718799472,0.2215852439,-0.055479154,-0.4672245979,-0.3613803983,-0.112116389,0.0817915946,-0.0246247575,0.1415217966,0.1522618234,0.0945477113,-0.1327701062,-0.031821005,0.0390705206,-0.006885164,-0.1174119264,0.077644065,0.2634224892,-0.3193992078,-0.0173510835,0.0327077061,-0.1240687966,-0.0698205903,-0.0495080277,-0.160652861,-0.3891847432,-0.0500397757,0.2463904023,0.0351430997,0.0531844385,0.0322598107,0.3560526073,-0.2187714428,-0.2059429288,0.2249713987,0.5275015235,0.0658603609,-0.0861707702,-0.1225061938,0.0980525315,0.3764437735,-0.2934448719,0.0175777264,0.1562343091,-0.011856542,0.4093005061,0.1135142967,0.0355978385,-0.3193938732,0.6556395292,0.2220013738,0.1110098809,-0.5080180168,-0.166004315,0.474457562,-0.0757983476,-0.0091725169,0.4678161442,-0.097515583,0.027177155,-0.0512219146,0.0440904051,0.1950539201,0.1647211015,0.0937380418,0.0728595555,-0.2874989212,0.3726451993,0.3998765945,0.2711029947,-0.0870173275,0.0192773398,0.3814283609,-0.0898479223,0.0075767599,-0.1937000901,0.1712236106,-0.1186542511,-0.1785885245,0.1602620184,-0.0270177461,-0.2150950581,0.0624956451,-0.0175416488,0.1291597486,0.2825272381,-0.051350981,0.1678371131,-0.3311964571,-0.2865743339,0.0728335902,-0.0231139455,-0.1399755776,0.2592064738,0.1491687149,0.0411834009,0.0421014056,0.4430114627,0.3218668997,0.2864172459,-0.2686313391,0.0015228509,0.0739781857,-0.0078135738,-0.231872648,0.519661963,0.1222428307,0.05969432,0.2745096385,0.2064560801,-0.3668713272,0.067281507,-0.1706021428,-0.1112614349,-0.0066816569,0.3858733177,0.1084976196,-0.0324997641,-0.188652873,0.0022835229,-0.4708926678,-0.2489531934,0.1562063694,-0.1883525252,0.1175481007,-0.1168772131,0.1839786023,-0.2057692707,0.3555119634,0.1809857637,-0.0371885076,-0.4695032239,-0.0737795681,-0.6462201476,0.3114566505,-0.0680140033,0.0381957293,-0.0369437113,0.3217431605,-0.0021356156,0.0471706986,0.0687011853,0.0923248529,-0.3843193352,-0.0723973066,-0.3665532768,0.1091312543,-0.0925436169,-0.1441311985,0.2491863221,-0.2177608609,0.2753472328,0.2990944386,0.2214644402,0.1656908989,-0.1268304139,0.1441126168,0.197955206,0.1718587279,0.1874152273,0.3140314519,-0.1425583065,-0.0248075742,-0.1302254051,-0.2573948205,-0.1985624731,0.0193509068,-0.0278101657,0.2541037202,0.0806493387,-0.0513484553,-0.2891745269,0.2457434237,0.3827511072,0.0045677144,-0.084671244,0.2094036937,0.2048760653,0.1019089371,0.0183392074,0.119390972,0.2462609708,0.2452925742,-0.3364291787,-0.2327938974,0.5535590649,-0.0860348865,-0.1422073245,-0.0817946866,0.3107659817,0.0481624827,-0.3635163009,-0.3502314687,-0.0957990214,0.2636708617,0.0149257472,-0.1011487916,0.1438498199,-0.1642894149,0.1807840914,-0.0321537815,0.1540705562,0.1616960615,-0.2038720548,-0.1178989187,-0.0627139509]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Not sure what could cause that on the `datasets` side. Could this be a `Trainer` issue ? cc @julien-c @sgugger ?","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":21,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nNot sure what could cause that on the `datasets` side. Could this be a `Trainer` issue ? cc @julien-c @sgugger ?","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"There was a memory leak issue fixed recently in master. You should install from source and see if it fixes your problem.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":22,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nThere was a memory leak issue fixed recently in master. You should install from source and see if it fixes your problem.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@lhoestq @sgugger Thanks for your comments. I have install from source code as you told, but the problem is still there.\r\nTo reproduce the issue, just replace [these lines](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L241-L258) with: \r\n(load_dataset and DataCollatorForDatasetsLanguageModeling as [above mentioned](https:\/\/github.com\/huggingface\/datasets\/issues\/633#issue-702440484))\r\n```python\r\n dataset = load_dataset(\"bookcorpus\")\r\n dataset = dataset.train_test_split(test_size=0.1)\r\n train_dataset = dataset['train']\r\n eval_dataset = dataset['test'] if training_args.do_eval else None\r\n\r\n data_collator = DataCollatorForDatasetsLanguageModeling(\r\n tokenizer=tokenizer,\r\n mlm=data_args.mlm,\r\n mlm_probability=data_args.mlm_probability,\r\n block_size=data_args.block_size\r\n )\r\n```\r\nand run by:\r\n```bash\r\npython run_language_modeling.py\r\n--output_dir=output \\\r\n--model_type=bert \\\r\n--model_name_or_path=bert-base-uncased \\\r\n--do_train \\\r\n--do_eval \\\r\n--mlm \r\n```","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":80,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@lhoestq @sgugger Thanks for your comments. I have install from source code as you told, but the problem is still there.\r\nTo reproduce the issue, just replace [these lines](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L241-L258) with: \r\n(load_dataset and DataCollatorForDatasetsLanguageModeling as [above mentioned](https:\/\/github.com\/huggingface\/datasets\/issues\/633#issue-702440484))\r\n```python\r\n dataset = load_dataset(\"bookcorpus\")\r\n dataset = dataset.train_test_split(test_size=0.1)\r\n train_dataset = dataset['train']\r\n eval_dataset = dataset['test'] if training_args.do_eval else None\r\n\r\n data_collator = DataCollatorForDatasetsLanguageModeling(\r\n tokenizer=tokenizer,\r\n mlm=data_args.mlm,\r\n mlm_probability=data_args.mlm_probability,\r\n block_size=data_args.block_size\r\n )\r\n```\r\nand run by:\r\n```bash\r\npython run_language_modeling.py\r\n--output_dir=output \\\r\n--model_type=bert \\\r\n--model_name_or_path=bert-base-uncased \\\r\n--do_train \\\r\n--do_eval \\\r\n--mlm \r\n```","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Same here. Pre-training on wikitext-103 to do some test. At the end of the training it takes 32GB of RAM + ~30GB of SWAP. I installed dataset==1.1.0, not built from source. I will try uninstalling and building from source when it finish.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":42,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nSame here. Pre-training on wikitext-103 to do some test. At the end of the training it takes 32GB of RAM + ~30GB of SWAP. I installed dataset==1.1.0, not built from source. I will try uninstalling and building from source when it finish.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"This seems to be on the `transformers` library side.\r\n\r\nIf you have more informations (pip env) or even better, a colab reproducing the error we can investigate.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":27,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nThis seems to be on the `transformers` library side.\r\n\r\nIf you have more informations (pip env) or even better, a colab reproducing the error we can investigate.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"It seems like it's solved with freshed versions of transformers. I have tried to replicate the error doing a fresh pip install transformers & datasets on colab and the error doesn't continue. On colab it keeps stable on 5GB! (Y)\r\n\r\nEdit: **Thanks for your great work**. Have a good day.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":50,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nIt seems like it's solved with freshed versions of transformers. I have tried to replicate the error doing a fresh pip install transformers & datasets on colab and the error doesn't continue. On colab it keeps stable on 5GB! (Y)\r\n\r\nEdit: **Thanks for your great work**. Have a good day.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@gaceladri witch version transformers and datasets are you using now? I want to try again. Thanks.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":16,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@gaceladri witch version transformers and datasets are you using now? I want to try again. Thanks.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"transformers==3.3.1\r\ndatasets==1.1.0\r\ntokenizers==0.8.1rc2\r\n","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":3,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \ntransformers==3.3.1\r\ndatasets==1.1.0\r\ntokenizers==0.8.1rc2\r\n","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"doing some modifications to mobilebert\r\nhttps:\/\/colab.research.google.com\/drive\/1ba09ZOpyHGAOQLcsxiQAHRXl10qnMU5o?usp=sharing ","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":6,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \ndoing some modifications to mobilebert\r\nhttps:\/\/colab.research.google.com\/drive\/1ba09ZOpyHGAOQLcsxiQAHRXl10qnMU5o?usp=sharing ","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"It does not happen to me anymore. Can we close? @leethu2012 ","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":11,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nIt does not happen to me anymore. Can we close? @leethu2012 ","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"It's happening to me again. After 4 hours of pre-training, my ram memory gets full and the kernel dies. I am using the last transformers version as today. 4.4.0 and the last version of datasets 1.2.1, both installed from master. The memory consumption keeps increasing.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":45,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nIt's happening to me again. After 4 hours of pre-training, my ram memory gets full and the kernel dies. I am using the last transformers version as today. 4.4.0 and the last version of datasets 1.2.1, both installed from master. The memory consumption keeps increasing.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"It looks like it is something from pytorch\/python itself :face_with_head_bandage: https:\/\/github.com\/pytorch\/pytorch\/issues\/13246 ","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":11,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nIt looks like it is something from pytorch\/python itself :face_with_head_bandage: https:\/\/github.com\/pytorch\/pytorch\/issues\/13246 ","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Thanks for the investigation @gaceladri \r\n\r\nApparently this happens when `num_workers>0` and has to do with objects being copied-on-write.\r\nDid you try setting num_workers to 0 @gaceladri ?\r\nIf the issue doesn't happen with `num_workers=0` then this would confirm that it's indeed related to this python\/pytorch issue.\r\n\r\nSince a `Dataset` object is a wrapper of a pyarrow Table, we should investigate if the data being copied comes from the Table itself or from metadata in the `Dataset` object. If it comes from the metadata in the `Dataset` object, we should be able to implement a workaround. But if it comes from the Table, we'll need to see with the pyarrow team what we can do... ","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":114,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nThanks for the investigation @gaceladri \r\n\r\nApparently this happens when `num_workers>0` and has to do with objects being copied-on-write.\r\nDid you try setting num_workers to 0 @gaceladri ?\r\nIf the issue doesn't happen with `num_workers=0` then this would confirm that it's indeed related to this python\/pytorch issue.\r\n\r\nSince a `Dataset` object is a wrapper of a pyarrow Table, we should investigate if the data being copied comes from the Table itself or from metadata in the `Dataset` object. If it comes from the metadata in the `Dataset` object, we should be able to implement a workaround. But if it comes from the Table, we'll need to see with the pyarrow team what we can do... ","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@lhoestq I have tried and it keeps increasing also with `dataloader_num_workers=0`","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":11,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@lhoestq I have tried and it keeps increasing also with `dataloader_num_workers=0`","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Hmmm so this might come from another issue...\r\nSince it doesn't seem to be related to multiprocessing it should be easier to investigate though.\r\nDo you have some ideas @gaceladri ?","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":31,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nHmmm so this might come from another issue...\r\nSince it doesn't seem to be related to multiprocessing it should be easier to investigate though.\r\nDo you have some ideas @gaceladri ?","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@lhoestq I looked quickly to a previously spoted bug in my env wandb \/sdk\/interface\/interface.py, because sometimes when I load the dataset I got a multiprocessing error at line 510 in wandb...interface.py\r\n\r\nThis bug is reported here https:\/\/github.com\/huggingface\/datasets\/issues\/847\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<timed eval> in <module>\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/transformers\/trainer.py in train(self, model_path, trial)\r\n 877 print(len(epoch_iterator))\r\n 878 \r\n--> 879 for step, inputs in enumerate(epoch_iterator):\r\n 880 \r\n 881 start_step = time.time()\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 433 if self._sampler_iter is None:\r\n 434 self._reset()\r\n--> 435 data = self._next_data()\r\n 436 self._num_yielded += 1\r\n 437 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 1083 else:\r\n 1084 del self._task_info[idx]\r\n-> 1085 return self._process_data(data)\r\n 1086 \r\n 1087 def _try_put_index(self):\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in _process_data(self, data)\r\n 1109 self._try_put_index()\r\n 1110 if isinstance(data, ExceptionWrapper):\r\n-> 1111 data.reraise()\r\n 1112 return data\r\n 1113 \r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/_utils.py in reraise(self)\r\n 426 # have message field\r\n 427 raise self.exc_type(message=msg)\r\n--> 428 raise self.exc_type(msg)\r\n 429 \r\n 430 \r\n\r\nAssertionError: Caught AssertionError in DataLoader worker process 0.\r\nOriginal Traceback (most recent call last):\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/worker.py\", line 198, in _worker_loop\r\n data = fetcher.fetch(index)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1083, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1070, in _getitem\r\n format_kwargs=format_kwargs,\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 886, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 847, in command\r\n return torch.tensor(x, **format_kwargs)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/warnings.py\", line 101, in _showwarnmsg\r\n _showwarnmsg_impl(msg)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/warnings.py\", line 30, in _showwarnmsg_impl\r\n file.write(text)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 729, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 186, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 191, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 510, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n```\r\n\r\nMy workaround was to just comment those lines without looking to much into consecuences:\r\n\r\n```\r\ndef _publish(self, record: pb.Record, local: bool = None) -> None:\r\n #if self._process and not self._process.is_alive():\r\n # raise Exception(\"The wandb backend process has shutdown\")\r\n```\r\n\r\nIt worked so far... I need to try running without wandb and see if it could be causing something wrong with multiprocessing. I am going to try to launch the training setting wandb to false and I will let you know again.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":396,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@lhoestq I looked quickly to a previously spoted bug in my env wandb \/sdk\/interface\/interface.py, because sometimes when I load the dataset I got a multiprocessing error at line 510 in wandb...interface.py\r\n\r\nThis bug is reported here https:\/\/github.com\/huggingface\/datasets\/issues\/847\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<timed eval> in <module>\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/transformers\/trainer.py in train(self, model_path, trial)\r\n 877 print(len(epoch_iterator))\r\n 878 \r\n--> 879 for step, inputs in enumerate(epoch_iterator):\r\n 880 \r\n 881 start_step = time.time()\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 433 if self._sampler_iter is None:\r\n 434 self._reset()\r\n--> 435 data = self._next_data()\r\n 436 self._num_yielded += 1\r\n 437 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 1083 else:\r\n 1084 del self._task_info[idx]\r\n-> 1085 return self._process_data(data)\r\n 1086 \r\n 1087 def _try_put_index(self):\r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/dataloader.py in _process_data(self, data)\r\n 1109 self._try_put_index()\r\n 1110 if isinstance(data, ExceptionWrapper):\r\n-> 1111 data.reraise()\r\n 1112 return data\r\n 1113 \r\n\r\n~\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/_utils.py in reraise(self)\r\n 426 # have message field\r\n 427 raise self.exc_type(message=msg)\r\n--> 428 raise self.exc_type(msg)\r\n 429 \r\n 430 \r\n\r\nAssertionError: Caught AssertionError in DataLoader worker process 0.\r\nOriginal Traceback (most recent call last):\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/worker.py\", line 198, in _worker_loop\r\n data = fetcher.fetch(index)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1083, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1070, in _getitem\r\n format_kwargs=format_kwargs,\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 886, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 847, in command\r\n return torch.tensor(x, **format_kwargs)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/warnings.py\", line 101, in _showwarnmsg\r\n _showwarnmsg_impl(msg)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/warnings.py\", line 30, in _showwarnmsg_impl\r\n file.write(text)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/lib\/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/wandb_run.py\", line 729, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 186, in publish_output\r\n self._publish_output(o)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 191, in _publish_output\r\n self._publish(rec)\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/site-packages\/wandb\/sdk\/interface\/interface.py\", line 510, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"\/home\/ad\/anaconda3\/envs\/tfm\/lib\/python3.6\/multiprocessing\/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n```\r\n\r\nMy workaround was to just comment those lines without looking to much into consecuences:\r\n\r\n```\r\ndef _publish(self, record: pb.Record, local: bool = None) -> None:\r\n #if self._process and not self._process.is_alive():\r\n # raise Exception(\"The wandb backend process has shutdown\")\r\n```\r\n\r\nIt worked so far... I need to try running without wandb and see if it could be causing something wrong with multiprocessing. I am going to try to launch the training setting wandb to false and I will let you know again.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@lhoestq But despite this, I got lost into the [class Dataset()](https:\/\/huggingface.co\/docs\/datasets\/_modules\/datasets\/arrow_dataset.html#Dataset) reading the pyarrow files.\r\n\r\nEdit: but you should be rigth, that it does not have to be related to multiprocessing since it keeps happening when `num_workers=0` ","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":37,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@lhoestq But despite this, I got lost into the [class Dataset()](https:\/\/huggingface.co\/docs\/datasets\/_modules\/datasets\/arrow_dataset.html#Dataset) reading the pyarrow files.\r\n\r\nEdit: but you should be rigth, that it does not have to be related to multiprocessing since it keeps happening when `num_workers=0` ","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Or maybe wandb uses multiprocessing ? One process for wandb logging and one for actual training ? If this is the case then even setting `num_workers=0` would cause the process to be forked for wandb and therefore cause the memory issue.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":41,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nOr maybe wandb uses multiprocessing ? One process for wandb logging and one for actual training ? If this is the case then even setting `num_workers=0` would cause the process to be forked for wandb and therefore cause the memory issue.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@lhoestq could be, but if we set wandb to false this should not happen. I am going to try.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":19,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@lhoestq could be, but if we set wandb to false this should not happen. I am going to try.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@lhoestq It keeps happening. I have uninstalled wandb from my env, setted `%env WANDB_DISABLED=true` on my notebook, and commented this func:\r\n\r\n```\r\ndef get_available_reporting_integrations():\r\n integrations = []\r\n if is_azureml_available():\r\n integrations.append(\"azure_ml\")\r\n if is_comet_available():\r\n integrations.append(\"comet_ml\")\r\n if is_mlflow_available():\r\n integrations.append(\"mlflow\")\r\n if is_tensorboard_available():\r\n integrations.append(\"tensorboard\")\r\n # if is_wandb_available():\r\n # integrations.append(\"wandb\")\r\n return integrations\r\n```\r\nAs a fast test and it keeps increasing the ram memory. Wandb could not be the blameworthy here.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":65,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@lhoestq It keeps happening. I have uninstalled wandb from my env, setted `%env WANDB_DISABLED=true` on my notebook, and commented this func:\r\n\r\n```\r\ndef get_available_reporting_integrations():\r\n integrations = []\r\n if is_azureml_available():\r\n integrations.append(\"azure_ml\")\r\n if is_comet_available():\r\n integrations.append(\"comet_ml\")\r\n if is_mlflow_available():\r\n integrations.append(\"mlflow\")\r\n if is_tensorboard_available():\r\n integrations.append(\"tensorboard\")\r\n # if is_wandb_available():\r\n # integrations.append(\"wandb\")\r\n return integrations\r\n```\r\nAs a fast test and it keeps increasing the ram memory. Wandb could not be the blameworthy here.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Thanks for checking @gaceladri . Let's investigate the single process setting then.\r\nIf you have some sort of colab notebook with a minimal code example that shows this behavior feel free to share it @gaceladri so that we can play around with it to find what causes this. Otherwise I'll probably try to reproduce on my side at one point","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":60,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nThanks for checking @gaceladri . Let's investigate the single process setting then.\r\nIf you have some sort of colab notebook with a minimal code example that shows this behavior feel free to share it @gaceladri so that we can play around with it to find what causes this. Otherwise I'll probably try to reproduce on my side at one point","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@lhoestq sure. Here you have https:\/\/colab.research.google.com\/drive\/1ba09ZOpyHGAOQLcsxiQAHRXl10qnMU5o?usp=sharing let me know if the link works and it reproduces the issue. To me, it reproduces the issue, since if you start the training the ram memory keeps increasing.\r\n\r\nLet me know. Thanks!","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":39,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@lhoestq sure. Here you have https:\/\/colab.research.google.com\/drive\/1ba09ZOpyHGAOQLcsxiQAHRXl10qnMU5o?usp=sharing let me know if the link works and it reproduces the issue. To me, it reproduces the issue, since if you start the training the ram memory keeps increasing.\r\n\r\nLet me know. Thanks!","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Could the bug be comming from tokenizers?\r\n\r\nI got this warning at the terminal from my jupyter notebook: \r\n```\r\nhuggingface\/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...\r\nTo disable this warning, you can either:\r\n\t- Avoid using `tokenizers` before the fork if possible\r\n\t- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)\r\n```","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":63,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nCould the bug be comming from tokenizers?\r\n\r\nI got this warning at the terminal from my jupyter notebook: \r\n```\r\nhuggingface\/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...\r\nTo disable this warning, you can either:\r\n\t- Avoid using `tokenizers` before the fork if possible\r\n\t- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)\r\n```","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"I've never experienced memory issues with tokenizers so I don't know\r\nCc @n1t0 are you aware of any issue that would cause memory to keep increasing when the tokenizer is used in the Data Collator for language modeling ?","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":39,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nI've never experienced memory issues with tokenizers so I don't know\r\nCc @n1t0 are you aware of any issue that would cause memory to keep increasing when the tokenizer is used in the Data Collator for language modeling ?","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"@lhoestq Thanks for pointing to n1t0, just to clarify. That warning was doing fine-tuning, without collator:\r\n```\r\n\r\nfrom datasets import load_dataset, load_metric\r\nimport numpy as np\r\n\r\nGLUE_TASKS = [\r\n \"cola\",\r\n \"mnli\",\r\n \"mnli-mm\",\r\n \"mrpc\",\r\n \"qnli\",\r\n \"qqp\",\r\n \"rte\",\r\n \"sst2\",\r\n \"stsb\",\r\n \"wnli\",\r\n]\r\ntask = \"mnli\"\r\nactual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(\"glue\", actual_task)\r\nmetric = load_metric(\"glue\", actual_task)\r\nbatch_size = 16\r\nattention_type = \"linear\"\r\n\r\nfrom transformers.models.mobilebert_mod import (\r\n MobileBertForSequenceClassification,\r\n MobileBertTokenizerFast,\r\n)\r\nfrom transformers.models.mobilebert_mod.configuration_mobilebert import (\r\n MobileBertConfigMod,\r\n)\r\nfrom transformers import TrainingArguments, Trainer\r\n\r\nnum_labels = 3 if task.startswith(\"mnli\") else 1 if task == \"stsb\" else 2\r\ntokenizer = MobileBertTokenizerFast.from_pretrained(\r\n \"\/media\/ad\/00b5422b-9d54-4449-8b5d-08eab5cdac8c\/training_trfm\/big_linear_layerdrop_shared\/checkpoint-23000\/\",\r\n max_len=512,\r\n)\r\nmodel = MobileBertForSequenceClassification.from_pretrained(\r\n \"\/media\/ad\/00b5422b-9d54-4449-8b5d-08eab5cdac8c\/training_trfm\/big_linear_layerdrop_shared\/checkpoint-23000\/\",\r\n num_labels=num_labels,\r\n)\r\nprint(model.num_parameters())\r\n\r\ntask_to_keys = {\r\n \"cola\": (\"sentence\", None),\r\n \"mnli\": (\"premise\", \"hypothesis\"),\r\n \"mnli-mm\": (\"premise\", \"hypothesis\"),\r\n \"mrpc\": (\"sentence1\", \"sentence2\"),\r\n \"qnli\": (\"question\", \"sentence\"),\r\n \"qqp\": (\"question1\", \"question2\"),\r\n \"rte\": (\"sentence1\", \"sentence2\"),\r\n \"sst2\": (\"sentence\", None),\r\n \"stsb\": (\"sentence1\", \"sentence2\"),\r\n \"wnli\": (\"sentence1\", \"sentence2\"),\r\n}\r\n\r\nsentence1_key, sentence2_key = task_to_keys[task]\r\nif sentence2_key is None:\r\n print(f\"Sentence: {dataset['train'][0][sentence1_key]}\")\r\nelse:\r\n print(f\"Sentence 1: {dataset['train'][0][sentence1_key]}\")\r\n print(f\"Sentence 2: {dataset['train'][0][sentence2_key]}\")\r\n\r\n\r\ndef preprocess_function(examples):\r\n if sentence2_key is None:\r\n return tokenizer(examples[sentence1_key], truncation=True)\r\n return tokenizer(examples[sentence1_key], examples[sentence2_key], truncation=True)\r\n\r\n\r\nencoded_dataset = dataset.map(preprocess_function, batched=True)\r\nmetric_name = (\r\n \"pearson\"\r\n if task == \"stsb\"\r\n else \"matthews_correlation\"\r\n if task == \"cola\"\r\n else \"accuracy\"\r\n)\r\n\r\nargs = TrainingArguments(\r\n f\"test-glue\/{task}_{attention_type}\",\r\n evaluation_strategy=\"steps\",\r\n learning_rate=1e-5,\r\n per_device_train_batch_size=batch_size,\r\n per_device_eval_batch_size=batch_size,\r\n logging_steps=200,\r\n num_train_epochs=5,\r\n gradient_accumulation_steps=1,\r\n warmup_steps=10000,\r\n fp16=True,\r\n dataloader_num_workers=10,\r\n weight_decay=0.1,\r\n load_best_model_at_end=True,\r\n metric_for_best_model=metric_name,\r\n)\r\n\r\n\r\ndef compute_metrics(eval_pred):\r\n predictions, labels = eval_pred\r\n if task != \"stsb\":\r\n predictions = np.argmax(predictions, axis=1)\r\n else:\r\n predictions = predictions[:, 0]\r\n return metric.compute(predictions=predictions, references=labels)\r\n\r\n\r\nvalidation_key = (\r\n \"validation_mismatched\"\r\n if task == \"mnli-mm\"\r\n else \"validation_matched\"\r\n if task == \"mnli\"\r\n else \"validation\"\r\n)\r\n\r\ntrainer = Trainer(\r\n model,\r\n args,\r\n train_dataset=encoded_dataset[\"train\"],\r\n eval_dataset=encoded_dataset[validation_key],\r\n tokenizer=tokenizer,\r\n compute_metrics=compute_metrics,\r\n)\r\n\r\ntrainer.train()\r\n```\r\n\r\nNow, I have come back to pre-training. The changes that I think I have done are: not formatting the dataset to torch: ~~`big_dataset.set_format(type='torch', columns=[\"text\", \"input_ids\", \"attention_mask\", \"token_type_ids\"])`~~ so maybe some column is dropped and not freezed in memory and now I have not setted any validation dataset in the trainer. \r\n\r\nMy validation dataset before:\r\n```\r\nbook_corpus_eval = load_dataset(\r\n \"bookcorpus\",\r\n \"plain_text\",\r\n cache_dir=\"\/home\/ad\/Desktop\/bookcorpus\",\r\n split=\"train[98:99%]\",\r\n)\r\nbook_corpus_eval = book_corpus_eval.map(encode, batched=True)\r\nbook_corpus_eval.set_format(\r\n type=\"torch\", columns=[\"text\", \"input_ids\", \"attention_mask\", \"token_type_ids\"]\r\n)\r\n**book_corpus_eval = book_corpus_eval.select([i for i in range(1500)])**\r\n```\r\nMaybe _selecting_ or indexing the dataset before feeding it to the trainer, do something strange.\r\n\r\nMy trainer now:\r\n```\r\n\r\nbig_dataset = load_from_disk(\"\/home\/ad\/Desktop\/35percent_data.arrow\/\")\r\n\r\nfrom transformers import DataCollatorForWholeWordMask\r\n\r\ndata_collator = DataCollatorForWholeWordMask(\r\n tokenizer=tokenizer, mlm=True, mlm_probability=0.15)\r\n\r\nfrom transformers import Trainer, TrainingArguments\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/big_linear_layerdrop_shared_silu_secondtry\",\r\n overwrite_output_dir=True,\r\n per_device_train_batch_size=60,\r\n per_device_eval_batch_size=60,\r\n save_steps=500,\r\n save_total_limit=10,\r\n logging_first_step=True,\r\n logging_steps=100,\r\n# evaluation_strategy='steps',\r\n# eval_steps=250,\r\n gradient_accumulation_steps=8,\r\n fp16=True,\r\n dataloader_num_workers=10,\r\n warmup_steps=15000,\r\n learning_rate=6e-4,\r\n adam_epsilon=1e-6,\r\n adam_beta2=0.98,\r\n weight_decay=0.01,\r\n max_grad_norm=1.0,\r\n max_steps=500000, \r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=big_dataset,\r\n# eval_dataset=book_corpus_eval,\r\n tokenizer=tokenizer)\r\n\r\nimport wandb\r\nwandb.login()\r\n\r\ntrainer.train()\r\n```\r\n\r\nAnd surprisingly, the ram now keeps going up and down. The training is up now for 12h without collapse the ram. I don't know what could cause the leakage. :mag: \r\n\r\nEdit: I didn't see the swap memory, that keeps increasing. So the problem persist. ","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":468,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \n@lhoestq Thanks for pointing to n1t0, just to clarify. That warning was doing fine-tuning, without collator:\r\n```\r\n\r\nfrom datasets import load_dataset, load_metric\r\nimport numpy as np\r\n\r\nGLUE_TASKS = [\r\n \"cola\",\r\n \"mnli\",\r\n \"mnli-mm\",\r\n \"mrpc\",\r\n \"qnli\",\r\n \"qqp\",\r\n \"rte\",\r\n \"sst2\",\r\n \"stsb\",\r\n \"wnli\",\r\n]\r\ntask = \"mnli\"\r\nactual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(\"glue\", actual_task)\r\nmetric = load_metric(\"glue\", actual_task)\r\nbatch_size = 16\r\nattention_type = \"linear\"\r\n\r\nfrom transformers.models.mobilebert_mod import (\r\n MobileBertForSequenceClassification,\r\n MobileBertTokenizerFast,\r\n)\r\nfrom transformers.models.mobilebert_mod.configuration_mobilebert import (\r\n MobileBertConfigMod,\r\n)\r\nfrom transformers import TrainingArguments, Trainer\r\n\r\nnum_labels = 3 if task.startswith(\"mnli\") else 1 if task == \"stsb\" else 2\r\ntokenizer = MobileBertTokenizerFast.from_pretrained(\r\n \"\/media\/ad\/00b5422b-9d54-4449-8b5d-08eab5cdac8c\/training_trfm\/big_linear_layerdrop_shared\/checkpoint-23000\/\",\r\n max_len=512,\r\n)\r\nmodel = MobileBertForSequenceClassification.from_pretrained(\r\n \"\/media\/ad\/00b5422b-9d54-4449-8b5d-08eab5cdac8c\/training_trfm\/big_linear_layerdrop_shared\/checkpoint-23000\/\",\r\n num_labels=num_labels,\r\n)\r\nprint(model.num_parameters())\r\n\r\ntask_to_keys = {\r\n \"cola\": (\"sentence\", None),\r\n \"mnli\": (\"premise\", \"hypothesis\"),\r\n \"mnli-mm\": (\"premise\", \"hypothesis\"),\r\n \"mrpc\": (\"sentence1\", \"sentence2\"),\r\n \"qnli\": (\"question\", \"sentence\"),\r\n \"qqp\": (\"question1\", \"question2\"),\r\n \"rte\": (\"sentence1\", \"sentence2\"),\r\n \"sst2\": (\"sentence\", None),\r\n \"stsb\": (\"sentence1\", \"sentence2\"),\r\n \"wnli\": (\"sentence1\", \"sentence2\"),\r\n}\r\n\r\nsentence1_key, sentence2_key = task_to_keys[task]\r\nif sentence2_key is None:\r\n print(f\"Sentence: {dataset['train'][0][sentence1_key]}\")\r\nelse:\r\n print(f\"Sentence 1: {dataset['train'][0][sentence1_key]}\")\r\n print(f\"Sentence 2: {dataset['train'][0][sentence2_key]}\")\r\n\r\n\r\ndef preprocess_function(examples):\r\n if sentence2_key is None:\r\n return tokenizer(examples[sentence1_key], truncation=True)\r\n return tokenizer(examples[sentence1_key], examples[sentence2_key], truncation=True)\r\n\r\n\r\nencoded_dataset = dataset.map(preprocess_function, batched=True)\r\nmetric_name = (\r\n \"pearson\"\r\n if task == \"stsb\"\r\n else \"matthews_correlation\"\r\n if task == \"cola\"\r\n else \"accuracy\"\r\n)\r\n\r\nargs = TrainingArguments(\r\n f\"test-glue\/{task}_{attention_type}\",\r\n evaluation_strategy=\"steps\",\r\n learning_rate=1e-5,\r\n per_device_train_batch_size=batch_size,\r\n per_device_eval_batch_size=batch_size,\r\n logging_steps=200,\r\n num_train_epochs=5,\r\n gradient_accumulation_steps=1,\r\n warmup_steps=10000,\r\n fp16=True,\r\n dataloader_num_workers=10,\r\n weight_decay=0.1,\r\n load_best_model_at_end=True,\r\n metric_for_best_model=metric_name,\r\n)\r\n\r\n\r\ndef compute_metrics(eval_pred):\r\n predictions, labels = eval_pred\r\n if task != \"stsb\":\r\n predictions = np.argmax(predictions, axis=1)\r\n else:\r\n predictions = predictions[:, 0]\r\n return metric.compute(predictions=predictions, references=labels)\r\n\r\n\r\nvalidation_key = (\r\n \"validation_mismatched\"\r\n if task == \"mnli-mm\"\r\n else \"validation_matched\"\r\n if task == \"mnli\"\r\n else \"validation\"\r\n)\r\n\r\ntrainer = Trainer(\r\n model,\r\n args,\r\n train_dataset=encoded_dataset[\"train\"],\r\n eval_dataset=encoded_dataset[validation_key],\r\n tokenizer=tokenizer,\r\n compute_metrics=compute_metrics,\r\n)\r\n\r\ntrainer.train()\r\n```\r\n\r\nNow, I have come back to pre-training. The changes that I think I have done are: not formatting the dataset to torch: ~~`big_dataset.set_format(type='torch', columns=[\"text\", \"input_ids\", \"attention_mask\", \"token_type_ids\"])`~~ so maybe some column is dropped and not freezed in memory and now I have not setted any validation dataset in the trainer. \r\n\r\nMy validation dataset before:\r\n```\r\nbook_corpus_eval = load_dataset(\r\n \"bookcorpus\",\r\n \"plain_text\",\r\n cache_dir=\"\/home\/ad\/Desktop\/bookcorpus\",\r\n split=\"train[98:99%]\",\r\n)\r\nbook_corpus_eval = book_corpus_eval.map(encode, batched=True)\r\nbook_corpus_eval.set_format(\r\n type=\"torch\", columns=[\"text\", \"input_ids\", \"attention_mask\", \"token_type_ids\"]\r\n)\r\n**book_corpus_eval = book_corpus_eval.select([i for i in range(1500)])**\r\n```\r\nMaybe _selecting_ or indexing the dataset before feeding it to the trainer, do something strange.\r\n\r\nMy trainer now:\r\n```\r\n\r\nbig_dataset = load_from_disk(\"\/home\/ad\/Desktop\/35percent_data.arrow\/\")\r\n\r\nfrom transformers import DataCollatorForWholeWordMask\r\n\r\ndata_collator = DataCollatorForWholeWordMask(\r\n tokenizer=tokenizer, mlm=True, mlm_probability=0.15)\r\n\r\nfrom transformers import Trainer, TrainingArguments\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\".\/big_linear_layerdrop_shared_silu_secondtry\",\r\n overwrite_output_dir=True,\r\n per_device_train_batch_size=60,\r\n per_device_eval_batch_size=60,\r\n save_steps=500,\r\n save_total_limit=10,\r\n logging_first_step=True,\r\n logging_steps=100,\r\n# evaluation_strategy='steps',\r\n# eval_steps=250,\r\n gradient_accumulation_steps=8,\r\n fp16=True,\r\n dataloader_num_workers=10,\r\n warmup_steps=15000,\r\n learning_rate=6e-4,\r\n adam_epsilon=1e-6,\r\n adam_beta2=0.98,\r\n weight_decay=0.01,\r\n max_grad_norm=1.0,\r\n max_steps=500000, \r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=big_dataset,\r\n# eval_dataset=book_corpus_eval,\r\n tokenizer=tokenizer)\r\n\r\nimport wandb\r\nwandb.login()\r\n\r\ntrainer.train()\r\n```\r\n\r\nAnd surprisingly, the ram now keeps going up and down. The training is up now for 12h without collapse the ram. I don't know what could cause the leakage. :mag: \r\n\r\nEdit: I didn't see the swap memory, that keeps increasing. So the problem persist. ","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Thanks for sharing your results.\r\nSo you still had the issue for fine-tuning ?\r\nAnd the issue still appears with a bare-bone dataset from an arrow file...","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":27,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nThanks for sharing your results.\r\nSo you still had the issue for fine-tuning ?\r\nAnd the issue still appears with a bare-bone dataset from an arrow file...","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/633","title":"Load large text file for LM pre-training resulting in OOM","comments":"Yes, on both cases. Fine-tuning a pre-trained model and pre-training from scratch with a local arrow file already pre-processed.","body":"I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. ","comment_length":19,"text":"Load large text file for LM pre-training resulting in OOM\nI tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n@dataclass\r\nclass DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling):\r\n \"\"\"\r\n Data collator used for language modeling based on DataCollatorForLazyLanguageModeling\r\n - collates batches of tensors, honoring their tokenizer's pad_token\r\n - preprocesses batches for masked language modeling\r\n \"\"\"\r\n\r\n block_size: int = 512\r\n\r\n def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]:\r\n examples = [example['text'] for example in examples]\r\n batch, attention_mask = self._tensorize_batch(examples)\r\n if self.mlm:\r\n inputs, labels = self.mask_tokens(batch)\r\n return {\"input_ids\": inputs, \"labels\": labels}\r\n else:\r\n labels = batch.clone().detach()\r\n if self.tokenizer.pad_token_id is not None:\r\n labels[labels == self.tokenizer.pad_token_id] = -100\r\n return {\"input_ids\": batch, \"labels\": labels}\r\n\r\n def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]:\r\n\r\n if self.tokenizer._pad_token is None:\r\n raise ValueError(\r\n \"You are attempting to pad samples but the tokenizer you are using\"\r\n f\" ({self.tokenizer.__class__.__name__}) does not have one.\"\r\n )\r\n\r\n tensor_examples = self.tokenizer.batch_encode_plus(\r\n [ex for ex in examples if ex],\r\n max_length=self.block_size,\r\n return_tensors=\"pt\",\r\n pad_to_max_length=True,\r\n return_attention_mask=True,\r\n truncation=True,\r\n )\r\n\r\n input_ids, attention_mask = tensor_examples[\"input_ids\"], tensor_examples[\"attention_mask\"]\r\n return input_ids, attention_mask\r\n\r\ndataset = load_dataset('text', data_files='train.txt',cache_dir=\".\/\", , split='train')\r\ndata_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, \r\n mlm_probability=0.15, block_size=tokenizer.max_len)\r\ntrainer = Trainer(model=model, args=args, data_collator=data_collator,\r\n train_dataset=train_dataset, prediction_loss_only=True, )\r\ntrainer.train(model_path=model_path)\r\n```\r\nThis train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. \r\nDuring training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/29704017\/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png)\r\n\r\nCould you please give me any suggestions on why this happened and how to fix it?\r\nThanks. \nYes, on both cases. Fine-tuning a pre-trained model and pre-training from scratch with a local arrow file already pre-processed.","embeddings":[-0.6339284182,-0.4775375128,0.0106934961,0.2986351252,0.360047698,-0.1518250406,0.5567328334,0.3738060296,0.010882902,0.0107194036,-0.1295929849,-0.1828003526,-0.2669845521,-0.1620899886,-0.0323722549,0.0232063718,-0.0998212025,0.1968754828,-0.277266562,-0.096666567,0.0570118688,-0.0776291639,-0.2127607018,-0.0297438148,-0.3879216313,-0.0617135428,0.3158175647,0.122121118,-0.164080888,-0.0958079994,-0.2297015041,0.1197901368,0.4393487871,0.4340215027,-0.0001147721,0.0066162376,0.189043954,-0.2363326699,-0.1672693491,0.0192164704,0.0775555968,-0.3022915423,-0.1064002961,-0.303617388,-0.0272462368,-0.1061251387,-0.0156819802,-0.2243732959,0.4987519383,0.5414040089,0.1587626338,0.2508126199,-0.2136609852,0.0723969489,0.2544373572,0.0488996431,0.1056831107,-0.2795141339,0.4687843025,-0.192318216,-0.4726884067,0.1478177905,-0.2247736454,-0.2016824782,-0.0950887725,0.0382044166,0.3660891652,0.0315693952,0.2641872764,0.3474211693,0.427207619,-0.0008118788,-0.2896353006,-0.3209854364,-0.1426243484,-0.0843160674,0.1892365068,0.2010775357,0.0598268472,0.1184589118,0.1071274877,-0.1253313571,-0.1519900858,0.0761608556,-0.2748079002,0.3397960365,-0.2012512684,-0.0380430035,0.1555751562,-0.0944791883,0.1961694509,-0.0997752547,0.0034040213,0.2567143738,-0.2454418987,-0.1123952121,-0.0716823488,-0.5194147229,0.1627843231,-0.2759447992,0.1981881857,0.297239095,-0.0876615867,-0.0676169693,0.0863937065,0.4202026427,-0.2524375618,0.2605278194,0.2272561342,0.1640114188,-0.1717065424,-0.0443987511,-0.3305119574,-0.1962455064,0.1015938595,-0.0773416013,-0.0110167162,-0.2546043396,-0.2404217273,0.0585469492,-0.1407701224,-0.0308455452,0.2748161852,0.3899729252,-0.3540612459,0.4294980168,0.1651093364,0.0523289666,-0.4855332375,-0.3365262449,-0.1744022071,0.1532574743,-0.2032443881,0.0389759205,0.1297923028,0.1396338344,0.0596543029,-0.0428832695,-0.0625896901,-0.4402190149,-0.0094448142,-0.0241036415,0.0207530651,-0.0425145514,0.0630237088,-0.0233655572,0.2347659767,-0.1249946058,-0.0253819562,0.3477917612,-0.2762119472,-0.2276407033,0.0509794205,0.2101766467,-0.0282483641,0.2091549635,0.0285065565,0.0768568069,0.5359784961,0.006653544,-0.0471478477,-0.3782060444,0.0945213884,0.0920173824,0.1619912833,0.1766304076,-0.0425883308,-0.0726022571,0.006389291,0.1035868526,0.255040735,0.3907330036,-0.2264698744,0.2906736135,0.0070044789,-0.0566558652,0.5319920182,-0.2620989978,-0.4484441876,0.2338527292,-0.3511554301,0.1081820279,0.2206698805,0.175902918,0.1138889194,-0.1089850739,0.2690151632,0.3656851351,0.0705158114,0.1209139228,-0.2366241813,-0.0922284573,-0.1715158075,0.4842026532,0.3627241552,-0.0255860128,-0.0730893835,0.3563797474,0.1833985001,-0.1992235482,0.1312804222,0.3693413436,-0.1744247079,-0.1261581182,0.0944813043,0.0655976534,-0.3000326455,-0.0709873512,-0.0908445567,0.0828670263,-0.1136056483,-0.0894262344,0.2825580239,0.0228704214,-0.0766655654,0.1131800264,0.0379414111,-0.1318219602,0.0248664208,-0.0704168379,0.1261966377,0.015421493,-0.0932066888,0.0873605087,-0.2184161246,-0.1239406168,0.0018026204,-0.1370613873,0.0769501179,0.1752732992,-0.0511878207,0.0384011976,-0.1730399132,0.1069574431,-0.1529778987,-0.2987073362,-0.5071573257,0.4358929098,0.1864788234,-0.338109225,0.3402600288,0.1040275991,-0.0295994803,-0.1062710211,-0.1198950335,0.5749334097,0.2730710804,0.0196176525,0.2990927994,-0.2454112023,0.2542806268,-0.2662391067,-0.1036156565,-0.0977647603,0.4209229648,0.0310557056,0.235062778,0.1482800692,-0.086945802,-0.232059598,0.4703815281,-0.1788095683,-0.0138725368,0.2665261328,-0.4247210622,-0.0115634901,-0.3480211198,-0.2494782507,0.0067336266,0.1732289195,-0.0944131836,-0.0554352179,0.0085378792,-0.2627930939,0.0078365039,0.1532371342,-0.0718523338,0.3198345006,0.1165149361,0.100595057,-0.1228921041,-0.1794070601,0.1028474271,0.178289935,-0.2751909494,0.1844432205,-0.0228225887,-0.0134467958,-0.2862635553,0.1904245317,-0.1268011034,0.0388152041,-0.0647433475,0.2586888969,-0.0779031143,0.1019119397,0.3001608551,0.3549227417,0.4629625678,-0.0780947357,0.3263479173,-0.082663402,-0.1796957552,0.0329435281,0.3345077932,-0.317992866,-0.0248905141,0.0053776167,0.0511768535,-0.098929964,-0.1660345495,0.0946889892,0.1968243569,-0.1661894023,-0.2573056817,-0.0721759722,0.1776906103,-0.0424321033,0.1197286993,-0.2324412018,0.0084075676,0.1450572908,-0.0062896051,-0.1843234897,0.0404453017,-0.119922094,-0.0470511988,-0.4116414189,-0.0535182059,0.107050769,0.136731714,0.359367758,0.1067891642,0.1944518089,0.2884452343,0.1585920751,0.0063675619,-0.0885991082,0.5141721964,-0.0175727438,-0.2971983254,-0.2908624113,-0.2397543192,0.0251130797,0.3155242205,-0.684260428,0.2337827832,-0.1138519347,-0.0940270722,-0.3955195546,0.0007256473,0.2522048652,-0.0761487111,0.0163396969,0.0954742506,0.2302770168,0.1546791047,-0.0387874991,0.385794431,-0.4130844176,0.3594082594,-0.1354581416,0.4855529368,-0.1221182346,-0.2300206572,0.0988747329,0.1504482478,-0.017115429,-0.0218022279,-0.0139687397,-0.2593793571,-0.0582376383,0.0227314793,0.5197216272,0.0168902613,-0.1655895859,-0.0158990715,0.0681640953,0.2563197017,-0.2606988549,0.490527451,-0.1907790899,0.018416198,-0.1996744126,0.0187458191,-0.2049811631,-0.2150857002,-0.1641348153,0.0934442803,-0.0303826444,-0.2240440547,-0.2116939425,0.1031402349,-0.3681346178,-0.043764852,0.1179902107,0.2293321937,0.1807704866,-0.1961438209,-0.0862887725,-0.2018710524,0.3574143946,0.1934529245,0.1124654561,0.1632591337,-0.0568783842,-0.2155374587,-0.1774520874,-0.2904544473,0.1419077367,0.650510788,0.6102463007,-0.216273874,-0.1585991383,0.0791451484,0.0026032848,-0.1652733237,-0.5487573743,-0.1285093129,-0.1654996127,-0.4410512149,0.0046065282,0.3584984541,0.3558538258,-0.2264457792,-0.0318755843,0.1387303919,-0.0007017415,0.3717511892,-0.1297639608,0.0150538199,-0.0664106682,0.2258288562,-0.1074168459,0.3673560321,0.2488897145,0.3563636839,0.1917262524,-0.3175574243,0.0706881359,-0.0298882648,0.2314307243,0.1151607484,0.1102282628,0.3546277285,0.1289709508,0.3322559297,-0.1909663528,0.6737229824,-0.0073443423,0.2068272829,-0.8374912143,-0.2673703432,0.3314560056,-0.0263579693,0.1994025856,0.2157180607,-0.4038519263,-0.3216699064,0.3766611814,0.0990833119,0.7675241232,-0.2972695529,0.396258086,-0.155298993,0.6611962318,0.1280306876,-0.3813102245,0.1585618109,-0.3987914026,-0.2663660944,0.2130050957,-0.0923117474,0.0365400203,0.1605038345,-0.2905771136,-0.0298116598,0.0862518772,0.2249330282,0.1306696385,0.5134937763,0.1209264472,-0.6006979346,-0.0828128979,0.0265692659,-0.0234002154,-0.1049021482,-0.0492791906,0.079946138,-0.0313266218,-0.2085642517,-0.0284665916,0.2052579373,-0.2474462092,0.0866156295,0.06375245,-0.0720531717,0.5367996693,0.1062301323,-0.0013605307,0.49889642,-0.0392867327,0.1321583837,-0.2274818122,-0.1652864814,0.2007082999,-0.0658492595,0.4217032194,0.0883790925,-0.0005881447,0.1141034439,0.063238509,-0.0466907471,0.0947634056,-0.3643093109,0.1054972857,-0.411231488,-0.2154277265,-0.0884305537,-0.3007826209,-0.100921981,0.1231095418,0.1373571604,-0.4644461274,0.1151452288,0.2652110159,-0.3750090599,0.0119970199,0.3838246465,0.1710590273,-0.1051386297,0.4998426735,0.4575876594,-0.0533619262,-0.2644975483,-0.1096294746,0.2944481373,-0.2731431425,0.0849164277,-0.2455589622,-0.3016007841,0.2093510926,0.0808756202,0.1000993326,-0.0149996774,-0.0632624626,-0.2510291636,-0.6380499601,-0.1814441979,-0.0027232091,-0.0035483204,0.1460837424,0.2738984525,-0.3100468516,0.5892773271,-0.2788558602,-0.0529421978,-0.2076086402,0.2782232165,-0.0555626303,-0.3855466247,0.2115356922,0.2033523023,0.1001041383,0.0805966482,0.0059094876,-0.2213536203,-0.1435203254,0.1127153859,0.236469999,-0.4196631312,-0.0159937777,-0.3129970431,-0.0924898982,-0.2700214982,0.0660679936,0.0170283876,0.0139454231,-0.2261790633,0.0093937991,0.0475260615,-0.2552624047,0.0420600101,-0.2545854449,0.1250910759,0.2465714067,0.1331033409,0.1178303957,0.0006370013,-0.3931023479,0.1796457469,0.6661381721,-0.0175324269,0.4190826416,-0.4357320666,-0.0715063885,0.2476596385,0.0498032048,0.0480659828,-0.3416773379,-0.0107185869,0.2935882509,0.2268190682,0.0152084939,-0.0147332903,0.2263547182,-0.4112985432,0.130148828,0.3453423381,-0.1119267642,-0.0674757361,0.1559655368,0.0913231447,0.5066782236,0.1149654835,0.0057114391,0.2489494383,-0.0012200999,0.1548569202,0.4856923223,0.2611194253,0.2386804968,0.3016282618,-0.0326663293,-0.0406999774,-0.2312842757,0.2294140607,-0.4356164038,-0.4151665866,-0.2674567699,0.5044153333,-0.0193237979,0.124209106,-0.3350303471,-0.0662413165,-0.0406535454,0.2541331649,-0.1192156971,0.2744085193,-0.4530942142,-0.1705989689,-0.0894243494,-0.1239616126,0.0115343612,0.1896333098,-0.0558579676,0.2722764611,-0.5796657801,0.229385525,0.0521362387,0.1451500803,-0.1346782744,0.1227787659,0.105501391,-0.5078569055,0.3433353007,0.2378561646,0.2918595076,0.1154054627,-0.2219066918,0.344370693,0.4366977513,-0.1908068657,-0.0406118706,-0.1209843084,-0.0356199592,-0.1499128789,0.1974353343,0.3490365744,0.1878990531,0.2596163452,0.0557278134,-0.2659661174,0.2785175145,0.0650138855,-0.3240590692,0.1786662638,-0.1329384595,0.4019323587,-0.3285809159,-0.4429644048,0.0098667834,-0.3941245675,0.2155870497,0.6670193672,-0.295422256,0.2878552973,-0.0254881084,0.0485503562,0.0018405904,0.4760500193,0.4697352946,0.2841595113,-0.3783710301,-0.0256672148,-0.3288835883,0.27655828,-0.2024301887,0.0669974238,0.0222478509,0.1707943231,0.1645782143,-0.0033026312,-0.0796228647,-0.0136218695,0.2188030481,-0.1950692683,-0.1675594449,-0.1280473471,0.0020945771,0.3573177457,-0.0259007961,-0.2975497842,0.1588654965,0.1150629818,0.0991567299,-0.3214533329,0.3626995385,-0.1984633952,-0.0203625746,0.0691283867,-0.1504709721,0.3792094588,-0.2204384655,-0.0330468602,-0.081827566,0.1281419247,-0.167077139,-0.2210000008,0.3402221799,0.3884244263,-0.1076363996,0.1082137674,-0.2551924884,-0.0864620805,0.0403044634,-0.3696909547,-0.3087159991,-0.0074877236,-0.0002531401,0.0085246228,0.1030979007,0.2458305508,0.1053371057,-0.0278402809,-0.2292290181,-0.1855800301,0.2094752491,-0.2946499288,-0.3130412102,-0.0353801101,0.0556249134,-0.1636290848,0.2511020303,-0.4624070823,0.1457487643,0.2297440767,-0.0255143084,-0.4540733695,0.2232973725,0.2540613115,0.0164721701,-0.2076968998,0.0534571446,0.1526236087,-0.1369921863,-0.2016399801,-0.1757193357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"Seems like it works when setting ```block_size=2100000000``` or something arbitrarily large though.","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":12,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nSeems like it works when setting ```block_size=2100000000``` or something arbitrarily large though.","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"Can you give us some stats on the data files you use as inputs?","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":14,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nCan you give us some stats on the data files you use as inputs?","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"Basically ~600MB txt files(UTF-8) * 59. \r\ncontents like ```\uc548\ub155\ud558\uc138\uc694, \uc774\uac83\uc740 \uc608\uc81c\ub85c \ud55c\ubc88 \ub9d0\ud574\ubcf4\ub294 \ud14d\uc2a4\ud2b8\uc785\ub2c8\ub2e4. \uadf8\ub0e5 \uc774\ub807\ub2e4\uace0\uc694.<|endoftext|>\\n```\r\n\r\nAlso, it gets stuck for a loooong time at ```Testing the mapped function outputs```, for more than 12 hours(currently ongoing)","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":36,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nBasically ~600MB txt files(UTF-8) * 59. \r\ncontents like ```\uc548\ub155\ud558\uc138\uc694, \uc774\uac83\uc740 \uc608\uc81c\ub85c \ud55c\ubc88 \ub9d0\ud574\ubcf4\ub294 \ud14d\uc2a4\ud2b8\uc785\ub2c8\ub2e4. \uadf8\ub0e5 \uc774\ub807\ub2e4\uace0\uc694.<|endoftext|>\\n```\r\n\r\nAlso, it gets stuck for a loooong time at ```Testing the mapped function outputs```, for more than 12 hours(currently ongoing)","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"It gets stuck while doing `.map()` ? Are you using multiprocessing ?\r\nIf you could provide a code snippet it could be very useful","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":24,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nIt gets stuck while doing `.map()` ? Are you using multiprocessing ?\r\nIf you could provide a code snippet it could be very useful","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"From transformers\/examples\/language-modeling\/run-language-modeling.py :\r\n```\r\ndef get_dataset(\r\n args: DataTrainingArguments,\r\n tokenizer: PreTrainedTokenizer,\r\n evaluate: bool = False,\r\n cache_dir: Optional[str] = None,\r\n):\r\n file_path = args.eval_data_file if evaluate else args.train_data_file\r\n if True:\r\n dataset = load_dataset(\"text\", data_files=glob.glob(file_path), split='train', use_threads=True, \r\n ignore_verifications=True, save_infos=True, block_size=104857600)\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n return dataset\r\n if args.line_by_line:\r\n return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\n else:\r\n return TextDataset(\r\n tokenizer=tokenizer,\r\n file_path=file_path,\r\n block_size=args.block_size,\r\n overwrite_cache=args.overwrite_cache,\r\n cache_dir=cache_dir,\r\n )\r\n```\r\n\r\nNo, I'm not using multiprocessing.","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":71,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nFrom transformers\/examples\/language-modeling\/run-language-modeling.py :\r\n```\r\ndef get_dataset(\r\n args: DataTrainingArguments,\r\n tokenizer: PreTrainedTokenizer,\r\n evaluate: bool = False,\r\n cache_dir: Optional[str] = None,\r\n):\r\n file_path = args.eval_data_file if evaluate else args.train_data_file\r\n if True:\r\n dataset = load_dataset(\"text\", data_files=glob.glob(file_path), split='train', use_threads=True, \r\n ignore_verifications=True, save_infos=True, block_size=104857600)\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n return dataset\r\n if args.line_by_line:\r\n return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\n else:\r\n return TextDataset(\r\n tokenizer=tokenizer,\r\n file_path=file_path,\r\n block_size=args.block_size,\r\n overwrite_cache=args.overwrite_cache,\r\n cache_dir=cache_dir,\r\n )\r\n```\r\n\r\nNo, I'm not using multiprocessing.","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"I am not able to reproduce on my side :\/\r\n\r\nCould you send the version of `datasets` and `pyarrow` you're using ?\r\nCould you try to update the lib and try again ?\r\nOr do you think you could try to reproduce it on google colab ?","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":47,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nI am not able to reproduce on my side :\/\r\n\r\nCould you send the version of `datasets` and `pyarrow` you're using ?\r\nCould you try to update the lib and try again ?\r\nOr do you think you could try to reproduce it on google colab ?","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"Huh, weird. It's fixed on my side too.\r\nBut now ```Caching processed dataset``` is taking forever - how can I disable it? Any flags?","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":24,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nHuh, weird. It's fixed on my side too.\r\nBut now ```Caching processed dataset``` is taking forever - how can I disable it? Any flags?","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"Right after `Caching processed dataset`, your function is applied to the dataset and there's a progress bar that shows how much time is left. How much time does it take for you ?\r\n\r\nAlso caching isn't supposed to slow down your processing. But if you still want to disable it you can do `.map(..., load_from_cache_file=False)`","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":55,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nRight after `Caching processed dataset`, your function is applied to the dataset and there's a progress bar that shows how much time is left. How much time does it take for you ?\r\n\r\nAlso caching isn't supposed to slow down your processing. But if you still want to disable it you can do `.map(..., load_from_cache_file=False)`","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"Ah, it\u2019s much faster now(Takes around 15~20min). \r\nBTW, any way to set default tensor output as plain tensors with distributed training? The ragged tensors are incompatible with tpustrategy :(","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":29,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\nAh, it\u2019s much faster now(Takes around 15~20min). \r\nBTW, any way to set default tensor output as plain tensors with distributed training? The ragged tensors are incompatible with tpustrategy :(","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":"> Ah, it\u2019s much faster now(Takes around 15~20min).\r\n\r\nGlad to see that it's faster now. What did you change exactly ?\r\n\r\n> BTW, any way to set default tensor output as plain tensors with distributed training? The ragged tensors are incompatible with tpustrategy :(\r\n\r\nOh I didn't know about that. Feel free to open an issue to mention that.\r\nI guess what you can do for now is set the dataset format to numpy instead of tensorflow, and use a wrapper of the dataset that converts the numpy arrays to tf tensors.\r\n\r\n","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":92,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\n> Ah, it\u2019s much faster now(Takes around 15~20min).\r\n\r\nGlad to see that it's faster now. What did you change exactly ?\r\n\r\n> BTW, any way to set default tensor output as plain tensors with distributed training? The ragged tensors are incompatible with tpustrategy :(\r\n\r\nOh I didn't know about that. Feel free to open an issue to mention that.\r\nI guess what you can do for now is set the dataset format to numpy instead of tensorflow, and use a wrapper of the dataset that converts the numpy arrays to tf tensors.\r\n\r\n","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/630","title":"Text dataset not working with large files","comments":">>> Glad to see that it's faster now. What did you change exactly ?\r\nI don't know, it just worked...? Sorry I couldn't be more helpful.\r\n\r\nSetting with numpy array is a great idea! Thanks.","body":"```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.","comment_length":35,"text":"Text dataset not working with large files\n```\r\nTraceback (most recent call last):\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 333, in <module>\r\n main()\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 262, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"examples\/language-modeling\/run_language_modeling.py\", line 144, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split='train+test')\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 469, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/ksjae\/.local\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/ksjae\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 104, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\n```\r\n\r\n**pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**\r\n\r\nIt gives the same message for both 200MB, 10GB .tx files but not for 700MB file.\r\nCan't upload due to size & copyright problem. sorry.\n>>> Glad to see that it's faster now. What did you change exactly ?\r\nI don't know, it just worked...? Sorry I couldn't be more helpful.\r\n\r\nSetting with numpy array is a great idea! Thanks.","embeddings":[-0.4925667346,-0.2310241014,-0.1198643222,0.2836014926,0.4663560987,-0.0735016093,0.3053681254,0.5961018801,-0.1138256192,0.0461653806,-0.0624196492,-0.0304052103,-0.1033419743,0.310179472,-0.1062136963,-0.0389501564,-0.2278844863,0.1130641699,-0.0946916267,0.0499166101,-0.112550877,0.1110998988,-0.1209788397,-0.0956026316,-0.4804550409,-0.0829951093,0.1185314655,0.2535659075,-0.2920109332,-0.3053144217,-0.1657067835,0.0965852961,0.1122046039,0.6015893817,-0.0001037621,0.129462406,0.2131118327,-0.1573347151,-0.2265827805,-0.0314556658,0.1559481472,-0.4493688941,-0.1823892593,-0.2179644257,0.0261511691,-0.0627441034,-0.1082879007,-0.117559962,0.265335381,0.5598000884,0.3412032723,0.170758754,0.1321692318,-0.1129611656,0.2507540286,-0.0872494131,-0.0887443796,0.0848500654,0.3108180165,0.0265774522,-0.2564010024,0.3199440241,0.0101718027,0.0377005637,0.0003141626,-0.0240623355,-0.0827939883,-0.2940048873,0.3711950183,0.2321464419,0.5224136114,-0.3169358075,-0.1641277969,-0.3288432062,-0.1824413538,-0.3512556851,0.226632759,0.1845538467,-0.1113308817,0.1161273718,-0.1436178535,-0.1014853194,-0.2063972354,-0.0293161273,-0.2615340352,0.1022737771,-0.2105868608,-0.1004619375,0.2575800121,-0.2702479959,0.0108298594,-0.1056223363,-0.1307831109,-0.0068054325,-0.247213304,-0.0840224698,0.2468504906,-0.199078232,0.2889420688,-0.0489982776,-0.0504212566,0.224761039,-0.0796672255,-0.0302130375,0.0054934151,0.3368757069,0.0041132052,0.0638572946,0.4085853696,0.2356281877,-0.379357934,-0.2320603579,-0.149626568,-0.5145174861,-0.0293558314,-0.1177171171,0.01687208,-0.0452369191,-0.1210255995,0.1521080732,0.1642943472,0.1547948122,0.10596627,0.4816172421,0.1049144343,0.2937650084,-0.1649460793,0.1770391911,-0.144326508,-0.1722711623,-0.2492276877,-0.0456115939,-0.0579169057,-0.0926888585,0.2676637173,0.1338510513,0.2967094481,-0.1395806223,0.1908528805,-0.1090161651,0.1168141365,-0.5029688478,0.1476960927,0.1410785764,-0.0354773737,0.1686936468,0.2287830412,-0.1788699031,-0.1023910046,0.050235346,-0.1035604551,-0.2971284389,0.1772404164,0.3213210106,0.1347744316,-0.0030032352,-0.018904794,0.2770770788,0.375772059,-0.2176397294,0.0281005073,-0.177321583,-0.2569280863,-0.125802964,0.2650002837,0.4393820763,-0.425060302,0.2090140134,0.0690165609,0.1223087683,0.0890245587,0.3739468455,-0.0650732517,0.3143593669,-0.1145715341,0.2009439021,0.3094802201,-0.285585016,-0.5090754032,0.5406064391,-0.1859303266,-0.1181871668,0.099257268,-0.0444510505,0.1034918204,0.1358725727,0.3047401011,0.2793403268,0.0113194669,0.2667300403,-0.2604504526,-0.0832502395,-0.0501141697,0.1559398472,0.3121375144,-0.2080699205,0.1951689571,-0.017231321,0.19437626,-0.044446256,0.2128829956,0.5059319735,0.1271578521,-0.0078406455,-0.0123689231,-0.2239847332,0.0316570215,0.1751381308,0.1210090369,0.0843074545,-0.2666850388,-0.0822329894,-0.3092454672,-0.1341452003,-0.2703467607,-0.193386212,0.2836054265,0.0813239738,-0.0114756655,0.2551389933,-0.0167013835,0.0253130831,-0.1602133662,-0.0655840337,-0.0120967701,-0.0018739073,-0.091083996,-0.2782087028,0.1229490936,0.1102730632,-0.041142635,-0.0616680495,-0.1109172329,0.3928830624,-0.0020204084,-0.0810614601,-0.0360698178,-0.0300612058,0.1020015851,-0.0952062309,0.0719461367,0.0007549835,0.1388051212,-0.1053132713,-0.1801620126,0.3006142974,-0.1398149133,0.2031631023,0.3438572288,-0.2090742588,0.2556621134,-0.0663511157,0.2276659161,-0.1381171942,0.2406677008,0.0831321999,0.136321649,-0.0013489889,-0.2719839215,-0.0127429022,0.6895840764,0.171065256,0.0928926244,0.1747953147,-0.2453399599,-0.0179574843,-0.2034236491,0.1299691498,0.3976373672,0.2744440138,0.2679811418,0.0357754268,0.1870050132,-0.3740030229,0.1397108883,0.0032486171,0.0111852335,0.5398927927,0.2758919895,-0.1295805275,-0.4534206688,-0.1563689411,0.1057760417,0.3134025633,-0.0895873681,0.0198824629,-0.2155629992,-0.2920289338,-0.2959762812,0.2103271335,-0.2627325952,-0.2253175527,0.0694883317,0.0018545482,-0.0096535487,0.0968664065,0.0594205372,0.1454249322,0.3997489512,-0.198824957,-0.0615104102,-0.2715903819,-0.1730497926,0.1567441374,0.2421877533,-0.0123151606,0.106642887,-0.1534677446,-0.1015961543,0.0038625724,-0.2853763402,0.0006907425,-0.029190002,0.1006606892,-0.0571642891,0.2626269162,0.1662754267,0.0769828781,0.312800765,-0.1297491789,-0.080774948,0.2233646214,0.0906755924,-0.0652652159,-0.1705733836,-0.4048942327,-0.1530372351,-0.6100059748,0.357511282,0.0955663547,0.0380034447,0.4022583663,0.3181683421,0.2154984921,-0.0212013293,0.2124930322,0.1002640873,-0.1950187981,0.2925721705,-0.2248299271,-0.3832399845,-0.0998249501,0.1606033593,0.16261594,0.007262093,-0.5234432817,0.0302410219,-0.4368826151,0.1623679101,-0.0637961328,0.168691963,0.0814080685,-0.0684886277,-0.2522610426,-0.0842753425,0.0830210894,0.0467453748,-0.0724342912,0.0489535592,-0.1854182333,0.4121846557,0.2402956635,0.4626801014,0.163782239,0.0814501196,0.3938990235,-0.121706374,0.4271225929,-0.3532859087,-0.3265112638,0.1239757761,-0.0912457034,-0.032902278,0.2716791928,0.0862800479,0.2699357569,-0.0800968334,-0.2192074358,0.0848959759,-0.27935341,0.202010408,-0.1586163789,0.1581823081,-0.1357393563,-0.0241032466,-0.0767385736,-0.1311157495,-0.0338697657,0.1800867468,-0.240611285,-0.0250576194,-0.3433842361,0.064881891,-0.4402942657,0.1162606105,-0.0008667396,0.3317792416,-0.1660775989,-0.1548693776,-0.1084446535,-0.1574716717,0.5587426424,-0.0700288638,0.050007239,0.0628436133,-0.0919483751,-0.3191085458,0.1609371305,-0.2735011876,0.1588300765,0.364718169,0.5645071268,-0.3012343347,-0.0509777814,-0.0399969779,0.3442745507,0.0167832561,-0.0334483795,-0.2761992514,-0.1711502522,-0.4797545075,0.1388336718,0.297947377,0.2846294641,-0.2693778574,0.0718105733,0.0067696143,-0.241671294,0.2142870724,0.0040949988,0.1445522904,-0.0965572298,0.2255922556,-0.1108148322,0.1980494261,0.1853599399,0.5712109208,-0.0331981257,-0.1837669611,-0.1003368199,-0.1977577955,0.3502214253,0.3210826814,-0.1226084605,0.2475554943,-0.0468919948,0.2047809064,0.0162528623,0.371440649,0.2987754643,-0.0353546627,-0.2595382929,-0.2792066634,0.1115558594,-0.084549889,0.1867009103,0.1408402324,-0.3193204403,-0.2557853162,0.0803829804,-0.3155545294,0.5819581151,0.0936629996,0.3404352963,-0.0566671602,0.1961902976,0.062907733,-0.529538393,0.2434755117,-0.3393689096,-0.268777281,0.0840474367,-0.0019825432,0.0826099217,0.1538453996,-0.2705694437,0.0997356027,0.1199411377,0.1092551425,-0.2874216139,0.3836100101,-0.4040075839,-0.120232746,-0.3170886934,0.1808964461,0.038462162,-0.1584917307,-0.0123262499,-0.1195177808,0.0203617811,-0.2523366511,-0.1452478021,-0.0493154302,-0.4198040366,0.1822546721,0.039505858,-0.3351989388,0.2435080856,0.177771166,0.0717558861,0.4507853389,-0.2375138253,0.1786590219,0.0119797494,-0.027082134,0.1566385478,-0.0020522338,0.3159705698,-0.0124907307,-0.2267416567,0.0910502449,-0.0267283469,0.1002250686,-0.0211739652,0.0054354095,-0.0417032167,-0.3602178097,-0.2473487854,-0.1221575066,-0.2448607832,-0.300704211,0.2255004197,-0.0533379354,-0.0544013232,0.168014273,0.0709897727,-0.3310471773,-0.1633546054,0.3258528709,-0.1954346448,0.178970322,0.5283616185,0.3944674134,-0.2252093256,-0.3289530873,0.3111713529,-0.0288908929,-0.3540940285,0.2759122849,0.0270211343,-0.0696801096,0.1265555173,0.2427475005,-0.1405765116,-0.3056271076,-0.1277884543,-0.238108933,-0.4304110408,0.0736415237,-0.014685858,0.1428983808,-0.0120183369,0.2134833634,0.1132786572,0.2139737904,-0.4020756781,0.1423479021,-0.1771862209,0.0996010154,0.0931536853,-0.1355485916,0.1392203867,0.0518148281,0.1112697497,0.2546557784,-0.0731492043,-0.3498377502,-0.1104320139,0.0802936926,0.0429649539,-0.1213981062,-0.0186528079,-0.3757633269,-0.2571214437,-0.2619501352,0.2055384666,-0.0595564842,-0.1060271189,0.1788904071,-0.1484446824,-0.0594712421,0.1162799075,-0.0307372585,-0.2113559395,0.1135027334,0.1723376662,0.0354118906,0.0073987548,0.001316295,-0.3159472942,0.0826841965,0.1583467573,-0.0741560161,0.324462533,-0.3757270277,-0.0566546358,0.1769374013,0.3609242737,0.4115839303,-0.2069198936,-0.0283430908,0.2429047376,0.2981537879,-0.2532995939,-0.100169003,0.0949342698,-0.2748050988,0.3003486395,0.1226505339,0.1055465266,-0.0864750743,-0.1073918417,0.2233810425,0.1748348027,-0.0378284678,-0.1495874822,0.3407848477,0.1178208962,-0.0499308445,0.3016181886,0.1821387112,0.3013967574,0.7971208692,0.0785115361,0.2538471222,-0.2110732347,-0.0938013792,-0.1365755349,-0.4474599957,0.0228371248,-0.0801561773,-0.0352986865,-0.1246766895,-0.2056946158,0.2049510926,-0.0154314805,0.2201544791,-0.1490969658,0.0436850823,-0.2486523837,-0.1612087488,0.0436246097,-0.3107736111,-0.0689439625,0.1061923951,-0.1068829075,0.3263956606,0.1242942289,0.062380258,0.1022240296,-0.4294129908,0.2035566121,-0.0626898557,0.0876704901,-0.3625392616,0.2839859724,0.2613130808,0.1896336377,0.1571779102,0.1909476668,0.5729644299,0.3944715559,-0.1363311261,-0.080993779,-0.275598973,-0.1103982478,-0.4154100716,0.2859199941,0.1979573816,0.355381757,0.4643360376,0.2075484693,-0.2680899799,-0.0660790429,0.1920687854,-0.2637651861,-0.2234546393,0.1047905385,0.0679993033,-0.2863512635,-0.3225553632,-0.0183769595,-0.4338229597,-0.0045649572,0.5934925675,-0.1135881692,0.3091260493,-0.3395166695,0.1448054612,0.005200055,0.6057991982,0.2953888476,-0.0178112928,-0.2954342961,-0.0195158366,-0.6181394458,0.126281172,-0.1511759907,0.0123337964,-0.0116538377,0.1425365657,0.1738922894,-0.0330511332,-0.0051008463,0.0362971015,-0.1498092115,-0.1097010076,-0.3313085139,-0.2676608562,-0.2250265926,0.1569494456,-0.0014627598,-0.3760662675,0.1550827771,-0.0955090672,0.2300336659,-0.179386884,0.0802163035,-0.020627711,-0.0355981886,0.1578445137,0.0340668336,0.3657603562,-0.2838610113,0.034045469,-0.1346201152,-0.0631753951,-0.1700670123,0.1941984594,0.2527386546,0.2654085457,-0.0962216258,-0.1707080156,-0.1380879581,0.1567032486,0.0587180555,0.037135195,0.1429637223,-0.0507429838,0.1262889504,0.0587171726,0.186367318,0.402974546,-0.1113762707,-0.0343026035,-0.3506342769,-0.5414934754,0.2185377181,-0.1180814579,-0.5158481002,0.0103078466,0.1300402284,0.1811083704,0.0267859064,-0.5836286545,0.3067767322,0.3096933365,0.1339992583,-0.3529006541,0.2580336928,-0.0505167134,0.0583431311,0.0119166169,0.1201330498,0.0242642723,-0.2702246904,-0.2056571245,-0.0322016366]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/629","title":"straddling object straddles two block boundaries","comments":"sorry it's an apache arrow issue.","body":"I am trying to read json data (it's an array with lots of dictionaries) and getting block boundaries issue as below : \r\n\r\nI tried calling read_json with readOptions but no luck .\r\n\r\n```\r\ntable = json.read_json(fn)\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"pyarrow\/_json.pyx\", line 246, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\n","comment_length":6,"text":"straddling object straddles two block boundaries\nI am trying to read json data (it's an array with lots of dictionaries) and getting block boundaries issue as below : \r\n\r\nI tried calling read_json with readOptions but no luck .\r\n\r\n```\r\ntable = json.read_json(fn)\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"pyarrow\/_json.pyx\", line 246, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\n\nsorry it's an apache arrow issue.","embeddings":[0.1221939102,-0.3380404115,-0.0981453061,0.5065849423,0.1361984611,-0.2670111656,-0.0138538983,0.4272662401,0.1391404122,-0.0226568021,0.2115778327,-0.0096409768,0.0531540029,-0.1932379156,-0.2861743271,-0.1781947017,-0.0429025553,0.0553168952,-0.062424019,0.163351804,-0.0237407479,0.1139382496,-0.1216934472,0.0911220014,0.3017581105,-0.1788771898,0.2082627565,-0.2800849378,-0.1157643199,-0.3911887109,0.1221905574,-0.1124015376,-0.142941609,0.2266660333,-0.0001247014,0.2448250949,0.1467627287,0.1846411824,0.0665711164,-0.1734286696,0.1785063148,-0.4670132995,0.24858284,-0.1834782958,0.0052145626,-0.6058434844,-0.3461125493,-0.0116603067,0.644257009,-0.1515834779,0.1053742245,-0.040877413,0.2541484535,0.7045570612,0.4632484019,0.1338707954,0.0207130145,0.2492012233,0.1725826114,0.1587036252,-0.186409235,-0.1618608981,-0.4059207737,0.120592989,-0.0443733595,-0.114846386,0.3264867067,0.0720032305,-0.0267124213,-0.0290750749,0.3404971361,-0.278259784,-0.0218001232,-0.0638636351,-0.1487136781,-0.0110273566,0.254838407,0.2484577745,-0.1102321669,0.1472475082,-0.0743856952,-0.0118273273,-0.35604617,0.3006097078,-0.0616173334,-0.0531551465,-0.0479199141,0.3025810719,0.1331458539,0.0405350886,0.3378864229,-0.0305671785,-0.2320637852,-0.1323006302,0.1542411745,-0.0163627826,-0.2203357071,-0.4461995661,0.2417398393,0.4176611602,0.3182554245,0.2934222817,0.0706386864,0.261790514,0.2064972371,-0.1980179995,0.125089705,0.1089444458,0.1624760926,0.2340981215,-0.0339395739,-0.0846276879,-0.011175504,-0.2418877631,0.027578529,-0.2664399147,0.3325451612,-0.3042002916,0.0047535156,0.0104778809,-1.0641465187,-0.1417254359,0.210038662,0.3148886859,0.0824864507,0.2365998179,0.2080058008,0.2767766416,0.1175692677,-0.208701551,0.034609437,-0.1284563392,0.0733916312,-0.0163344666,0.0706358701,0.3396988213,0.056065917,0.1893585771,0.0397035331,-0.4833970964,0.3182773888,-0.0891383588,0.1428472996,0.2726528347,0.171628207,-0.0133777289,-0.0868284181,-0.0046060551,-0.119203873,0.4557493925,-0.0606020913,0.1302318871,0.0755586699,0.0405135676,-0.16569148,0.4400503039,-0.0089537166,0.0385849886,0.3115188479,-0.3115566373,0.3006185591,0.061814744,0.0873002782,-0.4220349491,0.1002830714,-0.001686023,-0.4721928239,0.0587776415,0.24405469,0.1932098269,-0.0128270499,0.3430970609,-0.4101160765,-0.1552822143,-0.0651753098,-0.1131211817,0.2449532896,0.1813360602,-0.6026735306,0.1742914617,-0.0081336861,0.0093113659,0.1898839027,-0.0921245292,-0.0599458665,0.3408872485,-0.4663513303,0.2096236646,0.2477592081,-0.0099139409,-0.1577196419,0.0176545922,0.0931576714,-0.0727477521,-0.174492076,-0.4713683724,-0.062884897,-0.2623298764,0.1169764847,0.0942647457,0.3521212339,0.2784488201,0.0972624272,-0.1680398285,0.1321892291,-0.1653727889,-0.3835127056,-0.2485887706,-0.0693919286,-0.2169517279,-0.2988997102,0.0509469882,0.0694501549,0.2481665611,0.2258688509,0.1906953752,-0.0024215407,-0.0336212516,-0.0967527181,0.0357312039,0.0523115285,-0.1903065294,0.1146264896,0.2554110587,0.2228884995,0.5211620331,0.0738390833,-0.2304104865,-0.1845944077,0.1009224653,0.044892095,-0.0933890566,-0.105387114,0.1694660932,0.2695616782,-0.0613567494,-0.3427221179,0.0985910594,0.3854250908,-0.5530593991,0.1836281568,0.2986968756,0.1691129059,-0.2456722111,0.4044406414,0.5155094266,0.069604747,0.3588582873,-0.0732019618,-0.0779307038,-0.0176457614,-0.0648653954,-0.0558128431,0.0544279963,0.1490004063,0.4698218703,0.1728171557,-0.0089819422,-0.0361804962,0.0393672548,0.2482037842,-0.0796886683,-0.003262226,0.0618659519,-0.2461530119,0.0211713109,0.0982249305,-0.0413735732,-0.1085302457,0.207950905,-0.1306905597,0.0206551291,-0.0404692851,-0.2319751233,0.31867522,0.2557952702,0.0995961726,0.0774342641,0.4458798766,-0.239915967,-0.0773090497,-0.2417969853,0.1489981115,0.0632474571,-0.3035200238,0.187656492,-0.3129381239,-0.1466160566,-0.3122433722,-0.6046256423,-0.2564770877,-0.2040445805,-0.1508800536,0.5925183296,-0.1343509257,-0.0617584363,0.0831791535,0.180311963,0.0165884867,-0.010432913,-0.1291449815,-0.2433977276,-0.4313471913,-0.1282159984,0.4605234861,0.4605313838,0.057091672,0.4156341255,-0.0721921474,0.1785669476,-0.1890385598,0.0489960425,0.1157404408,-0.1011858284,-0.2396364957,0.2209964842,-0.2171897441,-0.1484886706,0.1497015506,-0.0297369529,-0.4243770838,0.2746404409,-0.0162073672,0.0346692875,-0.1876357943,0.1049747989,0.1979818195,-0.2337205857,0.3054261804,-0.0478487797,0.2120685726,-0.2157173157,0.1508844048,0.1353906095,0.1169366315,-0.0466811173,0.086959511,-0.2760988772,0.2423702031,-0.0015514338,-0.0227190517,0.245269388,0.0657812431,0.249670893,0.1348278224,-0.4404393733,0.1452090144,-0.0630370155,0.501529336,-0.2962933779,-0.5252681971,0.0639549643,0.3494590819,0.1108310074,0.0036078105,-0.0905916765,0.2919074893,-0.1711132377,0.2567440867,0.0960549116,0.4411299229,0.02595626,0.5720612407,-0.1326151192,0.0949236006,0.2477178276,0.054534778,-0.1643351912,0.2050507367,-0.0343188867,-0.250176698,0.021287635,-0.0559024028,-0.1338121146,0.0815450326,-0.0091672549,-0.006432421,-0.2689267099,-0.3499914408,-0.2205255032,0.2229492217,-0.0905289277,-0.1321556121,-0.413376689,-0.1213752553,-0.0264730714,-0.1451869607,-0.0792261288,-0.0483191572,0.0528997332,0.0103052892,-0.336163491,-0.4017824531,0.2254775614,0.3382680714,0.1010452136,0.1177205443,0.0396181308,-0.0006521379,0.255998224,-0.2732983232,0.5466594696,0.3759915829,0.1308572739,-0.0773963407,0.1328977495,-0.5146712065,-0.1026963368,0.0486877486,0.0311495829,0.2950701416,-0.0916153267,-0.3749934733,0.131488502,-0.1573142409,0.0987787321,0.0212784223,-0.2135723084,-0.2387360483,-0.3544844389,-0.299823463,0.0021503277,0.093649134,0.2110997587,-0.071586661,-0.4759413302,0.4967620373,-0.3534039259,-0.4186767042,0.0939669758,0.2077676356,-0.5599555373,0.0086787464,0.1510336548,0.1923869252,0.6953690648,0.3888013661,0.6559951305,-0.2825002968,-0.1104724705,0.1661477238,0.3288913369,-0.0838135853,0.1850601882,0.655685246,0.0640482679,-0.0591730475,0.4168734252,-0.2864123285,0.160256058,0.0553204156,0.0196712613,0.0451178662,0.3843131661,-0.0135487262,0.1344463974,0.5053291917,0.1057981923,-0.1729654223,0.5434190035,0.0249575879,1.1121329069,0.0095122783,-0.1423879564,0.513671577,-0.8654684424,0.4143802524,-0.3439680934,0.0110997092,-0.4318749011,0.5455959439,-0.115171589,-0.3545550406,0.4453093112,0.051615715,-0.1203936711,-0.2432658821,-0.4183788896,0.0770048574,-0.3075995743,-0.2649407685,-0.190422073,0.2452009022,-0.0847181007,-0.0194384158,-0.4681844115,-0.2433403432,0.0622381829,-0.2504560947,-0.3144944906,-0.3497321904,-0.4209700823,0.0624382757,0.2118715942,0.1973973811,0.2006926239,-0.2233040333,0.0153206671,-0.2899753153,-0.1028229818,-0.1185082272,-0.1754043251,0.1523054093,-0.0172151439,-0.2916840911,-0.1445857286,0.0430825278,-0.0209674966,-0.1501033902,-0.5218304396,0.2110175639,-0.0353178531,-0.1535421014,-0.1214236245,0.0237228051,0.0463080816,-0.0616906285,-0.0046062204,-0.1761158705,-0.0324023776,-0.1733210087,0.0465269126,-0.0962520242,0.0505587049,-0.5239298344,0.3960415423,-0.1638193727,-0.0358224362,0.1596411914,0.4159807563,0.3689696789,0.4184062779,0.3203200698,-0.1436672807,-0.1221150458,0.0282450076,0.4705425501,-0.3247518539,0.0017551805,-0.2252400219,0.0952780172,0.0659024417,0.0316695943,0.1898802221,0.4114486277,-0.1351121664,-0.4633244276,-0.249322474,0.3132324219,-0.2083242536,-0.005179673,-0.0121676782,0.5111253262,-0.0709793419,0.3726396263,-0.1093765125,0.1519524306,-0.0012051489,-0.0394005217,-0.6870328188,-0.193397373,-0.3393852413,0.1177887544,-0.0055272607,0.204425782,0.0755771548,0.0167975761,-0.0270376112,0.1244531199,0.0328277163,-0.0791937038,-0.3169907629,-0.1967821121,0.1188639924,0.1737050414,-0.2998007536,-0.1749807894,0.1304598302,0.0032781998,0.2392769456,0.0147159435,-0.2552894652,0.1036216393,-0.1780906916,-0.022617342,0.0666066855,0.2106164396,0.0550259165,-0.0736266375,-0.3986426592,-0.1101705581,0.2433808148,-0.0502628386,0.4622228146,-0.0536796041,-0.1809711307,-0.2317549288,0.1253713667,0.4037183225,-0.005937451,0.1116328537,-0.0372217111,0.0730807111,-0.0352924876,-0.1740953475,0.1986485571,0.0209514443,-0.2387384623,-0.0865653679,-0.121600844,-0.0765124932,-0.15919438,0.0802661479,0.1241780967,0.0889570042,-0.2377578467,0.2035538703,-0.1462118477,-0.2277026027,0.2171911448,-0.1676716954,0.5600355864,-0.4503755271,-0.0061121555,0.2324189246,0.1827725768,0.1352844536,0.0055989465,-0.3641906083,-0.2349604964,0.222199887,-0.1147741079,-0.1513827443,0.4180804193,0.2285613865,0.5125477314,-0.0601061396,-0.0667559505,0.3555901945,-0.1129636914,0.1070147976,-0.4054171145,-0.206018582,-0.068265304,-0.0098274667,0.014099963,-0.1731091887,0.3809698224,0.1298669577,-0.2413645089,-0.5284497738,-0.1671872437,0.0951377079,-0.00659509,-0.2939401865,0.0953662172,0.4171222448,-0.3732139468,-0.2264623195,-0.3205658793,0.1666924208,0.124835968,-0.2047858983,-0.2520410717,-0.0803067684,-0.1672213823,-0.1547547728,0.2970052958,0.5121389031,0.1777841151,0.0709629655,0.045319017,0.0256576631,0.1568625718,0.2892772555,0.1122660711,-0.1483584791,-0.0695853233,-0.241981566,-0.3659422398,-0.1267194003,-0.4494649172,-0.3581029177,-0.4238283634,0.177418381,-0.0791299641,-0.0909896865,-0.0906132162,-0.0311063919,-0.0000529972,0.5068749189,0.4130570292,0.3568481505,0.024220774,-0.2716104686,0.0795903131,0.4622479081,-0.3353607357,0.1716941744,0.1013360322,0.1934712231,0.0982410982,0.119971469,0.0176324546,0.4626224637,0.1162668541,0.0639665723,-0.4050982893,0.1216212735,0.3568625748,-0.2387486398,-0.1427961141,0.1235067174,0.4282862544,0.2081987262,0.0160211883,-0.1104383841,0.1777976006,0.2067870051,0.1004441231,0.1991128772,0.0972862765,0.4073148966,0.1640243977,0.1200995818,0.0315478593,-0.1210190654,-0.1331913322,0.5670006275,0.1444459409,0.3090570867,-0.176148504,-0.0645436719,0.0112789692,0.0061089913,-0.4022245407,-0.158985287,0.0736067221,-0.1281009167,-0.104061693,0.0811547711,-0.2171427608,-0.2646917105,0.1703396887,-0.0385037884,-0.196485728,0.0833187252,0.4148591161,-0.1362657547,-0.0126301022,0.5968663692,0.2562991977,0.0026542561,-0.326633215,-0.6718171835,-0.287239641,0.4440600574,-0.1358682364,-0.0217890795,0.1068910584,0.2573146522,0.149722293,-0.0075334744,0.5802325606,0.2388835847,-0.1956294626,0.0797595382,-0.2248768955]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"Indeed we convert tensors to list to be able to write in arrow format. Because of this conversion we lose the dtype information. We should add the dtype detection when we do type inference. However it would require a bit of refactoring since currently the conversion happens before the type inference..\r\n\r\nAnd then for your information, when reading from arrow format we have to cast from arrow to numpy (which is fast since pyarrow has a numpy integration), and then to torch.\r\n\r\nHowever there's one thing that can help you: we make sure that the dtypes correspond to what is defined in `features`.\r\nTherefore what you can do is provide `features` in `.map(preprocess, feature=...)` to specify the output types.\r\n\r\nFor example in your case:\r\n```python\r\nfrom datasets import Features, Value, Sequence\r\n\r\nfeatures = Features({\r\n \"input_ids\": Sequence(Value(\"int32\")),\r\n \"sembedding\": Sequence(Value(\"float32\"))\r\n})\r\npreprocessed_dataset = dataset.map(preprocess, features=features)\r\n\r\npreprocessed_dataset.set_format(\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\nprint(preprocessed_dataset[0][\"sembedding\"].dtype)\r\n# \"torch.float32\"\r\n```\r\n\r\nLet me know if it helps","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":156,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\nIndeed we convert tensors to list to be able to write in arrow format. Because of this conversion we lose the dtype information. We should add the dtype detection when we do type inference. However it would require a bit of refactoring since currently the conversion happens before the type inference..\r\n\r\nAnd then for your information, when reading from arrow format we have to cast from arrow to numpy (which is fast since pyarrow has a numpy integration), and then to torch.\r\n\r\nHowever there's one thing that can help you: we make sure that the dtypes correspond to what is defined in `features`.\r\nTherefore what you can do is provide `features` in `.map(preprocess, feature=...)` to specify the output types.\r\n\r\nFor example in your case:\r\n```python\r\nfrom datasets import Features, Value, Sequence\r\n\r\nfeatures = Features({\r\n \"input_ids\": Sequence(Value(\"int32\")),\r\n \"sembedding\": Sequence(Value(\"float32\"))\r\n})\r\npreprocessed_dataset = dataset.map(preprocess, features=features)\r\n\r\npreprocessed_dataset.set_format(\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\nprint(preprocessed_dataset[0][\"sembedding\"].dtype)\r\n# \"torch.float32\"\r\n```\r\n\r\nLet me know if it helps","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"If the arrow format is basically lists, why is the intermediate step to numpy necessary? I am a bit confused about that part.\r\n\r\nThanks for your suggestion. as I have currently implemented this, I cast to torch.Tensor in my collate_fn to save disk space (so I do not have to save padded tensors to max_len but can pad up to max batch len in collate_fn) at the cost of a bit slower processing. So for me this is not relevant anymore, but I am sure it is for others!","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":89,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\nIf the arrow format is basically lists, why is the intermediate step to numpy necessary? I am a bit confused about that part.\r\n\r\nThanks for your suggestion. as I have currently implemented this, I cast to torch.Tensor in my collate_fn to save disk space (so I do not have to save padded tensors to max_len but can pad up to max batch len in collate_fn) at the cost of a bit slower processing. So for me this is not relevant anymore, but I am sure it is for others!","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"I'm glad you managed to figure something out :)\r\n\r\nCasting from arrow to numpy can be 100x faster than casting from arrow to list.\r\nThis is because arrow has an integration with numpy that allows it to instantiate numpy arrays with zero-copy from arrow.\r\nOn the other hand to create python lists it is slow since it has to recreate the list object by iterating through each element in python.","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":70,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\nI'm glad you managed to figure something out :)\r\n\r\nCasting from arrow to numpy can be 100x faster than casting from arrow to list.\r\nThis is because arrow has an integration with numpy that allows it to instantiate numpy arrays with zero-copy from arrow.\r\nOn the other hand to create python lists it is slow since it has to recreate the list object by iterating through each element in python.","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"Ah that is interesting. I have no direct experience with arrow so I didn't know. ","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":15,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\nAh that is interesting. I have no direct experience with arrow so I didn't know. ","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"I encountered a simliar issue: `datasets` converted my float numpy array to `torch.float64` tensors, while many pytorch operations require `torch.float32` inputs and it's very troublesome. \r\n\r\nI tried @lhoestq 's solution, but since it's mixed with the preprocess function, it's not very intuitive. \r\n\r\nI just want to share another possible simpler solution: directly cast the dtype of the processed dataset.\r\n\r\nNow I want to change the type of `labels` in `train_dataset` from float64 to float32, I can do this.\r\n\r\n```\r\nfrom datasets import Value, Sequence, Features\r\nfeats = train_dataset.features.copy()\r\nfeats['labels'].feature = Value(dtype='float32')\r\nfeats = Features(feats)\r\ntrain_dataset.cast_(feats)\r\n```\r\n","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":96,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\nI encountered a simliar issue: `datasets` converted my float numpy array to `torch.float64` tensors, while many pytorch operations require `torch.float32` inputs and it's very troublesome. \r\n\r\nI tried @lhoestq 's solution, but since it's mixed with the preprocess function, it's not very intuitive. \r\n\r\nI just want to share another possible simpler solution: directly cast the dtype of the processed dataset.\r\n\r\nNow I want to change the type of `labels` in `train_dataset` from float64 to float32, I can do this.\r\n\r\n```\r\nfrom datasets import Value, Sequence, Features\r\nfeats = train_dataset.features.copy()\r\nfeats['labels'].feature = Value(dtype='float32')\r\nfeats = Features(feats)\r\ntrain_dataset.cast_(feats)\r\n```\r\n","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"Reopening since @bhavitvyamalik started looking into it !\r\n\r\nAlso I'm posting here a function that could be helpful to support preserving the dtype of tensors.\r\n\r\nIt's used to build a pyarrow array out of a numpy array and:\r\n- it doesn't convert the numpy array to a python list\r\n- it keeps the precision of the numpy array for the pyarrow array\r\n- it works with multidimensional arrays (while `pa.array` can only take a 1D array as input)\r\n- it builds the pyarrow ListArray from offsets created on-the-fly and values that come from the flattened numpy array\r\n\r\n```python\r\nfrom functools import reduce\r\nfrom operator import mul\r\n\r\nimport numpy as np\r\nimport pyarrow as pa\r\n\r\ndef pa_ndarray(a):\r\n \"\"\"Build a PyArrow ListArray from a multidimensional NumPy array\"\"\"\r\n values = pa.array(a.flatten()) \r\n for i in range(a.ndim - 1): \r\n n_offsets = reduce(mul, a.shape[:a.ndim - i - 1], 1) \r\n step_offsets = a.shape[a.ndim - i - 1] \r\n offsets = pa.array(np.arange(n_offsets + 1) * step_offsets, type=pa.int32()) \r\n values = pa.ListArray.from_arrays(offsets, values) \r\n return values \r\n\r\nnarr = np.arange(42).reshape(7, 2, 3).astype(np.uint8)\r\nparr = pa_ndarray(narr)\r\nassert isinstance(parr, pa.Array)\r\nassert parr.type == pa.list_(pa.list_(pa.uint8()))\r\nassert narr.tolist() == parr.to_pylist()\r\n```\r\n\r\nThe only costly operation is the offsets computations. Since it doesn't iterate on the numpy array values this function is pretty fast.","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":206,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\nReopening since @bhavitvyamalik started looking into it !\r\n\r\nAlso I'm posting here a function that could be helpful to support preserving the dtype of tensors.\r\n\r\nIt's used to build a pyarrow array out of a numpy array and:\r\n- it doesn't convert the numpy array to a python list\r\n- it keeps the precision of the numpy array for the pyarrow array\r\n- it works with multidimensional arrays (while `pa.array` can only take a 1D array as input)\r\n- it builds the pyarrow ListArray from offsets created on-the-fly and values that come from the flattened numpy array\r\n\r\n```python\r\nfrom functools import reduce\r\nfrom operator import mul\r\n\r\nimport numpy as np\r\nimport pyarrow as pa\r\n\r\ndef pa_ndarray(a):\r\n \"\"\"Build a PyArrow ListArray from a multidimensional NumPy array\"\"\"\r\n values = pa.array(a.flatten()) \r\n for i in range(a.ndim - 1): \r\n n_offsets = reduce(mul, a.shape[:a.ndim - i - 1], 1) \r\n step_offsets = a.shape[a.ndim - i - 1] \r\n offsets = pa.array(np.arange(n_offsets + 1) * step_offsets, type=pa.int32()) \r\n values = pa.ListArray.from_arrays(offsets, values) \r\n return values \r\n\r\nnarr = np.arange(42).reshape(7, 2, 3).astype(np.uint8)\r\nparr = pa_ndarray(narr)\r\nassert isinstance(parr, pa.Array)\r\nassert parr.type == pa.list_(pa.list_(pa.uint8()))\r\nassert narr.tolist() == parr.to_pylist()\r\n```\r\n\r\nThe only costly operation is the offsets computations. Since it doesn't iterate on the numpy array values this function is pretty fast.","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"@lhoestq Have you thought about this further?\r\n\r\nWe have a use case where we're attempting to load data containing numpy arrays using the `datasets` library.\r\n\r\nWhen using one of the \"standard\" methods (`[Value(...)]` or `Sequence()`) we see ~200 samples processed per second during the call to `_prepare_split`. This slowdown is caused by the vast number of calls to `encode_nested_example` (each sequence is converted to a list, and each element in the sequence...). \r\n\r\nUsing the `Feature` `ArrayND` improves this somewhat to ~500\/s as it now uses numpy's `tolist()` rather than iterating over each value in the array and converting them individually.\r\n\r\nHowever, it's still pretty slow and in theory it should be possible to avoid the `numpy -> python -> arrow` dance altogether. To demonstrate this, if you keep the `Feature` set to an `ArrayND` but instead return a `pa_ndarray(...)` in `_generate_examples` it skips the conversion (`return obj, False`) and hits ~11_000\/s. Two orders of magnitude speed up! The problem is this then fails later on when the `ArrowWriter` tries to write the examples to disk :-( \r\n\r\nIt would be nice to have first-class support for user-defined PyArrow objects. Is this a possibility? We have _large_ datasets where even an order of magnitude difference is important so settling on the middle ~500\/s is less than ideal! \r\n\r\nIs there a workaround for this or another method that should be used instead that gets near-to or equal performance to returning PyArrow arrays?","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":239,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\n@lhoestq Have you thought about this further?\r\n\r\nWe have a use case where we're attempting to load data containing numpy arrays using the `datasets` library.\r\n\r\nWhen using one of the \"standard\" methods (`[Value(...)]` or `Sequence()`) we see ~200 samples processed per second during the call to `_prepare_split`. This slowdown is caused by the vast number of calls to `encode_nested_example` (each sequence is converted to a list, and each element in the sequence...). \r\n\r\nUsing the `Feature` `ArrayND` improves this somewhat to ~500\/s as it now uses numpy's `tolist()` rather than iterating over each value in the array and converting them individually.\r\n\r\nHowever, it's still pretty slow and in theory it should be possible to avoid the `numpy -> python -> arrow` dance altogether. To demonstrate this, if you keep the `Feature` set to an `ArrayND` but instead return a `pa_ndarray(...)` in `_generate_examples` it skips the conversion (`return obj, False`) and hits ~11_000\/s. Two orders of magnitude speed up! The problem is this then fails later on when the `ArrowWriter` tries to write the examples to disk :-( \r\n\r\nIt would be nice to have first-class support for user-defined PyArrow objects. Is this a possibility? We have _large_ datasets where even an order of magnitude difference is important so settling on the middle ~500\/s is less than ideal! \r\n\r\nIs there a workaround for this or another method that should be used instead that gets near-to or equal performance to returning PyArrow arrays?","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"Note that manually generating the table using `pyarrow` achieves ~30_000\/s","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":10,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\nNote that manually generating the table using `pyarrow` achieves ~30_000\/s","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/625","title":"dtype of tensors should be preserved","comments":"Hi !\r\n\r\nIt would be awesome to achieve this speed for numpy arrays !\r\nFor now we have to use `encode_nested_example` to convert numpy arrays to python lists since pyarrow doesn't support multidimensional numpy arrays (only 1D).\r\n\r\nMaybe let's start a new PR from your PR @bhavitvyamalik (idk why we didn't answer your PR at that time, sorry about that).\r\nBasically the idea is to allow `TypedSequence` to support numpy arrays as you did, and remove the numpy->python casting in `_cast_to_python_objects`.\r\n\r\nThis is really important since we are starting to have a focus on other modalities than text as well (audio, images).\r\n\r\nThough until then @samgd, there is another feature that may interest you and that may give you the speed you want:\r\n\r\nIn a dataset script you can subclass either a GeneratorBasedBuilder (with the `_generate_examples ` method) or an ArrowBasedBuilder if you want. the ArrowBasedBuilder allows to yield arrow data by implementing the `_generate_tables` method (it's the same as `_generate_examples` except you must yield arrow tables). Since the data are already in arrow format, it doesn't call `encode_nested_example`. Let me know if that helps.","body":"After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.","comment_length":185,"text":"dtype of tensors should be preserved\nAfter switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https:\/\/discuss.pytorch.org\/t\/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32\/96221)). \r\n\r\nAs a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this:\r\n\r\n```python\r\ndef preprocess(sentences: List[str]):\r\n token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences]\r\n\r\n sembeddings = stransformer.encode(sentences)\r\n print(sembeddings.dtype)\r\n return {\"input_ids\": token_ids, \"sembedding\": sembeddings}\r\n```\r\n\r\nGiven a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column \"sembedding\" to a tensor that I as a user expect to be a float32.\r\n\r\nIt appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. \r\n\r\nMy model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64.\r\n\r\n```python\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n```\r\n\r\nThis happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation\/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. \r\n\r\n```python\r\nimport torch\r\nimport numpy as np\r\n\r\nl = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055]\r\ntorch_tensor = torch.tensor(l)\r\nnp_array = np.array(l)\r\nnp_to_torch = torch.from_numpy(np_array)\r\n\r\nprint(torch_tensor.dtype)\r\n# torch.float32\r\nprint(np_array.dtype)\r\n# float64\r\nprint(np_to_torch.dtype)\r\n# torch.float64\r\n```\r\n\r\nThis might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision.\r\n\r\nThe alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.\nHi !\r\n\r\nIt would be awesome to achieve this speed for numpy arrays !\r\nFor now we have to use `encode_nested_example` to convert numpy arrays to python lists since pyarrow doesn't support multidimensional numpy arrays (only 1D).\r\n\r\nMaybe let's start a new PR from your PR @bhavitvyamalik (idk why we didn't answer your PR at that time, sorry about that).\r\nBasically the idea is to allow `TypedSequence` to support numpy arrays as you did, and remove the numpy->python casting in `_cast_to_python_objects`.\r\n\r\nThis is really important since we are starting to have a focus on other modalities than text as well (audio, images).\r\n\r\nThough until then @samgd, there is another feature that may interest you and that may give you the speed you want:\r\n\r\nIn a dataset script you can subclass either a GeneratorBasedBuilder (with the `_generate_examples ` method) or an ArrowBasedBuilder if you want. the ArrowBasedBuilder allows to yield arrow data by implementing the `_generate_tables` method (it's the same as `_generate_examples` except you must yield arrow tables). Since the data are already in arrow format, it doesn't call `encode_nested_example`. Let me know if that helps.","embeddings":[-0.1134333536,-0.2211149335,-0.0097106956,0.2073052824,0.5532286763,0.1730130315,0.5313699841,0.1225805879,0.150482595,-0.0665395483,-0.0843995288,0.2457148135,-0.1175516769,-0.175145492,0.1026156992,-0.2025278956,0.2282533199,-0.0652568042,-0.1441734284,-0.205046013,-0.2274058759,-0.0888313875,-0.0027089356,-0.2082730979,-0.1601844877,-0.1640590578,0.2403164208,-0.1392812729,-0.1851640195,-0.0010071443,0.1801282763,-0.1652227491,0.4167289436,0.5717647672,-0.0001163156,0.2217250615,-0.0266850162,-0.0754260793,-0.1886361241,-0.0001533552,0.0046074335,-0.1931504607,-0.0723768249,-0.1798607856,-0.1541208327,-0.1833256185,-0.0845149383,-0.7935453057,0.2889484465,0.410826385,0.1519325227,0.2570824623,0.0712021217,0.1525935829,-0.1351298988,0.2725163996,-0.103012912,0.2802624404,-0.0208770931,0.3604668975,-0.0614813603,0.4459063411,-0.3614792526,-0.0750737414,0.0291567855,0.11278864,0.0541251488,-0.4639117122,-0.066283904,0.1654170305,0.1973419636,-0.2320809811,-0.2843651474,-0.1672888398,-0.0933570042,-0.3671621978,0.0662379041,-0.0845816359,0.1454873085,0.1031548381,-0.0847080871,-0.1715696007,-0.0345691293,0.1320960969,-0.4753398299,0.2185822874,-0.0089362506,0.1474191993,-0.1035433561,-0.0327189155,0.0743184015,-0.1159510016,0.2209848762,-0.0332037657,-0.1700056195,-0.2236013412,0.1068102047,-0.4505304396,-0.1278472394,-0.481867671,0.2282644808,0.1562227011,-0.2183923125,-0.0274901818,0.2101505846,0.3317389488,-0.2976427972,0.3678765297,0.2994463146,-0.0447378792,0.2111329287,0.0985190496,0.1282290816,0.0464336239,-0.0115374727,0.1037944704,0.5391173363,-0.1061271057,-0.315695256,0.2423450053,-0.4640916288,0.0854395926,0.0290507488,0.0762219429,-0.1580673456,0.446918875,0.1909368932,0.0856407434,-0.1537808329,-0.0482112318,-0.1408704817,-0.2463302314,0.0493125133,-0.1757228523,0.0473481752,0.1110769287,-0.0556811914,0.290984422,0.1304951161,0.1616608948,-0.0623956695,-0.1468250006,0.5156334639,0.294393152,-0.3300797641,0.1609635502,0.2148128599,-0.3500457704,-0.1941231787,0.34138304,-0.2835562229,0.000267259,-0.255297929,0.0896733627,-0.0626333654,-0.1942966878,0.0383375585,0.516879797,0.5945786238,0.0745409578,0.3335529566,-0.525503695,-0.2444197536,-0.2121095061,0.0651086941,-0.0023088125,-0.4651449621,-0.0084925191,0.2366493046,0.1842959225,0.356190145,0.4137149155,0.0704921186,0.1077738181,0.0749110356,-0.0053551802,0.4220858216,0.1117162034,-0.2204668969,-0.0649491996,-0.0586552843,0.3635667264,-0.0807049423,0.0079559609,0.3194953501,-0.2035728991,0.2078322172,0.0254099835,-0.2768169343,0.0815219134,-0.2061402947,0.0387928039,0.4647621214,-0.0825524181,-0.0355014801,-0.0381801277,-0.3315752745,0.3626123369,0.2424901426,-0.1368885785,0.0368469059,0.0449779332,0.0548131727,-0.0328165814,0.0832959637,-0.082889691,-0.5234786868,-0.0470567532,0.0362101495,0.2670900524,-0.050319504,-0.136253953,0.2245899737,0.083187528,-0.010379171,0.0391422622,0.068150878,-0.1717522889,-0.4256085157,0.0723535568,-0.0966651067,-0.2199121267,-0.0393039584,0.1064351872,-0.4977482855,0.0808114558,0.0075054457,-0.1966703385,-0.2570157945,0.285153985,0.0231736973,-0.0891861841,-0.2323273271,0.0862708464,0.1931117773,-0.0923557654,-0.6234652996,0.5602859855,0.5223332047,0.0226684622,0.196981594,0.4178994596,0.0167538133,0.0668812171,-0.1525678933,0.1664802283,0.1083381176,0.0184073467,-0.2958197296,0.0697329938,0.2497304231,0.1761268973,-0.2609070837,-0.1829364747,-0.1405118257,-0.0761567354,-0.0279818904,0.1387428492,-0.4120871723,0.335906148,0.6934578419,-0.0322736017,0.2642434537,0.0308322366,-0.3138761222,-0.1533606052,0.2285325527,-0.1215764284,0.2470545769,0.0851404145,0.3608531654,-0.1999171078,-0.1625255197,-0.0748273805,0.2153988332,0.0469212495,0.0663601533,0.0049412171,0.0751011446,0.1541916281,-0.1057289094,0.27096048,-0.050399024,0.203207612,-0.4259132743,0.2171180546,-0.4016545713,0.0512693189,-0.2296758741,-0.0917817354,-0.0777890086,-0.2456137538,-0.0293583423,0.0008401733,-0.2700945437,0.3935126066,0.2614880204,0.1166923568,0.0873444155,-0.2350945622,-0.1146133244,-0.1724233925,-0.2385483682,0.040034391,0.208354637,-0.3795832694,0.1849566102,0.3574534059,-0.2676813304,-0.2696791887,-0.7421143651,0.0485985689,-0.2355481237,0.0960817412,0.038121704,0.1293122768,-0.0246390663,0.0510537662,0.0254741795,0.0944768935,-0.0181039479,0.1180345491,-0.1916359514,0.0291311461,-0.2872902155,-0.2801524401,-0.0367826261,-0.1219953895,0.1812550128,-0.0798306614,-0.0676227063,0.0514861755,0.0912298411,0.0310504753,-0.0581302159,0.2310893536,-0.4207531214,0.0914393589,0.3349058032,-0.1597370058,-0.3068858087,-0.1432787478,0.0234636441,-0.018258201,0.0745135397,-0.2800920606,-0.0570334829,-0.2157623917,0.2658351958,-0.014034423,-0.1009145305,0.5142831206,0.232959345,0.0712706,-0.0603918321,0.0153217725,0.0630327985,0.2177911401,0.2721605003,-0.0094194403,0.3699365258,0.1042176038,0.3581045866,0.2943382859,-0.694893539,-0.1260363609,-0.0573534109,0.2153844386,-0.2174506038,-0.2911497355,0.0320028886,-0.2955263555,-0.1989385039,0.0065345895,-0.2766372561,-0.2770918906,0.1184649765,0.2331219167,0.1338419765,-0.4014648199,0.4034068286,-0.3055265546,-0.0188104045,0.0703196749,0.1725534201,-0.1366494894,-0.1540072262,0.1249556765,0.0767138898,0.1390779465,-0.0452391356,-0.1560620666,-0.1616583169,-0.5573336482,0.2960477769,0.0733386204,0.3998686671,-0.1019985527,-0.3138103485,-0.1236895174,0.2658509016,0.3583014011,-0.0651798546,-0.3085556924,-0.0340752453,0.0468399636,-0.1119084507,0.0908219814,-0.370539248,0.1399977952,0.1598600745,0.4799208939,-0.3152986467,-0.1501235217,0.3407086134,0.0990060121,-0.2651866078,0.1257002354,-0.3059802949,-0.2172537744,0.1029196084,-0.1605290771,0.3218425512,-0.1130138487,-0.1629989743,-0.1421160698,-0.1887492388,-0.0593864918,0.3024403751,-0.053796839,0.1746541262,0.273558408,-0.0524263084,0.1738081872,0.4484735429,0.0710393637,0.395806849,-0.2539944947,-0.4725269675,0.1205895767,0.2626002729,0.2191854864,0.3360373974,-0.0665040165,-0.0107505154,0.0495236404,0.2013030946,-0.4152363837,0.1450723261,0.3002059758,0.4395442903,-0.3818068802,-0.3916046619,0.3398995101,0.1049043387,-0.1118494123,0.3452863395,-0.34065184,-0.4210639298,0.4126035869,0.4866728783,0.943916142,-0.0515247025,0.3623064458,0.3886759281,0.0395120159,0.7027948499,-0.2195883393,0.0953722149,-0.1092881337,-0.2003380656,-0.1821270287,0.0283358023,0.1922813058,-0.1768277287,-0.3934646547,-0.0114229573,0.0450148582,0.3950944841,0.0918087959,0.1370074451,0.1337354332,-0.3839355111,0.0010400994,0.0962076709,-0.1420923769,-0.0245583728,0.025837006,-0.129616186,-0.2271977812,-0.0388554968,-0.2351932377,-0.2038244903,-0.6721233726,0.2279058993,-0.2112897038,-0.3824327588,0.2119167894,0.3805317879,0.4986090958,0.0699414685,-0.144413054,-0.0541003644,0.1406936198,0.2504908144,0.3674526215,-0.3385372162,0.5262497663,0.1788809001,-0.0170450285,-0.0523228385,0.0190299135,0.1195790544,-0.3839369118,-0.0444365293,0.0104465103,-0.4486308396,0.0180259272,0.0518191643,0.1929473579,-0.0719195753,0.0803572163,0.0033198879,0.060420122,0.439088881,-0.2384056002,-0.2049404085,-0.1360177696,0.2581453919,0.3120487332,-0.0276788119,0.3316818476,0.174864918,-0.2199062556,-0.1326004863,0.2696622312,0.0962474048,-0.2718690634,0.2938762605,-0.2294584066,-0.3275019526,-0.4029336274,0.0242936164,0.1296148896,0.0200838149,-0.4275782406,0.0481221378,-0.4269218445,0.1955787838,0.1994587332,0.229496792,-0.0968035236,0.2902593017,-0.1862215251,-0.0752579272,-0.2788594961,-0.1454710811,0.2015166283,0.1607856452,-0.0480510816,-0.0622587763,0.2153781503,-0.2433130592,0.0800853521,0.0955782831,0.0832244083,-0.1688759923,-0.1338006407,0.084402971,-0.0721534193,-0.1631305665,-0.0209849011,-0.1953462064,-0.1867278367,-0.167804122,0.1126665175,0.2799358368,0.1018421799,0.2763637006,0.0417334251,0.1626684368,0.3086152673,0.0054951869,-0.1270996332,0.2509422898,0.0635592416,0.3367979825,0.0193025582,-0.0522144847,-0.1994490176,0.2303504944,0.0502555631,-0.0898420513,0.4385258555,-0.1862919033,-0.13802737,0.0113268672,0.1839895248,0.2505350709,-0.312196523,-0.1436047107,-0.0271202698,0.1080143079,-0.1846054345,0.1011110097,0.1920875311,-0.1086047515,0.1267399937,0.4136988521,-0.0296261255,0.5634449124,-0.0397621803,-0.0074645346,0.1605869979,-0.2002462,0.2797859013,0.2631991506,-0.0898274258,0.1576702744,0.2914582491,0.0799149573,0.1357220411,0.1433382183,0.0475203246,0.20890899,0.0597947203,0.0278488416,0.4162573218,-0.1557866335,0.1227421314,0.090849556,-0.4397902787,0.2601419985,0.382668674,0.0751203969,-0.3052257895,-0.2359909266,-0.1305058002,0.1503945887,-0.4149083197,0.0025798059,-0.253952384,0.0859751105,-0.0619617775,-0.2450640351,-0.2087458372,-0.0473561175,0.2073629797,0.2033632994,-0.1458738744,0.0916262567,-0.2127766907,-0.1652701497,0.0895995423,0.0469890758,0.3118063509,0.3976686895,-0.1903300136,-0.1640629023,0.2795583904,0.2608894706,0.190088734,-0.2551006675,0.3188131154,0.1368780434,-0.282248944,0.0019039629,0.2594352663,0.022312453,-0.0181626063,0.215270713,0.1037106141,-0.028723089,0.1587679982,-0.0829131976,0.2346435934,0.0197709147,0.6455355883,-0.0429137461,-0.1467239261,-0.0784749091,-0.2073858827,-0.212395817,-0.2224061489,0.7710811496,-0.1710177213,0.2005575746,-0.0775002688,0.0657908693,0.2825333774,0.3724642694,0.0875464231,-0.0523024946,-0.2187703699,0.3671226501,-0.2781036198,0.1055850387,0.2141845375,-0.0478556752,-0.0218305513,0.2236844748,0.4392271638,0.2588708401,0.0447607264,-0.4173420668,0.2164113373,0.0331136659,-0.1054254174,-0.2507251203,-0.4146490693,0.180609718,0.0294060521,-0.3498017788,0.3316432536,0.1330267042,-0.0241069011,-0.4745323062,0.0894955322,-0.0053972993,0.2187125534,0.2378495932,0.437746495,0.1268995553,-0.1296063215,-0.2555925548,0.1482961327,-0.0524919294,-0.4803514481,0.4116386771,-0.1601348221,0.365714699,-0.301151365,-0.290764004,-0.1951592565,0.025860738,0.0851696506,-0.4364662468,-0.5503333807,0.3894611299,-0.1322181821,-0.0845068172,0.201791212,0.4802951515,0.0986784399,0.3818569779,-0.2019931227,-0.3419241011,0.640550077,-0.2690968513,-0.0412025042,-0.0623360835,-0.0725633353,-0.0648589134,0.0509315468,-0.8520143628,-0.2399327308,0.2974297106,-0.2665519416,-0.2482236624,0.3091590106,0.1196830124,0.1341458261,-0.3145248592,0.5315730572,-0.0456143729,-0.1522001177,0.0940460339,-0.188225776]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/623","title":"Custom feature types in `load_dataset` from CSV","comments":"Currently `csv` doesn't support the `features` attribute (unlike `json`).\r\nWhat you can do for now is cast the features using the in-place transform `cast_`\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'])\r\ndataset.cast_(emotion_features)\r\n```\r\n","body":"I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.","comment_length":38,"text":"Custom feature types in `load_dataset` from CSV\nI am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.\nCurrently `csv` doesn't support the `features` attribute (unlike `json`).\r\nWhat you can do for now is cast the features using the in-place transform `cast_`\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'])\r\ndataset.cast_(emotion_features)\r\n```\r\n","embeddings":[0.0802028403,-0.2782896161,-0.0531788729,0.3509230018,0.3172290027,-0.194310233,0.570133388,0.1113843918,0.4461252391,0.0253307857,0.094744429,0.3161773682,-0.0919099823,0.3901156783,-0.0581883378,0.0267204568,-0.1612745821,0.3348051906,-0.0091216536,-0.3497938216,-0.2713248432,0.1799540073,-0.0930471644,-0.0134863621,-0.3072171211,0.3395891786,0.2221690416,0.1434383839,0.0583117977,-0.3765774071,0.4237109721,0.1486945748,0.3030693531,0.0801393613,-0.0001151109,0.0768831447,0.0293591339,-0.1740051955,-0.0040852865,-0.3799194694,-0.0387842841,-0.202906549,0.424708277,-0.3651530445,-0.2385735363,-0.4564031661,-0.2810543478,-0.12927185,0.2869229019,0.3292911053,0.1512023062,-0.0502676293,-0.2750737369,0.2184944302,0.3587991297,0.6755337119,-0.2523884773,0.1870203614,-0.0406348482,-0.1123654619,0.0939737037,-0.102286227,-0.1438639611,0.3534234762,0.492610544,0.1710615605,-0.064391844,-0.1303502023,-0.1733280122,0.281424731,0.3474106789,0.0562936403,-0.0464353785,-0.3231960535,-0.1238626838,-0.2973287702,0.3824171126,0.0848603621,-0.1818491668,0.1604426801,-0.0789407417,0.4355748296,-0.0581504479,0.325290978,0.0186878722,-0.0715494603,-0.2392114401,0.1055342928,0.0308269262,-0.2173112333,0.0974367186,-0.3491620123,0.2218384594,0.1172008961,-0.0551008694,0.0995841473,0.0039355028,-0.0604649708,-0.1210799068,0.0352325253,0.2292758375,0.1932827532,-0.1932464093,0.0317142829,0.2777897716,0.2441826016,0.2002113014,-0.2317828387,0.1527378112,0.1047160104,-0.4232792258,0.0183332302,0.0277891699,-0.3080181777,0.3964600265,0.0869553015,0.5022855401,-0.2398283333,0.0576624535,-0.0980121568,0.0428094268,-0.0321919918,0.1175666079,0.1835571378,-0.0262655746,0.5809873939,-0.0656143278,0.1488876939,-0.3211797774,0.1283438653,-0.058127325,-0.2226473987,-0.037468221,-0.0493204929,0.4030344188,0.0941047966,0.2478337586,0.199026525,-0.0237019584,-0.3186688125,-0.12700589,-0.0959565863,0.1245454475,0.044335667,-0.441672951,0.3375264406,0.2412355691,-0.2255857587,-0.2444112748,0.095285885,-0.2509175837,-0.2328004986,0.1970367134,0.1555833071,-0.178460896,-0.0209625904,-0.1068924516,0.0988962203,0.1280106902,0.2582726479,0.0263719391,-0.4864165485,-0.3809947371,-0.3249777257,-0.0254871491,0.395431906,-0.4154390991,-0.2039041966,0.0394427255,-0.0463082902,0.1385587603,0.253090173,-0.3661015928,0.1451614648,-0.2393052876,0.2845743299,0.5265450478,0.0433984175,-0.2177205533,0.4294431508,0.1565902233,0.3267548084,0.1860678196,0.0742609575,0.0775937811,0.1260919273,0.1594997197,0.4206010401,0.1657205969,0.0673819482,-0.1837765872,-0.1553835273,0.1656042039,-0.0030714199,-0.3288172483,0.3090256751,0.2257003337,-0.5133963823,0.2171112299,-0.1282682717,-0.2264051139,-0.0468896255,0.4030794501,0.5202861428,0.0057925163,-0.0192435756,-0.4824976325,0.2246332169,-0.1493484527,-0.0522524752,-0.0067959139,-0.3252949715,-0.4934936166,-0.0567670278,-0.2471699268,0.1995169818,0.0646642148,0.3136295676,-0.3358725905,0.1588522345,-0.0870924741,0.0709819943,-0.2088876218,-0.2521634698,-0.0801553875,-0.0217766315,0.1720784605,-0.090050593,0.0157923438,0.0775028616,0.2781896889,0.0767477527,-0.3883281648,0.1277148426,0.2795608342,0.1613467485,-0.11299555,0.2116078585,-0.0030284876,-0.0095141912,-0.0278555136,0.173227042,0.1935628951,-0.0937048942,-0.0290854536,0.6565463543,0.1539311856,0.2489783466,-0.2678474784,-0.1128533781,0.2808355391,-0.0006696465,-0.1841066182,-0.0670059398,-0.2146361172,-0.0485210009,0.0014958512,0.4216481745,-0.3299655318,-0.1157782078,0.4788905382,0.0954588801,0.1291622519,-0.080403775,-0.0172782745,-0.0290924869,0.0167227238,-0.237050727,0.4729242623,-0.0610821843,-0.0803651959,-0.0082820412,0.0192120504,-0.224336803,0.0599999838,-0.1394983679,-0.1910201013,0.2915715277,0.0343268067,-0.1579163671,-0.3376892507,0.2412526906,-0.1466078162,-0.2058871537,-0.634283483,-0.0990623534,-0.5823052526,0.1683030874,-0.4637621641,-0.0822290853,-0.069795303,0.0163359847,-0.2120191008,0.1033786386,-0.0353961922,0.090043202,-0.1784474552,0.3625629246,0.0643079206,-0.8123736382,0.2222122103,0.1188194975,-0.449981004,-0.0581739321,0.1749914587,0.2886117697,0.0355603248,0.0289639011,-0.2870355844,0.0080763036,-0.0132933594,-0.1399178654,0.1707450449,0.5402772427,0.1732708216,0.1756921262,0.1394032091,-0.1260671914,0.3983735442,-0.0131169017,0.0591224469,-0.1250230521,-0.0263378602,0.1114321202,-0.2428705394,-0.6793348789,-0.150197506,-0.2003138959,0.2844246328,0.2139621377,0.0511261635,0.1971265525,0.3683053255,-0.2396501005,0.1546966583,-0.0263569597,-0.2073858082,-0.0725123882,0.4438348711,-0.1748046428,-0.2186147124,0.0093017723,-0.3093702197,-0.312223345,0.1340720356,-0.311917454,0.0612174384,-0.3492916822,0.5237395167,0.0541125759,0.0240740851,0.1082262248,-0.1130078137,0.0396237858,-0.1758637279,-0.3381882906,0.4521704912,0.3127176166,0.0991492271,0.4245265722,-0.0252545048,-0.4129971862,0.4985539317,-0.2855618596,-0.1919134259,0.5953598022,-0.1800899804,0.178718701,-0.0728689805,-0.2670245767,-0.112602815,-0.0425651111,-0.1116565093,0.2236911058,-0.0085567748,-0.2700242996,-0.1810299903,0.2123804986,-0.2176174968,-0.2608528435,0.2379053533,0.1473992467,-0.0105133103,-0.2122010142,-0.0562988408,-0.3047750294,-0.0011159627,-0.0551234484,0.4722039104,-0.1199500263,-0.0018577909,-0.320826441,0.2846590579,0.078286387,0.1430099905,0.2052050531,0.0887329131,0.0514349267,-0.2681927681,-0.0241440311,0.1990232766,0.2873644233,0.0269079208,0.1157544628,0.2087187916,-0.122569643,-0.0485267527,-0.3683553636,0.1046842858,-0.0980243459,-0.1456139088,0.4273831546,-0.0174692385,-0.2047403455,0.0689946637,0.1473543346,-0.3237196207,-0.3247000873,-0.0766784325,-0.103802748,0.0663044602,-0.1572970003,0.0566516109,0.2637958229,-0.4567130208,-0.2949294448,-0.3002611101,-0.0669396892,0.4277199507,0.0272209998,0.360886097,0.0342607722,-0.1682984829,-0.0919508412,0.3646993339,-0.2676113546,0.5217869282,-0.0273078158,-0.3972120583,-0.1356137246,-0.2784287333,0.1692486107,0.2122790515,-0.3540077507,0.0106414184,0.0537485257,-0.1274201721,-0.3462888896,0.4716665745,0.4681007862,-0.1359974295,0.1705080867,-0.770442605,0.3951091468,-0.2586807013,-0.1935910583,-0.0233426094,-0.271920234,-0.2325715125,0.3815220892,0.0811237544,0.6033577919,-0.1063161343,0.1100438684,0.2331691533,0.2517561316,0.0767841414,0.0193486623,-0.1155473888,-0.2250333875,0.1044467911,-0.0204281546,-0.0284996442,0.1987365633,0.4217336178,0.0612299591,0.0763242915,-0.2039118558,0.462703228,-0.1451129019,0.1012346074,0.0826516822,-0.2062903494,-0.1408660561,0.0782319382,-0.0795415342,0.2758697867,0.0676955059,-0.1689655632,0.0029448837,-0.2168522328,0.0900499076,-0.1180126444,-0.3544566035,0.1029871106,-0.0903585181,-0.1665305495,0.0494253188,0.4377081096,0.0453954935,0.072474502,0.0026919779,-0.089243412,0.3455891013,0.0723927319,-0.3652018905,0.0003101975,0.1250230819,0.1094890088,-0.4165462852,-0.2229629606,0.0246600267,-0.2354825586,-0.1175188348,-0.1140048951,-0.0075322078,-0.5912172198,-0.5814692974,-0.2086251378,0.1575358063,0.0230600778,0.086553596,-0.1536920816,0.0544932783,0.4086306691,0.0256565046,-0.293686837,-0.2414160073,0.1097040474,0.0934433118,-0.1074573472,0.4937070608,-0.1585168988,0.0675675422,-0.2334016711,0.1656762809,0.0291343275,-0.2786068022,0.1651807576,0.2246358842,-0.2610382438,-0.1285686195,0.669703424,0.0992187038,0.3524079919,0.0855977982,-0.3124063611,-0.3024978936,-0.1660955548,0.1188771501,0.3957426846,0.0464490056,0.16528292,0.1416232288,-0.2059974223,-0.2483381778,0.0333889201,0.1151640639,0.0171482135,0.1608874202,0.1793142706,0.0733208284,0.2779514492,0.1558933556,-0.0744743124,-0.0549426079,-0.1906136572,-0.243729338,0.1428411305,-0.1014106646,0.2073238939,-0.190238446,-0.2917285264,0.1587083787,-0.228608489,0.1630496234,0.269256413,-0.058433719,0.5097591281,-0.1074662656,0.2579497397,-0.0396139808,0.3279567659,-0.1831430197,0.0911633745,0.2839572728,-0.1068137884,-0.0950942859,0.0866206065,-0.1387923956,-0.2092005014,-0.2366230935,0.3477985859,-0.0584281273,-0.1488164663,-0.0404732265,0.2657982111,0.1655934006,0.4483644366,-0.0675904825,0.0337994769,0.1700197458,0.2013860345,-0.1373620927,0.0610724539,0.2482622862,0.0578231588,0.1259523183,0.5288023949,0.4163005352,-0.0609547384,-0.1263369769,-0.2910684049,0.1425026804,0.1937830746,0.1487294137,0.3334582746,0.0122656683,-0.0583516695,0.0439710692,0.160793677,0.1520167887,0.3418109715,0.029278215,-0.2028436959,0.0008464919,0.1873266101,0.0220549572,-0.6729871035,0.3362360895,0.1451904029,-0.28249228,0.3987858295,0.0277415942,0.304084152,-0.2217611521,-0.222658813,-0.1894535869,0.051500205,-0.0737111419,-0.2347966582,-0.1169679761,-0.1961244643,-0.3086974323,-0.1200800017,0.0403750576,0.1298111379,0.0409613885,0.067796424,-0.0571394302,-0.1646534204,0.1864993572,-0.2584422827,0.2858181298,0.0312479604,0.0454408675,-0.080177322,0.1768501252,0.2635409832,-0.0922866985,-0.0588778891,0.6529513597,-0.0708787143,-0.2127459943,-0.0827961788,0.0492598452,-0.1479174793,-0.18880108,-0.0592490956,0.5324174166,0.2739831507,0.1294231117,-0.1248222217,0.2267606854,-0.2451615632,0.263181597,-0.6641117334,0.3068289161,0.2461200655,0.0778743178,-0.1524754912,0.0674323216,-0.1703536958,-0.1933409423,0.4078930616,0.3887726068,0.5386596322,0.0947164744,0.0588996559,-0.2939717174,0.2791746259,-0.2840400636,0.0861398056,-0.0574960411,0.2992790937,-0.7319200635,-0.0987634361,0.2255541682,-0.1211581826,0.1216320693,0.2576985657,0.0417754911,0.1188103035,-0.155726254,-0.0532568693,0.3186100423,0.0825055987,-0.1304463148,-0.2680415809,-0.1397158355,0.1819897592,0.0978391692,-0.3734224141,-0.0446045958,0.0930457115,0.0255390443,-0.0383036286,-0.2225391865,0.0113506401,0.0509980023,0.2269794792,0.0419547632,0.1949174255,-0.010965066,0.1166931838,-0.2026650012,-0.0816119686,-0.1552953571,0.0546444021,-0.2456047535,0.5321156979,-0.3866825104,0.5693590641,-0.4407524168,0.1738172621,0.2310239524,-0.4541364014,0.0961408019,-0.0945509821,-0.1327079237,-0.1259726882,0.0628547668,0.3869289458,-0.1089815944,0.1471091509,0.1429865211,-0.3117977381,0.3475709558,-0.0603820272,-0.2231802791,-0.0993922278,0.2244294137,0.1183510944,0.3281213641,-0.3093108833,-0.0371792205,0.4952086508,-0.1033602804,0.2260007709,0.1840551496,-0.3150682747,0.0563162304,-0.1813354194,0.142545417,0.0503486358,0.2422543019,-0.2390481234,-0.477942735]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/623","title":"Custom feature types in `load_dataset` from CSV","comments":"Thanks for the clarification!","body":"I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.","comment_length":4,"text":"Custom feature types in `load_dataset` from CSV\nI am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.\nThanks for the clarification!","embeddings":[0.0802028403,-0.2782896161,-0.0531788729,0.3509230018,0.3172290027,-0.194310233,0.570133388,0.1113843918,0.4461252391,0.0253307857,0.094744429,0.3161773682,-0.0919099823,0.3901156783,-0.0581883378,0.0267204568,-0.1612745821,0.3348051906,-0.0091216536,-0.3497938216,-0.2713248432,0.1799540073,-0.0930471644,-0.0134863621,-0.3072171211,0.3395891786,0.2221690416,0.1434383839,0.0583117977,-0.3765774071,0.4237109721,0.1486945748,0.3030693531,0.0801393613,-0.0001151109,0.0768831447,0.0293591339,-0.1740051955,-0.0040852865,-0.3799194694,-0.0387842841,-0.202906549,0.424708277,-0.3651530445,-0.2385735363,-0.4564031661,-0.2810543478,-0.12927185,0.2869229019,0.3292911053,0.1512023062,-0.0502676293,-0.2750737369,0.2184944302,0.3587991297,0.6755337119,-0.2523884773,0.1870203614,-0.0406348482,-0.1123654619,0.0939737037,-0.102286227,-0.1438639611,0.3534234762,0.492610544,0.1710615605,-0.064391844,-0.1303502023,-0.1733280122,0.281424731,0.3474106789,0.0562936403,-0.0464353785,-0.3231960535,-0.1238626838,-0.2973287702,0.3824171126,0.0848603621,-0.1818491668,0.1604426801,-0.0789407417,0.4355748296,-0.0581504479,0.325290978,0.0186878722,-0.0715494603,-0.2392114401,0.1055342928,0.0308269262,-0.2173112333,0.0974367186,-0.3491620123,0.2218384594,0.1172008961,-0.0551008694,0.0995841473,0.0039355028,-0.0604649708,-0.1210799068,0.0352325253,0.2292758375,0.1932827532,-0.1932464093,0.0317142829,0.2777897716,0.2441826016,0.2002113014,-0.2317828387,0.1527378112,0.1047160104,-0.4232792258,0.0183332302,0.0277891699,-0.3080181777,0.3964600265,0.0869553015,0.5022855401,-0.2398283333,0.0576624535,-0.0980121568,0.0428094268,-0.0321919918,0.1175666079,0.1835571378,-0.0262655746,0.5809873939,-0.0656143278,0.1488876939,-0.3211797774,0.1283438653,-0.058127325,-0.2226473987,-0.037468221,-0.0493204929,0.4030344188,0.0941047966,0.2478337586,0.199026525,-0.0237019584,-0.3186688125,-0.12700589,-0.0959565863,0.1245454475,0.044335667,-0.441672951,0.3375264406,0.2412355691,-0.2255857587,-0.2444112748,0.095285885,-0.2509175837,-0.2328004986,0.1970367134,0.1555833071,-0.178460896,-0.0209625904,-0.1068924516,0.0988962203,0.1280106902,0.2582726479,0.0263719391,-0.4864165485,-0.3809947371,-0.3249777257,-0.0254871491,0.395431906,-0.4154390991,-0.2039041966,0.0394427255,-0.0463082902,0.1385587603,0.253090173,-0.3661015928,0.1451614648,-0.2393052876,0.2845743299,0.5265450478,0.0433984175,-0.2177205533,0.4294431508,0.1565902233,0.3267548084,0.1860678196,0.0742609575,0.0775937811,0.1260919273,0.1594997197,0.4206010401,0.1657205969,0.0673819482,-0.1837765872,-0.1553835273,0.1656042039,-0.0030714199,-0.3288172483,0.3090256751,0.2257003337,-0.5133963823,0.2171112299,-0.1282682717,-0.2264051139,-0.0468896255,0.4030794501,0.5202861428,0.0057925163,-0.0192435756,-0.4824976325,0.2246332169,-0.1493484527,-0.0522524752,-0.0067959139,-0.3252949715,-0.4934936166,-0.0567670278,-0.2471699268,0.1995169818,0.0646642148,0.3136295676,-0.3358725905,0.1588522345,-0.0870924741,0.0709819943,-0.2088876218,-0.2521634698,-0.0801553875,-0.0217766315,0.1720784605,-0.090050593,0.0157923438,0.0775028616,0.2781896889,0.0767477527,-0.3883281648,0.1277148426,0.2795608342,0.1613467485,-0.11299555,0.2116078585,-0.0030284876,-0.0095141912,-0.0278555136,0.173227042,0.1935628951,-0.0937048942,-0.0290854536,0.6565463543,0.1539311856,0.2489783466,-0.2678474784,-0.1128533781,0.2808355391,-0.0006696465,-0.1841066182,-0.0670059398,-0.2146361172,-0.0485210009,0.0014958512,0.4216481745,-0.3299655318,-0.1157782078,0.4788905382,0.0954588801,0.1291622519,-0.080403775,-0.0172782745,-0.0290924869,0.0167227238,-0.237050727,0.4729242623,-0.0610821843,-0.0803651959,-0.0082820412,0.0192120504,-0.224336803,0.0599999838,-0.1394983679,-0.1910201013,0.2915715277,0.0343268067,-0.1579163671,-0.3376892507,0.2412526906,-0.1466078162,-0.2058871537,-0.634283483,-0.0990623534,-0.5823052526,0.1683030874,-0.4637621641,-0.0822290853,-0.069795303,0.0163359847,-0.2120191008,0.1033786386,-0.0353961922,0.090043202,-0.1784474552,0.3625629246,0.0643079206,-0.8123736382,0.2222122103,0.1188194975,-0.449981004,-0.0581739321,0.1749914587,0.2886117697,0.0355603248,0.0289639011,-0.2870355844,0.0080763036,-0.0132933594,-0.1399178654,0.1707450449,0.5402772427,0.1732708216,0.1756921262,0.1394032091,-0.1260671914,0.3983735442,-0.0131169017,0.0591224469,-0.1250230521,-0.0263378602,0.1114321202,-0.2428705394,-0.6793348789,-0.150197506,-0.2003138959,0.2844246328,0.2139621377,0.0511261635,0.1971265525,0.3683053255,-0.2396501005,0.1546966583,-0.0263569597,-0.2073858082,-0.0725123882,0.4438348711,-0.1748046428,-0.2186147124,0.0093017723,-0.3093702197,-0.312223345,0.1340720356,-0.311917454,0.0612174384,-0.3492916822,0.5237395167,0.0541125759,0.0240740851,0.1082262248,-0.1130078137,0.0396237858,-0.1758637279,-0.3381882906,0.4521704912,0.3127176166,0.0991492271,0.4245265722,-0.0252545048,-0.4129971862,0.4985539317,-0.2855618596,-0.1919134259,0.5953598022,-0.1800899804,0.178718701,-0.0728689805,-0.2670245767,-0.112602815,-0.0425651111,-0.1116565093,0.2236911058,-0.0085567748,-0.2700242996,-0.1810299903,0.2123804986,-0.2176174968,-0.2608528435,0.2379053533,0.1473992467,-0.0105133103,-0.2122010142,-0.0562988408,-0.3047750294,-0.0011159627,-0.0551234484,0.4722039104,-0.1199500263,-0.0018577909,-0.320826441,0.2846590579,0.078286387,0.1430099905,0.2052050531,0.0887329131,0.0514349267,-0.2681927681,-0.0241440311,0.1990232766,0.2873644233,0.0269079208,0.1157544628,0.2087187916,-0.122569643,-0.0485267527,-0.3683553636,0.1046842858,-0.0980243459,-0.1456139088,0.4273831546,-0.0174692385,-0.2047403455,0.0689946637,0.1473543346,-0.3237196207,-0.3247000873,-0.0766784325,-0.103802748,0.0663044602,-0.1572970003,0.0566516109,0.2637958229,-0.4567130208,-0.2949294448,-0.3002611101,-0.0669396892,0.4277199507,0.0272209998,0.360886097,0.0342607722,-0.1682984829,-0.0919508412,0.3646993339,-0.2676113546,0.5217869282,-0.0273078158,-0.3972120583,-0.1356137246,-0.2784287333,0.1692486107,0.2122790515,-0.3540077507,0.0106414184,0.0537485257,-0.1274201721,-0.3462888896,0.4716665745,0.4681007862,-0.1359974295,0.1705080867,-0.770442605,0.3951091468,-0.2586807013,-0.1935910583,-0.0233426094,-0.271920234,-0.2325715125,0.3815220892,0.0811237544,0.6033577919,-0.1063161343,0.1100438684,0.2331691533,0.2517561316,0.0767841414,0.0193486623,-0.1155473888,-0.2250333875,0.1044467911,-0.0204281546,-0.0284996442,0.1987365633,0.4217336178,0.0612299591,0.0763242915,-0.2039118558,0.462703228,-0.1451129019,0.1012346074,0.0826516822,-0.2062903494,-0.1408660561,0.0782319382,-0.0795415342,0.2758697867,0.0676955059,-0.1689655632,0.0029448837,-0.2168522328,0.0900499076,-0.1180126444,-0.3544566035,0.1029871106,-0.0903585181,-0.1665305495,0.0494253188,0.4377081096,0.0453954935,0.072474502,0.0026919779,-0.089243412,0.3455891013,0.0723927319,-0.3652018905,0.0003101975,0.1250230819,0.1094890088,-0.4165462852,-0.2229629606,0.0246600267,-0.2354825586,-0.1175188348,-0.1140048951,-0.0075322078,-0.5912172198,-0.5814692974,-0.2086251378,0.1575358063,0.0230600778,0.086553596,-0.1536920816,0.0544932783,0.4086306691,0.0256565046,-0.293686837,-0.2414160073,0.1097040474,0.0934433118,-0.1074573472,0.4937070608,-0.1585168988,0.0675675422,-0.2334016711,0.1656762809,0.0291343275,-0.2786068022,0.1651807576,0.2246358842,-0.2610382438,-0.1285686195,0.669703424,0.0992187038,0.3524079919,0.0855977982,-0.3124063611,-0.3024978936,-0.1660955548,0.1188771501,0.3957426846,0.0464490056,0.16528292,0.1416232288,-0.2059974223,-0.2483381778,0.0333889201,0.1151640639,0.0171482135,0.1608874202,0.1793142706,0.0733208284,0.2779514492,0.1558933556,-0.0744743124,-0.0549426079,-0.1906136572,-0.243729338,0.1428411305,-0.1014106646,0.2073238939,-0.190238446,-0.2917285264,0.1587083787,-0.228608489,0.1630496234,0.269256413,-0.058433719,0.5097591281,-0.1074662656,0.2579497397,-0.0396139808,0.3279567659,-0.1831430197,0.0911633745,0.2839572728,-0.1068137884,-0.0950942859,0.0866206065,-0.1387923956,-0.2092005014,-0.2366230935,0.3477985859,-0.0584281273,-0.1488164663,-0.0404732265,0.2657982111,0.1655934006,0.4483644366,-0.0675904825,0.0337994769,0.1700197458,0.2013860345,-0.1373620927,0.0610724539,0.2482622862,0.0578231588,0.1259523183,0.5288023949,0.4163005352,-0.0609547384,-0.1263369769,-0.2910684049,0.1425026804,0.1937830746,0.1487294137,0.3334582746,0.0122656683,-0.0583516695,0.0439710692,0.160793677,0.1520167887,0.3418109715,0.029278215,-0.2028436959,0.0008464919,0.1873266101,0.0220549572,-0.6729871035,0.3362360895,0.1451904029,-0.28249228,0.3987858295,0.0277415942,0.304084152,-0.2217611521,-0.222658813,-0.1894535869,0.051500205,-0.0737111419,-0.2347966582,-0.1169679761,-0.1961244643,-0.3086974323,-0.1200800017,0.0403750576,0.1298111379,0.0409613885,0.067796424,-0.0571394302,-0.1646534204,0.1864993572,-0.2584422827,0.2858181298,0.0312479604,0.0454408675,-0.080177322,0.1768501252,0.2635409832,-0.0922866985,-0.0588778891,0.6529513597,-0.0708787143,-0.2127459943,-0.0827961788,0.0492598452,-0.1479174793,-0.18880108,-0.0592490956,0.5324174166,0.2739831507,0.1294231117,-0.1248222217,0.2267606854,-0.2451615632,0.263181597,-0.6641117334,0.3068289161,0.2461200655,0.0778743178,-0.1524754912,0.0674323216,-0.1703536958,-0.1933409423,0.4078930616,0.3887726068,0.5386596322,0.0947164744,0.0588996559,-0.2939717174,0.2791746259,-0.2840400636,0.0861398056,-0.0574960411,0.2992790937,-0.7319200635,-0.0987634361,0.2255541682,-0.1211581826,0.1216320693,0.2576985657,0.0417754911,0.1188103035,-0.155726254,-0.0532568693,0.3186100423,0.0825055987,-0.1304463148,-0.2680415809,-0.1397158355,0.1819897592,0.0978391692,-0.3734224141,-0.0446045958,0.0930457115,0.0255390443,-0.0383036286,-0.2225391865,0.0113506401,0.0509980023,0.2269794792,0.0419547632,0.1949174255,-0.010965066,0.1166931838,-0.2026650012,-0.0816119686,-0.1552953571,0.0546444021,-0.2456047535,0.5321156979,-0.3866825104,0.5693590641,-0.4407524168,0.1738172621,0.2310239524,-0.4541364014,0.0961408019,-0.0945509821,-0.1327079237,-0.1259726882,0.0628547668,0.3869289458,-0.1089815944,0.1471091509,0.1429865211,-0.3117977381,0.3475709558,-0.0603820272,-0.2231802791,-0.0993922278,0.2244294137,0.1183510944,0.3281213641,-0.3093108833,-0.0371792205,0.4952086508,-0.1033602804,0.2260007709,0.1840551496,-0.3150682747,0.0563162304,-0.1813354194,0.142545417,0.0503486358,0.2422543019,-0.2390481234,-0.477942735]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/623","title":"Custom feature types in `load_dataset` from CSV","comments":"Hi @lhoestq we've tried out your suggestion but are now running into the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-163-81ffd5ac18c9> in <module>\r\n----> 1 dataset.cast_(emotion_features)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/dataset_dict.py in cast_(self, features)\r\n 125 self._check_values_type()\r\n 126 for dataset in self.values():\r\n--> 127 dataset.cast_(features=features)\r\n 128 \r\n 129 def remove_columns_(self, column_names: Union[str, List[str]]):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in cast_(self, features)\r\n 602 self._info.features = features\r\n 603 schema = pa.schema(features.type)\r\n--> 604 self._data = self._data.cast(schema)\r\n 605 \r\n 606 @fingerprint(inplace=True)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.cast()\r\n\r\nValueError: Target schema's field names are not matching the table's field names: ['text', 'label'], ['label', 'text']\r\n```\r\n\r\nLooking at the types in `emotion_features` we see that `label` and `text` appear to be swapped in the Arrow table:\r\n\r\n```\r\nemotion_features.type\r\nStructType(struct<label: int64, text: string>)\r\n```\r\n\r\nDid we define the `emotion_features` incorrectly? We just followed the instructions from the [docs](https:\/\/huggingface.co\/docs\/datasets\/features.html?highlight=features#dataset-features), but perhaps we misunderstood something \ud83d\ude2c \r\n\r\n","body":"I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.","comment_length":168,"text":"Custom feature types in `load_dataset` from CSV\nI am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.\nHi @lhoestq we've tried out your suggestion but are now running into the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-163-81ffd5ac18c9> in <module>\r\n----> 1 dataset.cast_(emotion_features)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/dataset_dict.py in cast_(self, features)\r\n 125 self._check_values_type()\r\n 126 for dataset in self.values():\r\n--> 127 dataset.cast_(features=features)\r\n 128 \r\n 129 def remove_columns_(self, column_names: Union[str, List[str]]):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/datasets\/arrow_dataset.py in cast_(self, features)\r\n 602 self._info.features = features\r\n 603 schema = pa.schema(features.type)\r\n--> 604 self._data = self._data.cast(schema)\r\n 605 \r\n 606 @fingerprint(inplace=True)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.cast()\r\n\r\nValueError: Target schema's field names are not matching the table's field names: ['text', 'label'], ['label', 'text']\r\n```\r\n\r\nLooking at the types in `emotion_features` we see that `label` and `text` appear to be swapped in the Arrow table:\r\n\r\n```\r\nemotion_features.type\r\nStructType(struct<label: int64, text: string>)\r\n```\r\n\r\nDid we define the `emotion_features` incorrectly? We just followed the instructions from the [docs](https:\/\/huggingface.co\/docs\/datasets\/features.html?highlight=features#dataset-features), but perhaps we misunderstood something \ud83d\ude2c \r\n\r\n","embeddings":[0.0802028403,-0.2782896161,-0.0531788729,0.3509230018,0.3172290027,-0.194310233,0.570133388,0.1113843918,0.4461252391,0.0253307857,0.094744429,0.3161773682,-0.0919099823,0.3901156783,-0.0581883378,0.0267204568,-0.1612745821,0.3348051906,-0.0091216536,-0.3497938216,-0.2713248432,0.1799540073,-0.0930471644,-0.0134863621,-0.3072171211,0.3395891786,0.2221690416,0.1434383839,0.0583117977,-0.3765774071,0.4237109721,0.1486945748,0.3030693531,0.0801393613,-0.0001151109,0.0768831447,0.0293591339,-0.1740051955,-0.0040852865,-0.3799194694,-0.0387842841,-0.202906549,0.424708277,-0.3651530445,-0.2385735363,-0.4564031661,-0.2810543478,-0.12927185,0.2869229019,0.3292911053,0.1512023062,-0.0502676293,-0.2750737369,0.2184944302,0.3587991297,0.6755337119,-0.2523884773,0.1870203614,-0.0406348482,-0.1123654619,0.0939737037,-0.102286227,-0.1438639611,0.3534234762,0.492610544,0.1710615605,-0.064391844,-0.1303502023,-0.1733280122,0.281424731,0.3474106789,0.0562936403,-0.0464353785,-0.3231960535,-0.1238626838,-0.2973287702,0.3824171126,0.0848603621,-0.1818491668,0.1604426801,-0.0789407417,0.4355748296,-0.0581504479,0.325290978,0.0186878722,-0.0715494603,-0.2392114401,0.1055342928,0.0308269262,-0.2173112333,0.0974367186,-0.3491620123,0.2218384594,0.1172008961,-0.0551008694,0.0995841473,0.0039355028,-0.0604649708,-0.1210799068,0.0352325253,0.2292758375,0.1932827532,-0.1932464093,0.0317142829,0.2777897716,0.2441826016,0.2002113014,-0.2317828387,0.1527378112,0.1047160104,-0.4232792258,0.0183332302,0.0277891699,-0.3080181777,0.3964600265,0.0869553015,0.5022855401,-0.2398283333,0.0576624535,-0.0980121568,0.0428094268,-0.0321919918,0.1175666079,0.1835571378,-0.0262655746,0.5809873939,-0.0656143278,0.1488876939,-0.3211797774,0.1283438653,-0.058127325,-0.2226473987,-0.037468221,-0.0493204929,0.4030344188,0.0941047966,0.2478337586,0.199026525,-0.0237019584,-0.3186688125,-0.12700589,-0.0959565863,0.1245454475,0.044335667,-0.441672951,0.3375264406,0.2412355691,-0.2255857587,-0.2444112748,0.095285885,-0.2509175837,-0.2328004986,0.1970367134,0.1555833071,-0.178460896,-0.0209625904,-0.1068924516,0.0988962203,0.1280106902,0.2582726479,0.0263719391,-0.4864165485,-0.3809947371,-0.3249777257,-0.0254871491,0.395431906,-0.4154390991,-0.2039041966,0.0394427255,-0.0463082902,0.1385587603,0.253090173,-0.3661015928,0.1451614648,-0.2393052876,0.2845743299,0.5265450478,0.0433984175,-0.2177205533,0.4294431508,0.1565902233,0.3267548084,0.1860678196,0.0742609575,0.0775937811,0.1260919273,0.1594997197,0.4206010401,0.1657205969,0.0673819482,-0.1837765872,-0.1553835273,0.1656042039,-0.0030714199,-0.3288172483,0.3090256751,0.2257003337,-0.5133963823,0.2171112299,-0.1282682717,-0.2264051139,-0.0468896255,0.4030794501,0.5202861428,0.0057925163,-0.0192435756,-0.4824976325,0.2246332169,-0.1493484527,-0.0522524752,-0.0067959139,-0.3252949715,-0.4934936166,-0.0567670278,-0.2471699268,0.1995169818,0.0646642148,0.3136295676,-0.3358725905,0.1588522345,-0.0870924741,0.0709819943,-0.2088876218,-0.2521634698,-0.0801553875,-0.0217766315,0.1720784605,-0.090050593,0.0157923438,0.0775028616,0.2781896889,0.0767477527,-0.3883281648,0.1277148426,0.2795608342,0.1613467485,-0.11299555,0.2116078585,-0.0030284876,-0.0095141912,-0.0278555136,0.173227042,0.1935628951,-0.0937048942,-0.0290854536,0.6565463543,0.1539311856,0.2489783466,-0.2678474784,-0.1128533781,0.2808355391,-0.0006696465,-0.1841066182,-0.0670059398,-0.2146361172,-0.0485210009,0.0014958512,0.4216481745,-0.3299655318,-0.1157782078,0.4788905382,0.0954588801,0.1291622519,-0.080403775,-0.0172782745,-0.0290924869,0.0167227238,-0.237050727,0.4729242623,-0.0610821843,-0.0803651959,-0.0082820412,0.0192120504,-0.224336803,0.0599999838,-0.1394983679,-0.1910201013,0.2915715277,0.0343268067,-0.1579163671,-0.3376892507,0.2412526906,-0.1466078162,-0.2058871537,-0.634283483,-0.0990623534,-0.5823052526,0.1683030874,-0.4637621641,-0.0822290853,-0.069795303,0.0163359847,-0.2120191008,0.1033786386,-0.0353961922,0.090043202,-0.1784474552,0.3625629246,0.0643079206,-0.8123736382,0.2222122103,0.1188194975,-0.449981004,-0.0581739321,0.1749914587,0.2886117697,0.0355603248,0.0289639011,-0.2870355844,0.0080763036,-0.0132933594,-0.1399178654,0.1707450449,0.5402772427,0.1732708216,0.1756921262,0.1394032091,-0.1260671914,0.3983735442,-0.0131169017,0.0591224469,-0.1250230521,-0.0263378602,0.1114321202,-0.2428705394,-0.6793348789,-0.150197506,-0.2003138959,0.2844246328,0.2139621377,0.0511261635,0.1971265525,0.3683053255,-0.2396501005,0.1546966583,-0.0263569597,-0.2073858082,-0.0725123882,0.4438348711,-0.1748046428,-0.2186147124,0.0093017723,-0.3093702197,-0.312223345,0.1340720356,-0.311917454,0.0612174384,-0.3492916822,0.5237395167,0.0541125759,0.0240740851,0.1082262248,-0.1130078137,0.0396237858,-0.1758637279,-0.3381882906,0.4521704912,0.3127176166,0.0991492271,0.4245265722,-0.0252545048,-0.4129971862,0.4985539317,-0.2855618596,-0.1919134259,0.5953598022,-0.1800899804,0.178718701,-0.0728689805,-0.2670245767,-0.112602815,-0.0425651111,-0.1116565093,0.2236911058,-0.0085567748,-0.2700242996,-0.1810299903,0.2123804986,-0.2176174968,-0.2608528435,0.2379053533,0.1473992467,-0.0105133103,-0.2122010142,-0.0562988408,-0.3047750294,-0.0011159627,-0.0551234484,0.4722039104,-0.1199500263,-0.0018577909,-0.320826441,0.2846590579,0.078286387,0.1430099905,0.2052050531,0.0887329131,0.0514349267,-0.2681927681,-0.0241440311,0.1990232766,0.2873644233,0.0269079208,0.1157544628,0.2087187916,-0.122569643,-0.0485267527,-0.3683553636,0.1046842858,-0.0980243459,-0.1456139088,0.4273831546,-0.0174692385,-0.2047403455,0.0689946637,0.1473543346,-0.3237196207,-0.3247000873,-0.0766784325,-0.103802748,0.0663044602,-0.1572970003,0.0566516109,0.2637958229,-0.4567130208,-0.2949294448,-0.3002611101,-0.0669396892,0.4277199507,0.0272209998,0.360886097,0.0342607722,-0.1682984829,-0.0919508412,0.3646993339,-0.2676113546,0.5217869282,-0.0273078158,-0.3972120583,-0.1356137246,-0.2784287333,0.1692486107,0.2122790515,-0.3540077507,0.0106414184,0.0537485257,-0.1274201721,-0.3462888896,0.4716665745,0.4681007862,-0.1359974295,0.1705080867,-0.770442605,0.3951091468,-0.2586807013,-0.1935910583,-0.0233426094,-0.271920234,-0.2325715125,0.3815220892,0.0811237544,0.6033577919,-0.1063161343,0.1100438684,0.2331691533,0.2517561316,0.0767841414,0.0193486623,-0.1155473888,-0.2250333875,0.1044467911,-0.0204281546,-0.0284996442,0.1987365633,0.4217336178,0.0612299591,0.0763242915,-0.2039118558,0.462703228,-0.1451129019,0.1012346074,0.0826516822,-0.2062903494,-0.1408660561,0.0782319382,-0.0795415342,0.2758697867,0.0676955059,-0.1689655632,0.0029448837,-0.2168522328,0.0900499076,-0.1180126444,-0.3544566035,0.1029871106,-0.0903585181,-0.1665305495,0.0494253188,0.4377081096,0.0453954935,0.072474502,0.0026919779,-0.089243412,0.3455891013,0.0723927319,-0.3652018905,0.0003101975,0.1250230819,0.1094890088,-0.4165462852,-0.2229629606,0.0246600267,-0.2354825586,-0.1175188348,-0.1140048951,-0.0075322078,-0.5912172198,-0.5814692974,-0.2086251378,0.1575358063,0.0230600778,0.086553596,-0.1536920816,0.0544932783,0.4086306691,0.0256565046,-0.293686837,-0.2414160073,0.1097040474,0.0934433118,-0.1074573472,0.4937070608,-0.1585168988,0.0675675422,-0.2334016711,0.1656762809,0.0291343275,-0.2786068022,0.1651807576,0.2246358842,-0.2610382438,-0.1285686195,0.669703424,0.0992187038,0.3524079919,0.0855977982,-0.3124063611,-0.3024978936,-0.1660955548,0.1188771501,0.3957426846,0.0464490056,0.16528292,0.1416232288,-0.2059974223,-0.2483381778,0.0333889201,0.1151640639,0.0171482135,0.1608874202,0.1793142706,0.0733208284,0.2779514492,0.1558933556,-0.0744743124,-0.0549426079,-0.1906136572,-0.243729338,0.1428411305,-0.1014106646,0.2073238939,-0.190238446,-0.2917285264,0.1587083787,-0.228608489,0.1630496234,0.269256413,-0.058433719,0.5097591281,-0.1074662656,0.2579497397,-0.0396139808,0.3279567659,-0.1831430197,0.0911633745,0.2839572728,-0.1068137884,-0.0950942859,0.0866206065,-0.1387923956,-0.2092005014,-0.2366230935,0.3477985859,-0.0584281273,-0.1488164663,-0.0404732265,0.2657982111,0.1655934006,0.4483644366,-0.0675904825,0.0337994769,0.1700197458,0.2013860345,-0.1373620927,0.0610724539,0.2482622862,0.0578231588,0.1259523183,0.5288023949,0.4163005352,-0.0609547384,-0.1263369769,-0.2910684049,0.1425026804,0.1937830746,0.1487294137,0.3334582746,0.0122656683,-0.0583516695,0.0439710692,0.160793677,0.1520167887,0.3418109715,0.029278215,-0.2028436959,0.0008464919,0.1873266101,0.0220549572,-0.6729871035,0.3362360895,0.1451904029,-0.28249228,0.3987858295,0.0277415942,0.304084152,-0.2217611521,-0.222658813,-0.1894535869,0.051500205,-0.0737111419,-0.2347966582,-0.1169679761,-0.1961244643,-0.3086974323,-0.1200800017,0.0403750576,0.1298111379,0.0409613885,0.067796424,-0.0571394302,-0.1646534204,0.1864993572,-0.2584422827,0.2858181298,0.0312479604,0.0454408675,-0.080177322,0.1768501252,0.2635409832,-0.0922866985,-0.0588778891,0.6529513597,-0.0708787143,-0.2127459943,-0.0827961788,0.0492598452,-0.1479174793,-0.18880108,-0.0592490956,0.5324174166,0.2739831507,0.1294231117,-0.1248222217,0.2267606854,-0.2451615632,0.263181597,-0.6641117334,0.3068289161,0.2461200655,0.0778743178,-0.1524754912,0.0674323216,-0.1703536958,-0.1933409423,0.4078930616,0.3887726068,0.5386596322,0.0947164744,0.0588996559,-0.2939717174,0.2791746259,-0.2840400636,0.0861398056,-0.0574960411,0.2992790937,-0.7319200635,-0.0987634361,0.2255541682,-0.1211581826,0.1216320693,0.2576985657,0.0417754911,0.1188103035,-0.155726254,-0.0532568693,0.3186100423,0.0825055987,-0.1304463148,-0.2680415809,-0.1397158355,0.1819897592,0.0978391692,-0.3734224141,-0.0446045958,0.0930457115,0.0255390443,-0.0383036286,-0.2225391865,0.0113506401,0.0509980023,0.2269794792,0.0419547632,0.1949174255,-0.010965066,0.1166931838,-0.2026650012,-0.0816119686,-0.1552953571,0.0546444021,-0.2456047535,0.5321156979,-0.3866825104,0.5693590641,-0.4407524168,0.1738172621,0.2310239524,-0.4541364014,0.0961408019,-0.0945509821,-0.1327079237,-0.1259726882,0.0628547668,0.3869289458,-0.1089815944,0.1471091509,0.1429865211,-0.3117977381,0.3475709558,-0.0603820272,-0.2231802791,-0.0993922278,0.2244294137,0.1183510944,0.3281213641,-0.3093108833,-0.0371792205,0.4952086508,-0.1033602804,0.2260007709,0.1840551496,-0.3150682747,0.0563162304,-0.1813354194,0.142545417,0.0503486358,0.2422543019,-0.2390481234,-0.477942735]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/623","title":"Custom feature types in `load_dataset` from CSV","comments":"In general, I don't think there is any hard reason we don't allow to use `features` in the csv script, right @lhoestq?\r\n\r\nShould I add it?","body":"I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.","comment_length":26,"text":"Custom feature types in `load_dataset` from CSV\nI am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.\nIn general, I don't think there is any hard reason we don't allow to use `features` in the csv script, right @lhoestq?\r\n\r\nShould I add it?","embeddings":[0.0802028403,-0.2782896161,-0.0531788729,0.3509230018,0.3172290027,-0.194310233,0.570133388,0.1113843918,0.4461252391,0.0253307857,0.094744429,0.3161773682,-0.0919099823,0.3901156783,-0.0581883378,0.0267204568,-0.1612745821,0.3348051906,-0.0091216536,-0.3497938216,-0.2713248432,0.1799540073,-0.0930471644,-0.0134863621,-0.3072171211,0.3395891786,0.2221690416,0.1434383839,0.0583117977,-0.3765774071,0.4237109721,0.1486945748,0.3030693531,0.0801393613,-0.0001151109,0.0768831447,0.0293591339,-0.1740051955,-0.0040852865,-0.3799194694,-0.0387842841,-0.202906549,0.424708277,-0.3651530445,-0.2385735363,-0.4564031661,-0.2810543478,-0.12927185,0.2869229019,0.3292911053,0.1512023062,-0.0502676293,-0.2750737369,0.2184944302,0.3587991297,0.6755337119,-0.2523884773,0.1870203614,-0.0406348482,-0.1123654619,0.0939737037,-0.102286227,-0.1438639611,0.3534234762,0.492610544,0.1710615605,-0.064391844,-0.1303502023,-0.1733280122,0.281424731,0.3474106789,0.0562936403,-0.0464353785,-0.3231960535,-0.1238626838,-0.2973287702,0.3824171126,0.0848603621,-0.1818491668,0.1604426801,-0.0789407417,0.4355748296,-0.0581504479,0.325290978,0.0186878722,-0.0715494603,-0.2392114401,0.1055342928,0.0308269262,-0.2173112333,0.0974367186,-0.3491620123,0.2218384594,0.1172008961,-0.0551008694,0.0995841473,0.0039355028,-0.0604649708,-0.1210799068,0.0352325253,0.2292758375,0.1932827532,-0.1932464093,0.0317142829,0.2777897716,0.2441826016,0.2002113014,-0.2317828387,0.1527378112,0.1047160104,-0.4232792258,0.0183332302,0.0277891699,-0.3080181777,0.3964600265,0.0869553015,0.5022855401,-0.2398283333,0.0576624535,-0.0980121568,0.0428094268,-0.0321919918,0.1175666079,0.1835571378,-0.0262655746,0.5809873939,-0.0656143278,0.1488876939,-0.3211797774,0.1283438653,-0.058127325,-0.2226473987,-0.037468221,-0.0493204929,0.4030344188,0.0941047966,0.2478337586,0.199026525,-0.0237019584,-0.3186688125,-0.12700589,-0.0959565863,0.1245454475,0.044335667,-0.441672951,0.3375264406,0.2412355691,-0.2255857587,-0.2444112748,0.095285885,-0.2509175837,-0.2328004986,0.1970367134,0.1555833071,-0.178460896,-0.0209625904,-0.1068924516,0.0988962203,0.1280106902,0.2582726479,0.0263719391,-0.4864165485,-0.3809947371,-0.3249777257,-0.0254871491,0.395431906,-0.4154390991,-0.2039041966,0.0394427255,-0.0463082902,0.1385587603,0.253090173,-0.3661015928,0.1451614648,-0.2393052876,0.2845743299,0.5265450478,0.0433984175,-0.2177205533,0.4294431508,0.1565902233,0.3267548084,0.1860678196,0.0742609575,0.0775937811,0.1260919273,0.1594997197,0.4206010401,0.1657205969,0.0673819482,-0.1837765872,-0.1553835273,0.1656042039,-0.0030714199,-0.3288172483,0.3090256751,0.2257003337,-0.5133963823,0.2171112299,-0.1282682717,-0.2264051139,-0.0468896255,0.4030794501,0.5202861428,0.0057925163,-0.0192435756,-0.4824976325,0.2246332169,-0.1493484527,-0.0522524752,-0.0067959139,-0.3252949715,-0.4934936166,-0.0567670278,-0.2471699268,0.1995169818,0.0646642148,0.3136295676,-0.3358725905,0.1588522345,-0.0870924741,0.0709819943,-0.2088876218,-0.2521634698,-0.0801553875,-0.0217766315,0.1720784605,-0.090050593,0.0157923438,0.0775028616,0.2781896889,0.0767477527,-0.3883281648,0.1277148426,0.2795608342,0.1613467485,-0.11299555,0.2116078585,-0.0030284876,-0.0095141912,-0.0278555136,0.173227042,0.1935628951,-0.0937048942,-0.0290854536,0.6565463543,0.1539311856,0.2489783466,-0.2678474784,-0.1128533781,0.2808355391,-0.0006696465,-0.1841066182,-0.0670059398,-0.2146361172,-0.0485210009,0.0014958512,0.4216481745,-0.3299655318,-0.1157782078,0.4788905382,0.0954588801,0.1291622519,-0.080403775,-0.0172782745,-0.0290924869,0.0167227238,-0.237050727,0.4729242623,-0.0610821843,-0.0803651959,-0.0082820412,0.0192120504,-0.224336803,0.0599999838,-0.1394983679,-0.1910201013,0.2915715277,0.0343268067,-0.1579163671,-0.3376892507,0.2412526906,-0.1466078162,-0.2058871537,-0.634283483,-0.0990623534,-0.5823052526,0.1683030874,-0.4637621641,-0.0822290853,-0.069795303,0.0163359847,-0.2120191008,0.1033786386,-0.0353961922,0.090043202,-0.1784474552,0.3625629246,0.0643079206,-0.8123736382,0.2222122103,0.1188194975,-0.449981004,-0.0581739321,0.1749914587,0.2886117697,0.0355603248,0.0289639011,-0.2870355844,0.0080763036,-0.0132933594,-0.1399178654,0.1707450449,0.5402772427,0.1732708216,0.1756921262,0.1394032091,-0.1260671914,0.3983735442,-0.0131169017,0.0591224469,-0.1250230521,-0.0263378602,0.1114321202,-0.2428705394,-0.6793348789,-0.150197506,-0.2003138959,0.2844246328,0.2139621377,0.0511261635,0.1971265525,0.3683053255,-0.2396501005,0.1546966583,-0.0263569597,-0.2073858082,-0.0725123882,0.4438348711,-0.1748046428,-0.2186147124,0.0093017723,-0.3093702197,-0.312223345,0.1340720356,-0.311917454,0.0612174384,-0.3492916822,0.5237395167,0.0541125759,0.0240740851,0.1082262248,-0.1130078137,0.0396237858,-0.1758637279,-0.3381882906,0.4521704912,0.3127176166,0.0991492271,0.4245265722,-0.0252545048,-0.4129971862,0.4985539317,-0.2855618596,-0.1919134259,0.5953598022,-0.1800899804,0.178718701,-0.0728689805,-0.2670245767,-0.112602815,-0.0425651111,-0.1116565093,0.2236911058,-0.0085567748,-0.2700242996,-0.1810299903,0.2123804986,-0.2176174968,-0.2608528435,0.2379053533,0.1473992467,-0.0105133103,-0.2122010142,-0.0562988408,-0.3047750294,-0.0011159627,-0.0551234484,0.4722039104,-0.1199500263,-0.0018577909,-0.320826441,0.2846590579,0.078286387,0.1430099905,0.2052050531,0.0887329131,0.0514349267,-0.2681927681,-0.0241440311,0.1990232766,0.2873644233,0.0269079208,0.1157544628,0.2087187916,-0.122569643,-0.0485267527,-0.3683553636,0.1046842858,-0.0980243459,-0.1456139088,0.4273831546,-0.0174692385,-0.2047403455,0.0689946637,0.1473543346,-0.3237196207,-0.3247000873,-0.0766784325,-0.103802748,0.0663044602,-0.1572970003,0.0566516109,0.2637958229,-0.4567130208,-0.2949294448,-0.3002611101,-0.0669396892,0.4277199507,0.0272209998,0.360886097,0.0342607722,-0.1682984829,-0.0919508412,0.3646993339,-0.2676113546,0.5217869282,-0.0273078158,-0.3972120583,-0.1356137246,-0.2784287333,0.1692486107,0.2122790515,-0.3540077507,0.0106414184,0.0537485257,-0.1274201721,-0.3462888896,0.4716665745,0.4681007862,-0.1359974295,0.1705080867,-0.770442605,0.3951091468,-0.2586807013,-0.1935910583,-0.0233426094,-0.271920234,-0.2325715125,0.3815220892,0.0811237544,0.6033577919,-0.1063161343,0.1100438684,0.2331691533,0.2517561316,0.0767841414,0.0193486623,-0.1155473888,-0.2250333875,0.1044467911,-0.0204281546,-0.0284996442,0.1987365633,0.4217336178,0.0612299591,0.0763242915,-0.2039118558,0.462703228,-0.1451129019,0.1012346074,0.0826516822,-0.2062903494,-0.1408660561,0.0782319382,-0.0795415342,0.2758697867,0.0676955059,-0.1689655632,0.0029448837,-0.2168522328,0.0900499076,-0.1180126444,-0.3544566035,0.1029871106,-0.0903585181,-0.1665305495,0.0494253188,0.4377081096,0.0453954935,0.072474502,0.0026919779,-0.089243412,0.3455891013,0.0723927319,-0.3652018905,0.0003101975,0.1250230819,0.1094890088,-0.4165462852,-0.2229629606,0.0246600267,-0.2354825586,-0.1175188348,-0.1140048951,-0.0075322078,-0.5912172198,-0.5814692974,-0.2086251378,0.1575358063,0.0230600778,0.086553596,-0.1536920816,0.0544932783,0.4086306691,0.0256565046,-0.293686837,-0.2414160073,0.1097040474,0.0934433118,-0.1074573472,0.4937070608,-0.1585168988,0.0675675422,-0.2334016711,0.1656762809,0.0291343275,-0.2786068022,0.1651807576,0.2246358842,-0.2610382438,-0.1285686195,0.669703424,0.0992187038,0.3524079919,0.0855977982,-0.3124063611,-0.3024978936,-0.1660955548,0.1188771501,0.3957426846,0.0464490056,0.16528292,0.1416232288,-0.2059974223,-0.2483381778,0.0333889201,0.1151640639,0.0171482135,0.1608874202,0.1793142706,0.0733208284,0.2779514492,0.1558933556,-0.0744743124,-0.0549426079,-0.1906136572,-0.243729338,0.1428411305,-0.1014106646,0.2073238939,-0.190238446,-0.2917285264,0.1587083787,-0.228608489,0.1630496234,0.269256413,-0.058433719,0.5097591281,-0.1074662656,0.2579497397,-0.0396139808,0.3279567659,-0.1831430197,0.0911633745,0.2839572728,-0.1068137884,-0.0950942859,0.0866206065,-0.1387923956,-0.2092005014,-0.2366230935,0.3477985859,-0.0584281273,-0.1488164663,-0.0404732265,0.2657982111,0.1655934006,0.4483644366,-0.0675904825,0.0337994769,0.1700197458,0.2013860345,-0.1373620927,0.0610724539,0.2482622862,0.0578231588,0.1259523183,0.5288023949,0.4163005352,-0.0609547384,-0.1263369769,-0.2910684049,0.1425026804,0.1937830746,0.1487294137,0.3334582746,0.0122656683,-0.0583516695,0.0439710692,0.160793677,0.1520167887,0.3418109715,0.029278215,-0.2028436959,0.0008464919,0.1873266101,0.0220549572,-0.6729871035,0.3362360895,0.1451904029,-0.28249228,0.3987858295,0.0277415942,0.304084152,-0.2217611521,-0.222658813,-0.1894535869,0.051500205,-0.0737111419,-0.2347966582,-0.1169679761,-0.1961244643,-0.3086974323,-0.1200800017,0.0403750576,0.1298111379,0.0409613885,0.067796424,-0.0571394302,-0.1646534204,0.1864993572,-0.2584422827,0.2858181298,0.0312479604,0.0454408675,-0.080177322,0.1768501252,0.2635409832,-0.0922866985,-0.0588778891,0.6529513597,-0.0708787143,-0.2127459943,-0.0827961788,0.0492598452,-0.1479174793,-0.18880108,-0.0592490956,0.5324174166,0.2739831507,0.1294231117,-0.1248222217,0.2267606854,-0.2451615632,0.263181597,-0.6641117334,0.3068289161,0.2461200655,0.0778743178,-0.1524754912,0.0674323216,-0.1703536958,-0.1933409423,0.4078930616,0.3887726068,0.5386596322,0.0947164744,0.0588996559,-0.2939717174,0.2791746259,-0.2840400636,0.0861398056,-0.0574960411,0.2992790937,-0.7319200635,-0.0987634361,0.2255541682,-0.1211581826,0.1216320693,0.2576985657,0.0417754911,0.1188103035,-0.155726254,-0.0532568693,0.3186100423,0.0825055987,-0.1304463148,-0.2680415809,-0.1397158355,0.1819897592,0.0978391692,-0.3734224141,-0.0446045958,0.0930457115,0.0255390443,-0.0383036286,-0.2225391865,0.0113506401,0.0509980023,0.2269794792,0.0419547632,0.1949174255,-0.010965066,0.1166931838,-0.2026650012,-0.0816119686,-0.1552953571,0.0546444021,-0.2456047535,0.5321156979,-0.3866825104,0.5693590641,-0.4407524168,0.1738172621,0.2310239524,-0.4541364014,0.0961408019,-0.0945509821,-0.1327079237,-0.1259726882,0.0628547668,0.3869289458,-0.1089815944,0.1471091509,0.1429865211,-0.3117977381,0.3475709558,-0.0603820272,-0.2231802791,-0.0993922278,0.2244294137,0.1183510944,0.3281213641,-0.3093108833,-0.0371792205,0.4952086508,-0.1033602804,0.2260007709,0.1840551496,-0.3150682747,0.0563162304,-0.1813354194,0.142545417,0.0503486358,0.2422543019,-0.2390481234,-0.477942735]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/623","title":"Custom feature types in `load_dataset` from CSV","comments":"> In general, I don't think there is any hard reason we don't allow to use `features` in the csv script, right @lhoestq?\r\n> \r\n> Should I add it?\r\n\r\nSure let's add it. Setting the convert options should do the job\r\n\r\n> Hi @lhoestq we've tried out your suggestion but are now running into the following error:\r\n> \r\n> ```\r\n> ---------------------------------------------------------------------------\r\n> ValueError Traceback (most recent call last)\r\n> <ipython-input-163-81ffd5ac18c9> in <module>\r\n> ----> 1 dataset.cast_(emotion_features)\r\n>\r\n> \/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.cast()\r\n> \r\n> ValueError: Target schema's field names are not matching the table's field names: ['text', 'label'], ['label', 'text']\r\n> ```\r\n>\r\n> Did we define the `emotion_features` incorrectly? We just followed the instructions from the [docs](https:\/\/huggingface.co\/docs\/datasets\/features.html?highlight=features#dataset-features), but perhaps we misunderstood something \ud83d\ude2c\r\n\r\nThanks for reporting, that's a bug :) I'm fixing it right now","body":"I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.","comment_length":136,"text":"Custom feature types in `load_dataset` from CSV\nI am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.\n> In general, I don't think there is any hard reason we don't allow to use `features` in the csv script, right @lhoestq?\r\n> \r\n> Should I add it?\r\n\r\nSure let's add it. Setting the convert options should do the job\r\n\r\n> Hi @lhoestq we've tried out your suggestion but are now running into the following error:\r\n> \r\n> ```\r\n> ---------------------------------------------------------------------------\r\n> ValueError Traceback (most recent call last)\r\n> <ipython-input-163-81ffd5ac18c9> in <module>\r\n> ----> 1 dataset.cast_(emotion_features)\r\n>\r\n> \/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.cast()\r\n> \r\n> ValueError: Target schema's field names are not matching the table's field names: ['text', 'label'], ['label', 'text']\r\n> ```\r\n>\r\n> Did we define the `emotion_features` incorrectly? We just followed the instructions from the [docs](https:\/\/huggingface.co\/docs\/datasets\/features.html?highlight=features#dataset-features), but perhaps we misunderstood something \ud83d\ude2c\r\n\r\nThanks for reporting, that's a bug :) I'm fixing it right now","embeddings":[0.0802028403,-0.2782896161,-0.0531788729,0.3509230018,0.3172290027,-0.194310233,0.570133388,0.1113843918,0.4461252391,0.0253307857,0.094744429,0.3161773682,-0.0919099823,0.3901156783,-0.0581883378,0.0267204568,-0.1612745821,0.3348051906,-0.0091216536,-0.3497938216,-0.2713248432,0.1799540073,-0.0930471644,-0.0134863621,-0.3072171211,0.3395891786,0.2221690416,0.1434383839,0.0583117977,-0.3765774071,0.4237109721,0.1486945748,0.3030693531,0.0801393613,-0.0001151109,0.0768831447,0.0293591339,-0.1740051955,-0.0040852865,-0.3799194694,-0.0387842841,-0.202906549,0.424708277,-0.3651530445,-0.2385735363,-0.4564031661,-0.2810543478,-0.12927185,0.2869229019,0.3292911053,0.1512023062,-0.0502676293,-0.2750737369,0.2184944302,0.3587991297,0.6755337119,-0.2523884773,0.1870203614,-0.0406348482,-0.1123654619,0.0939737037,-0.102286227,-0.1438639611,0.3534234762,0.492610544,0.1710615605,-0.064391844,-0.1303502023,-0.1733280122,0.281424731,0.3474106789,0.0562936403,-0.0464353785,-0.3231960535,-0.1238626838,-0.2973287702,0.3824171126,0.0848603621,-0.1818491668,0.1604426801,-0.0789407417,0.4355748296,-0.0581504479,0.325290978,0.0186878722,-0.0715494603,-0.2392114401,0.1055342928,0.0308269262,-0.2173112333,0.0974367186,-0.3491620123,0.2218384594,0.1172008961,-0.0551008694,0.0995841473,0.0039355028,-0.0604649708,-0.1210799068,0.0352325253,0.2292758375,0.1932827532,-0.1932464093,0.0317142829,0.2777897716,0.2441826016,0.2002113014,-0.2317828387,0.1527378112,0.1047160104,-0.4232792258,0.0183332302,0.0277891699,-0.3080181777,0.3964600265,0.0869553015,0.5022855401,-0.2398283333,0.0576624535,-0.0980121568,0.0428094268,-0.0321919918,0.1175666079,0.1835571378,-0.0262655746,0.5809873939,-0.0656143278,0.1488876939,-0.3211797774,0.1283438653,-0.058127325,-0.2226473987,-0.037468221,-0.0493204929,0.4030344188,0.0941047966,0.2478337586,0.199026525,-0.0237019584,-0.3186688125,-0.12700589,-0.0959565863,0.1245454475,0.044335667,-0.441672951,0.3375264406,0.2412355691,-0.2255857587,-0.2444112748,0.095285885,-0.2509175837,-0.2328004986,0.1970367134,0.1555833071,-0.178460896,-0.0209625904,-0.1068924516,0.0988962203,0.1280106902,0.2582726479,0.0263719391,-0.4864165485,-0.3809947371,-0.3249777257,-0.0254871491,0.395431906,-0.4154390991,-0.2039041966,0.0394427255,-0.0463082902,0.1385587603,0.253090173,-0.3661015928,0.1451614648,-0.2393052876,0.2845743299,0.5265450478,0.0433984175,-0.2177205533,0.4294431508,0.1565902233,0.3267548084,0.1860678196,0.0742609575,0.0775937811,0.1260919273,0.1594997197,0.4206010401,0.1657205969,0.0673819482,-0.1837765872,-0.1553835273,0.1656042039,-0.0030714199,-0.3288172483,0.3090256751,0.2257003337,-0.5133963823,0.2171112299,-0.1282682717,-0.2264051139,-0.0468896255,0.4030794501,0.5202861428,0.0057925163,-0.0192435756,-0.4824976325,0.2246332169,-0.1493484527,-0.0522524752,-0.0067959139,-0.3252949715,-0.4934936166,-0.0567670278,-0.2471699268,0.1995169818,0.0646642148,0.3136295676,-0.3358725905,0.1588522345,-0.0870924741,0.0709819943,-0.2088876218,-0.2521634698,-0.0801553875,-0.0217766315,0.1720784605,-0.090050593,0.0157923438,0.0775028616,0.2781896889,0.0767477527,-0.3883281648,0.1277148426,0.2795608342,0.1613467485,-0.11299555,0.2116078585,-0.0030284876,-0.0095141912,-0.0278555136,0.173227042,0.1935628951,-0.0937048942,-0.0290854536,0.6565463543,0.1539311856,0.2489783466,-0.2678474784,-0.1128533781,0.2808355391,-0.0006696465,-0.1841066182,-0.0670059398,-0.2146361172,-0.0485210009,0.0014958512,0.4216481745,-0.3299655318,-0.1157782078,0.4788905382,0.0954588801,0.1291622519,-0.080403775,-0.0172782745,-0.0290924869,0.0167227238,-0.237050727,0.4729242623,-0.0610821843,-0.0803651959,-0.0082820412,0.0192120504,-0.224336803,0.0599999838,-0.1394983679,-0.1910201013,0.2915715277,0.0343268067,-0.1579163671,-0.3376892507,0.2412526906,-0.1466078162,-0.2058871537,-0.634283483,-0.0990623534,-0.5823052526,0.1683030874,-0.4637621641,-0.0822290853,-0.069795303,0.0163359847,-0.2120191008,0.1033786386,-0.0353961922,0.090043202,-0.1784474552,0.3625629246,0.0643079206,-0.8123736382,0.2222122103,0.1188194975,-0.449981004,-0.0581739321,0.1749914587,0.2886117697,0.0355603248,0.0289639011,-0.2870355844,0.0080763036,-0.0132933594,-0.1399178654,0.1707450449,0.5402772427,0.1732708216,0.1756921262,0.1394032091,-0.1260671914,0.3983735442,-0.0131169017,0.0591224469,-0.1250230521,-0.0263378602,0.1114321202,-0.2428705394,-0.6793348789,-0.150197506,-0.2003138959,0.2844246328,0.2139621377,0.0511261635,0.1971265525,0.3683053255,-0.2396501005,0.1546966583,-0.0263569597,-0.2073858082,-0.0725123882,0.4438348711,-0.1748046428,-0.2186147124,0.0093017723,-0.3093702197,-0.312223345,0.1340720356,-0.311917454,0.0612174384,-0.3492916822,0.5237395167,0.0541125759,0.0240740851,0.1082262248,-0.1130078137,0.0396237858,-0.1758637279,-0.3381882906,0.4521704912,0.3127176166,0.0991492271,0.4245265722,-0.0252545048,-0.4129971862,0.4985539317,-0.2855618596,-0.1919134259,0.5953598022,-0.1800899804,0.178718701,-0.0728689805,-0.2670245767,-0.112602815,-0.0425651111,-0.1116565093,0.2236911058,-0.0085567748,-0.2700242996,-0.1810299903,0.2123804986,-0.2176174968,-0.2608528435,0.2379053533,0.1473992467,-0.0105133103,-0.2122010142,-0.0562988408,-0.3047750294,-0.0011159627,-0.0551234484,0.4722039104,-0.1199500263,-0.0018577909,-0.320826441,0.2846590579,0.078286387,0.1430099905,0.2052050531,0.0887329131,0.0514349267,-0.2681927681,-0.0241440311,0.1990232766,0.2873644233,0.0269079208,0.1157544628,0.2087187916,-0.122569643,-0.0485267527,-0.3683553636,0.1046842858,-0.0980243459,-0.1456139088,0.4273831546,-0.0174692385,-0.2047403455,0.0689946637,0.1473543346,-0.3237196207,-0.3247000873,-0.0766784325,-0.103802748,0.0663044602,-0.1572970003,0.0566516109,0.2637958229,-0.4567130208,-0.2949294448,-0.3002611101,-0.0669396892,0.4277199507,0.0272209998,0.360886097,0.0342607722,-0.1682984829,-0.0919508412,0.3646993339,-0.2676113546,0.5217869282,-0.0273078158,-0.3972120583,-0.1356137246,-0.2784287333,0.1692486107,0.2122790515,-0.3540077507,0.0106414184,0.0537485257,-0.1274201721,-0.3462888896,0.4716665745,0.4681007862,-0.1359974295,0.1705080867,-0.770442605,0.3951091468,-0.2586807013,-0.1935910583,-0.0233426094,-0.271920234,-0.2325715125,0.3815220892,0.0811237544,0.6033577919,-0.1063161343,0.1100438684,0.2331691533,0.2517561316,0.0767841414,0.0193486623,-0.1155473888,-0.2250333875,0.1044467911,-0.0204281546,-0.0284996442,0.1987365633,0.4217336178,0.0612299591,0.0763242915,-0.2039118558,0.462703228,-0.1451129019,0.1012346074,0.0826516822,-0.2062903494,-0.1408660561,0.0782319382,-0.0795415342,0.2758697867,0.0676955059,-0.1689655632,0.0029448837,-0.2168522328,0.0900499076,-0.1180126444,-0.3544566035,0.1029871106,-0.0903585181,-0.1665305495,0.0494253188,0.4377081096,0.0453954935,0.072474502,0.0026919779,-0.089243412,0.3455891013,0.0723927319,-0.3652018905,0.0003101975,0.1250230819,0.1094890088,-0.4165462852,-0.2229629606,0.0246600267,-0.2354825586,-0.1175188348,-0.1140048951,-0.0075322078,-0.5912172198,-0.5814692974,-0.2086251378,0.1575358063,0.0230600778,0.086553596,-0.1536920816,0.0544932783,0.4086306691,0.0256565046,-0.293686837,-0.2414160073,0.1097040474,0.0934433118,-0.1074573472,0.4937070608,-0.1585168988,0.0675675422,-0.2334016711,0.1656762809,0.0291343275,-0.2786068022,0.1651807576,0.2246358842,-0.2610382438,-0.1285686195,0.669703424,0.0992187038,0.3524079919,0.0855977982,-0.3124063611,-0.3024978936,-0.1660955548,0.1188771501,0.3957426846,0.0464490056,0.16528292,0.1416232288,-0.2059974223,-0.2483381778,0.0333889201,0.1151640639,0.0171482135,0.1608874202,0.1793142706,0.0733208284,0.2779514492,0.1558933556,-0.0744743124,-0.0549426079,-0.1906136572,-0.243729338,0.1428411305,-0.1014106646,0.2073238939,-0.190238446,-0.2917285264,0.1587083787,-0.228608489,0.1630496234,0.269256413,-0.058433719,0.5097591281,-0.1074662656,0.2579497397,-0.0396139808,0.3279567659,-0.1831430197,0.0911633745,0.2839572728,-0.1068137884,-0.0950942859,0.0866206065,-0.1387923956,-0.2092005014,-0.2366230935,0.3477985859,-0.0584281273,-0.1488164663,-0.0404732265,0.2657982111,0.1655934006,0.4483644366,-0.0675904825,0.0337994769,0.1700197458,0.2013860345,-0.1373620927,0.0610724539,0.2482622862,0.0578231588,0.1259523183,0.5288023949,0.4163005352,-0.0609547384,-0.1263369769,-0.2910684049,0.1425026804,0.1937830746,0.1487294137,0.3334582746,0.0122656683,-0.0583516695,0.0439710692,0.160793677,0.1520167887,0.3418109715,0.029278215,-0.2028436959,0.0008464919,0.1873266101,0.0220549572,-0.6729871035,0.3362360895,0.1451904029,-0.28249228,0.3987858295,0.0277415942,0.304084152,-0.2217611521,-0.222658813,-0.1894535869,0.051500205,-0.0737111419,-0.2347966582,-0.1169679761,-0.1961244643,-0.3086974323,-0.1200800017,0.0403750576,0.1298111379,0.0409613885,0.067796424,-0.0571394302,-0.1646534204,0.1864993572,-0.2584422827,0.2858181298,0.0312479604,0.0454408675,-0.080177322,0.1768501252,0.2635409832,-0.0922866985,-0.0588778891,0.6529513597,-0.0708787143,-0.2127459943,-0.0827961788,0.0492598452,-0.1479174793,-0.18880108,-0.0592490956,0.5324174166,0.2739831507,0.1294231117,-0.1248222217,0.2267606854,-0.2451615632,0.263181597,-0.6641117334,0.3068289161,0.2461200655,0.0778743178,-0.1524754912,0.0674323216,-0.1703536958,-0.1933409423,0.4078930616,0.3887726068,0.5386596322,0.0947164744,0.0588996559,-0.2939717174,0.2791746259,-0.2840400636,0.0861398056,-0.0574960411,0.2992790937,-0.7319200635,-0.0987634361,0.2255541682,-0.1211581826,0.1216320693,0.2576985657,0.0417754911,0.1188103035,-0.155726254,-0.0532568693,0.3186100423,0.0825055987,-0.1304463148,-0.2680415809,-0.1397158355,0.1819897592,0.0978391692,-0.3734224141,-0.0446045958,0.0930457115,0.0255390443,-0.0383036286,-0.2225391865,0.0113506401,0.0509980023,0.2269794792,0.0419547632,0.1949174255,-0.010965066,0.1166931838,-0.2026650012,-0.0816119686,-0.1552953571,0.0546444021,-0.2456047535,0.5321156979,-0.3866825104,0.5693590641,-0.4407524168,0.1738172621,0.2310239524,-0.4541364014,0.0961408019,-0.0945509821,-0.1327079237,-0.1259726882,0.0628547668,0.3869289458,-0.1089815944,0.1471091509,0.1429865211,-0.3117977381,0.3475709558,-0.0603820272,-0.2231802791,-0.0993922278,0.2244294137,0.1183510944,0.3281213641,-0.3093108833,-0.0371792205,0.4952086508,-0.1033602804,0.2260007709,0.1840551496,-0.3150682747,0.0563162304,-0.1813354194,0.142545417,0.0503486358,0.2422543019,-0.2390481234,-0.477942735]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/623","title":"Custom feature types in `load_dataset` from CSV","comments":"PR is open for the `ValueError: Target schema's field names are not matching the table's field names` error.\r\n\r\nI'm adding the features parameter to csv","body":"I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.","comment_length":25,"text":"Custom feature types in `load_dataset` from CSV\nI am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.\nPR is open for the `ValueError: Target schema's field names are not matching the table's field names` error.\r\n\r\nI'm adding the features parameter to csv","embeddings":[0.0802028403,-0.2782896161,-0.0531788729,0.3509230018,0.3172290027,-0.194310233,0.570133388,0.1113843918,0.4461252391,0.0253307857,0.094744429,0.3161773682,-0.0919099823,0.3901156783,-0.0581883378,0.0267204568,-0.1612745821,0.3348051906,-0.0091216536,-0.3497938216,-0.2713248432,0.1799540073,-0.0930471644,-0.0134863621,-0.3072171211,0.3395891786,0.2221690416,0.1434383839,0.0583117977,-0.3765774071,0.4237109721,0.1486945748,0.3030693531,0.0801393613,-0.0001151109,0.0768831447,0.0293591339,-0.1740051955,-0.0040852865,-0.3799194694,-0.0387842841,-0.202906549,0.424708277,-0.3651530445,-0.2385735363,-0.4564031661,-0.2810543478,-0.12927185,0.2869229019,0.3292911053,0.1512023062,-0.0502676293,-0.2750737369,0.2184944302,0.3587991297,0.6755337119,-0.2523884773,0.1870203614,-0.0406348482,-0.1123654619,0.0939737037,-0.102286227,-0.1438639611,0.3534234762,0.492610544,0.1710615605,-0.064391844,-0.1303502023,-0.1733280122,0.281424731,0.3474106789,0.0562936403,-0.0464353785,-0.3231960535,-0.1238626838,-0.2973287702,0.3824171126,0.0848603621,-0.1818491668,0.1604426801,-0.0789407417,0.4355748296,-0.0581504479,0.325290978,0.0186878722,-0.0715494603,-0.2392114401,0.1055342928,0.0308269262,-0.2173112333,0.0974367186,-0.3491620123,0.2218384594,0.1172008961,-0.0551008694,0.0995841473,0.0039355028,-0.0604649708,-0.1210799068,0.0352325253,0.2292758375,0.1932827532,-0.1932464093,0.0317142829,0.2777897716,0.2441826016,0.2002113014,-0.2317828387,0.1527378112,0.1047160104,-0.4232792258,0.0183332302,0.0277891699,-0.3080181777,0.3964600265,0.0869553015,0.5022855401,-0.2398283333,0.0576624535,-0.0980121568,0.0428094268,-0.0321919918,0.1175666079,0.1835571378,-0.0262655746,0.5809873939,-0.0656143278,0.1488876939,-0.3211797774,0.1283438653,-0.058127325,-0.2226473987,-0.037468221,-0.0493204929,0.4030344188,0.0941047966,0.2478337586,0.199026525,-0.0237019584,-0.3186688125,-0.12700589,-0.0959565863,0.1245454475,0.044335667,-0.441672951,0.3375264406,0.2412355691,-0.2255857587,-0.2444112748,0.095285885,-0.2509175837,-0.2328004986,0.1970367134,0.1555833071,-0.178460896,-0.0209625904,-0.1068924516,0.0988962203,0.1280106902,0.2582726479,0.0263719391,-0.4864165485,-0.3809947371,-0.3249777257,-0.0254871491,0.395431906,-0.4154390991,-0.2039041966,0.0394427255,-0.0463082902,0.1385587603,0.253090173,-0.3661015928,0.1451614648,-0.2393052876,0.2845743299,0.5265450478,0.0433984175,-0.2177205533,0.4294431508,0.1565902233,0.3267548084,0.1860678196,0.0742609575,0.0775937811,0.1260919273,0.1594997197,0.4206010401,0.1657205969,0.0673819482,-0.1837765872,-0.1553835273,0.1656042039,-0.0030714199,-0.3288172483,0.3090256751,0.2257003337,-0.5133963823,0.2171112299,-0.1282682717,-0.2264051139,-0.0468896255,0.4030794501,0.5202861428,0.0057925163,-0.0192435756,-0.4824976325,0.2246332169,-0.1493484527,-0.0522524752,-0.0067959139,-0.3252949715,-0.4934936166,-0.0567670278,-0.2471699268,0.1995169818,0.0646642148,0.3136295676,-0.3358725905,0.1588522345,-0.0870924741,0.0709819943,-0.2088876218,-0.2521634698,-0.0801553875,-0.0217766315,0.1720784605,-0.090050593,0.0157923438,0.0775028616,0.2781896889,0.0767477527,-0.3883281648,0.1277148426,0.2795608342,0.1613467485,-0.11299555,0.2116078585,-0.0030284876,-0.0095141912,-0.0278555136,0.173227042,0.1935628951,-0.0937048942,-0.0290854536,0.6565463543,0.1539311856,0.2489783466,-0.2678474784,-0.1128533781,0.2808355391,-0.0006696465,-0.1841066182,-0.0670059398,-0.2146361172,-0.0485210009,0.0014958512,0.4216481745,-0.3299655318,-0.1157782078,0.4788905382,0.0954588801,0.1291622519,-0.080403775,-0.0172782745,-0.0290924869,0.0167227238,-0.237050727,0.4729242623,-0.0610821843,-0.0803651959,-0.0082820412,0.0192120504,-0.224336803,0.0599999838,-0.1394983679,-0.1910201013,0.2915715277,0.0343268067,-0.1579163671,-0.3376892507,0.2412526906,-0.1466078162,-0.2058871537,-0.634283483,-0.0990623534,-0.5823052526,0.1683030874,-0.4637621641,-0.0822290853,-0.069795303,0.0163359847,-0.2120191008,0.1033786386,-0.0353961922,0.090043202,-0.1784474552,0.3625629246,0.0643079206,-0.8123736382,0.2222122103,0.1188194975,-0.449981004,-0.0581739321,0.1749914587,0.2886117697,0.0355603248,0.0289639011,-0.2870355844,0.0080763036,-0.0132933594,-0.1399178654,0.1707450449,0.5402772427,0.1732708216,0.1756921262,0.1394032091,-0.1260671914,0.3983735442,-0.0131169017,0.0591224469,-0.1250230521,-0.0263378602,0.1114321202,-0.2428705394,-0.6793348789,-0.150197506,-0.2003138959,0.2844246328,0.2139621377,0.0511261635,0.1971265525,0.3683053255,-0.2396501005,0.1546966583,-0.0263569597,-0.2073858082,-0.0725123882,0.4438348711,-0.1748046428,-0.2186147124,0.0093017723,-0.3093702197,-0.312223345,0.1340720356,-0.311917454,0.0612174384,-0.3492916822,0.5237395167,0.0541125759,0.0240740851,0.1082262248,-0.1130078137,0.0396237858,-0.1758637279,-0.3381882906,0.4521704912,0.3127176166,0.0991492271,0.4245265722,-0.0252545048,-0.4129971862,0.4985539317,-0.2855618596,-0.1919134259,0.5953598022,-0.1800899804,0.178718701,-0.0728689805,-0.2670245767,-0.112602815,-0.0425651111,-0.1116565093,0.2236911058,-0.0085567748,-0.2700242996,-0.1810299903,0.2123804986,-0.2176174968,-0.2608528435,0.2379053533,0.1473992467,-0.0105133103,-0.2122010142,-0.0562988408,-0.3047750294,-0.0011159627,-0.0551234484,0.4722039104,-0.1199500263,-0.0018577909,-0.320826441,0.2846590579,0.078286387,0.1430099905,0.2052050531,0.0887329131,0.0514349267,-0.2681927681,-0.0241440311,0.1990232766,0.2873644233,0.0269079208,0.1157544628,0.2087187916,-0.122569643,-0.0485267527,-0.3683553636,0.1046842858,-0.0980243459,-0.1456139088,0.4273831546,-0.0174692385,-0.2047403455,0.0689946637,0.1473543346,-0.3237196207,-0.3247000873,-0.0766784325,-0.103802748,0.0663044602,-0.1572970003,0.0566516109,0.2637958229,-0.4567130208,-0.2949294448,-0.3002611101,-0.0669396892,0.4277199507,0.0272209998,0.360886097,0.0342607722,-0.1682984829,-0.0919508412,0.3646993339,-0.2676113546,0.5217869282,-0.0273078158,-0.3972120583,-0.1356137246,-0.2784287333,0.1692486107,0.2122790515,-0.3540077507,0.0106414184,0.0537485257,-0.1274201721,-0.3462888896,0.4716665745,0.4681007862,-0.1359974295,0.1705080867,-0.770442605,0.3951091468,-0.2586807013,-0.1935910583,-0.0233426094,-0.271920234,-0.2325715125,0.3815220892,0.0811237544,0.6033577919,-0.1063161343,0.1100438684,0.2331691533,0.2517561316,0.0767841414,0.0193486623,-0.1155473888,-0.2250333875,0.1044467911,-0.0204281546,-0.0284996442,0.1987365633,0.4217336178,0.0612299591,0.0763242915,-0.2039118558,0.462703228,-0.1451129019,0.1012346074,0.0826516822,-0.2062903494,-0.1408660561,0.0782319382,-0.0795415342,0.2758697867,0.0676955059,-0.1689655632,0.0029448837,-0.2168522328,0.0900499076,-0.1180126444,-0.3544566035,0.1029871106,-0.0903585181,-0.1665305495,0.0494253188,0.4377081096,0.0453954935,0.072474502,0.0026919779,-0.089243412,0.3455891013,0.0723927319,-0.3652018905,0.0003101975,0.1250230819,0.1094890088,-0.4165462852,-0.2229629606,0.0246600267,-0.2354825586,-0.1175188348,-0.1140048951,-0.0075322078,-0.5912172198,-0.5814692974,-0.2086251378,0.1575358063,0.0230600778,0.086553596,-0.1536920816,0.0544932783,0.4086306691,0.0256565046,-0.293686837,-0.2414160073,0.1097040474,0.0934433118,-0.1074573472,0.4937070608,-0.1585168988,0.0675675422,-0.2334016711,0.1656762809,0.0291343275,-0.2786068022,0.1651807576,0.2246358842,-0.2610382438,-0.1285686195,0.669703424,0.0992187038,0.3524079919,0.0855977982,-0.3124063611,-0.3024978936,-0.1660955548,0.1188771501,0.3957426846,0.0464490056,0.16528292,0.1416232288,-0.2059974223,-0.2483381778,0.0333889201,0.1151640639,0.0171482135,0.1608874202,0.1793142706,0.0733208284,0.2779514492,0.1558933556,-0.0744743124,-0.0549426079,-0.1906136572,-0.243729338,0.1428411305,-0.1014106646,0.2073238939,-0.190238446,-0.2917285264,0.1587083787,-0.228608489,0.1630496234,0.269256413,-0.058433719,0.5097591281,-0.1074662656,0.2579497397,-0.0396139808,0.3279567659,-0.1831430197,0.0911633745,0.2839572728,-0.1068137884,-0.0950942859,0.0866206065,-0.1387923956,-0.2092005014,-0.2366230935,0.3477985859,-0.0584281273,-0.1488164663,-0.0404732265,0.2657982111,0.1655934006,0.4483644366,-0.0675904825,0.0337994769,0.1700197458,0.2013860345,-0.1373620927,0.0610724539,0.2482622862,0.0578231588,0.1259523183,0.5288023949,0.4163005352,-0.0609547384,-0.1263369769,-0.2910684049,0.1425026804,0.1937830746,0.1487294137,0.3334582746,0.0122656683,-0.0583516695,0.0439710692,0.160793677,0.1520167887,0.3418109715,0.029278215,-0.2028436959,0.0008464919,0.1873266101,0.0220549572,-0.6729871035,0.3362360895,0.1451904029,-0.28249228,0.3987858295,0.0277415942,0.304084152,-0.2217611521,-0.222658813,-0.1894535869,0.051500205,-0.0737111419,-0.2347966582,-0.1169679761,-0.1961244643,-0.3086974323,-0.1200800017,0.0403750576,0.1298111379,0.0409613885,0.067796424,-0.0571394302,-0.1646534204,0.1864993572,-0.2584422827,0.2858181298,0.0312479604,0.0454408675,-0.080177322,0.1768501252,0.2635409832,-0.0922866985,-0.0588778891,0.6529513597,-0.0708787143,-0.2127459943,-0.0827961788,0.0492598452,-0.1479174793,-0.18880108,-0.0592490956,0.5324174166,0.2739831507,0.1294231117,-0.1248222217,0.2267606854,-0.2451615632,0.263181597,-0.6641117334,0.3068289161,0.2461200655,0.0778743178,-0.1524754912,0.0674323216,-0.1703536958,-0.1933409423,0.4078930616,0.3887726068,0.5386596322,0.0947164744,0.0588996559,-0.2939717174,0.2791746259,-0.2840400636,0.0861398056,-0.0574960411,0.2992790937,-0.7319200635,-0.0987634361,0.2255541682,-0.1211581826,0.1216320693,0.2576985657,0.0417754911,0.1188103035,-0.155726254,-0.0532568693,0.3186100423,0.0825055987,-0.1304463148,-0.2680415809,-0.1397158355,0.1819897592,0.0978391692,-0.3734224141,-0.0446045958,0.0930457115,0.0255390443,-0.0383036286,-0.2225391865,0.0113506401,0.0509980023,0.2269794792,0.0419547632,0.1949174255,-0.010965066,0.1166931838,-0.2026650012,-0.0816119686,-0.1552953571,0.0546444021,-0.2456047535,0.5321156979,-0.3866825104,0.5693590641,-0.4407524168,0.1738172621,0.2310239524,-0.4541364014,0.0961408019,-0.0945509821,-0.1327079237,-0.1259726882,0.0628547668,0.3869289458,-0.1089815944,0.1471091509,0.1429865211,-0.3117977381,0.3475709558,-0.0603820272,-0.2231802791,-0.0993922278,0.2244294137,0.1183510944,0.3281213641,-0.3093108833,-0.0371792205,0.4952086508,-0.1033602804,0.2260007709,0.1840551496,-0.3150682747,0.0563162304,-0.1813354194,0.142545417,0.0503486358,0.2422543019,-0.2390481234,-0.477942735]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/623","title":"Custom feature types in `load_dataset` from CSV","comments":"Thanks a lot for the PR and quick fix @lhoestq!","body":"I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.","comment_length":10,"text":"Custom feature types in `load_dataset` from CSV\nI am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. \r\n\r\nI am working with the local files from the emotion dataset. To get the data you can use the following code:\r\n\r\n```Python\r\nfrom pathlib import Path\r\nimport wget\r\n\r\nEMOTION_PATH = Path(\".\/data\/emotion\")\r\nDOWNLOAD_URLS = [\r\n \"https:\/\/www.dropbox.com\/s\/1pzkadrvffbqw6o\/train.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/2mzialpsgf9k5l3\/val.txt?dl=1\",\r\n \"https:\/\/www.dropbox.com\/s\/ikkqxfdbdec3fuj\/test.txt?dl=1\",\r\n]\r\n\r\nif not Path.is_dir(EMOTION_PATH):\r\n Path.mkdir(EMOTION_PATH)\r\nfor url in DOWNLOAD_URLS:\r\n wget.download(url, str(EMOTION_PATH))\r\n```\r\n\r\nThe first five lines of the train set are:\r\n```\r\ni didnt feel humiliated;sadness\r\ni can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness\r\nim grabbing a minute to post i feel greedy wrong;anger\r\ni am ever feeling nostalgic about the fireplace i will know that it is still on the property;love\r\ni am feeling grouchy;anger\r\n```\r\n\r\nHere the code to reproduce the issue:\r\n```Python\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\nclass_names = [\"sadness\", \"joy\", \"love\", \"anger\", \"fear\", \"surprise\"]\r\nemotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)})\r\nfile_dict = {'train': EMOTION_PATH\/'train.txt'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features)\r\n```\r\n\r\n**Observed behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': Value(dtype='string', id=None)}\r\n```\r\n**Expected behaviour:**\r\n```Python\r\ndataset['train'].features\r\n```\r\n```Python\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)}\r\n```\r\n\r\n**Things I've tried:**\r\n- deleting the cache\r\n- trying other types such as `int64`\r\n\r\nAm I missing anything? Thanks for any pointer in the right direction.\nThanks a lot for the PR and quick fix @lhoestq!","embeddings":[0.0802028403,-0.2782896161,-0.0531788729,0.3509230018,0.3172290027,-0.194310233,0.570133388,0.1113843918,0.4461252391,0.0253307857,0.094744429,0.3161773682,-0.0919099823,0.3901156783,-0.0581883378,0.0267204568,-0.1612745821,0.3348051906,-0.0091216536,-0.3497938216,-0.2713248432,0.1799540073,-0.0930471644,-0.0134863621,-0.3072171211,0.3395891786,0.2221690416,0.1434383839,0.0583117977,-0.3765774071,0.4237109721,0.1486945748,0.3030693531,0.0801393613,-0.0001151109,0.0768831447,0.0293591339,-0.1740051955,-0.0040852865,-0.3799194694,-0.0387842841,-0.202906549,0.424708277,-0.3651530445,-0.2385735363,-0.4564031661,-0.2810543478,-0.12927185,0.2869229019,0.3292911053,0.1512023062,-0.0502676293,-0.2750737369,0.2184944302,0.3587991297,0.6755337119,-0.2523884773,0.1870203614,-0.0406348482,-0.1123654619,0.0939737037,-0.102286227,-0.1438639611,0.3534234762,0.492610544,0.1710615605,-0.064391844,-0.1303502023,-0.1733280122,0.281424731,0.3474106789,0.0562936403,-0.0464353785,-0.3231960535,-0.1238626838,-0.2973287702,0.3824171126,0.0848603621,-0.1818491668,0.1604426801,-0.0789407417,0.4355748296,-0.0581504479,0.325290978,0.0186878722,-0.0715494603,-0.2392114401,0.1055342928,0.0308269262,-0.2173112333,0.0974367186,-0.3491620123,0.2218384594,0.1172008961,-0.0551008694,0.0995841473,0.0039355028,-0.0604649708,-0.1210799068,0.0352325253,0.2292758375,0.1932827532,-0.1932464093,0.0317142829,0.2777897716,0.2441826016,0.2002113014,-0.2317828387,0.1527378112,0.1047160104,-0.4232792258,0.0183332302,0.0277891699,-0.3080181777,0.3964600265,0.0869553015,0.5022855401,-0.2398283333,0.0576624535,-0.0980121568,0.0428094268,-0.0321919918,0.1175666079,0.1835571378,-0.0262655746,0.5809873939,-0.0656143278,0.1488876939,-0.3211797774,0.1283438653,-0.058127325,-0.2226473987,-0.037468221,-0.0493204929,0.4030344188,0.0941047966,0.2478337586,0.199026525,-0.0237019584,-0.3186688125,-0.12700589,-0.0959565863,0.1245454475,0.044335667,-0.441672951,0.3375264406,0.2412355691,-0.2255857587,-0.2444112748,0.095285885,-0.2509175837,-0.2328004986,0.1970367134,0.1555833071,-0.178460896,-0.0209625904,-0.1068924516,0.0988962203,0.1280106902,0.2582726479,0.0263719391,-0.4864165485,-0.3809947371,-0.3249777257,-0.0254871491,0.395431906,-0.4154390991,-0.2039041966,0.0394427255,-0.0463082902,0.1385587603,0.253090173,-0.3661015928,0.1451614648,-0.2393052876,0.2845743299,0.5265450478,0.0433984175,-0.2177205533,0.4294431508,0.1565902233,0.3267548084,0.1860678196,0.0742609575,0.0775937811,0.1260919273,0.1594997197,0.4206010401,0.1657205969,0.0673819482,-0.1837765872,-0.1553835273,0.1656042039,-0.0030714199,-0.3288172483,0.3090256751,0.2257003337,-0.5133963823,0.2171112299,-0.1282682717,-0.2264051139,-0.0468896255,0.4030794501,0.5202861428,0.0057925163,-0.0192435756,-0.4824976325,0.2246332169,-0.1493484527,-0.0522524752,-0.0067959139,-0.3252949715,-0.4934936166,-0.0567670278,-0.2471699268,0.1995169818,0.0646642148,0.3136295676,-0.3358725905,0.1588522345,-0.0870924741,0.0709819943,-0.2088876218,-0.2521634698,-0.0801553875,-0.0217766315,0.1720784605,-0.090050593,0.0157923438,0.0775028616,0.2781896889,0.0767477527,-0.3883281648,0.1277148426,0.2795608342,0.1613467485,-0.11299555,0.2116078585,-0.0030284876,-0.0095141912,-0.0278555136,0.173227042,0.1935628951,-0.0937048942,-0.0290854536,0.6565463543,0.1539311856,0.2489783466,-0.2678474784,-0.1128533781,0.2808355391,-0.0006696465,-0.1841066182,-0.0670059398,-0.2146361172,-0.0485210009,0.0014958512,0.4216481745,-0.3299655318,-0.1157782078,0.4788905382,0.0954588801,0.1291622519,-0.080403775,-0.0172782745,-0.0290924869,0.0167227238,-0.237050727,0.4729242623,-0.0610821843,-0.0803651959,-0.0082820412,0.0192120504,-0.224336803,0.0599999838,-0.1394983679,-0.1910201013,0.2915715277,0.0343268067,-0.1579163671,-0.3376892507,0.2412526906,-0.1466078162,-0.2058871537,-0.634283483,-0.0990623534,-0.5823052526,0.1683030874,-0.4637621641,-0.0822290853,-0.069795303,0.0163359847,-0.2120191008,0.1033786386,-0.0353961922,0.090043202,-0.1784474552,0.3625629246,0.0643079206,-0.8123736382,0.2222122103,0.1188194975,-0.449981004,-0.0581739321,0.1749914587,0.2886117697,0.0355603248,0.0289639011,-0.2870355844,0.0080763036,-0.0132933594,-0.1399178654,0.1707450449,0.5402772427,0.1732708216,0.1756921262,0.1394032091,-0.1260671914,0.3983735442,-0.0131169017,0.0591224469,-0.1250230521,-0.0263378602,0.1114321202,-0.2428705394,-0.6793348789,-0.150197506,-0.2003138959,0.2844246328,0.2139621377,0.0511261635,0.1971265525,0.3683053255,-0.2396501005,0.1546966583,-0.0263569597,-0.2073858082,-0.0725123882,0.4438348711,-0.1748046428,-0.2186147124,0.0093017723,-0.3093702197,-0.312223345,0.1340720356,-0.311917454,0.0612174384,-0.3492916822,0.5237395167,0.0541125759,0.0240740851,0.1082262248,-0.1130078137,0.0396237858,-0.1758637279,-0.3381882906,0.4521704912,0.3127176166,0.0991492271,0.4245265722,-0.0252545048,-0.4129971862,0.4985539317,-0.2855618596,-0.1919134259,0.5953598022,-0.1800899804,0.178718701,-0.0728689805,-0.2670245767,-0.112602815,-0.0425651111,-0.1116565093,0.2236911058,-0.0085567748,-0.2700242996,-0.1810299903,0.2123804986,-0.2176174968,-0.2608528435,0.2379053533,0.1473992467,-0.0105133103,-0.2122010142,-0.0562988408,-0.3047750294,-0.0011159627,-0.0551234484,0.4722039104,-0.1199500263,-0.0018577909,-0.320826441,0.2846590579,0.078286387,0.1430099905,0.2052050531,0.0887329131,0.0514349267,-0.2681927681,-0.0241440311,0.1990232766,0.2873644233,0.0269079208,0.1157544628,0.2087187916,-0.122569643,-0.0485267527,-0.3683553636,0.1046842858,-0.0980243459,-0.1456139088,0.4273831546,-0.0174692385,-0.2047403455,0.0689946637,0.1473543346,-0.3237196207,-0.3247000873,-0.0766784325,-0.103802748,0.0663044602,-0.1572970003,0.0566516109,0.2637958229,-0.4567130208,-0.2949294448,-0.3002611101,-0.0669396892,0.4277199507,0.0272209998,0.360886097,0.0342607722,-0.1682984829,-0.0919508412,0.3646993339,-0.2676113546,0.5217869282,-0.0273078158,-0.3972120583,-0.1356137246,-0.2784287333,0.1692486107,0.2122790515,-0.3540077507,0.0106414184,0.0537485257,-0.1274201721,-0.3462888896,0.4716665745,0.4681007862,-0.1359974295,0.1705080867,-0.770442605,0.3951091468,-0.2586807013,-0.1935910583,-0.0233426094,-0.271920234,-0.2325715125,0.3815220892,0.0811237544,0.6033577919,-0.1063161343,0.1100438684,0.2331691533,0.2517561316,0.0767841414,0.0193486623,-0.1155473888,-0.2250333875,0.1044467911,-0.0204281546,-0.0284996442,0.1987365633,0.4217336178,0.0612299591,0.0763242915,-0.2039118558,0.462703228,-0.1451129019,0.1012346074,0.0826516822,-0.2062903494,-0.1408660561,0.0782319382,-0.0795415342,0.2758697867,0.0676955059,-0.1689655632,0.0029448837,-0.2168522328,0.0900499076,-0.1180126444,-0.3544566035,0.1029871106,-0.0903585181,-0.1665305495,0.0494253188,0.4377081096,0.0453954935,0.072474502,0.0026919779,-0.089243412,0.3455891013,0.0723927319,-0.3652018905,0.0003101975,0.1250230819,0.1094890088,-0.4165462852,-0.2229629606,0.0246600267,-0.2354825586,-0.1175188348,-0.1140048951,-0.0075322078,-0.5912172198,-0.5814692974,-0.2086251378,0.1575358063,0.0230600778,0.086553596,-0.1536920816,0.0544932783,0.4086306691,0.0256565046,-0.293686837,-0.2414160073,0.1097040474,0.0934433118,-0.1074573472,0.4937070608,-0.1585168988,0.0675675422,-0.2334016711,0.1656762809,0.0291343275,-0.2786068022,0.1651807576,0.2246358842,-0.2610382438,-0.1285686195,0.669703424,0.0992187038,0.3524079919,0.0855977982,-0.3124063611,-0.3024978936,-0.1660955548,0.1188771501,0.3957426846,0.0464490056,0.16528292,0.1416232288,-0.2059974223,-0.2483381778,0.0333889201,0.1151640639,0.0171482135,0.1608874202,0.1793142706,0.0733208284,0.2779514492,0.1558933556,-0.0744743124,-0.0549426079,-0.1906136572,-0.243729338,0.1428411305,-0.1014106646,0.2073238939,-0.190238446,-0.2917285264,0.1587083787,-0.228608489,0.1630496234,0.269256413,-0.058433719,0.5097591281,-0.1074662656,0.2579497397,-0.0396139808,0.3279567659,-0.1831430197,0.0911633745,0.2839572728,-0.1068137884,-0.0950942859,0.0866206065,-0.1387923956,-0.2092005014,-0.2366230935,0.3477985859,-0.0584281273,-0.1488164663,-0.0404732265,0.2657982111,0.1655934006,0.4483644366,-0.0675904825,0.0337994769,0.1700197458,0.2013860345,-0.1373620927,0.0610724539,0.2482622862,0.0578231588,0.1259523183,0.5288023949,0.4163005352,-0.0609547384,-0.1263369769,-0.2910684049,0.1425026804,0.1937830746,0.1487294137,0.3334582746,0.0122656683,-0.0583516695,0.0439710692,0.160793677,0.1520167887,0.3418109715,0.029278215,-0.2028436959,0.0008464919,0.1873266101,0.0220549572,-0.6729871035,0.3362360895,0.1451904029,-0.28249228,0.3987858295,0.0277415942,0.304084152,-0.2217611521,-0.222658813,-0.1894535869,0.051500205,-0.0737111419,-0.2347966582,-0.1169679761,-0.1961244643,-0.3086974323,-0.1200800017,0.0403750576,0.1298111379,0.0409613885,0.067796424,-0.0571394302,-0.1646534204,0.1864993572,-0.2584422827,0.2858181298,0.0312479604,0.0454408675,-0.080177322,0.1768501252,0.2635409832,-0.0922866985,-0.0588778891,0.6529513597,-0.0708787143,-0.2127459943,-0.0827961788,0.0492598452,-0.1479174793,-0.18880108,-0.0592490956,0.5324174166,0.2739831507,0.1294231117,-0.1248222217,0.2267606854,-0.2451615632,0.263181597,-0.6641117334,0.3068289161,0.2461200655,0.0778743178,-0.1524754912,0.0674323216,-0.1703536958,-0.1933409423,0.4078930616,0.3887726068,0.5386596322,0.0947164744,0.0588996559,-0.2939717174,0.2791746259,-0.2840400636,0.0861398056,-0.0574960411,0.2992790937,-0.7319200635,-0.0987634361,0.2255541682,-0.1211581826,0.1216320693,0.2576985657,0.0417754911,0.1188103035,-0.155726254,-0.0532568693,0.3186100423,0.0825055987,-0.1304463148,-0.2680415809,-0.1397158355,0.1819897592,0.0978391692,-0.3734224141,-0.0446045958,0.0930457115,0.0255390443,-0.0383036286,-0.2225391865,0.0113506401,0.0509980023,0.2269794792,0.0419547632,0.1949174255,-0.010965066,0.1166931838,-0.2026650012,-0.0816119686,-0.1552953571,0.0546444021,-0.2456047535,0.5321156979,-0.3866825104,0.5693590641,-0.4407524168,0.1738172621,0.2310239524,-0.4541364014,0.0961408019,-0.0945509821,-0.1327079237,-0.1259726882,0.0628547668,0.3869289458,-0.1089815944,0.1471091509,0.1429865211,-0.3117977381,0.3475709558,-0.0603820272,-0.2231802791,-0.0993922278,0.2244294137,0.1183510944,0.3281213641,-0.3093108833,-0.0371792205,0.4952086508,-0.1033602804,0.2260007709,0.1840551496,-0.3150682747,0.0563162304,-0.1813354194,0.142545417,0.0503486358,0.2422543019,-0.2390481234,-0.477942735]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Can you give us more information on your os and pip environments (pip list)?","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":14,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nCan you give us more information on your os and pip environments (pip list)?","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"@thomwolf Sure. I'll try downgrading to 3.7 now even though Arrow say they support >=3.5.\r\n\r\nLinux (Ubuntu 18.04) - Python 3.8\r\n======================\r\nPackage - Version\r\n---------------------\r\ncertifi 2020.6.20\r\nchardet 3.0.4\r\nclick 7.1.2\r\ndatasets 1.0.1\r\ndill 0.3.2\r\nfasttext 0.9.2\r\nfilelock 3.0.12\r\nfuture 0.18.2\r\nidna 2.10\r\njoblib 0.16.0\r\nnltk 3.5\r\nnumpy 1.19.1\r\npackaging 20.4\r\npandas 1.1.2\r\npip 20.0.2\r\nprotobuf 3.13.0\r\npyarrow 1.0.1\r\npybind11 2.5.0\r\npyparsing 2.4.7\r\npython-dateutil 2.8.1\r\npytz 2020.1\r\nregex 2020.7.14\r\nrequests 2.24.0\r\nsacremoses 0.0.43\r\nscikit-learn 0.23.2\r\nscipy 1.5.2\r\nsentence-transformers 0.3.6\r\nsentencepiece 0.1.91\r\nsetuptools 46.1.3\r\nsix 1.15.0\r\nstanza 1.1.1\r\nthreadpoolctl 2.1.0\r\ntokenizers 0.8.1rc2\r\ntorch 1.6.0+cu101\r\ntqdm 4.48.2\r\ntransformers 3.1.0\r\nurllib3 1.25.10\r\nwheel 0.34.2\r\nxxhash 2.0.0\r\n\r\nWindows 10 - Python 3.8\r\n================\r\nPackage - Version\r\n----------------------------\r\ncertifi 2020.6.20\r\nchardet 3.0.4\r\nclick 7.1.2\r\ndatasets 1.0.1\r\ndill 0.3.2\r\nfasttext 0.9.2\r\nfilelock 3.0.12\r\nfuture 0.18.2\r\nidna 2.10\r\njoblib 0.16.0\r\nnlp 0.4.0\r\nnltk 3.5\r\nnumpy 1.19.1\r\npackaging 20.4\r\npandas 1.1.1\r\npip 20.0.2\r\nprotobuf 3.13.0\r\npyarrow 1.0.1\r\npybind11 2.5.0\r\npyparsing 2.4.7\r\npython-dateutil 2.8.1\r\npytz 2020.1\r\nregex 2020.7.14\r\nrequests 2.24.0\r\nsacremoses 0.0.43\r\nscikit-learn 0.23.2\r\nscipy 1.5.2\r\nsentence-transformers 0.3.5.1\r\nsentencepiece 0.1.91\r\nsetuptools 46.1.3\r\nsix 1.15.0\r\nstanza 1.1.1\r\nthreadpoolctl 2.1.0\r\ntokenizers 0.8.1rc1\r\ntorch 1.6.0+cu101\r\ntqdm 4.48.2\r\ntransformers 3.0.2\r\nurllib3 1.25.10\r\nwheel 0.34.2\r\nxxhash 2.0.0","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":194,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n@thomwolf Sure. I'll try downgrading to 3.7 now even though Arrow say they support >=3.5.\r\n\r\nLinux (Ubuntu 18.04) - Python 3.8\r\n======================\r\nPackage - Version\r\n---------------------\r\ncertifi 2020.6.20\r\nchardet 3.0.4\r\nclick 7.1.2\r\ndatasets 1.0.1\r\ndill 0.3.2\r\nfasttext 0.9.2\r\nfilelock 3.0.12\r\nfuture 0.18.2\r\nidna 2.10\r\njoblib 0.16.0\r\nnltk 3.5\r\nnumpy 1.19.1\r\npackaging 20.4\r\npandas 1.1.2\r\npip 20.0.2\r\nprotobuf 3.13.0\r\npyarrow 1.0.1\r\npybind11 2.5.0\r\npyparsing 2.4.7\r\npython-dateutil 2.8.1\r\npytz 2020.1\r\nregex 2020.7.14\r\nrequests 2.24.0\r\nsacremoses 0.0.43\r\nscikit-learn 0.23.2\r\nscipy 1.5.2\r\nsentence-transformers 0.3.6\r\nsentencepiece 0.1.91\r\nsetuptools 46.1.3\r\nsix 1.15.0\r\nstanza 1.1.1\r\nthreadpoolctl 2.1.0\r\ntokenizers 0.8.1rc2\r\ntorch 1.6.0+cu101\r\ntqdm 4.48.2\r\ntransformers 3.1.0\r\nurllib3 1.25.10\r\nwheel 0.34.2\r\nxxhash 2.0.0\r\n\r\nWindows 10 - Python 3.8\r\n================\r\nPackage - Version\r\n----------------------------\r\ncertifi 2020.6.20\r\nchardet 3.0.4\r\nclick 7.1.2\r\ndatasets 1.0.1\r\ndill 0.3.2\r\nfasttext 0.9.2\r\nfilelock 3.0.12\r\nfuture 0.18.2\r\nidna 2.10\r\njoblib 0.16.0\r\nnlp 0.4.0\r\nnltk 3.5\r\nnumpy 1.19.1\r\npackaging 20.4\r\npandas 1.1.1\r\npip 20.0.2\r\nprotobuf 3.13.0\r\npyarrow 1.0.1\r\npybind11 2.5.0\r\npyparsing 2.4.7\r\npython-dateutil 2.8.1\r\npytz 2020.1\r\nregex 2020.7.14\r\nrequests 2.24.0\r\nsacremoses 0.0.43\r\nscikit-learn 0.23.2\r\nscipy 1.5.2\r\nsentence-transformers 0.3.5.1\r\nsentencepiece 0.1.91\r\nsetuptools 46.1.3\r\nsix 1.15.0\r\nstanza 1.1.1\r\nthreadpoolctl 2.1.0\r\ntokenizers 0.8.1rc1\r\ntorch 1.6.0+cu101\r\ntqdm 4.48.2\r\ntransformers 3.0.2\r\nurllib3 1.25.10\r\nwheel 0.34.2\r\nxxhash 2.0.0","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Downgrading to 3.7 does not help. Here is a dummy text file:\r\n\r\n```text\r\nVerzekering weigert vaker te betalen\r\nBedrijven van verzekeringen erkennen steeds minder arbeidsongevallen .\r\nIn 2012 weigerden de bedrijven te betalen voor 21.055 ongevallen op het werk .\r\nDat is 11,8 % van alle ongevallen op het werk .\r\nNog nooit weigerden verzekeraars zoveel zaken .\r\nIn 2012 hadden 135.118 mensen een ongeval op het werk .\r\nDat zijn elke werkdag 530 mensen .\r\nBij die ongevallen stierven 67 mensen .\r\nBijna 12.000 hebben een handicap na het ongeval .\r\nGeen echt arbeidsongeval Bedrijven moeten een verzekering hebben voor hun werknemers .\r\n```\r\n\r\nA temporary work around for the \"text\" type, is\r\n\r\n```python\r\ndataset = Dataset.from_dict({\"text\": Path(dataset_f).read_text().splitlines()})\r\n```","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":120,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nDowngrading to 3.7 does not help. Here is a dummy text file:\r\n\r\n```text\r\nVerzekering weigert vaker te betalen\r\nBedrijven van verzekeringen erkennen steeds minder arbeidsongevallen .\r\nIn 2012 weigerden de bedrijven te betalen voor 21.055 ongevallen op het werk .\r\nDat is 11,8 % van alle ongevallen op het werk .\r\nNog nooit weigerden verzekeraars zoveel zaken .\r\nIn 2012 hadden 135.118 mensen een ongeval op het werk .\r\nDat zijn elke werkdag 530 mensen .\r\nBij die ongevallen stierven 67 mensen .\r\nBijna 12.000 hebben een handicap na het ongeval .\r\nGeen echt arbeidsongeval Bedrijven moeten een verzekering hebben voor hun werknemers .\r\n```\r\n\r\nA temporary work around for the \"text\" type, is\r\n\r\n```python\r\ndataset = Dataset.from_dict({\"text\": Path(dataset_f).read_text().splitlines()})\r\n```","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"![image](https:\/\/user-images.githubusercontent.com\/6847024\/92997714-d2add900-f532-11ea-83d4-e3473c2d94d7.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/6847024\/92997724-e22d2200-f532-11ea-951d-b1d8f4582ea3.png)\r\neven i am facing the same issue.","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":9,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n![image](https:\/\/user-images.githubusercontent.com\/6847024\/92997714-d2add900-f532-11ea-83d4-e3473c2d94d7.png)\r\n![image](https:\/\/user-images.githubusercontent.com\/6847024\/92997724-e22d2200-f532-11ea-951d-b1d8f4582ea3.png)\r\neven i am facing the same issue.","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"@banunitte Please do not post screenshots in the future but copy-paste your code and the errors. That allows others to copy-and-paste your code and test it. You may also want to provide the Python version that you are using.","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":39,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n@banunitte Please do not post screenshots in the future but copy-paste your code and the errors. That allows others to copy-and-paste your code and test it. You may also want to provide the Python version that you are using.","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"I have the exact same problem in Windows 10, Python 3.8.\r\n","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":11,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nI have the exact same problem in Windows 10, Python 3.8.\r\n","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"I have the same problem on Linux of the script crashing with a CSV error. This may be caused by 'CRLF', when changed 'CRLF' to 'LF', the problem solved.","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":29,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nI have the same problem on Linux of the script crashing with a CSV error. This may be caused by 'CRLF', when changed 'CRLF' to 'LF', the problem solved.","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"I pushed a fix for `pyarrow.lib.ArrowInvalid: CSV parse error`. Let me know if you still have this issue.\r\n\r\nNot sure about the windows one yet","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":25,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nI pushed a fix for `pyarrow.lib.ArrowInvalid: CSV parse error`. Let me know if you still have this issue.\r\n\r\nNot sure about the windows one yet","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"To complete what @lhoestq is saying, I think that to use the new version of the `text` processing script (which is on master right now) you need to either specify the version of the script to be the `master` one or to install the lib from source (in which case it uses the `master` version of the script by default):\r\n```python\r\ndataset = load_dataset('text', script_version='master', data_files=XXX)\r\n```\r\nWe do versioning by default, i.e. your version of the dataset lib will use the script with the same version by default (i.e. only the `1.0.1` version of the script if you have the PyPI version `1.0.1` of the lib).","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":107,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nTo complete what @lhoestq is saying, I think that to use the new version of the `text` processing script (which is on master right now) you need to either specify the version of the script to be the `master` one or to install the lib from source (in which case it uses the `master` version of the script by default):\r\n```python\r\ndataset = load_dataset('text', script_version='master', data_files=XXX)\r\n```\r\nWe do versioning by default, i.e. your version of the dataset lib will use the script with the same version by default (i.e. only the `1.0.1` version of the script if you have the PyPI version `1.0.1` of the lib).","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"![image](https:\/\/user-images.githubusercontent.com\/36957508\/93300760-fa9a8680-f829-11ea-9105-7a6f67ad8373.png)\r\nwin10, py3.6\r\n\r\n\r\n```\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\n\r\nfeatures = Features({'text': Value('string'), 'ctext': Value('string')})\r\nfile_dict = {'train': PATH\/'summary.csv'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, script_version='master', delimiter='\\t', column_names=['text', 'ctext'], features=features)\r\n```","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":31,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n![image](https:\/\/user-images.githubusercontent.com\/36957508\/93300760-fa9a8680-f829-11ea-9105-7a6f67ad8373.png)\r\nwin10, py3.6\r\n\r\n\r\n```\r\nfrom datasets import Features, Value, ClassLabel, load_dataset\r\n\r\n\r\nfeatures = Features({'text': Value('string'), 'ctext': Value('string')})\r\nfile_dict = {'train': PATH\/'summary.csv'}\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, script_version='master', delimiter='\\t', column_names=['text', 'ctext'], features=features)\r\n```","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"```python\r\nTraceback` (most recent call last):\r\n File \"main.py\", line 281, in <module>\r\n main()\r\n File \"main.py\", line 190, in main\r\n train_data, test_data = data_factory(\r\n File \"main.py\", line 129, in data_factory\r\n train_data = load_dataset('text', \r\n File \"\/home\/me\/Downloads\/datasets\/src\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/me\/Downloads\/datasets\/src\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/me\/Downloads\/datasets\/src\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/me\/Downloads\/datasets\/src\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/me\/.local\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/me\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7fcc649178b014\/text.py\", line 103, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 617, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 123, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nUnfortunately i am still getting this issue on Linux. I installed datasets from source and specified script_version to master.\r\n\r\n","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":135,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n```python\r\nTraceback` (most recent call last):\r\n File \"main.py\", line 281, in <module>\r\n main()\r\n File \"main.py\", line 190, in main\r\n train_data, test_data = data_factory(\r\n File \"main.py\", line 129, in data_factory\r\n train_data = load_dataset('text', \r\n File \"\/home\/me\/Downloads\/datasets\/src\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/me\/Downloads\/datasets\/src\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/me\/Downloads\/datasets\/src\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/me\/Downloads\/datasets\/src\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/me\/.local\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/me\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7fcc649178b014\/text.py\", line 103, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 617, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 123, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nUnfortunately i am still getting this issue on Linux. I installed datasets from source and specified script_version to master.\r\n\r\n","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"> ![image](https:\/\/user-images.githubusercontent.com\/36957508\/93300760-fa9a8680-f829-11ea-9105-7a6f67ad8373.png)\r\n> win10, py3.6\r\n> \r\n> ```\r\n> from datasets import Features, Value, ClassLabel, load_dataset\r\n> \r\n> \r\n> features = Features({'text': Value('string'), 'ctext': Value('string')})\r\n> file_dict = {'train': PATH\/'summary.csv'}\r\n> \r\n> dataset = load_dataset('csv', data_files=file_dict, script_version='master', delimiter='\\t', column_names=['text', 'ctext'], features=features)\r\n> ```\r\n\r\nSince #644 it should now work on windows @ScottishFold007 \r\n\r\n> Trying the following snippet, I get different problems on Linux and Windows.\r\n> \r\n> ```python\r\n> dataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n> # or \r\n> dataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n> ```\r\n>\r\n> Windows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n> \r\n> ```\r\n> Checking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\n> Found main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\n> Found specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\n> Found script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\n> Couldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\n> Found metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\n> Using custom data configuration default\r\n> ```\r\n\r\nSame for you @BramVanroy .\r\n\r\nNot sure about the one on linux though","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":184,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n> ![image](https:\/\/user-images.githubusercontent.com\/36957508\/93300760-fa9a8680-f829-11ea-9105-7a6f67ad8373.png)\r\n> win10, py3.6\r\n> \r\n> ```\r\n> from datasets import Features, Value, ClassLabel, load_dataset\r\n> \r\n> \r\n> features = Features({'text': Value('string'), 'ctext': Value('string')})\r\n> file_dict = {'train': PATH\/'summary.csv'}\r\n> \r\n> dataset = load_dataset('csv', data_files=file_dict, script_version='master', delimiter='\\t', column_names=['text', 'ctext'], features=features)\r\n> ```\r\n\r\nSince #644 it should now work on windows @ScottishFold007 \r\n\r\n> Trying the following snippet, I get different problems on Linux and Windows.\r\n> \r\n> ```python\r\n> dataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n> # or \r\n> dataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n> ```\r\n>\r\n> Windows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n> \r\n> ```\r\n> Checking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\n> Found main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\n> Found specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\n> Found script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\n> Couldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\n> Found metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\n> Using custom data configuration default\r\n> ```\r\n\r\nSame for you @BramVanroy .\r\n\r\nNot sure about the one on linux though","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"> To complete what @lhoestq is saying, I think that to use the new version of the `text` processing script (which is on master right now) you need to either specify the version of the script to be the `master` one or to install the lib from source (in which case it uses the `master` version of the script by default):\r\n> \r\n> ```python\r\n> dataset = load_dataset('text', script_version='master', data_files=XXX)\r\n> ```\r\n> \r\n> We do versioning by default, i.e. your version of the dataset lib will use the script with the same version by default (i.e. only the `1.0.1` version of the script if you have the PyPI version `1.0.1` of the lib).\r\n\r\nLinux here:\r\n\r\nI was using the 0.4.0 nlp library load_dataset to load a text dataset of 9-10Gb without collapsing the RAM memory. However, today I got the csv error message mentioned in this issue. After installing the new (datasets) library from source and specifying the script_verson = 'master' I'm still having this same error message. Furthermore, I cannot use the dictionary \"trick\" to load the dataset since the system kills the process due to a RAM out of memory problem. Is there any other solution to this error? Thank you in advance. ","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":206,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n> To complete what @lhoestq is saying, I think that to use the new version of the `text` processing script (which is on master right now) you need to either specify the version of the script to be the `master` one or to install the lib from source (in which case it uses the `master` version of the script by default):\r\n> \r\n> ```python\r\n> dataset = load_dataset('text', script_version='master', data_files=XXX)\r\n> ```\r\n> \r\n> We do versioning by default, i.e. your version of the dataset lib will use the script with the same version by default (i.e. only the `1.0.1` version of the script if you have the PyPI version `1.0.1` of the lib).\r\n\r\nLinux here:\r\n\r\nI was using the 0.4.0 nlp library load_dataset to load a text dataset of 9-10Gb without collapsing the RAM memory. However, today I got the csv error message mentioned in this issue. After installing the new (datasets) library from source and specifying the script_verson = 'master' I'm still having this same error message. Furthermore, I cannot use the dictionary \"trick\" to load the dataset since the system kills the process due to a RAM out of memory problem. Is there any other solution to this error? Thank you in advance. ","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Hi @raruidol \r\nTo fix the RAM issue you'll need to shard your text files into smaller files (see https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-691672919 for example)\r\n\r\nI'm not sure why you're having the csv error on linux.\r\nDo you think you could to to reproduce it on google colab for example ?\r\nOr send me a dummy .txt file that reproduces the issue ?","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":59,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nHi @raruidol \r\nTo fix the RAM issue you'll need to shard your text files into smaller files (see https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-691672919 for example)\r\n\r\nI'm not sure why you're having the csv error on linux.\r\nDo you think you could to to reproduce it on google colab for example ?\r\nOr send me a dummy .txt file that reproduces the issue ?","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"@lhoestq \r\n\r\nThe crash message shows up when loading the dataset:\r\n```\r\nprint('Loading corpus...') \r\nfiles = glob.glob('corpora\/shards\/*') \r\n-> dataset = load_dataset('text', script_version='master', data_files=files) \r\nprint('Corpus loaded.')\r\n```\r\nAnd this is the exact message:\r\n```\r\nTraceback (most recent call last):\r\n File \"run_language_modeling.py\", line 27, in <module>\r\n dataset = load_dataset('text', script_version='master', data_files=files)\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 471, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 548, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 892, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/jupyter-raruidol\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7fcc649178b014\/text.py\", line 107, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nAnd these are the pip packages I have atm and their versions:\r\n\r\n```\r\nPackage Version Location \r\n--------------- --------- -------------------------------------------------------------\r\ncertifi 2020.6.20 \r\nchardet 3.0.4 \r\nclick 7.1.2 \r\ndatasets 1.0.2 \r\ndill 0.3.2 \r\nfilelock 3.0.12 \r\nfuture 0.18.2 \r\nidna 2.10 \r\njoblib 0.16.0 \r\nnumpy 1.19.1 \r\npackaging 20.4 \r\npandas 1.1.1 \r\npip 19.0.3 \r\npyarrow 1.0.1 \r\npyparsing 2.4.7 \r\npython-dateutil 2.8.1 \r\npytz 2020.1 \r\nregex 2020.7.14 \r\nrequests 2.24.0 \r\nsacremoses 0.0.43 \r\nsentencepiece 0.1.91 \r\nsetuptools 40.8.0 \r\nsix 1.15.0 \r\ntokenizers 0.8.1rc2 \r\ntorch 1.6.0 \r\ntqdm 4.48.2 \r\ntransformers 3.0.2 \/home\/jupyter-raruidol\/DebatAnalyser\/env\/src\/transformers\/src\r\n```\r\n\r\n\r\n","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":207,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n@lhoestq \r\n\r\nThe crash message shows up when loading the dataset:\r\n```\r\nprint('Loading corpus...') \r\nfiles = glob.glob('corpora\/shards\/*') \r\n-> dataset = load_dataset('text', script_version='master', data_files=files) \r\nprint('Corpus loaded.')\r\n```\r\nAnd this is the exact message:\r\n```\r\nTraceback (most recent call last):\r\n File \"run_language_modeling.py\", line 27, in <module>\r\n dataset = load_dataset('text', script_version='master', data_files=files)\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 611, in load_dataset\r\n ignore_verifications=ignore_verifications,\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 471, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 548, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 892, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/jupyter-raruidol\/DebatAnalyser\/env\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/jupyter-raruidol\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7fcc649178b014\/text.py\", line 107, in _generate_tables\r\n convert_options=self.config.convert_options,\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nAnd these are the pip packages I have atm and their versions:\r\n\r\n```\r\nPackage Version Location \r\n--------------- --------- -------------------------------------------------------------\r\ncertifi 2020.6.20 \r\nchardet 3.0.4 \r\nclick 7.1.2 \r\ndatasets 1.0.2 \r\ndill 0.3.2 \r\nfilelock 3.0.12 \r\nfuture 0.18.2 \r\nidna 2.10 \r\njoblib 0.16.0 \r\nnumpy 1.19.1 \r\npackaging 20.4 \r\npandas 1.1.1 \r\npip 19.0.3 \r\npyarrow 1.0.1 \r\npyparsing 2.4.7 \r\npython-dateutil 2.8.1 \r\npytz 2020.1 \r\nregex 2020.7.14 \r\nrequests 2.24.0 \r\nsacremoses 0.0.43 \r\nsentencepiece 0.1.91 \r\nsetuptools 40.8.0 \r\nsix 1.15.0 \r\ntokenizers 0.8.1rc2 \r\ntorch 1.6.0 \r\ntqdm 4.48.2 \r\ntransformers 3.0.2 \/home\/jupyter-raruidol\/DebatAnalyser\/env\/src\/transformers\/src\r\n```\r\n\r\n\r\n","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"I tested on google colab which is also linux using this code:\r\n\r\n- first download an arbitrary text file\r\n```bash\r\nwget https:\/\/raw.githubusercontent.com\/abisee\/cnn-dailymail\/master\/url_lists\/all_train.txt\r\n```\r\n- then run\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nd = load_dataset(\"text\", data_files=\"all_train.txt\", script_version='master')\r\n```\r\nAnd I don't get this issue.\r\n\r\n\\> Could you test on your side if these lines work @raruidol ?\r\n\r\nalso cc @Skyy93 as it seems you have the same issue\r\n\r\nIf it works:\r\nIt could mean that the issue could come from unexpected patterns in the files you want to use.\r\nIn that case we should find a way to handle them.\r\n\r\nAnd if it doesn't work:\r\nIt could mean that it comes from the way pyarrow reads text files on linux.\r\nIn that case we should report it to pyarrow and find a workaround in the meantime\r\n\r\nEither way it should help to find where this bug comes from and fix it :)\r\n\r\nThank you in advance !","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":156,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nI tested on google colab which is also linux using this code:\r\n\r\n- first download an arbitrary text file\r\n```bash\r\nwget https:\/\/raw.githubusercontent.com\/abisee\/cnn-dailymail\/master\/url_lists\/all_train.txt\r\n```\r\n- then run\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nd = load_dataset(\"text\", data_files=\"all_train.txt\", script_version='master')\r\n```\r\nAnd I don't get this issue.\r\n\r\n\\> Could you test on your side if these lines work @raruidol ?\r\n\r\nalso cc @Skyy93 as it seems you have the same issue\r\n\r\nIf it works:\r\nIt could mean that the issue could come from unexpected patterns in the files you want to use.\r\nIn that case we should find a way to handle them.\r\n\r\nAnd if it doesn't work:\r\nIt could mean that it comes from the way pyarrow reads text files on linux.\r\nIn that case we should report it to pyarrow and find a workaround in the meantime\r\n\r\nEither way it should help to find where this bug comes from and fix it :)\r\n\r\nThank you in advance !","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Update: also tested the above code in a docker container from [jupyter\/minimal-notebook](https:\/\/hub.docker.com\/r\/jupyter\/minimal-notebook\/) (based on ubuntu) and still not able to reproduce","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":21,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nUpdate: also tested the above code in a docker container from [jupyter\/minimal-notebook](https:\/\/hub.docker.com\/r\/jupyter\/minimal-notebook\/) (based on ubuntu) and still not able to reproduce","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"It looks like with your text input file works without any problem. I have been doing some experiments this morning with my input files and I'm almost certain that the crash is caused by some unexpected pattern in the files. However, I've not been able to spot the main cause of it. What I find strange is that this same corpus was being loaded by the nlp 0.4.0 library without any problem... Where can I find the code where you structure the input text data in order to use it with pyarrow?","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":92,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nIt looks like with your text input file works without any problem. I have been doing some experiments this morning with my input files and I'm almost certain that the crash is caused by some unexpected pattern in the files. However, I've not been able to spot the main cause of it. What I find strange is that this same corpus was being loaded by the nlp 0.4.0 library without any problem... Where can I find the code where you structure the input text data in order to use it with pyarrow?","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Under the hood it does\r\n```python\r\nimport pyarrow as pa\r\nimport pyarrow.csv\r\n\r\n# Use csv reader from Pyarrow with one column for text files\r\n\r\n# To force the one-column setting, we set an arbitrary character\r\n# that is not in text files as delimiter, such as \\b or \\v.\r\n# The bell character, \\b, was used to make beeps back in the days\r\nparse_options = pa.csv.ParseOptions( \r\n delimiter=\"\\b\", \r\n quote_char=False, \r\n double_quote=False, \r\n escape_char=False, \r\n newlines_in_values=False, \r\n ignore_empty_lines=False, \r\n)\r\n\r\nread_options= pa.csv.ReadOptions(use_threads=True, column_names=[\"text\"])\r\n\r\npa_table = pa.csv.read_csv(\"all_train.txt\", read_options=read_options, parse_options=parse_options)\r\n```\r\n\r\nNote that we changed the parse options with datasets 1.0\r\nIn particular the delimiter used to be `\\r` but this delimiter doesn't work on windows.","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":107,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nUnder the hood it does\r\n```python\r\nimport pyarrow as pa\r\nimport pyarrow.csv\r\n\r\n# Use csv reader from Pyarrow with one column for text files\r\n\r\n# To force the one-column setting, we set an arbitrary character\r\n# that is not in text files as delimiter, such as \\b or \\v.\r\n# The bell character, \\b, was used to make beeps back in the days\r\nparse_options = pa.csv.ParseOptions( \r\n delimiter=\"\\b\", \r\n quote_char=False, \r\n double_quote=False, \r\n escape_char=False, \r\n newlines_in_values=False, \r\n ignore_empty_lines=False, \r\n)\r\n\r\nread_options= pa.csv.ReadOptions(use_threads=True, column_names=[\"text\"])\r\n\r\npa_table = pa.csv.read_csv(\"all_train.txt\", read_options=read_options, parse_options=parse_options)\r\n```\r\n\r\nNote that we changed the parse options with datasets 1.0\r\nIn particular the delimiter used to be `\\r` but this delimiter doesn't work on windows.","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Could you try with `\\a` instead of `\\b` ? It looks like the bell character is \\a in python and not \\b","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":22,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nCould you try with `\\a` instead of `\\b` ? It looks like the bell character is \\a in python and not \\b","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"I was just exploring if the crash was happening in every shard or not, and which shards were generating the error message. With \\b I got the following list of shards crashing:\r\n\r\n```\r\nErrors on files: ['corpora\/shards\/shard_0069', 'corpora\/shards\/shard_0043', 'corpora\/shards\/shard_0014', 'corpora\/shards\/shard_0032', 'corpora\/shards\/shard_0088', 'corpora\/shards\/shard_0018', 'corpora\/shards\/shard_0073', 'corpora\/shards\/shard_0079', 'corpora\/shards\/shard_0038', 'corpora\/shards\/shard_0041', 'corpora\/shards\/shard_0007', 'corpora\/shards\/shard_0004', 'corpora\/shards\/shard_0102', 'corpora\/shards\/shard_0096', 'corpora\/shards\/shard_0030', 'corpora\/shards\/shard_0076', 'corpora\/shards\/shard_0067', 'corpora\/shards\/shard_0052', 'corpora\/shards\/shard_0026', 'corpora\/shards\/shard_0024', 'corpora\/shards\/shard_0064', 'corpora\/shards\/shard_0044', 'corpora\/shards\/shard_0013', 'corpora\/shards\/shard_0062', 'corpora\/shards\/shard_0057', 'corpora\/shards\/shard_0097', 'corpora\/shards\/shard_0094', 'corpora\/shards\/shard_0078', 'corpora\/shards\/shard_0075', 'corpora\/shards\/shard_0039', 'corpora\/shards\/shard_0077', 'corpora\/shards\/shard_0021', 'corpora\/shards\/shard_0040', 'corpora\/shards\/shard_0009', 'corpora\/shards\/shard_0023', 'corpora\/shards\/shard_0095', 'corpora\/shards\/shard_0107', 'corpora\/shards\/shard_0063', 'corpora\/shards\/shard_0086', 'corpora\/shards\/shard_0047', 'corpora\/shards\/shard_0089', 'corpora\/shards\/shard_0037', 'corpora\/shards\/shard_0101', 'corpora\/shards\/shard_0093', 'corpora\/shards\/shard_0082', 'corpora\/shards\/shard_0091', 'corpora\/shards\/shard_0065', 'corpora\/shards\/shard_0020', 'corpora\/shards\/shard_0070', 'corpora\/shards\/shard_0008', 'corpora\/shards\/shard_0058', 'corpora\/shards\/shard_0060', 'corpora\/shards\/shard_0022', 'corpora\/shards\/shard_0059', 'corpora\/shards\/shard_0100', 'corpora\/shards\/shard_0027', 'corpora\/shards\/shard_0072', 'corpora\/shards\/shard_0098', 'corpora\/shards\/shard_0019', 'corpora\/shards\/shard_0066', 'corpora\/shards\/shard_0042', 'corpora\/shards\/shard_0053']\r\n```\r\n\r\nI also tried with \\a and the list decreased but there were still several crashes:\r\n\r\n```\r\nErrors on files: ['corpora\/shards\/shard_0069', 'corpora\/shards\/shard_0055', 'corpora\/shards\/shard_0043', 'corpora\/shards\/shard_0014', 'corpora\/shards\/shard_0073', 'corpora\/shards\/shard_0025', 'corpora\/shards\/shard_0068', 'corpora\/shards\/shard_0102', 'corpora\/shards\/shard_0096', 'corpora\/shards\/shard_0076', 'corpora\/shards\/shard_0067', 'corpora\/shards\/shard_0026', 'corpora\/shards\/shard_0024', 'corpora\/shards\/shard_0044', 'corpora\/shards\/shard_0087', 'corpora\/shards\/shard_0092', 'corpora\/shards\/shard_0074', 'corpora\/shards\/shard_0094', 'corpora\/shards\/shard_0078', 'corpora\/shards\/shard_0039', 'corpora\/shards\/shard_0077', 'corpora\/shards\/shard_0040', 'corpora\/shards\/shard_0009', 'corpora\/shards\/shard_0107', 'corpora\/shards\/shard_0063', 'corpora\/shards\/shard_0103', 'corpora\/shards\/shard_0047', 'corpora\/shards\/shard_0033', 'corpora\/shards\/shard_0089', 'corpora\/shards\/shard_0037', 'corpora\/shards\/shard_0082', 'corpora\/shards\/shard_0071', 'corpora\/shards\/shard_0091', 'corpora\/shards\/shard_0065', 'corpora\/shards\/shard_0070', 'corpora\/shards\/shard_0058', 'corpora\/shards\/shard_0081', 'corpora\/shards\/shard_0060', 'corpora\/shards\/shard_0002', 'corpora\/shards\/shard_0059', 'corpora\/shards\/shard_0027', 'corpora\/shards\/shard_0072', 'corpora\/shards\/shard_0098', 'corpora\/shards\/shard_0019', 'corpora\/shards\/shard_0045', 'corpora\/shards\/shard_0036', 'corpora\/shards\/shard_0066', 'corpora\/shards\/shard_0053']\r\n```\r\n\r\nWhich means that it is quite possible that the assumption of that some unexpected pattern in the files is causing the crashes is true. If I am able to reach any conclusion I will post It here asap.","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":205,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nI was just exploring if the crash was happening in every shard or not, and which shards were generating the error message. With \\b I got the following list of shards crashing:\r\n\r\n```\r\nErrors on files: ['corpora\/shards\/shard_0069', 'corpora\/shards\/shard_0043', 'corpora\/shards\/shard_0014', 'corpora\/shards\/shard_0032', 'corpora\/shards\/shard_0088', 'corpora\/shards\/shard_0018', 'corpora\/shards\/shard_0073', 'corpora\/shards\/shard_0079', 'corpora\/shards\/shard_0038', 'corpora\/shards\/shard_0041', 'corpora\/shards\/shard_0007', 'corpora\/shards\/shard_0004', 'corpora\/shards\/shard_0102', 'corpora\/shards\/shard_0096', 'corpora\/shards\/shard_0030', 'corpora\/shards\/shard_0076', 'corpora\/shards\/shard_0067', 'corpora\/shards\/shard_0052', 'corpora\/shards\/shard_0026', 'corpora\/shards\/shard_0024', 'corpora\/shards\/shard_0064', 'corpora\/shards\/shard_0044', 'corpora\/shards\/shard_0013', 'corpora\/shards\/shard_0062', 'corpora\/shards\/shard_0057', 'corpora\/shards\/shard_0097', 'corpora\/shards\/shard_0094', 'corpora\/shards\/shard_0078', 'corpora\/shards\/shard_0075', 'corpora\/shards\/shard_0039', 'corpora\/shards\/shard_0077', 'corpora\/shards\/shard_0021', 'corpora\/shards\/shard_0040', 'corpora\/shards\/shard_0009', 'corpora\/shards\/shard_0023', 'corpora\/shards\/shard_0095', 'corpora\/shards\/shard_0107', 'corpora\/shards\/shard_0063', 'corpora\/shards\/shard_0086', 'corpora\/shards\/shard_0047', 'corpora\/shards\/shard_0089', 'corpora\/shards\/shard_0037', 'corpora\/shards\/shard_0101', 'corpora\/shards\/shard_0093', 'corpora\/shards\/shard_0082', 'corpora\/shards\/shard_0091', 'corpora\/shards\/shard_0065', 'corpora\/shards\/shard_0020', 'corpora\/shards\/shard_0070', 'corpora\/shards\/shard_0008', 'corpora\/shards\/shard_0058', 'corpora\/shards\/shard_0060', 'corpora\/shards\/shard_0022', 'corpora\/shards\/shard_0059', 'corpora\/shards\/shard_0100', 'corpora\/shards\/shard_0027', 'corpora\/shards\/shard_0072', 'corpora\/shards\/shard_0098', 'corpora\/shards\/shard_0019', 'corpora\/shards\/shard_0066', 'corpora\/shards\/shard_0042', 'corpora\/shards\/shard_0053']\r\n```\r\n\r\nI also tried with \\a and the list decreased but there were still several crashes:\r\n\r\n```\r\nErrors on files: ['corpora\/shards\/shard_0069', 'corpora\/shards\/shard_0055', 'corpora\/shards\/shard_0043', 'corpora\/shards\/shard_0014', 'corpora\/shards\/shard_0073', 'corpora\/shards\/shard_0025', 'corpora\/shards\/shard_0068', 'corpora\/shards\/shard_0102', 'corpora\/shards\/shard_0096', 'corpora\/shards\/shard_0076', 'corpora\/shards\/shard_0067', 'corpora\/shards\/shard_0026', 'corpora\/shards\/shard_0024', 'corpora\/shards\/shard_0044', 'corpora\/shards\/shard_0087', 'corpora\/shards\/shard_0092', 'corpora\/shards\/shard_0074', 'corpora\/shards\/shard_0094', 'corpora\/shards\/shard_0078', 'corpora\/shards\/shard_0039', 'corpora\/shards\/shard_0077', 'corpora\/shards\/shard_0040', 'corpora\/shards\/shard_0009', 'corpora\/shards\/shard_0107', 'corpora\/shards\/shard_0063', 'corpora\/shards\/shard_0103', 'corpora\/shards\/shard_0047', 'corpora\/shards\/shard_0033', 'corpora\/shards\/shard_0089', 'corpora\/shards\/shard_0037', 'corpora\/shards\/shard_0082', 'corpora\/shards\/shard_0071', 'corpora\/shards\/shard_0091', 'corpora\/shards\/shard_0065', 'corpora\/shards\/shard_0070', 'corpora\/shards\/shard_0058', 'corpora\/shards\/shard_0081', 'corpora\/shards\/shard_0060', 'corpora\/shards\/shard_0002', 'corpora\/shards\/shard_0059', 'corpora\/shards\/shard_0027', 'corpora\/shards\/shard_0072', 'corpora\/shards\/shard_0098', 'corpora\/shards\/shard_0019', 'corpora\/shards\/shard_0045', 'corpora\/shards\/shard_0036', 'corpora\/shards\/shard_0066', 'corpora\/shards\/shard_0053']\r\n```\r\n\r\nWhich means that it is quite possible that the assumption of that some unexpected pattern in the files is causing the crashes is true. If I am able to reach any conclusion I will post It here asap.","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Hmmm I was expecting it to work with \\a, not sure why they appear in your text files though","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":19,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nHmmm I was expecting it to work with \\a, not sure why they appear in your text files though","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Hi @lhoestq, is there any input length restriction which was not before the update of the nlp library?","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":18,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nHi @lhoestq, is there any input length restriction which was not before the update of the nlp library?","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"No we never set any input length restriction on our side (maybe arrow but I don't think so)","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":18,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nNo we never set any input length restriction on our side (maybe arrow but I don't think so)","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"@lhoestq Can you ever be certain that a delimiter character is not present in a plain text file? In other formats (e.g. CSV) , rules are set of what is allowed and what isn't so that it actually constitutes a CSV file. In a text file you basically have \"anything goes\", so I don't think you can ever be entirely sure that the chosen delimiter does not exist in the text file, or am I wrong? \r\n\r\nIf I understand correctly you choose a delimiter that we hope does not exist in the file, so that when the CSV parser starts splitting into columns, it will only ever create one column? Why can't we use a newline character though?","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":118,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n@lhoestq Can you ever be certain that a delimiter character is not present in a plain text file? In other formats (e.g. CSV) , rules are set of what is allowed and what isn't so that it actually constitutes a CSV file. In a text file you basically have \"anything goes\", so I don't think you can ever be entirely sure that the chosen delimiter does not exist in the text file, or am I wrong? \r\n\r\nIf I understand correctly you choose a delimiter that we hope does not exist in the file, so that when the CSV parser starts splitting into columns, it will only ever create one column? Why can't we use a newline character though?","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"Okay, I have splitted the crashing shards into individual sentences and some examples of the inputs that are causing the crashes are the following ones:\r\n\r\n\r\n_4.\u2003DE L\u2019ORGANITZACI\u00d3 ESTAMENTAL A L\u2019ORGANITZACI\u00d3 EN CLASSES A mesura que es desenvolupava un sistema econ\u00f2mic capitalista i naixia una classe burgesa cada vegada m\u00e9s preparada per a substituir els dirigents de les velles monarquies absolutistes, es q\u00fcestionava l\u2019abund\u00e0ncia de b\u00e9ns amortitzats, que com s\u2019ha dit estaven fora del mercat i no pagaven tributs, pels perjudicis que ocasionaven a les finances p\u00fabliques i a l\u2019economia en general. Aquest estat d\u2019opini\u00f3 revolucionari va desembocar en un conjunt de mesures pr\u00e0ctiques de car\u00e0cter liberal. D\u2019una banda, les que intentaven desposseir les mans mortes del domini de b\u00e9ns acumulats, proc\u00e9s que acostumem a denominar desamortitzaci\u00f3, i que no \u00e9s m\u00e9s que la nacionalitzaci\u00f3 i venda d\u2019aquests b\u00e9ns eclesi\u00e0stics o civils en subhasta p\u00fablica al millor postor. D\u2019altra banda, les que redimien o redu\u00efen els censos i delmes o aixecaven les prohibicions de venda, \u00e9s a dir, les vinculacions. La desamortitzaci\u00f3, que va afectar b\u00e9ns dels ordes religiosos, dels pobles i d\u2019algunes corporacions civils, no va ser un cam\u00ed f\u00e0cil, perqu\u00e8 costava i costa trobar alg\u00fa que sigui indiferent a la p\u00e8rdua de b\u00e9ns, drets i privilegis. I t\u00e9 una gran transcend\u00e8ncia, va privar els antics estaments de les Espanyes, clero i pobles \u2014la noblesa en queda al marge\u2014, de la for\u00e7a econ\u00f2mica que els donaven bona part de les seves terres i, en \u00faltima inst\u00e0ncia, va preparar el terreny per a la substituci\u00f3 de la vella societat estamental per la nova societat classista. En aquesta societat, en teoria, les agrupacions socials s\u00f3n obertes, no tenen cap estatut jur\u00eddic privilegiat i estan definides per la possessi\u00f3 o no d\u2019uns b\u00e9ns econ\u00f2mics que s\u00f3n lliurement alienables. A les Espanyes la transformaci\u00f3 va afectar poc l\u2019aristocr\u00e0cia latifundista, all\u00e0 on n\u2019hi havia. Aquesta situaci\u00f3 va afavorir, en part, la persist\u00e8ncia de la vella cultura de la societat estamental en determinats ambients, i aix\u00f2 ha influ\u00eft decisivament en la manca de democr\u00e0cia que caracteritza la majoria de r\u00e8gims pol\u00edtics que s\u2019han anat succeint. Una manera de pensar que sempre sura en un moment o altre, i que de fet no acaba de desapar\u00e8ixer del tot. 5.\u2003INICI DE LA DESAMORTITZACI\u00d3 A LES ESPANYES Durant el segle xviii, dins d\u2019aquesta visi\u00f3 lliberal, va agafar for\u00e7a en alguns cercles de les Espanyes el corrent d\u2019opini\u00f3 contrari a les mans mortes. Durant el regnat de Carles III, s\u2019arbitraren les primeres mesures desamortitzadores proposades per alguns ministres il\u00b7lustrats. Aquestes disposicions foren modestes i poc eficaces, no van aturar l\u2019acumulaci\u00f3 de terres per part dels estaments que constitu\u00efen les mans mortes i varen afectar principalment b\u00e9ns dels pobles. L\u2019Esgl\u00e9sia no va ser tocada, excepte en el cas de 110_\r\n\r\n_la revoluci\u00f3 liberal, perqu\u00e8, encara que havia perdut els seus drets jurisdiccionals, havia conservat la majoria de terres i fins i tot les havia incrementat amb d\u2019altres que procedien de la desamortitzaci\u00f3. En la nova situaci\u00f3, les mans mortes del bosc p\u00fablic eren l\u2019Estat, que no cerca mai l\u2019autofinan\u00e7ament de les despeses de gesti\u00f3; els diners que manquin ja els posar\u00e0 l\u2019Estat. 9.\u2003DEFENSA I INTENTS DE RECUPERACI\u00d3 DELS B\u00c9NS COMUNALS DESAMORTITZATS El proc\u00e9s de centralitzaci\u00f3 no era senzill, perqu\u00e8, d\u2019una banda, la nova organitzaci\u00f3 apartava de la gesti\u00f3 moltes corporacions locals i molts ve\u00efns que l\u2019havien portada des de l\u2019edat mitjana, i, de l\u2019altra, era dif\u00edcil de coordinar la nova silvicultura amb moltes pr\u00e0ctiques forestals i drets tradicionals, com la pastura, fer llenya o tallar un arbre aqu\u00ed i un altre all\u00e0 quan tenia el gruix suficient, les pr\u00e0ctiques que s\u2019havien fet sempre. Les primeres passes de la nova organitzaci\u00f3 centralitzada varen tenir moltes dificultats en aquells indrets en qu\u00e8 els terrenys municipals i comunals tenien un paper important en l\u2019economia local. La desobedi\u00e8ncia a determinades normes imposades varen prendre formes diferents. Algunes institucions, com, per exemple, la Diputaci\u00f3 de Lleida, varen retardar la tramitaci\u00f3 d\u2019alguns expedients i varen evitar la venda de b\u00e9ns municipals. Molts pobles permeteren deixar que els ve\u00efns continuessin amb les seves pr\u00e0ctiques tradicionals, d\u2019altres varen boicotejar les subhastes d\u2019aprofitaments. L\u2019Estat va reaccionar encomanant a la Gu\u00e0rdia Civil el compliment de les noves directrius. Imposar el nou r\u00e8gim va costar a l\u2019Administraci\u00f3 un grapat d\u2019anys, per\u00f2 de mica en mica, amb molta, molta guarderia i gens de negociaci\u00f3, ho va aconseguir. La nova gesti\u00f3 estatal dels b\u00e9ns municipals va deixar, com hem comentat, molta gent sense uns recursos necessaris per a la superviv\u00e8ncia, sobre tot en \u00e0rees on predominaven les grans propietats, i on els pagesos sense terra treballaven de jornalers temporers. Aix\u00f2 va afavorir que, a bona part de les Espanyes, les primeres lluites camperoles de la segona meitat del segle xix defensessin la recuperaci\u00f3 dels comunals desamortitzats; per a molts aquella expropiaci\u00f3 i venda dirigida pels governs mon\u00e0rquics era la causa de molta mis\u00e8ria. D\u2019altres, m\u00e9s radicalitzats, varen entendre que l\u2019eliminaci\u00f3 de la propietat col\u00b7lectiva i la gesti\u00f3 estatal dels boscos no desamortitzats suposava una usurpaci\u00f3 pura i dura. En les zones m\u00e9s afectades per la desamortitzaci\u00f3 aix\u00f2 va donar lloc a un imaginari centrat en la defensa del comunal. La Segona Rep\u00fablica va arribar en una conjuntura econ\u00f2mica de crisi, generada pel crac del 1929. Al camp, aquesta situaci\u00f3 va produir una forta caiguda dels preus dels productes agraris i un increment important de l\u2019atur. QUADERNS AGRARIS 42\u2002(juny 2017), p. 105-126_\r\n\r\nI think that the main difference between the crashing samples and the rest is their length. Therefore, couldn't the length be causing the message errors? I hope with these samples you can identify what is causing the crashes considering that the 0.4.0 nlp library was loading them properly.","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":949,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nOkay, I have splitted the crashing shards into individual sentences and some examples of the inputs that are causing the crashes are the following ones:\r\n\r\n\r\n_4.\u2003DE L\u2019ORGANITZACI\u00d3 ESTAMENTAL A L\u2019ORGANITZACI\u00d3 EN CLASSES A mesura que es desenvolupava un sistema econ\u00f2mic capitalista i naixia una classe burgesa cada vegada m\u00e9s preparada per a substituir els dirigents de les velles monarquies absolutistes, es q\u00fcestionava l\u2019abund\u00e0ncia de b\u00e9ns amortitzats, que com s\u2019ha dit estaven fora del mercat i no pagaven tributs, pels perjudicis que ocasionaven a les finances p\u00fabliques i a l\u2019economia en general. Aquest estat d\u2019opini\u00f3 revolucionari va desembocar en un conjunt de mesures pr\u00e0ctiques de car\u00e0cter liberal. D\u2019una banda, les que intentaven desposseir les mans mortes del domini de b\u00e9ns acumulats, proc\u00e9s que acostumem a denominar desamortitzaci\u00f3, i que no \u00e9s m\u00e9s que la nacionalitzaci\u00f3 i venda d\u2019aquests b\u00e9ns eclesi\u00e0stics o civils en subhasta p\u00fablica al millor postor. D\u2019altra banda, les que redimien o redu\u00efen els censos i delmes o aixecaven les prohibicions de venda, \u00e9s a dir, les vinculacions. La desamortitzaci\u00f3, que va afectar b\u00e9ns dels ordes religiosos, dels pobles i d\u2019algunes corporacions civils, no va ser un cam\u00ed f\u00e0cil, perqu\u00e8 costava i costa trobar alg\u00fa que sigui indiferent a la p\u00e8rdua de b\u00e9ns, drets i privilegis. I t\u00e9 una gran transcend\u00e8ncia, va privar els antics estaments de les Espanyes, clero i pobles \u2014la noblesa en queda al marge\u2014, de la for\u00e7a econ\u00f2mica que els donaven bona part de les seves terres i, en \u00faltima inst\u00e0ncia, va preparar el terreny per a la substituci\u00f3 de la vella societat estamental per la nova societat classista. En aquesta societat, en teoria, les agrupacions socials s\u00f3n obertes, no tenen cap estatut jur\u00eddic privilegiat i estan definides per la possessi\u00f3 o no d\u2019uns b\u00e9ns econ\u00f2mics que s\u00f3n lliurement alienables. A les Espanyes la transformaci\u00f3 va afectar poc l\u2019aristocr\u00e0cia latifundista, all\u00e0 on n\u2019hi havia. Aquesta situaci\u00f3 va afavorir, en part, la persist\u00e8ncia de la vella cultura de la societat estamental en determinats ambients, i aix\u00f2 ha influ\u00eft decisivament en la manca de democr\u00e0cia que caracteritza la majoria de r\u00e8gims pol\u00edtics que s\u2019han anat succeint. Una manera de pensar que sempre sura en un moment o altre, i que de fet no acaba de desapar\u00e8ixer del tot. 5.\u2003INICI DE LA DESAMORTITZACI\u00d3 A LES ESPANYES Durant el segle xviii, dins d\u2019aquesta visi\u00f3 lliberal, va agafar for\u00e7a en alguns cercles de les Espanyes el corrent d\u2019opini\u00f3 contrari a les mans mortes. Durant el regnat de Carles III, s\u2019arbitraren les primeres mesures desamortitzadores proposades per alguns ministres il\u00b7lustrats. Aquestes disposicions foren modestes i poc eficaces, no van aturar l\u2019acumulaci\u00f3 de terres per part dels estaments que constitu\u00efen les mans mortes i varen afectar principalment b\u00e9ns dels pobles. L\u2019Esgl\u00e9sia no va ser tocada, excepte en el cas de 110_\r\n\r\n_la revoluci\u00f3 liberal, perqu\u00e8, encara que havia perdut els seus drets jurisdiccionals, havia conservat la majoria de terres i fins i tot les havia incrementat amb d\u2019altres que procedien de la desamortitzaci\u00f3. En la nova situaci\u00f3, les mans mortes del bosc p\u00fablic eren l\u2019Estat, que no cerca mai l\u2019autofinan\u00e7ament de les despeses de gesti\u00f3; els diners que manquin ja els posar\u00e0 l\u2019Estat. 9.\u2003DEFENSA I INTENTS DE RECUPERACI\u00d3 DELS B\u00c9NS COMUNALS DESAMORTITZATS El proc\u00e9s de centralitzaci\u00f3 no era senzill, perqu\u00e8, d\u2019una banda, la nova organitzaci\u00f3 apartava de la gesti\u00f3 moltes corporacions locals i molts ve\u00efns que l\u2019havien portada des de l\u2019edat mitjana, i, de l\u2019altra, era dif\u00edcil de coordinar la nova silvicultura amb moltes pr\u00e0ctiques forestals i drets tradicionals, com la pastura, fer llenya o tallar un arbre aqu\u00ed i un altre all\u00e0 quan tenia el gruix suficient, les pr\u00e0ctiques que s\u2019havien fet sempre. Les primeres passes de la nova organitzaci\u00f3 centralitzada varen tenir moltes dificultats en aquells indrets en qu\u00e8 els terrenys municipals i comunals tenien un paper important en l\u2019economia local. La desobedi\u00e8ncia a determinades normes imposades varen prendre formes diferents. Algunes institucions, com, per exemple, la Diputaci\u00f3 de Lleida, varen retardar la tramitaci\u00f3 d\u2019alguns expedients i varen evitar la venda de b\u00e9ns municipals. Molts pobles permeteren deixar que els ve\u00efns continuessin amb les seves pr\u00e0ctiques tradicionals, d\u2019altres varen boicotejar les subhastes d\u2019aprofitaments. L\u2019Estat va reaccionar encomanant a la Gu\u00e0rdia Civil el compliment de les noves directrius. Imposar el nou r\u00e8gim va costar a l\u2019Administraci\u00f3 un grapat d\u2019anys, per\u00f2 de mica en mica, amb molta, molta guarderia i gens de negociaci\u00f3, ho va aconseguir. La nova gesti\u00f3 estatal dels b\u00e9ns municipals va deixar, com hem comentat, molta gent sense uns recursos necessaris per a la superviv\u00e8ncia, sobre tot en \u00e0rees on predominaven les grans propietats, i on els pagesos sense terra treballaven de jornalers temporers. Aix\u00f2 va afavorir que, a bona part de les Espanyes, les primeres lluites camperoles de la segona meitat del segle xix defensessin la recuperaci\u00f3 dels comunals desamortitzats; per a molts aquella expropiaci\u00f3 i venda dirigida pels governs mon\u00e0rquics era la causa de molta mis\u00e8ria. D\u2019altres, m\u00e9s radicalitzats, varen entendre que l\u2019eliminaci\u00f3 de la propietat col\u00b7lectiva i la gesti\u00f3 estatal dels boscos no desamortitzats suposava una usurpaci\u00f3 pura i dura. En les zones m\u00e9s afectades per la desamortitzaci\u00f3 aix\u00f2 va donar lloc a un imaginari centrat en la defensa del comunal. La Segona Rep\u00fablica va arribar en una conjuntura econ\u00f2mica de crisi, generada pel crac del 1929. Al camp, aquesta situaci\u00f3 va produir una forta caiguda dels preus dels productes agraris i un increment important de l\u2019atur. QUADERNS AGRARIS 42\u2002(juny 2017), p. 105-126_\r\n\r\nI think that the main difference between the crashing samples and the rest is their length. Therefore, couldn't the length be causing the message errors? I hope with these samples you can identify what is causing the crashes considering that the 0.4.0 nlp library was loading them properly.","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"So we're using the csv reader to read text files because arrow doesn't have a text reader.\r\nTo workaround the fact that text files are just csv with one column, we want to set a delimiter that doesn't appear in text files.\r\nUntil now I thought that it would do the job but unfortunately it looks like even characters like \\a appear in text files.\r\n\r\nSo we have to option:\r\n- find another delimiter that does the job (maybe `\\x1b` esc or `\\x18` cancel)\r\n- don't use the csv reader from arrow but the text reader from pandas instead (or any other reader). The only important thing is that it must be fast (arrow's reader has a nice and fast multithreaded for csv that we're using now but hopefully we can find an alternative)\r\n\r\n\r\n\r\n> @lhoestq Can you ever be certain that a delimiter character is not present in a plain text file? In other formats (e.g. CSV) , rules are set of what is allowed and what isn't so that it actually constitutes a CSV file. In a text file you basically have \"anything goes\", so I don't think you can ever be entirely sure that the chosen delimiter does not exist in the text file, or am I wrong?\r\n\r\nAs long as the text file follows some encoding it wouldn't make sense to have characters such as the bell character. However I agree it can happen.\r\n\r\n> If I understand correctly you choose a delimiter that we hope does not exist in the file, so that when the CSV parser starts splitting into columns, it will only ever create one column? Why can't we use a newline character though?\r\n\r\nExactly. Arrow doesn't allow the newline character unfortunately.","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":289,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nSo we're using the csv reader to read text files because arrow doesn't have a text reader.\r\nTo workaround the fact that text files are just csv with one column, we want to set a delimiter that doesn't appear in text files.\r\nUntil now I thought that it would do the job but unfortunately it looks like even characters like \\a appear in text files.\r\n\r\nSo we have to option:\r\n- find another delimiter that does the job (maybe `\\x1b` esc or `\\x18` cancel)\r\n- don't use the csv reader from arrow but the text reader from pandas instead (or any other reader). The only important thing is that it must be fast (arrow's reader has a nice and fast multithreaded for csv that we're using now but hopefully we can find an alternative)\r\n\r\n\r\n\r\n> @lhoestq Can you ever be certain that a delimiter character is not present in a plain text file? In other formats (e.g. CSV) , rules are set of what is allowed and what isn't so that it actually constitutes a CSV file. In a text file you basically have \"anything goes\", so I don't think you can ever be entirely sure that the chosen delimiter does not exist in the text file, or am I wrong?\r\n\r\nAs long as the text file follows some encoding it wouldn't make sense to have characters such as the bell character. However I agree it can happen.\r\n\r\n> If I understand correctly you choose a delimiter that we hope does not exist in the file, so that when the CSV parser starts splitting into columns, it will only ever create one column? Why can't we use a newline character though?\r\n\r\nExactly. Arrow doesn't allow the newline character unfortunately.","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"> Okay, I have splitted the crashing shards into individual sentences and some examples of the inputs that are causing the crashes are the following ones\r\n\r\nThanks for digging into it !\r\n\r\nCharacters like \\a or \\b are not shown when printing the text, so as it is I can't tell if it contains unexpected characters.\r\nMaybe could could open the file in python and check if `\"\\b\" in open(\"path\/to\/file\", \"r\").read()` ?\r\n\r\n> I think that the main difference between the crashing samples and the rest is their length. Therefore, couldn't the length be causing the message errors? I hope with these samples you can identify what is causing the crashes considering that the 0.4.0 nlp library was loading them properly.\r\n\r\nTo check that you could try to run \r\n\r\n```python\r\nimport pyarrow as pa\r\nimport pyarrow.csv\r\n\r\nopen(\"dummy.txt\", \"w\").write(((\"a\" * 10_000) + \"\\n\") * 4) # 4 lines of 10 000 'a'\r\n\r\nparse_options = pa.csv.ParseOptions( \r\n delimiter=\"\\b\", \r\n quote_char=False, \r\n double_quote=False, \r\n escape_char=False, \r\n newlines_in_values=False, \r\n ignore_empty_lines=False, \r\n)\r\n\r\nread_options= pa.csv.ReadOptions(use_threads=True, column_names=[\"text\"])\r\n\r\npa_table = pa.csv.read_csv(\"dummy.txt\", read_options=read_options, parse_options=parse_options)\r\n```\r\n\r\non my side it runs without error though","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":178,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n> Okay, I have splitted the crashing shards into individual sentences and some examples of the inputs that are causing the crashes are the following ones\r\n\r\nThanks for digging into it !\r\n\r\nCharacters like \\a or \\b are not shown when printing the text, so as it is I can't tell if it contains unexpected characters.\r\nMaybe could could open the file in python and check if `\"\\b\" in open(\"path\/to\/file\", \"r\").read()` ?\r\n\r\n> I think that the main difference between the crashing samples and the rest is their length. Therefore, couldn't the length be causing the message errors? I hope with these samples you can identify what is causing the crashes considering that the 0.4.0 nlp library was loading them properly.\r\n\r\nTo check that you could try to run \r\n\r\n```python\r\nimport pyarrow as pa\r\nimport pyarrow.csv\r\n\r\nopen(\"dummy.txt\", \"w\").write(((\"a\" * 10_000) + \"\\n\") * 4) # 4 lines of 10 000 'a'\r\n\r\nparse_options = pa.csv.ParseOptions( \r\n delimiter=\"\\b\", \r\n quote_char=False, \r\n double_quote=False, \r\n escape_char=False, \r\n newlines_in_values=False, \r\n ignore_empty_lines=False, \r\n)\r\n\r\nread_options= pa.csv.ReadOptions(use_threads=True, column_names=[\"text\"])\r\n\r\npa_table = pa.csv.read_csv(\"dummy.txt\", read_options=read_options, parse_options=parse_options)\r\n```\r\n\r\non my side it runs without error though","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"That's true, It was my error printing the text that way. Maybe as a workaround, I can force all my input samples to have \"\\b\" at the end?","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":28,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\nThat's true, It was my error printing the text that way. Maybe as a workaround, I can force all my input samples to have \"\\b\" at the end?","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/622","title":"load_dataset for text files not working","comments":"> That's true, It was my error printing the text that way. Maybe as a workaround, I can force all my input samples to have \"\\b\" at the end?\r\n\r\nI don't think it would work since we only want one column, and \"\\b\" is set to be the delimiter between two columns, so it will raise the same issue again. Pyarrow would think that there is more than one column if the delimiter is found somewhere.\r\n\r\nAnyway, I I'll work on a new text reader if we don't find the right workaround about this delimiter issue.","body":"Trying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n","comment_length":96,"text":"load_dataset for text files not working\nTrying the following snippet, I get different problems on Linux and Windows.\r\n\r\n\r\n```python\r\ndataset = load_dataset(\"text\", data_files=\"data.txt\")\r\n# or \r\ndataset = load_dataset(\"text\", data_files=[\"data.txt\"])\r\n```\r\n\r\n(ps [This example](https:\/\/huggingface.co\/docs\/datasets\/loading_datasets.html#json-files) shows that you can use a string as input for data_files, but the signature is `Union[Dict, List]`.)\r\n\r\nThe problem on Linux is that the script crashes with a CSV error (even though it isn't a CSV file). On Windows the script just seems to freeze or get stuck after loading the config file.\r\n\r\nLinux stack trace:\r\n```\r\nPyTorch version 1.6.0+cu101 available.\r\nChecking \/home\/bram\/.cache\/huggingface\/datasets\/b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at \/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.json\r\nUsing custom data configuration default\r\nGenerating dataset text (\/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7)\r\nDownloading and preparing dataset text\/default-0907112cc6cd2a38 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/bram\/.cache\/huggingface\/datasets\/text\/default-0907112cc6cd2a38\/0.0.0\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7...\r\nDataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading took 0.0 min\r\nChecksum Computation took 0.0 min\r\nUnable to verify checksums.\r\nGenerating split train\r\nTraceback (most recent call last):\r\n File \"\/home\/bram\/Python\/projects\/dutch-simplification\/utils.py\", line 45, in prepare_data\r\n dataset = load_dataset(\"text\", data_files=dataset_f)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/load.py\", line 608, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 468, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 546, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/datasets\/builder.py\", line 888, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False, disable=not_verbose):\r\n File \"\/home\/bram\/.local\/share\/virtualenvs\/dutch-simplification-NcpPZtDF\/lib\/python3.8\/site-packages\/tqdm\/std.py\", line 1130, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/bram\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\/text.py\", line 100, in _generate_tables\r\n pa_table = pac.read_csv(\r\n File \"pyarrow\/_csv.pyx\", line 714, in pyarrow._csv.read_csv\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: CSV parse error: Expected 1 columns, got 2\r\n```\r\n\r\nWindows just seems to get stuck. Even with a tiny dataset of 10 lines, it has been stuck for 15 minutes already at this message:\r\n\r\n```\r\nChecking C:\\Users\\bramv\\.cache\\huggingface\\datasets\\b1d50a0e74da9a7b9822cea8ff4e4f217dd892e09eb14f6274a2169e5436e2ea.30c25842cda32b0540d88b7195147decf9671ee442f4bc2fb6ad74016852978e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py to C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.py\r\nCouldn't find dataset infos file at https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\\dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/datasets\/1.0.1\/datasets\/text\/text.py at C:\\Users\\bramv\\.cache\\huggingface\\modules\\datasets_modules\\datasets\\text\\7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7\\text.json\r\nUsing custom data configuration default\r\n```\r\n\n> That's true, It was my error printing the text that way. Maybe as a workaround, I can force all my input samples to have \"\\b\" at the end?\r\n\r\nI don't think it would work since we only want one column, and \"\\b\" is set to be the delimiter between two columns, so it will raise the same issue again. Pyarrow would think that there is more than one column if the delimiter is found somewhere.\r\n\r\nAnyway, I I'll work on a new text reader if we don't find the right workaround about this delimiter issue.","embeddings":[-0.2746651173,-0.4020575583,0.0175604951,0.3872526288,0.2696425617,-0.0386613607,0.3188881874,-0.0543565489,0.4263595045,-0.0580488145,0.0659720227,0.145524323,-0.1557630599,0.2742006779,0.0635561198,-0.035076078,0.1571835279,-0.0138417054,-0.2914433181,0.0402483419,-0.1144550294,0.3431209624,-0.2153358608,-0.1469300389,-0.4160707295,0.2740044296,-0.1050167531,0.4538784921,-0.0742620155,-0.2378563434,0.2710002363,0.1282079071,0.1639642566,0.6136498451,-0.0001269617,0.1100326851,0.2756481171,-0.1610112339,-0.2998276949,-0.5695467591,0.1585155129,-0.2294367254,0.2088484615,-0.0287806056,0.1301927119,-0.0045288531,0.1527263522,-0.3340618312,0.3551648259,0.3254779875,0.1009777859,0.3193374872,0.0620950386,-0.0146584613,0.0609620512,0.4765509367,-0.0917448476,0.3906496465,0.313786149,-0.1942874789,0.1169522405,-0.082400918,-0.173093155,-0.1696534455,0.3285271525,0.2871706486,-0.6134480238,-0.2057267129,0.0768413693,0.1859920323,0.4352324009,-0.4339176416,-0.221748367,-0.1591291726,-0.0756903589,-0.1254785657,0.4388749599,0.1783995628,-0.1985727847,0.0581210814,-0.2503564954,-0.0692397878,-0.1769828945,0.323066324,-0.0084304409,0.0508893542,-0.219182238,0.1040090844,0.3328258991,-0.0709382817,-0.2063759267,-0.2793429792,-0.0293538962,0.0611276552,-0.2661669552,0.1557293087,-0.2658762336,-0.0245267656,0.1117541268,0.1747620702,-0.0047685001,0.1417199671,0.0951443389,0.2306069136,0.2007873058,0.0978597626,0.4081541598,0.0843545943,0.2615751028,0.0263595376,-0.0160578508,-0.0952473879,-0.1963426471,-0.544190526,-0.069791168,-0.2300527245,0.4797008932,-0.2050437331,-0.1790357679,-0.0498552136,-0.0808124468,-0.0455941074,0.2161368579,0.6197093129,-0.0843391269,0.0338773318,0.1204905659,0.3182838857,-0.1168016568,0.1911230683,0.047579363,-0.0604269393,-0.0889259353,0.1795253456,0.5387113094,-0.3721969426,0.3424328566,0.1996889412,0.4671038091,-0.1905285269,-0.151450634,-0.1865481436,-0.0174010713,0.1886943281,0.0076277102,0.0419398174,0.2405289412,-0.1590055078,-0.1147044152,0.094007358,-0.2652543485,-0.0938472077,0.0322779305,0.0448435508,-0.09525875,-0.2330176234,-0.3073214889,0.0929248333,0.0373641104,-0.1054287404,0.0938850492,-0.1803099215,-0.2380830199,-0.1588562429,0.2716287971,0.6603162289,-0.2988535464,-0.1454449892,0.325542748,-0.1225560904,-0.1192727685,0.2550257146,-0.0943285748,0.0099470215,-0.2610239685,0.1993149221,0.1874883026,-0.4092430174,-0.1465579122,0.396353066,0.0584071279,0.1357305199,0.2080578804,0.0618107505,0.0415842496,0.0460110344,0.2806780338,0.0083062993,0.1047209352,-0.0915358141,-0.0034152512,-0.2037547082,0.0910752863,0.3670076728,-0.2113386095,0.0550341345,0.1075050905,-0.1117831096,0.1765403897,-0.1039660424,-0.034448199,0.5095410943,0.1362373233,0.3201455176,0.0793995485,-0.2116694301,-0.5979293585,0.1366195083,0.2275739461,-0.0535317548,-0.2638650537,-0.0957267582,-0.1770208925,0.0601343401,-0.2095396072,-0.0141126756,-0.0720959976,0.1533231437,0.2246878594,0.0292827152,-0.2047470063,0.4062135816,-0.1693811566,0.2454877943,-0.2644483447,0.1788318306,-0.0356653072,-0.2023017704,-0.0517953187,0.1018908471,-0.0046761343,-0.293756932,0.0954740718,0.4632967114,0.0546923168,0.08568497,-0.1847886443,-0.0953086168,0.2269652784,-0.0808336586,-0.0273770448,0.2919701338,0.2026348412,-0.1917808801,-0.2744134963,0.1912704259,-0.3355849385,0.1634294242,0.0298720635,-0.1083321646,-0.0484339148,0.1079697907,-0.3023562431,-0.0635969713,0.4508511424,-0.245456323,0.2605041265,0.001822158,-0.4244135916,-0.2174914628,0.4196653664,-0.0460522622,0.1226554886,0.1877402812,-0.2147010267,0.212012589,-0.0877679214,-0.0419770256,0.5764291883,0.1378007531,-0.2983497083,0.2298145294,-0.1166211367,-0.2212798446,0.289095819,-0.0472781882,-0.0982488245,0.0960121751,-0.1998828202,0.0083255861,-0.2600301802,-0.0678929389,-0.1015431955,0.0715258271,-0.412751019,0.222360611,-0.3261346519,-0.1620817333,-0.3533613086,0.1199700087,-0.2934476137,-0.0352531411,-0.2882407308,0.2256355137,0.1631828547,0.0368443988,-0.025598187,0.0039330921,0.1004287452,-0.5494058728,-0.1985674053,0.0081635946,-0.2308933288,-0.0934241638,0.3823769689,0.0441114455,0.2374777943,-0.3961957097,-0.1094290391,-0.0985937938,-0.1599082798,0.0557672121,-0.0614242442,0.1960316002,0.0531645492,0.2280198932,-0.123751916,-0.1310472041,0.4299977124,-0.0381334759,-0.1885631084,0.2453443706,0.418446213,-0.309718281,-0.2610920072,-0.3623771966,-0.1279267967,-0.2842009068,0.3581483066,0.1559180021,-0.0295690484,0.5024074316,0.4287550151,0.15696235,-0.1803515255,0.2165792435,0.0652968511,-0.1683056056,0.471650809,-0.1643847376,-0.6319022179,0.0775601417,0.4496120214,-0.2843745053,0.2175411284,-0.4422395527,-0.0620809942,-0.1192547455,0.092089884,0.0410850048,0.3046552837,0.1864622682,0.0804128125,0.0956785753,-0.0544941127,-0.3214206994,0.1158758476,-0.2110431045,0.03335388,0.2350959331,0.3713712096,-0.2012346834,0.4552719593,0.3391940594,-0.0664085746,0.2473822087,-0.5043381453,0.5094949603,-0.1640397608,-0.5347828269,0.0561735779,-0.2177346945,0.1750393808,0.2766551971,0.1536442488,0.3985494077,-0.3371795714,0.1042308062,-0.074585855,-0.1997736245,0.2598340511,-0.1172346398,0.0794808418,-0.1508025229,0.1711593866,0.1823592782,-0.2328867614,0.039038755,0.6069954634,-0.1873313934,0.051549565,-0.3987098336,0.0692123771,-0.3697613478,0.3847955465,-0.0440684631,0.3335777521,-0.2729938626,-0.1051141024,0.0177630913,-0.0372726545,0.5544432402,0.3075835109,-0.1952399313,0.0176452883,-0.2582427859,-0.3879300058,0.0970626697,-0.2831110656,0.3912253082,0.1797815561,0.6442822814,-0.3002554476,-0.2145057619,-0.0911262631,0.4764973521,-0.0228321832,-0.3114876747,-0.2785250545,-0.1634372324,-0.3262649477,-0.2423540801,0.0241883285,0.2216682285,-0.1815148294,0.2211253345,-0.057809554,-0.1648258865,0.2988142669,-0.0309885107,0.2692628205,-0.3283635974,0.2320739627,0.106232062,0.4693136215,0.2969624698,0.5515258312,-0.0897572115,-0.4601889551,0.0208050236,-0.1581919938,0.3489819467,0.1660993397,-0.2715944648,0.1426507235,0.2977319956,0.1101861969,-0.4385163784,-0.0669690967,0.5130977631,0.2113117278,-0.2971880138,-0.4682967961,0.3646911383,0.0281752609,0.149966687,0.2716005147,0.1743876189,-0.2929801643,0.0181149673,-0.3032903969,0.8145566583,-0.1351302564,0.4673897028,0.2345223725,-0.0871991217,0.3978260159,-0.0533247665,0.0656356812,-0.2903504372,-0.0147631858,-0.0129349958,-0.1748355329,0.443472594,0.1701350212,-0.4242368042,0.2192278653,-0.0133848321,0.310113132,-0.2710388303,0.1851529032,-0.4233773351,-0.3067678809,-0.3461351395,0.0314102881,0.1470375955,0.3159363568,-0.0193047076,0.0686776266,-0.1788800955,-0.2796393037,-0.3666658998,0.0574040636,-0.321728915,0.0584385805,0.2108019441,-0.3453936875,0.3396160901,0.4470129311,0.2227248847,0.1258746684,-0.1699065119,-0.0880349576,-0.2285286933,-0.1075692996,0.001792727,-0.2334611863,0.0256202221,0.0866714865,-0.2706852257,-0.0342282578,-0.0761413649,-0.1780479103,-0.0112773841,0.1302926689,-0.3514166176,-0.4001734853,-0.4443776906,-0.1643957794,0.1622943431,-0.0622147433,0.0125064095,0.1251550466,0.153515771,0.0424737409,0.0863078237,-0.1513924897,-0.1063951999,0.3808281422,-0.4325575531,-0.2435594797,0.6115919948,0.4478023946,-0.1363858581,-0.2047023922,0.3153443635,-0.0241039079,-0.4706380963,0.0600293167,0.3476564884,0.0469819419,0.021084819,0.2753241658,0.0409553722,-0.1721466184,0.1079835668,-0.57039994,-0.3437371254,0.1958448887,0.299157232,0.1324544549,0.2247095555,-0.151683256,-0.0144614223,-0.2124667168,-0.1799509674,0.0283627734,0.0711819381,-0.2090127319,0.4280432761,0.081355013,0.344661355,-0.2428899556,0.018016791,-0.008255586,0.0614189841,-0.0559105575,-0.132317096,0.1357380897,-0.0764971972,-0.180190295,-0.1695444733,-0.3009041548,-0.0054444694,-0.3061545193,0.1830037087,0.490568608,-0.1322172284,0.1660419554,-0.2308708429,0.2402737886,-0.3294675648,0.2159640789,-0.330162555,0.3099443913,0.1199942008,0.0645242259,-0.0035627526,0.0762148574,-0.1796475351,0.2804464102,0.1938064098,-0.0876487717,0.2333928645,-0.4173033535,0.1791363806,-0.1308648288,0.4586466551,0.6418234706,-0.2389519513,0.1770924479,0.3237511218,0.0375065692,-0.1066855043,0.0253855046,0.1342612356,-0.0756319016,0.0557794683,0.2839002907,-0.1099474356,-0.0767709538,0.0140856896,0.0420108885,0.301736474,0.1350494921,0.0422454402,0.3179253638,0.2821864784,0.2172951549,-0.0789582357,0.3243644536,0.168809697,0.3248771429,-0.1270244122,0.1264024526,-0.1879327297,0.3269269466,-0.0585651398,-0.3316073716,0.1236123517,0.0540438406,-0.2286887318,0.0388122387,-0.2160567045,0.5491757393,-0.5157114863,-0.0090491632,-0.1283324212,0.0684653968,-0.0507389456,-0.3157410026,0.0320397541,-0.2380458564,-0.0311427116,0.0246612746,-0.0586087257,-0.1354770064,0.0637513846,0.1260185093,-0.0337765217,-0.3310147226,0.121667631,0.0174299572,0.0148993013,0.0049083694,0.2710101008,0.1519322544,0.2196570933,0.3230521083,0.3475781977,0.5777654648,0.4073304832,0.1060493439,0.2475764155,-0.2693928778,-0.0126066068,-0.0394145399,0.3580520749,0.1128151789,0.0378045253,0.2164451629,0.0713129267,-0.0373921506,-0.0545146838,0.0653115958,-0.1178254336,-0.2989546657,0.232823506,-0.2536529005,-0.1291730106,-0.1700772047,-0.0467461348,-0.4412434697,-0.0388390608,0.5237540603,0.2658577561,0.1927789003,-0.0427044407,0.0265391581,0.3400238454,0.467133671,0.3830900192,0.2377198189,-0.0570629463,-0.011146294,-0.6012914181,-0.0115617318,0.0097210435,-0.1373764127,-0.2386875898,-0.153664276,0.1978022158,0.1551500559,0.0523215979,-0.1281724423,0.2108146548,-0.0120303463,-0.0534492657,-0.1669248343,-0.0151779186,0.1381872743,0.0049077021,-0.3023657203,0.0613835305,-0.1089829579,-0.0564865135,-0.3146940172,0.4238253236,-0.1058527976,-0.3355431855,0.1364640594,-0.0730648637,0.4701799154,0.0425091274,-0.1654959768,-0.1798643023,-0.2382656336,0.0213493612,0.4302535355,0.1545348465,0.0050935661,-0.4908640981,-0.4926251471,-0.26628986,0.2439431995,-0.0592638478,-0.3730340302,0.0246137008,-0.0186506398,-0.0354579501,0.0011535721,0.2333126962,0.4785373509,-0.2721097171,0.0106194569,-0.147819519,-0.1278119087,0.3023916781,-0.4510451853,-0.2669306993,0.0307972059,0.0089576747,0.0157020632,0.0496490262,-0.5663852692,0.1250867397,0.1359771192,-0.1033064723,-0.1399750113,0.0530345663,-0.1825748235,-0.0363837555,-0.1389525682,0.4549028575,0.088384077,-0.2588394582,0.0973682404,-0.2056961805]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"It seems that I ran into the same problem\r\n```\r\ndef tokenize(cols, example):\r\n for in_col, out_col in cols.items():\r\n example[out_col] = hf_tokenizer.convert_tokens_to_ids(hf_tokenizer.tokenize(example[in_col]))\r\n return example\r\ncola = datasets.load_dataset('glue', 'cola')\r\ntokenized_cola = cola.map(partial(tokenize, {'sentence': 'text_idxs'}),\r\n num_proc=2,)\r\n```\r\nand it outpus (exceprts)\r\n```\r\nConcatenating 2 shards from multiprocessing\r\nSet __getitem__(key) output type to python objects for ['idx', 'label', 'sentence', 'text_idxs'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nDone writing 532 indices in 4256 bytes .\r\nDone writing 531 indices in 4248 bytes .\r\nProcess #0 will write at \/home\/yisiang\/.cache\/huggingface\/datasets\/glue\/cola\/1.0.0\/930e9d141872db65102cabb9fa8ac01c11ffc8a1b72c2e364d8cdda4610df542\/tokenized_test_00000_of_00002.arrow\r\nProcess #1 will write at \/home\/yisiang\/.cache\/huggingface\/datasets\/glue\/cola\/1.0.0\/930e9d141872db65102cabb9fa8ac01c11ffc8a1b72c2e364d8cdda4610df542\/tokenized_test_00001_of_00002.arrow\r\nSpawning 2 processes\r\n```\r\nand then the program never stop.","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":121,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nIt seems that I ran into the same problem\r\n```\r\ndef tokenize(cols, example):\r\n for in_col, out_col in cols.items():\r\n example[out_col] = hf_tokenizer.convert_tokens_to_ids(hf_tokenizer.tokenize(example[in_col]))\r\n return example\r\ncola = datasets.load_dataset('glue', 'cola')\r\ntokenized_cola = cola.map(partial(tokenize, {'sentence': 'text_idxs'}),\r\n num_proc=2,)\r\n```\r\nand it outpus (exceprts)\r\n```\r\nConcatenating 2 shards from multiprocessing\r\nSet __getitem__(key) output type to python objects for ['idx', 'label', 'sentence', 'text_idxs'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nTesting the mapped function outputs\r\nTesting finished, running the mapping function on the dataset\r\nDone writing 532 indices in 4256 bytes .\r\nDone writing 531 indices in 4248 bytes .\r\nProcess #0 will write at \/home\/yisiang\/.cache\/huggingface\/datasets\/glue\/cola\/1.0.0\/930e9d141872db65102cabb9fa8ac01c11ffc8a1b72c2e364d8cdda4610df542\/tokenized_test_00000_of_00002.arrow\r\nProcess #1 will write at \/home\/yisiang\/.cache\/huggingface\/datasets\/glue\/cola\/1.0.0\/930e9d141872db65102cabb9fa8ac01c11ffc8a1b72c2e364d8cdda4610df542\/tokenized_test_00001_of_00002.arrow\r\nSpawning 2 processes\r\n```\r\nand then the program never stop.","embeddings":[-0.4042771459,-0.0761643201,0.0068098302,0.146670714,0.1475021839,-0.1896045357,0.3224094212,0.3429279327,0.1341744512,0.1369507015,0.0367483571,0.3943180144,-0.4054729342,0.310338974,-0.3056828976,0.0618968457,0.1283408403,-0.0522314869,-0.2669127584,0.1578018814,-0.279797256,0.2109823078,-0.3103196323,0.2153621763,-0.5755159855,0.0590776615,-0.0540032201,0.2517116368,0.004229615,-0.5668848157,0.3097843826,0.3128777742,0.0841088668,0.4231201708,-0.0001253461,0.1279418021,0.3116935492,-0.1059477702,-0.0364288501,-0.3833858371,-0.0744350031,-0.2384934276,0.2321601957,0.1469333619,0.2514846027,-0.0637059957,-0.1770320237,-0.1964325458,0.1855876446,0.3598256409,0.0798642561,0.2393260002,0.2558632493,0.106185101,0.0308736991,0.1623829901,-0.0530019514,-0.0189514477,0.2539156675,-0.4286129177,-0.0830043256,0.2592658103,-0.2532216311,-0.2118041664,-0.1340045184,-0.2208434492,0.5308262706,-0.540256381,0.2115586698,0.0486420356,-0.2676083148,-0.0402106382,-0.4429701567,-0.176854074,-0.2546718717,-0.5399870872,0.1709608883,0.0307247601,-0.1698826551,0.1184880957,-0.439244777,-0.2035617232,0.0078960899,-0.0298823081,-0.1458258331,0.6477392912,0.2138981819,0.3055250049,0.1323184669,0.130570069,0.1838415712,-0.0163391028,-0.0709623694,0.207008481,-0.4304935038,0.0555331931,0.0561705679,-0.2127608955,-0.1796363294,-0.0278097931,-0.2530755997,0.1307653189,-0.1172353625,0.0927313417,0.4056189954,-0.0962312669,0.1677099168,0.3280444443,0.2209367603,-0.1989021003,-0.0220488459,0.0117406221,0.2665813565,-0.0622575358,-0.0979441181,0.2686050832,0.0859605074,-0.1967092603,-0.2017096877,0.1901551634,-0.1704506874,-0.1557840407,0.076307945,0.2368407995,0.0790321007,0.7875010967,0.1054009572,0.2169736475,-0.3290952146,-0.1062655374,-0.069049105,-0.0170265641,-0.3850457072,0.1155861765,0.2116104662,0.1460444331,-0.0303646792,0.0347133093,-0.3158589602,-0.2540472746,0.2134025246,-0.3203782737,0.1630386114,0.6764134169,-0.0766727179,0.0704787001,0.0608326644,-0.2051453292,0.0493657514,0.196880579,-0.528067708,-0.1739725918,-0.2313921601,0.0630153418,0.0921197608,0.2983644307,-0.1968151927,0.2527915239,0.3752551079,-0.272626251,-0.2939334512,-0.4000747204,-0.3631501496,-0.2814882994,-0.0012882305,0.2978385687,-0.4101808965,-0.0337190032,0.1389254481,-0.1741313934,0.3642458618,0.2580123246,-0.0363178253,0.0802293941,-0.1671099961,0.1751141399,0.1112992838,-0.2420786768,-0.0535121113,0.2751089334,0.0924496725,0.3352847695,-0.088490203,-0.3326949179,0.3726418614,-0.1397295892,0.2458258718,-0.0516839549,-0.1721580476,0.0256981496,-0.4304659367,-0.0167047903,-0.0738441423,-0.1338193566,0.4579972923,0.2050948143,-0.1281719357,-0.4877860248,0.3257185519,-0.1268086135,0.3116859496,0.0617736131,0.1279980689,0.1767352819,0.1328583807,-0.2497059554,-0.3686510623,0.1384370476,-0.2625176907,0.0193939805,-0.0266870987,-0.0783315599,0.0837063417,0.2017332166,-0.2120802253,-0.2150038928,0.0355656222,0.0074174814,-0.3615643382,-0.0625980422,-0.1797469854,0.625947237,0.091997318,0.1794136018,0.0131107029,0.2692207396,-0.1198470592,-0.4355103672,-0.2855910659,0.3075087667,0.0316634066,-0.0748290792,-0.200615108,0.4490844011,0.5101372004,-0.1541983187,-0.0260389913,0.1566484571,0.2412418425,-0.1208864748,-0.1833622456,0.037562713,0.0654287338,-0.1723884791,0.3006968796,0.483391881,0.201399982,0.3852221966,0.25669837,0.2287194878,0.3226880431,0.0051959977,0.0316118337,-0.1580841541,0.1246569902,-0.1312185675,0.1465567648,0.0016201094,-0.1420030296,-0.195558086,0.1355320066,0.0595095456,-0.1583060324,0.0607259907,0.0952029526,-0.1042194143,0.1780013293,-0.0132490825,0.3717621267,-0.0238193534,-0.2368368059,0.1435663253,-0.2938762903,0.0356035046,0.0739235356,-0.0384558737,0.3780089915,0.3359934688,0.0807103962,0.0478107147,-0.2073845714,-0.3678789437,0.0716596916,0.3440575302,-0.5090512037,0.2327922434,-0.2995890379,0.3964595795,0.1238909364,-0.1198131517,-0.2803930938,-0.5962172747,-0.1310129613,0.5344035625,-0.0878550559,0.1409512162,-0.1044902727,0.1036587059,-0.2177422494,0.4064002335,-0.1127116904,-0.286685884,-0.2238863558,-0.143870756,0.3367703259,-0.2551400959,0.2371955961,0.1972360015,-0.2832110822,0.1011269614,-0.3656711876,0.1703359783,-0.0298877079,0.0133362152,0.110932447,-0.0670838058,-0.0422333665,-0.2127739042,0.1584321856,-0.1304798871,-0.2130654156,0.3102205694,-0.1451127082,0.0382950902,-0.1842952371,-0.2188405544,-0.4120937288,-0.1025576442,-0.1735719442,-0.2267582864,0.3224339187,0.0706637204,-0.0400239602,-0.0154559342,-0.0948967189,-0.0137680816,-0.1022756919,-0.0214018058,-0.163213551,0.0092593357,-0.1443979144,-0.0533849858,0.0301886387,0.1384504139,0.4326987267,-0.3306240737,0.0789940655,0.1064507589,-0.1260214746,0.1771831661,-0.1903131157,0.3407360017,0.3817727864,0.049549032,-0.0437264889,-0.1391005218,0.1047363207,-0.0750269368,0.0507857837,0.0907464176,0.3363541961,0.1733752191,0.680211544,0.3181328773,0.0758746341,0.4411141574,-0.0873885751,0.1266505122,-0.2016421407,-0.4437966645,-0.1668882966,-0.3328511119,0.0288301278,-0.1490571797,-0.1336886436,-0.4281876385,-0.2398667336,0.5580083728,-0.3731169701,-0.2661548555,-0.0114972563,-0.4910701215,0.294259727,-0.0721302629,0.0983663127,-0.068047978,-0.0286521465,-0.1811812222,0.241728127,0.1317715943,-0.2868562043,-0.3133730292,-0.2984551489,-0.4311941266,0.3018606901,0.2005129606,0.7738990188,0.0186728928,-0.0527879708,0.0111618182,-0.1868258715,0.9173187613,-0.5690357685,-0.3967113495,0.334510237,-0.4721472263,-0.2147320658,-0.2099572271,-0.20135203,0.5055507421,0.3819591701,0.5023783445,-0.1614258289,-0.3780629337,0.1530832052,-0.2570289373,-0.1191146821,0.0627677292,-0.3022753596,-0.0154425222,-0.2401999533,0.0927345231,-0.2525685728,0.1114537865,-0.217980817,-0.1276566982,-0.1641675234,-0.1095554158,0.2863313556,0.0590845011,-0.036182344,-0.1694262773,0.0408688895,-0.0105615938,0.1652771831,0.3650979996,0.1900334656,-0.3747362792,-0.172747463,0.1968021691,0.0158032663,0.3392807245,0.3174801171,0.2047665715,0.0000433281,-0.0467102788,0.2545279562,-0.313780278,0.0875377208,0.3993992209,-0.0182168111,-0.566470325,-0.0736856759,-0.0969469994,0.3861195445,-0.1427975297,0.5104857087,-0.3036777675,-0.2402489036,0.4625009298,0.0840426013,0.8406628966,-0.3637130857,0.013283791,0.0892839208,-0.0050920136,0.0694168583,-0.0628983378,-0.0200172663,-0.2953952849,0.004126376,-0.0972203314,0.0550890304,0.3370894194,0.2183585912,-0.1210360229,0.4495358169,-0.0287258793,-0.0096248183,-0.0391500108,0.1835768521,0.4047142565,-0.1066562012,0.1088414192,-0.0186672267,-0.0423479713,-0.1334843636,-0.0020877412,0.0470463037,0.2910001576,-0.1496485323,-0.3882095814,-0.0935217068,-0.1085731164,0.1244034618,0.2220883667,0.1331906468,0.0084557934,-0.0236116759,-0.0216732454,0.1052144989,-0.0715593323,-0.1477668881,0.1101583317,0.3326562643,0.1448520422,0.0032652961,0.3560247421,0.123672232,-0.2299612164,0.0653375387,-0.2153465003,-0.2262951732,0.0706195086,0.0178394821,0.1855371147,0.0778599977,0.152141124,-0.0188029315,-0.1115508303,-0.2047348768,0.0256376341,-0.1125884578,-0.1400844902,0.4181123972,-0.0282925516,-0.3238861561,-0.0096672634,0.4117647707,0.1827359796,-0.0839909986,0.1819818467,0.0767892003,-0.0807366744,-0.2318515927,-0.1778572202,0.3159954548,-0.1251131594,0.2839004993,0.0257563312,-0.2696927786,0.1455886811,0.0623940192,0.0602632463,0.2800891697,-0.1935098916,-0.4024766386,-0.434191674,0.3008685708,-0.0027213062,0.0832222924,-0.0944675952,0.2488301247,-0.1721978039,0.3336555064,-0.2000491023,0.0919543728,-0.1073144376,0.2210526913,-0.2320587188,0.0096003888,-0.1644627005,-0.0914378241,0.0653490052,0.0121605787,-0.2207485288,-0.0092055146,-0.0578293465,0.1484887302,0.1470884234,-0.1971917003,0.017915817,-0.0090579716,-0.1415070146,-0.2273655385,0.1838724464,0.3537555635,-0.2231854945,0.1348131895,0.3172183335,0.1472457796,0.0894243419,0.2858358026,-0.1238537803,0.0725882798,0.048357144,0.4121780992,0.1053218246,-0.1657755375,0.0245063882,0.1498433501,-0.0505776852,-0.0450951569,0.3554062843,-0.2137869895,0.068739295,-0.0204052757,0.3739389777,0.3854634166,-0.1282382309,0.0130777778,0.2367485315,0.0610653833,0.0084672477,-0.1731078327,0.4683952034,0.0407815352,0.0509879887,0.3392882943,0.3380618095,-0.1434528679,0.165693447,0.1148950458,-0.051844202,-0.0311292578,0.1816536635,0.1359544843,0.0247090515,0.3240545988,0.1971982419,-0.2997602224,-0.1530555189,0.2907688916,0.0458551124,0.5280521512,0.0002221263,0.1636801362,-0.1134032384,-0.6981275678,0.0208135601,0.1347817332,-0.4347136617,0.2099061459,-0.0852788091,0.2224349082,-0.226705119,-0.4809595048,-0.1522950381,0.479629904,-0.2569240332,-0.3121176958,-0.2767162621,-0.1225164533,-0.1777713448,0.0527482852,-0.0898243114,0.1594695598,0.7063724399,0.0123732593,-0.0512495749,-0.2058636993,-0.2766352594,-0.0424794443,0.3352660537,-0.0077611161,0.4334900081,-0.0677994117,0.1059244052,-0.0294918437,0.1589465439,0.4662959874,0.7281742692,-0.4269511998,-0.1200888902,0.118756257,-0.0645889863,-0.0706497952,0.3842765391,0.0765927583,0.225312382,0.3098933399,-0.088209562,0.0227853358,-0.2762964964,0.3280377388,0.3067688942,0.0891914666,0.4277430773,0.0766937509,-0.0542830378,0.2583425641,0.0579854734,-0.3257127106,-0.2193247825,0.5333292484,-0.4387640655,0.2860156,0.2186452001,0.0412960015,0.0746765733,0.3803938627,0.2904959619,0.0493487455,-0.2989239097,-0.2100402266,-0.4146656096,0.0077228141,-0.2228157967,0.2766835392,0.0260397773,0.0152188744,-0.0235646032,0.0177249424,-0.0496937223,-0.2223452479,0.3785585761,0.3334757984,-0.5266215205,0.1982720196,-0.0054435791,-0.026413355,-0.0235639121,-0.3387909234,0.3580256701,0.0833536014,-0.1214982271,0.1432868093,0.1241795272,0.0873626322,0.0999922678,0.0031533507,0.2782770395,0.5841792226,-0.0505016372,0.0479514897,-0.2000539154,0.0347690582,-0.1239147186,-0.010226557,-0.006463137,0.3861204088,-0.1547226906,-0.0461015962,0.0063532107,0.1711958349,0.0356916636,0.1453517228,-0.4596277177,0.3033055365,0.161960572,0.0160800479,-0.1337858737,0.4253684282,0.1762627363,0.3895446658,-0.3170860708,-0.2272410542,0.329297632,-0.0643050894,-0.1734612435,-0.6152847409,0.2426211387,-0.1712373793,0.0917392075,-0.5049587488,-0.3762700856,0.1952891201,-0.1857613176,-0.5628415942,0.2360658497,-0.0989015475,-0.1997238547,-0.0967540219,-0.1907408684,0.0916335136,-0.1020657495,0.2155115157,-0.0274184011]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"same problem.\r\n`encoded_dataset = core_data.map(lambda examples: tokenizer(examples[\"query\"], examples[\"document\"], padding=True, truncation='longest_first', return_tensors=\"pt\", max_length=384), num_proc=16, keep_in_memory=True)`\r\nit outputs:\r\n```\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787499 indices in 25568385696 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nSpawning 16 processes\r\n```","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":301,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nsame problem.\r\n`encoded_dataset = core_data.map(lambda examples: tokenizer(examples[\"query\"], examples[\"document\"], padding=True, truncation='longest_first', return_tensors=\"pt\", max_length=384), num_proc=16, keep_in_memory=True)`\r\nit outputs:\r\n```\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787500 indices in 25568400000 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nDone writing 1787499 indices in 25568385696 bytes .\r\nSet __getitem__(key) output type to python objects for ['document', 'is_random', 'query'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\nSpawning 16 processes\r\n```","embeddings":[-0.3812628984,-0.159982577,-0.0257185455,0.2450912893,0.1385162771,-0.1244136542,0.2277842462,0.3551832438,0.0971073061,0.1650547385,0.0904945731,0.4780934751,-0.437987715,0.3225879967,-0.3071635962,0.074354738,0.0885859951,-0.0340094566,-0.2489496171,0.2080955952,-0.240830496,0.2539783418,-0.3136201501,0.1324144155,-0.5448884964,0.0372945629,-0.0100327907,0.2946363986,0.0041617421,-0.5604776144,0.33509624,0.1862288415,0.0239839256,0.4058553278,-0.0001238069,0.1744511724,0.3322082162,-0.1320036054,-0.0576935597,-0.3747686148,-0.1845237017,-0.1774750054,0.2513703704,0.0584542006,0.1948399097,0.0846881494,-0.0990761369,-0.2675632834,0.2351568639,0.4671219587,0.0967172235,0.2684924901,0.2412029207,0.1626387089,-0.0700325146,0.1293651164,-0.0586703978,-0.0073627797,0.2534245551,-0.2842435539,-0.0843150318,0.2644785941,-0.2236341983,-0.2193562686,-0.1762116253,-0.1897657067,0.5956934094,-0.6405871511,0.2028373927,0.031608697,-0.2833524346,-0.0602123551,-0.3970498443,-0.1658379883,-0.2814686,-0.6259972453,0.2206950635,0.0547291599,-0.1740383953,0.1226082444,-0.4526482224,-0.2315917164,-0.0487773828,-0.058144398,-0.1401828974,0.5158827305,0.2073036283,0.307647109,0.1674176753,0.1691039801,0.140595302,-0.0439474881,-0.0630421638,0.1547047198,-0.3959687352,0.0444795601,0.0572552308,-0.1537994593,-0.2437874824,-0.0653126016,-0.08066836,0.1007320657,-0.1338203698,0.0974461958,0.3817778826,-0.0635505244,0.2099064738,0.3211019635,0.2443681359,-0.1982805729,-0.0408754162,0.0844115689,0.3090513945,-0.0766795278,-0.0829106718,0.3224020004,0.0780442804,-0.1507158279,-0.200878039,0.1684617549,-0.122961469,-0.1415818781,0.0955839083,0.1618539542,0.0173610169,0.7640119195,0.1510399133,0.1550600678,-0.3262269199,-0.1650597006,-0.0813882723,-0.0162959825,-0.383908242,0.1201186106,0.2784774303,-0.0458318479,-0.0025161454,-0.0125776334,-0.3755195141,-0.2901799381,0.1588222235,-0.2691520452,0.1290231645,0.6434388757,-0.0847705156,0.0378689505,-0.0311035067,-0.2311691344,0.0893548578,0.2703109682,-0.5454251766,-0.199884966,-0.1913015693,0.0830652118,-0.0090253511,0.2907006741,-0.1352845728,0.2626663744,0.4345033467,-0.3390246034,-0.3026486635,-0.3347218931,-0.4246219993,-0.3029883802,-0.0268523451,0.2384123504,-0.388689816,-0.0208522324,0.0716141015,-0.1603699178,0.3210649788,0.2287078202,-0.0067514163,0.022817051,-0.1003552526,0.1779856086,0.1289135963,-0.2334320247,-0.1171935499,0.229833886,0.0215960164,0.3336875737,-0.0718952119,-0.2325259596,0.3409836888,-0.1377684325,0.2706861496,-0.0474333502,-0.1684700102,0.0419768244,-0.445848763,-0.0475396179,-0.0104410807,-0.0907253698,0.3886668682,0.1698444337,-0.125853464,-0.4694304466,0.3113692999,-0.112603046,0.2983155847,0.1176054403,0.0700896457,-0.0038853672,0.1334573179,-0.2639538348,-0.4916706085,0.1975854933,-0.211062476,0.0598972403,-0.0548028573,0.0016005959,0.135636583,0.1703983545,-0.208983466,-0.2314667404,0.0527525395,-0.0362777449,-0.2970740497,-0.0225936715,-0.1978597641,0.6065127254,0.0651134476,0.1337962747,-0.1734601557,0.2274015099,-0.1264917105,-0.432724148,-0.3220412135,0.2733949125,-0.0145869879,-0.0576649196,-0.2054569572,0.412061125,0.4922932088,-0.0928073078,-0.1072134078,0.2081218958,0.2767289877,-0.2052212805,-0.1986125559,0.0897642225,0.0358788744,-0.1755609959,0.2432496101,0.4486861527,0.2031267583,0.3599848449,0.2320649624,0.2206025571,0.299762845,0.0055633401,-0.0274563059,-0.156145215,0.1594950259,-0.1439873427,0.1384494901,0.0725641176,-0.1243331432,-0.2617254853,0.1637731045,0.0659068897,-0.1640600413,0.0554417334,0.0439122729,-0.1209331304,0.1442405879,0.0353028253,0.3846464455,-0.0169444922,-0.1841903776,0.1875161231,-0.2923647761,-0.0041845525,0.1214142218,-0.060346216,0.4126350284,0.3325662017,0.0933479965,0.0121380128,-0.2061805278,-0.3470259011,0.1262339503,0.3581795096,-0.4770471156,0.2011184394,-0.2688544393,0.3921819627,0.1077217981,-0.0824889764,-0.251857698,-0.5163871646,-0.1252585202,0.5447285175,-0.0894417986,0.1139657944,-0.0750090778,0.0922435299,-0.1993822902,0.3349091411,-0.1236953363,-0.3494327962,-0.1855380386,-0.1118901297,0.3578799069,-0.3373328447,0.2036115825,0.1893173754,-0.2535944581,0.180045411,-0.3561492264,0.0978301764,0.0027099983,0.0007185338,0.0373477861,-0.0380866565,0.0021098915,-0.2566197813,0.1516730636,-0.0926751643,-0.2342160344,0.3143593967,-0.1155804098,0.0484064035,-0.2818495929,-0.3552564383,-0.4357712567,-0.1446245462,-0.1148808897,-0.2469650805,0.3538688421,0.2138126343,0.0279247425,0.0080096293,-0.0402045362,0.0558431558,-0.0504350811,-0.0470001921,-0.156967327,0.0019498187,-0.1449820697,-0.0489738211,0.0109178592,0.225982666,0.4416419864,-0.3533616662,0.1268649697,0.0998179391,-0.1105588973,0.2188432664,-0.2389684319,0.3640264273,0.3983311653,0.0564018935,-0.0557654351,-0.1833664924,0.1532932073,0.0631324276,0.0367054902,0.1003945768,0.3216459453,0.1986774653,0.6136776209,0.404703021,-0.0415371917,0.4049917161,-0.0742644444,0.0647051409,-0.1796070784,-0.3545342684,-0.1869882345,-0.3202335238,-0.0462964885,-0.152120322,-0.1201905012,-0.4882003069,-0.2875452936,0.5973172784,-0.3989580274,-0.22201702,-0.0257763881,-0.4343623519,0.2957084179,-0.0880863741,0.0619179569,-0.0284438133,-0.0533100441,-0.2206208557,0.1752651781,0.1265857518,-0.3411029577,-0.2711304724,-0.3298738003,-0.4660468996,0.2708948553,0.2392304689,0.6756355762,-0.0031590336,-0.0472990647,0.0279801767,-0.1610240191,0.9008379579,-0.5975762606,-0.3845264018,0.2998611629,-0.4416368008,-0.2805392146,-0.1464088708,-0.1916852593,0.5185416341,0.4178612828,0.4736289978,-0.1258116364,-0.2736027241,0.0931886956,-0.1973990798,-0.1291915923,0.0807636082,-0.2861379683,-0.01621731,-0.1824304014,0.0195708685,-0.2209484577,0.1472897679,-0.1967474967,-0.1496364325,-0.1174031794,-0.0945644453,0.2924083471,0.0564404689,0.0141238663,-0.168794319,0.0732964426,0.0023323405,0.1798488796,0.2803762853,0.2534046173,-0.421076864,-0.1872625798,0.1640625,0.0486611277,0.2841587663,0.2582045197,0.2260948271,-0.0701888129,0.0519922003,0.1801945567,-0.313906908,0.0786772221,0.3798128068,0.0034896117,-0.5512816906,-0.1097423136,-0.069743447,0.439383328,-0.1443434507,0.5627086163,-0.4565917552,-0.2755083144,0.3744552135,0.1028130427,0.757162571,-0.3558428884,-0.0179324318,0.1034672186,0.0332226492,0.102324903,-0.0298274159,0.0912708268,-0.2797409296,-0.0282336362,-0.0703882873,0.055682797,0.2865795195,0.1877313703,-0.1459397227,0.3820240796,0.015231939,-0.0896091238,-0.0244377702,0.21676144,0.4826403558,-0.0553401671,0.1282268614,-0.0104420511,-0.0659205168,-0.1713448465,0.0468599871,0.0239156894,0.2783243358,-0.1352676153,-0.3766929805,-0.03131124,-0.1759743243,0.1755469143,0.1576113105,0.1429025382,0.0067415112,0.2085309476,-0.0955410525,0.0447447076,-0.126219064,-0.1333509535,0.0964359865,0.3388732374,0.2169421762,0.0244919211,0.3706185818,0.089044705,-0.2019162774,0.0859582722,-0.2295126617,-0.2251865268,0.1175146475,0.0538832061,0.0550671667,0.1243347824,0.131767258,-0.0027598233,-0.1668452621,-0.1685528755,0.0406233966,-0.0621811673,-0.1579100192,0.4621092081,-0.0469860733,-0.3780895174,0.0285823178,0.4684546888,0.2376085222,-0.0665790737,0.1323938519,0.0757080764,-0.0553676151,-0.2345045805,-0.19471021,0.2041229904,-0.1385966688,0.3187769055,0.0618276484,-0.2533580661,0.1607473344,0.101940982,0.046793621,0.3771523535,-0.2085137069,-0.3417020142,-0.446218282,0.2933817208,0.0455734022,0.1013234705,-0.1182617322,0.1660129279,-0.1374008209,0.3116110861,-0.2197197229,0.1450562924,-0.0788094625,0.2332208008,-0.1382723302,0.0145995123,-0.0921101049,-0.1046821997,0.076147534,0.0331910513,-0.2527059019,-0.0344750173,0.0060301716,0.1244489402,0.1173038408,-0.2294049114,-0.0421263427,-0.0710528493,-0.1012940258,-0.2299918234,0.1830629855,0.3245702386,-0.2118426859,0.1726919115,0.2158292979,0.2231039107,0.0967910588,0.3091023862,-0.2309206128,0.0686679035,0.0627952516,0.4110857248,0.1045884565,-0.1674643159,-0.0104244305,0.0753285289,-0.0851463452,-0.0226071104,0.3046310246,-0.234646827,0.0113802264,0.0177814383,0.3636148572,0.3809760809,-0.1609293967,0.0975751877,0.1289530098,0.0942889974,0.0734337494,-0.1789153218,0.5030885935,0.0711812973,0.0788215026,0.2968461215,0.2957265079,-0.1840436906,0.1694109142,0.0986320972,-0.0717844591,-0.0473627523,0.2588550448,0.1763654649,0.0519246683,0.2686645687,0.1383133829,-0.3351877034,-0.1445140541,0.2858965993,0.0185991935,0.5381385684,0.0261384808,0.1412221342,-0.0872337744,-0.5816605687,0.0283281319,0.0647230148,-0.4965640008,0.2716895044,-0.0133592011,0.2167380452,-0.2738086879,-0.4163632691,-0.1943261176,0.4032902718,-0.276080966,-0.2871990502,-0.295316577,-0.1381362081,-0.0973222479,0.078092441,-0.0919771194,0.1750065535,0.741568327,0.0666250139,-0.1233505309,-0.0933173001,-0.3392289281,0.0101924296,0.3148339391,-0.0212049894,0.4458221197,0.0185701866,0.0203917306,-0.0606150255,0.0977773517,0.443887502,0.6963074803,-0.4553193152,-0.0343555473,0.1794538349,-0.0795084834,-0.0911586732,0.4181698561,0.1046571732,0.3097866774,0.2685209215,-0.0613815002,0.0192766469,-0.2396440506,0.3282628357,0.3821770847,0.0546371974,0.3476724327,0.1113528013,-0.0831607804,0.229930222,0.0156458877,-0.3268490136,-0.2684848905,0.5566337705,-0.4823427796,0.2635055482,0.2357005179,0.0446098223,0.0398492366,0.4240770638,0.2635849118,0.0626651272,-0.309312433,-0.1838689297,-0.3888843954,0.0516254939,-0.16014871,0.223379761,-0.005142733,0.0271895062,-0.0624481849,0.0350042656,-0.1013174504,-0.2399441749,0.3998103142,0.3118211925,-0.5812188387,0.1934361607,0.0182104204,-0.0433614887,0.0252557751,-0.3325001895,0.3341381848,0.0686642602,-0.1056523472,0.0980460197,0.1510054469,0.0682478845,0.1717232913,0.0574416034,0.2719837427,0.6113705635,-0.0759607553,-0.0164802223,-0.2286665738,0.1080481485,-0.1091564894,-0.0437677801,0.0593832172,0.4581898749,-0.1972041726,-0.0211739372,-0.0006724087,0.1767245531,0.0470644347,0.1126626357,-0.4025765061,0.2158271074,0.1916216761,0.064923428,-0.1029516682,0.4361132383,0.1671412885,0.3154135048,-0.2531251907,-0.2148840427,0.3275010288,-0.0833035707,-0.0832164884,-0.5843414664,0.2102661282,-0.1626892686,0.1443046033,-0.4967544973,-0.3728495538,0.1991523057,-0.2268794477,-0.6331874132,0.2344038785,-0.0566804782,-0.154506743,-0.062190216,-0.2077102959,0.136981681,-0.1371981651,0.1999458373,-0.1195721924]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Thanks for reporting.\r\n\r\n\r\nWhich tokenizers are you using ? What platform are you on ? Can you tell me which version of datasets and pyarrow you're using ? @timothyjlaurent @richarddwang @HuangLianzhe \r\n\r\nAlso if you're able to reproduce the issue on google colab that would be very helpful.\r\n\r\nI tried to run your code @richarddwang with the bert tokenizer and I wasn't able to reproduce","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":64,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nThanks for reporting.\r\n\r\n\r\nWhich tokenizers are you using ? What platform are you on ? Can you tell me which version of datasets and pyarrow you're using ? @timothyjlaurent @richarddwang @HuangLianzhe \r\n\r\nAlso if you're able to reproduce the issue on google colab that would be very helpful.\r\n\r\nI tried to run your code @richarddwang with the bert tokenizer and I wasn't able to reproduce","embeddings":[-0.4514070749,-0.0183560606,-0.0187173411,0.2539693415,0.1591782123,-0.1641509384,0.2910157144,0.3014787138,-0.0354786366,0.146110028,0.0482645482,0.4889172316,-0.4316676855,0.3358326554,-0.3009809554,0.0277109835,0.1179320961,-0.0190118216,-0.2634863555,0.2409516573,-0.2404624522,0.225755468,-0.401640892,0.1766656041,-0.6489459872,0.0602547824,-0.0938618779,0.2311900109,-0.0793512464,-0.586132586,0.3002739847,0.2776141763,0.070537962,0.4927061498,-0.0001242773,0.0863855034,0.3700333834,-0.068893984,-0.0755959973,-0.4253841937,-0.1398858577,-0.1962038279,0.326849699,0.1186022684,0.2297603786,-0.0144225005,-0.1105302274,-0.1358258724,0.295132339,0.4044710994,0.0998646319,0.2108288407,0.2679286003,0.1761285812,0.0266659819,0.1134238094,-0.029560836,0.0634082034,0.3047253788,-0.3775468469,-0.0981574059,0.1819476038,-0.2512390018,-0.2436913252,-0.1529686749,-0.1884289831,0.5462459326,-0.6417779922,0.1981923133,0.0507446788,-0.1269942969,-0.0952428058,-0.409552604,-0.1634819657,-0.2056479305,-0.5241117477,0.2612171471,0.0887136236,-0.1917433441,0.1744311899,-0.512126565,-0.1702480763,-0.0599607304,-0.0430708192,-0.1608616114,0.6147714853,0.1970791519,0.2824648619,0.1318354756,0.1094725952,0.3357935846,-0.0813709944,-0.1276591271,0.1960503757,-0.4639246464,0.0747697279,0.0073759221,-0.0823294818,-0.2194864303,0.020001635,-0.2418677807,0.1100883558,-0.0998958349,0.0787976682,0.384336561,-0.0076071243,0.2199829817,0.340882808,0.2288215309,-0.2596561313,0.0027711643,0.0371749103,0.2253895998,-0.1269138753,-0.0501925424,0.3050415814,0.1194529235,-0.2124556452,-0.2372740358,0.1194868237,-0.2387345582,-0.1723511666,0.094883658,0.1854165494,0.0307488758,0.7054215074,0.083328031,0.1961914897,-0.2631705105,-0.1573426425,-0.0654680729,-0.045228526,-0.4083966613,0.1598667949,0.2452349812,0.0504907779,-0.0003737888,0.050988391,-0.3588193655,-0.2416548431,0.1078233644,-0.2778948247,0.0396646708,0.6530490518,-0.0253033619,0.0413961373,-0.0589490347,-0.2250757962,0.0949930549,0.2815291286,-0.5119836926,-0.1841511428,-0.2612283528,0.0831530988,-0.0231683645,0.2442414463,-0.2944863737,0.2875611186,0.3829674125,-0.2585481405,-0.2988170981,-0.3133195937,-0.331846863,-0.3129889667,-0.0024773101,0.307767719,-0.4657447934,-0.0098974407,0.087454401,-0.1941268891,0.3770689368,0.1856148243,-0.0613604672,0.0669882447,-0.1267248392,0.215823561,0.140177384,-0.1279755533,-0.1124334782,0.2217224389,0.0198814105,0.2737292051,-0.0765961185,-0.2837847471,0.3305280805,-0.1273438632,0.240541622,-0.0251977723,-0.1801247448,0.055950325,-0.4307548404,-0.042464979,-0.0346674658,-0.0584328957,0.370148927,0.1658086181,-0.1280094981,-0.5494300723,0.3390696049,-0.0956562385,0.3101182878,0.0468285978,0.2010080665,0.0631357059,0.1250197142,-0.34169361,-0.4653599858,0.2084009796,-0.225909099,0.0564944372,-0.1141112745,-0.0186333749,0.0731931478,0.218865782,-0.2546346784,-0.1963988394,0.0548204556,-0.0569619983,-0.2888918519,-0.0090970388,-0.1978033781,0.6334373951,0.1330014765,0.1863329709,0.0088009676,0.3338902593,-0.1678679734,-0.4845399857,-0.2472648025,0.1851918697,0.0388390645,-0.0657138452,-0.2361524552,0.4898020327,0.4748798311,-0.1184359938,-0.0471902676,0.1707308143,0.238460511,-0.2020021379,-0.1777737737,0.0177077483,0.031491898,-0.1478212029,0.300881952,0.4878903329,0.1369510293,0.3793056309,0.1605117619,0.2333791256,0.2646214068,-0.0008948233,0.015010695,-0.1043363065,0.1605807543,-0.1033002213,0.16027686,0.0734191015,-0.0789115876,-0.2140072435,0.1174606234,-0.0077373013,-0.2083774805,0.0450219885,-0.0087575009,-0.1165949702,0.1818483919,0.0216803718,0.4007752836,0.0097523695,-0.2446662188,0.188806206,-0.2269085646,-0.0186138749,0.1428357065,-0.0703982338,0.4502498209,0.2981046438,0.1378419995,0.0256137382,-0.2035502344,-0.330483079,0.1311552525,0.411490649,-0.5120652318,0.2068114281,-0.3035531044,0.2663401067,0.0753615201,-0.1748274416,-0.3257455528,-0.5263910294,-0.1301469058,0.501065731,-0.0153540755,0.1526309401,-0.0430886745,0.075417839,-0.1909266114,0.3626380563,-0.1729192585,-0.3490327597,-0.2568057477,-0.096551016,0.4135924876,-0.226315558,0.2242790312,0.1385638565,-0.2426470965,0.1376253664,-0.3264600635,0.1218764111,-0.0456868373,0.0993344709,0.0476022996,-0.0871103406,-0.0611737072,-0.2198957205,0.1496847421,-0.1308501214,-0.2919811606,0.2979002297,-0.1592388898,0.0333513878,-0.2079053074,-0.2921704054,-0.4472483695,-0.0652050823,-0.1241436154,-0.2491881847,0.3339018524,0.1426371783,0.0502238907,-0.0268349778,-0.1156218126,-0.0214423239,-0.0339340977,-0.0353108235,-0.1688206941,0.0389618054,-0.1542485356,-0.0213817507,0.063800633,0.118798919,0.410574913,-0.3101975322,0.0174236409,0.0785765424,-0.021666104,0.2032386363,-0.2102433741,0.3423298001,0.4269163609,0.0653369948,-0.0130824624,-0.1810950488,0.0445570238,-0.0369701833,0.0857582092,0.0567320175,0.3177714348,0.1290831119,0.6365389824,0.3438287973,0.0245845001,0.3988358378,-0.1147427186,0.0726944283,-0.1767137796,-0.4179882407,-0.1399572194,-0.3385905027,-0.0034158151,-0.1102151573,-0.1246601418,-0.4634780586,-0.2985059023,0.5199305415,-0.4142768979,-0.2368219197,-0.043528907,-0.4586130381,0.293430239,-0.0736256763,0.0203414503,-0.026348738,-0.0830057338,-0.2813157141,0.2080483884,0.0749508515,-0.3284226358,-0.2657977641,-0.2576183975,-0.4092140496,0.3226155043,0.2142199278,0.6462939382,0.0636078864,0.0171444286,0.089739725,-0.1505380869,0.9170185328,-0.5295388103,-0.4050601125,0.2950895429,-0.3897665441,-0.2402521223,-0.2102429271,-0.2094303071,0.5488238931,0.4073247313,0.4270013273,-0.1365405917,-0.2981925905,0.1650775373,-0.1730670333,-0.0973168984,-0.0574685596,-0.322357595,-0.0002322448,-0.1945853084,0.101606153,-0.2478360981,0.1796011627,-0.2818056047,-0.1345826983,-0.0977982357,-0.1065054461,0.2231221944,0.1205080599,-0.0291966349,-0.2103309035,0.0621234365,0.0158566386,0.1463366896,0.402115792,0.2343917787,-0.3336496949,-0.2025294006,0.1285570562,0.027281329,0.3452246785,0.2670353055,0.1780945361,-0.0470145307,-0.0255398694,0.1622682661,-0.1858318597,0.0603776164,0.3804244101,-0.0274198055,-0.5772963762,-0.0675489083,-0.0502466634,0.4389664829,-0.1459271759,0.4760755897,-0.3540571928,-0.2528214157,0.3886637986,0.0871429965,0.8736892343,-0.3867269754,-0.0512384064,0.1110444143,0.0254610088,0.1342846453,0.0106231496,0.0341232941,-0.3378759921,-0.0024982758,-0.0574129112,0.0252357554,0.2827624679,0.2086385339,-0.1994738281,0.3970700502,-0.0344182625,-0.0258597638,0.034780439,0.1602527052,0.4071720243,-0.0319603384,0.102164574,-0.0038298646,-0.0508481376,-0.1375647187,-0.0000860812,0.036285419,0.2331690192,-0.1255358905,-0.4613443911,-0.0750199184,-0.201340571,0.1040432826,0.2047284395,0.0609578826,0.029219592,0.0799924433,-0.0750482827,0.0657527,-0.1150755808,-0.1086519882,0.1019145697,0.2996258736,0.1565519869,-0.0162426103,0.3197104633,0.0899088606,-0.2710874677,0.0669751465,-0.1930175126,-0.1885001957,0.0744520426,0.0479703993,0.0993245989,0.0851060152,0.1921526343,-0.0828155056,-0.0831783414,-0.2331730723,0.0493477285,-0.0488525182,-0.1012544408,0.4635450542,-0.0041761291,-0.3239700794,0.0002366261,0.4024417996,0.2230582088,-0.077550143,0.1868162453,0.121643737,-0.0351369455,-0.2248840928,-0.1783464104,0.2387915403,-0.1943719536,0.2779054344,0.0762504116,-0.1605111957,0.2321418673,0.1646016091,0.0615423359,0.366769284,-0.1725343615,-0.3573763669,-0.4265032709,0.25626719,0.0572028533,0.1014112234,-0.1279225498,0.286401242,-0.0851161405,0.3098717332,-0.2082177848,0.1187698245,-0.1360607296,0.2225152403,-0.1965074837,-0.0179603696,-0.161079511,-0.1194786355,0.0611644313,-0.0054542511,-0.1573917717,-0.0320138447,0.0094415033,0.1344663054,0.158411324,-0.2003346086,-0.0502524152,-0.0367936827,-0.1069698781,-0.2200430334,0.1807987243,0.3339798748,-0.2093783319,0.1154188439,0.3636812866,0.1378777772,0.0411139876,0.311267674,-0.2223465145,0.0580956265,0.0797488391,0.4565377235,0.1410527378,-0.1531626284,0.0351953283,0.0697182864,-0.0300755445,-0.0467673987,0.2754616439,-0.2803596258,0.0170983039,-0.0801739544,0.3852857351,0.4729864001,-0.1101132333,0.053667035,0.2419432551,0.1005078331,0.0127544561,-0.1003078818,0.5062838197,0.0754641294,0.0284595825,0.3011081517,0.2736853957,-0.1319648772,0.1596789062,0.143666774,-0.0089022182,0.074101463,0.2282645553,0.1235294193,0.0906235427,0.2836118639,0.2560003996,-0.3117290735,-0.1668026298,0.2614684999,-0.069992952,0.498898834,-0.0726140887,0.1265604049,-0.0666488856,-0.6106050014,0.0236301795,0.1232687309,-0.4722035527,0.1165864691,-0.0490897186,0.2964212,-0.2452362031,-0.4382399321,-0.1968893409,0.4906773269,-0.2664857507,-0.2746960223,-0.2911094725,-0.1555427313,-0.1222529709,0.0831959248,-0.1335173994,0.1632052213,0.7009337544,0.0375612602,-0.1453170329,-0.2588682771,-0.3074165583,0.062574558,0.2946562171,-0.0219730437,0.4159755707,-0.0591315366,0.0211349688,-0.0826200098,0.1767223924,0.502785027,0.741292417,-0.4434117675,-0.0956692547,0.1072113812,-0.0295981187,-0.0504573062,0.3902239203,0.1533710063,0.2753217518,0.2701470554,-0.057795018,0.0232255962,-0.211610198,0.3104964197,0.3349039853,0.0382675119,0.2798109353,0.0830269828,-0.0984350368,0.1646284312,0.070489943,-0.3213474154,-0.1564124227,0.5048130751,-0.4208494127,0.2050679028,0.1837756932,0.0468103439,0.0648642406,0.4589598477,0.3885909617,0.2099231631,-0.3096043169,-0.2707787156,-0.4379996061,0.0589852147,-0.214468196,0.2642849088,-0.0245214738,-0.0231492743,-0.0907870084,0.1096470803,-0.1216094494,-0.0915401578,0.3047743738,0.3852223754,-0.6141394973,0.2424553633,0.0491855107,-0.0252741948,-0.0407548025,-0.3058136702,0.3760474026,-0.0075479061,-0.1159946918,0.0928703323,0.1337736994,0.1216404662,0.145785436,0.0757278576,0.2860180438,0.6471262574,0.0260009784,0.1075971499,-0.2877633572,0.0207424071,-0.1586351544,-0.026822716,0.0917568132,0.3597247899,-0.1527473778,-0.0074944953,0.015790578,0.2090007514,0.0230568424,0.18942298,-0.4248897731,0.1550396681,0.2044277787,0.0485260189,-0.1781723797,0.3578691781,0.1304602623,0.3544123173,-0.3445574641,-0.2215111554,0.3429216444,-0.1219321862,-0.1181418151,-0.603833437,0.1825369298,-0.1230139434,0.0716556609,-0.4986050725,-0.3239954412,0.2401542068,-0.2024820894,-0.5594583154,0.2498544455,-0.0530508645,-0.1302678436,-0.1022438407,-0.2015973479,0.1135972068,-0.1165665388,0.1701204926,-0.0705464706]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Hi, Sorry that I forgot to see what my version was.\r\nBut after updating datasets to master (editable install), and latest pyarrow. \r\nIt works now ~","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":26,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nHi, Sorry that I forgot to see what my version was.\r\nBut after updating datasets to master (editable install), and latest pyarrow. \r\nIt works now ~","embeddings":[-0.4262234271,-0.028084619,-0.0307077449,0.1988400072,0.1331433058,-0.1375346929,0.3073057234,0.3333301246,-0.0064558061,0.1290375143,0.0286434963,0.4368834794,-0.4107373059,0.2967744768,-0.2640088201,0.028001057,0.1197290346,-0.0393833667,-0.3695648313,0.2227409482,-0.2434656173,0.2219716311,-0.3648296595,0.1880379766,-0.5168523192,0.0857324377,-0.0786062628,0.2741605937,-0.0307958685,-0.613079071,0.3389736414,0.2480798215,0.0578998439,0.4376125038,-0.0001256642,0.1473097205,0.391125828,-0.0815784633,-0.072228916,-0.3685734272,-0.1584385782,-0.1926398426,0.3192971647,0.0929173976,0.2308609635,-0.0479284786,-0.062031623,-0.1443226188,0.2170929909,0.4277088046,0.0917039514,0.2471802831,0.3233218193,0.1382610202,0.1143372357,0.1332390308,-0.0305649061,0.0948628336,0.3123919666,-0.3872240484,-0.1458236128,0.2094108909,-0.2584967017,-0.21596919,-0.1401004195,-0.1977108121,0.7033735514,-0.6043235064,0.1513261199,0.0687063262,-0.1421921402,-0.1105682552,-0.4584234655,-0.1897011548,-0.2329024971,-0.4672604501,0.2183406353,0.0801673532,-0.1841587573,0.1617284268,-0.4882292747,-0.1844171584,-0.0824079812,-0.0564404763,-0.1429248899,0.5605838895,0.2558191717,0.3145769835,0.1811260134,0.135893479,0.2715713084,-0.015385123,-0.1004991904,0.1912435591,-0.4333979189,0.0872442573,0.0809166357,-0.0300569236,-0.2048361301,-0.0127464198,-0.1876092702,0.0651656166,-0.0511016473,0.0766429603,0.4456846714,-0.0959976241,0.1797675639,0.2698792815,0.2719274461,-0.2075210959,0.0286727808,0.0807524174,0.3131726086,-0.1496840417,-0.0081414171,0.3094294965,0.1514962316,-0.1890536845,-0.1789469421,0.1035742611,-0.156428948,-0.2262350023,0.0571264736,0.2013350278,-0.0015099574,0.7359142303,0.0622512847,0.1686503291,-0.295674026,-0.0981428623,-0.0687071383,-0.0478754975,-0.4130519629,0.0763139352,0.2600796521,-0.0466352887,-0.0120455222,0.0436597578,-0.3395041823,-0.2060766071,0.0464951545,-0.2673592865,0.0518141054,0.6857979298,-0.0926611796,0.0158425719,-0.1071029082,-0.1668681949,0.0837772638,0.3180404902,-0.4997598827,-0.2156490535,-0.2343009859,0.0743628293,-0.0113553638,0.2890097499,-0.1569269449,0.2223335803,0.4243979752,-0.3984230459,-0.28088817,-0.3528041244,-0.2661394775,-0.3301196694,-0.0200220216,0.2311418504,-0.4519298673,-0.0070786849,0.0694646463,-0.2467421442,0.3404714167,0.2008070201,-0.0547829829,0.0209852308,-0.1204581261,0.1243193001,0.121008344,-0.1624820381,-0.170026511,0.2329113632,0.0909360722,0.3055431545,-0.0249607041,-0.2829441428,0.2935257554,-0.1453175694,0.1870567799,-0.0470500588,-0.1874552965,0.0637622327,-0.414737314,-0.0254800655,-0.0138666788,-0.0490526706,0.3853577077,0.1553201228,-0.0500654541,-0.5330550075,0.395567894,-0.1283024848,0.2941477299,0.0562965795,0.1294917762,0.0282950383,0.1651160866,-0.3353672028,-0.5447026491,0.2167762667,-0.3194432557,0.0381356739,-0.1189157814,-0.0113432445,0.1209759414,0.2229211479,-0.2015701085,-0.1836010665,0.0278964769,-0.0479693078,-0.2874882817,-0.0581980646,-0.2361510843,0.612225771,0.1214624867,0.1500258893,-0.090140231,0.2515523732,-0.1920496225,-0.4561068714,-0.2289396077,0.2139546871,-0.0094456999,-0.0951695293,-0.191960305,0.4771188498,0.4192665219,-0.0924054906,-0.1254468262,0.0974677205,0.263250947,-0.2146302909,-0.1903713793,0.0501568615,0.0394390374,-0.1361795366,0.2064519823,0.4271456897,0.1136481687,0.3887144923,0.2086599022,0.1980692595,0.2781986594,0.007067502,-0.011525684,-0.1324376613,0.1837469488,-0.1079408154,0.2337141633,0.0872981176,-0.1605630666,-0.2234201878,0.1846192926,-0.0384719744,-0.1715289205,0.0879378021,0.0761116371,-0.1311678737,0.1681315303,0.0018950755,0.3951360583,0.0263035204,-0.1902967095,0.1835365295,-0.3049235642,-0.0213664249,0.1568355113,-0.082507059,0.4652839303,0.3312572539,0.1206734404,-0.0191460829,-0.2211041749,-0.29788059,0.1129528359,0.3866342306,-0.5138962269,0.1831787229,-0.2872391343,0.3498784304,0.0920325518,-0.1949675083,-0.3359868526,-0.5282185078,-0.1070323214,0.5051035285,-0.0670930669,0.1802400053,-0.0582263619,0.0198349655,-0.1841710806,0.3257264793,-0.1978469044,-0.3439945579,-0.229653284,-0.1021254659,0.382671088,-0.3052901328,0.1771934777,0.1326709688,-0.2272641957,0.1391625851,-0.3415319622,0.0712252408,-0.0616249181,0.1284907758,0.1239015758,-0.0415203124,0.0358644836,-0.1854022592,0.1367173195,-0.1725316942,-0.2678056061,0.3011958897,-0.1624243408,0.0868372023,-0.2424321771,-0.3241862655,-0.4599994421,-0.0813023672,-0.0785735399,-0.2051275969,0.3805841804,0.1081921831,0.0498007089,-0.0413513668,-0.0785893053,0.0359862335,0.0005276261,-0.0447287709,-0.160464704,0.0560941435,-0.1445963532,-0.1038989797,0.0696532726,0.0801027864,0.4333178997,-0.3374870121,0.0384624526,0.081588611,-0.0312826969,0.2492198199,-0.1725509465,0.3564426303,0.446968317,0.0673353747,-0.0272414349,-0.191327557,0.0769577026,0.00462913,0.0519374236,0.1071616635,0.3089773059,0.1419169307,0.6148813367,0.3018051982,0.0772748068,0.4116166234,-0.1048776433,0.1169989556,-0.1796388328,-0.3834992647,-0.1815125644,-0.3323112428,0.0179434922,-0.1800160557,-0.1385158449,-0.5107507706,-0.29469046,0.508361578,-0.3593627214,-0.2251893282,-0.0294327512,-0.415838331,0.2737881541,-0.0482285991,0.0361039154,-0.013067211,-0.0562718883,-0.2096546888,0.1837164164,0.115562737,-0.2950021029,-0.2535847425,-0.3127793968,-0.4444827139,0.2949635983,0.2194478065,0.6344886422,0.0734112635,-0.0596753806,0.0646126792,-0.2058968246,0.9419652224,-0.5384346843,-0.3539143205,0.2767045498,-0.3353812993,-0.2960503697,-0.2124645859,-0.1766382903,0.51357162,0.3529110551,0.3994944394,-0.19023031,-0.3092124462,0.2457107753,-0.1942550391,-0.1138672084,-0.0568231568,-0.3268678784,-0.0149414688,-0.1858353615,0.0560887903,-0.3065467477,0.1166410297,-0.2546600699,-0.0739465505,-0.102065891,-0.0674841255,0.2043437511,0.0934561193,-0.0091585154,-0.2332925647,0.0409650169,0.0166023728,0.1896851659,0.3295447528,0.2695171237,-0.390627861,-0.1522639096,0.1227662861,0.0285160486,0.3404765129,0.2600503266,0.1857448667,-0.0897349492,0.0456196852,0.15678069,-0.2219396681,-0.0335156247,0.3762093484,-0.0683572367,-0.570235312,-0.1211678386,-0.0074177259,0.4382996559,-0.1868736744,0.4402972758,-0.3790718913,-0.25620839,0.3504360914,0.094413884,0.8294622898,-0.3685491383,0.007513775,0.1412825286,-0.0291989632,0.0943374485,0.0638992861,0.0500576459,-0.3382294476,0.010419582,-0.0711940974,0.0385398045,0.2984247804,0.1840958446,-0.1552243531,0.4556515217,0.0017167522,-0.0308157802,0.0527689382,0.1311539412,0.4662715793,-0.0224106051,0.0719884038,-0.0450139157,-0.0487071835,-0.2092828751,0.0159176793,-0.0253669843,0.2764601409,-0.1492100954,-0.3791146278,-0.0607022718,-0.1620898843,0.1172120571,0.2236527503,0.0434423275,-0.024061434,0.0903695673,-0.1618461609,-0.0048663532,-0.0905561224,-0.1171319559,0.0959542915,0.3593365252,0.1913869679,0.0502057225,0.3596635759,0.0885729864,-0.344055593,0.0759111792,-0.2765969634,-0.2300830781,0.0411389098,0.0753662586,0.0709391981,0.1175203174,0.129113704,-0.089992851,-0.1023082435,-0.2025319636,0.0411028937,-0.0845111161,-0.0814070776,0.5126656294,-0.0769249424,-0.34769696,-0.0207589064,0.4428584874,0.2637453973,0.0066066128,0.1803859174,0.1175712571,-0.0454971753,-0.2055634558,-0.2149389386,0.2369129062,-0.247435838,0.23628892,0.0953164697,-0.2004565299,0.2147379518,0.018135244,0.0912797078,0.3555853665,-0.2081534117,-0.3337844312,-0.4716787934,0.2893686295,0.0941013768,0.0819017217,-0.0503457375,0.2005216181,-0.0628676564,0.2897004187,-0.2136404961,0.1265349388,-0.053728573,0.2342295945,-0.1462177932,0.0019320027,-0.1427261978,-0.1322265565,0.0681190863,-0.0060307384,-0.2096684575,-0.0245789476,0.0344262421,0.1419171691,0.0874610618,-0.2029487938,-0.035299819,-0.0580195189,-0.0988910273,-0.1757137626,0.2244656682,0.3153636456,-0.1890924871,0.1739246398,0.2851654887,0.1327942908,0.0680408403,0.3327769339,-0.2285962552,0.0301887076,0.0431795828,0.4585475028,0.0703362375,-0.1614503413,0.0679132342,0.080925703,-0.0829224586,-0.0151953232,0.3467742503,-0.2715690136,-0.0050467663,-0.0690944269,0.457698673,0.440900296,-0.1034994572,0.0886817276,0.15685983,0.0879037827,-0.0480685681,-0.1489168704,0.5387001634,0.1000275016,0.0745915771,0.2959090173,0.3054138124,-0.1664310545,0.2006227076,0.1431748867,-0.0360420048,0.0572848357,0.1860654801,0.2724135518,0.1090565696,0.2569830418,0.2660528719,-0.2749657035,-0.1071283743,0.3567866981,-0.0522896014,0.4890805781,-0.0663105175,0.1197936088,-0.0846301913,-0.610498786,0.0734566078,0.1625179946,-0.4453828931,0.1463365406,0.0180167686,0.3133052886,-0.1979468316,-0.4583253562,-0.1947428882,0.4319702089,-0.2319599688,-0.3544824421,-0.2850638628,-0.1753609627,-0.0590662956,0.0898280516,-0.1454727203,0.1400508434,0.799793303,0.05380385,-0.0667098537,-0.2466627061,-0.3139773309,-0.00527613,0.3108201325,0.0006736455,0.4508585036,0.0546375737,-0.0182308983,-0.0618825443,0.152304098,0.5050356388,0.6981099248,-0.4474122226,-0.0480050482,0.1089005768,-0.0523091517,-0.0862674937,0.3647112548,0.1321835816,0.3019934595,0.2963252068,-0.0656322241,0.0374661982,-0.1313718706,0.3382485807,0.3452504873,0.0240487903,0.3023178279,0.0798646435,-0.1416295767,0.2565799654,0.0797422528,-0.3255626857,-0.1952850521,0.569218576,-0.4321996868,0.2212009877,0.1722597629,0.0486828275,0.1094018742,0.5082313418,0.3306356668,0.164489314,-0.252173692,-0.1733622849,-0.4439534843,0.0342976898,-0.1946024746,0.3066214323,0.051565852,0.0016597186,-0.0660617799,0.0451494455,0.0099889841,-0.1932028979,0.3324364126,0.3729103506,-0.585285008,0.2185807675,0.0498895384,-0.0063305278,-0.0412900075,-0.3142338395,0.3802556098,-0.0498876087,-0.1108638644,0.0644587874,0.1700888425,0.1745327711,0.2145899683,0.0503830388,0.2386428714,0.5973000526,-0.0063547953,0.0381645672,-0.2856546938,0.0393655896,-0.1230392903,-0.0162282586,0.0599050559,0.3647589386,-0.1982252747,-0.0922024325,0.0259149466,0.2058888823,0.0098139355,0.2247241288,-0.3983497322,0.1691613197,0.2343987077,-0.0028144224,-0.2005940378,0.3361578286,0.1607359648,0.3032528758,-0.3313165903,-0.2211698443,0.3441254795,-0.0998627171,-0.1244290248,-0.5944117904,0.1723879278,-0.1716984212,0.0901962221,-0.5051227212,-0.3045918643,0.2259376198,-0.1941107064,-0.5847123861,0.2045534849,-0.0807110816,-0.1077307239,-0.0857334062,-0.2042675465,0.1254522055,-0.1983261853,0.1375613213,-0.1398181468]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Sorry, I just noticed this.\r\nI'm running this on MACOS the version of datasets I'm was 1.0.0 but I've also tried it on 1.0.2. `pyarrow==1.0.1`, Python 3.6\r\n\r\nConsider this code:\r\n```python\r\n\r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path, name=\"prodigy-ds\", data_files=list(file_paths), cache_dir=cache_dir\r\n )[\"train\"]\r\n valid_relations = set(vocabulary.relation_types.keys())\r\n\r\n ds = ds.filter(filter_good_rows, fn_kwargs=dict(valid_rel_labels=valid_relations))\r\n\r\n ds = ds.map(map_bpe_encodings, batched=True, fn_kwargs=dict(tokenizer=vocabulary.tokenizer), num_proc=10)\r\n\r\n # add all feature data\r\n ner_ds: Dataset = ds.map(\r\n add_bio_tags,\r\n fn_kwargs=dict(ner_label_map=vocabulary.ner_labels, tokenizer=vocabulary.tokenizer),\r\n )\r\n rel_ds: Dataset = ner_ds.map(\r\n relation_ds_factory,\r\n batched=True,\r\n writer_batch_size=100,\r\n fn_kwargs=dict(tokenizer=vocabulary.tokenizer, vocabulary=vocabulary),\r\n )\r\n```\r\nThe loader is essentially a jsonloader with some extra error handling. The data is a jsonlines format with text field and a list of span objects and relation objects. \r\n\r\nIn the `ner_ds` a field, `ner_labels` is added, this is used in the downstream `relation_ds_factory`. It all runs fine in a single process but I get a KeyError error if run with num_proc set\r\n:\r\n\r\n```\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 348, in relation_ds_factory\r\n ner_labels = example[\"ner_labels\"]\r\nKeyError: 'ner_labels'\r\n``` \r\n\r\nThis is just one example of what goes wrong. I've started just saving the dataset as arrow at the end because it takes a long time to map\/filter\/shuffle and the caching isn't working (tracked it down to byte differences in the pickled functions). \r\n\r\n^^ Interestingly if I heed the warning from Tokenizers and set the environment variable, `TOKENIZERS_PARALLELISM=true` the map just hangs:\r\n\r\n```\r\n[I 200921 21:43:18 filelock:318] Lock 5694118768 released on \/Users\/timothy.laurent\/.cache\/huggingface\/datasets\/_Users_timothy.laurent_.cache_huggingface_datasets_prodigy_dataset_builder_prodigy-ds-5f34378723c4e83f_0.0.0_e67d9b43d5cd82c50b1eae8f2097daf95b601a04dc03ddd504f2b234a5fa247a.lock\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.34ba\/s]\r\n#0: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#1: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#2: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#3: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#4: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#5: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#6: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#7: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#8: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n```","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":289,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nSorry, I just noticed this.\r\nI'm running this on MACOS the version of datasets I'm was 1.0.0 but I've also tried it on 1.0.2. `pyarrow==1.0.1`, Python 3.6\r\n\r\nConsider this code:\r\n```python\r\n\r\n loader_path = str(Path(__file__).parent \/ \"prodigy_dataset_builder.py\")\r\n ds = load_dataset(\r\n loader_path, name=\"prodigy-ds\", data_files=list(file_paths), cache_dir=cache_dir\r\n )[\"train\"]\r\n valid_relations = set(vocabulary.relation_types.keys())\r\n\r\n ds = ds.filter(filter_good_rows, fn_kwargs=dict(valid_rel_labels=valid_relations))\r\n\r\n ds = ds.map(map_bpe_encodings, batched=True, fn_kwargs=dict(tokenizer=vocabulary.tokenizer), num_proc=10)\r\n\r\n # add all feature data\r\n ner_ds: Dataset = ds.map(\r\n add_bio_tags,\r\n fn_kwargs=dict(ner_label_map=vocabulary.ner_labels, tokenizer=vocabulary.tokenizer),\r\n )\r\n rel_ds: Dataset = ner_ds.map(\r\n relation_ds_factory,\r\n batched=True,\r\n writer_batch_size=100,\r\n fn_kwargs=dict(tokenizer=vocabulary.tokenizer, vocabulary=vocabulary),\r\n )\r\n```\r\nThe loader is essentially a jsonloader with some extra error handling. The data is a jsonlines format with text field and a list of span objects and relation objects. \r\n\r\nIn the `ner_ds` a field, `ner_labels` is added, this is used in the downstream `relation_ds_factory`. It all runs fine in a single process but I get a KeyError error if run with num_proc set\r\n:\r\n\r\n```\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 348, in relation_ds_factory\r\n ner_labels = example[\"ner_labels\"]\r\nKeyError: 'ner_labels'\r\n``` \r\n\r\nThis is just one example of what goes wrong. I've started just saving the dataset as arrow at the end because it takes a long time to map\/filter\/shuffle and the caching isn't working (tracked it down to byte differences in the pickled functions). \r\n\r\n^^ Interestingly if I heed the warning from Tokenizers and set the environment variable, `TOKENIZERS_PARALLELISM=true` the map just hangs:\r\n\r\n```\r\n[I 200921 21:43:18 filelock:318] Lock 5694118768 released on \/Users\/timothy.laurent\/.cache\/huggingface\/datasets\/_Users_timothy.laurent_.cache_huggingface_datasets_prodigy_dataset_builder_prodigy-ds-5f34378723c4e83f_0.0.0_e67d9b43d5cd82c50b1eae8f2097daf95b601a04dc03ddd504f2b234a5fa247a.lock\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1.34ba\/s]\r\n#0: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#1: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#2: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#3: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#4: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#5: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#6: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#7: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n#8: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\n```","embeddings":[-0.3902367055,-0.0592111722,-0.0349889807,0.2178634107,0.1645161361,-0.1262937039,0.2920169532,0.2759281397,0.0126609746,0.1588909626,0.027785698,0.5151471496,-0.4663677812,0.1931906193,-0.3737515807,0.0636929423,0.1268112361,-0.0707599223,-0.2591288388,0.2371894121,-0.2843744755,0.2678748965,-0.3191679418,0.132583499,-0.4713807702,0.0843237191,-0.0683147013,0.3312543929,-0.0138635291,-0.5806868672,0.313513577,0.2603117228,0.0385840237,0.4804727435,-0.0001283317,0.1993655264,0.3485915363,-0.0834867507,-0.1267864257,-0.4279059172,-0.1911366284,-0.1817952991,0.3719674647,0.061871592,0.2269573808,0.065886721,-0.1034057289,-0.2707560956,0.2035187483,0.4774683416,0.0735512674,0.2868750393,0.2374197096,0.1393536776,0.1045707241,0.2273843884,-0.0284487754,0.1546577513,0.3186120391,-0.3993650973,-0.0570379756,0.2126248181,-0.2640066445,-0.190993011,-0.1678403616,-0.1613083333,0.6544430852,-0.5508720875,0.1051117852,0.0620830394,-0.279481411,-0.0378321335,-0.4952349067,-0.2726466954,-0.202734828,-0.4610621333,0.2794249058,0.1045864448,-0.1968158334,0.1120820493,-0.533808291,-0.1942967772,-0.0470323078,-0.0213961508,-0.146976158,0.4835234284,0.201926887,0.3012487888,0.2571425736,0.1812760085,0.2967933416,-0.0528956242,-0.0266065784,0.1361504942,-0.3973773122,0.1082416475,0.0853555724,-0.0201199614,-0.2698212862,-0.1237291172,-0.2139736712,0.0115729384,-0.1222900674,0.0902782008,0.401918143,-0.1053700149,0.1600659192,0.3277313113,0.2415083647,-0.1646747738,-0.0319317169,0.049474217,0.3143293262,-0.0866266862,-0.0327453688,0.253067106,0.1316755861,-0.1706767082,-0.2822547555,0.1109131128,-0.1060232967,-0.1977370679,0.0632683411,0.186062023,0.0114508094,0.7769070864,0.1179558113,0.2578349113,-0.3525600135,-0.0100732474,-0.0376886874,-0.0341650657,-0.3728174269,0.1222595721,0.1942777336,-0.0074799247,-0.0251636375,0.0497132204,-0.376331687,-0.2689580321,0.0794303417,-0.237399295,0.0149759576,0.6824306846,-0.1149511337,0.0577007979,0.0015735601,-0.0498101749,0.0627911463,0.3190524578,-0.4979275465,-0.1897996068,-0.2720981538,0.0515523367,-0.0527851842,0.2867564857,-0.1496298462,0.1937617064,0.4530760646,-0.3586957455,-0.2541362047,-0.3936060667,-0.3620464206,-0.3363011181,-0.0959209949,0.2916580439,-0.4016878307,-0.0132419793,-0.0017041432,-0.1993289739,0.3723531961,0.1838982105,-0.0309001766,0.0506740324,-0.1230260506,0.0603297278,0.1363580674,-0.1564815044,-0.1171422005,0.2111741006,0.0261724386,0.338583082,-0.0083542829,-0.2690427899,0.2257904112,-0.1163293719,0.1201394945,-0.0425291732,-0.1851770431,0.0972474888,-0.3684078157,-0.085811004,-0.0229808185,0.0099370144,0.3281287253,0.2012072057,-0.0785196722,-0.5475058556,0.359230727,-0.137417838,0.2546218038,0.1185212508,0.0758690834,0.103861019,0.0685261711,-0.3592191935,-0.4807923734,0.1885324717,-0.2075800449,0.0658969432,-0.1936080009,-0.0030003553,0.1345405281,0.2470803112,-0.2068517506,-0.1502073258,-0.009307676,0.0411521383,-0.269040525,-0.009131371,-0.2266809344,0.5915706754,0.1010039225,0.1743608713,-0.1415229887,0.2607125342,-0.1526126117,-0.4264979661,-0.2687478364,0.1833890527,-0.0396083258,-0.1014267504,-0.2160593718,0.4433250427,0.526933372,-0.0824615657,-0.1035685763,0.0848543718,0.2232703418,-0.1494114995,-0.1771420538,0.0278939512,0.0121736052,-0.1099253148,0.1677223295,0.4997619092,0.1795986891,0.4273737669,0.1819782108,0.2141730487,0.2338954657,-0.0310655665,0.07588806,-0.1896802485,0.2151067704,-0.0762616768,0.1706488878,0.0909808427,-0.1406539679,-0.1778077036,0.1783758551,-0.0294325668,-0.223291412,0.1606327742,0.1228898689,-0.0887502432,0.1456647664,0.0523058921,0.5058743954,-0.0183926448,-0.1601222306,0.1790274084,-0.2569560409,-0.0544666089,0.1409515887,-0.0223163348,0.4753065109,0.2240649611,0.1380575448,0.0309247989,-0.1429341882,-0.3365243375,0.1123440415,0.4086907804,-0.5074980855,0.2802318335,-0.2813071012,0.35570997,0.100207448,-0.1912258863,-0.3339085281,-0.574213028,-0.1533732563,0.5668952465,-0.0479360223,0.112604849,0.0036664954,0.0170777421,-0.1395082027,0.2880706787,-0.1432930082,-0.2916991711,-0.1898858547,-0.1433071494,0.351377964,-0.3711665273,0.1469879597,0.1352257282,-0.2889918387,0.1344395876,-0.2641728818,0.078927435,-0.0919097662,0.0494198874,0.1642391533,-0.0096581923,0.0157181714,-0.1921673864,0.1474490762,-0.1172628701,-0.2671200931,0.2706723511,-0.0943267047,0.1018959284,-0.2709847987,-0.241150558,-0.4557641149,-0.1072819084,-0.0832757503,-0.2724333107,0.4007447362,0.095407702,0.1183815822,-0.0040317304,-0.1268735975,0.0669996664,-0.0192441586,-0.1305200756,-0.1743671447,0.060360048,-0.1189939752,-0.0972301364,0.0485861599,0.1815907657,0.4506176412,-0.4062321782,0.019671455,0.0632361919,0.0106000947,0.1800069809,-0.0984769315,0.3727628589,0.4366258979,0.0714963973,-0.0349689238,-0.1571934372,0.0263095386,-0.010599629,0.1163603961,0.1012194455,0.3563603163,0.0692335144,0.6154002547,0.3028970361,0.024001047,0.4207400084,-0.1140316427,0.1770393252,-0.1278588474,-0.388689816,-0.2190794051,-0.2585810423,-0.0364087708,-0.1498003453,-0.1143195704,-0.4521029592,-0.3066712022,0.5371510983,-0.3963173032,-0.227589339,-0.0357149169,-0.5229926705,0.3197062612,-0.1332027614,0.0923333168,-0.0210317355,-0.0123196105,-0.2608735263,0.1982918233,0.1756755263,-0.3334382176,-0.1989694238,-0.2547504604,-0.4235689938,0.2796332836,0.2617287934,0.6507911086,-0.0224347841,-0.0912636369,0.0490727499,-0.1589063853,0.9467081428,-0.5530065298,-0.3837415874,0.3207136989,-0.3916186094,-0.4114246368,-0.1919744164,-0.1682036221,0.5057574511,0.4449552298,0.426618576,-0.1363418698,-0.2412991673,0.1054762527,-0.1660167128,-0.0559677556,-0.1100163534,-0.3457888067,-0.0248545296,-0.2040703446,0.0148845473,-0.2748688757,0.136773482,-0.2351040542,-0.0453107171,-0.1065619811,-0.0417629145,0.2391806245,0.0945946723,-0.0620733239,-0.2242544293,0.0294657927,0.0364423804,0.0655658022,0.3162377477,0.3880427182,-0.4297848344,-0.2748398185,0.1313022524,0.0363301709,0.3011677265,0.2805289924,0.2016389966,-0.054517813,0.0339006037,0.1890442818,-0.2018795907,0.0521096326,0.3753204346,-0.006168121,-0.5099647641,-0.1230169013,-0.0394013822,0.4427990615,-0.1759002358,0.4935839176,-0.4076273739,-0.2237099558,0.3691267073,0.1325159371,0.8098657131,-0.3974618018,0.0769528225,0.1735082418,-0.0470371284,0.1863830388,0.0982898474,0.0292272866,-0.2433085144,0.0711705461,-0.0831505954,0.0216637403,0.3116820753,0.203614518,-0.121725373,0.3935730457,-0.0482354984,-0.1080324277,0.0144361984,0.1247010529,0.416215539,-0.0682364628,0.0542986318,-0.0463159978,-0.0153230019,-0.1686324924,0.0636389628,0.0377892517,0.208099708,-0.0833262205,-0.3949980736,-0.0183961522,-0.1994376779,0.1233626753,0.2540481389,0.0323194116,0.0004211348,0.1316601187,-0.1175600812,-0.0195802301,-0.0541591644,-0.1245338619,0.1197451577,0.3622776866,0.1966803372,-0.0266353656,0.3706836104,0.1262274981,-0.2895875871,0.0818761066,-0.2871003151,-0.3017195463,0.0900706798,0.1001078039,0.0619025789,0.0978557542,0.1198118627,-0.1417674571,-0.0813701749,-0.1729373187,0.0055496059,-0.0217465106,-0.0733123943,0.5219823718,-0.1842380166,-0.4113768637,-0.0405900553,0.475785464,0.2396560311,-0.0272116885,0.1421350986,0.1073061526,-0.0383118875,-0.191433832,-0.182170555,0.1619195491,-0.118245542,0.30265221,0.1182357892,-0.197013557,0.2240915895,0.0110590216,0.0495459735,0.3996921778,-0.2089018524,-0.4455861151,-0.4509825408,0.1496695727,0.0523024388,0.0951905176,-0.0578688085,0.1346494108,-0.0781704485,0.232291773,-0.1699384898,0.1046628058,-0.078807205,0.1839051992,-0.0657351613,-0.003337408,-0.0813558996,-0.1864002347,0.0610802025,-0.0536475703,-0.1819944233,0.0190771725,0.0090599749,0.159750849,0.0806204155,-0.1893983483,-0.040415328,-0.1286808401,-0.0983985364,-0.1971827,0.1859801561,0.296523571,-0.2234840542,0.1309629232,0.1407267004,0.1391110718,0.1234747693,0.2911407351,-0.261875689,0.081275396,0.0843400955,0.4735639095,0.1265994757,-0.1444475502,0.0068721795,0.0722144321,-0.0797407925,-0.0387322046,0.315366745,-0.214203462,0.0150712859,-0.038525369,0.4274146259,0.4411705732,-0.112010777,0.0696169734,0.1674492657,0.041415289,-0.0065671541,-0.1578631997,0.4595554769,0.1683235615,0.0615440682,0.3693101704,0.2341263741,-0.1762973517,0.2079518735,0.1163485795,0.009393272,0.1069398448,0.228132993,0.2055075765,0.0888195634,0.2472228855,0.2893257737,-0.2717821002,-0.1593868583,0.3207097352,-0.0138620324,0.5487515926,-0.1250101179,0.0803006738,-0.0502972677,-0.603381753,0.0717619807,0.065814808,-0.4713745713,0.1535793692,-0.0319702066,0.2851427197,-0.3424899578,-0.4615472853,-0.1285650581,0.4549022913,-0.2151285857,-0.2811872363,-0.2522721291,-0.1286940873,-0.0143393762,0.0968538374,-0.0987995341,0.076801464,0.8586918116,0.1223392934,-0.0633001328,-0.2654591203,-0.3154933751,-0.0107816188,0.311997354,-0.0386177674,0.4298146963,0.0076614562,0.0211525448,-0.0095239133,0.0986580849,0.5285570621,0.7020314336,-0.4172759056,-0.0881650522,0.0402847417,-0.0154563803,-0.076982148,0.3674937189,0.0732598752,0.444224745,0.30282408,-0.0867998153,0.054405991,-0.1054819524,0.3087967932,0.355712831,0.0208868068,0.3362545669,0.0463115051,-0.106468372,0.1838287115,0.0949918106,-0.3379248977,-0.2214144468,0.5490166545,-0.3919815719,0.2346206903,0.184907198,0.0182023123,0.0998835191,0.4574817717,0.3904993236,0.2112106085,-0.2984004617,-0.229690358,-0.4322270751,0.0492711738,-0.2420560867,0.1895724982,0.0139313191,-0.0174988266,-0.0689750686,0.0357956924,-0.0441640057,-0.1033590883,0.2798515558,0.312505126,-0.5940951109,0.1194919795,0.0456101708,-0.0892166793,-0.0052895411,-0.2988870442,0.3827373683,-0.0838379934,-0.1470669806,0.0633486286,0.1591054499,0.0582985096,0.2646428645,0.0672972575,0.2498590201,0.5691511035,0.0288309027,-0.0433519259,-0.2855929136,0.0926685333,-0.1503719389,-0.0287022684,0.0138799157,0.4206104279,-0.24538064,-0.0113225011,-0.0500105955,0.230975315,0.0593188405,0.1608801782,-0.3077153563,0.1874887645,0.2092629522,0.0709565729,-0.1115401909,0.2920920253,0.1933348924,0.35808447,-0.3929083645,-0.1799938232,0.3470683992,-0.1066254228,-0.1602694541,-0.4940546453,0.1251235753,-0.2230120003,0.1019254848,-0.6306145787,-0.3223098516,0.2625123858,-0.2387555093,-0.5590649247,0.1612033099,-0.0511401072,-0.1086177528,-0.105739437,-0.1461658925,0.0927547291,-0.1470592618,0.1850190312,-0.1125658154]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Thank you, I was able to reproduce :)\r\nI'm on it","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":11,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nThank you, I was able to reproduce :)\r\nI'm on it","embeddings":[-0.4101415575,-0.1790640652,-0.0430235974,0.2468688488,0.1407928318,-0.1577992886,0.2643332779,0.3121121526,0.1394723356,0.1607406586,0.0909069702,0.4094497561,-0.4560320973,0.3588190377,-0.343241632,0.0205528289,0.0853096023,-0.0600768961,-0.3569635153,0.2197032124,-0.2797016799,0.2653374374,-0.3396466374,0.1743323058,-0.519687891,0.0801019818,-0.0191707667,0.2728618383,-0.0031129692,-0.5582129359,0.3405413032,0.2201202214,0.0398693569,0.4308939278,-0.0001247126,0.122180067,0.3753521144,-0.1222560033,-0.0411701165,-0.3427047431,-0.2428434491,-0.1973312199,0.2993862331,0.0928841755,0.1936341375,0.0447824821,-0.1087390184,-0.2774657011,0.2237655818,0.4397250414,0.0937338918,0.2013761997,0.2791492641,0.1225062534,0.0096454462,0.1368162483,-0.0305004716,0.0473538935,0.3109394908,-0.3475981355,-0.0823922679,0.2536458671,-0.2262710482,-0.2042798698,-0.1474829465,-0.226615876,0.6760360003,-0.643080771,0.1846373975,0.0553103201,-0.1988876611,-0.013436961,-0.4216750264,-0.172252208,-0.2398082912,-0.5042340159,0.2336415648,0.0478539318,-0.1645748764,0.140189454,-0.4830957055,-0.2291193157,-0.0599837676,-0.0746501014,-0.1020793989,0.540238142,0.1948875189,0.3084171712,0.1625609696,0.117881462,0.1880409718,0.0175433774,-0.0749703646,0.1726723909,-0.434179455,0.0898015276,0.071250692,-0.0688220263,-0.253164649,0.0092102587,-0.1893303245,0.1016406715,-0.0969357416,0.0698183998,0.4214612246,-0.1020424366,0.1938752532,0.2934978902,0.2494896054,-0.2059928775,-0.0047403434,0.0405348763,0.3376119733,-0.1144567281,-0.0415709391,0.2997927666,0.1458345205,-0.2041462511,-0.1799636632,0.1493112445,-0.0752398148,-0.1993283033,0.0954562873,0.190964669,0.0444191024,0.7813181877,0.144151926,0.1712786108,-0.3040715158,-0.0613903217,-0.0545224324,-0.0482420735,-0.3861935735,0.0819059089,0.2661402225,0.0243291501,-0.0011291,0.0351777412,-0.3368368745,-0.2812681198,0.1315496564,-0.2978740633,0.0797573403,0.6891893148,-0.0959676951,0.020195974,-0.0513266958,-0.1086239889,0.0808543637,0.2715372443,-0.5269290805,-0.209209919,-0.2119809687,0.0803878233,0.0388847888,0.3201742172,-0.131025508,0.2245138139,0.4103449285,-0.3719108105,-0.2859771848,-0.357524693,-0.3740771413,-0.2960793078,-0.0368870273,0.2843802571,-0.3815568089,-0.026844291,0.0747342706,-0.2222239375,0.3424518108,0.1978043914,-0.0199165922,-0.0125282444,-0.1248900369,0.0557919703,0.1115825772,-0.2184216827,-0.0883419365,0.2907328904,0.0577507801,0.3254128098,-0.0300011262,-0.2500180602,0.2736208737,-0.1425734162,0.2413253635,-0.0663125142,-0.1639661342,0.0518228598,-0.4243109226,-0.0112346355,-0.0395526774,-0.0511120781,0.3793798685,0.2180943489,-0.1268110424,-0.5319942832,0.3141245544,-0.110349223,0.3085636199,0.0816225559,0.065418236,0.066239655,0.1429114491,-0.32364586,-0.4729792476,0.173680529,-0.2797944248,0.0216505453,-0.0442789495,0.026562389,0.0876156986,0.1970206052,-0.2083215564,-0.176524058,0.0325309187,-0.0281284042,-0.3058095276,-0.067377612,-0.2255303413,0.6929596663,0.084308289,0.1074263901,-0.1142612249,0.1818803996,-0.1542342305,-0.4138889015,-0.2585455775,0.2511281669,-0.0218194276,-0.0692520291,-0.1983412951,0.4451107383,0.4843270779,-0.118044965,-0.0518700294,0.1216176674,0.246603176,-0.2090547085,-0.138877511,0.068834506,0.0352383964,-0.1874007434,0.2327999622,0.4481291771,0.0947892815,0.407474488,0.2030845433,0.2327018827,0.3086680174,0.0127824163,-0.0362205207,-0.1380452365,0.2058536708,-0.1581202447,0.1850823909,0.0457534976,-0.1643611789,-0.2564615905,0.1085876077,0.0174395684,-0.1557406783,0.0850858316,0.1128454655,-0.0914609209,0.1244042814,0.0310198106,0.4274091125,-0.0059261983,-0.2235197425,0.1718454361,-0.2711071968,-0.0165314842,0.1195815802,-0.0777548179,0.4340060055,0.3442184329,0.0761829391,-0.0058277696,-0.1620981395,-0.3331285417,0.090616256,0.3732032478,-0.502589941,0.2049745619,-0.2303968966,0.3619090617,0.0869357735,-0.1609289795,-0.2941947877,-0.5459591746,-0.1432554573,0.4888988435,-0.0613883473,0.1217772514,-0.0845048651,0.0663760081,-0.201131165,0.3795134723,-0.1554935575,-0.3128970563,-0.1671391428,-0.1145062894,0.3428279161,-0.2638840377,0.2080451101,0.1412590593,-0.2588335872,0.1877717376,-0.2830425799,0.0676281154,-0.0259474441,0.0585001893,0.1284969151,-0.0208069049,0.0491969138,-0.2054762691,0.1398832798,-0.1417029351,-0.2139025182,0.2980799973,-0.1571695507,0.0964528173,-0.2438025773,-0.3256239891,-0.4506981373,-0.1157023087,-0.0904259905,-0.1948286444,0.3589168787,0.1306937784,0.0114295054,0.0026676615,-0.0306343716,0.0820514485,-0.0601099096,-0.0735768154,-0.1898320317,0.0512790456,-0.1240581051,-0.0503303744,0.0815546066,0.1412719637,0.4111151695,-0.3283545077,0.0388552323,0.0718703344,-0.1109181419,0.1700877249,-0.1913468242,0.340503037,0.4284879565,0.0626717135,-0.0406432599,-0.1832158267,0.1314997375,0.0006427078,0.050948903,0.1277842671,0.3174870908,0.0773281232,0.5648902655,0.4000612199,0.0795966387,0.3894059062,-0.0784598142,0.095120661,-0.1991408914,-0.3932906985,-0.1849010885,-0.2834628224,-0.0068197865,-0.228179574,-0.1353749186,-0.4457137883,-0.3392269909,0.5666782856,-0.3610299826,-0.2107529789,-0.0417289734,-0.5126770735,0.2882983685,-0.1212957203,0.0267894529,-0.0357249416,-0.0081121447,-0.2567384541,0.1977212578,0.1462723762,-0.3395597935,-0.2668709755,-0.2864095569,-0.4406820834,0.2940984368,0.2335691452,0.6908142567,-0.0002039793,-0.0424175039,0.0352362916,-0.1664053649,0.9200374484,-0.6156747937,-0.3640598059,0.3131287694,-0.3987411261,-0.2739586234,-0.1576876044,-0.1989460886,0.5483598709,0.3602884412,0.4070593417,-0.1227389276,-0.300540179,0.219314605,-0.2263223529,-0.1221460551,-0.007816893,-0.3369154036,-0.0060512894,-0.1695673913,0.0396813042,-0.2714727819,0.1052414253,-0.2288514078,-0.1012905389,-0.1274575293,-0.0521500893,0.2379688919,0.0548662581,-0.0112054925,-0.210375756,0.0438952222,0.0179141797,0.1625725776,0.2517501414,0.2416642606,-0.439756602,-0.1313535273,0.1798749119,0.0217344537,0.3080256581,0.2794195116,0.2328889817,-0.1114659458,0.0360400863,0.222242415,-0.300121516,-0.0078934748,0.3932848573,-0.0275971461,-0.559217155,-0.1329834163,-0.0475035273,0.4297206998,-0.1797455251,0.4996367395,-0.4091198146,-0.2604359686,0.3857953846,0.0881675035,0.821480751,-0.3865020275,-0.0374372937,0.0653712675,0.0169688612,0.0669762567,0.0731021017,0.032563407,-0.2790115178,-0.01404138,-0.0699305683,0.030944813,0.2947221398,0.1713699847,-0.1754851043,0.4398418069,-0.0194583144,-0.1161964536,-0.0484600104,0.1572265774,0.4406583607,-0.0374105312,0.0817843527,-0.0184295997,-0.0621318892,-0.1541798115,0.0185531974,0.0313328393,0.2975398898,-0.1252275854,-0.4192383587,-0.0724061728,-0.1181799397,0.1250344217,0.1846571565,0.1083898768,-0.0084211025,0.1368847042,-0.0964865685,0.0566756949,-0.1124495715,-0.1196193323,0.0853132382,0.3296992481,0.174145177,0.0646905527,0.3444856703,0.0899953991,-0.2917966545,0.0637751073,-0.2619931996,-0.2302082181,0.0379634649,0.0626608357,0.063704446,0.1068223566,0.0922640488,-0.0003191101,-0.1378253698,-0.1743955016,0.04270228,-0.0623484328,-0.1261004806,0.5267798305,-0.0455155782,-0.3807288706,0.004184695,0.4495490193,0.2468337715,-0.0429214723,0.1369284689,0.0942235142,-0.032328181,-0.2398495376,-0.2081017643,0.2037245929,-0.1616732031,0.3043036759,0.1201198474,-0.2429559082,0.2137769014,0.0304139517,0.0691135228,0.3609748781,-0.2050937116,-0.3565657735,-0.4723674655,0.2541821301,0.0660018101,0.0953551829,-0.1277286708,0.1256549656,-0.0732245222,0.308934927,-0.2165336311,0.1270089,-0.0945644751,0.217474848,-0.0744947717,-0.012730333,-0.124613516,-0.0959060714,0.0657045916,-0.0103376722,-0.2302361578,-0.0202948209,0.0522265658,0.140048027,0.0725103691,-0.2079423815,-0.057794895,-0.0867337659,-0.1270060092,-0.1857218742,0.1631090194,0.3482712507,-0.196799174,0.1733650714,0.2220966518,0.1730461717,0.1086621732,0.3437299728,-0.21691145,0.0295745693,0.0338103473,0.4157698154,0.1267511249,-0.1924056113,0.0462736785,0.0995301381,-0.0791856274,-0.0047398079,0.3188295662,-0.2237428129,0.0490523502,-0.0068439776,0.4192481935,0.4309894443,-0.1285291761,0.1167162061,0.1111284867,0.0837717205,0.0141894855,-0.1774628311,0.5001866221,0.0905035287,0.0548331179,0.3369073868,0.2816796005,-0.1774310023,0.225333482,0.1305469126,-0.075947687,-0.053749647,0.2161948979,0.1756877303,0.0883331075,0.2401508242,0.2119761258,-0.2828240395,-0.1160479859,0.3406655788,-0.0125602568,0.5239729881,-0.0488605872,0.1294172704,-0.1126944944,-0.6319570541,0.0555985272,0.1380709559,-0.5310005546,0.2147103399,-0.0109335668,0.2652157843,-0.3374674022,-0.4422847629,-0.1920214295,0.4297769666,-0.2591259778,-0.339897573,-0.2281831503,-0.1533527672,-0.0836689249,0.0735744908,-0.1097123325,0.2024549246,0.7951161861,0.0615023971,-0.0631391406,-0.1882469803,-0.3232829571,-0.0134190656,0.3367684186,-0.0098402947,0.430450052,0.0385267176,0.0455684103,-0.03910584,0.1618938446,0.4820715487,0.7032452822,-0.4767341018,-0.0073740524,0.1724205613,-0.0918179452,-0.0997022539,0.4091151953,0.1216854677,0.3184451461,0.2829155922,-0.0671014488,0.0288567878,-0.1637240648,0.3179637194,0.3598909676,0.0542587303,0.3573621511,0.0805680752,-0.0921949893,0.2188027352,0.0852720961,-0.3468811512,-0.2500665188,0.5011442304,-0.4767701924,0.2512305379,0.2434682101,0.0449358113,0.0634783804,0.4826915562,0.2787361443,0.1293309182,-0.2892729044,-0.1975532174,-0.4078818858,-0.0098155113,-0.1615679264,0.2877382338,0.0176803097,-0.0174050499,-0.0764769763,0.0200282652,-0.0329721719,-0.2333677709,0.386323005,0.3298452497,-0.5628172159,0.1965079308,0.0664201453,-0.0144150192,0.0069541987,-0.30533728,0.3622879982,0.0587223805,-0.1156783327,0.1099861935,0.170832023,0.0688161775,0.2018173933,0.0038733399,0.2783461511,0.5984800458,-0.031748727,0.0064989175,-0.2279817164,0.0421438776,-0.067610696,-0.0266452953,0.0801315904,0.3813818097,-0.2080049217,-0.0452977717,0.0238661915,0.1681547463,0.0005169157,0.2088668048,-0.4267875254,0.1841862947,0.2177486271,0.0126832118,-0.1509294063,0.3676734269,0.1753032953,0.323276341,-0.2848980427,-0.1847170442,0.3160332441,0.0194945056,-0.1091337875,-0.5905250311,0.2252316624,-0.2117346078,0.11794056,-0.5522873998,-0.3172584176,0.1953945011,-0.2018169165,-0.5665146708,0.2724033296,-0.1109253466,-0.1593625546,-0.0590723716,-0.2073619813,0.1310657412,-0.1207876131,0.1831817478,-0.0755577311]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"#659 should fix the `KeyError` issue. It was due to the formatting not getting updated the right way","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":18,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\n#659 should fix the `KeyError` issue. It was due to the formatting not getting updated the right way","embeddings":[-0.3512131572,-0.1855427474,-0.0195908081,0.21030581,0.1381857842,-0.1670842022,0.279554069,0.3684962094,0.0815791711,0.1086883396,0.0889100879,0.3803418577,-0.4003902972,0.400125742,-0.3610644341,-0.0263170786,0.1120805815,-0.0113873798,-0.3007359803,0.2906777561,-0.2718750834,0.2635469139,-0.3654167056,0.1494772881,-0.5365363955,0.0468143262,-0.1020090133,0.2186754048,-0.0090193618,-0.6350335479,0.3107408583,0.2939074039,-0.0312150251,0.4254748225,-0.0001246411,0.1012204811,0.3174884915,-0.1645227522,-0.0690915287,-0.2924408019,-0.222808972,-0.2399374247,0.2814206481,0.1030045003,0.1462618262,-0.0645411089,-0.168445006,-0.1902516633,0.2530600131,0.3587377369,0.1109459624,0.198188141,0.3393394649,0.1626249552,-0.0395517275,0.1583103985,-0.0288628805,0.020966731,0.302185595,-0.2966939509,-0.0835430175,0.3755783737,-0.2106487155,-0.2462289929,-0.0806416124,-0.2204282433,0.6230471134,-0.5610292554,0.1792760789,-0.0036662992,-0.1783303022,0.0002152335,-0.4601679146,-0.239700675,-0.2292336524,-0.4920301735,0.2561888695,-0.009299662,-0.138959378,0.1675009578,-0.4596162438,-0.2790907919,-0.0793253854,-0.0925054401,-0.045141831,0.5593509674,0.1581875086,0.2434732765,0.0868087932,0.1045118049,0.1365901828,-0.0745362714,-0.1486057192,0.1676543504,-0.3396682441,0.0580588914,0.0380162559,-0.053946726,-0.2356516272,-0.0823000744,-0.195446074,0.1293760091,-0.0923483148,-0.0000644441,0.4277984798,-0.0651589036,0.3168876469,0.2542020977,0.2367102802,-0.2161653489,0.0550286956,0.0133819832,0.3371013701,-0.1049287543,-0.0162288528,0.3913570642,0.1192442626,-0.1721612811,-0.1057623848,0.1788408011,-0.0347326808,-0.2633396685,0.0211248323,0.1464975923,0.0519356765,0.7958670855,0.1379258931,0.1675080657,-0.303982079,-0.0909318626,-0.0441626161,-0.0180010125,-0.4197016954,0.0411583148,0.2246312052,-0.0470206589,-0.0232576262,0.0513088591,-0.3942948282,-0.2987011373,0.0350022763,-0.2101656497,0.1784780771,0.6543557048,-0.131706804,0.0979316607,-0.0364758149,-0.1969342381,0.0777080879,0.2401845455,-0.5455412269,-0.1287513226,-0.2255560607,0.0737909228,0.0727319196,0.3116742373,-0.1367508918,0.2338903546,0.438331902,-0.3816615045,-0.2752705216,-0.3559553027,-0.3625250161,-0.3146935701,-0.0502027422,0.3140787482,-0.4769178033,-0.0701623186,0.1530882567,-0.1712597907,0.3403357267,0.1905321032,0.023639353,0.0471595265,-0.1083849296,0.1005764231,0.1277480125,-0.2211735994,-0.1791697741,0.3235459626,0.013238363,0.365031004,0.0487262979,-0.2932927012,0.3140540421,-0.0850169659,0.282943368,-0.1017530113,-0.1259889454,0.0704030991,-0.4282627702,0.015011901,-0.0669372231,-0.167552039,0.357986331,0.1379951835,-0.2026560605,-0.475399226,0.33156389,-0.1146088541,0.3142383695,0.0391068459,0.0674777627,0.0599834397,0.1719403118,-0.2382501364,-0.4738487899,0.1375685781,-0.2226947695,0.0047201421,-0.1448158026,0.0010650369,0.0444621667,0.1925702542,-0.2922160327,-0.2492041886,0.041765593,0.0107115628,-0.3038726449,-0.0823195204,-0.1843638122,0.5929731131,0.2195536196,0.1409046352,-0.0882686749,0.1312265992,-0.2019506991,-0.4209198654,-0.2407811284,0.2778545618,-0.0094550001,-0.1175537631,-0.2070064843,0.5246531367,0.5095104575,-0.1137293354,-0.1429157555,0.1915371865,0.268491447,-0.0770466253,-0.1743265837,0.0702149495,0.0227089003,-0.1169540212,0.0810186639,0.5475662947,0.0962426066,0.4146905541,0.1494486034,0.2452741563,0.2949585617,0.0570024736,-0.0602537394,-0.2418469638,0.1514021009,-0.2580640316,0.1053335667,0.001240272,-0.1381929964,-0.1894830465,0.1466972381,0.105109401,-0.109724246,0.017060576,0.0435087867,-0.1117243171,0.0992233604,0.0218311194,0.3506157994,0.0209108498,-0.2784491479,0.1740038544,-0.1862997711,-0.0493131541,0.1387522221,-0.0766849741,0.3364537656,0.3832502365,0.056158632,0.010623144,-0.1192159429,-0.2493800223,0.0324690156,0.3667414188,-0.5073884726,0.2152490467,-0.2861204743,0.2549757063,0.0421197563,-0.109275654,-0.1926590949,-0.5729332566,-0.1137760282,0.4209921658,-0.0548844673,0.1483545303,-0.0123480288,0.0344811529,-0.1731377095,0.3812194765,-0.1331987977,-0.3283342719,-0.1470113397,-0.1047791094,0.3488055766,-0.3240474164,0.2100392729,0.0738013089,-0.1691029519,0.236829266,-0.3342378139,0.1082952544,-0.0472081602,-0.0064860615,0.1615286767,0.0518070683,0.0494501293,-0.2052884996,0.1594745666,-0.1578568518,-0.2601367831,0.3101465702,-0.149607718,0.1178078204,-0.2731796205,-0.3018988371,-0.3529520035,-0.0495851599,-0.0805207863,-0.2224941552,0.3450743854,0.1496513486,0.0659467652,-0.0863049626,-0.1298864484,0.0984945223,-0.0306528844,0.0026161505,-0.1935723275,0.0507384501,-0.1470983773,-0.0774887279,0.0760762319,0.0738083199,0.3409603536,-0.2849748135,0.0146599235,0.095136486,-0.125904575,0.2607448399,-0.1134286672,0.3486637771,0.4518114328,0.0555604286,-0.0486735776,-0.1659440994,0.0462608039,-0.0057883211,-0.0026092781,0.1092260778,0.3223183155,0.13276802,0.6104079485,0.3465014696,0.1186999306,0.3046036065,-0.1118744835,0.1353441477,-0.2403228581,-0.3785359859,-0.2232074589,-0.2772703767,0.0575733781,-0.1354533881,-0.1057833061,-0.390625596,-0.2962539792,0.4868232608,-0.4268265069,-0.2603372931,0.0627299026,-0.5000072122,0.281824708,-0.0854153112,0.143267706,-0.041168265,-0.0294576567,-0.2238727212,0.1939019263,0.1314987987,-0.3822313845,-0.2747116387,-0.3799426258,-0.4724220932,0.3566471636,0.237246111,0.7411826849,0.0431601889,-0.0587891489,0.0283008721,-0.1665703505,0.8739997149,-0.6421237588,-0.4028041959,0.2552218437,-0.438818574,-0.2836793363,-0.2003698796,-0.2084669173,0.624573946,0.3335076272,0.5355913043,-0.1612277031,-0.3011726737,0.1604255289,-0.1607849896,-0.167784214,0.0375813991,-0.3154361546,0.0333147235,-0.2019890845,0.0484579876,-0.3043012917,0.1444120258,-0.133154422,0.0173656419,-0.0812017769,-0.0779063255,0.1874220073,0.088571474,0.0324916914,-0.106460236,-0.0421276763,0.0955709293,0.1971702278,0.231853351,0.2196662277,-0.4627836943,-0.0665591657,0.1394232064,-0.1067316681,0.3257451057,0.3188472986,0.2260964364,-0.0660924166,0.0870821849,0.1652499586,-0.3262059391,-0.0555948801,0.3966928124,-0.031492833,-0.5680369139,-0.2148531526,-0.1295146048,0.3982083201,-0.2285052836,0.4416521788,-0.3164745271,-0.2612662017,0.4667618275,0.0664021075,0.8066737056,-0.3608477414,0.0161097646,0.0260820016,-0.0325377546,0.1356405169,0.0996824801,0.0458455421,-0.365354389,-0.0213855281,-0.068143338,0.0656674951,0.2799925506,0.1120454967,-0.2234569788,0.4462039769,-0.0944939032,-0.0914893076,-0.0715271458,0.1839328706,0.4588896632,-0.0594944917,0.0677264556,-0.0470554866,-0.033569254,-0.141615063,0.0173166152,0.030314127,0.2864644825,-0.2391974628,-0.458938688,-0.1258006245,-0.0731941909,0.0951595604,0.1623540968,0.0272925608,-0.031251289,0.1535902619,-0.0549028032,0.1623506844,-0.0921655893,-0.1436306685,0.0721743405,0.3353672922,0.1709644049,0.0916816518,0.3252387643,0.1279744804,-0.2073791176,0.0669183657,-0.2403381765,-0.1906894147,0.0515942834,0.0379658863,0.0266993586,0.0648267567,0.0511722676,-0.0490261018,-0.099320516,-0.2013467401,0.0393683165,-0.0495411158,-0.0926299989,0.531072557,0.0182950124,-0.3574020267,0.0256605614,0.4819164574,0.2871125042,-0.0772084594,0.2314321697,0.1426067054,-0.0611013807,-0.2567400634,-0.156475082,0.2738894522,-0.1716051996,0.2461798489,0.0910365731,-0.1785513014,0.1953545958,0.1071571931,0.118932873,0.3926021457,-0.2192414105,-0.3378393054,-0.4110481143,0.2427576929,0.1089817658,0.0621586554,-0.1027741432,0.1274107546,-0.113509059,0.3102333248,-0.2243909538,0.0437067077,-0.00383811,0.2115470469,-0.0914864391,0.0675676391,-0.1171541363,-0.140081495,0.0643864051,0.0095132347,-0.1756792217,-0.0247321166,0.0113600716,0.1218328774,0.0662016571,-0.1837543249,-0.0681138709,0.0006087265,-0.0888127089,-0.2294077575,0.1706517339,0.323995769,-0.1849398017,0.2520669103,0.1697707772,0.1847819537,0.0892375559,0.3003495932,-0.2528879642,0.0888922587,0.0715160742,0.4085338116,0.0848133788,-0.1983519942,0.0351339951,0.1764115393,-0.1393329501,0.0583752543,0.2958631217,-0.2519658804,0.0858205557,0.0182251055,0.4392744601,0.4068154991,-0.1030400023,0.1577364057,0.0806601867,0.0935269147,0.0447812304,-0.2281359434,0.5149974823,0.0612499416,0.0703290403,0.3585091531,0.3325626552,-0.1416944861,0.1496500373,0.1260901392,-0.1130566746,0.008072515,0.1593117118,0.3051334321,0.1120931134,0.2470810264,0.2925590277,-0.2948818803,-0.1773032099,0.2537924349,0.0569481142,0.4753111303,0.0924650282,0.1063256338,-0.162421152,-0.6658914685,0.1066897959,0.0673350543,-0.456555903,0.2207169384,0.0476552136,0.2264577448,-0.3147785068,-0.4356147349,-0.132876873,0.3830533326,-0.2740252614,-0.3404760957,-0.2711844742,-0.1706426442,-0.0961931869,0.0628362745,-0.1690142751,0.2519634366,0.746922493,-0.0266137347,-0.057759285,-0.1461751461,-0.322909534,-0.0722546428,0.413166225,-0.089219436,0.3782847226,0.057088241,0.1401819587,0.0034944098,0.1684929132,0.4758540988,0.6625869274,-0.4762479663,-0.082555674,0.0828101039,-0.1063217595,-0.0610712357,0.4503815472,0.0963009298,0.2951018512,0.3415160179,-0.0530048683,0.0091973245,-0.1055935621,0.3893709779,0.3519671857,0.157383889,0.2242008448,0.0683892369,-0.1439942569,0.2735783756,0.0962108746,-0.3139980435,-0.3135891855,0.4868752956,-0.378231287,0.237613216,0.18786937,0.0577968657,0.1757684499,0.5463715792,0.3670721352,0.1067019925,-0.259142518,-0.1662559956,-0.3945916593,-0.0790592059,-0.1534976512,0.2001519501,-0.0425510705,-0.0055501298,0.0010906135,0.0817848742,-0.0310584549,-0.2819316089,0.2874446511,0.3652774096,-0.5556682348,0.2423667163,0.1740243435,-0.0015560285,-0.0395298712,-0.2589582801,0.3855098784,0.0197239649,-0.0677194148,0.073588185,0.1376985312,0.0231757239,0.1443931758,-0.0157724917,0.2998056412,0.6039760709,-0.0495157875,-0.0536487065,-0.2163181901,0.0924264193,-0.1226849407,0.0438340902,0.0929373875,0.5019168258,-0.1521149725,-0.0962430015,0.0546564721,0.2140313387,-0.0115536116,0.2152949572,-0.4327308238,0.2691527605,0.2196222097,-0.0632077903,-0.1603733152,0.4091362655,0.1511186361,0.2834834158,-0.2557684183,-0.1834850907,0.2162734568,-0.0245581344,-0.1608247608,-0.5821929574,0.1678624004,-0.1972016543,0.1791553795,-0.491659075,-0.2939208746,0.1788068712,-0.175419271,-0.5868240595,0.2050280869,-0.1429528743,-0.1216801405,-0.0698917061,-0.1655345857,0.096296601,-0.134757489,0.0342641808,-0.0679894462]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Also maybe @n1t0 knows why setting `TOKENIZERS_PARALLELISM=true` creates deadlock issues when calling `map` with multiprocessing ?","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":16,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nAlso maybe @n1t0 knows why setting `TOKENIZERS_PARALLELISM=true` creates deadlock issues when calling `map` with multiprocessing ?","embeddings":[-0.4357655942,-0.1091967374,-0.0417286009,0.2121776342,0.0695560127,-0.126509428,0.2546488643,0.3328145444,0.1938226819,0.1219439209,0.1160629243,0.4567498267,-0.4093019664,0.2884688675,-0.3752997816,0.0169866215,0.0992285386,-0.094248727,-0.3434426188,0.2482099235,-0.2782725096,0.2710521221,-0.3459435105,0.1593248844,-0.5353220701,0.0872885659,-0.0039736438,0.2528886497,-0.0188716929,-0.573179543,0.2949807346,0.2831947207,0.0004403761,0.4602254629,-0.0001217121,0.1132882759,0.3699000776,-0.0942805558,-0.0318265818,-0.3821792305,-0.2222522348,-0.2362090945,0.2945428491,0.103973031,0.2161948234,0.0128660435,-0.0772559792,-0.2985168695,0.1986693889,0.372841388,0.1237755194,0.2111014128,0.297875762,0.1220005155,-0.0423817784,0.1499406397,-0.0497977175,0.0485664085,0.3448382914,-0.3739497066,-0.1388048977,0.2796147764,-0.2362969965,-0.1811296493,-0.1206905842,-0.2400204986,0.6395820379,-0.5958400369,0.1814056486,0.0657546222,-0.2609025538,0.0006286997,-0.4477955699,-0.1799360365,-0.2648523152,-0.4586621821,0.2208007276,0.0218163226,-0.1726396382,0.1386403441,-0.5251956582,-0.2143575251,0.0352452695,-0.0544744171,-0.0366728529,0.5323663354,0.2305611372,0.3113503158,0.1949433684,0.1283584237,0.113835834,-0.0167996679,-0.0804867595,0.1571735591,-0.4042960703,0.1016928628,0.1372587383,-0.1121087521,-0.2406967133,-0.0560522564,-0.1629004329,0.0969965905,-0.0776492879,0.0668469444,0.4337721467,-0.1338667125,0.1751342565,0.2501194179,0.2519487739,-0.2092773765,0.0151370941,0.0317399129,0.3243446052,-0.1119964868,-0.0390037522,0.3069898784,0.0965447277,-0.1498585343,-0.1592805684,0.0878242329,-0.107860513,-0.2057652771,0.101189442,0.1439217776,0.0724177882,0.7728213072,0.1620066464,0.1372262686,-0.3340587616,-0.0898869187,-0.0551431254,-0.0666498616,-0.4002197087,0.1255273968,0.2563378513,0.0599527881,0.0080618942,0.0419398136,-0.3320827484,-0.21691598,0.1727746427,-0.2766205072,0.0906436667,0.631087482,-0.0971724987,0.0243760124,-0.0824455693,-0.0286204666,0.0511687733,0.2955014408,-0.5336505175,-0.1557149738,-0.1656827778,0.1078610867,0.0666458607,0.3060179949,-0.1915570945,0.254709214,0.3675672412,-0.3739640415,-0.2680085301,-0.3238871694,-0.4069840312,-0.3038038909,-0.0046002609,0.2971160412,-0.3287827969,-0.0234114919,0.0687612444,-0.2095829844,0.2784111798,0.2272229344,-0.0418685414,0.0260143224,-0.1166272312,0.1641643196,0.0441024639,-0.2357017696,-0.0901374593,0.2893678546,-0.008890464,0.329949826,-0.0470988005,-0.2168494463,0.3346088231,-0.1311822236,0.2851608992,-0.0148701631,-0.1969221085,0.1048761457,-0.4357783198,-0.0461987182,-0.0751762018,-0.082395494,0.3545601666,0.1582431942,-0.1317896545,-0.4724713862,0.2870244682,-0.071243085,0.3042713404,0.0659307986,0.0621412434,0.0573612377,0.1319958866,-0.2830133736,-0.5129542947,0.199440226,-0.2078189552,-0.0241884403,-0.0505439565,0.015270507,0.1671999097,0.1793141961,-0.1974955499,-0.1488039494,0.0632102564,-0.0676070079,-0.2822651863,-0.0553281121,-0.1842774153,0.6492671371,0.1441810876,0.0867221281,-0.0818904936,0.221170336,-0.1520411223,-0.4408725798,-0.2989481688,0.2226276249,-0.045403786,-0.0716135427,-0.2133348137,0.4630851448,0.4911286831,-0.0391312689,-0.037637122,0.1667612046,0.2644596696,-0.1828808486,-0.1570649445,0.1237259135,0.0492204279,-0.1837459803,0.2333987504,0.4180732071,0.0836041048,0.3880741,0.2444252819,0.238742426,0.2779892385,0.0294546597,-0.0278633907,-0.1282213926,0.2124024332,-0.1187339947,0.1564511061,0.0764603615,-0.0925443396,-0.2441881597,0.1627016068,0.0704164654,-0.1369761974,0.0474271998,0.0893530995,-0.1253041923,0.1578324735,0.0026259758,0.4480130076,0.0147289056,-0.1836158633,0.1479921937,-0.2042004019,-0.0055106869,0.088725321,-0.0780102462,0.4078099132,0.2825418711,0.0658046678,-0.0059386119,-0.152056396,-0.3619617522,0.1194626912,0.3093265593,-0.4322062135,0.1622132808,-0.2236522287,0.3789779544,0.140934065,-0.1527124643,-0.2679890394,-0.5417202115,-0.1163036153,0.5250684619,-0.1006801128,0.132585749,-0.0776599199,0.0356060602,-0.2600830793,0.3981772959,-0.1304369122,-0.3576631248,-0.1955149621,-0.0964715779,0.3104203343,-0.2298806906,0.2243646383,0.1590366215,-0.2553647459,0.1742384136,-0.246335566,0.0499618873,-0.0595836341,0.0070893858,0.1035203412,-0.0094830273,0.007403533,-0.1684115678,0.1527904123,-0.1796357632,-0.2490491569,0.2781174183,-0.1366073042,0.0774834082,-0.2758553922,-0.3428668976,-0.4242911637,-0.1626048833,-0.0231468678,-0.2679255307,0.3306127191,0.0996191949,0.0179547556,0.0415381156,-0.0787686855,0.0687170997,-0.0472171456,-0.062253952,-0.2065602243,0.0454690121,-0.1702650785,-0.0546774678,0.1007058099,0.0798726752,0.4131040275,-0.2472681254,0.076609537,0.0603867844,-0.0629385114,0.2250357568,-0.1975023001,0.3669516146,0.3852897882,0.0340522006,-0.0286275465,-0.1490013599,0.1410466731,0.0047071753,-0.0262567904,0.1507259905,0.309121877,0.1691416353,0.6082265973,0.3968889117,0.0640839487,0.3936433792,-0.0730804801,0.0929013193,-0.2089506835,-0.3948884904,-0.2156747133,-0.2659694552,-0.0087280031,-0.2322266847,-0.1142553985,-0.4359849691,-0.3072878122,0.5583676696,-0.361368984,-0.2352523208,-0.0361661762,-0.4950186014,0.2937211394,-0.1269934773,0.0426193736,-0.0449110419,-0.0112302788,-0.2169745564,0.2020891309,0.1722293645,-0.3479177356,-0.2839290798,-0.250210166,-0.4786651134,0.2983989716,0.2423682958,0.6819908023,-0.0241148807,-0.0504561439,0.0272309706,-0.1523956209,0.8559732437,-0.5813956857,-0.383443743,0.3133418858,-0.4112911522,-0.2568888664,-0.1569911391,-0.2074812651,0.557199955,0.4230445921,0.4213473201,-0.1724165827,-0.2897487581,0.122319296,-0.2117930502,-0.0765742213,-0.0049102004,-0.3257120848,0.0413429476,-0.1830744296,0.0275970884,-0.2412206233,0.092107892,-0.1844061166,-0.123599045,-0.0644754171,-0.0527578406,0.1900015026,0.0667347834,-0.0137832435,-0.2407394201,0.0838182196,0.0814906508,0.1122456491,0.2761133313,0.2073537707,-0.4210828543,-0.121202074,0.1534609348,0.0207233671,0.2747559845,0.2481248528,0.2212694287,-0.0763034597,0.0679313615,0.1537781358,-0.3496284485,-0.0344160013,0.4136733115,-0.0135288239,-0.4599664509,-0.1147129163,-0.0684803128,0.4527403414,-0.1746952981,0.5284518003,-0.4588576853,-0.2545072138,0.3466490507,0.0760228112,0.7190542221,-0.4016178846,-0.0562179498,0.0151910465,-0.0228415262,0.0593855605,0.0745066106,0.0395906419,-0.2484300435,0.0306452941,-0.0443410426,0.0463400222,0.3351760805,0.1809006333,-0.1599562615,0.3927199543,0.0175097156,-0.1283548474,-0.0871314183,0.2037595958,0.4618608356,-0.0916174725,0.1236859038,0.0090226475,-0.0373493209,-0.1418357939,0.029247595,0.0303373076,0.291194737,-0.1087180078,-0.4325701594,-0.0954679251,-0.1311917752,0.1541177779,0.1544950604,0.1990784258,-0.0326027721,0.1023308188,-0.1661847979,0.0416756794,-0.1075472459,-0.1571244448,0.0811328292,0.3711447716,0.1824040711,0.0543471128,0.3112122715,0.0866436809,-0.2523908019,0.038198445,-0.2390208691,-0.2315143049,0.0284906179,0.054431282,0.0770219862,0.1647556871,0.0821133181,-0.0087165451,-0.1461447179,-0.144038558,0.0647432134,-0.0179148894,-0.1472049206,0.4881018698,-0.0362947173,-0.3848839402,0.0069667697,0.498196125,0.2140212953,-0.0611994602,0.114951022,0.1187440977,-0.0506509021,-0.2443604022,-0.209525913,0.2826492786,-0.1479827613,0.2815427482,0.0740328208,-0.2316265255,0.2187848538,0.032504268,0.0637230575,0.3333226144,-0.2180020809,-0.3376714885,-0.4480600655,0.3275709152,0.0213086326,0.0697895661,-0.0914685279,0.0742159933,-0.0702163652,0.2902164161,-0.2416669428,0.1283608675,-0.0914117172,0.2105484307,-0.0862649158,-0.0200424325,-0.119935222,-0.0852740705,0.0836801603,-0.0180115327,-0.2818257213,-0.0459659956,0.0401631892,0.1124787629,0.0590033196,-0.2106230557,-0.027524801,-0.1331017315,-0.1302309334,-0.2069941163,0.184937343,0.3496144712,-0.2491645515,0.1554153115,0.2163188905,0.1715707332,0.0557660721,0.3391492665,-0.2267336398,-0.0104442909,-0.0211594682,0.4098553956,0.1665135324,-0.1714047641,-0.0185560826,0.0839513168,-0.0475137345,0.027463289,0.3284184933,-0.1623191386,0.08146099,0.0122954398,0.4144154191,0.3997435272,-0.0963426977,0.0735945776,0.104803212,0.1085935012,0.0461436361,-0.1891902834,0.5273979306,0.0765309483,0.1033681035,0.2709961832,0.3167985082,-0.1239268184,0.1547921747,0.1226018965,-0.1012954935,-0.0107697966,0.217214644,0.1855545342,0.0957496911,0.2719643116,0.1953198165,-0.3276026845,-0.1445708722,0.3017556369,-0.0117888264,0.6197060943,-0.0176018346,0.1720775217,-0.0681530833,-0.6532103419,0.1134988442,0.0888802707,-0.5180866122,0.2230882943,-0.0083675021,0.2138800621,-0.3701971471,-0.4355387688,-0.1738755554,0.397020936,-0.3046890199,-0.3131806552,-0.162599355,-0.1394881159,-0.0880024359,0.059706632,-0.1238612831,0.2549751997,0.8632911444,0.0616490729,-0.0298706759,-0.1592311114,-0.3787843883,-0.0492622666,0.3128960729,-0.0241103861,0.3672715425,-0.0567246526,0.0261306707,-0.0641734749,0.2130477875,0.4203889966,0.7252755165,-0.4599173665,-0.0259217937,0.1545582265,-0.0643241629,-0.1112351641,0.447611779,0.101650089,0.2901473343,0.2144193947,-0.0477454327,0.0076082312,-0.2236599773,0.2952461541,0.3430363238,0.067407243,0.3543583453,0.0387332253,-0.1062361971,0.2686705291,0.0939206257,-0.2781712413,-0.289506644,0.4929398894,-0.4646965563,0.1978830248,0.166608274,0.0611947812,0.0654450655,0.4523872733,0.2915227115,0.0668628886,-0.3113783002,-0.15977934,-0.4288888276,0.0130603965,-0.2220526189,0.3328104913,0.0194288474,-0.0061470252,-0.0825499073,0.0802581385,-0.0234007221,-0.2516870797,0.3622833192,0.3247748911,-0.5693955421,0.2066017687,0.0963991806,-0.0811117366,0.0033281583,-0.2955177128,0.356775254,0.0581232235,-0.0845478773,0.05988428,0.1237470433,0.0907379165,0.2181303352,-0.008436569,0.2849671543,0.5745168924,-0.0258132014,-0.0119340569,-0.2263609916,0.0990829319,-0.0619273111,-0.0422164425,0.0261936877,0.4040822387,-0.2270916253,-0.0838269219,-0.0307073109,0.188560605,-0.0535225794,0.166021511,-0.4402961731,0.1806248873,0.1818243563,0.0434451737,-0.1316507459,0.3896892965,0.1911218017,0.286524117,-0.2992208302,-0.1361809671,0.2220961154,-0.0719347447,-0.0803383142,-0.6504668593,0.208820492,-0.2121038139,0.1164390221,-0.5232219696,-0.3167284131,0.1987286806,-0.1904963553,-0.5479640961,0.2404860407,-0.1207567304,-0.1622239202,-0.0673251078,-0.1826601923,0.1011765823,-0.105663307,0.2155494839,-0.0217618681]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"@lhoestq \r\n\r\nThanks for taking a look. I pulled the master but I still see the key error.\r\n\r\n```\r\nTo disable this warning, you can either:\r\n - Avoid using `tokenizers` before the fork if possible\r\n - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)\r\n#0: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 21.56ba\/s]\r\n#1: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 17.71ba\/s]\r\n#2: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 20.45ba\/s]\r\n#3: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 26.05ba\/s]\r\n#4: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 26.83ba\/s]\r\n#5: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 27.00ba\/s]\r\n#6: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 27.40ba\/s]\r\n#7: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 25.91ba\/s]\r\n#8: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 22.46ba\/s]\r\n#9: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 20.15ba\/s]\r\n#10: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 26.81ba\/s]\r\n#11: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 27.45ba\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 322\/322 [00:00<00:00, 1462.85ex\/s]\r\nTraceback (most recent call last): | 0\/1 [00:00<?, ?ba\/s]\r\n File \"text2struct\/run_model.py\", line 372, in <module>\r\n main()\r\n File \"text2struct\/run_model.py\", line 368, in main | 0\/1 [00:00<?, ?ba\/s]\r\n run_model(auto_envvar_prefix=\"GFB_CIES\") # pragma: no cover\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs) | 0\/1 [00:00<?, ?ba\/s]\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 1236, in invoke\r\n return Command.invoke(self, ctx)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/decorators.py\", line 21, in new_func\r\n return f(get_current_context(), *args, **kwargs)\r\n File \"text2struct\/run_model.py\", line 136, in run_model\r\n ctx.invoke(ctx.command.commands[config_dict[\"mode\"]])\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/decorators.py\", line 21, in new_func\r\n return f(get_current_context(), *args, **kwargs)\r\n File \"text2struct\/run_model.py\", line 187, in train\r\n run_train_model(_parse_subcommand(ctx))\r\n File \"text2struct\/run_model.py\", line 241, in run_train_model\r\n checkpoint_steps=config.train.checkpoint_steps,\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/train.py\", line 153, in alternate_training\r\n max_len=config.model.dim.max_len,\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 466, in load_prodigy_tf_datasets\r\n folder, file_patterns, vocabulary, cache_dir=cache_dir, test_pct=test_pct\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 447, in load_prodigy_arrow_datasets\r\n fn_kwargs=dict(tokenizer=vocabulary.tokenizer, vocabulary=vocabulary),\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1224, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1195, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 348, in relation_ds_factory\r\n ner_labels = example[\"ner_labels\"]\r\nKeyError: 'ner_labels'\r\n\r\n```","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":299,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\n@lhoestq \r\n\r\nThanks for taking a look. I pulled the master but I still see the key error.\r\n\r\n```\r\nTo disable this warning, you can either:\r\n - Avoid using `tokenizers` before the fork if possible\r\n - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)\r\n#0: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 21.56ba\/s]\r\n#1: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 17.71ba\/s]\r\n#2: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 20.45ba\/s]\r\n#3: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 26.05ba\/s]\r\n#4: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 26.83ba\/s]\r\n#5: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 27.00ba\/s]\r\n#6: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 27.40ba\/s]\r\n#7: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 25.91ba\/s]\r\n#8: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 22.46ba\/s]\r\n#9: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 20.15ba\/s]\r\n#10: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 26.81ba\/s]\r\n#11: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 27.45ba\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 322\/322 [00:00<00:00, 1462.85ex\/s]\r\nTraceback (most recent call last): | 0\/1 [00:00<?, ?ba\/s]\r\n File \"text2struct\/run_model.py\", line 372, in <module>\r\n main()\r\n File \"text2struct\/run_model.py\", line 368, in main | 0\/1 [00:00<?, ?ba\/s]\r\n run_model(auto_envvar_prefix=\"GFB_CIES\") # pragma: no cover\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 829, in __call__\r\n return self.main(*args, **kwargs) | 0\/1 [00:00<?, ?ba\/s]\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 782, in main\r\n rv = self.invoke(ctx)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 1236, in invoke\r\n return Command.invoke(self, ctx)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 1066, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/decorators.py\", line 21, in new_func\r\n return f(get_current_context(), *args, **kwargs)\r\n File \"text2struct\/run_model.py\", line 136, in run_model\r\n ctx.invoke(ctx.command.commands[config_dict[\"mode\"]])\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/core.py\", line 610, in invoke\r\n return callback(*args, **kwargs)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/click\/decorators.py\", line 21, in new_func\r\n return f(get_current_context(), *args, **kwargs)\r\n File \"text2struct\/run_model.py\", line 187, in train\r\n run_train_model(_parse_subcommand(ctx))\r\n File \"text2struct\/run_model.py\", line 241, in run_train_model\r\n checkpoint_steps=config.train.checkpoint_steps,\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/train.py\", line 153, in alternate_training\r\n max_len=config.model.dim.max_len,\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 466, in load_prodigy_tf_datasets\r\n folder, file_patterns, vocabulary, cache_dir=cache_dir, test_pct=test_pct\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 447, in load_prodigy_arrow_datasets\r\n fn_kwargs=dict(tokenizer=vocabulary.tokenizer, vocabulary=vocabulary),\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1224, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1195, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 348, in relation_ds_factory\r\n ner_labels = example[\"ner_labels\"]\r\nKeyError: 'ner_labels'\r\n\r\n```","embeddings":[-0.4438695014,-0.1675047278,-0.040071819,0.1716096848,0.0889460295,-0.1188245341,0.207846269,0.3421575725,0.1180476472,0.157603845,0.1306535751,0.4856553674,-0.4362103641,0.2786436677,-0.3663627207,0.0951046795,0.111420311,-0.0841650814,-0.2514494061,0.2458094209,-0.2524542809,0.293427676,-0.2909142971,0.1944313943,-0.5435567498,0.0569215901,-0.0106121879,0.3160980046,0.005523209,-0.6119182706,0.2811330557,0.30997172,-0.0368735269,0.3930642903,-0.0001247698,0.1379997134,0.3410488963,-0.0913274884,-0.0481128283,-0.3334736228,-0.1873978972,-0.2403929383,0.267213732,0.0872503072,0.1997458488,-0.0002410833,-0.0975959301,-0.2515730858,0.2583797872,0.3503323197,0.0834934041,0.2617163062,0.2949203849,0.1206849888,-0.0550052226,0.1494259536,-0.0134723131,0.0250229966,0.3076893687,-0.3135743141,-0.1309461594,0.2591975331,-0.2412051111,-0.2611987889,-0.1408810616,-0.2327822149,0.6852400899,-0.6341823936,0.2206975073,0.0248721112,-0.282739073,-0.0443475544,-0.4056611061,-0.2426628619,-0.2770734131,-0.4417158663,0.2661078572,-0.021212779,-0.2025263309,0.1933264285,-0.4429095984,-0.266000241,0.0245455261,-0.1426336616,-0.049413491,0.570002079,0.2672023475,0.2987840176,0.2195273638,0.1560344845,0.1490322053,-0.020523157,-0.1217988506,0.1732535809,-0.421153605,0.0671509504,0.085211575,-0.1909997761,-0.2070730031,0.0080284579,-0.2620820999,0.0897589177,-0.1146723181,0.0613033324,0.3845333755,-0.1019945443,0.1810257286,0.3499873877,0.2183627039,-0.2259050012,0.0349917933,0.0682251081,0.2776801586,-0.0831107944,-0.057865832,0.3552602232,0.1289922595,-0.1625501066,-0.136644721,0.185089916,-0.050749585,-0.1519199163,0.0860182345,0.2178291678,0.0702168196,0.7802461386,0.1076600105,0.1804842949,-0.3048225641,-0.0502079763,-0.0702662542,-0.0836341903,-0.3501508832,0.0839562565,0.2303638756,0.0645156056,0.0073764003,0.0501154736,-0.3548419774,-0.2700493038,0.0857494026,-0.2150368989,0.0925924852,0.7242619991,-0.1044124588,0.0710303262,-0.0193017423,-0.1493759304,0.0569707677,0.3055911064,-0.5235205889,-0.2392239273,-0.1953841895,0.0776035339,-0.0194743387,0.2954956591,-0.1016284972,0.2702437937,0.4557403326,-0.3296603858,-0.268432796,-0.32086429,-0.365288198,-0.3097073436,0.0002421204,0.2654046714,-0.4719291627,-0.0701095015,0.0881139189,-0.2383557707,0.3207790554,0.2269821465,-0.0228191037,-0.0254130438,-0.1377120614,0.1407947391,0.0430819169,-0.2105707228,-0.1295746416,0.2465180904,-0.0217696242,0.3109727502,-0.0010192796,-0.2680616081,0.3701477945,-0.1722933501,0.2614529729,-0.122810021,-0.159778893,0.0373745412,-0.4497857392,-0.0502865836,-0.0986655951,-0.1323395222,0.4002127051,0.1608778834,-0.1596159935,-0.4839442968,0.3021200597,-0.069291614,0.345036149,0.0765082836,0.1038474068,0.0943718925,0.1545059681,-0.1950791925,-0.4406284988,0.2530566752,-0.2359959036,0.0414475277,-0.0916195959,-0.0021407029,0.0889056846,0.2467318475,-0.2203757167,-0.1574905962,0.0297026597,-0.0926260054,-0.2858552039,0.0036854541,-0.1975900531,0.5724753141,0.075917922,0.1623217762,-0.0595488586,0.1222857535,-0.2260802537,-0.3706210256,-0.2846194506,0.2243590057,0.01142901,-0.0456318222,-0.1860871613,0.4331724942,0.4528889358,-0.0840670317,-0.0524535701,0.1870767325,0.2882177234,-0.210293293,-0.1363984495,0.0566668138,-0.000644474,-0.200467363,0.2589430809,0.4819489419,0.1949970573,0.4277967215,0.1515733898,0.247333765,0.313282907,0.0197373424,-0.0024224338,-0.1308805048,0.1747965068,-0.1996351928,0.1646611989,-0.0104093347,-0.0819548815,-0.1719728261,0.0619493946,0.0059801014,-0.177711457,0.0690228119,0.106748715,-0.1143574268,0.1525004804,0.0578253046,0.3588520885,0.0116988104,-0.1743934751,0.1712271422,-0.2872713208,-0.0165318716,0.1565526873,-0.0600656755,0.3869134486,0.3310392201,0.029359499,-0.0537556335,-0.1345203519,-0.2860231996,0.1273595095,0.3807802796,-0.4825971723,0.2592690885,-0.2457223833,0.3851275146,0.169353053,-0.1360460371,-0.2819573283,-0.6221180558,-0.0652160794,0.4566330016,-0.1154796779,0.1542272866,-0.0666648746,0.0565893091,-0.2285957336,0.3749101162,-0.2166072577,-0.3992305696,-0.16723001,-0.1114408523,0.3502193689,-0.3183429539,0.2055116743,0.1797622591,-0.2845513225,0.193957448,-0.3226080239,0.1029553413,-0.123383224,0.0229367036,0.1417060643,-0.029571183,0.0921471864,-0.2455625087,0.1196155474,-0.2157841325,-0.270142138,0.253403157,-0.1736672521,0.0497094691,-0.2838630378,-0.2666267753,-0.4348909557,-0.1038086563,-0.061585471,-0.2117896378,0.3445335627,0.1505988836,0.0339488834,-0.0018496364,-0.0442048237,0.0065714754,-0.0929128677,-0.0397026315,-0.2193152755,0.0723461807,-0.0794137493,0.008724628,0.0408147462,0.1786851734,0.2683210671,-0.3155303299,0.1561607718,0.0617141463,-0.0350545458,0.1973545253,-0.2283177674,0.2923780382,0.438633889,0.0282353722,-0.0359945595,-0.1160214245,0.1322571039,-0.058225695,0.0320601165,0.0781751871,0.3204672635,0.1851849258,0.6763290763,0.3327035308,0.0967166722,0.3227106035,-0.0756793469,0.0984892473,-0.1708204597,-0.3568721116,-0.2422763854,-0.2925569117,-0.0492549837,-0.2442116737,-0.1278546154,-0.4333050847,-0.258937031,0.5570591688,-0.3641512394,-0.2965089381,-0.018592665,-0.5365771055,0.2150661796,-0.0889232308,0.0328725837,-0.0712174326,0.0218731966,-0.2308910042,0.2551981211,0.1453686655,-0.3435898423,-0.347494781,-0.2690316737,-0.5460581183,0.2908347547,0.1735797077,0.6667127609,0.0073579783,-0.0688790604,0.024522949,-0.1858625412,0.934643209,-0.5217699409,-0.3587026298,0.3193504512,-0.3986980021,-0.2629766464,-0.1964326948,-0.1744777709,0.5878407359,0.4124091268,0.4205800295,-0.1871404052,-0.3219944239,0.1479854137,-0.2433120161,-0.1495640874,-0.0031313982,-0.3532733023,0.001774338,-0.290709883,0.1065585166,-0.1895606667,0.1086736247,-0.3118028939,-0.0441961437,-0.0536382198,-0.112099953,0.1906818897,0.0595516935,-0.0004533102,-0.215348646,0.0571138747,0.051572714,0.05257456,0.2204810381,0.2181242108,-0.2953675985,-0.0804442242,0.1628425717,0.0561924912,0.2555772364,0.274300307,0.2173480392,-0.0336354785,0.1120121554,0.1660029441,-0.3386040926,0.007856058,0.2948587835,-0.0222808532,-0.5951184034,-0.1164301857,-0.0660492405,0.4181356728,-0.2006957233,0.4809945226,-0.3640756905,-0.2475471795,0.4040857852,0.1156168729,0.8487926722,-0.4349771142,-0.0354048721,0.0730250552,0.0347608551,0.0952215195,-0.0260099061,-0.0044387798,-0.3212984502,0.0248712692,-0.0477382168,0.045086313,0.268016398,0.1381977051,-0.173229903,0.3865528703,-0.0107104909,-0.1017483026,-0.0425728597,0.1595099121,0.4235086143,-0.03584693,0.2206964493,-0.0139766643,0.0037834931,-0.112647377,0.0064293607,0.0148405423,0.2116279006,-0.1840919852,-0.4811279774,-0.1047287807,-0.0924683213,0.1672160029,0.1426292062,0.1315797567,-0.0313535407,0.0471192226,-0.0821706727,0.1316116899,-0.079581432,-0.1066345125,0.0960278958,0.3185382485,0.1848517805,0.0571674109,0.2695898712,0.0950201005,-0.2720304132,0.0871152505,-0.2769041657,-0.2623147666,0.0128054395,0.0352960601,0.042424608,0.1547401845,0.0866833255,0.0098200366,-0.1576529592,-0.2034295946,0.0333738551,-0.0181387179,-0.1581920236,0.4551250339,-0.0940324217,-0.3601725996,0.0371090434,0.4425573945,0.2474647462,-0.0432838947,0.1479088217,0.119945161,-0.0285276603,-0.2268243879,-0.2088471651,0.212679252,-0.2034064382,0.276465714,0.0742246285,-0.3003955185,0.1926263273,0.0381830372,0.1105422005,0.3506318331,-0.2179453373,-0.3140774965,-0.5096750259,0.2873867452,-0.0424413122,0.1220761836,-0.0970072001,0.1696494073,-0.1525717229,0.3414262533,-0.2173339576,0.1055871621,-0.1629121304,0.1904532015,-0.1025944203,-0.0219331328,-0.1139365062,-0.1034621894,0.0645447075,0.0122849764,-0.2283412665,-0.0161116887,0.0347344875,0.1336135119,0.0958126262,-0.1613652259,-0.0570763201,-0.035073556,-0.0952621102,-0.1288368404,0.1950083673,0.3021523654,-0.1840545386,0.2073191106,0.2600148022,0.2184175402,0.0725976974,0.3586005867,-0.1578326374,0.07150013,0.0375997126,0.4504765868,0.1239431351,-0.1813465059,-0.0115456022,0.1517645419,-0.0749086514,0.0154511724,0.3202469349,-0.1412194669,0.0418410562,-0.0255629364,0.3905261159,0.4080471396,-0.1006334797,0.0914173201,0.1294074059,0.0766719803,0.0393775292,-0.2356768101,0.5182752609,0.0912385806,0.0712263957,0.3274708688,0.3438836038,-0.1482009292,0.2190123945,0.197262153,-0.0391973108,-0.0060767913,0.2387896627,0.2392288893,0.0759491175,0.2791828811,0.2188096344,-0.3160715699,-0.0826133713,0.3064985871,-0.0430859402,0.5566519499,-0.0231056958,0.1107530594,-0.0504675433,-0.6661481261,0.0684487522,0.0505349934,-0.4930523932,0.2005999535,0.0285409093,0.2260495871,-0.2843838632,-0.4488837719,-0.1733894497,0.4938540459,-0.3161694407,-0.3530267179,-0.2420186847,-0.1568669677,-0.0794109553,0.0364153571,-0.0944969505,0.2113800794,0.739854455,0.0275044981,-0.0592760406,-0.1394855976,-0.3253464103,0.0243773926,0.3143857419,-0.0302809794,0.3852975965,0.0602537356,0.023347998,-0.0070999078,0.1115665585,0.5153449774,0.64861691,-0.4715352058,-0.0051932791,0.1361397356,-0.0671252981,-0.1269532442,0.4410919249,0.0936702043,0.2746122479,0.2441114932,-0.0740941167,0.0241560079,-0.2729542851,0.3861781657,0.3558557332,0.164008975,0.379275471,0.0898746029,-0.1598764509,0.2703370154,0.1561388373,-0.2949358225,-0.2356030196,0.4741261601,-0.4620778561,0.2306520194,0.1898050904,0.053219717,0.0667013675,0.5093153715,0.3080870211,0.029837314,-0.3025135696,-0.1432604641,-0.4209337533,0.0029771344,-0.1952991933,0.2641439736,0.0434208326,0.0151952039,-0.0301825814,0.1026461869,-0.0260660741,-0.226930052,0.338041991,0.3651676476,-0.4900594056,0.2121445835,0.104848884,-0.0730708465,-0.01136606,-0.2620003521,0.3681489229,0.0782691911,-0.0984346867,0.1029912755,0.1684935838,0.0956896096,0.1215946823,0.0248371232,0.3121722937,0.5543993115,-0.0393269211,-0.0194798801,-0.2440269887,0.0761285052,-0.1203723997,-0.0872046575,0.0565832816,0.3794700503,-0.187867254,-0.0857282802,-0.0020314495,0.1167898849,0.0629818738,0.1787511706,-0.4754639566,0.266718179,0.163210094,0.0898106024,-0.1624226123,0.4008593261,0.2307242602,0.3632056713,-0.3065879345,-0.2034911513,0.3098656535,0.004514894,-0.0913119167,-0.551912725,0.1878937483,-0.180244863,0.1410410255,-0.5005385876,-0.3521081209,0.1755147874,-0.2155504525,-0.5762805343,0.2504348159,-0.0661979988,-0.2007854283,-0.0668061152,-0.1931968629,0.1094471961,-0.122339122,0.2190395296,-0.0942723155]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"The parallelism is automatically disabled on `tokenizers` when the process gets forked, while we already used the parallelism capabilities of a tokenizer. We have to do it in order to avoid having the process hang, because we cannot safely fork a multithreaded process (cf https:\/\/github.com\/huggingface\/tokenizers\/issues\/187).\r\nSo if possible, the tokenizers shouldn't be used before the fork, so that each process can then make use of the parallelism. Otherwise using `TOKENIZERS_PARALLELISM=false` is the way to go.","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":75,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nThe parallelism is automatically disabled on `tokenizers` when the process gets forked, while we already used the parallelism capabilities of a tokenizer. We have to do it in order to avoid having the process hang, because we cannot safely fork a multithreaded process (cf https:\/\/github.com\/huggingface\/tokenizers\/issues\/187).\r\nSo if possible, the tokenizers shouldn't be used before the fork, so that each process can then make use of the parallelism. Otherwise using `TOKENIZERS_PARALLELISM=false` is the way to go.","embeddings":[-0.469833523,-0.1674960703,-0.0140371053,0.1800090671,0.0854438543,-0.1905385852,0.2459647357,0.2756293416,0.1016369238,0.1726740003,0.0916070193,0.4495114982,-0.3661407232,0.2278046459,-0.3540114462,0.0023671666,0.1458534151,-0.1064164191,-0.1076287925,0.2664557099,-0.1931275576,0.324302882,-0.2955477238,0.1895089447,-0.5507256985,0.0559619181,-0.0354776047,0.2832084894,-0.0593082495,-0.6136327982,0.2357300073,0.3920495212,0.0007346575,0.4456866086,-0.0001240691,0.1468001455,0.3075380921,-0.0330980793,0.0126842633,-0.3793185055,0.00191086,-0.3051974773,0.2444296628,0.1721025556,0.2013462186,-0.1140399799,-0.1220822409,-0.1859960407,0.3015083671,0.297714591,0.0780777559,0.3740778565,0.2976615131,0.1366759241,-0.1064472646,0.1475401223,-0.0475674085,0.0469951853,0.3224133849,-0.2692290545,-0.1280239373,0.2051905096,-0.2646250427,-0.2231662869,-0.1501526535,-0.2274798453,0.533118248,-0.5728849173,0.2040055245,0.097276777,-0.2752830684,-0.0516938679,-0.4340721667,-0.2814349234,-0.3205261827,-0.5091427565,0.2326749265,-0.0793353468,-0.1704740673,0.177923426,-0.4278577566,-0.2340727746,0.0443510264,-0.101955086,-0.1272596419,0.5708979368,0.1884500533,0.3039306998,0.1742693335,0.1578971744,0.0501511022,-0.0234679058,-0.1019761935,0.1722763628,-0.4143748581,0.0462940373,0.1032643095,-0.3558402658,-0.1825723648,0.1217038929,-0.2746934891,0.1585308909,-0.1526545435,0.0980686024,0.3795083463,-0.0054288921,0.1715634614,0.2787450552,0.2465554625,-0.2089062184,0.0966488421,0.0370953567,0.2416516393,-0.0345837399,-0.1681513041,0.2515701056,0.1012417451,-0.149758637,-0.1228373423,0.1711271703,-0.0874946639,-0.1317547411,0.1335784197,0.2265292406,0.0358568653,0.7409479022,0.0959845632,0.1662244052,-0.3377198577,-0.0669960976,-0.0701267347,-0.0903126001,-0.3008594811,0.125000298,0.1836821288,0.133335948,0.0110817431,0.0407856591,-0.3716782629,-0.2365080118,0.1697461456,-0.246915251,0.0945088789,0.6521362662,-0.0997272208,0.0555373952,-0.0186683238,-0.0468873233,0.0317685418,0.1966137141,-0.4765743911,-0.18053855,-0.243328616,0.0614766963,-0.0035067094,0.2686824799,-0.268176496,0.3582110107,0.377532959,-0.339284271,-0.2765756249,-0.3204176724,-0.3733099103,-0.2689908147,0.0655429363,0.2235532403,-0.3428300619,-0.1210215762,0.137930885,-0.2942818701,0.342664808,0.3083378673,-0.0074371686,0.071666941,-0.1562374234,0.2892130315,0.0653889999,-0.2158596814,-0.0655001327,0.1830647439,-0.0553093627,0.3571151793,-0.0528864972,-0.293948859,0.4720177948,-0.0873196125,0.2116505504,-0.0297005158,-0.1419220567,0.0854478702,-0.4349107742,-0.0488808565,-0.0801716149,-0.1022236124,0.3707356751,0.1900942028,-0.1735134721,-0.4965021312,0.2677107155,-0.1114346385,0.3222943842,0.0715091303,0.1434360743,0.2017229795,0.1305195838,-0.2422705293,-0.377504468,0.2210634649,-0.2406433672,0.0430389866,-0.014628767,-0.0765243322,0.1771068722,0.1960383058,-0.2122108042,-0.1775559336,0.029469993,-0.0402090661,-0.2606339455,-0.0142164379,-0.220460847,0.5387184024,0.0587076545,0.2526376247,-0.0555030666,0.1792503744,-0.1390038282,-0.4171824753,-0.3341962695,0.2444838732,0.0404223092,-0.1081418023,-0.181263268,0.4281125665,0.4532399774,-0.0259470884,-0.078176409,0.1995645314,0.303053081,-0.1107944921,-0.1663929075,0.0268019866,-0.0005163518,-0.1456858218,0.3252193332,0.4752385914,0.2583381832,0.3575917184,0.2223508656,0.1967348456,0.3676941693,0.0150888059,0.0022732753,-0.076925613,0.2094924748,-0.0871469751,0.1809675097,-0.0041484931,-0.1150159165,-0.1338171363,0.0661967769,0.0200306438,-0.152868256,0.0386654325,0.0771412104,-0.1419483721,0.1957781911,-0.0908225253,0.324619323,0.0216593146,-0.1594498307,0.1472177804,-0.2333879024,0.0095722321,0.0928074569,0.0083769727,0.2946350873,0.3339911103,0.0658817887,0.0199006703,-0.29267326,-0.3580291867,0.1330217123,0.3469077647,-0.4571527541,0.2395317554,-0.2332216501,0.4305514097,0.1446413994,-0.1823011488,-0.3057281673,-0.5543479919,-0.0787178427,0.6065199971,-0.1380046755,0.1733262986,0.000842168,0.2153340429,-0.242401734,0.4286283255,-0.2241805494,-0.328969121,-0.2124576867,-0.1325620264,0.3449445069,-0.2445281148,0.2342283428,0.1797371209,-0.317370832,0.0695009604,-0.4021351933,0.1804139763,-0.0857736245,0.0844072625,0.1158489883,-0.0306755491,0.0407308005,-0.2014771551,0.1514792293,-0.2354545593,-0.2923106849,0.2084235847,-0.1431322545,-0.0121390671,-0.2302771956,-0.2263071537,-0.4417113066,-0.1428988278,0.0151706589,-0.2623472512,0.2979028225,0.0964594707,-0.0297172647,0.030512508,-0.1678612381,0.0696147978,-0.1205818728,-0.0331982225,-0.264649719,0.0089349076,-0.1214702502,-0.0549475551,0.0504694581,0.1772573441,0.3300919533,-0.2351266891,0.0866732523,0.0608826503,-0.0232305191,0.1820591241,-0.2267749608,0.3626005054,0.3413053751,0.0460003205,0.0077072466,-0.1542690843,0.0969479382,-0.1244616807,0.0159213953,0.1115942448,0.4085763097,0.2045243084,0.7586420178,0.3068211079,0.0815780014,0.4106377363,-0.1149264872,0.1306596696,-0.1736628562,-0.4086289704,-0.1645549536,-0.2788659334,-0.0181169398,-0.1671079695,-0.1037888527,-0.3872555196,-0.2170632482,0.4757826924,-0.3586519659,-0.2796050906,-0.0407272056,-0.3742516637,0.2615022957,-0.0808700845,0.085449338,-0.1107369736,0.0205805898,-0.193982169,0.3396611512,0.1362640113,-0.2989442647,-0.4563951194,-0.1417516172,-0.5637003183,0.2795660496,0.1750623733,0.7341640592,-0.0234585218,-0.0551584624,0.0587686002,-0.1976893693,0.9357404113,-0.3594240248,-0.4217449129,0.289485395,-0.5509144664,-0.2396155149,-0.1999230087,-0.1644481719,0.5423983335,0.5066506267,0.4541927278,-0.2640604675,-0.3756240904,0.1025584489,-0.2309686691,-0.1180360541,-0.0011557427,-0.329289645,0.0030473436,-0.3068460822,0.1469483227,-0.2366101146,0.1321084648,-0.2422953546,-0.1088757068,-0.1110606268,-0.1524031907,0.1520631313,0.0903228894,-0.0314770378,-0.278401494,0.0413450524,0.1173492596,0.0203277916,0.2515931726,0.1480960995,-0.1945677847,-0.2651031911,0.1867991537,0.1416079551,0.2887451351,0.2998694777,0.1709918976,0.0748204589,0.1025459096,0.2095711231,-0.3290915191,0.1400386393,0.3358182907,0.069967337,-0.5486969948,-0.0179219861,-0.0950674936,0.4445157945,-0.1411133856,0.5347307324,-0.3712587357,-0.2485485822,0.369253397,0.0327239521,0.8252356052,-0.3295942843,-0.0434363037,0.1655754298,0.0162167158,0.1495186985,-0.1232438385,0.0196426734,-0.3308779597,0.1058773324,-0.0595270321,0.0748168975,0.3638460636,0.1011833027,-0.0730094612,0.3764547706,0.0869753957,0.0188113209,-0.0523898043,0.2291526496,0.3855039775,-0.1186162606,0.123710148,-0.0242683552,-0.010819748,-0.0757304206,-0.007982471,0.049107179,0.2691777349,-0.171704486,-0.4417435229,-0.1571897864,-0.1704281271,0.1635919213,0.2262084484,0.1969398558,-0.0660275072,-0.1102877334,-0.0307533927,0.1097054854,-0.0953577831,-0.1034523174,0.0131335901,0.2929879725,0.1415536255,-0.0207570288,0.3205946982,0.0703554899,-0.1965799779,0.0415908359,-0.2161634713,-0.2890794277,0.0082086949,0.0118186623,0.1558249146,0.1179110631,0.1515677124,0.0131440293,-0.0991609693,-0.2529164255,0.0169841554,0.0426879786,-0.2005050927,0.3394840658,-0.0869794264,-0.3004573286,-0.0263283122,0.4269422591,0.1343546659,-0.1657799184,0.0890286714,0.1459604353,-0.0572540276,-0.1820950657,-0.2436644733,0.3081692457,-0.1138097048,0.2650239766,0.0591093674,-0.3618755341,0.1516514421,0.1528609246,0.0431237072,0.2574946284,-0.2684714496,-0.3503492773,-0.4898902774,0.2911831141,-0.0430532135,0.1423331946,0.0262090247,0.2362660021,-0.2149010301,0.3182534277,-0.2029466331,0.0799208432,-0.1292322278,0.1701392978,-0.1570792049,-0.1027873829,-0.1325781047,-0.0717893094,0.0402046405,-0.0084999753,-0.2677114904,-0.0053711818,-0.0168909021,0.1345700175,0.164226234,-0.2353510559,0.0058795013,0.0347650088,-0.1145366132,-0.2087295651,0.2275354415,0.2921721637,-0.2415676117,0.0413199998,0.2643428445,0.1993068159,0.0712755993,0.3201953471,-0.0814494118,-0.0106403297,0.072816655,0.4271273017,0.0593409352,-0.1600848734,0.0053648516,0.1601385325,0.1081877798,-0.0624859706,0.2911910713,-0.1417414099,-0.0191018991,-0.0337679274,0.3789577782,0.4425649047,-0.0806519836,0.0001757802,0.1910727173,0.0638121516,0.0837064981,-0.2308319062,0.6156703234,0.0289506037,0.1046122834,0.2869160473,0.3824906051,-0.0869988129,0.1666943878,0.1613331437,0.0975132436,-0.0389079154,0.2546490729,-0.0366277322,-0.0137625365,0.3162634671,0.2590506673,-0.3143553138,-0.1352953762,0.2599980533,-0.0384014137,0.5948092937,-0.0969623923,0.2248375267,-0.0666981414,-0.6823627353,0.0057196263,0.0434206612,-0.461548388,0.2214369625,-0.0417677499,0.2390712947,-0.3056819737,-0.4851911068,-0.1789880693,0.5870895386,-0.3731824756,-0.3623032272,-0.2752656341,-0.1022213846,-0.0483383648,0.0446069688,-0.0661511868,0.2078546882,0.6278228164,-0.0410969555,-0.0900360644,-0.2425192595,-0.334808439,-0.0144502288,0.2810928822,-0.0061646439,0.4593024552,-0.0634730831,0.0371081978,-0.0783298239,0.2297443748,0.4838189185,0.6397119164,-0.323633641,-0.1313558072,-0.0015709874,-0.0112166051,-0.1275711358,0.5054551363,0.0100009134,0.2311814874,0.1935999691,-0.0822726935,0.0042334157,-0.3318096399,0.3296576142,0.2684663534,0.0815627053,0.3986719549,0.0138111785,-0.0815168768,0.212375924,0.0896234959,-0.2625283897,-0.1877289116,0.4311327636,-0.4376075268,0.1901277453,0.1663139164,0.0485992022,0.1046998575,0.3792058527,0.3668218851,0.0105573684,-0.266384542,-0.1219752505,-0.4246735871,0.0550438762,-0.1286810935,0.3264997602,0.0281256288,-0.001522469,-0.0064913142,0.1175889149,-0.1066418588,-0.1334386319,0.3874142766,0.2544837892,-0.54399997,0.3324020505,0.0594327636,-0.0865356922,0.0030166127,-0.2445358634,0.4164734781,0.0530777127,-0.1391578317,0.139038533,0.1647678465,0.0155448886,-0.0387466922,0.0465328619,0.331342876,0.5452202559,-0.0247943699,0.0179172922,-0.2891092598,0.0822054446,-0.162248224,-0.0324375778,0.0186814629,0.4277773798,-0.1754865646,-0.0600115322,-0.063553676,0.1609639078,0.0716930553,0.092915751,-0.4696351886,0.3175766468,0.1130199134,0.0851163566,-0.1704145074,0.4451139569,0.2408338338,0.3864465058,-0.3660425246,-0.2012934089,0.3071762025,-0.1852543354,-0.0721664056,-0.6507634521,0.1840498447,0.0079308059,0.0638974607,-0.4796611965,-0.3711153865,0.148710072,-0.2079419494,-0.5344145298,0.2810973823,-0.1004102454,-0.244105354,-0.0879499093,-0.1413968951,0.1348972023,-0.0782699957,0.2719238102,-0.026052257]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"> Thanks for taking a look. I pulled the master but I still see the key error.\r\n\r\nI am no longer able to get the error since #659 was merged. Not sure why you still have it @timothyjlaurent \r\nMaybe it is a cache issue ? Could you try to use `load_from_cache_file=False` in your `.map()` calls ?","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":56,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\n> Thanks for taking a look. I pulled the master but I still see the key error.\r\n\r\nI am no longer able to get the error since #659 was merged. Not sure why you still have it @timothyjlaurent \r\nMaybe it is a cache issue ? Could you try to use `load_from_cache_file=False` in your `.map()` calls ?","embeddings":[-0.4314345121,-0.1244709343,-0.0213454217,0.2268492579,0.1299650222,-0.1009751707,0.216061607,0.4109111428,0.0611687526,0.1098547429,0.0194626674,0.4327996373,-0.3604695201,0.3772411644,-0.339530468,0.0885623991,0.0790837929,-0.0142771583,-0.3960187733,0.2743343115,-0.3150272071,0.2834353745,-0.2949822843,0.1656470895,-0.5235921144,0.0930623934,-0.1302800775,0.3110467494,-0.0348664373,-0.5802859664,0.300784409,0.2864485085,-0.0216589849,0.4401558936,-0.0001220013,0.0999585539,0.3847609162,-0.0793480724,-0.071606338,-0.3503724635,-0.247675553,-0.2170802653,0.2827697694,0.1046241671,0.2202919275,0.0048059016,-0.1133266315,-0.1922502816,0.2761889696,0.3521153033,0.119101733,0.1872767806,0.3880855441,0.0826243013,-0.0354129523,0.09376055,0.0023207404,0.0315022543,0.3187670708,-0.4253243804,-0.0663856417,0.3480861485,-0.2237595171,-0.2809122801,-0.0422265939,-0.2567038834,0.6990743876,-0.6044112444,0.215970993,0.0483471043,-0.2460377067,0.0421748459,-0.4069693983,-0.2098711282,-0.2622978687,-0.4920367301,0.2894101143,-0.0301969405,-0.2084701508,0.1472867131,-0.4892234802,-0.3158850372,0.0184610151,-0.1103789732,-0.0474434979,0.5334964395,0.2437919229,0.2392061055,0.2650356591,0.0902934447,0.1246623024,-0.0066084205,-0.1508881003,0.1766755432,-0.4339984059,0.0645188317,0.1089489758,-0.0413448699,-0.2220738232,-0.0226173643,-0.2547248602,0.1037639454,-0.0433884822,-0.0031188682,0.4134060442,-0.0021763786,0.2312474698,0.2805155218,0.208972007,-0.1722468436,-0.0612269156,0.0473840125,0.3236121833,-0.1497556567,-0.0274684988,0.3400017917,0.1807366014,-0.20466052,-0.0820394158,0.1177491322,-0.0378100127,-0.1827217788,0.0796588585,0.2012063563,0.0385181494,0.7639835477,0.1573572308,0.2139507681,-0.2573189437,-0.0196505692,-0.0640080944,-0.0725231767,-0.3861177266,0.0672839507,0.266523242,-0.1440191865,0.0057661026,0.0307274163,-0.2962273955,-0.2732139826,0.0389924683,-0.2264907956,0.1082785428,0.6682197452,-0.0568307154,-0.0038681626,-0.0184367131,-0.1838995218,0.102129519,0.2575799227,-0.5337067842,-0.2903423607,-0.1411552429,0.1126205102,0.0741786584,0.3089140654,-0.1461556107,0.2622054219,0.4196609557,-0.3520433009,-0.3468497992,-0.2722082734,-0.3974720538,-0.3425095379,-0.01547537,0.2808103263,-0.4167695642,-0.0224714801,0.0201118924,-0.1394563168,0.2266025692,0.1803773642,-0.067696631,-0.0393480286,-0.124489978,0.0694298148,0.0809170604,-0.2601244152,-0.2011919767,0.2847458422,0.0258586109,0.270404011,-0.0125878602,-0.2769052386,0.3264488578,-0.2117162198,0.2866227925,-0.097726129,-0.1917823404,0.0468083955,-0.4094309211,-0.0715955272,-0.0588161312,-0.1225052476,0.4039655924,0.1986887753,-0.1238547936,-0.5356959701,0.2904838026,-0.0942864716,0.3269191086,0.1106510311,0.0971123651,0.0743148476,0.1996762604,-0.2014030516,-0.4499591589,0.2379933894,-0.277695924,0.0345757082,-0.0698416159,0.0290762931,0.0377738364,0.1411239356,-0.2355711907,-0.1979332119,0.07768365,-0.1049174145,-0.2467697412,-0.0876589268,-0.2032495886,0.6826040745,0.1674461514,0.1178386435,-0.0962090567,0.143592,-0.2462816685,-0.3751718402,-0.2172571868,0.2264219671,0.0565494336,-0.0111123715,-0.2679614425,0.4512029588,0.3154532015,-0.0701540336,-0.0241982825,0.1356793195,0.2313723266,-0.2101558447,-0.1253693849,0.0301832147,0.0150996707,-0.2139487416,0.2164917588,0.4391390979,0.0700734332,0.4827993512,0.198990792,0.24452658,0.2599719763,0.0034953358,-0.005685681,-0.1511238217,0.2132588327,-0.2349467725,0.1760885119,0.0235666651,-0.0262152888,-0.1778630912,0.1258432865,0.0529481769,-0.146593824,0.0932001844,0.0199284125,-0.1221553534,0.055495549,0.0848882124,0.4538980126,0.0019589982,-0.1927702278,0.2290842682,-0.2108611763,0.0234586895,0.2167144865,-0.0559656136,0.3995956779,0.3053631783,0.0496963747,-0.0101163704,-0.1340932995,-0.1709618717,0.0689922422,0.3572278619,-0.4408538342,0.1790742427,-0.2723540068,0.2788523436,0.1270198375,-0.1176956519,-0.3178197742,-0.5795158148,-0.0903853327,0.4781474173,-0.0622863099,0.138661027,-0.0254734345,0.0500340797,-0.1990747005,0.377076447,-0.2009774745,-0.405901432,-0.1911114007,-0.0859188139,0.3534836769,-0.336589396,0.2275673151,0.1413680166,-0.2084483057,0.198938489,-0.3869027793,0.0617677942,-0.1056495979,0.0377923809,0.1049740314,0.0098565361,0.0452488326,-0.2347851843,0.1706171483,-0.2592055798,-0.2953844965,0.2534521222,-0.160486415,0.1175030544,-0.2453966439,-0.331682086,-0.3674536049,-0.1288785487,-0.0563551635,-0.230546087,0.3234847486,0.1497080028,0.0233152695,-0.0651255623,-0.0555437766,0.0701147541,-0.1063601524,-0.0443348512,-0.1975311637,0.0210301783,-0.0967616066,0.029542001,0.0621929914,0.2017765492,0.2890485525,-0.2750006914,0.0504876673,0.0505750328,-0.0595088117,0.1866527796,-0.1573631316,0.3616135716,0.4519875646,0.0313639119,-0.0434337296,-0.1598108411,0.0565444827,0.0406268053,0.0759693831,0.0609644838,0.330260545,0.1010796502,0.6614121795,0.3625734448,0.1548636109,0.3272961676,-0.0050969515,0.1325636655,-0.248461321,-0.4285871387,-0.2518708408,-0.3223823309,0.0676503181,-0.2475498319,-0.0832686424,-0.4537869692,-0.3686927259,0.4809542298,-0.4484333396,-0.2922631502,-0.0461664237,-0.4731544256,0.245474115,-0.0141626596,-0.0119094169,-0.0777527019,-0.0021200173,-0.2190319896,0.259960264,0.0750967786,-0.3458838463,-0.2463163435,-0.3370688856,-0.4683030248,0.3457066715,0.180027023,0.6090081334,0.0361394063,-0.0807513744,-0.0433470756,-0.1559733152,0.8825868368,-0.6140753627,-0.377945751,0.2825070918,-0.3890219331,-0.2425138205,-0.1773129553,-0.0925871432,0.6058717966,0.3386295736,0.4422083199,-0.1263378561,-0.2924227417,0.2054613084,-0.1899690181,-0.0901443884,0.012636954,-0.3751445115,-0.0160286427,-0.2455603331,0.0628633425,-0.2198968977,0.1254925579,-0.2943683267,-0.0085715391,-0.0713992268,-0.1021113545,0.1646036804,0.039050702,0.0290780086,-0.181882292,0.0904806629,0.0928516313,0.1154310256,0.2179584354,0.2940433025,-0.3483678401,-0.1547678858,0.1047963724,0.0128391553,0.2593703866,0.287494719,0.2066312283,-0.0999755561,0.1383375674,0.121935688,-0.2631670237,-0.0537210442,0.3026966453,-0.0508233756,-0.5315156579,-0.120104596,-0.0578831024,0.4363302588,-0.1913834065,0.5237945318,-0.4564827681,-0.2383583635,0.3694995642,0.015390167,0.8982658386,-0.406665653,-0.0524984263,0.0174785405,-0.0084767994,0.1157455593,0.0991774127,0.0434692092,-0.3183643222,-0.0740953162,-0.0107520884,0.0392740928,0.2415270358,0.1067193598,-0.1916190833,0.4355043471,0.0005615147,-0.0736709833,-0.017832147,0.1360589266,0.3495913148,-0.0522040911,0.110069558,-0.0046372432,-0.0166387986,-0.0884450153,-0.0157269053,-0.0205743834,0.2720359862,-0.2323769331,-0.4313637316,-0.058885552,-0.124622345,0.1750725508,0.1152960882,0.0633786991,-0.0505629294,0.089927569,-0.046570044,0.1361654103,-0.0849648416,-0.0713881478,0.1129709631,0.3522197008,0.1447092891,0.027493583,0.2665855885,0.079549849,-0.2798361182,0.0733124763,-0.2621832788,-0.2616207302,-0.0750687495,0.0655354112,-0.0213131998,0.1660890281,0.1111551449,0.0022645642,-0.1481415331,-0.2251985818,0.0727070644,-0.0243667271,-0.0966145992,0.5090479851,-0.0795824751,-0.3521074653,0.0703078434,0.4957988858,0.2781369686,-0.1033789143,0.1932804585,0.1098286435,0.0112323705,-0.2096003592,-0.2345390916,0.144998759,-0.1942834109,0.2652489841,0.0830777511,-0.1989101768,0.2920131087,0.0552815571,0.079301782,0.3829930127,-0.1570692956,-0.2903122604,-0.5277741551,0.2857143581,0.0204585902,0.0969087109,-0.1263668239,0.1305980682,-0.0521103702,0.3383786082,-0.2322706431,0.1291428655,-0.0689734071,0.1807765216,-0.0222007874,-0.0191036426,-0.1341444701,-0.1393241882,0.0646474585,0.0359888338,-0.204090625,-0.0415782295,0.0906358361,0.1251113266,0.0482723415,-0.1485891491,-0.1163560301,-0.0775001571,-0.0856593475,-0.1161112487,0.1988503635,0.2801808417,-0.1707325727,0.1351584345,0.1905764341,0.1141842753,0.1202034801,0.380643338,-0.2339485139,0.0605882518,0.0225697104,0.3719251156,0.121101521,-0.2014991641,0.0049367151,0.0348655172,-0.1434596479,0.0187816452,0.3351595998,-0.2365337163,-0.0248174686,0.0341962799,0.3707671165,0.4022577107,-0.0860794336,0.1595240086,0.1038326845,0.1220131516,-0.0071791317,-0.2172746658,0.4551473558,0.0618552454,0.039396219,0.2695762515,0.2744847536,-0.1645928025,0.2384148836,0.202052325,-0.1007001474,-0.034923967,0.2911992073,0.2524302602,0.1633526832,0.1991352439,0.2327963859,-0.2741599977,-0.122725822,0.2849218845,-0.0553044751,0.4940327704,-0.0550819635,0.0939499959,-0.0180264916,-0.6432769299,0.046890568,0.1129795462,-0.5432616472,0.2139881253,-0.0143762426,0.2575624287,-0.3294780254,-0.4762943089,-0.1783710122,0.4282780886,-0.2955001593,-0.3060948551,-0.1915060133,-0.1794612706,-0.0360155329,0.0738011152,-0.0777120292,0.2517944872,0.7757788301,0.043745704,-0.074963145,-0.1485755891,-0.3636286855,0.0480315946,0.3650012612,-0.0165604055,0.3460436463,0.0974886417,0.012390133,0.0072354665,0.223030746,0.5292618871,0.6800273657,-0.4986771047,0.02864106,0.1742101312,-0.0936603099,-0.1186063364,0.3716343939,0.1848436743,0.2645902932,0.2285882384,-0.0288324337,0.011145425,-0.1483448148,0.2338471115,0.383510977,0.0782244056,0.2400386333,0.1224801913,-0.1629377604,0.2473020852,0.1672240794,-0.3199607432,-0.2731571794,0.5410600305,-0.3435688913,0.1883387417,0.1087425351,0.0782318562,0.1328066438,0.506302774,0.3657382131,0.0652461126,-0.3326123953,-0.2014546841,-0.4707191288,-0.0253033228,-0.205199033,0.3350790739,0.0426931381,0.0332994536,-0.0573373437,0.0522836559,0.0825937092,-0.295743376,0.3538038135,0.3220769763,-0.5336776376,0.2130997628,0.1855689883,-0.0482999869,0.0373567939,-0.3035821319,0.3494933844,0.0925251767,-0.0680242479,0.0800444484,0.1540370136,0.1159981117,0.2276992798,0.0035316825,0.273909837,0.5416819453,-0.0236294158,-0.0608467087,-0.2412253469,0.0493029803,-0.0706125721,-0.0246931799,0.0966273174,0.3412883878,-0.219337374,-0.0735443905,-0.0155460425,0.102961041,0.0029367823,0.1824575365,-0.4736458361,0.1919768602,0.2427234501,0.0934370235,-0.1141403466,0.4848996103,0.1436389536,0.3039221764,-0.3430659175,-0.2384911478,0.3032128215,0.0595472828,-0.1561750919,-0.5542379618,0.2285564989,-0.1924533099,0.150055632,-0.5447407365,-0.3107988834,0.1632043123,-0.1011639833,-0.5235194564,0.2352226675,-0.1159075648,-0.1001254097,-0.0555109382,-0.2853479087,0.1060264707,-0.1125022024,0.1802150756,-0.0616510436]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"> The parallelism is automatically disabled on `tokenizers` when the process gets forked, while we already used the parallelism capabilities of a tokenizer. We have to do it in order to avoid having the process hang, because we cannot safely fork a multithreaded process (cf [huggingface\/tokenizers#187](https:\/\/github.com\/huggingface\/tokenizers\/issues\/187)).\r\n> So if possible, the tokenizers shouldn't be used before the fork, so that each process can then make use of the parallelism. Otherwise using `TOKENIZERS_PARALLELISM=false` is the way to go.\r\n\r\nOk thanks :)\r\n\r\nIs there something we should do on the `datasets` side to avoid that that the program hangs ?\r\n\r\nAlso when doing `.map` with a tokenizer, the tokenizer is called once on the first examples of the dataset to check the function output before spawning the processes. Is that compatible with how tokenizers are supposed to be used with multiprocessing ?","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":140,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\n> The parallelism is automatically disabled on `tokenizers` when the process gets forked, while we already used the parallelism capabilities of a tokenizer. We have to do it in order to avoid having the process hang, because we cannot safely fork a multithreaded process (cf [huggingface\/tokenizers#187](https:\/\/github.com\/huggingface\/tokenizers\/issues\/187)).\r\n> So if possible, the tokenizers shouldn't be used before the fork, so that each process can then make use of the parallelism. Otherwise using `TOKENIZERS_PARALLELISM=false` is the way to go.\r\n\r\nOk thanks :)\r\n\r\nIs there something we should do on the `datasets` side to avoid that that the program hangs ?\r\n\r\nAlso when doing `.map` with a tokenizer, the tokenizer is called once on the first examples of the dataset to check the function output before spawning the processes. Is that compatible with how tokenizers are supposed to be used with multiprocessing ?","embeddings":[-0.475686878,-0.1844159812,-0.0169958789,0.2049914598,0.0819928423,-0.1816118509,0.2112884074,0.2972383201,0.1219884381,0.1559848636,0.0783844516,0.4451454282,-0.3864643872,0.2127695829,-0.3733434379,-0.0014688398,0.1317390054,-0.115358308,-0.1595412195,0.2606368959,-0.2197670341,0.3095827401,-0.2703738213,0.1662527472,-0.5590302944,0.038462624,-0.0322675109,0.2863253653,-0.0549480692,-0.6010497808,0.2577683032,0.3555739224,-0.0043139998,0.4539803267,-0.0001233048,0.1498548985,0.303083688,-0.0515437759,0.0069506946,-0.3851677477,-0.0252814069,-0.277089119,0.2699133754,0.1735979319,0.2168494463,-0.0784442946,-0.1051158234,-0.2321279496,0.3105889857,0.3287059665,0.0884089097,0.371316731,0.3021605015,0.1215195507,-0.0908219069,0.1545794457,-0.0447724611,0.0636395812,0.3348871469,-0.2771942616,-0.1457321644,0.224081412,-0.2541387081,-0.2269081622,-0.1373940408,-0.2417161018,0.5890159011,-0.5875577927,0.2012948692,0.0768413022,-0.3183762133,-0.0541419387,-0.4318971038,-0.2669734955,-0.3097957671,-0.4587308466,0.2185549736,-0.0512987785,-0.1742665917,0.1737106591,-0.449732542,-0.2364093512,0.0335571207,-0.0911895186,-0.0923879296,0.592274785,0.1959981918,0.3030861914,0.1892191321,0.1432417631,0.0748298094,-0.0224433299,-0.0954662189,0.1679126769,-0.4035385549,0.0827436522,0.0919434577,-0.2895827889,-0.1892665476,0.0969899893,-0.2585215867,0.1280376464,-0.136983037,0.0878550261,0.384493947,-0.0407928117,0.1682704985,0.283328265,0.217635408,-0.1963160336,0.0822511017,0.0322841629,0.2650047541,-0.0385889374,-0.1799195558,0.2314647287,0.1087988988,-0.1337117553,-0.142317608,0.1458658278,-0.0681992397,-0.129616037,0.1331932545,0.2315075994,0.0405019708,0.7244006395,0.0960810855,0.1575881839,-0.3393119574,-0.0529349521,-0.0782698169,-0.0812074989,-0.3169773221,0.1219127476,0.2087178975,0.0912899151,0.0224034078,0.0469061136,-0.3484089971,-0.23330836,0.1524538547,-0.2576985359,0.1139875054,0.6639273763,-0.094220385,0.0524130799,-0.0007993848,-0.057426706,0.0372141451,0.2017600387,-0.4929462969,-0.1640477031,-0.2443074584,0.0716860816,0.004053521,0.2785006166,-0.2109918296,0.3353477418,0.3713299334,-0.3363822103,-0.262481451,-0.3061409891,-0.3732001185,-0.2744423747,0.0567431375,0.2622538805,-0.3388674259,-0.1239289865,0.0962545052,-0.2862996161,0.3363772929,0.300375849,0.0045218398,0.0158558898,-0.1642980874,0.2672314644,0.0184795447,-0.2350241691,-0.0645949543,0.2141822129,-0.0669817626,0.3825890124,-0.0254147574,-0.2721015513,0.4200842381,-0.082127206,0.2206921428,-0.0719713718,-0.1598526686,0.0723826289,-0.4469228089,-0.0518158935,-0.0820843726,-0.1019434556,0.3698016703,0.1915018857,-0.1476291567,-0.5076674819,0.2738440931,-0.0926945135,0.3278764784,0.0927436948,0.1244724765,0.174836427,0.1179584563,-0.2395222783,-0.4158412516,0.2211656868,-0.2326014191,0.0300400667,-0.0494024269,-0.0657086521,0.1601954401,0.2033800483,-0.2015704662,-0.1870108992,0.0405649096,-0.0309698377,-0.2700848877,-0.015052231,-0.1966944784,0.5803508162,0.0635141805,0.228332147,-0.0783324763,0.2001926899,-0.1512623429,-0.4348412156,-0.3166466951,0.2367548048,0.0390143655,-0.0944889039,-0.171459645,0.4456483424,0.4693530202,-0.0102848858,-0.0764804259,0.187821582,0.2870840132,-0.1167798489,-0.1751221865,0.0338017605,-0.0003312245,-0.1464110464,0.3196179569,0.4675672352,0.2330543101,0.3575600088,0.2539694607,0.2081482559,0.3439673185,0.0218104981,-0.0242783315,-0.1146609187,0.2205082923,-0.0994288772,0.2154528499,-0.0023469096,-0.111907579,-0.1728881747,0.0676754862,0.0434737615,-0.1648951918,0.0478622429,0.0696721673,-0.1340031028,0.2139618248,-0.0598307773,0.3442212939,0.0206978582,-0.1673970371,0.1541671008,-0.2430072576,-0.0026990883,0.0961449519,-0.0014894158,0.3433480263,0.3187732995,0.0600362606,0.0198895261,-0.2822582126,-0.3517171144,0.1048051715,0.3378388584,-0.4633424878,0.2322875261,-0.2338479906,0.3845455945,0.1340387315,-0.1681765318,-0.335367471,-0.5566539764,-0.0958939344,0.5814811587,-0.1384188384,0.1703567654,0.0081468765,0.1842138767,-0.2478107214,0.430931896,-0.2272590548,-0.3418935835,-0.2030451447,-0.1250691712,0.3325464129,-0.2567776144,0.2517679632,0.1759171337,-0.2938838303,0.0872063711,-0.3873684406,0.1526650935,-0.0731864944,0.0602894053,0.1090695411,-0.015785031,0.004555665,-0.2544361353,0.1518446803,-0.2350004613,-0.2803879976,0.2101806849,-0.1396773756,-0.0038290673,-0.255975306,-0.2211224735,-0.4681296647,-0.1458347142,0.0394946933,-0.2275548428,0.2909946144,0.0990902483,-0.0002406774,0.0341753587,-0.1882955581,0.0580974258,-0.0850370899,-0.0544365905,-0.2630152702,0.0116218552,-0.1198362038,-0.0642387271,0.0606498718,0.1806460023,0.3339760303,-0.2563076913,0.1001727358,0.0493990965,-0.037088979,0.1855917871,-0.2218176723,0.3342913985,0.3710938692,0.0414133444,-0.008143872,-0.1233919859,0.1171029359,-0.1319320798,0.0046587517,0.0892834142,0.3490759134,0.2013565749,0.7101199627,0.3344255686,0.0820886344,0.4097701609,-0.1185182706,0.1328537017,-0.1741563231,-0.4208975732,-0.1677476019,-0.2943623066,-0.0208642744,-0.1777105331,-0.1070094183,-0.3720279932,-0.2348513603,0.4854686558,-0.3625723124,-0.282332927,-0.0557362698,-0.3886998892,0.2692979574,-0.0867179781,0.0599014387,-0.1011542603,0.0085388208,-0.1931377053,0.3219674826,0.1452974826,-0.2985174656,-0.3974402547,-0.1347713172,-0.5677657723,0.2774596512,0.1943188757,0.7197498083,-0.0222193208,-0.0550324991,0.0424780697,-0.1738960296,0.9335653186,-0.3680361211,-0.4003145993,0.2850964367,-0.522855401,-0.2758197784,-0.1567967534,-0.182553798,0.5299737453,0.5033144355,0.4867070019,-0.2295978516,-0.3556586504,0.0830113143,-0.2391573787,-0.1092305109,0.0121597359,-0.3224848807,0.0144701926,-0.2834955752,0.1093770862,-0.2248125374,0.1295709312,-0.238933146,-0.0865680426,-0.0915002972,-0.1397095174,0.1736970842,0.0825343207,-0.009950445,-0.276592195,0.0427071676,0.1082732528,0.0614705272,0.2878067791,0.1701036543,-0.2898445129,-0.2364211977,0.17542395,0.1405265331,0.287101537,0.2971283495,0.1893925071,0.059080027,0.1080006957,0.2241963744,-0.3453711867,0.1258979142,0.3439379632,0.0436476022,-0.5239552855,-0.0237532742,-0.0868227556,0.4585831761,-0.1333044022,0.5331063271,-0.3373075128,-0.2532940805,0.377679646,0.0537825413,0.8072373867,-0.3265657425,-0.022922527,0.156232506,0.0173007287,0.1537858099,-0.0873929635,0.0119428635,-0.3142170906,0.063477017,-0.0452584811,0.0772980824,0.3754208684,0.1355510801,-0.1105250046,0.3550403416,0.0733519346,-0.0283137728,-0.0541696325,0.2516599,0.3914587796,-0.1080615595,0.105713293,-0.0169226322,-0.0128005994,-0.0900202319,-0.0018770141,0.0472421087,0.272573024,-0.1645089388,-0.4301676452,-0.1503295004,-0.1936087757,0.1699729413,0.1953967065,0.2028210908,-0.0608462989,-0.0689832419,-0.0499991477,0.1060868204,-0.1081335023,-0.0966314226,0.0212486107,0.2923203707,0.1452783644,0.0103634102,0.326582253,0.0780173168,-0.1968090683,0.0214372892,-0.2279840112,-0.2474514991,0.0210342314,0.0023376418,0.1069788113,0.1279899776,0.1304868013,-0.0097234882,-0.1147375032,-0.2371556163,0.0243882183,0.0176015515,-0.1802405566,0.3344123065,-0.0578747764,-0.3232226372,-0.0333039612,0.4687096775,0.1090976968,-0.1480857134,0.0871356577,0.1562364101,-0.0553741865,-0.1955142319,-0.2417651713,0.3024613857,-0.1383021474,0.2691172063,0.0501745827,-0.338375181,0.1524509192,0.1182908565,0.0824211016,0.2784633338,-0.2389619052,-0.34245646,-0.5004399419,0.3107463121,-0.022935247,0.1211043075,-0.0049041314,0.2242474556,-0.1838408858,0.345674336,-0.2167030573,0.0904567912,-0.1229842603,0.1570657641,-0.1399089843,-0.0782787427,-0.1073134318,-0.0705691203,0.0576645471,0.0072551416,-0.2576489151,-0.0122712329,-0.012342792,0.123655051,0.1437781751,-0.2519907057,-0.0008615344,0.0077841403,-0.1153548211,-0.2093736678,0.2162919641,0.3250153065,-0.2374074459,0.0713872388,0.2589741051,0.2015163898,0.0900138989,0.3069147766,-0.1149785519,0.009032934,0.0585633256,0.4187963903,0.0750647411,-0.1738540381,0.0440644249,0.1432626843,0.08318194,-0.057663992,0.2839287221,-0.1335043609,-0.0173590165,-0.0663808361,0.4031858742,0.4276712835,-0.0851340294,0.0137510141,0.1785940677,0.075170204,0.0693164244,-0.2387713045,0.6134283543,0.0490218699,0.1219353005,0.2915585339,0.357424438,-0.088723138,0.1632320732,0.1584088802,0.0630488247,-0.0033907278,0.2510570884,0.025180215,0.0141691612,0.3006717265,0.2103434056,-0.3132793307,-0.1205042154,0.2594989538,-0.0557238497,0.6039002538,-0.1152230278,0.2057898194,-0.0456069298,-0.6763283014,0.0114662955,0.0381749645,-0.4961703718,0.2350894213,-0.034553986,0.2672608793,-0.2967121005,-0.4756687582,-0.1563878208,0.5736584067,-0.3739906847,-0.3585554659,-0.2344927639,-0.1247158572,-0.0726851746,0.0357063375,-0.0670153722,0.1927326471,0.6713910103,-0.025535332,-0.0532723591,-0.2078116834,-0.3529209495,-0.0205989182,0.2887110412,-0.013829736,0.4510281682,-0.0193171389,0.0346368887,-0.0704786927,0.2078043073,0.4842162728,0.6428656578,-0.3257591724,-0.1031232178,0.0446996875,-0.0426338091,-0.1281030178,0.497990191,0.0402995385,0.243681401,0.2387828082,-0.0687588602,-0.0063498174,-0.3184901476,0.347750634,0.2925030291,0.0784480348,0.3953698277,-0.0102349278,-0.1058872342,0.2256115228,0.0787453875,-0.2978062332,-0.2154137492,0.4452059865,-0.4718310535,0.2071361542,0.1623845249,0.0543016754,0.0993550047,0.4279839098,0.3365131021,0.0190367829,-0.2900013924,-0.1771738082,-0.437306881,0.0562303066,-0.155432269,0.3129951358,0.0160792004,-0.0343838073,-0.0275724754,0.0940406322,-0.1088184044,-0.1512676626,0.3697281778,0.2707333565,-0.5438629389,0.3023406267,0.0414217748,-0.083915405,0.0237436146,-0.2340019494,0.411295414,0.0445201807,-0.1232690066,0.1092156097,0.1708803326,0.0304025132,0.0013582433,0.0325281173,0.3359393775,0.5460197926,-0.0598560274,-0.0129342908,-0.2739146352,0.0819800124,-0.1434948593,-0.043173641,0.029147109,0.4516671896,-0.1946020275,-0.0968337283,-0.0568246879,0.1586852521,0.0600403026,0.1141104102,-0.4758608341,0.3058161736,0.1352462173,0.0754725933,-0.1574523747,0.4557605982,0.2329504192,0.3832819462,-0.3385997117,-0.2078127265,0.3144842386,-0.1796894521,-0.0677396134,-0.6403833628,0.1724697351,-0.0376214497,0.0543130897,-0.4913704991,-0.3654240668,0.1549299955,-0.2062969059,-0.5398041606,0.2500121593,-0.1176326722,-0.237057656,-0.0799300075,-0.1416481137,0.1694273204,-0.0873231962,0.2898893058,-0.0397710688]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"#659 fixes the empty dict issue\r\n#688 fixes the hang issue","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":11,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\n#659 fixes the empty dict issue\r\n#688 fixes the hang issue","embeddings":[-0.4041630924,-0.1525838524,-0.0323358215,0.2152654678,0.1397531182,-0.1405668408,0.2812687755,0.3182485998,0.1930628568,0.1751671433,0.1018442884,0.4044229388,-0.4391215742,0.3682264984,-0.3513095081,0.0251598619,0.0772947669,-0.0434477814,-0.3377382457,0.2193848491,-0.2656829953,0.2496121973,-0.3628329039,0.0937624872,-0.5157561898,0.0884902924,-0.0346404091,0.2518197596,0.0153893642,-0.5729407072,0.3609697819,0.238561973,-0.0252481736,0.4412714839,-0.0001234641,0.1508419663,0.3837755322,-0.1290387511,-0.1070226878,-0.4026999474,-0.2377738804,-0.2075582296,0.2964994907,0.1294531971,0.1776448935,0.0457395129,-0.1031937078,-0.2820870876,0.200363636,0.4141037166,0.0991049036,0.1959061325,0.2409554869,0.0998634323,0.0113129942,0.1614626795,-0.0441629253,0.02273551,0.2922218144,-0.3577403724,-0.1025523245,0.2709384859,-0.2292333543,-0.1916504502,-0.1559479833,-0.2103705257,0.6568520665,-0.6255700588,0.1918476075,0.086867556,-0.2050861567,0.0055125826,-0.4269950986,-0.201799944,-0.2100611031,-0.4885864258,0.2076575309,0.0475860089,-0.1806461811,0.1380835176,-0.4365613163,-0.2257267684,-0.038511496,-0.0648878813,-0.1392575502,0.5241897702,0.2169024497,0.2822089195,0.1320563108,0.1124191284,0.1560850888,-0.0253071953,-0.0859848559,0.1553043574,-0.3993440568,0.0978845954,0.1229230464,-0.0766257048,-0.2633404434,-0.0058011683,-0.1869954765,0.1047539562,-0.0845100433,0.0537495725,0.4163246155,-0.1387190968,0.2336901426,0.2443135828,0.2504860759,-0.1965702474,-0.0089620063,0.0393141806,0.3112398982,-0.0724596307,-0.1022160202,0.2815064788,0.1592719853,-0.1656410545,-0.2270414829,0.1601279378,-0.0787440389,-0.1728204638,0.1050313339,0.1890412569,0.0321094356,0.7863182425,0.1603609174,0.1528146863,-0.3151907623,-0.098212488,-0.0827553123,-0.0469097458,-0.3591071069,0.0623312108,0.2653268874,0.0757654011,0.0225658584,0.0162762124,-0.3514234722,-0.2347065657,0.1464580297,-0.2979079187,0.1086895764,0.7072980404,-0.0892786458,0.0183770172,-0.0416308604,-0.130457297,0.1025350019,0.2937501669,-0.4762499332,-0.1720766276,-0.2216234058,0.0989384279,0.0158431679,0.3157363534,-0.1458758116,0.2584252059,0.3374360502,-0.3611263633,-0.3022662997,-0.3342594802,-0.3778937757,-0.2802392244,-0.0151137942,0.3497568369,-0.417835474,-0.0482937843,0.0290734842,-0.2109125108,0.282897979,0.1722494215,0.0034635623,0.0058178296,-0.1268529892,0.0859673098,0.0827628747,-0.2245172411,-0.0795051157,0.2871681452,0.0369178914,0.3691134751,-0.0136988107,-0.2268623561,0.3075447679,-0.1145909503,0.268471241,-0.061312288,-0.2099349052,0.0353067517,-0.4293358624,-0.0126746781,-0.0322019532,-0.0557020232,0.3776507974,0.1747308075,-0.1637310386,-0.4752461612,0.3148562908,-0.1022582427,0.3025348783,0.093907252,0.1037314832,0.0834248066,0.1367923766,-0.3164684474,-0.5455599427,0.1920397878,-0.2430796325,0.0453830436,-0.0629600585,0.0095055122,0.0856013522,0.2066118568,-0.2232861519,-0.1861708015,0.04369146,-0.0247203391,-0.291108042,-0.0545375906,-0.2148125619,0.6640715599,0.0832326934,0.127358675,-0.1278247684,0.2169323713,-0.1181343421,-0.4244371653,-0.2777647078,0.2951553762,-0.0546348393,-0.0944967568,-0.1768432409,0.4451701641,0.5167890787,-0.1166738048,-0.0630024597,0.1196777597,0.234720543,-0.158031702,-0.1404241025,0.089490518,0.0731599927,-0.2027816772,0.1714522839,0.4486739933,0.1051602811,0.403865844,0.200966239,0.2442136407,0.3083201647,0.0111940112,-0.0553299002,-0.1513994336,0.1954539865,-0.1391925216,0.2157170922,0.0422817245,-0.1442727745,-0.2772381902,0.1165647656,0.0492503643,-0.1768721491,0.0313884504,0.0579700768,-0.0970046297,0.1442263275,0.0139355492,0.4102200866,0.0368763059,-0.2197754532,0.1526475251,-0.2397209406,-0.0184339844,0.1210476756,-0.0923027992,0.3980694115,0.3248577416,0.0586439744,-0.0086753527,-0.1512736231,-0.3640149236,0.0955984816,0.354624927,-0.4936693609,0.1354499161,-0.2471874207,0.3136082292,0.1267848909,-0.1015333161,-0.3107508421,-0.5806995034,-0.1432543546,0.4852829576,-0.1077934355,0.1337595582,-0.0636859387,0.0561110936,-0.1849563867,0.3307796717,-0.1718563288,-0.3377627432,-0.179838419,-0.0932213962,0.3514887989,-0.2525600791,0.216876477,0.1445072591,-0.26963678,0.2316000164,-0.292008996,0.0788465515,-0.0409869291,0.0614089333,0.1522498131,0.0405817516,0.0468844324,-0.1451683491,0.1635687351,-0.1302802861,-0.2334637046,0.2752299309,-0.1006077975,0.1216905341,-0.2873381376,-0.3229163587,-0.4248741865,-0.1532507092,-0.0967191681,-0.2134873867,0.332819283,0.1174498722,0.0456912667,0.0309305955,-0.0491061993,0.0635498315,-0.0405400284,-0.0448912643,-0.1757173836,0.0618326403,-0.1460140347,-0.0274126492,0.0123623731,0.134441644,0.4099628627,-0.3408573568,0.0580804348,0.0781687945,-0.0750890523,0.2010138631,-0.2007391155,0.3472234905,0.4232900441,0.0538193099,-0.0540015213,-0.188864544,0.1221314818,0.0136502208,0.0058791912,0.1086636633,0.3014460206,0.1025300995,0.54163903,0.3708218634,0.0849515721,0.3666625023,-0.0932016596,0.1334334165,-0.2014301121,-0.3552421331,-0.1956765354,-0.2294409722,-0.010064004,-0.2117567807,-0.1094351262,-0.4185889363,-0.3560622334,0.5542834401,-0.3510013819,-0.2122955024,-0.0121958144,-0.604218781,0.2423480302,-0.0913701355,0.0661948845,-0.0325527936,0.0189963728,-0.2595744729,0.1695326269,0.1611236036,-0.3628127277,-0.2190208733,-0.2547181845,-0.4574095309,0.288962543,0.2186337858,0.659091711,0.0101674469,-0.0755763724,0.0008698013,-0.1559624076,0.9267110825,-0.6447675228,-0.3888559043,0.3332636952,-0.4086160064,-0.2881735265,-0.1551359892,-0.2108736634,0.5418133736,0.3744165897,0.402350843,-0.1822483987,-0.3309328556,0.2035425603,-0.1843895167,-0.1229218692,0.0353675298,-0.3275950551,-0.0117163882,-0.1551185697,0.0192968715,-0.2438773662,0.1271957159,-0.1892327815,-0.07987082,-0.153857857,-0.0616388321,0.2533315122,0.0459085479,0.0074975421,-0.2056071311,0.059068609,0.0173692927,0.1852119714,0.2850435972,0.2807537019,-0.4641651809,-0.150868535,0.1951137632,0.0033082617,0.308809936,0.2973106802,0.25360021,-0.081564419,0.0237846989,0.2183164209,-0.3107680082,-0.0008663741,0.370693028,-0.0225322749,-0.5087131262,-0.1511440277,-0.0834028348,0.4239025116,-0.1408558786,0.5061940551,-0.3755201697,-0.252050668,0.4052420259,0.0552656651,0.7573149204,-0.354886204,-0.016849773,0.1184910312,0.0098484363,0.0970328003,0.1131175831,0.032287553,-0.2598315179,0.0133907152,-0.062314976,0.0515733585,0.3200414181,0.1863402724,-0.1497122794,0.4255023599,-0.003150539,-0.0735292807,-0.0431129336,0.1802263111,0.4461591244,-0.0171469469,0.081188336,-0.0060850736,-0.0310882125,-0.1508318931,0.0155505911,0.0293797683,0.2699966729,-0.1261633188,-0.4167927206,-0.07930585,-0.119897835,0.1074991599,0.1958380044,0.105816938,-0.0161472913,0.1278051585,-0.1582110524,0.0715014935,-0.1213656068,-0.1072911099,0.089856416,0.3467809856,0.2066574544,0.0626638457,0.3774676323,0.1087837741,-0.2720771134,0.0607485622,-0.3036013246,-0.195413962,0.0736664683,0.0597293079,0.0521797501,0.0950557962,0.0682303533,-0.0060456595,-0.1241452992,-0.1683893204,0.0551886484,-0.0621769987,-0.1553435773,0.495839566,-0.0333332531,-0.381401211,-0.0075509883,0.4515570104,0.2331412733,-0.0708283484,0.1695889235,0.0922184363,-0.0275742374,-0.2567103803,-0.1982159019,0.2533377111,-0.1708274335,0.3515672088,0.1018628776,-0.2405954301,0.2189972252,0.0349058472,0.0780903473,0.3748520315,-0.206051901,-0.3172329664,-0.4248143733,0.2890491784,0.0804687366,0.1266635656,-0.139407739,0.1460566521,-0.0606247075,0.2729715705,-0.2356659919,0.1153295413,-0.0951582417,0.2406782806,-0.1023961082,0.0041404767,-0.0888964236,-0.1170448884,0.0980188921,-0.0400041677,-0.2261297405,-0.0314818844,0.0553728975,0.13033849,0.0484629273,-0.1824488044,-0.0334083065,-0.0827377737,-0.1225577295,-0.2159163803,0.1806539893,0.3406330347,-0.2332561463,0.1685999036,0.2338077575,0.2255002856,0.087296173,0.3448824286,-0.2215869427,0.01427281,-0.0002649394,0.4156591594,0.1165883765,-0.1762879491,0.0874333233,0.0882193819,-0.1290540993,-0.0214600172,0.3376171887,-0.2008489817,0.0574210621,0.0191689953,0.406169802,0.4292371869,-0.1080720201,0.1040020213,0.1315716654,0.105818443,-0.0005345822,-0.1020034179,0.5262361765,0.1090767309,0.1049852073,0.3122935295,0.2681319118,-0.1755151898,0.182104975,0.1462449729,-0.1108154356,-0.0302919671,0.2374853194,0.1952918321,0.0944705233,0.2267362773,0.1894975454,-0.3114218712,-0.137083441,0.3520288467,0.0038071831,0.5455647111,-0.0429404937,0.1299677938,-0.1348793358,-0.6490334868,0.0768085569,0.1382827014,-0.5358976126,0.1918130219,-0.0205574855,0.200024724,-0.3446400166,-0.4428057671,-0.1565843225,0.4058251381,-0.2986252308,-0.319827497,-0.2587651312,-0.1629883647,-0.0947454497,0.0425889827,-0.088688381,0.1690731496,0.7728458643,0.0407553501,-0.0316900983,-0.1772802621,-0.3225025535,-0.0538897552,0.3135891855,-0.0265445206,0.4111677408,0.0448730625,0.0383355916,-0.0227233879,0.1630438715,0.4615834057,0.713309586,-0.5018266439,-0.028844675,0.1895126402,-0.1302996129,-0.1272891611,0.4144805968,0.0985815674,0.3577187955,0.2936311066,-0.0437103994,0.0115096318,-0.1780918986,0.3067951202,0.3256412148,0.0407905579,0.3705882132,0.0735705867,-0.0960007533,0.2204148471,0.0571754724,-0.3301747143,-0.3269529343,0.4735249281,-0.4552642405,0.2488342971,0.2249587923,0.055293031,0.0392437391,0.4531626999,0.2978180051,0.1367695481,-0.2886734605,-0.1988475919,-0.4366715252,-0.0210789591,-0.2005471885,0.2436973602,-0.0142664621,0.0408304483,-0.04922143,0.0133763338,-0.0712217242,-0.2305485606,0.384347111,0.3094963729,-0.5917047262,0.1798881143,0.0557931513,-0.0403037928,-0.0081711719,-0.3062467575,0.3048745394,0.0619477294,-0.0930033848,0.1274387985,0.1538005322,0.0787632987,0.2079307735,-0.0278363898,0.2643871307,0.5697335005,-0.0152837159,-0.0272735599,-0.2518393695,0.0842742845,-0.0754962191,-0.0122487498,0.0922462717,0.419074297,-0.2056139559,-0.0286549963,0.0471718498,0.1599729359,0.0041817673,0.2315075397,-0.4313249886,0.2113007158,0.1868189722,0.0473328903,-0.1670436561,0.372158736,0.1542606801,0.2845135927,-0.2827651799,-0.185196504,0.2704379559,-0.0264111105,-0.1018258929,-0.6085389256,0.2107916921,-0.2392189503,0.1292657107,-0.5197768807,-0.293543607,0.2013055086,-0.2068508565,-0.5606351495,0.2407480925,-0.1010124981,-0.1556028724,-0.0872276723,-0.2131017894,0.1247436181,-0.1182470843,0.1601723135,-0.0955340192]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Hmmm I pulled the latest commit, `b93c5517f70a480533a44e0c42638392fd53d90`, and I'm still seeing both the hanging and the key error. ","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":18,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nHmmm I pulled the latest commit, `b93c5517f70a480533a44e0c42638392fd53d90`, and I'm still seeing both the hanging and the key error. ","embeddings":[-0.3769075871,-0.1904281527,-0.0374410823,0.1203723177,0.0553165115,-0.148819223,0.2511082292,0.3369261324,0.1473491788,0.1732964069,0.1061456203,0.4089758694,-0.4025794268,0.3307685554,-0.3633014262,0.0771415085,0.1139564142,-0.0854947194,-0.2521774173,0.254539609,-0.2573972642,0.3191765845,-0.3519818783,0.1498742551,-0.4931677878,0.0434545204,-0.0665596351,0.2185114771,0.0069515603,-0.5854371786,0.3268186152,0.3247092962,-0.0737521872,0.417165339,-0.0001270814,0.0950679332,0.3749818802,-0.056998264,-0.0621224381,-0.3229508698,-0.2003949583,-0.1734237373,0.3260558546,0.1571721435,0.2158144265,0.0487517938,-0.1209588051,-0.2518004179,0.2916254699,0.3332442045,0.0684020147,0.2374105155,0.2859758437,0.0629966259,-0.0574421883,0.1290603876,-0.0421773158,0.0746058747,0.390198797,-0.3858205974,-0.1070063636,0.2573369741,-0.2188557684,-0.2458664328,-0.1455921084,-0.2127299011,0.703522861,-0.6686064601,0.2560444772,0.0739463717,-0.1959520876,0.0557752773,-0.4045495987,-0.221463114,-0.1872924566,-0.3897366822,0.2885859311,-0.03843702,-0.2220059484,0.1508978456,-0.4666244984,-0.2990283072,-0.0254021641,-0.0910085812,-0.1253380924,0.5177974701,0.2228846252,0.3107624948,0.2196397185,0.1078673303,0.126888603,0.0262998249,-0.0399366096,0.146058172,-0.3875403106,0.1230902374,0.1270538718,-0.0346093364,-0.2650232315,-0.0606251359,-0.3166194856,0.1118736267,-0.0469082557,0.0298335701,0.4266043305,-0.120484747,0.1959402114,0.2868335843,0.2997545004,-0.1972702444,-0.0210937802,0.0749311671,0.3320276141,-0.0966067016,-0.0451822467,0.2595972419,0.1397739798,-0.2171106935,-0.1264463365,0.2199632376,-0.0300098434,-0.1575807631,0.0745053515,0.21255292,0.0536563322,0.7686829567,0.1733046919,0.1522127092,-0.3153272569,0.0211220104,-0.0567554981,-0.0981142521,-0.3359582722,0.0658256412,0.2510722876,0.0283318553,-0.0049024536,0.0512576886,-0.2201678306,-0.2804396451,0.0260843057,-0.2334932834,0.0855752304,0.7288818955,-0.131166622,0.0851642415,-0.0606707446,-0.239374429,0.0579013377,0.2445372045,-0.4499720633,-0.1594576985,-0.2083008885,0.0692318082,0.0207371358,0.2660789192,-0.0923078805,0.2476323247,0.3904178739,-0.3749101162,-0.2846361399,-0.3598318696,-0.3236688972,-0.2998531163,-0.0062102485,0.2738519609,-0.4148652554,-0.0681553185,0.038125366,-0.1951261014,0.2469145954,0.1701105088,0.0047404841,-0.0040635625,-0.1401057988,0.0848202184,0.0337081291,-0.2646328807,-0.1709015667,0.3487452865,0.0507284254,0.3495721221,0.0308210663,-0.2441286892,0.3451981544,-0.1635714918,0.2934062183,-0.1475645751,-0.1938306093,-0.0110447202,-0.3982699215,-0.0040063383,-0.1653377414,-0.0277196337,0.3931813538,0.160329029,-0.1570053548,-0.4206725359,0.3244021833,-0.0621509142,0.2917777598,0.0633131415,0.0832027793,0.0400557257,0.1159481853,-0.2447219044,-0.588796556,0.248418957,-0.2199754417,0.0512509719,-0.1405694634,0.0062952228,0.0613510758,0.2660131454,-0.2182736546,-0.1511555165,0.0010990939,-0.1101252288,-0.3123201132,-0.0099403458,-0.2225744724,0.6338652372,0.1779185534,0.1317066103,-0.0696569085,0.1322140545,-0.2103152126,-0.4017416537,-0.2050703317,0.2375408411,-0.0374558829,-0.0920631513,-0.1554378867,0.4663217962,0.4083088934,-0.0508940071,-0.0710795,0.146308437,0.2993747294,-0.0992844403,-0.1243374795,0.019627763,-0.048117552,-0.2307187468,0.1942163408,0.4785977304,0.1279977262,0.4742359221,0.1451377273,0.2324370891,0.2692558467,0.0385850593,-0.0693080127,-0.1486952603,0.266959995,-0.2604362071,0.2012297511,-0.0160269458,-0.1401325911,-0.1557421386,0.1133951172,0.0463226177,-0.2064407617,0.0521031469,0.1835554987,-0.1177733466,0.0970706642,0.0266081654,0.3838664293,-0.0120633869,-0.1649728864,0.1629096419,-0.3187038898,-0.0670593902,0.1021071151,-0.0751307309,0.3526102602,0.3592049778,0.0093879979,-0.0652726889,-0.1116801053,-0.2466428578,0.0869158506,0.4097516239,-0.4359943271,0.1810720861,-0.2512144148,0.3177016377,0.1722839028,-0.1079267114,-0.3008707464,-0.6829925776,-0.0988099426,0.4672749639,-0.0280366223,0.1316036433,-0.0585720874,0.0979007483,-0.2034888715,0.3690121472,-0.2601797879,-0.3610712588,-0.1307526529,-0.1165758669,0.3876147866,-0.3217427135,0.1939658374,0.1785333306,-0.2623199821,0.2360372096,-0.3630244434,0.0889376029,-0.1330858171,0.0983866975,0.203526929,0.0626997128,0.0286587421,-0.2158976346,0.1384954005,-0.2482049912,-0.210938409,0.2669925392,-0.1468042582,0.0815706849,-0.2717660964,-0.2617188096,-0.4254592061,-0.1215943471,-0.0129349232,-0.2455711365,0.288983047,0.0955772623,0.0731859878,0.0277666971,-0.0155605162,0.0322793983,-0.0874169692,-0.0393875055,-0.1897678971,0.0770542249,-0.0941975787,-0.0101405429,0.0946576521,0.1635846496,0.2471490353,-0.4137736261,0.0598174147,0.0181819126,0.0326778889,0.2200659364,-0.2045188993,0.2996485233,0.4151645601,0.0442237146,-0.0472322404,-0.1688159555,0.1457171887,0.0228527114,0.0040896293,0.0625493452,0.3118416667,0.138562724,0.5838057995,0.3471314609,0.1333993822,0.2966781557,-0.010763159,0.1625948995,-0.2375050187,-0.3424849808,-0.2309472412,-0.2596518993,-0.0357606858,-0.20820871,-0.0774802044,-0.4753356278,-0.342063576,0.5736966133,-0.3473555446,-0.2954928577,-0.0182284769,-0.4701861143,0.1867370754,-0.0581230745,0.0410769619,-0.0324110091,0.0159690604,-0.2220127583,0.2033531666,0.2477431595,-0.3888175488,-0.2757533789,-0.3214167356,-0.5434209108,0.3417310119,0.1708779782,0.6498444676,-0.0604271702,-0.0693535581,0.0671446174,-0.2798388302,0.9282134771,-0.6133777499,-0.4135127366,0.3380135894,-0.3858445585,-0.3094117343,-0.1624200344,-0.1247925907,0.6172782183,0.427493453,0.4958391786,-0.1949275136,-0.3422038257,0.1864223927,-0.2154033333,-0.1390995234,-0.0089291399,-0.357275635,-0.0246107522,-0.1836156398,0.0630400851,-0.2271406353,0.0721597299,-0.233251214,-0.0317146704,-0.0655671507,-0.1067605466,0.1498785019,0.0699225962,-0.0642870888,-0.1818167716,0.0786016285,0.1185861379,0.0782495886,0.1344025284,0.1924815476,-0.4369197786,-0.0408000611,0.1790243536,0.0244012307,0.2706047893,0.3768011332,0.2354343534,-0.0519160405,0.1326311231,0.2377825081,-0.3062327802,-0.0404676497,0.364269048,-0.0291333739,-0.5488160253,-0.1445972025,-0.0697215199,0.4107923806,-0.2437293977,0.4938353598,-0.3520704508,-0.2111810893,0.4083628356,0.0432296433,0.850925982,-0.4368037581,-0.0274585038,0.0738661438,0.0087487334,0.0746612772,0.0460166149,-0.0211691502,-0.2881321609,-0.0226640087,-0.0523689128,0.0165420063,0.2664729953,0.062055923,-0.1888653934,0.390216887,-0.0525774844,-0.0227682162,-0.0531621762,0.1566170454,0.3568082154,0.0011243765,0.120503068,-0.0319023691,0.0010160366,-0.1539350003,-0.0333331376,0.0376936086,0.2755374312,-0.1731656939,-0.462462455,-0.0863377154,-0.0337372422,0.1094930843,0.1800212562,0.0868441537,-0.0420005582,0.1443866938,-0.0876284391,0.1290786117,-0.1049812362,-0.0392893031,0.0913171396,0.3167607784,0.200742498,0.052756764,0.2512250841,0.0953507051,-0.27293697,0.0647852644,-0.2976250947,-0.2978365719,-0.0349186696,0.0663810223,0.0433263555,0.138147071,0.1084843352,0.0053986143,-0.1620916873,-0.2438485473,0.025362229,-0.0496192724,-0.0770037323,0.5529299378,-0.107359767,-0.3646469414,0.0220889151,0.4827080071,0.2201131135,-0.0684509128,0.179539457,0.170926705,-0.0470669195,-0.2419931442,-0.2201522887,0.1828523725,-0.1992957592,0.3358785808,0.0801732093,-0.2542218268,0.250511378,0.0631614327,0.1032474265,0.2757013738,-0.2074915618,-0.3126967549,-0.472281158,0.2779865861,0.0353676267,0.1646442711,-0.1440231055,0.0844866708,-0.0956170335,0.3184504807,-0.2017914653,0.110980168,-0.1178984344,0.1734066904,0.0233175643,-0.0683205128,-0.0564722531,-0.099685438,0.0616014786,0.0252546687,-0.2078841776,0.001862557,0.0787153319,0.1410406381,0.01534006,-0.1227155477,-0.0183771774,-0.0836907402,-0.0561203361,-0.1496049464,0.2362178266,0.2881514132,-0.1598276496,0.1271132082,0.1638927013,0.2709268332,0.1845809817,0.385547787,-0.1361583918,0.0546355359,0.0643814951,0.3762483597,0.1178337559,-0.1927624792,0.0755717978,0.1425991505,-0.1473907977,-0.0080327662,0.391315639,-0.1805911213,0.068083398,0.007701972,0.4710888267,0.4523408115,-0.0779246986,0.1208757237,0.1299077272,0.0696717203,0.0346482545,-0.1935385317,0.4906766713,0.1123969555,0.0491277054,0.3103135526,0.2718243599,-0.2247924954,0.2421731651,0.2108755857,-0.0402917005,-0.0609161444,0.1589468718,0.2681147754,0.1229021326,0.1887983829,0.2788735032,-0.2783479989,-0.1302317232,0.2561229765,-0.0551641174,0.4227167666,-0.0360201187,0.191689983,-0.0057650399,-0.5985743403,0.0967568532,0.1316299886,-0.5105832219,0.1886382699,0.0281333625,0.2563152313,-0.3215015829,-0.4784180522,-0.2237994224,0.4910321236,-0.2867060602,-0.361181885,-0.1705144793,-0.1840226799,-0.0509405695,0.0617613718,-0.1212099642,0.1888549477,0.7641544342,0.0984367803,0.0199344549,-0.2188363373,-0.356490612,-0.1301346719,0.3811824024,-0.0232565235,0.3999077082,0.0608853623,0.0769782141,-0.0178106409,0.1241748929,0.5050506592,0.6178365946,-0.5770907998,0.0544343032,0.1402111501,-0.0746737123,-0.1047943607,0.441591382,0.1400591433,0.3452727199,0.2643514872,-0.0792869702,0.0287589077,-0.1752514243,0.3240355551,0.4146078229,0.1198318973,0.3768998981,0.0824237987,-0.1472591013,0.3186635971,0.1207054257,-0.3364008367,-0.3069581985,0.3774551451,-0.4733802676,0.2193963081,0.2029016316,0.0530630387,0.0539451763,0.5389310122,0.3228668571,0.1460399777,-0.2956557572,-0.1743111163,-0.4652699232,-0.1036809087,-0.1289964318,0.2323549092,-0.0010729451,0.0631360486,0.0173696224,0.0486703403,-0.0112880338,-0.3573030531,0.3776764274,0.3754167259,-0.4881598949,0.2943690121,0.161005646,-0.0582351945,-0.0448180363,-0.2514109313,0.365336746,0.0154904779,-0.1032703891,0.0839191675,0.1313344985,0.063970916,0.1456246674,0.0059961588,0.2952332199,0.5333591104,-0.0039059427,-0.0923863649,-0.2899853289,-0.0016283304,-0.0063386052,-0.0786268786,0.0183875766,0.4141324759,-0.1897063851,-0.0324868076,0.0221517943,0.0365645513,0.0039271177,0.2086862028,-0.4341152012,0.2006957233,0.1827541143,0.0372736752,-0.1726938188,0.4718481302,0.2118583918,0.3240719736,-0.2758527696,-0.1918581128,0.3087020218,-0.0378631875,-0.1339868903,-0.4886540771,0.2032844275,-0.2365761548,0.114209421,-0.5074614882,-0.3361953497,0.1557460874,-0.1966831386,-0.6136543155,0.2853872776,-0.061989937,-0.2154586315,-0.0629374757,-0.2218505591,0.1091929302,-0.1340529025,0.1599437147,-0.0997045413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Hi @timothyjlaurent \r\n\r\nThe hanging fix just got merged, that why you still had it.\r\n\r\nFor the key error it's possible that the code you ran reused cached datasets from where the KeyError bug was still there.\r\nCould you try to clear your cache or make sure that it doesn't reuse cached data with `.map(..., load_from_cache=False)` ?\r\nLet me know if it it helps","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":63,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nHi @timothyjlaurent \r\n\r\nThe hanging fix just got merged, that why you still had it.\r\n\r\nFor the key error it's possible that the code you ran reused cached datasets from where the KeyError bug was still there.\r\nCould you try to clear your cache or make sure that it doesn't reuse cached data with `.map(..., load_from_cache=False)` ?\r\nLet me know if it it helps","embeddings":[-0.3729741573,-0.1859525442,-0.0339068882,0.1292066127,0.0980354324,-0.1325110644,0.2636093199,0.4240571856,0.1391321868,0.1339375675,0.0465279371,0.466599226,-0.3696385622,0.4466212392,-0.3270716071,-0.0237082001,0.0842828006,-0.086714305,-0.2671528161,0.2537027597,-0.333704412,0.2835071683,-0.3157280684,0.2287008464,-0.4680302739,-0.0196886603,-0.1571886837,0.2040823102,-0.0384008214,-0.5968233943,0.3087911606,0.2592989504,-0.0386928655,0.4125376344,-0.0001159086,0.0726363212,0.2915791273,-0.0694602653,-0.1238340586,-0.4056953788,-0.1630977094,-0.1920109689,0.2794992626,0.0924808457,0.1524776667,-0.0606593043,-0.137013346,-0.1714393348,0.3692105412,0.2973997593,0.1916394383,0.1836921275,0.3016476035,0.0850571617,-0.0235235803,-0.0432401225,-0.0450257324,-0.044133354,0.3562731445,-0.383523047,-0.1636612862,0.3938462138,-0.1591003686,-0.2558552027,0.0481436886,-0.2032792419,0.5871521831,-0.6714693308,0.1936377436,0.0861370713,-0.1640220582,0.0897193626,-0.4611234069,-0.2372850031,-0.1507684141,-0.415684998,0.2953735888,-0.0046022236,-0.2389934063,0.1260260642,-0.423813194,-0.2955448627,-0.0063166488,-0.1394233555,-0.0698230341,0.5199891925,0.2127799541,0.2648281753,0.238740027,0.0557517782,0.1047797352,-0.0716203079,-0.1317555606,0.2191713303,-0.4095770121,0.0330452286,0.1441633403,0.0376284756,-0.1656429321,-0.159757629,-0.2449796945,0.0867290273,0.0025472487,-0.0358658582,0.409196496,0.0254016053,0.2701698244,0.2604080439,0.2943361104,-0.1905028373,-0.0741190836,0.0486717373,0.2814122438,-0.1145023927,0.0346962586,0.3480422795,0.1715756059,-0.1595720202,-0.0572375171,0.0996338278,-0.0458171517,-0.1653964669,0.0432068668,0.2337245494,0.0642119199,0.6473721862,0.1239025667,0.1313348711,-0.232063666,-0.049682036,-0.1114419475,0.0055615781,-0.3369040489,0.0278137084,0.2164690197,-0.1401575804,0.0144445766,0.0348331258,-0.2584845424,-0.2159378082,0.0866359621,-0.1109285206,0.1053563952,0.6346384883,-0.0904199854,0.0813511536,-0.0073841261,-0.2797871828,0.0319329575,0.2214627564,-0.5432989001,-0.2066190541,-0.0062801335,0.1853593141,0.0336939134,0.1731328368,-0.1481188685,0.2331714928,0.3177970052,-0.2863822281,-0.2881161869,-0.286108166,-0.3216991425,-0.3718682528,-0.0147515312,0.2520573437,-0.4537152946,-0.0823263898,0.0158720501,-0.1219125167,0.1184691116,0.1592020839,-0.1201319024,0.1245121881,-0.1835378408,0.1869907081,0.0609243698,-0.2588416636,-0.2698888481,0.3641799092,-0.0383909978,0.2907693088,0.0354466215,-0.196684286,0.3934170008,-0.1237005889,0.3538429439,-0.0550936759,-0.1739271581,0.0194118936,-0.3822776973,-0.0390148051,-0.1381125599,-0.1211566627,0.4054738581,0.0988969877,-0.2085455209,-0.3918999434,0.2793352306,-0.1241683364,0.3125305474,0.0472516939,0.1341271698,-0.040825516,0.2059615701,-0.1856366992,-0.4679727852,0.322678715,-0.2243601233,0.0517389104,-0.1350090802,-0.0220244285,-0.0217192266,0.1625897139,-0.3280446529,-0.2965452671,0.146207422,-0.0975862667,-0.1941841394,-0.0213772766,-0.2148850411,0.515776217,0.3289866447,0.1015793607,-0.141281262,0.130711183,-0.3029308319,-0.3651212156,-0.1901479512,0.151596874,-0.0101102274,-0.0993063301,-0.2007431835,0.4787659049,0.385365814,-0.0566653013,-0.0514839068,0.2301508933,0.262424916,-0.0285461415,-0.1271883398,0.105215624,0.0398713276,-0.1369030923,0.1302661598,0.6240000129,0.1339046657,0.4230161309,0.1884403676,0.2562130988,0.2815244198,0.0092139337,-0.0580085665,-0.0960733443,0.1638165563,-0.3738693893,0.0694385767,-0.0546686575,-0.0387199372,-0.1118883416,0.2121147513,0.0672766641,-0.1295357943,0.0653589368,-0.0247472003,-0.0656456798,0.0686134025,0.1543727666,0.3296006024,0.0632075667,-0.2049697936,0.1575351954,-0.1924852729,-0.0033683754,0.2201360911,-0.1017463058,0.2729797661,0.3356056809,0.0168670844,-0.0137437163,-0.14149113,-0.1968515366,0.0353916101,0.4144649804,-0.4009288847,0.1306913942,-0.3032637239,0.1425554901,0.2121398449,0.0424723625,-0.1998408735,-0.6393460631,0.0158740748,0.3853761256,-0.0572635606,0.1523911953,0.0208081305,0.0435619354,-0.1224137396,0.3570632637,-0.1496761143,-0.4756741822,-0.1689337641,0.0046429839,0.3523488045,-0.339055866,0.2837488055,0.1046037599,-0.2289168537,0.2102584392,-0.4162888229,0.0956430212,-0.1570355743,0.0642079562,0.1687162966,0.1463493705,-0.015052679,-0.2420867682,0.1787250042,-0.2942126989,-0.3931431472,0.3215646446,-0.1034122184,0.1561695039,-0.2509377301,-0.4179036319,-0.3237200081,-0.2003942877,0.0025403553,-0.2512457967,0.252478987,0.168238759,0.0705053732,0.026711937,-0.0970103964,0.0801035315,-0.0571684986,0.076510556,-0.1497867405,0.0077674859,-0.2464174777,0.0273185894,-0.027702257,0.130397737,0.1635608673,-0.3969526589,-0.0190880354,0.0411077589,0.1432687789,0.268086195,-0.1667965055,0.3800825775,0.364736259,-0.0318775401,-0.0121165765,-0.0928804353,0.0071055205,0.073589839,0.0150000295,0.0612242334,0.3435931206,0.1424987018,0.685436964,0.2752219737,0.1665730923,0.2465547919,-0.0247274749,0.1074370742,-0.3065450191,-0.3919324279,-0.2665607631,-0.3089667559,0.0054316437,-0.1595831513,-0.0569786392,-0.4351715147,-0.2615182102,0.3561646342,-0.3579081595,-0.3693565428,0.0764668435,-0.4184983075,0.1878625154,0.0158674512,0.0569497272,-0.0557554215,0.0530093648,-0.1381237656,0.2112246901,0.2029326856,-0.4287493527,-0.2992621958,-0.3367242217,-0.4790402353,0.3524399102,0.1396892071,0.4911383688,0.0711003914,-0.1520290077,0.0255536102,-0.2880279422,0.8684406877,-0.6372084618,-0.4212183654,0.2932739556,-0.4309197068,-0.3340891004,-0.1677621454,-0.2115839571,0.6215441823,0.40474087,0.6619139314,-0.1441090405,-0.3067403436,0.0630455017,-0.0505907089,-0.1604545712,0.0117189288,-0.2843121588,-0.0188128799,-0.3132395148,0.0902739689,-0.2285679728,0.1387687624,-0.2473486811,-0.0448911302,-0.0667257234,-0.1827619821,0.0605005696,0.117720969,0.0207162946,-0.1789888144,0.1849313825,0.0713651851,0.1505000442,0.1448423117,0.2358928472,-0.387240231,-0.0403512828,0.1115831733,-0.0157178324,0.1632239223,0.3018364012,0.1191112176,0.0477821007,0.1097642854,0.1298591346,-0.2769138813,-0.1139095649,0.3287644088,-0.0388904437,-0.4032011926,-0.2041196227,-0.0893792138,0.3325070739,-0.2921119928,0.4312385917,-0.2718985975,-0.1908378601,0.488658309,-0.0745832324,0.7769594789,-0.3349074423,-0.0066604181,0.060916584,0.0300632343,0.1367226839,0.1239457205,0.0436868183,-0.3786579967,-0.0925659165,-0.0124078076,0.0585334823,0.2362622321,0.0588806346,-0.2678055167,0.3634712994,-0.0307003967,-0.0799854696,-0.1024518758,0.2062174529,0.2626480162,-0.03354045,0.1019342095,0.0622216202,0.0284089297,-0.0071151303,-0.0046363897,-0.0091663003,0.2356943339,-0.3067147434,-0.4391646683,-0.124404341,-0.0712271407,0.1604688913,0.1113618687,0.0259936471,-0.0110035427,0.1811235249,-0.0262442622,0.191625461,-0.1083165705,-0.0665973499,0.0885275081,0.2978014946,0.1572382748,0.0032185824,0.2674489915,0.009874803,-0.2708142996,0.0919783786,-0.2495938092,-0.2325697392,0.0245080087,0.0437904559,-0.0755675137,0.1149438918,0.1390471756,-0.0572589859,-0.2020321041,-0.2990484238,0.1250751615,-0.0670376346,-0.1480725557,0.5049770474,0.0004377814,-0.3364857137,0.060395062,0.5290411115,0.3188914955,-0.0273062177,0.3020719886,0.1057575792,-0.0224169847,-0.3198133409,-0.101100646,0.3219786882,-0.1888059378,0.2999127507,0.0288372003,-0.1429735124,0.2914672196,0.1589841247,0.2008523494,0.3241571188,-0.2165279537,-0.2855886519,-0.4925097823,0.3401563764,0.0692534298,0.1096189767,-0.1709636301,0.0120268539,-0.0885958895,0.3379904628,-0.3164950907,0.0990837216,-0.1227278858,0.176485613,0.0475692749,0.0167665444,-0.0642429963,-0.0726292878,0.1442051232,0.0324135274,-0.2023002356,-0.1341395825,0.056739822,0.0862880871,0.0451294146,-0.0931859165,-0.0640827417,-0.1119966358,-0.0435155146,-0.1591751426,0.2522532344,0.2442680895,-0.1731223017,0.1908839196,0.1939116269,0.1710810512,0.1297677457,0.3485626578,-0.2650255859,0.074702777,0.0642918423,0.344096005,0.1167639345,-0.1704270244,-0.0454205163,0.0493137278,-0.2381974161,0.0777189881,0.3738659322,-0.1637310386,-0.0200512428,0.0756099597,0.4149127007,0.3762215376,-0.0997755006,0.1957228184,0.1410349309,0.1938936114,-0.0605852902,-0.2219108492,0.3633992076,-0.0010872269,0.1359811127,0.2581396997,0.2731424868,-0.1010004506,0.1861305088,0.1735342294,-0.0997623652,0.0171632804,0.1727562696,0.3383671641,0.0944933742,0.112139307,0.2775925696,-0.3374082744,-0.1136412621,0.2190269977,-0.1272085309,0.418017,0.0461264886,0.1783936024,0.0522115938,-0.5796561241,0.2260048985,0.0386147238,-0.4828644693,0.0830069855,0.0876456052,0.1291488558,-0.2772055268,-0.4099140167,-0.1989888996,0.444693625,-0.3052487075,-0.3082734048,-0.2460544407,-0.2431775331,-0.0763234869,0.036475569,-0.1299948543,0.1892970055,0.6870554686,0.0264351461,-0.038080208,-0.1530599147,-0.2805673182,-0.0369321667,0.4131477773,-0.0874880925,0.3061192632,0.1116387472,0.0818897709,0.0151324989,0.1614595354,0.5225877762,0.6355858445,-0.5816590786,0.0809556693,0.1785973907,-0.1300255507,-0.1261998713,0.3898410797,0.1103923023,0.3445353508,0.3547881544,0.0350760706,-0.0407640859,-0.074059464,0.2384990901,0.392190665,0.0455203839,0.2923431396,0.1365070641,-0.1297212392,0.280379355,0.1797734648,-0.331474185,-0.3247245252,0.4075752199,-0.3387220204,0.2173644453,0.0483184084,0.116208449,0.0322411247,0.5574448109,0.3594413996,0.0940397829,-0.3318679929,-0.086582914,-0.5268977284,-0.0841363519,-0.2359142452,0.1518827528,-0.0689223185,0.1092913821,-0.0044610971,0.1244381219,-0.0152094327,-0.384454906,0.305098623,0.4490070641,-0.4907383919,0.181387946,0.306548804,-0.0227140784,-0.0300960913,-0.3004015982,0.3344733715,0.0385636874,0.0250894167,0.1233910546,0.0685886666,0.1777249426,0.0547109097,0.0470291935,0.2941527665,0.5331443548,-0.0617081076,-0.061541684,-0.2617748678,-0.0059646275,-0.117255792,-0.0979967043,0.1279917657,0.4231697023,-0.0840478614,-0.1946541518,-0.0175151788,0.1063797772,-0.0377557278,0.1182322428,-0.404391557,0.2538506985,0.2120008767,0.0213682782,-0.161151126,0.4416864812,0.1101305261,0.1553229243,-0.2664441466,-0.2475418448,0.2882629037,-0.0436089933,-0.1493017375,-0.4830526412,0.1681104898,-0.2236526161,0.1611625552,-0.4133827388,-0.289352566,0.231583342,-0.1507377476,-0.515560925,0.2117448598,-0.0465030782,-0.1064363346,-0.0427053869,-0.2824112773,0.0595694482,-0.1379961222,0.1203345805,-0.0893135667]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Hi @lhoestq , \r\n\r\nThanks for letting me know about the update.\r\n\r\nSo I don't think it's the caching - because hashing mechanism isn't stable for me -- but that's a different issue. In any case I `rm -rf ~\/.cache\/huggingface` to make a clean slate.\r\n\r\nI synced with master and I see the key error has gone away, I tried with and without the `TOKENIZERS_PARALLELISM` variable set and see the log line for setting the value false before the map.\r\n\r\nNow I'm seeing an issue with `.train_test_split()` on datasets that are the product of a multiprocess map.\r\n\r\nHere is the stack trace\r\n\r\n```\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 451, in load_prodigy_arrow_datasets\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/src\/datasets\/src\/datasets\/arrow_dataset.py\", line 168, in wrapper\r\n dataset.set_format(**new_format)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/src\/datasets\/src\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/src\/datasets\/src\/datasets\/arrow_dataset.py\", line 794, in set_format\r\n list(filter(lambda col: col not in self._data.column_names, columns)), self._data.column_names\r\nValueError: Columns ['train', 'test'] not in the dataset. Current columns in the dataset: ['_input_hash', '_task_hash', '_view_id', 'answer', 'encoding__ids', 'encoding__offsets', 'encoding__overflowing', 'encoding__tokens', 'encoding__words', 'ner_ids', 'ner_labels', 'relations', 'spans', 'text', 'tokens']\r\n```\r\n\r\n\r\n","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":174,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nHi @lhoestq , \r\n\r\nThanks for letting me know about the update.\r\n\r\nSo I don't think it's the caching - because hashing mechanism isn't stable for me -- but that's a different issue. In any case I `rm -rf ~\/.cache\/huggingface` to make a clean slate.\r\n\r\nI synced with master and I see the key error has gone away, I tried with and without the `TOKENIZERS_PARALLELISM` variable set and see the log line for setting the value false before the map.\r\n\r\nNow I'm seeing an issue with `.train_test_split()` on datasets that are the product of a multiprocess map.\r\n\r\nHere is the stack trace\r\n\r\n```\r\n File \"\/Users\/timothy.laurent\/src\/inv-text2struct\/text2struct\/model\/dataset.py\", line 451, in load_prodigy_arrow_datasets\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/src\/datasets\/src\/datasets\/arrow_dataset.py\", line 168, in wrapper\r\n dataset.set_format(**new_format)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/src\/datasets\/src\/datasets\/fingerprint.py\", line 163, in wrapper\r\n out = func(self, *args, **kwargs)\r\n File \"\/Users\/timothy.laurent\/.virtualenvs\/inv-text2struct\/src\/datasets\/src\/datasets\/arrow_dataset.py\", line 794, in set_format\r\n list(filter(lambda col: col not in self._data.column_names, columns)), self._data.column_names\r\nValueError: Columns ['train', 'test'] not in the dataset. Current columns in the dataset: ['_input_hash', '_task_hash', '_view_id', 'answer', 'encoding__ids', 'encoding__offsets', 'encoding__overflowing', 'encoding__tokens', 'encoding__words', 'ner_ids', 'ner_labels', 'relations', 'spans', 'text', 'tokens']\r\n```\r\n\r\n\r\n","embeddings":[-0.3455711901,-0.124476321,-0.0099463714,0.2436582297,0.132649079,-0.1085882708,0.2091186345,0.3842711151,0.1275978088,0.0554183125,0.0650868118,0.2885191441,-0.4160088003,0.3402267098,-0.2788427174,0.064945899,0.1176876798,-0.1377786398,-0.3737223148,0.2305669636,-0.2718407214,0.3353238404,-0.2338251621,0.1210066006,-0.5046049953,0.0449887589,-0.1111245677,0.3717213869,0.068482928,-0.5890556574,0.3367897868,0.2819594145,-0.0480297953,0.3480781317,-0.0001237825,0.1063452214,0.3847542405,-0.1089715138,-0.0622288585,-0.3263475299,-0.1737760901,-0.1924261302,0.2838899493,0.0872735232,0.1946082562,0.0350470878,-0.1431172341,-0.2287731469,0.3298036158,0.3584726453,0.1089558676,0.2274534702,0.2610171735,0.1091630161,0.0079336753,0.0144447973,0.0120129287,0.0105952481,0.2656100094,-0.3017512858,-0.1495733559,0.3722892404,-0.255100131,-0.2173593491,-0.0058807489,-0.2193851024,0.6108903885,-0.5381208062,0.2770029306,0.0419181697,-0.2300306261,0.0360986963,-0.4760312736,-0.2049938142,-0.2889748514,-0.5068151355,0.3482626975,-0.0966814458,-0.1204990447,0.1607678086,-0.5080752373,-0.2501933575,0.0702679902,-0.1309080273,-0.0440390669,0.4516026974,0.2043748349,0.2419659644,0.2371707857,0.1206098124,0.0828298554,0.0368181914,-0.1755397916,0.1539328247,-0.4410711527,0.0727010891,0.1501148492,-0.0112307305,-0.2254723161,0.0307366364,-0.1942606717,0.0992030576,-0.0916412547,-0.0203569066,0.3888459802,0.0562038198,0.1859543175,0.1347609758,0.2832926512,-0.1339153051,0.0400829352,0.0852165148,0.3456960917,-0.0811674669,-0.0124403508,0.4133488834,0.112997517,-0.2294027358,-0.0599842444,0.1502421945,-0.0300123934,-0.2148933411,0.0187901184,0.1588859707,-0.0029365425,0.7559267879,0.1216320768,0.1799938977,-0.2985371649,-0.1033544913,-0.0879793242,-0.0783394799,-0.4462857842,0.1121379957,0.2708312273,-0.1769164801,0.0029232351,-0.0293803643,-0.3343157172,-0.2672925591,-0.0057176384,-0.180957824,0.1032915786,0.6875529289,-0.1682294309,0.0362343602,-0.0051768241,-0.0747529343,0.0804438144,0.2214311063,-0.5367627144,-0.2646110058,-0.2048505247,0.0789481029,0.0069822115,0.3434107304,-0.0932201669,0.2554689646,0.4977132976,-0.3562063873,-0.2382134497,-0.257095933,-0.4370432198,-0.3161766231,0.0009105206,0.3135167956,-0.3315528929,-0.1214467958,0.0876626521,-0.1581772268,0.179208383,0.2633360922,0.0320779644,-0.0481657609,-0.1246599108,0.0071643279,0.0008243211,-0.2042131126,-0.2987396121,0.1712997258,0.0336488411,0.3593222499,0.0736718997,-0.2474052161,0.3209016025,-0.184307009,0.3235274553,-0.1109328866,-0.1499824524,0.0182688311,-0.4165968001,-0.0977840871,-0.1438877136,-0.1557713598,0.4019341171,0.2475280762,-0.0925158262,-0.4669560492,0.2758161724,-0.1199947596,0.3561442792,0.1221896559,0.0755639821,0.070230417,0.1730285585,-0.1853505522,-0.6047538519,0.3041417301,-0.2501339614,0.0431420989,-0.1477176547,-0.0157312006,0.1335422844,0.0891357735,-0.14906919,-0.2978046238,0.0218345337,-0.0204796121,-0.2045932263,-0.0690909475,-0.2441110313,0.7258079648,0.1483658403,0.1382077187,-0.1974518001,0.1050306037,-0.1945034415,-0.3886584938,-0.2837040126,0.2788521647,0.0128649343,-0.1279384792,-0.1980294138,0.3907418847,0.4001692235,-0.0490544103,-0.0479035042,0.299172312,0.2560943067,-0.1387450397,-0.0958262756,0.0244669486,-0.0276676193,-0.1636607051,0.2050244659,0.4712374806,0.0942035839,0.4115881324,0.1363151371,0.2084512264,0.3067332208,0.0051696063,0.0097786048,-0.1651752889,0.1741481572,-0.1718353182,0.2099791318,0.0183888171,0.0095019555,-0.1484384835,0.1744221747,0.0208028071,-0.1467273831,0.080833979,0.0518874079,-0.1830231547,0.1379190236,0.0632489324,0.3419297934,-0.0183102023,-0.1588670611,0.2417920679,-0.1436592042,0.0302835852,0.1401726007,-0.1044881418,0.2664687335,0.2962377071,0.0669785216,0.0796569288,-0.183586508,-0.1547777653,0.087174125,0.3825924993,-0.4576377869,0.1936836094,-0.2836962044,0.2762703001,0.2124966979,-0.0819847286,-0.3219939768,-0.5981811881,-0.0910403207,0.5325677395,-0.0922538713,0.1659387499,0.0012374313,0.1561725736,-0.1698634177,0.3994178772,-0.2692892849,-0.4219844043,-0.1239215508,-0.1292490661,0.3829295039,-0.4229315817,0.2115245908,-0.0087398402,-0.259565413,0.1289770454,-0.3988409936,0.0933481008,-0.1177257746,-0.0403881818,0.1544983983,-0.0220289361,0.0301756337,-0.161573723,0.1917395145,-0.3205642402,-0.3525020182,0.2628386617,-0.1099278182,0.1426189095,-0.3057217002,-0.2488423884,-0.3357530534,-0.0755689591,-0.0051657162,-0.265730828,0.3179179132,0.2084101737,-0.0123669673,-0.0700820461,-0.1287317872,0.0919010267,-0.1273197532,-0.1091576889,-0.2312927544,0.0972522646,-0.1596218795,0.033574909,0.0853756219,0.1164713725,0.3502861261,-0.3690614402,-0.0689097047,0.092549026,-0.0223239325,0.2291084081,-0.1680183113,0.3719292879,0.4319148362,0.0209557656,-0.0569459982,-0.2369533777,0.080542013,0.0453715585,0.0599122755,0.1212506443,0.3780147135,0.1420286298,0.7747809291,0.2866213918,0.0910940319,0.3499988019,-0.0283163488,0.2063764483,-0.2522653937,-0.393512398,-0.2120963484,-0.3562323451,0.0119807376,-0.1822084486,-0.1161310151,-0.3244532049,-0.2922693491,0.4655032158,-0.3842079937,-0.2641273439,0.0236350801,-0.5137665868,0.2076698542,-0.0357886925,0.0154689671,-0.0819851086,-0.0316874012,-0.126182586,0.2762473822,0.0724563301,-0.2466897517,-0.1857015043,-0.2622128427,-0.5433810353,0.3920854926,0.205896467,0.5801008344,0.0003778818,-0.0944950581,0.0310256258,-0.1828487217,0.9522311091,-0.4960359633,-0.3466887474,0.2083470374,-0.5455588698,-0.0968325734,-0.1612383127,-0.1376571506,0.5284411907,0.3635904491,0.4567363262,-0.1420131028,-0.3185372651,0.1275012493,-0.2275508344,-0.1594369859,0.0970046669,-0.3764307499,-0.0169399306,-0.2257393152,0.0497246124,-0.2841274142,0.0744781494,-0.2121138573,0.0725718811,-0.0414897986,-0.1092957035,0.0846452639,0.012382485,0.0425305255,-0.1751033366,0.0125152795,0.1235080063,0.2027816921,0.2298823595,0.3496079743,-0.4466882646,-0.0730663985,0.1103931665,0.0605578311,0.1754440814,0.3381383717,0.189244926,-0.0406991914,0.2022452801,0.1370728165,-0.2748180032,-0.0869929045,0.3566794097,0.0427998863,-0.4777282774,-0.1417573541,-0.0164189879,0.3961031735,-0.2092303634,0.4785840511,-0.3369475305,-0.2643272281,0.3273872137,0.0282706507,0.8502163887,-0.39525491,0.0811006054,0.0051095486,0.002954385,0.2348654568,0.067219846,0.0788503289,-0.2991178334,-0.0682635307,-0.0635966137,0.0854782835,0.2161601782,0.0640435666,-0.1342250109,0.474964112,0.0191980656,-0.0096894195,-0.0696131587,0.065487951,0.4006583989,-0.0159193072,0.0287466086,-0.0263755322,-0.0330142453,-0.015010234,-0.0109568126,-0.0315054469,0.2435775101,-0.2060918659,-0.3993578255,-0.0928605348,-0.1005685776,0.1672329158,0.2263244092,0.0712780952,-0.1681398153,-0.0114671262,0.0234831255,0.0354606025,-0.095732443,-0.1205083355,0.0795378089,0.3757038713,0.1605424732,0.1067806035,0.324285239,0.0916084275,-0.24716115,0.0538606569,-0.2858956754,-0.3169012666,0.1146847904,0.0705760419,0.0022970743,0.1854238361,0.1068824083,-0.022249246,-0.1199303195,-0.2102943808,0.0326752886,0.0129163507,-0.1103709564,0.3514623046,-0.1062302813,-0.3452733457,0.0387974381,0.5714837909,0.1792283505,-0.1365397871,0.1708634347,0.108427994,0.0061463984,-0.1767272949,-0.3147130907,0.1856324971,-0.182486251,0.2496794462,0.0504752621,-0.3122133315,0.2068657726,0.1525930166,0.1529288888,0.4311659336,-0.1888735443,-0.3375231326,-0.5552377701,0.3313269019,0.0134999221,0.0687698871,-0.0746563971,0.0384713411,-0.096709393,0.3323038518,-0.2229068726,0.0763912499,-0.0844008923,0.1326912791,-0.0582747646,-0.0666019917,-0.0561770871,-0.1351225078,0.0327473469,-0.0399704427,-0.3238811493,-0.0252359994,0.0375823751,0.1396390051,0.0788782611,-0.1851934791,-0.081432946,-0.0081887683,-0.045556508,-0.1865731627,0.2877725065,0.3394424021,-0.1544642001,0.2089521289,0.2331169397,0.1577552855,0.0578177571,0.3292011321,-0.1893489063,-0.0028628858,-0.0109865628,0.3587944806,0.0093151974,-0.1553913504,0.1088603064,0.1274763644,-0.1252475679,0.0000302324,0.3466128409,-0.1869863272,-0.0326906219,-0.0120867305,0.4094203115,0.3499452174,-0.0911726132,0.0539570414,0.0707102567,0.087418139,-0.0165350568,-0.2206016481,0.4769599438,0.1111516505,0.0399643295,0.3331834972,0.3857435882,-0.2357192487,0.3037241399,0.1749231815,-0.0326313972,-0.0436578393,0.2259695679,0.1887942106,0.1417442411,0.2702078223,0.3282844126,-0.2589625418,-0.1217816696,0.3115177155,0.0219227057,0.5094074607,-0.0482705459,0.1318750232,-0.1260092407,-0.6924303174,0.1770563275,0.1194026247,-0.43260023,0.1958500594,-0.0499767773,0.2462944388,-0.2728314996,-0.4944812357,-0.1602723449,0.4801970124,-0.2971932888,-0.3529887795,-0.1438146979,-0.168584004,0.0178841278,0.1029445231,-0.1220702007,0.2627573013,0.870564878,0.0013406149,-0.0940670744,-0.1132528931,-0.4790520966,-0.0430619903,0.3615504205,0.0103334766,0.4478519857,0.1272521764,0.0100145331,0.0303875431,0.2453121692,0.461440444,0.5924241543,-0.4128950536,0.0152744967,0.1953640133,-0.0158198234,-0.1098350286,0.4632724822,0.1278033406,0.2036515474,0.1503856629,-0.08785294,0.0157422759,-0.0646654665,0.2157691568,0.324413538,0.095329836,0.3323197961,0.0612213835,-0.1504763365,0.2467202246,0.2314826548,-0.2850705683,-0.3031418324,0.4654031098,-0.4126179218,0.2292830944,0.179342255,0.0684705749,0.1095157489,0.5616859198,0.4272277653,0.0376742333,-0.246272549,-0.1341667473,-0.4806021452,0.0521214716,-0.145783022,0.3216115534,0.029244516,-0.0412232578,-0.0230936781,-0.0139862206,0.0510885976,-0.2876891196,0.324234128,0.2073321193,-0.6163627505,0.2550119758,0.0732985735,-0.0790685788,0.0205268878,-0.2467685789,0.3759613335,0.0571835153,-0.0936128646,0.1096086353,0.1519048065,0.0314318724,0.1592660099,-0.0070399866,0.2722210586,0.5062955022,0.0325848944,-0.109738335,-0.2741948068,0.0390069373,-0.0975881517,-0.0104550356,0.0155986007,0.3920153677,-0.2077857405,-0.0883225501,-0.0510440879,0.3311134875,0.0445089377,0.0752524361,-0.4925262928,0.2965702116,0.1937415153,0.0449457653,-0.0557095818,0.4923805296,0.1975069046,0.2907029986,-0.3964489698,-0.1536383778,0.3064299226,-0.1346029937,-0.1018177792,-0.5227096081,0.2753640413,-0.2127102613,0.028195519,-0.5799015164,-0.2829042673,0.1240261123,-0.1634383202,-0.6094953418,0.2030107975,-0.1839464754,-0.1443089396,-0.0127475085,-0.2461772561,0.1673196107,-0.1142709851,0.1095995679,-0.1046536192]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Thanks for reporting.\r\nI'm going to fix that and add a test case so that it doesn't happen again :) \r\nI'll let you know when it's done\r\n\r\nIn the meantime if you could make a google colab that reproduces the issue it would be helpful ! @timothyjlaurent ","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":47,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nThanks for reporting.\r\nI'm going to fix that and add a test case so that it doesn't happen again :) \r\nI'll let you know when it's done\r\n\r\nIn the meantime if you could make a google colab that reproduces the issue it would be helpful ! @timothyjlaurent ","embeddings":[-0.4121118188,-0.140652433,-0.0332481563,0.2547091842,0.1527833194,-0.157567516,0.3005402088,0.2842245102,0.1220513582,0.1786371171,0.1057421714,0.4333913922,-0.4509957433,0.407086879,-0.3537718654,0.0146760466,0.1161233112,-0.0957586467,-0.3474512994,0.2302014083,-0.2869656384,0.2845570743,-0.3517576456,0.1021793038,-0.5433183908,0.1019151434,-0.0670114979,0.2673825622,-0.0081465319,-0.524277389,0.3323918283,0.2354477793,0.0166100841,0.447376281,-0.0001247604,0.1054002121,0.4095135331,-0.0857340321,-0.0827235281,-0.3554016948,-0.2497746497,-0.1529170126,0.3264889717,0.1437782049,0.1948261261,0.1137633622,-0.1099196374,-0.2486581653,0.2269985825,0.40951249,0.0959753469,0.1484627575,0.2599979639,0.1074940115,0.0507759787,0.1133370921,-0.0529266484,0.0704243332,0.3206124008,-0.3827876747,-0.0285132341,0.2651742101,-0.2205007672,-0.22886464,-0.1529174149,-0.1964246631,0.5769401193,-0.6709399223,0.2063273787,0.057501141,-0.1395634264,0.0107133789,-0.4084628224,-0.1748313308,-0.2404256612,-0.5073007345,0.2954607904,0.0732229277,-0.1645460725,0.1377577037,-0.5124667883,-0.2298987359,-0.0507234111,-0.067041114,-0.1249317899,0.5671178699,0.1926734298,0.2503543198,0.1725292355,0.1166622117,0.2252912819,0.0004574432,-0.1084869653,0.1727093309,-0.4143948257,0.0668925643,0.0559624061,-0.0301124845,-0.2576623857,0.0049349824,-0.2148026973,0.1003776491,-0.1257175356,0.0608673617,0.4142850339,-0.0701470375,0.190014869,0.2748544812,0.2734546959,-0.2022541016,0.0313611962,0.0276476629,0.3244475722,-0.0992103443,-0.0978157744,0.3083826303,0.165708974,-0.2255903929,-0.1998536885,0.1551358998,-0.143567875,-0.1698981971,0.0583345853,0.1578019708,-0.0012234198,0.7389056087,0.169774428,0.2071860433,-0.2773303688,-0.0702123269,-0.0462489948,-0.0458043441,-0.3989473283,0.1445712596,0.2839930058,0.0619038716,-0.0171851069,0.0610632859,-0.3289617598,-0.2898176312,0.1121152937,-0.2676969171,0.0600716434,0.6653485894,-0.0666875541,0.0173054561,-0.0513036102,-0.1929867119,0.1065107211,0.2754726708,-0.5094785094,-0.1835179329,-0.2386031747,0.085321933,0.001696066,0.2611215115,-0.1982413083,0.2716643214,0.382997632,-0.321161449,-0.2798917592,-0.325404942,-0.399545908,-0.3155680597,-0.0366556421,0.3317928612,-0.4338258803,-0.0076308553,0.0181093421,-0.1475040168,0.323269397,0.1963200122,0.009014735,-0.0334046558,-0.0844932869,0.0571835041,0.0135906227,-0.1737552285,-0.0926607773,0.2674301863,0.0592565164,0.3397592902,-0.0784268156,-0.2674646974,0.2675084472,-0.1609191447,0.3010762036,-0.0502837971,-0.1990887076,0.0589140579,-0.4480958283,-0.0225977562,-0.0723860115,-0.0604242012,0.3800568283,0.2501862347,-0.1262995899,-0.5661561489,0.2953267992,-0.1131389514,0.3038969636,0.0671970025,0.172843799,0.0754977018,0.1271312982,-0.2823401988,-0.5102094412,0.2077345401,-0.2473968863,0.0669737533,-0.0537822023,0.0269371588,0.1027752012,0.1855367124,-0.1827423722,-0.1951306164,0.0312554464,-0.0589818098,-0.3378646672,-0.0385565795,-0.1812935323,0.6933577061,0.120234631,0.1226822734,-0.1053099483,0.2600259483,-0.1650939882,-0.4068416655,-0.2427138686,0.1782928556,-0.0046848473,-0.0679927245,-0.1938952357,0.4690292776,0.4765939415,-0.0957903117,-0.022742046,0.1581677943,0.2636249065,-0.2274559736,-0.1249526367,0.0475181825,0.0494193174,-0.190145582,0.1769471169,0.4724966884,0.0835908428,0.4071725011,0.177128911,0.2576017976,0.2519412041,-0.0027743236,-0.0431507677,-0.12740165,0.1881029755,-0.1360896975,0.1354992241,0.0153029328,-0.1911497861,-0.1967406571,0.0656926036,0.0238865037,-0.1781348437,0.0698431954,0.0506462269,-0.0983653367,0.1394207925,0.0915659815,0.41578421,-0.0159224998,-0.207404539,0.20070903,-0.2526382506,-0.0316543803,0.1466479003,-0.074633345,0.4457933903,0.2905400991,0.117027685,0.0396163352,-0.1530807465,-0.2957579195,0.0840767249,0.3804645836,-0.5017094016,0.2247674763,-0.2366137207,0.3143704236,0.0710460097,-0.1635515392,-0.3381688893,-0.5497383475,-0.1408770233,0.5135917664,-0.0627140924,0.0747210011,-0.0036524192,0.0951412097,-0.2184601873,0.3544715941,-0.1666016281,-0.366450429,-0.1646612883,-0.098206751,0.351431042,-0.2684016526,0.2071605027,0.1605095416,-0.2898548543,0.1963260025,-0.3250332177,0.0755517259,-0.0649642497,0.0667983219,0.0770854875,-0.0536139533,0.0180142652,-0.1801133007,0.1070772335,-0.1326913834,-0.2237788439,0.2805934846,-0.1523412317,0.0745679364,-0.2608846724,-0.3176510334,-0.4118557572,-0.0885941088,-0.0634901822,-0.1984794736,0.3380073607,0.1842394471,0.0384362675,-0.0361018851,0.0183381308,0.0001203987,-0.0855408609,-0.0190588236,-0.1976469457,0.0187298004,-0.1217797697,0.0045977086,0.0779646263,0.088039279,0.4316266179,-0.3409133554,0.0082468623,0.0334051587,-0.1302686185,0.2284380496,-0.217625767,0.3201880455,0.4128198624,0.0741938055,-0.0453657173,-0.2277775109,0.0828587487,0.0422931984,0.0735568628,0.1102715358,0.295491755,0.0971603468,0.5779762864,0.3713933527,0.1077595949,0.3553503454,-0.0479933582,0.1098847315,-0.1997596025,-0.3643081784,-0.1469741166,-0.3254588842,-0.0515602455,-0.2147618383,-0.1002346128,-0.4813914299,-0.3708153367,0.6098026633,-0.3687629104,-0.2261610031,-0.0518780649,-0.5061081052,0.258156091,-0.1039812565,0.0180868022,0.0079532955,-0.0503300019,-0.2855006158,0.1941619068,0.1236160621,-0.3415813446,-0.188784495,-0.297268182,-0.4389134645,0.3179171979,0.2152687013,0.6413885951,-0.0004070778,-0.033509057,0.0526658483,-0.1484397799,0.8724436164,-0.6391437054,-0.3901263475,0.3212072253,-0.389272362,-0.2305482477,-0.1832404584,-0.2340836823,0.5545520186,0.3665707707,0.4201633036,-0.0816426128,-0.2458472401,0.2316583842,-0.1978904307,-0.1097211465,0.0331489593,-0.3254034519,-0.0265200306,-0.1062940881,0.0646469966,-0.243336603,0.0923802853,-0.2778440118,-0.0833832622,-0.063862212,-0.0533597544,0.2049310654,0.0473098867,-0.0129926018,-0.1951615214,0.047244478,0.0210986566,0.1435592026,0.2934773266,0.3029738367,-0.4163863361,-0.081562601,0.1727877408,-0.0032636863,0.2762569785,0.3233975768,0.2592428923,-0.1373685002,0.0007327193,0.1996777505,-0.2637888789,-0.0252387766,0.3556055725,-0.0455441736,-0.5671122074,-0.1211096495,0.0051767728,0.4199864566,-0.1998379529,0.5653181076,-0.3575085402,-0.2795487046,0.4127807021,0.1079265401,0.8483618498,-0.3829786777,-0.0730301142,0.0671945438,0.0268004797,0.117967777,0.106417045,0.08790645,-0.3112837374,-0.0391732566,-0.056288708,0.0275167674,0.2479702383,0.1572481692,-0.2271204144,0.4100708067,-0.0520965233,-0.1381069124,-0.0085790716,0.1033741385,0.3932394981,-0.0094016651,0.1235437095,0.0028071036,-0.0387005731,-0.1200740263,0.0052087829,0.0790054351,0.2419164628,-0.1296007633,-0.5135247111,-0.0410001539,-0.1335330009,0.0750278234,0.2297867388,0.0510988049,0.0216668099,0.1184398904,-0.0464344546,0.0676530749,-0.1435889006,-0.0928341299,0.0835629255,0.3170045018,0.1958108693,0.0367340334,0.3408838212,0.0936860889,-0.2502159774,0.0252612624,-0.2338944972,-0.2264738679,-0.0180612244,0.0726361349,-0.0585121252,0.1486910582,0.1381905973,0.0091719348,-0.1144135594,-0.1911193579,0.047596138,-0.0709196776,-0.0892757177,0.5210028291,-0.0350488871,-0.3730113506,0.0119760102,0.4310206473,0.2444379926,-0.0667294487,0.1065042764,0.1603962779,0.0003414901,-0.2310717106,-0.2428833097,0.1716859937,-0.1189794019,0.3094241321,0.038129285,-0.1944820732,0.2528368235,0.0619290695,0.0322724991,0.3911171556,-0.1844837219,-0.3591224551,-0.4126452804,0.2500081062,0.0484349206,0.110688366,-0.1846468449,0.1446647942,-0.0302458797,0.2589550614,-0.2122109681,0.1527973861,-0.1018271074,0.2051259726,-0.0535921454,-0.0008352815,-0.0993243679,-0.143149823,0.0514928959,-0.0185622051,-0.1838881522,-0.0198913906,0.0312998742,0.1376205683,0.0847332254,-0.1993035227,-0.0601025037,-0.0665640235,-0.1236637905,-0.2029262781,0.2158674002,0.3959042132,-0.207852006,0.1407669485,0.1893225908,0.1952367872,0.1439821422,0.3243099153,-0.2149887085,0.0778922662,0.0814792216,0.4155632555,0.1775707006,-0.1909879148,0.0620319508,0.1037419811,-0.1445652992,-0.0004130547,0.3068567216,-0.2390923053,0.045528613,-0.0586902499,0.4123044014,0.4355723262,-0.1179853454,0.111096859,0.1410018057,0.1002049819,0.0493217856,-0.1471814811,0.4749159515,0.1149623394,0.0238514151,0.3524885178,0.2848343551,-0.1469901055,0.1850551814,0.1636666805,-0.0338790417,-0.0026765261,0.2539984882,0.0815500394,0.0690928549,0.2622000575,0.1822576821,-0.3103900552,-0.1121328697,0.2957111299,0.0121205077,0.46735695,-0.069283627,0.065176338,-0.0891270489,-0.632989347,0.0578954034,0.1388348639,-0.5441825986,0.2122361362,-0.0096857306,0.2614074349,-0.386408329,-0.4462724626,-0.1971577853,0.4130579829,-0.2505246103,-0.3078994751,-0.2264644653,-0.1703550369,-0.1269653887,0.109610863,-0.0965497568,0.1930417866,0.8255369067,0.0335467346,-0.1240112782,-0.1953440905,-0.3356648088,0.0009228773,0.3389827013,-0.0118953567,0.3929478228,-0.0182970427,0.0001007127,-0.0541470535,0.1687270254,0.4923535883,0.7186236978,-0.5203197598,-0.0086596869,0.1379933655,-0.065894939,-0.0626947358,0.369022131,0.1728776544,0.2969072163,0.256148845,-0.0660468712,0.045555748,-0.1087195352,0.3275897503,0.3738709986,0.0244508293,0.2912442684,0.0596285686,-0.0805852413,0.2058669776,0.0789363533,-0.3390073776,-0.2918817401,0.4837103784,-0.4562891424,0.2314359695,0.2614198625,0.0443421789,0.1067170575,0.4937187433,0.3471804261,0.1562869698,-0.2608995736,-0.2374811172,-0.3877949417,-0.0001012414,-0.1960657388,0.3422498405,-0.0291248262,0.0112806745,-0.0700277761,0.0646049529,-0.0437061004,-0.2140718102,0.3408595324,0.3193295002,-0.5734247565,0.2057787776,0.0849361271,-0.059248507,-0.0117022693,-0.2920572758,0.3183326721,0.0405557193,-0.1181763336,0.0914621949,0.1584923267,0.0675964132,0.1648001671,0.0278841928,0.3071469665,0.6216836572,0.0024168892,-0.0157921668,-0.267776072,0.0197454952,-0.0921239704,-0.0592219718,0.0696678534,0.3804354072,-0.158795923,-0.0497493707,0.047622066,0.2289706618,0.0303712487,0.2470553666,-0.4424959421,0.1602875292,0.1941824704,0.0280080475,-0.1409187168,0.3804939091,0.1256975234,0.3354586363,-0.3165165782,-0.1907719076,0.2904995382,-0.0006388366,-0.1257482171,-0.6035087705,0.210559383,-0.1881111562,0.1225874275,-0.5265963078,-0.3557211161,0.1513366848,-0.1979534626,-0.5773916245,0.2608500421,-0.1053138524,-0.1814596504,-0.0679983273,-0.2073145956,0.1469498873,-0.1275481284,0.1519545913,-0.0671171099]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Sure thing, @lhoestq.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1lg4fbyrUO6m8ssQ2dNdVFaUqMUfA2zZ3?usp=sharing","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":4,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nSure thing, @lhoestq.\r\n\r\nhttps:\/\/colab.research.google.com\/drive\/1lg4fbyrUO6m8ssQ2dNdVFaUqMUfA2zZ3?usp=sharing","embeddings":[-0.4047035575,-0.1519752741,-0.0442972817,0.2092177272,0.1404305995,-0.1517882198,0.2771425247,0.3226374984,0.1497403532,0.1770489514,0.0803230926,0.4027684927,-0.452278614,0.3576361239,-0.3363932073,0.0439927764,0.0808749646,-0.0507654138,-0.3425464928,0.2297851443,-0.2877653539,0.2560468912,-0.3207340837,0.1290315688,-0.524396956,0.0934435204,-0.0118205724,0.2905389965,0.0102327932,-0.5569396019,0.3371310234,0.2326480597,0.0166103113,0.4065383971,-0.0001224699,0.1519722641,0.3582121432,-0.1318755299,-0.0577861369,-0.3551974595,-0.2268612385,-0.1915116161,0.2823707759,0.0964912847,0.1760086715,0.0437405482,-0.0894626155,-0.2767762244,0.2215551436,0.4360563159,0.110325411,0.2143336385,0.2522470951,0.1237717569,0.0072554876,0.1321102083,-0.0447281599,0.0270251259,0.3032941818,-0.3398368955,-0.0824702382,0.2699809968,-0.2196102291,-0.1953542531,-0.1629343629,-0.2304776013,0.6503714323,-0.6132762432,0.1957882643,0.0520638376,-0.222206071,-0.0039733942,-0.4121744037,-0.164399296,-0.2564831972,-0.4938956499,0.2126664519,0.0338227488,-0.1657253206,0.1377286613,-0.4768919647,-0.2225684971,-0.0341195092,-0.091705896,-0.098672457,0.545316875,0.2081365287,0.3022448421,0.1664265096,0.1238008514,0.1767909378,0.0367395096,-0.0688051581,0.1585240215,-0.4171749055,0.0883202553,0.0731538907,-0.0767670348,-0.248809427,0.0137146255,-0.1459393054,0.1057629585,-0.1057846993,0.0684987158,0.4067669809,-0.1171153709,0.1763373464,0.2674952149,0.255762279,-0.2099135965,0.0012335845,0.0450764559,0.3482174873,-0.1142116934,-0.0534105822,0.3068736196,0.147450462,-0.1945344359,-0.2118851095,0.1415082216,-0.0858705789,-0.1818106472,0.0940789878,0.1864383519,0.0569611192,0.7809958458,0.1382970959,0.1527548134,-0.3145497441,-0.1108476967,-0.0687057748,-0.0359514505,-0.3893536329,0.0698695034,0.2773252428,0.0142148761,0.0182809122,0.0041857273,-0.3702150881,-0.2565169036,0.1380714476,-0.2805086076,0.1044341251,0.6892309785,-0.0815829113,0.0075701526,-0.0534430481,-0.1306143403,0.0908776596,0.2581395209,-0.5104535222,-0.1893363446,-0.2075139433,0.0948870406,0.0435233526,0.3142111897,-0.1236462966,0.234451443,0.3833895326,-0.3764632344,-0.3001309335,-0.3596326709,-0.3816435635,-0.2924422026,-0.0350013934,0.2845458388,-0.3902513683,-0.0100644697,0.0667534173,-0.2195940763,0.3322684169,0.2005055249,-0.0202976167,-0.0210816078,-0.1190663204,0.0851615965,0.1007616594,-0.1957113445,-0.0883903727,0.2751072347,0.0508202352,0.3149879575,-0.0319327042,-0.2479927838,0.2682059407,-0.1619299948,0.2473702878,-0.04896551,-0.1799808294,0.0629266351,-0.4437901974,-0.0159065183,-0.0448208302,-0.0507712029,0.3856495321,0.1990354955,-0.1183740646,-0.4984902442,0.2931684554,-0.1042717323,0.3077528775,0.0823402181,0.0826190114,0.0551424995,0.1370140016,-0.3303594887,-0.4922583401,0.1733207107,-0.2664736807,0.0233625434,-0.0476175398,0.0322719552,0.0871725529,0.1912230551,-0.2058448941,-0.1962890774,0.0503740646,-0.0041906731,-0.2929547131,-0.0685327575,-0.2092014253,0.6694061756,0.1017912552,0.1152240708,-0.1284542829,0.1986725926,-0.1571337879,-0.4181500375,-0.2600544989,0.2615119815,-0.0259595811,-0.0657595396,-0.1867568344,0.4662875831,0.4772807062,-0.1277229935,-0.0621904097,0.1310781986,0.2439221889,-0.1862971038,-0.1580088139,0.0920259804,0.0379152112,-0.17450656,0.2006775439,0.4543550313,0.0985514373,0.3889692128,0.2271975279,0.2363935113,0.3221971989,0.013306397,-0.0277118478,-0.1510295272,0.1916356683,-0.1242921427,0.1721647233,0.0444665961,-0.169449091,-0.2546758354,0.110984467,0.0467454568,-0.1627622098,0.0583128817,0.0916021615,-0.0923540667,0.1328216344,0.0741846934,0.418497026,0.0107104704,-0.2050214112,0.1692454517,-0.2978027165,-0.0218858197,0.1124465242,-0.0777798593,0.4277511239,0.3414438963,0.0796985328,-0.0264739841,-0.1785031855,-0.3693338037,0.1029621884,0.3623557985,-0.5024702549,0.2006637603,-0.2493201047,0.3571793735,0.0770362988,-0.1340099871,-0.284524858,-0.5692771077,-0.119645521,0.50300318,-0.0992926955,0.1078938693,-0.0909410566,0.0646631718,-0.2059529573,0.3792887032,-0.1428103149,-0.3044557571,-0.1719958484,-0.0960277095,0.3353699744,-0.2427075952,0.2141651511,0.1393330544,-0.2615479529,0.190169692,-0.2960283458,0.0758780465,-0.0298639573,0.0488731302,0.1228439584,-0.0161502156,0.056866955,-0.2068219185,0.1461416185,-0.1532940865,-0.2158495933,0.2972332537,-0.1320170611,0.0963791385,-0.2746717334,-0.3468492329,-0.4326058328,-0.1261550039,-0.0812491775,-0.1907311529,0.350278914,0.1516907364,0.0276564881,-0.006403185,-0.0266876016,0.0756083205,-0.0578590296,-0.0563744456,-0.1711387932,0.0382254757,-0.1299647838,-0.03975337,0.0631706864,0.1299539208,0.4353449941,-0.2987812757,0.0272044279,0.0736795366,-0.126980111,0.1909008324,-0.1969956607,0.3493245244,0.4263823926,0.0442796014,-0.0353780352,-0.1740325242,0.1409507394,0.005688814,0.0488800444,0.1144257113,0.3136581779,0.1149055138,0.572329402,0.3782865405,0.0552950464,0.3922972679,-0.0838295892,0.105117999,-0.1933934689,-0.3614845872,-0.1856548488,-0.2923091054,-0.0170609597,-0.1943405271,-0.13094531,-0.4439553022,-0.330755353,0.5747679472,-0.3650678098,-0.2178163975,-0.0196026154,-0.496961534,0.2732592821,-0.1022956148,0.0436764061,-0.028450381,-0.0016492289,-0.2562008202,0.1830534041,0.1333388388,-0.3510716856,-0.2712600827,-0.2851646841,-0.4547207057,0.3030996025,0.2419728339,0.6977062225,0.0154547393,-0.062961109,0.008775061,-0.1732638776,0.9061996937,-0.6183616519,-0.3526533544,0.2977403998,-0.3949985504,-0.2654436529,-0.1656398326,-0.2222298682,0.5450920463,0.3819300532,0.401558727,-0.1398145854,-0.2926309407,0.2078938782,-0.2269920856,-0.1192756519,0.0306445695,-0.3259885013,-0.0138709173,-0.170777604,0.0273522399,-0.2482544482,0.1038410366,-0.223843798,-0.098688215,-0.1209340245,-0.0832896233,0.2511928678,0.0587145127,-0.0094820457,-0.1853179336,0.04731277,0.0248339027,0.192078501,0.2725405097,0.2676166594,-0.4400148094,-0.1200309619,0.1890266538,0.0168126356,0.2981960177,0.2841264606,0.2306372225,-0.1168344617,0.0428154804,0.2038948834,-0.2922552824,0.0041800165,0.3734890819,-0.0242690779,-0.5242649317,-0.1489549428,-0.0643860996,0.4139291048,-0.1621257812,0.4663180411,-0.4161088169,-0.2600950003,0.3970193267,0.0965410098,0.7863522172,-0.379188776,-0.0430715047,0.0583794527,0.0188015886,0.0802370235,0.0516023189,0.0389571004,-0.2663422823,-0.0149150062,-0.0648703277,0.0501480885,0.3107153773,0.1760373116,-0.1739809513,0.4373416603,0.01944899,-0.1177512035,-0.0529404432,0.1573265642,0.4464294016,-0.0455259532,0.0865266621,-0.0057777762,-0.0627519265,-0.1834568679,0.0314057693,0.0305048898,0.2921484709,-0.1282947361,-0.4060802162,-0.0696388409,-0.1310393214,0.1203927249,0.1890586168,0.1288673133,-0.0020236559,0.1303165853,-0.1130681336,0.0628703609,-0.1331321448,-0.1186146066,0.0914787799,0.3381004035,0.171694845,0.0706265345,0.3721080124,0.0954347104,-0.2663465738,0.0591174252,-0.2657741308,-0.2451483607,0.0679274574,0.0540541224,0.0551687591,0.1124071479,0.1114096045,0.0026680261,-0.159927696,-0.1883961409,0.0539297387,-0.078941837,-0.1275245845,0.4977596998,-0.018710969,-0.3787944019,0.0045321463,0.4575691521,0.2522782385,-0.0326194949,0.1495291144,0.096856989,-0.0432519764,-0.2423483431,-0.2222360224,0.2177442312,-0.1519189328,0.3033234179,0.0832955614,-0.2459680587,0.2281862646,0.0457830504,0.0632474795,0.3542215228,-0.2240121961,-0.3499901295,-0.457872659,0.2571275234,0.060101375,0.1075876951,-0.1160909235,0.1448788643,-0.0795240775,0.2947714031,-0.2364096642,0.1169478372,-0.1051279306,0.2275077999,-0.1088565961,-0.0068889325,-0.1135795042,-0.0885211602,0.0847740173,0.0059529264,-0.2423905879,-0.0330701992,0.0504651293,0.1298284978,0.0734717697,-0.2102362216,-0.055923067,-0.072836481,-0.1290493309,-0.1997785419,0.1627172381,0.3316245377,-0.2109104544,0.1653403193,0.2289645523,0.1676353067,0.1192844659,0.321856916,-0.2426526397,0.0338934325,0.0273177959,0.4187778533,0.1041275933,-0.1888513416,0.0242512915,0.093254298,-0.0825406462,-0.0000861904,0.3171823919,-0.2159916162,0.0221473183,-0.0018463288,0.42916435,0.4252501428,-0.1147810295,0.1191147193,0.131373331,0.097173214,0.0054815267,-0.1514835954,0.5023469925,0.0903691575,0.0928137973,0.3471127152,0.310718596,-0.1864662915,0.2238954753,0.1389076859,-0.1034619585,-0.0447146818,0.2090824097,0.1781285703,0.0867964923,0.2475539744,0.2225087583,-0.3072215021,-0.0937571824,0.3514227867,-0.0099820886,0.5492145419,-0.0604326874,0.1238287464,-0.1222563609,-0.6341024041,0.0741259009,0.1342924833,-0.5317959189,0.2296218574,-0.0112149157,0.2613585293,-0.3300201893,-0.4241015315,-0.2059581578,0.4338274896,-0.2633914649,-0.3475887775,-0.2358143181,-0.1638201177,-0.1081883088,0.091538474,-0.1079184487,0.197051838,0.79980129,0.034697745,-0.0692217648,-0.1735396981,-0.336678654,-0.0381818637,0.3228043318,-0.0258335117,0.4213103354,0.0350082181,0.0291371774,-0.0274272002,0.1536231041,0.4629772305,0.6900582314,-0.4682578146,-0.0346733779,0.1684129238,-0.0958714783,-0.1157678664,0.4023528695,0.1192984954,0.3145189583,0.3011546433,-0.0492697321,0.012480312,-0.173207283,0.318105787,0.3608618975,0.0571437813,0.3494888842,0.0968413502,-0.0829770565,0.2331873029,0.0611548945,-0.3592247367,-0.2743960619,0.4677306116,-0.4814819396,0.2664362192,0.2378314734,0.056812916,0.0688586906,0.480858326,0.2674568892,0.0946621671,-0.2661913335,-0.2027421445,-0.4044625461,-0.0160115287,-0.1812925339,0.2821110487,0.0072077545,0.0094185723,-0.0579225235,0.0353193991,-0.0638366118,-0.2363543808,0.3875297606,0.3233847618,-0.5686668158,0.183165729,0.070558995,-0.0353785828,0.0169519857,-0.3133646846,0.3429664373,0.0454765223,-0.0935216174,0.1039120704,0.1673183739,0.0731395632,0.2092420608,-0.0025735351,0.2695357502,0.5883299708,-0.0498463176,-0.0116383769,-0.2533767819,0.0700618476,-0.0786610395,-0.0423866063,0.0892889798,0.4106647372,-0.1987763345,-0.0295353923,0.0421900004,0.1762108803,0.0006465491,0.2341559082,-0.445741266,0.1941167116,0.2012751848,0.0194862243,-0.1482251883,0.3713886142,0.159867838,0.2976709604,-0.2920514047,-0.1958272457,0.2922280729,0.0048716362,-0.1147936657,-0.6067380905,0.2108431607,-0.2288656831,0.1436238736,-0.5422935486,-0.3122961819,0.1793645173,-0.1928045005,-0.5784544349,0.2572570443,-0.1100066155,-0.1766641587,-0.0610997379,-0.2316268682,0.1409604996,-0.1291955709,0.1669092476,-0.0736563802]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Thanks @timothyjlaurent ! I just merged a fix on master. I also checked your notebook and it looks like it's working now.\r\nI added some tests to make sure it works as expected now :)","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":35,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nThanks @timothyjlaurent ! I just merged a fix on master. I also checked your notebook and it looks like it's working now.\r\nI added some tests to make sure it works as expected now :)","embeddings":[-0.4193556011,-0.1071404666,-0.0392603949,0.2147689611,0.1458776742,-0.1598080397,0.2604388893,0.3937074542,0.0989538953,0.1513863951,0.0195596758,0.4504702091,-0.3637407124,0.4398677647,-0.319118619,0.0345809273,0.0604736246,-0.0270151775,-0.3903119564,0.2088315785,-0.2636983693,0.2989494503,-0.3305636048,0.1506016999,-0.5216234326,0.1386964023,-0.1130683497,0.2653171122,-0.1030749157,-0.4939137399,0.2864601016,0.2612058818,0.0879322216,0.4689686894,-0.000114397,0.1324707568,0.3333232403,-0.1248081103,-0.0798088238,-0.4356539845,-0.1579086185,-0.2550667822,0.2872832417,0.1007006541,0.0749271736,0.0389388353,-0.090324752,-0.1414096206,0.2836947739,0.4652611017,0.1934158057,0.2085710764,0.2621919215,0.0558455847,0.0510044582,0.0678695366,-0.0874518901,0.0078280475,0.3059768975,-0.4177977145,-0.1298090667,0.3312835991,-0.1521880329,-0.2131036073,-0.0349628963,-0.192424655,0.5803083181,-0.5957256556,0.2125930339,0.0949263498,-0.1266435832,-0.0087137632,-0.3816846311,-0.1348068714,-0.2131182551,-0.4743331075,0.2450659275,0.0600356758,-0.2198235989,0.1639676094,-0.4446834028,-0.1741554737,-0.0448266342,-0.130474031,-0.1257179081,0.5451292396,0.1489355117,0.2062472105,0.2068864256,0.1561340839,0.2369428277,0.0344263203,-0.056958802,0.1895682663,-0.4668980837,0.0611613691,0.0921528935,-0.0954745263,-0.2770744562,0.0083476817,-0.088118881,0.1364300549,-0.0016214101,0.0187218487,0.3237471581,-0.029831551,0.1760365963,0.2339707762,0.2529138923,-0.1339113861,0.0106226373,-0.032067012,0.2401254177,-0.1618898809,-0.1238920242,0.2203139216,0.1702696532,-0.2124930322,-0.212342456,0.1235531121,-0.0652692989,-0.1341298968,0.0883261785,0.2103799134,-0.0055050431,0.6364831924,0.2363899052,0.1668306589,-0.3084965646,-0.074295789,-0.1081771702,-0.0752082616,-0.4380477369,0.0819578096,0.3093464375,0.0654760227,0.0247549154,0.0464761332,-0.3424578011,-0.1757465005,0.1699210703,-0.2142257243,0.0935268775,0.6487029791,-0.0624318235,-0.0729875714,-0.0194960609,-0.1211733371,0.1072037518,0.2887684703,-0.4147422314,-0.1756033152,-0.1448325366,0.1784071028,0.0624776594,0.1968678981,-0.1354858726,0.2495191991,0.3412889242,-0.2765164375,-0.295017302,-0.2627283931,-0.3219247758,-0.264962256,0.0088395793,0.1723242551,-0.4067887068,-0.0042888788,0.0150060412,-0.1794583201,0.2717075646,0.1691123396,-0.0282548945,0.0070827706,-0.0883457139,0.0955991149,0.0768802315,-0.1691041142,-0.1678660512,0.2952840924,0.0339029543,0.2496963441,-0.0105180228,-0.2364145517,0.3208428025,-0.0797638372,0.244471252,0.0074716737,-0.1422511488,0.0394449793,-0.4914548099,-0.0142364688,-0.079917632,-0.002936464,0.4351004958,0.1397288442,-0.19181934,-0.5129804015,0.3801808655,-0.1166738868,0.3654141128,0.0820588022,0.1434988976,0.1009431332,0.1733538955,-0.2495142519,-0.3573037684,0.2332869023,-0.2872871161,0.1226692721,-0.0040788511,0.0095777353,0.0459842421,0.0689831972,-0.2158465236,-0.2534330785,0.1461217999,-0.0497329086,-0.2649739385,-0.0695729107,-0.2448036522,0.6643331647,0.1081313491,0.1286108643,-0.1205222532,0.2059649974,-0.2012664825,-0.4136802554,-0.2300351411,0.1771996021,-0.0337785371,-0.0788786635,-0.196377039,0.4693377316,0.4146348238,-0.2028420717,-0.0868468285,0.1255912632,0.1912124008,-0.1506284326,-0.1443390846,0.1170540452,0.0710575134,-0.1876674891,0.1343067437,0.5086156726,0.0340318941,0.3584803045,0.2193984985,0.2494549453,0.3088573515,-0.0572914593,-0.0569906831,-0.1041017026,0.1401890814,-0.1657108814,0.1885280013,0.0321423002,-0.1542482376,-0.2205114812,0.0744976252,0.0366966315,-0.2033707052,0.0992500782,0.0222155657,-0.0738715306,0.1572184414,0.0868034586,0.435272038,0.0814119726,-0.1514406502,0.2476586252,-0.2685739696,-0.0336246639,0.1898126304,-0.1886319965,0.3337744474,0.3131489158,0.0016618175,-0.064507246,-0.2526164353,-0.2312238514,0.1290824711,0.3988618553,-0.4570883811,0.1193113253,-0.2042877078,0.2391608059,0.0630442053,-0.1325523555,-0.2879113555,-0.4914489686,-0.0683125779,0.5388637781,-0.0008215093,0.1363489777,0.0390867554,0.1202900931,-0.1639818847,0.3817952573,-0.1218742654,-0.339301914,-0.2346854359,0.0099810669,0.3615128994,-0.1558511108,0.2152867764,0.0898219794,-0.2461383492,0.1259487867,-0.3414521813,0.0274587981,-0.1188350245,0.0474604592,0.0826992169,-0.0263408255,-0.0158582143,-0.1403867304,0.146448195,-0.1827648282,-0.2847910225,0.3372772336,-0.1016341746,0.0589413941,-0.2651586235,-0.3587749898,-0.3944595754,-0.1789947748,-0.0360180959,-0.2031382173,0.2997399271,0.1594230682,-0.0117753418,0.0170342978,-0.0087831635,0.0573236682,-0.108346872,0.0103760883,-0.1428316981,-0.0703980029,-0.2448671162,0.0341876782,0.1254526079,0.1972046494,0.3103440404,-0.3343865573,0.0510897487,-0.0141258808,-0.0532010905,0.2020154595,-0.137279138,0.4295086265,0.3030903339,-0.018551955,-0.0289912019,-0.1327067763,0.0661371946,-0.0121595031,0.1117382348,0.026613513,0.2794807255,0.1757091582,0.5257003903,0.3726461232,0.1225586161,0.3527100384,-0.0770341083,0.11402601,-0.2462852001,-0.4222671986,-0.1674271375,-0.3445949256,-0.030351514,-0.1905115098,-0.1223335266,-0.4686021507,-0.3261251152,0.4740751684,-0.2929596305,-0.2703279853,-0.0009388852,-0.3815755546,0.2553579807,-0.0103183324,0.0509463958,-0.0073614311,0.0065027364,-0.234153524,0.2373700291,0.0275295991,-0.343007654,-0.2495888919,-0.2013953924,-0.4698574841,0.325558126,0.0736435428,0.5758503079,0.0664786845,-0.15570876,0.0528421551,-0.1451820582,0.8551930189,-0.5291436315,-0.4163536429,0.3530488908,-0.4446538389,-0.298630625,-0.229309991,-0.2401954979,0.4804597199,0.3630558848,0.4405975938,-0.0835954547,-0.3271422088,0.207335487,-0.2588736415,-0.0956456587,-0.0193275213,-0.3289504051,0.0026642485,-0.1439438313,0.0806328878,-0.2556236982,0.1615923494,-0.2231432199,-0.1014842689,-0.1360713392,-0.1502159536,0.1863693595,0.131548062,0.0114303567,-0.236631915,0.0945365131,0.0136122499,0.1998776048,0.2887659967,0.2445150316,-0.4175812304,-0.1409902871,0.1858459711,-0.098127909,0.2522603571,0.3151225746,0.2014820576,-0.121371761,0.0429509096,0.1488642842,-0.2819696367,-0.0143473018,0.3474958837,-0.0276895687,-0.4713943005,-0.1396845728,-0.0640943497,0.3640647233,-0.186322704,0.4500076175,-0.4243415296,-0.303974092,0.4157571495,0.0357806794,0.8078531623,-0.351523906,-0.0082466621,0.0785525739,-0.0528231896,0.0444090888,0.1244342029,0.130584389,-0.3808036447,-0.1677162498,-0.0221058801,0.0908225179,0.300483048,0.1597411782,-0.2783438861,0.3571475744,0.0343680941,-0.1328841597,-0.0370978229,0.1615804583,0.3707350492,-0.0322963372,0.0534359254,0.0826026201,0.0088990666,-0.1095326245,0.0225268602,0.006069296,0.2187006921,-0.1285784245,-0.3923016489,-0.1209193766,-0.1604751348,0.1002305672,0.1534647942,0.0861688107,-0.0856876597,0.0591864847,-0.049535498,0.1008600369,-0.1607711166,-0.0901113227,0.0909276754,0.3196435869,0.1687979251,-0.0326881856,0.3603781164,0.0404223241,-0.2821494043,0.0309956204,-0.1847845167,-0.2302516252,0.0924392045,0.0527896732,0.0068405238,0.0545469597,0.103542231,-0.008926061,-0.1424869895,-0.2472645044,0.1489107013,-0.1039120927,-0.2403700799,0.4878776371,-0.040604502,-0.3871281445,0.0060362765,0.5196280479,0.2512229681,-0.0421018302,0.1161166653,0.1134916469,-0.0052116779,-0.2603788674,-0.1053166166,0.3555198014,-0.1472779065,0.2697661817,0.0772921592,-0.2323991358,0.2793722451,0.1730698943,0.1864338815,0.343708843,-0.2330510765,-0.2502618134,-0.5016092658,0.2002421319,0.0906714872,0.1555784643,-0.1567183435,0.1033436358,0.0227641687,0.2380063981,-0.3144082725,0.1618686765,-0.1278701127,0.1995831728,-0.0874239877,-0.0639711544,-0.1013243124,-0.128878206,0.1278573573,0.0110290479,-0.2183108479,-0.1269757599,0.0414432287,0.0959978923,0.0903681964,-0.1801791042,-0.0829938054,-0.0894424841,-0.1912938058,-0.1733568907,0.2130332142,0.3244560361,-0.2227032036,0.1553198099,0.2211142927,0.0520623177,0.1144620404,0.2124032378,-0.2451878637,-0.0374416448,0.0753425658,0.339186281,0.1671214849,-0.1564786881,0.0699888468,0.0722213909,-0.1351470649,0.0325764231,0.3185409009,-0.2506429851,-0.0078421384,0.0133410953,0.4798089266,0.4465430379,-0.0354574621,0.0300003476,0.1677158177,0.2217455655,-0.0638004839,-0.1488431692,0.4273242056,0.0264337659,0.0895147175,0.2835226655,0.2823504508,-0.1218487993,0.1831276268,0.195711419,-0.1121933982,-0.0428237692,0.2549173832,0.1816561222,0.0929491669,0.2404345572,0.2450756133,-0.3198019862,-0.1236821488,0.3709716499,-0.0579796173,0.4727436006,-0.0739560127,0.1659107357,-0.1547366381,-0.6035438776,0.0466978624,0.0995771512,-0.576584816,0.1664244086,-0.058970511,0.2135984302,-0.3429754674,-0.4492863417,-0.2299197167,0.3815024793,-0.3193082511,-0.278416872,-0.2968035638,-0.1927315146,-0.1144058108,0.0155423097,-0.0980919227,0.2728915215,0.7926209569,0.0167066269,-0.0766709894,-0.1761127263,-0.3733610511,-0.0104631726,0.2841352522,-0.052612137,0.3761051297,-0.0376650766,0.0078545539,-0.0527078547,0.2472108454,0.4495357871,0.758901298,-0.4396671355,0.0291388761,0.1385423243,-0.0769141093,-0.1264016926,0.390437752,0.2272910178,0.3436399102,0.2906418741,0.0186776556,-0.0276867487,-0.0531580523,0.1893318146,0.2868379653,-0.0592094697,0.2951444387,0.1460803449,-0.0888965726,0.168343097,0.1084670499,-0.3899188936,-0.2188164145,0.5099228621,-0.4190782309,0.2469058633,0.143342793,0.0981121436,0.0651598424,0.5027341247,0.2623902261,0.0760018528,-0.3008476198,-0.2206918597,-0.4730150104,-0.030117074,-0.2012250274,0.2934896946,0.0006068708,0.0507429689,-0.0133373113,0.0263473317,-0.0222064499,-0.3289549947,0.2724009752,0.3368607759,-0.5882843137,0.2075416744,0.132759884,-0.0550442524,0.0063749403,-0.3181932271,0.3201171756,0.0733071491,-0.0068846676,0.1621217728,0.1773038059,0.1711476743,0.2259214222,0.0468780957,0.2819208205,0.4912543595,0.0361809768,-0.000960627,-0.3127283454,0.1062100753,-0.1507226676,-0.0046324618,0.1694760025,0.4342525899,-0.1230455935,-0.1056581363,-0.0087013729,0.2750091851,0.0663075149,0.1348055899,-0.3960781097,0.2369547486,0.2314292043,0.0715242177,-0.1251013726,0.4445693195,0.1168591455,0.2360392362,-0.3626091182,-0.2346958518,0.26160267,-0.007833004,-0.1696405113,-0.6140580177,0.1814789623,-0.2047005296,0.1055041105,-0.5150931478,-0.3300754428,0.1696694046,-0.169114992,-0.5518653989,0.280289948,-0.03587332,-0.1203283668,-0.0870618448,-0.2423843592,0.1596472114,-0.1952757835,0.0383121967,-0.1134013236]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"Great, @lhoestq . I'm trying to verify in the colab:\r\nchanged\r\n```\r\n!pip install datasets\r\n```\r\nto \r\n\r\n```\r\n!pip install git+https:\/\/github.com\/huggingface\/datasets@master\r\n```\r\n\r\nBut I'm still seeing the error - I wonder why?","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":32,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nGreat, @lhoestq . I'm trying to verify in the colab:\r\nchanged\r\n```\r\n!pip install datasets\r\n```\r\nto \r\n\r\n```\r\n!pip install git+https:\/\/github.com\/huggingface\/datasets@master\r\n```\r\n\r\nBut I'm still seeing the error - I wonder why?","embeddings":[-0.4416673183,-0.1638828814,-0.033368811,0.2726554275,0.1398596466,-0.1488540173,0.2888736427,0.2428048104,0.1233460158,0.2103493214,0.0198594723,0.5049683452,-0.3794004917,0.366486311,-0.3114085793,0.0422243029,0.082374312,-0.0164952278,-0.3751588166,0.2027810365,-0.1774687022,0.2914339304,-0.3229137957,0.1249253452,-0.5289024115,0.0481469445,-0.0232043732,0.2264141887,-0.1429692358,-0.5102818608,0.4412212372,0.2774354517,0.0068003652,0.4950698614,-0.0001267765,0.1621522009,0.4178577662,-0.12817958,-0.1998692304,-0.4242440164,-0.2842195034,-0.2430701256,0.3733290136,0.0529889204,0.1932425946,0.1888237149,-0.0460889228,-0.2160014063,0.2681325972,0.3944459558,0.0823459849,0.2147485912,0.2474290878,0.104615882,0.0297514275,0.1779660583,-0.0468287989,0.0709813312,0.3152913153,-0.2751640379,-0.0240561645,0.2880619466,-0.2587496936,-0.0838409141,-0.2140988559,-0.1705383509,0.4943110645,-0.6263384819,0.1955993026,0.081557855,-0.14066495,-0.074899599,-0.4533854425,-0.2244773209,-0.2576598823,-0.5030200481,0.2255763113,0.0383145958,-0.2167304158,0.118720226,-0.5351097584,-0.2030429691,0.0226368159,-0.0981145278,-0.1145919189,0.5615178943,0.0796506554,0.3092218637,0.2317216545,0.2045359164,0.2542188764,-0.0397200175,-0.0713268816,0.1538144201,-0.3840873837,0.0325841345,0.0669131801,0.1032212079,-0.199235037,-0.0264021195,-0.2357920557,0.0283500869,-0.1383860558,0.0962060317,0.4001834691,0.0043299948,0.138394773,0.287150383,0.2604061067,-0.1034703553,0.0161071625,0.0157799069,0.3302517235,-0.0691931769,-0.1319175959,0.3151136041,0.225154534,-0.2654107213,-0.243190214,0.0499477126,-0.0663372204,-0.2566161156,0.0794131011,0.1817845702,-0.0491336361,0.6901754141,0.0495948121,0.146594733,-0.3301025331,-0.0150012281,-0.0359648019,-0.0239534974,-0.3363146484,0.071367763,0.2595929205,-0.0516902432,0.0886395276,0.032740254,-0.3011175692,-0.2104311436,0.1050219163,-0.3483010828,0.054035604,0.7567947507,-0.0330230221,0.1069527194,0.0345672928,-0.1978487968,0.0832137614,0.2143735737,-0.5603205562,-0.2310544103,-0.253677398,0.0324034058,-0.0642949715,0.2737276554,-0.3409034312,0.2217778116,0.4327417016,-0.3949703872,-0.2841382921,-0.3484717011,-0.3798283935,-0.2361614555,-0.039976377,0.3014243543,-0.4129752517,-0.0434269309,-0.0027410423,-0.2266343832,0.412774235,0.2082967609,-0.0585957058,0.0484883599,-0.175472483,-0.0414133742,0.1345539391,-0.2448756397,-0.1387614906,0.2229841501,0.0678844824,0.3529622257,0.0010165133,-0.2929701209,0.2707001567,-0.1585672945,0.2899981141,-0.1811337024,-0.2222993821,0.0575292185,-0.427988708,-0.0444761291,-0.1362227798,-0.0758409053,0.3197173476,0.1904804111,-0.0917359963,-0.4803735018,0.291433841,-0.0974822417,0.2749807239,0.0469423085,0.1833526492,0.0583575293,0.0860807151,-0.2779419422,-0.5560165644,0.1758852005,-0.1128361821,0.0734608397,-0.181999594,-0.0033157822,-0.0047651408,0.2145400345,-0.2078028768,-0.187245205,0.0109829074,0.0678195879,-0.2397839576,-0.0476466753,-0.1803411692,0.7802555561,0.0948861465,0.1407016963,-0.1968691349,0.218893528,-0.1770649403,-0.3737702966,-0.2619193196,0.2147626728,-0.0545192882,-0.1538217217,-0.2200833708,0.3961577713,0.4616061449,-0.1174414009,0.005523534,0.10436479,0.2650132775,-0.1628609449,-0.194704473,-0.0392282978,0.0907205492,-0.1257073581,0.1508945972,0.3900067508,0.0292396881,0.3952741027,0.1306958497,0.1663772613,0.3614394069,0.0658566505,-0.0800634846,-0.125314787,0.1883155107,-0.091673322,0.190022409,0.0967751741,-0.1025893465,-0.1702999324,0.1979158521,0.0816696659,-0.2172452062,0.0747730136,0.199974522,-0.1187727079,0.2165575325,0.1139418259,0.4971821606,-0.0219785459,-0.1903031468,0.1335120201,-0.279881984,-0.0160766132,0.125905782,-0.0240719486,0.3989955187,0.2772562802,0.1201241314,0.0499581583,-0.1512105465,-0.3175322711,0.0898070037,0.4272669256,-0.4888765812,0.2710611224,-0.2688693404,0.2669654191,0.052004803,-0.2310130149,-0.2184808254,-0.5585610867,-0.138685137,0.4793069363,-0.0087354407,0.1128482893,0.0227441136,0.0532073863,-0.1948923469,0.3044198751,-0.2504921257,-0.3013234735,-0.1368227303,-0.1110471934,0.3174388409,-0.2189763933,0.1903361976,0.0410231017,-0.2808040679,0.161413312,-0.3627012968,0.1494130939,-0.1711054295,0.212898016,0.1890973896,0.0244966391,-0.0382207111,-0.1943714023,0.1044634655,-0.1192779019,-0.2969190478,0.1680939794,-0.194265008,0.0368101113,-0.2712195218,-0.3120834231,-0.4605070055,-0.0989578143,-0.0250403397,-0.1841688752,0.2989388108,0.164873749,0.0333095007,-0.0042152,-0.033041168,0.0822724625,-0.1137121916,-0.2012418807,-0.2248912901,0.0750421882,-0.0986689627,-0.0048671509,0.0550270677,0.1905474663,0.3111845851,-0.3943756819,-0.1212726831,0.0237769987,-0.0814216658,0.2374360263,-0.0792715922,0.340854615,0.3825591803,0.101941593,-0.0926316977,-0.2722113132,0.0375120528,0.0219186805,0.101978533,0.066500023,0.3042426407,0.1053648293,0.6797077656,0.4086302817,0.0721286535,0.3929714561,-0.108514674,0.2568466067,-0.2528787851,-0.501732409,-0.1630624533,-0.3097382784,-0.0505203977,-0.1380239427,-0.1240742877,-0.4375977218,-0.368029505,0.6295220852,-0.4248677194,-0.2594050169,-0.1028579324,-0.3583771288,0.2096515596,0.028438421,0.0086305374,-0.0358935669,0.0070894775,-0.2760520875,0.3160626888,0.0908876061,-0.2581804693,-0.1830148846,-0.3940327764,-0.4978414774,0.3102759421,0.282206744,0.5873106718,-0.0067518819,0.054021813,0.0516171269,-0.1117561162,0.8583497405,-0.5854833126,-0.4201727211,0.240455851,-0.3516251445,-0.2097997069,-0.1688064337,-0.2396241724,0.5918940902,0.4185769856,0.398537755,-0.2252151221,-0.3151477873,0.3050112426,-0.1486110389,-0.0294339824,0.0809886009,-0.4181891382,-0.1203428134,-0.2003679723,0.0380931012,-0.2143436521,0.0938273892,-0.2853966057,-0.0301857498,-0.0720043629,0.0014538943,0.2060996443,0.111216478,-0.0534040518,-0.0441464931,0.0129579902,-0.0271878615,0.1129581183,0.2813891768,0.4119318724,-0.454277724,-0.1205614209,0.1787274331,0.1203821525,0.2637929022,0.2700818777,0.1726480424,-0.1320698112,-0.008581114,0.2180736363,-0.1424892843,0.1053133458,0.3505748212,0.0558373705,-0.5953503251,-0.15591681,0.0195554327,0.4437178075,-0.1576193869,0.5006455779,-0.2792442739,-0.2009059936,0.307248503,0.0240254756,0.7952352166,-0.3570336699,-0.0737249851,0.1223784685,0.0380153917,0.1837303042,0.1191256568,0.0349602588,-0.2109855711,-0.0451516062,-0.0598007888,0.0279016122,0.3046393991,0.071254842,-0.2756308019,0.4917845726,0.1518167108,-0.1268104613,0.0093977274,0.0813791901,0.3517754674,-0.0840343162,0.0386381261,-0.0358656496,-0.0800582096,-0.0291139521,-0.0343141109,0.0512452349,0.1818655878,-0.1677499563,-0.461477071,-0.1657685637,-0.2538037896,0.0696959198,0.1904789805,0.0104359742,-0.0155543527,0.1314590126,-0.0035096828,0.0700660199,-0.1248801425,-0.1123269275,0.0263518523,0.3060845435,0.1026720852,0.0255047791,0.3119353056,0.122100234,-0.2198850662,-0.028698815,-0.301032573,-0.2245673835,0.209878698,0.1740579307,0.0645144507,0.0788516626,0.0688026845,0.0294610634,-0.0752449408,-0.2182411104,0.0113635249,-0.056478586,-0.077633068,0.5352757573,-0.0748242661,-0.4235496819,-0.0209020805,0.4053623974,0.2002405375,-0.0302876327,0.116605103,0.0365920514,-0.0573128127,-0.1606634706,-0.2089412659,0.0354487225,-0.2038571984,0.2944544256,0.0864201784,-0.2173199505,0.2485121191,0.1789842695,0.0880031437,0.3290330768,-0.1933051646,-0.345064193,-0.451895088,0.2483118325,0.0230607968,0.1391384304,-0.0045212847,0.1524775326,-0.0317642577,0.2493462265,-0.1979568154,0.1260655671,-0.1137511805,0.2610041201,-0.0130677037,-0.0183456633,-0.0460454747,-0.0965024531,0.0362858288,0.0321509987,-0.2552461028,-0.0227031615,0.0939757302,0.1423315406,0.0806934983,-0.1288041323,-0.0636431947,-0.006253757,-0.0939160213,-0.2363372743,0.257829845,0.2937585711,-0.1800923347,0.1154794395,0.2336885333,0.206710875,0.1157901213,0.3178339005,-0.1584948003,0.097422339,0.0603504591,0.4072882235,0.0689705759,-0.1501168758,0.1449741572,0.0592921413,0.0115344794,0.0536512434,0.2454900444,-0.2691059113,0.1378958374,-0.0138419336,0.3978846669,0.3953821659,-0.0577862449,0.1139660701,0.1548532546,0.0471147709,0.0295295678,-0.0768276826,0.5202010274,0.1809118092,0.1295005828,0.4198645055,0.3191749454,-0.1101618037,0.3200751543,0.1504372954,0.0289672036,0.0775546059,0.2483988702,0.1680116653,0.0022479652,0.2887734175,0.2247946858,-0.2614195347,-0.1135329902,0.3078891337,-0.0695222691,0.5286960602,-0.0701683015,0.1286160946,-0.0728626326,-0.6161952615,0.1335549504,0.0361777358,-0.4510501027,0.1827563792,0.0080106203,0.2531333566,-0.3910349011,-0.3703324199,-0.2195663899,0.4271663427,-0.2437702864,-0.3405461609,-0.1598486453,-0.1082896441,-0.0566914231,0.0873434395,-0.1157142147,0.1754128486,0.7584131956,0.0234504957,-0.1393943578,-0.2420698702,-0.3078980148,0.0167937428,0.2915057838,0.0562274456,0.5065606833,-0.0626962781,-0.0322502032,-0.1617343426,0.2836578488,0.5438070893,0.6444215775,-0.3587302268,-0.1026917249,0.1454260945,-0.0588432886,-0.1368190348,0.395586282,0.0659537166,0.3027462661,0.2671024799,-0.0589361563,0.0205621552,-0.1275097728,0.2445561439,0.2430919558,-0.0351936854,0.308142364,0.0229520649,-0.047231894,0.2128295749,0.1174853295,-0.3980328739,-0.1747769415,0.4242067039,-0.5149039626,0.2643513083,0.1827848107,0.0437714234,0.0698603541,0.4881626964,0.3512020409,0.1873046458,-0.2638376057,-0.2302778065,-0.4505082071,-0.0517942496,-0.2045984417,0.2934902012,-0.0522416867,-0.0020607342,-0.0752915964,0.1396797895,-0.0422460437,-0.3041798472,0.4681904018,0.3093081713,-0.5451641679,0.12304838,0.0579696819,-0.1336415857,0.0283419508,-0.318873167,0.2898431122,-0.0201225486,-0.1442963481,0.1392993778,0.1771517396,0.0797943994,0.0918298364,-0.001413028,0.2944311798,0.5707865357,-0.0234766118,0.0339960083,-0.2556169331,-0.0185798574,-0.1071838439,0.036086008,0.061275091,0.4165352881,-0.1944236159,0.0195666254,0.0290838368,0.1536452174,0.1593728364,0.2374943048,-0.3469750881,0.1436149478,0.1193897948,0.042502258,-0.177137658,0.3741260767,0.173696667,0.3392372727,-0.337323308,-0.2500857711,0.3955120742,-0.0497901738,-0.1436159164,-0.592513442,0.1727865487,-0.1346159279,-0.0207734928,-0.594837606,-0.2907797396,0.1976029426,-0.2115624696,-0.5768405199,0.3014444411,-0.0852302909,-0.0988461599,-0.0711718723,-0.0490694307,0.1106259152,-0.1035622805,0.2021356076,-0.0847773775]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"It works on my side @timothyjlaurent on google colab.\r\nDid you try to uninstall datasets first, before updating it to master's version ?","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":23,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nIt works on my side @timothyjlaurent on google colab.\r\nDid you try to uninstall datasets first, before updating it to master's version ?","embeddings":[-0.4354046881,-0.0213313587,-0.037903402,0.2603287995,0.1033043042,-0.0524397716,0.2046660185,0.3077459931,0.0026610049,0.1889890134,0.0298017878,0.4250423312,-0.4757447243,0.3481321633,-0.3869570792,0.0821060538,0.0981038436,-0.0260525178,-0.4865825474,0.2010209113,-0.2182771564,0.2801361382,-0.3317174315,0.1242954805,-0.5650153756,0.1543955207,-0.0827687308,0.2532321513,-0.0937662646,-0.5869698524,0.3462376297,0.2859282494,0.0827236399,0.5430799127,-0.0001225756,0.1585456878,0.3996106088,-0.0728133023,-0.1234328896,-0.4519862235,-0.2428455949,-0.2225296944,0.3537067771,0.1101013869,0.1572968662,0.0896834061,-0.0942772552,-0.1369983554,0.2185709327,0.4498015046,0.1162435338,0.1494448334,0.2604189813,0.0649833754,0.0276055876,0.0701150596,-0.0304330997,0.0448700115,0.2290813774,-0.3252796531,-0.0746488497,0.2701995373,-0.1925476491,-0.2442042977,-0.0867799595,-0.2524146438,0.7048782706,-0.6269378662,0.2332711816,0.029663248,-0.1000972167,-0.0182876438,-0.4334952533,-0.1587342471,-0.2270739377,-0.4209839702,0.2868672013,0.0445033088,-0.1510620266,0.1726670414,-0.5132811666,-0.2258363813,-0.0053110146,-0.084507823,-0.1335171908,0.5947369337,0.1725467443,0.2310225964,0.2328880727,0.0781651512,0.2869796157,0.0087678758,-0.139462024,0.1362142563,-0.4196230769,0.0641667694,0.0493193865,-0.053837467,-0.3091590405,-0.0214068759,-0.1810193807,0.0600139126,-0.0766369179,0.0922531784,0.4507497847,-0.1021423489,0.1322092414,0.2775731087,0.2553703487,-0.1338891238,0.0027398535,0.0537189879,0.2788274288,-0.1379814148,-0.0712780058,0.3527615666,0.1793913841,-0.2538718879,-0.157178998,0.1081768796,-0.1086774468,-0.2043838054,0.0778672174,0.1800428629,-0.0431438498,0.7406883836,0.1121363863,0.1405053437,-0.2476970404,-0.0563073866,-0.0780576468,-0.0575003587,-0.4616798162,0.0806494653,0.2490377426,-0.1137034371,0.0169104487,0.0314534046,-0.3257333934,-0.2158705592,0.0585300364,-0.22576271,0.0498758517,0.6758981347,-0.0760784373,-0.0045044385,-0.0882379711,-0.1388807595,0.1334872991,0.2678618133,-0.4715874493,-0.2385824919,-0.2375298738,0.1271954775,0.057777863,0.2616375685,-0.1870081723,0.2372547984,0.4167127013,-0.3055348396,-0.3097494245,-0.342787236,-0.3820715249,-0.281580627,-0.0010893425,0.2608670592,-0.4083782136,-0.0015206833,-0.0218615569,-0.2062689513,0.2619020939,0.1431881934,0.0295007117,-0.0487717651,-0.1298925877,-0.0752124637,0.0204903893,-0.1485111713,-0.1331747174,0.2066442817,0.0465193242,0.2810730934,0.006318138,-0.2740145326,0.2846451402,-0.178880617,0.2588830292,-0.0733952597,-0.1972064078,0.0146349808,-0.4012226164,-0.049829226,-0.105110839,-0.0243016202,0.4020916522,0.2022488564,-0.0971084312,-0.457763195,0.3154033422,-0.0619272478,0.3068206608,0.0584407039,0.1021676362,0.0915568173,0.1256657392,-0.2957902849,-0.4701082408,0.2083050758,-0.1131582335,0.1168778315,-0.0765494332,0.0008789665,0.1114092246,0.1604783237,-0.1929918975,-0.1629264057,0.0659718439,0.0154281938,-0.3101081848,-0.1127226129,-0.212090984,0.658310771,0.0757853687,0.1861138493,-0.1015593559,0.1751902848,-0.1800919771,-0.3921669424,-0.1826357543,0.2310134172,-0.0450028963,-0.1228731275,-0.2446096987,0.4539425373,0.4125307202,-0.0828963295,-0.0622485429,0.1302368641,0.2469462901,-0.1775524169,-0.1807737499,0.1462675929,-0.0164524857,-0.1234347448,0.1983512342,0.4298544228,0.0611753911,0.4018045366,0.1947902739,0.2197562754,0.2534512579,0.0414370708,-0.020723721,-0.2292792499,0.1567949802,-0.0927781612,0.1558268666,0.0785802081,-0.0827602893,-0.2234297544,0.0998938456,0.0262255501,-0.2332392335,0.0764150545,0.0616142265,-0.0839660168,0.155496031,0.1190403402,0.3780833781,0.0167540628,-0.1266406924,0.2620253563,-0.2395739704,0.000429896,0.166756928,-0.0738915205,0.3823560178,0.2637552023,0.0877105594,0.0352480896,-0.1782208532,-0.2680314779,0.0835176632,0.4059555531,-0.4702152312,0.129532516,-0.3255926967,0.2863405943,0.1105974689,-0.153678,-0.3101344109,-0.5067864656,-0.1523490697,0.5275390148,-0.0916635394,0.1440250874,-0.0322289728,0.0580476932,-0.2180202007,0.3890382946,-0.147766009,-0.3442543447,-0.2100168616,-0.0725422427,0.4129150212,-0.2999781668,0.168494463,0.0808765143,-0.2183597237,0.1564550251,-0.297708571,0.0967226177,-0.10171929,0.0640537068,0.1419488937,-0.1134019196,0.0132322097,-0.1293158531,0.0996487066,-0.1607406437,-0.2497457266,0.2247122824,-0.1426769644,0.0508382879,-0.2256073654,-0.32131356,-0.4195499718,-0.1224398687,-0.1029029936,-0.1816918403,0.3723811507,0.1235641837,0.0041529187,-0.0326516591,-0.0467817187,0.0780996159,-0.0517186038,-0.1482006311,-0.1875604242,-0.011597123,-0.1723945737,0.0447336286,0.0954322666,0.148869738,0.3521501422,-0.3665966094,0.0323754475,0.0986392796,-0.1541324556,0.18215774,-0.1341321468,0.2472576052,0.4272384942,0.0512183569,-0.0059532123,-0.2139765471,0.0332701579,0.0367042646,0.0944873914,0.0269596092,0.3396880329,0.108480379,0.6143607497,0.3531054854,0.0673893318,0.3380681872,-0.0831012353,0.1371391714,-0.1871892214,-0.4327064455,-0.1445248127,-0.3046976626,-0.0695492029,-0.167898193,-0.1297107786,-0.3863485157,-0.3299003243,0.5365660787,-0.301944226,-0.2281478792,-0.0869774148,-0.476388067,0.2397387028,-0.0459842347,0.0330855511,-0.0160615966,-0.0340643451,-0.2282374352,0.2794628143,0.0463742875,-0.2925071716,-0.2189249694,-0.3621679842,-0.43940714,0.3069598079,0.1948440224,0.6096240282,0.0758864731,0.0055935015,0.0157366358,-0.1097050905,0.8984593749,-0.5756089091,-0.3717859983,0.29711321,-0.4429655075,-0.1966269463,-0.2272342891,-0.1986096948,0.5277136564,0.3542860746,0.3339732587,-0.0812628865,-0.3351702988,0.2341611087,-0.1970132291,-0.1019083261,-0.0012264983,-0.3534617126,0.038346976,-0.1581908315,0.0760400221,-0.3219123781,0.1301034093,-0.2366093248,-0.1010333449,-0.0990142152,-0.1143918261,0.2294454426,0.1132523194,-0.0070779212,-0.1349930912,0.0776519552,-0.0002367644,0.2038151026,0.3503992558,0.3510702252,-0.4083848596,-0.1374305338,0.2050534487,0.0272702184,0.3153603375,0.2631453574,0.2029474378,-0.1187973097,0.0076829516,0.1836533695,-0.2178966105,0.0136351623,0.3250062168,-0.0021273945,-0.6167440414,-0.1871480495,-0.0275554024,0.4484910071,-0.1468028277,0.4348120987,-0.3938328624,-0.3060362935,0.3708006442,0.1220897436,0.7906650901,-0.3841534555,-0.0390991457,0.0074553401,-0.0294349343,0.175647527,0.0881974921,0.0723472387,-0.3037419617,-0.0416939408,-0.0644234866,0.1050786898,0.2288899124,0.1262172312,-0.2223207057,0.4922861159,0.0313615389,-0.0393866338,0.0081446338,0.1416226774,0.3827841878,-0.0575260371,0.1069019884,0.0146028539,0.0199213754,-0.1362499595,-0.0200799368,-0.0057204128,0.2628923059,-0.066271767,-0.3939189911,-0.0826958418,-0.2232789248,0.1065090597,0.1593918353,0.0691582114,0.0179594513,-0.0173596032,-0.0089868437,0.087547563,-0.0935938507,-0.121605359,0.06405738,0.3566723764,0.1275506765,0.0769577324,0.3548250198,0.0887997001,-0.2294704318,0.0196582023,-0.2300751209,-0.229510814,0.0692367107,0.0351720043,-0.0158813596,0.1362736374,0.0685127974,-0.0160010215,-0.134108454,-0.1983484924,0.0759320557,-0.0742574856,-0.1341287047,0.4565622807,-0.0208648909,-0.4040488601,-0.0150763486,0.4753894508,0.2630131245,-0.0304916948,0.1754943728,0.1332140118,0.0530629121,-0.2314902395,-0.1961837411,0.2538499534,-0.1257029176,0.3030168116,0.0551358834,-0.273740828,0.3017933965,0.1004654318,0.1316060424,0.401843518,-0.2024610192,-0.3165369928,-0.5207990408,0.2257960588,0.0114972964,0.0770637542,-0.1637192369,0.2023658156,-0.0716061592,0.2509379983,-0.2399387211,0.1434378028,-0.1084161848,0.1871653348,-0.0967848673,0.0084109893,-0.1806605905,-0.17707026,0.0653354898,0.0199866947,-0.260219276,-0.0486227311,0.0486737788,0.1250516921,0.130723998,-0.120311074,-0.1024015546,-0.0336570628,-0.1377159804,-0.1636200696,0.2318588346,0.2656706572,-0.1671265662,0.1718910038,0.1938103288,0.0712786764,0.0635041147,0.3051727116,-0.1433326453,-0.0388203822,0.0464690104,0.4356119037,0.1066291481,-0.182919398,0.0858234093,0.0566577353,-0.1773483902,0.0550203212,0.2996478677,-0.2549988031,0.0197277125,0.0032343275,0.4577387571,0.4496193528,-0.0286767241,0.0523707233,0.1740736961,0.1175023839,-0.0192276351,-0.0907893106,0.4671577513,0.1339300275,0.0730500221,0.2822966278,0.285706073,-0.1302193552,0.170700863,0.2068179101,-0.1275929213,0.0618724525,0.2945271134,0.3045637012,0.1483749896,0.2929367423,0.2344320863,-0.2614428699,-0.1006951779,0.4330565333,-0.0023198863,0.6118899584,0.0232094862,0.0663693771,-0.0421509817,-0.5977569818,0.112870805,0.1524534672,-0.4710716903,0.1441162974,-0.02790571,0.2607676089,-0.361738801,-0.5522366762,-0.227829203,0.3568048775,-0.261992842,-0.3144665062,-0.2318210304,-0.1654921919,-0.1205222085,0.091098547,-0.114941366,0.1981247067,0.7187823653,0.0533016659,-0.0966015905,-0.2032718211,-0.4901356697,0.0202735364,0.2725242376,0.0038027358,0.3858621418,0.0100093251,-0.0448270291,-0.0597179458,0.2000953704,0.4813216329,0.7049614191,-0.4624646306,-0.0323620103,0.1701081544,-0.0370740369,-0.071241267,0.3653138876,0.1443501115,0.2370886356,0.2170631588,-0.0534210168,0.0103764711,-0.0988132209,0.2725346088,0.2391126901,-0.0500678346,0.3155852854,0.1682040542,-0.0819731355,0.2457941324,0.1070546806,-0.3547039628,-0.2583478391,0.5720741153,-0.4759394228,0.2169442624,0.1774109006,0.0621652044,0.0866575018,0.5702397227,0.3672221601,0.1261858493,-0.2507759929,-0.236335516,-0.4916426539,0.0022981011,-0.1355818361,0.3526295424,0.0493117794,-0.0417464823,-0.0722346902,0.0806352124,-0.0273710415,-0.1785648912,0.3432694972,0.316331923,-0.5723845363,0.1845902801,0.0346430801,-0.077348657,-0.0028168422,-0.2534410655,0.3371368349,-0.0350740477,-0.1019873843,0.1381351501,0.186068356,0.1227261648,0.1683132499,0.0822113678,0.2360449731,0.5106275082,0.0284205135,0.0158381853,-0.3051035702,0.0315123685,-0.1612470001,-0.0157732852,0.0998519436,0.4244027138,-0.1579049379,0.0052359942,-0.0106178876,0.2532088757,0.0556070805,0.2211031318,-0.4375537038,0.2328777164,0.2318993807,0.0851716176,-0.138164714,0.4167876244,0.1708544046,0.3439673781,-0.3627950549,-0.2012809813,0.326341033,-0.043199487,-0.1527235806,-0.6441495419,0.19367145,-0.1285030097,0.024587363,-0.5834943056,-0.2899975181,0.1339089572,-0.1852582097,-0.5932970643,0.2438433319,-0.0322695114,-0.0596765913,-0.074753955,-0.2190410793,0.1366812587,-0.1838207692,0.146508351,-0.0944253504]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/620","title":"map\/filter multiprocessing raises errors and corrupts datasets","comments":"I didn't -- it was a new sessions --- buuut - look like it's working today -- woot! I'll close this issue. Thanks @lhoestq ","body":"After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```","comment_length":24,"text":"map\/filter multiprocessing raises errors and corrupts datasets\nAfter upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing.\r\n\r\n```python\r\n ...\r\n ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n ner_ds_dict[\"validation\"] = ner_ds_dict[\"test\"]\r\n rel_ds_dict = rel_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed)\r\n rel_ds_dict[\"validation\"] = rel_ds_dict[\"test\"]\r\n return ner_ds_dict, rel_ds_dict\r\n```\r\n\r\nThe first train_test_split, `ner_ds`\/`ner_ds_dict`, returns a `train` and `test` split that are iterable.\r\nThe second, `rel_ds`\/`rel_ds_dict` in this case, returns a Dataset dict that has rows but if selected from or sliced into into returns an empty dictionary. eg `rel_ds_dict['train'][0] == {}` and `rel_ds_dict['train'][0:100] == {}`.\r\n\r\nOk I think I know the problem -- the rel_ds was mapped though a mapper with `num_proc=12`. If I remove `num_proc`. The dataset loads.\r\n\r\nI also see errors with other map and filter functions when `num_proc` is set.\r\n\r\n```\r\nDone writing 67 indices in 536 bytes .\r\nDone writing 67 indices in 536 bytes .\r\nFatal Python error: PyCOND_WAIT(gil_cond) failed\r\n```\nI didn't -- it was a new sessions --- buuut - look like it's working today -- woot! I'll close this issue. Thanks @lhoestq ","embeddings":[-0.4389111698,-0.1553174555,-0.0303596556,0.2238802314,0.1435325444,-0.1580417901,0.3082613945,0.3411384821,0.1361238062,0.1270737648,0.064798221,0.3830269277,-0.4311425686,0.3586114347,-0.3158304095,0.0192878153,0.1044041142,-0.0822964981,-0.3726718426,0.2437452078,-0.2811784446,0.2419260889,-0.3431835473,0.1499836445,-0.5202993751,0.0915086567,-0.0610799044,0.2870166004,0.000186195,-0.5671350956,0.3406110704,0.2375900447,0.0439229496,0.4339787662,-0.0001241726,0.1413946152,0.3641113043,-0.1141221449,-0.0690170452,-0.342663914,-0.2272138447,-0.2039410025,0.2903558612,0.117430985,0.191928193,0.0396893807,-0.1186003909,-0.2761327326,0.2109070867,0.428350538,0.1021492332,0.2034933418,0.2628211379,0.1077969521,0.0287278574,0.1473955661,-0.0351305194,0.0102035301,0.3186465502,-0.3860552907,-0.0874852315,0.263018012,-0.254226476,-0.1928121299,-0.1764389873,-0.2241109312,0.6580686569,-0.5945964456,0.1860852391,0.0570210777,-0.1738320887,-0.0090262825,-0.4382345378,-0.1861442477,-0.2577209771,-0.5019298792,0.2167701572,0.064054653,-0.1605045348,0.1461489052,-0.4861242473,-0.2162235826,-0.0459571183,-0.0600842275,-0.0794198811,0.526899159,0.2126140147,0.3030981123,0.166735962,0.1282711774,0.2003577054,0.0367891975,-0.107438609,0.1688026041,-0.4264659286,0.0932108015,0.1012027264,-0.0469355881,-0.2404868603,0.0284216087,-0.196073845,0.0912130326,-0.0908493996,0.0443773977,0.4214399159,-0.1040274724,0.2137225866,0.297825247,0.2598975599,-0.2024503052,0.0061802757,0.040315412,0.3496516645,-0.1308580339,-0.0845075995,0.3169029355,0.1554721147,-0.1958630085,-0.1778360158,0.1278182715,-0.0874687806,-0.2110477537,0.0750089139,0.193967849,0.0216418114,0.7758275867,0.1422030926,0.1933227479,-0.325258255,-0.1087506413,-0.0756798685,-0.0402470008,-0.3965352476,0.0837784857,0.2747548223,0.0335661359,-0.0190135147,0.0139808906,-0.3854869604,-0.2746291161,0.1181668565,-0.2540867329,0.1059900075,0.6713737249,-0.0934601873,0.0071930033,-0.03516322,-0.1314365119,0.0976828188,0.2614506781,-0.5445335507,-0.2043902725,-0.225669384,0.0868669897,0.0191376135,0.2917738855,-0.1509467363,0.2350901365,0.4053868651,-0.3626489937,-0.3123522103,-0.3348920345,-0.34825477,-0.2950326204,-0.0426826887,0.2990987599,-0.3928132355,-0.0415192172,0.0763777643,-0.2193003297,0.306273222,0.2014248669,-0.0193768181,-0.0061092004,-0.1035313383,0.0675832331,0.0662998483,-0.207234785,-0.108417809,0.2526930869,0.0651985481,0.3254318237,-0.0205413736,-0.2612626851,0.3218259811,-0.1641808897,0.240795359,-0.077146858,-0.1891902387,0.0664506033,-0.4434908628,-0.0198268779,-0.0087880632,-0.0699773058,0.3942247629,0.1964439303,-0.1109895408,-0.4963757694,0.3344274163,-0.1099407226,0.3142298758,0.0408367217,0.0857003853,0.0698186755,0.144445017,-0.2950135171,-0.482003957,0.1784681827,-0.2608712912,0.0483104698,-0.0037992212,0.0216593929,0.0957438424,0.1924470067,-0.1963997632,-0.2043780684,0.0327275656,-0.0337023139,-0.315082103,-0.0680290833,-0.2182216793,0.6899927258,0.1175851524,0.131913051,-0.1111836284,0.2003579736,-0.1506473571,-0.410149157,-0.2707152367,0.2542414963,-0.0004456755,-0.0678377077,-0.1897198707,0.4699810445,0.4786747098,-0.1128505841,-0.0575907826,0.1334883124,0.2497288585,-0.1686914861,-0.1394185126,0.0356437862,0.0225014333,-0.1713116169,0.2028626353,0.4330973029,0.1223600507,0.4227701426,0.2035799921,0.2512256503,0.2895075083,-0.0033044198,-0.0225608218,-0.1272280067,0.2133682817,-0.169662863,0.1399856955,0.0221167691,-0.1458357275,-0.239419058,0.0672901869,0.0190604422,-0.1369379461,0.0566276088,0.0696791783,-0.0945376456,0.1550387144,0.0438794903,0.4365884662,0.008224993,-0.208840996,0.17390351,-0.2554391325,-0.0099013038,0.1101606414,-0.0709502548,0.4103217721,0.2982468307,0.0862267837,0.0025430527,-0.155993849,-0.3185812533,0.070630677,0.3644186854,-0.5294398665,0.1767967641,-0.2681471705,0.3429842591,0.1035520285,-0.1245050207,-0.284629494,-0.5637062788,-0.107589379,0.5049539208,-0.0873982012,0.1322872192,-0.081022948,0.0878848657,-0.2123987973,0.3743552864,-0.1540359855,-0.3298156261,-0.1779154539,-0.0936025903,0.3544712663,-0.2819004357,0.2225762308,0.1541030705,-0.2573418021,0.1878952235,-0.3045693934,0.0626317486,-0.0436366238,0.0536772646,0.1156656668,-0.0004303984,0.0411712565,-0.1813838333,0.1464889646,-0.1574028134,-0.2222725302,0.2964487672,-0.1393526495,0.0939636976,-0.2634342313,-0.3453348875,-0.4375620186,-0.1150623783,-0.0753434226,-0.1841026992,0.3403443396,0.1291127503,0.030945722,-0.029971648,-0.0697862059,0.0478993617,-0.0658791587,-0.0593994111,-0.2077104598,0.0360214114,-0.1212216094,-0.0526950434,0.0805642009,0.1198307574,0.4075064063,-0.3314512074,0.0480158068,0.0899490044,-0.1362224072,0.2196607292,-0.1817476749,0.3469201028,0.410412699,0.0584697798,-0.0340174958,-0.1860706657,0.0954233557,0.0386616513,0.0524459779,0.0951317549,0.3037011325,0.1426620185,0.5477454066,0.3771878481,0.1038324311,0.384401232,-0.0821647644,0.1198285967,-0.202344507,-0.3933930993,-0.1818884909,-0.3160487115,-0.0118764536,-0.2425557226,-0.0927242488,-0.4461107254,-0.3504076004,0.5543606877,-0.3758941591,-0.2395128608,-0.0468255505,-0.5213514566,0.2646471858,-0.083806321,0.0509237275,-0.0283501819,-0.0298486948,-0.2563850582,0.1997668147,0.1217582226,-0.3251347542,-0.2877398133,-0.2596097887,-0.4335175753,0.3148641288,0.2012095302,0.6812641025,0.0003933008,-0.0801635832,0.0085703162,-0.1337454617,0.8955234885,-0.6052264571,-0.3748029172,0.3144158125,-0.3783261478,-0.2645667791,-0.162588492,-0.1791370511,0.5422021747,0.3443920016,0.3918618262,-0.1067367196,-0.2833149731,0.2422168553,-0.2013667971,-0.100880824,0.0209742524,-0.3194952309,-0.0250378326,-0.1747475564,0.03787205,-0.2602269351,0.0839427859,-0.237989381,-0.0681499392,-0.1017658859,-0.0635964572,0.255875051,0.0516586453,-0.0036192581,-0.2103956491,0.0366047285,0.0464498624,0.1680696309,0.273856014,0.2778927684,-0.4287841022,-0.1004448533,0.1635333598,-0.0033510413,0.3238947093,0.2844062448,0.2353754342,-0.12338496,0.0471062213,0.1985632777,-0.2943037748,-0.007224563,0.3591537774,-0.0389200374,-0.5711473227,-0.1325413436,-0.0589139536,0.4101545215,-0.1745319813,0.5028032064,-0.4040952325,-0.2674122751,0.3955002725,0.0618786365,0.7986565232,-0.3684019446,-0.0315077417,0.0784261599,0.0157469157,0.096581392,0.0716537163,0.0351687633,-0.2841925919,0.0027696232,-0.0598041788,0.0394074917,0.3132289052,0.1772978157,-0.1591722667,0.433293134,-0.0048889397,-0.0966085717,-0.0331121273,0.1443445683,0.4444709718,-0.0378570184,0.0888577029,-0.0030983374,-0.0598698407,-0.1694535017,0.0205887277,0.0446931832,0.2799759507,-0.1480804682,-0.4104651213,-0.0688648447,-0.110800527,0.1305981576,0.2134426087,0.1203013286,-0.0266283099,0.1129795909,-0.1301922351,0.0459450521,-0.0973097533,-0.122939527,0.0945897773,0.3547369242,0.2294875383,0.0524013191,0.3715539277,0.0996862724,-0.2941745818,0.0591278113,-0.2497494072,-0.2445216924,0.0230628941,0.074578099,0.0911555365,0.0852727965,0.113917686,-0.0113202874,-0.1307717562,-0.1835198253,0.0477007516,-0.0449272208,-0.1178248078,0.5261170864,-0.0499873385,-0.3697991669,-0.0020985359,0.4567427933,0.2305098176,-0.1154204533,0.1137665883,0.1260139793,-0.0325596072,-0.2368058562,-0.1776185036,0.2296192944,-0.1732086241,0.2777266204,0.0771603286,-0.2246045917,0.2320952863,0.0277601685,0.0488728024,0.3711372614,-0.1884256452,-0.3409450054,-0.4504369497,0.2852000594,0.0569391288,0.0860524699,-0.1097672284,0.1406216621,-0.0625022203,0.28600806,-0.221798569,0.1086066514,-0.0434016436,0.2306844443,-0.0811153278,0.0209754091,-0.1002751663,-0.1217826977,0.0783503428,-0.016135212,-0.222209841,-0.02997371,0.0373681858,0.136416465,0.0792154148,-0.2170054317,-0.0580647662,-0.0694184974,-0.1116668582,-0.1975589842,0.170011878,0.3778556287,-0.2157562375,0.1627099961,0.2334394157,0.1576212347,0.1360860765,0.3367206752,-0.218963027,0.0361099355,0.0200930908,0.4062175155,0.1348040104,-0.1911143512,0.048034057,0.1183557287,-0.0878900588,0.0132838851,0.3321454227,-0.2417400479,0.0466573797,0.002765158,0.4242886007,0.4040002525,-0.124279283,0.1138933003,0.1284148097,0.0989409611,0.0134859988,-0.157008186,0.5547395945,0.1114389598,0.0611254312,0.3319323063,0.2720370591,-0.1747553349,0.2071678191,0.1542920768,-0.0786119848,-0.076816678,0.254914254,0.1522733718,0.0538589619,0.2410249561,0.1996329427,-0.3166682422,-0.1305838525,0.3212500811,0.010067475,0.4989767373,-0.0399586894,0.1095738411,-0.1257620305,-0.6536566615,0.0496174954,0.1127596051,-0.5217781067,0.2226065248,-0.0324496515,0.2831968963,-0.3367138803,-0.4498470128,-0.1828360111,0.3819921017,-0.2711340189,-0.3221406341,-0.2267083228,-0.1483098119,-0.0997902825,0.0662750006,-0.1178686246,0.1854425967,0.801915586,0.0322909504,-0.0746422857,-0.1906512231,-0.3439487517,-0.0163974967,0.314108938,-0.019810196,0.4053681493,0.0181751698,0.0303027388,-0.0187587719,0.1585478634,0.4757374227,0.7030801177,-0.4632856846,-0.0224412363,0.1462654024,-0.0747058243,-0.0866851211,0.4076161683,0.1262261271,0.331307143,0.2883501947,-0.0535501316,0.0369687453,-0.1178287044,0.3061879575,0.3639971018,0.0329927392,0.3444562256,0.0636600852,-0.1106197834,0.2326204181,0.0831907615,-0.3255212903,-0.3061917126,0.5170181394,-0.4688683152,0.2489011586,0.1927312165,0.0471717641,0.1036801934,0.465349704,0.288744092,0.1077416018,-0.2872965038,-0.1678347439,-0.4076788127,-0.0197939295,-0.1969337612,0.3094287813,-0.0087308353,0.0166617129,-0.0470239036,0.047243394,-0.0202065594,-0.2586666346,0.3506315947,0.295622468,-0.5638790131,0.1922887266,0.0746317282,0.0110439016,0.0230899584,-0.3029471338,0.3537074625,0.0172166433,-0.1114221513,0.1150784492,0.1453231424,0.1238250583,0.1865614206,0.0095808003,0.2730889022,0.5970849991,-0.0058609061,0.0081697265,-0.2979804575,0.0210414585,-0.0924886763,-0.0315574035,0.0570214763,0.4098131359,-0.2148821354,-0.0458690822,0.0307063051,0.1876495034,0.0146622099,0.2140069604,-0.4316045046,0.196834147,0.2346977144,0.0222959742,-0.1508830637,0.3642426133,0.1588985026,0.3398591578,-0.2817399204,-0.1986920089,0.2920212746,-0.0160546247,-0.1066015065,-0.6062425971,0.2379884571,-0.1984817386,0.1442448497,-0.5549508929,-0.3425116837,0.1964059621,-0.1812261492,-0.5432785749,0.2515481412,-0.1097087637,-0.1704103798,-0.0524099357,-0.2285161018,0.1160430834,-0.1258507073,0.1942717433,-0.0692155361]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/619","title":"Mistakes in MLQA features names","comments":"Indeed you're right ! Thanks for reporting that\r\n\r\nCould you open a PR to fix the features names ?","body":"I think the following features in MLQA shouldn't be named the way they are:\r\n1. `questions` (should be `question`)\r\n2. `ids` (should be `id`)\r\n3. `start` (should be `answer_start`)\r\n\r\nThe reasons I'm suggesting these features be renamed are:\r\n* To make them consistent with other QA datasets like SQuAD, XQuAD, TyDiQA etc. and hence make it easier to concatenate multiple QA datasets.\r\n* The features names are not the same as the ones provided in the original MLQA datasets (it uses the names I suggested).\r\n\r\nI know these columns can be renamed using using `Dataset.rename_column_`, `questions` and `ids` can be easily renamed but `start` on the other hand is annoying to rename since it's nested inside the feature `answers`.\r\n","comment_length":19,"text":"Mistakes in MLQA features names\nI think the following features in MLQA shouldn't be named the way they are:\r\n1. `questions` (should be `question`)\r\n2. `ids` (should be `id`)\r\n3. `start` (should be `answer_start`)\r\n\r\nThe reasons I'm suggesting these features be renamed are:\r\n* To make them consistent with other QA datasets like SQuAD, XQuAD, TyDiQA etc. and hence make it easier to concatenate multiple QA datasets.\r\n* The features names are not the same as the ones provided in the original MLQA datasets (it uses the names I suggested).\r\n\r\nI know these columns can be renamed using using `Dataset.rename_column_`, `questions` and `ids` can be easily renamed but `start` on the other hand is annoying to rename since it's nested inside the feature `answers`.\r\n\nIndeed you're right ! Thanks for reporting that\r\n\r\nCould you open a PR to fix the features names ?","embeddings":[0.2796131968,-0.0672781244,-0.0831842795,0.1901165694,0.2244755626,0.3354707658,0.5049102902,0.1621009558,-0.2027406991,-0.0173162166,0.1145086139,0.189746365,0.358310312,0.4374582171,-0.0695425272,-0.2454162985,0.2463854551,-0.0470877476,0.2421891391,-0.0504073612,-0.2164766192,0.2113196552,-0.0955329239,0.6324017048,0.0162046198,0.0617662631,-0.0821974203,0.0513545536,0.1023243293,-0.2556810081,-0.3580749631,0.0417952128,-0.1288834363,0.0629301444,-0.0001121551,-0.5027588606,0.3252595067,-0.0481603742,0.0559792109,0.195929721,-0.4208875597,-0.4244924188,0.0467999913,-0.3844260871,0.2206073552,-0.2236407399,-0.2441880554,-0.1287856698,-0.1785168946,0.1382639706,0.2021540403,-0.0150762191,0.2414555103,0.0380664207,0.156750083,-0.0216935109,-0.2375083417,0.0635238811,0.1898536682,-0.3023804724,0.3947696686,0.4994888604,-0.2198397964,0.1615970284,0.2153594345,0.1484520584,0.4425870478,-0.4972687364,0.2903123796,0.2118736655,0.3128450811,-0.1872746497,-0.2283304334,-0.2488459945,0.1180141941,-0.4643637538,0.1259203851,0.1799848676,0.0420384631,-0.2551813722,0.0971525013,-0.2348621786,0.24766168,-0.0904301181,-0.1196242496,-0.1861042827,0.0913874209,0.103745617,-0.2814544141,-0.3552182913,0.3920393288,0.1443683803,-0.1899144053,-0.0709320381,-0.44919613,-0.1471944898,-0.0656530187,0.1372911781,-0.183368057,-0.0760083646,-0.2866718471,-0.1546825767,0.2513923645,-0.0729935318,-0.1349556446,0.2213105261,0.2043882757,-0.1287403554,0.2029762715,-0.0088791093,-0.1400403678,-0.0656036437,0.0924017653,-0.0184069425,-0.4182289541,0.2236414552,0.1739640385,-0.2739310861,-0.318123132,0.3076445162,-0.4477353394,-0.1837328821,-0.1012836769,0.1104021817,0.058475215,0.1314394772,-0.0421452299,-0.0141479019,-0.1282830089,-0.236476317,-0.016901277,-0.0350200832,-0.2116493136,0.288821429,-0.1696249694,0.312384665,0.2091035396,-0.3090595007,0.0306379795,0.0261928122,0.1595200598,0.244290188,-0.1613461375,0.1271407902,-0.1375322193,0.0108513646,0.077898398,0.0282534733,-0.0115864566,0.154052034,-0.4452196658,-0.0020242357,-0.3583761454,0.2011039853,0.2075180262,0.0912626609,0.5391229987,0.3284410536,0.2340393215,-0.1146792397,0.101470165,0.0638985857,0.4210463464,-0.0761173815,-0.037354596,0.3379053175,-0.2224426419,-0.2421078682,-0.2108462602,0.2244683057,0.0717405602,0.2689327896,0.1512112468,-0.2363493741,-0.0420139469,0.1127992198,-0.0634274334,-0.1214621142,0.0940467492,-0.00918076,0.0562501661,-0.4241338372,0.1855673194,-0.0839033425,-0.2598454058,-0.1348484755,0.0801539049,-0.0857665017,-0.1918994933,-0.0042833462,-0.2674156129,-0.3934065998,0.2177091241,-0.0340570621,-0.1972797364,0.2775914073,-0.1092591435,-0.372240603,0.0812074021,0.0164585207,-0.2207569331,-0.2087622881,0.2250334471,-0.0180336181,0.1622039825,0.1617988348,-0.4457290173,-0.2008742988,-0.4516760409,0.0267479271,0.1582198143,-0.0885870978,-0.1440156549,-0.0856709853,0.1257941425,0.0197070986,0.1156930104,0.1000402942,-0.3079137206,-0.0546972193,-0.2999891341,-0.0695870221,0.0166508127,0.0884377584,-0.2995299101,0.0845690593,0.006773775,0.0245156139,0.0497617871,0.3868393004,0.2614257336,0.2171546817,-0.1582842767,0.2827342153,-0.1130648404,0.1268773079,0.2862364054,0.3946857154,0.1731519848,0.0114114676,0.1446131617,0.1174112111,-0.0964709371,0.2288684994,-0.1158267558,0.2167550623,0.08290869,0.2774330378,0.0137833301,0.3125977218,-0.0607790686,-0.3890331089,-0.1617789865,-0.2932703793,0.2806172669,0.1219407544,-0.2302554846,0.0286681671,-0.4353847206,0.3081432581,0.4657300711,0.0086177075,-0.0998735279,-0.1286444962,-0.0286652073,0.019542262,0.3403352201,-0.133849442,0.2067207247,-0.0138759902,-0.3927894235,0.0305581223,0.2801202238,-0.170256272,0.1206268668,0.3795162439,-0.4961919188,-0.0308025293,0.399510622,-0.1206594482,-0.1075789556,0.2712964416,0.3922182322,-0.1185784265,-0.3020518422,0.1671589911,-0.0590958185,-0.4547241032,0.0114507945,-0.4382162094,-0.2242616266,-0.3971214294,0.2313409299,-0.1345754862,-0.1287674904,0.4221910834,0.3064456582,0.3904313445,-0.1393619031,0.5181222558,-0.2536642849,-0.0292948205,-0.0275446456,0.0332499146,-0.0914422944,0.2400507331,0.1717616767,-0.0654342473,-0.0571082905,-0.4989719093,-0.4638471901,0.136767298,-0.1452696472,0.5637741685,-0.0681429952,-0.248379752,-0.1504255831,-0.2278665006,0.181530714,0.4881224334,0.0109627135,-0.0555702411,-0.0401342958,-0.0752229914,-0.1130447835,-0.4911833405,-0.1803040504,-0.1180189773,0.1634392589,-0.1793171316,0.1086950079,-0.0388177298,-0.6115885377,-0.2516984046,0.0535071939,-0.1561785787,-0.3134723902,0.0295977537,0.1696384102,-0.2660049796,-0.0662297085,-0.0142088495,0.114947632,0.0425457731,-0.4492481351,-0.2512800992,-0.440128237,-0.0633307248,0.1214749515,0.3817400336,-0.048870448,0.0095455674,0.2082291692,-0.0724832341,0.1266065687,-0.4921844006,0.0524862967,0.5331341624,0.6120025516,-0.0265210047,0.4932017624,-0.3559103608,0.3424315751,0.0715392604,0.3521508574,0.0787941664,0.0839557871,-0.278308332,-0.0198826902,-0.2667330801,0.4488635063,-0.090237312,-0.1634920239,0.1174073815,-0.1591564566,-0.0831244737,-0.1478431225,0.6585765481,0.0095528327,-0.1561867446,-0.3136458397,-0.546618402,0.5313019753,0.1734657884,0.2710705101,-0.1492184401,-0.1392339319,0.3164743781,0.083672598,0.2949906588,-0.1113810465,-1.016127944,-0.6102489233,-0.1619787961,0.2193025798,-0.1704930812,0.3395325243,-0.0004180623,-0.1577079743,0.0694309697,0.2542560697,0.3741117716,0.0146666896,-0.5943673849,0.4133254588,0.4034888446,0.3657948971,-0.4275993705,0.0984599367,0.3059115112,0.040393021,0.0077427058,-0.3111066818,-0.1033150852,0.195250988,0.0858672336,-0.4044755101,-0.0101801818,-0.2013160735,0.003373092,-0.2211666256,-0.17501688,-0.227051869,0.0023751946,-0.1999277622,0.0937116221,0.1762762964,-0.1498912871,0.2022976726,0.5323475003,0.2188075483,0.0430123955,0.3364474177,-0.0893387645,0.1986302286,-0.1301907599,0.1045793295,-0.1586585492,-0.4007582366,0.1648655683,-0.4002197385,0.4479199946,-0.0035482824,-0.0747099817,0.3443922102,-0.1640865058,0.2891595066,-0.0436086319,-0.1024332866,0.4039686024,0.0920990631,-0.0113360221,-0.2493542284,0.3904542625,-0.1281461269,-0.1608403027,-0.0669844225,0.1257349551,0.0692488104,0.2802542448,0.0964806974,0.8481325507,0.0873071179,-0.3693995178,-0.0362952426,-0.0430645794,0.3490489125,0.2155731171,0.1329014003,-0.2732764482,-0.2228893489,0.1439788789,0.0114854127,0.0663692132,-0.0088117411,-0.0859190673,0.437845856,-0.2496571392,-0.1334144026,0.0046632318,-0.151967898,0.3291784227,-0.0919128656,-0.1032706499,0.00130365,-0.0551790334,0.0320165828,0.0673653334,0.0134253306,0.153214097,0.2280713022,0.0703620911,-0.2106656432,-0.1434587091,-0.2633118927,0.2165816128,-0.4240284562,-0.0710135922,0.1545516849,0.5141527653,0.1313969344,-0.2693907619,0.1191040725,0.4850041866,0.1883931458,0.0794615746,0.0329505205,-0.0805016086,-0.0658364668,0.0872921199,0.3662008345,0.100315392,0.1588852853,-0.8413519263,-0.0958494842,0.4370794594,0.0240396727,0.0303876754,-0.1506916136,0.199559167,-0.2023027986,0.1225018948,-0.0251039211,0.0037182942,0.2275635451,-0.1892841607,0.080350168,-0.0721300095,-0.1647093743,0.0514403135,0.0719822645,0.3237184286,0.2217674851,-0.0306719542,-0.0739218071,-0.0580770187,-0.2502267659,-0.1972156316,0.0762856603,0.1585503668,-0.4756890237,-0.1611995697,0.2974226177,-0.0639711469,0.2787192762,0.0751365796,-0.2167111337,-0.2680914998,0.3928009868,-0.0263237432,-0.1625745445,-0.0601109304,0.1347094923,-0.3619488776,0.1015697867,-0.2566002309,0.1478848755,-0.23831366,0.1147255823,-0.0440849401,-0.1134197786,-0.175256744,0.016719358,0.0728173479,-0.2535341382,0.1008443385,-0.0949607417,0.0469542816,0.158400014,0.0966844782,-0.0210421812,-0.0975233912,-0.2031390369,0.0063378289,-0.0154474555,-0.0230280384,-0.0880545527,0.1500523686,0.306299299,-0.227608934,-0.1806183606,-0.0943460688,0.2853701711,0.3514631987,-0.177541241,0.2365429848,0.1413549632,-0.1868919581,0.1068338975,-0.0975228697,0.1227636784,-0.3421072066,0.074615635,0.3677304089,-0.1180884838,-0.0392939374,0.1700601578,-0.1394600123,0.4927856028,-0.2133514434,0.0260656923,0.3135805726,0.1641212255,-0.3212099373,-0.2611582577,-0.3593670428,-0.0353641883,-0.0426317453,0.0828726739,0.4474897087,-0.1211829185,0.0169705674,0.1829270571,0.2554732561,0.0731597841,0.0739677176,0.3740552962,0.0700124651,0.4004065394,0.5137106776,-0.0227790643,-0.3062428236,0.1996742785,0.2639440894,-0.0143205402,0.6973946691,0.184651047,0.3648731112,0.2844782472,0.0337651186,-0.2174209505,0.1698569208,0.104714945,0.1198455542,-0.0501658581,0.0078805713,-0.440472573,-0.1537647247,-0.0058658663,-0.0703190342,-0.3327952623,0.0734068975,-0.0668125004,0.0654351413,-0.0247056484,-0.280800581,-0.1099703908,0.6754775643,-0.136412859,-0.1822154075,-0.152316764,-0.1406022757,0.4321360886,0.4079167545,-0.0529470779,0.10723795,-0.0599153228,-0.0789159685,0.2173178494,0.0980094075,-0.0897988379,0.1774738431,-0.4331849515,-0.2410427332,0.0076308441,-0.1011978611,0.3845257759,0.1653137952,0.1407289803,0.3100249469,0.1142037734,0.050196819,-0.0555058904,0.1272185296,-0.0078623975,0.1683947146,-0.4091054201,0.102230534,-0.3346336186,-0.0675831139,-0.1973170191,-0.1512133181,-0.2300191224,0.0502036586,0.194010362,0.0131134773,-0.0536456108,0.1401582658,0.0819327533,-0.3603378534,0.3539222181,0.2670244873,0.05645537,-0.2570889294,-0.1675048321,-0.3037042618,0.123196274,0.4289314449,0.3278815746,0.3765978217,-0.0242367554,0.098928012,-0.164277181,-0.3755149543,-0.1622452885,-0.1620221436,0.1337087899,-0.6391876936,-0.3137764633,-0.0577578358,0.2944856584,-0.0273586679,0.2157590687,0.4554787874,0.3199615777,-0.1616204232,-0.0226118639,-0.3401670754,0.3182308674,0.68069309,0.1832422167,0.2632365227,0.1262327433,0.0127973119,-0.2782557309,0.0611338168,-0.4556921124,-0.0571736693,0.0371971503,-0.1770212799,-0.0471637025,0.3318123221,0.2001005858,-0.1217118502,0.3218123615,0.119296208,0.0384386554,-0.4937000573,0.2479211837,0.4585892856,0.040730916,0.0988557935,0.0653117076,0.1100179106,0.3921360373,-0.0505034961,-0.2685379982,0.6965746284,-0.0760933757,-0.0218541529,-0.0188633949,0.4410728812,-0.0197553746,-0.1611001343,-0.4324763119,-0.4226664901,0.3406488001,-0.3558457196,-0.1867022812,0.2179405689,0.1376256496,-0.1097855642,-0.0341533683,0.1473587304,0.3316616416,-0.145902887,0.1566539705,0.1557341069]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/617","title":"Compare different Rouge implementations ","comments":"Updates - the differences between the following three\r\n(1) https:\/\/github.com\/bheinzerling\/pyrouge (previously popular. The one I trust the most)\r\n(2) https:\/\/github.com\/google-research\/google-research\/tree\/master\/rouge\r\n(3) https:\/\/github.com\/pltrdy\/files2rouge (used in fairseq)\r\ncan be explained by two things, stemming and handling multiple sentences.\r\n\r\nStemming: \r\n(1), (2): default is no stemming. (3): default is with stemming ==> No stemming is the correct default as you did [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py#L84)\r\n\r\nMultiple sentences:\r\n(1) `rougeL` splits text using `\\n`\r\n(2) `rougeL` ignores `\\n`. \r\n(2) `rougeLsum` splits text using `\\n`\r\n(3) `rougeL` splits text using `.`\r\n\r\nFor (2), `rougeL` and `rougeLsum` are identical if the sequence doesn't contain `\\n`. With `\\n`, it is `rougeLsum` that matches (1) not `rougeL`. \r\n\r\nOverall, and as far as I understand, for your implementation here https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py#L65 to match the default, you only need to change `rougeL` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py#L86) to `rougeLsum` to correctly compute metrics for text with newlines.\r\n\r\nTagging @sshleifer who might be interested.","body":"I used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n","comment_length":145,"text":"Compare different Rouge implementations \nI used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n\nUpdates - the differences between the following three\r\n(1) https:\/\/github.com\/bheinzerling\/pyrouge (previously popular. The one I trust the most)\r\n(2) https:\/\/github.com\/google-research\/google-research\/tree\/master\/rouge\r\n(3) https:\/\/github.com\/pltrdy\/files2rouge (used in fairseq)\r\ncan be explained by two things, stemming and handling multiple sentences.\r\n\r\nStemming: \r\n(1), (2): default is no stemming. (3): default is with stemming ==> No stemming is the correct default as you did [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py#L84)\r\n\r\nMultiple sentences:\r\n(1) `rougeL` splits text using `\\n`\r\n(2) `rougeL` ignores `\\n`. \r\n(2) `rougeLsum` splits text using `\\n`\r\n(3) `rougeL` splits text using `.`\r\n\r\nFor (2), `rougeL` and `rougeLsum` are identical if the sequence doesn't contain `\\n`. With `\\n`, it is `rougeLsum` that matches (1) not `rougeL`. \r\n\r\nOverall, and as far as I understand, for your implementation here https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py#L65 to match the default, you only need to change `rougeL` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py#L86) to `rougeLsum` to correctly compute metrics for text with newlines.\r\n\r\nTagging @sshleifer who might be interested.","embeddings":[-0.0971114635,-0.1454416215,-0.1018338054,0.2446954995,-0.2790500522,-0.6873996258,0.0966215879,0.0406908654,-0.2773724794,0.2449082136,-0.1052665338,0.1245480254,0.156569168,0.0299380291,0.1508821547,-0.2753460407,0.0824456811,0.0195159987,-0.0070209946,-0.1224717423,-0.0989000425,0.710848093,-0.1582221389,0.0294798966,-0.1360562593,0.0632249787,-0.065176338,0.1977922171,-0.1465395987,-0.1077265888,-0.4238384366,0.2580074668,-0.0102514327,0.5024155378,-0.0001064565,-0.2057293504,0.1464959681,-0.0821161121,0.0163352154,-0.0362428129,-0.1412746161,-0.1513666213,-0.0882195458,-0.1276598573,-0.1156207323,-0.1115842536,-0.1406907737,0.1013708264,0.3020978868,0.461953491,0.2020174712,-0.2780742645,0.1017406732,-0.1463398635,0.5364743471,-0.0650649816,0.178373754,0.4277877212,0.1088055074,-0.0589308254,-0.2776232958,0.3784532249,0.0933002383,-0.2227546126,-0.0347062945,0.1359006912,-0.0522903986,-0.1840516031,0.0191417355,0.5499538779,-0.0278989412,-0.1120534241,-0.5472528338,-0.0820501149,-0.1682856977,-0.3088389933,-0.0079201665,0.2469446361,0.0827694684,-0.0889340937,-0.0894185379,0.0640295371,-0.106120877,-0.1750179976,-0.4134925008,0.0716322884,-0.148103863,-0.0499196239,0.2254970819,-0.138600409,-0.4518758357,0.0884429961,-0.2105749547,-0.224803865,-0.2263540477,0.030409513,0.2012510896,0.2470627427,0.1548953056,0.3268068135,0.1311066896,0.0373832807,0.100532271,0.0954623744,-0.2064218968,0.02000219,0.1247547567,0.0741545409,0.4574242532,0.1187108085,-0.0139872888,0.1224831119,0.2310451865,-0.4499968588,-0.2169406265,0.0168354549,-0.1520794034,-0.2141431421,-0.447525233,0.0682152659,-0.4359628558,-0.0626249835,0.052083049,0.2803235352,-0.0924392194,0.1899290681,0.0190020502,0.0347934514,-0.3867938221,-0.4065914452,-0.2543853819,0.1753966212,-0.1582213193,-0.0041399603,0.1424891353,0.0595285818,0.1201603115,0.0901783407,0.1998115033,-0.1900429577,0.1238479465,-0.3567660451,-0.1528307498,-0.1231164187,-0.1686126739,-0.1810043752,-0.072272785,-0.0784543008,-0.108654812,-0.0798542425,-0.3861472011,-0.0888710842,0.2821951807,0.2549304664,-0.1683550626,-0.0753798783,-0.0070643811,0.3652495742,0.0471455082,-0.2536486685,0.2029476911,-0.3402794003,-0.3170870245,-0.1564066559,0.1576577723,0.1496517211,-0.329675138,-0.1588445455,0.4732919931,0.0124880495,0.2169860601,0.4738480747,0.0712074786,0.1423042268,0.0155601818,0.1054048166,0.1755568385,-0.3649576604,-0.2462287545,0.2770293951,0.1828013361,-0.1516365409,0.1605482548,-0.1878699064,0.289907068,-0.0099040708,0.4108474553,0.2934286594,0.2333953381,0.1149007306,-0.0389558189,0.0885279849,0.1400673985,0.0026700818,-0.1767756641,-0.2594023645,-0.0072355797,-0.5122947097,0.2567877769,-0.219949618,-0.0949148536,0.437561512,0.0547254719,0.2224579304,0.3824210763,-0.0100100078,0.0522029512,-0.1480383873,-0.9107893109,0.1051268354,-0.1624167264,-0.1286664754,-0.1426205635,-0.2644803226,0.2685120106,-0.3495296836,0.2174311876,0.1194116175,0.2412713766,0.1883353144,0.0180596076,0.4312154651,-0.0966070518,0.136684075,-0.1902314574,0.0697116554,-0.0828463957,-0.1460764259,0.3277026415,0.4177888632,0.1070450917,0.1123865917,0.1384594589,0.4820681214,0.4255564809,0.2667169273,0.0809953362,-0.0636455938,0.2385511249,-0.1907710582,0.1061717644,0.4484041333,0.1561098844,0.0829993039,-0.5478582978,0.4688169658,-0.2570522428,0.0163856652,0.1720330417,-0.1505869776,0.1022734419,0.0802513286,-0.3891167045,-0.0443931371,0.3786428869,0.2883983552,0.5072104931,0.0236743093,-0.2626601458,0.0863702819,0.4717354178,-0.217466265,0.238318488,0.0318508707,-0.1205308437,-0.2995311916,0.0457245074,0.0068179835,-0.0083268434,0.2829852998,0.1845151484,-0.0496972837,-0.0539589524,-0.4153895676,0.2375267744,-0.1250955462,-0.2692331374,0.2895718813,0.3184883296,-0.0660476908,-0.4066005349,-0.4403741658,-0.1574134976,-0.0045677782,-0.1147072092,0.2354545444,-0.0973481089,-0.0767505765,0.0552370027,-0.3195218742,-0.3229712248,-0.2448723018,0.2438192666,0.0047329529,-0.1418255121,-0.1242140979,0.3198378384,0.1610330343,0.1131194308,-0.0448290743,-0.0473403968,-0.099969171,-0.1550525725,0.1665180326,-0.3360775709,0.0314207785,0.0443984643,-0.3856732547,-0.1498594284,-0.0498602055,-0.4121781886,0.3080253899,-0.1008325145,0.0904304236,0.3280991316,-0.3643452525,0.0567746982,0.3632389009,0.1514820158,-0.2763025761,-0.2511568964,0.2864522338,-0.1681543291,0.1224940717,-0.2411346585,-0.405803442,-0.0488448031,-0.2470820397,0.3765846491,0.1368347257,0.1807420105,0.2938701808,-0.0960831866,0.0734526515,0.2460882366,0.3446050882,-0.2459904402,-0.0772866383,0.2710967064,-0.4211298227,-0.4233298004,-0.1781837642,0.0295938309,-0.2226561457,0.3266169131,-0.0901057944,-0.2683164477,-0.2974977493,-0.0205990821,0.0652812421,0.1605580896,0.4207520485,-0.1087337881,-0.1601273865,-0.2147703618,0.3006651103,0.1324690878,0.0576618239,0.1184684932,-0.111734882,0.3356950581,0.1246398091,0.1817935556,0.6110421419,0.4109737873,-0.1472970694,-0.1327874362,0.1051609591,-0.2374237776,0.4501044154,0.2883417606,0.1164882332,-0.1492541134,0.4155920148,0.1704951972,0.0853628442,-0.0921643451,0.0817558914,-0.092612274,-0.3090031147,0.0528819002,0.2603550851,0.2658606768,-0.2009761184,0.1422980875,0.1658272445,-0.1431714892,0.0257825181,-0.0636698902,0.436113447,0.1187671646,-0.1927919835,0.2547782362,-0.5697327256,0.3782298863,0.2520797551,-0.102536723,-0.0779546723,-0.1554358006,-0.014631398,-0.2424529344,0.23066926,-0.3393939435,-0.0417412855,0.103797242,-0.3388761878,-0.4969877303,0.0106606996,-0.6937622428,-0.0959676057,0.0893724188,-0.0482525118,-0.500387609,0.1068418697,0.1828927547,-0.0426491089,0.0415886343,0.1271324009,-0.2671121955,-0.0529950932,0.0766839758,0.095223248,0.064646475,0.2550480962,-0.2173064649,0.0498057492,0.2746774256,-0.2561138272,0.3525841832,0.1796699166,0.3466412723,-0.3354488015,-0.3113321066,0.0776261315,-0.1116546318,0.1052936688,0.3015514612,-0.0927387998,-0.3703180552,0.2278022021,-0.2204487473,0.1226779073,0.3488481343,-0.1216979846,0.0580517575,-0.1321872324,0.2628260851,-0.2526238859,-0.1727537513,0.1578547955,0.208111465,-0.1084816605,-0.262350738,0.1070522964,0.3056500256,-0.21375072,-0.1419991702,-0.0868672505,-0.3515120149,0.2254666239,-0.1824957877,0.8882467151,0.2881525755,0.2318202853,0.2464224696,-0.2060979456,0.4856419861,-0.5538768768,0.3664364815,-0.3987361491,-0.1991071403,-0.0963082314,-0.0561261214,-0.0862100422,-0.0280217826,-0.3337031007,0.4100314975,0.0071538971,0.0599104948,-0.0034882368,0.1867198795,-0.0764986724,-0.0226583406,0.074322328,0.188435033,0.0009857749,0.0434769392,-0.1266750693,0.0128384102,-0.4018338621,-0.1078986898,-0.0795296058,-0.0119307535,-0.0163799059,0.226796329,-0.0228107218,-0.6391391754,-0.3914237916,0.4576609731,0.1660065651,0.1854424626,-0.1440284401,0.1402640045,-0.3095242679,0.3136017025,0.1122794226,-0.0075607598,0.1653447747,-0.0874927565,-0.1345496476,-0.0730030388,0.0828046203,-0.2731730938,-0.4576719999,0.1336611211,-0.4114170671,-0.1679706872,0.3816081583,0.1423403919,-0.28379035,-0.0725672171,0.166345939,0.1209417954,-0.0679295063,0.1607418656,-0.0545489416,-0.3273955882,-0.1977744848,0.2310941666,0.0633952096,0.0514847785,-0.0855836943,0.1746193618,-0.1256533116,-0.3106576204,-0.1957192272,0.0224111136,-0.1284235716,0.2936967015,-0.2116024345,-0.0801142752,0.1554585099,0.3048990965,0.0072228261,0.1150606722,0.1041508168,-0.3179338872,-0.1029892489,0.4243269563,0.0835595578,-0.0209632609,0.1780699044,-0.063766025,0.1291882098,-0.0087507926,-0.3982552886,0.0700935945,0.0439028814,0.0695231706,-0.1887413263,-0.491079241,0.0962919369,0.134873122,0.1042610556,0.3069246113,-0.0105625698,-0.2076698691,-0.2314288616,0.0660093874,-0.1553931981,0.062584959,0.1309311241,0.1137333512,-0.0862929747,-0.3337721527,0.3982598186,0.2252207249,0.2101087421,0.1649029404,-0.2435775548,0.1802204251,-0.1114250794,0.0854099244,0.096819669,0.1067475304,-0.0056563462,0.0913318321,-0.1165051609,-0.1035879627,-0.1371754259,0.2953248322,0.1966885477,0.0181525368,0.2846370935,0.0069291661,-0.0362474546,-0.3335184157,0.6811598539,0.2897186577,-0.1604288965,-0.012849737,-0.1669511348,0.3059040904,-0.1640089899,-0.0910124853,0.1018706635,0.0283455849,0.2761340439,0.404127568,0.1747854203,-0.0776226148,0.1023048237,-0.0922658518,0.0974028483,-0.1086211056,0.0788895935,-0.1504365355,0.0693043172,0.0775826275,0.4464603662,-0.0243503358,0.2357943058,0.5418562889,0.170206964,0.0721821263,0.2276646793,0.3588780463,0.2913191319,0.1977207959,-0.270734638,0.2905136049,0.0959035605,-0.0549213775,-0.1055883393,0.1782448739,-0.1429117173,0.2414572686,-0.2461590767,0.1060929,-0.116565451,-0.3853595853,-0.290648222,-0.2895007133,-0.2302589267,-0.0070825587,-0.1021535844,0.0588495247,0.2445739657,0.1722999662,0.1638800502,-0.2813107073,-0.157796517,-0.2306461334,0.6589990258,-0.2258089632,0.3789855242,0.0600176677,0.2099437118,0.298242569,0.2825274169,0.2726281285,-0.0003005813,-0.1618120223,-0.006298013,-0.2738713622,0.002062964,-0.1665722877,0.1828665137,0.2584967613,-0.0665344447,0.245697394,0.1705545783,-0.1593528092,0.2583978474,0.1558142751,0.1162889525,-0.172827214,0.2136123329,0.0240418185,0.1233218461,-0.0855042264,-0.127245605,-0.4410176575,-0.1121004447,-0.1680351049,0.1084946543,0.025160525,0.0635279939,0.1330225468,0.1738585681,-0.074643217,0.0802953467,0.3782275617,0.1270335317,-0.1432542652,-0.092751734,-0.1001958027,-0.1376857013,0.059174858,-0.2534724474,0.450779587,-0.0930621997,0.26419276,0.1324884146,0.2620571256,-0.1844151914,0.103899017,-0.3080265522,0.2352761775,-0.1922248155,-0.0887037888,0.007962279,0.0237388015,-0.0437168293,-0.0535066277,0.0951554105,0.0235950705,-0.2101692855,0.1189062223,0.1491817385,0.232622236,0.5534223914,0.2988824844,-0.0902613252,-0.2370071262,-0.0626338497,-0.0594509616,-0.1318840683,0.4863539934,-0.0554535538,0.2033436745,0.1254243553,-0.5329968929,0.1028760672,0.7695531249,0.0413449705,-0.1699378192,-0.2594554424,-0.0285567343,-0.0032677792,-0.1611649096,-0.371461302,0.55631423,-0.0945035666,0.0448718965,-0.0440746918,-0.5466017127,0.239241749,-0.4795344174,0.0947285295,-0.1087267324,0.3924561441,0.4902783036,-0.0042809066,-0.0192579683,0.0582683347,0.1101683676,0.0432714,-0.2417525649,0.0718722343,0.0064345235,-0.0960608348,-0.2161065638,0.4115979671,0.2002281845,-0.3074797988,0.0446102209,-0.3132918775]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/617","title":"Compare different Rouge implementations ","comments":"Thanks for the clarification !\r\nWe're adding Rouge Lsum in #701 ","body":"I used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n","comment_length":11,"text":"Compare different Rouge implementations \nI used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n\nThanks for the clarification !\r\nWe're adding Rouge Lsum in #701 ","embeddings":[-0.1528427899,-0.0611458682,-0.1072614789,0.2750061154,-0.1931881607,-0.5852844119,0.1004895791,0.1216422245,-0.210647136,0.3737171888,-0.0761662871,-0.048032824,0.0633076355,0.2503480017,0.2926432788,-0.1131119952,0.0960009918,-0.0602212101,-0.2101655155,-0.1339818984,-0.1771398932,0.65410465,-0.0409100465,0.046367567,0.0632486269,0.2988618612,0.0673700571,0.2181526124,-0.1791144609,-0.1109263599,-0.2365793288,0.0613305755,0.0596034452,0.4658019245,-0.0001062587,-0.0672374666,0.1123889461,-0.0673811361,0.0600942522,0.0123098725,-0.0240003616,-0.2413208336,-0.1353582144,-0.1181622669,-0.2568468153,-0.1838687658,-0.3425980806,0.1531993151,0.0362444147,0.4438218176,0.2555426061,-0.127364412,0.1330001652,-0.1620425582,0.5634287,-0.0155348452,0.0679570436,0.396589607,0.1207176521,-0.0458484292,-0.1549121737,0.3379276693,0.2295182198,-0.1192707717,0.082954511,0.1697649062,0.0986734033,-0.0716408789,0.084492214,0.4691805243,-0.0288786571,-0.1069886833,-0.4058154225,0.0811476558,-0.1984739006,-0.3261063993,0.0500599705,0.0659596771,0.2496011853,-0.1849741638,-0.0812022835,0.116930075,-0.0369448923,-0.1823112965,-0.3639786541,-0.1119451448,-0.2039190233,-0.0089327479,0.2250459492,-0.1471475363,-0.5862848163,0.2469500601,-0.1263728589,-0.1741847396,-0.2890440524,-0.0751185343,0.345415324,0.4011195004,0.2002052218,0.3864662647,0.0959936678,0.0460763164,0.1159628481,0.1419698447,-0.3884547949,0.1357728839,-0.2166532129,0.008830836,0.4617735147,0.1950901449,0.100604035,0.0780770183,0.3994865119,-0.6182559729,-0.2151797563,0.1035709456,-0.2407171279,-0.4329729676,-0.574124217,0.0799981058,-0.2702183723,0.0079455692,0.1249134094,0.1930324137,-0.1957811266,0.0234519057,0.1231587157,-0.0426531695,-0.3626556695,-0.213628605,-0.2660005987,0.2370864302,-0.2587068975,0.0174170043,0.1940762103,0.0170851946,0.1839222908,0.0883413553,0.1430536509,-0.0371214151,0.3083031774,-0.1375155449,-0.1332628578,0.0662642196,-0.0820600167,-0.2116745561,-0.0183004681,-0.0290100612,-0.1287070364,-0.3225089908,-0.4527665973,-0.0673796311,0.3049278557,0.2760229111,-0.1430872977,-0.0287758932,-0.2365648896,0.277312398,-0.1067823321,-0.3479142487,0.0465568714,-0.2719525695,-0.3659673333,-0.1976127923,0.3096289039,0.1657142341,-0.1389899254,-0.0832702741,0.4742568135,-0.0099184532,0.1288123131,0.4362281859,0.0585098825,0.031231394,-0.161280781,-0.2734947503,0.2090555727,-0.4712717831,-0.3539271355,0.0425381437,0.1100050136,-0.3289382458,0.2141660005,-0.1195287928,0.0704513639,-0.1367423683,0.2866173685,0.1662424356,0.0901039019,0.1118936613,-0.1190808937,0.0559900776,0.3076560199,0.0289101657,-0.0009170442,-0.1683111191,0.1649904698,-0.5784717798,0.1853294075,-0.0626728982,-0.2203243524,0.3127512336,0.0251160767,-0.0002623535,0.3996472359,0.1340130568,0.046976842,-0.0518117659,-0.86396873,0.076780118,0.0732079223,-0.0382693,-0.299898237,-0.2111523449,0.3324357569,-0.4365283847,0.2283637971,-0.0982639715,0.2926132083,0.211512059,0.1612141579,0.2776581347,0.0296151359,0.0386080407,-0.2308572233,0.1521981359,-0.0588073581,-0.1017889827,0.3684000969,0.3248451948,0.1957282126,0.0123637374,-0.0107964687,0.3345304132,0.4417462349,0.2129287571,0.4284333587,0.3503247201,0.1484453827,-0.1505494416,-0.0047803945,0.5567100048,0.1401019394,0.1286562234,-0.5229117274,0.5289228559,-0.1727775335,0.0106472764,0.2033688277,0.0133677265,0.1232714057,0.133248657,-0.2716521919,0.0922405645,0.4088835418,0.3970427811,0.5696486235,-0.0024944362,-0.0945209935,0.0824232176,0.3581399322,-0.1263414472,0.1422861069,0.0946246088,-0.1648318768,-0.3421795964,0.0798388124,-0.0422697179,0.0349530876,0.2659430504,0.1358959228,-0.0178906191,-0.156768918,-0.2804732025,0.0579801947,-0.1918255538,-0.2144970149,0.2159801126,0.2755699754,0.0026511506,-0.404502064,-0.4074856639,-0.2298941165,0.2143649608,-0.1705447882,0.2009943724,0.0029078168,0.2751621008,0.037484102,-0.2373501658,-0.3371591568,-0.305901587,0.1972179115,0.0902575776,-0.0896178782,-0.0846618786,0.1339703947,0.1649670899,0.0457634516,0.0893999636,-0.0306954719,0.0717854202,-0.2854025364,0.1797916442,-0.2597793937,-0.0702542141,0.1722550392,-0.3645096123,0.1022890657,-0.1851407588,-0.5716658235,0.1637839079,0.0234540589,0.1886906475,0.2875541449,-0.3760790229,-0.0028093592,0.5482458472,0.1620009392,-0.4129064381,-0.2315890491,0.3086678386,-0.2798323333,0.1963503361,-0.1302864105,-0.3486282229,-0.1011128575,-0.0998548865,0.2986157537,0.2311086953,0.0784683898,0.2138189375,-0.0090619046,0.0340805687,0.4312954247,0.1299258173,-0.369964838,-0.2476148307,0.3482755423,-0.3980797231,-0.5546488762,-0.0805934891,0.0342733413,-0.1124177799,0.2715864182,-0.1769687831,-0.3452818096,-0.1302178055,-0.019770151,0.08100532,0.1199632362,0.3832591176,-0.1203208193,-0.2382375598,-0.2333061546,0.1195922419,0.0808405727,-0.0105793411,0.1975070983,-0.1426150054,0.279304266,0.0342684947,0.2377271354,0.4797061682,0.1795962155,0.0138947116,-0.0955122113,0.269938767,-0.1557110697,0.3139441311,0.3259946406,0.172384724,-0.1392376721,0.2765818834,0.2353680879,0.1125058904,-0.1708000153,0.1469275355,-0.0446894169,-0.2448344827,-0.0556753576,0.4001224637,0.2454077452,-0.1311144233,0.1294218302,-0.0364400297,-0.1666232944,0.3285706043,0.1519385427,0.2806546688,0.1473815739,-0.1559613645,0.2134525925,-0.6486700177,0.4603625238,-0.0308881551,-0.1085290015,-0.1376658678,0.0097837942,-0.1110118702,-0.2807649374,0.1993288845,-0.4603848755,0.066716902,0.1134239882,-0.3132347465,-0.3941717744,-0.078683041,-0.448538959,-0.1814924777,0.1171901152,0.1012273878,-0.3530376852,0.0611478798,0.0917448178,-0.0677018762,-0.0456716567,0.2226566374,-0.3362945914,-0.1970995665,0.236171633,-0.0676992312,0.0040912726,0.1672487855,-0.0757879391,-0.0333576873,0.1294903457,-0.2365455329,0.3288589418,0.1696354896,0.3636464477,-0.1086681038,-0.1674969494,0.1322385073,-0.0828189105,0.230282709,0.3414802551,-0.0538132824,-0.2775624394,0.4081151485,-0.1304142624,0.0880350769,0.1630932987,-0.1618663222,0.0499245971,-0.1587803364,0.2338882387,-0.3514137268,-0.0381976515,0.1602584571,0.2581945658,0.0972512364,-0.2841448784,0.2031993866,0.3612407446,-0.2494311184,-0.0636799335,-0.1197088361,-0.1709538102,0.148012042,0.0317917876,0.8107596636,0.1888261288,0.1929830015,0.2305129319,-0.3339792192,0.1820219308,-0.5704451203,0.1644662023,-0.4657248259,-0.2136604786,-0.1034941152,-0.1194690913,-0.033537522,-0.2235874385,-0.0782997832,0.4557689428,0.0419424102,0.0984256417,-0.091877982,0.1851922721,0.1044481993,-0.0350807682,0.0957682058,0.241348505,0.1102142408,0.1529450417,-0.111172393,-0.0246540718,-0.4451616406,-0.063958019,-0.0506898388,-0.0694252104,-0.059554901,0.2939298153,-0.0059691607,-0.6267854571,-0.5150988102,0.3526248336,0.2375436574,0.0704839453,-0.2019281238,0.1187909991,-0.4478484392,0.372811377,0.0806427076,0.0280544292,0.0842144936,-0.1626178771,0.0089347987,0.0749125779,0.1274729669,-0.2324500978,-0.375960201,-0.0138020786,-0.4278290868,-0.1879983842,0.5705935955,-0.0379408374,-0.2297326773,-0.0642030537,0.1978012472,0.3043072522,-0.1228984669,0.245670557,0.0372307673,-0.3480582535,-0.1357951015,0.2920525372,0.0463437252,0.0283095594,-0.1438573301,0.0265813638,-0.0365618207,-0.1978595108,-0.1090139374,0.1963453591,-0.0926556215,0.1375515014,-0.0818655342,0.0089192223,0.1224693581,0.2921477556,0.1473442763,0.0158533435,0.2272143513,-0.3467158079,-0.3288124204,0.3265758753,0.1177244037,0.017075168,-0.0256711189,-0.2779202759,0.1513644308,-0.0372230411,-0.4082359672,-0.0061759483,0.0333656557,-0.0127202477,-0.2629878223,-0.5039400458,0.1784015745,0.107167162,0.1114238352,0.0729708225,-0.282649368,-0.178284511,-0.2898870707,0.0639187172,-0.1217119172,0.0552221909,0.0187350735,0.176216498,-0.1442463398,-0.1706786603,0.1979034543,0.1417275816,0.1822058409,-0.0597179607,-0.2957048714,0.0236565247,-0.0693095252,0.0667957366,0.2711741328,0.0337897316,-0.0607240461,-0.0845252872,-0.0609109364,-0.0903814659,-0.0388481021,0.1906573325,0.0916739404,-0.0715399086,0.4211240411,0.0183396824,0.0686740056,-0.2104288638,0.5500888228,0.0981052145,-0.1150973216,-0.0581533425,-0.1049967781,0.2794933021,-0.2724590898,-0.0852908939,0.1035323292,0.1525912583,0.1581961364,0.3453665674,0.3301294148,0.0249643195,0.2019703537,-0.0580631271,0.0443283468,-0.1161082685,-0.1079584956,-0.3503504694,-0.0416158997,0.2297309041,0.5151963234,0.0099860849,0.1777613461,0.3620508015,0.2042513788,0.1290772706,0.0592902042,0.1612062752,0.0066647744,0.0222550239,-0.2425492704,0.2160447985,0.119064793,0.0967961252,-0.035247989,0.1933453381,-0.1556632072,0.1701268554,-0.2499852777,-0.0600799695,-0.1198159605,-0.3046157062,-0.1571251154,-0.3565097153,-0.2864809334,0.1135010794,-0.0126525695,0.086265102,0.3268326223,0.0907069817,0.0376075581,-0.1472859681,-0.3715554476,0.0859847218,0.4950945973,-0.0470159948,0.1837365627,0.0535122529,0.1549271345,0.1815846562,0.2320251018,0.2359657139,-0.0020984812,-0.2461183071,-0.1635944247,-0.0284720864,0.06258104,-0.2277621925,0.1807807088,0.1693140566,-0.2391952872,0.1032404676,0.1489448696,-0.178013891,0.2535608411,-0.1593280882,0.1564447135,-0.1166262031,0.1449630558,-0.0829523504,0.1041056812,-0.2034757882,-0.1244164333,-0.2963523269,-0.2545482516,-0.2615458071,0.2130655795,0.0604420751,-0.1268168241,0.1362889111,0.0446989015,-0.012362211,-0.0413504764,0.3326857686,-0.0229386855,-0.1541642547,-0.1858228594,-0.032166291,-0.097354643,0.0753589645,-0.2445172369,0.3561956584,-0.0256970748,0.1959177405,0.4269831181,0.1973309517,0.1189972833,0.1664416343,-0.35968104,0.1796678454,-0.3289890587,-0.0835999325,0.051192183,0.0580967404,-0.0648388714,-0.1481257677,0.2130869478,-0.1255076528,-0.2675185204,-0.0073577827,0.2974867225,0.5228182077,0.4153002501,0.3001864851,-0.1712741703,-0.2625159621,-0.1079299599,-0.0495198965,0.0065987632,0.2437066287,-0.2474615276,0.0755785406,0.240685001,-0.6320130825,-0.0358067192,0.7777664661,0.1705209762,-0.2056149542,-0.1771870255,-0.0467948653,0.0601412319,-0.1613179594,-0.3666918576,0.4293400943,0.0193279032,0.1457222402,-0.0775825828,-0.3843090236,0.4043550193,-0.4059817791,0.2161092907,-0.1777913272,0.4841455519,0.4587677717,0.0775431246,-0.0362339206,0.1240529045,0.0769750327,0.1865629703,-0.0513241328,0.067648761,-0.0532729253,-0.0345410481,-0.0965212062,0.4296952784,-0.0051104231,-0.1669147909,0.172662884,-0.2040516436]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/617","title":"Compare different Rouge implementations ","comments":"This is a real issue, sorry for missing the mention @ibeltagy\r\n\r\nWe implemented a more involved [solution](https:\/\/github.com\/huggingface\/transformers\/blob\/99cb924bfb6c4092bed9232bea3c242e27c6911f\/examples\/seq2seq\/utils.py#L481) that enforces that sentences are split with `\\n` so that rougeLsum scores match papers even if models don't generate newlines. \r\n\r\nUnfortunately, the best\/laziest way I found to do this introduced an `nltk` dependency (For sentence splitting, all sentences don't end in `.`!!!), but this might be avoidable with some effort.\r\n\r\n#### Sidebar: Wouldn't Deterministic Be Better?\r\n\r\n`rouge_scorer.scoring.BootstrapAggregator` is well named but is not deterministic which I would like to change for my mental health, unless there is some really good reason to sample 500 observations before computing f-scores.\r\n\r\nI have a fix on a branch, but I wanted to get some context before introducting a 4th way to compute rouge. Scores are generally within .03 Rouge2 of boostrap after multiplying by 100, e.g 22.05 vs 22.08 Rouge2.\r\n\r\n","body":"I used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n","comment_length":144,"text":"Compare different Rouge implementations \nI used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n\nThis is a real issue, sorry for missing the mention @ibeltagy\r\n\r\nWe implemented a more involved [solution](https:\/\/github.com\/huggingface\/transformers\/blob\/99cb924bfb6c4092bed9232bea3c242e27c6911f\/examples\/seq2seq\/utils.py#L481) that enforces that sentences are split with `\\n` so that rougeLsum scores match papers even if models don't generate newlines. \r\n\r\nUnfortunately, the best\/laziest way I found to do this introduced an `nltk` dependency (For sentence splitting, all sentences don't end in `.`!!!), but this might be avoidable with some effort.\r\n\r\n#### Sidebar: Wouldn't Deterministic Be Better?\r\n\r\n`rouge_scorer.scoring.BootstrapAggregator` is well named but is not deterministic which I would like to change for my mental health, unless there is some really good reason to sample 500 observations before computing f-scores.\r\n\r\nI have a fix on a branch, but I wanted to get some context before introducting a 4th way to compute rouge. Scores are generally within .03 Rouge2 of boostrap after multiplying by 100, e.g 22.05 vs 22.08 Rouge2.\r\n\r\n","embeddings":[-0.1379877776,0.0537790135,-0.065096423,0.2345050275,-0.0384498984,-0.6865086555,-0.1374980062,-0.0325634032,-0.2779031098,0.395267278,-0.0672829151,0.1716644168,0.1589004546,0.363702029,0.2211120129,-0.0784863532,0.1345872134,-0.0168794338,-0.1050260067,-0.1088072285,0.0128314812,0.6053342223,-0.0310589336,0.1092312336,-0.1490680575,0.1322920471,0.2132202089,0.0989613906,-0.1409740895,-0.10045214,-0.1728941202,0.1978754103,-0.109423846,0.4348340333,-0.0001064469,-0.1936359555,0.1715347022,-0.1881962866,0.0778008103,-0.1544618458,0.0580871999,-0.0233251452,-0.1111885011,-0.038434051,-0.2476809025,-0.0726091638,-0.2509014606,0.0748237446,0.1554621905,0.3872460723,0.2037396878,-0.0639967397,0.1957696378,-0.2286606133,0.5280544758,-0.0566640608,0.0506690852,0.5248132944,-0.12089324,-0.107549496,-0.3474807739,0.3372290134,0.1901493073,-0.3185483217,0.1235745698,0.0154710636,-0.0340329148,-0.2020000219,-0.1000618935,0.6137260199,-0.1752004027,0.0161953811,-0.5487056375,-0.0667547062,0.0393242054,-0.1948316991,0.0913067833,-0.0565147027,0.036862243,-0.1985515356,-0.0645336062,0.3034345508,-0.0474751815,-0.1840778142,-0.1046651229,-0.0378048271,-0.1254391521,-0.0757325441,0.4095121622,-0.1220963448,-0.5111020207,0.3069493175,-0.1545218378,-0.0891522393,-0.5262218714,-0.0610734038,0.1598901302,0.4157986641,0.2494131476,0.4495767057,0.3137447238,0.053171549,0.0750584751,0.0652157143,-0.3272762299,0.0348073766,-0.0133848796,-0.0332113691,0.3852156997,0.131689325,0.0124615915,0.1276688725,0.3607757986,-0.5237594247,-0.2875634432,0.1228606254,-0.4410630465,-0.3779605329,-0.4348496199,0.1786829531,-0.58181566,-0.0216637384,0.0735606402,0.1862042248,-0.1235137582,0.2579784691,0.01279213,-0.0408552103,-0.3306750655,-0.2836543024,-0.2149548233,0.0749734938,-0.1546937078,0.1803346425,0.0942670256,-0.0190296117,0.1723968834,0.0128592616,0.1753636003,-0.2587369978,0.0640308708,-0.2548103034,-0.1448667794,-0.0185460914,-0.1496222019,-0.2562123835,-0.0164534934,-0.1931044012,-0.130374521,-0.1442642361,-0.1164204851,-0.0432634428,0.4312921166,0.2439413667,-0.1952553988,-0.166735217,-0.0696855113,0.2667293549,-0.2485964149,-0.1255652905,0.1539250165,-0.2475672364,-0.2683533728,-0.0707558915,0.3554469347,0.1554047316,-0.1645865887,-0.1988318264,0.5341036916,-0.0445965305,0.0099409474,0.3509596288,0.144003734,0.3776940107,-0.0017380156,-0.1347105801,0.2774407268,-0.3410018384,-0.276017189,-0.0057499236,0.1186543182,-0.0679779425,0.1401611567,-0.0949994102,0.0232433118,-0.0039749565,0.1682724208,0.2526464164,0.0599440262,0.0217985604,-0.1456855536,0.0653613731,0.2068350762,0.1683458388,-0.0439079329,-0.2805389762,-0.0583609864,-0.608435154,0.330196321,-0.1606402695,-0.3503609002,0.3265835047,0.0191348307,0.165494442,0.2853588462,-0.0028769886,0.1243408024,-0.0277984142,-0.8134545088,0.1548524648,0.125845179,-0.1853113025,-0.326454699,-0.1159099936,0.3186032772,-0.4248637259,0.1786702722,-0.0310022291,0.4205991328,0.1584344208,0.0026366166,0.3995909989,-0.1581787318,0.1119308248,-0.3919112682,-0.0032044484,-0.0094384365,-0.1483703554,0.2668654621,0.4590812922,0.1804880053,0.1636718512,-0.0739021748,0.4581432343,0.3668297529,0.0855815038,0.1537888348,0.1852805465,0.0301368963,0.0182526968,-0.1596736908,0.5856784582,0.1365930885,0.1660449803,-0.343705982,0.5487450957,-0.4739865661,0.0545277968,0.180081591,-0.1455039084,0.0748512819,0.1776161045,-0.4026848078,0.0675391853,0.4525692463,0.1796044707,0.5824773312,0.0392693095,-0.2232962996,-0.0138708372,0.296569556,-0.2222269028,0.1048379168,0.0292703211,-0.0729205385,-0.1611636728,0.0832695216,-0.04691967,-0.0286678821,0.2086730599,0.0139425118,-0.0416535251,-0.1332062185,-0.4041922092,0.1018150374,-0.1700228602,0.0515731499,0.2663269043,0.2414890677,0.1121663079,-0.3807005286,-0.3593528271,-0.360661149,-0.0638499707,-0.1678470969,0.1870641261,-0.1131305397,0.0091959471,-0.1322094053,-0.2192118615,-0.4324485958,-0.298661083,0.1582986712,0.0585571602,-0.1180024147,-0.0339629017,0.3236483634,0.2754863203,0.0101452358,0.0402028747,-0.022171773,-0.0944166854,-0.2344562411,0.1769817322,-0.2460259348,-0.1812212616,0.0618322268,-0.3107549846,-0.0086973663,0.0190707985,-0.4526582956,0.2786203623,-0.1029074192,0.2484103441,0.2135023028,-0.4276442826,0.0032276656,0.4406077564,0.1218789071,-0.3287892044,-0.3009264469,0.1931613237,-0.0724636465,0.1929080039,-0.087208651,-0.2683808804,0.0207664911,-0.0987430736,0.5206075907,0.114422828,0.1009172648,0.3030361831,-0.091273807,0.0563136749,0.0778944567,0.1816982329,-0.2014032602,-0.0219408758,0.3603029847,-0.3418867886,-0.4680043757,-0.1450920701,0.0597500168,-0.2673578858,0.1555731744,-0.1215312704,-0.3317548335,-0.0737196356,0.0234594159,-0.0890874565,0.0716887563,0.4092116654,-0.2179155201,-0.0893819481,-0.1051583812,0.0778067634,0.1946632862,-0.1805390269,0.1870252341,-0.1966016889,0.3595887125,0.0451525673,0.0217612889,0.5290668011,0.3862200379,-0.2149225473,-0.0533157438,0.0830888152,-0.2630346715,0.4513287842,0.3552639782,0.0063207261,-0.1950836629,0.2531484962,0.0723757818,0.3628646731,-0.1058622524,0.0236285031,0.0601574332,-0.3445747793,0.0368500128,0.0886248723,0.1257892996,-0.2093370408,0.2825529277,0.0262495931,-0.1072156057,0.1978482753,-0.0251671094,0.381575048,0.1027464718,-0.1733493805,0.2707933486,-0.4753193855,0.3523247242,0.119090654,-0.0710646957,0.0614489019,-0.2489382625,-0.0822234899,-0.2329117805,0.334376663,-0.2879667282,-0.0426405556,0.1241561398,-0.2301528156,-0.3302980065,-0.1288579404,-0.3942030668,-0.3302254975,0.0731104985,0.0721996129,-0.2989219129,0.0522131138,0.3209913075,-0.1172250286,0.0026303825,0.109691143,-0.2671667635,-0.1202243865,0.1696534157,0.1458065659,-0.0550408736,0.2817126513,-0.1646098047,0.0965656266,0.0507510789,-0.1786045432,0.3109716475,0.2856430411,0.3082370758,-0.2026073635,-0.1595377028,0.1090664864,-0.1977570057,-0.1745832264,0.3606675565,-0.0489384606,-0.3824974298,0.1961781979,-0.2080855221,0.2019234598,0.3150054216,-0.0898815468,0.1575016677,-0.2980253398,0.1455921084,-0.4090366662,-0.0803997144,0.0376452394,0.1423522383,0.1401459873,-0.2033582479,0.3560957611,0.3175186813,-0.221390754,-0.3108106554,-0.0388196446,-0.197138086,0.4677298367,0.0941878036,0.8448509574,0.1214060038,0.1850350499,0.1718835831,-0.2384848744,0.482398808,-0.5614017844,0.1663890183,-0.5115357637,-0.2749329209,-0.0329950415,0.0543622449,-0.000091703,0.2027971745,0.0014260281,0.5391435027,0.2287288755,0.2562922537,0.1209904924,0.176376611,0.1043923348,-0.1710025668,-0.1731213033,0.2126554251,0.0154755535,0.0278288908,-0.0779891983,0.1016246602,-0.5096171498,0.0388418846,-0.0475021303,0.0537024476,-0.3614259064,-0.0153733632,0.2476603687,-0.5977811217,-0.4231728911,0.4159291983,0.2904119194,0.329241097,-0.0346556827,0.1431695372,-0.1709618568,0.3200541139,0.180507794,-0.0252600126,0.2272320837,-0.0795198306,-0.1334786415,-0.0351764634,0.2026481777,-0.3420901,-0.2875040174,-0.0931458026,-0.3542186618,-0.1526328027,0.3578611314,-0.0374417305,-0.0467738286,-0.0479708612,0.1659757048,0.0911061391,-0.035161972,0.0553895384,0.0553387441,-0.1962457448,-0.1203996986,0.1445234567,-0.1053355411,0.0237730294,-0.2270730138,0.0759275481,-0.0387644991,-0.207332626,-0.0831038654,0.1247793436,-0.159474194,0.0426362716,-0.1619380862,-0.2588415742,0.0776389688,0.5247109532,0.1725976169,-0.0272152536,0.2617967725,-0.247686401,-0.1904671937,0.3711900115,0.3751148582,-0.0460038967,0.1523190886,-0.1036455855,-0.1489115357,0.1098958775,-0.3918213248,-0.2247714102,-0.0893317908,-0.0843495131,-0.1334334761,-0.5113669634,0.1355942339,0.111934498,0.1190609708,0.4008902013,-0.038710054,-0.1745301932,-0.2668007314,0.0672653392,-0.0373619273,0.1258926541,-0.0388484448,0.2830671072,-0.0477955118,-0.387814641,0.2133150846,0.0649926886,0.3588874936,0.1206027195,-0.2820934653,0.3386889398,-0.2158180177,0.0978473872,0.1097206473,0.2145822197,-0.0102847042,0.0225072615,-0.0935600474,-0.0362783521,-0.0905458704,0.2599566579,0.1787956506,0.0101687144,0.347940892,0.0176668242,-0.1987026781,-0.264678359,0.5470200181,0.244130969,-0.0872399732,-0.1567054987,-0.1239550933,0.2401517332,-0.2691858113,0.0803485811,0.0650192872,-0.0343372412,0.1643128097,0.5011052489,0.1254830956,-0.0478974618,0.1490359604,-0.0842955038,-0.076876305,0.0619587377,-0.0179571696,-0.3792600632,0.2421419472,0.081241332,0.5509499907,-0.1418097168,0.2926924527,0.1170540228,0.1912472844,0.0523612313,0.0357893072,0.3498686552,0.0717564747,0.0685830489,-0.0695201233,0.3497830331,0.1589658111,-0.011422866,-0.1220556051,0.163286224,-0.1536619961,0.0304485504,-0.320453912,-0.0838658288,0.0096937977,-0.4271037281,-0.2785019875,-0.2624120712,-0.4468614161,0.1273812652,-0.1581125706,0.0449713506,0.1046799943,0.1011696681,0.1346336454,-0.1836872995,-0.3204913437,-0.089729242,0.7516458035,-0.2739139199,0.3097505569,0.014521786,0.2173309773,0.1310406029,0.2928340733,0.0901149064,0.0360212959,-0.2581721842,-0.0432589427,-0.1751989275,0.1129976511,-0.0306167416,0.2174670696,0.1622592658,-0.0668792129,0.3745427728,0.122553803,-0.1779455543,0.1882741451,0.0499314554,0.31998384,-0.2425409555,0.1575743109,0.1437201053,0.259103179,-0.1944749355,-0.0446774438,-0.259272486,-0.1768307388,-0.4694056809,0.1417616457,0.0533981211,-0.0886670575,0.123110339,0.1777946353,0.1758434772,-0.0021955296,0.2403035909,0.0196204111,-0.1153639629,-0.1746169627,-0.0410737284,-0.0697482154,0.1757575423,-0.1810772717,0.2745839953,0.0295503121,0.3539460301,0.0539209582,0.185622558,-0.0278201364,0.1713773161,-0.2799511254,0.1228574812,-0.4049214423,-0.2996936142,0.0162159093,0.0775864944,-0.1081045493,-0.0085341623,0.1224963292,0.148757562,-0.4328809679,0.1646026671,0.1242287979,0.3373118639,0.4153500199,0.2004424781,-0.012265428,-0.4298881292,-0.1146634966,-0.0157832131,-0.2943872511,0.3858157098,-0.3126731813,0.2461764067,0.2729097009,-0.2843562365,-0.0202984251,0.6423115134,0.0609293766,-0.3174879253,-0.1846447885,-0.0446751192,0.111122556,-0.2338804156,-0.3199441433,0.4819262624,-0.0924898684,0.3255455494,-0.0406468213,-0.2320198715,0.1530934721,-0.3321328163,0.337657094,-0.0510522537,0.3423324227,0.4322223067,0.0168738477,-0.2412815541,0.029936092,0.0586193912,0.1642422974,0.0302766897,0.1001572683,-0.134432748,-0.0817966834,-0.2281965315,0.4084760845,0.0062790331,0.0131902518,0.1105067134,-0.3058946431]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/617","title":"Compare different Rouge implementations ","comments":"> This is a real issue, sorry for missing the mention @ibeltagy\r\n> \r\n> We implemented a more involved [solution](https:\/\/github.com\/huggingface\/transformers\/blob\/99cb924bfb6c4092bed9232bea3c242e27c6911f\/examples\/seq2seq\/utils.py#L481) that enforces that sentences are split with `\\n` so that rougeLsum scores match papers even if models don't generate newlines.\r\n> \r\n> Unfortunately, the best\/laziest way I found to do this introduced an `nltk` dependency (For sentence splitting, all sentences don't end in `.`!!!), but this might be avoidable with some effort.\r\n\r\nThanks for the details, I didn't know about that. Maybe we should consider adding this processing step or at least mention it somewhere in the library or the documentation\r\n\r\n> #### Sidebar: Wouldn't Deterministic Be Better?\r\n> `rouge_scorer.scoring.BootstrapAggregator` is well named but is not deterministic which I would like to change for my mental health, unless there is some really good reason to sample 500 observations before computing f-scores.\r\n> \r\n> I have a fix on a branch, but I wanted to get some context before introducting a 4th way to compute rouge. Scores are generally within .03 Rouge2 of boostrap after multiplying by 100, e.g 22.05 vs 22.08 Rouge2.\r\n\r\nI think the default `n_samples` of the aggregator is 1000. We could increase it or at least allow users to change it if they want more precise results.","body":"I used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n","comment_length":210,"text":"Compare different Rouge implementations \nI used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n\n> This is a real issue, sorry for missing the mention @ibeltagy\r\n> \r\n> We implemented a more involved [solution](https:\/\/github.com\/huggingface\/transformers\/blob\/99cb924bfb6c4092bed9232bea3c242e27c6911f\/examples\/seq2seq\/utils.py#L481) that enforces that sentences are split with `\\n` so that rougeLsum scores match papers even if models don't generate newlines.\r\n> \r\n> Unfortunately, the best\/laziest way I found to do this introduced an `nltk` dependency (For sentence splitting, all sentences don't end in `.`!!!), but this might be avoidable with some effort.\r\n\r\nThanks for the details, I didn't know about that. Maybe we should consider adding this processing step or at least mention it somewhere in the library or the documentation\r\n\r\n> #### Sidebar: Wouldn't Deterministic Be Better?\r\n> `rouge_scorer.scoring.BootstrapAggregator` is well named but is not deterministic which I would like to change for my mental health, unless there is some really good reason to sample 500 observations before computing f-scores.\r\n> \r\n> I have a fix on a branch, but I wanted to get some context before introducting a 4th way to compute rouge. Scores are generally within .03 Rouge2 of boostrap after multiplying by 100, e.g 22.05 vs 22.08 Rouge2.\r\n\r\nI think the default `n_samples` of the aggregator is 1000. We could increase it or at least allow users to change it if they want more precise results.","embeddings":[-0.1199247912,0.0578211658,-0.0259994827,0.247537449,-0.0194163173,-0.7114323378,-0.0715660974,-0.0210268293,-0.2410765439,0.3681191504,-0.0042177765,0.1752641201,0.1229400113,0.3598643839,0.2975478172,-0.1921318769,0.0986103565,-0.0162566714,-0.0499816351,-0.0920571387,-0.0059650606,0.6030076146,-0.0572601967,0.1056353524,-0.1943653226,0.1431973875,0.1966951936,0.0866381153,-0.1192162037,-0.1853498966,-0.2451326549,0.2077442408,-0.0349605605,0.4523561895,-0.0001091167,-0.2276299447,0.2159281522,-0.2167396098,0.1104495153,-0.1605566889,0.0378186703,-0.1098205,-0.0797066987,-0.081466049,-0.2882637382,0.00725001,-0.1965866834,0.096768409,0.1674869955,0.356051892,0.1705368161,-0.0147535512,0.1806708574,-0.2480346411,0.5668400526,-0.0728992522,0.0628357455,0.6113482714,-0.0754710585,-0.0680768639,-0.3386306465,0.3610008359,0.1657429338,-0.3372178674,0.0628842488,0.1117008552,-0.0137623856,-0.1807054728,-0.0311415475,0.6794782281,-0.1486427635,0.0029165316,-0.4949690402,-0.0983795896,0.0643410534,-0.3069671988,0.1008419693,0.0712011755,0.0650982261,-0.1525472999,-0.1313446015,0.2446857393,-0.1071738079,-0.1755681932,-0.1494294852,-0.0151282949,-0.0904636607,-0.0794895366,0.3360760808,-0.0673165768,-0.5943489075,0.3753002584,-0.1992860883,-0.1545648128,-0.4956997633,-0.0426405221,0.1276846379,0.433683604,0.3025707603,0.4458981454,0.3535336852,0.0588151217,0.1667743474,0.0529164411,-0.2282678634,0.0254596267,-0.0179456957,-0.0743466988,0.4144559503,0.1811351031,0.0121966442,0.1405445039,0.4187540114,-0.5451748967,-0.1845691502,0.1008505449,-0.3514792323,-0.3432897627,-0.3676083088,0.1872015446,-0.5367393494,-0.0158157442,-0.0174540486,0.1968393773,-0.1532119215,0.2220282555,0.0514642969,0.0718086883,-0.3348800242,-0.3064831197,-0.1950988621,0.0791321099,-0.1510117799,0.2155078202,0.0868384466,-0.010045832,0.2042098194,0.0597339235,0.1382122189,-0.366733253,0.0484401695,-0.2148968875,-0.1565047055,-0.0596122667,-0.1266817898,-0.2758532465,-0.0598979481,-0.2508656383,-0.1477456987,-0.1589676142,-0.1754635721,-0.0020290541,0.4212627709,0.2016867548,-0.2421361953,-0.1874312162,-0.0104133161,0.293857336,-0.1305388212,-0.0845900849,0.1509460062,-0.2787753642,-0.324457109,-0.0561419316,0.3232987821,0.2026695162,-0.1975643933,-0.247759819,0.6061241031,0.040327698,0.0336668044,0.3809573352,0.1653750241,0.3888631761,0.0549484938,-0.0502892025,0.1929959655,-0.270024091,-0.2422163486,-0.0212492794,0.0832388774,-0.0768052712,0.1449465603,-0.1102794558,0.1698947251,0.0691119581,0.1431253254,0.2125446796,0.105232574,0.0390273184,-0.1694833636,0.0231248569,0.2083104551,0.182622835,-0.0441775657,-0.247535497,-0.0602207705,-0.5506232381,0.3593122065,-0.2336244881,-0.2559227049,0.3169041574,0.0317749605,0.1331079006,0.2627291381,0.0109732095,0.1177085638,-0.0579709969,-0.8542650342,0.2441546172,0.0617075711,-0.1692024767,-0.3705717921,-0.1805855781,0.300323844,-0.4371094406,0.1600407958,-0.0143354591,0.3412575424,0.2341626287,0.0369942561,0.4147161543,-0.1173598021,0.1140119284,-0.4648653269,0.00194202,-0.0095589487,-0.2148297876,0.3159090281,0.4534549713,0.1284361184,0.1765092313,-0.0494395345,0.4062876105,0.339990586,0.1041297242,0.1333085597,0.0938198641,0.0694995672,-0.0890278891,-0.1304594874,0.5995194912,0.124730587,0.1686528623,-0.3227340877,0.5980030298,-0.3787491322,0.046395205,0.1247782335,-0.1300297976,0.1045121178,0.105768241,-0.3970209956,0.0756273046,0.448471576,0.1915378124,0.5518527031,0.0433866121,-0.3257263303,-0.0421123505,0.3595901132,-0.1903334707,0.1797349006,0.0371973775,-0.0867273584,-0.1773941517,-0.021960875,-0.0783407465,0.0025477135,0.2195889205,0.0197215639,-0.0007429314,-0.1408705413,-0.4390758276,0.0764092132,-0.1429211348,-0.0605827421,0.3043664396,0.2423512638,0.0617750138,-0.4098843932,-0.2927568853,-0.3892460167,-0.042088937,-0.1957097948,0.1807440072,-0.0944136754,0.1418761909,-0.1394220442,-0.1282729506,-0.5086534023,-0.2359361649,0.2334479392,0.076352343,-0.1304826438,-0.0245747045,0.3039119542,0.3134801984,0.0245656967,0.0222486332,0.0049761296,-0.1759691834,-0.2359918803,0.146082893,-0.2856435776,-0.1068563312,0.0836335942,-0.3754407465,-0.0247780848,0.0332225077,-0.4409699142,0.2545731366,-0.0748175904,0.2755791545,0.2000907958,-0.3662665188,-0.0426011495,0.4201520383,0.1592188478,-0.3122100532,-0.2852800786,0.2487560064,-0.0951729193,0.2048257738,-0.1195692271,-0.412758857,0.0095930491,-0.1093281582,0.4913555086,0.1069434434,0.0888677239,0.2736880481,-0.0542880371,0.0500341319,-0.0098215146,0.2073385566,-0.1839255691,-0.0335972644,0.3422642648,-0.3354378641,-0.5135300159,-0.1427979469,0.0637115762,-0.201381281,0.2057778686,-0.1288751066,-0.2995453477,-0.0593241528,-0.0508388728,-0.055118192,0.082038857,0.43496418,-0.2504155934,-0.0757332966,-0.1027090251,0.0838338882,0.3216798306,-0.1883363128,0.2084263116,-0.158868283,0.3865533173,0.037282519,0.0425752699,0.6114041209,0.3491770029,-0.1800512373,-0.0786046907,0.0960942432,-0.2694198489,0.4766199291,0.2780717611,-0.0155186728,-0.1733074784,0.3527806699,0.0544787422,0.2919141054,-0.1228647456,-0.0032974731,-0.0022116969,-0.4111377299,0.0852207169,0.20457609,0.2007604241,-0.2420281172,0.2248528153,0.0827754289,-0.1478873789,0.1690865457,0.0599398911,0.4045791626,0.1620069444,-0.2010132074,0.2774576843,-0.4924324155,0.3232818246,0.1231625378,0.0037518095,0.0346342772,-0.2645382285,-0.0398273319,-0.1963596195,0.2720928788,-0.3343458176,-0.039872285,0.0852226987,-0.2270893157,-0.2626549304,-0.0394852906,-0.4668119252,-0.4294916987,0.1189720854,0.0257227942,-0.4162077904,0.140876919,0.2949523628,-0.0939940214,-0.0376576595,0.0707491934,-0.3008877337,-0.1250918508,0.1245121956,0.2273996174,0.0153751764,0.281619668,-0.1882916987,0.0350891948,0.0939707309,-0.1933543235,0.2874557972,0.3182862699,0.3587849438,-0.2967594564,-0.1858540624,0.1272571087,-0.1283434629,-0.1919141412,0.3262422979,-0.0692248046,-0.4199408591,0.2371010333,-0.2747509778,0.1988111585,0.3517254889,-0.0434976183,0.2319827527,-0.229451701,0.1682388932,-0.3555620313,-0.1029994413,-0.0069567328,0.1039782092,0.1005926058,-0.1379529536,0.3263968229,0.2669793069,-0.1933409423,-0.4485458136,-0.0590689331,-0.2420430034,0.4088690877,0.0335077159,0.9406026006,0.1146588251,0.2159931511,0.1509407312,-0.2938795686,0.4857125878,-0.4107207954,0.1659612656,-0.5914515853,-0.154063046,-0.0236732084,0.0152196316,0.0090304846,0.2254938781,-0.1336908042,0.527787149,0.1549180001,0.2343395948,0.0684799552,0.2131845355,0.0914743543,-0.1363695413,-0.2000812888,0.154839769,-0.0257828347,0.0350447595,-0.0692338496,0.016598925,-0.4950160086,0.0172243919,-0.001831464,0.0330695622,-0.2493982166,-0.040530581,0.2403572798,-0.6468834281,-0.362872988,0.4100298584,0.3001859486,0.3615040481,-0.1065691039,0.1137762144,-0.2730528116,0.2960500717,0.246479243,-0.0022383605,0.1830041707,-0.0635612085,-0.1858989596,-0.0480877645,0.1917336881,-0.3508482277,-0.2901604176,-0.0844340324,-0.3248621225,-0.1663077772,0.3957933784,-0.0780093148,-0.0337803066,-0.0825269297,0.1380172074,0.1615869403,0.021926837,0.0516949557,0.0785887018,-0.2714337707,-0.1466721594,0.0999819636,-0.0620632134,0.0250695739,-0.1733118445,0.1120142713,-0.0616239421,-0.2199273556,-0.0949828848,0.1216549501,-0.3167487085,0.0629125312,-0.0662863031,-0.1874003857,0.0848553628,0.454149574,0.1611184329,-0.0432567336,0.2365885079,-0.2844915986,-0.1792447716,0.4154628515,0.338074863,-0.0232430715,0.1925075352,-0.0853180066,-0.0687433556,0.092952542,-0.3633159697,-0.2334938198,-0.0807271153,-0.0226740334,-0.0575194135,-0.5269913077,0.0452138707,0.1302620173,0.0717526004,0.3796459734,-0.0466158986,-0.1393401474,-0.294829607,0.0863279402,-0.0445367508,0.2001894116,-0.0187490899,0.2959342599,-0.0253691785,-0.3763462007,0.2337668389,0.0076598921,0.3363522589,0.1284157336,-0.2368224561,0.3274032772,-0.2537733316,0.0768029615,0.0251885541,0.1082152799,0.006650384,0.0160390474,-0.111523971,-0.0190689284,-0.054685533,0.295034647,0.2611958981,0.0595220998,0.3306888342,-0.0011622612,-0.2251050174,-0.2662786245,0.5229455829,0.2373593152,-0.0539456271,-0.2033505142,-0.0982936099,0.2226356864,-0.2943924963,0.0724182054,0.0778271556,-0.0986256078,0.1442730278,0.4975623488,0.1614879221,-0.0648140907,0.1025289446,-0.0809710398,-0.0199590847,-0.0075993696,0.0267790705,-0.3497162759,0.240950495,0.0704931915,0.5140094757,-0.1148750708,0.2195925117,0.087900281,0.2013388127,0.0343093425,0.0216832869,0.4369777441,0.0568811931,0.1682256609,-0.1973410696,0.3928687572,0.1391462684,0.0145082278,-0.1283886284,0.2978079915,-0.1089272648,0.0387913845,-0.3456054628,-0.0703533366,0.0274058245,-0.4653739631,-0.2823560834,-0.2791447341,-0.4714227319,0.1094112843,-0.119545266,0.0771694705,0.1154011637,0.1312566251,0.1319941282,-0.2858121991,-0.3366743028,-0.1074163839,0.6733065248,-0.3037390411,0.2952612638,-0.040176753,0.2007911354,0.1491851062,0.3282278776,0.0973951668,0.0338949785,-0.201730907,-0.0645841286,-0.157863915,0.16612719,-0.0251393765,0.1989769638,0.2382318079,-0.1389902532,0.298578918,0.0918387845,-0.1330918521,0.2325234711,0.0410569422,0.3143246472,-0.2229613513,0.1764203161,0.1007426158,0.2382917553,-0.2092223167,-0.1073144302,-0.307757467,-0.1275487989,-0.4239983857,0.1715766191,0.0421766154,-0.1228059605,0.1027341038,0.1181352735,0.1602743268,-0.0267960634,0.2585558891,0.0192663874,-0.074778825,-0.1981179118,-0.0183102768,-0.0648413822,0.1657658219,-0.2037339509,0.3187336922,-0.0221429914,0.3684749603,-0.0281960927,0.2310350984,-0.0696298108,0.1423705518,-0.3201333284,0.1295230091,-0.3543378711,-0.2733688056,0.0231842417,0.1113182828,-0.0601809062,-0.0097250659,0.0663767532,0.1673560441,-0.3448729217,0.1648918986,0.1763146222,0.2991235256,0.4876230657,0.2544553876,-0.0168120358,-0.3954351544,-0.1899197847,-0.0650434196,-0.2933443487,0.4509838521,-0.2019188255,0.305975914,0.2800111771,-0.2772436142,-0.0059781666,0.7327027321,0.0470344834,-0.3717382252,-0.1414091438,-0.1147762686,0.1800146848,-0.2592808902,-0.3121394217,0.404166013,-0.0878885239,0.2905737758,-0.0122388499,-0.3303335309,0.1637311876,-0.3309814334,0.2920036614,-0.0520888939,0.388861984,0.449703902,0.0056046429,-0.309206754,0.0215132385,0.0359867439,0.1832429022,0.0314046107,0.1259144396,-0.0964607,-0.1132313758,-0.2234429121,0.4332582653,0.0445491597,-0.0269824769,0.0888876468,-0.359804213]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/617","title":"Compare different Rouge implementations ","comments":"Hi, thanks for the solution. \r\n\r\nI am not sure if this is a bug, but on line [510](https:\/\/github.com\/huggingface\/transformers\/blob\/99cb924bfb6c4092bed9232bea3c242e27c6911f\/examples\/seq2seq\/utils.py#L510), are pred, tgt supposed to be swapped?","body":"I used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n","comment_length":25,"text":"Compare different Rouge implementations \nI used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n\nHi, thanks for the solution. \r\n\r\nI am not sure if this is a bug, but on line [510](https:\/\/github.com\/huggingface\/transformers\/blob\/99cb924bfb6c4092bed9232bea3c242e27c6911f\/examples\/seq2seq\/utils.py#L510), are pred, tgt supposed to be swapped?","embeddings":[-0.0462399237,-0.3402347565,-0.0140987523,0.379104495,-0.0787526146,-0.6177638173,0.0081204483,0.0337361768,-0.3628861308,0.3499982357,-0.2657788098,-0.1302823424,0.1970662624,0.2637876868,0.260407567,-0.2404887378,0.0870682597,-0.015829904,-0.3073508739,-0.1472104639,-0.125898093,0.7282140255,-0.1187530011,0.1422054619,-0.1178556606,0.3190756738,0.0754643902,0.1428857297,-0.07691098,0.0503387451,-0.0773925483,0.1364061683,0.0535053238,0.6273003817,-0.0001120195,-0.0545938276,0.0474308021,-0.1357920468,0.069316633,-0.0228229351,-0.1206796393,-0.1080061793,-0.1051929966,-0.1453908086,-0.2766059041,-0.1439552605,-0.2889887393,0.2026909292,0.2424778342,0.4388734996,0.1830443293,0.037156608,0.1999942809,-0.1163281426,0.5332654715,-0.0113954572,0.0095240204,0.4414730668,0.0845701098,0.0121362219,-0.1068575233,0.2801780999,0.2016871572,-0.2245485634,0.0831993297,0.1899203956,0.0646672696,-0.0947259068,0.0609512553,0.4944716394,-0.033806514,-0.2444932461,-0.4565926492,0.0317237042,-0.2688064575,-0.334459424,0.0806375667,0.0214642473,0.3556173146,-0.0614489019,-0.2099113315,0.0630044863,-0.0412148051,-0.2200860828,-0.4216421545,-0.0584213063,-0.1846101433,0.0430524424,0.1635533422,-0.1405409724,-0.5604932904,0.2374844104,-0.1599112004,-0.1976955086,-0.3531385362,-0.1151283458,0.1373972148,0.3342143595,0.0371527448,0.4631068408,0.1254409701,0.1216638759,0.0856851414,0.0586970858,-0.3227404058,0.2932855487,-0.1890389472,0.172027275,0.355443418,0.0926138535,0.0749712437,0.0584116131,0.4156941772,-0.6989457011,-0.2002404928,0.0524525307,-0.167194441,-0.4326241612,-0.4299024642,-0.0271917507,-0.1767520905,-0.0308983214,0.1167363971,0.1503123492,-0.1495216936,0.0424442813,0.1605858803,-0.0055337274,-0.5098376274,-0.1614808142,-0.2780581713,0.1651213616,-0.1665259153,0.1229581088,0.0699660331,0.1001441777,0.2009737194,0.096145995,0.1871931851,-0.0009884766,0.1759388894,-0.0594167598,-0.1663112491,0.137147814,-0.1835516989,-0.0147337569,0.0321947448,0.0244991519,-0.1782720834,-0.2502812147,-0.4921041727,-0.1060738936,0.2954384089,0.1907461584,-0.2942804694,-0.0692167655,-0.1578979194,0.2298372984,-0.0357162841,-0.5121923685,0.0543526635,-0.3018949032,-0.3901182115,-0.1606630385,0.3361691236,0.1987190396,-0.1800176501,-0.1948147118,0.5356314778,-0.0672032684,0.110706456,0.4747158885,0.0492928214,0.0055499384,-0.174530983,-0.2681955695,0.1999701262,-0.4523166418,-0.3830577135,0.0476567633,0.0821941793,-0.2138216645,0.1366918385,-0.1346656978,-0.0554777682,-0.1711653918,0.145175308,0.0700774342,0.0844067633,0.0982008427,-0.2546145916,0.0316351429,0.2501725256,0.0316305794,-0.0061235502,-0.0144204451,-0.0123525774,-0.4560586512,0.1696414351,-0.0568991266,-0.1102579534,0.3962090313,-0.025027737,0.0397525132,0.3803070188,0.1711030155,-0.1741192788,-0.0921760648,-0.9340215921,0.1069794446,0.1560128778,-0.017484121,-0.4241196513,-0.2808053493,0.3445939422,-0.4135504663,0.1371365488,-0.0728048682,0.3419049084,0.1799040586,0.2065040767,0.3033022285,0.2343675196,0.0536500812,-0.3724871576,0.1102484018,-0.0694722682,-0.0858074278,0.2577384412,0.3375180662,0.2463002056,-0.022180615,-0.0713821277,0.4225411415,0.2885462344,0.3854725659,0.3367461562,0.4276495278,0.2519129813,-0.2109387517,-0.1562893838,0.5646339059,0.1081745997,0.0383515358,-0.4680090547,0.5065370202,-0.2661363482,-0.012103837,0.1127237529,-0.0463253818,0.0814128891,0.091087237,-0.3921841085,0.1565761864,0.291778326,0.3464295566,0.6304529309,0.1406931132,-0.0967639983,0.1004109383,0.3997428119,-0.1017156243,-0.0067254175,0.1135695502,-0.1569254696,-0.2928291857,0.1107685864,-0.0893801227,0.0598920658,0.2097007781,0.1487043947,0.0371913947,-0.1481204927,-0.3523471653,0.0504463166,-0.1629639864,-0.215439558,0.1390164644,0.1744087487,0.0353199206,-0.24279733,-0.2507183254,-0.2253893763,0.1806869805,-0.2774429619,0.2457858622,0.0038503595,0.2574873269,-0.074709937,-0.3186459541,-0.3954653144,-0.2954204082,0.2648135722,0.0942249298,-0.1397356391,0.08062388,0.2114585042,0.1704198271,0.1263626665,-0.0609066598,-0.2691183984,0.0670473799,-0.3442390859,0.1047258601,-0.2446593642,-0.0788622051,0.08067099,-0.429623872,0.0975872129,-0.1162439436,-0.5498434305,0.1776524335,0.0356237218,0.3770739138,0.1881416589,-0.3002074361,-0.0184461065,0.4710696638,0.2002328038,-0.3231211007,-0.1657702625,0.3012695312,-0.2839270234,0.2294908613,-0.2235215604,-0.3297278583,0.0143189803,-0.0862902552,0.2834471166,0.1472921818,0.0925280601,0.2298971713,0.064915359,-0.1279930472,0.3870963156,0.1898573488,-0.3099277914,-0.2520888746,0.3893834054,-0.350076139,-0.5373863578,-0.2022144198,0.0193111096,-0.0161610916,0.2700971365,-0.1697930992,-0.3686186373,-0.3688879013,-0.1145493239,0.0201562624,0.1319062859,0.4350543618,-0.1713933051,-0.1104349494,-0.2714865506,0.052683685,0.1724307984,-0.0486953072,0.219272241,-0.0858299881,0.2024034709,-0.0671014264,0.3914042413,0.4602386057,0.1244401336,-0.0473767966,-0.1539421827,0.2863889933,-0.1996272355,0.2588322759,0.1851464808,0.2490403801,-0.061674051,0.3672216833,0.2207795978,0.2248217911,-0.1708768308,0.2625947297,-0.1587223858,-0.3359983265,-0.0841349512,0.3974853754,0.1898150593,-0.0296094436,0.0822151229,-0.0941661149,-0.1437630653,0.2429082096,0.1454965472,0.3763527572,0.2039707899,-0.1141542792,0.198753953,-0.6007112861,0.4602781236,-0.0555289499,-0.0449594446,-0.0817770585,-0.1381318718,-0.0442636944,-0.2185066938,0.2711826563,-0.2302881628,0.047912851,0.1509724706,-0.4011395872,-0.3824820518,-0.0977538154,-0.4528385401,-0.2298427373,0.2018155754,0.2029628307,-0.5372371674,-0.0735699087,0.0599508062,-0.1079464629,-0.1870282143,0.2506589293,-0.2709747553,-0.2244587988,0.2324512601,-0.0348381624,0.0112213539,0.2210061699,-0.2367370725,0.1307562739,0.2223874927,-0.1823404282,0.2928214669,0.2518833876,0.483147651,-0.2346968204,-0.1617089957,0.1818938255,-0.0874767303,0.0406564809,0.334354341,0.0864176974,-0.2289610505,0.4154287875,-0.0384389833,-0.0139006274,0.2383216918,-0.1080386117,0.0137072895,-0.1961783767,0.3133017123,-0.427752763,-0.0448636077,0.2073244303,0.1391386837,-0.0488648154,-0.2333494872,0.2371948659,0.3299573064,-0.2377563268,-0.2284530848,0.0542810224,-0.200175941,0.1643245369,0.0023741995,1.0280889273,0.1047327071,0.2302791178,0.2844590545,-0.3621208966,0.2803664804,-0.4588991404,0.1027164459,-0.5660182238,-0.2690147758,-0.0265094992,-0.0703129321,-0.1234997883,-0.1072537974,-0.1337423027,0.4592174888,0.1445995569,0.1331365407,-0.0539653152,0.1167642325,0.1683036834,0.0048505748,0.1752374768,0.1566207558,0.0639585406,0.2994736433,-0.1798166484,-0.0644773915,-0.5498411059,-0.0796332061,-0.0305482242,-0.1351214796,0.0250814781,0.2051972598,-0.0030890766,-0.5478059053,-0.4278310835,0.4673524201,0.3638890088,0.1543482989,-0.2965244055,0.168576479,-0.4701938033,0.3899487257,0.0743408203,0.1156048551,0.0724762082,-0.0994993597,-0.0437542833,-0.0537650846,0.1490102112,-0.2506878376,-0.6245259047,-0.1295894682,-0.5371041894,-0.3104643226,0.5415868163,-0.0805539191,-0.1622419059,-0.0381718166,0.1411792636,0.2169870287,-0.0275866278,0.3249526918,0.0393105038,-0.4234240055,-0.1621247381,0.3064823151,0.0121897357,-0.0430703051,-0.0259228665,0.0730394125,0.0162068866,-0.1115276217,-0.1409133226,0.1159520671,-0.4342990518,0.2139217556,-0.0991930664,-0.025300283,0.1100037396,0.4212907851,0.2335816473,0.0027090081,0.2833829224,-0.1890311539,-0.314574182,0.4179911911,0.1682231873,0.1861319691,0.0181394089,-0.1133889481,0.1064385176,0.0218627695,-0.3656166196,0.0785602257,-0.0261600632,0.0308772046,-0.1768048257,-0.4855466187,0.1199008524,0.1508072168,0.0167068485,0.0663632825,-0.2826023102,-0.1009670272,-0.3385397196,0.0959462747,-0.1007133797,-0.0454229079,-0.0591188967,0.2864868939,-0.0228679124,-0.2581018806,0.2471311539,0.1842892915,0.3011257052,0.141524449,-0.1482146978,0.0849044472,-0.1509348452,0.1452336907,0.3150301576,0.0186424404,-0.0567418225,0.0243451763,-0.1893321276,-0.1604849249,0.0903149098,0.1853345484,0.2324611694,0.0039607738,0.3097013533,0.0568216816,0.0860439092,-0.1042043418,0.6678875685,0.1300587952,-0.092155911,-0.0505931117,-0.1103082225,0.1822526902,-0.1761769652,0.0433171727,0.1966630667,0.0390417278,0.1300091296,0.3549925685,0.2215704918,0.0620077029,0.4383097589,-0.0137921646,0.1863242239,-0.1359265298,-0.066981934,-0.2568200529,-0.0755849853,0.2242777795,0.5384526253,-0.0414161496,0.2130909115,0.2482748181,0.1608664095,0.0557454862,-0.0978935957,0.165934667,0.1322992444,-0.0367492735,-0.35957551,0.2460223287,0.1544196457,0.0825498104,-0.0643204227,0.2218964547,-0.1357979327,0.1727041751,-0.3701008856,0.014339908,-0.1660201699,-0.4361305237,-0.1228215769,-0.3909063041,-0.2636038065,0.0903260261,-0.092061244,0.0655294284,0.2612419128,0.0224434957,-0.0183030814,-0.0816773996,-0.4220667481,0.055042278,0.4008122683,0.0048687882,0.2830740213,0.1144412234,0.0623963512,0.2156860828,0.3841763437,0.1169887558,-0.0475933217,-0.1749850214,-0.0584877916,0.1209590808,0.1552110761,-0.1109542623,0.2898354232,0.1443898827,-0.308845371,0.1412970275,0.1111898869,-0.1289555132,0.2223316133,-0.2095448673,0.2623898983,-0.2080373466,0.1602804661,-0.1594026685,0.1173656806,-0.1060334519,-0.188470155,-0.3168028891,-0.1629392803,-0.3557917178,0.1665894836,0.1886276007,-0.0896939486,0.0911407694,0.1815398186,0.0131882308,0.0751547739,0.2211915553,-0.0654822364,-0.1103832349,-0.2407244444,0.1426548213,0.0655630305,0.3086328506,-0.2443654537,0.2922245562,-0.0720491931,0.1290788203,0.3750624061,0.1218327731,0.1178663149,0.2167738676,-0.3038434088,0.1798717231,-0.3895221651,-0.1668026149,0.0952639431,0.0987606645,0.057372354,0.0368031636,0.1220119894,-0.1249684319,-0.1437060535,0.1236508787,0.219398886,0.3691051602,0.521009028,0.2177212834,-0.1560079455,-0.1310566515,-0.1762015969,-0.114132151,-0.0443406142,0.1646036953,-0.3079448044,0.0857178271,0.1169650778,-0.5630621314,-0.1023762971,0.6308071017,0.0810671225,-0.3098793328,-0.1672507972,-0.0631320253,0.027787555,-0.3235362172,-0.2412043661,0.5562442541,0.0396711417,0.2578611672,-0.108027555,-0.3692850471,0.4570891559,-0.337315768,0.2436262369,-0.1098594517,0.3534843624,0.4649350941,-0.006232278,0.0651132539,0.1634217203,0.0482459702,0.1417987347,0.0699143782,0.1563726664,-0.0514429733,-0.1225459278,-0.1321689188,0.6715171337,0.1798330247,-0.1915443242,0.2135783732,-0.1527657807]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/617","title":"Compare different Rouge implementations ","comments":"This looks like a bug in an old version of the examples in `transformers`","body":"I used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n","comment_length":14,"text":"Compare different Rouge implementations \nI used RougeL implementation provided in `datasets` [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/metrics\/rouge\/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https:\/\/arxiv.org\/pdf\/1909.03186.pdf) for example.\r\nCan you make sure the google-research implementation you are using matches the official perl implementation? \r\nThere are a couple of python wrappers around the perl implementation, [this](https:\/\/pypi.org\/project\/pyrouge\/) has been commonly used, and [this](https:\/\/github.com\/pltrdy\/files2rouge) is used in fairseq). \r\nThere's also a python reimplementation [here](https:\/\/github.com\/pltrdy\/rouge) but its RougeL numbers are way off. \r\n\nThis looks like a bug in an old version of the examples in `transformers`","embeddings":[-0.140459314,-0.1666320115,-0.0553813092,0.3105680048,-0.149799183,-0.6479229331,0.2100273371,0.1188927218,-0.2537645996,0.3880321681,-0.0877828896,-0.0299108252,0.0843544006,0.1594460309,0.2132623494,-0.2509655356,0.1060832068,0.0105334604,-0.3251257241,-0.1782176644,-0.0890656561,0.6648769379,-0.061421141,0.0988482684,-0.1389408708,0.3187175989,0.1271303743,0.0459250323,-0.1077860072,0.0159843825,-0.094622843,0.0917224213,0.1321463734,0.5899976492,-0.0001094993,0.0157451816,0.1682445556,-0.0624584332,0.0470181927,-0.018273754,-0.1216712445,-0.2771709263,-0.1256235391,-0.2079274505,-0.1945716143,-0.1707388908,-0.265355438,0.2898335755,0.2048230469,0.3818970919,0.2263326794,-0.0042516924,0.1527947187,-0.1370229423,0.5805529356,0.1356012225,-0.0316655338,0.284219712,0.0450859852,-0.0527756065,-0.1867342442,0.324550122,0.2612424195,-0.2018049657,0.0724962279,0.2060273588,0.0747279674,-0.0663036332,0.0913229361,0.4707176387,0.0233657584,-0.1266350448,-0.489599824,0.0418849364,-0.3643784523,-0.2244064659,0.0833456293,-0.0110313483,0.3246656656,-0.1701801568,-0.1815205961,0.1783741415,-0.064758718,-0.2148351967,-0.4338491559,0.0277846511,-0.1527896076,0.0546473339,0.2302142531,-0.159029305,-0.3135477602,0.1870863885,-0.0669977441,-0.1427354217,-0.3490323722,-0.0911592022,0.2589839995,0.235766381,0.0674767271,0.3427844346,0.1162336841,0.0155994464,0.1035922542,0.1199542284,-0.2748477757,0.2085956186,-0.2685882449,0.2685345411,0.41790241,0.1459230632,0.0934004933,0.0385954157,0.3267495334,-0.5915048122,-0.1274477094,0.1129143089,-0.2300903201,-0.3986751139,-0.4284497499,-0.0359057002,-0.1448966712,-0.0074620582,0.1447640955,0.2116010487,-0.1558980942,0.1693690568,0.0817303956,-0.0755675286,-0.3663372099,-0.161319375,-0.2818154991,0.1702909917,-0.2420499921,-0.0504421704,0.1341944784,0.1889711171,0.2005510926,0.0232765172,0.1577116996,-0.0325178206,0.3207731247,-0.1235968322,-0.1431233585,0.1563705206,-0.1899004877,-0.060989026,0.0123061286,-0.0718971491,-0.1436998099,-0.2365230024,-0.3871378899,-0.0814949349,0.3326226175,0.23539491,-0.2534626722,-0.0449967682,-0.2175577432,0.2076278776,-0.0137538761,-0.5182260871,-0.0316572599,-0.3485686481,-0.2829996943,-0.1514376402,0.2765402198,0.1790951341,-0.1195084453,-0.1744687408,0.3528505266,-0.0298555028,0.1209404245,0.460804671,-0.0477344729,0.1285098791,-0.1554632783,-0.3561001718,0.2292717248,-0.4726471603,-0.4487233758,0.098542653,0.0920476466,-0.2741014361,0.163638249,-0.0543407537,0.156550318,-0.1553846896,0.1627203673,0.1870120019,0.0996166393,0.1863832325,-0.1037271917,0.0179649834,0.3783858418,0.0370759442,0.124311924,0.0043320027,0.1280157864,-0.3829339147,0.1957710385,-0.0211014748,-0.1177803576,0.3883109391,-0.0810940266,-0.0843806639,0.4730922282,0.1003664806,-0.041200459,-0.0953178108,-0.932923317,0.0115268379,0.0609400533,-0.0260806214,-0.3884185255,-0.1697681546,0.3657742441,-0.3250147402,0.1779837608,-0.121784687,0.2849994302,0.0537274405,0.1716847271,0.3031176329,0.1910099536,-0.0329288356,-0.2859045267,0.1995883435,-0.0860240832,-0.0311399549,0.2784815133,0.2441465855,0.2374116331,0.0078564659,-0.0418153182,0.301055938,0.3818162084,0.2707740664,0.4018670917,0.5288132429,0.2022923231,-0.22041291,0.0105669852,0.682559669,0.1541398019,0.0756026879,-0.6245843768,0.3878796995,-0.2864593565,0.0218402892,0.1345091611,0.0416975543,0.2145663798,0.125926882,-0.3729444742,0.1620088369,0.2599889338,0.3659470677,0.5652703643,0.0797621906,0.023756383,-0.0223620385,0.4147714376,-0.1673337072,0.1158461496,0.1196923032,-0.2158836722,-0.2814480066,0.0423129499,0.1184287295,-0.011801742,0.2057365626,0.1355306655,-0.1048883572,-0.1813168377,-0.2795502543,0.0790659711,-0.1222576424,-0.2349578291,0.2068306208,0.2374871969,0.0216496866,-0.3385473192,-0.4927966297,-0.1319088042,0.1682430059,-0.1311575919,0.2452857494,0.1315874308,0.3513325453,-0.0500165112,-0.2913888097,-0.3512006402,-0.2989152968,0.1913448423,0.1012569815,-0.236868307,0.0062926118,0.2165568918,0.174219057,0.1517305225,-0.0063204947,-0.1217777207,-0.0028830515,-0.2498833686,0.1547408551,-0.3555179834,-0.1370898783,0.0151836267,-0.3733907342,0.1072005108,-0.1417631209,-0.5271461606,0.1599103808,0.0694267228,0.1767097712,0.2472654283,-0.3571827114,0.051858034,0.4219163954,0.1763402671,-0.4172794223,-0.0968092754,0.3796289861,-0.2984886765,0.1824893355,-0.1590180993,-0.3141199946,-0.0487632789,-0.1140678898,0.216879487,0.2070592046,0.114015758,0.3755501211,0.0943801031,-0.0652550235,0.4522592127,0.0725239217,-0.2860426009,-0.160229519,0.4685708284,-0.325740844,-0.583126843,-0.2247514129,-0.0959939435,-0.21045959,0.3764496744,-0.0859153867,-0.2828555405,-0.3474115133,-0.1848403066,0.0478576794,0.1752902865,0.4554934204,-0.0593205541,-0.1585807055,-0.259126693,0.0869004652,0.133362785,0.0391326658,0.2878628075,-0.2082828432,0.1774826199,-0.0289999098,0.3227186799,0.4356993139,0.1382726133,-0.1102497652,-0.0465923585,0.2375448048,-0.1244765222,0.3198717833,0.1858309507,0.1360304952,-0.2313198149,0.2614921927,0.2246889919,0.0275757741,-0.1081355512,0.1774949282,-0.0751854852,-0.2988441288,-0.0620236173,0.3878315091,0.2875950336,-0.0108029228,0.1198496446,-0.1270591766,-0.1759402901,0.3057725132,0.057769794,0.3418780863,0.0763313249,-0.1282086819,0.2223423868,-0.5552514195,0.3345120251,0.0034007397,-0.035866227,-0.0494262204,-0.0936667621,-0.1167508215,-0.2071635872,0.418795526,-0.3053105175,0.1183129326,0.1568459719,-0.3314479589,-0.490508467,-0.1315387785,-0.5271983147,-0.1681184471,0.066269584,0.1094376147,-0.4069784582,0.0757463127,0.0755857602,-0.028494291,-0.1376329064,0.2367667854,-0.3199499547,-0.1647028923,0.2301306576,-0.0741211623,0.0616876259,0.1909112781,-0.1816860437,-0.0029250809,0.1423424929,-0.1735828668,0.1682807207,0.2642251849,0.4746654034,-0.2311570495,-0.151428625,0.1915057003,-0.1537448317,0.0010695052,0.2898985744,0.0700860918,-0.3075793982,0.4397188723,-0.1389171481,-0.0520110317,0.1567512751,-0.1381558627,-0.0701358616,-0.1340237856,0.2845629454,-0.3592057526,0.0099666715,0.125578016,0.2509867847,-0.0201491639,-0.2465477735,0.2323414236,0.2913419306,-0.2732951045,-0.1077323332,-0.1257587373,-0.1864857227,0.361958921,-0.0481471792,0.8872042298,0.1305503398,0.2061064392,0.1292333156,-0.3247427642,0.2913177609,-0.627014935,0.1491041929,-0.5457879901,-0.3125401735,-0.0628122017,-0.1228341088,-0.1515688449,-0.0876082703,-0.1433578581,0.4506320655,0.0110827051,0.2384200692,-0.0270395521,0.1372466534,0.1373324394,-0.0530092604,0.1951386929,0.2209436446,0.1261187941,0.2572033107,-0.1929622889,-0.0507814959,-0.5379667282,-0.112677753,-0.0320926532,-0.0386576839,0.1749575883,0.3920626044,0.076835908,-0.4889959693,-0.4969815016,0.4634498656,0.2521160245,0.1133243665,-0.2704553306,0.2213884145,-0.6116552949,0.432575047,0.0852647349,0.093595311,0.003086976,-0.1200725958,-0.0867533311,-0.037977539,0.0616706274,-0.2570353746,-0.402227819,0.0451146103,-0.5482831001,-0.2410297692,0.6665737033,-0.0371829756,-0.3550752997,-0.075047709,0.1873496324,0.226628691,-0.1244190484,0.3558245301,0.0965182409,-0.3208404779,-0.0744342878,0.2973371744,0.0736225769,0.0455231369,-0.0885564834,-0.0254566409,0.0029294083,-0.1764137,-0.1845006049,0.0570728183,-0.3326191902,0.2488167286,-0.2575883567,-0.1470044255,0.0640050769,0.3307500482,0.2159747481,0.0328276306,0.080697909,-0.3202642202,-0.3630156219,0.3839605451,0.1443711519,0.0858181864,-0.0343972333,-0.2342176884,0.2249379009,-0.0274526272,-0.3845295906,0.0147956014,-0.0490898229,0.1008044183,-0.2878468037,-0.5978873968,0.175905183,0.0339905657,0.0445986092,0.0768869445,-0.2370624393,-0.1731183231,-0.3442937136,0.0728561655,-0.082426101,-0.0302475411,-0.098523207,0.1997029483,-0.065100424,-0.2678388655,0.1854625046,0.2121262699,0.2411040217,0.0900132582,-0.174793914,0.0168741532,-0.1733338982,0.0960159451,0.3030309975,0.0049456344,0.0276710577,0.0468240045,-0.0140855042,-0.0952168405,-0.0796341151,0.200768441,0.199640736,-0.0460666418,0.4824616313,0.0616698749,0.1044236869,-0.1122396663,0.6408972144,0.1171953976,-0.0763341486,-0.0306168348,-0.1639598012,0.2521238029,-0.2336520106,-0.1958367974,0.1895846426,0.0778738186,0.1405644268,0.2735181749,0.2996415496,-0.0402463935,0.3364651501,-0.0438093357,0.2142460942,-0.1532717794,-0.0398328416,-0.2394900024,-0.0024303426,0.2620862126,0.4848607779,-0.1418489814,0.2315233797,0.4170896709,0.2315203547,0.068332836,-0.0535917617,0.1876856834,0.0663209781,0.0585886873,-0.2804885209,0.1473125815,0.1691794395,0.0715128183,-0.0331155173,-0.0190528445,-0.1150934249,0.342561394,-0.4064356685,-0.0979620516,-0.1164645106,-0.3692030907,-0.0296394117,-0.3280085623,-0.3222164214,0.0211851522,0.0127775157,0.1143509597,0.2493865192,0.0425938256,0.0610232763,-0.0388201773,-0.2169640958,0.157073766,0.4833334386,-0.092377767,0.1519687027,0.1248638183,0.1286894679,0.1089349389,0.2941928208,0.1496448815,-0.0210881941,-0.1819668263,-0.0438222215,0.0163531937,0.0222242586,-0.2347956896,0.2804518938,0.1001922563,-0.2225456685,0.1461078376,0.1460948586,-0.1917055547,0.1861434877,-0.1703131795,0.1885773391,-0.1269661635,0.2490632385,-0.0790562704,-0.0058247615,-0.1160630211,-0.1424039006,-0.2969498038,-0.2562906146,-0.2089937925,0.1151152328,0.1614728421,-0.0665084049,0.106802769,0.0499309599,-0.0055007548,0.0550657026,0.3586404622,-0.0733449757,-0.0848831758,-0.1756513268,0.0053798771,-0.0561794862,0.2387333065,-0.1711377203,0.3506372571,-0.1625084132,0.1431863904,0.3953286409,0.1969376653,0.06627132,0.1920223236,-0.2419171929,0.1664195061,-0.3209232688,-0.0830787867,0.0728232786,-0.0677918196,-0.0238168649,0.0634296238,0.2007507682,-0.1379491091,-0.267044425,0.0755303651,0.2180370241,0.397223711,0.4823642969,0.3243678212,-0.1961496323,-0.1460195482,-0.2011576295,-0.117978327,0.0040793163,0.087112233,-0.2644960284,0.1056127474,0.1152235717,-0.503434062,-0.0775501207,0.5961461663,0.1305889636,-0.1809835136,-0.2484394908,-0.0947576389,-0.0782276168,-0.1786729097,-0.3747298419,0.4665136933,-0.0113817342,0.1560565829,0.0009976829,-0.3328554034,0.3972482383,-0.3965352178,0.0717057511,-0.1677920967,0.3951998055,0.404414475,0.0822079256,0.0593160763,0.1665311009,0.0266806465,0.0770976022,0.0073117428,0.0022869471,0.0094990944,-0.0595953129,-0.1005195975,0.5810002685,0.1054085791,-0.1928119957,0.1333193779,-0.18812567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"I have the same issue","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":5,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\nI have the same issue","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"Same issue here when Trying to load a dataset from disk.","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":11,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\nSame issue here when Trying to load a dataset from disk.","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"I am also experiencing this issue, and don't know if it's affecting my training.","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":14,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\nI am also experiencing this issue, and don't know if it's affecting my training.","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"Same here. I hope the dataset is not being modified in-place.","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":11,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\nSame here. I hope the dataset is not being modified in-place.","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"I think the only way to avoid this warning would be to do a copy of the numpy array before providing it.\r\n\r\nThis would slow down a bit the iteration over the dataset but maybe it would be safer. We could disable the copy with a flag on the `set_format` command.\r\n\r\nIn most typical cases of training a NLP model, PyTorch shouldn't modify the input so it's ok to have a non-writable array but I can understand the warning is a bit scary so maybe we could choose the side of non-warning\/slower by default and have an option to speedup.\r\n\r\nWhat do you think @lhoestq ? ","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":106,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\nI think the only way to avoid this warning would be to do a copy of the numpy array before providing it.\r\n\r\nThis would slow down a bit the iteration over the dataset but maybe it would be safer. We could disable the copy with a flag on the `set_format` command.\r\n\r\nIn most typical cases of training a NLP model, PyTorch shouldn't modify the input so it's ok to have a non-writable array but I can understand the warning is a bit scary so maybe we could choose the side of non-warning\/slower by default and have an option to speedup.\r\n\r\nWhat do you think @lhoestq ? ","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"@thomwolf Would it be possible to have the array look writeable, but raise an error if it is actually written to?\r\n\r\nI would like to keep my code free of warning, but I also wouldn't like to slow down the program because of unnecessary copy operations. ","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":46,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\n@thomwolf Would it be possible to have the array look writeable, but raise an error if it is actually written to?\r\n\r\nI would like to keep my code free of warning, but I also wouldn't like to slow down the program because of unnecessary copy operations. ","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"@AndreasMadsen probably not I would guess (no free lunch hahah)","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":10,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\n@AndreasMadsen probably not I would guess (no free lunch hahah)","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"@thomwolf Why not? Writable is checked with `arr.flags.writeable`, and writing is done via magic methods.","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":15,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\n@thomwolf Why not? Writable is checked with `arr.flags.writeable`, and writing is done via magic methods.","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"Well because I don't know the internal of numpy as well as you I guess hahahah, do you want to try to open a PR proposing a solution?","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":28,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\nWell because I don't know the internal of numpy as well as you I guess hahahah, do you want to try to open a PR proposing a solution?","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"@thomwolf @AndreasMadsen I think this is a terrible idea, n\/o, and I am very much against it. Modifying internals of an array in such a hacky way is bound to run into other (user) issues down the line. To users it would not be clear at all what is going on e.g. when they check for write access (which will return True) but then they get a warning that the array is not writeable. That's extremely confusing. \r\n\r\nIf your only goal is to get rid of warnings in your code, then you can just use a [simplefilter](https:\/\/docs.python.org\/3.8\/library\/warnings.html#temporarily-suppressing-warnings) for UserWarnings in your own code. Changing the code-base in such an intuitive way does not seem like a good way to go and sets a bad precedent, imo. \r\n\r\n(Feel free to disagree, of course.)\r\n\r\nIMO a warning can stay (as they can be filtered by users anyway), but it can be clarified why the warning takes place.","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":155,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\n@thomwolf @AndreasMadsen I think this is a terrible idea, n\/o, and I am very much against it. Modifying internals of an array in such a hacky way is bound to run into other (user) issues down the line. To users it would not be clear at all what is going on e.g. when they check for write access (which will return True) but then they get a warning that the array is not writeable. That's extremely confusing. \r\n\r\nIf your only goal is to get rid of warnings in your code, then you can just use a [simplefilter](https:\/\/docs.python.org\/3.8\/library\/warnings.html#temporarily-suppressing-warnings) for UserWarnings in your own code. Changing the code-base in such an intuitive way does not seem like a good way to go and sets a bad precedent, imo. \r\n\r\n(Feel free to disagree, of course.)\r\n\r\nIMO a warning can stay (as they can be filtered by users anyway), but it can be clarified why the warning takes place.","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"> To users it would not be clear at all what is going on e.g. when they check for write access (which will return True) but then they get a warning that the array is not writeable. That's extremely confusing.\r\n\r\nConfusion can be resolved with a helpful error message. In this case, that error message can be controlled by huggingface\/datasets. The right argument here is that if code depends on `.flags.writable` being truthful (not just for warnings), then it will cause unavoidable errors. Although, I can't imagine such a use-case.\r\n\r\n> If your only goal is to get rid of warnings in your code, then you can just use a simplefilter for UserWarnings in your own code. Changing the code-base in such an intuitive way does not seem like a good way to go and sets a bad precedent, imo.\r\n\r\nI don't want to ignore all `UserWarnings`, nor all warnings regarding non-writable arrays. Ignoring warnings leads to hard to debug issues.\r\n\r\n> IMO a warning can stay (as they can be filtered by users anyway), but it can be clarified why the warning takes place.\r\n\r\nPlain use cases should really not generate warnings. It teaches developers to ignore warnings which is a terrible practice.\r\n\r\n---\r\n\r\nThe best solution would be to allow non-writable arrays in `DataLoader`, but that is a PyTorch issue.","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":222,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\n> To users it would not be clear at all what is going on e.g. when they check for write access (which will return True) but then they get a warning that the array is not writeable. That's extremely confusing.\r\n\r\nConfusion can be resolved with a helpful error message. In this case, that error message can be controlled by huggingface\/datasets. The right argument here is that if code depends on `.flags.writable` being truthful (not just for warnings), then it will cause unavoidable errors. Although, I can't imagine such a use-case.\r\n\r\n> If your only goal is to get rid of warnings in your code, then you can just use a simplefilter for UserWarnings in your own code. Changing the code-base in such an intuitive way does not seem like a good way to go and sets a bad precedent, imo.\r\n\r\nI don't want to ignore all `UserWarnings`, nor all warnings regarding non-writable arrays. Ignoring warnings leads to hard to debug issues.\r\n\r\n> IMO a warning can stay (as they can be filtered by users anyway), but it can be clarified why the warning takes place.\r\n\r\nPlain use cases should really not generate warnings. It teaches developers to ignore warnings which is a terrible practice.\r\n\r\n---\r\n\r\nThe best solution would be to allow non-writable arrays in `DataLoader`, but that is a PyTorch issue.","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"> The right argument here is that if code depends on `.flags.writable` being truthful (not just for warnings), then it will cause unavoidable errors. Although, I can't imagine such a use-case.\r\n\r\nThat's exactly the argument in my first sentence. Too often someone \"cannot think of a use-case\", but you can not foresee the use-cases of a whole research community.\r\n \r\n> I don't want to ignore all `UserWarnings`, nor all warnings regarding non-writable arrays. Ignoring warnings leads to hard to debug issues.\r\n\r\nThat's fair.\r\n\r\n> Plain use cases should really not generate warnings. It teaches developers to ignore warnings which is a terrible practice.\r\n\r\nBut this is not a plain use-case (because Pytorch does not support these read-only tensors). Manually setting the flag to writable will solve the issue on the surface but is basically just a hack to compensate for something that is not allowed in another library. \r\n\r\nWhat about an \"ignore_warnings\" flag in `set_format` that when True wraps the offending code in a block to ignore userwarnings at that specific step in [_convert_outputs](https:\/\/github.com\/huggingface\/datasets\/blob\/880c2c76a8223a00c303eab2909371e857113063\/src\/datasets\/arrow_dataset.py#L821)? Something like:\r\n\r\n```python\r\ndef _convert_outputs(..., ignore_warnings=True):\r\n ...\r\n with warnings.catch_warnings():\r\n if ignore_warnings:\r\n warnings.simplefilter(\"ignore\", UserWarning)\r\n return torch.tensor(...)\r\n# continues without warning filter after context manager...\r\n```","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":198,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\n> The right argument here is that if code depends on `.flags.writable` being truthful (not just for warnings), then it will cause unavoidable errors. Although, I can't imagine such a use-case.\r\n\r\nThat's exactly the argument in my first sentence. Too often someone \"cannot think of a use-case\", but you can not foresee the use-cases of a whole research community.\r\n \r\n> I don't want to ignore all `UserWarnings`, nor all warnings regarding non-writable arrays. Ignoring warnings leads to hard to debug issues.\r\n\r\nThat's fair.\r\n\r\n> Plain use cases should really not generate warnings. It teaches developers to ignore warnings which is a terrible practice.\r\n\r\nBut this is not a plain use-case (because Pytorch does not support these read-only tensors). Manually setting the flag to writable will solve the issue on the surface but is basically just a hack to compensate for something that is not allowed in another library. \r\n\r\nWhat about an \"ignore_warnings\" flag in `set_format` that when True wraps the offending code in a block to ignore userwarnings at that specific step in [_convert_outputs](https:\/\/github.com\/huggingface\/datasets\/blob\/880c2c76a8223a00c303eab2909371e857113063\/src\/datasets\/arrow_dataset.py#L821)? Something like:\r\n\r\n```python\r\ndef _convert_outputs(..., ignore_warnings=True):\r\n ...\r\n with warnings.catch_warnings():\r\n if ignore_warnings:\r\n warnings.simplefilter(\"ignore\", UserWarning)\r\n return torch.tensor(...)\r\n# continues without warning filter after context manager...\r\n```","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"> But this is not a plain use-case (because Pytorch does not support these read-only tensors).\r\n\r\nBy \"plain\", I mean the recommended way to use `datasets` with PyTorch according to the `datasets` documentation.","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":33,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\n> But this is not a plain use-case (because Pytorch does not support these read-only tensors).\r\n\r\nBy \"plain\", I mean the recommended way to use `datasets` with PyTorch according to the `datasets` documentation.","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/616","title":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors","comments":"This error is what I see when I run the first lines of the Pytorch Quickstart. It should also say that it should be ignored and\/or how to fix it. BTW, this is a Pytorch error message -- not a Huggingface error message. My code runs anyway.","body":"I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.","comment_length":47,"text":"UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors\nI am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this strange Userwarning without a stack trace:\r\n\r\n> Set __getitem__(key) output type to torch for ['input_ids', 'sembedding'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n> C:\\Users\\bramv\\.virtualenvs\\dutch-simplification-nbNdqK9u\\lib\\site-packages\\datasets\\arrow_dataset.py:835: UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors. This means you can write to the underlying (supposedly non-writeable) NumPy array using the tensor. You may want to copy the array to protect its data or make it writeable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ..\\torch\\csrc\\utils\\tensor_numpy.cpp:141.)\r\n> return torch.tensor(x, **format_kwargs)\r\n\r\nThe first one might not be related to the warning, but it is odd that it is shown, too. It is unclear whether that is something that I should do or something that that the program is doing at that moment.\r\n\r\nSnippet:\r\n```\r\n dataset = Dataset.from_dict(torch.load(\"data\/dummy.pt.pt\"))\r\n print(dataset)\r\n tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\n keys_to_retain = {\"input_ids\", \"sembedding\"}\r\n dataset = dataset.map(lambda example: tokenizer(example[\"text\"], padding='max_length'), batched=True)\r\n dataset.remove_columns_(set(dataset.column_names) - keys_to_retain)\r\n\r\n dataset.set_format(type=\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\n dataloader = torch.utils.data.DataLoader(dataset, batch_size=2)\r\n\r\n print(next(iter(dataloader)))\r\n```\r\n\r\nPS: the input type for `remove_columns_` should probably be an Iterable rather than just a List.\nThis error is what I see when I run the first lines of the Pytorch Quickstart. It should also say that it should be ignored and\/or how to fix it. BTW, this is a Pytorch error message -- not a Huggingface error message. My code runs anyway.","embeddings":[0.2382437438,-0.3691007793,0.0651260018,0.1298748255,0.4380830228,0.0914000422,0.6286588907,0.2343869954,0.2976013124,0.0604875013,-0.1968635917,0.4237814248,-0.3091077805,-0.2186422497,-0.256572783,-0.2071904987,-0.019928148,0.1609406769,0.0264154281,-0.0268473886,-0.0556126274,0.0349227302,-0.1734228432,0.3379645646,-0.2255875468,-0.4098304808,0.2081689388,-0.0633970499,-0.0799697861,-0.3318474889,0.1405587047,-0.2741777301,-0.0424051508,0.1276783198,-0.0001266726,0.1157198995,0.0211008638,-0.0969710872,-0.0743481889,-0.1103790775,0.7791628242,-0.3006570041,0.2598286569,-0.5041794181,-0.0403003171,-0.2053057849,-0.0674014911,-0.2887423038,0.6341741681,0.2032751888,0.0752525479,0.5057702661,0.1049243435,0.3881487846,-0.066052407,0.3452899456,-0.2550107539,-0.0293222163,0.0295194145,0.1392835528,-0.0000625259,0.2642535865,-0.5914075375,0.0171453189,0.152644217,-0.0334382728,-0.349470824,-0.1997036189,-0.0568565764,0.4435260892,0.3706563115,-0.2254174352,-0.046209719,-0.3330504894,-0.1216260865,-0.1587676555,0.1649687737,0.047861699,-0.1130082533,0.1626306921,0.0142950015,-0.0069782236,-0.1369986981,0.2631300986,0.0543763451,0.1673310399,-0.1042311937,0.3020676374,-0.070318155,-0.0098416703,0.0525531098,0.0074251285,0.1901581138,0.196217075,-0.1164147928,-0.3119843602,-0.2599380016,-0.7474357486,-0.0870942324,-0.2492239624,0.3341346085,0.2540561259,-0.1607192904,0.0971489623,0.1469385177,0.0135778785,0.1208697632,0.303039968,-0.0211054645,-0.2641310394,0.4715952277,0.0567326099,0.1360289454,-0.2395026833,0.0944854543,0.0855747908,0.0345738009,0.0435557812,-0.4489809275,0.1235723495,-0.2449784279,0.0512567237,-0.0606115498,0.043743331,-0.0110608209,0.008455541,0.0809402019,0.3331671059,-0.2172211111,0.2093106508,0.0050662151,0.2291956097,-0.1825448275,-0.1943188906,-0.0641829967,0.2339443862,0.0189591032,0.261413455,0.0302922912,0.0009383697,0.1234879196,-0.2359409928,0.6530778408,0.2942767441,-0.1781089008,0.2948439717,0.3621653318,-0.0307599809,-0.3972631097,0.2834928632,-0.3104486465,-0.2276656926,-0.5769506097,0.0020175837,-0.1900073141,0.1614836454,0.1787664145,-0.2559919059,0.6716828942,-0.1626638323,0.4028995037,-0.437556386,-0.0927184001,-0.3298016787,-0.0530223027,0.3662085533,-0.3259876072,-0.064038679,0.258772701,-0.0162683986,0.4336455464,0.4094212055,-0.1056687534,0.1454505175,-0.1220321953,0.2255036682,0.3607753515,-0.1224518269,-0.0684338585,0.188791126,-0.1807031184,0.3171916902,0.1701442301,0.0854641721,-0.1042878255,0.1792410761,0.2188509107,0.0108668376,0.3695318699,-0.0156879947,-0.3240776658,-0.2027052045,0.4832918048,-0.1357165277,-0.0672293007,-0.1775234044,-0.1403667331,0.0221992601,0.3671857715,-0.1119876876,0.0290920548,-0.0411325768,-0.167381376,0.2396441549,-0.140446946,0.1050849631,-0.1710649133,-0.0152083375,0.0745915994,0.1085796431,-0.2709115148,-0.1766555607,-0.105172202,0.0071568051,-0.3517931402,-0.0465408675,-0.1111049056,-0.1885502487,-0.1136639938,0.1317765117,-0.1785905063,-0.1946374923,-0.5005659461,0.1517338455,-0.4120471478,-0.1020534709,-0.1012726128,-0.212443158,-0.3982954323,0.3267838955,0.232444495,-0.035569869,-0.0395304337,0.0819807947,0.2321104854,-0.1997635812,-0.917568922,0.1638941765,0.1763911694,-0.090930514,0.2472078055,0.1546784341,-0.1466421038,-0.096070841,-0.0257543679,0.6306298971,-0.0646008551,0.0608098134,-0.5003480315,0.261929512,-0.1911876351,0.1908106506,-0.1049912944,-0.4325117767,-0.0143134892,0.1304034591,-0.2320648581,-0.0843060389,-0.4490361214,-0.1313851327,0.0598223768,0.0488689318,0.2781392932,-0.080569014,0.1933736503,-0.2550162077,0.2152756453,-0.5420799255,0.1550077051,-0.1119838879,-0.0120686553,-0.0852294937,-0.1558951139,-0.2969710827,-0.0197968502,0.0393947251,-0.0165408477,0.1850436181,-0.0722502619,0.1588746011,-0.3456683457,0.1977573484,-0.1231665015,0.0587118678,-0.498372376,0.2802625299,-0.1423191577,0.0969332159,-0.2249715179,-0.2606532276,0.0266311578,-0.0663827807,-0.0077088084,0.0847671181,-0.2427747101,0.1668353975,-0.0719061345,0.0822579786,0.1976013184,-0.1022716314,-0.0642502904,0.090562731,-0.0684159249,0.021913901,-0.0892241076,-0.1524734795,0.2516383827,-0.0162396878,-0.2449783832,-0.100350745,-0.2109436393,0.1382278502,-0.1868399531,-0.1263592839,0.3529059291,0.1456685066,-0.2808859944,-0.2442344129,-0.0112630334,-0.0327507965,0.0331098884,0.1553820074,0.0217528194,0.13360475,-0.0648512319,-0.1858778149,-0.423663497,-0.0582789257,0.1405140013,-0.1072303206,0.4788049161,0.2815490067,0.25558272,0.154078871,-0.0803942084,0.2228560597,-0.2713088393,-0.004655215,0.314692229,-0.1752711385,-0.1823733896,-0.2065906823,0.1696434915,0.0142171262,0.0761997998,-0.3303349316,-0.1466784775,-0.0949350446,0.120766893,-0.2022130787,-0.1867319793,0.241856724,0.228496924,0.0300903972,0.1885070801,0.2500557899,0.0332654417,-0.2170521766,-0.0861481056,-0.043779742,0.1183455214,0.2379999757,0.4617770314,-0.0420077182,-0.2846428156,0.1601472199,0.0010739247,0.5781360269,-0.1647348851,-0.3022521436,0.0216061492,0.0767859966,0.0009509391,-0.1333919168,-0.3836979866,-0.0001525639,0.2423481941,0.0825087354,0.1085122973,-0.2214963287,0.4477949739,-0.3171699345,0.0087407725,-0.184847936,0.7597178221,-0.1085421294,-0.2938226461,-0.1163830236,0.2062700391,0.2668504715,-0.0254098065,-0.4820899963,-0.33496207,-0.320260644,0.2719300687,0.1470129639,0.6929338574,-0.1348299086,-0.2319810539,0.3621513844,0.2398638278,0.2768639624,0.0184221789,-0.21673581,0.2924694121,0.0791204423,-0.3491106033,-0.1140718311,-0.0095900726,0.10742075,-0.008857673,0.4121200442,-0.1308528483,-0.224253118,0.1573220044,0.18824175,-0.1615239233,0.0613937974,-0.2136746645,-0.2623850107,-0.3775331378,0.1652603447,0.2398947924,0.2294291407,-0.2508634627,0.0488715582,-0.0084929885,0.2456668019,0.3178789616,-0.0956501886,0.1522449106,0.1483534276,-0.3418127298,0.3096505404,0.3398795128,-0.0277442764,-0.1095983088,-0.1247121915,0.1159105226,0.1117187664,-0.0217788871,0.2172193527,0.4647905231,0.1280823648,0.4410915375,-0.1509378701,-0.109577395,-0.1212495044,0.1708003283,0.4047023654,-0.0328604095,-0.31236884,-0.4118332267,0.2598879039,0.1042165533,-0.063375935,0.3375407159,-0.0369729847,-0.1301325858,0.6268120408,0.7443354726,1.2204271555,-0.2085440904,0.5956696868,0.3273388743,-0.0674326643,0.6187640429,-0.0102658663,0.0424034074,-0.2683011591,0.1548624933,-0.1980779767,-0.1577159613,0.2015676051,-0.0019605998,-0.0411037952,-0.0536004789,-0.5109754205,0.3334375322,0.0068873172,0.0026783019,-0.1011275947,-0.2555357218,0.2363882661,-0.1040957049,-0.0791460425,0.3209193349,-0.035044264,0.0317475349,0.1873889118,-0.276840508,-0.2541232705,-0.0294097736,0.104480207,0.5604252815,0.011987716,-0.6092010736,0.288367182,0.4530893266,0.380844146,0.1035281196,0.0959937498,-0.1774269491,0.2179634422,-0.3489566147,-0.0485725515,-0.0852727145,0.4823373258,0.1212893352,-0.0862490386,0.4521552324,-0.0263082851,-0.1532792151,-0.3243345916,0.0973289087,0.4660355449,-0.6561595798,-0.0559296906,-0.2054635137,0.0001007449,-0.062172208,-0.0448218994,0.2190756947,0.0646790937,0.1037345827,-0.2519076169,-0.4340581,-0.0367065184,0.2222432941,0.2893690169,-0.1179222539,0.8795531392,0.0664856508,-0.3234813511,-0.1874033362,0.0371702835,0.2153204679,-0.1092455164,-0.0018606688,0.1581522226,-0.1175392792,-0.3384714723,0.3617875278,0.0083906306,0.2841091454,-0.3869198263,-0.1965322047,-0.4224758148,-0.0321242064,0.2779535353,0.3617933095,0.0696609989,0.1996619999,-0.1247273609,-0.1013154536,-0.1560569406,-0.0752204508,0.0202211682,0.0376657322,0.0879771262,-0.031965889,-0.0273690168,-0.1754416227,-0.0822161213,0.0666393414,-0.1285916269,-0.0146796834,-0.2238384783,0.2570936382,0.1262628734,-0.4150517583,-0.1527789533,-0.2237043977,0.0586874336,-0.1579180211,0.1590578705,0.5828675032,0.106329836,0.3782195747,-0.2470912635,0.5658977628,-0.0748700127,-0.0877132118,-0.3030565977,0.0571165234,-0.0390429869,-0.032036595,0.0130811473,-0.0603507608,0.0060969284,0.1890248805,0.2629617751,0.0186595172,-0.0052156285,-0.0120392656,-0.1070672199,-0.3753265738,-0.0517683066,0.1893552244,-0.2479217201,-0.5765469074,0.2642939389,0.0123997042,0.0169079881,0.0137433065,-0.433645457,0.0795572028,-0.0310480967,0.1702327877,0.0805722475,0.1417361498,-0.0181903541,0.0981455594,0.2180118412,-0.3502313793,0.1294845194,0.4828864336,-0.2283924967,-0.0452727899,0.4522112012,-0.0312916078,0.2341943085,-0.1334193945,0.0915355012,0.4512062371,0.1692256778,-0.0341868773,-0.3178960979,-0.3285262883,0.1474927217,0.2378775775,0.097056739,0.356723547,0.2927555442,0.401902169,-0.4451820552,-0.020946689,0.0176274572,0.236085102,-0.3745492399,-0.0451738238,-0.3808806539,0.0096501093,0.0484978519,-0.1260535568,-0.1597313136,0.0313257836,0.0340275392,-0.089949064,-0.2625426948,0.1525455415,0.2857149541,0.0725883096,0.187876299,-0.2299559563,-0.18709144,0.0890278295,0.1067848876,-0.0448153988,0.0374787226,0.4447658062,0.3587413728,-0.1035056785,-0.0842831284,-0.1757000089,-0.0139693785,0.1993029863,0.4816280603,0.0186823793,0.0138761038,0.2879366577,-0.0183105506,-0.0949819461,0.0105205821,0.2952638268,0.2227564156,0.1748559475,0.3795814514,-0.0568387136,-0.4856645167,0.0033314647,-0.0617628731,-0.1641254127,0.1717436165,0.3564147055,0.2143037021,0.1797748953,-0.1373207867,0.0054709427,0.3706022501,0.4430555701,0.2518374324,-0.0295529477,-0.3219549358,0.3352751434,-0.3268108964,0.4080525935,-0.0225135982,-0.3146441579,-0.0225457698,-0.1043131277,0.3645895123,0.0440228879,-0.3006745279,0.1305758804,-0.2481038123,0.0429529957,-0.004868452,0.1862841249,0.0995331779,-0.0594904311,-0.0888574347,-0.3397574723,0.4442159235,0.1383636594,-0.0315220468,-0.4207597077,0.1421369463,-0.2364388108,0.2782038152,0.3304501772,0.1277323365,0.4267577231,-0.0317621566,-0.1237796992,-0.0141851408,0.0862124041,-0.1058098525,0.2842879295,0.2063486129,0.2731178999,0.0385229923,0.0542448908,-0.2951193452,0.5469299555,-0.1160326302,-0.3950821459,-0.0497787036,0.2726551592,-0.2424457371,-0.135364756,0.0816417411,0.235943377,0.3239526749,0.2044604272,-0.3263271749,-0.3508826494,0.5388799906,-0.3649677932,-0.1342417747,-0.1359106153,0.086955972,0.3645115793,0.1180236042,-0.7096754313,-0.0097282091,0.236346513,-0.1331491619,-0.040950045,-0.1120019704,-0.1369571537,-0.1903096288,-0.2534228563,0.6608255506,-0.057132192,-0.042245701,-0.1349737197,-0.3494077623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/615","title":"Offset overflow when slicing a big dataset with an array of indices in Pyarrow >= 1.0.0","comments":"Related: https:\/\/issues.apache.org\/jira\/browse\/ARROW-9773\r\n\r\nIt's definitely a size thing. I took a smaller dataset with 87000 rows and did:\r\n```\r\nfor i in range(10,1000,20):\r\n table = pa.concat_tables([dset._data]*i)\r\n table.take([0])\r\n```\r\nand it broke at around i=300.\r\n\r\nAlso when `_indices` is not None, this breaks indexing by slice. E.g. `dset.shuffle()[:1]` breaks.\r\n\r\nLuckily so far I haven't seen `_indices.column(0).take` break, which means it doesn't break `select` or anything like that which is where the speed really matters, it's just `_getitem`. So I'm currently working around it by just doing the arrow v0 method in `_getitem`:\r\n```\r\n#if PYARROW_V0:\r\ndata_subset = pa.concat_tables(\r\n self._data.slice(indices_array[i].as_py(), 1) for i in range(len(indices_array))\r\n)\r\n#else:\r\n #data_subset = self._data.take(indices_array)\r\n```","body":"How to reproduce:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nwiki = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\")\r\nwiki[[0]]\r\n\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-13-381aedc9811b> in <module>\r\n----> 1 wikipedia[[0]]\r\n\r\n~\/Desktop\/hf\/nlp\/src\/datasets\/arrow_dataset.py in __getitem__(self, key)\r\n 1069 format_columns=self._format_columns,\r\n 1070 output_all_columns=self._output_all_columns,\r\n-> 1071 format_kwargs=self._format_kwargs,\r\n 1072 )\r\n 1073 \r\n\r\n~\/Desktop\/hf\/nlp\/src\/datasets\/arrow_dataset.py in _getitem(self, key, format_type, format_columns, output_all_columns, format_kwargs)\r\n 1037 )\r\n 1038 else:\r\n-> 1039 data_subset = self._data.take(indices_array)\r\n 1040 \r\n 1041 if format_type is not None:\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.take()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in take(data, indices, boundscheck)\r\n 266 \"\"\"\r\n 267 options = TakeOptions(boundscheck)\r\n--> 268 return call_function('take', [data, indices], options)\r\n 269 \r\n 270 \r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.call_function()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.Function.call()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: offset overflow while concatenating arrays\r\n```\r\n\r\nIt seems to work fine with small datasets or with pyarrow 0.17.1","comment_length":108,"text":"Offset overflow when slicing a big dataset with an array of indices in Pyarrow >= 1.0.0\nHow to reproduce:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nwiki = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\")\r\nwiki[[0]]\r\n\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-13-381aedc9811b> in <module>\r\n----> 1 wikipedia[[0]]\r\n\r\n~\/Desktop\/hf\/nlp\/src\/datasets\/arrow_dataset.py in __getitem__(self, key)\r\n 1069 format_columns=self._format_columns,\r\n 1070 output_all_columns=self._output_all_columns,\r\n-> 1071 format_kwargs=self._format_kwargs,\r\n 1072 )\r\n 1073 \r\n\r\n~\/Desktop\/hf\/nlp\/src\/datasets\/arrow_dataset.py in _getitem(self, key, format_type, format_columns, output_all_columns, format_kwargs)\r\n 1037 )\r\n 1038 else:\r\n-> 1039 data_subset = self._data.take(indices_array)\r\n 1040 \r\n 1041 if format_type is not None:\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.take()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in take(data, indices, boundscheck)\r\n 266 \"\"\"\r\n 267 options = TakeOptions(boundscheck)\r\n--> 268 return call_function('take', [data, indices], options)\r\n 269 \r\n 270 \r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.call_function()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.Function.call()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: offset overflow while concatenating arrays\r\n```\r\n\r\nIt seems to work fine with small datasets or with pyarrow 0.17.1\nRelated: https:\/\/issues.apache.org\/jira\/browse\/ARROW-9773\r\n\r\nIt's definitely a size thing. I took a smaller dataset with 87000 rows and did:\r\n```\r\nfor i in range(10,1000,20):\r\n table = pa.concat_tables([dset._data]*i)\r\n table.take([0])\r\n```\r\nand it broke at around i=300.\r\n\r\nAlso when `_indices` is not None, this breaks indexing by slice. E.g. `dset.shuffle()[:1]` breaks.\r\n\r\nLuckily so far I haven't seen `_indices.column(0).take` break, which means it doesn't break `select` or anything like that which is where the speed really matters, it's just `_getitem`. So I'm currently working around it by just doing the arrow v0 method in `_getitem`:\r\n```\r\n#if PYARROW_V0:\r\ndata_subset = pa.concat_tables(\r\n self._data.slice(indices_array[i].as_py(), 1) for i in range(len(indices_array))\r\n)\r\n#else:\r\n #data_subset = self._data.take(indices_array)\r\n```","embeddings":[-0.3078599274,-0.1624817699,-0.0384140052,0.1917552799,0.1244793832,-0.0056269262,0.2420143485,0.1383444518,-0.3982105851,0.361923933,0.3079869449,0.4601465166,0.0280525908,0.0383535996,0.0241384003,-0.0889734328,-0.0426646657,0.0609558672,-0.1887903959,-0.0936412439,-0.0416765921,-0.3072259426,-0.2502094805,-0.0529374555,-0.1218299344,-0.2309742272,0.1539252698,0.1009820625,-0.1768930852,-0.2024064958,0.3342114687,-0.2671542764,0.0081014773,-0.0058169933,-0.0001175907,0.105093196,0.3507690728,0.0269848593,-0.0879454762,0.36186257,-0.1608607024,-0.3452382684,0.0494743697,-0.4793946147,0.2600331306,-0.1769219935,-0.2077457309,0.0676089823,0.0854020566,0.2236685753,0.1557555348,-0.0401188843,0.1585735977,0.2635318935,0.3263535202,0.0234108511,-0.0072753155,-0.2422338426,-0.167885378,0.0705335438,-0.2242981493,-0.0016125748,-0.0947054923,-0.0284246001,0.0539905503,0.0067401864,-0.023341896,0.0207042638,-0.1426187307,0.0567588471,0.2461706698,-0.2803447843,-0.1456433535,-0.2259111553,-0.0329124369,-0.3547538817,0.1401233077,0.3031680286,-0.0155505668,0.0339879505,-0.2461449653,0.2101866305,-0.2522290647,0.2505633831,-0.1889606565,0.2698267996,0.1512380987,0.1663444042,0.1357590854,-0.1141221076,0.4339353442,-0.1705885679,-0.0047746915,0.1696425527,-0.2869728804,0.1074186414,0.1948883533,-0.1316955835,0.500872314,-0.3496707678,0.0059454655,0.044077985,0.2339591086,0.0869400725,0.2841058373,0.1922837794,-0.5885927081,0.1779753715,-0.0207448937,0.0505632646,0.0146464435,0.0611452945,0.2007251531,-0.1749119461,0.3466876447,-0.080063574,0.0683387741,-0.1202837303,-0.2903328538,0.1888431013,-0.4950132966,0.3146744967,0.2096391916,-0.0979561433,0.0071371985,-0.0668994933,-0.2257014215,0.0162179675,-0.2137298733,0.0277413074,-0.1721847653,-0.0789436623,-0.097892873,-0.1193299741,0.0237431806,0.1679458022,-0.0813873485,0.3148367107,0.3989399672,-0.1049106643,-0.1097599268,-0.1568699181,0.4402474463,0.2657408416,-0.037249025,0.0719270855,-0.1062029973,0.0841025859,-0.0842686966,0.5176148415,-0.2612574995,-0.3617993593,-0.1042305902,0.1721643656,-0.1236087084,0.1671472192,-0.2392052561,0.1035816893,0.2674862444,-0.0605446771,0.186840266,-0.1679096073,-0.0465435795,-0.1870518327,0.2324185967,0.0651653484,-0.5403259993,0.3515728414,0.046457205,0.2777175009,0.3522078097,0.5056605935,-0.3103647828,-0.0813853443,-0.2419678718,0.054232087,0.1743374467,0.0021221105,-0.8825092912,-0.0902406722,-0.1311735809,-0.0398749858,0.2627778053,-0.1125264093,0.2651942372,0.0228159875,-0.0259930156,0.4810886681,-0.2747821212,0.2685475051,-0.4711420536,-0.1202219874,0.0109113958,0.0561424904,0.0007090906,-0.4467764199,-0.0522028096,-0.2801512182,0.2582900226,0.0480177514,0.2383665144,0.2446468771,0.0595670231,-0.2582804263,0.1781593412,-0.2147070318,-0.0670889765,-0.0253214389,-0.0054617655,-0.0934444442,-0.3762587905,-0.2153983414,-0.1107436791,0.3079881072,-0.1378659904,0.2697454691,0.1490946114,-0.3101868927,0.3051522672,0.1536781639,0.144846037,-0.1100105867,-0.1788464487,-0.0944767296,-0.2770965993,0.3967710435,-0.1217573658,-0.3430443704,-0.0394438095,0.3696358204,0.1240580752,-0.1210040003,-0.08161515,0.247591719,0.1220882535,-0.1240401193,-0.1544880718,0.201361239,0.1901596636,-0.6690385938,0.1178058013,-0.0190334246,-0.0259280931,0.073074013,-0.214742884,0.3537632227,-0.3796572685,0.1163655147,-0.0941120759,-0.0308949854,0.1070786119,-0.0249294546,0.1990388334,-0.0057956716,-0.1353929043,0.2408878505,-0.0255299453,-0.05827583,0.085350886,0.2082442045,0.4890706241,-0.1404571235,-0.0235346928,0.0426869988,-0.2933749259,-0.0679957122,0.1969684809,-0.2779559195,-0.0518844575,0.4408052564,0.0995412171,-0.234200567,-0.0678505599,-0.1258074045,0.236642167,0.3833583295,0.0870453045,0.1480712146,0.4960047007,-0.1378153265,-0.3675361574,0.0245736167,0.2212060094,0.4458927512,-0.1369953752,0.3641352952,-0.1575205475,-0.2291455865,-0.1539961249,-0.5479347706,-0.0495434105,-0.5558484793,-0.1171553135,0.2539521158,-0.2222500741,0.3176632822,0.028670188,-0.0788918287,0.4207021296,-0.2142774016,-0.0134648597,-0.2432038486,-0.1864147335,0.0226918533,0.3362286985,0.0649512634,0.0128789227,0.4299613833,0.0976502746,-0.0092566181,-0.3197989464,0.0494458601,0.1751476824,0.2215363234,0.0291967578,-0.2809969187,-0.3562772572,-0.2338242084,-0.0001732999,0.1419413835,-0.1411358565,0.4679219723,-0.1558834314,0.0448507592,0.008400552,-0.0066716825,-0.0028689625,-0.3170768321,-0.106845133,-0.0941461399,0.146625787,0.1291240901,0.2173149884,0.0217559095,0.1971845925,-0.3625352681,-0.2118215263,0.0719039589,0.3074546754,-0.0073131048,-0.1995082647,-0.1844004095,-0.4688217044,-0.120662719,0.3121152222,-0.43527022,0.1149598956,-0.165476054,0.5477545261,-0.4002652466,0.0760845542,0.1913095415,0.1314117759,-0.0754614696,-0.1794867814,-0.0041376762,0.0360306315,-0.1388119757,-0.0993702859,-0.0103422347,0.35558483,0.1919984221,0.6874073744,0.182875067,-0.0826884285,0.1499190331,-0.103769809,0.0607788451,-0.0972274169,-0.2453378737,0.1346120238,0.2759500742,-0.2638726234,0.0069096205,-0.0608507358,-0.3277964294,0.3485786915,-0.1818958968,-0.1257545352,-0.3210374415,0.2611432076,-0.1045739502,0.0411942899,0.1294791102,-0.1142401099,-0.1942567676,-0.2883715928,0.0373693146,-0.1895458251,0.2498661429,-0.1295817643,0.0257857721,-0.165540725,-0.108615227,0.0432148464,0.2535862625,0.1290031672,0.2761618495,-0.0774622336,0.214434281,-0.087756373,1.052972436,0.1035215929,0.0862727538,-0.0843415782,0.3270575702,-0.5469658375,0.045502834,-0.3395446539,0.4172218144,0.1493502408,0.0795401409,-0.2137515992,0.1312854886,0.6030226946,0.2673247755,-0.0757204667,-0.1873874664,-0.1004454195,-0.1217115819,0.2724308074,0.3786853552,0.0342793576,0.3319784701,0.0867872387,-0.2884069383,-0.4074480236,-0.1794433147,-0.43626073,0.0392235778,0.3683811128,-0.2323276997,0.0919802189,-0.2026369274,-0.1233670264,0.5874598622,0.276170522,0.1106542721,-0.1481909901,-0.114576228,0.1147041842,0.296895653,-0.1863588691,0.0224657096,-0.00429929,-0.0666999668,0.2870538831,-0.0751791298,-0.2727073133,-0.0565402173,0.0645271242,-0.0863803253,0.0467156321,0.535130024,-0.0962854698,-0.0200151559,0.2606781125,0.4524257481,-0.2280533165,0.2177149504,0.2168310881,0.6593003273,-0.0582131296,-0.0079521947,0.3936114609,0.0186922234,0.3969335556,0.1132596657,0.2290652841,-0.2633156776,-0.2673666477,0.1683330089,-0.2955305874,-0.2516088188,0.0523138903,-0.1367732882,0.0307242963,-0.1729731709,0.0640651882,-0.1427879781,-0.0798084587,0.1897869259,-0.2027525604,-0.3587452769,0.0043281298,0.0839706734,-0.4553226531,-0.0062032384,-0.0558736995,-0.0727324113,-0.0626004785,-0.2099886984,-0.0158432331,-0.3024449944,0.2118612826,0.1587907523,-0.6693410277,0.0959851071,-0.1027466953,-0.0909658521,-0.003784352,0.0125186779,-0.0979868025,0.0787511766,-0.1514410079,-0.093180418,0.1588275582,0.3725361526,0.1117558777,-0.3891111612,0.0792206377,0.1022791415,-0.060583692,0.0916546956,0.1698605716,0.1858794838,-0.3513030708,0.0801825821,-0.2744289935,-0.0276282039,-0.2533023357,0.0510548986,-0.0321576819,-0.3189631402,0.3559044302,0.1106919721,-0.1777253002,-0.0625040457,0.3362928033,0.1386843175,0.3069220185,0.4120994806,0.2120607942,-0.2112435251,-0.0923749208,0.2197960466,0.2482061535,-0.2466107905,-0.1335739493,-0.3357232809,0.2979923785,0.0040589105,-0.1968062818,0.0943774879,0.2139061689,-0.4765343964,-0.2514716685,-0.2843219638,0.211540699,0.0218784977,-0.1386735141,0.1883488894,-0.1097696796,-0.1664049178,0.0380355753,-0.2583703101,0.2426104844,0.015891118,0.2187202126,-0.2371936738,-0.1162924543,0.0712876394,-0.1096849814,0.0967871845,0.0323481523,-0.0849000812,-0.2896191776,-0.0747356862,0.068775028,0.1048449799,-0.2686823308,-0.1610892564,-0.0881018713,-0.197155416,-0.0708957389,0.199687615,-0.040991839,-0.0030111203,0.0278002545,0.6172444224,-0.2400876284,-0.4222017229,-0.0536632203,-0.0288906079,0.1443918198,0.1050772145,0.508179903,0.0785977244,-0.0513165146,-0.138573572,0.0988981724,0.0796134993,-0.0999370441,0.4434025586,-0.3727136254,-0.2157142907,0.248406738,0.3717080653,0.2496133149,-0.0126267867,-0.3776681721,0.1215229258,0.2283975482,-0.032063961,-0.2781687975,0.1422376782,0.1158116311,0.1839352548,0.3635499477,-0.045644626,-0.3120152354,0.0750242844,0.0862318501,0.5116409659,-0.0492971055,-0.2612501085,0.5733863711,-0.1759341806,-0.216320619,0.3265039623,-0.0425852649,0.5641289949,0.6425901055,-0.0141267804,0.2967489958,0.2060596198,-0.3082503974,-0.0604013316,-0.126505807,-0.2884389758,0.256336391,-0.0069088591,0.1854427606,0.1793788224,0.210778147,0.4293079674,0.0859497711,0.0137428362,-0.0070639001,-0.2963067591,0.0951610506,-0.3648280501,-0.0339967124,0.1075608581,-0.0892927498,-0.0629893318,0.1336776316,0.5363907814,0.073693946,0.0753084496,-0.1682991832,0.472537905,0.3535591662,0.2539380789,-0.1741780341,0.3685833514,0.3450315893,0.0600092635,-0.3759499192,0.2836781442,0.4476296008,0.0661306903,-0.2917457521,-0.1754899472,0.1431080997,-0.1993087679,-0.3571977913,0.0562635325,-0.02232638,0.269728303,0.1601201296,0.1058418378,-0.1833210886,0.0344817415,0.5079106092,-0.0918377116,-0.1485922933,-0.3756018281,-0.0136666046,-0.4743428826,0.0703353211,-0.0579412803,0.0409056284,0.0677907988,0.1807413399,-0.0876699984,0.0531846061,0.2307295203,0.035296768,-0.3808389604,0.7067388892,0.5049709082,0.2196255922,-0.2470973134,0.0792752951,0.0175036658,0.4386537969,-0.2930751443,-0.0511545241,0.5098760724,0.3039680719,-0.1780640036,0.2502115369,0.1689564884,0.1316731125,-0.19541502,0.4371808171,-0.5424678922,-0.2461664677,0.0678663552,-0.4354547262,-0.1111012176,-0.4641906321,0.2038557976,-0.0356637537,0.0513345823,-0.0334171131,0.237526536,-0.1416114271,0.1121383458,0.2387252599,0.4911783934,0.1525250673,0.0015406716,0.1109365895,-0.1395282894,-0.1394063681,-0.5160037279,0.5627593994,-0.1268523186,0.2740676701,-0.1731084436,-0.7187899351,0.1381969154,0.1420859098,0.1563244611,0.0118768979,0.0317550935,0.1862273067,-0.2070299238,0.0572317094,-0.0017846138,0.0907566845,0.1579876691,0.2495679855,-0.0379342251,-0.2104880363,0.1006878912,-0.2144787163,-0.1678662449,-0.0442154594,0.1772096902,0.2280100584,-0.3719458282,-0.254488498,0.0146524208,0.4591278434,-0.1570975333,-0.1533426195,0.1323523521,0.1654485613,0.0945475921,0.0275233816,0.2726256847,0.0628592223,-0.3466641903,-0.0776627585,-0.2243146002]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/615","title":"Offset overflow when slicing a big dataset with an array of indices in Pyarrow >= 1.0.0","comments":"Let me know if you meet other offset overflow issues @joeddav ","body":"How to reproduce:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nwiki = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\")\r\nwiki[[0]]\r\n\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-13-381aedc9811b> in <module>\r\n----> 1 wikipedia[[0]]\r\n\r\n~\/Desktop\/hf\/nlp\/src\/datasets\/arrow_dataset.py in __getitem__(self, key)\r\n 1069 format_columns=self._format_columns,\r\n 1070 output_all_columns=self._output_all_columns,\r\n-> 1071 format_kwargs=self._format_kwargs,\r\n 1072 )\r\n 1073 \r\n\r\n~\/Desktop\/hf\/nlp\/src\/datasets\/arrow_dataset.py in _getitem(self, key, format_type, format_columns, output_all_columns, format_kwargs)\r\n 1037 )\r\n 1038 else:\r\n-> 1039 data_subset = self._data.take(indices_array)\r\n 1040 \r\n 1041 if format_type is not None:\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.take()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in take(data, indices, boundscheck)\r\n 266 \"\"\"\r\n 267 options = TakeOptions(boundscheck)\r\n--> 268 return call_function('take', [data, indices], options)\r\n 269 \r\n 270 \r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.call_function()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.Function.call()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: offset overflow while concatenating arrays\r\n```\r\n\r\nIt seems to work fine with small datasets or with pyarrow 0.17.1","comment_length":11,"text":"Offset overflow when slicing a big dataset with an array of indices in Pyarrow >= 1.0.0\nHow to reproduce:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nwiki = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\")\r\nwiki[[0]]\r\n\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-13-381aedc9811b> in <module>\r\n----> 1 wikipedia[[0]]\r\n\r\n~\/Desktop\/hf\/nlp\/src\/datasets\/arrow_dataset.py in __getitem__(self, key)\r\n 1069 format_columns=self._format_columns,\r\n 1070 output_all_columns=self._output_all_columns,\r\n-> 1071 format_kwargs=self._format_kwargs,\r\n 1072 )\r\n 1073 \r\n\r\n~\/Desktop\/hf\/nlp\/src\/datasets\/arrow_dataset.py in _getitem(self, key, format_type, format_columns, output_all_columns, format_kwargs)\r\n 1037 )\r\n 1038 else:\r\n-> 1039 data_subset = self._data.take(indices_array)\r\n 1040 \r\n 1041 if format_type is not None:\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.take()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/compute.py in take(data, indices, boundscheck)\r\n 266 \"\"\"\r\n 267 options = TakeOptions(boundscheck)\r\n--> 268 return call_function('take', [data, indices], options)\r\n 269 \r\n 270 \r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.call_function()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/_compute.pyx in pyarrow._compute.Function.call()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: offset overflow while concatenating arrays\r\n```\r\n\r\nIt seems to work fine with small datasets or with pyarrow 0.17.1\nLet me know if you meet other offset overflow issues @joeddav ","embeddings":[-0.3078599274,-0.1624817699,-0.0384140052,0.1917552799,0.1244793832,-0.0056269262,0.2420143485,0.1383444518,-0.3982105851,0.361923933,0.3079869449,0.4601465166,0.0280525908,0.0383535996,0.0241384003,-0.0889734328,-0.0426646657,0.0609558672,-0.1887903959,-0.0936412439,-0.0416765921,-0.3072259426,-0.2502094805,-0.0529374555,-0.1218299344,-0.2309742272,0.1539252698,0.1009820625,-0.1768930852,-0.2024064958,0.3342114687,-0.2671542764,0.0081014773,-0.0058169933,-0.0001175907,0.105093196,0.3507690728,0.0269848593,-0.0879454762,0.36186257,-0.1608607024,-0.3452382684,0.0494743697,-0.4793946147,0.2600331306,-0.1769219935,-0.2077457309,0.0676089823,0.0854020566,0.2236685753,0.1557555348,-0.0401188843,0.1585735977,0.2635318935,0.3263535202,0.0234108511,-0.0072753155,-0.2422338426,-0.167885378,0.0705335438,-0.2242981493,-0.0016125748,-0.0947054923,-0.0284246001,0.0539905503,0.0067401864,-0.023341896,0.0207042638,-0.1426187307,0.0567588471,0.2461706698,-0.2803447843,-0.1456433535,-0.2259111553,-0.0329124369,-0.3547538817,0.1401233077,0.3031680286,-0.0155505668,0.0339879505,-0.2461449653,0.2101866305,-0.2522290647,0.2505633831,-0.1889606565,0.2698267996,0.1512380987,0.1663444042,0.1357590854,-0.1141221076,0.4339353442,-0.1705885679,-0.0047746915,0.1696425527,-0.2869728804,0.1074186414,0.1948883533,-0.1316955835,0.500872314,-0.3496707678,0.0059454655,0.044077985,0.2339591086,0.0869400725,0.2841058373,0.1922837794,-0.5885927081,0.1779753715,-0.0207448937,0.0505632646,0.0146464435,0.0611452945,0.2007251531,-0.1749119461,0.3466876447,-0.080063574,0.0683387741,-0.1202837303,-0.2903328538,0.1888431013,-0.4950132966,0.3146744967,0.2096391916,-0.0979561433,0.0071371985,-0.0668994933,-0.2257014215,0.0162179675,-0.2137298733,0.0277413074,-0.1721847653,-0.0789436623,-0.097892873,-0.1193299741,0.0237431806,0.1679458022,-0.0813873485,0.3148367107,0.3989399672,-0.1049106643,-0.1097599268,-0.1568699181,0.4402474463,0.2657408416,-0.037249025,0.0719270855,-0.1062029973,0.0841025859,-0.0842686966,0.5176148415,-0.2612574995,-0.3617993593,-0.1042305902,0.1721643656,-0.1236087084,0.1671472192,-0.2392052561,0.1035816893,0.2674862444,-0.0605446771,0.186840266,-0.1679096073,-0.0465435795,-0.1870518327,0.2324185967,0.0651653484,-0.5403259993,0.3515728414,0.046457205,0.2777175009,0.3522078097,0.5056605935,-0.3103647828,-0.0813853443,-0.2419678718,0.054232087,0.1743374467,0.0021221105,-0.8825092912,-0.0902406722,-0.1311735809,-0.0398749858,0.2627778053,-0.1125264093,0.2651942372,0.0228159875,-0.0259930156,0.4810886681,-0.2747821212,0.2685475051,-0.4711420536,-0.1202219874,0.0109113958,0.0561424904,0.0007090906,-0.4467764199,-0.0522028096,-0.2801512182,0.2582900226,0.0480177514,0.2383665144,0.2446468771,0.0595670231,-0.2582804263,0.1781593412,-0.2147070318,-0.0670889765,-0.0253214389,-0.0054617655,-0.0934444442,-0.3762587905,-0.2153983414,-0.1107436791,0.3079881072,-0.1378659904,0.2697454691,0.1490946114,-0.3101868927,0.3051522672,0.1536781639,0.144846037,-0.1100105867,-0.1788464487,-0.0944767296,-0.2770965993,0.3967710435,-0.1217573658,-0.3430443704,-0.0394438095,0.3696358204,0.1240580752,-0.1210040003,-0.08161515,0.247591719,0.1220882535,-0.1240401193,-0.1544880718,0.201361239,0.1901596636,-0.6690385938,0.1178058013,-0.0190334246,-0.0259280931,0.073074013,-0.214742884,0.3537632227,-0.3796572685,0.1163655147,-0.0941120759,-0.0308949854,0.1070786119,-0.0249294546,0.1990388334,-0.0057956716,-0.1353929043,0.2408878505,-0.0255299453,-0.05827583,0.085350886,0.2082442045,0.4890706241,-0.1404571235,-0.0235346928,0.0426869988,-0.2933749259,-0.0679957122,0.1969684809,-0.2779559195,-0.0518844575,0.4408052564,0.0995412171,-0.234200567,-0.0678505599,-0.1258074045,0.236642167,0.3833583295,0.0870453045,0.1480712146,0.4960047007,-0.1378153265,-0.3675361574,0.0245736167,0.2212060094,0.4458927512,-0.1369953752,0.3641352952,-0.1575205475,-0.2291455865,-0.1539961249,-0.5479347706,-0.0495434105,-0.5558484793,-0.1171553135,0.2539521158,-0.2222500741,0.3176632822,0.028670188,-0.0788918287,0.4207021296,-0.2142774016,-0.0134648597,-0.2432038486,-0.1864147335,0.0226918533,0.3362286985,0.0649512634,0.0128789227,0.4299613833,0.0976502746,-0.0092566181,-0.3197989464,0.0494458601,0.1751476824,0.2215363234,0.0291967578,-0.2809969187,-0.3562772572,-0.2338242084,-0.0001732999,0.1419413835,-0.1411358565,0.4679219723,-0.1558834314,0.0448507592,0.008400552,-0.0066716825,-0.0028689625,-0.3170768321,-0.106845133,-0.0941461399,0.146625787,0.1291240901,0.2173149884,0.0217559095,0.1971845925,-0.3625352681,-0.2118215263,0.0719039589,0.3074546754,-0.0073131048,-0.1995082647,-0.1844004095,-0.4688217044,-0.120662719,0.3121152222,-0.43527022,0.1149598956,-0.165476054,0.5477545261,-0.4002652466,0.0760845542,0.1913095415,0.1314117759,-0.0754614696,-0.1794867814,-0.0041376762,0.0360306315,-0.1388119757,-0.0993702859,-0.0103422347,0.35558483,0.1919984221,0.6874073744,0.182875067,-0.0826884285,0.1499190331,-0.103769809,0.0607788451,-0.0972274169,-0.2453378737,0.1346120238,0.2759500742,-0.2638726234,0.0069096205,-0.0608507358,-0.3277964294,0.3485786915,-0.1818958968,-0.1257545352,-0.3210374415,0.2611432076,-0.1045739502,0.0411942899,0.1294791102,-0.1142401099,-0.1942567676,-0.2883715928,0.0373693146,-0.1895458251,0.2498661429,-0.1295817643,0.0257857721,-0.165540725,-0.108615227,0.0432148464,0.2535862625,0.1290031672,0.2761618495,-0.0774622336,0.214434281,-0.087756373,1.052972436,0.1035215929,0.0862727538,-0.0843415782,0.3270575702,-0.5469658375,0.045502834,-0.3395446539,0.4172218144,0.1493502408,0.0795401409,-0.2137515992,0.1312854886,0.6030226946,0.2673247755,-0.0757204667,-0.1873874664,-0.1004454195,-0.1217115819,0.2724308074,0.3786853552,0.0342793576,0.3319784701,0.0867872387,-0.2884069383,-0.4074480236,-0.1794433147,-0.43626073,0.0392235778,0.3683811128,-0.2323276997,0.0919802189,-0.2026369274,-0.1233670264,0.5874598622,0.276170522,0.1106542721,-0.1481909901,-0.114576228,0.1147041842,0.296895653,-0.1863588691,0.0224657096,-0.00429929,-0.0666999668,0.2870538831,-0.0751791298,-0.2727073133,-0.0565402173,0.0645271242,-0.0863803253,0.0467156321,0.535130024,-0.0962854698,-0.0200151559,0.2606781125,0.4524257481,-0.2280533165,0.2177149504,0.2168310881,0.6593003273,-0.0582131296,-0.0079521947,0.3936114609,0.0186922234,0.3969335556,0.1132596657,0.2290652841,-0.2633156776,-0.2673666477,0.1683330089,-0.2955305874,-0.2516088188,0.0523138903,-0.1367732882,0.0307242963,-0.1729731709,0.0640651882,-0.1427879781,-0.0798084587,0.1897869259,-0.2027525604,-0.3587452769,0.0043281298,0.0839706734,-0.4553226531,-0.0062032384,-0.0558736995,-0.0727324113,-0.0626004785,-0.2099886984,-0.0158432331,-0.3024449944,0.2118612826,0.1587907523,-0.6693410277,0.0959851071,-0.1027466953,-0.0909658521,-0.003784352,0.0125186779,-0.0979868025,0.0787511766,-0.1514410079,-0.093180418,0.1588275582,0.3725361526,0.1117558777,-0.3891111612,0.0792206377,0.1022791415,-0.060583692,0.0916546956,0.1698605716,0.1858794838,-0.3513030708,0.0801825821,-0.2744289935,-0.0276282039,-0.2533023357,0.0510548986,-0.0321576819,-0.3189631402,0.3559044302,0.1106919721,-0.1777253002,-0.0625040457,0.3362928033,0.1386843175,0.3069220185,0.4120994806,0.2120607942,-0.2112435251,-0.0923749208,0.2197960466,0.2482061535,-0.2466107905,-0.1335739493,-0.3357232809,0.2979923785,0.0040589105,-0.1968062818,0.0943774879,0.2139061689,-0.4765343964,-0.2514716685,-0.2843219638,0.211540699,0.0218784977,-0.1386735141,0.1883488894,-0.1097696796,-0.1664049178,0.0380355753,-0.2583703101,0.2426104844,0.015891118,0.2187202126,-0.2371936738,-0.1162924543,0.0712876394,-0.1096849814,0.0967871845,0.0323481523,-0.0849000812,-0.2896191776,-0.0747356862,0.068775028,0.1048449799,-0.2686823308,-0.1610892564,-0.0881018713,-0.197155416,-0.0708957389,0.199687615,-0.040991839,-0.0030111203,0.0278002545,0.6172444224,-0.2400876284,-0.4222017229,-0.0536632203,-0.0288906079,0.1443918198,0.1050772145,0.508179903,0.0785977244,-0.0513165146,-0.138573572,0.0988981724,0.0796134993,-0.0999370441,0.4434025586,-0.3727136254,-0.2157142907,0.248406738,0.3717080653,0.2496133149,-0.0126267867,-0.3776681721,0.1215229258,0.2283975482,-0.032063961,-0.2781687975,0.1422376782,0.1158116311,0.1839352548,0.3635499477,-0.045644626,-0.3120152354,0.0750242844,0.0862318501,0.5116409659,-0.0492971055,-0.2612501085,0.5733863711,-0.1759341806,-0.216320619,0.3265039623,-0.0425852649,0.5641289949,0.6425901055,-0.0141267804,0.2967489958,0.2060596198,-0.3082503974,-0.0604013316,-0.126505807,-0.2884389758,0.256336391,-0.0069088591,0.1854427606,0.1793788224,0.210778147,0.4293079674,0.0859497711,0.0137428362,-0.0070639001,-0.2963067591,0.0951610506,-0.3648280501,-0.0339967124,0.1075608581,-0.0892927498,-0.0629893318,0.1336776316,0.5363907814,0.073693946,0.0753084496,-0.1682991832,0.472537905,0.3535591662,0.2539380789,-0.1741780341,0.3685833514,0.3450315893,0.0600092635,-0.3759499192,0.2836781442,0.4476296008,0.0661306903,-0.2917457521,-0.1754899472,0.1431080997,-0.1993087679,-0.3571977913,0.0562635325,-0.02232638,0.269728303,0.1601201296,0.1058418378,-0.1833210886,0.0344817415,0.5079106092,-0.0918377116,-0.1485922933,-0.3756018281,-0.0136666046,-0.4743428826,0.0703353211,-0.0579412803,0.0409056284,0.0677907988,0.1807413399,-0.0876699984,0.0531846061,0.2307295203,0.035296768,-0.3808389604,0.7067388892,0.5049709082,0.2196255922,-0.2470973134,0.0792752951,0.0175036658,0.4386537969,-0.2930751443,-0.0511545241,0.5098760724,0.3039680719,-0.1780640036,0.2502115369,0.1689564884,0.1316731125,-0.19541502,0.4371808171,-0.5424678922,-0.2461664677,0.0678663552,-0.4354547262,-0.1111012176,-0.4641906321,0.2038557976,-0.0356637537,0.0513345823,-0.0334171131,0.237526536,-0.1416114271,0.1121383458,0.2387252599,0.4911783934,0.1525250673,0.0015406716,0.1109365895,-0.1395282894,-0.1394063681,-0.5160037279,0.5627593994,-0.1268523186,0.2740676701,-0.1731084436,-0.7187899351,0.1381969154,0.1420859098,0.1563244611,0.0118768979,0.0317550935,0.1862273067,-0.2070299238,0.0572317094,-0.0017846138,0.0907566845,0.1579876691,0.2495679855,-0.0379342251,-0.2104880363,0.1006878912,-0.2144787163,-0.1678662449,-0.0442154594,0.1772096902,0.2280100584,-0.3719458282,-0.254488498,0.0146524208,0.4591278434,-0.1570975333,-0.1533426195,0.1323523521,0.1654485613,0.0945475921,0.0275233816,0.2726256847,0.0628592223,-0.3466641903,-0.0776627585,-0.2243146002]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/611","title":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648","comments":"Can you give us stats\/information on your pandas DataFrame?","body":"Hi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```","comment_length":9,"text":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\nHi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```\nCan you give us stats\/information on your pandas DataFrame?","embeddings":[-0.2813221216,-0.0744086578,-0.2195497006,0.4159464538,0.2823321819,-0.0299135447,0.4081522524,0.2113159895,0.3901511431,0.047444731,0.0003714884,0.3082345426,-0.0869823992,0.0685475916,0.000835677,-0.3419586718,-0.0933910608,0.2279389799,-0.275508076,0.0938671231,-0.2262557447,0.0042208233,-0.2458188832,0.2069006562,-0.099163264,-0.1639172286,0.0946853608,-0.0185281597,-0.0302411988,-0.4899121225,0.3556722403,-0.2945443392,0.0409597605,0.1827444583,-0.0001102196,0.0405716449,0.3709658682,0.0665661767,-0.2005666047,-0.1553708464,-0.3287505507,-0.4413967729,0.4300814569,-0.0768523142,0.2361876369,-0.2503988445,-0.2843058705,-0.3228805363,0.3386787176,0.0333642215,0.2106780857,0.2027079761,0.4377660751,0.121098958,0.0229463894,-0.0303455219,-0.1591518968,0.4214888215,0.1034795642,-0.2736412883,-0.0080701439,0.0098081287,-0.142464757,0.3750428259,0.148845315,-0.0841920599,-0.1104783043,-0.3776778579,-0.2088728249,0.2992431521,0.6146755219,-0.2124876976,-0.3339965045,-0.0728103667,-0.0512901135,-0.3944200575,0.0467973463,0.2284300476,-0.0249728207,0.055753652,0.0425886177,0.0693220422,-0.2296036631,0.0915905237,-0.0219255704,0.1695368588,-0.1137782633,0.2228228748,0.3527334929,-0.1476604491,0.1830546558,0.1812524498,0.0453062244,0.1507354528,-0.4711746573,-0.0076138237,-0.2046742588,-0.4235938787,0.0187764075,0.3424940407,0.2327027917,0.0266376585,0.0264474768,0.2103405595,0.1510000825,0.0093807904,-0.1516859233,0.0805711374,-0.2187978178,-0.0026093989,0.1261629015,-0.0250724573,-0.0848933235,-0.1750027239,0.1665083319,-0.1399342269,0.3136582673,-0.1227992103,-0.4663731754,0.2898985445,-0.494212687,-0.1571940184,0.106676884,0.1941330135,0.1898524016,0.3169753551,-0.0033455505,0.1713694483,-0.0771825686,0.1586159617,-0.1784182787,0.2248283327,0.0160623174,0.1040193215,-0.0073826094,0.0587322824,0.1619348079,0.1394265741,0.3225094676,-0.1406376809,0.2424962372,-0.4746845663,-0.0491963215,0.4604407549,-0.0665069968,0.137398988,0.4816637933,0.0661066324,-0.0525114462,0.287420392,-0.0509188175,-0.1669092923,-0.307931006,0.2034970522,0.0864146128,0.1019393206,-0.5072076321,-0.1107438877,0.2790833116,0.0325417966,0.2052158564,-0.1718201488,-0.0322390348,-0.237004444,-0.0113939112,0.2493876666,-0.6265670657,0.1540756524,-0.0097450269,-0.004250614,0.2222199887,0.3802011907,-0.371242255,0.0840109363,-0.1422454864,0.1392559856,-0.0864757076,0.0490292422,-0.4882147908,0.1294535846,0.0305430219,-0.0098824194,0.136037454,0.1243114844,0.0164619777,0.0188371874,-0.0198179577,0.2659706473,0.1596460044,-0.1190212071,-0.0785414279,-0.0919746459,0.1880332828,0.0425336324,-0.1309635341,-0.1111637652,-0.037798252,-0.058344353,0.0506681018,-0.0628661066,-0.1312288791,0.3077427745,0.102279827,0.0275660791,-0.1993941218,-0.1144554392,-0.2583920658,-0.0028342633,0.0060153445,-0.2196477205,-0.3686653972,0.0322350599,-0.3015362918,0.4142707586,-0.003701322,0.1467631906,0.1555080861,0.0351495408,-0.1392640173,-0.3405443728,-0.1125302464,0.3829867542,-0.2520798147,0.117400378,-0.2488704324,0.6389963627,-0.1225095615,-0.2358955741,0.102029413,-0.0589136705,0.0078237727,-0.0617898889,-0.0541578941,0.3003418446,0.2589096129,-0.0539924949,-0.1232170388,-0.1712924838,0.1711663902,-0.4476631284,-0.0194845684,0.3259800076,0.2496611625,0.0040201214,0.2005787194,0.2582695484,-0.230536446,0.1718457788,0.0606855899,0.0135376034,0.2570950389,0.2372244,-0.0010136174,-0.1002517715,-0.0585765764,0.1710237116,0.2922351956,0.0836870968,-0.3078921139,-0.0896601155,0.0380948409,0.0779544115,0.1145797297,0.1024587899,-0.1011391133,-0.1450705826,0.2871288657,-0.15945144,0.36046803,0.2623209059,-0.3645984828,-0.3160845041,-0.1206738278,0.0692813694,0.3395498991,0.2408055216,0.4563713074,-0.0912684724,0.0961696357,-0.0320504308,-0.2791270316,-0.5238831639,0.0624611005,0.4196495712,-0.3285613656,0.2295600623,-0.1541727185,-0.1055464894,0.2673347592,-0.6564484835,0.0172794033,-0.1459008306,-0.3326733708,0.3113354743,0.0069784429,0.098078616,-0.3459626734,-0.1520460844,0.3765098751,0.0735324323,0.0101614492,-0.0400554463,0.1425979286,0.1107221246,0.2702673376,0.0741861239,0.2685500979,0.3350822628,0.0054364279,-0.1947278678,-0.0174658187,-0.0798835009,-0.1669342369,0.0517802648,0.0623224191,0.3742077351,-0.5855019093,-0.5999208093,0.1264113486,0.0736583471,-0.1970529705,0.1744908839,0.0701362565,0.1771164238,0.1904895902,-0.1944365352,-0.314404428,-0.3713174164,0.1869551688,0.223639369,0.3277160227,0.3159987032,0.2656185925,0.1239757687,0.2631948292,-0.1057830453,-0.1982412785,0.1272568703,0.1754867584,0.0087331478,-0.2702606916,-0.1913998574,-0.267470181,0.2551690936,0.0270455424,-0.3893484771,-0.2998656631,-0.361228317,-0.08038757,-0.3379433453,-0.0242435355,0.270458132,0.1402743757,-0.2321830541,0.0587796047,-0.1001932099,-0.0318213888,-0.1266248375,0.1443426162,-0.1426805705,0.5946801901,-0.1982656568,0.5423558354,0.1506075263,-0.0378133692,0.4440318942,-0.0917442888,0.1699853688,-0.3620955348,-0.4971083701,-0.0669029206,-0.1251891553,0.066418387,-0.1115148067,-0.2821611166,-0.1304080635,0.1161029786,-0.0095336782,-0.1359714419,0.081121996,0.1733745337,-0.0260819588,0.1395895034,-0.1173078716,-0.0486916602,-0.1003917232,-0.0609221458,-0.2855698466,-0.2101400048,0.2482716441,-0.1466301233,-0.2170103639,-0.2675659955,-0.0345852487,0.3139541745,0.3056481481,0.5451367497,0.1436631083,-0.0632694811,0.0515470766,0.0490171835,0.3031839728,0.1406436861,-0.2015923113,0.2815236449,-0.1118085608,-0.5275207758,0.0068123387,-0.2888289392,0.2861074507,-0.1511148065,-0.1364953071,-0.2830283046,-0.0095327608,0.1217485815,0.3055149317,-0.0608977675,-0.1197311059,-0.3016712368,-0.0870330483,-0.5179740787,0.1697142571,0.1463099718,0.2763128877,0.0918487087,-0.2823677063,-0.2318428606,-0.5319346786,0.0363810956,0.0667778403,0.3335453272,0.1646634936,0.1819217652,-0.2661644816,-0.117607601,0.5514033437,0.3901548088,0.2355774045,-0.0744238794,-0.0169190038,0.0978515372,0.225010246,0.0370268635,-0.0905061439,0.1164837182,-0.2906730771,0.1484722197,0.1189413071,-0.0457758009,0.046827618,-0.0387605838,-0.2239279747,-0.2735214233,0.4360409081,0.0355576724,0.1602492183,0.4475951791,0.2731461227,-0.014615803,0.1969361454,0.2358176112,0.8210725784,-0.3222990036,0.0012435849,0.3847110868,0.0425249226,0.4504195452,0.3941969275,0.1207178086,-0.3513602614,-0.2400814742,-0.109383814,-0.2209489644,0.1930115074,0.1323995143,-0.2190535814,0.1336609721,-0.1972189993,0.1078306064,0.0800062492,0.0967849493,-0.0671035945,-0.3313188255,-0.5120387077,0.1155115589,-0.2454337031,-0.1842787713,-0.154357776,-0.0854888558,-0.0121683301,-0.068457216,-0.1600129902,0.1471405327,-0.2108902186,0.5190977454,0.2066063881,-0.5216561556,0.0591063499,-0.081766732,-0.2114801556,-0.0654445961,0.0227802098,-0.1752675772,-0.1868136674,-0.2421340793,-0.148311168,0.0007692617,0.276091665,0.0007844663,-0.1791667789,0.1693535894,-0.0637616292,-0.4431357384,0.2583422661,0.3664193451,0.2560353577,-0.3944438398,-0.3703472018,-0.1771428436,-0.3999643624,-0.2454397231,0.1234320551,0.0697477609,-0.0844346657,0.1232956648,0.223929584,-0.1587904543,-0.0421671905,0.2827085257,0.2458965927,0.1164725795,0.6250904202,0.1420297027,-0.2136124223,-0.2528681457,0.0410644673,0.2418822646,-0.4038007855,0.2405816764,0.0724531636,0.0953915119,0.028948307,0.0958370343,0.2089332789,0.4507569671,-0.1446702182,-0.5494536161,-0.5519043803,0.3133969903,-0.3650940955,0.1240868196,0.0728808716,0.0119035337,-0.0238186233,0.0976890922,-0.2888030708,-0.0474176705,-0.3507489264,0.059047889,-0.0160403736,-0.0977914259,-0.012049037,-0.0036973096,0.2137143612,0.2369950712,-0.1554045677,-0.3096164763,0.1651388556,0.0450443551,0.1219118088,-0.4085612595,-0.1422498524,-0.1819286346,-0.0714023635,-0.130444333,-0.1986453831,0.0088194683,-0.0786494166,-0.137713328,0.2962082326,-0.0935740992,-0.2675312161,0.0401337668,-0.2507911325,0.2575215995,0.0955288559,0.0755477995,0.3524129093,0.025937954,-0.2118298709,0.0399078913,-0.0894020721,-0.1462773085,0.3332202137,-0.3912499547,-0.3031314015,0.0155778779,0.4696730077,0.2980320156,-0.1497734338,0.0706827417,-0.0487037972,0.2152475268,-0.3896286488,-0.166516915,-0.0059126685,0.0592268333,-0.0484120399,0.3196217418,0.2594074011,-0.2182893902,0.2811603844,0.2165706158,0.2245479077,0.0140968384,-0.1251775175,0.6010939479,-0.1161141619,-0.2117932141,0.4168370068,0.2926371396,0.5112424493,0.26060763,-0.1071603373,0.1381422877,0.2688388228,0.1799882352,-0.0583727658,-0.4001861513,-0.1858804226,0.1077909619,0.1066513956,-0.138240248,0.1256545335,-0.028011376,0.2579557598,0.0361022688,0.1478007436,0.0859167874,-0.1818326712,0.0219459999,0.000807327,0.1249998361,0.0484733023,-0.2068495601,0.0252287332,-0.3063671291,0.2777527571,0.0016674157,-0.0858000293,-0.4399663508,0.1806315184,0.5117088556,-0.1094691604,-0.1741171777,0.328587383,0.4242916107,-0.0237483382,-0.1503320634,0.3860247135,0.5804914832,0.2409934551,-0.0238724966,0.2077873796,-0.0081494143,-0.1690758169,-0.0604744181,0.0909238458,0.0978051424,0.1481955349,0.1423253566,0.2105998993,-0.1486494541,0.2147817761,0.3060647845,0.106562227,-0.2027076781,0.1401877254,-0.116266422,-0.2585432827,-0.0850696713,-0.0174536537,-0.4101569951,0.1649530232,0.4182043672,-0.120069094,-0.016555516,0.1644590348,0.0681711882,-0.3404383957,0.4425357878,0.2781794965,0.2539971471,-0.1694387048,0.0014491276,-0.4343959689,0.2145670801,-0.3108062744,-0.1265728623,0.1883103102,0.3053250313,0.2690074742,0.3885180056,0.1593928039,0.1569662392,0.0389915444,0.1325366944,-0.3590078056,-0.1170711219,0.1671163887,0.115293242,-0.171261549,-0.1662117839,0.3117127419,0.0199156646,0.10100732,-0.2547494173,-0.0726020187,-0.2228352875,0.3414654732,0.3372592926,0.1324576437,0.3376508355,0.0862659737,0.2375318557,-0.0411623679,-0.4088019729,-0.0082033034,0.5613906384,0.3403669596,0.5017957091,-0.1312077641,-0.1257840246,-0.143103987,0.3379398584,-0.2317003608,-0.042705778,-0.3095024228,0.0799118504,-0.1820611507,-0.0606567971,-0.3343918025,-0.0018313729,0.0890034661,0.2038212121,-0.2043199539,-0.5852287412,0.4368210137,-0.1386185437,-0.4015676379,0.0768632218,0.0833692923,0.0400837772,-0.0228591729,-0.2593250275,0.1473932862,0.2686134577,-0.0831846893,-0.3365375698,0.0610045604,-0.028715238,0.2706465721,-0.1467019916,0.3606089652,-0.1064436436,-0.0384901576,-0.1178414598,-0.1704109609]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/611","title":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648","comments":"```\r\n<class 'pandas.core.frame.DataFrame'>\r\nInt64Index: 17136104 entries, 0 to 17136103\r\nData columns (total 6 columns):\r\n # Column Dtype \r\n--- ------ ----- \r\n 0 item_id int64 \r\n 1 item_titl object \r\n 2 start_price float64\r\n 3 shipping_fee float64\r\n 4 picture_url object \r\n 5 embeddings object \r\ndtypes: float64(2), int64(1), object(3)\r\nmemory usage: 915.2+ MB\r\n```","body":"Hi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```","comment_length":47,"text":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\nHi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```\n```\r\n<class 'pandas.core.frame.DataFrame'>\r\nInt64Index: 17136104 entries, 0 to 17136103\r\nData columns (total 6 columns):\r\n # Column Dtype \r\n--- ------ ----- \r\n 0 item_id int64 \r\n 1 item_titl object \r\n 2 start_price float64\r\n 3 shipping_fee float64\r\n 4 picture_url object \r\n 5 embeddings object \r\ndtypes: float64(2), int64(1), object(3)\r\nmemory usage: 915.2+ MB\r\n```","embeddings":[-0.2813221216,-0.0744086578,-0.2195497006,0.4159464538,0.2823321819,-0.0299135447,0.4081522524,0.2113159895,0.3901511431,0.047444731,0.0003714884,0.3082345426,-0.0869823992,0.0685475916,0.000835677,-0.3419586718,-0.0933910608,0.2279389799,-0.275508076,0.0938671231,-0.2262557447,0.0042208233,-0.2458188832,0.2069006562,-0.099163264,-0.1639172286,0.0946853608,-0.0185281597,-0.0302411988,-0.4899121225,0.3556722403,-0.2945443392,0.0409597605,0.1827444583,-0.0001102196,0.0405716449,0.3709658682,0.0665661767,-0.2005666047,-0.1553708464,-0.3287505507,-0.4413967729,0.4300814569,-0.0768523142,0.2361876369,-0.2503988445,-0.2843058705,-0.3228805363,0.3386787176,0.0333642215,0.2106780857,0.2027079761,0.4377660751,0.121098958,0.0229463894,-0.0303455219,-0.1591518968,0.4214888215,0.1034795642,-0.2736412883,-0.0080701439,0.0098081287,-0.142464757,0.3750428259,0.148845315,-0.0841920599,-0.1104783043,-0.3776778579,-0.2088728249,0.2992431521,0.6146755219,-0.2124876976,-0.3339965045,-0.0728103667,-0.0512901135,-0.3944200575,0.0467973463,0.2284300476,-0.0249728207,0.055753652,0.0425886177,0.0693220422,-0.2296036631,0.0915905237,-0.0219255704,0.1695368588,-0.1137782633,0.2228228748,0.3527334929,-0.1476604491,0.1830546558,0.1812524498,0.0453062244,0.1507354528,-0.4711746573,-0.0076138237,-0.2046742588,-0.4235938787,0.0187764075,0.3424940407,0.2327027917,0.0266376585,0.0264474768,0.2103405595,0.1510000825,0.0093807904,-0.1516859233,0.0805711374,-0.2187978178,-0.0026093989,0.1261629015,-0.0250724573,-0.0848933235,-0.1750027239,0.1665083319,-0.1399342269,0.3136582673,-0.1227992103,-0.4663731754,0.2898985445,-0.494212687,-0.1571940184,0.106676884,0.1941330135,0.1898524016,0.3169753551,-0.0033455505,0.1713694483,-0.0771825686,0.1586159617,-0.1784182787,0.2248283327,0.0160623174,0.1040193215,-0.0073826094,0.0587322824,0.1619348079,0.1394265741,0.3225094676,-0.1406376809,0.2424962372,-0.4746845663,-0.0491963215,0.4604407549,-0.0665069968,0.137398988,0.4816637933,0.0661066324,-0.0525114462,0.287420392,-0.0509188175,-0.1669092923,-0.307931006,0.2034970522,0.0864146128,0.1019393206,-0.5072076321,-0.1107438877,0.2790833116,0.0325417966,0.2052158564,-0.1718201488,-0.0322390348,-0.237004444,-0.0113939112,0.2493876666,-0.6265670657,0.1540756524,-0.0097450269,-0.004250614,0.2222199887,0.3802011907,-0.371242255,0.0840109363,-0.1422454864,0.1392559856,-0.0864757076,0.0490292422,-0.4882147908,0.1294535846,0.0305430219,-0.0098824194,0.136037454,0.1243114844,0.0164619777,0.0188371874,-0.0198179577,0.2659706473,0.1596460044,-0.1190212071,-0.0785414279,-0.0919746459,0.1880332828,0.0425336324,-0.1309635341,-0.1111637652,-0.037798252,-0.058344353,0.0506681018,-0.0628661066,-0.1312288791,0.3077427745,0.102279827,0.0275660791,-0.1993941218,-0.1144554392,-0.2583920658,-0.0028342633,0.0060153445,-0.2196477205,-0.3686653972,0.0322350599,-0.3015362918,0.4142707586,-0.003701322,0.1467631906,0.1555080861,0.0351495408,-0.1392640173,-0.3405443728,-0.1125302464,0.3829867542,-0.2520798147,0.117400378,-0.2488704324,0.6389963627,-0.1225095615,-0.2358955741,0.102029413,-0.0589136705,0.0078237727,-0.0617898889,-0.0541578941,0.3003418446,0.2589096129,-0.0539924949,-0.1232170388,-0.1712924838,0.1711663902,-0.4476631284,-0.0194845684,0.3259800076,0.2496611625,0.0040201214,0.2005787194,0.2582695484,-0.230536446,0.1718457788,0.0606855899,0.0135376034,0.2570950389,0.2372244,-0.0010136174,-0.1002517715,-0.0585765764,0.1710237116,0.2922351956,0.0836870968,-0.3078921139,-0.0896601155,0.0380948409,0.0779544115,0.1145797297,0.1024587899,-0.1011391133,-0.1450705826,0.2871288657,-0.15945144,0.36046803,0.2623209059,-0.3645984828,-0.3160845041,-0.1206738278,0.0692813694,0.3395498991,0.2408055216,0.4563713074,-0.0912684724,0.0961696357,-0.0320504308,-0.2791270316,-0.5238831639,0.0624611005,0.4196495712,-0.3285613656,0.2295600623,-0.1541727185,-0.1055464894,0.2673347592,-0.6564484835,0.0172794033,-0.1459008306,-0.3326733708,0.3113354743,0.0069784429,0.098078616,-0.3459626734,-0.1520460844,0.3765098751,0.0735324323,0.0101614492,-0.0400554463,0.1425979286,0.1107221246,0.2702673376,0.0741861239,0.2685500979,0.3350822628,0.0054364279,-0.1947278678,-0.0174658187,-0.0798835009,-0.1669342369,0.0517802648,0.0623224191,0.3742077351,-0.5855019093,-0.5999208093,0.1264113486,0.0736583471,-0.1970529705,0.1744908839,0.0701362565,0.1771164238,0.1904895902,-0.1944365352,-0.314404428,-0.3713174164,0.1869551688,0.223639369,0.3277160227,0.3159987032,0.2656185925,0.1239757687,0.2631948292,-0.1057830453,-0.1982412785,0.1272568703,0.1754867584,0.0087331478,-0.2702606916,-0.1913998574,-0.267470181,0.2551690936,0.0270455424,-0.3893484771,-0.2998656631,-0.361228317,-0.08038757,-0.3379433453,-0.0242435355,0.270458132,0.1402743757,-0.2321830541,0.0587796047,-0.1001932099,-0.0318213888,-0.1266248375,0.1443426162,-0.1426805705,0.5946801901,-0.1982656568,0.5423558354,0.1506075263,-0.0378133692,0.4440318942,-0.0917442888,0.1699853688,-0.3620955348,-0.4971083701,-0.0669029206,-0.1251891553,0.066418387,-0.1115148067,-0.2821611166,-0.1304080635,0.1161029786,-0.0095336782,-0.1359714419,0.081121996,0.1733745337,-0.0260819588,0.1395895034,-0.1173078716,-0.0486916602,-0.1003917232,-0.0609221458,-0.2855698466,-0.2101400048,0.2482716441,-0.1466301233,-0.2170103639,-0.2675659955,-0.0345852487,0.3139541745,0.3056481481,0.5451367497,0.1436631083,-0.0632694811,0.0515470766,0.0490171835,0.3031839728,0.1406436861,-0.2015923113,0.2815236449,-0.1118085608,-0.5275207758,0.0068123387,-0.2888289392,0.2861074507,-0.1511148065,-0.1364953071,-0.2830283046,-0.0095327608,0.1217485815,0.3055149317,-0.0608977675,-0.1197311059,-0.3016712368,-0.0870330483,-0.5179740787,0.1697142571,0.1463099718,0.2763128877,0.0918487087,-0.2823677063,-0.2318428606,-0.5319346786,0.0363810956,0.0667778403,0.3335453272,0.1646634936,0.1819217652,-0.2661644816,-0.117607601,0.5514033437,0.3901548088,0.2355774045,-0.0744238794,-0.0169190038,0.0978515372,0.225010246,0.0370268635,-0.0905061439,0.1164837182,-0.2906730771,0.1484722197,0.1189413071,-0.0457758009,0.046827618,-0.0387605838,-0.2239279747,-0.2735214233,0.4360409081,0.0355576724,0.1602492183,0.4475951791,0.2731461227,-0.014615803,0.1969361454,0.2358176112,0.8210725784,-0.3222990036,0.0012435849,0.3847110868,0.0425249226,0.4504195452,0.3941969275,0.1207178086,-0.3513602614,-0.2400814742,-0.109383814,-0.2209489644,0.1930115074,0.1323995143,-0.2190535814,0.1336609721,-0.1972189993,0.1078306064,0.0800062492,0.0967849493,-0.0671035945,-0.3313188255,-0.5120387077,0.1155115589,-0.2454337031,-0.1842787713,-0.154357776,-0.0854888558,-0.0121683301,-0.068457216,-0.1600129902,0.1471405327,-0.2108902186,0.5190977454,0.2066063881,-0.5216561556,0.0591063499,-0.081766732,-0.2114801556,-0.0654445961,0.0227802098,-0.1752675772,-0.1868136674,-0.2421340793,-0.148311168,0.0007692617,0.276091665,0.0007844663,-0.1791667789,0.1693535894,-0.0637616292,-0.4431357384,0.2583422661,0.3664193451,0.2560353577,-0.3944438398,-0.3703472018,-0.1771428436,-0.3999643624,-0.2454397231,0.1234320551,0.0697477609,-0.0844346657,0.1232956648,0.223929584,-0.1587904543,-0.0421671905,0.2827085257,0.2458965927,0.1164725795,0.6250904202,0.1420297027,-0.2136124223,-0.2528681457,0.0410644673,0.2418822646,-0.4038007855,0.2405816764,0.0724531636,0.0953915119,0.028948307,0.0958370343,0.2089332789,0.4507569671,-0.1446702182,-0.5494536161,-0.5519043803,0.3133969903,-0.3650940955,0.1240868196,0.0728808716,0.0119035337,-0.0238186233,0.0976890922,-0.2888030708,-0.0474176705,-0.3507489264,0.059047889,-0.0160403736,-0.0977914259,-0.012049037,-0.0036973096,0.2137143612,0.2369950712,-0.1554045677,-0.3096164763,0.1651388556,0.0450443551,0.1219118088,-0.4085612595,-0.1422498524,-0.1819286346,-0.0714023635,-0.130444333,-0.1986453831,0.0088194683,-0.0786494166,-0.137713328,0.2962082326,-0.0935740992,-0.2675312161,0.0401337668,-0.2507911325,0.2575215995,0.0955288559,0.0755477995,0.3524129093,0.025937954,-0.2118298709,0.0399078913,-0.0894020721,-0.1462773085,0.3332202137,-0.3912499547,-0.3031314015,0.0155778779,0.4696730077,0.2980320156,-0.1497734338,0.0706827417,-0.0487037972,0.2152475268,-0.3896286488,-0.166516915,-0.0059126685,0.0592268333,-0.0484120399,0.3196217418,0.2594074011,-0.2182893902,0.2811603844,0.2165706158,0.2245479077,0.0140968384,-0.1251775175,0.6010939479,-0.1161141619,-0.2117932141,0.4168370068,0.2926371396,0.5112424493,0.26060763,-0.1071603373,0.1381422877,0.2688388228,0.1799882352,-0.0583727658,-0.4001861513,-0.1858804226,0.1077909619,0.1066513956,-0.138240248,0.1256545335,-0.028011376,0.2579557598,0.0361022688,0.1478007436,0.0859167874,-0.1818326712,0.0219459999,0.000807327,0.1249998361,0.0484733023,-0.2068495601,0.0252287332,-0.3063671291,0.2777527571,0.0016674157,-0.0858000293,-0.4399663508,0.1806315184,0.5117088556,-0.1094691604,-0.1741171777,0.328587383,0.4242916107,-0.0237483382,-0.1503320634,0.3860247135,0.5804914832,0.2409934551,-0.0238724966,0.2077873796,-0.0081494143,-0.1690758169,-0.0604744181,0.0909238458,0.0978051424,0.1481955349,0.1423253566,0.2105998993,-0.1486494541,0.2147817761,0.3060647845,0.106562227,-0.2027076781,0.1401877254,-0.116266422,-0.2585432827,-0.0850696713,-0.0174536537,-0.4101569951,0.1649530232,0.4182043672,-0.120069094,-0.016555516,0.1644590348,0.0681711882,-0.3404383957,0.4425357878,0.2781794965,0.2539971471,-0.1694387048,0.0014491276,-0.4343959689,0.2145670801,-0.3108062744,-0.1265728623,0.1883103102,0.3053250313,0.2690074742,0.3885180056,0.1593928039,0.1569662392,0.0389915444,0.1325366944,-0.3590078056,-0.1170711219,0.1671163887,0.115293242,-0.171261549,-0.1662117839,0.3117127419,0.0199156646,0.10100732,-0.2547494173,-0.0726020187,-0.2228352875,0.3414654732,0.3372592926,0.1324576437,0.3376508355,0.0862659737,0.2375318557,-0.0411623679,-0.4088019729,-0.0082033034,0.5613906384,0.3403669596,0.5017957091,-0.1312077641,-0.1257840246,-0.143103987,0.3379398584,-0.2317003608,-0.042705778,-0.3095024228,0.0799118504,-0.1820611507,-0.0606567971,-0.3343918025,-0.0018313729,0.0890034661,0.2038212121,-0.2043199539,-0.5852287412,0.4368210137,-0.1386185437,-0.4015676379,0.0768632218,0.0833692923,0.0400837772,-0.0228591729,-0.2593250275,0.1473932862,0.2686134577,-0.0831846893,-0.3365375698,0.0610045604,-0.028715238,0.2706465721,-0.1467019916,0.3606089652,-0.1064436436,-0.0384901576,-0.1178414598,-0.1704109609]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/611","title":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648","comments":"Thanks and some more on the `embeddings` and `picture_url` would be nice as well (type and max lengths of the elements)","body":"Hi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```","comment_length":21,"text":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\nHi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```\nThanks and some more on the `embeddings` and `picture_url` would be nice as well (type and max lengths of the elements)","embeddings":[-0.2813221216,-0.0744086578,-0.2195497006,0.4159464538,0.2823321819,-0.0299135447,0.4081522524,0.2113159895,0.3901511431,0.047444731,0.0003714884,0.3082345426,-0.0869823992,0.0685475916,0.000835677,-0.3419586718,-0.0933910608,0.2279389799,-0.275508076,0.0938671231,-0.2262557447,0.0042208233,-0.2458188832,0.2069006562,-0.099163264,-0.1639172286,0.0946853608,-0.0185281597,-0.0302411988,-0.4899121225,0.3556722403,-0.2945443392,0.0409597605,0.1827444583,-0.0001102196,0.0405716449,0.3709658682,0.0665661767,-0.2005666047,-0.1553708464,-0.3287505507,-0.4413967729,0.4300814569,-0.0768523142,0.2361876369,-0.2503988445,-0.2843058705,-0.3228805363,0.3386787176,0.0333642215,0.2106780857,0.2027079761,0.4377660751,0.121098958,0.0229463894,-0.0303455219,-0.1591518968,0.4214888215,0.1034795642,-0.2736412883,-0.0080701439,0.0098081287,-0.142464757,0.3750428259,0.148845315,-0.0841920599,-0.1104783043,-0.3776778579,-0.2088728249,0.2992431521,0.6146755219,-0.2124876976,-0.3339965045,-0.0728103667,-0.0512901135,-0.3944200575,0.0467973463,0.2284300476,-0.0249728207,0.055753652,0.0425886177,0.0693220422,-0.2296036631,0.0915905237,-0.0219255704,0.1695368588,-0.1137782633,0.2228228748,0.3527334929,-0.1476604491,0.1830546558,0.1812524498,0.0453062244,0.1507354528,-0.4711746573,-0.0076138237,-0.2046742588,-0.4235938787,0.0187764075,0.3424940407,0.2327027917,0.0266376585,0.0264474768,0.2103405595,0.1510000825,0.0093807904,-0.1516859233,0.0805711374,-0.2187978178,-0.0026093989,0.1261629015,-0.0250724573,-0.0848933235,-0.1750027239,0.1665083319,-0.1399342269,0.3136582673,-0.1227992103,-0.4663731754,0.2898985445,-0.494212687,-0.1571940184,0.106676884,0.1941330135,0.1898524016,0.3169753551,-0.0033455505,0.1713694483,-0.0771825686,0.1586159617,-0.1784182787,0.2248283327,0.0160623174,0.1040193215,-0.0073826094,0.0587322824,0.1619348079,0.1394265741,0.3225094676,-0.1406376809,0.2424962372,-0.4746845663,-0.0491963215,0.4604407549,-0.0665069968,0.137398988,0.4816637933,0.0661066324,-0.0525114462,0.287420392,-0.0509188175,-0.1669092923,-0.307931006,0.2034970522,0.0864146128,0.1019393206,-0.5072076321,-0.1107438877,0.2790833116,0.0325417966,0.2052158564,-0.1718201488,-0.0322390348,-0.237004444,-0.0113939112,0.2493876666,-0.6265670657,0.1540756524,-0.0097450269,-0.004250614,0.2222199887,0.3802011907,-0.371242255,0.0840109363,-0.1422454864,0.1392559856,-0.0864757076,0.0490292422,-0.4882147908,0.1294535846,0.0305430219,-0.0098824194,0.136037454,0.1243114844,0.0164619777,0.0188371874,-0.0198179577,0.2659706473,0.1596460044,-0.1190212071,-0.0785414279,-0.0919746459,0.1880332828,0.0425336324,-0.1309635341,-0.1111637652,-0.037798252,-0.058344353,0.0506681018,-0.0628661066,-0.1312288791,0.3077427745,0.102279827,0.0275660791,-0.1993941218,-0.1144554392,-0.2583920658,-0.0028342633,0.0060153445,-0.2196477205,-0.3686653972,0.0322350599,-0.3015362918,0.4142707586,-0.003701322,0.1467631906,0.1555080861,0.0351495408,-0.1392640173,-0.3405443728,-0.1125302464,0.3829867542,-0.2520798147,0.117400378,-0.2488704324,0.6389963627,-0.1225095615,-0.2358955741,0.102029413,-0.0589136705,0.0078237727,-0.0617898889,-0.0541578941,0.3003418446,0.2589096129,-0.0539924949,-0.1232170388,-0.1712924838,0.1711663902,-0.4476631284,-0.0194845684,0.3259800076,0.2496611625,0.0040201214,0.2005787194,0.2582695484,-0.230536446,0.1718457788,0.0606855899,0.0135376034,0.2570950389,0.2372244,-0.0010136174,-0.1002517715,-0.0585765764,0.1710237116,0.2922351956,0.0836870968,-0.3078921139,-0.0896601155,0.0380948409,0.0779544115,0.1145797297,0.1024587899,-0.1011391133,-0.1450705826,0.2871288657,-0.15945144,0.36046803,0.2623209059,-0.3645984828,-0.3160845041,-0.1206738278,0.0692813694,0.3395498991,0.2408055216,0.4563713074,-0.0912684724,0.0961696357,-0.0320504308,-0.2791270316,-0.5238831639,0.0624611005,0.4196495712,-0.3285613656,0.2295600623,-0.1541727185,-0.1055464894,0.2673347592,-0.6564484835,0.0172794033,-0.1459008306,-0.3326733708,0.3113354743,0.0069784429,0.098078616,-0.3459626734,-0.1520460844,0.3765098751,0.0735324323,0.0101614492,-0.0400554463,0.1425979286,0.1107221246,0.2702673376,0.0741861239,0.2685500979,0.3350822628,0.0054364279,-0.1947278678,-0.0174658187,-0.0798835009,-0.1669342369,0.0517802648,0.0623224191,0.3742077351,-0.5855019093,-0.5999208093,0.1264113486,0.0736583471,-0.1970529705,0.1744908839,0.0701362565,0.1771164238,0.1904895902,-0.1944365352,-0.314404428,-0.3713174164,0.1869551688,0.223639369,0.3277160227,0.3159987032,0.2656185925,0.1239757687,0.2631948292,-0.1057830453,-0.1982412785,0.1272568703,0.1754867584,0.0087331478,-0.2702606916,-0.1913998574,-0.267470181,0.2551690936,0.0270455424,-0.3893484771,-0.2998656631,-0.361228317,-0.08038757,-0.3379433453,-0.0242435355,0.270458132,0.1402743757,-0.2321830541,0.0587796047,-0.1001932099,-0.0318213888,-0.1266248375,0.1443426162,-0.1426805705,0.5946801901,-0.1982656568,0.5423558354,0.1506075263,-0.0378133692,0.4440318942,-0.0917442888,0.1699853688,-0.3620955348,-0.4971083701,-0.0669029206,-0.1251891553,0.066418387,-0.1115148067,-0.2821611166,-0.1304080635,0.1161029786,-0.0095336782,-0.1359714419,0.081121996,0.1733745337,-0.0260819588,0.1395895034,-0.1173078716,-0.0486916602,-0.1003917232,-0.0609221458,-0.2855698466,-0.2101400048,0.2482716441,-0.1466301233,-0.2170103639,-0.2675659955,-0.0345852487,0.3139541745,0.3056481481,0.5451367497,0.1436631083,-0.0632694811,0.0515470766,0.0490171835,0.3031839728,0.1406436861,-0.2015923113,0.2815236449,-0.1118085608,-0.5275207758,0.0068123387,-0.2888289392,0.2861074507,-0.1511148065,-0.1364953071,-0.2830283046,-0.0095327608,0.1217485815,0.3055149317,-0.0608977675,-0.1197311059,-0.3016712368,-0.0870330483,-0.5179740787,0.1697142571,0.1463099718,0.2763128877,0.0918487087,-0.2823677063,-0.2318428606,-0.5319346786,0.0363810956,0.0667778403,0.3335453272,0.1646634936,0.1819217652,-0.2661644816,-0.117607601,0.5514033437,0.3901548088,0.2355774045,-0.0744238794,-0.0169190038,0.0978515372,0.225010246,0.0370268635,-0.0905061439,0.1164837182,-0.2906730771,0.1484722197,0.1189413071,-0.0457758009,0.046827618,-0.0387605838,-0.2239279747,-0.2735214233,0.4360409081,0.0355576724,0.1602492183,0.4475951791,0.2731461227,-0.014615803,0.1969361454,0.2358176112,0.8210725784,-0.3222990036,0.0012435849,0.3847110868,0.0425249226,0.4504195452,0.3941969275,0.1207178086,-0.3513602614,-0.2400814742,-0.109383814,-0.2209489644,0.1930115074,0.1323995143,-0.2190535814,0.1336609721,-0.1972189993,0.1078306064,0.0800062492,0.0967849493,-0.0671035945,-0.3313188255,-0.5120387077,0.1155115589,-0.2454337031,-0.1842787713,-0.154357776,-0.0854888558,-0.0121683301,-0.068457216,-0.1600129902,0.1471405327,-0.2108902186,0.5190977454,0.2066063881,-0.5216561556,0.0591063499,-0.081766732,-0.2114801556,-0.0654445961,0.0227802098,-0.1752675772,-0.1868136674,-0.2421340793,-0.148311168,0.0007692617,0.276091665,0.0007844663,-0.1791667789,0.1693535894,-0.0637616292,-0.4431357384,0.2583422661,0.3664193451,0.2560353577,-0.3944438398,-0.3703472018,-0.1771428436,-0.3999643624,-0.2454397231,0.1234320551,0.0697477609,-0.0844346657,0.1232956648,0.223929584,-0.1587904543,-0.0421671905,0.2827085257,0.2458965927,0.1164725795,0.6250904202,0.1420297027,-0.2136124223,-0.2528681457,0.0410644673,0.2418822646,-0.4038007855,0.2405816764,0.0724531636,0.0953915119,0.028948307,0.0958370343,0.2089332789,0.4507569671,-0.1446702182,-0.5494536161,-0.5519043803,0.3133969903,-0.3650940955,0.1240868196,0.0728808716,0.0119035337,-0.0238186233,0.0976890922,-0.2888030708,-0.0474176705,-0.3507489264,0.059047889,-0.0160403736,-0.0977914259,-0.012049037,-0.0036973096,0.2137143612,0.2369950712,-0.1554045677,-0.3096164763,0.1651388556,0.0450443551,0.1219118088,-0.4085612595,-0.1422498524,-0.1819286346,-0.0714023635,-0.130444333,-0.1986453831,0.0088194683,-0.0786494166,-0.137713328,0.2962082326,-0.0935740992,-0.2675312161,0.0401337668,-0.2507911325,0.2575215995,0.0955288559,0.0755477995,0.3524129093,0.025937954,-0.2118298709,0.0399078913,-0.0894020721,-0.1462773085,0.3332202137,-0.3912499547,-0.3031314015,0.0155778779,0.4696730077,0.2980320156,-0.1497734338,0.0706827417,-0.0487037972,0.2152475268,-0.3896286488,-0.166516915,-0.0059126685,0.0592268333,-0.0484120399,0.3196217418,0.2594074011,-0.2182893902,0.2811603844,0.2165706158,0.2245479077,0.0140968384,-0.1251775175,0.6010939479,-0.1161141619,-0.2117932141,0.4168370068,0.2926371396,0.5112424493,0.26060763,-0.1071603373,0.1381422877,0.2688388228,0.1799882352,-0.0583727658,-0.4001861513,-0.1858804226,0.1077909619,0.1066513956,-0.138240248,0.1256545335,-0.028011376,0.2579557598,0.0361022688,0.1478007436,0.0859167874,-0.1818326712,0.0219459999,0.000807327,0.1249998361,0.0484733023,-0.2068495601,0.0252287332,-0.3063671291,0.2777527571,0.0016674157,-0.0858000293,-0.4399663508,0.1806315184,0.5117088556,-0.1094691604,-0.1741171777,0.328587383,0.4242916107,-0.0237483382,-0.1503320634,0.3860247135,0.5804914832,0.2409934551,-0.0238724966,0.2077873796,-0.0081494143,-0.1690758169,-0.0604744181,0.0909238458,0.0978051424,0.1481955349,0.1423253566,0.2105998993,-0.1486494541,0.2147817761,0.3060647845,0.106562227,-0.2027076781,0.1401877254,-0.116266422,-0.2585432827,-0.0850696713,-0.0174536537,-0.4101569951,0.1649530232,0.4182043672,-0.120069094,-0.016555516,0.1644590348,0.0681711882,-0.3404383957,0.4425357878,0.2781794965,0.2539971471,-0.1694387048,0.0014491276,-0.4343959689,0.2145670801,-0.3108062744,-0.1265728623,0.1883103102,0.3053250313,0.2690074742,0.3885180056,0.1593928039,0.1569662392,0.0389915444,0.1325366944,-0.3590078056,-0.1170711219,0.1671163887,0.115293242,-0.171261549,-0.1662117839,0.3117127419,0.0199156646,0.10100732,-0.2547494173,-0.0726020187,-0.2228352875,0.3414654732,0.3372592926,0.1324576437,0.3376508355,0.0862659737,0.2375318557,-0.0411623679,-0.4088019729,-0.0082033034,0.5613906384,0.3403669596,0.5017957091,-0.1312077641,-0.1257840246,-0.143103987,0.3379398584,-0.2317003608,-0.042705778,-0.3095024228,0.0799118504,-0.1820611507,-0.0606567971,-0.3343918025,-0.0018313729,0.0890034661,0.2038212121,-0.2043199539,-0.5852287412,0.4368210137,-0.1386185437,-0.4015676379,0.0768632218,0.0833692923,0.0400837772,-0.0228591729,-0.2593250275,0.1473932862,0.2686134577,-0.0831846893,-0.3365375698,0.0610045604,-0.028715238,0.2706465721,-0.1467019916,0.3606089652,-0.1064436436,-0.0384901576,-0.1178414598,-0.1704109609]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/611","title":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648","comments":"`embedding` is `np.array` of shape `(128,)`. `picture_url` is url, such as 'https:\/\/i.ebayimg.com\/00\/s\/MTE5OVgxNjAw\/z\/ZOsAAOSwAG9fHQq5\/$_12.JPG?set_id=880000500F;https:\/\/i.ebayimg.com\/00\/s\/MTE5OVgxNjAw\/z\/OSgAAOSwokBfHQq8\/$_12.JPG?set_id=880000500F'","body":"Hi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```","comment_length":12,"text":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\nHi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```\n`embedding` is `np.array` of shape `(128,)`. `picture_url` is url, such as 'https:\/\/i.ebayimg.com\/00\/s\/MTE5OVgxNjAw\/z\/ZOsAAOSwAG9fHQq5\/$_12.JPG?set_id=880000500F;https:\/\/i.ebayimg.com\/00\/s\/MTE5OVgxNjAw\/z\/OSgAAOSwokBfHQq8\/$_12.JPG?set_id=880000500F'","embeddings":[-0.2813221216,-0.0744086578,-0.2195497006,0.4159464538,0.2823321819,-0.0299135447,0.4081522524,0.2113159895,0.3901511431,0.047444731,0.0003714884,0.3082345426,-0.0869823992,0.0685475916,0.000835677,-0.3419586718,-0.0933910608,0.2279389799,-0.275508076,0.0938671231,-0.2262557447,0.0042208233,-0.2458188832,0.2069006562,-0.099163264,-0.1639172286,0.0946853608,-0.0185281597,-0.0302411988,-0.4899121225,0.3556722403,-0.2945443392,0.0409597605,0.1827444583,-0.0001102196,0.0405716449,0.3709658682,0.0665661767,-0.2005666047,-0.1553708464,-0.3287505507,-0.4413967729,0.4300814569,-0.0768523142,0.2361876369,-0.2503988445,-0.2843058705,-0.3228805363,0.3386787176,0.0333642215,0.2106780857,0.2027079761,0.4377660751,0.121098958,0.0229463894,-0.0303455219,-0.1591518968,0.4214888215,0.1034795642,-0.2736412883,-0.0080701439,0.0098081287,-0.142464757,0.3750428259,0.148845315,-0.0841920599,-0.1104783043,-0.3776778579,-0.2088728249,0.2992431521,0.6146755219,-0.2124876976,-0.3339965045,-0.0728103667,-0.0512901135,-0.3944200575,0.0467973463,0.2284300476,-0.0249728207,0.055753652,0.0425886177,0.0693220422,-0.2296036631,0.0915905237,-0.0219255704,0.1695368588,-0.1137782633,0.2228228748,0.3527334929,-0.1476604491,0.1830546558,0.1812524498,0.0453062244,0.1507354528,-0.4711746573,-0.0076138237,-0.2046742588,-0.4235938787,0.0187764075,0.3424940407,0.2327027917,0.0266376585,0.0264474768,0.2103405595,0.1510000825,0.0093807904,-0.1516859233,0.0805711374,-0.2187978178,-0.0026093989,0.1261629015,-0.0250724573,-0.0848933235,-0.1750027239,0.1665083319,-0.1399342269,0.3136582673,-0.1227992103,-0.4663731754,0.2898985445,-0.494212687,-0.1571940184,0.106676884,0.1941330135,0.1898524016,0.3169753551,-0.0033455505,0.1713694483,-0.0771825686,0.1586159617,-0.1784182787,0.2248283327,0.0160623174,0.1040193215,-0.0073826094,0.0587322824,0.1619348079,0.1394265741,0.3225094676,-0.1406376809,0.2424962372,-0.4746845663,-0.0491963215,0.4604407549,-0.0665069968,0.137398988,0.4816637933,0.0661066324,-0.0525114462,0.287420392,-0.0509188175,-0.1669092923,-0.307931006,0.2034970522,0.0864146128,0.1019393206,-0.5072076321,-0.1107438877,0.2790833116,0.0325417966,0.2052158564,-0.1718201488,-0.0322390348,-0.237004444,-0.0113939112,0.2493876666,-0.6265670657,0.1540756524,-0.0097450269,-0.004250614,0.2222199887,0.3802011907,-0.371242255,0.0840109363,-0.1422454864,0.1392559856,-0.0864757076,0.0490292422,-0.4882147908,0.1294535846,0.0305430219,-0.0098824194,0.136037454,0.1243114844,0.0164619777,0.0188371874,-0.0198179577,0.2659706473,0.1596460044,-0.1190212071,-0.0785414279,-0.0919746459,0.1880332828,0.0425336324,-0.1309635341,-0.1111637652,-0.037798252,-0.058344353,0.0506681018,-0.0628661066,-0.1312288791,0.3077427745,0.102279827,0.0275660791,-0.1993941218,-0.1144554392,-0.2583920658,-0.0028342633,0.0060153445,-0.2196477205,-0.3686653972,0.0322350599,-0.3015362918,0.4142707586,-0.003701322,0.1467631906,0.1555080861,0.0351495408,-0.1392640173,-0.3405443728,-0.1125302464,0.3829867542,-0.2520798147,0.117400378,-0.2488704324,0.6389963627,-0.1225095615,-0.2358955741,0.102029413,-0.0589136705,0.0078237727,-0.0617898889,-0.0541578941,0.3003418446,0.2589096129,-0.0539924949,-0.1232170388,-0.1712924838,0.1711663902,-0.4476631284,-0.0194845684,0.3259800076,0.2496611625,0.0040201214,0.2005787194,0.2582695484,-0.230536446,0.1718457788,0.0606855899,0.0135376034,0.2570950389,0.2372244,-0.0010136174,-0.1002517715,-0.0585765764,0.1710237116,0.2922351956,0.0836870968,-0.3078921139,-0.0896601155,0.0380948409,0.0779544115,0.1145797297,0.1024587899,-0.1011391133,-0.1450705826,0.2871288657,-0.15945144,0.36046803,0.2623209059,-0.3645984828,-0.3160845041,-0.1206738278,0.0692813694,0.3395498991,0.2408055216,0.4563713074,-0.0912684724,0.0961696357,-0.0320504308,-0.2791270316,-0.5238831639,0.0624611005,0.4196495712,-0.3285613656,0.2295600623,-0.1541727185,-0.1055464894,0.2673347592,-0.6564484835,0.0172794033,-0.1459008306,-0.3326733708,0.3113354743,0.0069784429,0.098078616,-0.3459626734,-0.1520460844,0.3765098751,0.0735324323,0.0101614492,-0.0400554463,0.1425979286,0.1107221246,0.2702673376,0.0741861239,0.2685500979,0.3350822628,0.0054364279,-0.1947278678,-0.0174658187,-0.0798835009,-0.1669342369,0.0517802648,0.0623224191,0.3742077351,-0.5855019093,-0.5999208093,0.1264113486,0.0736583471,-0.1970529705,0.1744908839,0.0701362565,0.1771164238,0.1904895902,-0.1944365352,-0.314404428,-0.3713174164,0.1869551688,0.223639369,0.3277160227,0.3159987032,0.2656185925,0.1239757687,0.2631948292,-0.1057830453,-0.1982412785,0.1272568703,0.1754867584,0.0087331478,-0.2702606916,-0.1913998574,-0.267470181,0.2551690936,0.0270455424,-0.3893484771,-0.2998656631,-0.361228317,-0.08038757,-0.3379433453,-0.0242435355,0.270458132,0.1402743757,-0.2321830541,0.0587796047,-0.1001932099,-0.0318213888,-0.1266248375,0.1443426162,-0.1426805705,0.5946801901,-0.1982656568,0.5423558354,0.1506075263,-0.0378133692,0.4440318942,-0.0917442888,0.1699853688,-0.3620955348,-0.4971083701,-0.0669029206,-0.1251891553,0.066418387,-0.1115148067,-0.2821611166,-0.1304080635,0.1161029786,-0.0095336782,-0.1359714419,0.081121996,0.1733745337,-0.0260819588,0.1395895034,-0.1173078716,-0.0486916602,-0.1003917232,-0.0609221458,-0.2855698466,-0.2101400048,0.2482716441,-0.1466301233,-0.2170103639,-0.2675659955,-0.0345852487,0.3139541745,0.3056481481,0.5451367497,0.1436631083,-0.0632694811,0.0515470766,0.0490171835,0.3031839728,0.1406436861,-0.2015923113,0.2815236449,-0.1118085608,-0.5275207758,0.0068123387,-0.2888289392,0.2861074507,-0.1511148065,-0.1364953071,-0.2830283046,-0.0095327608,0.1217485815,0.3055149317,-0.0608977675,-0.1197311059,-0.3016712368,-0.0870330483,-0.5179740787,0.1697142571,0.1463099718,0.2763128877,0.0918487087,-0.2823677063,-0.2318428606,-0.5319346786,0.0363810956,0.0667778403,0.3335453272,0.1646634936,0.1819217652,-0.2661644816,-0.117607601,0.5514033437,0.3901548088,0.2355774045,-0.0744238794,-0.0169190038,0.0978515372,0.225010246,0.0370268635,-0.0905061439,0.1164837182,-0.2906730771,0.1484722197,0.1189413071,-0.0457758009,0.046827618,-0.0387605838,-0.2239279747,-0.2735214233,0.4360409081,0.0355576724,0.1602492183,0.4475951791,0.2731461227,-0.014615803,0.1969361454,0.2358176112,0.8210725784,-0.3222990036,0.0012435849,0.3847110868,0.0425249226,0.4504195452,0.3941969275,0.1207178086,-0.3513602614,-0.2400814742,-0.109383814,-0.2209489644,0.1930115074,0.1323995143,-0.2190535814,0.1336609721,-0.1972189993,0.1078306064,0.0800062492,0.0967849493,-0.0671035945,-0.3313188255,-0.5120387077,0.1155115589,-0.2454337031,-0.1842787713,-0.154357776,-0.0854888558,-0.0121683301,-0.068457216,-0.1600129902,0.1471405327,-0.2108902186,0.5190977454,0.2066063881,-0.5216561556,0.0591063499,-0.081766732,-0.2114801556,-0.0654445961,0.0227802098,-0.1752675772,-0.1868136674,-0.2421340793,-0.148311168,0.0007692617,0.276091665,0.0007844663,-0.1791667789,0.1693535894,-0.0637616292,-0.4431357384,0.2583422661,0.3664193451,0.2560353577,-0.3944438398,-0.3703472018,-0.1771428436,-0.3999643624,-0.2454397231,0.1234320551,0.0697477609,-0.0844346657,0.1232956648,0.223929584,-0.1587904543,-0.0421671905,0.2827085257,0.2458965927,0.1164725795,0.6250904202,0.1420297027,-0.2136124223,-0.2528681457,0.0410644673,0.2418822646,-0.4038007855,0.2405816764,0.0724531636,0.0953915119,0.028948307,0.0958370343,0.2089332789,0.4507569671,-0.1446702182,-0.5494536161,-0.5519043803,0.3133969903,-0.3650940955,0.1240868196,0.0728808716,0.0119035337,-0.0238186233,0.0976890922,-0.2888030708,-0.0474176705,-0.3507489264,0.059047889,-0.0160403736,-0.0977914259,-0.012049037,-0.0036973096,0.2137143612,0.2369950712,-0.1554045677,-0.3096164763,0.1651388556,0.0450443551,0.1219118088,-0.4085612595,-0.1422498524,-0.1819286346,-0.0714023635,-0.130444333,-0.1986453831,0.0088194683,-0.0786494166,-0.137713328,0.2962082326,-0.0935740992,-0.2675312161,0.0401337668,-0.2507911325,0.2575215995,0.0955288559,0.0755477995,0.3524129093,0.025937954,-0.2118298709,0.0399078913,-0.0894020721,-0.1462773085,0.3332202137,-0.3912499547,-0.3031314015,0.0155778779,0.4696730077,0.2980320156,-0.1497734338,0.0706827417,-0.0487037972,0.2152475268,-0.3896286488,-0.166516915,-0.0059126685,0.0592268333,-0.0484120399,0.3196217418,0.2594074011,-0.2182893902,0.2811603844,0.2165706158,0.2245479077,0.0140968384,-0.1251775175,0.6010939479,-0.1161141619,-0.2117932141,0.4168370068,0.2926371396,0.5112424493,0.26060763,-0.1071603373,0.1381422877,0.2688388228,0.1799882352,-0.0583727658,-0.4001861513,-0.1858804226,0.1077909619,0.1066513956,-0.138240248,0.1256545335,-0.028011376,0.2579557598,0.0361022688,0.1478007436,0.0859167874,-0.1818326712,0.0219459999,0.000807327,0.1249998361,0.0484733023,-0.2068495601,0.0252287332,-0.3063671291,0.2777527571,0.0016674157,-0.0858000293,-0.4399663508,0.1806315184,0.5117088556,-0.1094691604,-0.1741171777,0.328587383,0.4242916107,-0.0237483382,-0.1503320634,0.3860247135,0.5804914832,0.2409934551,-0.0238724966,0.2077873796,-0.0081494143,-0.1690758169,-0.0604744181,0.0909238458,0.0978051424,0.1481955349,0.1423253566,0.2105998993,-0.1486494541,0.2147817761,0.3060647845,0.106562227,-0.2027076781,0.1401877254,-0.116266422,-0.2585432827,-0.0850696713,-0.0174536537,-0.4101569951,0.1649530232,0.4182043672,-0.120069094,-0.016555516,0.1644590348,0.0681711882,-0.3404383957,0.4425357878,0.2781794965,0.2539971471,-0.1694387048,0.0014491276,-0.4343959689,0.2145670801,-0.3108062744,-0.1265728623,0.1883103102,0.3053250313,0.2690074742,0.3885180056,0.1593928039,0.1569662392,0.0389915444,0.1325366944,-0.3590078056,-0.1170711219,0.1671163887,0.115293242,-0.171261549,-0.1662117839,0.3117127419,0.0199156646,0.10100732,-0.2547494173,-0.0726020187,-0.2228352875,0.3414654732,0.3372592926,0.1324576437,0.3376508355,0.0862659737,0.2375318557,-0.0411623679,-0.4088019729,-0.0082033034,0.5613906384,0.3403669596,0.5017957091,-0.1312077641,-0.1257840246,-0.143103987,0.3379398584,-0.2317003608,-0.042705778,-0.3095024228,0.0799118504,-0.1820611507,-0.0606567971,-0.3343918025,-0.0018313729,0.0890034661,0.2038212121,-0.2043199539,-0.5852287412,0.4368210137,-0.1386185437,-0.4015676379,0.0768632218,0.0833692923,0.0400837772,-0.0228591729,-0.2593250275,0.1473932862,0.2686134577,-0.0831846893,-0.3365375698,0.0610045604,-0.028715238,0.2706465721,-0.1467019916,0.3606089652,-0.1064436436,-0.0384901576,-0.1178414598,-0.1704109609]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/611","title":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648","comments":"It looks like a Pyarrow limitation.\r\nI was able to reproduce the error with \r\n\r\n```python\r\nimport pandas as pd\r\nimport numpy as np\r\nimport pyarrow as pa\r\n\r\n n = 1713614\r\ndf = pd.DataFrame.from_dict({\"a\": list(np.zeros((n, 128))), \"b\": range(n)})\r\npa.Table.from_pandas(df)\r\n```\r\n\r\nI also tried with 50% of the dataframe and it actually works.\r\nI created an issue on Apache Arrow's JIRA [here](https:\/\/issues.apache.org\/jira\/browse\/ARROW-9976)\r\n\r\nOne way to fix that would be to chunk the dataframe and concatenate arrow tables.","body":"Hi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```","comment_length":75,"text":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\nHi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```\nIt looks like a Pyarrow limitation.\r\nI was able to reproduce the error with \r\n\r\n```python\r\nimport pandas as pd\r\nimport numpy as np\r\nimport pyarrow as pa\r\n\r\n n = 1713614\r\ndf = pd.DataFrame.from_dict({\"a\": list(np.zeros((n, 128))), \"b\": range(n)})\r\npa.Table.from_pandas(df)\r\n```\r\n\r\nI also tried with 50% of the dataframe and it actually works.\r\nI created an issue on Apache Arrow's JIRA [here](https:\/\/issues.apache.org\/jira\/browse\/ARROW-9976)\r\n\r\nOne way to fix that would be to chunk the dataframe and concatenate arrow tables.","embeddings":[-0.2813221216,-0.0744086578,-0.2195497006,0.4159464538,0.2823321819,-0.0299135447,0.4081522524,0.2113159895,0.3901511431,0.047444731,0.0003714884,0.3082345426,-0.0869823992,0.0685475916,0.000835677,-0.3419586718,-0.0933910608,0.2279389799,-0.275508076,0.0938671231,-0.2262557447,0.0042208233,-0.2458188832,0.2069006562,-0.099163264,-0.1639172286,0.0946853608,-0.0185281597,-0.0302411988,-0.4899121225,0.3556722403,-0.2945443392,0.0409597605,0.1827444583,-0.0001102196,0.0405716449,0.3709658682,0.0665661767,-0.2005666047,-0.1553708464,-0.3287505507,-0.4413967729,0.4300814569,-0.0768523142,0.2361876369,-0.2503988445,-0.2843058705,-0.3228805363,0.3386787176,0.0333642215,0.2106780857,0.2027079761,0.4377660751,0.121098958,0.0229463894,-0.0303455219,-0.1591518968,0.4214888215,0.1034795642,-0.2736412883,-0.0080701439,0.0098081287,-0.142464757,0.3750428259,0.148845315,-0.0841920599,-0.1104783043,-0.3776778579,-0.2088728249,0.2992431521,0.6146755219,-0.2124876976,-0.3339965045,-0.0728103667,-0.0512901135,-0.3944200575,0.0467973463,0.2284300476,-0.0249728207,0.055753652,0.0425886177,0.0693220422,-0.2296036631,0.0915905237,-0.0219255704,0.1695368588,-0.1137782633,0.2228228748,0.3527334929,-0.1476604491,0.1830546558,0.1812524498,0.0453062244,0.1507354528,-0.4711746573,-0.0076138237,-0.2046742588,-0.4235938787,0.0187764075,0.3424940407,0.2327027917,0.0266376585,0.0264474768,0.2103405595,0.1510000825,0.0093807904,-0.1516859233,0.0805711374,-0.2187978178,-0.0026093989,0.1261629015,-0.0250724573,-0.0848933235,-0.1750027239,0.1665083319,-0.1399342269,0.3136582673,-0.1227992103,-0.4663731754,0.2898985445,-0.494212687,-0.1571940184,0.106676884,0.1941330135,0.1898524016,0.3169753551,-0.0033455505,0.1713694483,-0.0771825686,0.1586159617,-0.1784182787,0.2248283327,0.0160623174,0.1040193215,-0.0073826094,0.0587322824,0.1619348079,0.1394265741,0.3225094676,-0.1406376809,0.2424962372,-0.4746845663,-0.0491963215,0.4604407549,-0.0665069968,0.137398988,0.4816637933,0.0661066324,-0.0525114462,0.287420392,-0.0509188175,-0.1669092923,-0.307931006,0.2034970522,0.0864146128,0.1019393206,-0.5072076321,-0.1107438877,0.2790833116,0.0325417966,0.2052158564,-0.1718201488,-0.0322390348,-0.237004444,-0.0113939112,0.2493876666,-0.6265670657,0.1540756524,-0.0097450269,-0.004250614,0.2222199887,0.3802011907,-0.371242255,0.0840109363,-0.1422454864,0.1392559856,-0.0864757076,0.0490292422,-0.4882147908,0.1294535846,0.0305430219,-0.0098824194,0.136037454,0.1243114844,0.0164619777,0.0188371874,-0.0198179577,0.2659706473,0.1596460044,-0.1190212071,-0.0785414279,-0.0919746459,0.1880332828,0.0425336324,-0.1309635341,-0.1111637652,-0.037798252,-0.058344353,0.0506681018,-0.0628661066,-0.1312288791,0.3077427745,0.102279827,0.0275660791,-0.1993941218,-0.1144554392,-0.2583920658,-0.0028342633,0.0060153445,-0.2196477205,-0.3686653972,0.0322350599,-0.3015362918,0.4142707586,-0.003701322,0.1467631906,0.1555080861,0.0351495408,-0.1392640173,-0.3405443728,-0.1125302464,0.3829867542,-0.2520798147,0.117400378,-0.2488704324,0.6389963627,-0.1225095615,-0.2358955741,0.102029413,-0.0589136705,0.0078237727,-0.0617898889,-0.0541578941,0.3003418446,0.2589096129,-0.0539924949,-0.1232170388,-0.1712924838,0.1711663902,-0.4476631284,-0.0194845684,0.3259800076,0.2496611625,0.0040201214,0.2005787194,0.2582695484,-0.230536446,0.1718457788,0.0606855899,0.0135376034,0.2570950389,0.2372244,-0.0010136174,-0.1002517715,-0.0585765764,0.1710237116,0.2922351956,0.0836870968,-0.3078921139,-0.0896601155,0.0380948409,0.0779544115,0.1145797297,0.1024587899,-0.1011391133,-0.1450705826,0.2871288657,-0.15945144,0.36046803,0.2623209059,-0.3645984828,-0.3160845041,-0.1206738278,0.0692813694,0.3395498991,0.2408055216,0.4563713074,-0.0912684724,0.0961696357,-0.0320504308,-0.2791270316,-0.5238831639,0.0624611005,0.4196495712,-0.3285613656,0.2295600623,-0.1541727185,-0.1055464894,0.2673347592,-0.6564484835,0.0172794033,-0.1459008306,-0.3326733708,0.3113354743,0.0069784429,0.098078616,-0.3459626734,-0.1520460844,0.3765098751,0.0735324323,0.0101614492,-0.0400554463,0.1425979286,0.1107221246,0.2702673376,0.0741861239,0.2685500979,0.3350822628,0.0054364279,-0.1947278678,-0.0174658187,-0.0798835009,-0.1669342369,0.0517802648,0.0623224191,0.3742077351,-0.5855019093,-0.5999208093,0.1264113486,0.0736583471,-0.1970529705,0.1744908839,0.0701362565,0.1771164238,0.1904895902,-0.1944365352,-0.314404428,-0.3713174164,0.1869551688,0.223639369,0.3277160227,0.3159987032,0.2656185925,0.1239757687,0.2631948292,-0.1057830453,-0.1982412785,0.1272568703,0.1754867584,0.0087331478,-0.2702606916,-0.1913998574,-0.267470181,0.2551690936,0.0270455424,-0.3893484771,-0.2998656631,-0.361228317,-0.08038757,-0.3379433453,-0.0242435355,0.270458132,0.1402743757,-0.2321830541,0.0587796047,-0.1001932099,-0.0318213888,-0.1266248375,0.1443426162,-0.1426805705,0.5946801901,-0.1982656568,0.5423558354,0.1506075263,-0.0378133692,0.4440318942,-0.0917442888,0.1699853688,-0.3620955348,-0.4971083701,-0.0669029206,-0.1251891553,0.066418387,-0.1115148067,-0.2821611166,-0.1304080635,0.1161029786,-0.0095336782,-0.1359714419,0.081121996,0.1733745337,-0.0260819588,0.1395895034,-0.1173078716,-0.0486916602,-0.1003917232,-0.0609221458,-0.2855698466,-0.2101400048,0.2482716441,-0.1466301233,-0.2170103639,-0.2675659955,-0.0345852487,0.3139541745,0.3056481481,0.5451367497,0.1436631083,-0.0632694811,0.0515470766,0.0490171835,0.3031839728,0.1406436861,-0.2015923113,0.2815236449,-0.1118085608,-0.5275207758,0.0068123387,-0.2888289392,0.2861074507,-0.1511148065,-0.1364953071,-0.2830283046,-0.0095327608,0.1217485815,0.3055149317,-0.0608977675,-0.1197311059,-0.3016712368,-0.0870330483,-0.5179740787,0.1697142571,0.1463099718,0.2763128877,0.0918487087,-0.2823677063,-0.2318428606,-0.5319346786,0.0363810956,0.0667778403,0.3335453272,0.1646634936,0.1819217652,-0.2661644816,-0.117607601,0.5514033437,0.3901548088,0.2355774045,-0.0744238794,-0.0169190038,0.0978515372,0.225010246,0.0370268635,-0.0905061439,0.1164837182,-0.2906730771,0.1484722197,0.1189413071,-0.0457758009,0.046827618,-0.0387605838,-0.2239279747,-0.2735214233,0.4360409081,0.0355576724,0.1602492183,0.4475951791,0.2731461227,-0.014615803,0.1969361454,0.2358176112,0.8210725784,-0.3222990036,0.0012435849,0.3847110868,0.0425249226,0.4504195452,0.3941969275,0.1207178086,-0.3513602614,-0.2400814742,-0.109383814,-0.2209489644,0.1930115074,0.1323995143,-0.2190535814,0.1336609721,-0.1972189993,0.1078306064,0.0800062492,0.0967849493,-0.0671035945,-0.3313188255,-0.5120387077,0.1155115589,-0.2454337031,-0.1842787713,-0.154357776,-0.0854888558,-0.0121683301,-0.068457216,-0.1600129902,0.1471405327,-0.2108902186,0.5190977454,0.2066063881,-0.5216561556,0.0591063499,-0.081766732,-0.2114801556,-0.0654445961,0.0227802098,-0.1752675772,-0.1868136674,-0.2421340793,-0.148311168,0.0007692617,0.276091665,0.0007844663,-0.1791667789,0.1693535894,-0.0637616292,-0.4431357384,0.2583422661,0.3664193451,0.2560353577,-0.3944438398,-0.3703472018,-0.1771428436,-0.3999643624,-0.2454397231,0.1234320551,0.0697477609,-0.0844346657,0.1232956648,0.223929584,-0.1587904543,-0.0421671905,0.2827085257,0.2458965927,0.1164725795,0.6250904202,0.1420297027,-0.2136124223,-0.2528681457,0.0410644673,0.2418822646,-0.4038007855,0.2405816764,0.0724531636,0.0953915119,0.028948307,0.0958370343,0.2089332789,0.4507569671,-0.1446702182,-0.5494536161,-0.5519043803,0.3133969903,-0.3650940955,0.1240868196,0.0728808716,0.0119035337,-0.0238186233,0.0976890922,-0.2888030708,-0.0474176705,-0.3507489264,0.059047889,-0.0160403736,-0.0977914259,-0.012049037,-0.0036973096,0.2137143612,0.2369950712,-0.1554045677,-0.3096164763,0.1651388556,0.0450443551,0.1219118088,-0.4085612595,-0.1422498524,-0.1819286346,-0.0714023635,-0.130444333,-0.1986453831,0.0088194683,-0.0786494166,-0.137713328,0.2962082326,-0.0935740992,-0.2675312161,0.0401337668,-0.2507911325,0.2575215995,0.0955288559,0.0755477995,0.3524129093,0.025937954,-0.2118298709,0.0399078913,-0.0894020721,-0.1462773085,0.3332202137,-0.3912499547,-0.3031314015,0.0155778779,0.4696730077,0.2980320156,-0.1497734338,0.0706827417,-0.0487037972,0.2152475268,-0.3896286488,-0.166516915,-0.0059126685,0.0592268333,-0.0484120399,0.3196217418,0.2594074011,-0.2182893902,0.2811603844,0.2165706158,0.2245479077,0.0140968384,-0.1251775175,0.6010939479,-0.1161141619,-0.2117932141,0.4168370068,0.2926371396,0.5112424493,0.26060763,-0.1071603373,0.1381422877,0.2688388228,0.1799882352,-0.0583727658,-0.4001861513,-0.1858804226,0.1077909619,0.1066513956,-0.138240248,0.1256545335,-0.028011376,0.2579557598,0.0361022688,0.1478007436,0.0859167874,-0.1818326712,0.0219459999,0.000807327,0.1249998361,0.0484733023,-0.2068495601,0.0252287332,-0.3063671291,0.2777527571,0.0016674157,-0.0858000293,-0.4399663508,0.1806315184,0.5117088556,-0.1094691604,-0.1741171777,0.328587383,0.4242916107,-0.0237483382,-0.1503320634,0.3860247135,0.5804914832,0.2409934551,-0.0238724966,0.2077873796,-0.0081494143,-0.1690758169,-0.0604744181,0.0909238458,0.0978051424,0.1481955349,0.1423253566,0.2105998993,-0.1486494541,0.2147817761,0.3060647845,0.106562227,-0.2027076781,0.1401877254,-0.116266422,-0.2585432827,-0.0850696713,-0.0174536537,-0.4101569951,0.1649530232,0.4182043672,-0.120069094,-0.016555516,0.1644590348,0.0681711882,-0.3404383957,0.4425357878,0.2781794965,0.2539971471,-0.1694387048,0.0014491276,-0.4343959689,0.2145670801,-0.3108062744,-0.1265728623,0.1883103102,0.3053250313,0.2690074742,0.3885180056,0.1593928039,0.1569662392,0.0389915444,0.1325366944,-0.3590078056,-0.1170711219,0.1671163887,0.115293242,-0.171261549,-0.1662117839,0.3117127419,0.0199156646,0.10100732,-0.2547494173,-0.0726020187,-0.2228352875,0.3414654732,0.3372592926,0.1324576437,0.3376508355,0.0862659737,0.2375318557,-0.0411623679,-0.4088019729,-0.0082033034,0.5613906384,0.3403669596,0.5017957091,-0.1312077641,-0.1257840246,-0.143103987,0.3379398584,-0.2317003608,-0.042705778,-0.3095024228,0.0799118504,-0.1820611507,-0.0606567971,-0.3343918025,-0.0018313729,0.0890034661,0.2038212121,-0.2043199539,-0.5852287412,0.4368210137,-0.1386185437,-0.4015676379,0.0768632218,0.0833692923,0.0400837772,-0.0228591729,-0.2593250275,0.1473932862,0.2686134577,-0.0831846893,-0.3365375698,0.0610045604,-0.028715238,0.2706465721,-0.1467019916,0.3606089652,-0.1064436436,-0.0384901576,-0.1178414598,-0.1704109609]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/611","title":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648","comments":"It looks like it's going to be fixed in pyarrow 2.0.0 :)\r\n\r\nIn the meantime I suggest to chunk big dataframes to create several small datasets, and then concatenate them using [concatenate_datasets](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?highlight=concatenate#datasets.concatenate_datasets)","body":"Hi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```","comment_length":32,"text":"ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\nHi, I'm trying to load a dataset from Dataframe, but I get the error:\r\n```bash\r\n---------------------------------------------------------------------------\r\nArrowCapacityError Traceback (most recent call last)\r\n<ipython-input-7-146b6b495963> in <module>\r\n----> 1 dataset = Dataset.from_pandas(emb)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in from_pandas(cls, df, features, info, split)\r\n 223 info.features = features\r\n 224 pa_table: pa.Table = pa.Table.from_pandas(\r\n--> 225 df=df, schema=pa.schema(features.type) if features is not None else None\r\n 226 )\r\n 227 return cls(pa_table, info=info, split=split)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/table.pxi in pyarrow.lib.Table.from_pandas()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in dataframe_to_arrays(df, schema, preserve_index, nthreads, columns, safe)\r\n 591 for i, maybe_fut in enumerate(arrays):\r\n 592 if isinstance(maybe_fut, futures.Future):\r\n--> 593 arrays[i] = maybe_fut.result()\r\n 594 \r\n 595 types = [x.type for x in arrays]\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 426 raise CancelledError()\r\n 427 elif self._state == FINISHED:\r\n--> 428 return self.__get_result()\r\n 429 \r\n 430 self._condition.wait(timeout)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/concurrent\/futures\/thread.py in run(self)\r\n 55 \r\n 56 try:\r\n---> 57 result = self.fn(*self.args, **self.kwargs)\r\n 58 except BaseException as exc:\r\n 59 self.future.set_exception(exc)\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py in convert_column(col, field)\r\n 557 \r\n 558 try:\r\n--> 559 result = pa.array(col, type=type_, from_pandas=True, safe=safe)\r\n 560 except (pa.ArrowInvalid,\r\n 561 pa.ArrowNotImplementedError,\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._ndarray_to_array()\r\n\r\n~\/miniconda3\/envs\/dev\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648\r\n```\r\nMy code is :\r\n```python\r\nfrom nlp import Dataset\r\ndataset = Dataset.from_pandas(emb)\r\n```\nIt looks like it's going to be fixed in pyarrow 2.0.0 :)\r\n\r\nIn the meantime I suggest to chunk big dataframes to create several small datasets, and then concatenate them using [concatenate_datasets](https:\/\/huggingface.co\/docs\/datasets\/package_reference\/main_classes.html?highlight=concatenate#datasets.concatenate_datasets)","embeddings":[-0.2813221216,-0.0744086578,-0.2195497006,0.4159464538,0.2823321819,-0.0299135447,0.4081522524,0.2113159895,0.3901511431,0.047444731,0.0003714884,0.3082345426,-0.0869823992,0.0685475916,0.000835677,-0.3419586718,-0.0933910608,0.2279389799,-0.275508076,0.0938671231,-0.2262557447,0.0042208233,-0.2458188832,0.2069006562,-0.099163264,-0.1639172286,0.0946853608,-0.0185281597,-0.0302411988,-0.4899121225,0.3556722403,-0.2945443392,0.0409597605,0.1827444583,-0.0001102196,0.0405716449,0.3709658682,0.0665661767,-0.2005666047,-0.1553708464,-0.3287505507,-0.4413967729,0.4300814569,-0.0768523142,0.2361876369,-0.2503988445,-0.2843058705,-0.3228805363,0.3386787176,0.0333642215,0.2106780857,0.2027079761,0.4377660751,0.121098958,0.0229463894,-0.0303455219,-0.1591518968,0.4214888215,0.1034795642,-0.2736412883,-0.0080701439,0.0098081287,-0.142464757,0.3750428259,0.148845315,-0.0841920599,-0.1104783043,-0.3776778579,-0.2088728249,0.2992431521,0.6146755219,-0.2124876976,-0.3339965045,-0.0728103667,-0.0512901135,-0.3944200575,0.0467973463,0.2284300476,-0.0249728207,0.055753652,0.0425886177,0.0693220422,-0.2296036631,0.0915905237,-0.0219255704,0.1695368588,-0.1137782633,0.2228228748,0.3527334929,-0.1476604491,0.1830546558,0.1812524498,0.0453062244,0.1507354528,-0.4711746573,-0.0076138237,-0.2046742588,-0.4235938787,0.0187764075,0.3424940407,0.2327027917,0.0266376585,0.0264474768,0.2103405595,0.1510000825,0.0093807904,-0.1516859233,0.0805711374,-0.2187978178,-0.0026093989,0.1261629015,-0.0250724573,-0.0848933235,-0.1750027239,0.1665083319,-0.1399342269,0.3136582673,-0.1227992103,-0.4663731754,0.2898985445,-0.494212687,-0.1571940184,0.106676884,0.1941330135,0.1898524016,0.3169753551,-0.0033455505,0.1713694483,-0.0771825686,0.1586159617,-0.1784182787,0.2248283327,0.0160623174,0.1040193215,-0.0073826094,0.0587322824,0.1619348079,0.1394265741,0.3225094676,-0.1406376809,0.2424962372,-0.4746845663,-0.0491963215,0.4604407549,-0.0665069968,0.137398988,0.4816637933,0.0661066324,-0.0525114462,0.287420392,-0.0509188175,-0.1669092923,-0.307931006,0.2034970522,0.0864146128,0.1019393206,-0.5072076321,-0.1107438877,0.2790833116,0.0325417966,0.2052158564,-0.1718201488,-0.0322390348,-0.237004444,-0.0113939112,0.2493876666,-0.6265670657,0.1540756524,-0.0097450269,-0.004250614,0.2222199887,0.3802011907,-0.371242255,0.0840109363,-0.1422454864,0.1392559856,-0.0864757076,0.0490292422,-0.4882147908,0.1294535846,0.0305430219,-0.0098824194,0.136037454,0.1243114844,0.0164619777,0.0188371874,-0.0198179577,0.2659706473,0.1596460044,-0.1190212071,-0.0785414279,-0.0919746459,0.1880332828,0.0425336324,-0.1309635341,-0.1111637652,-0.037798252,-0.058344353,0.0506681018,-0.0628661066,-0.1312288791,0.3077427745,0.102279827,0.0275660791,-0.1993941218,-0.1144554392,-0.2583920658,-0.0028342633,0.0060153445,-0.2196477205,-0.3686653972,0.0322350599,-0.3015362918,0.4142707586,-0.003701322,0.1467631906,0.1555080861,0.0351495408,-0.1392640173,-0.3405443728,-0.1125302464,0.3829867542,-0.2520798147,0.117400378,-0.2488704324,0.6389963627,-0.1225095615,-0.2358955741,0.102029413,-0.0589136705,0.0078237727,-0.0617898889,-0.0541578941,0.3003418446,0.2589096129,-0.0539924949,-0.1232170388,-0.1712924838,0.1711663902,-0.4476631284,-0.0194845684,0.3259800076,0.2496611625,0.0040201214,0.2005787194,0.2582695484,-0.230536446,0.1718457788,0.0606855899,0.0135376034,0.2570950389,0.2372244,-0.0010136174,-0.1002517715,-0.0585765764,0.1710237116,0.2922351956,0.0836870968,-0.3078921139,-0.0896601155,0.0380948409,0.0779544115,0.1145797297,0.1024587899,-0.1011391133,-0.1450705826,0.2871288657,-0.15945144,0.36046803,0.2623209059,-0.3645984828,-0.3160845041,-0.1206738278,0.0692813694,0.3395498991,0.2408055216,0.4563713074,-0.0912684724,0.0961696357,-0.0320504308,-0.2791270316,-0.5238831639,0.0624611005,0.4196495712,-0.3285613656,0.2295600623,-0.1541727185,-0.1055464894,0.2673347592,-0.6564484835,0.0172794033,-0.1459008306,-0.3326733708,0.3113354743,0.0069784429,0.098078616,-0.3459626734,-0.1520460844,0.3765098751,0.0735324323,0.0101614492,-0.0400554463,0.1425979286,0.1107221246,0.2702673376,0.0741861239,0.2685500979,0.3350822628,0.0054364279,-0.1947278678,-0.0174658187,-0.0798835009,-0.1669342369,0.0517802648,0.0623224191,0.3742077351,-0.5855019093,-0.5999208093,0.1264113486,0.0736583471,-0.1970529705,0.1744908839,0.0701362565,0.1771164238,0.1904895902,-0.1944365352,-0.314404428,-0.3713174164,0.1869551688,0.223639369,0.3277160227,0.3159987032,0.2656185925,0.1239757687,0.2631948292,-0.1057830453,-0.1982412785,0.1272568703,0.1754867584,0.0087331478,-0.2702606916,-0.1913998574,-0.267470181,0.2551690936,0.0270455424,-0.3893484771,-0.2998656631,-0.361228317,-0.08038757,-0.3379433453,-0.0242435355,0.270458132,0.1402743757,-0.2321830541,0.0587796047,-0.1001932099,-0.0318213888,-0.1266248375,0.1443426162,-0.1426805705,0.5946801901,-0.1982656568,0.5423558354,0.1506075263,-0.0378133692,0.4440318942,-0.0917442888,0.1699853688,-0.3620955348,-0.4971083701,-0.0669029206,-0.1251891553,0.066418387,-0.1115148067,-0.2821611166,-0.1304080635,0.1161029786,-0.0095336782,-0.1359714419,0.081121996,0.1733745337,-0.0260819588,0.1395895034,-0.1173078716,-0.0486916602,-0.1003917232,-0.0609221458,-0.2855698466,-0.2101400048,0.2482716441,-0.1466301233,-0.2170103639,-0.2675659955,-0.0345852487,0.3139541745,0.3056481481,0.5451367497,0.1436631083,-0.0632694811,0.0515470766,0.0490171835,0.3031839728,0.1406436861,-0.2015923113,0.2815236449,-0.1118085608,-0.5275207758,0.0068123387,-0.2888289392,0.2861074507,-0.1511148065,-0.1364953071,-0.2830283046,-0.0095327608,0.1217485815,0.3055149317,-0.0608977675,-0.1197311059,-0.3016712368,-0.0870330483,-0.5179740787,0.1697142571,0.1463099718,0.2763128877,0.0918487087,-0.2823677063,-0.2318428606,-0.5319346786,0.0363810956,0.0667778403,0.3335453272,0.1646634936,0.1819217652,-0.2661644816,-0.117607601,0.5514033437,0.3901548088,0.2355774045,-0.0744238794,-0.0169190038,0.0978515372,0.225010246,0.0370268635,-0.0905061439,0.1164837182,-0.2906730771,0.1484722197,0.1189413071,-0.0457758009,0.046827618,-0.0387605838,-0.2239279747,-0.2735214233,0.4360409081,0.0355576724,0.1602492183,0.4475951791,0.2731461227,-0.014615803,0.1969361454,0.2358176112,0.8210725784,-0.3222990036,0.0012435849,0.3847110868,0.0425249226,0.4504195452,0.3941969275,0.1207178086,-0.3513602614,-0.2400814742,-0.109383814,-0.2209489644,0.1930115074,0.1323995143,-0.2190535814,0.1336609721,-0.1972189993,0.1078306064,0.0800062492,0.0967849493,-0.0671035945,-0.3313188255,-0.5120387077,0.1155115589,-0.2454337031,-0.1842787713,-0.154357776,-0.0854888558,-0.0121683301,-0.068457216,-0.1600129902,0.1471405327,-0.2108902186,0.5190977454,0.2066063881,-0.5216561556,0.0591063499,-0.081766732,-0.2114801556,-0.0654445961,0.0227802098,-0.1752675772,-0.1868136674,-0.2421340793,-0.148311168,0.0007692617,0.276091665,0.0007844663,-0.1791667789,0.1693535894,-0.0637616292,-0.4431357384,0.2583422661,0.3664193451,0.2560353577,-0.3944438398,-0.3703472018,-0.1771428436,-0.3999643624,-0.2454397231,0.1234320551,0.0697477609,-0.0844346657,0.1232956648,0.223929584,-0.1587904543,-0.0421671905,0.2827085257,0.2458965927,0.1164725795,0.6250904202,0.1420297027,-0.2136124223,-0.2528681457,0.0410644673,0.2418822646,-0.4038007855,0.2405816764,0.0724531636,0.0953915119,0.028948307,0.0958370343,0.2089332789,0.4507569671,-0.1446702182,-0.5494536161,-0.5519043803,0.3133969903,-0.3650940955,0.1240868196,0.0728808716,0.0119035337,-0.0238186233,0.0976890922,-0.2888030708,-0.0474176705,-0.3507489264,0.059047889,-0.0160403736,-0.0977914259,-0.012049037,-0.0036973096,0.2137143612,0.2369950712,-0.1554045677,-0.3096164763,0.1651388556,0.0450443551,0.1219118088,-0.4085612595,-0.1422498524,-0.1819286346,-0.0714023635,-0.130444333,-0.1986453831,0.0088194683,-0.0786494166,-0.137713328,0.2962082326,-0.0935740992,-0.2675312161,0.0401337668,-0.2507911325,0.2575215995,0.0955288559,0.0755477995,0.3524129093,0.025937954,-0.2118298709,0.0399078913,-0.0894020721,-0.1462773085,0.3332202137,-0.3912499547,-0.3031314015,0.0155778779,0.4696730077,0.2980320156,-0.1497734338,0.0706827417,-0.0487037972,0.2152475268,-0.3896286488,-0.166516915,-0.0059126685,0.0592268333,-0.0484120399,0.3196217418,0.2594074011,-0.2182893902,0.2811603844,0.2165706158,0.2245479077,0.0140968384,-0.1251775175,0.6010939479,-0.1161141619,-0.2117932141,0.4168370068,0.2926371396,0.5112424493,0.26060763,-0.1071603373,0.1381422877,0.2688388228,0.1799882352,-0.0583727658,-0.4001861513,-0.1858804226,0.1077909619,0.1066513956,-0.138240248,0.1256545335,-0.028011376,0.2579557598,0.0361022688,0.1478007436,0.0859167874,-0.1818326712,0.0219459999,0.000807327,0.1249998361,0.0484733023,-0.2068495601,0.0252287332,-0.3063671291,0.2777527571,0.0016674157,-0.0858000293,-0.4399663508,0.1806315184,0.5117088556,-0.1094691604,-0.1741171777,0.328587383,0.4242916107,-0.0237483382,-0.1503320634,0.3860247135,0.5804914832,0.2409934551,-0.0238724966,0.2077873796,-0.0081494143,-0.1690758169,-0.0604744181,0.0909238458,0.0978051424,0.1481955349,0.1423253566,0.2105998993,-0.1486494541,0.2147817761,0.3060647845,0.106562227,-0.2027076781,0.1401877254,-0.116266422,-0.2585432827,-0.0850696713,-0.0174536537,-0.4101569951,0.1649530232,0.4182043672,-0.120069094,-0.016555516,0.1644590348,0.0681711882,-0.3404383957,0.4425357878,0.2781794965,0.2539971471,-0.1694387048,0.0014491276,-0.4343959689,0.2145670801,-0.3108062744,-0.1265728623,0.1883103102,0.3053250313,0.2690074742,0.3885180056,0.1593928039,0.1569662392,0.0389915444,0.1325366944,-0.3590078056,-0.1170711219,0.1671163887,0.115293242,-0.171261549,-0.1662117839,0.3117127419,0.0199156646,0.10100732,-0.2547494173,-0.0726020187,-0.2228352875,0.3414654732,0.3372592926,0.1324576437,0.3376508355,0.0862659737,0.2375318557,-0.0411623679,-0.4088019729,-0.0082033034,0.5613906384,0.3403669596,0.5017957091,-0.1312077641,-0.1257840246,-0.143103987,0.3379398584,-0.2317003608,-0.042705778,-0.3095024228,0.0799118504,-0.1820611507,-0.0606567971,-0.3343918025,-0.0018313729,0.0890034661,0.2038212121,-0.2043199539,-0.5852287412,0.4368210137,-0.1386185437,-0.4015676379,0.0768632218,0.0833692923,0.0400837772,-0.0228591729,-0.2593250275,0.1473932862,0.2686134577,-0.0831846893,-0.3365375698,0.0610045604,-0.028715238,0.2706465721,-0.1467019916,0.3606089652,-0.1064436436,-0.0384901576,-0.1178414598,-0.1704109609]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Could you try\r\n```python\r\nload_dataset('text', data_files='test.txt',cache_dir=\".\/\", split=\"train\")\r\n```\r\n?\r\n\r\n`load_dataset` returns a dictionary by default, like {\"train\": your_dataset}","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":18,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nCould you try\r\n```python\r\nload_dataset('text', data_files='test.txt',cache_dir=\".\/\", split=\"train\")\r\n```\r\n?\r\n\r\n`load_dataset` returns a dictionary by default, like {\"train\": your_dataset}","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Hi @lhoestq\r\nThanks for your suggestion.\r\n\r\nI tried \r\n```\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\", split=\"train\")\r\nprint(dataset)\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut it still doesn't work and got error:\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-7-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/arrow_dataset.py in __getitem__(self, key)\r\n 1069 format_columns=self._format_columns,\r\n 1070 output_all_columns=self._output_all_columns,\r\n-> 1071 format_kwargs=self._format_kwargs,\r\n 1072 )\r\n 1073 \r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/arrow_dataset.py in _getitem(self, key, format_type, format_columns, output_all_columns, format_kwargs)\r\n 1056 format_columns=format_columns,\r\n 1057 output_all_columns=output_all_columns,\r\n-> 1058 format_kwargs=format_kwargs,\r\n 1059 )\r\n 1060 return outputs\r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/arrow_dataset.py in _convert_outputs(self, outputs, format_type, format_columns, output_all_columns, format_kwargs)\r\n 872 continue\r\n 873 if format_columns is None or k in format_columns:\r\n--> 874 v = map_nested(command, v, **map_nested_kwargs)\r\n 875 output_dict[k] = v\r\n 876 return output_dict\r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 214 # Singleton\r\n 215 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 216 return function(data_struct)\r\n 217 \r\n 218 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/arrow_dataset.py in command(x)\r\n 833 if x.dtype == np.object: # pytorch tensors cannot be instantied from an array of objects\r\n 834 return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n--> 835 return torch.tensor(x, **format_kwargs)\r\n 836 \r\n 837 elif format_type == \"tensorflow\":\r\n\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nI found type can be ['numpy', 'torch', 'tensorflow', 'pandas'] only, how can I deal with the string type?","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":312,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nHi @lhoestq\r\nThanks for your suggestion.\r\n\r\nI tried \r\n```\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\", split=\"train\")\r\nprint(dataset)\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut it still doesn't work and got error:\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-7-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/arrow_dataset.py in __getitem__(self, key)\r\n 1069 format_columns=self._format_columns,\r\n 1070 output_all_columns=self._output_all_columns,\r\n-> 1071 format_kwargs=self._format_kwargs,\r\n 1072 )\r\n 1073 \r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/arrow_dataset.py in _getitem(self, key, format_type, format_columns, output_all_columns, format_kwargs)\r\n 1056 format_columns=format_columns,\r\n 1057 output_all_columns=output_all_columns,\r\n-> 1058 format_kwargs=format_kwargs,\r\n 1059 )\r\n 1060 return outputs\r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/arrow_dataset.py in _convert_outputs(self, outputs, format_type, format_columns, output_all_columns, format_kwargs)\r\n 872 continue\r\n 873 if format_columns is None or k in format_columns:\r\n--> 874 v = map_nested(command, v, **map_nested_kwargs)\r\n 875 output_dict[k] = v\r\n 876 return output_dict\r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types)\r\n 214 # Singleton\r\n 215 if not isinstance(data_struct, dict) and not isinstance(data_struct, types):\r\n--> 216 return function(data_struct)\r\n 217 \r\n 218 disable_tqdm = bool(logger.getEffectiveLevel() > INFO)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/datasets-0.4.0-py3.7.egg\/datasets\/arrow_dataset.py in command(x)\r\n 833 if x.dtype == np.object: # pytorch tensors cannot be instantied from an array of objects\r\n 834 return [map_nested(command, i, **map_nested_kwargs) for i in x]\r\n--> 835 return torch.tensor(x, **format_kwargs)\r\n 836 \r\n 837 elif format_type == \"tensorflow\":\r\n\r\nTypeError: new(): invalid data type 'str'\r\n```\r\n\r\nI found type can be ['numpy', 'torch', 'tensorflow', 'pandas'] only, how can I deal with the string type?","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"You need to tokenize the string inputs to convert them in integers before you can feed them to a pytorch dataloader.\r\n\r\nYou can read the quicktour of the datasets or the transformers libraries to know more about that:\r\n- transformers: https:\/\/huggingface.co\/transformers\/quicktour.html\r\n- dataset: https:\/\/huggingface.co\/docs\/datasets\/quicktour.html","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":44,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nYou need to tokenize the string inputs to convert them in integers before you can feed them to a pytorch dataloader.\r\n\r\nYou can read the quicktour of the datasets or the transformers libraries to know more about that:\r\n- transformers: https:\/\/huggingface.co\/transformers\/quicktour.html\r\n- dataset: https:\/\/huggingface.co\/docs\/datasets\/quicktour.html","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Hey @chiyuzhang94, I was also having trouble in loading a large text file (11GB).\r\nBut finally got it working. This is what I did after looking into the documentation.\r\n\r\n1. split the whole dataset file into smaller files\r\n```bash\r\nmkdir .\/shards\r\nsplit -a 4 -l 256000 -d full_raw_corpus.txt .\/shards\/shard_\r\n````\r\n2. Pass paths of small data files to `load_dataset`\r\n```python\r\nfiles = glob.glob('shards\/*')\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files=files, split='train')\r\n```\r\n(On passing the whole dataset file (11GB) directly to `load_dataset` was resulting into RAM issue)\r\n\r\n3. Tokenization\r\n```python\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\ndataset = dataset.map(encode, batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n```\r\n Now you can pass `dataset` to `Trainer` or `pytorch DataLoader`\r\n```python\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\nnext(iter(dataloader))\r\n```\r\nHope this helps\r\n","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":125,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nHey @chiyuzhang94, I was also having trouble in loading a large text file (11GB).\r\nBut finally got it working. This is what I did after looking into the documentation.\r\n\r\n1. split the whole dataset file into smaller files\r\n```bash\r\nmkdir .\/shards\r\nsplit -a 4 -l 256000 -d full_raw_corpus.txt .\/shards\/shard_\r\n````\r\n2. Pass paths of small data files to `load_dataset`\r\n```python\r\nfiles = glob.glob('shards\/*')\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files=files, split='train')\r\n```\r\n(On passing the whole dataset file (11GB) directly to `load_dataset` was resulting into RAM issue)\r\n\r\n3. Tokenization\r\n```python\r\ndef encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\ndataset = dataset.map(encode, batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n```\r\n Now you can pass `dataset` to `Trainer` or `pytorch DataLoader`\r\n```python\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\nnext(iter(dataloader))\r\n```\r\nHope this helps\r\n","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Thanks, @thomwolf and @sipah00 ,\r\n\r\nI tried to implement your suggestions in my scripts. \r\nNow, I am facing some connection time-out error. I am using my local file, I have no idea why the module request s3 database.\r\n\r\nThe log is:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 449, in send\r\n raise err\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 74, in create_connection\r\n timeout=timeout\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 720, in urlopen\r\n sock.connect(sa)\r\nTimeoutError: [Errno 110] Connection timed out\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 672, in urlopen\r\n method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py\", line 436, in increment\r\n chunked=chunked,\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 376, in _make_request\r\n raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\nurllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/text\/text.py (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection obj\r\nect at 0x7fff401e0e48>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\nTraceback (most recent call last):\r\n File \"\/scratch\/roberta_emohash\/run_language_modeling.py\", line 1019, in <module>\r\n main()\r\n File \"\/scratch\/roberta_emohash\/run_language_modeling.py\", line 962, in main\r\n train_dataset = load_and_cache_examples(args, tokenizer, evaluate=False)\r\n File \"\/scratch\/roberta_emohash\/run_language_modeling.py\", line 177, in load_and_cache_examples\r\n return HG_Datasets(tokenizer, file_path, args)\r\n File \"\/scratch\/roberta_emohash\/run_language_modeling.py\", line 117, in HG_Datasets\r\n dataset = load_dataset('text', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")\r\n File \"\/arc\/project\/evn_py36\/datasets\/datasets\/src\/datasets\/load.py\", line 590, in load_dataset\r\n self._validate_conn(conn)\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 994, in _validate_conn\r\n conn.connect()\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 300, in connect\r\n conn = self._new_conn()\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 169, in _new_conn\r\n self, \"Failed to establish a new connection: %s\" % e\r\nurllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x7fff401e0da0>: Failed to establish a new connection: [Errno 110] Connection timed out\r\n\r\n``` \r\n\r\nDo you have any experience on this issue?","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":248,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nThanks, @thomwolf and @sipah00 ,\r\n\r\nI tried to implement your suggestions in my scripts. \r\nNow, I am facing some connection time-out error. I am using my local file, I have no idea why the module request s3 database.\r\n\r\nThe log is:\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/requests\/adapters.py\", line 449, in send\r\n raise err\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/util\/connection.py\", line 74, in create_connection\r\n timeout=timeout\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 720, in urlopen\r\n sock.connect(sa)\r\nTimeoutError: [Errno 110] Connection timed out\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 672, in urlopen\r\n method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/util\/retry.py\", line 436, in increment\r\n chunked=chunked,\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 376, in _make_request\r\n raise MaxRetryError(_pool, url, error or ResponseError(cause))\r\nurllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: \/datasets.huggingface.co\/datasets\/datasets\/text\/text.py (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection obj\r\nect at 0x7fff401e0e48>: Failed to establish a new connection: [Errno 110] Connection timed out',))\r\n\r\nTraceback (most recent call last):\r\n File \"\/scratch\/roberta_emohash\/run_language_modeling.py\", line 1019, in <module>\r\n main()\r\n File \"\/scratch\/roberta_emohash\/run_language_modeling.py\", line 962, in main\r\n train_dataset = load_and_cache_examples(args, tokenizer, evaluate=False)\r\n File \"\/scratch\/roberta_emohash\/run_language_modeling.py\", line 177, in load_and_cache_examples\r\n return HG_Datasets(tokenizer, file_path, args)\r\n File \"\/scratch\/roberta_emohash\/run_language_modeling.py\", line 117, in HG_Datasets\r\n dataset = load_dataset('text', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")\r\n File \"\/arc\/project\/evn_py36\/datasets\/datasets\/src\/datasets\/load.py\", line 590, in load_dataset\r\n self._validate_conn(conn)\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connectionpool.py\", line 994, in _validate_conn\r\n conn.connect()\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 300, in connect\r\n conn = self._new_conn()\r\n File \"\/home\/.local\/lib\/python3.6\/site-packages\/urllib3\/connection.py\", line 169, in _new_conn\r\n self, \"Failed to establish a new connection: %s\" % e\r\nurllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x7fff401e0da0>: Failed to establish a new connection: [Errno 110] Connection timed out\r\n\r\n``` \r\n\r\nDo you have any experience on this issue?","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"No, I didn't encounter this problem, it seems to me a network problem","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":13,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nNo, I didn't encounter this problem, it seems to me a network problem","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"I noticed this is because I use a cloud server where does not provide for connections from our standard compute nodes to outside resources. \r\n\r\nFor the `datasets` package, it seems that if the loading script is not already cached in the library it will attempt to connect to an AWS resource to download the dataset loading script. \r\n\r\nI am wondering why the package works in this way. Do you have any suggestions to solve this issue? ","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":76,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nI noticed this is because I use a cloud server where does not provide for connections from our standard compute nodes to outside resources. \r\n\r\nFor the `datasets` package, it seems that if the loading script is not already cached in the library it will attempt to connect to an AWS resource to download the dataset loading script. \r\n\r\nI am wondering why the package works in this way. Do you have any suggestions to solve this issue? ","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"I solved the above issue by downloading text.py manually and passing the path to the `load_dataset` function. \r\n\r\nNow, I have a new issue with the Read-only file system.\r\n\r\nThe error is: \r\n```\r\nI0916 22:14:38.453380 140737353971520 filelock.py:274] Lock 140734268996072 acquired on \/scratch\/chiyuzh\/roberta\/text.py.lock\r\nFound main folder for dataset \/scratch\/chiyuzh\/roberta\/text.py at \/home\/chiyuzh\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nCreating specific version folder for dataset \/scratch\/chiyuzh\/roberta\/text.py at \/home\/chiyuzh\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7fcc649178b014\r\nI0916 22:14:38.530371 140737353971520 filelock.py:318] Lock 140734268996072 released on \/scratch\/chiyuzh\/roberta\/text.py.lock\r\nTraceback (most recent call last):\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling_hg.py\", line 1019, in <module>\r\n main()\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling_hg.py\", line 962, in main\r\n train_dataset = load_and_cache_examples(args, tokenizer, evaluate=False)\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling_hg.py\", line 177, in load_and_cache_examples\r\n return HG_Datasets(tokenizer, file_path, args)\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling_hg.py\", line 117, in HG_Datasets\r\n dataset = load_dataset('\/scratch\/chiyuzh\/roberta\/text.py', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")\r\n File \"\/arc\/project\/chiyuzh\/evn_py36\/datasets\/src\/datasets\/load.py\", line 590, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/arc\/project\/chiyuzh\/evn_py36\/datasets\/src\/datasets\/load.py\", line 385, in prepare_module\r\n os.makedirs(hash_folder_path)\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/home\/chiyuzh\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7fcc649178b014'\r\n\r\n```\r\n\r\nI installed datasets at \/project\/chiyuzh\/evn_py36\/datasets\/src where is a writable directory.\r\nI also tried change the environment variables to the writable directory:\r\n`export HF_MODULES_PATH=\/project\/chiyuzh\/evn_py36\/datasets\/cache_dir\/`\r\n`export HF_DATASETS_CACHE=\/project\/chiyuzh\/evn_py36\/datasets\/cache_dir\/`\r\n \r\nIn my scripts, I also changed to:\r\n`dataset = load_dataset('\/scratch\/chiyuzh\/roberta\/text.py', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")`\r\n`data_cache_dir = $TMPDIR\/data\/` that also a writable directory.\r\n \r\nBut it still try to make directory at \/home\/chiyuzh\/.cache\/huggingface\/modules\/.\r\nDo you have any idea about this issue? @thomwolf \r\n","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":214,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nI solved the above issue by downloading text.py manually and passing the path to the `load_dataset` function. \r\n\r\nNow, I have a new issue with the Read-only file system.\r\n\r\nThe error is: \r\n```\r\nI0916 22:14:38.453380 140737353971520 filelock.py:274] Lock 140734268996072 acquired on \/scratch\/chiyuzh\/roberta\/text.py.lock\r\nFound main folder for dataset \/scratch\/chiyuzh\/roberta\/text.py at \/home\/chiyuzh\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\r\nCreating specific version folder for dataset \/scratch\/chiyuzh\/roberta\/text.py at \/home\/chiyuzh\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7fcc649178b014\r\nI0916 22:14:38.530371 140737353971520 filelock.py:318] Lock 140734268996072 released on \/scratch\/chiyuzh\/roberta\/text.py.lock\r\nTraceback (most recent call last):\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling_hg.py\", line 1019, in <module>\r\n main()\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling_hg.py\", line 962, in main\r\n train_dataset = load_and_cache_examples(args, tokenizer, evaluate=False)\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling_hg.py\", line 177, in load_and_cache_examples\r\n return HG_Datasets(tokenizer, file_path, args)\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling_hg.py\", line 117, in HG_Datasets\r\n dataset = load_dataset('\/scratch\/chiyuzh\/roberta\/text.py', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")\r\n File \"\/arc\/project\/chiyuzh\/evn_py36\/datasets\/src\/datasets\/load.py\", line 590, in load_dataset\r\n path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True\r\n File \"\/arc\/project\/chiyuzh\/evn_py36\/datasets\/src\/datasets\/load.py\", line 385, in prepare_module\r\n os.makedirs(hash_folder_path)\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/home\/chiyuzh\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/text\/512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7fcc649178b014'\r\n\r\n```\r\n\r\nI installed datasets at \/project\/chiyuzh\/evn_py36\/datasets\/src where is a writable directory.\r\nI also tried change the environment variables to the writable directory:\r\n`export HF_MODULES_PATH=\/project\/chiyuzh\/evn_py36\/datasets\/cache_dir\/`\r\n`export HF_DATASETS_CACHE=\/project\/chiyuzh\/evn_py36\/datasets\/cache_dir\/`\r\n \r\nIn my scripts, I also changed to:\r\n`dataset = load_dataset('\/scratch\/chiyuzh\/roberta\/text.py', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")`\r\n`data_cache_dir = $TMPDIR\/data\/` that also a writable directory.\r\n \r\nBut it still try to make directory at \/home\/chiyuzh\/.cache\/huggingface\/modules\/.\r\nDo you have any idea about this issue? @thomwolf \r\n","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> Hey @chiyuzhang94, I was also having trouble in loading a large text file (11GB).\r\n> But finally got it working. This is what I did after looking into the documentation.\r\n> \r\n> 1. split the whole dataset file into smaller files\r\n> \r\n> ```shell\r\n> mkdir .\/shards\r\n> split -a 4 -l 256000 -d full_raw_corpus.txt .\/shards\/shard_\r\n> ```\r\n> \r\n> 1. Pass paths of small data files to `load_dataset`\r\n> \r\n> ```python\r\n> files = glob.glob('shards\/*')\r\n> from datasets import load_dataset\r\n> dataset = load_dataset('text', data_files=files, split='train')\r\n> ```\r\n> \r\n> (On passing the whole dataset file (11GB) directly to `load_dataset` was resulting into RAM issue)\r\n> \r\n> 1. Tokenization\r\n> \r\n> ```python\r\n> def encode(examples):\r\n> return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> dataset = dataset.map(encode, batched=True)\r\n> dataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n> ```\r\n> \r\n> Now you can pass `dataset` to `Trainer` or `pytorch DataLoader`\r\n> \r\n> ```python\r\n> dataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\n> next(iter(dataloader))\r\n> ```\r\n> \r\n> Hope this helps\r\n\r\nWhen I run 'dataset = dataset.map(encode, batched=True)',\r\nI encountered a problem like this:\r\n\r\n> Testing the mapped function outputs\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 300, in map\r\n for k, dataset in self.items()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 300, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1224, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1195, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"<stdin>\", line 3, in encode\r\nTypeError: __init__() takes 1 positional argument but 2 were given","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":254,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> Hey @chiyuzhang94, I was also having trouble in loading a large text file (11GB).\r\n> But finally got it working. This is what I did after looking into the documentation.\r\n> \r\n> 1. split the whole dataset file into smaller files\r\n> \r\n> ```shell\r\n> mkdir .\/shards\r\n> split -a 4 -l 256000 -d full_raw_corpus.txt .\/shards\/shard_\r\n> ```\r\n> \r\n> 1. Pass paths of small data files to `load_dataset`\r\n> \r\n> ```python\r\n> files = glob.glob('shards\/*')\r\n> from datasets import load_dataset\r\n> dataset = load_dataset('text', data_files=files, split='train')\r\n> ```\r\n> \r\n> (On passing the whole dataset file (11GB) directly to `load_dataset` was resulting into RAM issue)\r\n> \r\n> 1. Tokenization\r\n> \r\n> ```python\r\n> def encode(examples):\r\n> return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> dataset = dataset.map(encode, batched=True)\r\n> dataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n> ```\r\n> \r\n> Now you can pass `dataset` to `Trainer` or `pytorch DataLoader`\r\n> \r\n> ```python\r\n> dataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\n> next(iter(dataloader))\r\n> ```\r\n> \r\n> Hope this helps\r\n\r\nWhen I run 'dataset = dataset.map(encode, batched=True)',\r\nI encountered a problem like this:\r\n\r\n> Testing the mapped function outputs\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 300, in map\r\n for k, dataset in self.items()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 300, in <dictcomp>\r\n for k, dataset in self.items()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1224, in map\r\n update_data = does_function_return_dict(test_inputs, test_indices)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1195, in does_function_return_dict\r\n function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n File \"<stdin>\", line 3, in encode\r\nTypeError: __init__() takes 1 positional argument but 2 were given","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> > Hey @chiyuzhang94, I was also having trouble in loading a large text file (11GB).\r\n> > But finally got it working. This is what I did after looking into the documentation.\r\n> > \r\n> > 1. split the whole dataset file into smaller files\r\n> > \r\n> > ```shell\r\n> > mkdir .\/shards\r\n> > split -a 4 -l 256000 -d full_raw_corpus.txt .\/shards\/shard_\r\n> > ```\r\n> > \r\n> > \r\n> > \r\n> > 1. Pass paths of small data files to `load_dataset`\r\n> > \r\n> > ```python\r\n> > files = glob.glob('shards\/*')\r\n> > from datasets import load_dataset\r\n> > dataset = load_dataset('text', data_files=files, split='train')\r\n> > ```\r\n> > \r\n> > \r\n> > (On passing the whole dataset file (11GB) directly to `load_dataset` was resulting into RAM issue)\r\n> > \r\n> > 1. Tokenization\r\n> > \r\n> > ```python\r\n> > def encode(examples):\r\n> > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > dataset = dataset.map(encode, batched=True)\r\n> > dataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n> > ```\r\n> > \r\n> > \r\n> > Now you can pass `dataset` to `Trainer` or `pytorch DataLoader`\r\n> > ```python\r\n> > dataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\n> > next(iter(dataloader))\r\n> > ```\r\n> > \r\n> > \r\n> > Hope this helps\r\n> \r\n> When I run 'dataset = dataset.map(encode, batched=True)',\r\n> I encountered a problem like this:\r\n> \r\n> > Testing the mapped function outputs\r\n> > Traceback (most recent call last):\r\n> > File \"\", line 1, in \r\n> > File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 300, in map\r\n> > for k, dataset in self.items()\r\n> > File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 300, in \r\n> > for k, dataset in self.items()\r\n> > File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1224, in map\r\n> > update_data = does_function_return_dict(test_inputs, test_indices)\r\n> > File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1195, in does_function_return_dict\r\n> > function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n> > File \"\", line 3, in encode\r\n> > TypeError: **init**() takes 1 positional argument but 2 were given\r\n\r\nWhat is your encoder function?","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":331,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> > Hey @chiyuzhang94, I was also having trouble in loading a large text file (11GB).\r\n> > But finally got it working. This is what I did after looking into the documentation.\r\n> > \r\n> > 1. split the whole dataset file into smaller files\r\n> > \r\n> > ```shell\r\n> > mkdir .\/shards\r\n> > split -a 4 -l 256000 -d full_raw_corpus.txt .\/shards\/shard_\r\n> > ```\r\n> > \r\n> > \r\n> > \r\n> > 1. Pass paths of small data files to `load_dataset`\r\n> > \r\n> > ```python\r\n> > files = glob.glob('shards\/*')\r\n> > from datasets import load_dataset\r\n> > dataset = load_dataset('text', data_files=files, split='train')\r\n> > ```\r\n> > \r\n> > \r\n> > (On passing the whole dataset file (11GB) directly to `load_dataset` was resulting into RAM issue)\r\n> > \r\n> > 1. Tokenization\r\n> > \r\n> > ```python\r\n> > def encode(examples):\r\n> > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > dataset = dataset.map(encode, batched=True)\r\n> > dataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n> > ```\r\n> > \r\n> > \r\n> > Now you can pass `dataset` to `Trainer` or `pytorch DataLoader`\r\n> > ```python\r\n> > dataloader = torch.utils.data.DataLoader(dataset, batch_size=4)\r\n> > next(iter(dataloader))\r\n> > ```\r\n> > \r\n> > \r\n> > Hope this helps\r\n> \r\n> When I run 'dataset = dataset.map(encode, batched=True)',\r\n> I encountered a problem like this:\r\n> \r\n> > Testing the mapped function outputs\r\n> > Traceback (most recent call last):\r\n> > File \"\", line 1, in \r\n> > File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 300, in map\r\n> > for k, dataset in self.items()\r\n> > File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/dataset_dict.py\", line 300, in \r\n> > for k, dataset in self.items()\r\n> > File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1224, in map\r\n> > update_data = does_function_return_dict(test_inputs, test_indices)\r\n> > File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1195, in does_function_return_dict\r\n> > function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)\r\n> > File \"\", line 3, in encode\r\n> > TypeError: **init**() takes 1 positional argument but 2 were given\r\n\r\nWhat is your encoder function?","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> ```python\r\n> def encode(examples):\r\n> return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> ```\r\n\r\nIt is the same as suggested:\r\n\r\n> def encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":25,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> ```python\r\n> def encode(examples):\r\n> return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> ```\r\n\r\nIt is the same as suggested:\r\n\r\n> def encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> > ```python\r\n> > def encode(examples):\r\n> > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > ```\r\n> \r\n> It is the same as suggested:\r\n> \r\n> > def encode(examples):\r\n> > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\nDo you use this function in a `class` object? \r\n\r\ninit() takes 1 positional argument but 2 were given. I guess the additional argument is self?","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":60,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> > ```python\r\n> > def encode(examples):\r\n> > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > ```\r\n> \r\n> It is the same as suggested:\r\n> \r\n> > def encode(examples):\r\n> > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\nDo you use this function in a `class` object? \r\n\r\ninit() takes 1 positional argument but 2 were given. I guess the additional argument is self?","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> > > ```python\r\n> > > def encode(examples):\r\n> > > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > > ```\r\n> > \r\n> > \r\n> > It is the same as suggested:\r\n> > > def encode(examples):\r\n> > > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> \r\n> Do you use this function in a `class` object?\r\n> \r\n> init() takes 1 positional argument but 2 were given. I guess the additional argument is self?\r\n\r\nThanks for your reply.\r\nCould you provide some simple example here?\r\nCurrently, I do not use this function in a class object. \r\nI think you are right and I was wondering how to construct this class.\r\nI try to modify it based on transformers' LineByLineTextDataset. Am I correct?\r\n\r\n> class LineByLineTextDataset(Dataset):\r\n \"\"\"\r\n This will be superseded by a framework-agnostic approach\r\n soon.\r\n \"\"\"\r\n\r\n def __init__(self, tokenizer: PreTrainedTokenizer, file_path: str, block_size: int):\r\n assert os.path.isfile(file_path), f\"Input file path {file_path} not found\"\r\n # Here, we do not cache the features, operating under the assumption\r\n # that we will soon use fast multithreaded tokenizers from the\r\n # `tokenizers` repo everywhere =)\r\n #logger.info(\"Creating features from dataset file at %s\", file_path)\r\n #with open(file_path, encoding=\"utf-8\") as f:\r\n # lines = [line for line in f.read().splitlines() if (len(line) > 0 and not line.isspace())]\r\n #batch_encoding = tokenizer(lines, add_special_tokens=True, truncation=True, max_length=block_size)\r\n\r\n\timport glob\r\n\tfiles = glob.glob('\/home\/mtzhang111\/fairseq\/cs_doc\/shards\/shard_003*')\r\n\tfrom datasets import load_dataset\r\n\tdataset = load_dataset('text', data_files=files)\r\n batch_encoding= dataset.map(encode, batched=True)\r\n self.examples = batch_encoding[\"input_ids\"]\r\n\t\r\n\r\n def encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\n def __len__(self):\r\n return len(self.examples)\r\n\r\n def __getitem__(self, i) -> torch.Tensor:\r\n return torch.tensor(self.examples[i], dtype=torch.long)\r\n","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":250,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> > > ```python\r\n> > > def encode(examples):\r\n> > > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > > ```\r\n> > \r\n> > \r\n> > It is the same as suggested:\r\n> > > def encode(examples):\r\n> > > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> \r\n> Do you use this function in a `class` object?\r\n> \r\n> init() takes 1 positional argument but 2 were given. I guess the additional argument is self?\r\n\r\nThanks for your reply.\r\nCould you provide some simple example here?\r\nCurrently, I do not use this function in a class object. \r\nI think you are right and I was wondering how to construct this class.\r\nI try to modify it based on transformers' LineByLineTextDataset. Am I correct?\r\n\r\n> class LineByLineTextDataset(Dataset):\r\n \"\"\"\r\n This will be superseded by a framework-agnostic approach\r\n soon.\r\n \"\"\"\r\n\r\n def __init__(self, tokenizer: PreTrainedTokenizer, file_path: str, block_size: int):\r\n assert os.path.isfile(file_path), f\"Input file path {file_path} not found\"\r\n # Here, we do not cache the features, operating under the assumption\r\n # that we will soon use fast multithreaded tokenizers from the\r\n # `tokenizers` repo everywhere =)\r\n #logger.info(\"Creating features from dataset file at %s\", file_path)\r\n #with open(file_path, encoding=\"utf-8\") as f:\r\n # lines = [line for line in f.read().splitlines() if (len(line) > 0 and not line.isspace())]\r\n #batch_encoding = tokenizer(lines, add_special_tokens=True, truncation=True, max_length=block_size)\r\n\r\n\timport glob\r\n\tfiles = glob.glob('\/home\/mtzhang111\/fairseq\/cs_doc\/shards\/shard_003*')\r\n\tfrom datasets import load_dataset\r\n\tdataset = load_dataset('text', data_files=files)\r\n batch_encoding= dataset.map(encode, batched=True)\r\n self.examples = batch_encoding[\"input_ids\"]\r\n\t\r\n\r\n def encode(examples):\r\n return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n\r\n def __len__(self):\r\n return len(self.examples)\r\n\r\n def __getitem__(self, i) -> torch.Tensor:\r\n return torch.tensor(self.examples[i], dtype=torch.long)\r\n","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> > > > ```python\r\n> > > > def encode(examples):\r\n> > > > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > > > ```\r\n> > > \r\n> > > \r\n> > > It is the same as suggested:\r\n> > > > def encode(examples):\r\n> > > > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > \r\n> > \r\n> > Do you use this function in a `class` object?\r\n> > init() takes 1 positional argument but 2 were given. I guess the additional argument is self?\r\n> \r\n> Thanks for your reply.\r\n> Could you provide some simple example here?\r\n> Currently, I do not use this function in a class object.\r\n> I think you are right and I was wondering how to construct this class.\r\n> I try to modify it based on transformers' LineByLineTextDataset. Am I correct?\r\n> \r\n> > class LineByLineTextDataset(Dataset):\r\n> > \"\"\"\r\n> > This will be superseded by a framework-agnostic approach\r\n> > soon.\r\n> > \"\"\"\r\n> \r\n> ```\r\n> def __init__(self, tokenizer: PreTrainedTokenizer, file_path: str, block_size: int):\r\n> assert os.path.isfile(file_path), f\"Input file path {file_path} not found\"\r\n> # Here, we do not cache the features, operating under the assumption\r\n> # that we will soon use fast multithreaded tokenizers from the\r\n> # `tokenizers` repo everywhere =)\r\n> #logger.info(\"Creating features from dataset file at %s\", file_path)\r\n> #with open(file_path, encoding=\"utf-8\") as f:\r\n> # lines = [line for line in f.read().splitlines() if (len(line) > 0 and not line.isspace())]\r\n> #batch_encoding = tokenizer(lines, add_special_tokens=True, truncation=True, max_length=block_size)\r\n> \r\n> import glob\r\n> files = glob.glob('\/home\/mtzhang111\/fairseq\/cs_doc\/shards\/shard_003*')\r\n> from datasets import load_dataset\r\n> dataset = load_dataset('text', data_files=files)\r\n> batch_encoding= dataset.map(encode, batched=True)\r\n> self.examples = batch_encoding[\"input_ids\"]\r\n> \r\n> \r\n> def encode(examples):\r\n> return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> \r\n> def __len__(self):\r\n> return len(self.examples)\r\n> \r\n> def __getitem__(self, i) -> torch.Tensor:\r\n> return torch.tensor(self.examples[i], dtype=torch.long)\r\n> ```\r\n\r\nI am also struggling with this adaptation. \r\nI am not sure whether I am right.\r\n\r\nI think you don't need to construct `class LazyLineByLineTextDataset(Dataset)` at all. \r\ntorch.utils.data.Dataset is a generator. \r\n\r\nNow, we use `dataset = dataset.map(encode, batched=True)` as a generator. So we just pass dataset to torch.utils.data.DataLoader. ","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":357,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> > > > ```python\r\n> > > > def encode(examples):\r\n> > > > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > > > ```\r\n> > > \r\n> > > \r\n> > > It is the same as suggested:\r\n> > > > def encode(examples):\r\n> > > > return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> > \r\n> > \r\n> > Do you use this function in a `class` object?\r\n> > init() takes 1 positional argument but 2 were given. I guess the additional argument is self?\r\n> \r\n> Thanks for your reply.\r\n> Could you provide some simple example here?\r\n> Currently, I do not use this function in a class object.\r\n> I think you are right and I was wondering how to construct this class.\r\n> I try to modify it based on transformers' LineByLineTextDataset. Am I correct?\r\n> \r\n> > class LineByLineTextDataset(Dataset):\r\n> > \"\"\"\r\n> > This will be superseded by a framework-agnostic approach\r\n> > soon.\r\n> > \"\"\"\r\n> \r\n> ```\r\n> def __init__(self, tokenizer: PreTrainedTokenizer, file_path: str, block_size: int):\r\n> assert os.path.isfile(file_path), f\"Input file path {file_path} not found\"\r\n> # Here, we do not cache the features, operating under the assumption\r\n> # that we will soon use fast multithreaded tokenizers from the\r\n> # `tokenizers` repo everywhere =)\r\n> #logger.info(\"Creating features from dataset file at %s\", file_path)\r\n> #with open(file_path, encoding=\"utf-8\") as f:\r\n> # lines = [line for line in f.read().splitlines() if (len(line) > 0 and not line.isspace())]\r\n> #batch_encoding = tokenizer(lines, add_special_tokens=True, truncation=True, max_length=block_size)\r\n> \r\n> import glob\r\n> files = glob.glob('\/home\/mtzhang111\/fairseq\/cs_doc\/shards\/shard_003*')\r\n> from datasets import load_dataset\r\n> dataset = load_dataset('text', data_files=files)\r\n> batch_encoding= dataset.map(encode, batched=True)\r\n> self.examples = batch_encoding[\"input_ids\"]\r\n> \r\n> \r\n> def encode(examples):\r\n> return tokenizer(examples['text'], truncation=True, padding='max_length')\r\n> \r\n> def __len__(self):\r\n> return len(self.examples)\r\n> \r\n> def __getitem__(self, i) -> torch.Tensor:\r\n> return torch.tensor(self.examples[i], dtype=torch.long)\r\n> ```\r\n\r\nI am also struggling with this adaptation. \r\nI am not sure whether I am right.\r\n\r\nI think you don't need to construct `class LazyLineByLineTextDataset(Dataset)` at all. \r\ntorch.utils.data.Dataset is a generator. \r\n\r\nNow, we use `dataset = dataset.map(encode, batched=True)` as a generator. So we just pass dataset to torch.utils.data.DataLoader. ","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"@chiyuzhang94 Thanks for your reply. After some changes, currently, I managed to make the data loading process running.\r\nI published it in case you might want to take a look. Thanks for your help!\r\nhttps:\/\/github.com\/shizhediao\/Transformers_TPU","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":35,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n@chiyuzhang94 Thanks for your reply. After some changes, currently, I managed to make the data loading process running.\r\nI published it in case you might want to take a look. Thanks for your help!\r\nhttps:\/\/github.com\/shizhediao\/Transformers_TPU","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Hi @shizhediao ,\r\n\r\nThanks! It looks great!\r\n\r\nBut my problem still is the cache directory is a read-only file system. \r\n[As I mentioned](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-693912285), I tried to change the cache directory but it didn't work. \r\n\r\nDo you have any suggestions?\r\n\r\n","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":39,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nHi @shizhediao ,\r\n\r\nThanks! It looks great!\r\n\r\nBut my problem still is the cache directory is a read-only file system. \r\n[As I mentioned](https:\/\/github.com\/huggingface\/datasets\/issues\/610#issuecomment-693912285), I tried to change the cache directory but it didn't work. \r\n\r\nDo you have any suggestions?\r\n\r\n","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> I installed datasets at \/project\/chiyuzh\/evn_py36\/datasets\/src where is a writable directory.\r\n> I also tried change the environment variables to the writable directory:\r\n> `export HF_MODULES_PATH=\/project\/chiyuzh\/evn_py36\/datasets\/cache_dir\/`\r\n\r\nI think it is `HF_MODULES_CACHE` and not `HF_MODULES_PATH` @chiyuzhang94 .\r\nCould you try again and let me know if it fixes your issue ?\r\n","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":50,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> I installed datasets at \/project\/chiyuzh\/evn_py36\/datasets\/src where is a writable directory.\r\n> I also tried change the environment variables to the writable directory:\r\n> `export HF_MODULES_PATH=\/project\/chiyuzh\/evn_py36\/datasets\/cache_dir\/`\r\n\r\nI think it is `HF_MODULES_CACHE` and not `HF_MODULES_PATH` @chiyuzhang94 .\r\nCould you try again and let me know if it fixes your issue ?\r\n","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"We should probably add a section in the doc on the caching system with the env variables in particular.","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":19,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nWe should probably add a section in the doc on the caching system with the env variables in particular.","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Hi @thomwolf , @lhoestq ,\r\n\r\nThanks for your suggestions. With the latest version of this package, I can load text data without Internet. \r\n\r\nBut I found the speed of dataset loading is very slow. \r\n\r\nMy scrips like this: \r\n```\r\n def token_encode(examples):\r\n tokenizer_out = tokenizer(examples['text'], truncation=True, padding=\"max_length\", add_special_tokens=True, max_length=args.block_size)\r\n return tokenizer_out\r\n \r\n path = Path(file_path)\r\n files = sorted(path.glob('*'))\r\n dataset = load_dataset('.\/text.py', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")\r\n dataset = dataset.map(token_encode, batched=True)\r\n\r\n dataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n```\r\n\r\nI have 1,123,870,657 lines in my input directory. \r\nI can find the processing speed as following. It is very slow. \r\n```\r\n| 13\/1123871 [00:02<62:37:39, 4.98ba\/s]^M 0%| \r\n| 14\/1123871 [00:03<61:27:31, 5.08ba\/s]^M 0%| \r\n| 15\/1123871 [00:03<66:34:19, 4.69ba\/s]^M 0%| \r\n| 16\/1123871 [00:03<68:25:01, 4.56ba\/s]^M 0%| \r\n| 17\/1123871 [00:03<72:00:03, 4.34ba\/s]^M 0%| \r\n```\r\nDo you have any suggestions to accelerate this loading process?","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":129,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nHi @thomwolf , @lhoestq ,\r\n\r\nThanks for your suggestions. With the latest version of this package, I can load text data without Internet. \r\n\r\nBut I found the speed of dataset loading is very slow. \r\n\r\nMy scrips like this: \r\n```\r\n def token_encode(examples):\r\n tokenizer_out = tokenizer(examples['text'], truncation=True, padding=\"max_length\", add_special_tokens=True, max_length=args.block_size)\r\n return tokenizer_out\r\n \r\n path = Path(file_path)\r\n files = sorted(path.glob('*'))\r\n dataset = load_dataset('.\/text.py', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")\r\n dataset = dataset.map(token_encode, batched=True)\r\n\r\n dataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n```\r\n\r\nI have 1,123,870,657 lines in my input directory. \r\nI can find the processing speed as following. It is very slow. \r\n```\r\n| 13\/1123871 [00:02<62:37:39, 4.98ba\/s]^M 0%| \r\n| 14\/1123871 [00:03<61:27:31, 5.08ba\/s]^M 0%| \r\n| 15\/1123871 [00:03<66:34:19, 4.69ba\/s]^M 0%| \r\n| 16\/1123871 [00:03<68:25:01, 4.56ba\/s]^M 0%| \r\n| 17\/1123871 [00:03<72:00:03, 4.34ba\/s]^M 0%| \r\n```\r\nDo you have any suggestions to accelerate this loading process?","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"You can use multiprocessing by specifying `num_proc=` in `.map()`\r\n\r\nAlso it looks like you have `1123871` batches of 1000 elements (default batch size), i.e. 1,123,871,000 lines in total.\r\nAm I right ?","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":32,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nYou can use multiprocessing by specifying `num_proc=` in `.map()`\r\n\r\nAlso it looks like you have `1123871` batches of 1000 elements (default batch size), i.e. 1,123,871,000 lines in total.\r\nAm I right ?","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> You can use multiprocessing by specifying `num_proc=` in `.map()`\r\n> \r\n> Also it looks like you have `1123871` batches of 1000 elements (default batch size), i.e. 1,123,871,000 lines in total.\r\n> Am I right ?\r\n\r\nHi @lhoestq ,\r\n\r\nThanks. I will try it.\r\n\r\nYou are right. I have 1,123,870,657 lines totally in the path. I split the large file into 440 small files. Each file has 2,560,000 lines.\r\n\r\nI have another question. Because I am using a cloud server where only allows running a job up to 7 days. Hence, I need to resume my model every week. If the script needs to load and process the dataset every time. It is very low efficient based on the current processing speed. Is it possible that I process the dataset once and use the process cache to in the future work? \r\n","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":141,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> You can use multiprocessing by specifying `num_proc=` in `.map()`\r\n> \r\n> Also it looks like you have `1123871` batches of 1000 elements (default batch size), i.e. 1,123,871,000 lines in total.\r\n> Am I right ?\r\n\r\nHi @lhoestq ,\r\n\r\nThanks. I will try it.\r\n\r\nYou are right. I have 1,123,870,657 lines totally in the path. I split the large file into 440 small files. Each file has 2,560,000 lines.\r\n\r\nI have another question. Because I am using a cloud server where only allows running a job up to 7 days. Hence, I need to resume my model every week. If the script needs to load and process the dataset every time. It is very low efficient based on the current processing speed. Is it possible that I process the dataset once and use the process cache to in the future work? \r\n","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Hi @lhoestq ,\r\n\r\nI tried to use multi-processor, but I got errors as follow: \r\nBecause I am using python distributed training, it seems some conflicts with the distributed job. \r\n\r\nDo you have any suggestions?\r\n```\r\nI0925 10:19:35.603023 140737353971520 filelock.py:318] Lock 140737229443368 released on \/tmp\/pbs.1120510.pbsha.ib.sockeye\/cache\/_tmp_pbs.1120510.pbsha.ib.sockeye_cache_text_default-7fb934ed6fac5d01_0.0.0_512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7\r\nfcc649178b014.lock\r\nTraceback (most recent call last):\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling.py\", line 1024, in <module>\r\n main()\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling.py\", line 967, in main\r\n train_dataset = load_and_cache_examples(args, tokenizer, evaluate=False)\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling.py\", line 180, in load_and_cache_examples\r\n return HG_Datasets(tokenizer, file_path, args)\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling.py\", line 119, in HG_Datasets\r\n dataset = dataset.map(token_encode, batched=True, batch_size = 10000, num_proc = 16)\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1287, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1287, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/multiprocessing\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/multiprocessing\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/multiprocessing\/connection.py\", line 206, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/multiprocessing\/reduction.py\", line 51, in dumps\r\n cls(buf, protocol).dump(obj)\r\nAttributeError: Can't pickle local object 'HG_Datasets.<locals>.token_encode'\r\n```","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":157,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nHi @lhoestq ,\r\n\r\nI tried to use multi-processor, but I got errors as follow: \r\nBecause I am using python distributed training, it seems some conflicts with the distributed job. \r\n\r\nDo you have any suggestions?\r\n```\r\nI0925 10:19:35.603023 140737353971520 filelock.py:318] Lock 140737229443368 released on \/tmp\/pbs.1120510.pbsha.ib.sockeye\/cache\/_tmp_pbs.1120510.pbsha.ib.sockeye_cache_text_default-7fb934ed6fac5d01_0.0.0_512f465342e4f4cd07a8791428a629c043bb89d55ad7817cbf7\r\nfcc649178b014.lock\r\nTraceback (most recent call last):\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling.py\", line 1024, in <module>\r\n main()\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling.py\", line 967, in main\r\n train_dataset = load_and_cache_examples(args, tokenizer, evaluate=False)\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling.py\", line 180, in load_and_cache_examples\r\n return HG_Datasets(tokenizer, file_path, args)\r\n File \"\/scratch\/chiyuzh\/roberta\/run_language_modeling.py\", line 119, in HG_Datasets\r\n dataset = dataset.map(token_encode, batched=True, batch_size = 10000, num_proc = 16)\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1287, in map\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/site-packages\/datasets\/arrow_dataset.py\", line 1287, in <listcomp>\r\n transformed_shards = [r.get() for r in results]\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/multiprocessing\/pool.py\", line 644, in get\r\n raise self._value\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/multiprocessing\/pool.py\", line 424, in _handle_tasks\r\n put(task)\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/multiprocessing\/connection.py\", line 206, in send\r\n self._send_bytes(_ForkingPickler.dumps(obj))\r\n File \"\/project\/chiyuzh\/evn_py36\/lib\/python3.6\/multiprocessing\/reduction.py\", line 51, in dumps\r\n cls(buf, protocol).dump(obj)\r\nAttributeError: Can't pickle local object 'HG_Datasets.<locals>.token_encode'\r\n```","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"For multiprocessing, the function given to `map` must be picklable.\r\nMaybe you could try to define `token_encode` outside `HG_Datasets` ?\r\n\r\nAlso maybe #656 could make functions defined locally picklable for multiprocessing, once it's merged.","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":34,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nFor multiprocessing, the function given to `map` must be picklable.\r\nMaybe you could try to define `token_encode` outside `HG_Datasets` ?\r\n\r\nAlso maybe #656 could make functions defined locally picklable for multiprocessing, once it's merged.","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> I have another question. Because I am using a cloud server where only allows running a job up to 7 days. Hence, I need to resume my model every week. If the script needs to load and process the dataset every time. It is very low efficient based on the current processing speed. Is it possible that I process the dataset once and use the process cache to in the future work?\r\n\r\nFeel free to save your processed dataset using `dataset.save_to_disk(\"path\/to\/save\/directory\")`.\r\n\r\nThen you'll be able to reload it again using\r\n```python\r\nfrom datasets import load_from_disk\r\n\r\ndataset = load_from_disk(\"path\/to\/save\/directory\")\r\n```","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":100,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> I have another question. Because I am using a cloud server where only allows running a job up to 7 days. Hence, I need to resume my model every week. If the script needs to load and process the dataset every time. It is very low efficient based on the current processing speed. Is it possible that I process the dataset once and use the process cache to in the future work?\r\n\r\nFeel free to save your processed dataset using `dataset.save_to_disk(\"path\/to\/save\/directory\")`.\r\n\r\nThen you'll be able to reload it again using\r\n```python\r\nfrom datasets import load_from_disk\r\n\r\ndataset = load_from_disk(\"path\/to\/save\/directory\")\r\n```","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Hi @lhoestq ,\r\n\r\nThanks for your suggestion. \r\nI tried to process the dataset and save it to disk. \r\nI have 1.12B samples in the raw dataset. I used 16 processors.\r\nI run this process job for 7 days. But it didn't finish. I don't why the processing is such slow. \r\n\r\nThe log shows that some processors (\\#12, \\#14, \\#15) are very slow. The different processor has a different speed. These slow processors look like a bottleneck. \r\n\r\nCould you please give me any suggestion to improve the processing speed? \r\n\r\nThanks. \r\nChiyu\r\n\r\nHere is my code:\r\n```\r\ndef token_encode(examples):\r\n tokenizer_out = tokenizer(examples['text'], truncation=True, padding=\"max_length\", add_special_tokens=True, max_length=args.block_size)\r\n return tokenizer_out\r\n\r\npath = Path(file_path)\r\nfiles = sorted(path.glob('*'))\r\ndataset = load_dataset('.\/text.py', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")\r\ndataset = dataset.map(token_encode, batched=True, batch_size = 16384, num_proc = 16)\r\ndataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\ndataset.save_to_disk(output_dir)\r\n```\r\nHere is the log. \r\n```\r\n^M#6: 1%|\u258f | 59\/4288 [55:10<66:11:58, 56.35s\/ba]\r\n^M#1: 8%|\u258a | 356\/4288 [55:39<10:40:02, 9.77s\/ba]\r\n^M#2: 5%|\u258d | 210\/4288 [55:33<17:47:19, 15.70s\/ba]\r\n^M#0: 19%|\u2588\u2589 | 836\/4288 [55:53<4:08:56, 4.33s\/ba]\r\n^M#0: 20%|\u2588\u2589 | 837\/4288 [55:57<4:01:52, 4.21s\/ba]\r\n^M#1: 8%|\u258a | 357\/4288 [55:48<10:38:09, 9.74s\/ba]\r\n^M#0: 20%|\u2588\u2589 | 838\/4288 [56:01<4:02:56, 4.23s\/ba]\r\n^M#3: 4%|\u258e | 155\/4288 [55:43<24:41:20, 21.51s\/ba]\r\n^M#0: 20%|\u2588\u2589 | 839\/4288 [56:05<4:04:48, 4.26s\/ba]\r\n^M#12: 1%| | 29\/4288 [54:50<133:20:53, 112.72s\/ba]\r\n^M#2: 5%|\u258d | 211\/4288 [55:48<17:40:33, 15.61s\/ba]\r\n^M#14: 0%| | 2\/4288 [04:24<157:17:50, 132.12s\/ba]\r\n^M#15: 0%| | 1\/4288 [02:24<172:11:37, 144.60s\/ba]\r\n```","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":219,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nHi @lhoestq ,\r\n\r\nThanks for your suggestion. \r\nI tried to process the dataset and save it to disk. \r\nI have 1.12B samples in the raw dataset. I used 16 processors.\r\nI run this process job for 7 days. But it didn't finish. I don't why the processing is such slow. \r\n\r\nThe log shows that some processors (\\#12, \\#14, \\#15) are very slow. The different processor has a different speed. These slow processors look like a bottleneck. \r\n\r\nCould you please give me any suggestion to improve the processing speed? \r\n\r\nThanks. \r\nChiyu\r\n\r\nHere is my code:\r\n```\r\ndef token_encode(examples):\r\n tokenizer_out = tokenizer(examples['text'], truncation=True, padding=\"max_length\", add_special_tokens=True, max_length=args.block_size)\r\n return tokenizer_out\r\n\r\npath = Path(file_path)\r\nfiles = sorted(path.glob('*'))\r\ndataset = load_dataset('.\/text.py', data_files=files, cache_dir = args.data_cache_dir, split=\"train\")\r\ndataset = dataset.map(token_encode, batched=True, batch_size = 16384, num_proc = 16)\r\ndataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\ndataset.save_to_disk(output_dir)\r\n```\r\nHere is the log. \r\n```\r\n^M#6: 1%|\u258f | 59\/4288 [55:10<66:11:58, 56.35s\/ba]\r\n^M#1: 8%|\u258a | 356\/4288 [55:39<10:40:02, 9.77s\/ba]\r\n^M#2: 5%|\u258d | 210\/4288 [55:33<17:47:19, 15.70s\/ba]\r\n^M#0: 19%|\u2588\u2589 | 836\/4288 [55:53<4:08:56, 4.33s\/ba]\r\n^M#0: 20%|\u2588\u2589 | 837\/4288 [55:57<4:01:52, 4.21s\/ba]\r\n^M#1: 8%|\u258a | 357\/4288 [55:48<10:38:09, 9.74s\/ba]\r\n^M#0: 20%|\u2588\u2589 | 838\/4288 [56:01<4:02:56, 4.23s\/ba]\r\n^M#3: 4%|\u258e | 155\/4288 [55:43<24:41:20, 21.51s\/ba]\r\n^M#0: 20%|\u2588\u2589 | 839\/4288 [56:05<4:04:48, 4.26s\/ba]\r\n^M#12: 1%| | 29\/4288 [54:50<133:20:53, 112.72s\/ba]\r\n^M#2: 5%|\u258d | 211\/4288 [55:48<17:40:33, 15.61s\/ba]\r\n^M#14: 0%| | 2\/4288 [04:24<157:17:50, 132.12s\/ba]\r\n^M#15: 0%| | 1\/4288 [02:24<172:11:37, 144.60s\/ba]\r\n```","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Hi !\r\n\r\nAs far as I can tell, there could be several reasons for your processes to have different speeds:\r\n- some parts of your dataset have short passages while some have longer passages, that take more time to be processed\r\n- OR there are other processes running that prevent some of them to run at full speed\r\n- OR the value of `num_proc` is higher than the number of actual processes that you can run in parallel at full speed.\r\n\r\nSo I'd suggest you to check that you have nothing else running in parallel to your processing job, and also maybe take a look at the slow parts of the datasets.\r\nWhen doing multiprocessing, the dataset is sharded in `num_proc` contiguous parts that are processed individually in each process. If you want to take a look at the dataset processed in the 12th shard of 16 for example, you can do:\r\n\r\n```python\r\nmy_shard = dataset.shard(num_shards=16, index=12, contiguous=True)\r\n```\r\n\r\nHope this helps, let me know if you find what is causing this slow down.","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":174,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nHi !\r\n\r\nAs far as I can tell, there could be several reasons for your processes to have different speeds:\r\n- some parts of your dataset have short passages while some have longer passages, that take more time to be processed\r\n- OR there are other processes running that prevent some of them to run at full speed\r\n- OR the value of `num_proc` is higher than the number of actual processes that you can run in parallel at full speed.\r\n\r\nSo I'd suggest you to check that you have nothing else running in parallel to your processing job, and also maybe take a look at the slow parts of the datasets.\r\nWhen doing multiprocessing, the dataset is sharded in `num_proc` contiguous parts that are processed individually in each process. If you want to take a look at the dataset processed in the 12th shard of 16 for example, you can do:\r\n\r\n```python\r\nmy_shard = dataset.shard(num_shards=16, index=12, contiguous=True)\r\n```\r\n\r\nHope this helps, let me know if you find what is causing this slow down.","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Do you use a fast or a slow tokenizer from the `transformers` library @chiyuzhang94?","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":14,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nDo you use a fast or a slow tokenizer from the `transformers` library @chiyuzhang94?","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> Do you use a fast or a slow tokenizer from the `transformers` library @chiyuzhang94?\r\n\r\nHi @thomwolf ,\r\n I use this: \r\n```\r\nfrom transformers import\r\nAutoTokenizer.from_pretrained(args.model_name_or_path, cache_dir=args.cache_dir)\r\n```\r\n\r\nI guess this is a slow one, let me explore the fast tokenizer. ","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":41,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> Do you use a fast or a slow tokenizer from the `transformers` library @chiyuzhang94?\r\n\r\nHi @thomwolf ,\r\n I use this: \r\n```\r\nfrom transformers import\r\nAutoTokenizer.from_pretrained(args.model_name_or_path, cache_dir=args.cache_dir)\r\n```\r\n\r\nI guess this is a slow one, let me explore the fast tokenizer. ","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"> Hi !\r\n> \r\n> As far as I can tell, there could be several reasons for your processes to have different speeds:\r\n> \r\n> * some parts of your dataset have short passages while some have longer passages, that take more time to be processed\r\n> * OR there are other processes running that prevent some of them to run at full speed\r\n> * OR the value of `num_proc` is higher than the number of actual processes that you can run in parallel at full speed.\r\n> \r\n> So I'd suggest you to check that you have nothing else running in parallel to your processing job, and also maybe take a look at the slow parts of the datasets.\r\n> When doing multiprocessing, the dataset is sharded in `num_proc` contiguous parts that are processed individually in each process. If you want to take a look at the dataset processed in the 12th shard of 16 for example, you can do:\r\n> \r\n> ```python\r\n> my_shard = dataset.shard(num_shards=16, index=12, contiguous=True)\r\n> ```\r\n> \r\n> Hope this helps, let me know if you find what is causing this slow down.\r\n\r\nHi @lhoestq ,\r\n\r\nThanks for your suggestions. \r\nI don't think my problem is due to any one of these seasons. \r\n\r\n1. I have 1,123,870,657 lines totally in the path. I split the large file into 440 small files. Each file has 2,560,000 lines. The last file is smaller a little bit. But they are similar. I randomly shuffled all the 1,123,870,657 lines. Hence, the sequences should also be similar across all the files. \r\n\r\n2. I run this script on the entire node. I requested all the resources on the nodes (40 CPUs, 384GB memory). Hence, these were not any other processes. \r\n\r\n 3. As I say, the node has 40 CPUs, but I set num_proc = 16. This should not be a problem.","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":312,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\n> Hi !\r\n> \r\n> As far as I can tell, there could be several reasons for your processes to have different speeds:\r\n> \r\n> * some parts of your dataset have short passages while some have longer passages, that take more time to be processed\r\n> * OR there are other processes running that prevent some of them to run at full speed\r\n> * OR the value of `num_proc` is higher than the number of actual processes that you can run in parallel at full speed.\r\n> \r\n> So I'd suggest you to check that you have nothing else running in parallel to your processing job, and also maybe take a look at the slow parts of the datasets.\r\n> When doing multiprocessing, the dataset is sharded in `num_proc` contiguous parts that are processed individually in each process. If you want to take a look at the dataset processed in the 12th shard of 16 for example, you can do:\r\n> \r\n> ```python\r\n> my_shard = dataset.shard(num_shards=16, index=12, contiguous=True)\r\n> ```\r\n> \r\n> Hope this helps, let me know if you find what is causing this slow down.\r\n\r\nHi @lhoestq ,\r\n\r\nThanks for your suggestions. \r\nI don't think my problem is due to any one of these seasons. \r\n\r\n1. I have 1,123,870,657 lines totally in the path. I split the large file into 440 small files. Each file has 2,560,000 lines. The last file is smaller a little bit. But they are similar. I randomly shuffled all the 1,123,870,657 lines. Hence, the sequences should also be similar across all the files. \r\n\r\n2. I run this script on the entire node. I requested all the resources on the nodes (40 CPUs, 384GB memory). Hence, these were not any other processes. \r\n\r\n 3. As I say, the node has 40 CPUs, but I set num_proc = 16. This should not be a problem.","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/610","title":"Load text file for RoBERTa pre-training. ","comments":"Hi @thomwolf \r\nI am using `RobertaTokenizerFast` now. \r\n\r\nBut the speed is still imbalanced, some processors are still slow. \r\nHere is the part of the log. #0 is always much fast than lower rank processors. \r\n\r\n```\r\n#15: 3%|\u258e | 115\/3513 [3:18:36<98:01:33, 103.85s\/ba]\r\n#2: 24%|\u2588\u2588\u258d | 847\/3513 [3:20:43<11:06:49, 15.01s\/ba]\r\n#1: 37%|\u2588\u2588\u2588\u258b | 1287\/3513 [3:20:52<6:19:02, 10.22s\/ba]\r\n#0: 72%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f | 2546\/3513 [3:20:52<1:51:03, 6.89s\/ba]\r\n#3: 18%|\u2588\u258a | 617\/3513 [3:20:36<15:50:30, 19.69s\/ba]\r\n#0: 73%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258e | 2547\/3513 [3:20:59<1:50:25, 6.86s\/ba]\r\n#1: 37%|\u2588\u2588\u2588\u258b | 1288\/3513 [3:21:02<6:21:13, 10.28s\/ba]\r\n#7: 7%|\u258b | 252\/3513 [3:20:09<44:09:03, 48.74s\/ba]\r\n#12: 4%|\u258d | 144\/3513 [3:19:19<78:00:54, 83.36s\/ba]\r\n#4: 14%|\u2588\u258d | 494\/3513 [3:20:37<20:46:06, 24.77s\/ba]\r\n#0: 73%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258e | 2548\/3513 [3:21:06<1:49:26, 6.80s\/ba]\r\n#2: 24%|\u2588\u2588\u258d | 848\/3513 [3:20:58<11:06:17, 15.00s\/ba]\r\n```\r\nHere is my script related to the datasets processing, \r\n\r\n```\r\ntokenizer = RobertaTokenizerFast.from_pretrained(args.model_name_or_path, cache_dir=args.cache_dir)\r\n\r\ndef token_encode(examples):\r\n tokenizer_out = tokenizer(examples['text'], truncation=True, padding=\"max_length\", add_special_tokens=True, max_length=128)\r\n return tokenizer_out\r\n\r\ndef HG_Datasets(tokenizer, file_path, args):\r\n \r\n path = Path(file_path)\r\n files = sorted(path.glob('*'))\r\n dataset = load_dataset('.\/text.py', data_files=files, cache_dir = \"\".\/, split=\"train\")\r\n dataset = dataset.map(token_encode, batched=True, batch_size = 20000, num_proc = 16)\r\n\r\n dataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n return dataset\r\n\r\n```\r\nI have 1,123,870,657 lines totally in the path. I split the large file into 440 small files. Each file has 2,560,000 lines.\r\n\r\nCould you please give any suggestion? Thanks very much!!","body":"I migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1","comment_length":198,"text":"Load text file for RoBERTa pre-training. \nI migrate my question from https:\/\/github.com\/huggingface\/transformers\/pull\/4009#issuecomment-690039444\r\n\r\nI tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. \r\nAccording to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file. This test.txt is a simple sample where each line is a sentence.\r\n```\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('text', data_files='test.txt',cache_dir=\".\/\")\r\ndataset.set_format(type='torch',columns=[\"text\"])\r\ndataloader = torch.utils.data.DataLoader(dataset, batch_size=8)\r\nnext(iter(dataloader))\r\n```\r\n\r\nBut dataload cannot yield sample and error is:\r\n```\r\n---------------------------------------------------------------------------\r\nKeyError Traceback (most recent call last)\r\n<ipython-input-12-388aca337e2f> in <module>\r\n----> 1 next(iter(dataloader))\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in __next__(self)\r\n 361 \r\n 362 def __next__(self):\r\n--> 363 data = self._next_data()\r\n 364 self._num_yielded += 1\r\n 365 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/dataloader.py in _next_data(self)\r\n 401 def _next_data(self):\r\n 402 index = self._next_index() # may raise StopIteration\r\n--> 403 data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n 404 if self._pin_memory:\r\n 405 data = _utils.pin_memory.pin_memory(data)\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in fetch(self, possibly_batched_index)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\n\/Library\/Python\/3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py in <listcomp>(.0)\r\n 42 def fetch(self, possibly_batched_index):\r\n 43 if self.auto_collation:\r\n---> 44 data = [self.dataset[idx] for idx in possibly_batched_index]\r\n 45 else:\r\n 46 data = self.dataset[possibly_batched_index]\r\n\r\nKeyError: 0\r\n```\r\n\r\n`dataset.set_format(type='torch',columns=[\"text\"])` returns a log says:\r\n```\r\nSet __getitem__(key) output type to torch for ['text'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```\r\n\r\nI noticed the dataset is `DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None)}, num_rows: 44)})`.\r\nEach sample can be accessed by `dataset[\"train\"][\"text\"]` instead of `dataset[\"text\"]`. \r\n\r\nCould you please give me any suggestions on how to modify this code to load the text file?\r\n\r\nVersions:\r\nPython version 3.7.3\r\nPyTorch version 1.6.0 \r\nTensorFlow version 2.3.0 \r\ndatasets version: 1.0.1\nHi @thomwolf \r\nI am using `RobertaTokenizerFast` now. \r\n\r\nBut the speed is still imbalanced, some processors are still slow. \r\nHere is the part of the log. #0 is always much fast than lower rank processors. \r\n\r\n```\r\n#15: 3%|\u258e | 115\/3513 [3:18:36<98:01:33, 103.85s\/ba]\r\n#2: 24%|\u2588\u2588\u258d | 847\/3513 [3:20:43<11:06:49, 15.01s\/ba]\r\n#1: 37%|\u2588\u2588\u2588\u258b | 1287\/3513 [3:20:52<6:19:02, 10.22s\/ba]\r\n#0: 72%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f | 2546\/3513 [3:20:52<1:51:03, 6.89s\/ba]\r\n#3: 18%|\u2588\u258a | 617\/3513 [3:20:36<15:50:30, 19.69s\/ba]\r\n#0: 73%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258e | 2547\/3513 [3:20:59<1:50:25, 6.86s\/ba]\r\n#1: 37%|\u2588\u2588\u2588\u258b | 1288\/3513 [3:21:02<6:21:13, 10.28s\/ba]\r\n#7: 7%|\u258b | 252\/3513 [3:20:09<44:09:03, 48.74s\/ba]\r\n#12: 4%|\u258d | 144\/3513 [3:19:19<78:00:54, 83.36s\/ba]\r\n#4: 14%|\u2588\u258d | 494\/3513 [3:20:37<20:46:06, 24.77s\/ba]\r\n#0: 73%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258e | 2548\/3513 [3:21:06<1:49:26, 6.80s\/ba]\r\n#2: 24%|\u2588\u2588\u258d | 848\/3513 [3:20:58<11:06:17, 15.00s\/ba]\r\n```\r\nHere is my script related to the datasets processing, \r\n\r\n```\r\ntokenizer = RobertaTokenizerFast.from_pretrained(args.model_name_or_path, cache_dir=args.cache_dir)\r\n\r\ndef token_encode(examples):\r\n tokenizer_out = tokenizer(examples['text'], truncation=True, padding=\"max_length\", add_special_tokens=True, max_length=128)\r\n return tokenizer_out\r\n\r\ndef HG_Datasets(tokenizer, file_path, args):\r\n \r\n path = Path(file_path)\r\n files = sorted(path.glob('*'))\r\n dataset = load_dataset('.\/text.py', data_files=files, cache_dir = \"\".\/, split=\"train\")\r\n dataset = dataset.map(token_encode, batched=True, batch_size = 20000, num_proc = 16)\r\n\r\n dataset.set_format(type='torch', columns=['input_ids', 'attention_mask'])\r\n return dataset\r\n\r\n```\r\nI have 1,123,870,657 lines totally in the path. I split the large file into 440 small files. Each file has 2,560,000 lines.\r\n\r\nCould you please give any suggestion? Thanks very much!!","embeddings":[-0.2351640612,-0.2028229982,-0.0119608929,0.408447057,0.3831166625,-0.1324438751,0.5139924288,0.4229528308,-0.1767195165,0.0487692058,-0.2432833612,0.1370509565,-0.144555673,-0.389241904,-0.0388365388,-0.3020621836,-0.1219130456,0.1924059838,-0.4711352587,-0.0367562771,0.1829230189,0.0783707425,-0.137515828,0.1973749697,-0.6427964568,0.1069567874,0.2162100375,0.2214321196,-0.0749238133,-0.3002363741,0.2062293738,0.0434491038,0.5629116297,0.7953851223,-0.0001255157,0.1221125871,0.248756811,-0.2410901189,-0.312382251,-0.1114152744,0.4535882175,-0.2304576486,0.1303864568,-0.3562261164,0.0759145692,-0.1436130852,0.0986075252,-0.0940504894,0.5390230417,0.3707350791,0.0733966455,0.4406839907,-0.0109579563,-0.0613048933,0.3599255383,0.1098705828,-0.0299898889,0.0779794678,0.3455821574,-0.2738672793,-0.4754483998,0.1371811926,-0.1577266455,0.0799182877,-0.056934163,0.127282083,0.1293712556,-0.0320416205,0.0433062501,0.3101655841,0.5464712381,-0.1133196056,-0.3787391484,-0.3641907275,-0.1538963616,-0.1453783661,0.3427860439,0.0990873426,-0.2295717299,0.1603964567,-0.1188358143,0.0387966447,-0.4248600304,0.0132333254,-0.16641289,0.3269487023,-0.1371064335,0.0592869185,0.3322789371,-0.0902749598,0.2790078521,0.0059216856,0.1468862295,0.3295742571,-0.3641236424,-0.0306546167,-0.1202296764,-0.397066474,0.2847504616,0.059628021,0.2434100658,0.1548983753,-0.1783426106,-0.0348099098,0.2422806919,0.1821570992,-0.180077076,0.2752799094,0.2457235157,0.3165408075,-0.1399267316,-0.140717715,-0.4243340492,-0.2287802696,0.0969789028,-0.1354229897,0.1783904135,-0.1036362052,0.0023347768,0.0298269074,-0.0968155116,-0.1359277219,0.1534668803,0.396576494,-0.1335892081,0.3401686251,0.156048581,0.0891145095,-0.2729740441,-0.2191158235,-0.0154704899,-0.2309262753,-0.2201966196,0.1377543509,0.2425263524,-0.0458625555,0.2399176359,-0.3079228699,0.2493965626,-0.158296898,-0.1100429296,-0.3570912778,0.0135691743,0.2636014819,-0.1435478032,0.2023341358,0.1805712581,-0.0803246349,-0.0784907416,0.3265265822,-0.1811527014,-0.4468552768,0.1354891807,0.0617074184,-0.1038749516,-0.0352106765,-0.0966458619,0.0569574088,0.4921211302,-0.319745928,-0.0241774004,-0.1626605541,-0.1030522734,0.0550209507,0.2268247902,0.5365388989,-0.1438212693,0.020377133,-0.0253133215,-0.0516009592,0.1682346612,0.6376840472,-0.2747702599,0.5127100945,-0.136062175,0.1061270759,0.2240159214,-0.0344659239,-0.3717764914,0.3262194693,-0.0751044452,0.0406257249,0.2201387882,-0.0307628587,0.052055411,-0.0443849154,0.2690665424,0.2103522122,0.1113788635,0.1413730085,-0.1208721325,-0.1993711889,-0.0637961626,0.5253330469,0.2333912998,0.0096206898,-0.2704213262,0.0675348863,0.2778604031,-0.2674022317,0.1534748375,0.1924356371,-0.2048864365,0.2004879117,0.1048062071,-0.1451075971,-0.1432375461,0.0395660326,0.0148979388,-0.0170760285,-0.1478383094,0.1222516596,-0.2059447616,0.073311761,-0.1966186017,0.0283276383,-0.0339356661,-0.0356443822,-0.096428737,-0.1215626895,-0.2722823322,0.2359145582,-0.1684353948,0.1812613457,-0.2048251629,0.1474659592,-0.0961617008,-0.2596963048,-0.0846058205,0.1086725593,-0.1160761043,-0.0557811707,-0.1087903157,0.1681584567,-0.0896209404,-0.1042771265,-0.1853775233,0.0577392504,0.1356106848,-0.2802486122,0.1246793121,0.2198598534,0.1722252071,0.012317379,-0.1304285973,0.239719674,-0.0239515845,0.1762232035,0.1257120371,-0.1457850486,0.2596863806,-0.1022094563,-0.1274696141,0.1805504411,0.4158622921,0.2873030007,0.2258876562,0.0142894555,-0.1222361177,-0.5006818771,0.274925977,-0.1510628462,0.0672706962,0.2920909226,-0.377589047,0.0398272388,-0.2509226203,-0.2540422976,0.3311235607,0.0367435068,-0.0564246066,-0.0796815231,0.0952707604,-0.1589733064,0.1148993522,0.2903901637,0.0871184096,0.4159710407,0.0195508339,-0.01286125,-0.2850967646,-0.2793557644,-0.0324503444,0.0772638097,-0.2523705363,0.3918679953,0.181742996,0.1658732146,-0.3981604576,-0.0665213987,-0.0594438277,0.1600899696,-0.0328797214,0.2245890498,-0.1531229019,0.1375098228,0.2761565447,0.2386553884,0.531009078,-0.3450524807,-0.0228330046,-0.2657009661,-0.1711072922,-0.0281735435,0.1564720124,-0.3800223768,-0.0227340013,-0.1593811363,0.0414373428,-0.0410522707,-0.2128789574,0.1461291462,0.095018141,-0.0657052845,-0.0873159245,0.3188772202,0.2235686481,-0.2584017813,0.165287137,-0.3514517248,-0.0836746767,0.1345466673,-0.0258615669,-0.0039209398,0.1339139193,-0.4912370741,-0.0777524337,-0.4011791348,0.206659466,0.2536621094,0.1190750003,0.2927311957,0.3777467906,0.2591692209,-0.1536995769,0.2036137134,0.0069752932,-0.2739513814,0.6016744375,0.037279278,-0.3855601549,-0.273537755,-0.1876644045,-0.0212981012,0.1292121708,-0.501755774,0.2092529982,-0.1055911705,-0.0392211415,-0.2687577605,0.2106827796,0.3517278433,-0.0582844391,0.0220018271,0.0811398253,-0.1444359869,0.0589835458,0.0612709299,0.2923367918,-0.2327822298,0.0906195641,-0.1289009154,0.7936124206,0.0342532881,-0.0950503871,0.1772329956,0.0900984257,0.1697518826,-0.1435037255,-0.1156521142,-0.2848767042,-0.3349307179,-0.0750081539,0.3109183311,0.0489610545,-0.2725122571,-0.1710483134,0.0355686136,0.1201001108,-0.3826608658,0.5341356397,0.0394279882,0.1441112608,-0.1620216519,0.0803804696,-0.2024175227,-0.0343527719,-0.1732664704,0.2794306576,0.1588231176,-0.1944961846,-0.3677433729,0.2286981344,-0.1797410697,-0.0017897844,0.1497692317,0.2188696414,0.1418789625,-0.4290395379,0.0551718809,-0.1176693514,0.5121035576,0.4852637351,-0.19270733,-0.03602276,-0.1814633757,-0.3889468014,-0.1276862025,-0.1543958038,0.1139600277,0.2410908639,0.377393961,-0.2511388958,-0.0662893504,0.2503164709,0.1687663496,-0.1369205564,-0.3428625762,-0.2601178586,-0.1211403981,-0.5566355586,0.0615769401,0.069577232,0.2340845913,-0.1890149564,0.1321268529,0.0574261546,0.0725398883,0.1174274683,0.0908756107,0.1282272339,-0.2521229386,0.1319787502,0.3987477124,-0.0123182926,0.0352788381,0.496422559,0.1588187069,-0.3698637784,0.0866975039,0.0357630737,0.1273487657,0.3553589582,-0.0171525907,0.0441162698,0.2130622417,0.1465008408,-0.295083791,0.278614223,0.0300531965,-0.0124504035,-0.4403938353,-0.3575322628,0.431155771,-0.205529809,0.1844107211,-0.1050693467,-0.1353550255,-0.0892196596,0.46472615,-0.101550214,1.1124306917,-0.3223896921,0.5987325907,-0.0135056814,0.4439116418,0.223923251,-0.4978284836,0.1245432124,-0.4053942859,-0.1702314615,0.0305951592,-0.1809033751,0.03196913,0.225775212,-0.2543109655,0.17267102,0.2434322238,0.1995401084,0.1136990637,0.0503351465,-0.1298247874,-0.4798262715,-0.5485130548,-0.0774983168,-0.0896058157,0.3039267361,-0.006455916,-0.0907665268,-0.1625556052,-0.2407623678,-0.1279716641,0.0543005504,-0.1000754237,0.1193536967,0.2212517112,-0.2080654204,0.2713838816,0.4447632432,0.0335667394,0.2296931893,-0.1550184488,0.0855645463,-0.3620647192,-0.2912548184,-0.149007827,-0.0249443557,0.3309189081,0.1130107343,-0.2030445188,0.0722931474,-0.0686616749,-0.2512925565,0.1684594452,-0.0013506027,0.3717734516,-0.5535949469,-0.0597360395,-0.0040133162,-0.2479250431,-0.0405743569,0.0385403484,0.2612255216,-0.2834444046,0.1107986346,0.3466929197,-0.2172766179,-0.0156595707,0.6410550475,-0.0254197214,-0.0714917928,0.6801123023,0.3254789114,-0.1144692302,-0.1269172579,0.0653540045,0.1271479726,-0.8447719216,0.1113626435,-0.1274243146,-0.1004357561,-0.0756642967,0.0183767714,0.2501248717,-0.3937130868,-0.131799981,-0.3471087813,-0.8210464716,0.1725009531,-0.0540787317,0.0583463274,0.1969920099,0.0994361117,-0.06796702,0.2009984255,-0.1649546921,0.2213954031,-0.2336621732,0.2557599545,0.4548303485,-0.4371107519,0.1852941066,0.1128205582,0.0162504707,0.2322521508,-0.0405183956,-0.1479799598,-0.2513855398,0.1854471713,0.1747972071,-0.3302823901,-0.0397498123,-0.1643350273,-0.0223631021,-0.3079045117,0.1542821825,0.0038331202,0.0557835586,0.1906662881,0.1657608598,0.2118629813,-0.5373828411,0.113368012,-0.2964808643,0.2011533529,0.2831169367,-0.0739614442,-0.0867490172,-0.0841353163,-0.2059238553,0.2081248909,0.7083237171,0.0150301335,0.3763867915,-0.3561716974,0.1149772331,-0.02245632,0.1344812959,0.5138250589,-0.3078081012,0.0999430493,0.3407734632,0.1181182787,-0.0901404768,-0.077459082,0.0208086912,-0.4548580945,0.0252234451,0.2397970259,-0.1746069938,0.1946325153,0.1334871948,0.2194302231,0.1392875165,0.0803815722,0.0127503593,0.4389444888,-0.05677801,0.0051423009,0.3662571609,0.1114442125,0.2706191838,0.6254372001,-0.2381746769,0.2555933893,-0.2969632447,0.2216154635,-0.0832466409,-0.5557139516,-0.1067691967,0.2992471159,0.2346816063,-0.0115802409,-0.1024339125,0.2693872452,-0.0783937871,0.4590183496,-0.2419998199,-0.0667976066,-0.1246631816,-0.3170201778,0.0169500671,-0.1958136111,-0.0729703307,0.0799741894,-0.2137755156,0.073472023,0.0290785823,0.2068382204,-0.0088843359,-0.3853543401,-0.0500598438,0.2007181048,-0.0043411236,-0.3296885788,0.2033612579,0.3960950673,0.0450604036,0.2269716561,0.1732403338,0.4248790741,0.2449591905,-0.2228327394,0.2395240515,-0.2593833804,-0.1199424937,-0.0664998889,0.3457730114,0.2252686173,0.2683982849,0.1735759825,0.0053322124,-0.1643180251,0.2192570716,-0.0700120777,-0.1345784664,-0.1802104115,0.177402705,0.1630974114,-0.3449825644,-0.2759726048,-0.0461293794,-0.315161705,0.2501935065,0.4753973186,-0.0478396863,0.141846329,0.1386902183,-0.0118243285,0.0551333837,0.6322231889,0.4257953465,0.2179269493,-0.3433360159,0.1284049004,-0.3904696107,0.0644460097,0.0279973131,0.2212458998,0.0685954839,0.1235559136,0.1988864243,0.2944620252,0.024347119,-0.6205528378,0.2357997596,0.1450716704,-0.1094386801,-0.2260716408,0.242035836,0.2628989816,-0.0525474325,-0.5544046164,0.0243350063,0.1460006088,-0.0373800211,-0.2378278822,0.081513375,0.0553997569,0.096497491,0.1544813812,-0.1375415027,0.5417482257,-0.0090198806,-0.0410705097,-0.2688592672,-0.1199016497,-0.0654932186,-0.0751951709,0.4655777812,0.5816173553,-0.2886754572,-0.0158442277,-0.3237230182,-0.019135965,-0.0375781357,-0.240942508,-0.3693308532,-0.0854671448,-0.0077545331,0.0852633119,-0.0350231864,0.3017436862,-0.1090993509,-0.0184354875,-0.402327776,-0.219107464,0.4515859187,-0.3525932431,-0.3459249139,0.0132858073,-0.1101940498,0.1301019192,0.2072258145,-0.7693996429,0.4182786047,0.2454119474,-0.114542149,-0.5120350122,0.2250448614,0.0633463189,-0.0521315262,-0.2652809918,0.6368246675,0.0586002506,-0.1705967486,-0.2721941471,-0.2963877618]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/608","title":"Don't use the old NYU GLUE dataset URLs","comments":"Feel free to open the PR ;)\r\nThanks for updating the dataset_info.json file !","body":"NYU is switching dataset hosting from Google to FB. Initial changes to `datasets` are in https:\/\/github.com\/jeswan\/nlp\/commit\/b7d4a071d432592ded971e30ef73330529de25ce. What tests do you suggest I run before opening a PR?\r\n\r\nSee: https:\/\/github.com\/jiant-dev\/jiant\/issues\/161 and https:\/\/github.com\/nyu-mll\/jiant\/pull\/1112","comment_length":14,"text":"Don't use the old NYU GLUE dataset URLs\nNYU is switching dataset hosting from Google to FB. Initial changes to `datasets` are in https:\/\/github.com\/jeswan\/nlp\/commit\/b7d4a071d432592ded971e30ef73330529de25ce. What tests do you suggest I run before opening a PR?\r\n\r\nSee: https:\/\/github.com\/jiant-dev\/jiant\/issues\/161 and https:\/\/github.com\/nyu-mll\/jiant\/pull\/1112\nFeel free to open the PR ;)\r\nThanks for updating the dataset_info.json file !","embeddings":[-0.0815305412,0.0461987108,-0.0440148339,-0.0894913226,0.0934377089,-0.2022670954,0.2342824489,0.1845259815,-0.1438383758,0.0541750826,-0.2520014346,-0.2150709629,-0.0051280679,0.2300495505,0.0581797324,0.0094579235,-0.0636636242,0.0759056732,-0.2601184547,-0.1705842614,-0.2852198184,0.418702364,-0.1797442734,0.2601876855,-0.1021881476,-0.1262492239,-0.1382658184,0.4150830209,-0.3163621426,-0.2958871424,-0.0239996947,0.361481756,-0.1615878493,0.2830239832,-0.0001059319,-0.0397417583,-0.007462488,-0.0459720418,-0.1444257498,0.1503904611,0.1618950516,-0.184408471,-0.0377814323,0.0120939128,-0.4440310597,-0.3115535676,-0.0171919167,0.1068836674,0.0678983554,0.5909659863,0.2548412085,0.1707350761,-0.0331392027,-0.1445780247,-0.0079360669,-0.2772675157,0.0276968069,0.0290516056,0.2536796331,-0.2268133014,-0.1864950955,0.0406798087,0.2399950773,0.1024083942,0.1733943075,0.035082113,0.0000902436,-0.4717931449,0.071963273,0.3538927436,0.4369125664,0.0246256758,-0.1944542378,0.032328546,0.0150423599,0.061215397,0.4388726354,0.09892115,0.1462051123,0.2380239815,-0.0965913832,-0.3549345136,-0.0877736434,0.2519305348,-0.1447557658,0.298366636,0.0057976893,0.0242887866,0.0825955644,-0.3219253719,-0.1192052886,0.0524778888,-0.1663146317,-0.1329897344,-0.1643300056,-0.4506448805,0.3536868393,0.1017425954,0.0475164726,0.4340528846,-0.0757113546,0.1664914191,-0.2894127965,-0.0448380522,0.0639835671,0.0710309669,0.0457126796,-0.0002517102,0.4585216939,0.0386852883,0.0170890223,0.1182725802,0.2544286847,-0.2075129598,-0.5383527875,-0.211567834,0.2822569311,-0.3265925646,-0.0379544161,0.02608856,-0.1356073618,0.1298278868,-0.1020926908,0.2583024204,-0.0753673017,0.1692406833,-0.0180653371,-0.154808253,-0.2725767195,-0.146990031,-0.1563768089,0.0604998916,-0.0364803895,0.0084098261,0.0403368957,-0.0248805005,0.2173555046,0.0164257977,0.0505648963,0.3126538396,0.0204903074,0.0291887354,0.431512326,0.3250582516,-0.11652641,-0.0154463416,-0.1604562104,-0.1946548074,-0.1703055054,-0.173536405,-0.1577408761,-0.2864385545,-0.2885171771,0.2177111953,-0.0902739614,-0.0691076368,0.2576452196,0.3756667078,-0.2998924851,-0.2100192904,0.0355548672,0.040084675,0.0788224414,-0.1525305808,0.1606201977,0.2989221215,-0.2976539135,0.0279228017,-0.0770008713,-0.1545489132,0.0132196108,0.4863993824,-0.0569466241,-0.0750187188,-0.2804278433,0.1376403719,0.1677537113,-0.2060002238,-0.4310607016,0.2578151226,-0.2372572422,-0.1785892546,-0.2062732279,0.0894152448,0.1643792242,-0.2230277956,0.1707339287,0.5363100171,0.1727464348,0.0084679192,-0.5109096766,-0.2141594142,-0.3837043941,-0.1436802149,-0.0190733764,-0.0882952511,-0.00121413,0.0030274438,-0.025061924,-0.0132165756,-0.0703783855,-0.0113289552,0.5132507682,0.2801051736,-0.0812110603,-0.0080364719,-0.1904176772,0.0624640323,-0.5583847165,0.5273717642,0.0337747894,-0.0735982209,-0.0710749254,-0.156639114,-0.0135976952,-0.3969288766,0.2133641839,0.244327262,0.0980440378,0.2189209461,-0.07281477,0.1337826848,0.1923353225,0.1394000947,-0.1091362238,0.473647207,-0.0882739201,-0.0194885898,0.2572568357,0.2507839203,-0.0481287725,-0.2667835355,0.0487868227,0.6014677882,-0.2787816525,0.011807479,0.6889582276,-0.0399518721,0.4487111568,-0.1293281317,0.0818937942,0.0635509491,-0.1860887259,0.2565319836,0.0718627945,0.4460614622,0.1660874188,-0.1216769516,0.2087979913,0.2171675563,0.3715538085,-0.1284996271,-0.0608859174,-0.2927063406,0.0431978963,0.4350191355,0.0389897525,0.0884483308,-0.4351103902,0.2315612286,0.4314901829,-0.0811433941,0.0094957892,0.1501520127,-0.0112165501,-0.328945756,0.3498878181,-0.0315191858,-0.0044684149,0.1885190606,0.0455307513,0.0625536069,-0.3268223405,-0.3474856615,0.1979938,-0.0364803895,-0.2264385074,-0.1659120321,0.1251381189,-0.120339781,-0.4058655202,0.1149158925,-0.0358049348,0.2504002452,-0.285117507,-0.1018844917,-0.4362300634,-0.2920592725,0.0750426874,-0.2055503428,-0.36782974,-0.5351566076,0.309666723,0.0906980485,-0.218876332,0.3924735487,0.3090710342,0.5903583765,0.1133562401,0.1819350719,0.0464658774,-0.1354600638,-0.3340174854,0.2396905571,0.0517389253,0.3521155119,0.2807171643,-0.135337919,0.0936932713,-0.6681101322,-0.5598103404,0.112448588,-0.0143426107,0.3170069158,0.065813899,0.1472982168,0.2208315581,0.0077801589,0.0372733511,-0.1913273335,0.009852048,0.1321816742,-0.1632446051,0.0967765749,0.043182712,-0.444024384,-0.2044060081,-0.301263988,0.1185916662,0.1675300151,0.1398147643,-0.2043167502,0.0650282577,0.1584970206,-0.1760678887,-0.3235881031,-0.2564825118,-0.2009974718,0.2326965183,-0.4235156476,-0.4898581803,0.0594712943,0.1228051186,0.0415698439,-0.1438213587,-0.5145233274,-0.3850465119,-0.1311661303,0.4452172518,0.3657035232,-0.2018443197,0.1979298741,-0.0596164614,-0.3212858737,0.0893809497,-0.0843456089,0.1911910772,0.3078683317,0.2809933722,0.2219635695,0.2662307918,-0.0706426203,0.5673808455,0.2374493182,-0.1604406983,0.0660618991,-0.2345511019,0.3870394528,-0.0971257612,0.0053814203,0.2535950243,0.1460389197,0.2252920717,0.5125534534,0.0588065125,0.154195413,-0.22995992,0.527372539,0.0651183203,-0.4477466941,0.0302900039,-0.5027089119,0.0159131084,0.1773450524,0.0364188701,-0.0556848273,-0.1066928953,0.1129028052,0.1002189368,0.0557843745,-0.1214634404,-0.2502906322,0.1471623629,-0.5470259786,0.2997021675,-0.2632034719,0.5383787155,-0.2531147003,-0.0838700756,0.171139881,0.0929245502,0.254989028,-0.612283349,-0.1580459476,0.0556144379,0.0428078175,-0.3103716671,-0.0006890253,-0.4070082307,0.2171870023,0.1095143631,0.5468440652,-0.3629087508,-0.2147279233,0.2097549886,-0.2006596923,-0.0924147218,0.3507802784,-0.30965662,-0.3522347808,-0.2478978783,-0.0312763229,-0.2292221785,-0.0800451711,-0.1789912283,0.029603811,0.1523659527,-0.2168798447,0.2246189862,0.2519338131,0.2106309682,0.1249459237,-0.1091201454,0.1588959843,-0.1786711812,0.0325732604,0.5116541982,-0.3596472442,-0.0967076644,0.1853242666,0.0542922094,0.0287307296,0.4790836573,0.0499225073,-0.0466564149,0.1348198503,0.3730485439,-0.1307112724,0.3003566265,0.2641623914,0.0895337239,-0.2209350318,-0.2228918523,0.2768338919,-0.0238442533,-0.1229161471,0.2535097599,0.4700386226,-0.0882977247,0.4154875278,0.0797054544,0.7730290294,-0.0685592368,0.0282212347,0.4132522047,-0.066907458,0.3546436429,-0.1212833822,0.1169231385,-0.2277104408,-0.1538530886,-0.159098953,0.1360472739,0.0637281835,0.1171999499,-0.3691408634,0.2436383069,0.4324183762,0.3205875754,0.1924601942,0.4231338501,0.201855123,-0.0234868117,-0.1792185158,0.2642117143,-0.0219123531,0.3322498202,-0.0014296358,-0.2950154245,-0.1409733295,0.0563821197,-0.1340952516,-0.0797036588,-0.1981858909,0.231848821,-0.0570408665,-0.23352772,-0.247396484,-0.1146622002,0.157467097,0.208422333,-0.4108068943,0.01340729,0.00492793,-0.1079556867,0.0123557411,0.1106174216,0.3964450657,-0.1370070726,-0.2611355782,0.360835135,0.0577628836,-0.1249186248,-0.1385422498,-0.2958687842,-0.2823418081,-0.2117191255,-0.073841244,0.0980251431,-0.466658175,-0.1307080984,0.1791718453,-0.1544831097,0.1049045175,-0.0922380462,0.3735853136,-0.2578256726,-0.2293982655,0.2003886104,-0.1755502969,0.3283308148,0.080559805,0.1840776801,-0.0736283734,-0.1741015464,-0.1624800861,0.1388251483,-0.2557908297,0.1647531986,-0.0596325547,-0.330878526,0.0415149443,-0.0609535538,0.2164603919,0.0540034734,-0.2681308389,-0.0051541561,-0.2905800045,0.3610091805,-0.1422675699,0.1314106435,-0.0606693886,0.2106359005,0.1941748112,0.0791461468,-0.4309192598,0.1573156565,-0.3467265368,-0.0020664367,-0.312349081,-0.3435248733,0.0159708969,0.0064552659,0.0808409601,-0.222954616,-0.4124655724,-0.1862889826,0.0986650586,0.0800680518,0.1222157404,-0.180616051,-0.1171298772,-0.0043930016,-0.1248991564,-0.0786545724,0.1159434095,-0.2558644712,-0.0912729576,0.2219664454,0.3599924147,-0.3889544904,0.0836923718,0.1316675395,0.1729891151,0.1188572422,0.0864446685,0.0319050513,-0.2861537933,-0.1994773895,-0.18700324,0.333829999,-0.1015449911,-0.0157813504,0.1965237856,0.0709180459,0.0737632364,0.1506053358,0.2433271408,0.3122338057,-0.1787494868,-0.3073067665,0.146731779,0.1745702773,-0.3503434658,0.1167185381,0.0384443589,0.0524408817,0.2978146672,-0.0002457084,0.3503715694,-0.1357497722,0.0389633998,0.2801171839,0.1539386511,-0.3935673237,0.0852732807,0.2973479629,-0.1966025829,0.1672002375,0.3930957317,-0.0804626718,-0.0040490762,0.2333777249,0.1263065636,0.3260559142,0.0080614975,0.2269499004,0.3948718607,-0.3018023074,0.0483877473,0.3830452859,0.0156559497,0.0265776124,0.3089621067,0.4004354775,-0.4279942513,-0.3731962144,-0.1386023611,0.3819883168,-0.1616796851,-0.0353819989,0.0208015479,-0.0453277789,-0.057607837,0.0447756387,-0.2469455898,-0.2255786806,0.1555900127,-0.0265023429,-0.2033315301,-0.4346763194,0.1021979675,-0.3050889373,0.2492520213,0.0552985668,0.5072445273,0.0841786489,-0.0730943531,0.2232636362,0.2728118896,0.3121759295,0.0022240367,-0.2196980715,0.2568450272,-0.0332002714,-0.1180585846,0.0816432014,0.1168226227,0.1694725603,-0.162883535,0.0781476796,0.2136855572,-0.1565593481,0.0451759957,0.0161516741,0.2040765285,-0.0704761893,0.4516065419,0.1947488338,-0.2316397578,-0.0795545876,-0.0169315375,-0.3097386956,0.0304273628,0.1017192602,-0.2119780481,0.0930374116,-0.2143671066,0.1277318299,0.2142469883,0.7379848957,0.1987554729,-0.1023311093,0.0755141973,-0.3565084934,-0.7066305876,0.0679287463,0.0174250919,0.0417785794,0.0690785199,0.2518094182,0.1095759571,0.245705232,-0.1399485618,0.0239122622,-0.1392636895,-0.388055414,-0.6486176252,0.1816150844,-0.2430571914,0.4243817627,-0.0136386715,0.0583824888,0.4174169302,-0.0171294827,0.1109085903,0.1680636704,-0.0776835307,0.1956486553,0.1928188056,0.1089505255,0.3909157217,-0.0563453436,-0.0455234647,-0.1173101515,-0.2512851655,-0.2828775644,-0.081135653,-0.0378569439,-0.0082919253,-0.0191515908,0.2348590344,-0.1698960215,-0.004528767,0.0854043588,0.0664771274,0.2939960063,-0.1623925567,-0.1506487429,0.3404094577,-0.1265919209,-0.0157877076,0.3781648576,-0.061296504,-0.1356811225,-0.1804108769,-0.4401238561,0.6395123005,0.0316221602,0.1483445615,-0.0129707167,0.2195350528,-0.035402786,-0.0743952692,-0.281844914,-0.059774112,0.1437632293,-0.0679257587,-0.0947337896,0.2494254559,-0.0485475957,0.172566697,-0.134756133,0.1339234561,0.1812656671,-0.4520428479,-0.0652324855,-0.0073448205]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/600","title":"Pickling error when loading dataset","comments":"When I change from python3.6 to python3.8, it works! ","body":"Hi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```","comment_length":9,"text":"Pickling error when loading dataset\nHi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\nWhen I change from python3.6 to python3.8, it works! ","embeddings":[-0.2029621452,-0.1918641627,0.1510746926,0.2650444508,0.1417573392,-0.1736165881,0.2733694315,0.3465311229,0.0597782321,-0.0787867382,0.0942319855,0.3698832691,-0.2363304645,0.0360416137,0.0861706808,-0.4138222039,-0.0133403046,0.1228954196,-0.172864154,0.0111276107,-0.1992149353,0.319937855,-0.2243003845,0.1471229792,-0.505433321,-0.0055762944,0.075616017,0.1886179298,-0.0382539779,-0.3906199336,0.357863754,-0.0553878881,0.2542710304,0.4166387916,-0.0001185715,0.1009830087,0.4450067282,-0.1613719016,-0.2186082751,-0.5004016757,0.3419595361,-0.1968306452,0.1952064484,-0.119422853,-0.138964057,0.0235723462,0.0966473296,0.0984587669,0.3546349704,0.2370629162,0.2198283076,0.7399469614,0.0196053255,-0.004001346,-0.1397824585,0.3455076218,-0.0702364743,0.1682084203,0.0740607306,0.0049477555,-0.250533253,0.2688044012,0.0097407419,-0.1106548905,0.0340497196,0.0430935025,-0.0089143366,-0.0071578994,0.2315981686,0.1808640808,0.3163917959,-0.1332210153,-0.4018171132,-0.4381490052,-0.0321017392,-0.2538236082,0.3569175601,0.0049056802,-0.1719962656,-0.0149831194,-0.1687421799,-0.2313759178,0.0281378422,0.3235939145,0.1266870797,0.6207981706,-0.0286829732,0.2285003811,0.3370732367,-0.188007012,0.1313721091,-0.0277588796,0.0349624828,0.4023726285,-0.238850221,-0.1197175607,0.0664065331,-0.1022803336,0.040619906,0.0130149191,-0.3214137554,0.0603422858,-0.2768704593,-0.0223208498,0.2950415611,0.2889094651,0.0581622198,0.5016816258,-0.0262932908,-0.1388157159,0.2159735262,0.0168195665,-0.124156177,-0.281819135,-0.1952973008,0.0977016166,0.0427054726,-0.2523407638,-0.0277266987,-0.243488133,-0.0807851404,0.0121760387,0.1265031397,0.482233882,0.0803427994,0.1158533469,0.0164868422,0.1796439886,-0.4184333384,-0.2533637583,-0.1892677546,0.022043325,-0.3372351229,-0.1492154151,0.2223524898,0.1535892189,0.3354985416,-0.1102682278,-0.0604240373,0.1364385486,-0.0179690253,-0.1492116302,-0.0685894415,0.2175128609,-0.2264653891,0.1727866828,0.2571268976,-0.1251242161,-0.2969128489,0.1636125892,-0.0111897727,-0.2530519068,-0.1209304109,0.0851275399,-0.1824117154,-0.2599703074,-0.3591947556,0.2333444357,0.3632029891,-0.263068974,-0.1501649469,-0.3836107254,-0.1859427094,-0.1210468039,0.1095169187,0.6124694347,-0.1049165651,-0.4261669219,-0.1083565578,-0.1307385713,0.386395663,0.4650295377,-0.4532147944,0.153607592,-0.2186063975,0.5694765449,0.383651495,-0.1736365557,-0.2988277078,0.2628951669,-0.2244315892,0.2072925121,-0.0707563609,-0.1237978861,-0.1637935787,0.0578438863,-0.1647842675,0.3448475599,0.1585469395,0.2666162848,-0.2812247872,-0.2925919592,0.4360112846,0.2920269966,0.216148749,0.1050296426,-0.1254457384,0.3208654523,0.2130340934,-0.2217202485,0.191267103,0.0267760381,0.0222752038,-0.1574062705,-0.032853514,0.0872845724,-0.6515653729,0.1748918146,-0.2471687347,0.3482395709,-0.3800010085,0.1222224906,-0.0670937449,-0.1169426218,-0.3418495059,-0.2125000358,0.0444627553,0.183524698,-0.0876024067,-0.0400484204,-0.0541984327,0.0558037721,0.0123563483,0.1081894338,-0.4298703969,0.1635243893,-0.1746212542,-0.3540358245,-0.0586851798,0.2108436823,0.2123282403,-0.2704420388,-0.2660779059,0.2112212032,0.0572887398,0.2215671539,0.0190088879,-0.00503748,0.1972181201,-0.192833066,0.0177532136,0.3558892608,0.0182690863,0.0179571509,0.0842206031,0.3661319017,0.0188044794,0.0989182368,0.1062778458,0.0629072785,0.1784609258,-0.0767172575,0.163967669,-0.1630000323,0.300399065,0.0798098519,0.4254424274,0.0218044408,-0.336406678,-0.2099469602,0.4499590993,0.0788158178,0.1665172726,0.3088004887,-0.0811552033,-0.1771025658,-0.0583941601,0.0550881997,0.2792567611,0.0804542229,-0.0513546728,0.0372813307,-0.0395417921,-0.1005213484,0.1776282787,0.101741232,-0.0386786871,0.3335563242,-0.0069130347,0.0972452164,-0.4322930276,-0.2849202156,0.1184080318,0.3154613972,-0.2130566686,0.1573490947,-0.1570317745,-0.0666516423,-0.4013566077,-0.2381341457,-0.4762080014,-0.1907107383,-0.2355116904,0.1614567786,-0.1278426498,0.2946344018,-0.0560883917,0.0913452059,0.0177859832,-0.0834005475,-0.1864178628,-0.2087877989,-0.4619050622,-0.0359911546,0.2449781895,0.021026792,0.1043828949,0.123727344,-0.0020178452,0.1463040262,-0.3587393761,0.2713183761,0.0121457307,0.187657088,0.1870482713,0.0780553222,-0.1918313205,-0.4377750158,0.2089122981,-0.062545836,-0.2025256455,0.0551075153,0.0965651721,-0.0389919728,-0.1555091143,-0.3513694406,-0.3187605441,-0.4107227623,0.1541299969,-0.1068721488,0.192685917,0.6526251435,-0.0405006818,0.3030754626,-0.23926799,0.0254845191,-0.1506410986,0.1499269158,0.0620459914,-0.1284027398,-0.3454410732,-0.3472175002,-0.0850819349,0.2574238479,-0.0885448754,-0.2285953015,0.3524973392,-0.1272809058,-0.0153178964,-0.0792646483,-0.0165644158,0.437337935,0.1797020435,0.0240603052,-0.0108518275,-0.0252295118,0.2277865112,-0.0137914531,0.3982838094,-0.0115693389,0.121883519,0.0862016305,0.8433319926,0.1893257499,-0.3235135376,0.3949736059,0.3149036169,-0.0733484626,-0.256616503,-0.3480521142,-0.0821666047,-0.0401236229,0.1737650484,0.223294124,-0.0486487336,-0.3717193007,-0.0359347127,-0.1308296472,-0.1652549207,-0.3224005401,0.2523370087,-0.1592554748,0.2457666844,0.0005499725,0.0926778764,-0.2677298188,0.0144890798,0.0368714817,0.2266474813,0.1431090385,0.0896599293,-0.5136061907,-0.0663289055,-0.4781812429,0.2796219587,0.1847020835,0.564268887,-0.1529383361,-0.168003425,0.1045318618,-0.0186034441,0.7028640509,0.02301687,-0.0897498056,0.1123473346,-0.2468478978,-0.3765416741,0.0356233455,-0.045827385,0.1709315032,0.3891173005,0.5012796521,-0.4591732621,-0.3562014401,0.0085638128,-0.0293875299,-0.0819561556,-0.201848805,-0.186429441,0.0265050717,-0.2668227553,0.1215890199,-0.0974667743,0.3548803031,0.0250712298,0.2982490361,-0.0434097834,-0.1342862248,-0.1364925653,0.1130342782,0.1991302222,-0.0183561854,-0.0332275294,0.3735995293,0.1696552038,-0.2558128536,0.270385325,0.0274356678,-0.2911652327,0.0742891133,0.2565268576,0.1025480554,0.2148583978,0.1469199359,0.2874315977,0.1137224138,0.0473645777,0.0363879763,0.226997897,0.2227705419,0.1510111094,-0.3832801878,-0.2247336209,0.4746959507,-0.2865239978,0.0148015302,0.2330476195,-0.1075512543,-0.3312224448,0.6692470908,-0.021685563,1.0043586493,0.1279561073,0.2349968553,0.3570190072,0.0152032506,0.4015209675,0.075397931,0.092209965,-0.4636116326,0.1191376597,-0.036317274,-0.0786606148,0.1102758348,0.2392281592,-0.1544450521,0.1322183609,-0.0604646206,-0.2274056077,0.1711773425,0.4276742339,0.1397120804,-0.4334684908,-0.2043327689,0.0076698088,-0.266140908,0.5549460649,-0.0314720273,-0.2316812873,-0.0094632655,-0.3618118465,-0.2293406725,0.301956892,-0.0994984582,0.3744971454,0.600733459,-0.1288872063,0.4571174383,0.1026240066,0.2113724202,0.0299900398,-0.2630378902,0.1922145486,-0.0337365195,-0.0982527435,0.0589332096,0.0264072064,0.356651932,0.0146899167,-0.2336515188,-0.0800436065,-0.1291650534,-0.2312926501,-0.0259570889,-0.1611709595,0.019038491,-0.1472074986,-0.3155699968,0.0277754236,0.0255545061,-0.1649843454,0.0687859654,0.0484742932,-0.2449036688,-0.2078677565,0.3131470978,-0.2475472242,-0.0739669204,0.4810358882,-0.0992319584,-0.0194906276,0.6444162726,0.3653110266,-0.206257537,-0.1943459809,-0.0836617425,0.1650459766,-0.8099958897,-0.0325143486,-0.001692626,-0.2046627253,-0.0655456856,0.5908412933,0.2419272214,0.0629566014,-0.1193100661,-0.3331790566,-0.3890790939,0.2196124196,-0.222279042,0.1357290745,0.0955775231,0.3911072314,-0.0347451828,0.1781407148,-0.2182646394,-0.0300164614,-0.2331596166,0.252122879,0.2451464981,-0.4850661159,0.2257371098,0.0978112072,0.0273082107,0.1323976815,-0.3137067258,-0.1233846545,-0.4189220667,0.1245738044,0.2815958261,-0.2863855958,-0.0942837596,0.1666618735,0.1116882786,-0.1594538987,0.0924320891,-0.0615465902,-0.1313707232,0.2736981511,0.1938243657,0.0533949435,-0.0768333897,0.0873866379,-0.1187309399,-0.313446641,0.2363160253,0.2007078677,-0.1965818852,0.0685271248,-0.2107036561,0.0794220269,0.3125178814,0.1303780377,0.0038125271,-0.3625322282,0.179345876,-0.036080543,0.0760999322,0.031151671,-0.1425473988,-0.2864573002,0.4078353047,0.1651120633,-0.2894612849,-0.2251743972,0.0795918107,0.0420671366,-0.0174697917,0.0502241552,0.0443150178,0.0742180571,0.0834893435,0.0840106532,0.4601431787,-0.2134629041,0.2142047435,0.3518062234,-0.1893409789,0.1552309543,0.464518249,0.0193825513,0.1395028383,0.2465680093,-0.2924775779,0.2774638236,-0.26682657,0.290597111,0.0077866502,-0.5513201952,-0.1049560905,0.2481084913,0.1737886071,0.1777735054,-0.066557765,0.445124656,-0.1363940984,-0.1493584365,-0.048861865,0.3371517658,-0.2716991603,-0.3933604956,-0.1717560142,0.1165925711,-0.394910574,-0.0003244607,-0.1727865338,-0.0516136475,0.156663537,0.1176726744,-0.105841279,-0.3330611587,-0.0519959256,0.0413667373,-0.1215798855,-0.2785531878,0.3481361866,-0.0297522619,0.0341189764,-0.1454355419,0.133100912,0.286752671,0.2867029905,-0.3025152385,-0.1334831566,-0.1729675233,0.1225762069,-0.030438317,0.3490459025,0.1919632107,0.4685676694,0.3146248162,0.0043297336,-0.156321317,-0.1256596297,0.1083557904,0.1264597625,-0.0214302167,0.5825879574,-0.1685422212,-0.0980616063,-0.260268867,0.1421924382,-0.4087181091,0.3401703238,0.2832738459,0.0236689746,0.1059991643,-0.0358005501,0.043437358,0.1233691052,0.5698880553,0.3689297438,0.1243647039,-0.5338953137,-0.0682754219,-0.4477844834,0.082789883,0.193966046,0.1000257283,0.0117921755,0.0065873745,0.0200608894,-0.0266447198,-0.3352196217,0.3204243779,0.0809512511,0.0154610472,-0.0622798093,0.1577226818,-0.2189983875,0.0669474602,0.0925963596,-0.4667535424,0.0259801149,0.1531939656,-0.0335598029,-0.2191779166,-0.1638471335,0.0668924451,0.0369472876,-0.0208345447,0.1326977015,0.7262605429,-0.0420248248,-0.1220177785,-0.4175515175,-0.4629760683,-0.1109927222,0.0634776875,-0.1209399998,0.6496319771,-0.0621025972,0.3009475768,-0.4772535264,0.1635540873,-0.1076173261,-0.1320490986,-0.3146496117,0.1489403099,-0.2459841222,-0.0108380876,0.1265085787,0.3707412183,0.1228420958,0.0308370795,-0.3600005805,-0.2014899254,0.6447339058,-0.7064890862,-0.1418036073,-0.091070056,0.0048363954,0.2531350553,-0.2619392574,-0.7180592418,-0.0282277595,0.1826342046,-0.0192403179,-0.1775765568,0.0684963912,0.230887115,-0.0765093267,-0.1325629354,0.656732142,0.1076977178,-0.2209262699,0.1161853224,-0.1894451827]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/600","title":"Pickling error when loading dataset","comments":"Does it work when you install `nlp` from source on python 3.6?","body":"Hi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```","comment_length":12,"text":"Pickling error when loading dataset\nHi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\nDoes it work when you install `nlp` from source on python 3.6?","embeddings":[-0.2029621452,-0.1918641627,0.1510746926,0.2650444508,0.1417573392,-0.1736165881,0.2733694315,0.3465311229,0.0597782321,-0.0787867382,0.0942319855,0.3698832691,-0.2363304645,0.0360416137,0.0861706808,-0.4138222039,-0.0133403046,0.1228954196,-0.172864154,0.0111276107,-0.1992149353,0.319937855,-0.2243003845,0.1471229792,-0.505433321,-0.0055762944,0.075616017,0.1886179298,-0.0382539779,-0.3906199336,0.357863754,-0.0553878881,0.2542710304,0.4166387916,-0.0001185715,0.1009830087,0.4450067282,-0.1613719016,-0.2186082751,-0.5004016757,0.3419595361,-0.1968306452,0.1952064484,-0.119422853,-0.138964057,0.0235723462,0.0966473296,0.0984587669,0.3546349704,0.2370629162,0.2198283076,0.7399469614,0.0196053255,-0.004001346,-0.1397824585,0.3455076218,-0.0702364743,0.1682084203,0.0740607306,0.0049477555,-0.250533253,0.2688044012,0.0097407419,-0.1106548905,0.0340497196,0.0430935025,-0.0089143366,-0.0071578994,0.2315981686,0.1808640808,0.3163917959,-0.1332210153,-0.4018171132,-0.4381490052,-0.0321017392,-0.2538236082,0.3569175601,0.0049056802,-0.1719962656,-0.0149831194,-0.1687421799,-0.2313759178,0.0281378422,0.3235939145,0.1266870797,0.6207981706,-0.0286829732,0.2285003811,0.3370732367,-0.188007012,0.1313721091,-0.0277588796,0.0349624828,0.4023726285,-0.238850221,-0.1197175607,0.0664065331,-0.1022803336,0.040619906,0.0130149191,-0.3214137554,0.0603422858,-0.2768704593,-0.0223208498,0.2950415611,0.2889094651,0.0581622198,0.5016816258,-0.0262932908,-0.1388157159,0.2159735262,0.0168195665,-0.124156177,-0.281819135,-0.1952973008,0.0977016166,0.0427054726,-0.2523407638,-0.0277266987,-0.243488133,-0.0807851404,0.0121760387,0.1265031397,0.482233882,0.0803427994,0.1158533469,0.0164868422,0.1796439886,-0.4184333384,-0.2533637583,-0.1892677546,0.022043325,-0.3372351229,-0.1492154151,0.2223524898,0.1535892189,0.3354985416,-0.1102682278,-0.0604240373,0.1364385486,-0.0179690253,-0.1492116302,-0.0685894415,0.2175128609,-0.2264653891,0.1727866828,0.2571268976,-0.1251242161,-0.2969128489,0.1636125892,-0.0111897727,-0.2530519068,-0.1209304109,0.0851275399,-0.1824117154,-0.2599703074,-0.3591947556,0.2333444357,0.3632029891,-0.263068974,-0.1501649469,-0.3836107254,-0.1859427094,-0.1210468039,0.1095169187,0.6124694347,-0.1049165651,-0.4261669219,-0.1083565578,-0.1307385713,0.386395663,0.4650295377,-0.4532147944,0.153607592,-0.2186063975,0.5694765449,0.383651495,-0.1736365557,-0.2988277078,0.2628951669,-0.2244315892,0.2072925121,-0.0707563609,-0.1237978861,-0.1637935787,0.0578438863,-0.1647842675,0.3448475599,0.1585469395,0.2666162848,-0.2812247872,-0.2925919592,0.4360112846,0.2920269966,0.216148749,0.1050296426,-0.1254457384,0.3208654523,0.2130340934,-0.2217202485,0.191267103,0.0267760381,0.0222752038,-0.1574062705,-0.032853514,0.0872845724,-0.6515653729,0.1748918146,-0.2471687347,0.3482395709,-0.3800010085,0.1222224906,-0.0670937449,-0.1169426218,-0.3418495059,-0.2125000358,0.0444627553,0.183524698,-0.0876024067,-0.0400484204,-0.0541984327,0.0558037721,0.0123563483,0.1081894338,-0.4298703969,0.1635243893,-0.1746212542,-0.3540358245,-0.0586851798,0.2108436823,0.2123282403,-0.2704420388,-0.2660779059,0.2112212032,0.0572887398,0.2215671539,0.0190088879,-0.00503748,0.1972181201,-0.192833066,0.0177532136,0.3558892608,0.0182690863,0.0179571509,0.0842206031,0.3661319017,0.0188044794,0.0989182368,0.1062778458,0.0629072785,0.1784609258,-0.0767172575,0.163967669,-0.1630000323,0.300399065,0.0798098519,0.4254424274,0.0218044408,-0.336406678,-0.2099469602,0.4499590993,0.0788158178,0.1665172726,0.3088004887,-0.0811552033,-0.1771025658,-0.0583941601,0.0550881997,0.2792567611,0.0804542229,-0.0513546728,0.0372813307,-0.0395417921,-0.1005213484,0.1776282787,0.101741232,-0.0386786871,0.3335563242,-0.0069130347,0.0972452164,-0.4322930276,-0.2849202156,0.1184080318,0.3154613972,-0.2130566686,0.1573490947,-0.1570317745,-0.0666516423,-0.4013566077,-0.2381341457,-0.4762080014,-0.1907107383,-0.2355116904,0.1614567786,-0.1278426498,0.2946344018,-0.0560883917,0.0913452059,0.0177859832,-0.0834005475,-0.1864178628,-0.2087877989,-0.4619050622,-0.0359911546,0.2449781895,0.021026792,0.1043828949,0.123727344,-0.0020178452,0.1463040262,-0.3587393761,0.2713183761,0.0121457307,0.187657088,0.1870482713,0.0780553222,-0.1918313205,-0.4377750158,0.2089122981,-0.062545836,-0.2025256455,0.0551075153,0.0965651721,-0.0389919728,-0.1555091143,-0.3513694406,-0.3187605441,-0.4107227623,0.1541299969,-0.1068721488,0.192685917,0.6526251435,-0.0405006818,0.3030754626,-0.23926799,0.0254845191,-0.1506410986,0.1499269158,0.0620459914,-0.1284027398,-0.3454410732,-0.3472175002,-0.0850819349,0.2574238479,-0.0885448754,-0.2285953015,0.3524973392,-0.1272809058,-0.0153178964,-0.0792646483,-0.0165644158,0.437337935,0.1797020435,0.0240603052,-0.0108518275,-0.0252295118,0.2277865112,-0.0137914531,0.3982838094,-0.0115693389,0.121883519,0.0862016305,0.8433319926,0.1893257499,-0.3235135376,0.3949736059,0.3149036169,-0.0733484626,-0.256616503,-0.3480521142,-0.0821666047,-0.0401236229,0.1737650484,0.223294124,-0.0486487336,-0.3717193007,-0.0359347127,-0.1308296472,-0.1652549207,-0.3224005401,0.2523370087,-0.1592554748,0.2457666844,0.0005499725,0.0926778764,-0.2677298188,0.0144890798,0.0368714817,0.2266474813,0.1431090385,0.0896599293,-0.5136061907,-0.0663289055,-0.4781812429,0.2796219587,0.1847020835,0.564268887,-0.1529383361,-0.168003425,0.1045318618,-0.0186034441,0.7028640509,0.02301687,-0.0897498056,0.1123473346,-0.2468478978,-0.3765416741,0.0356233455,-0.045827385,0.1709315032,0.3891173005,0.5012796521,-0.4591732621,-0.3562014401,0.0085638128,-0.0293875299,-0.0819561556,-0.201848805,-0.186429441,0.0265050717,-0.2668227553,0.1215890199,-0.0974667743,0.3548803031,0.0250712298,0.2982490361,-0.0434097834,-0.1342862248,-0.1364925653,0.1130342782,0.1991302222,-0.0183561854,-0.0332275294,0.3735995293,0.1696552038,-0.2558128536,0.270385325,0.0274356678,-0.2911652327,0.0742891133,0.2565268576,0.1025480554,0.2148583978,0.1469199359,0.2874315977,0.1137224138,0.0473645777,0.0363879763,0.226997897,0.2227705419,0.1510111094,-0.3832801878,-0.2247336209,0.4746959507,-0.2865239978,0.0148015302,0.2330476195,-0.1075512543,-0.3312224448,0.6692470908,-0.021685563,1.0043586493,0.1279561073,0.2349968553,0.3570190072,0.0152032506,0.4015209675,0.075397931,0.092209965,-0.4636116326,0.1191376597,-0.036317274,-0.0786606148,0.1102758348,0.2392281592,-0.1544450521,0.1322183609,-0.0604646206,-0.2274056077,0.1711773425,0.4276742339,0.1397120804,-0.4334684908,-0.2043327689,0.0076698088,-0.266140908,0.5549460649,-0.0314720273,-0.2316812873,-0.0094632655,-0.3618118465,-0.2293406725,0.301956892,-0.0994984582,0.3744971454,0.600733459,-0.1288872063,0.4571174383,0.1026240066,0.2113724202,0.0299900398,-0.2630378902,0.1922145486,-0.0337365195,-0.0982527435,0.0589332096,0.0264072064,0.356651932,0.0146899167,-0.2336515188,-0.0800436065,-0.1291650534,-0.2312926501,-0.0259570889,-0.1611709595,0.019038491,-0.1472074986,-0.3155699968,0.0277754236,0.0255545061,-0.1649843454,0.0687859654,0.0484742932,-0.2449036688,-0.2078677565,0.3131470978,-0.2475472242,-0.0739669204,0.4810358882,-0.0992319584,-0.0194906276,0.6444162726,0.3653110266,-0.206257537,-0.1943459809,-0.0836617425,0.1650459766,-0.8099958897,-0.0325143486,-0.001692626,-0.2046627253,-0.0655456856,0.5908412933,0.2419272214,0.0629566014,-0.1193100661,-0.3331790566,-0.3890790939,0.2196124196,-0.222279042,0.1357290745,0.0955775231,0.3911072314,-0.0347451828,0.1781407148,-0.2182646394,-0.0300164614,-0.2331596166,0.252122879,0.2451464981,-0.4850661159,0.2257371098,0.0978112072,0.0273082107,0.1323976815,-0.3137067258,-0.1233846545,-0.4189220667,0.1245738044,0.2815958261,-0.2863855958,-0.0942837596,0.1666618735,0.1116882786,-0.1594538987,0.0924320891,-0.0615465902,-0.1313707232,0.2736981511,0.1938243657,0.0533949435,-0.0768333897,0.0873866379,-0.1187309399,-0.313446641,0.2363160253,0.2007078677,-0.1965818852,0.0685271248,-0.2107036561,0.0794220269,0.3125178814,0.1303780377,0.0038125271,-0.3625322282,0.179345876,-0.036080543,0.0760999322,0.031151671,-0.1425473988,-0.2864573002,0.4078353047,0.1651120633,-0.2894612849,-0.2251743972,0.0795918107,0.0420671366,-0.0174697917,0.0502241552,0.0443150178,0.0742180571,0.0834893435,0.0840106532,0.4601431787,-0.2134629041,0.2142047435,0.3518062234,-0.1893409789,0.1552309543,0.464518249,0.0193825513,0.1395028383,0.2465680093,-0.2924775779,0.2774638236,-0.26682657,0.290597111,0.0077866502,-0.5513201952,-0.1049560905,0.2481084913,0.1737886071,0.1777735054,-0.066557765,0.445124656,-0.1363940984,-0.1493584365,-0.048861865,0.3371517658,-0.2716991603,-0.3933604956,-0.1717560142,0.1165925711,-0.394910574,-0.0003244607,-0.1727865338,-0.0516136475,0.156663537,0.1176726744,-0.105841279,-0.3330611587,-0.0519959256,0.0413667373,-0.1215798855,-0.2785531878,0.3481361866,-0.0297522619,0.0341189764,-0.1454355419,0.133100912,0.286752671,0.2867029905,-0.3025152385,-0.1334831566,-0.1729675233,0.1225762069,-0.030438317,0.3490459025,0.1919632107,0.4685676694,0.3146248162,0.0043297336,-0.156321317,-0.1256596297,0.1083557904,0.1264597625,-0.0214302167,0.5825879574,-0.1685422212,-0.0980616063,-0.260268867,0.1421924382,-0.4087181091,0.3401703238,0.2832738459,0.0236689746,0.1059991643,-0.0358005501,0.043437358,0.1233691052,0.5698880553,0.3689297438,0.1243647039,-0.5338953137,-0.0682754219,-0.4477844834,0.082789883,0.193966046,0.1000257283,0.0117921755,0.0065873745,0.0200608894,-0.0266447198,-0.3352196217,0.3204243779,0.0809512511,0.0154610472,-0.0622798093,0.1577226818,-0.2189983875,0.0669474602,0.0925963596,-0.4667535424,0.0259801149,0.1531939656,-0.0335598029,-0.2191779166,-0.1638471335,0.0668924451,0.0369472876,-0.0208345447,0.1326977015,0.7262605429,-0.0420248248,-0.1220177785,-0.4175515175,-0.4629760683,-0.1109927222,0.0634776875,-0.1209399998,0.6496319771,-0.0621025972,0.3009475768,-0.4772535264,0.1635540873,-0.1076173261,-0.1320490986,-0.3146496117,0.1489403099,-0.2459841222,-0.0108380876,0.1265085787,0.3707412183,0.1228420958,0.0308370795,-0.3600005805,-0.2014899254,0.6447339058,-0.7064890862,-0.1418036073,-0.091070056,0.0048363954,0.2531350553,-0.2619392574,-0.7180592418,-0.0282277595,0.1826342046,-0.0192403179,-0.1775765568,0.0684963912,0.230887115,-0.0765093267,-0.1325629354,0.656732142,0.1076977178,-0.2209262699,0.1161853224,-0.1894451827]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/600","title":"Pickling error when loading dataset","comments":"No, still the pickling error.","body":"Hi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```","comment_length":5,"text":"Pickling error when loading dataset\nHi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\nNo, still the pickling error.","embeddings":[-0.2029621452,-0.1918641627,0.1510746926,0.2650444508,0.1417573392,-0.1736165881,0.2733694315,0.3465311229,0.0597782321,-0.0787867382,0.0942319855,0.3698832691,-0.2363304645,0.0360416137,0.0861706808,-0.4138222039,-0.0133403046,0.1228954196,-0.172864154,0.0111276107,-0.1992149353,0.319937855,-0.2243003845,0.1471229792,-0.505433321,-0.0055762944,0.075616017,0.1886179298,-0.0382539779,-0.3906199336,0.357863754,-0.0553878881,0.2542710304,0.4166387916,-0.0001185715,0.1009830087,0.4450067282,-0.1613719016,-0.2186082751,-0.5004016757,0.3419595361,-0.1968306452,0.1952064484,-0.119422853,-0.138964057,0.0235723462,0.0966473296,0.0984587669,0.3546349704,0.2370629162,0.2198283076,0.7399469614,0.0196053255,-0.004001346,-0.1397824585,0.3455076218,-0.0702364743,0.1682084203,0.0740607306,0.0049477555,-0.250533253,0.2688044012,0.0097407419,-0.1106548905,0.0340497196,0.0430935025,-0.0089143366,-0.0071578994,0.2315981686,0.1808640808,0.3163917959,-0.1332210153,-0.4018171132,-0.4381490052,-0.0321017392,-0.2538236082,0.3569175601,0.0049056802,-0.1719962656,-0.0149831194,-0.1687421799,-0.2313759178,0.0281378422,0.3235939145,0.1266870797,0.6207981706,-0.0286829732,0.2285003811,0.3370732367,-0.188007012,0.1313721091,-0.0277588796,0.0349624828,0.4023726285,-0.238850221,-0.1197175607,0.0664065331,-0.1022803336,0.040619906,0.0130149191,-0.3214137554,0.0603422858,-0.2768704593,-0.0223208498,0.2950415611,0.2889094651,0.0581622198,0.5016816258,-0.0262932908,-0.1388157159,0.2159735262,0.0168195665,-0.124156177,-0.281819135,-0.1952973008,0.0977016166,0.0427054726,-0.2523407638,-0.0277266987,-0.243488133,-0.0807851404,0.0121760387,0.1265031397,0.482233882,0.0803427994,0.1158533469,0.0164868422,0.1796439886,-0.4184333384,-0.2533637583,-0.1892677546,0.022043325,-0.3372351229,-0.1492154151,0.2223524898,0.1535892189,0.3354985416,-0.1102682278,-0.0604240373,0.1364385486,-0.0179690253,-0.1492116302,-0.0685894415,0.2175128609,-0.2264653891,0.1727866828,0.2571268976,-0.1251242161,-0.2969128489,0.1636125892,-0.0111897727,-0.2530519068,-0.1209304109,0.0851275399,-0.1824117154,-0.2599703074,-0.3591947556,0.2333444357,0.3632029891,-0.263068974,-0.1501649469,-0.3836107254,-0.1859427094,-0.1210468039,0.1095169187,0.6124694347,-0.1049165651,-0.4261669219,-0.1083565578,-0.1307385713,0.386395663,0.4650295377,-0.4532147944,0.153607592,-0.2186063975,0.5694765449,0.383651495,-0.1736365557,-0.2988277078,0.2628951669,-0.2244315892,0.2072925121,-0.0707563609,-0.1237978861,-0.1637935787,0.0578438863,-0.1647842675,0.3448475599,0.1585469395,0.2666162848,-0.2812247872,-0.2925919592,0.4360112846,0.2920269966,0.216148749,0.1050296426,-0.1254457384,0.3208654523,0.2130340934,-0.2217202485,0.191267103,0.0267760381,0.0222752038,-0.1574062705,-0.032853514,0.0872845724,-0.6515653729,0.1748918146,-0.2471687347,0.3482395709,-0.3800010085,0.1222224906,-0.0670937449,-0.1169426218,-0.3418495059,-0.2125000358,0.0444627553,0.183524698,-0.0876024067,-0.0400484204,-0.0541984327,0.0558037721,0.0123563483,0.1081894338,-0.4298703969,0.1635243893,-0.1746212542,-0.3540358245,-0.0586851798,0.2108436823,0.2123282403,-0.2704420388,-0.2660779059,0.2112212032,0.0572887398,0.2215671539,0.0190088879,-0.00503748,0.1972181201,-0.192833066,0.0177532136,0.3558892608,0.0182690863,0.0179571509,0.0842206031,0.3661319017,0.0188044794,0.0989182368,0.1062778458,0.0629072785,0.1784609258,-0.0767172575,0.163967669,-0.1630000323,0.300399065,0.0798098519,0.4254424274,0.0218044408,-0.336406678,-0.2099469602,0.4499590993,0.0788158178,0.1665172726,0.3088004887,-0.0811552033,-0.1771025658,-0.0583941601,0.0550881997,0.2792567611,0.0804542229,-0.0513546728,0.0372813307,-0.0395417921,-0.1005213484,0.1776282787,0.101741232,-0.0386786871,0.3335563242,-0.0069130347,0.0972452164,-0.4322930276,-0.2849202156,0.1184080318,0.3154613972,-0.2130566686,0.1573490947,-0.1570317745,-0.0666516423,-0.4013566077,-0.2381341457,-0.4762080014,-0.1907107383,-0.2355116904,0.1614567786,-0.1278426498,0.2946344018,-0.0560883917,0.0913452059,0.0177859832,-0.0834005475,-0.1864178628,-0.2087877989,-0.4619050622,-0.0359911546,0.2449781895,0.021026792,0.1043828949,0.123727344,-0.0020178452,0.1463040262,-0.3587393761,0.2713183761,0.0121457307,0.187657088,0.1870482713,0.0780553222,-0.1918313205,-0.4377750158,0.2089122981,-0.062545836,-0.2025256455,0.0551075153,0.0965651721,-0.0389919728,-0.1555091143,-0.3513694406,-0.3187605441,-0.4107227623,0.1541299969,-0.1068721488,0.192685917,0.6526251435,-0.0405006818,0.3030754626,-0.23926799,0.0254845191,-0.1506410986,0.1499269158,0.0620459914,-0.1284027398,-0.3454410732,-0.3472175002,-0.0850819349,0.2574238479,-0.0885448754,-0.2285953015,0.3524973392,-0.1272809058,-0.0153178964,-0.0792646483,-0.0165644158,0.437337935,0.1797020435,0.0240603052,-0.0108518275,-0.0252295118,0.2277865112,-0.0137914531,0.3982838094,-0.0115693389,0.121883519,0.0862016305,0.8433319926,0.1893257499,-0.3235135376,0.3949736059,0.3149036169,-0.0733484626,-0.256616503,-0.3480521142,-0.0821666047,-0.0401236229,0.1737650484,0.223294124,-0.0486487336,-0.3717193007,-0.0359347127,-0.1308296472,-0.1652549207,-0.3224005401,0.2523370087,-0.1592554748,0.2457666844,0.0005499725,0.0926778764,-0.2677298188,0.0144890798,0.0368714817,0.2266474813,0.1431090385,0.0896599293,-0.5136061907,-0.0663289055,-0.4781812429,0.2796219587,0.1847020835,0.564268887,-0.1529383361,-0.168003425,0.1045318618,-0.0186034441,0.7028640509,0.02301687,-0.0897498056,0.1123473346,-0.2468478978,-0.3765416741,0.0356233455,-0.045827385,0.1709315032,0.3891173005,0.5012796521,-0.4591732621,-0.3562014401,0.0085638128,-0.0293875299,-0.0819561556,-0.201848805,-0.186429441,0.0265050717,-0.2668227553,0.1215890199,-0.0974667743,0.3548803031,0.0250712298,0.2982490361,-0.0434097834,-0.1342862248,-0.1364925653,0.1130342782,0.1991302222,-0.0183561854,-0.0332275294,0.3735995293,0.1696552038,-0.2558128536,0.270385325,0.0274356678,-0.2911652327,0.0742891133,0.2565268576,0.1025480554,0.2148583978,0.1469199359,0.2874315977,0.1137224138,0.0473645777,0.0363879763,0.226997897,0.2227705419,0.1510111094,-0.3832801878,-0.2247336209,0.4746959507,-0.2865239978,0.0148015302,0.2330476195,-0.1075512543,-0.3312224448,0.6692470908,-0.021685563,1.0043586493,0.1279561073,0.2349968553,0.3570190072,0.0152032506,0.4015209675,0.075397931,0.092209965,-0.4636116326,0.1191376597,-0.036317274,-0.0786606148,0.1102758348,0.2392281592,-0.1544450521,0.1322183609,-0.0604646206,-0.2274056077,0.1711773425,0.4276742339,0.1397120804,-0.4334684908,-0.2043327689,0.0076698088,-0.266140908,0.5549460649,-0.0314720273,-0.2316812873,-0.0094632655,-0.3618118465,-0.2293406725,0.301956892,-0.0994984582,0.3744971454,0.600733459,-0.1288872063,0.4571174383,0.1026240066,0.2113724202,0.0299900398,-0.2630378902,0.1922145486,-0.0337365195,-0.0982527435,0.0589332096,0.0264072064,0.356651932,0.0146899167,-0.2336515188,-0.0800436065,-0.1291650534,-0.2312926501,-0.0259570889,-0.1611709595,0.019038491,-0.1472074986,-0.3155699968,0.0277754236,0.0255545061,-0.1649843454,0.0687859654,0.0484742932,-0.2449036688,-0.2078677565,0.3131470978,-0.2475472242,-0.0739669204,0.4810358882,-0.0992319584,-0.0194906276,0.6444162726,0.3653110266,-0.206257537,-0.1943459809,-0.0836617425,0.1650459766,-0.8099958897,-0.0325143486,-0.001692626,-0.2046627253,-0.0655456856,0.5908412933,0.2419272214,0.0629566014,-0.1193100661,-0.3331790566,-0.3890790939,0.2196124196,-0.222279042,0.1357290745,0.0955775231,0.3911072314,-0.0347451828,0.1781407148,-0.2182646394,-0.0300164614,-0.2331596166,0.252122879,0.2451464981,-0.4850661159,0.2257371098,0.0978112072,0.0273082107,0.1323976815,-0.3137067258,-0.1233846545,-0.4189220667,0.1245738044,0.2815958261,-0.2863855958,-0.0942837596,0.1666618735,0.1116882786,-0.1594538987,0.0924320891,-0.0615465902,-0.1313707232,0.2736981511,0.1938243657,0.0533949435,-0.0768333897,0.0873866379,-0.1187309399,-0.313446641,0.2363160253,0.2007078677,-0.1965818852,0.0685271248,-0.2107036561,0.0794220269,0.3125178814,0.1303780377,0.0038125271,-0.3625322282,0.179345876,-0.036080543,0.0760999322,0.031151671,-0.1425473988,-0.2864573002,0.4078353047,0.1651120633,-0.2894612849,-0.2251743972,0.0795918107,0.0420671366,-0.0174697917,0.0502241552,0.0443150178,0.0742180571,0.0834893435,0.0840106532,0.4601431787,-0.2134629041,0.2142047435,0.3518062234,-0.1893409789,0.1552309543,0.464518249,0.0193825513,0.1395028383,0.2465680093,-0.2924775779,0.2774638236,-0.26682657,0.290597111,0.0077866502,-0.5513201952,-0.1049560905,0.2481084913,0.1737886071,0.1777735054,-0.066557765,0.445124656,-0.1363940984,-0.1493584365,-0.048861865,0.3371517658,-0.2716991603,-0.3933604956,-0.1717560142,0.1165925711,-0.394910574,-0.0003244607,-0.1727865338,-0.0516136475,0.156663537,0.1176726744,-0.105841279,-0.3330611587,-0.0519959256,0.0413667373,-0.1215798855,-0.2785531878,0.3481361866,-0.0297522619,0.0341189764,-0.1454355419,0.133100912,0.286752671,0.2867029905,-0.3025152385,-0.1334831566,-0.1729675233,0.1225762069,-0.030438317,0.3490459025,0.1919632107,0.4685676694,0.3146248162,0.0043297336,-0.156321317,-0.1256596297,0.1083557904,0.1264597625,-0.0214302167,0.5825879574,-0.1685422212,-0.0980616063,-0.260268867,0.1421924382,-0.4087181091,0.3401703238,0.2832738459,0.0236689746,0.1059991643,-0.0358005501,0.043437358,0.1233691052,0.5698880553,0.3689297438,0.1243647039,-0.5338953137,-0.0682754219,-0.4477844834,0.082789883,0.193966046,0.1000257283,0.0117921755,0.0065873745,0.0200608894,-0.0266447198,-0.3352196217,0.3204243779,0.0809512511,0.0154610472,-0.0622798093,0.1577226818,-0.2189983875,0.0669474602,0.0925963596,-0.4667535424,0.0259801149,0.1531939656,-0.0335598029,-0.2191779166,-0.1638471335,0.0668924451,0.0369472876,-0.0208345447,0.1326977015,0.7262605429,-0.0420248248,-0.1220177785,-0.4175515175,-0.4629760683,-0.1109927222,0.0634776875,-0.1209399998,0.6496319771,-0.0621025972,0.3009475768,-0.4772535264,0.1635540873,-0.1076173261,-0.1320490986,-0.3146496117,0.1489403099,-0.2459841222,-0.0108380876,0.1265085787,0.3707412183,0.1228420958,0.0308370795,-0.3600005805,-0.2014899254,0.6447339058,-0.7064890862,-0.1418036073,-0.091070056,0.0048363954,0.2531350553,-0.2619392574,-0.7180592418,-0.0282277595,0.1826342046,-0.0192403179,-0.1775765568,0.0684963912,0.230887115,-0.0765093267,-0.1325629354,0.656732142,0.1076977178,-0.2209262699,0.1161853224,-0.1894451827]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/600","title":"Pickling error when loading dataset","comments":"I wasn't able to reproduce on google colab (python 3.6.9 as well) with \r\n\r\npickle==4.0\r\ndill=0.3.2\r\ntransformers==3.1.0\r\ndatasets=1.0.1 (also tried nlp 0.4.0)\r\n\r\nIf I try\r\n\r\n```python\r\nfrom datasets import load_dataset # or from nlp\r\nfrom transformers import BertTokenizer\r\n\r\ntokenizer = BertTokenizer.from_pretrained(\"bert-base-uncased\")\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=512), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\n```\r\nIt runs without error","body":"Hi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```","comment_length":61,"text":"Pickling error when loading dataset\nHi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\nI wasn't able to reproduce on google colab (python 3.6.9 as well) with \r\n\r\npickle==4.0\r\ndill=0.3.2\r\ntransformers==3.1.0\r\ndatasets=1.0.1 (also tried nlp 0.4.0)\r\n\r\nIf I try\r\n\r\n```python\r\nfrom datasets import load_dataset # or from nlp\r\nfrom transformers import BertTokenizer\r\n\r\ntokenizer = BertTokenizer.from_pretrained(\"bert-base-uncased\")\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=512), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\n```\r\nIt runs without error","embeddings":[-0.2029621452,-0.1918641627,0.1510746926,0.2650444508,0.1417573392,-0.1736165881,0.2733694315,0.3465311229,0.0597782321,-0.0787867382,0.0942319855,0.3698832691,-0.2363304645,0.0360416137,0.0861706808,-0.4138222039,-0.0133403046,0.1228954196,-0.172864154,0.0111276107,-0.1992149353,0.319937855,-0.2243003845,0.1471229792,-0.505433321,-0.0055762944,0.075616017,0.1886179298,-0.0382539779,-0.3906199336,0.357863754,-0.0553878881,0.2542710304,0.4166387916,-0.0001185715,0.1009830087,0.4450067282,-0.1613719016,-0.2186082751,-0.5004016757,0.3419595361,-0.1968306452,0.1952064484,-0.119422853,-0.138964057,0.0235723462,0.0966473296,0.0984587669,0.3546349704,0.2370629162,0.2198283076,0.7399469614,0.0196053255,-0.004001346,-0.1397824585,0.3455076218,-0.0702364743,0.1682084203,0.0740607306,0.0049477555,-0.250533253,0.2688044012,0.0097407419,-0.1106548905,0.0340497196,0.0430935025,-0.0089143366,-0.0071578994,0.2315981686,0.1808640808,0.3163917959,-0.1332210153,-0.4018171132,-0.4381490052,-0.0321017392,-0.2538236082,0.3569175601,0.0049056802,-0.1719962656,-0.0149831194,-0.1687421799,-0.2313759178,0.0281378422,0.3235939145,0.1266870797,0.6207981706,-0.0286829732,0.2285003811,0.3370732367,-0.188007012,0.1313721091,-0.0277588796,0.0349624828,0.4023726285,-0.238850221,-0.1197175607,0.0664065331,-0.1022803336,0.040619906,0.0130149191,-0.3214137554,0.0603422858,-0.2768704593,-0.0223208498,0.2950415611,0.2889094651,0.0581622198,0.5016816258,-0.0262932908,-0.1388157159,0.2159735262,0.0168195665,-0.124156177,-0.281819135,-0.1952973008,0.0977016166,0.0427054726,-0.2523407638,-0.0277266987,-0.243488133,-0.0807851404,0.0121760387,0.1265031397,0.482233882,0.0803427994,0.1158533469,0.0164868422,0.1796439886,-0.4184333384,-0.2533637583,-0.1892677546,0.022043325,-0.3372351229,-0.1492154151,0.2223524898,0.1535892189,0.3354985416,-0.1102682278,-0.0604240373,0.1364385486,-0.0179690253,-0.1492116302,-0.0685894415,0.2175128609,-0.2264653891,0.1727866828,0.2571268976,-0.1251242161,-0.2969128489,0.1636125892,-0.0111897727,-0.2530519068,-0.1209304109,0.0851275399,-0.1824117154,-0.2599703074,-0.3591947556,0.2333444357,0.3632029891,-0.263068974,-0.1501649469,-0.3836107254,-0.1859427094,-0.1210468039,0.1095169187,0.6124694347,-0.1049165651,-0.4261669219,-0.1083565578,-0.1307385713,0.386395663,0.4650295377,-0.4532147944,0.153607592,-0.2186063975,0.5694765449,0.383651495,-0.1736365557,-0.2988277078,0.2628951669,-0.2244315892,0.2072925121,-0.0707563609,-0.1237978861,-0.1637935787,0.0578438863,-0.1647842675,0.3448475599,0.1585469395,0.2666162848,-0.2812247872,-0.2925919592,0.4360112846,0.2920269966,0.216148749,0.1050296426,-0.1254457384,0.3208654523,0.2130340934,-0.2217202485,0.191267103,0.0267760381,0.0222752038,-0.1574062705,-0.032853514,0.0872845724,-0.6515653729,0.1748918146,-0.2471687347,0.3482395709,-0.3800010085,0.1222224906,-0.0670937449,-0.1169426218,-0.3418495059,-0.2125000358,0.0444627553,0.183524698,-0.0876024067,-0.0400484204,-0.0541984327,0.0558037721,0.0123563483,0.1081894338,-0.4298703969,0.1635243893,-0.1746212542,-0.3540358245,-0.0586851798,0.2108436823,0.2123282403,-0.2704420388,-0.2660779059,0.2112212032,0.0572887398,0.2215671539,0.0190088879,-0.00503748,0.1972181201,-0.192833066,0.0177532136,0.3558892608,0.0182690863,0.0179571509,0.0842206031,0.3661319017,0.0188044794,0.0989182368,0.1062778458,0.0629072785,0.1784609258,-0.0767172575,0.163967669,-0.1630000323,0.300399065,0.0798098519,0.4254424274,0.0218044408,-0.336406678,-0.2099469602,0.4499590993,0.0788158178,0.1665172726,0.3088004887,-0.0811552033,-0.1771025658,-0.0583941601,0.0550881997,0.2792567611,0.0804542229,-0.0513546728,0.0372813307,-0.0395417921,-0.1005213484,0.1776282787,0.101741232,-0.0386786871,0.3335563242,-0.0069130347,0.0972452164,-0.4322930276,-0.2849202156,0.1184080318,0.3154613972,-0.2130566686,0.1573490947,-0.1570317745,-0.0666516423,-0.4013566077,-0.2381341457,-0.4762080014,-0.1907107383,-0.2355116904,0.1614567786,-0.1278426498,0.2946344018,-0.0560883917,0.0913452059,0.0177859832,-0.0834005475,-0.1864178628,-0.2087877989,-0.4619050622,-0.0359911546,0.2449781895,0.021026792,0.1043828949,0.123727344,-0.0020178452,0.1463040262,-0.3587393761,0.2713183761,0.0121457307,0.187657088,0.1870482713,0.0780553222,-0.1918313205,-0.4377750158,0.2089122981,-0.062545836,-0.2025256455,0.0551075153,0.0965651721,-0.0389919728,-0.1555091143,-0.3513694406,-0.3187605441,-0.4107227623,0.1541299969,-0.1068721488,0.192685917,0.6526251435,-0.0405006818,0.3030754626,-0.23926799,0.0254845191,-0.1506410986,0.1499269158,0.0620459914,-0.1284027398,-0.3454410732,-0.3472175002,-0.0850819349,0.2574238479,-0.0885448754,-0.2285953015,0.3524973392,-0.1272809058,-0.0153178964,-0.0792646483,-0.0165644158,0.437337935,0.1797020435,0.0240603052,-0.0108518275,-0.0252295118,0.2277865112,-0.0137914531,0.3982838094,-0.0115693389,0.121883519,0.0862016305,0.8433319926,0.1893257499,-0.3235135376,0.3949736059,0.3149036169,-0.0733484626,-0.256616503,-0.3480521142,-0.0821666047,-0.0401236229,0.1737650484,0.223294124,-0.0486487336,-0.3717193007,-0.0359347127,-0.1308296472,-0.1652549207,-0.3224005401,0.2523370087,-0.1592554748,0.2457666844,0.0005499725,0.0926778764,-0.2677298188,0.0144890798,0.0368714817,0.2266474813,0.1431090385,0.0896599293,-0.5136061907,-0.0663289055,-0.4781812429,0.2796219587,0.1847020835,0.564268887,-0.1529383361,-0.168003425,0.1045318618,-0.0186034441,0.7028640509,0.02301687,-0.0897498056,0.1123473346,-0.2468478978,-0.3765416741,0.0356233455,-0.045827385,0.1709315032,0.3891173005,0.5012796521,-0.4591732621,-0.3562014401,0.0085638128,-0.0293875299,-0.0819561556,-0.201848805,-0.186429441,0.0265050717,-0.2668227553,0.1215890199,-0.0974667743,0.3548803031,0.0250712298,0.2982490361,-0.0434097834,-0.1342862248,-0.1364925653,0.1130342782,0.1991302222,-0.0183561854,-0.0332275294,0.3735995293,0.1696552038,-0.2558128536,0.270385325,0.0274356678,-0.2911652327,0.0742891133,0.2565268576,0.1025480554,0.2148583978,0.1469199359,0.2874315977,0.1137224138,0.0473645777,0.0363879763,0.226997897,0.2227705419,0.1510111094,-0.3832801878,-0.2247336209,0.4746959507,-0.2865239978,0.0148015302,0.2330476195,-0.1075512543,-0.3312224448,0.6692470908,-0.021685563,1.0043586493,0.1279561073,0.2349968553,0.3570190072,0.0152032506,0.4015209675,0.075397931,0.092209965,-0.4636116326,0.1191376597,-0.036317274,-0.0786606148,0.1102758348,0.2392281592,-0.1544450521,0.1322183609,-0.0604646206,-0.2274056077,0.1711773425,0.4276742339,0.1397120804,-0.4334684908,-0.2043327689,0.0076698088,-0.266140908,0.5549460649,-0.0314720273,-0.2316812873,-0.0094632655,-0.3618118465,-0.2293406725,0.301956892,-0.0994984582,0.3744971454,0.600733459,-0.1288872063,0.4571174383,0.1026240066,0.2113724202,0.0299900398,-0.2630378902,0.1922145486,-0.0337365195,-0.0982527435,0.0589332096,0.0264072064,0.356651932,0.0146899167,-0.2336515188,-0.0800436065,-0.1291650534,-0.2312926501,-0.0259570889,-0.1611709595,0.019038491,-0.1472074986,-0.3155699968,0.0277754236,0.0255545061,-0.1649843454,0.0687859654,0.0484742932,-0.2449036688,-0.2078677565,0.3131470978,-0.2475472242,-0.0739669204,0.4810358882,-0.0992319584,-0.0194906276,0.6444162726,0.3653110266,-0.206257537,-0.1943459809,-0.0836617425,0.1650459766,-0.8099958897,-0.0325143486,-0.001692626,-0.2046627253,-0.0655456856,0.5908412933,0.2419272214,0.0629566014,-0.1193100661,-0.3331790566,-0.3890790939,0.2196124196,-0.222279042,0.1357290745,0.0955775231,0.3911072314,-0.0347451828,0.1781407148,-0.2182646394,-0.0300164614,-0.2331596166,0.252122879,0.2451464981,-0.4850661159,0.2257371098,0.0978112072,0.0273082107,0.1323976815,-0.3137067258,-0.1233846545,-0.4189220667,0.1245738044,0.2815958261,-0.2863855958,-0.0942837596,0.1666618735,0.1116882786,-0.1594538987,0.0924320891,-0.0615465902,-0.1313707232,0.2736981511,0.1938243657,0.0533949435,-0.0768333897,0.0873866379,-0.1187309399,-0.313446641,0.2363160253,0.2007078677,-0.1965818852,0.0685271248,-0.2107036561,0.0794220269,0.3125178814,0.1303780377,0.0038125271,-0.3625322282,0.179345876,-0.036080543,0.0760999322,0.031151671,-0.1425473988,-0.2864573002,0.4078353047,0.1651120633,-0.2894612849,-0.2251743972,0.0795918107,0.0420671366,-0.0174697917,0.0502241552,0.0443150178,0.0742180571,0.0834893435,0.0840106532,0.4601431787,-0.2134629041,0.2142047435,0.3518062234,-0.1893409789,0.1552309543,0.464518249,0.0193825513,0.1395028383,0.2465680093,-0.2924775779,0.2774638236,-0.26682657,0.290597111,0.0077866502,-0.5513201952,-0.1049560905,0.2481084913,0.1737886071,0.1777735054,-0.066557765,0.445124656,-0.1363940984,-0.1493584365,-0.048861865,0.3371517658,-0.2716991603,-0.3933604956,-0.1717560142,0.1165925711,-0.394910574,-0.0003244607,-0.1727865338,-0.0516136475,0.156663537,0.1176726744,-0.105841279,-0.3330611587,-0.0519959256,0.0413667373,-0.1215798855,-0.2785531878,0.3481361866,-0.0297522619,0.0341189764,-0.1454355419,0.133100912,0.286752671,0.2867029905,-0.3025152385,-0.1334831566,-0.1729675233,0.1225762069,-0.030438317,0.3490459025,0.1919632107,0.4685676694,0.3146248162,0.0043297336,-0.156321317,-0.1256596297,0.1083557904,0.1264597625,-0.0214302167,0.5825879574,-0.1685422212,-0.0980616063,-0.260268867,0.1421924382,-0.4087181091,0.3401703238,0.2832738459,0.0236689746,0.1059991643,-0.0358005501,0.043437358,0.1233691052,0.5698880553,0.3689297438,0.1243647039,-0.5338953137,-0.0682754219,-0.4477844834,0.082789883,0.193966046,0.1000257283,0.0117921755,0.0065873745,0.0200608894,-0.0266447198,-0.3352196217,0.3204243779,0.0809512511,0.0154610472,-0.0622798093,0.1577226818,-0.2189983875,0.0669474602,0.0925963596,-0.4667535424,0.0259801149,0.1531939656,-0.0335598029,-0.2191779166,-0.1638471335,0.0668924451,0.0369472876,-0.0208345447,0.1326977015,0.7262605429,-0.0420248248,-0.1220177785,-0.4175515175,-0.4629760683,-0.1109927222,0.0634776875,-0.1209399998,0.6496319771,-0.0621025972,0.3009475768,-0.4772535264,0.1635540873,-0.1076173261,-0.1320490986,-0.3146496117,0.1489403099,-0.2459841222,-0.0108380876,0.1265085787,0.3707412183,0.1228420958,0.0308370795,-0.3600005805,-0.2014899254,0.6447339058,-0.7064890862,-0.1418036073,-0.091070056,0.0048363954,0.2531350553,-0.2619392574,-0.7180592418,-0.0282277595,0.1826342046,-0.0192403179,-0.1775765568,0.0684963912,0.230887115,-0.0765093267,-0.1325629354,0.656732142,0.1076977178,-0.2209262699,0.1161853224,-0.1894451827]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/600","title":"Pickling error when loading dataset","comments":"Closing since it looks like it's working on >= 3.6.9\r\nFeel free to re-open if you have other questions :)","body":"Hi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```","comment_length":20,"text":"Pickling error when loading dataset\nHi,\r\n\r\nI modified line 136 in the original [run_language_modeling.py](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py) as:\r\n\r\n```\r\n# line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with transformers (3.1.0) and nlp (0.4.0), I get the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"src\/run_language_modeling.py\", line 319, in <module>\r\n main()\r\n File \"src\/run_language_modeling.py\", line 248, in main\r\n get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None\r\n File \"src\/run_language_modeling.py\", line 139, in get_dataset\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size), batched=True)\r\n File \"\/data\/nlp\/src\/nlp\/arrow_dataset.py\", line 1136, in map\r\n new_fingerprint=new_fingerprint,\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 158, in wrapper\r\n self._fingerprint, transform, kwargs_for_fingerprint\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 105, in update_fingerprint\r\n hasher.update(transform_args[key])\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/data\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 362, in dumps\r\n dump(obj, file)\r\n File \"\/data\/nlp\/src\/nlp\/utils\/py_utils.py\", line 339, in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 446, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 409, in dump\r\n self.save(obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1438, in save_function\r\n obj.__dict__, fkwdefaults), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1170, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 736, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 521, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 605, in save_reduce\r\n save(cls)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 1365, in save_type\r\n obj.__bases__, _dict), obj=obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 610, in save_reduce\r\n save(args)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 751, in save_tuple\r\n save(element)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 476, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/site-packages\/dill\/_dill.py\", line 933, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 821, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 847, in _batch_setitems\r\n save(v)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 507, in save\r\n self.save_global(obj, rv)\r\n File \"\/root\/miniconda3\/envs\/py3.6\/lib\/python3.6\/pickle.py\", line 927, in save_global\r\n (obj, module_name, name))\r\n_pickle.PicklingError: Can't pickle typing.Union[str, NoneType]: it's not the same object as typing.Union\r\n```\nClosing since it looks like it's working on >= 3.6.9\r\nFeel free to re-open if you have other questions :)","embeddings":[-0.2029621452,-0.1918641627,0.1510746926,0.2650444508,0.1417573392,-0.1736165881,0.2733694315,0.3465311229,0.0597782321,-0.0787867382,0.0942319855,0.3698832691,-0.2363304645,0.0360416137,0.0861706808,-0.4138222039,-0.0133403046,0.1228954196,-0.172864154,0.0111276107,-0.1992149353,0.319937855,-0.2243003845,0.1471229792,-0.505433321,-0.0055762944,0.075616017,0.1886179298,-0.0382539779,-0.3906199336,0.357863754,-0.0553878881,0.2542710304,0.4166387916,-0.0001185715,0.1009830087,0.4450067282,-0.1613719016,-0.2186082751,-0.5004016757,0.3419595361,-0.1968306452,0.1952064484,-0.119422853,-0.138964057,0.0235723462,0.0966473296,0.0984587669,0.3546349704,0.2370629162,0.2198283076,0.7399469614,0.0196053255,-0.004001346,-0.1397824585,0.3455076218,-0.0702364743,0.1682084203,0.0740607306,0.0049477555,-0.250533253,0.2688044012,0.0097407419,-0.1106548905,0.0340497196,0.0430935025,-0.0089143366,-0.0071578994,0.2315981686,0.1808640808,0.3163917959,-0.1332210153,-0.4018171132,-0.4381490052,-0.0321017392,-0.2538236082,0.3569175601,0.0049056802,-0.1719962656,-0.0149831194,-0.1687421799,-0.2313759178,0.0281378422,0.3235939145,0.1266870797,0.6207981706,-0.0286829732,0.2285003811,0.3370732367,-0.188007012,0.1313721091,-0.0277588796,0.0349624828,0.4023726285,-0.238850221,-0.1197175607,0.0664065331,-0.1022803336,0.040619906,0.0130149191,-0.3214137554,0.0603422858,-0.2768704593,-0.0223208498,0.2950415611,0.2889094651,0.0581622198,0.5016816258,-0.0262932908,-0.1388157159,0.2159735262,0.0168195665,-0.124156177,-0.281819135,-0.1952973008,0.0977016166,0.0427054726,-0.2523407638,-0.0277266987,-0.243488133,-0.0807851404,0.0121760387,0.1265031397,0.482233882,0.0803427994,0.1158533469,0.0164868422,0.1796439886,-0.4184333384,-0.2533637583,-0.1892677546,0.022043325,-0.3372351229,-0.1492154151,0.2223524898,0.1535892189,0.3354985416,-0.1102682278,-0.0604240373,0.1364385486,-0.0179690253,-0.1492116302,-0.0685894415,0.2175128609,-0.2264653891,0.1727866828,0.2571268976,-0.1251242161,-0.2969128489,0.1636125892,-0.0111897727,-0.2530519068,-0.1209304109,0.0851275399,-0.1824117154,-0.2599703074,-0.3591947556,0.2333444357,0.3632029891,-0.263068974,-0.1501649469,-0.3836107254,-0.1859427094,-0.1210468039,0.1095169187,0.6124694347,-0.1049165651,-0.4261669219,-0.1083565578,-0.1307385713,0.386395663,0.4650295377,-0.4532147944,0.153607592,-0.2186063975,0.5694765449,0.383651495,-0.1736365557,-0.2988277078,0.2628951669,-0.2244315892,0.2072925121,-0.0707563609,-0.1237978861,-0.1637935787,0.0578438863,-0.1647842675,0.3448475599,0.1585469395,0.2666162848,-0.2812247872,-0.2925919592,0.4360112846,0.2920269966,0.216148749,0.1050296426,-0.1254457384,0.3208654523,0.2130340934,-0.2217202485,0.191267103,0.0267760381,0.0222752038,-0.1574062705,-0.032853514,0.0872845724,-0.6515653729,0.1748918146,-0.2471687347,0.3482395709,-0.3800010085,0.1222224906,-0.0670937449,-0.1169426218,-0.3418495059,-0.2125000358,0.0444627553,0.183524698,-0.0876024067,-0.0400484204,-0.0541984327,0.0558037721,0.0123563483,0.1081894338,-0.4298703969,0.1635243893,-0.1746212542,-0.3540358245,-0.0586851798,0.2108436823,0.2123282403,-0.2704420388,-0.2660779059,0.2112212032,0.0572887398,0.2215671539,0.0190088879,-0.00503748,0.1972181201,-0.192833066,0.0177532136,0.3558892608,0.0182690863,0.0179571509,0.0842206031,0.3661319017,0.0188044794,0.0989182368,0.1062778458,0.0629072785,0.1784609258,-0.0767172575,0.163967669,-0.1630000323,0.300399065,0.0798098519,0.4254424274,0.0218044408,-0.336406678,-0.2099469602,0.4499590993,0.0788158178,0.1665172726,0.3088004887,-0.0811552033,-0.1771025658,-0.0583941601,0.0550881997,0.2792567611,0.0804542229,-0.0513546728,0.0372813307,-0.0395417921,-0.1005213484,0.1776282787,0.101741232,-0.0386786871,0.3335563242,-0.0069130347,0.0972452164,-0.4322930276,-0.2849202156,0.1184080318,0.3154613972,-0.2130566686,0.1573490947,-0.1570317745,-0.0666516423,-0.4013566077,-0.2381341457,-0.4762080014,-0.1907107383,-0.2355116904,0.1614567786,-0.1278426498,0.2946344018,-0.0560883917,0.0913452059,0.0177859832,-0.0834005475,-0.1864178628,-0.2087877989,-0.4619050622,-0.0359911546,0.2449781895,0.021026792,0.1043828949,0.123727344,-0.0020178452,0.1463040262,-0.3587393761,0.2713183761,0.0121457307,0.187657088,0.1870482713,0.0780553222,-0.1918313205,-0.4377750158,0.2089122981,-0.062545836,-0.2025256455,0.0551075153,0.0965651721,-0.0389919728,-0.1555091143,-0.3513694406,-0.3187605441,-0.4107227623,0.1541299969,-0.1068721488,0.192685917,0.6526251435,-0.0405006818,0.3030754626,-0.23926799,0.0254845191,-0.1506410986,0.1499269158,0.0620459914,-0.1284027398,-0.3454410732,-0.3472175002,-0.0850819349,0.2574238479,-0.0885448754,-0.2285953015,0.3524973392,-0.1272809058,-0.0153178964,-0.0792646483,-0.0165644158,0.437337935,0.1797020435,0.0240603052,-0.0108518275,-0.0252295118,0.2277865112,-0.0137914531,0.3982838094,-0.0115693389,0.121883519,0.0862016305,0.8433319926,0.1893257499,-0.3235135376,0.3949736059,0.3149036169,-0.0733484626,-0.256616503,-0.3480521142,-0.0821666047,-0.0401236229,0.1737650484,0.223294124,-0.0486487336,-0.3717193007,-0.0359347127,-0.1308296472,-0.1652549207,-0.3224005401,0.2523370087,-0.1592554748,0.2457666844,0.0005499725,0.0926778764,-0.2677298188,0.0144890798,0.0368714817,0.2266474813,0.1431090385,0.0896599293,-0.5136061907,-0.0663289055,-0.4781812429,0.2796219587,0.1847020835,0.564268887,-0.1529383361,-0.168003425,0.1045318618,-0.0186034441,0.7028640509,0.02301687,-0.0897498056,0.1123473346,-0.2468478978,-0.3765416741,0.0356233455,-0.045827385,0.1709315032,0.3891173005,0.5012796521,-0.4591732621,-0.3562014401,0.0085638128,-0.0293875299,-0.0819561556,-0.201848805,-0.186429441,0.0265050717,-0.2668227553,0.1215890199,-0.0974667743,0.3548803031,0.0250712298,0.2982490361,-0.0434097834,-0.1342862248,-0.1364925653,0.1130342782,0.1991302222,-0.0183561854,-0.0332275294,0.3735995293,0.1696552038,-0.2558128536,0.270385325,0.0274356678,-0.2911652327,0.0742891133,0.2565268576,0.1025480554,0.2148583978,0.1469199359,0.2874315977,0.1137224138,0.0473645777,0.0363879763,0.226997897,0.2227705419,0.1510111094,-0.3832801878,-0.2247336209,0.4746959507,-0.2865239978,0.0148015302,0.2330476195,-0.1075512543,-0.3312224448,0.6692470908,-0.021685563,1.0043586493,0.1279561073,0.2349968553,0.3570190072,0.0152032506,0.4015209675,0.075397931,0.092209965,-0.4636116326,0.1191376597,-0.036317274,-0.0786606148,0.1102758348,0.2392281592,-0.1544450521,0.1322183609,-0.0604646206,-0.2274056077,0.1711773425,0.4276742339,0.1397120804,-0.4334684908,-0.2043327689,0.0076698088,-0.266140908,0.5549460649,-0.0314720273,-0.2316812873,-0.0094632655,-0.3618118465,-0.2293406725,0.301956892,-0.0994984582,0.3744971454,0.600733459,-0.1288872063,0.4571174383,0.1026240066,0.2113724202,0.0299900398,-0.2630378902,0.1922145486,-0.0337365195,-0.0982527435,0.0589332096,0.0264072064,0.356651932,0.0146899167,-0.2336515188,-0.0800436065,-0.1291650534,-0.2312926501,-0.0259570889,-0.1611709595,0.019038491,-0.1472074986,-0.3155699968,0.0277754236,0.0255545061,-0.1649843454,0.0687859654,0.0484742932,-0.2449036688,-0.2078677565,0.3131470978,-0.2475472242,-0.0739669204,0.4810358882,-0.0992319584,-0.0194906276,0.6444162726,0.3653110266,-0.206257537,-0.1943459809,-0.0836617425,0.1650459766,-0.8099958897,-0.0325143486,-0.001692626,-0.2046627253,-0.0655456856,0.5908412933,0.2419272214,0.0629566014,-0.1193100661,-0.3331790566,-0.3890790939,0.2196124196,-0.222279042,0.1357290745,0.0955775231,0.3911072314,-0.0347451828,0.1781407148,-0.2182646394,-0.0300164614,-0.2331596166,0.252122879,0.2451464981,-0.4850661159,0.2257371098,0.0978112072,0.0273082107,0.1323976815,-0.3137067258,-0.1233846545,-0.4189220667,0.1245738044,0.2815958261,-0.2863855958,-0.0942837596,0.1666618735,0.1116882786,-0.1594538987,0.0924320891,-0.0615465902,-0.1313707232,0.2736981511,0.1938243657,0.0533949435,-0.0768333897,0.0873866379,-0.1187309399,-0.313446641,0.2363160253,0.2007078677,-0.1965818852,0.0685271248,-0.2107036561,0.0794220269,0.3125178814,0.1303780377,0.0038125271,-0.3625322282,0.179345876,-0.036080543,0.0760999322,0.031151671,-0.1425473988,-0.2864573002,0.4078353047,0.1651120633,-0.2894612849,-0.2251743972,0.0795918107,0.0420671366,-0.0174697917,0.0502241552,0.0443150178,0.0742180571,0.0834893435,0.0840106532,0.4601431787,-0.2134629041,0.2142047435,0.3518062234,-0.1893409789,0.1552309543,0.464518249,0.0193825513,0.1395028383,0.2465680093,-0.2924775779,0.2774638236,-0.26682657,0.290597111,0.0077866502,-0.5513201952,-0.1049560905,0.2481084913,0.1737886071,0.1777735054,-0.066557765,0.445124656,-0.1363940984,-0.1493584365,-0.048861865,0.3371517658,-0.2716991603,-0.3933604956,-0.1717560142,0.1165925711,-0.394910574,-0.0003244607,-0.1727865338,-0.0516136475,0.156663537,0.1176726744,-0.105841279,-0.3330611587,-0.0519959256,0.0413667373,-0.1215798855,-0.2785531878,0.3481361866,-0.0297522619,0.0341189764,-0.1454355419,0.133100912,0.286752671,0.2867029905,-0.3025152385,-0.1334831566,-0.1729675233,0.1225762069,-0.030438317,0.3490459025,0.1919632107,0.4685676694,0.3146248162,0.0043297336,-0.156321317,-0.1256596297,0.1083557904,0.1264597625,-0.0214302167,0.5825879574,-0.1685422212,-0.0980616063,-0.260268867,0.1421924382,-0.4087181091,0.3401703238,0.2832738459,0.0236689746,0.1059991643,-0.0358005501,0.043437358,0.1233691052,0.5698880553,0.3689297438,0.1243647039,-0.5338953137,-0.0682754219,-0.4477844834,0.082789883,0.193966046,0.1000257283,0.0117921755,0.0065873745,0.0200608894,-0.0266447198,-0.3352196217,0.3204243779,0.0809512511,0.0154610472,-0.0622798093,0.1577226818,-0.2189983875,0.0669474602,0.0925963596,-0.4667535424,0.0259801149,0.1531939656,-0.0335598029,-0.2191779166,-0.1638471335,0.0668924451,0.0369472876,-0.0208345447,0.1326977015,0.7262605429,-0.0420248248,-0.1220177785,-0.4175515175,-0.4629760683,-0.1109927222,0.0634776875,-0.1209399998,0.6496319771,-0.0621025972,0.3009475768,-0.4772535264,0.1635540873,-0.1076173261,-0.1320490986,-0.3146496117,0.1489403099,-0.2459841222,-0.0108380876,0.1265085787,0.3707412183,0.1228420958,0.0308370795,-0.3600005805,-0.2014899254,0.6447339058,-0.7064890862,-0.1418036073,-0.091070056,0.0048363954,0.2531350553,-0.2619392574,-0.7180592418,-0.0282277595,0.1826342046,-0.0192403179,-0.1775765568,0.0684963912,0.230887115,-0.0765093267,-0.1325629354,0.656732142,0.1076977178,-0.2209262699,0.1161853224,-0.1894451827]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/598","title":"The current version of the package on github has an error when loading dataset","comments":"Thanks for reporting !\r\nWhich version of transformers are you using ?\r\nIt looks like it doesn't have the PreTrainedTokenizerBase class","body":"Instead of downloading the package from pip, downloading the version from source will result in an error when loading dataset (the pip version is completely fine):\r\n\r\nTo recreate the error: \r\nFirst, installing nlp directly from source:\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp.git\r\ncd nlp\r\npip install -e .\r\n```\r\nThen run:\r\n```\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train') \r\n```\r\nwill give error:\r\n\r\n```\r\n>>> dataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train')\r\nChecking \/home\/zeyuy\/.cache\/huggingface\/datasets\/84a754b488511b109e2904672d809c041008416ae74e38f9ee0c80a8dffa1383.2e21f48d63b5572d19c97e441fbb802257cf6a4c03fbc5ed8fae3d2c2273f59e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.py\r\nFound dataset infos file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/dataset_infos.json to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.json\r\nLoading Dataset Infos from \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nOverwrite dataset info from restored data version.\r\nLoading Dataset info from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nReusing dataset wikitext (\/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d)\r\nConstructing Dataset for split train, from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/load.py\", line 600, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 611, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 631, in _build_single_dataset\r\n ds = self._as_dataset(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 704, in _as_dataset\r\n return Dataset(**dataset_kwargs)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/arrow_dataset.py\", line 188, in __init__\r\n self._fingerprint = generate_fingerprint(self)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 91, in generate_fingerprint\r\n hasher.update(key)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 361, in dumps\r\n with _no_cache_fields(obj):\r\n File \"\/home\/zeyuy\/miniconda3\/lib\/python3.8\/contextlib.py\", line 113, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 348, in _no_cache_fields\r\n if isinstance(obj, tr.PreTrainedTokenizerBase) and hasattr(obj, \"cache\") and isinstance(obj.cache, dict):\r\nAttributeError: module 'transformers' has no attribute 'PreTrainedTokenizerBase'\r\n\r\n```\r\n\r\n\r\n","comment_length":21,"text":"The current version of the package on github has an error when loading dataset\nInstead of downloading the package from pip, downloading the version from source will result in an error when loading dataset (the pip version is completely fine):\r\n\r\nTo recreate the error: \r\nFirst, installing nlp directly from source:\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp.git\r\ncd nlp\r\npip install -e .\r\n```\r\nThen run:\r\n```\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train') \r\n```\r\nwill give error:\r\n\r\n```\r\n>>> dataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train')\r\nChecking \/home\/zeyuy\/.cache\/huggingface\/datasets\/84a754b488511b109e2904672d809c041008416ae74e38f9ee0c80a8dffa1383.2e21f48d63b5572d19c97e441fbb802257cf6a4c03fbc5ed8fae3d2c2273f59e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.py\r\nFound dataset infos file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/dataset_infos.json to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.json\r\nLoading Dataset Infos from \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nOverwrite dataset info from restored data version.\r\nLoading Dataset info from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nReusing dataset wikitext (\/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d)\r\nConstructing Dataset for split train, from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/load.py\", line 600, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 611, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 631, in _build_single_dataset\r\n ds = self._as_dataset(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 704, in _as_dataset\r\n return Dataset(**dataset_kwargs)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/arrow_dataset.py\", line 188, in __init__\r\n self._fingerprint = generate_fingerprint(self)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 91, in generate_fingerprint\r\n hasher.update(key)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 361, in dumps\r\n with _no_cache_fields(obj):\r\n File \"\/home\/zeyuy\/miniconda3\/lib\/python3.8\/contextlib.py\", line 113, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 348, in _no_cache_fields\r\n if isinstance(obj, tr.PreTrainedTokenizerBase) and hasattr(obj, \"cache\") and isinstance(obj.cache, dict):\r\nAttributeError: module 'transformers' has no attribute 'PreTrainedTokenizerBase'\r\n\r\n```\r\n\r\n\r\n\nThanks for reporting !\r\nWhich version of transformers are you using ?\r\nIt looks like it doesn't have the PreTrainedTokenizerBase class","embeddings":[-0.2237439454,-0.2012975067,-0.0425649211,0.0305607617,0.0355139226,-0.0190251414,-0.1124798357,0.2452018559,-0.001779821,-0.2171765715,0.2775869966,0.3542352021,-0.2464630753,0.2480493188,0.3240589201,-0.2881516814,0.0670975968,0.3065455556,-0.2375215143,0.052771423,-0.0257973578,0.4857282341,-0.2306242734,0.3760013878,-0.0834271014,0.0187894478,0.0690767914,0.1313544363,-0.290976733,-0.5385344028,0.6074130535,-0.0616439879,0.050687056,0.1241366863,-0.0001099683,0.0639232248,0.4524998665,0.0468346737,-0.2115150392,-0.5344376564,-0.0258459039,-0.4982024729,0.2533906996,-0.0385189801,-0.0332290567,-0.1539615691,0.1237362921,0.214582786,0.5887579322,0.2153976262,0.2569969594,0.3765678108,0.3491860032,-0.0690731034,-0.0353674218,-0.2311026156,-0.0577439554,0.1733411849,0.0412600711,-0.0806533322,0.2716600895,0.1918052435,0.0599418171,-0.0913049728,0.4098450243,-0.161454156,-0.1533882469,-0.1004508808,-0.106811516,0.2216578424,0.2371093333,-0.3243218064,-0.4073660076,-0.3286474347,0.187576741,-0.2094341666,0.4702663124,0.0190892965,-0.186120376,0.1364380717,-0.2939095795,-0.3428574502,-0.1923741549,0.2609534264,0.1397571862,0.3448449671,-0.1708793938,0.1244879737,0.2748580575,-0.1763910651,-0.1969624013,0.1317265481,-0.1033229306,0.3190252185,-0.1158773601,-0.0795881972,-0.1095688045,0.1889611632,0.0918167755,0.0788195804,-0.1971868873,-0.1108211726,-0.0314432606,0.1409585923,0.2720894814,0.2717589438,0.17719993,-0.1189834997,-0.0032822043,0.1774206609,0.3424126506,0.0457538143,-0.0453365743,-0.330901742,-0.4002842307,-0.1914319694,0.1130491421,-0.1737478673,-0.3322196305,-0.081326887,0.0919059366,0.0365613513,0.1109031439,0.0576236881,-0.1283797175,0.292615056,0.0234916713,0.1923785359,-0.3092001081,0.0621640123,-0.1687899083,0.1422373801,-0.2686138749,-0.1336476505,0.1294742972,-0.2735540867,0.5659924746,-0.0004660598,-0.0992649272,0.0367269032,-0.2505792677,0.2368855476,-0.3236500919,0.1910729706,-0.0812207609,0.2065933347,-0.0171954483,-0.079016231,-0.2939078808,-0.1111744195,-0.3565986454,-0.2549518347,-0.1689952463,0.2119121999,-0.2016157508,-0.2702847123,-0.28041628,0.0198180582,-0.0658054501,-0.3608138859,-0.0046903952,-0.2430276424,-0.1055391654,-0.0930548757,0.3117880523,0.2317885309,0.0926560387,-0.3515031338,-0.2359995544,-0.3532510996,0.085713774,0.232038483,-0.2740488648,0.2068709433,-0.1880133897,0.1380047649,0.6235006452,-0.5698119998,-0.5247965455,0.2389446348,-0.1837093979,0.1494831592,-0.1065205559,-0.0558203906,-0.2375170439,-0.130237326,0.081613116,0.2938145995,0.0514477938,-0.1161220297,-0.1398784667,-0.4660894871,0.2725190222,0.0642341301,-0.0643576309,-0.0080756107,-0.0887142718,0.4918843508,0.1930050403,0.1258874834,-0.0510848165,0.197620526,0.0127383536,0.0237861667,-0.3512737453,-0.2618550062,-0.4162878692,0.1538453549,-0.2101598233,-0.0290679373,-0.1326701939,-0.0577359386,-0.3078288436,-0.0738359988,-0.1671573669,-0.2715986669,0.160721153,0.0048288829,0.295542568,0.1357762218,-0.170800522,0.3214167953,0.0418822467,0.1781069189,-0.4866747558,0.0572806597,-0.1335321963,-0.04602357,0.1729026735,0.4731178582,0.1276734173,-0.2077899277,0.0151710575,0.4866738915,-0.3806300759,0.2787311375,-0.0023203127,-0.1690996885,0.1775311083,-0.3081972003,-0.1312314123,0.0708644241,-0.0935908034,0.0851684138,-0.0395264849,0.0182773862,-0.0922875032,0.0961840451,0.2280829251,0.1834034473,0.3326526284,-0.0641416088,-0.0241613369,-0.1820687354,0.2553633749,-0.0649808645,0.1488990933,-0.1461163461,-0.0508810841,-0.0003214711,0.492379874,0.1942920536,0.0343220159,0.2010842264,0.0021548562,-0.0955146253,0.0481096655,0.2638821304,0.2911488116,0.2004128695,-0.0828938261,0.1889248341,-0.2043614239,-0.14940539,0.2577418983,0.1075863168,-0.0307695027,0.1811572015,-0.133785665,0.0114306705,-0.2375520319,0.094807364,-0.0222688876,0.2971349061,-0.0851938948,-0.0681524053,-0.4806866348,-0.3239578605,-0.3728132248,-0.2951772809,-0.4344964921,-0.1589564532,-0.0362031534,0.1149727404,0.0485616215,0.1649898738,-0.0553438179,-0.0526930653,-0.2266286761,-0.0447862186,-0.0267593414,-0.048416011,-0.3857755959,0.0981594473,0.1967359185,0.2448918372,0.4569402337,-0.2492747158,-0.1252939254,-0.0644187778,-0.2730083168,0.1938122362,-0.2926279902,0.2592026591,0.3556085229,-0.0516292565,-0.2703355253,-0.3359704614,0.3509817421,-0.4124706089,-0.1979120374,-0.0850107968,0.0744214356,-0.0316736996,-0.201679945,-0.3083908558,-0.2766191661,-0.3505234122,0.311194241,0.0806484818,0.0984401703,0.4184328318,-0.0125570213,0.2687783539,-0.1211048067,0.0469820276,-0.1270651519,-0.2123743892,0.0569491833,-0.3140262067,-0.3949981332,0.0235566385,0.228068009,0.3154696226,0.08416567,-0.4360543191,-0.2520715296,0.0119622005,0.0624023601,0.0597277135,0.3059398532,0.3308093846,-0.0383228585,-0.0712782368,-0.161549747,-0.0299329776,0.0604027025,-0.1182855293,0.3523581624,0.1799397916,0.4678028822,-0.0238170847,0.6063979864,0.4972340763,0.0376835503,0.3265112936,-0.0553836562,0.3800790012,-0.1730224341,-0.2814178467,0.1348730773,0.085837245,0.5628989339,0.2523513734,0.125241071,0.1085704789,-0.2215523273,-0.1356805861,-0.271579951,-0.1474337429,-0.1767597944,0.0646157488,0.257738322,0.1145814806,-0.0049042036,-0.0982689559,-0.2549663484,0.1954672933,0.2483882308,-0.0700774714,0.2564755976,-0.3562459648,-0.2916217446,-0.6304395199,0.0574149936,0.1644297242,0.5169088244,0.025179103,0.1565306336,0.1662591845,-0.1031456068,0.3516110778,0.0002638189,0.0848304108,0.0737704039,0.1700430065,-0.5438058376,-0.0358665027,-0.0565546528,0.3793546557,0.4606749713,0.0374549143,-0.489433527,-0.398358494,0.313881278,0.1002968997,-0.1521896273,-0.0328312591,-0.1738864928,-0.2557458878,-0.2024965435,-0.202711463,-0.016621666,0.3583607972,0.0794930756,0.2841619849,-0.0602859966,-0.0682130009,-0.0234563984,0.0180871058,0.3587939441,0.0872806609,0.0495740697,0.1722739935,0.1227548495,-0.1981005818,0.740023911,0.0505527258,-0.0480798073,-0.1474427432,0.2945861816,0.1420302838,0.1440692991,-0.1842709631,-0.059040878,-0.0059903301,0.1297979653,-0.0680281222,0.111773856,0.5068559647,0.0243972335,-0.0904783085,-0.2858389914,0.4987763762,0.0972947478,-0.0226619169,0.2015023082,0.4368650913,-0.2149178237,-0.0692137554,-0.1267646998,0.6320818067,0.2031877041,0.1952810436,0.2906047106,-0.039530091,0.7745858431,-0.0314161181,0.0939436853,-0.2058902085,0.1828342825,0.0607479811,-0.0090651298,0.0709609389,0.0198440943,-0.2359975576,0.3186231256,0.2848711014,0.0379710048,0.0395126343,0.4954665005,-0.1353971064,-0.1887606382,-0.4930029511,0.1709432006,-0.1057606488,0.4444634318,-0.202564612,-0.1389370859,0.098843582,-0.3229313195,-0.3750478923,0.1910779774,0.003019549,0.3070128858,0.2578696012,-0.1358375698,0.3146021068,0.1584174931,0.3754766583,0.0336967111,-0.3622069955,0.078557618,-0.3673884571,-0.3306555152,0.0866527483,0.327881366,0.2063789815,-0.2947437167,-0.1042833403,-0.0982757062,-0.0447290875,-0.1725851446,-0.0250437204,-0.0658384413,0.0086648474,-0.0605948903,-0.2262761444,0.140776217,0.1795838028,-0.2310500741,0.1475728601,0.1545291692,-0.1493359059,0.041291669,0.1740124077,0.0057264324,0.037649516,0.4061413109,0.0011240054,-0.1326890588,0.5396635532,-0.0424128249,-0.0977002457,-0.2147788256,-0.1864573658,0.0674128905,-0.4528866112,-0.0735907033,0.1367858946,-0.0382699668,0.1060438603,0.3980601728,0.2150791883,-0.0018265424,0.0810615718,-0.3938004673,-0.1874099076,0.3194365203,0.046502009,0.033769574,0.1690135598,0.2051030546,0.2865497172,0.0706636831,-0.2923620343,0.0620421693,-0.0556405149,0.0586599559,0.2769283652,-0.0889653713,0.1965150982,-0.0237101316,0.0793573186,0.0398360826,-0.3567083776,-0.1860345006,-0.1730073988,0.089326635,0.1075413972,-0.3162682056,-0.0202372503,-0.1181695387,-0.0508417711,0.0308042206,0.1071531922,-0.0029971588,0.0161075369,0.1494882107,0.1483577043,0.2294033766,-0.0439704396,0.0712110847,-0.0725781769,0.2434771806,0.0977990478,0.0999879465,0.0849315599,-0.02075978,-0.1256193221,-0.0588641316,0.0226386767,0.1156224012,0.0954875797,-0.4526156485,0.1014503986,0.09202905,0.3228099346,0.4633265436,-0.1390681416,-0.0120361289,0.5312450528,0.2185307592,-0.4312898517,-0.1404989064,0.2825394571,0.0634499192,-0.0981675163,-0.0459189713,0.2540195882,-0.2216081321,0.1616757065,0.0993240848,0.2928048372,-0.1385902315,0.3533585072,0.2495930344,0.0416501835,-0.2093863934,-0.078408882,0.1208259836,0.1530651152,0.2093198001,-0.2409861237,0.0943248495,-0.2601040602,0.2324693203,0.0799039602,-0.1778873354,-0.2385259122,0.1060007513,0.236590445,0.1821846664,0.0217416883,0.5176038742,-0.3201269805,-0.2440446168,-0.2050433457,0.4098780453,-0.0548688136,-0.3036973774,-0.0718473941,0.0295325629,-0.1447568983,-0.0243562199,-0.0471460819,-0.2858159542,0.159594059,0.138734296,-0.1831247658,-0.4008322656,-0.1868460476,0.1542181373,0.1049749032,0.0005528335,0.4406995177,0.0576149933,-0.1394481063,-0.0410802476,0.5153031945,0.4115845561,0.2033145428,0.0158385634,-0.1346996725,-0.0550060943,0.1704122424,-0.1088863462,0.2642565072,0.11354094,0.2260669619,0.2382231653,0.1221168563,-0.1751365066,-0.0902137384,-0.0171184503,-0.0275768079,-0.2278973907,0.5078484416,-0.1254352331,0.019597752,-0.2496704906,0.3484070897,-0.4911293089,0.2872688174,0.020130394,0.1537850797,-0.0028905068,-0.3591858447,0.1024304181,0.0927831382,0.6194455028,0.2813887298,-0.0012827755,0.0550916977,-0.4740493596,-0.8101565838,-0.0658303797,-0.2427225858,-0.1460841596,0.2658505142,-0.047519248,-0.4228576124,0.2343580276,0.0174471736,0.1221750304,-0.055587966,-0.2802506983,-0.1547045708,0.0316942483,0.0309373476,-0.0327144712,0.1080197692,-0.1543244869,0.0754257441,-0.2737427354,-0.0130389687,0.0053766966,0.3522352576,-0.2582994699,-0.2497126758,0.2114720494,0.2341263294,0.0735757649,0.0139634823,-0.1919136494,-0.4351018965,-0.2454770505,-0.356677115,0.3163404465,-0.1414402723,0.3165807724,-0.2512359917,-0.070954062,-0.3026753068,0.0841596872,0.0305440687,0.0684219524,-0.1361841112,0.0888380185,-0.1519759446,0.1070061997,0.2659898698,0.4927127361,-0.140616402,0.1378587633,-0.3419418335,-0.3935497999,0.4931782484,-0.5361915827,-0.037550699,0.0461831018,0.2863286436,0.184512198,-0.3383826613,-0.6405612826,0.0665391684,0.1711781025,0.1276189089,0.032047905,0.2320660502,-0.1947587579,-0.1220137924,0.0623129085,0.3027658463,0.0421605147,-0.3858789206,0.3996355832,-0.1751168668]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/598","title":"The current version of the package on github has an error when loading dataset","comments":"I was using transformer 2.9. And I switch to the latest transformer package. Everything works just fine!!\r\n\r\nThanks for helping! I should look more carefully next time. Didn't realize loading the data part requires using tokenizer.\r\n","body":"Instead of downloading the package from pip, downloading the version from source will result in an error when loading dataset (the pip version is completely fine):\r\n\r\nTo recreate the error: \r\nFirst, installing nlp directly from source:\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp.git\r\ncd nlp\r\npip install -e .\r\n```\r\nThen run:\r\n```\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train') \r\n```\r\nwill give error:\r\n\r\n```\r\n>>> dataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train')\r\nChecking \/home\/zeyuy\/.cache\/huggingface\/datasets\/84a754b488511b109e2904672d809c041008416ae74e38f9ee0c80a8dffa1383.2e21f48d63b5572d19c97e441fbb802257cf6a4c03fbc5ed8fae3d2c2273f59e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.py\r\nFound dataset infos file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/dataset_infos.json to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.json\r\nLoading Dataset Infos from \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nOverwrite dataset info from restored data version.\r\nLoading Dataset info from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nReusing dataset wikitext (\/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d)\r\nConstructing Dataset for split train, from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/load.py\", line 600, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 611, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 631, in _build_single_dataset\r\n ds = self._as_dataset(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 704, in _as_dataset\r\n return Dataset(**dataset_kwargs)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/arrow_dataset.py\", line 188, in __init__\r\n self._fingerprint = generate_fingerprint(self)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 91, in generate_fingerprint\r\n hasher.update(key)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 361, in dumps\r\n with _no_cache_fields(obj):\r\n File \"\/home\/zeyuy\/miniconda3\/lib\/python3.8\/contextlib.py\", line 113, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 348, in _no_cache_fields\r\n if isinstance(obj, tr.PreTrainedTokenizerBase) and hasattr(obj, \"cache\") and isinstance(obj.cache, dict):\r\nAttributeError: module 'transformers' has no attribute 'PreTrainedTokenizerBase'\r\n\r\n```\r\n\r\n\r\n","comment_length":36,"text":"The current version of the package on github has an error when loading dataset\nInstead of downloading the package from pip, downloading the version from source will result in an error when loading dataset (the pip version is completely fine):\r\n\r\nTo recreate the error: \r\nFirst, installing nlp directly from source:\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp.git\r\ncd nlp\r\npip install -e .\r\n```\r\nThen run:\r\n```\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train') \r\n```\r\nwill give error:\r\n\r\n```\r\n>>> dataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train')\r\nChecking \/home\/zeyuy\/.cache\/huggingface\/datasets\/84a754b488511b109e2904672d809c041008416ae74e38f9ee0c80a8dffa1383.2e21f48d63b5572d19c97e441fbb802257cf6a4c03fbc5ed8fae3d2c2273f59e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.py\r\nFound dataset infos file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/dataset_infos.json to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.json\r\nLoading Dataset Infos from \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nOverwrite dataset info from restored data version.\r\nLoading Dataset info from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nReusing dataset wikitext (\/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d)\r\nConstructing Dataset for split train, from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/load.py\", line 600, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 611, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 631, in _build_single_dataset\r\n ds = self._as_dataset(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 704, in _as_dataset\r\n return Dataset(**dataset_kwargs)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/arrow_dataset.py\", line 188, in __init__\r\n self._fingerprint = generate_fingerprint(self)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 91, in generate_fingerprint\r\n hasher.update(key)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 361, in dumps\r\n with _no_cache_fields(obj):\r\n File \"\/home\/zeyuy\/miniconda3\/lib\/python3.8\/contextlib.py\", line 113, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 348, in _no_cache_fields\r\n if isinstance(obj, tr.PreTrainedTokenizerBase) and hasattr(obj, \"cache\") and isinstance(obj.cache, dict):\r\nAttributeError: module 'transformers' has no attribute 'PreTrainedTokenizerBase'\r\n\r\n```\r\n\r\n\r\n\nI was using transformer 2.9. And I switch to the latest transformer package. Everything works just fine!!\r\n\r\nThanks for helping! I should look more carefully next time. Didn't realize loading the data part requires using tokenizer.\r\n","embeddings":[-0.2237439454,-0.2012975067,-0.0425649211,0.0305607617,0.0355139226,-0.0190251414,-0.1124798357,0.2452018559,-0.001779821,-0.2171765715,0.2775869966,0.3542352021,-0.2464630753,0.2480493188,0.3240589201,-0.2881516814,0.0670975968,0.3065455556,-0.2375215143,0.052771423,-0.0257973578,0.4857282341,-0.2306242734,0.3760013878,-0.0834271014,0.0187894478,0.0690767914,0.1313544363,-0.290976733,-0.5385344028,0.6074130535,-0.0616439879,0.050687056,0.1241366863,-0.0001099683,0.0639232248,0.4524998665,0.0468346737,-0.2115150392,-0.5344376564,-0.0258459039,-0.4982024729,0.2533906996,-0.0385189801,-0.0332290567,-0.1539615691,0.1237362921,0.214582786,0.5887579322,0.2153976262,0.2569969594,0.3765678108,0.3491860032,-0.0690731034,-0.0353674218,-0.2311026156,-0.0577439554,0.1733411849,0.0412600711,-0.0806533322,0.2716600895,0.1918052435,0.0599418171,-0.0913049728,0.4098450243,-0.161454156,-0.1533882469,-0.1004508808,-0.106811516,0.2216578424,0.2371093333,-0.3243218064,-0.4073660076,-0.3286474347,0.187576741,-0.2094341666,0.4702663124,0.0190892965,-0.186120376,0.1364380717,-0.2939095795,-0.3428574502,-0.1923741549,0.2609534264,0.1397571862,0.3448449671,-0.1708793938,0.1244879737,0.2748580575,-0.1763910651,-0.1969624013,0.1317265481,-0.1033229306,0.3190252185,-0.1158773601,-0.0795881972,-0.1095688045,0.1889611632,0.0918167755,0.0788195804,-0.1971868873,-0.1108211726,-0.0314432606,0.1409585923,0.2720894814,0.2717589438,0.17719993,-0.1189834997,-0.0032822043,0.1774206609,0.3424126506,0.0457538143,-0.0453365743,-0.330901742,-0.4002842307,-0.1914319694,0.1130491421,-0.1737478673,-0.3322196305,-0.081326887,0.0919059366,0.0365613513,0.1109031439,0.0576236881,-0.1283797175,0.292615056,0.0234916713,0.1923785359,-0.3092001081,0.0621640123,-0.1687899083,0.1422373801,-0.2686138749,-0.1336476505,0.1294742972,-0.2735540867,0.5659924746,-0.0004660598,-0.0992649272,0.0367269032,-0.2505792677,0.2368855476,-0.3236500919,0.1910729706,-0.0812207609,0.2065933347,-0.0171954483,-0.079016231,-0.2939078808,-0.1111744195,-0.3565986454,-0.2549518347,-0.1689952463,0.2119121999,-0.2016157508,-0.2702847123,-0.28041628,0.0198180582,-0.0658054501,-0.3608138859,-0.0046903952,-0.2430276424,-0.1055391654,-0.0930548757,0.3117880523,0.2317885309,0.0926560387,-0.3515031338,-0.2359995544,-0.3532510996,0.085713774,0.232038483,-0.2740488648,0.2068709433,-0.1880133897,0.1380047649,0.6235006452,-0.5698119998,-0.5247965455,0.2389446348,-0.1837093979,0.1494831592,-0.1065205559,-0.0558203906,-0.2375170439,-0.130237326,0.081613116,0.2938145995,0.0514477938,-0.1161220297,-0.1398784667,-0.4660894871,0.2725190222,0.0642341301,-0.0643576309,-0.0080756107,-0.0887142718,0.4918843508,0.1930050403,0.1258874834,-0.0510848165,0.197620526,0.0127383536,0.0237861667,-0.3512737453,-0.2618550062,-0.4162878692,0.1538453549,-0.2101598233,-0.0290679373,-0.1326701939,-0.0577359386,-0.3078288436,-0.0738359988,-0.1671573669,-0.2715986669,0.160721153,0.0048288829,0.295542568,0.1357762218,-0.170800522,0.3214167953,0.0418822467,0.1781069189,-0.4866747558,0.0572806597,-0.1335321963,-0.04602357,0.1729026735,0.4731178582,0.1276734173,-0.2077899277,0.0151710575,0.4866738915,-0.3806300759,0.2787311375,-0.0023203127,-0.1690996885,0.1775311083,-0.3081972003,-0.1312314123,0.0708644241,-0.0935908034,0.0851684138,-0.0395264849,0.0182773862,-0.0922875032,0.0961840451,0.2280829251,0.1834034473,0.3326526284,-0.0641416088,-0.0241613369,-0.1820687354,0.2553633749,-0.0649808645,0.1488990933,-0.1461163461,-0.0508810841,-0.0003214711,0.492379874,0.1942920536,0.0343220159,0.2010842264,0.0021548562,-0.0955146253,0.0481096655,0.2638821304,0.2911488116,0.2004128695,-0.0828938261,0.1889248341,-0.2043614239,-0.14940539,0.2577418983,0.1075863168,-0.0307695027,0.1811572015,-0.133785665,0.0114306705,-0.2375520319,0.094807364,-0.0222688876,0.2971349061,-0.0851938948,-0.0681524053,-0.4806866348,-0.3239578605,-0.3728132248,-0.2951772809,-0.4344964921,-0.1589564532,-0.0362031534,0.1149727404,0.0485616215,0.1649898738,-0.0553438179,-0.0526930653,-0.2266286761,-0.0447862186,-0.0267593414,-0.048416011,-0.3857755959,0.0981594473,0.1967359185,0.2448918372,0.4569402337,-0.2492747158,-0.1252939254,-0.0644187778,-0.2730083168,0.1938122362,-0.2926279902,0.2592026591,0.3556085229,-0.0516292565,-0.2703355253,-0.3359704614,0.3509817421,-0.4124706089,-0.1979120374,-0.0850107968,0.0744214356,-0.0316736996,-0.201679945,-0.3083908558,-0.2766191661,-0.3505234122,0.311194241,0.0806484818,0.0984401703,0.4184328318,-0.0125570213,0.2687783539,-0.1211048067,0.0469820276,-0.1270651519,-0.2123743892,0.0569491833,-0.3140262067,-0.3949981332,0.0235566385,0.228068009,0.3154696226,0.08416567,-0.4360543191,-0.2520715296,0.0119622005,0.0624023601,0.0597277135,0.3059398532,0.3308093846,-0.0383228585,-0.0712782368,-0.161549747,-0.0299329776,0.0604027025,-0.1182855293,0.3523581624,0.1799397916,0.4678028822,-0.0238170847,0.6063979864,0.4972340763,0.0376835503,0.3265112936,-0.0553836562,0.3800790012,-0.1730224341,-0.2814178467,0.1348730773,0.085837245,0.5628989339,0.2523513734,0.125241071,0.1085704789,-0.2215523273,-0.1356805861,-0.271579951,-0.1474337429,-0.1767597944,0.0646157488,0.257738322,0.1145814806,-0.0049042036,-0.0982689559,-0.2549663484,0.1954672933,0.2483882308,-0.0700774714,0.2564755976,-0.3562459648,-0.2916217446,-0.6304395199,0.0574149936,0.1644297242,0.5169088244,0.025179103,0.1565306336,0.1662591845,-0.1031456068,0.3516110778,0.0002638189,0.0848304108,0.0737704039,0.1700430065,-0.5438058376,-0.0358665027,-0.0565546528,0.3793546557,0.4606749713,0.0374549143,-0.489433527,-0.398358494,0.313881278,0.1002968997,-0.1521896273,-0.0328312591,-0.1738864928,-0.2557458878,-0.2024965435,-0.202711463,-0.016621666,0.3583607972,0.0794930756,0.2841619849,-0.0602859966,-0.0682130009,-0.0234563984,0.0180871058,0.3587939441,0.0872806609,0.0495740697,0.1722739935,0.1227548495,-0.1981005818,0.740023911,0.0505527258,-0.0480798073,-0.1474427432,0.2945861816,0.1420302838,0.1440692991,-0.1842709631,-0.059040878,-0.0059903301,0.1297979653,-0.0680281222,0.111773856,0.5068559647,0.0243972335,-0.0904783085,-0.2858389914,0.4987763762,0.0972947478,-0.0226619169,0.2015023082,0.4368650913,-0.2149178237,-0.0692137554,-0.1267646998,0.6320818067,0.2031877041,0.1952810436,0.2906047106,-0.039530091,0.7745858431,-0.0314161181,0.0939436853,-0.2058902085,0.1828342825,0.0607479811,-0.0090651298,0.0709609389,0.0198440943,-0.2359975576,0.3186231256,0.2848711014,0.0379710048,0.0395126343,0.4954665005,-0.1353971064,-0.1887606382,-0.4930029511,0.1709432006,-0.1057606488,0.4444634318,-0.202564612,-0.1389370859,0.098843582,-0.3229313195,-0.3750478923,0.1910779774,0.003019549,0.3070128858,0.2578696012,-0.1358375698,0.3146021068,0.1584174931,0.3754766583,0.0336967111,-0.3622069955,0.078557618,-0.3673884571,-0.3306555152,0.0866527483,0.327881366,0.2063789815,-0.2947437167,-0.1042833403,-0.0982757062,-0.0447290875,-0.1725851446,-0.0250437204,-0.0658384413,0.0086648474,-0.0605948903,-0.2262761444,0.140776217,0.1795838028,-0.2310500741,0.1475728601,0.1545291692,-0.1493359059,0.041291669,0.1740124077,0.0057264324,0.037649516,0.4061413109,0.0011240054,-0.1326890588,0.5396635532,-0.0424128249,-0.0977002457,-0.2147788256,-0.1864573658,0.0674128905,-0.4528866112,-0.0735907033,0.1367858946,-0.0382699668,0.1060438603,0.3980601728,0.2150791883,-0.0018265424,0.0810615718,-0.3938004673,-0.1874099076,0.3194365203,0.046502009,0.033769574,0.1690135598,0.2051030546,0.2865497172,0.0706636831,-0.2923620343,0.0620421693,-0.0556405149,0.0586599559,0.2769283652,-0.0889653713,0.1965150982,-0.0237101316,0.0793573186,0.0398360826,-0.3567083776,-0.1860345006,-0.1730073988,0.089326635,0.1075413972,-0.3162682056,-0.0202372503,-0.1181695387,-0.0508417711,0.0308042206,0.1071531922,-0.0029971588,0.0161075369,0.1494882107,0.1483577043,0.2294033766,-0.0439704396,0.0712110847,-0.0725781769,0.2434771806,0.0977990478,0.0999879465,0.0849315599,-0.02075978,-0.1256193221,-0.0588641316,0.0226386767,0.1156224012,0.0954875797,-0.4526156485,0.1014503986,0.09202905,0.3228099346,0.4633265436,-0.1390681416,-0.0120361289,0.5312450528,0.2185307592,-0.4312898517,-0.1404989064,0.2825394571,0.0634499192,-0.0981675163,-0.0459189713,0.2540195882,-0.2216081321,0.1616757065,0.0993240848,0.2928048372,-0.1385902315,0.3533585072,0.2495930344,0.0416501835,-0.2093863934,-0.078408882,0.1208259836,0.1530651152,0.2093198001,-0.2409861237,0.0943248495,-0.2601040602,0.2324693203,0.0799039602,-0.1778873354,-0.2385259122,0.1060007513,0.236590445,0.1821846664,0.0217416883,0.5176038742,-0.3201269805,-0.2440446168,-0.2050433457,0.4098780453,-0.0548688136,-0.3036973774,-0.0718473941,0.0295325629,-0.1447568983,-0.0243562199,-0.0471460819,-0.2858159542,0.159594059,0.138734296,-0.1831247658,-0.4008322656,-0.1868460476,0.1542181373,0.1049749032,0.0005528335,0.4406995177,0.0576149933,-0.1394481063,-0.0410802476,0.5153031945,0.4115845561,0.2033145428,0.0158385634,-0.1346996725,-0.0550060943,0.1704122424,-0.1088863462,0.2642565072,0.11354094,0.2260669619,0.2382231653,0.1221168563,-0.1751365066,-0.0902137384,-0.0171184503,-0.0275768079,-0.2278973907,0.5078484416,-0.1254352331,0.019597752,-0.2496704906,0.3484070897,-0.4911293089,0.2872688174,0.020130394,0.1537850797,-0.0028905068,-0.3591858447,0.1024304181,0.0927831382,0.6194455028,0.2813887298,-0.0012827755,0.0550916977,-0.4740493596,-0.8101565838,-0.0658303797,-0.2427225858,-0.1460841596,0.2658505142,-0.047519248,-0.4228576124,0.2343580276,0.0174471736,0.1221750304,-0.055587966,-0.2802506983,-0.1547045708,0.0316942483,0.0309373476,-0.0327144712,0.1080197692,-0.1543244869,0.0754257441,-0.2737427354,-0.0130389687,0.0053766966,0.3522352576,-0.2582994699,-0.2497126758,0.2114720494,0.2341263294,0.0735757649,0.0139634823,-0.1919136494,-0.4351018965,-0.2454770505,-0.356677115,0.3163404465,-0.1414402723,0.3165807724,-0.2512359917,-0.070954062,-0.3026753068,0.0841596872,0.0305440687,0.0684219524,-0.1361841112,0.0888380185,-0.1519759446,0.1070061997,0.2659898698,0.4927127361,-0.140616402,0.1378587633,-0.3419418335,-0.3935497999,0.4931782484,-0.5361915827,-0.037550699,0.0461831018,0.2863286436,0.184512198,-0.3383826613,-0.6405612826,0.0665391684,0.1711781025,0.1276189089,0.032047905,0.2320660502,-0.1947587579,-0.1220137924,0.0623129085,0.3027658463,0.0421605147,-0.3858789206,0.3996355832,-0.1751168668]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/598","title":"The current version of the package on github has an error when loading dataset","comments":"Yes it shouldn\u2019t fail with older version of transformers since this is only a special feature to make caching more efficient when using transformers for tokenization.\r\nWe\u2019ll update this.","body":"Instead of downloading the package from pip, downloading the version from source will result in an error when loading dataset (the pip version is completely fine):\r\n\r\nTo recreate the error: \r\nFirst, installing nlp directly from source:\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp.git\r\ncd nlp\r\npip install -e .\r\n```\r\nThen run:\r\n```\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train') \r\n```\r\nwill give error:\r\n\r\n```\r\n>>> dataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train')\r\nChecking \/home\/zeyuy\/.cache\/huggingface\/datasets\/84a754b488511b109e2904672d809c041008416ae74e38f9ee0c80a8dffa1383.2e21f48d63b5572d19c97e441fbb802257cf6a4c03fbc5ed8fae3d2c2273f59e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.py\r\nFound dataset infos file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/dataset_infos.json to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.json\r\nLoading Dataset Infos from \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nOverwrite dataset info from restored data version.\r\nLoading Dataset info from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nReusing dataset wikitext (\/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d)\r\nConstructing Dataset for split train, from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/load.py\", line 600, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 611, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 631, in _build_single_dataset\r\n ds = self._as_dataset(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 704, in _as_dataset\r\n return Dataset(**dataset_kwargs)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/arrow_dataset.py\", line 188, in __init__\r\n self._fingerprint = generate_fingerprint(self)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 91, in generate_fingerprint\r\n hasher.update(key)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 361, in dumps\r\n with _no_cache_fields(obj):\r\n File \"\/home\/zeyuy\/miniconda3\/lib\/python3.8\/contextlib.py\", line 113, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 348, in _no_cache_fields\r\n if isinstance(obj, tr.PreTrainedTokenizerBase) and hasattr(obj, \"cache\") and isinstance(obj.cache, dict):\r\nAttributeError: module 'transformers' has no attribute 'PreTrainedTokenizerBase'\r\n\r\n```\r\n\r\n\r\n","comment_length":29,"text":"The current version of the package on github has an error when loading dataset\nInstead of downloading the package from pip, downloading the version from source will result in an error when loading dataset (the pip version is completely fine):\r\n\r\nTo recreate the error: \r\nFirst, installing nlp directly from source:\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp.git\r\ncd nlp\r\npip install -e .\r\n```\r\nThen run:\r\n```\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train') \r\n```\r\nwill give error:\r\n\r\n```\r\n>>> dataset = load_dataset('wikitext', 'wikitext-2-v1',split = 'train')\r\nChecking \/home\/zeyuy\/.cache\/huggingface\/datasets\/84a754b488511b109e2904672d809c041008416ae74e38f9ee0c80a8dffa1383.2e21f48d63b5572d19c97e441fbb802257cf6a4c03fbc5ed8fae3d2c2273f59e.py for additional imports.\r\nFound main folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\r\nFound specific version folder for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nFound script file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.py\r\nFound dataset infos file from https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/dataset_infos.json to \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/dataset_infos.json\r\nFound metadata file for dataset https:\/\/raw.githubusercontent.com\/huggingface\/nlp\/0.4.0\/datasets\/wikitext\/wikitext.py at \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\/wikitext.json\r\nLoading Dataset Infos from \/home\/zeyuy\/.cache\/huggingface\/modules\/nlp_modules\/datasets\/wikitext\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nOverwrite dataset info from restored data version.\r\nLoading Dataset info from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nReusing dataset wikitext (\/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d)\r\nConstructing Dataset for split train, from \/home\/zeyuy\/.cache\/huggingface\/datasets\/wikitext\/wikitext-2-v1\/1.0.0\/5de6e79516446f747fcccc09aa2614fa159053b75909594d28d262395f72d89d\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/load.py\", line 600, in load_dataset\r\n ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 611, in as_dataset\r\n datasets = utils.map_nested(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 631, in _build_single_dataset\r\n ds = self._as_dataset(\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/builder.py\", line 704, in _as_dataset\r\n return Dataset(**dataset_kwargs)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/arrow_dataset.py\", line 188, in __init__\r\n self._fingerprint = generate_fingerprint(self)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 91, in generate_fingerprint\r\n hasher.update(key)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 57, in update\r\n self.m.update(self.hash(value).encode(\"utf-8\"))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 53, in hash\r\n return cls.hash_default(value)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/fingerprint.py\", line 46, in hash_default\r\n return cls.hash_bytes(dumps(value))\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 361, in dumps\r\n with _no_cache_fields(obj):\r\n File \"\/home\/zeyuy\/miniconda3\/lib\/python3.8\/contextlib.py\", line 113, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/zeyuy\/transformers\/examples\/language-modeling\/nlp\/src\/nlp\/utils\/py_utils.py\", line 348, in _no_cache_fields\r\n if isinstance(obj, tr.PreTrainedTokenizerBase) and hasattr(obj, \"cache\") and isinstance(obj.cache, dict):\r\nAttributeError: module 'transformers' has no attribute 'PreTrainedTokenizerBase'\r\n\r\n```\r\n\r\n\r\n\nYes it shouldn\u2019t fail with older version of transformers since this is only a special feature to make caching more efficient when using transformers for tokenization.\r\nWe\u2019ll update this.","embeddings":[-0.2237439454,-0.2012975067,-0.0425649211,0.0305607617,0.0355139226,-0.0190251414,-0.1124798357,0.2452018559,-0.001779821,-0.2171765715,0.2775869966,0.3542352021,-0.2464630753,0.2480493188,0.3240589201,-0.2881516814,0.0670975968,0.3065455556,-0.2375215143,0.052771423,-0.0257973578,0.4857282341,-0.2306242734,0.3760013878,-0.0834271014,0.0187894478,0.0690767914,0.1313544363,-0.290976733,-0.5385344028,0.6074130535,-0.0616439879,0.050687056,0.1241366863,-0.0001099683,0.0639232248,0.4524998665,0.0468346737,-0.2115150392,-0.5344376564,-0.0258459039,-0.4982024729,0.2533906996,-0.0385189801,-0.0332290567,-0.1539615691,0.1237362921,0.214582786,0.5887579322,0.2153976262,0.2569969594,0.3765678108,0.3491860032,-0.0690731034,-0.0353674218,-0.2311026156,-0.0577439554,0.1733411849,0.0412600711,-0.0806533322,0.2716600895,0.1918052435,0.0599418171,-0.0913049728,0.4098450243,-0.161454156,-0.1533882469,-0.1004508808,-0.106811516,0.2216578424,0.2371093333,-0.3243218064,-0.4073660076,-0.3286474347,0.187576741,-0.2094341666,0.4702663124,0.0190892965,-0.186120376,0.1364380717,-0.2939095795,-0.3428574502,-0.1923741549,0.2609534264,0.1397571862,0.3448449671,-0.1708793938,0.1244879737,0.2748580575,-0.1763910651,-0.1969624013,0.1317265481,-0.1033229306,0.3190252185,-0.1158773601,-0.0795881972,-0.1095688045,0.1889611632,0.0918167755,0.0788195804,-0.1971868873,-0.1108211726,-0.0314432606,0.1409585923,0.2720894814,0.2717589438,0.17719993,-0.1189834997,-0.0032822043,0.1774206609,0.3424126506,0.0457538143,-0.0453365743,-0.330901742,-0.4002842307,-0.1914319694,0.1130491421,-0.1737478673,-0.3322196305,-0.081326887,0.0919059366,0.0365613513,0.1109031439,0.0576236881,-0.1283797175,0.292615056,0.0234916713,0.1923785359,-0.3092001081,0.0621640123,-0.1687899083,0.1422373801,-0.2686138749,-0.1336476505,0.1294742972,-0.2735540867,0.5659924746,-0.0004660598,-0.0992649272,0.0367269032,-0.2505792677,0.2368855476,-0.3236500919,0.1910729706,-0.0812207609,0.2065933347,-0.0171954483,-0.079016231,-0.2939078808,-0.1111744195,-0.3565986454,-0.2549518347,-0.1689952463,0.2119121999,-0.2016157508,-0.2702847123,-0.28041628,0.0198180582,-0.0658054501,-0.3608138859,-0.0046903952,-0.2430276424,-0.1055391654,-0.0930548757,0.3117880523,0.2317885309,0.0926560387,-0.3515031338,-0.2359995544,-0.3532510996,0.085713774,0.232038483,-0.2740488648,0.2068709433,-0.1880133897,0.1380047649,0.6235006452,-0.5698119998,-0.5247965455,0.2389446348,-0.1837093979,0.1494831592,-0.1065205559,-0.0558203906,-0.2375170439,-0.130237326,0.081613116,0.2938145995,0.0514477938,-0.1161220297,-0.1398784667,-0.4660894871,0.2725190222,0.0642341301,-0.0643576309,-0.0080756107,-0.0887142718,0.4918843508,0.1930050403,0.1258874834,-0.0510848165,0.197620526,0.0127383536,0.0237861667,-0.3512737453,-0.2618550062,-0.4162878692,0.1538453549,-0.2101598233,-0.0290679373,-0.1326701939,-0.0577359386,-0.3078288436,-0.0738359988,-0.1671573669,-0.2715986669,0.160721153,0.0048288829,0.295542568,0.1357762218,-0.170800522,0.3214167953,0.0418822467,0.1781069189,-0.4866747558,0.0572806597,-0.1335321963,-0.04602357,0.1729026735,0.4731178582,0.1276734173,-0.2077899277,0.0151710575,0.4866738915,-0.3806300759,0.2787311375,-0.0023203127,-0.1690996885,0.1775311083,-0.3081972003,-0.1312314123,0.0708644241,-0.0935908034,0.0851684138,-0.0395264849,0.0182773862,-0.0922875032,0.0961840451,0.2280829251,0.1834034473,0.3326526284,-0.0641416088,-0.0241613369,-0.1820687354,0.2553633749,-0.0649808645,0.1488990933,-0.1461163461,-0.0508810841,-0.0003214711,0.492379874,0.1942920536,0.0343220159,0.2010842264,0.0021548562,-0.0955146253,0.0481096655,0.2638821304,0.2911488116,0.2004128695,-0.0828938261,0.1889248341,-0.2043614239,-0.14940539,0.2577418983,0.1075863168,-0.0307695027,0.1811572015,-0.133785665,0.0114306705,-0.2375520319,0.094807364,-0.0222688876,0.2971349061,-0.0851938948,-0.0681524053,-0.4806866348,-0.3239578605,-0.3728132248,-0.2951772809,-0.4344964921,-0.1589564532,-0.0362031534,0.1149727404,0.0485616215,0.1649898738,-0.0553438179,-0.0526930653,-0.2266286761,-0.0447862186,-0.0267593414,-0.048416011,-0.3857755959,0.0981594473,0.1967359185,0.2448918372,0.4569402337,-0.2492747158,-0.1252939254,-0.0644187778,-0.2730083168,0.1938122362,-0.2926279902,0.2592026591,0.3556085229,-0.0516292565,-0.2703355253,-0.3359704614,0.3509817421,-0.4124706089,-0.1979120374,-0.0850107968,0.0744214356,-0.0316736996,-0.201679945,-0.3083908558,-0.2766191661,-0.3505234122,0.311194241,0.0806484818,0.0984401703,0.4184328318,-0.0125570213,0.2687783539,-0.1211048067,0.0469820276,-0.1270651519,-0.2123743892,0.0569491833,-0.3140262067,-0.3949981332,0.0235566385,0.228068009,0.3154696226,0.08416567,-0.4360543191,-0.2520715296,0.0119622005,0.0624023601,0.0597277135,0.3059398532,0.3308093846,-0.0383228585,-0.0712782368,-0.161549747,-0.0299329776,0.0604027025,-0.1182855293,0.3523581624,0.1799397916,0.4678028822,-0.0238170847,0.6063979864,0.4972340763,0.0376835503,0.3265112936,-0.0553836562,0.3800790012,-0.1730224341,-0.2814178467,0.1348730773,0.085837245,0.5628989339,0.2523513734,0.125241071,0.1085704789,-0.2215523273,-0.1356805861,-0.271579951,-0.1474337429,-0.1767597944,0.0646157488,0.257738322,0.1145814806,-0.0049042036,-0.0982689559,-0.2549663484,0.1954672933,0.2483882308,-0.0700774714,0.2564755976,-0.3562459648,-0.2916217446,-0.6304395199,0.0574149936,0.1644297242,0.5169088244,0.025179103,0.1565306336,0.1662591845,-0.1031456068,0.3516110778,0.0002638189,0.0848304108,0.0737704039,0.1700430065,-0.5438058376,-0.0358665027,-0.0565546528,0.3793546557,0.4606749713,0.0374549143,-0.489433527,-0.398358494,0.313881278,0.1002968997,-0.1521896273,-0.0328312591,-0.1738864928,-0.2557458878,-0.2024965435,-0.202711463,-0.016621666,0.3583607972,0.0794930756,0.2841619849,-0.0602859966,-0.0682130009,-0.0234563984,0.0180871058,0.3587939441,0.0872806609,0.0495740697,0.1722739935,0.1227548495,-0.1981005818,0.740023911,0.0505527258,-0.0480798073,-0.1474427432,0.2945861816,0.1420302838,0.1440692991,-0.1842709631,-0.059040878,-0.0059903301,0.1297979653,-0.0680281222,0.111773856,0.5068559647,0.0243972335,-0.0904783085,-0.2858389914,0.4987763762,0.0972947478,-0.0226619169,0.2015023082,0.4368650913,-0.2149178237,-0.0692137554,-0.1267646998,0.6320818067,0.2031877041,0.1952810436,0.2906047106,-0.039530091,0.7745858431,-0.0314161181,0.0939436853,-0.2058902085,0.1828342825,0.0607479811,-0.0090651298,0.0709609389,0.0198440943,-0.2359975576,0.3186231256,0.2848711014,0.0379710048,0.0395126343,0.4954665005,-0.1353971064,-0.1887606382,-0.4930029511,0.1709432006,-0.1057606488,0.4444634318,-0.202564612,-0.1389370859,0.098843582,-0.3229313195,-0.3750478923,0.1910779774,0.003019549,0.3070128858,0.2578696012,-0.1358375698,0.3146021068,0.1584174931,0.3754766583,0.0336967111,-0.3622069955,0.078557618,-0.3673884571,-0.3306555152,0.0866527483,0.327881366,0.2063789815,-0.2947437167,-0.1042833403,-0.0982757062,-0.0447290875,-0.1725851446,-0.0250437204,-0.0658384413,0.0086648474,-0.0605948903,-0.2262761444,0.140776217,0.1795838028,-0.2310500741,0.1475728601,0.1545291692,-0.1493359059,0.041291669,0.1740124077,0.0057264324,0.037649516,0.4061413109,0.0011240054,-0.1326890588,0.5396635532,-0.0424128249,-0.0977002457,-0.2147788256,-0.1864573658,0.0674128905,-0.4528866112,-0.0735907033,0.1367858946,-0.0382699668,0.1060438603,0.3980601728,0.2150791883,-0.0018265424,0.0810615718,-0.3938004673,-0.1874099076,0.3194365203,0.046502009,0.033769574,0.1690135598,0.2051030546,0.2865497172,0.0706636831,-0.2923620343,0.0620421693,-0.0556405149,0.0586599559,0.2769283652,-0.0889653713,0.1965150982,-0.0237101316,0.0793573186,0.0398360826,-0.3567083776,-0.1860345006,-0.1730073988,0.089326635,0.1075413972,-0.3162682056,-0.0202372503,-0.1181695387,-0.0508417711,0.0308042206,0.1071531922,-0.0029971588,0.0161075369,0.1494882107,0.1483577043,0.2294033766,-0.0439704396,0.0712110847,-0.0725781769,0.2434771806,0.0977990478,0.0999879465,0.0849315599,-0.02075978,-0.1256193221,-0.0588641316,0.0226386767,0.1156224012,0.0954875797,-0.4526156485,0.1014503986,0.09202905,0.3228099346,0.4633265436,-0.1390681416,-0.0120361289,0.5312450528,0.2185307592,-0.4312898517,-0.1404989064,0.2825394571,0.0634499192,-0.0981675163,-0.0459189713,0.2540195882,-0.2216081321,0.1616757065,0.0993240848,0.2928048372,-0.1385902315,0.3533585072,0.2495930344,0.0416501835,-0.2093863934,-0.078408882,0.1208259836,0.1530651152,0.2093198001,-0.2409861237,0.0943248495,-0.2601040602,0.2324693203,0.0799039602,-0.1778873354,-0.2385259122,0.1060007513,0.236590445,0.1821846664,0.0217416883,0.5176038742,-0.3201269805,-0.2440446168,-0.2050433457,0.4098780453,-0.0548688136,-0.3036973774,-0.0718473941,0.0295325629,-0.1447568983,-0.0243562199,-0.0471460819,-0.2858159542,0.159594059,0.138734296,-0.1831247658,-0.4008322656,-0.1868460476,0.1542181373,0.1049749032,0.0005528335,0.4406995177,0.0576149933,-0.1394481063,-0.0410802476,0.5153031945,0.4115845561,0.2033145428,0.0158385634,-0.1346996725,-0.0550060943,0.1704122424,-0.1088863462,0.2642565072,0.11354094,0.2260669619,0.2382231653,0.1221168563,-0.1751365066,-0.0902137384,-0.0171184503,-0.0275768079,-0.2278973907,0.5078484416,-0.1254352331,0.019597752,-0.2496704906,0.3484070897,-0.4911293089,0.2872688174,0.020130394,0.1537850797,-0.0028905068,-0.3591858447,0.1024304181,0.0927831382,0.6194455028,0.2813887298,-0.0012827755,0.0550916977,-0.4740493596,-0.8101565838,-0.0658303797,-0.2427225858,-0.1460841596,0.2658505142,-0.047519248,-0.4228576124,0.2343580276,0.0174471736,0.1221750304,-0.055587966,-0.2802506983,-0.1547045708,0.0316942483,0.0309373476,-0.0327144712,0.1080197692,-0.1543244869,0.0754257441,-0.2737427354,-0.0130389687,0.0053766966,0.3522352576,-0.2582994699,-0.2497126758,0.2114720494,0.2341263294,0.0735757649,0.0139634823,-0.1919136494,-0.4351018965,-0.2454770505,-0.356677115,0.3163404465,-0.1414402723,0.3165807724,-0.2512359917,-0.070954062,-0.3026753068,0.0841596872,0.0305440687,0.0684219524,-0.1361841112,0.0888380185,-0.1519759446,0.1070061997,0.2659898698,0.4927127361,-0.140616402,0.1378587633,-0.3419418335,-0.3935497999,0.4931782484,-0.5361915827,-0.037550699,0.0461831018,0.2863286436,0.184512198,-0.3383826613,-0.6405612826,0.0665391684,0.1711781025,0.1276189089,0.032047905,0.2320660502,-0.1947587579,-0.1220137924,0.0623129085,0.3027658463,0.0421605147,-0.3858789206,0.3996355832,-0.1751168668]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/597","title":"Indices incorrect with multiprocessing","comments":"I fixed a bug that could cause this issue earlier today. Could you pull the latest version and try again ?","body":"When `num_proc` > 1, the indices argument passed to the map function is incorrect:\r\n\r\n```python\r\nd = load_dataset('imdb', split='test[:1%]')\r\n\r\ndef fn(x, inds):\r\n print(inds)\r\n return x\r\n\r\nd.select(range(10)).map(fn, with_indices=True, batched=True)\r\n# [0, 1]\r\n# [0, 1, 2, 3, 4, 5, 6, 7, 8, 9]\r\n\r\nd.select(range(10)).map(fn, with_indices=True, batched=True, num_proc=2)\r\n# [0, 1]\r\n# [0, 1]\r\n# [0, 1, 2, 3, 4]\r\n# [0, 1, 2, 3, 4]\r\n```\r\n\r\nAs you can see, the subset passed to each thread is indexed from 0 to N which doesn't reflect their positions in `d`.","comment_length":21,"text":"Indices incorrect with multiprocessing\nWhen `num_proc` > 1, the indices argument passed to the map function is incorrect:\r\n\r\n```python\r\nd = load_dataset('imdb', split='test[:1%]')\r\n\r\ndef fn(x, inds):\r\n print(inds)\r\n return x\r\n\r\nd.select(range(10)).map(fn, with_indices=True, batched=True)\r\n# [0, 1]\r\n# [0, 1, 2, 3, 4, 5, 6, 7, 8, 9]\r\n\r\nd.select(range(10)).map(fn, with_indices=True, batched=True, num_proc=2)\r\n# [0, 1]\r\n# [0, 1]\r\n# [0, 1, 2, 3, 4]\r\n# [0, 1, 2, 3, 4]\r\n```\r\n\r\nAs you can see, the subset passed to each thread is indexed from 0 to N which doesn't reflect their positions in `d`.\nI fixed a bug that could cause this issue earlier today. Could you pull the latest version and try again ?","embeddings":[-0.4318974614,-0.3361007869,-0.1858976632,0.2841148973,-0.2439694554,-0.0441068672,0.4962969422,0.0556306951,0.2259730697,0.3695326447,-0.0362011306,0.3468486965,0.1309847087,0.132168293,-0.2639608979,0.125157252,-0.0425510369,0.0501092598,-0.1806087345,-0.1643419564,-0.3834481537,0.2268317491,-0.4945385754,-0.1573975533,-0.4757076502,-0.1956421882,-0.0759058446,0.1099578887,0.3210601807,-0.160506621,0.009957877,0.022106763,-0.0914489552,0.7085773945,-0.0001031148,-0.1593160182,0.0751710683,-0.1524419636,0.1996055543,-0.0002129417,-0.0909554511,0.1350019425,0.2246314734,-0.2731075287,0.0537289158,0.0532294475,-0.0010817386,-0.5365207195,0.188287884,0.114147231,0.2736756802,0.2315301895,-0.1976775974,0.1363347322,-0.0576324798,0.0317330249,0.0174953546,0.1988241076,0.5325119495,-0.0010723029,-0.0761025771,0.2661803365,-0.2675154209,0.1340358108,-0.055068668,0.2426737398,0.1910540611,-0.5292800069,0.0899396464,0.0624273382,0.2397211492,-0.014680095,-0.3417310417,-0.1670342833,-0.0660365671,-0.2750051916,-0.1892737597,0.1301205754,0.0398401432,-0.1010042951,-0.3078900874,0.2415540963,0.1040982306,-0.0423946232,0.1155258343,0.2264727652,0.1334076822,0.0960728675,0.1119297966,0.0600199662,0.0332671404,0.0170928892,0.2272539884,0.0101899169,-0.2228100002,0.06591966,0.1626019627,-0.1847358346,-0.2487383336,-0.032761436,-0.0716658756,0.1557407379,0.0264391396,0.1709657758,0.2026150674,-0.1219339147,-0.0715444162,0.0419611968,0.3141613305,-0.1366681457,-0.2969742715,-0.0143852308,0.2374504507,-0.2750277221,0.1658841968,0.2358232588,-0.1792279631,0.0275973007,-0.3413961232,0.0850766897,-0.059376616,-0.1049611866,0.0169430971,0.154386282,-0.0021363408,0.2345193923,0.1239192858,0.4017362893,-0.4142128229,0.1561328024,-0.3718011379,-0.108310394,-0.0729487687,0.1471422315,0.1215673387,0.0886527449,0.2377531976,0.2813608944,0.2871539891,-0.2365341783,0.0768127069,-0.2243103236,0.2108921558,0.1928569078,0.2893156409,0.257409215,0.1592210978,-0.08742062,-0.1630048156,0.2510741651,-0.2757888734,-0.0819021389,-0.1048828438,0.2451403886,0.3177548349,0.0950774923,-0.3571501374,0.0302917883,0.3217176795,0.0474879816,-0.0547763146,-0.2276129425,-0.264395386,-0.2511063814,0.1363691092,0.2460228652,-0.1555967331,0.131892994,0.0226910915,0.1686579585,0.2406096011,0.3759451509,0.0269461777,-0.0760053471,-0.1439495683,0.4667575359,-0.1113358811,-0.4037052393,-0.1221790835,0.321708113,-0.166230917,-0.0430526957,0.1146674678,0.1156925857,0.3915439844,0.0218701866,0.5178156495,0.1550714672,-0.0904079825,0.3614135385,-0.2356896102,0.0310664084,-0.0882619843,0.2377884388,0.1266495585,-0.1160868704,-0.1165002361,-0.7145783901,0.5741788149,0.0097651007,0.0703204125,0.2118837386,-0.0618280582,0.3727458715,-0.0360418856,-0.3215188682,0.0664668232,0.1910869628,0.1904696077,0.1615994722,0.0930091366,-0.2018201053,-0.0757172033,0.0433045439,-0.0996581092,-0.0145305907,0.2032821476,0.0761818215,-0.273105979,-0.0606174059,-0.2873538435,0.1573041826,0.2398821563,-0.1844333261,-0.3069197536,0.2067868561,-0.2794632614,-0.4271344244,0.0233322047,0.2178389728,0.3047004044,-0.0238305163,0.1050538942,0.3498241305,0.1846109182,-0.0830978602,0.0375301018,0.1117563024,0.1906618774,-0.2784384787,0.1273007095,0.0858666003,-0.0252231322,-0.187318936,0.0788250342,0.348264873,-0.5426028371,0.2951740623,-0.0744437799,-0.0522126071,0.1970887482,0.0442084596,0.158654213,-0.2298734188,-0.0192227215,0.1169385538,-0.004867496,-0.0748983994,-0.4469334781,0.3034744263,0.1718006283,0.2147850543,0.0117487311,0.0673645809,-0.143504411,0.0583524108,-0.0029330356,-0.1886029243,0.1613481045,0.3425671756,0.0278087314,-0.1712353379,0.0145638864,-0.1994079202,0.0625373498,-0.008083256,0.2212474048,0.241286993,-0.0798238367,-0.1361546963,-0.1645224839,-0.1607881337,-0.0919127911,-0.0036566989,-0.3200867474,-0.0344343521,-0.2061431855,0.1468798816,-0.0806964263,-0.3674132228,0.1646693349,-0.2895928919,0.0637264922,0.4193501174,-0.0656610951,0.1587560922,-0.0810854957,-0.0964359269,-0.1005364954,0.097990416,0.0630263761,-0.1917273402,-0.1016526222,0.1198314577,0.2083981633,0.2010071278,0.1114135906,-0.0428648926,-0.2291794568,0.0643692464,0.1898302436,-0.2369539738,-0.1981433481,0.236365512,-0.2259238809,0.2451674193,-0.0284197386,-0.4693149924,0.0851324871,-0.144576177,-0.0585238338,0.0074144285,0.0539941266,-0.2731432617,-0.1177036464,-0.2989680469,-0.386208266,-0.1980379522,0.2820741534,0.0527222864,0.2734357715,-0.0726147443,0.0076205987,-0.3040592968,0.4685657322,0.1552221924,-0.3286643326,-0.0194332562,0.1838425547,-0.2203141898,-0.1887140721,-0.0422976166,-0.0904345363,-0.0526839793,-0.1176268011,-0.2137936652,0.0841793418,-0.1675969064,0.1001660526,0.0828324854,0.3064706922,0.3374299109,0.1993242651,-0.2635637522,-0.0856798887,-0.3538393974,0.0745084733,0.1289100051,-0.1440953761,0.0223594662,0.3712635934,-0.2230851352,0.2971282005,0.4055191576,0.0535829403,0.3756560087,-0.1485211551,-0.1126590595,-0.1953375787,-0.3681219816,-0.0093424153,-0.1500668228,-0.2315351069,0.2093336731,-0.0445223078,-0.4889579117,-0.0024669825,0.2656967938,-0.8140271306,-0.2107980698,0.1814643294,-0.1368268281,0.0374781862,-0.0845679864,0.0923027769,-0.2396165431,-0.1102337912,-0.0795100555,-0.2985563874,0.2515534759,-0.3308257163,0.0888076723,-0.1440134048,-0.39743343,0.293114543,0.0634500906,0.2471883148,-0.0068137106,-0.1064689457,0.0457209982,-0.0695268363,0.6664331555,-0.3575475514,-0.0807216614,0.3131140172,-0.3457060754,-0.3597396314,-0.1085784212,-0.2446026653,0.4223017395,0.2490160912,0.1651697606,-0.1902461052,0.0813660249,0.3031409383,-0.0155177386,-0.0886725783,-0.1500977725,-0.3474948704,0.0120630777,-0.1366435736,0.1032244861,0.1718515307,0.1410544217,-0.2319599986,-0.2071465105,0.1057457253,-0.0937179625,0.0874392614,0.1244039387,0.3212257922,0.2401578724,0.1785146743,0.2920944691,-0.1577493846,0.3126628101,0.2225326896,-0.2178101838,0.3429147899,0.1411981136,-0.4355608821,0.260258168,0.2999686897,0.2798713744,-0.1091568917,0.0233674105,0.3030960858,-0.3033768833,-0.3331570923,0.3790987432,-0.0484517887,0.1102813184,-0.3353120387,0.3447393477,0.1321620494,-0.0587188229,0.5077552795,-0.0573802181,-0.1065005511,0.4608893991,0.2137577534,0.7910378575,-0.1878990829,-0.0348225236,-0.1420487016,0.0561312251,-0.1971440166,0.0901202261,0.2725308537,-0.5461354852,-0.4394000173,0.0830271021,-0.1587143689,-0.4350050688,0.2416206747,-0.1687009335,0.1569959819,-0.1435948461,-0.4053603411,-0.3324616551,0.0233393554,0.1497024149,-0.1628324986,0.1608368009,0.2037873417,0.2574531734,-0.1328361779,-0.0363434441,0.1462354362,-0.2572633624,0.0861458108,-0.5129283667,-0.1046794727,-0.2424630821,0.4165171981,-0.0738143474,-0.0363195613,-0.2102032155,0.2446124703,0.0141952373,-0.1521608829,0.0803564712,-0.0889950097,0.3105021715,0.2046589255,0.2126334459,-0.0884664506,0.2994721532,-0.0039053399,-0.4743807316,-0.1928721815,0.1596226692,-0.3719160557,-0.0219203867,0.337133944,-0.1180267483,0.0194581933,-0.265550524,0.087341249,0.0585143827,-0.187759012,0.200604856,0.0517514125,-0.0782908723,0.7023655176,-0.127011314,-0.2312026918,-0.0480040871,0.4856057465,0.2352045774,-0.0405891277,0.1515128613,0.3260006905,-0.3293747008,-0.2541389465,0.0042923414,0.0762187243,0.0773218125,0.1749631464,-0.1473270208,-0.1165867746,0.2391569167,0.2125855833,0.0134929288,0.3130646348,-0.2659531832,-0.2406717986,-0.165567711,0.2730347514,0.0203964598,0.1812422127,-0.0341896228,-0.0327733494,0.0266865175,0.1719718575,-0.4117128551,0.0941283256,-0.3399032056,0.1016599834,0.0688778087,-0.1732464284,0.0709691271,0.0608960949,0.2325665802,0.1283658445,0.041657649,-0.2969101667,0.0495159701,0.043257039,-0.1128758341,-0.181707859,-0.0096662268,-0.0501000732,-0.3024848402,-0.3113754988,0.0393278636,0.2253233939,0.0662996545,0.0374973863,0.4365172386,-0.0279010795,0.037096411,0.0683789775,-0.1565673202,0.2994945943,-0.0888125151,0.0638457686,0.1475529969,-0.1218965575,0.0224344581,-0.0766597092,0.0941209868,0.0960456729,0.1418560594,-0.3297684789,-0.1255578399,0.3771896064,0.6326016188,0.1965842396,0.0646505579,0.0106704198,-0.2082908601,0.3497166634,-0.1520095021,-0.0465256758,0.3921131194,0.0849537328,0.0894509256,0.2458789051,0.096635744,0.1112938598,-0.0250349659,0.1211655065,0.1282778829,-0.4195347726,0.0072613303,-0.2902785242,0.1103401557,0.0754344612,0.1623439044,0.0146120209,0.2888438404,0.4516721368,0.1128160134,0.287920326,0.0965842381,-0.2431024909,-0.086193338,-0.3390611112,0.0753159374,-0.1352417618,0.0778428912,0.1244155914,0.2083724141,0.0993629619,-0.4265171289,-0.1266115159,-0.21953924,0.240100652,-0.4704624116,-0.0629105195,-0.0076147602,-0.0582302772,-0.0885787532,-0.1483084708,-0.0676898807,0.1363838464,0.4371735156,0.2658163607,-0.0693716705,-0.2387353629,-0.2568921447,-0.2201757282,0.494489491,-0.197959438,0.0106244842,0.3356945813,0.297298342,-0.0094715515,0.4283316433,0.4837703109,0.2592352033,-0.551934123,0.05091713,0.0397088341,-0.1065053716,-0.1573697031,0.0114666512,-0.060625121,0.3748301268,0.1408799291,0.3006327748,-0.2232872546,0.0280751251,0.3141002059,0.2852795422,0.2101730108,-0.1244818717,-0.1865977347,-0.3570149541,-0.0979559496,-0.2014219463,-0.2138448954,0.0823482871,0.2129533291,-0.4234229624,0.040182054,0.0786066949,0.1550254822,0.1217676848,0.2978001535,0.0791388825,0.3265963495,-0.3561427295,-0.1397109628,-0.4843374193,0.1872500032,0.0585229173,-0.0798806846,-0.0666162819,0.5402233005,0.0532770827,0.0000194208,0.1980641633,-0.1719197184,0.2607088983,0.7007670403,-0.373498261,0.2200193256,-0.2280750871,-0.1617432684,0.1120655835,-0.1107742786,0.0504586734,-0.3848444521,0.2862801552,-0.0855060965,-0.0635394529,0.080091469,0.4555427432,0.3495667577,-0.0937796906,0.4544940591,-0.3589242995,0.0061670542,0.0604162738,-0.1165310964,-0.2515924573,0.0916561559,0.0739616752,0.4379518926,-0.0688909963,-0.0576679409,-0.1749537289,0.0739880204,-0.0019082994,-0.0922780931,-0.2533753812,0.0331603512,0.0801898465,-0.1115740389,0.0184371378,0.3207235634,-0.0257730577,0.3453643918,-0.079788506,-0.4359313846,0.3135235906,-0.1023566574,-0.3047317564,-0.4569961429,0.339549005,-0.1360758096,0.1617169082,-0.0694949552,0.0883439705,0.1941779107,0.0357638486,-0.3385346234,-0.0157026947,-0.0470935926,-0.141555354,-0.1656773686,0.0031907659,0.1155250221,-0.2539273798,-0.2057242543,-0.2042402178]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/597","title":"Indices incorrect with multiprocessing","comments":"Still the case on master.\r\nI guess we should have an offset in the multi-procs indeed (hopefully it's enough).\r\n\r\nAlso, side note is that we should add some logging before the \"test\" to say we are testing the function otherwise its confusing for the user to see two outputs I think. Proposal (see the \"Testing the mapped function outputs:\" lines):\r\n```\r\n>>> d.select(range(10)).map(fn, with_indices=True, batched=True, num_proc=2)\r\nDone writing 10 indices in 80 bytes .\r\nDone writing 5 indices in 41 bytes .\r\nDone writing 5 indices in 41 bytes .\r\nSpawning 2 processes\r\nTesting the mapped function outputs:\r\ninds: [0, 1]\r\ninds: [0, 1]\r\nTesting finished, running the mapped function on the dataset:\r\n#0: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\ninds: [0, 1, 2, 3, 4] inds: [0, 1, 2, 3, 4] | 0\/1 [00:00<?, ?ba\/s]\r\n#0: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1321.04ba\/s]\r\n#1: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1841.22ba\/s]\r\nConcatenating 2 shards from multiprocessing\r\nDataset(features: {'text': Value(dtype='string', id=None), 'label': ClassLabel(num_classes=2, names=['neg', 'pos'], names_file=None, id=None)}, num_rows: 10)\r\n```","body":"When `num_proc` > 1, the indices argument passed to the map function is incorrect:\r\n\r\n```python\r\nd = load_dataset('imdb', split='test[:1%]')\r\n\r\ndef fn(x, inds):\r\n print(inds)\r\n return x\r\n\r\nd.select(range(10)).map(fn, with_indices=True, batched=True)\r\n# [0, 1]\r\n# [0, 1, 2, 3, 4, 5, 6, 7, 8, 9]\r\n\r\nd.select(range(10)).map(fn, with_indices=True, batched=True, num_proc=2)\r\n# [0, 1]\r\n# [0, 1]\r\n# [0, 1, 2, 3, 4]\r\n# [0, 1, 2, 3, 4]\r\n```\r\n\r\nAs you can see, the subset passed to each thread is indexed from 0 to N which doesn't reflect their positions in `d`.","comment_length":163,"text":"Indices incorrect with multiprocessing\nWhen `num_proc` > 1, the indices argument passed to the map function is incorrect:\r\n\r\n```python\r\nd = load_dataset('imdb', split='test[:1%]')\r\n\r\ndef fn(x, inds):\r\n print(inds)\r\n return x\r\n\r\nd.select(range(10)).map(fn, with_indices=True, batched=True)\r\n# [0, 1]\r\n# [0, 1, 2, 3, 4, 5, 6, 7, 8, 9]\r\n\r\nd.select(range(10)).map(fn, with_indices=True, batched=True, num_proc=2)\r\n# [0, 1]\r\n# [0, 1]\r\n# [0, 1, 2, 3, 4]\r\n# [0, 1, 2, 3, 4]\r\n```\r\n\r\nAs you can see, the subset passed to each thread is indexed from 0 to N which doesn't reflect their positions in `d`.\nStill the case on master.\r\nI guess we should have an offset in the multi-procs indeed (hopefully it's enough).\r\n\r\nAlso, side note is that we should add some logging before the \"test\" to say we are testing the function otherwise its confusing for the user to see two outputs I think. Proposal (see the \"Testing the mapped function outputs:\" lines):\r\n```\r\n>>> d.select(range(10)).map(fn, with_indices=True, batched=True, num_proc=2)\r\nDone writing 10 indices in 80 bytes .\r\nDone writing 5 indices in 41 bytes .\r\nDone writing 5 indices in 41 bytes .\r\nSpawning 2 processes\r\nTesting the mapped function outputs:\r\ninds: [0, 1]\r\ninds: [0, 1]\r\nTesting finished, running the mapped function on the dataset:\r\n#0: 0%| | 0\/1 [00:00<?, ?ba\/s]\r\ninds: [0, 1, 2, 3, 4] inds: [0, 1, 2, 3, 4] | 0\/1 [00:00<?, ?ba\/s]\r\n#0: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1321.04ba\/s]\r\n#1: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 1841.22ba\/s]\r\nConcatenating 2 shards from multiprocessing\r\nDataset(features: {'text': Value(dtype='string', id=None), 'label': ClassLabel(num_classes=2, names=['neg', 'pos'], names_file=None, id=None)}, num_rows: 10)\r\n```","embeddings":[-0.3910766244,-0.4112408459,-0.1836059242,0.2711610198,-0.3301049173,-0.1160226539,0.5005037189,0.116604574,0.1060393304,0.4203519523,0.0210538805,0.3124271631,0.0579331666,0.1175346673,-0.2667937875,0.2599059641,-0.0655998066,0.0601689331,-0.3335288465,-0.1833892167,-0.3181264102,0.2125172466,-0.4628998339,-0.1326190233,-0.3110152781,-0.2076466978,-0.066436477,0.1109300777,0.2402825207,-0.1356410086,0.062812075,-0.0520633645,-0.2454714328,0.8725656867,-0.0001055827,-0.1686014384,0.1327808201,-0.1767240614,0.2557112873,-0.0440281928,-0.1471693963,0.1887268722,0.0456941016,-0.3676322997,-0.0554586388,0.0771236047,-0.0901453346,-0.562695384,0.0713642761,0.1461571753,0.2426275015,0.0921505615,-0.2653035223,0.1799373776,-0.0514845178,0.0000009364,0.0786735266,0.1380770206,0.3338859379,0.0206266865,-0.1120761037,0.2574878633,-0.1998972595,0.1658307612,0.0479150973,0.1884689927,0.4429793954,-0.4548024833,0.0140016992,0.1714254171,0.0765016004,-0.0286791045,-0.2724144459,-0.2252686471,-0.2483912706,-0.2456665933,-0.0806914344,0.0676270798,-0.0155447358,-0.1526164263,-0.4177740812,0.2804908156,0.062269792,0.0412678458,-0.0223894585,0.2801387608,0.1688672751,0.1247981191,0.2155305892,-0.0031879039,-0.0951534882,-0.0528317653,0.215305537,0.0714306608,-0.1718955487,-0.0526902415,0.2076423764,-0.2216391414,-0.2764878869,0.0082333321,-0.1167063937,0.1178073958,-0.0458209738,0.1933470666,0.2213090211,0.0341462828,-0.0692190379,0.1834474355,0.2160073817,-0.0012567108,-0.3653990626,0.1079142094,0.3758712709,-0.2926915884,0.2313117534,0.3381733894,-0.3253424466,-0.1112418547,-0.2383358479,0.1973326653,-0.10213238,-0.153431043,0.0871516168,0.1127257422,-0.0100677069,0.2547240257,0.0523730107,0.2664981484,-0.4644142091,0.2895242274,-0.2862216234,-0.076608412,-0.2670091987,0.125899449,0.0665823221,0.0638546273,0.1573699415,0.2249409854,0.4051101804,-0.137728408,0.0269698892,-0.2504670918,0.4778960049,0.137529403,0.1411994845,0.1892673373,0.0875435919,-0.0034042746,-0.2521336377,0.2119447291,-0.1446776837,-0.0629120246,0.0058162473,0.2167235017,0.151655063,0.3152199984,-0.2772162557,0.1823466867,0.3284090459,-0.1004108489,0.0519218817,-0.2757975757,-0.3509438038,-0.2752364874,0.203074947,0.1950675398,-0.0062630852,0.1461381614,0.1258711964,0.2690039575,0.2144625932,0.3467648923,0.0196750183,-0.1099930182,-0.1407660097,0.3324015141,-0.1847784668,-0.3128069639,-0.1643877178,0.3284085691,-0.1849719733,0.0688828528,0.1857278645,-0.0957635194,0.3820743859,-0.0666145459,0.4744650126,0.1066794395,-0.1397284269,0.3166435063,-0.3250187039,0.0013281275,-0.039994549,0.1066809818,0.145159483,-0.2366889417,-0.0859722048,-0.6444997787,0.5220373869,-0.0885261521,0.0595652163,0.1397806406,-0.0939240977,0.2035678774,-0.0551493652,-0.3799473941,0.128973946,0.2552985251,0.0092973942,0.2037033588,0.0678791627,-0.1114592627,0.0459924079,-0.0456701331,-0.0376253836,-0.0912924558,0.1730395555,0.0572230294,-0.2121395767,-0.0726199448,-0.1362882853,0.1295445859,0.2848334014,-0.2879860997,-0.3537170589,0.0829535946,-0.2009760588,-0.2908987999,-0.0033921464,0.3161678016,0.3665759861,-0.1081898138,0.0095778676,0.2655245364,0.2048649639,-0.0258755796,0.0107743731,0.2131205052,0.2367673814,-0.324680984,0.1559289396,0.2504885495,-0.05550991,-0.1659860462,-0.0835673511,0.4199184775,-0.4968352616,0.2824420333,-0.1163877472,-0.1537791491,0.0523724295,0.0249147415,-0.0073133968,-0.2020616382,0.060329359,0.1109678075,-0.1008910239,-0.0723247528,-0.2344715148,0.2854291797,0.1713829488,0.0873312801,-0.0693238974,-0.0081384573,-0.0807934105,-0.0266584493,0.1020047441,-0.2638395727,0.1642602533,0.3359586895,0.1982012987,-0.1459436268,-0.034151081,-0.0403983556,0.1023056805,-0.0844201148,0.2667040229,0.2048464417,-0.0976207331,-0.1244319081,-0.0333177745,-0.1147602573,-0.0911622345,-0.0124763437,-0.3146415949,-0.0698381141,-0.1532041878,0.2369310111,-0.0572197996,-0.3602379858,0.1795164943,-0.3023767471,-0.0132756997,0.4150726497,-0.0780604929,0.1847166419,-0.0846399739,-0.1025816277,-0.1331312209,-0.0831397325,0.0670240298,-0.1013770923,-0.059189938,0.1163101718,0.1773077995,0.21440956,0.1332989633,0.094060801,-0.1922879964,0.1254293919,0.0153614385,-0.2188111693,-0.2710222602,0.1883387864,-0.170957908,0.0261903554,-0.02336278,-0.3288132846,0.1242882162,-0.2972332835,-0.0278342664,-0.1320944875,-0.014627656,-0.2380347699,-0.1538012475,-0.1921212077,-0.2937818468,-0.139625147,0.3042711318,0.1084468514,0.3475036621,-0.1840423346,0.0001092674,-0.2826226652,0.3912380636,0.0776480287,-0.4431112409,-0.141406402,0.1247887164,-0.2343481034,-0.1426165998,-0.1725436747,-0.1278419197,-0.1158872694,-0.0639464557,-0.226484254,0.0080261091,-0.1173777804,0.1169526353,0.0998287424,0.3103832901,0.3031769693,0.2281727195,-0.2444687486,-0.0598863736,-0.4934159219,0.0906015858,0.281075865,-0.1371043324,0.0189824738,0.2403505594,-0.1708362997,0.3510899842,0.5338469744,-0.0504777357,0.1747781336,-0.2027628273,-0.0765639693,-0.0659675151,-0.2839648426,0.160335362,-0.0037910792,-0.2403055429,0.1769826263,-0.0580698997,-0.5434809327,0.0354697257,0.266335696,-0.7126568556,-0.3004765809,0.2838755846,-0.3026221395,-0.081135802,-0.114325799,0.0966883227,-0.3498638868,-0.027823614,-0.0748815462,-0.4129073918,0.2545217574,-0.3162638843,0.2005377412,-0.1877770275,-0.4116976857,0.3155356348,0.2320745885,0.145292446,0.0548219644,-0.0290690511,-0.0356313251,-0.0862541199,0.6001468897,-0.5158306956,-0.0338336304,0.3574095368,-0.3718165159,-0.3988589942,-0.0688081533,-0.1937642843,0.4010441899,0.1016469225,0.1860628575,-0.237863481,0.076669164,0.2436867952,-0.0697648823,-0.0611624271,-0.1438084692,-0.2677903771,0.0517562963,-0.0405487642,0.2658193409,0.078835845,0.1630345434,-0.1532448977,-0.1279492229,0.1677364707,0.0161563251,0.1190767139,0.1811914146,0.2898448408,0.241830796,0.0977225229,0.3397431076,-0.1506508589,0.1656900495,0.2621916234,-0.2122148722,0.513061285,0.2099909335,-0.3418715298,0.2145708054,0.304305315,0.249788031,-0.234781757,-0.013493224,0.3976300359,-0.3324228227,-0.3621887565,0.2209208161,0.0979548469,0.06776537,-0.3353708982,0.1729924232,-0.000670323,-0.1629149169,0.528234005,-0.1989287734,-0.1802892685,0.4125843942,0.3348652422,0.7906064391,-0.2270302474,-0.0294168871,-0.1285948157,0.1680881083,-0.2251191586,-0.0527229384,0.2099865377,-0.3576921523,-0.3432038426,0.0697159991,-0.1616348326,-0.429599762,0.1389186978,-0.1863941103,0.1940371394,0.0250117946,-0.2598969042,-0.3387441635,0.0050125774,0.4230815172,-0.1667315513,0.1342784613,0.2063212246,0.4948371053,-0.125334844,0.0594346076,0.1302832514,-0.350545913,0.1049800441,-0.4203336537,-0.0742862225,-0.3331219256,0.2314799577,-0.2039553225,-0.0209642407,-0.1774664521,0.1904473603,-0.0350683369,-0.1221124232,-0.0045948867,-0.1546927392,0.3599348962,0.334202677,0.1538676918,-0.1454571187,0.3058020473,0.0406036191,-0.435968399,-0.1341864169,0.1346479505,-0.3190563619,0.1046988815,0.2870694101,-0.2308180332,0.0233366154,-0.152099058,0.1466380209,-0.0560922213,-0.0413321145,0.1845853776,0.0453629866,-0.2868357003,0.8450873494,-0.1319475472,-0.4104748666,-0.0337706208,0.4405317605,0.2351767719,0.0902175605,0.0502426699,0.1424871385,-0.2120419741,-0.2564302981,-0.0198031459,0.1026873142,0.1887370199,0.1092246473,-0.2817033231,-0.1780264974,0.2748164535,0.1344033778,0.2005607784,0.4134202302,-0.2734942138,-0.1706080139,-0.1743008643,0.0849121734,-0.0985695571,0.1442488134,-0.0243018698,0.0028561614,0.0284101814,0.1862794459,-0.4245746732,0.1236378327,-0.3775283694,0.1838730425,0.0018026798,-0.1312131137,0.1206206903,-0.0559811182,0.2369425744,0.1135868952,-0.156391561,-0.2942231894,0.0190924983,0.0661202371,-0.0740262941,-0.1426882148,-0.0690259263,0.0293130707,-0.2474026978,-0.354954809,0.077244319,0.1281712055,-0.0082229022,0.0159807261,0.4201312661,-0.1592504829,-0.1846594959,0.0934563205,-0.1307508051,0.2019440532,-0.2125237137,0.1375518888,0.0694650486,-0.1816458702,0.144771859,0.0715746358,-0.0212698225,0.2221682668,0.2003697455,-0.1095193699,-0.0639297441,0.2514486611,0.6699182987,0.1478274912,-0.0035355128,0.0018870659,-0.2060385793,0.315081805,-0.0202073436,0.0604920425,0.3824723959,0.0946361348,0.2593531907,0.3717333972,0.2219747156,0.0257648937,0.0981079042,0.0108806044,-0.0218820013,-0.3711984158,-0.0489234552,-0.1159257367,0.0799683034,0.1274136454,0.3494866788,-0.0602412634,0.3144944906,0.4296078086,0.2094264776,0.3505512178,0.0553855002,-0.2268168777,-0.1631907076,-0.2393540293,0.1994140297,-0.0032796389,0.0517141335,0.2685461938,0.2488972694,0.1601736248,-0.5418087244,-0.1194700673,-0.221173346,0.1755756587,-0.3712435067,-0.0089359693,-0.0270312522,-0.1240661517,-0.1185614467,-0.162326321,-0.111340344,0.1844199896,0.5847884417,0.2405048013,-0.0501866937,-0.0658198893,-0.4881676733,-0.2559200823,0.5576311946,-0.1734602302,0.0864135101,0.3961074352,0.2808990777,-0.049303595,0.4639000297,0.4753144681,0.1773909926,-0.6370459199,0.087715514,0.1971673071,-0.0477365255,-0.2354268134,0.0636435151,-0.0134590846,0.205356285,0.1832570583,0.2913367152,-0.2434183508,0.0983013958,0.2694970667,0.108328931,0.3435380161,-0.1505181193,-0.0212081317,-0.2053917795,0.0279427916,-0.1470689029,-0.1231900379,0.0789557323,0.164856866,-0.4540506005,0.2122903466,0.1642923653,0.1567545384,0.1446087211,0.3897768855,-0.0588466562,0.2958666086,-0.304759413,-0.1029704362,-0.5643629432,0.2311453372,-0.0094629256,-0.1113327593,0.0587619804,0.4745410085,0.2023715079,0.0816897526,0.220358327,-0.0922070742,0.2018158883,0.7049713731,-0.4164034426,0.20996508,-0.2217618227,-0.174158752,0.1488358825,-0.1768952161,0.0444816872,-0.372952491,0.3136249483,-0.0362442844,0.0237336494,-0.0098154387,0.4138358533,0.4208663702,0.0215061996,0.4189323783,-0.4326876998,-0.0012490955,0.0622589327,0.2253861725,-0.3149060309,0.0232639909,0.0010216222,0.3829727471,-0.1051049456,-0.1056710109,-0.2199730426,0.0568579882,0.121207267,-0.0350118056,-0.2539077103,0.0556137636,0.0466686748,-0.0884667486,-0.0444155969,0.4599960744,-0.0279925931,0.3139511943,-0.0548091978,-0.2318788618,0.2483801693,-0.1875474751,-0.2203119099,-0.3956333995,0.277998656,-0.2484607548,0.1722343415,-0.0953582302,0.0872727334,0.1940007955,0.0580677278,-0.4339934886,-0.0281170662,-0.1502599269,-0.1491986215,-0.135263145,-0.0007944359,0.1186170876,-0.2021310776,-0.1844308972,-0.1668343842]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/595","title":"`Dataset`\/`DatasetDict` has no attribute 'save_to_disk'","comments":"`pip install git+https:\/\/github.com\/huggingface\/nlp.git` should have done the job.\r\n\r\nDid you uninstall `nlp` before installing from github ?","body":"Hi,\r\n\r\nAs the title indicates, both `Dataset` and `DatasetDict` classes don't seem to have the `save_to_disk` method. While the file [`arrow_dataset.py`](https:\/\/github.com\/huggingface\/nlp\/blob\/34bf0b03bfe03e7f77b8fec1cd48f5452c4fc7c1\/src\/nlp\/arrow_dataset.py) in the repo here has the method, the file `arrow_dataset.py` which is saved after `pip install nlp -U` in my `conda` environment DOES NOT contain the `save_to_disk` method. I even tried `pip install git+https:\/\/github.com\/huggingface\/nlp.git ` and still no luck. Do I need to install the library in another way?","comment_length":17,"text":"`Dataset`\/`DatasetDict` has no attribute 'save_to_disk'\nHi,\r\n\r\nAs the title indicates, both `Dataset` and `DatasetDict` classes don't seem to have the `save_to_disk` method. While the file [`arrow_dataset.py`](https:\/\/github.com\/huggingface\/nlp\/blob\/34bf0b03bfe03e7f77b8fec1cd48f5452c4fc7c1\/src\/nlp\/arrow_dataset.py) in the repo here has the method, the file `arrow_dataset.py` which is saved after `pip install nlp -U` in my `conda` environment DOES NOT contain the `save_to_disk` method. I even tried `pip install git+https:\/\/github.com\/huggingface\/nlp.git ` and still no luck. Do I need to install the library in another way?\n`pip install git+https:\/\/github.com\/huggingface\/nlp.git` should have done the job.\r\n\r\nDid you uninstall `nlp` before installing from github ?","embeddings":[-0.0576853193,0.3064997196,-0.024444893,0.095245108,0.1892760843,0.2053219229,-0.0553886071,0.1952711195,-0.1948140264,-0.0407715924,0.1610014737,0.6358869076,-0.2991715968,-0.0846181512,0.4080449939,0.0153641645,0.2040481418,0.4670255482,0.0720534474,-0.2329986393,-0.1809488535,0.3781033754,0.0341222435,0.0287979152,-0.1356095374,-0.1028221548,-0.2556759119,-0.1338198632,-0.0067095151,-0.4790412784,0.5973632932,0.0164337642,-0.1387768835,0.1500869244,-0.0001247489,-0.1039686576,0.1948121935,-0.090165101,-0.5097925663,-0.1813150495,-0.095452413,-0.4069551528,0.4231102169,-0.3911996186,0.0021168832,-0.0709338486,0.3261789083,-0.0896711871,0.2776719332,0.239450857,0.1396989077,0.0691265687,0.2781968415,-0.0320287347,-0.0457666405,0.3639844358,-0.2870591581,0.168753773,-0.0774382949,-0.0247770045,0.272967279,0.0327313095,-0.0990402102,-0.1614608765,0.3065212071,0.15005292,-0.2726331949,-0.1285836548,-0.1275789887,-0.123738572,0.4632946551,-0.4910497963,-0.3580473959,-0.1172125936,0.1763523668,-0.3220726252,0.1518509239,0.2760243118,-0.2431336492,0.0617489628,-0.1482972503,-0.4829071462,-0.2292247266,0.3118004501,0.1502443701,-0.011266958,-0.2004208416,0.0965088382,0.2891165316,0.0886989757,0.0511457473,-0.044186309,0.1211484522,0.0314178355,0.1531008631,-0.1660218686,-0.0288153552,-0.0116273686,-0.0508583374,0.3493904769,0.1532950848,-0.2702512443,-0.0151150012,0.2592653632,0.0581006818,0.2494726926,0.5383609533,0.3887923658,0.4372162223,-0.1594181806,0.2229388058,-0.0370935649,0.1210599616,0.0154879522,0.1832456887,-0.2957262099,0.2924045026,-0.3235220909,-0.284522891,0.0290026981,0.1464640796,-0.075319469,-0.0601791367,0.280354321,0.1070963517,0.2418390512,0.1395728588,0.359318465,0.0801763237,-0.0368359983,-0.1544838101,0.0541916825,0.0157968849,0.1448024362,0.5917967558,-0.0685569197,0.3948677182,-0.0441050641,-0.3212051392,0.0737070963,0.0234104004,-0.1391793936,0.0231919531,0.2737842798,0.0101857483,0.2499607652,0.0163397007,-0.2614886761,-0.3220079839,0.1296747923,-0.0567860641,-0.0743324012,-0.043953985,0.0210302379,0.114560917,-0.3268068433,-0.2669492066,0.1397685409,0.3172675073,-0.3323731124,0.0321769118,-0.1229317635,0.0017140181,-0.4182407558,0.372102648,0.0798885375,-0.191876173,-0.178402856,-0.0792943984,-0.0416536443,0.2084515393,0.3483615816,-0.0795303956,0.1954371482,-0.1444522589,0.0965556651,0.8378008604,-0.5005511642,-0.5021432042,-0.0146165071,0.2384306043,-0.139271602,-0.1509952843,0.2623471618,0.0058822003,-0.1259303689,0.144475624,0.2695761919,0.1040392816,0.01575079,-0.3102188408,-0.0621160753,0.2738034427,0.0696571618,-0.1182746589,0.0085191336,-0.0438954644,0.0914581865,0.2146680951,0.0336318016,0.1904893368,0.2619516253,0.2920945287,-0.0155354124,-0.295684725,-0.1719345152,-0.3159604967,-0.0912840366,0.0618651658,-0.2314924449,-0.1292692125,-0.0992313102,-0.04523791,-0.0977917463,-0.1335943788,-0.2724175155,0.009216818,0.0959903449,0.2064702362,0.3455800116,-0.4039347768,0.2961389124,-0.0212742984,0.1568083465,-0.2580255866,0.2367762774,-0.2979541421,-0.3281771839,-0.092259787,0.4159758091,0.0844305381,-0.0270584468,-0.0332022049,0.3201881945,-0.3143803477,-0.1482999623,0.1487742662,-0.0822434127,0.200581342,-0.4163779318,-0.0879674181,0.0980151892,0.1353491843,0.0386444479,-0.3379626274,0.1252183765,0.152457878,0.1615891457,0.2309930027,0.2231138945,0.2986209691,-0.0681873709,-0.092169404,-0.0746713132,0.2888090014,-0.1447555721,0.4032860994,-0.2652861178,-0.3689486682,0.1862704903,0.573530674,0.1868016422,0.5875403881,0.2163973749,-0.1049691439,-0.0852610841,-0.04492075,0.469073683,0.6300858259,0.1020102575,0.0310124718,-0.1670400202,-0.4748409986,-0.2145324498,0.2900676429,-0.0415595211,0.1297402531,0.1205121353,0.1305897683,0.0523800328,-0.3836540878,-0.0486110412,-0.1777878404,0.0709195286,-0.068803139,-0.042856276,-0.396068275,-0.1929874122,-0.1187582165,-0.3759918511,-0.2696293294,-0.363648802,-0.048149325,0.1673814505,0.0596651398,-0.0209342111,-0.3365979791,0.3685643673,-0.3781700432,-0.098981075,-0.3116406202,-0.1323969811,-0.2740909159,-0.0800352097,0.1156490371,0.1323539019,0.4838153422,-0.1280393749,0.0651177838,-0.1661777496,-0.12330845,0.0426873192,-0.1347209811,0.2439864129,0.0614738241,0.1860231459,-0.3289710581,-0.2357398421,0.3705380559,-0.1777910739,-0.192727685,0.0620921217,0.0706929937,-0.3187223375,-0.2763979435,-0.2736289799,-0.4865009487,-0.4724040627,0.4291748703,0.0022045847,0.2098549753,0.1633350402,-0.1868857145,0.3642731011,-0.0969282761,0.3241606653,-0.0344818309,-0.1116034612,0.4596477151,-0.4573006928,-0.5065928698,0.0659337789,0.1178522334,0.3186712265,-0.0784476399,-0.476791501,-0.0632539764,-0.3380720317,0.1348123997,-0.0090778386,0.1956044436,0.3882959485,0.0982562602,0.0456519388,-0.1662946641,-0.4008443654,0.3753771782,0.1130128503,0.3920342922,-0.1350152493,0.1833522469,-0.0676996037,0.290586412,0.2983123362,0.0744908974,0.4115737081,0.3516529202,0.4719513357,-0.1254168302,-0.5036275983,0.041128505,0.0504853241,0.1892404407,0.0555853732,0.069116503,-0.0055980133,-0.1221466959,-0.0195501279,-0.2420723289,-0.2064187676,-0.0650286227,-0.0861940607,0.1582929641,0.033402577,0.0539824329,0.0585006662,0.030189978,-0.0715040416,0.5122178793,-0.0429132357,0.2273364663,-0.5592311025,-0.4526760876,-0.705928266,0.1493568569,-0.2508864403,0.4395971894,-0.1966259032,0.1666858047,0.1066768393,-0.0994888619,0.3977792561,0.0314260572,-0.2973347008,0.1747869998,0.1483101547,-0.6878681183,0.0245514065,-0.0301458836,0.0808392838,0.1477579772,0.2117551267,-0.3747902513,-0.23223041,0.0214668587,0.0772843882,-0.1316453665,0.0094017293,-0.1647864729,-0.4430541098,-0.2546954155,-0.0852768943,-0.0267475862,0.1087700427,0.0268110353,-0.0069708149,-0.2921901047,-0.154745087,-0.0299637914,-0.1773817986,0.4007046819,-0.1765723825,0.0162970461,-0.1245777234,0.2342345268,0.0951604694,0.3477655947,-0.2530573606,-0.0623925738,0.1805556417,-0.0157294385,0.4948554337,0.0750751048,-0.0007128461,-0.1263858676,-0.1417503655,-0.1267608851,-0.0193339158,-0.1548913121,0.1679998189,-0.1604673415,-0.3305188417,-0.1659294963,0.4436079562,0.1086935624,0.081320256,0.3405788839,0.261169225,-0.1473250687,0.4154503047,-0.0161066819,0.821379602,0.1217929646,0.6255540848,0.1412681192,-0.3488842845,0.6622038484,0.194335416,0.3420631289,-0.2074201405,0.0528821573,0.0063349996,-0.1722123027,0.1727356017,0.0399780348,-0.5027002096,0.3164102733,0.1324860603,-0.5379626155,-0.023886038,0.1237440258,-0.1276747882,-0.2993037701,-0.4431585968,0.1207480878,-0.2031995356,0.3285003603,-0.0563368015,-0.0577970892,-0.3582288027,-0.1257172823,-0.2703658342,0.2586686611,-0.1589595079,0.6079386473,0.0830378532,-0.3375372887,0.0971336663,0.2447013408,0.5418917537,0.0794953927,-0.3354737461,-0.0393897295,-0.2932339609,-0.0141564822,0.0426776931,0.0584299937,0.3304018676,-0.1745100915,0.1848755181,-0.2277125865,-0.2636816204,-0.3943451047,-0.0752204582,0.0595055781,0.0209824704,0.0394042768,-0.2589337826,0.174989745,-0.0322130695,-0.3993843794,0.0444162041,0.3888375461,0.1419640034,0.3495690227,0.021703247,-0.0110557172,0.0761491507,0.1599625349,-0.0485818833,-0.1361466795,0.6374306083,-0.2584861815,-0.1061139628,-0.2650417387,0.0691273734,-0.0883132368,-0.3532879949,0.2797167003,-0.0579953454,-0.1244320124,0.0427911505,0.1967556477,-0.002386742,-0.1989201605,-0.1143311113,-0.3767811656,-0.4714045227,0.1940356195,0.1989244074,0.2183165252,0.2220207453,0.0166334957,0.35495314,0.0191396363,-0.1581320316,-0.1340224892,-0.2352516204,-0.0110493638,0.3107497692,-0.0593305379,0.1848315746,-0.0213299338,-0.0679536611,0.0031086674,-0.1043018848,-0.1208000332,-0.2125733644,0.1969395429,0.1923291981,0.1092870459,-0.1313858181,-0.239730984,0.0439847149,-0.1399281174,-0.4426007867,0.0799328238,-0.088881135,0.3664726913,0.0575383827,0.2464978397,0.1441567391,0.0473046415,-0.0154911885,-0.0473275334,0.0799996257,0.082598716,-0.0391018651,-0.0425978862,-0.444765389,-0.0654554144,-0.122284174,0.1487655789,0.0366333462,-0.380248785,0.0044799317,-0.0247702114,0.1926673204,0.3444615304,-0.0721338317,0.1401769072,0.3488791585,0.0576965213,-0.1214595586,-0.1246099547,0.294968307,0.0958726779,-0.3309558928,-0.0350722596,-0.0882490501,-0.1412103623,-0.2311881632,0.0485368632,0.2422637492,-0.5197973847,0.2657166421,0.2405587882,0.2196430117,0.1018091887,0.1921789348,0.0511509068,0.2464906573,0.2721923292,-0.2552722096,0.3102762103,0.2736412585,-0.0309983566,0.3919941187,0.2052751184,-0.1697526723,-0.1164802611,0.2805294096,0.2118795663,0.0934594795,0.485799998,-0.5065950155,-0.3788502514,-0.068605639,0.4413181245,-0.0941783711,-0.0603893995,-0.1129988581,0.1150919944,0.2628128827,-0.1528170109,-0.2369879931,-0.1597889662,0.1075635999,0.2128528953,0.0177869685,-0.5339607596,0.2798570693,0.1498137712,0.0744921118,0.0285126641,0.2995051146,-0.1057929173,-0.2620378733,-0.4199061692,0.6136077642,0.336355716,0.4281281233,0.1993324012,0.0982783586,-0.0779116526,0.0065818145,-0.1714570969,0.2047605515,0.4125988185,-0.0030339947,0.2666642964,0.0383117348,0.038198933,0.0080967015,0.0181995742,0.1316437721,-0.0388864987,0.7803910375,-0.1914765835,-0.5222330689,0.0024654716,0.3051825464,-0.4095509648,0.1250982732,0.1826255471,0.0405016541,0.0143567417,-0.279943049,0.0108898813,0.1269236058,0.3953199089,0.2955888212,0.0432539769,-0.1514935046,-0.3134215474,-0.5588613749,-0.1181414872,0.1788040549,0.1157040745,-0.0851334855,0.2500913143,-0.1083274409,-0.0275565628,0.1249398813,0.2288715243,0.1746481955,-0.0397468023,-0.3313179612,0.1571530253,0.1166601107,0.1270059496,0.1858627498,-0.1724362671,0.1059407815,0.0519053712,-0.0609420612,0.091280736,0.295347631,0.0733895525,0.2903586626,0.3931970894,0.0932849422,0.3351484537,-0.0508647896,-0.3590811789,0.1334628016,-0.0794146508,-0.1128136292,0.163249284,-0.0702653974,0.3910033703,-0.4936114848,0.0850493237,-0.1855799258,0.14908126,-0.1309607774,0.1026666164,-0.0199993905,0.1764273793,0.0601875968,0.1169024557,0.2232003212,0.139383018,-0.1337046176,0.1963238567,-0.6758767962,-0.2496201098,0.3895045817,-0.2825267315,-0.1214378402,0.1087150425,0.3584265709,0.1816571057,-0.0227926839,-0.587895155,-0.3907533586,0.2333760113,-0.0768755749,-0.3335980475,0.3162534237,-0.334321171,-0.019226091,0.0485633127,0.7234624028,0.1110987067,-0.3040368557,0.1088188067,-0.1862653792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/595","title":"`Dataset`\/`DatasetDict` has no attribute 'save_to_disk'","comments":"> Did you uninstall `nlp` before installing from github ?\r\n\r\nI did not. I created a new environment and installed `nlp` directly from `github` and it worked!\r\n\r\nThanks.\r\n","body":"Hi,\r\n\r\nAs the title indicates, both `Dataset` and `DatasetDict` classes don't seem to have the `save_to_disk` method. While the file [`arrow_dataset.py`](https:\/\/github.com\/huggingface\/nlp\/blob\/34bf0b03bfe03e7f77b8fec1cd48f5452c4fc7c1\/src\/nlp\/arrow_dataset.py) in the repo here has the method, the file `arrow_dataset.py` which is saved after `pip install nlp -U` in my `conda` environment DOES NOT contain the `save_to_disk` method. I even tried `pip install git+https:\/\/github.com\/huggingface\/nlp.git ` and still no luck. Do I need to install the library in another way?","comment_length":28,"text":"`Dataset`\/`DatasetDict` has no attribute 'save_to_disk'\nHi,\r\n\r\nAs the title indicates, both `Dataset` and `DatasetDict` classes don't seem to have the `save_to_disk` method. While the file [`arrow_dataset.py`](https:\/\/github.com\/huggingface\/nlp\/blob\/34bf0b03bfe03e7f77b8fec1cd48f5452c4fc7c1\/src\/nlp\/arrow_dataset.py) in the repo here has the method, the file `arrow_dataset.py` which is saved after `pip install nlp -U` in my `conda` environment DOES NOT contain the `save_to_disk` method. I even tried `pip install git+https:\/\/github.com\/huggingface\/nlp.git ` and still no luck. Do I need to install the library in another way?\n> Did you uninstall `nlp` before installing from github ?\r\n\r\nI did not. I created a new environment and installed `nlp` directly from `github` and it worked!\r\n\r\nThanks.\r\n","embeddings":[-0.0511759333,0.3495239913,-0.0176616888,0.1092848703,0.1668482125,0.1748827398,-0.0380238071,0.1976730078,-0.1885312051,-0.043285016,0.1474733651,0.6693127155,-0.3041138947,-0.0753769055,0.4125314057,0.016558148,0.2039823532,0.4532092214,0.0589582846,-0.2329033464,-0.1727200449,0.3684439659,0.0128143821,0.0299812425,-0.1168765053,-0.0699494556,-0.2774166167,-0.1472542584,0.0213420279,-0.4878751338,0.6232535839,0.025658058,-0.0995816737,0.1401851177,-0.0001252176,-0.0947313309,0.2191047817,-0.0743310079,-0.4842466116,-0.186062932,-0.1118605658,-0.4106766582,0.4256405532,-0.4007542729,-0.0077748084,-0.1093160957,0.3234489858,-0.0821146891,0.2835487127,0.2425789386,0.1348460019,0.0490633287,0.2647358179,-0.0296148248,-0.0352693498,0.3740502894,-0.2868793011,0.1654687673,-0.0698279291,-0.0520019419,0.2786721885,0.0281591397,-0.1160386205,-0.148350805,0.2933444381,0.1480827779,-0.2648943067,-0.1451873034,-0.1403990388,-0.1053526253,0.4972685874,-0.4533620775,-0.3359110355,-0.0857649148,0.1459582001,-0.33955881,0.1564502567,0.2751291096,-0.2350642234,0.0538978092,-0.1645874381,-0.4610552192,-0.2480334491,0.3096344173,0.1412266195,0.0225567967,-0.1946240664,0.1092327237,0.2540315688,0.0773193613,0.0809601694,-0.039671652,0.1217784435,0.0294953454,0.1341829747,-0.1684680432,-0.0206556562,-0.025548039,-0.0454425476,0.3416440189,0.1140734181,-0.2621801198,-0.0336799212,0.2551231682,0.0709185377,0.2513377964,0.5684151649,0.3965986669,0.45133394,-0.180692777,0.1973213255,-0.0392242447,0.1051931754,-0.0114816437,0.1841273308,-0.2710450292,0.2821151018,-0.3572694361,-0.2509005964,0.0311010964,0.1265180856,-0.0671607479,-0.0512965843,0.2768733203,0.1140203625,0.2726812065,0.1592994779,0.3604344428,0.0709906071,-0.0292262845,-0.1366823316,0.0375475399,-0.000931723,0.152721554,0.5961227417,-0.0289045926,0.4048804641,-0.0496452525,-0.329703033,0.0721500665,0.0249023866,-0.1425147653,0.0376485698,0.2746228576,-0.0059602386,0.2406769544,0.0008031203,-0.2513996065,-0.3149660528,0.1724728793,-0.0685228184,-0.0784183443,-0.047056064,0.0120060993,0.1201016381,-0.3300995529,-0.2568689883,0.1341209412,0.3234741092,-0.319558531,0.026189249,-0.1375035048,-0.0047139255,-0.4191738963,0.3578749895,0.0753017217,-0.2224448323,-0.1512532532,-0.0840386152,-0.0440831594,0.2268172503,0.3499010503,-0.1063032374,0.1825851202,-0.1158606783,0.0924044997,0.8516139388,-0.489674896,-0.4979109168,-0.0161963794,0.2707594633,-0.1674655676,-0.1509110034,0.2612154186,0.0241094604,-0.1258374453,0.1311420649,0.2887346447,0.1377986521,0.0192628074,-0.3267795742,-0.0591942333,0.3038318753,0.0636950061,-0.1093660668,0.0202697143,-0.0482570603,0.0815662667,0.221841231,0.0228147805,0.1908590049,0.2665262222,0.3088770509,-0.0222313702,-0.2593230009,-0.1783483773,-0.2849615812,-0.0821536556,0.0372208245,-0.211175397,-0.0952329338,-0.0784115791,-0.0413034707,-0.0949671715,-0.1408421993,-0.2969379127,-0.0042201378,0.0681393966,0.206966117,0.3199113607,-0.4070872962,0.2959427536,-0.01905833,0.1253743768,-0.228302747,0.253534615,-0.3011040092,-0.3270973563,-0.094833374,0.3961188793,0.0923491791,-0.0210860353,-0.0583561398,0.3021885753,-0.298795104,-0.1604407281,0.1566125453,-0.1066597253,0.21144633,-0.4306330681,-0.0665082633,0.0895551071,0.1336699575,0.0361181013,-0.3419333696,0.1272631884,0.1734529436,0.1750142872,0.2303904295,0.2305631787,0.3229705989,-0.0843192935,-0.0659552217,-0.069305107,0.2749518156,-0.1217547134,0.4015532732,-0.2916473746,-0.3590727448,0.2103360742,0.555231154,0.2151245475,0.5608869791,0.2083569914,-0.1154359728,-0.1199423149,-0.0386915579,0.4832510352,0.6548285484,0.0990174189,0.0429351702,-0.1476528794,-0.4620834589,-0.2071966678,0.3073485792,-0.0605245456,0.134557873,0.1047619283,0.1706264466,0.0544817634,-0.3766708076,-0.0536191016,-0.1458656192,0.0812437162,-0.0552989841,-0.0335780419,-0.4216027856,-0.2106356025,-0.100279443,-0.3712030649,-0.2571751475,-0.3521026373,-0.0483965129,0.1493587047,0.0526550226,-0.0050586625,-0.3272660375,0.3506123722,-0.3900284767,-0.0734015554,-0.3050143421,-0.1633831412,-0.2843654752,-0.0878532529,0.1165301204,0.1370327622,0.465826571,-0.1015403792,0.0480973721,-0.1767470092,-0.1433378458,0.0440041907,-0.1066029817,0.2553695142,0.0543655939,0.1757211238,-0.3232786059,-0.2217730731,0.3675338626,-0.1886543483,-0.1849095672,0.0674366131,0.0623550378,-0.319128722,-0.2779297233,-0.2873338163,-0.5013769865,-0.4498511255,0.374222219,0.0061319438,0.2335651666,0.162081033,-0.2169314623,0.3564762175,-0.080360204,0.2904996276,-0.0246064477,-0.1150031388,0.4654689729,-0.4859730899,-0.5153086782,0.0681334138,0.1344354749,0.3015755713,-0.0698592365,-0.4759966135,-0.0519197769,-0.3486807346,0.1344190687,0.0311377272,0.1858066618,0.4061759114,0.1082538962,0.042835217,-0.1794448644,-0.404648453,0.3869394362,0.1595667154,0.4086622298,-0.1234142706,0.1844185442,-0.0775087178,0.3114279509,0.2893519998,0.0900779441,0.4319112599,0.3623682857,0.4333250225,-0.1328260303,-0.4879168272,0.0329262801,0.0417347625,0.1875583977,0.0343163423,0.0796160996,-0.0590602942,-0.1343892068,-0.0549027547,-0.2365455478,-0.1970096081,-0.0738754421,-0.0758688226,0.1651827097,0.0359006785,0.0366294496,0.0521710552,0.012173322,-0.094113186,0.5079709291,-0.0324252956,0.2057265192,-0.5988534093,-0.4365368783,-0.6885825396,0.1605245769,-0.280225426,0.4618181288,-0.1839823276,0.1600230187,0.1078511477,-0.118913047,0.3920049667,0.0264730193,-0.2874025404,0.1963045299,0.1640446335,-0.6690198183,0.0100666294,-0.022141505,0.077490516,0.1449770629,0.1754987389,-0.334723413,-0.2417745739,0.0204197653,0.0821193159,-0.1349849999,0.0013745015,-0.1567968875,-0.4167924523,-0.2565547526,-0.0940174982,-0.0602870174,0.1025560349,-0.0005179733,-0.0125447791,-0.2849370837,-0.1403102875,-0.0548360832,-0.1625193357,0.4008567333,-0.2031629235,0.0290592238,-0.1423399299,0.216147393,0.1099639013,0.3306187093,-0.2481497973,-0.0624475218,0.1909237206,-0.0227074064,0.5063283443,0.0863353461,-0.0102373585,-0.1358548552,-0.1407329589,-0.1478153616,-0.0070227915,-0.1694612801,0.1523397863,-0.1671595424,-0.3186496198,-0.1829833686,0.4536823332,0.1151347831,0.069384791,0.3714612722,0.2221020311,-0.1585942656,0.4316373467,-0.0247000642,0.828610599,0.1389380544,0.6248913407,0.1214362159,-0.3604215384,0.6662311554,0.2118095607,0.3587363362,-0.233470872,0.0666393191,-0.0030154111,-0.1960456222,0.1963265091,0.0513781272,-0.4806956649,0.328730762,0.1060484201,-0.5526173115,-0.0421291776,0.1051839739,-0.08625862,-0.2951768041,-0.4481681287,0.1107847914,-0.1988897771,0.3293923438,-0.0434377119,-0.0704307184,-0.358810097,-0.1231829673,-0.2659948468,0.2920839489,-0.1017145142,0.6084001064,0.0909377709,-0.3310434818,0.1179759651,0.2404784858,0.5367372632,0.0682954192,-0.3451304436,-0.0286538508,-0.2723813355,0.0116608087,0.0727143958,0.0613146611,0.3213412762,-0.1756274253,0.1509183496,-0.2304440141,-0.2583704889,-0.3924555779,-0.0645197853,0.0728890896,0.0564477183,0.0609993301,-0.2666788697,0.1987739801,-0.0558424741,-0.391569525,0.0442522541,0.3813622594,0.1286257356,0.3738574386,0.0004818655,-0.0010865016,0.0977271572,0.1469855309,-0.0333756916,-0.1118766218,0.6497093439,-0.2514716089,-0.1003414989,-0.2615315616,0.0792426169,-0.1083051413,-0.3478527665,0.2586364448,-0.0585949942,-0.0978532955,0.0621515997,0.1924937516,-0.0522924103,-0.1827872247,-0.1236050725,-0.398281306,-0.4713499546,0.1672695428,0.1461129636,0.2080514282,0.2146076262,-0.0032129656,0.3686027825,0.0062746587,-0.1483167559,-0.1474383473,-0.2444085479,0.0028681315,0.2821365297,-0.0443981886,0.1526605487,-0.0413823761,-0.0892752931,-0.0028670512,-0.1005368531,-0.1137332618,-0.2085812539,0.2042339742,0.1988555491,0.0803739056,-0.1361831725,-0.2762062252,0.0311327185,-0.1081657559,-0.4312265217,0.0681482702,-0.0990016311,0.3777635992,0.0549809858,0.2136466205,0.133639589,0.071166195,0.0015579856,-0.0568110459,0.0871941522,0.0867152065,-0.0195388831,-0.0442278236,-0.4699700475,-0.0777489319,-0.1660931557,0.1743431538,0.0513615906,-0.3840723336,0.0083663762,-0.0289265141,0.1852200031,0.368046999,-0.0486997366,0.1466160417,0.3838507831,0.0455298796,-0.1440126449,-0.1348591149,0.3171977997,0.1027926803,-0.3474206328,-0.0502427816,-0.0638296679,-0.1537359059,-0.2240031958,0.0584658943,0.2263319194,-0.5590143204,0.2600129545,0.2483329475,0.2128687948,0.1100928485,0.205261901,0.0609771349,0.2157885283,0.2888847589,-0.2264429331,0.2971996367,0.3332091272,-0.0283934493,0.3937422037,0.2103585899,-0.1808766425,-0.1364284605,0.2831426561,0.1936683208,0.0886232331,0.4431955814,-0.4910355508,-0.3830980062,-0.0617898852,0.4575684965,-0.0753065795,-0.0721779764,-0.1289487928,0.1098828912,0.2873890996,-0.1455773413,-0.2623184323,-0.1291571856,0.1120995879,0.2135374844,0.0100195324,-0.5514196754,0.2878574431,0.1470237076,0.0715338364,0.0286219008,0.3166800141,-0.1223571301,-0.2435407788,-0.4203837216,0.6235864162,0.3259895742,0.4648938179,0.1818107218,0.0967918038,-0.0837049633,0.0049342127,-0.1630791128,0.2107363194,0.4334670305,-0.0168520696,0.252065748,0.0186228771,0.063038148,0.0313766226,0.0300566237,0.1394676715,-0.0340466537,0.7894078493,-0.1301584095,-0.5200198889,0.0189709738,0.2907419205,-0.3882836699,0.139614746,0.2263669819,0.0553828664,0.0139351068,-0.2838944495,0.0048551373,0.1097202525,0.4014550149,0.2755790055,0.0409511514,-0.1465991139,-0.2888380885,-0.5644543767,-0.1432638317,0.1710913628,0.1132695749,-0.0880268514,0.2399018258,-0.0980883986,-0.0342172496,0.1278224289,0.2346996665,0.1664561182,-0.044813294,-0.3491308093,0.1687341481,0.1297890842,0.1550358534,0.1488062143,-0.1924624741,0.0946161821,0.0623641722,-0.0870193467,0.1077696979,0.2719551921,0.0959563479,0.3114097416,0.3811461926,0.1047642529,0.3348769546,-0.0284328051,-0.3347637653,0.1108492836,-0.095137693,-0.0980143622,0.1521931738,-0.0495534725,0.3603442013,-0.4953613877,0.0990428478,-0.1936009973,0.1300020069,-0.1606476009,0.0955838263,-0.0119370827,0.1552874446,0.0668632388,0.1224892363,0.2099285126,0.1312448829,-0.1441489905,0.167751506,-0.689008832,-0.2295207679,0.3672010601,-0.2939682305,-0.1553182453,0.0923160091,0.3587689996,0.163801387,-0.017119823,-0.5786731839,-0.4102850854,0.2423870265,-0.0847046971,-0.3228687048,0.29540959,-0.3000365794,0.004037411,0.0361742899,0.6977641582,0.121036768,-0.294203788,0.0681874678,-0.1811480522]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/590","title":"The process cannot access the file because it is being used by another process (windows)","comments":"Hi, which version of `nlp` are you using?\r\n\r\nBy the way we'll be releasing today a significant update fixing many issues (but also comprising a few breaking changes).\r\nYou can see more informations here #545 and try it by installing from source from the master branch.","body":"Hi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```","comment_length":46,"text":"The process cannot access the file because it is being used by another process (windows)\nHi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```\nHi, which version of `nlp` are you using?\r\n\r\nBy the way we'll be releasing today a significant update fixing many issues (but also comprising a few breaking changes).\r\nYou can see more informations here #545 and try it by installing from source from the master branch.","embeddings":[-0.1765670329,0.1705409884,-0.0338350125,0.1972525716,0.276063025,0.046986036,0.2461643964,0.3142565191,0.1317553222,0.1372045875,-0.0612205826,0.529302001,-0.3115153611,-0.0680440515,0.0767634884,0.014277596,-0.0274664685,0.1223226339,0.0646729916,0.1522342712,-0.4235683978,0.089819245,-0.321685195,0.3932027221,-0.2467512786,-0.2833744586,-0.0073881098,0.4002689421,-0.1476258785,-0.3065513074,-0.1500901431,-0.1016815007,0.1635933965,0.339730531,-0.0001092507,0.1117590815,0.1657463014,0.0723258033,-0.0584406592,-0.186177507,-0.0081050647,-0.4370200336,-0.0492740385,-0.0949192494,0.2509933114,-0.0299171451,0.2664938271,-0.2296128422,0.3667313457,0.4969288409,0.2682473063,0.2371559888,0.1719648689,-0.0822591037,0.4717095494,-0.091223307,-0.0806378946,0.5170618892,0.2832086682,-0.3384489715,-0.0735353529,0.1981331557,-0.1567203254,-0.2741836011,-0.0703398734,-0.1646665037,0.3293884099,-0.5745841861,0.3498039246,-0.0944107249,0.1831559241,-0.0103389062,-0.1127805263,-0.1288834214,-0.0901363567,-0.1023518667,0.2645036876,0.4485737979,-0.2464579493,0.1231353655,-0.1331927925,-0.1977988333,-0.2744648159,0.0432025567,0.4913312495,0.1742404103,-0.0321537815,0.2832345068,0.3321835101,0.0047879368,0.1627507508,0.1906650066,0.082251437,0.2991720438,-0.2648767829,0.2358040959,-0.1145857275,0.3475693464,-0.2456605732,0.211557731,-0.0407701842,-0.2607629299,0.2219271809,0.3335291743,-0.096292004,0.365791291,-0.1115012839,0.0800432563,0.3007517457,-0.0706384107,-0.1508823186,0.0309130736,-0.5127479434,-0.7366874814,-0.3002051413,0.2035432458,0.3316034377,0.019516971,-0.2784527838,-0.2108523101,-0.1304586828,0.1237716749,0.2527556419,0.3760255575,-0.0118372235,0.1665927768,0.2711794078,0.2126277685,-0.1514804661,0.1475888342,0.0255004298,0.2315755039,-0.8735941052,-0.0115483757,0.1191699207,-0.0363761596,0.0999692082,-0.0447639152,-0.2247330546,-0.2866726518,0.2228051871,-0.1009206548,-0.0311283488,0.1471240222,0.0334274806,0.0902855694,0.1703851372,-0.2340905219,0.000517272,0.2663551271,-0.470720917,-0.3566271663,-0.1959459037,0.1416383684,-0.0774970502,-0.0170253646,0.5023498535,-0.0425135083,0.2484270781,-0.3122112453,-0.0222024266,-0.0065074591,-0.3388591707,-0.3757831156,-0.1752881259,0.1712938845,-0.4227336049,0.0585829876,-0.0057373825,-0.1998024583,0.2495300621,0.4242102802,-0.2043909878,0.3054891527,-0.0658724383,0.0704578534,0.8549776673,-0.1990090311,-0.6082813144,0.3079836667,-0.5397832394,-0.3694066107,0.1779018641,0.24567765,0.2107000053,-0.0261017065,0.3786222041,0.0901168063,0.2324507833,-0.0657918081,-0.2402687073,-0.1774432659,0.285887301,0.0218282063,-0.1026386321,0.2008702904,0.2385696173,0.2402634323,0.3049709499,-0.1066568941,0.330407232,0.218122527,0.2690805793,0.3724405766,-0.1290147156,-0.0760848969,0.1096108705,0.1873374283,-0.2275865525,0.0113488091,0.0424690694,-0.1055788919,-0.1242938191,0.1277856976,-0.4303869307,-0.3422091603,0.1733956784,0.1383132041,-0.0282927584,-0.0741612837,-0.00461961,0.0543021001,-0.066528976,-0.0914480835,-0.3181314468,0.1229491159,-0.4717287719,-0.1494160891,-0.1340329945,0.3342444599,0.1710793376,0.1060616747,-0.1678994298,0.1558741927,0.0076921401,0.0440265015,0.1992292851,-0.0338902585,-0.0364170112,-0.1528815776,0.2365458906,0.0526504964,0.0528609902,0.1382517368,0.1897073239,0.1646210551,0.0175435692,0.2983325422,0.0722362548,0.4031214118,0.0345559493,-0.0202592239,0.0355991833,0.1939075738,0.339620322,0.0514374934,-0.0347824581,0.0088161211,-0.0511058494,-0.0156220803,0.5277141333,0.1092472225,-0.0153419496,0.0442291871,-0.1107063815,-0.0175625384,0.0518276393,0.3424906135,0.4299674332,0.0601907298,0.0533262081,0.1188571304,-0.0037312207,-0.0990176797,0.0872748718,0.1116725355,0.3432447016,0.1183098555,0.154489547,0.2098272741,-0.1425979286,-0.3123348355,0.1767480224,0.4526069462,-0.2081012577,-0.0024063524,-0.2591174841,-0.3028211594,-0.3574784994,0.080826886,-0.2272000015,-0.0903564543,0.0413530469,0.2309697717,0.0392375439,0.2571948767,-0.1960360855,0.0636123046,-0.1070971116,-0.0212115739,0.1065067649,-0.2841461897,-0.3487016857,0.0732532516,0.4635375142,-0.0886287913,0.3015563488,0.0722924098,0.1173654124,-0.1981844008,-0.1850028336,-0.1156247705,-0.0971706361,0.0256881621,-0.0124279102,0.1032664031,0.023947414,-0.409684211,0.3573969603,-0.3606680036,-0.2367313504,0.0684053376,-0.1406067759,-0.322173506,-0.2491959482,-0.3021603227,-0.3738672733,-0.5032862425,0.4350882471,-0.176232174,0.2365990728,0.2102874368,-0.0092038447,0.4069841504,0.0855106711,0.1652594805,-0.0485855862,0.1821389943,-0.0066839829,-0.2967914641,-0.3691343963,0.1906411201,0.185684368,0.0145070814,0.104168646,-0.2383521497,-0.1529042572,-0.3123936057,-0.0752274767,-0.0741516799,-0.0384576321,0.4369110167,0.1654579043,-0.1200055033,-0.0298598167,-0.236050874,0.1802698821,0.0876838937,0.1502386034,0.1737029105,0.509760499,0.2883393764,0.2158128172,0.3241437376,0.1112618819,0.1727733314,-0.2193650454,0.1895129979,0.0585503243,-0.3553052843,0.3438940048,-0.095478937,-0.2766896784,0.2590985,-0.1013752893,-0.2876358926,-0.206346482,-0.0922211856,-0.3289086521,-0.1647435129,0.0548565425,0.1095771268,0.2776196599,0.02032982,-0.1122778952,-0.1518608332,-0.3898843825,0.0563465282,0.3294500709,-0.1117935628,0.1413694769,-0.3306563199,-0.3258556128,-0.4211726189,0.412471801,0.1170853525,0.5280153751,-0.3805490136,-0.2255387902,0.3277384341,0.028947074,0.431058228,-0.0674410164,0.0125309695,0.0216141045,0.2006269246,-0.259078294,-0.0499866605,-0.2014165521,0.5041893721,0.5344513655,0.3871465921,0.1936192513,-0.1236492321,0.0587879643,-0.0701735243,-0.2266270071,0.0727827549,-0.3008034825,-0.2470260561,-0.6267353296,0.0150931804,0.0658958554,0.3500673473,0.1190825999,0.1042059436,0.0870116577,0.1025859043,0.0400874801,0.0641186461,0.2440540493,0.1596351862,-0.023858266,0.0440107137,0.4587688148,-0.1278713495,0.3438452482,-0.1523287892,-0.0505895652,0.2836275101,-0.0745740682,0.4772113264,0.4196680784,-0.1903258115,0.1202337816,-0.0999226943,-0.109733738,-0.2003026605,0.0211654697,0.4794917107,-0.1645486057,-0.1696098298,-0.1381510496,0.0144466637,0.0097681079,-0.1500312537,0.0664378703,-0.2481898069,-0.145701468,0.3004769981,0.0781548098,1.1246336699,-0.1336105466,0.2934183478,0.0346855782,-0.1152120754,0.3822289407,0.0316152014,0.4024692476,-0.3729643822,-0.0356489643,0.044933483,-0.1520719826,-0.1320183128,0.1621411145,-0.0202923436,0.0472327285,0.3583298028,-0.081522353,0.10233628,0.1004131064,0.2062204182,-0.0612243414,-0.15373604,0.107619904,-0.0525326431,0.4005298615,-0.150371328,-0.1527941227,-0.1039494202,0.1493449658,-0.3186149299,0.234241277,-0.5883267522,0.2398847342,-0.2838946581,-0.2035966218,0.1377361119,0.5363996029,0.0978983939,0.1275990009,-0.1082234532,0.0722134113,0.237733528,-0.2076972276,0.1523761153,0.1767375469,0.1692968011,-0.2880898714,-0.2455843538,-0.0238138903,-0.0056499019,-0.2746348977,-0.1823810041,0.2214816809,0.1758712977,0.0375191011,-0.2288469821,-0.1714289784,-0.2919126749,-0.1695638597,0.1304813623,-0.1744417101,-0.1661675125,0.0666684657,-0.2520959079,-0.5136223435,0.2282321453,0.2680213749,-0.2405706942,0.04339296,0.5279790163,0.0291188695,-0.2210389227,-0.2394233346,0.0924208984,-0.3640986383,-0.5032579303,-0.0143474825,0.0835267678,0.1502227783,0.1606289744,0.100751929,0.2525140941,0.0363744758,-0.1265896857,-0.685240984,-0.4013647139,0.0656941235,-0.151758045,0.0423026718,-0.10146036,0.1632276624,0.1880227029,-0.2616633177,-0.304915905,0.1068875492,-0.3397385776,0.069059357,0.2392174155,-0.3161026835,0.0863329321,0.1420860291,-0.0048740553,0.1060876846,-0.0925553143,-0.2245420069,-0.0560308024,0.1410079151,0.1974923611,-0.0461962745,-0.0985007584,-0.4103367031,0.021262981,0.0283581745,-0.081216231,0.141697675,-0.0814264491,0.0727273524,-0.1813595593,0.0592008978,0.1271419227,0.2357475013,-0.2389324754,0.0665901974,0.2181263864,0.134544909,0.2189338505,0.005199634,-0.3545107841,-0.1517054737,0.17426835,0.1952820867,-0.2249659598,-0.2974476516,0.0157680493,-0.126891017,0.2684794962,0.3564190269,0.0442749411,0.1074714959,0.4834800363,0.1019813344,-0.2300039083,0.1293830127,0.2237648517,-0.4923118949,-0.2539443672,0.1021161601,0.1005869061,0.0964753702,0.0691590235,0.0250120237,-0.1789847016,-0.2635631859,-0.0562896281,0.3038937151,0.043119859,0.0624701865,0.1846094131,0.1837253273,0.2011337429,0.3358926773,-0.1404509395,0.1026413739,0.1053097993,-0.0080889091,0.1069152728,-0.3594929278,-0.2406992018,0.0367142521,0.0672627091,-0.0023503369,-0.2880562246,0.7975200415,-0.271224767,-0.1419475377,0.0676128715,0.3134138286,-0.2785675228,-0.193380475,-0.3274287581,-0.1071232259,-0.3263842463,0.0946334675,-0.0282573234,0.2516313493,-0.1173728183,0.058729656,-0.1699640155,-0.1321097463,-0.2452425808,0.3337428272,-0.1339598,-0.2808953226,0.1638755798,-0.2397953272,0.2477649599,0.0722718239,0.1565960646,0.401740253,0.3074324131,-0.347771883,-0.0361203104,-0.1223844662,0.0071289768,-0.2038355172,0.4214951098,0.4806911349,0.2369429022,0.2757431865,0.0185546931,-0.1177505776,0.1282254159,-0.0405920595,-0.1940850317,0.2781760991,0.3976898193,-0.0181025416,-0.4306953549,-0.1541188359,-0.0883549526,-0.2297391146,-0.0503083766,0.3019017875,-0.0576792508,0.1849604845,-0.2418810874,0.0751061589,-0.0019745003,0.4621947408,0.4583948255,0.1190907806,-0.1391515434,-0.3664035499,-0.6616654992,0.4916216135,-0.1032006219,-0.0950771645,-0.1530798972,-0.0902869552,-0.3845091164,-0.0181287583,0.2054770589,-0.2124761939,0.1697883606,0.2388371527,-0.1559887975,0.1548927277,0.2879478633,-0.0596503727,0.1704817414,-0.4716244936,0.0754617974,-0.1548440456,-0.0108604115,-0.1652823836,-0.0244545247,-0.3091744781,0.0255910307,0.4873985052,0.1285872459,0.4149319828,-0.0879802182,-0.0702042207,-0.6123086214,-0.1760347337,-0.2088954151,-0.1943843216,0.1757498235,0.2102415115,-0.1574782282,0.3392831385,-0.2442167848,0.1438356936,-0.0030851378,0.0850602463,0.0815796405,0.0405254588,0.1516462564,0.058792837,0.4189053476,0.122528024,0.0522084534,0.0806520283,-0.6217884421,-0.1854175627,0.5325632691,-0.5758374929,-0.4868765175,0.4769593775,0.2119816691,-0.050043188,0.0520830974,-0.3509189785,-0.0963341892,0.1097563729,0.0829489231,-0.3734717667,0.1202867627,-0.4446396232,-0.1011976525,0.1041453406,0.0656334907,-0.017513888,-0.1992147863,0.0896269381,-0.2572648227]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/590","title":"The process cannot access the file because it is being used by another process (windows)","comments":"I'm using version 0.4.0.\r\n\r\n","body":"Hi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```","comment_length":4,"text":"The process cannot access the file because it is being used by another process (windows)\nHi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```\nI'm using version 0.4.0.\r\n\r\n","embeddings":[-0.1683559269,-0.0665037259,-0.058606714,0.3233726621,0.3460412323,0.1262852997,0.3516989946,0.1841173917,0.2367685586,0.228607893,-0.1739535183,0.2335168868,-0.1790038049,-0.1726325154,-0.1070728078,-0.0009102829,-0.0121794567,0.0106416997,0.2435708642,0.2740671337,-0.4829724133,0.0663556755,-0.3408662081,0.4542468488,-0.2576961815,-0.3462219834,0.0952632427,0.3354336917,-0.041139666,-0.2197328955,-0.1430921555,-0.0509854555,0.1279723048,0.4750158787,-0.0001184053,0.1985581666,0.0216586143,0.0237138327,0.0576845929,-0.10134276,-0.0837891102,-0.1880415827,-0.1013302654,-0.0465593152,0.1593774557,-0.0084868437,0.2261129618,-0.4198738933,0.3096064031,0.4290079474,0.1890557706,0.2119928002,0.1282654256,-0.1794866771,0.5450937748,-0.0760362968,-0.0465740599,0.6440474987,0.542137444,-0.2162204981,-0.1142839566,0.2094042748,-0.2758595645,-0.1710480601,-0.1054810733,-0.2453682125,0.4545497894,-0.6501967907,0.5259749293,-0.0407179296,0.2336618006,0.0001739793,-0.1074666455,-0.0369744562,-0.2111619562,0.0139018204,0.2603872418,0.4270632267,-0.1555250436,0.2764199674,-0.1647695601,-0.1036969051,-0.3396034539,-0.0619628504,0.4541213512,0.0330004729,-0.0294504613,0.334918052,0.3636845946,0.0688177124,0.2349050641,0.10775619,0.0165614691,0.2802183032,-0.3352690339,0.2753875554,-0.21687828,0.4066414237,-0.2250238657,0.1406475008,-0.1133254468,-0.2810035348,0.2276980132,0.3724148571,-0.082705006,0.3144809008,-0.2183321714,0.1310916692,0.3253979981,-0.1139138788,-0.2957125008,-0.0043965569,-0.5373074412,-0.8001700044,-0.1677862406,0.3091683686,0.3589131236,0.0512561537,-0.4111965299,-0.1760664731,-0.1746020615,0.1022373736,0.2993425131,0.3742713332,-0.0035445441,0.1493280679,0.2024606019,0.2030694634,-0.0717559308,0.0990766138,0.0868042037,0.2365092486,-0.7948428392,-0.0169023741,-0.0563215278,-0.0860329494,0.0142100221,0.0564989559,-0.1430816799,-0.4108214676,0.306450367,-0.2622085214,-0.0073554111,0.2176884115,0.0135578886,0.0815238878,0.2845109999,-0.1568186581,0.0446159765,0.279817313,-0.6352659464,-0.4118247628,-0.1171019152,0.0655519441,-0.008947677,0.1711052805,0.5125060678,-0.0775057226,0.2796494663,-0.3804197609,0.0139469495,0.0073155775,-0.3754882812,-0.4266764224,-0.1833852381,0.2180030942,-0.4303310215,0.1589080393,0.1137764603,-0.2989491522,0.2729409337,0.383821547,-0.2031650245,0.3783925772,-0.1257641762,-0.1394321173,0.6555793881,-0.2254822254,-0.6815656424,0.3027138412,-0.5042111874,-0.2812467515,0.3899071515,0.2309723943,0.2967833877,-0.0403140672,0.427816838,-0.0505089797,0.1854620874,-0.1355486512,-0.1072022319,-0.2046245635,0.254586339,-0.0409692191,-0.0365293585,0.3108257055,0.3782010674,0.1141820103,0.272682935,-0.1219407246,0.2833160758,0.2152990699,0.3744134009,0.4507679045,-0.1247661635,-0.1505818218,0.1773387194,0.2243200839,-0.1530564129,-0.2940297127,0.0297405254,-0.1902048588,-0.0928086266,0.2330319434,-0.5529106855,-0.232395038,0.0733022541,0.1303833872,-0.035827823,-0.154259488,0.0943478271,0.0186200608,-0.086517863,-0.0962306783,-0.1170991063,0.0998261496,-0.4068972766,-0.1053081006,-0.2468519807,0.1757128537,0.2501446903,0.1586503983,-0.1043516994,0.209416002,0.2204131484,0.1608942598,0.2281097919,0.1137862504,-0.0207808185,-0.0444634333,0.2164999694,0.0135146119,0.083534427,0.0445596762,0.3502482474,0.064571932,-0.0134633668,0.3468743861,0.0251265373,0.3043022752,-0.0381456278,0.0403178819,0.1159390584,0.2593650818,0.1849548668,0.1052650437,-0.0335859843,0.052886039,0.0364254899,-0.0702797696,0.4182111919,0.0776030421,-0.025333181,0.0172193516,-0.1475397497,0.0676750168,0.095228985,0.2567639351,0.386685431,-0.04522679,-0.0111292955,0.0676973537,0.1851275712,-0.0623180568,0.1568654329,0.2089354396,0.3409472704,0.1761184335,0.0284427684,0.1947261542,0.0325387605,-0.3516673446,0.1559220999,0.3932161927,-0.2919055521,0.09871573,-0.1704436988,-0.0487230532,-0.3162109256,0.1017432883,-0.0886317268,0.0138823474,-0.0659559295,0.4510789514,0.0182547532,0.1845785379,-0.2166780233,-0.0281688012,-0.10000965,-0.0435397439,0.2168788761,-0.2350558639,-0.1619689763,0.0241767671,0.4536184967,-0.3816851377,0.348900646,-0.0118682161,0.1506352127,-0.3339325488,-0.1208776906,-0.1029922292,0.0178811122,-0.0115059521,0.0123582976,0.1807662994,0.1051459983,-0.4779096246,0.3853621185,-0.3077895641,-0.1952047199,0.0766071975,-0.2386821806,-0.3542949259,-0.1024158671,-0.1599394083,-0.3263009489,-0.4099702239,0.5021161437,-0.4058033824,0.2138349712,-0.0109839719,0.0272820126,0.3683962524,0.1026370078,0.2066823542,0.0061423085,0.0328125954,-0.1294892132,-0.152264908,-0.1565936357,0.2663016617,0.2404064983,-0.1145104989,0.2804791629,-0.1421534717,-0.2617709339,-0.315502435,0.0156001886,-0.0963245407,-0.0800268129,0.4200828373,0.2286265641,-0.0692826137,0.036644198,-0.1768316925,0.1716952473,-0.0450437479,0.0379256159,0.3204523027,0.5157552958,0.2461374104,0.3286038935,0.3779637814,0.2196093649,0.153156966,-0.2200354189,0.1603693366,0.1155514717,-0.3801929355,0.2663734555,-0.148611322,-0.5358139277,0.242661193,-0.163316682,-0.05188765,-0.1535043865,-0.3002959788,-0.3137047589,-0.1303146929,-0.0120202797,0.0720596164,0.3666327298,-0.1845205575,-0.1246650219,-0.0719175488,-0.3121961057,0.0652872846,0.4072082341,0.0080556944,0.1276382208,-0.1796524227,-0.3431226313,-0.2896775007,0.4998687506,0.1998041123,0.4337497652,-0.4395886958,-0.0841463283,0.2949256301,0.1051820517,0.5569274426,-0.0480566993,0.0709237531,-0.1268905401,0.0368487537,-0.0983099788,-0.0492763519,-0.1567501873,0.4721143246,0.4147341251,0.4088441432,0.2693660557,-0.0369830057,-0.0258483347,-0.0225203019,-0.2241665572,0.0590212755,-0.2850130796,-0.2577957511,-0.652913928,0.0974239185,0.1285099536,0.1824299991,0.2045988888,0.0873662978,0.1023598537,0.1243088469,0.1723734736,0.0094596175,0.3051178753,0.0648782998,0.0201379731,0.0872995183,0.395452559,0.1560705602,0.2930589914,-0.2067009509,-0.1698085368,0.236114189,-0.1357057542,0.5472165346,0.4346119165,-0.2435641289,0.2091375589,-0.0529830121,-0.1419092268,-0.2629524469,-0.1420802325,0.4830339551,-0.0691600144,-0.1351070553,-0.1938246489,-0.2032421082,0.0327319689,-0.2673163414,0.101483427,-0.2433861792,-0.2042539716,0.2343826592,0.1028464437,1.2042719126,-0.1956301779,0.2059898078,-0.109746635,-0.17496337,0.03261647,-0.0183188524,0.3367493451,-0.3396266103,-0.269625783,-0.0623504519,-0.2532959282,0.002693699,0.1907986999,0.0182638001,0.0528908633,0.1262963265,0.0988033339,0.1389982402,-0.1095457822,0.2430292666,0.0786470398,0.036378447,0.0170277283,-0.0439792089,0.3078232706,-0.2032174468,-0.0412934124,-0.0195513871,0.182983622,-0.3517112136,0.1284044832,-0.8502941728,0.282841146,-0.4515869021,-0.1424523741,0.0752003118,0.6410470605,0.1879467219,0.0880810544,-0.0671910048,0.1171678156,0.3748848736,-0.1736275852,0.126029104,0.1074739397,0.177980274,-0.2654752433,-0.3125857413,0.1091150939,-0.0219592731,-0.3042628765,-0.309363246,0.2580755353,0.2914443016,-0.0487825163,-0.2490363121,-0.3367034793,-0.2572714388,-0.1167422906,0.0533073992,-0.1758656204,-0.1289635748,0.0984499529,-0.4632661343,-0.6499096751,0.1977008283,0.2591704428,-0.1170660108,-0.0004449372,0.4651483595,-0.1092951149,-0.3071674109,-0.1599643528,0.1493697315,-0.3395326436,-0.445417881,0.064187333,0.0195414145,0.1605978608,0.044273857,0.0065489844,0.3367238045,0.0611893907,-0.1546613723,-0.6555976272,-0.3565403521,0.0098683611,-0.0278430227,0.0367065854,-0.2027579546,0.0384865999,0.1131630167,-0.3166759312,-0.2159463912,0.1716905832,-0.239005968,0.0439687595,0.2175382078,-0.24448511,0.0781362206,0.1097826734,-0.0697380304,0.0688158572,-0.0681940615,-0.1211003289,-0.0841800496,0.1743065417,0.0447640158,-0.0176660661,-0.0359380767,-0.451624006,0.1036339775,-0.0222537369,0.046317704,0.3773230314,-0.0025829277,-0.0799712315,-0.143952474,0.141102165,-0.0422016233,0.2110962123,-0.2846925557,0.1056279317,0.2381057888,0.0629950166,0.1032851338,0.0129841669,-0.2728833556,-0.1838634461,0.4452696145,0.043535836,-0.1954203248,-0.2052733302,-0.0591521747,-0.026170196,0.2771117687,0.2001664639,0.0438181087,0.1724247634,0.280736059,0.0301997121,-0.1525176018,0.1081285551,0.2587609887,-0.5703699589,-0.3613182008,0.1702863574,0.1269246042,0.1254055053,0.1761239171,-0.0334668756,-0.0383137763,-0.219499737,-0.0505345091,0.3082654178,0.1283427775,0.1847287118,0.1590383947,0.2140697688,0.2449125051,0.3593891263,-0.221719861,0.0731185526,0.1545165032,-0.0548595563,0.18455562,-0.3908846378,-0.118002072,-0.012178285,-0.1136472225,-0.0759458393,-0.332574755,0.7090557218,-0.292937398,-0.0986420587,0.0627890676,0.2578258812,-0.3293049037,-0.2495585531,-0.1790318936,-0.1409697384,-0.2487565279,0.0881023034,0.0744135082,0.3233660161,-0.055286631,-0.0054630814,-0.1097642332,0.0885164365,-0.316724807,0.4021060467,-0.0722892955,-0.3876360953,0.0505573973,-0.0482507348,0.1872232854,0.2544696629,0.1024527028,0.4234913886,0.2186218351,-0.3185593486,-0.0988921747,-0.1685783118,0.0541355088,-0.148203209,0.3754410148,0.3232997954,0.2739565372,0.1933223307,-0.0278413016,-0.064884074,0.0421760306,-0.0792151317,-0.1229761541,0.201464504,0.3277260065,-0.0659311265,-0.550971508,-0.008958607,-0.0945113227,-0.136499837,-0.1253489256,0.3165857196,-0.2261432409,0.1828007698,-0.0912380219,0.0270687882,0.0310945865,0.276268214,0.4487712383,-0.0375610963,-0.234589383,-0.2642634809,-0.5896722078,0.5715637207,-0.0231021084,-0.009993922,-0.116618149,-0.2483124584,-0.3610644042,-0.0094561968,0.3248535395,-0.4543974102,0.1970541179,0.2760403752,-0.0760146752,0.1297848225,0.3004381657,-0.0648898184,0.1752381772,-0.491355747,0.2305054665,-0.175346598,-0.0446764678,-0.2447554022,-0.0075050662,-0.2934405804,-0.1602525413,0.5000418425,0.1223413423,0.4307357073,-0.1199941635,-0.0033828781,-0.4428308308,-0.104728274,-0.0662060454,-0.2328478694,0.1750132591,0.1236138046,-0.2591575086,0.1216331273,-0.2312772572,0.2309749573,0.0189096648,0.074720107,0.0592006296,-0.0067313169,0.1975616664,0.0203099959,0.4189380407,0.0718174651,0.1294226944,0.0923717618,-0.523077786,-0.1269270182,0.4997888803,-0.5733703971,-0.530098021,0.3877124786,0.235075295,-0.1829036772,0.1305138767,-0.2387893349,-0.0302693751,0.2207904607,0.0986206532,-0.3708654642,0.0756688714,-0.3987150192,0.0192831215,0.1876833141,0.0324167907,-0.0699852258,-0.2188092172,0.0850296542,-0.2074954808]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/590","title":"The process cannot access the file because it is being used by another process (windows)","comments":"Ok, it's probably fixed on master. Otherwise if you can give me a fully self-contained exemple to reproduce the error, I can try to investigate.","body":"Hi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```","comment_length":25,"text":"The process cannot access the file because it is being used by another process (windows)\nHi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```\nOk, it's probably fixed on master. Otherwise if you can give me a fully self-contained exemple to reproduce the error, I can try to investigate.","embeddings":[-0.2114655524,0.0665597841,-0.0459682867,0.2789514065,0.3627562821,0.1312585175,0.3703974187,0.2482227832,0.2258239537,0.219642058,-0.1714614034,0.2673845291,-0.1824055612,-0.1288236976,-0.1086891592,0.0654447675,-0.0072369794,0.0088248989,0.0774251744,0.226959154,-0.4474639297,0.0641376451,-0.3092188239,0.4365319908,-0.3265941441,-0.2484493554,0.0702646598,0.4150611758,-0.0526370592,-0.285081327,-0.1506447941,-0.0810049251,0.1276627034,0.5536198616,-0.0001143734,0.2240380496,0.0947395191,0.0214060824,0.0279271845,-0.1374551207,-0.0875019953,-0.2366851121,-0.1152373105,-0.0333837569,0.1649273336,0.031001538,0.2015654296,-0.3285351992,0.2990384102,0.4722038209,0.2349824011,0.2076999396,0.1914312392,-0.2376170754,0.5633089542,-0.1176098436,-0.0370919183,0.6386207342,0.4471789896,-0.242445454,-0.1647862792,0.1773202717,-0.2169586867,-0.1768907905,-0.0892447755,-0.264503479,0.4385124743,-0.6413016319,0.4716545939,-0.0395026579,0.237539947,-0.0037552058,-0.1493100226,-0.0266971048,-0.1550510824,0.1002503559,0.2798582017,0.4880426824,-0.2075369507,0.2474027276,-0.1368513256,-0.0809995905,-0.2716522515,-0.0632070899,0.3949955106,0.0777700618,-0.0472148731,0.262033999,0.3292509019,0.0647502691,0.2178885192,0.0957395509,-0.0180257894,0.2244765759,-0.3569532931,0.3265299797,-0.1940017343,0.349080652,-0.2772836089,0.1277076155,-0.1150758639,-0.2551479936,0.2213638872,0.4073951244,-0.0367275998,0.298748374,-0.2618215382,0.1614613235,0.3607685864,-0.0533831976,-0.2936809361,0.0442496389,-0.534307003,-0.832441926,-0.1913453341,0.3128830791,0.344866544,0.0278751533,-0.3136811256,-0.141304031,-0.1441437602,0.139582485,0.2897669077,0.3550514877,-0.0385161415,0.1184187606,0.1915986389,0.1892247647,-0.0693101808,0.096791856,0.0672732964,0.1931974739,-0.8466009498,-0.0646453723,-0.0300680194,-0.151851818,0.0300290622,0.1117195636,-0.0966770649,-0.4035180807,0.2738670111,-0.1954499185,-0.0170250703,0.2679279745,0.0631162301,0.0826185793,0.3132300973,-0.1889096797,0.0930427983,0.3087938428,-0.5649157166,-0.4224681556,-0.1513073295,0.1013720408,-0.0662613809,0.1933198124,0.5052144527,-0.1071624979,0.310756892,-0.3346762657,0.0081839086,-0.045209568,-0.3426704705,-0.4344490767,-0.116665341,0.2212156504,-0.4432276487,0.1203273088,0.1230368912,-0.3053646088,0.1948789209,0.3802189827,-0.1538111418,0.3382058144,-0.1127017066,-0.1150262505,0.5909101367,-0.1668607295,-0.6615204215,0.2475111336,-0.5012634993,-0.2549396455,0.3151992261,0.1936203986,0.3255032897,-0.0650660768,0.4324889481,-0.0485434271,0.1648044884,-0.0730378628,-0.1286816448,-0.1494332403,0.1266527623,0.0218284857,-0.0161906332,0.2625211179,0.3569929898,0.0693524405,0.2791787386,-0.1492733955,0.3342377245,0.2405438125,0.3433425725,0.4428647459,-0.0851452649,-0.1356274933,0.1946950108,0.2756517529,-0.0956630334,-0.1632959098,0.0504540987,-0.1325701326,-0.0460985787,0.2153440267,-0.5356934667,-0.2269077897,0.1207040995,0.1248268113,-0.0949283689,-0.0859355479,0.0596954674,-0.004747049,-0.053257335,-0.0512145683,-0.1109489426,0.0484345406,-0.416023314,-0.1190474778,-0.2173052579,0.1694720834,0.2049477249,0.1270708144,-0.1877919436,0.1887207031,0.1872302443,0.131691888,0.1440542042,0.0508290343,-0.060366109,0.0016104854,0.197579816,0.084423244,0.0869315714,0.0391190425,0.2750764787,0.0635229126,-0.0350247808,0.3883447051,0.0055476683,0.3557156026,0.0042983671,0.0541647822,0.0657000467,0.2314148098,0.2577147782,0.0834213123,-0.0048109335,0.0656048879,0.0971830562,-0.0874451622,0.4135709405,0.0413592532,-0.0040190318,0.0519632809,-0.1366141438,0.0770386085,0.0497035757,0.3433394432,0.3794067204,-0.0573430583,0.0114713805,0.0805661678,0.1906318367,-0.0809611976,0.1559361219,0.1926278919,0.3218732178,0.2225374579,0.0483527742,0.1897525638,0.0217799656,-0.2853983939,0.1454841495,0.3537414372,-0.2817568183,0.0333628543,-0.1760063171,0.03251791,-0.2888410687,0.0986624807,-0.11784596,-0.0183148552,-0.0627996475,0.3851552904,-0.0559987165,0.2293783128,-0.1767348349,-0.0484613702,-0.0998528749,-0.0012479118,0.2512572408,-0.3172087371,-0.1889255196,0.0308867637,0.4870844781,-0.3395869136,0.2789709866,0.0210346933,0.1365533024,-0.3264134228,-0.1483816355,-0.0809008777,-0.0233134478,0.0468306616,0.0199408177,0.0718649849,0.1895557642,-0.3815601766,0.4461351931,-0.2553527057,-0.213452816,0.1011984199,-0.2133298516,-0.3726608157,-0.1734943986,-0.1339586526,-0.3441823423,-0.4939570129,0.4266889989,-0.3828506172,0.1781271994,-0.0357736275,0.055256635,0.2962765396,0.1091315448,0.1583042145,-0.0177807063,0.0172436275,-0.101785332,-0.249112308,-0.2508819997,0.2443485707,0.2196863741,-0.0482334644,0.2001391202,-0.1813926697,-0.1735420227,-0.3676159978,-0.0372530073,-0.1183127314,-0.0274094045,0.3658221662,0.2106409967,-0.1014410928,-0.0020055287,-0.1801065207,0.1635891944,-0.0637449473,-0.0035238403,0.2608179152,0.5538358688,0.2435319424,0.310675621,0.2992294729,0.2581651807,0.1126635522,-0.2037102431,0.1955663264,0.0714910403,-0.3873255849,0.3228870332,-0.2160300165,-0.5179752111,0.221157819,-0.1499301791,-0.0986209139,-0.1731103659,-0.2908778191,-0.2867597342,-0.1963751614,0.044630684,0.0739586875,0.3265121281,-0.1473403126,-0.1058715209,-0.0832556337,-0.2712127864,0.1070031002,0.3940211535,-0.0297030173,0.0916698128,-0.2047254294,-0.2930417359,-0.2580559552,0.4496949017,0.1703341156,0.4296236336,-0.3886586428,-0.1273224801,0.2763598859,0.0712186322,0.5536616445,-0.0467513651,0.1151600406,-0.0569405071,0.0121359015,-0.107147187,-0.1470273435,-0.1440141499,0.5119684935,0.3892018497,0.3635127544,0.2263512015,-0.0852317438,-0.0016644315,-0.0544380173,-0.2536138594,0.0420827754,-0.3382838666,-0.2160536647,-0.6214618683,0.0946314037,0.1319582164,0.2071800381,0.1676148325,0.0907708183,0.0802081302,0.1334478855,0.0519934744,-0.0032124184,0.2480492145,0.062913157,0.0978419334,0.0849561393,0.3432382047,0.103663817,0.3674907088,-0.2215961069,-0.2028971612,0.2907598615,-0.1647239327,0.466694355,0.390570581,-0.2132820636,0.1104503497,-0.0619390532,-0.1088378727,-0.278885901,-0.1341560334,0.4840845466,-0.0348863192,-0.1999316067,-0.1988426,-0.1572070867,-0.0048592216,-0.1904501319,0.1059810743,-0.2338134199,-0.1743928641,0.2608867288,0.1183208898,1.1062227488,-0.2257890254,0.279361397,-0.1018452346,-0.1728658378,0.0449693017,-0.0019078937,0.3326909542,-0.3763323426,-0.299559474,-0.0191268735,-0.2436331064,-0.1120934784,0.1470809579,0.0192079451,0.0236797575,0.1171261668,0.153285861,0.1730033606,-0.1390999407,0.2013514638,0.0400507301,0.0612145439,0.0750313252,0.0416849926,0.2896731794,-0.1694939435,-0.0740855932,-0.0342113338,0.2264031619,-0.3800475597,0.1525072604,-0.7754430175,0.2625685036,-0.4186415672,-0.1962836683,0.0945626646,0.5298982263,0.1237987131,0.1372974068,-0.0150658833,0.1529947817,0.3583410084,-0.1340484172,0.125883624,0.1337571591,0.1848993301,-0.2221436948,-0.3096384108,0.0611799993,-0.0552839451,-0.2310341597,-0.2675893903,0.2978301048,0.2179475129,-0.0421867855,-0.2948163152,-0.3396803737,-0.2047763467,-0.1398226619,0.0923409835,-0.2002799511,-0.148007527,0.0953225344,-0.4863420129,-0.6295734644,0.1926597506,0.30287081,-0.1706923097,0.0375143066,0.4928511679,-0.0302347615,-0.2388455123,-0.1716964692,0.1937442869,-0.3293118179,-0.4841332436,0.1162542552,-0.0370853506,0.1057279557,0.0720478296,0.024576623,0.3253953457,0.0260174219,-0.1965755224,-0.6410691142,-0.3526356518,-0.049519904,-0.0757587329,0.0325629413,-0.2115785331,0.0357925668,0.0924601033,-0.3269482553,-0.2626305521,0.1730850786,-0.2984365821,0.0239123758,0.2066274732,-0.2245717049,0.0567710474,0.0920963809,-0.0277760606,0.1124165654,-0.0668178126,-0.1739186496,-0.0778120682,0.1674265414,0.089033857,0.0575484335,-0.0051389965,-0.4036240578,0.0469921976,-0.0048165782,0.0552514344,0.3545493484,-0.0326883346,-0.0850429684,-0.2105123103,0.076686807,-0.0223545972,0.2580421567,-0.2664374113,0.0524817221,0.1899424195,0.0978460312,0.0852969661,0.0487020984,-0.2524018586,-0.1314856112,0.2411220968,0.1285726726,-0.1531641185,-0.2609266043,-0.0011969288,-0.0239949003,0.3382184803,0.2167876959,0.0252133664,0.1930728853,0.3364773989,0.0848474354,-0.1381895393,0.0676321983,0.1782607883,-0.5768575668,-0.2878244221,0.1503683627,0.0809060484,0.1030313522,0.1729649752,0.0152334217,-0.0828049034,-0.1475142986,-0.0355978459,0.3810205758,0.115998283,0.2030673772,0.185452655,0.0992241651,0.236952126,0.4189277589,-0.1544699967,0.0909946337,0.1707572937,-0.0959754735,0.207654044,-0.4515888989,-0.0706638023,-0.018240517,-0.0657051578,-0.089073956,-0.3714002669,0.7061873078,-0.3409683406,-0.1721334755,0.0367105864,0.2421206385,-0.3435501158,-0.1791521162,-0.1876557469,-0.1753399372,-0.2762936056,0.1574510634,0.0459495857,0.3376576602,-0.0722652003,-0.0009427762,-0.1236012504,0.0642235875,-0.3750308156,0.3611934483,-0.11063198,-0.3491284847,0.0585408695,-0.001129319,0.2037260383,0.2221876383,0.1219835132,0.5111351013,0.3227023184,-0.3435425162,-0.070783928,-0.1407616138,0.0528826267,-0.1593224406,0.3650752902,0.3437562883,0.2447632998,0.2015528232,0.0124921193,-0.1023597643,0.091474928,-0.0523324087,-0.1686676592,0.2005506009,0.3143181205,-0.0290102791,-0.5234512091,-0.0481195524,-0.1434713453,-0.1753615141,-0.0986288488,0.3696943521,-0.1777478009,0.1657351851,-0.1102023348,0.0658401623,0.0460827239,0.3879244328,0.4450755119,-0.012807576,-0.2195875049,-0.2608731985,-0.6123731732,0.5736035705,-0.0702377632,0.0302800983,-0.1296097785,-0.2331713736,-0.3260962069,-0.0178500116,0.3657417595,-0.4090032876,0.168955043,0.3255272806,-0.1080323532,0.0970966592,0.2898111939,-0.1113478988,0.1933068633,-0.5271305442,0.1646500826,-0.1570419222,0.0108298929,-0.1827662736,-0.019554954,-0.2236522436,-0.1941230744,0.5261103511,0.0835381076,0.4081564248,-0.0912900418,-0.0587922446,-0.5051499009,-0.1380488575,-0.1489104629,-0.2095574588,0.1882743388,0.1756819338,-0.2250097394,0.1258345097,-0.3072805107,0.2281695008,0.0165924728,0.1367890686,0.0588472523,0.0183328763,0.2320972532,0.0202527605,0.4795720875,0.127306208,0.103776671,0.1305953413,-0.5822556615,-0.1587870866,0.513585031,-0.6223827004,-0.5242946148,0.3736459315,0.2037199885,-0.1904453039,0.115077883,-0.2549934387,-0.0020231917,0.1618896276,0.1187794432,-0.4265726507,0.0556961112,-0.39171803,0.0160407294,0.1347406805,0.0209409352,-0.0386478454,-0.2768895924,0.0921132192,-0.1791355312]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/590","title":"The process cannot access the file because it is being used by another process (windows)","comments":"I get the same behavior, on Windows, when `map`ping a function to a loaded dataset. \r\nThe error doesn't occur if I re-run the cell a second time though! \r\nI'm on version 1.0.1.","body":"Hi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```","comment_length":32,"text":"The process cannot access the file because it is being used by another process (windows)\nHi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```\nI get the same behavior, on Windows, when `map`ping a function to a loaded dataset. \r\nThe error doesn't occur if I re-run the cell a second time though! \r\nI'm on version 1.0.1.","embeddings":[-0.14136222,0.0115144076,-0.0121005736,0.0894324929,0.3479508162,0.0587963611,0.4676170945,0.1985688657,0.2368670702,0.2460598201,-0.2052183598,0.2554389834,-0.0708039179,-0.2149474472,0.0211809669,0.1801772118,0.004898577,0.0978128985,0.1116587073,0.2412271351,-0.5601937175,0.0767828226,-0.3426276147,0.4758161306,-0.2404768169,-0.2781504095,0.0295200013,0.3297816217,-0.0508581176,-0.2468936443,-0.0585948527,-0.0388170071,0.0610766411,0.5021083355,-0.0001234325,0.2076142132,0.0500871316,0.0183689892,-0.0317803435,-0.1070131138,-0.0684424564,-0.1512527317,-0.0919808596,-0.06773635,0.1647817194,0.1099002808,0.1434138715,-0.4373567104,0.3229967058,0.4147151113,0.1281574517,0.2726634741,0.1071127132,-0.1961139441,0.6379075646,-0.0293761101,-0.1111278534,0.6946662068,0.5680832863,-0.409153372,-0.1346031725,0.2250454128,-0.2191309035,-0.0264014807,-0.1551886946,-0.1384028345,0.4357500374,-0.4799388349,0.4535281658,-0.0303509794,0.1233582273,-0.0895661861,-0.0924161375,-0.1571206003,-0.1803879887,-0.0489244312,0.3667474687,0.3262472153,-0.1456913948,0.1990912855,-0.2490430772,-0.0421518348,-0.2591157854,-0.0333371423,0.3093458712,-0.0745700523,-0.0314113349,0.3234914541,0.3730352223,0.1414939016,0.2741088867,0.0841259658,-0.01832016,0.2918751538,-0.2928037643,0.3861653209,-0.0639367774,0.3130858243,-0.1918380558,-0.0625624359,-0.0166269969,-0.3786750436,0.1871845424,0.3874210417,0.0990446582,0.1986857504,-0.2627533078,0.1062393636,0.3773981333,-0.2037296742,-0.3466935158,-0.0822355822,-0.3972129524,-0.6542537212,-0.0959610492,0.2318628579,0.3864630461,0.0201445948,-0.444794476,-0.130255267,-0.2066166252,0.1450045407,0.2346306294,0.3613013625,0.1014283523,0.0251276139,0.2232877314,0.1528564543,-0.1587819606,0.0670550242,0.1297106445,0.0575867407,-0.8975262046,-0.0086455056,0.0262226686,-0.0616105609,-0.0616382845,0.0595624484,-0.0952669755,-0.4462268353,0.2453478575,-0.2969215512,0.1318940818,0.3331855834,-0.035110198,0.0989937037,0.3047444522,-0.2101074755,0.0419480726,0.3383021653,-0.542812109,-0.3435719013,-0.0544381738,-0.0100679323,-0.1112358421,0.2478314638,0.4594936669,-0.0881026462,0.2433748096,-0.5887208581,-0.0351895951,-0.1100902855,-0.4130099714,-0.4058870971,-0.1700188965,0.3269718289,-0.34586972,0.1046348438,0.0815209448,-0.2160977721,0.1933882684,0.3409868479,-0.2419815511,0.4212700129,-0.2040447146,-0.0338634886,0.5636239648,-0.1880694479,-0.752253592,0.283264488,-0.4704799354,-0.1629840285,0.1556879282,0.2498376071,0.3320500553,-0.0803198889,0.4212530851,-0.0354684442,0.1602882892,-0.0513002351,-0.0560669564,-0.0933916569,0.209737137,-0.1015670672,-0.0983515009,0.3848287463,0.5134427547,0.0934621617,0.2899844646,-0.1401347667,0.2673074305,0.0757894069,0.2811129987,0.4180851281,-0.114678748,-0.1326086074,0.0769589394,0.2450481951,-0.0843716487,-0.2340932935,-0.0087467162,-0.2115938514,-0.0939519778,0.2580208778,-0.4916742444,-0.1750731766,0.0038237234,0.1637924314,-0.0698042959,-0.0209143776,0.0334407724,0.0008902663,-0.1489134133,-0.0832627788,-0.0076971375,0.0859409049,-0.2953228354,-0.1756304055,-0.2129376531,0.1347346753,0.1840849817,0.1390432566,-0.1797684282,0.2243450433,0.3209833503,0.2022684813,0.0562611297,0.0483096763,-0.0491747931,0.1920040101,0.1274475157,-0.001757763,0.1362547576,-0.0406481251,0.1508340389,0.1049964502,0.0537456982,0.375671804,-0.0394800678,0.26376912,-0.0308439601,0.0898995325,-0.0284792911,0.2370778024,0.1620222181,0.060071934,-0.031735599,0.0925721526,0.021815503,-0.1094775796,0.4865920842,-0.0019061774,0.1156216115,-0.0405843444,-0.1285411566,0.0856793076,0.1505112201,0.3509045541,0.4571465254,-0.1010679305,-0.0190003589,0.0774334818,0.0760467872,-0.0636864901,0.0982598513,0.1598389894,0.3917525113,0.1816726476,0.0771106333,0.1335733384,-0.1835931242,-0.2560079396,0.2024307549,0.3051225543,-0.3449656963,0.0733584017,-0.1770001203,0.0318625942,-0.3168502152,0.0179167707,-0.0927894935,-0.0392646044,-0.1229689866,0.5164975524,-0.0034537071,0.2352937311,-0.2392572761,0.0669594407,-0.1224464104,-0.0871035233,0.220988825,-0.2567580342,-0.2006324232,-0.0195788965,0.4293872118,-0.3859557807,0.3175410032,-0.0251149498,0.2133794725,-0.2417886108,-0.1630088389,-0.0314556696,0.0066923751,0.075885877,-0.0006341499,0.162471056,0.0390705802,-0.2285880446,0.380076617,-0.3706241846,-0.2118948102,0.1213602722,-0.1508995146,-0.3524764776,-0.1128026918,-0.2050598562,-0.2362649441,-0.3290404975,0.5150355101,-0.5015131235,0.1188741103,0.0526286513,0.1022956967,0.3818595707,0.0674314052,0.1462170333,-0.0838080272,0.0460403748,-0.0927362815,-0.2266786546,-0.1957887262,0.1894262135,0.3228176236,-0.151099056,0.4219771922,-0.2286889553,-0.2187833786,-0.2935786545,-0.0576859079,-0.0539284572,-0.0149472523,0.3952738643,0.2193965763,-0.0485511832,-0.0531065166,-0.2103042901,0.0815929845,-0.0166885015,-0.0519202724,0.3699101806,0.5677775145,0.2121956348,0.2237595469,0.3477529883,0.2361708581,0.196256429,-0.2543728948,0.3022272289,-0.0147841973,-0.3478395045,0.3301239908,-0.2493269145,-0.5216106176,0.2343190014,-0.1332305372,-0.140432179,-0.0753043219,-0.2685903609,-0.3589956462,-0.2149880528,0.1313045174,0.0430365726,0.3650546968,-0.0159004163,-0.0638611466,-0.0945850238,-0.2662542462,0.0111900549,0.2958978117,0.2029915452,0.1099668741,-0.1891301721,-0.3375439942,-0.3593934178,0.4088298082,0.1948037148,0.4997758567,-0.3862688839,-0.0939888731,0.3094781935,0.0615850203,0.5285878181,-0.0825698152,0.1349604279,-0.0322112553,0.0120380111,-0.1825372726,-0.0618384108,-0.1464896947,0.5905237198,0.3163489401,0.4090490043,0.3880321681,0.0856932849,-0.0564417504,0.0261868909,-0.1849884838,-0.0189290997,-0.3187396228,-0.3213723302,-0.6887792349,0.1587914526,0.1606612951,0.0763273016,0.1857764423,0.0524575599,-0.022573255,0.1010531858,0.10766761,0.1295807809,0.22067298,0.038789846,0.0728812143,0.1941767782,0.2760866582,0.1475396007,0.4105932713,-0.2495783865,-0.1668436825,0.2950470746,-0.1544923037,0.4194769263,0.5211333036,-0.1932347864,0.1478028893,-0.1029863432,-0.2066300958,-0.3486858308,-0.2006395608,0.5689414144,0.0225589927,-0.1061350405,-0.2542788684,-0.1583705693,0.019726593,-0.3631269634,0.1986816525,-0.2120752335,-0.1715662777,0.3051836491,0.027049087,1.1691449881,-0.2686498463,0.1820832491,-0.0748525113,-0.1111440435,0.0132018207,-0.004612254,0.3997065425,-0.2680267096,-0.3571724892,-0.0744407251,-0.3069595695,0.0026623616,0.2075480521,0.1080101877,0.0051191123,0.1454626471,0.2447905093,0.1421995759,-0.194520697,0.2674925327,0.028727252,-0.0151185161,0.002805376,0.080376856,0.3495514095,-0.1559372544,0.1015472859,-0.0940516815,0.1380604208,-0.27982831,0.1844419241,-0.7715752125,0.2749507725,-0.5094804764,-0.0246033687,0.0325674638,0.6402313709,0.0685770661,0.0868128538,0.0277775992,0.1095401794,0.4914098978,-0.0713762119,0.0799370185,0.0126030724,0.1288146973,-0.1386258006,-0.3554722369,0.0933273956,-0.0085450513,-0.3653726876,-0.3121630251,0.3467276692,0.2849783301,-0.1010493711,-0.1500522792,-0.340785861,-0.1899007857,-0.1435968131,0.0098658372,-0.2104025036,-0.0666450039,0.3015584946,-0.5401701927,-0.6096476912,0.1981391758,0.3259804845,-0.1335199922,0.0663746595,0.3299743235,-0.1386146098,-0.2458222061,-0.1201138943,0.2088449597,-0.2566815615,-0.4841573238,0.1436632872,-0.0594276711,0.1939190328,-0.0004122374,0.1493328512,0.2973934412,0.0471616574,-0.2445894182,-0.6123518348,-0.2343069017,0.0820054486,0.0850910619,0.1012048647,0.0122905457,0.1258261651,0.1343888342,-0.2901144028,-0.1751637161,0.2391397506,-0.1433006525,0.08209759,0.2798568606,-0.1230386645,0.1190674677,-0.0504880548,-0.1129848436,0.234979108,-0.0249794088,-0.0788338482,-0.0615954958,0.2095419616,-0.0111298375,0.085523732,-0.1009887457,-0.4917930067,0.139608413,-0.1762103885,0.0901895165,0.3931539357,-0.0345814638,-0.2588270009,-0.0482468754,0.0897355229,0.1152819917,0.299190551,-0.3786828518,-0.0274622459,0.1826839745,0.1161412597,0.1735876203,0.14321208,-0.1663254052,-0.0768376589,0.2709827423,-0.0188512001,-0.1418463141,-0.2272294909,0.0257441122,-0.1293246448,0.3927751482,0.252884537,-0.0798219219,0.0150173567,0.2868748605,-0.0068378178,-0.1479762942,-0.063216202,0.2872813344,-0.4038873911,-0.2592488527,0.2804384232,0.1411351413,0.0427056178,0.0749328136,-0.0996244177,0.0239488948,-0.1973247081,-0.1192630678,0.3578242064,0.0735373721,0.2194062918,0.2079422772,0.1932835281,0.2636315227,0.5280952454,-0.1284553111,-0.0686653554,0.1418910921,0.0015388825,0.1378089041,-0.4184558392,0.0150376577,0.0566635206,-0.1932702065,0.028106425,-0.2981097996,0.6454676986,-0.357894063,-0.1090263873,-0.0337382406,0.2910708487,-0.3422365487,-0.2771368027,-0.23138991,-0.1520056278,-0.2557014823,0.06016846,0.0127155706,0.4008627236,0.1555402875,-0.0556645766,-0.029253874,0.0869758874,-0.4900072217,0.2940213084,-0.0374816768,-0.4188626707,0.0647214949,-0.1684449315,0.3638537824,0.263474822,0.1484508514,0.5313634276,0.3240067959,-0.1957427114,-0.0670815632,-0.1455582827,0.0605515577,-0.2047076821,0.3779831231,0.3348797858,0.2564017475,0.2689193487,-0.0570321344,-0.0056080353,0.070185028,-0.0868096873,-0.1274071634,0.1538698673,0.233924225,-0.0588724352,-0.449020803,-0.0687281713,-0.0442024656,-0.1037912071,-0.0279977489,0.2987698317,-0.3428556919,0.2525935471,-0.041470632,-0.0017833511,0.0940829441,0.218576774,0.3783856928,-0.0119014774,-0.2470856756,-0.3395846784,-0.7406719923,0.4986650944,-0.2982273102,0.0753232986,-0.1138961539,-0.2813333273,-0.250682056,0.0077749663,0.3160145283,-0.4026606083,0.0189629085,0.3225497305,-0.0602930114,0.0331290551,0.2629623711,-0.0624234527,0.1254675537,-0.4483182132,0.2661900222,-0.1243150383,-0.056860555,-0.2836568952,0.0359411873,-0.3244712055,-0.1425289214,0.6262717843,0.2383572012,0.4493779838,-0.0864796042,-0.0794856921,-0.50342381,-0.1842192709,-0.08061032,-0.1888899356,0.1566803306,0.0228779539,-0.2726192772,0.0991907641,-0.2502710223,0.0577518679,-0.0042232876,0.3306563199,0.0701657683,-0.0507742614,0.1258136034,0.1061976552,0.2523438334,0.1627786756,0.1021654457,0.2257297188,-0.4350597262,-0.1813194603,0.4802578986,-0.6469630599,-0.5080171824,0.34633407,0.1946897954,-0.2940411568,0.131604135,-0.3184368312,-0.0499446504,0.2210285515,0.1013968214,-0.4337932765,0.1087736636,-0.4045119286,-0.0239032339,0.0889079124,0.1531774551,-0.0511756651,-0.3144827187,0.0913261473,-0.2533950806]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/590","title":"The process cannot access the file because it is being used by another process (windows)","comments":"This is going to be fixed by #644 ","body":"Hi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```","comment_length":8,"text":"The process cannot access the file because it is being used by another process (windows)\nHi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```\nThis is going to be fixed by #644 ","embeddings":[-0.1896061301,-0.0111349421,-0.0756933913,0.2970290482,0.3587117493,0.088923797,0.3750119805,0.2285797149,0.2834442258,0.2133695036,-0.1578124613,0.2428354174,-0.1462750733,-0.1303370297,-0.0942254364,-0.0009319866,0.0219870172,-0.0113319866,0.2287476361,0.2580116391,-0.4923674464,0.0478530228,-0.3434571028,0.4212952852,-0.2977575958,-0.323761642,0.1236702353,0.3587287366,-0.0439824052,-0.2376574874,-0.1439123899,-0.0272726081,0.1506108791,0.4492019415,-0.0001159419,0.1736665964,0.064376004,0.0334825106,0.028815724,-0.090987809,-0.0780475959,-0.2314954251,-0.1092303097,-0.0287026409,0.1302566677,-0.0331770182,0.2041799277,-0.4126203358,0.2992787957,0.4082279205,0.2239761502,0.1960004717,0.2185512483,-0.1635704488,0.5425683856,-0.0646535829,-0.0171096008,0.6168642044,0.5340331197,-0.2707248628,-0.1320349872,0.2320161164,-0.2581591606,-0.1883072257,-0.1337163299,-0.2620303631,0.439724803,-0.6819570065,0.4670731127,-0.0644351989,0.1965785325,0.0006998722,-0.1244240925,-0.0585001111,-0.1212099865,0.0553540848,0.2755756378,0.4438889623,-0.1700520515,0.2390906513,-0.1645203978,-0.1246395782,-0.3203935921,-0.0721634328,0.4729402363,0.0403171331,-0.0400329269,0.2791860402,0.3224700987,0.0812803209,0.2239244878,0.0799820647,-0.0072506974,0.2848716974,-0.3404327035,0.2747806311,-0.2307688892,0.4361918867,-0.2410114706,0.1475558728,-0.1577611268,-0.2450217456,0.2613366544,0.3526161909,-0.087301001,0.2992543876,-0.2182901204,0.1356960088,0.3698437214,-0.0791944563,-0.3039358258,0.0077803996,-0.5414533615,-0.8275242448,-0.174916923,0.3600537479,0.3753437102,0.0566429496,-0.3996003866,-0.135838449,-0.1608099341,0.0838092268,0.3063353598,0.330157429,-0.0011319065,0.1408398896,0.2080480754,0.2109631747,-0.0986399949,0.0873964727,0.0808432177,0.2141329646,-0.7662721276,-0.0566555485,-0.027641736,-0.0573615842,0.0372833386,0.1071389914,-0.1373636574,-0.4314849377,0.2857490778,-0.2481935769,-0.0149705047,0.2371587753,0.0255416054,0.0817830712,0.2865552604,-0.2029199749,0.0603021421,0.2916236222,-0.6150487065,-0.3869928718,-0.1512694806,0.1052987203,0.0149285225,0.1913467795,0.5124511123,-0.070266746,0.2584140599,-0.3458579481,0.0130124632,-0.0032866132,-0.3444340229,-0.4121506214,-0.1347711533,0.2288802862,-0.4172494113,0.1363769323,0.1412187368,-0.308960706,0.2556519508,0.3206567764,-0.1912593544,0.4152080417,-0.12969248,-0.150142923,0.616882205,-0.1896093488,-0.7148285508,0.3209627271,-0.4731826484,-0.2585818172,0.3463978469,0.249239549,0.3209867477,-0.0305974465,0.4143554568,-0.01453989,0.1810656488,-0.0908435136,-0.1414228678,-0.1961360127,0.2201836258,-0.0356382467,-0.0282833464,0.3136174679,0.3815901577,0.1029664502,0.3185447752,-0.1075206473,0.2873400748,0.2236991823,0.4133407474,0.4305747151,-0.1363721639,-0.1505951881,0.137310341,0.2030469477,-0.1471232474,-0.2582456172,0.0107231149,-0.1929414421,-0.0791551992,0.2075111717,-0.5668330789,-0.2339009047,0.1157646254,0.0857649446,-0.0660511553,-0.1405810267,0.0952803493,0.0259515233,-0.0711172074,-0.0563448332,-0.08067891,0.158817932,-0.4011130631,-0.0834553316,-0.2308870852,0.1840308011,0.2591130733,0.1328862458,-0.1389790177,0.2086849511,0.2051519006,0.1770069748,0.1973045021,0.0809802711,-0.0062009734,-0.00139797,0.1984378546,0.0143010365,0.077195555,0.0239399839,0.3437654376,0.0484436229,-0.0895605385,0.3721993864,0.0026957719,0.3128175437,-0.0401710644,0.0415618196,0.0933356434,0.2612961829,0.1691417843,0.1460463703,0.0013181582,0.0580662675,0.0566093251,-0.0829296708,0.4079308808,0.0973261297,-0.0512248054,0.0069874683,-0.1214834824,0.0851215124,0.0877695233,0.2791482806,0.3573869467,-0.0515190437,-0.0487530716,0.0468466952,0.1772335768,-0.0813328847,0.1824625134,0.2199662477,0.333073765,0.197972402,-0.0262230113,0.1845230609,0.0187025201,-0.3304832578,0.1524261981,0.362606734,-0.2700324059,0.0480062738,-0.1534359604,-0.0647667274,-0.3453166187,0.0846009851,-0.0805019736,0.0207725093,-0.052902814,0.4374658465,-0.0180324707,0.2181233168,-0.2056616396,-0.0750188529,-0.1085773855,-0.0424443185,0.2220038325,-0.2138205916,-0.1462878138,0.040974874,0.466835469,-0.4126318991,0.3311639428,-0.0089588286,0.1229583472,-0.3438409865,-0.131356895,-0.1208957359,0.0233945418,0.0503964797,-0.0138244471,0.157964021,0.1163952351,-0.4896101058,0.402169764,-0.2863936424,-0.1860496849,0.0402320959,-0.2192755938,-0.3489832878,-0.1271799505,-0.1704589576,-0.3118916452,-0.4371007681,0.4584638476,-0.4091207385,0.2009373456,-0.0198623948,0.0256998539,0.2885017395,0.1298436373,0.1835536212,0.018969411,0.0362323485,-0.1417939961,-0.1800798327,-0.1633633822,0.2370503992,0.2226043493,-0.1065144166,0.249179706,-0.1292481124,-0.2445161492,-0.3292141557,0.042586498,-0.0924032256,-0.0455230474,0.3911231756,0.2523777187,-0.0939145237,0.0553876944,-0.1743300408,0.1766394526,-0.0234863423,0.0002402305,0.3245328665,0.4985196292,0.2388265133,0.2680182755,0.3834841847,0.2575274408,0.1435801685,-0.1849066168,0.1471119076,0.1244134009,-0.3952119648,0.2767937183,-0.1646655798,-0.5181784034,0.2146025747,-0.1655102819,-0.0592270456,-0.1364012063,-0.2954134047,-0.3502160907,-0.1632698476,-0.0229336917,0.0676756427,0.3478930295,-0.1615187973,-0.1144526526,-0.0358070061,-0.276796937,0.0947901309,0.388982296,0.0498090275,0.0985774025,-0.226214245,-0.3501341045,-0.3017319739,0.4599396288,0.1883821487,0.4383776188,-0.3809550703,-0.0893540755,0.284237504,0.1223752052,0.5683745146,-0.026748348,0.057535816,-0.0798436701,0.0605767369,-0.0562408827,-0.1057902128,-0.1263247579,0.4915861189,0.3959472179,0.4006907344,0.2572445273,-0.0490139835,0.0089127552,-0.0362959094,-0.2471695244,0.0495031029,-0.276899904,-0.2413869053,-0.6287124157,0.1296342164,0.1370023191,0.2113829702,0.1636599302,0.0683230981,0.0761281699,0.1197075844,0.1185496822,-0.026315527,0.3146895766,0.0835891291,0.0587244667,0.0975228325,0.4170951545,0.1270496696,0.2529617548,-0.2789255977,-0.1786984652,0.2059531957,-0.1952288747,0.5352224708,0.4229588211,-0.213161394,0.1796359271,-0.0204989295,-0.11762622,-0.3314145803,-0.166853562,0.4844257534,-0.0585105754,-0.1294094026,-0.2044247985,-0.2076360881,-0.0163758174,-0.2521146834,0.0805362687,-0.2273979187,-0.2350527346,0.2302858233,0.1059424505,1.1456122398,-0.1900105774,0.1850304753,-0.145409137,-0.1881723851,0.0017539792,-0.0123074446,0.3503583372,-0.3939956427,-0.2862128615,-0.036814291,-0.2177464813,-0.0108658839,0.1750418842,0.0357100889,0.0514370054,0.1216165349,0.0900072306,0.1099772602,-0.1032304838,0.2439352423,0.0765866488,0.0418984294,0.0394213684,-0.0274434797,0.34195292,-0.2197959274,-0.0362634547,0.0230115019,0.1515326053,-0.341991514,0.1189646721,-0.8483833671,0.2523172498,-0.480419755,-0.1707520634,0.0791154057,0.6273828745,0.189651534,0.1059432328,-0.0358217843,0.1230019554,0.3821809888,-0.1302404404,0.1349033266,0.1317630857,0.1397355646,-0.236889407,-0.3161583245,0.0502963141,-0.0153915118,-0.2801125944,-0.3621993363,0.2437868267,0.3324080408,-0.0607228614,-0.2289867103,-0.3431330025,-0.185107857,-0.1717157066,0.0858601034,-0.1853400767,-0.1129531413,0.1379953176,-0.4442261755,-0.6407071352,0.1717633456,0.2699073553,-0.1510042697,-0.0068800291,0.4472545981,-0.0919443369,-0.331350565,-0.1815640032,0.1480621099,-0.3299973905,-0.464541465,0.059743531,0.0009870142,0.1521107554,-0.0070225792,-0.0052872635,0.3180846274,0.0379500128,-0.120712623,-0.6573610902,-0.3386734724,0.0191726536,-0.0206397865,0.040766798,-0.1958699375,-0.0196514148,0.1052604318,-0.3278910816,-0.2433198839,0.1522271037,-0.203645736,0.0448070578,0.1969258487,-0.2285685539,0.0729076862,0.1142854318,-0.0270666722,0.0274338853,-0.0261156298,-0.1521071196,-0.0744706169,0.1521098763,0.0068894699,-0.0080266884,-0.0383798219,-0.4461027682,0.1059710309,-0.0272988416,0.0387744866,0.3677967191,0.0281005297,-0.0466673337,-0.1792620718,0.1431598067,-0.025476763,0.2266397178,-0.2805318534,0.1027583405,0.2131870538,0.0715154186,0.1267507374,0.0260523763,-0.2476917058,-0.205533728,0.3830507398,0.0796730667,-0.191581741,-0.2063257545,-0.0470338166,0.0264300238,0.3033035696,0.1985917389,0.0537995137,0.1742466539,0.303324312,0.0780278966,-0.1515975744,0.0899005011,0.2466272414,-0.5950181484,-0.3395097256,0.1666785032,0.1241646335,0.1425308287,0.1615913957,-0.0152577758,-0.0547617152,-0.1816307157,-0.032209117,0.317443639,0.1438305825,0.1623511165,0.1246612445,0.1553144753,0.1836323142,0.3777534068,-0.1998728663,0.0496703424,0.178208515,-0.0570675619,0.2044968754,-0.3648743629,-0.0889386237,-0.0023365763,-0.1426094621,-0.0876635239,-0.316404134,0.7064428926,-0.2875232697,-0.1225691512,0.0914211199,0.2239280641,-0.3808161914,-0.2152669877,-0.1820080578,-0.1646178216,-0.2508289814,0.0798106194,0.0746361613,0.2998779118,-0.067602098,-0.0105463928,-0.1181614324,0.0879397839,-0.2843604386,0.4009625018,-0.0433599427,-0.3832948506,0.0510897227,-0.0367777534,0.1896565259,0.2288068086,0.14984487,0.3896231055,0.2460693717,-0.3263424337,-0.112716049,-0.1452726275,0.0587323532,-0.1744680852,0.3734346926,0.2820680439,0.263736397,0.1822232157,0.001769617,-0.0829799101,0.0228774939,-0.0832157433,-0.1693642586,0.2377011925,0.3378268778,-0.0502939746,-0.5564652085,-0.0021069206,-0.1404892653,-0.0693333,-0.1546571255,0.3172262907,-0.2066260725,0.1849721074,-0.1165553257,0.0548716113,0.0295996722,0.2365799844,0.4449949861,-0.0482608117,-0.2467894852,-0.2740619183,-0.5804633498,0.5542895794,0.0031319174,0.0288267806,-0.1241859719,-0.2573463917,-0.3960330188,-0.015950283,0.3581181467,-0.4992668033,0.2199922502,0.288664937,-0.0687615648,0.1695595384,0.3354432881,-0.0688330531,0.1673910916,-0.5421072245,0.2476441413,-0.1331788599,-0.019655196,-0.2666824758,0.0422583446,-0.2738879621,-0.1893415004,0.5030882955,0.1479177326,0.3974019587,-0.1574472189,-0.0179086868,-0.4367474914,-0.0827972591,-0.0853548124,-0.1981009543,0.1627022028,0.1680724174,-0.235877797,0.111070253,-0.2700627744,0.2445556223,0.0262844414,0.1150946245,0.0124340784,-0.0368950181,0.2171852142,-0.0020051079,0.4336349964,0.0786347017,0.136746645,0.1077812165,-0.5093563199,-0.142151922,0.4349456728,-0.5569652319,-0.5172345042,0.3507876992,0.2624551058,-0.167357564,0.150177598,-0.1966997385,-0.012168075,0.2336731255,0.0868020132,-0.3278173208,0.1140776649,-0.3924230337,0.0234629195,0.166490972,0.0418686047,-0.0626884624,-0.2209634632,0.1038445458,-0.2059338987]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/590","title":"The process cannot access the file because it is being used by another process (windows)","comments":"@saareliad I got the same issue that troubled me quite a while. Unfortunately, there are no good answers to this issue online, I tried it on Linux and that's absolutely fine. After hacking the source code, I solved this problem as follows.\r\n\r\nIn the source code file: arrow_dataset.py -> _map_single(...)\r\n\r\nchange\r\n```python\r\nif update_data and tmp_file is not None:\r\n shutil.move(tmp_file.name, cache_file_name)\r\n```\r\nto\r\n```python\r\ntmp_file.close()\r\nif update_data and tmp_file is not None:\r\n shutil.move(tmp_file.name, cache_file_name)\r\n```\r\n\r\nThen it works without needing multiple times runs to avoid the permission error.\r\nI know this solution is unusual since it changes the source code. Hopefully, the lib's contributors can have better solutions in the future.\r\n","body":"Hi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```","comment_length":111,"text":"The process cannot access the file because it is being used by another process (windows)\nHi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```\n@saareliad I got the same issue that troubled me quite a while. Unfortunately, there are no good answers to this issue online, I tried it on Linux and that's absolutely fine. After hacking the source code, I solved this problem as follows.\r\n\r\nIn the source code file: arrow_dataset.py -> _map_single(...)\r\n\r\nchange\r\n```python\r\nif update_data and tmp_file is not None:\r\n shutil.move(tmp_file.name, cache_file_name)\r\n```\r\nto\r\n```python\r\ntmp_file.close()\r\nif update_data and tmp_file is not None:\r\n shutil.move(tmp_file.name, cache_file_name)\r\n```\r\n\r\nThen it works without needing multiple times runs to avoid the permission error.\r\nI know this solution is unusual since it changes the source code. Hopefully, the lib's contributors can have better solutions in the future.\r\n","embeddings":[-0.1046754122,0.1074618921,-0.0233315229,0.2269619256,0.3214992881,0.0007905434,0.4211469889,0.1908867508,0.1951685995,0.1922733337,-0.1767421812,0.3607426584,-0.174327001,-0.2670363486,-0.1233475432,0.0355832614,0.0496971495,0.013052769,0.2309346646,0.2061697394,-0.3807088435,-0.0331440866,-0.3299262524,0.3217106164,-0.2837905884,-0.3542702496,0.0100989966,0.4027675092,0.0295142867,-0.288077116,-0.1288164705,0.0231063496,0.1196947098,0.5545585752,-0.0001154608,0.1748197377,-0.046863202,0.0502059348,0.0135354763,-0.0613390207,0.061984878,-0.2356380969,-0.0451326482,-0.1107083708,0.1963998228,0.0137783512,0.2351698577,-0.4139168859,0.4707049131,0.4032635391,0.1879202724,0.2586315572,0.1164235771,-0.0457814038,0.4986590743,-0.0801039413,-0.0995624736,0.5888605118,0.4468511343,-0.2438413948,-0.1711247116,0.3385624886,-0.2344774902,-0.1508841813,-0.1414978206,-0.2691135406,0.2938320339,-0.6156047583,0.4682242274,-0.0681593418,0.0907823145,-0.0633870959,-0.2490687221,-0.1350107938,-0.0936150178,-0.1098759249,0.2257647365,0.4395219982,-0.1495494992,0.2055870593,-0.0682476908,-0.1247708127,-0.3324677944,-0.1159746796,0.4712693393,0.1288682073,-0.0207260437,0.2954647243,0.3008286655,0.1232295334,0.2916143537,0.17420654,0.0111910403,0.2981752753,-0.3884970248,0.3316394687,-0.1883032471,0.2630648017,-0.3031633198,0.2108683884,-0.013326996,-0.1471738964,0.245511815,0.3030039966,-0.0604699589,0.2595026195,-0.2948144376,0.2529595792,0.4194628894,-0.1222944558,-0.2796795964,0.0336038806,-0.5552170277,-0.7762519717,-0.0632790774,0.3375050724,0.4028723538,0.1093554124,-0.3044160903,-0.132616356,-0.2338767499,0.208268255,0.2471144944,0.4144230485,0.0425139293,0.1491455883,0.2400327325,0.1745164543,-0.1405508071,0.1698295623,0.0849562585,0.149600625,-0.7198398113,-0.0348419026,0.0289795231,-0.0626349375,0.0019792835,0.0834667459,-0.1175395846,-0.411811769,0.3406249881,-0.2460239381,0.1409452856,0.3160174489,0.0178084187,0.1441170573,0.2787432373,-0.1227599531,0.071984686,0.2836942375,-0.4965551198,-0.3666153848,-0.0984075516,0.0370425209,-0.0430696607,0.1910544634,0.4627479017,-0.1594442725,0.3161435723,-0.3744051754,0.1293583661,-0.0766193196,-0.3452595472,-0.4303708673,-0.1973848492,0.3492318988,-0.2921376824,0.1789638847,0.1128091514,-0.1752108783,0.1565735042,0.3864723444,-0.2189261317,0.2811246216,-0.1052904055,-0.0941804722,0.6837505102,-0.2806610465,-0.6752010584,0.2573552728,-0.5013668537,-0.1445819587,0.4097953439,0.2358091772,0.2997538149,-0.0794482082,0.3253081143,-0.0315070562,0.1878620088,0.0564687699,-0.1306353956,-0.1833696067,0.1399638653,-0.0709243044,-0.0423557572,0.1124436036,0.4389398098,0.0218531154,0.3322861195,-0.1725773066,0.3336981535,0.2270243317,0.3746827245,0.3700474203,-0.1073651388,-0.070631884,0.0883536935,0.2813763916,-0.0693755671,-0.2177545428,-0.0564315692,-0.218089655,0.0088667516,0.1772125065,-0.5243186951,-0.2778482139,0.1006738693,0.1897175312,-0.0861744285,-0.1850094199,0.0722659305,0.0684186518,-0.074706614,-0.0877977982,-0.1331679076,-0.0027411866,-0.3007888794,-0.141679287,-0.2466587722,0.1272350699,0.230346024,0.1158881038,-0.1657595187,0.2577354014,0.1691140532,0.2231011242,0.0879704133,0.0498900078,-0.0814919546,0.1020352989,0.158797577,0.0954468995,0.0937408209,0.0492445976,0.1746814102,0.0925376788,0.0108560948,0.3938214779,0.0800941512,0.3215031922,0.0077006505,0.0573983341,-0.0739320442,0.2918778956,0.1616856754,0.1740435958,0.1012408882,0.0455260538,0.0264242776,0.0338476188,0.4503731132,0.0911366045,-0.0042150235,-0.0510589108,-0.1219831035,0.0876495838,0.0985331684,0.4043946862,0.3523017168,-0.0472353771,0.0283693671,-0.0014660567,0.1685306877,-0.0364185572,0.1677808166,0.1691948622,0.313115865,0.1606737673,0.0856153145,0.2174445987,-0.0743992031,-0.3199755251,0.2708792388,0.2420294732,-0.2740249038,0.1205577478,-0.1823228151,0.1474021226,-0.3863972127,0.0602678768,-0.0597069375,0.0446175709,-0.0311288331,0.4068607986,-0.1384714395,0.3061871827,-0.2012708634,-0.0706980675,-0.0700489953,-0.2389224321,0.2810952067,-0.1928637624,-0.1357761621,-0.0142605733,0.5200597048,-0.3461202979,0.3348174691,0.1310818046,0.1168422252,-0.4002668262,-0.1492413282,-0.1489002556,-0.0322173573,-0.0071632434,-0.05770652,0.082321614,0.0533612333,-0.3870067596,0.2943769097,-0.3648621738,-0.1971298605,0.0926634222,-0.1414785832,-0.2842856646,-0.2090806812,-0.1819683909,-0.2725278437,-0.4515674412,0.5456200242,-0.3829762936,0.2032080144,0.0438957587,0.1365900189,0.3035441935,0.1075681672,0.1809549034,-0.0206499323,0.0749729201,-0.0540520772,-0.2144558281,-0.2169629186,0.2142417282,0.2077107579,-0.0423310585,0.3718558252,-0.1384136528,-0.2154281288,-0.3418720961,0.0235143956,-0.1657322794,0.0905483663,0.4791874588,0.2240640819,-0.1020764261,-0.0121041443,-0.0576759577,0.1700657755,-0.0863168687,-0.0665369779,0.2340048552,0.4373008609,0.2596574426,0.4526251554,0.2906372249,0.2908763885,0.0928928256,-0.1935479641,0.2867951393,0.1278290898,-0.316806525,0.2851980031,-0.2700603306,-0.4956001639,0.2250652611,-0.1028251126,-0.1102602109,-0.0529992878,-0.2232200503,-0.2524748743,-0.3226301968,0.1074711755,0.0381970257,0.3169091642,-0.1035541669,-0.003930991,-0.1443039924,-0.2486412227,0.0522293448,0.3439494967,0.0143345166,0.0409765616,-0.1483407021,-0.266276747,-0.3803992867,0.4236616194,0.1634957343,0.4395972788,-0.2032951266,-0.1815982759,0.3152520955,0.0375741199,0.4597308636,-0.0057463311,0.0575062595,-0.0323106982,0.0679840744,-0.1410298347,-0.1038109064,-0.172436893,0.4739091098,0.3283730149,0.3694810867,0.1661260128,-0.0821782202,-0.0833798423,-0.0226210002,-0.2463689744,0.1327558458,-0.4000185132,-0.3016052544,-0.7419366837,0.1716010123,0.2297468036,0.112628229,0.0381696261,0.0588924028,-0.0580217391,0.1704585999,0.0642805994,0.0105290124,0.315136075,0.1231191754,0.039739076,0.0812842995,0.3094007075,0.1304791719,0.3722923994,-0.2756938636,-0.2082936615,0.2713947892,-0.1358176023,0.4039910138,0.571254909,-0.2827678323,0.2027136832,-0.0318034291,-0.1041084751,-0.2370406091,-0.0971106589,0.4168572128,-0.0461195447,-0.0898677483,-0.251837492,-0.1096379682,0.0374985598,-0.2676094174,0.2155771255,-0.1272413433,-0.1754097492,0.28136608,0.0854727924,1.1278810501,-0.1698310524,0.3113788068,0.024305813,-0.2841434777,0.0148329502,-0.0074551646,0.3938042819,-0.3412923515,-0.2458896488,0.0506225713,-0.1569850147,-0.1203958616,0.1006703824,0.0696705878,0.0256377589,0.179423824,0.0938359499,0.0987193659,-0.0899840221,0.1407271475,-0.0488823019,0.0103355786,0.0164595582,0.1089332774,0.2242488116,-0.1826723814,-0.0617940836,-0.0259379428,0.0945737585,-0.3339995742,0.082649529,-0.8847638369,0.2795906067,-0.5237190127,-0.1967059523,0.0685229748,0.4826606214,0.1073735431,0.1613194644,-0.0332433246,0.1616610885,0.5459961295,-0.0964553133,0.0235433392,-0.0513969995,0.2007148713,-0.1294846684,-0.2107151896,0.0201185532,-0.0996212587,-0.3269338012,-0.2084340155,0.4338733852,0.2239693552,0.0029142192,-0.1283280551,-0.1845458001,-0.2664197683,-0.1747767627,0.0659529492,-0.1907125264,-0.1345563382,0.2007707506,-0.4961914718,-0.6412081718,0.2126495987,0.2699280083,-0.1814335734,0.0538589656,0.5651237965,-0.1387382448,-0.2504096627,-0.1887463778,0.1547123939,-0.3191350996,-0.5244018435,0.1637370735,-0.0067657875,0.1218091175,0.0170326605,-0.0523423664,0.2090021074,0.1088749841,-0.1975927353,-0.5801056027,-0.3217696548,0.0204440672,-0.0270668752,0.0711365193,-0.192868799,0.0724066347,-0.0021002609,-0.3044886291,-0.2470369339,0.2354738712,-0.1681189984,0.0797645077,0.1445333809,-0.2157518268,0.0955558941,0.1201053336,-0.0588402376,0.16466932,-0.0294885132,-0.1312187463,-0.1284130514,0.1852892637,0.0036394394,0.0240727719,-0.0250859149,-0.5244274139,0.0952147096,-0.069600381,0.0878882334,0.3729644716,-0.0400488712,-0.181628868,-0.1431842744,0.2240914404,0.0290738158,0.1879339218,-0.3237573206,0.0476397313,0.151443243,0.1387570351,0.0818122625,0.0803238228,-0.1212570295,-0.1599752307,0.3411812782,0.1043951139,-0.2243653983,-0.1746525019,-0.1480349451,-0.0895791277,0.3540413976,0.3658653498,0.0026513885,0.0847171098,0.3677080274,0.0505968854,-0.1701986194,0.1099457517,0.0946589634,-0.6445772648,-0.210069567,0.307020396,0.0962799415,-0.0129726203,0.1107003838,0.0291551482,-0.1314839721,-0.2131491005,-0.0594348051,0.299570471,0.0880743489,0.0795415044,0.2046574354,0.136989817,0.3034844697,0.3255365789,-0.2799254656,0.1091093048,0.1488380134,-0.0644914508,0.1530715525,-0.3837516904,-0.139533028,-0.1055234969,-0.1366044581,-0.0333435088,-0.3139841855,0.6491677165,-0.3166701198,-0.1552521735,0.0183505304,0.3227443397,-0.2984808683,-0.2541522384,-0.172483474,-0.1527472436,-0.2804698348,0.102204673,0.0635208488,0.3036481738,-0.025176473,0.0312171653,-0.0720614046,0.0474079512,-0.3547353148,0.3512820005,-0.0333033726,-0.4196984768,0.0542930365,0.0418187492,0.1526836455,0.1329924613,0.0545922779,0.4572569132,0.2488875538,-0.3006018102,-0.0312285405,-0.0436881632,0.0892655551,-0.1867302507,0.3113063574,0.3029294312,0.1951557249,0.1481228918,-0.0564433299,-0.1130298302,0.0454416759,-0.0977411047,-0.1588467956,0.1349367201,0.2942933142,-0.0361746624,-0.4966941774,-0.1367792487,-0.1217938364,-0.0749000087,-0.0802467763,0.3662536144,-0.2292342782,0.1579884738,-0.1954775155,0.051634647,0.0634429678,0.3620239794,0.4153779149,-0.0330788828,-0.2347083241,-0.1802056283,-0.7036268711,0.5146067739,-0.1440577805,0.0011260638,-0.196144253,-0.1153308377,-0.3221107125,-0.0725278184,0.4101197124,-0.4489621222,0.2706626952,0.3859274089,-0.1417349726,0.1673413217,0.2593232393,-0.1482378393,0.296464026,-0.6339463592,0.1755181402,-0.1820917428,-0.0187496226,-0.0969291925,-0.0914125368,-0.1733268946,-0.0728183389,0.6395336986,0.0932448059,0.1818899661,-0.0565464608,-0.1298192292,-0.606308341,-0.0375937894,-0.2250327915,-0.2564142942,0.2086795568,0.1860785484,-0.2442077547,0.0818965137,-0.2877624929,0.1981612146,0.0348179005,0.0372907035,0.0996131077,0.0312338807,0.1810390949,0.148149237,0.2882189751,0.1698300391,0.013602009,0.1833907664,-0.5974610448,-0.2791656554,0.5037991405,-0.6794379354,-0.525578618,0.293538481,0.3454275131,-0.1101910397,0.1374966353,-0.3126282394,0.0519279726,0.1169415414,0.1351990849,-0.4604374766,0.0032775588,-0.4373631477,0.0121843945,0.1431403309,0.1368299574,-0.0396286212,-0.2378731519,0.0738469064,-0.2436505258]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/590","title":"The process cannot access the file because it is being used by another process (windows)","comments":"@wangcongcong123 thanks for sharing.\n(BTW I also solved it locally on windows by putting the problematic line under try except and not using cache... On windows I just needed 1% of the dataset anyway)","body":"Hi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```","comment_length":34,"text":"The process cannot access the file because it is being used by another process (windows)\nHi, I consistently get the following error when developing in my PC (windows 10):\r\n\r\n```\r\n train_dataset = train_dataset.map(convert_to_features, batched=True)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\site-packages\\nlp\\arrow_dataset.py\", line 970, in map\r\n shutil.move(tmp_file.name, cache_file_name)\r\n File \"C:\\Users\\saareliad\\AppData\\Local\\Continuum\\miniconda3\\envs\\py38\\lib\\shutil.py\", line 803, in move\r\n os.unlink(src)\r\nPermissionError: [WinError 32] The process cannot access the file because it is being used by another process: 'C:\\\\Users\\\\saareliad\\\\.cache\\\\huggingface\\\\datasets\\\\squad\\\\plain_text\\\\1.0.0\\\\408a8fa46a1e2805445b793f1022e743428ca739a34809fce872f0c7f17b44ab\\\\tmpsau1bep1'\r\n\r\n```\n@wangcongcong123 thanks for sharing.\n(BTW I also solved it locally on windows by putting the problematic line under try except and not using cache... On windows I just needed 1% of the dataset anyway)","embeddings":[-0.1660128385,-0.0378365368,-0.0557806827,0.2940466702,0.3540613651,0.1522651315,0.3368764222,0.1749328077,0.2806835771,0.2250535935,-0.1342353523,0.2158219069,-0.1404852122,-0.1683883667,-0.1173466891,0.0108611938,0.0012655875,0.044212468,0.2459802777,0.2743896544,-0.4984654486,0.0607167557,-0.3295137882,0.423772186,-0.2919598818,-0.3280044198,0.1058607846,0.3697049916,-0.073203221,-0.1974390894,-0.1002669632,-0.0513668768,0.1314433068,0.4546838403,-0.0001180091,0.1998518556,0.0130158756,0.0264493786,0.0207782388,-0.0899392739,-0.0563950725,-0.2078053057,-0.1073655263,-0.066042006,0.1141617745,0.0234638974,0.2015790343,-0.4571410716,0.3327825367,0.4084486961,0.1968295425,0.2311187088,0.1027954519,-0.1659160107,0.535613656,-0.1112563685,-0.0627501085,0.633163631,0.5471028686,-0.2220690846,-0.1172490567,0.2611917853,-0.2444720715,-0.1755072474,-0.1392346472,-0.2325413078,0.3693843186,-0.6541210413,0.4994104505,-0.0672556311,0.2061076164,0.0079516917,-0.1065307781,-0.1031479985,-0.1989775598,-0.0291665532,0.2501770556,0.4200428426,-0.1939842254,0.3057061434,-0.1787290424,-0.1260132641,-0.3501567245,-0.0605799817,0.4386653304,0.0405456498,-0.0260064825,0.3118020296,0.3198761046,0.134772256,0.255161196,0.1011163965,0.0105901724,0.2738932073,-0.3511509597,0.263247937,-0.2067010701,0.4223089218,-0.2271409184,0.1090166867,-0.109516643,-0.2546097338,0.2531674206,0.3653867245,-0.0747126415,0.314797163,-0.252848357,0.1199978143,0.3119694293,-0.1192552969,-0.2826243043,-0.0470495634,-0.5266233087,-0.8045742512,-0.1368512362,0.3325532973,0.354899019,0.0067585111,-0.4248265624,-0.1458636522,-0.1495648474,0.1357555538,0.2901575565,0.3832624853,0.0116868084,0.1563884169,0.2046037763,0.2144590616,-0.1127433702,0.0718330964,0.0921355113,0.2212531865,-0.8064484596,-0.0484174006,-0.0237104427,-0.0383469947,0.0575435348,0.0691565275,-0.1150588319,-0.393950969,0.3609554172,-0.2968108654,0.0476291552,0.2490938604,0.0270233378,0.1089756787,0.3133922219,-0.1210688651,0.0360730924,0.2741183043,-0.594530344,-0.4111284018,-0.0860513076,0.0659642816,-0.0545929894,0.1561235189,0.4824938774,-0.0974189118,0.2904032469,-0.3547206223,0.0470065475,0.0024051284,-0.3527668715,-0.4089997113,-0.1699237972,0.2275535166,-0.4454240799,0.1663402319,0.1123914272,-0.2967844009,0.2837726474,0.4242579341,-0.2254429162,0.3697239459,-0.1372352839,-0.1389300972,0.6785027981,-0.1987152994,-0.7276574373,0.3511730134,-0.49825266,-0.2206837982,0.3715967238,0.2742252946,0.3183171153,-0.0356604569,0.4256315827,-0.011666554,0.1258437932,-0.1144285873,-0.1041864529,-0.2354744673,0.2226243168,-0.0457357429,-0.0473345406,0.3152661324,0.3644979894,0.0500470363,0.2891131938,-0.1426452249,0.3079093993,0.2466266304,0.3382252157,0.4254822135,-0.1479532272,-0.1629213691,0.1703911126,0.2562148273,-0.1407578737,-0.2786255777,-0.0121229701,-0.2053186148,-0.1437116414,0.2086337209,-0.5470626354,-0.265106827,0.076324977,0.0634872615,-0.0284603648,-0.1560453475,0.0889445841,0.0087253191,-0.0728094056,-0.097151041,-0.12181364,0.0821682215,-0.3950124085,-0.0846489146,-0.2371075153,0.1842457354,0.2675737739,0.1738244593,-0.1304078102,0.1893813461,0.1834195554,0.1690430045,0.231851697,0.0930386782,-0.026241364,0.0120444121,0.2170071155,-0.0181245822,0.0920294449,0.0306708012,0.3035518229,0.0972554833,-0.0145244868,0.3563963473,0.0089260023,0.2917680442,-0.029410949,0.0325310305,0.1224771515,0.2409925312,0.2025087327,0.1058615595,-0.0261074025,0.0663528293,0.039524354,-0.0594210848,0.4154398739,0.094932422,-0.0369529687,-0.0006532009,-0.1412418038,0.0572685376,0.0906977803,0.2913498878,0.3895952404,-0.0294962283,-0.0036336074,0.0386686549,0.1876131147,-0.068315886,0.1674478948,0.204569295,0.3450443447,0.1418389678,0.0323923305,0.1896013767,0.0150764957,-0.3164551854,0.175363332,0.3392398655,-0.2660369277,0.1201162115,-0.1579214931,-0.041079253,-0.3458783329,0.122357212,-0.07665243,0.0178315733,-0.0643588826,0.4527756274,0.0092966976,0.1905372441,-0.2618510425,-0.0504557639,-0.1107772216,-0.1444194615,0.2394903451,-0.2877012193,-0.1475091577,0.0261554532,0.4310839474,-0.3701074719,0.3875534236,0.0265530068,0.1025777906,-0.3318633735,-0.0847977027,-0.1104743555,0.0069137928,0.0146128265,-0.0169821903,0.1785103381,0.116805993,-0.4084207714,0.3876773417,-0.2849356532,-0.1894249469,0.038214393,-0.2024391145,-0.329521209,-0.0973660871,-0.1877679527,-0.3236171603,-0.3973131478,0.4519200325,-0.4148595631,0.1817364395,-0.0535670221,0.0407032557,0.2566218674,0.1344563961,0.1836619079,-0.0362395085,-0.0234081801,-0.1205165833,-0.1703297347,-0.1716614068,0.2624780238,0.2322476208,-0.1153168082,0.3312906325,-0.1592457145,-0.2347825021,-0.3380510211,0.0444339402,-0.0777586699,-0.0934742317,0.4242709279,0.1935469061,-0.0781683996,0.0072097885,-0.1666396409,0.1691445708,-0.0822771415,0.0318536609,0.2952131331,0.5357869864,0.2394001484,0.4260253608,0.3088305593,0.2428982258,0.1612043679,-0.2039854228,0.178150177,0.08118321,-0.3624286652,0.3248820603,-0.1848437935,-0.5356467366,0.2135555893,-0.1303966343,-0.1177752316,-0.1384094357,-0.2869176567,-0.345482558,-0.1585712433,0.0306948461,0.0713225305,0.3649418652,-0.1111489981,-0.0920349061,-0.0814012066,-0.2938336432,0.0752838179,0.3776858151,0.0201362818,0.1086257175,-0.1931687891,-0.3319662809,-0.3213911653,0.4562102258,0.1692189425,0.4438434839,-0.4118531048,-0.0896592811,0.3133984208,0.1247214228,0.5701313615,-0.0519171543,0.0445696525,-0.1224263906,0.0376479104,-0.0938312411,-0.0465224646,-0.1349647939,0.5151660442,0.4262576103,0.3858271241,0.2570055723,-0.0518756025,-0.0383305661,0.0106332088,-0.253995508,0.0637404099,-0.2831496298,-0.237943247,-0.655844152,0.1182360426,0.1682165414,0.1619084775,0.1417960227,0.0898206234,0.0581239313,0.1918324083,0.144683063,0.021833485,0.2823253274,0.0339359976,0.0422183238,0.106946446,0.4538537562,0.1563448757,0.3275200725,-0.2118463814,-0.137298882,0.2419723272,-0.15978764,0.5036764741,0.4627988338,-0.2284407765,0.1845660806,0.0050520063,-0.1472487152,-0.3052193522,-0.1386516839,0.4399252534,-0.0257143322,-0.1587971151,-0.2194195539,-0.1881217659,0.0239704605,-0.2828882039,0.0737632811,-0.2547082901,-0.2068543881,0.293056041,0.0863483101,1.1771047115,-0.167923525,0.2396931499,-0.1062488928,-0.1760233343,0.0520626418,-0.0099003194,0.3472304046,-0.3507127166,-0.2937557995,-0.0666391328,-0.2492950559,-0.0179036949,0.1763150841,0.0670560896,0.0652056485,0.1141454205,0.0944143683,0.1545519233,-0.1133685634,0.2104421705,0.0443569645,0.0469817333,0.0256701596,-0.0007110463,0.3377893567,-0.227586925,-0.041726917,-0.0867518187,0.1854973137,-0.3375629783,0.1208254844,-0.8556818962,0.2725380659,-0.509742856,-0.1789045334,0.1088634804,0.5916884542,0.2322082669,0.0948024243,-0.07321845,0.1410889626,0.4008791149,-0.1539737731,0.1750455201,0.0856609419,0.1558368653,-0.2459352911,-0.3225769401,0.0867207199,-0.0374830328,-0.2703434527,-0.2978816926,0.2684123516,0.3056305349,-0.0906069502,-0.3078661561,-0.3610241115,-0.1902948171,-0.1502473205,0.05816935,-0.1451313645,-0.1091106758,0.1210773289,-0.4654102027,-0.6315049529,0.1588025391,0.2614533901,-0.1378861219,-0.0487119257,0.4710257947,-0.1170888096,-0.3067562878,-0.1757038087,0.1382467002,-0.3328721523,-0.4727419913,0.0567256734,0.0327367969,0.1449125111,-0.0046044397,-0.0250376761,0.3608408272,0.0328282416,-0.138251707,-0.6702950001,-0.3509264588,-0.0255259704,-0.0082491795,0.0349920839,-0.2053298354,0.0162954628,0.103214018,-0.3115543723,-0.2290213853,0.1851828545,-0.2252602428,0.0900642499,0.2395976484,-0.2171790004,0.1010475531,0.0843688846,-0.0757305101,0.0543475896,-0.0631411076,-0.1334279925,-0.0891866982,0.1684433967,0.0173843503,-0.0289023053,-0.0389247276,-0.4502808452,0.1195157021,-0.0200904664,0.0714456066,0.3995411694,0.0297894813,-0.0857914835,-0.1079765037,0.1343511939,-0.0954786018,0.2518918514,-0.3170964718,0.1222263575,0.2177685946,0.1003891379,0.1017368883,0.045182433,-0.2046897858,-0.218756631,0.4051786065,0.0770388767,-0.172435984,-0.2021098137,-0.0457873121,0.0469361246,0.3182510138,0.2145389169,0.0239803772,0.1379553676,0.3166840076,0.0569055527,-0.1337971836,0.094368346,0.2925770581,-0.5845674872,-0.3134317398,0.1808789968,0.1375522465,0.1393166929,0.1402137727,-0.0156518202,-0.0274777841,-0.2047486752,-0.0686047599,0.2648796737,0.192453444,0.1821684837,0.1630934626,0.2138556242,0.2589559555,0.391667217,-0.1995453686,0.0571524128,0.2089228779,-0.0143969804,0.1258680373,-0.4021158218,-0.0961162746,-0.0148467701,-0.1273846626,-0.0578614287,-0.3358650208,0.6804255247,-0.2341683507,-0.111340113,0.106093213,0.2795393765,-0.3370931447,-0.2484958619,-0.1900666803,-0.1407715082,-0.238608405,0.0732296631,0.0666421875,0.3015091419,-0.0395285152,0.0005210226,-0.1078353152,0.0987060294,-0.28450194,0.3806562424,-0.0815793574,-0.3826133013,0.0141809741,-0.0543736778,0.1845443696,0.2824126482,0.0918021128,0.421800673,0.2285348028,-0.3425934017,-0.0770236328,-0.1309373528,0.0451752506,-0.2144874036,0.3496632576,0.2967481911,0.2587933242,0.2340887487,-0.0217040628,-0.0714710206,0.0370586701,-0.0747130513,-0.1455681175,0.172225073,0.3018600643,-0.035478875,-0.5675351024,-0.03133801,-0.0884139612,-0.1225686744,-0.1041583121,0.3175881207,-0.2436788082,0.2089742124,-0.0893611833,0.03598658,0.0207205471,0.2716797888,0.4336242378,-0.0144852465,-0.2830018699,-0.2416051775,-0.6190906167,0.5306822062,-0.0851750001,-0.0155680086,-0.1069778726,-0.2263483107,-0.3635822833,-0.0004375665,0.3084354699,-0.3861480355,0.1735990644,0.3289051354,-0.0795237944,0.1127683073,0.2845210731,-0.0676444545,0.1665804386,-0.5369420648,0.2253514081,-0.1526335031,-0.0399996527,-0.2405831963,-0.0501343869,-0.2626915872,-0.1540222019,0.5076757073,0.1767824888,0.4080907404,-0.1407747418,-0.0308866948,-0.4318880141,-0.1114625409,-0.0953979045,-0.2588637471,0.2124850601,0.1325124502,-0.2664830089,0.1283311546,-0.2662896514,0.2366710156,0.0138930809,0.053983368,0.0401121415,-0.0055823824,0.1839079857,0.0592538007,0.4290370643,0.1012134179,0.1110894904,0.1096933335,-0.5186725855,-0.1349044442,0.4938797355,-0.5808687806,-0.5136085153,0.3696513772,0.2703973353,-0.1672361195,0.1247319505,-0.1787780225,-0.010374669,0.2156502306,0.1081310585,-0.3498935103,0.0447333381,-0.3796355128,0.0116808517,0.1946570128,0.0808168873,-0.0623395704,-0.2462138087,0.0799528137,-0.2100162357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/580","title":"nlp re-creates already-there caches when using a script, but not within a shell","comments":"Couln't reproduce on my side :\/ \r\nlet me know if you manage to reproduce on another env (colab for example)","body":"`nlp` keeps creating new caches for the same file when launching `filter` from a script, and behaves correctly from within the shell.\r\n\r\nExample: try running\r\n\r\n```\r\nimport nlp\r\n\r\nhans_easy_data = nlp.load_dataset('hans', split=\"validation\").filter(lambda x: x['label'] == 0)\r\nhans_hard_data = nlp.load_dataset('hans', split=\"validation\").filter(lambda x: x['label'] == 1)\r\n```\r\n\r\ntwice. If launched from a `file.py` script, the cache will be re-created the second time. If launched as 3 shell\/`ipython` commands, `nlp` will correctly re-use the cache.\r\nAs observed with @lhoestq.","comment_length":20,"text":"nlp re-creates already-there caches when using a script, but not within a shell\n`nlp` keeps creating new caches for the same file when launching `filter` from a script, and behaves correctly from within the shell.\r\n\r\nExample: try running\r\n\r\n```\r\nimport nlp\r\n\r\nhans_easy_data = nlp.load_dataset('hans', split=\"validation\").filter(lambda x: x['label'] == 0)\r\nhans_hard_data = nlp.load_dataset('hans', split=\"validation\").filter(lambda x: x['label'] == 1)\r\n```\r\n\r\ntwice. If launched from a `file.py` script, the cache will be re-created the second time. If launched as 3 shell\/`ipython` commands, `nlp` will correctly re-use the cache.\r\nAs observed with @lhoestq.\nCouln't reproduce on my side :\/ \r\nlet me know if you manage to reproduce on another env (colab for example)","embeddings":[0.0140547734,0.1271244735,0.0085562682,0.035959471,-0.0078010503,-0.263797164,0.1551121324,0.1553373635,0.3577845991,-0.1724361777,-0.1574480087,0.2879821062,-0.0159159172,-0.1409790665,0.3365516365,0.1459569633,-0.0274409931,0.1052429006,0.1590004116,-0.0974338725,-0.0484839082,0.1789569855,-0.1397798955,-0.168349281,-0.0173317902,0.0266732313,-0.0224758815,0.155035004,0.1056409478,-0.4688453674,0.2854130864,-0.0673121512,-0.095046483,0.1259214133,-0.0001123684,-0.1467732191,0.2711631358,-0.082623899,-0.2608337402,-0.1274135113,0.1139277816,-0.3116106689,0.3685198724,-0.1763194054,-0.0037387295,0.1525944471,0.137779966,-0.167603299,0.5938943028,0.3666618466,0.2137849182,0.2000362128,-0.2147941738,0.197613284,0.2465297282,-0.008641189,-0.1001662463,-0.1787401438,-0.0123234885,-0.3991587162,-0.1550955474,0.4522329569,-0.2373663634,0.0255578868,0.1741417646,0.1520873308,0.169741556,-0.3322789967,0.0382096209,-0.1236505732,0.0047663492,-0.3702811897,-0.1622067541,-0.3918675482,-0.1547731161,-0.5328938961,0.2281316966,0.2824671268,-0.0714175254,0.0352046788,-0.1941294372,0.1388888806,0.172004953,0.0363139659,0.2497341186,0.4306173027,0.0625300854,0.0532180816,0.1106561273,-0.0548592508,0.2229954898,-0.3569903672,-0.0537067614,0.3228704035,-0.2641911507,0.0885866731,0.3014730215,0.4047347307,-0.0454725511,0.2235714048,0.173475638,-0.04801891,-0.0876520574,0.225335598,0.0472554676,0.3800198138,0.3655150831,0.0478012338,0.0904157385,-0.010813619,-0.3588793874,-0.0273606144,0.3107827008,0.1221773773,0.4281068146,-0.0614237674,-0.0855408087,-0.0838730782,-0.2211291343,0.1604165435,-0.1021044776,-0.1083700731,0.2620150745,0.3968604207,0.0097243367,0.1738788635,0.0508838631,-0.0545453094,-0.4351886511,-0.1983556896,-0.174329102,0.1057347655,-0.2974101603,0.187769264,0.4734048247,0.2011061907,0.449016273,0.1165864989,-0.2418131381,-0.2520127296,0.3037593365,-0.1120486856,0.4608233273,0.2307967097,0.196698159,0.0357480943,0.2566693127,-0.2837709188,-0.1765892208,0.1323595643,-0.1480898261,0.0681454539,0.446412921,0.1733050793,-0.0703175217,0.0103104915,0.3509064615,0.0753671378,0.3642429113,-0.5678489208,-0.0792800933,-0.4126440585,-0.363712877,-0.251506865,-0.0087802894,0.4585021138,-0.1346430629,-0.1648874581,-0.1391745508,0.1942669749,0.2277397811,0.3964073062,-0.1332827061,0.0929469988,-0.1476305723,0.0661578849,0.6278782487,-0.4900433421,-0.3707662821,0.1228375509,0.0793688223,0.3197263777,0.1245157644,0.2428554446,-0.0670710653,-0.2258004695,0.2498877048,0.3515277207,0.0406160578,0.082237117,-0.2879509926,0.3968786597,0.325894624,-0.2016420513,-0.1354915053,0.2156065553,-0.1007497683,0.0996777192,0.0922810361,0.013635885,0.1314320713,0.040105816,-0.0570404343,-0.0098581789,0.1718413234,-0.0933756903,-0.0365489274,0.033453308,-0.2740916014,-0.2214099765,0.1858896166,-0.1368455738,0.0164683778,-0.1372162998,-0.1522569656,-0.4593999088,0.1264474988,0.3427806497,0.346229583,-0.1067051068,-0.219979316,0.5316773057,0.1187795922,-0.0748461634,-0.2673670948,-0.0871285871,-0.151425764,-0.2447949052,-0.2198594064,0.1518070996,0.2090099454,0.098716788,-0.155787915,0.192658186,0.2483176738,-0.1676100791,-0.2118837237,0.0143151721,0.0387563407,0.3125903904,0.0876144692,0.1030734256,0.1594628841,-0.1032630876,-0.1650417894,0.3819607794,0.1783604622,0.072118029,0.1815746725,0.0295195691,0.3982230723,-0.2489920259,-0.025376942,-0.2632983923,0.4047840834,-0.1850556433,0.089964658,0.045469474,-0.2181368172,0.3232727051,0.4830141664,0.2757501006,0.0691799819,0.0878766105,-0.0159705598,-0.3959516585,-0.0724546015,0.4940246344,0.4645235837,0.2245407701,-0.0549589097,0.1367132813,-0.1678971052,-0.2303491235,0.1029078588,-0.1981987059,-0.0195112955,0.1492987871,0.3859745264,0.0612337776,-0.3336632848,0.0843749121,-0.0284442119,-0.1334321648,-0.1216012686,0.3592764437,-0.370267123,-0.4869056344,-0.1300432235,0.2459070086,-0.1161158234,-0.0988213941,0.0744908452,0.3453322351,0.0831707343,0.1217840612,-0.0779095143,0.2753459811,-0.1485340148,-0.1314699948,-0.0340485126,-0.2705312073,-0.3422706425,-0.0438280068,0.1777164489,0.0886241719,0.337908417,0.0205061566,-0.0477558561,-0.1575678438,-0.3425697088,0.0794402957,0.0691463202,0.2530013323,-0.0915529504,-0.3343288302,-0.3508038819,-0.0153951058,0.1838683039,-0.3609211743,-0.1059565619,0.0877094343,0.0262161009,-0.1499068588,-0.3578155637,-0.339805454,-0.2054053992,-0.1251423955,-0.0889979973,0.2696052194,0.2463214695,0.5280355811,-0.4274564385,-0.185481146,-0.3738567531,0.2974689901,-0.2761282921,0.1026844755,0.0667226389,-0.2120818049,-0.2694002092,-0.1537080854,-0.2264178246,-0.0149639407,0.1810590029,-0.4297280908,-0.4037624896,-0.1235879362,-0.1860659719,0.0473243669,0.1130345613,0.4327495098,-0.0400290005,-0.112348929,-0.4418010712,-0.3570384383,0.1656191945,-0.0183242783,0.2703091502,-0.1238624826,0.3400833905,-0.0012423571,0.0214991551,0.2903298438,0.0951482877,0.3572483659,0.123428911,0.3901816308,-0.2690267563,-0.2034891993,-0.090188168,0.0914639235,-0.0907981247,0.3345515132,0.1480843276,-0.3454196453,-0.13688384,-0.1049648523,-0.2018754482,-0.1158565134,0.0805864036,0.053337805,0.2972946465,0.5297690034,0.1124679372,-0.3232437968,-0.144285515,-0.3683000803,-0.0390040539,0.3265341818,0.1358303726,-0.2818974555,-0.089818418,-0.2725964189,-0.0929437876,0.0633993223,0.3947662413,-0.2736207843,-0.1881750673,0.1198272631,-0.0901692212,0.3416710496,-0.1776978672,0.1468310654,0.214221254,-0.017210396,-0.3004528284,0.094697386,-0.1667273045,0.3412582278,0.6077418327,0.1224251613,-0.0647705346,-0.0097881537,-0.1268860102,0.2940235138,-0.2101613134,-0.232992962,-0.1707641184,-0.1079775691,-0.1880622059,-0.0429782234,-0.1812502742,-0.1158419177,0.1952055842,0.3438968062,0.1889823377,0.0690553263,0.054589659,0.1227550358,0.3098184466,-0.0418688357,0.1788340062,-0.3252730072,-0.0026634883,-0.5377194881,0.3805280924,-0.1568697095,0.0011065565,-0.1874194294,0.1442615092,-0.0284360461,0.229206115,0.0389027335,-0.0094723785,-0.1985989809,-0.1174351797,0.1762857437,-0.1198255867,0.0579108186,-0.1924186647,0.0621135235,-0.4259145856,0.2208210826,0.0800805315,-0.0958824903,0.1784726083,-0.4109885693,-0.2472111881,0.5547166467,-0.1747338325,0.6807311177,0.2325862795,0.3097771704,0.2408888638,0.0989019498,0.6356295943,-0.0390549488,0.4298944175,-0.3387111127,0.2573567927,0.1301668584,-0.1957311481,-0.2928292155,-0.1590916216,-0.298291415,0.4382796288,0.0669864938,-0.2236717045,0.1064552516,0.2793874741,-0.0087800827,-0.2717784047,-0.4540945888,0.1975494623,-0.1337623,0.1832060814,-0.0059471414,0.0134427641,0.1246393248,-0.2137923092,0.0419076011,0.3979530632,0.0236713309,0.1598170549,0.1413331181,-0.2615765631,-0.3126150668,0.0931331962,0.0415089428,0.012331767,-0.0579652227,0.2015196979,0.2126502246,0.2799979746,0.0238377079,-0.0728585497,0.4331111908,-0.1247008294,-0.0739282221,-0.4255559742,-0.1259181648,-0.2914919555,-0.0023657572,0.4447730482,0.1387238204,0.1159527972,-0.1770390719,0.1464039683,-0.2158010453,-0.1990838796,0.1694985926,-0.0036466999,-0.1287610531,0.2478405535,-0.0036251941,-0.1904655993,0.0074085365,0.0963304937,0.0952242017,0.1120963916,0.3333137929,0.0039070877,-0.1420276165,-0.2184360921,-0.4068237841,-0.3128982186,-0.2059400231,0.250510335,0.0564306602,-0.0420041196,0.2730363905,-0.2061298639,-0.1348140538,0.0696013346,-0.1859170347,-0.4265003502,0.1855749339,-0.0669688955,-0.2016600668,-0.2573540509,0.4571641684,0.277107656,0.0285784472,0.2128786296,-0.3180309832,-0.135241732,0.1126771569,0.3225851059,0.0394788124,-0.1526788622,-0.1495839357,0.0955018327,-0.0041965409,0.1972425729,-0.319049269,-0.2811469436,-0.0855115801,0.1044004261,-0.0046493318,-0.1303962171,-0.2635141015,-0.041873958,0.0824362859,-0.3643066585,-0.030125007,-0.0617691316,-0.1133384556,-0.0641006753,-0.15008232,0.141518712,0.3738201857,0.0577138588,-0.3782220185,-0.1485735029,0.0266934615,0.1841719151,0.0111783221,0.0454672016,-0.3978869021,-0.1515819281,-0.1233687177,0.1552700549,0.1819794327,-0.3702733517,-0.0786006153,-0.099212274,0.2283598632,0.2684122026,0.0206625108,-0.0115412287,0.2399541587,0.1466327459,-0.2967821956,0.023371188,0.3732441366,0.15947631,0.064662464,0.1382667869,0.258161664,-0.0944325849,0.3471428156,0.1362633258,0.4831183553,-0.5962883234,-0.0016912008,-0.1406365782,0.2338176519,0.0332976356,0.2633271515,-0.176309377,0.0219205618,0.4350858629,0.1924680471,0.3204328418,0.3400508761,-0.0985441133,-0.2971983254,-0.1915215701,-0.1900567412,-0.0022808823,0.0494239964,0.1495681405,-0.001623097,0.0589980707,-0.1581043601,-0.1077885926,-0.1957950443,0.4848647416,-0.0400336199,-0.1616549194,-0.0043452582,0.0572854057,0.2180641741,0.1310000271,0.1196314692,0.4419143498,0.0766541958,0.0482812449,-0.0525682233,-0.2858624756,-0.2747980952,0.0684224144,0.1314973831,-0.2465025634,0.4550008774,0.1307494044,0.1033761725,-0.1723936945,0.1569487453,0.5075445771,0.4611142874,-0.0035801916,0.0155276209,-0.2227503955,-0.0329935588,-0.3223071694,0.3511278927,0.2794367671,-0.0107299574,-0.004351981,0.0595913641,-0.0882378742,0.1745178849,0.2891914845,-0.2743089199,-0.1138385385,0.3731920719,0.2556148171,0.0462459922,-0.4124015272,0.2220692635,-0.4528362751,0.1879248917,0.3555687964,0.1741958112,0.2492036074,-0.365042448,0.0916589722,-0.0503007211,0.2403859049,0.1078092307,0.0505630597,-0.084949024,0.0670368299,-0.70586133,0.0267197583,-0.4562807083,-0.1184796169,-0.1871311963,0.3274886906,-0.3748757839,0.0732662007,-0.1305035502,-0.2199995816,0.1314416081,-0.3093394041,-0.2394201905,-0.0231264252,0.168838948,0.0829377547,-0.0250741914,-0.2326982468,-0.021565659,-0.4717831016,0.2197501808,0.0933376551,-0.2499940544,0.1657356918,0.6163634658,0.1083637699,0.4907679856,0.4504414797,-0.0541110113,-0.0049738516,-0.3906283081,-0.1123801246,-0.0764643699,-0.3105053306,0.0192334764,0.39261204,-0.4182898998,-0.087288253,-0.2374890745,0.4055718184,-0.1793858409,-0.0698678941,0.1784398258,0.0141679645,-0.119570449,0.3620304167,-0.2111615241,0.4743294716,-0.0432452224,0.3824567795,-0.3174717724,-0.3478336334,0.3471221626,-0.6071430445,-0.3471948504,0.0734205469,0.2991023362,0.0560424067,-0.4155086279,-0.2385803163,-0.0634935573,0.0880297124,-0.0643564388,-0.2337215394,0.1836695373,-0.3016317487,-0.0839672387,0.0482020043,-0.2365951985,0.3630095124,-0.3723286688,0.2081933171,-0.0050064325]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/580","title":"nlp re-creates already-there caches when using a script, but not within a shell","comments":"Fixed with a clean re-install!","body":"`nlp` keeps creating new caches for the same file when launching `filter` from a script, and behaves correctly from within the shell.\r\n\r\nExample: try running\r\n\r\n```\r\nimport nlp\r\n\r\nhans_easy_data = nlp.load_dataset('hans', split=\"validation\").filter(lambda x: x['label'] == 0)\r\nhans_hard_data = nlp.load_dataset('hans', split=\"validation\").filter(lambda x: x['label'] == 1)\r\n```\r\n\r\ntwice. If launched from a `file.py` script, the cache will be re-created the second time. If launched as 3 shell\/`ipython` commands, `nlp` will correctly re-use the cache.\r\nAs observed with @lhoestq.","comment_length":5,"text":"nlp re-creates already-there caches when using a script, but not within a shell\n`nlp` keeps creating new caches for the same file when launching `filter` from a script, and behaves correctly from within the shell.\r\n\r\nExample: try running\r\n\r\n```\r\nimport nlp\r\n\r\nhans_easy_data = nlp.load_dataset('hans', split=\"validation\").filter(lambda x: x['label'] == 0)\r\nhans_hard_data = nlp.load_dataset('hans', split=\"validation\").filter(lambda x: x['label'] == 1)\r\n```\r\n\r\ntwice. If launched from a `file.py` script, the cache will be re-created the second time. If launched as 3 shell\/`ipython` commands, `nlp` will correctly re-use the cache.\r\nAs observed with @lhoestq.\nFixed with a clean re-install!","embeddings":[-0.0163848605,0.1808854491,-0.0174822975,0.0194336046,0.0402135178,-0.2356126457,0.1533257961,0.2469057292,0.308850795,-0.1499484479,-0.1733098328,0.2903880775,-0.0571840368,-0.138614729,0.2896094024,0.0798135847,-0.0757226571,0.1186784953,0.1688123047,-0.0410380438,-0.0695227459,0.2125500292,-0.109913893,-0.1152573228,-0.0178042073,0.0234994497,0.0401806645,0.1861426532,0.0676668733,-0.4850822091,0.2590028048,-0.0646582246,-0.1197368726,0.1224610209,-0.0001095846,-0.1284681559,0.2958439887,-0.0514415279,-0.2693364918,-0.1360889673,0.1393997073,-0.2887373865,0.3314529359,-0.1660370976,0.010441849,0.1462468952,0.1281618625,-0.1483440697,0.5328407288,0.3832762241,0.252543807,0.2368161678,-0.1922276318,0.1802837402,0.2751609087,-0.0337940939,-0.0853516534,-0.1610592455,0.033648897,-0.4193161726,-0.1620030999,0.4727592468,-0.233646825,0.0124069862,0.1643406004,0.1199935973,0.1583965272,-0.300606668,0.0513211079,-0.1485339254,0.0105904732,-0.3625773191,-0.1241604909,-0.4010258913,-0.1763983667,-0.5047300458,0.2171524018,0.2590265572,-0.0639419854,0.0217726491,-0.1965945065,0.1315354854,0.1612571627,0.0520058051,0.2808975875,0.3880451024,0.0478264093,0.0564263463,0.1282859147,-0.0597739294,0.179985404,-0.311213553,-0.0448573343,0.312299341,-0.2241398692,0.0471305773,0.3617567122,0.3881236911,-0.0609293506,0.2444334775,0.1549114734,-0.0542739183,-0.0703350753,0.1804027557,0.0495468304,0.3550102115,0.3591275513,0.0375033021,0.0850110203,0.0241857842,-0.2733801901,-0.0229116306,0.3330259323,0.0689406246,0.3928884566,-0.032821998,-0.0112956772,-0.0948278159,-0.2497217506,0.1604578495,-0.077647768,-0.1087362096,0.2466953248,0.3870065212,0.0304694492,0.2157053798,0.0933611393,-0.0205758754,-0.4181882143,-0.2028553635,-0.2009446025,0.139373675,-0.2775062323,0.1811218262,0.4646053612,0.1827657074,0.4519316256,0.1055932418,-0.239763841,-0.2302652299,0.317193687,-0.1082947925,0.3961496353,0.2584691346,0.198184222,0.0409629717,0.2748934925,-0.2642138302,-0.1747832,0.1178346053,-0.1373906285,0.0160626639,0.3764325082,0.2090303749,-0.0295771826,0.048584491,0.3228157759,0.0816322342,0.3573574722,-0.57915169,-0.0860106349,-0.4147101343,-0.3481357098,-0.2728207111,0.0054101981,0.4563548267,-0.1826176494,-0.1603648365,-0.1639048308,0.150614813,0.2062464356,0.3887485266,-0.1500096321,0.1604389399,-0.1398696154,0.1020540297,0.632163167,-0.4810979366,-0.3584729135,0.1374000609,0.0357045531,0.3112778962,0.0491752811,0.2283137888,-0.0330055095,-0.2177517712,0.269087106,0.3208866417,0.0252104141,0.084854506,-0.3005706668,0.3707406521,0.2937579453,-0.1620469242,-0.1098939925,0.2290362567,-0.1051124036,0.1186838001,0.1017028168,0.026356956,0.1409219652,0.0459504575,-0.0326981694,0.048349373,0.1568864435,-0.1516822577,-0.0476460606,0.0333223306,-0.2704016268,-0.1751034856,0.1228271127,-0.1482572854,-0.0298038162,-0.1538022459,-0.1173644736,-0.4119267762,0.1680992544,0.3050995171,0.3428573012,-0.1079819575,-0.2254121155,0.5092610121,0.0937320143,-0.0801807046,-0.3031412959,-0.0758911073,-0.1527346522,-0.2435359359,-0.1666399091,0.1333344579,0.2314141989,0.1225419492,-0.1618217677,0.2263352573,0.204769209,-0.1820966154,-0.190355733,0.0334212631,-0.0118048601,0.2786030173,0.0784483925,0.108813487,0.1585495472,-0.1108423173,-0.1280681044,0.4049761593,0.1913359016,0.0457262509,0.1564854681,0.067398034,0.3959316313,-0.2378958464,-0.0090031791,-0.2537346184,0.421733588,-0.1836773455,0.0917784497,0.0225657821,-0.2371750772,0.3287149072,0.4718925059,0.2593461573,0.0505435541,0.0899181589,0.0066971066,-0.3520186245,-0.1030310094,0.4618034661,0.4859337211,0.2363109887,-0.101813145,0.1147034764,-0.1891305298,-0.2176958919,0.0888578221,-0.1688321829,-0.0391281694,0.1118563041,0.3624937236,0.0706143454,-0.3688943684,0.0535768308,-0.0646191165,-0.0716492385,-0.1583639085,0.3442477286,-0.3108969629,-0.4783163071,-0.0931905955,0.2242314965,-0.1344704479,-0.1384939253,0.1109563485,0.3637188077,0.0270671938,0.1130055934,-0.1080298722,0.3232314885,-0.1177449226,-0.1193372533,-0.0852189064,-0.2791736722,-0.3344272673,-0.0053019435,0.2029283643,0.0920410976,0.3395648003,-0.0037639008,-0.0486115664,-0.1372338384,-0.3448315561,0.0217181984,0.0257772952,0.2438466698,-0.0504440293,-0.2631428838,-0.3051371276,-0.0383138917,0.2150997072,-0.3725755215,-0.104474619,0.0748903751,0.0130425077,-0.1462853849,-0.3992411494,-0.3698316216,-0.1797045171,-0.1653824747,-0.0471812673,0.2538398802,0.2492615283,0.4601083696,-0.4260555208,-0.1431338787,-0.3873452246,0.3033135235,-0.286631763,0.0584384799,0.0519692488,-0.2260954529,-0.2945150733,-0.1549081504,-0.1850316823,-0.0141755398,0.1260770708,-0.4229622483,-0.3771382272,-0.1413822919,-0.1743396521,0.0361652784,0.0850762948,0.4376715124,-0.010772327,-0.1220099553,-0.4587093294,-0.3348917663,0.1501431018,-0.0568234324,0.2300712615,-0.1559386402,0.3591262102,0.0320761316,0.092784971,0.242428571,0.1294802278,0.3537249863,0.1401235163,0.4094706476,-0.2524750531,-0.2242056876,-0.1014506742,0.0600208119,-0.0378232971,0.3222278953,0.135826841,-0.3781397343,-0.1549308747,-0.0794192851,-0.2426045388,-0.0980465189,0.1346310228,0.0384272635,0.2618247569,0.5380387306,0.1150059402,-0.3600819707,-0.1497326046,-0.3338670433,-0.0641177222,0.2761481702,0.1247978434,-0.2871512771,-0.0670799315,-0.3112904131,-0.0799868479,0.079359822,0.3748309016,-0.2727255821,-0.1902433485,0.101990141,-0.062186081,0.3638741672,-0.1732073277,0.1371023357,0.2531927228,0.0323906057,-0.3218532205,0.082708165,-0.2083588392,0.3547706902,0.654278338,0.1858098507,-0.094383277,-0.0220677685,-0.115590319,0.2448103428,-0.1969492733,-0.2081068009,-0.1477949768,-0.1416478455,-0.1754560322,-0.0206798408,-0.1288987994,-0.0114646312,0.1552335322,0.3832678199,0.2340631336,0.0542438515,0.0185277928,0.1447094679,0.3353135288,-0.0281673167,0.1445558667,-0.2956347466,0.0099893939,-0.5055728555,0.3856001198,-0.1293625534,0.0128435306,-0.1667750627,0.1344214827,-0.0486471169,0.2458149642,0.0305995345,-0.0187623762,-0.2070365101,-0.0503099933,0.141648531,-0.0696873143,0.0809978098,-0.2289444208,0.0494824387,-0.3814747036,0.2394323647,0.087516509,-0.0794157237,0.2394650131,-0.3620915711,-0.2482351661,0.5374693871,-0.181423977,0.6791018844,0.2271070927,0.2542057335,0.2552260458,0.0510992929,0.6143396497,-0.0561419353,0.4320717752,-0.3459079266,0.2442657202,0.133020997,-0.1812045872,-0.2793732882,-0.1547368169,-0.2938479781,0.3962006569,0.0454188734,-0.2541280687,0.1254790872,0.2795797586,-0.0148631874,-0.2426230758,-0.4268494248,0.2331208289,-0.1737080961,0.1800774783,-0.0243471321,-0.0024650376,0.1094544828,-0.2520020902,0.02674588,0.3773344457,0.0279505309,0.1672255397,0.1589614898,-0.2091281563,-0.3004730642,0.0372649804,0.0361641273,-0.0020329815,-0.0824225098,0.1884731352,0.2241062671,0.2597307861,0.0844131857,-0.0049434113,0.4326538742,-0.111655429,-0.0248407722,-0.3890338838,-0.1430266351,-0.3102144003,-0.0071311248,0.4139478207,0.1080872566,0.0694315061,-0.2178801447,0.1068437025,-0.1847339869,-0.2082326114,0.2030316591,0.0151202418,-0.1358782947,0.2105736881,-0.0204818174,-0.1737279594,0.0137052359,0.0930986255,0.0616886131,0.0948032662,0.3192842603,0.0177386235,-0.1626371145,-0.2111219913,-0.4049542248,-0.2701284885,-0.2443625778,0.2841250896,0.047801327,-0.0323062874,0.2544871569,-0.1804123521,-0.1308072507,0.0768236071,-0.1604642719,-0.4598062634,0.1663903147,-0.0342476964,-0.2320073992,-0.2524180114,0.439481914,0.2995864749,0.0456276573,0.1986074895,-0.357958436,-0.1303530484,0.0404821038,0.3190901577,0.0338478535,-0.1714097112,-0.1262425184,0.0893070325,0.0155827813,0.223327592,-0.3173328936,-0.3138656616,-0.0652142987,0.0867029577,-0.0242532492,-0.1462834477,-0.2677054703,-0.0252692662,0.0763793662,-0.3268355131,-0.0598268695,-0.0397243351,-0.1265114397,-0.0502296239,-0.1224550083,0.1055080295,0.3949957788,0.0473887511,-0.3532531857,-0.1791968793,0.0049133673,0.1968066096,-0.0024144198,0.0390588567,-0.3728185296,-0.125752762,-0.1506801397,0.1872487962,0.2027030587,-0.3648320735,-0.0955219343,-0.0626938865,0.2578339875,0.2819534838,-0.0164126363,0.0140372487,0.2067246288,0.1710170507,-0.2935049832,0.0198231023,0.344925344,0.173107177,0.059071295,0.0947375,0.2336835265,-0.1005313993,0.3182949424,0.1716630161,0.4767451584,-0.5991134644,0.0040631369,-0.1387659609,0.2316904664,-0.0037187128,0.275524348,-0.1571354121,0.0111683076,0.4345908463,0.1800847948,0.3217641413,0.2501147985,-0.1283851713,-0.314143151,-0.1930382699,-0.1985817105,0.0017594633,0.0229259692,0.1305348873,-0.0038566948,0.0566038899,-0.1757298857,-0.143805176,-0.1628782451,0.4128829837,-0.0973007828,-0.1288700402,-0.0412256308,0.0709436536,0.21974805,0.1207777336,0.1164015681,0.4147906899,0.0769294947,0.029880628,-0.0849478245,-0.2961078882,-0.3043581545,0.0618518591,0.1312175691,-0.2512078881,0.4048873484,0.1072721183,0.0542848371,-0.1469626725,0.157033816,0.49894315,0.466345042,0.0071017738,0.0020131045,-0.2259946018,-0.0420803837,-0.318002522,0.3545324206,0.3158841729,0.0203844439,0.0118556591,0.100838162,-0.1233556867,0.1598566175,0.2425422221,-0.2012317628,-0.0997380763,0.3651478291,0.1975165308,0.011045048,-0.4154348969,0.2065033168,-0.4409568012,0.1596950591,0.3317349851,0.22333318,0.2513045669,-0.4059313536,0.1068980545,-0.0158918202,0.2367562801,0.1420716494,0.0245290436,-0.104107216,0.0101916464,-0.6859769821,0.040206939,-0.4289596677,-0.1105654314,-0.1612906009,0.3549923897,-0.3604249358,0.0779914111,-0.1011489257,-0.2092361301,0.0973486528,-0.3575369716,-0.2522353232,-0.0033853403,0.1499662846,0.0586648062,-0.0092397323,-0.2017405778,0.0029419207,-0.475674808,0.2480868995,0.0562626198,-0.275380373,0.1618616581,0.6741089821,0.0306341015,0.4666767716,0.4286803007,-0.067371279,0.0113932714,-0.4140633941,-0.1138645038,-0.0865443125,-0.3312041461,0.0058459416,0.4544873834,-0.4172289371,-0.0608514696,-0.2611103654,0.4375708997,-0.1769771427,-0.0408848561,0.1512593031,0.035054706,-0.1236311272,0.3606989384,-0.1900017112,0.4287849665,-0.0275687873,0.394289732,-0.3357802331,-0.383400619,0.3927365541,-0.5918278694,-0.3927876949,0.1133143902,0.2804989815,0.0448979437,-0.4229475558,-0.2715828717,-0.086101383,0.0792752579,-0.0398889668,-0.2370935827,0.1901362091,-0.3297051787,-0.0778002217,0.0747225061,-0.2223463506,0.3650073111,-0.3238643408,0.2303884178,-0.0210236125]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"Some wikipedia languages have already been processed by us and are hosted on our google storage. This is the case for \"fr\" and \"en\" for example.\r\n\r\nFor other smaller languages (in terms of bytes), they are directly downloaded and parsed from the wikipedia dump site.\r\nParsing can take some time for languages with hundreds of MB of xml.\r\n\r\nLet me know if you encounter an error or if you feel that is is taking too long for you.\r\nWe could process those that really take too much time","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":88,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nSome wikipedia languages have already been processed by us and are hosted on our google storage. This is the case for \"fr\" and \"en\" for example.\r\n\r\nFor other smaller languages (in terms of bytes), they are directly downloaded and parsed from the wikipedia dump site.\r\nParsing can take some time for languages with hundreds of MB of xml.\r\n\r\nLet me know if you encounter an error or if you feel that is is taking too long for you.\r\nWe could process those that really take too much time","embeddings":[0.2373758107,-0.2088861763,-0.1352212578,0.4252154529,0.1233771443,0.2758935392,0.1313822865,0.1356951892,0.7319019437,-0.2289887667,0.0522107147,0.0235065091,0.1044930071,-0.2209710777,0.067934826,-0.1553103626,0.0418751948,-0.1114457995,0.1551072299,-0.3610844612,-0.3178491592,0.2384437621,-0.3044137657,0.0057737767,-0.2436927557,0.2308339477,0.0614723749,-0.1755262315,0.0371080935,-0.3282756805,0.1453267783,0.2548292279,0.2869493663,0.1309534311,-0.0001185968,-0.1286964715,0.704265058,-0.1099032238,-0.4838129878,-0.121251151,-0.101447098,-0.4541423023,0.1855032146,-0.1290799081,0.0664206818,-0.2606805563,0.2823228836,-0.5186074972,0.1401310861,-0.0058512911,0.1931751817,-0.2319425642,0.0404679999,0.1306501776,0.3112449348,0.0921855792,0.172114253,0.1484159082,0.0676715821,-0.2165156007,-0.0645688474,0.1268210262,-0.164436087,-0.0737825707,0.1959549189,-0.1150276139,-0.1769626737,-0.6844722629,0.1788195372,0.2517311573,0.7159077525,0.0819102451,-0.3249859512,-0.2314273566,0.0833554491,0.0490160696,0.1674252898,0.3484603465,-0.104491137,-0.0879922211,-0.045023784,-0.3324614167,-0.0504330061,0.5696833134,-0.1381946951,0.3546818197,0.1198177859,0.1225396991,0.0333931856,-0.1997120678,-0.2386387736,-0.2273808122,0.423584193,0.4048978686,-0.2517802119,0.4409765303,-0.0247038472,0.2985454202,0.026323637,-0.1559597701,-0.227429226,0.2422037423,-0.1045706049,0.1780498922,0.133090362,-0.0985689536,0.2461398393,-0.1657909006,0.3742342591,-0.0247806441,-0.1900200099,0.0311534312,-0.3043862283,-0.1849526614,-0.19125247,-0.0971256346,-0.0666854456,-0.3103949428,0.1589362919,0.157227993,-0.2838952839,-0.261797756,-0.1157063544,0.2324323803,-0.110389702,0.2289988548,0.179744184,0.3324601054,-0.4585287869,-0.0347173102,0.0760531202,-0.0932939351,-0.1863993108,0.1857512146,0.3903912604,0.1105753854,0.3006482124,-0.0655988231,0.0075084828,-0.1033298746,-0.0583558753,-0.2353228629,-0.0578199178,0.0070611779,0.1893652081,0.5287054181,0.0204354096,-0.268887341,-0.1559168845,0.1354001462,-0.1534598321,0.3921814859,0.0703725144,0.1048097163,-0.3071572781,-0.1301625818,-0.275441587,0.2882411778,0.1407090575,-0.1476809829,-0.0730014816,-0.2850129604,-0.1553779542,-0.0928238928,0.3056333363,0.5948296785,-0.2702881098,-0.1211088523,-0.169947505,0.1433556229,0.3089423776,0.2673920691,0.0071027861,0.2469750345,-0.1199265495,0.1096144766,0.273052305,0.0004276375,0.0255027823,-0.0254085846,0.2584200501,0.2935872078,-0.0431686603,0.0379178226,0.2038619369,0.1188984886,0.1142759398,0.5594838858,0.1990445405,-0.0339212455,-0.3283992708,0.0008387555,0.557341218,0.3541518748,0.1461993307,-0.2276218385,0.0513215214,0.303647697,0.1359987706,-0.0113095017,0.0927649662,0.3572766185,-0.3393714726,0.2750453353,0.2403738499,0.0269897766,-0.552510798,0.2369062901,0.0318737924,0.1359116584,-0.0013061247,0.1318639964,-0.2870848179,-0.1605425328,-0.2156870961,0.2611429393,0.0749469623,0.1703058183,-0.11367625,0.4636848867,0.0636296049,-0.1318305582,0.0043708328,-0.1513058692,-0.5514069796,0.4124667346,0.046518404,-0.0663193241,0.0748853236,0.1963200569,0.2826358378,-0.1605145335,-0.0965554863,-0.2543490529,0.3377592266,0.2528760135,0.2167096883,-0.07920596,0.2825820148,-0.4112240374,0.2441625744,0.229599759,0.1425693184,-0.2165514231,-0.0493407734,0.2202326953,-0.1231868044,0.4874270856,0.0636299476,0.0318964235,0.3125784099,0.0732511804,-0.1109998748,-0.076010786,0.5720253587,0.1005094498,0.0688266531,0.2869985402,-0.208210513,-0.1159858629,0.5724138021,0.0600974187,0.0017771302,0.268658936,-0.1495233625,-0.0875654295,-0.0348374359,-0.2696869671,-0.2388126552,0.1612473577,0.0583887622,-0.433024615,0.0328254029,-0.1703622937,-0.0219371021,0.1148250327,0.2563506663,0.1282822043,0.170076713,-0.0760187954,-0.2888402045,-0.0338465795,0.0780994147,0.1424240619,-0.1322094947,0.0680787638,-0.3862009645,-0.3865298927,-0.2589558661,-0.2271352112,-0.4857493043,-0.4379450679,-0.1076668277,-0.0235601161,0.1001374274,0.0911559463,-0.0419983156,-0.0005249568,-0.2282491475,-0.1849108189,-0.3860189021,-0.434535712,-0.3712606132,-0.0285548996,0.5293226242,0.15528135,0.1042223945,-0.0408951119,-0.179421559,-0.1444323063,-0.3140221536,-0.243980214,-0.3181396425,0.1540388018,-0.3206973374,0.4155473113,-0.1617505699,-0.052578643,0.3501430154,0.0088333795,0.0303261727,0.2139815837,0.2482554615,0.1368112415,0.0138354339,-0.4436653852,-0.2476265579,-0.2580005527,-0.0753559396,0.1386438757,-0.0286487266,0.0223311484,-0.108826004,0.0438175462,0.1129069924,0.013307808,-0.0412299968,0.3525742292,0.5122914314,-0.0197107866,-0.3502418399,0.1122015491,0.1237486452,0.0851646215,0.1428358555,-0.5253459811,0.2483905107,-0.227443859,0.089688845,0.1307523102,-0.1152568534,-0.0421737619,-0.162675187,0.1069582924,0.0575448573,-0.1961786598,-0.215112105,-0.022867471,0.1928330809,0.0768851489,0.3748746812,-0.3542456031,0.6905225515,0.1530368179,0.1087186933,0.2486219108,0.2776161432,-0.0156387351,-0.1505483985,-0.2539342642,0.0796210393,-0.052961316,-0.0789796039,0.3414691091,0.0570947453,-0.4630891979,-0.1659130454,0.183831349,-0.3000289202,-0.2447899133,0.1687377244,-0.276878655,0.1899990141,-0.0978071317,-0.0024241572,0.0266600251,-0.4019005895,-0.0604734421,0.2074945122,0.459682703,0.1586940289,-0.3370335698,-0.4819998741,-0.2602394223,0.2074870169,-0.0078574931,0.5056961775,-0.288292855,0.0243108403,0.1711665541,0.1254115403,0.190741837,-0.2676234543,-0.1190828681,0.4583571255,0.0570353754,-0.6286123395,-0.0129506662,-0.131342724,0.1177796349,0.3597909212,0.1823291183,-0.4848775864,-0.2624340951,-0.1122146398,0.6234146357,-0.1808881015,-0.0617818721,-0.4719067216,-0.1900830418,-0.3897494376,-0.17545937,-0.1965720803,0.1376363188,-0.2524032593,-0.0068889749,-0.0342991389,0.2318183333,0.1261013895,0.1136706546,0.118502371,0.2835496366,-0.0647996739,0.1833137572,0.1066991612,-0.1316339374,0.3084096611,0.0355661213,-0.1813290119,-0.0142122284,-0.1458048522,0.2586133182,0.0000785487,0.1404329836,0.07072822,0.0779155418,-0.1616079658,-0.0759618655,0.0272244923,0.2273463905,-0.0392969362,-0.3217107356,-0.3995402455,0.6012827754,0.0243771039,-0.0773069859,0.2754061818,-0.0577723011,-0.3797739148,0.4766834676,0.0389671773,0.8272898197,-0.4226615429,0.0835661814,-0.1444535106,0.1735061109,0.4089493752,0.008670602,0.0363030136,-0.2705720365,0.3124822974,-0.0021859307,0.0716995299,0.0022460518,0.1210937202,-0.1676385105,0.2851983011,-0.1164297536,-0.0431564488,0.0004537301,0.5521711707,0.1862845123,-0.3387275934,-0.3292571008,0.0565183721,-0.190905422,0.3705632389,-0.0544358641,0.1684122235,-0.0676842779,0.0371417478,-0.2191146463,0.1383495778,-0.0738876089,0.2030120939,-0.2177364528,-0.2613622546,0.3470251262,0.4555535018,-0.0610570163,0.2965242863,-0.1833679229,0.1614168435,-0.1743264645,-0.2580781579,-0.0258767828,0.039228972,0.0128863938,0.0260216706,-0.3502269685,0.0602129027,-0.0239021908,0.0278149582,-0.4295760393,0.1425749511,0.0463308506,0.1478120089,-0.2561358213,-0.015678782,-0.1960348338,0.0551938601,0.1045234725,0.0453430191,0.2451972812,0.1529280394,0.4280240834,-0.2785580158,-0.0932602286,0.2778621018,0.0488083288,0.032817591,0.7834230065,0.3793965578,-0.3779433668,-0.2083092779,-0.303198576,-0.4560174346,-0.3899913728,-0.2934160531,0.1242031157,0.383982569,-0.2075721174,-0.1481448412,-0.0268963333,-0.4575170875,-0.0013409202,-0.6665328145,0.054934714,0.3396061957,0.0496392772,-0.0022722993,-0.0495142639,-0.4748337567,0.1891742051,0.0927148089,-0.1784785688,-0.0465924889,-0.2098364085,0.1940191239,0.290284127,-0.0305012781,-0.1945335269,-0.1245620474,0.0715634152,-0.2864050567,0.1808196902,-0.1599361897,0.0942240506,0.1508226097,0.1461357176,-0.104981333,-0.1656461805,-0.4103103578,0.4030146897,-0.3441690505,-0.1329703331,-0.1333861053,-0.0661587343,0.2610643208,-0.0318381526,0.230648756,-0.0626973957,0.1877810359,-0.0556442775,-0.0242141876,0.2736472189,0.0275777522,0.6309036016,0.127067551,-0.4748721719,-0.0672629252,-0.0513816252,0.1632102132,0.251910001,-0.1646305323,0.3837778568,0.3636279106,-0.0786023289,0.5536407232,0.0051677246,-0.0831237957,0.2831914723,0.0924541429,-0.1920320541,0.1026499122,0.6308951378,0.1232615411,-0.1636996269,0.1788730472,0.1540611684,-0.2622277439,0.0061379955,-0.0735289156,0.1782835722,-0.0653238967,0.1384174973,-0.0279031768,0.2722364664,0.0861174315,0.0418895259,0.102583997,-0.1573581398,0.0006607387,0.1196801737,0.1837936193,-0.0361657888,0.1555347294,-0.2045336515,-0.3261486292,0.0495486781,0.4282365739,0.0327397659,0.2258841544,-0.0281022601,0.1899497956,0.1641092747,-0.2246441394,-0.1284000576,0.2254067808,0.0386079662,-0.2214490622,-0.159989655,-0.0641025305,-0.1283914149,-0.0061341254,-0.0202885177,-0.2115823179,0.0185392406,0.4602557123,-0.1359467357,-0.5251308084,0.5226477385,-0.2327724993,0.2248285115,-0.1313632131,0.1586133391,-0.0202140193,-0.0204820521,-0.0203272905,0.1130882949,0.0961166993,0.3027553558,-0.4491692483,0.0732996687,-0.0639269054,0.1251825839,-0.0428359807,0.5011216998,0.0970254689,0.2795985341,0.2483770996,0.1094286889,-0.0767133459,-0.2166141421,0.2603474557,-0.0639564693,-0.1150573045,0.0645175651,-0.1318162531,0.0212891418,-0.3001458049,-0.0776860118,-0.3329944909,0.195884034,0.2172612846,0.1616084874,0.032801453,-0.0893495083,0.0220451001,-0.1579569429,0.6827074885,0.5152642727,0.3482388556,-0.2263831049,-0.1950800419,-0.5486900806,-0.0545453727,-0.3532513976,-0.0403818302,0.2484322786,0.2327834815,0.1065830961,0.06349691,0.0282984972,-0.1706920117,-0.1625119895,0.2500012219,-0.3407201469,-0.1809644103,-0.1340546161,0.0768587813,-0.3858617544,-0.1097125486,0.1934162229,-0.258818239,-0.0511220805,-0.0512895994,-0.0344718993,0.0920769647,0.2395557314,0.227883935,0.0322224535,0.6150705218,0.0347039811,0.0201276224,-0.2817748487,-0.2000225633,0.1077022627,0.386624366,-0.2164957672,0.297326088,-0.1123961285,0.1421588361,-0.2420965582,0.3953529596,-0.1706999093,0.0467420183,0.0303892605,-0.2690484226,-0.0076742023,-0.1595900506,0.0385246649,0.2603835166,-0.0615825541,0.041365657,-0.3012881577,-0.1222678572,0.1400857121,-0.3539153039,-0.4278773665,0.0277092066,0.1384716481,0.1929815859,0.2279366851,-0.3475384712,0.0305330995,0.4040951431,-0.0450155474,0.0737330914,0.1564313024,0.0955429971,-0.2031619549,-0.1827109158,-0.1231424436,0.0558167957,-0.0470121875,-0.1533836722,-0.1262530833]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"Ok, thanks for clarifying, that makes sense. I will time those examples later today and post back here.\r\n\r\nAlso, it seems that not all dumps should use the same date. For instance, I was checking the Spanish dump doing the following:\r\n```\r\ndata = nlp.load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner', split='train')\r\n```\r\n\r\nI got the error below because this URL does not exist: https:\/\/dumps.wikimedia.org\/eswiki\/20200501\/dumpstatus.json. So I checked the actual available dates here https:\/\/dumps.wikimedia.org\/eswiki\/ and there is no 20200501. If one tries for a date available in the link, then the nlp library does not allow such a request because is not in the list of expected datasets.\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.es (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.es\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 965, in _download_and_prepare\r\n super(BeamBasedBuilder, self)._download_and_prepare(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 422, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract({\"info\": info_url})\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/download_manager.py\", line 155, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 163, in map_nested\r\n return {\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 164, in <dictcomp>\r\n k: map_nested(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 191, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/download_manager.py\", line 156, in <lambda>\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/file_utils.py\", line 191, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/dumps.wikimedia.org\/eswiki\/20200501\/dumpstatus.json\r\n```","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":252,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nOk, thanks for clarifying, that makes sense. I will time those examples later today and post back here.\r\n\r\nAlso, it seems that not all dumps should use the same date. For instance, I was checking the Spanish dump doing the following:\r\n```\r\ndata = nlp.load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner', split='train')\r\n```\r\n\r\nI got the error below because this URL does not exist: https:\/\/dumps.wikimedia.org\/eswiki\/20200501\/dumpstatus.json. So I checked the actual available dates here https:\/\/dumps.wikimedia.org\/eswiki\/ and there is no 20200501. If one tries for a date available in the link, then the nlp library does not allow such a request because is not in the list of expected datasets.\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.es (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.es\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 965, in _download_and_prepare\r\n super(BeamBasedBuilder, self)._download_and_prepare(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 422, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract({\"info\": info_url})\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/download_manager.py\", line 155, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 163, in map_nested\r\n return {\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 164, in <dictcomp>\r\n k: map_nested(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 191, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/download_manager.py\", line 156, in <lambda>\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/file_utils.py\", line 191, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/gaguilar\/.conda\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/dumps.wikimedia.org\/eswiki\/20200501\/dumpstatus.json\r\n```","embeddings":[0.3074045777,-0.113163203,-0.1143275797,0.2666960657,0.0679331198,0.2159503251,0.1886423379,0.1882861257,0.6027131081,-0.2003641725,0.032980103,0.0883672684,0.1432431787,-0.2862560451,0.0379251018,-0.1857149005,0.0711626112,-0.0526695661,0.0999960229,-0.3232582808,-0.3721318245,0.2316435426,-0.3498652279,0.0395479575,-0.2949983478,0.3205533326,-0.1095639318,-0.1376003921,0.0048030014,-0.3297347724,0.1528253704,0.2533638775,0.3197317719,0.2253631353,-0.0001195317,-0.1390187144,0.7226005197,-0.140297249,-0.5811544657,-0.1576965749,-0.3304528296,-0.474478662,0.1427856833,-0.2372422367,0.1085572839,-0.1610044986,0.29019472,-0.372911334,0.2277098149,0.1040740609,0.1720932275,-0.0161261316,0.0532660335,0.0327017643,0.310792774,0.1675486565,0.1112300977,0.0960423201,0.0196501911,-0.2150571346,0.1037457734,0.0984232426,-0.1336673051,-0.1492940038,0.1253485233,-0.0472491719,-0.1610984057,-0.616353929,0.1744622141,0.2040951699,0.6709717512,0.0023055098,-0.4189354181,-0.1440745294,-0.022451736,0.0209414754,0.1702196002,0.3295253515,-0.0454614237,0.0398623869,-0.0027117806,-0.2038349658,-0.0404362828,0.7206988931,-0.0264509171,0.4018862247,0.1202324033,0.1144707501,-0.0478507318,-0.1456205547,-0.1257717609,-0.3335688412,0.4209320843,0.4332474768,-0.1695616692,0.2906306982,-0.0857509226,0.2263407409,-0.0529668033,-0.1909829229,-0.1185327768,0.2109177113,-0.0961131901,0.1913405508,0.3146859109,-0.1448129117,0.2309992313,-0.0167630315,0.3436278105,0.0785777792,-0.1617887169,-0.023676727,-0.2893207371,-0.2270640582,-0.2285067141,-0.0829101056,0.0350426435,-0.2333207279,0.2474828213,0.1096672639,-0.238432169,-0.2751595676,-0.1687636971,0.3228849769,-0.1710451692,0.2898728251,0.2068962008,0.3334832788,-0.4393334091,-0.098703973,0.0050612511,-0.0328752212,-0.2345456481,0.2383075505,0.2733896673,0.1644213349,0.3222702444,0.0639748201,-0.2206522822,-0.1693531275,-0.0159264226,-0.1536651552,-0.0792210624,0.0881713629,0.2245720029,0.5646492839,0.0242779851,-0.2164819539,-0.1830213815,0.1098443568,-0.1811427623,0.2454306185,-0.0224614739,0.0989767835,-0.3706935048,-0.053505253,-0.2598139346,0.1856046021,0.1031376943,-0.1530772299,-0.0132446084,-0.2585122287,-0.1633183956,-0.0153553355,0.2881461084,0.5672477484,-0.1877700835,-0.1831033081,-0.2109172642,0.1560409516,0.3122498691,0.2041439116,-0.1045734137,0.3262509406,-0.1384691745,0.0980197638,0.3381272256,-0.0435735025,0.0776512176,0.0450996086,0.2526730895,0.2135817856,-0.0926935524,-0.0261055101,0.1963848472,0.0178859457,0.1415536255,0.4359198213,0.3257624507,-0.0453435518,-0.374733448,-0.0472105034,0.3917998672,0.2488121539,0.1009588838,-0.1692661792,0.1102248505,0.5450981259,0.1534947455,0.0131801693,0.1428936273,0.4438541532,-0.3199238479,0.2711729407,0.2049239427,0.0090840152,-0.5281934142,0.2005945742,0.0112970527,0.1782017499,0.0175968874,0.052591607,-0.3461596668,-0.113277562,-0.201107353,0.2239731997,0.0632472038,0.2544429302,-0.0102858078,0.5279006958,0.0372149125,-0.3473691642,-0.000489692,-0.1019139811,-0.59104985,0.4307764769,-0.0491725318,-0.0058684056,0.0674812496,0.2370138764,0.3196126521,-0.1044643372,-0.0402181149,-0.223232165,0.3271289468,0.3333978951,0.187366575,-0.096820496,0.308155328,-0.3724738657,0.1896894872,0.3602362871,0.1580131799,-0.2589322925,0.0363305509,0.3114947677,-0.0588045307,0.3653423488,0.0718835965,0.0265556257,0.3908677995,-0.0897051767,-0.1538592875,-0.2190165222,0.475998491,0.1982602775,0.0277488716,0.3858515024,-0.1408335567,-0.1522122175,0.5553575158,-0.0403646119,-0.050155852,0.2276264429,-0.1491426229,-0.1826263368,0.0057154684,-0.2520778179,-0.1125383675,0.1228652298,0.0823420882,-0.3532349765,-0.0523493327,-0.249039501,0.0300016236,0.0911749154,0.0588251203,0.1508001536,0.1172384769,-0.0549026728,-0.3138919473,0.0181929115,0.0604091324,0.1874241233,-0.1832011491,0.0711913705,-0.4476852417,-0.3566983044,-0.2708434165,-0.3444032967,-0.3639487028,-0.5259601474,-0.0327019654,0.0293999556,-0.0621467941,0.2045760006,-0.1936113983,0.0379354395,-0.2525600791,-0.0846444443,-0.4149157703,-0.4787170887,-0.3776288033,-0.0720942989,0.5312838554,0.2609661222,0.0445938744,-0.112190865,-0.2606809735,-0.216250062,-0.3026230335,-0.1276061535,-0.1894437373,0.1767017096,-0.2689674199,0.4213791192,-0.0727718398,-0.2691791654,0.3827326298,-0.0300239883,0.0335943997,0.1892677993,0.1611525416,0.1392670125,-0.0332661197,-0.4139075875,-0.3171475828,-0.2624527812,-0.1804500818,0.0568670258,-0.009888825,0.1248526648,-0.0625321344,0.044401776,-0.0467396528,-0.028973205,-0.1634233147,0.2337409705,0.5412666202,-0.1276019514,-0.4130738974,0.1248721778,-0.102524437,0.0197148491,0.0213591382,-0.4710836112,0.1845600605,-0.1556138545,-0.0356668569,0.2318301201,-0.0825889111,-0.0668094903,-0.1799345613,0.0941499844,-0.0223419052,-0.106119521,-0.1568667442,-0.0057118949,0.2745938897,0.1639134735,0.3346804082,-0.3372355103,0.746999383,0.2176746428,0.0936311558,0.3313852549,0.2281756401,0.1476372182,-0.1205980852,-0.300983727,0.0003936335,0.0358179696,-0.1016240939,0.4037686884,0.1191242263,-0.4371238649,-0.1979615837,0.1191071793,-0.3471530974,-0.363963455,0.1728973538,-0.0917733163,0.1023441553,-0.0137912948,0.0246041585,-0.0789720491,-0.3898311257,-0.092544429,0.2616250217,0.3869504929,0.1271241754,-0.507219255,-0.3176085055,-0.3696794808,0.1878490001,-0.015675908,0.5833176374,-0.2090268582,0.0465389304,0.1606329679,0.3036927879,0.251665324,-0.3141660392,-0.0480247773,0.3816942871,-0.0033965562,-0.6274206638,0.0392707698,-0.1355001628,0.1793410778,0.4498906434,0.2873638868,-0.4655566514,-0.2426208854,-0.0170389377,0.5938124061,-0.1884110272,-0.1671379358,-0.3821137547,-0.2638597786,-0.3691657186,-0.3094538748,-0.0639619157,0.1736786962,-0.1798742265,-0.0858214945,-0.0537276417,0.2198533416,0.0736369863,0.0636089891,0.1167483702,0.2804917395,-0.0188217536,0.2255704701,-0.0031966686,0.0051286826,0.3203221262,-0.0955734923,-0.3046878874,-0.0183821246,-0.1032423005,0.252004385,0.1287858486,0.2227884233,0.0742222741,0.1264163107,-0.1507891566,-0.068966873,0.0607990138,0.2132912427,-0.0428382978,-0.3614948988,-0.4229004383,0.5458523035,0.0162435807,0.0175946094,0.3845964372,-0.0831711665,-0.4012175798,0.4881597459,-0.0436768904,1.00380373,-0.3012083769,0.1526129395,-0.0668854639,0.3107737601,0.4644118547,0.1293584406,-0.0028781679,-0.2208916098,0.297934413,-0.010266915,0.1015672684,0.0109867034,0.1999209672,-0.1830207556,0.3646388352,-0.0610989779,0.1480944604,0.070935905,0.5119069219,0.2190013975,-0.2706773877,-0.4574085772,0.0575842038,-0.1003017649,0.4405318201,-0.0420637392,0.051094912,0.0012199177,-0.0505814031,-0.2098079473,0.11462529,-0.0174278095,0.2365611047,-0.156231299,-0.147705391,0.0835242942,0.5369327068,-0.0238237809,0.2082618028,-0.1498301178,0.1094385758,-0.2827109396,-0.2716318071,-0.1630763412,0.1041024849,0.0436499752,0.0404535048,-0.4021268785,0.1497724652,-0.0398562141,-0.0782060027,-0.2885349989,0.1645684987,-0.1855711639,0.1062678173,-0.434903264,-0.0446732566,-0.1654358953,-0.0457042344,0.1009831205,0.0029672848,0.2235193104,0.1006203517,0.3073587418,-0.2967174351,-0.1264645606,0.307308197,-0.129716754,0.0496772565,0.7562810183,0.2978057265,-0.33279109,-0.1673114151,-0.1610654294,-0.4561047852,-0.4126517773,-0.2062857598,0.1005489975,0.3791306913,-0.2177530676,0.021458054,0.0231132954,-0.372857064,-0.0067505464,-0.6713995934,0.1130530983,0.3629444838,0.1078162342,0.0478669778,-0.0269311145,-0.3612900972,0.1439280808,0.0542994626,-0.1796963662,-0.0027685822,-0.2716397047,0.2428219169,0.1177801043,-0.0710611343,-0.1670647711,-0.1781167835,0.0492537022,-0.2977864146,0.1928614378,-0.1315274537,-0.0634015948,0.1811517775,0.2159472555,-0.129636541,-0.204641223,-0.376655072,0.4240162969,-0.3951970935,-0.2933062315,0.1064593047,-0.0215580482,0.3240570128,0.223929882,0.1515512168,-0.0383066535,0.1920634657,0.0011907161,0.1995295137,0.3148193657,0.0791809037,0.5858976841,0.1241865531,-0.3300997913,-0.0889594704,-0.107926026,0.0873936713,0.2444917411,-0.0481134504,0.321102351,0.4410486221,-0.0376987904,0.5469148159,-0.0492980555,-0.0506540164,0.3139183223,0.0774447247,-0.1749257296,0.0777029321,0.6115117073,0.0843544006,-0.1762633175,0.1499617249,0.1087770537,-0.3841848969,0.1678217202,0.0111341467,0.1982858479,-0.041244939,0.2548063695,0.0230463054,0.248361215,0.1717458665,-0.0276453681,0.0706359297,-0.0424115025,-0.1481902897,0.0432384498,0.1219860986,-0.0025727646,0.238838926,-0.2227486521,-0.3457959592,0.0852577314,0.4066886902,-0.0035728463,0.2401835918,-0.0093099903,0.1591867805,0.1050682515,0.0565769859,-0.2103706449,0.3899438679,0.0976409093,-0.2719575167,-0.2775010765,-0.0745846927,-0.1095886752,-0.0321725458,0.0246661659,-0.2972523272,0.0244527776,0.4213555753,-0.1656297594,-0.530297935,0.6537271738,-0.2060681134,0.1800951511,-0.1433719248,0.1685187668,-0.0142209409,-0.0609263442,0.1010964811,0.0816319808,0.1371220797,0.3256593347,-0.3377506733,0.0915213525,-0.0351753496,0.0738730058,-0.0911912918,0.5608187318,0.1278090924,0.1370948553,0.245810166,0.0882411674,-0.0985300243,-0.2725244761,0.2466867715,0.0393860973,-0.1093163639,0.1839756966,-0.1468201429,0.0436571911,-0.1639619917,-0.0617335923,-0.3652877212,0.1352901906,0.1807592362,0.0783859119,0.087484315,-0.1390055269,0.0064075883,-0.2053585202,0.5153891444,0.5274784565,0.1966546476,-0.2682808638,-0.2076094449,-0.6327536106,-0.0379971936,-0.4089773595,-0.076909408,0.2512984872,0.1556714624,0.0449994393,0.1244791672,0.0400832668,-0.1354501843,-0.1715441346,0.2276878059,-0.412297368,-0.1674693674,-0.0974029452,0.2109037042,-0.3606422544,-0.1823797971,0.238732338,-0.2613445818,-0.0877228007,-0.1229695603,-0.0824445933,0.0811957717,0.3395742178,0.1680944413,0.1144550815,0.5315930843,-0.0349051617,0.0782145411,-0.3346693516,-0.2836706042,0.1601029038,0.3635137081,-0.2720755339,0.1439711899,-0.16739133,0.0388079807,-0.2867976129,0.4059892595,-0.2051696777,-0.0799489021,0.0568463989,-0.2158320844,-0.0427997559,-0.137801379,0.092725426,0.2488564551,-0.1494413018,0.1475373358,-0.1498866528,-0.1672076732,0.1603717357,-0.4102844894,-0.3715366721,-0.0028676393,0.0432940014,0.0619986095,0.2191030681,-0.3153202832,0.0515870862,0.4503411651,-0.0099273054,0.1241251156,0.1163679585,0.056241259,-0.1464502513,-0.1525178552,0.0188840646,0.1500613391,-0.0950559452,-0.0833321065,-0.1693374813]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"Thanks ! This will be very helpful.\r\n\r\nAbout the date issue, I think it's possible to use another date with\r\n\r\n```python\r\nload_dataset(\"wikipedia\", language=\"es\", date=\"...\", beam_runner=\"...\")\r\n```\r\n\r\nHowever we've not processed wikipedia dumps for other dates than 20200501 (yet ?)\r\n\r\nOne more thing that is specific to 20200501.es: it was available once but the `mwparserfromhell` was not able to parse it for some reason, so we didn't manage to get a processed version of 20200501.es (see #321 )","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":77,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nThanks ! This will be very helpful.\r\n\r\nAbout the date issue, I think it's possible to use another date with\r\n\r\n```python\r\nload_dataset(\"wikipedia\", language=\"es\", date=\"...\", beam_runner=\"...\")\r\n```\r\n\r\nHowever we've not processed wikipedia dumps for other dates than 20200501 (yet ?)\r\n\r\nOne more thing that is specific to 20200501.es: it was available once but the `mwparserfromhell` was not able to parse it for some reason, so we didn't manage to get a processed version of 20200501.es (see #321 )","embeddings":[0.1844169199,-0.1612697095,-0.14086321,0.3425940871,0.0797085017,0.2668132186,0.1986873895,0.1461935341,0.7380709052,-0.2608357668,0.117104128,0.1079088822,0.0482873581,-0.1868229806,0.0673659146,-0.2168420255,0.1130851358,-0.1392923295,0.0121253384,-0.2986245155,-0.375538826,0.2062976807,-0.3456492424,-0.0187360477,-0.2716134489,0.2526730895,-0.0286885835,-0.1416581571,0.0758292824,-0.3684397042,0.1234105751,0.2239328474,0.3555548787,0.1723899692,-0.0001185649,-0.1505396366,0.660580039,-0.1414551139,-0.5648554564,-0.101155445,-0.1321353316,-0.4575914145,0.2321635932,-0.1510665417,0.015325794,-0.2116499543,0.2019940317,-0.495457083,0.1956826001,0.1200146303,0.1927446127,-0.093378976,0.0861300603,0.0984007418,0.2837163806,0.0862573832,0.1070939824,0.091626972,0.0906863883,-0.2530971169,0.0317131244,0.1068276763,-0.1099942774,-0.1209026426,0.1942079663,-0.0663056523,-0.0168333966,-0.6403782964,0.1275741011,0.1657272279,0.7213552594,0.0507114045,-0.3950336277,-0.223338291,-0.0087345932,0.0409741402,0.167893976,0.2475132942,-0.0853280202,-0.0411406644,0.0339016505,-0.2376619428,-0.0329024345,0.6365787983,-0.0692424849,0.4427610636,0.1174650714,0.1288728416,0.0202501975,-0.253596276,-0.1812103689,-0.2587616444,0.4181684554,0.4655200839,-0.174560383,0.3636753559,-0.0123904804,0.3851032257,-0.038643036,-0.15390791,-0.2357905805,0.2513465881,-0.1333470494,0.1615580171,0.2684187889,-0.1073283479,0.2236352265,-0.1620290279,0.3467641175,0.0117434561,-0.1397831887,0.0552117676,-0.2117568851,-0.2181311399,-0.1720184833,-0.0451148674,-0.0281924997,-0.2950689793,0.2572841644,0.1664818823,-0.2195028663,-0.3274429142,-0.1749036163,0.2692579329,-0.1951353103,0.2675597966,0.1724327058,0.2840561867,-0.4226954281,-0.1039365456,0.0161602814,-0.0995109975,-0.257117182,0.2204846144,0.3558721244,0.1486310512,0.2602377832,0.0523075499,-0.0767712891,-0.1491639316,-0.0647039264,-0.163499251,-0.0249815471,0.0091658114,0.1766334772,0.5606009364,0.0988832936,-0.2323977351,-0.1752841026,0.1039439142,-0.0761017725,0.3082154691,0.0369334295,0.1233540773,-0.3571022153,-0.104668878,-0.298697412,0.2197835743,0.1202819943,-0.200995326,-0.0416015312,-0.3253329098,-0.1707216948,-0.0548555851,0.3113587797,0.5726601481,-0.2665934861,-0.1806461811,-0.1147188693,0.1087629795,0.207697615,0.2036616057,-0.0525589585,0.257996738,-0.1294812411,-0.0633769482,0.2858005166,-0.0579142123,0.10676568,-0.0260217004,0.2796825767,0.3171775043,-0.1271193177,0.0283081699,0.1985153556,0.0353313871,0.0832561329,0.4870526791,0.2065672874,-0.0626081824,-0.3746045232,-0.0362061933,0.5943416953,0.3290396333,0.1499209255,-0.1518611312,0.0305371415,0.3674328327,0.1763261408,0.0274736993,0.1374637932,0.3568587005,-0.3556275666,0.3041117489,0.2620894611,-0.0356626809,-0.5659899116,0.2383675426,0.1369163841,0.1958183497,0.0540325716,0.0960886702,-0.2962388992,-0.1357355118,-0.2376848906,0.2135418653,0.0795716047,0.1882343739,-0.0467985235,0.5260180235,0.0267842729,-0.0761934221,0.0167776328,-0.1406954825,-0.5490181446,0.3648965657,-0.0192004647,-0.078961879,0.0225067642,0.2183268368,0.329241395,-0.1560719758,-0.0868927315,-0.2810145617,0.3929997683,0.2565650642,0.1176106855,-0.064565137,0.2742711604,-0.3607213199,0.2317126989,0.3237192631,0.1243609935,-0.2508164942,-0.0573165268,0.3349547386,-0.1451161206,0.4635535181,0.035760574,0.0688562021,0.3251148462,0.0063651716,-0.1108129844,-0.1346651912,0.5287485123,0.1029547378,0.0637287572,0.2845212817,-0.1806773543,-0.0947574303,0.5067564249,0.0490307026,-0.0480579324,0.2475455403,-0.2007910758,-0.1672690511,-0.0358605571,-0.2484871894,-0.182742551,0.132170409,0.0703616589,-0.3738316596,0.0458124578,-0.1439269185,-0.0153811239,0.1464487612,0.1807280481,0.0824580863,0.1403475404,-0.000978288,-0.2181272358,0.0029938754,0.0622046776,0.1825273335,-0.1979463845,0.0681489632,-0.3990391195,-0.3739188015,-0.2093369812,-0.2234026939,-0.4862132072,-0.4510011077,-0.1158518568,0.0748049617,0.0489552431,0.1520549506,-0.0781797916,0.035830643,-0.244780466,-0.091454424,-0.4282477498,-0.5143586397,-0.4257921278,-0.025491789,0.540037334,0.1605172604,0.0958300158,-0.0363360979,-0.265871197,-0.1976131648,-0.2801333666,-0.1823983938,-0.2643185556,0.2399842441,-0.3431635499,0.3949019313,-0.0874270499,-0.0930264741,0.3740706444,-0.0331536047,0.0601067021,0.1564411521,0.1854311973,0.2013899237,-0.0076232972,-0.4389497936,-0.369666338,-0.2770028412,-0.0789736286,0.1606952548,-0.058080636,0.0915826932,-0.1584427953,-0.0235236343,0.020878952,0.0305209514,-0.1258708239,0.2937546968,0.4754168093,-0.0671438798,-0.4101143777,0.0925857797,0.0022497803,0.0188389048,0.037371181,-0.4818988144,0.110514991,-0.1167071015,0.0338356383,0.1245819107,-0.124604784,-0.0315149836,-0.1618935913,0.118665278,0.0365787409,-0.1993729621,-0.1834090501,-0.0256385431,0.1785925329,0.0766231343,0.3736392856,-0.3757041097,0.6833336353,0.191984728,0.0517833717,0.2668404877,0.2044528127,0.0208218321,-0.1411595494,-0.2687708139,0.0289678834,-0.0435255691,-0.0692294836,0.3616577685,0.0561509505,-0.447046876,-0.2149753571,0.1601336151,-0.3118796647,-0.3062430024,0.1863856018,-0.2279086411,0.1966601759,-0.0439397506,0.0396871828,-0.078578718,-0.4171133339,-0.0730284378,0.2312418222,0.420150727,0.1305899322,-0.3580099344,-0.3635906279,-0.2295687795,0.147547096,0.0181840267,0.5453318357,-0.3247435093,-0.025975652,0.1364896894,0.1869921982,0.2491882592,-0.3277454376,-0.0784446225,0.5329350829,0.0247936212,-0.6945915222,-0.0137459598,-0.1425858289,0.1327881217,0.3927876651,0.1861748099,-0.4104098082,-0.3094126582,-0.0620852597,0.6232078075,-0.2286628783,-0.1218044609,-0.3604113162,-0.0846646056,-0.3967116475,-0.2190544605,-0.146990478,0.145607233,-0.2334452868,0.0663505867,-0.0082136532,0.2877655029,0.1197932288,0.1248495206,0.1368210018,0.3288369179,0.0264689922,0.1864501387,0.0177560356,-0.0739143342,0.2431001514,-0.0446397513,-0.2214606255,-0.0219967254,-0.1734540164,0.2541014254,0.0243116394,0.16866754,0.0354034305,0.0936079323,-0.1219545454,-0.1373637319,0.0995345041,0.2646071613,-0.042721305,-0.3857752979,-0.4617741406,0.5700258613,0.0440340377,-0.0771908164,0.3294095397,-0.0840583295,-0.4005016983,0.5531386137,-0.0370687656,0.8670908809,-0.3276717365,0.1661259383,-0.1091105342,0.2213766128,0.5047303438,0.0582498051,-0.0689954609,-0.2617022991,0.3780922294,-0.0033926887,0.0951020792,-0.0139448009,0.138204366,-0.1986902058,0.3340370953,-0.1274997741,0.0509022139,0.0275454968,0.5285962224,0.2370521426,-0.3180035353,-0.3905282617,0.078416422,-0.1627967805,0.3956042528,-0.0741905943,0.1031677127,-0.0298173074,0.0155839538,-0.2829713821,0.1142401695,-0.0419712216,0.2151598483,-0.1797875762,-0.2497597337,0.2762182951,0.4374958873,0.0755615011,0.1677909344,-0.1664403081,0.1357033849,-0.2448333949,-0.2289774716,-0.0646029413,0.0835580826,0.1101592034,0.028770417,-0.3292125463,0.0522801317,-0.0567030758,-0.0233200677,-0.3895906508,0.1673524827,-0.0041857511,0.0608698651,-0.3358550966,0.0115341768,-0.1193821877,-0.0414843336,0.1193093285,0.0352866501,0.2250262648,0.2422863543,0.3342757821,-0.2829225659,-0.113866806,0.3032532334,-0.0697611421,-0.0558016859,0.7682930231,0.4745148122,-0.3394963741,-0.2241128087,-0.2571645677,-0.438836813,-0.4500105679,-0.2344160527,0.1606298685,0.4009587765,-0.1865049303,0.0555912293,-0.0034108192,-0.4493925869,-0.0002202857,-0.6345475912,0.034030132,0.3447431922,0.1035904288,0.0054722549,0.0012738223,-0.3435917199,0.1298700571,0.0612002499,-0.1972837299,-0.0626491457,-0.1562690884,0.2272823453,0.3119395673,-0.0378487371,-0.1576277763,-0.1320439726,0.0860211477,-0.2972199917,0.1732495725,-0.1805264205,-0.0063105435,0.1574811041,0.1737737209,-0.1680445373,-0.238455981,-0.3589776754,0.4282383919,-0.3507373631,-0.1986349225,-0.0201145709,-0.0225042384,0.3182732463,0.0657492653,0.2486482114,-0.0504165031,0.1881613284,0.0012930608,0.1054658219,0.320145905,0.0821959451,0.5838451385,0.1199807897,-0.414164573,-0.068112731,-0.0772438869,0.0852589607,0.2501686811,-0.165885672,0.325609535,0.3960368037,-0.0062784785,0.5301900506,-0.0131567996,-0.0744047612,0.2766919136,0.1095913723,-0.1725284457,0.082023181,0.7015462518,0.1404868662,-0.1471320838,0.1354475319,0.1613201499,-0.2493498176,0.1213673204,-0.0278707966,0.2471284121,-0.099020496,0.2043269426,-0.0573814772,0.2665820718,0.1177336276,-0.033000011,0.1238400415,-0.1190438122,-0.0065648393,0.1209407076,0.1066178828,-0.0044656801,0.1893348098,-0.1750041842,-0.3500475585,0.1187172085,0.3595323265,0.0404182971,0.2055849284,-0.0243211854,0.2554510236,0.1185597181,-0.1806191355,-0.1621672511,0.2564994395,0.0781321451,-0.2098520696,-0.1287288964,-0.0651434734,-0.1099868342,-0.0071921428,-0.0591130555,-0.2668320239,0.0281604454,0.4564834535,-0.0903670564,-0.550486207,0.4089471698,-0.1690150797,0.2195013463,-0.0892808065,0.1430746764,0.0180570558,-0.07557486,-0.018593628,0.0626493543,0.1614682972,0.2659322321,-0.4696862698,0.1626980603,-0.0513631888,0.0850127414,-0.0890451744,0.4808033705,0.103676267,0.297078371,0.2520151138,0.1200079992,-0.0791587085,-0.2093758732,0.2073232979,-0.0279633533,-0.0967573449,0.0871886685,-0.0559912845,0.0332673676,-0.2572042644,-0.0831968635,-0.3097164035,0.150976792,0.2160118818,0.2096531987,0.0658720061,-0.0658787265,0.016535107,-0.145769164,0.5989708304,0.5052396059,0.3864729404,-0.2828418314,-0.2142058611,-0.6079564691,-0.0683486909,-0.3300475478,-0.0620986968,0.2401628941,0.1515762806,0.1486020386,0.0755125806,0.0893741176,-0.1815302372,-0.127714619,0.2712645531,-0.3947217166,-0.1702540368,-0.1462319642,0.165957734,-0.3174644411,-0.0818439722,0.1850228757,-0.2791994214,-0.0646481812,-0.1395654529,-0.1430629492,0.0477134436,0.287676096,0.2277488858,0.0640403479,0.6330654025,0.0302951075,0.0548836514,-0.3189073801,-0.2455743253,0.1151759252,0.3448651731,-0.3442027569,0.2105502933,-0.177417621,0.0714088902,-0.2868709266,0.3694992661,-0.1817532033,-0.0142912976,-0.0039675795,-0.2246591151,0.0378588326,-0.1431713402,0.0891864821,0.2404401302,-0.1001728624,0.1090268269,-0.2864095569,-0.1072931886,0.1480803937,-0.3834710419,-0.4022498131,-0.0258693378,0.0717194602,0.0885525048,0.180499658,-0.386585474,0.01289263,0.4387580454,-0.0302377734,0.1058728844,0.1398379952,0.1077253222,-0.1861807555,-0.1685120314,-0.1288342029,0.0912464485,-0.0665118545,-0.1031779647,-0.1249067634]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"Cool! Thanks for the trick regarding different dates!\r\n\r\nI checked the download\/processing time for retrieving the Arabic Wikipedia dump, and it took about 3.2 hours. I think that this may be a bit impractical when it comes to working with multiple languages (although I understand that storing those datasets in your Google storage may not be very appealing either). \r\n\r\nFor the record, here's what I did:\r\n```python\r\nimport nlp\r\nimport time\r\n\r\ndef timeit(filename):\r\n elapsed = time.time()\r\n data = nlp.load_dataset('wikipedia', filename, beam_runner='DirectRunner', split='train')\r\n elapsed = time.time() - elapsed\r\n print(f\"Loading the '{filename}' data took {elapsed:,.1f} seconds...\")\r\n return data\r\n\r\ndata = timeit('20200501.ar')\r\n```\r\n\r\nHere's the output:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 13.0k\/13.0k [00:00<00:00, 8.34MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 28.7k\/28.7k [00:00<00:00, 954kB\/s]\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguil20\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 47.4k\/47.4k [00:00<00:00, 1.40MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 79.8M\/79.8M [00:15<00:00, 5.13MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 171M\/171M [00:33<00:00, 5.13MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 103M\/103M [00:20<00:00, 5.14MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 227M\/227M [00:44<00:00, 5.06MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 140M\/140M [00:28<00:00, 4.96MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 160M\/160M [00:30<00:00, 5.20MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 97.5M\/97.5M [00:19<00:00, 5.06MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 222M\/222M [00:42<00:00, 5.21MB\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [03:16<00:00, 196.39s\/sources]\r\nDataset wikipedia downloaded and prepared to \/home\/gaguil20\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50. Subsequent calls will reuse this data.\r\nLoading the '20200501.ar' data took 11,582.7 seconds...\r\n````","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":202,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nCool! Thanks for the trick regarding different dates!\r\n\r\nI checked the download\/processing time for retrieving the Arabic Wikipedia dump, and it took about 3.2 hours. I think that this may be a bit impractical when it comes to working with multiple languages (although I understand that storing those datasets in your Google storage may not be very appealing either). \r\n\r\nFor the record, here's what I did:\r\n```python\r\nimport nlp\r\nimport time\r\n\r\ndef timeit(filename):\r\n elapsed = time.time()\r\n data = nlp.load_dataset('wikipedia', filename, beam_runner='DirectRunner', split='train')\r\n elapsed = time.time() - elapsed\r\n print(f\"Loading the '{filename}' data took {elapsed:,.1f} seconds...\")\r\n return data\r\n\r\ndata = timeit('20200501.ar')\r\n```\r\n\r\nHere's the output:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 13.0k\/13.0k [00:00<00:00, 8.34MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 28.7k\/28.7k [00:00<00:00, 954kB\/s]\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguil20\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 47.4k\/47.4k [00:00<00:00, 1.40MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 79.8M\/79.8M [00:15<00:00, 5.13MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 171M\/171M [00:33<00:00, 5.13MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 103M\/103M [00:20<00:00, 5.14MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 227M\/227M [00:44<00:00, 5.06MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 140M\/140M [00:28<00:00, 4.96MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 160M\/160M [00:30<00:00, 5.20MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 97.5M\/97.5M [00:19<00:00, 5.06MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 222M\/222M [00:42<00:00, 5.21MB\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [03:16<00:00, 196.39s\/sources]\r\nDataset wikipedia downloaded and prepared to \/home\/gaguil20\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50. Subsequent calls will reuse this data.\r\nLoading the '20200501.ar' data took 11,582.7 seconds...\r\n````","embeddings":[0.2451880127,-0.1499561667,-0.1770797819,0.4052183628,0.0632281676,0.2607442439,0.2458155751,0.1112545356,0.6888960004,-0.2664470673,-0.0213993397,0.0063388622,0.0856510848,-0.2514183521,0.0070183449,-0.1802873462,0.0203865115,-0.1383438259,0.1124566123,-0.2367403209,-0.2798018456,0.2774860263,-0.2395900041,-0.0487777628,-0.1759125739,0.2128080875,-0.0351585783,-0.2206657529,0.1427014172,-0.3392617404,0.1575160921,0.2800291777,0.2215889841,0.1722692549,-0.0001170091,-0.0706660226,0.6258941889,-0.0953066498,-0.4801300168,-0.1560158134,-0.108958818,-0.4129000306,0.1846741289,-0.1244983152,0.0033279601,-0.2184667885,0.2557478845,-0.5376159549,0.1487024873,0.1018920243,0.1965543479,-0.1393259913,-0.0304985978,0.1324132234,0.2393972725,0.0642504022,0.151079759,0.1061608791,0.1205729097,-0.2250445187,-0.0113991546,0.148513779,-0.1639057994,-0.0380540639,0.1549713314,-0.0637798086,-0.1280984879,-0.6488245726,0.1541220397,0.1783592552,0.6348540187,0.0936589912,-0.3493359387,-0.2358499914,0.0220760126,0.0087388158,0.1717161685,0.3690199256,-0.0816911906,-0.0795899257,-0.0399339795,-0.1963656247,-0.0322664864,0.5931959152,-0.0880079716,0.3404202163,0.1230519637,0.1242888495,0.0851862431,-0.1302557439,-0.2010220289,-0.3093886375,0.4458721876,0.3907281458,-0.2508179545,0.4231896996,0.0125241587,0.2542313933,-0.0182336401,-0.1473664641,-0.1184965745,0.2253141105,-0.1155808792,0.1374964118,0.1883317232,-0.124688454,0.1700281799,-0.1154406294,0.3546131849,0.0449151173,-0.1789989918,0.0283960514,-0.1855123341,-0.1656094342,-0.2223558724,-0.0903871953,-0.0302750859,-0.2254109085,0.1415824294,0.1084441096,-0.3119995892,-0.3404396474,-0.1642335355,0.2511446178,-0.1807250381,0.2364733964,0.1775927693,0.3219090402,-0.4573749602,-0.0067457012,0.0449120812,-0.0690974742,-0.2206031978,0.2642526329,0.4220621884,0.0948154628,0.2164129764,0.0365128182,-0.0398966447,-0.0909143984,0.0710607097,-0.2667188942,-0.0585894696,-0.0462982021,0.2065355182,0.5741945505,0.0112998402,-0.2553770244,-0.2352945656,0.070635505,-0.0643666163,0.3766923547,0.0205886737,0.1219099164,-0.2673357427,-0.1256442368,-0.2556728423,0.3742669821,0.1674688905,-0.1810794771,-0.0655050129,-0.2110247314,-0.2137752175,-0.0543820895,0.2509810328,0.5896446109,-0.2892338634,-0.1646264493,-0.1988548636,0.2035034448,0.1886523813,0.2982170284,-0.0958646461,0.2337156236,-0.1778505743,0.0835026726,0.2486936152,-0.0395822115,-0.0014971978,-0.0257272013,0.2922070622,0.3353715837,-0.007831594,0.0580954328,0.2001075298,0.0750987157,0.1405767798,0.621952951,0.2040380388,-0.0073912088,-0.3208083808,-0.050289724,0.584579587,0.3552081585,0.1517485082,-0.2365605384,0.0374267027,0.2481883466,0.1562710702,-0.0131509518,0.0936758518,0.3756870329,-0.4390864372,0.2866098881,0.1611397713,0.0143171204,-0.6067688465,0.2630679607,0.0607769564,0.1707911491,-0.031874612,0.1079960689,-0.2593958378,-0.2113682032,-0.2820684314,0.1819368005,0.0974049121,0.2778010368,-0.0736682266,0.5513207316,0.1415987462,-0.0580300502,0.1464496255,-0.1990238428,-0.6403774619,0.4351526201,-0.02461222,-0.0105862133,0.075666368,0.1637992114,0.2957056165,-0.1776614934,-0.0669073537,-0.3276809156,0.2493428886,0.3236111999,0.225586772,-0.0355046913,0.3311399817,-0.3649567068,0.2097979635,0.2917325497,0.1693488955,-0.2859678268,-0.0535754189,0.312384963,-0.0520770177,0.5205019712,0.0842649043,0.0086803883,0.3458456397,-0.0037226875,-0.1675105989,-0.0433898382,0.5325737596,0.0772654563,0.0950237438,0.2842740715,-0.2170914263,-0.0199629348,0.5617513657,0.1191320047,-0.0381754115,0.1991221607,-0.1175560355,-0.1439589709,-0.0568963364,-0.2177955955,-0.1218238845,0.1635314077,0.111062631,-0.4647266865,-0.0015269194,-0.1823050082,-0.0400638804,0.0966213271,0.1555220783,0.1344235539,0.1967048496,-0.0348842032,-0.2852691412,-0.1075898558,0.143174246,0.1040027365,-0.0965625718,0.1613909155,-0.403183341,-0.3990436196,-0.2320873141,-0.3176559508,-0.4566809237,-0.4382843375,-0.1172616631,0.0020064681,0.0365824588,0.1779958308,-0.0860112309,-0.0040276083,-0.214853406,-0.1572204083,-0.3402951956,-0.4157637656,-0.3499432504,-0.0113514634,0.6082346439,0.1408375353,0.1521293223,-0.0444762409,-0.3557099104,-0.1870456636,-0.2607334256,-0.2199682742,-0.2619196177,0.1499442905,-0.3553150892,0.335590601,-0.1427180022,-0.1477808505,0.3654433787,0.0682374239,0.0449981689,0.1869262755,0.2114941031,0.2520879209,0.0519438162,-0.4182102084,-0.2767595053,-0.297721535,-0.0779388696,0.0720985383,-0.0323851146,0.0740771145,-0.1342922002,0.0277838763,0.0462004617,0.0540389381,-0.1266372651,0.3575998545,0.498567462,-0.0664858893,-0.3655041754,0.0541130602,0.0312132072,0.0838306397,0.0972238481,-0.4861781597,0.1801841259,-0.2643375993,0.0509781353,0.1153757274,-0.1262599528,-0.0208539218,-0.2068902999,0.0848828331,0.0921390206,-0.2170365453,-0.1743117273,-0.0443453491,0.1687156707,0.0408560038,0.3214994371,-0.3043214083,0.7088569403,0.1946599185,0.0952943116,0.2897074819,0.2598703802,0.0279643163,-0.1108570918,-0.3611119986,0.0237769596,-0.0480476879,-0.0266053341,0.3631151021,-0.0062626242,-0.5170608163,-0.1891023666,0.1743405908,-0.3713749349,-0.2986271083,0.2226099819,-0.3242891133,0.1422244161,-0.0391420349,-0.0105264932,0.0027756859,-0.3711197972,-0.1053499728,0.1779324412,0.3246054947,0.1485723257,-0.3348633647,-0.3951420486,-0.3076943457,0.2421050966,0.0071049645,0.4984691143,-0.2654772103,0.0471523143,0.1594441831,0.1530167013,0.2365149409,-0.2597925961,-0.1111958921,0.5279589891,-0.0271047242,-0.6189777851,-0.0526653528,-0.1377344131,0.1269148588,0.3149394095,0.2196885049,-0.4838066101,-0.288721323,-0.1226008087,0.5876404047,-0.1701470017,-0.0847838297,-0.4336578548,-0.2338644117,-0.3693840504,-0.1450404525,-0.1572351903,0.1778553575,-0.2349481881,-0.045978114,-0.0219067577,0.2416180074,0.1035632119,0.0267473385,0.131609723,0.2954697311,0.0465458483,0.1607414633,0.0174135957,-0.1359119266,0.3154527247,-0.0129849585,-0.2080694735,0.0275538582,-0.1306412965,0.1875984818,0.0917407274,0.245194599,0.1307401508,0.1095636636,-0.1335561126,-0.048985146,0.0352487229,0.2923878729,-0.0489346869,-0.3870002031,-0.4622569084,0.534645617,0.0381880365,-0.0413823426,0.337988019,-0.1091617346,-0.3768290877,0.5159648061,-0.0695693269,0.850066781,-0.3642225266,0.2199822962,-0.1245952994,0.172277078,0.4675637484,0.0763167888,0.0761720613,-0.2263677567,0.281131655,0.0344469883,0.0823061988,-0.047271207,0.128032133,-0.1903810054,0.3003567159,-0.1230533868,-0.0278718304,-0.0277450588,0.5177565813,0.1137029454,-0.3444422185,-0.3903771341,0.0782159194,-0.1911295205,0.3340757489,-0.0570767075,0.1440505534,-0.0817054138,0.044823017,-0.2575119138,0.1331407875,-0.0768298134,0.2991917729,-0.2658872008,-0.254170835,0.2688462734,0.4031136334,0.0046637477,0.1715748757,-0.1435803473,0.1161303669,-0.2655657828,-0.2525236905,-0.0746075734,-0.1055534482,0.0396233983,0.0475947298,-0.3035375476,0.0346583053,0.0085531212,-0.0259419251,-0.4309224784,0.2589807808,-0.0142876813,0.1846861541,-0.3098225296,0.020759318,-0.1681016833,0.0122457668,0.1065094322,0.0701263472,0.1924309731,0.214222461,0.2856585979,-0.269775629,-0.0905393437,0.3106727898,0.0634701699,-0.0143201249,0.750733912,0.4354947805,-0.3659291267,-0.2016963214,-0.2751033306,-0.3988817334,-0.4118560255,-0.1979347467,0.0786666572,0.4254753292,-0.234116897,-0.0405565463,-0.0630697682,-0.4103242755,0.0456098244,-0.665674746,0.1718959957,0.3561232388,0.0507008322,0.0035444102,-0.0290383361,-0.4292141199,0.1859495938,0.1663058251,-0.1968049109,-0.0389883742,-0.2543281913,0.2168639749,0.2401294708,-0.1219029725,-0.109031409,-0.0833920464,0.0853056833,-0.2562125325,0.1659389287,-0.1899263561,0.0360816456,0.1547968835,0.1552375257,-0.0410484225,-0.1149520427,-0.44504264,0.4579294622,-0.3163775206,-0.1755691767,0.0163528584,-0.0156325623,0.2527423799,0.0262188036,0.1651684791,-0.0539301299,0.1831568331,-0.1122229025,-0.02502493,0.3604780436,0.0616053827,0.6041914225,0.1429425329,-0.4622181058,-0.0460664369,-0.0973878801,0.1787067503,0.1860468388,-0.12291421,0.3573132753,0.3914485276,0.0029296197,0.5594460368,-0.0614701435,-0.1044644862,0.2407747507,0.0994288474,-0.175363645,0.0176679492,0.6052831411,0.027210936,-0.0985855609,0.1960463971,0.1467482299,-0.2734449804,0.0870824233,-0.0509235673,0.1829746068,-0.0898048431,0.1844636053,-0.0684339926,0.2767162025,0.079066284,0.0812537596,0.0991478488,-0.1257116944,-0.0263299048,0.0368317999,0.1100983545,-0.0732965693,0.142131567,-0.1959451288,-0.3418853581,0.1488083303,0.3849465549,-0.0385040715,0.1966334581,-0.0197482686,0.2618835568,0.0570272207,-0.1767059565,-0.1262668818,0.2815194726,0.0860372633,-0.1537542492,-0.1103939787,-0.0374433175,-0.1708141863,0.0302489288,-0.0098094968,-0.1767525673,0.1158330962,0.4291350842,-0.143288359,-0.5412380695,0.5037273169,-0.1810054034,0.1998375952,-0.1252733916,0.1471012384,0.0336151049,-0.0736716613,-0.0792783126,0.1109912917,0.0890679434,0.2411633581,-0.4174546003,0.1338170171,-0.0459812582,0.1110811383,-0.0803475752,0.4521905184,0.1720502675,0.2157218158,0.297624588,0.1016492993,-0.1120321304,-0.1626937836,0.2548759282,-0.0762876198,-0.174043864,0.0921575949,-0.1568003893,0.0316732749,-0.2665814459,-0.0060913172,-0.3949311376,0.1219009086,0.154734984,0.130067125,0.0728919208,-0.0983196795,0.0153386127,-0.0937213823,0.6074228883,0.4792847037,0.3430700302,-0.2939924002,-0.1760926247,-0.6380918622,-0.0678525567,-0.3921573758,0.0492006205,0.2024064064,0.3104743958,0.1381627023,0.0823643208,0.1153787225,-0.1095290557,-0.1194224581,0.2652045786,-0.4071400166,-0.1527133584,-0.1296408623,0.0743366256,-0.3396193385,-0.1283410788,0.2265630513,-0.2407397926,-0.0277350489,-0.1027848795,-0.1051416323,0.072988607,0.2741319239,0.20153144,0.0922099352,0.5899716616,0.0433661081,0.0089176092,-0.2871464193,-0.1425205469,0.104171738,0.3150423765,-0.2458575368,0.251447618,-0.1794993877,0.0906048119,-0.2311390042,0.4558018148,-0.1921847761,0.048700612,0.0483478233,-0.232985884,0.032991454,-0.0703178346,0.028270619,0.2565557659,-0.0535291843,0.0707786754,-0.3227969706,-0.1854753047,0.073579669,-0.336276114,-0.4813310206,0.0286505036,0.1353667378,0.1401510239,0.2730078101,-0.2980144322,0.0082408777,0.4361495674,-0.0367801748,0.0740418211,0.2029283643,0.1121998429,-0.1739456654,-0.1829123199,-0.1639410257,0.1568365842,-0.0951531455,-0.1418355852,-0.1823761165]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"> About the date issue, I think it's possible to use another date with\r\n> ```python\r\n> load_dataset(\"wikipedia\", language=\"es\", date=\"...\", beam_runner=\"...\")\r\n> ```\r\n\r\nI tried your suggestion about the date and the function does not accept the language and date keywords. I tried both on `nlp` v0.4 and the new `datasets` library (v1.0.2):\r\n```\r\nload_dataset(\"wikipedia\", language=\"es\", date=\"20200601\", beam_runner='DirectRunner', split='train')\r\n```\r\nFor now, my quick workaround to keep things moving was to simply change the date inside the library at this line: [https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wikipedia\/wikipedia.py#L403](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wikipedia\/wikipedia.py#L403)\r\n\r\nNote that the date and languages are valid: [https:\/\/dumps.wikimedia.org\/eswiki\/20200601\/dumpstatus.json](https:\/\/dumps.wikimedia.org\/eswiki\/20200601\/dumpstatus.json)\r\n\r\nAny suggestion is welcome :) @lhoestq \r\n\r\n\r\n## **[UPDATE]**\r\n\r\nThe workaround I mentioned fetched the data, but then I faced another issue (even the log says to report this as bug):\r\n```\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\n```\r\n\r\nHere's the full stack (which says that there is a key error caused by this key: `KeyError: '000nbsp'`):\r\n\r\n```Downloading and preparing dataset wikipedia\/20200601.es (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gustavoag\/.cache\/huggingface\/datasets\/wikipedia\/20200601.es\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 74.7k\/74.7k [00:00<00:00, 1.53MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 232M\/232M [00:48<00:00, 4.75MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 442M\/442M [01:39<00:00, 4.44MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 173M\/173M [00:33<00:00, 5.12MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 344M\/344M [01:14<00:00, 4.59MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 541M\/541M [01:59<00:00, 4.52MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 476M\/476M [01:31<00:00, 5.18MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 545M\/545M [02:02<00:00, 4.46MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 299M\/299M [01:01<00:00, 4.89MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 9.60M\/9.60M [00:01<00:00, 4.84MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 423M\/423M [01:36<00:00, 4.38MB\/s]\r\nWARNING:apache_beam.options.pipeline_options:Discarding unparseable args: ['--lang', 'es', '--date', '20200601', '--tokenizer', 'bert-base-multilingual-cased', '--cache', 'train', 'valid', '--max_dataset_length', '200000', '10000']\r\n\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\nTraceback (most recent call last):\r\n File \"apache_beam\/runners\/common.py\", line 961, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 553, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1095, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 500, in _clean_content\r\n text = _parse_and_clean_wikicode(raw_content, parser=mwparserfromhell)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 556, in _parse_and_clean_wikicode\r\n section_text.append(section.strip_code().strip())\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/wikicode.py\", line 643, in strip_code\r\n stripped = node.__strip__(**kwargs)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 63, in __strip__\r\n return self.normalize()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 178, in normalize\r\n return chrfunc(htmlentities.name2codepoint[self.value])\r\nKeyError: '000nbsp'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/runpy.py\", line 194, in _run_module_as_main\r\n return _run_code(code, main_globals, None,\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/runpy.py\", line 87, in _run_code\r\n exec(code, run_globals)\r\n File \"\/raid\/data\/gustavoag\/projects\/char2subword\/research\/preprocessing\/split_wiki.py\", line 96, in <module>\r\n main()\r\n File \"\/raid\/data\/gustavoag\/projects\/char2subword\/research\/preprocessing\/split_wiki.py\", line 65, in main\r\n data = nlp.load_dataset('wikipedia', f'{args.date}.{args.lang}', beam_runner='DirectRunner', split='train')\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/direct\/direct_runner.py\", line 119, in run_pipeline\r\n return runner.run_pipeline(pipeline, options)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 172, in run_pipeline\r\n self._latest_run_result = self.run_via_runner_api(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 183, in run_via_runner_api\r\n return self.run_stages(stage_context, stages)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 338, in run_stages\r\n stage_results = self._run_stage(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 512, in _run_stage\r\n last_result, deferred_inputs, fired_timers = self._run_bundle(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 556, in _run_bundle\r\n result, splits = bundle_manager.process_bundle(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 940, in process_bundle\r\n for result, split_result in executor.map(execute, zip(part_inputs, # pylint: disable=zip-builtin-not-iterating\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/concurrent\/futures\/_base.py\", line 611, in result_iterator\r\n yield fs.pop().result()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/concurrent\/futures\/_base.py\", line 439, in result\r\n return self.__get_result()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/concurrent\/futures\/_base.py\", line 388, in __get_result\r\n raise self._exception\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/utils\/thread_pool_executor.py\", line 44, in run\r\n self._future.set_result(self._fn(*self._fn_args, **self._fn_kwargs))\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 932, in execute\r\n return bundle_manager.process_bundle(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 837, in process_bundle\r\n result_future = self._worker_handler.control_conn.push(process_bundle_req)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/worker_handlers.py\", line 352, in push\r\n response = self.worker.do_instruction(request)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/worker\/sdk_worker.py\", line 479, in do_instruction\r\n return getattr(self, request_type)(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/worker\/sdk_worker.py\", line 515, in process_bundle\r\n bundle_processor.process_bundle(instruction_id))\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/worker\/bundle_processor.py\", line 977, in process_bundle\r\n input_op_by_transform_id[element.transform_id].process_encoded(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/worker\/bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 330, in apache_beam.runners.worker.operations.Operation.output\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 332, in apache_beam.runners.worker.operations.Operation.output\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 195, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 670, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 671, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 963, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1030, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"apache_beam\/runners\/common.py\", line 961, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 553, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1122, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 195, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 670, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 671, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 963, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1030, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"apache_beam\/runners\/common.py\", line 961, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 553, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1122, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 195, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 670, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 671, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 963, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1045, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/future\/utils\/__init__.py\", line 446, in raise_with_traceback\r\n raise exc.with_traceback(traceback)\r\n File \"apache_beam\/runners\/common.py\", line 961, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 553, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1095, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 500, in _clean_content\r\n text = _parse_and_clean_wikicode(raw_content, parser=mwparserfromhell)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 556, in _parse_and_clean_wikicode\r\n section_text.append(section.strip_code().strip())\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/wikicode.py\", line 643, in strip_code\r\n stripped = node.__strip__(**kwargs)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 63, in __strip__\r\n return self.normalize()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 178, in normalize\r\n return chrfunc(htmlentities.name2codepoint[self.value])\r\nKeyError: \"000nbsp [while running 'train\/Clean content']\"```","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":841,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\n> About the date issue, I think it's possible to use another date with\r\n> ```python\r\n> load_dataset(\"wikipedia\", language=\"es\", date=\"...\", beam_runner=\"...\")\r\n> ```\r\n\r\nI tried your suggestion about the date and the function does not accept the language and date keywords. I tried both on `nlp` v0.4 and the new `datasets` library (v1.0.2):\r\n```\r\nload_dataset(\"wikipedia\", language=\"es\", date=\"20200601\", beam_runner='DirectRunner', split='train')\r\n```\r\nFor now, my quick workaround to keep things moving was to simply change the date inside the library at this line: [https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wikipedia\/wikipedia.py#L403](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/datasets\/wikipedia\/wikipedia.py#L403)\r\n\r\nNote that the date and languages are valid: [https:\/\/dumps.wikimedia.org\/eswiki\/20200601\/dumpstatus.json](https:\/\/dumps.wikimedia.org\/eswiki\/20200601\/dumpstatus.json)\r\n\r\nAny suggestion is welcome :) @lhoestq \r\n\r\n\r\n## **[UPDATE]**\r\n\r\nThe workaround I mentioned fetched the data, but then I faced another issue (even the log says to report this as bug):\r\n```\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\n```\r\n\r\nHere's the full stack (which says that there is a key error caused by this key: `KeyError: '000nbsp'`):\r\n\r\n```Downloading and preparing dataset wikipedia\/20200601.es (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gustavoag\/.cache\/huggingface\/datasets\/wikipedia\/20200601.es\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 74.7k\/74.7k [00:00<00:00, 1.53MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 232M\/232M [00:48<00:00, 4.75MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 442M\/442M [01:39<00:00, 4.44MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 173M\/173M [00:33<00:00, 5.12MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 344M\/344M [01:14<00:00, 4.59MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 541M\/541M [01:59<00:00, 4.52MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 476M\/476M [01:31<00:00, 5.18MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 545M\/545M [02:02<00:00, 4.46MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 299M\/299M [01:01<00:00, 4.89MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 9.60M\/9.60M [00:01<00:00, 4.84MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 423M\/423M [01:36<00:00, 4.38MB\/s]\r\nWARNING:apache_beam.options.pipeline_options:Discarding unparseable args: ['--lang', 'es', '--date', '20200601', '--tokenizer', 'bert-base-multilingual-cased', '--cache', 'train', 'valid', '--max_dataset_length', '200000', '10000']\r\n\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\nERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.\r\nTraceback (most recent call last):\r\n File \"apache_beam\/runners\/common.py\", line 961, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 553, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1095, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 500, in _clean_content\r\n text = _parse_and_clean_wikicode(raw_content, parser=mwparserfromhell)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 556, in _parse_and_clean_wikicode\r\n section_text.append(section.strip_code().strip())\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/wikicode.py\", line 643, in strip_code\r\n stripped = node.__strip__(**kwargs)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 63, in __strip__\r\n return self.normalize()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 178, in normalize\r\n return chrfunc(htmlentities.name2codepoint[self.value])\r\nKeyError: '000nbsp'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/runpy.py\", line 194, in _run_module_as_main\r\n return _run_code(code, main_globals, None,\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/runpy.py\", line 87, in _run_code\r\n exec(code, run_globals)\r\n File \"\/raid\/data\/gustavoag\/projects\/char2subword\/research\/preprocessing\/split_wiki.py\", line 96, in <module>\r\n main()\r\n File \"\/raid\/data\/gustavoag\/projects\/char2subword\/research\/preprocessing\/split_wiki.py\", line 65, in main\r\n data = nlp.load_dataset('wikipedia', f'{args.date}.{args.lang}', beam_runner='DirectRunner', split='train')\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/direct\/direct_runner.py\", line 119, in run_pipeline\r\n return runner.run_pipeline(pipeline, options)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 172, in run_pipeline\r\n self._latest_run_result = self.run_via_runner_api(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 183, in run_via_runner_api\r\n return self.run_stages(stage_context, stages)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 338, in run_stages\r\n stage_results = self._run_stage(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 512, in _run_stage\r\n last_result, deferred_inputs, fired_timers = self._run_bundle(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 556, in _run_bundle\r\n result, splits = bundle_manager.process_bundle(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 940, in process_bundle\r\n for result, split_result in executor.map(execute, zip(part_inputs, # pylint: disable=zip-builtin-not-iterating\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/concurrent\/futures\/_base.py\", line 611, in result_iterator\r\n yield fs.pop().result()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/concurrent\/futures\/_base.py\", line 439, in result\r\n return self.__get_result()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/concurrent\/futures\/_base.py\", line 388, in __get_result\r\n raise self._exception\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/utils\/thread_pool_executor.py\", line 44, in run\r\n self._future.set_result(self._fn(*self._fn_args, **self._fn_kwargs))\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 932, in execute\r\n return bundle_manager.process_bundle(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 837, in process_bundle\r\n result_future = self._worker_handler.control_conn.push(process_bundle_req)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/worker_handlers.py\", line 352, in push\r\n response = self.worker.do_instruction(request)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/worker\/sdk_worker.py\", line 479, in do_instruction\r\n return getattr(self, request_type)(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/worker\/sdk_worker.py\", line 515, in process_bundle\r\n bundle_processor.process_bundle(instruction_id))\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/worker\/bundle_processor.py\", line 977, in process_bundle\r\n input_op_by_transform_id[element.transform_id].process_encoded(\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/apache_beam\/runners\/worker\/bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 330, in apache_beam.runners.worker.operations.Operation.output\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 332, in apache_beam.runners.worker.operations.Operation.output\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 195, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 670, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 671, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 963, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1030, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"apache_beam\/runners\/common.py\", line 961, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 553, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1122, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 195, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 670, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 671, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 963, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1030, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"apache_beam\/runners\/common.py\", line 961, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 553, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1122, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 195, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 670, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 671, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 963, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1045, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/future\/utils\/__init__.py\", line 446, in raise_with_traceback\r\n raise exc.with_traceback(traceback)\r\n File \"apache_beam\/runners\/common.py\", line 961, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 553, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1095, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 500, in _clean_content\r\n text = _parse_and_clean_wikicode(raw_content, parser=mwparserfromhell)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/nlp\/datasets\/wikipedia\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50\/wikipedia.py\", line 556, in _parse_and_clean_wikicode\r\n section_text.append(section.strip_code().strip())\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/wikicode.py\", line 643, in strip_code\r\n stripped = node.__strip__(**kwargs)\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 63, in __strip__\r\n return self.normalize()\r\n File \"\/home\/gustavoag\/anaconda3\/envs\/pytorch\/lib\/python3.8\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 178, in normalize\r\n return chrfunc(htmlentities.name2codepoint[self.value])\r\nKeyError: \"000nbsp [while running 'train\/Clean content']\"```","embeddings":[0.2467407435,-0.0418637209,-0.121031262,0.312300235,0.1007470265,0.200199604,0.1836347133,0.1085749045,0.6142448187,-0.2803642154,0.1189876348,0.2007731944,0.034528438,-0.1363366842,0.0880576223,-0.2634973526,0.0546107702,-0.1040422618,0.0303132087,-0.2449040413,-0.3982990086,0.2610419989,-0.3530089557,0.0145684741,-0.1812228709,0.2081045955,-0.0124996239,-0.1226630807,0.0534755848,-0.4016560614,0.1271000654,0.1967625618,0.3313166201,0.197395429,-0.0001186253,-0.1977957487,0.6336007118,-0.162105158,-0.5641887784,-0.1122205406,-0.1143344641,-0.436532259,0.2821566463,-0.1752434075,-0.0461251885,-0.3043342829,0.161864087,-0.5166298747,0.1445365399,0.22253187,0.1954718828,-0.0924769938,0.0923195779,0.1296254545,0.2619315684,0.0538863093,0.0945693552,0.1151472703,0.0538561195,-0.3287509978,0.0647287369,0.0624158122,-0.1051380038,-0.1172371209,0.3672706485,-0.1092405543,-0.1217817441,-0.5950627327,0.0641387179,0.0947354436,0.7301491499,-0.0377021991,-0.3581014872,-0.1889031976,-0.006987181,0.0573049374,0.172826916,0.1559323519,-0.0244679321,-0.1372971684,0.0292161051,-0.2469111979,-0.0619474538,0.743431747,-0.0167284198,0.3858577013,0.0737346858,0.1732941717,0.0626394004,-0.3136678934,-0.1730396897,-0.219293803,0.3790635467,0.4839769602,-0.1226430684,0.3475767076,-0.0123350741,0.3825616837,-0.0521783829,-0.158069253,-0.2352990359,0.1662348509,-0.1662418544,0.142632857,0.2290004045,-0.0134099592,0.3087078631,-0.135526821,0.2983295918,-0.064053908,-0.0458319858,0.0196876992,-0.1417012215,-0.199209854,-0.1657811403,-0.0544730239,-0.0478284135,-0.3128400147,0.2918701172,0.1117087528,-0.2736557126,-0.3444135487,-0.1295112967,0.3133083284,-0.1838987172,0.2302260101,0.1439910084,0.2676378489,-0.3891966045,-0.0696639568,-0.0016796045,0.0119818524,-0.2927108705,0.1612307727,0.3973585963,0.0833454207,0.2268027514,0.074819766,-0.014652444,-0.1550946087,-0.0783686116,-0.2006682754,0.0307389423,-0.0432604589,0.1019617915,0.5389502048,0.0627788231,-0.2294377238,-0.1606117338,0.0834421068,-0.109654516,0.3419541717,-0.0056775445,0.1212532222,-0.3625359833,-0.1406649351,-0.2977912128,0.2064987868,0.0987751856,-0.2669972479,-0.022539977,-0.3690509796,-0.1852134615,-0.1132443026,0.2871932089,0.4969678819,-0.2388131768,-0.2270080596,-0.0786340609,0.1175903901,0.1469183862,0.2245139778,-0.0805962011,0.1937002093,-0.1273546219,-0.0071627893,0.4132717252,-0.1051344052,0.0572991483,0.0223826915,0.2743159533,0.2108414769,-0.1830105931,0.0946493596,0.1899325997,0.0379426181,0.064643465,0.5457744002,0.2533354759,-0.0215168614,-0.3435814083,-0.0450481027,0.6537558436,0.3368974626,0.1342661381,-0.1086801291,0.05735882,0.3899138272,0.1507699043,0.0243369956,0.1013401523,0.3543474972,-0.3717959821,0.2649192512,0.309679687,-0.0367005803,-0.5847884417,0.2162316889,0.1441769451,0.2161539346,0.0928700194,0.1093354151,-0.3285709321,-0.1730718166,-0.2318118662,0.2040987909,0.0549687706,0.2256593853,-0.0271209273,0.5444970727,0.0311963856,0.0278387927,0.0366641656,-0.1735834479,-0.5712451935,0.3580466509,-0.0180360805,-0.0789082348,0.0505060814,0.2768629491,0.3914181292,-0.1247505546,-0.1328833103,-0.2583323121,0.3054012656,0.285580188,0.0374736823,-0.0770552754,0.2321472019,-0.3242640495,0.2337021232,0.3573760986,0.1669023335,-0.2340485007,-0.0023754444,0.33484447,-0.1550549567,0.4224194884,0.0630625188,0.0345317647,0.2765224874,-0.0269019473,-0.1470502466,-0.1411589533,0.5319138765,0.0347096138,0.0096832868,0.2768490314,-0.1995253563,-0.1459798813,0.5744214058,0.1044289395,-0.0474535115,0.2356465757,-0.1933622509,-0.2013408244,-0.0289734472,-0.1284898221,-0.1779083312,0.1603291482,0.0171624683,-0.3485314846,-0.0272583812,-0.1230650768,-0.0213758126,0.1418350786,0.1582038552,0.0385224223,0.2043384463,0.0613393486,-0.2022944093,-0.032809563,0.0632717684,0.1949344426,-0.2581511736,0.1352342963,-0.4075537622,-0.5762093663,-0.1227530763,-0.1689467728,-0.495135963,-0.4757381082,-0.1143935025,0.0920536369,0.1348278672,0.1173320115,0.0116768619,0.0618192367,-0.2605223656,-0.1536409855,-0.3928337991,-0.4991735518,-0.4106238484,-0.0192187112,0.511539638,0.2060919553,0.1550516188,0.0181787405,-0.2731821239,-0.2046999186,-0.3014617562,-0.1553871632,-0.2485721558,0.2819806039,-0.3579009175,0.3568585217,-0.1203139722,-0.0770154297,0.3586685956,-0.0063796281,0.0426307246,0.1374920458,0.1480117887,0.2510440052,-0.1259438097,-0.4387935698,-0.3378150463,-0.2986637652,-0.0107980268,0.1471051723,-0.0097866282,0.1274691075,-0.1898834705,-0.0101257004,-0.0537630767,0.0646510273,-0.1916065961,0.3349404335,0.4417556822,-0.1034759954,-0.4246730208,0.0266974531,0.0199538432,0.055486802,0.0543134846,-0.4534945488,0.048951108,-0.1315477341,0.016514061,0.1188752204,-0.1290848553,-0.0327551737,-0.1807679683,0.1202562675,0.0126705309,-0.2608505785,-0.0970647559,0.0698099807,0.1594927311,0.1414654106,0.3957380056,-0.3709970713,0.7079386711,0.1465926766,-0.0535260104,0.3381591439,0.1850141287,0.0966084599,-0.1214873642,-0.3148031533,0.0809997991,0.0567982905,0.0732563287,0.3039269447,0.0184212606,-0.4587872624,-0.2295135111,0.0927793235,-0.3207525909,-0.2435476035,0.2442715615,-0.2609005272,0.1809426397,-0.0212023165,0.0700073689,-0.0617877319,-0.4386396706,-0.0769583806,0.1941025257,0.3789052963,0.1480207443,-0.3692039549,-0.392105937,-0.2418220192,0.1042949557,0.0268055145,0.5830655694,-0.368201673,0.0149908122,0.0973754749,0.1464220583,0.2079623044,-0.3628023267,-0.0967256874,0.5620808005,0.0523689687,-0.750012219,0.0660218745,-0.1347543597,0.2047205716,0.3208766282,0.204546541,-0.3868639171,-0.3456487358,-0.1134215742,0.6048507094,-0.1742549241,-0.1486738771,-0.3017244041,-0.0627289563,-0.4483117461,-0.2812884152,-0.1597108245,0.1530341059,-0.1605032384,0.0950325653,-0.0373122357,0.3016240895,0.1328446567,0.1292181313,0.1982297301,0.3423752487,0.0227591209,0.1644464284,-0.0577660426,-0.1283718497,0.1258242577,-0.0418089852,-0.1127712429,-0.040416237,-0.1151963025,0.2561216652,0.0040566246,0.1212824211,0.073747687,0.0824491978,-0.1730267406,-0.1732717603,0.0891194865,0.3258433044,-0.0751111582,-0.3365181386,-0.4986970425,0.5371443033,0.0989008993,-0.0740019754,0.4017368257,-0.0626515299,-0.3506148756,0.5301766396,-0.0509453453,0.8433184028,-0.3003081679,0.1955537498,-0.0325539708,0.2135919034,0.5460360646,0.1363722384,-0.0340170786,-0.2968798876,0.4123263061,-0.0095589086,0.054063715,-0.0665577576,0.1078643128,-0.1792192757,0.3505465686,-0.1377652436,0.0290447157,0.028029805,0.5921821594,0.2505412102,-0.3041813374,-0.3877245784,0.0808078647,-0.2429037243,0.3783740699,-0.0661635101,0.0325785391,-0.0062684868,-0.0077963285,-0.2571872473,0.1725372672,-0.0441095307,0.1861193031,-0.1785834581,-0.2531176805,0.320086211,0.4341292381,0.1388460696,0.1387241036,-0.1515464485,0.1208060235,-0.2238152474,-0.2308393717,-0.0556879006,0.0883085802,0.0925779864,-0.0438492969,-0.2501573563,0.0243038423,-0.0591951907,-0.0486490801,-0.3381797969,0.2409883589,-0.0426510945,0.0921945944,-0.380848527,0.0418863557,-0.1134269014,-0.0094744619,0.1064318493,0.0402998328,0.2752071917,0.2200922221,0.373134613,-0.2605766952,-0.1001689807,0.2874048352,-0.0598819293,-0.0850018039,0.7583626509,0.4538274109,-0.3374438584,-0.2229510993,-0.2997133732,-0.4052827954,-0.4894793928,-0.200089559,0.1830621213,0.3600777686,-0.1642893553,0.114218168,-0.1042966694,-0.2964008749,-0.0067335591,-0.6568121314,0.0636315718,0.340077728,0.0016580726,-0.0359528325,0.0945209935,-0.3717746735,0.1721281856,0.0088422736,-0.2004660517,-0.0362885222,-0.1256379634,0.2398033291,0.3705247343,-0.0229529683,-0.0745555982,-0.1351184845,0.0613332801,-0.2670684755,0.1481943727,-0.1666263491,-0.0288465414,0.1589289755,0.1681956798,-0.1475324333,-0.2843917906,-0.4324966669,0.3609696329,-0.3049426079,-0.1883181334,0.0068975892,-0.0595958084,0.3868705034,-0.0060751485,0.2022577524,0.0264697205,0.1902355701,-0.0078729503,0.0779852644,0.3223080933,0.0898681507,0.5199073553,0.1076181456,-0.5469010472,-0.0673875287,-0.1689727008,0.0710244551,0.2085427344,-0.1280460954,0.3231802285,0.3659036458,-0.0152706876,0.5192853808,-0.0380621217,-0.0792635307,0.2815095484,0.0779809058,-0.1643521786,0.0108782286,0.5853736997,0.1379610598,-0.1037727669,0.1389348954,0.1245171428,-0.1958184838,0.1128970683,-0.0578760803,0.2018037587,-0.1756234318,0.2006062865,-0.0206894968,0.2387910634,0.0520915017,0.0392177925,0.1447448134,-0.146544978,0.0226366799,0.1288404167,0.1109216064,0.1582996547,0.2208364755,-0.0854603276,-0.2918560505,0.0891618952,0.3781286478,0.0758501664,0.2025578767,0.0398822911,0.3129250407,0.1630699337,-0.1602582484,-0.0606194772,0.2834801972,0.0812045708,-0.1401235461,-0.125184074,-0.0776649266,-0.1392726004,-0.0548244789,-0.1382765025,-0.2507811189,0.1163996235,0.4009120464,-0.0892092437,-0.5508674383,0.4159518182,-0.1234508976,0.1944138408,-0.0673287436,0.1497879922,-0.0897911862,-0.060251493,0.0327401273,0.0837589055,0.1919623911,0.2536801994,-0.4132226408,0.1657594144,-0.0451342575,0.0403790399,-0.0329474993,0.4479345977,0.1915175915,0.290791899,0.284160167,0.1063368544,-0.0679247752,-0.1725246012,0.1881706417,0.0243560039,-0.0728111416,0.1567289978,-0.0406156592,0.0821505412,-0.3063538373,-0.03688021,-0.3028011322,0.1123585105,0.2028055191,0.318768084,0.0568406954,-0.1566620469,0.0131232152,-0.1144666821,0.5851653814,0.4834053218,0.4793633223,-0.2608214915,-0.2378094345,-0.6629287004,-0.1066004038,-0.3062576652,-0.0114738764,0.2245665044,0.1316018701,0.1677433997,0.0255722813,0.0976621807,-0.1200578362,-0.0860859603,0.1423184723,-0.4113579988,-0.140954569,-0.1071040556,0.1539100707,-0.2812532783,-0.055890467,0.1590658277,-0.358704567,-0.0662567094,-0.1490184814,-0.2088910639,-0.020084668,0.3185063004,0.2466904223,0.0142018599,0.6204614639,0.0284747425,0.0257635992,-0.2770076394,-0.2164008617,0.1149386391,0.3116105497,-0.3073529601,0.2667536736,-0.2045121193,0.1605451107,-0.3032949567,0.2787632048,-0.2977600098,0.0285213087,0.0239710249,-0.1633809209,0.0643670335,-0.1165079325,-0.0065042675,0.2059287429,-0.0801009089,0.080695875,-0.3005992174,-0.1112572551,0.2099957615,-0.3703149259,-0.3459795117,0.0325196497,0.0496533476,0.0896233916,0.0790840983,-0.4082634449,-0.001955423,0.4207478166,0.0091002127,0.1098377556,0.1136834994,0.0344162509,-0.1132737994,-0.1676362902,-0.1806420237,0.1454222053,-0.0464864708,-0.0824349374,-0.114144668]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"@lhoestq Any updates on this? I have similar issues with the Romanian dump, tnx.","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":14,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\n@lhoestq Any updates on this? I have similar issues with the Romanian dump, tnx.","embeddings":[0.231198594,-0.2242944241,-0.1356965452,0.4085657001,0.1686417609,0.2521318495,0.1360733211,0.1498604268,0.723131299,-0.2435539067,-0.0025092987,0.0499951094,0.0790529177,-0.1926254481,0.0988619551,-0.1272140294,0.0742337033,-0.1101482436,0.0643479899,-0.3931700885,-0.3301622868,0.2772267759,-0.3170908689,0.013749768,-0.2585511804,0.3180563152,0.0473584272,-0.1556025892,0.0706285089,-0.2903396189,0.1405127347,0.2378277779,0.3354895413,0.1179040968,-0.0001206645,-0.1217459962,0.6853445768,-0.1523891389,-0.47875458,-0.1736898273,-0.1555294544,-0.4171535075,0.2216451615,-0.1337185502,0.0947497189,-0.2408287674,0.290120542,-0.4414692521,0.1594606787,0.0582847558,0.1969513595,-0.0886724889,0.0097790509,0.0977566317,0.2844187915,0.1132531762,0.1416764408,0.1000140458,0.040876735,-0.2274515927,-0.0296262987,0.1250065267,-0.099125497,-0.1418463886,0.1874495298,-0.1044107825,-0.1825484931,-0.6795156598,0.1505747139,0.2180954218,0.6746116281,0.0789948925,-0.3274583817,-0.1519758552,0.0827558041,0.0301468372,0.1337465793,0.3308246732,-0.1446182728,-0.0348880775,-0.0176542234,-0.2813016772,-0.0381006859,0.5898107886,-0.0856798664,0.3503937423,0.1098657772,0.1091291085,0.0978617445,-0.1932920814,-0.2175076306,-0.200551942,0.419277519,0.3955248892,-0.2924030423,0.4274419546,-0.0163236745,0.3145042658,-0.0185765233,-0.1301147044,-0.2027206272,0.2560156882,-0.1353626251,0.1668282449,0.1679094434,-0.0707909241,0.2392725348,-0.1217976362,0.3387943506,-0.0180113241,-0.1833731085,0.0006215786,-0.3191644549,-0.1907746494,-0.1819385439,-0.1187120825,-0.0959826112,-0.2975000143,0.1685823649,0.1892309189,-0.2003194392,-0.2690446675,-0.144501999,0.2856930792,-0.0736447424,0.2356275916,0.1955014318,0.3997506201,-0.4723612666,-0.0308196228,0.0448191799,-0.1213264465,-0.2197831124,0.1135416776,0.3503095806,0.1606917828,0.3111136258,-0.0547717251,0.0977851003,-0.1135627776,-0.037034709,-0.2121035755,-0.012876709,0.0634204894,0.1746838689,0.5128020644,0.0763656572,-0.2571019828,-0.1138145775,0.1319322735,-0.1282844543,0.3864475489,0.0198635906,0.0940805599,-0.3066650927,-0.1252387017,-0.2649210691,0.2721215487,0.1236329898,-0.1182551607,-0.0191597901,-0.3594434261,-0.1210303158,-0.0704220757,0.3130854964,0.5921543241,-0.2285267413,-0.1441316456,-0.14203538,0.1011906117,0.3070222139,0.2433531284,-0.0493321195,0.212595284,-0.1290672421,0.0899810195,0.3893499374,-0.0143294493,0.0731272846,-0.0496316254,0.2841254473,0.2779330611,-0.0617489405,0.0406794213,0.2077564001,0.1313802153,0.1106554717,0.508436799,0.2564708591,-0.0609084517,-0.3005440235,0.0058271801,0.5861533284,0.3145479262,0.1586782932,-0.1745739728,0.0255689267,0.2987791598,0.1941149682,0.0221879985,0.1027114466,0.3475764096,-0.31955567,0.2637336254,0.269665122,-0.0283845402,-0.5010654926,0.2056977302,0.0246757194,0.1705879122,-0.0161359254,0.1363363415,-0.3015045226,-0.1278837621,-0.21779266,0.2042318881,0.0642038211,0.1754682958,-0.1498828381,0.4917678833,0.0062578912,-0.1132715642,0.0277848225,-0.1034393385,-0.5904014111,0.4204047024,0.0447374061,-0.0683451518,0.026934918,0.2241107225,0.2653141022,-0.1953164935,-0.0897284523,-0.2533890903,0.3834173679,0.1810073107,0.2013305277,-0.0953278169,0.2865664363,-0.4601859748,0.232817173,0.2740482092,0.1453388631,-0.236702621,0.0109160254,0.244015038,-0.1516830325,0.4815168083,0.0450009108,0.0630140379,0.370328635,0.047732465,-0.1183162332,-0.1085374728,0.5623392463,0.1224461719,0.0598740131,0.3069185615,-0.2205646634,-0.1518557519,0.5031884909,0.0648300424,-0.0705102757,0.2817807496,-0.1720614433,-0.0867058858,-0.0030905202,-0.2673522234,-0.1269577146,0.1260490268,0.0995406955,-0.4170101583,0.0572301857,-0.1871452779,-0.0305897184,0.0715115294,0.1501082927,0.1372815669,0.127883926,-0.0901373923,-0.3193377554,0.0392471626,0.0928857625,0.1727453172,-0.1639804244,0.0656810775,-0.3811748028,-0.3921128511,-0.3133763373,-0.2514913678,-0.4660907984,-0.4384667575,-0.1204760671,-0.0224817637,0.1090618521,0.1760476977,-0.086326018,-0.0385478102,-0.2503173649,-0.0976119116,-0.3828794956,-0.3937867284,-0.3590959609,-0.0364689156,0.4997276366,0.2098444253,0.0333123729,-0.062270999,-0.2002071589,-0.1230201349,-0.3137001097,-0.254696399,-0.2786341906,0.1677387357,-0.2932590246,0.4170515537,-0.1636611074,-0.124575831,0.3797829449,-0.0041149789,0.0368141532,0.2570533156,0.2325877249,0.1426032186,-0.011750292,-0.4150223732,-0.3233534098,-0.2617139518,-0.1023240536,0.1570174694,-0.0733203813,0.1016905308,-0.0929102153,0.0308587831,0.0636849105,-0.0488569513,-0.0597054958,0.3979922831,0.5003034472,-0.040162269,-0.3833507001,0.078635253,0.0871256888,0.0742943436,0.0883349478,-0.4891473055,0.2518662512,-0.21074678,0.0501815379,0.1444415003,-0.0712631121,-0.0805053562,-0.2114443332,0.1255307198,0.0356351584,-0.1564305723,-0.202678293,-0.0504760817,0.211713016,0.0801400989,0.3157870173,-0.350479722,0.7058744431,0.1921321154,0.0783533156,0.2524917722,0.2384284288,0.0716558322,-0.1704698801,-0.3559666276,0.0721511543,-0.0299708098,-0.0661529675,0.3289129138,0.0281502921,-0.4627867341,-0.1496144533,0.2344335765,-0.2886903882,-0.2572989762,0.2169904113,-0.3028122187,0.1202919483,-0.0432905629,0.0069090663,0.0268970132,-0.4060660899,-0.1030900255,0.2451839596,0.3711899817,0.1058327556,-0.3993130922,-0.456694603,-0.2887693346,0.1615059972,-0.0510986969,0.5551277399,-0.2875941992,-0.0101261344,0.1905988306,0.2336499542,0.2854257822,-0.2474084347,-0.1490550786,0.4652762711,0.0380238555,-0.6231976151,-0.0137864519,-0.1637412012,0.1167109534,0.4188989103,0.189781636,-0.4300957024,-0.2656518817,-0.0388561226,0.5694671869,-0.2006335109,-0.0765785873,-0.4818706214,-0.1291991174,-0.3692965209,-0.211612314,-0.1627228707,0.1440255344,-0.2743734121,0.0132847009,-0.0120167769,0.2674705982,0.1548913568,0.1069529355,0.1238630116,0.3254612982,-0.0935813263,0.1908935159,0.0388204344,-0.1643135548,0.2609273791,-0.0883052275,-0.3104438782,-0.0590113141,-0.1672922224,0.3420469463,0.0068755383,0.1067957357,0.062208496,0.0659936741,-0.1949153095,-0.0121012954,0.0520721972,0.2775200307,-0.0034069226,-0.3329298794,-0.3923076093,0.5533442497,-0.0269399341,-0.0629671961,0.2945230603,-0.0988428518,-0.3665857613,0.5410224199,-0.0520403646,0.8773372173,-0.3767125309,0.1709647775,-0.1589401811,0.1940221786,0.4191613793,0.0529320389,-0.0331091955,-0.29065907,0.2451390773,-0.0272327233,0.0478736348,0.0130078169,0.1026575416,-0.1659946293,0.2503292263,-0.1539670527,0.0555314682,-0.0032756096,0.470423311,0.2476812005,-0.3381187618,-0.3266291916,0.0521252528,-0.1748310328,0.3651746213,-0.0714523047,0.1397518367,-0.0135867754,0.0383513346,-0.2488345504,0.16719459,-0.0630080104,0.2196301669,-0.1021787301,-0.2639827728,0.2934743166,0.5170730352,0.0049232636,0.2732410431,-0.1656500548,0.1434072256,-0.1472896487,-0.208339408,0.0009446074,0.0735448897,0.0129080191,0.0618136413,-0.4331110418,0.1039604023,0.0251912456,0.0301294774,-0.3280943632,0.1763676256,0.0083154496,0.0472772159,-0.3259102404,-0.0325167291,-0.1480147392,0.0478771441,0.106988363,0.0693052411,0.1727194935,0.1488998979,0.4217494428,-0.2977423966,-0.0764734372,0.2921284139,0.0234571267,-0.0033932982,0.7687141895,0.3801554143,-0.3644343317,-0.1992286295,-0.2732874453,-0.4397443831,-0.4488041103,-0.293792069,0.1458128691,0.3572792411,-0.226086393,-0.0863305107,-0.0328396931,-0.4674906731,-0.0309248678,-0.6850931644,-0.0121289995,0.3154999018,0.0248071402,0.0398602895,-0.0786067694,-0.4837468266,0.2007858157,0.0884571299,-0.1669360101,-0.0469215587,-0.2045617253,0.2097292989,0.247848317,-0.0509754904,-0.1838492751,-0.1308384538,0.0582081676,-0.2709689736,0.1328424513,-0.1545865834,0.0486888401,0.1522944421,0.1716973037,-0.1316655874,-0.2065613866,-0.3695727587,0.4112119079,-0.3657695949,-0.1414703131,-0.0633799583,-0.0426632911,0.3686000109,0.0971762612,0.1706913561,-0.0226006657,0.2021639049,-0.027214298,-0.0087886686,0.2765034735,0.0335513428,0.57838732,0.1145752147,-0.4434992969,-0.0427372158,-0.0738745257,0.1745398641,0.2246326655,-0.1643905789,0.4204715788,0.3435082734,-0.0410754792,0.5861054659,0.0238779075,-0.0492135137,0.2720599174,0.0776563361,-0.2052349746,0.0896208584,0.6872704625,0.1966203749,-0.1943619847,0.1523566693,0.1524021029,-0.3699932694,0.1739002764,-0.0646921545,0.1782352179,-0.1115218624,0.1892976165,-0.0273525212,0.2786221206,0.1114368364,0.0014676531,0.1102849767,-0.1714064777,0.0010153082,0.0789081976,0.1409815103,-0.0708356351,0.0918019637,-0.2058390975,-0.3436744809,0.0298494138,0.4436971247,-0.0097170305,0.2399620265,-0.0022458159,0.2647947967,0.1681948453,-0.2060249895,-0.1525399387,0.2341702878,-0.0007224944,-0.2436980903,-0.1196931228,-0.023726128,-0.1587028056,-0.0361844413,-0.0088331113,-0.2057434469,0.0867247954,0.4444304705,-0.1563536674,-0.5520169735,0.4912546277,-0.251940012,0.1365534663,-0.1183473691,0.1486435831,0.0011070532,0.0170863103,0.0413629971,0.1502188146,0.1181169078,0.2742418647,-0.3787223399,0.0990929753,-0.0767315403,0.1506683081,-0.09215983,0.5184403062,0.1304636598,0.3427439034,0.21580109,0.0967306122,-0.069748871,-0.2339534163,0.2786284983,-0.0512115434,-0.154633224,0.1226312742,-0.1372813582,0.025306562,-0.3407315314,-0.0936911255,-0.340880245,0.2450985014,0.2256659269,0.1566182375,0.0654929876,-0.0737389028,0.0150649175,-0.1530887634,0.6583413482,0.544218421,0.2883574963,-0.2854521275,-0.2108085752,-0.6145953536,-0.0102881864,-0.2837764025,-0.0010293288,0.2485931963,0.1864219606,0.1245619953,0.0498291217,0.0217252932,-0.1410493851,-0.1814438552,0.2570175827,-0.3932396472,-0.1755452305,-0.2044725269,0.1232550815,-0.3827182651,-0.093016237,0.1924562156,-0.2498030066,-0.0817600712,-0.0604010336,-0.1103816554,0.0622184835,0.2647798657,0.1904657036,0.0234022979,0.6052153111,0.0422620177,0.0667470023,-0.2617509365,-0.2737630606,0.1296344101,0.3509278893,-0.2317469716,0.2168164849,-0.1663697511,0.1687141657,-0.2534519434,0.3687879145,-0.1417267174,-0.0694631785,-0.0132057425,-0.2106514126,0.0056493999,-0.1943420619,0.0662859455,0.303384155,-0.1049032658,0.0794686824,-0.2618261874,-0.1129093021,0.1582173407,-0.3588274419,-0.4618912339,0.0201520473,0.0898669586,0.1170257926,0.1931000501,-0.4079059064,0.0256375223,0.39058882,-0.056125544,0.1096264645,0.1606512666,0.0728284866,-0.1497803479,-0.2000660896,-0.0632709116,0.1109582484,-0.0173983276,-0.0911841393,-0.1203394309]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"Hey @gaguilar ,\r\n\r\nI just found the [\"char2subword\" paper](https:\/\/arxiv.org\/pdf\/2010.12730.pdf) and I'm really interested in trying it out on own vocabs\/datasets like for historical texts (I've already [trained some lms](https:\/\/github.com\/stefan-it\/europeana-bert) on newspaper articles with OCR errors).\r\n\r\nDo you plan to release the code for your paper or is it possible to get the implementation \ud83e\udd14 Many thanks :hugs: ","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":57,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nHey @gaguilar ,\r\n\r\nI just found the [\"char2subword\" paper](https:\/\/arxiv.org\/pdf\/2010.12730.pdf) and I'm really interested in trying it out on own vocabs\/datasets like for historical texts (I've already [trained some lms](https:\/\/github.com\/stefan-it\/europeana-bert) on newspaper articles with OCR errors).\r\n\r\nDo you plan to release the code for your paper or is it possible to get the implementation \ud83e\udd14 Many thanks :hugs: ","embeddings":[0.3906110525,-0.2458596528,-0.0904512033,0.4170152545,0.0234599393,0.1634708643,0.1849508882,0.1755524725,0.511062026,-0.265707761,0.0157987792,-0.1771205217,-0.016337797,-0.1042658538,0.083665736,-0.0257931259,0.1556571722,-0.0064998199,0.050713893,-0.2967506945,-0.2989606559,0.329172194,-0.3816176355,0.059065327,-0.2614990175,0.2898676991,0.0583659597,-0.135697782,-0.0120732691,-0.2747810185,0.2623194456,0.2832196057,0.4896444082,0.1554790288,-0.0001210404,-0.104849793,0.5296111107,-0.2100765109,-0.434851557,-0.0876918733,0.1109086275,-0.2407892942,0.2297366709,-0.170543626,0.0351688936,-0.0798513591,0.3073665202,-0.4942201078,0.1781291962,0.0284844358,0.1619759947,-0.2675949335,0.085838072,0.1176090389,0.3122581244,0.0666575953,0.1999560893,-0.0564155653,0.1693619341,-0.2023465633,-0.0613853633,0.2206851691,-0.1636947393,-0.1409708411,0.1139511168,-0.031828437,-0.2651315033,-0.7127404213,0.262950778,0.142591089,0.6615582705,0.0197381247,-0.2775624692,-0.0204803683,0.0560813993,0.1707977653,0.2110823989,0.2694203854,-0.1647181809,0.0527461357,0.0525142513,-0.2845051587,-0.0732934996,0.5155893564,0.0218740199,0.3449914157,0.0481329337,0.1534670293,0.0292204823,-0.3066637814,-0.2476733923,-0.1166796684,0.3090465665,0.441865772,-0.3259526491,0.4467229545,-0.08317855,0.3020559251,0.1693530381,-0.2793391943,-0.2269317359,0.216872707,-0.1235042289,0.1648903936,0.1029393673,-0.1278946698,0.1356350631,-0.0819741115,0.4372631907,0.0055121942,-0.2835946977,0.0510559864,-0.3815603554,-0.2719697058,-0.1230667159,-0.1756993979,0.0755281523,-0.5338534117,0.1250914782,0.1443604231,-0.4663891792,-0.2009459138,-0.1746475399,0.3171269,-0.1507844329,0.1184732392,0.3016981781,0.2242139727,-0.5199456215,0.00496571,0.0979952067,-0.1183567867,-0.2367059737,0.1520155072,0.4170891345,0.0140928002,0.315595001,-0.0823279098,0.2053255886,-0.1119857803,-0.1444021612,-0.2095376253,0.0242700428,-0.0174173359,0.1545064747,0.483289212,-0.0132773435,-0.2494764477,-0.1152053177,0.1660108417,-0.1901400536,0.4341748059,-0.0041948585,0.0465076081,-0.3614522815,-0.1916782707,-0.1078745052,0.3816649616,0.2352099121,-0.1657788754,0.0189094581,-0.3135446012,-0.2109045684,-0.0437980704,0.5165929794,0.5619415641,-0.2702687085,-0.1021604165,0.1821817011,0.2397642732,0.2302520871,0.3319924176,0.0419325233,0.1244721636,-0.1256992519,0.0838077664,0.3810041249,-0.0728212073,-0.0444147699,-0.1370721906,0.2496331185,0.1740491092,0.0639344677,-0.0021652472,0.1435866207,-0.0364138857,0.0722809359,0.477065295,0.256054908,-0.0451618433,-0.3485659659,-0.186723575,0.6346889734,0.2341763079,0.1699862629,-0.2361823767,-0.0794716403,0.2613799572,0.2289856672,-0.1427284181,0.1414875388,0.2766788602,-0.2538872957,0.2242304236,0.4858230352,0.0357992798,-0.32580778,0.2015612274,0.0289686881,0.2578414381,-0.0120714763,0.1157361791,-0.2651404142,-0.2199261487,-0.2003055066,0.0093617775,0.0058349576,0.115889132,-0.0417724997,0.5396041274,-0.1149114147,-0.2178815603,-0.0612491183,-0.0855938643,-0.4907000959,0.4045851231,0.0090365233,-0.1398370862,-0.0711571723,0.3765223324,0.2453237474,-0.060257908,-0.1392203569,-0.2804969251,0.2026958317,0.0408371016,0.2288630456,0.0024843169,0.2272130996,-0.527145505,0.3992442489,0.3866473436,0.1387699842,-0.1339149475,0.0012375616,0.2392908633,-0.0441633612,0.4396706223,0.0859869123,-0.0385911949,0.4223067164,0.0905772597,-0.1323942244,-0.0435808003,0.4852474928,0.0732430741,0.1626705378,0.1913674176,-0.1442712396,-0.1936851889,0.5177627206,0.0909318551,-0.0187756233,0.3626833856,-0.2295614779,-0.0533266217,-0.1021808088,-0.2229631245,-0.127301231,0.1584008932,0.0243470501,-0.4226490557,0.0394837186,-0.1722596735,-0.0449526794,0.049581755,0.165425688,0.0041067321,0.1203465313,0.0365821086,-0.3161322773,0.0196039006,-0.0089468528,0.1602902561,-0.244757846,0.0077647972,-0.2301369458,-0.5507711768,-0.2677677572,-0.0249456838,-0.3480504751,-0.1943135411,-0.0272616148,-0.1914621145,0.1517439783,0.132422179,-0.0981023833,0.1305907965,-0.192523852,-0.0035208424,-0.2965407073,-0.3732478023,-0.2692627907,-0.0616008639,0.538720727,0.1602941304,0.0685506836,0.0189862493,-0.2335125506,-0.1795647144,-0.5753451586,-0.1952230781,-0.2593540847,0.1230406165,-0.2514222264,0.2690456808,-0.1243936345,-0.2165589184,0.3017782569,-0.0531293564,-0.052869536,0.2159952074,0.2227418572,0.0537780225,-0.0508998223,-0.4319815636,-0.1463396996,-0.1911932975,0.0201552194,0.2135230452,-0.0891561508,0.1526998729,-0.2282402515,0.1837090403,0.1528959423,0.1669873446,-0.0652987063,0.3019635975,0.4984689057,0.0652890876,-0.4329943359,0.0519565679,-0.0042816373,-0.0170327723,0.0980360135,-0.4875886142,0.1571108997,-0.1949493885,-0.108396098,0.2408302873,-0.0894410461,-0.0174036752,-0.1026525795,0.1484975219,0.064368844,-0.241184026,-0.0574180409,0.0351131521,0.1684246212,0.1630053967,0.4710479081,-0.2816250026,0.5095044374,0.2555075288,0.0217603724,0.234055087,0.3127952516,-0.0030243618,-0.2253617495,-0.3853049278,0.2120362818,-0.1305787116,-0.2007102221,0.4930437803,0.1017165631,-0.4706683457,-0.1700117141,0.1370811611,-0.4081181288,-0.2867386639,0.1878869534,-0.3789225221,0.0918688998,-0.1119489819,0.0144185079,0.0737813786,-0.4300969839,0.0140181202,0.1315654069,0.4339990318,0.0714245066,-0.3578455746,-0.1737831682,-0.4213225245,0.279715091,-0.0169851184,0.515320003,-0.2730410695,-0.0953856036,0.2180935889,0.0899831802,0.1950146109,-0.1942401528,-0.0948112532,0.3745181859,0.0353590026,-0.542188406,-0.0446794555,-0.1850262433,0.0787377059,0.4650762677,0.2198137343,-0.370888412,-0.2298645824,0.0967989489,0.3545049131,-0.1607287824,-0.0920621455,-0.4532794952,-0.2630352676,-0.4781172574,-0.1487607956,-0.1451946646,0.1708305478,-0.2001547664,-0.0280962661,-0.0539931841,0.1814184785,0.2792246044,0.1019654647,0.0412670374,0.3717172742,0.0310328081,0.2131714225,-0.0054113953,-0.2380092144,0.2672728598,-0.1106412634,-0.262287885,-0.0459533371,-0.2169791162,0.500957191,0.0594177321,0.0803892985,0.2558214366,0.0931416079,-0.1713900268,-0.1147550792,0.0068125748,0.372397095,0.1456341445,-0.4133529067,-0.4464229643,0.4641010463,0.01441269,-0.1129802987,0.2672722638,0.144834727,-0.423982054,0.5985949039,0.131193921,0.9971320629,-0.3668914437,0.2018922567,-0.2075106353,0.2028149962,0.3728070855,0.0160654802,-0.1227189749,-0.3841163218,0.2088970244,-0.0226503965,0.1443276405,0.0302236974,-0.0453519635,-0.3547632992,0.1877815574,0.1019592732,0.2176272869,0.088415578,0.3985004425,0.2086004466,-0.4390072227,-0.3299594522,0.0061997185,-0.1620272994,0.2169494033,-0.0881794244,0.0771138966,-0.0005553461,-0.0640535131,-0.3029723763,0.2212537974,-0.2876311839,0.1381070167,-0.125560984,-0.3206408024,0.4721293747,0.5965858102,0.3510197699,0.2904988825,-0.1735066772,0.0105503732,-0.1729169637,-0.2807391286,0.0505279265,0.0023192547,0.0547874346,-0.0329862498,-0.4186165631,0.1099379063,0.1539148539,-0.0091176303,-0.3345573246,0.0581942871,0.052010607,-0.0574338473,-0.2535780966,0.0690897331,-0.1837307513,-0.0002323558,0.0812404603,0.0987496451,-0.0213150512,0.2302994877,0.5099390149,-0.2384019643,-0.0022814586,0.2756131589,-0.0330426395,0.0459997803,0.8242175579,0.3531859815,-0.2382497042,-0.160487175,-0.2077345103,-0.350625366,-0.3734713495,-0.2744709253,0.1705955714,0.1926055998,-0.1020293534,0.1096994653,0.0579543859,-0.413662523,-0.0560124256,-0.6720861793,-0.1262620091,0.3311512172,-0.1153274849,0.0324970186,-0.0142588876,-0.5099301934,0.1948366165,0.1946291476,-0.1821058542,-0.0208346564,-0.2921030223,0.245923847,0.2934960127,-0.1317529678,-0.0726341233,-0.0188716371,0.0475855917,-0.1795159876,0.1140779555,-0.1495864987,-0.0031640094,0.155625701,0.1842788458,-0.162033841,-0.2362392247,-0.3010374308,0.2912217975,-0.3832778633,-0.1784518659,-0.101108402,-0.0268252008,0.3944601715,0.0486964323,0.2048484534,-0.168572247,0.1366659999,-0.1157348305,0.0159154832,0.4320169687,0.0235158019,0.414054662,0.1401090026,-0.2484562844,0.0704439208,-0.0075616548,0.0395555571,0.2657284439,-0.2449883819,0.3485542536,0.1776705533,0.0376950502,0.4427368939,0.0143102119,-0.0834108591,0.4294576049,0.0554975271,-0.2726632059,0.2773701847,0.7467560172,0.020000238,-0.2252029926,0.2666770816,0.1365985274,-0.2904500961,0.1645984799,-0.0856710449,0.1493192613,-0.1826094389,0.0764398724,0.0454641618,0.2334896475,0.0234495122,-0.0325806178,0.0579319894,-0.0885806531,0.0422131903,0.0480658188,0.236126408,-0.1614205986,0.3138409257,-0.065184474,-0.2827222347,0.1595359296,0.4230858684,0.067978546,0.0578721203,-0.002941784,0.5064715147,0.2450966686,-0.1024085432,-0.2211094648,0.317928344,-0.0194962639,-0.3253653049,-0.2263767421,-0.0569535829,-0.1090921015,-0.0277843606,-0.0417712294,-0.2031984627,0.0063877692,0.4082840085,-0.1870365739,-0.4960336685,0.3953443766,-0.2119192332,0.2540350854,-0.2087828666,0.2732511163,0.0641731471,0.0511596426,0.0361319855,0.3524009287,0.2060094178,0.2829681933,-0.5382399559,0.1005950719,0.0231573675,0.1314054579,0.0017922762,0.492710799,0.1881440431,0.358055979,0.2191223055,0.045040302,-0.0004441419,-0.1535635889,0.2608041465,-0.1749703288,-0.1246768311,-0.087945506,-0.0300967637,0.0313685387,-0.3653179109,-0.0056127962,-0.3547827899,0.119356133,0.2188405693,0.0083387047,-0.0003710155,0.011787314,-0.0110135023,-0.00262405,0.6466359496,0.6027101874,0.3966038525,-0.235325262,-0.196337685,-0.6136168838,-0.0775855556,-0.1869963109,-0.0533818267,0.1000551805,0.120770745,0.2142696828,-0.067432493,0.0162823237,-0.1170122474,-0.1347685456,0.2401026636,-0.3762452602,-0.0697817504,-0.0522630066,0.1122822464,-0.4041323066,-0.0911911055,0.0363357626,-0.2400300205,-0.1073957086,-0.007648055,-0.094588317,-0.0707323253,0.1332762241,0.2698799074,0.0466827303,0.6554793715,-0.0406153984,0.0991711542,-0.1925159544,-0.2858042121,0.0517743491,0.2309688032,-0.0677301884,0.2449161112,-0.18349877,0.0795507953,-0.1450666189,0.3292119205,-0.1423070431,-0.0384935252,-0.0850031972,-0.1523203105,-0.1048032939,-0.1643165946,-0.0506860912,0.1647798568,-0.1563882679,0.0143320672,-0.31517452,-0.0461168662,0.2633312643,-0.4905767143,-0.3067497909,0.0222112238,0.2202163488,0.1670764089,0.1308675259,-0.5824353695,0.0124707595,0.296341449,0.1136178672,0.020018464,0.2618848979,0.2026687413,-0.1604249477,-0.2549130023,-0.1259838939,0.0105358409,0.0325701125,-0.1665249616,-0.2512551248]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"Hi @stefan-it! Thanks for your interest in our work! We do plan to release the code, but we will make it available once the paper has been published at a conference. Sorry for the inconvenience!\r\n\r\nHi @lhoestq, do you have any insights for this issue by any chance? Thanks!","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":49,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nHi @stefan-it! Thanks for your interest in our work! We do plan to release the code, but we will make it available once the paper has been published at a conference. Sorry for the inconvenience!\r\n\r\nHi @lhoestq, do you have any insights for this issue by any chance? Thanks!","embeddings":[0.2403038889,-0.2484169006,-0.1631048322,0.4403241575,0.1737348139,0.2488844991,0.1497551203,0.1210620478,0.7601813078,-0.2220047265,0.022221826,0.0713406503,0.0954832956,-0.1848984063,0.1068781391,-0.1616459787,0.04406159,-0.1030339375,0.0729285851,-0.3541831076,-0.3202170134,0.286111474,-0.3100569844,0.0425740592,-0.2333234549,0.2517737746,0.0728946924,-0.1765368283,0.1209294796,-0.3202476799,0.1578963548,0.2314836532,0.3251018524,0.0852677822,-0.0001190338,-0.0872882307,0.6546165347,-0.1592825651,-0.4487918615,-0.1851799786,-0.097332187,-0.389665246,0.2533375025,-0.1280304044,0.0163537487,-0.2522680759,0.264759779,-0.5146146417,0.1419245154,0.0559386946,0.2139525115,-0.1379256099,0.0227346662,0.1068883613,0.2769060731,0.0649670586,0.1694057435,0.1623741984,0.0854746848,-0.2234982997,-0.0387618989,0.1137027219,-0.1003309861,-0.1278005987,0.1507626325,-0.1206091344,-0.1666315347,-0.6842166781,0.1447825879,0.2033780366,0.6822656393,0.0585358813,-0.2979769409,-0.1612643152,0.0933199599,0.0761077255,0.1409383118,0.3160327971,-0.0881594121,-0.0372593552,-0.0245255753,-0.2631927133,-0.0459631719,0.5973466635,-0.0228506625,0.3351396918,0.0907268748,0.1467968673,0.0506305471,-0.1951539963,-0.2783495486,-0.19472,0.4301373661,0.381080091,-0.3087663352,0.4801419973,-0.0456432663,0.3292756081,0.0133883953,-0.0930944905,-0.1806409508,0.2417501956,-0.1000083834,0.1367214024,0.1078261212,-0.0790935829,0.2475861162,-0.1791079044,0.3294415772,0.0208850019,-0.1913853139,0.0133500993,-0.325407356,-0.1911403537,-0.1798246652,-0.1045292094,-0.0958530679,-0.2829456627,0.1404173076,0.1610734165,-0.2561336458,-0.2864630222,-0.1066169813,0.2845076323,-0.0664205775,0.2138198912,0.1953395307,0.3847078979,-0.4297505021,-0.0301164277,0.040981587,-0.0935204998,-0.2020628154,0.1292860955,0.3580398262,0.1207701415,0.3291495442,-0.0694355294,0.081036225,-0.1557005346,-0.0373727046,-0.2140842229,-0.0087239351,0.0356196873,0.1680486351,0.5185453892,0.0808475316,-0.2173630893,-0.1059135795,0.1191341281,-0.1258127391,0.3667230904,0.0681520626,0.1092630178,-0.2834845781,-0.1489905417,-0.2708299756,0.2827830911,0.1446893364,-0.1428505629,0.0068381266,-0.3387593627,-0.1361780018,-0.0932921469,0.3333103955,0.6020081043,-0.237949878,-0.1239936128,-0.1097203121,0.0625886098,0.2680113614,0.226448223,-0.0461876057,0.2239888608,-0.1566924155,0.0619398914,0.3468519747,-0.0469791926,0.0503417365,-0.0675895661,0.2577605247,0.2884011865,-0.0052862037,0.0463495813,0.1783651859,0.137341395,0.1054223925,0.5019161701,0.264608711,-0.0576131083,-0.2766193748,0.0062879673,0.5969915986,0.3590583205,0.1414652169,-0.1886006296,0.024151193,0.2261442542,0.1959433854,-0.0028576991,0.0716466829,0.3509297073,-0.3852471411,0.2675403655,0.254868269,-0.0209833253,-0.5135547519,0.2265412062,0.0287272409,0.1059741005,0.0159471445,0.1545713693,-0.3214335144,-0.1303369701,-0.227605015,0.2190585434,0.0747418031,0.1921162307,-0.1218720078,0.4916851223,0.0129682105,-0.0874283537,-0.0117388181,-0.1381868422,-0.5985741019,0.4235678315,0.0482075028,-0.0562231094,0.0593153723,0.2330970466,0.3044797182,-0.2093183547,-0.0807318836,-0.2575537562,0.3645226955,0.2046219558,0.1982116997,-0.0947811082,0.240541175,-0.4418725073,0.2254377753,0.2603181899,0.151606515,-0.2144683003,0.0301996786,0.2297265083,-0.1430050433,0.537518084,0.0430869013,0.0495939739,0.3477125764,0.0680927485,-0.1327029318,-0.0794157013,0.5763515234,0.0573305711,0.0827741325,0.2758433521,-0.2210941613,-0.1473471969,0.5132134557,0.076164417,-0.02231328,0.2689856291,-0.167255044,-0.0592591539,-0.0539216697,-0.2574972212,-0.1739653051,0.1562022716,0.0415365025,-0.4466871023,0.050426688,-0.1968023479,-0.0201259404,0.1313177347,0.2322646827,0.1344462931,0.1571954787,-0.097515367,-0.3240645528,-0.0681784749,0.0862878859,0.1361801624,-0.1638785005,0.0931756496,-0.3838435411,-0.4041524827,-0.2602187097,-0.2542757392,-0.4711836576,-0.3847303987,-0.1355746239,-0.0394922197,0.125409469,0.1541991234,-0.0528748259,-0.0324826166,-0.2601850927,-0.1441574246,-0.3858620524,-0.4119680524,-0.3634985387,-0.0191759914,0.4967757165,0.1376648992,0.1182281673,-0.0453858525,-0.2144778371,-0.1219040528,-0.2909273803,-0.2819651961,-0.2903445661,0.1524647772,-0.2907551825,0.4180134833,-0.15801844,-0.0648819134,0.3832049668,0.0132899303,0.0320150256,0.2098772228,0.2292372733,0.1488263905,-0.0189769212,-0.4778474867,-0.2821603715,-0.2814055979,-0.0662077442,0.174546957,-0.0539825596,0.0485925972,-0.106930919,0.0735311434,0.0665916353,0.0155837629,-0.0805284381,0.3945896924,0.4707878232,-0.0173793323,-0.3659881949,0.0674857795,0.074989967,0.1151342541,0.0657223761,-0.5204170346,0.1854240745,-0.2083954066,0.0451812297,0.098008588,-0.0894570872,-0.0342035927,-0.1984021515,0.1101851761,0.0285058245,-0.2111654729,-0.1685480028,-0.0847698972,0.1542387754,0.0776105896,0.3722039163,-0.4040284753,0.6679967046,0.180007413,0.0724654719,0.26361534,0.2328600436,0.0247048941,-0.1845100969,-0.3040192723,0.0815445632,-0.0401307642,-0.0343115479,0.3433051407,0.0115958247,-0.4260087311,-0.1549558342,0.1568580419,-0.3191002011,-0.2304741144,0.2014162391,-0.3361041546,0.1445452124,-0.0662023723,-0.0007370066,0.0227244254,-0.3679701984,-0.0963786393,0.231851846,0.4085189998,0.1300846189,-0.3945157826,-0.4773107469,-0.2277921587,0.2020534277,0.0023551821,0.5626490116,-0.309643954,0.0228691008,0.2210660875,0.1677704155,0.3177931905,-0.2400725335,-0.1420292258,0.4882227778,0.0156564564,-0.638061285,-0.0204223264,-0.1622190624,0.0833286569,0.3625949025,0.1878807545,-0.4364580512,-0.2926006615,-0.0732093155,0.5856533647,-0.1869585067,-0.0772839412,-0.5057544112,-0.1394651532,-0.4415773153,-0.1582535803,-0.1779350787,0.1417432576,-0.2761985958,0.0048866146,-0.0101856114,0.2405772507,0.115362376,0.1174696684,0.1481307894,0.3268494308,-0.0886294022,0.1726894677,0.0501057282,-0.1758479774,0.2731843293,-0.0112895584,-0.2234648019,-0.0317070186,-0.1399449259,0.3491233289,-0.0059773135,0.1135569885,0.095323287,0.0818393379,-0.1781454682,-0.028683586,0.013591554,0.2783601582,-0.038218841,-0.3387447894,-0.4012156129,0.5765026212,0.0070208567,-0.089207761,0.2786211371,-0.0551364794,-0.3789120018,0.5338388681,-0.0165801272,0.849290967,-0.3891259432,0.1554397494,-0.1432348192,0.1330961138,0.4436136484,0.0589538887,-0.0184575766,-0.3015095294,0.2767418325,-0.0101542082,0.0423014536,-0.0207288601,0.0980931669,-0.1530329883,0.2524519861,-0.143523857,-0.0090275519,0.0044084932,0.4702911973,0.2433363199,-0.3317686915,-0.3470061421,0.0545624793,-0.203655839,0.3451610506,-0.0672624484,0.144468084,-0.043210566,0.0378945395,-0.2116959542,0.1651288569,-0.1171961054,0.2133906484,-0.1344571859,-0.2575912476,0.3556829393,0.5074017644,0.0215176195,0.2731773555,-0.1743643582,0.1470681727,-0.1626449227,-0.2417931557,-0.0147237713,0.0709897205,0.0129226465,0.0283667929,-0.3950636089,0.0442974046,0.0037675798,0.0431344211,-0.3927102387,0.1890946925,0.1144265011,0.083700411,-0.3093764484,-0.0286046136,-0.1517617851,0.0350534804,0.1089377254,0.0696154013,0.2089784145,0.1458031684,0.459457159,-0.2960391641,-0.0933225602,0.3001584113,0.0638604686,-0.0011977139,0.7996009588,0.3861696422,-0.359421283,-0.22984384,-0.2924443185,-0.3995828032,-0.4833544791,-0.2949711382,0.209821701,0.3583879173,-0.2310598642,-0.0729671791,-0.0307935886,-0.4497263432,-0.0161587913,-0.7090858221,0.016518617,0.3147452772,-0.0000944205,0.0260789823,-0.0277001858,-0.4851160645,0.2010199428,0.0916262269,-0.1756435484,-0.0473138168,-0.2248123288,0.1954904348,0.2992232144,-0.0352142975,-0.1304019392,-0.0882096291,0.071346283,-0.2821350694,0.1325382292,-0.1703797728,0.0527293682,0.143451497,0.1451938301,-0.1037199348,-0.2014237344,-0.392580241,0.4189735055,-0.3449261785,-0.1525970846,-0.1066151783,-0.0266448297,0.3194226325,-0.0140296016,0.1901133209,-0.0774290115,0.1578541696,-0.0850618333,-0.0823891982,0.3056702912,-0.0068961293,0.5810577869,0.10998597,-0.5075945854,-0.0799196139,-0.0467283949,0.1994897425,0.2299571633,-0.156562537,0.3871649504,0.3396727741,-0.0693697929,0.5151048303,0.0073441952,-0.052873753,0.254553467,0.0867435336,-0.2415274382,0.0920685083,0.6816080809,0.1594073027,-0.1839639693,0.1778033525,0.1591592729,-0.329059273,0.138337791,-0.0923922062,0.1674383581,-0.1172065735,0.158010602,-0.0585483536,0.2875808179,0.0745178983,0.043279063,0.1063264087,-0.1758418977,0.0457456373,0.0872724429,0.1847973913,-0.0610205494,0.1015653014,-0.201975733,-0.3225002885,0.0728746876,0.4022855163,0.0096956445,0.2069603652,-0.0112266624,0.2832611501,0.1767344028,-0.2089669704,-0.1305618584,0.2014626265,0.0201620981,-0.2395565063,-0.1525935382,-0.0021782047,-0.1380704641,-0.0669315159,-0.0068496787,-0.202361092,0.0999128297,0.4635050595,-0.1395789683,-0.5846256018,0.43340379,-0.2078950703,0.1689862758,-0.1067220196,0.1678973138,0.0310393069,0.0004861061,0.0150999473,0.127366066,0.1169172972,0.2159251124,-0.352630347,0.0818491206,-0.0694537237,0.1497204751,-0.0708671436,0.5127887726,0.1474301368,0.363458395,0.2605467737,0.1120828986,-0.0764614344,-0.1830803007,0.2877781391,-0.0758491904,-0.1495366693,0.1175031662,-0.1551468223,0.0124112656,-0.388663888,-0.051238738,-0.3217366934,0.2560248971,0.1722012013,0.1620510817,0.0402544662,-0.109150514,0.0207824968,-0.1413675845,0.6663071513,0.5066701174,0.3194197118,-0.275891304,-0.1868968606,-0.5659415126,-0.0348336883,-0.2592558563,-0.0048590265,0.2345603406,0.152818203,0.1511091292,0.0310091507,0.0640192553,-0.1681585908,-0.1787694246,0.2347784936,-0.3656275272,-0.1523028761,-0.2002972215,0.1293934137,-0.3538476527,-0.0602661446,0.1703132987,-0.2967703342,-0.0597511493,-0.0768338144,-0.0956023335,0.031084409,0.2247756422,0.2009019554,0.0253069233,0.6212067008,0.0268289,0.0655915663,-0.2723119259,-0.2518820465,0.1372406036,0.3818343878,-0.2070990056,0.2753737569,-0.1499731094,0.1425309032,-0.2214936614,0.3350886106,-0.1331698895,0.0281577893,0.0193346124,-0.2070406824,0.0183900893,-0.1782643646,0.0649191141,0.2715238631,-0.0717397779,0.0544042699,-0.2975769937,-0.1280845255,0.184530139,-0.3519735634,-0.454113245,0.055749014,0.1558382362,0.1332850903,0.1779460758,-0.3800539374,0.0439228229,0.4062698483,-0.0453390256,0.1094525829,0.1945025921,0.0544758849,-0.1300237775,-0.1876983047,-0.1143447608,0.0784830973,-0.011728568,-0.0870878622,-0.1433933973]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"This is an issue on the `mwparserfromhell` side. You could try to update `mwparserfromhell` and see if it fixes the issue. If it doesn't we'll have to create an issue on their repo for them to fix it.\r\nBut first let's see if the latest version of `mwparserfromhell` does the job.","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":51,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nThis is an issue on the `mwparserfromhell` side. You could try to update `mwparserfromhell` and see if it fixes the issue. If it doesn't we'll have to create an issue on their repo for them to fix it.\r\nBut first let's see if the latest version of `mwparserfromhell` does the job.","embeddings":[0.081004858,-0.1919707954,-0.1455876678,0.357018739,0.1417111307,0.2244725823,0.1284718215,0.1698514074,0.7591428757,-0.1323028803,0.1258179694,0.0651477799,0.0723554045,-0.225640744,0.1120552346,-0.1160828695,0.0668614954,-0.0926693827,-0.119749248,-0.3306937516,-0.378772378,0.2551675141,-0.3586254418,0.0707237497,-0.2304441184,0.2964781225,0.0954792574,-0.0933470801,0.0668037683,-0.354336828,0.1113957912,0.1501275301,0.3143440187,0.0501736812,-0.0001221014,-0.0832126215,0.7176688313,-0.1283557862,-0.417483598,-0.1124236509,-0.1387168765,-0.4225863516,0.1965474933,-0.0338458717,0.1579807252,-0.0946854874,0.2541137636,-0.4594916105,0.1522707194,0.1381663382,0.1686459035,-0.0437433794,0.037888892,0.1722898483,0.2732972801,0.0850297958,0.1379061788,0.0558566637,0.1328401268,-0.2986986339,-0.0256934054,0.1416166276,-0.1255062521,-0.074288018,0.1094742343,-0.0571211763,0.0959105417,-0.6544166803,0.1758662611,0.1662172973,0.7025114894,0.1253193319,-0.2536801696,-0.1987040043,0.0069871801,0.1365217417,0.2370205373,0.4124928117,-0.1284551024,0.0225707535,-0.0646742135,-0.1853040457,0.0156961363,0.6050239801,-0.0556667522,0.5664605498,0.1225819662,0.194029972,0.096586369,-0.2046299726,-0.2477408946,-0.2428136617,0.3923789859,0.3300985992,-0.133600533,0.4396103919,-0.0657655746,0.3765960932,-0.017255893,-0.1677450538,-0.3217781186,0.3225260973,-0.0982338339,0.1468278468,0.2242142856,-0.1507669389,0.3737666309,-0.3395674527,0.3315449059,0.0707732588,-0.3030894101,0.0902165771,-0.2847909927,-0.2116484493,-0.2382559925,-0.0894366726,-0.0229720268,-0.2458191216,0.2052039802,0.1451718658,-0.1429190934,-0.4003151953,-0.1844200492,0.3283189535,-0.0951183513,0.3528809547,0.2199086994,0.279997915,-0.471241802,-0.1743672341,0.0938560963,-0.1050391197,-0.3094645143,0.2047819495,0.311884284,0.2557216883,0.2917598188,-0.0396315493,0.092054002,-0.2491992712,-0.117688641,-0.0920443237,-0.0362549052,0.0806826577,0.2162978649,0.5053890347,0.1520207375,-0.2430995256,-0.173117891,0.1904238462,-0.1428010464,0.3200359046,0.0128071336,0.0700483397,-0.2752066255,-0.0271426681,-0.3407329917,0.2371091247,0.0794993863,-0.1422304362,-0.0680596605,-0.4404708445,-0.284509629,-0.0630098954,0.2751733661,0.6824002862,-0.2337113619,-0.0756816491,-0.1166577637,0.0289645363,0.2386087924,0.1880309433,-0.1038100198,0.2789823413,-0.1744201332,0.0498729125,0.3225488365,-0.0450440086,0.0905496404,-0.0659107715,0.2943434715,0.2915300727,-0.0669319257,-0.076311402,0.0243998226,0.0208586454,0.050952971,0.4110230505,0.1989032179,0.0261734091,-0.234259367,-0.1510098726,0.5229259133,0.3386918604,0.0948804393,-0.1500118226,0.0498552509,0.340354532,0.2524162531,-0.0655561164,0.1375987232,0.3369295001,-0.3582295179,0.2677420974,0.2269896865,-0.1253042817,-0.554710269,0.230648309,0.0456744879,0.2032351941,0.0385237113,0.2390528321,-0.2087599188,-0.0982533023,-0.2538189292,0.3036789298,0.0383167453,0.2109606117,-0.1044602767,0.3884797394,0.0119993826,-0.0699289888,-0.0308246613,-0.0951843336,-0.4704369903,0.3934468925,0.0150187491,-0.135568589,0.0250264332,0.2128004134,0.3387266099,-0.1139988452,-0.1212918684,-0.2969061434,0.4140083492,0.2411861569,0.1252679527,-0.0791362524,0.2526995838,-0.4128444791,0.2862435281,0.1946456432,0.1356674284,-0.2306513935,-0.0101967659,0.2143626511,-0.0276604202,0.5676321983,0.0566371791,0.0776869804,0.3403141797,0.0193824731,-0.0210541133,-0.1676110923,0.546212852,0.0987184644,0.0150502147,0.2127047777,-0.0319072343,-0.0564370602,0.5448497534,0.0253492389,0.0208523683,0.2844620049,-0.2476340681,-0.0999411419,-0.0892350823,-0.2863973677,-0.0896072984,0.1175357625,0.0390074588,-0.2837412655,0.0367046148,-0.1309304982,0.0369652845,0.214072004,0.2218250334,0.0949605778,0.1513848454,-0.1968063265,-0.2821566164,0.0594320521,0.0114388391,0.1479180604,-0.1853646785,0.075736925,-0.4506164193,-0.1568676084,-0.3748697042,-0.2476449758,-0.5724465251,-0.378277868,-0.1889107078,0.0723680109,0.0509037264,0.0925706699,-0.0886129066,0.0059156343,-0.3178310692,0.0326358005,-0.3865585327,-0.4102072418,-0.4526662529,-0.1088003442,0.5771566033,0.151923418,0.0683279783,-0.0507632196,-0.2250634879,-0.1655669659,-0.3020949364,-0.1577259004,-0.2327002287,0.2592408657,-0.2609623969,0.3641046584,-0.0438750237,-0.1165322736,0.3952418864,-0.1275030226,0.0709459558,0.1280759424,0.1700384915,0.1128410324,0.0035246571,-0.4678861499,-0.358343482,-0.1989178509,-0.1360440254,0.2233173698,-0.0345684141,-0.1036565453,-0.1498719156,0.0007474066,-0.0653173178,-0.0321483277,-0.1435091347,0.3579942286,0.4484443069,-0.0049899849,-0.3743659556,0.1497588754,0.0967432335,0.0119670462,-0.0661526024,-0.5379254818,0.0361202471,-0.1953772306,-0.0451239944,0.0857066661,-0.0456166975,-0.0016662735,-0.1290929317,0.1255060732,-0.0732008666,-0.3776082993,-0.2028995603,-0.0218626857,0.1950786859,-0.0038253164,0.3918199539,-0.3685402274,0.6188606024,0.1327216029,0.0883559212,0.1921495497,0.195550248,-0.0770062804,-0.1292403191,-0.2143438458,0.1045699269,-0.1352505237,-0.0306809023,0.4353137314,0.0935673118,-0.470320344,-0.1284866631,0.2265396416,-0.354759872,-0.3016400933,0.2073978931,-0.0769027025,0.0912332833,-0.0251529608,0.004347994,-0.0949109942,-0.4156622589,-0.0764416382,0.226227507,0.3356660604,0.1157969013,-0.4246564209,-0.478282243,-0.2039898634,0.1658692062,0.1202297732,0.5696461797,-0.3223559558,-0.0003044132,0.2053594291,0.0649921149,0.4050049186,-0.2661755085,-0.0219342839,0.4310590029,0.0408875942,-0.5107001662,-0.093852073,-0.1408032924,-0.0567444712,0.4831960499,0.1177918911,-0.3134800196,-0.2617876232,-0.1015229821,0.5601978302,-0.2487789243,-0.0578359328,-0.51548177,-0.1347944885,-0.5276181102,-0.1725111753,-0.1721640825,0.1211989373,-0.1933709234,0.0334696323,0.1754015833,0.2532631159,0.1501697153,0.0883519575,0.1000320688,0.3204565644,-0.0820175782,0.0797392204,0.1786417216,-0.10248027,0.3204133809,0.0459115691,-0.1989246607,-0.0307989828,-0.1371486634,0.3870814145,0.102564536,0.1812932789,-0.0358679034,0.1762811989,-0.0651327968,-0.1640777886,0.1738256365,0.2553316951,-0.0520867072,-0.3193418682,-0.3537043333,0.6563943624,-0.0665214509,-0.0533949584,0.2429166585,-0.0126091084,-0.4147311747,0.4613971412,0.0934210047,0.9715640545,-0.3171394765,0.1448168755,-0.1763695031,0.0304253642,0.5069677234,0.0396601371,-0.0675011799,-0.274096936,0.3756844401,-0.0274236146,0.004009617,0.0358705856,0.1747505665,-0.1739991158,0.2709539533,-0.011291747,0.0670562163,0.0594732463,0.4678330421,0.253262192,-0.3550417125,-0.3557913601,0.0304090362,-0.2388121486,0.3524136543,-0.0693884715,0.0893608555,0.0727595389,0.0014335,-0.2758075893,0.2437413633,-0.0760525316,0.1676732749,-0.1628613025,-0.0867026001,0.3922525942,0.3192015588,0.0237114895,0.1971574873,-0.2213934064,0.2150328457,-0.245354414,-0.1812199801,-0.0596248358,0.1299696416,-0.0255760364,-0.005978344,-0.3947524428,-0.0175610445,0.0111840107,-0.0713633001,-0.4252784252,0.1303715408,0.186726436,0.096724838,-0.377607286,0.0327802934,-0.0694380105,-0.0047398317,0.079503864,0.0087578027,0.148255676,0.3051056266,0.4470944405,-0.296082139,-0.0706328824,0.3325310349,-0.1022136807,-0.1241696551,0.7613788843,0.5131797194,-0.3496618271,-0.2417659163,-0.2614117265,-0.4295880795,-0.4725286365,-0.2641901076,0.2634420693,0.3396519125,-0.1439696401,0.0455074161,-0.0183112342,-0.4574539065,-0.0471401699,-0.6610638499,0.0393412076,0.2092491388,-0.0146018732,-0.0092405584,0.0129330279,-0.3130429089,0.1598887295,0.1986601651,-0.1436172724,-0.0569279566,-0.2126782238,0.1778737903,0.3161405921,-0.0280059129,-0.2068298608,-0.0318055004,-0.0100409221,-0.2475087643,0.0706069395,-0.1326598227,0.0354800709,0.1784733832,0.166611895,-0.1664200872,-0.2140848786,-0.332883656,0.4437115192,-0.366109401,-0.1377566606,-0.160803467,-0.0988249183,0.290040195,0.111507006,0.1816402972,-0.0278570112,0.1566685587,-0.0921218768,0.0552503504,0.3821152747,0.0441885516,0.5659002066,0.1548046172,-0.2928225994,-0.0903576538,0.008749444,0.236912474,0.2314941436,-0.1619158685,0.2521121204,0.3678916991,-0.1160195693,0.4805481136,0.0019996453,-0.0912085623,0.268261224,0.0529294834,-0.2074800283,0.0183697827,0.7501780391,0.1219283417,-0.2131173909,0.1073694602,0.2034209371,-0.3128164709,0.1306407005,-0.0431428924,0.351377368,-0.1563558578,0.0493195578,-0.0115216859,0.2406993359,0.1322971731,0.0122128623,0.0604780093,-0.0349901393,0.1902000308,0.1454406381,0.179483071,-0.0096066566,0.1245269477,-0.2170787454,-0.3147494495,0.0829714313,0.4683617651,0.1482069641,0.1615779698,-0.0154721141,0.1307284981,0.0450580455,-0.2188173085,-0.2239313126,0.2758842111,0.093747057,-0.2819700241,-0.2536243796,0.0600826666,-0.1051290333,0.0900983289,0.0291752964,-0.2653272152,0.0694409609,0.4912320971,-0.1592136025,-0.4413060844,0.1000303328,-0.2558698654,0.0785363764,-0.1379108578,0.1403271258,0.0559332781,-0.0345929377,-0.0192603543,-0.0161289107,0.2129900753,0.3521134853,-0.4400802255,0.0628664717,-0.0654182658,0.1685570776,-0.1672759652,0.5017226338,0.1632207334,0.4650028348,0.2088683844,0.0879559889,-0.0324948579,-0.2866955101,0.2699802816,-0.1593201458,-0.0804591998,0.120146364,-0.0552680604,0.0659209788,-0.3422133923,-0.0335798413,-0.2467506826,0.2937825322,0.242466554,0.0778462589,0.0621773824,-0.0159029271,-0.012686722,-0.1081871986,0.6601423621,0.5105100274,0.291929841,-0.2291159332,-0.2408699989,-0.6080421209,-0.0933289081,-0.3303572237,-0.0876863673,0.2293175757,0.1256251484,0.1672124714,0.0515519083,0.0977146998,-0.0765203834,-0.1580450982,0.2019281238,-0.3592011929,-0.1981224567,-0.0247905366,0.1949086636,-0.3859387338,-0.1294591129,0.2386542708,-0.2436130196,-0.1144219562,-0.1381229907,-0.1011749953,0.0008640788,0.2453065068,0.2526297569,0.1153825969,0.6777515411,-0.0509353802,0.0873871818,-0.3224086761,-0.2137747258,0.1348675191,0.182836175,-0.354296118,0.2720734775,-0.1901830584,0.3042676449,-0.3099836111,0.3408549428,-0.0733300149,0.1433740556,-0.016136704,-0.2343817055,-0.0079792496,-0.1627307236,0.1087230444,0.2059752792,-0.0367440507,0.0858358815,-0.255394429,-0.0138550932,0.2228540033,-0.3553275764,-0.6054439545,-0.0821306854,0.1580117345,-0.0303069931,0.1140199229,-0.4135737717,-0.062473841,0.3374500275,0.0369989574,0.0575542375,0.2023040205,0.0482080169,-0.2381666601,-0.1790001243,-0.1499950737,0.1198802143,-0.1001126841,-0.1120221093,-0.173882544]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"I think the work around as suggested in the issue [#886] is not working for several languages, such as `id`. For example, I tried all the dates to download dataset for `id` langauge from the following link: (https:\/\/github.com\/huggingface\/datasets\/pull\/886) [https:\/\/dumps.wikimedia.org\/idwiki\/](https:\/\/dumps.wikimedia.org\/idwiki\/ )\r\n\r\n> >>> dataset = load_dataset('wikipedia', language='id', date=\"20210501\", beam_runner='DirectRunner')\r\nWARNING:datasets.builder:Using custom data configuration 20210501.id-date=20210501,language=id\r\nDownloading and preparing dataset wikipedia\/20210501.id (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/Users\/.cache\/huggingface\/datasets\/wikipedia\/20210501.id-date=20210501,language=id\/0.0.0\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 745, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 1139, in _download_and_prepare\r\n super(BeamBasedBuilder, self)._download_and_prepare(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 630, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/Users\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 420, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract({\"info\": info_url})\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py\", line 287, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py\", line 195, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/py_utils.py\", line 203, in map_nested\r\n mapped = [\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/py_utils.py\", line 204, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py\", line 218, in _download\r\n return cached_path(url_or_filename, download_config=download_config)\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py\", line 281, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/dumps.wikimedia.org\/idwiki\/20210501\/dumpstatus.json\r\n\r\nMoreover the downloading speed for `non-en` language is very very slow. And interestingly the download stopped after approx a couple minutes due to the read time-out. I tried numerous times and the results is same. Is there any feasible way to download non-en language using huggingface?\r\n\r\n> File \"\/Users\/miislamg\/opt\/anaconda3\/envs\/proj-semlm\/lib\/python3.9\/site-packages\/requests\/models.py\", line 760, in generate\r\n raise ConnectionError(e)\r\nrequests.exceptions.ConnectionError: HTTPSConnectionPool(host='dumps.wikimedia.org', port=443): Read timed out.\r\nDownloading: 7%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258e | 10.2M\/153M [03:35<50:07, 47.4kB\/s]","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":274,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nI think the work around as suggested in the issue [#886] is not working for several languages, such as `id`. For example, I tried all the dates to download dataset for `id` langauge from the following link: (https:\/\/github.com\/huggingface\/datasets\/pull\/886) [https:\/\/dumps.wikimedia.org\/idwiki\/](https:\/\/dumps.wikimedia.org\/idwiki\/ )\r\n\r\n> >>> dataset = load_dataset('wikipedia', language='id', date=\"20210501\", beam_runner='DirectRunner')\r\nWARNING:datasets.builder:Using custom data configuration 20210501.id-date=20210501,language=id\r\nDownloading and preparing dataset wikipedia\/20210501.id (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/Users\/.cache\/huggingface\/datasets\/wikipedia\/20210501.id-date=20210501,language=id\/0.0.0\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/load.py\", line 745, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 574, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 1139, in _download_and_prepare\r\n super(BeamBasedBuilder, self)._download_and_prepare(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/builder.py\", line 630, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/Users\/.cache\/huggingface\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 420, in _split_generators\r\n downloaded_files = dl_manager.download_and_extract({\"info\": info_url})\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py\", line 287, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py\", line 195, in download\r\n downloaded_path_or_paths = map_nested(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/py_utils.py\", line 203, in map_nested\r\n mapped = [\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/py_utils.py\", line 204, in <listcomp>\r\n _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/py_utils.py\", line 142, in _single_map_nested\r\n return function(data_struct)\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/download_manager.py\", line 218, in _download\r\n return cached_path(url_or_filename, download_config=download_config)\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py\", line 281, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/Users\/opt\/anaconda3\/envs\/proj\/lib\/python3.9\/site-packages\/datasets\/utils\/file_utils.py\", line 623, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/dumps.wikimedia.org\/idwiki\/20210501\/dumpstatus.json\r\n\r\nMoreover the downloading speed for `non-en` language is very very slow. And interestingly the download stopped after approx a couple minutes due to the read time-out. I tried numerous times and the results is same. Is there any feasible way to download non-en language using huggingface?\r\n\r\n> File \"\/Users\/miislamg\/opt\/anaconda3\/envs\/proj-semlm\/lib\/python3.9\/site-packages\/requests\/models.py\", line 760, in generate\r\n raise ConnectionError(e)\r\nrequests.exceptions.ConnectionError: HTTPSConnectionPool(host='dumps.wikimedia.org', port=443): Read timed out.\r\nDownloading: 7%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258e | 10.2M\/153M [03:35<50:07, 47.4kB\/s]","embeddings":[0.1800149828,-0.1523410529,-0.0989488661,0.4668173194,0.1006341428,0.2423329055,0.1486937255,0.1840970814,0.7133534551,-0.2562913597,-0.0293328445,0.0879321471,0.1379880607,-0.0704495609,0.0778176859,-0.2036105841,0.083321698,-0.1366672218,0.0640653819,-0.2263311893,-0.3286511898,0.3770771027,-0.3152113259,-0.033993151,-0.3056198657,0.2438495308,0.000596139,-0.1118378639,0.1123140231,-0.4247998893,0.2647848129,0.3021344244,0.312507391,0.175404191,-0.0001204848,-0.0556482673,0.704850018,-0.1893924773,-0.52306813,-0.2352164388,-0.0554581322,-0.4046289921,0.2281317711,-0.1170848757,-0.0360485911,-0.2035437673,0.1830866933,-0.390080601,0.1351653188,0.0376099162,0.1968803406,0.0330657512,0.1856203824,0.0348662511,0.2729909122,0.1124169081,0.1574663818,0.1152702719,-0.0042906459,-0.1470837593,0.0060088797,0.1632699221,-0.0448037684,-0.1061679423,0.2156065851,-0.0627294928,-0.1983726472,-0.6513531804,0.2012463361,0.2551749945,0.6670199037,-0.0362070054,-0.4080487788,-0.2918708622,0.008493525,0.0092041604,0.2186987996,0.2262768894,-0.1332885921,-0.0248336159,-0.0666920543,-0.2193876654,-0.0538428053,0.6153132915,-0.0989302322,0.3721018136,0.0412183702,0.152586624,0.1164748147,-0.2843738794,-0.3300948739,-0.1842395365,0.308218807,0.4228442907,-0.2161461562,0.3507112861,-0.0911564007,0.3997006714,0.0851336569,-0.1119415164,-0.2023396641,0.2832679749,-0.0947817862,0.1178443879,0.2353040576,-0.0526314341,0.2513489723,-0.084284395,0.3670358658,0.1552246362,-0.1794159114,0.0032228411,-0.2408719063,-0.2111485898,-0.2467986941,-0.1010616198,-0.056251429,-0.270691663,0.2040662915,0.1117742062,-0.2330821306,-0.288991034,-0.0754254609,0.2926601171,-0.1363919079,0.1686062813,0.1217710078,0.3821804821,-0.4291238785,-0.0510994084,0.0246106163,-0.0835722834,-0.1841303259,0.1816866994,0.4236653447,-0.0112116886,0.3521850109,-0.0009735774,-0.0124729387,-0.1143772751,-0.13916789,-0.1302518398,-0.0192519668,0.0520948581,0.1551325321,0.5195742249,0.0761604831,-0.2763779759,-0.2286849022,0.0425432995,-0.14053303,0.2529540062,0.0347801484,0.0855048075,-0.4118609428,-0.157736361,-0.2202835381,0.3061152697,0.1613567173,-0.1992474198,-0.0212133266,-0.2634349465,-0.1758718491,-0.0303868521,0.3723707497,0.7429201007,-0.2238100767,-0.1922197342,-0.0357124098,-0.0061537009,0.1992998868,0.246270895,-0.0146878511,0.2034340352,-0.1720418781,0.0657365918,0.2481043041,-0.1104744747,-0.0709453002,0.0185716581,0.2829237878,0.3354863524,-0.0245479904,0.0082481876,0.1751103103,0.0824777931,0.0755898803,0.4516778588,0.2592275441,-0.0364522003,-0.2878525555,-0.0955098271,0.4473000765,0.3634183109,0.1111965775,-0.1489389688,0.0520696752,0.3142115772,0.2028692216,-0.0173192602,0.1274423599,0.3553062975,-0.412081331,0.2675910592,0.1979930252,-0.0266612824,-0.6482115388,0.3039937019,0.0955562815,0.1926107109,-0.0523342416,0.0743491352,-0.3088146746,-0.176993385,-0.277630806,0.152644828,0.0550505407,0.1985759437,0.0841121599,0.4623348415,-0.0035472361,-0.0708874241,0.0051468941,-0.0730241761,-0.625194788,0.3269507587,0.0571064837,-0.0073050014,0.0225015488,0.2638174891,0.3264119327,-0.1936039031,-0.0540632159,-0.2034285814,0.291021198,0.341198355,0.1482165605,-0.0214505158,0.3269128799,-0.3324068785,0.2147027105,0.2711626291,0.109329395,-0.2314001322,-0.0771375597,0.3427499235,-0.176651448,0.5679537058,0.0784945339,0.0279703494,0.349321425,0.05606848,-0.1829757988,-0.2069450021,0.4769160151,0.0130715976,0.1207293496,0.2523759305,-0.2548974752,-0.0791640878,0.4247357845,0.0567899123,-0.0763744786,0.3021983802,-0.1847210675,-0.0454660542,0.0024955848,-0.1456357092,-0.1451688111,0.1742578447,0.1809597313,-0.3056588471,0.0288486108,-0.2651314437,0.0159962103,0.1129490957,0.093893528,0.1430916786,0.1447440237,-0.0515068546,-0.3176003098,0.0371384881,0.05871718,0.1647081524,-0.2192389965,0.0716491863,-0.4835647047,-0.4850093424,-0.2407976985,-0.2680956721,-0.5067273974,-0.4104590714,-0.1513090879,0.0539763756,0.0409823917,0.1727270186,-0.0090345284,-0.0408731215,-0.2526899874,-0.2420945019,-0.4165675938,-0.4249743819,-0.3751249611,-0.0479990691,0.5167554021,0.1541214287,0.1876805127,-0.1379991919,-0.2960053086,-0.2453897446,-0.3184579611,-0.1335677058,-0.2716222405,0.2574396431,-0.2281763256,0.363048166,-0.1546144634,-0.0545480475,0.439535886,0.0442846529,0.0154841458,0.1544584036,0.1573001891,0.208096832,-0.0512999631,-0.3745378554,-0.3137008548,-0.3135613203,0.0784837976,0.1661369205,-0.0123981154,0.1932970285,-0.1125321835,0.0462635607,-0.0381767265,0.0103718117,-0.1559996009,0.1697083861,0.4793511927,-0.1100016087,-0.393274188,0.0673182309,0.0674255788,0.1897398531,0.0910398215,-0.5468358994,0.1703419536,-0.1504279077,0.1138713583,0.2040820271,0.0073746983,-0.0466771312,-0.2249495387,0.1329428554,0.0290109813,-0.2154764235,-0.099813439,-0.1156409532,0.1483947337,0.1870363951,0.3458723724,-0.3495883346,0.7946072221,0.3428243697,0.0824462622,0.3724907339,0.1699153334,0.1285421401,-0.1782819033,-0.3313499689,0.0166985262,-0.0605239123,0.0615107417,0.3438498974,0.0552456863,-0.3861548305,-0.2321020365,0.0446626581,-0.2911091745,-0.341812551,0.1781959832,-0.1911976188,0.2100485265,-0.0755979121,0.0897875428,-0.0583659522,-0.3991448581,-0.0373492949,0.2391900569,0.3913217783,0.2014875263,-0.4114879072,-0.4465774298,-0.3870896101,0.2356893271,-0.0073544085,0.4826688468,-0.2531711459,0.0031656921,0.1816846877,0.1617052853,0.3192493618,-0.2490258962,-0.1349231005,0.4210767448,-0.1227573231,-0.7164517641,0.0497482307,-0.0690234303,0.0806783065,0.3638060093,0.3140451908,-0.5319975019,-0.3132551014,-0.0475540012,0.5816972852,-0.1379837245,-0.1933632642,-0.3705926239,-0.1753130853,-0.510584116,-0.1583568752,-0.1646462381,0.1698280126,-0.1786434799,0.0079970658,-0.018816039,0.2356257588,0.077410832,0.1168572828,0.1895963252,0.3470451534,0.0704391077,0.2878077328,0.0471596457,-0.0871025324,0.393461585,-0.0413503461,-0.225859791,-0.0686275735,-0.0754133388,0.2024580836,0.1237270236,0.1806011349,0.0705815405,0.1341579109,-0.1330598742,-0.139954716,0.0249319617,0.1640877128,-0.0207960308,-0.336977452,-0.4289216995,0.629592061,0.051031813,-0.0160415117,0.2564948797,0.0874255002,-0.4474819899,0.3799237907,-0.0563119985,0.8428797126,-0.2478800267,0.1902820319,-0.0803552568,0.0971005782,0.5161353946,0.1319610775,-0.0354957953,-0.2667683363,0.371740967,0.0381695814,0.0976983011,-0.0479693785,0.065086484,-0.1759572178,0.3413991928,-0.1395687312,0.0184768271,-0.0198481679,0.537340045,0.0486165285,-0.4129085839,-0.386579752,0.0466350764,-0.1613312215,0.4640792012,-0.0780518353,0.0684453398,-0.024586428,-0.0271558687,-0.290538311,0.0905472636,-0.1602765024,0.1231572032,-0.1205277964,-0.3054324687,0.319547534,0.4627718925,0.1373737305,0.1259135902,-0.2365532368,0.1082673967,-0.2404059023,-0.2273076773,-0.0074580358,0.054956425,0.0817355663,-0.0056701694,-0.3299396038,-0.0615572147,-0.0077630426,0.0784517676,-0.429923296,0.1066842675,-0.0029672214,0.0635950193,-0.3002030849,-0.0961654857,-0.0863905475,-0.0379311517,0.0927072167,0.0291343406,0.1981609315,0.0762013793,0.3280578256,-0.25856933,-0.135601759,0.3534421325,-0.0675859004,-0.0445446968,0.7959251404,0.3751738667,-0.3174501956,-0.1682408005,-0.2172058821,-0.3076891303,-0.5152919888,-0.2519545853,0.1604455262,0.3308833241,-0.2394684106,0.0721554607,0.0040699379,-0.4090076685,0.0556098409,-0.6211547256,-0.0471359529,0.3892129958,0.0198453609,0.0419546664,-0.0981767923,-0.3809345663,0.1713947207,0.0565869361,-0.1696271896,-0.0085768998,-0.1524821222,0.1862353086,0.4146291316,0.0372427776,-0.1246515587,-0.1072086468,0.0359362885,-0.1875638217,0.0902870297,-0.148058176,-0.0431228317,0.1577670723,0.1764922589,-0.1455526799,-0.1077943295,-0.3282791078,0.3568211496,-0.4021494091,-0.0671140477,-0.0220964011,-0.0244817454,0.334454,-0.0031857789,0.2768300772,-0.0731008202,0.1625114083,0.0563128851,0.0489860773,0.2679505944,0.0181406774,0.3847280741,0.1170411631,-0.3176272511,-0.0271483269,-0.0381819755,0.1088418514,0.2414458096,-0.1534071267,0.2959816456,0.4175151885,-0.0091351261,0.5071926713,0.0014607772,-0.0576087013,0.3259932101,0.0661523193,-0.1879680902,0.0828081369,0.6814025044,0.0829468295,-0.1184814945,0.2207801491,0.2282129079,-0.2920019031,0.1522591114,-0.0557174459,0.2191081345,-0.1131620482,0.2034062296,0.0151080964,0.2112383544,0.106737271,0.0779896602,0.1231906563,-0.112757653,0.0520134866,0.014187742,0.169598937,-0.0638837889,0.2166299224,-0.1179412752,-0.3135267794,0.0741367787,0.3286192417,-0.0034967174,0.216874063,-0.0233666487,0.3721559942,0.0895652771,-0.1539684236,-0.1557576656,0.2957776487,0.0512530394,-0.2696965933,-0.2077621371,-0.1345248669,-0.1193171963,0.0236816891,-0.0822676718,-0.3273004293,0.0912085623,0.3711731136,-0.1423135698,-0.6611459851,0.3607368469,-0.1965745538,0.2174201161,-0.1351703256,0.2060368508,0.0114979707,-0.0433632955,0.0537618585,0.163427189,0.2086670697,0.2028922737,-0.2874934673,0.0840255916,-0.0873381943,0.130036667,-0.1065585166,0.5063302517,0.2059764266,0.2298507094,0.3066175282,0.0775057971,-0.065345414,-0.3077969849,0.2216909528,0.017084403,-0.2135260105,0.1237168759,-0.2414343804,0.0527078919,-0.235443905,-0.1045847684,-0.3741672933,0.2458755821,0.1488333791,0.2014051378,0.1116238758,-0.1091585532,0.016785318,-0.1219202057,0.6659029722,0.4709475338,0.242037788,-0.3118558228,-0.2013340741,-0.6826386452,-0.1007702053,-0.2332904339,0.0286014099,0.206920594,0.1498171538,0.1325330138,0.0392778665,0.0202885736,-0.2368140072,-0.0766944587,0.1923868656,-0.3165422976,-0.150176242,-0.2111931145,0.076106742,-0.3056559265,-0.0772878081,0.2012035996,-0.2959721386,-0.1056284755,-0.0678512603,0.0083138663,0.0190564562,0.1428830475,0.1978755444,0.0781710595,0.6389302611,0.0099746147,0.0193108376,-0.3615393341,-0.2315559983,0.0306947567,0.4513472915,-0.3157905042,0.3092701733,-0.1534179151,-0.0080641843,-0.2956069112,0.3202164471,-0.1158288568,0.0076612709,0.0380929299,-0.1818344742,0.007573891,-0.147587806,0.1352748722,0.3294763565,-0.0791000873,0.0743841603,-0.3277178109,-0.1718485653,0.2044254094,-0.4374215007,-0.3669041991,0.045949854,0.0716151372,0.21265921,0.086010851,-0.4164581299,0.0903612971,0.3952999413,-0.0862375721,0.0810386762,0.1426228136,0.0373411961,-0.1932459027,-0.229464069,-0.1520406604,0.1861823499,-0.1080208868,-0.0250993613,-0.1390414834]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"Hi ! The link https:\/\/dumps.wikimedia.org\/idwiki\/20210501\/dumpstatus.json seems to be working fine for me.\r\n\r\nRegarding the time outs, it must come either from an issue on the wikimedia host side, or from your internet connection.\r\nFeel free to try again several times.","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":40,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nHi ! The link https:\/\/dumps.wikimedia.org\/idwiki\/20210501\/dumpstatus.json seems to be working fine for me.\r\n\r\nRegarding the time outs, it must come either from an issue on the wikimedia host side, or from your internet connection.\r\nFeel free to try again several times.","embeddings":[0.1920681745,-0.2602579296,-0.1416171342,0.3762246966,0.1157500148,0.2121071965,0.1645559072,0.1551883221,0.7062768936,-0.2246863097,0.11798051,0.0496410765,0.2343845218,-0.2096705884,-0.0094013372,-0.2210128754,0.0925963148,-0.195391655,0.0648471639,-0.2729018927,-0.2977039218,0.2671948969,-0.397896409,0.027171284,-0.1857133359,0.1696379483,0.0080057522,-0.1815737188,0.0773589015,-0.3585591912,0.1663145572,0.2452066392,0.3375380635,0.1511198282,-0.0001189297,-0.1920975596,0.6491462588,-0.1044014916,-0.4709958136,-0.0881614313,-0.1650997996,-0.3635079265,0.2519365549,-0.1299187392,0.0853196234,-0.1647308618,0.2627398074,-0.4774175286,0.1545994282,0.0544664823,0.1764225662,-0.0442770943,0.0150865419,0.1050192267,0.1392659098,0.0897769555,0.1414282769,0.0952253267,0.1480285823,-0.332827121,0.0266098175,0.1536878496,-0.1374328434,-0.1619034559,0.2353250384,-0.1162593514,-0.2261555344,-0.671456337,0.1671416461,0.2753604352,0.6968308091,0.1771871299,-0.3633895814,-0.2033320814,0.0303305909,0.1108860597,0.2197708786,0.3188805282,-0.1332755089,-0.0756762326,-0.0223537441,-0.1838998497,-0.0571159758,0.6162976027,-0.0327543244,0.4338471293,0.1757819206,0.2188185155,0.0114605315,-0.2223478407,-0.2297370881,-0.2822299302,0.2777375877,0.2912098467,-0.229287982,0.3861026764,-0.0092938757,0.3389609158,-0.0212205537,-0.1730937809,-0.1703128666,0.2720246017,-0.0565527044,0.1500321776,0.2194157392,-0.1049375907,0.2648623884,-0.05036043,0.3448389471,0.0464350432,-0.2778557837,-0.0219871625,-0.2163523585,-0.1501250714,-0.1527246833,-0.1844341904,-0.1040594503,-0.3015193343,0.2841693759,0.117397055,-0.1959318817,-0.2869795561,-0.1895691007,0.2224102914,-0.1789711416,0.3431453407,0.2166811675,0.318279326,-0.4698387384,-0.0308829788,0.0582000241,-0.0295045059,-0.2047911286,0.2661013007,0.2671023309,0.0934590027,0.3063895106,-0.0349103101,0.0334962569,-0.1906967908,0.0378600135,-0.19088763,0.0086274091,-0.0087980423,0.1492285728,0.6243290901,0.1069346964,-0.240696758,-0.1143351868,0.0173425358,-0.1332241297,0.3218250871,-0.0038698744,0.1082379073,-0.2380398959,-0.0695936158,-0.2211727202,0.2241665572,0.1904148757,-0.2009399086,-0.0108837662,-0.3347891271,-0.1034783125,-0.0116115492,0.2275258154,0.5787898898,-0.2940459847,0.0078724902,-0.0782736838,0.2460217774,0.2944015265,0.2770064175,-0.1376061738,0.224449873,-0.1103304029,0.1050942391,0.2300578058,-0.0963964611,0.063338019,-0.0674745888,0.2769455016,0.2506719828,-0.101845704,0.07155402,0.3002308309,0.0578249432,0.1811822504,0.5470150113,0.1700084656,-0.0198371634,-0.3321273029,0.0105104512,0.4842863977,0.2271985263,0.0975403637,-0.2832230031,-0.0301945675,0.3478222191,0.1292705387,0.0006163185,0.1187987328,0.3935267627,-0.4654506445,0.2700213194,0.3104889095,-0.0189492647,-0.4321421385,0.220550999,0.1768899113,0.2310065031,-0.0370158181,0.1332352906,-0.2768265903,-0.0912002176,-0.2021674514,0.2402665019,0.078774482,0.1118167341,0.0146720903,0.5503216386,0.0871952623,-0.234334752,-0.0230131373,-0.0929099247,-0.5333667994,0.4093146026,0.0253338981,-0.0747516975,0.0875979662,0.1141729727,0.3228402138,-0.156650275,-0.1230947673,-0.2445740551,0.24396725,0.3421175778,0.176921621,-0.0512005314,0.2956596017,-0.4270628095,0.2064131796,0.230202809,0.2040228695,-0.294174403,-0.0031867428,0.2778055966,-0.1314051747,0.579092443,0.0430002473,0.0254932567,0.3273768723,0.0562182739,-0.1040597111,-0.1395439655,0.4879278839,0.0925179869,-0.1234057173,0.2163667828,-0.1414022297,0.0244224835,0.5701662898,0.0576848425,-0.1411892027,0.174388811,-0.1842634827,-0.2130209208,0.0451659337,-0.1988901794,-0.1464824677,0.1640647352,0.119608447,-0.4188327193,0.0040637571,-0.1981055439,0.0572085343,0.0680594742,0.1505713761,0.1231385544,0.1096363813,-0.0984696671,-0.2919638753,-0.0995791927,0.0575046167,0.1817233711,-0.125602901,0.1521684825,-0.3066744208,-0.335704565,-0.142748028,-0.2836947143,-0.4758937657,-0.5971008539,-0.1310851127,0.0254992601,0.0646237209,0.1711786985,-0.0614617653,-0.0896035582,-0.1060794964,-0.1029613987,-0.4092002809,-0.4708696604,-0.3191872537,-0.0478072725,0.4986441135,0.2129051983,0.0927867144,-0.0574292466,-0.3174837828,-0.1957260519,-0.2670004964,-0.1926907897,-0.2869168222,0.1864595413,-0.302075088,0.4951706231,-0.1321394891,-0.1355850101,0.4025033116,-0.0661303475,0.1005680189,0.1314277947,0.1966309994,0.1588290632,-0.0201147087,-0.3963410854,-0.2530008256,-0.2210238278,-0.1711599678,0.0928545967,-0.0977484211,-0.0573782437,-0.0299816933,-0.0126924701,0.0298288558,-0.0714874566,-0.0779576674,0.2201895714,0.5501049161,-0.0635473356,-0.4144967198,0.1688749641,0.0828099772,0.1014464945,0.039773535,-0.5010535121,0.1527956128,-0.2344356179,0.0599981397,0.1821861565,-0.083700113,-0.0569189191,-0.2782068849,0.0634085312,0.0754753873,-0.1482956409,-0.1136238128,-0.0790371597,0.1974343807,0.0912215263,0.3415472507,-0.3307110071,0.6799039841,0.1604367793,0.0958411172,0.1546342075,0.2151065469,-0.0109759877,-0.1354507953,-0.2767878771,0.0811419487,-0.0431491956,-0.0215320047,0.2655888498,0.0784381106,-0.4907026887,-0.1525413245,0.1452765316,-0.321346581,-0.3399422467,0.1406853497,-0.2062221617,0.1402599663,0.0673843846,0.0754931048,-0.0615367293,-0.417648226,-0.0513381213,0.1821236759,0.3251450956,0.0753133148,-0.4876290858,-0.4883402884,-0.3488652706,0.1095598489,0.0329043679,0.5476915836,-0.2390423864,0.0497859791,0.2448563427,0.1036564782,0.2784717977,-0.2613422871,-0.0725740641,0.3867153823,0.0527335219,-0.5512059331,0.0222637318,-0.069646053,0.0761990249,0.3663045764,0.1582707614,-0.4561689198,-0.251511991,-0.1542205065,0.5933499932,-0.1971588433,-0.1010307595,-0.3331771195,-0.2064864933,-0.4048658311,-0.1718534529,-0.1741798967,0.1794888079,-0.2179675996,-0.051005125,0.0569905229,0.3927876949,0.1286990941,0.1028494388,0.0724306181,0.1700817645,-0.0082414886,0.1224792004,0.03874854,-0.0678741857,0.2467940748,0.1005030945,-0.0703677014,-0.0824481994,-0.0939305201,0.2078879923,0.0814419389,0.2122065723,0.1155673265,0.0986324027,0.0036280567,-0.0799865052,-0.0119700693,0.2462151796,0.0859593973,-0.3830393553,-0.3961808681,0.4086367786,0.0116261598,-0.0380274989,0.4489441514,0.0282257367,-0.3888244033,0.4838344753,0.0946419314,0.969438076,-0.3626310229,0.2214490324,-0.0437711515,0.2138673365,0.4991881549,0.1092955321,-0.0625862256,-0.3294620216,0.3004903197,-0.0175757967,0.0838006362,-0.0148678971,0.0289212372,-0.1912825406,0.2490026355,-0.0175562501,0.0332377888,-0.1363829225,0.4640840888,0.1043737009,-0.3234977424,-0.4103520513,0.0503707863,-0.1101789847,0.2641601861,-0.0269806497,0.1360906661,-0.0234080218,-0.0133636752,-0.2455650717,0.1273057014,0.0046779914,0.1741744876,-0.2429156303,-0.1915960312,0.3354978561,0.496309042,-0.040124815,0.2873746455,-0.1373690069,0.1501461715,-0.3548809886,-0.268871963,-0.072710678,0.0108712623,-0.0164028127,0.04684655,-0.4290374517,0.189470008,-0.0311859567,0.0892986357,-0.4014641047,0.1624378562,-0.0486572459,0.0198917259,-0.3122407198,-0.0576990359,-0.1851567179,-0.0049890135,0.096187748,0.0825844035,0.1815800667,0.1937353313,0.4425471723,-0.2791804969,-0.0839969814,0.3254942298,-0.0386321284,0.0233748443,0.8597972989,0.5083207488,-0.4090457857,-0.2138406336,-0.1994021684,-0.4488832653,-0.3732959628,-0.1985151619,0.188700825,0.4621225297,-0.1589322388,-0.0379793495,0.0002930261,-0.4264071286,0.0035265896,-0.7153940797,0.0773031414,0.3578203022,-0.0380160324,-0.0327807479,-0.1383401006,-0.5434535146,0.1512978226,0.1065283939,-0.1819632947,-0.0730728954,-0.3117240071,0.2369899601,0.2586434484,-0.0669716373,-0.1493398249,-0.1213868931,0.041902747,-0.2143432796,0.1584369689,-0.1586949378,0.0270057097,0.1492963284,0.1782796532,-0.2000199705,-0.1121389717,-0.4700894952,0.4432995617,-0.3254799545,-0.1196746677,-0.1218463555,-0.1160033196,0.2512051165,0.040689569,0.2366195321,-0.0806230083,0.1934547722,-0.0462981835,0.1295404285,0.2762299478,0.0621566884,0.6145495176,0.1069134027,-0.5067756176,0.0503829494,-0.1780497283,0.1562207639,0.4244710803,-0.1226939708,0.354563266,0.3840480745,-0.0265253168,0.5288011432,-0.0064136838,-0.0363181978,0.2458725125,0.077463761,-0.1324833483,0.0445871949,0.6764924526,-0.0212143753,-0.1660683304,0.1263344735,0.0910311639,-0.4747356772,0.0532801561,-0.0454299822,0.2685058713,-0.1508270949,0.164114669,0.0040681898,0.2596120238,0.1298170835,0.072023429,0.1332854182,-0.1690908521,-0.0071402094,0.0926575288,0.1842784584,-0.0161300208,0.2660954297,-0.2436091006,-0.2792325616,0.0486121438,0.3852074146,-0.0197930336,0.1365385652,0.028611999,0.1218474582,0.0750943124,-0.1625012606,-0.0949182436,0.3073095679,0.0666261092,-0.2399256229,-0.2406448126,-0.0946002901,-0.0292487871,0.044901032,0.079350926,-0.1638033837,0.0914223865,0.445600301,-0.0483419858,-0.5520182252,0.4726909101,-0.2859683335,0.2255390137,-0.2370163351,0.0579245687,0.0388138257,-0.0223017596,0.0073356535,0.0700337589,0.1823276579,0.2420694381,-0.3882776499,0.1533538252,-0.0922935978,0.0992061868,0.0095501728,0.5086563826,0.1092301086,0.2271664143,0.2142227739,0.0890798718,-0.1114998832,-0.3165863752,0.2277709842,-0.0068163141,-0.0732114986,-0.0116788521,-0.1161157787,0.0383870453,-0.2987958491,-0.0176453758,-0.2505201697,0.1779468656,0.188217625,0.1119516194,0.0408331119,-0.2005807161,0.0125092361,-0.1838635653,0.7236377597,0.585983336,0.2835956216,-0.2090541124,-0.2439518869,-0.455547601,-0.081756182,-0.4325778484,-0.0857660249,0.2726193964,0.1955885887,0.1392155439,0.0434404574,0.1211521924,-0.0834992528,-0.192889601,0.3312895,-0.3818550408,-0.2281175554,-0.1718607396,0.1640768796,-0.4031395614,-0.1043485403,0.2561421692,-0.3159589469,-0.0706806183,-0.0782782659,-0.0037422285,-0.0126484679,0.2914525867,0.1636818945,0.1555778682,0.5637648702,0.1061396748,0.0176236909,-0.2532347739,-0.2525511682,0.1693687588,0.3212299645,-0.2114066482,0.3466905951,-0.1603897661,0.0920932367,-0.1886446923,0.3205472827,-0.2373628914,0.0067877513,0.0499625094,-0.243648842,0.0094702272,-0.0830182731,0.0681333765,0.2913365364,-0.02881689,0.0511911698,-0.1973848045,-0.132645309,0.1089980602,-0.3755273223,-0.5066684484,0.054144159,-0.0013109103,0.1312768906,0.1871966571,-0.3093905449,-0.0325960554,0.3663615882,-0.006638214,0.0827036202,0.1522185057,0.0932112634,-0.0943486542,-0.2015009224,-0.1099107116,0.0467755497,-0.0614159256,-0.1624554247,-0.1030104086]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"I was trying to download dataset for `es` language, however I am getting the following error:\r\n```\r\ndataset = load_dataset('wikipedia', language='es', date=\"20210320\", beam_runner='DirectRunner') \r\n```\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20210320.es (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/scratch\/user_name\/datasets\/wikipedia\/20210320.es-date=20210320,language=es\/0.0.0\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1...\r\nTraceback (most recent call last):\r\n File \"apache_beam\/runners\/common.py\", line 1233, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1368, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"\/scratch\/user_name\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 492, in _clean_content\r\n text = _parse_and_clean_wikicode(raw_content, parser=mwparserfromhell)\r\n File \"\/scratch\/user_name\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 548, in _parse_and_clean_wikicode\r\n section_text.append(section.strip_code().strip())\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/wikicode.py\", line 639, in strip_code\r\n stripped = node.__strip__(**kwargs)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 60, in __strip__\r\n return self.normalize()\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 150, in normalize\r\n return chr(htmlentities.name2codepoint[self.value])\r\nKeyError: '000nbsp'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"download_dataset_all.py\", line 8, in <module>\r\n dataset = load_dataset('wikipedia', language=language, date=\"20210320\", beam_runner='DirectRunner') \r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1152, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/pipeline.py\", line 564, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/direct\/direct_runner.py\", line 131, in run_pipeline\r\n return runner.run_pipeline(pipeline, options)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 190, in run_pipeline\r\n pipeline.to_runner_api(default_environment=self._default_environment))\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 200, in run_via_runner_api\r\n return self.run_stages(stage_context, stages)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 366, in run_stages\r\n bundle_context_manager,\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 562, in _run_stage\r\n bundle_manager)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 602, in _run_bundle\r\n data_input, data_output, input_timers, expected_timer_output)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 903, in process_bundle\r\n result_future = self._worker_handler.control_conn.push(process_bundle_req)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/worker_handlers.py\", line 378, in push\r\n response = self.worker.do_instruction(request)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/worker\/sdk_worker.py\", line 610, in do_instruction\r\n getattr(request, request_type), request.instruction_id)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/worker\/sdk_worker.py\", line 647, in process_bundle\r\n bundle_processor.process_bundle(instruction_id))\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/worker\/bundle_processor.py\", line 1001, in process_bundle\r\n element.data)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/worker\/bundle_processor.py\", line 229, in process_encoded\r\n self.output(decoded_value)\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 356, in apache_beam.runners.worker.operations.Operation.output\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 358, in apache_beam.runners.worker.operations.Operation.output\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 717, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 718, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 1235, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"apache_beam\/runners\/common.py\", line 1233, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 717, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 718, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 1235, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"apache_beam\/runners\/common.py\", line 1233, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 717, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 718, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 1235, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1315, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/future\/utils\/__init__.py\", line 446, in raise_with_traceback\r\n raise exc.with_traceback(traceback)\r\n File \"apache_beam\/runners\/common.py\", line 1233, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1368, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"\/scratch\/user_name\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 492, in _clean_content\r\n text = _parse_and_clean_wikicode(raw_content, parser=mwparserfromhell)\r\n File \"\/scratch\/user_name\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 548, in _parse_and_clean_wikicode\r\n section_text.append(section.strip_code().strip())\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/wikicode.py\", line 639, in strip_code\r\n stripped = node.__strip__(**kwargs)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 60, in __strip__\r\n return self.normalize()\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 150, in normalize\r\n return chr(htmlentities.name2codepoint[self.value])\r\nKeyError: \"000nbsp [while running 'train\/Clean content']\"\r\n```","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":481,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nI was trying to download dataset for `es` language, however I am getting the following error:\r\n```\r\ndataset = load_dataset('wikipedia', language='es', date=\"20210320\", beam_runner='DirectRunner') \r\n```\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20210320.es (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/scratch\/user_name\/datasets\/wikipedia\/20210320.es-date=20210320,language=es\/0.0.0\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1...\r\nTraceback (most recent call last):\r\n File \"apache_beam\/runners\/common.py\", line 1233, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1368, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"\/scratch\/user_name\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 492, in _clean_content\r\n text = _parse_and_clean_wikicode(raw_content, parser=mwparserfromhell)\r\n File \"\/scratch\/user_name\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 548, in _parse_and_clean_wikicode\r\n section_text.append(section.strip_code().strip())\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/wikicode.py\", line 639, in strip_code\r\n stripped = node.__strip__(**kwargs)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 60, in __strip__\r\n return self.normalize()\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 150, in normalize\r\n return chr(htmlentities.name2codepoint[self.value])\r\nKeyError: '000nbsp'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"download_dataset_all.py\", line 8, in <module>\r\n dataset = load_dataset('wikipedia', language=language, date=\"20210320\", beam_runner='DirectRunner') \r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/load.py\", line 748, in load_dataset\r\n use_auth_token=use_auth_token,\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 575, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/datasets\/builder.py\", line 1152, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/pipeline.py\", line 564, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/direct\/direct_runner.py\", line 131, in run_pipeline\r\n return runner.run_pipeline(pipeline, options)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 190, in run_pipeline\r\n pipeline.to_runner_api(default_environment=self._default_environment))\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 200, in run_via_runner_api\r\n return self.run_stages(stage_context, stages)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 366, in run_stages\r\n bundle_context_manager,\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 562, in _run_stage\r\n bundle_manager)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 602, in _run_bundle\r\n data_input, data_output, input_timers, expected_timer_output)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/fn_runner.py\", line 903, in process_bundle\r\n result_future = self._worker_handler.control_conn.push(process_bundle_req)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/portability\/fn_api_runner\/worker_handlers.py\", line 378, in push\r\n response = self.worker.do_instruction(request)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/worker\/sdk_worker.py\", line 610, in do_instruction\r\n getattr(request, request_type), request.instruction_id)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/worker\/sdk_worker.py\", line 647, in process_bundle\r\n bundle_processor.process_bundle(instruction_id))\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/worker\/bundle_processor.py\", line 1001, in process_bundle\r\n element.data)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/apache_beam\/runners\/worker\/bundle_processor.py\", line 229, in process_encoded\r\n self.output(decoded_value)\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 356, in apache_beam.runners.worker.operations.Operation.output\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 358, in apache_beam.runners.worker.operations.Operation.output\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 717, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 718, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 1235, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"apache_beam\/runners\/common.py\", line 1233, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 717, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 718, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 1235, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"apache_beam\/runners\/common.py\", line 1233, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 717, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/worker\/operations.py\", line 718, in apache_beam.runners.worker.operations.DoOperation.process\r\n File \"apache_beam\/runners\/common.py\", line 1235, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 1315, in apache_beam.runners.common.DoFnRunner._reraise_augmented\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/future\/utils\/__init__.py\", line 446, in raise_with_traceback\r\n raise exc.with_traceback(traceback)\r\n File \"apache_beam\/runners\/common.py\", line 1233, in apache_beam.runners.common.DoFnRunner.process\r\n File \"apache_beam\/runners\/common.py\", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process\r\n File \"apache_beam\/runners\/common.py\", line 1368, in apache_beam.runners.common._OutputProcessor.process_outputs\r\n File \"\/scratch\/user_name\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 492, in _clean_content\r\n text = _parse_and_clean_wikicode(raw_content, parser=mwparserfromhell)\r\n File \"\/scratch\/user_name\/modules\/datasets_modules\/datasets\/wikipedia\/2fe8db1405aef67dff9fcc51e133e1f9c5b0106f9d9e9638188176d278fd5ff1\/wikipedia.py\", line 548, in _parse_and_clean_wikicode\r\n section_text.append(section.strip_code().strip())\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/wikicode.py\", line 639, in strip_code\r\n stripped = node.__strip__(**kwargs)\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 60, in __strip__\r\n return self.normalize()\r\n File \"\/opt\/conda\/lib\/python3.7\/site-packages\/mwparserfromhell\/nodes\/html_entity.py\", line 150, in normalize\r\n return chr(htmlentities.name2codepoint[self.value])\r\nKeyError: \"000nbsp [while running 'train\/Clean content']\"\r\n```","embeddings":[0.1763544828,-0.1677631289,-0.1718394458,0.4323384762,0.1707598567,0.2831906378,0.1634615809,0.2103996426,0.7203700542,-0.2296734005,0.0827716962,0.0324949548,0.0824812278,-0.1710957587,0.1047614142,-0.1762412637,0.040906731,-0.1311677247,0.0619521402,-0.3332505226,-0.3264745176,0.2989555299,-0.2558051348,0.0192325842,-0.2063501179,0.2564130425,0.0552222505,-0.1168860421,0.0501078293,-0.3417122662,0.1863566786,0.2099172026,0.3298726082,0.1184753329,-0.0001155041,-0.0839764923,0.6460733414,-0.1828085482,-0.4777913392,-0.2485990077,-0.1580174863,-0.4117628932,0.1903592348,-0.1931241453,0.085821338,-0.2423582971,0.2660349309,-0.5352807045,0.1104979366,0.1375032067,0.2382594645,-0.1910074055,0.0760797933,0.071904026,0.3408456445,0.0599353276,0.1818883121,0.1272431165,0.0240778606,-0.1747232676,0.0576144196,0.1426581889,-0.1027082428,-0.0544828437,0.1305572391,-0.0746287107,-0.093501769,-0.6982363462,0.1663738638,0.2483552247,0.7527777553,0.0537605397,-0.3433137238,-0.1573726535,0.0807394534,0.0541956611,0.1600305289,0.2955708802,-0.1713262051,-0.08595182,-0.0020556361,-0.2566790283,-0.0361364037,0.6079903245,-0.0604825802,0.3638177514,0.1059381887,0.1484155953,0.0372281857,-0.2130876333,-0.2027879804,-0.208680585,0.3671191633,0.3883555532,-0.2663078606,0.4091638923,-0.0272477381,0.2662196159,0.0512345843,-0.1118468046,-0.171297729,0.2166763991,-0.1110823005,0.1603291929,0.2141034156,-0.1045850217,0.2297871709,-0.0994358137,0.282810539,0.0523053743,-0.10580156,0.0235038307,-0.2767423987,-0.2106363624,-0.1899599135,-0.0080819987,-0.0186589397,-0.3135660589,0.1164932996,0.1050030962,-0.2387083322,-0.2927421331,-0.1104368269,0.2609954178,-0.1404688805,0.2598820925,0.1735314131,0.3614093363,-0.4104596376,-0.0088007478,-0.0065336893,-0.0349614099,-0.2180782855,0.1077685952,0.3512766659,0.1387448013,0.3635692,-0.044670552,-0.0229277033,-0.1345911473,-0.0369594246,-0.2337093353,-0.0712124258,-0.0027836114,0.1963965744,0.5106027722,0.0689720139,-0.2143213302,-0.1292956918,0.1026896834,-0.136296019,0.333781451,-0.0152218118,0.1500983089,-0.3008658588,-0.1998071223,-0.2625985444,0.2575986981,0.1306094229,-0.1939145327,-0.0293582138,-0.3026609421,-0.1415430009,-0.0959850922,0.320959121,0.5890487432,-0.2810034454,-0.1207376346,-0.1368455738,0.0293474011,0.2669731379,0.2324832678,-0.0744377524,0.2620208263,-0.1818814278,0.0912397951,0.3925758302,-0.0914892703,-0.033330705,0.0226022806,0.2543193102,0.2409895658,-0.0208703689,0.0407411717,0.185290426,0.1152122691,0.0883842483,0.5103183985,0.1995154321,-0.0357514285,-0.2663413584,-0.0222022217,0.5889864564,0.3162772059,0.1695824713,-0.1865864843,0.0164659489,0.3192102909,0.2271896303,-0.0213697348,0.1035802662,0.3533834219,-0.3805092871,0.3051334023,0.1942701191,-0.0205473118,-0.5447943807,0.2211422026,0.0312379785,0.1221075654,-0.0230360944,0.1085037813,-0.3356630504,-0.1956944764,-0.210438922,0.2059174925,0.1153013036,0.2285695523,-0.0890166312,0.4337516129,0.0359382555,-0.074444063,0.0314512327,-0.0982374176,-0.5547204018,0.4431158006,0.0008211986,-0.0520538166,0.0469282679,0.1857425421,0.2999597788,-0.1909607202,-0.1065998077,-0.2020519823,0.3487930894,0.1899758577,0.1270586699,-0.1278118193,0.253605634,-0.3870398104,0.1879015118,0.3023718596,0.1857383102,-0.1931846142,-0.0406670645,0.1949964166,-0.12047977,0.4849776924,0.045072291,0.0869900361,0.3448004425,0.0223699026,-0.0912765786,-0.1143773645,0.5354329348,0.0885521397,0.0355971567,0.2150045931,-0.1734351814,-0.1120484024,0.5037209988,0.112896882,0.0008994567,0.2261939347,-0.1032246575,-0.0716235861,-0.01897466,-0.170152083,-0.1271852851,0.1564020216,0.0838888958,-0.3795368075,0.0664776638,-0.1950324625,0.001023291,0.1105647981,0.2348871529,0.1291821301,0.0971046537,-0.0754145384,-0.2708680332,-0.1090857089,0.0891002491,0.2237171382,-0.1484306157,0.0937630162,-0.4318659008,-0.4758825302,-0.2640722096,-0.2109764814,-0.4392780662,-0.4192179143,-0.1311877221,-0.0238214266,0.0547445118,0.1354451627,-0.078701742,-0.1223691925,-0.2312014997,-0.1154662892,-0.3434458673,-0.4542329907,-0.3919792175,0.0169904344,0.4950798154,0.141469121,0.1429104656,-0.0854893997,-0.2279533744,-0.1821111888,-0.253644973,-0.2276076972,-0.3089286685,0.1341107935,-0.2234796435,0.4357998669,-0.1184876263,-0.1121197715,0.3430631161,0.0052084159,0.0267195683,0.175797388,0.2352900058,0.1736939251,0.002161144,-0.5106552839,-0.3253031075,-0.3045636714,-0.1049587354,0.1949892938,-0.0025412478,0.1089148521,-0.1016926765,0.0883664936,0.0876663551,-0.0099650202,-0.0963250697,0.2964367568,0.4350730479,-0.0457885973,-0.3770374656,0.1137984246,0.0461983941,0.1506627649,0.1252917349,-0.5083814859,0.2113194168,-0.1774634719,0.0534333959,0.0646804944,-0.0603528544,0.0183729641,-0.1751566082,0.0705562979,0.0486055724,-0.1803525239,-0.1996345222,-0.0730896443,0.2446131706,0.0437166244,0.3491378129,-0.3166738451,0.6470992565,0.1936473697,0.0915279239,0.3157174289,0.1844841838,0.0257805251,-0.1023911536,-0.3294164538,0.0942497775,-0.0175162666,-0.0070647388,0.326117605,-0.0540145002,-0.4404801428,-0.1798976362,0.1821159422,-0.3427726924,-0.2643786073,0.1805134416,-0.3352513611,0.1539508253,-0.0801706314,0.0648976937,-0.0291185547,-0.3887671828,-0.104192093,0.2153419852,0.3641248047,0.1478046626,-0.41537112,-0.467261672,-0.2456073463,0.2328764051,0.0414019153,0.5650767088,-0.2956768572,0.0696779862,0.1657708287,0.1735635251,0.3264964521,-0.3261795342,-0.1534575373,0.4508708715,0.0462354757,-0.7034316063,-0.0042074956,-0.2033417225,0.166332677,0.3576461971,0.1388449073,-0.468118459,-0.2272915989,-0.0402059555,0.5781701207,-0.1646542698,-0.1139071286,-0.4263759851,-0.2196538597,-0.4382930994,-0.1795291007,-0.1568226814,0.1889148206,-0.2022859007,0.0113070179,-0.0283576772,0.22435987,0.0859734714,0.0765161589,0.1680495143,0.3729920983,-0.103243567,0.1669536233,0.0377365202,-0.1140469387,0.3032959998,0.0118585285,-0.2465838343,-0.0403529741,-0.1105409712,0.2490494549,0.0052949544,0.1370437443,0.0612594895,0.1077449173,-0.181401372,0.013224774,0.0830996037,0.2614063323,-0.0501082055,-0.3356160522,-0.5175657868,0.5426709652,0.0492941029,-0.0429762378,0.293823272,-0.138583228,-0.3869178593,0.4827125072,-0.0211485289,0.8443021774,-0.2888145745,0.140182361,-0.1172740459,0.1775506586,0.5058164001,0.0213002898,0.0218098667,-0.2752049565,0.2812980115,-0.0359645821,0.0768368542,-0.0160988122,0.1130458191,-0.2303787321,0.340659827,-0.1429936588,-0.0339909457,-0.0085052103,0.5127865076,0.1584011614,-0.333658278,-0.351277411,0.0931842178,-0.2316280007,0.3745656312,-0.0766454339,0.1081154048,-0.0638462603,-0.018038746,-0.296582073,0.1969098151,-0.12895675,0.2717767656,-0.2028959394,-0.2525972426,0.3149382472,0.4847461581,0.0797066391,0.2703463137,-0.1776283681,0.1221787333,-0.1631240845,-0.26545766,-0.0193037484,0.0598704182,0.0388399959,0.0270219687,-0.4331537783,0.1020945534,0.0171888508,0.0461759791,-0.3746176064,0.1316771358,0.0650545731,0.0574659184,-0.293877542,-0.019136522,-0.1510985345,0.0080400193,0.1236542165,0.0664095208,0.1730903983,0.1505156606,0.3648106754,-0.2949020863,-0.083943598,0.3448216915,0.0631317198,-0.0334385484,0.7968848348,0.3733598292,-0.3334774375,-0.2379008085,-0.2758555114,-0.4379781783,-0.4079407454,-0.2622452378,0.1215666234,0.416859448,-0.2348003983,-0.0616233386,-0.0043553095,-0.3975146711,0.0172493439,-0.6908323169,0.0198201314,0.3004770279,-0.0004534828,0.0362745114,-0.0524352826,-0.4288681746,0.2243539542,0.0289952345,-0.2083336562,-0.0463955626,-0.2517012954,0.199966073,0.3432351947,-0.0210956316,-0.1768621802,-0.1202896014,0.0963073894,-0.2485558838,0.1086606309,-0.1992165297,0.0105908448,0.1252682805,0.159891963,-0.1354900301,-0.1576244831,-0.4065170586,0.400698781,-0.28696087,-0.148242265,-0.0836087614,0.014024171,0.372095108,0.0178923551,0.2017331123,-0.0937442631,0.1883346885,-0.0662293956,-0.0049582385,0.2702949047,0.087179631,0.6029509306,0.0909569561,-0.4829334319,-0.0918360576,-0.0459741764,0.1904551536,0.19062379,-0.1822914183,0.4023688436,0.3664537966,-0.0583340116,0.5254053473,-0.0139515689,-0.0432630442,0.2768571675,0.1318716705,-0.2207181752,0.0970370322,0.6332995892,0.1795684993,-0.1412868351,0.16560781,0.1306850314,-0.290212512,0.1639255136,-0.0739221722,0.1292051524,-0.0764234737,0.2255380005,-0.0071146125,0.2125224918,0.0758070275,0.011429511,0.0831690878,-0.1004327163,0.080598861,0.096041292,0.1570519656,-0.0790934041,0.102584295,-0.2248828262,-0.3368074894,0.0790255293,0.3334820271,0.0219750199,0.184964776,0.0201572441,0.2418786138,0.1360520124,-0.1676202267,-0.1336253136,0.2248082608,0.0342127979,-0.1987567097,-0.2119292468,-0.0630901083,-0.1566896141,-0.0681767091,-0.0568807535,-0.2547958195,0.0527102761,0.4515684545,-0.173305735,-0.6086337566,0.4393877983,-0.1677065045,0.092408739,-0.1355500966,0.169407174,0.0415965542,-0.0366634391,0.0079007763,0.1931378692,0.1218981892,0.2390928417,-0.3821927309,0.0667040572,-0.0202773828,0.1204774231,-0.0778521895,0.4424696267,0.152680248,0.3361578882,0.308671087,0.1398470998,-0.0929352343,-0.1569002569,0.2654617429,-0.0708080009,-0.1257316023,0.1575004458,-0.1317599714,0.0889852345,-0.3517001569,-0.0725112781,-0.4017906189,0.1466517895,0.172634542,0.1591138244,0.0643569008,-0.1085583568,0.0359785929,-0.1587149501,0.6551538706,0.4563866854,0.3284854293,-0.2881631553,-0.1525837034,-0.5963298678,-0.011808495,-0.3060020804,-0.0847931281,0.2159911245,0.1992806643,0.1125136316,0.103334941,0.0917911232,-0.1117607728,-0.1631113291,0.2765766382,-0.3883314729,-0.214669466,-0.1558448225,0.0969343185,-0.3057290912,-0.104794696,0.1783330739,-0.2916110158,-0.0240953267,-0.1326552033,-0.1322903037,0.0373812579,0.1854144335,0.256907016,0.0480230786,0.6660428047,-0.0019647924,0.0365837701,-0.3340020478,-0.2727919817,0.0568973832,0.4294123054,-0.181642741,0.286981225,-0.1499210447,0.1362273544,-0.2817760408,0.3479064703,-0.1118230969,0.0423602648,0.0446400754,-0.2241065502,0.020761257,-0.161987558,0.0570326447,0.2371427566,-0.0722227395,0.0700838491,-0.2826004624,-0.1661403775,0.191840753,-0.2951719463,-0.4527609646,0.0211081374,0.111452572,0.1165713817,0.1614441276,-0.3354124427,0.1012782529,0.4468919933,-0.0410032906,0.0854315683,0.144568786,0.0365336165,-0.1233106181,-0.1775791198,-0.1447811872,0.0762892514,-0.0864648297,-0.055475153,-0.1618865579]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/577","title":"Some languages in wikipedia dataset are not loading","comments":"Hi ! This looks related to this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1994\r\nBasically the parser that is used (mwparserfromhell) has some issues for some pages in `es`.\r\nWe already reported some issues for `es` on their repo at https:\/\/github.com\/earwig\/mwparserfromhell\/issues\/247 but it looks like there are still a few issues. Might be a good idea to open a new issue on the mwparserfromhell repo","body":"Hi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!","comment_length":60,"text":"Some languages in wikipedia dataset are not loading\nHi,\r\n\r\nI am working with the `wikipedia` dataset and I have a script that goes over 92 of the available languages in that dataset. So far I have detected that `ar`, `af`, `an` are not loading. Other languages like `fr` and `en` are working fine. Here's how I am loading them:\r\n\r\n```\r\nimport nlp\r\n\r\nlangs = ['ar'. 'af', 'an']\r\n\r\nfor lang in langs:\r\n data = nlp.load_dataset('wikipedia', f'20200501.{lang}', beam_runner='DirectRunner', split='train') \r\n print(lang, len(data))\r\n```\r\n\r\nHere's what I see for 'ar' (it gets stuck there):\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.ar (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/gaguilar\/.cache\/huggingface\/datasets\/wikipedia\/20200501.ar\/1.0.0\/7be7f4324255faf70687be8692de57cf79197afdc33ff08d6a04ed602df32d50...\r\n```\r\n\r\nNote that those languages are indeed in the list of expected languages. Any suggestions on how to work around this? Thanks!\nHi ! This looks related to this issue: https:\/\/github.com\/huggingface\/datasets\/issues\/1994\r\nBasically the parser that is used (mwparserfromhell) has some issues for some pages in `es`.\r\nWe already reported some issues for `es` on their repo at https:\/\/github.com\/earwig\/mwparserfromhell\/issues\/247 but it looks like there are still a few issues. Might be a good idea to open a new issue on the mwparserfromhell repo","embeddings":[0.1454113871,-0.3527697027,-0.1282091141,0.4522044063,0.1266885996,0.2579191029,0.0820914432,0.1208299994,0.664529264,-0.2317667156,0.049036961,0.0963193104,0.1007603258,-0.1433963627,0.1095914394,-0.1608642042,0.0873809457,-0.0652693734,-0.0717887953,-0.2942489982,-0.3200691044,0.3018416166,-0.3130692542,0.0875176564,-0.2290617824,0.2680266201,0.0768611804,-0.0958316401,0.1281856298,-0.3754602671,0.1456995755,0.1750696898,0.3053931296,0.1023045182,-0.0001208771,-0.0671264678,0.6932794452,-0.1155171245,-0.4495871663,-0.166202873,-0.06001845,-0.394503057,0.2241929024,-0.084403865,0.0897282213,-0.2101305872,0.2128771394,-0.4753671288,0.2147086412,0.0394958816,0.1839958131,-0.0195652153,0.0651311576,0.1096665487,0.2768373191,0.1622897238,0.1529465616,0.0777904242,0.0675142407,-0.1927975267,-0.0473742187,0.1472345144,-0.0759023055,-0.1603672206,0.1972974986,-0.0518286191,-0.0422015786,-0.582298398,0.254065305,0.1718512774,0.6346393824,0.0478740036,-0.3067273498,-0.2932403684,-0.0394684263,0.1125655249,0.2497185469,0.3170331717,-0.1362361312,0.0208639465,-0.0583339259,-0.2783628404,-0.0473206192,0.5574242473,-0.0775968954,0.4045908749,0.0481628925,0.1386425197,0.1338667423,-0.2032577395,-0.3406938612,-0.2172077,0.3812316954,0.3883214593,-0.1246935874,0.4484196603,-0.0046273577,0.3969624043,0.0827118382,-0.1514822394,-0.3183687031,0.2936990857,-0.1225302592,0.1603396684,0.2655622661,-0.1362960488,0.3388786316,-0.2905006707,0.3430874646,0.1240720376,-0.1800464541,0.0506780557,-0.2567569017,-0.1705679595,-0.236676231,-0.1381655335,-0.0410395376,-0.3294737637,0.196850881,0.2522233725,-0.154537946,-0.329650104,-0.1120171845,0.3492595255,-0.0588776991,0.2394414991,0.1553758234,0.3203840852,-0.4523286223,-0.1098876074,0.0471827909,-0.1493367106,-0.2431429178,0.1791200191,0.2842279971,0.0920019746,0.3260169923,-0.0165743381,0.0259938464,-0.2473278493,-0.1506370306,-0.1562276036,-0.0563953407,0.0323725231,0.1693221033,0.5196843743,0.1437484324,-0.2705621719,-0.1778963953,0.0902567655,-0.1203154847,0.3405405581,0.0907867849,0.0816247761,-0.2965570688,-0.0704906061,-0.3920736909,0.2990905941,0.0731613636,-0.1684184372,-0.0265507791,-0.3151744902,-0.2208946794,-0.0374681614,0.3430621326,0.6845704913,-0.211151883,-0.1792090535,-0.0998364091,0.0061344467,0.2455126941,0.2722980082,-0.0244780108,0.2045751214,-0.1845206171,0.123340562,0.2107272446,0.0210101921,0.1327968687,-0.0070712422,0.2727644444,0.3687072098,-0.0597575903,-0.1002426818,0.0652208403,0.033174023,0.040587008,0.4009352326,0.2103778869,0.0254881736,-0.2740682065,-0.0816094279,0.5256812572,0.3201315403,0.0770762041,-0.208517462,-0.0055628195,0.258412689,0.1945173144,-0.1087042987,0.1237082928,0.4082092643,-0.3513307273,0.3550541997,0.2252843231,-0.0785182938,-0.5690230727,0.1859130859,0.0625603274,0.1939194202,-0.0308892075,0.1396826059,-0.2773541808,-0.0942796841,-0.2347115725,0.1973899305,0.0538190715,0.2056227624,-0.0723793805,0.5458549261,0.0506740585,-0.1096699163,-0.0347334743,-0.0858458504,-0.5625824332,0.3580300212,0.0369766764,-0.0827883482,0.0142246261,0.2626745105,0.267306149,-0.1898435801,-0.0969693884,-0.247234419,0.4029215574,0.2461492717,0.1654469818,-0.0673913807,0.2773109078,-0.4852219224,0.1915935874,0.2021669149,0.1328281462,-0.1937077045,-0.0900983438,0.2346535921,-0.1538479626,0.5538975,0.074362509,0.0472388752,0.3715311885,0.045037251,-0.0550123379,-0.1765595526,0.5834852457,0.0442728251,0.1136139408,0.251668036,-0.1351589859,-0.1290346235,0.4401016831,0.0481502451,-0.0139508341,0.3489096761,-0.224578321,-0.0870222077,-0.080037564,-0.2888080478,-0.177338779,0.1297315806,0.0773875341,-0.2715661228,0.0271356776,-0.1570096314,0.0504053123,0.1856997013,0.1723355204,0.1115908846,0.1352159679,-0.1144849509,-0.3306603432,0.0014811121,-0.0210893042,0.1773554087,-0.2405599803,0.1249641553,-0.4194021225,-0.2736737728,-0.3203495443,-0.1592349112,-0.597266376,-0.3544646204,-0.1591806412,0.0280090757,0.0243867189,0.0893844068,-0.0452939719,0.0177030582,-0.2964751422,-0.0144438595,-0.468862921,-0.4974044561,-0.4261675477,-0.0808675513,0.6184802651,0.1411379129,0.1409076899,-0.0959097594,-0.2744810581,-0.1694307923,-0.3183217049,-0.156931594,-0.3113234937,0.2710371017,-0.2807469666,0.393448472,-0.134000808,-0.0908795148,0.4661806524,-0.0452605486,0.0351128615,0.1680601537,0.2521027327,0.0955855101,0.0170567296,-0.4298453629,-0.2396096438,-0.2278456241,0.0001125239,0.1121791154,-0.0594212562,-0.0182170179,-0.1800003648,0.0478013605,-0.0729807541,0.0590053089,-0.1639425308,0.3648957908,0.4119184613,-0.0110468855,-0.3786582947,0.053966105,0.1473717988,0.1111441329,-0.0056716437,-0.5760472417,0.0858578756,-0.1520605087,-0.0183307882,0.1165752038,-0.0038573905,-0.066689752,-0.1485970914,0.1208328083,0.0229585543,-0.306251049,-0.2113108784,-0.1515654773,0.1288844794,0.0644821674,0.4482483864,-0.3527316153,0.6243929863,0.2692997754,0.1014882401,0.2874146998,0.2123984843,-0.0006011522,-0.1808139235,-0.2832011282,0.0454875752,-0.1020772755,-0.0436686464,0.4483745098,0.0932997167,-0.4209004939,-0.1743419021,0.0814675987,-0.3827198744,-0.3302599192,0.1829960197,-0.0985627547,0.1848843992,-0.1032559499,0.0673331022,-0.0198075119,-0.39537251,-0.0184431151,0.2780789733,0.3779979944,0.1212863773,-0.4214282036,-0.4577128887,-0.2371254265,0.1816522628,0.0365826897,0.5437663794,-0.3557375371,-0.0216215942,0.2409396917,0.0927205905,0.383803308,-0.1757911891,-0.0598697998,0.3640905619,-0.0865457952,-0.5796944499,0.0129828947,-0.1262015104,0.0152768772,0.4794662595,0.2567058206,-0.4715684354,-0.3081721067,-0.1061527282,0.5398074389,-0.2224415392,-0.0887349695,-0.4245489538,-0.1331251413,-0.4907981753,-0.1154399589,-0.1945579648,0.110185869,-0.1219692901,0.0694763437,0.0580264963,0.2612292767,0.1675082445,0.1626870483,0.1256118268,0.236199826,-0.0506516211,0.2241119742,0.1598333716,-0.0596056841,0.3734050691,0.0134114316,-0.3509371281,-0.0528820865,-0.136341095,0.3402988613,0.0525180325,0.1857918352,-0.0086663021,0.2165903598,-0.0618351549,-0.1990320832,0.0680585131,0.2766777277,0.0226415731,-0.3399432898,-0.3216896653,0.6216524839,-0.0022299318,-0.0463795029,0.191045776,0.1186406985,-0.4334776402,0.484128803,-0.0268211234,0.915192306,-0.3280844986,0.1606160402,-0.1644029021,0.0693101957,0.5103273392,0.0728498027,-0.1079437807,-0.281508863,0.3742105067,0.003761518,0.0704976395,0.033776205,0.1312715709,-0.1726237237,0.2494290024,-0.0444497503,-0.0069418112,0.0265839621,0.5273493528,0.2180223316,-0.3456304967,-0.3993367255,0.0495016649,-0.2260796875,0.3657533824,-0.0629485548,0.1194542125,-0.0033383435,-0.0138054388,-0.3177850246,0.1582010388,-0.0728236809,0.152575314,-0.0260325614,-0.2204855233,0.4314934909,0.3578050733,0.0291130506,0.188619107,-0.2568198144,0.1398090571,-0.2945701778,-0.2012758851,-0.0186440907,0.0743671656,0.0361814015,0.0337958001,-0.3190800548,-0.0103380475,0.0058311033,0.0496031567,-0.3906801641,0.0969273895,0.0281796437,0.0357890539,-0.3012332618,-0.0508096479,-0.0378191136,-0.0096781589,0.0933457389,0.0314240456,0.1742428541,0.1721737683,0.4657342434,-0.2537176907,-0.0828507394,0.2836849988,-0.0905733332,-0.1029881686,0.7523550987,0.4931761622,-0.3861753941,-0.2160488665,-0.2555176616,-0.386367172,-0.5297489762,-0.2308185548,0.2378482223,0.3199221492,-0.2337437719,-0.0487273037,0.0475786701,-0.4764851332,-0.0238332208,-0.6159587502,0.0629197359,0.3354687393,0.0165780745,0.0593027361,0.0062833158,-0.3525325358,0.1037407741,0.1478776634,-0.1578246951,-0.0258048233,-0.1869847775,0.1717973351,0.3937838078,-0.0427252129,-0.1469962597,-0.1072945371,0.0563528202,-0.2338115424,0.0880920365,-0.1381164044,0.0456751958,0.154764697,0.1228262037,-0.1809514314,-0.2163016647,-0.2853405476,0.4532654285,-0.4488101602,-0.1005927473,-0.1114415452,-0.0505845919,0.247631982,0.0785312802,0.1711519957,-0.0566083454,0.1552759856,-0.0537956543,0.0145780938,0.3479129672,0.0301182233,0.4769906104,0.1282318085,-0.3364941776,0.032993257,0.073609367,0.1639910936,0.2181762755,-0.1429115981,0.2568368316,0.470287621,-0.0622325391,0.5298502445,0.032063853,-0.0739812627,0.3156300187,0.0666023493,-0.1692597568,0.0446893871,0.7771174312,0.1248701662,-0.1889654249,0.163543269,0.149710983,-0.2596126497,0.1042393073,-0.0925645307,0.3295483291,-0.0656349882,0.1110406443,-0.0727938414,0.2612842023,0.1104629785,0.0010078148,0.0843648612,-0.0359288715,0.0308814645,0.0707209781,0.0810002238,-0.1426289678,0.2048092335,-0.1930038184,-0.3593955636,-0.0145362867,0.3893146515,0.1058749706,0.188455686,-0.021011427,0.324328512,0.1207644567,-0.2275565565,-0.2272682935,0.3058841825,0.0331070907,-0.3350078166,-0.1665449142,-0.0003768264,-0.1447977722,0.0586103983,-0.0320913643,-0.2932665646,0.0943524614,0.4384005964,-0.1160318032,-0.5470771194,0.2070728242,-0.2495485991,0.1478572339,-0.0930462927,0.211356774,0.040819563,-0.0305342339,-0.0404403806,0.0705895573,0.2195025682,0.2346307933,-0.3896163702,0.0523686111,-0.1126590371,0.2038098425,-0.1255104393,0.5091893673,0.0827486515,0.4134193361,0.229246065,0.0984183624,-0.037211366,-0.3240758181,0.2217305601,-0.0853826404,-0.0885720327,0.0381478295,-0.1624500602,0.1032264829,-0.3157377541,-0.148256436,-0.2848929167,0.2957913876,0.2464161664,0.1293192655,0.0400482975,0.0059815715,0.008170234,-0.0770003796,0.6391988397,0.6014124155,0.2962319255,-0.2728562355,-0.231626913,-0.6162995696,-0.0720535442,-0.2881959379,-0.0741292015,0.2476618439,0.112710081,0.1472309679,0.0724601671,0.0799306184,-0.1244617552,-0.1621920913,0.2225466371,-0.3230166435,-0.1215390339,-0.1404101253,0.0720485449,-0.3405055404,-0.0398606025,0.2299981266,-0.2287095338,-0.1217885688,-0.0672967359,0.0563757941,-0.1051207185,0.1474519372,0.2411967069,0.0859884247,0.5995777249,0.0011793002,0.0553466603,-0.2797983587,-0.2594980896,0.1524222195,0.4133249223,-0.2980041504,0.2303025723,-0.206395179,0.0778888613,-0.289984405,0.360212326,-0.0527691469,0.0901128501,-0.0611301735,-0.1782128215,-0.0139319394,-0.1953758746,0.0780582726,0.2493192405,-0.0704437196,0.111326389,-0.2750208676,-0.1183611602,0.2797544599,-0.3547440767,-0.436190635,0.027380567,0.1413471699,0.1676131487,0.0812713355,-0.4344432652,-0.0146427136,0.3558734059,-0.0144727807,0.007326785,0.1548235565,0.0178533047,-0.284368813,-0.1955544353,-0.1332451403,0.1044576839,-0.0845366493,0.0304160826,-0.1153535545]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/575","title":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.","comments":"Update:\r\n\r\nThe imdb download completed after a long time (about 45 mins). Ofcourse once download loading was instantaneous. Also, the loaded object was of type `arrow_dataset`. \r\n\r\nThe urls for glue still doesn't work though.","body":"Hi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n","comment_length":34,"text":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.\nHi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n\nUpdate:\r\n\r\nThe imdb download completed after a long time (about 45 mins). Ofcourse once download loading was instantaneous. Also, the loaded object was of type `arrow_dataset`. \r\n\r\nThe urls for glue still doesn't work though.","embeddings":[-0.112636596,-0.0148180015,-0.0514469258,0.2499045283,0.2150651067,-0.1080487967,-0.0023050923,0.1171098575,0.144177258,-0.1348839551,-0.425668329,-0.0689154193,0.1447221488,0.1289693564,0.274225086,-0.0121180899,-0.1593028903,-0.0744765624,-0.0916164145,0.017701054,-0.2575627863,0.3606531918,-0.2036087364,0.1764502376,-0.1922911257,-0.0935443118,-0.1606316715,0.1398071498,-0.1229681745,-0.3245210052,0.0947918966,0.3048549592,0.0481997393,0.3857811987,-0.0001219456,-0.0564453825,0.5723124146,-0.0555414632,-0.3910080492,-0.4353305697,-0.2779269516,-0.0589925274,0.355984658,-0.0667517558,0.130753994,0.1634606421,0.0878987014,-0.2419316471,0.2720347047,0.26318115,0.1377917379,0.0733033642,0.1359027624,0.0714743361,0.3174403608,-0.2632397711,0.0046695629,0.3708156645,0.3326888382,-0.291418612,0.0472086743,0.0577220842,-0.205646798,0.0583759286,0.1841480434,0.1635258496,-0.25784266,-0.5462059379,0.0292282049,0.3213647902,0.2748629451,0.0140512539,-0.4840654433,-0.3165412545,-0.0286721792,0.1173614413,0.2789925039,0.1334600002,-0.2538383007,0.0309729725,-0.5450456142,-0.1321876645,0.0033698108,0.5532464981,0.1101105139,-0.0098115858,0.0639427006,0.1689395458,0.3128620982,-0.0395253114,0.1148340628,-0.0232361071,0.2994754314,0.1872622669,-0.0450580493,-0.0265388358,0.0512051396,0.4236578643,0.1454071999,0.1372276694,0.1992809027,0.0405518636,-0.1392972767,0.1321616024,0.15483661,-0.0140559683,-0.0127521874,0.0798756033,0.6293030977,0.3603568673,-0.1680348217,0.0711006969,-0.1136864573,-0.0785518587,-0.3989927769,-0.0077087539,-0.1199624836,-0.284093976,-0.1622226536,-0.0771386921,-0.1368774027,0.0150720552,0.0852001831,0.3928332329,-0.3108938336,0.2066407949,-0.087643005,0.3793320358,-0.1213530898,0.0319333375,-0.0930832773,0.1856428087,-0.1889153719,0.2517310679,0.4859232306,0.0784364417,0.3209574521,-0.3183521628,-0.1498351097,-0.2291524559,0.0177800152,-0.3440533578,0.0556664392,0.0532136112,0.3014571071,0.3389762938,-0.0610685609,-0.1870903969,-0.1006215811,-0.2109535187,-0.2240741253,-0.1662695557,0.3674656749,0.1121519953,-0.4041662812,-0.1559494436,-0.1344938278,-0.004829342,-0.0961971432,-0.0948827863,-0.1139309928,-0.0595751256,-0.1584257782,-0.200951606,0.3062122166,0.6965598464,-0.1322715133,-0.2766490877,-0.1275884658,0.041212976,-0.0782895386,0.502037406,-0.0070996187,0.1006353199,-0.4256356955,0.2437322885,0.5672342777,-0.2894794941,-0.3902561069,0.5320960879,-0.0154582411,0.0364813991,0.1849975735,0.0225609727,0.131248787,-0.0723032951,0.4729860723,0.5410932899,-0.0449328944,-0.0970429629,-0.2779184282,-0.1605929285,0.1289730817,0.1257047504,0.223503828,0.0212896317,-0.0032729721,-0.079706423,0.233262375,0.2292817086,0.0552493818,0.2484004796,-0.1013187617,0.1788349301,-0.104797557,-0.1381653398,-0.3142169118,0.2160550207,-0.0629844964,0.1218500361,-0.3327453732,0.2367781252,-0.1883798987,-0.1930935532,0.0038398847,-0.0187660865,0.0700894296,0.2044757456,0.131302923,0.0753203556,-0.2522730231,0.5734852552,0.0272469651,0.1724773198,-0.316007942,0.6796264052,-0.1542854309,-0.191763401,0.1090584621,0.3169533312,0.2635817826,-0.2635529935,0.0134699494,0.3022545278,-0.1289988309,0.1917850226,0.2702765465,0.0180154573,0.3637427986,-0.2720222771,0.1126302034,0.0491894484,0.1340863556,-0.1252143681,0.1797020435,0.2376711965,-0.2186419368,0.6319778562,0.4755757749,-0.0091407485,0.1447475404,-0.1831095815,-0.2134258151,-0.2173276395,0.6007691026,0.2402251363,0.2317631692,-0.1197275817,-0.148830682,0.1870426685,0.3490749896,0.0156644378,0.05569648,0.2178222686,-0.1637909412,-0.1998849362,-0.1010185257,0.3804479241,0.2209589183,0.1598923951,0.1135150269,0.021280108,-0.2521316409,-0.3306078017,0.2030390054,0.1683373004,-0.0889684558,0.0819643363,0.2019196898,0.1494818777,-0.0947707817,-0.4382834136,0.0053283786,-0.0233714301,-0.3714630306,-0.0647940934,-0.445727706,-0.6604163051,0.0467646122,-0.1720999628,-0.2792548835,-0.646451056,-0.1037925184,0.2034947723,0.0090118954,0.0356713496,-0.2063030154,-0.0062600239,-0.2111667097,-0.0129495598,-0.1801105589,-0.3260889649,-0.062903434,0.0395989567,0.3534008265,0.0574423596,0.1485339999,-0.2441104352,-0.1611087322,-0.1612910032,-0.1101251915,0.19283548,-0.1821724176,0.3608709574,0.2109528184,0.4903079271,0.2029083073,-0.2494547963,0.3121628463,-0.1433433294,0.0774954781,-0.0519493446,0.0077687465,-0.0471082143,0.1043553874,-0.0099998713,-0.471308589,-0.4650456309,0.2359144241,-0.0598598197,0.2176345289,0.1203419119,-0.041052755,0.2929945886,-0.0266634505,0.1280004233,-0.041178517,-0.3354517817,0.5650700331,-0.2066883743,-0.2000689805,-0.0150570692,0.0398590788,-0.2016781569,0.0807290226,-0.706532836,-0.2792837024,-0.4299277067,0.0273614805,0.2272491306,-0.0665315464,-0.0501204729,-0.2217652798,-0.0760565326,-0.0140878279,-0.4060663879,-0.0029456718,0.3619810045,0.3475250006,0.1065041944,0.5122767091,-0.0753890201,0.272541374,-0.0739138052,0.0747887865,0.6189810038,0.0821451545,0.3144127727,-0.1913370788,-0.269040674,0.1089694574,-0.3142771423,0.0521760285,0.1350190192,0.2109739482,-0.3150362074,-0.5289624333,0.172768876,-0.4613894224,-0.1889805794,0.0283963811,-0.0316216201,-0.2145553231,0.2021680921,0.035619285,-0.0524334349,-0.2252569646,-0.0457172729,0.1494725794,0.2521126568,0.039624352,-0.1105956882,-0.1379540712,-0.4256444275,0.2749575973,0.1958890855,0.3104603589,-0.2696570158,-0.0499456488,0.0475213565,-0.0248350445,0.5041225553,-0.4135855436,0.2516334057,0.1580861807,0.0192981642,-0.6474230289,0.0887410268,-0.1247085482,0.1403576285,0.6630892158,0.2826699615,-0.1967948675,0.1196002215,-0.0139652062,0.1940147877,0.1074987426,0.0746590942,-0.0067794248,-0.2686884701,-0.2585899532,-0.2295411676,-0.0385957845,0.0707760081,-0.0640727878,-0.3239626288,0.3134526014,-0.4003784955,-0.0395804793,0.0364644267,0.1761945337,-0.107365407,0.0827875137,0.2307296842,0.1615782082,-0.1797794253,0.4915136993,-0.3460828662,0.0986247435,0.5306717157,-0.0260191523,0.1837768108,0.6057760715,0.1359435916,0.1512895077,0.0835235864,0.1909211874,0.2960536778,0.3395085037,0.2108195573,-0.0150642218,-0.282137394,-0.3250028491,0.3341210485,-0.0021029513,-0.0231160782,0.2038309872,0.0800835639,-0.2553889751,0.2971126735,-0.3381642997,1.148299098,0.0942292139,0.1294936687,-0.1617401689,-0.252526015,0.664984107,-0.1067686155,0.1237175688,-0.046933569,-0.1145089716,-0.1157739386,-0.1703818738,-0.211443305,0.0383760333,-0.2149146348,0.5331195593,0.075566791,-0.0113159101,0.1618145704,0.4351913631,-0.1023060456,-0.0456134863,-0.5257042646,0.0877486318,-0.0442999899,0.2518552244,-0.0140865427,-0.1485861987,-0.2059692591,0.005314081,-0.4601523578,0.1955182999,-0.6179574132,0.2456762493,-0.0609878637,-0.3103728592,-0.1943768859,-0.0575052127,0.1287858635,-0.0164992493,-0.3511244655,0.0874209404,0.2061311007,-0.1411281377,0.1736903042,0.0441385135,0.1641439348,-0.1359499395,-0.0259148777,-0.0478292927,0.0252410229,-0.3272763789,0.0321033821,-0.1285140514,0.0396090299,0.0648923889,-0.1876660287,-0.1389926225,-0.1790039539,-0.2322729975,0.0341174826,-0.0430926159,0.1417649537,-0.085842438,0.3707483113,-0.2654457986,-0.16239582,0.5142537951,-0.305388242,0.0012647071,0.339337945,0.3020931184,-0.3583815396,-0.0925878659,-0.1114680618,-0.1822286099,-0.3933390677,-0.0003344097,-0.2642868161,0.0666865036,-0.0891283676,0.0259221327,0.1505150199,0.1860919148,-0.1181158945,-0.6274639964,-0.1790275872,0.4015606344,-0.0270387996,0.0485015772,-0.2964761555,0.1652342081,0.4207263887,-0.0386223979,-0.2305992246,0.1560420096,-0.44154194,0.0104385708,0.3247368336,-0.3346687257,0.0485885106,-0.0424531437,0.0563483611,0.0833296478,-0.1790658981,-0.0608192533,-0.1101867259,0.1974409521,0.0511796996,-0.3207353354,-0.0964926332,-0.0979493782,0.2066013664,-0.0818443969,-0.0759881213,-0.0132816853,0.0211339984,-0.2951012254,0.1266241819,-0.0999270082,-0.1372247636,0.3212115169,-0.0065158373,0.2825662494,0.3173856437,0.1064764485,0.0883074254,-0.0717145652,-0.0931007713,-0.1241110042,-0.0829002708,0.1949430406,0.4736431241,-0.0916405395,0.0512022041,0.0569887497,0.3979811966,0.5058012009,0.0556126125,-0.1075799018,0.1179754958,0.0908082947,0.0191558711,0.251812458,0.2627199292,0.1216957122,0.0234896373,0.097651422,0.0498040468,-0.0724791065,0.0851857215,0.075487569,0.6589386463,0.201536268,-0.0219140518,-0.1459535807,0.0476535223,-0.0527363233,0.0410787389,-0.0996239334,0.0687503144,-0.126663357,0.0534229763,-0.0639401227,-0.1313260347,0.3249314427,0.2482800186,-0.2003667951,-0.0873381868,0.2849595845,0.0983348265,0.1369969398,-0.220414117,0.3333672583,-0.1611330211,0.0659117475,-0.1380514205,0.4353375435,0.1926468909,-0.3216380179,0.0582479835,-0.3349765539,-0.4865922332,0.0540139452,-0.1999975294,-0.1715221703,0.1560159922,0.2826383114,0.0135858925,-0.7253261805,-0.1404952407,-0.1912470907,0.232984975,-0.3082934618,0.4661635458,-0.1038244143,0.0777674317,-0.0195026733,0.1526183188,0.3781436086,0.3050530851,-0.4279482365,0.0964921415,0.2019785941,-0.0506129041,-0.0096543431,0.230298087,0.4857646525,-0.1341400892,0.492813617,0.0891050696,-0.1842339635,0.2686845064,-0.0354273506,0.1412737072,-0.0751045942,0.2763312161,-0.0875300467,-0.0513294786,-0.2288811356,-0.1396883428,-0.5556668043,-0.109657459,-0.0335210189,-0.5973433256,0.0323545001,-0.0905735418,0.04113717,0.0053061331,0.6298923492,0.569093585,0.5683688521,-0.380954206,-0.3364817202,-0.3252815902,0.0403116867,-0.1251098812,-0.0124700721,-0.1701546311,0.2908905745,-0.174534604,0.2047268748,-0.0137992641,0.6748144627,-0.1523853242,0.1743044406,-0.2713821232,-0.0484785065,-0.0313812792,0.0573393144,-0.1030781344,-0.0273058582,0.0516437143,-0.5177112818,-0.0516986661,0.2759017646,-0.2825129032,0.133970499,0.236651957,0.3119557798,0.1032110676,0.7628632784,-0.122793287,-0.1470048726,-0.4071486294,-0.6882551908,-0.0240677781,-0.1192925423,0.0928619131,0.1546040028,-0.0933201984,0.2021304667,-0.1746087968,0.190074876,-0.0879273117,0.4516707063,0.1424894631,-0.3678409457,0.065007776,-0.0495072305,0.0439469479,-0.0063381037,-0.2057438493,0.1007037237,-0.1159259304,-0.3511934578,0.5208379626,-0.0540384129,-0.0175133403,-0.1145703346,0.2055766732,0.0688287094,0.2129953802,-0.3332384527,-0.2800841331,0.344409585,0.058339078,-0.4353078008,0.2324451804,-0.0068449518,-0.0778456554,-0.0766064376,-0.0418283306,0.2216898501,-0.218432188,-0.192988202,-0.0678091273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/575","title":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.","comments":"Thanks for the report, I'll give a look!","body":"Hi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n","comment_length":8,"text":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.\nHi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n\nThanks for the report, I'll give a look!","embeddings":[-0.112636596,-0.0148180015,-0.0514469258,0.2499045283,0.2150651067,-0.1080487967,-0.0023050923,0.1171098575,0.144177258,-0.1348839551,-0.425668329,-0.0689154193,0.1447221488,0.1289693564,0.274225086,-0.0121180899,-0.1593028903,-0.0744765624,-0.0916164145,0.017701054,-0.2575627863,0.3606531918,-0.2036087364,0.1764502376,-0.1922911257,-0.0935443118,-0.1606316715,0.1398071498,-0.1229681745,-0.3245210052,0.0947918966,0.3048549592,0.0481997393,0.3857811987,-0.0001219456,-0.0564453825,0.5723124146,-0.0555414632,-0.3910080492,-0.4353305697,-0.2779269516,-0.0589925274,0.355984658,-0.0667517558,0.130753994,0.1634606421,0.0878987014,-0.2419316471,0.2720347047,0.26318115,0.1377917379,0.0733033642,0.1359027624,0.0714743361,0.3174403608,-0.2632397711,0.0046695629,0.3708156645,0.3326888382,-0.291418612,0.0472086743,0.0577220842,-0.205646798,0.0583759286,0.1841480434,0.1635258496,-0.25784266,-0.5462059379,0.0292282049,0.3213647902,0.2748629451,0.0140512539,-0.4840654433,-0.3165412545,-0.0286721792,0.1173614413,0.2789925039,0.1334600002,-0.2538383007,0.0309729725,-0.5450456142,-0.1321876645,0.0033698108,0.5532464981,0.1101105139,-0.0098115858,0.0639427006,0.1689395458,0.3128620982,-0.0395253114,0.1148340628,-0.0232361071,0.2994754314,0.1872622669,-0.0450580493,-0.0265388358,0.0512051396,0.4236578643,0.1454071999,0.1372276694,0.1992809027,0.0405518636,-0.1392972767,0.1321616024,0.15483661,-0.0140559683,-0.0127521874,0.0798756033,0.6293030977,0.3603568673,-0.1680348217,0.0711006969,-0.1136864573,-0.0785518587,-0.3989927769,-0.0077087539,-0.1199624836,-0.284093976,-0.1622226536,-0.0771386921,-0.1368774027,0.0150720552,0.0852001831,0.3928332329,-0.3108938336,0.2066407949,-0.087643005,0.3793320358,-0.1213530898,0.0319333375,-0.0930832773,0.1856428087,-0.1889153719,0.2517310679,0.4859232306,0.0784364417,0.3209574521,-0.3183521628,-0.1498351097,-0.2291524559,0.0177800152,-0.3440533578,0.0556664392,0.0532136112,0.3014571071,0.3389762938,-0.0610685609,-0.1870903969,-0.1006215811,-0.2109535187,-0.2240741253,-0.1662695557,0.3674656749,0.1121519953,-0.4041662812,-0.1559494436,-0.1344938278,-0.004829342,-0.0961971432,-0.0948827863,-0.1139309928,-0.0595751256,-0.1584257782,-0.200951606,0.3062122166,0.6965598464,-0.1322715133,-0.2766490877,-0.1275884658,0.041212976,-0.0782895386,0.502037406,-0.0070996187,0.1006353199,-0.4256356955,0.2437322885,0.5672342777,-0.2894794941,-0.3902561069,0.5320960879,-0.0154582411,0.0364813991,0.1849975735,0.0225609727,0.131248787,-0.0723032951,0.4729860723,0.5410932899,-0.0449328944,-0.0970429629,-0.2779184282,-0.1605929285,0.1289730817,0.1257047504,0.223503828,0.0212896317,-0.0032729721,-0.079706423,0.233262375,0.2292817086,0.0552493818,0.2484004796,-0.1013187617,0.1788349301,-0.104797557,-0.1381653398,-0.3142169118,0.2160550207,-0.0629844964,0.1218500361,-0.3327453732,0.2367781252,-0.1883798987,-0.1930935532,0.0038398847,-0.0187660865,0.0700894296,0.2044757456,0.131302923,0.0753203556,-0.2522730231,0.5734852552,0.0272469651,0.1724773198,-0.316007942,0.6796264052,-0.1542854309,-0.191763401,0.1090584621,0.3169533312,0.2635817826,-0.2635529935,0.0134699494,0.3022545278,-0.1289988309,0.1917850226,0.2702765465,0.0180154573,0.3637427986,-0.2720222771,0.1126302034,0.0491894484,0.1340863556,-0.1252143681,0.1797020435,0.2376711965,-0.2186419368,0.6319778562,0.4755757749,-0.0091407485,0.1447475404,-0.1831095815,-0.2134258151,-0.2173276395,0.6007691026,0.2402251363,0.2317631692,-0.1197275817,-0.148830682,0.1870426685,0.3490749896,0.0156644378,0.05569648,0.2178222686,-0.1637909412,-0.1998849362,-0.1010185257,0.3804479241,0.2209589183,0.1598923951,0.1135150269,0.021280108,-0.2521316409,-0.3306078017,0.2030390054,0.1683373004,-0.0889684558,0.0819643363,0.2019196898,0.1494818777,-0.0947707817,-0.4382834136,0.0053283786,-0.0233714301,-0.3714630306,-0.0647940934,-0.445727706,-0.6604163051,0.0467646122,-0.1720999628,-0.2792548835,-0.646451056,-0.1037925184,0.2034947723,0.0090118954,0.0356713496,-0.2063030154,-0.0062600239,-0.2111667097,-0.0129495598,-0.1801105589,-0.3260889649,-0.062903434,0.0395989567,0.3534008265,0.0574423596,0.1485339999,-0.2441104352,-0.1611087322,-0.1612910032,-0.1101251915,0.19283548,-0.1821724176,0.3608709574,0.2109528184,0.4903079271,0.2029083073,-0.2494547963,0.3121628463,-0.1433433294,0.0774954781,-0.0519493446,0.0077687465,-0.0471082143,0.1043553874,-0.0099998713,-0.471308589,-0.4650456309,0.2359144241,-0.0598598197,0.2176345289,0.1203419119,-0.041052755,0.2929945886,-0.0266634505,0.1280004233,-0.041178517,-0.3354517817,0.5650700331,-0.2066883743,-0.2000689805,-0.0150570692,0.0398590788,-0.2016781569,0.0807290226,-0.706532836,-0.2792837024,-0.4299277067,0.0273614805,0.2272491306,-0.0665315464,-0.0501204729,-0.2217652798,-0.0760565326,-0.0140878279,-0.4060663879,-0.0029456718,0.3619810045,0.3475250006,0.1065041944,0.5122767091,-0.0753890201,0.272541374,-0.0739138052,0.0747887865,0.6189810038,0.0821451545,0.3144127727,-0.1913370788,-0.269040674,0.1089694574,-0.3142771423,0.0521760285,0.1350190192,0.2109739482,-0.3150362074,-0.5289624333,0.172768876,-0.4613894224,-0.1889805794,0.0283963811,-0.0316216201,-0.2145553231,0.2021680921,0.035619285,-0.0524334349,-0.2252569646,-0.0457172729,0.1494725794,0.2521126568,0.039624352,-0.1105956882,-0.1379540712,-0.4256444275,0.2749575973,0.1958890855,0.3104603589,-0.2696570158,-0.0499456488,0.0475213565,-0.0248350445,0.5041225553,-0.4135855436,0.2516334057,0.1580861807,0.0192981642,-0.6474230289,0.0887410268,-0.1247085482,0.1403576285,0.6630892158,0.2826699615,-0.1967948675,0.1196002215,-0.0139652062,0.1940147877,0.1074987426,0.0746590942,-0.0067794248,-0.2686884701,-0.2585899532,-0.2295411676,-0.0385957845,0.0707760081,-0.0640727878,-0.3239626288,0.3134526014,-0.4003784955,-0.0395804793,0.0364644267,0.1761945337,-0.107365407,0.0827875137,0.2307296842,0.1615782082,-0.1797794253,0.4915136993,-0.3460828662,0.0986247435,0.5306717157,-0.0260191523,0.1837768108,0.6057760715,0.1359435916,0.1512895077,0.0835235864,0.1909211874,0.2960536778,0.3395085037,0.2108195573,-0.0150642218,-0.282137394,-0.3250028491,0.3341210485,-0.0021029513,-0.0231160782,0.2038309872,0.0800835639,-0.2553889751,0.2971126735,-0.3381642997,1.148299098,0.0942292139,0.1294936687,-0.1617401689,-0.252526015,0.664984107,-0.1067686155,0.1237175688,-0.046933569,-0.1145089716,-0.1157739386,-0.1703818738,-0.211443305,0.0383760333,-0.2149146348,0.5331195593,0.075566791,-0.0113159101,0.1618145704,0.4351913631,-0.1023060456,-0.0456134863,-0.5257042646,0.0877486318,-0.0442999899,0.2518552244,-0.0140865427,-0.1485861987,-0.2059692591,0.005314081,-0.4601523578,0.1955182999,-0.6179574132,0.2456762493,-0.0609878637,-0.3103728592,-0.1943768859,-0.0575052127,0.1287858635,-0.0164992493,-0.3511244655,0.0874209404,0.2061311007,-0.1411281377,0.1736903042,0.0441385135,0.1641439348,-0.1359499395,-0.0259148777,-0.0478292927,0.0252410229,-0.3272763789,0.0321033821,-0.1285140514,0.0396090299,0.0648923889,-0.1876660287,-0.1389926225,-0.1790039539,-0.2322729975,0.0341174826,-0.0430926159,0.1417649537,-0.085842438,0.3707483113,-0.2654457986,-0.16239582,0.5142537951,-0.305388242,0.0012647071,0.339337945,0.3020931184,-0.3583815396,-0.0925878659,-0.1114680618,-0.1822286099,-0.3933390677,-0.0003344097,-0.2642868161,0.0666865036,-0.0891283676,0.0259221327,0.1505150199,0.1860919148,-0.1181158945,-0.6274639964,-0.1790275872,0.4015606344,-0.0270387996,0.0485015772,-0.2964761555,0.1652342081,0.4207263887,-0.0386223979,-0.2305992246,0.1560420096,-0.44154194,0.0104385708,0.3247368336,-0.3346687257,0.0485885106,-0.0424531437,0.0563483611,0.0833296478,-0.1790658981,-0.0608192533,-0.1101867259,0.1974409521,0.0511796996,-0.3207353354,-0.0964926332,-0.0979493782,0.2066013664,-0.0818443969,-0.0759881213,-0.0132816853,0.0211339984,-0.2951012254,0.1266241819,-0.0999270082,-0.1372247636,0.3212115169,-0.0065158373,0.2825662494,0.3173856437,0.1064764485,0.0883074254,-0.0717145652,-0.0931007713,-0.1241110042,-0.0829002708,0.1949430406,0.4736431241,-0.0916405395,0.0512022041,0.0569887497,0.3979811966,0.5058012009,0.0556126125,-0.1075799018,0.1179754958,0.0908082947,0.0191558711,0.251812458,0.2627199292,0.1216957122,0.0234896373,0.097651422,0.0498040468,-0.0724791065,0.0851857215,0.075487569,0.6589386463,0.201536268,-0.0219140518,-0.1459535807,0.0476535223,-0.0527363233,0.0410787389,-0.0996239334,0.0687503144,-0.126663357,0.0534229763,-0.0639401227,-0.1313260347,0.3249314427,0.2482800186,-0.2003667951,-0.0873381868,0.2849595845,0.0983348265,0.1369969398,-0.220414117,0.3333672583,-0.1611330211,0.0659117475,-0.1380514205,0.4353375435,0.1926468909,-0.3216380179,0.0582479835,-0.3349765539,-0.4865922332,0.0540139452,-0.1999975294,-0.1715221703,0.1560159922,0.2826383114,0.0135858925,-0.7253261805,-0.1404952407,-0.1912470907,0.232984975,-0.3082934618,0.4661635458,-0.1038244143,0.0777674317,-0.0195026733,0.1526183188,0.3781436086,0.3050530851,-0.4279482365,0.0964921415,0.2019785941,-0.0506129041,-0.0096543431,0.230298087,0.4857646525,-0.1341400892,0.492813617,0.0891050696,-0.1842339635,0.2686845064,-0.0354273506,0.1412737072,-0.0751045942,0.2763312161,-0.0875300467,-0.0513294786,-0.2288811356,-0.1396883428,-0.5556668043,-0.109657459,-0.0335210189,-0.5973433256,0.0323545001,-0.0905735418,0.04113717,0.0053061331,0.6298923492,0.569093585,0.5683688521,-0.380954206,-0.3364817202,-0.3252815902,0.0403116867,-0.1251098812,-0.0124700721,-0.1701546311,0.2908905745,-0.174534604,0.2047268748,-0.0137992641,0.6748144627,-0.1523853242,0.1743044406,-0.2713821232,-0.0484785065,-0.0313812792,0.0573393144,-0.1030781344,-0.0273058582,0.0516437143,-0.5177112818,-0.0516986661,0.2759017646,-0.2825129032,0.133970499,0.236651957,0.3119557798,0.1032110676,0.7628632784,-0.122793287,-0.1470048726,-0.4071486294,-0.6882551908,-0.0240677781,-0.1192925423,0.0928619131,0.1546040028,-0.0933201984,0.2021304667,-0.1746087968,0.190074876,-0.0879273117,0.4516707063,0.1424894631,-0.3678409457,0.065007776,-0.0495072305,0.0439469479,-0.0063381037,-0.2057438493,0.1007037237,-0.1159259304,-0.3511934578,0.5208379626,-0.0540384129,-0.0175133403,-0.1145703346,0.2055766732,0.0688287094,0.2129953802,-0.3332384527,-0.2800841331,0.344409585,0.058339078,-0.4353078008,0.2324451804,-0.0068449518,-0.0778456554,-0.0766064376,-0.0418283306,0.2216898501,-0.218432188,-0.192988202,-0.0678091273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/575","title":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.","comments":"I am also seeing a similar error when running the following:\r\n\r\n```\r\nimport nlp\r\ndataset = load_dataset('cola')\r\n```\r\nError:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/js11133\/.conda\/envs\/jiant\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 509, in load_dataset\r\n module_path = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/home\/js11133\/.conda\/envs\/jiant\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 248, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/js11133\/.conda\/envs\/jiant\/lib\/python3.8\/site-packages\/nlp\/utils\/file_utils.py\", line 191, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/js11133\/.conda\/envs\/jiant\/lib\/python3.8\/site-packages\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cola\/cola.py\r\n```","body":"Hi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n","comment_length":76,"text":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.\nHi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n\nI am also seeing a similar error when running the following:\r\n\r\n```\r\nimport nlp\r\ndataset = load_dataset('cola')\r\n```\r\nError:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/js11133\/.conda\/envs\/jiant\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 509, in load_dataset\r\n module_path = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/home\/js11133\/.conda\/envs\/jiant\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 248, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/home\/js11133\/.conda\/envs\/jiant\/lib\/python3.8\/site-packages\/nlp\/utils\/file_utils.py\", line 191, in cached_path\r\n output_path = get_from_cache(\r\n File \"\/home\/js11133\/.conda\/envs\/jiant\/lib\/python3.8\/site-packages\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cola\/cola.py\r\n```","embeddings":[-0.112636596,-0.0148180015,-0.0514469258,0.2499045283,0.2150651067,-0.1080487967,-0.0023050923,0.1171098575,0.144177258,-0.1348839551,-0.425668329,-0.0689154193,0.1447221488,0.1289693564,0.274225086,-0.0121180899,-0.1593028903,-0.0744765624,-0.0916164145,0.017701054,-0.2575627863,0.3606531918,-0.2036087364,0.1764502376,-0.1922911257,-0.0935443118,-0.1606316715,0.1398071498,-0.1229681745,-0.3245210052,0.0947918966,0.3048549592,0.0481997393,0.3857811987,-0.0001219456,-0.0564453825,0.5723124146,-0.0555414632,-0.3910080492,-0.4353305697,-0.2779269516,-0.0589925274,0.355984658,-0.0667517558,0.130753994,0.1634606421,0.0878987014,-0.2419316471,0.2720347047,0.26318115,0.1377917379,0.0733033642,0.1359027624,0.0714743361,0.3174403608,-0.2632397711,0.0046695629,0.3708156645,0.3326888382,-0.291418612,0.0472086743,0.0577220842,-0.205646798,0.0583759286,0.1841480434,0.1635258496,-0.25784266,-0.5462059379,0.0292282049,0.3213647902,0.2748629451,0.0140512539,-0.4840654433,-0.3165412545,-0.0286721792,0.1173614413,0.2789925039,0.1334600002,-0.2538383007,0.0309729725,-0.5450456142,-0.1321876645,0.0033698108,0.5532464981,0.1101105139,-0.0098115858,0.0639427006,0.1689395458,0.3128620982,-0.0395253114,0.1148340628,-0.0232361071,0.2994754314,0.1872622669,-0.0450580493,-0.0265388358,0.0512051396,0.4236578643,0.1454071999,0.1372276694,0.1992809027,0.0405518636,-0.1392972767,0.1321616024,0.15483661,-0.0140559683,-0.0127521874,0.0798756033,0.6293030977,0.3603568673,-0.1680348217,0.0711006969,-0.1136864573,-0.0785518587,-0.3989927769,-0.0077087539,-0.1199624836,-0.284093976,-0.1622226536,-0.0771386921,-0.1368774027,0.0150720552,0.0852001831,0.3928332329,-0.3108938336,0.2066407949,-0.087643005,0.3793320358,-0.1213530898,0.0319333375,-0.0930832773,0.1856428087,-0.1889153719,0.2517310679,0.4859232306,0.0784364417,0.3209574521,-0.3183521628,-0.1498351097,-0.2291524559,0.0177800152,-0.3440533578,0.0556664392,0.0532136112,0.3014571071,0.3389762938,-0.0610685609,-0.1870903969,-0.1006215811,-0.2109535187,-0.2240741253,-0.1662695557,0.3674656749,0.1121519953,-0.4041662812,-0.1559494436,-0.1344938278,-0.004829342,-0.0961971432,-0.0948827863,-0.1139309928,-0.0595751256,-0.1584257782,-0.200951606,0.3062122166,0.6965598464,-0.1322715133,-0.2766490877,-0.1275884658,0.041212976,-0.0782895386,0.502037406,-0.0070996187,0.1006353199,-0.4256356955,0.2437322885,0.5672342777,-0.2894794941,-0.3902561069,0.5320960879,-0.0154582411,0.0364813991,0.1849975735,0.0225609727,0.131248787,-0.0723032951,0.4729860723,0.5410932899,-0.0449328944,-0.0970429629,-0.2779184282,-0.1605929285,0.1289730817,0.1257047504,0.223503828,0.0212896317,-0.0032729721,-0.079706423,0.233262375,0.2292817086,0.0552493818,0.2484004796,-0.1013187617,0.1788349301,-0.104797557,-0.1381653398,-0.3142169118,0.2160550207,-0.0629844964,0.1218500361,-0.3327453732,0.2367781252,-0.1883798987,-0.1930935532,0.0038398847,-0.0187660865,0.0700894296,0.2044757456,0.131302923,0.0753203556,-0.2522730231,0.5734852552,0.0272469651,0.1724773198,-0.316007942,0.6796264052,-0.1542854309,-0.191763401,0.1090584621,0.3169533312,0.2635817826,-0.2635529935,0.0134699494,0.3022545278,-0.1289988309,0.1917850226,0.2702765465,0.0180154573,0.3637427986,-0.2720222771,0.1126302034,0.0491894484,0.1340863556,-0.1252143681,0.1797020435,0.2376711965,-0.2186419368,0.6319778562,0.4755757749,-0.0091407485,0.1447475404,-0.1831095815,-0.2134258151,-0.2173276395,0.6007691026,0.2402251363,0.2317631692,-0.1197275817,-0.148830682,0.1870426685,0.3490749896,0.0156644378,0.05569648,0.2178222686,-0.1637909412,-0.1998849362,-0.1010185257,0.3804479241,0.2209589183,0.1598923951,0.1135150269,0.021280108,-0.2521316409,-0.3306078017,0.2030390054,0.1683373004,-0.0889684558,0.0819643363,0.2019196898,0.1494818777,-0.0947707817,-0.4382834136,0.0053283786,-0.0233714301,-0.3714630306,-0.0647940934,-0.445727706,-0.6604163051,0.0467646122,-0.1720999628,-0.2792548835,-0.646451056,-0.1037925184,0.2034947723,0.0090118954,0.0356713496,-0.2063030154,-0.0062600239,-0.2111667097,-0.0129495598,-0.1801105589,-0.3260889649,-0.062903434,0.0395989567,0.3534008265,0.0574423596,0.1485339999,-0.2441104352,-0.1611087322,-0.1612910032,-0.1101251915,0.19283548,-0.1821724176,0.3608709574,0.2109528184,0.4903079271,0.2029083073,-0.2494547963,0.3121628463,-0.1433433294,0.0774954781,-0.0519493446,0.0077687465,-0.0471082143,0.1043553874,-0.0099998713,-0.471308589,-0.4650456309,0.2359144241,-0.0598598197,0.2176345289,0.1203419119,-0.041052755,0.2929945886,-0.0266634505,0.1280004233,-0.041178517,-0.3354517817,0.5650700331,-0.2066883743,-0.2000689805,-0.0150570692,0.0398590788,-0.2016781569,0.0807290226,-0.706532836,-0.2792837024,-0.4299277067,0.0273614805,0.2272491306,-0.0665315464,-0.0501204729,-0.2217652798,-0.0760565326,-0.0140878279,-0.4060663879,-0.0029456718,0.3619810045,0.3475250006,0.1065041944,0.5122767091,-0.0753890201,0.272541374,-0.0739138052,0.0747887865,0.6189810038,0.0821451545,0.3144127727,-0.1913370788,-0.269040674,0.1089694574,-0.3142771423,0.0521760285,0.1350190192,0.2109739482,-0.3150362074,-0.5289624333,0.172768876,-0.4613894224,-0.1889805794,0.0283963811,-0.0316216201,-0.2145553231,0.2021680921,0.035619285,-0.0524334349,-0.2252569646,-0.0457172729,0.1494725794,0.2521126568,0.039624352,-0.1105956882,-0.1379540712,-0.4256444275,0.2749575973,0.1958890855,0.3104603589,-0.2696570158,-0.0499456488,0.0475213565,-0.0248350445,0.5041225553,-0.4135855436,0.2516334057,0.1580861807,0.0192981642,-0.6474230289,0.0887410268,-0.1247085482,0.1403576285,0.6630892158,0.2826699615,-0.1967948675,0.1196002215,-0.0139652062,0.1940147877,0.1074987426,0.0746590942,-0.0067794248,-0.2686884701,-0.2585899532,-0.2295411676,-0.0385957845,0.0707760081,-0.0640727878,-0.3239626288,0.3134526014,-0.4003784955,-0.0395804793,0.0364644267,0.1761945337,-0.107365407,0.0827875137,0.2307296842,0.1615782082,-0.1797794253,0.4915136993,-0.3460828662,0.0986247435,0.5306717157,-0.0260191523,0.1837768108,0.6057760715,0.1359435916,0.1512895077,0.0835235864,0.1909211874,0.2960536778,0.3395085037,0.2108195573,-0.0150642218,-0.282137394,-0.3250028491,0.3341210485,-0.0021029513,-0.0231160782,0.2038309872,0.0800835639,-0.2553889751,0.2971126735,-0.3381642997,1.148299098,0.0942292139,0.1294936687,-0.1617401689,-0.252526015,0.664984107,-0.1067686155,0.1237175688,-0.046933569,-0.1145089716,-0.1157739386,-0.1703818738,-0.211443305,0.0383760333,-0.2149146348,0.5331195593,0.075566791,-0.0113159101,0.1618145704,0.4351913631,-0.1023060456,-0.0456134863,-0.5257042646,0.0877486318,-0.0442999899,0.2518552244,-0.0140865427,-0.1485861987,-0.2059692591,0.005314081,-0.4601523578,0.1955182999,-0.6179574132,0.2456762493,-0.0609878637,-0.3103728592,-0.1943768859,-0.0575052127,0.1287858635,-0.0164992493,-0.3511244655,0.0874209404,0.2061311007,-0.1411281377,0.1736903042,0.0441385135,0.1641439348,-0.1359499395,-0.0259148777,-0.0478292927,0.0252410229,-0.3272763789,0.0321033821,-0.1285140514,0.0396090299,0.0648923889,-0.1876660287,-0.1389926225,-0.1790039539,-0.2322729975,0.0341174826,-0.0430926159,0.1417649537,-0.085842438,0.3707483113,-0.2654457986,-0.16239582,0.5142537951,-0.305388242,0.0012647071,0.339337945,0.3020931184,-0.3583815396,-0.0925878659,-0.1114680618,-0.1822286099,-0.3933390677,-0.0003344097,-0.2642868161,0.0666865036,-0.0891283676,0.0259221327,0.1505150199,0.1860919148,-0.1181158945,-0.6274639964,-0.1790275872,0.4015606344,-0.0270387996,0.0485015772,-0.2964761555,0.1652342081,0.4207263887,-0.0386223979,-0.2305992246,0.1560420096,-0.44154194,0.0104385708,0.3247368336,-0.3346687257,0.0485885106,-0.0424531437,0.0563483611,0.0833296478,-0.1790658981,-0.0608192533,-0.1101867259,0.1974409521,0.0511796996,-0.3207353354,-0.0964926332,-0.0979493782,0.2066013664,-0.0818443969,-0.0759881213,-0.0132816853,0.0211339984,-0.2951012254,0.1266241819,-0.0999270082,-0.1372247636,0.3212115169,-0.0065158373,0.2825662494,0.3173856437,0.1064764485,0.0883074254,-0.0717145652,-0.0931007713,-0.1241110042,-0.0829002708,0.1949430406,0.4736431241,-0.0916405395,0.0512022041,0.0569887497,0.3979811966,0.5058012009,0.0556126125,-0.1075799018,0.1179754958,0.0908082947,0.0191558711,0.251812458,0.2627199292,0.1216957122,0.0234896373,0.097651422,0.0498040468,-0.0724791065,0.0851857215,0.075487569,0.6589386463,0.201536268,-0.0219140518,-0.1459535807,0.0476535223,-0.0527363233,0.0410787389,-0.0996239334,0.0687503144,-0.126663357,0.0534229763,-0.0639401227,-0.1313260347,0.3249314427,0.2482800186,-0.2003667951,-0.0873381868,0.2849595845,0.0983348265,0.1369969398,-0.220414117,0.3333672583,-0.1611330211,0.0659117475,-0.1380514205,0.4353375435,0.1926468909,-0.3216380179,0.0582479835,-0.3349765539,-0.4865922332,0.0540139452,-0.1999975294,-0.1715221703,0.1560159922,0.2826383114,0.0135858925,-0.7253261805,-0.1404952407,-0.1912470907,0.232984975,-0.3082934618,0.4661635458,-0.1038244143,0.0777674317,-0.0195026733,0.1526183188,0.3781436086,0.3050530851,-0.4279482365,0.0964921415,0.2019785941,-0.0506129041,-0.0096543431,0.230298087,0.4857646525,-0.1341400892,0.492813617,0.0891050696,-0.1842339635,0.2686845064,-0.0354273506,0.1412737072,-0.0751045942,0.2763312161,-0.0875300467,-0.0513294786,-0.2288811356,-0.1396883428,-0.5556668043,-0.109657459,-0.0335210189,-0.5973433256,0.0323545001,-0.0905735418,0.04113717,0.0053061331,0.6298923492,0.569093585,0.5683688521,-0.380954206,-0.3364817202,-0.3252815902,0.0403116867,-0.1251098812,-0.0124700721,-0.1701546311,0.2908905745,-0.174534604,0.2047268748,-0.0137992641,0.6748144627,-0.1523853242,0.1743044406,-0.2713821232,-0.0484785065,-0.0313812792,0.0573393144,-0.1030781344,-0.0273058582,0.0516437143,-0.5177112818,-0.0516986661,0.2759017646,-0.2825129032,0.133970499,0.236651957,0.3119557798,0.1032110676,0.7628632784,-0.122793287,-0.1470048726,-0.4071486294,-0.6882551908,-0.0240677781,-0.1192925423,0.0928619131,0.1546040028,-0.0933201984,0.2021304667,-0.1746087968,0.190074876,-0.0879273117,0.4516707063,0.1424894631,-0.3678409457,0.065007776,-0.0495072305,0.0439469479,-0.0063381037,-0.2057438493,0.1007037237,-0.1159259304,-0.3511934578,0.5208379626,-0.0540384129,-0.0175133403,-0.1145703346,0.2055766732,0.0688287094,0.2129953802,-0.3332384527,-0.2800841331,0.344409585,0.058339078,-0.4353078008,0.2324451804,-0.0068449518,-0.0778456554,-0.0766064376,-0.0418283306,0.2216898501,-0.218432188,-0.192988202,-0.0678091273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/575","title":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.","comments":"@jeswan `\"cola\"` is not a valid dataset identifier (you can check the up-to-date list on https:\/\/huggingface.co\/datasets) but you can find cola inside glue.","body":"Hi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n","comment_length":23,"text":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.\nHi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n\n@jeswan `\"cola\"` is not a valid dataset identifier (you can check the up-to-date list on https:\/\/huggingface.co\/datasets) but you can find cola inside glue.","embeddings":[-0.112636596,-0.0148180015,-0.0514469258,0.2499045283,0.2150651067,-0.1080487967,-0.0023050923,0.1171098575,0.144177258,-0.1348839551,-0.425668329,-0.0689154193,0.1447221488,0.1289693564,0.274225086,-0.0121180899,-0.1593028903,-0.0744765624,-0.0916164145,0.017701054,-0.2575627863,0.3606531918,-0.2036087364,0.1764502376,-0.1922911257,-0.0935443118,-0.1606316715,0.1398071498,-0.1229681745,-0.3245210052,0.0947918966,0.3048549592,0.0481997393,0.3857811987,-0.0001219456,-0.0564453825,0.5723124146,-0.0555414632,-0.3910080492,-0.4353305697,-0.2779269516,-0.0589925274,0.355984658,-0.0667517558,0.130753994,0.1634606421,0.0878987014,-0.2419316471,0.2720347047,0.26318115,0.1377917379,0.0733033642,0.1359027624,0.0714743361,0.3174403608,-0.2632397711,0.0046695629,0.3708156645,0.3326888382,-0.291418612,0.0472086743,0.0577220842,-0.205646798,0.0583759286,0.1841480434,0.1635258496,-0.25784266,-0.5462059379,0.0292282049,0.3213647902,0.2748629451,0.0140512539,-0.4840654433,-0.3165412545,-0.0286721792,0.1173614413,0.2789925039,0.1334600002,-0.2538383007,0.0309729725,-0.5450456142,-0.1321876645,0.0033698108,0.5532464981,0.1101105139,-0.0098115858,0.0639427006,0.1689395458,0.3128620982,-0.0395253114,0.1148340628,-0.0232361071,0.2994754314,0.1872622669,-0.0450580493,-0.0265388358,0.0512051396,0.4236578643,0.1454071999,0.1372276694,0.1992809027,0.0405518636,-0.1392972767,0.1321616024,0.15483661,-0.0140559683,-0.0127521874,0.0798756033,0.6293030977,0.3603568673,-0.1680348217,0.0711006969,-0.1136864573,-0.0785518587,-0.3989927769,-0.0077087539,-0.1199624836,-0.284093976,-0.1622226536,-0.0771386921,-0.1368774027,0.0150720552,0.0852001831,0.3928332329,-0.3108938336,0.2066407949,-0.087643005,0.3793320358,-0.1213530898,0.0319333375,-0.0930832773,0.1856428087,-0.1889153719,0.2517310679,0.4859232306,0.0784364417,0.3209574521,-0.3183521628,-0.1498351097,-0.2291524559,0.0177800152,-0.3440533578,0.0556664392,0.0532136112,0.3014571071,0.3389762938,-0.0610685609,-0.1870903969,-0.1006215811,-0.2109535187,-0.2240741253,-0.1662695557,0.3674656749,0.1121519953,-0.4041662812,-0.1559494436,-0.1344938278,-0.004829342,-0.0961971432,-0.0948827863,-0.1139309928,-0.0595751256,-0.1584257782,-0.200951606,0.3062122166,0.6965598464,-0.1322715133,-0.2766490877,-0.1275884658,0.041212976,-0.0782895386,0.502037406,-0.0070996187,0.1006353199,-0.4256356955,0.2437322885,0.5672342777,-0.2894794941,-0.3902561069,0.5320960879,-0.0154582411,0.0364813991,0.1849975735,0.0225609727,0.131248787,-0.0723032951,0.4729860723,0.5410932899,-0.0449328944,-0.0970429629,-0.2779184282,-0.1605929285,0.1289730817,0.1257047504,0.223503828,0.0212896317,-0.0032729721,-0.079706423,0.233262375,0.2292817086,0.0552493818,0.2484004796,-0.1013187617,0.1788349301,-0.104797557,-0.1381653398,-0.3142169118,0.2160550207,-0.0629844964,0.1218500361,-0.3327453732,0.2367781252,-0.1883798987,-0.1930935532,0.0038398847,-0.0187660865,0.0700894296,0.2044757456,0.131302923,0.0753203556,-0.2522730231,0.5734852552,0.0272469651,0.1724773198,-0.316007942,0.6796264052,-0.1542854309,-0.191763401,0.1090584621,0.3169533312,0.2635817826,-0.2635529935,0.0134699494,0.3022545278,-0.1289988309,0.1917850226,0.2702765465,0.0180154573,0.3637427986,-0.2720222771,0.1126302034,0.0491894484,0.1340863556,-0.1252143681,0.1797020435,0.2376711965,-0.2186419368,0.6319778562,0.4755757749,-0.0091407485,0.1447475404,-0.1831095815,-0.2134258151,-0.2173276395,0.6007691026,0.2402251363,0.2317631692,-0.1197275817,-0.148830682,0.1870426685,0.3490749896,0.0156644378,0.05569648,0.2178222686,-0.1637909412,-0.1998849362,-0.1010185257,0.3804479241,0.2209589183,0.1598923951,0.1135150269,0.021280108,-0.2521316409,-0.3306078017,0.2030390054,0.1683373004,-0.0889684558,0.0819643363,0.2019196898,0.1494818777,-0.0947707817,-0.4382834136,0.0053283786,-0.0233714301,-0.3714630306,-0.0647940934,-0.445727706,-0.6604163051,0.0467646122,-0.1720999628,-0.2792548835,-0.646451056,-0.1037925184,0.2034947723,0.0090118954,0.0356713496,-0.2063030154,-0.0062600239,-0.2111667097,-0.0129495598,-0.1801105589,-0.3260889649,-0.062903434,0.0395989567,0.3534008265,0.0574423596,0.1485339999,-0.2441104352,-0.1611087322,-0.1612910032,-0.1101251915,0.19283548,-0.1821724176,0.3608709574,0.2109528184,0.4903079271,0.2029083073,-0.2494547963,0.3121628463,-0.1433433294,0.0774954781,-0.0519493446,0.0077687465,-0.0471082143,0.1043553874,-0.0099998713,-0.471308589,-0.4650456309,0.2359144241,-0.0598598197,0.2176345289,0.1203419119,-0.041052755,0.2929945886,-0.0266634505,0.1280004233,-0.041178517,-0.3354517817,0.5650700331,-0.2066883743,-0.2000689805,-0.0150570692,0.0398590788,-0.2016781569,0.0807290226,-0.706532836,-0.2792837024,-0.4299277067,0.0273614805,0.2272491306,-0.0665315464,-0.0501204729,-0.2217652798,-0.0760565326,-0.0140878279,-0.4060663879,-0.0029456718,0.3619810045,0.3475250006,0.1065041944,0.5122767091,-0.0753890201,0.272541374,-0.0739138052,0.0747887865,0.6189810038,0.0821451545,0.3144127727,-0.1913370788,-0.269040674,0.1089694574,-0.3142771423,0.0521760285,0.1350190192,0.2109739482,-0.3150362074,-0.5289624333,0.172768876,-0.4613894224,-0.1889805794,0.0283963811,-0.0316216201,-0.2145553231,0.2021680921,0.035619285,-0.0524334349,-0.2252569646,-0.0457172729,0.1494725794,0.2521126568,0.039624352,-0.1105956882,-0.1379540712,-0.4256444275,0.2749575973,0.1958890855,0.3104603589,-0.2696570158,-0.0499456488,0.0475213565,-0.0248350445,0.5041225553,-0.4135855436,0.2516334057,0.1580861807,0.0192981642,-0.6474230289,0.0887410268,-0.1247085482,0.1403576285,0.6630892158,0.2826699615,-0.1967948675,0.1196002215,-0.0139652062,0.1940147877,0.1074987426,0.0746590942,-0.0067794248,-0.2686884701,-0.2585899532,-0.2295411676,-0.0385957845,0.0707760081,-0.0640727878,-0.3239626288,0.3134526014,-0.4003784955,-0.0395804793,0.0364644267,0.1761945337,-0.107365407,0.0827875137,0.2307296842,0.1615782082,-0.1797794253,0.4915136993,-0.3460828662,0.0986247435,0.5306717157,-0.0260191523,0.1837768108,0.6057760715,0.1359435916,0.1512895077,0.0835235864,0.1909211874,0.2960536778,0.3395085037,0.2108195573,-0.0150642218,-0.282137394,-0.3250028491,0.3341210485,-0.0021029513,-0.0231160782,0.2038309872,0.0800835639,-0.2553889751,0.2971126735,-0.3381642997,1.148299098,0.0942292139,0.1294936687,-0.1617401689,-0.252526015,0.664984107,-0.1067686155,0.1237175688,-0.046933569,-0.1145089716,-0.1157739386,-0.1703818738,-0.211443305,0.0383760333,-0.2149146348,0.5331195593,0.075566791,-0.0113159101,0.1618145704,0.4351913631,-0.1023060456,-0.0456134863,-0.5257042646,0.0877486318,-0.0442999899,0.2518552244,-0.0140865427,-0.1485861987,-0.2059692591,0.005314081,-0.4601523578,0.1955182999,-0.6179574132,0.2456762493,-0.0609878637,-0.3103728592,-0.1943768859,-0.0575052127,0.1287858635,-0.0164992493,-0.3511244655,0.0874209404,0.2061311007,-0.1411281377,0.1736903042,0.0441385135,0.1641439348,-0.1359499395,-0.0259148777,-0.0478292927,0.0252410229,-0.3272763789,0.0321033821,-0.1285140514,0.0396090299,0.0648923889,-0.1876660287,-0.1389926225,-0.1790039539,-0.2322729975,0.0341174826,-0.0430926159,0.1417649537,-0.085842438,0.3707483113,-0.2654457986,-0.16239582,0.5142537951,-0.305388242,0.0012647071,0.339337945,0.3020931184,-0.3583815396,-0.0925878659,-0.1114680618,-0.1822286099,-0.3933390677,-0.0003344097,-0.2642868161,0.0666865036,-0.0891283676,0.0259221327,0.1505150199,0.1860919148,-0.1181158945,-0.6274639964,-0.1790275872,0.4015606344,-0.0270387996,0.0485015772,-0.2964761555,0.1652342081,0.4207263887,-0.0386223979,-0.2305992246,0.1560420096,-0.44154194,0.0104385708,0.3247368336,-0.3346687257,0.0485885106,-0.0424531437,0.0563483611,0.0833296478,-0.1790658981,-0.0608192533,-0.1101867259,0.1974409521,0.0511796996,-0.3207353354,-0.0964926332,-0.0979493782,0.2066013664,-0.0818443969,-0.0759881213,-0.0132816853,0.0211339984,-0.2951012254,0.1266241819,-0.0999270082,-0.1372247636,0.3212115169,-0.0065158373,0.2825662494,0.3173856437,0.1064764485,0.0883074254,-0.0717145652,-0.0931007713,-0.1241110042,-0.0829002708,0.1949430406,0.4736431241,-0.0916405395,0.0512022041,0.0569887497,0.3979811966,0.5058012009,0.0556126125,-0.1075799018,0.1179754958,0.0908082947,0.0191558711,0.251812458,0.2627199292,0.1216957122,0.0234896373,0.097651422,0.0498040468,-0.0724791065,0.0851857215,0.075487569,0.6589386463,0.201536268,-0.0219140518,-0.1459535807,0.0476535223,-0.0527363233,0.0410787389,-0.0996239334,0.0687503144,-0.126663357,0.0534229763,-0.0639401227,-0.1313260347,0.3249314427,0.2482800186,-0.2003667951,-0.0873381868,0.2849595845,0.0983348265,0.1369969398,-0.220414117,0.3333672583,-0.1611330211,0.0659117475,-0.1380514205,0.4353375435,0.1926468909,-0.3216380179,0.0582479835,-0.3349765539,-0.4865922332,0.0540139452,-0.1999975294,-0.1715221703,0.1560159922,0.2826383114,0.0135858925,-0.7253261805,-0.1404952407,-0.1912470907,0.232984975,-0.3082934618,0.4661635458,-0.1038244143,0.0777674317,-0.0195026733,0.1526183188,0.3781436086,0.3050530851,-0.4279482365,0.0964921415,0.2019785941,-0.0506129041,-0.0096543431,0.230298087,0.4857646525,-0.1341400892,0.492813617,0.0891050696,-0.1842339635,0.2686845064,-0.0354273506,0.1412737072,-0.0751045942,0.2763312161,-0.0875300467,-0.0513294786,-0.2288811356,-0.1396883428,-0.5556668043,-0.109657459,-0.0335210189,-0.5973433256,0.0323545001,-0.0905735418,0.04113717,0.0053061331,0.6298923492,0.569093585,0.5683688521,-0.380954206,-0.3364817202,-0.3252815902,0.0403116867,-0.1251098812,-0.0124700721,-0.1701546311,0.2908905745,-0.174534604,0.2047268748,-0.0137992641,0.6748144627,-0.1523853242,0.1743044406,-0.2713821232,-0.0484785065,-0.0313812792,0.0573393144,-0.1030781344,-0.0273058582,0.0516437143,-0.5177112818,-0.0516986661,0.2759017646,-0.2825129032,0.133970499,0.236651957,0.3119557798,0.1032110676,0.7628632784,-0.122793287,-0.1470048726,-0.4071486294,-0.6882551908,-0.0240677781,-0.1192925423,0.0928619131,0.1546040028,-0.0933201984,0.2021304667,-0.1746087968,0.190074876,-0.0879273117,0.4516707063,0.1424894631,-0.3678409457,0.065007776,-0.0495072305,0.0439469479,-0.0063381037,-0.2057438493,0.1007037237,-0.1159259304,-0.3511934578,0.5208379626,-0.0540384129,-0.0175133403,-0.1145703346,0.2055766732,0.0688287094,0.2129953802,-0.3332384527,-0.2800841331,0.344409585,0.058339078,-0.4353078008,0.2324451804,-0.0068449518,-0.0778456554,-0.0766064376,-0.0418283306,0.2216898501,-0.218432188,-0.192988202,-0.0678091273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/575","title":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.","comments":"Ah right. Thanks!","body":"Hi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n","comment_length":3,"text":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.\nHi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n\nAh right. Thanks!","embeddings":[-0.112636596,-0.0148180015,-0.0514469258,0.2499045283,0.2150651067,-0.1080487967,-0.0023050923,0.1171098575,0.144177258,-0.1348839551,-0.425668329,-0.0689154193,0.1447221488,0.1289693564,0.274225086,-0.0121180899,-0.1593028903,-0.0744765624,-0.0916164145,0.017701054,-0.2575627863,0.3606531918,-0.2036087364,0.1764502376,-0.1922911257,-0.0935443118,-0.1606316715,0.1398071498,-0.1229681745,-0.3245210052,0.0947918966,0.3048549592,0.0481997393,0.3857811987,-0.0001219456,-0.0564453825,0.5723124146,-0.0555414632,-0.3910080492,-0.4353305697,-0.2779269516,-0.0589925274,0.355984658,-0.0667517558,0.130753994,0.1634606421,0.0878987014,-0.2419316471,0.2720347047,0.26318115,0.1377917379,0.0733033642,0.1359027624,0.0714743361,0.3174403608,-0.2632397711,0.0046695629,0.3708156645,0.3326888382,-0.291418612,0.0472086743,0.0577220842,-0.205646798,0.0583759286,0.1841480434,0.1635258496,-0.25784266,-0.5462059379,0.0292282049,0.3213647902,0.2748629451,0.0140512539,-0.4840654433,-0.3165412545,-0.0286721792,0.1173614413,0.2789925039,0.1334600002,-0.2538383007,0.0309729725,-0.5450456142,-0.1321876645,0.0033698108,0.5532464981,0.1101105139,-0.0098115858,0.0639427006,0.1689395458,0.3128620982,-0.0395253114,0.1148340628,-0.0232361071,0.2994754314,0.1872622669,-0.0450580493,-0.0265388358,0.0512051396,0.4236578643,0.1454071999,0.1372276694,0.1992809027,0.0405518636,-0.1392972767,0.1321616024,0.15483661,-0.0140559683,-0.0127521874,0.0798756033,0.6293030977,0.3603568673,-0.1680348217,0.0711006969,-0.1136864573,-0.0785518587,-0.3989927769,-0.0077087539,-0.1199624836,-0.284093976,-0.1622226536,-0.0771386921,-0.1368774027,0.0150720552,0.0852001831,0.3928332329,-0.3108938336,0.2066407949,-0.087643005,0.3793320358,-0.1213530898,0.0319333375,-0.0930832773,0.1856428087,-0.1889153719,0.2517310679,0.4859232306,0.0784364417,0.3209574521,-0.3183521628,-0.1498351097,-0.2291524559,0.0177800152,-0.3440533578,0.0556664392,0.0532136112,0.3014571071,0.3389762938,-0.0610685609,-0.1870903969,-0.1006215811,-0.2109535187,-0.2240741253,-0.1662695557,0.3674656749,0.1121519953,-0.4041662812,-0.1559494436,-0.1344938278,-0.004829342,-0.0961971432,-0.0948827863,-0.1139309928,-0.0595751256,-0.1584257782,-0.200951606,0.3062122166,0.6965598464,-0.1322715133,-0.2766490877,-0.1275884658,0.041212976,-0.0782895386,0.502037406,-0.0070996187,0.1006353199,-0.4256356955,0.2437322885,0.5672342777,-0.2894794941,-0.3902561069,0.5320960879,-0.0154582411,0.0364813991,0.1849975735,0.0225609727,0.131248787,-0.0723032951,0.4729860723,0.5410932899,-0.0449328944,-0.0970429629,-0.2779184282,-0.1605929285,0.1289730817,0.1257047504,0.223503828,0.0212896317,-0.0032729721,-0.079706423,0.233262375,0.2292817086,0.0552493818,0.2484004796,-0.1013187617,0.1788349301,-0.104797557,-0.1381653398,-0.3142169118,0.2160550207,-0.0629844964,0.1218500361,-0.3327453732,0.2367781252,-0.1883798987,-0.1930935532,0.0038398847,-0.0187660865,0.0700894296,0.2044757456,0.131302923,0.0753203556,-0.2522730231,0.5734852552,0.0272469651,0.1724773198,-0.316007942,0.6796264052,-0.1542854309,-0.191763401,0.1090584621,0.3169533312,0.2635817826,-0.2635529935,0.0134699494,0.3022545278,-0.1289988309,0.1917850226,0.2702765465,0.0180154573,0.3637427986,-0.2720222771,0.1126302034,0.0491894484,0.1340863556,-0.1252143681,0.1797020435,0.2376711965,-0.2186419368,0.6319778562,0.4755757749,-0.0091407485,0.1447475404,-0.1831095815,-0.2134258151,-0.2173276395,0.6007691026,0.2402251363,0.2317631692,-0.1197275817,-0.148830682,0.1870426685,0.3490749896,0.0156644378,0.05569648,0.2178222686,-0.1637909412,-0.1998849362,-0.1010185257,0.3804479241,0.2209589183,0.1598923951,0.1135150269,0.021280108,-0.2521316409,-0.3306078017,0.2030390054,0.1683373004,-0.0889684558,0.0819643363,0.2019196898,0.1494818777,-0.0947707817,-0.4382834136,0.0053283786,-0.0233714301,-0.3714630306,-0.0647940934,-0.445727706,-0.6604163051,0.0467646122,-0.1720999628,-0.2792548835,-0.646451056,-0.1037925184,0.2034947723,0.0090118954,0.0356713496,-0.2063030154,-0.0062600239,-0.2111667097,-0.0129495598,-0.1801105589,-0.3260889649,-0.062903434,0.0395989567,0.3534008265,0.0574423596,0.1485339999,-0.2441104352,-0.1611087322,-0.1612910032,-0.1101251915,0.19283548,-0.1821724176,0.3608709574,0.2109528184,0.4903079271,0.2029083073,-0.2494547963,0.3121628463,-0.1433433294,0.0774954781,-0.0519493446,0.0077687465,-0.0471082143,0.1043553874,-0.0099998713,-0.471308589,-0.4650456309,0.2359144241,-0.0598598197,0.2176345289,0.1203419119,-0.041052755,0.2929945886,-0.0266634505,0.1280004233,-0.041178517,-0.3354517817,0.5650700331,-0.2066883743,-0.2000689805,-0.0150570692,0.0398590788,-0.2016781569,0.0807290226,-0.706532836,-0.2792837024,-0.4299277067,0.0273614805,0.2272491306,-0.0665315464,-0.0501204729,-0.2217652798,-0.0760565326,-0.0140878279,-0.4060663879,-0.0029456718,0.3619810045,0.3475250006,0.1065041944,0.5122767091,-0.0753890201,0.272541374,-0.0739138052,0.0747887865,0.6189810038,0.0821451545,0.3144127727,-0.1913370788,-0.269040674,0.1089694574,-0.3142771423,0.0521760285,0.1350190192,0.2109739482,-0.3150362074,-0.5289624333,0.172768876,-0.4613894224,-0.1889805794,0.0283963811,-0.0316216201,-0.2145553231,0.2021680921,0.035619285,-0.0524334349,-0.2252569646,-0.0457172729,0.1494725794,0.2521126568,0.039624352,-0.1105956882,-0.1379540712,-0.4256444275,0.2749575973,0.1958890855,0.3104603589,-0.2696570158,-0.0499456488,0.0475213565,-0.0248350445,0.5041225553,-0.4135855436,0.2516334057,0.1580861807,0.0192981642,-0.6474230289,0.0887410268,-0.1247085482,0.1403576285,0.6630892158,0.2826699615,-0.1967948675,0.1196002215,-0.0139652062,0.1940147877,0.1074987426,0.0746590942,-0.0067794248,-0.2686884701,-0.2585899532,-0.2295411676,-0.0385957845,0.0707760081,-0.0640727878,-0.3239626288,0.3134526014,-0.4003784955,-0.0395804793,0.0364644267,0.1761945337,-0.107365407,0.0827875137,0.2307296842,0.1615782082,-0.1797794253,0.4915136993,-0.3460828662,0.0986247435,0.5306717157,-0.0260191523,0.1837768108,0.6057760715,0.1359435916,0.1512895077,0.0835235864,0.1909211874,0.2960536778,0.3395085037,0.2108195573,-0.0150642218,-0.282137394,-0.3250028491,0.3341210485,-0.0021029513,-0.0231160782,0.2038309872,0.0800835639,-0.2553889751,0.2971126735,-0.3381642997,1.148299098,0.0942292139,0.1294936687,-0.1617401689,-0.252526015,0.664984107,-0.1067686155,0.1237175688,-0.046933569,-0.1145089716,-0.1157739386,-0.1703818738,-0.211443305,0.0383760333,-0.2149146348,0.5331195593,0.075566791,-0.0113159101,0.1618145704,0.4351913631,-0.1023060456,-0.0456134863,-0.5257042646,0.0877486318,-0.0442999899,0.2518552244,-0.0140865427,-0.1485861987,-0.2059692591,0.005314081,-0.4601523578,0.1955182999,-0.6179574132,0.2456762493,-0.0609878637,-0.3103728592,-0.1943768859,-0.0575052127,0.1287858635,-0.0164992493,-0.3511244655,0.0874209404,0.2061311007,-0.1411281377,0.1736903042,0.0441385135,0.1641439348,-0.1359499395,-0.0259148777,-0.0478292927,0.0252410229,-0.3272763789,0.0321033821,-0.1285140514,0.0396090299,0.0648923889,-0.1876660287,-0.1389926225,-0.1790039539,-0.2322729975,0.0341174826,-0.0430926159,0.1417649537,-0.085842438,0.3707483113,-0.2654457986,-0.16239582,0.5142537951,-0.305388242,0.0012647071,0.339337945,0.3020931184,-0.3583815396,-0.0925878659,-0.1114680618,-0.1822286099,-0.3933390677,-0.0003344097,-0.2642868161,0.0666865036,-0.0891283676,0.0259221327,0.1505150199,0.1860919148,-0.1181158945,-0.6274639964,-0.1790275872,0.4015606344,-0.0270387996,0.0485015772,-0.2964761555,0.1652342081,0.4207263887,-0.0386223979,-0.2305992246,0.1560420096,-0.44154194,0.0104385708,0.3247368336,-0.3346687257,0.0485885106,-0.0424531437,0.0563483611,0.0833296478,-0.1790658981,-0.0608192533,-0.1101867259,0.1974409521,0.0511796996,-0.3207353354,-0.0964926332,-0.0979493782,0.2066013664,-0.0818443969,-0.0759881213,-0.0132816853,0.0211339984,-0.2951012254,0.1266241819,-0.0999270082,-0.1372247636,0.3212115169,-0.0065158373,0.2825662494,0.3173856437,0.1064764485,0.0883074254,-0.0717145652,-0.0931007713,-0.1241110042,-0.0829002708,0.1949430406,0.4736431241,-0.0916405395,0.0512022041,0.0569887497,0.3979811966,0.5058012009,0.0556126125,-0.1075799018,0.1179754958,0.0908082947,0.0191558711,0.251812458,0.2627199292,0.1216957122,0.0234896373,0.097651422,0.0498040468,-0.0724791065,0.0851857215,0.075487569,0.6589386463,0.201536268,-0.0219140518,-0.1459535807,0.0476535223,-0.0527363233,0.0410787389,-0.0996239334,0.0687503144,-0.126663357,0.0534229763,-0.0639401227,-0.1313260347,0.3249314427,0.2482800186,-0.2003667951,-0.0873381868,0.2849595845,0.0983348265,0.1369969398,-0.220414117,0.3333672583,-0.1611330211,0.0659117475,-0.1380514205,0.4353375435,0.1926468909,-0.3216380179,0.0582479835,-0.3349765539,-0.4865922332,0.0540139452,-0.1999975294,-0.1715221703,0.1560159922,0.2826383114,0.0135858925,-0.7253261805,-0.1404952407,-0.1912470907,0.232984975,-0.3082934618,0.4661635458,-0.1038244143,0.0777674317,-0.0195026733,0.1526183188,0.3781436086,0.3050530851,-0.4279482365,0.0964921415,0.2019785941,-0.0506129041,-0.0096543431,0.230298087,0.4857646525,-0.1341400892,0.492813617,0.0891050696,-0.1842339635,0.2686845064,-0.0354273506,0.1412737072,-0.0751045942,0.2763312161,-0.0875300467,-0.0513294786,-0.2288811356,-0.1396883428,-0.5556668043,-0.109657459,-0.0335210189,-0.5973433256,0.0323545001,-0.0905735418,0.04113717,0.0053061331,0.6298923492,0.569093585,0.5683688521,-0.380954206,-0.3364817202,-0.3252815902,0.0403116867,-0.1251098812,-0.0124700721,-0.1701546311,0.2908905745,-0.174534604,0.2047268748,-0.0137992641,0.6748144627,-0.1523853242,0.1743044406,-0.2713821232,-0.0484785065,-0.0313812792,0.0573393144,-0.1030781344,-0.0273058582,0.0516437143,-0.5177112818,-0.0516986661,0.2759017646,-0.2825129032,0.133970499,0.236651957,0.3119557798,0.1032110676,0.7628632784,-0.122793287,-0.1470048726,-0.4071486294,-0.6882551908,-0.0240677781,-0.1192925423,0.0928619131,0.1546040028,-0.0933201984,0.2021304667,-0.1746087968,0.190074876,-0.0879273117,0.4516707063,0.1424894631,-0.3678409457,0.065007776,-0.0495072305,0.0439469479,-0.0063381037,-0.2057438493,0.1007037237,-0.1159259304,-0.3511934578,0.5208379626,-0.0540384129,-0.0175133403,-0.1145703346,0.2055766732,0.0688287094,0.2129953802,-0.3332384527,-0.2800841331,0.344409585,0.058339078,-0.4353078008,0.2324451804,-0.0068449518,-0.0778456554,-0.0766064376,-0.0418283306,0.2216898501,-0.218432188,-0.192988202,-0.0678091273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/575","title":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.","comments":"Hi. Closing this one since #626 updated the glue urls.\r\n\r\n> 1. Why is it still blocking? Is it still downloading?\r\n\r\nAfter downloading it generates the arrow file by iterating through the examples.\r\nThe number of examples processed by second is shown during the processing (not sure why it was not the case for you)\r\n\r\n> 2. I specified split as train, so why is the test folder being populated?\r\n\r\nIt downloads every split\r\n\r\n\r\n\r\n","body":"Hi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n","comment_length":74,"text":"Couldn't reach certain URLs and for the ones that can be reached, code just blocks after downloading.\nHi,\r\n\r\nI'm following the [quick tour](https:\/\/huggingface.co\/nlp\/quicktour.html) and tried to load the glue dataset:\r\n```\r\n>>> from nlp import load_dataset\r\n>>> dataset = load_dataset('glue', 'mrpc', split='train')\r\n```\r\n\r\nHowever, this ran into a `ConnectionError` saying it could not reach the URL (just pasting the last few lines):\r\n```\r\n\r\n\/net\/vaosl01\/opt\/NFS\/su0\/miniconda3\/envs\/hf\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only)\r\n 354 \" to False.\"\r\n 355 )\r\n--> 356 raise ConnectionError(\"Couldn't reach {}\".format(url))\r\n 357 \r\n 358 # From now on, connected is True.\r\n\r\nConnectionError: Couldn't reach https:\/\/firebasestorage.googleapis.com\/v0\/b\/mtl-sentence-representations.appspot.com\/o\/data%2Fmrpc_dev_ids.tsv?alt=media&token=ec5c0836-31d5-48f4-b431-7480817f1adc\r\n```\r\n\r\nI tried glue with cola and sst2. I got the same error, just instead of mrpc in the URL, it was replaced with cola and sst2.\r\n\r\nSince this was not working, I thought I'll try another dataset. So I tried downloading the imdb dataset:\r\n```\r\nds = load_dataset('imdb', split='train')\r\n```\r\nThis downloads the data, but it just blocks after that:\r\n```\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.56k\/4.56k [00:00<00:00, 1.38MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.07k\/2.07k [00:00<00:00, 1.15MB\/s]\r\nDownloading and preparing dataset imdb\/plain_text (download: 80.23 MiB, generated: 127.06 MiB, post-processed: Unknown sizetotal: 207.28 MiB) to \/net\/vaosl01\/opt\/NFS\/su0\/huggingface\/datasets\/imdb\/plain_text\/1.0.0\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 84.1M\/84.1M [00:07<00:00, 11.1MB\/s]\r\n```\r\n\r\nI checked the folder `$HF_HOME\/datasets\/downloads\/extracted\/<id>\/aclImdb`. This folder is constantly growing in size. When I navigated to the train folder within, there was no file. However, the test folder seemed to be populating. The last time I checked it was 327M. I thought the Imdb dataset was smaller than that. My questions are:\r\n1. Why is it still blocking? Is it still downloading?\r\n2. I specified split as train, so why is the test folder being populated?\r\n3. I read somewhere that after downloading, `nlp` converts the text files into some sort of `arrow` files, which will also take a while. Is this also happening here?\r\n\r\nThanks.\r\n\nHi. Closing this one since #626 updated the glue urls.\r\n\r\n> 1. Why is it still blocking? Is it still downloading?\r\n\r\nAfter downloading it generates the arrow file by iterating through the examples.\r\nThe number of examples processed by second is shown during the processing (not sure why it was not the case for you)\r\n\r\n> 2. I specified split as train, so why is the test folder being populated?\r\n\r\nIt downloads every split\r\n\r\n\r\n\r\n","embeddings":[-0.112636596,-0.0148180015,-0.0514469258,0.2499045283,0.2150651067,-0.1080487967,-0.0023050923,0.1171098575,0.144177258,-0.1348839551,-0.425668329,-0.0689154193,0.1447221488,0.1289693564,0.274225086,-0.0121180899,-0.1593028903,-0.0744765624,-0.0916164145,0.017701054,-0.2575627863,0.3606531918,-0.2036087364,0.1764502376,-0.1922911257,-0.0935443118,-0.1606316715,0.1398071498,-0.1229681745,-0.3245210052,0.0947918966,0.3048549592,0.0481997393,0.3857811987,-0.0001219456,-0.0564453825,0.5723124146,-0.0555414632,-0.3910080492,-0.4353305697,-0.2779269516,-0.0589925274,0.355984658,-0.0667517558,0.130753994,0.1634606421,0.0878987014,-0.2419316471,0.2720347047,0.26318115,0.1377917379,0.0733033642,0.1359027624,0.0714743361,0.3174403608,-0.2632397711,0.0046695629,0.3708156645,0.3326888382,-0.291418612,0.0472086743,0.0577220842,-0.205646798,0.0583759286,0.1841480434,0.1635258496,-0.25784266,-0.5462059379,0.0292282049,0.3213647902,0.2748629451,0.0140512539,-0.4840654433,-0.3165412545,-0.0286721792,0.1173614413,0.2789925039,0.1334600002,-0.2538383007,0.0309729725,-0.5450456142,-0.1321876645,0.0033698108,0.5532464981,0.1101105139,-0.0098115858,0.0639427006,0.1689395458,0.3128620982,-0.0395253114,0.1148340628,-0.0232361071,0.2994754314,0.1872622669,-0.0450580493,-0.0265388358,0.0512051396,0.4236578643,0.1454071999,0.1372276694,0.1992809027,0.0405518636,-0.1392972767,0.1321616024,0.15483661,-0.0140559683,-0.0127521874,0.0798756033,0.6293030977,0.3603568673,-0.1680348217,0.0711006969,-0.1136864573,-0.0785518587,-0.3989927769,-0.0077087539,-0.1199624836,-0.284093976,-0.1622226536,-0.0771386921,-0.1368774027,0.0150720552,0.0852001831,0.3928332329,-0.3108938336,0.2066407949,-0.087643005,0.3793320358,-0.1213530898,0.0319333375,-0.0930832773,0.1856428087,-0.1889153719,0.2517310679,0.4859232306,0.0784364417,0.3209574521,-0.3183521628,-0.1498351097,-0.2291524559,0.0177800152,-0.3440533578,0.0556664392,0.0532136112,0.3014571071,0.3389762938,-0.0610685609,-0.1870903969,-0.1006215811,-0.2109535187,-0.2240741253,-0.1662695557,0.3674656749,0.1121519953,-0.4041662812,-0.1559494436,-0.1344938278,-0.004829342,-0.0961971432,-0.0948827863,-0.1139309928,-0.0595751256,-0.1584257782,-0.200951606,0.3062122166,0.6965598464,-0.1322715133,-0.2766490877,-0.1275884658,0.041212976,-0.0782895386,0.502037406,-0.0070996187,0.1006353199,-0.4256356955,0.2437322885,0.5672342777,-0.2894794941,-0.3902561069,0.5320960879,-0.0154582411,0.0364813991,0.1849975735,0.0225609727,0.131248787,-0.0723032951,0.4729860723,0.5410932899,-0.0449328944,-0.0970429629,-0.2779184282,-0.1605929285,0.1289730817,0.1257047504,0.223503828,0.0212896317,-0.0032729721,-0.079706423,0.233262375,0.2292817086,0.0552493818,0.2484004796,-0.1013187617,0.1788349301,-0.104797557,-0.1381653398,-0.3142169118,0.2160550207,-0.0629844964,0.1218500361,-0.3327453732,0.2367781252,-0.1883798987,-0.1930935532,0.0038398847,-0.0187660865,0.0700894296,0.2044757456,0.131302923,0.0753203556,-0.2522730231,0.5734852552,0.0272469651,0.1724773198,-0.316007942,0.6796264052,-0.1542854309,-0.191763401,0.1090584621,0.3169533312,0.2635817826,-0.2635529935,0.0134699494,0.3022545278,-0.1289988309,0.1917850226,0.2702765465,0.0180154573,0.3637427986,-0.2720222771,0.1126302034,0.0491894484,0.1340863556,-0.1252143681,0.1797020435,0.2376711965,-0.2186419368,0.6319778562,0.4755757749,-0.0091407485,0.1447475404,-0.1831095815,-0.2134258151,-0.2173276395,0.6007691026,0.2402251363,0.2317631692,-0.1197275817,-0.148830682,0.1870426685,0.3490749896,0.0156644378,0.05569648,0.2178222686,-0.1637909412,-0.1998849362,-0.1010185257,0.3804479241,0.2209589183,0.1598923951,0.1135150269,0.021280108,-0.2521316409,-0.3306078017,0.2030390054,0.1683373004,-0.0889684558,0.0819643363,0.2019196898,0.1494818777,-0.0947707817,-0.4382834136,0.0053283786,-0.0233714301,-0.3714630306,-0.0647940934,-0.445727706,-0.6604163051,0.0467646122,-0.1720999628,-0.2792548835,-0.646451056,-0.1037925184,0.2034947723,0.0090118954,0.0356713496,-0.2063030154,-0.0062600239,-0.2111667097,-0.0129495598,-0.1801105589,-0.3260889649,-0.062903434,0.0395989567,0.3534008265,0.0574423596,0.1485339999,-0.2441104352,-0.1611087322,-0.1612910032,-0.1101251915,0.19283548,-0.1821724176,0.3608709574,0.2109528184,0.4903079271,0.2029083073,-0.2494547963,0.3121628463,-0.1433433294,0.0774954781,-0.0519493446,0.0077687465,-0.0471082143,0.1043553874,-0.0099998713,-0.471308589,-0.4650456309,0.2359144241,-0.0598598197,0.2176345289,0.1203419119,-0.041052755,0.2929945886,-0.0266634505,0.1280004233,-0.041178517,-0.3354517817,0.5650700331,-0.2066883743,-0.2000689805,-0.0150570692,0.0398590788,-0.2016781569,0.0807290226,-0.706532836,-0.2792837024,-0.4299277067,0.0273614805,0.2272491306,-0.0665315464,-0.0501204729,-0.2217652798,-0.0760565326,-0.0140878279,-0.4060663879,-0.0029456718,0.3619810045,0.3475250006,0.1065041944,0.5122767091,-0.0753890201,0.272541374,-0.0739138052,0.0747887865,0.6189810038,0.0821451545,0.3144127727,-0.1913370788,-0.269040674,0.1089694574,-0.3142771423,0.0521760285,0.1350190192,0.2109739482,-0.3150362074,-0.5289624333,0.172768876,-0.4613894224,-0.1889805794,0.0283963811,-0.0316216201,-0.2145553231,0.2021680921,0.035619285,-0.0524334349,-0.2252569646,-0.0457172729,0.1494725794,0.2521126568,0.039624352,-0.1105956882,-0.1379540712,-0.4256444275,0.2749575973,0.1958890855,0.3104603589,-0.2696570158,-0.0499456488,0.0475213565,-0.0248350445,0.5041225553,-0.4135855436,0.2516334057,0.1580861807,0.0192981642,-0.6474230289,0.0887410268,-0.1247085482,0.1403576285,0.6630892158,0.2826699615,-0.1967948675,0.1196002215,-0.0139652062,0.1940147877,0.1074987426,0.0746590942,-0.0067794248,-0.2686884701,-0.2585899532,-0.2295411676,-0.0385957845,0.0707760081,-0.0640727878,-0.3239626288,0.3134526014,-0.4003784955,-0.0395804793,0.0364644267,0.1761945337,-0.107365407,0.0827875137,0.2307296842,0.1615782082,-0.1797794253,0.4915136993,-0.3460828662,0.0986247435,0.5306717157,-0.0260191523,0.1837768108,0.6057760715,0.1359435916,0.1512895077,0.0835235864,0.1909211874,0.2960536778,0.3395085037,0.2108195573,-0.0150642218,-0.282137394,-0.3250028491,0.3341210485,-0.0021029513,-0.0231160782,0.2038309872,0.0800835639,-0.2553889751,0.2971126735,-0.3381642997,1.148299098,0.0942292139,0.1294936687,-0.1617401689,-0.252526015,0.664984107,-0.1067686155,0.1237175688,-0.046933569,-0.1145089716,-0.1157739386,-0.1703818738,-0.211443305,0.0383760333,-0.2149146348,0.5331195593,0.075566791,-0.0113159101,0.1618145704,0.4351913631,-0.1023060456,-0.0456134863,-0.5257042646,0.0877486318,-0.0442999899,0.2518552244,-0.0140865427,-0.1485861987,-0.2059692591,0.005314081,-0.4601523578,0.1955182999,-0.6179574132,0.2456762493,-0.0609878637,-0.3103728592,-0.1943768859,-0.0575052127,0.1287858635,-0.0164992493,-0.3511244655,0.0874209404,0.2061311007,-0.1411281377,0.1736903042,0.0441385135,0.1641439348,-0.1359499395,-0.0259148777,-0.0478292927,0.0252410229,-0.3272763789,0.0321033821,-0.1285140514,0.0396090299,0.0648923889,-0.1876660287,-0.1389926225,-0.1790039539,-0.2322729975,0.0341174826,-0.0430926159,0.1417649537,-0.085842438,0.3707483113,-0.2654457986,-0.16239582,0.5142537951,-0.305388242,0.0012647071,0.339337945,0.3020931184,-0.3583815396,-0.0925878659,-0.1114680618,-0.1822286099,-0.3933390677,-0.0003344097,-0.2642868161,0.0666865036,-0.0891283676,0.0259221327,0.1505150199,0.1860919148,-0.1181158945,-0.6274639964,-0.1790275872,0.4015606344,-0.0270387996,0.0485015772,-0.2964761555,0.1652342081,0.4207263887,-0.0386223979,-0.2305992246,0.1560420096,-0.44154194,0.0104385708,0.3247368336,-0.3346687257,0.0485885106,-0.0424531437,0.0563483611,0.0833296478,-0.1790658981,-0.0608192533,-0.1101867259,0.1974409521,0.0511796996,-0.3207353354,-0.0964926332,-0.0979493782,0.2066013664,-0.0818443969,-0.0759881213,-0.0132816853,0.0211339984,-0.2951012254,0.1266241819,-0.0999270082,-0.1372247636,0.3212115169,-0.0065158373,0.2825662494,0.3173856437,0.1064764485,0.0883074254,-0.0717145652,-0.0931007713,-0.1241110042,-0.0829002708,0.1949430406,0.4736431241,-0.0916405395,0.0512022041,0.0569887497,0.3979811966,0.5058012009,0.0556126125,-0.1075799018,0.1179754958,0.0908082947,0.0191558711,0.251812458,0.2627199292,0.1216957122,0.0234896373,0.097651422,0.0498040468,-0.0724791065,0.0851857215,0.075487569,0.6589386463,0.201536268,-0.0219140518,-0.1459535807,0.0476535223,-0.0527363233,0.0410787389,-0.0996239334,0.0687503144,-0.126663357,0.0534229763,-0.0639401227,-0.1313260347,0.3249314427,0.2482800186,-0.2003667951,-0.0873381868,0.2849595845,0.0983348265,0.1369969398,-0.220414117,0.3333672583,-0.1611330211,0.0659117475,-0.1380514205,0.4353375435,0.1926468909,-0.3216380179,0.0582479835,-0.3349765539,-0.4865922332,0.0540139452,-0.1999975294,-0.1715221703,0.1560159922,0.2826383114,0.0135858925,-0.7253261805,-0.1404952407,-0.1912470907,0.232984975,-0.3082934618,0.4661635458,-0.1038244143,0.0777674317,-0.0195026733,0.1526183188,0.3781436086,0.3050530851,-0.4279482365,0.0964921415,0.2019785941,-0.0506129041,-0.0096543431,0.230298087,0.4857646525,-0.1341400892,0.492813617,0.0891050696,-0.1842339635,0.2686845064,-0.0354273506,0.1412737072,-0.0751045942,0.2763312161,-0.0875300467,-0.0513294786,-0.2288811356,-0.1396883428,-0.5556668043,-0.109657459,-0.0335210189,-0.5973433256,0.0323545001,-0.0905735418,0.04113717,0.0053061331,0.6298923492,0.569093585,0.5683688521,-0.380954206,-0.3364817202,-0.3252815902,0.0403116867,-0.1251098812,-0.0124700721,-0.1701546311,0.2908905745,-0.174534604,0.2047268748,-0.0137992641,0.6748144627,-0.1523853242,0.1743044406,-0.2713821232,-0.0484785065,-0.0313812792,0.0573393144,-0.1030781344,-0.0273058582,0.0516437143,-0.5177112818,-0.0516986661,0.2759017646,-0.2825129032,0.133970499,0.236651957,0.3119557798,0.1032110676,0.7628632784,-0.122793287,-0.1470048726,-0.4071486294,-0.6882551908,-0.0240677781,-0.1192925423,0.0928619131,0.1546040028,-0.0933201984,0.2021304667,-0.1746087968,0.190074876,-0.0879273117,0.4516707063,0.1424894631,-0.3678409457,0.065007776,-0.0495072305,0.0439469479,-0.0063381037,-0.2057438493,0.1007037237,-0.1159259304,-0.3511934578,0.5208379626,-0.0540384129,-0.0175133403,-0.1145703346,0.2055766732,0.0688287094,0.2129953802,-0.3332384527,-0.2800841331,0.344409585,0.058339078,-0.4353078008,0.2324451804,-0.0068449518,-0.0778456554,-0.0766064376,-0.0418283306,0.2216898501,-0.218432188,-0.192988202,-0.0678091273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/568","title":"`metric.compute` throws `ArrowInvalid` error","comments":"Hmm might be related to what we are solving in #564","body":"I get the following error with `rouge.compute`. It happens only with distributed training, and it occurs randomly I can't easily reproduce it. This is using `nlp==0.4.0`\r\n\r\n```\r\n File \"\/home\/beltagy\/trainer.py\", line 92, in validation_step\r\n rouge_scores = rouge.compute(predictions=generated_str, references=gold_str, rouge_types=['rouge2', 'rouge1', 'rougeL'])\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 224, in compute\r\n self.finalize(timeout=timeout)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 213, in finalize\r\n self.data = Dataset(**reader.read_files(node_files))\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 217, in read_files\r\n dataset_kwargs = self._read_files(files=files, info=self._info, original_instructions=original_instructions)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 162, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 276, in _get_dataset_from_filename\r\n f = pa.ipc.open_stream(mmap)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 173, in open_stream\r\n return RecordBatchStreamReader(source)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 64, in __init__\r\n self._open(source)\r\n File \"pyarrow\/ipc.pxi\", line 469, in pyarrow.lib._RecordBatchStreamReader._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Tried reading schema message, was null or length 0\r\n```","comment_length":11,"text":"`metric.compute` throws `ArrowInvalid` error\nI get the following error with `rouge.compute`. It happens only with distributed training, and it occurs randomly I can't easily reproduce it. This is using `nlp==0.4.0`\r\n\r\n```\r\n File \"\/home\/beltagy\/trainer.py\", line 92, in validation_step\r\n rouge_scores = rouge.compute(predictions=generated_str, references=gold_str, rouge_types=['rouge2', 'rouge1', 'rougeL'])\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 224, in compute\r\n self.finalize(timeout=timeout)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 213, in finalize\r\n self.data = Dataset(**reader.read_files(node_files))\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 217, in read_files\r\n dataset_kwargs = self._read_files(files=files, info=self._info, original_instructions=original_instructions)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 162, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 276, in _get_dataset_from_filename\r\n f = pa.ipc.open_stream(mmap)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 173, in open_stream\r\n return RecordBatchStreamReader(source)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 64, in __init__\r\n self._open(source)\r\n File \"pyarrow\/ipc.pxi\", line 469, in pyarrow.lib._RecordBatchStreamReader._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Tried reading schema message, was null or length 0\r\n```\nHmm might be related to what we are solving in #564","embeddings":[-0.4075993598,-0.2417839468,0.0564272702,0.2862923145,0.3149873018,-0.1689801663,-0.159183085,0.2923860252,-0.1545246542,0.4045127332,0.0525604412,0.5000802875,-0.1109343246,-0.3641472757,-0.094238542,-0.1911460012,-0.1322497874,0.0535943359,0.0623223409,-0.2779391706,-0.3765873313,0.0985771418,-0.2131596059,0.3267936409,-0.1932066679,-0.3521064818,0.1751866639,0.1816710234,-0.3589941561,-0.5263907909,0.2709598839,-0.2937383056,0.0037990306,0.1183886901,-0.0001191873,0.1296570301,0.3363063931,-0.0975792706,-0.0041522533,-0.1153502688,-0.0072667021,-0.0410663113,0.3097956181,-0.1023338959,0.1306812316,-0.1691236049,-0.2776666284,-0.2377279848,0.2970281839,0.3777075708,0.0835815147,0.2802878916,-0.0036599415,-0.2270237654,-0.1957388967,-0.2623844147,-0.0339903161,0.7621635795,0.0858863145,-0.212425679,0.0804970711,0.2712044716,0.3043489158,0.1928760856,0.434317559,-0.2398328185,0.2804588079,-0.002974591,-0.1896120757,0.1256083846,-0.0713665262,-0.1000525355,-0.2646160424,0.2437917441,0.2630293965,-0.5339521766,-0.0313146561,0.2237594873,-0.0697997287,0.0458759703,-0.240044713,0.0889871642,-0.1897262484,0.0984046757,0.163955614,0.130643785,0.015154738,0.2536816597,0.0861735046,0.0687298328,-0.3007571697,0.3731757998,-0.2901461422,-0.0234646499,-0.4769316614,-0.1539025456,-0.0954827592,0.0382494144,0.0169564392,0.1846893579,0.6731569171,-0.0968763605,0.2253452539,0.3182927072,0.1315817535,0.2608215213,-0.2110044658,0.5256132483,0.0850065351,-0.0524344891,-0.0642045215,0.0298821982,0.2078827918,-0.6180186272,0.1823914796,0.3520441949,0.0247808602,0.1011312529,-0.5585077405,-0.187474221,-0.4208103716,0.1441128701,0.1162040085,0.070241943,-0.0098986542,0.1733299792,0.2674604952,0.2471818775,-0.1467236131,-0.3565670252,-0.2359515429,0.1969593465,-0.2147193253,0.049003385,-0.0045695757,-0.0573178567,0.1841375381,0.090225853,0.0255612973,-0.1761726886,0.5448206067,-0.2204472274,-0.0769289583,0.0257570073,-0.3781884015,-0.2873996794,0.229777202,0.1334006041,0.037450254,-0.1065518111,-0.2053185105,-0.3397361338,0.2455221862,0.1057567075,0.3015190959,-0.0896550715,0.2021171451,0.1027184129,0.1186710373,-0.1679324359,0.1719890237,-0.2009641826,-0.1101601794,0.0065554334,0.0725460723,-0.0001086169,-0.1014669389,0.1107227728,0.1415624768,0.1545994133,0.0349537246,0.4689890444,-0.2500044107,0.2202896327,0.0234921817,-0.3029279113,0.5286731124,-0.9485337734,-0.2145374864,0.0507860072,-0.1609332114,-0.5299543142,-0.1245944723,-0.1013837904,-0.1401650608,0.1376924664,0.3832190335,0.0421500206,-0.1681721956,-0.2119553238,-0.2906453907,0.0716145709,0.246895358,0.1549103707,0.0768484026,0.0673309341,0.0960901529,-0.2710840702,0.0719325691,0.031280566,-0.0309009049,0.2132062167,-0.1415386796,-0.143858254,0.4165867269,-0.0793790221,0.2842078209,0.1112132147,-0.3692346513,-0.1870997995,-0.2376635224,0.0109891947,-0.3217407465,0.0674759075,-0.0672575831,-0.0084727677,0.0498084016,0.0204183981,0.2208724916,0.1332473457,-0.3149588406,-0.221074149,-0.4542842507,-0.030646855,-0.1260975599,0.0629270151,0.0274376441,-0.2461945862,0.0837185308,0.1811012775,0.160761103,-0.0457629487,-0.2479192317,0.4671181738,0.0894628614,0.0979057625,0.4763180912,-0.1720580012,0.1261303127,-0.2955779731,-0.3338622749,0.2377325296,0.3096678555,0.0850776434,0.1167073324,0.2165964097,-0.3274505138,-0.0149451001,-0.0602276884,0.2954143882,0.219982028,0.1660233587,-0.1664334536,0.0484345555,0.1911974251,-0.1303867102,0.39263165,-0.2581726313,-0.3667012453,-0.2043807954,0.1814842671,0.1247731745,0.0273581091,-0.2630428374,0.2203611284,0.2895263433,0.057939209,0.1685934812,0.4083440602,-0.08389882,0.0149456374,0.2547494471,-0.0987793058,-0.1508776844,0.0551872589,0.0896368101,0.1410100907,0.3735880852,0.2065856308,0.0663291514,-0.3672625124,-0.1021530479,0.2359729409,0.2093068361,-0.0221330989,0.4142645001,-0.0017026064,0.277654469,-0.1761566401,-0.1490000039,-0.1172153726,-0.1813320518,0.0745567977,0.0275700875,-0.0813113302,0.2263077348,0.2115996629,0.34382689,0.3297126591,0.0593655221,0.069641985,-0.1464833021,-0.1094480455,-0.12097238,-0.1135568023,0.0799365714,0.1879225671,-0.079188183,0.2124749422,-0.1500804126,-0.0069223507,-0.0119671673,0.0123387277,-0.0033971411,0.3124577403,-0.419359833,-0.6186656356,0.1612169147,0.3857408762,-0.1651768237,-0.1729445606,0.1802288145,0.0099464897,-0.0189911984,-0.1486451179,-0.3686364293,-0.408726573,-0.3623014987,0.3295700252,0.0106309811,-0.1517916471,0.1586492658,0.0680436641,0.1036152095,0.2316931784,-0.0726489872,0.020813426,-0.0302379169,0.2664110959,-0.2160875201,-0.2734997869,-0.0737783611,-0.1136969328,0.3361850679,0.4906616807,-0.1465961337,-0.2178466171,0.1189218536,0.1014851928,-0.5226741433,0.1076532006,0.5523734093,-0.1283311099,-0.0396117494,-0.1679284573,0.3205580413,0.4238615036,-0.0645174012,0.0601091981,-0.1110722274,0.1943630576,0.105065316,0.750133872,0.2554729879,-0.0024984078,0.1955074072,0.04087184,0.1205502972,0.1490536183,-0.0575668849,0.2131001651,0.1321832836,-0.1164219603,-0.0873289034,-0.1998218447,0.0960988998,-0.1304616034,-0.0857842565,0.1583218127,0.081030108,-0.117395632,-0.7013334036,0.5111743212,0.2526737154,0.175617069,-0.1056380644,0.0872496217,0.2066690624,-0.1311411709,0.4093682468,0.0597588904,0.0744302198,-0.3496944606,-0.0763837025,0.3868108988,0.1509348899,0.5822174549,-0.0802684203,0.2114307731,0.1476576775,0.0032646088,0.176494956,-0.4874759316,-0.0024385985,0.0024493767,-0.3182797432,-0.7047178745,0.0278705899,-0.2246014029,0.2069823146,-0.0390860923,0.2247414142,-0.0905078501,-0.0705309287,0.0914265737,-0.0246223025,-0.2360409647,-0.04375346,-0.43846187,-0.1007978246,-0.0079877311,0.3548945487,0.215847224,0.0242722016,0.2047313899,0.0260860771,-0.4139418304,0.2024679184,-0.1164642423,-0.1058029979,0.2377334386,-0.2303404808,-0.1086247787,0.1344448179,0.0973659903,0.1317891777,0.3259266615,0.1949752271,-0.7740002871,-0.0881980807,-0.0107653225,0.0912959501,0.1395350248,-0.1086555645,-0.0830888301,-0.1653802395,-0.0371893421,-0.1362710595,-0.0624066703,0.2667120993,-0.0437857248,0.2142005563,-0.0918841064,0.3918713927,0.0683675706,-0.0616462566,0.6337679029,-0.088559255,-0.1122555882,0.0026838945,0.4313238561,0.6851304173,0.3991342783,-0.2286525965,0.5048461556,-0.2030970156,0.2501061559,-0.2921949327,0.3559749126,-0.3496271372,-0.291167438,0.0264542475,-0.1216660365,0.0053279358,0.1125444621,0.0203569159,0.2939416468,-0.0912707224,-0.2852716744,-0.0292429142,0.0008494384,-0.2626026273,-0.0537064783,0.0257876385,-0.0062576835,-0.1543589681,-0.02531958,-0.1340193897,-0.2451969981,-0.1659554243,-0.101049982,0.1498215646,-0.0198502187,-0.2273249477,0.4922190309,0.372472465,-0.126603201,-0.1150018275,0.2851442099,0.4384118021,-0.0976752639,-0.090534538,0.3126313388,-0.1261661053,-0.2087913156,-0.1333062649,0.2157023996,0.0356794074,-0.1506471932,-0.2951408029,0.0754740834,-0.301020205,-0.2734054923,0.0300637912,0.0591305681,-0.2400632203,0.2178396136,0.1667966396,-0.1111845225,-0.3868939877,-0.0963925719,0.0177952964,0.1052126661,-0.3568505049,0.0832281485,-0.0021250511,-0.3438087702,0.3042221367,0.3032570481,-0.1221484467,0.0216180794,0.3507742584,-0.2816231251,0.0514557138,-0.2250773758,-0.0476729162,0.1816169769,-0.3708324134,0.3871153891,-0.1974066198,0.2173268646,0.1861213297,0.5506527424,0.1998642981,0.0305276476,0.0216813646,-0.3809008002,-0.281195879,0.0749440119,-0.10570243,-0.1554643065,-0.0670688301,0.0661698952,-0.063046284,0.1681356132,-0.1711985916,0.0182608329,-0.4026435018,0.1385002881,-0.2358004749,-0.1427798569,-0.2918173969,0.1429328918,-0.105355382,0.47483325,-0.1834557205,-0.0251481198,0.0208589584,0.1604123265,0.1491889954,0.2476718426,0.0554537214,-0.0948632434,-0.4114228487,-0.0237532984,-0.1994509399,-0.1094015688,-0.1021039709,0.1169447973,-0.148319602,0.2587156296,-0.0825612769,-0.102657862,0.2263108045,0.0078636818,-0.4149520099,-0.0237018876,0.1374997199,-0.1295656264,-0.4047162831,-0.071770817,-0.1746266186,0.032731384,0.235680297,-0.3694477081,-0.0406388976,-0.3534323275,0.3480403125,0.4453604221,-0.4016061723,-0.0821943805,0.0196640231,0.0796486735,-0.2390367687,0.0348827839,-0.0927760378,0.4075059891,-0.0362933762,0.0683815405,0.2137533724,-0.3845719993,-0.0951184109,0.1793109775,-0.1105760559,-0.5657693744,0.183415845,-0.0904031694,-0.0515371934,0.0342611969,0.4253365993,0.3455851972,0.2773367465,0.378754884,-0.1718459576,0.3203332722,0.2994095683,-0.0327489302,-0.0350370258,-0.0182347279,-0.324139595,-0.0569081418,0.1216445416,-0.2019240558,0.1269879639,0.4522322714,-0.5222976804,-0.5407586694,-0.1329667121,-0.1093748659,-0.0489517786,0.2321892083,-0.5116025209,-0.1040367335,-0.0905237794,-0.0982390195,0.2170716226,0.5896323919,0.5254977345,0.2307590097,0.0998491794,-0.4661764801,-0.0672634095,0.5261714458,0.3482179344,-0.3474153578,0.0702713281,0.2197810411,-0.0478885323,-0.1836316586,0.468246758,0.1106709316,0.4221267104,-0.3368279338,-0.0574068166,-0.0913393646,0.055816371,0.188318342,0.4640157223,0.4181191623,-0.1294232756,0.4254341722,-0.0807892233,-0.0253640153,0.3410935402,-0.1127055511,0.5322974324,-0.1863979995,0.2962688506,-0.1552479714,0.1272639781,-0.1232938096,0.0020000769,-0.5153283477,0.3860246539,-0.291657418,0.0725813955,0.0253607891,-0.271500349,0.0217014272,-0.2019307315,0.4118304849,0.254483372,-0.0213224012,-0.1645658463,-0.364146024,-0.3480188847,0.1979256719,-0.0347262323,-0.0323350802,0.0377677195,0.2458748519,-0.0899563655,0.2194443345,0.0136135528,0.3184034824,0.1633765846,0.4191622138,-0.4301507175,-0.081525147,-0.3555106819,-0.1731373817,0.2258825153,-0.1960830986,0.0569185689,-0.0774393007,-0.0041136811,0.2366613001,-0.4406211674,0.0182626452,0.4891805351,0.4408990741,0.1522389948,0.2984647155,0.0991083309,0.0148841208,-0.1326742172,-0.360555619,0.0081420112,0.3292993903,0.0611585565,0.3952183425,-0.1376777887,0.1447617561,-0.2830953002,0.372395575,-0.1469091177,-0.0667040795,-0.0228478238,0.0460149981,-0.0089644669,0.1270906031,-0.0131379813,0.1258742809,0.1562907994,0.0732177496,-0.2892949283,-0.2071141303,0.5095154047,-0.152072221,-0.4998908937,0.1289652586,0.3653587103,0.8499709368,-0.111030139,-0.4012586176,0.0702343807,0.1562007815,-0.045343522,-0.0882446542,-0.0701677501,0.0586670339,0.1032955199,0.1094179526,0.0976541489,-0.05990557,0.0193283074,0.2159411013,-0.1393848062]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/568","title":"`metric.compute` throws `ArrowInvalid` error","comments":"Could you try to update to `datasets>=1.0.0` (we changed the name of the library) and try again ?\r\nIf is was related to the distributed setup settings it must be fixed.\r\nIf it was related to empty metric inputs it's going to be fixed in #654 ","body":"I get the following error with `rouge.compute`. It happens only with distributed training, and it occurs randomly I can't easily reproduce it. This is using `nlp==0.4.0`\r\n\r\n```\r\n File \"\/home\/beltagy\/trainer.py\", line 92, in validation_step\r\n rouge_scores = rouge.compute(predictions=generated_str, references=gold_str, rouge_types=['rouge2', 'rouge1', 'rougeL'])\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 224, in compute\r\n self.finalize(timeout=timeout)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 213, in finalize\r\n self.data = Dataset(**reader.read_files(node_files))\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 217, in read_files\r\n dataset_kwargs = self._read_files(files=files, info=self._info, original_instructions=original_instructions)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 162, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 276, in _get_dataset_from_filename\r\n f = pa.ipc.open_stream(mmap)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 173, in open_stream\r\n return RecordBatchStreamReader(source)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 64, in __init__\r\n self._open(source)\r\n File \"pyarrow\/ipc.pxi\", line 469, in pyarrow.lib._RecordBatchStreamReader._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Tried reading schema message, was null or length 0\r\n```","comment_length":46,"text":"`metric.compute` throws `ArrowInvalid` error\nI get the following error with `rouge.compute`. It happens only with distributed training, and it occurs randomly I can't easily reproduce it. This is using `nlp==0.4.0`\r\n\r\n```\r\n File \"\/home\/beltagy\/trainer.py\", line 92, in validation_step\r\n rouge_scores = rouge.compute(predictions=generated_str, references=gold_str, rouge_types=['rouge2', 'rouge1', 'rougeL'])\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 224, in compute\r\n self.finalize(timeout=timeout)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 213, in finalize\r\n self.data = Dataset(**reader.read_files(node_files))\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 217, in read_files\r\n dataset_kwargs = self._read_files(files=files, info=self._info, original_instructions=original_instructions)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 162, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 276, in _get_dataset_from_filename\r\n f = pa.ipc.open_stream(mmap)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 173, in open_stream\r\n return RecordBatchStreamReader(source)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 64, in __init__\r\n self._open(source)\r\n File \"pyarrow\/ipc.pxi\", line 469, in pyarrow.lib._RecordBatchStreamReader._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Tried reading schema message, was null or length 0\r\n```\nCould you try to update to `datasets>=1.0.0` (we changed the name of the library) and try again ?\r\nIf is was related to the distributed setup settings it must be fixed.\r\nIf it was related to empty metric inputs it's going to be fixed in #654 ","embeddings":[-0.4075993598,-0.2417839468,0.0564272702,0.2862923145,0.3149873018,-0.1689801663,-0.159183085,0.2923860252,-0.1545246542,0.4045127332,0.0525604412,0.5000802875,-0.1109343246,-0.3641472757,-0.094238542,-0.1911460012,-0.1322497874,0.0535943359,0.0623223409,-0.2779391706,-0.3765873313,0.0985771418,-0.2131596059,0.3267936409,-0.1932066679,-0.3521064818,0.1751866639,0.1816710234,-0.3589941561,-0.5263907909,0.2709598839,-0.2937383056,0.0037990306,0.1183886901,-0.0001191873,0.1296570301,0.3363063931,-0.0975792706,-0.0041522533,-0.1153502688,-0.0072667021,-0.0410663113,0.3097956181,-0.1023338959,0.1306812316,-0.1691236049,-0.2776666284,-0.2377279848,0.2970281839,0.3777075708,0.0835815147,0.2802878916,-0.0036599415,-0.2270237654,-0.1957388967,-0.2623844147,-0.0339903161,0.7621635795,0.0858863145,-0.212425679,0.0804970711,0.2712044716,0.3043489158,0.1928760856,0.434317559,-0.2398328185,0.2804588079,-0.002974591,-0.1896120757,0.1256083846,-0.0713665262,-0.1000525355,-0.2646160424,0.2437917441,0.2630293965,-0.5339521766,-0.0313146561,0.2237594873,-0.0697997287,0.0458759703,-0.240044713,0.0889871642,-0.1897262484,0.0984046757,0.163955614,0.130643785,0.015154738,0.2536816597,0.0861735046,0.0687298328,-0.3007571697,0.3731757998,-0.2901461422,-0.0234646499,-0.4769316614,-0.1539025456,-0.0954827592,0.0382494144,0.0169564392,0.1846893579,0.6731569171,-0.0968763605,0.2253452539,0.3182927072,0.1315817535,0.2608215213,-0.2110044658,0.5256132483,0.0850065351,-0.0524344891,-0.0642045215,0.0298821982,0.2078827918,-0.6180186272,0.1823914796,0.3520441949,0.0247808602,0.1011312529,-0.5585077405,-0.187474221,-0.4208103716,0.1441128701,0.1162040085,0.070241943,-0.0098986542,0.1733299792,0.2674604952,0.2471818775,-0.1467236131,-0.3565670252,-0.2359515429,0.1969593465,-0.2147193253,0.049003385,-0.0045695757,-0.0573178567,0.1841375381,0.090225853,0.0255612973,-0.1761726886,0.5448206067,-0.2204472274,-0.0769289583,0.0257570073,-0.3781884015,-0.2873996794,0.229777202,0.1334006041,0.037450254,-0.1065518111,-0.2053185105,-0.3397361338,0.2455221862,0.1057567075,0.3015190959,-0.0896550715,0.2021171451,0.1027184129,0.1186710373,-0.1679324359,0.1719890237,-0.2009641826,-0.1101601794,0.0065554334,0.0725460723,-0.0001086169,-0.1014669389,0.1107227728,0.1415624768,0.1545994133,0.0349537246,0.4689890444,-0.2500044107,0.2202896327,0.0234921817,-0.3029279113,0.5286731124,-0.9485337734,-0.2145374864,0.0507860072,-0.1609332114,-0.5299543142,-0.1245944723,-0.1013837904,-0.1401650608,0.1376924664,0.3832190335,0.0421500206,-0.1681721956,-0.2119553238,-0.2906453907,0.0716145709,0.246895358,0.1549103707,0.0768484026,0.0673309341,0.0960901529,-0.2710840702,0.0719325691,0.031280566,-0.0309009049,0.2132062167,-0.1415386796,-0.143858254,0.4165867269,-0.0793790221,0.2842078209,0.1112132147,-0.3692346513,-0.1870997995,-0.2376635224,0.0109891947,-0.3217407465,0.0674759075,-0.0672575831,-0.0084727677,0.0498084016,0.0204183981,0.2208724916,0.1332473457,-0.3149588406,-0.221074149,-0.4542842507,-0.030646855,-0.1260975599,0.0629270151,0.0274376441,-0.2461945862,0.0837185308,0.1811012775,0.160761103,-0.0457629487,-0.2479192317,0.4671181738,0.0894628614,0.0979057625,0.4763180912,-0.1720580012,0.1261303127,-0.2955779731,-0.3338622749,0.2377325296,0.3096678555,0.0850776434,0.1167073324,0.2165964097,-0.3274505138,-0.0149451001,-0.0602276884,0.2954143882,0.219982028,0.1660233587,-0.1664334536,0.0484345555,0.1911974251,-0.1303867102,0.39263165,-0.2581726313,-0.3667012453,-0.2043807954,0.1814842671,0.1247731745,0.0273581091,-0.2630428374,0.2203611284,0.2895263433,0.057939209,0.1685934812,0.4083440602,-0.08389882,0.0149456374,0.2547494471,-0.0987793058,-0.1508776844,0.0551872589,0.0896368101,0.1410100907,0.3735880852,0.2065856308,0.0663291514,-0.3672625124,-0.1021530479,0.2359729409,0.2093068361,-0.0221330989,0.4142645001,-0.0017026064,0.277654469,-0.1761566401,-0.1490000039,-0.1172153726,-0.1813320518,0.0745567977,0.0275700875,-0.0813113302,0.2263077348,0.2115996629,0.34382689,0.3297126591,0.0593655221,0.069641985,-0.1464833021,-0.1094480455,-0.12097238,-0.1135568023,0.0799365714,0.1879225671,-0.079188183,0.2124749422,-0.1500804126,-0.0069223507,-0.0119671673,0.0123387277,-0.0033971411,0.3124577403,-0.419359833,-0.6186656356,0.1612169147,0.3857408762,-0.1651768237,-0.1729445606,0.1802288145,0.0099464897,-0.0189911984,-0.1486451179,-0.3686364293,-0.408726573,-0.3623014987,0.3295700252,0.0106309811,-0.1517916471,0.1586492658,0.0680436641,0.1036152095,0.2316931784,-0.0726489872,0.020813426,-0.0302379169,0.2664110959,-0.2160875201,-0.2734997869,-0.0737783611,-0.1136969328,0.3361850679,0.4906616807,-0.1465961337,-0.2178466171,0.1189218536,0.1014851928,-0.5226741433,0.1076532006,0.5523734093,-0.1283311099,-0.0396117494,-0.1679284573,0.3205580413,0.4238615036,-0.0645174012,0.0601091981,-0.1110722274,0.1943630576,0.105065316,0.750133872,0.2554729879,-0.0024984078,0.1955074072,0.04087184,0.1205502972,0.1490536183,-0.0575668849,0.2131001651,0.1321832836,-0.1164219603,-0.0873289034,-0.1998218447,0.0960988998,-0.1304616034,-0.0857842565,0.1583218127,0.081030108,-0.117395632,-0.7013334036,0.5111743212,0.2526737154,0.175617069,-0.1056380644,0.0872496217,0.2066690624,-0.1311411709,0.4093682468,0.0597588904,0.0744302198,-0.3496944606,-0.0763837025,0.3868108988,0.1509348899,0.5822174549,-0.0802684203,0.2114307731,0.1476576775,0.0032646088,0.176494956,-0.4874759316,-0.0024385985,0.0024493767,-0.3182797432,-0.7047178745,0.0278705899,-0.2246014029,0.2069823146,-0.0390860923,0.2247414142,-0.0905078501,-0.0705309287,0.0914265737,-0.0246223025,-0.2360409647,-0.04375346,-0.43846187,-0.1007978246,-0.0079877311,0.3548945487,0.215847224,0.0242722016,0.2047313899,0.0260860771,-0.4139418304,0.2024679184,-0.1164642423,-0.1058029979,0.2377334386,-0.2303404808,-0.1086247787,0.1344448179,0.0973659903,0.1317891777,0.3259266615,0.1949752271,-0.7740002871,-0.0881980807,-0.0107653225,0.0912959501,0.1395350248,-0.1086555645,-0.0830888301,-0.1653802395,-0.0371893421,-0.1362710595,-0.0624066703,0.2667120993,-0.0437857248,0.2142005563,-0.0918841064,0.3918713927,0.0683675706,-0.0616462566,0.6337679029,-0.088559255,-0.1122555882,0.0026838945,0.4313238561,0.6851304173,0.3991342783,-0.2286525965,0.5048461556,-0.2030970156,0.2501061559,-0.2921949327,0.3559749126,-0.3496271372,-0.291167438,0.0264542475,-0.1216660365,0.0053279358,0.1125444621,0.0203569159,0.2939416468,-0.0912707224,-0.2852716744,-0.0292429142,0.0008494384,-0.2626026273,-0.0537064783,0.0257876385,-0.0062576835,-0.1543589681,-0.02531958,-0.1340193897,-0.2451969981,-0.1659554243,-0.101049982,0.1498215646,-0.0198502187,-0.2273249477,0.4922190309,0.372472465,-0.126603201,-0.1150018275,0.2851442099,0.4384118021,-0.0976752639,-0.090534538,0.3126313388,-0.1261661053,-0.2087913156,-0.1333062649,0.2157023996,0.0356794074,-0.1506471932,-0.2951408029,0.0754740834,-0.301020205,-0.2734054923,0.0300637912,0.0591305681,-0.2400632203,0.2178396136,0.1667966396,-0.1111845225,-0.3868939877,-0.0963925719,0.0177952964,0.1052126661,-0.3568505049,0.0832281485,-0.0021250511,-0.3438087702,0.3042221367,0.3032570481,-0.1221484467,0.0216180794,0.3507742584,-0.2816231251,0.0514557138,-0.2250773758,-0.0476729162,0.1816169769,-0.3708324134,0.3871153891,-0.1974066198,0.2173268646,0.1861213297,0.5506527424,0.1998642981,0.0305276476,0.0216813646,-0.3809008002,-0.281195879,0.0749440119,-0.10570243,-0.1554643065,-0.0670688301,0.0661698952,-0.063046284,0.1681356132,-0.1711985916,0.0182608329,-0.4026435018,0.1385002881,-0.2358004749,-0.1427798569,-0.2918173969,0.1429328918,-0.105355382,0.47483325,-0.1834557205,-0.0251481198,0.0208589584,0.1604123265,0.1491889954,0.2476718426,0.0554537214,-0.0948632434,-0.4114228487,-0.0237532984,-0.1994509399,-0.1094015688,-0.1021039709,0.1169447973,-0.148319602,0.2587156296,-0.0825612769,-0.102657862,0.2263108045,0.0078636818,-0.4149520099,-0.0237018876,0.1374997199,-0.1295656264,-0.4047162831,-0.071770817,-0.1746266186,0.032731384,0.235680297,-0.3694477081,-0.0406388976,-0.3534323275,0.3480403125,0.4453604221,-0.4016061723,-0.0821943805,0.0196640231,0.0796486735,-0.2390367687,0.0348827839,-0.0927760378,0.4075059891,-0.0362933762,0.0683815405,0.2137533724,-0.3845719993,-0.0951184109,0.1793109775,-0.1105760559,-0.5657693744,0.183415845,-0.0904031694,-0.0515371934,0.0342611969,0.4253365993,0.3455851972,0.2773367465,0.378754884,-0.1718459576,0.3203332722,0.2994095683,-0.0327489302,-0.0350370258,-0.0182347279,-0.324139595,-0.0569081418,0.1216445416,-0.2019240558,0.1269879639,0.4522322714,-0.5222976804,-0.5407586694,-0.1329667121,-0.1093748659,-0.0489517786,0.2321892083,-0.5116025209,-0.1040367335,-0.0905237794,-0.0982390195,0.2170716226,0.5896323919,0.5254977345,0.2307590097,0.0998491794,-0.4661764801,-0.0672634095,0.5261714458,0.3482179344,-0.3474153578,0.0702713281,0.2197810411,-0.0478885323,-0.1836316586,0.468246758,0.1106709316,0.4221267104,-0.3368279338,-0.0574068166,-0.0913393646,0.055816371,0.188318342,0.4640157223,0.4181191623,-0.1294232756,0.4254341722,-0.0807892233,-0.0253640153,0.3410935402,-0.1127055511,0.5322974324,-0.1863979995,0.2962688506,-0.1552479714,0.1272639781,-0.1232938096,0.0020000769,-0.5153283477,0.3860246539,-0.291657418,0.0725813955,0.0253607891,-0.271500349,0.0217014272,-0.2019307315,0.4118304849,0.254483372,-0.0213224012,-0.1645658463,-0.364146024,-0.3480188847,0.1979256719,-0.0347262323,-0.0323350802,0.0377677195,0.2458748519,-0.0899563655,0.2194443345,0.0136135528,0.3184034824,0.1633765846,0.4191622138,-0.4301507175,-0.081525147,-0.3555106819,-0.1731373817,0.2258825153,-0.1960830986,0.0569185689,-0.0774393007,-0.0041136811,0.2366613001,-0.4406211674,0.0182626452,0.4891805351,0.4408990741,0.1522389948,0.2984647155,0.0991083309,0.0148841208,-0.1326742172,-0.360555619,0.0081420112,0.3292993903,0.0611585565,0.3952183425,-0.1376777887,0.1447617561,-0.2830953002,0.372395575,-0.1469091177,-0.0667040795,-0.0228478238,0.0460149981,-0.0089644669,0.1270906031,-0.0131379813,0.1258742809,0.1562907994,0.0732177496,-0.2892949283,-0.2071141303,0.5095154047,-0.152072221,-0.4998908937,0.1289652586,0.3653587103,0.8499709368,-0.111030139,-0.4012586176,0.0702343807,0.1562007815,-0.045343522,-0.0882446542,-0.0701677501,0.0586670339,0.1032955199,0.1094179526,0.0976541489,-0.05990557,0.0193283074,0.2159411013,-0.1393848062]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/568","title":"`metric.compute` throws `ArrowInvalid` error","comments":"Closing this one as it was fixed in #654 \r\nFeel free to re-open if you have other questions","body":"I get the following error with `rouge.compute`. It happens only with distributed training, and it occurs randomly I can't easily reproduce it. This is using `nlp==0.4.0`\r\n\r\n```\r\n File \"\/home\/beltagy\/trainer.py\", line 92, in validation_step\r\n rouge_scores = rouge.compute(predictions=generated_str, references=gold_str, rouge_types=['rouge2', 'rouge1', 'rougeL'])\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 224, in compute\r\n self.finalize(timeout=timeout)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 213, in finalize\r\n self.data = Dataset(**reader.read_files(node_files))\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 217, in read_files\r\n dataset_kwargs = self._read_files(files=files, info=self._info, original_instructions=original_instructions)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 162, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 276, in _get_dataset_from_filename\r\n f = pa.ipc.open_stream(mmap)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 173, in open_stream\r\n return RecordBatchStreamReader(source)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 64, in __init__\r\n self._open(source)\r\n File \"pyarrow\/ipc.pxi\", line 469, in pyarrow.lib._RecordBatchStreamReader._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Tried reading schema message, was null or length 0\r\n```","comment_length":18,"text":"`metric.compute` throws `ArrowInvalid` error\nI get the following error with `rouge.compute`. It happens only with distributed training, and it occurs randomly I can't easily reproduce it. This is using `nlp==0.4.0`\r\n\r\n```\r\n File \"\/home\/beltagy\/trainer.py\", line 92, in validation_step\r\n rouge_scores = rouge.compute(predictions=generated_str, references=gold_str, rouge_types=['rouge2', 'rouge1', 'rougeL'])\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 224, in compute\r\n self.finalize(timeout=timeout)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 213, in finalize\r\n self.data = Dataset(**reader.read_files(node_files))\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 217, in read_files\r\n dataset_kwargs = self._read_files(files=files, info=self._info, original_instructions=original_instructions)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 162, in _read_files\r\n pa_table: pa.Table = self._get_dataset_from_filename(f_dict)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py\", line 276, in _get_dataset_from_filename\r\n f = pa.ipc.open_stream(mmap)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 173, in open_stream\r\n return RecordBatchStreamReader(source)\r\n File \"\/home\/beltagy\/miniconda3\/envs\/allennlp\/lib\/python3.7\/site-packages\/pyarrow\/ipc.py\", line 64, in __init__\r\n self._open(source)\r\n File \"pyarrow\/ipc.pxi\", line 469, in pyarrow.lib._RecordBatchStreamReader._open\r\n File \"pyarrow\/error.pxi\", line 122, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow\/error.pxi\", line 84, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Tried reading schema message, was null or length 0\r\n```\nClosing this one as it was fixed in #654 \r\nFeel free to re-open if you have other questions","embeddings":[-0.4075993598,-0.2417839468,0.0564272702,0.2862923145,0.3149873018,-0.1689801663,-0.159183085,0.2923860252,-0.1545246542,0.4045127332,0.0525604412,0.5000802875,-0.1109343246,-0.3641472757,-0.094238542,-0.1911460012,-0.1322497874,0.0535943359,0.0623223409,-0.2779391706,-0.3765873313,0.0985771418,-0.2131596059,0.3267936409,-0.1932066679,-0.3521064818,0.1751866639,0.1816710234,-0.3589941561,-0.5263907909,0.2709598839,-0.2937383056,0.0037990306,0.1183886901,-0.0001191873,0.1296570301,0.3363063931,-0.0975792706,-0.0041522533,-0.1153502688,-0.0072667021,-0.0410663113,0.3097956181,-0.1023338959,0.1306812316,-0.1691236049,-0.2776666284,-0.2377279848,0.2970281839,0.3777075708,0.0835815147,0.2802878916,-0.0036599415,-0.2270237654,-0.1957388967,-0.2623844147,-0.0339903161,0.7621635795,0.0858863145,-0.212425679,0.0804970711,0.2712044716,0.3043489158,0.1928760856,0.434317559,-0.2398328185,0.2804588079,-0.002974591,-0.1896120757,0.1256083846,-0.0713665262,-0.1000525355,-0.2646160424,0.2437917441,0.2630293965,-0.5339521766,-0.0313146561,0.2237594873,-0.0697997287,0.0458759703,-0.240044713,0.0889871642,-0.1897262484,0.0984046757,0.163955614,0.130643785,0.015154738,0.2536816597,0.0861735046,0.0687298328,-0.3007571697,0.3731757998,-0.2901461422,-0.0234646499,-0.4769316614,-0.1539025456,-0.0954827592,0.0382494144,0.0169564392,0.1846893579,0.6731569171,-0.0968763605,0.2253452539,0.3182927072,0.1315817535,0.2608215213,-0.2110044658,0.5256132483,0.0850065351,-0.0524344891,-0.0642045215,0.0298821982,0.2078827918,-0.6180186272,0.1823914796,0.3520441949,0.0247808602,0.1011312529,-0.5585077405,-0.187474221,-0.4208103716,0.1441128701,0.1162040085,0.070241943,-0.0098986542,0.1733299792,0.2674604952,0.2471818775,-0.1467236131,-0.3565670252,-0.2359515429,0.1969593465,-0.2147193253,0.049003385,-0.0045695757,-0.0573178567,0.1841375381,0.090225853,0.0255612973,-0.1761726886,0.5448206067,-0.2204472274,-0.0769289583,0.0257570073,-0.3781884015,-0.2873996794,0.229777202,0.1334006041,0.037450254,-0.1065518111,-0.2053185105,-0.3397361338,0.2455221862,0.1057567075,0.3015190959,-0.0896550715,0.2021171451,0.1027184129,0.1186710373,-0.1679324359,0.1719890237,-0.2009641826,-0.1101601794,0.0065554334,0.0725460723,-0.0001086169,-0.1014669389,0.1107227728,0.1415624768,0.1545994133,0.0349537246,0.4689890444,-0.2500044107,0.2202896327,0.0234921817,-0.3029279113,0.5286731124,-0.9485337734,-0.2145374864,0.0507860072,-0.1609332114,-0.5299543142,-0.1245944723,-0.1013837904,-0.1401650608,0.1376924664,0.3832190335,0.0421500206,-0.1681721956,-0.2119553238,-0.2906453907,0.0716145709,0.246895358,0.1549103707,0.0768484026,0.0673309341,0.0960901529,-0.2710840702,0.0719325691,0.031280566,-0.0309009049,0.2132062167,-0.1415386796,-0.143858254,0.4165867269,-0.0793790221,0.2842078209,0.1112132147,-0.3692346513,-0.1870997995,-0.2376635224,0.0109891947,-0.3217407465,0.0674759075,-0.0672575831,-0.0084727677,0.0498084016,0.0204183981,0.2208724916,0.1332473457,-0.3149588406,-0.221074149,-0.4542842507,-0.030646855,-0.1260975599,0.0629270151,0.0274376441,-0.2461945862,0.0837185308,0.1811012775,0.160761103,-0.0457629487,-0.2479192317,0.4671181738,0.0894628614,0.0979057625,0.4763180912,-0.1720580012,0.1261303127,-0.2955779731,-0.3338622749,0.2377325296,0.3096678555,0.0850776434,0.1167073324,0.2165964097,-0.3274505138,-0.0149451001,-0.0602276884,0.2954143882,0.219982028,0.1660233587,-0.1664334536,0.0484345555,0.1911974251,-0.1303867102,0.39263165,-0.2581726313,-0.3667012453,-0.2043807954,0.1814842671,0.1247731745,0.0273581091,-0.2630428374,0.2203611284,0.2895263433,0.057939209,0.1685934812,0.4083440602,-0.08389882,0.0149456374,0.2547494471,-0.0987793058,-0.1508776844,0.0551872589,0.0896368101,0.1410100907,0.3735880852,0.2065856308,0.0663291514,-0.3672625124,-0.1021530479,0.2359729409,0.2093068361,-0.0221330989,0.4142645001,-0.0017026064,0.277654469,-0.1761566401,-0.1490000039,-0.1172153726,-0.1813320518,0.0745567977,0.0275700875,-0.0813113302,0.2263077348,0.2115996629,0.34382689,0.3297126591,0.0593655221,0.069641985,-0.1464833021,-0.1094480455,-0.12097238,-0.1135568023,0.0799365714,0.1879225671,-0.079188183,0.2124749422,-0.1500804126,-0.0069223507,-0.0119671673,0.0123387277,-0.0033971411,0.3124577403,-0.419359833,-0.6186656356,0.1612169147,0.3857408762,-0.1651768237,-0.1729445606,0.1802288145,0.0099464897,-0.0189911984,-0.1486451179,-0.3686364293,-0.408726573,-0.3623014987,0.3295700252,0.0106309811,-0.1517916471,0.1586492658,0.0680436641,0.1036152095,0.2316931784,-0.0726489872,0.020813426,-0.0302379169,0.2664110959,-0.2160875201,-0.2734997869,-0.0737783611,-0.1136969328,0.3361850679,0.4906616807,-0.1465961337,-0.2178466171,0.1189218536,0.1014851928,-0.5226741433,0.1076532006,0.5523734093,-0.1283311099,-0.0396117494,-0.1679284573,0.3205580413,0.4238615036,-0.0645174012,0.0601091981,-0.1110722274,0.1943630576,0.105065316,0.750133872,0.2554729879,-0.0024984078,0.1955074072,0.04087184,0.1205502972,0.1490536183,-0.0575668849,0.2131001651,0.1321832836,-0.1164219603,-0.0873289034,-0.1998218447,0.0960988998,-0.1304616034,-0.0857842565,0.1583218127,0.081030108,-0.117395632,-0.7013334036,0.5111743212,0.2526737154,0.175617069,-0.1056380644,0.0872496217,0.2066690624,-0.1311411709,0.4093682468,0.0597588904,0.0744302198,-0.3496944606,-0.0763837025,0.3868108988,0.1509348899,0.5822174549,-0.0802684203,0.2114307731,0.1476576775,0.0032646088,0.176494956,-0.4874759316,-0.0024385985,0.0024493767,-0.3182797432,-0.7047178745,0.0278705899,-0.2246014029,0.2069823146,-0.0390860923,0.2247414142,-0.0905078501,-0.0705309287,0.0914265737,-0.0246223025,-0.2360409647,-0.04375346,-0.43846187,-0.1007978246,-0.0079877311,0.3548945487,0.215847224,0.0242722016,0.2047313899,0.0260860771,-0.4139418304,0.2024679184,-0.1164642423,-0.1058029979,0.2377334386,-0.2303404808,-0.1086247787,0.1344448179,0.0973659903,0.1317891777,0.3259266615,0.1949752271,-0.7740002871,-0.0881980807,-0.0107653225,0.0912959501,0.1395350248,-0.1086555645,-0.0830888301,-0.1653802395,-0.0371893421,-0.1362710595,-0.0624066703,0.2667120993,-0.0437857248,0.2142005563,-0.0918841064,0.3918713927,0.0683675706,-0.0616462566,0.6337679029,-0.088559255,-0.1122555882,0.0026838945,0.4313238561,0.6851304173,0.3991342783,-0.2286525965,0.5048461556,-0.2030970156,0.2501061559,-0.2921949327,0.3559749126,-0.3496271372,-0.291167438,0.0264542475,-0.1216660365,0.0053279358,0.1125444621,0.0203569159,0.2939416468,-0.0912707224,-0.2852716744,-0.0292429142,0.0008494384,-0.2626026273,-0.0537064783,0.0257876385,-0.0062576835,-0.1543589681,-0.02531958,-0.1340193897,-0.2451969981,-0.1659554243,-0.101049982,0.1498215646,-0.0198502187,-0.2273249477,0.4922190309,0.372472465,-0.126603201,-0.1150018275,0.2851442099,0.4384118021,-0.0976752639,-0.090534538,0.3126313388,-0.1261661053,-0.2087913156,-0.1333062649,0.2157023996,0.0356794074,-0.1506471932,-0.2951408029,0.0754740834,-0.301020205,-0.2734054923,0.0300637912,0.0591305681,-0.2400632203,0.2178396136,0.1667966396,-0.1111845225,-0.3868939877,-0.0963925719,0.0177952964,0.1052126661,-0.3568505049,0.0832281485,-0.0021250511,-0.3438087702,0.3042221367,0.3032570481,-0.1221484467,0.0216180794,0.3507742584,-0.2816231251,0.0514557138,-0.2250773758,-0.0476729162,0.1816169769,-0.3708324134,0.3871153891,-0.1974066198,0.2173268646,0.1861213297,0.5506527424,0.1998642981,0.0305276476,0.0216813646,-0.3809008002,-0.281195879,0.0749440119,-0.10570243,-0.1554643065,-0.0670688301,0.0661698952,-0.063046284,0.1681356132,-0.1711985916,0.0182608329,-0.4026435018,0.1385002881,-0.2358004749,-0.1427798569,-0.2918173969,0.1429328918,-0.105355382,0.47483325,-0.1834557205,-0.0251481198,0.0208589584,0.1604123265,0.1491889954,0.2476718426,0.0554537214,-0.0948632434,-0.4114228487,-0.0237532984,-0.1994509399,-0.1094015688,-0.1021039709,0.1169447973,-0.148319602,0.2587156296,-0.0825612769,-0.102657862,0.2263108045,0.0078636818,-0.4149520099,-0.0237018876,0.1374997199,-0.1295656264,-0.4047162831,-0.071770817,-0.1746266186,0.032731384,0.235680297,-0.3694477081,-0.0406388976,-0.3534323275,0.3480403125,0.4453604221,-0.4016061723,-0.0821943805,0.0196640231,0.0796486735,-0.2390367687,0.0348827839,-0.0927760378,0.4075059891,-0.0362933762,0.0683815405,0.2137533724,-0.3845719993,-0.0951184109,0.1793109775,-0.1105760559,-0.5657693744,0.183415845,-0.0904031694,-0.0515371934,0.0342611969,0.4253365993,0.3455851972,0.2773367465,0.378754884,-0.1718459576,0.3203332722,0.2994095683,-0.0327489302,-0.0350370258,-0.0182347279,-0.324139595,-0.0569081418,0.1216445416,-0.2019240558,0.1269879639,0.4522322714,-0.5222976804,-0.5407586694,-0.1329667121,-0.1093748659,-0.0489517786,0.2321892083,-0.5116025209,-0.1040367335,-0.0905237794,-0.0982390195,0.2170716226,0.5896323919,0.5254977345,0.2307590097,0.0998491794,-0.4661764801,-0.0672634095,0.5261714458,0.3482179344,-0.3474153578,0.0702713281,0.2197810411,-0.0478885323,-0.1836316586,0.468246758,0.1106709316,0.4221267104,-0.3368279338,-0.0574068166,-0.0913393646,0.055816371,0.188318342,0.4640157223,0.4181191623,-0.1294232756,0.4254341722,-0.0807892233,-0.0253640153,0.3410935402,-0.1127055511,0.5322974324,-0.1863979995,0.2962688506,-0.1552479714,0.1272639781,-0.1232938096,0.0020000769,-0.5153283477,0.3860246539,-0.291657418,0.0725813955,0.0253607891,-0.271500349,0.0217014272,-0.2019307315,0.4118304849,0.254483372,-0.0213224012,-0.1645658463,-0.364146024,-0.3480188847,0.1979256719,-0.0347262323,-0.0323350802,0.0377677195,0.2458748519,-0.0899563655,0.2194443345,0.0136135528,0.3184034824,0.1633765846,0.4191622138,-0.4301507175,-0.081525147,-0.3555106819,-0.1731373817,0.2258825153,-0.1960830986,0.0569185689,-0.0774393007,-0.0041136811,0.2366613001,-0.4406211674,0.0182626452,0.4891805351,0.4408990741,0.1522389948,0.2984647155,0.0991083309,0.0148841208,-0.1326742172,-0.360555619,0.0081420112,0.3292993903,0.0611585565,0.3952183425,-0.1376777887,0.1447617561,-0.2830953002,0.372395575,-0.1469091177,-0.0667040795,-0.0228478238,0.0460149981,-0.0089644669,0.1270906031,-0.0131379813,0.1258742809,0.1562907994,0.0732177496,-0.2892949283,-0.2071141303,0.5095154047,-0.152072221,-0.4998908937,0.1289652586,0.3653587103,0.8499709368,-0.111030139,-0.4012586176,0.0702343807,0.1562007815,-0.045343522,-0.0882446542,-0.0701677501,0.0586670339,0.1032955199,0.1094179526,0.0976541489,-0.05990557,0.0193283074,0.2159411013,-0.1393848062]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/565","title":"No module named 'nlp.logging'","comments":"Thanks for reporting.\r\n\r\nApparently this is a versioning issue: the lib downloaded the `bleurt` script from the master branch where we did this change recently. We'll fix that in a new release this week or early next week. Cc @thomwolf \r\n\r\nUntil that, I'd suggest you to download the right bleurt folder from github ([this one](https:\/\/github.com\/huggingface\/nlp\/tree\/0.4.0\/metrics\/bleurt)) and do\r\n\r\n```python\r\nfrom nlp import load_metric\r\n\r\nbleurt = load_metric(\"path\/to\/bleurt\/folder\")\r\n```\r\n\r\nTo download it you can either clone the repo or download the `bleurt.py` file and place it in a folder named `bleurt` ","body":"Hi, I am using nlp version 0.4.0. Trying to use bleurt as an eval metric, however, the bleurt script imports nlp.logging which creates the following error. What am I missing?\r\n\r\n```\r\n>>> import nlp\r\n2020-09-02 13:47:09.210310: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n>>> bleurt = nlp.load_metric(\"bleurt\")\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 443, in load_metric\r\n metric_cls = import_main_class(module_path, dataset=False)\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 61, in import_main_class\r\n module = importlib.import_module(module_path)\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/metrics\/bleurt\/43448cf2959ea81d3ae0e71c5c8ee31dc15eed9932f197f5f50673cbcecff2b5\/bleurt.py\", line 20, in <module>\r\n from nlp.logging import get_logger\r\nModuleNotFoundError: No module named 'nlp.logging'\r\n```\r\n\r\nJust to show once again that I can't import the logging module:\r\n\r\n```\r\n>>> import nlp\r\n2020-09-02 13:48:38.190621: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n>>> nlp.__version__\r\n'0.4.0'\r\n>>> from nlp.logging import get_logger\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nModuleNotFoundError: No module named 'nlp.logging'\r\n```","comment_length":88,"text":"No module named 'nlp.logging'\nHi, I am using nlp version 0.4.0. Trying to use bleurt as an eval metric, however, the bleurt script imports nlp.logging which creates the following error. What am I missing?\r\n\r\n```\r\n>>> import nlp\r\n2020-09-02 13:47:09.210310: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n>>> bleurt = nlp.load_metric(\"bleurt\")\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 443, in load_metric\r\n metric_cls = import_main_class(module_path, dataset=False)\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 61, in import_main_class\r\n module = importlib.import_module(module_path)\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/metrics\/bleurt\/43448cf2959ea81d3ae0e71c5c8ee31dc15eed9932f197f5f50673cbcecff2b5\/bleurt.py\", line 20, in <module>\r\n from nlp.logging import get_logger\r\nModuleNotFoundError: No module named 'nlp.logging'\r\n```\r\n\r\nJust to show once again that I can't import the logging module:\r\n\r\n```\r\n>>> import nlp\r\n2020-09-02 13:48:38.190621: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n>>> nlp.__version__\r\n'0.4.0'\r\n>>> from nlp.logging import get_logger\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nModuleNotFoundError: No module named 'nlp.logging'\r\n```\nThanks for reporting.\r\n\r\nApparently this is a versioning issue: the lib downloaded the `bleurt` script from the master branch where we did this change recently. We'll fix that in a new release this week or early next week. Cc @thomwolf \r\n\r\nUntil that, I'd suggest you to download the right bleurt folder from github ([this one](https:\/\/github.com\/huggingface\/nlp\/tree\/0.4.0\/metrics\/bleurt)) and do\r\n\r\n```python\r\nfrom nlp import load_metric\r\n\r\nbleurt = load_metric(\"path\/to\/bleurt\/folder\")\r\n```\r\n\r\nTo download it you can either clone the repo or download the `bleurt.py` file and place it in a folder named `bleurt` ","embeddings":[0.0177445877,-0.3961698413,-0.0208087899,-0.2511645257,0.2170929313,-0.0493598208,0.2021470666,0.3259567022,0.0938878134,-0.0865905285,0.157289505,0.3769112527,-0.678425312,0.1394217014,0.2663362026,-0.1221125945,-0.1415552497,0.1301270425,0.4967287779,-0.0840892345,-0.1551878899,0.289421618,-0.125125289,0.1515601426,-0.237595126,0.1572917402,0.1200002357,-0.061919082,-0.0776275024,-0.3836641014,0.070940122,-0.3669742346,0.3694146574,0.0429327115,-0.0001146313,-0.2010705024,0.3691684306,-0.0789471492,-0.307600379,-0.4952634275,0.0295944158,-0.5176201463,0.2748705149,-0.1576724201,0.149302572,-0.2864575684,0.2366618663,-0.3097525537,0.1938056946,0.4161177576,0.2068959326,-0.0221976209,-0.0768320784,0.0798196793,0.0021625236,-0.0918753371,-0.0400303714,0.3646017313,0.1682526171,-0.2020527273,-0.2094080299,0.0075306385,0.0123415431,-0.0307616368,0.4454035163,0.0410990827,0.4342164695,0.0744407326,0.1132750139,0.0331307426,0.3260391355,-0.2414316237,0.2124530524,0.2793803215,0.2856973112,-0.6526633501,0.0322943069,0.3493673801,-0.1642590314,-0.1975704432,-0.3211485744,-0.1605102569,-0.0767731816,0.2774338126,-0.0267945323,0.2369682789,-0.1171348989,-0.0183080826,0.0479584634,0.0457221493,0.051273074,0.3074576557,0.0579455532,0.2783448696,-0.0295917876,0.0939961076,0.3766542375,0.0809352994,-0.3641400933,0.037301816,-0.0792932957,-0.2028973997,0.0826363415,0.1005701274,0.128095448,0.482413888,0.5802249908,0.0861540437,0.1211153269,0.1061822399,0.0224334728,0.1806164533,0.0178397074,-0.2114959955,-0.0291198101,-0.0329339616,-0.0712683052,-0.0087648984,-0.2257304788,-0.0556228422,-0.2265833467,-0.0307782069,0.1380367726,0.3203542829,-0.2769455314,-0.3906171024,0.5350926518,0.1231341138,-0.2232646644,-0.032176964,-0.1262366623,0.5046454072,-0.4369259179,-0.1370981336,0.143480733,-0.0797784775,0.4327978492,-0.1558860689,-0.1006639972,-0.2113872468,-0.0947681069,0.2198672891,-0.0771770924,-0.0099433828,-0.0664815083,-0.3026097715,0.0548935197,-0.1666546017,-0.1033997461,-0.0218892228,-0.1572464854,-0.3142599761,-0.2244267315,0.1509290785,-0.3393890858,-0.2686007321,0.2011347711,0.0782153308,-0.0370585285,-0.3826788962,-0.1565061957,-0.1386644989,0.0589941666,-0.0779406652,0.1836316586,-0.0987101644,0.07261803,-0.3945173919,-0.3643250465,0.2603411674,0.3422808349,-0.1949792653,-0.2013522387,0.2731038332,-0.1028451994,-0.1197198927,0.625600636,-0.2656618953,0.2711018324,-0.2094727308,-0.0848327279,-0.3129795194,-0.0258243866,-0.0342767574,-0.1933761388,-0.0780976713,0.3543556333,0.2127129138,-0.0401689,-0.294375211,-0.1962094605,-0.1480362713,0.1277350634,0.1170276999,0.3017060757,0.042928759,0.1312465668,0.5985461473,-0.0901134312,-0.0685134977,-0.0750644207,0.1411289424,0.180771023,-0.2709906995,0.1386940926,-0.1707386076,0.119585596,-0.1768057793,-0.3409104049,0.0959557295,0.0197773781,0.2479205132,-0.1413602382,-0.1239601895,0.117582202,-0.3142851293,0.1030290574,0.0294074677,-0.1288951039,0.3554201126,-0.382553339,-0.0465960838,-0.2218794823,-0.1060575098,-0.1710862815,-0.1487460881,-0.0094743446,-0.2253688574,0.1783795655,0.6324031949,-0.0850064158,0.2346969992,-0.0037051507,-0.1035135016,-0.0135534313,0.2673344016,0.0565778017,-0.0577418916,-0.0482448675,-0.0602159128,0.1752789021,0.0162852742,-0.096297048,-0.0791670084,0.4494024813,0.0644367188,0.2025585771,0.0629727095,0.1511247158,0.2264125198,-0.0125715872,-0.2510657012,0.0872598439,-0.1092128158,0.8314300776,-0.1710357815,0.2515512705,-0.2746801674,0.0517181419,-0.200482592,0.6626394391,0.2042042166,0.0820650309,0.180214867,-0.2810432315,-0.0692387745,0.1406123638,0.0014656645,0.3870842755,0.2310897112,-0.0509548113,0.2067657709,-0.1034539565,-0.1705688387,0.1831772625,-0.0443811901,-0.2436577231,0.0247675907,0.1182230189,-0.130331859,-0.1481409669,-0.1530951858,-0.123313278,0.1996001899,-0.0350606553,0.1603679359,-0.3734546304,-0.1339826435,-0.2300897241,-0.1811053157,-0.229476884,-0.2983353436,0.0284396969,0.0945010856,0.0437802263,0.085607782,0.4398249984,0.2297273874,-0.2515345216,-0.1649123132,-0.0883319676,-0.4656353891,-0.468177855,0.0709280223,0.1150501594,0.2313400209,0.2546295226,-0.1552510709,0.1433707327,-0.0437558442,-0.5664678216,-0.0181738697,0.0860506445,0.070925042,0.3076506257,-0.06650085,-0.1981247663,-0.3041720986,0.1178149879,-0.2248552144,0.0074069975,0.2838937044,-0.1443242431,-0.0391858928,-0.3543541729,-0.3782945871,-0.2264765948,-0.5012181401,0.0852806047,0.0689325258,-0.0958226547,0.438611716,0.0011936188,0.5463566184,-0.0049784738,0.2645167112,0.0071177036,0.3665676117,0.1443119496,-0.2059247047,-0.3020400703,0.1247395277,-0.123570092,0.3829789162,0.0842039064,-0.3999665678,-0.3124961853,0.3001262546,-0.4379544556,0.0871436968,0.1368221492,0.2238348871,0.017371418,-0.0351178348,-0.1492095739,-0.1678654104,0.1437755525,0.137040928,0.7387797832,-0.0642652363,0.2167179286,0.0948870853,0.3520082235,0.1043622345,-0.3068554103,0.0734939575,0.044195313,0.2859508991,0.1100366488,-0.2359181792,0.1601347625,0.0702000782,-0.1433645785,0.3142095208,-0.0822944865,-0.2094067186,-0.3659453392,-0.0670050904,0.1722664833,-0.0579462089,0.0224081315,-0.0850380585,0.2522856891,-0.0257230792,-0.1219104826,-0.3413435221,-0.1392955631,0.0958547443,0.269105345,-0.0436150208,0.1610604078,0.2585424483,-0.1735367626,-0.4186548889,0.3763270974,-0.0102599775,0.0753204748,-0.1960619837,-0.0525866486,0.0923788324,0.1480776072,-0.4450248182,0.0221821908,-0.0322811268,0.1799473614,0.220301196,-0.4206765294,0.0985987782,-0.0531907082,0.1326414943,0.6327114105,-0.028860243,-0.1421376914,0.001626344,0.3982037604,0.0126886014,0.0350451991,-0.2267851382,-0.4677221477,-0.3101116419,0.0240105502,-0.4039107859,-0.1081059352,0.1898400038,0.4054775536,0.2305119187,0.1628723741,-0.2111327499,0.2652073205,-0.1094020456,-0.0186426993,-0.2338207513,-0.1241565645,-0.176509887,0.0746410117,-0.2352043986,-0.0406512618,-0.1481791288,-0.4993696213,0.034539137,0.0694435984,0.3608239889,-0.0557444803,-0.1037609503,0.1168950126,-0.2295798659,0.043443732,0.1421991736,0.344650954,0.2187676281,0.247687608,0.1366506368,-0.0933897495,0.3914400041,-0.0074757617,0.2628164589,0.6665168405,-0.2765217721,-0.2091833055,0.0268203486,0.0259910617,0.6081098318,0.3980511725,-0.121754691,0.5636966825,-0.0332950503,0.3709040284,-0.2529835999,0.3597998321,-0.208351329,0.2281372994,0.1237296388,-0.2150111496,0.133933872,-0.3755695224,-0.0285154656,0.1180475801,-0.1512495428,-0.1804251373,0.0978511125,0.2777616978,-0.0858836398,0.0670812503,-0.3402472138,0.1049265191,-0.5141498446,0.3371415138,-0.1234817132,-0.1297787577,0.0850417539,-0.0571581684,-0.0586516,0.2024144828,0.1764393747,0.2453874052,0.179105252,0.0015667555,0.1329327524,0.3271038532,0.4160194397,-0.1411470771,-0.1413773745,0.1801221818,-0.2844997048,-0.0953674316,0.2117823064,0.2493564188,0.3500951231,-0.320893079,0.0167744104,0.0921131968,0.0947501957,-0.3090747893,0.2534045875,0.3062739372,-0.0042506745,0.3142991364,0.0549887866,0.4683715105,0.0037165184,0.203883633,0.1448802948,0.1391287595,-0.1887374371,0.1346669942,-0.0655604526,-0.2411875725,0.0385503098,0.1038153023,0.068636097,-0.0115708206,0.3635194898,-0.171493575,-0.1635908335,-0.0536273643,-0.516215384,-0.0766493827,-0.2516139746,0.2087752819,0.2589109242,-0.1144478396,0.3502965271,0.3396171331,0.0072795763,-0.003563327,-0.1966385394,-0.1979463696,-0.3698661029,0.3572338223,-0.4561422169,-0.0821598545,0.0321734101,0.5907017589,0.0789095238,0.0997670293,-0.2613565028,-0.1251948327,-0.1099203005,0.1504797935,0.1698433012,-0.3847261965,0.1303604245,0.224498868,0.0167798679,0.0390081592,-0.3336628973,-0.1411159635,-0.264118135,0.1118496731,-0.0036447519,0.2566160262,-0.0333230793,-0.2517702579,-0.2034108639,-0.0182718504,-0.1530154347,0.100002557,-0.1967065781,0.1052413881,-0.3052292168,0.1879130304,0.2350843549,-0.252890259,0.0142704416,0.1697072238,0.0677078366,-0.0516888388,0.1439801306,0.1308486462,-0.2021969706,-0.1405560076,-0.0514319278,0.0384286083,0.2726108432,0.0369232222,0.0304564293,-0.5750373602,-0.2175094336,0.2021585107,-0.0762687102,-0.0241153091,0.187298134,0.0907590091,-0.2569817305,0.202582106,-0.0290613882,0.3365930915,-0.2204916626,-0.0248667747,0.1196049005,0.0173409712,0.3703121841,0.0830854401,-0.1299896091,-0.624268055,0.1826901883,-0.3209093511,-0.2156856358,0.0443556421,0.1597321779,0.2159297019,0.1625348479,-0.1061165184,-0.1998499781,0.2159614861,-0.1669279486,-0.0119243506,0.080601573,-0.0823673233,-0.1484625638,0.2061640769,0.0952175781,0.1660230309,-0.1503381133,0.5230813622,-0.2694711685,0.1774445623,-0.1627534628,0.2485528141,-0.0211527608,0.3340616226,0.3533409536,0.1310346723,-0.064106226,-0.0746059045,0.3371078968,0.1211457551,-0.2316764742,0.2225821912,-0.2420724481,0.0484229624,-0.4349020422,0.2174109221,0.1711486727,-0.0655467808,-0.0372691602,-0.1865079105,0.0532967374,-0.0720687285,0.2679019272,0.1563833952,0.4642439783,0.0486471318,-0.2285404652,0.2022954673,0.0982777849,0.0707495064,0.3525562584,0.6207296848,0.3232081234,-0.0066832323,0.0765518174,-0.1554659158,0.1863847375,-0.0107102385,-0.1345583498,-0.144550696,0.5972315073,0.1582023799,0.077412948,-0.539167285,0.0336949863,-0.5927145481,-0.0046077524,0.1251922697,0.3890294433,-0.1128291711,-0.1896718591,0.0589730777,-0.1500441134,-0.072370261,0.269099474,0.277995944,-0.0097625824,-0.2528842688,-0.2057820261,0.2209814638,0.0981387123,-0.1787582636,-0.1262218654,-0.0857611448,-0.1896122843,-0.129774034,-0.1223468035,0.1015101299,0.0939781964,-0.5398161411,-0.3409730792,-0.1714624763,0.0252823122,-0.0521862358,-0.0143680945,-0.0892978907,-0.2821855545,-0.4951393008,0.0742339045,-0.0983189568,-0.0900792107,0.2250673622,0.5810584426,0.2559539378,-0.1868476421,0.4078856707,0.1008466557,-0.1221649349,-0.202649951,0.0749394596,-0.0977010354,-0.1044102013,0.2023274899,0.2365569323,-0.3165027499,0.1566904932,-0.3306831717,0.8244410157,0.03512327,-0.2451610118,-0.114677757,0.0505557247,-0.098811157,0.309276849,0.2174607962,0.4579276145,-0.1223986596,0.0678027794,-0.3379885852,-0.0043132771,0.4632381797,-0.6978577971,-0.2602540553,0.3898676038,0.4401413798,0.3855003417,-0.1703182757,-0.686550498,-0.2105173469,0.0655455589,0.0370597355,0.1663676053,0.335321337,-0.0993318483,0.1353846937,0.0923244283,-0.2173095047,0.2460513115,-0.022815764,0.0020172163,-0.2244128436]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/565","title":"No module named 'nlp.logging'","comments":"Actually we can fix this on our side, this script didn't had to be updated. I'll do it in a few minutes","body":"Hi, I am using nlp version 0.4.0. Trying to use bleurt as an eval metric, however, the bleurt script imports nlp.logging which creates the following error. What am I missing?\r\n\r\n```\r\n>>> import nlp\r\n2020-09-02 13:47:09.210310: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n>>> bleurt = nlp.load_metric(\"bleurt\")\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 443, in load_metric\r\n metric_cls = import_main_class(module_path, dataset=False)\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 61, in import_main_class\r\n module = importlib.import_module(module_path)\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/metrics\/bleurt\/43448cf2959ea81d3ae0e71c5c8ee31dc15eed9932f197f5f50673cbcecff2b5\/bleurt.py\", line 20, in <module>\r\n from nlp.logging import get_logger\r\nModuleNotFoundError: No module named 'nlp.logging'\r\n```\r\n\r\nJust to show once again that I can't import the logging module:\r\n\r\n```\r\n>>> import nlp\r\n2020-09-02 13:48:38.190621: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n>>> nlp.__version__\r\n'0.4.0'\r\n>>> from nlp.logging import get_logger\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nModuleNotFoundError: No module named 'nlp.logging'\r\n```","comment_length":22,"text":"No module named 'nlp.logging'\nHi, I am using nlp version 0.4.0. Trying to use bleurt as an eval metric, however, the bleurt script imports nlp.logging which creates the following error. What am I missing?\r\n\r\n```\r\n>>> import nlp\r\n2020-09-02 13:47:09.210310: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n>>> bleurt = nlp.load_metric(\"bleurt\")\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 443, in load_metric\r\n metric_cls = import_main_class(module_path, dataset=False)\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 61, in import_main_class\r\n module = importlib.import_module(module_path)\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n File \"\/home\/melody\/anaconda3\/envs\/transformers\/lib\/python3.6\/site-packages\/nlp\/metrics\/bleurt\/43448cf2959ea81d3ae0e71c5c8ee31dc15eed9932f197f5f50673cbcecff2b5\/bleurt.py\", line 20, in <module>\r\n from nlp.logging import get_logger\r\nModuleNotFoundError: No module named 'nlp.logging'\r\n```\r\n\r\nJust to show once again that I can't import the logging module:\r\n\r\n```\r\n>>> import nlp\r\n2020-09-02 13:48:38.190621: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1\r\n>>> nlp.__version__\r\n'0.4.0'\r\n>>> from nlp.logging import get_logger\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nModuleNotFoundError: No module named 'nlp.logging'\r\n```\nActually we can fix this on our side, this script didn't had to be updated. I'll do it in a few minutes","embeddings":[0.0177445877,-0.3961698413,-0.0208087899,-0.2511645257,0.2170929313,-0.0493598208,0.2021470666,0.3259567022,0.0938878134,-0.0865905285,0.157289505,0.3769112527,-0.678425312,0.1394217014,0.2663362026,-0.1221125945,-0.1415552497,0.1301270425,0.4967287779,-0.0840892345,-0.1551878899,0.289421618,-0.125125289,0.1515601426,-0.237595126,0.1572917402,0.1200002357,-0.061919082,-0.0776275024,-0.3836641014,0.070940122,-0.3669742346,0.3694146574,0.0429327115,-0.0001146313,-0.2010705024,0.3691684306,-0.0789471492,-0.307600379,-0.4952634275,0.0295944158,-0.5176201463,0.2748705149,-0.1576724201,0.149302572,-0.2864575684,0.2366618663,-0.3097525537,0.1938056946,0.4161177576,0.2068959326,-0.0221976209,-0.0768320784,0.0798196793,0.0021625236,-0.0918753371,-0.0400303714,0.3646017313,0.1682526171,-0.2020527273,-0.2094080299,0.0075306385,0.0123415431,-0.0307616368,0.4454035163,0.0410990827,0.4342164695,0.0744407326,0.1132750139,0.0331307426,0.3260391355,-0.2414316237,0.2124530524,0.2793803215,0.2856973112,-0.6526633501,0.0322943069,0.3493673801,-0.1642590314,-0.1975704432,-0.3211485744,-0.1605102569,-0.0767731816,0.2774338126,-0.0267945323,0.2369682789,-0.1171348989,-0.0183080826,0.0479584634,0.0457221493,0.051273074,0.3074576557,0.0579455532,0.2783448696,-0.0295917876,0.0939961076,0.3766542375,0.0809352994,-0.3641400933,0.037301816,-0.0792932957,-0.2028973997,0.0826363415,0.1005701274,0.128095448,0.482413888,0.5802249908,0.0861540437,0.1211153269,0.1061822399,0.0224334728,0.1806164533,0.0178397074,-0.2114959955,-0.0291198101,-0.0329339616,-0.0712683052,-0.0087648984,-0.2257304788,-0.0556228422,-0.2265833467,-0.0307782069,0.1380367726,0.3203542829,-0.2769455314,-0.3906171024,0.5350926518,0.1231341138,-0.2232646644,-0.032176964,-0.1262366623,0.5046454072,-0.4369259179,-0.1370981336,0.143480733,-0.0797784775,0.4327978492,-0.1558860689,-0.1006639972,-0.2113872468,-0.0947681069,0.2198672891,-0.0771770924,-0.0099433828,-0.0664815083,-0.3026097715,0.0548935197,-0.1666546017,-0.1033997461,-0.0218892228,-0.1572464854,-0.3142599761,-0.2244267315,0.1509290785,-0.3393890858,-0.2686007321,0.2011347711,0.0782153308,-0.0370585285,-0.3826788962,-0.1565061957,-0.1386644989,0.0589941666,-0.0779406652,0.1836316586,-0.0987101644,0.07261803,-0.3945173919,-0.3643250465,0.2603411674,0.3422808349,-0.1949792653,-0.2013522387,0.2731038332,-0.1028451994,-0.1197198927,0.625600636,-0.2656618953,0.2711018324,-0.2094727308,-0.0848327279,-0.3129795194,-0.0258243866,-0.0342767574,-0.1933761388,-0.0780976713,0.3543556333,0.2127129138,-0.0401689,-0.294375211,-0.1962094605,-0.1480362713,0.1277350634,0.1170276999,0.3017060757,0.042928759,0.1312465668,0.5985461473,-0.0901134312,-0.0685134977,-0.0750644207,0.1411289424,0.180771023,-0.2709906995,0.1386940926,-0.1707386076,0.119585596,-0.1768057793,-0.3409104049,0.0959557295,0.0197773781,0.2479205132,-0.1413602382,-0.1239601895,0.117582202,-0.3142851293,0.1030290574,0.0294074677,-0.1288951039,0.3554201126,-0.382553339,-0.0465960838,-0.2218794823,-0.1060575098,-0.1710862815,-0.1487460881,-0.0094743446,-0.2253688574,0.1783795655,0.6324031949,-0.0850064158,0.2346969992,-0.0037051507,-0.1035135016,-0.0135534313,0.2673344016,0.0565778017,-0.0577418916,-0.0482448675,-0.0602159128,0.1752789021,0.0162852742,-0.096297048,-0.0791670084,0.4494024813,0.0644367188,0.2025585771,0.0629727095,0.1511247158,0.2264125198,-0.0125715872,-0.2510657012,0.0872598439,-0.1092128158,0.8314300776,-0.1710357815,0.2515512705,-0.2746801674,0.0517181419,-0.200482592,0.6626394391,0.2042042166,0.0820650309,0.180214867,-0.2810432315,-0.0692387745,0.1406123638,0.0014656645,0.3870842755,0.2310897112,-0.0509548113,0.2067657709,-0.1034539565,-0.1705688387,0.1831772625,-0.0443811901,-0.2436577231,0.0247675907,0.1182230189,-0.130331859,-0.1481409669,-0.1530951858,-0.123313278,0.1996001899,-0.0350606553,0.1603679359,-0.3734546304,-0.1339826435,-0.2300897241,-0.1811053157,-0.229476884,-0.2983353436,0.0284396969,0.0945010856,0.0437802263,0.085607782,0.4398249984,0.2297273874,-0.2515345216,-0.1649123132,-0.0883319676,-0.4656353891,-0.468177855,0.0709280223,0.1150501594,0.2313400209,0.2546295226,-0.1552510709,0.1433707327,-0.0437558442,-0.5664678216,-0.0181738697,0.0860506445,0.070925042,0.3076506257,-0.06650085,-0.1981247663,-0.3041720986,0.1178149879,-0.2248552144,0.0074069975,0.2838937044,-0.1443242431,-0.0391858928,-0.3543541729,-0.3782945871,-0.2264765948,-0.5012181401,0.0852806047,0.0689325258,-0.0958226547,0.438611716,0.0011936188,0.5463566184,-0.0049784738,0.2645167112,0.0071177036,0.3665676117,0.1443119496,-0.2059247047,-0.3020400703,0.1247395277,-0.123570092,0.3829789162,0.0842039064,-0.3999665678,-0.3124961853,0.3001262546,-0.4379544556,0.0871436968,0.1368221492,0.2238348871,0.017371418,-0.0351178348,-0.1492095739,-0.1678654104,0.1437755525,0.137040928,0.7387797832,-0.0642652363,0.2167179286,0.0948870853,0.3520082235,0.1043622345,-0.3068554103,0.0734939575,0.044195313,0.2859508991,0.1100366488,-0.2359181792,0.1601347625,0.0702000782,-0.1433645785,0.3142095208,-0.0822944865,-0.2094067186,-0.3659453392,-0.0670050904,0.1722664833,-0.0579462089,0.0224081315,-0.0850380585,0.2522856891,-0.0257230792,-0.1219104826,-0.3413435221,-0.1392955631,0.0958547443,0.269105345,-0.0436150208,0.1610604078,0.2585424483,-0.1735367626,-0.4186548889,0.3763270974,-0.0102599775,0.0753204748,-0.1960619837,-0.0525866486,0.0923788324,0.1480776072,-0.4450248182,0.0221821908,-0.0322811268,0.1799473614,0.220301196,-0.4206765294,0.0985987782,-0.0531907082,0.1326414943,0.6327114105,-0.028860243,-0.1421376914,0.001626344,0.3982037604,0.0126886014,0.0350451991,-0.2267851382,-0.4677221477,-0.3101116419,0.0240105502,-0.4039107859,-0.1081059352,0.1898400038,0.4054775536,0.2305119187,0.1628723741,-0.2111327499,0.2652073205,-0.1094020456,-0.0186426993,-0.2338207513,-0.1241565645,-0.176509887,0.0746410117,-0.2352043986,-0.0406512618,-0.1481791288,-0.4993696213,0.034539137,0.0694435984,0.3608239889,-0.0557444803,-0.1037609503,0.1168950126,-0.2295798659,0.043443732,0.1421991736,0.344650954,0.2187676281,0.247687608,0.1366506368,-0.0933897495,0.3914400041,-0.0074757617,0.2628164589,0.6665168405,-0.2765217721,-0.2091833055,0.0268203486,0.0259910617,0.6081098318,0.3980511725,-0.121754691,0.5636966825,-0.0332950503,0.3709040284,-0.2529835999,0.3597998321,-0.208351329,0.2281372994,0.1237296388,-0.2150111496,0.133933872,-0.3755695224,-0.0285154656,0.1180475801,-0.1512495428,-0.1804251373,0.0978511125,0.2777616978,-0.0858836398,0.0670812503,-0.3402472138,0.1049265191,-0.5141498446,0.3371415138,-0.1234817132,-0.1297787577,0.0850417539,-0.0571581684,-0.0586516,0.2024144828,0.1764393747,0.2453874052,0.179105252,0.0015667555,0.1329327524,0.3271038532,0.4160194397,-0.1411470771,-0.1413773745,0.1801221818,-0.2844997048,-0.0953674316,0.2117823064,0.2493564188,0.3500951231,-0.320893079,0.0167744104,0.0921131968,0.0947501957,-0.3090747893,0.2534045875,0.3062739372,-0.0042506745,0.3142991364,0.0549887866,0.4683715105,0.0037165184,0.203883633,0.1448802948,0.1391287595,-0.1887374371,0.1346669942,-0.0655604526,-0.2411875725,0.0385503098,0.1038153023,0.068636097,-0.0115708206,0.3635194898,-0.171493575,-0.1635908335,-0.0536273643,-0.516215384,-0.0766493827,-0.2516139746,0.2087752819,0.2589109242,-0.1144478396,0.3502965271,0.3396171331,0.0072795763,-0.003563327,-0.1966385394,-0.1979463696,-0.3698661029,0.3572338223,-0.4561422169,-0.0821598545,0.0321734101,0.5907017589,0.0789095238,0.0997670293,-0.2613565028,-0.1251948327,-0.1099203005,0.1504797935,0.1698433012,-0.3847261965,0.1303604245,0.224498868,0.0167798679,0.0390081592,-0.3336628973,-0.1411159635,-0.264118135,0.1118496731,-0.0036447519,0.2566160262,-0.0333230793,-0.2517702579,-0.2034108639,-0.0182718504,-0.1530154347,0.100002557,-0.1967065781,0.1052413881,-0.3052292168,0.1879130304,0.2350843549,-0.252890259,0.0142704416,0.1697072238,0.0677078366,-0.0516888388,0.1439801306,0.1308486462,-0.2021969706,-0.1405560076,-0.0514319278,0.0384286083,0.2726108432,0.0369232222,0.0304564293,-0.5750373602,-0.2175094336,0.2021585107,-0.0762687102,-0.0241153091,0.187298134,0.0907590091,-0.2569817305,0.202582106,-0.0290613882,0.3365930915,-0.2204916626,-0.0248667747,0.1196049005,0.0173409712,0.3703121841,0.0830854401,-0.1299896091,-0.624268055,0.1826901883,-0.3209093511,-0.2156856358,0.0443556421,0.1597321779,0.2159297019,0.1625348479,-0.1061165184,-0.1998499781,0.2159614861,-0.1669279486,-0.0119243506,0.080601573,-0.0823673233,-0.1484625638,0.2061640769,0.0952175781,0.1660230309,-0.1503381133,0.5230813622,-0.2694711685,0.1774445623,-0.1627534628,0.2485528141,-0.0211527608,0.3340616226,0.3533409536,0.1310346723,-0.064106226,-0.0746059045,0.3371078968,0.1211457551,-0.2316764742,0.2225821912,-0.2420724481,0.0484229624,-0.4349020422,0.2174109221,0.1711486727,-0.0655467808,-0.0372691602,-0.1865079105,0.0532967374,-0.0720687285,0.2679019272,0.1563833952,0.4642439783,0.0486471318,-0.2285404652,0.2022954673,0.0982777849,0.0707495064,0.3525562584,0.6207296848,0.3232081234,-0.0066832323,0.0765518174,-0.1554659158,0.1863847375,-0.0107102385,-0.1345583498,-0.144550696,0.5972315073,0.1582023799,0.077412948,-0.539167285,0.0336949863,-0.5927145481,-0.0046077524,0.1251922697,0.3890294433,-0.1128291711,-0.1896718591,0.0589730777,-0.1500441134,-0.072370261,0.269099474,0.277995944,-0.0097625824,-0.2528842688,-0.2057820261,0.2209814638,0.0981387123,-0.1787582636,-0.1262218654,-0.0857611448,-0.1896122843,-0.129774034,-0.1223468035,0.1015101299,0.0939781964,-0.5398161411,-0.3409730792,-0.1714624763,0.0252823122,-0.0521862358,-0.0143680945,-0.0892978907,-0.2821855545,-0.4951393008,0.0742339045,-0.0983189568,-0.0900792107,0.2250673622,0.5810584426,0.2559539378,-0.1868476421,0.4078856707,0.1008466557,-0.1221649349,-0.202649951,0.0749394596,-0.0977010354,-0.1044102013,0.2023274899,0.2365569323,-0.3165027499,0.1566904932,-0.3306831717,0.8244410157,0.03512327,-0.2451610118,-0.114677757,0.0505557247,-0.098811157,0.309276849,0.2174607962,0.4579276145,-0.1223986596,0.0678027794,-0.3379885852,-0.0043132771,0.4632381797,-0.6978577971,-0.2602540553,0.3898676038,0.4401413798,0.3855003417,-0.1703182757,-0.686550498,-0.2105173469,0.0655455589,0.0370597355,0.1663676053,0.335321337,-0.0993318483,0.1353846937,0.0923244283,-0.2173095047,0.2460513115,-0.022815764,0.0020172163,-0.2244128436]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/560","title":"Using custom DownloadConfig results in an error","comments":"From my limited understanding, part of the issue seems related to the `prepare_module` and `download_and_prepare` functions each handling the case where no config is passed. For example, `prepare_module` does mutate the object passed and forces the flags `extract_compressed_file` and `force_extract` to `True`.\r\n\r\nSee:\r\n* https:\/\/github.com\/huggingface\/nlp\/blob\/5fb61e1012bda724a9b6b847307d90a1380abfa5\/src\/nlp\/load.py#L227\r\n* https:\/\/github.com\/huggingface\/nlp\/blob\/5fb61e1012bda724a9b6b847307d90a1380abfa5\/src\/nlp\/builder.py#L388\r\n\r\nMaybe a cleaner solution would be to always instantiate a default `DownloadConfig` object at the top-level, have it as non-optional for the lower-level functions and treat it as immutable. ","body":"## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n","comment_length":76,"text":"Using custom DownloadConfig results in an error\n## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n\nFrom my limited understanding, part of the issue seems related to the `prepare_module` and `download_and_prepare` functions each handling the case where no config is passed. For example, `prepare_module` does mutate the object passed and forces the flags `extract_compressed_file` and `force_extract` to `True`.\r\n\r\nSee:\r\n* https:\/\/github.com\/huggingface\/nlp\/blob\/5fb61e1012bda724a9b6b847307d90a1380abfa5\/src\/nlp\/load.py#L227\r\n* https:\/\/github.com\/huggingface\/nlp\/blob\/5fb61e1012bda724a9b6b847307d90a1380abfa5\/src\/nlp\/builder.py#L388\r\n\r\nMaybe a cleaner solution would be to always instantiate a default `DownloadConfig` object at the top-level, have it as non-optional for the lower-level functions and treat it as immutable. ","embeddings":[-0.1150195375,0.1608139873,0.1123033017,0.0204855502,-0.0762844905,-0.0592143387,0.2878563702,0.0386445262,0.3867911994,-0.1992649138,-0.1571436226,0.3636115789,-0.2807902694,-0.0396447405,0.1373905092,0.1345225722,-0.335785538,0.191671595,0.2066785991,-0.0992470011,-0.4401432276,0.4904036224,-0.1363261491,-0.1055808887,-0.1104631573,0.2965768874,-0.0853590146,0.6002389789,-0.0052087544,-0.3369760513,0.5871984363,0.0767923817,0.2937197685,0.0780296773,-0.0001159093,0.0705925748,0.2336821258,-0.3101004958,-0.3799623847,-0.2622923553,-0.0511955321,-0.1243966147,0.2814418674,-0.1946797967,-0.0140370019,-0.0242791511,0.0815578252,0.0334612019,0.0941410288,0.2972890437,0.1419871897,-0.0229717921,-0.2864381075,-0.1697261631,0.2262740731,0.260076493,0.0480878502,0.5292792916,0.3167858422,-0.2207205594,0.0189583451,-0.3049541116,-0.1514441222,0.0622548983,0.5781335831,0.0206799582,0.2644710541,-0.0439114347,-0.1765135378,0.3859302104,0.2035041451,-0.2281819582,-0.0819195062,-0.3655665219,0.144602403,-0.4327120185,0.1780243218,0.1168847531,-0.273757726,0.0803424343,-0.2586563826,-0.1680175215,-0.0763951689,0.5087516308,0.2171331793,0.3228867352,-0.0183474217,0.1714263111,-0.0767275244,-0.0576790571,0.252420336,-0.3967368007,0.0253677983,0.1348518133,-0.0958460644,-0.0770932958,0.1277973652,0.0256882552,-0.0327643231,0.1246374547,0.0682523176,-0.1068671271,0.0644042119,0.0538509563,0.0521950275,0.286323905,-0.076600723,0.1167676002,-0.2036602199,0.1669653356,-0.0508592837,-0.0176646784,0.0861193165,-0.075323768,-0.3973380923,-0.1976338178,0.168750599,0.0014936652,-0.0629578158,-0.339681685,-0.0813496038,0.1825856417,0.338355571,0.2477036715,-0.0689882264,-0.2663639188,0.1952618361,0.2603036761,-0.0893417671,0.1586500257,-0.088268958,-0.1036935523,-0.1506859362,-0.0623486452,0.4825169444,0.2577634156,0.6017351151,0.0311783329,0.0021425504,0.0281839706,-0.0155198788,0.1216142625,0.0396912619,0.1164999828,0.004311326,0.0528843515,0.0625880286,-0.2535326481,-0.2557506263,-0.0347330607,-0.1973161101,-0.5240719318,-0.32106933,0.1228109673,-0.3273592293,-0.3346843123,0.2478318959,-0.0070183282,0.060555689,-0.3129452467,-0.0337526128,0.029548144,-0.4143957198,-0.2400684655,-0.089732632,0.3797775209,0.0012102607,-0.2380746603,-0.4831041098,-0.1488144845,0.2068390101,-0.0621503629,-0.4448852539,0.2460324764,-0.3705919683,0.1259263307,0.7066502571,-0.593824029,-0.475862354,0.7060877681,-0.1530996859,0.1541841626,0.1148285046,0.3047769666,0.3644037843,-0.1938655227,0.2743984759,0.449900806,0.1645402014,-0.0376653224,-0.1342462301,-0.2366191745,-0.0452952273,0.069085665,-0.2126618922,0.3807032704,0.0322402194,0.228694275,0.2805713415,0.0719816536,-0.0035698258,0.0699649006,-0.0190699659,0.2078530788,-0.216038838,0.2683116198,-0.5918141603,0.3081945777,-0.2088387758,0.0111979246,0.1865449697,0.0156147871,-0.3768417537,-0.2749233842,-0.1442956775,-0.391390115,0.0948154703,0.3689242601,0.1382951736,-0.1243658364,-0.1281298399,0.3340533972,-0.2191484272,-0.0176077504,-0.1604177207,-0.0221684948,0.1522677392,-0.1637913138,0.0569932424,0.105084911,0.2022019923,0.1713188738,-0.0114865042,0.4094794691,-0.1540582776,0.0669990554,-0.08875563,-0.2120270282,0.0864794329,-0.0025839424,0.2240317017,0.0819461644,0.1684664637,-0.227870211,-0.0677169561,0.1494398415,0.0992672071,0.3279950321,0.0210084673,0.1423203647,0.1164696813,-0.2840330005,-0.0270047877,-0.34892869,0.0761514902,0.1613445729,0.3522972167,0.0295899063,-0.2389683723,-0.0452459715,0.2232688069,0.1981347203,0.0204542782,0.1191779003,-0.0028204762,-0.0736611784,0.029122578,0.6464995742,0.816970706,0.110755451,-0.0627681091,0.4516937435,-0.2712000906,-0.2540049255,0.279292047,-0.1479594558,0.1289023012,0.1667049378,0.0079044867,0.0080792001,-0.1439014524,-0.4434364438,-0.0547379367,-0.0084888637,-0.4380910993,-0.0696294531,-0.2960466444,-0.4512906969,-0.1279855371,-0.2876825333,-0.0485401414,-0.2141084671,-0.159273237,0.1131819934,0.1047847271,-0.0083428686,-0.2631918788,0.019749023,-0.2949643731,-0.8628898859,0.0819824636,-0.0314406157,-0.3450305164,-0.0835407823,0.0386174619,0.1652457118,0.073131986,-0.1557795554,-0.2440009415,0.1602346897,-0.1029197425,0.1122968048,0.1613503247,0.3075020313,0.3883806169,-0.1511534303,0.2565552294,-0.4360593557,0.1637900174,0.2162101418,-0.0127829853,0.0424747802,0.0326359421,-0.0875902697,-0.368976295,-0.2115421742,-0.1649634391,-0.450252533,-0.0885402486,0.28491202,0.1961458027,0.4581585228,0.1651855111,0.3243179917,-0.1645777971,0.0822794884,-0.0674815327,-0.3117092252,0.2791596651,-0.4733329117,-0.0789996907,0.085931167,-0.2227194011,0.2154388875,0.1335568577,-0.3884005845,0.199653998,-0.1469651461,0.0002483857,-0.0470809862,-0.0934197456,0.2494487315,-0.082257472,0.0479591452,-0.2609181702,-0.3588173091,0.3365956545,0.3105373681,0.5375539064,0.275968343,0.2986249626,-0.1452536583,0.4846876264,0.1508263201,-0.1560512483,0.6103883386,-0.1947124898,0.4601784647,-0.0382337458,-0.2387869209,-0.0032227396,0.3633525074,-0.0172826368,0.2315363288,-0.0611758232,-0.3007305861,-0.1902965009,-0.1062506512,-0.1545367837,-0.230031535,-0.0597753413,0.0295013599,0.1523663849,0.0646302924,0.1584214121,-0.083147265,-0.1170615926,-0.187199682,0.435128808,-0.0597942136,0.2172851712,-0.0276041552,0.2997868061,-0.3493392169,0.1421994418,-0.2189187407,0.3044458926,-0.1182061136,0.0613894761,-0.0159715731,0.116941765,0.2156404406,-0.2203439325,0.2306120545,0.262530148,0.1650993526,-0.6942793131,-0.0943868831,0.0507083125,0.3705742955,0.6355696321,0.1778835207,-0.1156038195,-0.2360606492,0.1376715899,-0.1754219085,-0.0924998894,-0.4230398536,-0.066459544,-0.3104561269,-0.2801861167,-0.3828448951,-0.1292264163,0.230558604,0.2301347256,0.1135466546,0.0101915412,0.0370249003,0.0223113801,-0.0750334486,0.0257843025,0.1655019671,0.2535585463,-0.3498473167,0.0041104979,0.0184270665,0.455697149,-0.1090064123,-0.1824812293,-0.0907652378,0.0760888234,0.1698767692,0.2375928909,-0.0469807722,-0.2788476646,-0.0843373016,-0.2431958616,0.061178796,0.1471870542,0.2561111152,-0.5553205013,0.181288138,-0.3989014328,0.5427388549,-0.1028674021,-0.0532116704,0.1602564305,-0.0507060289,-0.3970303535,0.324452132,-0.0401610807,0.5586260557,0.2404643893,0.0567302257,0.412304014,-0.433357656,0.3407605588,0.3820564151,0.0827065259,-0.2056452781,0.2356447577,0.1081556752,-0.2434303612,0.1182638481,0.4598899782,-0.1859720945,0.3259095848,-0.1559953988,-0.4174781144,0.0909639597,0.4379138649,-0.2495400608,-0.3017905354,-0.3178763986,0.1256023496,-0.0598930605,0.3444620967,-0.0450952463,-0.313447684,-0.3213606477,-0.1474078596,-0.2323348522,0.1951860338,-0.0521365926,-0.1184979379,0.1656085402,0.0278660897,0.1757669449,0.3797243237,-0.2905163169,-0.0837098733,-0.2389895022,0.0830149725,-0.1039216667,-0.0695943534,0.1519480944,0.0730860904,0.2014408559,-0.1141102687,-0.3429861069,-0.1494403929,-0.0445597135,-0.2015192062,0.4617881179,-0.0481094755,0.0759342238,-0.1242546812,-0.4921830595,0.0628540292,-0.0180091802,0.1029136181,0.0785280094,0.0157993361,-0.0287095681,0.1853824854,-0.1957309395,-0.337682426,-0.0281161107,0.3746060133,-0.3746849597,0.2907844186,0.4278508723,-0.1933617741,0.1264047176,-0.0584542118,0.0083092405,-0.5848215222,-0.216162622,0.1031734273,0.2297327518,0.164756462,0.1329749525,0.2493012846,-0.003495506,0.1802944839,0.2565452456,-0.3308797181,-0.1663021296,-0.0038664439,-0.3690316081,0.1466334462,-0.1980221123,0.2377348244,0.3796907067,-0.2016597241,-0.2258139551,0.1875278056,-0.1729793102,0.0496794544,0.1979008466,0.1081392094,0.0217493102,0.1827962399,0.0240138508,-0.062509872,-0.2874262929,-0.1339973062,-0.3909825087,0.1758333296,0.2264897525,-0.1660330445,-0.0547830835,-0.2572702467,-0.2181168497,-0.031453073,-0.0373602435,0.1865928322,-0.0764180124,0.3890280724,-0.1978252083,0.243919909,-0.0793271363,0.1648311615,-0.1698613018,0.3258364797,0.0379703566,0.1343825907,-0.1742454469,0.0088649197,-0.1068822294,-0.231389761,-0.2788546085,0.0202061161,0.2709775567,-0.0176596604,0.2543975413,0.2019496262,-0.0415051319,0.3494026363,-0.2206249982,-0.1269973963,0.2625512481,0.1083602384,-0.1582949609,0.2222321928,0.3854727447,0.3658417761,0.0349594019,0.2504579127,0.1602328718,-0.3160717785,-0.0662192702,-0.045041766,0.2565685511,0.0481726713,-0.0075777015,0.3368707597,-0.2633071244,0.200918898,0.0543518886,0.2817530036,0.2158435732,0.4725821316,-0.0022776665,-0.1878003627,-0.2764528692,0.0665216371,0.1926874369,-0.5811086893,0.1973800063,-0.0308278725,0.1413285732,0.2872923315,-0.1325430572,0.0486670956,-0.1231485605,0.4628397524,-0.0270578172,0.2395097613,0.0963629335,-0.1682436913,-0.1834777743,0.043465618,-0.0890507922,0.0064130211,-0.1697317511,-0.2001911998,-0.0572066903,0.0989688635,-0.2617083192,-0.2644017339,0.2126554251,-0.0461103767,-0.0920265764,-0.28864941,0.1431363076,-0.0127508091,0.1704822779,0.3351086974,0.2608134151,0.2379633188,0.5772158504,-0.0407342091,-0.1141022816,0.1889009327,0.0600242428,-0.2130747139,0.0773516297,0.3676773608,0.3506335914,0.3967398107,0.0209638998,-0.0882208273,-0.0021246681,0.2406388372,0.2035897821,-0.1885028034,0.7938399911,0.0524095446,0.1809535921,-0.116335772,0.3631625772,-0.4275179803,-0.2053982168,0.287856102,0.2680921257,0.3093080819,-0.0219096225,0.0199616514,-0.1861562878,0.5015246272,0.0297628138,-0.1064067036,-0.1711036861,-0.0851538107,-0.5885362029,-0.109198913,-0.0785844773,-0.3384315372,0.0276821163,0.4448843002,-0.272295475,0.0116580725,-0.1669053137,0.1294880956,0.0158839952,0.1184924766,0.0914537683,-0.176231131,-0.2772181332,0.058955051,0.2688049078,-0.3439932764,-0.2598087788,-0.4583736658,0.028446652,-0.0672117546,-0.1894518435,0.2401178181,0.2790070474,0.4852349758,-0.0436648764,0.3151144683,-0.0392435901,-0.2778841257,-0.396420747,0.1328745633,-0.1942772418,-0.0794956088,-0.001046693,0.1259427369,-0.327412039,0.3908373713,-0.2904158235,0.1902526468,0.054690212,-0.145225063,-0.0034682259,0.1812205017,-0.3123649955,0.1539282352,0.2389645278,0.3021642864,-0.1769243479,0.2898986638,-0.2135694921,-0.2442600131,0.5382840633,-0.1240976006,-0.0283331275,0.1594706923,0.2759436369,-0.1947040707,-0.1216937676,-0.2669958472,0.0924191847,0.4286957085,-0.2330366075,0.1069559753,-0.0202895533,-0.2478488237,0.028868109,0.0702999756,0.069040589,0.119177945,-0.1778608859,-0.2479636669,-0.2421284765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/560","title":"Using custom DownloadConfig results in an error","comments":"Thanks for the report, I'll take a look.\r\n\r\nWhat is your specific use-case for providing a DownloadConfig object?\r\n","body":"## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n","comment_length":18,"text":"Using custom DownloadConfig results in an error\n## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n\nThanks for the report, I'll take a look.\r\n\r\nWhat is your specific use-case for providing a DownloadConfig object?\r\n","embeddings":[-0.1150195375,0.1608139873,0.1123033017,0.0204855502,-0.0762844905,-0.0592143387,0.2878563702,0.0386445262,0.3867911994,-0.1992649138,-0.1571436226,0.3636115789,-0.2807902694,-0.0396447405,0.1373905092,0.1345225722,-0.335785538,0.191671595,0.2066785991,-0.0992470011,-0.4401432276,0.4904036224,-0.1363261491,-0.1055808887,-0.1104631573,0.2965768874,-0.0853590146,0.6002389789,-0.0052087544,-0.3369760513,0.5871984363,0.0767923817,0.2937197685,0.0780296773,-0.0001159093,0.0705925748,0.2336821258,-0.3101004958,-0.3799623847,-0.2622923553,-0.0511955321,-0.1243966147,0.2814418674,-0.1946797967,-0.0140370019,-0.0242791511,0.0815578252,0.0334612019,0.0941410288,0.2972890437,0.1419871897,-0.0229717921,-0.2864381075,-0.1697261631,0.2262740731,0.260076493,0.0480878502,0.5292792916,0.3167858422,-0.2207205594,0.0189583451,-0.3049541116,-0.1514441222,0.0622548983,0.5781335831,0.0206799582,0.2644710541,-0.0439114347,-0.1765135378,0.3859302104,0.2035041451,-0.2281819582,-0.0819195062,-0.3655665219,0.144602403,-0.4327120185,0.1780243218,0.1168847531,-0.273757726,0.0803424343,-0.2586563826,-0.1680175215,-0.0763951689,0.5087516308,0.2171331793,0.3228867352,-0.0183474217,0.1714263111,-0.0767275244,-0.0576790571,0.252420336,-0.3967368007,0.0253677983,0.1348518133,-0.0958460644,-0.0770932958,0.1277973652,0.0256882552,-0.0327643231,0.1246374547,0.0682523176,-0.1068671271,0.0644042119,0.0538509563,0.0521950275,0.286323905,-0.076600723,0.1167676002,-0.2036602199,0.1669653356,-0.0508592837,-0.0176646784,0.0861193165,-0.075323768,-0.3973380923,-0.1976338178,0.168750599,0.0014936652,-0.0629578158,-0.339681685,-0.0813496038,0.1825856417,0.338355571,0.2477036715,-0.0689882264,-0.2663639188,0.1952618361,0.2603036761,-0.0893417671,0.1586500257,-0.088268958,-0.1036935523,-0.1506859362,-0.0623486452,0.4825169444,0.2577634156,0.6017351151,0.0311783329,0.0021425504,0.0281839706,-0.0155198788,0.1216142625,0.0396912619,0.1164999828,0.004311326,0.0528843515,0.0625880286,-0.2535326481,-0.2557506263,-0.0347330607,-0.1973161101,-0.5240719318,-0.32106933,0.1228109673,-0.3273592293,-0.3346843123,0.2478318959,-0.0070183282,0.060555689,-0.3129452467,-0.0337526128,0.029548144,-0.4143957198,-0.2400684655,-0.089732632,0.3797775209,0.0012102607,-0.2380746603,-0.4831041098,-0.1488144845,0.2068390101,-0.0621503629,-0.4448852539,0.2460324764,-0.3705919683,0.1259263307,0.7066502571,-0.593824029,-0.475862354,0.7060877681,-0.1530996859,0.1541841626,0.1148285046,0.3047769666,0.3644037843,-0.1938655227,0.2743984759,0.449900806,0.1645402014,-0.0376653224,-0.1342462301,-0.2366191745,-0.0452952273,0.069085665,-0.2126618922,0.3807032704,0.0322402194,0.228694275,0.2805713415,0.0719816536,-0.0035698258,0.0699649006,-0.0190699659,0.2078530788,-0.216038838,0.2683116198,-0.5918141603,0.3081945777,-0.2088387758,0.0111979246,0.1865449697,0.0156147871,-0.3768417537,-0.2749233842,-0.1442956775,-0.391390115,0.0948154703,0.3689242601,0.1382951736,-0.1243658364,-0.1281298399,0.3340533972,-0.2191484272,-0.0176077504,-0.1604177207,-0.0221684948,0.1522677392,-0.1637913138,0.0569932424,0.105084911,0.2022019923,0.1713188738,-0.0114865042,0.4094794691,-0.1540582776,0.0669990554,-0.08875563,-0.2120270282,0.0864794329,-0.0025839424,0.2240317017,0.0819461644,0.1684664637,-0.227870211,-0.0677169561,0.1494398415,0.0992672071,0.3279950321,0.0210084673,0.1423203647,0.1164696813,-0.2840330005,-0.0270047877,-0.34892869,0.0761514902,0.1613445729,0.3522972167,0.0295899063,-0.2389683723,-0.0452459715,0.2232688069,0.1981347203,0.0204542782,0.1191779003,-0.0028204762,-0.0736611784,0.029122578,0.6464995742,0.816970706,0.110755451,-0.0627681091,0.4516937435,-0.2712000906,-0.2540049255,0.279292047,-0.1479594558,0.1289023012,0.1667049378,0.0079044867,0.0080792001,-0.1439014524,-0.4434364438,-0.0547379367,-0.0084888637,-0.4380910993,-0.0696294531,-0.2960466444,-0.4512906969,-0.1279855371,-0.2876825333,-0.0485401414,-0.2141084671,-0.159273237,0.1131819934,0.1047847271,-0.0083428686,-0.2631918788,0.019749023,-0.2949643731,-0.8628898859,0.0819824636,-0.0314406157,-0.3450305164,-0.0835407823,0.0386174619,0.1652457118,0.073131986,-0.1557795554,-0.2440009415,0.1602346897,-0.1029197425,0.1122968048,0.1613503247,0.3075020313,0.3883806169,-0.1511534303,0.2565552294,-0.4360593557,0.1637900174,0.2162101418,-0.0127829853,0.0424747802,0.0326359421,-0.0875902697,-0.368976295,-0.2115421742,-0.1649634391,-0.450252533,-0.0885402486,0.28491202,0.1961458027,0.4581585228,0.1651855111,0.3243179917,-0.1645777971,0.0822794884,-0.0674815327,-0.3117092252,0.2791596651,-0.4733329117,-0.0789996907,0.085931167,-0.2227194011,0.2154388875,0.1335568577,-0.3884005845,0.199653998,-0.1469651461,0.0002483857,-0.0470809862,-0.0934197456,0.2494487315,-0.082257472,0.0479591452,-0.2609181702,-0.3588173091,0.3365956545,0.3105373681,0.5375539064,0.275968343,0.2986249626,-0.1452536583,0.4846876264,0.1508263201,-0.1560512483,0.6103883386,-0.1947124898,0.4601784647,-0.0382337458,-0.2387869209,-0.0032227396,0.3633525074,-0.0172826368,0.2315363288,-0.0611758232,-0.3007305861,-0.1902965009,-0.1062506512,-0.1545367837,-0.230031535,-0.0597753413,0.0295013599,0.1523663849,0.0646302924,0.1584214121,-0.083147265,-0.1170615926,-0.187199682,0.435128808,-0.0597942136,0.2172851712,-0.0276041552,0.2997868061,-0.3493392169,0.1421994418,-0.2189187407,0.3044458926,-0.1182061136,0.0613894761,-0.0159715731,0.116941765,0.2156404406,-0.2203439325,0.2306120545,0.262530148,0.1650993526,-0.6942793131,-0.0943868831,0.0507083125,0.3705742955,0.6355696321,0.1778835207,-0.1156038195,-0.2360606492,0.1376715899,-0.1754219085,-0.0924998894,-0.4230398536,-0.066459544,-0.3104561269,-0.2801861167,-0.3828448951,-0.1292264163,0.230558604,0.2301347256,0.1135466546,0.0101915412,0.0370249003,0.0223113801,-0.0750334486,0.0257843025,0.1655019671,0.2535585463,-0.3498473167,0.0041104979,0.0184270665,0.455697149,-0.1090064123,-0.1824812293,-0.0907652378,0.0760888234,0.1698767692,0.2375928909,-0.0469807722,-0.2788476646,-0.0843373016,-0.2431958616,0.061178796,0.1471870542,0.2561111152,-0.5553205013,0.181288138,-0.3989014328,0.5427388549,-0.1028674021,-0.0532116704,0.1602564305,-0.0507060289,-0.3970303535,0.324452132,-0.0401610807,0.5586260557,0.2404643893,0.0567302257,0.412304014,-0.433357656,0.3407605588,0.3820564151,0.0827065259,-0.2056452781,0.2356447577,0.1081556752,-0.2434303612,0.1182638481,0.4598899782,-0.1859720945,0.3259095848,-0.1559953988,-0.4174781144,0.0909639597,0.4379138649,-0.2495400608,-0.3017905354,-0.3178763986,0.1256023496,-0.0598930605,0.3444620967,-0.0450952463,-0.313447684,-0.3213606477,-0.1474078596,-0.2323348522,0.1951860338,-0.0521365926,-0.1184979379,0.1656085402,0.0278660897,0.1757669449,0.3797243237,-0.2905163169,-0.0837098733,-0.2389895022,0.0830149725,-0.1039216667,-0.0695943534,0.1519480944,0.0730860904,0.2014408559,-0.1141102687,-0.3429861069,-0.1494403929,-0.0445597135,-0.2015192062,0.4617881179,-0.0481094755,0.0759342238,-0.1242546812,-0.4921830595,0.0628540292,-0.0180091802,0.1029136181,0.0785280094,0.0157993361,-0.0287095681,0.1853824854,-0.1957309395,-0.337682426,-0.0281161107,0.3746060133,-0.3746849597,0.2907844186,0.4278508723,-0.1933617741,0.1264047176,-0.0584542118,0.0083092405,-0.5848215222,-0.216162622,0.1031734273,0.2297327518,0.164756462,0.1329749525,0.2493012846,-0.003495506,0.1802944839,0.2565452456,-0.3308797181,-0.1663021296,-0.0038664439,-0.3690316081,0.1466334462,-0.1980221123,0.2377348244,0.3796907067,-0.2016597241,-0.2258139551,0.1875278056,-0.1729793102,0.0496794544,0.1979008466,0.1081392094,0.0217493102,0.1827962399,0.0240138508,-0.062509872,-0.2874262929,-0.1339973062,-0.3909825087,0.1758333296,0.2264897525,-0.1660330445,-0.0547830835,-0.2572702467,-0.2181168497,-0.031453073,-0.0373602435,0.1865928322,-0.0764180124,0.3890280724,-0.1978252083,0.243919909,-0.0793271363,0.1648311615,-0.1698613018,0.3258364797,0.0379703566,0.1343825907,-0.1742454469,0.0088649197,-0.1068822294,-0.231389761,-0.2788546085,0.0202061161,0.2709775567,-0.0176596604,0.2543975413,0.2019496262,-0.0415051319,0.3494026363,-0.2206249982,-0.1269973963,0.2625512481,0.1083602384,-0.1582949609,0.2222321928,0.3854727447,0.3658417761,0.0349594019,0.2504579127,0.1602328718,-0.3160717785,-0.0662192702,-0.045041766,0.2565685511,0.0481726713,-0.0075777015,0.3368707597,-0.2633071244,0.200918898,0.0543518886,0.2817530036,0.2158435732,0.4725821316,-0.0022776665,-0.1878003627,-0.2764528692,0.0665216371,0.1926874369,-0.5811086893,0.1973800063,-0.0308278725,0.1413285732,0.2872923315,-0.1325430572,0.0486670956,-0.1231485605,0.4628397524,-0.0270578172,0.2395097613,0.0963629335,-0.1682436913,-0.1834777743,0.043465618,-0.0890507922,0.0064130211,-0.1697317511,-0.2001911998,-0.0572066903,0.0989688635,-0.2617083192,-0.2644017339,0.2126554251,-0.0461103767,-0.0920265764,-0.28864941,0.1431363076,-0.0127508091,0.1704822779,0.3351086974,0.2608134151,0.2379633188,0.5772158504,-0.0407342091,-0.1141022816,0.1889009327,0.0600242428,-0.2130747139,0.0773516297,0.3676773608,0.3506335914,0.3967398107,0.0209638998,-0.0882208273,-0.0021246681,0.2406388372,0.2035897821,-0.1885028034,0.7938399911,0.0524095446,0.1809535921,-0.116335772,0.3631625772,-0.4275179803,-0.2053982168,0.287856102,0.2680921257,0.3093080819,-0.0219096225,0.0199616514,-0.1861562878,0.5015246272,0.0297628138,-0.1064067036,-0.1711036861,-0.0851538107,-0.5885362029,-0.109198913,-0.0785844773,-0.3384315372,0.0276821163,0.4448843002,-0.272295475,0.0116580725,-0.1669053137,0.1294880956,0.0158839952,0.1184924766,0.0914537683,-0.176231131,-0.2772181332,0.058955051,0.2688049078,-0.3439932764,-0.2598087788,-0.4583736658,0.028446652,-0.0672117546,-0.1894518435,0.2401178181,0.2790070474,0.4852349758,-0.0436648764,0.3151144683,-0.0392435901,-0.2778841257,-0.396420747,0.1328745633,-0.1942772418,-0.0794956088,-0.001046693,0.1259427369,-0.327412039,0.3908373713,-0.2904158235,0.1902526468,0.054690212,-0.145225063,-0.0034682259,0.1812205017,-0.3123649955,0.1539282352,0.2389645278,0.3021642864,-0.1769243479,0.2898986638,-0.2135694921,-0.2442600131,0.5382840633,-0.1240976006,-0.0283331275,0.1594706923,0.2759436369,-0.1947040707,-0.1216937676,-0.2669958472,0.0924191847,0.4286957085,-0.2330366075,0.1069559753,-0.0202895533,-0.2478488237,0.028868109,0.0702999756,0.069040589,0.119177945,-0.1778608859,-0.2479636669,-0.2421284765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/560","title":"Using custom DownloadConfig results in an error","comments":"Thanks. Our use case involves running a training job behind a corporate firewall with no access to any external resources (S3, GCP or other web resources).\r\n\r\nI was thinking about a 2-steps process:\r\n1) Download the resources \/ artifacts using some secure corporate channel, ie run `nlp.load_dataset()` without a specific `DownloadConfig`. After that, collect the files from the `$HF_HOME` folder\r\n2) Copy the `$HF_HOME` folder in the firewalled environment. Run `nlp.load_dataset()` with a custom config `DownloadConfig(local_files_only=True)`\r\n\r\nHowever this ends up a bit clunky in practice, even when solving the `DownloadConfig` issue above. For example, the `filename` hash computed in `get_from_cache()` differs in the `local_files_only=False` vs `local_files_only=True` case (local case defaults `etag` to `None`, which results in a different hash). So effectively step 2) above doesn't work because the hash computed differs from the hash in the cache folder. Some hacks \/ workaround are possible but this solution becomes very convoluted.\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/c214aa5a4430c1df1bcd0619fd94d6abdf9d2da7\/src\/nlp\/utils\/file_utils.py#L417\r\n\r\nWould you recommend a different path?\r\n","body":"## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n","comment_length":157,"text":"Using custom DownloadConfig results in an error\n## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n\nThanks. Our use case involves running a training job behind a corporate firewall with no access to any external resources (S3, GCP or other web resources).\r\n\r\nI was thinking about a 2-steps process:\r\n1) Download the resources \/ artifacts using some secure corporate channel, ie run `nlp.load_dataset()` without a specific `DownloadConfig`. After that, collect the files from the `$HF_HOME` folder\r\n2) Copy the `$HF_HOME` folder in the firewalled environment. Run `nlp.load_dataset()` with a custom config `DownloadConfig(local_files_only=True)`\r\n\r\nHowever this ends up a bit clunky in practice, even when solving the `DownloadConfig` issue above. For example, the `filename` hash computed in `get_from_cache()` differs in the `local_files_only=False` vs `local_files_only=True` case (local case defaults `etag` to `None`, which results in a different hash). So effectively step 2) above doesn't work because the hash computed differs from the hash in the cache folder. Some hacks \/ workaround are possible but this solution becomes very convoluted.\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/c214aa5a4430c1df1bcd0619fd94d6abdf9d2da7\/src\/nlp\/utils\/file_utils.py#L417\r\n\r\nWould you recommend a different path?\r\n","embeddings":[-0.1150195375,0.1608139873,0.1123033017,0.0204855502,-0.0762844905,-0.0592143387,0.2878563702,0.0386445262,0.3867911994,-0.1992649138,-0.1571436226,0.3636115789,-0.2807902694,-0.0396447405,0.1373905092,0.1345225722,-0.335785538,0.191671595,0.2066785991,-0.0992470011,-0.4401432276,0.4904036224,-0.1363261491,-0.1055808887,-0.1104631573,0.2965768874,-0.0853590146,0.6002389789,-0.0052087544,-0.3369760513,0.5871984363,0.0767923817,0.2937197685,0.0780296773,-0.0001159093,0.0705925748,0.2336821258,-0.3101004958,-0.3799623847,-0.2622923553,-0.0511955321,-0.1243966147,0.2814418674,-0.1946797967,-0.0140370019,-0.0242791511,0.0815578252,0.0334612019,0.0941410288,0.2972890437,0.1419871897,-0.0229717921,-0.2864381075,-0.1697261631,0.2262740731,0.260076493,0.0480878502,0.5292792916,0.3167858422,-0.2207205594,0.0189583451,-0.3049541116,-0.1514441222,0.0622548983,0.5781335831,0.0206799582,0.2644710541,-0.0439114347,-0.1765135378,0.3859302104,0.2035041451,-0.2281819582,-0.0819195062,-0.3655665219,0.144602403,-0.4327120185,0.1780243218,0.1168847531,-0.273757726,0.0803424343,-0.2586563826,-0.1680175215,-0.0763951689,0.5087516308,0.2171331793,0.3228867352,-0.0183474217,0.1714263111,-0.0767275244,-0.0576790571,0.252420336,-0.3967368007,0.0253677983,0.1348518133,-0.0958460644,-0.0770932958,0.1277973652,0.0256882552,-0.0327643231,0.1246374547,0.0682523176,-0.1068671271,0.0644042119,0.0538509563,0.0521950275,0.286323905,-0.076600723,0.1167676002,-0.2036602199,0.1669653356,-0.0508592837,-0.0176646784,0.0861193165,-0.075323768,-0.3973380923,-0.1976338178,0.168750599,0.0014936652,-0.0629578158,-0.339681685,-0.0813496038,0.1825856417,0.338355571,0.2477036715,-0.0689882264,-0.2663639188,0.1952618361,0.2603036761,-0.0893417671,0.1586500257,-0.088268958,-0.1036935523,-0.1506859362,-0.0623486452,0.4825169444,0.2577634156,0.6017351151,0.0311783329,0.0021425504,0.0281839706,-0.0155198788,0.1216142625,0.0396912619,0.1164999828,0.004311326,0.0528843515,0.0625880286,-0.2535326481,-0.2557506263,-0.0347330607,-0.1973161101,-0.5240719318,-0.32106933,0.1228109673,-0.3273592293,-0.3346843123,0.2478318959,-0.0070183282,0.060555689,-0.3129452467,-0.0337526128,0.029548144,-0.4143957198,-0.2400684655,-0.089732632,0.3797775209,0.0012102607,-0.2380746603,-0.4831041098,-0.1488144845,0.2068390101,-0.0621503629,-0.4448852539,0.2460324764,-0.3705919683,0.1259263307,0.7066502571,-0.593824029,-0.475862354,0.7060877681,-0.1530996859,0.1541841626,0.1148285046,0.3047769666,0.3644037843,-0.1938655227,0.2743984759,0.449900806,0.1645402014,-0.0376653224,-0.1342462301,-0.2366191745,-0.0452952273,0.069085665,-0.2126618922,0.3807032704,0.0322402194,0.228694275,0.2805713415,0.0719816536,-0.0035698258,0.0699649006,-0.0190699659,0.2078530788,-0.216038838,0.2683116198,-0.5918141603,0.3081945777,-0.2088387758,0.0111979246,0.1865449697,0.0156147871,-0.3768417537,-0.2749233842,-0.1442956775,-0.391390115,0.0948154703,0.3689242601,0.1382951736,-0.1243658364,-0.1281298399,0.3340533972,-0.2191484272,-0.0176077504,-0.1604177207,-0.0221684948,0.1522677392,-0.1637913138,0.0569932424,0.105084911,0.2022019923,0.1713188738,-0.0114865042,0.4094794691,-0.1540582776,0.0669990554,-0.08875563,-0.2120270282,0.0864794329,-0.0025839424,0.2240317017,0.0819461644,0.1684664637,-0.227870211,-0.0677169561,0.1494398415,0.0992672071,0.3279950321,0.0210084673,0.1423203647,0.1164696813,-0.2840330005,-0.0270047877,-0.34892869,0.0761514902,0.1613445729,0.3522972167,0.0295899063,-0.2389683723,-0.0452459715,0.2232688069,0.1981347203,0.0204542782,0.1191779003,-0.0028204762,-0.0736611784,0.029122578,0.6464995742,0.816970706,0.110755451,-0.0627681091,0.4516937435,-0.2712000906,-0.2540049255,0.279292047,-0.1479594558,0.1289023012,0.1667049378,0.0079044867,0.0080792001,-0.1439014524,-0.4434364438,-0.0547379367,-0.0084888637,-0.4380910993,-0.0696294531,-0.2960466444,-0.4512906969,-0.1279855371,-0.2876825333,-0.0485401414,-0.2141084671,-0.159273237,0.1131819934,0.1047847271,-0.0083428686,-0.2631918788,0.019749023,-0.2949643731,-0.8628898859,0.0819824636,-0.0314406157,-0.3450305164,-0.0835407823,0.0386174619,0.1652457118,0.073131986,-0.1557795554,-0.2440009415,0.1602346897,-0.1029197425,0.1122968048,0.1613503247,0.3075020313,0.3883806169,-0.1511534303,0.2565552294,-0.4360593557,0.1637900174,0.2162101418,-0.0127829853,0.0424747802,0.0326359421,-0.0875902697,-0.368976295,-0.2115421742,-0.1649634391,-0.450252533,-0.0885402486,0.28491202,0.1961458027,0.4581585228,0.1651855111,0.3243179917,-0.1645777971,0.0822794884,-0.0674815327,-0.3117092252,0.2791596651,-0.4733329117,-0.0789996907,0.085931167,-0.2227194011,0.2154388875,0.1335568577,-0.3884005845,0.199653998,-0.1469651461,0.0002483857,-0.0470809862,-0.0934197456,0.2494487315,-0.082257472,0.0479591452,-0.2609181702,-0.3588173091,0.3365956545,0.3105373681,0.5375539064,0.275968343,0.2986249626,-0.1452536583,0.4846876264,0.1508263201,-0.1560512483,0.6103883386,-0.1947124898,0.4601784647,-0.0382337458,-0.2387869209,-0.0032227396,0.3633525074,-0.0172826368,0.2315363288,-0.0611758232,-0.3007305861,-0.1902965009,-0.1062506512,-0.1545367837,-0.230031535,-0.0597753413,0.0295013599,0.1523663849,0.0646302924,0.1584214121,-0.083147265,-0.1170615926,-0.187199682,0.435128808,-0.0597942136,0.2172851712,-0.0276041552,0.2997868061,-0.3493392169,0.1421994418,-0.2189187407,0.3044458926,-0.1182061136,0.0613894761,-0.0159715731,0.116941765,0.2156404406,-0.2203439325,0.2306120545,0.262530148,0.1650993526,-0.6942793131,-0.0943868831,0.0507083125,0.3705742955,0.6355696321,0.1778835207,-0.1156038195,-0.2360606492,0.1376715899,-0.1754219085,-0.0924998894,-0.4230398536,-0.066459544,-0.3104561269,-0.2801861167,-0.3828448951,-0.1292264163,0.230558604,0.2301347256,0.1135466546,0.0101915412,0.0370249003,0.0223113801,-0.0750334486,0.0257843025,0.1655019671,0.2535585463,-0.3498473167,0.0041104979,0.0184270665,0.455697149,-0.1090064123,-0.1824812293,-0.0907652378,0.0760888234,0.1698767692,0.2375928909,-0.0469807722,-0.2788476646,-0.0843373016,-0.2431958616,0.061178796,0.1471870542,0.2561111152,-0.5553205013,0.181288138,-0.3989014328,0.5427388549,-0.1028674021,-0.0532116704,0.1602564305,-0.0507060289,-0.3970303535,0.324452132,-0.0401610807,0.5586260557,0.2404643893,0.0567302257,0.412304014,-0.433357656,0.3407605588,0.3820564151,0.0827065259,-0.2056452781,0.2356447577,0.1081556752,-0.2434303612,0.1182638481,0.4598899782,-0.1859720945,0.3259095848,-0.1559953988,-0.4174781144,0.0909639597,0.4379138649,-0.2495400608,-0.3017905354,-0.3178763986,0.1256023496,-0.0598930605,0.3444620967,-0.0450952463,-0.313447684,-0.3213606477,-0.1474078596,-0.2323348522,0.1951860338,-0.0521365926,-0.1184979379,0.1656085402,0.0278660897,0.1757669449,0.3797243237,-0.2905163169,-0.0837098733,-0.2389895022,0.0830149725,-0.1039216667,-0.0695943534,0.1519480944,0.0730860904,0.2014408559,-0.1141102687,-0.3429861069,-0.1494403929,-0.0445597135,-0.2015192062,0.4617881179,-0.0481094755,0.0759342238,-0.1242546812,-0.4921830595,0.0628540292,-0.0180091802,0.1029136181,0.0785280094,0.0157993361,-0.0287095681,0.1853824854,-0.1957309395,-0.337682426,-0.0281161107,0.3746060133,-0.3746849597,0.2907844186,0.4278508723,-0.1933617741,0.1264047176,-0.0584542118,0.0083092405,-0.5848215222,-0.216162622,0.1031734273,0.2297327518,0.164756462,0.1329749525,0.2493012846,-0.003495506,0.1802944839,0.2565452456,-0.3308797181,-0.1663021296,-0.0038664439,-0.3690316081,0.1466334462,-0.1980221123,0.2377348244,0.3796907067,-0.2016597241,-0.2258139551,0.1875278056,-0.1729793102,0.0496794544,0.1979008466,0.1081392094,0.0217493102,0.1827962399,0.0240138508,-0.062509872,-0.2874262929,-0.1339973062,-0.3909825087,0.1758333296,0.2264897525,-0.1660330445,-0.0547830835,-0.2572702467,-0.2181168497,-0.031453073,-0.0373602435,0.1865928322,-0.0764180124,0.3890280724,-0.1978252083,0.243919909,-0.0793271363,0.1648311615,-0.1698613018,0.3258364797,0.0379703566,0.1343825907,-0.1742454469,0.0088649197,-0.1068822294,-0.231389761,-0.2788546085,0.0202061161,0.2709775567,-0.0176596604,0.2543975413,0.2019496262,-0.0415051319,0.3494026363,-0.2206249982,-0.1269973963,0.2625512481,0.1083602384,-0.1582949609,0.2222321928,0.3854727447,0.3658417761,0.0349594019,0.2504579127,0.1602328718,-0.3160717785,-0.0662192702,-0.045041766,0.2565685511,0.0481726713,-0.0075777015,0.3368707597,-0.2633071244,0.200918898,0.0543518886,0.2817530036,0.2158435732,0.4725821316,-0.0022776665,-0.1878003627,-0.2764528692,0.0665216371,0.1926874369,-0.5811086893,0.1973800063,-0.0308278725,0.1413285732,0.2872923315,-0.1325430572,0.0486670956,-0.1231485605,0.4628397524,-0.0270578172,0.2395097613,0.0963629335,-0.1682436913,-0.1834777743,0.043465618,-0.0890507922,0.0064130211,-0.1697317511,-0.2001911998,-0.0572066903,0.0989688635,-0.2617083192,-0.2644017339,0.2126554251,-0.0461103767,-0.0920265764,-0.28864941,0.1431363076,-0.0127508091,0.1704822779,0.3351086974,0.2608134151,0.2379633188,0.5772158504,-0.0407342091,-0.1141022816,0.1889009327,0.0600242428,-0.2130747139,0.0773516297,0.3676773608,0.3506335914,0.3967398107,0.0209638998,-0.0882208273,-0.0021246681,0.2406388372,0.2035897821,-0.1885028034,0.7938399911,0.0524095446,0.1809535921,-0.116335772,0.3631625772,-0.4275179803,-0.2053982168,0.287856102,0.2680921257,0.3093080819,-0.0219096225,0.0199616514,-0.1861562878,0.5015246272,0.0297628138,-0.1064067036,-0.1711036861,-0.0851538107,-0.5885362029,-0.109198913,-0.0785844773,-0.3384315372,0.0276821163,0.4448843002,-0.272295475,0.0116580725,-0.1669053137,0.1294880956,0.0158839952,0.1184924766,0.0914537683,-0.176231131,-0.2772181332,0.058955051,0.2688049078,-0.3439932764,-0.2598087788,-0.4583736658,0.028446652,-0.0672117546,-0.1894518435,0.2401178181,0.2790070474,0.4852349758,-0.0436648764,0.3151144683,-0.0392435901,-0.2778841257,-0.396420747,0.1328745633,-0.1942772418,-0.0794956088,-0.001046693,0.1259427369,-0.327412039,0.3908373713,-0.2904158235,0.1902526468,0.054690212,-0.145225063,-0.0034682259,0.1812205017,-0.3123649955,0.1539282352,0.2389645278,0.3021642864,-0.1769243479,0.2898986638,-0.2135694921,-0.2442600131,0.5382840633,-0.1240976006,-0.0283331275,0.1594706923,0.2759436369,-0.1947040707,-0.1216937676,-0.2669958472,0.0924191847,0.4286957085,-0.2330366075,0.1069559753,-0.0202895533,-0.2478488237,0.028868109,0.0702999756,0.069040589,0.119177945,-0.1778608859,-0.2479636669,-0.2421284765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/560","title":"Using custom DownloadConfig results in an error","comments":"I see.\r\n\r\nProbably the easiest way for you would be that we add simple serialization\/deserialization methods to the Dataset and DatasetDict objects once the data files have been downloaded and all the dataset is processed.\r\n\r\nWhat do you think @lhoestq ?","body":"## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n","comment_length":41,"text":"Using custom DownloadConfig results in an error\n## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n\nI see.\r\n\r\nProbably the easiest way for you would be that we add simple serialization\/deserialization methods to the Dataset and DatasetDict objects once the data files have been downloaded and all the dataset is processed.\r\n\r\nWhat do you think @lhoestq ?","embeddings":[-0.1150195375,0.1608139873,0.1123033017,0.0204855502,-0.0762844905,-0.0592143387,0.2878563702,0.0386445262,0.3867911994,-0.1992649138,-0.1571436226,0.3636115789,-0.2807902694,-0.0396447405,0.1373905092,0.1345225722,-0.335785538,0.191671595,0.2066785991,-0.0992470011,-0.4401432276,0.4904036224,-0.1363261491,-0.1055808887,-0.1104631573,0.2965768874,-0.0853590146,0.6002389789,-0.0052087544,-0.3369760513,0.5871984363,0.0767923817,0.2937197685,0.0780296773,-0.0001159093,0.0705925748,0.2336821258,-0.3101004958,-0.3799623847,-0.2622923553,-0.0511955321,-0.1243966147,0.2814418674,-0.1946797967,-0.0140370019,-0.0242791511,0.0815578252,0.0334612019,0.0941410288,0.2972890437,0.1419871897,-0.0229717921,-0.2864381075,-0.1697261631,0.2262740731,0.260076493,0.0480878502,0.5292792916,0.3167858422,-0.2207205594,0.0189583451,-0.3049541116,-0.1514441222,0.0622548983,0.5781335831,0.0206799582,0.2644710541,-0.0439114347,-0.1765135378,0.3859302104,0.2035041451,-0.2281819582,-0.0819195062,-0.3655665219,0.144602403,-0.4327120185,0.1780243218,0.1168847531,-0.273757726,0.0803424343,-0.2586563826,-0.1680175215,-0.0763951689,0.5087516308,0.2171331793,0.3228867352,-0.0183474217,0.1714263111,-0.0767275244,-0.0576790571,0.252420336,-0.3967368007,0.0253677983,0.1348518133,-0.0958460644,-0.0770932958,0.1277973652,0.0256882552,-0.0327643231,0.1246374547,0.0682523176,-0.1068671271,0.0644042119,0.0538509563,0.0521950275,0.286323905,-0.076600723,0.1167676002,-0.2036602199,0.1669653356,-0.0508592837,-0.0176646784,0.0861193165,-0.075323768,-0.3973380923,-0.1976338178,0.168750599,0.0014936652,-0.0629578158,-0.339681685,-0.0813496038,0.1825856417,0.338355571,0.2477036715,-0.0689882264,-0.2663639188,0.1952618361,0.2603036761,-0.0893417671,0.1586500257,-0.088268958,-0.1036935523,-0.1506859362,-0.0623486452,0.4825169444,0.2577634156,0.6017351151,0.0311783329,0.0021425504,0.0281839706,-0.0155198788,0.1216142625,0.0396912619,0.1164999828,0.004311326,0.0528843515,0.0625880286,-0.2535326481,-0.2557506263,-0.0347330607,-0.1973161101,-0.5240719318,-0.32106933,0.1228109673,-0.3273592293,-0.3346843123,0.2478318959,-0.0070183282,0.060555689,-0.3129452467,-0.0337526128,0.029548144,-0.4143957198,-0.2400684655,-0.089732632,0.3797775209,0.0012102607,-0.2380746603,-0.4831041098,-0.1488144845,0.2068390101,-0.0621503629,-0.4448852539,0.2460324764,-0.3705919683,0.1259263307,0.7066502571,-0.593824029,-0.475862354,0.7060877681,-0.1530996859,0.1541841626,0.1148285046,0.3047769666,0.3644037843,-0.1938655227,0.2743984759,0.449900806,0.1645402014,-0.0376653224,-0.1342462301,-0.2366191745,-0.0452952273,0.069085665,-0.2126618922,0.3807032704,0.0322402194,0.228694275,0.2805713415,0.0719816536,-0.0035698258,0.0699649006,-0.0190699659,0.2078530788,-0.216038838,0.2683116198,-0.5918141603,0.3081945777,-0.2088387758,0.0111979246,0.1865449697,0.0156147871,-0.3768417537,-0.2749233842,-0.1442956775,-0.391390115,0.0948154703,0.3689242601,0.1382951736,-0.1243658364,-0.1281298399,0.3340533972,-0.2191484272,-0.0176077504,-0.1604177207,-0.0221684948,0.1522677392,-0.1637913138,0.0569932424,0.105084911,0.2022019923,0.1713188738,-0.0114865042,0.4094794691,-0.1540582776,0.0669990554,-0.08875563,-0.2120270282,0.0864794329,-0.0025839424,0.2240317017,0.0819461644,0.1684664637,-0.227870211,-0.0677169561,0.1494398415,0.0992672071,0.3279950321,0.0210084673,0.1423203647,0.1164696813,-0.2840330005,-0.0270047877,-0.34892869,0.0761514902,0.1613445729,0.3522972167,0.0295899063,-0.2389683723,-0.0452459715,0.2232688069,0.1981347203,0.0204542782,0.1191779003,-0.0028204762,-0.0736611784,0.029122578,0.6464995742,0.816970706,0.110755451,-0.0627681091,0.4516937435,-0.2712000906,-0.2540049255,0.279292047,-0.1479594558,0.1289023012,0.1667049378,0.0079044867,0.0080792001,-0.1439014524,-0.4434364438,-0.0547379367,-0.0084888637,-0.4380910993,-0.0696294531,-0.2960466444,-0.4512906969,-0.1279855371,-0.2876825333,-0.0485401414,-0.2141084671,-0.159273237,0.1131819934,0.1047847271,-0.0083428686,-0.2631918788,0.019749023,-0.2949643731,-0.8628898859,0.0819824636,-0.0314406157,-0.3450305164,-0.0835407823,0.0386174619,0.1652457118,0.073131986,-0.1557795554,-0.2440009415,0.1602346897,-0.1029197425,0.1122968048,0.1613503247,0.3075020313,0.3883806169,-0.1511534303,0.2565552294,-0.4360593557,0.1637900174,0.2162101418,-0.0127829853,0.0424747802,0.0326359421,-0.0875902697,-0.368976295,-0.2115421742,-0.1649634391,-0.450252533,-0.0885402486,0.28491202,0.1961458027,0.4581585228,0.1651855111,0.3243179917,-0.1645777971,0.0822794884,-0.0674815327,-0.3117092252,0.2791596651,-0.4733329117,-0.0789996907,0.085931167,-0.2227194011,0.2154388875,0.1335568577,-0.3884005845,0.199653998,-0.1469651461,0.0002483857,-0.0470809862,-0.0934197456,0.2494487315,-0.082257472,0.0479591452,-0.2609181702,-0.3588173091,0.3365956545,0.3105373681,0.5375539064,0.275968343,0.2986249626,-0.1452536583,0.4846876264,0.1508263201,-0.1560512483,0.6103883386,-0.1947124898,0.4601784647,-0.0382337458,-0.2387869209,-0.0032227396,0.3633525074,-0.0172826368,0.2315363288,-0.0611758232,-0.3007305861,-0.1902965009,-0.1062506512,-0.1545367837,-0.230031535,-0.0597753413,0.0295013599,0.1523663849,0.0646302924,0.1584214121,-0.083147265,-0.1170615926,-0.187199682,0.435128808,-0.0597942136,0.2172851712,-0.0276041552,0.2997868061,-0.3493392169,0.1421994418,-0.2189187407,0.3044458926,-0.1182061136,0.0613894761,-0.0159715731,0.116941765,0.2156404406,-0.2203439325,0.2306120545,0.262530148,0.1650993526,-0.6942793131,-0.0943868831,0.0507083125,0.3705742955,0.6355696321,0.1778835207,-0.1156038195,-0.2360606492,0.1376715899,-0.1754219085,-0.0924998894,-0.4230398536,-0.066459544,-0.3104561269,-0.2801861167,-0.3828448951,-0.1292264163,0.230558604,0.2301347256,0.1135466546,0.0101915412,0.0370249003,0.0223113801,-0.0750334486,0.0257843025,0.1655019671,0.2535585463,-0.3498473167,0.0041104979,0.0184270665,0.455697149,-0.1090064123,-0.1824812293,-0.0907652378,0.0760888234,0.1698767692,0.2375928909,-0.0469807722,-0.2788476646,-0.0843373016,-0.2431958616,0.061178796,0.1471870542,0.2561111152,-0.5553205013,0.181288138,-0.3989014328,0.5427388549,-0.1028674021,-0.0532116704,0.1602564305,-0.0507060289,-0.3970303535,0.324452132,-0.0401610807,0.5586260557,0.2404643893,0.0567302257,0.412304014,-0.433357656,0.3407605588,0.3820564151,0.0827065259,-0.2056452781,0.2356447577,0.1081556752,-0.2434303612,0.1182638481,0.4598899782,-0.1859720945,0.3259095848,-0.1559953988,-0.4174781144,0.0909639597,0.4379138649,-0.2495400608,-0.3017905354,-0.3178763986,0.1256023496,-0.0598930605,0.3444620967,-0.0450952463,-0.313447684,-0.3213606477,-0.1474078596,-0.2323348522,0.1951860338,-0.0521365926,-0.1184979379,0.1656085402,0.0278660897,0.1757669449,0.3797243237,-0.2905163169,-0.0837098733,-0.2389895022,0.0830149725,-0.1039216667,-0.0695943534,0.1519480944,0.0730860904,0.2014408559,-0.1141102687,-0.3429861069,-0.1494403929,-0.0445597135,-0.2015192062,0.4617881179,-0.0481094755,0.0759342238,-0.1242546812,-0.4921830595,0.0628540292,-0.0180091802,0.1029136181,0.0785280094,0.0157993361,-0.0287095681,0.1853824854,-0.1957309395,-0.337682426,-0.0281161107,0.3746060133,-0.3746849597,0.2907844186,0.4278508723,-0.1933617741,0.1264047176,-0.0584542118,0.0083092405,-0.5848215222,-0.216162622,0.1031734273,0.2297327518,0.164756462,0.1329749525,0.2493012846,-0.003495506,0.1802944839,0.2565452456,-0.3308797181,-0.1663021296,-0.0038664439,-0.3690316081,0.1466334462,-0.1980221123,0.2377348244,0.3796907067,-0.2016597241,-0.2258139551,0.1875278056,-0.1729793102,0.0496794544,0.1979008466,0.1081392094,0.0217493102,0.1827962399,0.0240138508,-0.062509872,-0.2874262929,-0.1339973062,-0.3909825087,0.1758333296,0.2264897525,-0.1660330445,-0.0547830835,-0.2572702467,-0.2181168497,-0.031453073,-0.0373602435,0.1865928322,-0.0764180124,0.3890280724,-0.1978252083,0.243919909,-0.0793271363,0.1648311615,-0.1698613018,0.3258364797,0.0379703566,0.1343825907,-0.1742454469,0.0088649197,-0.1068822294,-0.231389761,-0.2788546085,0.0202061161,0.2709775567,-0.0176596604,0.2543975413,0.2019496262,-0.0415051319,0.3494026363,-0.2206249982,-0.1269973963,0.2625512481,0.1083602384,-0.1582949609,0.2222321928,0.3854727447,0.3658417761,0.0349594019,0.2504579127,0.1602328718,-0.3160717785,-0.0662192702,-0.045041766,0.2565685511,0.0481726713,-0.0075777015,0.3368707597,-0.2633071244,0.200918898,0.0543518886,0.2817530036,0.2158435732,0.4725821316,-0.0022776665,-0.1878003627,-0.2764528692,0.0665216371,0.1926874369,-0.5811086893,0.1973800063,-0.0308278725,0.1413285732,0.2872923315,-0.1325430572,0.0486670956,-0.1231485605,0.4628397524,-0.0270578172,0.2395097613,0.0963629335,-0.1682436913,-0.1834777743,0.043465618,-0.0890507922,0.0064130211,-0.1697317511,-0.2001911998,-0.0572066903,0.0989688635,-0.2617083192,-0.2644017339,0.2126554251,-0.0461103767,-0.0920265764,-0.28864941,0.1431363076,-0.0127508091,0.1704822779,0.3351086974,0.2608134151,0.2379633188,0.5772158504,-0.0407342091,-0.1141022816,0.1889009327,0.0600242428,-0.2130747139,0.0773516297,0.3676773608,0.3506335914,0.3967398107,0.0209638998,-0.0882208273,-0.0021246681,0.2406388372,0.2035897821,-0.1885028034,0.7938399911,0.0524095446,0.1809535921,-0.116335772,0.3631625772,-0.4275179803,-0.2053982168,0.287856102,0.2680921257,0.3093080819,-0.0219096225,0.0199616514,-0.1861562878,0.5015246272,0.0297628138,-0.1064067036,-0.1711036861,-0.0851538107,-0.5885362029,-0.109198913,-0.0785844773,-0.3384315372,0.0276821163,0.4448843002,-0.272295475,0.0116580725,-0.1669053137,0.1294880956,0.0158839952,0.1184924766,0.0914537683,-0.176231131,-0.2772181332,0.058955051,0.2688049078,-0.3439932764,-0.2598087788,-0.4583736658,0.028446652,-0.0672117546,-0.1894518435,0.2401178181,0.2790070474,0.4852349758,-0.0436648764,0.3151144683,-0.0392435901,-0.2778841257,-0.396420747,0.1328745633,-0.1942772418,-0.0794956088,-0.001046693,0.1259427369,-0.327412039,0.3908373713,-0.2904158235,0.1902526468,0.054690212,-0.145225063,-0.0034682259,0.1812205017,-0.3123649955,0.1539282352,0.2389645278,0.3021642864,-0.1769243479,0.2898986638,-0.2135694921,-0.2442600131,0.5382840633,-0.1240976006,-0.0283331275,0.1594706923,0.2759436369,-0.1947040707,-0.1216937676,-0.2669958472,0.0924191847,0.4286957085,-0.2330366075,0.1069559753,-0.0202895533,-0.2478488237,0.028868109,0.0702999756,0.069040589,0.119177945,-0.1778608859,-0.2479636669,-0.2421284765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/560","title":"Using custom DownloadConfig results in an error","comments":"This use-case will be solved with #571 ","body":"## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n","comment_length":7,"text":"Using custom DownloadConfig results in an error\n## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n\nThis use-case will be solved with #571 ","embeddings":[-0.1150195375,0.1608139873,0.1123033017,0.0204855502,-0.0762844905,-0.0592143387,0.2878563702,0.0386445262,0.3867911994,-0.1992649138,-0.1571436226,0.3636115789,-0.2807902694,-0.0396447405,0.1373905092,0.1345225722,-0.335785538,0.191671595,0.2066785991,-0.0992470011,-0.4401432276,0.4904036224,-0.1363261491,-0.1055808887,-0.1104631573,0.2965768874,-0.0853590146,0.6002389789,-0.0052087544,-0.3369760513,0.5871984363,0.0767923817,0.2937197685,0.0780296773,-0.0001159093,0.0705925748,0.2336821258,-0.3101004958,-0.3799623847,-0.2622923553,-0.0511955321,-0.1243966147,0.2814418674,-0.1946797967,-0.0140370019,-0.0242791511,0.0815578252,0.0334612019,0.0941410288,0.2972890437,0.1419871897,-0.0229717921,-0.2864381075,-0.1697261631,0.2262740731,0.260076493,0.0480878502,0.5292792916,0.3167858422,-0.2207205594,0.0189583451,-0.3049541116,-0.1514441222,0.0622548983,0.5781335831,0.0206799582,0.2644710541,-0.0439114347,-0.1765135378,0.3859302104,0.2035041451,-0.2281819582,-0.0819195062,-0.3655665219,0.144602403,-0.4327120185,0.1780243218,0.1168847531,-0.273757726,0.0803424343,-0.2586563826,-0.1680175215,-0.0763951689,0.5087516308,0.2171331793,0.3228867352,-0.0183474217,0.1714263111,-0.0767275244,-0.0576790571,0.252420336,-0.3967368007,0.0253677983,0.1348518133,-0.0958460644,-0.0770932958,0.1277973652,0.0256882552,-0.0327643231,0.1246374547,0.0682523176,-0.1068671271,0.0644042119,0.0538509563,0.0521950275,0.286323905,-0.076600723,0.1167676002,-0.2036602199,0.1669653356,-0.0508592837,-0.0176646784,0.0861193165,-0.075323768,-0.3973380923,-0.1976338178,0.168750599,0.0014936652,-0.0629578158,-0.339681685,-0.0813496038,0.1825856417,0.338355571,0.2477036715,-0.0689882264,-0.2663639188,0.1952618361,0.2603036761,-0.0893417671,0.1586500257,-0.088268958,-0.1036935523,-0.1506859362,-0.0623486452,0.4825169444,0.2577634156,0.6017351151,0.0311783329,0.0021425504,0.0281839706,-0.0155198788,0.1216142625,0.0396912619,0.1164999828,0.004311326,0.0528843515,0.0625880286,-0.2535326481,-0.2557506263,-0.0347330607,-0.1973161101,-0.5240719318,-0.32106933,0.1228109673,-0.3273592293,-0.3346843123,0.2478318959,-0.0070183282,0.060555689,-0.3129452467,-0.0337526128,0.029548144,-0.4143957198,-0.2400684655,-0.089732632,0.3797775209,0.0012102607,-0.2380746603,-0.4831041098,-0.1488144845,0.2068390101,-0.0621503629,-0.4448852539,0.2460324764,-0.3705919683,0.1259263307,0.7066502571,-0.593824029,-0.475862354,0.7060877681,-0.1530996859,0.1541841626,0.1148285046,0.3047769666,0.3644037843,-0.1938655227,0.2743984759,0.449900806,0.1645402014,-0.0376653224,-0.1342462301,-0.2366191745,-0.0452952273,0.069085665,-0.2126618922,0.3807032704,0.0322402194,0.228694275,0.2805713415,0.0719816536,-0.0035698258,0.0699649006,-0.0190699659,0.2078530788,-0.216038838,0.2683116198,-0.5918141603,0.3081945777,-0.2088387758,0.0111979246,0.1865449697,0.0156147871,-0.3768417537,-0.2749233842,-0.1442956775,-0.391390115,0.0948154703,0.3689242601,0.1382951736,-0.1243658364,-0.1281298399,0.3340533972,-0.2191484272,-0.0176077504,-0.1604177207,-0.0221684948,0.1522677392,-0.1637913138,0.0569932424,0.105084911,0.2022019923,0.1713188738,-0.0114865042,0.4094794691,-0.1540582776,0.0669990554,-0.08875563,-0.2120270282,0.0864794329,-0.0025839424,0.2240317017,0.0819461644,0.1684664637,-0.227870211,-0.0677169561,0.1494398415,0.0992672071,0.3279950321,0.0210084673,0.1423203647,0.1164696813,-0.2840330005,-0.0270047877,-0.34892869,0.0761514902,0.1613445729,0.3522972167,0.0295899063,-0.2389683723,-0.0452459715,0.2232688069,0.1981347203,0.0204542782,0.1191779003,-0.0028204762,-0.0736611784,0.029122578,0.6464995742,0.816970706,0.110755451,-0.0627681091,0.4516937435,-0.2712000906,-0.2540049255,0.279292047,-0.1479594558,0.1289023012,0.1667049378,0.0079044867,0.0080792001,-0.1439014524,-0.4434364438,-0.0547379367,-0.0084888637,-0.4380910993,-0.0696294531,-0.2960466444,-0.4512906969,-0.1279855371,-0.2876825333,-0.0485401414,-0.2141084671,-0.159273237,0.1131819934,0.1047847271,-0.0083428686,-0.2631918788,0.019749023,-0.2949643731,-0.8628898859,0.0819824636,-0.0314406157,-0.3450305164,-0.0835407823,0.0386174619,0.1652457118,0.073131986,-0.1557795554,-0.2440009415,0.1602346897,-0.1029197425,0.1122968048,0.1613503247,0.3075020313,0.3883806169,-0.1511534303,0.2565552294,-0.4360593557,0.1637900174,0.2162101418,-0.0127829853,0.0424747802,0.0326359421,-0.0875902697,-0.368976295,-0.2115421742,-0.1649634391,-0.450252533,-0.0885402486,0.28491202,0.1961458027,0.4581585228,0.1651855111,0.3243179917,-0.1645777971,0.0822794884,-0.0674815327,-0.3117092252,0.2791596651,-0.4733329117,-0.0789996907,0.085931167,-0.2227194011,0.2154388875,0.1335568577,-0.3884005845,0.199653998,-0.1469651461,0.0002483857,-0.0470809862,-0.0934197456,0.2494487315,-0.082257472,0.0479591452,-0.2609181702,-0.3588173091,0.3365956545,0.3105373681,0.5375539064,0.275968343,0.2986249626,-0.1452536583,0.4846876264,0.1508263201,-0.1560512483,0.6103883386,-0.1947124898,0.4601784647,-0.0382337458,-0.2387869209,-0.0032227396,0.3633525074,-0.0172826368,0.2315363288,-0.0611758232,-0.3007305861,-0.1902965009,-0.1062506512,-0.1545367837,-0.230031535,-0.0597753413,0.0295013599,0.1523663849,0.0646302924,0.1584214121,-0.083147265,-0.1170615926,-0.187199682,0.435128808,-0.0597942136,0.2172851712,-0.0276041552,0.2997868061,-0.3493392169,0.1421994418,-0.2189187407,0.3044458926,-0.1182061136,0.0613894761,-0.0159715731,0.116941765,0.2156404406,-0.2203439325,0.2306120545,0.262530148,0.1650993526,-0.6942793131,-0.0943868831,0.0507083125,0.3705742955,0.6355696321,0.1778835207,-0.1156038195,-0.2360606492,0.1376715899,-0.1754219085,-0.0924998894,-0.4230398536,-0.066459544,-0.3104561269,-0.2801861167,-0.3828448951,-0.1292264163,0.230558604,0.2301347256,0.1135466546,0.0101915412,0.0370249003,0.0223113801,-0.0750334486,0.0257843025,0.1655019671,0.2535585463,-0.3498473167,0.0041104979,0.0184270665,0.455697149,-0.1090064123,-0.1824812293,-0.0907652378,0.0760888234,0.1698767692,0.2375928909,-0.0469807722,-0.2788476646,-0.0843373016,-0.2431958616,0.061178796,0.1471870542,0.2561111152,-0.5553205013,0.181288138,-0.3989014328,0.5427388549,-0.1028674021,-0.0532116704,0.1602564305,-0.0507060289,-0.3970303535,0.324452132,-0.0401610807,0.5586260557,0.2404643893,0.0567302257,0.412304014,-0.433357656,0.3407605588,0.3820564151,0.0827065259,-0.2056452781,0.2356447577,0.1081556752,-0.2434303612,0.1182638481,0.4598899782,-0.1859720945,0.3259095848,-0.1559953988,-0.4174781144,0.0909639597,0.4379138649,-0.2495400608,-0.3017905354,-0.3178763986,0.1256023496,-0.0598930605,0.3444620967,-0.0450952463,-0.313447684,-0.3213606477,-0.1474078596,-0.2323348522,0.1951860338,-0.0521365926,-0.1184979379,0.1656085402,0.0278660897,0.1757669449,0.3797243237,-0.2905163169,-0.0837098733,-0.2389895022,0.0830149725,-0.1039216667,-0.0695943534,0.1519480944,0.0730860904,0.2014408559,-0.1141102687,-0.3429861069,-0.1494403929,-0.0445597135,-0.2015192062,0.4617881179,-0.0481094755,0.0759342238,-0.1242546812,-0.4921830595,0.0628540292,-0.0180091802,0.1029136181,0.0785280094,0.0157993361,-0.0287095681,0.1853824854,-0.1957309395,-0.337682426,-0.0281161107,0.3746060133,-0.3746849597,0.2907844186,0.4278508723,-0.1933617741,0.1264047176,-0.0584542118,0.0083092405,-0.5848215222,-0.216162622,0.1031734273,0.2297327518,0.164756462,0.1329749525,0.2493012846,-0.003495506,0.1802944839,0.2565452456,-0.3308797181,-0.1663021296,-0.0038664439,-0.3690316081,0.1466334462,-0.1980221123,0.2377348244,0.3796907067,-0.2016597241,-0.2258139551,0.1875278056,-0.1729793102,0.0496794544,0.1979008466,0.1081392094,0.0217493102,0.1827962399,0.0240138508,-0.062509872,-0.2874262929,-0.1339973062,-0.3909825087,0.1758333296,0.2264897525,-0.1660330445,-0.0547830835,-0.2572702467,-0.2181168497,-0.031453073,-0.0373602435,0.1865928322,-0.0764180124,0.3890280724,-0.1978252083,0.243919909,-0.0793271363,0.1648311615,-0.1698613018,0.3258364797,0.0379703566,0.1343825907,-0.1742454469,0.0088649197,-0.1068822294,-0.231389761,-0.2788546085,0.0202061161,0.2709775567,-0.0176596604,0.2543975413,0.2019496262,-0.0415051319,0.3494026363,-0.2206249982,-0.1269973963,0.2625512481,0.1083602384,-0.1582949609,0.2222321928,0.3854727447,0.3658417761,0.0349594019,0.2504579127,0.1602328718,-0.3160717785,-0.0662192702,-0.045041766,0.2565685511,0.0481726713,-0.0075777015,0.3368707597,-0.2633071244,0.200918898,0.0543518886,0.2817530036,0.2158435732,0.4725821316,-0.0022776665,-0.1878003627,-0.2764528692,0.0665216371,0.1926874369,-0.5811086893,0.1973800063,-0.0308278725,0.1413285732,0.2872923315,-0.1325430572,0.0486670956,-0.1231485605,0.4628397524,-0.0270578172,0.2395097613,0.0963629335,-0.1682436913,-0.1834777743,0.043465618,-0.0890507922,0.0064130211,-0.1697317511,-0.2001911998,-0.0572066903,0.0989688635,-0.2617083192,-0.2644017339,0.2126554251,-0.0461103767,-0.0920265764,-0.28864941,0.1431363076,-0.0127508091,0.1704822779,0.3351086974,0.2608134151,0.2379633188,0.5772158504,-0.0407342091,-0.1141022816,0.1889009327,0.0600242428,-0.2130747139,0.0773516297,0.3676773608,0.3506335914,0.3967398107,0.0209638998,-0.0882208273,-0.0021246681,0.2406388372,0.2035897821,-0.1885028034,0.7938399911,0.0524095446,0.1809535921,-0.116335772,0.3631625772,-0.4275179803,-0.2053982168,0.287856102,0.2680921257,0.3093080819,-0.0219096225,0.0199616514,-0.1861562878,0.5015246272,0.0297628138,-0.1064067036,-0.1711036861,-0.0851538107,-0.5885362029,-0.109198913,-0.0785844773,-0.3384315372,0.0276821163,0.4448843002,-0.272295475,0.0116580725,-0.1669053137,0.1294880956,0.0158839952,0.1184924766,0.0914537683,-0.176231131,-0.2772181332,0.058955051,0.2688049078,-0.3439932764,-0.2598087788,-0.4583736658,0.028446652,-0.0672117546,-0.1894518435,0.2401178181,0.2790070474,0.4852349758,-0.0436648764,0.3151144683,-0.0392435901,-0.2778841257,-0.396420747,0.1328745633,-0.1942772418,-0.0794956088,-0.001046693,0.1259427369,-0.327412039,0.3908373713,-0.2904158235,0.1902526468,0.054690212,-0.145225063,-0.0034682259,0.1812205017,-0.3123649955,0.1539282352,0.2389645278,0.3021642864,-0.1769243479,0.2898986638,-0.2135694921,-0.2442600131,0.5382840633,-0.1240976006,-0.0283331275,0.1594706923,0.2759436369,-0.1947040707,-0.1216937676,-0.2669958472,0.0924191847,0.4286957085,-0.2330366075,0.1069559753,-0.0202895533,-0.2478488237,0.028868109,0.0702999756,0.069040589,0.119177945,-0.1778608859,-0.2479636669,-0.2421284765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/560","title":"Using custom DownloadConfig results in an error","comments":"Thank you very much @thomwolf and @lhoestq we will give it a try","body":"## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n","comment_length":13,"text":"Using custom DownloadConfig results in an error\n## Version \/ Environment\r\n\r\nUbuntu 18.04\r\nPython 3.6.8\r\nnlp 0.4.0\r\n\r\n## Description\r\n\r\nLoading `imdb` dataset works fine when when I don't specify any `download_config` argument. When I create a custom `DownloadConfig` object and pass it to the `nlp.load_dataset` function, this results in an error.\r\n\r\n## How to reproduce\r\n\r\n### Example without DownloadConfig --> works\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-without-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n imdb = nlp.load_dataset(path=\"imdb\")\r\n```\r\n\r\n### Example with DownloadConfig --> doesn't work\r\n\r\n```python\r\nimport os\r\n\r\nos.environ[\"HF_HOME\"] = \"\/data\/hf-test-with-dl-config-01\/\"\r\n\r\nimport logging\r\nimport nlp\r\nfrom nlp.utils import DownloadConfig\r\n\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nif __name__ == \"__main__\":\r\n download_config = DownloadConfig()\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n```\r\n\r\nError traceback:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/...\/example_with_dl_config.py\", line 13, in <module>\r\n imdb = nlp.load_dataset(path=\"imdb\", download_config=download_config)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 463, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/builder.py\", line 518, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/datasets\/imdb\/76cdbd7249ea3548c928bbf304258dab44d09cd3638d9da8d42480d1d1be3743\/imdb.py\", line 86, in _split_generators\r\n arch_path = dl_manager.download_and_extract(_DOWNLOAD_URL)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 220, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 158, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 108, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n File \"\/...\/python3.6\/python3.6\/site-packages\/nlp\/utils\/info_utils.py\", line 79, in get_size_checksum_dict\r\n with open(path, \"rb\") as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '\/data\/hf-test-with-dl-config-01\/datasets\/extracted\/b6802c5b61824b2c1f7dbf7cda6696b5f2e22214e18d171ce1ed3be90c931ce5'\r\n```\r\n\r\n\nThank you very much @thomwolf and @lhoestq we will give it a try","embeddings":[-0.1150195375,0.1608139873,0.1123033017,0.0204855502,-0.0762844905,-0.0592143387,0.2878563702,0.0386445262,0.3867911994,-0.1992649138,-0.1571436226,0.3636115789,-0.2807902694,-0.0396447405,0.1373905092,0.1345225722,-0.335785538,0.191671595,0.2066785991,-0.0992470011,-0.4401432276,0.4904036224,-0.1363261491,-0.1055808887,-0.1104631573,0.2965768874,-0.0853590146,0.6002389789,-0.0052087544,-0.3369760513,0.5871984363,0.0767923817,0.2937197685,0.0780296773,-0.0001159093,0.0705925748,0.2336821258,-0.3101004958,-0.3799623847,-0.2622923553,-0.0511955321,-0.1243966147,0.2814418674,-0.1946797967,-0.0140370019,-0.0242791511,0.0815578252,0.0334612019,0.0941410288,0.2972890437,0.1419871897,-0.0229717921,-0.2864381075,-0.1697261631,0.2262740731,0.260076493,0.0480878502,0.5292792916,0.3167858422,-0.2207205594,0.0189583451,-0.3049541116,-0.1514441222,0.0622548983,0.5781335831,0.0206799582,0.2644710541,-0.0439114347,-0.1765135378,0.3859302104,0.2035041451,-0.2281819582,-0.0819195062,-0.3655665219,0.144602403,-0.4327120185,0.1780243218,0.1168847531,-0.273757726,0.0803424343,-0.2586563826,-0.1680175215,-0.0763951689,0.5087516308,0.2171331793,0.3228867352,-0.0183474217,0.1714263111,-0.0767275244,-0.0576790571,0.252420336,-0.3967368007,0.0253677983,0.1348518133,-0.0958460644,-0.0770932958,0.1277973652,0.0256882552,-0.0327643231,0.1246374547,0.0682523176,-0.1068671271,0.0644042119,0.0538509563,0.0521950275,0.286323905,-0.076600723,0.1167676002,-0.2036602199,0.1669653356,-0.0508592837,-0.0176646784,0.0861193165,-0.075323768,-0.3973380923,-0.1976338178,0.168750599,0.0014936652,-0.0629578158,-0.339681685,-0.0813496038,0.1825856417,0.338355571,0.2477036715,-0.0689882264,-0.2663639188,0.1952618361,0.2603036761,-0.0893417671,0.1586500257,-0.088268958,-0.1036935523,-0.1506859362,-0.0623486452,0.4825169444,0.2577634156,0.6017351151,0.0311783329,0.0021425504,0.0281839706,-0.0155198788,0.1216142625,0.0396912619,0.1164999828,0.004311326,0.0528843515,0.0625880286,-0.2535326481,-0.2557506263,-0.0347330607,-0.1973161101,-0.5240719318,-0.32106933,0.1228109673,-0.3273592293,-0.3346843123,0.2478318959,-0.0070183282,0.060555689,-0.3129452467,-0.0337526128,0.029548144,-0.4143957198,-0.2400684655,-0.089732632,0.3797775209,0.0012102607,-0.2380746603,-0.4831041098,-0.1488144845,0.2068390101,-0.0621503629,-0.4448852539,0.2460324764,-0.3705919683,0.1259263307,0.7066502571,-0.593824029,-0.475862354,0.7060877681,-0.1530996859,0.1541841626,0.1148285046,0.3047769666,0.3644037843,-0.1938655227,0.2743984759,0.449900806,0.1645402014,-0.0376653224,-0.1342462301,-0.2366191745,-0.0452952273,0.069085665,-0.2126618922,0.3807032704,0.0322402194,0.228694275,0.2805713415,0.0719816536,-0.0035698258,0.0699649006,-0.0190699659,0.2078530788,-0.216038838,0.2683116198,-0.5918141603,0.3081945777,-0.2088387758,0.0111979246,0.1865449697,0.0156147871,-0.3768417537,-0.2749233842,-0.1442956775,-0.391390115,0.0948154703,0.3689242601,0.1382951736,-0.1243658364,-0.1281298399,0.3340533972,-0.2191484272,-0.0176077504,-0.1604177207,-0.0221684948,0.1522677392,-0.1637913138,0.0569932424,0.105084911,0.2022019923,0.1713188738,-0.0114865042,0.4094794691,-0.1540582776,0.0669990554,-0.08875563,-0.2120270282,0.0864794329,-0.0025839424,0.2240317017,0.0819461644,0.1684664637,-0.227870211,-0.0677169561,0.1494398415,0.0992672071,0.3279950321,0.0210084673,0.1423203647,0.1164696813,-0.2840330005,-0.0270047877,-0.34892869,0.0761514902,0.1613445729,0.3522972167,0.0295899063,-0.2389683723,-0.0452459715,0.2232688069,0.1981347203,0.0204542782,0.1191779003,-0.0028204762,-0.0736611784,0.029122578,0.6464995742,0.816970706,0.110755451,-0.0627681091,0.4516937435,-0.2712000906,-0.2540049255,0.279292047,-0.1479594558,0.1289023012,0.1667049378,0.0079044867,0.0080792001,-0.1439014524,-0.4434364438,-0.0547379367,-0.0084888637,-0.4380910993,-0.0696294531,-0.2960466444,-0.4512906969,-0.1279855371,-0.2876825333,-0.0485401414,-0.2141084671,-0.159273237,0.1131819934,0.1047847271,-0.0083428686,-0.2631918788,0.019749023,-0.2949643731,-0.8628898859,0.0819824636,-0.0314406157,-0.3450305164,-0.0835407823,0.0386174619,0.1652457118,0.073131986,-0.1557795554,-0.2440009415,0.1602346897,-0.1029197425,0.1122968048,0.1613503247,0.3075020313,0.3883806169,-0.1511534303,0.2565552294,-0.4360593557,0.1637900174,0.2162101418,-0.0127829853,0.0424747802,0.0326359421,-0.0875902697,-0.368976295,-0.2115421742,-0.1649634391,-0.450252533,-0.0885402486,0.28491202,0.1961458027,0.4581585228,0.1651855111,0.3243179917,-0.1645777971,0.0822794884,-0.0674815327,-0.3117092252,0.2791596651,-0.4733329117,-0.0789996907,0.085931167,-0.2227194011,0.2154388875,0.1335568577,-0.3884005845,0.199653998,-0.1469651461,0.0002483857,-0.0470809862,-0.0934197456,0.2494487315,-0.082257472,0.0479591452,-0.2609181702,-0.3588173091,0.3365956545,0.3105373681,0.5375539064,0.275968343,0.2986249626,-0.1452536583,0.4846876264,0.1508263201,-0.1560512483,0.6103883386,-0.1947124898,0.4601784647,-0.0382337458,-0.2387869209,-0.0032227396,0.3633525074,-0.0172826368,0.2315363288,-0.0611758232,-0.3007305861,-0.1902965009,-0.1062506512,-0.1545367837,-0.230031535,-0.0597753413,0.0295013599,0.1523663849,0.0646302924,0.1584214121,-0.083147265,-0.1170615926,-0.187199682,0.435128808,-0.0597942136,0.2172851712,-0.0276041552,0.2997868061,-0.3493392169,0.1421994418,-0.2189187407,0.3044458926,-0.1182061136,0.0613894761,-0.0159715731,0.116941765,0.2156404406,-0.2203439325,0.2306120545,0.262530148,0.1650993526,-0.6942793131,-0.0943868831,0.0507083125,0.3705742955,0.6355696321,0.1778835207,-0.1156038195,-0.2360606492,0.1376715899,-0.1754219085,-0.0924998894,-0.4230398536,-0.066459544,-0.3104561269,-0.2801861167,-0.3828448951,-0.1292264163,0.230558604,0.2301347256,0.1135466546,0.0101915412,0.0370249003,0.0223113801,-0.0750334486,0.0257843025,0.1655019671,0.2535585463,-0.3498473167,0.0041104979,0.0184270665,0.455697149,-0.1090064123,-0.1824812293,-0.0907652378,0.0760888234,0.1698767692,0.2375928909,-0.0469807722,-0.2788476646,-0.0843373016,-0.2431958616,0.061178796,0.1471870542,0.2561111152,-0.5553205013,0.181288138,-0.3989014328,0.5427388549,-0.1028674021,-0.0532116704,0.1602564305,-0.0507060289,-0.3970303535,0.324452132,-0.0401610807,0.5586260557,0.2404643893,0.0567302257,0.412304014,-0.433357656,0.3407605588,0.3820564151,0.0827065259,-0.2056452781,0.2356447577,0.1081556752,-0.2434303612,0.1182638481,0.4598899782,-0.1859720945,0.3259095848,-0.1559953988,-0.4174781144,0.0909639597,0.4379138649,-0.2495400608,-0.3017905354,-0.3178763986,0.1256023496,-0.0598930605,0.3444620967,-0.0450952463,-0.313447684,-0.3213606477,-0.1474078596,-0.2323348522,0.1951860338,-0.0521365926,-0.1184979379,0.1656085402,0.0278660897,0.1757669449,0.3797243237,-0.2905163169,-0.0837098733,-0.2389895022,0.0830149725,-0.1039216667,-0.0695943534,0.1519480944,0.0730860904,0.2014408559,-0.1141102687,-0.3429861069,-0.1494403929,-0.0445597135,-0.2015192062,0.4617881179,-0.0481094755,0.0759342238,-0.1242546812,-0.4921830595,0.0628540292,-0.0180091802,0.1029136181,0.0785280094,0.0157993361,-0.0287095681,0.1853824854,-0.1957309395,-0.337682426,-0.0281161107,0.3746060133,-0.3746849597,0.2907844186,0.4278508723,-0.1933617741,0.1264047176,-0.0584542118,0.0083092405,-0.5848215222,-0.216162622,0.1031734273,0.2297327518,0.164756462,0.1329749525,0.2493012846,-0.003495506,0.1802944839,0.2565452456,-0.3308797181,-0.1663021296,-0.0038664439,-0.3690316081,0.1466334462,-0.1980221123,0.2377348244,0.3796907067,-0.2016597241,-0.2258139551,0.1875278056,-0.1729793102,0.0496794544,0.1979008466,0.1081392094,0.0217493102,0.1827962399,0.0240138508,-0.062509872,-0.2874262929,-0.1339973062,-0.3909825087,0.1758333296,0.2264897525,-0.1660330445,-0.0547830835,-0.2572702467,-0.2181168497,-0.031453073,-0.0373602435,0.1865928322,-0.0764180124,0.3890280724,-0.1978252083,0.243919909,-0.0793271363,0.1648311615,-0.1698613018,0.3258364797,0.0379703566,0.1343825907,-0.1742454469,0.0088649197,-0.1068822294,-0.231389761,-0.2788546085,0.0202061161,0.2709775567,-0.0176596604,0.2543975413,0.2019496262,-0.0415051319,0.3494026363,-0.2206249982,-0.1269973963,0.2625512481,0.1083602384,-0.1582949609,0.2222321928,0.3854727447,0.3658417761,0.0349594019,0.2504579127,0.1602328718,-0.3160717785,-0.0662192702,-0.045041766,0.2565685511,0.0481726713,-0.0075777015,0.3368707597,-0.2633071244,0.200918898,0.0543518886,0.2817530036,0.2158435732,0.4725821316,-0.0022776665,-0.1878003627,-0.2764528692,0.0665216371,0.1926874369,-0.5811086893,0.1973800063,-0.0308278725,0.1413285732,0.2872923315,-0.1325430572,0.0486670956,-0.1231485605,0.4628397524,-0.0270578172,0.2395097613,0.0963629335,-0.1682436913,-0.1834777743,0.043465618,-0.0890507922,0.0064130211,-0.1697317511,-0.2001911998,-0.0572066903,0.0989688635,-0.2617083192,-0.2644017339,0.2126554251,-0.0461103767,-0.0920265764,-0.28864941,0.1431363076,-0.0127508091,0.1704822779,0.3351086974,0.2608134151,0.2379633188,0.5772158504,-0.0407342091,-0.1141022816,0.1889009327,0.0600242428,-0.2130747139,0.0773516297,0.3676773608,0.3506335914,0.3967398107,0.0209638998,-0.0882208273,-0.0021246681,0.2406388372,0.2035897821,-0.1885028034,0.7938399911,0.0524095446,0.1809535921,-0.116335772,0.3631625772,-0.4275179803,-0.2053982168,0.287856102,0.2680921257,0.3093080819,-0.0219096225,0.0199616514,-0.1861562878,0.5015246272,0.0297628138,-0.1064067036,-0.1711036861,-0.0851538107,-0.5885362029,-0.109198913,-0.0785844773,-0.3384315372,0.0276821163,0.4448843002,-0.272295475,0.0116580725,-0.1669053137,0.1294880956,0.0158839952,0.1184924766,0.0914537683,-0.176231131,-0.2772181332,0.058955051,0.2688049078,-0.3439932764,-0.2598087788,-0.4583736658,0.028446652,-0.0672117546,-0.1894518435,0.2401178181,0.2790070474,0.4852349758,-0.0436648764,0.3151144683,-0.0392435901,-0.2778841257,-0.396420747,0.1328745633,-0.1942772418,-0.0794956088,-0.001046693,0.1259427369,-0.327412039,0.3908373713,-0.2904158235,0.1902526468,0.054690212,-0.145225063,-0.0034682259,0.1812205017,-0.3123649955,0.1539282352,0.2389645278,0.3021642864,-0.1769243479,0.2898986638,-0.2135694921,-0.2442600131,0.5382840633,-0.1240976006,-0.0283331275,0.1594706923,0.2759436369,-0.1947040707,-0.1216937676,-0.2669958472,0.0924191847,0.4286957085,-0.2330366075,0.1069559753,-0.0202895533,-0.2478488237,0.028868109,0.0702999756,0.069040589,0.119177945,-0.1778608859,-0.2479636669,-0.2421284765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/554","title":"nlp downloads to its module path","comments":"Indeed this is a known issue arising from the fact that we try to be compatible with cloupickle.\r\n\r\nDoes this also happen if you are installing in a virtual environment?","body":"I am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?","comment_length":30,"text":"nlp downloads to its module path\nI am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?\nIndeed this is a known issue arising from the fact that we try to be compatible with cloupickle.\r\n\r\nDoes this also happen if you are installing in a virtual environment?","embeddings":[0.0326843746,0.2539580166,0.1499526352,0.0345963463,0.214757815,-0.1813214272,-0.0204587802,-0.0045512239,0.1024628431,-0.0508642942,0.1380285323,0.8880224228,-0.442250967,0.0815139189,0.3362323046,0.3320013881,-0.1796586066,-0.0002901663,-0.4403924942,-0.0244457889,-0.212073341,0.345549643,0.0295308251,0.1748972386,0.0630624965,-0.2714599073,0.1764766276,0.2207587212,-0.2017823905,-0.2983646989,0.2262380868,0.0677955747,0.0233549811,0.2312595993,-0.0001246572,-0.1266600043,0.1466834694,-0.2211489081,-0.5942364335,-0.4229247868,0.0836751238,-0.1504780948,0.3097313643,-0.2473908216,0.4054462016,0.6272526979,0.2100753635,-0.0133366985,0.0938679725,0.180040136,0.129822284,0.3722446263,-0.146121189,0.354054302,0.1991157532,0.1898444295,-0.2509313822,0.5630124211,-0.0468491726,-0.1564094275,0.1237194836,-0.0159906466,-0.1453780383,-0.2201298177,0.4053889215,0.0232429337,0.542170763,-0.1580370963,-0.1846934557,0.002779186,0.1580805629,-0.6352776885,-0.2875439227,-0.3459047377,0.0451218262,-0.1250391304,0.520231545,0.3565153182,-0.2196206152,0.1207837164,0.062723279,-0.3631262481,-0.1046481654,0.3527787626,0.0758769885,0.4675630331,-0.0018829404,0.0300988741,0.3376191258,0.2649231553,0.3683661222,0.1396361142,0.0817191228,0.2940796912,0.2760203183,-0.0600535199,-0.0974647105,0.3331544697,-0.0598653778,0.0230617691,0.1331875026,-0.1023270339,-0.0971473083,0.2347497195,0.240801245,-0.2937859595,-0.0199316982,0.1705410331,0.2132526338,0.1521571428,0.2788050175,0.0625127256,-0.0701544359,-0.0998438001,-0.1074586064,-0.1125402004,0.0609444045,-0.1470344216,-0.0614198893,-0.1838048697,-0.1485045552,0.2996825278,0.2301567495,0.0253079124,-0.1921526641,0.1033578962,0.2294237614,0.0773207694,-0.3032056093,0.4586708248,0.0294333175,0.1733400971,-0.3338633478,0.0488014333,0.1534634978,0.1618083417,0.3029888868,-0.2545301914,-0.0548669472,-0.1869945824,0.0137971649,0.2578770518,-0.231493175,0.3605966866,0.0254138689,0.3670630753,-0.2235779911,-0.2853199542,-0.459127903,0.1197424158,-0.2980464101,-0.3935777247,0.1862976104,0.0347367786,-0.2341649532,-0.0807254016,0.1144634932,-0.6620383859,0.2572256327,-0.2190913707,0.1985624731,-0.2899947166,-0.2044749856,-0.2870658338,-0.0835918561,0.4170087576,0.1406819522,-0.2434347421,-0.2343941927,-0.0062076612,0.4222149253,0.1802350581,-0.1540425867,-0.0480853803,-0.2764587998,-0.2796525359,0.8423376083,-0.3605712056,-0.3386266232,0.2357316464,-0.5933491588,-0.0297031067,0.0297485106,0.1759302914,-0.4824045002,-0.1026944295,0.0016464071,0.7592406273,0.2715176642,0.0893569589,-0.3122752309,-0.2933425307,0.2515257597,-0.1365662664,-0.0736148879,0.0352933146,-0.0354187638,0.5409154892,0.5577620864,0.2064168006,0.0009637355,0.3216338158,0.5599628091,0.1643122435,-0.2873654068,-0.1587284207,0.0221837405,-0.0262412354,-0.8242051005,0.1558991969,-0.5611915588,-0.2444065213,-0.104234457,0.1596709043,0.1721691191,0.0421344712,-0.0329916552,0.2325987965,0.0532537661,0.1800094694,-0.1128176898,0.4923970401,-0.1988110244,0.1270056963,-0.1246593744,0.1227730662,-0.1702817678,-0.1980679631,-0.0759830251,0.4849922657,0.1350174248,-0.1688594818,0.0287453476,0.5297238827,-0.1243889034,0.1040727273,-0.3746854961,0.0564400889,0.3235250711,0.0784604698,0.0241341349,0.2581119537,-0.0472500548,0.0969391465,-0.1821463704,0.1139857545,-0.3834557831,0.0970005989,0.1707639843,0.2843174636,0.033099696,0.053851448,-0.1857933253,-0.0878794566,0.2567599416,0.2180497795,0.4287627041,0.3316879869,0.0636417195,-0.0161924902,0.3589437008,0.1764912158,0.1916058064,0.0315664075,0.041478321,-0.038178131,0.0644056499,0.2960349917,0.1246425807,0.0143525209,0.115152061,0.2016739547,-0.3002697825,-0.1990064383,0.1411170363,0.002523032,0.2384322286,0.1320196837,0.1465292722,-0.1913343221,-0.1618216187,0.0934488699,0.0271345172,0.1175148338,-0.3330697417,0.0698605552,-0.5621652007,-0.3618030846,-0.6628373265,0.1497510076,-0.5517084002,-0.2448141128,-0.1711098552,0.3884853125,-0.0105235167,-0.0930337682,-0.0906846225,-0.087428771,-0.4265940189,-0.5033159852,0.0046081385,0.0666370764,-0.3229703009,-0.1529540569,0.2286383063,0.1363472044,0.1716546863,-0.0452005267,-0.4361425042,0.0560572073,-0.2377965301,0.0539311469,0.3451829553,0.1509552449,0.2952367961,-0.1043479443,0.0696856603,-0.2573923171,0.0952086374,-0.1837004572,0.1074125841,0.0065734577,0.1170465201,0.0201885458,-0.3216100335,-0.1229667962,-0.4191868007,-0.4624139071,0.3291345835,0.2890335917,0.3142392635,0.8311154842,-0.1648727506,-0.0336302593,-0.1469506174,0.1657740474,0.2017695904,0.035377536,0.2355373502,-0.226958558,-0.1271612644,-0.1235052645,0.018370498,0.0342529044,0.3961649537,-0.2507331967,-0.0285671856,-0.0195147749,0.0932909027,0.1410595924,0.0818161294,0.5214660764,0.2803615928,0.1878141314,-0.0578506142,-0.2552235723,0.2268039137,0.5429323316,0.4109835029,0.3115747571,0.1435751021,0.2280550152,0.312443763,-0.0587832779,0.215649724,0.279889673,0.014291686,0.5034257174,0.0840158239,-0.0997243598,-0.2104303986,0.0449197255,-0.1026230082,0.050526429,0.1502986699,-0.2044912279,-0.1009589583,0.0065251496,-0.0164254867,-0.408855021,-0.0000797209,0.0778424069,0.3124973774,0.1592760086,0.0183886923,-0.1750243902,-0.2969209552,0.2853292823,0.3175711334,-0.0364395864,0.3643212318,-0.2949274778,0.0840503648,-0.3474314809,-0.3622843027,-0.102194421,0.3788607121,-0.0956839025,-0.0347492956,0.1840121895,-0.1730255038,0.1940382868,-0.0323353074,0.2411331087,0.1861349791,-0.0809582248,-0.5383049846,-0.0579554327,0.0232510064,0.4209911525,0.2981879711,-0.0265945476,-0.0678070262,-0.0809438005,0.0771075487,0.0870723054,-0.1857691705,-0.1419165283,-0.2486450225,-0.4294263721,-0.225902766,0.0182488058,0.0481487215,0.171020627,0.2736572325,-0.0725753009,0.1815531105,-0.1232747436,-0.0012338248,-0.2885781527,0.2251169086,0.0260172505,0.1176228225,-0.1871997565,0.151719138,-0.270637542,0.3499566913,-0.0563949607,0.1945851296,-0.2746193707,-0.1085656807,0.3069177866,0.5553452373,-0.0173649639,0.0041446337,-0.0703744665,-0.4417226315,0.0162285175,0.167255491,-0.1102345735,-0.3083549142,0.1300924271,-0.2451559305,0.2288478464,0.1490292549,-0.2295273095,0.3482482135,-0.1365222186,-0.4190997779,0.1697119176,0.2539562881,1.1712492704,-0.1969950348,0.2815849483,0.132633552,-0.3524650931,0.2458818555,-0.2211242318,0.2156592011,-0.4574078918,0.2059419751,0.0255226791,-0.3373981118,-0.0713266656,-0.0014053899,0.1140248105,0.2740053833,-0.0155984191,-0.2410455048,0.1407306343,0.3760009408,-0.1842535287,-0.2960529327,-0.3168869317,-0.0070974226,-0.0626246333,0.4845359623,-0.1284253746,-0.1354673356,0.1815513819,-0.1964697838,0.0611094497,0.1740922034,-0.3605582118,0.0303994119,-0.0645299479,-0.0112136304,-0.0753134489,0.1288613975,-0.0585702658,0.5918822289,0.0438423343,-0.2288346142,-0.0650869831,-0.3218739629,0.2259808332,0.431986779,0.0406467095,-0.0441016555,0.0701378882,-0.4085543752,-0.0115288999,-0.3302136064,0.2451237291,-0.0626667663,0.248924911,0.197911039,-0.3371074498,0.0066840062,-0.0553533696,-0.0486362129,0.0257920083,0.1970212311,-0.0889837667,0.1510649323,-0.1321663857,-0.3250288963,-0.0004810002,0.5062596202,-0.0911182687,-0.0040776185,0.2302753329,-0.2699753046,-0.2008228898,-0.1109198779,-0.1100176051,-0.4767761827,-0.1046166867,-0.0447584726,-0.1264206022,-0.3185576797,-0.3657461107,0.1917575747,0.2100533098,0.0222781841,-0.3211946487,-0.379904449,-0.1829975396,0.071736224,0.2581097186,0.0427502915,0.222627148,0.0151273236,0.074371703,-0.1445606202,-0.1302337199,-0.0362577327,-0.2002604008,0.0194208343,0.1271392703,-0.0886328891,0.0229694098,0.0672450289,0.0520129241,0.1141588688,0.023423072,-0.0221919138,-0.2302063257,0.2198514044,0.1898466349,-0.1088619828,0.3095646799,-0.4991005957,-0.0709012151,-0.0725439265,-0.2607088685,0.3616882861,0.146986872,-0.0562004633,0.1453669965,0.2952059507,-0.0725108981,0.3919354975,-0.3460874259,-0.1231111586,0.0871474072,0.1882703453,0.0799198374,0.0217772312,-0.2326595485,-0.2026038319,-0.1023408175,0.0921725109,0.16762577,-0.1014085338,-0.1661898494,0.2122140676,-0.0144442283,0.4409593642,0.0101484992,-0.1732783467,0.5315682888,-0.0094009107,-0.0644601136,-0.052649349,0.1593038887,-0.1526360512,-0.0896263793,0.4959955812,0.2406762987,-0.1837686896,0.2814666629,0.0132143889,0.2747500539,-0.064692758,0.0044511063,0.0725781694,-0.0291589294,0.0261632279,-0.1802540123,0.0871636271,0.2104332298,-0.0309836678,-0.0112776151,-0.1122564748,-0.0845983922,-0.1000115499,-0.0228976887,-0.0171520319,-0.2293910682,-0.0004272533,0.5783107877,0.2522877753,0.1274656951,0.2180206925,-0.409868598,0.0148701798,0.1073543876,0.1376320273,-0.1755608618,-0.0841577053,-0.1308661252,-0.0059653907,-0.3466261327,0.076287277,-0.0426454544,-0.3841930926,0.3021792471,0.1633160412,0.2153522223,-0.3880352974,0.0622002818,0.1634439975,0.0764943212,-0.2189601064,0.1823263913,-0.4646622241,0.3215572834,-0.2943394482,-0.230681017,0.1723097116,0.3910271525,-0.0443320163,-0.1313524246,0.0344272591,0.1779143363,-0.273719281,-0.0303489063,-0.0306203701,0.7043981552,0.2260794938,-0.0646468103,-0.0352959819,-0.3021515608,0.1523119062,-0.2608647048,0.025533773,0.6669088006,0.3248124421,-0.0320298858,-0.0970158949,0.1974645406,-0.1992254108,0.063569881,0.1859070361,0.2536439598,0.1504038274,0.0420225635,0.0411132984,0.3362576962,0.3680169582,0.0272582453,0.1248697191,0.0767663196,-0.0560112409,-0.5223753452,0.2010111809,-0.2894560397,-0.3216926754,-0.0663250387,-0.1535455137,-0.0217733514,0.0549932159,-0.2688905001,0.212206319,0.1099134758,-0.1559635699,0.1555137932,-0.0372734666,0.1464596093,-0.0764593482,0.0325149298,-0.4078032076,-0.067325905,-0.1238289773,-0.0607972033,-0.1650110036,-0.1264009625,-0.1908155978,0.228384465,0.1445746124,-0.0110217677,0.0957019404,-0.0102241542,-0.4300072789,-0.3392841816,0.0320978463,-0.290073365,-0.0707346648,-0.1296246648,0.0680457056,-0.2968210876,-0.0652114749,-0.6669611931,0.3160361946,-0.0314049087,0.110165,0.3075265288,0.0258793961,-0.2131533474,0.3563804924,0.1680597067,0.0822218433,-0.0048588188,0.0684648156,-0.3529636264,-0.1970294267,0.4783398807,-0.05543302,-0.1137767881,0.1693719476,0.3043698072,-0.1784659922,-0.0747164786,-0.3926655054,-0.2959657311,0.1449327469,-0.162750721,0.05887172,0.0743249208,-0.3249977827,0.0506935008,-0.0538545474,0.415384084,-0.0118987951,-0.2158793211,0.0724133924,-0.3163650632]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/554","title":"nlp downloads to its module path","comments":"> Indeed this is a know issue with the fact that we try to be compatible with cloupickle.\r\n> \r\n> Does this also happen if you are installing in a virtual environment?\r\n\r\nThen it would work, because the package is in a writable path.","body":"I am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?","comment_length":44,"text":"nlp downloads to its module path\nI am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?\n> Indeed this is a know issue with the fact that we try to be compatible with cloupickle.\r\n> \r\n> Does this also happen if you are installing in a virtual environment?\r\n\r\nThen it would work, because the package is in a writable path.","embeddings":[0.0366535224,0.2192299068,0.1506813467,0.0508494563,0.2112177163,-0.1820767671,-0.0289928615,-0.0183257349,0.1189896837,-0.0385014191,0.1243856326,0.8661038876,-0.4588842094,0.1108861566,0.3196983039,0.3289014399,-0.1757936031,0.0106611783,-0.3965616226,-0.0296090506,-0.2183236629,0.3573447168,0.0127435839,0.1647962034,0.0456613861,-0.2805224657,0.1733621806,0.2096358538,-0.2175506502,-0.2872889638,0.2205725312,0.0743843615,0.019858567,0.2375182509,-0.0001259466,-0.140705511,0.1535024792,-0.2212117612,-0.5928941965,-0.4169439375,0.0931646377,-0.1434482336,0.3012827635,-0.2460523099,0.375099957,0.6685689688,0.1899013072,-0.0168130044,0.0882839486,0.1640619189,0.1206227168,0.3852365315,-0.1497676969,0.355763793,0.1676640064,0.1658295542,-0.2532820106,0.5958601236,-0.0119575001,-0.1434496194,0.1398182958,-0.0086906506,-0.1408176571,-0.2264499068,0.3995679915,0.0035117115,0.5514386296,-0.1675674915,-0.1826414019,0.0073816292,0.1411041468,-0.6322299242,-0.2780581415,-0.3192709088,0.0291315671,-0.1716597229,0.5286079645,0.3607387841,-0.1927578896,0.1301563531,0.0535914674,-0.33201316,-0.0955119804,0.363376677,0.0818298087,0.4446329474,-0.0106202653,0.0516017005,0.3404615521,0.2785791755,0.3875044584,0.1370323002,0.065361701,0.2994748056,0.2602428198,-0.0839847103,-0.0904842168,0.3794725537,-0.0536357053,0.0383315831,0.1583432108,-0.1037540361,-0.0878680497,0.2539205551,0.2582300901,-0.2878369093,-0.0004792726,0.1772267222,0.2212037593,0.1572444588,0.2616731524,0.063905254,-0.0639304668,-0.1136361361,-0.1016861498,-0.0943214074,0.0696439072,-0.1461735815,-0.0453581512,-0.1766080111,-0.1582485139,0.2957185507,0.2490972877,0.0295461416,-0.2259839177,0.0910730213,0.2279285491,0.0887091979,-0.2838964164,0.4456208944,0.0281991865,0.1882098019,-0.3258990943,0.0462065153,0.1481841803,0.1662860364,0.3029676676,-0.2558451593,-0.0446325429,-0.2242345065,0.018395761,0.2445895374,-0.2678960264,0.3420213163,0.0084259668,0.3626168072,-0.2094111145,-0.3191005588,-0.4718196988,0.1166787669,-0.3187381029,-0.3920495808,0.1970468313,0.021062497,-0.2289592326,-0.0609708056,0.0684654936,-0.6516927481,0.2548456192,-0.2134694755,0.2012834549,-0.295047462,-0.2088278681,-0.2876152992,-0.0944900587,0.4052164853,0.0979430377,-0.2253578752,-0.2361149788,-0.010029288,0.4378164709,0.182920441,-0.1644482464,-0.0694433227,-0.282744199,-0.2839719057,0.830616653,-0.3531689644,-0.3510722816,0.2237607986,-0.5753187537,-0.0287069101,0.0173475128,0.1781503856,-0.4840010703,-0.1129441932,-0.030246133,0.7501533628,0.2822034061,0.0799725801,-0.3088632822,-0.2825063765,0.2679076791,-0.1572669894,-0.0863526389,0.0594158471,-0.0044740187,0.5208021998,0.5502972603,0.2040090412,0.003794922,0.3402992189,0.5700147152,0.1585389078,-0.288918972,-0.1508489698,0.0248760767,-0.0060572233,-0.8670706153,0.1508536041,-0.5519198179,-0.2643413246,-0.1136568189,0.1670417935,0.1619913578,0.0302645303,-0.0453914404,0.2329876423,0.0791513547,0.1599001884,-0.108599849,0.4938775897,-0.2198509723,0.1199445203,-0.0938762575,0.1448790133,-0.1620064378,-0.2122442126,-0.0867494196,0.4839967489,0.1669929922,-0.1863855422,0.014285584,0.5541186333,-0.0953236669,0.1087051928,-0.3931443393,0.0577358343,0.3354277015,0.0622458309,0.0238070022,0.2601512969,-0.0441208966,0.0759531558,-0.156748414,0.111864455,-0.3385421932,0.0956282839,0.1660507619,0.2892729938,0.0081913136,0.0418537594,-0.1756075025,-0.0711489022,0.2505498528,0.2442206889,0.4438206255,0.3397476673,0.0398138948,-0.0256215241,0.3678896725,0.1768845469,0.1766762137,0.029185148,0.0396356061,-0.0324893668,0.0751423538,0.2753876448,0.1064353511,-0.0058809309,0.0884655491,0.2145641297,-0.2644822896,-0.1744257212,0.1345820725,0.0081139337,0.2572434545,0.1063861623,0.1480322033,-0.1938437819,-0.1374715269,0.0872050896,0.0261675101,0.1036969721,-0.3324563205,0.0644233748,-0.557071507,-0.3485966921,-0.6488167644,0.1673888862,-0.5379951,-0.2224393785,-0.1973872185,0.4104060829,0.0084693497,-0.0865177736,-0.0707248375,-0.0870466828,-0.4106627405,-0.4929046035,-0.0081484821,0.0776349828,-0.3279862106,-0.1643287987,0.2320948243,0.1121879742,0.1739886105,-0.0162180196,-0.4475990832,0.0440152287,-0.254773438,0.0593611263,0.3497127593,0.1541694999,0.3018980026,-0.093814455,0.0188857596,-0.2698601782,0.1019585282,-0.1533108056,0.1081808209,0.0174670145,0.1186528802,0.0067654438,-0.3135100007,-0.1002925336,-0.4666461647,-0.445424825,0.3380752206,0.2953378856,0.3194355369,0.824881196,-0.1670692712,-0.0477720015,-0.1108454391,0.1610222608,0.2100711912,0.0273991562,0.221004039,-0.2460757196,-0.138740018,-0.1334884018,0.0053342599,0.0291796364,0.3621094823,-0.2343441099,-0.0264554434,-0.0387189537,0.1095168814,0.1662641764,0.0689311698,0.534698844,0.2771008611,0.2051319778,-0.071721673,-0.2466054112,0.2251665592,0.5450434089,0.4341472983,0.3294350207,0.14637582,0.2293754816,0.2922331989,-0.0532068275,0.2080352902,0.30007568,0.0098279063,0.4775596261,0.0632710755,-0.1031333208,-0.1840748787,0.0382701196,-0.1106667742,0.0486713164,0.1554290205,-0.2069404125,-0.112356253,0.0205846895,-0.0134232612,-0.3826886415,-0.0242390651,0.1152460426,0.2954922616,0.1438838392,-0.0172862597,-0.148543328,-0.2984665334,0.2888365984,0.3253398538,-0.0598448291,0.3793082535,-0.2952449918,0.0804322064,-0.3175659776,-0.3601869047,-0.110040307,0.3189067245,-0.0847240761,-0.0077190404,0.1835466474,-0.1647548079,0.1903630793,-0.0420522057,0.2470883876,0.184795171,-0.0830094218,-0.518452704,-0.0562215857,0.0118575981,0.3997411132,0.3311586976,-0.0270577911,-0.0743688196,-0.0558985844,0.0720927268,0.1218219697,-0.1918270439,-0.1297147423,-0.2415909767,-0.4315784574,-0.2292883247,0.0169676561,0.0482570417,0.2081787139,0.2774986923,-0.1133425832,0.17753537,-0.1554814428,0.0177854989,-0.2938522995,0.2565215826,0.0394327305,0.1155607849,-0.2399397045,0.1914774626,-0.2706723809,0.3414493501,-0.0520600192,0.2024663836,-0.2684369683,-0.1000598148,0.3315455616,0.5321868062,-0.0422708243,-0.0014328217,-0.0665892437,-0.4235523939,0.0184181947,0.2030992657,-0.1293684542,-0.2833474874,0.0891486406,-0.2291783839,0.2372550666,0.1529622078,-0.2294103801,0.373108089,-0.1355731636,-0.4136408269,0.1329635978,0.2320306748,1.2185707092,-0.1927911192,0.2424693257,0.1323702335,-0.3454881608,0.2222171128,-0.2093658149,0.1942041516,-0.4596819282,0.1864029914,0.0184311699,-0.334203124,-0.050121922,0.0156164737,0.0858067572,0.2734682858,-0.0548892543,-0.2729023695,0.1304855347,0.3938907385,-0.1766320169,-0.2877826691,-0.2902982831,-0.01443935,-0.070849739,0.507183075,-0.1328954846,-0.1240093112,0.1772804409,-0.2180301398,0.0646021068,0.1719403565,-0.4313856959,0.0417577103,-0.0812024996,-0.0232601464,-0.0917146429,0.1257975399,-0.0543595962,0.62542063,0.0450867414,-0.2213436216,-0.0199678056,-0.3359385133,0.2360227108,0.4263322055,0.037567921,-0.0540942587,0.0606901832,-0.4100448787,-0.0237482041,-0.3558067381,0.2627709806,-0.0814676955,0.2663211823,0.1987922192,-0.3775652647,-0.0022953146,-0.0742054209,-0.0449427962,0.0188488476,0.1913132519,-0.0827542618,0.1424546838,-0.1317510903,-0.3155436218,-0.0179329254,0.4924048781,-0.0810655281,-0.0276896525,0.2180949897,-0.2610344887,-0.2072010934,-0.1047283337,-0.1250659823,-0.4378283918,-0.0794605091,-0.0168344826,-0.1192844734,-0.3126403987,-0.3789984286,0.2657550871,0.2184222341,0.0590272024,-0.3117892146,-0.3809970915,-0.2000356317,0.0596004464,0.2855507731,0.0585901923,0.2201538086,0.0488655679,0.075291954,-0.1427547038,-0.1186264083,-0.0432412215,-0.1961247176,0.013206521,0.1663921475,-0.0661026388,0.0380035229,0.0921193883,0.0476537533,0.1081354767,0.0495403334,-0.0159849692,-0.2462609857,0.2203298658,0.1938226223,-0.1061006486,0.3261564076,-0.5064128041,-0.0603542849,-0.0794403329,-0.2599948645,0.3589471579,0.1483557373,-0.0642248094,0.1416981667,0.2805728316,-0.0697368979,0.3821462095,-0.3312809765,-0.1311706156,0.0909305066,0.1834527403,0.1079597548,0.0135394232,-0.2332727164,-0.2288060635,-0.1173919141,0.0736956969,0.1737177372,-0.0846077204,-0.1903154701,0.2157856375,-0.0114631942,0.4147981405,0.0116175888,-0.1680883914,0.5315957665,-0.0207544994,-0.0809613019,-0.0303738043,0.1509318352,-0.135805577,-0.0864589587,0.499910295,0.2174673229,-0.1777748168,0.2785038352,-0.007734518,0.2855094671,-0.0789785683,0.0152459182,0.0686137453,-0.0156895518,0.0382688902,-0.1985386461,0.0799785256,0.197394371,-0.0911445841,-0.0596216992,-0.1054480597,-0.096978873,-0.1131463274,0.0123385433,-0.0061766189,-0.1821185052,0.0117927408,0.5577607155,0.2451309711,0.1157423705,0.260414809,-0.4050930738,0.0172739122,0.0973192826,0.1587707251,-0.1763161868,-0.0776242167,-0.1327058375,0.0050082388,-0.3201335371,0.0667285994,-0.0235008914,-0.3816544712,0.3042977452,0.1711448282,0.2156491429,-0.3642803133,0.0291381404,0.1577450782,0.0846864581,-0.2103460878,0.1931721568,-0.4692875743,0.3293414712,-0.3130377829,-0.2359605283,0.1507844031,0.3931719065,-0.0258135982,-0.1243555769,0.0420358405,0.167362541,-0.2569442689,-0.049286291,-0.0278110318,0.6900135875,0.2152639031,-0.0723849759,-0.0323289558,-0.303787291,0.1399613172,-0.2608659565,0.0227423403,0.6522600055,0.3411045074,-0.0329658426,-0.1084382012,0.1946373135,-0.1974798739,0.0580817163,0.187431708,0.2687219679,0.1433303654,0.045193132,0.0345933661,0.3350121379,0.3339106441,0.0289559606,0.1249742284,0.0653203651,-0.0672075972,-0.5001581311,0.2113154978,-0.2728497088,-0.3391133547,-0.0894578919,-0.1681342125,-0.0510367565,0.0347784981,-0.2608060539,0.2128754854,0.1028019711,-0.1845181137,0.1697342247,-0.0235912018,0.1486403495,-0.0666163936,0.0496670827,-0.3975166678,-0.0580948256,-0.1013715267,-0.0718592107,-0.1895344555,-0.1180964783,-0.2055093497,0.2318049967,0.1495648026,-0.0137110893,0.12013226,-0.0202559177,-0.4006500542,-0.3265463412,0.0394963063,-0.2855172157,-0.0721657202,-0.121519044,0.0657677799,-0.2892018855,-0.0710093081,-0.6805310845,0.3385013044,-0.0133686895,0.0413125455,0.3364284635,0.0010193314,-0.2102988362,0.3626990616,0.1839002967,0.0758148059,-0.0131362006,0.0722300038,-0.3415880799,-0.2063355446,0.4868858457,-0.0624346249,-0.1025512367,0.1755742133,0.3126603961,-0.2155442089,-0.0865192041,-0.3908149302,-0.3085576296,0.1468036473,-0.1520264894,0.0587397031,0.0959872752,-0.3098630011,0.0676000267,-0.0619198717,0.416359961,0.003358959,-0.213535741,0.0681846663,-0.3184007108]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/554","title":"nlp downloads to its module path","comments":"If it's fine for you then this is the recommended way to solve this issue.","body":"I am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?","comment_length":15,"text":"nlp downloads to its module path\nI am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?\nIf it's fine for you then this is the recommended way to solve this issue.","embeddings":[0.0456970446,0.2198795229,0.1330279857,0.0576157607,0.2453001887,-0.0993736237,-0.0618805103,0.0330491066,0.0334914885,-0.0734839961,0.1636885852,0.9088417888,-0.3889652193,-0.0022776662,0.4177902341,0.3236100376,-0.1697185785,-0.0225905124,-0.5341246724,-0.0169948544,-0.1622969657,0.3050960004,0.0772611424,0.2103696913,0.115923956,-0.2541028559,0.1606364846,0.215590626,-0.1639485955,-0.3389035463,0.2224670351,0.0083092237,0.0655623823,0.2705921829,-0.0001232511,-0.1123505458,0.0861978605,-0.2068321407,-0.5517083406,-0.4239936471,-0.0148001676,-0.0849360824,0.2973421812,-0.2733477354,0.4307845533,0.4361686409,0.2285035402,0.024311211,0.1034945324,0.234602809,0.148006916,0.2664437294,-0.0894249678,0.343288213,0.266503185,0.2373018414,-0.1961114258,0.4667908251,-0.1984425485,-0.1247534677,0.1270865351,0.0056848656,-0.1805005968,-0.2069909573,0.4597918689,0.0496077649,0.5410174131,-0.1459678113,-0.1784220785,-0.0178554058,0.2908133566,-0.642926991,-0.2781715393,-0.3530682623,0.0910798311,0.0200187806,0.4550758898,0.3568699062,-0.2100289315,0.154193908,0.1279545575,-0.4636644721,-0.181461215,0.3170287907,0.1041276902,0.5098668933,-0.0027671116,0.0066675376,0.3294624388,0.2134521306,0.3141112626,0.1163369492,0.0860578641,0.2740933597,0.2820371985,-0.0066126664,-0.1560246199,0.1707146764,-0.0778533965,-0.1184973493,0.005213452,-0.1214013249,-0.1242593378,0.175099194,0.1527309269,-0.2403070629,-0.0165108424,0.1584413648,0.1670554429,0.1226088256,0.2449633926,0.0753710195,-0.099551551,-0.0186606813,-0.0943079591,-0.1567019075,0.0298855547,-0.1276332438,-0.077237986,-0.2306555808,-0.1844921112,0.29718256,0.1867804378,0.0180612132,-0.0898969024,0.1105458215,0.2327719927,0.0169168767,-0.2663823664,0.4370716214,0.0268502049,0.1493208557,-0.2784195244,0.0691128522,0.113911204,0.0878558457,0.3548499048,-0.2265578359,-0.1389373243,-0.1703401953,-0.0085247327,0.2640251517,-0.1764732599,0.3757030368,0.080115132,0.4583417177,-0.2369888574,-0.1655579209,-0.401953578,0.1424208134,-0.2953974605,-0.3840682507,0.1258054227,0.0609360076,-0.2546836436,-0.1480506808,0.1975680739,-0.6603124142,0.2777021229,-0.1704701781,0.2110695541,-0.2004638612,-0.1851413548,-0.3039751053,-0.0456876419,0.4764195383,0.2460522652,-0.2469320446,-0.24819085,0.0314670093,0.3878192902,0.1698340476,-0.0847766548,0.1193863004,-0.2714425921,-0.2908033431,0.9045638442,-0.3688507378,-0.3031209409,0.3059940636,-0.6601226926,-0.0448579788,0.1377703547,0.1706227809,-0.4674173295,-0.0175080728,0.0730022341,0.7036175728,0.2801552117,0.0956122801,-0.2633234859,-0.3003543913,0.2514546514,-0.1232614964,-0.0369932204,0.0041029491,-0.1603769362,0.6015458107,0.5774344206,0.1854399741,0.0041121887,0.3577729166,0.5448876023,0.2054187208,-0.2792643309,-0.2789790034,0.0539979115,-0.0863599032,-0.6585479379,0.1137281433,-0.5175483823,-0.1787605882,-0.1170039102,0.1333517581,0.1825945973,0.067067273,-0.0122842267,0.2413033098,-0.0135435704,0.2098301053,-0.101972267,0.4941106439,-0.1045441478,0.1626832485,-0.2112010866,0.01906294,-0.1806180328,-0.1541095674,-0.0723478794,0.4531132281,0.1245922595,-0.1282081008,0.0420721546,0.4814319313,-0.2115142792,0.0788517818,-0.2875140309,0.0465633087,0.2747341096,0.0362552032,0.0267068092,0.2531223595,-0.0366810076,0.1950973123,-0.2124860287,0.0844998881,-0.5008102059,0.0955987647,0.1026367769,0.2530409992,0.0514631569,0.0589089952,-0.1706527174,-0.1424614638,0.2965351045,0.0837095603,0.2630125284,0.3538713455,0.1563922316,-0.020731397,0.4070531428,0.1956209838,0.191443041,0.04286981,0.0112981237,-0.0135624073,0.018914884,0.3087257743,0.1408744901,0.0657754317,0.1651693881,0.1302620471,-0.2708735466,-0.1907940358,0.2119492739,-0.0162467565,0.1602199525,0.1799028963,0.097437948,-0.2070076168,-0.1689727604,0.1133483574,0.0608205572,0.1836171299,-0.33942312,0.0510006174,-0.5935868621,-0.3907923698,-0.7110666037,0.037933778,-0.5656552315,-0.2330485582,-0.1597855091,0.2962647676,0.0249872878,-0.1106696278,-0.0746926367,-0.2433951944,-0.4009862542,-0.5734475255,0.0617551729,0.0969658941,-0.2814296186,-0.1439194679,0.1710545868,0.2177450359,0.1586261094,-0.1363528222,-0.386477977,0.0480798669,-0.1734461039,0.0060339919,0.3010382056,0.1417226344,0.2841223478,-0.091314517,0.2047014385,-0.1859254688,0.0481789336,-0.228292048,0.1044923291,0.0083979126,0.1126394868,0.0613930933,-0.3540355563,-0.1800745279,-0.2471634895,-0.4988822639,0.2726937532,0.222217381,0.3294353187,0.7713518739,-0.0952416509,0.0119741159,-0.2259766608,0.1741758734,0.2116490752,-0.0090330727,0.2586947381,-0.1175372526,-0.0637982562,-0.0302557312,0.0894203931,0.0423465483,0.4483211637,-0.2921323478,-0.0350470468,0.0425993949,0.0514859557,0.0139471516,0.1357073337,0.4397869408,0.2903017998,0.1430885196,-0.0312579535,-0.2862818837,0.21901761,0.567012012,0.3764134347,0.3127657473,0.2143055499,0.1781481057,0.3430245519,-0.0284148976,0.2479142696,0.2442087531,-0.0003234924,0.5273303986,0.186770767,-0.1364533305,-0.3223749697,0.137450859,-0.0484680049,0.0627109408,0.1228740364,-0.103636317,-0.042507194,-0.1285205781,-0.0395408385,-0.4146597385,0.083072342,-0.1149965823,0.3773278296,0.1543293446,0.0765544623,-0.2605673373,-0.2827227712,0.2984002233,0.3421392143,-0.0011311391,0.355271101,-0.2587130964,0.0276148655,-0.3369044363,-0.3431702256,-0.1087240428,0.4992229044,-0.1390259862,0.0123777147,0.2169144154,-0.1580486745,0.3271794319,0.0402868465,0.2686864436,0.1256740391,-0.1135066971,-0.5608112216,-0.0047471635,0.0803990215,0.4261993468,0.1998753548,0.0197881218,-0.0617796071,-0.1569174826,0.0660890415,0.0395535603,-0.2014623731,-0.2333533019,-0.2762470543,-0.4220377803,-0.2192889303,-0.0132892523,0.0533251241,0.0956818983,0.3092010617,-0.036390081,0.1939858049,-0.0716445595,-0.0535451882,-0.2910580933,0.1496711075,-0.0201249067,0.0750904754,-0.0155829042,0.0130528985,-0.2452145964,0.3458244503,-0.0562756695,0.1639468223,-0.3325324059,-0.0741356611,0.279383719,0.4921968281,-0.0107336976,0.0987110883,-0.0386388265,-0.4512882531,0.0407353342,0.0797198713,-0.0383371264,-0.314784348,0.2086417377,-0.3498779535,0.1923879534,0.1596284062,-0.223557204,0.2690446675,-0.1165081039,-0.4240179956,0.2375945896,0.2285971493,1.0424617529,-0.2574047148,0.2977075279,0.0459924787,-0.420176357,0.3538000882,-0.2609346509,0.2716051042,-0.455794543,0.2838845849,0.0053345095,-0.340769887,-0.14362818,-0.0919755548,0.1232366189,0.2716785371,0.099196583,-0.0758423954,0.1716447175,0.341688931,-0.1615092456,-0.2997382581,-0.3987140656,-0.0007041405,-0.1077941507,0.426487267,-0.0939497054,-0.2394241691,0.1695456207,-0.1685712188,0.0342485867,0.1492774487,-0.1472021192,-0.0380650051,-0.0214800332,0.0143170226,0.0093234256,0.201551035,-0.1073320061,0.5296409726,0.0582509227,-0.2741683424,-0.2537510097,-0.289816618,0.0962789878,0.4569170177,-0.0292652976,-0.0181855373,-0.0026280531,-0.3400768936,0.0205047168,-0.2476282716,0.2225717306,-0.0037762085,0.2904999256,0.2086617947,-0.1985602081,-0.0134211946,0.0079634478,-0.0590843111,0.0345073789,0.2139196396,-0.1648346931,0.1604242027,-0.1242000461,-0.3431697786,0.0752575472,0.5427678823,-0.07378041,0.1747149527,0.3054579794,-0.307828486,-0.1662594527,-0.1395645291,-0.0570528992,-0.6076746583,-0.112245664,-0.0983480886,-0.1259804517,-0.3058276772,-0.2966697514,-0.0808568895,0.1902513355,-0.0221330747,-0.3743106127,-0.3863301873,-0.1754004508,0.0660792366,0.1239999011,0.0203877371,0.2021502256,-0.0533307791,0.0813578144,-0.1209565997,-0.1392894983,-0.0207172018,-0.259252876,0.0485611893,0.0330525488,-0.1571983695,-0.0284792464,0.0261264294,0.0567316562,0.1544045657,-0.0211543851,-0.0197581369,-0.2072848827,0.2216863781,0.1483040005,-0.0777833834,0.2621794939,-0.4664173126,-0.0840954259,-0.0185536314,-0.2682810128,0.3110802174,0.1818186194,0.0034045356,0.1886448562,0.2731965184,-0.1343635917,0.4006821513,-0.4106796384,-0.1736482382,0.037620414,0.1838478744,-0.0011923912,0.0375297628,-0.3261893094,-0.2109837681,0.0090062767,0.1013555676,0.1518164575,-0.1277240813,-0.1144634113,0.2225053757,-0.0724136904,0.4809968174,0.0414337553,-0.1515541673,0.518280983,0.0086917812,-0.0037865452,-0.0856490806,0.188626647,-0.2147376835,-0.116630353,0.490335077,0.2236653566,-0.1959444433,0.2941580415,0.0378090963,0.3125507832,0.0300271083,-0.0830405354,0.1131492853,-0.0781269222,-0.0035307191,-0.0880292729,0.1135217845,0.2326314449,0.17474325,0.1302459389,-0.1373953968,-0.011856189,-0.1405991167,-0.0688528344,-0.0179467667,-0.3787089884,-0.0136349276,0.6411446333,0.2231981754,0.1396004707,0.0995074287,-0.3772587478,0.0074626766,0.2150182426,0.1076058373,-0.1582237482,-0.0823406503,-0.1276501715,0.0096934168,-0.3812961876,0.0734745562,-0.044588577,-0.4292425811,0.3016229272,0.1419322193,0.1501521319,-0.4597578645,0.0719524249,0.2341798842,0.0053280871,-0.2607336938,0.1430935413,-0.4472833276,0.3017169237,-0.2727394998,-0.2731342614,0.1973454356,0.3301753104,-0.1082480922,-0.1058890969,0.0760612115,0.1924533099,-0.2649707198,-0.0006441565,-0.0838015899,0.7792975903,0.1702766567,-0.0565587506,-0.027538389,-0.3019994497,0.1790735424,-0.2257461995,0.0285292044,0.6861563325,0.2728618383,-0.0443648398,-0.1223117635,0.1910347342,-0.2395915687,0.0919959024,0.1993154734,0.1808580905,0.122345522,0.0167048741,0.0403666385,0.2590352297,0.4418447018,0.0534510463,0.1397840679,0.0687002093,0.0058006048,-0.516361773,0.1673299074,-0.3077678382,-0.2685191929,-0.025481414,-0.1394604743,0.0067782444,0.0480157994,-0.2367467731,0.1812340021,0.1115838811,-0.0226234701,0.1221251637,-0.0881153867,0.1404328644,-0.0390868001,0.0191259868,-0.4098718166,-0.0251322798,-0.1634407043,-0.0496990904,-0.1162369326,-0.1644612104,-0.1083747372,0.2024492919,0.1680300385,-0.0431594141,0.0067169536,0.0016484803,-0.4647762775,-0.3578769267,0.0641016513,-0.3016726971,-0.1293732226,-0.0791570321,0.0803886726,-0.337318182,-0.0587875061,-0.6380975842,0.2748585045,-0.0884445459,0.2332716882,0.2388482541,0.0650880635,-0.2396909148,0.2882930636,0.1480465233,0.0977215394,0.0497382767,0.0467106737,-0.4029242098,-0.1428598315,0.4646067619,-0.0074647726,-0.1955696344,0.1968471855,0.2791340649,-0.0200119298,-0.0709803328,-0.4080962539,-0.2064034492,0.2197978944,-0.169077009,0.0961122662,0.0229503047,-0.3149964213,0.0802039504,0.0203058869,0.3826418221,-0.0619480684,-0.1994216144,0.1121303365,-0.3041766882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/554","title":"nlp downloads to its module path","comments":"> If it's fine for you then this is the recommended way to solve this issue.\r\n\r\nI don't want to use a virtual environment, because Nix is fully reproducible, and virtual environments are not. And I am the maintainer of the `transformers` in nixpkgs, so sooner or later I will have to package `nlp`, since it is becoming a dependency of `transformers` ;).","body":"I am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?","comment_length":63,"text":"nlp downloads to its module path\nI am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?\n> If it's fine for you then this is the recommended way to solve this issue.\r\n\r\nI don't want to use a virtual environment, because Nix is fully reproducible, and virtual environments are not. And I am the maintainer of the `transformers` in nixpkgs, so sooner or later I will have to package `nlp`, since it is becoming a dependency of `transformers` ;).","embeddings":[0.046116367,0.234247297,0.1267920136,0.0296455473,0.2665544748,-0.1070262492,-0.0495609231,0.0400402658,0.0180516001,-0.0675463974,0.1628862768,0.9074655771,-0.3907279372,-0.0176893678,0.4088222384,0.325348407,-0.1871856004,-0.0095229726,-0.5038885474,-0.0106999353,-0.1629397571,0.3172166646,0.0928088799,0.2191578299,0.1168668419,-0.2506914735,0.1828442365,0.2161523998,-0.1610955298,-0.3452356458,0.2223353237,0.0172989201,0.0685411766,0.2536882162,-0.0001222753,-0.1021162271,0.0917157605,-0.2052916735,-0.5588748455,-0.4225788116,-0.0087841656,-0.0983675718,0.2872693837,-0.2772430182,0.4104694128,0.4269267917,0.2175244987,-0.0110216606,0.1032749563,0.2261535525,0.1559865922,0.2630299628,-0.0832814127,0.3464424908,0.2658575475,0.2363731116,-0.1880696565,0.4692555666,-0.195657894,-0.1145167723,0.1284811795,0.0002660673,-0.188188374,-0.2027669996,0.4447340071,0.0490654148,0.5470705628,-0.1268076301,-0.1714593172,-0.0210750327,0.2907862067,-0.645606935,-0.2828428447,-0.3490971625,0.0826938525,0.0141357509,0.4553363621,0.3532887399,-0.2175490409,0.1618506908,0.1237481236,-0.4460406899,-0.1745378077,0.3221803606,0.1006399542,0.5085071921,0.0016213604,0.00436628,0.3273635209,0.2121271938,0.310200572,0.1156195328,0.0912705809,0.2791387737,0.2880519331,-0.0127185676,-0.1458795518,0.1641432941,-0.0892470777,-0.1161925867,0.006875712,-0.1094334498,-0.1109065935,0.1684023142,0.1201752499,-0.2382375151,-0.0224934928,0.1595935076,0.1828632057,0.1355985999,0.2601260841,0.0825105309,-0.0887886435,-0.033735089,-0.1104378924,-0.1315760463,0.0319398306,-0.123421289,-0.087278612,-0.2416331768,-0.1768742949,0.3003259599,0.1871353984,-0.000803103,-0.0788418576,0.1259497851,0.23518911,0.0070649926,-0.272452414,0.4379551709,0.0300905649,0.1560270637,-0.2845471203,0.078573294,0.131322071,0.079952769,0.3415975273,-0.2330340147,-0.1420311928,-0.1482820064,-0.0054023978,0.2611575723,-0.1751811057,0.3640359342,0.0621220618,0.4477259517,-0.2313425243,-0.1666173935,-0.4070670605,0.1336168498,-0.284452945,-0.402718544,0.1284139901,0.0695235208,-0.2449092269,-0.1395531148,0.1998022348,-0.6429062486,0.2645700574,-0.179700911,0.2135895789,-0.182408303,-0.1748671234,-0.3008734882,-0.0477913022,0.4714634418,0.229443565,-0.2523702681,-0.2462406605,0.0334514081,0.3789203465,0.1632091105,-0.1067515537,0.1069479734,-0.2686387002,-0.2919618189,0.9047156572,-0.3906837404,-0.3091709912,0.2915870249,-0.6707193851,-0.0479654893,0.1383062303,0.1629870683,-0.471446991,-0.0274625439,0.072186172,0.7123209834,0.2875203788,0.1052506864,-0.2627530098,-0.3227465451,0.222829774,-0.1149718761,-0.038980417,-0.0065271291,-0.1487011313,0.597994864,0.5738622546,0.1660397798,-0.0016501561,0.3604936898,0.5375530124,0.1958610117,-0.27916044,-0.2807016671,0.0373023115,-0.0844608471,-0.6614479423,0.107897684,-0.5330183506,-0.1834053397,-0.1122304201,0.1228865534,0.1824475229,0.0702087656,-0.0002824996,0.2549110651,0.0022279522,0.2083616406,-0.0960031152,0.4704718292,-0.1031578407,0.1595535725,-0.1962206364,0.0275265183,-0.1773788333,-0.1482795328,-0.0782647207,0.4618581235,0.12504071,-0.1315963715,0.0474949852,0.4746610522,-0.2096363902,0.0889927298,-0.2957380414,0.0512470976,0.2695703208,0.0370287374,0.0307874102,0.2559651434,-0.0543590561,0.2016519308,-0.2219351679,0.0880995989,-0.4998498559,0.0931912512,0.1127002537,0.2424832135,0.063227728,0.0523401685,-0.1859665513,-0.1516279876,0.2849550843,0.100478597,0.2885392606,0.3445744216,0.161164254,-0.0273502786,0.3952870071,0.1889089197,0.1825788766,0.0537530147,0.0252611041,-0.014876904,0.0243096557,0.3026709259,0.1458862126,0.0798739642,0.1653839201,0.1257158965,-0.2704928219,-0.1986791939,0.2159199417,-0.0208988097,0.1586565226,0.1924609542,0.1183894873,-0.2064065486,-0.1669658422,0.1002587825,0.0627789795,0.168289572,-0.3316838145,0.0543968603,-0.595418334,-0.3879919648,-0.7191829681,0.0387097411,-0.5594838858,-0.2267316431,-0.1493154168,0.2969737053,0.0152958473,-0.0988679603,-0.0740788728,-0.2233131528,-0.4157322645,-0.6021327972,0.0733015463,0.0929680392,-0.287340343,-0.1341560185,0.1705031097,0.1962793618,0.1610697657,-0.1474251747,-0.3979492188,0.0391923487,-0.1772091687,0.0086356094,0.3191351891,0.1386234015,0.2791619599,-0.1118232235,0.2117307484,-0.199268803,0.0509133562,-0.2132970989,0.1117410809,-0.0033845371,0.1127620935,0.0831770897,-0.344376564,-0.1749547869,-0.253012538,-0.5123717785,0.2787902653,0.216633141,0.3217908144,0.7720021605,-0.102707155,0.0091882246,-0.2122785002,0.1872679591,0.2074654847,0.0008651846,0.2596406043,-0.1210708171,-0.0713396743,-0.0554444529,0.0765059814,0.0564994514,0.4624179602,-0.2822106481,-0.033298634,0.0441931263,0.0560681969,0.0114322631,0.1246184781,0.4576794505,0.2989203334,0.1324489564,-0.0288429037,-0.2807248831,0.2095328569,0.5572180748,0.3679825366,0.3192533255,0.2056295127,0.1678475142,0.3601109385,-0.0422435254,0.2414954007,0.2436206937,0.0004374397,0.529665947,0.1801270843,-0.1364840865,-0.3310603499,0.1352423877,-0.0557458214,0.0593938008,0.1209199876,-0.1072509959,-0.0436491817,-0.116557084,-0.0315997824,-0.4213882089,0.0777139813,-0.1087860689,0.3773988783,0.154431507,0.0712340996,-0.2669484615,-0.2816902399,0.2958913445,0.3518319428,-0.014598893,0.3535113633,-0.2625626326,0.0431502983,-0.3250785172,-0.326254338,-0.0907096863,0.4920764565,-0.1388789117,0.012444092,0.205597043,-0.1642439365,0.312672317,0.0556151532,0.2724235654,0.1352795511,-0.110595569,-0.5586287975,-0.0161936451,0.0890263766,0.4203826189,0.1868071705,0.0291638169,-0.0691932067,-0.1596040875,0.0529489741,0.0410507098,-0.189956516,-0.2359449565,-0.2774136066,-0.4206409156,-0.2196006626,-0.0223434251,0.0532167032,0.1078637168,0.3100388944,-0.0294070691,0.1922949553,-0.0837131888,-0.0561042167,-0.2963161767,0.1525681019,-0.0145515911,0.0554600134,-0.0099012936,0.0194986425,-0.2392356247,0.34317711,-0.0608595088,0.1602139324,-0.3263559341,-0.0778776109,0.2681139112,0.5132370591,-0.0173178166,0.0941148773,-0.044365339,-0.4583930969,0.0289774165,0.0902791768,-0.0448994711,-0.3154542744,0.2143292725,-0.3476989269,0.1942738742,0.1721813381,-0.2360344231,0.2812296152,-0.1251872927,-0.4402067959,0.2245350182,0.2467519343,1.0411486626,-0.2406390905,0.3079272807,0.0620334484,-0.424690187,0.3527469337,-0.2589766085,0.2856745124,-0.4480985105,0.2983219624,0.01414782,-0.3479577005,-0.1444007456,-0.0763517693,0.1431494206,0.2755274177,0.1016653478,-0.09200982,0.163508743,0.3466672301,-0.1620665938,-0.3068075776,-0.4199129045,0.0050213635,-0.1081741229,0.4152198136,-0.096290119,-0.2268120795,0.1657084376,-0.1818028986,0.0427781492,0.1374816149,-0.1462484449,-0.0159092918,-0.0259536915,0.0052117859,0.0082688238,0.1796968132,-0.116181545,0.5233233571,0.0573027506,-0.2882836163,-0.2484291047,-0.2790073454,0.0953644291,0.4519754648,-0.0233347379,-0.0182631742,0.009847654,-0.3462279737,0.0162903089,-0.2407739758,0.2226673216,-0.0171753224,0.2993600368,0.1988860965,-0.189571619,-0.0237381067,0.0130642438,-0.0495463386,0.0436640754,0.188868016,-0.1576395035,0.1557563394,-0.1305058002,-0.334508568,0.0824721083,0.5543608665,-0.0746291354,0.1600217521,0.3207663894,-0.3217406869,-0.1697060913,-0.1431185901,-0.055456277,-0.601500988,-0.1151209548,-0.0963397473,-0.1213515699,-0.3210611641,-0.291046977,-0.07755889,0.1969454139,-0.0023167764,-0.3674696982,-0.3863797486,-0.1825271249,0.0739893094,0.1213882193,0.0162813403,0.2047555149,-0.0354256853,0.0652357861,-0.1276288629,-0.1498263329,-0.0104739787,-0.2554082274,0.0443133339,0.0218519233,-0.1618617326,-0.0238606706,0.0383478403,0.0634287447,0.1393598914,-0.0296473317,-0.0312707052,-0.2038730085,0.2177937329,0.1383657455,-0.0901478007,0.2828359902,-0.4831421077,-0.0757835135,-0.0179250818,-0.2721808553,0.3135299385,0.1926708519,-0.0050087501,0.1831313819,0.2811107039,-0.1277559102,0.4007119536,-0.4252769947,-0.1594098508,0.0407718308,0.2038413882,-0.0085452357,0.0371544994,-0.3193021119,-0.2105541229,0.0072943857,0.1087435037,0.1364179999,-0.1161944196,-0.1163043752,0.2203354388,-0.0702840388,0.463568449,0.0348291397,-0.1621770561,0.5252125859,0.0169345904,-0.0106113181,-0.0934469402,0.1800398231,-0.2295191735,-0.0997845083,0.4896005094,0.2366866618,-0.2012722343,0.298592329,0.0290996637,0.3146234751,0.036692366,-0.0767072067,0.1179007515,-0.0764643177,-0.0103188986,-0.0833898261,0.1200806573,0.2478801459,0.1602066308,0.111409232,-0.1465801001,-0.0346585065,-0.1222596318,-0.0823650435,-0.0275228322,-0.3606378436,-0.0261692964,0.6504030824,0.2208254635,0.1487555802,0.0993580222,-0.3846209943,0.0283917505,0.2322043926,0.1140445396,-0.1544247717,-0.0870945454,-0.1269276887,0.0022708324,-0.3815245032,0.0682170764,-0.0539320409,-0.4044136405,0.2946982682,0.1397668719,0.1493916661,-0.441901952,0.0937094539,0.2336538732,0.0106891412,-0.2599323392,0.1499800533,-0.4215279222,0.3116586208,-0.2540074885,-0.2645650506,0.1834292114,0.3263542354,-0.1018268168,-0.1180417091,0.0643484369,0.1971223205,-0.2784435451,-0.0048409891,-0.0827920437,0.7673559785,0.1852514744,-0.0460717045,-0.0333991759,-0.3045271933,0.1852636635,-0.2388017327,0.0338929854,0.6955245733,0.2933859229,-0.0328547359,-0.1110075489,0.1918168366,-0.2263256013,0.0878710896,0.1948159635,0.1879599392,0.1277544498,0.039190203,0.0471434854,0.2659832239,0.4307887852,0.0303577445,0.1286846101,0.0742659494,0.0063470788,-0.5193608999,0.159716785,-0.3004153073,-0.2859917581,-0.022470152,-0.1499352306,0.027986899,0.0471543334,-0.2507306337,0.1889437586,0.1043841764,-0.0416014902,0.1218470857,-0.0788706169,0.1488678902,-0.0447730683,0.0158178415,-0.4161309302,-0.0405766591,-0.1405721456,-0.0442634858,-0.1302486658,-0.1614128798,-0.1218257472,0.1957513392,0.1491968334,-0.0281863715,0.0009697044,0.0020336218,-0.4676629007,-0.3582865894,0.0724993423,-0.312836647,-0.1168833598,-0.0736809447,0.0873701349,-0.3420382142,-0.071032919,-0.6486016512,0.287296474,-0.0913354531,0.2416810393,0.2341255844,0.0768069476,-0.2380164415,0.2777349651,0.1358683556,0.1072093844,0.046515964,0.0664102286,-0.4112359583,-0.1507118046,0.4748026431,-0.0082358513,-0.1946942657,0.1906995624,0.2914386392,-0.0253211483,-0.0612463318,-0.3841492534,-0.2049587369,0.2253571749,-0.1689433008,0.1071541011,0.027488742,-0.3181130886,0.0742192119,0.0069107385,0.3483940363,-0.055765707,-0.1817986667,0.0871192887,-0.3096677363]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/554","title":"nlp downloads to its module path","comments":"Ok interesting. We could have another check to see if it's possible to download and import the datasets script at another location than the module path. I think this would probably involve tweaking the python system path dynamically.\r\n\r\nI don't know anything about Nix so if you want to give this a try your self we can guide you or you can give us more information on your general project and how this works.\r\n\r\nRegarding `nlp` and `transformers`, we are not sure `nlp` will become a required dependency for `transformers`. It will probably be used a lot in the examples but I think it probably won't be a required dependency for the main package since we try to keep it as light as possible in terms of deps.\r\n\r\nHappy to help you make all these things work better for your use-case ","body":"I am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?","comment_length":141,"text":"nlp downloads to its module path\nI am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?\nOk interesting. We could have another check to see if it's possible to download and import the datasets script at another location than the module path. I think this would probably involve tweaking the python system path dynamically.\r\n\r\nI don't know anything about Nix so if you want to give this a try your self we can guide you or you can give us more information on your general project and how this works.\r\n\r\nRegarding `nlp` and `transformers`, we are not sure `nlp` will become a required dependency for `transformers`. It will probably be used a lot in the examples but I think it probably won't be a required dependency for the main package since we try to keep it as light as possible in terms of deps.\r\n\r\nHappy to help you make all these things work better for your use-case ","embeddings":[-0.0510633774,0.3298898339,0.1181987748,0.0430239737,0.2496246547,-0.1069276407,-0.0243597999,0.0844147205,0.0435677618,-0.1043166742,0.1716201752,0.8944575787,-0.3751601577,0.0496552885,0.4546215534,0.3453479707,-0.2024181634,-0.0113467798,-0.5702016354,-0.0118620675,-0.1689658463,0.3025683165,0.089546077,0.2100641578,0.0889092609,-0.2382594198,0.1604039371,0.2349795997,-0.1778022349,-0.3359404802,0.25264588,0.0436157882,0.1131191775,0.3050610125,-0.000122344,-0.1219699681,0.1226148382,-0.205461815,-0.5831148624,-0.431515038,-0.0276235566,-0.1339646727,0.2938423455,-0.3086282015,0.4444714785,0.4372788668,0.2576222718,-0.0503479801,0.1188958585,0.2702376246,0.153328225,0.3109493256,-0.1096139774,0.2995907068,0.2616927922,0.2531508505,-0.1965018064,0.3940299749,-0.1449605823,-0.1727851331,0.1455776691,-0.0195415467,-0.1950080544,-0.2205276787,0.4399972856,0.0483359061,0.4507356882,-0.1525630653,-0.1464647651,-0.0134353014,0.343753159,-0.6811647415,-0.3214661181,-0.3661260605,0.1135828197,0.0105409035,0.4246498644,0.3493974805,-0.2088426352,0.1319354326,0.0785173923,-0.4541172683,-0.1879279763,0.3013216257,0.0854256898,0.4778109193,0.0132613247,0.0066711451,0.2882016897,0.2142534852,0.3760667443,0.0814083666,0.1321113706,0.2844476104,0.2851187289,-0.0205892194,-0.1099530458,0.1793450415,-0.0300104544,-0.0847236365,0.0232693981,-0.1208940148,-0.1493709832,0.1514124721,0.2035764009,-0.247454226,0.0086239614,0.1453725249,0.1995931119,0.1786696166,0.2614904046,0.0819995105,-0.1421168,-0.0424142443,-0.1242887452,-0.1236252785,0.0732089132,-0.1548983753,-0.1208454743,-0.2121127695,-0.1051156819,0.2327277511,0.1520649642,0.0448909439,-0.0963122547,0.1335330158,0.2258944064,0.0212671552,-0.2801139951,0.3634817302,0.0146192536,0.1916169226,-0.2859227061,0.0753069296,0.1577265859,0.1072308496,0.341599673,-0.2339811474,-0.1381294131,-0.1220097318,-0.0012774215,0.2555344999,-0.1804871261,0.3423956335,0.1419740915,0.4394195974,-0.2550589144,-0.163176015,-0.4166961014,0.1174447984,-0.2558010221,-0.3332113624,0.1345807761,0.0606480204,-0.2976812422,-0.1739300936,0.1332337856,-0.6403436661,0.2606746554,-0.200510487,0.1565443128,-0.229875043,-0.1998524815,-0.2740626931,-0.0539591908,0.5535969138,0.1542535424,-0.2462250143,-0.2222314179,0.016472077,0.3207409978,0.1756643206,-0.1314175576,0.140968442,-0.2844077647,-0.2637615502,0.9092653394,-0.4382293522,-0.3080140054,0.3058415651,-0.6308793426,-0.0351980515,0.1344608217,0.177456215,-0.4078784883,-0.025490094,0.048203256,0.7276788354,0.2470006943,0.0803027824,-0.2167281657,-0.3240265846,0.2392051369,-0.0756338313,-0.0015354637,-0.0259253923,-0.129698649,0.5696974993,0.5710898042,0.1803895235,0.0185479186,0.3647143543,0.4654217064,0.1909317374,-0.2792431414,-0.3242795169,-0.0685014948,-0.0869885534,-0.6172335744,0.083127521,-0.5409413576,-0.1930327863,-0.1252249181,0.1388610154,0.1718779057,0.0744480193,-0.0156721268,0.2380425334,-0.002327444,0.208499521,-0.1346627623,0.4996421337,-0.0731389448,0.1621944904,-0.2590222359,0.0580387004,-0.1896460056,-0.1445167065,-0.0115874903,0.4375527203,0.1188165992,-0.1396409124,0.0758898184,0.4306365848,-0.1802239567,0.0920963138,-0.2294408679,0.0527636111,0.2509596944,0.0611949265,0.0663021579,0.2502961159,-0.0160635021,0.1760639399,-0.2467821985,0.093126975,-0.5122020841,0.1198067963,0.1599705219,0.2397339195,0.0859846994,0.0437981896,-0.1961194426,-0.1598126739,0.2977919579,0.1301416308,0.2628808618,0.3170697391,0.1482849121,0.0396019407,0.4440465271,0.1459557414,0.1853348613,0.0026250521,-0.0029060233,0.0064198836,0.0199254919,0.3390770257,0.1651233286,0.0997665226,0.1520308107,0.1469431818,-0.27741763,-0.2139859945,0.1724490523,-0.0141214132,0.1878112257,0.1951902956,0.0787312761,-0.1810029,-0.181981802,0.0489849597,0.0554885603,0.1607269049,-0.3464353979,0.0695948675,-0.6128370166,-0.4567539394,-0.7045779824,0.109661594,-0.5078395605,-0.2648012042,-0.1699569225,0.2932728827,0.0135282818,-0.1058647335,-0.1143262759,-0.1898264438,-0.419467181,-0.5510396361,0.0548674949,0.0549337864,-0.2664960027,-0.1325835884,0.2395806909,0.2251581848,0.1956444383,-0.1559059918,-0.3618601263,0.010488946,-0.1625342518,0.0291622337,0.2996201813,0.1283689141,0.2688087225,-0.0790974423,0.2188508064,-0.1453215182,0.033265464,-0.213767305,0.0759833679,-0.0186203849,0.1018570811,-0.0267567094,-0.3278627396,-0.2171003073,-0.2940075099,-0.5226594806,0.2773306668,0.2654492855,0.3282835186,0.7198580503,-0.1003150046,0.0236577746,-0.2349622548,0.1796551049,0.1849345565,0.0104816621,0.3286431134,-0.1557531804,-0.1278235316,-0.0141998576,0.0795360282,0.0719748586,0.4339595437,-0.3098500967,-0.0570606105,0.0136772105,0.085111782,0.048378326,0.1605340391,0.4640448987,0.2874497175,0.1406031847,-0.0581677184,-0.285225898,0.147121802,0.5211663842,0.3329201043,0.2887787819,0.2531733215,0.1671905965,0.3276456892,0.003027312,0.2196825296,0.244699344,-0.0253733657,0.5758994222,0.2007047534,-0.1775512844,-0.3279164732,0.1077798903,-0.0169585086,0.0566400811,0.0813843608,-0.0753141195,-0.073992379,-0.1050532833,-0.0269179568,-0.4237570167,0.0826270059,-0.0962546095,0.3312710524,0.223852694,0.1235564649,-0.2656936347,-0.264118433,0.2756395936,0.3231788278,0.0153704481,0.3640689552,-0.3138623238,0.0321954153,-0.3797568977,-0.3239942789,-0.1088466421,0.484484762,-0.1609846652,0.0305011272,0.2107773721,-0.1735964119,0.3358698189,-0.0021739537,0.2514642775,0.1559422463,-0.0984622091,-0.6052802801,-0.0064762565,0.0294505116,0.4140464067,0.2112326473,0.0046945848,-0.0695414618,-0.1452973038,0.0879957154,0.066471979,-0.2027299106,-0.2055308372,-0.261906445,-0.4269976914,-0.2829475403,-0.0546450429,0.0684246719,0.0742831081,0.2629223168,-0.0338506065,0.1958602518,-0.0462727137,-0.0393274538,-0.2717521489,0.1334659457,0.0085925926,0.1045139953,-0.0316489302,0.0205000043,-0.1956700236,0.4103145301,-0.0548782311,0.1217812821,-0.3207540214,-0.0974599645,0.3002343476,0.5241619945,0.0014360932,0.0281782467,-0.0456096642,-0.4427618682,0.0347291008,0.0793786421,-0.0407574587,-0.3023775816,0.2004400194,-0.4026096165,0.1797274947,0.181704849,-0.2110960931,0.2459633052,-0.1306901574,-0.3887245953,0.2153541893,0.1915543675,1.0236365795,-0.2396771014,0.3599079251,0.0731763467,-0.3844380677,0.401240617,-0.260976851,0.2450611591,-0.4522320926,0.2662657499,-0.0015371201,-0.3288942575,-0.1334097087,-0.0633254498,0.133866623,0.330349803,0.0987858623,-0.0731307864,0.156696707,0.3976758122,-0.1640682518,-0.3028284907,-0.3594765365,0.0098461779,-0.1150171533,0.4567201436,-0.0818569362,-0.2235649526,0.1479725838,-0.1842615008,0.0058870418,0.2117668539,-0.1513441056,0.0323051848,-0.0858151987,0.0216811616,0.0550218746,0.23313272,-0.0863150433,0.5087731481,0.0393289663,-0.2616960108,-0.2504293323,-0.2981612682,0.1080787778,0.4550248981,0.0318250321,-0.025415251,0.0077714031,-0.3527024388,0.066333428,-0.2313762009,0.1986922771,0.0110727753,0.2796635926,0.167022571,-0.2244542688,-0.0015972749,0.0190638024,-0.0449290909,0.0379766785,0.2209531814,-0.1191035584,0.1960103959,-0.063323617,-0.3193187714,0.072658658,0.5721097589,-0.1113176569,0.1076373532,0.31508255,-0.2308070809,-0.191078946,-0.1671551615,-0.0601866506,-0.5337637067,-0.1317318827,-0.1004977301,-0.1209860817,-0.2803724706,-0.286631465,-0.1232302338,0.1319796741,-0.0445765704,-0.350990653,-0.4225005209,-0.149766162,0.1053398326,0.1956525445,-0.0248330049,0.2506741881,-0.0674876794,0.0741262585,-0.1372200549,-0.1543278843,-0.0214239303,-0.1949546635,0.0507559665,0.120476529,-0.1465531588,-0.0282829553,0.0064559672,0.0490756072,0.119357869,-0.0334131494,-0.0371213555,-0.1996513754,0.2175134867,0.1549658775,-0.0837438032,0.2546332479,-0.4744584858,-0.0889240131,-0.0554784574,-0.2485372573,0.3140683472,0.1804844141,0.0298968572,0.1788478196,0.2484346479,-0.1243826449,0.4295685589,-0.4073117971,-0.1309642643,0.036665082,0.2387412935,0.0165815894,0.05330351,-0.3374192715,-0.1995884925,0.0303788837,0.1637501717,0.1828977764,-0.1355603337,-0.0794109553,0.2315355837,-0.0689691156,0.5128157735,-0.007258669,-0.1286974102,0.503020227,0.0117332907,-0.0458938219,-0.0972716361,0.2089644074,-0.1990007758,-0.1058805212,0.4660559595,0.2210081369,-0.1808511615,0.3089650869,0.0340255387,0.3652097881,0.0222477783,-0.0381954052,0.1044046432,-0.0497451201,-0.0288082398,-0.1028469205,0.1517295539,0.2119782716,0.2232206464,0.0762060359,-0.1425379515,-0.018712962,-0.1013332009,-0.0659233704,-0.0458261333,-0.360807389,-0.0269393399,0.627656281,0.2042374015,0.1547156721,0.0871560872,-0.34294644,0.0554116145,0.1697133482,0.1552277654,-0.1366814673,-0.0642936081,-0.1878053099,0.0336256959,-0.3583999574,0.0959847569,-0.0487383753,-0.4258345068,0.2607402205,0.160765484,0.1171657518,-0.4328006506,0.1650243998,0.1692375988,-0.020743696,-0.2386969477,0.1550823152,-0.4366209209,0.3467750847,-0.2618815601,-0.2523040473,0.2180014998,0.3444643915,-0.0700987503,-0.0922983512,0.0812103674,0.1643684804,-0.3064896762,-0.0062732636,-0.1060760319,0.7569422722,0.2321663648,-0.0359248891,-0.0273309089,-0.3327254653,0.2151863873,-0.2924382389,0.0483573861,0.689119339,0.2632908821,-0.0477681272,-0.1519185603,0.1965536773,-0.2658856809,0.0746650696,0.2499924153,0.1845074147,0.1303051561,-0.0365408957,0.0434275977,0.2541340888,0.425416559,0.058415845,0.1500281543,0.0744764283,0.0282467641,-0.576366961,0.1385060102,-0.3032060266,-0.3131721914,-0.0199992917,-0.101450108,-0.0216269828,0.0585193262,-0.2070562392,0.1820068657,0.1419270784,-0.0399258733,0.078320466,-0.0992768854,0.151590541,-0.0067264214,0.0335787572,-0.4286132753,-0.0621461272,-0.2182703018,-0.0557151847,-0.142843172,-0.1603760272,-0.1300808936,0.175480634,0.1766574085,-0.0793760642,0.0945205688,-0.0011483638,-0.4309166372,-0.4071859121,0.0395411104,-0.3112210333,-0.0475023575,-0.0774854273,0.0218865871,-0.350371927,-0.0836389363,-0.5981824398,0.2577419877,-0.1080349609,0.2534945905,0.2370180339,0.0655613244,-0.2741672099,0.3343589902,0.110822849,0.0519285686,0.0080320165,0.0687494203,-0.3790957332,-0.1683959663,0.4932083189,-0.0010689513,-0.1690285802,0.1351238638,0.2747317851,-0.0612607487,-0.120022662,-0.3803345263,-0.1763632894,0.1848460585,-0.1816511005,0.1024647504,0.0878539383,-0.3482228518,0.0780851766,-0.0038820552,0.3832035363,-0.075927034,-0.2382911891,0.0604641996,-0.2934118509]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/554","title":"nlp downloads to its module path","comments":"@danieldk modules are now installed in a different location (by default in the cache directory of the lib, in `~\/.cache\/huggingface\/modules`). You can also change that using the environment variable `HF_MODULES_PATH`\r\n\r\nFeel free to play with this change from the master branch for now, and let us know if it sounds good for you :)\r\nWe plan to do a release in the next coming days","body":"I am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?","comment_length":65,"text":"nlp downloads to its module path\nI am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?\n@danieldk modules are now installed in a different location (by default in the cache directory of the lib, in `~\/.cache\/huggingface\/modules`). You can also change that using the environment variable `HF_MODULES_PATH`\r\n\r\nFeel free to play with this change from the master branch for now, and let us know if it sounds good for you :)\r\nWe plan to do a release in the next coming days","embeddings":[0.0594276041,0.2854037285,0.1192257926,0.0074428776,0.2506986558,-0.1265826225,-0.0366810672,0.0295479912,0.0179713909,-0.119238168,0.145190075,0.9340407252,-0.348520726,-0.0550438054,0.448151648,0.3217197955,-0.1729900539,-0.0191877522,-0.5424788594,-0.0087753208,-0.1322143078,0.343559891,0.0899610296,0.2194026858,0.1086243019,-0.2739531398,0.1533594579,0.1835228354,-0.1502880007,-0.361440748,0.2200702876,0.0330717936,0.0396199599,0.2460679263,-0.0001222595,-0.0723031759,0.0618373081,-0.1991177052,-0.5908346772,-0.4036874473,0.0073660119,-0.1110637337,0.331789583,-0.2588049471,0.4212759435,0.4604949653,0.2072204202,-0.0632975549,0.0779818371,0.1990575045,0.1536840349,0.2625417411,-0.1157127097,0.3399074078,0.2558580935,0.206086278,-0.2066236734,0.4231725633,-0.1807647943,-0.1173191592,0.1137165353,0.0151541848,-0.1605414748,-0.2232352942,0.4218443334,0.0965171233,0.5065210462,-0.1041269973,-0.178357169,-0.057623636,0.2488173693,-0.6838291883,-0.3021194637,-0.3956843615,0.0583829917,-0.011045713,0.4367433786,0.2950358689,-0.1969232112,0.1277929246,0.1174732447,-0.4387536049,-0.1516216099,0.2879826128,0.0643961206,0.5356558561,-0.0010861605,-0.0201419462,0.3124943674,0.2233442515,0.3028106093,0.1413413584,0.1304742694,0.2637385428,0.306614846,-0.0090490058,-0.126691401,0.1287931651,-0.1076702848,-0.0864102021,0.0486149788,-0.1003772989,-0.100146845,0.1405427754,0.128412351,-0.2314269096,-0.0331277698,0.1172136813,0.1847460717,0.1608623266,0.237539351,0.0487547666,-0.0829556063,-0.0250671338,-0.0955912769,-0.1017725915,0.0238328222,-0.1320399195,-0.1010759845,-0.2369559705,-0.1393790841,0.2759144604,0.1524428576,0.0107965935,-0.0849052668,0.1406838745,0.2397924215,0.0124138677,-0.3015604913,0.4846943617,0.0200684797,0.1443138719,-0.3159932196,0.0632190555,0.1285133362,0.0914971083,0.3637309074,-0.2417959124,-0.1056750491,-0.1247909218,-0.0109360656,0.284550339,-0.1785971522,0.3613257706,0.0448554307,0.4277518094,-0.2470026016,-0.1423707306,-0.4103898406,0.1094860882,-0.2655018866,-0.351354003,0.1702574342,0.0741485283,-0.2385264635,-0.1675650924,0.228550598,-0.6515527368,0.2796333432,-0.1777371764,0.1845386773,-0.2027804404,-0.1684630066,-0.2977358103,-0.055017367,0.4658385813,0.2440615594,-0.2764915824,-0.2273944169,0.0714991987,0.3334225714,0.157590881,-0.0960775763,0.1057573482,-0.2538484037,-0.2903698981,0.9286288619,-0.4024112523,-0.2970966101,0.2756410539,-0.6943036914,-0.0525972098,0.1200872064,0.1547679156,-0.4531956017,-0.0136170126,0.0777316689,0.7065265775,0.2650775313,0.128693521,-0.2549385428,-0.3223257065,0.2341711223,-0.1441075802,-0.0270406026,-0.0166181941,-0.129373461,0.6533600092,0.5752850175,0.1956337392,0.0006461617,0.3474607468,0.5321191549,0.2032576501,-0.2729887068,-0.2872490287,0.003228341,-0.0935835093,-0.6751148701,0.1077949554,-0.5177379847,-0.1886842549,-0.1189373508,0.1137020066,0.1747476161,0.0378866047,0.0076451357,0.2549017966,0.022808522,0.2261755615,-0.132650122,0.457488209,-0.0661611483,0.1869132668,-0.1980352551,0.0420323275,-0.2060025334,-0.1463294029,-0.0704797655,0.5003626943,0.1262582242,-0.1380066425,0.0808468759,0.4534180462,-0.2065954953,0.0868909508,-0.2848199606,0.0214223545,0.3208904862,0.0965320095,0.1042212471,0.2300522178,-0.0680665001,0.1967657506,-0.202443406,0.1004064009,-0.4833717346,0.0858843699,0.1622382998,0.2138819098,0.0588472784,0.040826492,-0.1509973109,-0.1954513937,0.2688579559,0.1386809647,0.2791650593,0.366545409,0.1968338192,0.0444832966,0.3903595805,0.2215695381,0.1931091547,0.0381867364,0.0259094518,-0.0182071514,0.0148508698,0.2840825021,0.1803556532,0.0792310834,0.1897486299,0.1379857808,-0.2874861956,-0.2119708955,0.2023925185,-0.0441415682,0.1804206222,0.2209619582,0.1378938556,-0.1611262709,-0.1813827306,0.1015159264,0.0782349482,0.1337401569,-0.2980790734,0.0790636539,-0.6222834587,-0.3927905262,-0.7182297707,0.0430182479,-0.5489847064,-0.247586295,-0.1140798628,0.2927651107,0.0120862331,-0.1029171571,-0.0963459015,-0.1998202503,-0.4088550806,-0.5224940181,0.0758768991,0.0744953901,-0.3028307557,-0.1217320487,0.1361252815,0.1972023994,0.1763104349,-0.1033913568,-0.3834490478,0.0270648189,-0.2044581473,0.0011427145,0.3371047974,0.1227916777,0.264877975,-0.1386473328,0.2038497329,-0.1997368932,0.0460605621,-0.2331735641,0.0697179586,-0.008256549,0.1032613739,0.0749440789,-0.3153726161,-0.2225984186,-0.2827696502,-0.4827471375,0.2971625626,0.2415864319,0.3336547613,0.7670289874,-0.144687593,-0.0249776915,-0.2470210344,0.1891744137,0.1809889525,0.0252696518,0.2319223434,-0.1236328185,-0.0485379882,-0.0837118626,0.0633222982,0.059877146,0.4791371524,-0.2922404408,-0.0563170575,0.0676808879,0.0439293869,0.0678461567,0.1174633205,0.472961992,0.2859099805,0.1283735335,-0.0220060572,-0.3026004434,0.187647447,0.5915962458,0.3701991141,0.2775197029,0.1701226085,0.2223974168,0.3195584714,-0.0285745673,0.2068920135,0.2589379549,0.0239454638,0.5233561993,0.1765196323,-0.1383137852,-0.3046542108,0.1189453676,-0.052725073,0.04060366,0.1409616172,-0.1012789682,-0.0248239189,-0.0608646236,-0.041363515,-0.4126467705,0.0877082422,-0.1293936968,0.3402701914,0.193472296,0.0796755403,-0.2661873698,-0.2592256069,0.2928960621,0.3117831945,0.0069993003,0.3332320452,-0.2916759253,0.0138884857,-0.3250705898,-0.3397919834,-0.0627191663,0.4760490954,-0.1372724473,-0.0144195417,0.2336534858,-0.1431980282,0.2587844133,0.0479314476,0.2314875573,0.1666573435,-0.1175597832,-0.5710583925,-0.0377743505,0.0776699334,0.4122169912,0.1997942924,0.0090891495,-0.0728257746,-0.1482959539,0.0782654807,0.0250604134,-0.2103389353,-0.1921575963,-0.3132322729,-0.4200137854,-0.2545804381,-0.0016287579,0.0169383492,0.0680180714,0.2967764437,-0.0510970317,0.2507216036,-0.0551032387,-0.0773958936,-0.2792865336,0.1545903087,-0.0144263627,0.0327636823,-0.0014920747,0.0513724536,-0.249586612,0.3455182314,-0.0801326185,0.1714082211,-0.3153802752,-0.0383841917,0.2609153092,0.5330729485,0.02689448,0.1530143023,-0.0604951009,-0.4698383808,0.0383233204,0.0581184812,-0.0631178766,-0.3346704841,0.2662449181,-0.3325277567,0.1799117774,0.1377166957,-0.2574649155,0.2627463639,-0.1769567281,-0.4302590787,0.2455020398,0.2469624579,0.9723243117,-0.2307320088,0.3222673535,0.0736390799,-0.4327769279,0.3743034601,-0.2826933265,0.2952537835,-0.4656220675,0.3089422882,0.0315963775,-0.3489761353,-0.1434809417,-0.0949161425,0.1160013899,0.2775852978,0.1108724773,-0.0856860727,0.1542301327,0.3288413286,-0.1667702645,-0.308303833,-0.4076279402,0.008780499,-0.0911314636,0.4811152816,-0.1528789997,-0.2010472566,0.1455898136,-0.1341896057,0.0559144393,0.1484689415,-0.1410245299,0.0270280242,-0.0384529196,-0.0294565223,-0.0405968577,0.1738117188,-0.0472843982,0.5029783845,0.0930437148,-0.2813098133,-0.232560277,-0.2815986574,0.1130905524,0.410629034,-0.0128927669,-0.0304554235,-0.0021414417,-0.3768609762,0.0323989205,-0.2656219006,0.2018044144,0.0022114515,0.2960530221,0.1907166392,-0.1325136125,0.0046178191,0.0488902889,-0.0576300323,0.0454691425,0.213784188,-0.1503866017,0.1730715483,-0.0952475965,-0.3525271714,0.0890106931,0.5636550784,-0.0558686666,0.130743131,0.2681160271,-0.3183728755,-0.1945942193,-0.1448384374,-0.1017150357,-0.5875363946,-0.0966898724,-0.1024175808,-0.1494128853,-0.3403916657,-0.2876192927,-0.1150793284,0.190659225,-0.0029870479,-0.3976166844,-0.3932400644,-0.1758130044,0.1078059822,0.1168417931,-0.0234045256,0.226348713,-0.0421316437,0.0578667596,-0.123500213,-0.151301384,0.0047081551,-0.2057931125,0.0646609664,0.0125340233,-0.2309372127,-0.0588082522,0.0386312157,0.0674866587,0.1651312113,-0.0121785197,-0.0329803228,-0.1857112348,0.2187648118,0.1187468246,-0.0821009502,0.2349191308,-0.4461464286,-0.0323818214,-0.0367250592,-0.228381142,0.2963550985,0.2207998186,-0.0084141744,0.1888852417,0.2444496155,-0.1081682742,0.3947142065,-0.4404299557,-0.1914960295,0.058791317,0.2005114108,-0.0385495797,0.0262689926,-0.3012150526,-0.1798777878,0.0153009146,0.1391173601,0.155297637,-0.1375177801,-0.1048076525,0.2277946174,-0.0768609196,0.4794976115,0.0657889917,-0.1560293138,0.4810068905,0.018120762,-0.0185829643,-0.0961592719,0.2216251791,-0.201333642,-0.1024582535,0.4729754925,0.2948073149,-0.2150394917,0.3210662305,0.0439506397,0.338032335,0.0131287733,-0.0645014197,0.0703040734,-0.0233831424,0.0092220865,-0.072838217,0.1210112646,0.2201230228,0.1502483189,0.1225652322,-0.1162338853,-0.0255551822,-0.1099669412,-0.0900656283,-0.0236122273,-0.3888558447,-0.0020131138,0.6453694105,0.1920175552,0.1655847877,0.0818606168,-0.3722477853,-0.0144479368,0.2004949003,0.128369838,-0.2168528885,-0.0728575513,-0.1325110495,0.0121029718,-0.3438655138,0.0954862162,-0.0437260121,-0.3884956837,0.3022679389,0.1114626452,0.1269278377,-0.469509691,0.1270013303,0.200374797,0.019215703,-0.210467875,0.1976362318,-0.4272415638,0.2984758317,-0.2829221785,-0.2194400728,0.1775113344,0.3179076016,-0.1029378772,-0.0928037018,0.0829827785,0.1744609475,-0.3017723262,-0.0103362668,-0.077597782,0.7949570417,0.1794158369,-0.0431595333,-0.0475341491,-0.3129450083,0.22062397,-0.2494146824,0.0387258194,0.6695898175,0.3283213079,-0.0491800234,-0.0740498751,0.2219728529,-0.261903882,0.0619016252,0.2014833838,0.1449313611,0.132254824,0.0189826135,0.0547857396,0.3298667073,0.381329447,-0.0070947567,0.1342583597,0.063984558,-0.0140789133,-0.5347217321,0.1711659431,-0.3061721325,-0.2194457948,-0.0245554112,-0.1132671162,-0.0000205536,0.0607558191,-0.2270715535,0.1703407019,0.130868122,-0.0588848181,0.0626302361,-0.0209633075,0.184641853,-0.0497381538,0.0059668887,-0.4095641375,-0.0503586829,-0.1564158946,-0.0454779305,-0.1061545834,-0.1463037431,-0.1245374456,0.2262842804,0.1479841918,-0.0311849713,-0.0018010786,0.0141691752,-0.4841278791,-0.3667883873,0.0356950499,-0.2991974056,-0.1208483726,-0.1179776713,0.0703835338,-0.3258701861,-0.0642839596,-0.6216781735,0.2913423479,-0.0578833483,0.2728645504,0.2130309194,0.0763408914,-0.2483003289,0.2943148017,0.1480212808,0.106659919,0.0531164259,0.0504271388,-0.3799441159,-0.1667602509,0.4578514397,0.0323451497,-0.1924961209,0.1639104635,0.2948584855,-0.0421203598,-0.0757892951,-0.3868128657,-0.2089805603,0.2247839272,-0.1893402785,0.1238151491,0.0422796756,-0.33436957,0.0257927645,0.0008597435,0.360099256,-0.0796993598,-0.1856142431,0.0677081794,-0.3200276494]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/554","title":"nlp downloads to its module path","comments":"Awesome! I\u2019ll hopefully have some time in the coming days to try this.","body":"I am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?","comment_length":13,"text":"nlp downloads to its module path\nI am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?\nAwesome! I\u2019ll hopefully have some time in the coming days to try this.","embeddings":[0.0418294333,0.2534075975,0.1386140138,0.0216108505,0.2504269183,-0.1144135669,-0.0517433174,0.0558355376,0.0386522263,-0.0650730282,0.1688313335,0.9240578413,-0.4022577405,0.0137117188,0.4288115203,0.3343911171,-0.159665525,-0.0184178744,-0.5421508551,-0.0123548927,-0.1628659517,0.3238161802,0.0957886577,0.2096389979,0.1090578362,-0.2200091779,0.1640121788,0.2077167481,-0.1559758931,-0.3526194394,0.2312970459,0.0157164801,0.0543910824,0.2630090415,-0.0001235746,-0.1145020351,0.1038903445,-0.2037020624,-0.5521075726,-0.4115805626,-0.0187189709,-0.0749409348,0.3090879023,-0.2563981116,0.4332640469,0.4154006541,0.2282980382,0.0107338969,0.095377095,0.2215437442,0.1400552243,0.2679061592,-0.1248426065,0.3356770277,0.2586003244,0.2509728968,-0.2028630376,0.4553247094,-0.186783582,-0.1706300676,0.1252813488,-0.0036334826,-0.1767319888,-0.2037964612,0.4682290852,0.0628647506,0.5204929113,-0.1328703165,-0.1806992441,-0.0017642822,0.2549144626,-0.6548467278,-0.2750431001,-0.3721725941,0.0868566632,0.0163589641,0.4685966671,0.3566254079,-0.233925283,0.1553521603,0.1223921627,-0.4636808634,-0.1816964447,0.310940057,0.0992298126,0.5150282383,0.0094051603,-0.0067259036,0.3597558141,0.2328117341,0.3040569723,0.1261950433,0.0981685966,0.2708328366,0.3027562499,-0.0150519703,-0.1233629733,0.1666315049,-0.0726540685,-0.1228725091,0.0088159824,-0.1305527687,-0.1255352646,0.168219924,0.1629052758,-0.2460231632,-0.026040135,0.1456226259,0.1787384748,0.122557655,0.2616567314,0.0845641941,-0.0827915743,-0.0212746784,-0.1153428853,-0.1527855843,0.0262177084,-0.1295469999,-0.0660198107,-0.2291292697,-0.1755687892,0.2998301089,0.1762029529,0.011194095,-0.0716518164,0.1207123995,0.2215410173,0.0266553983,-0.2817135751,0.4243544042,0.0362362042,0.1453109086,-0.2836043537,0.0695054457,0.1400603354,0.1051636934,0.3405663669,-0.2266487479,-0.1252206713,-0.1481317878,-0.0158760212,0.286700964,-0.1732510328,0.3688527346,0.071478717,0.459888339,-0.2371465713,-0.1731324345,-0.4142155051,0.1436846703,-0.2608460784,-0.3819936216,0.119232215,0.0517110266,-0.2623910308,-0.1453406662,0.2002359182,-0.6477709413,0.2685735822,-0.1585607976,0.2101445645,-0.1979130954,-0.1973253787,-0.2802213728,-0.0451070815,0.4881328046,0.2346744984,-0.2558528781,-0.2276197374,0.0361940041,0.3606449366,0.1716537774,-0.1026235968,0.1151047945,-0.255463779,-0.2956459522,0.8869797587,-0.3626984954,-0.2903615832,0.2899290323,-0.671682179,-0.0425353236,0.1173964962,0.1682763845,-0.481184721,-0.0058447146,0.0944169015,0.7163159847,0.2666470706,0.1134107336,-0.2747299671,-0.3304677308,0.2352412194,-0.1098986641,-0.0446129777,-0.0039056356,-0.16745767,0.6007237434,0.5775532126,0.1740926653,0.0019759503,0.3332244158,0.5358592868,0.1875705272,-0.2864373326,-0.2668869197,0.0456505232,-0.0837819576,-0.6365544796,0.1228748485,-0.5194808841,-0.1926686466,-0.108366847,0.143728286,0.1867934614,0.0767503679,-0.0179266296,0.2434986979,-0.0329059139,0.2278067619,-0.1108156145,0.4810334146,-0.1100192815,0.1544666737,-0.206017971,-0.0101343552,-0.1779963523,-0.1433139592,-0.0821668208,0.4535774589,0.0965785682,-0.1240254045,0.0504825898,0.4710916281,-0.2074751705,0.0969225764,-0.2732140422,0.0560357943,0.275598824,0.0331924297,0.0300531331,0.2590866089,-0.0559849218,0.1919047982,-0.2442449927,0.0993752033,-0.5002464056,0.1031692699,0.0998342857,0.2525646687,0.056296505,0.0603295751,-0.1821353436,-0.1413209885,0.2916157544,0.1009642333,0.2522548735,0.356440872,0.1555127054,0.002077484,0.3812526166,0.2035042644,0.1896533966,0.0407917276,0.0208965912,-0.0188344084,0.0191379227,0.3209028244,0.1342677474,0.0689531267,0.173417598,0.1450206041,-0.2985064387,-0.2035327256,0.2024057209,-0.0200218,0.1469393373,0.1711248308,0.1204817742,-0.202411741,-0.1841471046,0.1275900602,0.0419996716,0.1727479547,-0.3343026936,0.070667237,-0.5970048308,-0.3824313581,-0.7140731215,0.0368406996,-0.5593503118,-0.2496697903,-0.1294646114,0.2926515043,-0.00612973,-0.1176145598,-0.0907773897,-0.1927859932,-0.4273487628,-0.5884398818,0.0605809204,0.0762898177,-0.2748950422,-0.1477163732,0.1880694777,0.2169873714,0.1580551416,-0.1176227257,-0.3935468495,0.059942849,-0.1835197806,0.0021817121,0.3144090176,0.1501932144,0.27812621,-0.0920722038,0.2115061432,-0.1820324659,0.0415893644,-0.2296114713,0.1198756769,-0.0053982739,0.1131510884,0.0682690293,-0.365133673,-0.1788724363,-0.2470638901,-0.4987287521,0.2901611626,0.2217643857,0.3247456849,0.7784491777,-0.0884122476,-0.0025166725,-0.2445620894,0.1509108245,0.216164425,-0.0044993637,0.2655749023,-0.1308584362,-0.0609765388,-0.0401885621,0.0847922564,0.0286857598,0.4407421052,-0.3026358783,-0.0272076912,0.0240075625,0.054623872,0.0339477472,0.1121748313,0.4431363046,0.2924706936,0.1394312978,-0.0211779103,-0.2869519293,0.2251290828,0.5765174031,0.3781866431,0.3043326139,0.1885630935,0.1878098249,0.3516527414,-0.0412553921,0.2518646419,0.2450685501,0.0015089759,0.5418968797,0.1687188447,-0.1285502315,-0.3250176907,0.1292764693,-0.0497953445,0.0587641448,0.1341176778,-0.1080719456,-0.0335135423,-0.0938217714,-0.0089664301,-0.4311963022,0.1054149792,-0.0967562273,0.3484645784,0.1630801409,0.0772229657,-0.252517879,-0.2752417624,0.2950673401,0.3228208423,-0.0040059108,0.359187901,-0.2693516016,0.0255640373,-0.3555164039,-0.360248208,-0.105548203,0.4851176143,-0.1355840862,-0.021680763,0.2077476829,-0.1623816341,0.324357748,0.0446485989,0.264189899,0.1277220845,-0.1164147109,-0.5647357702,-0.016786363,0.0926255956,0.398650229,0.1968774945,0.0096786814,-0.0432436578,-0.163420096,0.0706112385,0.0418122075,-0.1947576106,-0.2161845416,-0.2714056373,-0.4215853512,-0.2126562446,-0.0120736966,0.0613648072,0.0829081014,0.2916682363,-0.0307235848,0.2108202428,-0.0576977022,-0.0559911318,-0.275480181,0.1352327615,-0.0130693121,0.0731906593,-0.0041119563,0.0217297785,-0.2675389946,0.3485262394,-0.0655389279,0.1675309092,-0.3315232694,-0.0775304437,0.284427762,0.5340335965,0.0009662221,0.0879936069,-0.0464801714,-0.4646082819,0.0249907859,0.0986376256,-0.0510616452,-0.3298466504,0.2297488749,-0.3413712382,0.2045945674,0.1476123184,-0.2355486304,0.2855746448,-0.1147062033,-0.4128862619,0.2616344392,0.2328268886,1.0417244434,-0.2599345148,0.3303723037,0.0488393568,-0.4149959385,0.345830828,-0.2619602382,0.273679316,-0.4296852052,0.2920298278,0.0090353293,-0.3381702304,-0.1593915373,-0.0936512202,0.1404439211,0.2590028346,0.0970626548,-0.073985979,0.1680105031,0.3345319331,-0.1661510915,-0.313397944,-0.399148941,-0.0042323661,-0.078919813,0.4227870405,-0.0893053338,-0.2319464684,0.161267966,-0.1722237766,0.0232148021,0.1553335041,-0.1369795948,-0.0463397764,-0.0289707556,0.026320722,0.0076603736,0.1985774934,-0.0971941873,0.5324209332,0.0375580154,-0.2742050588,-0.2472143471,-0.2783737183,0.104185842,0.435143888,-0.0381262153,-0.0085787447,0.0203016102,-0.3510388434,0.0346772783,-0.2601581514,0.1960866898,-0.0072529116,0.2796385586,0.2240684479,-0.1997314394,-0.0108407941,0.0175201166,-0.0499265119,0.0316208862,0.2114604712,-0.152279526,0.162178576,-0.1226647571,-0.3572569191,0.082637921,0.5585677028,-0.0889019445,0.1565210819,0.2886305451,-0.2973922789,-0.1511554122,-0.1393899024,-0.0482732691,-0.5979011059,-0.108062759,-0.1069821566,-0.1324701458,-0.330191493,-0.3016616702,-0.0693702251,0.1817860305,-0.036913123,-0.3676624596,-0.3775153458,-0.1715250909,0.085571304,0.1245198697,0.0355435796,0.1998342872,-0.0467664227,0.0735006332,-0.129897818,-0.1353505701,-0.0194109455,-0.2274007201,0.0429360121,0.0298347697,-0.1467767209,-0.0054611238,0.0218470488,0.0475532152,0.1438240707,-0.0244408548,-0.0205011796,-0.2081116289,0.2245577425,0.1470132917,-0.0681883693,0.2743360102,-0.4622973204,-0.080645673,-0.0222837646,-0.2554251254,0.3135326505,0.1799019128,0.0002641503,0.1845735908,0.2836195827,-0.1086298078,0.3980067372,-0.3936973214,-0.1676221192,0.0364725702,0.1844539046,-0.0148262847,0.0430262983,-0.2990981042,-0.1867211461,-0.005764097,0.1293311417,0.1288925707,-0.1097215265,-0.1166316718,0.2188687921,-0.0670458451,0.4941369891,0.0274463072,-0.1444521099,0.5177966356,0.0042800773,0.0047590104,-0.0952822939,0.1944553107,-0.2102113068,-0.1120998785,0.498182565,0.2302597761,-0.2052747607,0.297090441,0.0339482129,0.2981280386,0.0149077531,-0.0677287802,0.1218806654,-0.0938096792,-0.0141498828,-0.1019033641,0.1188948527,0.2447907776,0.1939441562,0.1302758157,-0.1481094509,-0.0166179053,-0.1342697889,-0.0799317509,-0.0193673652,-0.3802731633,-0.013060499,0.6425284743,0.2309107631,0.1620884091,0.1023210138,-0.3735554814,-0.0009796345,0.1975531131,0.1168378741,-0.1429667026,-0.0987295881,-0.1239557788,0.0058877715,-0.376064837,0.0942515656,-0.0441192314,-0.4241124094,0.2915341258,0.137667805,0.1545638144,-0.4633343518,0.0875329226,0.2162381262,0.0116310716,-0.2519981265,0.1276583672,-0.4688633084,0.3034558892,-0.2724653482,-0.2711786032,0.1869259328,0.3480255604,-0.1217410788,-0.1052358598,0.0512160622,0.1935057938,-0.2802423835,-0.0102825211,-0.0678559467,0.7720012665,0.190816164,-0.0627545491,-0.0238217544,-0.3058003783,0.1673726141,-0.2202323824,0.0346398279,0.6944566369,0.3011940122,-0.0317995884,-0.1120656431,0.1925322562,-0.2314776629,0.0762315765,0.1957195252,0.1862598956,0.1375096142,0.0208844058,0.0399171486,0.2841912508,0.437877357,0.0324794091,0.1352613121,0.0646249056,0.0047267559,-0.5239402652,0.1511453688,-0.316700846,-0.2397229671,-0.0230512712,-0.1251464039,0.0257826112,0.0624909587,-0.2607123256,0.1912573427,0.097368896,-0.0156503711,0.1242326126,-0.0774747804,0.1549631059,-0.0386974551,0.0149905188,-0.4072253108,-0.0331648327,-0.1638901532,-0.0601973683,-0.1039701626,-0.1665753275,-0.1165012419,0.2075195462,0.1594305634,-0.049736239,0.0028709415,0.0118721519,-0.4753716886,-0.3776991367,0.0536743738,-0.3117780685,-0.1399779767,-0.0947524756,0.0672410429,-0.3473833203,-0.0534331501,-0.6327044368,0.2650748491,-0.0769568011,0.2418436706,0.2353070676,0.0647470653,-0.2434095591,0.3095802665,0.1225509644,0.0916947052,0.0450530685,0.0583181828,-0.3954268396,-0.1425477117,0.4672878683,-0.0111658759,-0.1972273141,0.1872034967,0.2757283449,-0.0325110294,-0.0666256845,-0.3964265287,-0.2190314382,0.1923856437,-0.1665027887,0.0879366249,0.0400611311,-0.3278409243,0.0703914315,0.0002715769,0.390011996,-0.0552369095,-0.1973922998,0.0939456001,-0.3245262802]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/554","title":"nlp downloads to its module path","comments":"> Feel free to play with this change from the master branch for now, and let us know if it sounds good for you :)\r\n> We plan to do a release in the next coming days\r\n\r\nThanks for making this change! I just packaged the latest commit on master and it works like a charm now! :partying_face: ","body":"I am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?","comment_length":58,"text":"nlp downloads to its module path\nI am trying to package `nlp` for Nix, because it is now an optional dependency for `transformers`. The problem that I encounter is that the `nlp` library downloads to the module path, which is typically not writable in most package management systems:\r\n\r\n```>>> import nlp\r\n>>> squad_dataset = nlp.load_dataset('squad')\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 530, in load_dataset\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n File \"\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/load.py\", line 329, in prepare_module\r\n os.makedirs(main_folder_path, exist_ok=True)\r\n File \"\/nix\/store\/685kq8pyhrvajah1hdsfn4q7gm3j4yd4-python3-3.8.5\/lib\/python3.8\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nOSError: [Errno 30] Read-only file system: '\/nix\/store\/2yhik0hhqayksmkkfb0ylqp8cf5wa5wp-python3-3.8.5-env\/lib\/python3.8\/site-packages\/nlp\/datasets\/squad'\r\n```\r\n\r\nDo you have any suggested workaround for this issue?\r\n\r\nPerhaps overriding the default value for `force_local_path` of `prepare_module`?\n> Feel free to play with this change from the master branch for now, and let us know if it sounds good for you :)\r\n> We plan to do a release in the next coming days\r\n\r\nThanks for making this change! I just packaged the latest commit on master and it works like a charm now! :partying_face: ","embeddings":[0.0606999062,0.2831610143,0.1252468526,0.0024886292,0.2875494063,-0.1234022155,-0.0438722596,0.0497939438,0.0141634243,-0.107714884,0.1856339574,0.9319191575,-0.3687045276,-0.005124581,0.4456731677,0.3321512341,-0.158552438,-0.0046792314,-0.5357556939,-0.0219645798,-0.1532678306,0.3509792089,0.1030698493,0.2226558477,0.0984963104,-0.2493528277,0.1723134667,0.1874399036,-0.1824496984,-0.3391270339,0.2290158421,0.0284767486,0.0377187617,0.2479455024,-0.0001226053,-0.0975798666,0.096115537,-0.1961422861,-0.5711393952,-0.4336972237,-0.0093921246,-0.1244058684,0.3227511644,-0.2763148546,0.4087745249,0.4453664124,0.2210507691,-0.0199714135,0.0893469825,0.1858502775,0.1542526633,0.274738282,-0.1103928015,0.32424891,0.2490096837,0.2277175188,-0.2165386677,0.4389002025,-0.1975561082,-0.1600362659,0.1131767407,0.0045972471,-0.1425084025,-0.244683221,0.4304543436,0.0933979899,0.5288747549,-0.1197954044,-0.1822494715,-0.0372463465,0.2761249542,-0.6750584841,-0.2873580754,-0.3752099276,0.0746391714,0.0016228288,0.457609117,0.3321336806,-0.2250935286,0.1359217018,0.1363037676,-0.4285431206,-0.1708862633,0.3129571676,0.0732560158,0.5276486874,0.0064699282,-0.0034219967,0.3391944468,0.2261256278,0.2892914414,0.1331435293,0.1072316021,0.262378782,0.3028283417,-0.0299122743,-0.1308882385,0.1408717185,-0.0703875124,-0.1001491323,0.0285716392,-0.1111956388,-0.1021454558,0.1614953876,0.17541188,-0.2487050444,-0.0226982366,0.1262587011,0.1851589531,0.1101535782,0.2593814433,0.067362532,-0.090057455,0.001816177,-0.1098765507,-0.1200204492,0.0314019285,-0.1485593766,-0.0833677575,-0.2215364724,-0.1621479839,0.2782630026,0.1563169956,0.0119718453,-0.0685706958,0.1434453279,0.2104525715,0.0210198835,-0.2872292697,0.4619431794,0.0285878666,0.1335110515,-0.3166640997,0.0505026132,0.1373167336,0.1125014499,0.3570070565,-0.2416785359,-0.0993690491,-0.1350381225,-0.0010034338,0.267819941,-0.1936892569,0.3649393022,0.0505852997,0.4417575598,-0.2347773463,-0.1650624126,-0.4237526059,0.1138925701,-0.2390219271,-0.3653302789,0.1551286429,0.0705930963,-0.2658124566,-0.1641084701,0.2171801627,-0.6330216527,0.2544854283,-0.1724122614,0.1936621368,-0.1905896068,-0.1788431853,-0.2897791564,-0.0612880662,0.4643811285,0.2399652153,-0.2634062767,-0.2285619229,0.0405276641,0.3470140994,0.184764266,-0.1119706258,0.1091829762,-0.243764773,-0.2768899798,0.9239950776,-0.3892824948,-0.2849107087,0.2982855439,-0.6965087056,-0.0430955291,0.0987514853,0.1680971384,-0.4586606622,-0.0215316545,0.075115636,0.7128710151,0.2529002726,0.1142326668,-0.2597278059,-0.3190846443,0.2476548702,-0.1181304902,-0.045669429,-0.0101613291,-0.1488674283,0.6455153823,0.5735569,0.1820348501,-0.0125398561,0.3438820541,0.5681894422,0.1878331006,-0.2743800581,-0.2664017975,-0.0003910377,-0.0994902328,-0.6493749619,0.1542912275,-0.504535377,-0.1930664778,-0.1210341454,0.1389087439,0.1891831458,0.061525736,0.0021858034,0.2332508266,-0.0239607375,0.2157777697,-0.1198194772,0.4618180096,-0.1038123816,0.1886592358,-0.1736332476,0.0234374367,-0.2014029175,-0.1311799288,-0.0557637513,0.4927198887,0.1181429699,-0.137855947,0.0671340823,0.4556130171,-0.1938720793,0.0698337629,-0.273149699,0.0493071303,0.2993729413,0.0655336976,0.0606363118,0.2522670031,-0.0486279018,0.1616556346,-0.2174688876,0.1048420817,-0.4903993309,0.0865472406,0.1171607673,0.2364952266,0.0667830855,0.0605360344,-0.1961008906,-0.1744933873,0.2643865943,0.1245055199,0.2692700624,0.3537789881,0.1622397006,0.0071045556,0.4008791447,0.1960743666,0.1870089322,0.0447349623,0.0367527194,-0.0143876653,0.0241524521,0.2787659764,0.1582866609,0.0996200293,0.1799615026,0.1286295503,-0.3068122566,-0.2155839056,0.2058727145,-0.0370233655,0.1797150224,0.1938584298,0.1194465235,-0.1883077323,-0.1963335127,0.1248443052,0.0569100343,0.1447522044,-0.3052254319,0.06636329,-0.5841510296,-0.4200511277,-0.6971724629,0.0577419028,-0.5651504397,-0.2432687432,-0.1315678656,0.2803404629,0.0072184922,-0.1125031263,-0.0989015698,-0.1876918823,-0.4081627429,-0.5757272243,0.0602040142,0.0518222526,-0.2901481688,-0.1236888468,0.1656201929,0.2025422454,0.1851045638,-0.0894983187,-0.3911500275,0.0396242253,-0.2137362659,0.0148326224,0.2922559977,0.1402152479,0.2720702291,-0.1263947934,0.1923101693,-0.1844576597,0.0526951179,-0.2295178324,0.0896231309,-0.0221743863,0.1062931493,0.0566955768,-0.3506991267,-0.1971296519,-0.2604816556,-0.497148633,0.3187969029,0.2221008539,0.3196673393,0.7840191722,-0.1315129101,-0.0195281021,-0.2262925655,0.1622522622,0.184805125,0.0358595662,0.2581927776,-0.1146698445,-0.0830991343,-0.0601112396,0.0697410107,0.0492029712,0.4465402365,-0.2789008915,-0.0321907923,0.0596203431,0.053092245,0.0721440166,0.1278874129,0.4659276903,0.2913442552,0.136693418,-0.0429818034,-0.3192870617,0.2071259171,0.581269145,0.3781496286,0.2848648727,0.198588416,0.1895773858,0.3258741498,-0.0334559791,0.2295010835,0.2446362823,0.0202223547,0.5292630196,0.1749498993,-0.1418644935,-0.3167340159,0.1077736914,-0.0488310903,0.046114672,0.1397987902,-0.1316756308,-0.0335369259,-0.054395549,-0.0411403403,-0.4204459488,0.0880989432,-0.0913277492,0.3229241967,0.187847659,0.0720073879,-0.2425556332,-0.2681694925,0.2999157608,0.3304556608,-0.0219774134,0.3394544125,-0.2850991488,0.0375343561,-0.3237114847,-0.3741632402,-0.0990820974,0.4699094296,-0.1400239468,-0.0109422747,0.2265978307,-0.1689763218,0.3078696132,0.0618509054,0.2381908596,0.1564677209,-0.1248227656,-0.5593657494,-0.0138598075,0.0608834475,0.4183636308,0.2125428021,0.0158925112,-0.0787872672,-0.1674569994,0.0450293683,0.0172659215,-0.186270684,-0.2216734737,-0.3028729558,-0.4074144363,-0.2037132978,-0.0045979018,0.0468649045,0.0947290957,0.2849694788,-0.0375725515,0.231184572,-0.0778232738,-0.0490505993,-0.2754959166,0.1342771649,-0.0307983924,0.0601019599,-0.0201706607,0.0564728118,-0.2783775926,0.337590307,-0.043848969,0.1900481284,-0.3183274269,-0.0657870173,0.2830541432,0.531448245,0.0186985154,0.0987232476,-0.0702538788,-0.4763909876,0.0237488337,0.0817925036,-0.0662815198,-0.3168521225,0.2155394852,-0.3315450549,0.1977359354,0.1427229345,-0.252774179,0.2816808224,-0.133167088,-0.4260431528,0.2371695638,0.2485246062,0.9905763268,-0.2436975986,0.3380008638,0.0949408785,-0.4214064479,0.3640119135,-0.2840444148,0.291951865,-0.4717691839,0.3271916807,0.0250540413,-0.346821785,-0.1333052963,-0.0808614269,0.1263559163,0.2611092031,0.1430902332,-0.0839995295,0.1720122993,0.3603963852,-0.16223526,-0.3209450543,-0.4041652083,0.0072639654,-0.0752794221,0.4611846507,-0.1371786296,-0.2295722067,0.1437276602,-0.1819764078,0.0348361023,0.15389961,-0.1240403354,-0.0089389216,-0.0296094324,-0.0094486214,0.008872306,0.1726734638,-0.0743677542,0.5297319889,0.0576445498,-0.2772844136,-0.229626134,-0.2684810162,0.1140008792,0.4309659004,-0.0211047158,-0.033872202,0.0209359135,-0.379983902,0.0266886093,-0.2590024471,0.1962019056,0.0089064892,0.2537415028,0.1914273798,-0.1751874685,0.0173283927,0.0195889622,-0.040992409,0.0452312604,0.2079636157,-0.1550946832,0.1690049917,-0.0811460242,-0.3309135139,0.0828366503,0.5577000976,-0.0600151755,0.1329279095,0.2639296055,-0.3103864789,-0.1690565795,-0.1576101929,-0.0620957203,-0.5718548298,-0.0864445642,-0.1173255295,-0.1262333095,-0.3579061925,-0.2837916315,-0.0830994248,0.1901943833,-0.0206201766,-0.3535207212,-0.3866328001,-0.1874477267,0.1063271239,0.1185432598,-0.0102188569,0.21497491,-0.0358131714,0.0734799504,-0.1466747969,-0.1507854164,-0.014602134,-0.2211660147,0.0565903038,0.0260124262,-0.201083079,-0.0164540503,0.0352763794,0.066640377,0.1325965077,-0.0256871562,-0.0348101072,-0.1901123822,0.2159983963,0.1355572194,-0.0695012659,0.2680897415,-0.4432977438,-0.0740878358,-0.0213888288,-0.2436885387,0.2915946245,0.2023130208,0.0082690576,0.1977012455,0.2854662836,-0.1080868021,0.3883495629,-0.407450676,-0.1672105789,0.0570151433,0.18978706,-0.0244023409,0.036105115,-0.3062739372,-0.172618255,0.001314131,0.1376888454,0.1503579021,-0.1284583956,-0.1020126045,0.2298369259,-0.0594988242,0.4817076325,0.0433282219,-0.1527951062,0.4984946251,0.0178825706,-0.00218212,-0.0885135904,0.2093964666,-0.2022422552,-0.102338016,0.4958066046,0.2620566487,-0.1986763924,0.2865676284,0.0292145424,0.3147399724,0.0256736446,-0.05744306,0.0836313888,-0.0550510921,0.0174757484,-0.1141674444,0.118449524,0.219936654,0.1767617315,0.1238600314,-0.1511375606,-0.0160604846,-0.1166117489,-0.0808954611,0.0021123425,-0.3904837072,-0.0148650734,0.6448432207,0.1958512515,0.151601702,0.1036148965,-0.3628794253,0.0036809165,0.2076685131,0.093315959,-0.195606783,-0.0703347474,-0.1616565883,0.0210513212,-0.3600159883,0.0797492191,-0.0567715801,-0.4050351679,0.2851819098,0.1234713569,0.1535702348,-0.443541795,0.1153307706,0.1981543899,0.0157187786,-0.2311701924,0.1667555273,-0.4791303277,0.3052455187,-0.2733854055,-0.2546569109,0.1634381562,0.3410562277,-0.1024389863,-0.112022467,0.0654759035,0.1713257581,-0.2860589027,0.0029460776,-0.0672470853,0.7874848843,0.1904533505,-0.0534625463,-0.0291887019,-0.3258475363,0.220803082,-0.2670594454,0.0503324643,0.6900346875,0.3053019345,-0.031440679,-0.0856998339,0.2216227204,-0.2574249208,0.0926635116,0.1983491033,0.1986164749,0.1361198127,0.0112204272,0.0524485111,0.293497026,0.3950175643,-0.0037643451,0.1215990111,0.066153273,0.0040794043,-0.5283036828,0.1576404124,-0.3230124414,-0.2439773679,-0.0172331184,-0.1373329461,0.0265005175,0.07364209,-0.2676175535,0.1852428913,0.1084587052,-0.0692891404,0.0906953514,-0.0457017273,0.1658130586,-0.0438650176,0.0029077958,-0.3991202712,-0.0527482368,-0.1488348842,-0.0447633825,-0.1245916337,-0.1464538127,-0.1560395807,0.2114462703,0.134619385,-0.0255455356,0.0093025016,0.0284068771,-0.4777077734,-0.356910497,0.0273276083,-0.3115734458,-0.1141432747,-0.1120087579,0.0585229099,-0.316860944,-0.0571983829,-0.6203392148,0.2801817656,-0.0818860754,0.2548214793,0.2169451863,0.0841783807,-0.2436900437,0.2907480299,0.109635964,0.1155238152,0.0551214777,0.0580844469,-0.3958208263,-0.1631210595,0.4747029543,-0.002312015,-0.1922682524,0.1842209846,0.276714325,-0.0266984086,-0.0891864374,-0.3841156065,-0.2153951526,0.2019433379,-0.1862558573,0.0974386409,0.0430322923,-0.3200283647,0.032103315,-0.0103072012,0.3574393392,-0.071097225,-0.1901919693,0.0787406489,-0.334531188]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"When you load a text file for the first time with `nlp`, the file is converted into Apache Arrow format. Arrow allows to use memory-mapping, which means that you can load an arbitrary large dataset.\r\n\r\nNote that as soon as the conversion has been done once, the next time you'll load the dataset it will be much faster.\r\n\r\nHowever for a 1TB dataset, the conversion can indeed take time. You could try to load parts of it in parallel, and then use `nlp.concatenate_datasets` to get your full dataset.","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":88,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nWhen you load a text file for the first time with `nlp`, the file is converted into Apache Arrow format. Arrow allows to use memory-mapping, which means that you can load an arbitrary large dataset.\r\n\r\nNote that as soon as the conversion has been done once, the next time you'll load the dataset it will be much faster.\r\n\r\nHowever for a 1TB dataset, the conversion can indeed take time. You could try to load parts of it in parallel, and then use `nlp.concatenate_datasets` to get your full dataset.","embeddings":[-0.2580243051,-0.1429562569,-0.0743959993,0.201565221,-0.1192451119,0.0688815638,0.0702806711,0.448238492,0.1473057866,-0.2636481225,0.148645401,0.2279233038,-0.106906116,0.061183054,0.1534855515,0.0742182583,-0.0200049654,0.2010743171,-0.1128234789,-0.1518828869,0.0267371889,0.0252636019,-0.1835510284,-0.2006816268,-0.1286405623,-0.0606779456,-0.0646085069,0.0253585782,-0.0933855772,-0.2799025178,-0.0734302998,-0.0075133867,0.0779141262,0.163817063,-0.0001190489,-0.1963580251,0.3564788699,0.0899229571,-0.2416128963,-0.1499068588,0.2623203397,-0.6448490024,0.2374050021,-0.1103647798,0.0207275543,-0.1262657791,0.1428301334,-0.1979647428,0.1357072145,0.0908239335,0.1139508039,-0.0867789313,-0.2106594443,0.3208089769,0.1221834123,0.1906038821,-0.0291808713,0.2637990713,0.2932355404,-0.2001265585,-0.3337846398,0.0332033969,-0.1195470244,0.1559601575,0.3333066404,0.0464646779,-0.0898462012,-0.076048851,0.1180850491,0.284802258,0.5425196886,-0.0258227885,-0.0519602261,-0.4517751336,-0.0019575551,-0.1291745007,0.1924464405,0.3310850263,-0.2503350079,-0.2429402322,-0.2714089155,-0.3056858182,-0.126796633,0.2447672635,0.0714997202,0.059363354,-0.0026139019,0.2128312439,0.2367528975,0.0035537938,0.0124043925,-0.197156623,0.2276108563,0.4581683576,-0.3844109774,-0.0025614724,0.0357512124,0.1673823744,0.1374229044,0.1301303357,0.1820942909,0.0464357771,-0.2015420198,0.0214452725,0.2738535106,0.342161119,-0.2031341344,-0.1676602513,0.3533450067,-0.2663768828,-0.1084761694,0.2691608071,-0.1542274356,-0.1744618267,0.0330999903,-0.4120718241,-0.1778144836,-0.3051295578,-0.0733598024,-0.0831516236,-0.0601924062,-0.2440093905,-0.0297536664,0.1259436011,-0.3379230201,0.2886785269,0.1437388062,-0.1046234518,-0.3226682842,-0.0085545257,-0.0555738546,-0.0189568829,-0.1516158581,0.3027732968,0.5116083622,0.0325876735,0.1436165571,0.0451546311,-0.2622653842,-0.0435528383,-0.177493304,-0.2493155152,-0.0659802258,-0.0313736387,-0.0227674898,0.1939994693,0.0823346153,0.3259328306,-0.4238597155,0.1759598851,-0.3380049467,-0.0592473708,0.0427806303,0.0781180412,-0.3854233623,-0.3015832305,-0.3347423077,0.4295698702,-0.060089048,-0.1445232779,-0.2657331228,-0.0977050513,-0.2564182878,-0.002068582,0.0752363428,0.1738006324,-0.1466029882,-0.0385921225,-0.0385037363,0.4130982757,0.2468641102,0.5918000937,-0.1988864988,0.2207202166,-0.1733455658,0.0261158813,0.5016420484,0.124737151,-0.3987202644,0.5826674104,-0.2134343833,0.0198439807,0.1535961777,0.3465153575,0.010509843,-0.0216892827,0.2502728105,0.700524509,0.2258731574,0.1835709065,-0.5376799703,-0.1484410465,0.2145620733,0.3103547394,-0.1488063633,0.0026408879,0.0340144858,0.0508759953,0.4399492741,0.1979080886,-0.0152008375,0.3547917604,-0.3042240441,-0.0217803176,0.0680044815,0.2160368264,-0.0974014997,0.1468753815,0.0538267344,0.087903969,0.3113023341,0.1676249653,-0.2224716693,-0.2288695425,0.0802318975,0.2670134306,-0.0229946245,0.0762374923,0.0345356502,0.1919613034,-0.1195787191,0.3874516785,-0.0643284991,-0.1230421886,-0.438457191,-0.0962342992,0.2748400569,-0.0464416407,0.4039375484,0.2337038219,-0.1707218587,0.0219945908,-0.1359869242,-0.0038257232,-0.3297897875,0.2558680177,0.0886448249,0.0330987796,0.0563279726,-0.1082670093,0.395526588,0.2351761758,0.3025710583,-0.3483243585,-0.2962740958,0.4585195184,-0.1163601726,0.3341569901,0.1840611398,-0.3987696469,0.0885084271,-0.0222317837,0.0819798484,0.3170708418,0.8408857584,0.0845447481,0.6601690054,0.2273520827,-0.1195096523,-0.0499757677,0.4336675107,-0.1112653166,-0.2259491682,0.5013071895,-0.1831347346,-0.406235069,0.0281976368,-0.1438335925,0.391063869,0.2931973934,0.1178413928,0.0659755021,-0.1019050628,-0.1797468215,0.2066102773,-0.1422030032,0.3015500903,0.0961891562,0.3848954737,0.0827190503,-0.4828436673,-0.2860980928,-0.0393602513,0.3457221687,-0.0580390021,0.108785294,-0.0568862036,-0.3092046976,-0.2312429249,0.0379159488,-0.3179530203,-0.2344306707,-0.010179257,0.0390949994,0.4102935493,-0.176364392,-0.0082466369,0.1628456563,-0.0902125314,-0.4379183054,-0.2815780044,-0.1494336426,-0.4388256371,-0.0467550047,0.3423187733,0.2407114059,0.3009675145,0.1920312345,-0.1717655063,0.2372120619,-0.0423781425,-0.1797998101,-0.0757500753,-0.0906863809,-0.2423758209,0.1320564747,0.1255113333,0.0224761236,0.0044578537,-0.2911377549,-0.0708213076,0.210793227,-0.0324204527,-0.0122498367,0.0540511534,-0.265748471,-0.1826431751,-0.2270224094,0.4667129517,0.1411596537,0.1653353274,0.2913052738,-0.1021565497,0.2856541276,0.1465475112,0.2479116172,-0.0135069024,-0.1057498008,0.3281100094,0.2865343988,-0.2873679399,-0.3175397813,0.0452958569,0.1026076004,0.1219858006,-0.7564432025,0.4603051841,-0.4373180866,-0.0935813338,-0.1840261519,0.1150645688,0.1560970396,-0.2291910648,-0.0480020866,0.2379166037,-0.0687173679,-0.0098802671,-0.0480650626,0.2736463845,-0.0779336616,0.3746143579,-0.0265049282,0.1672289968,0.1788381189,-0.0635445639,0.1850762814,-0.0453564562,0.0589768477,-0.2291685045,-0.1360009164,0.1213740259,-0.1993957162,-0.0526299067,0.3622400165,0.0995582789,-0.4794427752,-0.268851459,-0.1049255803,-0.0438407101,-0.1263776422,0.2061017007,-0.5250543356,-0.0895582885,0.0315530896,-0.1230122298,-0.2496896237,-0.7182970643,-0.1159708574,0.0851065964,0.1345787644,0.0686195269,-0.2203113139,0.1092063636,-0.8535843492,0.1522729546,-0.2927674353,0.4942768216,0.1012598723,-0.2130983621,0.3597968519,-0.1164099202,0.2636369765,0.3609699309,0.1064669192,0.1501381546,-0.0586203597,-0.5524616241,-0.0408009924,-0.1757726967,0.2366239727,0.4474132657,0.3369255066,-0.2200979292,-0.1212978214,0.0034103254,0.3156624138,-0.0433465727,-0.5378076434,-0.4096361697,-0.1558472365,-0.02519482,0.025566766,-0.3112889528,0.1965810508,-0.1206171438,0.0235489588,-0.0312657245,-0.1023792401,0.2745970786,0.0140290419,-0.2206919491,0.0643055364,0.3005280197,-0.0779168233,0.3423224986,-0.4795140326,0.463601321,0.3183017969,0.061168503,0.0151963895,0.0597666055,0.3986094892,0.29318434,0.0773352608,0.0504431874,-0.2372925729,0.19027403,-0.1804479957,0.063873902,0.1414530277,-0.0688966811,-0.4214698672,-0.3904386759,0.9887099862,0.1678908765,0.0653035417,0.4738778174,-0.3927305937,0.0651942641,0.335981667,-0.3441711962,0.8150549531,-0.478022933,0.3495225608,-0.0670739338,0.0347133204,0.4023123384,-0.1422496438,0.0654206797,-0.3965914249,0.2229334414,0.2521339655,-0.0400774404,-0.0607664138,0.305562973,-0.1380981356,0.3873783052,0.4162447453,-0.3467713594,-0.151673913,0.7324685454,-0.1925999075,-0.2279392928,-0.5543438792,0.0806716979,-0.2827617526,0.4859900475,0.0569211282,0.0594941489,-0.1382347047,-0.0232668389,-0.2746174037,0.0934718996,-0.0221110806,0.0251461789,-0.1356394291,-0.3379199207,0.1115451679,0.0942991301,-0.3294048905,0.0775134042,-0.2826444209,0.0387522876,-0.4085694849,0.2268725187,0.4534935653,-0.0715886429,0.1698086709,-0.1747025996,0.02782543,-0.1191384792,-0.0530398972,-0.2230220586,-0.0910930187,0.1972079277,0.1085301414,0.141653344,-0.3056222498,0.2965962291,-0.0367921852,0.0183540806,0.0337991677,0.361548841,0.0726236776,0.2760682106,0.156723246,-0.4740467072,0.0187790357,0.5012875199,0.095610179,-0.1507552564,0.4508935511,0.1583354473,-0.1157471314,-0.1204261333,0.1693951637,0.0386524759,-0.2350002676,0.0467653759,0.0816519111,0.1031811759,0.2221589535,0.0563776642,-0.067580305,-0.1099897027,0.0928842425,-0.4718827009,-0.2744015753,0.0782025233,-0.0862439349,-0.0269839447,0.0571640059,-0.0591558591,0.2229402959,0.3382492661,-0.1543870866,0.1469291449,0.0962574035,0.0974756479,-0.0990156457,-0.2196989208,-0.00717609,0.3180195689,0.0732910633,-0.074076131,0.1194782779,-0.1856727898,0.0396323055,0.1955276579,0.0100293392,-0.0282179695,-0.0172089282,-0.3136977851,-0.0298536643,-0.2065971941,-0.1353016645,-0.0746403113,-0.0685594007,-0.0881907046,0.0215052441,0.0047976268,-0.0432660878,0.2160219699,-0.4406600893,0.2270668745,0.1653146744,0.0807469115,0.3621737361,-0.0014526105,-0.5631071925,-0.0575418547,0.1354710609,0.0851349011,0.1073212624,-0.1598534882,0.1474939883,-0.1557088643,0.1999237388,0.5425590277,0.0573213547,-0.1970003843,0.1085044667,0.04933111,-0.2381014526,-0.0085443845,0.2233696878,0.0445110798,-0.0989393219,0.1071872115,0.1314359754,0.0226621311,-0.3565069437,-0.357015729,0.2992474139,-0.1193618774,-0.2936942875,0.1449043304,0.1741228402,-0.146610871,0.2117778361,0.0921040773,0.0983328447,0.4157552123,0.1702169031,0.3688651323,0.108068116,0.4503047764,-0.3991969228,-0.3068241477,-0.1270230561,0.4518741667,0.1279358864,0.198924005,-0.0008423597,0.2880346775,-0.0131749138,0.0271890406,-0.175269857,0.0176179968,0.0247530285,-0.1582579166,0.148402974,0.0871673375,-0.2405381501,0.2644072473,-0.3471239805,-0.0155325392,-0.0242402479,0.3398851156,-0.0073964354,-0.4118709862,0.1741108149,0.0005494563,0.1975174248,-0.1807393581,0.1701130718,-0.2927471101,0.1100096777,0.0769691467,0.1250455976,0.2397668064,0.372766614,-0.0875177011,0.1372534335,-0.1563856453,-0.0505314954,-0.2074498385,0.3958819509,0.2734569311,0.3311544955,0.1413858533,-0.0814320371,-0.0653947592,0.167497322,-0.1196422875,-0.3277775645,-0.2014134973,0.2485801727,0.1475756168,-0.2981002033,-0.5778741837,0.1361485273,-0.3145260215,-0.2423122525,0.197441563,-0.1156183109,-0.0432786793,-0.0786160082,0.022219792,-0.3155360818,0.6768491864,0.4696147442,0.3979502916,-0.0979028568,-0.2553620338,-0.4972056746,-0.0835588649,-0.3635211885,-0.0955648571,0.0487689674,0.312739253,0.1088964567,0.1623489708,-0.3987654746,0.2395132333,-0.1641736329,-0.0754955783,-0.2630473971,-0.0352834202,-0.149156034,0.1646288186,-0.2576058507,-0.2837482691,0.1094364747,-0.1651513875,-0.1021280289,-0.137277171,0.0079127913,0.0893052742,0.4669982195,0.3119837046,0.0175567884,0.3307690322,0.0810269043,-0.105068028,-0.4045045972,0.1325248927,-0.1420278549,0.2934335172,0.0654425099,-0.02041151,-0.0650469214,0.2211392522,0.1623337418,0.2604758441,-0.2782451808,-0.0273939464,0.054119952,-0.3129333854,-0.0963110924,0.3950642645,-0.0569718704,0.4237586856,-0.0561881028,-0.1047647893,-0.4859927893,0.0311776251,0.232667312,-0.3553600013,-0.0956417173,-0.0655777678,0.3098990321,0.1066006348,0.2938704491,-0.397606194,-0.1224124953,0.3562179208,0.0164163746,-0.0080445902,0.4049390554,0.0255446862,-0.0593668558,-0.0508872233,0.308942914,-0.1203017309,-0.1957064122,-0.3198836148,-0.2703247964]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"Humm, we can give a look at these large scale datasets indeed.\r\n\r\nDo you mind sharing a few stats on your dataset so I can try to test on a similar one?\r\n\r\nIn particular some orders of magnitudes for the number of files, number of lines per files, line lengths.","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":50,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nHumm, we can give a look at these large scale datasets indeed.\r\n\r\nDo you mind sharing a few stats on your dataset so I can try to test on a similar one?\r\n\r\nIn particular some orders of magnitudes for the number of files, number of lines per files, line lengths.","embeddings":[-0.2486386895,-0.2108440548,-0.130509004,0.2715967298,-0.0312591419,0.0532569923,0.1728605479,0.3125979006,0.3433538675,-0.2542510331,0.1040127277,0.2262744755,-0.1480088234,0.2491922528,0.1742842495,0.0440121144,-0.0694732442,0.1338013709,-0.0940501094,-0.2089585066,0.0764913261,-0.0042570992,-0.2356909961,-0.2695990503,-0.1871654689,-0.0503750071,-0.0101462258,0.0359051377,-0.1234337389,-0.2105946243,0.0055685746,0.093107231,0.0121268416,0.3132479787,-0.0001156001,-0.2746174634,0.3784097731,0.0973197743,-0.1762195677,-0.0817299411,0.208563149,-0.6034761071,0.1212058067,-0.1240255088,0.0174892824,-0.0432460383,0.076217033,-0.1090654433,-0.0044657835,0.1309253275,0.1321333945,-0.0904468521,-0.2628418207,0.2112321109,0.0767723173,0.2716992795,-0.0006690567,0.357282728,0.4154173136,-0.1923882812,-0.278337121,0.0211374331,-0.1434020549,0.0957906246,0.2736075819,0.0057202582,-0.1591182649,-0.0868941173,0.1164136529,0.4194946885,0.5542607903,-0.1248918623,-0.0682062954,-0.4857752025,0.0235811286,-0.1223733574,0.1659093499,0.2761653662,-0.1462365985,-0.2896139622,-0.4326581061,-0.2402228564,-0.07956025,0.1657622755,0.0045831068,0.0249823406,-0.013176525,0.1476900578,0.1474984437,-0.0551853701,-0.0441146642,-0.2092203051,0.2359211296,0.4067614675,-0.4368546903,-0.1065014154,0.0749983415,0.1771083176,0.1731025875,0.067776598,0.1592899114,0.0109073902,-0.1891322881,0.0263830461,0.2703831792,0.3900518119,-0.1767294258,-0.1684203744,0.3515302837,-0.1697862595,-0.2296455353,0.2596701682,-0.1128102168,-0.1672840863,0.0136890551,-0.4427077472,-0.2214354128,-0.352612406,-0.0619462952,-0.0342046469,0.0437178351,-0.2788405418,0.1019065306,0.2122916579,-0.3088967204,0.290129602,0.0159919169,-0.0993956551,-0.2755247653,0.0564937964,-0.1495098025,0.04312874,-0.1205473617,0.2141251713,0.5559558868,-0.0202070978,0.1875703633,0.1515731514,-0.2232275158,-0.1161600724,-0.2009130865,-0.342775315,-0.0972758606,-0.0436135381,-0.0318155065,0.1685258895,0.0070512062,0.3376305401,-0.3838861883,0.1070853099,-0.3090755343,-0.0513507724,0.0715455115,0.1486166865,-0.4272139668,-0.3566762507,-0.2842693329,0.3984337747,-0.102804184,-0.1434931308,-0.2055092156,-0.1970776767,-0.2360456586,-0.0038905607,0.0610405579,0.2080202997,-0.1028612778,-0.0373454466,-0.0420119278,0.327589184,0.3008471429,0.5783417821,-0.1911014467,0.2444197983,-0.1414471567,0.1041969508,0.4420722723,0.0621602535,-0.4157561064,0.5438098907,-0.1831786335,0.0356256217,0.224320814,0.3146437705,0.0846073031,-0.0453319326,0.3129044175,0.6097602844,0.2616622448,0.2236879766,-0.5192388892,-0.0816205963,0.2502227128,0.4377360046,-0.0662628189,-0.110010758,0.0123595959,0.021278359,0.4793024957,0.1545644701,-0.1717208624,0.3233118355,-0.2008062452,-0.0290947109,0.0442672558,0.1464744657,-0.0480411388,0.2284378856,0.1344222724,0.0384671986,0.3972015977,0.1483384818,-0.1765311062,-0.1649584025,0.0871813297,0.1871472597,0.0015195954,-0.0436734445,0.073898837,0.2142150402,-0.1339357942,0.3165629506,-0.2223711163,-0.1686550677,-0.3253329098,-0.0225288868,0.3222882748,-0.0935231373,0.3534987271,0.2709269524,-0.1474179626,0.0001545947,-0.1234338731,-0.102943942,-0.1997460574,0.2632876933,0.1220244989,0.1304640919,0.0262938794,-0.0328233205,0.4114790559,0.1348858625,0.2995357513,-0.3384072185,-0.3134112954,0.412679255,-0.1818640232,0.2255385816,0.1676179767,-0.512196362,0.1029440761,-0.1461997926,0.1794183105,0.3618108332,0.9970802665,0.0423891991,0.6520142555,0.3005900681,-0.157351464,-0.0620913506,0.4747378826,-0.0766688213,-0.1524769813,0.4595927298,-0.2282647789,-0.3162681162,0.0290650055,-0.0418973565,0.3655613363,0.2885782719,0.05707426,0.0865556076,-0.0182121042,-0.2234887928,0.206138432,-0.1537454426,0.1727264971,0.2333604395,0.3821172416,0.0576562136,-0.4399826229,-0.2300445288,-0.0233045202,0.3337501585,-0.0937670544,-0.012236774,0.0247635879,-0.2615176737,-0.1847613454,0.0775963962,-0.3184903264,-0.2108866721,-0.0157956146,-0.0252601802,0.4601272345,-0.191271171,-0.0180885643,0.1107639596,-0.1913944632,-0.499602437,-0.2530306876,-0.1412377506,-0.3140839338,0.0136330742,0.2918311059,0.2350362688,0.3330124915,0.1568288356,-0.1925464123,0.2865509689,-0.0626537725,-0.1530089378,-0.0221395642,-0.0233355109,-0.1805532277,0.1107464135,0.2681524754,0.0269478168,0.0560398772,-0.3479304612,0.0260532834,0.1491565704,-0.044270277,-0.0498942733,0.0474365987,-0.2053926289,-0.2814620137,-0.217321381,0.5193728805,0.1146609113,0.1249854565,0.2194504887,-0.1814265251,0.2895323932,0.0622400455,0.2473640293,-0.0041090539,-0.2301887572,0.2904568911,0.3158654869,-0.2225532532,-0.2283245325,0.025532674,0.0695408583,0.0832812041,-0.7475146651,0.4059768319,-0.4893417358,-0.0323973782,-0.0399835818,0.0885111019,0.1408664435,-0.2680480182,-0.0466729552,0.2261449993,-0.1216790006,-0.0299028456,-0.0066040559,0.2154010236,-0.217801258,0.3159351349,-0.1042943895,0.0511522889,0.1029468924,-0.1928379238,0.1229181513,-0.0626878962,0.0938585922,-0.1282745004,-0.2376913577,0.2467371225,-0.0836677924,-0.1745881885,0.3359709084,0.0408466719,-0.4224254489,-0.2895857394,-0.1774409562,-0.0099108936,-0.094366625,0.1670954525,-0.2766082585,-0.0619886741,-0.0030025798,-0.1154807061,-0.2226512581,-0.7354319096,-0.1276739091,0.1004891917,0.1220178977,-0.0587942488,-0.1328059137,0.0061601438,-0.7636799216,0.1280788034,-0.1921784282,0.427696377,0.0147736175,-0.1615127325,0.3043775558,-0.1195620671,0.3274208009,0.2310962379,0.2009230852,0.1975379884,-0.1139947474,-0.4052624404,-0.0265557393,-0.2160053849,0.2019543797,0.5574670434,0.3640386462,-0.1233929023,-0.1083953083,-0.0000924866,0.3028017282,-0.0054603601,-0.647436738,-0.2758657932,-0.0769907907,0.13114959,0.0913962051,-0.3576986194,0.1455600858,-0.2545612454,0.0722507462,-0.026150804,-0.0077564884,0.2913787663,0.0363023803,-0.2095211446,0.0593112856,0.3716659248,-0.1332944632,0.3029575348,-0.4588218629,0.4044708312,0.1889938563,0.1596022546,-0.0264290161,0.0294620655,0.4311202168,0.2654605508,0.0792784244,0.0055525829,-0.3309392929,0.2348129749,-0.2141076475,0.0845479593,0.1717277765,0.0702350959,-0.4354387522,-0.3807117343,0.875483036,0.1672880054,0.0593722649,0.3597486913,-0.4834032953,0.0001693089,0.1790502965,-0.3303281367,0.73957026,-0.4816345572,0.2712956369,-0.141192928,0.1208735406,0.3883980811,-0.106818758,-0.0140716629,-0.4078050554,0.0012217812,0.2903621197,-0.0737047866,-0.0712144002,0.262981832,-0.0949847326,0.4188192487,0.3380481005,-0.3610144854,-0.162867412,0.7147482038,-0.1446309984,-0.1830555052,-0.5921898484,0.1191916838,-0.3006485403,0.4313510656,0.0359367728,0.1080778018,-0.1226973385,0.0350912511,-0.3221391141,0.1057326794,-0.0214598309,-0.1039429829,-0.0558501109,-0.3602657318,0.2104269266,0.1329291463,-0.3163472414,0.1919997334,-0.2904361784,0.1694648266,-0.4294074178,0.2215346694,0.5690383911,0.0400773399,0.0920963138,-0.1330712736,-0.016483916,-0.1876693666,-0.0030463687,-0.084252581,-0.083965838,0.128794685,-0.0102347406,0.0207697581,-0.3285716474,0.2691439986,0.0174874607,0.0442583747,0.0795395225,0.3560287654,0.0995687768,0.3716541529,0.0781682283,-0.4419523478,-0.007497557,0.5539149642,0.0446981527,-0.093468681,0.3801753819,0.2489362508,-0.1149856672,-0.1516567022,0.0772769898,-0.1090517119,-0.1209493801,0.0214856509,0.0952614769,0.1124397814,0.215258494,0.0036678049,-0.0396408327,-0.1684331447,0.0969597325,-0.4598970413,-0.270339191,0.0183253828,0.0163764302,-0.0594286658,0.0189467538,-0.0193063524,0.2558468282,0.2553028762,-0.2078297138,0.1938476861,0.0379539318,0.0008026081,0.0161528941,-0.2550886571,0.0776679888,0.134703815,0.13364923,-0.169539094,0.131592378,-0.2541551292,0.0428642519,0.1551256776,-0.050730858,0.0469622277,0.0386367664,-0.29087165,-0.0510552935,-0.1894154251,-0.126094982,-0.037235681,-0.0687760115,-0.0327706896,-0.0266296323,-0.0183576681,-0.1332514584,0.2710002661,-0.4025324583,0.1420174688,0.1175244898,-0.0147002209,0.3519359231,0.0105445655,-0.5127433538,-0.1042865664,0.1264260858,0.1025161296,0.1936853379,-0.1674493104,0.1844198555,0.0059566456,0.278980732,0.5280631781,-0.0062940582,-0.1858847588,0.0784659609,0.0974324644,-0.2273055762,-0.0144933471,0.3085356653,0.2134364098,-0.0792190358,0.1308327615,0.1391678751,-0.0282361247,-0.3540787399,-0.3104559183,0.406298548,-0.154144451,-0.2605230212,0.142459482,0.190019086,-0.125462383,0.1808305532,0.1629980654,0.0297829751,0.4247863591,0.2490755022,0.3453607559,0.0390368067,0.4415322542,-0.3563576341,-0.2952813506,-0.1471569091,0.508295238,0.0578821972,0.1391096711,-0.1498195529,0.2389546782,-0.0552015677,0.0499007516,-0.2160600722,0.0952132344,0.0089178393,-0.0742415637,0.1654215008,0.0643649325,-0.1956367046,0.2332742363,-0.3532126844,0.1447060406,-0.0955291167,0.3290522695,-0.0386093445,-0.4128954709,0.1281604171,-0.0209207945,0.232691735,-0.2117918581,0.1328069568,-0.2030369043,0.0903227031,0.0802235976,0.1932073683,0.265167594,0.2962996662,-0.0642591342,0.0547845215,-0.1846187562,0.0069711646,-0.2256305069,0.4160029888,0.1715600938,0.3349130154,0.1330730319,-0.0182283837,-0.0888316259,0.2212150097,-0.1641310155,-0.2552423477,-0.1376208812,0.2702530622,0.1682336777,-0.2660666704,-0.6023842692,0.1053807661,-0.3172764778,-0.193238169,0.1913498789,-0.0232282281,-0.0075537083,-0.0448897444,0.0491049588,-0.3418805897,0.6926574707,0.4073372483,0.4532288909,-0.0886882022,-0.223482132,-0.492606461,-0.1371570528,-0.3056133687,-0.1503495872,0.1317792237,0.2089385092,0.0859815776,0.0431547835,-0.41997841,0.2226309627,-0.1943285316,-0.1410339624,-0.3337698281,-0.030956693,-0.3584557176,0.0855696872,-0.1483549625,-0.1733203083,0.0769327134,-0.2930166423,-0.0804651678,-0.1545779407,0.1351483464,-0.0031841991,0.4502994716,0.2374556512,-0.0331187099,0.4438133538,0.0437533557,-0.0190128218,-0.3446159363,0.1138082966,-0.1254460365,0.2371196449,0.0708208978,-0.0549821183,-0.0471285991,0.2642510533,0.2522796094,0.1962285787,-0.2932772636,-0.0220112149,-0.0116341645,-0.4068407714,0.0198660903,0.2453898489,-0.0170996487,0.4612652361,-0.1009520888,-0.1236659959,-0.3529674113,0.0987660289,0.2898667753,-0.3197882175,-0.0076201493,-0.2301364541,0.3354278505,-0.0057791467,0.3162283301,-0.3158343136,-0.033767473,0.3471738398,0.0945029706,0.0538208298,0.4641377628,0.0120115336,-0.1097201407,0.0060082669,0.2013753653,-0.157434985,-0.1576907933,-0.3613035381,-0.1312541515]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"@lhoestq Yes, I understand that the first time requires more time. The concatenate_datasets seems to be a workaround, but I believe a multi-processing method should be integrated into load_dataset to make it easier and more efficient for users.\r\n\r\n@thomwolf Sure, here are the statistics:\r\nNumber of lines: 4.2 Billion\r\nNumber of files: 6K\r\nNumber of tokens: 800 Billion\r\nThe number of lines is distributed equally across these 6k files.\r\nThe line length varies between 100 tokens to 40k tokens.\r\n","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":79,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\n@lhoestq Yes, I understand that the first time requires more time. The concatenate_datasets seems to be a workaround, but I believe a multi-processing method should be integrated into load_dataset to make it easier and more efficient for users.\r\n\r\n@thomwolf Sure, here are the statistics:\r\nNumber of lines: 4.2 Billion\r\nNumber of files: 6K\r\nNumber of tokens: 800 Billion\r\nThe number of lines is distributed equally across these 6k files.\r\nThe line length varies between 100 tokens to 40k tokens.\r\n","embeddings":[-0.2211722136,-0.069399111,-0.0759154335,0.1695175469,-0.0901319832,0.1008208171,0.2206684351,0.35938555,0.2234858721,-0.1874328107,0.1198083907,0.2741488814,-0.0474218689,0.258961916,0.1557987183,0.1031196713,-0.0261339173,0.2411929667,0.0420610458,-0.0941148102,0.0570230298,-0.0064296168,-0.2886626422,-0.3357908726,-0.2980341613,0.0214096457,-0.1646300852,0.0981199443,-0.0391979553,-0.2644697428,-0.0098630665,0.2474714071,0.0183744244,0.3060204387,-0.0001189172,-0.2587305009,0.3416219056,0.0662196055,-0.2302351892,-0.1102391034,0.2941592336,-0.623270154,0.1834743619,-0.1261273175,0.0808529034,-0.0354745463,0.1360356659,-0.2005755305,0.0374361277,-0.0441890098,0.0706425011,-0.0996934995,-0.2695122659,0.2410738319,-0.0642084703,0.1584541947,0.1085236445,0.2832097113,0.3556242585,-0.2277128398,-0.2626972497,0.0034155196,-0.1359223872,0.1137017906,0.2706817985,0.0631936714,-0.1407952905,-0.1139549166,0.0160542745,0.3350263238,0.5140841603,-0.0896612927,-0.076942943,-0.6029696465,0.0666935295,-0.223353073,0.1713902354,0.2411392033,-0.1910333633,-0.2759585381,-0.3890081048,-0.3037951291,-0.066812776,0.1294828355,0.106554538,0.107335411,0.0603799447,0.183560282,0.2886107266,0.05188445,0.1139529124,-0.2195518464,0.2720116973,0.4078280926,-0.524374783,0.0028063587,0.0755731687,0.1262906641,0.1804340035,0.134829089,0.1348883361,0.0588204302,-0.129231438,0.0144624915,0.279881835,0.3531160355,-0.1694983989,-0.1542180032,0.3105047345,-0.2360600233,-0.1550752372,0.21929425,-0.1841177791,-0.2138880342,-0.0357904285,-0.387501359,-0.2708041072,-0.2539885342,-0.1084271148,-0.1540873796,-0.0759232044,-0.2951635122,0.1072275192,0.1636065245,-0.3497135043,0.4387899637,0.0694812536,-0.0635824502,-0.3524895608,0.0796268955,-0.1247897372,-0.0223259125,-0.1649293303,0.2461105138,0.5220390558,-0.0665994883,0.0762934014,0.1878412962,-0.1300724894,-0.1539223194,-0.1207379475,-0.2807799578,-0.0596730858,-0.0509844087,0.1413926035,0.2089808136,-0.0016483591,0.2444801331,-0.3578422368,0.1410300136,-0.4562772512,-0.1309579909,0.0828592405,0.1161715761,-0.3526193202,-0.3164808452,-0.3232347667,0.4442940354,-0.0430524684,-0.1600922644,-0.3826977909,-0.0633413047,-0.283069551,-0.0609633923,0.156142205,0.2751457691,-0.1106501445,-0.0312856287,-0.1239522472,0.3189246058,0.3326700628,0.5696031451,-0.2447539121,0.3569869697,-0.2544994354,0.0757549554,0.4739772677,0.1069935411,-0.3331349492,0.5554558039,-0.2506341636,0.0948168859,0.2838944197,0.3763917685,0.0054065464,-0.0773129091,0.3804531395,0.6409078836,0.0993211567,0.2818810344,-0.4149831831,-0.1698682606,0.3565430641,0.320394814,-0.1597365737,-0.0079932474,-0.0604932308,0.0265476257,0.5553174615,0.0830677673,-0.0824094042,0.3210537434,-0.2720299065,0.0587018356,-0.0290385094,0.2288731188,-0.1514300406,0.2201864719,0.171594739,0.0485988632,0.3051662147,0.0657510608,-0.1193264052,-0.2602834404,0.0140938004,0.2311348766,-0.0359791517,0.1465760916,0.0275787972,0.0845088735,-0.1976344138,0.4954994917,-0.1594910771,-0.1767944098,-0.4130460322,-0.0086111827,0.2972275615,-0.0934247896,0.3136305809,0.2108672261,-0.0928665251,-0.0361031741,-0.0675518885,0.0511842184,-0.1797370166,0.2204971313,0.0708774626,0.0440140367,0.0122543173,0.0982807577,0.2722420096,0.0909829587,0.2742013931,-0.4083100855,-0.247096017,0.4286514223,-0.1855553091,0.3100507557,0.2059784085,-0.4950374961,0.096775569,-0.0787006319,0.1401259154,0.3078310788,0.9347122312,0.0532533452,0.5731901526,0.2683214247,-0.2199278772,0.0297283325,0.389318198,0.0419210158,-0.2104264945,0.4030595422,-0.0934763327,-0.3390423357,-0.0685087368,-0.0761765987,0.4614566565,0.2961539626,0.0580204912,0.0992351845,-0.0641596764,-0.2159323394,0.1537287086,-0.0942450613,0.2555584311,0.2294766456,0.3906415403,0.0528665856,-0.4248518348,-0.2571054697,0.1408696473,0.2379993051,-0.1475533992,0.0597534142,-0.1099086925,-0.2331698537,-0.2371135205,-0.0225572698,-0.3098423183,-0.2584304214,-0.0043134238,0.1100212708,0.3845556974,-0.1346055865,0.0247181803,0.1140392944,-0.2102863193,-0.5112171769,-0.2711211145,-0.0879051611,-0.3018433452,-0.0550501943,0.3290278912,0.2225958556,0.3865032196,0.151008293,-0.2263593972,0.2597491741,-0.1077080071,-0.2157651335,-0.0335097052,-0.0349800214,-0.1383125186,0.0841221288,0.2307423055,-0.1095758229,0.0561306104,-0.2220048904,0.0077633588,0.11203219,-0.0254599061,-0.0381971747,-0.0416819192,-0.2548500299,-0.2413040251,-0.2299583703,0.4161978364,0.0681597814,0.1740722209,0.135834232,-0.1965063959,0.1943809986,0.0443454981,0.2354702204,-0.0281360261,-0.2152674645,0.2673341334,0.265014559,-0.2199538797,-0.2671394646,0.0153983338,0.0858292282,0.1488799751,-0.5809306502,0.3345518708,-0.4182727337,-0.0692416579,-0.2077288777,0.1764719486,0.1968632042,-0.2409963459,-0.0182339493,0.1923283488,-0.1248925477,0.003407666,-0.0392328985,0.2761059701,-0.1794133782,0.3650551438,-0.0341829583,0.0906497538,0.1667857319,-0.0816928521,0.1674102247,-0.1193677858,0.1143756285,-0.2352804095,-0.2032779902,0.162955001,-0.1864682138,-0.0897848308,0.2568531334,0.0274924636,-0.4070498049,-0.3549591303,-0.0481115356,-0.0825889707,-0.085396193,0.1855717301,-0.3649887145,-0.0593573861,0.0400199108,-0.1426536292,-0.3144992888,-0.6197455525,-0.2300228626,0.0844626352,0.1375828981,-0.0740299672,-0.2278306782,0.0216147676,-0.8412263989,0.1526020616,-0.141340822,0.4096388221,0.1637924463,-0.1827286184,0.2759926915,-0.1852129251,0.4483211339,0.3038882315,0.0324006565,0.2335909903,-0.1467618793,-0.3926106691,-0.0477919281,-0.1070294231,0.2448954135,0.5613288283,0.352704227,-0.2720772326,-0.1386839449,0.0024804352,0.3695443571,-0.0126941716,-0.6122847795,-0.3650853932,-0.135060519,-0.0601901449,-0.0026265185,-0.2886678576,0.2873368263,-0.2353353202,0.0492122285,-0.0060792184,-0.0807549283,0.271171242,-0.0706987306,-0.1623620242,0.1187895983,0.4574926496,-0.0586015582,0.256457299,-0.3785607815,0.509550333,0.1829781234,0.0380830914,0.0153014949,0.0238473117,0.4698845148,0.3147840202,0.048596479,0.0663202628,-0.2757831216,0.2096984982,-0.2540392876,0.1294779629,0.1517028362,-0.0418042839,-0.5193066597,-0.5292105079,0.7771810889,0.2270498425,0.0676864982,0.4543674886,-0.5446687341,-0.0262049045,0.1825098395,-0.4340154231,0.7889279723,-0.4708952904,0.3348888755,-0.1298365295,0.1203687415,0.4411838949,-0.1899067461,0.1310924888,-0.3771680593,0.0294764508,0.2731670141,-0.0956341103,-0.1184744015,0.3176466227,-0.0738034099,0.4722570479,0.2584544122,-0.355227381,-0.1696262211,0.758644104,-0.1207198277,-0.2305506021,-0.5395560265,0.064011544,-0.2327133417,0.3239288926,0.0217368025,0.097968556,-0.0537597612,-0.0609819777,-0.2237212062,0.110068202,-0.0275562666,-0.0205850136,-0.1457604617,-0.2827556431,0.152599439,0.1017946452,-0.3295610845,0.156399399,-0.2382220626,0.1088370308,-0.3978130519,0.2081929147,0.4564130604,-0.1122383848,0.1258047819,-0.1088593304,0.014994855,-0.2972338796,-0.0178750344,-0.2510601878,-0.1294376552,0.2271600813,0.126854524,0.0994769856,-0.3163398504,0.3577563465,0.08114665,-0.0535732917,0.0244537331,0.370223403,0.0564930961,0.3644560575,-0.0608210489,-0.4419072568,-0.0018406257,0.5231863856,0.1518363804,-0.2070405036,0.4211270511,0.2474677265,-0.0902631804,-0.0967121795,0.041541744,-0.0495926626,-0.2053091973,0.0923493207,0.120153375,0.1430458874,0.231980592,0.0783926174,-0.1750056446,-0.1093824282,0.1624648124,-0.423237592,-0.2539897859,-0.024852762,0.0481915437,0.0019835988,0.0486336201,0.0797554627,0.1869950891,0.3776380718,-0.1662532091,0.1743819267,0.072122328,0.1036484838,-0.0061725359,-0.243062675,0.0630403236,0.2165745795,0.1129323095,-0.0198081173,0.1745865196,-0.2009208351,0.0026038252,0.1835123152,-0.0430871285,0.0688635185,0.0614928827,-0.3432898521,-0.1306156367,-0.2249423712,-0.0999411717,0.0214994084,-0.0673013851,-0.1040214077,0.0291912798,-0.0268023685,-0.0449196845,0.294644326,-0.3951108754,0.1822452545,0.0805596933,0.0715904981,0.2153650373,0.0801527575,-0.4667008519,-0.1507309079,0.2663052082,0.0661004409,0.0761099309,-0.1938565522,0.0803808942,-0.0789445564,0.3015627861,0.6326431632,0.0356377512,-0.1744022369,0.1561800987,0.0443391092,-0.2226135582,-0.0299927127,0.2348788679,0.1107997745,-0.0944953412,0.1726058125,0.1578608304,0.0348748639,-0.3649460971,-0.2931927443,0.3783940673,-0.1450100541,-0.2155480236,0.0737625584,0.1551503092,-0.1362436563,0.2214229554,0.1097290218,-0.0541347228,0.3969765306,0.1784482896,0.2801536024,0.0857571065,0.3843370974,-0.413248986,-0.2857423127,-0.1096890047,0.5274931192,0.0916665122,0.1056273133,-0.1077562496,0.3064687252,-0.0640582889,0.0382292345,-0.1252681017,0.1632192582,0.0184704475,-0.1065423042,0.2033777982,0.0467290357,-0.2128214836,0.2236432433,-0.2771652341,0.0626529157,-0.0550700128,0.3192573488,0.0674424395,-0.490560323,0.1210725158,0.0442174077,0.2686516643,-0.2866969407,0.0787747204,-0.2237661332,0.0632781759,0.1702563614,0.1406056285,0.3410817981,0.3432904184,0.028105706,0.1186328605,-0.2185475826,0.0354694277,-0.2167068273,0.3741964698,0.2538721859,0.2906402647,0.1658261567,-0.0742582306,-0.0946067125,0.1677904427,-0.0720718652,-0.1949005723,-0.2488558888,0.3140272796,0.1776565611,-0.2495557964,-0.530820787,0.1827208996,-0.3358587325,-0.2405874878,0.2575833499,-0.0709406361,-0.0413871296,-0.0808177888,0.0254606847,-0.2634350061,0.6369584203,0.4109855294,0.4801579416,-0.1210205257,-0.2793284357,-0.5621271729,-0.1291811764,-0.2751346529,-0.0435366668,0.0579457469,0.3335485756,0.0830451921,0.1002028063,-0.4502342939,0.2044606656,-0.1037470177,-0.0798092782,-0.3132740557,-0.0057977419,-0.2921862304,0.0728381351,-0.1411017179,-0.2275684327,0.0896720141,-0.2099714577,-0.1017426923,-0.1241573989,0.1359547526,0.1137386411,0.4492316842,0.2314134538,0.0325365514,0.5327629447,0.0228703897,-0.0580756553,-0.4057278931,0.112114802,-0.2275319993,0.2010342479,0.1649188995,-0.0808065981,-0.1234430447,0.3457123637,0.0963352993,0.2587437034,-0.1959899813,-0.0336919054,-0.0136601916,-0.2710326612,0.0340362675,0.3747008145,-0.0256210621,0.5051813126,-0.0297325738,-0.07425607,-0.3696877658,0.0132709024,0.2512350976,-0.2922117412,-0.0986907631,-0.2751006186,0.350518316,0.0055118473,0.3568725884,-0.3684661388,-0.0565424263,0.3812297583,0.1000538468,-0.0715708882,0.571169734,0.0044753863,-0.1593976021,-0.055820547,0.1501427293,-0.0609393045,-0.1890611649,-0.2985860705,-0.212580964]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"@agemagician you can give a try at a multithreaded version if you want (currently on the #548).\r\n\r\nTo test it, you just need to copy the new `text` processing script which is [here](https:\/\/github.com\/huggingface\/nlp\/blob\/07d92a82b7594498ff702f3cca55c074e2052257\/datasets\/text\/text.py) somewhere on your drive and give it's local path instead of `text` to `load_dataset`. E.g. in your example:\r\n```python\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('.\/datasets\/text.py', # path to where you've dowloaded the multi-threaded text loading script\r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":76,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\n@agemagician you can give a try at a multithreaded version if you want (currently on the #548).\r\n\r\nTo test it, you just need to copy the new `text` processing script which is [here](https:\/\/github.com\/huggingface\/nlp\/blob\/07d92a82b7594498ff702f3cca55c074e2052257\/datasets\/text\/text.py) somewhere on your drive and give it's local path instead of `text` to `load_dataset`. E.g. in your example:\r\n```python\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('.\/datasets\/text.py', # path to where you've dowloaded the multi-threaded text loading script\r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```","embeddings":[-0.2351232022,-0.1589976102,-0.1092258841,0.1558210254,-0.0522058196,0.0935270935,0.2101726234,0.3627994359,0.2417389005,-0.1184581593,0.0937155634,0.2659687698,-0.1745844036,0.2608832419,0.2182195783,0.1295302063,-0.0509276241,0.1763873547,0.0324817151,-0.0707732961,0.0224847477,0.0642834753,-0.2663993835,-0.2938196063,-0.1675427556,-0.0311636087,-0.060530778,0.1297042221,-0.0731489435,-0.2301614434,-0.033592727,0.1957894713,0.0279497448,0.2765226066,-0.0001104622,-0.2245029658,0.3785304427,0.0646175668,-0.1945024282,-0.1081684455,0.3321503997,-0.5667905807,0.1606936753,-0.156064555,0.0011674038,0.0118144704,0.0906925872,-0.0626586527,-0.0046143015,0.0926827937,0.1748110205,-0.0288208555,-0.2254574299,0.2356196195,0.0194670707,0.1631803662,-0.0351904556,0.3379810154,0.4009987712,-0.1895385832,-0.300290823,0.1376228929,-0.1286110282,0.1814782917,0.280821979,0.0355102047,-0.1779794246,-0.123710148,0.088163048,0.3048611581,0.4366061985,-0.0966573954,-0.0096868891,-0.4623732269,0.0237911232,-0.2291672081,0.2377023995,0.2097762823,-0.2134128809,-0.3197594881,-0.4350693822,-0.2194339782,-0.050846979,0.1706949174,0.1299187988,0.1563623399,-0.0357263796,0.0971484333,0.3201568425,-0.0055307108,-0.0444112532,-0.070935674,0.2302666754,0.4172829688,-0.4461006224,-0.0729122311,0.1141119599,0.2020266205,0.0926702619,0.0474264696,0.1258238852,0.0924155712,-0.1682846695,0.0449279882,0.2149930745,0.3550407887,-0.2170166969,-0.2151004672,0.3845054209,-0.0908288732,-0.1448483914,0.2591442466,-0.0805850029,-0.1868590266,-0.0264053736,-0.3573792577,-0.1944363117,-0.3192492723,-0.0887169391,-0.1874363124,0.0210011899,-0.2330123484,0.1086945757,0.2244624197,-0.3101556897,0.273878634,0.0386037901,-0.0659090355,-0.3392502069,0.0695986524,-0.1426309943,0.081539318,-0.18821235,0.1823238134,0.6098544002,-0.0081199119,0.1737595499,0.1366225481,-0.2034400851,-0.1409626305,-0.1883977056,-0.2797947526,-0.1239791363,-0.0474042222,0.0543628745,0.1805571914,-0.0143887866,0.2696852386,-0.3400957882,0.1253848523,-0.3222605586,-0.1604731381,0.0425276905,0.1878402382,-0.40412274,-0.4088690877,-0.2737722993,0.4251868427,-0.0619101822,-0.1103430912,-0.258369267,-0.0493566953,-0.3161037266,0.0233812016,0.1262766868,0.2047499269,-0.049246192,-0.0198602993,-0.1246586367,0.3684651852,0.2619371414,0.5896151662,-0.1822747439,0.2504915297,-0.1912335902,0.0575083308,0.4228338301,0.0254763179,-0.3447416127,0.4530771971,-0.2493126094,0.0352213383,0.1712365597,0.335291326,0.0411991738,-0.0678871721,0.2889809906,0.6735248566,0.1797504127,0.2687721252,-0.4803863168,-0.1913736761,0.2381407171,0.4988001585,-0.100481905,-0.1164523736,-0.0320142917,0.0593519621,0.4168198407,0.1106979772,-0.0889548659,0.3573492467,-0.2799513638,0.0601352081,0.0224748477,0.0664092824,-0.1698401123,0.2152332067,0.1656542122,0.1306633204,0.3433740437,0.1143689454,-0.1496990025,-0.3231546879,0.0520347878,0.1409707814,0.0762310252,0.0999655798,0.060535837,0.2033236474,-0.1631735861,0.3446237147,-0.149008736,-0.2035680562,-0.3692210615,-0.0732146874,0.1707623452,-0.1606675237,0.3976022005,0.3120849133,-0.1375889331,-0.0085790148,-0.124214448,0.014577386,-0.317317903,0.3243600726,0.0493608341,0.1176916808,0.0104946699,0.0003519012,0.3571496904,0.1619955301,0.233458221,-0.3475341797,-0.3124429584,0.3943000138,-0.2278010994,0.2335012257,0.1397653371,-0.4423324466,0.0909141377,-0.1508065313,0.1609528214,0.2870859206,0.864359796,0.10864795,0.4909005761,0.2559530437,-0.2232835889,0.0353816114,0.5316964984,-0.0115907546,-0.1453190893,0.4955512881,-0.1591791064,-0.3933447301,-0.076403372,-0.0008987524,0.3892951012,0.3365332484,0.0801728591,0.0598104633,-0.1004606485,-0.2265397459,0.21119757,-0.1728012115,0.2205049843,0.1607719362,0.3817356825,0.07273601,-0.5193353891,-0.2708147168,-0.0174346864,0.3317105174,-0.0837924108,-0.0142663177,-0.0677763447,-0.332364887,-0.2030872852,0.0228839312,-0.3021137416,-0.2011318803,0.0522101261,0.1582282186,0.3566643298,-0.1291055828,0.03498872,0.1396400332,-0.2011191398,-0.4455116689,-0.2782613337,-0.1495083421,-0.3876416087,0.0376855396,0.2981480062,0.2329722196,0.3325830102,0.1779067218,-0.2270571589,0.2563956976,-0.0381544344,-0.2229510248,-0.1001670808,-0.027787013,-0.1903454363,0.0827777013,0.1761137694,-0.1057824567,0.0916429833,-0.3298029006,-0.0209475365,0.1437876225,0.048124779,0.0051461952,-0.0873239487,-0.2857705951,-0.2702842951,-0.3121495545,0.5135660768,0.0740733221,0.1946208477,0.2714239359,-0.2012180984,0.2871980965,0.0897094682,0.2429691106,-0.0052619549,-0.1578310728,0.3171799481,0.2386637479,-0.2387669981,-0.2388914376,0.0595538206,0.0963346809,-0.0034868771,-0.6804850101,0.2912909985,-0.4513459206,-0.0845335871,-0.0776610747,0.1385241151,0.2351496965,-0.1911541522,-0.1271463782,0.2458197623,-0.1807982773,0.0519634373,0.0566413142,0.178040117,-0.1369871944,0.3521686196,-0.0042590508,0.026034398,0.0612689145,-0.1209980249,0.2219278365,-0.038306646,0.0854202658,-0.1794352978,-0.2658483684,0.2797511518,-0.1560489535,-0.1178876162,0.3768692613,0.0356276184,-0.4968029559,-0.2630113065,-0.1910326332,-0.080294542,-0.0930124521,0.2119566202,-0.3439237475,-0.1305854023,-0.0440253764,-0.1847305298,-0.1962886006,-0.7099232674,-0.1090595648,0.1553484797,0.0283267237,0.0216540489,-0.1448139399,0.0565858148,-0.7600155473,0.092953749,-0.1794436425,0.4511081874,0.0435856357,-0.101014033,0.3006531596,-0.2252385914,0.3230228424,0.2441086918,0.2112258226,0.2297372371,-0.0903326645,-0.3802280724,-0.0638860539,-0.2193334997,0.1761723459,0.5705080032,0.3222826123,-0.1027431488,-0.106600143,0.0648815185,0.2419088483,-0.0228410494,-0.5537737608,-0.3349891901,-0.199630633,0.0151826218,0.1195352897,-0.2522740066,0.1908494234,-0.1686496139,0.00703615,0.0828008503,-0.1175255552,0.2057752013,0.0423089489,-0.21212475,0.132438302,0.3194184899,-0.0765133128,0.1944128722,-0.4677106142,0.3661045134,0.2831432521,0.2096608281,-0.011277712,0.0569320619,0.3561211526,0.2310208976,0.1218743995,0.0553161316,-0.2382798046,0.2259386629,-0.2462039441,0.0201344751,0.1647601128,-0.0066913078,-0.3034945428,-0.3959756792,0.8766254187,0.175106734,0.0521594547,0.4086862504,-0.454288125,-0.0554090068,0.185663566,-0.2767317593,0.739502728,-0.4225721359,0.2307497561,-0.1247346401,0.0336233936,0.3825359643,-0.2189726979,0.0912722945,-0.3892379701,0.0464172103,0.3155104816,0.0045717214,-0.1580930054,0.2595574558,-0.2366605848,0.4072613716,0.4061505795,-0.3258392513,-0.1774658263,0.774253428,-0.173397705,-0.2909516394,-0.5964768529,0.1502722949,-0.3037243485,0.425758183,0.0358881168,0.0840641558,0.0225063395,-0.0011621045,-0.3838971555,0.1141907573,-0.1107715443,-0.0717734098,-0.1724258661,-0.2370329946,0.1862777472,0.1481084675,-0.2201445699,0.2130778581,-0.2657666802,0.1285834461,-0.3506118357,0.2198165804,0.5145677328,-0.0815347731,0.1458852887,-0.2349336147,-0.0623743422,-0.2554595768,0.0540961288,-0.2006596923,-0.0338574499,0.1091721058,-0.0494350307,0.1636491418,-0.2356083989,0.2690001726,0.026838731,-0.0462337248,0.1172098964,0.3300510645,0.0246529877,0.3350321054,-0.0102442428,-0.4108898342,-0.0247596987,0.5550806522,0.0905227512,-0.1743538976,0.3781972826,0.2993063331,-0.102679044,-0.193521589,-0.0058145109,-0.1136119813,-0.1300796419,0.0919476375,0.1440708637,0.001486711,0.2302419543,0.0934481993,-0.0601610318,-0.0756294504,0.0322803706,-0.449411869,-0.3595046997,-0.0184067115,-0.0575397946,-0.0257059652,0.0484162755,0.0251541641,0.318577677,0.2946865857,-0.2691206038,0.1614511311,-0.0746617392,0.0227461495,0.0471829139,-0.3493124843,0.1104752421,0.2076112032,0.1626275033,-0.1240937859,0.1149851233,-0.2853878438,0.0394423679,0.1062259004,-0.0581014715,0.0430499241,-0.0194173679,-0.325753212,-0.0764435455,-0.1878139675,-0.052564688,-0.0375624411,-0.0514752679,0.0023722257,-0.0455615222,-0.0308670811,-0.0019326862,0.1513373256,-0.4418312311,0.1661905199,0.1071057618,0.0150326351,0.3286926448,0.0265068263,-0.5275723338,-0.0933599919,0.0859325975,0.1513931602,0.0765262544,-0.1655911505,0.0297098048,-0.0720439926,0.2578231394,0.4545097649,0.0649751499,-0.2551617324,0.1589473337,0.1518911868,-0.2063771635,0.0199879725,0.234076336,0.0335199125,0.0116149606,0.1474227309,0.1710616648,0.0347312875,-0.2572361231,-0.2264292985,0.3829537928,-0.2253789902,-0.2730332613,-0.002302998,0.1371928304,-0.1771082729,0.2033013105,0.0657119453,0.0528958961,0.4379095137,0.2583200932,0.2723615766,0.1238265559,0.4825105071,-0.3734045923,-0.1923393607,-0.1380871981,0.3719217479,0.0837011859,0.1760033816,-0.0947280526,0.3289039135,-0.1408207268,0.0054064882,-0.2425810993,0.1206111163,-0.050064981,-0.123220779,0.167399317,0.0422044657,-0.2815217972,0.2524554133,-0.2612040937,0.1297660321,0.0048753424,0.3172589242,-0.0420136526,-0.3678248525,0.03653422,0.0090389904,0.2566407025,-0.2119986266,0.211887762,-0.1680599004,0.0850642845,0.038012635,0.2695977092,0.3048685491,0.3242583871,-0.0846756995,0.0877341777,-0.1773621291,-0.0074821846,-0.2701363266,0.4044509232,0.1985896975,0.3803492785,0.1028906032,0.0262566172,-0.1636014581,0.1692243218,-0.1378523111,-0.2667510509,-0.1030694842,0.2986758053,0.2648009062,-0.2539790571,-0.5705643296,0.149351716,-0.341912359,-0.2652081251,0.1564534903,-0.0367871746,-0.025259139,-0.1077066809,0.0758045018,-0.2157666385,0.7528712749,0.3798870444,0.3102044463,-0.0567047819,-0.2296943516,-0.5407270193,-0.129448846,-0.2332152575,-0.1430045813,0.0734249428,0.3392794728,0.0097257895,0.0686180443,-0.3661653697,0.242590934,-0.0832142159,-0.0967292562,-0.3036373556,0.1143033504,-0.2806533575,0.0782235637,-0.1154286936,-0.2335688323,0.0654680729,-0.2150258124,-0.0165010039,-0.1261469871,0.1125936583,0.0025023872,0.3830724061,0.2040139735,-0.0181007404,0.3712948561,-0.0314275548,-0.1291222572,-0.4532833695,0.1737650931,-0.2454369217,0.1342475712,0.146778062,0.0351161249,-0.0224410743,0.2664222419,0.0916809514,0.2971769273,-0.2757699192,0.0320414305,0.0875488222,-0.2770710289,0.0480133779,0.3402582705,0.033714857,0.5124368072,-0.0249810722,-0.0983120948,-0.5657388568,-0.0551331863,0.2666742206,-0.2849320173,-0.1079978123,-0.2226086706,0.360327661,0.031440299,0.3329227567,-0.4231390953,-0.0123163741,0.2397252023,0.1389221102,-0.0445491895,0.5026028156,-0.010092264,-0.1497851014,0.0342618078,0.0675461888,-0.0645567402,-0.0972398296,-0.3322935104,-0.1884772182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"I have already generated the dataset, but now I tried to reload it and it is still very slow.\r\n\r\nI also have installed your commit and it is slow, even after the dataset was already generated.\r\n`pip install git+https:\/\/github.com\/huggingface\/nlp.git@07d92a82b7594498ff702f3cca55c074e2052257`\r\n\r\nIt uses only a single thread.\r\n\r\nDid I miss something ?","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":50,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nI have already generated the dataset, but now I tried to reload it and it is still very slow.\r\n\r\nI also have installed your commit and it is slow, even after the dataset was already generated.\r\n`pip install git+https:\/\/github.com\/huggingface\/nlp.git@07d92a82b7594498ff702f3cca55c074e2052257`\r\n\r\nIt uses only a single thread.\r\n\r\nDid I miss something ?","embeddings":[-0.199171856,-0.1527013481,-0.0914235115,0.1850308776,-0.0580269136,0.0069911219,0.0542506613,0.3648399115,0.226775974,-0.1864666641,0.1637936831,0.271445334,-0.0985700637,0.165866375,0.1985837221,0.1851770878,-0.0191418119,0.2842988074,0.1013761908,-0.188750878,0.0956259891,0.0487127341,-0.184758231,-0.2295959294,-0.1763940454,0.0025341206,-0.027706068,0.1009524688,-0.1621125638,-0.2173818052,0.0553882159,0.1236786172,-0.0306907948,0.3531843424,-0.0001217054,-0.2242824882,0.3527976871,0.1074252874,-0.1829125136,-0.0307793859,0.2109853476,-0.6483900547,0.1963076144,-0.1187645718,-0.0270373039,0.0886713043,0.086174272,-0.0851289555,0.0995737836,0.1071528643,0.0966105983,-0.1017713249,-0.2455634326,0.2750653625,0.018805882,0.3272772431,-0.0368346199,0.3568100035,0.3645662665,-0.1647836119,-0.1937814802,0.054430034,-0.1813870668,0.1767542511,0.3036622405,0.0144899813,-0.0835126117,-0.1341192871,0.1963970661,0.317099452,0.5264444351,-0.118086569,-0.0595808811,-0.4169434011,0.0415285006,-0.172948122,0.144746393,0.260304451,-0.2250424176,-0.2841115594,-0.4277854264,-0.41022259,-0.0714328289,0.2077246755,0.1785079688,0.0898235813,-0.0192157719,0.1839961261,0.2658657432,0.1081896275,-0.1406799108,-0.0842886716,0.1884384006,0.4555005133,-0.3959882855,-0.0676118731,0.148156926,0.1502484828,0.1318481565,0.082677573,-0.0320088454,0.0502000451,-0.1457344443,-0.0421881117,0.2815548778,0.388105005,-0.2023485452,-0.1266419739,0.4440859854,-0.2445262969,-0.1741784364,0.2413170338,-0.1810796708,-0.1101007089,0.0550281592,-0.3759531975,-0.1100725979,-0.3834215105,-0.1334659457,-0.0778839961,0.0038247406,-0.298048526,0.0217744652,0.222545132,-0.3536616862,0.2660466731,0.038554579,-0.1337033659,-0.4125122726,0.0827439129,-0.1261990517,0.0016385531,-0.1777823716,0.2467430234,0.5104455948,-0.1568821967,0.1790259182,0.1848551035,-0.2046812028,-0.2262265831,-0.2678073347,-0.2743785679,-0.2211675644,0.0030917912,-0.0159184523,0.2825623453,0.0864949673,0.2100139707,-0.3664640188,0.1766580045,-0.4176154733,-0.1564798653,-0.0340588056,0.0695470273,-0.4505048394,-0.297231406,-0.2967454493,0.3501704037,-0.021043634,-0.0672935098,-0.2913905978,-0.0671825707,-0.2015482783,-0.0603885539,0.1400727779,0.2322684675,-0.0928435847,-0.0590415448,-0.1035439819,0.2933353186,0.3807816505,0.6156035662,-0.1922990084,0.3422890902,-0.1217292547,-0.0452178679,0.5736959577,-0.0108839273,-0.4533556998,0.5298498869,-0.2530936897,0.0212491974,0.1293158382,0.3815018237,0.0154520711,-0.0342910774,0.2500491738,0.5158036351,0.1687353998,0.2006347179,-0.4838428795,-0.1462139189,0.2273876816,0.3783594668,-0.103270568,0.015090947,0.0358741879,0.0942553431,0.4295170307,0.1692921966,-0.1244556606,0.3971879184,-0.1047697663,0.0351576433,0.0449835397,0.1743526608,-0.1608433127,0.1670681685,0.1669221222,0.1247466952,0.4018214047,0.1500276774,-0.2643034756,-0.1956247538,0.0506964028,0.1468177736,-0.0498642996,0.0479860418,0.0912837088,0.179148227,-0.185681209,0.3405303955,-0.1098366678,-0.216119349,-0.3428982496,-0.0301032979,0.2130297869,-0.1742124557,0.4311957359,0.3232764304,-0.1331785768,-0.0647700354,-0.0922236368,-0.0369056426,-0.2966367006,0.248290807,0.1047014147,0.143235147,-0.003089936,-0.0303133354,0.3299259543,0.023629114,0.2452347577,-0.4048092365,-0.1249653175,0.3379196227,-0.2271563709,0.2508988082,0.0562171228,-0.5008863807,0.1578982323,-0.1253249943,0.1313193738,0.3620710969,0.9713236094,0.0677562281,0.5750047565,0.2706082761,-0.1945349574,0.0007566935,0.5275571942,-0.0308702756,-0.1748841703,0.5408584476,-0.1377526224,-0.3145546317,-0.0478421077,-0.1603901237,0.4037880898,0.2320174873,-0.0110661304,0.0500108562,-0.0768371075,-0.2358856797,0.1037636623,-0.122410059,0.3007444739,0.1260949522,0.3335475028,0.0369228572,-0.4046323597,-0.158119455,-0.0177756529,0.5295583606,-0.0765737295,0.0730884224,-0.024825152,-0.3180570304,-0.218727693,-0.0239542779,-0.2930161059,-0.1844211668,-0.0439070277,0.0747429952,0.5305743217,-0.1234557554,0.1065886468,0.1333356947,-0.2151924074,-0.4275615513,-0.3017472029,-0.1941060722,-0.3608124852,-0.0455354378,0.3237594068,0.2041276395,0.3716340661,0.1838134825,-0.2097916156,0.1934048831,-0.078715533,-0.2579253018,-0.1157392934,0.0656484738,-0.2042783201,0.254460752,0.1804481596,0.0209541935,0.1526819021,-0.3680243492,-0.0320829935,0.1338160634,-0.1131739467,-0.0788569599,-0.0251730029,-0.1890720576,-0.2600283623,-0.1960692853,0.5295091867,0.1719647348,0.202089414,0.214591831,-0.1746442616,0.2627972364,0.1632849723,0.1996219307,-0.0905221254,-0.2076040655,0.2065794617,0.3381420672,-0.2271890342,-0.1761685014,0.1211566851,0.1823153347,-0.0823850557,-0.7940250635,0.2953012586,-0.4839649498,-0.0214370377,-0.0950324461,0.1529513299,0.2004410326,-0.2212822735,-0.0338631682,0.2205282003,-0.2697232366,-0.0121434135,0.0616766922,0.1756334752,-0.162468031,0.4543955922,-0.0773452744,0.1778048575,0.2052347809,-0.1453509629,0.1505452693,-0.0238698591,0.0915137529,-0.1570835263,-0.2992313504,0.260060966,-0.1520539969,-0.1015936211,0.3296049237,0.0612494908,-0.4600823224,-0.2959480286,-0.1237133369,-0.0436814912,-0.043080993,0.130120188,-0.3121578395,-0.1345962584,0.0089058485,-0.1670843214,-0.2617808282,-0.7116495371,-0.1529487967,0.1397790164,0.098907195,0.0223071147,-0.2154083699,-0.0797727853,-0.8438497186,0.0574550368,-0.2956528366,0.4102829993,0.0209305584,0.0098647038,0.463324517,-0.2124309987,0.3210779428,0.3303005397,0.0987618789,0.1878583282,-0.1014783978,-0.4424764812,-0.0420524217,-0.1406910121,0.3415518403,0.6328332424,0.2685382366,-0.2692616284,-0.1277536899,0.0687531158,0.2400995195,-0.1106044278,-0.5722797513,-0.4723233283,-0.2107154131,-0.0132700028,0.1024186462,-0.3304749727,0.1581961513,-0.1131990775,0.0735707805,0.0565336086,-0.0732529834,0.2725447416,0.0096538505,-0.2714055479,0.1710288227,0.2738582492,-0.0345490463,0.169529736,-0.4581918418,0.4365380704,0.2314756513,0.2155783474,0.000707105,0.0947542563,0.4200230539,0.2714297473,0.1474982798,0.0664408281,-0.18561019,0.2110603303,-0.3316260278,0.0767130032,0.175939858,-0.0090968907,-0.4444129169,-0.4641419053,0.8058819175,0.1324350536,0.0663396865,0.4682039022,-0.3776952922,0.0978343114,0.0659835115,-0.2738637924,0.8955380321,-0.4456989467,0.1810887903,-0.0647239834,0.0101055726,0.4141217172,-0.14846389,-0.0020365668,-0.4385489821,0.0917849243,0.2670578957,-0.0993265957,-0.0731003061,0.1387941688,-0.1598206311,0.4082352519,0.3856756985,-0.3247514367,-0.120925501,0.8059942126,-0.000917161,-0.1387105137,-0.511818409,0.0789531991,-0.3942382932,0.5236879587,-0.0254260171,0.078804642,-0.1073717251,0.0096574361,-0.3741385639,0.1029963195,-0.0440151207,-0.0804702714,-0.0706518814,-0.3025927842,0.1593977958,0.048073709,-0.1583145559,0.1963740438,-0.2845977247,0.1651973426,-0.3597008586,0.2295217365,0.5117166042,0.0027045181,0.0890644118,-0.2288412005,-0.083658956,-0.1240887046,-0.0090777865,-0.310236752,-0.0539712645,0.165418148,0.0732497871,0.1553654969,-0.2900008857,0.3394241631,0.0487428941,-0.0190633237,0.0211600456,0.4879475832,0.0691234395,0.3877803385,0.0445620753,-0.4567271769,-0.0220169649,0.4734851718,0.19885616,-0.1774843782,0.4827728868,0.2412371337,-0.130726397,-0.1122941673,0.0090100402,-0.1025874317,-0.2773761749,-0.0096223205,0.1534974873,0.0580185801,0.2951738834,0.1091601923,-0.1204330027,-0.061847683,0.0590344071,-0.3825007677,-0.3001973033,-0.0500293002,-0.0319829695,-0.0089847334,0.0168450139,0.0097812144,0.2448109388,0.4505028725,-0.1600066423,0.1078561321,-0.0533717275,0.0784185901,0.1320558339,-0.2580537498,0.0839033201,0.1609606594,0.0528897792,-0.1663526446,0.0882008448,-0.183482796,0.1023485214,0.1997906864,-0.0797450319,0.0218733475,0.0000607917,-0.2671490312,-0.0195674933,-0.1199635118,-0.1016683355,-0.0600429252,-0.0506259538,-0.0036463942,0.0608458966,-0.0245200135,0.046186246,0.197410062,-0.421626538,0.2693496943,0.1194001436,0.0213582572,0.4252481163,0.0188261811,-0.6070492864,-0.123491846,0.208340168,0.1969070733,0.1300879866,-0.2446459532,0.1431639344,-0.0218960829,0.2003124505,0.4992823005,0.070760563,-0.2289083898,0.1936346292,0.0427840836,-0.1854317188,-0.0000548364,0.3791795373,0.1669329405,-0.1386950165,0.1918570548,0.1223263517,0.0853966549,-0.3114671111,-0.2537027001,0.4503023326,-0.2387213111,-0.2689567506,-0.014466228,0.1872873306,-0.1875882149,0.1477086395,0.0865142569,-0.0531236827,0.3869537413,0.1852475107,0.2312433273,0.1526881903,0.4042868018,-0.3817747831,-0.1503842026,-0.1685385555,0.4807797968,0.0616285093,0.1368676871,-0.0440534316,0.2922136486,-0.0140513796,-0.0222861953,-0.2037829459,0.1579188257,-0.0241177585,-0.0707793981,0.1406330168,0.12906982,-0.1040128022,0.2433003634,-0.2980495393,0.1007594094,-0.0613966137,0.3115348816,-0.0987393782,-0.4200946391,0.0597672909,-0.0506384857,0.2242970318,-0.1350816637,0.2209181637,-0.2688324451,0.0531478822,-0.0043940619,0.2688487172,0.2770287991,0.423879385,0.0000271925,0.1043703556,-0.0800285712,-0.0206314493,-0.185152784,0.4359554946,0.1467037797,0.3372706175,-0.0445446633,-0.0802007988,-0.0651489124,0.131826669,-0.146918878,-0.2350205183,-0.1072516069,0.3373669982,0.1562323868,-0.3151374459,-0.5273731351,0.1796395332,-0.1902775317,-0.1569943428,0.0874983072,-0.0946276039,-0.0411558524,-0.1593942493,0.0105488263,-0.3326226473,0.712417841,0.3719665706,0.4264705777,-0.1453033686,-0.2639220953,-0.5644286275,-0.0651560277,-0.348398298,-0.1616449505,0.0900787264,0.3021501303,0.0062262765,0.0611698776,-0.3685857654,0.0544767641,-0.0754302815,-0.0999560431,-0.2846449912,0.0190533735,-0.3252052665,0.0777173862,-0.1635840088,-0.1889843792,0.1147741899,-0.2919634879,-0.125784725,-0.1764401644,0.0499908812,0.0072851889,0.5267725587,0.2606075406,0.0163785685,0.3779647052,-0.0012552078,-0.1727335751,-0.3434295356,0.0821246207,-0.1888545901,0.2389385104,0.0655279383,0.0062899254,-0.1153211445,0.2998861969,0.0443255976,0.2522705793,-0.2702216506,0.0204271376,0.0721706524,-0.2926415801,0.0333512239,0.3275591135,0.0152923996,0.3754202724,-0.011364013,-0.0416368693,-0.5524223447,0.134272784,0.3410869539,-0.270577997,-0.0809869617,-0.1680973172,0.3409724832,0.0285032876,0.2374149114,-0.3947706223,-0.0690630898,0.3056742847,0.1608585715,-0.0737948343,0.5093422532,-0.038003087,-0.1087642759,0.0135922441,0.1959216297,-0.1505623907,-0.1879007071,-0.2074455023,-0.2052045912]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"As mentioned in #548 , each time you call `load_dataset` with `data_files=`, they are hashed to get the cache directory name. Hashing can be too slow with 1TB of data. I feel like we should have a faster way of getting a hash that identifies the input data files","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":49,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nAs mentioned in #548 , each time you call `load_dataset` with `data_files=`, they are hashed to get the cache directory name. Hashing can be too slow with 1TB of data. I feel like we should have a faster way of getting a hash that identifies the input data files","embeddings":[-0.1172169596,-0.0360578671,-0.0959250852,0.3137763441,0.0166612789,0.1080764607,0.1881460994,0.562651217,0.3573107123,-0.0834074765,0.0841233954,0.0788894296,-0.2751655877,0.1463639289,0.3032574654,0.1522518843,0.040683534,0.2266365737,0.0538096651,-0.1176902652,0.0696330369,-0.0204687137,-0.029910475,-0.2968647778,-0.301710397,0.030994568,-0.0536712706,0.0801734254,-0.092462942,-0.2358043343,-0.0621912293,0.1432615668,0.0518749729,0.2695997059,-0.0001230621,-0.1986707598,0.3916837275,-0.0107746581,-0.2250648588,-0.1616011262,0.0600100383,-0.7300632,0.1203292385,-0.1630459428,0.0337214544,-0.0094950749,0.2076354325,-0.2160127759,-0.2010137886,-0.0743039697,0.062980175,-0.2378575504,-0.3024275601,0.4499882758,0.315975219,0.1362914592,-0.0163585339,0.3796997368,0.2599540055,-0.2834985554,-0.344534874,0.147704199,-0.0814258009,0.0042502461,0.3230665922,0.2134281099,-0.1169731915,0.0059481687,-0.0009590726,0.3054632545,0.5086736083,-0.2065645456,-0.2373914272,-0.5030111074,-0.0787002593,-0.1927808821,0.2358420938,0.1443256587,-0.2236379087,-0.2345628887,-0.4230131805,-0.3044024408,-0.0414100289,0.1779715419,0.2704318464,-0.1131949946,0.1272668839,0.0756366104,0.3568819165,0.0150676705,0.1043341458,-0.1023497954,0.3182216585,0.3189169765,-0.355796814,-0.0773666799,0.079976894,0.3400668204,0.0997809544,0.2149660289,0.2113414407,0.1314075887,-0.1452826113,0.0403852463,0.1335639358,0.4111125469,-0.1875095516,-0.2380570322,0.3621019721,-0.3224426806,-0.3378063142,0.2185865343,-0.1644865125,-0.2672842145,0.2383487076,-0.2126469165,-0.3028368056,-0.293504715,0.0493218601,-0.0200048089,-0.090031594,-0.2638342083,0.2086578757,0.2353861481,-0.2940382957,0.2273739129,-0.0614142865,-0.1680902541,-0.3403705657,0.1153089926,-0.150528729,0.1385418922,-0.1539921314,0.1784461588,0.6293757558,-0.1461412013,0.1557849646,-0.1494501233,-0.036918398,-0.0314942226,-0.1652005762,-0.2768216133,0.0819735304,-0.047023762,-0.1023334786,0.1132050082,0.0330055654,0.2430368811,-0.4355452061,0.0362548828,-0.3797471523,-0.1476532817,0.2258044928,0.0793962032,-0.4669116139,-0.2244969308,-0.2262626588,0.3716507554,0.0188341029,-0.2251745611,-0.3301511705,0.0527155064,-0.2251861691,-0.0896676183,0.0870079026,0.2484913468,-0.0564145036,-0.0262250379,0.0206191614,0.2286192775,0.191990912,0.4548192322,-0.1003708243,0.3429888189,-0.269059062,0.0256046727,0.5262586474,-0.0462650843,-0.5724012256,0.4323171675,-0.2308463454,0.0318439975,0.3767980933,0.5546629429,0.0004532505,-0.1959045082,0.3050215542,0.5151875615,0.0486500151,0.265817821,-0.4664846957,-0.1919652075,0.3387056589,0.3695990443,-0.0656158701,-0.0110721439,0.1641367078,0.1056739762,0.3326959312,0.187718153,-0.0577580594,0.3255420327,-0.1506641656,-0.0861674026,0.0373497158,0.1127424836,-0.2608890533,0.1496976167,0.1009583399,0.0266809575,0.0260290429,0.092009522,-0.0673766434,-0.2983823419,0.1009331718,0.3866150379,-0.0716902763,0.0765951723,0.1257007867,0.0621488467,-0.1450602114,0.4783859253,-0.0661613941,-0.1453858018,-0.3675588965,-0.0822641477,0.2519709468,-0.1136748567,0.2339874357,0.3054568172,-0.1155244857,-0.1129526123,-0.0965614468,-0.0622179583,-0.193963334,0.1687045395,0.1451293677,0.3294298351,0.046190735,0.0611431971,0.3061034083,0.1154011935,0.2048128992,-0.4098871052,-0.2786921263,0.4903713167,-0.2560999095,0.1333206296,0.1243943721,-0.4574455619,0.0698688775,-0.1131155118,0.1526701748,0.1416293085,0.7772764564,-0.0267719477,0.599798739,0.3419163525,0.0416853651,-0.0370966978,0.6666430235,-0.0325513631,-0.2055317163,0.3865835667,-0.0312755629,-0.4225259423,-0.091039516,0.0016564556,0.469394803,0.2598968446,0.1467089504,0.0442964621,-0.0494520031,-0.2745335698,0.1561003178,-0.1106311679,0.0547729172,0.1620090008,0.3263850212,-0.0729023665,-0.4668158293,-0.2270137221,0.1389245689,0.3100939989,-0.1783640087,0.0604971647,-0.0893567204,-0.310862571,-0.1810174733,-0.0770312026,-0.2673085928,-0.2871685326,0.1371500492,0.1534987539,0.3889503777,-0.0441602916,-0.1306312084,0.2002463341,-0.1349127442,-0.6252069473,-0.2875651419,-0.1997583061,-0.2606434524,-0.0950763673,0.3254851699,0.0230568014,0.2869828045,0.1275070757,-0.0611164086,0.0269521158,-0.1847222,-0.1780194044,-0.0560671762,0.0458907075,-0.1653107405,0.1088286489,0.1011015996,-0.0643062964,-0.1143616661,-0.2909546793,-0.1170977354,0.1513289958,0.0563451201,0.0040036067,-0.0683834478,-0.2516080141,-0.3148128092,-0.1504773945,0.3833199441,0.1087944806,0.2968517244,0.2592844367,-0.234059155,0.1469753683,-0.0350159854,0.2759186924,-0.0761280879,-0.5129765272,0.3502998054,0.29409042,-0.115003109,-0.2911066413,-0.0380456932,0.0169119053,0.3138881624,-0.6159600019,0.1260202378,-0.269446224,0.0532075167,0.0712784454,0.126404807,0.1990044117,-0.1272869408,0.0033351153,0.051857505,-0.0789894611,0.100439705,-0.0399759114,0.4138163328,0.0149863241,0.3838113844,0.1389642209,0.3002887368,-0.0128754489,-0.2340246737,0.3244723678,-0.072200194,0.0982686579,-0.1152264848,-0.2440123111,0.1621457487,-0.0477000438,-0.0663540214,0.3179939687,0.0898097083,-0.3988576233,-0.3091432452,-0.0631492287,-0.1957660317,-0.1362979859,0.1651681215,-0.6871646643,-0.1743243188,0.0545405187,-0.0697959289,-0.3932175338,-0.7218485475,-0.0207654461,0.160268113,0.1696252972,0.0409314148,-0.1226970032,0.029551696,-0.6921350956,0.1490304619,-0.0126851778,0.4575088024,0.0520592667,-0.2797318995,0.2788511217,-0.0648148581,0.5138838887,0.1857994795,0.2223902494,-0.0494035706,-0.1781798303,-0.1109013036,0.0213453807,-0.0597058088,0.378093183,0.3873068392,0.3428022563,-0.3467059731,-0.0881019756,-0.0436140858,0.2210691124,-0.0473368205,-0.4539403021,-0.2696037889,-0.0743694827,0.1142491698,0.1016498879,-0.2978039384,0.1837015599,-0.0865080804,0.1280921996,-0.1037397534,-0.0666290894,0.4675747156,0.0323397182,-0.1322664768,0.0966264382,0.3669077158,-0.0064787832,0.3931185007,-0.4296896756,0.3949216604,0.0239030272,0.0105793625,-0.0990382731,0.1186465025,0.4053674042,0.3409672976,0.0405170918,0.035396006,-0.181134522,0.031817615,-0.2229655087,0.1625323743,-0.0378648378,-0.0219848473,-0.5102128983,-0.5090435743,0.83614856,0.2132326216,-0.0450975224,0.4602642357,-0.4012110829,-0.1145940274,0.2288123667,-0.409059763,0.9492385983,-0.5258201957,0.2740208209,-0.1382442266,0.1987670064,0.3029145002,-0.2896644473,0.1577135473,-0.4653742313,-0.2272246033,0.2176132947,-0.0880728811,-0.1694068462,0.4042203128,-0.0187939238,0.5223017335,0.3355797827,-0.1276986152,-0.1191428602,0.7030899525,-0.0966557711,-0.2274332643,-0.5007805228,0.0651399866,-0.2991484106,0.434127748,0.0483541749,0.0430695079,-0.1179756895,-0.0800280198,-0.1191719696,0.0320823044,-0.0518309176,0.1584090441,-0.1272395253,-0.2793522775,-0.2007530034,0.0972549543,-0.1604655683,0.1582180858,-0.2683205307,0.1277573556,-0.1909477562,0.3028258085,0.4667535722,-0.0477823094,0.1865968853,-0.0292198788,0.0592971109,-0.2993100882,-0.0307529867,-0.5082600117,-0.1132696792,0.1326718479,-0.0908717737,0.2093599439,-0.3355440199,0.0894006491,-0.116632551,-0.00953128,0.0105146756,0.4713151455,0.0110353194,0.394682616,0.1198588312,-0.45771873,0.0139833372,0.6548168063,0.103077352,-0.3002205193,0.367087692,-0.0034525909,-0.2541691363,-0.0705163702,0.0114987716,-0.1161025539,-0.0931541696,0.0365684964,0.109834075,0.045358602,0.1772182882,-0.1044460759,-0.0608146079,-0.0454838313,0.0847871676,-0.4854914248,-0.3651283979,0.1256477386,-0.1291414648,0.111748904,0.0809478238,-0.1771707386,0.0301037673,0.3462210298,-0.1482207328,0.1582172215,-0.0119688567,0.1362667382,0.0018274705,-0.2031135261,0.1816446483,0.2041499913,0.0785746425,-0.1187301651,0.0751340315,-0.1347316802,0.0523106493,0.2363779694,-0.0419003814,0.1236183122,-0.1118106321,-0.2583374381,-0.0259358734,-0.2587968111,-0.0806114376,0.1525801718,0.0079997424,0.0056520109,0.1232375726,-0.11546579,0.0744762272,0.1694041491,-0.4375497997,0.2435422838,0.0464208461,0.0212838631,0.1374116689,0.0515323058,-0.4352950454,-0.0022803459,0.2835547328,0.0239242632,0.1941166967,-0.0768841505,0.0947942734,0.1372132301,0.2040147334,0.3987971246,0.0571495071,-0.2171076536,0.2979225516,-0.0044385539,-0.1637790799,-0.1087215543,0.3308369517,0.0532171689,-0.0702323839,0.2231256217,0.1704673767,0.093730852,-0.3291569948,-0.3950857222,0.4794484973,-0.2557059228,-0.2780338824,0.1011341885,0.1931038201,-0.1679010987,0.3428748548,0.020051742,-0.0306278914,0.5752463937,0.0383024141,0.3500930965,0.0070865392,0.416941762,-0.5145714283,-0.2326480299,-0.1306800842,0.5559139252,0.0862722248,0.1915236861,-0.0950585082,0.3922818005,0.0793460682,0.2054959387,-0.1484098881,0.0668458045,0.045548901,-0.1972443312,0.25827232,0.0165173393,-0.2255829126,0.3079741299,-0.279541254,0.0819223225,0.0677575842,0.4223356843,0.1127082184,-0.2670770884,0.1762161106,0.0826564208,0.2910770774,-0.2383849025,0.1545838714,-0.3150224388,0.1111706942,0.2030689716,-0.0098552508,0.1738347709,0.3405084312,-0.077023223,-0.0338248722,-0.1383889169,-0.0386209339,-0.319622159,0.5069381595,0.1659612954,0.3182131946,0.0938626602,-0.0991064757,-0.0015627245,0.2202218175,0.0395093895,-0.1838704199,-0.0821900666,0.3029431701,0.1810439229,-0.2337032408,-0.5124337077,0.1375777423,-0.2991976142,-0.4470985532,0.2444852889,0.0163898505,0.150172621,-0.0656899288,0.0049451329,-0.2424595505,0.6127952337,0.4382681549,0.3431379795,-0.2574048638,-0.0538627841,-0.5327308178,-0.1590914428,-0.1744761765,0.0389422961,-0.1136806682,0.2780540586,0.047677923,0.0992183313,-0.4922326803,0.193495661,-0.0417263545,-0.0505757779,-0.3259424567,-0.0463680401,-0.265906781,0.0837051868,-0.1113722697,-0.322360158,0.012878756,-0.1545253396,-0.0974247083,-0.1571799368,0.0748673677,0.0284899157,0.478715539,0.3693390489,-0.0176857151,0.4180803597,0.0488855541,-0.0095489221,-0.4613138139,0.0245149732,-0.1619866937,0.2181374431,0.0159047097,0.0216298476,-0.0940036923,0.3084729314,0.1127119139,0.3075918853,-0.3190087974,0.0016715898,0.056877967,-0.3979631066,-0.0713029653,0.4347881973,-0.0173200257,0.4235863686,-0.0750871152,0.0365405828,-0.4108154178,0.0092510087,0.2968645394,-0.325345993,-0.0109927366,-0.1295346469,0.3831002414,-0.0136334775,0.2570008337,-0.3716460466,-0.0795824155,0.3040837049,-0.0451909527,-0.0746097565,0.4503560066,0.0829839334,0.0080557372,-0.0372341499,0.0042993464,0.025876781,-0.0560947992,-0.3571121693,-0.1623252332]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"I believe this is really a very important feature, otherwise, we will still have the issue of too slow loading problems even if the data cache generation is fast.","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":29,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nI believe this is really a very important feature, otherwise, we will still have the issue of too slow loading problems even if the data cache generation is fast.","embeddings":[-0.1708837897,-0.1193052456,-0.108886078,0.2128687948,-0.0477886051,0.0981999189,0.1322444677,0.4315571785,0.3260004818,-0.1955100894,0.1550403088,0.2180582285,-0.1044813842,0.1303240508,0.2459831089,0.0549582057,0.0207399093,0.271002084,0.054923296,-0.1215847731,0.0354797617,-0.0033766872,-0.2422816604,-0.2905796766,-0.2007866055,0.0008391837,-0.0093753505,-0.0193931069,-0.1012573987,-0.2493915856,-0.016847536,0.0840887427,0.0404026546,0.1795384139,-0.000121539,-0.288066864,0.479102999,0.1190052107,-0.299382031,-0.0663854405,0.3308892548,-0.6440515518,0.1856696159,-0.1526859552,-0.0242588725,0.0225023907,0.1224434227,-0.1704470813,-0.0260150451,0.0512879379,0.0980148762,-0.076127857,-0.2565419078,0.2432713807,0.1302573234,0.2178975642,-0.0313274562,0.2933136821,0.4627696574,-0.2035045624,-0.2461470813,0.0658878833,-0.1829822361,0.0979476199,0.2545043826,0.0872411802,-0.1545416564,-0.0985595062,0.0784101933,0.3017698526,0.6385815144,-0.1194592118,-0.0844112113,-0.5116711855,0.1060038283,-0.2073562443,0.1937394291,0.2269611061,-0.2527628541,-0.23120749,-0.3389865756,-0.3271773458,-0.0391761065,0.2343063504,0.1697995514,0.0866003186,-0.0181806758,0.1299438924,0.189647615,0.1051436961,0.0536866523,-0.1221146211,0.256398201,0.380194515,-0.3880558908,-0.0825358033,0.0302411728,0.2858523428,0.1697465181,0.1421570927,0.1205964163,-0.0010591052,-0.1231785417,0.0225766636,0.2663192153,0.3897802532,-0.1854939759,-0.1530641019,0.4266661406,-0.1951057911,-0.1227803379,0.213754043,-0.1022178084,-0.0929636657,0.1184921041,-0.3602445126,-0.2023613751,-0.3070412278,-0.0508143753,-0.0440401994,0.0419510193,-0.2672597468,0.0199080184,0.1974839717,-0.3789928555,0.2880629301,0.0326742083,-0.082256645,-0.3433534205,0.0434473604,-0.132945925,0.019297367,-0.1289641857,0.237637639,0.5631651282,-0.0453249775,0.1705602854,0.0663459599,-0.1947800815,-0.0836630762,-0.1964970231,-0.3833451867,-0.0816078112,-0.0069685495,-0.0666886419,0.2241399139,0.0425942913,0.350884527,-0.439255774,0.1806218624,-0.3687042594,-0.1353670806,0.0462310985,0.1133784577,-0.4325674474,-0.3168877959,-0.2426804304,0.3940674961,-0.009996214,-0.2400588095,-0.17383039,-0.0357681774,-0.2878268957,-0.0154794501,0.093803063,0.2874269485,-0.0541599654,-0.0180748347,-0.0591328964,0.2659279406,0.3198640347,0.495621711,-0.1973638684,0.2581637502,-0.2065196186,-0.0021846569,0.5175064206,0.0865026414,-0.4842506945,0.5351499915,-0.2245032638,0.0454430096,0.2786035538,0.4554198682,0.0155416271,-0.0804585889,0.2262421399,0.6157296896,0.1381959319,0.2069051415,-0.4678855836,-0.1348220408,0.3957818747,0.414177537,-0.0689775348,0.0771701708,0.0468459278,0.1267206818,0.4656170011,0.1988309622,-0.0619081482,0.2839721739,-0.2368707359,0.0469793715,0.0830770507,0.1733320951,-0.2105592787,0.1638224274,0.155417949,0.0450230278,0.3792930841,0.1307999492,-0.1590068787,-0.211835295,0.1023089439,0.2358945608,-0.0275605209,-0.0554423295,0.0593646839,0.2304652631,-0.2147925347,0.440354079,-0.1765075177,-0.2186828852,-0.3693207204,-0.0035073732,0.2707007527,-0.049532406,0.2977488339,0.3167923987,-0.1896380633,0.0506399535,-0.131039381,-0.0754616484,-0.1972570419,0.1816370636,-0.0243195985,0.0862565935,0.0342186131,0.0019279764,0.3978556693,0.1988737136,0.2990909219,-0.4056387842,-0.2126014531,0.3579401076,-0.1928501427,0.2532295287,0.1134444475,-0.4938903451,0.1087489203,-0.1259760857,0.1698671281,0.2656657398,0.903899312,0.034859851,0.6662992239,0.2545955479,-0.1685732901,0.0134953149,0.4577048123,-0.0228779409,-0.0987972021,0.4839093983,-0.0615176596,-0.3843647838,0.0158465989,-0.111004211,0.4429910779,0.2537376285,0.0886205807,0.0565742515,-0.0263063088,-0.2527052462,0.1614331752,-0.1258850545,0.2274756581,0.2149233669,0.3197237849,0.1339851767,-0.4487345219,-0.2647607327,-0.0168763604,0.3032929599,-0.0794268474,-0.0173537135,-0.0245067421,-0.3835326433,-0.148156032,0.0173822474,-0.2176360339,-0.2085160613,0.0413723812,0.0394828171,0.4439233541,-0.1601712853,-0.006861045,0.1897861809,-0.2549465001,-0.5058605075,-0.3831742406,-0.1712242216,-0.3200800419,-0.0412952639,0.2373730987,0.1959146112,0.3190254271,0.1532439739,-0.1738554388,0.170615226,-0.1240507141,-0.1911413074,-0.0086410912,-0.005972161,-0.185142979,0.1018374562,0.2547336221,-0.0169275142,-0.0134271495,-0.3910280764,-0.059179917,0.1095030159,-0.0506862924,-0.0129666999,-0.0616958104,-0.2215944827,-0.3127224147,-0.2671322525,0.4247215986,0.0532290004,0.1396271139,0.157479763,-0.2318089604,0.2191645354,0.0809980482,0.2858385444,-0.040409822,-0.2382849306,0.2692323029,0.2617773414,-0.1944067031,-0.2885739207,0.0042175879,0.1090979725,0.0836814493,-0.7654343843,0.3025305867,-0.4086547196,-0.0389932729,-0.1383836418,0.169416219,0.2052452117,-0.2655721009,-0.0068380916,0.2257292122,-0.1746085882,-0.0823598206,-0.0131540671,0.2187665105,-0.1661929637,0.3839714229,-0.0843010768,0.0593582876,0.065373525,-0.0612383522,0.2289560139,-0.0226926059,0.1704556495,-0.2278119624,-0.2003519684,0.2423760146,-0.1477457881,-0.153709963,0.2911832929,0.0405597799,-0.523668468,-0.2941627502,-0.1583211124,-0.0664444417,-0.0404028744,0.158853963,-0.3776305318,-0.1067281589,0.0405201055,-0.1254660189,-0.2737348974,-0.7370407581,-0.1140795127,0.1241322905,0.193688795,0.02667946,-0.1326932311,-0.0473712683,-0.7960611582,0.0488155261,-0.157657519,0.4344090819,0.0749644116,-0.1837304384,0.3018525243,-0.089074105,0.3479284644,0.2913413346,0.1778750718,0.1246942058,-0.1579191536,-0.3158029914,-0.032115601,-0.0669330284,0.2634797096,0.4881898165,0.3126024008,-0.2156249285,-0.1253401339,0.0418464355,0.3029629588,-0.0100264894,-0.7628208995,-0.3536115885,-0.0647104159,0.0765328482,0.054641109,-0.2676482201,0.1482436061,-0.1760107726,-0.0063907593,-0.0748459175,-0.0502955727,0.2330686599,-0.0276344214,-0.2499347031,0.0962873995,0.3787204623,-0.0933473706,0.2527825832,-0.5406363606,0.4283747077,0.2015638798,0.190618664,0.0158231296,0.0758719221,0.4232341945,0.2755126953,0.0652987808,0.0898814946,-0.2765953243,0.1825899333,-0.2460808307,0.0451995879,0.0425409898,-0.0536249727,-0.4035268426,-0.4539190233,0.948481679,0.1471525133,0.0550566874,0.393635869,-0.4269388318,0.0688719377,0.1636858433,-0.3242270052,0.7935565114,-0.4691546559,0.3224421144,-0.1204101294,0.1067986861,0.5137325525,-0.1806548983,0.0926414579,-0.410043329,0.0047129057,0.2545189857,-0.027902808,-0.1416480988,0.280543834,-0.105655022,0.486440599,0.3472347558,-0.3024031222,-0.1157685593,0.6819103956,-0.1476734877,-0.2155460268,-0.5637030602,0.0641234219,-0.2749715149,0.4820745289,0.0241862331,0.142963171,-0.0827580839,-0.0242400896,-0.2062787414,0.1470733732,-0.0777715221,-0.0442319363,-0.0703182146,-0.3486212194,0.1795981526,0.1700843722,-0.2536796331,0.2336205989,-0.293781817,0.098264955,-0.342051357,0.2043760568,0.4629022479,0.0172054581,0.1355712861,-0.2207510918,0.0362211093,-0.1873764694,-0.1038953513,-0.2375708073,-0.0763846487,0.1820119619,0.0744049475,0.1673855186,-0.3001605868,0.290170908,0.0506366603,-0.032030981,0.0332940184,0.4427126646,0.0551371127,0.3508199751,0.0021837661,-0.4148605466,-0.0583070889,0.5324238539,0.1278883219,-0.1993600726,0.4141848385,0.1323919296,-0.1521719694,-0.1286331564,-0.0661162883,-0.089160867,-0.1918783039,0.0175546706,0.0948015898,0.0830990374,0.1859890372,-0.0450856574,-0.0991526544,-0.1269467771,0.1224898025,-0.4464525878,-0.3540443778,-0.003919614,-0.1062919572,-0.0885119885,0.0354365073,-0.0504938103,0.2558650076,0.3383277357,-0.1661081612,0.125355497,0.0610268191,0.0751438141,0.0404821336,-0.2260240614,0.034596134,0.2679469287,0.0642320514,-0.1063543037,0.1544245034,-0.1853528172,0.0156966373,0.1970503926,-0.0105966292,0.1639119238,0.0115315802,-0.2834026814,-0.0786420852,-0.2003032863,-0.0915802121,-0.0367644355,-0.0581973828,-0.025506055,0.0380092338,0.0103589455,-0.064276211,0.3120422661,-0.3961381912,0.190691039,0.0510309413,0.0342136063,0.3081752062,0.1043488309,-0.5492603183,-0.1659027785,0.1145075858,0.0709213838,0.1866657436,-0.1495612115,0.1903330684,0.0521307699,0.1587469876,0.5096777678,0.0740132704,-0.2529695928,0.2039005905,0.0371939242,-0.2015983611,-0.0274777487,0.2153618187,0.170675084,-0.106366694,0.1950926334,0.2131561488,0.0300229751,-0.2574487925,-0.3286288977,0.4000739753,-0.2123188525,-0.2883189321,0.0732744485,0.2181686014,-0.1305529475,0.2378870994,0.1110629141,0.013348476,0.3842390776,0.2560364902,0.3813843727,0.092795372,0.4703731835,-0.4008374214,-0.2251325697,-0.1482995749,0.5131306052,0.0866047069,0.134749189,-0.1501096338,0.24548015,-0.0429761969,0.1373547465,-0.1762051433,-0.0240671914,-0.0081065129,-0.0769458115,0.2742763162,0.0929182768,-0.140264228,0.1897150129,-0.303906858,0.0690493956,-0.0252105799,0.3115069866,-0.0612560362,-0.4201532602,0.1340332329,0.0630377829,0.2313789576,-0.2374030799,0.0885190666,-0.3801608086,0.0390045531,0.0805716515,0.1637848318,0.2026797384,0.3916238248,-0.0344123393,0.0159595832,-0.1590322554,-0.0349935554,-0.2816363871,0.4697636366,0.3143095672,0.3289676905,0.0662651286,-0.0791719779,-0.0623052083,0.2926388681,-0.0918344557,-0.2164207995,-0.1209673136,0.284450233,0.2031002492,-0.3142252564,-0.5752778649,0.1504357308,-0.276434958,-0.1777211279,0.1595373154,-0.0912328959,-0.0382678658,-0.0871387646,0.0184830464,-0.3198889196,0.6624272466,0.4286255538,0.4453484714,-0.1340379715,-0.2282948345,-0.5174891949,-0.0722060725,-0.2385068536,-0.0622383133,0.1356020123,0.264682591,-0.0000905103,0.1179217994,-0.4758056104,0.1493461728,-0.1230763197,-0.1929820925,-0.3171249926,-0.0792832822,-0.27109164,0.1241759211,-0.1513261497,-0.2715722322,0.0279217288,-0.2592737079,-0.1142924801,-0.1937058717,0.1246498674,-0.044099804,0.5869351029,0.3407252133,-0.0067029567,0.3802830279,0.0303096846,-0.0634885728,-0.4289747179,0.0660752207,-0.1503823698,0.1862996668,0.1481111944,-0.0619395003,-0.0912795514,0.3369539678,0.1819931865,0.2498596013,-0.2586732209,0.0191845689,0.0775856078,-0.309153676,-0.0328679122,0.358189702,0.0256290082,0.3765022159,-0.0346233658,-0.1105372384,-0.4577965736,0.0815941989,0.2500174642,-0.2777269781,-0.0528303273,-0.2340158373,0.3957236409,0.0628948435,0.2202779949,-0.3804607987,-0.0504566841,0.3263394833,0.1096779704,0.0260347445,0.5247809291,-0.0093297092,-0.1259953976,-0.0275588483,0.2195758671,-0.1400582939,-0.061450392,-0.3122717142,-0.2382739484]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"Hmm ok then maybe it's the hashing step indeed.\r\n\r\nLet's see if we can improve this as well.\r\n\r\n(you will very likely have to regenerate your dataset if we change this part of the lib though since I expect modifications on this part of the lib to results in new hashes)","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":51,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nHmm ok then maybe it's the hashing step indeed.\r\n\r\nLet's see if we can improve this as well.\r\n\r\n(you will very likely have to regenerate your dataset if we change this part of the lib though since I expect modifications on this part of the lib to results in new hashes)","embeddings":[-0.1601843834,-0.1073325351,-0.1088037193,0.2356132567,0.0101085901,0.0855073333,0.0967391133,0.4174629152,0.2409227192,-0.2387262434,0.1788458377,0.2212671041,-0.1573331654,0.2012864649,0.2440309823,0.0954292566,-0.0814827681,0.2566314042,-0.0297538601,-0.150974825,0.0977721512,-0.0070644319,-0.2352129817,-0.3177333474,-0.1379797757,-0.0537621565,-0.0029595352,0.1181511506,-0.1523030698,-0.2356422842,-0.0744617581,0.0946115702,-0.0337065309,0.1301782578,-0.000117367,-0.254348129,0.4368434548,0.103002578,-0.178303346,-0.0722184181,0.2452519536,-0.6293510199,0.1563462615,-0.113575235,0.0100137265,0.0221238304,0.0306232534,-0.0951652974,0.0151720811,0.0258699339,0.124812901,-0.157610178,-0.2547904849,0.3403293192,0.1736136824,0.1329304278,-0.0348422751,0.411365211,0.3614937663,-0.1394686401,-0.2442426682,0.0927665606,-0.1439820975,0.0500793457,0.299335897,0.0624082498,-0.0894317627,-0.0703959763,0.0655167624,0.3119708598,0.5350320935,-0.1155160889,-0.0738632008,-0.3971727192,0.0449689142,-0.1798521131,0.2664926052,0.1910128295,-0.1697357595,-0.291092962,-0.3915231824,-0.3189979792,-0.005201444,0.2116544843,0.2382818013,0.0413417444,0.0166872703,0.1217721477,0.3245376348,-0.0493633337,-0.0690267235,-0.0392776802,0.1879114807,0.3829360306,-0.3432565629,-0.0498863161,0.0962742269,0.2656586766,0.0248609781,0.1510502845,0.111714758,0.1768588722,-0.1563506573,0.006316388,0.2069072425,0.283395797,-0.0738158971,-0.2024073899,0.3430435956,-0.1749476045,-0.1971488148,0.2683002949,-0.1556073427,-0.218659997,-0.0255660247,-0.342505157,-0.2920804918,-0.3515900671,0.0043010889,-0.0825531855,-0.058711037,-0.2975071967,0.1371726394,0.2337220013,-0.4289889038,0.1917135715,-0.0391213372,-0.2029282302,-0.3366799057,0.0984217227,-0.1477579325,0.0849541873,-0.2443003058,0.2111695558,0.5522867441,-0.1529755741,0.2215201706,0.128443256,-0.1704710126,-0.0670168996,-0.2042763978,-0.2548729777,-0.059762869,-0.0222387612,-0.0546885282,0.2033077329,-0.0341567919,0.3675209284,-0.42348966,0.1777499765,-0.3505127728,-0.1370326877,0.0521859601,0.1265480518,-0.4527393878,-0.2849046886,-0.1414549798,0.3504903615,-0.0039889622,-0.2377236933,-0.2531360686,-0.0554263704,-0.1817226559,-0.0687226728,0.1630561203,0.2110370696,-0.1411571801,-0.0157532096,-0.0991911963,0.3512475789,0.3494640589,0.5109770298,-0.2431333363,0.2242118716,-0.1391341537,0.0753151849,0.5104481578,0.1495012492,-0.4814804792,0.5334195495,-0.1864397824,0.066399537,0.1667491645,0.4503003657,-0.015408475,-0.0686402321,0.1876243353,0.657384336,0.1865057796,0.2055634707,-0.5292838216,-0.1271956712,0.3831863701,0.3882476091,-0.1265484691,-0.0448073931,0.0644438118,0.2115140855,0.4432809651,0.1463390738,-0.0802837387,0.3024356365,-0.1130207554,-0.008691892,0.0369792543,0.1256117523,-0.0781993195,0.1612611562,0.1250467002,0.1854242384,0.2704234719,0.15310058,-0.137463361,-0.3148970306,0.0726239979,0.19421345,-0.0297198072,-0.0767292157,-0.0417796895,0.1080722511,-0.1495256722,0.2906084657,-0.2226118296,-0.1651922613,-0.4035986662,-0.0779212937,0.2123479843,-0.1582205445,0.3385802209,0.3289043903,-0.1488137245,-0.0629239455,-0.1064999029,-0.0856109187,-0.309491843,0.2200084329,0.0904118344,0.1272014976,0.0426448509,-0.0571574755,0.4239549041,0.1180785373,0.1862309575,-0.4027871192,-0.2185846269,0.4604168832,-0.2816222608,0.195881173,0.0863028616,-0.4638467431,0.065761894,-0.1463585347,0.1708156168,0.2818573415,0.8769376874,0.0479910634,0.6095963717,0.3669523001,-0.0704372153,0.0417420529,0.529009223,0.0008987521,-0.2038211375,0.521261692,-0.0849338695,-0.4068973958,-0.0294889249,-0.0923200771,0.3518699706,0.2258281559,0.0709662363,0.1058856919,-0.05677066,-0.2607472539,0.1747760922,-0.2040489167,0.2004164308,0.1718323082,0.2445509732,0.1183077246,-0.4640571475,-0.1943505853,0.0368423387,0.4481584728,-0.1012153774,-0.0404050685,-0.1280096173,-0.2470637858,-0.2019764185,-0.0580903403,-0.3640301228,-0.239800483,-0.0023724409,0.0400862098,0.383977294,-0.1228239685,0.0325509384,0.0749538541,-0.1560163945,-0.5239127278,-0.2754573524,-0.1642967016,-0.3267016113,-0.0128085306,0.3502834737,0.2617174089,0.2895052731,0.1520948857,-0.2226365209,0.2321971059,-0.198388651,-0.2286720276,-0.0677984878,0.0007592857,-0.1222772971,0.1360549033,0.1543964744,-0.048221536,0.0794742629,-0.3189553916,-0.0575904027,0.1524783075,-0.0264733545,0.0079263281,-0.0531881191,-0.3366261423,-0.1873835921,-0.2386597246,0.5235024095,0.0800590664,0.2400283366,0.2175990045,-0.1314043701,0.2389299721,0.0015136065,0.263714999,0.0267378297,-0.318177402,0.3042290807,0.3015608788,-0.1725576669,-0.1673830748,0.0746447816,0.114982307,0.1365856081,-0.7411957979,0.2916590571,-0.3721135557,-0.0532522425,-0.040381439,0.1372641474,0.161243856,-0.1977793127,-0.0433081537,0.2443802357,-0.2286248803,0.0600735471,-0.0032308525,0.2971607745,-0.1220101938,0.4379106462,-0.0273348112,0.1570204496,0.1014796421,-0.1560010314,0.1715011448,0.0441330485,0.0634028614,-0.1760016829,-0.2444697022,0.2238000631,-0.0720512792,-0.0700969175,0.3466936052,0.1279879361,-0.4035211802,-0.2299666852,-0.1928111911,-0.134305492,-0.0957183391,0.2891788781,-0.5735127926,-0.1390632242,-0.0276078265,-0.1219029501,-0.2293453366,-0.710714817,-0.0334077179,0.1148639917,0.1338197887,0.0355896838,-0.1537472755,0.0088324463,-0.6964544654,0.0708515644,-0.1736069322,0.4671595991,0.0477878451,-0.1481667459,0.3559194505,-0.1329542398,0.3270044625,0.2354595959,0.1489394903,0.1138510108,-0.1929860562,-0.2955775261,-0.0183350053,-0.1081781909,0.2827035785,0.5091449022,0.296872288,-0.2459940165,-0.1142121553,0.0738170743,0.2307710648,-0.0188426655,-0.5913386941,-0.299145937,-0.1378331929,0.0352209434,0.0417165421,-0.4061606228,0.1900845617,-0.1016412824,0.0375565477,0.0361320116,-0.0489720181,0.3762347698,0.0298861917,-0.2006280571,0.1272472739,0.4202657044,-0.1220819503,0.2732324302,-0.5304989219,0.4065213501,0.238378644,0.2210581899,-0.1334275305,0.0646393001,0.4113091826,0.2798776627,0.0700514987,0.1080987379,-0.2257949561,0.1607625782,-0.1429771781,0.1101176515,0.1430761963,0.0137173478,-0.4057683945,-0.4632485211,0.8867708445,0.1272925735,0.0569163263,0.448053509,-0.4334327877,0.0480899103,0.1586894393,-0.2592352927,0.8431890011,-0.4389936328,0.1994746476,-0.1670083702,0.1202710122,0.3556024134,-0.1527603418,0.0230371933,-0.5007897615,0.0173591301,0.2982678115,-0.0431024209,-0.0816722214,0.2605397105,-0.0874306113,0.3969627023,0.3509871364,-0.2249670327,-0.0968426988,0.7268240452,-0.1192472354,-0.214669168,-0.5672605038,0.1038514376,-0.3063289523,0.4123212099,0.0262241475,-0.0132764131,-0.0762001052,0.0199680254,-0.2648591995,0.1553102285,0.044124838,-0.0842340887,-0.0550622605,-0.2805173993,0.1886934042,0.0249195192,-0.2194394618,0.2818987668,-0.2779997885,0.1446072459,-0.4240193665,0.162023142,0.5813894868,0.0342736095,0.166753009,-0.1988769472,-0.0346663296,-0.2172168791,-0.0073398957,-0.3404229879,-0.0938075036,0.1256667525,0.0258531068,0.1218631417,-0.3384752572,0.2759360671,0.0006478658,-0.042335134,0.058821138,0.4667872488,-0.0157506876,0.4426684976,0.079273589,-0.3779088855,-0.0581216738,0.5386013985,0.1215896159,-0.1944125295,0.3733441532,0.244712621,-0.1593827009,-0.1300231814,-0.0689207986,-0.1411026269,-0.1078483686,0.0062505063,0.0818998516,-0.0025040552,0.2305915803,-0.0075667338,-0.0698211938,-0.0405272134,0.0405026414,-0.441070199,-0.3163051009,0.0238047615,-0.0345412381,0.0053882813,-0.1201699972,0.0403421,0.2122024745,0.3924250901,-0.2014086396,0.1785957813,0.0456348881,0.14701882,0.0360876545,-0.283774972,0.0253578741,0.1983897239,0.0944816098,-0.042813357,0.1105436385,-0.1971140057,0.0569720976,0.1666144431,-0.0283113476,0.056809146,-0.0241401158,-0.2199422717,-0.1199150831,-0.1984608322,-0.1657455117,-0.0402068421,-0.0369198471,-0.004049866,0.0812607035,0.0007808972,-0.0361267924,0.2397200167,-0.4830047786,0.1856575161,0.0966672897,-0.0207795277,0.2858249843,0.0304301307,-0.4567317069,-0.101353012,0.1237872317,0.1051801667,0.219739452,-0.1102239713,0.1212710738,0.0106820259,0.3315263987,0.5235652924,0.1338108182,-0.2428768277,0.2041597664,0.0784935281,-0.2715134323,-0.0141104227,0.2915894389,0.1058764458,-0.1336043924,0.1146291792,0.1202777848,0.0585417114,-0.3464828432,-0.3014535308,0.3380943239,-0.2441135794,-0.29797104,0.0643428341,0.0163056739,-0.2179073691,0.2900067866,0.0778814405,-0.0167018659,0.4894806743,0.1676052064,0.3421228528,0.0653250068,0.5332942009,-0.4446782768,-0.2065306753,-0.1768009067,0.5822597146,0.2081183046,0.0829707906,-0.0738840029,0.3838200867,0.0277765281,-0.0072327335,-0.1711737365,0.1092461199,-0.0463983342,-0.1083678454,0.1905538291,0.090356648,-0.2495438457,0.2571478784,-0.2787741721,0.0264761802,0.1017144918,0.3117561042,-0.0405870676,-0.4505388737,0.1253461689,-0.0486578457,0.2256065458,-0.2179592997,0.1314529926,-0.2777087688,0.0332339108,0.0723273903,0.1033718139,0.1764124781,0.3884591162,-0.0186529569,0.0564335287,-0.1125284135,0.0122587364,-0.2258788496,0.4924060106,0.2195760608,0.4541970789,0.0401571393,-0.0623214394,-0.1163429618,0.1922288686,-0.1800906062,-0.1461562812,-0.0361529663,0.3346709609,0.2543663979,-0.30171749,-0.5345801115,0.1077119038,-0.2916919589,-0.4246152639,0.1893160641,-0.0572944246,-0.0502091348,-0.1278002709,0.0379529819,-0.295509845,0.7509638071,0.4251607358,0.3682785332,-0.1710943431,-0.1999723911,-0.5557056665,-0.0755980387,-0.181099847,-0.0966887549,0.0442199074,0.3024515212,0.0710492581,0.0783822984,-0.4742082655,0.2333523482,-0.0662329271,-0.1416818351,-0.2677322626,0.0943398401,-0.339109689,0.0613313615,-0.1751963347,-0.198219955,0.1026215106,-0.2987852991,-0.0779157057,-0.1858755797,-0.0222114399,0.0004873693,0.4985751212,0.3433731496,-0.0345308408,0.4027955234,0.0035729299,-0.0944419727,-0.4025659859,0.050253462,-0.1398017704,0.2345340103,0.0715540573,-0.0042178878,0.041400481,0.4242451787,0.0843154415,0.2722791731,-0.3205932975,0.0311020724,-0.0058254027,-0.3514074683,0.0110239275,0.3562553227,-0.0384769551,0.4106634855,-0.0319841132,-0.1470578909,-0.5312728882,0.0306694843,0.3248066008,-0.3233669996,-0.0022172532,-0.1551621407,0.3223170042,0.0688256025,0.2539067864,-0.4685249329,-0.0492620803,0.241620183,0.0816294476,-0.0064782654,0.5046198964,-0.0128519703,-0.0934122279,-0.0103990762,0.0301523022,-0.0991655067,-0.1319022626,-0.3541839123,-0.1793479174]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"Also, @agemagician you have to follow the step I indicate in my previous message [here](https:\/\/github.com\/huggingface\/nlp\/issues\/546#issuecomment-684648927) to use the new text loading script.\r\n\r\nJust doing `pip install git+https:\/\/github.com\/huggingface\/nlp.git@07d92a82b7594498ff702f3cca55c074e2052257` like you did won't use the new script (they are not inside the library but hosted on our hub).","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":46,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nAlso, @agemagician you have to follow the step I indicate in my previous message [here](https:\/\/github.com\/huggingface\/nlp\/issues\/546#issuecomment-684648927) to use the new text loading script.\r\n\r\nJust doing `pip install git+https:\/\/github.com\/huggingface\/nlp.git@07d92a82b7594498ff702f3cca55c074e2052257` like you did won't use the new script (they are not inside the library but hosted on our hub).","embeddings":[-0.1538135111,-0.2194951475,-0.0950479582,0.1538355052,-0.0239236262,0.0457142889,0.1359891295,0.4179357886,0.3139533103,-0.176857695,0.1352542192,0.2417795211,-0.1323707402,0.2570710182,0.2852427363,0.0912736505,-0.0661661029,0.2590822577,0.0090136901,-0.1174314693,0.1052107662,0.0947560593,-0.2691370547,-0.2267983854,-0.1341264844,0.0427628644,-0.017962059,0.0809901729,-0.1078435481,-0.2789711058,0.0466638505,0.1192347333,-0.013505592,0.2386166304,-0.0001188216,-0.2400173396,0.4164008498,0.0876647085,-0.2083815336,-0.1413666755,0.3876317739,-0.589399755,0.221742332,-0.0976361707,-0.0658547804,0.0419179946,0.0641734824,-0.0316150971,0.0141374767,0.0813605562,0.1186160296,-0.069691956,-0.1954965889,0.2457896918,0.0393077806,0.2178897113,0.0103265932,0.355119437,0.4591064453,-0.2080248147,-0.2608448863,0.0574853346,-0.1576683521,0.1157764345,0.2842022181,0.053822007,-0.1553236246,-0.0795961469,0.1041076407,0.316495508,0.5134063959,-0.1090567484,0.012628559,-0.4890187085,0.014008943,-0.2325054556,0.2363929898,0.2235964835,-0.2313871235,-0.2666046917,-0.4041782022,-0.3050924242,-0.0624962598,0.2165593654,0.1770122647,0.0690088496,-0.0753798783,0.1108511388,0.2798572779,0.0148910899,-0.1217617914,-0.0262428243,0.2419862747,0.438036114,-0.3837111592,-0.1077051163,0.1144530848,0.2436068058,0.1071792543,0.1331069916,0.0522230975,0.0372730419,-0.1837209165,-0.0216732118,0.2654947042,0.4267695844,-0.1390134692,-0.1967364997,0.3785618842,-0.0578008182,-0.1065250263,0.2549192309,-0.1018854827,-0.1470218301,-0.0624277145,-0.4120223224,-0.1541306674,-0.4001076818,-0.0897248015,-0.1155261397,0.1008113995,-0.2847557068,0.0834816098,0.2124907076,-0.3558643162,0.289180696,0.110858269,-0.0932926536,-0.3905107379,0.1106145307,-0.0973509327,0.0845343545,-0.15649046,0.2319569737,0.5620113611,-0.1151951328,0.1986724883,0.1585441977,-0.1802897751,-0.1654041857,-0.306299895,-0.2818942964,-0.1220208928,-0.0347553603,-0.0456170775,0.2278311402,0.0319258347,0.2941394448,-0.3541756868,0.122178942,-0.360491097,-0.1590794921,0.0083839772,0.1173997149,-0.4411828816,-0.4041283429,-0.2906878591,0.403341502,-0.0657472387,-0.1836446822,-0.2509905696,-0.0534896962,-0.3046061695,0.0265602935,0.145907864,0.2639088929,-0.0665433556,-0.0997292176,-0.0743318498,0.3339369297,0.30014503,0.5857614279,-0.1756265014,0.2770580649,-0.177849561,0.0134650553,0.4331519902,0.0287085902,-0.3839407861,0.4948644638,-0.2551279664,0.0936986282,0.1792030334,0.3430300057,-0.021170212,-0.1265482903,0.2647036016,0.5780928135,0.210837543,0.2118260711,-0.4929418862,-0.1691846251,0.2549579442,0.4343913198,-0.1443443298,-0.0401955657,-0.07403256,0.1021754965,0.4322358668,0.1786887497,-0.1159980893,0.3622051775,-0.1837661564,0.1298588663,0.0351084508,0.1390739828,-0.1245252043,0.157169804,0.1711615473,0.1409959495,0.342733413,0.1587989479,-0.2499694973,-0.2745953798,0.1144898161,0.1902351528,-0.0246140845,-0.043202661,0.109130539,0.2025135458,-0.2353196442,0.3263877928,-0.1443891823,-0.1601638943,-0.4143637121,-0.0626252517,0.2259680033,-0.1562429965,0.36173895,0.3640159965,-0.119945243,-0.03927017,-0.1009056866,-0.0439311415,-0.3276138604,0.2787549794,0.0462335013,0.0630746409,0.0254493598,-0.0205107369,0.3717831969,0.1491747648,0.2595142424,-0.360008806,-0.2161252797,0.3663825393,-0.2371299416,0.2518211007,0.1492191404,-0.4946300685,0.0410462096,-0.1400924325,0.1856495738,0.26307863,0.9809722304,0.0491882898,0.5438917279,0.2374862581,-0.1860270947,0.0135343093,0.4733050168,-0.0671336725,-0.1612819731,0.5493712425,-0.0972786024,-0.3391580284,-0.0454707332,-0.1855878681,0.3958618641,0.2632412612,0.027101215,0.134681195,-0.0331405513,-0.2451714426,0.1936379522,-0.1551497281,0.1723720878,0.1338720918,0.3242029548,0.0977158695,-0.4654926956,-0.1708809137,-0.0920765027,0.3634902835,-0.1205447838,-0.0147363553,-0.0754011124,-0.3507580161,-0.2301577181,-0.0362827294,-0.3193768561,-0.1914129555,0.0126554789,0.1060666591,0.418626368,-0.1617042124,0.0398750491,0.1301965564,-0.2356345356,-0.4469167888,-0.3646012843,-0.1023060977,-0.311170429,-0.0221492946,0.2604712546,0.2491402775,0.3346447647,0.1370214373,-0.2228450477,0.2680353224,-0.1485192925,-0.2124429941,-0.1045691669,-0.009521598,-0.1503900439,0.2072898299,0.2030453831,0.017159624,0.1038346887,-0.4042710662,-0.0231148154,0.1369468719,-0.025765216,0.0110459067,-0.0730471164,-0.264147073,-0.2755765617,-0.2684983313,0.547409296,0.1586174965,0.1890622526,0.236809358,-0.2227825373,0.2654804289,0.049927853,0.2723809779,-0.0391039401,-0.1843655556,0.2653585672,0.2940625548,-0.2037299126,-0.2006460279,0.1070316434,0.1367928982,-0.0888872966,-0.7485929728,0.2411628515,-0.4384424686,-0.0879486427,-0.057007391,0.1646114588,0.2118792385,-0.2826988995,-0.0507269502,0.2520308495,-0.227019459,-0.0051456485,-0.0192309879,0.1749121249,-0.1399839669,0.4712072015,-0.071143508,0.0566901155,0.1988454908,-0.0808589086,0.2197439671,-0.0378409512,0.1163278893,-0.2227664441,-0.2527053356,0.2082594782,-0.0971232653,-0.0405654721,0.3472695947,0.1087502986,-0.3780517876,-0.3060775101,-0.1786088347,-0.0703601688,-0.0841788724,0.1630303264,-0.2872516215,-0.1635906845,-0.0280779079,-0.1237446964,-0.2175312191,-0.7590795159,-0.1072119623,0.1878442615,0.1177792102,0.0067719817,-0.1637411863,0.0266660005,-0.8222570419,0.1097549349,-0.2407436222,0.3917474449,0.0051581846,-0.1589802504,0.3964145184,-0.1695563495,0.2948711514,0.3480378985,0.1049225405,0.1405104846,-0.1382962912,-0.4310269356,-0.0184935704,-0.1026415527,0.2042777836,0.600551784,0.3204458356,-0.2217426151,-0.1375665963,0.141061008,0.2688470483,-0.0296869781,-0.6407162547,-0.3850316107,-0.2039022297,0.0176880211,0.0526278503,-0.2620922327,0.1665073931,-0.1924744993,0.0582242347,0.1080174223,-0.0836961269,0.2440255731,-0.0485575795,-0.2378832698,0.1179621965,0.3290274143,-0.0206172671,0.214784205,-0.5225408673,0.4845505953,0.2467738986,0.1944158822,-0.0512727536,0.1031717584,0.4254483879,0.3236665428,0.1285507083,0.1109720767,-0.2379472405,0.2317786068,-0.2935094535,-0.0006207124,0.1391830593,-0.0066116774,-0.3307585418,-0.394857049,0.8833647966,0.1604795754,0.0862204507,0.374040544,-0.3879625499,0.0101389205,0.1129832342,-0.3719643056,0.8148231506,-0.4314712584,0.2765333951,-0.0590740219,0.0877641365,0.3967078626,-0.1042747572,0.0209738985,-0.4423851371,0.0922846273,0.2870862484,-0.0151306065,-0.117352128,0.1940272003,-0.1858384758,0.4905821383,0.319214642,-0.2694951892,-0.1522089094,0.7354371548,-0.1749265194,-0.2167393118,-0.645986557,0.087080203,-0.3232801855,0.5647565722,0.0320597515,0.1008213609,-0.0531559251,-0.0877471343,-0.2761627138,0.1468392015,-0.0159174874,-0.1591123492,-0.0253691357,-0.2758975327,0.2373277247,0.0997473523,-0.1956424117,0.2086945027,-0.2587711811,0.1135695651,-0.4159818888,0.1700321585,0.5306617618,0.0424905159,0.128677398,-0.1998658925,-0.0250678863,-0.2198510915,0.0235133693,-0.1915742457,-0.1013419703,0.132784009,0.0029748997,0.0669255853,-0.2910996079,0.2608261704,0.1093739867,0.0212026834,0.050393112,0.431158334,0.0453743525,0.3068483472,0.0473052934,-0.3876747191,-0.0522999093,0.536680162,0.1274870336,-0.2194700539,0.454970777,0.3048579395,-0.1194844022,-0.1240122914,-0.0843766928,-0.0591268726,-0.2533002496,0.0543560833,0.2050640136,0.0701524913,0.2456106991,0.0874527395,-0.1702506393,-0.1359898746,0.0609174147,-0.4060442448,-0.3884435296,0.0203083884,-0.0279723871,-0.0565355718,0.0271126237,0.0002156237,0.2553329766,0.3067405522,-0.1939477324,0.156815201,0.0301953461,0.0572671369,0.1338682771,-0.3019939065,0.0996770635,0.1949741691,0.0877972469,-0.1048567444,0.1284299195,-0.202447772,0.0640576258,0.1629898697,-0.0517631136,0.0146521917,-0.0578785166,-0.2419179082,-0.0619424209,-0.1997778118,-0.0647026449,-0.0572800785,-0.032186117,-0.0201481804,-0.0409631394,-0.0384059772,0.0375934802,0.2064875215,-0.4180062711,0.2450839877,0.1308741271,-0.0194787402,0.2973195612,0.0806586593,-0.5015334487,-0.0753707811,0.1877058893,0.1760762036,0.1198439375,-0.1874102801,0.0878273025,-0.0222263876,0.2007060349,0.5044090748,0.0197257828,-0.2055368423,0.1735659391,0.079379946,-0.1997905076,0.0575193986,0.2935539484,0.1656487286,-0.0971027836,0.181015864,0.1531108171,0.0339189954,-0.2126951516,-0.2405851334,0.3971414566,-0.2115814537,-0.2819989026,0.0333866477,0.1631562263,-0.2190025598,0.1968732178,0.1078827456,0.0027401319,0.3443065584,0.1866302341,0.2403716296,-0.008709331,0.4980347455,-0.428935945,-0.2252036035,-0.22066167,0.491269201,0.1481189728,0.1297488958,-0.0765360072,0.368855834,-0.0934270397,0.0200453978,-0.2058987021,0.1213426143,-0.0305216573,-0.0982712582,0.1794662774,0.1005722433,-0.2097426355,0.2337902784,-0.2923447192,0.072654739,-0.0779219195,0.2915318012,-0.0786120594,-0.4484071732,-0.0169525985,-0.0345751494,0.2662855983,-0.1513136178,0.2016660124,-0.2848907113,0.0933898464,0.0807095021,0.2364780009,0.288494736,0.360955745,0.0017196281,0.0992551073,-0.1842439473,0.0128813656,-0.1870212704,0.4430474639,0.2176151723,0.4096033871,0.0002743271,-0.0485134237,-0.0907899216,0.1589909345,-0.1634992361,-0.2090654969,-0.1326423138,0.3491035998,0.2128050774,-0.2817459702,-0.6433497071,0.088339664,-0.2993735373,-0.1925557703,0.1242974848,-0.0390458591,-0.0400079489,-0.1342176795,0.0340582579,-0.2491402626,0.6754311323,0.398175627,0.3670075536,-0.0854026973,-0.2425558567,-0.5165573359,-0.1479375362,-0.2220449746,-0.1068291217,0.0941355154,0.2349832952,0.0108161196,0.0494082868,-0.4032297432,0.1061641425,-0.0794841573,-0.2023797035,-0.267960757,0.0747052655,-0.2873657346,0.0908478945,-0.13992019,-0.1393684596,0.0178873334,-0.2966466248,-0.1048668921,-0.1611986607,0.1833269745,-0.0955511406,0.4189105928,0.2183757722,-0.0168883279,0.3727635145,0.0274847094,-0.1051087305,-0.4040213227,0.0599322803,-0.2096919417,0.2013243139,0.0908446163,-0.0278294161,-0.0458412096,0.2639580667,0.095299162,0.268160224,-0.2644178271,0.014308569,0.0921965837,-0.3141652942,-0.0095516322,0.3425393105,-0.0030010494,0.4836660326,-0.0296046585,-0.0528521985,-0.521684885,0.0340327919,0.3458130956,-0.2806152999,-0.0370056182,-0.1729828268,0.3743341565,0.0503041223,0.2137223929,-0.4686451852,-0.0288100187,0.2371795177,0.157663241,-0.0020752207,0.5520824194,-0.06353046,-0.1372726858,0.0172430705,0.202183038,-0.061425928,-0.1073466241,-0.2837381959,-0.153096348]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"No problem, I will regenerate it. This will make us see if we solved both issues and now both the data generation step, as well as the hashing step, is fast.","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":31,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nNo problem, I will regenerate it. This will make us see if we solved both issues and now both the data generation step, as well as the hashing step, is fast.","embeddings":[-0.151921019,-0.1488330811,-0.1157126203,0.2396368831,0.0376232602,0.0494575314,0.1378421187,0.4157027304,0.2257434875,-0.2202196121,0.1811740845,0.2084628195,-0.1486444473,0.2225388736,0.2400526702,0.1334556341,-0.0106428927,0.2384750545,-0.0394382626,-0.1812114716,0.0848225802,-0.0164611079,-0.1985274255,-0.2775098979,-0.2173950672,-0.0564304069,-0.0055379877,0.0889179185,-0.1562844366,-0.246810019,-0.0427505188,0.065219067,-0.0512204841,0.1408244818,-0.0001163276,-0.2455975711,0.4251794219,0.1135242581,-0.1921709031,-0.0729340538,0.2216125429,-0.623847127,0.0891681835,-0.0979589522,-0.0205341838,-0.0389111713,0.0668587983,-0.0750634819,0.0484487824,0.0367645472,0.1141063571,-0.1823955625,-0.2229845375,0.2734904587,0.0973575711,0.2033148557,-0.0078115729,0.4100628495,0.3281129003,-0.1822448522,-0.2425540239,0.0311107002,-0.152671814,0.0686310306,0.2688613236,0.0768638253,-0.0380756818,-0.1262125522,0.117283076,0.2772989273,0.5142145157,-0.074316822,-0.0818330944,-0.3553854525,0.0407789089,-0.1829119474,0.3108657598,0.2620792687,-0.2431778312,-0.2519623935,-0.3767651021,-0.310742408,-0.035888847,0.1691657752,0.1969483644,-0.0173001047,0.0154272225,0.111554116,0.2200426757,-0.0238085818,-0.0098397965,-0.0847477019,0.2152493298,0.3841085434,-0.3390983343,-0.0486883745,0.0806409791,0.2237048447,0.0989988446,0.1023349762,0.0983228534,0.0823981762,-0.0574505329,-0.000183278,0.1947009563,0.2265612036,-0.1181398332,-0.2156523764,0.3537078202,-0.241489321,-0.2157759666,0.2924395204,-0.1502350569,-0.164395541,0.0223640148,-0.3545993567,-0.2640551031,-0.3176326454,0.0175317023,-0.0621748343,-0.0178758595,-0.3224639297,0.0862378106,0.2377759814,-0.4192526937,0.1554706991,-0.0714039877,-0.1302153915,-0.2981193662,0.1277148724,-0.1737806499,0.0912984237,-0.2290061861,0.2068091035,0.5647193193,-0.1109810919,0.1604576707,0.2225558013,-0.2360689342,-0.1250260323,-0.2276332825,-0.3216445148,-0.0550676398,-0.0739355162,-0.0266831201,0.2162742764,-0.0251332857,0.378280431,-0.3872926533,0.1844607294,-0.3624752164,-0.1348638237,0.0935598612,0.1424205601,-0.4249199927,-0.2871745527,-0.1541262567,0.3690338433,-0.0186956637,-0.1709997356,-0.3164302707,-0.0107481014,-0.1685535312,-0.0905547142,0.1508934796,0.1897432357,-0.0913726017,0.0287128892,-0.0428013168,0.3228217661,0.3777242303,0.5567439795,-0.2224965096,0.2703514695,-0.1818801761,0.0393540747,0.5186567307,0.1584905088,-0.4370102286,0.5346621871,-0.2084816396,0.0218877029,0.2098893076,0.356403321,0.0203293487,-0.02317366,0.3230544329,0.6156225204,0.1618902087,0.2482868731,-0.5206919909,-0.1039238572,0.3324407637,0.3731040657,-0.1081128865,-0.000086043,0.030773703,0.1343970597,0.4153932929,0.1156594455,-0.0639218614,0.3338144124,-0.1479671746,-0.0461843796,0.0276160985,0.1599299312,-0.1134731323,0.160213992,0.160799697,0.1595616192,0.2463900149,0.1536947638,-0.1623039842,-0.2808181047,0.1097992957,0.2023738772,-0.0130556077,-0.0812413245,0.0570364483,0.1287442595,-0.1415074021,0.3597144485,-0.2277530432,-0.1650277674,-0.3817080259,-0.0339083448,0.2300840914,-0.1821877211,0.3948881328,0.3099307716,-0.131473437,-0.0184682682,-0.1039199531,-0.1602776349,-0.2824917138,0.2352555543,0.0922763199,0.1354871094,0.0085610785,-0.0969768763,0.4391179979,0.0276454967,0.223895818,-0.4000404477,-0.1980136633,0.4823370874,-0.2769957483,0.1535592973,0.0893522128,-0.4861752391,0.165191859,-0.1509474516,0.2142757028,0.2887521982,0.8977122903,0.0871162489,0.6330669522,0.3522446454,-0.0897753462,0.0587165989,0.5588106513,0.0003701936,-0.2278028131,0.4814253747,-0.1175043285,-0.4113434851,-0.0957972556,-0.1145308167,0.3423905373,0.2279917747,0.0552585199,0.0877815932,-0.0244908687,-0.2676412761,0.2117271125,-0.1949855238,0.219550252,0.1812288314,0.3146229386,0.0641861632,-0.4541725516,-0.1606372297,0.0818958655,0.4434013665,-0.0880121142,-0.0837822482,-0.065564476,-0.2384914905,-0.1835364699,0.042312637,-0.3184978068,-0.2116201371,-0.0081897108,0.0785562247,0.3982796967,-0.1201592758,0.0863153487,0.0470030606,-0.2222497761,-0.4293803871,-0.3244926929,-0.178023234,-0.2633098066,-0.0021639171,0.3373177648,0.2433393449,0.357989639,0.1530640423,-0.2724926472,0.2025842816,-0.1350709647,-0.2082981318,-0.0527464487,0.0306908339,-0.1651145816,0.1683881879,0.1188065708,-0.0561857074,0.118415527,-0.3244405687,-0.1042575091,0.1782276928,-0.0425552502,0.0216902774,-0.0231979545,-0.2697674632,-0.2457210869,-0.2183193415,0.4573231637,0.0718535185,0.2080427259,0.1711173505,-0.1983994991,0.1929040849,0.0591406003,0.2377710491,-0.0346497893,-0.3612533212,0.3110781908,0.2732957304,-0.2055792212,-0.2129628658,-0.000386686,0.1451835185,0.1592358202,-0.6935850978,0.3073457181,-0.4094820917,0.0009773812,-0.0250767004,0.1169303656,0.1280314624,-0.1906652898,-0.0838195235,0.2492346168,-0.1603615284,0.0423882455,-0.0349028073,0.2069481313,-0.0972284749,0.493847996,-0.0372571051,0.1877983958,0.1699877828,-0.1237929985,0.1305236965,0.007146819,0.0334258601,-0.1676997095,-0.2454053611,0.2036920339,-0.0900606513,-0.0877127051,0.277251184,0.0523031913,-0.4173797369,-0.2477504313,-0.1811407208,-0.1873638779,-0.091823414,0.2865215242,-0.5284758806,-0.0481063426,0.0373417586,-0.1261239946,-0.2916545272,-0.7598503232,-0.0795493424,0.0665853396,0.1641322672,-0.00177769,-0.1557506174,-0.0200949591,-0.8061320186,0.041666612,-0.1593798995,0.4539009333,0.0435040109,-0.1204279214,0.3541276157,-0.2013698071,0.3818622231,0.2417010665,0.2068860233,0.0798383653,-0.2099682242,-0.2572870851,0.0196315087,-0.0897798166,0.2821121514,0.5095995665,0.3365418315,-0.2804056108,-0.169502303,0.0357937105,0.245988071,-0.021690093,-0.5882351995,-0.2670210898,-0.1094954163,0.0767495334,0.0446323082,-0.3609039783,0.2183086574,-0.2051842362,0.0153711094,-0.007032813,0.0009391541,0.329998374,0.0006128092,-0.2278040946,0.0404113233,0.4593442976,-0.1085021123,0.2746824622,-0.4829615355,0.4173160791,0.2848142087,0.1655387282,-0.1102580205,0.0440889075,0.4188216627,0.2915816009,0.0836194307,0.0606903508,-0.1767797023,0.1607070565,-0.2304643989,0.0047920961,0.1330128163,0.0144372834,-0.4678266346,-0.4470437169,0.820092082,0.1840217561,0.075641863,0.4255055189,-0.4305782318,0.0649195835,0.1462249011,-0.3074306548,0.8372446895,-0.452318728,0.2127172947,-0.1174648181,0.1125611514,0.4033430517,-0.117192544,0.0232849047,-0.5333122015,0.0126796644,0.3159544766,-0.0936960354,-0.0982322991,0.3226341605,-0.0720538571,0.3586013317,0.3544499576,-0.2024106234,-0.1047167107,0.7358256578,-0.0964388102,-0.2277330607,-0.610696137,0.0926401168,-0.3161284328,0.3682957292,0.0124187684,0.0283616185,-0.0909684971,0.0002190753,-0.2489505112,0.1964833289,0.0222944692,-0.0129739027,-0.0664415509,-0.3269301355,0.1552107632,0.0258802082,-0.1196146309,0.2387782037,-0.2393494099,0.1571929753,-0.3865939677,0.1956171095,0.5479066372,0.0184888635,0.1701104343,-0.2476156503,-0.0154640693,-0.1794599146,-0.0350059234,-0.2775325477,-0.1322488785,0.0984091833,-0.0219792295,0.1278960407,-0.2948842943,0.2470289022,-0.0133962659,-0.0488163903,0.0624469891,0.4836609364,0.0340440944,0.3923449516,0.0793342665,-0.3864131272,-0.0471248589,0.5738090873,0.1745922267,-0.183954522,0.4202536941,0.2087324113,-0.1922680736,-0.161516577,-0.0472017266,-0.1382559985,-0.1656136662,0.0393201746,0.1203709319,0.0757086128,0.2407544702,-0.0242269803,-0.096984528,-0.0998341888,0.1119725332,-0.4448240697,-0.3277871907,0.0251970291,-0.0319350399,-0.0786001831,-0.0284308717,0.0876648426,0.1883689165,0.4070272446,-0.2081248462,0.1860581934,0.0399078466,0.1396618485,0.0374499634,-0.3408958316,0.0609892569,0.2001645714,0.1273373216,-0.0768772438,0.1126628965,-0.2216008604,0.0783965737,0.1619157046,-0.0161785055,0.0360280611,-0.0057047773,-0.2285242677,-0.1336786151,-0.1288510859,-0.133219853,-0.0244696718,-0.02715482,-0.0701318607,0.0374011919,0.0630754456,-0.0289538521,0.2631289661,-0.4573256075,0.1790242046,0.040820349,0.0016271655,0.2796972692,-0.0150453234,-0.5266523361,-0.1142414063,0.1430064887,0.0916510671,0.2667956948,-0.1470529437,0.0916664377,0.0101627139,0.2162152082,0.4975883365,0.1279038191,-0.2263847888,0.2040489316,0.0754474029,-0.2107714415,-0.0211809333,0.2762135863,0.0733466148,-0.1190045923,0.1031397134,0.1193346754,0.0505637005,-0.3744419217,-0.3314027488,0.3522191942,-0.3175426126,-0.3191577792,0.0585369803,0.1417465955,-0.1946662664,0.262262851,0.1232298166,-0.028596852,0.5310922265,0.2104306221,0.3036752343,0.0958800465,0.4896481633,-0.3807523549,-0.2418751121,-0.1851466596,0.5272130966,0.0734564736,0.0661835372,-0.1291834563,0.3042574227,0.029554544,0.0364185087,-0.2261734605,0.1182015315,-0.050620243,-0.0577845871,0.2026021183,0.0818910822,-0.20654127,0.2879891992,-0.3173886836,0.0886031091,0.0695467889,0.3331947029,0.0021915608,-0.3964429796,0.1359227002,-0.0196419638,0.2442291081,-0.2184237093,0.0984248444,-0.3111554086,0.0590207167,0.0256308783,0.1395989358,0.2233558893,0.3780044913,-0.032432247,0.0409342013,-0.0881008506,-0.0551777147,-0.2662888765,0.5282589197,0.2188618034,0.4217073619,0.0443362556,-0.0360290967,-0.1194537655,0.0805527568,-0.0876427591,-0.178139627,0.007475995,0.2960134149,0.2574980557,-0.3041255176,-0.4800012112,0.0982157066,-0.2988273501,-0.3871194422,0.185552001,-0.0563837551,-0.0452754162,-0.1175369397,0.0496161208,-0.3259602487,0.7584056854,0.4055704176,0.3671917319,-0.1711865962,-0.2273601294,-0.4747660756,-0.106892027,-0.1883242726,-0.0390505195,0.0362219587,0.31191504,0.1116375998,0.058551196,-0.5351706147,0.2161413282,-0.1605439037,-0.0492740683,-0.2908783853,0.1094960645,-0.3576453924,0.0739117712,-0.1917953491,-0.2154418379,0.1510277539,-0.2720613182,-0.0632472932,-0.166684106,0.0175186172,0.0176632311,0.4537422359,0.3330732584,0.0163753256,0.4121362567,-0.0482990406,-0.0997279286,-0.3684396744,0.0448549539,-0.1481306702,0.3600824773,0.1196820438,-0.0243292283,0.0052560931,0.365052402,0.1344748586,0.2595983446,-0.3519541323,0.0271038767,0.0509630702,-0.2954756618,0.0557381213,0.3249191046,-0.0039188527,0.4142499268,-0.0009483723,-0.124733001,-0.4926358759,0.0283457898,0.2743467093,-0.3181958199,-0.0327874832,-0.1831097007,0.3310082555,0.0620222725,0.2665617764,-0.3928840458,-0.0350352265,0.2672197223,0.0293984022,0.0250360798,0.5127064586,0.0603040978,-0.1070545912,-0.0263854284,0.0339891724,-0.1410609931,-0.1633292288,-0.3431574106,-0.1546515673]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"Any news for the hashing ?","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":6,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nAny news for the hashing ?","embeddings":[-0.135568127,-0.1679757982,-0.1173763201,0.2645759284,0.0156154856,0.1071171463,0.1318227798,0.4035707712,0.2499975562,-0.2276438475,0.153519094,0.1644650996,-0.1642705798,0.1887554675,0.279314369,0.1575817764,0.0111409007,0.2219391167,0.0500924885,-0.1764005274,0.0771498904,-0.018003013,-0.2502294183,-0.2750237882,-0.1785914451,-0.0341135338,-0.0176673066,0.0391661897,-0.1399887651,-0.2255966663,-0.0493646264,0.0539096072,-0.0536811873,0.2000684291,-0.0001201887,-0.2316441536,0.4392504096,0.1204971969,-0.1374653727,-0.0789382085,0.2026335597,-0.6502266526,0.1439642906,-0.0968083218,-0.0048188483,-0.0350757241,0.0832937211,-0.0844039917,-0.0057960656,-0.0064956443,0.0963453129,-0.1689699441,-0.2549466193,0.2965782881,0.1082332805,0.1542338133,0.005129376,0.4148212671,0.3381083012,-0.200744763,-0.2361472249,0.0227255337,-0.1363456249,0.0548584834,0.2568830848,0.0695583895,-0.0719247609,-0.0397266746,0.0974143967,0.2698521614,0.5360155702,-0.0819927529,-0.0952928439,-0.3762821555,0.0037675439,-0.1763651967,0.2902004123,0.246731028,-0.1815931499,-0.271686703,-0.4254911542,-0.3020046949,-0.0138869444,0.2052422315,0.1689676344,0.036951147,0.0318495631,0.1177117974,0.2571713924,-0.0395128019,-0.0641419366,-0.0925879031,0.2376888692,0.3826837242,-0.3412059546,-0.0669870377,0.1042862684,0.2512679994,0.0475549176,0.1004544422,0.1189336553,0.1413401514,-0.0965512693,0.0191545282,0.2491918355,0.2782239318,-0.1381891668,-0.1781016141,0.3427100182,-0.252668947,-0.2478661537,0.2569594979,-0.1433359832,-0.1694235653,0.0766896084,-0.4076299071,-0.2522549927,-0.3669342101,0.001111347,-0.0732373372,-0.0532624871,-0.3043573797,0.0951319113,0.2250630707,-0.407710731,0.1626761407,-0.0679235756,-0.1291411966,-0.3446945846,0.1559618115,-0.1562259346,0.0649410337,-0.243956551,0.1859498024,0.5268715024,-0.114823319,0.1601588577,0.1569849551,-0.1583259106,-0.1121533886,-0.2454874814,-0.2966223061,-0.0374716334,-0.0696249157,-0.0581662953,0.248914063,-0.0114082592,0.3411039114,-0.409265548,0.1407239437,-0.3690372705,-0.1041257828,0.0877106711,0.1162682399,-0.4426725209,-0.3416030109,-0.1552458256,0.3505348563,-0.012420577,-0.2204845995,-0.2775160372,-0.0361937284,-0.1941386014,-0.0967229158,0.1024558321,0.1707753986,-0.0811160356,-0.0264615919,-0.0180685427,0.3286243081,0.3275488019,0.5740589499,-0.2077241838,0.179547146,-0.1706608981,0.0484036878,0.4910611808,0.0906848162,-0.4995357692,0.5159525275,-0.1963765323,0.0352322161,0.2432277352,0.4338689744,-0.0227990746,-0.0569686964,0.3088056445,0.5780259967,0.1788394302,0.1990534961,-0.5083053708,-0.0790513679,0.3344772458,0.3809754848,-0.0626497492,-0.0601875521,0.0689005032,0.1813963205,0.4310930073,0.200332284,-0.0826641843,0.3457258344,-0.1189662591,-0.0042554899,0.0638419315,0.1519672275,-0.0779205114,0.1330109388,0.1708758324,0.1144196093,0.2358637154,0.1610815227,-0.1391715556,-0.2834704518,0.0781170055,0.2152483165,-0.0628132299,-0.1076933146,-0.0069244057,0.1504240632,-0.183795318,0.3051694632,-0.1730572134,-0.16839616,-0.3719194233,-0.0280093439,0.2734823823,-0.198860094,0.3423899114,0.323748976,-0.1399327368,-0.0346578434,-0.0710451454,-0.151983887,-0.2454446703,0.2121457756,0.1087777168,0.2038850933,0.0421254821,-0.1089183837,0.4345690608,0.1402393878,0.2208811939,-0.3924812973,-0.2543740571,0.4574328959,-0.2766771913,0.1356613934,0.1249099374,-0.5178820491,0.1155586466,-0.1283958554,0.186171025,0.2633789778,0.9024999738,0.04700353,0.6106613278,0.3528787494,-0.0961636379,0.0051083597,0.5726713538,0.0018944095,-0.1883677989,0.4641438425,-0.1548915654,-0.3614988923,-0.0803396031,-0.1647896469,0.3335800469,0.2316286117,0.127688244,0.0942627639,-0.0393796787,-0.2729051113,0.1720833778,-0.1892972291,0.1557424366,0.1858523339,0.2948763371,0.0795812383,-0.4504948258,-0.162116006,0.0834264383,0.3833135366,-0.1354967058,-0.0613132417,-0.1100228652,-0.2578619719,-0.1649397314,-0.0368242413,-0.3088656962,-0.2576654851,0.0094126649,0.0716180503,0.4421849251,-0.1275012493,-0.0002466782,0.0727500394,-0.1718100756,-0.5184136033,-0.267909646,-0.1434024572,-0.2706848979,-0.0223430227,0.3221218288,0.1714126766,0.2907131314,0.1560173333,-0.2325373739,0.2141088098,-0.1682055593,-0.2460450381,-0.0307588559,0.0144992024,-0.1766727418,0.1484038383,0.0856401473,0.0115054809,0.1122067422,-0.3551222384,-0.0898993537,0.1611426771,-0.0001170794,0.0259947367,0.0000682386,-0.2243504077,-0.2405676544,-0.188049227,0.4681168199,0.1170293391,0.1744459867,0.2214355618,-0.201141715,0.2010892481,-0.0262036081,0.1733050793,-0.0309348274,-0.3551903665,0.3177434504,0.2951759696,-0.1763511449,-0.1780849248,0.05974048,0.1136544645,0.1374809295,-0.6833629608,0.288846314,-0.407065779,0.0210614,-0.0059787366,0.1246227473,0.1238094568,-0.2376222759,-0.0676046982,0.2496976554,-0.1606474966,0.0179347452,-0.0063020289,0.2356674522,-0.1279481351,0.4619129598,-0.0283046551,0.1844602376,0.1522844881,-0.1400704086,0.1942747086,-0.0534415692,0.0801875144,-0.1562188566,-0.2302116752,0.150962159,-0.0423329435,-0.0957653373,0.3187417686,0.1066991985,-0.4129915237,-0.2681288123,-0.2341492176,-0.1965827644,-0.0688151866,0.2436381429,-0.550116539,-0.1006577983,0.0060553774,-0.1188995913,-0.2579843998,-0.8107978702,-0.0886439681,0.1115385666,0.1418068111,-0.0186519045,-0.1254350841,-0.0019514477,-0.7301461697,0.0859234035,-0.2048450112,0.4765409827,0.015086744,-0.165925011,0.3607977331,-0.1390093118,0.3432276845,0.229436934,0.231170997,0.1268695295,-0.2358857691,-0.2924330533,-0.0200933572,-0.1341980398,0.2910181582,0.4701869786,0.3386704624,-0.2109950036,-0.1122326925,0.0331881233,0.2007946223,-0.0463350415,-0.5885095,-0.3338903189,-0.1042778641,0.1251729876,0.0498284437,-0.3312629163,0.1846339107,-0.1419173777,0.014819039,-0.0186878201,-0.0644956753,0.3712440133,-0.0144758048,-0.2155698687,0.0864332393,0.3805233836,-0.1101738811,0.3068843484,-0.5190032125,0.4141854942,0.2951724231,0.1443726718,-0.1044755876,0.0681107268,0.48499614,0.2805309594,0.0727250427,0.1074703187,-0.1483933479,0.1615279168,-0.2212729603,0.0046956353,0.1552949548,0.0559348986,-0.5429232121,-0.4145946801,0.8792427778,0.2034608126,0.0537672825,0.4225312471,-0.4378096461,0.0195140261,0.1285548508,-0.3023575842,0.8844261765,-0.4876252115,0.2667918205,-0.1427880228,0.1045815498,0.3689923882,-0.1400716305,0.0535977483,-0.4890018702,-0.0224748459,0.2861877084,-0.0741394162,-0.0718906149,0.3116643727,-0.026214214,0.4243609905,0.2834996283,-0.1549219787,-0.1281777769,0.6898174286,-0.1313003898,-0.1389394552,-0.600728631,0.0725087747,-0.2503744364,0.4260059297,0.0095727928,0.0590007305,-0.0897774249,-0.0092580384,-0.2213560939,0.1506018788,0.0860445127,-0.0383481421,-0.0656749234,-0.2873577476,0.1353088319,0.0469483137,-0.1573356092,0.2268480211,-0.2573876083,0.1667392254,-0.4367384911,0.1700863242,0.5596879721,0.0737887844,0.1408279985,-0.2069168538,-0.0311694909,-0.1966684461,0.0104995212,-0.3072821796,-0.1878343076,0.1687066704,-0.0427182429,0.0860314891,-0.3359212875,0.1928967685,-0.0220473278,0.0073892628,0.0282421857,0.4713738859,0.0229731351,0.4426311851,0.138444066,-0.4195038974,-0.0564388409,0.6509244442,0.1540878415,-0.2020782083,0.4344594479,0.1784682274,-0.2042498887,-0.1239009053,-0.0381312333,-0.053481929,-0.1030353904,0.0106898546,0.1027106494,0.1370224953,0.2356937826,-0.0274023227,-0.0967865959,-0.0574710481,0.0797141269,-0.4758243263,-0.2814335823,0.0782545134,-0.0346593671,-0.0560276844,-0.0708443299,0.0637029633,0.2404703349,0.382733494,-0.1732861251,0.2091056705,0.0490193367,0.1297856718,0.034028478,-0.2758023441,0.0635286048,0.175900504,0.1001752093,-0.0893801153,0.1463343799,-0.1821585745,0.0805238858,0.1911786944,-0.0182574484,0.0710047856,-0.0147511484,-0.249136731,-0.0985987335,-0.1257417649,-0.1691789031,0.0209515467,-0.0121735428,-0.0336073563,0.0476470366,0.0073601366,-0.0108680762,0.2618748248,-0.4449562728,0.2136015147,0.0884900317,-0.047475215,0.3296180069,0.0567104444,-0.5133446455,-0.1297393888,0.1511341631,0.029143272,0.2455508858,-0.1812510639,0.1469950378,0.0154703502,0.2825738192,0.5169383883,0.0923560709,-0.2505999804,0.2057130784,0.0472314283,-0.2324267626,-0.0325670578,0.3367089033,0.1220629364,-0.1604412794,0.1270446479,0.1400305033,0.0594025105,-0.3971574306,-0.3388273418,0.3963354826,-0.2949801683,-0.288862288,0.0447505303,0.1424228996,-0.2148334682,0.2856868207,0.1387457997,-0.0496772937,0.5338299274,0.2140261978,0.353611052,0.0886826739,0.5044412017,-0.4456042945,-0.2826678753,-0.194241792,0.6284957528,0.129330039,0.0893078744,-0.110012114,0.2932817638,0.016308954,0.0755492374,-0.1834852397,0.1059910059,-0.0417010561,-0.1042094231,0.2512892783,0.0551970638,-0.2106814981,0.2818473577,-0.2898305058,0.0781681985,0.0393487886,0.347394377,-0.0138222789,-0.3718743026,0.1553680301,0.0059544649,0.2821175158,-0.2320890427,0.0946820304,-0.3501392901,0.0913081393,0.0703391358,0.1336841583,0.2132326663,0.3437464535,-0.0388279036,0.061717771,-0.1379773617,-0.036578428,-0.239729926,0.5453297496,0.2327238172,0.4280001819,0.0116979284,-0.0574458987,-0.0723947659,0.1846771091,-0.1404439211,-0.1664669514,-0.058276929,0.2892153263,0.1981066763,-0.3058029413,-0.5087031126,0.0924151763,-0.2841328382,-0.3969328403,0.2010991573,-0.0269717537,-0.0412665047,-0.1026284397,0.0212468132,-0.2606879175,0.7240054011,0.4203976989,0.4147620499,-0.1415319443,-0.2133690715,-0.5053450465,-0.0611190461,-0.2168268561,-0.0413791835,0.0282436796,0.3009122908,0.0783494189,0.0319598131,-0.4840411246,0.2393326014,-0.1437499523,-0.0717962533,-0.2601799071,0.1210918203,-0.3646515906,0.1039534435,-0.1859138459,-0.1800041646,0.1082032397,-0.3293187916,-0.1053128615,-0.1933625937,0.0335705839,-0.0486839786,0.4685698152,0.3732513189,-0.0485884026,0.4595887363,0.0025365683,-0.0833146423,-0.3543555439,0.0255611856,-0.139668107,0.347266674,0.0696113259,-0.0716930479,-0.0028261342,0.4050496221,0.1833652556,0.2690808475,-0.3769444525,0.0071518463,0.0666602552,-0.3617617488,0.0437066853,0.3213766813,-0.0183436219,0.4127100408,-0.0525357053,-0.1152974293,-0.4713505805,0.0648956597,0.2717746794,-0.3392553926,-0.0139824729,-0.1822091937,0.3339839876,0.0270641632,0.2821818292,-0.4494110942,-0.0429609232,0.3149370849,0.0588434674,0.0364785381,0.4986689687,0.0495593734,-0.1199783087,-0.0335841849,0.022612825,-0.1671622843,-0.1060022414,-0.3730045259,-0.1559014469]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"I'm working on it today :)","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":6,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nI'm working on it today :)","embeddings":[-0.2098070532,-0.1893767416,-0.113343589,0.2468183488,-0.0262854192,0.0757775232,0.1170608774,0.3745322526,0.2817263603,-0.2054487914,0.1460434496,0.2481243312,-0.1347287595,0.2060510218,0.2234463692,0.14441441,-0.0241965633,0.2383002639,0.061371401,-0.167665258,0.0590009131,0.0084360261,-0.2544986606,-0.2704230249,-0.1512416005,-0.0178332217,-0.0153628662,0.0058342228,-0.1236878857,-0.2144337445,-0.0112812538,0.0760494396,0.0394974649,0.2516771555,-0.0001200161,-0.2742325366,0.4378759563,0.1163683608,-0.2017989159,-0.0746774971,0.2788412869,-0.5852707624,0.1812155247,-0.1189995855,-0.0257356036,-0.0232756138,0.0798631012,-0.1075416058,-0.0041389503,0.1263385117,0.1005879268,-0.1010209844,-0.2386603206,0.262165904,0.0556278229,0.2222626507,0.0208074935,0.3326944113,0.4490742683,-0.2129424065,-0.1926452965,0.0332611725,-0.1859663576,0.120190911,0.2755359113,0.0395931639,-0.1301336735,-0.0843460038,0.1561899632,0.3102736771,0.6140022278,-0.0900589079,0.0060042092,-0.4219771624,0.0546173267,-0.1753690392,0.1996639371,0.2796439528,-0.2141847014,-0.2940831482,-0.375416249,-0.3069724739,-0.0686154589,0.2325564623,0.0984311774,0.1251184195,-0.018151002,0.1386836618,0.2178703099,-0.0160284694,0.0087076677,-0.1008289084,0.2292426974,0.4415272176,-0.4179868996,-0.0657523274,0.0669129044,0.2529183328,0.0957408398,0.0816614255,0.070471257,0.038979657,-0.130032137,-0.0045284815,0.2867873907,0.3769712448,-0.1578958631,-0.1333108544,0.3634670079,-0.1974665374,-0.1964277029,0.2243645936,-0.1412547976,-0.1446076035,0.0773197338,-0.4156198204,-0.1715791076,-0.3830442131,-0.0599191412,-0.0850499868,0.0217893124,-0.2608987391,0.0659524053,0.2079597414,-0.3821007013,0.2496022433,0.0471194945,-0.1011523604,-0.3087279499,0.0834954679,-0.1267711967,0.0560515821,-0.1418029517,0.2036444843,0.5197031498,-0.0364491828,0.1447392404,0.1584238857,-0.2018302232,-0.1832618862,-0.2201519459,-0.3624844253,-0.1113837138,-0.03578775,-0.0305555146,0.2423553765,0.0230871532,0.2990721762,-0.3848541975,0.1741216928,-0.3638705015,-0.0939875841,0.0216194056,0.1248797551,-0.3923646212,-0.394733876,-0.2705399096,0.3753013909,-0.0406454243,-0.1865315288,-0.2639097869,-0.0909239873,-0.250361532,-0.0041788416,0.1197530478,0.187686339,-0.1207809597,0.0192685556,-0.0918100551,0.3721483052,0.3164002299,0.5578086972,-0.1813742965,0.266457051,-0.1738400757,0.0095365364,0.4789331555,0.0644591153,-0.4383442402,0.5403292775,-0.2256928682,0.0082729496,0.2308749408,0.4123356938,-0.0086284028,-0.0470253788,0.2735259235,0.6046065688,0.1983832121,0.2263249308,-0.4755065143,-0.1029583141,0.3067094088,0.4413370192,-0.0706171691,-0.0403977595,-0.0164076667,0.085239768,0.4993160367,0.1882215738,-0.1121634841,0.3725619912,-0.1432316899,0.1016244739,0.0538382679,0.1351826638,-0.0944543034,0.1391943395,0.189027071,0.0678952783,0.3592125177,0.1419591606,-0.229540363,-0.2463929057,0.0661270469,0.2134019732,-0.0346749611,-0.0496127196,0.025119599,0.1891441494,-0.1815380007,0.3669069409,-0.1746489257,-0.1781435311,-0.3642722964,-0.0405800752,0.2872920334,-0.1342209876,0.3709872067,0.2745962441,-0.1193276644,0.0176679529,-0.1126552522,-0.1123550087,-0.2146866769,0.2168321311,0.0553183109,0.0420257561,-0.0065958761,-0.0405090116,0.4049243033,0.2132036686,0.2860561609,-0.3814387023,-0.2571118474,0.3721030354,-0.217678085,0.2709050179,0.1141863167,-0.5126156807,0.0569842681,-0.129486829,0.1939579993,0.3392424881,0.9656714201,0.0456668213,0.6300940514,0.2854171395,-0.1924778372,-0.0010384654,0.4612407386,-0.0501372814,-0.163708657,0.4812838435,-0.1392501742,-0.3082069755,-0.0178891644,-0.1155175418,0.3902979493,0.2552175224,0.0363571495,0.0570871159,-0.0317927264,-0.2319685817,0.1577322781,-0.1610462368,0.2213312984,0.1611674726,0.3317386508,0.072955668,-0.4321333468,-0.2334658951,-0.022566421,0.3878911138,-0.0865401328,-0.0148838563,-0.0389268845,-0.3274970651,-0.207703948,0.0031670474,-0.2956582606,-0.2096622735,-0.0238746367,0.0736641735,0.4856424928,-0.1858491451,0.0393577628,0.0810066387,-0.1939608306,-0.5081317425,-0.3052713275,-0.1355334222,-0.3142745793,-0.0324282609,0.274527967,0.2505613565,0.3233252168,0.1425894797,-0.2181750536,0.2533923686,-0.0973753184,-0.2711670399,-0.0217919797,0.0079427259,-0.1830547303,0.1949309707,0.2241320908,0.0066699376,0.0605065785,-0.3427268565,-0.0099307476,0.1180213094,-0.0169266537,-0.0263429228,0.0103817508,-0.1915089339,-0.2911600173,-0.2170344442,0.4692827463,0.1007853374,0.1657878757,0.1899107248,-0.1947635412,0.2380948663,0.1045479551,0.226798147,-0.0202727839,-0.2011308819,0.3028926551,0.3215899765,-0.191694051,-0.2206761688,0.0572917052,0.131881088,0.0054171411,-0.7303994298,0.3203904629,-0.4778620303,-0.0439355709,-0.0965827405,0.1214755103,0.1433648914,-0.2556490004,-0.0355197303,0.2620173395,-0.1638049781,-0.0342946164,-0.010903554,0.2172041535,-0.1644603759,0.4016971886,-0.1255955547,0.0320092589,0.1228725538,-0.1266494095,0.176912412,-0.0605884418,0.0956800058,-0.173530519,-0.2571721673,0.2069671005,-0.0855906382,-0.1333476156,0.301848799,0.0581423491,-0.4478368163,-0.3004540205,-0.1842099875,-0.0725727826,-0.0466597676,0.1685478389,-0.3596469164,-0.1433980316,0.0026167538,-0.1110582203,-0.237141937,-0.7687709928,-0.1511241496,0.1655917615,0.1081735566,0.001464834,-0.1475750357,-0.0249993354,-0.7787162066,0.0981520638,-0.2656366229,0.4183078706,0.0329815112,-0.1077656299,0.3655878305,-0.1556669027,0.3255587518,0.2863793373,0.1948199421,0.1694852114,-0.1598326415,-0.4076752961,-0.0259759892,-0.1253756881,0.2234366834,0.5260003209,0.353985399,-0.1578964293,-0.1078254059,0.0526447743,0.2960776091,-0.0549800284,-0.6808643937,-0.3612894416,-0.1574700922,0.1035207286,0.0703474581,-0.2828344405,0.1640522331,-0.1810534596,0.0081304545,-0.0192600898,-0.0591020025,0.2874442637,-0.0436985977,-0.2140493691,0.0804895684,0.3570564389,-0.1244407967,0.2762258351,-0.4865377843,0.3951712847,0.2636784017,0.1463648826,-0.0683527663,0.0342716575,0.4541308582,0.2233950496,0.0810940936,0.0921797007,-0.2465676218,0.2296372056,-0.2370931804,0.0421344079,0.1452109367,0.0572675429,-0.435898006,-0.4449151158,0.8915281892,0.1878013313,0.0680455267,0.4211841524,-0.4836359322,0.0357966796,0.1167383716,-0.3481820822,0.8028689027,-0.4478403032,0.2572024167,-0.1275613755,0.0433622189,0.3825463951,-0.143786788,0.0227239858,-0.4348393083,0.0426642224,0.2968159914,-0.0333771855,-0.1097169369,0.2699379027,-0.1399709284,0.4599863291,0.284915477,-0.292106837,-0.1518308669,0.7412165403,-0.1335546523,-0.1520693153,-0.6122818589,0.0824930966,-0.3149623275,0.4781811833,0.011763881,0.1157456413,-0.0905005783,-0.0099272365,-0.2647446096,0.1292331666,-0.0143130915,-0.1058596298,-0.104117915,-0.3387745917,0.1879661828,0.1769638658,-0.2429625243,0.2280129045,-0.2703282535,0.143042922,-0.3931135833,0.1809400022,0.4963358641,0.0009661404,0.0641657859,-0.2031046748,-0.054290764,-0.1872370392,-0.0195037127,-0.1797208637,-0.1137329489,0.1796028465,0.0184304267,0.088445507,-0.3176743686,0.2540959716,0.0361604132,0.0058600316,0.0410415605,0.4165132344,0.0538435914,0.3842563629,0.0479822867,-0.4513024986,-0.0395936444,0.5395545959,0.1217087507,-0.1459848285,0.46058622,0.230971545,-0.1452953964,-0.150770992,0.0093044871,-0.0773529708,-0.1405992061,0.0291640479,0.126384303,0.1079675481,0.2421905845,-0.0150626209,-0.1432790607,-0.1289118528,0.0818041414,-0.4670028985,-0.3291833997,-0.0196267851,-0.0225507505,-0.086156927,0.0031834806,0.0001066097,0.2867251337,0.3378161788,-0.1679943353,0.2015372813,0.0312243868,0.0945095047,0.0571008138,-0.2799992859,0.055386018,0.1939717531,0.100555338,-0.1244738623,0.2145989835,-0.2028178871,0.0425761007,0.1768122762,-0.040383134,0.0893462598,-0.0179558918,-0.320666492,-0.0656435266,-0.1436055601,-0.1465188265,-0.0102272592,-0.0139074074,-0.0317225531,0.0413852409,-0.0422095805,-0.0521733351,0.2796012759,-0.4064957798,0.179345876,0.0939539671,-0.0049920846,0.3704088926,0.0597345084,-0.5080822706,-0.1573055834,0.1321980059,0.0726969838,0.1698604673,-0.1766264141,0.1661752611,-0.0170798972,0.2522969842,0.5463634133,0.0730456337,-0.2292596102,0.1549790204,0.0638316795,-0.2190780193,0.0226387773,0.2881912887,0.1615083814,-0.1306631118,0.1573886573,0.0634807199,0.0630052686,-0.3247476816,-0.2944059074,0.4152236879,-0.1919692159,-0.2931559086,0.0538345948,0.1764658242,-0.1903890073,0.2280211002,0.1371686608,-0.0384180434,0.4552154839,0.2348764092,0.2986885011,0.0860926583,0.4523495138,-0.3968259096,-0.210934937,-0.1973252296,0.5488992333,0.0790239051,0.0954906642,-0.1007674485,0.2776524723,-0.0599859357,0.030911129,-0.1595167518,0.0613578185,-0.0253112335,-0.0722122788,0.2184247971,0.0980783328,-0.1762621403,0.2171435952,-0.2743106484,0.0813130289,-0.1001011729,0.3290943205,-0.0401935317,-0.4175705016,0.0833957493,0.0355694331,0.2196297795,-0.2002089024,0.1089139804,-0.3450181484,0.1050757542,0.0972498804,0.1603935957,0.2714065611,0.3865174055,-0.0512234569,0.1216503382,-0.1435828209,-0.0518962033,-0.225732103,0.4070631564,0.2187854946,0.3817570508,0.073944591,-0.0461466312,-0.0790976882,0.218038395,-0.15374358,-0.2063079029,-0.1114474386,0.2788775265,0.2117234021,-0.31155321,-0.6116627455,0.0983043388,-0.3025065362,-0.222488299,0.1431492567,-0.0557410829,-0.0467907563,-0.0728458241,0.0198447853,-0.3255864978,0.6861280799,0.4153162837,0.4743106067,-0.1483798772,-0.2850852013,-0.5462821722,-0.1125694811,-0.2654057741,-0.1604689956,0.0840704963,0.2368223518,0.0313297138,0.0340133645,-0.3972458243,0.2277527004,-0.1215414926,-0.1078157946,-0.3063416183,0.0153031154,-0.329115659,0.1270498782,-0.1299075186,-0.2076819241,0.0782118216,-0.2764892876,-0.1064276025,-0.1683818102,0.0965333357,-0.0417372361,0.5272362828,0.278573066,-0.0798405185,0.4496697783,0.0107723773,-0.0453968346,-0.3342131078,0.1106920689,-0.1908749342,0.2570905685,0.1343574077,-0.0862838924,-0.0524103381,0.33993572,0.1672205776,0.273142606,-0.2656617761,0.002923727,0.0692944676,-0.3554351926,0.0256454721,0.3276981115,0.0054620756,0.4216554463,-0.0394147225,-0.0911680907,-0.4780692458,0.0759683475,0.3117601275,-0.2633885145,-0.0499419086,-0.225554347,0.3488414884,0.0019851748,0.2699088454,-0.3635504842,-0.0128209135,0.3278622329,0.1145427823,0.0332036503,0.4925608635,0.0243919007,-0.0954440013,-0.011889928,0.189648211,-0.1009322926,-0.096513994,-0.3658975661,-0.1802264452]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"Ok so now the text files won't be hashed.\r\n\r\nI also updated #548 to include this change.\r\nLet us know if it helps @agemagician :)","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":25,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nOk so now the text files won't be hashed.\r\n\r\nI also updated #548 to include this change.\r\nLet us know if it helps @agemagician :)","embeddings":[-0.1468623132,-0.1677474678,-0.1159891412,0.2507619858,0.0116750384,0.1286732107,0.2028606385,0.4360566139,0.2853423357,-0.2157860845,0.1411366612,0.1851586252,-0.1777909547,0.1772634238,0.2101206928,0.1729486883,-0.0373955444,0.2307506204,0.0664635375,-0.1615090966,0.0738641471,0.0152549492,-0.2379418314,-0.2696356773,-0.1859089434,0.008016401,0.0108226752,0.0459067747,-0.1581698656,-0.2418571413,-0.040229924,0.1110756323,-0.0026757144,0.2269131988,-0.0001188803,-0.2230950445,0.4324368834,0.0995655954,-0.1644918025,-0.078722775,0.2473833114,-0.5807968378,0.1536241919,-0.1115248278,-0.0279732812,-0.0242866836,0.0849890262,-0.0908508673,0.0091997487,0.0936385319,0.0956710428,-0.15014714,-0.2433137596,0.2775405347,0.1566742361,0.1584055722,0.0178573802,0.3768332303,0.384021312,-0.1982916445,-0.2379698902,0.0691012144,-0.204436332,0.0671158209,0.2822948098,0.0695497766,-0.0627496392,-0.1067551076,0.0847101435,0.2903830409,0.5367581844,-0.0819913149,-0.0145144658,-0.4157767892,-0.0333935805,-0.1956962943,0.2631214261,0.2212334722,-0.1917852759,-0.2598748803,-0.426959604,-0.2712003589,-0.0346584283,0.1871253699,0.1697664857,0.0125535596,0.0180831663,0.1133252084,0.2346084565,-0.0057958774,-0.0303427596,-0.1031939611,0.2266772091,0.4136863053,-0.3679282665,-0.0799880922,0.0814782083,0.2051116526,0.0756677389,0.1108645052,0.0774825737,0.1088117063,-0.1353557855,0.000163895,0.2267829329,0.3041849732,-0.1098384783,-0.1708535403,0.3388783932,-0.1978961229,-0.239296183,0.2261637598,-0.1341163218,-0.1949007958,0.0409322158,-0.3911283016,-0.218679592,-0.3882549703,-0.0161750708,-0.1021418571,-0.0117927101,-0.2803953886,0.0991439447,0.2328170687,-0.3714700639,0.1882843673,0.0256628543,-0.1213439927,-0.3319734931,0.1176175922,-0.1348504871,0.0648253337,-0.2003657669,0.2102786452,0.5573926568,-0.1081117839,0.1549529582,0.1407556832,-0.1908708513,-0.1235960722,-0.2451802045,-0.3195974529,-0.051716309,-0.0391326062,-0.060400635,0.2393612266,0.0170847718,0.2906760275,-0.3809715211,0.1462773234,-0.3813503087,-0.0917466953,0.0715365335,0.125502184,-0.4116356969,-0.3418392241,-0.1967470944,0.3686625957,-0.0151023939,-0.229745239,-0.2660705149,-0.070772402,-0.2337767333,-0.0221741013,0.1182159856,0.2390013635,-0.1117661893,0.0041146092,-0.070294939,0.3350042701,0.2934179306,0.5889776945,-0.1977968216,0.2587456405,-0.2076937258,0.0664704442,0.4533969164,0.0793929473,-0.4505611062,0.5086809993,-0.2272730172,0.0262120198,0.2360358685,0.4096684158,0.0190483537,-0.0865272582,0.330994308,0.5486661792,0.2094087154,0.1983354837,-0.5275840759,-0.1045790389,0.3166627884,0.4256461859,-0.0732965469,-0.0420415327,0.0342000425,0.1110450551,0.454603523,0.1974325329,-0.0950349644,0.3816902041,-0.1414045841,0.0289914031,0.0420624726,0.1356700212,-0.0766887665,0.1253725439,0.1800880879,0.1044865474,0.2555927932,0.1586179137,-0.1795031428,-0.2985091209,0.0688033924,0.1966096312,-0.041799482,-0.0983957499,-0.0063333763,0.1894557476,-0.175567314,0.3249536753,-0.226368323,-0.1705373824,-0.3540986478,-0.05081366,0.2618776858,-0.1658668667,0.34876436,0.2691206634,-0.128028214,-0.0214635041,-0.0956853256,-0.1553688943,-0.262642175,0.2162464708,0.0798505247,0.1474854499,-0.0038794365,-0.0686583519,0.4277429879,0.1416683793,0.2356412411,-0.3945387304,-0.2611134946,0.4239822328,-0.2399422824,0.1925337315,0.100040786,-0.5276620984,0.0952800363,-0.1499476582,0.1792489141,0.3012718856,0.9152841568,0.0748972222,0.5796502233,0.3361260593,-0.0965634063,0.0237046275,0.5802968144,-0.0135958046,-0.2307458073,0.5299873948,-0.1376429349,-0.3367550671,-0.0907449573,-0.1596767306,0.3672909141,0.2584278584,0.0612425245,0.0831917599,-0.0067600305,-0.2745862305,0.1860840917,-0.1748491526,0.1555586904,0.1842088103,0.3061496019,0.0672388375,-0.4513900876,-0.1900628209,0.0111349635,0.3954620659,-0.1219163463,-0.0600742809,-0.0622061528,-0.2860005498,-0.1999253035,0.0116269998,-0.3100402355,-0.2185954452,0.039121449,0.0945368037,0.443469137,-0.1405885667,0.0237838235,0.0919803455,-0.1460936666,-0.5046122074,-0.3176472485,-0.1240732446,-0.3057529628,-0.0356729403,0.3436005116,0.2066265643,0.2857854664,0.1580767035,-0.2393473834,0.2140695602,-0.1337425709,-0.2464983314,-0.0655812323,0.0074957688,-0.1984963864,0.1913417727,0.1513044536,-0.0254358482,0.0704560801,-0.3434503078,-0.0498985499,0.1691686511,0.0141745796,0.0251855552,-0.0353945158,-0.2966687977,-0.2575834095,-0.1942804903,0.5056717396,0.1282690614,0.1918306798,0.2333414853,-0.1564410627,0.2238167226,0.0130965756,0.196070835,-0.0103010377,-0.3355054259,0.3539385796,0.3149785995,-0.1909895241,-0.184513256,0.0784603804,0.0577202477,0.0648786202,-0.7292475104,0.3334608078,-0.424723357,-0.0589508042,-0.0232272297,0.0763256326,0.1828999817,-0.2181565166,-0.0653541088,0.2649844885,-0.1422681659,0.0547602102,-0.041454345,0.2459864169,-0.1490635127,0.4429358244,-0.0416023247,0.1310745329,0.1429376751,-0.1276002228,0.1752118766,-0.0469571017,0.0878356099,-0.2085817307,-0.255073607,0.1639944464,-0.0698838159,-0.1226182058,0.3286094069,0.1117333397,-0.4102245569,-0.257324636,-0.188120082,-0.1360911727,-0.091328539,0.2339009196,-0.4632798135,-0.1695217639,-0.0002382079,-0.1234507114,-0.2706368864,-0.8070899248,-0.0899768919,0.1939152777,0.113490276,-0.0124874366,-0.1416276693,0.0572747402,-0.7214740515,0.0916469172,-0.2255845517,0.4330693781,0.0095027471,-0.1530495733,0.3658154905,-0.1416426748,0.3453766704,0.2615406215,0.2158099711,0.164853543,-0.1457090378,-0.3058147132,-0.0606587417,-0.1359806806,0.2373658419,0.524466455,0.3955211341,-0.1596056074,-0.1255625486,0.03219622,0.2663289607,-0.0455251709,-0.6147637963,-0.340033561,-0.123643443,0.0839363262,0.0548790731,-0.2561672032,0.164159283,-0.2225336134,0.0019965328,0.0160446987,-0.0569870137,0.3316158652,-0.0649500936,-0.197247088,0.0483908504,0.382894218,-0.1095412225,0.2603668571,-0.5135046244,0.4005385339,0.2878218293,0.1498518437,-0.098192364,0.0158507023,0.4233320653,0.2882662416,0.056611944,0.1032960936,-0.1901738942,0.2038579583,-0.2070806175,0.0365809947,0.1552279741,0.0373079702,-0.4535947442,-0.4440155029,0.8658829927,0.2004580349,0.052307874,0.4413743317,-0.4890833199,0.0076253279,0.1646187305,-0.3255107105,0.8962662816,-0.5091392994,0.2615803182,-0.1717269421,0.1543432325,0.3635105193,-0.2261977643,0.0469628237,-0.4782652855,-0.0128275976,0.2797729969,-0.0558130816,-0.1093785763,0.2966302037,-0.0840289518,0.4026567638,0.3037196398,-0.1816541851,-0.1350027919,0.7126305103,-0.1246581003,-0.1495569795,-0.5936140418,0.0717839673,-0.249015972,0.4504694641,0.0290423222,0.0551492348,-0.1124364585,-0.0363451503,-0.2671859562,0.1421435475,0.0412773229,-0.0683745071,-0.1079094112,-0.2744807303,0.161703825,0.1341018826,-0.1970690042,0.245426923,-0.2713269591,0.1479254961,-0.3887501657,0.1796724647,0.5466579199,0.0269680116,0.1233063415,-0.2026598454,-0.0371569693,-0.2263115048,-0.0037361865,-0.2732928693,-0.1686253846,0.1459264457,-0.0254487321,0.0775291398,-0.3267630041,0.2190438956,-0.01498463,0.019492656,0.0393402725,0.4278938174,0.0573043115,0.4107716084,0.0781801343,-0.4328205585,-0.0341705866,0.6329444051,0.1193458959,-0.171577543,0.4617879391,0.2462400496,-0.1851257533,-0.1439626068,-0.0316445008,-0.0844542533,-0.1185495034,0.0457433164,0.1380202621,0.1327204406,0.2338372916,-0.0158675201,-0.1340325326,-0.1210011169,0.0679353923,-0.4891684055,-0.332213968,0.007664307,-0.0354981534,-0.0495295674,-0.0455133058,0.0394646674,0.2536679208,0.2950505316,-0.1840224713,0.2337615788,0.010318852,0.1074576229,0.0471791998,-0.2894695699,0.0775341392,0.1480062306,0.0994741097,-0.088141337,0.1734413058,-0.1998885423,0.0783429518,0.1771930903,-0.0222311355,0.0684999749,-0.0260577109,-0.2911944985,-0.090316698,-0.1383881867,-0.1463042945,0.0333831683,-0.0204251669,-0.0187389031,-0.0062021152,0.0183193274,-0.017955143,0.2220660448,-0.4845607281,0.2126890421,0.1188967749,-0.0434448309,0.3119729459,0.068291977,-0.501971662,-0.1139496118,0.1637864858,0.1204588488,0.1987317801,-0.171328634,0.1428079903,-0.0002148531,0.2472883016,0.5397869349,0.0465881675,-0.2343889922,0.2082855403,0.0702079982,-0.2324064821,0.0128430426,0.3182410598,0.12216416,-0.1434263587,0.0990469977,0.106835112,0.0493921414,-0.3408548832,-0.2861553133,0.4140724838,-0.2528637052,-0.2983031869,0.0619954988,0.1794878691,-0.2136211991,0.2452967167,0.1134573221,-0.0485684164,0.5281426311,0.1692796946,0.2819338143,0.0298364945,0.4990845025,-0.4591498673,-0.2564749718,-0.1732317507,0.5806372762,0.1013704836,0.08646819,-0.1049983948,0.321266681,0.0133718085,0.066419892,-0.1577326506,0.1102186963,-0.0187561437,-0.0704561248,0.2236236781,0.0472387709,-0.206590414,0.2458727658,-0.2708841562,0.1115668118,-0.0127021726,0.3698680401,-0.038898278,-0.3657515943,0.0976007059,-0.0232986733,0.2659726143,-0.21682778,0.0947543383,-0.3053723276,0.1046342552,0.1304863989,0.1334237158,0.2483450472,0.3836371899,-0.0247401819,0.1180631295,-0.1708301008,-0.0525136329,-0.2167859077,0.4916284382,0.2272069454,0.4188089669,0.0320710093,-0.0444031395,-0.0905760601,0.1935408711,-0.1724515259,-0.1869644225,-0.0721828043,0.294770658,0.2757712305,-0.3323739171,-0.5637335181,0.0603600815,-0.3046702147,-0.3551775217,0.218863368,-0.011657726,-0.0033615704,-0.0865995362,0.0295070596,-0.2765630782,0.6839308143,0.4079569578,0.3934640586,-0.1450067014,-0.1829031855,-0.5345113873,-0.1190320998,-0.2366613895,-0.0726826116,0.0406126939,0.2603267729,0.0788448229,0.029218873,-0.4430757165,0.2248760164,-0.1540221572,-0.0911423191,-0.3024024665,0.1116739661,-0.343049854,0.1367995441,-0.1714632511,-0.1733539999,0.1039347127,-0.2731744647,-0.0943435878,-0.1843815744,0.0913889706,-0.0099158613,0.4788247049,0.3028303683,-0.0568209067,0.4388084114,-0.0038316105,-0.0684952587,-0.3777605891,0.1048204154,-0.1790893227,0.2524610162,0.1182070673,-0.0653510317,-0.0488358028,0.324619472,0.1262114197,0.2831217647,-0.3425390422,0.0193986874,0.0338451713,-0.355753392,0.0484087802,0.3573252261,-0.0367289782,0.4229919314,-0.0241790656,-0.1041419059,-0.4996803403,0.0355128646,0.3133938611,-0.336089313,-0.0400696285,-0.1442221254,0.3503611386,0.0657299012,0.3043915033,-0.4382010102,-0.0326372385,0.3086366057,0.0936469659,0.0248136707,0.4935049415,0.0502892211,-0.0830160156,-0.0098908562,0.0670778155,-0.0936904773,-0.0860048085,-0.3788420856,-0.1776272357]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/546","title":"Very slow data loading on large dataset","comments":"Perfect thanks for your amazing work.","body":"I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?","comment_length":6,"text":"Very slow data loading on large dataset\nI made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data.\r\nIt has been 8 hours and still, it is on the loading steps.\r\nIt does work when the text dataset size is small about 1 GB, but it doesn't scale.\r\nIt also uses a single thread during the data loading step.\r\n\r\n```\r\ntrain_files = glob.glob(\"xxx\/*.txt\",recursive=True)\r\nrandom.shuffle(train_files)\r\n\r\nprint(train_files)\r\n\r\ndataset = nlp.load_dataset('text', \r\n data_files=train_files,\r\n name=\"customDataset\",\r\n version=\"1.0.0\",\r\n cache_dir=\"xxx\/nlp\")\r\n```\r\n\r\nIs there something that I am missing ?\nPerfect thanks for your amazing work.","embeddings":[-0.2064628303,-0.191352576,-0.1161669418,0.2584391832,-0.0166993514,0.0694551691,0.1227240413,0.3733961284,0.277638495,-0.2095820755,0.1358122528,0.2592593133,-0.1206243634,0.1935770065,0.2182747871,0.1384231746,-0.0271036793,0.2464279383,0.042670615,-0.1855962276,0.0708172992,0.0068790377,-0.2594532073,-0.2643780112,-0.1522012949,-0.0241370983,-0.0093098748,-0.0021653678,-0.1319070011,-0.2185436636,-0.0120398365,0.0722602829,0.0323081203,0.2687521875,-0.0001201538,-0.2754677534,0.4317723513,0.1101457179,-0.2094584554,-0.0653743371,0.265617311,-0.5675516725,0.1750622988,-0.1118939072,-0.0366132781,-0.0363442376,0.087481916,-0.1074082553,0.0000491529,0.116626583,0.098656565,-0.0950340852,-0.232998848,0.2596722841,0.0443578847,0.2215188593,0.0145913698,0.3317924738,0.4489250481,-0.2127762437,-0.20530653,0.0369005948,-0.1930972785,0.1105765477,0.2802088857,0.0238176547,-0.1257042736,-0.0892179087,0.1556457132,0.309625119,0.6144813299,-0.0750001669,0.0055949525,-0.4087952375,0.0390406586,-0.1719481945,0.2035648376,0.2900739312,-0.2112865001,-0.2875551581,-0.3922846019,-0.2995085716,-0.083483696,0.2278725207,0.0845008418,0.1186696514,-0.0207585152,0.1399671882,0.1877581775,-0.0146750277,0.0072085904,-0.1221777126,0.242464155,0.4261233509,-0.4225408137,-0.0898995325,0.0612629578,0.2489886582,0.0980462506,0.0847832039,0.0684842318,0.0461013317,-0.1279117018,-0.0003683823,0.2870486081,0.3781379759,-0.1434829384,-0.1271819621,0.365924716,-0.1858035326,-0.2054266781,0.23506625,-0.1375425309,-0.1246394515,0.0513974577,-0.4164976478,-0.181899026,-0.3800926507,-0.0651322752,-0.0909197107,0.0130475573,-0.2702613175,0.0655763,0.2181974649,-0.3839210272,0.2502197623,0.0446670279,-0.1034948155,-0.3093996048,0.0787635744,-0.132761687,0.0576256514,-0.1399493217,0.2081627995,0.5051515102,-0.0375990644,0.1512376368,0.1518910527,-0.1967155337,-0.1858692765,-0.2169855982,-0.3718828857,-0.1069850698,-0.044450406,-0.0281949583,0.2323598415,0.0208683666,0.3179561198,-0.3754476905,0.1769740582,-0.3685427308,-0.0838963017,0.0220656618,0.1250694096,-0.3910247684,-0.3990053236,-0.2742531002,0.3662090898,-0.0562002175,-0.1692617089,-0.2579728067,-0.0897923857,-0.2547347844,-0.0041089617,0.1076058969,0.1818695217,-0.1181416288,0.0106083453,-0.0677141845,0.3591398895,0.3241712451,0.5574554801,-0.1863472462,0.2728256881,-0.1752620339,0.0091050547,0.4866777956,0.0829760432,-0.4417930543,0.5455459356,-0.2194483876,-0.0028435949,0.2277935445,0.4081596732,-0.0008754987,-0.0429577902,0.2643415928,0.6023539305,0.2153676897,0.2257461846,-0.4878566563,-0.1062959358,0.3043478131,0.4239815176,-0.0688364431,-0.0343533307,0.0015620632,0.0795565397,0.5111004114,0.1921368688,-0.1154027432,0.3828218579,-0.1355111897,0.0938008726,0.0537529625,0.145175606,-0.0752621591,0.1439171135,0.1857670695,0.0709754974,0.3642826378,0.1555156857,-0.2276854366,-0.235580653,0.0621323958,0.2167879641,-0.0350674652,-0.0493195541,0.0391655453,0.1868165135,-0.173986733,0.3615138531,-0.1930289268,-0.1738186181,-0.3660267293,-0.021476943,0.3063131273,-0.1251061112,0.3712162673,0.2631581128,-0.1138974428,0.0209255479,-0.1069299579,-0.1212789416,-0.2221746445,0.233674109,0.0789354071,0.0400482602,-0.0047038198,-0.0481526367,0.4212553501,0.2092886716,0.3031222224,-0.3683668375,-0.2481105775,0.3731607795,-0.2025850862,0.2553523481,0.122979112,-0.52782166,0.0661758855,-0.1304350346,0.2004536986,0.3381605744,0.9652521014,0.0636015534,0.6257626414,0.2902112603,-0.189384982,0.0040102387,0.4691151083,-0.0487566851,-0.1561121643,0.4776482582,-0.1420037001,-0.3130761385,-0.0172846895,-0.1247274131,0.3959932625,0.2597151101,0.0544795431,0.062444523,-0.013506772,-0.2285393029,0.1708576232,-0.1525458843,0.2292471975,0.1671680957,0.3377499282,0.0688356459,-0.4404555559,-0.2154637128,-0.016360702,0.3812622726,-0.0752003565,-0.0163315386,-0.0316186734,-0.3333665133,-0.2110594362,-0.0008296487,-0.3015737534,-0.2051253766,-0.0146287624,0.067433089,0.4984069169,-0.1755091697,0.0339419283,0.0879183263,-0.1963795573,-0.497081399,-0.3022202253,-0.1287800372,-0.3219399452,-0.032224793,0.2718434632,0.2464666069,0.3294966817,0.1526928246,-0.2276367545,0.2478709519,-0.0929404199,-0.2565830946,-0.0203611925,0.0057101962,-0.195731312,0.2047291398,0.2184639871,0.0228784308,0.0512853712,-0.3602461219,-0.0164382979,0.1010769755,-0.0197798461,-0.0177017525,0.0117470762,-0.1965187937,-0.2843403816,-0.2083163112,0.4595793784,0.0998379588,0.1667434424,0.1743380278,-0.2117479891,0.233548075,0.0886028334,0.2258562446,-0.0302584302,-0.2114747167,0.2963692248,0.3256415725,-0.1863032281,-0.2155823261,0.0611813478,0.1278829575,-0.0006413636,-0.743596375,0.317579776,-0.476834625,-0.0534886494,-0.0955720767,0.1247513741,0.1486904174,-0.2628227472,-0.0381240211,0.2550986111,-0.172903195,-0.0446070433,-0.0048791072,0.2243495882,-0.162244767,0.4023034275,-0.112766318,0.0493468568,0.1407847404,-0.1338742226,0.1805111766,-0.048480913,0.0868628472,-0.1794621646,-0.253467083,0.2013685107,-0.0784404501,-0.1278539598,0.3122014403,0.06096825,-0.459828645,-0.3008589745,-0.2008235306,-0.0769508481,-0.0372448936,0.1563156545,-0.3443288207,-0.1422681063,0.0066039083,-0.1126352698,-0.2343132943,-0.7801941037,-0.148019895,0.1600371003,0.1125817448,-0.0083014276,-0.1504102498,-0.0132112391,-0.7703263164,0.1091150343,-0.2647887766,0.4115989506,0.0373066142,-0.113067843,0.3877170682,-0.1512145102,0.34461537,0.2999116182,0.2056921422,0.1474537551,-0.151949659,-0.3923360705,-0.0222849213,-0.1228467077,0.2193900943,0.5247529149,0.3656102419,-0.1424369514,-0.1183733866,0.0536124967,0.3070941269,-0.0524015129,-0.6875514388,-0.3652831316,-0.1508550495,0.1094107553,0.0704938993,-0.2956984639,0.15960446,-0.1940315813,-0.0047353469,-0.0233209226,-0.0642552897,0.2793583572,-0.0439920574,-0.2117697895,0.0831487253,0.3607195318,-0.1277569234,0.2866908014,-0.4680342376,0.3948204815,0.2614140511,0.1467720568,-0.0700745583,0.033893168,0.4634920061,0.2501182258,0.0796178877,0.0967945457,-0.2493149191,0.2269737422,-0.235381186,0.0233110078,0.1442375481,0.060453441,-0.4236147702,-0.4395122826,0.8902724981,0.1811902672,0.0623181611,0.4138205647,-0.480238378,0.0377800688,0.1142240912,-0.3401150107,0.8368625045,-0.4432654977,0.2591164708,-0.1400171369,0.0483906604,0.3820872605,-0.1390073001,0.0342390426,-0.4515461922,0.0333423838,0.2900369465,-0.0402598791,-0.1052457392,0.2715468705,-0.1172636002,0.4659103751,0.2807489634,-0.3035242856,-0.1498831511,0.7195215821,-0.1341894567,-0.1395349503,-0.6194486022,0.0764224455,-0.3083485961,0.482448101,0.0068099033,0.1065992191,-0.0870498717,-0.0087678712,-0.2511349916,0.1276197731,-0.0258046743,-0.1027815267,-0.101699546,-0.3439658284,0.1902310997,0.171478197,-0.2547635734,0.2254244983,-0.2585760355,0.1459002495,-0.390473783,0.1918963343,0.4995245337,0.0087637855,0.0664011687,-0.2146486491,-0.0609343201,-0.1872716695,-0.0148167377,-0.1918154508,-0.1141339689,0.1834645122,0.0337257385,0.0938417539,-0.3134108186,0.2523222864,0.0440040603,0.0136085879,0.0403702259,0.4243261814,0.0501849912,0.3822975457,0.0569057353,-0.4591709077,-0.041731637,0.5529173613,0.1220748425,-0.1444573551,0.4646651149,0.2205794454,-0.1569383293,-0.1496025026,0.0140058435,-0.0909275189,-0.1398652941,0.0169717968,0.1331249326,0.1355783641,0.2341016233,-0.0165840276,-0.1585317105,-0.1221946925,0.0776385665,-0.4789117575,-0.3148927689,-0.021592252,-0.0395429358,-0.0992661715,0.0102130519,-0.0187297929,0.2890526354,0.3210760355,-0.1713255048,0.2040016055,0.0325825438,0.0890282467,0.0507720485,-0.2775058746,0.050678093,0.1893563569,0.0980357006,-0.1321193129,0.2135874182,-0.2048989683,0.0590254739,0.1803741902,-0.0431040972,0.0905844346,-0.0247171465,-0.3258624971,-0.0739968419,-0.1366931945,-0.1403192133,-0.0154805165,-0.01115721,-0.0265323296,0.0315493532,-0.0397341363,-0.0592306964,0.2786622643,-0.4045325816,0.1877349615,0.0925105065,-0.0058284509,0.3725189865,0.0494416095,-0.5098102689,-0.1651767194,0.1253519654,0.0901752487,0.1678463519,-0.1826974899,0.1556201279,-0.0161207356,0.2403425723,0.5511682034,0.0618132614,-0.2310562879,0.1535741389,0.0622787215,-0.2256807983,0.0080805821,0.2881081998,0.1601737142,-0.1310749501,0.1656140834,0.0789514259,0.0528526679,-0.3153249323,-0.3029812574,0.4106685817,-0.201972276,-0.2914353609,0.0570095405,0.1789778322,-0.1870132536,0.2215230912,0.1601164043,-0.0457926393,0.4443448186,0.2455278337,0.3096632957,0.0872968063,0.4421386421,-0.3956040144,-0.2177019864,-0.1913025677,0.5468568206,0.0612315834,0.0892824158,-0.1022688299,0.2711954415,-0.0430963412,0.0265310872,-0.1645523906,0.0514973216,-0.0290644988,-0.0676097944,0.2003665566,0.1010733545,-0.1695903242,0.2238950133,-0.2845290601,0.0803637877,-0.1069702506,0.331656009,-0.0543398447,-0.4111993015,0.0762251467,0.0444639474,0.2213985324,-0.1953843087,0.1019844264,-0.3446280062,0.1163390875,0.0851747543,0.1644843519,0.2667880356,0.3782955706,-0.0437734649,0.1145027131,-0.1458956897,-0.0496128201,-0.2269505858,0.4177313745,0.2095606625,0.3627350032,0.0654634908,-0.0501536503,-0.0740136728,0.2251682729,-0.1489551067,-0.1787031889,-0.1153090969,0.2779479623,0.2091939747,-0.3128177226,-0.6102100015,0.0925913826,-0.301151216,-0.2116357386,0.1579613835,-0.0606069937,-0.0398115888,-0.083304666,0.0225803629,-0.3347791135,0.6765967011,0.4208290875,0.4656966031,-0.1368144602,-0.2752452493,-0.526186645,-0.1093009785,-0.2534067929,-0.1427280158,0.0700333565,0.2333860695,0.0469843335,0.0231817774,-0.4038151205,0.2104362249,-0.1211484522,-0.1006539986,-0.3098424971,0.0135060521,-0.3242448866,0.1387134194,-0.1441217661,-0.2085773349,0.0802914351,-0.2874164283,-0.1106137633,-0.1579367965,0.1118856892,-0.0435148552,0.5238573551,0.2668916583,-0.0683869421,0.4558729529,0.0049808444,-0.0348179229,-0.3222616315,0.1173090488,-0.196580708,0.2605994642,0.1420747638,-0.101772204,-0.0483227372,0.3244952559,0.1827436537,0.269146353,-0.26703462,-0.0045680297,0.0809299275,-0.3678593934,0.0271077994,0.3268606365,0.0068257586,0.4156354666,-0.0424332134,-0.1212046221,-0.4768878818,0.0792683139,0.3031967878,-0.2803261876,-0.0627861694,-0.2345437557,0.3547855616,0.009086513,0.2646028101,-0.3561140299,-0.0145371696,0.3365519047,0.1021107212,0.0432854071,0.5073423982,0.0272974763,-0.0846960619,-0.0080113485,0.1900116354,-0.1091895029,-0.0974700004,-0.3643222153,-0.1790897101]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/545","title":"New release coming up for this library","comments":"Update: release is planed mid-next week.","body":"Hi all,\r\nA few words on the roadmap for this library.\r\n\r\nThe next release will be a big one and is planed at the end of this week.\r\n\r\nIn addition to the support for indexed datasets (useful for non-parametric models like REALM, RAG, DPR, knn-LM and many other fast dataset retrieval technics), it will:\r\n- have support for multi-modal datasets\r\n- include various significant improvements on speed for standard processing (map, shuffling, ...)\r\n- have a better support for metrics (better caching, and a robust API) and a bigger focus on reproductibility\r\n- change the name to the final name (voted by the community): `datasets`\r\n- be the 1.0.0 release as we think the API will be mostly stabilized from now on","comment_length":6,"text":"New release coming up for this library\nHi all,\r\nA few words on the roadmap for this library.\r\n\r\nThe next release will be a big one and is planed at the end of this week.\r\n\r\nIn addition to the support for indexed datasets (useful for non-parametric models like REALM, RAG, DPR, knn-LM and many other fast dataset retrieval technics), it will:\r\n- have support for multi-modal datasets\r\n- include various significant improvements on speed for standard processing (map, shuffling, ...)\r\n- have a better support for metrics (better caching, and a robust API) and a bigger focus on reproductibility\r\n- change the name to the final name (voted by the community): `datasets`\r\n- be the 1.0.0 release as we think the API will be mostly stabilized from now on\nUpdate: release is planed mid-next week.","embeddings":[-0.5525949597,0.1822217703,-0.2811431885,-0.1986802816,-0.1143226996,-0.1640757173,-0.1367889196,0.6200951934,-0.1699171364,0.0294172,0.0153851993,0.3776414692,-0.3390047848,0.3672705889,0.1043167859,-0.1752639115,0.3210148215,-0.0416744687,-0.0560550578,-0.1332110018,-0.1862286329,0.2256812453,-0.0742022246,0.305655539,-0.025860263,-0.3233274221,-0.2814056277,-0.0468009673,-0.7868933678,-0.3462535441,0.1888869107,0.4110548496,0.2405812144,0.3264506757,-0.0000951948,-0.2936264277,0.5146219134,0.1920410842,-0.079160817,-0.3204219937,-0.1224488243,-0.3491452932,0.0560886823,-0.1020491645,0.1166390926,-0.2722469866,0.1088623405,-0.2063395679,-0.2791921496,-0.1226604059,0.3595248163,-0.0403240994,0.1006082445,-0.3042125404,0.2677444816,0.0366288163,-0.3399615586,-0.3794840574,0.439116478,0.0565735102,-0.1106676906,0.3273255825,-0.2656116188,-0.0698944852,0.3687957525,0.0736092627,-0.077460587,-0.1565372199,-0.0378533788,0.3712140918,0.8543296456,-0.203365609,-0.361828506,0.1668990403,-0.1461803317,-0.0615086444,0.0901183486,-0.0335535631,0.0501466356,0.0294789132,-0.113444984,-0.2867079079,-0.2518987358,0.2157980949,0.063842766,0.5765627027,-0.0002895344,0.106184423,0.3398706615,-0.0698452368,0.3526054621,-0.071973525,-0.3607054949,0.1509602517,-0.2323001176,-0.3804147542,0.2010494024,0.1966285259,0.4493124783,0.1516240835,-0.1423086226,0.0707952529,-0.0682594627,-0.0653670654,0.5567094684,-0.1930360645,0.0366998725,0.01267651,0.4196163118,0.0051412745,0.3992218077,0.0681891963,0.0675145835,0.0127409194,-0.2168606967,-0.2780488133,-0.016046565,-0.3419430554,0.2002929151,-0.1372896582,0.1807544231,-0.2711821198,-0.218774274,-0.0011165665,-0.1216599867,0.2512157559,-0.1377485693,-0.2890484035,-0.0483944118,0.0792820826,-0.2816410065,-0.0127244536,-0.3121898174,-0.0594221987,-0.1738409251,-0.1839885265,0.1321397275,0.046903681,0.0009806285,0.1286846101,0.1825882941,0.1170641482,0.0148533732,0.289239794,-0.3106588423,0.0824632943,-0.2464152426,0.0368597172,-0.305450052,0.1414190084,-0.1102194116,-0.2021080703,-0.3258401453,0.3091481626,-0.0477108061,-0.21213907,-0.0631367341,0.4591263831,-0.3820885122,-0.13325423,0.0042152554,0.2282761484,0.0872695148,-0.2194212973,-0.13909401,-0.0700578839,-0.3947921097,0.0547359549,-0.1158564687,-0.1072246656,-0.0948885977,0.1636662036,-0.1402417272,-0.1771206409,-0.0444038101,-0.0273472667,0.4289875031,-0.3376486897,-0.0377146378,0.2789900899,0.1299616843,-0.5447012782,0.1780115068,0.3231274784,0.1334217489,-0.1073059589,-0.0951223448,-0.0487553477,-0.1796957552,-0.2302102,-0.1669443101,-0.3608306944,0.0364060663,0.0344279334,0.1733640432,-0.1003014222,0.1994162202,0.2221633643,0.1984939128,0.2028439194,0.0216834173,-0.0346924104,0.3472035527,-0.1308937073,0.2749940455,-0.2045080364,-0.2304095477,-0.0655287951,0.0821050033,0.0985777825,0.208558023,-0.0077779959,0.0437852927,0.1023807526,0.0772849545,-0.1399913728,0.1944874227,-0.1108801588,0.0441658497,0.1128684208,-0.5708701015,-0.1422139555,-0.0409016423,-0.011565391,-0.2097584158,0.0943340957,-0.1159503385,0.0136138303,-0.0058277794,0.0820855126,-0.1341051459,-0.0776012093,0.1627473533,0.3813057244,0.097632423,-0.0464064293,0.4080612063,0.557303071,0.2257407606,0.0983065143,0.4142620862,-0.2146224529,0.0041502379,0.1701409817,0.109858878,0.3809589744,-0.2284814268,-0.0142261256,0.0327303745,-0.0299099758,0.3424819708,0.1720540226,0.0183281433,-0.0145885888,-0.029121846,-0.1088026762,0.1273904294,-0.101237677,-0.1904339641,0.1093806326,0.6179309487,-0.1382458806,-0.2296858579,0.032828778,-0.0760989636,-0.177809298,0.1390634179,0.3136151135,0.1576402932,0.2383032143,0.0138473818,-0.2718754411,-0.0079595447,-0.1654439718,0.2090682089,0.1957863718,-0.0159053542,-0.0170692652,-0.2003262043,0.095954217,-0.1916884482,-0.2032521814,0.0164188091,0.2885518968,0.3955717683,-0.1231115237,-0.0884466469,-0.1384778619,0.155727163,-0.1707956791,-0.2725146413,-0.3333197236,0.175023675,0.0680003688,-0.0579224117,0.2110824734,-0.1394922733,0.2071819752,-0.208182469,0.4079120755,-0.3300776184,-0.1575813144,-0.1485504061,0.3032838702,0.0779435486,-0.2391232997,0.3620595336,0.1123809069,0.2344872206,-0.57665205,-0.4527537525,0.2418867946,-0.2270611674,0.0799562111,0.3840842545,-0.1142220795,-0.0015719796,-0.1495757252,0.0527871139,-0.5272467732,-0.2587989867,-0.2288229465,-0.1957895309,-0.0818963125,0.0151876584,-0.3909979761,-0.1447191089,-0.3071348965,0.0158708002,-0.0844822004,0.0223496091,0.2620143294,-0.1104593948,0.2041597962,-0.1820643246,-0.1730185896,-0.2022863179,-0.2899111807,0.0300656129,-0.0976327732,-0.3040081263,0.0888758823,0.1953358799,0.1734179854,-0.1269642711,-0.4470668435,-0.1476915479,0.1828003228,0.2557521164,0.2825198174,-0.0221764334,0.414331913,0.1158304662,-0.3213061988,-0.1094314456,-0.244275555,-0.3254814148,0.1904073358,0.3608268797,0.1292877644,0.1915955544,0.0172287002,0.3703114688,0.0349856541,-0.1337102205,0.1464427859,0.2402999699,0.2746374905,-0.2306236327,-0.0648512766,0.0903673619,0.0024939389,0.0453705043,0.3252563179,0.1467186064,-0.5317548513,-0.3780499697,0.0321054198,0.0566413961,0.0927298442,0.0188038927,0.0903562009,0.0390443653,0.09636233,0.0396672301,-0.1069975868,-0.1464128643,0.1663310826,0.1741354018,0.1542081535,-0.1314745843,-0.2837042212,0.1488749534,-0.3126962185,0.2888373435,0.0940982029,-0.0762261674,-0.0122887436,-0.0238202941,0.1239950806,0.0012919521,0.4169732034,-0.2627556622,-0.4117499292,0.1502257138,-0.0550670251,0.0589873344,0.1477471292,-0.1672425568,0.1402536631,0.0541079715,0.3949808478,0.1117444634,-0.5463081598,0.2634912729,-0.0684778094,0.0224114154,-0.1537786722,0.0242978409,-0.1235016212,-0.0399253294,-0.095754005,-0.2760098875,-0.1445921063,-0.0889273211,-0.4988796413,0.2125972509,-0.0768793598,-0.0553389527,-0.0515610799,0.1934533715,-0.0504479893,0.1060009748,0.2567024231,0.3126895428,0.2861280441,0.0490766354,-0.1889261156,0.1667094827,0.2929116488,-0.0371816233,0.2208798677,0.2186275721,-0.0038450703,0.2150868028,-0.0819933936,0.0542903207,-0.2279860824,-0.2671776414,0.4450004399,0.0652883351,-0.4337330163,-0.5365271568,0.378606081,0.1446326971,-0.2591173649,0.5277165771,0.1524919719,0.1614003181,0.0280129127,0.2135293782,0.8639208078,-0.1850996614,0.0836001858,-0.1999328136,-0.202820316,0.4804623723,-0.138188675,0.0716976523,-0.3027890623,-0.0018237492,-0.1338136345,0.1474702209,0.1555192769,-0.0278938729,-0.1692372411,0.1561857164,0.3681625426,0.1515266895,-0.1770182401,0.3562440872,0.0653228983,-0.0722198412,-0.048105184,0.2932276428,-0.143599838,-0.0328862555,-0.1239669099,-0.0509333648,0.1094553173,-0.0095549878,-0.4072631598,-0.2251214683,0.2899646163,0.2076732814,0.0375397764,-0.1894572377,-0.0047233161,-0.0291570537,0.2489641607,0.1302993149,-0.1448979378,0.1563552916,-0.2412340194,0.0311048981,0.0991197675,-0.0473654978,0.405184716,-0.2712770104,-0.1328590512,-0.0519573726,-0.0518799387,-0.3155253232,-0.3936416209,0.1564204842,0.2460119575,-0.2070404589,0.3774801195,0.1909811795,-0.1287969649,-0.1125301868,0.2530385554,0.3683555424,-0.00176344,0.2307890803,0.2673897147,-0.2230395526,-0.1719782203,0.1677080542,0.2834619284,-0.1555492878,-0.1427919716,0.0467762984,-0.2155649066,-0.335378319,0.0409709252,0.2777130902,-0.0587936714,0.1719716787,0.0423018038,-0.0938845873,0.488861382,0.4687161446,0.1909093857,-0.0609669909,-0.4412037432,0.0431878231,-0.183502987,0.3752722144,-0.3523418009,0.1656083614,-0.1600692272,-0.1868311167,0.1454935819,-0.0234716833,-0.4344523549,-0.0707686916,0.0956689566,0.1055879593,-0.1305017173,0.0482375622,-0.0995093063,0.006623629,0.1659105122,-0.2924628258,-0.4715195298,-0.3030092716,-0.1147016063,0.0841460675,-0.0112629663,0.0638138875,0.0832069814,-0.101579614,-0.3627947271,0.0337761119,0.1067777798,-0.0321456566,0.1437891871,0.2426048666,0.3892826736,-0.2242628038,0.2370267659,0.1925827414,0.3193385899,0.2376919538,0.1441548169,0.306489259,0.1250895262,-0.2121041417,0.0447688736,0.200383991,0.0023225565,0.2071258426,0.2989008129,0.0541724451,-0.0414831415,0.3938335478,0.4007311165,0.1513117403,0.266503185,-0.358305186,0.1713320017,0.2957594097,-0.1262369454,-0.3396953642,-0.0716839805,0.3122537434,-0.0973759368,-0.0660254359,0.3646315336,0.0768963099,0.2859972417,-0.0031084735,0.2607701123,-0.3374049067,0.2920914292,0.1450373083,0.1032851487,-0.0239671804,0.2845978737,0.0130656874,-0.0130039835,0.4786172211,0.0482535921,0.3492493331,0.4752145112,0.23886244,0.1737211794,-0.0064965878,0.0596412495,0.0789453909,0.0299087726,0.1065132245,0.2205812484,0.3414841294,-0.0383950733,0.0787723586,-0.2200274169,-0.1344373077,-0.0283808485,-0.2461355925,-0.1330412328,-0.3000856936,-0.0052363523,-0.0242454894,-0.0576829948,0.2010937929,0.4837923348,0.050885845,0.1850036085,-0.2544620931,0.0733862668,-0.0140565075,0.2912755907,-0.0830312222,0.2164705992,0.0972768366,-0.0480329171,0.1145717651,0.2155565917,0.0952679887,-0.1161056161,-0.302626282,-0.002402121,-0.2100431323,-0.1574826092,0.0693127885,0.2608692944,0.1986811459,-0.0381656811,0.1386571079,0.1715403795,-0.1287280321,0.2629286051,-0.0664159209,0.0477668047,-0.1233182997,0.1711016744,0.2559481561,-0.1452478319,-0.1696902514,0.3436050117,-0.4000928998,-0.2778905928,0.2851356864,-0.4175385833,-0.1929260343,-0.095417887,0.1736502647,-0.0776407123,0.3035697341,0.2355186641,0.25424698,0.123435773,-0.3660645187,-0.0971119776,0.2088132203,0.1059997901,0.2483169138,-0.2235615999,0.1951987296,0.1649823636,0.1255246252,0.1915245652,-0.243078813,-0.0355301872,-0.0267846193,-0.4530907571,0.152463913,0.0027476905,0.2091756463,-0.0884147361,0.0541524403,-0.0936823115,-0.2901253104,0.0448731482,-0.2098949999,-0.0150591424,0.3648504019,0.1683992893,0.2890778482,0.0904652476,-0.0242440216,-0.148339957,-0.268348664,-0.3204602301,-0.2356704772,-0.2177486718,0.0561867319,0.1189127117,0.106680952,-0.1155577227,-0.0096346503,0.0852128789,0.3258015215,-0.3670690954,0.2174212039,-0.4504807293,-0.1004456431,-0.0250268653,-0.3050011694,-0.0882938951,0.2485762089,0.089781113,-0.0485085845,-0.2607584,-0.2336314768,0.3598339558,-0.076755777,-0.0876961499,-0.0206118822,-0.2099297494,0.0735640749,-0.0189702678,-0.6225720048,0.0144729037,0.14648588,0.2227266282,-0.022387458,0.1843177825,0.2618782222,-0.0096289543,-0.1621606499,0.1101030111,0.1062785238,-0.3068648577,0.1189224496,-0.1422388256]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/543","title":"nlp.load_dataset is not safe for multi processes when loading from local files","comments":"I'll take a look!","body":"Loading from local files, e.g., `dataset = nlp.load_dataset('csv', data_files=['file_1.csv', 'file_2.csv'])`\r\nconcurrently from multiple processes, will raise `FileExistsError` from builder's line 430, https:\/\/github.com\/huggingface\/nlp\/blob\/6655008c738cb613c522deb3bd18e35a67b2a7e5\/src\/nlp\/builder.py#L423-L438\r\n\r\nLikely because multiple processes step into download_and_prepare, https:\/\/github.com\/huggingface\/nlp\/blob\/6655008c738cb613c522deb3bd18e35a67b2a7e5\/src\/nlp\/load.py#L550-L554\r\n\r\nThis can happen when launching distributed training with commands like `python -m torch.distributed.launch --nproc_per_node 4` on a new collection of files never loaded before.\r\n\r\nI can create a PR that puts in some file locks. It would be helpful if I can be informed of the convention for naming and placement of the lock.","comment_length":4,"text":"nlp.load_dataset is not safe for multi processes when loading from local files\nLoading from local files, e.g., `dataset = nlp.load_dataset('csv', data_files=['file_1.csv', 'file_2.csv'])`\r\nconcurrently from multiple processes, will raise `FileExistsError` from builder's line 430, https:\/\/github.com\/huggingface\/nlp\/blob\/6655008c738cb613c522deb3bd18e35a67b2a7e5\/src\/nlp\/builder.py#L423-L438\r\n\r\nLikely because multiple processes step into download_and_prepare, https:\/\/github.com\/huggingface\/nlp\/blob\/6655008c738cb613c522deb3bd18e35a67b2a7e5\/src\/nlp\/load.py#L550-L554\r\n\r\nThis can happen when launching distributed training with commands like `python -m torch.distributed.launch --nproc_per_node 4` on a new collection of files never loaded before.\r\n\r\nI can create a PR that puts in some file locks. It would be helpful if I can be informed of the convention for naming and placement of the lock.\nI'll take a look!","embeddings":[-0.0677058026,0.246800378,0.0098843975,-0.0446524099,-0.0798522085,-0.2330475897,0.5056552291,0.1000798643,0.3607544303,0.0900502354,0.2382265478,0.2282434851,-0.1981795132,-0.3228935599,0.1710017622,-0.1084064543,0.0410026573,0.0717097819,0.0613710657,0.0232866276,-0.2557511926,0.1316337734,0.0588490814,0.3450223505,-0.3524826765,-0.0700740665,-0.2764480114,0.5191791654,-0.0313581116,-0.4743485749,0.1093649864,0.3782074153,-0.0096424678,0.3501272202,-0.0001107842,0.0113410093,0.109303534,-0.0824821591,-0.1725915074,-0.2966860235,-0.1778950691,-0.6893734336,0.3947988153,-0.2921565771,0.2900379002,-0.3698517382,0.5056335926,0.0171923265,0.4170382917,0.2521377802,0.1046826467,0.2477042228,-0.1231489107,0.0176402908,0.0486078076,-0.0763560683,-0.1369857192,0.3870864213,0.3630029261,-0.3143343329,-0.4267943799,-0.2233087569,-0.0590115599,0.0248145182,0.3551026285,0.0978184119,0.0094909109,-0.2972874045,-0.5111076832,0.1290315092,0.0657319576,-0.0298541039,-0.2860132158,-0.7358392477,0.2892959714,-0.0180504285,0.2564793527,0.3986741602,-0.2367146313,-0.1155090407,-0.228845194,-0.0933738574,0.0202806909,0.0442148857,0.3830667436,0.2889916897,0.2357245535,0.1337042749,0.3594629169,-0.0206270348,0.1386083364,0.0982034802,0.0450825505,0.1986579597,-0.3583192825,0.0624857955,-0.0066759055,0.1486155689,0.0010228946,0.2065602988,-0.0266604554,-0.4352199733,0.1166021004,0.321532011,0.0537072457,-0.0653092712,0.00281737,0.1300088018,0.2937302887,0.188105464,-0.289681673,0.1395388097,-0.1073758453,-0.5000427961,-0.2825135589,0.0657903105,-0.0961981937,0.1333681494,-0.0136662098,0.0805119276,0.0203790534,-0.2405004501,0.1502341926,0.3017907143,0.3954746127,0.0460775346,0.0997079015,0.0829486102,-0.3044987917,-0.0490257703,-0.1852066517,0.0995360687,-0.5700981021,0.1342210323,0.183630228,-0.0900456831,0.4895657003,0.1473776251,0.0073029739,-0.104641974,0.4991342127,0.1455028206,-0.0002605886,0.2877076566,0.079690814,-0.2237643898,-0.0495640598,-0.4024512172,-0.1604245901,0.0952543616,-0.4077451229,-0.1332660317,0.0501317196,0.1329575032,-0.0524365567,-0.0044572675,0.4236879349,0.0989752784,0.3883138895,-0.1148898304,0.3558266461,-0.2230216265,-0.1846059412,-0.0505362712,-0.2495144159,0.4267528951,0.1101217791,-0.0429347903,0.099271968,-0.2750361264,-0.1786448956,0.2227526903,-0.2187804729,0.387401402,-0.1692819744,0.2657566071,0.307842046,-0.1192565411,-0.2177829146,0.4583076537,-0.4290387928,-0.3405137062,0.3833592832,0.2986827791,0.0345679447,-0.2598641515,0.1739815325,0.3889186084,-0.0972505212,0.1560527831,0.0702670068,-0.0258696862,0.1519034207,0.2459802777,-0.1345254332,-0.045698572,-0.0676070601,0.4174708724,0.114783816,-0.1391945928,-0.1387461424,0.2798424065,-0.4744606316,0.1197502911,-0.3098706901,0.256192565,-0.4087414443,0.1604101807,-0.4327827692,-0.0157175567,0.1223010421,-0.0545525104,0.1128135398,-0.1596733779,-0.4367004037,-0.1595250666,0.0468971655,0.3212048709,-0.1085798144,-0.0195358843,-0.220359832,0.4198933244,-0.3391633928,0.0943924561,-0.3506613076,-0.0524570048,-0.1677867323,-0.1643053144,-0.0063866093,0.1288645864,0.0568443649,-0.1765687019,-0.0985044315,0.5064197183,0.2953180671,0.3453960121,0.2189708799,0.0352551714,0.1399111748,0.3919066489,0.0280317292,-0.1242520139,-0.0447771959,0.0266181957,-0.1831101626,0.5922919512,-0.1155326292,-0.2013705969,0.1205875203,-0.051510185,-0.0160528217,-0.2434578687,-0.0382225327,0.1014510691,0.4043351114,0.4942108095,0.320148319,0.299767524,-0.3885113895,-0.2745434344,0.2222897261,-0.0186026823,-0.0443459861,-0.0303745568,0.567299664,0.0034676124,0.0062054489,0.3909807205,0.7118954062,0.0966428444,-0.1056087539,0.1117979661,-0.2135190219,-0.5867146254,-0.0622233897,0.059476465,-0.0794513971,0.2391099781,0.0959490538,-0.1687955707,-0.5147601962,-0.3524712026,0.1926231682,-0.0421268009,-0.214689374,0.0376821682,-0.0766104758,-0.1698518395,-0.4905445874,0.1203018948,-0.2531323135,-0.2034400403,0.2605524957,0.3021129668,-0.3972657323,0.0991030633,-0.2988418639,0.039942041,-0.3393559456,0.2586064041,0.0830561221,-0.1486356556,-0.1588339955,-0.0001141317,0.4003840685,0.0571696199,0.4111907184,0.1599784642,-0.3651053309,-0.0219705794,0.0036741027,-0.0344845019,-0.1788079292,-0.0536114089,0.0660381019,0.0097992159,-0.1585678607,-0.4596485496,-0.1014436781,-0.2386858761,-0.1224188805,-0.2336742133,0.0630361587,-0.1799706221,-0.4012879431,-0.4286587536,-0.4084974527,-0.1953618973,0.1994516402,0.1581364125,0.0222846642,0.1596483588,-0.2513945699,0.1856624484,0.1069012955,0.2143770456,0.0466150679,-0.3738225102,0.0883309394,-0.240737617,-0.1872440428,-0.2338407785,0.0396939181,-0.4090135694,0.0997457653,-0.3041002452,-0.3219039142,0.0314073302,0.0792534128,0.1068549007,-0.0106761027,0.1220357567,-0.1001482755,-0.1499199271,-0.0254638214,0.3709861338,0.5164568424,0.029937705,0.1074117348,0.0517379157,-0.0859827548,0.014896092,0.7045516372,0.226897344,-0.1163077801,0.3795087039,-0.053894788,0.0138993226,-0.0779856965,0.1365267485,0.3311722577,0.2960718572,-0.1130415201,0.1512904018,-0.1583374143,-0.0680848882,-0.0717345998,0.1306525022,-0.2383367419,-0.2801853716,0.0209427495,0.070697017,0.1888109446,-0.2280309498,-0.068931967,0.2999784946,-0.1525603682,-0.0595790446,0.3854104578,0.0108649433,-0.1215480715,-0.2265806198,0.0134218819,-0.333181262,0.0400681458,0.1074087843,0.3775999248,-0.1081189141,-0.0683751628,0.1759139299,0.0612722337,0.5162094235,0.1663949788,0.0035996442,0.0225815903,0.1820749193,0.252925992,-0.2765934169,-0.1958997399,0.2423358858,0.2367173135,0.3072757423,-0.1091757566,-0.2227484435,-0.3072417378,-0.1019435376,-0.3562960923,-0.2899579108,-0.1835357547,-0.098529987,-0.536139667,0.352673471,-0.0695096627,0.016628867,-0.1489235014,-0.0282961167,0.1993024051,0.2633692622,0.0155256661,0.0939764306,-0.1137077138,0.2371772081,0.0202278737,0.1834658682,0.2923348844,0.0043769898,0.3856642842,-0.3102437258,0.0460615903,0.1041700467,-0.2689587176,0.3282300234,0.6207247376,0.1886660159,-0.0114408126,0.0680550486,0.0658749342,-0.0052560475,0.157963112,0.2851806581,-0.3106063902,0.210256964,-0.5296699405,0.1316988915,0.2349701971,-0.1436539888,0.1230359748,-0.3661715984,-0.2558087707,0.3320649564,0.1715979725,0.7989926934,-0.3118830025,0.3623324931,0.1557096243,0.1445892453,0.5249934196,-0.388669312,0.0334546454,-0.5821710229,-0.4371654987,-0.0152599718,-0.1928639412,0.1633029282,0.0860542879,-0.136310026,0.2617340684,0.1066394746,-0.2371177971,-0.1794357151,0.5527573228,0.0086119836,-0.3393853605,-0.2361819744,-0.0228141937,0.0989314914,0.1342336237,-0.1996278018,-0.0922464728,0.1922352612,-0.1958047301,-0.0866862163,0.1465521753,0.1842530668,0.2953158021,0.061376363,-0.1837503612,-0.1603994071,0.6436176896,0.0268322229,-0.1328427345,0.0613591745,0.2885336876,-0.1704771668,0.0597193651,-0.1822382212,0.0483225696,0.500823617,-0.095921658,-0.0703815892,-0.3279990554,0.0532332659,0.0114413779,-0.0001163411,0.1302745193,0.1692200154,-0.0173928551,-0.3229168653,-0.0186924282,-0.1779464334,-0.0375133567,0.0526874922,-0.0052405377,-0.0809332877,-0.3094137609,-0.2059779465,-0.2088387161,0.144138813,0.3705526888,-0.0248552617,-0.172698617,0.501003325,0.2646502554,-0.0197116192,-0.1917690188,0.135245949,-0.0674841926,-0.158984974,0.059772525,0.1769757569,-0.1514421999,-0.0304230209,0.1934452802,0.2045143247,0.1152872592,0.1111493334,-0.3941908777,-0.4443686903,-0.0494336002,-0.2992429137,0.2751091421,0.2731451392,0.1927789301,0.2921340168,0.1022768244,-0.22815983,-0.1151247621,-0.4438696802,0.0175744705,0.139432326,0.0236053132,-0.1094143391,0.1466816813,-0.0040078894,0.2691300511,-0.146657005,-0.16732198,-0.267160207,0.1688487232,0.2614877522,0.021701226,-0.1825638264,-0.5259767175,0.1659283936,-0.0053436533,0.093394734,-0.0436030813,-0.0631880984,0.4087051451,0.042960614,0.3635098338,0.0720632747,0.0284395386,-0.3225614429,-0.1357703507,0.1890716404,0.3749207258,-0.1030851454,-0.0514150523,-0.529805541,0.1874416322,0.143492043,0.4414678812,-0.2233514041,0.1548980922,-0.261159867,-0.1127183512,0.1975956112,0.5131251216,-0.247412324,-0.094698675,0.3112428486,0.1242288202,-0.0823097676,-0.1215076298,0.3611730635,0.1266506016,-0.1136336476,0.1124421358,0.4369507432,-0.0951650068,-0.6982483268,0.0278137904,-0.0936950594,-0.0088491319,0.1576332599,0.3655929267,-0.044892773,0.0950798541,0.3773104548,0.0222483482,0.07148958,0.1269896477,-0.1359091699,0.3199797273,0.0014308718,0.1654050797,0.1159876809,-0.3079258502,0.2142968178,0.2307337224,0.3312051296,-0.1587822586,0.094684124,0.6083934307,0.0255625676,-0.1024966761,0.0092920028,0.5487231016,-0.1987001151,-0.2935665846,-0.2630930543,0.007810662,-0.2928235233,0.2008779943,-0.0292935632,0.0938627571,0.3164236844,0.3376310468,0.2317691445,-0.260727495,0.3160362542,0.125739336,-0.0488046408,-0.4051788151,0.0433712527,0.2779586911,-0.0664090887,0.2960099578,-0.1292642206,0.259168148,0.1429398954,-0.1105441302,0.1586129516,-0.0575407632,0.169416815,-0.2873646915,0.388222605,-0.0768504441,0.3020793796,0.5372684002,0.0154454205,-0.0874223411,0.1215621307,0.4625393748,-0.2685354948,0.0558237992,0.5503655076,0.4539985657,-0.1767344624,0.1436665058,0.1611837149,-0.449647367,-0.087602295,0.1769050658,0.0735859722,0.2607281804,-0.1873313189,0.0331540667,0.1965666264,0.4934288263,0.2552929521,0.2330641598,-0.2920608819,-0.2963970602,-0.6911859512,0.4115105569,0.0342653655,-0.2891732454,-0.0248576943,0.1318686604,-0.2231898308,0.0142105855,-0.2697614133,0.0453522466,-0.0008936492,0.4122497141,-0.2083510906,0.1298426688,-0.0750634,0.0155947208,-0.0237902552,-0.0193383768,-0.0967874601,-0.2765207887,-0.098830387,-0.1264156252,-0.0923763961,0.0179766994,0.2360464483,0.2000232786,0.0317676514,0.4896837473,-0.2605349422,0.0843976513,-0.0306944344,0.0794843286,-0.3207927346,-0.2721894979,-0.015254546,0.1783100218,-0.3969019055,0.4975353181,-0.3603973687,0.1672009677,-0.2821972072,0.2074938864,-0.3503920138,-0.0065578162,-0.0577528477,0.0158862285,0.207890138,0.3653307259,0.0557827316,0.0192423742,-0.1916590631,0.0805798247,0.3538522124,-0.6020537615,-0.0252846442,-0.1411203295,0.1768609136,-0.1382386982,0.1548046619,-0.4462213218,0.1186364442,0.11312069,-0.2991206348,-0.4804895818,0.2095164508,-0.0415919609,-0.2597482502,-0.051907625,-0.0431860648,0.1077789441,-0.4856520891,-0.0588252917,-0.2727823555]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/539","title":"[Dataset] `NonMatchingChecksumError` due to an update in the LinCE benchmark data","comments":"Hi @gaguilar \r\n\r\nIf you want to take care of this, it very simple, you just need to regenerate the `dataset_infos.json` file as indicated [in the doc](https:\/\/huggingface.co\/nlp\/share_dataset.html#adding-metadata) by [installing from source](https:\/\/huggingface.co\/nlp\/installation.html#installing-from-source) and running the following command from the root of the repo:\r\n```bash\r\npython nlp-cli test .\/datasets\/lince --save_infos --all_configs\r\n```\r\nAnd then you can open a pull-request with the updated json file.\r\n\r\nOtherwise we'll do it sometime this week.","body":"Hi,\r\n\r\nThere is a `NonMatchingChecksumError` error for the `lid_msaea` (language identification for Modern Standard Arabic - Egyptian Arabic) dataset from the LinCE benchmark due to a minor update on that dataset. \r\n\r\nHow can I update the checksum of the library to solve this issue? The error is below and it also appears in the [nlp viewer](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=lince&config=lid_msaea):\r\n\r\n```python\r\nimport nlp\r\nnlp.load_dataset('lince', 'lid_msaea')\r\n```\r\n\r\nOutput:\r\n```\r\nNonMatchingChecksumError: ['https:\/\/ritual.uh.edu\/lince\/libaccess\/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9\/lid_msaea.zip']\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 196, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 150, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n download_config.force_download = download_mode == FORCE_REDOWNLOAD\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 469, in _download_and_prepare\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 36, in verify_checksums\r\n raise NonMatchingChecksumError(str(bad_urls))\r\n```\r\n\r\nThank you in advance!\r\n\r\n@lhoestq ","comment_length":68,"text":"[Dataset] `NonMatchingChecksumError` due to an update in the LinCE benchmark data\nHi,\r\n\r\nThere is a `NonMatchingChecksumError` error for the `lid_msaea` (language identification for Modern Standard Arabic - Egyptian Arabic) dataset from the LinCE benchmark due to a minor update on that dataset. \r\n\r\nHow can I update the checksum of the library to solve this issue? The error is below and it also appears in the [nlp viewer](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=lince&config=lid_msaea):\r\n\r\n```python\r\nimport nlp\r\nnlp.load_dataset('lince', 'lid_msaea')\r\n```\r\n\r\nOutput:\r\n```\r\nNonMatchingChecksumError: ['https:\/\/ritual.uh.edu\/lince\/libaccess\/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9\/lid_msaea.zip']\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 196, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 150, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n download_config.force_download = download_mode == FORCE_REDOWNLOAD\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 469, in _download_and_prepare\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 36, in verify_checksums\r\n raise NonMatchingChecksumError(str(bad_urls))\r\n```\r\n\r\nThank you in advance!\r\n\r\n@lhoestq \nHi @gaguilar \r\n\r\nIf you want to take care of this, it very simple, you just need to regenerate the `dataset_infos.json` file as indicated [in the doc](https:\/\/huggingface.co\/nlp\/share_dataset.html#adding-metadata) by [installing from source](https:\/\/huggingface.co\/nlp\/installation.html#installing-from-source) and running the following command from the root of the repo:\r\n```bash\r\npython nlp-cli test .\/datasets\/lince --save_infos --all_configs\r\n```\r\nAnd then you can open a pull-request with the updated json file.\r\n\r\nOtherwise we'll do it sometime this week.","embeddings":[-0.0616527833,0.4437022507,-0.0581613593,0.0664172098,-0.307210058,0.1153274924,-0.2732498646,0.5395020843,-0.0016354375,-0.1196584553,0.0857631639,0.2251154929,0.0789773464,-0.1222939044,0.0061919261,0.1934710592,0.0321503244,0.0933852121,0.0987019464,0.0393903553,-0.2053777874,0.2148771733,-0.057483878,-0.2985480428,0.0893565938,-0.004181175,0.0692205578,0.0182900559,0.1299763471,-0.4194515646,0.3206056654,0.2793268561,0.0003896341,0.1324239075,-0.0001272806,-0.0694950148,0.3210234344,-0.0919738933,-0.2670393586,-0.0846346244,0.0034744749,-0.4629070163,-0.0613037869,-0.2885442674,0.1637886763,0.2399886698,0.1029127687,-0.4139173031,-0.095096156,0.1851278096,0.1083362922,0.5190732479,0.0809718668,0.3380495608,0.256370753,-0.42114079,0.0306582432,0.3837837577,0.3393696845,-0.1366900653,-0.4413134158,0.3681635559,-0.3522348702,0.0423129387,-0.0161721241,-0.3124465644,0.0194569509,-0.0667776987,0.302544415,0.1907632947,0.3051871955,-0.4455226064,-0.2008479834,0.0067092571,-0.2116422057,-0.1155177131,0.4846940637,0.1011212692,-0.0301548149,-0.1416909546,-0.0993822813,-0.0432990901,0.1072188467,0.2136925161,0.3474704325,0.4857030511,0.243153125,0.1118917316,0.1624684483,0.0055500744,0.503978312,-0.0932113156,-0.1411801726,0.1578629613,-0.3675528467,-0.1388427168,0.0753963515,0.3629109859,0.1303834319,0.4638726413,0.5369125009,0.247657612,-0.1883206666,0.1930721104,-0.1384759098,0.2871352434,0.4059982002,-0.1246177554,0.1441257894,0.4248977005,0.0256022699,0.065224953,0.1155529246,-0.2771501243,0.3135475516,0.4061032236,0.0826807097,-0.4132894278,-0.227190122,0.307826966,-0.1956929564,-0.3644892573,0.211998567,0.1644034833,-0.3546676934,0.2263011485,0.1475031525,0.0085846949,-0.1239643097,-0.5568845868,-0.0674783885,0.2780528069,-0.1873663217,0.1560589522,0.2345919013,-0.3751237094,0.4669339359,-0.3525468111,0.0226076655,-0.0944171771,0.2752314806,-0.1536362767,0.0063340105,0.2864765227,-0.0740386918,-0.1444516182,0.1842181534,0.1451916844,-0.3300545812,0.2802166045,0.1353553981,-0.1413926184,-0.1740648896,-0.0231370293,-0.4117465615,-0.2353129536,0.272849828,-0.0858844668,0.390968442,-0.6479693055,0.0522488616,-0.0388518497,-0.2448469102,-0.1984824985,-0.1250327229,0.0340124369,-0.1752650887,0.0889368281,-0.2065021992,0.1218558773,0.301831007,0.1827165931,0.059068989,-0.2488514483,-0.1102877334,0.1323599666,0.2490991056,-0.3191206157,-0.6921525002,-0.0797567293,0.1015781909,0.3253774941,0.2345360368,0.4361954927,-0.1628785282,-0.2904224396,-0.0313170739,0.1885817349,-0.0350644253,-0.0714681223,-0.3921247721,-0.1106406301,0.5248131156,0.0953375474,0.0534198955,-0.0261638723,0.1034081951,-0.0051165503,0.2173566222,0.0385111012,0.0118800262,0.0543003716,0.0686132163,0.0348475017,0.0741711408,-0.0369870961,-0.4227679074,0.1104315296,-0.2665937245,0.2791252434,0.0841261968,-0.1472602785,-0.0798559189,-0.2010103911,0.008968005,0.0517491885,0.0013107359,0.3164534271,0.1445411444,-0.0932148546,0.0056738132,0.036909584,-0.2276382744,-0.0102180643,-0.5147677064,0.2930194736,-0.0692704618,-0.2226178199,0.2267709523,0.2620928288,0.1244557053,-0.067651242,-0.3514721692,0.379909277,-0.049288936,-0.139429003,0.0233200267,0.4658017457,-0.1902141124,-0.1928856522,0.2801755071,0.7407286763,-0.0178896319,-0.0470744595,0.188155219,0.4245472848,-0.0047320961,0.1563244313,-0.0887772217,0.1785681695,0.1582686752,-0.2391062677,-0.256729424,-0.0941786394,0.4943873882,-0.148961395,0.178952679,0.0457770936,-0.1842954308,-0.0852906406,0.3507559299,0.0191176645,0.0700724646,0.2472043037,0.1055887192,-0.3084563613,0.1519667804,0.34983477,0.2286102325,-0.0166477971,0.006565324,0.2571461201,-0.0707734674,0.0059679616,0.0090975491,0.0203417037,0.2311230451,0.4381026924,0.2708524466,0.0421677716,-0.4537835121,-0.2937007248,-0.0823328644,0.3356816173,-0.0985868648,0.0412338115,-0.2519593835,-0.3669678867,-0.1616517901,-0.3328652084,-0.5280544162,-0.3471707702,0.0056135268,0.3331266046,-0.0336648524,0.3375812769,-0.2607325315,-0.041509334,-0.2104506344,-0.055516988,-0.0397895426,-0.3035803735,-0.3481827378,-0.100549072,0.6255894899,0.1226193532,0.3256288469,-0.2633631825,-0.0386355855,-0.4359060526,-0.4488388598,-0.2283978611,-0.0557631403,0.0680425093,0.3516705632,0.0911212862,-0.0205849521,-0.3282206953,0.1320411414,-0.0937089622,-0.1445062608,0.0734271407,-0.1742323339,0.1782113463,0.1171157733,-0.1628405005,-0.0572150797,-0.2265317887,-0.0431701057,-0.1390309632,0.3625038266,0.0243449211,-0.3539898098,-0.0659313425,0.0604897924,0.4428186417,-0.1127821505,-0.5371189117,0.183861807,0.1153431833,-0.1968841404,0.0928090662,-0.1693492532,0.6429797411,-0.092325367,-0.509729445,-0.0690603107,-0.1017558351,-0.0644859299,-0.0095773246,-0.0312164482,0.4292153716,-0.0215043649,-0.0031932807,-0.2097126395,-0.4482651353,0.1087318808,0.0067476174,0.437325418,0.073648639,0.3563734889,0.170651257,0.0788869634,0.1601974666,-0.162990272,0.4158077538,0.4866502881,0.1476314217,0.0428027995,-0.0988665894,-0.1026581004,-0.0358230583,-0.0471012406,0.0812134668,-0.057063546,-0.1882273108,-0.116942361,-0.3559015989,-0.3291288018,-0.2423394322,0.1743614674,-0.3096680939,0.3644334972,0.2393897623,-0.1058761254,-0.3316639662,-0.420277983,0.1958696395,0.3060153723,-0.1326766461,-0.0060011023,-0.550457418,-0.34876284,-0.1723356545,0.7032877803,0.2243926227,0.6603319049,0.0806642771,-0.0668382496,0.0195838772,-0.1764271855,0.0931476653,-0.365216881,0.2800974548,0.1581119597,0.0958499089,-0.0339047648,-0.2032892555,0.1098054051,0.2863220572,0.4885847867,-0.178142339,-0.2392450124,-0.121640332,0.2182543576,0.2945373058,-0.0311449543,-0.2273663133,-0.4379222095,-0.2903141081,-0.114300631,-0.0357312597,-0.2543928623,0.320625633,0.2463235408,0.177491352,0.1452334076,-0.2266955674,0.3136151433,-0.0948358923,0.2140125483,0.3180748522,0.1195525751,-0.3462470174,0.130244419,-0.0941932723,0.2956397831,-0.1126083434,-0.1375366449,-0.0734804794,0.2040028274,0.0879630968,0.1910957098,0.2005631477,0.2737550437,0.4517329633,-0.1058313847,0.1607081145,-0.1901838034,-0.0732441768,0.1468066275,-0.2490932643,-0.2411087006,0.2625308037,0.1931416541,-0.0387197845,0.3418625891,-0.4028802216,-0.1004320756,0.2029738277,0.4714302421,1.0572549105,0.0201435611,0.0580873601,-0.050939858,-0.1898695081,0.0285251606,-0.0933944061,0.2850465775,-0.4167386591,-0.4778108299,-0.0199747197,-0.1191670597,0.3133391738,-0.2553495765,-0.2167989016,0.41465801,-0.0257352516,-0.4536561668,0.1333514303,0.2383010685,-0.425064832,-0.0571774803,-0.1014586166,0.0327990949,-0.3412192464,0.3648760021,-0.1134908721,0.0406587645,0.1785674095,-0.3359895647,-0.3500402272,0.0750798509,-0.0747927055,0.0608153604,0.3293658793,0.1280973852,-0.0595258772,-0.1196693555,0.3529457152,0.0838501155,0.0882170871,0.0468815304,0.49015522,-0.0665743724,0.3549170196,0.2023560554,0.3399666846,-0.1613616198,-0.1293440312,0.0918295458,-0.123580642,-0.5137318969,0.3060742319,-0.2194625437,0.0945610851,0.0228990261,-0.1769923866,0.205457747,-0.0713040233,-0.2482821047,0.0278570279,0.1989021003,-0.3135830462,0.1156932488,-0.0807046816,-0.2498116493,-0.0033673241,0.4579544961,0.3928121328,-0.2363060266,0.3262052536,0.0302985553,-0.1058911532,-0.1204885095,-0.2305814624,-0.1916443259,-0.2923538685,0.1083471105,-0.0822686255,-0.075918518,-0.1271994263,0.5551509261,0.2674683928,0.6689737439,-0.12501131,-0.498616308,-0.1051060259,0.1533489376,0.0695057958,0.1367290616,-0.289671123,0.0720115378,-0.0632774085,-0.0139155621,-0.1498123407,0.0530563369,-0.2679403424,0.1874872595,-0.0079384195,0.16731821,-0.0482473932,0.1539581865,-0.0541640557,-0.0090228152,-0.244091928,-0.010713323,-0.0693433359,0.1767442375,0.0797735229,-0.0332644545,-0.1182662398,-0.3067573011,-0.0969026387,-0.0597471595,-0.4343479574,-0.0200659856,-0.1304022223,0.2618319392,-0.0712140724,0.0125173153,0.1222654134,-0.0130180223,-0.3334136903,-0.1952343285,-0.098141849,0.1256816834,-0.0123606939,-0.0134816281,0.2143861204,0.0588422567,-0.2207695693,-0.1099382117,-0.0495147519,0.00431058,-0.058556892,0.4181289077,0.1668494642,0.3067296743,0.0485336855,-0.3271096051,0.358933717,0.0716522858,-0.5490003228,0.0469181165,0.2238919884,-0.0183319878,0.0063809059,0.0733909532,0.2378414869,-0.1913523227,0.6728568077,0.068653591,-0.0561068282,-0.3147752583,0.3004775047,0.1758289039,0.0488543585,0.0803689286,0.1535321921,-0.2527367473,0.0332416743,-0.2883599401,-0.044377137,0.4211072326,0.05685772,0.3054702282,-0.4572608769,-0.2989241779,0.2259770036,0.4199058712,0.4424139857,0.0849668682,0.2154006064,0.5814492702,0.0252286009,0.0394732319,-0.2616809309,0.1191078871,0.0596153103,0.0078946827,-0.163883999,-0.0199991167,-0.0829944387,0.0370736197,0.1350004822,0.384144187,0.2583817542,0.039419584,-0.0428873338,-0.3136141598,-0.4687738717,0.0511496328,-0.1495767534,-0.0921679139,0.3379406631,0.311460048,-0.0182744265,0.0465470329,0.3662718832,0.260745585,0.149502784,-0.0323063731,-0.0989735574,-0.1910018027,0.1689160019,-0.0160105582,0.4543611407,0.3557760417,-0.0146972686,0.0548875332,-0.0971449018,-0.0204035416,0.3564491868,-0.1523542404,-0.097025387,0.1846250147,0.3677533865,0.0412646942,0.1355872005,-0.2689230442,0.2020162791,-0.2849611044,-0.2672598958,0.0389268734,-0.0312434658,-0.053507641,-0.0286826678,0.0271092784,-0.1850095838,0.3126622438,0.4369837642,-0.1364959925,-0.3239024878,-0.6861960292,-0.6801717281,0.3550527096,-0.0194597989,-0.1567206681,-0.0008682099,-0.1191424131,-0.0471941046,0.0927399695,0.3334364295,-0.0994437411,-0.1262345612,-0.1114645153,-0.1337472796,0.0834585503,0.1662860066,-0.1219962165,-0.0319091529,-0.2020152062,-0.1274804473,0.0751948282,-0.014412974,-0.2027576715,-0.2078402787,-0.0922850296,0.6267009974,0.2511222661,0.1438152492,0.5188578367,0.0033948012,-0.1690084487,-0.158147946,-0.168882966,-0.1216177046,0.304931581,0.0201929696,0.666085422,-0.1472443938,0.4769164324,-0.1770951301,0.5503355861,-0.1630852222,-0.1171028316,-0.380998075,0.2751182616,-0.1488752663,0.3362069726,-0.0528538749,0.1610537618,0.3596596718,0.4432800412,-0.6372888684,-0.097549431,0.6288667917,-0.3102293909,-0.2225486785,-0.1355645657,0.2629569471,-0.3097366691,-0.3672787845,-0.4621838629,-0.1992236823,0.1436507255,0.2374666929,-0.0912956744,0.1890782416,-0.4163116217,0.1561918408,0.0135606779,-0.0843809247,0.0612948202,-0.3806418478,-0.2585653961,-0.2395602763]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/539","title":"[Dataset] `NonMatchingChecksumError` due to an update in the LinCE benchmark data","comments":"Hi @thomwolf \r\n\r\nThanks for the details! I just created a PR with the updated `dataset_infos.json` file (#550).","body":"Hi,\r\n\r\nThere is a `NonMatchingChecksumError` error for the `lid_msaea` (language identification for Modern Standard Arabic - Egyptian Arabic) dataset from the LinCE benchmark due to a minor update on that dataset. \r\n\r\nHow can I update the checksum of the library to solve this issue? The error is below and it also appears in the [nlp viewer](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=lince&config=lid_msaea):\r\n\r\n```python\r\nimport nlp\r\nnlp.load_dataset('lince', 'lid_msaea')\r\n```\r\n\r\nOutput:\r\n```\r\nNonMatchingChecksumError: ['https:\/\/ritual.uh.edu\/lince\/libaccess\/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9\/lid_msaea.zip']\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 196, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 150, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n download_config.force_download = download_mode == FORCE_REDOWNLOAD\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 469, in _download_and_prepare\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 36, in verify_checksums\r\n raise NonMatchingChecksumError(str(bad_urls))\r\n```\r\n\r\nThank you in advance!\r\n\r\n@lhoestq ","comment_length":17,"text":"[Dataset] `NonMatchingChecksumError` due to an update in the LinCE benchmark data\nHi,\r\n\r\nThere is a `NonMatchingChecksumError` error for the `lid_msaea` (language identification for Modern Standard Arabic - Egyptian Arabic) dataset from the LinCE benchmark due to a minor update on that dataset. \r\n\r\nHow can I update the checksum of the library to solve this issue? The error is below and it also appears in the [nlp viewer](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=lince&config=lid_msaea):\r\n\r\n```python\r\nimport nlp\r\nnlp.load_dataset('lince', 'lid_msaea')\r\n```\r\n\r\nOutput:\r\n```\r\nNonMatchingChecksumError: ['https:\/\/ritual.uh.edu\/lince\/libaccess\/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9\/lid_msaea.zip']\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 196, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 150, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n download_config.force_download = download_mode == FORCE_REDOWNLOAD\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 469, in _download_and_prepare\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 36, in verify_checksums\r\n raise NonMatchingChecksumError(str(bad_urls))\r\n```\r\n\r\nThank you in advance!\r\n\r\n@lhoestq \nHi @thomwolf \r\n\r\nThanks for the details! I just created a PR with the updated `dataset_infos.json` file (#550).","embeddings":[-0.0616527833,0.4437022507,-0.0581613593,0.0664172098,-0.307210058,0.1153274924,-0.2732498646,0.5395020843,-0.0016354375,-0.1196584553,0.0857631639,0.2251154929,0.0789773464,-0.1222939044,0.0061919261,0.1934710592,0.0321503244,0.0933852121,0.0987019464,0.0393903553,-0.2053777874,0.2148771733,-0.057483878,-0.2985480428,0.0893565938,-0.004181175,0.0692205578,0.0182900559,0.1299763471,-0.4194515646,0.3206056654,0.2793268561,0.0003896341,0.1324239075,-0.0001272806,-0.0694950148,0.3210234344,-0.0919738933,-0.2670393586,-0.0846346244,0.0034744749,-0.4629070163,-0.0613037869,-0.2885442674,0.1637886763,0.2399886698,0.1029127687,-0.4139173031,-0.095096156,0.1851278096,0.1083362922,0.5190732479,0.0809718668,0.3380495608,0.256370753,-0.42114079,0.0306582432,0.3837837577,0.3393696845,-0.1366900653,-0.4413134158,0.3681635559,-0.3522348702,0.0423129387,-0.0161721241,-0.3124465644,0.0194569509,-0.0667776987,0.302544415,0.1907632947,0.3051871955,-0.4455226064,-0.2008479834,0.0067092571,-0.2116422057,-0.1155177131,0.4846940637,0.1011212692,-0.0301548149,-0.1416909546,-0.0993822813,-0.0432990901,0.1072188467,0.2136925161,0.3474704325,0.4857030511,0.243153125,0.1118917316,0.1624684483,0.0055500744,0.503978312,-0.0932113156,-0.1411801726,0.1578629613,-0.3675528467,-0.1388427168,0.0753963515,0.3629109859,0.1303834319,0.4638726413,0.5369125009,0.247657612,-0.1883206666,0.1930721104,-0.1384759098,0.2871352434,0.4059982002,-0.1246177554,0.1441257894,0.4248977005,0.0256022699,0.065224953,0.1155529246,-0.2771501243,0.3135475516,0.4061032236,0.0826807097,-0.4132894278,-0.227190122,0.307826966,-0.1956929564,-0.3644892573,0.211998567,0.1644034833,-0.3546676934,0.2263011485,0.1475031525,0.0085846949,-0.1239643097,-0.5568845868,-0.0674783885,0.2780528069,-0.1873663217,0.1560589522,0.2345919013,-0.3751237094,0.4669339359,-0.3525468111,0.0226076655,-0.0944171771,0.2752314806,-0.1536362767,0.0063340105,0.2864765227,-0.0740386918,-0.1444516182,0.1842181534,0.1451916844,-0.3300545812,0.2802166045,0.1353553981,-0.1413926184,-0.1740648896,-0.0231370293,-0.4117465615,-0.2353129536,0.272849828,-0.0858844668,0.390968442,-0.6479693055,0.0522488616,-0.0388518497,-0.2448469102,-0.1984824985,-0.1250327229,0.0340124369,-0.1752650887,0.0889368281,-0.2065021992,0.1218558773,0.301831007,0.1827165931,0.059068989,-0.2488514483,-0.1102877334,0.1323599666,0.2490991056,-0.3191206157,-0.6921525002,-0.0797567293,0.1015781909,0.3253774941,0.2345360368,0.4361954927,-0.1628785282,-0.2904224396,-0.0313170739,0.1885817349,-0.0350644253,-0.0714681223,-0.3921247721,-0.1106406301,0.5248131156,0.0953375474,0.0534198955,-0.0261638723,0.1034081951,-0.0051165503,0.2173566222,0.0385111012,0.0118800262,0.0543003716,0.0686132163,0.0348475017,0.0741711408,-0.0369870961,-0.4227679074,0.1104315296,-0.2665937245,0.2791252434,0.0841261968,-0.1472602785,-0.0798559189,-0.2010103911,0.008968005,0.0517491885,0.0013107359,0.3164534271,0.1445411444,-0.0932148546,0.0056738132,0.036909584,-0.2276382744,-0.0102180643,-0.5147677064,0.2930194736,-0.0692704618,-0.2226178199,0.2267709523,0.2620928288,0.1244557053,-0.067651242,-0.3514721692,0.379909277,-0.049288936,-0.139429003,0.0233200267,0.4658017457,-0.1902141124,-0.1928856522,0.2801755071,0.7407286763,-0.0178896319,-0.0470744595,0.188155219,0.4245472848,-0.0047320961,0.1563244313,-0.0887772217,0.1785681695,0.1582686752,-0.2391062677,-0.256729424,-0.0941786394,0.4943873882,-0.148961395,0.178952679,0.0457770936,-0.1842954308,-0.0852906406,0.3507559299,0.0191176645,0.0700724646,0.2472043037,0.1055887192,-0.3084563613,0.1519667804,0.34983477,0.2286102325,-0.0166477971,0.006565324,0.2571461201,-0.0707734674,0.0059679616,0.0090975491,0.0203417037,0.2311230451,0.4381026924,0.2708524466,0.0421677716,-0.4537835121,-0.2937007248,-0.0823328644,0.3356816173,-0.0985868648,0.0412338115,-0.2519593835,-0.3669678867,-0.1616517901,-0.3328652084,-0.5280544162,-0.3471707702,0.0056135268,0.3331266046,-0.0336648524,0.3375812769,-0.2607325315,-0.041509334,-0.2104506344,-0.055516988,-0.0397895426,-0.3035803735,-0.3481827378,-0.100549072,0.6255894899,0.1226193532,0.3256288469,-0.2633631825,-0.0386355855,-0.4359060526,-0.4488388598,-0.2283978611,-0.0557631403,0.0680425093,0.3516705632,0.0911212862,-0.0205849521,-0.3282206953,0.1320411414,-0.0937089622,-0.1445062608,0.0734271407,-0.1742323339,0.1782113463,0.1171157733,-0.1628405005,-0.0572150797,-0.2265317887,-0.0431701057,-0.1390309632,0.3625038266,0.0243449211,-0.3539898098,-0.0659313425,0.0604897924,0.4428186417,-0.1127821505,-0.5371189117,0.183861807,0.1153431833,-0.1968841404,0.0928090662,-0.1693492532,0.6429797411,-0.092325367,-0.509729445,-0.0690603107,-0.1017558351,-0.0644859299,-0.0095773246,-0.0312164482,0.4292153716,-0.0215043649,-0.0031932807,-0.2097126395,-0.4482651353,0.1087318808,0.0067476174,0.437325418,0.073648639,0.3563734889,0.170651257,0.0788869634,0.1601974666,-0.162990272,0.4158077538,0.4866502881,0.1476314217,0.0428027995,-0.0988665894,-0.1026581004,-0.0358230583,-0.0471012406,0.0812134668,-0.057063546,-0.1882273108,-0.116942361,-0.3559015989,-0.3291288018,-0.2423394322,0.1743614674,-0.3096680939,0.3644334972,0.2393897623,-0.1058761254,-0.3316639662,-0.420277983,0.1958696395,0.3060153723,-0.1326766461,-0.0060011023,-0.550457418,-0.34876284,-0.1723356545,0.7032877803,0.2243926227,0.6603319049,0.0806642771,-0.0668382496,0.0195838772,-0.1764271855,0.0931476653,-0.365216881,0.2800974548,0.1581119597,0.0958499089,-0.0339047648,-0.2032892555,0.1098054051,0.2863220572,0.4885847867,-0.178142339,-0.2392450124,-0.121640332,0.2182543576,0.2945373058,-0.0311449543,-0.2273663133,-0.4379222095,-0.2903141081,-0.114300631,-0.0357312597,-0.2543928623,0.320625633,0.2463235408,0.177491352,0.1452334076,-0.2266955674,0.3136151433,-0.0948358923,0.2140125483,0.3180748522,0.1195525751,-0.3462470174,0.130244419,-0.0941932723,0.2956397831,-0.1126083434,-0.1375366449,-0.0734804794,0.2040028274,0.0879630968,0.1910957098,0.2005631477,0.2737550437,0.4517329633,-0.1058313847,0.1607081145,-0.1901838034,-0.0732441768,0.1468066275,-0.2490932643,-0.2411087006,0.2625308037,0.1931416541,-0.0387197845,0.3418625891,-0.4028802216,-0.1004320756,0.2029738277,0.4714302421,1.0572549105,0.0201435611,0.0580873601,-0.050939858,-0.1898695081,0.0285251606,-0.0933944061,0.2850465775,-0.4167386591,-0.4778108299,-0.0199747197,-0.1191670597,0.3133391738,-0.2553495765,-0.2167989016,0.41465801,-0.0257352516,-0.4536561668,0.1333514303,0.2383010685,-0.425064832,-0.0571774803,-0.1014586166,0.0327990949,-0.3412192464,0.3648760021,-0.1134908721,0.0406587645,0.1785674095,-0.3359895647,-0.3500402272,0.0750798509,-0.0747927055,0.0608153604,0.3293658793,0.1280973852,-0.0595258772,-0.1196693555,0.3529457152,0.0838501155,0.0882170871,0.0468815304,0.49015522,-0.0665743724,0.3549170196,0.2023560554,0.3399666846,-0.1613616198,-0.1293440312,0.0918295458,-0.123580642,-0.5137318969,0.3060742319,-0.2194625437,0.0945610851,0.0228990261,-0.1769923866,0.205457747,-0.0713040233,-0.2482821047,0.0278570279,0.1989021003,-0.3135830462,0.1156932488,-0.0807046816,-0.2498116493,-0.0033673241,0.4579544961,0.3928121328,-0.2363060266,0.3262052536,0.0302985553,-0.1058911532,-0.1204885095,-0.2305814624,-0.1916443259,-0.2923538685,0.1083471105,-0.0822686255,-0.075918518,-0.1271994263,0.5551509261,0.2674683928,0.6689737439,-0.12501131,-0.498616308,-0.1051060259,0.1533489376,0.0695057958,0.1367290616,-0.289671123,0.0720115378,-0.0632774085,-0.0139155621,-0.1498123407,0.0530563369,-0.2679403424,0.1874872595,-0.0079384195,0.16731821,-0.0482473932,0.1539581865,-0.0541640557,-0.0090228152,-0.244091928,-0.010713323,-0.0693433359,0.1767442375,0.0797735229,-0.0332644545,-0.1182662398,-0.3067573011,-0.0969026387,-0.0597471595,-0.4343479574,-0.0200659856,-0.1304022223,0.2618319392,-0.0712140724,0.0125173153,0.1222654134,-0.0130180223,-0.3334136903,-0.1952343285,-0.098141849,0.1256816834,-0.0123606939,-0.0134816281,0.2143861204,0.0588422567,-0.2207695693,-0.1099382117,-0.0495147519,0.00431058,-0.058556892,0.4181289077,0.1668494642,0.3067296743,0.0485336855,-0.3271096051,0.358933717,0.0716522858,-0.5490003228,0.0469181165,0.2238919884,-0.0183319878,0.0063809059,0.0733909532,0.2378414869,-0.1913523227,0.6728568077,0.068653591,-0.0561068282,-0.3147752583,0.3004775047,0.1758289039,0.0488543585,0.0803689286,0.1535321921,-0.2527367473,0.0332416743,-0.2883599401,-0.044377137,0.4211072326,0.05685772,0.3054702282,-0.4572608769,-0.2989241779,0.2259770036,0.4199058712,0.4424139857,0.0849668682,0.2154006064,0.5814492702,0.0252286009,0.0394732319,-0.2616809309,0.1191078871,0.0596153103,0.0078946827,-0.163883999,-0.0199991167,-0.0829944387,0.0370736197,0.1350004822,0.384144187,0.2583817542,0.039419584,-0.0428873338,-0.3136141598,-0.4687738717,0.0511496328,-0.1495767534,-0.0921679139,0.3379406631,0.311460048,-0.0182744265,0.0465470329,0.3662718832,0.260745585,0.149502784,-0.0323063731,-0.0989735574,-0.1910018027,0.1689160019,-0.0160105582,0.4543611407,0.3557760417,-0.0146972686,0.0548875332,-0.0971449018,-0.0204035416,0.3564491868,-0.1523542404,-0.097025387,0.1846250147,0.3677533865,0.0412646942,0.1355872005,-0.2689230442,0.2020162791,-0.2849611044,-0.2672598958,0.0389268734,-0.0312434658,-0.053507641,-0.0286826678,0.0271092784,-0.1850095838,0.3126622438,0.4369837642,-0.1364959925,-0.3239024878,-0.6861960292,-0.6801717281,0.3550527096,-0.0194597989,-0.1567206681,-0.0008682099,-0.1191424131,-0.0471941046,0.0927399695,0.3334364295,-0.0994437411,-0.1262345612,-0.1114645153,-0.1337472796,0.0834585503,0.1662860066,-0.1219962165,-0.0319091529,-0.2020152062,-0.1274804473,0.0751948282,-0.014412974,-0.2027576715,-0.2078402787,-0.0922850296,0.6267009974,0.2511222661,0.1438152492,0.5188578367,0.0033948012,-0.1690084487,-0.158147946,-0.168882966,-0.1216177046,0.304931581,0.0201929696,0.666085422,-0.1472443938,0.4769164324,-0.1770951301,0.5503355861,-0.1630852222,-0.1171028316,-0.380998075,0.2751182616,-0.1488752663,0.3362069726,-0.0528538749,0.1610537618,0.3596596718,0.4432800412,-0.6372888684,-0.097549431,0.6288667917,-0.3102293909,-0.2225486785,-0.1355645657,0.2629569471,-0.3097366691,-0.3672787845,-0.4621838629,-0.1992236823,0.1436507255,0.2374666929,-0.0912956744,0.1890782416,-0.4163116217,0.1561918408,0.0135606779,-0.0843809247,0.0612948202,-0.3806418478,-0.2585653961,-0.2395602763]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/539","title":"[Dataset] `NonMatchingChecksumError` due to an update in the LinCE benchmark data","comments":"Thanks for updating the json file. Closing this one","body":"Hi,\r\n\r\nThere is a `NonMatchingChecksumError` error for the `lid_msaea` (language identification for Modern Standard Arabic - Egyptian Arabic) dataset from the LinCE benchmark due to a minor update on that dataset. \r\n\r\nHow can I update the checksum of the library to solve this issue? The error is below and it also appears in the [nlp viewer](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=lince&config=lid_msaea):\r\n\r\n```python\r\nimport nlp\r\nnlp.load_dataset('lince', 'lid_msaea')\r\n```\r\n\r\nOutput:\r\n```\r\nNonMatchingChecksumError: ['https:\/\/ritual.uh.edu\/lince\/libaccess\/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9\/lid_msaea.zip']\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 196, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 150, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n download_config.force_download = download_mode == FORCE_REDOWNLOAD\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 469, in _download_and_prepare\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 36, in verify_checksums\r\n raise NonMatchingChecksumError(str(bad_urls))\r\n```\r\n\r\nThank you in advance!\r\n\r\n@lhoestq ","comment_length":9,"text":"[Dataset] `NonMatchingChecksumError` due to an update in the LinCE benchmark data\nHi,\r\n\r\nThere is a `NonMatchingChecksumError` error for the `lid_msaea` (language identification for Modern Standard Arabic - Egyptian Arabic) dataset from the LinCE benchmark due to a minor update on that dataset. \r\n\r\nHow can I update the checksum of the library to solve this issue? The error is below and it also appears in the [nlp viewer](https:\/\/huggingface.co\/nlp\/viewer\/?dataset=lince&config=lid_msaea):\r\n\r\n```python\r\nimport nlp\r\nnlp.load_dataset('lince', 'lid_msaea')\r\n```\r\n\r\nOutput:\r\n```\r\nNonMatchingChecksumError: ['https:\/\/ritual.uh.edu\/lince\/libaccess\/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9\/lid_msaea.zip']\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 196, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 150, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n download_config.force_download = download_mode == FORCE_REDOWNLOAD\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 469, in _download_and_prepare\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 36, in verify_checksums\r\n raise NonMatchingChecksumError(str(bad_urls))\r\n```\r\n\r\nThank you in advance!\r\n\r\n@lhoestq \nThanks for updating the json file. Closing this one","embeddings":[-0.0616527833,0.4437022507,-0.0581613593,0.0664172098,-0.307210058,0.1153274924,-0.2732498646,0.5395020843,-0.0016354375,-0.1196584553,0.0857631639,0.2251154929,0.0789773464,-0.1222939044,0.0061919261,0.1934710592,0.0321503244,0.0933852121,0.0987019464,0.0393903553,-0.2053777874,0.2148771733,-0.057483878,-0.2985480428,0.0893565938,-0.004181175,0.0692205578,0.0182900559,0.1299763471,-0.4194515646,0.3206056654,0.2793268561,0.0003896341,0.1324239075,-0.0001272806,-0.0694950148,0.3210234344,-0.0919738933,-0.2670393586,-0.0846346244,0.0034744749,-0.4629070163,-0.0613037869,-0.2885442674,0.1637886763,0.2399886698,0.1029127687,-0.4139173031,-0.095096156,0.1851278096,0.1083362922,0.5190732479,0.0809718668,0.3380495608,0.256370753,-0.42114079,0.0306582432,0.3837837577,0.3393696845,-0.1366900653,-0.4413134158,0.3681635559,-0.3522348702,0.0423129387,-0.0161721241,-0.3124465644,0.0194569509,-0.0667776987,0.302544415,0.1907632947,0.3051871955,-0.4455226064,-0.2008479834,0.0067092571,-0.2116422057,-0.1155177131,0.4846940637,0.1011212692,-0.0301548149,-0.1416909546,-0.0993822813,-0.0432990901,0.1072188467,0.2136925161,0.3474704325,0.4857030511,0.243153125,0.1118917316,0.1624684483,0.0055500744,0.503978312,-0.0932113156,-0.1411801726,0.1578629613,-0.3675528467,-0.1388427168,0.0753963515,0.3629109859,0.1303834319,0.4638726413,0.5369125009,0.247657612,-0.1883206666,0.1930721104,-0.1384759098,0.2871352434,0.4059982002,-0.1246177554,0.1441257894,0.4248977005,0.0256022699,0.065224953,0.1155529246,-0.2771501243,0.3135475516,0.4061032236,0.0826807097,-0.4132894278,-0.227190122,0.307826966,-0.1956929564,-0.3644892573,0.211998567,0.1644034833,-0.3546676934,0.2263011485,0.1475031525,0.0085846949,-0.1239643097,-0.5568845868,-0.0674783885,0.2780528069,-0.1873663217,0.1560589522,0.2345919013,-0.3751237094,0.4669339359,-0.3525468111,0.0226076655,-0.0944171771,0.2752314806,-0.1536362767,0.0063340105,0.2864765227,-0.0740386918,-0.1444516182,0.1842181534,0.1451916844,-0.3300545812,0.2802166045,0.1353553981,-0.1413926184,-0.1740648896,-0.0231370293,-0.4117465615,-0.2353129536,0.272849828,-0.0858844668,0.390968442,-0.6479693055,0.0522488616,-0.0388518497,-0.2448469102,-0.1984824985,-0.1250327229,0.0340124369,-0.1752650887,0.0889368281,-0.2065021992,0.1218558773,0.301831007,0.1827165931,0.059068989,-0.2488514483,-0.1102877334,0.1323599666,0.2490991056,-0.3191206157,-0.6921525002,-0.0797567293,0.1015781909,0.3253774941,0.2345360368,0.4361954927,-0.1628785282,-0.2904224396,-0.0313170739,0.1885817349,-0.0350644253,-0.0714681223,-0.3921247721,-0.1106406301,0.5248131156,0.0953375474,0.0534198955,-0.0261638723,0.1034081951,-0.0051165503,0.2173566222,0.0385111012,0.0118800262,0.0543003716,0.0686132163,0.0348475017,0.0741711408,-0.0369870961,-0.4227679074,0.1104315296,-0.2665937245,0.2791252434,0.0841261968,-0.1472602785,-0.0798559189,-0.2010103911,0.008968005,0.0517491885,0.0013107359,0.3164534271,0.1445411444,-0.0932148546,0.0056738132,0.036909584,-0.2276382744,-0.0102180643,-0.5147677064,0.2930194736,-0.0692704618,-0.2226178199,0.2267709523,0.2620928288,0.1244557053,-0.067651242,-0.3514721692,0.379909277,-0.049288936,-0.139429003,0.0233200267,0.4658017457,-0.1902141124,-0.1928856522,0.2801755071,0.7407286763,-0.0178896319,-0.0470744595,0.188155219,0.4245472848,-0.0047320961,0.1563244313,-0.0887772217,0.1785681695,0.1582686752,-0.2391062677,-0.256729424,-0.0941786394,0.4943873882,-0.148961395,0.178952679,0.0457770936,-0.1842954308,-0.0852906406,0.3507559299,0.0191176645,0.0700724646,0.2472043037,0.1055887192,-0.3084563613,0.1519667804,0.34983477,0.2286102325,-0.0166477971,0.006565324,0.2571461201,-0.0707734674,0.0059679616,0.0090975491,0.0203417037,0.2311230451,0.4381026924,0.2708524466,0.0421677716,-0.4537835121,-0.2937007248,-0.0823328644,0.3356816173,-0.0985868648,0.0412338115,-0.2519593835,-0.3669678867,-0.1616517901,-0.3328652084,-0.5280544162,-0.3471707702,0.0056135268,0.3331266046,-0.0336648524,0.3375812769,-0.2607325315,-0.041509334,-0.2104506344,-0.055516988,-0.0397895426,-0.3035803735,-0.3481827378,-0.100549072,0.6255894899,0.1226193532,0.3256288469,-0.2633631825,-0.0386355855,-0.4359060526,-0.4488388598,-0.2283978611,-0.0557631403,0.0680425093,0.3516705632,0.0911212862,-0.0205849521,-0.3282206953,0.1320411414,-0.0937089622,-0.1445062608,0.0734271407,-0.1742323339,0.1782113463,0.1171157733,-0.1628405005,-0.0572150797,-0.2265317887,-0.0431701057,-0.1390309632,0.3625038266,0.0243449211,-0.3539898098,-0.0659313425,0.0604897924,0.4428186417,-0.1127821505,-0.5371189117,0.183861807,0.1153431833,-0.1968841404,0.0928090662,-0.1693492532,0.6429797411,-0.092325367,-0.509729445,-0.0690603107,-0.1017558351,-0.0644859299,-0.0095773246,-0.0312164482,0.4292153716,-0.0215043649,-0.0031932807,-0.2097126395,-0.4482651353,0.1087318808,0.0067476174,0.437325418,0.073648639,0.3563734889,0.170651257,0.0788869634,0.1601974666,-0.162990272,0.4158077538,0.4866502881,0.1476314217,0.0428027995,-0.0988665894,-0.1026581004,-0.0358230583,-0.0471012406,0.0812134668,-0.057063546,-0.1882273108,-0.116942361,-0.3559015989,-0.3291288018,-0.2423394322,0.1743614674,-0.3096680939,0.3644334972,0.2393897623,-0.1058761254,-0.3316639662,-0.420277983,0.1958696395,0.3060153723,-0.1326766461,-0.0060011023,-0.550457418,-0.34876284,-0.1723356545,0.7032877803,0.2243926227,0.6603319049,0.0806642771,-0.0668382496,0.0195838772,-0.1764271855,0.0931476653,-0.365216881,0.2800974548,0.1581119597,0.0958499089,-0.0339047648,-0.2032892555,0.1098054051,0.2863220572,0.4885847867,-0.178142339,-0.2392450124,-0.121640332,0.2182543576,0.2945373058,-0.0311449543,-0.2273663133,-0.4379222095,-0.2903141081,-0.114300631,-0.0357312597,-0.2543928623,0.320625633,0.2463235408,0.177491352,0.1452334076,-0.2266955674,0.3136151433,-0.0948358923,0.2140125483,0.3180748522,0.1195525751,-0.3462470174,0.130244419,-0.0941932723,0.2956397831,-0.1126083434,-0.1375366449,-0.0734804794,0.2040028274,0.0879630968,0.1910957098,0.2005631477,0.2737550437,0.4517329633,-0.1058313847,0.1607081145,-0.1901838034,-0.0732441768,0.1468066275,-0.2490932643,-0.2411087006,0.2625308037,0.1931416541,-0.0387197845,0.3418625891,-0.4028802216,-0.1004320756,0.2029738277,0.4714302421,1.0572549105,0.0201435611,0.0580873601,-0.050939858,-0.1898695081,0.0285251606,-0.0933944061,0.2850465775,-0.4167386591,-0.4778108299,-0.0199747197,-0.1191670597,0.3133391738,-0.2553495765,-0.2167989016,0.41465801,-0.0257352516,-0.4536561668,0.1333514303,0.2383010685,-0.425064832,-0.0571774803,-0.1014586166,0.0327990949,-0.3412192464,0.3648760021,-0.1134908721,0.0406587645,0.1785674095,-0.3359895647,-0.3500402272,0.0750798509,-0.0747927055,0.0608153604,0.3293658793,0.1280973852,-0.0595258772,-0.1196693555,0.3529457152,0.0838501155,0.0882170871,0.0468815304,0.49015522,-0.0665743724,0.3549170196,0.2023560554,0.3399666846,-0.1613616198,-0.1293440312,0.0918295458,-0.123580642,-0.5137318969,0.3060742319,-0.2194625437,0.0945610851,0.0228990261,-0.1769923866,0.205457747,-0.0713040233,-0.2482821047,0.0278570279,0.1989021003,-0.3135830462,0.1156932488,-0.0807046816,-0.2498116493,-0.0033673241,0.4579544961,0.3928121328,-0.2363060266,0.3262052536,0.0302985553,-0.1058911532,-0.1204885095,-0.2305814624,-0.1916443259,-0.2923538685,0.1083471105,-0.0822686255,-0.075918518,-0.1271994263,0.5551509261,0.2674683928,0.6689737439,-0.12501131,-0.498616308,-0.1051060259,0.1533489376,0.0695057958,0.1367290616,-0.289671123,0.0720115378,-0.0632774085,-0.0139155621,-0.1498123407,0.0530563369,-0.2679403424,0.1874872595,-0.0079384195,0.16731821,-0.0482473932,0.1539581865,-0.0541640557,-0.0090228152,-0.244091928,-0.010713323,-0.0693433359,0.1767442375,0.0797735229,-0.0332644545,-0.1182662398,-0.3067573011,-0.0969026387,-0.0597471595,-0.4343479574,-0.0200659856,-0.1304022223,0.2618319392,-0.0712140724,0.0125173153,0.1222654134,-0.0130180223,-0.3334136903,-0.1952343285,-0.098141849,0.1256816834,-0.0123606939,-0.0134816281,0.2143861204,0.0588422567,-0.2207695693,-0.1099382117,-0.0495147519,0.00431058,-0.058556892,0.4181289077,0.1668494642,0.3067296743,0.0485336855,-0.3271096051,0.358933717,0.0716522858,-0.5490003228,0.0469181165,0.2238919884,-0.0183319878,0.0063809059,0.0733909532,0.2378414869,-0.1913523227,0.6728568077,0.068653591,-0.0561068282,-0.3147752583,0.3004775047,0.1758289039,0.0488543585,0.0803689286,0.1535321921,-0.2527367473,0.0332416743,-0.2883599401,-0.044377137,0.4211072326,0.05685772,0.3054702282,-0.4572608769,-0.2989241779,0.2259770036,0.4199058712,0.4424139857,0.0849668682,0.2154006064,0.5814492702,0.0252286009,0.0394732319,-0.2616809309,0.1191078871,0.0596153103,0.0078946827,-0.163883999,-0.0199991167,-0.0829944387,0.0370736197,0.1350004822,0.384144187,0.2583817542,0.039419584,-0.0428873338,-0.3136141598,-0.4687738717,0.0511496328,-0.1495767534,-0.0921679139,0.3379406631,0.311460048,-0.0182744265,0.0465470329,0.3662718832,0.260745585,0.149502784,-0.0323063731,-0.0989735574,-0.1910018027,0.1689160019,-0.0160105582,0.4543611407,0.3557760417,-0.0146972686,0.0548875332,-0.0971449018,-0.0204035416,0.3564491868,-0.1523542404,-0.097025387,0.1846250147,0.3677533865,0.0412646942,0.1355872005,-0.2689230442,0.2020162791,-0.2849611044,-0.2672598958,0.0389268734,-0.0312434658,-0.053507641,-0.0286826678,0.0271092784,-0.1850095838,0.3126622438,0.4369837642,-0.1364959925,-0.3239024878,-0.6861960292,-0.6801717281,0.3550527096,-0.0194597989,-0.1567206681,-0.0008682099,-0.1191424131,-0.0471941046,0.0927399695,0.3334364295,-0.0994437411,-0.1262345612,-0.1114645153,-0.1337472796,0.0834585503,0.1662860066,-0.1219962165,-0.0319091529,-0.2020152062,-0.1274804473,0.0751948282,-0.014412974,-0.2027576715,-0.2078402787,-0.0922850296,0.6267009974,0.2511222661,0.1438152492,0.5188578367,0.0033948012,-0.1690084487,-0.158147946,-0.168882966,-0.1216177046,0.304931581,0.0201929696,0.666085422,-0.1472443938,0.4769164324,-0.1770951301,0.5503355861,-0.1630852222,-0.1171028316,-0.380998075,0.2751182616,-0.1488752663,0.3362069726,-0.0528538749,0.1610537618,0.3596596718,0.4432800412,-0.6372888684,-0.097549431,0.6288667917,-0.3102293909,-0.2225486785,-0.1355645657,0.2629569471,-0.3097366691,-0.3672787845,-0.4621838629,-0.1992236823,0.1436507255,0.2374666929,-0.0912956744,0.1890782416,-0.4163116217,0.1561918408,0.0135606779,-0.0843809247,0.0612948202,-0.3806418478,-0.2585653961,-0.2395602763]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"`NonMatchingChecksumError` means that the checksum of the downloaded file is not the expected one.\r\nEither the file you downloaded was corrupted along the way, or the host updated the file.\r\nCould you try to clear your cache and run `load_dataset` again ? If the error is still there, it means that there was an update in the data, and we may have to update the expected checksum value.","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":68,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\n`NonMatchingChecksumError` means that the checksum of the downloaded file is not the expected one.\r\nEither the file you downloaded was corrupted along the way, or the host updated the file.\r\nCould you try to clear your cache and run `load_dataset` again ? If the error is still there, it means that there was an update in the data, and we may have to update the expected checksum value.","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"I just cleared the cache an run it again. The error persists ):\r\n\r\n```\r\n nlp (master) $ rm -rf \/Users\/abarbosa\/.cache\/huggingface\/\r\n nlp (master) $ python\r\nPython 3.8.5 (default, Aug 5 2020, 03:39:04)\r\n[Clang 10.0.0 ] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import nlp\r\n>>> dataset = nlp.load_dataset(\"race\")\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.39k\/4.39k [00:00<00:00, 661kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1.81k\/1.81k [00:00<00:00, 644kB\/s]\r\nUsing custom data configuration default\r\nDownloading and preparing dataset race\/default (download: 84.52 MiB, generated: 132.61 MiB, post-processed: Unknown size, total: 217.13 MiB) to \/Users\/abarbosa\/.cache\/huggingface\/datasets\/race\/default\/0.1.0\/5461327f1a83549ca0d845a3159c806d2baf4f8d0d8f7d657157ce7cdf3899c2...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 25.4M\/25.4M [01:03<00:00, 401kB\/s]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/abarbosa\/Documents\/nlp\/src\/nlp\/load.py\", line 550, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/Users\/abarbosa\/Documents\/nlp\/src\/nlp\/builder.py\", line 471, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/Users\/abarbosa\/Documents\/nlp\/src\/nlp\/builder.py\", line 530, in _download_and_prepare\r\n verify_checksums(\r\n File \"\/Users\/abarbosa\/Documents\/nlp\/src\/nlp\/utils\/info_utils.py\", line 38, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\nnlp.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n>>>\r\n```","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":147,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\nI just cleared the cache an run it again. The error persists ):\r\n\r\n```\r\n nlp (master) $ rm -rf \/Users\/abarbosa\/.cache\/huggingface\/\r\n nlp (master) $ python\r\nPython 3.8.5 (default, Aug 5 2020, 03:39:04)\r\n[Clang 10.0.0 ] :: Anaconda, Inc. on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import nlp\r\n>>> dataset = nlp.load_dataset(\"race\")\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.39k\/4.39k [00:00<00:00, 661kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1.81k\/1.81k [00:00<00:00, 644kB\/s]\r\nUsing custom data configuration default\r\nDownloading and preparing dataset race\/default (download: 84.52 MiB, generated: 132.61 MiB, post-processed: Unknown size, total: 217.13 MiB) to \/Users\/abarbosa\/.cache\/huggingface\/datasets\/race\/default\/0.1.0\/5461327f1a83549ca0d845a3159c806d2baf4f8d0d8f7d657157ce7cdf3899c2...\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 25.4M\/25.4M [01:03<00:00, 401kB\/s]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/abarbosa\/Documents\/nlp\/src\/nlp\/load.py\", line 550, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/Users\/abarbosa\/Documents\/nlp\/src\/nlp\/builder.py\", line 471, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/Users\/abarbosa\/Documents\/nlp\/src\/nlp\/builder.py\", line 530, in _download_and_prepare\r\n verify_checksums(\r\n File \"\/Users\/abarbosa\/Documents\/nlp\/src\/nlp\/utils\/info_utils.py\", line 38, in verify_checksums\r\n raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\nnlp.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n>>>\r\n```","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"Dealing with the same issue please update the checksum on nlp library end. The data seems to have changed on their end.","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":22,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\nDealing with the same issue please update the checksum on nlp library end. The data seems to have changed on their end.","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"We have a discussion on this datasets here: https:\/\/github.com\/huggingface\/nlp\/pull\/540\r\n\r\nFeel free to participate if you have some opinion on the scope of data which should be included in this dataset.","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":30,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\nWe have a discussion on this datasets here: https:\/\/github.com\/huggingface\/nlp\/pull\/540\r\n\r\nFeel free to participate if you have some opinion on the scope of data which should be included in this dataset.","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"At least for me, the file that was downloaded from CMU isn't the complete dataset, but a small subset of it (~25MB vs ~85MB). I've previously downloaded the dataset directly, so for my personal needs I could just swap out the corrupted file with the correct one. Perhaps you could host it like you do for the Wikipedia and BookCorpus datasets.\r\n\r\n","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":61,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\nAt least for me, the file that was downloaded from CMU isn't the complete dataset, but a small subset of it (~25MB vs ~85MB). I've previously downloaded the dataset directly, so for my personal needs I could just swap out the corrupted file with the correct one. Perhaps you could host it like you do for the Wikipedia and BookCorpus datasets.\r\n\r\n","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"> At least for me, the file that was downloaded from CMU isn't the complete dataset, but a small subset of it (~25MB vs ~85MB). I've previously downloaded the dataset directly, so for my personal needs I could just swap out the corrupted file with the correct one. Perhaps you could host it like you do for the Wikipedia and BookCorpus datasets.\r\n\r\nCould you upload this please?","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":67,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\n> At least for me, the file that was downloaded from CMU isn't the complete dataset, but a small subset of it (~25MB vs ~85MB). I've previously downloaded the dataset directly, so for my personal needs I could just swap out the corrupted file with the correct one. Perhaps you could host it like you do for the Wikipedia and BookCorpus datasets.\r\n\r\nCould you upload this please?","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"> > At least for me, the file that was downloaded from CMU isn't the complete dataset, but a small subset of it (~25MB vs ~85MB). I've previously downloaded the dataset directly, so for my personal needs I could just swap out the corrupted file with the correct one. Perhaps you could host it like you do for the Wikipedia and BookCorpus datasets.\r\n> \r\n> Could you upload this please?\r\n\r\nNot sure if I can upload it according to their license (\"You agree not to reproduce, duplicate, copy, sell, trade, resell or exploit for any commercial purpose, any portion of the contexts and any portion of derived data.\").","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":108,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\n> > At least for me, the file that was downloaded from CMU isn't the complete dataset, but a small subset of it (~25MB vs ~85MB). I've previously downloaded the dataset directly, so for my personal needs I could just swap out the corrupted file with the correct one. Perhaps you could host it like you do for the Wikipedia and BookCorpus datasets.\r\n> \r\n> Could you upload this please?\r\n\r\nNot sure if I can upload it according to their license (\"You agree not to reproduce, duplicate, copy, sell, trade, resell or exploit for any commercial purpose, any portion of the contexts and any portion of derived data.\").","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"I managed to fix it in #540 :)","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":8,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\nI managed to fix it in #540 :)","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/537","title":"[Dataset] RACE dataset Checksums error","comments":"Closing since @540 is merged\r\n\r\nThanks again @abarbosa94 ","body":"Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```","comment_length":8,"text":"[Dataset] RACE dataset Checksums error\nHi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps:\r\n\r\n```\r\ndataset = nlp.load_dataset(\"race\")\r\nlen(dataset[\"train\"]), len(dataset[\"validation\"])\r\n```\r\n\r\nBut then I got the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-15-8bf7603ce0ed> in <module>\r\n----> 1 dataset = nlp.load_dataset(\"race\")\r\n 2 len(dataset[\"train\"]), len(dataset[\"validation\"])\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 546 \r\n 547 # Download and prepare data\r\n--> 548 builder_instance.download_and_prepare(\r\n 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 460 logger.info(\"Dataset not on Hf google storage. Downloading and preparing it from source\")\r\n 461 if not downloaded_from_gcs:\r\n--> 462 self._download_and_prepare(\r\n 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n--> 521 verify_checksums(\r\n 522 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), \"dataset source files\"\r\n 523 )\r\n\r\n~\/miniconda3\/envs\/masters\/lib\/python3.8\/site-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name)\r\n 36 if len(bad_urls) > 0:\r\n 37 error_msg = \"Checksums didn't match\" + for_verification_name + \":\\n\"\r\n---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls))\r\n 39 logger.info(\"All the checksums matched successfully\" + for_verification_name)\r\n 40 \r\n\r\nNonMatchingChecksumError: Checksums didn't match for dataset source files:\r\n['http:\/\/www.cs.cmu.edu\/~glai1\/data\/race\/RACE.tar.gz']\r\n```\nClosing since @540 is merged\r\n\r\nThanks again @abarbosa94 ","embeddings":[-0.3484401107,0.3749217391,-0.0223544165,0.2424075007,0.2333744168,0.0591794364,0.2440315485,0.4674772024,0.2553084195,-0.1403686702,-0.0667250529,-0.0904459357,-0.3057557046,0.1228690147,0.0535939559,-0.0088576656,-0.0977109894,0.0319078714,-0.1895279884,0.0566937141,-0.1161630303,-0.0643239096,-0.1031743437,0.0158479381,-0.0917348042,-0.1198823452,0.2069767118,0.1899706721,-0.2297097296,-0.457950443,0.2975064218,0.1518295407,0.2854932845,0.4001206756,-0.0001180434,0.1198679805,0.4619445801,-0.0878783837,-0.4845319688,-0.3111259639,-0.2444009781,-0.0946979746,0.0279894862,-0.1625524312,-0.1754536331,0.3403500319,0.1200167835,-0.3093126714,0.2425986379,0.415902704,0.1814554483,0.4850332439,-0.0983780921,0.1419311464,0.2775052786,0.0405619107,-0.0977338254,0.4429739118,0.0883837417,-0.2895773947,-0.0681834817,0.1827246845,-0.1868143529,0.2293165624,0.2373239398,-0.031059809,0.050049074,-0.1472805589,0.0172663014,0.3044555187,0.228475675,-0.4627510011,-0.262774229,-0.198295325,-0.0720933005,-0.663425982,0.309194535,0.2011436969,-0.239018321,-0.1425317824,-0.5278380513,0.0906221122,-0.0025051723,0.116995424,0.1789869219,0.4105425179,0.2487375885,0.1001004428,0.1208602414,0.0906594023,0.1509335041,-0.2384599894,-0.1124067903,0.2280863971,-0.5045996308,-0.167716071,0.1644721627,0.1875035167,0.0313857719,0.3799568415,0.1418694258,0.084726125,0.0732125565,0.0911256745,0.2244704515,0.1387256682,0.1841384023,0.2849820256,0.384634763,0.1717725396,0.2062670141,0.1483997554,-0.0711589828,-0.1690804064,0.0967379138,0.0995636731,-0.0724148899,-0.3175649941,-0.550168097,0.2802394927,0.0591629855,-0.10878966,0.3846683204,0.2568905056,-0.2069443613,-0.3697853684,0.0991448984,0.2307722121,-0.2812793255,-0.0285786558,-0.1864865869,0.2175771296,-0.3917308152,-0.0531366691,0.3407814205,-0.0546818748,0.3931039572,-0.1745660603,0.0894263536,0.2351038456,0.1652139723,-0.2833355367,-0.372736454,0.4511757195,0.0721225217,-0.0609272681,0.2574218214,-0.1946486831,-0.3575665653,-0.0453447327,-0.0227217954,-0.1541518718,-0.2694600523,0.1336429864,-0.5319498181,-0.1265995502,-0.1047882885,-0.0980740264,0.4659026563,-0.2218079269,0.0720761493,-0.2355969846,-0.4272306859,-0.2113928646,0.2053787857,0.2566948831,-0.2136314958,-0.0465420075,-0.2215101272,-0.0849393308,0.3301586509,0.2536371946,-0.1534425318,0.2487071902,-0.237203747,0.1597365588,0.4282442331,-0.0966892466,-0.377505362,0.0635767803,0.0420482419,0.0293667056,-0.0294797532,0.1122540906,-0.0142485276,-0.067274861,0.1905495077,0.3296660483,-0.1435996741,-0.1567834467,-0.2640849948,0.0402008817,0.4529900849,0.3493306935,0.2753326297,-0.1353898197,0.0652282834,0.1668525487,0.1153825596,0.0573682748,0.0887684748,0.1663028598,0.1947500855,-0.0009249717,-0.3076634705,-0.1934029013,-0.428485781,0.2985518277,-0.4023680389,0.3442194164,-0.0052058366,0.1163645461,-0.2743442655,-0.0884458572,-0.0779779553,-0.0349813402,0.0552481115,-0.0362682566,0.118636325,-0.1575817913,-0.1216600239,0.4596948028,-0.3044737279,0.0306052417,-0.203237772,0.5589396954,-0.2494065017,-0.3483188748,0.1816463619,0.1678283066,-0.0748553053,-0.0887406096,-0.1464426816,0.4584696591,0.1846816242,-0.2115025371,0.0676155612,0.2728219628,0.2086811811,0.0457450002,0.1133401245,0.3485707939,-0.0103805559,-0.1062241271,-0.225029394,0.5668245554,-0.1356490254,0.1098725349,-0.0194805656,0.1030010805,0.1647879183,-0.1018060818,-0.1466487348,-0.0287135188,0.5365939736,0.1954443306,0.2387123257,0.0966354981,-0.3509839475,-0.1509943455,0.5099392533,0.0974254385,0.037500862,0.3271868825,0.1645047069,-0.2205292135,0.1525416374,0.0362701751,0.5619978905,0.0448566377,-0.0008600512,0.3147718906,-0.3628351986,-0.0569822639,0.0665312409,-0.158066377,0.3175257444,0.323420912,0.0579751097,0.0612480752,-0.3958386183,-0.2288500965,0.1042920575,0.4989686012,-0.2964100242,0.066174075,-0.3261881471,-0.1147774085,-0.0888150036,-0.1403396726,-0.5619576573,-0.2250494063,0.0893958807,0.3187947571,0.090573065,0.350672394,-0.4854868352,-0.0084516723,0.1274918616,-0.3142074049,-0.0129661616,-0.1558169574,-0.2733362615,0.0355255343,0.7237205505,0.0963545367,0.3842619359,0.0550471693,-0.2218773365,-0.0491370149,-0.3143673539,-0.0223673265,-0.2295466214,0.338960737,0.3319755793,0.1100180224,-0.1604092568,-0.2559396625,0.2994101048,-0.0435467698,-0.1957678199,0.1944876164,-0.0658890158,-0.0019956648,-0.1117209345,-0.0121067651,-0.1920267642,-0.2215815634,-0.1455163062,0.1978730559,0.3210581243,0.3196024299,0.0293593053,0.1503638178,0.0955186337,-0.0287066139,-0.179673478,-0.3023663461,0.354970634,0.0689084977,-0.4053297043,-0.214728415,-0.1006930247,0.2305930257,0.1018887013,-0.5546712875,0.1395835578,-0.0337636657,0.0982766151,0.1617600173,-0.0188597664,0.4739741981,-0.0388860777,0.056005761,-0.1978461742,-0.2456507236,0.1355912387,0.1616822779,0.6563411951,-0.0545418188,0.4591037631,-0.1254960299,0.5254045129,0.2798558772,-0.1763413548,0.3539974391,0.2002420872,0.2736523151,-0.1534941196,-0.2914330065,0.2855210304,-0.0495390184,-0.1825995594,0.0115922466,-0.0829059854,-0.2005023956,-0.186934337,0.3586075306,-0.1521629244,-0.2032236159,0.0759045109,-0.4192081392,0.1117390022,0.1238377988,0.0417894572,-0.2293193489,-0.2097422928,-0.0342514552,0.3359264731,-0.029353939,0.0341083445,-0.5879100561,-0.2072460651,-0.2761886418,0.3000999391,-0.0228021778,0.5082477331,-0.1271038949,-0.0221780334,0.102294907,-0.1911925822,0.1879758388,-0.43067801,0.1604577899,0.2485894859,0.1959752142,-0.2174742222,-0.3243355751,-0.1971815377,0.0899364948,0.3619167805,0.239043653,-0.4133508205,-0.1795630753,0.307611376,-0.1772216558,-0.0317560434,-0.2116952538,-0.5768941045,-0.3119826317,-0.2455371022,0.1072793156,-0.1011514515,0.309596926,-0.0959005877,0.2520931959,0.1111570522,0.1200232729,0.3118346632,-0.0258780681,0.1249620318,0.2003544718,0.2429250181,-0.2824971676,-0.0304986704,0.1372507811,0.4655221105,-0.0030381177,0.1497278661,0.2462175041,0.1032721847,0.0787187591,0.1354254037,0.1089805216,0.0048270817,0.2942673266,0.1080225632,-0.1681431979,0.2124031782,0.1513140202,-0.0376338847,-0.2570491135,-0.4218550324,0.38398242,0.093876496,0.0786313564,0.5386753082,-0.1991007924,-0.2274917513,0.1776132137,0.4289579093,0.783896327,-0.0462472104,0.050589487,0.2714215815,-0.3885199726,0.1734306216,-0.1278073788,0.2138846517,-0.4140084088,-0.2300041467,-0.1147606447,-0.2370069176,0.2867196798,0.0709195957,0.0210513715,0.2501484454,0.0637665913,0.2134827971,0.0473113768,0.3004758656,-0.2035632432,-0.0009061639,-0.2438044101,0.1599234641,-0.2949710488,0.3841906488,-0.182770893,-0.1416789889,-0.2252994031,-0.2297966331,-0.5931190848,0.0420734324,-0.0345498845,0.1219109669,0.1612068415,-0.2970359027,0.2536320686,-0.0125447102,0.3052423894,0.2149487883,-0.2189770937,0.0320769846,0.1596261412,0.063277714,0.1346657127,0.0577453375,0.3857584298,-0.2114268541,-0.2823188901,-0.0503009036,-0.0693830997,-0.5146650076,0.2868447602,-0.1723500937,0.0800506026,-0.062182311,-0.083649151,0.013985334,0.0304551311,-0.1959529966,0.1261596829,0.0669346377,-0.0748330727,0.2518903613,-0.1257235259,-0.1455212831,-0.1935207397,0.3756965995,0.1373264641,0.0165086873,0.4529263079,0.2720410824,-0.1528951377,-0.2304186672,-0.1056392565,0.1786421239,0.0247844514,-0.0148171419,-0.1295316517,-0.0842389688,-0.1829866022,0.1187394038,0.2865358889,0.0998412445,0.2927297056,-0.5119155645,-0.5008373857,0.2192024887,-0.1976202726,0.1925922036,-0.4042046964,0.3515561819,-0.0026992236,0.0052047288,-0.2980507016,-0.1789057702,-0.4618512392,0.110935986,0.0729268342,-0.2021183968,-0.0560514852,-0.1882601529,0.1196749508,-0.1352256089,-0.391865015,-0.1660435349,-0.1048925593,0.1586326361,0.0087997038,-0.3563891351,0.1701889336,-0.148850292,-0.2751868665,-0.015599017,-0.2245164216,0.0084071597,-0.0516109169,0.1526714116,0.2826909423,-0.0229276065,0.0159272756,-0.1590148509,-0.17387788,0.0574296564,0.0226225108,0.177499637,0.1828509569,0.0482546501,-0.210406661,0.1155915633,-0.103203021,0.0354531817,0.2705411017,-0.3210694194,-0.1103250533,-0.0056250701,0.470459342,0.5312490463,-0.1288326979,-0.1956436783,0.2631398141,0.14437747,-0.4927864373,0.0960755795,0.1032330915,0.0822240859,-0.0551963672,0.1345909089,0.1972732544,0.234277457,0.0172124207,0.1755637974,0.164355889,-0.2493069917,0.0796920657,0.3898656666,-0.0801882744,-0.0386280864,0.376491636,-0.1423221678,-0.0126708494,-0.0343081951,0.1067916229,0.286095798,-0.1919635087,0.0537921451,-0.4108016789,-0.4486485124,-0.1012229621,0.3432513475,0.0036119174,-0.0192605332,-0.0278019346,0.2828312516,-0.1828968376,-0.0578758977,-0.2266147584,0.1052788347,0.0238650609,-0.0718203709,-0.4027289748,-0.1296691298,-0.2279717773,0.1654794812,0.0540963486,0.1262377352,0.3088742197,0.100350827,-0.2129182369,-0.3295906186,0.1171260998,0.3001696467,-0.3030859232,-0.2791534662,0.3936853111,0.2128652781,0.015566417,-0.2271831781,0.3640896976,0.3236168921,0.388795346,-0.2514789104,-0.2520096004,0.0765974447,0.0299055818,-0.0827059522,0.4978212714,0.3275831044,0.0998728797,0.2096665055,0.0794471875,-0.0956615284,0.1966316402,-0.0840153098,-0.2941376269,0.0035990309,0.6603893638,0.1717773676,0.0049782558,-0.0732265711,0.0308114681,-0.1909760088,-0.040956676,0.3082460761,0.1975506544,0.1833478063,-0.0998536721,0.044190418,-0.1410416365,0.6444203854,0.2349773198,0.1140373796,-0.4878853559,-0.3304450512,-0.6936562061,-0.0819301009,-0.2694924176,-0.1117674783,0.105363138,0.2048398256,-0.0446271598,0.0618335828,0.1968583614,0.0232926272,-0.1839202344,-0.0317240842,-0.3047983646,0.1097688451,-0.2589264512,0.0314820558,-0.1542465985,-0.3109222353,0.1440677047,0.0020723504,0.033082813,-0.1179745942,-0.0353142545,0.2205820978,0.3118729591,0.2911404073,0.0490069166,0.6657663584,0.1145598143,-0.1374766827,-0.3619854748,-0.1498403847,-0.4374695122,0.1409903616,-0.1612324715,0.6160255671,0.0332119875,-0.1389366388,-0.3363064528,0.5645500422,0.0902393833,-0.3458330035,-0.085558556,-0.0303031225,-0.1786027253,0.2308140844,-0.0440805741,0.0040956177,0.2179864794,0.454369992,-0.4882106483,-0.1645536423,0.5545018911,-0.071817264,-0.2701316178,-0.1823329479,0.1575057209,-0.0962590948,-0.1653824449,-0.7271353006,-0.0263308324,0.2373397052,-0.132682249,-0.2621173859,0.2379527241,-0.1139862984,0.1122611612,0.0536026247,0.2132699341,-0.3363862336,-0.3369754553,-0.1634086967,-0.193607226]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/534","title":"`list_datasets()` is broken.","comments":"Thanks for reporting !\r\nThis has been fixed in #475 and the fix will be available in the next release","body":"version = '0.4.0'\r\n\r\n`list_datasets()` is broken. It results in the following error : \r\n\r\n```\r\nIn [3]: nlp.list_datasets()\r\nOut[3]: ---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/core\/formatters.py in __call__(self, obj)\r\n 700 type_pprinters=self.type_printers,\r\n 701 deferred_pprinters=self.deferred_printers)\r\n--> 702 printer.pretty(obj)\r\n 703 printer.flush()\r\n 704 return stream.getvalue()\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 375 if cls in self.type_pprinters:\r\n 376 # printer registered in self.type_pprinters\r\n--> 377 return self.type_pprinters[cls](obj, self, cycle)\r\n 378 else:\r\n 379 # deferred printer\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in inner(obj, p, cycle)\r\n 553 p.text(',')\r\n 554 p.breakable()\r\n--> 555 p.pretty(x)\r\n 556 if len(obj) == 1 and type(obj) is tuple:\r\n 557 # Special case for 1-item tuples.\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 392 if cls is not object \\\r\n 393 and callable(cls.__dict__.get('__repr__')):\r\n--> 394 return _repr_pprint(obj, self, cycle)\r\n 395\r\n 396 return _default_pprint(obj, self, cycle)\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in _repr_pprint(obj, p, cycle)\r\n 698 \"\"\"A pprint that just redirects to the normal repr function.\"\"\"\r\n 699 # Find newlines and replace them with p.break_()\r\n--> 700 output = repr(obj)\r\n 701 lines = output.splitlines()\r\n 702 with p.group():\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/nlp\/hf_api.py in __repr__(self)\r\n 110\r\n 111 def __repr__(self):\r\n--> 112 single_line_description = self.description.replace(\"\\n\", \"\")\r\n 113 return f\"nlp.ObjectInfo(id='{self.id}', description='{single_line_description}', files={self.siblings})\"\r\n 114\r\n\r\nAttributeError: 'NoneType' object has no attribute 'replace'\r\n```","comment_length":20,"text":"`list_datasets()` is broken.\nversion = '0.4.0'\r\n\r\n`list_datasets()` is broken. It results in the following error : \r\n\r\n```\r\nIn [3]: nlp.list_datasets()\r\nOut[3]: ---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/core\/formatters.py in __call__(self, obj)\r\n 700 type_pprinters=self.type_printers,\r\n 701 deferred_pprinters=self.deferred_printers)\r\n--> 702 printer.pretty(obj)\r\n 703 printer.flush()\r\n 704 return stream.getvalue()\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 375 if cls in self.type_pprinters:\r\n 376 # printer registered in self.type_pprinters\r\n--> 377 return self.type_pprinters[cls](obj, self, cycle)\r\n 378 else:\r\n 379 # deferred printer\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in inner(obj, p, cycle)\r\n 553 p.text(',')\r\n 554 p.breakable()\r\n--> 555 p.pretty(x)\r\n 556 if len(obj) == 1 and type(obj) is tuple:\r\n 557 # Special case for 1-item tuples.\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 392 if cls is not object \\\r\n 393 and callable(cls.__dict__.get('__repr__')):\r\n--> 394 return _repr_pprint(obj, self, cycle)\r\n 395\r\n 396 return _default_pprint(obj, self, cycle)\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in _repr_pprint(obj, p, cycle)\r\n 698 \"\"\"A pprint that just redirects to the normal repr function.\"\"\"\r\n 699 # Find newlines and replace them with p.break_()\r\n--> 700 output = repr(obj)\r\n 701 lines = output.splitlines()\r\n 702 with p.group():\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/nlp\/hf_api.py in __repr__(self)\r\n 110\r\n 111 def __repr__(self):\r\n--> 112 single_line_description = self.description.replace(\"\\n\", \"\")\r\n 113 return f\"nlp.ObjectInfo(id='{self.id}', description='{single_line_description}', files={self.siblings})\"\r\n 114\r\n\r\nAttributeError: 'NoneType' object has no attribute 'replace'\r\n```\nThanks for reporting !\r\nThis has been fixed in #475 and the fix will be available in the next release","embeddings":[-0.2925504744,0.1287720501,-0.1103174761,0.2530123293,0.1547410935,0.0697497725,0.3828648925,0.4221462011,-0.0766394809,-0.049442023,-0.1751877069,0.4431259036,-0.2499777824,-0.0311814267,-0.0670694038,-0.4052232206,-0.0786968246,0.3645731211,0.0530656092,0.1719869971,-0.177527532,0.3013062477,-0.2915241718,0.1099653244,0.0032130617,-0.3575332165,-0.0393835679,-0.0953138992,-0.2738591433,-0.6101930737,0.2050484121,0.0413203202,0.145760715,0.191788584,-0.0001070091,-0.1000377238,0.3466987014,0.1209742874,-0.4599771798,-0.1070150882,-0.3484742939,-0.1636984348,0.15418455,-0.1577059478,0.2177172452,-0.298948288,0.0492456704,-0.0324081853,0.1428057849,0.4736910462,0.2920954823,0.1383364052,0.3362931013,-0.0987469628,0.1389920712,0.1220614463,-0.2137827873,0.1378371567,0.2005366683,-0.1644205153,0.1876337528,0.3156543076,-0.1838986576,0.0428843275,0.080845803,-0.1026863679,-0.0092912083,-0.2145709842,0.1760112792,0.0995607078,0.4164287746,-0.2796163559,-0.3078490496,0.0164624508,0.058969751,-0.7774239182,0.1081925631,0.0224134699,0.0358126685,0.019904213,-0.3528318703,-0.0658771247,-0.2318228632,0.4673488736,-0.0962115526,0.2898914516,-0.1140294299,0.0171308834,-0.0981983617,-0.0868966356,0.0178055316,0.1038976535,-0.2775534987,-0.006325121,-0.1692347378,-0.1642162651,0.2114341259,0.0902850628,-0.0042494638,0.4375266731,-0.0416292585,-0.1178216562,0.0063491771,0.1518926173,0.1883146763,0.1584148854,0.2687661946,0.3543733954,0.1175390035,0.2619755566,0.1657694131,-0.0674194694,0.1508336514,-0.2813037336,0.1861238033,0.0103147179,0.34960109,-0.110645175,-0.2098476887,0.2666292191,-0.1843719929,-0.0758719295,0.0960812718,0.2046851963,-0.0793879256,0.2663103342,0.1156045943,-0.0302527286,-0.1726293713,-0.0930825919,0.0089965528,0.0878812671,-0.0294366628,-0.0644476563,0.0349587202,-0.0605724901,0.2429809868,-0.1208946779,0.1779672801,0.0197550375,0.1838933676,-0.4000001848,0.1542435884,0.2231269926,-0.1593793035,0.1788142025,0.2923768163,-0.4119538963,-0.1074468568,-0.1714627445,-0.076747492,-0.2104943842,-0.4990203083,0.2266946137,0.1809665859,-0.0194095653,-0.096170783,-0.1299234927,0.3491951823,-0.2621610463,0.0378679335,-0.5697956681,-0.099115558,-0.1339209676,0.1721034348,0.300033778,-0.3002201021,0.1677255332,-0.2391221225,-0.0513175763,0.0095718605,0.2239179015,-0.1983066946,0.2528689802,-0.1809735298,-0.0589219183,0.2944368422,-0.1372349709,-0.2301748246,0.2795769572,0.2491778433,-0.0680247694,-0.2227915227,-0.0755549893,0.2689787447,0.0481586978,0.239718467,0.1333486587,0.0076559032,-0.1522696316,-0.1668646932,-0.0934467167,0.381395787,-0.1442018896,0.1337888688,0.1369581521,-0.0157831535,0.1134748012,0.1468215734,-0.1264295727,-0.0556389503,0.1905207187,0.2522134185,0.1714027375,0.0542701334,-0.3509204388,-0.4462583363,-0.0797010213,-0.2040190697,0.0652356073,-0.2956978679,0.1840091497,-0.229904905,0.1598493457,-0.1872381717,-0.2770086527,0.199253276,0.2211045772,0.0691296235,-0.0560180396,-0.1147577688,-0.0540781133,-0.0689443648,0.081534408,-0.4202454388,0.3430917263,-0.2603977621,-0.2390125692,0.0644561425,0.0965775475,0.1995361745,-0.1361991763,-0.2320535332,0.4403236806,0.1650057137,-0.2004151046,-0.1622930318,0.077113837,0.1574232131,-0.1858457625,0.0220011715,0.3202284276,0.098340556,0.057860136,-0.0076144696,0.2805489302,-0.1693376452,0.1688168645,0.2262055576,0.3731301129,0.2607486248,0.1023362055,-0.0886058956,-0.2534463406,0.4879236221,0.0339732207,0.2914642394,-0.2156838924,-0.5336611271,0.0760332271,0.3073705435,0.148070097,0.1422265172,-0.0100683542,-0.1057600006,0.0769602954,0.2739711702,0.1093880683,0.3183995187,0.0714439526,-0.2827411294,-0.06643226,-0.268722266,-0.1903395504,0.2828251719,0.0425259769,0.1601824164,0.1422438473,0.2664467096,0.0376957208,-0.38587901,-0.4234346151,-0.0569498241,0.3714588881,-0.1407547593,0.2643475235,-0.3491425514,-0.4249199927,-0.2960482836,-0.2388049364,-0.2406001985,-0.3623484075,-0.0960663483,0.1309019029,0.0306267813,0.23544842,-0.1495063454,-0.0630283803,0.1627140343,0.1025079563,-0.0452507511,-0.1686990857,-0.2626588941,0.1136495322,0.1955718696,-0.0830842555,0.376789093,-0.0422195643,-0.1625643373,-0.2199974954,-0.4036245048,0.1120402217,-0.0680637285,0.198832646,0.5051586032,0.0949800089,-0.0435776599,-0.1564034969,0.4378064275,-0.056366723,0.1829486191,0.1709958315,-0.0545085445,-0.3030057251,-0.1642799526,-0.4381222725,-0.1721548587,-0.1750888675,0.1431864202,0.3657177091,0.367105037,0.5672655702,0.0486722551,0.1627244651,-0.1498387605,0.1697799414,-0.0384772494,0.0926329643,0.1827802211,-0.2975689173,-0.4033008814,-0.205091387,0.1107111797,0.0534857623,-0.0109257596,-0.4335919619,-0.1561717689,-0.2316404432,0.2016340494,-0.0561200008,0.0157572627,0.2922447622,0.1445402205,-0.0896595716,-0.0230043251,-0.0041051153,-0.0687814355,-0.1858177781,0.1621867567,0.02480096,0.2922298312,-0.0115637807,0.2237012982,0.3424477279,-0.21517995,0.127971068,-0.1084780023,0.4514208436,-0.1578459591,-0.4963950813,0.2596535981,-0.0834074542,-0.0708971173,0.1117999703,-0.0996082053,-0.2902067304,-0.1338139623,0.3741458654,-0.0089206379,-0.070991613,0.0830765516,-0.1850357056,0.3181610405,0.0026355914,0.2742580771,-0.1590847522,-0.2411731929,-0.1272432357,0.1199280769,-0.009803867,-0.1198805571,-0.5893691778,-0.3102943003,-0.247439459,0.3407361805,0.1813260019,0.7312636375,-0.0435702987,0.1915239096,-0.2322877198,-0.0519199781,0.2175544798,-0.326344192,-0.1254757047,0.5302116871,-0.0307003837,-0.118570663,-0.1065544412,-0.4047205448,0.3956201375,0.1058988646,0.2219056487,-0.0387296975,-0.0528825969,0.3625879288,0.1771215349,-0.1430920213,0.0908260271,-0.062347617,-0.0892777219,-0.1557316929,-0.1180961952,-0.192764461,0.1466251016,-0.0049093072,0.1919082701,-0.1392062753,-0.1257688999,-0.0748906657,0.1233940497,0.1236951351,0.0033262207,-0.0421622843,-0.0984143764,0.2123972028,0.1988953501,0.111814931,-0.3817193508,-0.220465377,-0.1041044816,-0.0807111338,0.2407717109,0.0738160312,-0.2533874512,0.0336981937,-0.0934980437,0.1150830686,-0.0288711935,-0.2338213772,0.5240261555,0.0251873322,-0.3135377169,-0.4343085885,0.2066931278,0.0136043653,0.1050095633,0.3896684647,-0.2068133503,-0.3086203039,0.1892786175,0.1139311194,0.6405631304,0.0851485878,0.0652173385,0.2540678382,-0.0935600325,0.6293442845,0.154667303,0.2402301282,-0.3571700752,-0.2086233944,-0.0803108662,-0.1555121988,0.4733545184,0.050789699,-0.3131177723,0.2505069971,-0.0391839445,0.1385234892,0.0359818265,0.2782002389,0.0778178126,-0.1967809945,-0.5601501465,0.1838223338,-0.2324073017,-0.0873937979,-0.3194411397,0.0811043978,0.1707180589,-0.1656274945,-0.4413450956,0.1346510947,-0.1211197376,0.3450767696,0.3261592686,-0.3975536823,0.1371126026,0.1328669786,-0.0263587907,0.2817955017,-0.1486691386,-0.0861459672,-0.04510298,-0.1509840935,0.0990662873,0.1168280244,0.4106481075,-0.0275178254,0.0686404184,0.1827409565,0.0321697704,-0.1284331381,-0.0911368951,-0.0259144288,0.1404956877,-0.2732411027,-0.1399982572,-0.0654593408,-0.1451391429,-0.3277772665,0.160689801,0.1787601113,-0.1430871338,0.2235409617,-0.064017199,0.0266124457,-0.0206706487,0.1997918636,0.0545464233,-0.0542282909,0.4666308463,0.4435594678,-0.2183807343,-0.3694681823,0.055177819,-0.1608375758,-0.3175583184,0.2667807043,-0.1879154742,0.1919151545,-0.0624178648,0.1855025887,-0.1084150523,0.2329244018,0.0201102067,-0.6296340227,-0.1027127504,0.0046861414,0.1599672288,0.1682860702,0.0211270824,0.163348332,0.0623625815,0.1665311903,-0.3393384218,-0.0928668007,-0.2718740404,0.1751075983,0.0623882599,0.0062928703,0.0212407559,-0.0425847955,0.1828920096,0.1107411385,-0.0693013147,-0.2771647274,-0.2256083786,0.0743487477,0.1718113869,-0.507412076,0.0664311126,-0.0953751802,-0.2018558085,-0.2209293842,-0.1938804686,0.0060284715,-0.1222522259,0.3884958923,-0.091798149,-0.0195536781,0.0311304871,0.088862054,-0.0639946982,0.1822521836,0.0906904861,0.3359932899,0.1552698314,-0.0223114416,-0.4787735641,0.0881573036,-0.1612366438,0.0551174805,0.4794063866,-0.5633710623,-0.0244878326,-0.0482899845,0.5955436826,0.3907880783,-0.2083756775,0.0840869248,0.2443304658,0.2077737153,-0.3391135931,-0.1119453683,0.2847105563,0.1861761063,0.0454925708,0.0799686164,0.0038469923,-0.0595146827,-0.1182622686,0.029866742,0.1543686539,-0.5287248492,0.2331763953,0.3293747306,0.1726527959,-0.0026453021,0.166285947,0.1339374334,0.1855650842,0.3369790018,0.0011743638,0.2679281831,0.2355300933,0.0644427314,0.236332953,-0.129755646,0.0706022382,0.1460008174,0.2153232545,0.2068515718,0.0903030559,0.3038617074,-0.3223173916,-0.092349872,0.0277668256,0.2207098454,-0.1179727167,-0.0328400135,-0.222666204,-0.0682456419,-0.1485171914,0.0043878197,-0.2459124029,-0.0806655288,0.5146118999,0.1833415926,-0.1658653319,-0.8891292214,0.2164248377,0.1544321924,0.0647152662,-0.314719975,0.2495400906,0.1710821092,0.0437222868,-0.2234827876,0.7008237839,0.5018439889,0.3941646218,0.1825341433,-0.0488634109,-0.2682225406,-0.1042501628,0.0419572145,0.1700416356,0.1123287678,0.0595856346,0.3015963137,0.133301422,-0.149044916,0.2988526821,0.2556347251,0.0375600196,-0.0993071422,0.4731218219,-0.0470735952,-0.2465269268,-0.135557279,-0.0301100947,-0.3524824679,0.1153910905,0.540610373,0.2814173102,0.0571114309,-0.0885857344,0.1267948747,0.1267552376,0.4031401277,0.475553751,0.3269433081,-0.1795827299,-0.1937531084,-0.2992210388,0.157597959,-0.2197696865,-0.0460958704,0.1135376394,0.0597481094,-0.0773707703,0.2345739752,0.0274125598,0.0945409834,-0.0248964783,0.0568600632,-0.3780333996,-0.1162496805,0.2867147028,0.0270568617,-0.0278433077,-0.196413368,0.1512077898,-0.1286519319,0.1126163304,-0.3674302101,-0.0651417524,-0.1094167382,0.2097267956,0.0312755145,0.4317313135,0.2470489889,-0.0309630502,-0.2971615791,-0.2883206904,-0.4377296567,-0.1834001243,0.5462127328,0.1380329877,0.53362602,-0.1627794355,-0.243976742,-0.4112980664,0.433110863,-0.3177213073,-0.0555907972,-0.2225956321,0.2077827901,-0.2010600567,-0.0666386336,0.0329128429,0.2839014828,-0.0416875146,0.2112437934,-0.4552669227,-0.2773829699,0.259721756,-0.7163981199,-0.3755799532,-0.1382011771,0.1000654697,-0.0333187468,-0.1368526518,-0.3827458918,-0.0606800728,0.336499691,-0.1395791024,-0.2957063317,0.073035188,0.1560716629,0.1417677701,-0.0939191952,0.6182773113,0.0997412354,-0.3169441819,0.0998760164,-0.0598640777]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/534","title":"`list_datasets()` is broken.","comments":"What you can do instead to get the list of the datasets is call\r\n\r\n```python\r\nprint([dataset.id for dataset in nlp.list_datasets()])\r\n```","body":"version = '0.4.0'\r\n\r\n`list_datasets()` is broken. It results in the following error : \r\n\r\n```\r\nIn [3]: nlp.list_datasets()\r\nOut[3]: ---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/core\/formatters.py in __call__(self, obj)\r\n 700 type_pprinters=self.type_printers,\r\n 701 deferred_pprinters=self.deferred_printers)\r\n--> 702 printer.pretty(obj)\r\n 703 printer.flush()\r\n 704 return stream.getvalue()\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 375 if cls in self.type_pprinters:\r\n 376 # printer registered in self.type_pprinters\r\n--> 377 return self.type_pprinters[cls](obj, self, cycle)\r\n 378 else:\r\n 379 # deferred printer\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in inner(obj, p, cycle)\r\n 553 p.text(',')\r\n 554 p.breakable()\r\n--> 555 p.pretty(x)\r\n 556 if len(obj) == 1 and type(obj) is tuple:\r\n 557 # Special case for 1-item tuples.\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 392 if cls is not object \\\r\n 393 and callable(cls.__dict__.get('__repr__')):\r\n--> 394 return _repr_pprint(obj, self, cycle)\r\n 395\r\n 396 return _default_pprint(obj, self, cycle)\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in _repr_pprint(obj, p, cycle)\r\n 698 \"\"\"A pprint that just redirects to the normal repr function.\"\"\"\r\n 699 # Find newlines and replace them with p.break_()\r\n--> 700 output = repr(obj)\r\n 701 lines = output.splitlines()\r\n 702 with p.group():\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/nlp\/hf_api.py in __repr__(self)\r\n 110\r\n 111 def __repr__(self):\r\n--> 112 single_line_description = self.description.replace(\"\\n\", \"\")\r\n 113 return f\"nlp.ObjectInfo(id='{self.id}', description='{single_line_description}', files={self.siblings})\"\r\n 114\r\n\r\nAttributeError: 'NoneType' object has no attribute 'replace'\r\n```","comment_length":21,"text":"`list_datasets()` is broken.\nversion = '0.4.0'\r\n\r\n`list_datasets()` is broken. It results in the following error : \r\n\r\n```\r\nIn [3]: nlp.list_datasets()\r\nOut[3]: ---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/core\/formatters.py in __call__(self, obj)\r\n 700 type_pprinters=self.type_printers,\r\n 701 deferred_pprinters=self.deferred_printers)\r\n--> 702 printer.pretty(obj)\r\n 703 printer.flush()\r\n 704 return stream.getvalue()\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 375 if cls in self.type_pprinters:\r\n 376 # printer registered in self.type_pprinters\r\n--> 377 return self.type_pprinters[cls](obj, self, cycle)\r\n 378 else:\r\n 379 # deferred printer\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in inner(obj, p, cycle)\r\n 553 p.text(',')\r\n 554 p.breakable()\r\n--> 555 p.pretty(x)\r\n 556 if len(obj) == 1 and type(obj) is tuple:\r\n 557 # Special case for 1-item tuples.\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 392 if cls is not object \\\r\n 393 and callable(cls.__dict__.get('__repr__')):\r\n--> 394 return _repr_pprint(obj, self, cycle)\r\n 395\r\n 396 return _default_pprint(obj, self, cycle)\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in _repr_pprint(obj, p, cycle)\r\n 698 \"\"\"A pprint that just redirects to the normal repr function.\"\"\"\r\n 699 # Find newlines and replace them with p.break_()\r\n--> 700 output = repr(obj)\r\n 701 lines = output.splitlines()\r\n 702 with p.group():\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/nlp\/hf_api.py in __repr__(self)\r\n 110\r\n 111 def __repr__(self):\r\n--> 112 single_line_description = self.description.replace(\"\\n\", \"\")\r\n 113 return f\"nlp.ObjectInfo(id='{self.id}', description='{single_line_description}', files={self.siblings})\"\r\n 114\r\n\r\nAttributeError: 'NoneType' object has no attribute 'replace'\r\n```\nWhat you can do instead to get the list of the datasets is call\r\n\r\n```python\r\nprint([dataset.id for dataset in nlp.list_datasets()])\r\n```","embeddings":[-0.2925504744,0.1287720501,-0.1103174761,0.2530123293,0.1547410935,0.0697497725,0.3828648925,0.4221462011,-0.0766394809,-0.049442023,-0.1751877069,0.4431259036,-0.2499777824,-0.0311814267,-0.0670694038,-0.4052232206,-0.0786968246,0.3645731211,0.0530656092,0.1719869971,-0.177527532,0.3013062477,-0.2915241718,0.1099653244,0.0032130617,-0.3575332165,-0.0393835679,-0.0953138992,-0.2738591433,-0.6101930737,0.2050484121,0.0413203202,0.145760715,0.191788584,-0.0001070091,-0.1000377238,0.3466987014,0.1209742874,-0.4599771798,-0.1070150882,-0.3484742939,-0.1636984348,0.15418455,-0.1577059478,0.2177172452,-0.298948288,0.0492456704,-0.0324081853,0.1428057849,0.4736910462,0.2920954823,0.1383364052,0.3362931013,-0.0987469628,0.1389920712,0.1220614463,-0.2137827873,0.1378371567,0.2005366683,-0.1644205153,0.1876337528,0.3156543076,-0.1838986576,0.0428843275,0.080845803,-0.1026863679,-0.0092912083,-0.2145709842,0.1760112792,0.0995607078,0.4164287746,-0.2796163559,-0.3078490496,0.0164624508,0.058969751,-0.7774239182,0.1081925631,0.0224134699,0.0358126685,0.019904213,-0.3528318703,-0.0658771247,-0.2318228632,0.4673488736,-0.0962115526,0.2898914516,-0.1140294299,0.0171308834,-0.0981983617,-0.0868966356,0.0178055316,0.1038976535,-0.2775534987,-0.006325121,-0.1692347378,-0.1642162651,0.2114341259,0.0902850628,-0.0042494638,0.4375266731,-0.0416292585,-0.1178216562,0.0063491771,0.1518926173,0.1883146763,0.1584148854,0.2687661946,0.3543733954,0.1175390035,0.2619755566,0.1657694131,-0.0674194694,0.1508336514,-0.2813037336,0.1861238033,0.0103147179,0.34960109,-0.110645175,-0.2098476887,0.2666292191,-0.1843719929,-0.0758719295,0.0960812718,0.2046851963,-0.0793879256,0.2663103342,0.1156045943,-0.0302527286,-0.1726293713,-0.0930825919,0.0089965528,0.0878812671,-0.0294366628,-0.0644476563,0.0349587202,-0.0605724901,0.2429809868,-0.1208946779,0.1779672801,0.0197550375,0.1838933676,-0.4000001848,0.1542435884,0.2231269926,-0.1593793035,0.1788142025,0.2923768163,-0.4119538963,-0.1074468568,-0.1714627445,-0.076747492,-0.2104943842,-0.4990203083,0.2266946137,0.1809665859,-0.0194095653,-0.096170783,-0.1299234927,0.3491951823,-0.2621610463,0.0378679335,-0.5697956681,-0.099115558,-0.1339209676,0.1721034348,0.300033778,-0.3002201021,0.1677255332,-0.2391221225,-0.0513175763,0.0095718605,0.2239179015,-0.1983066946,0.2528689802,-0.1809735298,-0.0589219183,0.2944368422,-0.1372349709,-0.2301748246,0.2795769572,0.2491778433,-0.0680247694,-0.2227915227,-0.0755549893,0.2689787447,0.0481586978,0.239718467,0.1333486587,0.0076559032,-0.1522696316,-0.1668646932,-0.0934467167,0.381395787,-0.1442018896,0.1337888688,0.1369581521,-0.0157831535,0.1134748012,0.1468215734,-0.1264295727,-0.0556389503,0.1905207187,0.2522134185,0.1714027375,0.0542701334,-0.3509204388,-0.4462583363,-0.0797010213,-0.2040190697,0.0652356073,-0.2956978679,0.1840091497,-0.229904905,0.1598493457,-0.1872381717,-0.2770086527,0.199253276,0.2211045772,0.0691296235,-0.0560180396,-0.1147577688,-0.0540781133,-0.0689443648,0.081534408,-0.4202454388,0.3430917263,-0.2603977621,-0.2390125692,0.0644561425,0.0965775475,0.1995361745,-0.1361991763,-0.2320535332,0.4403236806,0.1650057137,-0.2004151046,-0.1622930318,0.077113837,0.1574232131,-0.1858457625,0.0220011715,0.3202284276,0.098340556,0.057860136,-0.0076144696,0.2805489302,-0.1693376452,0.1688168645,0.2262055576,0.3731301129,0.2607486248,0.1023362055,-0.0886058956,-0.2534463406,0.4879236221,0.0339732207,0.2914642394,-0.2156838924,-0.5336611271,0.0760332271,0.3073705435,0.148070097,0.1422265172,-0.0100683542,-0.1057600006,0.0769602954,0.2739711702,0.1093880683,0.3183995187,0.0714439526,-0.2827411294,-0.06643226,-0.268722266,-0.1903395504,0.2828251719,0.0425259769,0.1601824164,0.1422438473,0.2664467096,0.0376957208,-0.38587901,-0.4234346151,-0.0569498241,0.3714588881,-0.1407547593,0.2643475235,-0.3491425514,-0.4249199927,-0.2960482836,-0.2388049364,-0.2406001985,-0.3623484075,-0.0960663483,0.1309019029,0.0306267813,0.23544842,-0.1495063454,-0.0630283803,0.1627140343,0.1025079563,-0.0452507511,-0.1686990857,-0.2626588941,0.1136495322,0.1955718696,-0.0830842555,0.376789093,-0.0422195643,-0.1625643373,-0.2199974954,-0.4036245048,0.1120402217,-0.0680637285,0.198832646,0.5051586032,0.0949800089,-0.0435776599,-0.1564034969,0.4378064275,-0.056366723,0.1829486191,0.1709958315,-0.0545085445,-0.3030057251,-0.1642799526,-0.4381222725,-0.1721548587,-0.1750888675,0.1431864202,0.3657177091,0.367105037,0.5672655702,0.0486722551,0.1627244651,-0.1498387605,0.1697799414,-0.0384772494,0.0926329643,0.1827802211,-0.2975689173,-0.4033008814,-0.205091387,0.1107111797,0.0534857623,-0.0109257596,-0.4335919619,-0.1561717689,-0.2316404432,0.2016340494,-0.0561200008,0.0157572627,0.2922447622,0.1445402205,-0.0896595716,-0.0230043251,-0.0041051153,-0.0687814355,-0.1858177781,0.1621867567,0.02480096,0.2922298312,-0.0115637807,0.2237012982,0.3424477279,-0.21517995,0.127971068,-0.1084780023,0.4514208436,-0.1578459591,-0.4963950813,0.2596535981,-0.0834074542,-0.0708971173,0.1117999703,-0.0996082053,-0.2902067304,-0.1338139623,0.3741458654,-0.0089206379,-0.070991613,0.0830765516,-0.1850357056,0.3181610405,0.0026355914,0.2742580771,-0.1590847522,-0.2411731929,-0.1272432357,0.1199280769,-0.009803867,-0.1198805571,-0.5893691778,-0.3102943003,-0.247439459,0.3407361805,0.1813260019,0.7312636375,-0.0435702987,0.1915239096,-0.2322877198,-0.0519199781,0.2175544798,-0.326344192,-0.1254757047,0.5302116871,-0.0307003837,-0.118570663,-0.1065544412,-0.4047205448,0.3956201375,0.1058988646,0.2219056487,-0.0387296975,-0.0528825969,0.3625879288,0.1771215349,-0.1430920213,0.0908260271,-0.062347617,-0.0892777219,-0.1557316929,-0.1180961952,-0.192764461,0.1466251016,-0.0049093072,0.1919082701,-0.1392062753,-0.1257688999,-0.0748906657,0.1233940497,0.1236951351,0.0033262207,-0.0421622843,-0.0984143764,0.2123972028,0.1988953501,0.111814931,-0.3817193508,-0.220465377,-0.1041044816,-0.0807111338,0.2407717109,0.0738160312,-0.2533874512,0.0336981937,-0.0934980437,0.1150830686,-0.0288711935,-0.2338213772,0.5240261555,0.0251873322,-0.3135377169,-0.4343085885,0.2066931278,0.0136043653,0.1050095633,0.3896684647,-0.2068133503,-0.3086203039,0.1892786175,0.1139311194,0.6405631304,0.0851485878,0.0652173385,0.2540678382,-0.0935600325,0.6293442845,0.154667303,0.2402301282,-0.3571700752,-0.2086233944,-0.0803108662,-0.1555121988,0.4733545184,0.050789699,-0.3131177723,0.2505069971,-0.0391839445,0.1385234892,0.0359818265,0.2782002389,0.0778178126,-0.1967809945,-0.5601501465,0.1838223338,-0.2324073017,-0.0873937979,-0.3194411397,0.0811043978,0.1707180589,-0.1656274945,-0.4413450956,0.1346510947,-0.1211197376,0.3450767696,0.3261592686,-0.3975536823,0.1371126026,0.1328669786,-0.0263587907,0.2817955017,-0.1486691386,-0.0861459672,-0.04510298,-0.1509840935,0.0990662873,0.1168280244,0.4106481075,-0.0275178254,0.0686404184,0.1827409565,0.0321697704,-0.1284331381,-0.0911368951,-0.0259144288,0.1404956877,-0.2732411027,-0.1399982572,-0.0654593408,-0.1451391429,-0.3277772665,0.160689801,0.1787601113,-0.1430871338,0.2235409617,-0.064017199,0.0266124457,-0.0206706487,0.1997918636,0.0545464233,-0.0542282909,0.4666308463,0.4435594678,-0.2183807343,-0.3694681823,0.055177819,-0.1608375758,-0.3175583184,0.2667807043,-0.1879154742,0.1919151545,-0.0624178648,0.1855025887,-0.1084150523,0.2329244018,0.0201102067,-0.6296340227,-0.1027127504,0.0046861414,0.1599672288,0.1682860702,0.0211270824,0.163348332,0.0623625815,0.1665311903,-0.3393384218,-0.0928668007,-0.2718740404,0.1751075983,0.0623882599,0.0062928703,0.0212407559,-0.0425847955,0.1828920096,0.1107411385,-0.0693013147,-0.2771647274,-0.2256083786,0.0743487477,0.1718113869,-0.507412076,0.0664311126,-0.0953751802,-0.2018558085,-0.2209293842,-0.1938804686,0.0060284715,-0.1222522259,0.3884958923,-0.091798149,-0.0195536781,0.0311304871,0.088862054,-0.0639946982,0.1822521836,0.0906904861,0.3359932899,0.1552698314,-0.0223114416,-0.4787735641,0.0881573036,-0.1612366438,0.0551174805,0.4794063866,-0.5633710623,-0.0244878326,-0.0482899845,0.5955436826,0.3907880783,-0.2083756775,0.0840869248,0.2443304658,0.2077737153,-0.3391135931,-0.1119453683,0.2847105563,0.1861761063,0.0454925708,0.0799686164,0.0038469923,-0.0595146827,-0.1182622686,0.029866742,0.1543686539,-0.5287248492,0.2331763953,0.3293747306,0.1726527959,-0.0026453021,0.166285947,0.1339374334,0.1855650842,0.3369790018,0.0011743638,0.2679281831,0.2355300933,0.0644427314,0.236332953,-0.129755646,0.0706022382,0.1460008174,0.2153232545,0.2068515718,0.0903030559,0.3038617074,-0.3223173916,-0.092349872,0.0277668256,0.2207098454,-0.1179727167,-0.0328400135,-0.222666204,-0.0682456419,-0.1485171914,0.0043878197,-0.2459124029,-0.0806655288,0.5146118999,0.1833415926,-0.1658653319,-0.8891292214,0.2164248377,0.1544321924,0.0647152662,-0.314719975,0.2495400906,0.1710821092,0.0437222868,-0.2234827876,0.7008237839,0.5018439889,0.3941646218,0.1825341433,-0.0488634109,-0.2682225406,-0.1042501628,0.0419572145,0.1700416356,0.1123287678,0.0595856346,0.3015963137,0.133301422,-0.149044916,0.2988526821,0.2556347251,0.0375600196,-0.0993071422,0.4731218219,-0.0470735952,-0.2465269268,-0.135557279,-0.0301100947,-0.3524824679,0.1153910905,0.540610373,0.2814173102,0.0571114309,-0.0885857344,0.1267948747,0.1267552376,0.4031401277,0.475553751,0.3269433081,-0.1795827299,-0.1937531084,-0.2992210388,0.157597959,-0.2197696865,-0.0460958704,0.1135376394,0.0597481094,-0.0773707703,0.2345739752,0.0274125598,0.0945409834,-0.0248964783,0.0568600632,-0.3780333996,-0.1162496805,0.2867147028,0.0270568617,-0.0278433077,-0.196413368,0.1512077898,-0.1286519319,0.1126163304,-0.3674302101,-0.0651417524,-0.1094167382,0.2097267956,0.0312755145,0.4317313135,0.2470489889,-0.0309630502,-0.2971615791,-0.2883206904,-0.4377296567,-0.1834001243,0.5462127328,0.1380329877,0.53362602,-0.1627794355,-0.243976742,-0.4112980664,0.433110863,-0.3177213073,-0.0555907972,-0.2225956321,0.2077827901,-0.2010600567,-0.0666386336,0.0329128429,0.2839014828,-0.0416875146,0.2112437934,-0.4552669227,-0.2773829699,0.259721756,-0.7163981199,-0.3755799532,-0.1382011771,0.1000654697,-0.0333187468,-0.1368526518,-0.3827458918,-0.0606800728,0.336499691,-0.1395791024,-0.2957063317,0.073035188,0.1560716629,0.1417677701,-0.0939191952,0.6182773113,0.0997412354,-0.3169441819,0.0998760164,-0.0598640777]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/534","title":"`list_datasets()` is broken.","comments":"Thanks @lhoestq . ","body":"version = '0.4.0'\r\n\r\n`list_datasets()` is broken. It results in the following error : \r\n\r\n```\r\nIn [3]: nlp.list_datasets()\r\nOut[3]: ---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/core\/formatters.py in __call__(self, obj)\r\n 700 type_pprinters=self.type_printers,\r\n 701 deferred_pprinters=self.deferred_printers)\r\n--> 702 printer.pretty(obj)\r\n 703 printer.flush()\r\n 704 return stream.getvalue()\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 375 if cls in self.type_pprinters:\r\n 376 # printer registered in self.type_pprinters\r\n--> 377 return self.type_pprinters[cls](obj, self, cycle)\r\n 378 else:\r\n 379 # deferred printer\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in inner(obj, p, cycle)\r\n 553 p.text(',')\r\n 554 p.breakable()\r\n--> 555 p.pretty(x)\r\n 556 if len(obj) == 1 and type(obj) is tuple:\r\n 557 # Special case for 1-item tuples.\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 392 if cls is not object \\\r\n 393 and callable(cls.__dict__.get('__repr__')):\r\n--> 394 return _repr_pprint(obj, self, cycle)\r\n 395\r\n 396 return _default_pprint(obj, self, cycle)\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in _repr_pprint(obj, p, cycle)\r\n 698 \"\"\"A pprint that just redirects to the normal repr function.\"\"\"\r\n 699 # Find newlines and replace them with p.break_()\r\n--> 700 output = repr(obj)\r\n 701 lines = output.splitlines()\r\n 702 with p.group():\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/nlp\/hf_api.py in __repr__(self)\r\n 110\r\n 111 def __repr__(self):\r\n--> 112 single_line_description = self.description.replace(\"\\n\", \"\")\r\n 113 return f\"nlp.ObjectInfo(id='{self.id}', description='{single_line_description}', files={self.siblings})\"\r\n 114\r\n\r\nAttributeError: 'NoneType' object has no attribute 'replace'\r\n```","comment_length":3,"text":"`list_datasets()` is broken.\nversion = '0.4.0'\r\n\r\n`list_datasets()` is broken. It results in the following error : \r\n\r\n```\r\nIn [3]: nlp.list_datasets()\r\nOut[3]: ---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/core\/formatters.py in __call__(self, obj)\r\n 700 type_pprinters=self.type_printers,\r\n 701 deferred_pprinters=self.deferred_printers)\r\n--> 702 printer.pretty(obj)\r\n 703 printer.flush()\r\n 704 return stream.getvalue()\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 375 if cls in self.type_pprinters:\r\n 376 # printer registered in self.type_pprinters\r\n--> 377 return self.type_pprinters[cls](obj, self, cycle)\r\n 378 else:\r\n 379 # deferred printer\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in inner(obj, p, cycle)\r\n 553 p.text(',')\r\n 554 p.breakable()\r\n--> 555 p.pretty(x)\r\n 556 if len(obj) == 1 and type(obj) is tuple:\r\n 557 # Special case for 1-item tuples.\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in pretty(self, obj)\r\n 392 if cls is not object \\\r\n 393 and callable(cls.__dict__.get('__repr__')):\r\n--> 394 return _repr_pprint(obj, self, cycle)\r\n 395\r\n 396 return _default_pprint(obj, self, cycle)\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/IPython\/lib\/pretty.py in _repr_pprint(obj, p, cycle)\r\n 698 \"\"\"A pprint that just redirects to the normal repr function.\"\"\"\r\n 699 # Find newlines and replace them with p.break_()\r\n--> 700 output = repr(obj)\r\n 701 lines = output.splitlines()\r\n 702 with p.group():\r\n\r\n~\/.virtualenvs\/san-lgUCsFg_\/lib\/python3.8\/site-packages\/nlp\/hf_api.py in __repr__(self)\r\n 110\r\n 111 def __repr__(self):\r\n--> 112 single_line_description = self.description.replace(\"\\n\", \"\")\r\n 113 return f\"nlp.ObjectInfo(id='{self.id}', description='{single_line_description}', files={self.siblings})\"\r\n 114\r\n\r\nAttributeError: 'NoneType' object has no attribute 'replace'\r\n```\nThanks @lhoestq . ","embeddings":[-0.2925504744,0.1287720501,-0.1103174761,0.2530123293,0.1547410935,0.0697497725,0.3828648925,0.4221462011,-0.0766394809,-0.049442023,-0.1751877069,0.4431259036,-0.2499777824,-0.0311814267,-0.0670694038,-0.4052232206,-0.0786968246,0.3645731211,0.0530656092,0.1719869971,-0.177527532,0.3013062477,-0.2915241718,0.1099653244,0.0032130617,-0.3575332165,-0.0393835679,-0.0953138992,-0.2738591433,-0.6101930737,0.2050484121,0.0413203202,0.145760715,0.191788584,-0.0001070091,-0.1000377238,0.3466987014,0.1209742874,-0.4599771798,-0.1070150882,-0.3484742939,-0.1636984348,0.15418455,-0.1577059478,0.2177172452,-0.298948288,0.0492456704,-0.0324081853,0.1428057849,0.4736910462,0.2920954823,0.1383364052,0.3362931013,-0.0987469628,0.1389920712,0.1220614463,-0.2137827873,0.1378371567,0.2005366683,-0.1644205153,0.1876337528,0.3156543076,-0.1838986576,0.0428843275,0.080845803,-0.1026863679,-0.0092912083,-0.2145709842,0.1760112792,0.0995607078,0.4164287746,-0.2796163559,-0.3078490496,0.0164624508,0.058969751,-0.7774239182,0.1081925631,0.0224134699,0.0358126685,0.019904213,-0.3528318703,-0.0658771247,-0.2318228632,0.4673488736,-0.0962115526,0.2898914516,-0.1140294299,0.0171308834,-0.0981983617,-0.0868966356,0.0178055316,0.1038976535,-0.2775534987,-0.006325121,-0.1692347378,-0.1642162651,0.2114341259,0.0902850628,-0.0042494638,0.4375266731,-0.0416292585,-0.1178216562,0.0063491771,0.1518926173,0.1883146763,0.1584148854,0.2687661946,0.3543733954,0.1175390035,0.2619755566,0.1657694131,-0.0674194694,0.1508336514,-0.2813037336,0.1861238033,0.0103147179,0.34960109,-0.110645175,-0.2098476887,0.2666292191,-0.1843719929,-0.0758719295,0.0960812718,0.2046851963,-0.0793879256,0.2663103342,0.1156045943,-0.0302527286,-0.1726293713,-0.0930825919,0.0089965528,0.0878812671,-0.0294366628,-0.0644476563,0.0349587202,-0.0605724901,0.2429809868,-0.1208946779,0.1779672801,0.0197550375,0.1838933676,-0.4000001848,0.1542435884,0.2231269926,-0.1593793035,0.1788142025,0.2923768163,-0.4119538963,-0.1074468568,-0.1714627445,-0.076747492,-0.2104943842,-0.4990203083,0.2266946137,0.1809665859,-0.0194095653,-0.096170783,-0.1299234927,0.3491951823,-0.2621610463,0.0378679335,-0.5697956681,-0.099115558,-0.1339209676,0.1721034348,0.300033778,-0.3002201021,0.1677255332,-0.2391221225,-0.0513175763,0.0095718605,0.2239179015,-0.1983066946,0.2528689802,-0.1809735298,-0.0589219183,0.2944368422,-0.1372349709,-0.2301748246,0.2795769572,0.2491778433,-0.0680247694,-0.2227915227,-0.0755549893,0.2689787447,0.0481586978,0.239718467,0.1333486587,0.0076559032,-0.1522696316,-0.1668646932,-0.0934467167,0.381395787,-0.1442018896,0.1337888688,0.1369581521,-0.0157831535,0.1134748012,0.1468215734,-0.1264295727,-0.0556389503,0.1905207187,0.2522134185,0.1714027375,0.0542701334,-0.3509204388,-0.4462583363,-0.0797010213,-0.2040190697,0.0652356073,-0.2956978679,0.1840091497,-0.229904905,0.1598493457,-0.1872381717,-0.2770086527,0.199253276,0.2211045772,0.0691296235,-0.0560180396,-0.1147577688,-0.0540781133,-0.0689443648,0.081534408,-0.4202454388,0.3430917263,-0.2603977621,-0.2390125692,0.0644561425,0.0965775475,0.1995361745,-0.1361991763,-0.2320535332,0.4403236806,0.1650057137,-0.2004151046,-0.1622930318,0.077113837,0.1574232131,-0.1858457625,0.0220011715,0.3202284276,0.098340556,0.057860136,-0.0076144696,0.2805489302,-0.1693376452,0.1688168645,0.2262055576,0.3731301129,0.2607486248,0.1023362055,-0.0886058956,-0.2534463406,0.4879236221,0.0339732207,0.2914642394,-0.2156838924,-0.5336611271,0.0760332271,0.3073705435,0.148070097,0.1422265172,-0.0100683542,-0.1057600006,0.0769602954,0.2739711702,0.1093880683,0.3183995187,0.0714439526,-0.2827411294,-0.06643226,-0.268722266,-0.1903395504,0.2828251719,0.0425259769,0.1601824164,0.1422438473,0.2664467096,0.0376957208,-0.38587901,-0.4234346151,-0.0569498241,0.3714588881,-0.1407547593,0.2643475235,-0.3491425514,-0.4249199927,-0.2960482836,-0.2388049364,-0.2406001985,-0.3623484075,-0.0960663483,0.1309019029,0.0306267813,0.23544842,-0.1495063454,-0.0630283803,0.1627140343,0.1025079563,-0.0452507511,-0.1686990857,-0.2626588941,0.1136495322,0.1955718696,-0.0830842555,0.376789093,-0.0422195643,-0.1625643373,-0.2199974954,-0.4036245048,0.1120402217,-0.0680637285,0.198832646,0.5051586032,0.0949800089,-0.0435776599,-0.1564034969,0.4378064275,-0.056366723,0.1829486191,0.1709958315,-0.0545085445,-0.3030057251,-0.1642799526,-0.4381222725,-0.1721548587,-0.1750888675,0.1431864202,0.3657177091,0.367105037,0.5672655702,0.0486722551,0.1627244651,-0.1498387605,0.1697799414,-0.0384772494,0.0926329643,0.1827802211,-0.2975689173,-0.4033008814,-0.205091387,0.1107111797,0.0534857623,-0.0109257596,-0.4335919619,-0.1561717689,-0.2316404432,0.2016340494,-0.0561200008,0.0157572627,0.2922447622,0.1445402205,-0.0896595716,-0.0230043251,-0.0041051153,-0.0687814355,-0.1858177781,0.1621867567,0.02480096,0.2922298312,-0.0115637807,0.2237012982,0.3424477279,-0.21517995,0.127971068,-0.1084780023,0.4514208436,-0.1578459591,-0.4963950813,0.2596535981,-0.0834074542,-0.0708971173,0.1117999703,-0.0996082053,-0.2902067304,-0.1338139623,0.3741458654,-0.0089206379,-0.070991613,0.0830765516,-0.1850357056,0.3181610405,0.0026355914,0.2742580771,-0.1590847522,-0.2411731929,-0.1272432357,0.1199280769,-0.009803867,-0.1198805571,-0.5893691778,-0.3102943003,-0.247439459,0.3407361805,0.1813260019,0.7312636375,-0.0435702987,0.1915239096,-0.2322877198,-0.0519199781,0.2175544798,-0.326344192,-0.1254757047,0.5302116871,-0.0307003837,-0.118570663,-0.1065544412,-0.4047205448,0.3956201375,0.1058988646,0.2219056487,-0.0387296975,-0.0528825969,0.3625879288,0.1771215349,-0.1430920213,0.0908260271,-0.062347617,-0.0892777219,-0.1557316929,-0.1180961952,-0.192764461,0.1466251016,-0.0049093072,0.1919082701,-0.1392062753,-0.1257688999,-0.0748906657,0.1233940497,0.1236951351,0.0033262207,-0.0421622843,-0.0984143764,0.2123972028,0.1988953501,0.111814931,-0.3817193508,-0.220465377,-0.1041044816,-0.0807111338,0.2407717109,0.0738160312,-0.2533874512,0.0336981937,-0.0934980437,0.1150830686,-0.0288711935,-0.2338213772,0.5240261555,0.0251873322,-0.3135377169,-0.4343085885,0.2066931278,0.0136043653,0.1050095633,0.3896684647,-0.2068133503,-0.3086203039,0.1892786175,0.1139311194,0.6405631304,0.0851485878,0.0652173385,0.2540678382,-0.0935600325,0.6293442845,0.154667303,0.2402301282,-0.3571700752,-0.2086233944,-0.0803108662,-0.1555121988,0.4733545184,0.050789699,-0.3131177723,0.2505069971,-0.0391839445,0.1385234892,0.0359818265,0.2782002389,0.0778178126,-0.1967809945,-0.5601501465,0.1838223338,-0.2324073017,-0.0873937979,-0.3194411397,0.0811043978,0.1707180589,-0.1656274945,-0.4413450956,0.1346510947,-0.1211197376,0.3450767696,0.3261592686,-0.3975536823,0.1371126026,0.1328669786,-0.0263587907,0.2817955017,-0.1486691386,-0.0861459672,-0.04510298,-0.1509840935,0.0990662873,0.1168280244,0.4106481075,-0.0275178254,0.0686404184,0.1827409565,0.0321697704,-0.1284331381,-0.0911368951,-0.0259144288,0.1404956877,-0.2732411027,-0.1399982572,-0.0654593408,-0.1451391429,-0.3277772665,0.160689801,0.1787601113,-0.1430871338,0.2235409617,-0.064017199,0.0266124457,-0.0206706487,0.1997918636,0.0545464233,-0.0542282909,0.4666308463,0.4435594678,-0.2183807343,-0.3694681823,0.055177819,-0.1608375758,-0.3175583184,0.2667807043,-0.1879154742,0.1919151545,-0.0624178648,0.1855025887,-0.1084150523,0.2329244018,0.0201102067,-0.6296340227,-0.1027127504,0.0046861414,0.1599672288,0.1682860702,0.0211270824,0.163348332,0.0623625815,0.1665311903,-0.3393384218,-0.0928668007,-0.2718740404,0.1751075983,0.0623882599,0.0062928703,0.0212407559,-0.0425847955,0.1828920096,0.1107411385,-0.0693013147,-0.2771647274,-0.2256083786,0.0743487477,0.1718113869,-0.507412076,0.0664311126,-0.0953751802,-0.2018558085,-0.2209293842,-0.1938804686,0.0060284715,-0.1222522259,0.3884958923,-0.091798149,-0.0195536781,0.0311304871,0.088862054,-0.0639946982,0.1822521836,0.0906904861,0.3359932899,0.1552698314,-0.0223114416,-0.4787735641,0.0881573036,-0.1612366438,0.0551174805,0.4794063866,-0.5633710623,-0.0244878326,-0.0482899845,0.5955436826,0.3907880783,-0.2083756775,0.0840869248,0.2443304658,0.2077737153,-0.3391135931,-0.1119453683,0.2847105563,0.1861761063,0.0454925708,0.0799686164,0.0038469923,-0.0595146827,-0.1182622686,0.029866742,0.1543686539,-0.5287248492,0.2331763953,0.3293747306,0.1726527959,-0.0026453021,0.166285947,0.1339374334,0.1855650842,0.3369790018,0.0011743638,0.2679281831,0.2355300933,0.0644427314,0.236332953,-0.129755646,0.0706022382,0.1460008174,0.2153232545,0.2068515718,0.0903030559,0.3038617074,-0.3223173916,-0.092349872,0.0277668256,0.2207098454,-0.1179727167,-0.0328400135,-0.222666204,-0.0682456419,-0.1485171914,0.0043878197,-0.2459124029,-0.0806655288,0.5146118999,0.1833415926,-0.1658653319,-0.8891292214,0.2164248377,0.1544321924,0.0647152662,-0.314719975,0.2495400906,0.1710821092,0.0437222868,-0.2234827876,0.7008237839,0.5018439889,0.3941646218,0.1825341433,-0.0488634109,-0.2682225406,-0.1042501628,0.0419572145,0.1700416356,0.1123287678,0.0595856346,0.3015963137,0.133301422,-0.149044916,0.2988526821,0.2556347251,0.0375600196,-0.0993071422,0.4731218219,-0.0470735952,-0.2465269268,-0.135557279,-0.0301100947,-0.3524824679,0.1153910905,0.540610373,0.2814173102,0.0571114309,-0.0885857344,0.1267948747,0.1267552376,0.4031401277,0.475553751,0.3269433081,-0.1795827299,-0.1937531084,-0.2992210388,0.157597959,-0.2197696865,-0.0460958704,0.1135376394,0.0597481094,-0.0773707703,0.2345739752,0.0274125598,0.0945409834,-0.0248964783,0.0568600632,-0.3780333996,-0.1162496805,0.2867147028,0.0270568617,-0.0278433077,-0.196413368,0.1512077898,-0.1286519319,0.1126163304,-0.3674302101,-0.0651417524,-0.1094167382,0.2097267956,0.0312755145,0.4317313135,0.2470489889,-0.0309630502,-0.2971615791,-0.2883206904,-0.4377296567,-0.1834001243,0.5462127328,0.1380329877,0.53362602,-0.1627794355,-0.243976742,-0.4112980664,0.433110863,-0.3177213073,-0.0555907972,-0.2225956321,0.2077827901,-0.2010600567,-0.0666386336,0.0329128429,0.2839014828,-0.0416875146,0.2112437934,-0.4552669227,-0.2773829699,0.259721756,-0.7163981199,-0.3755799532,-0.1382011771,0.1000654697,-0.0333187468,-0.1368526518,-0.3827458918,-0.0606800728,0.336499691,-0.1395791024,-0.2957063317,0.073035188,0.1560716629,0.1417677701,-0.0939191952,0.6182773113,0.0997412354,-0.3169441819,0.0998760164,-0.0598640777]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"I am facing probably facing similar issues with \r\n\r\n`wiki40b_en_100_0`","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":9,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nI am facing probably facing similar issues with \r\n\r\n`wiki40b_en_100_0`","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Could you try to run `dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")` once before calling the script ?\r\n\r\nIt looks like several processes try to create the dataset in arrow format at the same time. If the dataset is already created it should be fine","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":43,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nCould you try to run `dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")` once before calling the script ?\r\n\r\nIt looks like several processes try to create the dataset in arrow format at the same time. If the dataset is already created it should be fine","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Thanks! I tested on 328MB text data on `n1-standard-8 (8 vCPUs, 30 GB memory)`. The main script ran without any issue, but it seems to require a huge space in the drive.\r\n\r\nAs suggested, I ran the following script before running the pre-training command with `xla_spawn.py`.\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n\r\nfile_path=\"your_file_name\"\r\nload_dataset(\"text\", data_files=file_path, split=\"train\")\r\n```\r\nThis will create `text-train.arrow` under the default cache directory. Then, I run the script with `xla_spawn.py`. It will load data from the cached file. My understanding is that there's no other way but to do this two-step process with the current version (0.4) of `nlp`.\r\n\r\nDuring another caching process that happens in the main script:\r\n\r\n```\r\n08\/26\/2020 09:19:51 - INFO - nlp.utils.info_utils - All the checksums matched successfully for post processing resources\r\n08\/26\/2020 09:19:53 - INFO - nlp.arrow_dataset - Caching processed dataset at \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\/cache-f90f341e5308a7469\r\n8d872bcc88f9c0e.arrow\r\n```\r\n\r\n`nlp` generates a temporary file per core, each of which is three times larger than the original text data. If each process is actually writing on the disk, you will need a huge amount of space in your drive. (Maybe I'm missing something.)\r\n\r\n```\r\n-rw-r--r-- 1 ***** ***** 674 Aug 26 09:19 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 09:19 LICENSE\r\n-rw-r--r-- 1 ***** ***** 332M Aug 26 09:10 text-train.arrow\r\n-rw------- 1 ***** ***** 940M Aug 26 09:31 tmp0k43sazw\r\n-rw------- 1 ***** ***** 940M Aug 26 09:31 tmp7sxs9mj5\r\n-rw------- 1 ***** ***** 939M Aug 26 09:31 tmpbbiqw2vp\r\n-rw------- 1 ***** ***** 937M Aug 26 09:31 tmpjxb5ptyu\r\n-rw------- 1 ***** ***** 933M Aug 26 09:31 tmpk3hkdh0e\r\n-rw------- 1 ***** ***** 944M Aug 26 09:31 tmpnoalwftz\r\n-rw------- 1 ***** ***** 931M Aug 26 09:31 tmpuxdr_dz3\r\n-rw------- 1 ***** ***** 945M Aug 26 09:31 tmpxjyuy6dk\r\n```\r\nAfter the caching process, they seem to be merged into one file.\r\n\r\n```\r\n-rw------- 1 ***** ***** 989M Aug 26 09:32 cache-f90f341e5308a74698d872bcc88f9c0e.arrow\r\n-rw-r--r-- 1 ***** ***** 674 Aug 26 09:19 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 09:19 LICENSE\r\n-rw-r--r-- 1 ***** ***** 332M Aug 26 09:10 text-train.arrow\r\n```","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":336,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nThanks! I tested on 328MB text data on `n1-standard-8 (8 vCPUs, 30 GB memory)`. The main script ran without any issue, but it seems to require a huge space in the drive.\r\n\r\nAs suggested, I ran the following script before running the pre-training command with `xla_spawn.py`.\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n\r\nfile_path=\"your_file_name\"\r\nload_dataset(\"text\", data_files=file_path, split=\"train\")\r\n```\r\nThis will create `text-train.arrow` under the default cache directory. Then, I run the script with `xla_spawn.py`. It will load data from the cached file. My understanding is that there's no other way but to do this two-step process with the current version (0.4) of `nlp`.\r\n\r\nDuring another caching process that happens in the main script:\r\n\r\n```\r\n08\/26\/2020 09:19:51 - INFO - nlp.utils.info_utils - All the checksums matched successfully for post processing resources\r\n08\/26\/2020 09:19:53 - INFO - nlp.arrow_dataset - Caching processed dataset at \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\/cache-f90f341e5308a7469\r\n8d872bcc88f9c0e.arrow\r\n```\r\n\r\n`nlp` generates a temporary file per core, each of which is three times larger than the original text data. If each process is actually writing on the disk, you will need a huge amount of space in your drive. (Maybe I'm missing something.)\r\n\r\n```\r\n-rw-r--r-- 1 ***** ***** 674 Aug 26 09:19 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 09:19 LICENSE\r\n-rw-r--r-- 1 ***** ***** 332M Aug 26 09:10 text-train.arrow\r\n-rw------- 1 ***** ***** 940M Aug 26 09:31 tmp0k43sazw\r\n-rw------- 1 ***** ***** 940M Aug 26 09:31 tmp7sxs9mj5\r\n-rw------- 1 ***** ***** 939M Aug 26 09:31 tmpbbiqw2vp\r\n-rw------- 1 ***** ***** 937M Aug 26 09:31 tmpjxb5ptyu\r\n-rw------- 1 ***** ***** 933M Aug 26 09:31 tmpk3hkdh0e\r\n-rw------- 1 ***** ***** 944M Aug 26 09:31 tmpnoalwftz\r\n-rw------- 1 ***** ***** 931M Aug 26 09:31 tmpuxdr_dz3\r\n-rw------- 1 ***** ***** 945M Aug 26 09:31 tmpxjyuy6dk\r\n```\r\nAfter the caching process, they seem to be merged into one file.\r\n\r\n```\r\n-rw------- 1 ***** ***** 989M Aug 26 09:32 cache-f90f341e5308a74698d872bcc88f9c0e.arrow\r\n-rw-r--r-- 1 ***** ***** 674 Aug 26 09:19 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 09:19 LICENSE\r\n-rw-r--r-- 1 ***** ***** 332M Aug 26 09:10 text-train.arrow\r\n```","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Again it looks like every process tries to tokenize the full dataset at the same time.\r\nIf you do the tokenization before calling `xla_spawn.py` once, then each process will then use the tokenized cached file `cache-f90f341e5308a74698d872bcc88f9c0e.arrow` and not recompute it.\r\n\r\nNot sure if there's a better way to do that cc @julien-c @thomwolf ","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":53,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nAgain it looks like every process tries to tokenize the full dataset at the same time.\r\nIf you do the tokenization before calling `xla_spawn.py` once, then each process will then use the tokenized cached file `cache-f90f341e5308a74698d872bcc88f9c0e.arrow` and not recompute it.\r\n\r\nNot sure if there's a better way to do that cc @julien-c @thomwolf ","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"I wrote a separate script just for preparing a cached file, including tokenization. Each process did use the tokenized cached file.\r\n\r\nCurrently I'm testing the pipeline on 24GB text data. It took about 1.5 hour to create a cached file on `n1-highmem-16 (16 vCPUs, 104 GB memory)`. I assume loading this cached file in the main script with `xla_spawn.py` won't be an issue (even if there are 8 processes).\r\n\r\n```\r\ntotal 98G\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 26 13:38 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 26 12:24 ..\r\n-rw------- 1 ***** ***** 74G Aug 26 13:38 cache-a7aa04134ba7b1aff5d9710f14a4e334.arrow\r\n-rw-r--r-- 1 ***** ***** 681 Aug 26 12:24 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 12:24 LICENSE\r\n-rw-r--r-- 1 ***** ***** 25G Aug 26 12:24 text-train.arrow\r\n```","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":127,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nI wrote a separate script just for preparing a cached file, including tokenization. Each process did use the tokenized cached file.\r\n\r\nCurrently I'm testing the pipeline on 24GB text data. It took about 1.5 hour to create a cached file on `n1-highmem-16 (16 vCPUs, 104 GB memory)`. I assume loading this cached file in the main script with `xla_spawn.py` won't be an issue (even if there are 8 processes).\r\n\r\n```\r\ntotal 98G\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 26 13:38 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 26 12:24 ..\r\n-rw------- 1 ***** ***** 74G Aug 26 13:38 cache-a7aa04134ba7b1aff5d9710f14a4e334.arrow\r\n-rw-r--r-- 1 ***** ***** 681 Aug 26 12:24 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 12:24 LICENSE\r\n-rw-r--r-- 1 ***** ***** 25G Aug 26 12:24 text-train.arrow\r\n```","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Yes loading the cached file should be fine from different processes","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":11,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nYes loading the cached file should be fine from different processes","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Sorry, I thought it was working, but actually the second call doesn't use the cached file that was generated separately, and it will generate another cache-****.arrorw file with a different name. If I run the training script again (with `xla_spawn.py`), it will use the second cached file, which was generated by the training script itself in the previous run.\r\n\r\n```\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 26 15:35 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 26 15:29 ..\r\n-rw------- 1 ***** ***** 99M Aug 26 15:35 cache-0d77dfce704493dbe63f071eed6a5431.arrow\r\n-rw------- 1 ***** ***** 99M Aug 26 15:29 cache-69633651476e943b93c89ace715f9487.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 26 15:33 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 15:33 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 26 15:29 text-train.arrow\r\n```","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":124,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nSorry, I thought it was working, but actually the second call doesn't use the cached file that was generated separately, and it will generate another cache-****.arrorw file with a different name. If I run the training script again (with `xla_spawn.py`), it will use the second cached file, which was generated by the training script itself in the previous run.\r\n\r\n```\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 26 15:35 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 26 15:29 ..\r\n-rw------- 1 ***** ***** 99M Aug 26 15:35 cache-0d77dfce704493dbe63f071eed6a5431.arrow\r\n-rw------- 1 ***** ***** 99M Aug 26 15:29 cache-69633651476e943b93c89ace715f9487.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 26 15:33 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 15:33 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 26 15:29 text-train.arrow\r\n```","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"So if I understand correctly it means that the cached file generated by your separated script is different by the one used by the training script ?","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":27,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nSo if I understand correctly it means that the cached file generated by your separated script is different by the one used by the training script ?","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Yes.\r\n\r\n1. `cache-69633651476e943b93c89ace715f9487.arrow` was generated with a separate script. \r\n2. I ran the entire script with `xla_spawn.py`.\r\n3. `cache-69633651476e943b93c89ace715f9487.arrow` is not used.\r\n4. `cache-0d77dfce704493dbe63f071eed6a5431.arrow` is created.\r\n5. training starts...\r\n\r\nNow, if I kill the process at step 5, and do the step 2 again, it will use `cache-0d77dfce704493dbe63f071eed6a5431.arrow` (cached file created at step 4) without any issue.\r\n\r\nI used the following to generate the first cached file.\r\n```python\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\n```","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":85,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nYes.\r\n\r\n1. `cache-69633651476e943b93c89ace715f9487.arrow` was generated with a separate script. \r\n2. I ran the entire script with `xla_spawn.py`.\r\n3. `cache-69633651476e943b93c89ace715f9487.arrow` is not used.\r\n4. `cache-0d77dfce704493dbe63f071eed6a5431.arrow` is created.\r\n5. training starts...\r\n\r\nNow, if I kill the process at step 5, and do the step 2 again, it will use `cache-0d77dfce704493dbe63f071eed6a5431.arrow` (cached file created at step 4) without any issue.\r\n\r\nI used the following to generate the first cached file.\r\n```python\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\n```","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"1. Here's the log from the first step.\r\n```\r\nDownloading and preparing dataset text\/default-e84dd29acc4ad9ef (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDataset text downloaded and prepared to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d. Subsequent calls will reuse this data.\r\n```\r\nThere's a file named `cache-7b1440ba7077af0f0d9035b5a55d01fc.arrow`, so it did create a cached file.\r\n```\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 26 15:59 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 26 15:58 ..\r\n-rw------- 1 ***** ***** 99M Aug 26 15:59 cache-7b1440ba7077af0f0d9035b5a55d01fc.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 26 15:58 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 15:58 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 26 15:58 text-train.arrow\r\n```\r\n2. Ideally, `cache-7b1440ba7077af0f0d9035b5a55d01fc.arrow` should be used in `run_language_modeling.py` (modified version using `nlp`) with `xla_spawn.py`. But it looks like it's creating a new cached file.\r\n\r\n```\r\n08\/26\/2020 16:13:03 - INFO - filelock - Lock 139635836351096 released on \/home\/*****\/.cache\/huggingface\/datasets\/3e34209a2741375a1db1ff03bf1abba1a9bd0e6016912d3ead0114b9d1ca2685.202fa4f84f552bff1f5400ae012663839c61efb3de068c6c8722d34ac0ea6192\r\n.py.lock\r\n08\/26\/2020 16:13:03 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Overwrite dataset info from restored data version.\r\n08\/26\/2020 16:13:03 - INFO - nlp.info - Loading Dataset info from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Reusing dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Constructing Dataset for split train, from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:13:03 - INFO - nlp.utils.info_utils - All the checksums matched successfully for post processing resources\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Overwrite dataset info from restored data version.\r\n08\/26\/2020 16:13:03 - INFO - nlp.info - Loading Dataset info from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Reusing dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Constructing Dataset for split train, from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:13:03 - INFO - nlp.utils.info_utils - All the checksums matched successfully for post processing resources\r\n08\/26\/2020 16:13:05 - INFO - nlp.arrow_dataset - Caching processed dataset at \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\/cache-0d77dfce704493dbe\r\n63f071eed6a5431.arrow\r\n^M 0%| | 0\/100 [00:00<?, ?it\/s]08\/26\/2020 16:13:05 - INFO - nlp.arrow_dataset - Caching processed dataset at \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6\r\nfe661fe4d070d380d\/cache-0d77dfce704493dbe63f071eed6a5431.arrow\r\n```\r\n\r\nThere are two cached files in the directory:\r\n\r\n```\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 26 16:14 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 26 15:58 ..\r\n-rw------- 1 ***** ***** 99M Aug 26 16:14 cache-0d77dfce704493dbe63f071eed6a5431.arrow\r\n-rw------- 1 ***** ***** 99M Aug 26 15:59 cache-7b1440ba7077af0f0d9035b5a55d01fc.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 26 16:13 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 16:13 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 26 15:58 text-train.arrow\r\n```\r\n\r\nIf I kill the process, and run it again, it will use the second cached file.\r\n\r\n```\r\n08\/26\/2020 16:19:52 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/26\/2020 16:19:52 - INFO - nlp.builder - Overwrite dataset info from restored data version.\r\n08\/26\/2020 16:19:52 - INFO - nlp.info - Loading Dataset info from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:19:52 - INFO - nlp.builder - Reusing dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/26\/2020 16:19:52 - INFO - nlp.builder - Constructing Dataset for split train, from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:19:52 - INFO - nlp.utils.info_utils - All the checksums matched successfully for post processing resources\r\n08\/26\/2020 16:19:53 - INFO - nlp.arrow_dataset - Loading cached processed dataset at \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\/cache-0d77dfce70\r\n4493dbe63f071eed6a5431.arrow\r\n08\/26\/2020 16:19:53 - INFO - nlp.arrow_dataset - Set __getitem__(key) output type to torch for ['input_ids'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":539,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\n1. Here's the log from the first step.\r\n```\r\nDownloading and preparing dataset text\/default-e84dd29acc4ad9ef (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDataset text downloaded and prepared to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d. Subsequent calls will reuse this data.\r\n```\r\nThere's a file named `cache-7b1440ba7077af0f0d9035b5a55d01fc.arrow`, so it did create a cached file.\r\n```\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 26 15:59 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 26 15:58 ..\r\n-rw------- 1 ***** ***** 99M Aug 26 15:59 cache-7b1440ba7077af0f0d9035b5a55d01fc.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 26 15:58 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 15:58 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 26 15:58 text-train.arrow\r\n```\r\n2. Ideally, `cache-7b1440ba7077af0f0d9035b5a55d01fc.arrow` should be used in `run_language_modeling.py` (modified version using `nlp`) with `xla_spawn.py`. But it looks like it's creating a new cached file.\r\n\r\n```\r\n08\/26\/2020 16:13:03 - INFO - filelock - Lock 139635836351096 released on \/home\/*****\/.cache\/huggingface\/datasets\/3e34209a2741375a1db1ff03bf1abba1a9bd0e6016912d3ead0114b9d1ca2685.202fa4f84f552bff1f5400ae012663839c61efb3de068c6c8722d34ac0ea6192\r\n.py.lock\r\n08\/26\/2020 16:13:03 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Overwrite dataset info from restored data version.\r\n08\/26\/2020 16:13:03 - INFO - nlp.info - Loading Dataset info from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Reusing dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Constructing Dataset for split train, from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:13:03 - INFO - nlp.utils.info_utils - All the checksums matched successfully for post processing resources\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Overwrite dataset info from restored data version.\r\n08\/26\/2020 16:13:03 - INFO - nlp.info - Loading Dataset info from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Reusing dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/26\/2020 16:13:03 - INFO - nlp.builder - Constructing Dataset for split train, from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:13:03 - INFO - nlp.utils.info_utils - All the checksums matched successfully for post processing resources\r\n08\/26\/2020 16:13:05 - INFO - nlp.arrow_dataset - Caching processed dataset at \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\/cache-0d77dfce704493dbe\r\n63f071eed6a5431.arrow\r\n^M 0%| | 0\/100 [00:00<?, ?it\/s]08\/26\/2020 16:13:05 - INFO - nlp.arrow_dataset - Caching processed dataset at \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6\r\nfe661fe4d070d380d\/cache-0d77dfce704493dbe63f071eed6a5431.arrow\r\n```\r\n\r\nThere are two cached files in the directory:\r\n\r\n```\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 26 16:14 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 26 15:58 ..\r\n-rw------- 1 ***** ***** 99M Aug 26 16:14 cache-0d77dfce704493dbe63f071eed6a5431.arrow\r\n-rw------- 1 ***** ***** 99M Aug 26 15:59 cache-7b1440ba7077af0f0d9035b5a55d01fc.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 26 16:13 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 26 16:13 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 26 15:58 text-train.arrow\r\n```\r\n\r\nIf I kill the process, and run it again, it will use the second cached file.\r\n\r\n```\r\n08\/26\/2020 16:19:52 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/26\/2020 16:19:52 - INFO - nlp.builder - Overwrite dataset info from restored data version.\r\n08\/26\/2020 16:19:52 - INFO - nlp.info - Loading Dataset info from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:19:52 - INFO - nlp.builder - Reusing dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/26\/2020 16:19:52 - INFO - nlp.builder - Constructing Dataset for split train, from \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\n08\/26\/2020 16:19:52 - INFO - nlp.utils.info_utils - All the checksums matched successfully for post processing resources\r\n08\/26\/2020 16:19:53 - INFO - nlp.arrow_dataset - Loading cached processed dataset at \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\/cache-0d77dfce70\r\n4493dbe63f071eed6a5431.arrow\r\n08\/26\/2020 16:19:53 - INFO - nlp.arrow_dataset - Set __getitem__(key) output type to torch for ['input_ids'] columns (when key is int or slice) and don't output other (un-formatted) columns.\r\n```","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Thanks for all the details.\r\nThe two cached files are supposed to be the same. I suspect that the caching has a problem with the tokenizer.\r\nWhich tokenizer did you use ?","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":32,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nThanks for all the details.\r\nThe two cached files are supposed to be the same. I suspect that the caching has a problem with the tokenizer.\r\nWhich tokenizer did you use ?","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"I trained a byte-level BPE tokenizer on my data with `tokenziers` library following this [example](https:\/\/github.com\/huggingface\/tokenizers\/blob\/master\/bindings\/python\/examples\/train_bytelevel_bpe.py).\r\n\r\nAnd I put these model files in a directory named `\"model_name\"`. I also put config.json, which is the original RoBERTa config file.\r\n\r\n```bash\r\n%ls model_name\r\nconfig.json merges.txt vocab.json\r\n```\r\n\r\n[This](https:\/\/github.com\/huggingface\/transformers\/blob\/4bd7be9a4268221d2a0000c7e8033aaeb365c03b\/examples\/language-modeling\/run_language_modeling.py#L196) is the line where `run_language_modeling.py` loads the tokenier.\r\n\r\n```python\r\ntokenizer = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir)\r\n```\r\n\r\nI use `\"model_name\"` for `model_args.tokenizer_name`. I don't specify `model_args.cache_dir`. It is 'None' by default.","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":73,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nI trained a byte-level BPE tokenizer on my data with `tokenziers` library following this [example](https:\/\/github.com\/huggingface\/tokenizers\/blob\/master\/bindings\/python\/examples\/train_bytelevel_bpe.py).\r\n\r\nAnd I put these model files in a directory named `\"model_name\"`. I also put config.json, which is the original RoBERTa config file.\r\n\r\n```bash\r\n%ls model_name\r\nconfig.json merges.txt vocab.json\r\n```\r\n\r\n[This](https:\/\/github.com\/huggingface\/transformers\/blob\/4bd7be9a4268221d2a0000c7e8033aaeb365c03b\/examples\/language-modeling\/run_language_modeling.py#L196) is the line where `run_language_modeling.py` loads the tokenier.\r\n\r\n```python\r\ntokenizer = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir)\r\n```\r\n\r\nI use `\"model_name\"` for `model_args.tokenizer_name`. I don't specify `model_args.cache_dir`. It is 'None' by default.","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"In my separated script for caching, I'm using `use_fast=True` when initializing a tokenizer.\r\n\r\n```python\r\ntokenizer = AutoTokenizer.from_pretrained(args.config_name, use_fast=True)\r\n```\r\nI wasn't using that option in the main script. That could be the reason...","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":33,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nIn my separated script for caching, I'm using `use_fast=True` when initializing a tokenizer.\r\n\r\n```python\r\ntokenizer = AutoTokenizer.from_pretrained(args.config_name, use_fast=True)\r\n```\r\nI wasn't using that option in the main script. That could be the reason...","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Yea it could definitely explain why you have two different cache files.\r\nLet me know if using the same tokenizers on both sides fixes the issue","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":26,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nYea it could definitely explain why you have two different cache files.\r\nLet me know if using the same tokenizers on both sides fixes the issue","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"It still creates a new file even if I remove `use_fast=True`... \r\n\r\nHere's the script used to create a cached file.\r\n```python \r\n#!\/usr\/bin\/env python3\r\n\r\nimport argparse\r\n\r\nfrom transformers import AutoTokenizer\r\n\r\nfrom nlp import load_dataset\r\n\r\n\r\ndef main():\r\n parser = argparse.ArgumentParser(description='description')\r\n parser.add_argument('--config_name', type=str, help='Pretrained config name or path if not the same as model_name')\r\n parser.add_argument('--data_file', type=str, help='The input data file (a text file).')\r\n parser.add_argument('--block_size', type=int, default=-1, help='The training dataset will be truncated in block of this size for training')\r\n args = parser.parse_args()\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(args.config_name)\r\n\r\n dataset = load_dataset(\"text\", data_files=args.data_file, split=\"train\")\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\nHere's how the data is loaded in the modified `run_language_modeling.py`. [[original function](https:\/\/github.com\/huggingface\/transformers\/blob\/971d1802d009d9996b36a34a34477cee849ef39f\/examples\/language-modeling\/run_language_modeling.py#L128-L135)]\r\n\r\n```python\r\ndef get_dataset(args: DataTrainingArguments, tokenizer: PreTrainedTokenizer, evaluate=False):\r\n file_path = args.eval_data_file if evaluate else args.train_data_file\r\n split = \"validation\" if evaluate else \"train\"\r\n if args.line_by_line:\r\n # return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n return dataset\r\n\r\n else:\r\n return TextDataset(\r\n tokenizer=tokenizer, file_path=file_path, block_size=args.block_size, overwrite_cache=args.overwrite_cache\r\n )\r\n```\r\n\r\nProbably I don't need this part in the main script,\r\n\r\n```python\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n```\r\nand simply do this?\r\n```python\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\nreturn dataset\r\n```","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":207,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nIt still creates a new file even if I remove `use_fast=True`... \r\n\r\nHere's the script used to create a cached file.\r\n```python \r\n#!\/usr\/bin\/env python3\r\n\r\nimport argparse\r\n\r\nfrom transformers import AutoTokenizer\r\n\r\nfrom nlp import load_dataset\r\n\r\n\r\ndef main():\r\n parser = argparse.ArgumentParser(description='description')\r\n parser.add_argument('--config_name', type=str, help='Pretrained config name or path if not the same as model_name')\r\n parser.add_argument('--data_file', type=str, help='The input data file (a text file).')\r\n parser.add_argument('--block_size', type=int, default=-1, help='The training dataset will be truncated in block of this size for training')\r\n args = parser.parse_args()\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(args.config_name)\r\n\r\n dataset = load_dataset(\"text\", data_files=args.data_file, split=\"train\")\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\nHere's how the data is loaded in the modified `run_language_modeling.py`. [[original function](https:\/\/github.com\/huggingface\/transformers\/blob\/971d1802d009d9996b36a34a34477cee849ef39f\/examples\/language-modeling\/run_language_modeling.py#L128-L135)]\r\n\r\n```python\r\ndef get_dataset(args: DataTrainingArguments, tokenizer: PreTrainedTokenizer, evaluate=False):\r\n file_path = args.eval_data_file if evaluate else args.train_data_file\r\n split = \"validation\" if evaluate else \"train\"\r\n if args.line_by_line:\r\n # return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n return dataset\r\n\r\n else:\r\n return TextDataset(\r\n tokenizer=tokenizer, file_path=file_path, block_size=args.block_size, overwrite_cache=args.overwrite_cache\r\n )\r\n```\r\n\r\nProbably I don't need this part in the main script,\r\n\r\n```python\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n```\r\nand simply do this?\r\n```python\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\nreturn dataset\r\n```","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"You need this part in the main script or it will use the dataset that is not tokenized\r\n\r\n","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":18,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nYou need this part in the main script or it will use the dataset that is not tokenized\r\n\r\n","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"I can see that the tokenizer in `run_language_modeling.py` is not instantiated the same way as in your separated script.\r\nIndeed we can see L196:\r\n```python\r\ntokenizer = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir)\r\n```\r\nCould you try to make it so they are instantiated the exact same way please ?","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":46,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nI can see that the tokenizer in `run_language_modeling.py` is not instantiated the same way as in your separated script.\r\nIndeed we can see L196:\r\n```python\r\ntokenizer = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir)\r\n```\r\nCould you try to make it so they are instantiated the exact same way please ?","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"I updated my separated script, but it's creating a cached file again. If I don't use the `model_args.cache_dir`, both will get `None`, so they should be the same.\r\n\r\n```python\r\n#!\/usr\/bin\/env python3\r\nimport argparse\r\n\r\nfrom transformers import AutoTokenizer\r\nfrom nlp import load_dataset\r\n\r\ndef main():\r\n parser = argparse.ArgumentParser(description='description')\r\n parser.add_argument('--tokenizer_name', type=str, help='Pretrained tokenizer name or path if not the same as model_name')\r\n parser.add_argument('--data_file', type=str, help='The input data file (a text file).')\r\n parser.add_argument('--cache_dir', type=str, default=None, help='Where do you want to store the pretrained models downloaded from s3')\r\n parser.add_argument('--block_size', type=int, default=-1, help='The training dataset will be truncated in block of this size for training')\r\n\r\n model_args = parser.parse_args()\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir)\r\n\r\n dataset = load_dataset(\"text\", data_files=model_args.data_file, split=\"train\")\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=model_args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\nIs there a way to specify the cache file to load, and skip the re-computation?","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":143,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nI updated my separated script, but it's creating a cached file again. If I don't use the `model_args.cache_dir`, both will get `None`, so they should be the same.\r\n\r\n```python\r\n#!\/usr\/bin\/env python3\r\nimport argparse\r\n\r\nfrom transformers import AutoTokenizer\r\nfrom nlp import load_dataset\r\n\r\ndef main():\r\n parser = argparse.ArgumentParser(description='description')\r\n parser.add_argument('--tokenizer_name', type=str, help='Pretrained tokenizer name or path if not the same as model_name')\r\n parser.add_argument('--data_file', type=str, help='The input data file (a text file).')\r\n parser.add_argument('--cache_dir', type=str, default=None, help='Where do you want to store the pretrained models downloaded from s3')\r\n parser.add_argument('--block_size', type=int, default=-1, help='The training dataset will be truncated in block of this size for training')\r\n\r\n model_args = parser.parse_args()\r\n\r\n tokenizer = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir)\r\n\r\n dataset = load_dataset(\"text\", data_files=model_args.data_file, split=\"train\")\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=model_args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\nIs there a way to specify the cache file to load, and skip the re-computation?","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Could you also check that the `args.block_size` used in the lambda function is the same as well ?","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":18,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nCould you also check that the `args.block_size` used in the lambda function is the same as well ?","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"Here's a minimal working example to reproduce this issue.\r\n\r\nAssumption:\r\n- You have access to TPU.\r\n- You have installed `transformers` and `nlp`.\r\n- You have tokenizer files (`config.json`, `merges.txt`, `vocab.json`) under the directory named `model_name`.\r\n- You have `xla_spawn.py` (Download from https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py).\r\n- You have saved the following script as `prepare_cached_dataset.py`.\r\n\r\n```python\r\n#!\/usr\/bin\/env python3\r\nimport argparse\r\nfrom transformers import AutoTokenizer\r\nfrom nlp import load_dataset\r\n\r\ndef main():\r\n parser = argparse.ArgumentParser(description='description')\r\n parser.add_argument('--tokenizer_name', type=str, help='Pretrained tokenizer name or path if not the same as model_name')\r\n parser.add_argument('--data_file', type=str, help='The input data file (a text file).')\r\n parser.add_argument('--cache_dir', type=str, default=None, help='Where do you want to store the pretrained models downloaded from s3')\r\n parser.add_argument('--block_size', type=int, default=-1, help='The training dataset will be truncated in block of this size for training')\r\n parser.add_argument('--tpu_num_cores', type=int, default=1, help='Number of TPU cores to use (1 or 8). For xla_apwan.py')\r\n model_args = parser.parse_args()\r\n \r\n tokenizer = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir, use_fast=True)\r\n \r\n dataset = load_dataset(\"text\", data_files=model_args.data_file, split=\"train\")\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=model_args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n\r\ndef _mp_fn(index):\r\n # For xla_spawn (TPUs)\r\n main()\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\n- Run the following command. Replace `your_training_data` with some text file.\r\n\r\n```bash\r\nexport TRAIN_DATA=your_training_data\r\n\r\npython prepare_cached_dataset.py \\\r\n--tokenizer_name=model_name \\\r\n--block_size=512 \\\r\n--data_file=$TRAIN_DATA\r\n```\r\n- Check the cached directory.\r\n```bash\r\nls -lha \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\ntotal 132M\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 28 13:08 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 28 13:08 ..\r\n-rw------- 1 ***** ***** 99M Aug 28 13:08 cache-bfc7cb0702426d19242db5e8c079f04b.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 28 13:08 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 28 13:08 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 28 13:08 text-train.arrow\r\n```\r\n\r\n- Run the same script again. (The output should be just `Using custom data configuration default`.)\r\n```\r\npython prepare_cached_dataset.py \\\r\n--tokenizer_name=model_name \\\r\n--block_size=512 \\\r\n--data_file=$TRAIN_DATA\r\n```\r\n- Check the cached directory.\r\n```bash\r\nls -lha \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\ntotal 132M\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 28 13:08 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 28 13:08 ..\r\n-rw------- 1 ***** ***** 99M Aug 28 13:08 cache-bfc7cb0702426d19242db5e8c079f04b.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 28 13:20 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 28 13:20 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 28 13:08 text-train.arrow\r\n```\r\n- The cached file (`cache-bfc7cb0702426d19242db5e8c079f04b.arrow`) is reused.\r\n- Now, run this script with `xla_spawn.py`. Ideally, it should reuse the cached file, however, you will see each process is creating a cache file again.\r\n\r\n```bash\r\npython xla_spawn.py --num_cores 8 \\\r\nprepare_cached_dataset.py \\\r\n--tokenizer_name=model_name \\\r\n--block_size=512 \\\r\n--data_file=$TRAIN_DATA\r\n```\r\n\r\n- Check the cached directory. There are two arrrow files.\r\n```bash\r\nls -lha \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\ntotal 230M\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 28 13:25 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 28 13:08 ..\r\n-rw------- 1 ***** ***** 99M Aug 28 13:08 cache-bfc7cb0702426d19242db5e8c079f04b.arrow\r\n-rw------- 1 ***** ***** 99M Aug 28 13:25 cache-e0e2313e49c8a110aafcc8133154c19a.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 28 13:24 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 28 13:24 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 28 13:08 text-train.arrow\r\n```\r\n","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":482,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nHere's a minimal working example to reproduce this issue.\r\n\r\nAssumption:\r\n- You have access to TPU.\r\n- You have installed `transformers` and `nlp`.\r\n- You have tokenizer files (`config.json`, `merges.txt`, `vocab.json`) under the directory named `model_name`.\r\n- You have `xla_spawn.py` (Download from https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py).\r\n- You have saved the following script as `prepare_cached_dataset.py`.\r\n\r\n```python\r\n#!\/usr\/bin\/env python3\r\nimport argparse\r\nfrom transformers import AutoTokenizer\r\nfrom nlp import load_dataset\r\n\r\ndef main():\r\n parser = argparse.ArgumentParser(description='description')\r\n parser.add_argument('--tokenizer_name', type=str, help='Pretrained tokenizer name or path if not the same as model_name')\r\n parser.add_argument('--data_file', type=str, help='The input data file (a text file).')\r\n parser.add_argument('--cache_dir', type=str, default=None, help='Where do you want to store the pretrained models downloaded from s3')\r\n parser.add_argument('--block_size', type=int, default=-1, help='The training dataset will be truncated in block of this size for training')\r\n parser.add_argument('--tpu_num_cores', type=int, default=1, help='Number of TPU cores to use (1 or 8). For xla_apwan.py')\r\n model_args = parser.parse_args()\r\n \r\n tokenizer = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir, use_fast=True)\r\n \r\n dataset = load_dataset(\"text\", data_files=model_args.data_file, split=\"train\")\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=model_args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n\r\ndef _mp_fn(index):\r\n # For xla_spawn (TPUs)\r\n main()\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\n- Run the following command. Replace `your_training_data` with some text file.\r\n\r\n```bash\r\nexport TRAIN_DATA=your_training_data\r\n\r\npython prepare_cached_dataset.py \\\r\n--tokenizer_name=model_name \\\r\n--block_size=512 \\\r\n--data_file=$TRAIN_DATA\r\n```\r\n- Check the cached directory.\r\n```bash\r\nls -lha \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\ntotal 132M\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 28 13:08 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 28 13:08 ..\r\n-rw------- 1 ***** ***** 99M Aug 28 13:08 cache-bfc7cb0702426d19242db5e8c079f04b.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 28 13:08 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 28 13:08 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 28 13:08 text-train.arrow\r\n```\r\n\r\n- Run the same script again. (The output should be just `Using custom data configuration default`.)\r\n```\r\npython prepare_cached_dataset.py \\\r\n--tokenizer_name=model_name \\\r\n--block_size=512 \\\r\n--data_file=$TRAIN_DATA\r\n```\r\n- Check the cached directory.\r\n```bash\r\nls -lha \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\ntotal 132M\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 28 13:08 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 28 13:08 ..\r\n-rw------- 1 ***** ***** 99M Aug 28 13:08 cache-bfc7cb0702426d19242db5e8c079f04b.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 28 13:20 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 28 13:20 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 28 13:08 text-train.arrow\r\n```\r\n- The cached file (`cache-bfc7cb0702426d19242db5e8c079f04b.arrow`) is reused.\r\n- Now, run this script with `xla_spawn.py`. Ideally, it should reuse the cached file, however, you will see each process is creating a cache file again.\r\n\r\n```bash\r\npython xla_spawn.py --num_cores 8 \\\r\nprepare_cached_dataset.py \\\r\n--tokenizer_name=model_name \\\r\n--block_size=512 \\\r\n--data_file=$TRAIN_DATA\r\n```\r\n\r\n- Check the cached directory. There are two arrrow files.\r\n```bash\r\nls -lha \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-e84dd29acc4ad9ef\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d\r\ntotal 230M\r\ndrwxr-xr-x 2 ***** ***** 4.0K Aug 28 13:25 .\r\ndrwxr-xr-x 3 ***** ***** 4.0K Aug 28 13:08 ..\r\n-rw------- 1 ***** ***** 99M Aug 28 13:08 cache-bfc7cb0702426d19242db5e8c079f04b.arrow\r\n-rw------- 1 ***** ***** 99M Aug 28 13:25 cache-e0e2313e49c8a110aafcc8133154c19a.arrow\r\n-rw-r--r-- 1 ***** ***** 670 Aug 28 13:24 dataset_info.json\r\n-rw-r--r-- 1 ***** ***** 0 Aug 28 13:24 LICENSE\r\n-rw-r--r-- 1 ***** ***** 33M Aug 28 13:08 text-train.arrow\r\n```\r\n","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/532","title":"File exists error when used with TPU","comments":"I ended up specifying the `cache_file_name` argument when I call `map` function.\r\n\r\n```python\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size),\r\n batched=True,\r\n cache_file_name=cache_file_name)\r\n```\r\n\r\nNote:\r\n- `text` dataset in `nlp` does not strip `\"\\n\"`. If you want the same output as in [`LineByLineTextDataset`](https:\/\/github.com\/huggingface\/transformers\/blob\/afc4ece462ad83a090af620ff4da099a0272e171\/src\/transformers\/data\/datasets\/language_modeling.py#L88-L111), you would need to create your own dataset class where you replace `line` to `line.strip()` [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/text\/text.py#L35).\r\n","body":"Hi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n","comment_length":59,"text":"File exists error when used with TPU\nHi,\r\n\r\nI'm getting a \"File exists\" error when I use [text dataset](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).\r\n\r\nI modified [line 131 in the original `run_language_modeling.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/language-modeling\/run_language_modeling.py#L131) as follows:\r\n\r\n```python\r\n# line 131: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\ndataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\ndataset.set_format(type='torch', columns=['input_ids'])\r\nreturn dataset\r\n```\r\n\r\nWhen I run this with [`xla_spawn.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/master\/examples\/xla_spawn.py), I get the following error (it produces one message per core in TPU, which I believe is fine).\r\n\r\nIt seems the current version doesn't take into account distributed training processes as in [this example](https:\/\/github.com\/huggingface\/transformers\/blob\/a573777901e662ec2e565be312ffaeedef6effec\/src\/transformers\/data\/datasets\/language_modeling.py#L35-L38)?\r\n\r\n```\r\n08\/25\/2020 13:59:41 - WARNING - nlp.builder - Using custom data configuration default\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\n08\/25\/2020 13:59:43 - INFO - nlp.builder - Generating dataset text (\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d)\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:6: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:4: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:1: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:7: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:3: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nDownloading and preparing dataset text\/default-b0932b2bdbb63283 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to \/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/\r\n447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d...\r\nException in device=TPU:2: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nException in device=TPU:0: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nTraceback (most recent call last):\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\nTraceback (most recent call last):\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/torch_xla\/distributed\/xla_multiprocessing.py\", line 231, in _start_fn\r\n fn(gindex, *args)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 300, in _mp_fn\r\n main()\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 240, in main\r\n train_dataset = get_dataset(data_args, tokenizer=tokenizer) if training_args.do_train else None\r\n File \"\/home\/*****\/huggingface_roberta\/run_language_modeling.py\", line 134, in get_dataset\r\n dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 546, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 450, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/contextlib.py\", line 81, in __enter__\r\n return next(self.gen)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/site-packages\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/anaconda3\/envs\/torch-xla-1.6\/lib\/python3.6\/os.py\", line 220, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/*****\/.cache\/huggingface\/datasets\/text\/default-b0932b2bdbb63283\/0.0.0\/447f2bcfa2a721a37bc8fdf23800eade1523cf07f7eada6fe661fe4d070d380d.incomplete'\r\n```\r\n\r\n\nI ended up specifying the `cache_file_name` argument when I call `map` function.\r\n\r\n```python\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True, truncation=True, max_length=args.block_size),\r\n batched=True,\r\n cache_file_name=cache_file_name)\r\n```\r\n\r\nNote:\r\n- `text` dataset in `nlp` does not strip `\"\\n\"`. If you want the same output as in [`LineByLineTextDataset`](https:\/\/github.com\/huggingface\/transformers\/blob\/afc4ece462ad83a090af620ff4da099a0272e171\/src\/transformers\/data\/datasets\/language_modeling.py#L88-L111), you would need to create your own dataset class where you replace `line` to `line.strip()` [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/text\/text.py#L35).\r\n","embeddings":[-0.0798375383,-0.1442335248,0.1469174325,0.1047480032,0.2392238975,-0.183026433,0.5247991085,0.2587524354,-0.2769027948,0.1412049085,0.1099315658,-0.2591763437,-0.0396562777,-0.2304338664,0.2745446861,-0.1979209632,0.1962260306,0.3209769726,-0.1909585297,-0.0990928262,-0.0947270691,0.2928031087,-0.0418691412,0.1614839286,-0.6791219711,0.1413165182,-0.1830788553,0.2864961624,0.1081077009,-0.3061461747,0.1703537852,0.0626990348,0.2858562171,0.5413048863,-0.0001269419,0.3936147988,0.1948049366,-0.2527728081,-0.0595969483,-0.1641455293,0.3079850674,-0.0324931145,0.0716107562,-0.1129939109,-0.1293354481,0.1014449,0.3417716622,0.0041012326,0.4713276625,0.2538045943,-0.0111972932,0.7012689114,0.0628258288,-0.1866247505,-0.0218900703,0.2796743512,-0.1007430106,0.1228495017,0.3792977631,-0.2468184531,0.0520924367,0.0928169936,0.220889613,-0.209236607,-0.0801462457,0.2455779612,0.4375521541,-0.0800734386,0.0164037794,0.2247600257,0.2114127874,-0.2419933081,-0.4570935369,-0.4064103663,-0.0087269936,-0.3136442304,0.5325062275,0.1598000973,-0.2045152038,0.27584216,-0.0676135495,-0.0035868529,-0.0381930172,-0.0560378879,0.1150068715,0.2927227318,0.1143157631,-0.0198008679,-0.0402777568,0.0766948983,-0.2098604888,-0.089777112,0.1380930543,0.3785318732,-0.2391121984,-0.253087908,-0.1679145396,-0.5764261484,-0.1399088353,0.0170649737,-0.0117709413,-0.3582884967,-0.27250126,0.0793670937,0.5551717281,0.175221771,-0.1873039603,0.0450787358,0.0471253842,0.1027877629,-0.131499216,-0.0840475485,-0.3446531892,-0.2317399234,-0.2921254933,-0.0994022787,0.0419315882,0.0865661278,-0.0301301107,-0.1701417565,-0.3134407699,0.0887426212,0.1542658508,0.0939973891,0.1844564378,0.3348243237,0.4495030046,0.0126924552,-0.5675520301,-0.326512903,-0.1946421266,-0.1032909453,-0.3978535235,-0.0439434163,0.2442614734,-0.1168974414,0.2883491516,-0.260386914,0.2765443623,-0.0780453756,-0.1356095821,-0.0873277634,0.0491121672,0.3619853258,-0.0605050921,-0.103941977,0.0495957844,-0.3493099809,-0.1817699224,0.3404088914,-0.1699376851,-0.3592523634,0.0314719565,0.0184519328,-0.1031052247,0.0612264127,0.1891774684,0.094912298,0.5039544702,-0.5416415334,-0.0628860444,0.0121024968,-0.3074621856,0.1219959557,0.1785236001,0.5345861316,-0.008132427,-0.16372329,-0.1448525041,-0.3543258905,0.0238805059,0.4770064354,-0.1161619872,0.4217163324,-0.2756722569,0.1975223124,0.2805541754,-0.3485553563,-0.0487581454,0.2919351757,-0.2538606524,-0.1722111553,0.1669990718,-0.279571861,-0.2911915183,-0.3079813421,0.0637209266,0.051063668,-0.000889027,0.0401973054,-0.2602468133,-0.1193800643,-0.4364037216,0.1627896279,0.0766466632,0.1036475599,-0.092042774,0.6391347051,0.2819698453,0.0595624037,0.2897178531,-0.0073050591,0.1470348239,0.1178078055,-0.0899871364,0.1348540038,0.0085599395,0.1045292765,-0.031775713,0.3992015123,-0.0099070687,-0.0652452186,0.2647108138,0.1742208302,0.0090645598,-0.3494711816,-0.0604715943,0.0241230484,-0.1790014207,0.1490033269,-0.3081002831,0.0328932963,-0.1788248569,0.0819587409,-0.2277283072,0.2888807952,-0.1237493753,-0.2570460141,-0.1084304973,0.3896348476,-0.4102819562,0.0260165893,-0.0665694252,0.3695499599,-0.1494660527,0.2167463154,0.0701968446,0.2750688195,0.4112743139,0.1022071391,-0.1609389633,0.2704153061,0.0733857602,-0.0277392995,-0.1982143223,0.2000836134,0.0843485445,-0.0878728479,0.4576553702,-0.1464757472,0.1857974678,-0.0860165507,-0.0115564875,-0.1726840585,0.158526808,0.3137045801,0.0797688439,0.2871515453,-0.2362147272,-0.3226622641,0.48514238,-0.1081836447,0.0846824422,0.1235769615,0.0567587055,0.0045667798,-0.0996449441,0.1038167328,0.7618492246,0.0303733144,0.0746840164,0.2042744458,0.002426604,-0.4016212523,0.3145742118,0.0199703872,0.1478983015,0.3603564501,0.0456466228,0.1358236223,-0.2201711088,-0.2086426169,0.0125130415,0.1076703444,-0.3380920291,-0.0305709019,-0.1294817477,-0.2260117084,-0.613512218,0.3738873303,-0.2147596031,0.0300704781,0.1859934032,-0.1521373242,-0.3798165321,0.2276350409,0.2427494526,0.4120456576,-0.0400349014,-0.1004720181,0.2312813997,-0.2532883286,-0.413808614,-0.0716314539,0.1834110618,-0.1839823872,-0.1943470687,-0.1855919212,0.051695589,0.1973858327,-0.283944726,0.2684098184,0.1496329755,0.0128650814,0.0611016154,-0.0220944546,-0.0020250932,-0.338844955,0.3798138499,-0.3485740423,-0.3173532784,-0.0116605479,0.236385107,-0.1102096513,-0.3155271411,-0.3423826993,-0.0144651989,-0.2746013105,0.1298458278,-0.0374870226,0.2161976397,0.4914858937,0.3568271995,0.1663624346,-0.1525254846,-0.1984069198,0.153575018,-0.2659078538,0.0295194909,-0.2206698954,-0.2314794362,0.0058998899,-0.2214410901,0.3120176196,0.0577066243,-0.2935745716,-0.2030821145,0.1279263198,-0.1166286319,0.1098140776,0.329225421,0.5302072763,0.0346470438,-0.0924844742,-0.0594452322,-0.0121635487,0.4074875116,-0.0231232122,0.3068994284,0.0181849729,0.2183344066,0.1459243,0.702552259,0.2315587401,-0.0831463337,0.119037807,-0.034518227,0.3092693985,-0.253087014,0.0555535518,0.0235939417,0.2866981924,-0.0594139695,0.5606921911,0.187789008,-0.0019274236,-0.2192113101,0.0443913825,-0.1045405194,-0.3789369464,0.3051457703,0.0155064594,0.5026896596,-0.106448248,-0.2906706333,0.109794192,0.0504850112,-0.164696753,0.4930084646,0.1573979259,0.0962497443,0.0156962387,0.0435917005,-0.1954207122,-0.0942171365,0.1772780865,0.530616641,-0.2424412072,-0.2974810898,0.1551135629,0.136534363,0.1886734515,-0.0277821198,-0.1924500614,-0.0162192471,-0.3138192892,-0.1919985712,0.0206927881,-0.019027723,-0.2417087704,0.4024551213,0.5562255979,-0.1023862064,-0.4468138516,0.0831837207,-0.2228112072,-0.1595280319,-0.2530989051,-0.4939982593,-0.2606785297,-0.426384002,-0.0009276246,0.1753772646,0.0783521831,-0.0456291474,0.4315594137,-0.115151912,-0.021490097,-0.0176863167,-0.0781199262,0.0555077419,-0.228934899,0.1036133841,0.2859187722,0.0224685576,-0.1199823543,0.2650687397,-0.0706764683,-0.1486153007,0.0402937233,0.092425026,0.3039663136,0.5088429451,-0.0222058352,-0.0461166464,0.3289165795,-0.033121936,-0.3946834505,-0.0668153688,0.1291236281,-0.2472635508,-0.0779907405,-0.2480993122,0.2846378684,-0.0126131866,0.0781783685,-0.0091101499,-0.1303603351,-0.3914082646,0.3210028708,-0.0969046801,1.0393942595,-0.0084022731,0.612011373,0.3818511963,0.1126778424,0.3598698676,-0.4737111032,0.3548025787,-0.2686953843,-0.2874269187,-0.0151406601,-0.0034192204,0.2282916754,0.5558609962,-0.1965659112,0.0478182621,0.5781379342,0.109081611,0.267947346,0.121124059,-0.1075649038,-0.3451633453,-0.3952378929,-0.1394962668,0.2283513844,0.4478593171,-0.1105583385,-0.2278492302,-0.074955076,-0.0793944672,-0.0609492511,0.0613094047,-0.22407116,0.0624809749,0.5027500987,0.0188727342,0.2975089252,0.3734672666,0.1682242006,0.1691643596,-0.0724182054,0.4037234783,-0.1006249189,-0.024204826,-0.1672062576,0.0561545789,0.3336468339,-0.1572693884,-0.1032642499,-0.267070353,-0.2405528873,-0.2115935236,-0.0652293339,-0.1552081108,-0.1209754944,-0.0391939394,-0.097068347,0.3202849627,-0.3403377235,0.0686925203,0.0345718674,0.0811051652,-0.368298918,-0.3365427256,-0.1137615591,-0.1586219817,0.0343489274,0.4634477496,0.0143184615,-0.0709296688,0.5357791781,0.1047844067,0.094094336,-0.1901984364,0.063811779,0.1296214461,-0.8252724409,0.2144873738,0.1549895406,-0.2786587477,0.1183840185,0.3002387583,0.2092146128,-0.2025872171,0.0438221991,-0.2727172077,-0.4122874439,0.1511340141,-0.2708778679,0.0814899951,0.2648313642,0.68531847,0.3507699072,0.0534955822,-0.1497532129,-0.0178970154,-0.2738274932,0.1723219901,0.1717136502,-0.1761118323,0.2574409842,-0.0473734997,-0.1264455169,0.2257917672,-0.2392767221,0.0320567824,-0.0798017085,0.2339298874,0.3233177066,-0.0893468857,0.2869093418,0.1396989673,-0.1093263477,-0.1490410715,0.1398923844,-0.1106196865,-0.1481669396,0.2082501203,0.2583065033,0.2423363477,-0.3391033411,-0.1234730929,-0.061359819,0.0870398358,0.304703027,-0.1691262424,-0.1774098724,-0.050303001,-0.0367336795,0.1991208494,-0.0548996106,0.1663125753,0.1998898238,-0.3054782748,-0.0036393276,-0.0824467167,0.2388523966,-0.1146253645,-0.228988409,-0.1010643765,0.4654880166,0.0056850226,-0.2708635628,0.3376211226,0.0864608213,-0.1025362313,-0.1397997737,0.2574537694,0.1826211512,0.009457401,-0.0797461048,0.4228531718,0.0686199516,-0.5925223827,0.2610255182,0.5560619831,0.170055598,0.4229312837,-0.0492201187,-0.1060884893,0.1530782878,0.3082480729,-0.1963417828,0.1678284705,-0.0404758863,0.158505559,0.1119916141,-0.5258697867,-0.2627845705,0.2734742165,0.041289635,-0.0313573368,-0.2307530344,0.4621376395,-0.4583755136,0.2152919769,-0.2082636952,0.2460870743,-0.18534787,-0.2498941123,-0.066849865,-0.0791250095,-0.2840770185,0.034546487,0.0929383934,0.1996177286,-0.1173853874,0.1672763675,0.1528889835,-0.391109556,-0.2510310411,-0.0109409913,0.2041033059,-0.1220552698,0.4520687759,0.1890475005,-0.0954333693,0.2551528513,0.4810509086,0.2450400293,0.2664690316,0.1609449238,-0.0054430221,-0.2634566724,0.1757067442,0.032055743,0.3873583972,0.497679621,0.3165107667,0.0717504397,-0.1194694191,-0.0642404407,-0.0427799262,0.1871768981,-0.1899292916,-0.3418011367,0.8261855841,-0.109123975,-0.1319714636,-0.0133555094,0.1182518452,-0.4318698645,0.3141445518,0.3554555774,-0.0011689821,0.1497472972,-0.1496789455,-0.0445798412,0.2240770608,0.4062027931,0.284313798,-0.3281782568,-0.2428979725,-0.2061026692,-0.5585113764,0.0270148125,0.0461116582,0.2493058145,-0.0349739753,0.0079351477,-0.0681045949,0.0497218072,-0.3419698477,-0.3155051172,0.0513183773,-0.1820664704,-0.1156256199,-0.1244633198,-0.2046041489,0.046072159,0.0653910339,-0.270229727,-0.2537830472,0.0498041213,-0.0264444202,0.0422568247,0.1493170857,0.1925272495,0.1814150959,-0.0711630806,0.0884157866,0.0676685646,0.2310806811,0.0728290603,-0.39859429,-0.1142987087,-0.3612944186,0.0032336561,0.0695548952,0.495339632,-0.4000498056,0.0474147089,-0.2902721167,-0.2899895906,-0.0752145052,-0.2781928182,-0.2372091562,-0.0139884325,-0.0827318281,0.122254476,0.4128880501,0.4169913232,-0.1042444855,-0.1076245978,-0.5772278309,-0.360363245,0.3100019097,-0.559237659,-0.0874062255,0.1202496514,-0.0655824319,0.0252231155,0.1328889132,-0.642411828,0.0706516281,0.1075027138,0.1937766373,-0.3184123635,0.0537423715,0.0611682348,-0.2190911621,0.1069517285,0.3757636845,0.0110927885,-0.4208606184,0.1000858918,-0.1757013351]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/525","title":"wmt download speed example","comments":"Thanks for creating the issue :)\r\nThe download link for wmt-en-de raw looks like a mirror. We should use that instead of the current url.\r\nIs this mirror official ?\r\n\r\nAlso it looks like for `ro-en` it tried to download other languages. If we manage to only download the one that is asked it'd be cool\r\n\r\nAlso cc @patrickvonplaten ","body":"Continuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n","comment_length":59,"text":"wmt download speed example\nContinuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n\nThanks for creating the issue :)\r\nThe download link for wmt-en-de raw looks like a mirror. We should use that instead of the current url.\r\nIs this mirror official ?\r\n\r\nAlso it looks like for `ro-en` it tried to download other languages. If we manage to only download the one that is asked it'd be cool\r\n\r\nAlso cc @patrickvonplaten ","embeddings":[-0.0366919525,-0.2923784852,0.0385637581,0.2077883631,0.0263427533,-0.1215780005,0.1395124793,0.1727898419,0.1227243766,-0.0240065344,0.1071392298,0.4486669302,-0.074683398,0.0759545639,0.2661069036,-0.0970808119,-0.0178977884,-0.0835175291,-0.5572077632,-0.265684247,0.0741957724,0.3135083914,-0.1048664227,0.0379503109,0.0109728323,0.1809777915,-0.0905235335,0.012872465,-0.4572179019,-0.2325666249,0.0901376083,0.1737001687,-0.0498700477,0.1833204329,-0.0001086784,-0.1280477494,0.4216489494,-0.0375629552,-0.4472953975,-0.3175246716,-0.5728065968,-0.5532028079,-0.1477105469,-0.0608193167,0.1857579648,-0.0601900183,0.0039492603,0.0888582096,0.1687783897,0.0085925143,0.2087498754,0.2460334301,-0.382253021,-0.1306511909,0.5218646526,0.0253502559,-0.2318369895,0.486392498,0.3931060135,0.15960747,0.092266418,0.4000581801,-0.1544285864,0.2864332199,0.1128638312,0.0945444852,0.3612880707,-0.1099048033,0.0165801048,0.5688591599,0.3632158935,0.1589251459,-0.3924889565,-0.2711546719,-0.0051723788,-0.2193635553,0.105397284,0.3335303664,-0.3128455877,0.0895097032,-0.6589964032,-0.0388941467,0.0859431848,0.2939275503,0.0006860329,0.6809962392,-0.0953051895,0.0456331149,0.089819476,-0.0484639853,0.0121703213,-0.3499272764,0.0040101497,0.0348172933,0.0384136029,-0.0830790699,0.1507284045,-0.0253082607,0.0151570877,0.0298281834,-0.1212169379,0.2414697558,0.2694712579,-0.0043602251,0.2079116106,0.3571501076,-0.1800067127,-0.17567496,0.1628447026,0.0752706677,0.2268755734,0.2538921237,-0.2276644111,-0.067733258,-0.4933923781,-0.0276074652,-0.0981749594,-0.0790215135,-0.2791447937,-0.2397160381,0.0282093734,-0.1965026259,-0.1278362423,0.1890400499,-0.253284514,0.2782042325,-0.1491139829,0.045209568,-0.228411451,-0.1067471728,-0.0300925989,-0.0015241646,-0.2437842041,0.005183408,0.3144951463,0.2160779536,0.3192485571,0.1389926374,-0.2071329802,-0.0875620544,0.0687246248,0.1165935397,-0.0168886948,-0.0155766355,-0.1657674611,0.4662278593,-0.1186365783,0.1973176897,-0.351909399,0.1186806411,-0.3961490989,-0.1821977049,0.0933934003,0.1816111207,-0.3348970711,-0.2620611489,-0.0394127741,0.3635922074,-0.3202870786,0.2390497923,-0.0523409694,-0.0484774262,-0.2543555796,-0.0335065424,0.0671493411,0.408264786,-0.1602657437,-0.1924747676,-0.2034654021,0.1912580281,0.4898460507,0.4522068501,-0.2085849196,0.0941871926,-0.2111016959,0.037608929,0.1644468904,-0.1547364444,-0.2068759054,0.6244184375,-0.4453573525,0.1312571466,0.1468585879,0.0212514997,0.1801609844,-0.1096653715,0.1793670654,0.51820153,0.0115399733,0.1022072658,-0.4862095416,-0.405929327,0.2895371318,0.01182546,-0.1852208674,-0.1393261999,-0.0287491605,0.1552169472,0.3241948783,-0.0678680167,0.0529374294,0.0178252198,-0.1040143296,-0.361774683,-0.0266505722,0.2734881043,-0.2059174627,0.3234938681,-0.0726450458,0.309057951,0.230760172,0.2014811337,-0.2200478762,-0.1866343021,0.0362695307,-0.0985003188,0.1168618798,0.0804219842,-0.0565657429,0.0912403613,0.1351668835,0.0237788055,0.013343866,0.0601711608,0.3700859547,-0.1242327094,0.0169894006,-0.0740822777,0.2616152763,-0.0662718266,0.120517835,-0.0700983778,-0.1816136539,0.1176898256,-0.083994396,0.7728492618,0.2703164518,0.0602892451,0.2631710768,-0.3408568799,0.4670611024,0.4881790578,0.1534876525,-0.0567406788,-0.2976082861,0.631728828,-0.0665884688,0.1452822089,0.2575868666,0.0677094236,0.0057374337,-0.0056524235,-0.2457089424,-0.0185158066,0.476483047,0.1203885823,-0.0962068141,0.1273318082,0.1313000172,0.2547520697,0.635476768,-0.1735519916,0.0543273315,0.203287065,-0.3028753698,-0.4658342004,0.2684465349,0.3569935262,0.2943990231,0.1340411752,0.5332397819,0.1486283988,-0.0177149959,-0.2474839985,0.4499844611,0.0999943614,-0.343003273,0.0635734499,-0.0105902851,-0.0121998955,-0.2686423063,0.2848038673,0.0428778343,0.0149842594,-0.1176051795,0.072036922,-0.493404597,-0.2570087314,-0.3033087254,-0.2698554993,-0.2456465214,-0.3047164679,0.0502166674,-0.1831916124,-0.1301475763,0.0301142652,0.3293049335,0.0620103702,-0.3602292836,0.1799545437,-0.1737008095,-0.102861166,-0.4780779779,0.0354307815,0.2430090904,0.1615803987,0.3787631691,-0.3717290461,-0.1113354042,-0.051209487,-0.3588240445,0.1746948212,-0.0494212955,0.2054582387,0.0113350134,-0.0139566725,-0.244441852,0.071715571,-0.0249675866,-0.3179233372,0.1241822168,0.0388134904,-0.2473672181,-0.0790558681,-0.2507551312,-0.500261426,-0.2906683683,-0.0813751072,0.2634072602,-0.1414362937,0.1935472935,0.2922198772,-0.151379928,0.3148866594,-0.0831439644,0.1852230877,-0.1199272051,-0.4302172661,0.320438832,-0.0602751672,-0.266969651,0.3288850486,0.1126204059,0.0716884211,0.0390772596,-0.4180586338,0.3023874462,-0.0928385407,-0.0431097597,0.4418205023,0.1617445648,0.1722646803,-0.0778144449,-0.029338304,-0.0650472268,-0.3291345239,-0.0873753503,0.3501098156,0.2656290233,-0.2338151038,-0.0085863555,0.3358736932,0.5412426591,0.1123502031,0.3109045029,0.0478649512,0.1983861476,-0.1089453921,0.2531492114,0.0189189799,0.3179726899,-0.2049615234,-0.0589671284,0.3475881219,0.2113371491,-0.3529017866,-0.1316321939,0.1703730822,0.0732095763,-0.2385872155,-0.0146643771,-0.0693894625,-0.1037624925,0.1401767135,0.0954217687,-0.1875465214,-0.2585598528,0.0589300618,0.3068724275,-0.0914940983,0.1234385148,0.0384723172,0.0400008336,-0.4884056151,-0.1577084064,0.0374623425,0.3167340159,0.0552034937,0.0052320845,0.1884696186,-0.1565472782,0.0461756252,-0.2120104432,0.1149504632,-0.0534555279,-0.2566125095,-0.4716103077,-0.3695564866,-0.3273006976,-0.4731156826,0.4219290614,0.0111074839,-0.1668891609,-0.1243621707,0.1625125706,0.1825753897,0.0208011232,0.0075457776,0.0194824431,-0.5261372328,-0.0277475603,-0.1128445119,-0.2303444147,-0.1688616127,-0.0842981115,0.0826326162,0.3962228596,0.3614590168,-0.0180386212,-0.0073240451,0.0085880673,0.043760743,0.1732918471,0.1010830849,0.3227288723,-0.2572561502,0.4116533101,0.2514832616,0.1263488382,-0.085150674,-0.0144902924,0.096261315,0.6376786232,0.1410153508,0.114758946,-0.1640266031,0.1280244291,-0.217405811,0.1239241883,0.4413773715,-0.0589338206,-0.0277079325,-0.2384822667,0.3543862104,-0.0186971743,-0.1848835051,0.0306657385,-0.4643349051,-0.1061646491,-0.1553113312,0.1650105715,1.2208110094,-0.0829151943,0.2892374694,0.2155918479,-0.4171025753,0.5484615564,-0.111516647,-0.1139390394,0.1213186607,0.4160411954,0.0257534441,0.081648469,0.1442889422,0.0930332243,-0.4381765127,0.2913718522,0.2732536197,0.2148755789,0.2208146602,0.3266221583,-0.1780341864,-0.2599669099,0.0081866039,0.1460518986,-0.0396314561,0.4048943222,-0.0165237281,-0.1932680756,-0.1302768141,-0.147046864,-0.5490785837,0.0887485892,-0.1451573521,-0.1727003306,-0.3821457624,-0.3444069922,0.1687267423,-0.2724861205,-0.1259727776,0.1509996951,-0.2945027053,0.204981938,-0.3947577477,0.2506480217,0.3349378407,-0.165231511,-0.1220938787,-0.3081336319,-0.1625579298,-0.2768587768,0.0607701838,0.2732270062,-0.2541185319,-0.2172360867,-0.1065584794,0.5261734724,0.0016012331,0.3198694885,-0.0687206537,-0.2394001186,0.1680699885,0.1484867483,-0.2169896811,0.340495795,0.0559044778,-0.2652062178,-0.3284946382,0.4559430182,-0.1275828928,-0.095876053,0.2058307528,0.2329422981,-0.1831436455,-0.2538066506,-0.0710104331,-0.3330278397,-0.2490781397,0.0377656259,-0.2313555628,-0.0287278574,-0.0312701128,0.223993957,0.423027277,0.0452817604,-0.0434645191,-0.1492175758,-0.0193058793,0.1178322062,0.0930558443,0.0369653627,-0.523183167,-0.2003670186,0.1825504899,0.1160079539,-0.3140278459,0.1554561406,-0.0174042564,-0.0162281096,0.1054595411,-0.1625567824,0.1020605564,0.1132232547,0.057504043,-0.2987834811,-0.1324883401,-0.1951033622,-0.0476425178,0.1375062466,0.0506283008,-0.6086687446,0.417444706,-0.1578948945,-0.0103832847,0.0280093495,0.061262887,0.0869585276,-0.1946710348,0.2316184044,0.0213860311,0.1267999113,-0.1131836027,0.1549940854,-0.0874431059,0.4376253784,0.1472379863,0.0859618634,0.4909369051,-0.076023981,0.0008736272,-0.0867419243,-0.0854293928,0.3044659793,0.6014089584,-0.0512759313,0.3350890875,-0.0219937824,0.1548386067,0.3572499156,-0.012053201,-0.1645054966,0.3825442195,0.1769101322,-0.004785432,-0.2509565353,0.0224978775,0.1511775255,-0.1174377799,0.0811753422,0.2170943916,0.1162599772,-0.1577135623,-0.1268215925,0.1819760799,0.0390516855,0.0071632452,0.3407666385,0.0460038297,0.2685064971,-0.2568445504,-0.0754603818,0.0026271916,-0.1139533967,-0.0525700077,0.2767501175,0.1438092142,0.6271647811,-0.039635282,-0.189007014,-0.0726458654,0.4148306251,-0.1195017695,0.2051081806,0.1833916157,-0.1066178456,-0.0427800193,-0.1508678198,-0.3581852317,0.3616628349,0.0624043234,-0.3463210464,-0.1909940392,-0.1027354598,-0.3462980092,0.3681491613,-0.3697713017,0.0707769319,-0.0027193339,0.3390327394,0.0976682454,-0.3586314023,-0.3370804787,-0.4636339545,0.168369174,-0.2557735145,0.1014713794,-0.1461806744,0.0710592046,-0.1298072189,0.1876079738,0.0947513878,0.2715924084,-0.196673736,-0.12965177,0.102822192,0.0874680802,-0.1183481514,0.4209769964,0.1580832601,0.1502567977,0.1274961233,0.0148599362,-0.2582024038,0.1020779833,-0.2203730196,-0.8334546685,-0.2312723547,0.3081340194,-0.0655826554,0.0679112002,-0.3168427348,0.038859006,-0.2322292924,-0.071452029,0.4818241298,0.0246453751,0.1847480536,-0.0579721518,0.0740562305,-0.1456452161,0.5916481614,0.2572513819,0.2669019401,-0.3744655848,-0.2374792993,-0.3805504739,0.0291999597,-0.2586583197,0.0700792447,0.0953942761,0.068146266,-0.0346758999,0.421176821,-0.0340851359,0.6065480113,-0.1240187734,0.0428498164,-0.1137506068,-0.1234534085,0.1284444779,0.0653606653,-0.1536687464,-0.2195087224,0.2254100144,-0.0676793307,-0.0850570723,-0.1279168576,0.2348797172,0.1530579031,0.269656539,0.1503949761,0.0401460864,0.3597563207,0.0910031646,-0.3492451906,-0.4259172976,-0.0400484726,-0.0875401422,0.0470683537,-0.4361440539,-0.0421792455,0.1169766039,0.2007253915,-0.2572182417,0.0876617506,-0.08609505,0.1506073326,-0.1392784268,0.0463102125,-0.229709968,0.1670975089,0.0448346846,0.2075980306,-0.13091667,-0.0254996829,-0.3452730179,-0.0338524096,0.2426211685,-0.2107735276,-0.1762984246,-0.3393878341,0.2527181506,-0.2134812176,0.1959014535,-0.0284391344,-0.0017706392,0.1980074495,-0.0186707173,-0.2138498425,0.3829642236,0.2446406186,-0.2085780948,0.052731365,0.3852646053,0.0374385454,-0.2234347314,-0.1060093641,-0.2093635201]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/525","title":"wmt download speed example","comments":"Mirror is not official.","body":"Continuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n","comment_length":4,"text":"wmt download speed example\nContinuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n\nMirror is not official.","embeddings":[-0.2030677646,-0.2355915904,0.0204493832,0.0908323899,-0.0058562788,-0.2086308897,0.2170621008,0.1441665143,0.1470083743,-0.0152362473,0.014190061,0.4619309902,-0.1260588169,0.0828568116,0.1583827585,0.0030627702,0.008713237,0.024945017,-0.6374734044,-0.2368490845,0.1026754677,0.2861510515,-0.0466097668,-0.1556268483,-0.0598167665,0.2199995071,-0.1058428437,0.0222445019,-0.4653730989,-0.2268990427,0.1078721508,0.1963041276,-0.1250354052,0.2288564891,-0.0001047362,-0.0788546205,0.3874823153,0.04121124,-0.5057906508,-0.2190839797,-0.4556691349,-0.5933763385,-0.0999460369,0.004230509,0.0806585625,0.0239088628,0.0372118428,0.1130727828,0.205658868,-0.0920939893,0.2260437608,0.4178249836,-0.3471935987,-0.1840334982,0.475616008,-0.0356496833,-0.2780715525,0.4869729578,0.3253695667,0.1012507081,0.0083049824,0.402605772,-0.2482906133,0.3595455885,0.1433266699,0.0522795431,0.2765185833,0.0225311108,-0.0642793104,0.5838644505,0.3547344506,0.1416072696,-0.4659505188,-0.3094413579,0.0167103689,-0.132569164,0.1006563678,0.2839250267,-0.3353668749,0.0723492503,-0.6540669799,0.0040230467,0.0759235397,0.2350191623,-0.1124524549,0.6518868804,-0.1092210561,0.0321764573,0.0594284125,-0.0640894175,-0.0341035239,-0.2815542817,0.0075940336,-0.0036619953,-0.0351191387,-0.1704579592,0.1861032695,-0.1868005991,-0.0518975891,0.126724422,-0.1136157811,0.2357570529,0.2769711018,-0.0024387406,0.1659559309,0.3462160826,-0.243555218,-0.158763364,0.1297718734,-0.0493339784,0.1967822015,0.1981655359,-0.1342841089,-0.0498470552,-0.4596246481,-0.128510803,-0.0441143364,-0.0432467721,-0.2834776938,-0.2843770087,0.1155992895,-0.1418762952,-0.1883973628,0.2164369226,-0.1706951261,0.2608783245,-0.1016232371,0.0099950377,-0.2099787295,-0.0593416542,-0.1121548787,-0.0728264302,-0.172431916,0.0067751128,0.3686103225,0.1807249486,0.3187747598,0.0368449874,-0.1568606943,0.0875714123,0.1097929627,0.1756331474,-0.0442888625,0.0572112873,-0.1499746591,0.3965829611,-0.1831968576,0.2844672799,-0.3124317229,-0.011195167,-0.3218106031,-0.166082263,-0.0125757409,0.2197827697,-0.3754304349,-0.2757881582,-0.0477951802,0.2472575009,-0.3778367341,0.2222268432,-0.1033454388,-0.0365001298,-0.2740734816,0.0088664712,0.033799991,0.3266341984,-0.0545243658,-0.1670298427,-0.1565573215,0.2004750222,0.3898042142,0.4546078742,-0.1952291876,-0.0324841291,-0.158452034,0.1124154031,0.0894827992,-0.1717761606,-0.2374877483,0.6228610873,-0.4740535617,0.0374255441,0.1954982579,0.0994876102,0.2371068299,-0.1745349914,0.1396849304,0.5552384257,0.0367418863,0.1466036737,-0.4413136244,-0.4089230299,0.2830276489,0.0621594191,-0.1908890754,-0.1757435799,0.0149924243,0.2861718237,0.2952894568,-0.0368655249,-0.0050558359,-0.0181964189,-0.0472911522,-0.386303544,-0.0618518293,0.2278930843,-0.1751128733,0.3031557202,-0.00080793,0.3377199769,0.2991241515,0.173042044,-0.1038531587,-0.1400363743,0.0785016119,-0.0950670019,0.1608621776,0.1519647539,0.0267883893,-0.0081564607,0.1190481409,0.0688812062,0.0581147783,0.0821766928,0.3590968549,-0.1101092994,0.0602873415,-0.0981409699,0.3463968337,-0.0911270753,0.0179644674,-0.008664283,-0.0887940675,0.1124151126,-0.1725164205,0.7700487971,0.2590562105,0.0300503001,0.2304239422,-0.2941974699,0.5189212561,0.5327334404,0.1456305832,-0.0524913371,-0.3895391822,0.5775874853,-0.0035945368,0.1051050797,0.230160147,0.0483533517,-0.0014539611,0.0465171374,-0.2578205764,-0.0955373347,0.4237414896,0.062680088,-0.0326520912,0.1789205819,0.1140270233,0.2815776467,0.7040287256,-0.1659501344,0.0809741169,0.1443807781,-0.3343804479,-0.4499129653,0.2610890567,0.3194567859,0.339109391,0.2026924491,0.4832705557,0.1593817174,0.0029307599,-0.2737599909,0.4070984125,0.0565173738,-0.308146596,0.0962212309,0.1062801033,0.0079617705,-0.2985522747,0.2640021145,0.0471200757,-0.0346504487,-0.0714860409,0.0349122509,-0.435046196,-0.1857003719,-0.2186580002,-0.1814424098,-0.1903096884,-0.3750255704,0.0464654081,-0.0609224699,-0.1039060429,0.0167489368,0.2858650386,0.049713891,-0.2917511761,0.113459833,-0.1752455384,-0.0542134084,-0.3069823086,0.0982840806,0.1957153231,0.1089966595,0.4032247961,-0.2869220972,-0.1240653992,-0.0845244527,-0.329834938,0.1257891804,-0.0305351596,0.1863965988,0.0730807483,-0.005486846,-0.1466668099,0.0993516445,-0.0094522955,-0.3443151712,-0.0099093616,0.0588234439,-0.2532224953,-0.1032923982,-0.2634672523,-0.4303185344,-0.2605269253,-0.1366133243,0.3005034626,-0.098001942,0.1187162772,0.2674123347,-0.1572344303,0.247794956,-0.1187513769,0.1500257105,-0.1252908558,-0.4138742387,0.3190301657,-0.0239780638,-0.291213274,0.2572585642,0.1610604078,0.0016408355,0.037707366,-0.4150870442,0.2317053378,-0.0877555609,-0.0214301068,0.5156495571,0.1321832538,0.2328296304,-0.0975098088,-0.0899272114,-0.0736709759,-0.2760532796,-0.1630902439,0.3570804596,0.133291617,-0.3352710009,0.0222012959,0.3192939162,0.4842571318,0.069580026,0.2317884266,0.0492185205,0.1620238125,-0.0513266586,0.1169483066,0.0127544971,0.3907426298,-0.160404548,-0.0167384576,0.3219852149,0.1498742104,-0.4193620384,-0.1510782987,0.0797455311,0.0803811625,-0.172223717,-0.0388418436,-0.0167368017,-0.1407767385,0.2423205525,0.1315241009,-0.1471152455,-0.3114785254,0.01538759,0.1863207668,-0.0866228938,0.0855802596,0.0695009232,0.1179367006,-0.6063817739,-0.1542641222,-0.0685446262,0.2889143527,0.1370920539,-0.0978058726,0.1473590583,-0.2178400606,0.0666501969,-0.1105863899,0.1872760952,-0.0713115185,-0.3396683037,-0.5359287858,-0.3630856574,-0.3715191782,-0.485385865,0.3254081011,0.1912355572,-0.129969731,-0.1490300596,0.1764754206,0.0945740789,0.0584355853,-0.0000525341,-0.0111107985,-0.4407349527,-0.0774047598,-0.1474565864,-0.3060890734,-0.1363726854,-0.0297644939,0.040525496,0.2626615763,0.3169988692,-0.0464752503,0.0569102429,-0.0429550223,0.0745248646,0.2735492289,0.0500823073,0.3529339135,-0.3185132742,0.3776083589,0.3099298477,0.1814811826,0.0091792196,-0.0054329997,0.0079816412,0.6054229736,0.160768643,0.0625661761,-0.208040148,0.2155769914,-0.290186882,0.078422524,0.377163738,-0.0670928583,0.055972185,-0.1702265739,0.3322433233,-0.0637213811,-0.0912123993,-0.0710737407,-0.4786358476,-0.0016499622,-0.0984324887,0.1886273921,1.0390402079,-0.0755165815,0.3711608648,0.3059725761,-0.3486093581,0.5862109065,-0.1547619998,-0.085341461,0.0472474359,0.4046235681,0.0505167097,0.0924850255,0.0321560986,0.0790044516,-0.3292599916,0.2617874444,0.3181463182,0.2266500592,0.1160334796,0.4343652427,-0.2055250704,-0.2904225588,0.0136180585,0.2201698571,-0.0661943629,0.3490396142,0.0008137732,-0.2109444141,-0.04541035,-0.1260855943,-0.4736896455,0.1355124563,-0.0336435549,-0.0731765106,-0.3062026501,-0.3103531301,0.0721733198,-0.3765782416,-0.187753275,0.1742203832,-0.2950953245,0.1948688179,-0.407622844,0.2341154367,0.3771949708,-0.1980637312,-0.0825071409,-0.3397155404,-0.0878528208,-0.3407077491,0.0762117356,0.1503628045,-0.1418018043,-0.1267127395,-0.1475165188,0.4593782127,-0.0075173294,0.3161019087,-0.1039775461,-0.2321657687,0.1945485771,0.159803316,-0.2318720967,0.3469981551,0.1494224966,-0.3162741661,-0.353184253,0.5173122287,-0.1119871289,0.0045890189,0.2409374863,0.1820444763,-0.2036862075,-0.2895685732,-0.1300400943,-0.2313759923,-0.2454027534,0.0778879896,-0.1830804944,-0.0719831511,0.1315717548,0.2217419744,0.3250101507,0.0210559778,-0.0970687419,-0.1306852549,0.0424077921,0.0706218779,-0.0118636629,0.0847761184,-0.4070269763,-0.1470526308,0.0843899623,0.0668876544,-0.3677631617,0.1075294465,0.0043594129,0.0078446753,0.0812266767,-0.2476647049,0.1771439612,0.0814750195,0.1172733903,-0.2359526902,-0.1907861233,-0.2570080459,-0.0205823705,0.1092094257,0.0071423827,-0.5400050879,0.3640684187,-0.1244470626,-0.1203373969,-0.0195689183,0.1560299546,0.1142219603,-0.183584258,0.149993062,0.0365353823,0.0977839306,-0.0774508119,0.2121670097,-0.0221414082,0.3368267417,0.0046161721,0.0827447474,0.4533126056,-0.0265100189,0.0557888895,-0.0638632774,-0.1286212057,0.325085938,0.6440168619,-0.11487744,0.2489048243,-0.0756189227,0.2732082903,0.294485271,0.0193971805,-0.2231171578,0.2930339873,0.2291125804,-0.0576169789,-0.2947650552,0.0438688211,0.2509971857,-0.0686888173,0.0246287622,0.1997220218,0.0882238895,-0.1875220537,-0.0427077524,0.22703363,0.0460261814,-0.0419350713,0.2957999408,0.0564476028,0.1600303203,-0.1975458264,-0.040084932,0.0208602548,0.0245117117,0.0127702095,0.2903151214,0.1194299012,0.5836248398,-0.0883156136,-0.1250924617,-0.0209416617,0.3985710144,-0.1158476993,0.2192596197,0.1453162432,-0.1544575691,-0.0461346544,-0.1769136935,-0.362264961,0.4138174057,0.0343768969,-0.3142726123,-0.1528639197,-0.0100868689,-0.255779773,0.4200840294,-0.3458042145,0.0508207083,0.0428394862,0.3194565177,0.133680135,-0.2487619817,-0.273147136,-0.409527868,0.1432676464,-0.2912506759,0.0554055311,-0.1470518261,0.1090874895,-0.073454842,0.2886704803,0.1487576514,0.2091506422,-0.1682263613,-0.068734929,0.0507089198,-0.0098396661,-0.2144845724,0.3839369416,0.0661013573,0.0571958087,0.1309463382,0.0848970488,-0.2866131365,0.0795324817,-0.2396415919,-0.7498160601,-0.2254345566,0.1953789145,-0.089516148,0.0765395239,-0.3475219309,0.1744411588,-0.2625537813,-0.0650928766,0.4861157238,0.0434888154,0.1831423342,-0.0555383265,0.1059816852,-0.1421706975,0.5879185796,0.2074314505,0.1932649463,-0.2036723942,-0.1913641691,-0.3645216525,-0.0168635268,-0.2121161669,0.1181735173,0.0549207591,0.225411877,0.0104192756,0.4055377245,-0.109431237,0.6497071385,-0.1581714302,0.0473337732,-0.0550364777,-0.0664439499,0.0982952788,-0.0524399653,-0.1089148745,-0.2765333056,0.1618120968,-0.118818529,-0.0074267201,-0.0835497975,0.2241306752,0.1550568789,0.2814682126,0.1628462076,0.014370624,0.3227446377,0.0761917308,-0.3740495443,-0.4074919522,-0.0494565926,-0.143452391,-0.0040914593,-0.3519194126,0.0242210925,0.1271557212,0.2725497186,-0.1739250869,0.06750305,-0.161978066,0.1350411922,-0.1252853125,0.1008634344,-0.3387354314,0.200956732,0.0158406198,0.2104011923,-0.0665495768,-0.1487017721,-0.3379896879,-0.1380699426,0.2594902217,-0.2753849924,-0.0893143117,-0.3876660764,0.181074813,-0.1819102466,0.1680427641,-0.0815771669,0.0866722018,0.1364199519,0.0080791265,-0.1554384679,0.3992872238,0.2895398438,-0.1864658445,-0.0165691581,0.3685422838,-0.0161026772,-0.2213472128,-0.1653291583,-0.1419237554]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/525","title":"wmt download speed example","comments":"Shall we host the files ourselves or it is fine to use this mirror in your opinion ?","body":"Continuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n","comment_length":18,"text":"wmt download speed example\nContinuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n\nShall we host the files ourselves or it is fine to use this mirror in your opinion ?","embeddings":[-0.0970186219,-0.2302699536,0.0760004371,0.2327990234,-0.1930162907,-0.1516977102,0.4099264741,0.0660621449,0.2584859431,0.0476587974,-0.0551418997,0.2684978545,-0.1681045443,0.0384293236,0.2093773037,0.0515972152,0.0763443038,0.0341624543,-0.6077142358,-0.2628243268,0.1051906794,0.1649488509,-0.0395846143,-0.0544994175,-0.096891135,0.2049501091,-0.065912649,0.019605305,-0.3745268583,-0.1257733554,0.1418088824,0.2571884394,0.0026763303,0.3433766365,-0.0001119774,0.0217483398,0.3440234065,-0.0119890571,-0.3946137428,-0.1340293139,-0.5043621063,-0.5133453012,-0.0980625674,-0.0678515211,0.0444890261,-0.1336168796,-0.0031023214,0.0976151079,0.2608041465,-0.1199490204,0.1531336159,0.3764331639,-0.553996861,-0.0628617033,0.5288295746,0.0120592564,-0.1928016245,0.7287992835,0.4486648738,0.2702367902,0.0053199776,0.3857813776,-0.2627511621,0.2417167574,0.2182659209,0.143761754,0.3119071722,0.0835657045,-0.2040938735,0.4565157294,0.3446314335,0.1195577309,-0.5175538063,-0.3979912102,-0.0086078746,-0.0808486566,0.1689206958,0.3056764007,-0.1893408448,0.301427573,-0.6494131684,-0.2435513884,0.0007555283,0.2693094611,-0.0311490763,0.5902032256,-0.0753811598,0.1342080235,-0.1156806871,0.0863806382,0.1333897263,-0.3744549751,0.0500809476,-0.1554635614,0.1346354783,-0.2157566249,0.0398157686,-0.2851393819,-0.1335162371,0.2030915916,0.0032352936,0.0864800587,0.126071319,0.0952772573,0.1827128828,0.3722531199,-0.3179148734,-0.1064494848,0.1842977256,0.0189920552,0.2428282052,0.1555784047,-0.1604377776,-0.0409256518,-0.5879229307,-0.086891681,-0.1165796444,-0.1082954705,-0.1772287935,-0.1111123934,0.1398072243,-0.1824859977,-0.0964270607,0.1045272201,-0.1154746786,0.3642384112,-0.0578075163,0.0434145629,-0.1267211884,-0.1380361319,-0.0077105891,-0.0665884987,-0.093504481,0.0864177197,0.3612859547,0.1556963474,0.3075782657,0.0156511683,0.0400045998,0.1399882734,0.1401408166,0.1032531708,-0.0399017371,0.1006247252,-0.1518379748,0.3922111094,-0.2451858968,0.3385667503,-0.3105312288,0.0323683619,-0.339302808,-0.3392120898,0.0903845355,0.1400927901,-0.2683833838,-0.1725406349,-0.1806355566,0.1373317391,-0.391908586,0.0645885915,-0.0769555941,-0.0070266626,-0.2558456659,-0.0696706101,-0.032681942,0.3766858876,-0.1034617871,0.0147951078,-0.0211647227,0.2590466142,0.4362417161,0.5246804953,-0.2769944072,0.1391410083,-0.1622255743,0.0690056831,0.0558938012,-0.2430541068,-0.2646181583,0.6330524683,-0.5914660692,-0.0062272605,0.3196915388,0.037497934,0.2989468873,-0.2110330909,0.2449943423,0.5267232656,0.1126668304,0.0809040293,-0.4941202104,-0.4558441937,0.2412892133,0.1632135361,-0.2086906284,-0.1027116254,-0.0507310182,0.2361124754,0.2248370647,-0.1360492557,-0.0092619238,-0.1300798059,-0.0910939276,-0.3411992788,-0.1972526312,0.2972263098,-0.184872061,0.2816551626,0.0172871873,0.250207305,0.3418705165,0.0803093389,-0.0942329466,-0.2476306707,0.0873019844,-0.0315813683,0.0378059782,0.1351808608,-0.0566881597,-0.0686647668,0.0052511431,0.1243499741,-0.0264909919,0.1590866745,0.3992736638,-0.0466005318,0.0200522244,0.0172211025,0.2545689344,-0.0975966305,0.037316788,-0.1404786259,-0.0191654563,0.1281843334,-0.2272661924,0.8664548993,0.3883675933,-0.0091440883,0.4069119096,-0.3548854291,0.5632680655,0.4918876886,0.0938520357,-0.0801300406,-0.3097838461,0.530461669,-0.0195397921,0.1419156045,0.2477449626,-0.0110155512,-0.1525164396,0.0978008136,-0.2506643236,-0.0653886348,0.4053527117,0.1671307236,0.0045260834,0.2160785645,0.077575177,0.1888605058,0.481002897,-0.1825459152,-0.0097448779,0.0617801361,-0.1871864051,-0.3809093535,0.3489971459,0.3133142889,0.4159398377,0.2065536082,0.4286356866,0.1747942865,0.0932023525,-0.2560546994,0.4070366025,0.0821523443,-0.3458167911,0.1380351633,0.0604515634,0.0187940393,-0.2305562645,0.178094849,0.0391163118,-0.1586551517,-0.1597523093,0.0542651601,-0.2796344161,-0.137821123,-0.386287123,-0.1363766193,-0.2332428545,-0.2318859845,0.0297000054,-0.0364917666,-0.2091930062,-0.1093644574,0.3059449196,0.1775646955,-0.511921525,0.0521571189,-0.1740999818,-0.0387714207,-0.3981323242,-0.0045327442,0.2409882098,0.0839544907,0.4800804853,-0.2831428647,-0.0959866494,-0.1280893981,-0.2729632258,0.1172757372,0.0654704869,0.1425894499,0.0450697206,0.0209899656,-0.148674652,0.0222847611,-0.058536347,-0.2918953896,0.1592782736,-0.0876687467,-0.2113157958,-0.050128635,-0.2477100044,-0.5144851804,-0.2921965718,-0.1058580354,0.163472414,-0.197786957,0.2133414596,0.1712711751,-0.1417769492,0.1947842985,-0.1191123575,0.1587194353,-0.0597623587,-0.5474858284,0.2177067995,-0.0215540417,-0.2499404699,0.1782260835,0.2267056108,-0.0758668929,0.133033365,-0.4908933043,0.2066096216,0.0889606774,-0.0867545307,0.5087793469,0.2504207492,0.1344650239,-0.0427322648,-0.012821747,-0.0286974013,-0.2431729883,-0.0593144186,0.4615968168,0.0892490149,-0.0717798695,-0.0344243273,0.2145276666,0.7158355117,0.1123102084,0.2709057629,0.0734363049,0.1690087914,-0.0442204662,0.1032898054,-0.0450689606,0.3228220046,-0.0915730968,-0.1981370002,0.3320971727,0.1605146974,-0.3140473664,-0.2965776026,-0.0691014081,-0.083694011,-0.2231552005,-0.0285991021,0.0346235149,-0.0391559117,0.2152923793,0.035167072,-0.1530727893,-0.1862273365,0.0510190539,0.2082631439,0.0232878625,0.1214372665,0.1869958788,0.0174502842,-0.5837151408,-0.1514384896,-0.0604083873,0.4268853366,0.1196801215,-0.0053087068,0.2003970742,-0.1578143686,0.1050109565,-0.1754297465,0.1773994118,-0.2610222399,-0.292027086,-0.5091144443,-0.2712901831,-0.2293028384,-0.4954594076,0.3333487511,0.1774192005,-0.2093194425,-0.0772734284,0.1002501249,0.1798802614,-0.0299486127,-0.047129944,0.090334177,-0.4981131554,-0.0542902462,-0.2347117066,-0.2449136674,-0.2359820008,0.1032374054,0.1892080903,0.3253815472,0.3559901118,0.0577487461,0.0654738545,-0.1120626852,0.0077309157,0.1831941158,0.1641881764,0.5088049173,-0.2714176178,0.5144896507,0.3122367263,0.277520299,-0.1294936538,-0.0580586642,0.1091769785,0.7778987288,0.1361853331,-0.0957429633,-0.2226422429,0.0911439583,-0.2652897239,0.0761789605,0.2636010051,-0.1587104797,0.0467036292,-0.1856402904,0.4024966061,0.0445795581,-0.1112493575,-0.1283761412,-0.3259797692,-0.0579226762,0.0264297128,0.1786329746,1.2413955927,-0.0512667336,0.5479646921,0.339854151,-0.373105973,0.6129063964,-0.3549629152,-0.2105882317,0.1111223698,0.4099706709,-0.0888913348,0.0101735676,0.1415470988,0.138372466,-0.1926465631,0.321441561,0.3675510287,0.2141028047,0.1016231254,0.4243979454,-0.1762019545,-0.2723134756,0.1168755293,0.0460686907,-0.0769592524,0.3433996141,-0.0870346799,-0.2236351967,-0.1582718194,-0.1783522218,-0.3901975751,0.0252667256,0.0243971124,-0.0944943652,-0.4595492184,-0.3507035971,0.1282486916,-0.4450742304,-0.2020016909,0.178225711,-0.2774980068,0.0638343245,-0.4113684893,0.2809457183,0.3688465357,-0.1442646086,-0.0672983453,-0.3414596319,-0.1593785286,-0.2439575642,0.1144457608,0.0914485529,-0.1264486164,-0.1907732189,-0.1291674376,0.4585465491,-0.1421941668,0.3873198628,-0.1405364275,-0.143874824,0.1080984473,0.0851507112,-0.2322066128,0.2715298533,0.0208108239,-0.2724520266,-0.3984046578,0.5281153917,0.0161830932,0.0225141738,0.1650773436,0.204115361,-0.1229282022,-0.2184978276,-0.0596865229,-0.3233775496,-0.2938448191,-0.0270117298,-0.2106590569,-0.0880985782,-0.1055648327,0.1250372827,0.559166193,0.0456969589,-0.0706096143,-0.0698530003,-0.0359770134,0.1490237415,0.1137292907,0.113996461,-0.2447698712,-0.1570647657,0.0539070815,-0.0921541229,-0.2827051282,0.0909492224,0.0752789974,-0.0176336244,0.1012666225,-0.1367356479,0.0663047284,0.0776910111,0.0151347797,-0.2517248094,-0.1639629751,-0.1377808899,0.042820029,0.1708699465,0.1362432837,-0.5324262381,0.3317977786,-0.228767097,-0.1382037103,0.1105390787,0.1144325957,0.101974234,-0.1439725757,0.1564760953,0.0012162264,0.0148285879,-0.2591868639,0.4146894813,-0.0019636666,0.3680515587,-0.0486135259,0.0644156411,0.4446602464,-0.0495102182,-0.0428194217,-0.0697860345,-0.1663296819,0.4053024948,0.6680960655,-0.0827416703,0.3558437526,0.0900747553,0.2813591659,0.342007041,-0.0242601074,-0.2278636396,0.3319764137,0.1144141331,-0.0572645366,-0.2742274702,0.1136649773,0.2460724413,-0.1159611046,0.1479768306,0.3108568192,0.1180342361,-0.4081661403,-0.0566990413,0.1665659994,0.2172143757,-0.0733055323,0.3288076818,0.0730559528,0.2956299782,-0.2882792354,0.0651037544,-0.0272113737,0.0072313184,-0.0644333214,0.3254032135,0.1335479766,0.5295645595,-0.015241893,-0.2173381597,-0.0067767552,0.3404940963,0.0472745858,0.1829468906,0.1153544858,-0.1629121453,-0.0913250446,-0.168131277,-0.3184599578,0.3788322508,0.0995471105,-0.227999121,-0.1665635705,0.0182778612,-0.2023857832,0.4663447142,-0.410246551,-0.0335330926,-0.0610462949,0.4058499336,0.1479474157,-0.226137355,-0.1839331686,-0.4843294322,0.1700115651,-0.2348323315,0.0972104296,-0.1264779419,0.0262408871,0.0391526073,0.4013298452,0.0752618685,0.1574715972,-0.1435383111,-0.0534454919,-0.0087841712,0.0822140723,-0.2174326032,0.2964996397,0.0206260551,0.099622637,0.198422581,0.0122812483,-0.1984613985,0.029300848,-0.260643661,-0.7364006639,-0.2951452136,0.2905284464,-0.103632316,-0.0496596061,-0.2225229442,0.0644874424,-0.3825047016,-0.0347700529,0.6124904752,0.1132045761,0.2108598351,-0.0199695267,0.032223545,-0.2134573907,0.6463638544,0.2980593443,0.2459907383,-0.3212292492,-0.1414684802,-0.3462895453,0.0655435547,-0.215190351,0.1030265316,0.0468518734,0.0800823793,0.0053181364,0.4956417084,-0.0943778753,0.6934795976,-0.3345608413,0.0004036961,0.0100862542,-0.1352155209,0.0052141664,0.0981654897,-0.1735398769,-0.2624406815,0.1998807341,0.0391980633,-0.1149830595,-0.0844619796,0.2595461607,0.1401650459,0.1990090907,0.2034509629,-0.0767749175,0.2250240892,0.10662736,-0.3222239912,-0.4003611505,-0.0525873676,-0.1452838928,-0.0425835773,-0.372171104,-0.0040717563,0.0323500633,0.3106036186,-0.1983379722,-0.0282453764,-0.0853505209,0.1447116435,-0.1080791354,0.0099683646,-0.4353299439,0.0790129304,0.1022066921,0.2661632299,-0.1070330739,-0.2484223098,-0.3806074262,-0.0880279914,0.3034051657,-0.2912916541,-0.0510536507,-0.2912946641,0.3219889104,-0.1367405802,0.1849550158,-0.0974545702,0.0435680486,0.1183546633,-0.1040178612,-0.170769915,0.2843350768,0.2103105336,-0.2942461371,0.0086486926,0.3904806077,-0.1679489762,-0.2728066444,-0.2695900798,-0.1056654528]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/525","title":"wmt download speed example","comments":"Should we add an argument in `load_dataset` to override some URL with a custom URL (e.g. mirror) or a local path?\r\n\r\nThis could also be used to provide local files instead of the original files as requested by some users (e.g. when you made a dataset with the same format than SQuAD and what to use it instead of the official dataset files).","body":"Continuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n","comment_length":63,"text":"wmt download speed example\nContinuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n\nShould we add an argument in `load_dataset` to override some URL with a custom URL (e.g. mirror) or a local path?\r\n\r\nThis could also be used to provide local files instead of the original files as requested by some users (e.g. when you made a dataset with the same format than SQuAD and what to use it instead of the official dataset files).","embeddings":[-0.1401740611,-0.0430655032,0.0662449524,0.0506724678,0.0602064915,-0.1812299788,0.2489060909,0.1119349077,0.1134405211,-0.013772957,0.0660708994,0.5972366929,-0.057672888,0.0454577804,0.2306551784,0.0531793498,-0.1269133091,-0.0680238828,-0.5435426831,-0.1649612188,-0.0084484592,0.1539152563,-0.0286870617,-0.0853236392,-0.0160792787,0.1635005921,-0.0522697493,0.1442735791,-0.32329005,-0.422385633,0.1229837611,0.2223754674,0.0290668588,0.2388714105,-0.0001091673,-0.034856528,0.3812175989,-0.0277624056,-0.5606554151,-0.206246078,-0.5388450623,-0.5287117958,-0.051505778,-0.150037542,0.1338907182,0.0201803222,0.0294659454,-0.052528102,0.2164442092,0.0395456143,0.1784766912,0.3661825359,-0.5167102814,-0.0938478112,0.4847781658,0.002749362,-0.2234309316,0.4499016404,0.4752109349,0.1483808309,-0.0139931412,0.3062309623,-0.2024827003,0.3198318183,0.2504799664,0.0922966748,0.2866598964,0.0172408521,-0.0795295462,0.5750318766,0.4227974117,0.0624039099,-0.5002784133,-0.3649969995,0.1189359054,-0.1521324068,0.1861514449,0.279928416,-0.3316590488,0.1248505339,-0.5772380233,-0.0171520803,-0.0086234557,0.3052995801,0.0751177222,0.636567831,-0.0298028067,0.088975735,0.0519584753,0.023126917,0.0600558296,-0.3546413779,0.1290397644,0.0254483409,-0.064096719,-0.1827630997,0.1450038105,-0.1434669644,0.0556561686,0.0483117253,0.0375731401,0.2358700186,0.2863411903,0.041620519,0.2809550166,0.1880232841,-0.2490612119,-0.1966480166,0.1953296065,-0.0647456199,0.174302727,0.1942556053,-0.1582804024,-0.0995948687,-0.4025150239,-0.1299947351,0.004997449,0.0445326194,-0.2315447181,-0.3442460597,0.16343382,-0.1543385237,-0.0295567196,0.2383633107,-0.2319650352,0.2009649128,-0.236982584,0.0273457747,-0.2427757829,-0.0867126435,-0.0774694309,-0.075891301,-0.1686362773,0.133510083,0.401701659,0.1725014597,0.3206163645,-0.0123998737,-0.2200240046,0.0505857244,0.1651203632,0.1433419287,-0.0575465783,-0.04798159,-0.0609802231,0.3564445972,-0.1713732183,0.1529389322,-0.3698183894,0.1584869623,-0.4505815208,-0.3000794351,0.0098978011,0.1790207475,-0.3993910849,-0.3566007316,-0.1363359839,0.3336696923,-0.4130515456,0.1365327537,-0.1254845262,0.0593281426,-0.3018197417,-0.0340455696,0.0507970564,0.4687490761,-0.224972412,-0.1960880458,-0.1931681633,0.2358486205,0.3037902117,0.4155172408,-0.3779424131,0.1080230027,-0.1842556596,0.0571966544,0.3586405218,-0.3323921859,-0.2837767601,0.5307893157,-0.5233163834,0.0617463291,0.2024645805,0.1216159612,0.2539978325,-0.1279946119,0.0827525482,0.6500003338,-0.0270980354,0.1977931559,-0.458914727,-0.4428385794,0.2827785909,0.1533723176,-0.2588041723,-0.0776052997,0.0641294047,0.2633578181,0.2396986336,-0.0735427812,0.0113799656,-0.0566637255,-0.0633372068,-0.3452426791,-0.1362152696,0.3386700451,-0.4280269742,0.3024757504,-0.0730890036,0.3111579716,0.198628366,0.0526191257,-0.1303733438,-0.204859823,-0.0220026597,-0.0455685183,0.0781750455,0.2087099105,-0.0010314339,-0.0329714008,0.0532725044,-0.0190537535,-0.0692303181,0.1288835555,0.3759290874,-0.0387995914,0.0337791778,-0.0809425041,0.3571797609,-0.0392466597,0.1040321514,-0.0517284125,-0.1456834227,0.2341835201,-0.137287423,0.7774727941,0.3542417884,0.0056494446,0.2782478929,-0.2661018074,0.480173856,0.450812608,0.0903477073,-0.004709295,-0.4777366519,0.5367771387,-0.1658912152,0.2051068097,0.2896354198,-0.0087829707,0.0856239945,0.0277637858,-0.3223411739,-0.0495687351,0.2209577709,0.1809858978,0.0623554848,0.1957708001,0.1109706983,0.3044623733,0.6227391958,-0.1718568504,0.0422483385,0.2157825679,-0.2499338388,-0.5171298981,0.3028015494,0.3569072485,0.4761772156,0.2474783063,0.4310215116,0.0737211183,-0.0636577085,-0.2712914646,0.3826601207,0.1265354306,-0.374804616,0.0553989448,0.1359221339,-0.0237425622,-0.2967222631,0.1882191449,0.0841449648,0.0228420403,-0.1887816489,0.0943800062,-0.4816093743,-0.244949609,-0.3461509347,-0.1799567193,-0.2855333984,-0.3380047977,0.0620545037,-0.1093808338,-0.1883797348,0.0807610974,0.2616901994,0.2746520936,-0.3171329498,-0.039859578,-0.1123803109,-0.1213930696,-0.3910529315,0.0253416337,0.265301317,0.2031013519,0.3739596903,-0.2555414736,-0.0685693249,-0.1352080107,-0.3251989186,0.1363383979,0.0960098505,0.3435523808,0.1109589636,0.1049076021,-0.1508765221,-0.0654419661,0.0444721617,-0.2533891499,0.1706563681,-0.1076092795,-0.2372186184,-0.0109760957,-0.1955933124,-0.5634587407,-0.3295453191,-0.2169276029,0.1890612245,-0.0816294104,0.1998165697,0.3314262927,-0.0365300886,0.3603949845,-0.1359031498,0.0994871035,-0.1641945541,-0.5906572938,0.3921245337,-0.1721303016,-0.2118633091,0.2214899659,0.0735787451,0.1024077386,0.0603677891,-0.4876346886,0.1473144442,0.0086398572,0.0074478332,0.3440036178,0.2092840225,0.197977677,-0.1226305515,-0.019643236,-0.0573194697,-0.373526603,-0.0599367134,0.4258570671,0.2736263871,-0.1724301428,0.0042299689,0.2495697439,0.621309936,0.008626014,0.2401248515,0.1506208926,0.0887310132,0.0395380706,0.2544414103,0.0111293914,0.4417012632,-0.0943876728,0.0097343251,0.3715402186,0.2294644713,-0.3094246984,-0.1197520867,0.0856709555,0.041375462,-0.3028289676,0.0665811226,0.0872152075,-0.1653722376,0.20366247,0.1021770164,-0.3788698316,-0.2287276089,0.0009728362,0.3664315045,-0.0112813944,0.1826778352,-0.1534882486,0.1495300382,-0.5660690665,-0.1106044129,0.0032998549,0.2913611531,0.1697706431,-0.0129302358,0.0722882822,-0.0817571506,0.1548693627,-0.2576476336,0.08386603,-0.1553104073,-0.1887220442,-0.5956742764,-0.2519658208,-0.1097417772,-0.3701532781,0.3256337941,0.1633917391,-0.1536768228,-0.2543477118,0.2115948349,0.0952994525,0.0143090533,-0.0936372951,0.0478198156,-0.5150456429,-0.1141292453,-0.1453862637,-0.2715370357,-0.1408784986,0.0269033276,-0.0670617595,0.2952195406,0.3303209245,-0.0917467698,0.0781196803,0.0124163218,0.2065200806,0.2116667926,0.1049581468,0.3575665355,-0.2163476646,0.4708114564,0.1069052368,0.0359313488,-0.0078181438,-0.0741849765,0.1217956841,0.679254055,0.143856883,0.0894057453,-0.1710984856,0.0827874914,-0.2717101574,0.2893840075,0.3266997039,-0.1793211997,0.0642348081,-0.3848208189,0.343978256,-0.0726121217,-0.1075433567,0.1265313625,-0.377702117,-0.1238894463,-0.054910548,0.2150656879,1.1090852022,-0.1453067511,0.3599414229,0.3286302686,-0.3125862777,0.6720408797,-0.1128684208,-0.1041186824,0.132355243,0.3963906169,0.0688303486,0.0030712653,0.1413909197,0.32410115,-0.3143125176,0.310819149,0.3265374899,0.2781631947,0.1273645312,0.5954557657,-0.1624188721,-0.4131713212,-0.1340104938,0.1710876226,-0.1431821436,0.2317877114,0.0013285313,-0.1586711556,-0.0011413035,-0.1303226501,-0.3256799579,-0.0010128296,-0.2405777574,-0.1992855966,-0.3163982034,-0.3503407538,0.174093917,-0.1777689457,-0.254717797,0.1785103977,-0.3482359648,0.2497154474,-0.495749861,0.2895105481,0.2319345772,-0.1472203583,-0.0692785978,-0.2525032163,-0.1565076262,-0.2393179536,0.0925187096,-0.0974793211,-0.0912379399,-0.245271042,-0.0557976402,0.3949801028,-0.1176541746,0.2038953453,-0.1565934718,-0.2492966801,0.1574815214,0.1143657789,-0.1629769653,0.3960427344,0.1069387943,-0.2614466846,-0.341365099,0.488540411,-0.2765487432,-0.0615248047,0.2336362451,0.1557389349,-0.1110067144,-0.2471162826,-0.0362695344,-0.24580203,-0.1622366607,0.0235787071,-0.2357708514,-0.0581379533,0.0265014768,0.2130108327,0.3274288177,0.156762898,-0.025536187,-0.1540929824,-0.0428745635,0.1159028485,0.1328278184,0.0901324973,-0.3304200768,-0.1285469383,0.0739919096,0.0358344689,-0.3404176533,0.1331428885,0.001846363,0.0540982299,0.0885093808,-0.1253241599,0.1922050565,0.0758357719,0.0379020236,-0.2574374974,-0.1625723094,-0.2088868171,-0.1289773285,0.1288842261,0.0279289931,-0.5542566776,0.2583650053,-0.2548000216,-0.0465644449,-0.0423148945,0.1137969717,0.0943844169,-0.2149706632,0.1778300554,-0.0196649414,0.0237758309,-0.1117223576,0.1866171658,-0.0113710668,0.3861462772,0.1349964887,0.2166286856,0.3578079343,-0.0335595794,0.0262251403,-0.03449912,-0.0337326489,0.4142051935,0.5385565162,-0.0902963206,0.2103455961,0.1377086639,0.1006859466,0.3548391461,-0.0678966716,-0.1434862167,0.3465021849,0.1763028949,0.0150441993,-0.3085129857,0.049649816,0.1041296124,0.01535815,0.1124413684,0.1885599345,0.1076685786,-0.2401627898,-0.1056161076,0.3327112496,0.0869314745,-0.0052429759,0.2966088653,-0.1117819324,0.1312162429,-0.1709469259,0.0042504086,0.0453569591,-0.0272480175,-0.0770828053,0.2865386307,0.0881540254,0.5736840963,-0.103993617,-0.1535383165,-0.0302681923,0.4111757278,-0.1179981008,0.231750235,0.1915976852,-0.0341953486,-0.0863341913,-0.1675744653,-0.1613700092,0.2868330479,0.0892394185,-0.3066066504,-0.3026069105,-0.1101242751,-0.240702495,0.341642499,-0.3660486937,-0.0802694708,-0.053052485,0.2778140903,0.1537456512,-0.2622273862,-0.1393932253,-0.3666598797,0.0990124047,-0.2527454793,0.0838493928,-0.1308987886,0.1243994385,0.0224082731,0.1503724605,0.1136518866,0.2444439977,-0.1679368913,-0.0952028558,0.094759047,0.0610182695,-0.2260663509,0.2288854569,0.0501586087,-0.0193012394,0.2327535897,0.0663528517,-0.2215143591,0.0119334003,-0.1773615927,-0.7652970552,-0.2818597555,0.2427919358,0.0895274952,0.0186084099,-0.2629157603,0.1194325015,-0.1581221372,-0.0575796701,0.5074003935,0.0057014157,0.2013276219,-0.0759596527,0.0824250951,-0.1424697191,0.696980238,0.1067084968,0.1756349951,-0.3695183396,-0.1531882137,-0.555154264,0.020267617,-0.1759876609,0.0439421833,0.181924805,0.2019901425,-0.0927359313,0.5005363226,-0.0832173675,0.5214187503,-0.0795298293,-0.0151403947,-0.1287299395,-0.2067317367,0.0540204868,0.1199586391,-0.0565971322,-0.2893303931,0.1644916832,0.0172600653,-0.1056062803,-0.1385487169,0.195702374,0.1395480782,0.1905782372,0.1874487698,-0.0867070481,0.2968438566,0.0934024602,-0.2792479694,-0.441105634,-0.1013871059,-0.2070908546,-0.0262069516,-0.4229453504,-0.0029901285,0.060950961,0.214826405,-0.3219360411,0.054262206,-0.0594870336,0.1481987536,-0.0824299604,0.1172231138,-0.2113015205,0.1959909797,-0.0120273354,0.1964811683,-0.0904243365,-0.0009058007,-0.2757561803,0.0070145857,0.2989305556,-0.260360986,-0.1206921041,-0.3642936349,0.2702487111,-0.207114026,0.203606993,-0.1141023561,0.032281965,0.2379381359,-0.0222251862,-0.1997592449,0.3995940387,0.1431195736,-0.1232108995,0.0237852018,0.2609974742,0.0480974652,-0.1497206837,-0.165486306,-0.1931341141]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/525","title":"wmt download speed example","comments":"@lhoestq I think we should host it ourselves. I'll put the subset of wmt (without preprocessed files) that we need on s3 and post a link over the weekend.","body":"Continuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n","comment_length":29,"text":"wmt download speed example\nContinuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n\n@lhoestq I think we should host it ourselves. I'll put the subset of wmt (without preprocessed files) that we need on s3 and post a link over the weekend.","embeddings":[-0.2155659497,-0.1206225529,-0.0204935316,0.1300849617,0.0658164695,-0.1955837756,0.2002224177,0.1801560968,0.1433986127,0.1119243875,-0.0341102295,0.4751238227,-0.1775703728,0.2044598758,0.2067526579,0.0962957665,-0.0070732124,0.0183814168,-0.5470451713,-0.1375177205,0.0861985013,0.2179209441,-0.0510405712,-0.1955765486,-0.0300692152,0.1657574326,-0.1316437721,0.0727209076,-0.4188373387,-0.2630718052,0.1180561185,0.2656253278,0.0648002177,0.3028047681,-0.0001020328,-0.1106325164,0.2632767856,-0.0324820802,-0.4443925619,-0.1743226349,-0.5166602731,-0.5105953813,-0.2215472609,0.0091433972,-0.060807012,-0.0149904126,-0.0065327538,-0.013225141,0.2297478914,0.0044426573,0.2328479588,0.359841466,-0.4252631664,-0.0844932944,0.4815312922,0.0049271262,-0.2498247623,0.4009498656,0.3564257026,0.0796066523,0.0113519337,0.4700137377,-0.2561063766,0.2787464857,0.2031781226,0.0228023194,0.2285641879,-0.0028507665,-0.1551726609,0.5608808398,0.3373489976,0.1065415591,-0.5177890658,-0.2909383774,0.0493709818,-0.2340979427,0.1028875336,0.2994498312,-0.3247202039,0.1200974882,-0.5581644773,-0.0929765403,-0.0255202781,0.2339274883,0.0102101695,0.7002000213,-0.1331142336,-0.0084219025,-0.0204181653,-0.0180142485,0.0947145075,-0.1942599565,0.0571601465,-0.0769828856,0.0418055803,-0.2353392243,0.1806431413,-0.1294105798,-0.0580373444,0.1132813096,-0.133883819,0.1659895033,0.4308972955,0.0214825198,0.1853033602,0.1846225113,-0.2198336124,-0.3102083206,0.2138497382,0.1062673479,0.207175687,0.193554759,-0.2216509879,-0.0671727359,-0.4991372228,-0.0334163196,-0.0477717929,-0.0327215046,-0.1735000014,-0.2733834684,0.1939501911,-0.1617196202,-0.1389788389,0.2310477495,-0.2283917069,0.2743051052,-0.1372373849,0.044894442,-0.1253168732,-0.1350657791,-0.0852319673,0.025963204,-0.1604275256,0.0628022477,0.4178491235,0.1151154041,0.2968902886,0.0482058227,-0.1270541698,0.0623152032,0.1218127832,0.24790214,0.0074638347,0.0301758628,-0.0424834564,0.2363862395,-0.2150469273,0.2390692085,-0.3477389216,-0.0609671921,-0.2003092021,-0.2398068309,-0.0375852101,0.2353386134,-0.3309755325,-0.2809259593,-0.0166015718,0.2984886765,-0.340136379,0.224585399,-0.0754440054,0.0480984673,-0.2222734839,-0.0174944568,0.0495946966,0.3618355989,-0.0601066723,-0.0221713018,-0.1672596782,0.3471522331,0.4115245938,0.3610854745,-0.1718593538,-0.0356686562,-0.1001421884,0.113896668,0.1149026901,-0.2124214321,-0.1597549319,0.5290461779,-0.4272652864,-0.0512516908,0.1980796307,0.0377272069,0.1451046169,-0.1989156604,0.10155157,0.6101759672,0.0147081343,0.0896477997,-0.4038211107,-0.4023844898,0.2538384795,0.1133946627,-0.2388010919,-0.225761503,-0.0191058647,0.2652854025,0.2038480937,-0.0947933644,0.0399125442,-0.0602308176,-0.0004274945,-0.3014577627,-0.2138361335,0.1486183554,-0.1586669832,0.2525816858,0.0318034999,0.3712976277,0.2999790907,0.0660214499,-0.0779223368,-0.1569643617,0.0843522847,-0.0943670049,0.1866601259,0.2646548748,0.0709149018,-0.0910296664,-0.0194461998,-0.1746232808,0.1247932091,0.0471184812,0.346439153,-0.2572759688,-0.0442039818,-0.0858551264,0.4426569343,-0.0285973251,-0.0212582517,0.0165666994,-0.0540228747,0.1077196226,-0.1630091071,0.7219814062,0.2625408471,0.090217568,0.1276442707,-0.188795343,0.4084606767,0.3476678729,0.0464663953,0.0551670007,-0.3467504382,0.5725204945,-0.0478213318,0.1236320138,0.1868003458,0.1262732148,0.0195048098,0.055102542,-0.2233821601,-0.0770522431,0.3090645075,0.0208185948,-0.1342979372,0.1040955707,0.0351863764,0.2854640782,0.6748992205,-0.1631390154,0.1368694454,0.1020002142,-0.2426295578,-0.5028501749,0.3370275795,0.2667830884,0.2922967076,0.2082756013,0.5471039414,0.1239030212,0.1133713201,-0.2837970257,0.441649884,0.0307938959,-0.24003914,0.0520266294,0.0643799454,0.0097814091,-0.2793463469,0.2093320638,-0.0416769572,-0.0475103185,-0.0696106702,0.1332087368,-0.3933753967,-0.2082936317,-0.1213198006,-0.2465979159,-0.2604210973,-0.2351294607,0.1347340047,-0.0801403672,-0.1687469184,-0.0055047055,0.3391518891,0.1928296387,-0.3416621983,0.0859147161,-0.2080182135,-0.1062614843,-0.3279315233,0.1355035156,0.2595315576,0.0796539113,0.4632073939,-0.2562026083,-0.0677856952,0.0012308706,-0.3010240793,0.1154059619,-0.0636522472,0.297429651,0.1763873547,0.0849846005,-0.2226348966,0.0782181993,-0.0378867052,-0.2860150337,0.0537933968,-0.0335327461,-0.1982925683,-0.149812445,-0.2401798964,-0.463853538,-0.2347120792,-0.1960784644,0.2384764403,-0.1457492709,0.1154222563,0.1634501964,-0.132123813,0.2078085244,-0.0888709202,0.1999312192,-0.1722768247,-0.5507377982,0.3365274668,-0.1226580814,-0.2967382371,0.2950409055,0.1803974211,0.163167119,0.0276233703,-0.4810869396,0.2160256803,0.0755470917,-0.0122779887,0.4312326312,0.1811036915,0.2614403367,-0.0525122471,-0.1331841648,-0.116903089,-0.2547282279,-0.1903392375,0.3488098681,0.0875853747,-0.28283149,-0.0525928028,0.3150904179,0.4403005242,0.0352593958,0.2692222297,0.145313397,0.1900500357,0.0205959342,0.181396842,0.0188444369,0.4231962264,-0.0633487329,-0.0132917324,0.3236269355,0.1291986555,-0.3245406449,-0.2914395928,0.0539339259,0.0542901084,-0.199147433,0.0365903564,0.03283678,-0.0825024471,0.260483712,0.113652125,-0.1314818114,-0.2285270691,0.1083636656,0.2110732794,0.0231594499,0.1771333367,0.0534082241,0.0787190348,-0.6460815072,-0.111149773,-0.0496117547,0.2459914833,0.0804355145,-0.1045289934,0.0938564464,-0.2556418777,0.0522537306,-0.2582125962,0.1134897545,-0.0836109444,-0.260478735,-0.4620932639,-0.2324958593,-0.3129477799,-0.4964567125,0.278830111,0.1503121704,-0.1244363934,-0.2181077749,0.0399517529,0.0615079105,0.0328092612,-0.0039066817,0.0036900153,-0.5525987148,-0.1009244025,-0.087520428,-0.191000998,-0.1118557826,0.1091226637,0.1032060012,0.2399267405,0.4019559324,-0.005945425,0.1635181308,-0.1407597512,0.1461706907,0.1538026184,0.0355157927,0.3521102667,-0.2427490056,0.4205362797,0.2550263405,0.133293286,-0.0101003498,0.0425596386,0.1045064181,0.5606143475,0.1336904019,0.0361020714,-0.1573226899,0.1971204877,-0.4761061668,0.1635749638,0.3196149766,-0.0963663086,0.1028660983,-0.2305859476,0.2454120815,-0.1227336675,-0.0561032891,-0.0822792873,-0.419978857,-0.0470854416,-0.0305168852,0.3228857815,1.087644577,-0.0548614599,0.4298871756,0.4172224402,-0.2629695535,0.6519911885,-0.1784108877,-0.0568831898,0.062689878,0.425978452,0.0377705768,0.1224648431,-0.0300477967,0.1886978,-0.336374253,0.280685991,0.393296808,0.1404631734,0.0706053451,0.4829426706,-0.2904085219,-0.3686588705,-0.131205976,0.216467157,-0.1190492958,0.366171211,-0.0052992525,-0.2352270931,-0.0096470974,-0.0482901782,-0.3888449967,0.1304736286,-0.025182547,-0.1556475013,-0.4032443762,-0.2059352398,0.1470164657,-0.2993979752,-0.1754350215,0.2969132066,-0.3061264455,0.1365481615,-0.373716116,0.1065821052,0.3472371399,-0.1877133846,-0.0940727144,-0.3946021795,-0.0860472918,-0.3269335628,0.1791285872,0.0974243283,-0.0737682208,-0.2169006914,-0.0856583789,0.4046719968,0.0004224118,0.2502942979,-0.0447824188,-0.2648874521,0.214547053,0.1384792924,-0.2282404155,0.3342369199,0.0763717219,-0.1768038422,-0.3621092737,0.407646805,-0.1357071251,-0.0548254699,0.1094947085,0.3518755734,-0.1896471977,-0.3585042655,-0.0407099463,-0.2579796612,-0.1610672623,0.0826614499,-0.0515572727,-0.1386895329,0.044767078,0.147168532,0.4218549728,-0.063857317,-0.0254352819,-0.1077773422,-0.0039006157,0.0178715531,0.1565532386,0.116929844,-0.2704853415,-0.2178095281,0.0143177658,0.1008339003,-0.4075272083,0.0065378011,-0.0030575469,-0.0312984884,-0.0204132758,-0.2406107187,0.1910777092,0.1239069626,0.1061534882,-0.2774294615,-0.1806873083,-0.2569988668,-0.0778445527,0.1051139832,0.0400192775,-0.5048238635,0.4163051844,-0.1469412595,-0.2252707779,-0.062245924,0.1531935334,0.0653287619,-0.1851453483,0.1893167347,0.010625124,0.0315577127,-0.1294826865,0.155529961,0.0254080426,0.3431404531,-0.0174623691,0.1370590329,0.498419106,-0.136776343,0.0979610384,-0.0604100525,-0.1392354965,0.3413677514,0.6376524568,-0.0923818946,0.1332418174,0.0038226275,0.2270390689,0.3653770685,0.0729362369,-0.105343923,0.3341254294,0.2490955293,-0.0806622058,-0.2883702815,-0.1841811389,0.1503634304,-0.0476041809,0.0749274492,0.1278429329,0.0845283717,-0.2489455044,-0.0464622863,0.0413438603,0.1420660764,-0.0882272646,0.2352517843,-0.040843159,0.0806747079,-0.2251591831,0.0038357256,0.0022387612,-0.0712820739,0.0497234501,0.2862336934,0.0757821128,0.5007494092,-0.1257101893,-0.1692750454,-0.0077233957,0.3311446309,-0.0134001123,0.2483821958,0.1064843014,-0.1253101379,-0.1222495735,-0.1940726191,-0.2696334124,0.4114986956,0.0255447179,-0.2943881154,-0.1661920696,-0.0054864082,-0.2019101679,0.4600009322,-0.2653143108,-0.0194640514,-0.0762196258,0.3743437231,0.1899861544,-0.1610188633,-0.2972479761,-0.3647537529,0.175100565,-0.2916701436,0.0634526983,-0.0151320137,0.0053437832,-0.1055820361,0.1762616932,0.221974656,0.1720624119,-0.2952432334,-0.0167615637,0.0213408507,0.0078680022,-0.2192863822,0.3557500541,0.0686776564,0.1814699024,0.1829157919,0.0935287029,-0.2586449087,0.0309435483,-0.2813492417,-0.6952272654,-0.0633432865,0.1894435138,0.0416124165,0.0927127153,-0.3185550869,0.2004051656,-0.2808116674,-0.034728989,0.4659156799,0.0250449833,0.1504021436,-0.1765357256,0.1219728589,-0.1205238849,0.727730751,0.2135393769,0.2480608672,-0.219800517,-0.201987952,-0.3592083156,-0.0052028485,-0.1304275542,0.0303927567,-0.0436069295,0.2344098836,0.0169191901,0.4801867008,-0.2524454296,0.5650542974,-0.1560819298,0.1449367106,-0.0684496984,0.0334084071,0.1315193325,0.0886436626,-0.0565040298,-0.2338403314,0.1833182722,-0.0248792022,-0.0018763784,-0.1556955725,0.2322485745,0.1646092385,0.2284878641,0.1511097997,-0.0296817832,0.1483923793,-0.0339799747,-0.4519455135,-0.4413381815,0.0062540607,-0.1830948293,-0.0221662782,-0.2848102152,0.0437055342,0.1046818495,0.2613090277,-0.2944231927,0.0885809064,-0.1664045304,0.277145803,0.012628451,0.050910987,-0.3258062303,0.2498874664,0.0871119797,0.1541566402,-0.0772977769,-0.1914646775,-0.4286545813,-0.1590392739,0.3215028048,-0.3492078483,-0.1736403555,-0.3517514467,0.1742374897,-0.1791326553,0.1163301617,-0.1438179314,0.1457781643,0.0896577165,0.0786117315,-0.1525877565,0.4270353019,0.1251838654,-0.1900885552,-0.013504371,0.4377389252,-0.0389466956,-0.18233715,-0.2686680555,-0.2382282764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/525","title":"wmt download speed example","comments":"Is there a solution yet? The download speed is still too slow. 60-70kbps download for wmt16 and around 100kbps for wmt19. @sshleifer ","body":"Continuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n","comment_length":22,"text":"wmt download speed example\nContinuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n\nIs there a solution yet? The download speed is still too slow. 60-70kbps download for wmt16 and around 100kbps for wmt19. @sshleifer ","embeddings":[-0.1923046708,-0.2106527686,0.0381692983,0.1611495763,0.0975149423,-0.1585473865,0.1904044002,0.153340131,0.1204154789,0.000796883,0.0680536181,0.4522775412,-0.1424290836,0.1149549261,0.1379803121,-0.0190718397,-0.0316072032,-0.0164662972,-0.5539655685,-0.1806611121,0.1243808717,0.2241039574,-0.0454688706,-0.1134556755,0.013256277,0.1571440846,-0.0064908247,-0.0526825115,-0.4414258003,-0.2438793778,0.1700959951,0.2271818817,-0.0814935416,0.2266278118,-0.0001094717,-0.1007321849,0.3573346436,0.0219913293,-0.471912086,-0.2320852578,-0.4433206022,-0.5319789052,-0.1851849109,0.0123840664,0.0689449012,-0.0021641352,-0.007084114,0.1026333496,0.1497532278,-0.0618960895,0.1908744127,0.3501648307,-0.3927136064,-0.0803053454,0.529024601,0.0538763329,-0.3179961443,0.4086429179,0.4026593566,0.0647772253,0.0284354519,0.3826526999,-0.2017960846,0.3520322144,0.1550590545,0.1255157441,0.3314122558,0.0501096062,-0.0919991508,0.5831670165,0.31319502,0.1831501722,-0.4971561134,-0.3578057289,0.0633992627,-0.2700966895,0.0803164318,0.285651058,-0.424318254,0.0002229383,-0.6766803265,-0.0099584851,0.0651908144,0.2772797942,-0.0481214151,0.666228354,-0.120648168,0.0215347987,0.1117916033,-0.0764692426,0.0303589907,-0.2781360149,0.0698558316,-0.0057149301,-0.0622209907,-0.1293959916,0.1594761759,-0.030263897,-0.0472638793,0.033974696,-0.0991022661,0.2560758591,0.312438041,-0.0474327244,0.2130236626,0.3228138387,-0.2606376112,-0.1927301586,0.2396687567,0.1491972506,0.1560321599,0.2299154997,-0.1587983221,-0.0781511515,-0.4115286171,-0.1184404567,-0.0473972,0.0037349404,-0.3469017446,-0.2562147677,0.105670467,-0.1855091304,-0.0941440016,0.2317569107,-0.2538494766,0.3382097483,-0.1181777567,0.0032467169,-0.2547437251,-0.1260493845,-0.0535420962,-0.0828829706,-0.1512470245,0.0678527132,0.402818054,0.152606979,0.2217156589,0.1495414823,-0.2563756704,0.0244262554,0.1367171109,0.1140657514,-0.0522736944,0.0211516861,-0.1438651532,0.4553885758,-0.1439696699,0.2233110964,-0.3718619943,0.0244954433,-0.3538338244,-0.0992959216,0.0683534518,0.1765669286,-0.3633994162,-0.3125863075,-0.0774948969,0.3425288796,-0.325663954,0.2113520056,-0.1457198411,-0.0741149113,-0.3165704906,-0.0088636633,0.0709337443,0.3774990737,-0.0508103073,-0.1849677712,-0.2460460216,0.2898678482,0.483340919,0.450823009,-0.1895449162,0.0414339602,-0.1766331196,0.0714003295,0.1194597632,-0.161826238,-0.2467792332,0.5846036077,-0.3631800115,0.1026232988,0.2272954136,0.0897727907,0.1548352689,-0.1884676516,0.1260327101,0.5456567407,-0.0167231113,0.143302694,-0.4377728105,-0.4620153606,0.4027626514,0.0353230201,-0.2536877394,-0.1711742729,-0.0318978466,0.1660171598,0.2982515693,-0.0939555019,0.0602231808,-0.0077166804,-0.0792511404,-0.4218617976,-0.1064630449,0.2571693659,-0.3071238995,0.3156714737,-0.0298684128,0.2846216261,0.2721892595,0.1418253481,-0.1290594041,-0.1995245665,0.0606609173,-0.0541771352,0.1089612842,0.1821487248,-0.0141481729,0.0092422804,0.1536193937,-0.0705281645,0.127756387,0.0534543879,0.3099103272,-0.1456305981,0.0676016062,-0.0970410854,0.3784598112,-0.0925220177,0.0166757554,0.0030209613,-0.1428598166,0.0351032317,-0.1620597094,0.8205000162,0.1636327058,0.0556998923,0.2263664901,-0.2033575326,0.4189575613,0.4614840448,0.1375055611,-0.0583207384,-0.3883259892,0.6559360027,-0.0334870443,0.2081570923,0.2344060838,0.0253950432,-0.0108922059,0.0331262574,-0.1978801489,-0.0266735852,0.452554673,0.0287113376,-0.0495551042,0.1335161179,0.1221131906,0.337290436,0.699052453,-0.0764462352,0.080249615,0.1611669213,-0.2440985739,-0.4479449093,0.2832095325,0.3346195817,0.4011619687,0.1236729845,0.479391098,0.0982858911,0.0149974059,-0.2614249587,0.4538693726,0.0683598965,-0.3148661852,0.0691406503,0.0558245704,-0.0407564677,-0.3203515708,0.3138863742,0.0323735178,0.0395363085,-0.1711006314,0.106843479,-0.5006799698,-0.2801331282,-0.2517411411,-0.2207892835,-0.2238320708,-0.3181209564,0.0076522077,-0.0296564326,-0.1530618668,-0.0261147674,0.3023250699,0.0973892361,-0.325288713,0.198079288,-0.2525206804,-0.0247696787,-0.3617876768,0.0325186402,0.2561215758,0.1363514513,0.3704488873,-0.2708317041,-0.1832274497,-0.0131913498,-0.3252643049,0.0813810453,-0.0044514914,0.3099692166,0.086993739,0.0535680875,-0.2448784113,0.0258565545,0.0441320315,-0.3550921381,0.0461037233,-0.0059964391,-0.2256912887,-0.0581977405,-0.2201523185,-0.410479784,-0.2594747245,-0.1429173201,0.2594441175,-0.143910408,0.1373718083,0.2676988542,-0.2405003458,0.215484038,-0.0677714571,0.1764876693,-0.1195020974,-0.4075124264,0.3521413803,-0.0469389968,-0.2553837299,0.2142441422,0.1334388256,0.077306591,0.0538562052,-0.3868004382,0.3585605323,-0.1362276971,-0.04135729,0.4282892644,0.2188338786,0.1871518493,-0.1434068382,-0.0398893021,-0.0842349008,-0.3309119642,-0.1504511088,0.3599236906,0.1746121794,-0.2924568951,-0.0139145358,0.333658129,0.4254917204,0.1031931341,0.2084677368,0.1242858171,0.1685888618,-0.0807216838,0.2745329738,-0.0265402086,0.3414583206,-0.1584328264,-0.0061025694,0.3130585253,0.131487295,-0.2813426852,-0.1579692066,0.150936231,0.15226686,-0.2612183392,-0.0356048495,0.0205150861,-0.1144809797,0.2356669754,0.1436486542,-0.194329381,-0.3548479974,0.0296808947,0.1559302956,-0.0876217932,0.124720782,0.0522010364,0.1158501655,-0.6414423585,-0.1147344783,0.0603273772,0.341902554,0.182667017,-0.0428260937,0.1333038956,-0.1815695763,0.0717385039,-0.1534901708,0.2029654831,-0.0383706354,-0.291760236,-0.5185977221,-0.381261766,-0.2496387959,-0.4734687805,0.345001936,0.12635234,-0.1400874704,-0.1853709817,0.1125074998,0.165415287,0.023246374,-0.0432436392,-0.0088309757,-0.4730167091,0.0405270271,-0.1378014684,-0.2883766294,-0.1090498716,-0.027274739,-0.0055790385,0.3268518448,0.2950001955,-0.0195189584,0.0309066717,-0.014820029,0.1408236921,0.2824568748,0.0278673824,0.3156514764,-0.2493624389,0.4222184122,0.2564003766,0.1388325393,0.0171244871,0.0414955691,0.0653407201,0.6348186731,0.1393073201,0.1914890856,-0.1340216994,0.2048901618,-0.3286897242,0.0856200531,0.4198349416,-0.092277512,0.0315125659,-0.1679935306,0.3272449374,-0.0432739481,-0.0682050362,-0.019567579,-0.4723746479,-0.0389499031,-0.1813560873,0.1990232617,1.0325306654,-0.1032289937,0.3606827259,0.2462729365,-0.3267297149,0.6611087918,-0.0010683155,-0.0463623442,0.079313159,0.3475842774,0.0780697465,0.0713927075,0.1006940678,0.0853022635,-0.3384716511,0.3028821945,0.2931594849,0.2341822088,0.1250130385,0.3649465442,-0.2309813201,-0.3035395145,-0.0039909203,0.1476071179,-0.1462842375,0.438497752,0.0112073347,-0.2127481699,-0.0808171928,-0.0827344581,-0.4961404204,0.0781482384,-0.1791175902,-0.1322704554,-0.3198828995,-0.3072761595,0.1382165253,-0.3125723004,-0.1823169738,0.159379214,-0.3063563406,0.2015749216,-0.4997559786,0.3228515685,0.3548268378,-0.2214930356,-0.1207584292,-0.3069573939,-0.0201910362,-0.3673645556,0.1121911183,0.1790436059,-0.1402601898,-0.1554662585,-0.1174491942,0.4760375619,0.0256161373,0.2536866963,0.0014986837,-0.2755518854,0.1603969187,0.1830125451,-0.1511212438,0.424906075,0.0306918696,-0.2530908585,-0.3022610545,0.5446170568,-0.1712573916,-0.0762043595,0.2353501469,0.237866953,-0.2018400431,-0.2567597032,-0.0353071913,-0.2382573187,-0.2796830535,0.1312979162,-0.2013621479,0.0162847731,0.1016576439,0.2791966498,0.2899504304,-0.0184432659,-0.0281561315,-0.1501911432,0.0254226588,0.0378251262,0.0631622747,0.0463948026,-0.4493823349,-0.2314696312,0.0730350092,0.1662112176,-0.3018161058,0.0727958977,0.0534257032,-0.0370062813,0.0468185842,-0.2304476351,0.1648779362,0.0983268917,0.0627444312,-0.2889100909,-0.1000972167,-0.2096280307,-0.0944980457,0.1422278434,0.0093972636,-0.6041991711,0.4268814325,-0.12095128,-0.0704817399,-0.0491632,0.0633513406,0.1322060525,-0.1895681173,0.1917333901,0.0043710205,0.1408359557,-0.1195366308,0.1781885475,-0.0502229258,0.3720971346,0.1173602417,0.1187564209,0.5305382609,-0.0467230566,0.0031677224,-0.1742536873,-0.1299127787,0.2519765496,0.6152692437,-0.1011678055,0.2166440338,0.0312855691,0.1953369677,0.4234856963,0.0186632555,-0.1360410303,0.3578154445,0.1634841561,-0.0076250718,-0.3044868112,-0.0711654723,0.1949174106,-0.143606782,0.0686585307,0.2110058963,0.1086741686,-0.2150646299,-0.1451815814,0.2016157359,0.0481323749,0.0283961091,0.3459692895,0.039086666,0.1616563797,-0.2308831364,-0.0689180419,0.0217670742,-0.0943104476,-0.0131876199,0.2198145837,0.0897098109,0.6077306867,-0.1097400039,-0.2152640671,-0.0743065998,0.4268311858,-0.1261548549,0.2197944224,0.2022066265,-0.117249988,-0.0472153611,-0.2399027646,-0.2813337147,0.3914323449,0.0404247567,-0.3339931369,-0.1844970435,-0.0089884587,-0.3252817988,0.4570534229,-0.3683613837,0.055662889,-0.012479254,0.3678564131,0.1208803579,-0.3303735554,-0.3447899222,-0.4250264466,0.2262216508,-0.2514730394,0.0538899861,-0.1363190114,0.0155507075,-0.1111817285,0.2048872411,0.1433780342,0.2517197728,-0.1315740943,-0.1068135723,0.0552501753,0.0517467298,-0.1826298535,0.3563995361,0.1094767004,0.1614317894,0.1396194249,0.0032214553,-0.2617654502,0.050464429,-0.2575356364,-0.7597655058,-0.2148529142,0.2591416836,-0.0997466668,0.1027271003,-0.3110260665,0.1342175156,-0.1453998983,-0.0192849077,0.4423729479,-0.0146612348,0.2050583363,-0.1046234518,0.0775277615,-0.1276626438,0.6228075027,0.1438279599,0.1733986437,-0.3027555645,-0.2098262012,-0.436092943,-0.0435048342,-0.2079172283,0.1201574802,0.0933167487,0.1345236301,-0.0540953055,0.4037422538,-0.0514872521,0.6126378775,-0.0678340048,0.0476843379,-0.1062775627,-0.1528729796,0.1337752193,-0.0528776534,-0.0957044512,-0.3004838824,0.2328521013,-0.105712302,-0.098573193,-0.17859447,0.2746801376,0.0828837156,0.3282959163,0.1484969556,0.0690785348,0.3220796883,0.0548950806,-0.3768774271,-0.4569468796,-0.0288916472,-0.1553301364,0.0526762642,-0.3676460087,-0.0331748985,0.1107920185,0.2575419247,-0.3100862801,0.0315032825,-0.1456474513,0.2497520149,-0.0561202802,0.1214539334,-0.2628327906,0.2197058201,0.07602337,0.0962246507,-0.0901902467,-0.0787441954,-0.4243402481,-0.058229275,0.2731252611,-0.2771432996,-0.0997497663,-0.4166157842,0.2395465374,-0.1594294161,0.1861555427,-0.0203019995,0.0482706316,0.1561702043,0.0054629184,-0.1677879393,0.4472099245,0.2200207859,-0.2076552808,0.0784489512,0.4788645208,0.090583697,-0.2312944233,-0.1678174585,-0.1623740047]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/525","title":"wmt download speed example","comments":"I'm working on mirror links which will provide high download speed :)\r\nSee https:\/\/github.com\/huggingface\/datasets\/issues\/1892","body":"Continuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n","comment_length":14,"text":"wmt download speed example\nContinuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine.\r\n\r\n```\r\nimport nlp\r\nnlp.load_dataset('wmt16', 'de-en')\r\n```\r\nDownloads at 49.1 KB\/S\r\n\r\nWhereas \r\n```\r\npip install gdown # download from google drive\r\n!gdown https:\/\/drive.google.com\/uc?id=1iO7um-HWoNoRKDtw27YUSgyeubn9uXqj\r\n```\r\nDownloads at 127 MB\/s. (The file is a copy of wmt-en-de raw).\r\n\r\n\r\n```\r\nnlp.load_dataset('wmt16', 'ro-en')\r\n```\r\ngoes at 27 MB\/s, much faster. \r\n\r\nif we wget the same data from s3 is the same download speed, but \u00bc the file size:\r\n```\r\nwget https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/translation\/wmt_en_ro_packed_200_rand.tgz\r\n```\r\n\r\nFinally,\r\n```\r\nnlp.load_dataset('wmt19', 'zh-en')\r\n```\r\nStarts fast, but broken. (duplicate of #493 )\r\n\r\n\r\n\nI'm working on mirror links which will provide high download speed :)\r\nSee https:\/\/github.com\/huggingface\/datasets\/issues\/1892","embeddings":[-0.2278036177,-0.2449893504,0.0185677521,0.1584457606,0.0320302658,-0.1575241089,0.1905912012,0.0970176831,0.1366793662,-0.0322623625,-0.0301572215,0.4099149406,-0.1177745759,0.1829386503,0.1914406568,-0.0394568294,-0.0119318385,0.005328408,-0.6386652589,-0.2011507601,0.0932161361,0.2979103625,-0.0429115705,-0.1650554836,-0.0661923438,0.2229908854,-0.0590161271,0.0514757298,-0.4083237648,-0.2387069911,0.1231774166,0.2106235772,-0.1339845806,0.220405817,-0.0001054537,-0.0657783449,0.376950711,0.033743374,-0.5496273637,-0.2370949984,-0.4597283304,-0.5416636467,-0.1422523111,-0.0169353019,0.0744045526,0.0202442091,-0.0083170719,0.0779429823,0.1964875311,-0.0705952197,0.2312697321,0.4556207359,-0.3298271,-0.2033065706,0.4846257567,0.029638242,-0.2757789791,0.449013561,0.3262709677,0.1614215374,0.0324654579,0.4066756666,-0.2176024765,0.3635308743,0.1626876593,0.0891347975,0.1535993963,0.0596102662,-0.0476298779,0.6126980782,0.3530480564,0.1144229099,-0.5037577152,-0.3654456139,0.0321991742,-0.1781254858,0.1096832827,0.252188623,-0.3264876306,0.1029541567,-0.6757367253,0.0071950215,0.0801198035,0.2405779362,-0.0958345607,0.5124458075,-0.1917394102,0.0185796414,0.1207336262,-0.0760184973,-0.1117576584,-0.3014387488,0.0582753047,-0.0340041965,-0.0091549726,-0.147555694,0.1987005174,-0.0720231757,-0.0001358254,0.0965328813,-0.1229065284,0.2326779962,0.2343299687,0.0015982855,0.1701946557,0.3186388612,-0.292313993,-0.2117571086,0.214874804,0.0600864999,0.1816572249,0.1990931183,-0.0792862475,-0.0394480377,-0.4545413554,-0.1924396753,-0.027673291,-0.0321898051,-0.3439838886,-0.2110289335,0.1149932072,-0.1661013216,-0.1334947795,0.2426365763,-0.2476663738,0.2095007896,-0.1532128602,0.0308107994,-0.2253645211,-0.0402443446,-0.1318982542,-0.079831928,-0.1713083088,0.064422749,0.385776341,0.125925988,0.2977269888,0.0849479064,-0.1636789888,0.0525746644,0.0743273944,0.157902196,-0.0629546419,0.0580417104,-0.1229703352,0.4222218692,-0.1633229554,0.2488654256,-0.3343653083,-0.0583569817,-0.3180196881,-0.1380146891,0.0561024845,0.2109673321,-0.4015089869,-0.3108105958,-0.1036499143,0.2906546891,-0.3915573955,0.2090374976,-0.1009516791,-0.0047388463,-0.2759314477,0.0106548667,0.0637835264,0.4282468259,-0.0520715006,-0.2128608078,-0.1155939922,0.1778731495,0.3643668592,0.4724942744,-0.2038295418,-0.0417919978,-0.1704837531,0.0978999659,0.119241409,-0.147101447,-0.2474419624,0.605315268,-0.4485521019,0.0897633135,0.2088180631,0.0756400675,0.1797626317,-0.1697110832,0.1137100607,0.5087568164,0.0487930775,0.1427264512,-0.4088692963,-0.4231189787,0.2868409455,0.0688659027,-0.2364816517,-0.1682803631,0.0562717393,0.2454385012,0.2586451173,-0.1157126427,0.0539253503,0.0350077823,-0.1110686958,-0.3403765261,-0.1116258949,0.1777840406,-0.2960335314,0.3040241599,-0.0235911142,0.3127228916,0.2371614426,0.1213500202,-0.1315324306,-0.1860589832,0.0823288783,-0.0988180637,0.148211509,0.1687115282,0.0616521537,0.0360628441,0.0838806704,0.041177623,0.1131678224,0.0947726369,0.2632141709,-0.0812514871,0.0735539645,-0.052074194,0.381257385,-0.0157662071,-0.0093577057,-0.0125891641,-0.0947264731,0.0907350406,-0.1830511987,0.8025314212,0.272098124,0.0939422846,0.2580421865,-0.2719812989,0.4948805869,0.4337244928,0.1414253861,-0.043901667,-0.4614189565,0.5947166681,-0.0451983251,0.1581170559,0.280058682,0.0370074958,0.0368618742,0.0252845604,-0.2235080451,-0.1100451723,0.4200837016,0.0462405682,0.016249964,0.1825931668,0.0382542051,0.3386175334,0.667566061,-0.1506392956,0.0633299425,0.1952248067,-0.3557453752,-0.4706017971,0.2449205071,0.327067256,0.3597590327,0.2294591814,0.5250665545,0.1483656466,0.0093970234,-0.3147375882,0.3857565522,0.0697214976,-0.3105335236,0.0706083849,0.0608081892,0.0283504482,-0.3427206874,0.2494564503,0.0331769362,-0.006988246,-0.1296771914,0.0788350925,-0.4767616987,-0.2588635981,-0.2372522652,-0.1784440428,-0.2159552574,-0.3778603375,0.0095547689,-0.0134861628,-0.137324065,-0.0049512549,0.2772053182,0.0808114633,-0.2957013249,0.082334511,-0.2489618361,-0.0301616117,-0.3238693178,0.1076501161,0.2081461698,0.1659701616,0.4080638587,-0.2855162621,-0.0910396576,-0.1187173873,-0.3343368173,0.1271797568,-0.040857356,0.2249114662,0.1474662274,0.0433119982,-0.2114515156,0.0851855725,0.0726592764,-0.2664498687,0.0273105241,-0.0056126676,-0.1886042207,-0.1013278291,-0.2349665016,-0.402267009,-0.2286574095,-0.1718753576,0.347735405,-0.1114025041,0.0955403969,0.2808482647,-0.1624500006,0.2866710126,-0.1147175282,0.1812588722,-0.1842008084,-0.4297692776,0.3252176046,-0.1045261547,-0.3016571701,0.2031452805,0.1394266188,0.0589346215,0.0365041159,-0.4557619393,0.1851967275,-0.1401437372,-0.0016700282,0.4980579913,0.2480312288,0.1707411557,-0.1192915589,-0.1042258143,-0.1025302857,-0.2610042989,-0.1600720584,0.2890143096,0.1259987056,-0.3124397397,0.01922906,0.2898628712,0.4665854573,0.1380418539,0.2042038888,0.1532810181,0.112296015,0.0494045429,0.1615774632,0.006058333,0.3598295748,-0.1729433686,0.054837741,0.3241954744,0.1822603196,-0.3001067936,-0.1878323257,0.0821350887,0.0865831152,-0.2322179526,-0.0105779972,0.0602586381,-0.118030481,0.2514720559,0.101618968,-0.1213715002,-0.3142239749,0.0596020557,0.2048887163,-0.0589055791,0.0958552659,-0.0036129549,0.1284652203,-0.6818811893,-0.1018031761,-0.013808378,0.2594186962,0.1238761693,-0.0882578939,0.1322746873,-0.1943133771,0.1308400631,-0.1205150187,0.1508388072,-0.0956884921,-0.3354518116,-0.5460032821,-0.2531987429,-0.3399115205,-0.4505849481,0.3139680624,0.20353809,-0.2062508315,-0.2051536888,0.1976651102,0.13864173,0.0675811619,-0.0118779512,0.0161650199,-0.4566856921,-0.0901722237,-0.1357225329,-0.2822875679,-0.0956170261,-0.0227305423,-0.0000190935,0.2632417679,0.3073829412,-0.0230842624,0.0680411533,-0.0422005355,0.1442096382,0.2557236552,0.1089208573,0.3666503131,-0.2508639991,0.4638072848,0.2075849622,0.1192968488,-0.0114535475,0.0544860214,0.0298431423,0.6052960753,0.1604882926,0.0943573713,-0.1188786924,0.2208076864,-0.3463250995,0.1177888587,0.3901830614,-0.0476539955,-0.0062914141,-0.1977857202,0.3890604675,-0.0571213067,-0.0659734607,-0.0588164777,-0.3356775045,-0.0270456392,-0.1858955175,0.1706058979,1.0278722048,-0.060218554,0.4367638826,0.3248566687,-0.3266931772,0.6418299675,-0.071072571,-0.1140852645,0.0515699834,0.3660745919,0.0199636444,0.0860152841,0.0628612936,0.0946907923,-0.3033452332,0.2658703923,0.3495605886,0.2627804279,0.0972579196,0.4591771364,-0.2718316317,-0.3510262668,-0.0624151677,0.2142834514,-0.0555100664,0.3949986398,-0.0050089671,-0.2030993998,-0.0693379417,-0.1403748393,-0.4506738484,0.0963908657,-0.1663417518,-0.0905042514,-0.296546936,-0.2882033288,0.0998995379,-0.3189271092,-0.1619473398,0.1835816503,-0.3336291015,0.1971904337,-0.4918015599,0.2803612053,0.3689106405,-0.2524164319,-0.0521563254,-0.2918370068,-0.0449748002,-0.3346067071,0.1048029885,0.1860169321,-0.0777023733,-0.1449819654,-0.2520608306,0.3574699759,-0.0045615621,0.2694210708,-0.0711416081,-0.2473730594,0.1777150184,0.1737534851,-0.1759852469,0.3784638345,0.123722367,-0.2874700427,-0.3646201789,0.555654645,-0.2211151868,-0.0746378824,0.1888568848,0.2037159353,-0.2276465595,-0.271186024,-0.0756600276,-0.2039837688,-0.3105270863,0.1179706603,-0.1493556947,-0.002851438,0.0959103703,0.2575080395,0.2992128134,-0.0066142427,-0.0539706014,-0.160785526,0.0413149074,0.0796266347,0.0545341484,0.1091516241,-0.3637549281,-0.1627520472,0.0806313679,0.0779912472,-0.366843909,0.1193602607,0.038441129,-0.0379380211,0.1485626251,-0.2323998064,0.2769100964,0.0887587145,0.0994679108,-0.267179817,-0.2000130713,-0.258035779,-0.0852898881,0.1098863482,-0.0028103001,-0.5658128858,0.3338009715,-0.1063118875,-0.0801787004,-0.0777132809,0.1904405057,0.1456781924,-0.1856242269,0.0778659582,0.0141146882,0.0765159801,-0.103840977,0.1878687441,0.0025075017,0.3287600875,0.0459085517,0.0893998891,0.420478493,-0.0448304825,0.0523184463,-0.06698028,-0.0565757267,0.2288537472,0.6191875935,-0.1097382531,0.1721948981,0.0255373884,0.2925862074,0.3629858196,0.0282979626,-0.201383993,0.3330614865,0.2093964219,-0.0313396119,-0.2624979913,0.0301429406,0.1827001423,-0.0406459793,0.0517310277,0.2309838384,0.1488845348,-0.1897253394,-0.1042694524,0.2326614261,0.0802102312,0.0098737255,0.2750046849,0.0387028567,0.1858868152,-0.1554759592,0.0000232353,0.0554645807,0.0237765741,-0.0431482382,0.2586733103,0.0503741987,0.5822404027,-0.1120570526,-0.1585723013,-0.024453301,0.3635825217,-0.153592065,0.2110055834,0.1513534039,-0.0533913746,-0.1040357724,-0.1802143455,-0.315443635,0.4559774995,0.0405236408,-0.3467871249,-0.1222010106,-0.0654060468,-0.2915532887,0.4230563939,-0.3619393408,-0.0024486091,0.0420045368,0.2771652341,0.1799012721,-0.3586443067,-0.2854553759,-0.3778004944,0.1740759611,-0.2307738662,0.0618738048,-0.0507417731,0.0652787387,-0.0931409076,0.2732791305,0.1874116957,0.1284462214,-0.1075265184,-0.1084764972,0.0289548934,0.0404368937,-0.2206260711,0.3556927443,0.0622880161,0.0106409509,0.1830691546,0.0717911422,-0.282649368,0.0629991665,-0.2800224125,-0.7281924486,-0.2565510571,0.2272079736,-0.1281121075,0.1408048719,-0.3589399457,0.1272005737,-0.2325616777,-0.0105592432,0.4713356793,-0.0338360928,0.1761631519,-0.0890103728,0.1087087244,-0.0700897351,0.6682075262,0.2622384429,0.1787069738,-0.2517367601,-0.171428293,-0.4347377121,-0.045670297,-0.1706930101,0.1424691677,0.041926045,0.1706548184,-0.068718113,0.4308171868,-0.1109828129,0.6124202609,-0.0850846618,0.0120496778,-0.0391465612,-0.0882054418,0.062454924,-0.0615847521,-0.0893645063,-0.2551354766,0.2007423788,-0.1077137962,-0.0421427116,-0.1737249643,0.3040222526,0.0186933577,0.1620146036,0.1972310394,0.0636834577,0.3214316666,0.0415646248,-0.3932589293,-0.4192876518,-0.1037833691,-0.1198080406,0.0221333914,-0.3533702195,0.0102672148,0.1278913319,0.1947753131,-0.2015191913,0.0742885098,-0.0825213939,0.17393215,-0.0748161376,0.1018521264,-0.3297917545,0.1890606731,0.0741769448,0.2282841951,-0.0693751276,-0.0602217987,-0.3319128454,-0.1795633584,0.3032065928,-0.2977585196,-0.0802857727,-0.3585752845,0.1850948185,-0.0952108651,0.0915893093,-0.1230558231,0.1083016247,0.1734558642,0.0168662462,-0.1803049296,0.4372309744,0.2133683711,-0.220958814,-0.020436706,0.4119206965,0.0250335895,-0.2520858943,-0.1073636636,-0.1374615133]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/524","title":"Some docs are missing parameter names","comments":"Indeed, good catch!","body":"See https:\/\/huggingface.co\/nlp\/master\/package_reference\/main_classes.html#nlp.Dataset.map. I believe this is because the parameter names are enclosed in backticks in the docstrings, maybe it's an old docstring format that doesn't work with the current Sphinx version.","comment_length":3,"text":"Some docs are missing parameter names\nSee https:\/\/huggingface.co\/nlp\/master\/package_reference\/main_classes.html#nlp.Dataset.map. I believe this is because the parameter names are enclosed in backticks in the docstrings, maybe it's an old docstring format that doesn't work with the current Sphinx version.\nIndeed, good catch!","embeddings":[0.0628682002,-0.0433962457,-0.0054094824,0.0561148748,-0.0059910305,-0.0295540057,0.1999278963,0.4539961517,0.2492182106,0.0325166732,0.1219933555,0.6139661074,0.0151463421,-0.04828519,0.0946684256,-0.3679223061,-0.0413447022,0.1911993027,-0.2250861675,-0.2712671459,-0.3365822136,0.3780422211,-0.1641474366,0.0467107445,0.0763089508,0.014768146,-0.2275708467,0.089565143,-0.2776342034,-0.3085414171,-0.0840431228,0.0355368741,-0.032798931,0.2277625352,-0.0000943017,0.053817872,0.3881798387,-0.0519943088,-0.0900975019,-0.1755924672,-0.3330280483,-0.1342014223,0.0633359626,-0.3746837974,-0.2000793964,0.0020665161,0.1586944461,0.1028646082,0.1669650227,0.27784428,0.3879548311,0.3652001321,0.0805283114,-0.1907944679,0.1919032186,0.0112478295,-0.158015132,-0.0081744315,0.1677753627,0.0117605189,-0.1981260926,0.3487674594,-0.0335769951,-0.1270306855,0.0912752673,0.0228958465,0.3306500018,-0.2602231503,0.1103549525,0.0815445334,0.3750848472,-0.0407979451,-0.1712891161,-0.0712532252,-0.0176735166,0.0816727206,0.02726724,0.1386444569,-0.0373102017,-0.0029236607,0.0607418343,-0.2159264386,-0.0393716991,0.0391702875,0.1895159632,0.2502148747,-0.0674789548,0.0968678966,0.0978348702,-0.2426709086,-0.169504717,-0.0290025566,-0.1453015357,0.3370956182,-0.0330661573,-0.2480003685,0.4959839582,0.2177780718,0.1278256476,0.0936785489,-0.0444428921,-0.0160674304,-0.047792539,0.2593005896,0.3291020989,-0.0312931836,0.0479714312,0.2602094412,0.1010690778,0.0382992774,0.0025920162,-0.0064151264,0.2158514559,-0.1287643015,-0.2308710217,-0.0086847069,0.2820471823,-0.2423503995,0.0819844976,-0.1205413565,-0.0564104766,-0.0868748948,-0.0599075928,0.4010887146,0.1936464757,-0.1863579303,0.021494804,0.0554836579,-0.1838386655,-0.2315294743,-0.3607816696,0.0154758915,-0.21154131,0.036182113,0.11836081,0.0081933523,0.3796702623,0.0008453781,-0.3515478671,0.2706520557,0.0011043713,0.0933905393,0.1892453134,0.0638374463,0.0222189929,0.0691046789,0.0850324258,0.117045477,-0.2626607418,0.0885317698,-0.2630006671,-0.12629731,-0.0534425601,0.3122201562,0.1234615445,-0.0544315353,0.2737950981,0.4334540963,0.0788257793,-0.2548095584,-0.1546880454,0.0111236516,-0.2115375251,-0.082582511,0.0029456185,0.3026934862,0.0649312437,-0.2023724914,0.1770929098,0.1783075184,0.069093287,0.4216291606,-0.3640195727,0.1404823065,-0.0973162055,0.5855336189,0.2480907142,-0.1942642331,-0.2627233863,0.2184170932,-0.016579967,-0.3338091075,-0.0472718403,-0.1368414164,0.2561867237,-0.2056462914,0.2610388994,0.0909274071,-0.0123440092,0.2640562952,-0.243307054,-0.0016728654,0.1938843876,-0.0861319005,-0.1338000447,0.1523420811,0.1117983237,-0.0891390964,0.1651969701,-0.1432699561,-0.0353187695,0.4682686925,0.1101841778,-0.1352331191,0.1601189524,0.0344843306,-0.1264109612,-0.1708427072,-0.0076441052,0.5041703582,0.0981615558,0.0320689008,-0.2362539768,-0.0538687147,-0.0158204846,-0.2176575661,0.3247867823,-0.0274653416,-0.053430371,0.1699585617,-0.113807179,-0.1902680248,0.017070381,0.0124505749,-0.3312921822,-0.0177408978,-0.2616075277,-0.0271646641,0.3867134154,0.1059383824,0.3655080497,0.132095933,-0.1069639549,0.1437412202,-0.22529535,-0.0336580127,0.1385602355,-0.0694537237,-0.0570986271,-0.2589862645,0.1110754162,0.2538810074,0.1795717925,-0.1547504663,-0.1851053834,-0.0422321223,0.3821730316,0.0460618548,0.2500041723,0.1638450474,0.4482704997,-0.1731150597,-0.2488027513,-0.2248718739,-0.0038285111,0.1902564317,0.0708730668,0.1519927382,-0.2562674582,0.1361551434,0.6354427934,-0.0388662182,0.1455650926,0.0965380222,-0.1823533773,-0.2138780951,0.018907778,0.2719705999,0.3874413073,0.2988602519,0.1694372296,0.004984993,-0.2936942577,-0.1879739165,0.2797504365,-0.0404854603,-0.2513740957,-0.0267895721,0.1000496969,-0.1038120314,-0.5854811668,0.0020221653,0.0391267277,0.2726356089,-0.2894683778,-0.0249401294,-0.0286240112,-0.4095839858,0.0044695386,-0.1197503954,-0.1913067102,-0.3558319211,0.1935454309,0.0901996195,-0.2658030987,0.1966329217,-0.1987885684,0.2901074588,0.1728898734,0.2685265243,-0.2108505815,-0.3858404458,-0.2006872892,0.1579169333,-0.0837163851,-0.0566349067,0.178487584,-0.0687878504,-0.0080929631,-0.4439879656,-0.4284653962,0.2221723944,-0.2963480055,0.2433605045,0.2802430093,0.1485808045,-0.0136910649,-0.1551654786,0.2944475114,-0.3795534372,-0.1851131767,-0.0739662424,-0.129482165,-0.2353095114,-0.4094422758,-0.573145628,0.1810149401,-0.4400359392,0.263777554,0.0900903046,0.0756786168,0.5857785344,-0.2159049213,0.1823563129,-0.213693276,0.0434022248,-0.370095998,-0.3949865699,0.4569822252,-0.3954302669,-0.5962331891,-0.0275319535,-0.1372983903,0.325715512,-0.1918362081,-0.028337229,0.0750623047,-0.0949579105,-0.3480111957,0.2233451903,0.1860450208,0.250352025,0.0386431366,-0.2813532352,-0.3812724352,-0.3411063254,0.2201342881,0.4511543512,0.2440176606,-0.0910940766,0.172260046,0.0420067273,0.1752705574,-0.0486862026,0.1346620768,0.4269753098,-0.0009697472,0.3349584341,-0.0424609929,-0.2787276208,-0.024839744,0.0083691962,0.1272156984,0.365039736,0.1872544736,-0.4560109079,-0.0561935082,-0.2614572644,-0.1351989508,-0.4804874957,-0.0560083315,0.122215569,0.2794422507,0.3057087958,0.0908026099,-0.3829132318,-0.0323803909,0.2345427424,-0.0556550883,0.1493997574,0.0177200083,-0.2862962782,0.238162905,-0.6978186369,0.1285312921,-0.0987040251,0.109173134,-0.1890649498,-0.1098020747,-0.0050551626,0.1034774631,0.2256796807,-0.1032917574,-0.0013749271,0.2911800742,-0.0723464116,-0.5309328437,0.1137332097,-0.2167556286,0.290411979,0.1475860476,0.3199315071,-0.3405298591,-0.1966498494,0.3041075766,-0.272908926,-0.1176361144,-0.1070504412,-0.056596268,-0.3931778371,-0.1838053316,-0.0633544028,0.1356905252,-0.0489896648,0.0554337911,-0.0496728271,0.1472901404,-0.0605058372,-0.0547474436,0.2616469562,0.3203641474,0.0949045047,0.0333652832,0.2025404871,0.1975250691,-0.1986178458,0.3617630601,-0.1472912431,0.0574957505,0.0957698897,-0.1896416843,0.2630931139,0.1637955159,0.0131633347,0.0937188491,-0.0686819553,0.057687778,0.0099818408,-0.009923459,0.4256134033,-0.0595787168,-0.044544179,-0.3294442892,0.2152658403,-0.0387011431,-0.2017594129,0.2908856273,0.0802787691,-0.2245468497,0.2985441089,-0.0371613167,0.7064017057,-0.0130908312,0.1374562979,0.3505438566,0.0543041527,0.3327179551,0.0999648198,0.283293277,-0.4260006845,0.1154696792,0.0709054396,0.0414117239,0.0265496969,0.1329632699,-0.2965044081,0.2703728676,0.2569546402,0.1245926917,0.0807386786,0.0622727163,0.0349497795,-0.1866335422,-0.2183836848,0.3004314601,0.0121538369,-0.0146930479,0.0094238399,-0.357822597,0.133039996,-0.5363782048,0.012093164,0.1545147896,-0.0472424403,0.1293155104,0.1900507659,0.0983438715,0.0552583858,0.0519033894,-0.2251222283,-0.0856423527,-0.3513787389,0.1759295166,-0.0685012415,-0.0062787221,-0.0193726681,0.3234482408,0.3062043786,-0.0289366823,-0.0775615573,-0.0972653925,-0.1621435881,-0.160091415,-0.6762113571,0.0499671549,-0.1572362483,-0.2606279254,-0.1849138737,0.0626824945,-0.1394041479,-0.1912921667,0.2908982635,0.0977703333,-0.0685869083,0.1029531583,0.3817556798,-0.1735851616,0.0956107378,0.2950510681,-0.1177664623,0.3298107386,0.4312812984,-0.0568810292,-0.1748978794,-0.3068621159,-0.294975847,-0.0716785714,-0.1337492019,0.2489179224,-0.0342203788,0.0369397961,0.2245219201,0.2149932384,-0.2076091468,-0.142354697,-0.1892432868,-0.2330184877,-0.2516133487,0.1799746156,-0.4823427796,0.0754192099,0.242918089,0.0406243131,0.2717315853,0.0518045798,-0.4888215363,-0.0854716375,-0.2728893161,0.2689625323,0.5146235824,-0.1100345105,0.0594814196,0.023793254,0.2333922088,0.0390443727,-0.2569611073,-0.3460226357,-0.2283363342,0.0461504124,0.1952754706,0.2140471041,-0.0652123392,-0.2087744325,-0.1781663299,-0.0716409683,-0.102102384,-0.0183339436,-0.1543915421,0.4785360396,-0.0965858996,-0.1623399258,0.1956464052,0.1033246964,0.0579966493,0.0496830307,-0.0368608721,-0.0716243386,-0.3414184749,-0.0494689345,0.0254664272,0.2632015944,0.106594488,0.2901563644,0.3755254447,0.0325429924,0.1704351753,-0.0186907165,0.2338708341,-0.1688820571,-0.21524854,0.0730602294,0.1095789298,0.3673935533,-0.4389020205,-0.1706762761,0.4112866819,-0.0746603757,0.2276984006,-0.0600300282,0.0697369352,-0.1663137376,0.0118316505,-0.0557967685,0.2267986238,0.063423261,-0.0209737401,0.4840454161,-0.2329082936,0.0294100642,0.3392560482,0.0579338744,0.2325547636,0.477473557,0.2663329244,0.2024491429,-0.0766731426,0.1219920963,0.0999335498,-0.0806785449,0.0758294463,-0.0134713203,0.3640440404,0.1704614311,0.0876415074,0.2305372357,0.0931595787,-0.2066823244,-0.0995828509,0.3394984007,-0.3266471028,-0.2113064677,-0.1174526811,-0.0136164129,0.1796867251,-0.0407513566,-0.0473648794,0.0874932557,0.2167153805,-0.0428101979,0.2276267558,-0.284073323,-0.1026748717,-0.0099467058,0.0798375607,-0.27833426,0.2344743013,-0.430375427,-0.2172937542,-0.0017671336,0.2062956244,0.2022680342,0.2648886442,-0.0887944549,-0.1533178836,-0.0695641637,-0.1544545144,-0.1303304285,0.1290793866,0.1804339588,0.0676665977,0.312472403,0.1955773532,-0.1827995032,0.017275231,0.1076394841,0.1277784556,0.0841553286,0.2617289126,0.0533078611,-0.3432200253,-0.0612493418,-0.1697010547,-0.2882397473,-0.1312055439,0.371848613,0.0207056049,0.1917959005,-0.3995906115,0.181109786,0.1819731146,0.5042255521,0.0607642531,0.026332356,-0.1550440639,-0.2067809552,-0.4626918435,-0.1130296886,-0.0199178904,0.0013284191,0.1126116887,0.1656373292,0.2262961268,0.0369245037,0.1365008652,-0.066496782,-0.1879077852,-0.3055305183,-0.1954663247,-0.0910405293,-0.2693823874,-0.0683294535,0.1780770421,-0.1524440497,-0.0265657268,-0.1620629132,0.2357372642,-0.2194620818,0.109467715,0.0032236781,0.2965334952,0.1166567057,0.1024166644,0.3244428039,-0.1177698299,0.030162815,-0.0786029249,-0.2072636783,-0.0942791253,0.1347608715,0.0182611495,0.1156786978,-0.1170998812,-0.2267826498,-0.2457654029,0.0164517798,-0.0103578856,0.0477803051,-0.4082534909,0.4157607257,-0.1320361048,-0.0686884299,-0.1663314551,0.1633168459,-0.0815187097,-0.0702010393,-0.0226975102,-0.4287277758,0.697381556,-0.6235144138,-0.3743567467,0.0434880927,0.1356387585,0.0774052665,-0.175743103,-0.4462452233,-0.0781868324,0.1165619791,0.0660160556,-0.2818761468,0.1060353741,-0.0049958746,0.016674215,0.0020250892,0.2288940698,0.2931384742,-0.1358071864,0.1091111898,-0.1866802275]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/522","title":"dictionnary typo in docs","comments":"Thanks!","body":"Many places dictionary is spelled dictionnary, not sure if its on purpose or not.\r\nFixed in this pr: \r\nhttps:\/\/github.com\/huggingface\/nlp\/pull\/521 ","comment_length":1,"text":"dictionnary typo in docs\nMany places dictionary is spelled dictionnary, not sure if its on purpose or not.\r\nFixed in this pr: \r\nhttps:\/\/github.com\/huggingface\/nlp\/pull\/521 \nThanks!","embeddings":[0.3587005734,-0.3545765281,-0.2223300636,-0.2630307376,0.140349552,-0.398886025,-0.0674041659,0.212411955,-0.1955124587,0.1194985732,0.2530528009,0.296327889,-0.065179579,0.2114567757,-0.0711993277,-0.1157141179,0.1378538162,0.2619175613,-0.2184101641,-0.0921173617,-0.2020973712,0.3565544784,-0.2098657042,0.0498674735,-0.1133758575,0.1161388606,-0.0525025167,0.0468954109,-0.3057415485,-0.3307763934,0.0761769265,0.3635496199,-0.2451307476,0.1198551133,-0.0000952093,-0.2132453173,0.3693211675,-0.0809950009,-0.2063624263,-0.1515925676,0.014650682,-0.3123210371,0.1644353271,-0.1392316371,-0.3472953737,-0.0801296085,0.1261037886,0.1333982497,0.2693256736,0.1532718092,0.3806523681,-0.2222915143,0.1204043254,-0.0587888993,-0.0795299932,0.0032434354,-0.0400482751,0.0758900344,0.1288776696,-0.3612804711,0.0703516006,0.3308737278,0.0159155894,0.0817301869,0.0707374886,-0.1677332669,0.0400108062,-0.0085477792,-0.0724102333,0.1406291723,0.3047914207,-0.2547594011,-0.157816574,-0.0016170866,-0.1036706045,-0.1919446588,0.4462951124,-0.0670760125,0.1089749858,-0.0562485121,0.392126441,-0.2383960485,-0.000032787,0.1683438867,-0.081306234,0.3950176239,-0.1599489301,-0.225615114,0.0831821486,-0.0942531526,-0.2736730576,0.3813836277,0.0257066116,-0.0822668523,0.120024018,-0.0887872875,0.4061689675,0.1286649555,-0.042661842,-0.1114573851,0.0367415436,0.093777515,-0.0891250297,-0.0849263817,0.3083674908,-0.1229334846,0.5367205143,0.1444904506,0.2461080104,0.4068067968,0.0600885823,0.0785755366,0.0940218791,-0.238121286,-0.4381729364,0.0609313436,0.0667709112,-0.0492633693,0.1972508132,-0.0500917733,-0.1692086607,-0.0309881046,-0.2101082653,0.2411364615,0.038278129,0.3279162645,0.0740160421,0.0855914578,-0.2642291784,-0.2058856636,-0.2273866385,0.1758704185,-0.2015406638,-0.2029173523,0.0487663895,0.1653040349,0.246852085,0.1704600751,-0.1126245633,-0.1249997541,-0.1567352116,0.1494103372,-0.1186634526,0.1831264049,-0.0406295806,0.1166332737,0.0724483058,-0.2196998596,-0.0562823787,-0.0000055898,0.271122694,0.1028206944,-0.4017218947,0.371009022,0.118401669,0.0209003203,0.2497308552,0.1755276322,0.2323905975,-0.1794672012,0.1188362539,0.0982439891,-0.0980633572,-0.1533243805,-0.1752437502,-0.0751351193,-0.0056991102,-0.1400620788,0.0909508243,-0.0716144294,0.1829174161,0.3020287752,-0.0144054433,-0.072907418,-0.0618362688,0.206602037,0.1407347769,-0.4541842043,-0.1866090447,-0.23754403,-0.1558150798,-0.1652465016,0.00175105,-0.2202851325,0.1951823235,0.0743585825,0.2652040124,0.0967794508,0.3137081563,0.2137850225,-0.4150006175,0.0238655899,0.0378904492,-0.2268217206,-0.198802039,-0.1778355539,-0.0289106853,0.115461573,0.3011429012,-0.037438564,-0.0656037107,0.232609719,0.2657708824,0.1940825731,0.2921639681,-0.0228685513,-0.1810616553,-0.1660659909,-0.0832297355,0.3246611059,0.1117984653,-0.2199345231,-0.0558941886,-0.232590422,-0.0754393563,-0.0447475538,0.4443050027,-0.0773178488,0.2194670737,0.3036084175,-0.2616411448,-0.0133399311,-0.076384604,0.1611550301,-0.0659555718,-0.0682130009,-0.2509696186,-0.0128874416,0.0788689554,0.2272311151,0.1735945195,0.0337055773,-0.0960087851,0.1204606742,0.1225416437,0.0220873896,-0.0384328552,0.1191855893,0.1054834276,-0.0853783861,0.0528492108,0.1078932583,0.0742532536,-0.0941914394,-0.0998583809,0.2809100747,0.3096565008,-0.0879690275,0.1490744501,0.2589907944,0.4525316954,-0.0482026376,-0.1384425759,-0.4277294278,0.3056055009,-0.0331923179,-0.0149675487,-0.1056647524,-0.030044945,0.0822712258,0.4250644147,0.0327860229,0.1407711655,0.1179195419,0.0729044899,-0.1670283675,-0.0700069815,0.2628031671,-0.0776843056,0.2543607056,0.0095589114,-0.0475780219,-0.0495783426,-0.1965793073,0.1910380125,-0.0287678223,-0.0105817467,0.0323610716,-0.0147069143,-0.0597276874,-0.3590884209,0.1922037899,0.0347007439,0.0303251334,-0.2053412497,-0.290017873,-0.2483954132,-0.5460711718,0.183891803,-0.3163733482,-0.0875919387,-0.3729589283,0.2239165455,0.055394683,-0.3214116693,0.0761545897,0.408817023,-0.0527508929,-0.0846164823,0.3830052912,0.0451193489,-0.2713432014,-0.1207542941,0.2075693309,-0.0691147745,-0.1028549224,0.3539296091,-0.2171601951,-0.3885513544,-0.2643555403,-0.5480153561,0.1998875737,-0.043771401,-0.1827266067,0.4485645592,0.1917069554,0.1376475841,0.0927064568,0.2383589298,-0.0830198228,-0.1930603087,0.0863664672,0.067202203,0.1753711849,-0.3989144862,-0.3207933903,0.0526168942,-0.2788453698,0.1970179081,-0.004539052,0.1223443374,0.1139198393,-0.217316404,0.0776681304,-0.2542801499,0.2332898974,-0.1136211827,0.3815391362,-0.1034781933,-0.1452323794,-0.421354115,-0.3025188446,-0.1252241135,0.2458587885,-0.4489496946,-0.1066291779,-0.211348787,-0.1909855306,0.1034664512,0.0129044438,-0.0682421327,0.4294717014,0.0250318199,-0.3092312813,-0.1494918317,-0.1602897048,0.1953799874,0.3329664171,0.2177693099,-0.073700048,0.3767351508,0.2399127334,0.300486207,0.11997132,0.2741595209,0.2794904411,-0.0089178067,0.1620101184,-0.0586978234,0.0534331128,0.0122992024,-0.0312858783,0.0744371712,0.4062526226,0.0892488956,-0.1184165478,-0.1406348199,-0.1182560623,-0.2793468237,-0.2786314487,-0.049665615,-0.2679958642,-0.2474013567,0.1784951985,-0.1133020967,0.1499215662,0.0998854861,0.0679003447,0.1162071228,-0.2524591684,-0.0178851839,-0.0920106918,-0.097111553,-0.2810994387,0.0990464687,0.1042749211,0.0053954837,0.0477891341,-0.2302932441,0.0524196737,-0.153009668,0.5334825516,-0.136541605,-0.0380601548,0.1183671355,0.0877855048,-0.1142830774,-0.1522427499,-0.3694849014,0.351817131,0.3402408063,-0.0521717444,-0.2318309844,-0.1961204857,-0.1051916927,-0.3576316833,0.0495790504,-0.2015541792,-0.1938968152,-0.1823664457,-0.4465910494,-0.1765600294,0.3099935651,-0.1602675617,-0.2349596024,-0.2199610472,0.4754160345,-0.1492206156,0.0160707105,-0.0385668948,0.1424299031,-0.2041657567,-0.1756755859,0.1313582808,0.1315849721,-0.1828974038,0.1152895913,-0.0448753648,-0.2160423249,-0.1256476641,-0.2073916644,0.2154329866,0.0541692153,0.2568556368,0.2954950631,0.0023553011,0.2537415326,0.0791216642,-0.1273266971,0.5048317313,0.1187232807,-0.005064961,-0.2076712698,0.4079175293,0.1827067435,-0.0453181267,0.246779412,0.6372480392,-0.1674930155,0.0257304627,-0.1214112341,0.6533981562,-0.0497799069,0.1991159618,0.326208204,-0.43094787,0.6647866964,0.0519848131,0.093295306,-0.3564256132,0.3866100013,0.0531563871,0.1804052889,-0.123823382,-0.0980105326,-0.4044474363,0.2972835004,-0.4174153209,0.0785363615,0.0244962834,0.0212280545,0.0745268464,-0.0530430749,-0.0677921027,0.2886642218,-0.0790652931,0.1530763805,-0.1258597523,-0.0942815319,0.1232808754,-0.3312439919,-0.3503991961,-0.0272403993,-0.0002248676,0.0416146182,0.1012657285,0.2254308909,-0.0609300584,0.3121614158,0.1937425584,0.1495437324,-0.1617564112,-0.0787139162,0.2791174054,-0.0343762413,0.0731021166,0.2186136842,0.2879087031,-0.2690938115,-0.1359627843,-0.1328984797,0.0098965773,0.1237257347,-0.0764705017,0.1195588112,-0.0955873281,-0.094862394,-0.0209636837,0.1234704107,-0.3269657493,-0.2366624475,0.3216083944,-0.0280996542,-0.073496379,-0.0191678703,0.2231615484,-0.0780106485,-0.0029689944,0.1977265626,0.2243113369,-0.1875480115,0.1811676621,0.1492102146,-0.3577132821,-0.3238386512,-0.3995308578,-0.4052521586,-0.0174206421,0.4085692167,0.1829524487,-0.3351935744,0.062289495,0.0523298308,-0.1001646295,0.18829377,-0.2762586772,-0.0807788596,-0.1194825247,0.2098890543,-0.5044342875,-0.0852226093,0.1969268769,0.2349262089,0.1826912761,-0.0196985248,-0.4821089804,-0.0789413303,-0.0722287595,0.0603507943,-0.0679569542,-0.2283620983,-0.1917209923,0.2052067965,0.3593305051,0.044931788,0.0501377247,-0.4372414947,0.0147310626,0.0388126485,0.0878315121,0.3122624159,0.3309977055,-0.0589462183,-0.2077880502,0.0429156981,-0.0884640813,-0.0355092175,0.0882305428,0.4268828034,0.1192724183,0.0925840065,0.3773431182,0.0304833949,0.0366265811,-0.0902752131,-0.3747606575,-0.0059610731,-0.2606042624,-0.0281818993,0.0649607405,0.0397988036,-0.0645314828,0.0756227896,0.0396836586,-0.2111315131,-0.0527091734,-0.0145160388,0.3393570185,0.0669117644,-0.3110976219,-0.2300228775,0.0715501085,0.4436367452,-0.3233917952,0.0854075328,0.0109705264,0.1966338754,0.233128354,0.1413669735,0.0711833015,-0.2619955838,0.1295348257,0.0615778752,-0.3814848065,-0.3158189654,0.1005208045,-0.0769912973,-0.2061733752,-0.1606106758,-0.018278081,-0.2332965285,0.0757700354,0.4601027071,0.3002338707,-0.047670681,0.1190609112,0.1264192313,0.3010521233,0.1099580899,0.1883099675,0.3340133429,0.0173881222,0.0851220414,0.1656866223,0.2589431107,-0.0871893838,0.0889940634,-0.3799975216,0.1014832556,-0.2343044281,-0.0551607162,0.0139422696,0.0682705566,0.0699217841,-0.1866624802,-0.1210024208,0.1984088123,0.0410507917,0.0153644457,0.1347726882,-0.4416974783,0.1423007846,-0.2295614034,0.5110232234,-0.1497814655,0.2062227577,0.2159491181,-0.008260888,-0.1609378457,0.1708160192,0.4443235099,0.3519482613,-0.3813742101,-0.2036667317,-0.0766380876,-0.2080165446,0.0152836973,0.148558408,0.2521484792,-0.0071606892,0.1464250088,0.266102463,-0.3210030198,0.0390323959,0.243116051,0.153342545,0.1799675375,-0.0679010898,-0.0462379046,0.115794234,-0.1948282719,0.1297761947,-0.5846956968,-0.2928275466,0.0502817929,0.153637886,0.0701901019,-0.2074181437,0.1891690344,0.2547892034,0.3610678613,0.1872622222,0.234521538,0.2902305722,-0.286449492,-0.176219523,0.008623424,-0.0962433591,-0.0375820175,-0.0943745077,0.1014781445,0.1268817186,-0.1137221232,0.1742957234,0.072382845,-0.2082746029,-0.1433112919,-0.2704357505,0.0466555692,0.0390395634,-0.2034730017,0.1740653366,-0.0487040095,0.0167111028,-0.0453776792,0.1741844416,-0.0798861384,0.4019079208,0.1752224714,0.2872333825,-0.4105449021,0.1976094246,0.2863583267,-0.1496382803,-0.3218604326,0.0833370015,0.1479766667,-0.4299286902,0.2514698505,-0.0593686216,0.0954717845,-0.0172929522,0.0399137177,0.1453296244,0.2383128703,0.204175517,0.1324253082,-0.5074682832,0.0806016847,0.3620995283,-0.0156448279,-0.2938103676,0.180825457,-0.0796122327,-0.0165058691,-0.3276174366,-0.6123373508,0.4645321965,0.1676053405,-0.1407639086,0.0270636696,0.36031425,0.1239807531,-0.2485746443,-0.4118592143,-0.1234909669,0.0752765387,0.0962753594,-0.3423342705,0.2890592813,-0.0409025103,-0.0613061562,-0.0924196988,0.1474770606,0.217100352,-0.1061338112,0.0251950473,-0.2410076559]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/519","title":"[BUG] Metrics throwing new error on master since 0.4.0","comments":"Update - maybe this is only failing on bleu because I was not tokenizing inputs to the metric","body":"The following error occurs when passing in references of type `List[List[str]]` to metrics like bleu.\r\nWasn't happening on 0.4.0 but happening now on master.\r\n\r\n```\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 226, in compute\r\n self.add_batch(predictions=predictions, references=references)\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 242, in add_batch\r\n batch = self.info.features.encode_batch(batch)\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 527, in encode_batch\r\n encoded_batch[key] = [encode_nested_example(self[key], cast_to_python_objects(obj)) for obj in column]\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 527, in <listcomp>\r\n encoded_batch[key] = [encode_nested_example(self[key], cast_to_python_objects(obj)) for obj in column]\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 456, in encode_nested_example\r\n raise ValueError(\"Got a string but expected a list instead: '{}'\".format(obj))\r\n```","comment_length":18,"text":"[BUG] Metrics throwing new error on master since 0.4.0\nThe following error occurs when passing in references of type `List[List[str]]` to metrics like bleu.\r\nWasn't happening on 0.4.0 but happening now on master.\r\n\r\n```\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 226, in compute\r\n self.add_batch(predictions=predictions, references=references)\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 242, in add_batch\r\n batch = self.info.features.encode_batch(batch)\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 527, in encode_batch\r\n encoded_batch[key] = [encode_nested_example(self[key], cast_to_python_objects(obj)) for obj in column]\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 527, in <listcomp>\r\n encoded_batch[key] = [encode_nested_example(self[key], cast_to_python_objects(obj)) for obj in column]\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 456, in encode_nested_example\r\n raise ValueError(\"Got a string but expected a list instead: '{}'\".format(obj))\r\n```\nUpdate - maybe this is only failing on bleu because I was not tokenizing inputs to the metric","embeddings":[-0.1062732562,-0.007290211,0.0222176891,0.2218198925,0.3199834824,-0.0804266036,0.1759800762,0.3901971579,0.0343043692,0.1756983399,-0.0845413059,0.2548736334,-0.4288854301,0.0719972551,-0.028758537,-0.0816463232,-0.0013923431,0.0024744708,-0.0758355707,-0.0748338327,-0.4520992935,0.1437273473,-0.1970937401,0.0110387532,0.0288925469,0.0917715877,-0.1796934754,0.0378206633,-0.2270138264,-0.6951411366,0.0842948109,-0.1058674529,0.0286651198,0.5098218918,-0.0001164395,0.004818005,0.5615224242,-0.0969978422,-0.2757411599,-0.6296263933,-0.1194718406,-0.3785984516,0.2123195082,-0.0551587865,-0.1251197904,-0.4211822152,-0.1600915343,-0.0245703179,-0.2853537798,0.5097929239,0.1826967448,0.3170953095,-0.1970277429,-0.3702695668,0.1295918822,-0.4435131848,-0.0542089976,0.2791592181,0.0688031539,0.0893617868,-0.3565911353,0.2778955102,-0.16722022,0.097716637,0.4126311839,-0.0599433407,0.7112179995,-0.0169462506,-0.0244267806,0.0683050305,0.26930058,-0.0594066605,-0.326531589,0.1645624489,-0.0124039454,-0.5957030058,0.1381164193,-0.0336661972,-0.1020849347,-0.2063252032,-0.3794083595,-0.1562570184,-0.0071801376,0.2209166288,-0.2712134719,0.3870169818,-0.1360195428,-0.0998500958,0.5003242493,-0.0388078876,0.1800909787,-0.0616373532,-0.3171218038,-0.0552313365,-0.270742178,-0.1092254743,0.094979845,0.2227651924,-0.0316815488,-0.0119900247,0.0349352881,-0.0800457671,0.0522040986,0.1384292692,0.3684639335,0.4071957469,0.2165304273,-0.1421268731,0.2275400907,0.2707014084,0.0833954141,-0.0756214187,0.2968890667,-0.2845106125,-0.0679679215,0.3694419861,0.0975967422,-0.2575008869,-0.3548736274,0.1660371125,-0.2789395154,-0.0550797619,0.4045330286,0.4178312719,-0.3578873277,0.2668999434,0.2166235298,0.3325395584,-0.032659106,-0.0041469946,-0.1672389358,0.0001239339,-0.300743252,-0.1598049402,-0.1282812655,-0.3283575773,0.3711374402,0.1049494967,0.4938421547,-0.1525180042,-0.1196286753,-0.0951244906,0.0839676037,0.1391263902,0.0361173414,-0.1184324995,0.2583685815,-0.4067990482,0.0937797576,0.1375787258,-0.3253022134,-0.1255240887,-0.0379998982,0.1357336044,-0.1714640707,0.1331057101,-0.2497183383,0.1519280821,0.1851925105,0.0763436854,-0.2300545126,-0.2365286946,-0.2112006247,-0.27439785,0.0838536769,-0.067543067,-0.0719906315,-0.0643367991,-0.0047448212,0.2580306828,0.0987973884,0.1628796309,0.0601906888,0.1947447211,-0.1199280471,-0.4076298475,0.2281286269,-0.4222308695,-0.0846987516,0.0330415629,-0.1003820747,-0.1777200848,0.0184921157,-0.0879658461,0.1359378695,-0.2058136314,0.5759731531,-0.2368007451,-0.0427050069,-0.1251954138,-0.3218889534,-0.225340724,0.0200865585,-0.1174828857,0.4351535738,0.3503578007,0.047109846,0.4298745394,0.0101366341,-0.3278263211,-0.013037323,0.2420766652,0.3168124855,-0.288885057,0.2373013645,-0.0589065328,-0.0143643785,-0.0380503275,-0.1371418089,0.2937998474,0.0554157272,0.0134860519,-0.1908234805,0.0431649238,0.0552901067,0.2678282857,0.1259252578,-0.0428628549,-0.2989124358,0.2762337625,0.0067476421,-0.3360529542,-0.2650056481,0.0583856106,0.2053063214,0.0922112316,-0.125864923,-0.2059825063,0.1046516597,0.396045059,0.3106920719,0.0560849868,-0.1756464541,0.117861703,0.2674864233,-0.2520056665,-0.078571476,0.4476377368,0.4362250865,-0.013738309,0.0325252376,0.4991463721,0.0985601917,-0.1266162395,0.2538799942,0.4040437639,0.0748697892,0.176336363,0.0878888667,0.2562079132,0.0964245051,0.0888751,0.0462802276,-0.2134583145,0.3034289181,-0.2395786494,-0.0124711236,-0.3341283202,-0.0273536295,-0.0680245906,0.4389280677,0.1797917038,0.1811548024,0.0823640153,-0.046151381,-0.0485799685,-0.1481972486,-0.0649200678,0.3559918702,0.0830437019,0.1323611587,0.3042525649,0.0315914974,-0.2434706837,0.2299399525,0.1900543123,-0.2009668797,-0.0143520478,0.188591212,-0.1266460866,0.1589463651,0.1349481344,-0.0589856319,0.352466315,-0.386965394,-0.0252722129,-0.3195974529,0.2235095352,-0.0465928316,-0.2895771563,-0.2419312447,-0.383918792,0.0785665885,0.5633406043,-0.172150597,0.2836770415,0.2250238657,0.2812371552,0.0664489865,0.4272817075,0.2186665684,-0.3022622764,-0.1205074936,0.0455397107,0.2007766664,-0.3718587458,0.1317168921,0.0621055104,0.0519474186,-0.3125457168,-0.4706139863,0.0956862569,0.0487275198,0.5173864961,0.30605492,-0.1674262732,-0.1922731698,0.0310937576,0.2899303734,-0.0361479819,-0.0908919647,0.2110298723,-0.2417314053,-0.0862906054,-0.3249903023,-0.0833851174,-0.0758275166,-0.4211907089,-0.0845593214,0.1301460564,0.1021899655,0.1177106649,0.0031333398,0.3794730008,0.3393113315,-0.2537638843,-0.131380111,0.0278131291,-0.1283527166,-0.2453269958,-0.5491186976,-0.0731105581,0.2241434157,0.2258107215,-0.0659760088,-0.3508491516,-0.7318812013,0.3000254333,-0.0032866569,0.1589718759,0.1296351552,-0.0681936294,0.1262823641,-0.1781275868,-0.2250598371,-0.0352242924,0.1025236174,0.0331365652,0.2410041243,0.1788245291,-0.0349804163,0.3686442971,0.7059157491,0.2269937545,-0.2379549146,0.3036684394,0.1353710145,0.4072172642,-0.0264269281,-0.308899045,-0.173540771,0.1025480703,-0.1601964086,0.3309653103,-0.1620676219,-0.1369139403,-0.0541627668,-0.0518586524,0.2931030691,-0.0336522423,-0.0468551368,-0.4473406672,0.0542359501,-0.005066412,0.1837350279,-0.192781508,-0.2258315235,-0.0249956194,0.1985562146,-0.0877370015,-0.1020570993,-0.0756953657,-0.551887989,-0.2223026454,0.4270265102,0.2587658465,0.3232921362,-0.2660307884,0.189486891,-0.1109078601,-0.0832203701,0.4132605195,-0.3680095673,-0.2397209853,0.0916128904,0.0217401162,-0.1438007355,0.1161426902,-0.1778071523,0.1660716087,0.3169516623,0.3760707676,-0.0251740869,-0.2788380086,-0.0359645747,-0.0113211814,-0.0535286553,0.102918461,-0.1676890403,-0.4124310315,0.0450770073,0.292370677,-0.0652451664,0.2904182076,-0.1708579361,0.0626641512,0.0301399063,0.1009008139,-0.0013059684,-0.1361410618,0.1180525348,-0.1802487671,0.0078084944,0.0204729773,0.0319023319,0.4550202489,0.0060106646,-0.2917804122,-0.181306228,0.1308020502,-0.1310534924,0.2443943173,0.2665787041,-0.2280216664,0.1538259983,0.0599574633,0.1282739937,-0.0326259173,-0.2780759633,0.1886828691,0.1657046974,-0.2918078005,-0.2890518606,0.2140416801,0.1921145171,0.0039767325,0.3337952793,-0.145944491,-0.3020854294,0.2901753485,0.2580891848,0.7199829221,0.1772864312,0.0621616878,0.267811656,-0.0248080771,0.3369649649,-0.0011279862,0.2272084653,-0.3251259029,-0.1819219291,-0.0510066003,-0.0912292004,0.0903947651,-0.2549564838,-0.4692076743,0.3571285009,-0.3386453688,0.068428047,-0.0027007095,-0.0616370998,0.0340327062,-0.1055297554,-0.1815976053,0.1450162679,-0.0471081398,-0.0551434606,-0.1813331842,-0.1496152133,-0.1351154149,-0.2093958855,-0.2244066447,0.0030924173,-0.0500720553,0.079242371,0.092234835,-0.1890149564,0.0159309823,0.0483718254,0.6843813658,0.1389763355,-0.1221339181,-0.0896775946,-0.3120189905,0.3602885008,0.1824044734,0.0228049047,0.1903815567,-0.1521885693,-0.1710986793,-0.0263634082,0.0382207371,-0.2362463325,0.2422467172,0.1380507201,-0.5968794227,0.0490746647,-0.0457348041,-0.1342943311,0.073613219,-0.1502982974,0.126263231,0.0150179854,-0.3940709233,0.1426136196,0.0255682077,-0.1432815045,0.1711678952,0.4133885801,0.382507056,-0.0730059817,0.2826308608,0.079398334,-0.0198941529,-0.1597783864,-0.1873869449,-0.1054579467,-0.116109319,0.5724226832,-0.0795715377,-0.1791669279,0.056435205,0.5587019324,0.0625022203,0.406067431,-0.2007863373,-0.2350217998,-0.2174002826,0.1329993159,-0.3143025637,-0.0420992114,-0.3033979237,0.096973002,0.001345976,-0.0066056787,-0.2827903926,0.0291473083,-0.5816066861,0.2390802205,0.0347585939,-0.0798224658,0.2138299197,0.1546038985,-0.0672048554,-0.0371850617,-0.5296412706,-0.1155348569,-0.2056086212,0.1222660169,0.2622719109,-0.0186739955,-0.0507431105,0.0401780494,-0.0510658734,0.0489721261,0.0812328309,0.2357848138,-0.1211950853,0.1747644842,0.0279376563,0.2638652325,0.4427062273,0.1367002279,0.1742110401,0.0412107892,-0.1324889958,0.2207814157,-0.0025748478,-0.0217302255,0.0557328127,0.0362121724,0.0179564469,0.1621977687,0.1982881725,-0.2473767847,-0.0484851822,0.1497213244,0.2405904084,-0.1302976012,-0.037473809,-0.0905087143,0.0600294843,0.1330467165,-0.1818069369,-0.2732081413,0.1540732235,0.266150564,0.0918546617,0.1680219471,0.5722925067,-0.1812608242,0.4379871488,-0.0577976741,0.2492556274,-0.5429469347,0.2392246425,0.2238966227,0.2023130357,0.2100307047,0.2175001502,-0.0853962749,0.0664067194,0.3406583667,0.0033979581,0.3295253217,0.4985058308,-0.2298929989,0.0517689623,-0.2210611701,0.1668427885,0.1557105035,0.1768080443,-0.039093174,-0.2176530063,0.1408508867,-0.2812756598,-0.3672321141,-0.0805910081,0.0971923769,-0.2728293836,-0.0136384852,-0.2616145313,-0.2562116086,0.0258018933,-0.1278202385,0.1950095743,0.0901674852,0.1426076889,0.0604190268,0.2087846398,-0.1284009963,-0.7600818276,0.3565040827,0.0087326197,-0.1113043576,0.036103446,0.155718714,0.10526409,0.1129840091,0.4749213159,0.5415930152,0.4865358174,-0.184635818,-0.202948451,-0.0598557889,-0.0964690447,0.0054023042,0.3943559825,0.2869703174,-0.2051620185,0.0358298309,0.0237964131,-0.1001323089,0.6779952645,0.0031576373,-0.1738666743,-0.083685711,0.6400130391,0.1307485849,-0.1503636837,-0.2051586211,-0.2563599944,-0.3835419714,0.2625057101,0.6904423237,-0.1350450665,0.1891409606,-0.0441588461,0.0795023963,0.1670886129,0.2144993842,0.2570817173,0.1379097253,-0.419914633,0.093659997,-0.2452676445,0.1963988245,-0.020285802,0.6180791855,0.0736085549,0.0504818894,-0.0926955342,0.062262129,-0.100486815,-0.0657914355,0.1433024108,-0.0475096181,-0.6162129641,-0.0039124438,0.1456247121,-0.2014154345,-0.0295913685,-0.1896532625,-0.2828024626,-0.1986451298,0.0802242085,-0.1811144054,0.0271706153,-0.043273367,0.3224055767,0.1463052481,0.1675937176,0.5402649641,0.1617095917,-0.3007182479,-0.2196441144,-0.3226026893,-0.3023864031,-0.0873757079,0.0282489825,0.4219988286,-0.2259738892,-0.1336125731,-0.3583192825,0.8097123504,0.0864828452,0.0152092641,-0.2573971748,0.4139159024,-0.2747877836,-0.0174310524,0.1034038737,0.1995464116,0.2240753472,0.3738697469,-0.3511281013,-0.152645722,0.607324779,-0.6849967837,-0.2127241045,-0.0282142945,0.4172566533,-0.0590864643,-0.4034624398,-0.5686315298,-0.2560310364,0.2383514941,0.0717151761,-0.2924547791,0.0656686872,0.0870311037,0.1815014184,-0.058249224,-0.5435273051,0.0497679524,-0.3589709997,0.1848375052,-0.2681566477]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/519","title":"[BUG] Metrics throwing new error on master since 0.4.0","comments":"Closing - seems to be just forgetting to tokenize. And found the helpful discussion in #137 ","body":"The following error occurs when passing in references of type `List[List[str]]` to metrics like bleu.\r\nWasn't happening on 0.4.0 but happening now on master.\r\n\r\n```\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 226, in compute\r\n self.add_batch(predictions=predictions, references=references)\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 242, in add_batch\r\n batch = self.info.features.encode_batch(batch)\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 527, in encode_batch\r\n encoded_batch[key] = [encode_nested_example(self[key], cast_to_python_objects(obj)) for obj in column]\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 527, in <listcomp>\r\n encoded_batch[key] = [encode_nested_example(self[key], cast_to_python_objects(obj)) for obj in column]\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 456, in encode_nested_example\r\n raise ValueError(\"Got a string but expected a list instead: '{}'\".format(obj))\r\n```","comment_length":16,"text":"[BUG] Metrics throwing new error on master since 0.4.0\nThe following error occurs when passing in references of type `List[List[str]]` to metrics like bleu.\r\nWasn't happening on 0.4.0 but happening now on master.\r\n\r\n```\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 226, in compute\r\n self.add_batch(predictions=predictions, references=references)\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 242, in add_batch\r\n batch = self.info.features.encode_batch(batch)\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 527, in encode_batch\r\n encoded_batch[key] = [encode_nested_example(self[key], cast_to_python_objects(obj)) for obj in column]\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 527, in <listcomp>\r\n encoded_batch[key] = [encode_nested_example(self[key], cast_to_python_objects(obj)) for obj in column]\r\n File \"\/usr\/local\/lib\/python3.7\/site-packages\/nlp\/features.py\", line 456, in encode_nested_example\r\n raise ValueError(\"Got a string but expected a list instead: '{}'\".format(obj))\r\n```\nClosing - seems to be just forgetting to tokenize. And found the helpful discussion in #137 ","embeddings":[-0.1025830954,0.0307526477,0.0034684436,0.1723006815,0.2973582745,-0.0995549634,0.1647171527,0.3669822514,-0.0618152656,0.1189627945,-0.0670210794,0.3194746971,-0.378850162,0.0334373973,0.0053102658,-0.0455288813,0.0488767996,0.0500882454,-0.0283139199,-0.0415438786,-0.4338151217,0.1561397314,-0.203751877,0.067439042,-0.001488513,0.0320698358,-0.2056008428,0.0124587305,-0.2085186392,-0.7494331002,-0.0237540212,-0.0226647723,0.048785802,0.4582443833,-0.0001134856,-0.0649087876,0.5408959985,-0.0713074431,-0.2891862392,-0.570089519,-0.0161664598,-0.4225831926,0.2314783186,-0.1026178747,-0.0931410789,-0.4966783226,-0.0944120288,-0.0052618203,-0.1745480597,0.4718379974,0.2129917592,0.3863579333,-0.1674005687,-0.3329547942,0.0664207265,-0.4447818398,-0.054023169,0.263297677,0.0874250159,0.0430792719,-0.3499209583,0.2789092958,-0.1636088341,0.0674255937,0.3732295632,-0.0351286754,0.6463177204,-0.0362074375,-0.0175520871,-0.0167520549,0.26169011,-0.0798288211,-0.3455686271,0.0525409356,-0.0400219597,-0.6671746969,0.1051290482,-0.0665210485,-0.1253444254,-0.2333725393,-0.3577315807,-0.1747198701,0.0175905209,0.190661177,-0.2718409896,0.4762349129,-0.1664230078,-0.0759524927,0.3964941204,-0.0569942407,0.119490236,-0.0522871055,-0.2902086675,-0.0657731295,-0.1954233497,-0.1534586251,0.1102752239,0.1753437668,-0.0519042537,0.0104955677,-0.0231150035,-0.0207916964,0.0387965031,0.0917936191,0.384759903,0.4587340951,0.2474343926,-0.1352373213,0.2305123657,0.2567485273,0.1074836627,-0.1054218933,0.2179353088,-0.1991687268,-0.1020575389,0.3530691266,0.0771287456,-0.1486461759,-0.3942013383,0.1839554012,-0.2494340688,-0.0700132847,0.3418893516,0.4307641387,-0.3583085537,0.2780347466,0.1595157087,0.292285651,-0.0373388864,-0.0542074069,-0.1841020286,0.0287250727,-0.3161766529,-0.1576248258,-0.1969667375,-0.2060832083,0.401879847,0.0567915179,0.4840884507,-0.1494250894,-0.05778699,-0.0592000559,0.102693826,0.13774997,0.034052752,-0.0404240116,0.2528172433,-0.4652852714,0.0712756813,0.1340927184,-0.2421035618,-0.1026693806,-0.0363444276,0.1629735976,-0.2597141564,0.1191634834,-0.2604870498,0.1404055655,0.2097462565,0.1001923531,-0.2383135408,-0.2353683859,-0.1932164282,-0.276163131,0.0805804953,-0.0998494327,-0.0549946502,-0.0592913292,0.0014287456,0.2687463462,0.173439309,0.2330849916,0.0622671619,0.2976830602,-0.1386229992,-0.2723981738,0.2804914713,-0.3891468346,-0.0451328531,0.0075003677,-0.1507262141,-0.1308654696,-0.0063750376,-0.1555448472,0.2743073106,-0.2414910197,0.6387081742,-0.2098978907,-0.0234917328,-0.0919859931,-0.3614304662,-0.1873827428,0.0221896563,-0.1704645604,0.3997805715,0.2793895602,-0.000480628,0.4381350577,0.0573236793,-0.3086840212,0.0102445967,0.1429016441,0.3598019183,-0.2083673924,0.2066050768,0.0043909783,0.0301526766,-0.0903610364,-0.1430296451,0.2886415124,0.0005858635,-0.053415779,-0.1537871659,0.0622516945,0.0656035319,0.2067133784,0.1687725186,-0.065676406,-0.3247308433,0.2614595294,-0.0777015537,-0.3801072538,-0.2696731985,0.0930509493,0.1923234165,0.0451431498,-0.1129506677,-0.1805782765,0.1130102277,0.4350059927,0.3111210167,-0.0321298465,-0.14475362,0.1643628627,0.2113064826,-0.2326555401,-0.0815256387,0.4597387016,0.4157818556,0.060364753,0.0278821774,0.478833288,0.0670261011,-0.1000906602,0.2037340403,0.4495797753,0.1506353319,0.1320396811,0.1151866093,0.2695251405,0.1095170975,0.1239702478,-0.0123747867,-0.2585503757,0.244444862,-0.2514766157,-0.0599756688,-0.3223254383,-0.0368457474,-0.0281212442,0.4343467355,0.1494999677,0.1949959546,0.0555024892,-0.0568944216,-0.0893000662,-0.0794319138,-0.1411330253,0.3704999983,0.1090402976,0.1751407087,0.2766388953,0.0335929655,-0.2884184122,0.2643554807,0.2360635847,-0.2386018187,-0.0717694387,0.2106545269,-0.0873401836,0.1041121036,0.1218275428,-0.0150989862,0.3402748704,-0.3700836301,-0.0135867521,-0.3047521412,0.1540881842,-0.0464738682,-0.2822912931,-0.2647284567,-0.4586750865,0.1253719479,0.5365670919,-0.2035598457,0.3060360253,0.2197273076,0.3430376649,-0.0040900237,0.4481122792,0.2011852264,-0.273815304,-0.1592060626,0.0607995242,0.2161760479,-0.4423589408,0.1926897168,0.023108067,0.0449754335,-0.3382922113,-0.5410550833,0.1346272379,0.0187823512,0.5171115994,0.3091272712,-0.1555643529,-0.2226472646,0.0494303964,0.2618091106,-0.0502664559,-0.07128869,0.1240149289,-0.2022914439,-0.0897509232,-0.330417186,-0.1338534504,-0.0362834632,-0.4213833809,-0.0046855952,0.1267534345,0.1458474845,0.162497744,-0.0917065516,0.3000639975,0.2218308449,-0.2381519675,-0.1674543917,0.0321863256,-0.1012241617,-0.2741954029,-0.526902914,-0.0320391245,0.1731408089,0.277420342,-0.141188547,-0.266263932,-0.7050465345,0.2729977667,0.0149008073,0.1656219214,0.145554617,-0.0736351758,0.122906372,-0.2186596841,-0.1856951863,-0.0137087023,0.0625974834,0.0388252959,0.2608799934,0.1186240166,-0.0265427232,0.3920225203,0.7582386136,0.207714498,-0.2172785401,0.2250673324,0.089022249,0.4650042057,-0.0440357476,-0.243343547,-0.1683036089,0.0978556052,-0.1787171066,0.410587877,-0.1375833005,-0.1326861084,0.0024825288,-0.0907583311,0.2721269727,-0.1075215936,-0.0025520641,-0.4442729354,0.0620432198,0.0377654433,0.2141904235,-0.1999805272,-0.2447486669,-0.073861517,0.1722290218,-0.0853310674,-0.1359672844,-0.1858635694,-0.5356898904,-0.2991837561,0.4534092546,0.1427959502,0.4406175911,-0.191241771,0.1218818128,-0.0776239932,-0.075408116,0.4426033795,-0.2954182327,-0.3221557438,0.0678027421,0.0504134968,-0.1277242154,0.0715707093,-0.18364577,0.2224704921,0.3943437636,0.4560619295,-0.0618279614,-0.368231535,-0.0891581103,-0.0150082773,-0.0613364391,0.1047880724,-0.1913542598,-0.4135424793,0.0008691555,0.3236663043,-0.0232242998,0.2896872759,-0.1676885635,0.0767333359,0.0855688378,-0.0248729717,0.0633710325,-0.1107609123,0.0915858001,-0.1242247075,-0.0088207275,0.0352127813,0.0975973383,0.4031941295,-0.0614821911,-0.2436503172,-0.1201555878,0.1217566878,-0.0821399465,0.3428159356,0.3620158136,-0.2411073595,0.1716371477,-0.0573204607,0.2196196765,-0.0515444912,-0.1894413382,0.1967514455,0.1235135123,-0.329777658,-0.2535516322,0.1763489544,0.1899455488,-0.0582579412,0.3350798786,-0.173222512,-0.3365857899,0.3295777142,0.2607652247,0.7838844061,0.158783868,0.1453652233,0.2241971195,0.0076919533,0.4496666491,-0.0851001218,0.2854069471,-0.3297614753,-0.1069127247,-0.0175761934,-0.028336877,0.0765313879,-0.2744090259,-0.4212642312,0.3958389759,-0.3130154312,0.1119685322,0.0476605929,-0.0193150453,0.0017608975,-0.1346183568,-0.1763059795,0.1434059888,-0.0509915613,-0.0638151094,-0.1985662282,-0.1533885449,-0.0887423605,-0.2491589636,-0.2029888034,-0.046508234,0.000816797,0.0500569604,0.1021752059,-0.1325033903,0.0101590771,-0.0453761071,0.6511393189,0.2024534345,-0.0967987552,-0.0861033872,-0.3375155926,0.3422828913,0.1619880795,0.0377236642,0.2105181962,-0.1257282048,-0.0671116859,0.0216029584,0.0543297827,-0.1683906764,0.1029615477,0.1180802882,-0.4745998383,-0.0123661896,0.0147312535,-0.1293436289,0.022709582,-0.1549342722,0.1451229304,0.0443018898,-0.4271368384,0.2010450959,-0.008381553,-0.1178950965,0.137421295,0.3740793169,0.3873314857,-0.1296368986,0.3213075995,0.0294983014,-0.0242383797,-0.1934430301,-0.1587308943,-0.1021229178,-0.0283009652,0.5267846584,-0.0159637537,-0.1582407653,0.0393529199,0.6108918786,0.0679978654,0.3737491965,-0.2151808143,-0.2402611822,-0.2137254626,0.1524187028,-0.3421464264,0.0357887596,-0.2687770724,0.1598456651,-0.0483096205,-0.0675327852,-0.3256031871,-0.0053279209,-0.5822659135,0.2281893492,0.0019555746,-0.0447117947,0.1777951866,0.1672752202,-0.0269951951,0.0087781474,-0.5163072944,-0.1283080727,-0.1641362309,0.1332288831,0.2568814754,0.0015648699,-0.0337989889,0.0868801251,-0.0358155556,0.0137622543,0.0513418317,0.2249253243,-0.1380153,0.1345108598,0.0531318039,0.3093575239,0.4541187584,0.0168514308,0.2114289105,0.1103707626,-0.1310707033,0.2171839327,-0.049100697,-0.0213714913,0.0826736614,0.0668933764,0.0013303271,0.207570672,0.1935934573,-0.2006028593,-0.1441448927,0.1552373916,0.228674382,0.0008656189,-0.057990741,-0.1737775654,0.1117397994,0.1524030119,-0.1537358761,-0.2560847402,0.082869418,0.2873782516,0.0435165353,0.2376539856,0.64144665,-0.1512835473,0.3893934488,-0.0593384653,0.3311495483,-0.5056810975,0.2554982603,0.1255850792,0.1752373278,0.2068281472,0.2748238444,-0.1083727553,0.0928033739,0.3665814698,-0.0058930591,0.4253592491,0.5238871574,-0.1939542443,0.0416306928,-0.2242094576,0.1587412357,0.1086469144,0.2729485929,-0.0232509114,-0.2108722627,0.0915459245,-0.2629575133,-0.3100191057,-0.1211332157,0.1296381354,-0.3028589785,-0.0189107526,-0.2903178334,-0.2848831415,0.0099470709,-0.1718783975,0.1598358005,0.0589966848,0.0714218765,-0.0249303244,0.193334952,-0.0902368501,-0.6897425056,0.2785591781,0.047724057,-0.1806768328,0.0971988291,0.0878421292,0.102454409,0.0077937664,0.4235361516,0.5571835041,0.5265235305,-0.1513025463,-0.2116869241,-0.1038864106,-0.0789821371,0.0074872235,0.4405223429,0.2474873513,-0.2089954913,0.0488182306,0.0464874916,-0.1254559457,0.6335610151,0.0618402623,-0.2150242329,-0.0572171025,0.6917786598,0.1169635728,-0.1868626177,-0.1524780989,-0.2617040873,-0.4149160087,0.2753247321,0.7080239654,-0.0909690633,0.1992948949,-0.1455474645,0.1045816466,0.1784872562,0.2287333608,0.2525074184,0.1226940379,-0.4430888891,0.0635415688,-0.2749469876,0.2051919997,-0.0078023714,0.5662417412,0.0571242794,0.0373441726,-0.0386761539,0.0657408834,-0.1301491708,-0.0293788575,0.0640195385,-0.1114696637,-0.5865784287,0.0228971429,0.1776908189,-0.1765088737,-0.018786218,-0.1661435217,-0.2150130421,-0.2838037312,0.0908251479,-0.2027358264,0.036149893,0.0451818183,0.3070928156,0.1492668688,0.2019896656,0.4724553823,0.1372263283,-0.262591213,-0.1933424026,-0.2765476406,-0.3023959994,-0.0688934177,0.0031622718,0.3741799295,-0.2008823007,-0.1248794794,-0.3668100834,0.6668751836,0.082066454,0.021716956,-0.2482246608,0.4603601396,-0.3204938173,-0.0269623417,0.0846239552,0.2409331501,0.1729374826,0.3077320457,-0.3895661533,-0.2265737653,0.6256468892,-0.7567941546,-0.3048583269,-0.0389306583,0.4279524684,-0.0262921378,-0.3801739812,-0.5356045365,-0.2704559863,0.2905351222,0.0874241814,-0.3660974205,0.038787879,0.0568868294,0.127649188,-0.0671716928,-0.4462124109,-0.0022067823,-0.3939641416,0.1510462314,-0.3153620064]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/517","title":"add MLDoc dataset","comments":"Any updates on this?","body":"Hi,\r\n\r\nI am recommending that someone add MLDoc, a multilingual news topic classification dataset.\r\n\r\n- Here's a link to the Github: https:\/\/github.com\/facebookresearch\/MLDoc\r\n- and the paper: http:\/\/www.lrec-conf.org\/proceedings\/lrec2018\/pdf\/658.pdf\r\n\r\nLooks like the dataset contains news stories in multiple languages that can be classified into four hierarchical groups: CCAT (Corporate\/Industrial), ECAT (Economics), GCAT (Government\/Social) and MCAT (Markets). There are 13 languages: Dutch, French, German, Chinese, Japanese, Russian, Portuguese, Spanish, Latin American Spanish, Italian, Danish, Norwegian, and Swedish","comment_length":4,"text":"add MLDoc dataset\nHi,\r\n\r\nI am recommending that someone add MLDoc, a multilingual news topic classification dataset.\r\n\r\n- Here's a link to the Github: https:\/\/github.com\/facebookresearch\/MLDoc\r\n- and the paper: http:\/\/www.lrec-conf.org\/proceedings\/lrec2018\/pdf\/658.pdf\r\n\r\nLooks like the dataset contains news stories in multiple languages that can be classified into four hierarchical groups: CCAT (Corporate\/Industrial), ECAT (Economics), GCAT (Government\/Social) and MCAT (Markets). There are 13 languages: Dutch, French, German, Chinese, Japanese, Russian, Portuguese, Spanish, Latin American Spanish, Italian, Danish, Norwegian, and Swedish\nAny updates on this?","embeddings":[-0.3161039352,-0.0893218145,-0.1245369315,0.214871347,-0.107491456,0.1644279957,0.194639951,-0.087495625,-0.3016334772,-0.1355899125,0.009924151,0.1497566998,-0.3895456195,0.1229820102,0.0908919126,-0.3055920303,0.0582694449,-0.2608481646,0.0298086572,-0.1544269174,-0.3026458919,0.230475992,-0.1727487147,-0.285104692,-0.2030512691,-0.0418056697,-0.4262661338,0.0079488996,-0.0299918652,-0.364109844,0.0089092972,0.0911869854,0.0826076567,0.2956672907,-0.0001156171,-0.3821992278,0.0775891766,-0.0406755209,0.0948487148,-0.3528190255,-0.2796533108,-0.5650494099,-0.0812690482,-0.0793565884,-0.2362531424,0.0122220237,0.026979249,-0.5049071908,-0.0939009041,0.5147284865,0.1726725996,0.0839323997,0.2308064103,0.0049986602,0.3583601713,0.1408225,-0.1425605714,0.6420926452,0.4798176885,-0.0604548603,0.0168123432,0.2931489944,0.0421966575,-0.2560436428,-0.2170831561,-0.0284569729,0.381862402,-0.501193583,0.1867676675,0.3797077239,0.6146844029,-0.0419261716,0.0085203145,0.0057386472,0.2317281663,-0.1519537121,-0.2325933427,0.4703545272,-0.1052659899,0.0667393953,0.2554001808,-0.4361554682,-0.1804952472,0.226756528,-0.0226617008,0.671412468,-0.0351593085,0.0550902598,0.0852232501,0.1282985657,0.2609470189,-0.1687908471,0.0168185532,0.1735041291,-0.0038143203,-0.147827208,0.1127892956,0.0940269828,-0.0451913141,-0.5917299986,-0.235865429,-0.0109092817,0.0110553838,0.1982402205,0.1455835402,-0.1086364537,-0.0592280068,-0.1474559158,0.0702522919,-0.7701852918,0.0110461786,0.0256299861,-0.0082955472,0.1740553975,-0.3149286509,0.2394474,0.0864422545,0.0434206538,-0.0782243907,-0.036794316,-0.2303143889,-0.2392066568,-0.0330223292,-0.0967061073,0.2632063031,0.2904962599,-0.2941029072,0.4184141755,-0.0505080298,-0.2463105023,0.0626778826,0.3465834558,-0.4992511272,0.1855049431,0.1302926689,0.3321517408,-0.150966078,0.0019153988,-0.1573825926,-0.0108601218,0.0073115323,0.0017641741,0.0674608648,-0.1892834306,-0.2082765847,-0.1419896185,0.1662976295,0.1365501136,-0.3003061414,0.179481104,-0.2951506972,-0.1200680658,-0.4736914337,0.0589620508,0.3256321549,-0.2470995039,-0.0215181801,0.8453760743,0.1973769367,-0.0543208048,0.188983649,0.0476840809,-0.380841434,-0.1075221449,-0.3781074584,0.3273080289,-0.343932569,0.1803003103,-0.4260459542,0.2609502673,0.4254829288,-0.1539027244,-0.3010749519,-0.0665272847,-0.0464957133,0.1353734881,-0.01507298,-0.3035830557,-0.0384749919,-0.05943726,-0.1453637481,-0.1665870547,0.3075957298,0.389929086,-0.1793880612,-0.0141258994,-0.3963181973,0.0599814169,-0.2059464604,0.0849484205,-0.165683493,0.0729985535,0.4850760102,0.201497063,0.2406222671,-0.1482796222,0.4237134457,-0.094915539,0.2769164741,-0.0061897631,-0.0065056114,0.030554682,0.3268515468,-0.2674290538,0.0986035615,-0.1658446193,-0.4614525735,-0.1292781383,0.1482153982,0.4656696022,0.0604783185,-0.1954339147,0.0505195148,-0.1745963991,-0.0820847973,-0.2763258219,0.0971261188,0.1257524937,-0.0883501247,0.3386182189,-0.0342889093,-0.1686956733,-0.1286529005,-0.1933522671,-0.0445879512,0.375148803,-0.0341820158,0.1758744717,0.1766217947,0.404379487,0.0938167274,-0.1556471884,0.0129879247,0.0501605533,-0.1438335329,0.077091895,0.4314051569,0.3916873038,0.3239655495,-0.6390002966,0.2010955364,-0.070761621,-0.0024600863,-0.1207505316,-0.0984818041,0.4895502329,0.2501735389,-0.0187818781,-0.181460157,-0.0729531497,0.2061074525,0.1915643513,0.1190116405,-0.1220583841,0.2413107455,0.2145991474,0.2902033031,-0.1197289303,-0.3318169415,0.2389162779,0.2643521428,0.175537914,0.3217074871,0.1474555582,-0.2005413473,-0.1058891341,0.1603130996,0.3257518709,0.4092631638,0.2118619531,0.2023309767,-0.0715541691,0.2099483907,-0.1586280763,0.0675414056,-0.2297553718,0.0300469808,0.059542235,0.3063521683,-0.0639320612,-0.3512089849,0.062524654,0.2352854162,0.072668843,0.1900253147,0.1033867523,-0.0185034834,-0.338891536,-0.2298248261,-0.0730810165,-0.4730736911,-0.1141252667,0.1308929473,-0.159829542,0.2420696616,0.1111933962,0.083578147,0.640763104,-0.2147794962,0.305125922,0.1394613683,-0.3233281076,-0.0682101771,0.115930751,0.5082378983,-0.0749257132,0.3822109401,0.0137335509,0.015730435,-0.120505482,-0.4053385854,0.0882741883,-0.1581764072,-0.0384416953,-0.2868129015,-0.3743537962,-0.112620078,-0.1910167038,-0.0253328104,0.2829367518,0.0647037104,0.0033893187,-0.1799674481,-0.1633210033,0.0157219507,-0.4868681431,-0.6651721597,-0.1253872663,0.094665952,-0.4282265604,0.039020244,-0.1049834639,-0.2204387188,-0.0104438001,-0.1392693967,-0.1619330347,-0.0431790315,-0.1534897983,0.1165015474,-0.3612948656,-0.2658743858,-0.1386337429,-0.0454623587,0.3293384612,-0.1388217807,-0.3296888471,0.0221587569,0.0404297002,0.3258478343,0.2121469378,-0.1123452932,0.462041676,0.0930000469,-0.0612702966,0.1632756889,0.217102468,-0.051813826,0.177677229,0.3534026742,-0.125522837,0.141625315,0.0429778844,0.6912635565,-0.2343098372,-0.2613500953,0.3329302669,0.0194754712,0.3425001502,-0.075519301,-0.5011281371,0.1568206847,-0.1730667055,0.034949284,0.3226716518,0.1199663654,-0.5944864154,-0.157412529,0.0397807211,-0.1590979397,-0.0414499789,0.3013286293,-0.1439544261,0.4696543515,-0.0229035802,-0.4530291855,-0.1752995104,-0.5101118088,-0.0881998688,0.4122723341,-0.0594952442,-0.2304231077,-0.3894881606,-0.3638186157,-0.4789706767,0.1596678197,0.3044630885,-0.0218189545,-0.266164273,0.2572978437,-0.0067450409,0.0422039516,0.0497872271,-0.2639575303,-0.1723994464,0.2928647995,0.2116479725,0.395342648,0.0958597362,-0.2447618544,0.5235510468,0.3451586962,-0.0004440844,0.0253913682,0.0386365019,0.1010094956,-0.0744101182,-0.0554908961,-0.2960104346,0.165689975,0.3303752244,0.1773315966,0.1852131039,0.0097000999,-0.0583237559,-0.0833136365,-0.0180910099,0.3147129714,-0.2639171779,0.2121656537,-0.0325664319,-0.294737041,0.4083919525,-0.1198516041,-0.3297536969,0.1513791084,0.6235747337,0.2800120413,-0.2948048115,-0.2377313673,0.1123452783,-0.6080301404,0.6139953136,0.104289189,0.2320532352,0.1332834363,-0.0831372812,-0.2133960724,-0.0936293826,0.1459916234,0.1380143911,0.2048009485,-0.3589913845,-0.0805861354,0.4891933501,-0.1749478281,-0.0832789168,0.0736307502,0.0458414182,-0.1333594471,0.0220078994,-0.0773705393,1.0203342438,-0.027741462,-0.0243967827,-0.2225931436,-0.0184968505,0.4742819369,-0.5000018477,0.1094364747,0.000319174,0.3724808395,-0.092702046,0.1415572166,0.0847819895,0.3037308753,-0.1524434984,-0.1611654609,-0.1422030628,-0.4079784453,-0.1738759875,0.3820859194,0.2963570356,-0.1577022374,0.0380823091,0.0782410502,0.2242311239,-0.0880518556,-0.1604145169,-0.0196751747,0.2165701687,0.0187603608,-0.6114597917,-0.1172697321,-0.1171946675,0.0620078184,-0.4068563581,-0.4923069179,0.5100695491,0.1960898638,0.1159565896,-0.1482759416,-0.0833958834,-0.0696772411,0.1068154201,-0.1484178901,0.053054437,-0.3272459805,0.1321708262,-0.1233905107,-0.1433402449,0.391320616,0.0964204744,-0.3590767682,-0.2523808181,0.1072655171,0.4264556766,0.2977100611,0.1206474304,0.5238121152,-0.0642690361,-0.0226505455,0.1405712664,0.1584642231,-0.0598184429,-0.0308504142,-0.0566266477,-0.0236375891,-0.1803921461,0.1119929403,0.3372855186,0.0049513881,-0.0230812151,0.4470150471,-0.1572818905,-0.2920029759,0.3027466536,0.4329997301,-0.216969192,-0.3101950288,0.1183742136,-0.0909729376,-0.0092925187,0.0530085117,0.0262459703,-0.0910780206,-0.0474470742,-0.3197403252,0.0955983177,-0.1870209128,-0.2502359748,-0.1092918292,-0.1190732494,-0.0182713903,0.2193838358,0.5964455605,-0.2485315353,0.1027370393,0.0482021905,0.2361399233,0.4438053668,0.3064292669,0.1618046314,0.019932108,0.0146322949,-0.265053153,-0.1053809598,-0.1213192865,0.0971526057,0.1713612974,0.0713539422,0.0907989889,0.2676156163,-0.2629383504,-0.090688765,0.1967404485,-0.228774026,-0.069117479,-0.1597829908,0.2473544031,0.0444684811,0.0614271723,0.1779852659,0.2258719355,-0.016533833,0.3224899173,0.2143685818,0.3924399316,0.3542926013,0.0629471913,-0.2362529635,0.0326483883,-0.0684087425,-0.1090294123,0.097483255,0.2035556436,0.1560212523,0.2140376866,-0.0106847892,-0.2989867628,-0.1563165486,0.2392697781,0.1400107294,0.1323050857,-0.2825158238,-0.2303553671,0.0142130321,0.0033129596,0.1898577064,0.1890708804,0.5553250313,0.2151902467,0.2427761108,-0.163411811,0.3031857908,-0.052987285,0.1493083984,-0.2699905932,0.0616217554,0.1540008336,0.3009645343,0.109637849,0.1064473391,0.0362210646,0.176335752,0.4727225006,0.7046511173,-0.0300186146,0.1792778671,0.2635942101,0.4060668349,-0.0264790393,-0.0523511358,0.1175435409,0.1069884002,0.4829010665,-0.1459024251,-0.2232303321,0.0191114545,0.2593222558,0.0465453267,0.1243097708,0.1833484024,-0.1657890081,-0.2516209781,-0.0643500537,-0.0994573832,-0.1479308009,0.362319231,0.1877649277,-0.0497136861,0.0332476385,-0.1004197374,0.0958929658,0.1379150897,-0.1904796064,0.2643565238,-0.1586328596,-0.1244641542,0.0676679835,0.1825357527,0.0496900268,0.2114148289,-0.1839973778,0.0114792688,-0.2317993641,-0.0224277135,0.0708922595,0.1379727572,0.2204980701,0.2781646848,0.0646416619,0.0681328624,0.0754290223,0.2724816203,0.492637217,-0.1717422158,-0.1487839222,-0.3506985605,-0.0309094787,-0.0343429223,-0.4125754535,0.1749332845,-0.1814156622,0.0154436007,0.4532209039,-0.2591049373,-0.091868557,-0.1093944535,0.0361168981,0.0328455828,0.2338091135,0.2012881935,0.1852917075,-0.2891471684,-0.3444236219,-0.4017688334,0.0424986817,0.0432531834,0.0552963316,-0.3497747183,0.1036265045,0.3693175614,0.404561758,-0.169424966,-0.0747127607,0.0349067599,-0.0158190243,-0.5527945757,0.10617771,-0.2762549818,0.2607685626,-0.1199430749,0.2440975755,-0.0901693776,0.0380189903,-0.0315703601,-0.4046221077,-0.2596617937,-0.0601879098,0.4208044112,0.2802884877,0.0456372611,0.681664288,-0.0919639915,-0.190205425,0.0536521412,-0.1369406581,-0.0668199807,0.1436765641,0.2179225534,-0.0200488977,-0.4761185646,-0.394182831,0.2012684792,0.2407066673,-0.1699722409,-0.1474769861,-0.3377133906,0.0605628565,0.2438518107,-0.0211728644,0.0461079888,0.1928720772,0.1557125449,-0.1179665476,0.0537492298,0.0603771284,-0.0618129335,-0.2463539839,-0.0809222981,-0.1850476861,0.1449114531,0.0076224911,0.1643137634,-0.395676136,-0.3705459237,0.351051867,0.0713356212,0.0749224722,0.1792922318,0.161991924,-0.3336823881,-0.1534081846,-0.2520061731,-0.0690076053,-0.1168251261,-0.0603220239,-0.4609406292]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/517","title":"add MLDoc dataset","comments":"This request is still an open issue waiting to be addressed by any community member, @GuillemGSubies.","body":"Hi,\r\n\r\nI am recommending that someone add MLDoc, a multilingual news topic classification dataset.\r\n\r\n- Here's a link to the Github: https:\/\/github.com\/facebookresearch\/MLDoc\r\n- and the paper: http:\/\/www.lrec-conf.org\/proceedings\/lrec2018\/pdf\/658.pdf\r\n\r\nLooks like the dataset contains news stories in multiple languages that can be classified into four hierarchical groups: CCAT (Corporate\/Industrial), ECAT (Economics), GCAT (Government\/Social) and MCAT (Markets). There are 13 languages: Dutch, French, German, Chinese, Japanese, Russian, Portuguese, Spanish, Latin American Spanish, Italian, Danish, Norwegian, and Swedish","comment_length":16,"text":"add MLDoc dataset\nHi,\r\n\r\nI am recommending that someone add MLDoc, a multilingual news topic classification dataset.\r\n\r\n- Here's a link to the Github: https:\/\/github.com\/facebookresearch\/MLDoc\r\n- and the paper: http:\/\/www.lrec-conf.org\/proceedings\/lrec2018\/pdf\/658.pdf\r\n\r\nLooks like the dataset contains news stories in multiple languages that can be classified into four hierarchical groups: CCAT (Corporate\/Industrial), ECAT (Economics), GCAT (Government\/Social) and MCAT (Markets). There are 13 languages: Dutch, French, German, Chinese, Japanese, Russian, Portuguese, Spanish, Latin American Spanish, Italian, Danish, Norwegian, and Swedish\nThis request is still an open issue waiting to be addressed by any community member, @GuillemGSubies.","embeddings":[-0.3504601717,0.0476948284,-0.179826811,0.1780563295,-0.0401054062,0.0907025114,0.2102559358,0.0325836353,-0.2079457641,-0.0521601848,-0.0914575011,0.2889948189,-0.3860351443,0.2393020391,0.00817692,-0.275159061,0.0485755689,-0.1197368801,-0.0523765422,-0.1522308141,-0.2527983487,0.2394448966,-0.0691751987,-0.3075847328,-0.1811061352,-0.0353651531,-0.3153686821,0.0952790231,-0.0668420419,-0.2999739647,-0.0394034944,0.0626558885,0.0979469121,0.2279056162,-0.0001061797,-0.2723769248,-0.0213431939,-0.0787593275,-0.0399520546,-0.342012465,-0.192526713,-0.5444474816,-0.1018266603,-0.0767772272,-0.3466023803,-0.049074471,0.0083702337,-0.3613840938,0.0605379716,0.5007581711,0.2565388381,0.0859355852,0.102427505,-0.0976428464,0.2727009058,0.1102413088,-0.224194482,0.4792916179,0.3978053033,-0.0129565382,-0.1251174808,0.3853281736,0.0855682343,-0.1745478362,-0.235900104,-0.1055123433,0.3037681878,-0.4347405136,0.2402412146,0.4121541083,0.4717878997,-0.0578594469,-0.0302899815,0.0187110994,0.2011547238,-0.1232533678,-0.1241777465,0.42480281,-0.1183524132,0.2057818323,0.2912897766,-0.2708719373,-0.1782761812,0.2673748434,0.0062592584,0.5904614925,-0.0948563814,-0.0398360714,0.0286296066,0.0630985871,0.089431785,-0.0559796952,0.0216120202,0.0299717579,-0.0069941855,-0.1323748678,0.1900782585,0.0338908285,-0.0118479012,-0.5495747328,-0.2824370265,0.1264005452,0.0002330515,0.2171889991,0.1047805846,-0.1002691984,-0.0895598084,-0.1077378467,0.0900662839,-0.6283724904,-0.041423589,0.134491533,-0.0186324716,0.2106272131,-0.2073094696,0.2827202082,0.0501628965,0.0374758169,-0.1497849524,0.02024775,-0.0945357606,-0.1005709767,-0.0626533106,-0.0025815703,0.2395416796,0.3365658522,-0.2852636874,0.4120237529,-0.0436523892,-0.2059766799,-0.012076308,0.2986410558,-0.4672502875,0.2174862772,0.2671689987,0.3142031133,-0.1296132654,0.0681928843,-0.1805099845,-0.0437338948,-0.0008034438,-0.0104817133,0.0942154899,-0.2559864819,-0.1792082936,-0.1735662818,0.1549897939,0.0668152347,-0.3196915686,0.1501157582,-0.2282118052,-0.1416861713,-0.4731708467,0.1578960866,0.2616533041,-0.2436589897,0.0642135367,0.863278985,0.0825070366,0.0251896027,0.1984636635,0.1562655717,-0.3449949324,-0.0244769491,-0.3151106834,0.3379444182,-0.2713032961,0.0985440686,-0.3621907532,0.1460315883,0.4985351861,-0.1491980255,-0.2632514834,-0.0646698102,-0.0307394043,0.1373834014,0.0982732922,-0.1820759624,-0.1346080452,-0.063598603,-0.274834305,-0.1773222387,0.3722852767,0.3560859561,-0.0993723348,0.0530079119,-0.2976567447,0.197338596,-0.1643756181,0.1864760369,-0.2438937128,0.0223280117,0.3751322925,0.2322484851,0.1834988594,-0.1514345556,0.3337680697,-0.1110692397,0.1882600337,-0.0859177709,0.0978523567,0.0397928469,0.3091090024,-0.2474258095,0.0880564377,-0.1876769513,-0.4251469076,-0.195760712,0.0680512935,0.5277488828,0.0006937525,-0.249694109,-0.1992037147,-0.1998839229,-0.0813732296,-0.2588812709,0.2315973639,0.2079891711,-0.1649474651,0.3441869617,0.044437509,-0.1457084119,-0.1748246998,-0.1963101923,0.0520573445,0.3023568094,-0.1225667968,0.1410601437,0.2456605583,0.3988873661,0.0367090739,-0.0794375688,-0.0863954872,-0.0437920429,-0.1423499584,0.1276963502,0.460562408,0.3530477285,0.2352763414,-0.5769765377,0.052005183,0.0045328485,-0.0322566405,-0.0442243777,-0.1316456348,0.5089737177,0.2918946445,-0.0439195074,-0.1299797744,0.0080994666,0.2802174091,0.1063909233,0.1076645255,-0.2521215975,0.1961372048,0.2229685634,0.223180145,0.0003108688,-0.2700794339,0.3575141728,0.4288192689,0.2195136547,0.4198083878,0.1980836242,-0.1720623076,-0.1587088257,0.1949118972,0.3316400349,0.3206742406,0.2881241739,0.2182106227,-0.1101256385,0.1997352391,-0.2270747423,0.0574108139,-0.2346522808,0.0194468722,0.1496636868,0.2977226675,-0.0787324831,-0.3853931725,-0.0284576844,0.168272227,0.1194262505,0.2478954494,0.0757955685,0.0107711926,-0.3028209805,-0.1325542033,-0.0371492356,-0.3319294155,-0.0881230086,0.2266663462,-0.1962447464,0.0135379657,0.1910297722,0.073874101,0.6659824252,-0.2408099025,0.3134451509,0.0946907178,-0.3747076094,-0.1084236056,0.1872622818,0.3665884137,0.0572728589,0.4213939607,-0.0802005529,0.0244263876,-0.1391052902,-0.3946224153,0.1376084685,-0.1326623261,0.0045509716,-0.2764537632,-0.2783454955,-0.2011666447,-0.2879025936,0.0679926872,0.2422076911,0.0488183051,0.096886687,-0.2544697225,-0.0422844253,-0.0170372613,-0.5840312243,-0.5971542597,-0.3213184774,0.1237818822,-0.4896151721,0.0652241856,-0.1523710787,-0.1793899387,-0.0761272013,-0.1064795256,0.0808062255,-0.1429606378,-0.1013497114,0.1082436442,-0.4424519837,-0.2948853374,-0.228261739,-0.1236577481,0.3014537692,-0.113021642,-0.3047885299,-0.0105106989,0.0000253713,0.2592225373,0.2193441689,-0.1279331297,0.3854506314,0.179912746,-0.2204765528,0.1150018796,0.1998904645,-0.0299933851,0.1024783179,0.1932315528,-0.1513638794,0.144139275,0.2045750618,0.6741436124,-0.2040857524,-0.2150021046,0.3979616165,0.0647444651,0.2752732038,-0.1986930966,-0.4880429208,0.0617456213,-0.146840483,0.1192880273,0.4349211454,0.1452074647,-0.6007018685,-0.1493844539,-0.0440626852,-0.237251699,-0.1192613468,0.4779384136,-0.1192407832,0.489343822,-0.0604410693,-0.4806213081,-0.1466921717,-0.463752687,-0.0524486266,0.3658185899,-0.0979318544,-0.2127047181,-0.3793241382,-0.2370933294,-0.4760767817,0.165868178,0.2960683703,-0.0070886873,-0.2773671746,0.1327594221,-0.0535997413,0.1378530115,-0.0471147671,-0.2517941296,-0.2112855613,0.2253705859,0.1360321641,0.2857651412,0.1675660312,-0.2783945501,0.3581987023,0.3128788173,0.0537659712,-0.1441449821,0.0092230309,0.10487739,-0.1199764609,-0.0164659284,-0.2579379976,0.2546908855,0.2651934326,0.04053149,0.206028536,0.077180177,0.0131417215,-0.0574987046,-0.079552412,0.2969717383,-0.1456534564,0.1984928399,-0.0233265497,-0.2609379888,0.4214571714,-0.0801086202,-0.4275275767,0.0767887384,0.4764680862,0.3444794714,-0.3277983963,-0.3228880167,0.0872002766,-0.6639154553,0.4696894884,0.1570418477,0.2391200513,0.2240027934,-0.0443972349,-0.1164847612,-0.1286984235,0.1328663975,-0.0400876813,0.1218543798,-0.2763405144,-0.0844313577,0.3336800337,-0.178390041,-0.1271011531,0.046332702,0.0429407209,-0.1024370864,0.1454384923,-0.1845743805,0.9991689324,0.0421384498,-0.0390668623,-0.1620026678,0.0415903069,0.3682968616,-0.4186359048,0.0810637474,0.0249310564,0.2886128128,-0.0106032062,0.2014366388,0.0217226706,0.4167997539,-0.259411335,-0.1328703314,-0.0949497968,-0.3697191179,-0.1294359267,0.3930044472,0.1445161402,-0.3100063205,-0.0702935159,0.151393801,0.1585782319,-0.0640837625,-0.0977445617,-0.0784077719,0.1131888106,-0.0263166185,-0.5851589441,-0.1633251309,-0.140444994,-0.0010684883,-0.3791236579,-0.4034354389,0.4643145502,0.1088788211,0.1197958812,-0.0816847011,-0.0747471377,-0.048137933,0.0761646405,-0.0824099705,0.1130241081,-0.2577242553,0.1517159641,-0.1179750711,-0.1456616372,0.3262199461,0.0533518866,-0.2974798679,-0.1739798188,-0.0022872791,0.3571199477,0.1794328988,0.20657368,0.4325196445,-0.1237897351,-0.1778423488,0.2551514208,0.0903268605,-0.0567391701,-0.0667492971,0.0631639063,-0.090490818,-0.2605319023,0.1119602248,0.358950913,0.0481433943,-0.0543599129,0.3596835136,-0.1761045158,-0.3980855048,0.2156861722,0.470317632,-0.3415983319,-0.2643133104,0.0983725041,-0.2792688012,0.0947731361,0.1272611916,0.0722855628,-0.076164335,-0.0338224731,-0.2948744893,0.0781944245,-0.1039999202,-0.1449818015,-0.0626107156,-0.1448184252,-0.0697873309,0.1207150444,0.5719767809,-0.3960479498,0.0132710887,-0.1148025766,0.3371840119,0.3385542631,0.1901922077,0.1771303415,0.1263930351,0.0809032544,-0.2876845896,-0.155265674,-0.2474000752,0.074627988,0.126308918,0.0042050579,0.1544843763,0.2634314299,-0.1682382971,-0.1326878369,0.0997382924,-0.0896566063,-0.0885651633,-0.2006301135,0.2484868467,-0.0242495127,0.1026362032,0.1923003048,0.0701182336,-0.0605509281,0.4121595919,0.1102798805,0.3135128319,0.2025724202,0.040665105,-0.2264906764,0.04808091,0.1162045449,-0.0264299251,0.019506922,0.312618345,0.0909894109,0.2852868438,-0.0029834234,-0.3067893684,-0.1718092859,0.2622524798,0.1899763197,0.2933300436,-0.2764184177,-0.1160557345,-0.01232439,-0.1363573968,0.3736110926,0.2231891453,0.5580360293,0.0558356605,0.3407676518,-0.128191635,0.212619707,-0.1000775769,0.1407826394,-0.2668406963,0.0731989369,0.1450913399,0.2389808148,0.0224512797,0.1655884832,0.0073102326,0.1722318381,0.449250102,0.5488793254,0.0749434456,0.1833158433,0.257196039,0.4761044681,-0.0145536661,-0.1138030961,0.105837591,0.1011008173,0.5873667598,-0.0996633619,-0.2136489153,0.0050924099,0.2641492486,0.0209781863,0.1223194227,0.1888942868,-0.1931831241,-0.2949824929,-0.1077043563,-0.0582860112,-0.0591637939,0.323507458,0.1896043271,-0.0172165036,-0.0435145609,-0.1825181991,0.0981011018,0.1939062178,-0.2571991384,0.2715384662,-0.0609120615,-0.1142501086,0.0420161262,0.2911749184,0.0636654347,0.1521029621,-0.1244268119,0.0080018304,-0.2051228583,-0.0687011778,-0.0833025798,0.1853264421,0.222143814,0.1677642316,0.0907906219,0.1673061103,-0.0428928286,0.1316726208,0.4447266459,-0.1616387963,-0.1065653265,-0.2647453845,0.056903027,0.0685199127,-0.3800063729,0.0482381955,-0.1828311533,-0.01349435,0.3872528672,-0.234101072,-0.0111235566,-0.2185673416,0.1019932255,0.0592904091,0.2190640718,0.1313597113,0.1556180865,-0.2932222486,-0.332588464,-0.3342855275,0.1244103238,0.0433671661,0.1419538409,-0.2438903749,0.0919462368,0.3528184891,0.4291576743,-0.2348553687,-0.0953566208,-0.0262033083,0.0503740795,-0.6079154611,0.1129190102,-0.1610679179,0.1781863421,-0.0642761216,0.0717877299,-0.0605343431,0.0821429789,0.0979240835,-0.3729559779,-0.2880815566,-0.001777434,0.3894966543,0.3264068961,0.0986554846,0.5398358703,-0.1891202629,-0.243223086,0.0601159371,-0.2482290566,-0.0344025679,0.0884416327,0.2808527052,0.0784208998,-0.4055997729,-0.3906467259,0.236811772,0.1843078583,-0.1191082895,-0.1633129865,-0.3069269657,0.1217953935,0.2452245951,-0.0172460973,-0.0886313096,0.2369781733,0.1804031134,-0.2136698812,-0.0545984693,-0.1283068806,-0.0078332387,-0.1371266097,-0.1735511869,-0.1268012971,0.2104526609,0.0174021572,0.1700179279,-0.2431251258,-0.2548590004,0.2900339663,0.1148306131,0.0799134448,0.25780949,0.1249188334,-0.2490222901,-0.1825180799,-0.224145025,-0.0429618433,-0.1331420839,-0.0737081319,-0.4327379763]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/514","title":"dataset.shuffle(keep_in_memory=True) is never allowed","comments":"This seems to be fixed in #513 for the filter function, replacing `cache_file_name` with `indices_cache_file_name` in the assert. Although not for the `map()` function @thomwolf ","body":"As of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.","comment_length":25,"text":"dataset.shuffle(keep_in_memory=True) is never allowed\nAs of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.\nThis seems to be fixed in #513 for the filter function, replacing `cache_file_name` with `indices_cache_file_name` in the assert. Although not for the `map()` function @thomwolf ","embeddings":[-0.1573929638,0.1101285964,-0.0568150766,0.1280068755,0.1705864519,-0.0684821829,-0.1548946053,0.2861289978,0.0338808484,0.2117180824,0.0658880919,0.5071401596,-0.3157268465,-0.4159291685,-0.0771163255,0.1456125528,0.0106629906,0.0505587161,0.0162348952,-0.1306301504,-0.3281734884,-0.137692526,-0.4366667271,-0.2426392883,-0.2989773154,-0.1640271246,0.0052831536,0.0470123179,-0.227496326,-0.4783442318,0.0199873708,0.3617694974,-0.4792946875,0.4746551812,-0.0001162967,-0.2271992266,0.2232827097,-0.0738838613,-0.2772014737,-0.0804563761,-0.3944146037,0.1710918993,-0.0294310097,-0.1011725441,0.2029477805,-0.1530759931,0.0347536691,-0.4851988554,0.2687033415,0.2786893845,0.1599706262,0.0159433503,0.0001701377,0.000820215,0.4894811511,0.1501083225,-0.1416635364,0.1507811248,0.1773031503,-0.2589056492,0.0207065959,0.2994833589,-0.2331489921,0.2211959064,0.2024488896,-0.2042768002,-0.2059482038,-0.2288378179,0.3310019374,0.2132777274,0.2408331186,-0.3785223365,-0.301278621,-0.1053988859,-0.037107069,-0.2631696463,0.3107367754,0.2316165417,-0.0896453559,-0.1278472841,-0.2277664691,-0.1627272516,0.1554199457,-0.0969653353,0.1520405263,0.3311278522,0.0854251981,0.0750735775,0.5438088775,-0.0434678234,0.1729149818,-0.11182715,-0.2220365405,0.3620202541,-0.400971055,-0.0684148818,0.1164387539,0.1214043126,0.2701940835,0.0775753558,0.2510643899,-0.0182149708,0.0698503405,0.2189764977,0.1172196269,0.3911765814,-0.0454112142,0.2291889191,0.202336967,-0.0665106401,-0.1798820794,0.0643799081,0.0984606296,-0.2280791253,0.0362020805,-0.1349883974,0.0791311935,-0.2575393319,-0.149314329,0.1455168575,-0.2016399056,-0.2749112844,0.1139304414,0.5352529883,0.21457313,-0.0850824714,0.0486496724,0.0245817658,-0.0372562222,0.1303484738,-0.3592341542,0.1180868447,-0.2858389318,0.0910925716,0.3312159777,0.0103797177,0.4083402455,-0.0821994916,-0.0767392218,0.1427711993,0.3175205588,-0.2982518971,0.4927133918,0.3478182256,-0.2853671014,0.0245101303,0.0233668201,0.085998103,-0.1818951964,0.4057045877,-0.1047065631,-0.315454632,0.3535188735,0.1046302915,-0.0984297544,-0.0372052342,0.0105707468,-0.0325796157,0.4916000664,-0.3913598657,0.3386237919,-0.2468027771,-0.0318350978,-0.3389356732,0.075642556,0.1905239671,-0.2994280159,0.0468248092,-0.3101512492,0.0093933139,0.1475354284,0.4033971429,-0.0408287086,-0.3672423959,-0.2317583263,0.1196080968,0.2779319584,-0.1551210284,-0.5335749388,-0.224050954,-0.0394641496,0.1237245053,0.1481851935,-0.0203214381,-0.0424621589,-0.1730810106,0.5085964203,0.0379993469,-0.0126274452,0.1150694937,-0.2760385275,-0.0769601688,0.3453753591,-0.1851587147,-0.2285199165,-0.1318752319,0.1136803031,-0.1366752088,0.3384538293,0.0672309846,-0.0678052083,-0.0848844573,0.5236661434,-0.2864767909,-0.1019460484,-0.095645085,-0.0910898,0.1867520064,0.1332996786,-0.1971724033,-0.0885401368,-0.2106115222,-0.0276733283,-0.0563779585,-0.1452712417,-0.3837918043,0.0689062327,0.101848647,0.1232281849,0.0783526897,-0.165017724,0.1610037535,-0.4800157547,-0.1713840216,-0.0109556094,0.14752312,-0.0071874508,-0.2062905282,0.0499242544,0.1624589413,-0.0845243782,0.0634023994,-0.3319532871,0.3970722556,0.2695005834,0.0723539144,0.0935901105,0.1672176123,-0.0943982229,-0.1139782593,0.2701926231,-0.1726678461,0.1205780655,-0.0815213993,-0.1144955158,0.2229023427,-0.2499862313,0.0974670276,0.2538131475,-0.1576067507,0.329996109,0.2417398542,-0.1405291259,-0.2411245406,0.359174639,0.1679524779,0.3795506656,0.2267789543,-0.2383547425,-0.0708889663,0.6992047429,0.2837002873,0.1699571311,0.0535838865,0.031355869,-0.3145792782,0.2185159177,0.4217031598,0.5327516198,0.1507366598,0.1366884261,-0.0458083674,-0.0098620364,-0.0731765032,0.3802345097,0.0488995537,0.0775541738,0.1109744608,0.2988072932,-0.1855213046,-0.5816820264,-0.1974030137,0.3076075017,0.1843895614,-0.1164078265,-0.1214051768,-0.3387998044,-0.13702178,-0.0874949917,-0.1034934446,-0.2300516516,-0.1256727278,0.1022123322,0.4080298841,-0.1954042614,0.3857203126,-0.3336474895,0.500220716,-0.0102241859,-0.1942370832,-0.2881383002,-0.2212348133,0.133469522,0.0746022686,0.1877979338,0.1455854475,0.3903840184,0.3359338641,-0.1787646562,-0.0312285833,-0.2958876193,0.06916558,-0.1598305106,0.1165891439,0.0000164303,0.0856649429,-0.0761800557,-0.379935354,-0.0507458635,-0.3272202909,-0.2472828925,0.1441958845,0.1873338372,-0.162285164,-0.3767653406,-0.3460093439,-0.2578058541,-0.4631391168,0.3171484768,-0.0848922208,0.1214853004,0.4202855825,0.0141147096,0.0793087855,-0.4065115452,0.1262601018,-0.4711131752,-0.3016932607,0.1513073444,-0.1811847687,-0.0848974362,-0.0582177415,-0.1400675029,0.2091389298,0.389939636,-0.1939363629,-0.0617895871,-0.4313779175,-0.1898041517,0.0714063868,0.1114423648,0.4826931059,0.1092969105,-0.1300193667,-0.2072770298,-0.1935292035,0.2210785598,0.3850803375,0.1657591313,0.0406268761,0.066813983,0.3262483478,0.6937778592,0.0598500744,0.2929897308,0.2368310243,0.1813685,0.3728828132,-0.3469637632,-0.1049738377,0.361243099,0.1402894109,-0.3288646042,0.0948680192,-0.2270585895,-0.3511334062,-0.0471122749,0.2294057012,0.0816844329,-0.1761454642,0.0910460502,-0.1346196681,0.0280176885,0.2118985355,0.0978678837,-0.2013153583,-0.1443096995,-0.174539268,-0.061266765,0.2783504128,-0.1113697961,-0.5371747017,-0.3369067311,-0.4133474231,0.3385955393,0.073298201,0.7409533858,-0.1150697395,-0.2570548058,0.1130611226,0.0311744157,0.5424150229,-0.1315325648,0.0703382045,0.2860825658,0.1088634953,-0.1734209955,0.0428576767,-0.3285349607,0.1347732842,-0.2887087464,-0.2275044024,-0.2670069337,-0.1154419035,-0.0717204586,0.0076806662,0.0275267642,0.1575459689,-0.3131513,0.118960917,-0.3537300527,0.2048887461,0.2223889679,-0.1562041044,-0.4568429291,0.0531481765,0.0876473039,-0.0443667918,-0.0011116486,0.183996737,0.6027074456,0.066846177,0.0654246882,0.2121333778,-0.3203272521,0.2301210016,0.3122932315,-0.1689057648,0.2302590013,-0.0336303264,-0.0269165579,0.3270866871,0.4371746182,-0.0666524917,0.2152834535,0.0755508691,-0.0700166523,-0.1579749137,-0.0572544448,0.2989966571,-0.1464774758,0.0034275136,-0.6915032864,0.2409238964,0.4322825968,-0.1686564684,0.8573561311,0.0784562752,-0.1668886244,0.3656103611,-0.0356802121,1.1166884899,0.1091565937,0.0970483273,0.1437730193,0.0438765064,0.0899275541,-0.0661742389,0.2910059988,-0.0725260079,-0.5402422547,0.1124229282,-0.3357086182,0.2155232728,-0.1074574292,-0.3066191077,0.2703738213,0.1088065058,-0.0759075209,-0.0636948496,0.0278354455,-0.3591059446,0.0095768431,-0.0428428352,0.0767821148,-0.0392099544,-0.0215328578,-0.0597066544,0.0223249774,0.0490411483,-0.2968662381,-0.1549022794,0.0383190513,-0.2098947316,0.1892452389,-0.245846957,-0.0086617349,0.0586981028,-0.1383514702,-0.1002738029,0.2575943768,-0.0381109156,0.4269349277,0.5315418243,0.2869336009,0.0922337025,0.2054480165,0.3232618868,0.049744077,-0.1855177134,0.0617486946,-0.1750717461,-0.521566987,-0.0883001387,0.0091216173,0.210127756,0.2104575634,-0.3420522213,0.246695742,-0.3974885345,-0.2043136805,0.1140341386,0.0952135026,-0.1253778785,0.4969868958,0.0045316797,-0.2676779032,-0.0105162375,0.4765641391,0.0281229261,0.2652397156,0.4883902371,-0.1547877192,-0.1877841055,-0.2324952483,-0.3721441925,-0.0979964659,-0.1788605899,0.238790676,-0.1320291907,-0.3678796887,0.1020920426,0.214132309,0.0809845552,0.4203977287,-0.153085649,-0.3298025429,-0.3210723102,0.4822471738,-0.1216563061,0.4010073543,-0.2554948032,0.1152670383,-0.1312909722,0.2543548644,-0.3306411207,0.0438495576,-0.086695306,0.2975469232,-0.0985952541,-0.0719029084,0.0443642139,-0.1761741787,0.0815104693,0.3535404205,-0.3803093731,-0.1887716502,0.12812455,0.1591945589,0.0792163387,0.0319469236,-0.1188534871,-0.5915091634,-0.0735405236,-0.2292191684,0.1576597691,0.1029230505,-0.0979311019,0.2407464236,0.3533526659,-0.0703480616,0.3893805444,0.2689373195,-0.1864769161,-0.4986565113,0.004750384,0.0578373633,-0.185445711,-0.1271495521,-0.1210962459,-0.0699628368,-0.3378582001,0.2191348523,0.0398983769,0.0895158425,-0.261916995,-0.1121973917,0.352163434,0.6336433887,-0.0485649817,-0.2793470621,0.3481562734,0.1860110611,-0.3188645542,-0.2030694634,0.0778351277,0.3197860718,0.201632902,-0.0069373567,-0.0450205468,-0.219085291,-0.0726040229,0.3098425865,0.2485373169,-0.596152246,0.2809727192,0.1141276285,0.0162179787,-0.150506705,0.1792109013,0.2326972038,0.3400190771,0.3597915173,0.0004890343,0.3931155205,0.4816125631,0.0684742704,0.087767899,-0.269335568,0.1169808134,0.5212513208,-0.0955215096,0.3057990372,0.2569494843,0.246852681,0.0996061191,-0.2025586665,-0.2280268967,-0.0831031054,-0.1025747135,0.0028625061,0.1434672028,-0.3137084842,0.0395571925,-0.1121045873,-0.0068380348,0.1534907967,0.6984449625,0.1048739851,0.0598904938,-0.3550285399,0.2080316097,0.1177922562,0.2147826701,-0.2703762949,0.1793623865,0.1657558084,-0.0415789224,0.3765405416,0.2736275196,0.2136911005,0.3082119823,-0.1811732054,0.1040918306,-0.0340609439,-0.1599441916,-0.1944903284,0.3913168311,0.3522192538,-0.1744027883,0.1435544789,0.1338281631,-0.1224542707,0.0188088119,0.1264186651,0.1370453238,-0.289732188,0.2310371399,0.2666769922,-0.0323876627,0.1337470859,-0.072519958,-0.4232139289,0.029578086,-0.0952586085,-0.2903544009,0.0419534966,0.0068691368,0.0609536991,-0.0908786654,0.2384780198,0.1523755789,-0.1010504588,-0.1375350952,-0.2638489902,-0.8945609331,0.1401766092,0.1064961553,0.1123239473,-0.0922296643,0.389256984,0.0441299863,-0.1401147395,0.0965879932,-0.1807502657,-0.0234658197,0.1317146271,-0.4472528994,0.026688775,-0.0620255098,-0.0675666407,0.1009897068,-0.2585665882,0.0665585622,-0.1518154144,0.0168735925,0.0684262812,-0.1055905744,0.1257440299,0.4385501444,0.3916816413,0.0588670708,0.3196925223,-0.0921977907,0.1568846107,-0.3073195815,-0.2492433786,0.0389408991,-0.3124742806,0.0258383173,0.5218631625,-0.2367622703,-0.1749369055,-0.2347989976,0.3725399077,-0.2896844149,0.0679834783,-0.3461996019,0.2034951895,-0.0127021214,0.4027408361,-0.3385981321,0.0093156183,-0.0445306599,0.1489071995,-0.2840517461,-0.3808189929,0.8212559223,-0.3352939785,-0.2349312603,-0.1738761812,0.5242074728,0.0427517109,-0.4556315839,-0.2740296125,-0.0482049696,0.1213063896,-0.0840213075,-0.2470399737,0.0919980034,-0.3285838366,0.1826313585,-0.0695369169,0.5346601605,-0.1046647429,-0.2599052489,0.0749412253,-0.1796701401]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/514","title":"dataset.shuffle(keep_in_memory=True) is never allowed","comments":"Maybe I'm a bit tired but I fail to see the issue here.\r\n\r\nSince `cache_file_name` is `None` by default, if you set `keep_in_memory` to `True`, the assert should pass, no?","body":"As of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.","comment_length":30,"text":"dataset.shuffle(keep_in_memory=True) is never allowed\nAs of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.\nMaybe I'm a bit tired but I fail to see the issue here.\r\n\r\nSince `cache_file_name` is `None` by default, if you set `keep_in_memory` to `True`, the assert should pass, no?","embeddings":[-0.0392353348,0.0279586129,-0.0443971194,0.2410495281,0.1197215095,-0.0215087403,-0.2454705983,0.2770644426,-0.0534341335,0.2102242708,0.2128321975,0.374445051,-0.2902280986,-0.566421926,-0.0628235266,0.2306990027,0.0152336266,0.0731291249,0.1307362616,-0.1656005234,-0.409574151,-0.131580174,-0.524050951,-0.1571469009,-0.428753376,-0.1573809236,-0.0444145277,0.117156662,-0.2388194799,-0.5017880797,0.0678522363,0.2654946446,-0.4624954164,0.4916145802,-0.0001249819,-0.3131682873,0.2609167993,-0.0434021205,-0.1833150089,-0.0600501485,-0.4701530933,0.3102053702,-0.1049670875,-0.1583370715,0.32724455,-0.17662099,0.0549891926,-0.4825304449,0.17302531,0.2994125187,0.0882786587,-0.0590316318,-0.0274076667,0.0266639832,0.3734092712,0.0743080527,-0.0033938757,0.2121803761,0.0776693746,-0.116405569,0.1800208092,0.1766789705,-0.3082759678,0.228798002,0.1727747917,-0.2034926265,-0.2115484774,-0.1800227016,0.3078238368,0.2102493942,0.1988276988,-0.3202574551,-0.1345006824,0.0092537589,-0.0782134682,-0.202175796,0.3165804446,0.286834687,-0.0235549174,-0.052101884,-0.1797533333,-0.2429484725,0.106021978,-0.0864027515,0.0385790765,0.2909682989,0.032297194,0.2232920974,0.510096848,0.1254252195,0.0449493974,-0.1143685654,-0.2260143906,0.33785671,-0.4495311081,-0.0888596252,0.1479341537,0.1570460945,0.3375706971,0.0762877688,0.1617085636,-0.0072407457,0.0201300737,0.1431003213,0.1532425284,0.4152951539,0.0578295439,0.1727067232,0.1973043829,-0.0983939022,-0.1155526266,0.1021902636,0.0477520376,-0.2676830292,0.0527646616,-0.0756467879,0.2449690402,-0.2947060764,-0.1255384982,0.1726624072,-0.3767850399,-0.3416075706,0.0408682935,0.5163072348,0.258212328,-0.1329903305,-0.0419685505,0.0868414119,0.0162561126,0.0719358474,-0.3517907858,0.175907433,-0.2351176739,0.0944050401,0.2889090776,-0.055646874,0.2828926444,-0.1205164939,-0.0619889721,0.1201131865,0.2415632904,-0.3670422733,0.5084925294,0.29857409,-0.3389808834,0.0358220562,0.0677113086,0.1633681059,-0.1894767731,0.4758026898,-0.0540290587,-0.3840192854,0.4273476005,0.0068682996,-0.2204317451,0.0482815169,0.0162271895,-0.0989364684,0.3985984921,-0.4084577858,0.4869121611,-0.1724495292,0.0578270145,-0.3036775291,0.069000423,0.1707658321,-0.2531927526,0.0957381651,-0.2667095661,0.0407610163,0.0872286707,0.3894156516,0.0169782303,-0.2980735004,-0.3070945144,0.1243021786,0.1682263315,-0.2866532803,-0.450091064,-0.2724568248,-0.0869327709,0.0789938644,0.2291227877,-0.0798648521,-0.1744347662,-0.1958563924,0.4720093608,-0.0606687814,0.0313487612,-0.013874433,-0.272058934,-0.0306105837,0.3783480227,-0.2080346793,-0.2740474045,-0.1138512269,0.0983329192,-0.1043224782,0.3244960308,0.0963459611,-0.1518213451,-0.1861663908,0.4907175899,-0.3960990906,-0.1179611385,-0.0278684553,-0.0874459073,0.2058205307,0.157132417,-0.1732318848,-0.1034319848,-0.1785058528,-0.0762262717,-0.0115153044,-0.1123682559,-0.3707544506,-0.0385702662,0.1194324344,0.1440198272,0.1694183052,-0.1760571897,0.1061042249,-0.5986754298,-0.2244273424,0.0427490026,0.1040752828,0.0717406422,-0.0567062646,-0.0475695245,0.1452860981,-0.0732137039,-0.0363402218,-0.2680030167,0.4468207061,0.2502593696,0.0327756293,0.0547514074,0.1621982157,-0.1792761832,-0.1653306186,0.2657489181,-0.2976213098,0.1159587651,-0.1690282226,-0.0656774491,0.1331111491,-0.3098518252,0.1112017483,0.1109440625,-0.1791750938,0.2296435237,0.2283749282,-0.1756385118,-0.1526007354,0.3091276586,0.2405741811,0.4722813368,0.2488897443,-0.1919386387,-0.1872116178,0.657630682,0.204226166,0.1307446808,0.0229937173,0.0130414646,-0.2385837287,0.2374875844,0.3303774297,0.6330350041,0.0619604848,0.1140250936,-0.0172825735,0.0254269727,-0.0293959361,0.3783991933,0.0832827911,-0.0130594857,0.0471797958,0.2299642861,-0.2011452615,-0.5206315517,-0.1508069783,0.316210717,0.2098184675,-0.1740718484,-0.1102673262,-0.31410411,-0.0405462272,-0.1490080208,-0.1887919307,-0.2578429878,-0.131034404,-0.0023850203,0.3185665607,-0.0844664276,0.3169512451,-0.3148087859,0.6230222583,-0.0225864425,-0.227621749,-0.2458033562,-0.1571443379,0.1797525138,0.0086066034,0.1364538074,0.1759322435,0.3079749048,0.2266804725,-0.2091944665,0.0181867294,-0.2771199942,0.1260360926,-0.3050534129,0.027729759,0.1027247533,0.0826681107,0.070201315,-0.3488433361,-0.0780072883,-0.2902079225,-0.2417974323,0.1993187666,0.2402353138,-0.256210804,-0.3174752593,-0.3330124021,-0.1823951304,-0.3930699825,0.4581944644,-0.0623161644,0.1128846854,0.3934788704,-0.0587169975,0.1582943648,-0.4475107491,0.12444406,-0.4678843319,-0.3296846151,0.1020309106,-0.0584399737,0.0074493526,0.0205647647,-0.0924885571,0.2599042952,0.2751397192,-0.2820496559,-0.0737171397,-0.3988437057,-0.1644539833,-0.056347128,0.0964503959,0.4778026342,0.0493492745,-0.0785903186,-0.1939099729,-0.216729179,0.2699857354,0.5140612721,0.0867329761,0.1301528215,-0.1105623916,0.3138841987,0.5761359334,0.0391610898,0.2408781201,0.1803646088,0.1420124471,0.4015878439,-0.2043007761,-0.1414286047,0.4393417537,0.2179206908,-0.409155637,0.0728612542,-0.3133540154,-0.1283579171,-0.0502659343,0.217462644,0.1644690037,-0.1821400821,0.1009646952,-0.2382576913,-0.005092016,0.207818374,0.1116479263,-0.1473661959,-0.0391802825,-0.1153479293,-0.0875029564,0.341039449,-0.1243541613,-0.5923788548,-0.3487130702,-0.3278659582,0.4281809032,0.1119026914,0.830755353,-0.1152659357,-0.1735564172,0.1640826166,0.0619445145,0.5362945795,-0.1172737181,0.0954474434,0.260638237,0.0896617025,-0.1401681304,0.0434231088,-0.3344217539,0.1724483818,-0.3914504051,-0.2697243392,-0.3254366517,-0.1192421615,0.0361473449,0.0663407147,0.0753052384,0.1080308333,-0.320833981,0.1973257512,-0.2422353476,0.2051680982,0.2163420469,-0.1862930506,-0.4946204424,0.1055210978,0.0382707864,0.0400774553,0.0240706969,0.0755532309,0.5680780411,0.0298866704,-0.0848728791,0.2088941485,-0.3640563488,0.2437746972,0.3006787896,-0.0664612055,0.2759126425,-0.0423175059,-0.0690606087,0.4477422237,0.4921619892,-0.048148673,0.2069589198,0.0724308342,-0.1094598845,-0.2873228788,-0.0450785309,0.4593095183,-0.0728085786,0.0014989327,-0.5875300169,0.2001548409,0.4760893285,-0.1860663891,0.8500694633,0.2287766188,-0.1439658552,0.214205429,-0.1160354689,1.221617341,0.2491054535,0.0933465734,0.1263584942,0.1488142461,0.1581909359,-0.0590580851,0.2269845754,-0.043982219,-0.589613378,0.0094675077,-0.3999624848,0.2146924287,-0.1705570221,-0.3709151745,0.2024341524,0.1345631033,-0.0681876242,0.020445887,-0.0177284162,-0.331456542,0.1697222888,-0.0034246368,-0.0226266868,-0.0847361088,0.007906368,-0.0950066447,0.0382414274,0.0030143308,-0.3511697948,-0.1641819626,-0.0277094673,-0.1303076595,0.1225664169,-0.2093459219,0.038927909,0.1519820839,-0.1092840284,0.0015473284,0.2733682096,-0.0448905863,0.3576461077,0.5594891906,0.2490703017,0.1969086677,0.158167094,0.2421550602,0.132776618,-0.2835871279,0.154061079,-0.2534992397,-0.5219041705,-0.1789191216,0.1445878297,0.2153963149,0.1876575351,-0.487937808,0.3411854804,-0.402575314,-0.1266578138,0.0396501683,0.1244203523,-0.1550627798,0.520971179,-0.0555498004,-0.4595393538,0.0229396038,0.4400493801,-0.0654214323,0.2711241543,0.4940776527,-0.266498208,-0.0770594329,-0.1505247056,-0.3368164897,-0.1000726447,-0.2525239289,0.0732148141,-0.1508962959,-0.2655437887,0.0139124524,0.3450862467,0.215937838,0.4373987317,-0.1543657035,-0.2863980234,-0.2581793368,0.5522170663,-0.1103368476,0.3832840919,-0.2081240863,0.1962575018,-0.1034545228,0.2826241851,-0.2426853925,0.0834269822,-0.1535305232,0.2829118967,-0.0633066148,-0.0240060687,0.0130652906,-0.2674836218,0.0173162054,0.3564035594,-0.3901502192,-0.0926478431,0.1129371598,0.2374341935,0.1665585339,0.1201585755,-0.0711926371,-0.6452884674,0.0608286709,-0.130581364,0.0980869159,0.1315955073,-0.0744963512,0.2508107126,0.389762193,0.0260038469,0.1697997302,0.2967478335,-0.1825957894,-0.5086311102,-0.0112055251,0.0410151295,-0.285535574,-0.1894867122,-0.1129680946,-0.0598156899,-0.3204421401,0.2510896921,-0.0417909808,0.1318593174,-0.2278374285,-0.2504897118,0.3681552112,0.6540330052,-0.1026201695,-0.3410202861,0.3324914277,0.0743103772,-0.2177842706,-0.1193661839,0.0007171264,0.3264860511,0.1794406921,0.0963232592,-0.065825507,-0.2668124735,-0.072548762,0.2635923624,0.2529502809,-0.5686261058,0.2056439817,0.0314374641,-0.1529376358,-0.1393414736,0.2492614239,0.2507662773,0.3166508675,0.1620587856,0.0046582087,0.3221443295,0.5274825692,-0.0197696183,0.1290693134,-0.1776991189,0.1891236752,0.5614771843,0.0008473446,0.1952776462,0.315161258,0.4734137654,-0.0253538992,-0.1588600576,-0.1766507179,-0.1852623224,-0.0416796133,0.1477966607,0.1525010169,-0.2881447375,0.1304527819,-0.2304118574,0.0018162299,0.1126257852,0.7740170956,0.2041043788,0.0308056325,-0.3100682795,0.1341643631,0.0680978596,0.2156365067,-0.2797543108,0.203261748,0.2326891273,-0.0048195217,0.4367749691,0.4031173885,0.2432021499,0.3272054791,-0.1946522593,0.0895833597,-0.0154555654,-0.0866339356,-0.1136948988,0.3165296912,0.2770948112,-0.1894355118,0.1903124601,0.0970126465,-0.0669464543,0.0287212171,0.133405596,0.1350352764,-0.3762663305,0.3123153746,0.2819025218,-0.0253002215,0.1395068914,-0.0156964324,-0.4565905035,0.071507223,-0.2928657234,-0.4036082327,-0.0115443952,0.0796277076,0.0128479386,-0.10030514,0.1913981438,0.2416291535,-0.0139174983,-0.2232582271,-0.2436425984,-0.8886808753,0.2838243842,0.2331370264,0.0113429287,-0.0413219891,0.2859717906,0.0587555878,-0.1878685355,0.0955033451,-0.2109869123,-0.1159204841,0.166045621,-0.4591845572,0.1372925937,0.0080654845,-0.1652604491,0.0057893577,-0.1957208216,0.173664555,-0.0851025879,-0.0832149833,0.161666587,-0.0848581716,0.0822808892,0.468652904,0.4728082716,0.0355510302,0.4646649659,-0.0505962819,0.1603971273,-0.2282314152,-0.1636154503,0.1063531637,-0.2522866428,-0.0127325235,0.4332020283,-0.239588052,-0.147265777,-0.1908494681,0.4235982597,-0.2601757944,0.0941417441,-0.3972090781,0.1711036265,-0.1830190718,0.3910830021,-0.2500678599,-0.0073044305,0.0230879411,0.1512428671,-0.2764293253,-0.3034215271,0.8547509909,-0.1976509839,-0.2707471848,-0.060435839,0.6034948826,0.036227189,-0.5679724813,-0.1976687461,-0.1313981712,0.1464742422,-0.0189043675,-0.3351228237,0.0778696537,-0.4213604033,0.1482851058,-0.013073829,0.5897585154,-0.1565413475,-0.2910609543,0.1241462976,-0.182650134]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/514","title":"dataset.shuffle(keep_in_memory=True) is never allowed","comments":"I failed to realise that this only applies to `shuffle()`. Whenever `keep_in_memory` is set to True, this is passed on to the `select()` function. However, if `cache_file_name` is None, it will be defined in the `shuffle()` function before it is passed on to `select()`. \r\n\r\nThus, `select()` is called with `keep_in_memory=True` and a not None value for `cache_file_name`. \r\nThis is essentially fixed in #513 \r\n\r\nEasily reproducible:\r\n```python\r\n>>> import nlp\r\n>>> data = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\nUsing custom data configuration default\r\n>>> data.shuffle(keep_in_memory=True)\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 1398, in shuffle\r\n verbose=verbose,\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 1178, in select\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\nAssertionError: Please use either `keep_in_memory` or `cache_file_name` but not both.\r\n>>>data.select([0], keep_in_memory=True)\r\n# No error\r\n```","body":"As of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.","comment_length":131,"text":"dataset.shuffle(keep_in_memory=True) is never allowed\nAs of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.\nI failed to realise that this only applies to `shuffle()`. Whenever `keep_in_memory` is set to True, this is passed on to the `select()` function. However, if `cache_file_name` is None, it will be defined in the `shuffle()` function before it is passed on to `select()`. \r\n\r\nThus, `select()` is called with `keep_in_memory=True` and a not None value for `cache_file_name`. \r\nThis is essentially fixed in #513 \r\n\r\nEasily reproducible:\r\n```python\r\n>>> import nlp\r\n>>> data = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\nUsing custom data configuration default\r\n>>> data.shuffle(keep_in_memory=True)\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 1398, in shuffle\r\n verbose=verbose,\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 1178, in select\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\nAssertionError: Please use either `keep_in_memory` or `cache_file_name` but not both.\r\n>>>data.select([0], keep_in_memory=True)\r\n# No error\r\n```","embeddings":[0.0582521483,0.1715381891,0.0236405991,0.1729352325,0.1541408747,-0.1635663211,-0.3068185747,0.3015698195,-0.1271577775,0.203600049,0.1368852705,0.5172116756,-0.3025810719,-0.431401819,0.0418817587,0.1355720311,0.0163559653,-0.0274757352,0.0766994432,-0.1642751247,-0.4050556719,-0.0679784268,-0.5050225258,-0.2724936903,-0.353969872,-0.0422089845,-0.0651145205,0.0619966909,-0.2063216567,-0.5102363825,0.046151191,0.363330096,-0.370182693,0.3270519972,-0.0001268262,-0.2589108348,0.3602529764,-0.0730153546,-0.336335063,-0.0587591082,-0.3150103986,0.2423155159,-0.0876233578,-0.1639830023,0.2200228721,-0.1693580896,0.05125577,-0.4195265472,0.1856562495,0.3109852374,0.0398746654,-0.0168228857,-0.0427034087,0.0539652221,0.400021553,0.050295461,-0.0710290521,0.1810719371,0.025824314,-0.1065656245,0.1176697016,0.1653529853,-0.2495699078,0.0350748859,0.1981938332,-0.1733987331,-0.1578304023,-0.1440363079,0.2231003642,0.226267755,0.1969543546,-0.3949074447,-0.144900009,-0.1585543305,0.0652295798,-0.3188783824,0.3673403263,0.2271049023,-0.0369679816,-0.141162172,-0.281786561,-0.1317581236,0.0459681116,0.0505823828,0.1565725356,0.4199050069,0.0662545115,0.2579324543,0.5411183834,0.0532425605,0.1429492086,-0.1205495819,-0.1223397031,0.3546843827,-0.4297558963,-0.1071819961,0.1272524744,0.2284680307,0.3260549009,0.1716279387,0.2923562229,-0.001856618,0.0666188002,0.1755289882,0.1559852511,0.4208346903,0.0823542625,0.0897043049,0.2520473301,-0.0248581469,-0.0233360268,0.1855761856,0.0089342995,-0.251447916,-0.0599062853,0.028395813,0.1699498892,-0.2244923413,-0.1456382424,0.1323855221,-0.4279551804,-0.2963871658,0.0615714565,0.4393815398,0.2439153194,-0.036311131,-0.0334248208,0.0826520845,0.0316745453,-0.0285779368,-0.2911154032,0.1890472174,-0.3774471879,0.1855505109,0.414896071,0.0236336868,0.4221939743,-0.1622270644,-0.0405370481,0.1626266539,0.0964482576,-0.3839091659,0.4567762613,0.2786706388,-0.2528736889,0.0173069797,-0.0764115006,0.1386786997,-0.2524915934,0.4537697434,-0.0491760075,-0.2829820812,0.3743939996,-0.0348970592,-0.2979338169,-0.034002915,-0.1033734307,0.0014509729,0.3998070955,-0.3425259888,0.5083597898,-0.2778913975,-0.0489257202,-0.3168386817,0.1293801218,0.2588361502,-0.1447185874,0.0277004186,-0.2785175145,0.0615408309,0.1749753505,0.3470169306,-0.0739551783,-0.0982237309,-0.3033793271,0.0451323539,0.2472506315,-0.233133778,-0.5170556307,-0.4015239775,-0.0577961132,0.1500942856,0.1228232384,-0.0361853391,-0.0466819741,-0.2397120595,0.569152832,0.1230766848,0.0507953987,0.0702164099,-0.2244808674,-0.144208923,0.3917486966,-0.1554684937,-0.2716077566,-0.1668291688,0.1186869442,0.0079841008,0.3367796242,0.0780103803,-0.0938457996,-0.0691634417,0.3975318074,-0.2684621811,-0.1448493302,-0.0017916842,-0.262835145,0.1890727282,0.1500018835,-0.0776878074,-0.0046193781,-0.1408882439,0.0069597056,-0.1317948103,-0.0786308944,-0.4655655921,-0.0289580505,0.2572417259,0.2690214515,0.1674773395,-0.1987805367,0.1474260539,-0.6084166169,-0.1418688595,-0.0942370966,0.1213890314,0.1407781988,-0.2250754982,-0.0015512937,0.2883162498,0.0494448617,0.0084298318,-0.2568846345,0.4210997522,0.2343772948,0.0555305518,-0.0846201852,0.2513121367,-0.0905506313,-0.1039312556,0.25464046,-0.227212593,0.1112728789,-0.1098298207,-0.1313947737,0.1791956276,-0.3276397586,0.0956532955,0.1494388431,-0.2250342667,0.1931090653,0.2141469121,-0.3467211127,-0.0755976886,0.3512639701,0.2432071269,0.3952656984,0.2722056508,-0.2244189084,-0.0795437843,0.7486112714,0.1755332649,0.1091258749,-0.0206410158,0.0418238193,-0.2721835971,0.2493070662,0.3991541564,0.6660166979,0.0948135853,0.0460690781,-0.0085152518,-0.1063189507,-0.1277070343,0.3653984368,0.1909220964,0.116327934,0.0705738366,0.345372498,-0.1046396941,-0.583958149,-0.2935898304,0.2861483097,0.1273398697,-0.1608408391,-0.1700657904,-0.3559508622,-0.0956454054,-0.197192952,-0.1961465329,-0.1914370507,-0.1775632054,-0.0734645426,0.4637481868,-0.1138138771,0.3132784963,-0.3044902086,0.5873445272,-0.1263481826,-0.3305932283,-0.2006736249,-0.1024089828,0.0209167693,-0.0236809943,0.0142460261,0.0744796768,0.2375768572,0.2396815717,-0.1397291273,0.0067295702,-0.4025407434,0.1071235389,-0.2851748466,0.0803081468,-0.0229946487,-0.0686065257,0.0633816049,-0.3273164034,-0.0840858668,-0.3485420346,-0.1023633555,0.129064545,0.210666284,-0.1978334188,-0.3947789073,-0.3335371315,-0.1437161863,-0.4087192118,0.408521384,-0.129422769,0.1063499674,0.4895442724,-0.070001848,0.2074606419,-0.5142788291,0.0601189658,-0.3819688857,-0.3399350643,0.2264350206,-0.1062783524,-0.0376657844,-0.0557117052,-0.1253292561,0.3742881715,0.3618136644,-0.1126945838,0.0698801205,-0.3953334987,-0.1428849101,-0.0798912719,0.0358972922,0.5086120367,0.011170717,0.0101618664,-0.1552386284,-0.3371166587,0.2735789418,0.5556554794,0.1936832964,0.1405683011,-0.1279440522,0.3053508699,0.6116023064,0.0180628728,0.2218025476,0.2513357699,0.198574245,0.3103548288,-0.1589478552,-0.0326582529,0.5106629133,0.2174625993,-0.3428010046,0.1145192236,-0.2571847141,-0.1489802599,-0.0304284375,0.1263510436,0.0973010957,-0.2774019241,0.1408428997,-0.0972375497,0.0415064059,0.1747030616,0.169573918,-0.0600337684,-0.1153497398,-0.1410131902,-0.0297326203,0.3056159317,-0.0441634431,-0.6732628942,-0.3755868077,-0.4721690416,0.3012936413,0.065272972,0.8658403754,-0.0815697312,-0.1442963481,0.227611497,0.0985192209,0.6172943115,-0.0488356426,0.0754870027,0.2543058395,0.1052025557,-0.2266443074,0.0009143866,-0.2905430496,0.0431483537,-0.3057165742,-0.1708203554,-0.3268437386,-0.1532397866,-0.0169361439,-0.0528797843,0.1032339633,0.1100655049,-0.2144611329,0.0156568121,-0.2628638744,0.1524929255,0.2070516348,-0.0559040122,-0.5712479949,0.1035744995,0.0509471893,-0.0004634329,0.0421867408,0.090068914,0.4630864263,0.091573365,0.0178620648,0.247365877,-0.4582463503,0.2574883103,0.1381069273,-0.0138703482,0.2481126338,-0.0344713777,0.0659726411,0.4264141023,0.5325857401,-0.0090929354,0.1813393384,-0.0978788212,-0.0824097916,-0.1879819632,-0.0353313237,0.3640075028,-0.1816083193,0.044892218,-0.6824074984,0.3029801548,0.5967734456,-0.1900641322,0.7536084056,0.2377085835,-0.2059596181,0.1781200171,-0.0468188487,1.312079668,0.2189839184,0.1824340671,0.2493997812,0.1198646054,0.3375508785,-0.1520350426,0.3163945973,-0.0312289838,-0.3117735982,0.0648642704,-0.3569341898,0.2322402,-0.2951771915,-0.2881934941,0.3330242634,0.1268987805,-0.1023201346,0.0835058466,-0.0195151605,-0.4196310639,-0.0224889237,-0.0607589521,-0.0592128895,-0.2439978272,0.0793126747,-0.0112399478,-0.0675592199,-0.0331189036,-0.2414554954,-0.0576565117,0.1237726659,-0.1919861734,0.0547811091,-0.1553829759,0.1071045473,0.1933681965,-0.0190596804,0.0026660927,0.2434748858,-0.0744905099,0.3823255301,0.429853797,0.2604591846,0.1087382883,0.0632611662,0.2432439476,0.1503244191,-0.3025839031,0.1687010974,-0.1743139327,-0.424126029,-0.354119271,0.0664166361,0.3193895817,0.2247797251,-0.2856481373,0.3518815637,-0.4087076485,-0.1726908237,0.017225232,0.0574396998,0.0024693217,0.4494126737,-0.1175222173,-0.3633414805,0.0476154126,0.5570681095,0.0076236627,0.3624717891,0.4736931324,-0.33117643,-0.0371003412,-0.068341352,-0.4076806307,-0.23671709,-0.2915999889,0.0075128335,-0.1156883165,-0.2336660922,-0.0219087452,0.3400224149,-0.0292224847,0.2901128531,-0.0539475791,-0.2933695912,-0.1853292435,0.5183563828,-0.2268459946,0.3059214652,-0.0526913218,0.3073262572,-0.0820918307,0.1352659762,-0.1991172731,0.0094818678,-0.0242254697,0.2965017557,0.0616707839,-0.0675209016,-0.0440090373,-0.2257984281,-0.0656407401,0.4734465778,-0.3245003223,-0.0415418968,0.008200136,0.2415739745,0.2567052245,0.0518080182,-0.0856154338,-0.6488208175,0.0862816349,-0.1840725094,0.1395010054,0.1251874864,0.0484997071,0.3103797436,0.323656857,0.1338423491,0.3690098226,0.3317423761,-0.1376085728,-0.547177732,-0.0068298765,0.0681181476,-0.3414595723,-0.172171697,-0.2888044715,-0.0765359998,-0.2675986886,0.3461116254,-0.0858969167,0.1340374947,-0.3176960349,-0.3257191777,0.1708032936,0.622497201,-0.1092253998,-0.4840922952,0.3311726749,0.0048595276,-0.0928247124,-0.0435001031,-0.1759013534,0.3503393233,0.1592321396,0.1100088283,-0.0759282112,-0.3464381397,-0.1687949151,0.1526178271,0.2878944874,-0.5417854786,0.2586516142,0.1237137094,-0.1214045361,-0.1668012589,0.2514119148,0.1404767185,0.3174530864,0.2154073268,0.0999704227,0.3795158267,0.5215622187,-0.0019801056,0.1691205502,-0.1758662164,-0.0089274561,0.6253475547,-0.0551837012,0.2040561885,0.2317167521,0.3403336704,0.107828185,-0.0912869498,-0.1490228772,-0.1434207708,-0.0192735512,0.0140529769,-0.0285806246,-0.2736736238,0.0153437862,-0.1142724007,-0.0629808679,0.0013237052,0.6738683581,0.0799124241,-0.0938675776,-0.4253213406,0.1122412384,0.1956027299,0.1990813017,-0.3041790128,0.255056411,0.1390112787,-0.1476669461,0.2324561924,0.2578600347,0.2080412507,0.3255351782,-0.0578252226,0.051676102,0.0448113196,-0.0683915317,-0.0784833878,0.2672591507,0.386857897,-0.2032080293,0.178051874,0.0537288338,-0.0439273976,-0.0059597502,0.2202218026,0.2924682498,-0.3709783256,0.3863028288,0.3180395961,0.1202207655,0.245804444,-0.0817094371,-0.4959750175,-0.107197009,-0.2609366775,-0.2421960831,0.0345698148,-0.0283225905,-0.0033037632,-0.1313527524,0.2376209944,0.1642875969,-0.2725536823,-0.2089897245,-0.2560702264,-0.7737237215,0.0829083622,0.188056916,0.0842100754,-0.0520197637,0.294865787,-0.0397689641,-0.1835630089,-0.0507862456,-0.0545091629,-0.0149621032,0.0740039274,-0.4018296003,0.0929490402,-0.0839924067,-0.0751556307,0.0935691968,-0.3197195232,0.102796033,-0.1501589715,-0.1391862035,0.0401842743,-0.0024999944,0.14774777,0.4713801742,0.4296171963,0.0616028979,0.3756872118,-0.0348588079,0.0771450922,-0.3971295357,-0.0284060501,0.0693795756,-0.3193969131,-0.090302147,0.4964284599,-0.203197673,-0.0359871313,-0.1933774352,0.4085067213,-0.2978166938,0.1286790669,-0.2891921699,0.1979857236,-0.1864554882,0.4920978248,-0.2932485938,-0.0035865554,0.0226559751,0.2047997117,-0.3808238506,-0.2601338923,0.8820676804,-0.3497811556,-0.2414846271,-0.1135065183,0.5502931476,0.0787510425,-0.4763632417,-0.1970194131,-0.0315152667,0.1201934963,-0.0565548651,-0.3065265417,0.2336234897,-0.3561180532,0.0286806524,-0.0238203611,0.4381566942,-0.1030241475,-0.2813260555,0.1388041824,-0.2105051726]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/514","title":"dataset.shuffle(keep_in_memory=True) is never allowed","comments":"Oh yes ok got it thanks. Should be fixed if we are happy with #513 indeed.","body":"As of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.","comment_length":16,"text":"dataset.shuffle(keep_in_memory=True) is never allowed\nAs of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.\nOh yes ok got it thanks. Should be fixed if we are happy with #513 indeed.","embeddings":[-0.1078727394,0.0314778425,-0.0592158996,0.1798339337,0.2271602452,-0.0569601916,-0.0781591311,0.3618608415,-0.017298406,0.310203135,0.0617669411,0.5769485235,-0.2770041823,-0.4012131095,-0.0302102193,0.0821130574,0.0426669531,0.0644076467,-0.0605757385,-0.1543073803,-0.4035514593,-0.1343421936,-0.4251638949,-0.1685622483,-0.2936737835,-0.1476836503,-0.0694249272,0.1123735085,-0.2651752532,-0.4221323133,-0.0507327616,0.2752957046,-0.4287704825,0.4618907571,-0.0001105258,-0.2425429821,0.2252237648,-0.0732617751,-0.2165711671,-0.0847063959,-0.3749551475,0.102944456,-0.0625902563,-0.1702314466,0.1931007504,-0.1649805158,0.050829947,-0.4214198887,0.2389518917,0.212985903,0.2309739739,0.0795905069,-0.0047133146,-0.0425943919,0.5332520604,0.0685249642,-0.1711269021,0.1763259321,0.1607525796,-0.2377842218,0.0212521441,0.2817619741,-0.1588008255,0.2197417319,0.2054112852,-0.078985557,-0.1350956708,-0.1896862388,0.309653163,0.179354459,0.2681084871,-0.3994344473,-0.2405280471,-0.1129989251,0.0084619131,-0.3494419754,0.3002124131,0.2281771451,-0.0264050737,-0.1879418641,-0.1457694769,-0.1069997698,0.163274169,-0.0938220322,0.2004605085,0.2594092786,0.0756183565,0.072026521,0.492928952,-0.0876783207,0.2041461468,-0.0494837165,-0.2196620554,0.311832726,-0.4002455771,-0.0920264721,0.1614938527,0.074483484,0.2304908633,0.0756535754,0.2575995326,0.0067211818,0.0132537782,0.2516729832,0.2001265883,0.3784680367,-0.0654057786,0.2134313285,0.1905719191,-0.0800729692,-0.1731542647,0.0880875438,0.0415702164,-0.3028277457,0.074925825,-0.0800301507,0.1167870611,-0.2463715225,-0.1012335047,0.1318372935,-0.1357847601,-0.2121957242,0.0453829467,0.5516231656,0.1866026372,-0.0767670795,0.0147760455,0.0687277243,-0.058700826,0.0963963494,-0.362719357,0.1153650358,-0.3626089096,0.0938812569,0.2376080155,0.1213927269,0.3731115162,-0.1051931605,-0.0658724904,0.1102133766,0.273516506,-0.2634017169,0.3925337791,0.2290444672,-0.2552441955,0.0241906587,0.0206683706,0.1493790001,-0.2203023881,0.4370922148,-0.1299475282,-0.2776747048,0.2739734352,0.1745291054,-0.0480163023,-0.0379783921,-0.003666705,-0.0730223134,0.446079582,-0.3608433902,0.3171112537,-0.2089282572,-0.0279092379,-0.3870055079,0.0059968843,0.1643785387,-0.2229548395,0.0757542104,-0.2734275162,-0.0148449326,0.1948697418,0.4441378117,-0.0910409391,-0.3584691286,-0.2399041206,0.173092097,0.2468248755,-0.1380795836,-0.5333728194,-0.2369209528,-0.1584784836,0.0361692458,0.0879132673,0.0584418848,-0.0627003312,-0.1163539216,0.4518163204,0.0210287981,0.0183452349,0.1343852878,-0.3099245727,-0.1400605291,0.3697386086,-0.15208067,-0.2401830852,-0.1730976254,0.1195153147,-0.1781291813,0.3807413876,0.0413312055,-0.0308136549,-0.0349831395,0.484125644,-0.2198288441,-0.084740378,-0.0747508183,-0.1592671126,0.1509011388,0.0692167953,-0.128204301,-0.1019469574,-0.2161985338,-0.050065089,-0.0720342547,-0.1936576068,-0.3217894733,0.1400498748,0.055104278,0.0714577511,0.055319421,-0.1687679589,0.092811428,-0.4236603677,-0.1620372534,-0.0887979195,0.1601252258,-0.1244970262,-0.2009057105,0.0920647532,0.2002391964,-0.0514342897,0.0513803363,-0.3045920432,0.4380547404,0.2062827349,0.0823625103,0.0751083344,0.0988028497,-0.1049174219,-0.1015241146,0.264321208,-0.1814331412,0.113137573,-0.1391750276,-0.0818582401,0.2396378219,-0.2278598845,0.0760872215,0.1914031506,-0.2140289843,0.2335218787,0.1947780401,-0.0763356611,-0.18927975,0.3093521893,0.2143834978,0.3748636544,0.2533474267,-0.2395371348,-0.0730855763,0.752476573,0.2206949443,0.1499158293,0.0220452789,0.0211722478,-0.3006110787,0.1650642008,0.4131223559,0.4986058176,0.1892418116,0.1480913907,-0.002885808,0.0206921361,-0.0762027353,0.3187276125,0.0252004452,0.1475012004,0.0877850279,0.276910454,-0.2225174606,-0.5638150573,-0.1646709442,0.3080355525,0.2424911112,-0.0820208788,-0.2388604134,-0.3599188328,-0.1361257285,-0.1408329159,-0.106000632,-0.212834239,-0.1299735904,0.092754364,0.3725082874,-0.2125671655,0.3739692271,-0.2482064068,0.4168656766,-0.0222027451,-0.1171981692,-0.2320803404,-0.2177790105,0.1436156332,0.1165102497,0.1697309017,0.0945722312,0.3899741173,0.2874011993,-0.1426670104,-0.0584663749,-0.215809986,0.0347240977,-0.1650770605,-0.0293394774,0.0274550952,0.0614132807,-0.1137502044,-0.3544992208,0.0075111478,-0.3451034725,-0.2569625676,0.1152637005,0.2479770482,-0.1941602677,-0.3728398681,-0.4107463658,-0.2554987073,-0.4678596258,0.4274529815,-0.0421122909,0.1697332114,0.4122061133,0.0268561561,0.0871611238,-0.3679885268,0.0467262566,-0.4801112711,-0.3200534582,0.1664246619,-0.2421019971,-0.1274799556,-0.0509109646,-0.0797985941,0.2987231314,0.2556759417,-0.1701513231,-0.014400091,-0.4049577415,-0.0595423244,0.0206536874,0.1705158949,0.5260155797,0.1508603543,-0.1882628798,-0.1759503633,-0.1672910303,0.2345252633,0.3991559744,0.197608456,0.030670708,-0.0013973066,0.2786433995,0.6611911058,0.0348750874,0.2650378644,0.2444595844,0.1208591983,0.3795444071,-0.3707634807,-0.0451818556,0.3426940143,0.1007099003,-0.342136234,0.0898478925,-0.1920036525,-0.366315484,-0.0647846088,0.2407469153,0.0730638057,-0.18177405,0.0891747102,-0.0761930719,-0.0610054955,0.2142038494,0.0979064479,-0.2363516688,-0.1197244152,-0.0830202177,-0.1502345204,0.2026253045,-0.1722979993,-0.5896502137,-0.2854010761,-0.4158503711,0.2997022867,0.0219847932,0.6304904819,-0.046244856,-0.2750856578,0.1421259493,-0.0136844702,0.4843626022,-0.1223758161,0.0978710428,0.2757172287,0.2364537269,-0.1469788849,0.0225874353,-0.3153304756,0.1052415892,-0.2829584181,-0.2083058655,-0.253583312,-0.079295449,-0.0413057804,-0.016632095,0.0280888975,0.0557618104,-0.322219789,0.1323811263,-0.3599127829,0.1306756735,0.209744826,-0.1717753112,-0.4578042924,0.0188158005,0.0959856957,0.0022871611,-0.050848309,0.1427773833,0.578363061,0.0532344468,0.0800341517,0.2153636962,-0.2208368778,0.2404068708,0.3562558889,-0.180362016,0.2077956498,-0.0196924731,-0.1094742715,0.3237188458,0.403108418,-0.0087547805,0.2421165556,0.0413068905,-0.0269217864,-0.1610459238,-0.089862071,0.3695458174,-0.1644634455,0.0096179945,-0.6999325156,0.2722674012,0.3819766939,-0.1795826852,0.7695519328,0.0342331119,-0.0926487967,0.3315494657,-0.0772523358,1.1195508242,0.0339704826,0.1771061122,0.1755619496,-0.0159260109,0.1148792878,-0.0396805741,0.2906040251,-0.2378376126,-0.521114409,0.1604729891,-0.269826889,0.1762300879,-0.0710654631,-0.2536159456,0.2452199608,0.0929339528,-0.0444948152,-0.172701776,0.0078682266,-0.3439272046,-0.0405512042,-0.0429272614,0.1322551072,-0.0335360952,0.0299750939,-0.0586016178,0.013780605,-0.063080512,-0.2973026335,-0.0944353789,0.1352884769,-0.2044692636,0.2324853987,-0.2904860079,0.0062023555,0.0109412214,-0.0547461621,-0.1283721477,0.2540833652,-0.0481442958,0.4096056819,0.5715175867,0.2314739525,0.0768476427,0.2145596147,0.2458197325,0.0255630333,-0.2269501984,0.0539178327,-0.1483024508,-0.4766715765,-0.1591783315,0.0363246277,0.2632266283,0.1623258442,-0.3679367602,0.2486217916,-0.3597947657,-0.2111902386,0.1669828594,0.1270516813,-0.1322497427,0.4700601101,0.0142770195,-0.2295577675,-0.0009838848,0.5051695704,0.0864669755,0.2383112609,0.5237416029,-0.0726725459,-0.1700122207,-0.287299484,-0.2536242008,-0.0360128619,-0.2044704705,0.1837618053,-0.0554690398,-0.2839893401,0.0688749775,0.2643089592,0.1066330671,0.4070765972,-0.094849661,-0.2292091697,-0.4201833308,0.4437668324,-0.099257037,0.4140265584,-0.2612751126,0.0826347619,-0.121506013,0.1916929632,-0.3876183033,0.0938353762,-0.0836898386,0.2934552133,0.0157555435,-0.111934647,0.0230321921,-0.1707738936,0.1524876654,0.3507134318,-0.3724937737,-0.2430212647,0.1457935274,0.1175747663,0.0835485458,0.0592063405,-0.0896506831,-0.5925161242,-0.1081837043,-0.1836387217,0.0946238935,0.1458480358,-0.0578490458,0.3071137667,0.3662775159,-0.0061752493,0.372161597,0.2892753184,-0.1997232139,-0.3840440512,-0.0466640592,-0.0214118846,-0.1907102615,-0.1097671241,-0.0990030393,-0.0674088299,-0.2541639507,0.2214464098,0.0556140132,-0.0175468232,-0.2199251652,-0.0980802625,0.4295175076,0.6297251582,-0.0764326081,-0.3640529513,0.4069697559,0.2642086148,-0.3455339074,-0.1629193723,0.0464293174,0.2752840221,0.1474657655,0.0083066886,-0.0759513602,-0.1910993606,-0.1815365255,0.2896525562,0.2205361128,-0.6552200913,0.2770577073,0.1268347055,0.0210315678,-0.17865026,0.2017375976,0.2562086582,0.3099995852,0.3726297319,-0.0460222177,0.3559654057,0.4608314633,0.0659659058,0.11708498,-0.2425633073,0.1153915748,0.5658167601,-0.081200026,0.276699245,0.2415809184,0.3678364754,0.0612211563,-0.1463450342,-0.1686167121,-0.04493187,-0.1831551343,0.0117362384,0.1710044593,-0.2913222909,-0.0583174303,-0.0793403909,-0.0683768541,0.1755081564,0.6497915983,0.1210021228,0.0610482395,-0.367777437,0.2369811833,0.1279889345,0.2782155871,-0.3036893904,0.1446888894,0.1185062081,-0.0752007365,0.3234330416,0.2481110692,0.1816097498,0.3829049766,-0.2008118927,0.0689979568,-0.050516706,-0.1265684962,-0.1566113681,0.3311425745,0.3847040236,-0.1390633136,0.1985670924,0.1633262336,-0.1767994761,0.0143265286,0.2011549026,0.1655360907,-0.1903313845,0.2306905538,0.1728613675,-0.150460586,0.0427806079,-0.1037800983,-0.4727163911,-0.03838934,-0.1418858916,-0.2031335682,0.0378007703,-0.0849466845,0.0919963866,-0.092162095,0.3033494353,0.1568582505,-0.1262314916,-0.1780654937,-0.2676553428,-0.9151129127,0.2245615274,0.0515014492,0.0624052435,-0.1349491775,0.3846207559,0.0082023777,-0.2036102265,0.1121895984,-0.0982702672,-0.0396633334,0.2232722938,-0.4372611344,0.0702404901,-0.0814529806,-0.0988089889,0.1465445012,-0.2478380352,0.0311280172,-0.1147615165,0.0479399525,0.0072398069,-0.0380123071,0.0359303467,0.3697406948,0.4667228758,0.0161225684,0.2574400902,-0.0802935436,0.0422933921,-0.297757417,-0.2828944325,0.0245156735,-0.2185552865,0.070948571,0.47692433,-0.1603500992,-0.1255030036,-0.2700118124,0.29123348,-0.2710494399,0.1183954701,-0.3323071003,0.2123012543,0.0901179686,0.3516814709,-0.2855340838,-0.0280278679,-0.036668472,0.1451439261,-0.2931612432,-0.3934074044,0.7955736518,-0.2120288461,-0.2552544475,-0.1243004128,0.5379705429,0.0219981745,-0.4403071105,-0.3389934003,-0.0300063975,0.1150769517,0.0005135578,-0.2319951206,0.1029696614,-0.3894504011,0.0518758185,-0.0874024555,0.5879086256,-0.1065413132,-0.2709550261,0.1191219836,-0.2413641661]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/514","title":"dataset.shuffle(keep_in_memory=True) is never allowed","comments":"My bad. This is actually not fixed in #513. Sorry about that...\r\nThe new `indices_cache_file_name` is set to a non-None value in the new `shuffle()` as well. \r\n\r\nThe buffer and caching mechanisms used in the `select()` function are too intricate for me to understand why the check is there at all. I've removed it in my local build and it seems to be working fine for my project, without really considering other implications of the change. \r\n\r\n","body":"As of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.","comment_length":76,"text":"dataset.shuffle(keep_in_memory=True) is never allowed\nAs of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.\nMy bad. This is actually not fixed in #513. Sorry about that...\r\nThe new `indices_cache_file_name` is set to a non-None value in the new `shuffle()` as well. \r\n\r\nThe buffer and caching mechanisms used in the `select()` function are too intricate for me to understand why the check is there at all. I've removed it in my local build and it seems to be working fine for my project, without really considering other implications of the change. \r\n\r\n","embeddings":[-0.1627905816,0.1008973792,-0.0472362339,0.141304642,0.1969874352,-0.0894875675,-0.0724679828,0.3746205568,0.0284968764,0.3297641873,0.054085426,0.5689108968,-0.2949666977,-0.4070610702,-0.063083522,0.1825318038,0.0914553255,0.0464378409,-0.06404268,-0.1149820238,-0.4023231566,-0.1355522275,-0.3807732165,-0.2072246373,-0.2809173167,-0.1180846766,-0.0617817342,0.0914643034,-0.2582835555,-0.4361536205,-0.0063434797,0.347877264,-0.409319073,0.4160298705,-0.0001099763,-0.1791142374,0.2073139101,-0.0738750994,-0.2709137797,-0.0989495292,-0.3752152026,0.1104811728,-0.0958124101,-0.0919574276,0.1756296158,-0.1854459643,0.0588805191,-0.5130761266,0.1786990166,0.2128080279,0.2111129314,0.0499816909,-0.0335805938,-0.0293106381,0.5861102939,0.0566133894,-0.1890409291,0.1663571447,0.1492808759,-0.1905910522,-0.0007994702,0.2823547721,-0.2549176812,0.2575418353,0.2611257136,-0.0905021727,-0.1236087754,-0.1846337467,0.3069356084,0.2646509707,0.2544342279,-0.3675505519,-0.2848781049,-0.1775666922,-0.0083235037,-0.3538299799,0.3373945653,0.1521790475,0.0035764987,-0.1769806594,-0.1892120242,-0.0270581804,0.2319644541,-0.0966159105,0.2491226494,0.2606118023,0.0594690368,0.0409907475,0.5399266481,-0.0905899554,0.2658842802,-0.1224403903,-0.2388854772,0.2838874161,-0.3870707452,-0.1137686595,0.1962217242,0.0459888168,0.2212799788,0.0704398379,0.2729631364,0.0708098859,0.0240048617,0.202840656,0.1022178531,0.4053637981,-0.0894306824,0.1724646837,0.1995564401,-0.0315999389,-0.140944764,0.0801540986,0.005148387,-0.2488575727,0.0470433049,-0.0482765324,0.099941209,-0.2141382694,-0.1254854649,0.1203748956,-0.1201372221,-0.2233006209,0.090140745,0.5079164505,0.1808762103,-0.0202656128,-0.0539890192,0.0603303425,-0.0707249343,0.1533785611,-0.3479059041,0.0679386929,-0.3014752865,0.1446823776,0.2900347412,0.0929114297,0.3374700546,-0.0530469567,-0.0744542181,0.1677108854,0.2409879118,-0.2988532186,0.4511573911,0.2090821862,-0.327780515,0.0116757499,0.0305169038,0.0661879629,-0.271343112,0.4143098891,-0.1308401823,-0.2582525611,0.2951088846,0.168598488,-0.078163974,-0.0627123937,-0.0282901675,0.0004497216,0.4258086979,-0.3920407295,0.3442200422,-0.2152245194,-0.0734057575,-0.3512366414,0.0223537236,0.1592847109,-0.2199881375,0.0661382601,-0.264898777,0.0545590147,0.1390299946,0.4122771025,-0.0707731023,-0.398648262,-0.1971273273,0.0905979946,0.2141714394,-0.1521358639,-0.5648226738,-0.2408738732,-0.1316463053,0.0403598472,0.1658996195,0.0475503877,-0.0199212302,-0.1683561504,0.4874163866,0.0159762781,0.0379024521,0.1605656594,-0.286796838,-0.2108647674,0.2988801003,-0.1238557547,-0.25221017,-0.1908814013,0.1199977994,-0.1736690849,0.3175570071,0.0462701507,-0.0111070732,-0.033735577,0.4648877978,-0.2766029239,-0.0674634948,-0.1012833118,-0.1785651147,0.1913275123,0.068677716,-0.2096097767,-0.1075913832,-0.1848170906,0.0236457866,-0.1183180138,-0.2057343572,-0.3201948106,0.114191398,0.1147771478,0.158302471,0.0917552337,-0.1539441496,0.0645098165,-0.3795572221,-0.1858312935,-0.0621419549,0.1152379289,-0.0434533209,-0.242235139,0.0461175591,0.2248344421,-0.0923197567,0.0642551631,-0.3035538495,0.4044660032,0.2118751705,0.0654413402,0.0497317687,0.2067206651,-0.0972361639,-0.0455045328,0.3066761494,-0.2456496656,0.1084035635,-0.0401805677,-0.0883613303,0.2758997381,-0.2985675931,0.1285168231,0.1937158853,-0.3063238561,0.2506430745,0.1813519746,-0.0775220171,-0.237853542,0.2764934599,0.2271322012,0.4078924954,0.2792389095,-0.168536976,-0.0030661023,0.8022277355,0.2215924859,0.1226186454,0.0282464344,-0.0099236406,-0.3311356604,0.1510760486,0.4368934929,0.5458803773,0.2066428065,0.139176324,-0.0523492135,0.0427910052,-0.1248243377,0.3032709658,-0.0051012756,0.0998852178,0.1477361619,0.2927755713,-0.1784620583,-0.5660678148,-0.117468372,0.3162509203,0.1801416129,-0.0468137003,-0.1696550548,-0.3915604055,-0.1134415269,-0.1453727484,-0.1362470835,-0.2019047141,-0.1397088468,0.1181998402,0.4280184209,-0.2240033746,0.4030801654,-0.2191064954,0.3761196434,-0.0820556134,-0.2009871602,-0.2323857844,-0.2473513186,0.113504082,0.1071829796,0.1801324487,0.0349476971,0.3226132095,0.2647101879,-0.1561774015,-0.1072735488,-0.2211646885,0.0503194369,-0.1555036157,-0.0216474384,-0.0479001515,0.0101211295,-0.0882760808,-0.3869800568,0.008692015,-0.3831766248,-0.2412016094,0.178995043,0.1787716001,-0.1435182989,-0.4018124938,-0.333520323,-0.2333360314,-0.4529966116,0.3741290569,-0.0639918149,0.0997937471,0.4208821952,0.0501933917,0.1192922294,-0.370349288,0.0566270426,-0.5147151351,-0.3384055495,0.1441303939,-0.1881360114,-0.0859256759,-0.1079170406,-0.1023709923,0.2616953552,0.3022656441,-0.1528601944,0.0201328509,-0.4135584831,-0.0220302772,0.0647840053,0.0957225263,0.5864396691,0.1653224379,-0.2120554745,-0.192295596,-0.138011992,0.178935796,0.3718593121,0.179812029,0.0572609268,0.0182418022,0.2684489489,0.7134554386,-0.0110729849,0.2548037171,0.2837665975,0.1679668427,0.4330673218,-0.4372291565,-0.0121131735,0.3658877015,0.0932755023,-0.3274992704,0.107803084,-0.2009227127,-0.4239680171,-0.0702934489,0.253644824,0.0882731155,-0.1928747892,0.1146664619,-0.1082742736,-0.0777749941,0.2088018805,0.1254799962,-0.2121367007,-0.1972232014,-0.0898095444,-0.1305542588,0.205319047,-0.1757034063,-0.5716443062,-0.3005228937,-0.4015974402,0.3316797018,0.0836056098,0.6012481451,-0.0574416742,-0.3226551712,0.0937577188,0.0277690608,0.4177288115,-0.1002545878,0.0553203933,0.2866081297,0.2566705346,-0.1767850965,-0.0243481882,-0.298376143,0.0748021081,-0.3461553752,-0.1568787545,-0.271336019,-0.105112426,-0.0974851623,0.0260276888,0.0498910099,0.1094591692,-0.4067121148,0.1649535745,-0.3717480302,0.1544073969,0.1667530537,-0.1572788954,-0.4636015892,-0.0095959837,0.1169116274,-0.0204659477,-0.0727098137,0.1309043616,0.5553306937,0.1349182874,0.0567920245,0.2547768056,-0.2617733777,0.2977277637,0.2745452821,-0.2015114576,0.2483657151,0.0267158933,-0.0410489924,0.249118939,0.4222625792,-0.0196001362,0.2453920394,0.1159087718,-0.0461063571,-0.2089996785,-0.0855993852,0.3329139352,-0.1725011319,0.0061698486,-0.7266321182,0.2700081766,0.4216602445,-0.2021493018,0.808962822,-0.0263066348,-0.0745423362,0.3130102456,-0.0293327607,1.0826152563,0.0523930714,0.1845881641,0.1523943245,-0.0149463499,0.1152192503,-0.0274769161,0.3338971734,-0.1610727608,-0.5241378546,0.1641458869,-0.2412721813,0.1434488893,-0.0853687972,-0.2555903792,0.2435743213,0.1431552023,-0.001405832,-0.1848627776,-0.0017581935,-0.4116509855,-0.0602565482,-0.0149683412,0.1189694554,-0.0288363397,0.0215758551,0.0040093213,0.0507694446,-0.0561807789,-0.2149414569,-0.092203401,0.0877368078,-0.2132854164,0.2512765825,-0.2927488685,0.0120274713,-0.0775300264,-0.0794104338,-0.132429719,0.2365807742,-0.0912311897,0.4156896472,0.5673074126,0.2921690643,0.0902456716,0.1129869968,0.2711985409,0.0623779967,-0.2033884227,0.0321929269,-0.137654081,-0.4710271955,-0.1352291107,-0.0016211983,0.2588437796,0.1652996689,-0.2814908624,0.2102949172,-0.3044535816,-0.2140780091,0.1672844291,0.0760878399,-0.0154467616,0.4147460163,-0.0723187923,-0.2643076777,-0.047985442,0.5450500846,0.0525799543,0.2975494564,0.4717647731,-0.1705194265,-0.1695048362,-0.2284467071,-0.2198898792,-0.0372731276,-0.1842297763,0.1633380353,-0.0290565416,-0.3322977722,0.1018249914,0.253638953,0.0716210827,0.4311787784,-0.1206997335,-0.1742140353,-0.3888826072,0.486436516,-0.1069338024,0.3767164946,-0.258910507,0.012109925,-0.1251518428,0.1198547482,-0.3764693439,0.0512257926,-0.0563943163,0.2495509237,-0.0418037698,-0.1335870773,0.065922305,-0.2019754052,0.1232356727,0.2920111418,-0.3878259957,-0.2230171114,0.1272722334,0.1292952597,0.0517727323,0.0483963937,-0.118335411,-0.6028356552,-0.075404048,-0.2451005429,0.1326666772,0.1149459407,-0.0639121234,0.2569665015,0.3623216748,0.0208954792,0.3931995928,0.2849186659,-0.2523910105,-0.3889371753,-0.0620705336,0.0018400797,-0.1961840689,-0.0952343643,-0.088657774,-0.0619137511,-0.2578052282,0.1668848544,0.0387530625,0.0319139957,-0.2349714786,-0.0317918286,0.3383492827,0.6381781101,0.015067107,-0.4197871387,0.3746500015,0.2426559776,-0.2449926436,-0.1985408217,0.0005011793,0.2810861766,0.1674740314,-0.0375039466,-0.0309758335,-0.2451088727,-0.0940177739,0.3114634156,0.2527227998,-0.6664285064,0.2603012621,0.116674073,0.0619484,-0.1633707434,0.2422200292,0.1889887005,0.3079882264,0.3570047021,-0.0369509682,0.3194255829,0.4252534807,0.1109740585,0.0593798645,-0.2671722174,0.1472589225,0.5640743375,-0.1756101698,0.2904323637,0.2577305138,0.3387796581,0.0252396408,-0.0847704113,-0.1462690383,-0.0492820032,-0.1576877087,-0.0292049702,0.1933458149,-0.2965596616,-0.021637002,-0.053490527,-0.0778494999,0.2190318853,0.6823831797,0.0787296817,0.0228161421,-0.3085376918,0.2365974188,0.1222058386,0.3123135865,-0.2915606797,0.1540491283,0.1867463589,-0.092655234,0.3304531574,0.2505855262,0.1957370639,0.3371891379,-0.2278551161,0.0878952891,-0.0286824834,-0.1443091929,-0.2209825069,0.3758554459,0.3537641168,-0.1627030224,0.1452498734,0.1758390516,-0.1829490215,-0.0167656653,0.1404101253,0.2012847364,-0.1829661429,0.1995000094,0.2446160913,-0.0855713263,0.1097931042,-0.0666055903,-0.4104937613,-0.0136313532,-0.1664497256,-0.2242798507,0.0342133902,-0.0789882541,0.0903614983,-0.0353063978,0.3293505013,0.1339177638,-0.1580812186,-0.1349669993,-0.2358578891,-0.8640738726,0.181851998,0.0682395846,0.1568344831,-0.0889477655,0.3957232237,0.0190338418,-0.1654411703,0.0644150376,-0.0863238126,-0.0428871773,0.1791071594,-0.443593502,0.0902754888,-0.085021995,-0.1068328544,0.129518643,-0.2946932316,0.0467493832,-0.103476122,0.0325065292,0.0049843709,-0.0208948143,0.0711631104,0.4219408929,0.4397476614,0.041528482,0.2196604013,-0.1005732715,0.0349753052,-0.2901998162,-0.2234974951,0.0042940467,-0.2173172534,0.0332012326,0.4783435166,-0.1900581121,-0.1743460149,-0.230085507,0.3456980884,-0.250929147,0.1229722723,-0.3572003245,0.1816216409,0.0742473528,0.3774432242,-0.3227165639,0.0056157433,0.0328293815,0.1458384842,-0.2736541331,-0.3756458163,0.7129778266,-0.2250519693,-0.2088698149,-0.1358782053,0.4948709011,0.0398080572,-0.3910283744,-0.2560963929,0.0281631667,0.1448637545,0.0447650515,-0.2136390805,0.0972757041,-0.3613094687,0.1303618848,-0.1091749892,0.5722794533,-0.0723561719,-0.277222842,0.1097422168,-0.2119771838]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/514","title":"dataset.shuffle(keep_in_memory=True) is never allowed","comments":"Ok I'll investigate and add a series of tests on the `keep_in_memory=True` settings which is under-tested atm","body":"As of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.","comment_length":17,"text":"dataset.shuffle(keep_in_memory=True) is never allowed\nAs of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.\nOk I'll investigate and add a series of tests on the `keep_in_memory=True` settings which is under-tested atm","embeddings":[-0.1645395905,0.0470636711,-0.0645132363,0.1816313863,0.2262281924,-0.0658221468,-0.0984032601,0.3320508301,-0.0075618103,0.3399212062,0.0444224738,0.5816398859,-0.2912329435,-0.4286259413,-0.076349996,0.1383523792,0.043693874,0.0938878059,-0.0431256704,-0.1475760788,-0.348734051,-0.1134749576,-0.4357581139,-0.1966107786,-0.3342380524,-0.1588747054,-0.0248811655,0.1103741303,-0.2372039557,-0.3972377479,-0.0687004775,0.2789515257,-0.4423449636,0.4937436581,-0.0001106297,-0.2429226488,0.1797085553,-0.0578649975,-0.2435786724,-0.1217864305,-0.3575274646,0.1902090311,-0.0383543484,-0.1658546627,0.1627754569,-0.1657254994,0.062461216,-0.4391321242,0.2739268541,0.1919246465,0.2325711399,0.0483618416,-0.0108647291,-0.0708407983,0.5056999326,0.0983686373,-0.1865060776,0.2023808509,0.0994287729,-0.1826705188,-0.0003844972,0.2622644007,-0.1502898932,0.236128211,0.2072391212,-0.0669974536,-0.1666733474,-0.1825502366,0.3554884493,0.1823657453,0.2292565554,-0.4242457747,-0.223812148,-0.1083013117,0.006393515,-0.2964934111,0.2984439135,0.1853243411,-0.0228906292,-0.1757450849,-0.1930435896,-0.0560921468,0.1507157087,-0.0823626518,0.2069921494,0.2619947195,0.0504164807,0.0567099378,0.5261256099,-0.0970137864,0.1877243519,-0.0374600217,-0.2361270636,0.302949965,-0.4230674207,-0.0624217503,0.1713781357,0.087275736,0.220587641,0.0613539331,0.3008513153,0.0446182489,0.0128952106,0.2106889188,0.1390131712,0.3808430731,-0.072035715,0.1835352629,0.1963306814,-0.0748435408,-0.2012492716,0.049445685,0.0088462615,-0.274849087,0.08455275,-0.134238869,0.1005304381,-0.2576947808,-0.1026129425,0.1227495745,-0.1149030998,-0.1920341849,0.0423770733,0.603325665,0.1603025496,-0.1170068979,0.0255678259,0.0903490186,-0.0460837707,0.1373942345,-0.3694564402,0.0849426389,-0.3214630485,0.1383457333,0.2982785106,0.150287658,0.395069629,-0.0642243326,-0.0251956563,0.1381615847,0.2827377617,-0.2803497314,0.3846621811,0.2303172797,-0.2404924482,0.0470300131,-0.0196590833,0.1189517751,-0.1882035434,0.4020310044,-0.0872791409,-0.2613456845,0.3605814278,0.1778236181,-0.1003840268,-0.0343159847,-0.0103852917,-0.0915901288,0.412330538,-0.3571671546,0.2896513641,-0.1929020882,-0.0031734719,-0.3660291433,0.0472139865,0.1411512494,-0.1684119701,0.0470237285,-0.2436847091,0.0236433167,0.1621584594,0.4457203746,-0.0874056965,-0.3389571011,-0.2185922712,0.1965607107,0.22976771,-0.1487362832,-0.5433413982,-0.2355117053,-0.1417669803,0.0443009064,0.1115820706,0.0318915658,-0.0390054062,-0.1002301201,0.4659562409,0.0601362363,0.0381388813,0.1441458017,-0.3020870388,-0.1312883049,0.328831017,-0.1144824699,-0.2394287735,-0.1633696705,0.0918087438,-0.1207452714,0.3615780175,0.065369688,-0.0403922722,-0.0620871745,0.4602644444,-0.2890093029,-0.0927340314,-0.0882570371,-0.1043707728,0.1592429578,0.0940951928,-0.1410191357,-0.1503851563,-0.2046491057,-0.0333179273,-0.1032901332,-0.242335245,-0.3282055259,0.154852137,0.1008489504,0.1014075354,0.0617164709,-0.1063408852,0.086782366,-0.4196273088,-0.1925985664,-0.0756413564,0.1260682344,-0.0884511992,-0.2383362204,0.0605838597,0.2049890608,-0.0931285471,0.009616022,-0.3148371875,0.4407109022,0.1878857315,0.0636855885,0.0666090772,0.096437104,-0.1263602227,-0.1158674881,0.2639688253,-0.214144513,0.1109960377,-0.1265970021,-0.1125413924,0.2180500329,-0.2423674315,0.0968880281,0.2393686324,-0.2473141849,0.263948679,0.1885925233,-0.0416331217,-0.2077914625,0.2843483984,0.1691623032,0.4292958975,0.2045132369,-0.2278526723,-0.0814930797,0.7257265449,0.2631525695,0.1692700535,-0.0289141275,-0.0314215757,-0.2949008048,0.1342020482,0.3861592114,0.5165145397,0.2118197978,0.1496459842,-0.0296350438,0.0169992521,-0.0878011957,0.2908545434,-0.0056398436,0.1365726143,0.1303857714,0.2946321964,-0.2110142112,-0.5754868984,-0.1763487756,0.3066597581,0.2283549607,-0.0538047813,-0.2455100417,-0.3411333561,-0.0883902833,-0.0771433786,-0.11255797,-0.191927284,-0.1004242226,0.1275380552,0.3741232157,-0.1871351898,0.3885907233,-0.2175397128,0.3834332526,-0.0255779549,-0.1685042977,-0.2406041175,-0.1994387656,0.124864921,0.1316697896,0.1653927565,0.1222232431,0.3936085999,0.3040719926,-0.1854711622,-0.0635928363,-0.1819154322,0.0548923053,-0.1391015351,-0.0073849428,-0.0143501479,0.0617336519,-0.1076501533,-0.3901997209,0.0434611812,-0.3487757742,-0.2571064234,0.1572529078,0.222551465,-0.2505584061,-0.401353687,-0.4012826383,-0.2319550514,-0.4734946787,0.4063262045,-0.0250745174,0.121636413,0.4188084602,0.0301111676,0.1068951562,-0.3666160107,0.1007186696,-0.4784907699,-0.2979616225,0.1623972803,-0.2508962154,-0.1209593266,-0.044663284,-0.0964062363,0.2673754692,0.3345435262,-0.197452426,0.001811086,-0.4214383662,-0.0430031754,0.0390927419,0.1247504428,0.5532547832,0.1390643418,-0.1910766661,-0.1802101433,-0.1579673886,0.2839006484,0.410150975,0.180718109,-0.0001552659,-0.0088671837,0.2447098494,0.6376883388,-0.0267477818,0.2307050824,0.2228118628,0.0709281787,0.3748376369,-0.3868434727,-0.0213324372,0.3736646473,0.1293521821,-0.3258839846,0.1407595575,-0.2224761695,-0.353887856,-0.080568172,0.2327488959,0.0461789444,-0.1896999925,0.0882823244,-0.0363680534,-0.0494111702,0.1891163439,0.132914111,-0.191675365,-0.1098934337,-0.0942801759,-0.1530971974,0.188177377,-0.1836776584,-0.5609346032,-0.2704069614,-0.4111950994,0.3094627261,-0.0331860855,0.6306986213,-0.0860467702,-0.2974003553,0.1229010075,-0.0088541107,0.4743195176,-0.1237043217,0.0845308304,0.2729361355,0.1821281314,-0.1336542666,0.0130962236,-0.3547095656,0.0956472978,-0.2860817611,-0.1827217788,-0.2694420815,-0.0852590501,-0.083034277,0.007515247,0.036567349,0.0690898821,-0.3085597157,0.1819417924,-0.3635666668,0.1643828452,0.1514687985,-0.1472990066,-0.42720294,0.00782347,0.0576614402,0.0153032914,-0.0405130647,0.2390465736,0.6286554933,0.0407699756,0.1030655727,0.2285547107,-0.1970311701,0.2719873488,0.3310125768,-0.1418705285,0.2264735252,0.0044248374,-0.1267071664,0.2463165373,0.4111555517,-0.0600575171,0.2025131732,0.0487426408,-0.0015606472,-0.1360408217,-0.0408391915,0.3589111865,-0.1728655994,0.0083105126,-0.653093636,0.2220593989,0.3576476872,-0.1815082431,0.7226055264,-0.0072309859,-0.1178559959,0.3815122843,-0.0998145044,1.0839064121,0.0566599481,0.1469870508,0.1830813289,0.0449310206,0.0791685358,-0.0625474826,0.2574712932,-0.1825030148,-0.5199652314,0.1725640148,-0.2586373687,0.1509541124,-0.0197495595,-0.2885986269,0.2232110053,0.0881950632,-0.0452224426,-0.2266037166,0.0252872128,-0.3053609729,-0.0573657565,-0.0778177008,0.1446905136,0.0186364911,0.003321284,-0.0188237373,0.0113276299,-0.0692390203,-0.278013587,-0.0984087065,0.1322057694,-0.2228197306,0.1897662282,-0.2956101,0.0582380593,0.030688649,-0.076763384,-0.1113330722,0.2837205529,-0.0794102252,0.4060233235,0.5098747015,0.2248028219,0.0827048048,0.1829490662,0.2328514606,0.0048803524,-0.1908226311,0.0659305304,-0.1313794851,-0.4533767104,-0.1187157333,0.0302377325,0.2131251991,0.1467329413,-0.3811898232,0.242221415,-0.3799881339,-0.2048636973,0.1745130867,0.1023661718,-0.1270205379,0.4684289992,0.0316039585,-0.2463270128,-0.0167555474,0.464419663,0.0397440009,0.2125464082,0.4721181691,-0.1330167949,-0.1437510401,-0.3098315299,-0.3066723943,-0.0307117403,-0.2263919562,0.1833345741,-0.0311103836,-0.3075294495,0.0913114324,0.2305913717,0.11419525,0.4145865142,-0.1433431804,-0.2366756797,-0.3798678517,0.4347459078,-0.0309177041,0.3997690976,-0.2919093668,0.1168431342,-0.1029943749,0.2220922112,-0.3882456124,0.0621822588,-0.0940001234,0.341976136,-0.1091265753,-0.1284656227,0.0838779062,-0.1945869923,0.1724512726,0.3311291933,-0.3868030012,-0.2547594011,0.1303526312,0.1155683324,0.0742471591,0.0859083831,-0.1114682257,-0.5825694799,-0.1071042791,-0.2249593884,0.1532945633,0.1011646315,-0.0949433222,0.2292374521,0.4127990305,0.000214695,0.3605824113,0.26484707,-0.2013716698,-0.3913645744,-0.0288558714,-0.0226959456,-0.2145078778,-0.1246571392,-0.087004967,-0.0855234638,-0.2674266696,0.2091461122,0.0080301696,0.011076618,-0.2223532498,-0.0892885849,0.4202598929,0.6356108189,-0.0435015783,-0.3844485581,0.370919615,0.2683918476,-0.3369072378,-0.1750692278,0.03922607,0.2460407764,0.2021853775,0.0122515811,-0.1007213145,-0.2154272944,-0.1817184091,0.2678736448,0.223695159,-0.6251196265,0.2195984125,0.1022555754,0.0065356418,-0.1585647613,0.19507581,0.2338111401,0.2964687943,0.3898302913,-0.0963084251,0.3289062679,0.4681845903,0.0587382615,0.1087260395,-0.2375671417,0.128678903,0.539804101,-0.078689456,0.2875324786,0.2556869984,0.3370885253,0.0629014075,-0.1188621894,-0.181439206,-0.054375127,-0.1782623231,0.0336606242,0.2041493505,-0.2808974981,-0.0744928196,-0.0966938138,-0.0654539689,0.1876657605,0.6637601256,0.1149747595,0.0587375388,-0.3534989655,0.2141942531,0.124332957,0.2556313872,-0.2843035161,0.1898604929,0.1373907328,-0.0432813503,0.3281436861,0.2436445653,0.1775020361,0.3849610686,-0.2102647275,0.1273862123,-0.0378346927,-0.1224706545,-0.2118448913,0.3429830074,0.4248510301,-0.1512331367,0.194469437,0.1775666475,-0.1910857111,-0.0204143506,0.1460638642,0.1708217114,-0.1709228754,0.1695912778,0.1908321232,-0.1284255385,0.0399029814,-0.0729861706,-0.4594965577,0.0175565463,-0.1854106337,-0.2413070202,0.0444920398,-0.0420019999,0.0901077762,-0.0942573249,0.3228235841,0.121725224,-0.0727368146,-0.1752489358,-0.2853651047,-0.8999342918,0.212852791,0.0689447001,0.0817750096,-0.1407223493,0.4141230583,0.0608885698,-0.2042655349,0.116466485,-0.0787590668,-0.0535233207,0.2270152271,-0.4681152999,0.1060831994,-0.083152771,-0.0797097236,0.1405967474,-0.2543372512,0.0337202288,-0.0785126835,0.0557509847,0.0533094667,-0.0191892199,0.0217641201,0.3450400829,0.4877483249,0.0153716886,0.2486279309,-0.1103384197,0.0711962879,-0.2749643326,-0.2630705535,0.0476426221,-0.2343212366,0.1461323351,0.4901224375,-0.115140453,-0.1558482647,-0.2355591059,0.2968493402,-0.2419131547,0.1062014252,-0.3029259443,0.2135644853,0.0751415491,0.3843677342,-0.3215219676,-0.0135815106,-0.0700051337,0.0796152428,-0.2733808458,-0.3939669132,0.7890161872,-0.2005458474,-0.2381666005,-0.1327884793,0.5532743931,-0.0033606619,-0.4281305671,-0.325409174,-0.0058574746,0.1068147197,0.0033221201,-0.2388474643,0.0915663689,-0.3629478216,0.0831769928,-0.0804674104,0.6088451147,-0.0973442867,-0.2683152556,0.0887286216,-0.2417260259]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/514","title":"dataset.shuffle(keep_in_memory=True) is never allowed","comments":"Hey, still seeing this issue with the latest version.","body":"As of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.","comment_length":9,"text":"dataset.shuffle(keep_in_memory=True) is never allowed\nAs of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)`\r\n\r\nThe commit added the lines\r\n```python\r\n# lines 994-996 in src\/nlp\/arrow_dataset.py\r\n assert (\r\n not keep_in_memory or cache_file_name is None\r\n ), \"Please use either `keep_in_memory` or `cache_file_name` but not both.\"\r\n```\r\n\r\nThis affects both `shuffle()` as `select()` is a sub-routine, and `map()` that has the same check. \r\n\r\nI'd love to fix this myself, but unsure what the intention of the assert is given the rest of the logic in the function concerning `ccache_file_name` and `keep_in_memory`.\nHey, still seeing this issue with the latest version.","embeddings":[-0.1523228437,0.0339478776,-0.0342748053,0.1510977745,0.1926056594,-0.0610219948,-0.135276258,0.3430135846,-0.0620456748,0.3255017102,0.0727649555,0.5818741322,-0.2995440364,-0.4434979558,-0.0462150611,0.1174887568,0.0790202767,0.053681504,-0.0152773503,-0.1284914613,-0.4000428915,-0.1399145275,-0.5022443533,-0.2024136037,-0.3272928894,-0.1066185459,-0.0746184215,0.0748594776,-0.2364875227,-0.4599909186,-0.0240407307,0.2838071287,-0.448526293,0.4509904385,-0.0001137035,-0.2428496033,0.259752959,-0.0400044061,-0.2704867721,-0.1955441833,-0.3517667949,0.1206050813,0.0015930043,-0.1001374573,0.208542794,-0.160144642,0.1233663261,-0.3773526549,0.2411485463,0.2169632614,0.2007220089,0.0904059783,-0.0087698577,-0.0907543972,0.5111327767,0.0963181481,-0.2116031051,0.1890374571,0.1623952687,-0.2321925908,0.0466896147,0.2401898205,-0.1758620739,0.2119930983,0.2218020409,-0.1177992672,-0.1893938184,-0.1611664742,0.3521120548,0.2186971456,0.2838217616,-0.3542183936,-0.2634583712,-0.1350060254,0.0199068002,-0.2959308028,0.3419173658,0.2008268833,-0.0197335873,-0.2188702673,-0.1846728176,-0.0664375275,0.1965449005,-0.0276694205,0.214738667,0.2922666073,0.0930605903,0.0706244558,0.5875075459,-0.0949660391,0.2055861056,-0.0087993229,-0.2543205023,0.359333396,-0.4376538098,-0.0451877862,0.1401899159,0.0613788813,0.2175680846,0.0860849619,0.2137533873,0.0137980599,0.005041765,0.2205571979,0.1762703806,0.3477258086,-0.0242539402,0.1956952363,0.2332122177,-0.0687558129,-0.1190759838,0.0791163072,0.0188619308,-0.2875858843,0.0422347561,-0.1583059579,0.121968247,-0.2386939675,-0.1071417034,0.0932660922,-0.1511609256,-0.1983094811,0.0409382284,0.5278738141,0.231161356,-0.0613679588,0.0376946814,0.0994711146,-0.0358398296,0.1054847836,-0.3571704924,0.0566590428,-0.3289223015,0.1080122665,0.270626992,0.0872269347,0.3897396624,-0.0677259266,-0.0655702353,0.1076240316,0.203519702,-0.2802420259,0.3659576774,0.2292896062,-0.2842022479,0.0416417569,-0.0224965923,0.1175699383,-0.2090827376,0.4473838806,-0.0905299634,-0.2553617954,0.2805505097,0.1263212115,-0.0353449695,-0.0864230916,-0.0176328477,-0.045697391,0.4209060073,-0.3810840249,0.2975797653,-0.2562085986,-0.0331429802,-0.3538597822,0.000572684,0.169678539,-0.2186419964,0.0424177088,-0.2883758843,0.0245569497,0.147283867,0.4558100402,-0.0763456449,-0.3804713786,-0.2105561495,0.1303927898,0.2415380776,-0.1398259401,-0.5529176593,-0.2419739813,-0.082620807,0.0301344655,0.1207709238,0.0470111445,-0.0735914409,-0.1391841322,0.4504775405,0.0730063394,0.0366889797,0.1229384392,-0.2851240039,-0.16177544,0.350148052,-0.1282299757,-0.2033945471,-0.1295629144,0.1091976687,-0.0902116224,0.372451514,0.0816698298,-0.065778397,-0.1148225218,0.5087441206,-0.2761410475,-0.0880227387,-0.1033044383,-0.1779095531,0.1717288792,0.1042102054,-0.1944554746,-0.1027309522,-0.186409533,-0.0191149823,-0.0761979595,-0.201859802,-0.3225047588,0.0775418952,0.087070331,0.0810527951,0.0876541808,-0.1546484977,0.0788050592,-0.4157167077,-0.1548125893,-0.0821588784,0.1792559475,-0.063744083,-0.2594814301,0.0559999123,0.2023085505,-0.0702371746,0.0287213027,-0.3243327141,0.4582817554,0.2021447271,0.0681334659,0.0419614501,0.0764612183,-0.0983850211,-0.0747128353,0.2502643466,-0.2000463754,0.0913555548,-0.1173189133,-0.1071683764,0.2105637342,-0.2357544303,0.1104635298,0.2167244554,-0.1977429092,0.2551853359,0.1924228668,-0.0328593031,-0.2152446061,0.325850606,0.2094761133,0.4402759671,0.2046201229,-0.2793903947,-0.0584812164,0.7611732483,0.2384541184,0.1763844192,0.0588796586,0.0152858552,-0.2919085026,0.1336955279,0.4024159014,0.4868173003,0.1555770785,0.1036976352,0.0257126614,0.0026808141,-0.1044200957,0.324460417,0.0093867807,0.1416457444,0.1130372509,0.3175455332,-0.209796086,-0.5585438013,-0.1362612098,0.3189341724,0.2285263836,-0.0461947024,-0.2156600654,-0.4359963536,-0.0923957378,-0.1327927411,-0.1201428771,-0.2868597209,-0.1623802036,0.1078386977,0.4112693369,-0.1793343127,0.4007072449,-0.276303798,0.456979543,-0.0333915055,-0.1323144287,-0.2382488549,-0.1784982681,0.0964083001,0.082038857,0.2115966976,0.0647570342,0.3535111547,0.3214106262,-0.1523463279,-0.0358725041,-0.2237986922,0.033532314,-0.1450598985,0.014280091,-0.0018231071,0.0015395263,-0.0859372392,-0.3589426577,0.0100798225,-0.4018694758,-0.238302812,0.1663739681,0.224153921,-0.2518361807,-0.3636648357,-0.3613270819,-0.2504192591,-0.4683120847,0.425293386,-0.0366797261,0.1553679109,0.4318975806,0.02744844,0.104342632,-0.3891735673,0.0103096534,-0.4546596408,-0.2552247941,0.1748842448,-0.2495827824,-0.0986717418,-0.0462999158,-0.0581564866,0.2457109839,0.3281246424,-0.1846050024,-0.0101480596,-0.458484441,-0.0360014066,0.0592043772,0.1351396739,0.5508520603,0.1246870682,-0.1678165942,-0.1834421307,-0.2020700425,0.2398934811,0.436304599,0.2045953125,0.0568298995,0.0321584158,0.3249320984,0.677400291,0.0171736851,0.2548112571,0.2403763384,0.09893585,0.4087170064,-0.4331912994,-0.0504418127,0.3493139148,0.1355533302,-0.2981446683,0.1082138494,-0.201066643,-0.3647350073,-0.0584440455,0.2224804759,0.0846283659,-0.1530602425,0.0843588784,-0.0898496583,-0.0680631399,0.2231617272,0.1578363627,-0.16475977,-0.1897262484,-0.1305512637,-0.1317124516,0.2131184936,-0.1253233254,-0.5683355927,-0.3025104403,-0.3843923211,0.3587282002,0.0077680307,0.716399014,-0.0925237685,-0.2888040841,0.1498601884,0.0160963126,0.5108730793,-0.0932937339,0.0652377903,0.2708751559,0.2064561397,-0.1678159088,-0.0069336849,-0.3076793253,0.1295476109,-0.2743956745,-0.1842459738,-0.2179804295,-0.1066235453,-0.0174835976,0.0133545501,0.0285961647,0.1179598644,-0.3406065404,0.1919500232,-0.3427577615,0.1778108627,0.162332356,-0.1747166812,-0.4698042572,0.0317597054,0.0598838218,-0.0519057103,-0.0918224007,0.1499222815,0.6042638421,0.0842968896,0.0794821382,0.2757429481,-0.2625181079,0.2678204477,0.3162001073,-0.1359219253,0.1941882968,-0.0001735234,-0.0907358974,0.3267127275,0.4540135264,-0.0492136888,0.2430979908,0.0559400581,-0.0458652265,-0.139558658,-0.1430489421,0.3889396489,-0.2146271765,0.0101498803,-0.6707347035,0.2762897313,0.3643588722,-0.1535010934,0.7643755674,0.028103061,-0.1020487249,0.3556153178,-0.1063872054,1.1211606264,0.0258130636,0.1256908774,0.1596155316,-0.0043916488,0.1172738671,-0.0135871945,0.3224063218,-0.2130192667,-0.5085250735,0.1628716439,-0.2738621533,0.1819158942,-0.0793801248,-0.2793267667,0.2721919417,0.0458343774,-0.0756527856,-0.1505372226,0.0042965598,-0.3476036489,-0.0240673255,-0.056309294,0.1081715673,-0.0107844044,0.0170135684,-0.0361415967,0.0405835994,-0.0278729126,-0.2842169702,-0.1301520616,0.1582309157,-0.1363284588,0.2298943251,-0.2055757791,0.0160835404,0.0426659994,-0.0999352857,-0.1288579404,0.2672466338,-0.0645720288,0.4335857332,0.4975020587,0.2277266532,0.0747373477,0.2080104947,0.2768876851,0.026823869,-0.2248021513,0.0345282964,-0.1238028109,-0.5388119221,-0.178457886,0.0204367414,0.2164949626,0.2068653554,-0.3735930622,0.2313926667,-0.3312829733,-0.2423234433,0.1336107254,0.1197270527,-0.0845568553,0.4332300723,-0.0095465258,-0.2326609045,0.0191134401,0.5020785332,0.058587525,0.2147290111,0.5386371613,-0.1106167957,-0.1566289216,-0.2457644194,-0.2554166913,-0.0507159829,-0.2370631993,0.2375958562,-0.0838666558,-0.293289572,0.0997491181,0.2747478187,0.0465968587,0.3615556657,-0.1312487572,-0.271863699,-0.3518284261,0.4678418934,-0.0877849087,0.3895817399,-0.2922976911,0.0941633582,-0.0807291046,0.2089201063,-0.3403148055,0.1033464149,-0.0456969365,0.3043443561,-0.0999647304,-0.1380433887,0.0574482232,-0.1807541102,0.1120159477,0.3410528302,-0.3929506242,-0.1955540478,0.0874767527,0.1476004124,0.1394308656,0.074552767,-0.1463611424,-0.6262488961,-0.0897290781,-0.2344642878,0.1293613166,0.1516275555,-0.093420133,0.2782840431,0.4359214306,0.0185726508,0.4053817689,0.2934599221,-0.1795536578,-0.3969342709,-0.036059238,-0.0113727851,-0.1850249916,-0.0867975503,-0.1468539834,-0.070751749,-0.272205919,0.2144118249,0.0021164098,-0.0204208866,-0.2396029085,-0.1220889688,0.4061265886,0.7123479843,-0.0289094523,-0.3819673657,0.4181921184,0.2048318088,-0.3366698027,-0.2132159919,0.0299866647,0.332080394,0.1233353019,-0.0265859663,-0.0423193052,-0.2569596171,-0.1488041878,0.2785340548,0.1979430318,-0.7157920599,0.2831477523,0.1322887838,0.0587160066,-0.159686178,0.2049797028,0.2281685024,0.3020779788,0.3666253686,-0.0500923693,0.3059238493,0.4769910276,0.0402668528,0.1446726322,-0.2815288603,0.1143624932,0.5387769341,-0.0664096251,0.2812642157,0.244680807,0.3454099596,0.0562110953,-0.1697476059,-0.2477837503,-0.0825507343,-0.1259756535,-0.0356952809,0.2152745724,-0.2906943262,-0.0603908896,-0.0935773104,-0.087406382,0.1444812715,0.6592633724,0.1226857081,0.0448960029,-0.4028156996,0.2300040573,0.1277373731,0.2612426579,-0.2904290259,0.1764186323,0.0867687166,-0.0512931384,0.3670182228,0.2784990966,0.180267334,0.3742346466,-0.2082800716,0.1063422784,-0.0788624138,-0.1257859021,-0.1450229585,0.3869370222,0.4001320004,-0.1100248545,0.1338261366,0.1298683584,-0.1429035962,0.0549935773,0.1743365228,0.1933814883,-0.2395916879,0.2564083934,0.2287568003,-0.1247138679,0.1150289029,-0.0646068677,-0.4569514394,0.0078606643,-0.1690598428,-0.2123014331,0.0144393472,-0.0521020778,0.0743694976,-0.0315297246,0.2818236947,0.182795614,-0.119732298,-0.1045880914,-0.2514837384,-0.8800290823,0.1997799426,0.0627546161,0.135307312,-0.0904738754,0.4021145999,0.021717824,-0.2124950886,0.1316484213,-0.159424901,-0.0323209018,0.2042614669,-0.4245313406,0.0600409508,-0.0586967468,-0.1281222701,0.1189167798,-0.2664861381,0.0399364494,-0.1581790149,0.0071571665,0.0432903804,-0.0430367738,0.0482696071,0.3790184557,0.4789620638,0.0235074647,0.2259089798,-0.0525604375,0.0418192334,-0.3906812668,-0.2969523966,0.0246264376,-0.2409005463,0.0434155986,0.4724800587,-0.1886805296,-0.1245852113,-0.266938597,0.3293886483,-0.3191959262,0.1335234344,-0.3065262437,0.2006523609,0.0196535904,0.3947871625,-0.2910099626,0.0128399907,-0.0714826733,0.1238861978,-0.3022344112,-0.3576339483,0.7856059074,-0.2647389472,-0.2542292476,-0.1567199975,0.5325220823,0.0140522141,-0.4511745572,-0.3033657074,-0.0438832678,0.1266727448,0.0193205811,-0.2119210064,0.0866604224,-0.3577595651,0.0882198066,-0.0911996439,0.6190120578,-0.1114943251,-0.3096784949,0.1415411383,-0.2388024181]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/511","title":"dataset.shuffle() and select() resets format. Intended?","comments":"Hi @vegarab yes feel free to open a discussion here.\r\n\r\nThis design choice was not very much thought about.\r\n\r\nSince `dataset.select()` (like all the method without a trailing underscore) is non-destructive and returns a new dataset it has most of its properties initialized from scratch (except the table and infos).\r\n\r\nThinking about it I don't see a strong reason against transmitting the format from the parent dataset to its newly created child. It's probably what's expected by the user in most cases. What do you think @lhoestq?\r\n\r\nBy the way, I've been working today on a refactoring of all the samples re-ordering\/selection methods (`select`, `sort`, `shuffle`, `shard`, `train_test_split`). The idea is to speed them up by a lot (like, really a lot) by working as much as possible with an indices mapping table instead of doing a deep copy of the full dataset as we've been doing currently. You can give it a look and try it here: https:\/\/github.com\/huggingface\/nlp\/pull\/513\r\nFeedbacks are very much welcome","body":"Calling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```","comment_length":164,"text":"dataset.shuffle() and select() resets format. Intended?\nCalling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```\nHi @vegarab yes feel free to open a discussion here.\r\n\r\nThis design choice was not very much thought about.\r\n\r\nSince `dataset.select()` (like all the method without a trailing underscore) is non-destructive and returns a new dataset it has most of its properties initialized from scratch (except the table and infos).\r\n\r\nThinking about it I don't see a strong reason against transmitting the format from the parent dataset to its newly created child. It's probably what's expected by the user in most cases. What do you think @lhoestq?\r\n\r\nBy the way, I've been working today on a refactoring of all the samples re-ordering\/selection methods (`select`, `sort`, `shuffle`, `shard`, `train_test_split`). The idea is to speed them up by a lot (like, really a lot) by working as much as possible with an indices mapping table instead of doing a deep copy of the full dataset as we've been doing currently. You can give it a look and try it here: https:\/\/github.com\/huggingface\/nlp\/pull\/513\r\nFeedbacks are very much welcome","embeddings":[-0.1382125914,-0.2247202247,0.0239643287,0.2476742417,0.1281202286,-0.1362012476,0.2032122612,0.0824569091,-0.6032932997,-0.0022735526,-0.2842715979,0.322750628,-0.1951235682,0.1738311946,0.1398987323,-0.1447517425,0.1145583838,0.2089668661,-0.0435798876,-0.115610823,-0.1563792527,-0.2866491675,-0.2859784365,-0.0606167801,-0.4345372915,0.0260992981,0.0050066882,0.1479002982,-0.0802040696,0.0760351792,0.0442324206,0.2561855018,-0.1987256706,0.5789396763,-0.0001289723,-0.1655297279,0.0554003008,-0.2835808396,-0.292496562,0.0555677451,-0.0595302619,0.2450539023,0.0335684307,-0.1791678071,0.0966271758,-0.3774504662,-0.0920630097,-0.1147696525,0.189046368,0.1550750881,0.0798439384,0.3368454278,-0.2491133064,0.2438392788,0.3745867014,0.557667613,-0.1313131005,0.2056202143,0.101867266,0.3548139632,0.4095387161,0.1268223077,-0.2900162339,-0.2419890761,0.2838993073,0.0142173441,-0.2483233958,-0.2783831358,0.1692030579,0.0764759481,0.7366990447,-0.5577983856,-0.1140442863,-0.430565536,0.1685490608,-0.3902668655,0.071951814,0.154959783,0.0929495022,-0.0118967332,-0.4119156003,-0.0944790542,0.1351483017,0.0751213357,0.2118790597,0.2510275543,-0.1262820661,0.2214883268,-0.2606827915,0.3279381096,0.2481326014,-0.3805216253,-0.0002255971,-0.0308669601,-0.1849082708,-0.0945181474,-0.0433808714,0.045279514,-0.0704916492,0.3731215,0.5257031322,0.2177648246,-0.1684404016,0.1170973629,0.375684768,0.1648922712,0.1534628421,0.062367551,0.0424035825,-0.3042173982,0.0391020514,0.0110796252,0.2551579475,-0.0433039553,0.3766871095,-0.0346268825,-0.0403555222,-0.1273494214,-0.0218834579,0.0662864149,-0.2807991803,-0.3609564006,-0.1409435421,0.111996375,0.0275993235,0.334613651,-0.2615431249,0.2003477812,-0.3364670575,0.0672368109,-0.1536249816,-0.1003374457,-0.2305383235,0.1230605617,0.1779459715,-0.341763556,0.2291339785,0.085795626,-0.055318255,0.0410853773,0.1837367415,-0.3757719398,0.4564963281,0.23024863,-0.1737115979,0.1937036365,0.0566022396,0.2335437834,-0.2558582723,0.497835964,-0.241155833,-0.2049873471,0.2311902791,0.0342101119,-0.1543439478,-0.0228702538,-0.208795324,0.1444380581,0.4477293193,-0.3578275442,0.2071652412,-0.5868683457,-0.2272828221,-0.4626278877,0.0286326837,0.2016877979,-0.3084818423,-0.0998898596,0.5335569382,-0.1290865541,0.2292512059,0.2242171019,-0.3500424922,0.3968575001,-0.1544257998,-0.063399151,0.5925039053,0.0699077174,-0.2872909904,-0.3488504887,-0.0791895986,0.3577805459,-0.1115627959,0.1525214911,-0.1749596149,-0.0800708011,0.2018082291,0.11498449,-0.0569889396,0.1131259128,-0.2052199543,0.0311108716,0.4739532173,0.1160250679,-0.085915342,-0.0094829211,-0.0761847273,0.158943966,0.1887992769,-0.0295025203,-0.2351531982,0.0499477834,-0.1133157834,0.1867135465,-0.0154641299,-0.1854670942,-0.2254036516,-0.1526828557,0.1799524128,-0.3040947616,0.3576213419,0.0391091369,0.010645817,-0.332868278,-0.4570793509,-0.3931130469,-0.0499445647,0.4065027535,0.0490477979,0.1186681837,-0.3195305169,0.0614476018,-0.2063674182,-0.0083991289,-0.4115723372,0.1248530522,0.1170983836,-0.2123429328,-0.3896256983,0.1242080256,-0.1243458912,-0.0857831985,-0.2129511684,0.3616207242,0.1442990601,-0.2647690475,-0.3784194589,0.4010251462,0.1274124384,0.2065180242,0.1345853209,-0.1895992756,0.3265334368,-0.1915585399,-0.2690972388,0.3493041098,-0.6575779319,-0.1802589595,-0.1800181866,-0.6270247698,-0.0393623821,0.2436851114,-0.38536641,0.045983322,0.0531655848,0.3270132244,0.1938665211,0.3799907863,-0.5700509548,-0.0476526283,0.7285046577,-0.2212718576,0.0553403497,0.0533638261,-0.1717451513,-0.3505660892,-0.1845951527,-0.0203250516,0.782093823,-0.0153637482,-0.0367374867,0.0553940795,0.0814553723,-0.3519118428,0.0955269486,0.2491502464,0.00695509,0.4411150217,0.3575190604,0.1117630824,-0.488229394,0.1302794069,0.1637688428,-0.0803554654,-0.3349176943,0.0944054872,-0.155654043,0.0596321747,-0.3772047162,-0.0254960172,0.0446690954,-0.0599994063,-0.081163533,0.0489457361,0.1002306193,-0.0627609044,0.2464330047,0.224451378,-0.0175593719,-0.3098870516,0.3108996451,-0.0712586492,-0.1993415505,-0.1147271618,0.0123944832,-0.0296001565,0.3543699086,-0.0558917895,-0.2055740356,-0.1391870677,-0.1954615563,0.2927092016,-0.0928981304,-0.0006683394,0.0193675347,-0.0350417793,0.0142719653,-0.1980372071,0.1020730063,0.1147153229,0.1975731552,0.0408127047,0.1796193719,0.1345437765,-0.1993232816,-0.291289866,-0.0313384794,0.088165395,0.2162626535,-0.1674561948,-0.0139829572,0.465908438,-0.0845903978,-0.1470181197,-0.6666821837,0.1076658294,-0.3695809245,-0.5208802819,0.129328534,-0.0555758886,-0.2137968689,-0.2652085423,0.1450389624,-0.0572913438,0.5463920832,-0.2382081747,-0.1953873187,-0.1374084204,-0.0648620725,-0.0691306964,0.4167664349,0.3020017147,0.0571569018,0.1945846677,0.038517151,-0.043403767,0.1414676905,0.3747738302,0.0798519477,0.0951740593,0.1910053194,-0.0178647935,0.5995121002,-0.1991274357,-0.1613339633,0.1348063946,0.1795824766,0.3882859349,-0.4722309709,-0.104178369,0.4822199345,-0.0516313165,-0.0903253704,0.2900754213,-0.2437630743,0.0153201325,-0.1181699112,0.1589726955,-0.2877340019,-0.2249055505,0.4279428422,-0.0042330185,0.2148872763,0.0903220549,0.3507176936,-0.1093763113,0.0532986932,-0.1550857127,0.1416816264,0.404302299,0.0007882579,-0.5444186926,-0.2212259918,-0.1243489757,0.0089226775,0.1123208329,0.5317481756,0.0306649636,-0.1736357659,0.1770093888,0.2656615376,0.5368555784,-0.0887493938,-0.0582991466,0.1731205434,-0.3876774609,-0.1465545893,-0.3178415,-0.2131336778,0.0677269697,-0.2510716021,0.263696909,-0.203009069,-0.0140729835,-0.0744816959,0.4079804122,-0.1185537726,0.0340827964,-0.0952488631,-0.0083027743,-0.1466766298,0.0440276712,-0.1954734176,-0.1743855923,-0.3775610924,0.3725061119,-0.0666091517,0.2309618145,-0.0352425128,-0.1375807226,0.265242964,0.1389311254,-0.1037343889,0.1086483896,0.0686223432,0.0677358881,0.0991382301,-0.0500395894,-0.0214681067,0.2940168083,-0.343662709,0.1544417888,0.4693882167,0.0623107404,0.0337494723,-0.420491457,0.0856411383,-0.2320627272,0.0723931193,0.347905755,-0.080632478,-0.2999900579,-1.069737196,0.2451432347,0.1815061867,0.0347042903,0.1759747416,-0.2191982418,-0.0421901494,0.2608632743,0.1543843597,1.0892894268,-0.0507256314,0.5218431354,0.3229362369,-0.0731944144,0.3819906414,-0.3416105509,0.1129657924,-0.244715184,-0.4424582422,0.0531473085,-0.2846353948,-0.0072025433,0.0078241881,-0.1367422193,0.3234067857,-0.2796405852,-0.0524001606,-0.0252828151,0.0340554975,-0.1102772057,-0.0948151127,-0.0633685216,-0.0598191917,-0.1997699589,0.1812173277,-0.0717719346,0.1181937978,-0.1108608022,0.0298046488,-0.04482181,-0.1666142493,-0.0512364395,0.0092817806,-0.3341157734,-0.6155846715,0.283614248,0.3402102292,0.0996392891,0.179265514,0.053441681,0.0836941898,0.2451819032,0.4168071449,-0.0029478443,-0.0521088839,0.4235515296,0.0652486905,0.0677940995,0.2524892688,-0.019039575,-0.096982412,-0.8286117911,0.1709132344,0.3791937232,-0.2473125607,-0.4441283047,0.0542826056,-0.0023440337,0.0265491717,-0.0408351608,0.1820113659,0.1740788072,0.4315118194,-0.2256633192,-0.4154630899,-0.1995775253,0.2852520943,0.2003929466,-0.045226451,0.4193188846,-0.0519285202,-0.2008045614,-0.0527363978,0.2371236384,-0.0974297449,-0.4542705119,-0.1012882963,-0.0841003731,0.1255544573,-0.019005673,0.2705136538,-0.2472783327,0.0637085661,-0.2091897577,-0.0291788038,0.1038118005,0.0176787004,0.4277179837,0.3437690735,0.3647261858,0.3129191101,0.0152425794,-0.0564862154,-0.1305907518,0.0297765024,0.3241150379,0.3977801502,0.0183637086,0.0064171935,-0.3131085336,-0.0604932867,-0.1109374538,0.1440346837,-0.1140713394,-0.0874803364,-0.1520568579,0.2300266922,0.0748610124,-0.0025268281,-0.169527486,-0.1676239073,-0.1742764413,-0.2396644801,0.2976013422,0.2611989975,0.0210243277,0.1257654279,0.4362260401,0.0285326708,0.1622669399,0.1904217899,-0.158964932,-0.2615143061,0.1234544814,0.0164786372,-0.2788673341,-0.0519074686,-0.599006176,0.143350035,0.2141147405,0.3800849319,0.0123016732,-0.3274235427,0.3317849338,-0.1678716242,0.4022538066,0.5396655202,0.0275575705,-0.4948884547,0.2594493628,-0.0720313266,0.2253415436,-0.1419828832,-0.0395640843,0.1817880869,-0.1139530316,0.1323440224,-0.1584558636,0.3500945866,-0.5146568418,-0.0992891118,0.1379054487,-0.1613038331,0.2236180604,0.4979375005,0.046598129,-0.0668177158,0.1828082502,-0.0771138519,-0.1445321739,0.2808187902,0.7301182747,0.5393793583,0.4825175703,0.0395526029,-0.0672645941,-0.4697543085,-0.2155254185,0.1976953298,0.0375333838,0.5161758065,0.2281161696,0.0463414192,0.1134193763,-0.0799599364,0.1270460784,0.336656481,-0.1223196089,-0.042072054,0.2104701251,-0.3991296887,0.0812123939,0.1484832466,-0.4271040857,0.2103952616,0.1051131934,0.170705542,-0.1856763512,-0.111417748,0.2426232994,-0.1366685182,0.1284887642,-0.1080194265,-0.0554121286,0.1767769009,0.0271324012,0.1898210347,0.1285141706,-0.051371187,0.3518345356,-0.1466553062,0.0571049117,-0.103277728,0.1627274603,0.0270891506,0.0942607075,0.3115408123,-0.3098682165,0.4923088849,0.0618408546,-0.0126401251,-0.1264349073,0.3408855796,0.2179021537,-0.5887755752,-0.0272438154,0.1398259848,0.0593104847,0.0964889973,0.1066811383,-0.4906990528,0.0379791483,0.0147474837,0.4293690026,0.1835746914,-0.0085616251,-0.0065229358,0.1072581634,0.4353663921,0.0826549903,-0.1552221775,-0.0474484824,0.0259117559,-0.428497076,0.0506531,0.3247933388,0.1212253124,0.2492528111,-0.1100715548,0.2826694548,0.1380500942,-0.1708900332,0.217368871,-0.0020115608,0.4330170155,-0.3053794503,-0.0873088166,-0.41097036,0.1781348139,-0.1709602028,-0.1909289807,0.1201006249,0.1029169634,0.0281698108,-0.2006617785,-0.1350136995,-0.0147139719,0.3194841444,0.2151916772,0.0709726214,0.2752601504,0.0149041666,0.0139682246,-0.1395524144,0.0865887702,-0.0084388126,0.2101175189,-0.2170508057,0.39561975,-0.1571722031,0.1808829904,-0.3153145313,0.2677330971,-0.1786733717,0.2012965232,-0.2949556708,-0.061833445,0.322712332,0.0421419963,0.0804623365,0.5686163306,-0.0181748904,0.2659771144,-0.2791575789,0.2963319719,0.4934827983,0.1330400407,-0.2201577425,-0.2162210643,-0.0268434361,0.1314427555,0.0620843135,-0.3097263277,0.0052932235,0.2125545293,-0.2343695909,-0.3316029906,0.2053728402,-0.054333806,-0.2618587911,-0.1250595897,0.5423796773,-0.0547792353,-0.0949317217,-0.0436527915,-0.03871084]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/511","title":"dataset.shuffle() and select() resets format. Intended?","comments":"I think it's ok to keep the format.\r\nIf we want to have this behavior for `.map` too we just have to make sure it doesn't keep a column that's been removed.","body":"Calling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```","comment_length":32,"text":"dataset.shuffle() and select() resets format. Intended?\nCalling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```\nI think it's ok to keep the format.\r\nIf we want to have this behavior for `.map` too we just have to make sure it doesn't keep a column that's been removed.","embeddings":[-0.1382125914,-0.2247202247,0.0239643287,0.2476742417,0.1281202286,-0.1362012476,0.2032122612,0.0824569091,-0.6032932997,-0.0022735526,-0.2842715979,0.322750628,-0.1951235682,0.1738311946,0.1398987323,-0.1447517425,0.1145583838,0.2089668661,-0.0435798876,-0.115610823,-0.1563792527,-0.2866491675,-0.2859784365,-0.0606167801,-0.4345372915,0.0260992981,0.0050066882,0.1479002982,-0.0802040696,0.0760351792,0.0442324206,0.2561855018,-0.1987256706,0.5789396763,-0.0001289723,-0.1655297279,0.0554003008,-0.2835808396,-0.292496562,0.0555677451,-0.0595302619,0.2450539023,0.0335684307,-0.1791678071,0.0966271758,-0.3774504662,-0.0920630097,-0.1147696525,0.189046368,0.1550750881,0.0798439384,0.3368454278,-0.2491133064,0.2438392788,0.3745867014,0.557667613,-0.1313131005,0.2056202143,0.101867266,0.3548139632,0.4095387161,0.1268223077,-0.2900162339,-0.2419890761,0.2838993073,0.0142173441,-0.2483233958,-0.2783831358,0.1692030579,0.0764759481,0.7366990447,-0.5577983856,-0.1140442863,-0.430565536,0.1685490608,-0.3902668655,0.071951814,0.154959783,0.0929495022,-0.0118967332,-0.4119156003,-0.0944790542,0.1351483017,0.0751213357,0.2118790597,0.2510275543,-0.1262820661,0.2214883268,-0.2606827915,0.3279381096,0.2481326014,-0.3805216253,-0.0002255971,-0.0308669601,-0.1849082708,-0.0945181474,-0.0433808714,0.045279514,-0.0704916492,0.3731215,0.5257031322,0.2177648246,-0.1684404016,0.1170973629,0.375684768,0.1648922712,0.1534628421,0.062367551,0.0424035825,-0.3042173982,0.0391020514,0.0110796252,0.2551579475,-0.0433039553,0.3766871095,-0.0346268825,-0.0403555222,-0.1273494214,-0.0218834579,0.0662864149,-0.2807991803,-0.3609564006,-0.1409435421,0.111996375,0.0275993235,0.334613651,-0.2615431249,0.2003477812,-0.3364670575,0.0672368109,-0.1536249816,-0.1003374457,-0.2305383235,0.1230605617,0.1779459715,-0.341763556,0.2291339785,0.085795626,-0.055318255,0.0410853773,0.1837367415,-0.3757719398,0.4564963281,0.23024863,-0.1737115979,0.1937036365,0.0566022396,0.2335437834,-0.2558582723,0.497835964,-0.241155833,-0.2049873471,0.2311902791,0.0342101119,-0.1543439478,-0.0228702538,-0.208795324,0.1444380581,0.4477293193,-0.3578275442,0.2071652412,-0.5868683457,-0.2272828221,-0.4626278877,0.0286326837,0.2016877979,-0.3084818423,-0.0998898596,0.5335569382,-0.1290865541,0.2292512059,0.2242171019,-0.3500424922,0.3968575001,-0.1544257998,-0.063399151,0.5925039053,0.0699077174,-0.2872909904,-0.3488504887,-0.0791895986,0.3577805459,-0.1115627959,0.1525214911,-0.1749596149,-0.0800708011,0.2018082291,0.11498449,-0.0569889396,0.1131259128,-0.2052199543,0.0311108716,0.4739532173,0.1160250679,-0.085915342,-0.0094829211,-0.0761847273,0.158943966,0.1887992769,-0.0295025203,-0.2351531982,0.0499477834,-0.1133157834,0.1867135465,-0.0154641299,-0.1854670942,-0.2254036516,-0.1526828557,0.1799524128,-0.3040947616,0.3576213419,0.0391091369,0.010645817,-0.332868278,-0.4570793509,-0.3931130469,-0.0499445647,0.4065027535,0.0490477979,0.1186681837,-0.3195305169,0.0614476018,-0.2063674182,-0.0083991289,-0.4115723372,0.1248530522,0.1170983836,-0.2123429328,-0.3896256983,0.1242080256,-0.1243458912,-0.0857831985,-0.2129511684,0.3616207242,0.1442990601,-0.2647690475,-0.3784194589,0.4010251462,0.1274124384,0.2065180242,0.1345853209,-0.1895992756,0.3265334368,-0.1915585399,-0.2690972388,0.3493041098,-0.6575779319,-0.1802589595,-0.1800181866,-0.6270247698,-0.0393623821,0.2436851114,-0.38536641,0.045983322,0.0531655848,0.3270132244,0.1938665211,0.3799907863,-0.5700509548,-0.0476526283,0.7285046577,-0.2212718576,0.0553403497,0.0533638261,-0.1717451513,-0.3505660892,-0.1845951527,-0.0203250516,0.782093823,-0.0153637482,-0.0367374867,0.0553940795,0.0814553723,-0.3519118428,0.0955269486,0.2491502464,0.00695509,0.4411150217,0.3575190604,0.1117630824,-0.488229394,0.1302794069,0.1637688428,-0.0803554654,-0.3349176943,0.0944054872,-0.155654043,0.0596321747,-0.3772047162,-0.0254960172,0.0446690954,-0.0599994063,-0.081163533,0.0489457361,0.1002306193,-0.0627609044,0.2464330047,0.224451378,-0.0175593719,-0.3098870516,0.3108996451,-0.0712586492,-0.1993415505,-0.1147271618,0.0123944832,-0.0296001565,0.3543699086,-0.0558917895,-0.2055740356,-0.1391870677,-0.1954615563,0.2927092016,-0.0928981304,-0.0006683394,0.0193675347,-0.0350417793,0.0142719653,-0.1980372071,0.1020730063,0.1147153229,0.1975731552,0.0408127047,0.1796193719,0.1345437765,-0.1993232816,-0.291289866,-0.0313384794,0.088165395,0.2162626535,-0.1674561948,-0.0139829572,0.465908438,-0.0845903978,-0.1470181197,-0.6666821837,0.1076658294,-0.3695809245,-0.5208802819,0.129328534,-0.0555758886,-0.2137968689,-0.2652085423,0.1450389624,-0.0572913438,0.5463920832,-0.2382081747,-0.1953873187,-0.1374084204,-0.0648620725,-0.0691306964,0.4167664349,0.3020017147,0.0571569018,0.1945846677,0.038517151,-0.043403767,0.1414676905,0.3747738302,0.0798519477,0.0951740593,0.1910053194,-0.0178647935,0.5995121002,-0.1991274357,-0.1613339633,0.1348063946,0.1795824766,0.3882859349,-0.4722309709,-0.104178369,0.4822199345,-0.0516313165,-0.0903253704,0.2900754213,-0.2437630743,0.0153201325,-0.1181699112,0.1589726955,-0.2877340019,-0.2249055505,0.4279428422,-0.0042330185,0.2148872763,0.0903220549,0.3507176936,-0.1093763113,0.0532986932,-0.1550857127,0.1416816264,0.404302299,0.0007882579,-0.5444186926,-0.2212259918,-0.1243489757,0.0089226775,0.1123208329,0.5317481756,0.0306649636,-0.1736357659,0.1770093888,0.2656615376,0.5368555784,-0.0887493938,-0.0582991466,0.1731205434,-0.3876774609,-0.1465545893,-0.3178415,-0.2131336778,0.0677269697,-0.2510716021,0.263696909,-0.203009069,-0.0140729835,-0.0744816959,0.4079804122,-0.1185537726,0.0340827964,-0.0952488631,-0.0083027743,-0.1466766298,0.0440276712,-0.1954734176,-0.1743855923,-0.3775610924,0.3725061119,-0.0666091517,0.2309618145,-0.0352425128,-0.1375807226,0.265242964,0.1389311254,-0.1037343889,0.1086483896,0.0686223432,0.0677358881,0.0991382301,-0.0500395894,-0.0214681067,0.2940168083,-0.343662709,0.1544417888,0.4693882167,0.0623107404,0.0337494723,-0.420491457,0.0856411383,-0.2320627272,0.0723931193,0.347905755,-0.080632478,-0.2999900579,-1.069737196,0.2451432347,0.1815061867,0.0347042903,0.1759747416,-0.2191982418,-0.0421901494,0.2608632743,0.1543843597,1.0892894268,-0.0507256314,0.5218431354,0.3229362369,-0.0731944144,0.3819906414,-0.3416105509,0.1129657924,-0.244715184,-0.4424582422,0.0531473085,-0.2846353948,-0.0072025433,0.0078241881,-0.1367422193,0.3234067857,-0.2796405852,-0.0524001606,-0.0252828151,0.0340554975,-0.1102772057,-0.0948151127,-0.0633685216,-0.0598191917,-0.1997699589,0.1812173277,-0.0717719346,0.1181937978,-0.1108608022,0.0298046488,-0.04482181,-0.1666142493,-0.0512364395,0.0092817806,-0.3341157734,-0.6155846715,0.283614248,0.3402102292,0.0996392891,0.179265514,0.053441681,0.0836941898,0.2451819032,0.4168071449,-0.0029478443,-0.0521088839,0.4235515296,0.0652486905,0.0677940995,0.2524892688,-0.019039575,-0.096982412,-0.8286117911,0.1709132344,0.3791937232,-0.2473125607,-0.4441283047,0.0542826056,-0.0023440337,0.0265491717,-0.0408351608,0.1820113659,0.1740788072,0.4315118194,-0.2256633192,-0.4154630899,-0.1995775253,0.2852520943,0.2003929466,-0.045226451,0.4193188846,-0.0519285202,-0.2008045614,-0.0527363978,0.2371236384,-0.0974297449,-0.4542705119,-0.1012882963,-0.0841003731,0.1255544573,-0.019005673,0.2705136538,-0.2472783327,0.0637085661,-0.2091897577,-0.0291788038,0.1038118005,0.0176787004,0.4277179837,0.3437690735,0.3647261858,0.3129191101,0.0152425794,-0.0564862154,-0.1305907518,0.0297765024,0.3241150379,0.3977801502,0.0183637086,0.0064171935,-0.3131085336,-0.0604932867,-0.1109374538,0.1440346837,-0.1140713394,-0.0874803364,-0.1520568579,0.2300266922,0.0748610124,-0.0025268281,-0.169527486,-0.1676239073,-0.1742764413,-0.2396644801,0.2976013422,0.2611989975,0.0210243277,0.1257654279,0.4362260401,0.0285326708,0.1622669399,0.1904217899,-0.158964932,-0.2615143061,0.1234544814,0.0164786372,-0.2788673341,-0.0519074686,-0.599006176,0.143350035,0.2141147405,0.3800849319,0.0123016732,-0.3274235427,0.3317849338,-0.1678716242,0.4022538066,0.5396655202,0.0275575705,-0.4948884547,0.2594493628,-0.0720313266,0.2253415436,-0.1419828832,-0.0395640843,0.1817880869,-0.1139530316,0.1323440224,-0.1584558636,0.3500945866,-0.5146568418,-0.0992891118,0.1379054487,-0.1613038331,0.2236180604,0.4979375005,0.046598129,-0.0668177158,0.1828082502,-0.0771138519,-0.1445321739,0.2808187902,0.7301182747,0.5393793583,0.4825175703,0.0395526029,-0.0672645941,-0.4697543085,-0.2155254185,0.1976953298,0.0375333838,0.5161758065,0.2281161696,0.0463414192,0.1134193763,-0.0799599364,0.1270460784,0.336656481,-0.1223196089,-0.042072054,0.2104701251,-0.3991296887,0.0812123939,0.1484832466,-0.4271040857,0.2103952616,0.1051131934,0.170705542,-0.1856763512,-0.111417748,0.2426232994,-0.1366685182,0.1284887642,-0.1080194265,-0.0554121286,0.1767769009,0.0271324012,0.1898210347,0.1285141706,-0.051371187,0.3518345356,-0.1466553062,0.0571049117,-0.103277728,0.1627274603,0.0270891506,0.0942607075,0.3115408123,-0.3098682165,0.4923088849,0.0618408546,-0.0126401251,-0.1264349073,0.3408855796,0.2179021537,-0.5887755752,-0.0272438154,0.1398259848,0.0593104847,0.0964889973,0.1066811383,-0.4906990528,0.0379791483,0.0147474837,0.4293690026,0.1835746914,-0.0085616251,-0.0065229358,0.1072581634,0.4353663921,0.0826549903,-0.1552221775,-0.0474484824,0.0259117559,-0.428497076,0.0506531,0.3247933388,0.1212253124,0.2492528111,-0.1100715548,0.2826694548,0.1380500942,-0.1708900332,0.217368871,-0.0020115608,0.4330170155,-0.3053794503,-0.0873088166,-0.41097036,0.1781348139,-0.1709602028,-0.1909289807,0.1201006249,0.1029169634,0.0281698108,-0.2006617785,-0.1350136995,-0.0147139719,0.3194841444,0.2151916772,0.0709726214,0.2752601504,0.0149041666,0.0139682246,-0.1395524144,0.0865887702,-0.0084388126,0.2101175189,-0.2170508057,0.39561975,-0.1571722031,0.1808829904,-0.3153145313,0.2677330971,-0.1786733717,0.2012965232,-0.2949556708,-0.061833445,0.322712332,0.0421419963,0.0804623365,0.5686163306,-0.0181748904,0.2659771144,-0.2791575789,0.2963319719,0.4934827983,0.1330400407,-0.2201577425,-0.2162210643,-0.0268434361,0.1314427555,0.0620843135,-0.3097263277,0.0052932235,0.2125545293,-0.2343695909,-0.3316029906,0.2053728402,-0.054333806,-0.2618587911,-0.1250595897,0.5423796773,-0.0547792353,-0.0949317217,-0.0436527915,-0.03871084]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/511","title":"dataset.shuffle() and select() resets format. Intended?","comments":"Shall we have this in the coming release by the way @lhoestq ?","body":"Calling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```","comment_length":13,"text":"dataset.shuffle() and select() resets format. Intended?\nCalling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```\nShall we have this in the coming release by the way @lhoestq ?","embeddings":[-0.1382125914,-0.2247202247,0.0239643287,0.2476742417,0.1281202286,-0.1362012476,0.2032122612,0.0824569091,-0.6032932997,-0.0022735526,-0.2842715979,0.322750628,-0.1951235682,0.1738311946,0.1398987323,-0.1447517425,0.1145583838,0.2089668661,-0.0435798876,-0.115610823,-0.1563792527,-0.2866491675,-0.2859784365,-0.0606167801,-0.4345372915,0.0260992981,0.0050066882,0.1479002982,-0.0802040696,0.0760351792,0.0442324206,0.2561855018,-0.1987256706,0.5789396763,-0.0001289723,-0.1655297279,0.0554003008,-0.2835808396,-0.292496562,0.0555677451,-0.0595302619,0.2450539023,0.0335684307,-0.1791678071,0.0966271758,-0.3774504662,-0.0920630097,-0.1147696525,0.189046368,0.1550750881,0.0798439384,0.3368454278,-0.2491133064,0.2438392788,0.3745867014,0.557667613,-0.1313131005,0.2056202143,0.101867266,0.3548139632,0.4095387161,0.1268223077,-0.2900162339,-0.2419890761,0.2838993073,0.0142173441,-0.2483233958,-0.2783831358,0.1692030579,0.0764759481,0.7366990447,-0.5577983856,-0.1140442863,-0.430565536,0.1685490608,-0.3902668655,0.071951814,0.154959783,0.0929495022,-0.0118967332,-0.4119156003,-0.0944790542,0.1351483017,0.0751213357,0.2118790597,0.2510275543,-0.1262820661,0.2214883268,-0.2606827915,0.3279381096,0.2481326014,-0.3805216253,-0.0002255971,-0.0308669601,-0.1849082708,-0.0945181474,-0.0433808714,0.045279514,-0.0704916492,0.3731215,0.5257031322,0.2177648246,-0.1684404016,0.1170973629,0.375684768,0.1648922712,0.1534628421,0.062367551,0.0424035825,-0.3042173982,0.0391020514,0.0110796252,0.2551579475,-0.0433039553,0.3766871095,-0.0346268825,-0.0403555222,-0.1273494214,-0.0218834579,0.0662864149,-0.2807991803,-0.3609564006,-0.1409435421,0.111996375,0.0275993235,0.334613651,-0.2615431249,0.2003477812,-0.3364670575,0.0672368109,-0.1536249816,-0.1003374457,-0.2305383235,0.1230605617,0.1779459715,-0.341763556,0.2291339785,0.085795626,-0.055318255,0.0410853773,0.1837367415,-0.3757719398,0.4564963281,0.23024863,-0.1737115979,0.1937036365,0.0566022396,0.2335437834,-0.2558582723,0.497835964,-0.241155833,-0.2049873471,0.2311902791,0.0342101119,-0.1543439478,-0.0228702538,-0.208795324,0.1444380581,0.4477293193,-0.3578275442,0.2071652412,-0.5868683457,-0.2272828221,-0.4626278877,0.0286326837,0.2016877979,-0.3084818423,-0.0998898596,0.5335569382,-0.1290865541,0.2292512059,0.2242171019,-0.3500424922,0.3968575001,-0.1544257998,-0.063399151,0.5925039053,0.0699077174,-0.2872909904,-0.3488504887,-0.0791895986,0.3577805459,-0.1115627959,0.1525214911,-0.1749596149,-0.0800708011,0.2018082291,0.11498449,-0.0569889396,0.1131259128,-0.2052199543,0.0311108716,0.4739532173,0.1160250679,-0.085915342,-0.0094829211,-0.0761847273,0.158943966,0.1887992769,-0.0295025203,-0.2351531982,0.0499477834,-0.1133157834,0.1867135465,-0.0154641299,-0.1854670942,-0.2254036516,-0.1526828557,0.1799524128,-0.3040947616,0.3576213419,0.0391091369,0.010645817,-0.332868278,-0.4570793509,-0.3931130469,-0.0499445647,0.4065027535,0.0490477979,0.1186681837,-0.3195305169,0.0614476018,-0.2063674182,-0.0083991289,-0.4115723372,0.1248530522,0.1170983836,-0.2123429328,-0.3896256983,0.1242080256,-0.1243458912,-0.0857831985,-0.2129511684,0.3616207242,0.1442990601,-0.2647690475,-0.3784194589,0.4010251462,0.1274124384,0.2065180242,0.1345853209,-0.1895992756,0.3265334368,-0.1915585399,-0.2690972388,0.3493041098,-0.6575779319,-0.1802589595,-0.1800181866,-0.6270247698,-0.0393623821,0.2436851114,-0.38536641,0.045983322,0.0531655848,0.3270132244,0.1938665211,0.3799907863,-0.5700509548,-0.0476526283,0.7285046577,-0.2212718576,0.0553403497,0.0533638261,-0.1717451513,-0.3505660892,-0.1845951527,-0.0203250516,0.782093823,-0.0153637482,-0.0367374867,0.0553940795,0.0814553723,-0.3519118428,0.0955269486,0.2491502464,0.00695509,0.4411150217,0.3575190604,0.1117630824,-0.488229394,0.1302794069,0.1637688428,-0.0803554654,-0.3349176943,0.0944054872,-0.155654043,0.0596321747,-0.3772047162,-0.0254960172,0.0446690954,-0.0599994063,-0.081163533,0.0489457361,0.1002306193,-0.0627609044,0.2464330047,0.224451378,-0.0175593719,-0.3098870516,0.3108996451,-0.0712586492,-0.1993415505,-0.1147271618,0.0123944832,-0.0296001565,0.3543699086,-0.0558917895,-0.2055740356,-0.1391870677,-0.1954615563,0.2927092016,-0.0928981304,-0.0006683394,0.0193675347,-0.0350417793,0.0142719653,-0.1980372071,0.1020730063,0.1147153229,0.1975731552,0.0408127047,0.1796193719,0.1345437765,-0.1993232816,-0.291289866,-0.0313384794,0.088165395,0.2162626535,-0.1674561948,-0.0139829572,0.465908438,-0.0845903978,-0.1470181197,-0.6666821837,0.1076658294,-0.3695809245,-0.5208802819,0.129328534,-0.0555758886,-0.2137968689,-0.2652085423,0.1450389624,-0.0572913438,0.5463920832,-0.2382081747,-0.1953873187,-0.1374084204,-0.0648620725,-0.0691306964,0.4167664349,0.3020017147,0.0571569018,0.1945846677,0.038517151,-0.043403767,0.1414676905,0.3747738302,0.0798519477,0.0951740593,0.1910053194,-0.0178647935,0.5995121002,-0.1991274357,-0.1613339633,0.1348063946,0.1795824766,0.3882859349,-0.4722309709,-0.104178369,0.4822199345,-0.0516313165,-0.0903253704,0.2900754213,-0.2437630743,0.0153201325,-0.1181699112,0.1589726955,-0.2877340019,-0.2249055505,0.4279428422,-0.0042330185,0.2148872763,0.0903220549,0.3507176936,-0.1093763113,0.0532986932,-0.1550857127,0.1416816264,0.404302299,0.0007882579,-0.5444186926,-0.2212259918,-0.1243489757,0.0089226775,0.1123208329,0.5317481756,0.0306649636,-0.1736357659,0.1770093888,0.2656615376,0.5368555784,-0.0887493938,-0.0582991466,0.1731205434,-0.3876774609,-0.1465545893,-0.3178415,-0.2131336778,0.0677269697,-0.2510716021,0.263696909,-0.203009069,-0.0140729835,-0.0744816959,0.4079804122,-0.1185537726,0.0340827964,-0.0952488631,-0.0083027743,-0.1466766298,0.0440276712,-0.1954734176,-0.1743855923,-0.3775610924,0.3725061119,-0.0666091517,0.2309618145,-0.0352425128,-0.1375807226,0.265242964,0.1389311254,-0.1037343889,0.1086483896,0.0686223432,0.0677358881,0.0991382301,-0.0500395894,-0.0214681067,0.2940168083,-0.343662709,0.1544417888,0.4693882167,0.0623107404,0.0337494723,-0.420491457,0.0856411383,-0.2320627272,0.0723931193,0.347905755,-0.080632478,-0.2999900579,-1.069737196,0.2451432347,0.1815061867,0.0347042903,0.1759747416,-0.2191982418,-0.0421901494,0.2608632743,0.1543843597,1.0892894268,-0.0507256314,0.5218431354,0.3229362369,-0.0731944144,0.3819906414,-0.3416105509,0.1129657924,-0.244715184,-0.4424582422,0.0531473085,-0.2846353948,-0.0072025433,0.0078241881,-0.1367422193,0.3234067857,-0.2796405852,-0.0524001606,-0.0252828151,0.0340554975,-0.1102772057,-0.0948151127,-0.0633685216,-0.0598191917,-0.1997699589,0.1812173277,-0.0717719346,0.1181937978,-0.1108608022,0.0298046488,-0.04482181,-0.1666142493,-0.0512364395,0.0092817806,-0.3341157734,-0.6155846715,0.283614248,0.3402102292,0.0996392891,0.179265514,0.053441681,0.0836941898,0.2451819032,0.4168071449,-0.0029478443,-0.0521088839,0.4235515296,0.0652486905,0.0677940995,0.2524892688,-0.019039575,-0.096982412,-0.8286117911,0.1709132344,0.3791937232,-0.2473125607,-0.4441283047,0.0542826056,-0.0023440337,0.0265491717,-0.0408351608,0.1820113659,0.1740788072,0.4315118194,-0.2256633192,-0.4154630899,-0.1995775253,0.2852520943,0.2003929466,-0.045226451,0.4193188846,-0.0519285202,-0.2008045614,-0.0527363978,0.2371236384,-0.0974297449,-0.4542705119,-0.1012882963,-0.0841003731,0.1255544573,-0.019005673,0.2705136538,-0.2472783327,0.0637085661,-0.2091897577,-0.0291788038,0.1038118005,0.0176787004,0.4277179837,0.3437690735,0.3647261858,0.3129191101,0.0152425794,-0.0564862154,-0.1305907518,0.0297765024,0.3241150379,0.3977801502,0.0183637086,0.0064171935,-0.3131085336,-0.0604932867,-0.1109374538,0.1440346837,-0.1140713394,-0.0874803364,-0.1520568579,0.2300266922,0.0748610124,-0.0025268281,-0.169527486,-0.1676239073,-0.1742764413,-0.2396644801,0.2976013422,0.2611989975,0.0210243277,0.1257654279,0.4362260401,0.0285326708,0.1622669399,0.1904217899,-0.158964932,-0.2615143061,0.1234544814,0.0164786372,-0.2788673341,-0.0519074686,-0.599006176,0.143350035,0.2141147405,0.3800849319,0.0123016732,-0.3274235427,0.3317849338,-0.1678716242,0.4022538066,0.5396655202,0.0275575705,-0.4948884547,0.2594493628,-0.0720313266,0.2253415436,-0.1419828832,-0.0395640843,0.1817880869,-0.1139530316,0.1323440224,-0.1584558636,0.3500945866,-0.5146568418,-0.0992891118,0.1379054487,-0.1613038331,0.2236180604,0.4979375005,0.046598129,-0.0668177158,0.1828082502,-0.0771138519,-0.1445321739,0.2808187902,0.7301182747,0.5393793583,0.4825175703,0.0395526029,-0.0672645941,-0.4697543085,-0.2155254185,0.1976953298,0.0375333838,0.5161758065,0.2281161696,0.0463414192,0.1134193763,-0.0799599364,0.1270460784,0.336656481,-0.1223196089,-0.042072054,0.2104701251,-0.3991296887,0.0812123939,0.1484832466,-0.4271040857,0.2103952616,0.1051131934,0.170705542,-0.1856763512,-0.111417748,0.2426232994,-0.1366685182,0.1284887642,-0.1080194265,-0.0554121286,0.1767769009,0.0271324012,0.1898210347,0.1285141706,-0.051371187,0.3518345356,-0.1466553062,0.0571049117,-0.103277728,0.1627274603,0.0270891506,0.0942607075,0.3115408123,-0.3098682165,0.4923088849,0.0618408546,-0.0126401251,-0.1264349073,0.3408855796,0.2179021537,-0.5887755752,-0.0272438154,0.1398259848,0.0593104847,0.0964889973,0.1066811383,-0.4906990528,0.0379791483,0.0147474837,0.4293690026,0.1835746914,-0.0085616251,-0.0065229358,0.1072581634,0.4353663921,0.0826549903,-0.1552221775,-0.0474484824,0.0259117559,-0.428497076,0.0506531,0.3247933388,0.1212253124,0.2492528111,-0.1100715548,0.2826694548,0.1380500942,-0.1708900332,0.217368871,-0.0020115608,0.4330170155,-0.3053794503,-0.0873088166,-0.41097036,0.1781348139,-0.1709602028,-0.1909289807,0.1201006249,0.1029169634,0.0281698108,-0.2006617785,-0.1350136995,-0.0147139719,0.3194841444,0.2151916772,0.0709726214,0.2752601504,0.0149041666,0.0139682246,-0.1395524144,0.0865887702,-0.0084388126,0.2101175189,-0.2170508057,0.39561975,-0.1571722031,0.1808829904,-0.3153145313,0.2677330971,-0.1786733717,0.2012965232,-0.2949556708,-0.061833445,0.322712332,0.0421419963,0.0804623365,0.5686163306,-0.0181748904,0.2659771144,-0.2791575789,0.2963319719,0.4934827983,0.1330400407,-0.2201577425,-0.2162210643,-0.0268434361,0.1314427555,0.0620843135,-0.3097263277,0.0052932235,0.2125545293,-0.2343695909,-0.3316029906,0.2053728402,-0.054333806,-0.2618587911,-0.1250595897,0.5423796773,-0.0547792353,-0.0949317217,-0.0436527915,-0.03871084]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/511","title":"dataset.shuffle() and select() resets format. Intended?","comments":"Yes sure !","body":"Calling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```","comment_length":3,"text":"dataset.shuffle() and select() resets format. Intended?\nCalling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```\nYes sure !","embeddings":[-0.1382125914,-0.2247202247,0.0239643287,0.2476742417,0.1281202286,-0.1362012476,0.2032122612,0.0824569091,-0.6032932997,-0.0022735526,-0.2842715979,0.322750628,-0.1951235682,0.1738311946,0.1398987323,-0.1447517425,0.1145583838,0.2089668661,-0.0435798876,-0.115610823,-0.1563792527,-0.2866491675,-0.2859784365,-0.0606167801,-0.4345372915,0.0260992981,0.0050066882,0.1479002982,-0.0802040696,0.0760351792,0.0442324206,0.2561855018,-0.1987256706,0.5789396763,-0.0001289723,-0.1655297279,0.0554003008,-0.2835808396,-0.292496562,0.0555677451,-0.0595302619,0.2450539023,0.0335684307,-0.1791678071,0.0966271758,-0.3774504662,-0.0920630097,-0.1147696525,0.189046368,0.1550750881,0.0798439384,0.3368454278,-0.2491133064,0.2438392788,0.3745867014,0.557667613,-0.1313131005,0.2056202143,0.101867266,0.3548139632,0.4095387161,0.1268223077,-0.2900162339,-0.2419890761,0.2838993073,0.0142173441,-0.2483233958,-0.2783831358,0.1692030579,0.0764759481,0.7366990447,-0.5577983856,-0.1140442863,-0.430565536,0.1685490608,-0.3902668655,0.071951814,0.154959783,0.0929495022,-0.0118967332,-0.4119156003,-0.0944790542,0.1351483017,0.0751213357,0.2118790597,0.2510275543,-0.1262820661,0.2214883268,-0.2606827915,0.3279381096,0.2481326014,-0.3805216253,-0.0002255971,-0.0308669601,-0.1849082708,-0.0945181474,-0.0433808714,0.045279514,-0.0704916492,0.3731215,0.5257031322,0.2177648246,-0.1684404016,0.1170973629,0.375684768,0.1648922712,0.1534628421,0.062367551,0.0424035825,-0.3042173982,0.0391020514,0.0110796252,0.2551579475,-0.0433039553,0.3766871095,-0.0346268825,-0.0403555222,-0.1273494214,-0.0218834579,0.0662864149,-0.2807991803,-0.3609564006,-0.1409435421,0.111996375,0.0275993235,0.334613651,-0.2615431249,0.2003477812,-0.3364670575,0.0672368109,-0.1536249816,-0.1003374457,-0.2305383235,0.1230605617,0.1779459715,-0.341763556,0.2291339785,0.085795626,-0.055318255,0.0410853773,0.1837367415,-0.3757719398,0.4564963281,0.23024863,-0.1737115979,0.1937036365,0.0566022396,0.2335437834,-0.2558582723,0.497835964,-0.241155833,-0.2049873471,0.2311902791,0.0342101119,-0.1543439478,-0.0228702538,-0.208795324,0.1444380581,0.4477293193,-0.3578275442,0.2071652412,-0.5868683457,-0.2272828221,-0.4626278877,0.0286326837,0.2016877979,-0.3084818423,-0.0998898596,0.5335569382,-0.1290865541,0.2292512059,0.2242171019,-0.3500424922,0.3968575001,-0.1544257998,-0.063399151,0.5925039053,0.0699077174,-0.2872909904,-0.3488504887,-0.0791895986,0.3577805459,-0.1115627959,0.1525214911,-0.1749596149,-0.0800708011,0.2018082291,0.11498449,-0.0569889396,0.1131259128,-0.2052199543,0.0311108716,0.4739532173,0.1160250679,-0.085915342,-0.0094829211,-0.0761847273,0.158943966,0.1887992769,-0.0295025203,-0.2351531982,0.0499477834,-0.1133157834,0.1867135465,-0.0154641299,-0.1854670942,-0.2254036516,-0.1526828557,0.1799524128,-0.3040947616,0.3576213419,0.0391091369,0.010645817,-0.332868278,-0.4570793509,-0.3931130469,-0.0499445647,0.4065027535,0.0490477979,0.1186681837,-0.3195305169,0.0614476018,-0.2063674182,-0.0083991289,-0.4115723372,0.1248530522,0.1170983836,-0.2123429328,-0.3896256983,0.1242080256,-0.1243458912,-0.0857831985,-0.2129511684,0.3616207242,0.1442990601,-0.2647690475,-0.3784194589,0.4010251462,0.1274124384,0.2065180242,0.1345853209,-0.1895992756,0.3265334368,-0.1915585399,-0.2690972388,0.3493041098,-0.6575779319,-0.1802589595,-0.1800181866,-0.6270247698,-0.0393623821,0.2436851114,-0.38536641,0.045983322,0.0531655848,0.3270132244,0.1938665211,0.3799907863,-0.5700509548,-0.0476526283,0.7285046577,-0.2212718576,0.0553403497,0.0533638261,-0.1717451513,-0.3505660892,-0.1845951527,-0.0203250516,0.782093823,-0.0153637482,-0.0367374867,0.0553940795,0.0814553723,-0.3519118428,0.0955269486,0.2491502464,0.00695509,0.4411150217,0.3575190604,0.1117630824,-0.488229394,0.1302794069,0.1637688428,-0.0803554654,-0.3349176943,0.0944054872,-0.155654043,0.0596321747,-0.3772047162,-0.0254960172,0.0446690954,-0.0599994063,-0.081163533,0.0489457361,0.1002306193,-0.0627609044,0.2464330047,0.224451378,-0.0175593719,-0.3098870516,0.3108996451,-0.0712586492,-0.1993415505,-0.1147271618,0.0123944832,-0.0296001565,0.3543699086,-0.0558917895,-0.2055740356,-0.1391870677,-0.1954615563,0.2927092016,-0.0928981304,-0.0006683394,0.0193675347,-0.0350417793,0.0142719653,-0.1980372071,0.1020730063,0.1147153229,0.1975731552,0.0408127047,0.1796193719,0.1345437765,-0.1993232816,-0.291289866,-0.0313384794,0.088165395,0.2162626535,-0.1674561948,-0.0139829572,0.465908438,-0.0845903978,-0.1470181197,-0.6666821837,0.1076658294,-0.3695809245,-0.5208802819,0.129328534,-0.0555758886,-0.2137968689,-0.2652085423,0.1450389624,-0.0572913438,0.5463920832,-0.2382081747,-0.1953873187,-0.1374084204,-0.0648620725,-0.0691306964,0.4167664349,0.3020017147,0.0571569018,0.1945846677,0.038517151,-0.043403767,0.1414676905,0.3747738302,0.0798519477,0.0951740593,0.1910053194,-0.0178647935,0.5995121002,-0.1991274357,-0.1613339633,0.1348063946,0.1795824766,0.3882859349,-0.4722309709,-0.104178369,0.4822199345,-0.0516313165,-0.0903253704,0.2900754213,-0.2437630743,0.0153201325,-0.1181699112,0.1589726955,-0.2877340019,-0.2249055505,0.4279428422,-0.0042330185,0.2148872763,0.0903220549,0.3507176936,-0.1093763113,0.0532986932,-0.1550857127,0.1416816264,0.404302299,0.0007882579,-0.5444186926,-0.2212259918,-0.1243489757,0.0089226775,0.1123208329,0.5317481756,0.0306649636,-0.1736357659,0.1770093888,0.2656615376,0.5368555784,-0.0887493938,-0.0582991466,0.1731205434,-0.3876774609,-0.1465545893,-0.3178415,-0.2131336778,0.0677269697,-0.2510716021,0.263696909,-0.203009069,-0.0140729835,-0.0744816959,0.4079804122,-0.1185537726,0.0340827964,-0.0952488631,-0.0083027743,-0.1466766298,0.0440276712,-0.1954734176,-0.1743855923,-0.3775610924,0.3725061119,-0.0666091517,0.2309618145,-0.0352425128,-0.1375807226,0.265242964,0.1389311254,-0.1037343889,0.1086483896,0.0686223432,0.0677358881,0.0991382301,-0.0500395894,-0.0214681067,0.2940168083,-0.343662709,0.1544417888,0.4693882167,0.0623107404,0.0337494723,-0.420491457,0.0856411383,-0.2320627272,0.0723931193,0.347905755,-0.080632478,-0.2999900579,-1.069737196,0.2451432347,0.1815061867,0.0347042903,0.1759747416,-0.2191982418,-0.0421901494,0.2608632743,0.1543843597,1.0892894268,-0.0507256314,0.5218431354,0.3229362369,-0.0731944144,0.3819906414,-0.3416105509,0.1129657924,-0.244715184,-0.4424582422,0.0531473085,-0.2846353948,-0.0072025433,0.0078241881,-0.1367422193,0.3234067857,-0.2796405852,-0.0524001606,-0.0252828151,0.0340554975,-0.1102772057,-0.0948151127,-0.0633685216,-0.0598191917,-0.1997699589,0.1812173277,-0.0717719346,0.1181937978,-0.1108608022,0.0298046488,-0.04482181,-0.1666142493,-0.0512364395,0.0092817806,-0.3341157734,-0.6155846715,0.283614248,0.3402102292,0.0996392891,0.179265514,0.053441681,0.0836941898,0.2451819032,0.4168071449,-0.0029478443,-0.0521088839,0.4235515296,0.0652486905,0.0677940995,0.2524892688,-0.019039575,-0.096982412,-0.8286117911,0.1709132344,0.3791937232,-0.2473125607,-0.4441283047,0.0542826056,-0.0023440337,0.0265491717,-0.0408351608,0.1820113659,0.1740788072,0.4315118194,-0.2256633192,-0.4154630899,-0.1995775253,0.2852520943,0.2003929466,-0.045226451,0.4193188846,-0.0519285202,-0.2008045614,-0.0527363978,0.2371236384,-0.0974297449,-0.4542705119,-0.1012882963,-0.0841003731,0.1255544573,-0.019005673,0.2705136538,-0.2472783327,0.0637085661,-0.2091897577,-0.0291788038,0.1038118005,0.0176787004,0.4277179837,0.3437690735,0.3647261858,0.3129191101,0.0152425794,-0.0564862154,-0.1305907518,0.0297765024,0.3241150379,0.3977801502,0.0183637086,0.0064171935,-0.3131085336,-0.0604932867,-0.1109374538,0.1440346837,-0.1140713394,-0.0874803364,-0.1520568579,0.2300266922,0.0748610124,-0.0025268281,-0.169527486,-0.1676239073,-0.1742764413,-0.2396644801,0.2976013422,0.2611989975,0.0210243277,0.1257654279,0.4362260401,0.0285326708,0.1622669399,0.1904217899,-0.158964932,-0.2615143061,0.1234544814,0.0164786372,-0.2788673341,-0.0519074686,-0.599006176,0.143350035,0.2141147405,0.3800849319,0.0123016732,-0.3274235427,0.3317849338,-0.1678716242,0.4022538066,0.5396655202,0.0275575705,-0.4948884547,0.2594493628,-0.0720313266,0.2253415436,-0.1419828832,-0.0395640843,0.1817880869,-0.1139530316,0.1323440224,-0.1584558636,0.3500945866,-0.5146568418,-0.0992891118,0.1379054487,-0.1613038331,0.2236180604,0.4979375005,0.046598129,-0.0668177158,0.1828082502,-0.0771138519,-0.1445321739,0.2808187902,0.7301182747,0.5393793583,0.4825175703,0.0395526029,-0.0672645941,-0.4697543085,-0.2155254185,0.1976953298,0.0375333838,0.5161758065,0.2281161696,0.0463414192,0.1134193763,-0.0799599364,0.1270460784,0.336656481,-0.1223196089,-0.042072054,0.2104701251,-0.3991296887,0.0812123939,0.1484832466,-0.4271040857,0.2103952616,0.1051131934,0.170705542,-0.1856763512,-0.111417748,0.2426232994,-0.1366685182,0.1284887642,-0.1080194265,-0.0554121286,0.1767769009,0.0271324012,0.1898210347,0.1285141706,-0.051371187,0.3518345356,-0.1466553062,0.0571049117,-0.103277728,0.1627274603,0.0270891506,0.0942607075,0.3115408123,-0.3098682165,0.4923088849,0.0618408546,-0.0126401251,-0.1264349073,0.3408855796,0.2179021537,-0.5887755752,-0.0272438154,0.1398259848,0.0593104847,0.0964889973,0.1066811383,-0.4906990528,0.0379791483,0.0147474837,0.4293690026,0.1835746914,-0.0085616251,-0.0065229358,0.1072581634,0.4353663921,0.0826549903,-0.1552221775,-0.0474484824,0.0259117559,-0.428497076,0.0506531,0.3247933388,0.1212253124,0.2492528111,-0.1100715548,0.2826694548,0.1380500942,-0.1708900332,0.217368871,-0.0020115608,0.4330170155,-0.3053794503,-0.0873088166,-0.41097036,0.1781348139,-0.1709602028,-0.1909289807,0.1201006249,0.1029169634,0.0281698108,-0.2006617785,-0.1350136995,-0.0147139719,0.3194841444,0.2151916772,0.0709726214,0.2752601504,0.0149041666,0.0139682246,-0.1395524144,0.0865887702,-0.0084388126,0.2101175189,-0.2170508057,0.39561975,-0.1571722031,0.1808829904,-0.3153145313,0.2677330971,-0.1786733717,0.2012965232,-0.2949556708,-0.061833445,0.322712332,0.0421419963,0.0804623365,0.5686163306,-0.0181748904,0.2659771144,-0.2791575789,0.2963319719,0.4934827983,0.1330400407,-0.2201577425,-0.2162210643,-0.0268434361,0.1314427555,0.0620843135,-0.3097263277,0.0052932235,0.2125545293,-0.2343695909,-0.3316029906,0.2053728402,-0.054333806,-0.2618587911,-0.1250595897,0.5423796773,-0.0547792353,-0.0949317217,-0.0436527915,-0.03871084]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/511","title":"dataset.shuffle() and select() resets format. Intended?","comments":"Since datasets 1.0.0 the format is not reset anymore.\r\nClosing this one, but feel free to re-open if you have other questions","body":"Calling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```","comment_length":22,"text":"dataset.shuffle() and select() resets format. Intended?\nCalling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight?\r\n\r\nWhen working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save(\"dataset.pt\")`. Later loading the dataset object using `torch.load(\"dataset.pt\")`, which conserves the defined format before saving. \r\nI do shuffling and selecting (for controlling dataset size) after loading the data from .pt-file, as it's convenient whenever you train multiple models with varying sizes of the same dataset. \r\n\r\nThe obvious workaround for this is to set the format again after using `dataset.select()` or `dataset.shuffle()`.\r\n\r\n_I guess this is more of a discussion on the design philosophy of the functions. Please let me know if this is not the right channel for these kinds of discussions or if they are not wanted at all!_\r\n\r\n#### How to reproduce:\r\n\r\n```python\r\nimport nlp\r\nfrom transformers import T5Tokenizer\r\n\r\ntokenizer = T5Tokenizer.from_pretrained(\"t5-base\")\r\ndef create_features(batch):\r\n context_encoding = tokenizer.batch_encode_plus(batch[\"context\"])\r\n return {\"input_ids\": context_encoding[\"input_ids\"]}\r\n\r\ndataset = nlp.load_dataset(\"cosmos_qa\", split=\"train\")\r\ndataset = dataset.map(create_features, batched=True)\r\ndataset.set_format(type=\"torch\", columns=[\"input_ids\"])\r\ndataset[0]\r\n# {'input_ids': tensor([ 1804, 3525, 1602, ... 0, 0])}\r\n\r\ndataset = dataset.shuffle()\r\ndataset[0]\r\n# {'id': '3Q9(...)20', 'context': \"Good Old War an (...) play ?', 'answer0': 'None of the above choices .', 'answer1': 'This person likes music and likes to see the show , they will see other bands play .', (...) 'input_ids': [1804, 3525, 1602, ... , 0, 0]}\r\n\r\n```\nSince datasets 1.0.0 the format is not reset anymore.\r\nClosing this one, but feel free to re-open if you have other questions","embeddings":[-0.1382125914,-0.2247202247,0.0239643287,0.2476742417,0.1281202286,-0.1362012476,0.2032122612,0.0824569091,-0.6032932997,-0.0022735526,-0.2842715979,0.322750628,-0.1951235682,0.1738311946,0.1398987323,-0.1447517425,0.1145583838,0.2089668661,-0.0435798876,-0.115610823,-0.1563792527,-0.2866491675,-0.2859784365,-0.0606167801,-0.4345372915,0.0260992981,0.0050066882,0.1479002982,-0.0802040696,0.0760351792,0.0442324206,0.2561855018,-0.1987256706,0.5789396763,-0.0001289723,-0.1655297279,0.0554003008,-0.2835808396,-0.292496562,0.0555677451,-0.0595302619,0.2450539023,0.0335684307,-0.1791678071,0.0966271758,-0.3774504662,-0.0920630097,-0.1147696525,0.189046368,0.1550750881,0.0798439384,0.3368454278,-0.2491133064,0.2438392788,0.3745867014,0.557667613,-0.1313131005,0.2056202143,0.101867266,0.3548139632,0.4095387161,0.1268223077,-0.2900162339,-0.2419890761,0.2838993073,0.0142173441,-0.2483233958,-0.2783831358,0.1692030579,0.0764759481,0.7366990447,-0.5577983856,-0.1140442863,-0.430565536,0.1685490608,-0.3902668655,0.071951814,0.154959783,0.0929495022,-0.0118967332,-0.4119156003,-0.0944790542,0.1351483017,0.0751213357,0.2118790597,0.2510275543,-0.1262820661,0.2214883268,-0.2606827915,0.3279381096,0.2481326014,-0.3805216253,-0.0002255971,-0.0308669601,-0.1849082708,-0.0945181474,-0.0433808714,0.045279514,-0.0704916492,0.3731215,0.5257031322,0.2177648246,-0.1684404016,0.1170973629,0.375684768,0.1648922712,0.1534628421,0.062367551,0.0424035825,-0.3042173982,0.0391020514,0.0110796252,0.2551579475,-0.0433039553,0.3766871095,-0.0346268825,-0.0403555222,-0.1273494214,-0.0218834579,0.0662864149,-0.2807991803,-0.3609564006,-0.1409435421,0.111996375,0.0275993235,0.334613651,-0.2615431249,0.2003477812,-0.3364670575,0.0672368109,-0.1536249816,-0.1003374457,-0.2305383235,0.1230605617,0.1779459715,-0.341763556,0.2291339785,0.085795626,-0.055318255,0.0410853773,0.1837367415,-0.3757719398,0.4564963281,0.23024863,-0.1737115979,0.1937036365,0.0566022396,0.2335437834,-0.2558582723,0.497835964,-0.241155833,-0.2049873471,0.2311902791,0.0342101119,-0.1543439478,-0.0228702538,-0.208795324,0.1444380581,0.4477293193,-0.3578275442,0.2071652412,-0.5868683457,-0.2272828221,-0.4626278877,0.0286326837,0.2016877979,-0.3084818423,-0.0998898596,0.5335569382,-0.1290865541,0.2292512059,0.2242171019,-0.3500424922,0.3968575001,-0.1544257998,-0.063399151,0.5925039053,0.0699077174,-0.2872909904,-0.3488504887,-0.0791895986,0.3577805459,-0.1115627959,0.1525214911,-0.1749596149,-0.0800708011,0.2018082291,0.11498449,-0.0569889396,0.1131259128,-0.2052199543,0.0311108716,0.4739532173,0.1160250679,-0.085915342,-0.0094829211,-0.0761847273,0.158943966,0.1887992769,-0.0295025203,-0.2351531982,0.0499477834,-0.1133157834,0.1867135465,-0.0154641299,-0.1854670942,-0.2254036516,-0.1526828557,0.1799524128,-0.3040947616,0.3576213419,0.0391091369,0.010645817,-0.332868278,-0.4570793509,-0.3931130469,-0.0499445647,0.4065027535,0.0490477979,0.1186681837,-0.3195305169,0.0614476018,-0.2063674182,-0.0083991289,-0.4115723372,0.1248530522,0.1170983836,-0.2123429328,-0.3896256983,0.1242080256,-0.1243458912,-0.0857831985,-0.2129511684,0.3616207242,0.1442990601,-0.2647690475,-0.3784194589,0.4010251462,0.1274124384,0.2065180242,0.1345853209,-0.1895992756,0.3265334368,-0.1915585399,-0.2690972388,0.3493041098,-0.6575779319,-0.1802589595,-0.1800181866,-0.6270247698,-0.0393623821,0.2436851114,-0.38536641,0.045983322,0.0531655848,0.3270132244,0.1938665211,0.3799907863,-0.5700509548,-0.0476526283,0.7285046577,-0.2212718576,0.0553403497,0.0533638261,-0.1717451513,-0.3505660892,-0.1845951527,-0.0203250516,0.782093823,-0.0153637482,-0.0367374867,0.0553940795,0.0814553723,-0.3519118428,0.0955269486,0.2491502464,0.00695509,0.4411150217,0.3575190604,0.1117630824,-0.488229394,0.1302794069,0.1637688428,-0.0803554654,-0.3349176943,0.0944054872,-0.155654043,0.0596321747,-0.3772047162,-0.0254960172,0.0446690954,-0.0599994063,-0.081163533,0.0489457361,0.1002306193,-0.0627609044,0.2464330047,0.224451378,-0.0175593719,-0.3098870516,0.3108996451,-0.0712586492,-0.1993415505,-0.1147271618,0.0123944832,-0.0296001565,0.3543699086,-0.0558917895,-0.2055740356,-0.1391870677,-0.1954615563,0.2927092016,-0.0928981304,-0.0006683394,0.0193675347,-0.0350417793,0.0142719653,-0.1980372071,0.1020730063,0.1147153229,0.1975731552,0.0408127047,0.1796193719,0.1345437765,-0.1993232816,-0.291289866,-0.0313384794,0.088165395,0.2162626535,-0.1674561948,-0.0139829572,0.465908438,-0.0845903978,-0.1470181197,-0.6666821837,0.1076658294,-0.3695809245,-0.5208802819,0.129328534,-0.0555758886,-0.2137968689,-0.2652085423,0.1450389624,-0.0572913438,0.5463920832,-0.2382081747,-0.1953873187,-0.1374084204,-0.0648620725,-0.0691306964,0.4167664349,0.3020017147,0.0571569018,0.1945846677,0.038517151,-0.043403767,0.1414676905,0.3747738302,0.0798519477,0.0951740593,0.1910053194,-0.0178647935,0.5995121002,-0.1991274357,-0.1613339633,0.1348063946,0.1795824766,0.3882859349,-0.4722309709,-0.104178369,0.4822199345,-0.0516313165,-0.0903253704,0.2900754213,-0.2437630743,0.0153201325,-0.1181699112,0.1589726955,-0.2877340019,-0.2249055505,0.4279428422,-0.0042330185,0.2148872763,0.0903220549,0.3507176936,-0.1093763113,0.0532986932,-0.1550857127,0.1416816264,0.404302299,0.0007882579,-0.5444186926,-0.2212259918,-0.1243489757,0.0089226775,0.1123208329,0.5317481756,0.0306649636,-0.1736357659,0.1770093888,0.2656615376,0.5368555784,-0.0887493938,-0.0582991466,0.1731205434,-0.3876774609,-0.1465545893,-0.3178415,-0.2131336778,0.0677269697,-0.2510716021,0.263696909,-0.203009069,-0.0140729835,-0.0744816959,0.4079804122,-0.1185537726,0.0340827964,-0.0952488631,-0.0083027743,-0.1466766298,0.0440276712,-0.1954734176,-0.1743855923,-0.3775610924,0.3725061119,-0.0666091517,0.2309618145,-0.0352425128,-0.1375807226,0.265242964,0.1389311254,-0.1037343889,0.1086483896,0.0686223432,0.0677358881,0.0991382301,-0.0500395894,-0.0214681067,0.2940168083,-0.343662709,0.1544417888,0.4693882167,0.0623107404,0.0337494723,-0.420491457,0.0856411383,-0.2320627272,0.0723931193,0.347905755,-0.080632478,-0.2999900579,-1.069737196,0.2451432347,0.1815061867,0.0347042903,0.1759747416,-0.2191982418,-0.0421901494,0.2608632743,0.1543843597,1.0892894268,-0.0507256314,0.5218431354,0.3229362369,-0.0731944144,0.3819906414,-0.3416105509,0.1129657924,-0.244715184,-0.4424582422,0.0531473085,-0.2846353948,-0.0072025433,0.0078241881,-0.1367422193,0.3234067857,-0.2796405852,-0.0524001606,-0.0252828151,0.0340554975,-0.1102772057,-0.0948151127,-0.0633685216,-0.0598191917,-0.1997699589,0.1812173277,-0.0717719346,0.1181937978,-0.1108608022,0.0298046488,-0.04482181,-0.1666142493,-0.0512364395,0.0092817806,-0.3341157734,-0.6155846715,0.283614248,0.3402102292,0.0996392891,0.179265514,0.053441681,0.0836941898,0.2451819032,0.4168071449,-0.0029478443,-0.0521088839,0.4235515296,0.0652486905,0.0677940995,0.2524892688,-0.019039575,-0.096982412,-0.8286117911,0.1709132344,0.3791937232,-0.2473125607,-0.4441283047,0.0542826056,-0.0023440337,0.0265491717,-0.0408351608,0.1820113659,0.1740788072,0.4315118194,-0.2256633192,-0.4154630899,-0.1995775253,0.2852520943,0.2003929466,-0.045226451,0.4193188846,-0.0519285202,-0.2008045614,-0.0527363978,0.2371236384,-0.0974297449,-0.4542705119,-0.1012882963,-0.0841003731,0.1255544573,-0.019005673,0.2705136538,-0.2472783327,0.0637085661,-0.2091897577,-0.0291788038,0.1038118005,0.0176787004,0.4277179837,0.3437690735,0.3647261858,0.3129191101,0.0152425794,-0.0564862154,-0.1305907518,0.0297765024,0.3241150379,0.3977801502,0.0183637086,0.0064171935,-0.3131085336,-0.0604932867,-0.1109374538,0.1440346837,-0.1140713394,-0.0874803364,-0.1520568579,0.2300266922,0.0748610124,-0.0025268281,-0.169527486,-0.1676239073,-0.1742764413,-0.2396644801,0.2976013422,0.2611989975,0.0210243277,0.1257654279,0.4362260401,0.0285326708,0.1622669399,0.1904217899,-0.158964932,-0.2615143061,0.1234544814,0.0164786372,-0.2788673341,-0.0519074686,-0.599006176,0.143350035,0.2141147405,0.3800849319,0.0123016732,-0.3274235427,0.3317849338,-0.1678716242,0.4022538066,0.5396655202,0.0275575705,-0.4948884547,0.2594493628,-0.0720313266,0.2253415436,-0.1419828832,-0.0395640843,0.1817880869,-0.1139530316,0.1323440224,-0.1584558636,0.3500945866,-0.5146568418,-0.0992891118,0.1379054487,-0.1613038331,0.2236180604,0.4979375005,0.046598129,-0.0668177158,0.1828082502,-0.0771138519,-0.1445321739,0.2808187902,0.7301182747,0.5393793583,0.4825175703,0.0395526029,-0.0672645941,-0.4697543085,-0.2155254185,0.1976953298,0.0375333838,0.5161758065,0.2281161696,0.0463414192,0.1134193763,-0.0799599364,0.1270460784,0.336656481,-0.1223196089,-0.042072054,0.2104701251,-0.3991296887,0.0812123939,0.1484832466,-0.4271040857,0.2103952616,0.1051131934,0.170705542,-0.1856763512,-0.111417748,0.2426232994,-0.1366685182,0.1284887642,-0.1080194265,-0.0554121286,0.1767769009,0.0271324012,0.1898210347,0.1285141706,-0.051371187,0.3518345356,-0.1466553062,0.0571049117,-0.103277728,0.1627274603,0.0270891506,0.0942607075,0.3115408123,-0.3098682165,0.4923088849,0.0618408546,-0.0126401251,-0.1264349073,0.3408855796,0.2179021537,-0.5887755752,-0.0272438154,0.1398259848,0.0593104847,0.0964889973,0.1066811383,-0.4906990528,0.0379791483,0.0147474837,0.4293690026,0.1835746914,-0.0085616251,-0.0065229358,0.1072581634,0.4353663921,0.0826549903,-0.1552221775,-0.0474484824,0.0259117559,-0.428497076,0.0506531,0.3247933388,0.1212253124,0.2492528111,-0.1100715548,0.2826694548,0.1380500942,-0.1708900332,0.217368871,-0.0020115608,0.4330170155,-0.3053794503,-0.0873088166,-0.41097036,0.1781348139,-0.1709602028,-0.1909289807,0.1201006249,0.1029169634,0.0281698108,-0.2006617785,-0.1350136995,-0.0147139719,0.3194841444,0.2151916772,0.0709726214,0.2752601504,0.0149041666,0.0139682246,-0.1395524144,0.0865887702,-0.0084388126,0.2101175189,-0.2170508057,0.39561975,-0.1571722031,0.1808829904,-0.3153145313,0.2677330971,-0.1786733717,0.2012965232,-0.2949556708,-0.061833445,0.322712332,0.0421419963,0.0804623365,0.5686163306,-0.0181748904,0.2659771144,-0.2791575789,0.2963319719,0.4934827983,0.1330400407,-0.2201577425,-0.2162210643,-0.0268434361,0.1314427555,0.0620843135,-0.3097263277,0.0052932235,0.2125545293,-0.2343695909,-0.3316029906,0.2053728402,-0.054333806,-0.2618587911,-0.1250595897,0.5423796773,-0.0547792353,-0.0949317217,-0.0436527915,-0.03871084]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/510","title":"Version of numpy to use the library","comments":"Seems like this method was added in 1.17. I'll add a requirement on this.","body":"Thank you so much for your excellent work! I would like to use nlp library in my project. While importing nlp, I am receiving the following error `AttributeError: module 'numpy.random' has no attribute 'Generator'` Numpy version in my project is 1.16.0. May I learn which numpy version is used for the nlp library.\r\n\r\nThanks in advance.","comment_length":14,"text":"Version of numpy to use the library\nThank you so much for your excellent work! I would like to use nlp library in my project. While importing nlp, I am receiving the following error `AttributeError: module 'numpy.random' has no attribute 'Generator'` Numpy version in my project is 1.16.0. May I learn which numpy version is used for the nlp library.\r\n\r\nThanks in advance.\nSeems like this method was added in 1.17. I'll add a requirement on this.","embeddings":[-0.0217308123,-0.035936866,-0.1325140893,0.028069254,-0.0138459513,-0.0955159143,0.1382413656,0.208564505,0.195153892,0.2883733809,0.3901232779,0.5647602081,-0.3264506757,-0.1772346944,0.2545482218,-0.4155167937,0.2558472753,0.1544232517,0.0641539171,-0.2498569191,-0.4465940893,0.2496136874,-0.0539416708,0.6086594462,-0.1304111928,-0.0638025776,-0.0659989566,0.0015847152,-0.3209117651,-0.4154162705,0.1371644139,-0.0586813726,0.4173369408,-0.1541468054,-0.0001149852,-0.2631111741,0.5817897916,0.0145270424,0.0947816595,-0.5415332317,0.3599846661,-0.2434728742,0.1161146685,-0.1320894212,0.1149252132,-0.0119524729,0.3403695226,0.0282699913,-0.0258892514,-0.1023480147,0.2299793661,0.062332198,0.5442869663,-0.0170853063,-0.1914930344,-0.0845065042,0.0838638023,0.1854483783,0.4158837497,-0.250597626,0.1194931343,0.1976879686,0.2649782896,0.0194779765,0.6274604797,-0.1202160269,0.4131525755,-0.2635687292,-0.2294682562,0.257350713,0.3068662882,0.19982256,-0.3169952929,0.1957339048,0.0827362388,-0.2381158322,0.0039732349,0.2316205055,-0.1859661937,-0.1766357422,-0.010863713,-0.2609300613,-0.2837365866,0.4536567926,0.1673924029,0.4271886349,0.1460914761,0.1792851537,0.2573533952,-0.1743065715,-0.109450601,0.1971662641,0.1735471338,0.2233343571,0.2122558653,-0.0583120957,0.1734813452,-0.0231115706,-0.2046331912,-0.0185167324,-0.3267122507,-0.1792520136,-0.4373709857,0.1720219702,0.2394918054,0.1562304944,0.3343299627,0.0465536825,0.2841745615,0.0305882599,0.1534157693,0.1653660685,-0.1474713385,-0.1831774712,-0.00756235,-0.010426091,0.0167404097,-0.073661074,0.1120843217,-0.4979269505,-0.1133158505,-0.1391735375,-0.1143532917,0.0925774947,-0.0698049664,0.3541526198,0.1640752107,-0.0629141256,-0.1002721488,-0.3546003401,-0.0616739281,0.3524659872,-0.2493607551,-0.470630914,-0.1337919086,-0.0117800301,0.323182404,-0.2745202184,-0.304130286,-0.2039431185,0.0386823192,-0.0989627615,0.000180358,-0.0810561329,-0.3463960588,-0.0622941144,-0.0390600935,-0.2904673219,-0.1889507771,0.2138080001,-0.3084082007,-0.2582431436,-0.110167332,0.0910696536,0.195925042,-0.1085563675,0.1426841617,0.0201348737,-0.1813662052,-0.200219512,-0.0157594513,-0.4307922423,-0.1945033967,-0.1663512439,-0.166892916,0.1571138054,0.09659224,-0.2537865639,-0.1218264624,-0.1699345857,-0.1338137388,0.2338297665,-0.1543863118,-0.2130621374,0.238309592,-0.0183516685,0.7947320342,-0.3379612863,-0.1125856861,0.0239763465,0.1958198696,-0.40612486,-0.059668161,0.4793464839,-0.3789378107,-0.1704158485,0.3203155696,0.5940742493,0.0044302568,-0.1441716701,-0.1885215044,-0.5312308669,0.4913970232,0.1049409881,0.1727520972,0.0651764125,0.0376867838,0.0804962367,0.2049772888,-0.0572540537,-0.247518599,-0.0131579135,0.2381892949,-0.012274459,-0.0345644802,-0.1364994198,0.254963249,-0.130390197,-0.3168933988,0.060618259,0.1783967763,0.2292550951,0.0531962775,0.0643447116,-0.1235537976,-0.0752409026,0.07203722,-0.0654183701,0.3974335194,0.0255452897,-0.2275915146,0.0880052,-0.2712872326,0.1049717963,-0.7872636914,0.0545687005,-0.3162102401,-0.2314693183,0.1856021583,0.321482867,-0.0412039906,-0.102369979,-0.0584234744,0.1327745914,-0.2669433355,0.2215995044,-0.1532467008,-0.0101036569,0.0659466833,-0.0739621595,0.3291057348,0.3399540186,0.0503917746,0.2798424661,0.5786486864,0.1387599707,0.0488368236,0.0087098181,0.2964704335,0.2027317137,-0.0454242975,0.22973077,0.0572115704,-0.1730433404,0.3563598096,0.1154691502,0.1029251069,-0.1136728749,-0.4433079958,-0.0563078448,0.5467044115,0.0596165955,0.2767464519,0.2117886692,-0.2334416807,-0.1059126556,0.0581276715,0.2287420779,0.3336503208,0.0175045729,-0.3252829909,-0.0313116685,-0.1317009926,-0.1519592255,0.1401927322,0.1278725266,-0.0018749802,-0.2997200787,-0.0092431568,-0.0920051634,-0.0392924733,-0.0707804114,-0.3826495707,0.2375277281,0.1331774592,0.3652309179,-0.2270892709,-0.4761755168,-0.2809788585,-0.1842054874,-0.5064703226,-0.0983931124,-0.0634630099,0.039598994,0.1867501736,-0.03597188,0.1318917572,-0.3172332942,-0.0591149069,0.236968562,-0.0762394369,-0.1973803639,-0.4693325162,0.0325085893,0.1293919533,0.1052084118,0.2717398405,-0.2043821365,0.1401276141,0.2600786984,-0.1006855294,-0.093265906,0.0404728688,0.3480695784,0.4335059524,0.0692791492,-0.0433901064,0.0906204879,0.2583431005,-0.6114416718,-0.0194202662,0.093266055,-0.0834079161,-0.1223266274,-0.0464182496,-0.3892589509,-0.5913951993,-0.1272018552,0.0857136026,0.3966119289,0.3112340569,0.3121299744,0.044671014,0.0525208898,0.0585839115,0.2106284797,0.0129805841,0.0795469657,0.2477669865,-0.073382318,-0.4251493514,-0.0947875977,-0.118824929,-0.1377221495,-0.1817935109,0.150801748,-0.3300594091,-0.0132797593,0.0642752275,0.0842116401,0.1823756993,0.3733659387,-0.0283229258,-0.0587640069,0.1823022068,-0.0523317195,0.2514538169,0.2805742025,0.4423525333,0.4937015474,0.1127205342,0.0391131453,0.1956380606,0.2122749835,-0.493734628,0.0764903873,0.2330483943,-0.0043500252,-0.0182413533,-0.1132464856,-0.0298094619,0.2662193775,0.3194458485,0.0936526954,0.0163295381,-0.0505381748,-0.3421443999,-0.3402908444,0.0367459357,0.1736919731,-0.1010194644,0.1027403325,0.2570572495,0.0516010076,-0.0178442169,-0.275296092,-0.293092221,0.1626864523,0.4659411013,0.2322887182,0.2617291808,-0.1011862308,-0.5353935957,-0.2774604857,0.2927829623,0.1867445111,0.3731660247,0.1184756905,0.08074411,0.1794382334,-0.1907123923,0.0994277894,-0.130810827,-0.1520885676,0.10337843,0.129181087,-0.0497887954,0.0449460633,-0.0041935709,-0.1487160027,0.3080886006,0.0802557245,0.1219829023,-0.094587341,0.1968966275,0.2126959711,-0.1705798805,-0.2043403387,-0.3111718893,-0.0917423964,-0.154169172,-0.2962228358,-0.3719168603,0.0915572345,0.2062623948,0.2402590066,0.06891343,-0.3013079166,0.0713753551,0.1266505867,0.0200586617,-0.1926350445,-0.3761070073,0.1470488608,0.1011740491,-0.2435718477,-0.3453870416,-0.0951398462,0.3017825782,-0.2071052641,0.0271513239,0.1802430898,0.0582813248,-0.2119587809,0.0973676741,-0.1636725813,-0.2545107305,0.0651434064,-0.6037247777,0.3066554666,0.0569636449,0.2135917842,-0.355055064,0.5274339318,0.0722524524,-0.005889459,0.3273260593,-0.4485221207,-0.0111734197,-0.0048128664,0.045761276,0.7797033787,-0.0000684387,0.0320620909,0.2554924488,-0.0658107474,0.6031081676,0.1565392315,0.1735357791,-0.4816448689,0.113457635,-0.0846624374,-0.117698729,0.6107546687,0.0118390843,-0.1073408872,0.6082800031,0.1969384998,-0.4833672047,0.0700743198,0.5598540306,0.0778667778,0.1294127256,-0.534416914,0.0015000344,-0.3213026524,0.5169433355,-0.1696436852,-0.1081719771,-0.0694788247,-0.0973937809,-0.0163922068,0.031836912,0.1168860048,0.2906290591,0.6350306869,0.0235443097,0.1665973216,0.1873728931,0.4075179994,-0.180853799,0.0739158168,0.0465074033,0.0770343617,-0.1910115629,0.1207102388,0.4612988234,0.4130579531,-0.394389689,-0.1931154877,0.2795176804,-0.0355959199,-0.2546185851,-0.3156577051,-0.0266513191,-0.0063767321,0.3824508488,-0.3049358726,-0.0716645345,-0.1520326287,0.0785025582,0.1014943048,0.2460110039,0.0845441744,0.0365168117,0.1721831411,0.0397952422,0.0595463365,-0.0969085842,0.2198208421,0.0516144931,0.1656813622,-0.0846495479,-0.0283992663,-0.0767615139,-0.0268838611,-0.3058400452,-0.2211679667,0.1387557685,0.2226415128,0.1124513894,0.0413435064,0.3080591857,-0.0981693268,0.1582844108,-0.1558645219,-0.2982054651,-0.1403629631,0.1777842045,-0.1337140203,-0.2131894976,-0.0327903181,-0.0219381135,0.3597118855,0.0215925127,-0.2203498781,0.0403460823,-0.0051171547,0.2088282704,0.1505788416,-0.0349669717,-0.3847933412,0.1274237335,-0.0014689481,-0.2296186239,-0.4922756553,-0.1202665791,-0.1882554144,0.0983695313,-0.0099100396,-0.2176237404,0.2163748443,-0.244579643,-0.0432640873,0.2704946697,-0.4097151458,-0.3173872828,-0.2183305323,0.4455193281,-0.1849891841,0.0334976204,-0.0859214142,0.3244090378,-0.3353485167,0.3678446114,0.0352950469,0.1887072474,0.3394546211,-0.1222919747,-0.8591285944,-0.0085647209,-0.6065402031,0.115559943,0.151867196,-0.1954690218,0.1737019867,-0.2344572842,-0.2220875174,0.2246530056,-0.0963442475,-0.0232462361,0.0901703984,0.088803418,-0.2576099932,0.1966322958,0.0064806221,0.1179969981,-0.0990390554,0.2117509991,0.2546635568,-0.0762074813,0.0249268729,0.0108025689,-0.0025676736,-0.6233060956,0.2217756063,0.3310471177,0.2918611467,0.041844409,0.1935825497,0.1287200898,0.0111987609,0.2633214295,0.0633068681,-0.037333671,0.2631056607,-0.0175369158,0.0696774349,0.1479154825,0.0500192307,0.5351456404,0.1083689108,0.1678683311,0.2207513005,0.6887670755,-0.3609095514,0.049786564,-0.1175009459,0.3411080837,0.1495263278,-0.2341774702,-0.2045268267,0.0543507896,-0.4532110989,0.0231887139,-0.0686198026,0.0434098504,0.0144623816,0.3055369556,0.0199728701,-0.4412520826,0.1589882821,-0.1048736274,0.316096276,-0.1741710305,0.0644668117,-0.0970513374,0.0203446858,0.150515452,0.2155666202,0.2434814423,0.2625699043,-0.1379041374,-0.0188902263,-0.2740520537,-0.2428899109,0.2518094778,0.0441819094,0.4233146906,0.1735565811,0.0239184145,0.0176734086,0.0093644634,0.0886444226,0.0032607815,0.3612545729,-0.3148952723,0.9499719739,-0.1240187436,-0.2319941819,-0.1835771799,0.2177206427,-0.4107674956,0.2448221743,-0.0682366714,0.2299148291,-0.1912906915,-0.4018255174,0.0153510021,-0.1770698875,0.093764849,0.2788654864,0.2421730757,0.1108698472,-0.3392418325,-0.4093896747,-0.1052724123,-0.0416229106,0.0879576802,-0.1182307974,-0.1000687703,-0.0936383232,0.0656758249,-0.1881935596,0.1174932942,0.0725576207,0.1553100199,-0.124355346,-0.1129889563,-0.0245068129,0.1557133496,-0.0293864105,-0.0712459832,-0.3220122755,-0.6207131147,-0.0384723619,-0.2078833431,0.0239670053,0.126977101,0.7133052349,0.2449607849,0.1892892718,0.2083445489,0.1137781888,-0.2316340953,0.0708007514,0.1615182012,-0.1521316767,0.1527262479,0.094780609,0.5412036777,-0.1381214261,0.2821508646,-0.342456162,0.3896830082,-0.5016524792,-0.0349056982,-0.3041777611,0.0458113626,-0.1040554121,-0.016064439,0.1319736242,0.0922811702,0.0648049638,-0.100699842,-0.475707978,0.0902403817,0.0372454748,-0.3849801719,-0.2723801136,-0.2732599974,0.0146991666,0.1048194095,-0.1656322479,-0.3864678144,-0.0123482384,0.2886557579,0.1602869034,0.1672171652,0.1097835079,0.0489088371,0.0592211001,-0.0579529442,0.2159820646,-0.1676485986,-0.0619653426,0.0967414752,0.0811994746]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/510","title":"Version of numpy to use the library","comments":"Thank you so much. After upgrading the numpy library, it worked.","body":"Thank you so much for your excellent work! I would like to use nlp library in my project. While importing nlp, I am receiving the following error `AttributeError: module 'numpy.random' has no attribute 'Generator'` Numpy version in my project is 1.16.0. May I learn which numpy version is used for the nlp library.\r\n\r\nThanks in advance.","comment_length":11,"text":"Version of numpy to use the library\nThank you so much for your excellent work! I would like to use nlp library in my project. While importing nlp, I am receiving the following error `AttributeError: module 'numpy.random' has no attribute 'Generator'` Numpy version in my project is 1.16.0. May I learn which numpy version is used for the nlp library.\r\n\r\nThanks in advance.\nThank you so much. After upgrading the numpy library, it worked.","embeddings":[-0.0230417755,0.0587460697,-0.1424418986,0.1102989241,0.0604320914,-0.0701470748,0.1647360474,0.2216649652,0.1902176142,0.2875782847,0.350508064,0.5082064867,-0.3525879383,-0.1883734763,0.2187626511,-0.4138630033,0.2175045013,0.1571477056,0.0268467572,-0.2394560575,-0.4282829463,0.2229597718,-0.0689513013,0.6512901783,-0.1564567387,-0.0709198862,-0.0293639209,-0.0077940007,-0.3166870773,-0.414119482,0.1287453175,-0.0636266321,0.4369371533,-0.10610421,-0.0001122024,-0.2479391843,0.5385162234,0.0261972956,0.1134954691,-0.5189345479,0.3627840579,-0.1946241111,0.0496621951,-0.1447806209,0.1409762949,0.0418988653,0.3101233542,-0.0375675373,-0.0392224826,-0.0628054142,0.2682030201,0.0548785701,0.5681495667,-0.0108820945,-0.2323837429,-0.0803296566,0.1129933,0.2051335275,0.3511664271,-0.2525427341,0.1581658125,0.1857919246,0.2253884971,0.0785815641,0.6507299542,-0.1633629054,0.3697710335,-0.235985145,-0.190450117,0.25384444,0.2850823402,0.2020891607,-0.2776203156,0.2536578476,0.0935304165,-0.220962584,0.038826257,0.2360924929,-0.1263049543,-0.1677855402,0.0095179137,-0.222492978,-0.3029130399,0.4443202019,0.1185114756,0.3425175548,0.1181033105,0.172468707,0.2419205606,-0.194723472,-0.0503590666,0.1865182966,0.160037294,0.2214049697,0.198676154,-0.0147862481,0.1255711019,-0.0477869548,-0.2059313208,-0.0504524298,-0.2985902429,-0.1345874667,-0.3723447621,0.1563756019,0.2399391383,0.148533687,0.3067619801,0.1004028991,0.2837051749,-0.0017343578,0.0874068588,0.1240988448,-0.2042235136,-0.2251146287,0.0308547672,-0.0646629184,0.0414180718,-0.0793984234,0.1022592336,-0.4533910453,-0.0564082712,-0.1387154609,-0.1632455587,0.1465843469,-0.1074685752,0.3444688618,0.2066928446,0.0019835648,-0.0874590576,-0.3411920369,-0.0679840669,0.3105290532,-0.2530358732,-0.4632886052,-0.1463994682,-0.059600614,0.318045944,-0.2581760287,-0.2658281326,-0.220504567,0.0606228039,-0.1223791614,-0.0598641485,-0.1104598865,-0.3411158621,-0.0006963323,-0.0175360832,-0.2709775865,-0.147640124,0.1943186671,-0.3192361593,-0.249897331,-0.1389815956,0.1173105463,0.2078158259,-0.0856893659,0.1085254326,-0.05005062,-0.1548241973,-0.2241758257,-0.0313768797,-0.4612494409,-0.1234215423,-0.1902691424,-0.1633706987,0.1466114819,0.1561618447,-0.2388308793,-0.1123123094,-0.1636349261,-0.1563417763,0.2766105533,-0.1320408285,-0.1959423125,0.2150252908,-0.0412018374,0.7778900266,-0.3370402455,-0.136914134,0.0417218544,0.2027535141,-0.4204610586,-0.1174209416,0.4351083934,-0.3173748851,-0.1413674653,0.2995675504,0.5585716367,0.0259602144,-0.1356610209,-0.2064825445,-0.4862895906,0.5391215682,0.1100346446,0.1921138167,0.0706528202,0.0265731532,0.0960283205,0.1711423993,-0.032008715,-0.2253991812,0.0053278054,0.1907286346,0.0062683439,-0.0691500604,-0.1269810647,0.2765134275,-0.113059938,-0.3539266586,0.0546824746,0.1870908439,0.2411067039,0.0551638044,0.0426049046,-0.1279129535,-0.091281414,0.1059543341,-0.0671251193,0.3774759173,0.0637505874,-0.1945491582,0.1010446921,-0.2528136671,0.1018720195,-0.7236810327,0.1024357975,-0.3602537215,-0.2146774232,0.1944534183,0.2678383291,-0.037194591,-0.1004959345,-0.0819853693,0.1089468226,-0.2692813873,0.2277349383,-0.1931604445,-0.0343493931,0.063938044,-0.138827011,0.3494164646,0.3695811033,0.0611981824,0.2475497425,0.6522328854,0.1395842433,0.0421520472,0.0599570684,0.261394918,0.2358780503,-0.0298087671,0.2004385293,0.1246725693,-0.140042901,0.3912900388,0.1416036785,0.146952644,-0.1146814898,-0.4749297798,-0.1010206193,0.5184028745,0.0785271376,0.2257183343,0.191310972,-0.3029630482,-0.0754425004,0.0619850233,0.1895535886,0.3275952041,0.0133151207,-0.3282137811,-0.0261393879,-0.1567196548,-0.1460415423,0.1264741272,0.1038151681,0.0259475093,-0.2788936794,-0.0403660983,-0.0712029487,-0.0075110281,-0.0732868016,-0.3836455941,0.2876871228,0.0600319952,0.3707722127,-0.2122280598,-0.4347816706,-0.2727116048,-0.2094180286,-0.507275939,-0.0856714845,-0.109485887,0.0260264184,0.2567792535,-0.0293544475,0.1680545509,-0.406385839,-0.0577113405,0.2357577682,-0.0959099159,-0.1958942115,-0.4667716622,0.0579143874,0.1934875399,0.135900259,0.2386782318,-0.1993270963,0.1487330198,0.3067723513,-0.0609575287,-0.1181347445,0.0323595069,0.3549681902,0.3849944174,0.1530881822,-0.0702148899,0.0501310751,0.3112064302,-0.6154255271,-0.0166003928,0.1463223696,-0.0514293872,-0.1141709015,-0.0077953162,-0.3921297491,-0.5957376361,-0.1448276788,0.0571159385,0.4058895707,0.2802901268,0.3435550928,0.0723526701,0.0955468565,0.0595486313,0.1977039725,-0.0157605615,0.0719789118,0.2731529176,-0.0992118567,-0.4692971408,-0.1003162414,-0.0988043919,-0.1109955609,-0.1834321469,0.1145538613,-0.2499616295,0.0147753898,0.0306580607,0.0666907877,0.1387410462,0.3455575705,-0.0424161255,-0.0628019646,0.1681746393,-0.0169535745,0.1997095346,0.2270037383,0.408618331,0.4637135863,0.1109261066,-0.0081360713,0.1975439936,0.2376259863,-0.5082518458,0.0706565604,0.2276420891,0.0297721978,0.0073426384,-0.1545264572,0.0294228345,0.2450495064,0.350466758,0.0585857593,-0.0150404181,-0.0278944876,-0.3644146621,-0.3135895729,0.0250879675,0.1697963774,-0.0836626068,0.0952752158,0.2599568665,0.0662055388,-0.0413242914,-0.2406982332,-0.30272156,0.1815474629,0.455545485,0.1982375383,0.2633092701,-0.1271222681,-0.4945287406,-0.2775929272,0.3088920414,0.1404982358,0.327193141,0.134544462,0.1270618439,0.1895319074,-0.1851771176,0.119768545,-0.1315535605,-0.1409491897,0.1163798794,0.1262321323,-0.1099247411,0.0826831833,-0.0179186147,-0.1430846155,0.2966104746,0.0266392194,0.1057083979,-0.0384860523,0.1525880694,0.2002092451,-0.1854775995,-0.2329517752,-0.3124728799,-0.0633684024,-0.1832767725,-0.3422147632,-0.3940199614,0.1095646173,0.2350468189,0.1883278489,0.0117759388,-0.2597732842,0.0698739067,0.1418337822,0.0521614701,-0.1969386041,-0.3536034822,0.1327338815,0.0695965737,-0.1270062476,-0.3291597366,-0.1160431281,0.3129285872,-0.2221635133,0.0408482216,0.1324957013,0.0055560288,-0.1910522282,0.0697894394,-0.1379254609,-0.2195637077,0.0628650039,-0.5455786586,0.3026946187,0.073039189,0.1392245144,-0.3735049665,0.5016009808,0.0415918529,0.021379808,0.320715785,-0.429572165,-0.0205138344,-0.0428588614,0.0772109553,0.7816240788,-0.0526061617,0.0390595905,0.236869663,-0.1141585335,0.5930085182,0.2134640962,0.0874213055,-0.4942667484,0.1170185357,-0.0795190856,-0.1355597973,0.6088715792,0.0074928408,-0.1240720078,0.5534377694,0.1863689274,-0.4640606046,0.0305760559,0.5398049355,0.0938539058,0.1707660854,-0.5159168243,0.0349022262,-0.361947149,0.5030906796,-0.1917951256,-0.0923645794,-0.0675736815,-0.0941405594,-0.0148989568,0.014044066,0.0095964754,0.2984943986,0.6154732704,-0.0070879324,0.2051034123,0.2178409547,0.3972374499,-0.1782265306,0.0462577827,0.0144773312,0.0759247765,-0.2579556108,0.1455678642,0.4604550898,0.3723807633,-0.3976374567,-0.1993945092,0.3092298806,-0.0282781199,-0.2474288493,-0.2715922296,0.0168984216,-0.0830980465,0.3377696574,-0.3125579357,-0.1008299217,-0.1501384825,0.0720314458,0.1252520233,0.2293660045,0.1259737462,0.0620651245,0.2113361955,0.0172228385,0.0411054343,-0.0494969301,0.2308980078,0.0634914488,0.173748672,-0.0897824764,-0.0534350239,-0.0948814154,0.0322818868,-0.2769626975,-0.2062477469,0.116926387,0.2247645259,0.1397359669,0.0428908914,0.2301014513,-0.089726232,0.1689947695,-0.1305616796,-0.2848933339,-0.1813863069,0.1465870887,-0.1556390077,-0.2419095486,-0.0386497751,0.0186890718,0.3789910078,0.0616707094,-0.2498445213,0.0608952604,-0.0065614977,0.177366659,0.1449210346,-0.0705221146,-0.3852585852,0.0741126835,0.0153508438,-0.2253526747,-0.5095077157,-0.1481477767,-0.1861064583,0.0788236111,-0.0014690303,-0.2550166547,0.2166714072,-0.2447014004,-0.0446043871,0.2481542677,-0.4033469558,-0.2692141533,-0.2411769927,0.4310518801,-0.1895548105,0.0182980057,-0.113694407,0.3035233319,-0.3714706302,0.3506653309,0.0334021188,0.157929033,0.405543685,-0.1212541759,-0.8910323381,-0.0295443423,-0.643196404,0.0685966685,0.2296698987,-0.2298865169,0.1877039224,-0.1671691537,-0.2457676828,0.2258531153,-0.104279764,-0.0346642435,0.1260725707,0.1110885665,-0.2570075393,0.2218847424,-0.0866067708,0.0725611076,-0.1129366681,0.2091326267,0.2320159525,-0.1538135111,0.0165422987,0.0045772051,0.0321045965,-0.6327697635,0.2310894281,0.3367103636,0.2593335509,0.0119667267,0.2150826305,0.1661189944,0.0212620553,0.2789312899,0.052566506,-0.026398804,0.1838980019,-0.0238123871,0.0428026319,0.1671018451,-0.0008793604,0.5695611835,0.0478159525,0.2049698979,0.2295559645,0.6759201884,-0.3925817609,0.0722685605,-0.0825579986,0.3268382251,0.1378169954,-0.1887001246,-0.1874413788,0.0686091557,-0.4343878329,-0.0010692619,-0.0576968826,0.0112648765,0.0882679671,0.355196625,-0.0478018597,-0.468141675,0.1822800487,-0.0686363727,0.263471663,-0.1859623045,0.0818911791,-0.0836759433,0.0305189397,0.1718582213,0.265704602,0.2976500392,0.259727478,-0.1915446967,-0.032517463,-0.2804623842,-0.2074335665,0.248292163,0.0114694983,0.3905882239,0.1543615311,0.0322782472,0.0336832367,-0.0208962113,0.0698728263,-0.0135587109,0.3720460534,-0.3081141412,0.9181989431,-0.166085273,-0.1990537941,-0.1876412928,0.1862151325,-0.4463690817,0.2666329741,-0.070490092,0.1882477999,-0.2064370811,-0.3993350565,0.0260310266,-0.1586426944,0.1307472885,0.2980472744,0.253210932,0.1032379568,-0.344661057,-0.454936564,-0.0538995937,-0.0321331397,0.1465165019,-0.0968419909,-0.1362642646,-0.123058632,0.0594331548,-0.1084254235,0.154209286,0.0470910482,0.0910913944,-0.1720979661,-0.0528792106,-0.0481772907,0.1676159054,-0.0192787237,-0.0418403149,-0.2824164331,-0.5900307298,-0.0056417077,-0.2167733759,0.0769632012,0.0858865455,0.743486464,0.2619650364,0.1940930635,0.2294193059,0.128071174,-0.1879013926,0.1068153679,0.1232542023,-0.1048304215,0.1902511269,0.1744700223,0.5403839946,-0.1481960565,0.1321694851,-0.326399684,0.3996653259,-0.473995477,-0.010505978,-0.304556191,0.081571281,-0.0730130225,-0.0019493393,0.1047257185,0.0796515048,0.0989068076,-0.0834675431,-0.4883708954,0.0584851392,0.0474420115,-0.3390646279,-0.2939167023,-0.2472036034,0.0090348283,0.0443163738,-0.1661784947,-0.4500996172,0.0133496402,0.2960585952,0.1564371586,0.1482731551,0.0560653247,-0.0014316748,0.1016428694,-0.0670200437,0.2455322444,-0.1820179075,-0.0703999549,0.119115375,0.1079816669]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/509","title":"Converting TensorFlow dataset example","comments":"Do you want to convert a dataset script to the tfds format ?\r\nIf so, we currently have a comversion script nlp\/commands\/convert.py but it is a conversion script that goes from tfds to nlp.\r\nI think it shouldn't be too hard to do the changes in reverse (at some manual adjustments).\r\nIf you manage to make it work in reverse, feel free to open a PR to share it with the community :)","body":"Hi,\r\nI want to use TensorFlow datasets with this repo, I noticed you made some conversion script,\r\ncan you give a simple example of using it?\r\n\r\nThanks\r\n","comment_length":73,"text":"Converting TensorFlow dataset example\nHi,\r\nI want to use TensorFlow datasets with this repo, I noticed you made some conversion script,\r\ncan you give a simple example of using it?\r\n\r\nThanks\r\n\nDo you want to convert a dataset script to the tfds format ?\r\nIf so, we currently have a comversion script nlp\/commands\/convert.py but it is a conversion script that goes from tfds to nlp.\r\nI think it shouldn't be too hard to do the changes in reverse (at some manual adjustments).\r\nIf you manage to make it work in reverse, feel free to open a PR to share it with the community :)","embeddings":[-0.1354902536,-0.0565532297,-0.1098584607,-0.1969328523,0.165716216,0.0915980563,0.3340383768,0.4940679669,-0.1472647488,0.1399837881,0.0302083679,0.3426104784,0.0195342824,0.3617252707,0.3380131125,-0.3144352734,0.0504774712,0.2795871794,-0.3302764893,-0.1137338728,-0.0358067453,0.10837356,-0.111526899,0.0685933232,0.1368517131,-0.0301686917,-0.0850050896,0.2507574558,0.1226198226,0.0422674604,0.1680342853,0.1560598016,0.3117433488,0.728592217,-0.0001057781,-0.0028073546,0.0753042623,-0.0946940407,-0.0748347789,-0.3386337757,-0.1552296132,-0.2183431983,0.0090980828,-0.3499976695,-0.2264173031,0.0812482461,0.1151307598,-0.1407833397,0.5385851264,0.5260947347,0.2700379193,0.1933607906,-0.0691369101,0.1015990376,0.0665854886,0.1731276512,-0.2012015432,0.044669345,0.3796443939,-0.1914269626,0.2758765817,0.23360838,-0.0874221846,-0.1543228477,0.3159394562,0.0481564514,0.131105572,-0.4717644453,0.008744332,0.5226515532,0.5459263921,-0.4398654997,-0.0945537388,-0.1390538067,0.0545596778,-0.3348767757,-0.4834914804,0.1121080294,-0.1392924339,0.2684394717,-0.5411630869,-0.2216578275,-0.402703464,0.1094899997,-0.0812656283,0.0930473581,0.0270113312,0.0200878419,-0.0979301855,-0.2529070973,0.0744435787,-0.0001649336,0.2594025731,0.1288956255,-0.129897818,-0.3237948418,-0.0438239872,-0.4010471404,0.0947273746,0.0584126823,-0.0523551367,0.2425812632,-0.3122434318,0.1642353833,-0.1862477213,0.2841514647,0.1887120306,0.2639867365,0.1826592833,-0.2470732033,0.0311288536,0.1899619997,-0.2797750235,-0.1054207385,-0.1855313182,-0.0917769894,0.3033269048,-0.1337188035,-0.1332099736,-0.2010025382,-0.0339619778,-0.0861754417,-0.1275319308,0.2554565668,-0.0343337655,0.097173132,0.2132084966,0.2463805079,0.0076521351,0.0132319778,-0.0899327248,0.1835042089,-0.2858375013,-0.1959553808,0.0537286289,0.1041620076,0.0944256186,-0.1755783111,0.0349992812,0.305578649,0.3406420052,0.369146347,0.1823938936,0.0859625936,0.154000029,-0.0796315372,0.237705946,-0.1891966909,-0.3463330269,0.3555153012,-0.3057293296,-0.1236113682,-0.1417661756,0.2352067381,-0.2167689949,-0.4702011645,-0.2125683427,0.4416574836,0.0098987501,0.0000363395,0.1393336207,-0.2634074092,-0.2979614139,-0.2006848603,0.0635969266,0.2533058822,-0.6270143986,0.0087208096,0.0280472562,-0.3241333961,0.2353955954,0.0472910367,-0.0420394689,0.1241582856,0.1520136893,-0.0321764946,0.8314973712,-0.4936859012,0.0705592558,0.375086844,-0.182338655,-0.1454542726,-0.1348911971,0.0472492017,0.1545653939,-0.1671396941,-0.2246347368,0.4862645864,-0.3266904354,0.2609779537,-0.0088492017,-0.187792778,0.3314785957,0.1990187913,0.0573097691,0.2233078778,-0.0590549372,0.0614016205,0.3003413379,-0.3258717954,0.0479201935,-0.0598340705,0.1043673828,-0.0128765935,0.1901646405,-0.2196418494,-0.3102653027,0.051224947,0.0617602915,0.1311339587,0.242563054,-0.1107348651,-0.2938144803,-0.1632141918,0.109819375,0.0496342853,0.1978273392,-0.2180907726,0.081186451,-0.127427429,-0.4316331446,-0.0382887237,-0.0973472819,0.0361173637,-0.1963023245,0.3646839559,0.148756057,-0.1214118525,-0.0107458867,0.2152694613,-0.0853616744,-0.5249539614,-0.0451358519,0.3602922559,-0.241701141,0.3648475707,0.2504191399,0.3131833673,0.2691744566,-0.1689711362,0.2763992548,0.1496813744,-0.0801793337,-0.0926614776,-0.3937797844,0.4031621516,0.0032226385,0.246151194,0.0937030539,0.1271241903,0.3030202687,-0.3468945324,-0.236931026,0.0371233709,0.0832642764,-0.0511141159,-0.292799592,-0.1821140796,-0.442122519,0.4492573142,0.4657908082,-0.4190252125,-0.0151990587,0.3167186379,-0.0855055675,-0.0136619536,0.1080683619,0.2586348057,0.021684086,0.1125841439,0.0199201442,0.2573532462,-0.1341613233,-0.0874178782,0.1846426725,-0.0575618483,0.4302837551,-0.2699303031,0.018453259,0.0372076929,0.0475731604,-0.2376663536,0.0429766402,0.3551800251,-0.1561983079,0.1930073798,-0.4067861438,-0.4654603302,-0.2318651378,0.0395088196,0.2529284656,-0.1253548861,0.0527370051,0.0213936809,-0.0492600612,0.1367943436,0.2264103889,0.2378944606,0.1030076072,0.0996955484,-0.2383313179,-0.2714831233,-0.37849015,0.2252317965,0.3144752979,0.1246517226,0.3925198019,-0.3350910544,-0.0369047485,-0.4471805692,-0.3675509691,0.1797019392,-0.3059438467,-0.1407954842,0.1060349494,0.1329557449,-0.0322531462,0.056858439,-0.1369748861,-0.0169070587,0.0253216904,0.205211997,-0.2348159254,-0.2529453039,-0.338911891,-0.2970342934,-0.205621779,-0.1967361569,0.1507043839,0.3304749429,0.3571401536,0.1313921362,0.2177330703,0.4648527801,0.2672098577,-0.1147541106,0.1104004532,-0.1542226374,0.2004641145,-0.39842242,-0.4548293352,0.2888484895,0.0177486539,0.1205377132,0.1832213849,-0.238778919,-0.2569132745,-0.1529293507,0.0101189315,0.0296911895,0.1653308123,0.2616771162,-0.0988902897,-0.1023724079,-0.1688304096,-0.0462651253,0.1744609475,-0.0326427966,0.1593564004,0.0931049585,0.5094342828,0.0120586604,0.57361871,0.0804745555,-0.4635077715,-0.1816733629,-0.3495344222,0.109951362,0.2126186639,0.0766795054,0.2035273761,0.1046320051,0.1543194056,0.2083739787,-0.0070091081,-0.2179463357,-0.0320741236,-0.0371238887,-0.1571473926,-0.194051668,0.0228558909,-0.3748559356,0.0867139399,0.0973369628,-0.2412240803,-0.257076472,-0.0799961388,-0.0409280099,0.2149729431,0.2032997608,0.0429554507,-0.534958005,-0.164212361,-0.4318338335,0.2228109539,-0.2035095841,-0.05206028,-0.0591831133,0.1090996116,0.0152425962,-0.129447028,0.5048212409,-0.2355551571,0.0115836812,0.2057328671,-0.1228235364,-0.3891153038,-0.0865115672,-0.5830419064,0.4415954947,0.1200191677,0.2497491539,-0.0956403688,-0.2873215377,0.3748546541,-0.028056873,-0.1595231295,0.0119609265,0.1875294447,-0.4287967682,-0.3779330254,-0.1595683992,0.0658505931,-0.2024733424,-0.4159742892,-0.2461671978,0.1638126671,0.2247572541,0.3919678926,0.2527739406,-0.079042621,0.3548322618,-0.1536442339,-0.1274601519,0.0106801186,0.0386051424,-0.0318938345,0.0507644713,-0.24217695,0.0336724445,0.0372343659,0.3352758884,0.2374717891,0.0662430599,-0.1106176972,-0.0915659666,0.1527935863,0.0967044234,0.1496431231,0.2475595325,0.1688019335,-0.4757455587,-0.5301387906,0.3401373327,-0.0074000182,-0.1359575987,0.0993587226,0.0928995684,-0.2327264398,0.1926203668,0.2273746878,0.9439206123,-0.2948634326,0.1038292348,0.5949648619,-0.0607054681,0.7162030339,-0.4931621552,-0.1701778919,-0.2644360065,-0.0217541698,-0.0256889146,-0.1684122086,-0.0368597545,0.2574602067,-0.6424680352,0.2470122725,0.0667691752,-0.1334486902,0.2724005878,0.2310857475,0.1183705032,-0.161688298,-0.1287461668,0.2684848905,0.0740971863,-0.0578411818,-0.109194167,-0.2582225204,-0.1783963591,0.2283305228,-0.2402149141,0.2407479286,-0.0256201569,-0.2215587199,0.1087244526,-0.1106748357,0.4480558038,0.3364578485,0.6660100222,0.1680409163,-0.3463957012,0.0017299299,-0.0454422086,0.1112882644,0.0587019995,-0.0112943053,0.4692011774,-0.2207797617,-0.2251426131,0.2291727811,0.11377763,0.0434798971,-0.1949654669,-0.1645002961,-0.0549053103,0.0029231485,-0.1764400303,0.4043123126,-0.4062635899,0.2102762759,0.1911529601,0.1750521809,-0.2124243677,0.1991522163,0.0095465155,-0.1557264179,0.1329835951,-0.265642345,0.2591842413,-0.2787064612,0.3880563378,0.3272640109,0.0369455107,-0.2179433554,-0.1133805215,0.1392152011,-0.0868159011,0.1708610058,-0.0673044547,-0.0607714057,0.1837001443,-0.026000835,-0.1884330511,0.1745701134,-0.3014700711,-0.2322060317,-0.1320873946,-0.1107660085,0.1990896612,0.0888256356,-0.1003729403,0.2689449191,0.0134300301,-0.244442448,-0.4105583727,0.2353582829,-0.0595583543,0.2885017693,0.433680892,0.1194698438,-0.1040643901,-0.1778852195,0.0551258065,-0.0050949729,-0.1838559061,-0.3351588845,-0.1824339926,0.1167357489,-0.0160745867,-0.492631793,0.3430458307,0.0909259841,-0.1951273531,-0.0338493474,0.1562356651,0.0174518786,-0.0989516079,0.4800694883,0.2818796933,-0.0314351395,0.2551215291,-0.246338889,0.1036114544,0.2019971609,-0.1880027354,0.4242720902,0.1399915069,-0.1804623902,-0.263830483,0.1186977699,0.2879514992,0.2130420059,0.427151978,-0.0986881703,0.2778207958,-0.0777332783,0.2292592674,0.0675179064,-0.384288758,0.1852208227,0.2729553282,0.2478898764,-0.1421965957,0.1829955876,0.2702974379,-0.1790209264,0.0693479404,0.151263237,-0.1563562602,0.3313253224,0.259458214,-0.1847726554,0.2648240626,-0.3637621701,0.1419525146,0.2273327857,-0.3223602772,-0.0151453139,-0.3122265041,0.12772654,0.0570511073,-0.119721204,0.1553027481,0.3168907464,0.1523033381,-0.0843116343,0.3610110581,-0.0898131281,-0.0412837453,0.1228873134,-0.1211333498,0.311942935,0.2081070095,0.1741911769,0.1185057387,0.070254758,-0.0391794667,0.3936656117,0.128722012,-0.0405922085,-0.2223962545,0.07702481,-0.1651801765,-0.209395498,-0.3505536318,-0.1053746939,-0.1378686279,0.1823372096,0.0087056905,0.1171004921,0.3869176805,-0.1468653083,0.0901405811,0.0904913321,-0.0567420088,-0.0716627538,-0.0677580684,-0.1516686827,0.2467569262,0.2833401859,0.2651734054,0.0742943063,0.4210430086,0.0679794848,-0.1373471469,0.2447863966,0.1359493434,0.0686763078,-0.0924702734,-0.1591879874,0.1929430217,-0.1376971155,0.1138626561,-0.0290909614,-0.2743697762,-0.2508378923,0.2464634925,0.1080305949,-0.2730606496,-0.2436656058,-0.0376395881,-0.2719661295,-0.1385140866,0.5426119566,0.123278439,0.1378570646,-0.1872280091,0.1152628437,0.0063549485,0.5613746643,0.0490062237,0.3272159994,0.0605118126,-0.0555821434,-0.3118554354,0.1494045109,0.003506416,-0.145392105,0.0505326763,-0.0098452689,0.2392953783,0.1602766663,0.0475229323,-0.4620170295,-0.2005768418,0.1359281838,-0.1267751008,-0.3016201556,-0.2518890202,0.3050577343,0.0338141955,-0.2266815305,0.2748782337,-0.0354946777,0.0481881686,-0.2989723682,0.3004085124,0.231010437,0.0054128179,-0.0243231375,0.0861264095,0.265409261,-0.1187250987,0.0633611381,0.0004630914,-0.0467339084,-0.4390634298,0.1593444198,-0.1768251359,-0.1780132055,-0.091569446,-0.2608183026,-0.0369734019,-0.2370560914,0.0015795584,-0.2869719863,-0.2905570865,0.1905444562,-0.0497277863,-0.0407313406,0.2042836994,0.2939431369,-0.018323008,0.1006949097,-0.3634259105,0.0011141129,0.4306078851,-0.0717418045,-0.3922431171,0.0657982081,-0.0065416344,-0.1159202978,-0.0193988997,-0.3849144876,-0.0626806766,0.1641977578,-0.0941297412,-0.124769263,0.2935926318,0.1921745688,-0.0533421002,-0.2185537517,0.1653592885,0.0157192238,-0.2285494357,-0.2731185257,-0.3632560968]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/509","title":"Converting TensorFlow dataset example","comments":"In our docs: [Using a Dataset with PyTorch\/Tensorflow](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html).","body":"Hi,\r\nI want to use TensorFlow datasets with this repo, I noticed you made some conversion script,\r\ncan you give a simple example of using it?\r\n\r\nThanks\r\n","comment_length":8,"text":"Converting TensorFlow dataset example\nHi,\r\nI want to use TensorFlow datasets with this repo, I noticed you made some conversion script,\r\ncan you give a simple example of using it?\r\n\r\nThanks\r\n\nIn our docs: [Using a Dataset with PyTorch\/Tensorflow](https:\/\/huggingface.co\/docs\/datasets\/torch_tensorflow.html).","embeddings":[-0.1419396251,-0.4034663737,-0.0607448779,0.0852320343,0.3067462444,0.0435966738,0.4041775465,0.2861265838,0.0479127206,0.2570740879,-0.054075554,0.3698243797,-0.0009652911,0.3926182389,0.3297069967,-0.6211851239,0.0841344222,0.0982295275,-0.3374058604,-0.0590412319,0.0004504519,0.1306580007,-0.1981158555,-0.0912891701,-0.0883246586,-0.0149719398,-0.0851007476,0.1221195012,0.0599690303,0.0893349275,0.1880473644,-0.0443612486,0.2641646564,0.9007080793,-0.0001139598,0.1199630648,0.0931030214,-0.1822628379,0.0719780177,-0.480730474,0.1892479807,-0.1038336307,-0.0568598248,-0.3334319592,-0.3843234181,0.0645748302,0.1572875679,-0.3219797015,0.5534271598,0.5040550232,0.2060557008,0.44551754,-0.0970500484,0.0058584339,0.1129866838,0.2664681971,-0.2348084599,0.1821740419,0.6163927913,0.1130483747,0.1574643999,0.2912979722,-0.0955424383,0.0635853782,0.2167607397,0.1121512875,-0.2433822751,-0.4371806383,0.0106477104,0.4891938865,0.421656996,-0.3476823568,-0.0473571569,-0.0615217611,0.0202838238,-0.2647205591,-0.3171337843,0.2513303459,-0.134668678,0.1689284593,-0.6511039138,-0.1361819655,-0.5397370458,0.0981215313,-0.0679835752,-0.1127799675,-0.1911243945,0.106501177,0.1124010608,-0.1721286327,0.0578110516,-0.015123304,0.3810296059,0.0074050757,-0.0050125308,-0.0974533334,-0.3282433152,-0.2150655836,0.2010757774,-0.0521318801,-0.2056275159,0.2634401321,-0.1623853445,0.169167161,-0.2516342402,0.2354124337,0.1147572398,0.2649154365,0.1880252212,-0.064839907,0.0431814045,0.0948212594,-0.4259136021,-0.2373490036,-0.2520835996,-0.1535047442,0.3791403174,-0.1321630478,-0.0826449022,-0.1657318473,-0.2046619505,-0.0091448044,0.0435200445,0.470269382,-0.1751130372,-0.1221272945,0.3200878799,0.4176295102,0.0208236948,-0.0591776296,-0.0305860769,0.2927513719,-0.2097145021,-0.1121565774,0.0737081543,-0.1980111003,0.2678852677,0.0399260372,0.3687272072,0.1337426156,0.471377641,0.2482379824,0.1647976488,0.10548722,0.1916526854,-0.0240901113,0.2377195209,0.1324255615,-0.2519478798,0.3220311999,-0.3951957524,-0.1754214466,-0.2657574117,0.1169715673,-0.1971892715,-0.3116089106,-0.2724729776,0.2645527422,0.1299701631,0.1083452851,0.1627785414,-0.2956992686,-0.3806507587,-0.1221136227,0.2463354766,0.1886753291,-0.2913604975,-0.0684511513,0.2763668895,-0.2319553196,0.3408975601,0.2413065135,-0.2008634657,0.0529082492,0.0848251805,-0.0772137567,0.5744962692,-0.4749907553,-0.0091406023,0.1762588918,-0.002118676,-0.0105754593,-0.0705233216,0.0582419075,0.2643561959,-0.0091893133,-0.1445628554,0.3317347765,-0.2078689039,0.1893875301,0.0350574292,-0.1614903808,0.2815487087,0.3152970076,0.0214223564,0.0818621665,-0.175478369,-0.0661310405,0.0987092778,-0.363574028,0.1297979355,0.092580609,0.0920756757,-0.0456274822,0.1608266234,-0.113676928,-0.2212644219,0.1580907851,0.0895892382,0.133517772,0.146191299,-0.1451093256,-0.3454348743,-0.0893263072,-0.1177043095,-0.0397462063,0.0770263076,-0.2347398549,0.1152591035,-0.1270985454,-0.3620249927,0.0942125171,-0.21035932,0.1608030796,-0.3820556402,0.4191981256,0.00319906,-0.1342918575,-0.0855259374,0.095537141,0.0397986248,-0.5280249715,0.0397994332,0.2589019835,-0.0416937657,0.3449777961,0.1212123483,0.2004817724,0.3560218811,-0.3472864926,0.3784770072,0.0378976241,-0.0651452094,-0.0879005194,-0.3373423517,0.4681508839,-0.028579073,0.1886000484,0.0525043271,0.1988107711,0.1932123452,-0.2062474936,-0.1798029244,0.1721576601,0.2735787034,-0.0746073797,-0.0449196696,-0.1628725678,-0.4812940359,0.1705408245,0.4338665605,-0.4058136642,0.0171609148,0.2353515625,-0.3179831207,-0.0222688317,0.1488210261,-0.1066819429,0.0330175608,0.094873026,0.1447566003,0.2694982588,-0.0074736997,0.003641818,0.1666660756,-0.0396006517,0.2774893343,-0.4029468894,-0.1732613593,-0.0961208567,0.0409534685,-0.0326560996,0.1245834902,0.2342983037,-0.2668986917,0.2614420652,-0.196455732,-0.2894860208,-0.2612662911,-0.0425411686,0.0863433629,-0.0342536159,-0.0489304028,0.1870986968,0.1103989631,0.2271821797,0.4216972291,0.1313470453,0.1294521987,-0.048740875,-0.2068424225,-0.1943178475,-0.2976552844,0.1769041568,0.2722431719,-0.0655120984,0.3640612364,-0.5403103828,0.0199046377,-0.5801346302,-0.2799916565,0.1661766917,-0.2741619945,-0.057270091,0.1001745015,0.272685051,-0.1112967879,0.1259271055,0.0443949364,-0.1228689998,-0.1098970175,0.1861411333,-0.3315242529,-0.1139253676,-0.217921406,-0.2989134789,-0.1934589446,-0.0730465651,0.1804597825,0.2051845491,0.2112950236,-0.0716288537,0.3016239405,0.5777567625,0.0754454806,-0.0516792238,-0.0320209339,-0.5341496468,0.2688440979,-0.3571660221,-0.5912207961,0.2791449428,0.1086079404,0.0809449032,0.185989216,-0.2324410379,-0.3693557978,-0.1038163975,0.168702215,0.094716005,0.2675015032,0.2145919204,-0.2249369621,0.0486897007,-0.2586696744,-0.0829070732,0.2337828726,-0.1042545959,0.1461910307,0.0838355124,0.514567256,0.1403859705,0.9278008938,0.1451210976,-0.4722534716,-0.0635634363,-0.4951373637,0.1148425341,0.100253731,-0.0753645375,0.2110219151,-0.1447956413,0.2078035176,0.0314414948,-0.0342130885,-0.0252426118,-0.0467316397,-0.1369517148,-0.2040889561,-0.1181780249,0.1360300034,-0.2051982433,0.034210477,-0.1197801754,-0.1203835756,-0.1574088335,-0.0765812173,0.0007567001,0.2622559965,0.1730743349,0.13913697,-0.470210433,-0.010819708,-0.5485517979,0.389210701,-0.2108340114,-0.1051846892,-0.0717360526,0.124061726,0.0932111144,-0.0315477699,0.6287962198,0.0950962976,-0.1282638162,0.046495147,-0.4806596339,-0.5406784415,0.0661441535,-0.5372265577,0.4077055156,0.0784085691,0.307515949,-0.0590989403,-0.2178789377,0.3302050829,0.0992684439,-0.033274129,0.1120470017,0.033794526,-0.436652869,-0.3812949359,-0.0584103204,0.109367609,-0.2219913155,-0.3196611404,-0.1692928821,0.0456870571,0.3272436559,0.3240923584,0.2321347743,0.0338977017,0.0894692317,-0.1072694436,0.2209588885,0.1798519492,0.3237094581,0.1538009346,-0.1019667983,-0.5030066371,0.1136437133,0.24474518,0.3203787804,0.0699985251,0.0279829241,-0.2068301737,0.1329062879,0.1377996802,0.0355992541,0.2912479937,0.3376977444,0.3697826862,-0.4719485044,-0.4816091359,0.3364977837,0.1585470289,0.0940654948,0.2888782024,0.2215318233,-0.0845305771,-0.0376100019,0.17399773,0.9268850684,-0.3884472251,0.1937304288,0.6682819724,-0.1171870828,0.5490179062,-0.4127282798,-0.1440768242,-0.3086013198,-0.1043901891,-0.1323443949,-0.325748533,0.1444610506,-0.0112937577,-0.6283471584,0.1100619286,-0.1700413823,0.0530837588,0.1594268233,0.1073896512,-0.036043372,-0.1297847182,-0.1155998632,0.1848646551,-0.0425269939,0.1928306967,0.0083858054,-0.0846688747,-0.2124248445,0.0906054527,-0.1398128122,0.1979661435,-0.2911018431,-0.2244481593,0.1196608171,-0.2888183296,0.3641314507,0.5134046078,0.5194079876,-0.1119468287,-0.3414388001,0.0104811713,-0.1244557351,-0.120263204,0.0882092193,-0.124864094,0.2406342328,-0.2304455489,-0.2713536918,0.2660363317,0.1424032301,0.088507317,-0.0301720556,-0.0515202917,0.0812926441,-0.1673688591,-0.2876319289,0.2644255161,-0.1902801245,0.0951652974,0.1086570323,0.2241567671,-0.1962133348,0.0090654055,-0.0183034912,-0.1826345474,0.0850333199,-0.107472159,0.1350952834,-0.3523226082,0.3006063998,0.2784617245,-0.0031427243,-0.1066997871,0.165723443,0.4650788903,-0.1808641404,0.0608933121,-0.1482497305,0.0530458875,-0.1478974223,0.0117602237,-0.0606757067,-0.0525686778,-0.1471507847,-0.2203410268,-0.3322900236,-0.0685096681,0.2230931818,0.1579010189,0.0302019753,0.0324824154,-0.0585046113,-0.2741895318,-0.2899427712,0.4187195301,0.0858277157,0.2107316554,0.5091471672,0.0528423823,0.1120233759,-0.188928619,-0.0177579001,0.0247932114,-0.2102365941,-0.2238136977,-0.2434844226,0.0806916505,-0.0878899544,-0.5453620553,0.2522554398,0.0458014011,-0.1293357313,-0.0976308882,0.3290349543,0.4259286225,-0.138724342,0.3641325235,0.1901960522,0.004459023,0.0281756371,-0.3390733302,0.0340817384,0.1690910757,0.0137140816,0.2272927761,0.1859781146,-0.1856757998,-0.1165088564,0.0884004831,0.456790179,0.0057267719,0.3786994815,-0.0328486674,0.2997865379,-0.1931182295,0.2061681896,0.0527811795,-0.4609765708,0.2732961476,0.2491721213,0.1461075395,-0.1432177424,0.0586394109,0.2845825851,-0.1844053864,0.0456527695,0.1825743616,-0.137268275,0.5687559843,0.1527211517,-0.2869208157,0.2279777825,-0.2513832748,0.3418418765,0.1223317981,-0.3513161242,0.0454616845,-0.1464015394,0.273139149,0.1062365323,-0.1196602136,-0.0287507325,0.5595580339,0.0440680496,-0.0061163185,0.4853105545,-0.0990713686,-0.1490478665,0.0275485311,-0.210336864,0.2472427189,-0.0690029263,0.2948076129,-0.0462247618,-0.0657863915,-0.119745411,0.3366292715,0.0504249111,-0.0976017118,-0.1421394795,-0.0330577344,-0.0678784847,-0.2075569183,-0.1646140367,-0.2626519501,0.0784786493,0.1773516089,-0.0057677813,0.2456191927,0.236184597,0.0417263433,0.0390921272,0.1026791111,0.0439673625,0.0822334886,-0.0699055269,-0.1577041298,0.1902589053,0.3686679304,0.0505389608,0.0059122839,0.3986721337,0.1243964806,-0.0913434625,0.2714133561,0.0493766367,0.1011300236,-0.0906838775,-0.1092575714,0.1201479882,-0.05142726,-0.0176146552,-0.0635297894,-0.0926341638,-0.3520623446,0.1811241657,-0.4167869091,-0.2689490914,-0.2787384093,-0.1304801255,-0.254093647,-0.0696130618,0.5097812414,0.178332001,0.1131891683,-0.0603247546,0.0520272404,0.0002671419,0.48874861,0.1057779491,0.3719137609,-0.0787951201,-0.0442272685,-0.2410942614,0.2722684443,0.0316483267,-0.1474875063,-0.0711346269,-0.1544221342,0.2219590694,0.1941971928,0.1148816645,-0.5466011763,0.1081469804,0.0861002728,-0.1881438196,-0.3130680919,-0.1160730124,0.2490881979,0.2208352536,-0.1634863764,0.2242891937,-0.0380784273,-0.0277576912,-0.4091936946,0.2718212306,0.1592110097,-0.1402521133,0.0108292913,-0.0242776182,0.3912292719,-0.0398149565,0.0080952598,0.2960894108,0.0299864132,-0.2361740321,0.1950451434,-0.1834563613,-0.1960112453,-0.0912033617,-0.5192207694,-0.1560113281,-0.1106303707,0.1810898483,-0.4407528043,-0.3398949206,0.1470361501,-0.0447697304,-0.1053203866,0.2532119155,0.4619872868,-0.0274411682,0.0461077131,-0.3709779382,0.2000123858,0.5724573731,-0.1720262021,-0.3385213912,-0.0874139369,-0.106363222,-0.0505819507,-0.0523046143,-0.5948602557,0.0122724213,0.1850887239,-0.134402737,-0.139373064,0.2362634689,0.2169276327,0.0268071704,-0.1975290328,0.6016830802,0.0007348063,-0.2036162317,-0.1385705173,-0.3663615584]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/508","title":"TypeError: Receiver() takes no arguments","comments":"Which version of Apache Beam do you have (can you copy your full environment info here)?","body":"I am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.","comment_length":16,"text":"TypeError: Receiver() takes no arguments\nI am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.\nWhich version of Apache Beam do you have (can you copy your full environment info here)?","embeddings":[-0.0946546644,0.3181282282,0.0874954313,0.362924248,0.2740803361,-0.0989540517,0.2871833742,0.2369655818,-0.0286748931,0.0688424557,0.2388483435,0.3297018707,-0.0908722356,-0.3363240659,0.1751256734,-0.5368747115,-0.2141874284,0.1947945654,-0.232953012,-0.1850881428,-0.1442272663,0.3994988501,-0.3325727284,-0.0724646524,-0.1412276328,0.1027451456,-0.0761774182,0.1753544658,-0.118464604,-0.3360362947,0.1966992915,-0.0673624724,0.0733791813,0.2750073373,-0.0001245085,0.3304243982,0.6380137801,-0.1410859525,-0.5410779715,-0.0631846488,-0.0811768249,-0.2349257469,0.1993419528,-0.2498530447,-0.2083238214,-0.2352619022,0.1023960337,-0.0581367277,0.3563393056,0.281555891,0.0985948294,0.1995158792,0.1826244593,-0.1368430406,0.5043628216,0.043215435,0.0561958142,0.0361742526,-0.2274192423,0.2372251302,-0.2068565637,0.0459311418,-0.2663237154,-0.0578701384,0.3306712508,-0.2204922587,0.0874133259,0.1752309054,-0.0942779183,-0.0104458947,0.8057671189,0.0052634547,0.0640940294,-0.201214239,0.0806409568,-0.1908854842,0.2383200824,0.0715610385,-0.4543537199,-0.1370184273,0.1143337935,-0.1281690449,-0.341394484,0.3352718651,-0.0335033908,0.4075168669,-0.0230080616,0.328975141,0.1422492564,0.1264762729,0.2072722167,-0.2324920595,-0.1268543601,0.1452392787,-0.3808562756,0.174837634,0.1763764918,0.2966895998,-0.0637696385,0.1829950064,0.0932282135,-0.1865078211,0.0735035315,0.1276494712,0.3398443162,0.1209282503,0.1319723129,0.0127151208,0.0047444995,-0.0747593865,0.0249654111,0.0320717469,-0.0236140732,-0.36054492,-0.2406122535,0.2245736271,0.2357540876,-0.0485627688,0.0452682115,-0.2116847932,-0.3131856322,-0.2082335055,0.1694273204,0.4560175538,0.0537248924,0.4907977581,0.3138802648,0.1613340378,-0.1267466247,-0.5722319484,-0.1803703606,0.3171550035,-0.161258176,-0.1637429446,0.3268589377,0.0613696501,0.2865219116,-0.2762343287,0.1872421801,0.1734672487,-0.0168005116,-0.1587234586,-0.2288744003,0.1743646711,0.1935789287,0.2789788842,0.2141343802,-0.2745317817,0.0189148672,0.2086554021,-0.2223849446,-0.1689475477,0.2648113072,0.042568069,0.107154578,-0.1384298354,-0.507543385,0.1115983129,-0.1357539892,-0.2659593821,-0.2610871792,-0.2019345164,-0.0996060669,-0.2297554612,0.3767790496,0.5294706821,-0.0989685282,-0.0291802213,-0.0063148174,0.1354220808,0.0829522833,0.0357792452,-0.311412245,0.492174834,-0.1760831773,-0.1047421619,0.8686538339,-0.5204976797,-0.0506761558,0.3639996946,0.1172987446,-0.071886532,0.3070545495,0.0618702248,0.1170750111,-0.0958950073,0.0317528211,0.4021973014,-0.0465446152,0.0234558284,-0.1533483863,0.0390015058,0.2969612777,0.3607648909,0.0934307128,0.2212417275,-0.2734820545,0.4947689474,0.2071569562,0.0037385139,0.0605673529,0.5123981833,-0.3901424408,-0.0827326104,0.0390271321,0.113077119,0.0266657285,-0.0488630608,-0.1335092783,0.1562322825,0.1626265943,0.1614460051,-0.4275047183,0.0230152868,-0.0830781758,0.0427630432,0.0567392223,0.2840376198,-0.1937215477,0.2580310404,0.0285491738,0.3317639828,0.0175104998,-0.0234073736,-0.3775363863,0.2045926601,-0.1200193763,-0.2554772794,0.282967478,-0.0228369217,0.4561701119,0.0420877747,-0.1734325886,0.1393243074,0.1799443066,-0.0209084228,0.0413391441,-0.2081551701,0.0283872373,-0.134917751,-0.0474718511,0.5747932792,0.2473167032,-0.193128854,0.0073986109,-0.0563839599,-0.0950333998,0.2234849483,-0.0520732105,0.1890359968,0.0732780173,0.2233697176,-0.0849315748,0.2273704857,0.204699859,0.1990070641,0.2141200006,-0.0926080421,-0.0686815456,-0.2286660373,0.2874112725,0.2148696929,0.0324278735,-0.0144205429,-0.2963917255,0.002503603,0.2251099199,-0.0638526604,0.2678775489,0.1726977378,-0.12468867,-0.0075488235,0.0018160535,-0.1072777137,0.4460032582,0.0564808063,0.261120379,-0.016631458,-0.1084316075,-0.0197839774,-0.2426254004,-0.2545631528,0.0278431252,0.2242687643,-0.4247313738,0.0356792621,-0.2087251991,-0.6600081921,-0.2359620631,-0.0494453721,-0.1605994701,-0.5055602193,-0.1612770408,0.3273543119,0.0146426978,0.1275832504,-0.120585151,0.2305703163,0.3319926858,-0.2339310497,-0.4185202122,-0.1734266132,-0.2996529937,-0.0336430594,0.1122232676,0.132373184,0.1385983676,0.113913238,-0.0306637753,-0.0660983101,-0.0057053682,0.2152480781,-0.1090568751,0.123509571,0.2650186121,0.3777044117,-0.2389560044,-0.0974703133,0.3813408017,0.0026626792,0.1714810729,0.3198922575,0.0355916694,-0.075479649,-0.1061445251,-0.2500555813,0.0514185093,-0.6115214825,-0.1775638014,0.237305969,0.000052158,0.1542863101,0.2415067703,0.1805426329,0.1287678778,0.3176608682,0.0169246066,-0.2387425154,0.3329898715,-0.1693131328,-0.3372408748,-0.1376382411,0.1169654056,0.2338880152,0.219890222,-0.4601656795,-0.3833651543,0.29610461,-0.034795884,0.0065297475,0.1694480926,0.3062468469,0.017128082,0.0924849436,-0.2547699511,-0.198965624,0.0468889549,-0.2222736031,0.2826959491,0.2882086039,0.3215004206,-0.0384758487,0.9122626185,0.237421751,0.143084988,0.5309516788,0.1021699384,0.2991394699,-0.0367710814,-0.2422951907,0.230656445,0.1812363267,0.1366699338,0.2091348171,-0.249240011,-0.255463928,-0.4086242914,-0.2131759226,-0.2362272888,-0.3552210331,0.1634069979,-0.3094646931,0.2537305355,-0.1385103017,0.2027332932,-0.3603169322,-0.0995280966,0.1494172961,0.1440065354,0.1602015197,0.1130502746,-0.075225547,-0.0607112311,-0.3460043967,0.3344669938,-0.1206776649,0.3733889461,-0.0566632375,0.0655358359,0.0389598571,-0.0580293387,0.7135971189,-0.3010065556,-0.0036403185,0.2482844591,0.2207377255,-0.6750039458,0.1139069796,-0.112287201,0.2548101842,0.0049686255,-0.3782780468,-0.4797842801,-0.1921132207,0.1742551476,-0.0723284334,-0.1652673632,-0.0882530063,-0.2201903909,-0.3247160912,-0.30970186,-0.105983533,-0.0444047228,0.4491245151,0.2301389277,0.2729759812,-0.1422615498,0.0037111251,-0.0231207926,-0.1272257864,0.144898206,-0.020322958,-0.2366642058,-0.1947486848,0.516156733,-0.5861030817,0.3650775254,0.2983977795,-0.6307681799,-0.0487274341,-0.0185211748,0.2841162086,0.2116642445,-0.3109889627,0.0528074764,-0.0490276553,-0.1705605388,-0.0002037241,0.2304157615,0.4107171595,-0.3813044727,-0.0521856807,-0.5629246831,0.4374673963,-0.1350968629,-0.0104099624,0.2742978334,-0.12945728,-0.1884309798,0.6053571701,-0.0318610035,0.974200964,0.2021540403,0.1885179579,0.4231554568,-0.1277760267,0.6227615476,-0.1243739352,0.3219046295,-0.3893945515,0.0414476544,-0.0638145506,-0.1746109426,-0.0930563435,0.1332767904,-0.2606726587,0.2774880826,0.0704492256,-0.0549795069,0.0889615119,0.1255862564,0.1116446927,-0.1067715287,-0.5512533784,0.057797011,-0.4502734542,0.0162889734,-0.0729946271,-0.1618123502,0.0773974955,-0.5212675929,-0.1029234454,0.1655472964,-0.240270555,0.1557008624,-0.0565365739,-0.2547738552,0.1854233742,0.3622989655,-0.0209172349,0.056942869,-0.1994664818,0.3114800751,-0.5459028482,-0.0695615858,-0.1593795568,0.3023845851,0.28486678,-0.0251290761,-0.146742776,0.3340931833,-0.1659708768,-0.6099496484,-0.0495170727,-0.0853261799,0.4431837499,-0.0368481465,-0.2056553364,0.0775685012,-0.1524201035,0.0355418697,0.0395073146,0.0244935751,-0.2060181797,0.0688497052,-0.0528987795,0.0025283047,0.1737148464,0.4771814048,0.0580924861,0.0253521185,0.5821675658,0.2740312815,-0.0257173087,-0.1586009413,-0.2322812825,-0.340931952,-0.6085496545,0.1990737617,0.3044446111,0.0621341728,0.0874242857,0.2038391829,-0.0779726431,0.170503661,0.0560895205,-0.3049573004,0.1134140939,0.2060197741,-0.2436185032,-0.0629765019,0.4305720031,-0.0323971473,0.3100137413,-0.1491773576,-0.1858063489,-0.1241254285,-0.1878904104,0.1358189285,0.160010308,-0.2263966352,-0.1157580242,0.2592052519,0.0706709921,-0.2440284193,-0.2439692467,-0.060357634,-0.1667722315,0.2002599239,0.2243479937,0.1214103028,-0.12901932,-0.4042980671,-0.123750478,-0.1816293448,0.0225903336,-0.1426826268,-0.2429104149,0.0557440706,0.1504145712,0.3674416244,-0.5081473589,0.1975682825,0.0760390013,-0.0926693827,0.2132937163,0.1916123182,0.262457788,-0.0496929251,-0.4662693143,0.1615186185,0.2004982233,-0.0344547965,0.5295451283,-0.1840267777,-0.1134769991,-0.1094243154,-0.0806771889,0.0807791799,-0.1513764858,0.0117968936,0.0692100525,0.0332740247,-0.1537947059,0.0494468063,0.3225930333,0.2039945722,-0.0651857853,0.0736154616,-0.0870866776,-0.366114676,0.2096334994,0.1163535565,0.3633618355,-0.1511389464,-0.1119756177,0.1339500695,-0.0349585712,-0.0854377896,-0.0527979434,0.2270284742,0.053739287,0.2903501391,-0.1444785446,0.0543756522,-0.0564408787,-0.3188199699,-0.0471042953,-0.1534538865,0.2008984089,0.4862774611,0.1516952962,0.1443457007,-0.1632328033,0.1997066289,0.0482344143,0.1141901091,-0.2721175551,0.1604944468,0.0182803292,0.0310199037,-0.6282174587,-0.0157810338,-0.4327570498,-0.1448920965,-0.004945952,-0.1630238593,-0.0300583038,0.211942181,-0.1169978753,-0.3793839514,0.2154737115,0.121847786,-0.0333387293,-0.1512890309,0.2037801892,0.0854873583,0.1178439409,0.1998904943,0.2706089318,0.3540655077,0.4170272052,-0.2698945403,-0.1693301797,0.2886002958,0.0005997109,-0.2072156966,-0.0614681467,0.0971704423,0.3210936785,0.3959818184,0.0738284066,0.0143997325,-0.0085142581,0.1275813282,-0.0613215007,-0.23826316,0.472381264,-0.0463526249,-0.0918838382,-0.3303665221,0.0329027697,-0.5663148761,0.3297737837,0.0595323928,0.3573003113,-0.0106479805,-0.2045569271,-0.0199389141,-0.0656273142,0.3586010933,0.3856936395,0.3740837276,-0.3091628253,-0.4248059392,-0.4309866726,0.158187449,0.1639747918,-0.2411166281,0.1070591733,0.1998973787,0.2831844389,0.2404049635,-0.1074828804,-0.2510265112,-0.0557258539,0.1786379665,-0.218937248,-0.2895013988,0.2064394653,0.0696631521,-0.1936567277,-0.2178909183,-0.1419939697,0.0083855763,0.0918040499,-0.1656745374,-0.3625209033,0.1477167904,0.0822409987,-0.0218153838,-0.0256654713,0.4474615753,0.0580903478,-0.0019750001,-0.1832550913,-0.0743170604,-0.1690218449,-0.2816382647,-0.1547074914,0.5438709259,-0.2213325948,0.2525574267,-0.3257875443,0.2445088774,-0.1334824413,0.3438494503,-0.1466025561,0.0648301616,-0.4684083462,0.3983730972,-0.1344812959,-0.0675862655,-0.0255529806,0.14621903,-0.4820397496,-0.2730072737,0.566480279,-0.5063444972,-0.2313738167,0.0063175997,0.2157868743,-0.4116660357,0.3354253471,-0.3584127426,-0.0386067405,0.2855788767,-0.067324236,-0.144484818,0.165515691,-0.0476102047,0.1122501269,-0.0831059441,-0.5315554142,0.0437982604,-0.1335093379,-0.1396399587,-0.2604459524]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/508","title":"TypeError: Receiver() takes no arguments","comments":"apache-beam==2.23.0\r\nnlp==0.4.0\r\n\r\nFor me this was resolved by running the same python script on Linux (or really WSL). ","body":"I am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.","comment_length":18,"text":"TypeError: Receiver() takes no arguments\nI am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.\napache-beam==2.23.0\r\nnlp==0.4.0\r\n\r\nFor me this was resolved by running the same python script on Linux (or really WSL). ","embeddings":[-0.0946546644,0.3181282282,0.0874954313,0.362924248,0.2740803361,-0.0989540517,0.2871833742,0.2369655818,-0.0286748931,0.0688424557,0.2388483435,0.3297018707,-0.0908722356,-0.3363240659,0.1751256734,-0.5368747115,-0.2141874284,0.1947945654,-0.232953012,-0.1850881428,-0.1442272663,0.3994988501,-0.3325727284,-0.0724646524,-0.1412276328,0.1027451456,-0.0761774182,0.1753544658,-0.118464604,-0.3360362947,0.1966992915,-0.0673624724,0.0733791813,0.2750073373,-0.0001245085,0.3304243982,0.6380137801,-0.1410859525,-0.5410779715,-0.0631846488,-0.0811768249,-0.2349257469,0.1993419528,-0.2498530447,-0.2083238214,-0.2352619022,0.1023960337,-0.0581367277,0.3563393056,0.281555891,0.0985948294,0.1995158792,0.1826244593,-0.1368430406,0.5043628216,0.043215435,0.0561958142,0.0361742526,-0.2274192423,0.2372251302,-0.2068565637,0.0459311418,-0.2663237154,-0.0578701384,0.3306712508,-0.2204922587,0.0874133259,0.1752309054,-0.0942779183,-0.0104458947,0.8057671189,0.0052634547,0.0640940294,-0.201214239,0.0806409568,-0.1908854842,0.2383200824,0.0715610385,-0.4543537199,-0.1370184273,0.1143337935,-0.1281690449,-0.341394484,0.3352718651,-0.0335033908,0.4075168669,-0.0230080616,0.328975141,0.1422492564,0.1264762729,0.2072722167,-0.2324920595,-0.1268543601,0.1452392787,-0.3808562756,0.174837634,0.1763764918,0.2966895998,-0.0637696385,0.1829950064,0.0932282135,-0.1865078211,0.0735035315,0.1276494712,0.3398443162,0.1209282503,0.1319723129,0.0127151208,0.0047444995,-0.0747593865,0.0249654111,0.0320717469,-0.0236140732,-0.36054492,-0.2406122535,0.2245736271,0.2357540876,-0.0485627688,0.0452682115,-0.2116847932,-0.3131856322,-0.2082335055,0.1694273204,0.4560175538,0.0537248924,0.4907977581,0.3138802648,0.1613340378,-0.1267466247,-0.5722319484,-0.1803703606,0.3171550035,-0.161258176,-0.1637429446,0.3268589377,0.0613696501,0.2865219116,-0.2762343287,0.1872421801,0.1734672487,-0.0168005116,-0.1587234586,-0.2288744003,0.1743646711,0.1935789287,0.2789788842,0.2141343802,-0.2745317817,0.0189148672,0.2086554021,-0.2223849446,-0.1689475477,0.2648113072,0.042568069,0.107154578,-0.1384298354,-0.507543385,0.1115983129,-0.1357539892,-0.2659593821,-0.2610871792,-0.2019345164,-0.0996060669,-0.2297554612,0.3767790496,0.5294706821,-0.0989685282,-0.0291802213,-0.0063148174,0.1354220808,0.0829522833,0.0357792452,-0.311412245,0.492174834,-0.1760831773,-0.1047421619,0.8686538339,-0.5204976797,-0.0506761558,0.3639996946,0.1172987446,-0.071886532,0.3070545495,0.0618702248,0.1170750111,-0.0958950073,0.0317528211,0.4021973014,-0.0465446152,0.0234558284,-0.1533483863,0.0390015058,0.2969612777,0.3607648909,0.0934307128,0.2212417275,-0.2734820545,0.4947689474,0.2071569562,0.0037385139,0.0605673529,0.5123981833,-0.3901424408,-0.0827326104,0.0390271321,0.113077119,0.0266657285,-0.0488630608,-0.1335092783,0.1562322825,0.1626265943,0.1614460051,-0.4275047183,0.0230152868,-0.0830781758,0.0427630432,0.0567392223,0.2840376198,-0.1937215477,0.2580310404,0.0285491738,0.3317639828,0.0175104998,-0.0234073736,-0.3775363863,0.2045926601,-0.1200193763,-0.2554772794,0.282967478,-0.0228369217,0.4561701119,0.0420877747,-0.1734325886,0.1393243074,0.1799443066,-0.0209084228,0.0413391441,-0.2081551701,0.0283872373,-0.134917751,-0.0474718511,0.5747932792,0.2473167032,-0.193128854,0.0073986109,-0.0563839599,-0.0950333998,0.2234849483,-0.0520732105,0.1890359968,0.0732780173,0.2233697176,-0.0849315748,0.2273704857,0.204699859,0.1990070641,0.2141200006,-0.0926080421,-0.0686815456,-0.2286660373,0.2874112725,0.2148696929,0.0324278735,-0.0144205429,-0.2963917255,0.002503603,0.2251099199,-0.0638526604,0.2678775489,0.1726977378,-0.12468867,-0.0075488235,0.0018160535,-0.1072777137,0.4460032582,0.0564808063,0.261120379,-0.016631458,-0.1084316075,-0.0197839774,-0.2426254004,-0.2545631528,0.0278431252,0.2242687643,-0.4247313738,0.0356792621,-0.2087251991,-0.6600081921,-0.2359620631,-0.0494453721,-0.1605994701,-0.5055602193,-0.1612770408,0.3273543119,0.0146426978,0.1275832504,-0.120585151,0.2305703163,0.3319926858,-0.2339310497,-0.4185202122,-0.1734266132,-0.2996529937,-0.0336430594,0.1122232676,0.132373184,0.1385983676,0.113913238,-0.0306637753,-0.0660983101,-0.0057053682,0.2152480781,-0.1090568751,0.123509571,0.2650186121,0.3777044117,-0.2389560044,-0.0974703133,0.3813408017,0.0026626792,0.1714810729,0.3198922575,0.0355916694,-0.075479649,-0.1061445251,-0.2500555813,0.0514185093,-0.6115214825,-0.1775638014,0.237305969,0.000052158,0.1542863101,0.2415067703,0.1805426329,0.1287678778,0.3176608682,0.0169246066,-0.2387425154,0.3329898715,-0.1693131328,-0.3372408748,-0.1376382411,0.1169654056,0.2338880152,0.219890222,-0.4601656795,-0.3833651543,0.29610461,-0.034795884,0.0065297475,0.1694480926,0.3062468469,0.017128082,0.0924849436,-0.2547699511,-0.198965624,0.0468889549,-0.2222736031,0.2826959491,0.2882086039,0.3215004206,-0.0384758487,0.9122626185,0.237421751,0.143084988,0.5309516788,0.1021699384,0.2991394699,-0.0367710814,-0.2422951907,0.230656445,0.1812363267,0.1366699338,0.2091348171,-0.249240011,-0.255463928,-0.4086242914,-0.2131759226,-0.2362272888,-0.3552210331,0.1634069979,-0.3094646931,0.2537305355,-0.1385103017,0.2027332932,-0.3603169322,-0.0995280966,0.1494172961,0.1440065354,0.1602015197,0.1130502746,-0.075225547,-0.0607112311,-0.3460043967,0.3344669938,-0.1206776649,0.3733889461,-0.0566632375,0.0655358359,0.0389598571,-0.0580293387,0.7135971189,-0.3010065556,-0.0036403185,0.2482844591,0.2207377255,-0.6750039458,0.1139069796,-0.112287201,0.2548101842,0.0049686255,-0.3782780468,-0.4797842801,-0.1921132207,0.1742551476,-0.0723284334,-0.1652673632,-0.0882530063,-0.2201903909,-0.3247160912,-0.30970186,-0.105983533,-0.0444047228,0.4491245151,0.2301389277,0.2729759812,-0.1422615498,0.0037111251,-0.0231207926,-0.1272257864,0.144898206,-0.020322958,-0.2366642058,-0.1947486848,0.516156733,-0.5861030817,0.3650775254,0.2983977795,-0.6307681799,-0.0487274341,-0.0185211748,0.2841162086,0.2116642445,-0.3109889627,0.0528074764,-0.0490276553,-0.1705605388,-0.0002037241,0.2304157615,0.4107171595,-0.3813044727,-0.0521856807,-0.5629246831,0.4374673963,-0.1350968629,-0.0104099624,0.2742978334,-0.12945728,-0.1884309798,0.6053571701,-0.0318610035,0.974200964,0.2021540403,0.1885179579,0.4231554568,-0.1277760267,0.6227615476,-0.1243739352,0.3219046295,-0.3893945515,0.0414476544,-0.0638145506,-0.1746109426,-0.0930563435,0.1332767904,-0.2606726587,0.2774880826,0.0704492256,-0.0549795069,0.0889615119,0.1255862564,0.1116446927,-0.1067715287,-0.5512533784,0.057797011,-0.4502734542,0.0162889734,-0.0729946271,-0.1618123502,0.0773974955,-0.5212675929,-0.1029234454,0.1655472964,-0.240270555,0.1557008624,-0.0565365739,-0.2547738552,0.1854233742,0.3622989655,-0.0209172349,0.056942869,-0.1994664818,0.3114800751,-0.5459028482,-0.0695615858,-0.1593795568,0.3023845851,0.28486678,-0.0251290761,-0.146742776,0.3340931833,-0.1659708768,-0.6099496484,-0.0495170727,-0.0853261799,0.4431837499,-0.0368481465,-0.2056553364,0.0775685012,-0.1524201035,0.0355418697,0.0395073146,0.0244935751,-0.2060181797,0.0688497052,-0.0528987795,0.0025283047,0.1737148464,0.4771814048,0.0580924861,0.0253521185,0.5821675658,0.2740312815,-0.0257173087,-0.1586009413,-0.2322812825,-0.340931952,-0.6085496545,0.1990737617,0.3044446111,0.0621341728,0.0874242857,0.2038391829,-0.0779726431,0.170503661,0.0560895205,-0.3049573004,0.1134140939,0.2060197741,-0.2436185032,-0.0629765019,0.4305720031,-0.0323971473,0.3100137413,-0.1491773576,-0.1858063489,-0.1241254285,-0.1878904104,0.1358189285,0.160010308,-0.2263966352,-0.1157580242,0.2592052519,0.0706709921,-0.2440284193,-0.2439692467,-0.060357634,-0.1667722315,0.2002599239,0.2243479937,0.1214103028,-0.12901932,-0.4042980671,-0.123750478,-0.1816293448,0.0225903336,-0.1426826268,-0.2429104149,0.0557440706,0.1504145712,0.3674416244,-0.5081473589,0.1975682825,0.0760390013,-0.0926693827,0.2132937163,0.1916123182,0.262457788,-0.0496929251,-0.4662693143,0.1615186185,0.2004982233,-0.0344547965,0.5295451283,-0.1840267777,-0.1134769991,-0.1094243154,-0.0806771889,0.0807791799,-0.1513764858,0.0117968936,0.0692100525,0.0332740247,-0.1537947059,0.0494468063,0.3225930333,0.2039945722,-0.0651857853,0.0736154616,-0.0870866776,-0.366114676,0.2096334994,0.1163535565,0.3633618355,-0.1511389464,-0.1119756177,0.1339500695,-0.0349585712,-0.0854377896,-0.0527979434,0.2270284742,0.053739287,0.2903501391,-0.1444785446,0.0543756522,-0.0564408787,-0.3188199699,-0.0471042953,-0.1534538865,0.2008984089,0.4862774611,0.1516952962,0.1443457007,-0.1632328033,0.1997066289,0.0482344143,0.1141901091,-0.2721175551,0.1604944468,0.0182803292,0.0310199037,-0.6282174587,-0.0157810338,-0.4327570498,-0.1448920965,-0.004945952,-0.1630238593,-0.0300583038,0.211942181,-0.1169978753,-0.3793839514,0.2154737115,0.121847786,-0.0333387293,-0.1512890309,0.2037801892,0.0854873583,0.1178439409,0.1998904943,0.2706089318,0.3540655077,0.4170272052,-0.2698945403,-0.1693301797,0.2886002958,0.0005997109,-0.2072156966,-0.0614681467,0.0971704423,0.3210936785,0.3959818184,0.0738284066,0.0143997325,-0.0085142581,0.1275813282,-0.0613215007,-0.23826316,0.472381264,-0.0463526249,-0.0918838382,-0.3303665221,0.0329027697,-0.5663148761,0.3297737837,0.0595323928,0.3573003113,-0.0106479805,-0.2045569271,-0.0199389141,-0.0656273142,0.3586010933,0.3856936395,0.3740837276,-0.3091628253,-0.4248059392,-0.4309866726,0.158187449,0.1639747918,-0.2411166281,0.1070591733,0.1998973787,0.2831844389,0.2404049635,-0.1074828804,-0.2510265112,-0.0557258539,0.1786379665,-0.218937248,-0.2895013988,0.2064394653,0.0696631521,-0.1936567277,-0.2178909183,-0.1419939697,0.0083855763,0.0918040499,-0.1656745374,-0.3625209033,0.1477167904,0.0822409987,-0.0218153838,-0.0256654713,0.4474615753,0.0580903478,-0.0019750001,-0.1832550913,-0.0743170604,-0.1690218449,-0.2816382647,-0.1547074914,0.5438709259,-0.2213325948,0.2525574267,-0.3257875443,0.2445088774,-0.1334824413,0.3438494503,-0.1466025561,0.0648301616,-0.4684083462,0.3983730972,-0.1344812959,-0.0675862655,-0.0255529806,0.14621903,-0.4820397496,-0.2730072737,0.566480279,-0.5063444972,-0.2313738167,0.0063175997,0.2157868743,-0.4116660357,0.3354253471,-0.3584127426,-0.0386067405,0.2855788767,-0.067324236,-0.144484818,0.165515691,-0.0476102047,0.1122501269,-0.0831059441,-0.5315554142,0.0437982604,-0.1335093379,-0.1396399587,-0.2604459524]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/508","title":"TypeError: Receiver() takes no arguments","comments":"Do you manage to run a dummy beam pipeline with python on windows ? \r\nYou can test a dummy pipeline with [this code](https:\/\/github.com\/apache\/beam\/blob\/master\/sdks\/python\/apache_beam\/examples\/wordcount_minimal.py)\r\n\r\nIf you get the same error, it means that the issue comes from apache beam.\r\nOtherwise we'll investigate what went wrong here","body":"I am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.","comment_length":45,"text":"TypeError: Receiver() takes no arguments\nI am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.\nDo you manage to run a dummy beam pipeline with python on windows ? \r\nYou can test a dummy pipeline with [this code](https:\/\/github.com\/apache\/beam\/blob\/master\/sdks\/python\/apache_beam\/examples\/wordcount_minimal.py)\r\n\r\nIf you get the same error, it means that the issue comes from apache beam.\r\nOtherwise we'll investigate what went wrong here","embeddings":[-0.0946546644,0.3181282282,0.0874954313,0.362924248,0.2740803361,-0.0989540517,0.2871833742,0.2369655818,-0.0286748931,0.0688424557,0.2388483435,0.3297018707,-0.0908722356,-0.3363240659,0.1751256734,-0.5368747115,-0.2141874284,0.1947945654,-0.232953012,-0.1850881428,-0.1442272663,0.3994988501,-0.3325727284,-0.0724646524,-0.1412276328,0.1027451456,-0.0761774182,0.1753544658,-0.118464604,-0.3360362947,0.1966992915,-0.0673624724,0.0733791813,0.2750073373,-0.0001245085,0.3304243982,0.6380137801,-0.1410859525,-0.5410779715,-0.0631846488,-0.0811768249,-0.2349257469,0.1993419528,-0.2498530447,-0.2083238214,-0.2352619022,0.1023960337,-0.0581367277,0.3563393056,0.281555891,0.0985948294,0.1995158792,0.1826244593,-0.1368430406,0.5043628216,0.043215435,0.0561958142,0.0361742526,-0.2274192423,0.2372251302,-0.2068565637,0.0459311418,-0.2663237154,-0.0578701384,0.3306712508,-0.2204922587,0.0874133259,0.1752309054,-0.0942779183,-0.0104458947,0.8057671189,0.0052634547,0.0640940294,-0.201214239,0.0806409568,-0.1908854842,0.2383200824,0.0715610385,-0.4543537199,-0.1370184273,0.1143337935,-0.1281690449,-0.341394484,0.3352718651,-0.0335033908,0.4075168669,-0.0230080616,0.328975141,0.1422492564,0.1264762729,0.2072722167,-0.2324920595,-0.1268543601,0.1452392787,-0.3808562756,0.174837634,0.1763764918,0.2966895998,-0.0637696385,0.1829950064,0.0932282135,-0.1865078211,0.0735035315,0.1276494712,0.3398443162,0.1209282503,0.1319723129,0.0127151208,0.0047444995,-0.0747593865,0.0249654111,0.0320717469,-0.0236140732,-0.36054492,-0.2406122535,0.2245736271,0.2357540876,-0.0485627688,0.0452682115,-0.2116847932,-0.3131856322,-0.2082335055,0.1694273204,0.4560175538,0.0537248924,0.4907977581,0.3138802648,0.1613340378,-0.1267466247,-0.5722319484,-0.1803703606,0.3171550035,-0.161258176,-0.1637429446,0.3268589377,0.0613696501,0.2865219116,-0.2762343287,0.1872421801,0.1734672487,-0.0168005116,-0.1587234586,-0.2288744003,0.1743646711,0.1935789287,0.2789788842,0.2141343802,-0.2745317817,0.0189148672,0.2086554021,-0.2223849446,-0.1689475477,0.2648113072,0.042568069,0.107154578,-0.1384298354,-0.507543385,0.1115983129,-0.1357539892,-0.2659593821,-0.2610871792,-0.2019345164,-0.0996060669,-0.2297554612,0.3767790496,0.5294706821,-0.0989685282,-0.0291802213,-0.0063148174,0.1354220808,0.0829522833,0.0357792452,-0.311412245,0.492174834,-0.1760831773,-0.1047421619,0.8686538339,-0.5204976797,-0.0506761558,0.3639996946,0.1172987446,-0.071886532,0.3070545495,0.0618702248,0.1170750111,-0.0958950073,0.0317528211,0.4021973014,-0.0465446152,0.0234558284,-0.1533483863,0.0390015058,0.2969612777,0.3607648909,0.0934307128,0.2212417275,-0.2734820545,0.4947689474,0.2071569562,0.0037385139,0.0605673529,0.5123981833,-0.3901424408,-0.0827326104,0.0390271321,0.113077119,0.0266657285,-0.0488630608,-0.1335092783,0.1562322825,0.1626265943,0.1614460051,-0.4275047183,0.0230152868,-0.0830781758,0.0427630432,0.0567392223,0.2840376198,-0.1937215477,0.2580310404,0.0285491738,0.3317639828,0.0175104998,-0.0234073736,-0.3775363863,0.2045926601,-0.1200193763,-0.2554772794,0.282967478,-0.0228369217,0.4561701119,0.0420877747,-0.1734325886,0.1393243074,0.1799443066,-0.0209084228,0.0413391441,-0.2081551701,0.0283872373,-0.134917751,-0.0474718511,0.5747932792,0.2473167032,-0.193128854,0.0073986109,-0.0563839599,-0.0950333998,0.2234849483,-0.0520732105,0.1890359968,0.0732780173,0.2233697176,-0.0849315748,0.2273704857,0.204699859,0.1990070641,0.2141200006,-0.0926080421,-0.0686815456,-0.2286660373,0.2874112725,0.2148696929,0.0324278735,-0.0144205429,-0.2963917255,0.002503603,0.2251099199,-0.0638526604,0.2678775489,0.1726977378,-0.12468867,-0.0075488235,0.0018160535,-0.1072777137,0.4460032582,0.0564808063,0.261120379,-0.016631458,-0.1084316075,-0.0197839774,-0.2426254004,-0.2545631528,0.0278431252,0.2242687643,-0.4247313738,0.0356792621,-0.2087251991,-0.6600081921,-0.2359620631,-0.0494453721,-0.1605994701,-0.5055602193,-0.1612770408,0.3273543119,0.0146426978,0.1275832504,-0.120585151,0.2305703163,0.3319926858,-0.2339310497,-0.4185202122,-0.1734266132,-0.2996529937,-0.0336430594,0.1122232676,0.132373184,0.1385983676,0.113913238,-0.0306637753,-0.0660983101,-0.0057053682,0.2152480781,-0.1090568751,0.123509571,0.2650186121,0.3777044117,-0.2389560044,-0.0974703133,0.3813408017,0.0026626792,0.1714810729,0.3198922575,0.0355916694,-0.075479649,-0.1061445251,-0.2500555813,0.0514185093,-0.6115214825,-0.1775638014,0.237305969,0.000052158,0.1542863101,0.2415067703,0.1805426329,0.1287678778,0.3176608682,0.0169246066,-0.2387425154,0.3329898715,-0.1693131328,-0.3372408748,-0.1376382411,0.1169654056,0.2338880152,0.219890222,-0.4601656795,-0.3833651543,0.29610461,-0.034795884,0.0065297475,0.1694480926,0.3062468469,0.017128082,0.0924849436,-0.2547699511,-0.198965624,0.0468889549,-0.2222736031,0.2826959491,0.2882086039,0.3215004206,-0.0384758487,0.9122626185,0.237421751,0.143084988,0.5309516788,0.1021699384,0.2991394699,-0.0367710814,-0.2422951907,0.230656445,0.1812363267,0.1366699338,0.2091348171,-0.249240011,-0.255463928,-0.4086242914,-0.2131759226,-0.2362272888,-0.3552210331,0.1634069979,-0.3094646931,0.2537305355,-0.1385103017,0.2027332932,-0.3603169322,-0.0995280966,0.1494172961,0.1440065354,0.1602015197,0.1130502746,-0.075225547,-0.0607112311,-0.3460043967,0.3344669938,-0.1206776649,0.3733889461,-0.0566632375,0.0655358359,0.0389598571,-0.0580293387,0.7135971189,-0.3010065556,-0.0036403185,0.2482844591,0.2207377255,-0.6750039458,0.1139069796,-0.112287201,0.2548101842,0.0049686255,-0.3782780468,-0.4797842801,-0.1921132207,0.1742551476,-0.0723284334,-0.1652673632,-0.0882530063,-0.2201903909,-0.3247160912,-0.30970186,-0.105983533,-0.0444047228,0.4491245151,0.2301389277,0.2729759812,-0.1422615498,0.0037111251,-0.0231207926,-0.1272257864,0.144898206,-0.020322958,-0.2366642058,-0.1947486848,0.516156733,-0.5861030817,0.3650775254,0.2983977795,-0.6307681799,-0.0487274341,-0.0185211748,0.2841162086,0.2116642445,-0.3109889627,0.0528074764,-0.0490276553,-0.1705605388,-0.0002037241,0.2304157615,0.4107171595,-0.3813044727,-0.0521856807,-0.5629246831,0.4374673963,-0.1350968629,-0.0104099624,0.2742978334,-0.12945728,-0.1884309798,0.6053571701,-0.0318610035,0.974200964,0.2021540403,0.1885179579,0.4231554568,-0.1277760267,0.6227615476,-0.1243739352,0.3219046295,-0.3893945515,0.0414476544,-0.0638145506,-0.1746109426,-0.0930563435,0.1332767904,-0.2606726587,0.2774880826,0.0704492256,-0.0549795069,0.0889615119,0.1255862564,0.1116446927,-0.1067715287,-0.5512533784,0.057797011,-0.4502734542,0.0162889734,-0.0729946271,-0.1618123502,0.0773974955,-0.5212675929,-0.1029234454,0.1655472964,-0.240270555,0.1557008624,-0.0565365739,-0.2547738552,0.1854233742,0.3622989655,-0.0209172349,0.056942869,-0.1994664818,0.3114800751,-0.5459028482,-0.0695615858,-0.1593795568,0.3023845851,0.28486678,-0.0251290761,-0.146742776,0.3340931833,-0.1659708768,-0.6099496484,-0.0495170727,-0.0853261799,0.4431837499,-0.0368481465,-0.2056553364,0.0775685012,-0.1524201035,0.0355418697,0.0395073146,0.0244935751,-0.2060181797,0.0688497052,-0.0528987795,0.0025283047,0.1737148464,0.4771814048,0.0580924861,0.0253521185,0.5821675658,0.2740312815,-0.0257173087,-0.1586009413,-0.2322812825,-0.340931952,-0.6085496545,0.1990737617,0.3044446111,0.0621341728,0.0874242857,0.2038391829,-0.0779726431,0.170503661,0.0560895205,-0.3049573004,0.1134140939,0.2060197741,-0.2436185032,-0.0629765019,0.4305720031,-0.0323971473,0.3100137413,-0.1491773576,-0.1858063489,-0.1241254285,-0.1878904104,0.1358189285,0.160010308,-0.2263966352,-0.1157580242,0.2592052519,0.0706709921,-0.2440284193,-0.2439692467,-0.060357634,-0.1667722315,0.2002599239,0.2243479937,0.1214103028,-0.12901932,-0.4042980671,-0.123750478,-0.1816293448,0.0225903336,-0.1426826268,-0.2429104149,0.0557440706,0.1504145712,0.3674416244,-0.5081473589,0.1975682825,0.0760390013,-0.0926693827,0.2132937163,0.1916123182,0.262457788,-0.0496929251,-0.4662693143,0.1615186185,0.2004982233,-0.0344547965,0.5295451283,-0.1840267777,-0.1134769991,-0.1094243154,-0.0806771889,0.0807791799,-0.1513764858,0.0117968936,0.0692100525,0.0332740247,-0.1537947059,0.0494468063,0.3225930333,0.2039945722,-0.0651857853,0.0736154616,-0.0870866776,-0.366114676,0.2096334994,0.1163535565,0.3633618355,-0.1511389464,-0.1119756177,0.1339500695,-0.0349585712,-0.0854377896,-0.0527979434,0.2270284742,0.053739287,0.2903501391,-0.1444785446,0.0543756522,-0.0564408787,-0.3188199699,-0.0471042953,-0.1534538865,0.2008984089,0.4862774611,0.1516952962,0.1443457007,-0.1632328033,0.1997066289,0.0482344143,0.1141901091,-0.2721175551,0.1604944468,0.0182803292,0.0310199037,-0.6282174587,-0.0157810338,-0.4327570498,-0.1448920965,-0.004945952,-0.1630238593,-0.0300583038,0.211942181,-0.1169978753,-0.3793839514,0.2154737115,0.121847786,-0.0333387293,-0.1512890309,0.2037801892,0.0854873583,0.1178439409,0.1998904943,0.2706089318,0.3540655077,0.4170272052,-0.2698945403,-0.1693301797,0.2886002958,0.0005997109,-0.2072156966,-0.0614681467,0.0971704423,0.3210936785,0.3959818184,0.0738284066,0.0143997325,-0.0085142581,0.1275813282,-0.0613215007,-0.23826316,0.472381264,-0.0463526249,-0.0918838382,-0.3303665221,0.0329027697,-0.5663148761,0.3297737837,0.0595323928,0.3573003113,-0.0106479805,-0.2045569271,-0.0199389141,-0.0656273142,0.3586010933,0.3856936395,0.3740837276,-0.3091628253,-0.4248059392,-0.4309866726,0.158187449,0.1639747918,-0.2411166281,0.1070591733,0.1998973787,0.2831844389,0.2404049635,-0.1074828804,-0.2510265112,-0.0557258539,0.1786379665,-0.218937248,-0.2895013988,0.2064394653,0.0696631521,-0.1936567277,-0.2178909183,-0.1419939697,0.0083855763,0.0918040499,-0.1656745374,-0.3625209033,0.1477167904,0.0822409987,-0.0218153838,-0.0256654713,0.4474615753,0.0580903478,-0.0019750001,-0.1832550913,-0.0743170604,-0.1690218449,-0.2816382647,-0.1547074914,0.5438709259,-0.2213325948,0.2525574267,-0.3257875443,0.2445088774,-0.1334824413,0.3438494503,-0.1466025561,0.0648301616,-0.4684083462,0.3983730972,-0.1344812959,-0.0675862655,-0.0255529806,0.14621903,-0.4820397496,-0.2730072737,0.566480279,-0.5063444972,-0.2313738167,0.0063175997,0.2157868743,-0.4116660357,0.3354253471,-0.3584127426,-0.0386067405,0.2855788767,-0.067324236,-0.144484818,0.165515691,-0.0476102047,0.1122501269,-0.0831059441,-0.5315554142,0.0437982604,-0.1335093379,-0.1396399587,-0.2604459524]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/508","title":"TypeError: Receiver() takes no arguments","comments":"Still, same error, so I guess it is on apache beam then. \r\nThanks for the investigation.","body":"I am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.","comment_length":16,"text":"TypeError: Receiver() takes no arguments\nI am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.\nStill, same error, so I guess it is on apache beam then. \r\nThanks for the investigation.","embeddings":[-0.0946546644,0.3181282282,0.0874954313,0.362924248,0.2740803361,-0.0989540517,0.2871833742,0.2369655818,-0.0286748931,0.0688424557,0.2388483435,0.3297018707,-0.0908722356,-0.3363240659,0.1751256734,-0.5368747115,-0.2141874284,0.1947945654,-0.232953012,-0.1850881428,-0.1442272663,0.3994988501,-0.3325727284,-0.0724646524,-0.1412276328,0.1027451456,-0.0761774182,0.1753544658,-0.118464604,-0.3360362947,0.1966992915,-0.0673624724,0.0733791813,0.2750073373,-0.0001245085,0.3304243982,0.6380137801,-0.1410859525,-0.5410779715,-0.0631846488,-0.0811768249,-0.2349257469,0.1993419528,-0.2498530447,-0.2083238214,-0.2352619022,0.1023960337,-0.0581367277,0.3563393056,0.281555891,0.0985948294,0.1995158792,0.1826244593,-0.1368430406,0.5043628216,0.043215435,0.0561958142,0.0361742526,-0.2274192423,0.2372251302,-0.2068565637,0.0459311418,-0.2663237154,-0.0578701384,0.3306712508,-0.2204922587,0.0874133259,0.1752309054,-0.0942779183,-0.0104458947,0.8057671189,0.0052634547,0.0640940294,-0.201214239,0.0806409568,-0.1908854842,0.2383200824,0.0715610385,-0.4543537199,-0.1370184273,0.1143337935,-0.1281690449,-0.341394484,0.3352718651,-0.0335033908,0.4075168669,-0.0230080616,0.328975141,0.1422492564,0.1264762729,0.2072722167,-0.2324920595,-0.1268543601,0.1452392787,-0.3808562756,0.174837634,0.1763764918,0.2966895998,-0.0637696385,0.1829950064,0.0932282135,-0.1865078211,0.0735035315,0.1276494712,0.3398443162,0.1209282503,0.1319723129,0.0127151208,0.0047444995,-0.0747593865,0.0249654111,0.0320717469,-0.0236140732,-0.36054492,-0.2406122535,0.2245736271,0.2357540876,-0.0485627688,0.0452682115,-0.2116847932,-0.3131856322,-0.2082335055,0.1694273204,0.4560175538,0.0537248924,0.4907977581,0.3138802648,0.1613340378,-0.1267466247,-0.5722319484,-0.1803703606,0.3171550035,-0.161258176,-0.1637429446,0.3268589377,0.0613696501,0.2865219116,-0.2762343287,0.1872421801,0.1734672487,-0.0168005116,-0.1587234586,-0.2288744003,0.1743646711,0.1935789287,0.2789788842,0.2141343802,-0.2745317817,0.0189148672,0.2086554021,-0.2223849446,-0.1689475477,0.2648113072,0.042568069,0.107154578,-0.1384298354,-0.507543385,0.1115983129,-0.1357539892,-0.2659593821,-0.2610871792,-0.2019345164,-0.0996060669,-0.2297554612,0.3767790496,0.5294706821,-0.0989685282,-0.0291802213,-0.0063148174,0.1354220808,0.0829522833,0.0357792452,-0.311412245,0.492174834,-0.1760831773,-0.1047421619,0.8686538339,-0.5204976797,-0.0506761558,0.3639996946,0.1172987446,-0.071886532,0.3070545495,0.0618702248,0.1170750111,-0.0958950073,0.0317528211,0.4021973014,-0.0465446152,0.0234558284,-0.1533483863,0.0390015058,0.2969612777,0.3607648909,0.0934307128,0.2212417275,-0.2734820545,0.4947689474,0.2071569562,0.0037385139,0.0605673529,0.5123981833,-0.3901424408,-0.0827326104,0.0390271321,0.113077119,0.0266657285,-0.0488630608,-0.1335092783,0.1562322825,0.1626265943,0.1614460051,-0.4275047183,0.0230152868,-0.0830781758,0.0427630432,0.0567392223,0.2840376198,-0.1937215477,0.2580310404,0.0285491738,0.3317639828,0.0175104998,-0.0234073736,-0.3775363863,0.2045926601,-0.1200193763,-0.2554772794,0.282967478,-0.0228369217,0.4561701119,0.0420877747,-0.1734325886,0.1393243074,0.1799443066,-0.0209084228,0.0413391441,-0.2081551701,0.0283872373,-0.134917751,-0.0474718511,0.5747932792,0.2473167032,-0.193128854,0.0073986109,-0.0563839599,-0.0950333998,0.2234849483,-0.0520732105,0.1890359968,0.0732780173,0.2233697176,-0.0849315748,0.2273704857,0.204699859,0.1990070641,0.2141200006,-0.0926080421,-0.0686815456,-0.2286660373,0.2874112725,0.2148696929,0.0324278735,-0.0144205429,-0.2963917255,0.002503603,0.2251099199,-0.0638526604,0.2678775489,0.1726977378,-0.12468867,-0.0075488235,0.0018160535,-0.1072777137,0.4460032582,0.0564808063,0.261120379,-0.016631458,-0.1084316075,-0.0197839774,-0.2426254004,-0.2545631528,0.0278431252,0.2242687643,-0.4247313738,0.0356792621,-0.2087251991,-0.6600081921,-0.2359620631,-0.0494453721,-0.1605994701,-0.5055602193,-0.1612770408,0.3273543119,0.0146426978,0.1275832504,-0.120585151,0.2305703163,0.3319926858,-0.2339310497,-0.4185202122,-0.1734266132,-0.2996529937,-0.0336430594,0.1122232676,0.132373184,0.1385983676,0.113913238,-0.0306637753,-0.0660983101,-0.0057053682,0.2152480781,-0.1090568751,0.123509571,0.2650186121,0.3777044117,-0.2389560044,-0.0974703133,0.3813408017,0.0026626792,0.1714810729,0.3198922575,0.0355916694,-0.075479649,-0.1061445251,-0.2500555813,0.0514185093,-0.6115214825,-0.1775638014,0.237305969,0.000052158,0.1542863101,0.2415067703,0.1805426329,0.1287678778,0.3176608682,0.0169246066,-0.2387425154,0.3329898715,-0.1693131328,-0.3372408748,-0.1376382411,0.1169654056,0.2338880152,0.219890222,-0.4601656795,-0.3833651543,0.29610461,-0.034795884,0.0065297475,0.1694480926,0.3062468469,0.017128082,0.0924849436,-0.2547699511,-0.198965624,0.0468889549,-0.2222736031,0.2826959491,0.2882086039,0.3215004206,-0.0384758487,0.9122626185,0.237421751,0.143084988,0.5309516788,0.1021699384,0.2991394699,-0.0367710814,-0.2422951907,0.230656445,0.1812363267,0.1366699338,0.2091348171,-0.249240011,-0.255463928,-0.4086242914,-0.2131759226,-0.2362272888,-0.3552210331,0.1634069979,-0.3094646931,0.2537305355,-0.1385103017,0.2027332932,-0.3603169322,-0.0995280966,0.1494172961,0.1440065354,0.1602015197,0.1130502746,-0.075225547,-0.0607112311,-0.3460043967,0.3344669938,-0.1206776649,0.3733889461,-0.0566632375,0.0655358359,0.0389598571,-0.0580293387,0.7135971189,-0.3010065556,-0.0036403185,0.2482844591,0.2207377255,-0.6750039458,0.1139069796,-0.112287201,0.2548101842,0.0049686255,-0.3782780468,-0.4797842801,-0.1921132207,0.1742551476,-0.0723284334,-0.1652673632,-0.0882530063,-0.2201903909,-0.3247160912,-0.30970186,-0.105983533,-0.0444047228,0.4491245151,0.2301389277,0.2729759812,-0.1422615498,0.0037111251,-0.0231207926,-0.1272257864,0.144898206,-0.020322958,-0.2366642058,-0.1947486848,0.516156733,-0.5861030817,0.3650775254,0.2983977795,-0.6307681799,-0.0487274341,-0.0185211748,0.2841162086,0.2116642445,-0.3109889627,0.0528074764,-0.0490276553,-0.1705605388,-0.0002037241,0.2304157615,0.4107171595,-0.3813044727,-0.0521856807,-0.5629246831,0.4374673963,-0.1350968629,-0.0104099624,0.2742978334,-0.12945728,-0.1884309798,0.6053571701,-0.0318610035,0.974200964,0.2021540403,0.1885179579,0.4231554568,-0.1277760267,0.6227615476,-0.1243739352,0.3219046295,-0.3893945515,0.0414476544,-0.0638145506,-0.1746109426,-0.0930563435,0.1332767904,-0.2606726587,0.2774880826,0.0704492256,-0.0549795069,0.0889615119,0.1255862564,0.1116446927,-0.1067715287,-0.5512533784,0.057797011,-0.4502734542,0.0162889734,-0.0729946271,-0.1618123502,0.0773974955,-0.5212675929,-0.1029234454,0.1655472964,-0.240270555,0.1557008624,-0.0565365739,-0.2547738552,0.1854233742,0.3622989655,-0.0209172349,0.056942869,-0.1994664818,0.3114800751,-0.5459028482,-0.0695615858,-0.1593795568,0.3023845851,0.28486678,-0.0251290761,-0.146742776,0.3340931833,-0.1659708768,-0.6099496484,-0.0495170727,-0.0853261799,0.4431837499,-0.0368481465,-0.2056553364,0.0775685012,-0.1524201035,0.0355418697,0.0395073146,0.0244935751,-0.2060181797,0.0688497052,-0.0528987795,0.0025283047,0.1737148464,0.4771814048,0.0580924861,0.0253521185,0.5821675658,0.2740312815,-0.0257173087,-0.1586009413,-0.2322812825,-0.340931952,-0.6085496545,0.1990737617,0.3044446111,0.0621341728,0.0874242857,0.2038391829,-0.0779726431,0.170503661,0.0560895205,-0.3049573004,0.1134140939,0.2060197741,-0.2436185032,-0.0629765019,0.4305720031,-0.0323971473,0.3100137413,-0.1491773576,-0.1858063489,-0.1241254285,-0.1878904104,0.1358189285,0.160010308,-0.2263966352,-0.1157580242,0.2592052519,0.0706709921,-0.2440284193,-0.2439692467,-0.060357634,-0.1667722315,0.2002599239,0.2243479937,0.1214103028,-0.12901932,-0.4042980671,-0.123750478,-0.1816293448,0.0225903336,-0.1426826268,-0.2429104149,0.0557440706,0.1504145712,0.3674416244,-0.5081473589,0.1975682825,0.0760390013,-0.0926693827,0.2132937163,0.1916123182,0.262457788,-0.0496929251,-0.4662693143,0.1615186185,0.2004982233,-0.0344547965,0.5295451283,-0.1840267777,-0.1134769991,-0.1094243154,-0.0806771889,0.0807791799,-0.1513764858,0.0117968936,0.0692100525,0.0332740247,-0.1537947059,0.0494468063,0.3225930333,0.2039945722,-0.0651857853,0.0736154616,-0.0870866776,-0.366114676,0.2096334994,0.1163535565,0.3633618355,-0.1511389464,-0.1119756177,0.1339500695,-0.0349585712,-0.0854377896,-0.0527979434,0.2270284742,0.053739287,0.2903501391,-0.1444785446,0.0543756522,-0.0564408787,-0.3188199699,-0.0471042953,-0.1534538865,0.2008984089,0.4862774611,0.1516952962,0.1443457007,-0.1632328033,0.1997066289,0.0482344143,0.1141901091,-0.2721175551,0.1604944468,0.0182803292,0.0310199037,-0.6282174587,-0.0157810338,-0.4327570498,-0.1448920965,-0.004945952,-0.1630238593,-0.0300583038,0.211942181,-0.1169978753,-0.3793839514,0.2154737115,0.121847786,-0.0333387293,-0.1512890309,0.2037801892,0.0854873583,0.1178439409,0.1998904943,0.2706089318,0.3540655077,0.4170272052,-0.2698945403,-0.1693301797,0.2886002958,0.0005997109,-0.2072156966,-0.0614681467,0.0971704423,0.3210936785,0.3959818184,0.0738284066,0.0143997325,-0.0085142581,0.1275813282,-0.0613215007,-0.23826316,0.472381264,-0.0463526249,-0.0918838382,-0.3303665221,0.0329027697,-0.5663148761,0.3297737837,0.0595323928,0.3573003113,-0.0106479805,-0.2045569271,-0.0199389141,-0.0656273142,0.3586010933,0.3856936395,0.3740837276,-0.3091628253,-0.4248059392,-0.4309866726,0.158187449,0.1639747918,-0.2411166281,0.1070591733,0.1998973787,0.2831844389,0.2404049635,-0.1074828804,-0.2510265112,-0.0557258539,0.1786379665,-0.218937248,-0.2895013988,0.2064394653,0.0696631521,-0.1936567277,-0.2178909183,-0.1419939697,0.0083855763,0.0918040499,-0.1656745374,-0.3625209033,0.1477167904,0.0822409987,-0.0218153838,-0.0256654713,0.4474615753,0.0580903478,-0.0019750001,-0.1832550913,-0.0743170604,-0.1690218449,-0.2816382647,-0.1547074914,0.5438709259,-0.2213325948,0.2525574267,-0.3257875443,0.2445088774,-0.1334824413,0.3438494503,-0.1466025561,0.0648301616,-0.4684083462,0.3983730972,-0.1344812959,-0.0675862655,-0.0255529806,0.14621903,-0.4820397496,-0.2730072737,0.566480279,-0.5063444972,-0.2313738167,0.0063175997,0.2157868743,-0.4116660357,0.3354253471,-0.3584127426,-0.0386067405,0.2855788767,-0.067324236,-0.144484818,0.165515691,-0.0476102047,0.1122501269,-0.0831059441,-0.5315554142,0.0437982604,-0.1335093379,-0.1396399587,-0.2604459524]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/508","title":"TypeError: Receiver() takes no arguments","comments":"Thanks for trying\r\nLet us know if you find clues of what caused this issue, or if you find a fix","body":"I am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.","comment_length":21,"text":"TypeError: Receiver() takes no arguments\nI am trying to load a wikipedia data set\r\n\r\n```\r\nimport nlp\r\nfrom nlp import load_dataset\r\n\r\ndataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=data_path, beam_runner='DirectRunner')\r\n#dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner')\r\n```\r\n\r\nThis fails in the apache beam runner. \r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"D:\/ML\/wikiembedding\/gpt2_sv.py\", line 36, in <module>\r\n dataset = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=my_cache_dir, beam_runner='DirectRunner')\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\load.py\", line 548, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 462, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\nlp\\builder.py\", line 969, in _download_and_prepare\r\n pipeline_results = pipeline.run()\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\pipeline.py\", line 534, in run\r\n return self.runner.run_pipeline(self, self._options)\r\n....\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\bundle_processor.py\", line 218, in process_encoded\r\n self.output(decoded_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\apache_beam\\runners\\worker\\operations.py\", line 332, in output\r\n cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)\r\n File \"C:\\Users\\seto\\AppData\\Local\\Programs\\Python\\Python38\\lib\\site-packages\\Cython\\Shadow.py\", line 167, in cast\r\n return type(*args)\r\nTypeError: Receiver() takes no arguments\r\n\r\n```\r\n\r\nThis is run on a Windows 10 machine with python 3.8. I get the same error loading the swedish wikipedia dump.\nThanks for trying\r\nLet us know if you find clues of what caused this issue, or if you find a fix","embeddings":[-0.0946546644,0.3181282282,0.0874954313,0.362924248,0.2740803361,-0.0989540517,0.2871833742,0.2369655818,-0.0286748931,0.0688424557,0.2388483435,0.3297018707,-0.0908722356,-0.3363240659,0.1751256734,-0.5368747115,-0.2141874284,0.1947945654,-0.232953012,-0.1850881428,-0.1442272663,0.3994988501,-0.3325727284,-0.0724646524,-0.1412276328,0.1027451456,-0.0761774182,0.1753544658,-0.118464604,-0.3360362947,0.1966992915,-0.0673624724,0.0733791813,0.2750073373,-0.0001245085,0.3304243982,0.6380137801,-0.1410859525,-0.5410779715,-0.0631846488,-0.0811768249,-0.2349257469,0.1993419528,-0.2498530447,-0.2083238214,-0.2352619022,0.1023960337,-0.0581367277,0.3563393056,0.281555891,0.0985948294,0.1995158792,0.1826244593,-0.1368430406,0.5043628216,0.043215435,0.0561958142,0.0361742526,-0.2274192423,0.2372251302,-0.2068565637,0.0459311418,-0.2663237154,-0.0578701384,0.3306712508,-0.2204922587,0.0874133259,0.1752309054,-0.0942779183,-0.0104458947,0.8057671189,0.0052634547,0.0640940294,-0.201214239,0.0806409568,-0.1908854842,0.2383200824,0.0715610385,-0.4543537199,-0.1370184273,0.1143337935,-0.1281690449,-0.341394484,0.3352718651,-0.0335033908,0.4075168669,-0.0230080616,0.328975141,0.1422492564,0.1264762729,0.2072722167,-0.2324920595,-0.1268543601,0.1452392787,-0.3808562756,0.174837634,0.1763764918,0.2966895998,-0.0637696385,0.1829950064,0.0932282135,-0.1865078211,0.0735035315,0.1276494712,0.3398443162,0.1209282503,0.1319723129,0.0127151208,0.0047444995,-0.0747593865,0.0249654111,0.0320717469,-0.0236140732,-0.36054492,-0.2406122535,0.2245736271,0.2357540876,-0.0485627688,0.0452682115,-0.2116847932,-0.3131856322,-0.2082335055,0.1694273204,0.4560175538,0.0537248924,0.4907977581,0.3138802648,0.1613340378,-0.1267466247,-0.5722319484,-0.1803703606,0.3171550035,-0.161258176,-0.1637429446,0.3268589377,0.0613696501,0.2865219116,-0.2762343287,0.1872421801,0.1734672487,-0.0168005116,-0.1587234586,-0.2288744003,0.1743646711,0.1935789287,0.2789788842,0.2141343802,-0.2745317817,0.0189148672,0.2086554021,-0.2223849446,-0.1689475477,0.2648113072,0.042568069,0.107154578,-0.1384298354,-0.507543385,0.1115983129,-0.1357539892,-0.2659593821,-0.2610871792,-0.2019345164,-0.0996060669,-0.2297554612,0.3767790496,0.5294706821,-0.0989685282,-0.0291802213,-0.0063148174,0.1354220808,0.0829522833,0.0357792452,-0.311412245,0.492174834,-0.1760831773,-0.1047421619,0.8686538339,-0.5204976797,-0.0506761558,0.3639996946,0.1172987446,-0.071886532,0.3070545495,0.0618702248,0.1170750111,-0.0958950073,0.0317528211,0.4021973014,-0.0465446152,0.0234558284,-0.1533483863,0.0390015058,0.2969612777,0.3607648909,0.0934307128,0.2212417275,-0.2734820545,0.4947689474,0.2071569562,0.0037385139,0.0605673529,0.5123981833,-0.3901424408,-0.0827326104,0.0390271321,0.113077119,0.0266657285,-0.0488630608,-0.1335092783,0.1562322825,0.1626265943,0.1614460051,-0.4275047183,0.0230152868,-0.0830781758,0.0427630432,0.0567392223,0.2840376198,-0.1937215477,0.2580310404,0.0285491738,0.3317639828,0.0175104998,-0.0234073736,-0.3775363863,0.2045926601,-0.1200193763,-0.2554772794,0.282967478,-0.0228369217,0.4561701119,0.0420877747,-0.1734325886,0.1393243074,0.1799443066,-0.0209084228,0.0413391441,-0.2081551701,0.0283872373,-0.134917751,-0.0474718511,0.5747932792,0.2473167032,-0.193128854,0.0073986109,-0.0563839599,-0.0950333998,0.2234849483,-0.0520732105,0.1890359968,0.0732780173,0.2233697176,-0.0849315748,0.2273704857,0.204699859,0.1990070641,0.2141200006,-0.0926080421,-0.0686815456,-0.2286660373,0.2874112725,0.2148696929,0.0324278735,-0.0144205429,-0.2963917255,0.002503603,0.2251099199,-0.0638526604,0.2678775489,0.1726977378,-0.12468867,-0.0075488235,0.0018160535,-0.1072777137,0.4460032582,0.0564808063,0.261120379,-0.016631458,-0.1084316075,-0.0197839774,-0.2426254004,-0.2545631528,0.0278431252,0.2242687643,-0.4247313738,0.0356792621,-0.2087251991,-0.6600081921,-0.2359620631,-0.0494453721,-0.1605994701,-0.5055602193,-0.1612770408,0.3273543119,0.0146426978,0.1275832504,-0.120585151,0.2305703163,0.3319926858,-0.2339310497,-0.4185202122,-0.1734266132,-0.2996529937,-0.0336430594,0.1122232676,0.132373184,0.1385983676,0.113913238,-0.0306637753,-0.0660983101,-0.0057053682,0.2152480781,-0.1090568751,0.123509571,0.2650186121,0.3777044117,-0.2389560044,-0.0974703133,0.3813408017,0.0026626792,0.1714810729,0.3198922575,0.0355916694,-0.075479649,-0.1061445251,-0.2500555813,0.0514185093,-0.6115214825,-0.1775638014,0.237305969,0.000052158,0.1542863101,0.2415067703,0.1805426329,0.1287678778,0.3176608682,0.0169246066,-0.2387425154,0.3329898715,-0.1693131328,-0.3372408748,-0.1376382411,0.1169654056,0.2338880152,0.219890222,-0.4601656795,-0.3833651543,0.29610461,-0.034795884,0.0065297475,0.1694480926,0.3062468469,0.017128082,0.0924849436,-0.2547699511,-0.198965624,0.0468889549,-0.2222736031,0.2826959491,0.2882086039,0.3215004206,-0.0384758487,0.9122626185,0.237421751,0.143084988,0.5309516788,0.1021699384,0.2991394699,-0.0367710814,-0.2422951907,0.230656445,0.1812363267,0.1366699338,0.2091348171,-0.249240011,-0.255463928,-0.4086242914,-0.2131759226,-0.2362272888,-0.3552210331,0.1634069979,-0.3094646931,0.2537305355,-0.1385103017,0.2027332932,-0.3603169322,-0.0995280966,0.1494172961,0.1440065354,0.1602015197,0.1130502746,-0.075225547,-0.0607112311,-0.3460043967,0.3344669938,-0.1206776649,0.3733889461,-0.0566632375,0.0655358359,0.0389598571,-0.0580293387,0.7135971189,-0.3010065556,-0.0036403185,0.2482844591,0.2207377255,-0.6750039458,0.1139069796,-0.112287201,0.2548101842,0.0049686255,-0.3782780468,-0.4797842801,-0.1921132207,0.1742551476,-0.0723284334,-0.1652673632,-0.0882530063,-0.2201903909,-0.3247160912,-0.30970186,-0.105983533,-0.0444047228,0.4491245151,0.2301389277,0.2729759812,-0.1422615498,0.0037111251,-0.0231207926,-0.1272257864,0.144898206,-0.020322958,-0.2366642058,-0.1947486848,0.516156733,-0.5861030817,0.3650775254,0.2983977795,-0.6307681799,-0.0487274341,-0.0185211748,0.2841162086,0.2116642445,-0.3109889627,0.0528074764,-0.0490276553,-0.1705605388,-0.0002037241,0.2304157615,0.4107171595,-0.3813044727,-0.0521856807,-0.5629246831,0.4374673963,-0.1350968629,-0.0104099624,0.2742978334,-0.12945728,-0.1884309798,0.6053571701,-0.0318610035,0.974200964,0.2021540403,0.1885179579,0.4231554568,-0.1277760267,0.6227615476,-0.1243739352,0.3219046295,-0.3893945515,0.0414476544,-0.0638145506,-0.1746109426,-0.0930563435,0.1332767904,-0.2606726587,0.2774880826,0.0704492256,-0.0549795069,0.0889615119,0.1255862564,0.1116446927,-0.1067715287,-0.5512533784,0.057797011,-0.4502734542,0.0162889734,-0.0729946271,-0.1618123502,0.0773974955,-0.5212675929,-0.1029234454,0.1655472964,-0.240270555,0.1557008624,-0.0565365739,-0.2547738552,0.1854233742,0.3622989655,-0.0209172349,0.056942869,-0.1994664818,0.3114800751,-0.5459028482,-0.0695615858,-0.1593795568,0.3023845851,0.28486678,-0.0251290761,-0.146742776,0.3340931833,-0.1659708768,-0.6099496484,-0.0495170727,-0.0853261799,0.4431837499,-0.0368481465,-0.2056553364,0.0775685012,-0.1524201035,0.0355418697,0.0395073146,0.0244935751,-0.2060181797,0.0688497052,-0.0528987795,0.0025283047,0.1737148464,0.4771814048,0.0580924861,0.0253521185,0.5821675658,0.2740312815,-0.0257173087,-0.1586009413,-0.2322812825,-0.340931952,-0.6085496545,0.1990737617,0.3044446111,0.0621341728,0.0874242857,0.2038391829,-0.0779726431,0.170503661,0.0560895205,-0.3049573004,0.1134140939,0.2060197741,-0.2436185032,-0.0629765019,0.4305720031,-0.0323971473,0.3100137413,-0.1491773576,-0.1858063489,-0.1241254285,-0.1878904104,0.1358189285,0.160010308,-0.2263966352,-0.1157580242,0.2592052519,0.0706709921,-0.2440284193,-0.2439692467,-0.060357634,-0.1667722315,0.2002599239,0.2243479937,0.1214103028,-0.12901932,-0.4042980671,-0.123750478,-0.1816293448,0.0225903336,-0.1426826268,-0.2429104149,0.0557440706,0.1504145712,0.3674416244,-0.5081473589,0.1975682825,0.0760390013,-0.0926693827,0.2132937163,0.1916123182,0.262457788,-0.0496929251,-0.4662693143,0.1615186185,0.2004982233,-0.0344547965,0.5295451283,-0.1840267777,-0.1134769991,-0.1094243154,-0.0806771889,0.0807791799,-0.1513764858,0.0117968936,0.0692100525,0.0332740247,-0.1537947059,0.0494468063,0.3225930333,0.2039945722,-0.0651857853,0.0736154616,-0.0870866776,-0.366114676,0.2096334994,0.1163535565,0.3633618355,-0.1511389464,-0.1119756177,0.1339500695,-0.0349585712,-0.0854377896,-0.0527979434,0.2270284742,0.053739287,0.2903501391,-0.1444785446,0.0543756522,-0.0564408787,-0.3188199699,-0.0471042953,-0.1534538865,0.2008984089,0.4862774611,0.1516952962,0.1443457007,-0.1632328033,0.1997066289,0.0482344143,0.1141901091,-0.2721175551,0.1604944468,0.0182803292,0.0310199037,-0.6282174587,-0.0157810338,-0.4327570498,-0.1448920965,-0.004945952,-0.1630238593,-0.0300583038,0.211942181,-0.1169978753,-0.3793839514,0.2154737115,0.121847786,-0.0333387293,-0.1512890309,0.2037801892,0.0854873583,0.1178439409,0.1998904943,0.2706089318,0.3540655077,0.4170272052,-0.2698945403,-0.1693301797,0.2886002958,0.0005997109,-0.2072156966,-0.0614681467,0.0971704423,0.3210936785,0.3959818184,0.0738284066,0.0143997325,-0.0085142581,0.1275813282,-0.0613215007,-0.23826316,0.472381264,-0.0463526249,-0.0918838382,-0.3303665221,0.0329027697,-0.5663148761,0.3297737837,0.0595323928,0.3573003113,-0.0106479805,-0.2045569271,-0.0199389141,-0.0656273142,0.3586010933,0.3856936395,0.3740837276,-0.3091628253,-0.4248059392,-0.4309866726,0.158187449,0.1639747918,-0.2411166281,0.1070591733,0.1998973787,0.2831844389,0.2404049635,-0.1074828804,-0.2510265112,-0.0557258539,0.1786379665,-0.218937248,-0.2895013988,0.2064394653,0.0696631521,-0.1936567277,-0.2178909183,-0.1419939697,0.0083855763,0.0918040499,-0.1656745374,-0.3625209033,0.1477167904,0.0822409987,-0.0218153838,-0.0256654713,0.4474615753,0.0580903478,-0.0019750001,-0.1832550913,-0.0743170604,-0.1690218449,-0.2816382647,-0.1547074914,0.5438709259,-0.2213325948,0.2525574267,-0.3257875443,0.2445088774,-0.1334824413,0.3438494503,-0.1466025561,0.0648301616,-0.4684083462,0.3983730972,-0.1344812959,-0.0675862655,-0.0255529806,0.14621903,-0.4820397496,-0.2730072737,0.566480279,-0.5063444972,-0.2313738167,0.0063175997,0.2157868743,-0.4116660357,0.3354253471,-0.3584127426,-0.0386067405,0.2855788767,-0.067324236,-0.144484818,0.165515691,-0.0476102047,0.1122501269,-0.0831059441,-0.5315554142,0.0437982604,-0.1335093379,-0.1396399587,-0.2604459524]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/507","title":"Errors when I use ","comments":"Looks like an issue with 3.0.2 transformers version. Works fine when I use \"master\" version of transformers.","body":"I tried the following example code from https:\/\/huggingface.co\/deepset\/roberta-base-squad2 and got errors \r\nI am using **transformers 3.0.2** code .\r\n\r\n\r\nfrom transformers.pipelines import pipeline\r\nfrom transformers.modeling_auto import AutoModelForQuestionAnswering\r\nfrom transformers.tokenization_auto import AutoTokenizer\r\n\r\nmodel_name = \"deepset\/roberta-base-squad2\"\r\n\r\nnlp = pipeline('question-answering', model=model_name, tokenizer=model_name)\r\nQA_input = {\r\n 'question': 'Why is model conversion important?',\r\n 'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'\r\n}\r\nres = nlp(QA_input)\r\n\r\nThe errors are :\r\n\r\nres = nlp(QA_input)\r\n File \".local\/lib\/python3.6\/site-packages\/transformers\/pipelines.py\", line 1316, in __call__\r\n for s, e, score in zip(starts, ends, scores)\r\n File \".local\/lib\/python3.6\/site-packages\/transformers\/pipelines.py\", line 1316, in <listcomp>\r\n for s, e, score in zip(starts, ends, scores)\r\nKeyError: 0\r\n\r\n","comment_length":17,"text":"Errors when I use \nI tried the following example code from https:\/\/huggingface.co\/deepset\/roberta-base-squad2 and got errors \r\nI am using **transformers 3.0.2** code .\r\n\r\n\r\nfrom transformers.pipelines import pipeline\r\nfrom transformers.modeling_auto import AutoModelForQuestionAnswering\r\nfrom transformers.tokenization_auto import AutoTokenizer\r\n\r\nmodel_name = \"deepset\/roberta-base-squad2\"\r\n\r\nnlp = pipeline('question-answering', model=model_name, tokenizer=model_name)\r\nQA_input = {\r\n 'question': 'Why is model conversion important?',\r\n 'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'\r\n}\r\nres = nlp(QA_input)\r\n\r\nThe errors are :\r\n\r\nres = nlp(QA_input)\r\n File \".local\/lib\/python3.6\/site-packages\/transformers\/pipelines.py\", line 1316, in __call__\r\n for s, e, score in zip(starts, ends, scores)\r\n File \".local\/lib\/python3.6\/site-packages\/transformers\/pipelines.py\", line 1316, in <listcomp>\r\n for s, e, score in zip(starts, ends, scores)\r\nKeyError: 0\r\n\r\n\nLooks like an issue with 3.0.2 transformers version. Works fine when I use \"master\" version of transformers.","embeddings":[0.1425897479,-0.3862664104,0.101900138,0.2882798612,-0.0408770628,-0.0037297707,0.360788852,0.286631912,-0.2326703668,-0.0361642577,0.1111190468,0.1749107987,-0.3020108342,-0.1347761005,-0.1869156063,-0.2722382843,0.2231000662,0.1648703665,-0.2249961048,-0.075563103,-0.0981416553,0.4294788837,-0.5823258162,0.3994284868,-0.6540243626,0.0634840205,-0.0976940095,-0.0882224217,0.2179895788,-0.4048303068,0.1664741337,-0.1379939169,0.4408763647,0.4950669706,-0.0001128557,0.0662931651,0.5166268349,0.0034337423,-0.1831671298,-0.4616500437,-0.0958666354,-0.021679448,0.2401293516,-0.1488670856,0.1099354774,-0.0095904814,0.1123630106,0.1161488667,0.6158462167,0.2604494393,0.2036687583,0.5679879785,0.1606728882,-0.1586021334,0.0386505686,0.0651195422,-0.1622345448,-0.2532220483,-0.0711850077,-0.1541238725,0.0513456874,0.3433673978,-0.1545834094,-0.319506824,0.1965100169,-0.0248659048,0.3771715462,-0.6403257251,0.1768134832,0.344715029,-0.0220762305,0.0297095384,-0.4338785112,-0.370990634,-0.0477737077,0.0846328735,0.0715651065,-0.0575927012,-0.1082190797,-0.0320369713,-0.3570268154,0.1110915169,-0.2577053308,-0.0387131535,-0.0056881788,0.7290472388,-0.0043151458,0.115419209,0.0263560321,-0.3988368511,0.1119359508,-0.0886761546,-0.1108078659,0.2810792029,-0.4623845816,-0.4678985178,-0.1054500863,-0.0141506167,-0.1258595437,0.132976383,-0.047635071,-0.0400705934,0.1146947145,0.050596036,0.2708026767,0.3776749372,0.3136990964,0.3700650334,0.3415612578,-0.0040864917,0.1595462114,-0.0196935255,-0.4887844622,-0.0226209201,0.0947434604,0.1682706028,0.2845754027,-0.0403796956,0.0695829839,-0.0889555886,-0.2725495696,0.0308969915,0.2877452075,0.3125798106,0.1607083231,0.4081320167,-0.0442818627,0.3618612885,-0.1540920287,-0.4788371027,-0.0482453443,-0.0248205196,-0.1305386573,0.0259037018,0.183268398,0.3635356128,0.2959641218,-0.3133236766,0.1574114859,0.1929110885,0.0696845949,-0.2652225494,-0.1392059326,0.222193867,0.0477906168,0.1235276908,0.2732222676,-0.1546162963,-0.0269144867,0.2888833284,-0.152669847,-0.0730928034,0.0382680893,0.128095746,-0.1564274281,-0.2049593031,-0.1068510935,0.1424414068,0.2601503134,-0.0196878687,-0.1511559039,-0.2680458724,-0.0712297335,-0.0273503531,0.0152345067,0.4272629619,0.1217842326,-0.2629700005,-0.0074936883,0.1439886689,-0.105313167,0.1660331339,-0.0594261736,0.239754051,-0.0067093009,0.3142331541,0.3387302756,-0.2347958386,0.0610828958,0.1511665434,-0.0109123942,-0.2349623442,0.055894576,-0.1820522696,0.1183111966,-0.2271659523,-0.0724461749,-0.1401990652,0.0083090318,0.1611102819,-0.0977460369,-0.2160190493,-0.0398633629,0.2065846771,0.1662975699,0.1019943804,-0.379703939,0.5313245058,-0.1683852822,-0.190780282,0.1780082732,-0.0916158035,0.4059320092,-0.241859749,0.0768683404,-0.1072881296,-0.2731477022,0.2203312218,-0.498021841,0.1964557916,-0.0521731004,0.2371753007,-0.1382563263,0.0803377777,-0.1817128807,-0.0770964175,0.170181185,-0.0776247233,0.1599178612,0.1136876568,-0.2996758223,-0.0866934285,0.4864613116,0.0755586326,-0.3800067008,0.2390677333,-0.256290853,-0.3221542835,-0.172824353,0.1800199896,0.4133439064,0.1605977416,-0.0603638999,0.2198885977,-0.2134710997,-0.0920353681,-0.1008523032,0.3315393627,0.2418498993,-0.3857061863,0.153454572,0.4965134859,0.1735297292,0.1489476264,0.1390587538,0.4117442369,0.049088195,0.1475323439,0.0893392488,0.1436101794,0.027317645,-0.1975849867,-0.1786641628,-0.0489501692,0.0124958521,0.0653677955,-0.0209296271,-0.3088859022,0.0854294524,-0.190109387,0.4991106093,-0.0220074784,0.1619540155,0.1475793719,-0.4284462333,0.040990755,-0.1600481272,0.2272354066,0.3109731674,0.1093931422,-0.1593066603,0.1008156389,-0.0423444621,-0.0339870341,0.330276072,0.0721159279,0.2826166451,0.1288802326,0.1914828718,0.1401966363,0.1415277123,-0.2043960243,-0.0989474133,0.203945294,-0.3245393932,0.1887784898,0.0387539268,0.0697252005,-0.1371349692,-0.243154645,-0.2385838777,-0.3978781402,0.0922368541,0.0347949155,0.0489942655,0.2307594121,0.0914829001,0.2400474399,0.2203381807,0.4928330481,-0.0906588361,-0.0119173201,-0.4879196286,-0.0100958925,0.0699659437,-0.5400472283,-0.0604566336,0.0447766706,-0.0678852722,0.2947699726,-0.3794869781,0.3721455038,-0.1519844085,0.0831870884,0.1768323928,-0.1078044176,-0.0183152072,-0.3880954683,0.429258436,-0.3619121611,-0.0171322078,0.3871999681,0.0423268676,-0.1268575191,-0.2220814228,-0.0811966434,0.0566206016,-0.1438731849,-0.0924244896,-0.134232834,0.1725471616,0.4180764258,0.02203325,0.4735864699,-0.2557699978,-0.1696811765,0.1678778827,0.1819523722,0.5343099236,-0.0216974895,-0.3780197799,0.1458331943,-0.2738334537,-0.0258694589,-0.5093508363,-0.0553693734,-0.25315997,-0.1245511249,-0.1565585136,0.1502877772,0.0557691082,0.2515535653,0.0170890093,0.00738212,0.0466912054,-0.1840882599,0.1286096573,0.2061227858,0.3723753393,-0.0267038718,0.4878305793,0.10480652,0.629336834,0.3555672169,-0.0857290626,0.1341524422,0.0791091472,-0.2296394855,0.0292939637,-0.2251011133,-0.3233571649,0.0312585235,-0.0029998603,0.1798377484,0.0321976617,-0.2912757695,-0.178267926,-0.1114204377,-0.1802411973,-0.2081420124,-0.1353159547,-0.179733187,0.1601634622,-0.2067034245,0.2087402195,-0.2168794423,0.0219858289,-0.0253866762,-0.0939738154,0.0660040006,-0.1992715746,-0.4379247725,-0.1425551176,-0.2592296004,0.4074148834,-0.1205333099,0.3221701682,-0.0108913295,-0.357421726,-0.0811125264,-0.1580724567,0.6426005363,-0.1390682906,-0.179114446,0.086758323,-0.2970716655,-0.4907179773,-0.1606562436,-0.3340063393,0.0472242311,0.4098540843,0.3750634491,-0.2564999759,-0.0656044483,0.0633754134,-0.5187550783,-0.1749586612,0.0412300415,-0.2884531617,0.0803984106,-0.1542812437,-0.1534996331,-0.0503883697,0.122549817,-0.1448912323,0.1505256593,0.1205115393,-0.2101489753,0.0912968218,0.4907368124,0.081281364,0.000574711,0.0183104184,0.2177863121,-0.0035043452,-0.1578535587,0.0480025783,0.2946721911,-0.3666608334,0.2672851086,0.1955073774,0.290158242,0.1951660514,-0.0157654602,0.0167655572,0.0211580414,0.2260190398,0.1106333807,-0.3405204117,0.3545123935,-0.012284725,-0.1532619596,0.0581079684,0.1402445734,-0.3238540292,0.1420631707,-0.4734607339,0.109843336,-0.1724623144,0.616448164,0.0426818393,0.9207695127,0.1358619332,0.0129353032,0.0848222449,-0.0169215295,0.5575109124,0.1966556162,0.1872054487,-0.4816451371,0.2159908265,0.022424018,0.054521434,0.0828108713,0.160476014,-0.5252497196,0.2450695187,-0.0070608263,0.1365507394,-0.0562600307,0.158865869,0.1024962664,-0.1845778823,-0.3168312013,0.159211114,0.0363697745,0.2058607787,-0.0185142681,-0.2774508297,-0.1049805209,-0.3573000431,-0.1582355052,0.3035014868,-0.1292399615,-0.08576525,0.7080216408,0.2028636932,0.6197367907,0.5606236458,0.5166891813,0.053359434,-0.4011372626,-0.1151210815,0.0267613884,-0.0420548581,-0.1256859452,0.0420052782,-0.0431938395,-0.1279369295,-0.2508966327,0.0883573741,0.1080322787,0.1712237298,-0.1111465618,-0.0848674402,0.1050732955,-0.1521772146,0.0199646577,-0.1242935583,-0.0758291259,0.0386434682,0.1679321676,-0.0929518715,-0.3017591834,-0.2810387909,0.1419226825,-0.0742103085,0.2143646628,0.3510378599,-0.0347982682,0.3840599954,0.6542611718,0.0626151636,0.1027651057,-0.1921772212,0.2396259457,0.0520185344,-0.6680747271,0.1615445465,0.0713285133,-0.4031158984,0.2372424155,0.3277624249,0.223409161,0.1068715006,-0.1623993218,0.0089589879,-0.3093211055,0.2055793107,0.0101804249,-0.1933353543,-0.1214145273,0.0279884711,0.1757922322,0.1185592636,-0.3084102571,-0.1066320762,-0.4146066606,0.0186440889,-0.4001933038,-0.6514692903,0.0922470242,0.0401496179,0.0978696421,-0.002570187,-0.2797648311,-0.1697243005,-0.1700242311,0.1263951808,0.3313119411,-0.2610855997,-0.2167741209,0.0373911113,-0.085464038,-0.1478115171,0.0160994362,0.0890379846,0.1297553629,0.4998903275,0.3431233466,0.1461975574,-0.3304530978,0.0130643472,-0.0009318416,-0.1115221977,0.3274611235,0.1754567921,0.072292462,-0.0946314707,-0.3950279355,0.2633705735,0.2868724167,-0.050698474,0.3815041184,-0.4005307257,-0.0263228714,0.0040735961,-0.127041012,-0.0061292849,-0.1277551949,0.0248195902,0.0619152039,0.109987475,-0.252478838,0.1654843688,-0.1811101437,-0.1280916631,-0.1367939711,0.2108073235,0.0977656469,-0.0270173959,0.2516926825,0.1957516372,0.5039551258,0.1116122976,0.2541822791,0.2503435016,-0.1214849353,0.1640223712,0.3342103958,-0.4417216778,-0.0433260351,-0.2629716396,-0.1226976886,0.2609669268,-0.2051965147,0.2217993885,0.3339415193,0.0073817614,0.0020573663,0.11502368,0.3605110645,-0.0688447058,-0.3044216037,-0.0567799956,0.025882626,0.1380682588,-0.5615828633,0.1502878666,-0.1157252043,-0.4203839302,-0.2101976573,-0.1604299992,-0.3518047631,-0.2396868467,-0.1430501789,-0.194539398,-0.2422722727,0.0659594312,-0.2627659738,0.0042390432,-0.0832630917,0.4153215289,0.1173381582,-0.119824931,0.3166148961,0.2233848572,0.1095568016,-0.074865751,0.3391101956,0.4799001217,0.5285980105,-0.5117017627,0.1337428093,0.0446158871,-0.1574282497,0.3019146025,0.164059788,0.1410888135,0.3766641617,-0.0836732537,0.0264850706,-0.1202347353,-0.0275304429,-0.3470242321,0.3050808907,-0.0769748837,0.6026279926,-0.1700717658,0.0705299377,-0.0380335301,0.0176379755,-0.6420901418,0.2209078521,0.2818742692,-0.0080038207,-0.0216587465,-0.0877609402,0.0681823865,-0.1623540521,0.3674947619,0.3344620764,-0.2495416552,-0.3301184475,0.0222446881,-0.235827595,0.1758079231,0.02058216,0.246853292,0.1701909453,0.1396845877,0.0877762735,0.0432502516,-0.2119872868,-0.2127484083,-0.3202986121,0.2325865179,-0.2264475971,0.1626768261,0.0917378813,0.1630987525,0.0970385745,-0.1210403666,-0.0209503509,0.1817311943,-0.041223336,0.1883678138,0.0586961061,0.032441888,0.1596479863,-0.0109967822,0.1578501165,0.4102101624,0.1294571161,-0.1036412194,-0.1486009061,-0.2190626711,-0.3594355583,-0.2469455302,-0.0665294454,0.4473225474,-0.036766164,-0.0715218782,-0.2113655508,-0.2242801785,0.0165896155,-0.1691882461,-0.3869923353,0.1617828608,-0.1489507109,0.0543511957,-0.0128800422,-0.090609543,-0.0042412016,-0.0085157072,-0.3282026649,0.0023106902,0.5907205343,-0.5755989552,-0.2532939315,-0.2007430345,0.0924839005,-0.1720797569,-0.0598422624,-0.4238317907,-0.2063874602,0.3250182569,-0.1363997459,-0.2507224381,0.178562209,0.1819871068,-0.1021631882,-0.0824737698,0.1303395927,0.1093230471,-0.0676947236,-0.0570605658,-0.0815464184]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/501","title":"Caching doesn't work for map (non-deterministic)","comments":"Thanks for reporting !\r\n\r\nTo store the cache file, we compute a hash of the function given in `.map`, using our own hashing function.\r\nThe hash doesn't seem to stay the same over sessions for the tokenizer.\r\nApparently this is because of the regex at `tokenizer.pat` is not well supported by our hashing function.\r\n\r\nI'm working on a fix","body":"The caching functionality doesn't work reliably when tokenizing a dataset. Here's a small example to reproduce it. \r\n\r\n```python\r\nimport nlp\r\nimport transformers\r\n\r\ndef main():\r\n ds = nlp.load_dataset(\"reddit\", split=\"train[:500]\")\r\n\r\n tokenizer = transformers.AutoTokenizer.from_pretrained(\"gpt2\")\r\n\r\n def convert_to_features(example_batch):\r\n input_str = example_batch[\"body\"]\r\n encodings = tokenizer(input_str, add_special_tokens=True, truncation=True)\r\n return encodings\r\n\r\n ds = ds.map(convert_to_features, batched=True)\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\nRoughly 3\/10 times, this example recomputes the tokenization.\r\n\r\nIs this expected behaviour?","comment_length":59,"text":"Caching doesn't work for map (non-deterministic)\nThe caching functionality doesn't work reliably when tokenizing a dataset. Here's a small example to reproduce it. \r\n\r\n```python\r\nimport nlp\r\nimport transformers\r\n\r\ndef main():\r\n ds = nlp.load_dataset(\"reddit\", split=\"train[:500]\")\r\n\r\n tokenizer = transformers.AutoTokenizer.from_pretrained(\"gpt2\")\r\n\r\n def convert_to_features(example_batch):\r\n input_str = example_batch[\"body\"]\r\n encodings = tokenizer(input_str, add_special_tokens=True, truncation=True)\r\n return encodings\r\n\r\n ds = ds.map(convert_to_features, batched=True)\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\nRoughly 3\/10 times, this example recomputes the tokenization.\r\n\r\nIs this expected behaviour?\nThanks for reporting !\r\n\r\nTo store the cache file, we compute a hash of the function given in `.map`, using our own hashing function.\r\nThe hash doesn't seem to stay the same over sessions for the tokenizer.\r\nApparently this is because of the regex at `tokenizer.pat` is not well supported by our hashing function.\r\n\r\nI'm working on a fix","embeddings":[-0.0136866393,0.0475987047,0.0673073158,-0.0456753708,0.0919944867,-0.183521077,0.2613239586,0.1327966154,0.0008185431,-0.3439307809,0.0885018855,0.2487123013,-0.1410467774,-0.2732979357,0.1312420666,0.1709653884,0.3097136021,0.044130858,0.3055776656,-0.0806424171,-0.0654457957,-0.0272676107,-0.3542217612,-0.0297061186,-0.5633124113,0.0520044453,-0.1355011463,-0.4004319906,0.2827106118,-0.3911952674,0.0991080254,0.1787096262,-0.0134690013,0.2490125597,-0.0001165723,-0.1295522749,0.3142707944,-0.0642244592,0.2027622312,-0.1374648809,-0.1056158543,-0.3216924667,-0.1589991748,-0.0220345352,-0.0848208889,-0.2071923018,0.2631583512,-0.1060144603,0.5493525863,0.0068813544,0.1311336756,0.17380023,-0.1037610248,0.3160282373,0.0371434055,0.0110689765,-0.1272770464,-0.2586758137,-0.108120963,-0.3813346922,-0.4259673953,0.1791107953,-0.1774550974,-0.1869017035,0.2517661154,0.2371277362,-0.059267953,-0.1703378409,0.2343092263,-0.1654309332,0.431982547,-0.2239157557,-0.2962221503,-0.0784381852,-0.2797979116,-0.1567181647,0.4439513683,-0.0860385969,-0.0714854971,0.0591570735,-0.4800797999,0.1590325534,0.1002617851,0.0925735608,0.0439992175,0.530156672,0.0888788626,0.109763898,-0.1824956685,-0.2295015305,-0.0440586209,-0.2285477072,-0.1261464804,0.4742988944,-0.1523884535,-0.2847386003,0.0262276139,-0.080077678,0.0544599853,0.2948219776,0.0504115596,0.2815189064,-0.0299814846,0.2159365416,0.0078934794,0.5478121638,0.1964381784,0.4425427914,0.1908519268,-0.558888793,-0.4889090955,-0.0812601894,-0.0298012141,0.0343143567,0.6790271401,0.0561272353,-0.3889093995,0.034137778,0.185283035,0.0584528856,-0.3547094464,-0.1792828739,0.0523776412,0.2285992056,-0.1112921014,0.0362102911,-0.5301241279,0.0330119133,0.0699479878,-0.1227097288,-0.2177218795,-0.1649406105,-0.4442878962,0.258090049,0.1851163954,0.4816352427,0.2829695344,-0.1187253073,-0.1610891521,-0.1224146932,0.1046068221,-0.2712718546,0.4402396381,-0.0034611151,-0.3136524558,0.1900182813,-0.0267266892,-0.0141044986,-0.3247203827,-0.0082997931,-0.0187004078,-0.084729135,0.4318655431,0.0427012108,-0.2380245626,-0.134765327,-0.0299305357,0.2931546271,0.5849563479,-0.2344622761,-0.0222435258,-0.0523846932,-0.3115319312,-0.3348666131,0.1609579176,0.2611805797,0.2704752386,-0.2919741869,0.4862928092,0.3912269771,0.3023074567,0.5853936672,-0.1874111891,0.1277573705,-0.0788884088,0.6074695587,0.2261340618,-0.1387712359,-0.5778239369,0.0163231231,0.0931184664,0.0922376215,-0.0854177698,-0.051496163,0.3476942182,-0.2769735754,0.0319776386,0.3163945079,0.0487544164,0.0056761322,-0.3302252293,-0.0635455847,0.4362440705,-0.2102127224,0.1992403716,-0.0479143448,-0.0353894569,0.1872395277,-0.2263924479,0.0907613039,0.2284286469,-0.0491260327,0.1635676622,-0.180558756,0.126012668,-0.0757405981,-0.1877246797,0.1920467019,-0.1871372163,-0.1807709485,0.0339365825,-0.1134052947,0.1624483615,-0.1431271136,-0.0662772954,-0.2908812165,0.1051157862,0.1116528437,0.3874667883,0.1515065134,-0.1465882659,0.1243401989,0.4645557106,0.0119360471,0.1125247702,-0.0979791507,0.0455767773,-0.2353745401,-0.3090813458,0.1631506234,0.2418068349,0.057264138,-0.1184217185,0.1625994593,-0.0288488586,0.2141636163,-0.0700809807,0.404435277,0.0787220895,-0.0859916061,0.0643275753,0.3019316792,-0.1264505833,-0.2796177566,0.0359570049,0.4145944118,0.5003684759,-0.0235094652,-0.1423827559,0.0557339229,0.3090839684,-0.1524005234,-0.1168998852,-0.1637314558,-0.0872419178,-0.0906469971,0.1349387914,0.0935462117,0.090276517,0.1426397264,0.3982763588,-0.0260449871,0.0850622952,0.0565835275,-0.3832198083,-0.5182855129,0.1002304256,-0.0206213463,0.0375779271,0.0057199551,0.266020298,-0.0137696704,-0.1720893979,-0.1545707136,0.1181880981,-0.0653091893,-0.2103818655,0.1019193605,0.1604079902,0.3006861508,-0.1281342357,-0.1357728988,0.247564584,0.007779927,-0.0536922999,0.0689047426,-0.2456957102,0.0005732215,0.2347133905,-0.0898354203,-0.2250187546,-0.2172689438,-0.1392380297,0.2956092954,0.0359280705,0.4442127645,-0.1402324438,0.0842037648,-0.0456857011,-0.1548352391,-0.1630382985,-0.3049255013,-0.2561385632,-0.0947617739,0.2327451557,-0.2959227562,0.2157171071,-0.0149192456,0.0495578647,-0.1273158491,-0.3565973639,0.164893046,0.0026206882,-0.0809675604,-0.1507011354,-0.3412349522,-0.5605165362,-0.0347067751,0.2581994236,-0.4532195628,-0.1404502988,0.222409606,0.1876201928,-0.1044222191,0.0156461075,0.1428369433,-0.0484263301,0.0104427459,0.1698827595,-0.0645508692,0.091172412,0.4004296958,-0.2932043076,0.0047200359,-0.4901485145,-0.1463362277,-0.3706046045,-0.2366392314,0.7362596393,-0.110546343,-0.0800710991,-0.0261626262,-0.1154209673,0.2398850471,0.5107489228,-0.1860038191,0.1623386145,-0.0766512305,0.1053569019,0.0962623581,-0.3081242144,0.3067164123,-0.0934617147,-0.0706831813,0.0511957854,-0.1466677636,0.1430201828,0.2608576715,0.3049932718,0.2392116487,0.091107592,0.3536170721,0.8350927234,0.0838100836,-0.1541403979,0.1790556163,0.2011922449,0.0561240166,-0.1676304936,0.0182382129,-0.0649655089,-0.1081262901,-0.0374112986,0.1951103806,0.1247869506,-0.4296417534,-0.0752571821,-0.2870825529,-0.0773975104,-0.3653160632,0.2314306498,-0.526329577,0.3536961377,0.2522393763,0.4212900996,-0.5082790852,-0.3702867031,0.0515362173,-0.1205800101,0.3768258095,0.0275495313,-0.6036427617,0.1598292142,-0.6390217543,0.225487411,-0.0314551331,0.6086584926,-0.0492966995,-0.1263486445,-0.1125274301,-0.0273518711,0.3081949651,-0.3621613979,0.1180663258,0.1659582555,-0.261436969,0.0024888604,-0.1347979307,0.0365083478,0.1394207925,0.3713922203,0.6505627632,-0.1984954476,-0.1003762484,-0.6000760794,-0.0540398136,0.0648796931,-0.2873961031,0.0734634623,0.3018555641,-0.1909439862,0.0821540356,0.1917179674,-0.0974146649,0.2428386658,-0.0313902162,-0.1940697879,-0.1073344722,0.2325009108,0.0415738374,0.3474744558,-0.0098961806,0.2071543038,0.2720626593,0.1065755561,-0.2526548207,-0.3211754262,0.2206668407,-0.2290359288,-0.0693346485,-0.0051415078,0.2123206407,0.1209633723,-0.0375125185,0.0440714248,-0.148895666,-0.0146017624,-0.2373541594,0.1671671271,0.0833694413,0.1954215467,-0.379445672,-0.3263631165,0.1258124858,0.1384030432,-0.180743441,0.2266241759,-0.3775848746,-0.4460231662,0.6914545894,0.2273261398,0.9793089032,0.1366843283,0.4066056013,-0.2610701919,0.245463118,0.3961776793,-0.2034376264,0.270539999,-0.2106606215,0.0620773435,-0.0508617014,-0.1326844096,0.0091233226,0.3172047436,-0.2797608376,0.3796658814,0.1443431675,0.8239807487,0.1463344842,0.0589602329,0.085770987,-0.2330422401,0.2865398228,0.1325599104,-0.0268880557,0.2350768894,0.0284592956,-0.0910462439,0.0394463018,0.0768702924,-0.1889703274,0.0690730214,-0.2721968591,0.172340855,0.1990005374,0.0919738859,-0.0357413143,0.1271830201,0.3098221421,0.1635277718,-0.2802835107,0.074983038,0.3586246669,0.2018197477,-0.0096237026,-0.3221385181,0.2645990849,0.0702608973,0.0540559478,-0.0745434165,-0.0545702912,-0.4214327335,-0.5006970763,-0.0484710149,0.1182112545,-0.0697837323,0.1867562383,0.0447833836,-0.2054714113,-0.132945329,0.1166347861,0.1708298624,-0.0546879545,0.4148551226,-0.1348784566,-0.0356569104,0.0283778924,0.4033699036,-0.0007586213,-0.1859470457,0.4934469759,-0.4574660361,-0.3638173342,-0.1566360742,-0.1653166264,-0.201334849,-0.2225387543,-0.0625699237,-0.3878855705,-0.1972301602,0.03753471,0.0448245332,0.1629532129,0.1089467034,-0.0389682129,-0.2716580033,-0.4200651348,-0.064196609,-0.0288735367,0.0158560798,-0.0422420092,0.2786876261,-0.0310846046,0.2577911317,-0.2614102066,-0.2017786801,0.0711543709,0.1786181033,-0.3131328821,-0.2629752457,0.117848888,-0.0698960051,-0.0040267641,0.341235429,-0.3315979242,-0.1127352193,-0.0775685608,0.1665164381,0.3449390233,-0.1828275323,-0.2137666941,0.1694703847,0.0159770437,-0.5175480247,-0.1979539245,0.308491677,0.0102923522,0.0097130397,0.1756763756,0.2858932912,0.1375933439,-0.097506389,-0.0166271329,0.088563554,-0.0466824435,0.2704510987,0.0063902349,0.0285551623,-0.2647026181,0.1746232063,0.1920980811,-0.0426922701,0.446392566,-0.0728394315,-0.1335800588,0.1615471542,-0.1888234764,0.184987992,0.0214437284,-0.4864824712,0.7638765574,0.1510971338,-0.016299231,-0.0525915846,0.3405271769,0.1530586183,-0.064923048,0.0520948805,0.4691830873,-0.0986259058,0.005894857,0.0498067886,0.5783456564,-0.4047413766,0.1436864883,-0.1437251568,0.0994394645,0.1940859407,0.3811707795,0.0763948038,0.0855784491,0.054260917,0.3561937511,0.2328817397,0.1432287693,0.2286987901,-0.3098290265,-0.2425928563,-0.2123387009,0.4649174809,0.0333407074,0.5086625218,-0.1132325232,0.2447827607,0.1591771245,-0.1347512752,-0.2319497764,0.303999573,-0.2283733487,-0.289488256,0.1007914618,-0.0436235331,-0.1505001038,0.1539199054,-0.1167203709,0.125316456,0.1442428827,-0.0083631463,-0.0354179777,0.0300402213,0.1455737054,-0.0579562411,0.3666917682,-0.3213549256,0.2709108591,-0.4509079456,-0.1322086304,-0.1833264381,0.1297281533,0.2746244967,0.3292092085,-0.3230865896,0.1327791661,-0.0697844401,0.0578023493,-0.1904818416,0.6730791926,0.0656595603,-0.4936017394,0.0241806433,-0.0414923653,-0.0289018285,-0.0774160847,0.0728832632,0.0147569347,0.1996299773,0.4930044413,0.0454483517,-0.138201654,0.2667305171,0.0464600325,-0.2540118396,-0.0944357216,0.1564908773,-0.0249837041,0.2481471151,-0.2527535558,0.0736774281,-0.0742076263,0.3581218719,0.3809100389,-0.1709431112,-0.2639420927,-0.064626433,-0.4245135486,0.402598232,-0.0921910107,0.2847801745,-0.2324804217,0.2075217217,-0.1105875671,-0.0049809776,-0.1382384449,0.0578960888,-0.320258677,0.0850499123,-0.1461157352,0.4202271998,-0.1469530314,0.2991520166,0.1546773165,-0.3179660738,0.2920421064,-0.0085773142,0.1318595707,-0.0959046111,-0.0870148093,0.0377331823,0.2726814151,0.5620135069,0.2327573597,0.0872584358,0.0696963519,0.1894726604,-0.3747022748,0.0216686875,-0.056568332,-0.0297164079,-0.1354915053,0.1955702007,-0.3513377309,0.1786046177,-0.1632318199,0.0025963881,0.0449981652,-0.4295828938,-0.0499900579,-0.0506925583,-0.1909506023,-0.0292213149,0.0003646286,0.3448747694,-0.1746665835,0.2510059476,-0.3981535137,-0.2146267593,0.229882136,-0.6583095193,-0.4971134067,-0.0440979488,0.0841892585,0.4080100656,0.0069476529,-0.2966827154,-0.215778172,0.276168257,0.0777905211,-0.2484920174,-0.1016387939,0.0264707115,-0.1246338785,-0.2842131555,0.2960050404,-0.0015489684,-0.2540474236,0.0405078344,-0.0360088758]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/501","title":"Caching doesn't work for map (non-deterministic)","comments":"Thanks everyone. Works great now.","body":"The caching functionality doesn't work reliably when tokenizing a dataset. Here's a small example to reproduce it. \r\n\r\n```python\r\nimport nlp\r\nimport transformers\r\n\r\ndef main():\r\n ds = nlp.load_dataset(\"reddit\", split=\"train[:500]\")\r\n\r\n tokenizer = transformers.AutoTokenizer.from_pretrained(\"gpt2\")\r\n\r\n def convert_to_features(example_batch):\r\n input_str = example_batch[\"body\"]\r\n encodings = tokenizer(input_str, add_special_tokens=True, truncation=True)\r\n return encodings\r\n\r\n ds = ds.map(convert_to_features, batched=True)\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\nRoughly 3\/10 times, this example recomputes the tokenization.\r\n\r\nIs this expected behaviour?","comment_length":5,"text":"Caching doesn't work for map (non-deterministic)\nThe caching functionality doesn't work reliably when tokenizing a dataset. Here's a small example to reproduce it. \r\n\r\n```python\r\nimport nlp\r\nimport transformers\r\n\r\ndef main():\r\n ds = nlp.load_dataset(\"reddit\", split=\"train[:500]\")\r\n\r\n tokenizer = transformers.AutoTokenizer.from_pretrained(\"gpt2\")\r\n\r\n def convert_to_features(example_batch):\r\n input_str = example_batch[\"body\"]\r\n encodings = tokenizer(input_str, add_special_tokens=True, truncation=True)\r\n return encodings\r\n\r\n ds = ds.map(convert_to_features, batched=True)\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```\r\n\r\nRoughly 3\/10 times, this example recomputes the tokenization.\r\n\r\nIs this expected behaviour?\nThanks everyone. Works great now.","embeddings":[-0.0653020144,-0.0202138368,0.0454191007,-0.046825923,0.0627297387,-0.2320473343,0.2753812373,0.1205503121,0.0586609133,-0.327898562,0.1037917882,0.3029954433,-0.0933355615,-0.227922067,0.0724128038,0.1814223677,0.3917804956,0.0900952965,0.2326811999,-0.1603681147,-0.122882165,-0.027148338,-0.4633234143,-0.1014404073,-0.6048887968,0.0553987771,-0.1914039552,-0.4963865578,0.3038679957,-0.3857289553,0.0703976676,0.2114217132,0.023792835,0.2152408361,-0.0001167598,-0.1829490215,0.326490581,-0.0239920318,0.1390181035,-0.077252768,-0.0638179183,-0.4001007378,-0.1070975289,-0.0578926317,-0.1952037662,-0.2246039659,0.2945444286,-0.0429317877,0.5579413772,-0.0348243117,0.1359409392,0.260347724,-0.002955504,0.3158698082,-0.0376345702,0.0729148984,-0.1443504095,-0.2609886527,-0.0028192133,-0.3891080022,-0.35265854,0.2260738164,-0.1551319808,-0.1614749134,0.2059939057,0.2399124652,-0.0442886315,-0.2419793159,0.2009918094,-0.2371627688,0.5421618819,-0.1216144115,-0.3020129204,-0.1723036915,-0.2303777933,-0.2011266798,0.4143952131,-0.0743337646,-0.0210168678,0.0497722551,-0.4701749086,0.1807063669,0.0494476892,0.1673293412,-0.0128928134,0.6631796956,0.0835289732,0.1248870715,-0.2742434144,-0.2743825912,0.0340177044,-0.2694793046,-0.0967003927,0.4743560851,-0.2160496861,-0.3108386993,-0.0229645222,-0.0951669514,0.0308344439,0.1300111413,-0.0840613842,0.2015557885,0.0515439212,0.1758266091,0.1100623012,0.5602411032,0.0718101189,0.5329988003,0.213250488,-0.5769847631,-0.4814466834,-0.0599973015,-0.0639258847,0.0747662485,0.6501581073,0.0530278236,-0.2369964719,0.060760919,0.0351736471,-0.0220580045,-0.2788511515,-0.1663747281,0.0423369072,0.2178771049,-0.1702241004,0.0511298217,-0.577872932,0.0975022316,0.0420040004,-0.1306936741,-0.2144740969,-0.2296663821,-0.4164915383,0.1852616519,0.1976918429,0.6082418561,0.2922904193,-0.120841004,-0.2218939215,-0.1572457999,0.14155294,-0.3102341592,0.3619471192,0.0378491655,-0.2648470998,0.2035553604,0.0164827071,-0.0282894801,-0.3012412786,0.0236020423,-0.1335204393,-0.0995507538,0.3618279397,0.0779589638,-0.172359243,-0.1862046719,-0.0284265913,0.3116667569,0.65926826,-0.1849716455,-0.0026543075,-0.0568991676,-0.323176682,-0.2410423756,0.1189938411,0.3158764541,0.2714449465,-0.2666344941,0.4864300489,0.4206514657,0.4093566835,0.6471379399,-0.2214821726,0.1638169289,-0.1300514489,0.4931869805,0.2474846989,-0.1693533063,-0.5759469867,0.0575645827,0.0326220803,0.0547599755,-0.1390762031,0.0522226878,0.3410359323,-0.2457308173,0.0337693505,0.3429498971,0.1386271417,0.0643612742,-0.3282088935,-0.037495587,0.4721898735,-0.2055945247,0.1758277118,0.0360431038,-0.0719527826,0.2154139429,-0.1821421832,0.0815458,0.2273689508,-0.0373418853,0.1772968471,-0.1875530183,0.1071577966,-0.0516093038,-0.234957844,0.20061405,-0.1871265322,-0.1867991984,0.0864345282,-0.1572236568,0.137879163,-0.1027239785,-0.1269734949,-0.2564879656,0.1115242466,0.1394729167,0.3687924147,0.1387793273,-0.1497005522,0.1301361471,0.3398651183,0.0486721992,0.1369336396,-0.0114841321,0.0763677433,-0.2256335318,-0.3104737699,0.1423216164,0.2774383724,0.0343515612,-0.1771728247,0.1392451972,-0.0472373776,0.3075819314,-0.1321673989,0.2997970581,0.1552118063,-0.121978797,0.0302317683,0.3692366183,-0.0214609876,-0.2681771815,-0.0106258169,0.4504292309,0.4662848413,0.025378922,-0.1023799926,-0.0095833289,0.2728211582,-0.1039049774,-0.0727568716,-0.187082395,-0.0203615278,-0.0841654986,0.1369693875,0.1256361157,-0.0706224144,0.1627443284,0.42398718,0.0171786528,0.0757417902,0.0526347123,-0.3284949958,-0.497736752,0.2024538219,-0.0602063127,0.0433367454,0.031004129,0.244054392,-0.110768266,-0.1210913137,-0.1613684446,0.1670750231,-0.1103777364,-0.1148020402,0.1138796136,0.2761962116,0.275218606,-0.2153663933,-0.0362360738,0.2743252814,0.0224871263,-0.0435669683,0.0263100211,-0.1545654684,-0.1472116411,0.2381845862,-0.0465882942,-0.2039805502,-0.1941530704,-0.1562561542,0.3320425451,0.0735759214,0.4605137408,-0.0963831544,0.069739446,-0.0676534399,-0.0607706681,-0.1348325461,-0.2721209824,-0.239028722,-0.0738152117,0.1467876583,-0.2578914165,0.2720677257,-0.0310348608,0.0491149388,-0.113248758,-0.3442519903,0.1562727243,-0.0086452216,-0.155209437,-0.2188283801,-0.2916173637,-0.4970089197,-0.011820077,0.2021970898,-0.363181293,-0.1355059594,0.1165074185,0.1536949873,-0.15706487,0.0029804937,0.1280414164,-0.1576476991,0.0278073754,0.1507251859,-0.015924396,0.0783389062,0.413304776,-0.2509163618,-0.0062365411,-0.3067879081,-0.1189336106,-0.3360534608,-0.1892143935,0.6777979136,-0.0741898045,-0.1459681988,-0.0839100629,-0.0895115659,0.2506249845,0.4015814364,-0.206268236,0.1616940051,-0.1717135608,0.1268459558,0.0643774047,-0.3153370917,0.2891644835,-0.135587275,-0.091763407,0.0690469444,-0.0290428698,0.1141546741,0.1650416702,0.2601776719,0.1370312124,0.1395323873,0.2431280762,0.8101244569,0.0562594905,-0.15360111,0.2286020368,0.1859832704,0.0851895288,-0.2001599669,-0.0640195757,-0.0191877503,-0.0938731357,-0.0172077306,0.1979723275,0.0642828047,-0.5166178346,-0.0707938001,-0.1951275319,-0.0425238237,-0.247399658,0.1434348971,-0.4774414897,0.4122231305,0.2807963192,0.3622750044,-0.5234687328,-0.4430350959,-0.0484388098,-0.1710771471,0.3679878414,-0.0597570054,-0.6432790756,0.1076242998,-0.6434779763,0.1648287177,-0.0803466737,0.6191594601,-0.061744716,-0.0851488039,-0.01580997,-0.0766576231,0.4049384296,-0.207270965,0.1418862492,0.1035571396,-0.2267690748,-0.0297382064,-0.173875019,-0.0404763557,0.1654497236,0.4327696264,0.7441766858,-0.1531309336,-0.1319165826,-0.574950695,-0.0251536313,0.082770139,-0.295710057,-0.0166319292,0.3887716234,-0.1977948248,0.1605420262,0.2259784639,-0.1305199713,0.1161974594,-0.1313422471,-0.2656112313,-0.1178572848,0.0973199457,0.0760862529,0.355240047,0.0089083137,0.2890256941,0.168001622,0.1620039344,-0.2283402085,-0.3112522662,0.1641118675,-0.2720577717,-0.0190150328,-0.0178305134,0.2943405509,0.1031521261,-0.0241001491,0.1051790789,-0.2099490911,0.0135326553,-0.3208743632,0.176351428,0.1901991814,0.2479486912,-0.3742703795,-0.3202723861,0.1891012043,0.1330007017,-0.150434643,0.194736138,-0.3314842284,-0.460759908,0.6001465321,0.2242605388,0.9640939236,0.1624002159,0.4029426873,-0.2921115756,0.207877636,0.5217230916,-0.0475481935,0.3225553334,-0.223012954,0.1038316563,-0.0360145085,-0.1235316247,-0.0341253355,0.3026357889,-0.3025872409,0.3806535304,0.0747208446,0.7538641691,0.0548072495,0.0850660279,0.1126422882,-0.2378841192,0.2222023457,0.1249258593,-0.0743058249,0.3012235165,-0.0182941742,-0.0443725213,0.0720108896,0.0631892309,-0.1468547583,0.1004685834,-0.1687465906,0.2136517763,0.1924710721,-0.0062141949,-0.0089089219,0.2136020064,0.3846445382,0.1958388537,-0.2509059906,0.0606473014,0.3297256231,0.2780996859,-0.0865225121,-0.2774796486,0.1973219514,0.035378553,0.0366462395,-0.1029066145,-0.0703406483,-0.3287983537,-0.4701607823,0.0011331926,0.2086943537,-0.0988461226,0.1384550333,0.0829358846,-0.1988267004,-0.1856831759,0.1304501295,0.1470264494,-0.0318883881,0.3176639676,-0.1083245799,-0.061388053,0.0287472382,0.4463638365,0.1126517504,-0.1061444879,0.6411355734,-0.4504158199,-0.3554027081,-0.1971056461,-0.1580515206,-0.2437925637,-0.3205919564,-0.0364055596,-0.2929611206,-0.1839295924,0.0279187337,0.1311285943,0.0982375965,0.0759622306,-0.0193501655,-0.2008481175,-0.4703054726,-0.1235761419,-0.0474192575,0.0260635894,-0.0510737896,0.2376328707,0.038884934,0.2088433206,-0.273247391,-0.2083369046,-0.0176674575,0.2203858793,-0.3480385542,-0.2401839197,-0.0420312956,-0.0459292345,0.0135918092,0.2336961925,-0.1944241077,-0.1440051645,-0.0415631905,0.1649190784,0.3418274224,-0.2668527663,-0.1786298752,0.0828090459,-0.0448026285,-0.509991467,-0.2129707336,0.2315692753,0.0281007905,0.0241041966,0.2009625584,0.3067676723,0.1727914959,-0.0417600945,0.0065668416,-0.0254179705,-0.1153657436,0.2973347008,0.0633657351,0.0603688657,-0.2951351702,0.1069441363,0.2014807314,-0.0188991353,0.4069460332,-0.1713302583,-0.120038785,0.1185930446,-0.1873949617,0.173545137,0.0442081913,-0.4994340539,0.769520998,0.1591219455,-0.1283281595,-0.0756162703,0.2212612629,0.1312406361,-0.0563526601,-0.0023197122,0.4812898338,-0.0320980586,0.0290491786,0.0776334107,0.5810437202,-0.3271012902,0.1836983711,-0.0958350822,0.1182298735,0.157428503,0.2900806665,0.078693904,0.0169539507,0.0191444661,0.3803720474,0.2586429417,0.1718867123,0.2382688075,-0.2620005012,-0.171695739,-0.1738492996,0.3552845418,-0.084082447,0.5323511958,-0.1053518206,0.1746243834,0.0578930676,-0.1648116708,-0.2164888829,0.3623829782,-0.2819154859,-0.3074509799,0.0138876755,-0.0396033563,-0.0742360353,0.1408207268,-0.1607384384,0.1386866421,-0.007048782,-0.0436915755,-0.0263910759,-0.0558442995,0.189499259,0.0098624155,0.3851006031,-0.2825767994,0.3480746746,-0.431471169,-0.0469177738,-0.2241779566,0.2256393135,0.2686682343,0.365216881,-0.3265911937,0.1069626957,-0.0359644778,0.0288718548,-0.1814087629,0.6722570658,0.0512520373,-0.5190725923,0.0183791816,-0.0541453622,-0.0261780713,0.0302471537,0.1335176378,-0.0765955597,0.1631425917,0.5349285603,-0.0189138427,-0.248439908,0.2411663085,0.0832035318,-0.3054163456,0.0202840511,0.1891672313,-0.0963486657,0.255572468,-0.2167486697,0.0756774172,-0.1731589437,0.3869381249,0.4390624166,-0.0682906508,-0.2560794055,-0.1151082814,-0.4415774345,0.4227825105,-0.1082160249,0.3445723355,-0.2482795119,0.1458862722,-0.188789323,-0.0780697688,-0.1451387703,0.0555390567,-0.2978533804,0.0803246871,-0.200373292,0.397572875,-0.1888707429,0.3042846024,0.1004484817,-0.2827378809,0.3277181983,-0.0082625421,0.1178579703,-0.0830326155,0.0254823733,-0.0151369954,0.3258376718,0.4636445343,0.2416368574,0.1604121476,0.0837587565,0.1618284732,-0.3611883223,-0.0951479077,-0.0798974782,-0.0534456819,-0.0638958067,0.1235970333,-0.3203250766,0.1852367073,-0.090777427,-0.0262238327,0.1159882322,-0.4581905901,-0.0087258331,-0.0698096007,-0.1827688515,-0.1153420955,0.0035540191,0.3067379594,-0.0786314681,0.2184354663,-0.3980624974,-0.2383288145,0.2559683323,-0.6370259523,-0.4952368736,-0.1327913404,0.1361172795,0.4375092387,0.05099345,-0.2765730321,-0.2089773118,0.358428359,0.0507428683,-0.1939691901,-0.0110155409,0.0950631723,-0.1305749118,-0.3030377328,0.426474303,-0.0248533692,-0.2435499877,0.0642276853,-0.0556413867]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/492","title":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema","comments":"In 0.4.0, the assertion in `concatenate_datasets ` is on the features, and not the schema.\r\nCould you try to update `nlp` ?\r\n\r\nAlso, since 0.4.0, you can use `dset_wikipedia.cast_(dset_books.features)` to avoid the schema cast hack.","body":"Here's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n","comment_length":35,"text":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema\nHere's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n\nIn 0.4.0, the assertion in `concatenate_datasets ` is on the features, and not the schema.\r\nCould you try to update `nlp` ?\r\n\r\nAlso, since 0.4.0, you can use `dset_wikipedia.cast_(dset_books.features)` to avoid the schema cast hack.","embeddings":[-0.0643963218,0.11732544,0.0268085655,0.1252689362,0.1567067057,-0.0773083195,0.3404867053,0.1692478806,-0.3219616711,-0.3533984423,0.0765748993,0.3673611879,-0.0683538541,0.3275818527,0.0926476717,-0.0278011616,0.20226641,0.2302506864,0.0969581977,-0.0026237783,-0.0930334851,0.1811446249,-0.3965626359,0.0402221829,-0.5902644396,0.2065486014,-0.0333003029,-0.0195062682,-0.1387539655,-0.5088027716,0.344329983,0.20197393,-0.1002951264,0.3447961807,-0.0001246,-0.1214664951,0.5814360976,-0.1061937883,-0.1357245892,-0.1315975189,-0.1061730608,-0.334274441,0.1451298594,-0.184059307,0.0441182479,-0.1426272988,-0.0075730518,0.3041886091,-0.5539480448,0.2694932222,0.0474009812,0.5145810246,0.4430456161,0.2505997121,0.3381246626,0.0039449814,-0.1633378565,0.0774524137,-0.2217185795,-0.2308957428,0.2946070433,-0.004089464,-0.3032649755,-0.2862060964,0.189220652,0.0053237276,0.036097832,-0.1486219764,-0.1327947378,0.4225979745,0.4490500093,-0.2722763419,-0.3485791087,-0.2052143365,0.0249234177,-0.2226937264,0.3266037405,0.4107818007,0.0495450981,0.0445891209,-0.4404838681,0.1101490781,-0.1068293229,0.2732964456,-0.3308849335,0.3821586967,0.423401922,0.1096609756,0.0128080724,-0.1292209476,0.2586248517,-0.3746066988,-0.0194233023,0.1626722962,-0.2908125818,-0.1841994226,-0.22538656,0.0221132934,0.430544436,-0.0290179849,-0.1455563158,0.0746988207,-0.1401955932,0.1076272652,0.1945328861,0.2544096708,0.2239787877,-0.1042137668,0.0038693603,0.076795876,-0.2166543156,0.2659810185,0.2274273485,-0.2757505774,0.0893795118,-0.0405187644,0.5651307702,-0.0694724619,-0.3042345047,0.0565479025,-0.7364634871,-0.1235802546,0.102914609,-0.0402498618,-0.0763520151,0.4491232932,0.0586440302,0.3635566235,-0.197696045,-0.3798347414,-0.0722624809,0.0145610329,0.016205024,0.06115219,0.1437865645,0.0790431872,0.1670461595,-0.0208609458,-0.4102073908,0.205223456,-0.2113888413,-0.1676452011,-0.1489060223,0.0289777331,-0.2191667557,-0.3602612317,0.0156736393,-0.3502401114,-0.2368439138,0.153339684,-0.0034028466,-0.3047691882,-0.2618936896,0.032093484,-0.3846156001,-0.1074053645,-0.0526561029,0.3227862716,0.2190195769,-0.0189926382,0.0711328015,-0.1152549013,-0.0018663049,-0.2445031852,0.1468636692,0.1550917625,-0.0509334356,-0.0164897721,0.1121340021,-0.1654553711,-0.0499183312,0.2630227506,-0.0839521363,0.0946981981,-0.1670563072,0.4398614168,0.3568460941,-0.0073680687,-0.0031482261,0.0458445586,0.2689749002,0.0852905959,0.1107114032,0.0083126724,-0.3739365041,0.0349341445,-0.0445479415,0.3406242728,-0.0643275231,-0.1324491948,-0.200401634,-0.329082936,0.6721149087,-0.1642895341,-0.0742964223,0.1388433725,-0.0933730677,0.2371647358,0.3771465123,-0.2393366247,-0.1351750642,-0.0320394002,0.2976107299,0.2248855978,-0.3573701382,-0.3245314062,-0.6215187907,0.1153126806,-0.2744531631,0.5649373531,-0.0713170171,-0.3467866778,-0.0226434506,0.1168041304,0.1842118055,0.0491243191,0.0502412207,0.1320669651,0.003475416,0.2040457278,-0.017224595,-0.0341619737,-0.0021852811,0.0743370801,-0.1543972343,0.5079219341,-0.050801307,-0.2447683811,0.0116013149,0.5782472491,0.2310594618,0.0018774219,0.0209860038,0.2861755192,-0.0210095346,0.0169910416,-0.1841734499,0.3597716987,0.0103445137,-0.108213231,-0.1197339594,0.2958289087,0.0824739859,-0.0362321101,0.106977649,0.5403357744,0.0890617818,-0.1274826825,0.2386863232,-0.0310108252,0.0641216114,-0.1456842124,-0.0917195156,-0.0687631965,0.1033114865,-0.029711565,-0.3782554865,0.4686181545,-0.5818831325,-0.1038340107,0.4880428314,-0.1803923398,0.2012025714,-0.1890496463,-0.1980809718,0.0216359086,0.3691531718,-0.36863482,0.1642698497,0.1738227159,0.0248235594,0.0485812575,-0.3041326702,0.0645630807,0.3759593666,-0.0363258719,0.1276257038,-0.0101232845,0.2944433391,-0.2063151449,-0.3622223139,0.1404734701,0.1248603091,-0.0001578323,-0.3340364397,-0.00883877,-0.3641410172,-0.0675907582,-0.4399199188,-0.2286157608,-0.5033985972,-0.4318349063,-0.148087725,0.7034394741,0.0043244301,0.2784733474,-0.6784133911,-0.2246820331,0.0438677296,-0.2637732327,-0.0135059636,0.0479043722,-0.1932651848,-0.0891905725,0.2925310135,0.1887551546,-0.2067164034,0.0206944756,-0.1609638333,-0.2025876343,-0.7249019146,0.0783891156,-0.3375943601,0.0019903327,0.2957773507,-0.0846194029,0.1590904295,-0.3467276096,0.1459008306,0.1718693674,-0.4136032164,0.4664402902,0.0333243459,-0.2877027988,-0.008931377,-0.2185085714,-0.0588836856,-0.2434386015,0.23347947,-0.0135711432,0.0751471817,0.4313396513,0.0159301758,0.0507930443,-0.1473100036,-0.0440619811,0.0670361146,0.4366896451,0.4299979508,0.2493715286,-0.1570455581,0.1447858959,-0.2302629799,-0.1630174965,0.2447808534,0.0009895115,-0.1102875695,-0.1580325514,0.5572417378,0.2054716796,-0.5082899332,0.3505128622,0.418728143,0.0752615407,0.0139181046,-0.232442379,-0.0090982756,0.4431093037,0.2931427658,0.0866401121,-0.1674508154,0.1715910435,0.052488599,-0.0354896784,-0.2688117921,0.5545909405,-0.1980309933,-0.0067355544,0.1651455164,-0.0011852155,-0.1469272524,0.0155081609,0.2795686722,0.1951107532,-0.24524647,-0.2343912721,0.1371358037,0.1229925752,-0.1862457395,-0.1444648355,0.1335587054,-0.3576019406,0.3650502264,-0.0697130859,-0.1677846164,-0.1720569581,-0.424606353,-0.024985997,0.1461025923,-0.0582427159,0.1105497256,-0.6636128426,-0.051517196,-0.0062472653,0.3996485174,0.1919797361,0.5644320846,0.2260040641,-0.4211959243,-0.1715785861,-0.1437278986,0.544991076,0.0099101411,-0.1428054422,0.2944712639,0.2399410009,-0.2892708182,-0.2785446346,0.0273507927,0.1203591451,0.356913805,0.2400013506,-0.5941352844,-0.2463476658,0.3794517517,0.2820280194,0.030324284,-0.0314473994,-0.1235699877,0.0520266034,0.0594212487,-0.1702728271,-0.0757285058,0.3546293676,-0.3836919963,0.0351629667,-0.3101992011,-0.0964160785,0.1639787704,0.3305377066,0.3377700746,-0.3164930344,0.0744247511,-0.3892452419,0.1455589831,0.1262887865,0.3810679913,0.3967331052,-0.4478458166,-0.4655646682,0.1159033403,0.3578379154,-0.027673535,0.1747054011,0.2886177003,-0.2007166296,-0.2645499706,-0.198375836,-0.028964065,0.4133416414,-0.0997605249,-0.1146218702,-0.064942196,0.5096235275,0.0212690458,0.1219204143,-0.588642776,0.2646948993,-0.4755286276,0.2535638213,-0.1144695878,0.9977645278,0.1156836897,0.1002285033,0.2691831291,0.1657368243,0.5099569559,0.0075562834,0.1577658355,-0.3320664167,0.2349310815,-0.1403216273,-0.203341946,0.1893981546,0.215536043,-0.284773469,0.2099850625,-0.0445333906,0.0899679959,-0.0312904008,0.2504956722,0.0605234243,0.2865279913,-0.3072863221,-0.0948877707,-0.0923763365,0.0832219347,0.0634090975,-0.2093326747,-0.3941881061,-0.1552625,0.0115259755,0.1810256094,0.0728999227,0.0601830669,0.4171114862,-0.2120634764,-0.0318059288,0.2774350047,-0.0092717484,-0.2507450879,-0.0398266986,-0.0581834465,0.2255801558,-0.1509781033,0.1982783228,-0.115557462,0.2459988892,0.1325130016,-0.3227518797,-0.1154311076,-0.1396712661,-0.1983374506,-0.1088741422,-0.1120207906,0.0315249898,-0.1432553381,-0.1874484718,0.0321699344,-0.0815727264,-0.0831200778,-0.0252199378,-0.2151886821,-0.0778528526,0.0383991562,0.0380341299,-0.1525835693,0.0448343791,0.241837129,0.2376711965,0.5448002219,0.4637742937,0.0376261435,-0.0040663434,0.0038768693,-0.2874631584,-0.2303058803,-0.3807948828,-0.1140156016,-0.0650251135,-0.1488628983,-0.1858001947,0.6646890044,0.0446055941,0.2915352881,0.0867566094,-0.154698506,-0.0842443556,0.3814484477,0.2085337341,0.2213026583,0.0042428169,0.5544479489,0.1517053396,0.0485914201,-0.1127221435,-0.0973064452,-0.3818844557,0.2463040203,0.1525021493,-0.0737936944,-0.1864910573,-0.3309208155,0.0815941542,-0.2978520691,0.1310171187,0.0108479671,-0.0611109957,0.2071423084,0.3853248656,0.0462933555,0.0434727594,0.0189264417,0.3019472063,-0.1228420511,-0.1633669585,0.1167983934,-0.1509326696,0.3340252042,0.3067085445,0.26797387,-0.389914453,0.0454359986,-0.1173832342,-0.0730447322,-0.0132277273,0.340855509,0.0195690617,0.1627985835,-0.0384200849,0.1229011863,-0.0323857851,0.0929692164,-0.0094262743,-0.1149352044,-0.332537204,-0.0618076138,-0.019270204,0.4199446738,-0.0150182582,-0.214532584,0.3441825211,0.0217387788,-0.4045166671,-0.056325864,0.1963019222,-0.327665627,0.0297584385,0.1409677118,0.4163489044,0.2480724603,-0.3645541966,0.1151564494,0.0950261354,0.0422360264,0.2047935277,0.2129473239,-0.015564491,0.1771935821,0.4805328548,-0.0208507907,0.0870032236,-0.1591030508,0.4040462375,0.1797994226,0.0218498986,-0.185967505,0.4318914711,0.1254309267,-0.3001895249,0.1034144759,0.4830498695,0.0874408111,0.1706008017,0.3680459261,0.0489859022,-0.2632360756,-0.1646349132,0.0922513828,-0.1182093248,-0.127955839,-0.2326815724,-0.2687022686,-0.126254946,-0.1826531142,0.0351066887,-0.1847102791,0.5446308255,0.0857566223,-0.0871325284,-0.3670359552,0.3395572603,0.2477977723,0.1671564728,-0.2275745571,0.0550816655,0.0380385481,-0.2480456978,0.2264219075,0.3592941165,0.2917849123,0.3263564706,-0.3158448637,-0.3911439776,0.0193011519,0.029535139,0.0963333771,0.0056228219,-0.0832962021,0.1238268614,0.1970171779,-0.0187276714,0.085687764,-0.0020505295,-0.0753250346,-0.2528958023,-0.3066676557,0.5929834843,-0.0182348024,-0.3518256843,0.0388039872,0.0725442916,-0.0117436759,-0.1955383271,0.1793350875,0.2753736377,0.2146021873,0.0860147998,0.0115738651,0.1936455965,0.1590189338,0.0958808139,0.1392926574,-0.0169128831,0.1692793518,-0.4543947279,0.0249606818,-0.0521974824,-0.210227102,0.0804067701,0.3008460999,0.0454501212,0.0251691341,-0.3979474306,-0.0292960349,0.2426876724,-0.0250514969,-0.2467198968,-0.1098271757,0.2442295402,-0.3090698123,-0.2742753625,-0.3003704548,0.3783368468,-0.2326954007,-0.0584372431,-0.0080521535,0.1860939562,0.4821601212,0.309761703,-0.0155441482,-0.00046733,0.724096477,0.0777884498,-0.0391623266,-0.276391387,-0.1431502402,-0.4282419086,0.2060937583,-0.2975790203,0.220962882,-0.077944845,0.5272229314,0.0625877082,0.5148112178,-0.1245449558,-0.2223350555,0.0077806949,0.2017616332,-0.3749049902,0.0519046485,0.1074221432,0.0230159648,-0.0262647849,0.0595195182,-0.0937539935,-0.4693756998,0.1555227488,-0.1344616413,-0.2337852418,0.02665109,-0.0185008477,0.1400673985,0.0730230659,-0.4839622974,-0.2024584115,0.3920608163,0.0701854825,0.0771596134,0.16482535,0.1535686702,-0.1244249418,-0.1648193896,-0.0260827411,-0.2262390852,-0.0990438834,-0.0559796952,-0.0403773747]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/492","title":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema","comments":"Or maybe the assertion comes from elsewhere ?","body":"Here's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n","comment_length":8,"text":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema\nHere's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n\nOr maybe the assertion comes from elsewhere ?","embeddings":[-0.004785981,-0.0092260875,0.0121712452,0.1523435563,0.1799283624,-0.0479315966,0.3413929641,0.0989464149,-0.2621549368,-0.3168407679,0.1227191687,0.2497010231,-0.0509567969,0.3013101518,0.0487219021,-0.0173008684,0.1966295987,0.2346923649,0.125324741,-0.0087767495,-0.017577067,0.1620617807,-0.3953854442,0.0968917161,-0.6276301742,0.2252174616,0.0908359438,-0.0662148446,-0.1551104933,-0.4570470154,0.3430460393,0.1146859154,-0.1609416604,0.3489036262,-0.000124564,-0.0775955617,0.4647088647,-0.1066646278,-0.0433508568,-0.098645471,-0.0813707784,-0.2661244869,0.1024675295,-0.1905415505,0.0551229641,0.0045433231,-0.0285391603,0.2679912448,-0.5301090479,0.2436398268,0.0364935137,0.6262816787,0.290717274,0.2483019233,0.4621179402,0.0874625668,-0.1938254684,0.0564850606,-0.1900108159,-0.2328702807,0.3879126012,-0.0413332582,-0.3467579186,-0.222642526,0.1681652963,0.0057512512,-0.0175436046,-0.1376878768,-0.2156186849,0.445777297,0.392876178,-0.2787118554,-0.3270686269,-0.2222522944,-0.0580471568,-0.1167403534,0.3168330491,0.4224161506,0.0072789425,0.0542970747,-0.3891451955,0.12580809,-0.0619560592,0.1554920524,-0.3865788281,0.3180312216,0.4267664552,0.1240807474,-0.0740206838,-0.0169427134,0.2130088359,-0.3554985225,0.1043031514,0.1468765587,-0.2892206907,-0.2219942063,-0.2340594381,-0.1161567494,0.4582234025,-0.0947525874,-0.061350137,0.0553176254,-0.1703644544,0.1864726543,0.1643102914,0.2252327353,0.142478466,-0.072898604,0.0196998604,0.0247821324,-0.3057262897,0.2383477986,0.2756439149,-0.2755093575,0.2168242484,-0.0460068844,0.5883056521,-0.1506343186,-0.3742639124,0.0914605632,-0.7003248334,-0.1180746034,0.1518756598,-0.0718429387,-0.0280574001,0.3667907715,-0.0813509226,0.3777570128,-0.210306555,-0.3545616865,-0.0953712314,0.0757203773,0.0505279899,0.0076210778,0.1007766426,0.0981725901,0.1554510444,-0.0018674864,-0.3395426571,0.1698251814,-0.101845637,-0.175190255,-0.0627804399,0.073054269,-0.2111307234,-0.3351984024,0.0902682692,-0.3248077035,-0.2701268494,0.1370531172,0.0182029307,-0.2734075785,-0.0920910686,0.0405926481,-0.459885329,-0.0727035478,-0.0395574123,0.2385687083,0.1865880936,-0.024166109,0.1401028037,-0.1149866283,-0.0109473672,-0.2565029263,0.1130178124,0.1123024672,-0.0841017142,0.0130253071,0.1801306307,-0.2737558186,-0.0878988057,0.2781911492,-0.0608813614,0.0718444437,-0.2008141577,0.4981017113,0.3322263062,0.0139211817,0.052833572,0.0429227054,0.2500374317,0.0773393214,0.134837091,0.0704396367,-0.3074423075,-0.0345277004,-0.1199408099,0.3504836559,-0.1293555945,-0.1255274415,-0.1489067823,-0.2664929032,0.6161352992,-0.1371898502,-0.0598265007,0.1151808053,-0.0845592171,0.2856226563,0.2711490393,-0.2172311693,-0.2185028046,-0.0149977347,0.2235634774,0.2422384024,-0.4363441765,-0.407987088,-0.5287966132,0.1642406285,-0.3140405118,0.5425207615,-0.0279314518,-0.3887187839,-0.1093801335,0.1772449017,0.2198752165,0.1401134729,0.0454144552,0.1312163025,-0.0008841763,0.2227061987,-0.0229123961,-0.1230301484,-0.0032845323,0.1289856583,-0.1605000198,0.5008702874,0.0087863021,-0.2006728351,-0.0149451792,0.5367492437,0.2159677297,0.0247019697,0.0529624745,0.2026316971,0.0284688938,0.0302868243,-0.1701388955,0.3373049498,-0.003091604,-0.1313240826,-0.1549595743,0.3086469173,0.1379088163,-0.0865259171,0.0398425534,0.5201992393,0.0603611767,-0.1567438543,0.2007970214,-0.0808709487,0.0061761569,-0.084780924,-0.1095373556,-0.0887270793,0.0949286371,0.0480127633,-0.5272288918,0.5609241724,-0.5941967368,-0.2167778015,0.4718075395,-0.2298101187,0.2134331465,-0.2478556037,-0.1618185043,0.0901276916,0.3462010324,-0.4085990787,0.1292859316,0.0920234025,0.0572300702,-0.0328137688,-0.3276786208,0.0914897025,0.3103349507,0.0041958848,0.0859684274,-0.0549475551,0.2178253531,-0.1422345489,-0.3497721553,0.1383843571,0.1406942755,-0.0966740847,-0.339165777,-0.0109021785,-0.312792033,-0.0769862011,-0.4011210203,-0.2129554898,-0.4600875378,-0.4381008446,-0.0991601422,0.6787856817,0.008392808,0.219869554,-0.6687480807,-0.2164129317,0.0293286089,-0.410890013,0.009428856,0.0023178034,-0.077090919,-0.0846884847,0.2842701674,0.2997267246,-0.1745077372,0.0016133441,-0.1770061255,-0.2176718414,-0.7185379863,0.1118253991,-0.3853941262,0.0592145249,0.3181407154,-0.0285103638,0.2587691844,-0.236976862,0.1502224505,0.1798454523,-0.4181735218,0.4772998691,0.0948120207,-0.3267707527,0.0040497603,-0.1232685223,-0.0183731653,-0.2264425308,0.1247773096,-0.0779058337,0.0419338979,0.3225084841,-0.0159120616,0.0139257722,-0.1609014124,-0.0137895253,0.104732275,0.3885356188,0.4683499634,0.2886269391,-0.0811112374,0.2098567784,-0.2750849426,-0.2032036781,0.2414142489,0.0098640742,-0.1784251481,-0.2571187615,0.5445712805,0.1864225417,-0.5103509426,0.3068909347,0.3403405249,0.093356289,0.0250833351,-0.2474646419,-0.0673394874,0.4448437393,0.2569540739,0.0509117804,-0.1634841263,0.1154749468,0.0934367329,-0.1198005825,-0.2392062545,0.5755367875,-0.2621307075,-0.0892771035,0.2125920802,0.0303992331,-0.1617897004,0.0345983878,0.1091296673,0.1588628292,-0.2512073815,-0.1643459797,0.1476332694,0.1202981174,-0.1716205031,-0.1736437678,0.1648007184,-0.3297954202,0.3428836763,-0.0809262097,-0.1796829402,-0.1892062575,-0.3381567895,-0.0080887703,0.1572022587,0.0161987673,0.1484356672,-0.6450247169,-0.0467488952,0.1192167103,0.3252415359,0.1660017818,0.4992073476,0.1365662217,-0.4519253075,-0.1594451964,-0.0761411488,0.5156096816,0.0471197441,-0.0622665659,0.206189394,0.1724051237,-0.277086705,-0.2917996347,-0.0324467868,0.0901252553,0.2900862098,0.265006125,-0.5857691169,-0.1961087584,0.3191227019,0.346391201,0.0433636978,-0.0548148602,-0.2070434541,0.0315168947,0.0440522768,-0.0993405208,0.0390652753,0.2321897149,-0.3441726863,-0.0220627319,-0.3575501442,-0.0682849437,0.2613855898,0.4202063084,0.2859714627,-0.3918973207,0.0590169355,-0.4968430698,0.1250277609,-0.0362324417,0.4893501103,0.491712451,-0.4919506013,-0.4625247717,0.1317973584,0.3717431724,-0.0450255275,0.1595862955,0.2731501758,-0.2304660231,-0.2730473578,-0.210694775,0.0200168807,0.3701804876,-0.0943464339,-0.1062540337,0.0066075739,0.529790163,0.0870433226,0.0882399902,-0.674552083,0.3732399642,-0.360951215,0.3093167841,-0.0302860681,0.9489668012,0.0464783944,0.0924320444,0.3146071434,0.1836227775,0.4811191857,0.0439551994,0.1295956969,-0.300341934,0.1218733788,-0.1950270981,-0.2801047266,0.1347260922,0.2211033404,-0.197102949,0.2304278761,0.0054383543,0.1858121604,-0.0063045416,0.1834717989,-0.0349599347,0.3689017594,-0.2534586191,-0.0905286297,-0.096865125,0.1494576782,0.1160180047,-0.1693805903,-0.4660926163,-0.1161099002,0.0283738617,0.1219599694,0.0840101317,-0.0038335077,0.4262957871,-0.2215324789,-0.0139567228,0.3097035587,0.0235393401,-0.2851578891,-0.0016903951,-0.0676000267,0.2272352129,-0.1379612386,0.1971857995,-0.1120719984,0.2513670027,0.1422253102,-0.2862040102,-0.1044700518,-0.1607238799,-0.1574973315,-0.1422558725,-0.10121499,0.032324072,-0.2076862901,-0.2260582447,-0.0701909661,-0.1010112166,-0.0630556419,-0.0440445878,-0.2107200474,-0.076437369,0.0926047191,0.0261376929,-0.1754809022,0.0728175938,0.097693406,0.1955206245,0.6283993125,0.4137364924,-0.0223704632,0.0191647634,0.0118647823,-0.2910834253,-0.2685158253,-0.3629368544,-0.0984063521,-0.1037652269,-0.2437645048,-0.2447677553,0.6959140897,0.144372642,0.2615063488,0.1239740476,-0.201725781,-0.0184849072,0.3531692326,0.233653754,0.2870007753,-0.0163041912,0.5312854052,0.131166175,-0.0030142094,-0.1023904756,-0.1204623878,-0.4402192533,0.2519258857,0.1189312935,0.0077896211,-0.2369001806,-0.3578852415,0.0864674151,-0.2533874214,0.1663862169,0.010761993,-0.0199149251,0.2223003209,0.3242277503,0.148856625,0.0614837408,0.1104591638,0.3345174193,-0.0988324285,-0.0945358276,0.153929323,-0.118721731,0.3594181538,0.2521878481,0.2769091725,-0.4655179083,0.057854943,-0.1198254526,-0.0794064999,0.0536935627,0.3234640956,0.0181332752,0.1620722562,-0.0243072398,0.1169343218,-0.0132531896,0.0521303155,-0.0634211078,-0.0673672482,-0.2948097587,0.024152929,-0.042577561,0.4346228838,-0.0470379032,-0.1739727706,0.3148883283,0.0339301936,-0.3890341222,0.0009835833,0.2133665979,-0.3481936157,0.0617878288,0.1933296323,0.4035304189,0.3520259261,-0.3485837281,0.0741624981,0.1290104538,0.1599664241,0.1441977173,0.2398770154,-0.0688115805,0.1877918243,0.4888006151,-0.0069864267,0.1112007573,-0.1352553368,0.436930865,0.2510242462,-0.0807874948,-0.1815100759,0.4903355837,0.1549818963,-0.3249863088,0.0498898439,0.5133996606,0.0654169247,0.1919138134,0.3587975204,0.0410558134,-0.2260779738,-0.2229132503,0.0586596318,-0.0895294622,-0.0686720163,-0.2873828709,-0.2645567656,-0.0657415166,-0.1849382967,0.0535662659,-0.1890936345,0.5253466368,0.0919665843,-0.0839039758,-0.2667464018,0.3988182545,0.2174451053,0.168542847,-0.21222651,0.017883338,0.0710401759,-0.2531982958,0.3045856655,0.2596724331,0.2583881617,0.2523325682,-0.3353277147,-0.3877333999,-0.0495089106,-0.0123766325,0.0955541134,-0.0399626568,-0.2151383013,0.1655564457,0.2279090881,0.0082411934,0.1072810963,-0.0226891693,-0.208313033,-0.3251972198,-0.3396722674,0.5955364108,-0.0267016161,-0.2827983499,-0.0416539125,0.012402988,-0.0129786804,-0.2670300305,0.1460541785,0.2412270904,0.2286090553,0.1157018915,-0.0063614291,0.1336998641,0.1090112776,0.1165920869,0.1160320938,-0.0008446695,0.2811250985,-0.558054924,0.0459786393,-0.0577492304,-0.3217475414,0.1343531907,0.3035083711,0.0490537398,0.091162473,-0.3798691332,0.0191324241,0.1508483738,-0.000143539,-0.132341221,-0.147115618,0.0964461938,-0.3864599168,-0.2627547681,-0.2521406114,0.3807266057,-0.2499854267,-0.0253344346,-0.0081773419,0.1695270538,0.4465702772,0.2856822312,-0.004552376,0.0142943505,0.7054159641,0.1145624146,0.0526766665,-0.1354610622,-0.0539685413,-0.4079678953,0.1682770997,-0.3026760817,0.2241445929,-0.020686971,0.6440569162,0.056498874,0.4612848759,-0.1040260121,-0.3525039852,0.0618606843,0.1703175008,-0.3997608125,0.1088516638,0.1823960245,0.0300158467,0.0273917112,0.0624746569,-0.045820564,-0.493575424,0.1477240622,-0.0513203144,-0.2195354104,0.0968590081,-0.0446412675,0.1350383908,0.0760717094,-0.4458818734,-0.1953661144,0.3465432227,0.0500855185,0.1546519995,0.1114577204,0.2195537388,-0.0848818794,-0.1300525069,-0.03994238,-0.2887550294,-0.0059880367,-0.0547763109,-0.0086849872]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/492","title":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema","comments":"I'm using the master branch. The assertion failure comes from the underlying `pa.concat_tables()`, which is in the pyarrow package. That method does check schemas.\r\n\r\nSince `features.type` does not contain information about nullable vs non-nullable features, the `cast_()` method won't resolve the schema mismatch. There is information in a schema which is not stored in features.","body":"Here's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n","comment_length":55,"text":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema\nHere's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n\nI'm using the master branch. The assertion failure comes from the underlying `pa.concat_tables()`, which is in the pyarrow package. That method does check schemas.\r\n\r\nSince `features.type` does not contain information about nullable vs non-nullable features, the `cast_()` method won't resolve the schema mismatch. There is information in a schema which is not stored in features.","embeddings":[-0.0000325116,0.0538676195,0.0309730265,0.1527771801,0.1315416694,-0.0097018266,0.2905566692,0.1126790717,-0.424089551,-0.3221019804,0.1324172914,0.3009293675,0.0746788085,0.2184081674,0.0105956672,0.0620799325,0.186666593,0.1985082179,0.1221262068,0.0144806318,-0.032848645,0.1648804247,-0.3214719892,0.0625458732,-0.619639039,0.2229527831,0.0357409976,-0.01967825,-0.1147865579,-0.4618887305,0.3706979454,0.1314748824,-0.1188881695,0.358096391,-0.0001218364,-0.0039295494,0.4484198689,-0.1272719949,-0.0842170715,-0.0493644662,-0.1476014405,-0.336984098,0.0708868057,-0.1506245583,0.0678572729,-0.0325273313,-0.0610251911,0.3406130075,-0.4604620636,0.1814389974,0.0470485501,0.5140041113,0.4450060427,0.2114942372,0.3964324296,-0.0114495941,-0.1496786773,0.1629143655,-0.2344385535,-0.1761411428,0.3071684539,-0.0447370484,-0.3355340362,-0.246884793,0.09197741,-0.0022425421,0.0193526782,-0.0927106887,-0.2154493332,0.401458621,0.325085789,-0.3767760992,-0.3620739281,-0.2539633811,-0.0721921921,-0.0025720007,0.3222019374,0.4332791269,-0.0434090719,0.0893690437,-0.3906940818,0.1698400378,-0.0055579445,0.0199997053,-0.375679642,0.2791659832,0.4180589616,0.125666514,-0.0578124225,-0.0325820111,0.2527773976,-0.3134332597,0.0239158869,0.0500919782,-0.4031199217,-0.1911042482,-0.1240607724,-0.1303989887,0.4479844868,-0.0571687929,0.0196361281,0.0487373024,-0.0874165744,0.1509540677,0.1608293653,0.2469576597,0.0690925047,-0.0032205575,-0.049184639,0.0776771754,-0.2310604602,0.267149806,0.2664994001,-0.272282511,0.1462367624,0.0156127978,0.5298162103,-0.19427827,-0.3017012179,0.1567925513,-0.7380691767,-0.1070708185,0.0951361805,-0.0827222392,-0.0859725624,0.4375620484,0.023168711,0.375118494,-0.1258000582,-0.3113953471,-0.0775917768,-0.0183433406,0.0185442045,0.0074304799,0.1500350684,0.1472896188,0.1175246388,-0.005865301,-0.3892932236,0.200388968,-0.2125387937,-0.1689075083,-0.052931156,0.1127571091,-0.1986382008,-0.3424559534,0.0748593658,-0.269253701,-0.2956694961,0.148986876,-0.0504913032,-0.2925234139,-0.0487848744,0.0491825826,-0.4548963606,-0.0138572296,-0.0014940753,0.1368769258,0.1710840911,-0.0737833232,0.1213525906,-0.1454526484,0.0488345399,-0.3303678334,0.227491051,0.0074431039,-0.1383932233,0.0149259549,0.2834969163,-0.2184307724,-0.0481670797,0.285407573,0.008305016,0.1808186024,-0.170597896,0.5086686015,0.2644163072,0.0803474262,-0.030159615,-0.0841616765,0.2482199371,0.1521852762,0.1921499074,0.0378289148,-0.4064255059,-0.0462312587,-0.013871355,0.3398268223,-0.1373070627,-0.1272159517,-0.2157674134,-0.2559677958,0.5326585174,-0.0319034196,-0.1721978933,0.0755647495,-0.0281263273,0.2179979235,0.2779777348,-0.225631088,-0.2787723243,0.0033528744,0.1888370514,0.1972596943,-0.3089707792,-0.2877030969,-0.4605284631,0.1879185587,-0.3435061276,0.5599398017,-0.2155990154,-0.3793883324,-0.0215970427,0.2495394945,0.2796551585,0.1542746723,0.0767219663,0.18392919,0.1065375805,0.1507077813,0.0433381014,-0.0635325313,-0.0650913715,0.0893567875,-0.2372569144,0.4350220263,-0.0643317178,-0.2545842826,-0.0205403138,0.5571249723,0.1251193434,0.0465454124,0.0292412508,0.2860116959,-0.0419971608,-0.0050192378,-0.200878188,0.2961167693,-0.1280407906,-0.1656190306,-0.2111010551,0.214900583,0.0744986534,-0.0052277958,0.1057888046,0.487539202,0.0662074462,-0.0467140749,0.2157475501,-0.1068633497,-0.0017402008,-0.0956156328,-0.0844757855,0.0191746522,0.2111942619,0.2053700238,-0.5470693111,0.5216994286,-0.4695622623,-0.0886455998,0.5151802301,-0.2779281735,0.228165105,-0.21243155,-0.1346039325,0.1312163323,0.398961544,-0.3695643544,0.1940737516,0.1023910046,0.0288279541,0.0448633432,-0.3958705366,0.0651666224,0.2654562593,0.0423330478,0.038110882,0.0511463545,0.2779502571,-0.1112808436,-0.2664991319,0.1818794906,0.0312794484,-0.1044894457,-0.3702565134,0.0883166417,-0.36469993,-0.0115047209,-0.3589561582,-0.2359383851,-0.5129183531,-0.3944704533,-0.0359419174,0.5990917683,-0.050605949,0.252838254,-0.6956558228,-0.2696458697,0.0114534106,-0.4257436395,0.051117897,-0.0740268081,-0.1123211086,-0.0595152602,0.2943375111,0.1861018986,-0.2042576969,-0.1255380511,-0.1582054794,-0.1919829845,-0.6355778575,0.1859968156,-0.4187495708,0.0109838648,0.3473017812,-0.1060544774,0.1912702024,-0.3261114657,0.1868243366,0.1711120307,-0.4117400348,0.5654469728,0.0669071302,-0.3507688344,0.0160049591,-0.1678054184,-0.0840316862,-0.2349204272,0.2703926265,-0.0993289128,0.0352934487,0.2802296281,0.0138519639,0.0089081423,-0.2474102229,0.1114722937,0.0248692911,0.390065223,0.3563617468,0.3765396476,-0.1372018456,0.1883707792,-0.3040184677,-0.1588795632,0.2909474075,-0.0703083426,-0.1315319836,-0.1706431359,0.6069602966,0.1364942491,-0.4531551898,0.3239976764,0.4185802341,0.0673283488,0.0007435285,-0.1860945672,-0.0540754125,0.4165595174,0.2537947297,0.0350470841,-0.1258539259,0.1467141807,0.0590124428,-0.0642360002,-0.2257235646,0.5036706328,-0.2651816607,0.0297761355,0.1756279767,0.0661862716,-0.1662025601,-0.017792549,0.0821510106,0.1531198621,-0.2464073449,-0.1709888428,0.0917296633,0.106661886,-0.1533644497,-0.2786035836,0.2364925593,-0.2413656712,0.3538047373,-0.0401275381,-0.1589080989,-0.1880382299,-0.3680251539,-0.0109519707,0.0784580708,-0.0820757076,0.0958490893,-0.6450893879,-0.3294062316,0.0752365515,0.3648801744,0.2579317689,0.574452579,0.1310128123,-0.4348478317,-0.2196396887,-0.0917119756,0.5359675884,-0.0084521472,-0.038445536,0.2916204929,0.2740372717,-0.2739708126,-0.2968041003,-0.0802817941,0.0834848136,0.2472426593,0.1379487216,-0.6362394691,-0.1290539652,0.284350127,0.3965991437,0.1000204682,0.0278409272,-0.2355177104,0.0123147592,-0.0486455821,-0.1425992846,0.0284252819,0.2278541178,-0.3272442222,0.0632215515,-0.3245686591,0.0213547926,0.1703502387,0.4805373549,0.2850087881,-0.3363882303,0.0398547873,-0.4539273977,0.203431204,0.1072443128,0.5703112483,0.5020774007,-0.5287027359,-0.3886255026,0.1841168404,0.3425621092,-0.0465759002,0.0925025567,0.1649524719,-0.1543335617,-0.255489856,-0.2398642749,-0.0500078648,0.3597525954,-0.0308982022,-0.1070574299,0.027835371,0.4318611622,0.158547163,0.0956576616,-0.6823592186,0.239442274,-0.4008047879,0.272321105,-0.0476907976,0.9379635453,0.1299600601,0.0923514962,0.3337065876,0.2017401457,0.4809494615,-0.0183208697,0.1827783436,-0.2460651845,-0.0113404812,-0.1870483756,-0.2756963074,0.0733893961,0.1745607853,-0.2565917969,0.3026056886,0.0013166665,0.1466768086,0.068698585,0.1166922301,-0.0081486907,0.3080254793,-0.3076499999,-0.082559973,-0.087612994,0.0282528643,0.1045907587,-0.1090517566,-0.5096393228,-0.1694669425,0.000223622,0.1846589744,-0.032826487,0.0137886144,0.3616638184,-0.1149209812,-0.1352647394,0.1706979722,-0.0022500334,-0.2959180176,0.017662717,-0.0759263039,0.2719798982,-0.0784948394,0.3047949672,-0.0825137421,0.3430141807,0.1229274794,-0.3146331608,-0.0677113533,-0.223583281,-0.1790149361,-0.0821914151,-0.1533217877,0.0760868192,-0.2540961206,-0.1125236377,-0.1276055276,-0.1201350689,-0.0320298821,-0.0108245816,-0.1855920851,-0.0641234517,0.1463435143,-0.0642082021,-0.1906718165,0.0457104705,0.104787387,0.1875716448,0.6064355969,0.3791364729,-0.0399996676,0.0133007457,0.0572318882,-0.3177098632,-0.3002565801,-0.4511320889,-0.1510186642,-0.1692899019,-0.2728013694,-0.2661795318,0.6759356856,0.2157923728,0.2065015435,0.058778055,-0.187387839,-0.0219804607,0.4642726779,0.3379494548,0.1834041923,0.0576505251,0.6058554053,0.1056460962,-0.0143961636,-0.1383230537,-0.1891865879,-0.5123186111,0.2607536912,-0.0032221302,-0.0168849062,-0.199546814,-0.3552440405,0.0996780172,-0.1902482659,0.1187714711,-0.001830496,0.0799269229,0.2026243657,0.3011369705,0.1307026744,0.0292954575,0.059095744,0.3133092523,-0.086357154,-0.1691776812,0.0408393741,-0.0871464387,0.2083610296,0.2513519824,0.2837483883,-0.4437426925,0.1296731681,-0.1673194915,-0.093604289,-0.020044595,0.3436528742,-0.1722467989,0.15085347,0.0111372229,0.1842427701,-0.003724643,0.090871267,0.0047094137,-0.0272976235,-0.3770060539,-0.0851903185,0.0857567117,0.3486763239,-0.0090898834,-0.1660287678,0.301148802,0.0432871804,-0.3755925894,-0.0223312713,0.2652483881,-0.3256576359,0.1198397577,0.2465086281,0.4056600928,0.2672122121,-0.2461304516,0.1336134821,0.194164753,0.1927344799,0.0810904801,0.3828414083,-0.0625906438,0.2826988995,0.4963555932,-0.0428271815,0.1783228368,-0.2052474022,0.3776833713,0.2292224318,-0.0082835564,-0.2543436587,0.4671546221,0.1503640264,-0.2536224127,-0.0102298101,0.5429743528,0.0719741955,0.0915775299,0.3733730912,-0.0238436814,-0.2738281488,-0.2085317373,0.0868316367,-0.0629947037,-0.0161900744,-0.2711746991,-0.1961763203,-0.0899881348,-0.2310900092,0.011707874,-0.1769199818,0.545314014,0.0538267121,-0.0699245781,-0.2815671861,0.292090714,0.2002325952,0.183766067,-0.2117910981,0.0637999922,0.1046726555,-0.3237306476,0.3169930279,0.3942148387,0.3608563244,0.2614827752,-0.2633312047,-0.3394249678,-0.0505207181,0.0534342937,0.0293796733,0.0123522794,-0.2310139239,-0.0348662324,0.1708759665,0.0397926383,0.0614803731,-0.1119551212,-0.197902739,-0.2635838687,-0.2899729013,0.5647652745,-0.0532339364,-0.3765743077,0.0905212238,0.0299441069,-0.0413339399,-0.1076731533,0.123944141,0.2545179427,0.1993421018,0.1536210924,0.0158996042,0.1261710376,0.1948054582,0.1311796457,0.1723545492,0.0539706834,0.2657205462,-0.5592803955,0.0607552603,-0.0314208455,-0.3658487797,0.2509436309,0.2590923905,0.1226243749,0.1718993336,-0.3156035542,0.0229119435,0.0584223866,0.00017417,-0.1930543184,-0.2240376472,0.1051742658,-0.3982798457,-0.295742929,-0.247790888,0.42068699,-0.1910123825,-0.005666377,0.0953531116,0.0705785826,0.464858979,0.2801660597,0.0737537518,0.0240755752,0.6520791054,0.0565084666,-0.0289480295,-0.1964107156,-0.162568599,-0.4249635041,0.2250287533,-0.3275966048,0.1831674576,-0.0568537526,0.5732348561,0.0306405239,0.3586791754,0.0138112884,-0.2656754255,0.0224262644,0.247553125,-0.3517124355,0.1354187578,0.1260948479,0.0590378083,0.0177512653,0.1408074796,-0.0773067698,-0.5094180703,0.1611294448,-0.0186845828,-0.1912284642,0.0852579474,-0.0545714684,-0.0550153777,0.0704414919,-0.4616407454,-0.1345546246,0.3279925883,0.0959778875,0.0850067809,0.088345021,0.136330232,-0.0962660164,-0.094611153,0.0125522511,-0.3227331638,-0.0914963335,-0.026031347,0.067021206]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/492","title":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema","comments":"I'm doing a refactor of type inference in #363 . Both text fields should match after that","body":"Here's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n","comment_length":17,"text":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema\nHere's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n\nI'm doing a refactor of type inference in #363 . Both text fields should match after that","embeddings":[-0.0161189064,0.0418332927,0.0497547574,0.1791476756,0.2209598422,-0.1551372409,0.3485466242,0.1771947145,-0.2337825149,-0.2496900111,0.0528562553,0.2427443117,0.0418508314,0.3009526134,-0.0142641636,-0.0392464958,0.1369884163,0.1927271336,0.057350155,-0.0133017655,0.0279206391,0.2128109336,-0.432697773,0.0798531547,-0.6052960753,0.2832012773,0.1217158511,-0.081533961,-0.1724741757,-0.3538147807,0.3396440744,0.0633926839,-0.1273943186,0.2941175103,-0.0001230619,-0.0787973627,0.4711290002,-0.1221959591,0.0056540486,-0.1541832536,0.0033247711,-0.2184462398,0.1188606918,-0.1594703943,-0.0122466227,0.0787842572,0.0574800968,0.2655121088,-0.485758543,0.2143756896,0.0313540623,0.5556202531,0.3611120582,0.2949572802,0.4820652604,0.1081078947,-0.1640469879,-0.0403706841,-0.1488456428,-0.2206766009,0.3500571847,0.0233771205,-0.3031662405,-0.2488754392,0.2104114145,0.1371807605,0.0132312728,-0.1658138186,-0.2161864638,0.3846069574,0.378490299,-0.2721752524,-0.2982418239,-0.2871522605,-0.063659139,-0.1074372679,0.3144471645,0.2532115877,-0.0958044007,0.0034741391,-0.2865937948,0.1025461406,-0.0919411853,0.1352296621,-0.3512834609,0.3031047285,0.4236579239,0.0838388056,-0.1090577915,-0.058107052,0.2037526518,-0.368291378,0.1821449399,0.1974302232,-0.3036396205,-0.2578523457,-0.1963276267,-0.1283039302,0.3435287476,-0.0971284658,-0.1011330709,0.0921988636,-0.1799586564,0.2135977298,0.2164671272,0.194020465,0.0921862796,-0.046675887,0.0324375629,0.0155875199,-0.2754154503,0.1777701378,0.2533297539,-0.3579941392,0.2222444415,-0.0222091675,0.5951607227,-0.2146412283,-0.3581434488,0.1046301797,-0.6537966728,-0.029283965,0.1376782507,-0.124590382,-0.1058189422,0.4346594214,-0.0013267018,0.3968073726,-0.2712625265,-0.363132298,-0.0425493792,-0.0084453467,0.0379900299,-0.011499295,0.1196713448,0.2067755312,0.1114208326,0.0167975277,-0.3512586355,0.2169020176,-0.1675958633,-0.2053929716,-0.0912243351,0.0087310197,-0.2506656051,-0.3312960863,0.1384344697,-0.3670900762,-0.2213854194,0.1448272616,0.0643103421,-0.1541718245,-0.1422813088,0.0524084941,-0.4513302147,-0.1260323673,-0.162942782,0.335595727,0.2153673619,0.0086010993,0.0566011667,-0.1638466418,-0.010453105,-0.2158825845,0.2360284179,0.1352060288,-0.0815292671,-0.0126469173,0.1752243489,-0.2601315379,-0.0406284444,0.2410740405,-0.0236211754,0.06748835,-0.1356779188,0.5942892432,0.3529723883,0.0254981369,0.1012028828,0.0385647714,0.2206334919,0.0668069571,0.0872135535,0.0233997442,-0.3188511729,0.0589516349,-0.0854753703,0.3682381809,-0.1921682507,-0.0841463506,-0.2003959566,-0.2794536054,0.5485409498,-0.1183056757,-0.0097655393,0.0945071802,-0.1517408937,0.2460001558,0.265645057,-0.2121187449,-0.1468266547,-0.0330821797,0.3460656106,0.3285520673,-0.425127387,-0.4670061767,-0.5462675095,0.1526118815,-0.351154387,0.5649885535,-0.0926044285,-0.3628206551,-0.0444890596,0.0957760215,0.2239724696,0.1940458715,0.0716375783,0.0665750653,-0.0451185443,0.1773070395,-0.0164547116,-0.146097675,0.029327767,0.1820035428,-0.1878122091,0.5039507151,-0.0505503863,-0.2504658997,-0.0486585088,0.5068321228,0.2212622017,0.0639989302,0.0746512488,0.1706355363,0.0844889209,0.0274592768,-0.2565710843,0.2689789534,-0.0357396677,-0.092710413,-0.1811239123,0.2658528686,0.1443213224,-0.0504332259,0.0820591524,0.478107959,0.1191560924,-0.168379426,0.2042897195,-0.1395613402,-0.0092684506,-0.118708849,-0.0515521392,-0.0625140518,0.0655582249,0.0588733144,-0.5917344689,0.5582193136,-0.6158355474,-0.136264056,0.4373213649,-0.1705741733,0.1929085851,-0.2389010042,-0.1187842935,0.0776206404,0.2546368241,-0.4221937656,0.1173963323,0.1194532886,0.048579853,-0.0174208209,-0.3542957008,0.1041503176,0.2996139228,0.0059601068,0.1607751846,-0.0010493792,0.2029417902,-0.0292646028,-0.3321914077,0.1610555351,0.1127512977,-0.0855798572,-0.3652036488,-0.051685445,-0.2765150666,-0.0831770301,-0.4930624068,-0.1210074648,-0.3591436148,-0.4036301076,-0.0751148537,0.7243289351,-0.0182254557,0.1753558666,-0.6069014072,-0.2430244237,-0.0019760968,-0.342610836,-0.0480912104,0.0056586768,-0.1899309903,-0.0440212749,0.349336803,0.2407254428,-0.2683977783,-0.0032677299,-0.2104715109,-0.2128612995,-0.8254592419,0.0923086256,-0.3423710763,0.0749224648,0.2931859493,-0.0475140214,0.2073671073,-0.2992775142,0.1691264212,0.1805768609,-0.4905545413,0.4479922354,0.0758826062,-0.289455086,-0.0016760846,-0.171876967,0.0003845845,-0.1767751426,0.0933616087,-0.0921217948,-0.0221567992,0.3520934582,-0.0063570072,-0.0300375801,-0.1265924871,0.0043456126,0.1129131913,0.4491636157,0.4396447837,0.3248732686,-0.0922381207,0.1975491196,-0.2271224707,-0.2233732939,0.1804771572,0.0070676259,-0.1562499404,-0.2261992991,0.5512505174,0.2355549634,-0.4834789038,0.3058188558,0.4202595651,0.077481702,0.0824681446,-0.2298171669,-0.0179771837,0.3961814344,0.3203352988,-0.0228164364,-0.1455749869,0.1353580505,0.0627212152,-0.1255173981,-0.2235631943,0.5573557615,-0.1932648718,-0.1299394816,0.1700903475,0.0713229105,-0.1176922396,-0.0582736693,0.1083228365,0.1492452919,-0.2013480216,-0.2157623172,0.084583655,0.0908726901,-0.2214397341,-0.27298823,0.163974613,-0.3020723462,0.2303423434,-0.0404953919,-0.1805935055,-0.2284918278,-0.3759212494,-0.0044902046,0.1142399311,0.0032656614,0.1018866822,-0.6307584643,-0.0709882826,0.0868610889,0.3399066031,0.1385790557,0.3695656359,0.1523776948,-0.502094388,-0.1680769175,-0.0866928473,0.4714750051,0.0638903603,-0.0974172726,0.1335677952,0.1817171425,-0.1957000941,-0.2683575749,-0.0107413279,0.0082232347,0.3010815382,0.2705413997,-0.639673233,-0.1791681051,0.3811224401,0.3375141919,0.0145524619,-0.0312188417,-0.1594942808,0.0915168077,0.0923595801,-0.1257299483,0.0637995005,0.2886252403,-0.3415796161,0.0139587224,-0.3368560374,-0.0377888158,0.2842560112,0.4143279195,0.223479107,-0.3767017424,0.0828052387,-0.4237990975,0.1871515661,-0.0236257929,0.4478932619,0.4763248563,-0.5722714067,-0.4386937618,0.137778163,0.3645659387,-0.101865001,0.214809224,0.2886378467,-0.1515352875,-0.2807994187,-0.180992648,0.0132280309,0.3981312215,-0.0156292748,-0.0768472925,-0.0258171428,0.5186474323,-0.0225814767,0.0883838981,-0.6236912608,0.3999246061,-0.2987964451,0.3228163421,-0.0470073372,0.8657002449,0.0237572528,0.0476168953,0.2893191576,0.1270741224,0.5067017674,0.007846795,0.1557168812,-0.2817581594,0.1701719761,-0.1489223093,-0.2062648386,0.1013187915,0.2628420293,-0.1684441715,0.142224431,0.0898132324,0.2186553031,0.0041729226,0.1318729818,-0.0253724027,0.3312389553,-0.2488351762,-0.0895866975,-0.0604133271,0.1597784311,0.1583166271,-0.1786233932,-0.4871671796,-0.0729666576,0.0176851489,0.233814925,0.0173136108,0.0315410271,0.4748317003,-0.1961058974,0.0424908884,0.3946954906,0.0842883065,-0.3070059121,-0.0172493588,-0.0754971504,0.2429916561,-0.1359571815,0.2157039493,-0.2212693989,0.2682464421,0.1090969294,-0.3341181278,-0.1193563566,-0.0863955766,-0.1597208083,-0.0778873712,-0.1271838993,0.1667787731,-0.1795397401,-0.2407258302,-0.0648114085,-0.0542013422,-0.0677198172,-0.0305903107,-0.2076230049,-0.0698593929,0.0501648821,0.015396446,-0.1575434655,0.0998773575,0.1503840536,0.2214140445,0.6293485165,0.4565291405,0.0286152996,-0.0017692659,0.0247230437,-0.2843742967,-0.2169724405,-0.3930873275,-0.1469142288,-0.0448106602,-0.2518891692,-0.1806902736,0.7333981991,0.1308616698,0.1295643896,0.0895379633,-0.1842827052,-0.0429358035,0.3645335138,0.1538838148,0.2513920963,0.0819069371,0.5773454309,0.1186793447,-0.0448270552,-0.1121121198,-0.0954188183,-0.3992855549,0.2623731196,0.0613640063,-0.055696331,-0.1916269511,-0.4026491046,0.1119806841,-0.257745266,0.2308489829,-0.0068080649,-0.033836022,0.2028183043,0.3138066828,0.0888000429,0.0690880418,0.1050522253,0.3505173326,-0.0829090253,-0.0861672014,0.1872447282,-0.0784431472,0.3959691823,0.1823231578,0.2939938605,-0.467949599,0.0343058705,-0.1881801784,-0.0878022164,0.0834835321,0.2765671015,-0.0083836187,0.1889642924,-0.0700783357,0.1228085086,-0.013631816,-0.0546037555,-0.0650930181,-0.0648845062,-0.3568446934,-0.002499182,-0.074743405,0.4072778523,0.0618506223,-0.1671941131,0.2969161272,0.0530839898,-0.3996775746,0.079594627,0.1522869915,-0.4152296185,0.0922972038,0.1816769838,0.3825950921,0.4038789272,-0.3096643388,0.0723771602,0.1195047498,0.1604404002,0.1794630885,0.283893764,0.030270664,0.1587478369,0.4817334414,0.0031789879,0.0648491904,-0.1722460389,0.4451206923,0.1695605069,-0.248500526,-0.1077877656,0.4351513088,0.0730048344,-0.2572182715,0.1032343656,0.4207418561,0.0211806241,0.1734536141,0.2988909781,0.1028895676,-0.1919418126,-0.2126992494,0.0856518,-0.1438970864,-0.1289997548,-0.2481292188,-0.337312758,-0.0672619417,-0.1897701323,0.0088829482,-0.1929847002,0.4818557799,0.0361586586,-0.137183696,-0.2450810373,0.3785538375,0.1929199696,0.2649739385,-0.1939347237,0.0645455718,-0.0002824203,-0.3305129409,0.3294852078,0.3417547345,0.2270214111,0.3470463157,-0.3818877041,-0.3901887238,0.0075492449,-0.0243573505,0.0469747894,-0.1039172187,-0.2061827183,0.1557299793,0.2123421282,0.0250740182,0.1044118404,-0.0466437899,-0.1971744448,-0.2442260981,-0.2792159617,0.5544175506,-0.0030109757,-0.1924303472,-0.0728186518,-0.0142637976,-0.0339003801,-0.3230066895,0.1046110317,0.2030600905,0.1489593536,0.0625698417,0.0109980088,0.1575793922,0.0550223365,0.1844745576,0.052803997,-0.04441173,0.3143443465,-0.4968889952,0.05857867,-0.0200466886,-0.2889113426,0.1667104363,0.3873151541,0.0609078668,0.0682779327,-0.3706538081,-0.0068576862,0.1168920845,0.060620103,-0.0962383524,-0.1325853765,0.0162220392,-0.3778547943,-0.2371814549,-0.24194704,0.3659837246,-0.1840184033,-0.0142632928,-0.0643141195,0.2227291912,0.4342676997,0.3516326845,-0.0617970042,0.0468503721,0.7053897381,0.127047196,-0.0117466683,-0.0867316276,-0.0395373031,-0.5143460631,0.2094183266,-0.297378391,0.2212361097,-0.002668574,0.601721108,0.1248425022,0.3921010494,-0.0645234808,-0.3328499794,-0.0212387461,0.1732373685,-0.3336331248,0.1193784699,0.1856160313,0.0193986632,0.0606359169,0.0533926636,-0.0882342756,-0.454374969,0.09792725,-0.0631977394,-0.1480974108,-0.0222428236,-0.1345863044,0.1181329712,0.1067835465,-0.581104219,-0.1921417564,0.3485527933,0.0206630826,0.1605139822,0.1254466474,0.255687654,-0.1377919465,-0.1797768474,0.0106499847,-0.2607831657,0.0892707705,-0.1012819782,-0.0876175985]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/492","title":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema","comments":"By default nullable will be set to True","body":"Here's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n","comment_length":8,"text":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema\nHere's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n\nBy default nullable will be set to True","embeddings":[-0.0597350635,0.020884037,0.0121693527,0.1386784464,0.2081823647,-0.0631272271,0.3283044994,0.1403906345,-0.2708768547,-0.3422797024,0.0534099638,0.2863906324,-0.0883676484,0.3610792756,-0.0066221873,-0.034791097,0.2013576776,0.2277596444,0.0959846452,0.0323305428,-0.0212752204,0.1505942047,-0.3719955981,0.0844485462,-0.6136106849,0.2484197468,0.0452213921,-0.0657066777,-0.1522474736,-0.3937349021,0.3249823749,0.147156924,-0.1463051587,0.3095084429,-0.0001255305,-0.0587978996,0.4344861209,-0.1438227296,-0.0635293275,-0.139529407,-0.0778823495,-0.2721914053,0.137548238,-0.2096318603,0.0241037998,-0.0078636324,0.0258459393,0.2503262162,-0.5159304738,0.2352961898,0.0368498079,0.6410489082,0.3111078143,0.2796728909,0.4912626445,0.0734693408,-0.2531262338,0.0278056674,-0.2257163823,-0.2351019531,0.3821918368,-0.0499913506,-0.3477748036,-0.2468803823,0.2217561603,0.0438821316,-0.0133317849,-0.1588592976,-0.210109666,0.4336486459,0.3740136921,-0.2935658395,-0.3100818098,-0.2875709236,-0.0185453016,-0.1425522566,0.2638463676,0.3826983273,0.0492892042,0.0221543144,-0.4096701443,0.1429872513,-0.0950998217,0.1707137376,-0.3570522368,0.3538344204,0.4198227823,0.11698699,-0.047091648,-0.0751353428,0.1888923347,-0.3439044952,0.1331655085,0.1567875445,-0.2868037522,-0.2121208161,-0.2721185684,-0.1057998165,0.4752843082,-0.0720683038,-0.0761020035,0.0844076276,-0.1772782207,0.1945656836,0.1480740756,0.2151602656,0.1647093147,-0.0699966028,0.0036941257,0.0037746921,-0.2714701295,0.2570689023,0.2788900137,-0.2678378224,0.2000959367,-0.0226920135,0.5936886668,-0.1271915436,-0.4148895144,0.0798956975,-0.6584081054,-0.1063159555,0.1283833385,-0.059987396,-0.0415098295,0.364797771,-0.0963850915,0.3918248415,-0.2237041891,-0.3302684128,-0.0726358369,0.0172925834,0.0747152567,0.0247185994,0.1330604702,0.0914207324,0.1651644558,-0.0021090873,-0.3132512569,0.1933936477,-0.0740096718,-0.1736866534,-0.0909896046,0.0827644914,-0.2067174911,-0.2952288389,0.1017626822,-0.3783075213,-0.2819790542,0.1599660069,-0.0022319723,-0.2642727196,-0.092016533,0.0463493839,-0.4435577989,-0.0898797363,-0.0604381785,0.2923828959,0.2130704224,-0.0076946858,0.0957759023,-0.1393678635,-0.0153531143,-0.2934083641,0.1128658131,0.1698661,-0.0875824541,0.0031146198,0.1884275526,-0.2909247577,-0.046344433,0.2726826072,-0.0307013076,0.065660961,-0.1804792434,0.4900422096,0.3232553899,0.0498991311,0.0509927347,0.0384092256,0.1993471384,0.0549693704,0.1226168051,0.1069110334,-0.2926638722,0.0047158161,-0.113419354,0.3440844417,-0.1385008693,-0.0900838971,-0.1269411892,-0.3193788826,0.6502860188,-0.1165814176,-0.066075705,0.1404481381,-0.1163600087,0.3061498106,0.2979385555,-0.2057720274,-0.159730196,-0.007160102,0.2123869509,0.2451272458,-0.4375897348,-0.4165979922,-0.5919030309,0.1541799605,-0.331294477,0.5389016271,-0.0569695681,-0.4040943682,-0.0834590122,0.1319044381,0.1622988284,0.1513739228,0.053196162,0.1483138949,-0.0127951596,0.1685631424,-0.0522252731,-0.0750359371,0.0526927002,0.1646139622,-0.2061636001,0.5347011685,0.00278376,-0.2295766026,0.0382909365,0.5389218926,0.2875726521,0.0099130003,0.0793321207,0.1966691613,-0.0044836793,0.0280737858,-0.1996313483,0.3866656125,0.0311243869,-0.0710657313,-0.1512087137,0.3293085098,0.1306029111,-0.0934778824,0.0377685875,0.5906512141,0.0618434325,-0.1477007866,0.2201209366,-0.1108726859,-0.0061383406,-0.1173455715,-0.1054376289,-0.1466505975,0.1040819585,0.0263595134,-0.5558634996,0.5491797924,-0.6089541316,-0.1616200507,0.4234130681,-0.2069806457,0.2208489627,-0.2433069497,-0.1418639123,0.0954222381,0.352286756,-0.4031302631,0.1232059076,0.1191231012,0.0808889046,-0.010419365,-0.3745527267,0.0898170471,0.2980148792,-0.0695844144,0.1288015246,-0.0245125648,0.2252023369,-0.1642980129,-0.3474866748,0.1326829046,0.1201903,-0.0800362304,-0.3314780891,-0.0324064568,-0.3120579422,-0.0835593417,-0.4015709758,-0.1815852225,-0.4237849414,-0.4418182075,-0.0933420509,0.7129380107,-0.0192967579,0.2138645649,-0.6736953855,-0.2574226558,0.0317691676,-0.4067466259,0.0088652382,0.0495842919,-0.0754573494,-0.0688127279,0.2525759935,0.2596936226,-0.1672881991,-0.0021023047,-0.1726115793,-0.242779538,-0.7182214856,0.0983697996,-0.3956941068,0.0439363457,0.2822448313,-0.0363323428,0.2014860064,-0.2762645483,0.1778001636,0.2180732638,-0.3957937658,0.4361875355,0.0818167031,-0.2895507514,-0.0024879673,-0.1117900908,-0.0155547671,-0.2048933357,0.1016196609,-0.061486531,0.0573266484,0.3110711277,0.0008290294,0.0044259732,-0.121200867,-0.0266568735,0.1183293462,0.4131828845,0.5135343075,0.2388001531,-0.0729243234,0.2006088048,-0.237606585,-0.2471648604,0.2664754093,0.0162625387,-0.1511445045,-0.1974180341,0.5477028489,0.2389688194,-0.5269469023,0.2817563415,0.3643774688,0.1024785861,0.0489011444,-0.2751307786,-0.0748274177,0.4428215325,0.2771472633,0.0625508428,-0.179861486,0.1103140265,0.0987932384,-0.0934011191,-0.1807136238,0.5831880569,-0.2477495074,-0.1143461391,0.1957502663,0.0393549427,-0.1652594656,-0.012801677,0.1371041536,0.1443104595,-0.1978074163,-0.1926635802,0.1154867485,0.1212906614,-0.1501743644,-0.163479358,0.1751876324,-0.2908284068,0.3398902118,-0.1055570096,-0.1747492552,-0.1910919994,-0.3639567196,0.0388681367,0.1693535298,-0.0408089235,0.1236696616,-0.6418358088,-0.0189553183,0.0676957518,0.296456337,0.1192647293,0.4459927082,0.1375071853,-0.4808485806,-0.1777559668,-0.1028230488,0.5364670753,0.0754496753,-0.0621176809,0.2279929072,0.1981808841,-0.2798018157,-0.3005771637,0.0178386904,0.0872575045,0.3005458117,0.2813715339,-0.6402693987,-0.1918012947,0.3040622473,0.3746522665,0.0101057477,-0.0328187086,-0.1761928201,0.0332199521,0.0236079339,-0.1423874199,-0.0096101044,0.2353884876,-0.3604550064,-0.0428208597,-0.3543444276,-0.0747422203,0.2858725488,0.4250281453,0.3026975989,-0.3614391387,0.1004581079,-0.4509877861,0.145721361,-0.043783769,0.5166503191,0.4580870271,-0.5108500123,-0.451820612,0.1119030267,0.3316417336,-0.0598896407,0.1692962497,0.3103397489,-0.1790432781,-0.2157631963,-0.2183538228,0.0341009572,0.3654386997,-0.1288332045,-0.1375283748,-0.054635901,0.5391080379,0.0835798383,0.0927888975,-0.6805208325,0.3257465065,-0.3508454561,0.3303340077,-0.0210061688,0.9474201202,0.00365512,0.0968813375,0.2851679623,0.1297618151,0.461297363,-0.0011565312,0.1350178868,-0.3423935175,0.1952522695,-0.1885187626,-0.2481838614,0.1240564361,0.2474047542,-0.1999174803,0.2174632102,-0.0367318727,0.202674672,-0.0587851629,0.179687649,-0.0286661591,0.3350063264,-0.2501300573,-0.0965211838,-0.0862456039,0.1361624748,0.1338602901,-0.1659347266,-0.4587306976,-0.0750424638,0.0559381954,0.1115299985,0.0395696796,0.0067251446,0.3984144032,-0.2534787953,-0.0386062376,0.3201803565,0.0173446648,-0.3192642331,0.0100973798,-0.1165294647,0.2353086025,-0.148594901,0.1582009196,-0.1754218489,0.2784201205,0.1414984018,-0.2654873431,-0.0944519565,-0.1022698879,-0.1389694512,-0.1327950954,-0.096614629,0.036653202,-0.2128440291,-0.1945607364,-0.0318936892,-0.0765330046,-0.0764766037,-0.0521522388,-0.1987257749,-0.0420050845,0.0444355533,0.0364369228,-0.1640881002,0.0557334572,0.1239170656,0.2603317499,0.6043723822,0.4075954854,0.0003877147,-0.0079445885,0.0111727938,-0.2432221621,-0.2579716742,-0.3546795547,-0.0937401131,-0.088268362,-0.2231822312,-0.2369000912,0.6876326203,0.1261771321,0.2493024468,0.119512327,-0.1730232537,-0.0839450806,0.3266427517,0.2323450595,0.2894643247,-0.0179405212,0.5015746355,0.1483290792,-0.0486037768,-0.0966143534,-0.1156853512,-0.4184356034,0.2435720563,0.1322387308,0.0216943026,-0.231021896,-0.3486384749,0.0928752497,-0.3038947284,0.1580753624,0.0155425705,-0.0421978571,0.2233418673,0.3277329504,0.0851535648,0.0095370421,0.0841738582,0.3361693323,-0.1082786396,-0.0538368858,0.1909888834,-0.1188639402,0.3479019105,0.2051142156,0.2374228835,-0.4334314466,0.0088830227,-0.0674506873,-0.0647538826,0.0734206736,0.3344576061,0.0009053297,0.1776964366,-0.0031253158,0.1273834556,-0.0140398033,0.0202651825,-0.0467816368,-0.0982367918,-0.3016041517,0.048854392,-0.0299970191,0.4649605751,-0.0470895432,-0.1494235694,0.3197104335,0.0288208202,-0.3654848933,-0.0192938093,0.229398787,-0.354333818,0.0419916809,0.1828685254,0.3889395297,0.3620495796,-0.3723714352,0.0627231151,0.1223541498,0.1302496195,0.162143141,0.2286185473,-0.0096520847,0.1690406054,0.5085991025,0.0031548138,0.0779023543,-0.1385605335,0.4601182342,0.2893788218,-0.0883513242,-0.1717671305,0.457403034,0.1604172885,-0.325040102,0.0842365697,0.4515223205,0.0954152867,0.2247812748,0.317558229,0.0805218443,-0.2432469428,-0.1687746793,0.072746709,-0.1049396768,-0.1071671322,-0.2583351135,-0.312818557,-0.0797531307,-0.1803425401,0.0150709469,-0.214799732,0.5172771811,0.0682889819,-0.0428715944,-0.2651297152,0.4386022389,0.2374718487,0.1793858558,-0.2224191427,-0.0288450178,0.0570785441,-0.2610165179,0.261731863,0.2343724519,0.2695312202,0.2613160014,-0.3779346645,-0.3774036169,-0.0181652829,-0.0124824829,0.0769639164,-0.0441963673,-0.2057920247,0.181586653,0.2518902421,0.0024753788,0.0910812914,-0.0059262225,-0.1712349504,-0.3060333431,-0.3038693666,0.5648995042,-0.036615096,-0.2871672511,0.0123636471,-0.0184461083,0.000947364,-0.2382259816,0.1934973598,0.2186689526,0.2357762158,0.1147131175,-0.0055373819,0.2090849578,0.0957027227,0.0842032209,0.1190257594,-0.0104587423,0.2857113779,-0.5487099886,0.0386896282,-0.0746961832,-0.2965120673,0.1383705586,0.294310689,0.0472973697,0.0960764363,-0.4000143111,-0.0193419438,0.1944928467,0.0035650863,-0.1289337724,-0.1481680423,0.1160731092,-0.3349030912,-0.2765428424,-0.2781659961,0.3767978847,-0.230714187,-0.0382883623,-0.0605705306,0.1945783049,0.4585614204,0.2639157772,-0.0288294274,0.0102321925,0.7164396644,0.113471061,0.0584762357,-0.1389955878,-0.0099027529,-0.4111581743,0.1441848278,-0.2882116139,0.2470389605,-0.0100424429,0.6224292517,0.0238287542,0.4472152293,-0.0675780326,-0.3026520014,0.0785099119,0.1804637164,-0.3547310233,0.0819835514,0.1507246047,0.0645892695,0.0359846354,0.0266789366,-0.0269720033,-0.4595888257,0.1576984227,-0.0472666807,-0.2222174257,0.0753072575,-0.1108874157,0.1452618241,0.1406121552,-0.4749511778,-0.1911162734,0.3411378562,0.0048264661,0.1430996954,0.0946592912,0.2379060686,-0.1284876019,-0.1627023369,-0.0176989362,-0.2349477261,-0.0161532518,-0.0426645577,-0.0656700507]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/492","title":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema","comments":"It should be good now. I was able to run\r\n\r\n```python\r\n>>> from nlp import concatenate_datasets, load_dataset\r\n>>>\r\n>>> bookcorpus = load_dataset(\"bookcorpus\", split=\"train\")\r\n>>> wiki = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\")\r\n>>> wiki.remove_columns_(\"title\") # only keep the text\r\n>>>\r\n>>> assert bookcorpus.features.type == wiki.features.type\r\n>>> bert_dataset = concatenate_datasets([bookcorpus, wiki])\r\n```","body":"Here's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n","comment_length":48,"text":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema\nHere's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n\nIt should be good now. I was able to run\r\n\r\n```python\r\n>>> from nlp import concatenate_datasets, load_dataset\r\n>>>\r\n>>> bookcorpus = load_dataset(\"bookcorpus\", split=\"train\")\r\n>>> wiki = load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\")\r\n>>> wiki.remove_columns_(\"title\") # only keep the text\r\n>>>\r\n>>> assert bookcorpus.features.type == wiki.features.type\r\n>>> bert_dataset = concatenate_datasets([bookcorpus, wiki])\r\n```","embeddings":[0.0195755307,0.1014376283,0.0100496262,0.1525079161,0.2617855966,-0.0262445733,0.3992982209,0.1835874319,-0.2757073045,-0.3310970366,0.0578915067,0.2993277311,-0.0961161107,0.3490905464,-0.0248573888,-0.0886692926,0.2123325616,0.2151870579,0.0904195979,-0.0006232074,-0.021691367,0.1471001059,-0.4190040827,0.0407256186,-0.5618267059,0.1993517131,0.0091327727,-0.0635869876,-0.1321011037,-0.3558357954,0.359895885,0.1632291377,-0.0964083001,0.3494224846,-0.000123727,-0.0465317518,0.4617714584,-0.0944593549,-0.1142689288,-0.1229414791,-0.0468562543,-0.2261640728,0.1428212374,-0.2239898592,0.0660102889,0.0396810472,-0.0237114392,0.2180909365,-0.4827174544,0.3005023897,0.0497984998,0.6041247249,0.3216882646,0.2887707055,0.4325879812,0.0260213073,-0.2378447503,0.0571174398,-0.2866386473,-0.240039289,0.3645997345,-0.0030348324,-0.4224071801,-0.2776353657,0.17048572,0.0321026519,0.0139935929,-0.1811416149,-0.1326822788,0.3763589561,0.3835633993,-0.3270166814,-0.2761546075,-0.2740529478,-0.0383335501,-0.1471807957,0.2313626409,0.3422130942,0.0426287167,0.0377751254,-0.3583473861,0.1332840919,-0.0611011237,0.1729693711,-0.357649982,0.2891921997,0.4066442549,0.126505509,-0.0574196093,-0.0956979245,0.239703387,-0.3863389492,0.1325589269,0.1334773153,-0.3427969813,-0.1534195989,-0.2158363461,-0.1752316058,0.4142799973,-0.1013137847,-0.0684355274,0.0484885499,-0.1178586558,0.158159852,0.1803003252,0.3018612862,0.1722828746,-0.0425041802,-0.0473931432,0.0539680012,-0.2548178434,0.268329978,0.278373152,-0.2501658499,0.1644538939,-0.0081232116,0.5877899528,-0.1211650744,-0.3953981698,0.1054600105,-0.6579975486,-0.0653714165,0.1158040315,-0.0750866905,-0.016488662,0.4329295158,-0.0250338651,0.4224492311,-0.2510067225,-0.2912806273,-0.1103081182,0.0483359583,-0.006532588,-0.028801525,0.1892850101,0.0969447643,0.1741938889,-0.0417579561,-0.3898367584,0.2016262263,-0.1408275366,-0.2009076327,-0.0840657949,0.0512061007,-0.1916183382,-0.2432677597,0.1291822046,-0.4263624549,-0.2597584426,0.1686255932,0.0584447272,-0.228746593,-0.1498566717,0.0685939789,-0.4282655418,-0.1048147976,-0.0640074462,0.2332176566,0.230550766,-0.0454131365,0.0735665709,-0.1591985375,0.021661913,-0.262665391,0.1199827045,0.2031269819,-0.0939802006,0.0117671741,0.1568477154,-0.2104033381,0.005221765,0.2258634269,-0.0269044489,0.115931727,-0.186822623,0.5118384361,0.3278872371,-0.0432593897,0.0168937389,0.0380694084,0.2024354339,0.0965592265,0.1285848469,0.044415094,-0.2982770503,-0.0274744742,-0.0779013932,0.3087241948,-0.0985608324,-0.0781478584,-0.1820751131,-0.2565665543,0.6591675878,-0.0943205431,-0.056405399,0.1404180378,-0.1382977962,0.3328261375,0.3180729449,-0.1969291568,-0.1213603988,0.0179033335,0.1878581941,0.220866397,-0.4277662635,-0.3764052689,-0.6329755187,0.1295931786,-0.2952991128,0.5714241862,-0.0909824297,-0.3621647954,-0.1214407682,0.1148455888,0.1560097188,0.0661059469,0.0689284652,0.1525991112,0.019554507,0.1733784527,-0.063270323,-0.0691425502,-0.0001670228,0.1299988627,-0.3026860654,0.479421109,-0.0680714473,-0.2394234985,0.0380316302,0.5497893691,0.2630757987,0.0291558802,0.0447435454,0.2267813236,-0.0474290662,-0.0320529081,-0.184887141,0.3059113622,-0.006551655,-0.0952601209,-0.1367530078,0.2938750982,0.1112039983,-0.1022120267,0.0039689308,0.5206984282,-0.0201631188,-0.0484421477,0.187858358,-0.0610092357,0.0482265204,-0.1584980637,-0.0894819722,-0.1095933318,0.1126954183,0.0823305026,-0.4957492948,0.5330727696,-0.5667440891,-0.1811655313,0.4938714802,-0.1976945847,0.1824948192,-0.2020718157,-0.1289619952,0.0910811648,0.3797803521,-0.4210346043,0.2070302069,0.1284208745,0.1107249856,0.0249741543,-0.3196204901,0.0812596232,0.3331664205,-0.0291224178,0.111473456,-0.0383761674,0.2469995618,-0.1389680803,-0.3223757446,0.098164089,0.1357905269,-0.0476688854,-0.3277385831,0.0029740955,-0.3342837989,-0.1165993363,-0.3874281347,-0.2271279544,-0.4085310102,-0.4292827845,-0.1264097542,0.6499423385,-0.0563761368,0.2817405462,-0.654840529,-0.294658035,0.07852678,-0.4065558612,-0.0109403497,0.0041679931,-0.125518918,-0.067432791,0.3089804053,0.2916698754,-0.1843713522,-0.0413820632,-0.2161363065,-0.1620108336,-0.7112454772,0.1253072172,-0.4536364079,0.0261319242,0.297493577,-0.0095083332,0.1389731616,-0.3155109584,0.186070472,0.1755691916,-0.3917913139,0.4901833236,0.1041634083,-0.2944886684,-0.0321601257,-0.1916290522,-0.0293731745,-0.2649886012,0.1147539094,-0.0891405493,0.0408561379,0.3860730827,0.021111479,0.0487828329,-0.0745107234,0.0348818749,0.1042637676,0.4250565767,0.4264548421,0.2812412679,-0.1241302937,0.1868460029,-0.2605314851,-0.2054006904,0.2208205462,-0.0510112084,-0.1454127729,-0.2333434671,0.5366603732,0.2374630123,-0.3884631693,0.2863419354,0.4431610107,0.0826120079,0.0101798903,-0.3189225197,-0.0405623019,0.384085685,0.2121475488,0.0399056971,-0.1200202778,0.1387341172,0.0387932435,-0.0864963159,-0.2054846883,0.530977726,-0.2393932194,-0.0613697208,0.1415787041,-0.03202191,-0.1077440456,-0.0494660847,0.1545826495,0.1559004784,-0.2657797635,-0.2518395185,0.1059899256,0.1672793776,-0.2113858014,-0.177841574,0.2012264878,-0.305159539,0.3195039332,-0.040607214,-0.1985891014,-0.1919311732,-0.3699122369,0.0338114835,0.1395127773,-0.0981299654,0.1026445851,-0.6497277617,0.0017366242,0.0208792258,0.3767075241,0.1342526972,0.4731307626,0.1460255086,-0.4508696198,-0.190236792,-0.0615375675,0.5640015006,0.0612318963,-0.0645726547,0.318395406,0.2116873562,-0.2911326587,-0.2674374878,-0.0787036195,0.0937376022,0.3564350903,0.2167785913,-0.6534016728,-0.2309086174,0.3677493036,0.3814469874,0.0330670439,0.0042817993,-0.1924687922,-0.0612924173,0.0153521318,-0.1625397056,0.0129422108,0.3118935525,-0.417218715,0.034411788,-0.3534698486,-0.1319494247,0.2542102635,0.3573930264,0.3443434238,-0.3219574094,0.0954724178,-0.4758375585,0.1821273267,-0.0383868031,0.5125539899,0.4208607674,-0.4859166741,-0.372866869,0.1200297624,0.3413616717,-0.0079405494,0.1713605523,0.2588386834,-0.2045013905,-0.2189020365,-0.1618321985,0.0314353369,0.3969662488,-0.1037935838,-0.1424650401,-0.0794328228,0.4981072247,0.0853006914,0.1210734323,-0.6400702,0.2907299399,-0.3826375008,0.3815762997,-0.118493706,0.9706950784,0.034426447,0.1159855723,0.3479290903,0.1673336923,0.5139724016,0.0721904635,0.1802600622,-0.300727129,0.1964896023,-0.1293438226,-0.2213265747,0.1310353279,0.1971330345,-0.2321762741,0.232596457,-0.0664608553,0.1716502309,-0.0171470158,0.2490389943,0.0054718279,0.2760457397,-0.3180602491,-0.1031709388,-0.1124358773,0.0892093256,0.1288127005,-0.1908584386,-0.4532469511,-0.1403726339,0.083198376,0.1300480664,0.0804386213,0.0620406158,0.3610145748,-0.2321573347,-0.0285358913,0.3560189903,0.0499160103,-0.3062996268,0.0172607098,-0.0950057432,0.2784456909,-0.1184562072,0.2219922096,-0.1196756512,0.319614321,0.1337704062,-0.2596149743,-0.0831990242,-0.1216642931,-0.1449956894,-0.160941422,-0.1105655506,0.0658315271,-0.2531045377,-0.1635769457,-0.024169473,-0.1154229641,-0.0930649042,-0.0350699835,-0.2148525417,-0.0962174833,0.0768126547,0.0298723057,-0.152810663,0.0419528782,0.1257472634,0.2709084451,0.6052499413,0.5042905807,0.0361744799,-0.0084300991,0.0055015595,-0.2731563747,-0.2976391315,-0.3696178496,-0.0714033917,-0.1066273451,-0.2701565027,-0.2218778282,0.6855999827,0.0500295833,0.2579543293,0.1075877026,-0.1760199815,-0.0886619836,0.3262965977,0.2517034411,0.272650063,0.01209151,0.5588988662,0.1151680276,-0.0451640226,-0.1183431596,-0.0941506699,-0.4501890838,0.2993203402,0.1259326786,-0.0273818597,-0.1964678913,-0.3474905789,0.1052604243,-0.2562189102,0.1677025706,-0.0052706935,-0.0697109774,0.2226893157,0.2814825475,0.0629889742,0.0083440505,0.0771827102,0.3129277229,-0.1115404963,-0.1160236597,0.1301070601,-0.1095884889,0.3613773286,0.1909679025,0.28765288,-0.3854857087,0.0116554582,-0.1146795452,-0.0839902163,0.0384314433,0.3189916015,0.025051346,0.1879182309,-0.0126740653,0.0783964768,-0.0395017006,0.0791227669,-0.0052776993,-0.1448149234,-0.3106230199,-0.0418730825,0.0175049622,0.4356743097,-0.0573865548,-0.2004944831,0.3062954247,0.0385014787,-0.3957902491,0.0481884629,0.2331674546,-0.3720861077,0.0785648897,0.1954507679,0.3485142589,0.2989061475,-0.3653692007,0.0893369392,0.1632991284,0.1465239823,0.1509815753,0.2636651099,-0.0648991019,0.1604724824,0.4644605517,-0.0115038315,0.0504283346,-0.1111410409,0.4408941269,0.2518627346,-0.1169424057,-0.1243630797,0.4053173661,0.1224788874,-0.2539821267,0.1131926477,0.4421788156,0.1059342474,0.1297671944,0.3113707602,0.0384411439,-0.226745382,-0.1211201921,0.1179408878,-0.0931440666,-0.0873481184,-0.2757014334,-0.2565480173,-0.0810904354,-0.1800367087,0.0243237633,-0.1540012211,0.5543488264,0.0592058115,-0.1635770351,-0.2852016389,0.4014184475,0.1921865344,0.1359365433,-0.1995634586,0.0366867743,0.0967628285,-0.2490332723,0.2249766588,0.3008886874,0.2749756277,0.2934953868,-0.4081037939,-0.3811703026,0.0252957884,0.0026776602,0.0431609824,0.0289606415,-0.1680408716,0.1478979141,0.2348801196,0.0013729372,0.0638586059,-0.0405183807,-0.1047833264,-0.281761229,-0.2665731907,0.5761907101,-0.0518483147,-0.2904804349,-0.0145231411,-0.0229910612,-0.0898443982,-0.1888554841,0.2122808099,0.2419331968,0.2618584931,0.1101134568,0.003034384,0.1769433618,0.1449158341,0.1025812849,0.1341080666,-0.041792158,0.2796098888,-0.5538759232,0.0183846448,-0.041544646,-0.3055502474,0.1791532785,0.2813614309,0.063994348,0.0571320988,-0.4404073358,-0.0674702227,0.1548766196,0.0214466378,-0.1898565143,-0.1347698718,0.125425145,-0.3303123415,-0.2333500236,-0.3570854068,0.3794100583,-0.1839249283,-0.0159746855,-0.0201080963,0.2072883546,0.4881543517,0.2732687294,-0.0220352542,0.0696333945,0.71887362,0.1258993745,-0.0476648584,-0.2302094251,-0.0618913695,-0.4107393026,0.1571617573,-0.247882247,0.2414243817,-0.0675194785,0.5831837058,0.0085766567,0.4599258304,-0.0380416848,-0.2560944259,0.0078375265,0.1765657663,-0.3624716699,0.1571974754,0.1328831613,0.1275197268,-0.0200873893,0.1368733346,-0.1184490621,-0.4956022501,0.2262371778,-0.1183772385,-0.2952270508,0.1455469579,-0.1058936566,0.1077498272,0.0919096023,-0.5134955049,-0.1520467401,0.3226170242,-0.0144651085,0.0524429977,0.0765091181,0.2326274663,-0.1332787871,-0.09630467,-0.0525508113,-0.2394361198,-0.0427297279,-0.0583460592,-0.0507116877]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/492","title":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema","comments":"Thanks!","body":"Here's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n","comment_length":1,"text":"nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema\nHere's the code I'm trying to run:\r\n\r\n```python\r\ndset_wikipedia = nlp.load_dataset(\"wikipedia\", \"20200501.en\", split=\"train\", cache_dir=args.cache_dir)\r\ndset_wikipedia.drop(columns=[\"title\"])\r\ndset_wikipedia.features.pop(\"title\")\r\ndset_books = nlp.load_dataset(\"bookcorpus\", split=\"train\", cache_dir=args.cache_dir)\r\ndset = nlp.concatenate_datasets([dset_wikipedia, dset_books])\r\n```\r\n\r\nThis fails because they have different schemas, despite having identical features.\r\n\r\n```python\r\nassert dset_wikipedia.features == dset_books.features # True\r\nassert dset_wikipedia._data.schema == dset_books._data.schema # False\r\n```\r\n\r\nThe Wikipedia dataset has 'text: string', while the BookCorpus dataset has 'text: string not null'. Currently I hack together a working schema match with the following line, but it would be better if this was handled in Features themselves.\r\n\r\n```python\r\ndset_wikipedia._data = dset_wikipedia.data.cast(dset_books._data.schema)\r\n```\r\n\nThanks!","embeddings":[-0.0448801108,0.0062618693,0.0173283126,0.1682711691,0.2230744362,-0.0621826947,0.3309408128,0.1448956728,-0.2532812059,-0.3528960049,0.0564280897,0.2673994303,-0.0631397888,0.3689028025,0.0229015611,-0.0474989824,0.2045090497,0.2224965543,0.1073413342,0.0204158705,-0.0038787983,0.153313905,-0.3694169819,0.1006374434,-0.6156262159,0.2590619326,0.0477673896,-0.0527918786,-0.1532708406,-0.3891787231,0.332998544,0.1165231615,-0.141970709,0.3460913897,-0.0001264128,-0.0468588397,0.4429562688,-0.1456463486,-0.0567967184,-0.1349925697,-0.0857358053,-0.2637895346,0.1227670461,-0.2072709054,0.021981135,0.017595334,0.0102824224,0.2515926957,-0.511700213,0.2599984705,0.029450953,0.6438952088,0.3012346625,0.2767096162,0.4791561365,0.0943052769,-0.232714355,0.0265360847,-0.2116011381,-0.2294592559,0.3904607892,-0.0502416641,-0.3566617668,-0.2715986371,0.2123684138,0.038955301,-0.00754125,-0.1563342959,-0.2058854699,0.4156657755,0.4066308141,-0.2864206433,-0.3169647753,-0.2571143806,-0.0219363067,-0.1467234939,0.2574432194,0.3813537657,0.047724735,0.0278937221,-0.3996604979,0.1394699067,-0.0945347175,0.1609137356,-0.3527618945,0.3311661184,0.4382251799,0.1159461886,-0.0664291456,-0.0793702677,0.2227964103,-0.3506995142,0.1299065799,0.1476262659,-0.2760589719,-0.2300761491,-0.2641409039,-0.1096549928,0.449169606,-0.0774305761,-0.0678217337,0.0793575048,-0.1761967987,0.1970984489,0.1830151379,0.1964344978,0.1761818528,-0.0575269237,-0.0108308438,-0.0116140163,-0.3057376146,0.2463934124,0.2433477193,-0.2666074038,0.1926814467,-0.0420936607,0.5749986768,-0.1320970803,-0.3799809515,0.0747070834,-0.6538074613,-0.1110766307,0.1407732069,-0.0510298796,-0.0291147139,0.353911221,-0.0687290877,0.3858068585,-0.2238057852,-0.3407226503,-0.0767959431,0.039024964,0.0556021631,0.0054286011,0.1059087589,0.0958222374,0.1772526205,-0.0211551748,-0.3178199232,0.161781013,-0.0995850116,-0.1669784635,-0.098948136,0.0823761672,-0.1805225313,-0.3181404173,0.1006147936,-0.3788382411,-0.2713882625,0.1607512087,0.0105452901,-0.2456015497,-0.1364842951,0.0304472204,-0.4371491969,-0.091694966,-0.0618584827,0.2654531598,0.2157913297,-0.014824409,0.0835113078,-0.1605887115,-0.0200645495,-0.2809292376,0.1177877411,0.1437309235,-0.0858941302,0.0024369627,0.1830232888,-0.2702117562,-0.0721942112,0.2852176726,-0.0352928489,0.0773441717,-0.1702269316,0.4980225265,0.378626883,0.0348261409,0.0589954667,0.0329786465,0.2058920562,0.0255773235,0.1180749089,0.1201005504,-0.2976595163,0.0073136995,-0.1126278043,0.3531253934,-0.1579978317,-0.0949935615,-0.1193357408,-0.3043027222,0.6342318058,-0.1070511788,-0.0537290908,0.1362027675,-0.1019034907,0.3085007966,0.2861429453,-0.2093664706,-0.1716703176,-0.0032346812,0.2184511721,0.2890660763,-0.4328747094,-0.4184077382,-0.5766484141,0.1472960711,-0.3304214776,0.5291690826,-0.0426753014,-0.3846941292,-0.1016397327,0.1355674714,0.1801891774,0.1740861386,0.0382286087,0.1262745559,-0.0057472298,0.1624972522,-0.0510249361,-0.0948095545,0.0513381176,0.160520032,-0.2124759853,0.5435077548,-0.0134333847,-0.2456252277,0.0123080881,0.5332690477,0.2788841724,0.0309909657,0.0797271729,0.1892461777,-0.0019995682,0.0271290317,-0.183599934,0.3623335958,0.0204167943,-0.0936032981,-0.1447377354,0.3266547024,0.1559484452,-0.0892912298,0.0254457574,0.5524482727,0.0685351193,-0.1579214931,0.2101214081,-0.0892066956,0.0011234476,-0.1303198189,-0.0838992298,-0.1063473597,0.0964014009,0.0178030524,-0.5440840125,0.567660749,-0.6094178557,-0.1653397679,0.4562682509,-0.2197403461,0.2057948709,-0.2460896224,-0.1635510623,0.102707684,0.3452097476,-0.4267811179,0.1155703813,0.1122339293,0.0700915754,-0.0156854969,-0.3419274092,0.0964458212,0.3073928654,-0.0566928498,0.1302871555,-0.0382126942,0.2196430266,-0.1677629054,-0.3320037723,0.1360586882,0.1255233735,-0.0951926559,-0.3470984697,-0.0371200591,-0.3094110489,-0.0912447497,-0.4194905162,-0.1826290637,-0.4198156297,-0.4413434565,-0.1236550212,0.7008601427,-0.0177299306,0.1978925467,-0.6700432301,-0.2774953246,0.0492400676,-0.3995980918,0.011442597,0.0148913879,-0.0733688772,-0.0870078579,0.2734766901,0.2917412817,-0.1897356063,-0.0008335778,-0.1651877761,-0.2308116704,-0.7176827788,0.1102686226,-0.3910573125,0.0441698767,0.2752370834,-0.0223062113,0.2003913969,-0.2722445428,0.1678510308,0.2062655836,-0.3991113305,0.4627359509,0.1078234687,-0.3002520502,-0.0092078708,-0.1099070311,-0.0276114792,-0.2135955393,0.0942308828,-0.0862506852,0.0614141598,0.3091311157,-0.004426809,0.0199072864,-0.1430472732,-0.0395360515,0.1178778857,0.4055967927,0.5117624998,0.26081267,-0.0915239081,0.2016879767,-0.2428741902,-0.2411470413,0.2358126342,0.040437378,-0.1346413791,-0.2199570835,0.5386496782,0.2356010526,-0.5058151484,0.3071929514,0.3656333089,0.1072588339,0.0360254608,-0.2801835239,-0.0534452721,0.442324549,0.2758016288,0.0487580784,-0.1615020186,0.0787836164,0.1008439139,-0.1156169325,-0.2195214629,0.5805271864,-0.2571693063,-0.109842822,0.1961529702,0.0343593955,-0.16452685,-0.0097312871,0.1441927999,0.1660127342,-0.2098439336,-0.1936238706,0.1229808256,0.1239542663,-0.1897653043,-0.1696639061,0.1699931026,-0.2906679213,0.3181858063,-0.1007414237,-0.1873039454,-0.1924325228,-0.3460583687,0.0191253759,0.1727534384,-0.0420382321,0.1250179112,-0.6560661793,-0.0469099507,0.1015988067,0.3088893294,0.1387881041,0.4464935362,0.1088171974,-0.4637120068,-0.1776310503,-0.0819064006,0.5365412831,0.0638087541,-0.0584613569,0.2199558467,0.1883281171,-0.2731577456,-0.2824517787,0.0083973529,0.073803246,0.2962682545,0.2742966712,-0.6037446856,-0.1916427612,0.3171719313,0.3793386221,-0.0011060352,-0.0368858315,-0.1793114245,0.0075943177,0.0369182117,-0.1407228857,0.0038686539,0.2534176409,-0.3429115415,-0.0489150807,-0.3486585319,-0.0808285475,0.2886826098,0.4028430283,0.312354058,-0.3529822826,0.0914018378,-0.4784142077,0.1507918388,-0.0419844016,0.4972965717,0.4666127861,-0.4932498038,-0.4400035739,0.104179047,0.3715502322,-0.0567502864,0.1880355328,0.2884448171,-0.1938230246,-0.2297523916,-0.1888608932,0.0138052497,0.3524463773,-0.1079066917,-0.1447208375,-0.0419296063,0.5425732136,0.0772326887,0.1056776345,-0.6856749058,0.3390678167,-0.3440513313,0.3295375407,-0.0359517597,0.9693979025,-0.0000762053,0.0945802033,0.2922485471,0.1335750669,0.4692514837,0.0192873664,0.1283658743,-0.3416178823,0.163527593,-0.1912623495,-0.2569366395,0.121740073,0.2627844214,-0.209023267,0.226187259,-0.0477842465,0.2124791145,-0.0430897288,0.1805104166,-0.0128857708,0.3465476036,-0.2673856914,-0.1008255035,-0.0844019502,0.1316176504,0.1436578333,-0.1747906208,-0.4823785424,-0.0664012656,0.0364054516,0.138597697,0.0684253275,0.0052224207,0.3954237998,-0.2528504431,-0.0209834743,0.3400256336,0.023033604,-0.3180303574,0.0055429828,-0.0876056775,0.2474800348,-0.1485331804,0.1685189754,-0.1524543166,0.2502440214,0.1381442398,-0.2733095288,-0.0980507582,-0.1025735512,-0.136797592,-0.1435686201,-0.0960542485,0.0197395794,-0.2391243726,-0.2245329916,-0.0364589691,-0.0671480298,-0.0688887611,-0.0574226491,-0.181526795,-0.0571524277,0.0491878614,0.0337344855,-0.1626971066,0.0789013356,0.1143804044,0.2522880733,0.6020578742,0.4128520787,0.0101298187,0.0075302147,0.016259877,-0.2461384684,-0.248567,-0.3736048937,-0.096110709,-0.0684267208,-0.2595949173,-0.2359576076,0.6898490787,0.1070465595,0.2580772042,0.1179959401,-0.1757451743,-0.0688841492,0.3435314,0.2259231955,0.2820385098,-0.0227693338,0.5188938379,0.1511959434,-0.0489223115,-0.0867049396,-0.1001048908,-0.4281012416,0.2733705044,0.1423183382,0.0057457848,-0.239862144,-0.3628188372,0.0821276531,-0.2926056385,0.1691567302,0.0212260652,-0.0225909688,0.228843227,0.3204802871,0.1147127077,0.0232967883,0.0890408829,0.3312734962,-0.0993806794,-0.0897174925,0.1752562821,-0.1079478711,0.356703788,0.2367999852,0.2442308664,-0.4354184866,0.0297569502,-0.0931765512,-0.0452798679,0.0526613779,0.3269843161,0.0137711931,0.1752380878,0.0000456941,0.1150661036,-0.0066957879,0.0290843341,-0.0375327654,-0.0877785012,-0.2991589308,0.0395985879,-0.0301435273,0.4621220231,-0.043208912,-0.1424031109,0.3055092096,0.0216555782,-0.381795764,-0.0124108046,0.2581444979,-0.3631215692,0.0267434865,0.181829527,0.3709528744,0.3668279946,-0.3642131388,0.0785256848,0.1434004754,0.1336973011,0.177464053,0.2514982224,-0.0073595229,0.1600061655,0.482205838,-0.0018297482,0.0793426856,-0.1413375884,0.4614363313,0.2659011483,-0.0982790962,-0.1667800546,0.4742590785,0.158373937,-0.3313771188,0.0797289833,0.4608873725,0.0788980946,0.1961143464,0.3415955603,0.0701612905,-0.2417647392,-0.17226322,0.0930664316,-0.1025390625,-0.1038741693,-0.2498863637,-0.2854084373,-0.092943877,-0.1711246222,0.0333198011,-0.1871662438,0.5072126985,0.076171726,-0.050884597,-0.2490593642,0.4247331917,0.2498681694,0.1526672095,-0.214864701,-0.0116734486,0.0636816695,-0.2632588744,0.2762016952,0.2402351648,0.2737585604,0.2493442148,-0.3756702542,-0.3619799316,-0.0038706516,-0.0174831599,0.0782139823,-0.0454818904,-0.190946728,0.1720745265,0.2336172611,-0.0049082381,0.1051338464,-0.0152658429,-0.17579934,-0.2990526855,-0.2980513573,0.5810395479,-0.0425819494,-0.2818369865,-0.0136865238,-0.0273596123,-0.0243216176,-0.2647331059,0.1848090291,0.2367518693,0.2285263985,0.1112204716,-0.0135025922,0.1970384866,0.1022106856,0.1120023206,0.13823241,-0.0138801094,0.2896609008,-0.5459221005,0.0369277038,-0.0639393702,-0.2957056165,0.1417855471,0.2952887714,0.0486421511,0.0740769356,-0.3884553313,-0.0020330837,0.1797571033,0.0071678376,-0.1268198192,-0.15685983,0.1063188389,-0.3369936347,-0.2708660066,-0.2781214714,0.3784303367,-0.2183064669,-0.0430415757,-0.0471045375,0.1650626212,0.4612142146,0.260065943,-0.0217502899,-0.0121351387,0.7294693589,0.1243342236,0.0588451922,-0.1143428013,-0.0374013707,-0.4124180079,0.1586294472,-0.2737038732,0.2427846938,-0.046895057,0.650971055,0.0292013679,0.4432554841,-0.1011844575,-0.3270037174,0.0743243098,0.1742576659,-0.3667991161,0.0924917534,0.1581842154,0.0422303788,0.0235246215,0.0521462075,-0.0285734721,-0.4694264829,0.1731987447,-0.03952685,-0.2418798506,0.0873610452,-0.1123388857,0.148655057,0.1342154443,-0.471054554,-0.2059438527,0.3357712626,0.0273692012,0.1469393969,0.107142888,0.244554773,-0.1285234839,-0.1475846469,-0.023469232,-0.2351099849,-0.0302709807,-0.0622365624,-0.045740217]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/491","title":"No 0.4.0 release on GitHub","comments":"I did the release on github, and updated the doc :)\r\nSorry for the delay","body":"0.4.0 was released on PyPi, but not on GitHub. This means [the documentation](https:\/\/huggingface.co\/nlp\/) is still displaying from 0.3.0, and that there's no tag to easily clone the 0.4.0 version of the repo.","comment_length":15,"text":"No 0.4.0 release on GitHub\n0.4.0 was released on PyPi, but not on GitHub. This means [the documentation](https:\/\/huggingface.co\/nlp\/) is still displaying from 0.3.0, and that there's no tag to easily clone the 0.4.0 version of the repo.\nI did the release on github, and updated the doc :)\r\nSorry for the delay","embeddings":[0.1097745523,-0.3050479591,-0.0885717496,-0.1643030196,-0.1300125867,-0.0987591743,0.0865198746,0.4773237109,-0.1674301326,0.2169478834,0.0148458024,0.3269321918,-0.1404925585,0.2279849499,0.4223875105,-0.0798372179,0.0108980471,0.2660887241,0.1807587296,-0.1086733118,0.0090757478,0.350238651,0.0049004401,0.0948055014,0.0502954274,0.0198652539,-0.0938507393,-0.0042202044,-0.2600695491,-0.3923913538,0.2254684269,0.0499741249,-0.1221126541,0.1499101967,-0.0001023581,-0.130769968,0.5131497383,0.0419392288,-0.0584902838,-0.3072885573,0.1793330461,-0.3557259738,0.087325789,-0.0528424866,-0.1879538298,-0.2122714818,0.2089814842,0.4363964498,0.0834233984,-0.0015605966,0.3164174855,0.2422754765,0.1821965724,0.128685087,-0.0037896198,0.0080133826,-0.1833246499,0.2676139176,0.082905449,0.0138120856,0.0411561243,0.285253346,0.0403593928,-0.2667756975,0.0017215441,-0.1252865493,0.212435469,-0.0066953753,0.1419259608,0.4129835665,0.198906824,-0.0550982691,-0.2195722312,0.0991801023,0.2249660492,-0.4488681257,0.0846580788,0.2182836831,-0.0215764362,-0.0697396547,-0.277572751,-0.3081166446,-0.2103670835,0.1567332447,0.1761054248,0.3889921308,-0.1189764217,-0.1299042255,0.3715721071,-0.1988338977,-0.3617920876,0.3160563707,-0.1641805619,0.397778064,0.3813743591,-0.0941342339,0.1411082,0.1664628386,0.2239449322,0.1943911761,-0.2435500473,-0.0012101909,-0.0434066281,-0.012569786,0.0547221862,0.1040658429,0.3095617294,-0.2089109421,0.27014938,0.3774684966,0.3149372637,0.0360752158,0.0144854123,0.1346196383,-0.2623892426,-0.1846634001,0.0161493476,-0.200036332,0.0824054703,-0.1196521744,-0.3849448562,-0.041502025,-0.2256126553,0.1304145008,0.0605039597,0.1848663986,0.2368445843,-0.0877474323,-0.296110183,-0.1991691887,-0.2883899808,0.1181430966,-0.2719077468,0.0692076385,-0.1561517417,-0.204786554,0.5439360142,-0.0663965717,0.1074594781,-0.0151376612,-0.379457444,0.3614737391,-0.219671756,0.2343245596,-0.1971852183,-0.124938339,-0.1101791784,-0.2274560928,-0.1792049259,-0.1253212392,-0.0893826187,-0.120694302,-0.2004615217,0.2397440821,0.0109059485,-0.0974185914,-0.1917046458,0.1730080694,-0.1352958083,-0.1200354844,0.1137594134,0.0372641459,0.0472297296,-0.1096290797,0.2630980909,0.1857754588,0.2908222377,-0.5492760539,0.0139395315,0.1185525358,0.0319557153,0.3558286726,-0.0319329761,-0.0431015044,-0.0484874249,0.2081918716,0.1572487652,-0.3020326197,-0.2928094864,-0.0664919987,-0.1253111511,0.0040421044,-0.0721188411,-0.2471802831,-0.0382762849,0.0733454302,-0.0109236762,-0.0481399707,0.1043432131,0.0405375771,-0.2897572815,-0.338878721,0.1504158825,-0.0068274508,0.035345085,-0.0763814524,-0.0652317479,0.1903578937,0.3292993009,-0.0432044044,-0.0597374104,0.0709005073,0.3344672024,0.0940981433,0.0377267338,-0.0419112109,-0.0775016248,-0.1542983055,-0.0709459856,0.4672296643,0.1858744025,-0.0400371775,-0.2073463053,-0.2887874842,-0.0182889327,-0.5060579777,0.2743088305,0.1064518169,0.0992989317,0.5235136151,-0.1043104604,0.1433347315,-0.213811174,-0.0019408015,-0.1942864656,0.1202153713,-0.3706581593,-0.0740378872,0.0327487625,0.5490599275,0.0392470807,-0.1397407502,0.1947565079,0.1055673733,-0.1941973269,0.1065057889,0.0679442883,0.2603227496,0.4372475147,-0.4418745637,0.0866045728,0.2407047004,-0.1129365042,0.2069445103,0.251131773,0.1673723906,0.3672575951,-0.0682953447,0.2847189605,0.1488604397,0.2286664248,-0.0552586243,-0.228206113,0.0107405037,0.2275051326,-0.1378390491,0.4040344656,-0.3127473593,-0.34728387,0.262714088,0.5641533732,-0.1695206761,-0.0861031562,0.113902092,-0.0907971337,-0.1966735423,0.0440082699,-0.1926125139,-0.2064304501,0.2989940345,-0.024359934,0.139004305,-0.0141642056,-0.144193694,0.1642326117,-0.2079316825,-0.000511441,-0.0757871419,0.1206286475,-0.0015773283,-0.354129225,0.1376797259,-0.167603597,0.0407221131,0.0345908403,-0.0687099621,-0.3943232894,-0.3571660221,-0.287330091,-0.3590350449,-0.6427389979,-0.2138649225,0.1016211808,0.1223013923,0.0134263337,0.2458707541,0.1434026361,0.1263971031,-0.1399241239,0.2547169924,-0.1460592896,-0.0223506838,-0.2459363341,0.1747088283,-0.1806385666,-0.1725101471,0.2963793576,-0.3268965781,-0.1507420093,-0.2478049248,-0.7388551235,0.1447146535,-0.1920887083,-0.2372150421,0.2504742742,-0.0594621487,-0.080723539,-0.1671815813,0.0646215081,-0.5400335789,-0.4812393188,-0.1131338105,-0.1418080181,0.0689396411,-0.2940815389,-0.3889843225,-0.1229447648,-0.3170722127,0.5989442468,0.0573755838,0.1157954782,0.5650062561,-0.147912398,0.3627459705,-0.1518739611,-0.0098365797,-0.2061063051,-0.0108421007,-0.111003682,-0.2450450957,-0.3855564296,0.0723696873,0.1048372984,0.2078466415,-0.3038978577,-0.4259847999,0.027693674,0.1715291291,0.2407047898,0.2690120935,0.0331775546,0.5288996696,0.0086928634,-0.1749554724,-0.0422586091,-0.2454692274,0.116104655,0.1143522039,0.3505427241,0.1157255471,-0.0185913201,0.0661516786,0.4575294554,0.1239591092,0.0245193727,0.1687532812,0.0751385987,0.5493571758,-0.2809480131,-0.1302001029,0.2312648743,0.2429668456,0.2144328058,0.4412685037,0.160688743,-0.2779148519,-0.0919688642,-0.3134173751,-0.1414399594,-0.1069787145,-0.3308624923,0.1480096132,0.0445893593,0.0287358388,-0.0184259359,0.1055414453,-0.0577918775,0.1747833788,0.3718698621,-0.0636548772,0.2283624858,-0.0832301974,-0.1287064105,-0.5055819154,0.0161579307,0.2024766952,-0.083603777,-0.0664268658,-0.2352582514,0.0601232871,-0.2679367065,0.1807168871,0.1352591664,0.0689557642,0.1784606576,-0.0017140832,-0.0325669013,0.014150572,0.1253139824,-0.1253104657,0.6306692958,0.2593958378,-0.307708174,-0.3269405067,0.394009918,-0.1323353648,-0.0988492742,0.2040722817,-0.2355337292,-0.4011268616,0.1220994145,-0.0779909045,-0.1810360104,0.1697601229,-0.115055941,0.2597642839,0.3329612315,-0.2047816366,-0.2088852376,0.2124983817,0.1674743444,-0.133961916,-0.1140326187,0.3180952072,0.1059656888,0.16154401,0.152409479,-0.108391583,-0.0609077476,-0.0609293655,-0.0114600826,0.201126039,0.1668871939,0.3579171896,0.1084419414,-0.0139036933,0.100376226,-0.0636817887,-0.2641143203,0.3138152957,0.0771819204,0.0081212018,-0.1721118987,0.3693093657,-0.1572976261,0.0606560744,0.2674817443,0.6381277442,-0.0449528433,-0.1045829952,-0.1302860081,0.953966856,0.1647224277,0.2874431312,0.2908834815,-0.3708800375,0.4390354156,-0.3018469214,0.1329322904,-0.3388608396,0.3271322846,0.1102457717,0.2086589336,0.0560023896,-0.2708139718,-0.2874607742,0.1225598529,0.1556996554,-0.0971819684,0.274913162,0.1121596918,0.1008199453,-0.0342895351,-0.3664432764,0.2223825306,-0.0100636687,0.2919954658,0.0386673883,-0.1770448238,-0.0510831699,-0.2008706033,-0.2632407248,0.2637700737,0.2101063579,0.0866717473,0.3839967251,-0.06181027,-0.249815315,-0.2748511732,0.1541501433,0.063792102,-0.2384619117,-0.0111059081,0.0196513552,-0.3123302162,-0.0727036074,0.2639507949,0.3362129331,-0.2214405686,-0.1780579984,-0.0249433406,-0.0148849208,-0.2538127005,-0.1790101677,-0.1139464229,-0.3573061526,0.2106627524,0.2832039595,0.3498182893,0.0044999896,-0.2452440709,0.2353584915,0.1341016889,-0.3595988154,-0.4762401879,0.2902620435,-0.0678669065,-0.0760647431,0.2286474407,0.1625002474,0.1668436676,0.2511849999,-0.0128596108,-0.1624667197,-0.2838717997,-0.1994093657,0.2845627069,-0.2356564403,-0.0301396735,-0.0812557563,-0.2637270689,0.1230460778,0.3419757187,-0.1201492101,0.0312552229,-0.3281012774,-0.0628147796,-0.3986290097,0.3053176999,-0.3547998667,-0.0592947081,0.3962648213,0.2847808003,0.3192140758,0.0803471282,-0.3862119913,0.0569406934,-0.2516603768,0.1005703285,0.3516659141,-0.3483695388,0.4213495553,0.1544224173,0.2180861831,-0.0490070954,-0.3039083481,-0.2468560636,-0.3900024891,0.0129166134,0.0842238814,0.0247717351,0.1852525324,0.0293016769,-0.1629063189,0.1576490849,0.1123787835,-0.0457968526,0.0867875144,0.2583108544,0.2601886988,0.116556406,0.0919354782,0.049188111,0.1172783002,0.2615178227,-0.1100450456,-0.0990911871,-0.0938261598,-0.0408470258,0.0406649411,0.030642841,-0.0057807397,0.1367981136,0.1692366153,-0.183579728,0.1041291952,-0.323900491,0.3886856139,-0.1118805408,0.0896470025,-0.0470413975,0.3413604796,0.2172344774,-0.3115528822,-0.0937284753,0.0950352252,-0.1359078437,0.0618778169,-0.0103676897,0.1412375569,0.035141781,0.2352087051,0.0181964263,0.0526243486,-0.4260637462,0.1352793574,0.0618010387,-0.1729567498,-0.0681790635,0.1155299321,-0.0718744919,0.0936689079,-0.2527161837,0.0877648667,0.0662895516,0.1149832234,0.1882536262,0.1526401788,0.1719412953,-0.2576741874,0.0418935679,0.3465334773,0.2193929404,0.0997698233,0.4396777749,-0.0215747356,-0.2867356837,-0.0852374211,0.356210202,-0.1950461119,-0.1695654243,0.1172563732,-0.0536040291,-0.1070455909,0.0169252064,0.1639561802,-0.1174354255,0.3473332226,0.1311643124,-0.0722794086,-0.2934361696,-0.5230307579,-0.0085486108,0.3465904593,-0.0338200554,0.2226702422,-0.1444329768,-0.2287691385,-0.2946460545,0.516330421,0.3497287035,-0.0294393171,0.0149057228,-0.1145617589,0.0986961126,-0.1106768847,0.0415712483,0.21072644,0.4277881384,0.0135360854,-0.072816357,0.1064063162,-0.2462691963,0.1955023706,0.1697788239,0.1245211437,0.0172020718,0.4382292926,-0.2332436591,-0.2389855385,-0.4067927897,0.2750797868,-0.4260049462,0.4545690119,-0.0158397723,0.1323325187,-0.1265829802,-0.2651042044,0.1389888078,0.2574091256,0.4421365559,0.1656535119,0.0395765677,0.2699441016,-0.3447089791,-0.0948262811,0.1524881572,-0.2356951535,-0.1537340581,-0.1378357261,0.1734246761,-0.0446181037,-0.1884547621,0.1801978499,-0.0332487598,0.3474586904,-0.3183619678,-0.2192788571,0.107491225,0.0731034353,-0.1810711622,-0.0453668274,0.0416041315,-0.072627902,-0.1101786494,0.1139271781,0.1941836625,0.366600275,0.0528934635,0.1189257428,-0.0451779738,0.0338155404,0.0821753144,-0.09143278,-0.6400892735,-0.2405815274,-0.0776412264,-0.3123889267,0.2235176712,-0.0625775158,0.324062407,-0.1591137797,-0.4136303365,-0.2498507351,0.4939595461,0.0433746763,0.0434581339,-0.2602043152,0.2506363094,-0.0942486823,-0.3265657723,0.1013785973,0.2748602033,-0.1918968707,-0.3348210156,-0.2755479813,-0.2184520513,0.5653696656,-0.6479350924,0.0795658976,-0.0208746716,0.1490263194,0.2167130411,-0.3155020773,-0.7494992018,0.1011732742,0.0981061012,0.2251111716,-0.3442803621,0.2452152222,-0.1376415342,-0.0555839576,0.0039039548,0.3601038754,0.3003624976,-0.1732675433,0.3295921087,-0.1269993633]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/491","title":"No 0.4.0 release on GitHub","comments":"Thanks!","body":"0.4.0 was released on PyPi, but not on GitHub. This means [the documentation](https:\/\/huggingface.co\/nlp\/) is still displaying from 0.3.0, and that there's no tag to easily clone the 0.4.0 version of the repo.","comment_length":1,"text":"No 0.4.0 release on GitHub\n0.4.0 was released on PyPi, but not on GitHub. This means [the documentation](https:\/\/huggingface.co\/nlp\/) is still displaying from 0.3.0, and that there's no tag to easily clone the 0.4.0 version of the repo.\nThanks!","embeddings":[0.102597177,-0.3189143538,-0.0914218649,-0.1304027587,-0.105916746,0.0324019194,0.1008440256,0.5270662308,-0.2064419687,0.2322422713,0.0247421842,0.3248824179,-0.1479322463,0.189981848,0.49399966,-0.0577390194,-0.0129243433,0.2527408004,0.2898946404,-0.0989583507,0.001360754,0.3206633627,0.1036469713,0.1322809607,0.0806212127,0.0750705302,-0.1015203744,-0.1028995439,-0.2486671507,-0.4138995111,0.2521547675,0.0289952271,-0.0960927531,0.1258235723,-0.000108978,-0.1391673982,0.504437387,0.0324757993,-0.0494054332,-0.3234680891,0.140457809,-0.3283075094,0.0879417658,-0.0666259974,-0.1368005574,-0.2492750734,0.1908358932,0.4418295026,0.0223844312,-0.0164556243,0.2758927941,0.2119890451,0.1430975944,0.116968818,0.0221572649,0.0366247706,-0.1728118211,0.3649065793,0.1149728596,0.0804958194,0.1116966382,0.2400978804,0.0083885863,-0.2872368991,-0.0165344216,-0.162736401,0.2648174465,0.0468895249,0.1800996363,0.5238986015,0.2491148412,-0.0627237856,-0.2056113034,0.1888357252,0.2495647818,-0.4218858182,0.0368560441,0.3052047789,-0.0331675299,-0.0526058413,-0.3091994524,-0.3730955422,-0.2898273766,0.1397529691,0.1658659875,0.4636186063,-0.1092152447,-0.0907096788,0.4064543843,-0.1483881325,-0.3956212401,0.3586896062,-0.2453745902,0.4179316461,0.5035131574,-0.195767954,0.0768727511,0.1385650784,0.2554308474,0.2427877933,-0.298483789,-0.0639237165,-0.0435014181,-0.0242921263,0.0437253006,0.0537166856,0.3417479992,-0.2457643747,0.2721821964,0.3996718228,0.3572723269,0.0607473925,-0.0387897231,0.2111050636,-0.2625677586,-0.2283099741,0.0430580862,-0.187689051,0.1198775247,-0.1379011571,-0.4639099538,-0.0492057241,-0.2567237914,0.1135348752,0.0830594823,0.1731575578,0.2467565984,-0.1684223861,-0.2331860065,-0.2860478759,-0.2448035479,0.1151394174,-0.2102530003,0.0279922076,-0.2582589388,-0.2927747071,0.5455408096,-0.0791900083,0.0628385767,-0.0667760968,-0.393419832,0.3676623404,-0.2599868476,0.2916986942,-0.2708552778,-0.1336937547,-0.094452031,-0.2200831771,-0.2300673574,-0.0862372965,-0.1097079664,-0.1890876293,-0.2654557526,0.1743678749,0.0099495286,-0.0716363415,-0.2897734344,0.1318877935,-0.2211062759,-0.0565425418,0.1784131378,0.0638817847,0.0729482621,-0.0837112367,0.2112747729,0.1900354177,0.2383177727,-0.645270884,-0.0575957373,0.059870299,-0.0322184153,0.3452118933,-0.0401319414,-0.0371165723,-0.0911474153,0.1051998809,0.2034024298,-0.3502502739,-0.2071926594,0.0002165051,-0.0997458622,-0.0042109229,-0.0550371222,-0.2034797221,-0.0926143378,0.0561160631,-0.0498744585,-0.1537473351,0.0780297443,-0.0561843067,-0.1593245417,-0.4058892429,0.1651533395,-0.0299847592,0.0768645927,-0.0779413134,-0.0952992961,0.1869044304,0.3230402172,-0.0161954965,-0.1417930871,0.0442288518,0.3997039497,0.1262935847,0.020568708,-0.0614216886,-0.0533894897,-0.1346270889,-0.0593777709,0.3599951863,0.1598887146,-0.0125271492,-0.199931249,-0.2544406056,0.0195470341,-0.5118592978,0.202625528,0.0791681856,0.1127088889,0.5140984654,-0.1080210283,0.1419098824,-0.266038388,-0.0210903361,-0.1570889652,0.1170041263,-0.2798521221,-0.0039684502,-0.0450268351,0.5799440145,0.0497388318,-0.1486122906,0.2430239171,0.0850753039,-0.1573781073,0.0808317587,0.0768423751,0.2643696964,0.4072155356,-0.4788786769,0.0546533987,0.2665328681,-0.1352819949,0.2430439144,0.2996871471,0.1254481971,0.4242119789,-0.0201390777,0.2742373943,0.1946041882,0.1996032149,-0.0104834782,-0.203049764,0.0021021231,0.253223002,-0.0659606382,0.4530445635,-0.2797189951,-0.3120288551,0.2432749867,0.5795442462,-0.2506395578,-0.0970787257,0.1537241042,-0.09141545,-0.1559164077,0.0445272364,-0.1709288508,-0.2031067908,0.2548325956,-0.0371754542,0.1896934658,-0.0004980463,-0.1127466187,0.2111118883,-0.2122817189,0.0105881961,-0.1822805703,0.0828148946,0.0195913836,-0.2730160654,0.0958604142,-0.1575440317,0.0624752603,0.0574482009,0.0058243787,-0.4826205373,-0.358871758,-0.3711575866,-0.3512035906,-0.7194263339,-0.2079680115,0.0947362259,0.1115996912,0.0759598985,0.2107636482,0.1338190138,0.0986313894,-0.1733289212,0.2477752268,-0.1205588356,-0.0582415983,-0.2185338885,0.1261024922,-0.1859385818,-0.2550871968,0.323251307,-0.3280813992,-0.1757607013,-0.3212037086,-0.7397776246,0.1518592685,-0.131895557,-0.2437610775,0.2550938129,-0.0488741808,-0.0384659134,-0.1786967516,0.0356130414,-0.4952868521,-0.4862889946,-0.1258718371,-0.2033585906,0.1319027245,-0.2275194824,-0.3433361053,-0.1581766158,-0.2642326951,0.6038829684,0.0069670505,0.1392574012,0.5786946416,-0.216115728,0.4295647442,-0.1003182158,-0.0040944722,-0.196646452,-0.0179589801,-0.1878767163,-0.1808229685,-0.2660261989,0.1575370431,0.1586238146,0.2137844861,-0.2672394514,-0.4849357307,0.0708013028,0.2028850615,0.2306467593,0.2996018231,-0.0461783148,0.5537978411,-0.0137493266,-0.1141976193,0.0163587593,-0.1979799122,0.0265278034,0.0998540446,0.3418708146,0.1520854533,-0.0146021694,0.0613076538,0.5553973317,0.1793523282,0.0025023723,0.0980270281,0.058013685,0.5926529765,-0.2933063805,-0.0997658297,0.1430508494,0.2676169574,0.2654038072,0.4272514582,0.2050083876,-0.3010185659,-0.1034662649,-0.307979852,-0.0655752718,-0.0443811044,-0.3953476846,0.101383999,0.0844503567,-0.0234676804,-0.0180369802,0.1251612306,-0.0993224531,0.1531407386,0.4418348372,-0.0530400239,0.2302622944,-0.0602186471,-0.178763479,-0.4065766931,-0.0030766025,0.2387730777,-0.0742024556,-0.0721416771,-0.1949095428,0.08688467,-0.2380864918,0.1219607517,0.1015281305,0.0995324403,0.2457046509,0.0112589635,-0.0426185727,-0.0321644954,0.1567409486,-0.125456661,0.6944035888,0.2326188236,-0.2560304403,-0.3276251554,0.4804562032,-0.1464721411,-0.1256499738,0.2329678386,-0.2644886971,-0.3830582201,0.2216239721,-0.0767169073,-0.2383896858,0.1421402246,-0.1657108217,0.3268590271,0.3001070321,-0.1968775541,-0.2097491175,0.135474354,0.0903123319,-0.1059464887,-0.1524704099,0.2834610939,0.08764375,0.2121315747,0.1146845743,-0.0333564617,-0.1276208162,-0.0076674051,0.0340650268,0.3031983972,0.1957875043,0.3953767419,0.1267798096,-0.0008948182,0.0430029109,-0.1074810848,-0.3722331524,0.3022150099,0.059464477,-0.0456112847,-0.1577154994,0.4044290483,-0.1012403145,0.096991241,0.291728735,0.6351006031,-0.0301879719,-0.2116839141,-0.1524653584,0.9944585562,0.1101112738,0.2621447742,0.2641974092,-0.4491031468,0.4552725852,-0.2986944914,0.0760364458,-0.3394431174,0.3364833295,0.0445122123,0.1574631929,0.1264816672,-0.2804222703,-0.2880457342,0.1245221421,0.1140610799,-0.046359662,0.3879903257,0.046622362,0.0652661845,0.0364158824,-0.3801704347,0.1721948087,-0.0875227228,0.3047935367,0.0300449319,-0.1137292087,-0.0074655833,-0.165659979,-0.3876800835,0.2127515674,0.2780766487,0.1369823813,0.4097833633,-0.0454344787,-0.2398765385,-0.2903537452,0.1745948792,0.0627010614,-0.240876928,-0.0201682337,0.0018813356,-0.3547647595,-0.0530697741,0.2605133951,0.3276875615,-0.1985084414,-0.1905314922,-0.0027053251,0.027954381,-0.3477607667,-0.1700111479,-0.1122286543,-0.3771436214,0.236813888,0.2775099576,0.343105644,0.0478277616,-0.2263687402,0.1692358106,0.1196099147,-0.3793692887,-0.5730995536,0.2525703311,-0.0613488778,-0.0380167589,0.1989709586,0.2275759578,0.1914633363,0.1827197224,-0.0537139103,-0.15131782,-0.2559074461,-0.0881881639,0.2829969823,-0.2363899946,-0.0048484844,-0.0613084696,-0.2136826515,0.0853600353,0.290302068,-0.1840700358,0.0787581131,-0.3594378829,-0.0522355773,-0.3737532794,0.2904631793,-0.3423633575,-0.0148046082,0.4303411245,0.2097094953,0.2804424465,0.0333671346,-0.3134224713,0.0775363892,-0.2408474535,0.1089970395,0.3259932101,-0.3064300418,0.4068439901,0.161337629,0.1556371152,-0.0741145834,-0.3083829284,-0.189187482,-0.4398370981,0.0297272354,0.023042433,0.071034804,0.2272569239,0.0089077139,-0.1332447827,0.1898778081,0.12429481,0.0067974138,0.1152632758,0.2857635319,0.3016123176,0.1246340498,0.009779443,0.0789509937,0.1211145446,0.3995328844,-0.0539069884,-0.0995980427,-0.0839346573,-0.0488277003,-0.0684996098,0.0609353669,-0.0637737364,0.083079651,0.1572601944,-0.1464657485,0.1443136185,-0.3757692873,0.3151191175,-0.162366733,0.0787194818,-0.0177903548,0.3756830692,0.145018816,-0.3024832308,-0.1376949698,0.1497020423,-0.0901474208,-0.0037557732,0.0065686377,0.1466800272,0.1059335023,0.3394371867,-0.0436157323,0.0290217139,-0.4164650738,0.22603935,0.0988224372,-0.0635407344,-0.0704642609,0.0560032129,-0.0290599279,0.115957208,-0.404395014,0.106789276,0.0656996742,0.108100459,0.1235304102,0.2358576506,0.1782613695,-0.3514133394,0.0823060647,0.3234928548,0.2612126768,0.0920623615,0.4751371741,0.0366661176,-0.300517261,-0.0924393982,0.3450289667,-0.2206608206,-0.1866651624,0.1963268071,-0.0294609107,-0.1059274077,0.095505774,0.2316699028,-0.1718136519,0.3419370949,0.1643057764,-0.1076841503,-0.2641296685,-0.5133778453,0.0090541253,0.3367300332,-0.0170049053,0.164040044,-0.1718665808,-0.2234795988,-0.2737706602,0.5090317726,0.3229248822,-0.1235986874,0.0529215373,-0.0933650732,0.1170440465,-0.1368620247,0.0931592286,0.1440073997,0.4201311171,0.0524283871,-0.1283063889,0.0616146401,-0.1583227217,0.195906058,0.2042183429,0.1306305826,-0.0438394174,0.5524852872,-0.2753658593,-0.2423334867,-0.4242610037,0.3719963133,-0.3707495034,0.5046244264,-0.0546651259,0.2100565583,-0.163602829,-0.1854290664,0.1024427786,0.2251119912,0.3638076186,0.1828972697,0.0375823081,0.326189369,-0.3778406382,0.0313029066,0.0647515953,-0.2829372883,-0.1411540955,-0.1084090844,0.1429117471,-0.0925330818,-0.2067071944,0.2160489857,-0.0754432678,0.3842659593,-0.2786131501,-0.1814004034,0.0089364639,0.1884468198,-0.1723127365,-0.0919545591,0.0293317623,-0.076006569,-0.1068360806,0.0256231092,0.2066315562,0.3692269325,0.0882759467,0.1640555561,-0.0240166895,-0.039690949,0.0908309519,0.0096078264,-0.6122999191,-0.2427603453,-0.0578636713,-0.2355371565,0.261251837,-0.0778996646,0.2396171689,-0.215865463,-0.414478153,-0.2225368619,0.5202875733,0.0211208817,0.0701085478,-0.263689667,0.223247543,-0.2067176551,-0.3037778139,0.1375869513,0.2304023504,-0.2690447867,-0.3416749537,-0.2533715963,-0.1207700744,0.549086988,-0.6042456627,0.0737414211,-0.0700733662,0.1491204798,0.2041278034,-0.2871294022,-0.7376785278,0.0840232968,0.1536599249,0.2368463874,-0.3714280725,0.2589824498,-0.1196193397,-0.0318596549,0.0056712036,0.3683803976,0.2700661123,-0.1816470623,0.3556435704,-0.1219635829]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/489","title":"ug","comments":"whoops","body":"","comment_length":1,"text":"ug\n\nwhoops","embeddings":[-0.0437335931,-0.4741887748,-0.3003547788,-0.1629350483,0.0372604728,-0.0372374021,0.2043906003,-0.0019466456,0.0315277949,0.1998867244,-0.1424533725,0.0655031055,-0.2205738723,0.2943479121,0.120513469,0.172912702,0.0708731934,0.0993628204,-0.1823475212,-0.2225968987,-0.130149141,0.040563792,-0.2111795247,0.1559441984,0.1208760589,0.1072827876,0.0032098226,0.2127170116,-0.4563615918,-0.034643352,0.0165239777,0.1705389619,-0.2063904405,0.264026016,-0.0000878093,0.0126513764,0.3597921133,0.0729773641,-0.2168055922,0.4580675662,-0.2148250341,0.2638809681,-0.3455985487,-0.0968011022,0.1008037478,-0.0850233212,-0.0350264274,-0.1757781655,0.0832863227,-0.2039538026,0.3940430582,0.074311018,-0.2154120654,-0.0616896525,0.2880034447,0.0153245898,-0.166237846,0.0616783611,0.25867787,-0.1615466326,0.1902516484,0.3292347789,-0.1588876992,-0.2584008873,-0.2828766108,0.0185421519,0.497629106,-0.0110662291,0.1027739197,0.0289559085,-0.1471133232,-0.0211334694,-0.1233600676,0.1610824317,0.0143270604,-0.1940425932,0.0361976251,-0.035722632,-0.0795215517,0.1578166038,-0.2808706164,0.0174016431,-0.0567696057,0.357375443,0.0685096234,0.4487261772,-0.1053036302,-0.0999850184,0.2034897506,0.1237521693,-0.228637293,0.0208590496,-0.03816377,0.1131238416,0.1296753883,-0.2042436302,0.2256514132,0.2724907398,0.0201368891,0.000650666,-0.0435418263,-0.0223200638,-0.0703345686,0.1566832662,-0.0330575854,0.0588102601,0.0962725952,0.0785322413,0.1694085598,0.1432204843,0.1349037737,0.0626626313,0.2574335039,0.0297010243,-0.0840968639,0.0829576105,-0.0074191699,-0.0152778951,-0.1459209472,-0.103936255,-0.3214039207,0.0317425653,-0.3057083786,0.1297670603,0.1003032625,0.1419233233,-0.0816384852,0.0985118002,-0.0421375297,-0.1868114471,-0.3403778076,0.0314551219,-0.0212156978,0.0439877883,-0.2481061071,0.033339709,0.1998006403,0.0031637046,0.1483091116,-0.1240902692,-0.1220684573,0.1152247265,0.0231835376,-0.1617975682,-0.0361182056,-0.0473178178,0.0230212398,-0.0766147524,0.1522587985,0.113484323,0.0382158495,-0.0948365554,-0.3537782133,0.4842614233,0.0703703165,0.0497278124,0.0357698128,-0.0555556901,-0.1171615869,0.1064464226,0.1325241774,0.1508265883,0.1449101716,-0.0845960751,-0.1934979856,-0.084959425,0.146370545,-0.2109400332,0.0445450395,-0.0492626391,0.363168031,0.2740994096,-0.1263138801,0.0850453079,-0.2304373831,-0.0446324013,0.0026378855,0.0288668033,-0.0507983118,-0.1671739221,-0.2695344985,0.1531313509,0.2103715688,-0.1152521223,-0.1767597795,0.0149409855,-0.0041658054,0.1452288181,-0.0145345721,-0.2418302745,-0.4350953996,-0.2537641823,-0.0045174677,0.0767621547,0.141152814,0.1233741194,-0.0493499935,0.0662548542,0.4393638968,-0.1901883632,-0.2248522192,-0.029596515,0.5213904977,0.0694475919,0.0288308915,0.105007425,-0.0114174914,-0.1064487994,-0.0677815452,0.3375574052,0.0903211683,-0.1621562093,-0.1194534153,0.2377368361,0.1027786359,-0.2287148833,0.3223569989,0.2831696272,0.0704061389,0.2150643766,-0.4414310753,0.0935094506,-0.1947179884,0.2110980004,-0.1233086362,0.0675557703,-0.1244945675,-0.0637161881,0.0244405326,0.0150878746,0.1813746542,0.0041591106,0.2334298939,0.0628275424,-0.1389272064,0.2019508034,0.2216833234,0.2437108904,0.1391994059,-0.2800720036,0.2148939669,0.2551388741,0.267783612,0.0733373389,0.2549755275,0.4662261009,-0.0583725907,-0.2441386878,-0.0001254437,-0.2864994407,-0.2513606846,-0.0155654754,0.0211671181,-0.2553600371,0.3830388188,-0.0884796456,-0.0358017571,-0.0792004243,0.0524031371,-0.0247284975,0.4124392867,-0.0639105365,0.0398375019,0.012129751,0.1867997646,0.0317918845,-0.0924808159,-0.3045138121,0.1658548862,0.2745782733,0.0876560137,-0.0258659888,0.0902649462,-0.2539885938,0.114267543,-0.0406764895,0.0253719222,-0.0194056127,0.051981654,0.12284787,-0.1887737215,0.1626519263,-0.256265521,-0.1443149149,0.0849698484,-0.0955053344,-0.0208497066,-0.4576800168,0.2308312058,0.0390063636,-0.1644046605,-0.3236570954,0.366615504,0.1799394041,-0.3069176972,0.2491975874,0.0205249712,0.1910999566,0.1895147264,0.3916306496,-0.2301266193,-0.1147523001,0.0316158384,0.3055847585,-0.2120417953,-0.2897216082,0.0761174858,-0.1229307726,-0.052886907,0.143644169,-0.4497205913,0.1900289208,0.1358484328,-0.4097729027,-0.1753462553,-0.1565844715,-0.0961436033,-0.0856111124,0.0893365219,-0.3264183998,-0.180543825,-0.0343786329,0.0132208057,-0.0509572439,-0.2734404206,-0.0146897919,0.0281014424,-0.1937847733,0.0390395932,0.1776456684,0.0411084257,0.1887172461,0.0072513535,0.1559181809,0.0621692948,-0.0966322199,-0.2221772373,-0.2917254269,0.2141328901,-0.0216672458,-0.0360821225,-0.2943255603,-0.0279093385,-0.3235606849,-0.3080965579,-0.6366423965,-0.125345856,-0.1576029658,-0.0687268376,-0.0538769476,-0.1257343441,-0.1245444641,0.0025909797,-0.5163835287,-0.0823611468,-0.1028211787,0.014385392,0.2097806633,0.1016123742,-0.0282182638,-0.1238856912,0.3325175345,0.4514806569,-0.0866181776,0.2484607399,0.071722731,0.1267457455,0.2088101357,0.1063401625,-0.113765493,0.1028844863,0.1706239879,-0.0536403656,0.1687076837,-0.0740579069,-0.1222453192,-0.0806709453,0.1042105481,0.033211112,-0.1977441162,-0.016454367,0.1480159163,-0.3006875217,-0.0166652035,-0.1785878241,0.2961150408,0.0249354616,0.3285531998,0.0239036381,0.1646786332,0.0308554675,-0.065307647,0.1685462296,0.2511477768,0.195661962,-0.0315416344,0.084502317,-0.1815800965,-0.3675236702,0.1683652401,0.1523882896,0.2268154472,-0.1105550602,-0.0333011299,-0.1404165328,0.0633396283,0.4071694911,0.1330967098,-0.2808613181,0.0369482599,0.3814129531,0.3610522747,0.0249680225,0.148885563,0.289773494,-0.2700679898,-0.0281420536,-0.0757454261,-0.3418281376,-0.1853839904,-0.2250660658,0.1290557086,0.200514257,-0.0731918663,-0.284793973,-0.035799738,0.1284313649,0.0252468009,-0.0484753698,-0.0217854194,-0.4748367965,-0.1980112344,-0.1887206435,-0.0365909338,0.1347164661,-0.2770785689,0.1683457196,0.3935664296,-0.2425721437,-0.1539837867,-0.0175411198,0.1554930657,0.2543304861,0.1836950779,0.1868779808,-0.0004481288,0.29876104,0.0911732838,-0.33375597,0.1621053815,0.0295221731,0.1525165141,0.197019428,0.176724866,0.073850818,-0.0649089366,0.4115722179,0.552200675,0.0847011581,0.2250091583,0.0954591185,1.195843339,-0.0160598401,0.0150613897,0.1525325179,-0.3276491165,0.1312006563,-0.0380184054,0.0747490153,-0.1460228711,0.2582153976,0.0152849527,0.0470920801,0.1734583378,-0.0004745425,-0.1285455823,-0.0682492703,-0.3274586797,0.2269608676,0.2470972687,-0.2852106094,0.180375874,0.0911331177,-0.0508207083,0.3412799835,0.0588763654,0.3017112315,-0.0682898238,-0.1512740254,-0.1257924289,0.018045878,-0.0583609082,0.1907788068,-0.3692568243,-0.0179527905,0.1272800863,-0.2687824667,-0.4043893218,0.042857863,0.0436830893,0.0317362621,0.2419072688,0.161079064,0.368494004,0.0642453209,-0.1015146971,0.1171675324,0.1417912692,-0.2053697854,-0.3467721939,0.2271955311,0.1084917784,-0.2317898721,-0.2366964966,0.0534512028,0.0938146561,0.0435206145,0.2445262969,-0.0534181893,-0.1400858015,-0.1893442273,0.3782517612,0.0396778286,-0.1399175227,-0.1099501997,0.2023342401,-0.204656601,-0.3025210798,0.164474234,0.0549390353,0.0113091171,0.0234976951,0.004780408,-0.3391121924,-0.3685500324,-0.1334164292,-0.2893949747,-0.1593419164,0.0456397869,0.0044545308,-0.1598891616,0.0239339005,-0.3307081759,0.0680078045,0.2852567732,-0.0621372722,-0.0611967519,0.0910929888,0.2363588959,-0.0361822434,-0.0551451705,0.0735602006,-0.0337257758,0.097147271,0.1783731282,-0.5886748433,-0.0688244179,-0.3334913254,-0.1889003813,-0.0556574985,-0.2527341843,0.0324334204,0.0203463826,0.5104571581,0.2304228544,-0.064210847,-0.4882788062,0.0582356863,0.096909374,0.223417148,0.0222984087,-0.070569247,0.1658992618,-0.1361043304,0.1644947529,0.1607033163,0.0974800289,0.3927878141,0.2807882726,0.0478528328,0.0730818585,0.0225934591,-0.2930831909,0.0770134106,0.3611442745,-0.1153204367,-0.2072908282,0.0810042769,-0.0856004879,0.3843890429,0.0191667061,0.4529004097,0.2017155886,0.3536551595,-0.3483547568,-0.0995001495,-0.2246514708,0.1612174809,-0.1800558716,-0.2713946998,-0.1327000409,-0.1108964831,0.4174641967,-0.0803432465,-0.0129744029,-0.0019395917,-0.127949357,-0.0114613939,0.2317105383,-0.2203080952,-0.0339113586,0.4150943458,-0.0323093422,-0.2213962078,-0.1742727309,-0.0765467435,-0.2775951922,-0.2242569327,0.1091209427,-0.0237828474,0.2019168288,0.0783613995,0.0341567732,0.1572858244,-0.2306452692,-0.1137423888,0.2455040663,0.0150245838,0.0841123611,-0.2900322974,0.4350042939,-0.1814204156,0.2838892341,0.0696543902,0.0759684816,0.1966336668,-0.1303307563,-0.1584434062,-0.2898952663,-0.0617264844,0.270403713,0.0174394976,-0.129190892,-0.568375051,-0.0273792874,0.0324698947,-0.1014007404,0.0562782586,-0.0123104909,0.0366902091,-0.3278360367,-0.2467757463,-0.044038184,0.2171133757,-0.3153653443,-0.0151790688,0.3517895937,0.0079167383,-0.1656942964,0.2585664093,0.0447287038,-0.0788884088,-0.2397040874,0.2174692154,0.1694951057,-0.2928570211,0.0100867813,0.1001555473,0.3597319424,-0.1373020113,-0.1130762622,0.401055783,-0.4852762222,-0.1666710824,0.3283014596,0.2628527284,-0.0499716215,0.1990734041,-0.4620003402,0.0208031684,-0.3443519175,0.0609698854,-0.2839337587,-0.3468373418,0.0982578397,-0.0938345417,-0.0748943612,-0.0737973824,0.219087556,0.0758348256,0.2237912118,0.4042210281,0.1214174181,0.0472104363,-0.212480858,0.0909303799,0.5390870571,0.1716981977,-0.1172630191,-0.0431797802,0.0888339356,0.3512161076,0.386287272,-0.0171315614,0.0357585587,-0.1284101158,0.0353125967,-0.1607109159,0.0252258461,-0.2964859605,-0.1575738788,0.0407147892,-0.1704651564,0.0476719812,-0.3185106516,0.3325026631,-0.028375918,-0.0514638126,-0.1414311677,0.3360973001,0.0217508096,-0.116924867,0.0903122351,-0.2452601641,-0.1850142628,0.5311390162,-0.1276957691,0.0373217463,0.3100013137,-0.0185543224,-0.1572021842,-0.1212829575,-0.3497071266,0.2189918309,0.2532023787,0.167312026,-0.2845161259,-0.13306126,0.0134299695,0.0724461451,-0.038484361,-0.0384799428,-0.0078570349,0.2023949325,-0.2250300199,-0.2332446128,-0.1535705626,0.2866339386,0.1143194586,-0.164140448,0.0640916824,0.2255733907,0.282151103,-0.1511053592,-0.1649658829,0.1279270947,-0.0194348674,-0.1609199494,0.0573712401,0.2651025653,0.1623861343,-0.0510671027,-0.180307433,0.1359896213,0.1391028017,0.0575792268,0.1468624324,-0.1183245555]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/489","title":"ug","comments":"please delete this","body":"","comment_length":3,"text":"ug\n\nplease delete this","embeddings":[0.3251394331,-0.3791508377,-0.3336610794,-0.0561220311,0.1526272744,0.2186691761,0.3307940662,-0.0663170591,-0.1441396326,0.140221566,-0.1578388214,0.0331792161,-0.0114352312,0.2071858048,-0.1967462897,0.2367383242,-0.0898705646,0.2015131712,-0.1144152209,-0.0954754129,-0.1858640313,-0.0492178947,-0.1193711981,0.2473709136,-0.1246762127,-0.2335108966,0.1072306633,0.1910086423,-0.3876164556,-0.0813365281,-0.011772858,0.0404558145,-0.4390583932,0.0761255473,-0.0000883204,-0.0557879657,0.3311251104,0.2945351601,-0.2524197996,0.4574931264,-0.1300058067,0.3443587124,-0.1396685541,-0.0088590309,0.0916793272,-0.1041545495,-0.1291739345,-0.0141843343,0.0953857899,0.136971578,0.4821210504,-0.1794409007,-0.2034163326,-0.0691115037,0.2562834918,0.1072602719,-0.0004478141,0.0608253926,0.0675326362,0.0523039401,0.2252196074,0.4574859142,-0.3876855969,-0.1712354273,-0.0275015924,-0.0932801291,0.7213936448,-0.2108874321,0.2820754051,-0.1039277911,-0.0270008203,-0.0687126964,0.2583593726,0.1448934376,0.0491171628,-0.0803311095,0.0630146265,0.111804381,0.0439120457,0.1976449043,-0.1149442419,-0.2982549965,0.0210008435,0.131087184,0.1317531615,0.114236109,-0.1264379919,-0.0301968828,0.2389686406,-0.0421871878,-0.14024885,-0.2163594216,-0.1058446988,0.0591184683,0.1367540061,-0.1960930228,0.036324542,0.1770200282,0.0571413971,0.0765688345,-0.0320647769,-0.0680347309,0.0055594542,0.1875848621,0.0861835107,0.1712066382,-0.0204155631,-0.0568385758,0.0671837479,-0.030154014,0.2558500171,0.2727033794,0.063344419,-0.0520095751,-0.1883606017,0.1752689481,0.0238495804,0.0187696889,-0.4426252842,0.0798812434,0.0204529259,0.2177879214,-0.082118459,0.1449830979,0.1004594043,0.2111973763,-0.0171897803,-0.0344948359,-0.06792216,-0.0652628839,-0.2328010499,0.3408983648,-0.1134666875,-0.1703284085,-0.108985588,-0.1000979096,0.1868377477,0.0484836251,0.0827930644,-0.0610400401,-0.051903151,-0.1555416435,0.0560428724,0.0275284741,-0.1855925918,-0.070332095,0.1447699517,-0.2539032996,-0.1067650169,0.3751991093,-0.2463761866,-0.2314403057,-0.2052222341,0.4683482051,0.3693299294,0.0890153497,-0.0814543068,-0.2418859005,-0.0449434221,0.213677153,0.3183338046,0.205916971,-0.1051854566,-0.0536891036,-0.3064975441,0.0015079349,0.0387710519,-0.0755133629,-0.1456528306,-0.1589523703,0.2147979438,0.4246843159,-0.097979635,0.0392812453,-0.3053821623,-0.0930024832,0.0725498646,-0.1353942007,-0.0634097978,0.2480850965,-0.2203394473,-0.0061449795,-0.1488687992,-0.0381881706,-0.2878954709,-0.105376482,0.0780235603,0.1825656146,0.0828857347,-0.2630689144,-0.3355977237,-0.4426750243,-0.059006162,-0.003988646,0.0561062209,0.0508368984,-0.0536621623,0.0240521859,0.3398604393,0.2351448089,-0.1207875162,0.0976374373,0.5795704722,0.040529754,-0.0067117382,0.1425935477,-0.1161237136,-0.2183266878,-0.1179018244,0.1455397308,0.0203538984,-0.4112123251,-0.1271644086,0.2515357733,-0.1423864514,-0.1083663553,0.3070005774,0.0004877531,-0.0716600418,-0.0497422777,-0.2991770506,0.0816960037,-0.3200016022,-0.03325231,-0.0640079454,-0.1356941313,-0.3131044507,0.165972203,0.1275187135,0.196452558,-0.091225177,0.0216572881,0.0758211389,0.0553935021,-0.1615127772,-0.0616081506,0.0576744452,0.2545337975,0.1118770614,-0.4514567554,-0.0789780319,0.2505477369,0.1681341529,0.0032466529,-0.0014197663,0.1629577726,0.0058144825,-0.1634030789,-0.2183650285,-0.1043599173,-0.1700589657,0.0162057262,0.162601769,-0.3757065535,0.1622658521,0.3434500694,-0.1365842074,0.159643501,0.1896733046,0.0385966934,0.2403078377,-0.086830169,0.1091373414,0.1381447464,0.1054539159,-0.1173324659,-0.0976103693,-0.3328694105,0.1424831748,0.0585771576,0.0357595794,0.0926951393,0.3814825714,-0.2665866315,0.3377171457,0.0647291765,0.0639992133,0.1283043474,0.0328483619,-0.0441631302,-0.1276735365,0.071390897,-0.2080574036,0.1267147511,0.0360628031,-0.046777498,-0.1818278879,-0.4615250528,0.4207829833,0.0347797647,-0.1625119746,-0.2041899711,0.4057887495,-0.0710207373,-0.2689083517,0.2175909877,-0.1019131541,0.5905122161,0.0948815122,0.278488189,-0.2523947954,-0.1869662106,-0.0373965837,0.363754034,-0.0802060813,-0.0889258236,0.2390879393,-0.0359181277,0.0511645116,-0.0090196598,-0.2777970135,0.1827554107,0.0595975704,0.0884666517,0.1359856576,0.1013269499,-0.0842593536,0.0731644332,-0.004779302,-0.2152835727,-0.055884432,-0.2364851683,-0.0271326695,0.0493136719,-0.1495669335,0.0098583736,0.0934685245,-0.1778974384,-0.0385912955,-0.0108415848,0.2044233233,0.2961528599,0.0206697881,0.2035495192,0.1580870599,0.1301713735,-0.2598519921,-0.3678024709,0.0255112648,0.1009464711,-0.0177399814,-0.0004752395,0.0321073569,-0.0042996411,-0.3270363212,-0.3279399872,-0.2578916848,0.0400787033,-0.0551853962,0.1146883667,-0.3947942853,-0.0742061287,-0.0957022756,-0.4122716188,0.050859645,-0.087503992,-0.1537881792,-0.0013713952,-0.0117306458,0.143323347,0.0515278503,0.2530711889,0.0411269292,-0.0388169214,0.2444158047,-0.0018923836,0.1705395877,0.2076675594,-0.0240250491,-0.2513823807,-0.0163877346,0.2980128527,-0.1319760084,0.2182841003,-0.0637666062,0.00224712,0.2001181394,-0.0124048749,-0.0730491951,-0.0012833687,0.0268074386,-0.041182749,-0.1878655553,-0.1471683681,-0.3359273374,0.208985433,-0.1906930506,0.1618019938,0.2124516666,-0.1331670731,0.0323413201,-0.0721993297,-0.0128038386,0.1715410501,0.2504726946,-0.0463598706,-0.0171995703,-0.1503163278,-0.1158861741,0.2789913118,-0.0467965007,0.120158121,-0.171595037,0.0663513541,-0.1298653036,0.1093037277,0.2789557576,0.1644321978,0.015830161,0.0832758173,0.3149493337,0.300809294,-0.1173831597,-0.0825533122,0.3112048805,-0.3307460248,-0.042446401,-0.1090372801,-0.3165953755,-0.2636817694,-0.3158059418,0.1001778692,0.3320193291,-0.0382373221,0.0226949453,0.0024403306,-0.0688270628,-0.0572654866,0.0941625684,-0.1841012836,-0.2320436537,0.0016789043,-0.4101068079,-0.2384145707,0.2154860049,-0.1578576118,0.2026024312,0.0652746111,0.1159142479,-0.0227048844,-0.1979093552,0.0840602219,0.2260818183,0.2111831158,0.0219174158,0.1401047856,0.3169317842,0.047784429,-0.1701558977,0.1047057807,-0.2770403028,0.0987522826,-0.0466908738,-0.0325941592,0.1226853654,-0.0998945385,0.2937782407,0.2456422597,0.0586323515,0.0800519288,0.1364177465,1.1753840446,0.0850313678,0.0662716776,0.1121813208,-0.0872185081,0.0198880751,-0.0402994007,0.216407299,-0.1395081133,0.223965168,0.0262213405,0.0380940549,0.0826084688,0.0732415318,-0.0138227874,-0.1126851588,-0.3934193254,0.2955793738,0.2734291553,-0.047941424,0.0734086782,0.1622211784,0.014039983,0.4344528615,0.0415304527,0.0556876436,-0.1100044027,0.222041741,0.1834473163,-0.0638237596,0.1523630023,0.1220489591,-0.1223076284,0.3025960624,-0.0329896435,-0.0897593424,-0.2651385963,-0.1217150912,0.1697660983,-0.0538143963,0.1581519097,0.2699386179,0.2912330329,-0.0882017314,-0.2344424129,0.0980652869,0.0519090332,-0.2399362028,-0.3485417962,0.2938335538,0.1646436304,-0.2159532905,-0.2259257734,0.0574521981,-0.032081794,0.124020718,-0.0676126406,-0.1232245266,-0.048535388,-0.0137177901,0.3800589442,-0.0710122585,-0.0585504919,-0.1455301046,0.1524885744,-0.3250125945,-0.1039428934,-0.1020726413,0.1998976618,-0.0439609736,0.1147711128,-0.1215743572,-0.4268776476,-0.6101433635,-0.0539863743,-0.312797904,-0.0384446047,0.0440305658,-0.0524906218,-0.2494985759,-0.094278723,-0.0943749249,-0.0186086502,0.1719747633,0.110307537,0.0799313486,-0.0905208737,0.01892961,-0.1584761143,0.1388210356,-0.1521743983,0.0697690472,0.0411811508,0.0830626562,-0.5751924515,0.1753800362,-0.2854382396,-0.2650830448,-0.0145746581,-0.172647357,0.051828932,-0.0190966092,0.40869385,0.0308847204,-0.1031549126,-0.537676692,0.0631717816,0.0061325263,0.1609380692,-0.1864877939,0.054256089,0.2653352916,-0.0494560115,0.2941940129,0.2368931472,0.0794241205,0.4217903316,0.1554763019,-0.1034312174,-0.2391563803,0.0790705159,-0.1563168764,0.1083447933,0.2677183449,0.1118262485,-0.2746810913,0.0854659826,-0.1668836176,0.2398767769,0.1627875268,0.4339314103,0.0506925955,-0.0262695439,-0.3419072032,-0.000419605,-0.0992942974,-0.1599562764,-0.1392900497,-0.3516272008,-0.3013123572,-0.0621304922,0.4348588586,-0.051698342,0.1915853173,-0.0162969064,-0.2472392768,-0.1202979535,0.1785375327,-0.0962109417,0.1412573159,0.0831884965,0.1641571075,0.1502377689,-0.3829547167,-0.1830313206,-0.2163675725,-0.2302640826,-0.0704438612,0.0912582353,0.2102478743,0.1002496928,-0.0101813944,-0.0462289266,-0.0208282936,0.1521176994,-0.0230620671,0.197187826,0.0045264391,-0.0652747974,0.5169909,-0.2934723198,-0.0192269832,0.3629436195,0.1340881586,0.3134340942,-0.3114856482,-0.108710669,0.1159977764,-0.2485215962,0.1382350624,-0.3137205243,0.0782085061,-0.3825937212,0.0755325109,-0.013792295,-0.01700739,0.1573326737,0.0297443885,-0.2642410994,-0.1893710345,0.0258671287,0.1212218553,0.0792187303,-0.1172309145,-0.1181297749,0.0933918357,0.2662143707,0.0479785204,0.1858396381,0.0750425458,-0.1254289895,-0.0376665555,-0.0838253051,-0.1040192991,-0.1915812641,0.0159620289,0.2132993042,0.0845386758,0.1651545614,-0.0871797428,0.4049282074,-0.3784354031,-0.0774089769,0.2708157301,0.1687759012,-0.1712133288,0.2934945524,-0.1978512555,-0.1034141928,-0.1832702011,0.2252779752,-0.3554995954,-0.0589331351,0.4378695786,-0.1454057544,0.0307991467,-0.1924993843,0.2079732567,0.04868358,0.1853008866,0.4956554472,0.0111678066,0.0165150259,-0.0989915729,-0.2457017004,0.3231140971,-0.0484446287,0.0831371099,0.110013403,0.0057344018,-0.055868987,0.2105262429,-0.0437424406,0.0518957339,-0.0757084563,-0.2156314403,-0.1111968383,0.275475353,0.0275162589,0.0748559833,0.1208234578,-0.3569108546,0.0328715146,0.0892233029,0.4182977974,-0.1020179093,-0.0666831359,-0.1363741159,0.1964920312,0.1060187444,0.072502166,-0.1719487756,-0.1550308466,-0.2656377554,0.2151388228,-0.1488396227,0.1206298694,0.1263219267,0.1700476706,-0.2076264024,-0.0807580426,-0.1758440286,0.0564217307,0.1772513837,0.0628673136,-0.0294672381,-0.1563623548,0.0995595232,0.0487221554,0.4552814364,-0.2773113847,-0.1309827119,0.1251428723,-0.1129915416,-0.2597572803,-0.1758491695,0.3955600262,0.138546586,-0.2705258131,0.1267334372,0.3834560215,0.2519322038,0.0163695663,-0.0781295225,0.0364481956,0.0075214836,0.0254518408,-0.1647969037,0.3209688962,-0.2271738797,-0.117752701,-0.0956274346,0.0125848548,0.0914252028,-0.0488387011,-0.0824891105,-0.2699761391]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/488","title":"issues with downloading datasets for wmt16 and wmt19","comments":"I found `UNv1.0.en-ru.tar.gz` here: https:\/\/conferences.unite.un.org\/uncorpus\/en\/downloadoverview, so it can be reconstructed with:\r\n```\r\nwget -c https:\/\/stuncorpusprod.blob.core.windows.net\/corpusfiles\/UNv1.0.en-ru.tar.gz.00\r\nwget -c https:\/\/stuncorpusprod.blob.core.windows.net\/corpusfiles\/UNv1.0.en-ru.tar.gz.01\r\nwget -c https:\/\/stuncorpusprod.blob.core.windows.net\/corpusfiles\/UNv1.0.en-ru.tar.gz.02\r\ncat UNv1.0.en-ru.tar.gz.0* > UNv1.0.en-ru.tar.gz\r\n```\r\nit has other languages as well, in case https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/ is gone","body":"I have encountered multiple issues while trying to:\r\n```\r\nimport nlp\r\ndataset = nlp.load_dataset('wmt16', 'ru-en')\r\nmetric = nlp.load_metric('wmt16')\r\n```\r\n1. I had to do `pip install -e \".[dev]\" ` on master, currently released nlp didn't work (sorry, didn't save the error) - I went back to the released version and now it worked. So it must have been some outdated dependencies that `pip install -e \".[dev]\" ` fixed.\r\n\r\n2. it was downloading at 60kbs - almost 5 hours to get the dataset. It was downloading all pairs and not just the one I asked for. \r\n\r\nI tried the same code with `wmt19` in parallel and it took a few secs to download and it only fetched data for the requested pair. (but it failed too, see below)\r\n\r\n3. my machine has crushed and when I retried I got:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/download.py\", line 9, in <module>\r\n dataset = nlp.load_dataset('wmt16', 'ru-en')\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/builder.py\", line 449, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/stas\/anaconda3\/envs\/main\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/stas\/anaconda3\/envs\/main\/lib\/python3.7\/os.py\", line 221, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ru-en\/1.0.0\/4d8269cdd971ed26984a9c0e4a158e0c7afc8135fac8fb8ee43ceecf38fd422d.incomplete'\r\n```\r\nit can't handle resumes. but neither allows a new start. Had to delete it manually.\r\n\r\n4. and finally when it downloaded the dataset, it then failed to fetch the metrics:\r\n```\r\nTraceback (most recent call last):\r\n File \".\/download.py\", line 15, in <module>\r\n metric = nlp.load_metric('wmt16')\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 442, in load_metric\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=False)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 258, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/metrics\/wmt16\/wmt16.py\r\n```\r\n\r\n5. If I run the same code with `wmt19`, it fails too:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-ru.tar.gz\r\n```","comment_length":37,"text":"issues with downloading datasets for wmt16 and wmt19\nI have encountered multiple issues while trying to:\r\n```\r\nimport nlp\r\ndataset = nlp.load_dataset('wmt16', 'ru-en')\r\nmetric = nlp.load_metric('wmt16')\r\n```\r\n1. I had to do `pip install -e \".[dev]\" ` on master, currently released nlp didn't work (sorry, didn't save the error) - I went back to the released version and now it worked. So it must have been some outdated dependencies that `pip install -e \".[dev]\" ` fixed.\r\n\r\n2. it was downloading at 60kbs - almost 5 hours to get the dataset. It was downloading all pairs and not just the one I asked for. \r\n\r\nI tried the same code with `wmt19` in parallel and it took a few secs to download and it only fetched data for the requested pair. (but it failed too, see below)\r\n\r\n3. my machine has crushed and when I retried I got:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/download.py\", line 9, in <module>\r\n dataset = nlp.load_dataset('wmt16', 'ru-en')\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/builder.py\", line 449, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/stas\/anaconda3\/envs\/main\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/stas\/anaconda3\/envs\/main\/lib\/python3.7\/os.py\", line 221, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ru-en\/1.0.0\/4d8269cdd971ed26984a9c0e4a158e0c7afc8135fac8fb8ee43ceecf38fd422d.incomplete'\r\n```\r\nit can't handle resumes. but neither allows a new start. Had to delete it manually.\r\n\r\n4. and finally when it downloaded the dataset, it then failed to fetch the metrics:\r\n```\r\nTraceback (most recent call last):\r\n File \".\/download.py\", line 15, in <module>\r\n metric = nlp.load_metric('wmt16')\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 442, in load_metric\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=False)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 258, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/metrics\/wmt16\/wmt16.py\r\n```\r\n\r\n5. If I run the same code with `wmt19`, it fails too:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-ru.tar.gz\r\n```\nI found `UNv1.0.en-ru.tar.gz` here: https:\/\/conferences.unite.un.org\/uncorpus\/en\/downloadoverview, so it can be reconstructed with:\r\n```\r\nwget -c https:\/\/stuncorpusprod.blob.core.windows.net\/corpusfiles\/UNv1.0.en-ru.tar.gz.00\r\nwget -c https:\/\/stuncorpusprod.blob.core.windows.net\/corpusfiles\/UNv1.0.en-ru.tar.gz.01\r\nwget -c https:\/\/stuncorpusprod.blob.core.windows.net\/corpusfiles\/UNv1.0.en-ru.tar.gz.02\r\ncat UNv1.0.en-ru.tar.gz.0* > UNv1.0.en-ru.tar.gz\r\n```\r\nit has other languages as well, in case https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/ is gone","embeddings":[-0.2756031156,-0.0938129723,-0.0165136941,0.5474312305,0.0739607885,-0.0815913901,0.0184081551,0.1925772578,0.1801728159,-0.0740146041,-0.0608425885,0.022961678,-0.1682554036,0.0706977993,0.1812847853,-0.0102873575,-0.0982710868,0.0546426848,-0.5839893818,-0.1066129804,-0.251978755,0.1597542912,-0.1889990717,0.0316224173,0.0291602258,-0.1472716779,-0.0056710104,0.0338853523,-0.4226700962,-0.3919169903,0.4446610212,0.1778903902,0.0486655273,0.2788916528,-0.0001182758,-0.2390260845,0.2288018614,0.0146836666,-0.302140981,-0.1883610785,-0.265650332,-0.5416650772,0.1796258837,0.0793974623,0.1905766726,-0.2577368617,-0.096332252,-0.1201571822,0.3336129189,0.1541093886,0.1808845848,0.4754364192,0.1862676293,0.0183087811,0.1463204622,0.1829875112,-0.0161829032,0.0414138436,0.3683440983,-0.1451096982,0.3042299747,0.2626513243,-0.1972892433,0.3405378163,0.2254597098,0.0342918411,0.2514481544,-0.3291010261,-0.0104288654,0.3739078939,0.0919398814,-0.0189400502,-0.28330791,-0.3328357339,-0.0607213192,-0.591956377,0.0915093645,0.3554451466,-0.3222121298,-0.0712231621,-0.5180068612,0.0008279618,0.0945125818,0.3394875526,-0.1083701104,0.3477387726,0.0328624025,0.1435943693,0.1835633516,-0.3430734873,-0.1032524481,-0.3144376874,0.2530840933,0.2212258428,-0.2552488148,0.0872775763,0.0101171061,0.0087072952,0.0032323,-0.0976789668,-0.0223416574,0.1039455757,-0.138657406,0.0357805528,0.1094087362,0.214331314,-0.0763300359,-0.0989684314,0.2642333806,-0.0348216146,0.2747761905,0.2433936149,0.1871807128,-0.1478894949,-0.3470747769,0.0696833655,0.1543291807,-0.2491889447,-0.3024494648,-0.0835244283,0.0652420372,-0.2730256915,0.0408257954,0.0257640518,-0.299441874,0.2130645663,0.0610918254,0.1647904664,-0.2455622852,-0.1874856204,-0.1245096624,-0.2107741833,-0.0994957685,0.1155639514,0.6232811213,0.0698022023,0.2070064247,0.1056845635,-0.1488355249,0.109546639,-0.0190164763,-0.1189081147,0.1222632751,0.2221131921,0.0451032706,0.4238990545,-0.1189093962,0.0161676593,-0.1315837801,-0.1299055666,-0.0583826154,0.0434743352,0.2906782329,0.2067490369,-0.3159318864,-0.0533832051,-0.2992135882,0.0907554924,-0.1152078286,-0.0338809527,-0.1886040717,-0.2438194454,-0.3752387464,-0.2526579201,0.376717478,0.2799056172,-0.166607812,-0.1649063677,-0.0837725848,0.0289434884,0.3123579025,0.4445025027,-0.2286306024,0.046422787,-0.1588803083,-0.0510554165,0.3584416807,-0.3500113189,-0.14593409,0.0142124686,-0.0908057168,0.211790964,-0.0302680898,0.0880928412,0.2926700115,-0.1595734209,0.2096131444,0.4093490839,-0.0834689662,-0.1282000095,-0.2435817719,-0.4912775755,0.6207739711,0.0323619507,0.1447010785,-0.3409759402,-0.0999755561,-0.1594638377,0.4071716368,0.1020297557,0.138751775,-0.0597009845,-0.0496113822,-0.0825747699,-0.0716121942,-0.0022220758,-0.4874588251,0.3761884272,-0.2441596836,0.1380985528,0.0267383829,0.2019913644,-0.0252641272,-0.1451207101,-0.1093342528,-0.1639743894,0.0372414552,0.0199695416,0.0985343978,-0.0397354141,0.0781109557,0.0820711926,-0.0402478613,0.1321014911,-0.2335660011,0.2222174406,0.0546635017,-0.028811371,0.3562957942,0.283198595,0.1775622219,-0.0596050918,-0.0277689975,-0.0303030014,-0.0066433311,0.4922422469,-0.104197897,-0.0141533799,0.3394637108,-0.4042288959,0.1655757278,0.2719747126,0.2212738097,-0.3597029746,0.0170894042,0.2387666702,-0.0915701389,0.3743444383,0.1100310087,0.0123178326,0.0113621661,0.0139470706,-0.089154087,-0.0341479592,0.4079688191,-0.0964392647,0.0220886227,0.0996986181,-0.0303346273,0.2387857437,0.2519143224,0.1178979278,0.0268229246,-0.0370242521,-0.1333160549,-0.1370724738,-0.1800833493,0.2379724383,0.5232086778,0.1820151508,0.190797016,0.0909063593,0.0877680108,-0.4019683599,0.3726394176,0.2383785248,-0.0019474661,0.1320959479,0.0430695452,-0.0163701978,-0.1104837582,0.4184349179,0.0917636901,0.1791213006,-0.2323913872,-0.2492549121,-0.5810912848,-0.4634234309,-0.4940530658,-0.3730905354,-0.3810460567,-0.3317370713,-0.1937682629,-0.0642361715,-0.0816402733,0.1759491414,0.168973878,0.0007634666,-0.355763495,0.4598483145,-0.027881328,-0.0493743308,-0.3101087511,0.0172245521,0.4009830952,0.1410500854,0.4067273736,0.0004268971,-0.0131987119,0.0081043374,-0.3204092085,-0.0419830829,-0.0738799348,0.2629978657,0.1375837922,0.2869311273,0.2675925493,-0.0940290391,0.1489828825,-0.2181366682,-0.0514863394,-0.1704598516,-0.0215497147,0.0187395047,-0.2237060517,-0.4680871069,-0.711363852,-0.3568281531,0.1662416756,0.0643555596,0.0964026153,0.3763373196,-0.2668996453,-0.0048240889,-0.043119669,0.1872792244,-0.1405702829,0.1058703735,0.3484725058,-0.043336086,-0.4463670552,0.0350139476,0.1576631516,0.2508547604,0.1007582992,-0.4295147955,0.3669915497,-0.1442829072,-0.1073541045,0.2240204513,0.1611544937,0.1435948461,-0.0059772413,-0.0148710888,-0.0385597683,-0.2719137371,-0.0486938916,-0.055932112,0.2735077143,0.0550433919,0.122125119,0.0208200738,0.5913627148,0.2570456564,-0.0870676115,0.1685414165,0.2239551544,0.0688810498,0.3214880526,-0.5713624358,0.0624389425,0.014921315,0.2643279731,0.2561017871,-0.024556065,-0.0789706185,-0.3335882723,0.0456078388,-0.2322384864,-0.3285506666,-0.0424311832,0.2228118926,0.2037611455,0.3823131621,0.1332704574,-0.0036640295,-0.3344186544,-0.1921590567,-0.0713421926,-0.0795905367,0.1410654634,-0.305629462,-0.3055719137,-0.3390577734,0.1853805482,0.01230607,0.6306856275,-0.1004613265,0.0602535829,0.0003582306,0.0535958745,0.2863096595,-0.0149867469,0.223822102,-0.0550027937,0.0418632627,-0.7075550556,-0.0788407773,-0.149083674,-0.1024419293,0.3979476988,0.2676806748,-0.3344471753,-0.3497644961,0.128687799,0.4699044526,-0.3002048731,-0.0902623162,-0.000184378,-0.1544873416,-0.1256392151,-0.1509202421,-0.2789451778,0.1925426424,-0.3451689482,0.0452711768,0.1801136434,0.0680580288,0.3399452269,-0.0267138015,0.0946662575,0.1880122721,0.4270004332,0.3394292891,0.1034898907,0.0907359049,0.3604844809,-0.0748465285,0.3105944097,-0.1079983413,-0.0581995025,0.2565919757,0.5723106265,-0.0812101364,0.2673694491,0.0279870499,0.0784473494,-0.4492246211,0.0794503167,0.3588107526,-0.1368251741,-0.2159993649,-0.2861743867,0.325907886,0.2019692212,-0.1109420136,0.2802770138,-0.1149723232,-0.3127885759,-0.1978135258,0.0408775061,0.950541079,0.2387174815,0.2275000066,-0.0095485104,-0.4000633061,0.4883543849,0.4271089435,-0.1436373293,-0.0481956117,0.319152385,-0.0039904923,-0.1169142127,0.0799548328,-0.0163567495,-0.1558759212,0.5566452146,0.3277597129,0.1902880073,0.0671359971,0.3798567951,0.008937126,-0.3773046434,-0.2538492084,0.0673444793,-0.3593877852,0.2783208489,-0.1393684,-0.3252620995,0.0053060064,-0.2072428614,-0.3205970824,0.0383541808,-0.4705601633,0.1999380738,-0.1084300652,-0.0484950431,0.191311717,-0.0939288884,0.0233553965,0.1939816773,-0.3421660364,0.0471422747,-0.4583493471,0.2351858616,0.041799169,0.0699069649,-0.001924105,-0.1840398312,-0.1144265607,-0.2535886765,0.0115255034,0.3370921314,-0.0575815439,0.0259476639,0.1794763356,0.173576355,-0.0933967158,0.1255259365,0.1240274906,-0.1376567334,0.1038089171,0.2577846944,-0.0941019878,0.592993021,0.1446930021,-0.0829000771,-0.1528961807,0.5256072283,0.0521146469,-0.146619007,0.4898711741,0.551068902,-0.1143533215,-0.2417031229,-0.0041621705,-0.3494568765,-0.4845702648,-0.0105194738,0.0329329371,0.3049015105,0.0559948571,0.5251324773,0.2182418853,-0.2787428498,0.157030493,-0.4990168214,-0.1411263794,0.3006083667,0.0930177122,0.0704476312,-0.1138189733,0.0523063913,0.0783459917,0.3154848516,-0.2897268236,0.19059892,-0.0451219566,0.1344131529,0.2483339757,-0.242375508,0.1088931113,-0.046524927,-0.0645425841,-0.1267765462,-0.4485759437,-0.1471533775,-0.2247590423,0.1452783346,-0.0036861061,-0.474239707,0.0590283237,-0.3070831001,0.0487826057,0.0607667863,-0.1974305212,0.0964792371,0.0146514121,0.1217674017,0.1369762421,0.171283409,0.0307994187,0.4250535965,-0.0532620959,0.2902286351,0.1436617225,0.1730932593,0.2801129222,-0.1230007559,-0.1648173481,-0.1834212095,-0.0149281854,-0.2281159312,0.2769261003,-0.2383158058,-0.0286652558,0.1843677312,0.3846011162,0.5537839532,0.1585487872,0.0089077326,-0.0664098114,0.1120337769,-0.1739787906,0.0375586152,0.2010296732,0.4397958219,-0.1216781586,0.130498603,0.3467441499,-0.2371853143,-0.1927649975,0.161176309,0.4595688879,-0.1366926283,0.0359057449,0.139931038,0.0546213537,-0.0069570635,-0.140351221,0.1283730865,0.0266115945,0.0465465412,-0.2296463996,0.409722209,0.1429581791,-0.0815549493,0.1501864195,-0.3475331664,0.0019549369,0.3410553038,-0.1524552256,-0.0454493687,0.1050023064,0.6036858559,-0.0399988405,-0.3411641717,-0.0343074873,0.5274829865,0.1017361432,-0.1091591194,-0.3369088173,-0.2871397734,-0.2651319206,0.0117885899,-0.113232024,-0.20115453,0.115309827,0.2378166318,-0.2011552006,-0.6774241328,-0.2983496487,0.0227554571,-0.1241075397,-0.1449491829,0.2098425776,0.4280456901,-0.0640470833,0.0502663814,0.5844964385,0.3379960656,0.3474577069,-0.2088452131,-0.0027260517,0.1122177839,0.0115934629,-0.2457560003,0.3455425203,0.0748100579,-0.0603161268,0.3822260797,0.0945058167,-0.1459354758,-0.2766559124,0.1253265589,-0.2008286566,-0.130469203,0.4320131838,-0.214146927,0.0450588353,-0.2039672583,0.0448366702,-0.3329181373,0.1637924463,0.5231636763,0.0549017303,0.1952826083,-0.3426962197,0.0593012683,-0.0701015666,0.7746468186,0.3278970122,0.1063381359,-0.3810437024,-0.2430894673,-0.9569471478,-0.2136810422,-0.1736372858,0.2115756571,-0.0691612735,0.2186158001,0.1369336843,0.2308200151,0.0730831102,0.4956247509,0.0183098894,-0.0512789004,-0.4442256987,0.026509909,0.1034276262,-0.1029869989,-0.0211806856,-0.1284394711,0.3681438267,-0.5790628195,-0.0293166954,-0.3203986287,0.0675374269,-0.1546666622,0.5107826591,0.1049053818,0.2914257348,0.4872877002,0.1722445935,-0.1365397722,-0.2456263304,-0.1868422329,0.0769094303,0.2122345418,-0.1695899218,0.2421000302,-0.1548054963,0.0563738011,-0.2551071942,-0.0378178731,-0.2682906091,-0.014333168,-0.0254047792,0.0842683092,-0.0810289979,0.0937445387,0.1031431407,0.171403721,-0.1260119975,0.1677506268,-0.3540552855,-0.4014781117,0.4523608983,-0.2661026716,-0.190934211,-0.2678647637,0.2520736456,0.1884660274,0.0888339654,-0.5078635216,-0.0687039346,0.3283727169,-0.2275226414,-0.1473045647,0.2823517919,-0.1936619878,-0.3195132613,-0.0002883481,0.1741587371,0.2486365736,-0.3690263331,0.0640236884,-0.0880565718]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/488","title":"issues with downloading datasets for wmt16 and wmt19","comments":"Further, `nlp.load_dataset('wmt19', 'ru-en')` has only the `train` and `val` datasets. `test` is missing.\r\n\r\nFixed locally for summarization needs, by running:\r\n```\r\npip install sacrebleu\r\nsacrebleu -t wmt19 -l ru-en --echo src > test.source\r\nsacrebleu -t wmt19 -l ru-en --echo ref > test.target\r\n```\r\nh\/t @sshleifer ","body":"I have encountered multiple issues while trying to:\r\n```\r\nimport nlp\r\ndataset = nlp.load_dataset('wmt16', 'ru-en')\r\nmetric = nlp.load_metric('wmt16')\r\n```\r\n1. I had to do `pip install -e \".[dev]\" ` on master, currently released nlp didn't work (sorry, didn't save the error) - I went back to the released version and now it worked. So it must have been some outdated dependencies that `pip install -e \".[dev]\" ` fixed.\r\n\r\n2. it was downloading at 60kbs - almost 5 hours to get the dataset. It was downloading all pairs and not just the one I asked for. \r\n\r\nI tried the same code with `wmt19` in parallel and it took a few secs to download and it only fetched data for the requested pair. (but it failed too, see below)\r\n\r\n3. my machine has crushed and when I retried I got:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/download.py\", line 9, in <module>\r\n dataset = nlp.load_dataset('wmt16', 'ru-en')\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/builder.py\", line 449, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/stas\/anaconda3\/envs\/main\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/stas\/anaconda3\/envs\/main\/lib\/python3.7\/os.py\", line 221, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ru-en\/1.0.0\/4d8269cdd971ed26984a9c0e4a158e0c7afc8135fac8fb8ee43ceecf38fd422d.incomplete'\r\n```\r\nit can't handle resumes. but neither allows a new start. Had to delete it manually.\r\n\r\n4. and finally when it downloaded the dataset, it then failed to fetch the metrics:\r\n```\r\nTraceback (most recent call last):\r\n File \".\/download.py\", line 15, in <module>\r\n metric = nlp.load_metric('wmt16')\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 442, in load_metric\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=False)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 258, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/metrics\/wmt16\/wmt16.py\r\n```\r\n\r\n5. If I run the same code with `wmt19`, it fails too:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-ru.tar.gz\r\n```","comment_length":45,"text":"issues with downloading datasets for wmt16 and wmt19\nI have encountered multiple issues while trying to:\r\n```\r\nimport nlp\r\ndataset = nlp.load_dataset('wmt16', 'ru-en')\r\nmetric = nlp.load_metric('wmt16')\r\n```\r\n1. I had to do `pip install -e \".[dev]\" ` on master, currently released nlp didn't work (sorry, didn't save the error) - I went back to the released version and now it worked. So it must have been some outdated dependencies that `pip install -e \".[dev]\" ` fixed.\r\n\r\n2. it was downloading at 60kbs - almost 5 hours to get the dataset. It was downloading all pairs and not just the one I asked for. \r\n\r\nI tried the same code with `wmt19` in parallel and it took a few secs to download and it only fetched data for the requested pair. (but it failed too, see below)\r\n\r\n3. my machine has crushed and when I retried I got:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \".\/download.py\", line 9, in <module>\r\n dataset = nlp.load_dataset('wmt16', 'ru-en')\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 549, in load_dataset\r\n download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/builder.py\", line 449, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/stas\/anaconda3\/envs\/main\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/builder.py\", line 422, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/stas\/anaconda3\/envs\/main\/lib\/python3.7\/os.py\", line 221, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/stas\/.cache\/huggingface\/datasets\/wmt16\/ru-en\/1.0.0\/4d8269cdd971ed26984a9c0e4a158e0c7afc8135fac8fb8ee43ceecf38fd422d.incomplete'\r\n```\r\nit can't handle resumes. but neither allows a new start. Had to delete it manually.\r\n\r\n4. and finally when it downloaded the dataset, it then failed to fetch the metrics:\r\n```\r\nTraceback (most recent call last):\r\n File \".\/download.py\", line 15, in <module>\r\n metric = nlp.load_metric('wmt16')\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 442, in load_metric\r\n module_path, hash = prepare_module(path, download_config=download_config, dataset=False)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/load.py\", line 258, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/mnt\/nvme1\/code\/huggingface\/nlp-master\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/metrics\/wmt16\/wmt16.py\r\n```\r\n\r\n5. If I run the same code with `wmt19`, it fails too:\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-ru.tar.gz\r\n```\nFurther, `nlp.load_dataset('wmt19', 'ru-en')` has only the `train` and `val` datasets. `test` is missing.\r\n\r\nFixed locally for summarization needs, by running:\r\n```\r\npip install sacrebleu\r\nsacrebleu -t wmt19 -l ru-en --echo src > test.source\r\nsacrebleu -t wmt19 -l ru-en --echo ref > test.target\r\n```\r\nh\/t @sshleifer ","embeddings":[-0.2756031156,-0.0938129723,-0.0165136941,0.5474312305,0.0739607885,-0.0815913901,0.0184081551,0.1925772578,0.1801728159,-0.0740146041,-0.0608425885,0.022961678,-0.1682554036,0.0706977993,0.1812847853,-0.0102873575,-0.0982710868,0.0546426848,-0.5839893818,-0.1066129804,-0.251978755,0.1597542912,-0.1889990717,0.0316224173,0.0291602258,-0.1472716779,-0.0056710104,0.0338853523,-0.4226700962,-0.3919169903,0.4446610212,0.1778903902,0.0486655273,0.2788916528,-0.0001182758,-0.2390260845,0.2288018614,0.0146836666,-0.302140981,-0.1883610785,-0.265650332,-0.5416650772,0.1796258837,0.0793974623,0.1905766726,-0.2577368617,-0.096332252,-0.1201571822,0.3336129189,0.1541093886,0.1808845848,0.4754364192,0.1862676293,0.0183087811,0.1463204622,0.1829875112,-0.0161829032,0.0414138436,0.3683440983,-0.1451096982,0.3042299747,0.2626513243,-0.1972892433,0.3405378163,0.2254597098,0.0342918411,0.2514481544,-0.3291010261,-0.0104288654,0.3739078939,0.0919398814,-0.0189400502,-0.28330791,-0.3328357339,-0.0607213192,-0.591956377,0.0915093645,0.3554451466,-0.3222121298,-0.0712231621,-0.5180068612,0.0008279618,0.0945125818,0.3394875526,-0.1083701104,0.3477387726,0.0328624025,0.1435943693,0.1835633516,-0.3430734873,-0.1032524481,-0.3144376874,0.2530840933,0.2212258428,-0.2552488148,0.0872775763,0.0101171061,0.0087072952,0.0032323,-0.0976789668,-0.0223416574,0.1039455757,-0.138657406,0.0357805528,0.1094087362,0.214331314,-0.0763300359,-0.0989684314,0.2642333806,-0.0348216146,0.2747761905,0.2433936149,0.1871807128,-0.1478894949,-0.3470747769,0.0696833655,0.1543291807,-0.2491889447,-0.3024494648,-0.0835244283,0.0652420372,-0.2730256915,0.0408257954,0.0257640518,-0.299441874,0.2130645663,0.0610918254,0.1647904664,-0.2455622852,-0.1874856204,-0.1245096624,-0.2107741833,-0.0994957685,0.1155639514,0.6232811213,0.0698022023,0.2070064247,0.1056845635,-0.1488355249,0.109546639,-0.0190164763,-0.1189081147,0.1222632751,0.2221131921,0.0451032706,0.4238990545,-0.1189093962,0.0161676593,-0.1315837801,-0.1299055666,-0.0583826154,0.0434743352,0.2906782329,0.2067490369,-0.3159318864,-0.0533832051,-0.2992135882,0.0907554924,-0.1152078286,-0.0338809527,-0.1886040717,-0.2438194454,-0.3752387464,-0.2526579201,0.376717478,0.2799056172,-0.166607812,-0.1649063677,-0.0837725848,0.0289434884,0.3123579025,0.4445025027,-0.2286306024,0.046422787,-0.1588803083,-0.0510554165,0.3584416807,-0.3500113189,-0.14593409,0.0142124686,-0.0908057168,0.211790964,-0.0302680898,0.0880928412,0.2926700115,-0.1595734209,0.2096131444,0.4093490839,-0.0834689662,-0.1282000095,-0.2435817719,-0.4912775755,0.6207739711,0.0323619507,0.1447010785,-0.3409759402,-0.0999755561,-0.1594638377,0.4071716368,0.1020297557,0.138751775,-0.0597009845,-0.0496113822,-0.0825747699,-0.0716121942,-0.0022220758,-0.4874588251,0.3761884272,-0.2441596836,0.1380985528,0.0267383829,0.2019913644,-0.0252641272,-0.1451207101,-0.1093342528,-0.1639743894,0.0372414552,0.0199695416,0.0985343978,-0.0397354141,0.0781109557,0.0820711926,-0.0402478613,0.1321014911,-0.2335660011,0.2222174406,0.0546635017,-0.028811371,0.3562957942,0.283198595,0.1775622219,-0.0596050918,-0.0277689975,-0.0303030014,-0.0066433311,0.4922422469,-0.104197897,-0.0141533799,0.3394637108,-0.4042288959,0.1655757278,0.2719747126,0.2212738097,-0.3597029746,0.0170894042,0.2387666702,-0.0915701389,0.3743444383,0.1100310087,0.0123178326,0.0113621661,0.0139470706,-0.089154087,-0.0341479592,0.4079688191,-0.0964392647,0.0220886227,0.0996986181,-0.0303346273,0.2387857437,0.2519143224,0.1178979278,0.0268229246,-0.0370242521,-0.1333160549,-0.1370724738,-0.1800833493,0.2379724383,0.5232086778,0.1820151508,0.190797016,0.0909063593,0.0877680108,-0.4019683599,0.3726394176,0.2383785248,-0.0019474661,0.1320959479,0.0430695452,-0.0163701978,-0.1104837582,0.4184349179,0.0917636901,0.1791213006,-0.2323913872,-0.2492549121,-0.5810912848,-0.4634234309,-0.4940530658,-0.3730905354,-0.3810460567,-0.3317370713,-0.1937682629,-0.0642361715,-0.0816402733,0.1759491414,0.168973878,0.0007634666,-0.355763495,0.4598483145,-0.027881328,-0.0493743308,-0.3101087511,0.0172245521,0.4009830952,0.1410500854,0.4067273736,0.0004268971,-0.0131987119,0.0081043374,-0.3204092085,-0.0419830829,-0.0738799348,0.2629978657,0.1375837922,0.2869311273,0.2675925493,-0.0940290391,0.1489828825,-0.2181366682,-0.0514863394,-0.1704598516,-0.0215497147,0.0187395047,-0.2237060517,-0.4680871069,-0.711363852,-0.3568281531,0.1662416756,0.0643555596,0.0964026153,0.3763373196,-0.2668996453,-0.0048240889,-0.043119669,0.1872792244,-0.1405702829,0.1058703735,0.3484725058,-0.043336086,-0.4463670552,0.0350139476,0.1576631516,0.2508547604,0.1007582992,-0.4295147955,0.3669915497,-0.1442829072,-0.1073541045,0.2240204513,0.1611544937,0.1435948461,-0.0059772413,-0.0148710888,-0.0385597683,-0.2719137371,-0.0486938916,-0.055932112,0.2735077143,0.0550433919,0.122125119,0.0208200738,0.5913627148,0.2570456564,-0.0870676115,0.1685414165,0.2239551544,0.0688810498,0.3214880526,-0.5713624358,0.0624389425,0.014921315,0.2643279731,0.2561017871,-0.024556065,-0.0789706185,-0.3335882723,0.0456078388,-0.2322384864,-0.3285506666,-0.0424311832,0.2228118926,0.2037611455,0.3823131621,0.1332704574,-0.0036640295,-0.3344186544,-0.1921590567,-0.0713421926,-0.0795905367,0.1410654634,-0.305629462,-0.3055719137,-0.3390577734,0.1853805482,0.01230607,0.6306856275,-0.1004613265,0.0602535829,0.0003582306,0.0535958745,0.2863096595,-0.0149867469,0.223822102,-0.0550027937,0.0418632627,-0.7075550556,-0.0788407773,-0.149083674,-0.1024419293,0.3979476988,0.2676806748,-0.3344471753,-0.3497644961,0.128687799,0.4699044526,-0.3002048731,-0.0902623162,-0.000184378,-0.1544873416,-0.1256392151,-0.1509202421,-0.2789451778,0.1925426424,-0.3451689482,0.0452711768,0.1801136434,0.0680580288,0.3399452269,-0.0267138015,0.0946662575,0.1880122721,0.4270004332,0.3394292891,0.1034898907,0.0907359049,0.3604844809,-0.0748465285,0.3105944097,-0.1079983413,-0.0581995025,0.2565919757,0.5723106265,-0.0812101364,0.2673694491,0.0279870499,0.0784473494,-0.4492246211,0.0794503167,0.3588107526,-0.1368251741,-0.2159993649,-0.2861743867,0.325907886,0.2019692212,-0.1109420136,0.2802770138,-0.1149723232,-0.3127885759,-0.1978135258,0.0408775061,0.950541079,0.2387174815,0.2275000066,-0.0095485104,-0.4000633061,0.4883543849,0.4271089435,-0.1436373293,-0.0481956117,0.319152385,-0.0039904923,-0.1169142127,0.0799548328,-0.0163567495,-0.1558759212,0.5566452146,0.3277597129,0.1902880073,0.0671359971,0.3798567951,0.008937126,-0.3773046434,-0.2538492084,0.0673444793,-0.3593877852,0.2783208489,-0.1393684,-0.3252620995,0.0053060064,-0.2072428614,-0.3205970824,0.0383541808,-0.4705601633,0.1999380738,-0.1084300652,-0.0484950431,0.191311717,-0.0939288884,0.0233553965,0.1939816773,-0.3421660364,0.0471422747,-0.4583493471,0.2351858616,0.041799169,0.0699069649,-0.001924105,-0.1840398312,-0.1144265607,-0.2535886765,0.0115255034,0.3370921314,-0.0575815439,0.0259476639,0.1794763356,0.173576355,-0.0933967158,0.1255259365,0.1240274906,-0.1376567334,0.1038089171,0.2577846944,-0.0941019878,0.592993021,0.1446930021,-0.0829000771,-0.1528961807,0.5256072283,0.0521146469,-0.146619007,0.4898711741,0.551068902,-0.1143533215,-0.2417031229,-0.0041621705,-0.3494568765,-0.4845702648,-0.0105194738,0.0329329371,0.3049015105,0.0559948571,0.5251324773,0.2182418853,-0.2787428498,0.157030493,-0.4990168214,-0.1411263794,0.3006083667,0.0930177122,0.0704476312,-0.1138189733,0.0523063913,0.0783459917,0.3154848516,-0.2897268236,0.19059892,-0.0451219566,0.1344131529,0.2483339757,-0.242375508,0.1088931113,-0.046524927,-0.0645425841,-0.1267765462,-0.4485759437,-0.1471533775,-0.2247590423,0.1452783346,-0.0036861061,-0.474239707,0.0590283237,-0.3070831001,0.0487826057,0.0607667863,-0.1974305212,0.0964792371,0.0146514121,0.1217674017,0.1369762421,0.171283409,0.0307994187,0.4250535965,-0.0532620959,0.2902286351,0.1436617225,0.1730932593,0.2801129222,-0.1230007559,-0.1648173481,-0.1834212095,-0.0149281854,-0.2281159312,0.2769261003,-0.2383158058,-0.0286652558,0.1843677312,0.3846011162,0.5537839532,0.1585487872,0.0089077326,-0.0664098114,0.1120337769,-0.1739787906,0.0375586152,0.2010296732,0.4397958219,-0.1216781586,0.130498603,0.3467441499,-0.2371853143,-0.1927649975,0.161176309,0.4595688879,-0.1366926283,0.0359057449,0.139931038,0.0546213537,-0.0069570635,-0.140351221,0.1283730865,0.0266115945,0.0465465412,-0.2296463996,0.409722209,0.1429581791,-0.0815549493,0.1501864195,-0.3475331664,0.0019549369,0.3410553038,-0.1524552256,-0.0454493687,0.1050023064,0.6036858559,-0.0399988405,-0.3411641717,-0.0343074873,0.5274829865,0.1017361432,-0.1091591194,-0.3369088173,-0.2871397734,-0.2651319206,0.0117885899,-0.113232024,-0.20115453,0.115309827,0.2378166318,-0.2011552006,-0.6774241328,-0.2983496487,0.0227554571,-0.1241075397,-0.1449491829,0.2098425776,0.4280456901,-0.0640470833,0.0502663814,0.5844964385,0.3379960656,0.3474577069,-0.2088452131,-0.0027260517,0.1122177839,0.0115934629,-0.2457560003,0.3455425203,0.0748100579,-0.0603161268,0.3822260797,0.0945058167,-0.1459354758,-0.2766559124,0.1253265589,-0.2008286566,-0.130469203,0.4320131838,-0.214146927,0.0450588353,-0.2039672583,0.0448366702,-0.3329181373,0.1637924463,0.5231636763,0.0549017303,0.1952826083,-0.3426962197,0.0593012683,-0.0701015666,0.7746468186,0.3278970122,0.1063381359,-0.3810437024,-0.2430894673,-0.9569471478,-0.2136810422,-0.1736372858,0.2115756571,-0.0691612735,0.2186158001,0.1369336843,0.2308200151,0.0730831102,0.4956247509,0.0183098894,-0.0512789004,-0.4442256987,0.026509909,0.1034276262,-0.1029869989,-0.0211806856,-0.1284394711,0.3681438267,-0.5790628195,-0.0293166954,-0.3203986287,0.0675374269,-0.1546666622,0.5107826591,0.1049053818,0.2914257348,0.4872877002,0.1722445935,-0.1365397722,-0.2456263304,-0.1868422329,0.0769094303,0.2122345418,-0.1695899218,0.2421000302,-0.1548054963,0.0563738011,-0.2551071942,-0.0378178731,-0.2682906091,-0.014333168,-0.0254047792,0.0842683092,-0.0810289979,0.0937445387,0.1031431407,0.171403721,-0.1260119975,0.1677506268,-0.3540552855,-0.4014781117,0.4523608983,-0.2661026716,-0.190934211,-0.2678647637,0.2520736456,0.1884660274,0.0888339654,-0.5078635216,-0.0687039346,0.3283727169,-0.2275226414,-0.1473045647,0.2823517919,-0.1936619878,-0.3195132613,-0.0002883481,0.1741587371,0.2486365736,-0.3690263331,0.0640236884,-0.0880565718]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/486","title":"Bookcorpus data contains pretokenized text","comments":"Yes indeed it looks like some `'` and spaces are missing (for example in `dont` or `didnt`).\r\nDo you know if there exist some copies without this issue ?\r\nHow would you fix this issue on the current data exactly ? I can see that the data is raw text (not tokenized) so I'm not sure I understand how you would do it. Could you provide more details ?","body":"It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575","comment_length":69,"text":"Bookcorpus data contains pretokenized text\nIt seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575\nYes indeed it looks like some `'` and spaces are missing (for example in `dont` or `didnt`).\r\nDo you know if there exist some copies without this issue ?\r\nHow would you fix this issue on the current data exactly ? I can see that the data is raw text (not tokenized) so I'm not sure I understand how you would do it. Could you provide more details ?","embeddings":[0.1711838841,0.225408867,-0.0682412684,0.348397702,-0.1143134087,-0.2559943497,-0.1209403202,0.2354474664,-0.4317489266,0.1244580597,-0.0878593251,0.226625219,0.3232100308,-0.1945853382,-0.0046418118,0.1363860369,0.2212239802,0.3154225349,0.1631603688,-0.2263966203,-0.1609849334,0.2912673056,0.1281992197,-0.0333504714,-0.2382831573,-0.3374491632,-0.2922897935,0.2494613975,-0.1388570219,-0.4692249596,0.0219636299,0.313759625,-0.0443142876,0.3362240791,-0.0001108157,-0.2952581942,-0.0561965853,-0.052477628,-0.2681582272,-0.0142242312,0.2646716833,-0.0555802546,0.3515130877,-0.4346842468,-0.3570377231,0.0960462764,-0.1705732793,-0.0543890223,0.3019437492,0.1815719455,0.2007733434,0.2229604572,-0.1164248139,0.2234891802,0.2198396772,-0.0061349124,-0.1732557118,0.1943600178,0.5178596377,0.2606931925,-0.0824627802,0.3769720197,-0.1626883,-0.1177665815,0.0906067342,0.034189146,0.0038020799,-0.1884816736,0.1414439082,0.1694197804,0.2140863389,-0.1607313454,-0.1241174042,-0.1997764558,0.0149029605,-0.1498395205,0.3236131668,-0.0014166802,0.1347049624,0.0760397166,-0.2752534449,-0.1492274702,-0.0142614702,0.3484376967,-0.0572859459,0.3520627916,-0.0388945304,-0.0381912291,-0.1972888112,-0.1449823827,-0.2161726952,0.1024223641,-0.3670475185,-0.0338461995,-0.0150918476,-0.1852909923,-0.1053255051,0.3172924221,-0.0513036028,0.1927736253,0.1294379979,0.1439675987,-0.1002326831,-0.0902262926,0.0809145719,0.1594159901,0.1486373544,0.2822788954,0.0600122698,0.003229551,-0.3305776417,-0.1897467226,0.0025670317,-0.0278814211,-0.0324557908,0.0732346922,-0.0350145772,-0.450147599,0.2254944891,0.1886160374,-0.0430544391,0.2807793021,-0.0074396078,0.1009571999,-0.096537061,0.0122182937,0.2783244848,0.1033927649,-0.1605129987,-0.1804709136,-0.1714458615,-0.0501717664,-0.303201288,0.0928248093,0.1805255115,0.0485841073,0.1871951669,0.089273721,-0.0540964268,0.0202349387,-0.1291193217,-0.1583362073,0.2625409663,0.0551409759,-0.0162990857,0.2405933589,-0.0453124866,-0.0048087128,-0.1340314597,0.0975835696,-0.3343697488,0.2123455107,-0.3812570572,0.1735198796,-0.3609201014,-0.2368099689,-0.0987068862,-0.0088431677,0.3678664267,-0.3480273187,0.3336296976,0.0685062632,-0.078144066,-0.0212075319,-0.1076178923,0.2046420425,-0.1081343442,-0.1003820226,-0.1654623598,0.2421965301,0.4265679717,0.1452926397,-0.0860859007,0.0135072637,0.1440583169,0.8647382855,0.1361807883,-0.1356698275,-0.4758000672,0.3001531661,0.018385727,0.0494007803,-0.0647801682,-0.1640854627,0.3293263614,0.0272320416,-0.175546214,0.120677717,0.4675885141,0.1378512383,-0.636767447,-0.1789665371,0.3183293939,0.0800496638,-0.047018636,-0.0316244401,0.169001326,0.0419173017,0.0185092632,-0.2328215837,0.2899684906,0.4480457306,0.2772974968,0.3473074138,0.1860031784,-0.0661409572,0.147070393,-0.0265920796,-0.6945602894,0.5669020414,-0.2257548124,-0.3093989491,0.0004434552,-0.3298888206,-0.0985443443,0.0407630838,0.2374934107,0.2225187719,-0.0592692494,0.3315542042,0.1590389311,-0.0570656545,0.1234892458,0.201455459,-0.2365746498,0.1906151325,-0.1024696156,-0.3099570572,-0.087907359,0.4058858454,0.4630976617,-0.0385711454,-0.2972506285,0.6062339544,-0.3367236853,0.2219973803,-0.207753554,0.1725815535,0.3421447873,0.0432178825,0.1004510596,0.640265584,0.2788161635,-0.006633033,0.1500734538,0.5099726319,0.3225123286,-0.0841197819,0.0058407453,-0.0700925589,0.1578987539,-0.380923003,0.2004105002,-0.1249167621,0.2203355134,-0.0428412259,0.1233304143,0.3034807146,-0.2406615913,0.2080424428,0.4111535251,-0.0495025255,0.1402168572,0.1452310681,-0.2254825681,-0.3757791221,0.096760124,0.0503878631,0.1027316079,0.1209968254,-0.0806207135,-0.0165553987,0.303905189,-0.3542489111,0.2387602031,0.1112501547,-0.2931630611,0.0804677457,0.2978738844,0.3431167006,-0.3636696935,0.0125109674,-0.0580732189,0.1891938001,-0.1028236076,0.2186203748,-0.3165397346,-0.7242130637,0.0452753268,0.011635324,-0.1649993807,-0.2254279256,0.3410965502,-0.0040558805,-0.146143809,0.0730768517,0.2235157937,0.3386718035,-0.3230760694,-0.1607567221,-0.2665176392,-0.3520283997,-0.7646338344,0.1563434005,0.1334807277,0.0910889506,-0.0997747481,-0.2050942183,0.0495026,-0.2066179961,-0.2803930342,0.0593728051,-0.1930378079,-0.1592743099,0.0604399331,-0.0513258129,-0.3704760373,-0.30636096,-0.0696584433,-0.2150731236,-0.1219453514,0.2372481823,-0.2543063462,-0.4921061397,-0.4656765461,-0.4422762394,-0.0322483145,-0.1254855841,0.2829438448,0.1891656518,0.0109011913,0.3011928797,-0.3569279015,0.0509781875,0.1335722059,0.3939691484,-0.2407732159,-0.4349057078,0.3319368958,-0.0415879153,-0.3880459368,-0.0198517106,0.1343091875,0.012537105,-0.0809675083,-0.4142768681,0.2543452382,-0.2746359408,0.0899759531,0.009953904,-0.0172472857,0.3932931423,0.0561426654,-0.2075251937,-0.0990403816,-0.0020561488,0.2808200717,0.305360198,0.3586889505,-0.3069497645,-0.1700131148,0.0090943836,-0.0627477542,-0.1406327039,-0.1806910783,0.4801080823,-0.1187062189,0.2573770583,-0.1555959433,-0.0260785371,0.15320611,-0.1746580452,-0.0243705586,0.7594629526,0.0098679028,-0.3053728044,-0.0623554513,-0.4967707098,-0.6557124257,-0.5054519773,-0.0710999891,-0.1472119391,0.1028703526,0.0756940022,-0.0286354516,0.0655078366,-0.5048431158,0.0048908638,0.178398937,-0.0184361283,-0.0303235166,-0.4487184584,0.2572936416,-0.3311731517,0.3807834983,-0.0455914326,0.0179858729,-0.1406439692,0.1126987934,0.2615388632,-0.1235289425,0.1620300859,0.0080644228,0.0107142553,0.1937716603,-0.1526547223,-0.0026583862,-0.0676651224,-0.1539684087,0.1622390896,0.3666578531,0.2105138302,-0.3340780735,-0.0206138268,-0.3664181828,0.1761317253,-0.1587234437,0.3018679321,-0.0259899814,-0.0875351876,-0.3132631183,0.0464815535,0.1347442865,0.5419646502,0.1058676764,-0.0064765355,0.0495667085,0.1654445976,0.2462301999,0.2353478819,0.5685104132,-0.2812584639,-0.1022603363,-0.0955822244,0.1699769944,0.0980623588,0.3154857755,0.0730682835,-0.0071165389,-0.0959687009,-0.3180464804,-0.1072642431,0.3794507384,0.04904918,0.3208692372,0.1861805618,-0.0182297118,0.09147688,0.1371534169,-0.0602616146,-0.0930709243,-0.0335601494,0.0151791759,0.1516720057,0.0272315592,-0.0395202339,-0.0180927198,0.1867927164,-0.503100872,0.620419085,-0.0008975349,1.2612766027,0.1932474077,0.2816486359,0.0721577853,0.087782748,0.2446509153,-0.5147622824,-0.1171675473,-0.2032101154,0.2356707901,-0.0472662598,0.4142568409,-0.0621234477,0.0443155766,-0.3159234822,0.1781927049,-0.1872994602,0.1955240518,-0.035326317,-0.0026298368,0.1325232834,-0.300635457,0.0733425021,0.1304015815,-0.1342473626,0.0691553503,0.0287463907,-0.3014602363,-0.3858850896,-0.1688604653,-0.33677122,0.1540474743,-0.0842109323,-0.0488001183,0.2964715064,-0.0255130697,-0.0934052691,0.2621548474,0.4480492473,0.3557026088,-0.3563840091,0.2410063595,0.0586012304,0.0303219743,0.167530328,0.280872643,0.2103385329,-0.1832425296,0.0490228757,-0.2144331783,-0.1476320326,-0.144391194,-0.1899322122,-0.0150394849,0.0474712551,-0.3482374549,-0.1098542213,0.1363227814,-0.3647044003,-0.3153707683,0.1473478675,-0.1926616877,-0.1967438459,-0.0320671499,0.0646404326,-0.2907950282,-0.1841213554,0.3845701516,0.1932658702,-0.0776366591,0.389677465,-0.0795742869,-0.1772791892,-0.3628854156,-0.0967231691,0.3678474724,-0.5188493729,0.2787390351,-0.0351923592,-0.2901853025,0.2774505317,0.7058497667,0.0832637995,0.3054120243,-0.1391423345,-0.3576372266,0.0650565624,0.0816209912,0.0107031884,0.1169683561,0.3884273767,0.4844404161,-0.0018329126,-0.2269229144,-0.3512559533,0.282918632,-0.3311580122,0.2968040705,-0.1983089298,-0.0599054396,-0.1090955287,-0.1000983566,0.0295821037,0.2926212549,0.0739406869,-0.2269210368,-0.3539001644,0.1085512415,0.2140667737,0.0486710519,-0.0785531029,0.0373881347,0.0986025408,-0.1802547425,-0.0454193838,-0.0931103006,0.0363427475,0.4058874249,0.1728239357,0.1893001944,-0.4371961057,-0.083361648,-0.1115267351,-0.0858326778,-0.1863231212,0.0574384667,-0.1623051912,-0.1390863359,0.0167729352,0.0274893325,0.0644909292,0.3385532498,0.2382492274,0.1366113275,0.0127728926,0.2071962953,0.1874638945,0.2976185083,-0.0923380628,-0.3938643932,0.2880011499,0.221812278,-0.1930512488,-0.3196573555,-0.0593457446,-0.3006494641,-0.0258532297,-0.0057220631,0.514333725,-0.2840704322,0.0530101061,0.0542978384,0.3861446381,-0.0326830521,0.2601012886,0.3640578091,-0.0215339083,0.0841228738,0.115810141,-0.0529191345,-0.2634904683,0.4093241096,0.0434785075,0.4805603027,0.4735280275,0.2181145698,0.009259766,0.1857922226,0.3121904433,0.0123445997,0.2001440674,0.348390758,0.0653688908,0.1877524555,-0.0765117779,-0.4006655812,-0.0514734387,0.1714372635,-0.4195893407,0.0145129776,-0.2791683674,-0.0921411589,-0.1563064158,-0.1374533921,-0.2533433735,0.1310721636,0.034561038,-0.0306947567,-0.0264238678,-0.3715113699,-0.1213361919,-0.2162144631,0.2671529949,-0.2426275462,0.4080795944,0.1121295691,-0.0916346535,0.014258543,0.1875007153,0.2888582647,0.2417786717,-0.0892018825,-0.0316662341,-0.0248266533,-0.1028054059,0.160505712,0.2988031507,0.2712019086,0.0629225299,-0.0137732606,0.0430957526,-0.178842932,0.2997131944,-0.1341922581,0.1837209761,0.1572059393,0.125420779,0.1399512589,-0.2528880537,-0.2671349049,-0.2466673255,-0.3726021051,0.0224107299,0.0989459604,0.0139142741,0.1304582506,-0.2182960957,0.1071203873,0.131856218,0.6128297448,0.3352708817,-0.1319619715,-0.1983515322,-0.2101863176,-0.3865143061,0.2486063242,0.1450932473,-0.2164658755,-0.0558161028,-0.1605853736,0.022555422,-0.0882061943,-0.3286328614,0.2831570804,-0.1476368308,-0.26384601,-0.4621167481,0.3192758858,0.1612539887,-0.130388245,-0.0027051677,-0.0093032671,0.1570312232,-0.0393295139,0.0818503499,-0.095719181,-0.5678967237,-0.0521933809,0.2463183701,0.1098825186,0.338601321,0.0294407737,-0.2972246408,0.1576794982,0.0301026646,-0.2671940923,-0.089721255,0.0611121804,0.2313503772,0.2754618824,0.074269101,0.3054203689,-0.1224624887,0.1058153212,0.1703940481,0.2043978721,-0.276596725,-0.0157868378,-0.0840039253,-0.0624207035,0.1036124751,0.5216318965,0.0051865005,-0.0991424173,-0.381359309,-0.5017678738,0.4677010477,-0.4028327465,-0.2314209193,-0.0074711251,0.1788390577,0.5742583871,-0.0790819973,-0.6015177369,-0.0861020237,0.2420421392,-0.0026647721,-0.4625404775,0.2797017992,-0.0554504506,-0.0989711508,-0.0140887406,0.3603188097,0.4206549227,-0.0952747762,0.1268050075,-0.2433051318]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/486","title":"Bookcorpus data contains pretokenized text","comments":"I'm afraid that I don't know how to obtain the original BookCorpus data. I believe this version came from an anonymous Google Drive link posted in another issue.\r\n\r\nGoing through the raw text in this version, it's apparent that NLTK's TreebankWordTokenizer was applied on it (I gave some examples in my original post), followed by:\r\n`' '.join(tokens)`\r\nYou can retrieve the tokenization by splitting on whitespace. You can then \"detokenize\" it with TreebankWordDetokenizer class of NLTK (though, as I suggested, use the fixed version in my repo). This will bring the text closer to its original form, but some steps of TreebankWordTokenizer are destructive, so it wouldn't be one-to-one. Something along the lines of the following should work:\r\n```\r\ntreebank_detokenizer = nltk.tokenize.treebank.TreebankWordDetokenizer()\r\ndb = nlp.load_dataset('bookcorpus', split=nlp.Split.TRAIN)\r\ndb = db.map(lambda x: treebank_detokenizer.detokenize(x['text'].split()))\r\n```\r\n\r\nRegarding other issues beyond the above, I'm afraid that I can't help with that.","body":"It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575","comment_length":146,"text":"Bookcorpus data contains pretokenized text\nIt seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575\nI'm afraid that I don't know how to obtain the original BookCorpus data. I believe this version came from an anonymous Google Drive link posted in another issue.\r\n\r\nGoing through the raw text in this version, it's apparent that NLTK's TreebankWordTokenizer was applied on it (I gave some examples in my original post), followed by:\r\n`' '.join(tokens)`\r\nYou can retrieve the tokenization by splitting on whitespace. You can then \"detokenize\" it with TreebankWordDetokenizer class of NLTK (though, as I suggested, use the fixed version in my repo). This will bring the text closer to its original form, but some steps of TreebankWordTokenizer are destructive, so it wouldn't be one-to-one. Something along the lines of the following should work:\r\n```\r\ntreebank_detokenizer = nltk.tokenize.treebank.TreebankWordDetokenizer()\r\ndb = nlp.load_dataset('bookcorpus', split=nlp.Split.TRAIN)\r\ndb = db.map(lambda x: treebank_detokenizer.detokenize(x['text'].split()))\r\n```\r\n\r\nRegarding other issues beyond the above, I'm afraid that I can't help with that.","embeddings":[0.0502685122,0.3206242919,-0.032077454,0.2856236994,-0.1109395847,-0.3141337633,0.0659819022,0.2292077094,-0.3773543835,0.1320857555,-0.1504071951,0.3783100247,0.2091655135,-0.0847186893,-0.0539451391,0.057601355,0.1612681448,0.1827619672,0.1995265633,-0.0846720189,-0.127008751,0.3466224372,0.0968654826,0.098527275,-0.2559473217,-0.3039924502,-0.172138229,0.2106493711,-0.1130991653,-0.2921229303,-0.0514739901,0.2297893167,-0.0205983426,0.3393049538,-0.0001025469,-0.2117831707,-0.0415515676,0.0248816162,-0.2459975481,-0.0421917364,0.3318285346,-0.0583667383,0.3199644685,-0.3479763269,-0.3073748052,0.0060196351,-0.150744319,0.0970791578,0.287876308,0.2866069674,0.256216675,0.1728054583,-0.1479442269,0.2911150157,0.2293517888,-0.2285568714,-0.1792050749,0.1154583991,0.3404251635,0.1576986462,-0.1479932815,0.3168506324,-0.1120618507,-0.0993605703,0.0107641099,0.0866040736,-0.0468522534,-0.1749296784,0.1741347313,0.1410867125,0.2253863961,-0.2929340899,-0.1760388315,-0.2555312514,0.0676854029,-0.0627081841,0.3575012684,0.1597668827,0.179701075,0.0991524979,-0.3076472878,-0.1184087023,0.0067069819,0.1861819029,-0.1890390068,0.35173738,-0.0275463499,-0.0521571487,-0.0401723534,-0.167187959,-0.230284363,0.0457763374,-0.2839516103,0.0290542562,-0.0194794647,-0.2148331702,-0.0329769664,0.2633591294,0.0583551377,0.2064031661,0.1382035911,0.2117533535,-0.1937957406,0.0104676699,0.1419467181,0.2237099856,0.072224699,0.1600134671,0.1111708805,0.0893621519,-0.2414986044,-0.0320880078,-0.0146005843,-0.0237487461,0.0396704897,0.0728731677,-0.0326367132,-0.3373933733,0.0546515249,0.1993586421,-0.2055116147,0.3033297956,-0.0416620336,0.1658537388,-0.0549800098,-0.0120549733,0.2828995585,0.0206976235,-0.1756383926,-0.1595058292,-0.1973602176,0.0380616337,-0.3233500719,0.1083543003,0.2031531334,0.0035495893,0.1434016228,0.0539375618,-0.1815329641,-0.043326702,0.005220796,-0.1907522529,0.1267917752,-0.0700531974,0.0332464091,0.1602682173,-0.0991465002,-0.1501461267,-0.2041988075,0.0204513539,-0.3056629598,0.1298172176,-0.3631924689,0.2384641618,-0.3120621741,-0.2469170988,-0.0951263607,0.1402453333,0.2168128788,-0.2126640677,0.2082401365,0.0645192713,-0.1796924174,-0.0139365802,-0.0532152057,0.2223668247,-0.189180851,-0.1539739519,-0.0387751758,0.0368672162,0.2192967087,0.204496786,-0.139821589,0.1316277832,0.0368637405,0.714405477,0.0238878001,-0.1102677584,-0.6006366014,0.2778592706,-0.1638279259,0.0712636337,-0.1692005247,-0.2280033976,0.1500805318,0.109369576,0.011197539,0.3238185644,0.476390332,0.1905881763,-0.5679761767,-0.2778165936,0.2300938219,0.0829983205,-0.1315531135,-0.0157731473,0.1510691792,-0.0035497241,0.0716411769,-0.1447981298,0.2080217898,0.4561246932,0.1945448816,0.2754332125,0.1379222125,-0.108359836,-0.0035321764,-0.0280957203,-0.6072434783,0.5558738112,-0.2705800831,-0.3538455367,-0.0436139405,-0.3035729825,-0.0848913714,-0.0426058136,0.3206100166,0.206859991,-0.0636004657,0.3263259232,0.1955290437,-0.0546719879,0.0610972568,0.0985155851,-0.4020552635,0.0388387218,-0.157295078,-0.24166435,-0.0188539699,0.2859687805,0.257676512,0.0964295715,-0.1627963483,0.512216866,-0.4067947865,0.3954562545,-0.1029731035,0.2249748111,0.2453071475,0.0368538499,0.1425614804,0.5524895191,0.2463425398,-0.020252496,-0.0108900508,0.5778490305,0.3246223629,-0.0902370438,0.0885414109,-0.116036281,0.1845515966,-0.3483377695,0.0513697453,-0.1625525802,0.1203030497,0.2504860163,0.1890166998,0.1793299317,-0.2093157917,0.362485826,0.4404110909,-0.0698240474,0.1454109401,0.1719345897,-0.316878736,-0.2977505028,0.0184552837,0.1694018394,0.016636176,0.1354276985,0.0464465618,0.0763096884,0.1118471324,-0.2814236581,0.2418965399,0.0813676268,-0.2202339619,0.1963642985,0.3998716474,0.1825363934,-0.4641732275,-0.0631450862,-0.0252164043,0.1623710394,-0.1716697812,0.3289465606,-0.3300445974,-0.6390396357,-0.1327872425,0.0547671132,-0.1161165982,-0.1989362091,0.2469795197,0.2347543836,-0.1379943043,0.0465457179,0.3245829046,0.3713593483,-0.2469572276,-0.141789943,-0.22850658,-0.3313556314,-0.6871263981,0.187860623,0.1113764718,0.0861142948,0.0377117433,-0.1410930306,0.016011877,-0.359277159,-0.3051671684,0.0756579638,-0.2178703099,0.0663701743,0.062879689,-0.0984486565,-0.3571196795,-0.2685876191,-0.024832895,-0.1988770068,-0.1679217517,0.0876095146,-0.211187467,-0.2750479877,-0.388892889,-0.4716968536,-0.0215845406,-0.1205599308,0.3580135107,0.2281888276,0.0478285849,0.2902409434,-0.2424391508,0.1059382334,0.3240024447,0.3039756715,-0.3152661622,-0.3232303858,0.3678918779,-0.0866242573,-0.4257172644,-0.0392808989,0.0964044705,-0.0240303054,-0.009690254,-0.3700835705,0.171119988,-0.2685033977,0.0882919729,0.075406827,-0.0319826454,0.3750682771,0.1537846476,-0.2389876246,-0.0523224026,-0.1377874464,0.2527928054,0.2869014144,0.3607076406,-0.2157279253,-0.0460317656,0.0932907611,0.1086268947,-0.2304874063,-0.1206873357,0.3998340666,-0.1271839589,0.1041408777,-0.1783699244,0.0119410465,0.0047446485,-0.2672373354,-0.0580006391,0.7442845106,-0.0562028252,-0.3819235563,-0.0194861293,-0.411041528,-0.6470891237,-0.451921314,-0.0588593259,-0.1319078356,0.0151589494,0.0825076476,0.0060698777,0.1075698137,-0.4360057116,0.0010979424,0.0672482699,-0.0700572133,0.0234593172,-0.2529757321,0.2467830777,-0.4832003117,0.2708823681,-0.0280913822,0.0848309994,0.0159986913,0.0260652192,0.2659101784,-0.3306407332,0.1358569562,0.0492858179,0.0616585761,0.2379518747,-0.2220129967,0.0675275624,-0.0936080962,-0.2850674689,0.0268585831,0.3058536053,0.2077633291,-0.4310358465,-0.0104457559,-0.2550184131,0.1295178384,-0.2020731866,0.3922217488,-0.0559090301,-0.2724835873,-0.4281615615,0.088595815,0.0837259591,0.5127240419,0.0020353547,0.0375510715,0.0914578959,0.0699014217,0.282825917,0.2051466554,0.5005566478,-0.3195250928,-0.0289351773,-0.0867306665,0.2290772945,0.1046167016,0.3962088823,0.1842536777,0.0687597245,-0.0918967351,-0.2497075349,-0.1034452841,0.4508930147,0.0933599845,0.298909992,0.1022295132,-0.0504258722,0.0568539612,0.0936959907,-0.0392993093,-0.0523021445,-0.0866834298,-0.0688020438,0.2607877254,0.0625126883,-0.1282547414,0.0160364993,0.1689428091,-0.5297275186,0.6358743906,0.0817720294,1.2664068937,0.1259853393,0.2774372995,0.1221759841,0.1035661176,0.242363289,-0.4548504949,0.0130382525,-0.3154565394,0.1917826086,-0.0502261855,0.3567132354,-0.1773014963,0.189102143,-0.2983834445,0.157602191,-0.1464833617,0.0954728723,0.1002279893,0.2019843012,0.055555854,-0.2359241843,0.0622628629,0.2006782591,-0.1212719679,0.09996631,-0.0230808184,-0.2644629776,-0.2481749207,-0.1299494058,-0.3037829697,0.2492029816,-0.1399762034,0.007249543,0.1680199802,-0.1097241193,-0.1510291994,0.2088519335,0.3138302565,0.2617882788,-0.383323729,0.3085211217,0.0770520121,-0.1476068944,0.220404774,0.1608800888,0.3945513666,-0.1488298178,-0.0294075422,-0.235410437,0.0630314872,-0.234452039,-0.1290118992,-0.0519688129,-0.0692102984,-0.2831617594,-0.1294769943,0.172229737,-0.3984479904,-0.2890193164,0.1957060546,-0.2353201658,-0.1418458819,-0.1430527121,0.0883559287,-0.3133049309,-0.1746404767,0.4973690212,0.2756565511,-0.0691615939,0.4139432609,0.0083497846,-0.1509474963,-0.3615687191,-0.2265955657,0.2644393742,-0.2990568876,0.189054206,-0.1237003431,-0.3279196024,0.2532061636,0.6107335091,-0.0646060482,0.328861773,-0.1273344755,-0.3482036293,-0.0388989821,0.0796699151,-0.0323107764,0.1915735304,0.3719386756,0.3681503236,-0.0934032425,-0.1438858509,-0.4209375978,0.2488287836,-0.3923137784,0.2334463745,-0.1301633865,-0.0854213983,-0.0414089486,0.0143725015,0.1004480198,0.3575179875,0.0731341541,-0.318749547,-0.298866421,0.0751723647,0.1684569269,-0.0097893039,-0.1823962629,0.0656283349,0.0205029678,-0.1436796933,0.1260128766,0.0330823846,0.1086089686,0.3633995354,0.1074027941,0.206494242,-0.2905073762,-0.2369407415,-0.0867214575,0.0155248707,-0.1578540504,0.1750104874,-0.1986194998,-0.0538297407,-0.1889486015,0.0618186034,0.1741078347,0.5176964402,0.2648685873,0.1763467938,-0.0339082107,0.022254141,0.1865950674,0.2837531269,-0.2215486467,-0.4664621949,0.3147236705,0.2928294539,-0.1725576222,-0.2323793322,-0.0907034129,-0.5723102689,-0.0013675223,0.0893005431,0.4483964443,-0.074014239,-0.0162083711,0.0517580807,0.4621597826,-0.1148688793,0.021964144,0.33470577,-0.046840895,0.0239960607,0.2213386297,0.0655297935,-0.0683628768,0.4197322428,0.0771020502,0.3663988411,0.4050022066,0.2621216178,-0.0193109792,0.1859980524,0.2309220284,0.0941111147,0.2074230611,0.3001117706,0.0419664718,0.1585603654,-0.2035882175,-0.2547902465,-0.0800383762,0.3237069249,-0.3548710942,-0.1016118452,-0.2776465118,-0.217257008,-0.2227424234,0.0889212713,-0.1792000383,0.1237813681,-0.0473928116,-0.0068240124,-0.0887009352,-0.3339597583,-0.0968102962,-0.2567664385,0.3450835943,-0.3081307709,0.3353688419,-0.0239988603,-0.0117398575,0.0344992168,0.2444020063,0.2280296534,0.1660609245,-0.2368104458,0.0084655369,-0.2022775114,-0.0699958727,0.0739986151,0.3090717196,0.2560390532,-0.0349520966,0.1519611925,0.1186423451,-0.2021938711,0.1569798738,-0.0205566473,0.0906974748,0.0912725851,-0.0068206666,0.0685005784,-0.1779577583,-0.2704570293,-0.2943276465,-0.210618183,-0.1027152464,0.1102602109,0.0450011194,0.2079066336,-0.2224058658,0.1511448324,0.2386326492,0.5989447832,0.3897720575,0.0290972106,-0.1135457382,-0.1662546545,-0.4377086461,0.228262201,0.1493016928,-0.2425405383,-0.0323457383,0.0235697422,-0.0043589096,-0.0257835835,-0.452344805,0.3621743023,-0.1675101817,-0.1839092523,-0.3424536586,0.3002201021,0.129579708,-0.1052916721,-0.0413144864,-0.1630313098,0.0921328589,-0.0677634031,0.166281417,-0.1584757119,-0.4944362938,-0.0925905704,0.2201500237,0.1942606121,0.3134510517,-0.0123511264,-0.3465572298,0.0573732331,-0.0488117896,-0.0741133839,-0.2141726911,0.0072930986,0.1437648982,0.2346392125,0.2637171447,0.3825136125,-0.1025282666,0.2364631295,0.184326455,0.1577815711,-0.3397211432,-0.068281427,-0.0784968436,-0.0010875044,-0.0723825693,0.6917637587,0.0642041117,0.0776449293,-0.3717277348,-0.422644347,0.4227739871,-0.3316350877,-0.2500548363,0.1223129556,0.1393047273,0.4561154544,-0.0476736091,-0.6378759146,-0.0586780012,0.1106198654,-0.0203513708,-0.6066350937,0.2190053314,-0.0008714436,-0.0870744064,-0.1062999666,0.2862390876,0.4174633622,0.0130852107,-0.0244562291,-0.1725815982]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/486","title":"Bookcorpus data contains pretokenized text","comments":"Ok I get it, that would be very cool indeed\r\n\r\nWhat kinds of patterns the detokenizer can't retrieve ?","body":"It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575","comment_length":19,"text":"Bookcorpus data contains pretokenized text\nIt seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575\nOk I get it, that would be very cool indeed\r\n\r\nWhat kinds of patterns the detokenizer can't retrieve ?","embeddings":[0.0604030266,0.2112906575,0.0049694735,0.4674902856,-0.1865660697,-0.3642662466,0.0308460351,0.2487041056,-0.4858066738,0.1431304365,-0.1751267612,0.4655150771,0.1896488965,-0.1176308468,-0.1269879937,0.2603909969,0.0707708225,0.2080192715,0.0444150902,-0.1715232879,-0.0605966896,0.3185583949,0.0735049993,-0.0343129635,-0.3340622485,-0.2613661587,-0.1723051667,0.2871147692,-0.1565576792,-0.3885434866,0.0394663811,0.2147516757,-0.0296005215,0.3092271984,-0.0001112136,-0.1212928072,0.0267073531,-0.1100272685,-0.1816276312,-0.0909245014,0.1232385263,-0.0185541585,0.3582623005,-0.5265985727,-0.2822774649,0.2261689901,-0.1437604278,0.1404644549,0.2870262265,0.4154458344,0.187061131,0.2642733753,-0.1532092392,0.1688100398,0.3264146149,-0.0386349112,-0.2777245939,0.0013342452,0.386443764,0.3084563315,-0.1662812531,0.2464140952,-0.1460218281,0.0169743355,0.1081547067,-0.0495128781,0.1588053852,-0.2349087,0.2553260326,0.3478803337,0.2573385835,-0.038626384,0.0085572451,-0.2316288054,-0.13998124,0.2030919194,0.2722402513,0.127852276,0.0821161717,0.1513773948,-0.1062148735,-0.0710781962,-0.0499277711,0.1270304024,-0.1186060011,0.3517419696,0.0026368664,-0.0285530593,-0.0115611814,0.0066976305,-0.1764365286,0.0958928987,-0.3348130584,0.0043225819,-0.0044847033,-0.2323170602,0.0973935872,0.2527470291,-0.045265615,0.4082658589,0.3445390165,0.2949661016,-0.1966409832,0.0793915167,0.0403271466,0.1573100388,0.0555436425,0.2430897802,0.1665278375,0.0107505927,-0.3805959225,-0.0743013397,0.0317922011,0.0189552568,0.0036563515,0.061177168,0.0374833718,-0.483212769,0.1632658392,0.1853301674,-0.2136845589,0.1784506589,0.0564471483,0.0898110941,-0.0515366346,0.0880343542,0.0918418765,-0.0330606177,-0.0517974459,-0.199009493,-0.1495576948,-0.1054217443,-0.3613607883,0.0929126143,0.2237260342,0.034839619,0.1418288052,-0.0173506569,-0.172521919,0.0216433182,0.020218771,-0.2706736922,0.4571626782,-0.0256435797,-0.1910112649,0.0324009992,0.0236856248,-0.1179529876,-0.303106606,-0.0246074237,-0.3003565371,0.1305866539,-0.2167600542,0.1804510504,-0.3390056193,-0.1201605648,0.0040062373,0.2269220352,0.1144161671,-0.1300052851,0.2640655041,-0.0077673234,-0.1130756289,-0.0804778636,-0.1571694463,0.2868357599,-0.291061759,-0.1324337721,-0.072908476,-0.0038544538,0.2922712564,0.1737480909,-0.0803821608,0.0385204628,0.0571063794,0.8963900208,-0.0338544361,-0.0474755131,-0.5306514502,0.2030206472,-0.0835102797,0.1309683174,0.0783679411,-0.100826405,0.2333675176,-0.0387079976,-0.1269482821,0.1347050071,0.5446136594,-0.0159926135,-0.5618402362,-0.2940213084,0.2203255743,0.052337084,0.0197996944,0.133662045,0.2531479299,-0.0214614328,0.0830839276,-0.230712682,0.1561270654,0.3136553168,0.3593850434,0.406624198,0.2494668514,-0.1262724549,0.1709288657,0.085590668,-0.7975932956,0.4854716361,-0.110042654,-0.2766066194,0.0316400863,-0.2976920903,-0.1390509605,-0.0308803059,0.2581126988,0.2348937541,0.0465355329,0.434281379,0.158879444,0.0390868373,0.1868077815,0.0412401073,-0.4006344676,0.1447036713,-0.1005987972,-0.2781329155,-0.0628803223,0.3709408343,0.2360568345,0.0312169883,-0.0716287792,0.289059639,-0.3318715096,0.1124550849,-0.2324206829,0.1594908535,0.1141446009,0.0137153827,-0.0172763169,0.6242741346,0.2798282206,-0.1008123234,0.3238973916,0.4158972204,0.377671212,0.0362377726,-0.0003744426,-0.1702014506,0.1261226833,-0.3321750164,0.0453589559,-0.2139383256,0.2807053626,0.1363817453,0.2800716162,0.1733032465,-0.2103452682,0.2834628522,0.3500041068,-0.0286059752,0.1028477326,0.2167858332,-0.240759626,-0.2014989555,0.0918725133,0.1266491264,0.1089738011,0.0998078138,0.0215668827,0.0722229928,0.0722113624,-0.3865079284,0.1687373817,0.2188985646,-0.2239869684,0.129643321,0.3718090057,0.1892755926,-0.4481157362,-0.2308838069,-0.1116982102,0.0805018842,0.0273709688,0.2684747577,-0.5430806279,-0.6649031043,0.0325669535,0.0656293556,-0.0501057059,-0.2324708253,0.2761966586,0.1472476274,-0.1875771284,-0.2033764124,0.2906611562,0.3883592784,-0.2766059041,-0.0873173848,-0.4001852274,-0.470796138,-0.6657466888,0.1560483277,0.1820809692,-0.0244103279,-0.0165222082,-0.1729425639,0.0603350289,-0.1272507757,-0.2706115842,0.066847831,-0.0042660078,-0.0731112137,0.0747511908,-0.1444064975,-0.4029604197,-0.2654490471,-0.0277516767,-0.1547914296,-0.2307020277,0.2833867371,-0.2646175623,-0.312178582,-0.3437876999,-0.4522251487,-0.1565519422,-0.1983984411,0.5167419314,0.3702398837,0.0481463149,0.4064434469,-0.348472327,0.1662015766,-0.0078285737,0.4200567305,-0.2951849997,-0.2476711422,0.449470073,-0.0966084152,-0.4395467639,-0.0071631572,0.0947204605,0.1041146815,-0.0838209838,-0.4395299852,0.2352150977,-0.4147758782,0.2214820385,0.0919786841,-0.0054008611,0.4141243994,0.1176143065,-0.1854997575,-0.0184925795,0.0200538039,0.38923648,0.2683612108,0.1220119298,-0.1243890002,-0.1852469444,0.2529534101,-0.0080480669,-0.154407531,-0.0834846944,0.535487175,-0.0166322477,0.1233316883,-0.2894328535,0.080849506,0.0365821421,-0.363463521,-0.0181334727,0.6952674985,-0.0387203954,-0.4206862152,-0.1525160372,-0.3626703322,-0.3947038054,-0.5091341138,-0.0947225317,-0.0894459859,-0.1103576869,0.1220048591,0.1227145791,0.1484939009,-0.3748821914,-0.0294180121,0.0291140676,-0.0574427545,0.101572223,-0.2822040617,0.1459388882,-0.5896186233,0.2196616381,-0.1175741255,-0.0568125919,-0.1042476371,-0.0132991001,0.2554581165,-0.1549900323,0.0596973673,0.0823835358,0.1571889222,0.3531690538,-0.4536797702,0.0466437675,-0.1597648412,-0.3209364414,0.0148513746,0.2244004309,0.1983937323,-0.4008465111,0.2173607349,-0.4708299637,0.2582645118,-0.1026700661,0.3557087779,-0.0527283102,-0.1844118834,-0.4908764362,0.0296219792,0.14421013,0.4182726443,0.0507552586,0.01312216,-0.007545691,0.0741120353,0.2711829841,0.3257134557,0.4172253907,-0.3802880049,-0.0506035537,0.0162402205,0.2785224617,-0.0386804193,0.3240661323,0.1429882497,-0.0012642809,-0.145013839,-0.2655173242,-0.2225731909,0.4299736619,0.1031809747,0.3510273993,0.1296644807,-0.0288923047,0.0275590029,0.2613199055,-0.2476960272,-0.1647700369,-0.1122042984,-0.0371275358,0.2096108347,0.0568736792,-0.0862637088,-0.0101329377,0.0833755732,-0.4478079379,0.5934388041,0.1470458061,1.3137550354,0.2581230104,0.3736341,0.1208668649,0.0678044483,0.1194816753,-0.4052529633,-0.0135768708,-0.2606821358,0.1679503769,-0.0823445469,0.2353165895,-0.1076756939,0.2000227571,-0.406671375,0.1032233983,-0.1559080034,0.2173660547,0.1010726765,-0.0092816548,0.0906760395,-0.1309901476,0.2300188541,0.1524742395,-0.0055047758,0.1577048153,-0.0476447791,-0.3615916371,-0.3194209337,-0.0812914222,-0.280269146,0.185831368,-0.144321084,0.1160686761,0.300259769,-0.0802204609,-0.2489446551,0.2821151614,0.3190871477,0.2485009134,-0.4652882516,0.1221020967,0.0649990663,-0.1323267817,0.3016616404,-0.021849772,0.4099272788,-0.2100420743,-0.0312505066,-0.0809152871,-0.0453299992,-0.2504413128,-0.2505156398,0.2192789763,-0.1018663496,-0.0866020694,-0.1273266077,0.1625170708,-0.5676107407,-0.2275832593,0.1299807131,-0.1229792982,-0.2572291791,0.0325813182,0.0439509563,-0.296992451,-0.217534095,0.4929182827,0.1279948801,0.1437836289,0.2659582794,0.0155057497,-0.2118877769,-0.3199820518,-0.0646491274,0.3073034883,-0.4513332844,0.3142532408,-0.052931767,-0.3912534714,0.2202435285,0.6401067376,-0.0648359284,0.2356961519,-0.3432988226,-0.3344877958,-0.1428493708,0.0723183304,-0.0759882927,0.1215203181,0.3601948321,0.3767555356,-0.1318208426,-0.1066307127,-0.3545944989,0.2168207914,-0.4204818904,0.2378807664,-0.3747776151,-0.2061500847,-0.1499353498,0.0259233397,0.0081206504,0.5068909526,0.0129051227,-0.2600454986,-0.3013978899,0.1099210307,0.2588219941,-0.0234448183,-0.1358154714,0.0719740465,0.0002931002,-0.1304661334,0.1453938335,0.1555185914,0.0607334077,0.3466275334,0.1490340382,0.1122617722,-0.4103755653,-0.0152569441,-0.168470189,-0.0485253409,-0.2342363745,0.1298634857,-0.2454683185,-0.097214222,-0.0906306952,0.1664320827,0.1093839481,0.3238173127,0.2405936569,0.2146869898,0.0495452322,0.0479723662,0.0919283181,0.2279502749,0.0072845952,-0.4190809727,0.4563711882,0.2205511034,-0.2072083056,-0.3670642376,0.0264971759,-0.3992231488,0.1342362911,0.1030328646,0.4256849289,-0.3307785094,0.044364661,0.0788631886,0.2585692406,-0.0978071094,0.2458579391,0.5148998499,0.0579986349,0.0307019856,0.2223694772,-0.0739789754,-0.0304962266,0.4679300189,0.1013525575,0.5709165335,0.492591083,0.158223331,0.175416559,0.2502923608,0.2761949897,0.2544514835,0.3463716507,0.4401054978,-0.0277228374,0.0877457485,-0.1820437759,-0.1817067116,-0.1663595736,0.1184738353,-0.4020214081,-0.0162040424,-0.2048482597,-0.2311948687,-0.1976766586,0.0143797258,-0.2363506407,0.2180002928,-0.101202026,-0.0192652233,-0.1090383604,-0.1648951471,-0.0554527789,-0.3457092643,0.3126297891,-0.2368155569,0.2367736548,-0.0280486289,-0.0491227731,0.1689564586,0.1625014991,0.1855438501,0.2687326074,-0.167534858,-0.0098342877,-0.1640844941,-0.2072373182,-0.0074246903,0.4367580712,0.1841107905,-0.0729749203,0.2002144903,0.0352903567,-0.1615734398,0.2720960975,-0.0540759452,0.0946178213,0.0394379981,-0.0451987535,0.1204200983,-0.2497218847,-0.145922482,-0.183353141,-0.2593675852,0.0367671512,0.0272676442,-0.0264102984,0.1567187011,-0.1840615273,0.0957191065,0.0984390229,0.5475634933,0.3088125288,-0.0921852812,-0.0028425066,-0.0847655162,-0.3762497008,0.2123697847,0.0569202676,-0.329870224,-0.0072314595,-0.0174498148,0.0328906476,0.0972705483,-0.3886375129,0.2236616462,-0.1885614991,-0.1668173224,-0.3574567437,0.2894044816,0.1753381193,-0.1492225677,-0.0366453081,-0.2643887997,0.1396127492,0.0647705197,0.143591702,-0.0932576135,-0.6359952688,-0.1098389626,0.3305190802,0.1894246638,0.388525486,-0.0265478157,-0.2530365884,0.0473787151,-0.0632921308,-0.0674686357,-0.0986117423,-0.1176718175,0.1615013629,0.2237387151,0.0077377781,0.402238518,0.0059390846,0.3810948133,0.1957579106,0.3764248192,-0.2912952006,-0.0638091713,-0.2182696611,0.0347629748,-0.0692657679,0.53244555,0.0608735941,-0.2226440459,-0.2715649903,-0.4242271185,0.5321038961,-0.3395157456,-0.2588925958,-0.0994960219,0.0125077246,0.3902791739,0.0195650421,-0.443123728,-0.1875344217,0.1218617186,-0.076257959,-0.5258943439,0.1829117835,0.0751643032,-0.1372605562,-0.1607269347,0.275514096,0.446867615,0.0316404328,0.0359923057,-0.2550365925]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/486","title":"Bookcorpus data contains pretokenized text","comments":"The TreebankTokenizer makes some assumptions about whitespace, parentheses, quotation marks, etc. For instance, while tokenizing the following text:\r\n```\r\nDwayne \"The Rock\" Johnson\r\n```\r\nwill result in:\r\n```\r\nDwayne `` The Rock '' Johnson\r\n```\r\nwhere the left and right quotation marks are turned into distinct symbols. Upon reconstruction, we can attach the left part to its token on the right, and respectively for the right part. However, the following texts would be tokenized exactly the same:\r\n```\r\nDwayne \" The Rock \" Johnson\r\nDwayne \" The Rock\" Johnson\r\nDwayne \" The Rock\" Johnson\r\n...\r\n```\r\nIn the above examples, the detokenizer would correct these inputs into the canonical text\r\n```\r\nDwayne \"The Rock\" Johnson\r\n```\r\nHowever, there are cases where there the solution cannot easily be inferred (at least without a true LM - this tokenizer is just a bunch of regexes). For instance, in cases where you have a fragment that contains the end of quote, but not its beginning, plus an accidental space:\r\n```\r\n... and it sounds fantastic, \" he said.\r\n```\r\nIn the above case, the tokenizer would assume that the quotes refer to the next token, and so upon detokenization it will result in the following mistake:\r\n```\r\n... and it sounds fantastic, \"he said.\r\n```\r\n\r\nWhile these are all odd edge cases (the basic assumptions do make sense), in noisy data they can occur, which is why I mentioned that the detokenizer cannot restore the original perfectly.\r\n","body":"It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575","comment_length":244,"text":"Bookcorpus data contains pretokenized text\nIt seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575\nThe TreebankTokenizer makes some assumptions about whitespace, parentheses, quotation marks, etc. For instance, while tokenizing the following text:\r\n```\r\nDwayne \"The Rock\" Johnson\r\n```\r\nwill result in:\r\n```\r\nDwayne `` The Rock '' Johnson\r\n```\r\nwhere the left and right quotation marks are turned into distinct symbols. Upon reconstruction, we can attach the left part to its token on the right, and respectively for the right part. However, the following texts would be tokenized exactly the same:\r\n```\r\nDwayne \" The Rock \" Johnson\r\nDwayne \" The Rock\" Johnson\r\nDwayne \" The Rock\" Johnson\r\n...\r\n```\r\nIn the above examples, the detokenizer would correct these inputs into the canonical text\r\n```\r\nDwayne \"The Rock\" Johnson\r\n```\r\nHowever, there are cases where there the solution cannot easily be inferred (at least without a true LM - this tokenizer is just a bunch of regexes). For instance, in cases where you have a fragment that contains the end of quote, but not its beginning, plus an accidental space:\r\n```\r\n... and it sounds fantastic, \" he said.\r\n```\r\nIn the above case, the tokenizer would assume that the quotes refer to the next token, and so upon detokenization it will result in the following mistake:\r\n```\r\n... and it sounds fantastic, \"he said.\r\n```\r\n\r\nWhile these are all odd edge cases (the basic assumptions do make sense), in noisy data they can occur, which is why I mentioned that the detokenizer cannot restore the original perfectly.\r\n","embeddings":[-0.0130970906,0.2090142518,-0.0026562868,0.3173747659,-0.1956932545,-0.2996238172,-0.0173011087,0.2006196082,-0.4918516278,0.0816216096,-0.172097221,0.2842743099,0.1636684388,-0.2349403501,-0.1071667597,0.2111709714,0.1792732924,0.2823951542,0.220247522,-0.153670311,-0.1823280305,0.2915745974,0.1904963702,-0.0035161814,-0.2633253932,-0.2643784583,-0.1785983741,0.3562456369,-0.1762152612,-0.3386277854,-0.1300115585,0.2864885032,-0.0567627177,0.3584989309,-0.0001062125,-0.2521669269,-0.0237554628,-0.0865369737,-0.3363088369,-0.0799458772,0.2673516572,-0.064941071,0.2785297632,-0.3779522181,-0.3191954195,0.0946913511,-0.1585107893,0.0207608454,0.2960724831,0.2338730991,0.2192092985,0.1864215434,-0.119373396,0.2506962121,0.3235376179,-0.1017344892,-0.1702123731,0.0435539894,0.3703986406,0.3260974288,-0.1662236303,0.3166057467,-0.1516396105,0.0288449693,0.0620671697,-0.1197543442,0.0588301048,-0.1836254001,0.1840665042,0.2314481139,0.2577361763,-0.1910746992,-0.1388308704,-0.2062944174,-0.0055133468,-0.0866998211,0.2901036143,0.1626656651,0.1006433815,0.1169619262,-0.2036015093,-0.1254278868,0.1330465078,0.2178181261,-0.0939432234,0.3291178942,-0.0937990099,-0.0315528624,-0.0638477504,-0.0639029443,-0.2752233744,0.0771327317,-0.3466393352,-0.0471422412,-0.0716824234,-0.2270145118,0.0099338954,0.3180498183,-0.0523345694,0.291467458,0.1656623334,0.20682019,-0.1103693247,0.0022941055,0.142885834,0.1520640403,0.0951022878,0.2186256051,0.1616232544,0.0547114499,-0.2744164765,-0.0449897759,0.0262068305,0.0409516282,0.0673123077,0.1175812185,-0.0945215449,-0.4092077911,0.1104818434,0.1627935916,-0.175558269,0.351349324,0.0304975193,0.1711644083,-0.1282082945,0.1095000282,0.2333882898,0.0160870925,-0.1464368999,-0.1457025409,-0.1949326694,-0.0282436144,-0.3139037788,0.1749850214,0.2165554911,-0.024981942,0.1767987311,0.0159949977,-0.0900460109,-0.0779916942,-0.0937322676,-0.1390446872,0.2256879061,-0.0904374868,0.0073019317,0.2255372107,-0.067043297,-0.0109519763,-0.272883147,0.0048361868,-0.3288038075,0.1000960693,-0.2406043559,0.2492175847,-0.3356083035,-0.1833829433,-0.1336333454,0.0967791751,0.249566853,-0.205788523,0.2461419255,0.0952634141,-0.0845146403,0.0397761501,-0.1700638384,0.1673566103,-0.1764993668,-0.166233927,-0.0244440436,0.2212650031,0.2749224603,0.1089541391,-0.1420920938,-0.0086042462,0.0813517869,0.7183502316,-0.0174182076,-0.0456807055,-0.5037994981,0.2438368201,-0.0503257662,0.0734820142,-0.0498397648,-0.1478775591,0.2721961737,-0.025697317,-0.1284986287,0.1694664806,0.5301482081,0.1767308861,-0.5494855046,-0.2097418457,0.1637216359,0.0617381781,-0.1836184412,-0.0272415951,0.2119203657,0.1857924163,0.0499813482,-0.2376085967,0.2587558031,0.3918693662,0.3059104085,0.3005813062,0.2135974318,-0.1132968068,0.1385641843,-0.0084675467,-0.5750662684,0.5548720956,-0.1922527254,-0.325068891,0.0255882666,-0.3308907747,-0.1080776826,-0.0202899445,0.2663990557,0.2701419294,-0.0265747681,0.2296952158,0.1098188907,-0.0887290388,0.0138865039,0.144550696,-0.2879339457,0.0096704438,-0.1162086129,-0.2436328828,-0.091394648,0.3883041143,0.3055767119,0.0095238565,-0.1216301098,0.5886529684,-0.2743594944,0.2116333544,-0.3422011733,0.1967356205,0.2266284823,0.0910278186,0.1045902595,0.6204184294,0.3134614527,-0.0619515926,0.1241833344,0.4847055972,0.5381141305,-0.0348015986,0.135763973,-0.2097647935,0.1407641023,-0.277854681,0.0349247418,-0.1607512832,0.1870877594,0.2222881168,0.2258031666,0.2511379421,-0.1797742993,0.3216660321,0.3824762404,-0.1144577339,0.1248017922,0.1802485585,-0.3248032629,-0.3172587156,0.0615759529,0.0733916759,0.0257054139,0.1317723989,0.076852113,0.0449810438,0.1044769511,-0.3250133991,0.2166339904,0.0959457681,-0.3224861324,0.1516514719,0.3414753675,0.1739849746,-0.4222617149,-0.0884152651,-0.0366610624,0.0930631384,-0.1447469741,0.2682285607,-0.425195843,-0.6615489125,-0.0893082097,-0.0604751445,-0.1013161093,-0.247446835,0.3624102473,0.2117522061,-0.1529857814,-0.0137812076,0.2746281624,0.3462324739,-0.3052076101,-0.1405730844,-0.1333603561,-0.3698670864,-0.606369853,0.172246784,0.0836452246,0.0029904635,-0.1088610515,-0.2052446008,0.0304429177,-0.3144642413,-0.2817363441,0.125975132,-0.1332973391,-0.0618762933,0.0989273265,-0.2075009495,-0.2802496552,-0.3656508625,-0.0915759131,-0.1999159604,-0.2207590342,0.2158221453,-0.2644087672,-0.3948630691,-0.3731863797,-0.4336093664,0.0073938593,-0.1918857098,0.3253252804,0.3003253937,0.015803067,0.3178790808,-0.3654292524,0.0570112951,0.2947160602,0.4680432975,-0.2738643885,-0.3697005212,0.379750818,-0.0870073736,-0.3737559021,0.0451309457,0.106417805,0.1011918783,0.0062330649,-0.3513376117,0.2975459695,-0.2787443697,0.118222557,0.1003810763,0.0857115015,0.3026743233,0.1589624733,-0.242217645,-0.0651043579,0.018234862,0.2500403821,0.3025486469,0.3333814442,-0.215423733,-0.1298262626,0.0965181664,0.0774964914,-0.1745658219,-0.2244542837,0.4646266401,-0.056080807,0.1519815624,-0.1868851781,0.0825462192,0.1256967634,-0.2974067032,-0.0397857651,0.7535393238,0.0184613653,-0.4435170889,-0.0814064294,-0.307174176,-0.5265576243,-0.4649650156,-0.1181864217,-0.0913815349,-0.0362730511,0.0899660364,0.0348156728,0.1896204948,-0.4391711652,-0.0119334599,0.1453507841,-0.0218619537,0.06755355,-0.2452216297,0.2796761394,-0.3746808171,0.286726892,-0.0188874938,-0.0390884355,0.0734120682,-0.0693436861,0.1641741097,-0.27022174,-0.0470541157,0.1129213497,0.1211184636,0.1840787381,-0.2732980549,0.0120059066,-0.0459975339,-0.2533962429,0.044019226,0.2973749638,0.2621649802,-0.4981845915,0.0783740804,-0.3845174313,0.1832409501,-0.0779349506,0.251042515,-0.0613877065,-0.1280373037,-0.4437142313,0.0530308336,0.1399629861,0.4243906736,0.0537542626,-0.0306482334,0.0338712744,0.0521035716,0.3206896186,0.2767496109,0.4186456501,-0.334975481,-0.0093225567,-0.1058883443,0.3103251755,0.136946857,0.3233544827,0.0445084758,0.0566172749,-0.1323209405,-0.2381839305,-0.1703961194,0.4261049032,0.0424848124,0.314878881,0.0364562385,-0.0454361625,0.0617155991,0.1312963367,-0.2217708975,-0.0027023989,-0.0312761888,-0.1476117969,0.1463031173,0.1031937897,-0.0746625438,-0.003503331,0.2231930047,-0.4792090058,0.5465048552,0.0245953836,1.2696266174,0.3159296811,0.3267770708,0.0482161492,0.1233346909,0.2043088675,-0.5028336048,0.0450372435,-0.2049015313,0.2078601122,-0.0692745596,0.3685301542,-0.1507875919,0.2397280633,-0.3311704993,0.2369765639,-0.2088413835,0.076066412,-0.0275174975,0.0753913671,0.0834389478,-0.3084034324,0.037002489,0.2059900314,-0.1129739434,0.0470509231,0.0031081992,-0.1953648627,-0.2502888143,-0.116970703,-0.2816466391,0.2171150893,-0.1992187202,-0.0228738319,0.2414246798,-0.0237631164,-0.171705097,0.1255424768,0.2219231874,0.3128904402,-0.2807663381,0.1846368462,0.1854190975,-0.042959135,0.236381337,0.0808084607,0.4306316674,-0.1359006315,0.0029212816,-0.1516211927,-0.0207546204,-0.2704412341,-0.0228738952,-0.0337700546,-0.0581843592,-0.2406965345,-0.0833918229,0.2814622819,-0.4552398622,-0.2789027393,0.1790005416,-0.2404043823,-0.2246223986,-0.1254339367,0.0349248163,-0.3544524908,-0.2354166955,0.4743593633,0.0659203529,-0.0302721709,0.2621461451,-0.0482762754,-0.2409835011,-0.3099688292,-0.1786750555,0.300883472,-0.3326336443,0.1876684874,-0.0916172341,-0.3863521516,0.3450787961,0.7153212428,0.0183653142,0.3967464566,-0.1848096699,-0.326711446,-0.0301632602,0.0570698828,-0.0381056368,0.1439881921,0.2901443839,0.3565283716,-0.1229695529,-0.1283666193,-0.4049934745,0.2145232409,-0.336237222,0.2910217047,-0.2069050521,-0.152196452,-0.0805535465,-0.022007918,0.0823573992,0.4147498906,-0.0205565579,-0.2648275495,-0.3594222665,0.0865644887,0.2694784701,-0.0352666043,-0.2333404273,0.0358312614,0.0373579822,-0.2132279724,0.095404312,0.017413374,0.0847225264,0.3056302667,0.1444253623,0.1671497077,-0.3926972151,-0.0462918878,-0.038390398,0.0012807852,-0.3028196096,0.1512007415,-0.2068940699,-0.0710864961,-0.0749292448,0.1925043762,0.2182451934,0.4009816349,0.2719069123,0.2186349928,0.0163994245,0.089021191,0.2961805761,0.2918317318,-0.0262057297,-0.4860494137,0.3494085073,0.2846329808,-0.1590370238,-0.2643704116,-0.0835703835,-0.4274363816,0.1194313392,0.1845214516,0.4698522389,-0.2557896078,0.0934513435,0.0261177719,0.4078741372,-0.1422617733,0.120197244,0.3639909327,-0.0351201631,0.0599592179,0.2685972154,0.0387329422,-0.0535021164,0.4079728127,0.0812922046,0.4704857469,0.5988804102,0.2576120496,0.0574812889,0.2749409974,0.3712730408,0.0921033323,0.2998262346,0.3460760117,-0.0133701079,0.1780267954,-0.2248471826,-0.2335674763,-0.0808882564,0.2483653128,-0.3571589589,-0.1191973835,-0.2130063176,-0.244584024,-0.2593747377,-0.0840540528,-0.2889247537,0.061946597,-0.0790872723,-0.0274676774,-0.029448403,-0.2832812369,-0.1334621608,-0.2766895294,0.3692995012,-0.2943421602,0.3083448708,0.0762209445,-0.0538068302,0.1717821211,0.0994597599,0.1822338253,0.215762049,-0.2081106156,0.0303518865,-0.1222161576,-0.1316477358,0.034915667,0.3396193087,0.1013623103,-0.093731761,0.1128105521,0.1167639345,-0.2326501757,0.2023671567,-0.0629938766,0.1703377217,0.0302392766,-0.0107635837,0.0460676923,-0.1949507892,-0.2688479125,-0.172262162,-0.2579215169,-0.0288864244,0.1489243358,-0.1003093868,0.1757889539,-0.2235441357,0.1416842937,0.216223985,0.5997519493,0.264531821,-0.100430876,-0.0401843823,-0.1288246512,-0.4038026631,0.2358362824,0.2191649079,-0.3714693487,-0.0989225656,-0.0245527755,0.0847208202,0.127822578,-0.4963508248,0.3194303513,-0.2276887447,-0.1976346821,-0.347738117,0.3086852431,0.1343068331,-0.1725870073,0.0300982017,-0.1894376725,0.0174396019,0.0781253576,0.1550009996,-0.0839313716,-0.6021875143,-0.0343673751,0.1993200034,0.1280792803,0.2976602316,0.0557811782,-0.3549658954,0.0256131105,0.0023079778,-0.1479882449,-0.2636102438,-0.0502678603,0.0962545201,0.219159469,0.1750769466,0.4257462323,-0.1178935766,0.264166981,0.2795372903,0.0778628811,-0.3049061298,-0.0054787672,-0.2572006285,-0.0134134972,-0.0083371578,0.6562623382,0.0288472194,-0.0379882939,-0.2665489614,-0.4903170168,0.4548585117,-0.2702224255,-0.1678291708,-0.0110594165,0.1426568776,0.4260194302,-0.0236694105,-0.5835841298,-0.1032111645,0.1385838836,-0.0550480001,-0.595993042,0.1988397986,-0.082914032,-0.159206748,-0.1237686202,0.3169107735,0.4044702947,-0.0247503165,0.0226831213,-0.2333716601]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/486","title":"Bookcorpus data contains pretokenized text","comments":"To confirm, since this is preprocessed, this was not the exact version of the Book Corpus used to actually train the models described here (particularly Distilbert)? https:\/\/huggingface.co\/datasets\/bookcorpus\r\n\r\nOr does this preprocessing exactly match that of the papers?","body":"It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575","comment_length":37,"text":"Bookcorpus data contains pretokenized text\nIt seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575\nTo confirm, since this is preprocessed, this was not the exact version of the Book Corpus used to actually train the models described here (particularly Distilbert)? https:\/\/huggingface.co\/datasets\/bookcorpus\r\n\r\nOr does this preprocessing exactly match that of the papers?","embeddings":[0.0296658985,0.1984409541,-0.0193816107,0.3028745055,-0.080079928,-0.2216101289,0.1067400351,0.1523843706,-0.3669324219,-0.007889973,-0.2142251432,0.3167914748,0.2326788157,-0.1111376584,0.0192023013,0.0697399378,0.2531403005,0.2138079405,0.0349262245,-0.2171892971,-0.1444738209,0.2366247177,0.0563932396,0.1152127907,-0.0726907924,-0.3140857518,-0.2279933542,0.3609324396,-0.2332173884,-0.3183438182,0.1385847479,0.3533370197,0.0491578691,0.4791388214,-0.0001063632,-0.0733194351,-0.1108100265,-0.0216152892,-0.1269199103,0.1278881729,0.2702474296,0.0159017872,0.3585571349,-0.4776166677,-0.3692617118,0.1781267524,-0.2240692377,0.1918198764,0.2211266309,0.3104521036,0.2462922782,0.0918430984,0.022231115,0.2674802542,0.1596523374,-0.133894369,-0.2298511714,0.0877113342,0.5446926951,0.2163472325,-0.1553095877,0.3055977225,-0.1081151217,-0.044440303,0.0253587738,-0.0049127494,0.18997702,-0.3225924373,0.1490077525,0.2697640657,0.277718395,-0.1194734648,-0.1294752061,-0.0601966754,-0.0922961086,0.1277790517,0.2424373329,0.0890644714,0.0441507921,0.2079998851,-0.1743232161,-0.1435809284,-0.0047822334,0.0986920595,-0.1490667015,0.3915131688,-0.0701674148,-0.0638668165,-0.0509070642,-0.0633310005,-0.3180634975,-0.0380534641,-0.2157049626,-0.0351227075,-0.0073074475,-0.219176352,-0.0236476567,0.4003375173,-0.1009498611,0.2117394358,0.1576360166,0.1187427044,-0.0770310834,-0.1793920994,0.1342777908,0.2484534085,0.1051763892,0.2507029176,0.0362859033,0.3239161968,-0.2562924623,0.0412143655,-0.1305710822,-0.0480502546,-0.1039604619,-0.086110495,-0.0972716585,-0.491089493,0.0703444704,0.2772708535,0.0333309099,0.3170373738,0.0583801717,0.0865695998,-0.2092938572,0.0372137092,0.2322643846,0.0183618534,-0.1913560778,-0.2068758011,-0.2237014323,0.0077761565,-0.3948843181,0.0062467833,0.1177478805,-0.0474415161,0.2660805285,-0.0383913964,-0.2094124854,-0.1035205051,-0.0290001296,-0.1996834725,0.2668502927,0.0191406272,-0.1366973072,0.2283186018,0.0493180603,-0.1478088349,-0.1507509798,0.0246911049,-0.3962158263,0.0323202685,-0.1493204981,0.232612744,-0.2601950765,-0.2491978258,-0.0344613791,0.281306982,0.4073366225,-0.1509707421,0.3375005126,-0.0243922994,-0.1385217905,0.0281361956,0.0209932309,0.4135779142,-0.22823973,-0.1458941996,-0.0486666076,0.0728444606,0.3752904534,0.0883204937,-0.2509766221,0.061753951,0.2696670294,0.6624370813,0.1866119504,-0.0779509097,-0.4325422943,0.2136226296,-0.0138359247,0.2117446512,-0.1904442906,-0.1700180471,0.180130586,-0.0564443059,-0.0000731084,0.0359709337,0.4290822744,0.1707169861,-0.6041297913,-0.3096036911,0.3253692985,0.2513689399,-0.1255486906,-0.0189415403,0.0184531193,0.0625682697,0.1516380012,-0.1676374972,0.1570030451,0.4770158529,0.180464223,0.3032194078,0.2585037351,-0.0054152706,0.2384948283,-0.0026990685,-0.4634099603,0.6628252268,0.0335819907,-0.2584096789,-0.0643427968,-0.2789319754,-0.1723700017,-0.2421185374,0.27399382,0.0556422584,0.0018351172,0.28350842,0.0382650867,-0.1210398674,-0.2297665924,0.1404663622,-0.5065470338,-0.0698980018,-0.2286457568,-0.0394805558,-0.2563893199,0.3594435155,0.178655073,0.0626103655,-0.1487026066,0.457759589,-0.3280336559,0.0037435619,-0.0326191001,0.2570511699,0.1916282028,-0.0837774202,0.0840337351,0.463596046,0.2697112262,-0.0392536707,0.2076291144,0.3762252033,0.4181046188,0.0358300246,0.0566478185,-0.1161638349,0.0447982661,-0.2119199336,0.0115555702,-0.1377198994,0.2113891095,0.1241783649,0.1486055702,0.2420843989,-0.2317489237,0.3815658689,0.259077698,-0.1092300713,0.0798091665,0.1837252975,-0.3506711423,-0.3496268988,-0.0154661844,0.1806977242,0.1021821201,0.0754515529,0.007575315,0.0766030326,-0.0453089289,-0.3259964287,0.285486877,0.151080206,-0.2582577467,0.0437834263,0.234524712,0.2017498165,-0.4386012852,-0.1384997815,-0.0466303416,0.1867870837,-0.1489841193,0.3591692746,-0.2862590551,-0.4356797636,-0.1512416303,0.048651915,-0.0093042972,-0.1176725402,0.2985786498,0.1213746741,-0.0993097052,0.0566712692,0.149135232,0.5113168359,-0.2363129556,-0.0650571138,-0.1264048517,-0.2729800642,-0.7907334566,0.1810279191,0.1005466953,-0.0117973955,-0.1401486248,-0.3721141219,-0.1473334134,-0.1202430502,-0.4255813956,0.0287551545,-0.1445327997,0.0502932929,0.1586819589,-0.1826369017,-0.2583869994,-0.1930638701,-0.0241111033,-0.111395292,-0.1810622811,0.1034460813,-0.2522014081,-0.4218706787,-0.3365124762,-0.5817195177,-0.0893881619,0.0237446446,0.2863693237,0.232751295,0.007716727,0.4235409796,-0.1916831881,-0.014036607,0.1494739056,0.452454865,-0.2753588855,-0.2926751077,0.3957184851,-0.0382814221,-0.4434627891,-0.1504343897,0.0747855902,0.0841897503,-0.0416624509,-0.3317264318,0.1113496125,-0.3541252315,-0.0893000141,0.0395525433,0.081608437,0.2867455482,0.0311784595,-0.2125947922,-0.1179542243,-0.0140804518,0.5295315385,0.2311033905,0.2775559425,-0.3133275211,-0.060712073,-0.0815422609,0.0009640065,-0.0211240016,-0.2409221232,0.3181604147,0.0015801466,0.0636589825,-0.1606695652,-0.032405939,0.0874614045,-0.2342914194,-0.1427820772,0.6714075804,-0.0063176523,-0.3302324414,-0.126548782,-0.2998776436,-0.4062518179,-0.3436610997,-0.1450318396,-0.081875287,0.0542494431,0.1594569236,0.0565785877,0.1285170317,-0.3857202232,0.0279503725,0.2911815047,-0.2177026272,0.0844202042,-0.0557054691,0.2240014076,-0.5900645852,0.1102208197,-0.1178842559,-0.0061592618,-0.00744646,-0.0192108881,0.2219082862,-0.2553706765,0.1108383536,-0.091390647,0.0196139812,0.1842551678,-0.4195017517,0.0429296903,-0.1706501544,-0.3483814001,-0.016740879,0.2624340057,0.219324246,-0.396091342,0.0259023942,-0.2092825472,-0.0174241643,-0.1271578223,0.3416578472,-0.0376446396,-0.3000449836,-0.2775603533,0.1199879348,0.0643891543,0.3676751256,-0.1024145186,-0.0005579214,0.0976182818,0.4271933436,0.2771632969,0.3257227838,0.316280961,-0.3522459567,0.033426974,-0.1422742605,0.2826834619,0.211571306,0.3943057358,-0.0231346861,0.0652965233,-0.0599542707,-0.2926258147,-0.3611947596,0.3583038747,-0.0038997887,0.3576032519,0.2419638187,-0.0357152075,0.0211483762,0.1509304792,0.0977677181,0.0407654867,-0.192247957,-0.2394361049,0.2312880307,0.0902263001,-0.1045805663,-0.0171595365,0.2092266381,-0.5008425117,0.5596808791,0.1802475154,1.1709780693,0.1551647186,0.4108209908,0.2056064308,0.1132090315,0.1977671385,-0.5068720579,-0.0735532567,-0.3077666163,0.1728138924,-0.0555335842,0.4235354662,-0.0200595222,0.2493421584,-0.2679891884,0.1992612332,-0.2963901162,0.2366586626,0.0645620972,-0.1335395426,0.2405453324,-0.0945096165,0.0363566279,0.2081566006,0.0397811905,0.1289013475,0.0057344651,-0.2461297661,-0.2890439928,-0.0531961136,-0.2310490757,0.2454655766,-0.0331804007,0.0494030677,0.2079698741,-0.1106896847,0.1575109661,0.138241291,0.471807152,0.1743507832,-0.3955404758,0.2477910668,0.1785717756,-0.1037529111,0.2984573543,0.0512138121,0.5131377578,-0.0922016278,-0.0203276556,-0.2224055529,-0.1007061079,-0.1466458738,-0.0992624313,-0.0249892287,0.1355513632,-0.4075537622,-0.00808019,0.3327296376,-0.3182326853,-0.3199116588,0.1742996126,-0.219223097,-0.1999281347,0.0848744884,0.1321092993,-0.2795988321,-0.1058832183,0.3580372334,0.2552371919,-0.0036921506,0.421399653,-0.0478948094,-0.2383197844,-0.3894928992,-0.3604232967,0.5178197026,-0.4499707818,0.2662649751,-0.0771513954,-0.3334920108,0.3360660374,0.7329288721,-0.0140918642,0.1992586106,-0.3125976622,-0.2560109198,-0.0607898757,-0.0324201435,-0.0447006598,0.1747518629,0.2742785215,0.3117230237,-0.1441117525,-0.1182435304,-0.3982656002,0.2483280152,-0.3359957933,0.1890867352,-0.1631184071,-0.098218523,-0.1337666959,0.0796321109,0.0410761908,0.258713305,-0.0245961845,-0.2936259508,-0.3275233507,0.0688280314,0.2236124277,-0.214241311,-0.0723753721,-0.0335410126,-0.1770566404,-0.1123383865,-0.0055618137,-0.0022374347,0.2331557572,0.2266034633,0.0625120252,0.1165823862,-0.4161798656,-0.0662279353,-0.1230773404,-0.077465333,-0.1583362073,0.0727250129,-0.2133881152,-0.1209276617,-0.1550257951,0.3376784027,0.164060846,0.4420236349,0.3606298864,0.0693640783,0.1109359786,0.115117386,0.2891586423,0.3181856275,-0.1142821014,-0.4719843268,0.3100068569,0.280436933,-0.1817850918,-0.3593287468,0.0826414153,-0.5128973126,0.0000176283,0.1591277868,0.4305146635,-0.1177245378,0.1454430223,0.0505663343,0.3363843262,0.0239165481,0.2004487514,0.4209345281,0.064042367,0.101402767,0.2356197089,-0.0335277058,-0.0871535763,0.4698125422,0.1750523001,0.5409213305,0.4480159879,0.2250310034,-0.0101082418,0.1249662861,0.4259962142,0.0221624263,0.1973182261,0.2906276584,-0.0926223621,0.2536987066,-0.3461321294,-0.316028595,-0.3414964378,0.2313448936,-0.4401254654,-0.2124316245,-0.2084479779,-0.2706871331,-0.0848739669,-0.011378292,-0.2570594549,0.2090296596,0.0766107291,-0.2347417474,-0.1947250664,-0.4968023598,-0.1306171119,-0.133172676,0.4301268756,-0.1012997851,0.4015232325,0.1030493975,-0.1810583621,0.0290284511,0.3175077736,0.087291874,0.1963389814,0.0133625669,-0.1889032423,0.0431263261,-0.1191530451,0.0723831132,0.3179276884,0.2134418786,-0.0753749683,0.1678413451,0.0687644929,-0.1467098892,0.2038977146,-0.0412294269,0.24318178,0.0998826548,0.0278642699,0.1857538819,-0.0540202707,-0.3054508865,-0.1482003331,-0.2258319855,0.0509598814,0.2498961836,-0.008111503,0.1863151193,-0.2863770425,0.1335958689,0.1357712001,0.6176176071,0.286067456,-0.142704621,-0.1430062205,-0.1510526389,-0.3049138188,0.3031997383,0.1856849641,-0.3401728272,-0.0325693488,-0.0787375495,0.0555833392,-0.0671912134,-0.3388928771,0.1711835265,-0.0554585941,-0.1162089184,-0.2399934381,0.3016031384,-0.1026963294,-0.0256432872,-0.0212991387,-0.083833836,0.0958451554,0.0370113738,0.1094705611,-0.0381377824,-0.4150056541,0.0061343596,-0.0140303131,0.1681959331,0.2728514075,-0.0538946763,-0.2741570771,0.0648501441,-0.0640594959,-0.1224150211,-0.2482182235,-0.1093807444,0.0175887495,0.1900931597,0.1669217646,0.2837260067,-0.0647188127,0.0451710448,0.3557327092,0.0490445495,-0.3398377597,-0.0906767994,-0.0053016781,-0.1250113398,0.2342025638,0.655759871,0.0375273488,-0.0007673817,-0.2772576511,-0.5082750916,0.5603707433,-0.3009026051,-0.1876616925,-0.0994003415,0.1073845401,0.4117279053,-0.0702464432,-0.7056644559,-0.1818267405,0.1257766634,-0.1075636819,-0.5928035378,0.3088395596,-0.0592664368,-0.0610288419,-0.1592838317,0.2810201943,0.1327218711,0.0317786559,-0.0160774272,-0.2202399969]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/486","title":"Bookcorpus data contains pretokenized text","comments":"I believe these are just artifacts of this particular source. It might be better to crawl it again, or use another preprocessed source, as found here: https:\/\/github.com\/soskek\/bookcorpus ","body":"It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575","comment_length":27,"text":"Bookcorpus data contains pretokenized text\nIt seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575\nI believe these are just artifacts of this particular source. It might be better to crawl it again, or use another preprocessed source, as found here: https:\/\/github.com\/soskek\/bookcorpus ","embeddings":[0.0448162481,0.1466449648,-0.0626623482,0.3189455867,-0.0481281355,-0.2803273499,-0.002858825,0.2543207407,-0.4264054596,0.0911981389,-0.1354174912,0.3492498696,0.2933695614,-0.1151346415,-0.0640322417,0.2069191933,0.2553399801,0.244030863,0.0813037232,-0.1928595752,-0.1757329255,0.3218994737,0.1353747696,0.1295599639,-0.1914875954,-0.231698364,-0.1364848465,0.296626091,-0.1697288454,-0.3685367405,0.0418670997,0.2373154461,-0.0457461327,0.4200381339,-0.0001042967,-0.1999749392,-0.0085521257,-0.0008110324,-0.2220437229,0.0396469124,0.2418955714,-0.1047132462,0.3316567838,-0.4221564233,-0.2050881237,0.0990941674,-0.1921876073,0.1155716628,0.1336494237,0.2984400392,0.268550992,0.2499953061,-0.1447290331,0.240590021,0.2765300572,-0.1052289456,-0.2680330276,0.1541532129,0.4808159471,0.2871724963,-0.1176711023,0.3252732456,-0.1483825892,-0.0550989546,0.0139530217,-0.0682798102,0.1332594901,-0.2030378431,0.2449139655,0.174347505,0.1978157312,-0.1787820309,-0.0374164693,-0.0773184374,-0.0317805521,-0.0305365752,0.3331260085,0.19728145,0.1977313459,0.0857232511,-0.2289130986,-0.1092083901,0.0344108492,0.1787556708,-0.069673188,0.3172315955,-0.126973331,-0.0139591452,-0.0726937652,-0.1028295532,-0.2077139616,0.0291778482,-0.3983161449,0.0002941578,-0.0310087353,-0.1343759447,-0.0526116453,0.3532701433,-0.029496694,0.1897891611,0.158452183,0.2260872871,-0.1797975451,-0.0180140603,0.1203735471,0.1379707754,0.1127826422,0.2411134392,0.0682670251,0.0640382171,-0.2253690809,-0.0351159014,0.039488893,0.014292608,0.0254294369,0.0347913317,-0.0737026259,-0.4883922338,0.1922175139,0.2083523124,-0.0955336019,0.294585824,0.0166280046,0.0861300454,-0.1078063175,0.0290524233,0.2102760077,0.0138786798,-0.1805820912,-0.1609926522,-0.1973941624,-0.0833952129,-0.3167269826,0.0978531688,0.149828583,-0.0653669089,0.215406552,0.0616454743,-0.1452436745,-0.0833908767,0.0132533731,-0.2287236601,0.2126800865,-0.0081226351,-0.0353080928,0.2613686919,0.0131319519,0.00745766,-0.1886086315,0.0436918736,-0.3702692688,0.074915871,-0.2921417356,0.2618963718,-0.268196106,-0.2209359705,-0.0918350518,0.0528980307,0.2958088815,-0.2276176959,0.3187195361,0.0037580112,-0.0961418077,0.0139309429,-0.2106834948,0.2404086441,-0.1392854154,-0.1497936845,-0.0291267857,0.012516669,0.2761663198,0.1072723567,-0.2123286128,-0.0086455001,0.1043876335,0.7325607538,0.0433354415,-0.077836439,-0.478520751,0.2862702906,-0.0392540433,0.0749829859,-0.1802366674,-0.1213595271,0.1312688142,0.0449119657,-0.1406109333,0.1511675417,0.6003320217,0.2346280664,-0.604428947,-0.2423948646,0.2666245103,0.1318976283,-0.1301935911,-0.0382740833,0.1433165371,-0.0083301011,0.094967261,-0.2224237025,0.2434530705,0.4617540836,0.1976676732,0.3330716193,0.2323134989,-0.0787156075,0.1970963776,-0.005560061,-0.6133110523,0.608982563,-0.0590002649,-0.29228881,-0.0787849426,-0.3615652323,-0.1525607407,-0.0115266228,0.2970635891,0.1812445521,-0.117087014,0.3826634884,0.1691549569,-0.0269447919,-0.0533163473,0.0236411877,-0.3362281024,0.043103192,-0.1681781411,-0.1970750093,-0.1390396804,0.2937299907,0.2984584868,0.0936225951,-0.2625220418,0.5021242499,-0.4160388708,0.2466996014,-0.1689228415,0.1719899476,0.232218042,-0.00927235,0.1116819829,0.5388678908,0.3786891997,-0.0529740602,0.2159281671,0.3990840912,0.4482703507,-0.0155954091,0.0430603065,-0.1934474558,0.1470884085,-0.2783168256,0.1191444099,-0.1288463026,0.2308624536,0.090961732,0.0945118144,0.202873975,-0.1953773946,0.3303447962,0.3287514746,-0.0927829742,0.0969554707,0.2161509395,-0.283090651,-0.2576099634,0.0114530865,0.1333328485,0.0847555548,0.0552718602,-0.0330925658,0.0875942111,0.0944515318,-0.328237772,0.2330359966,0.1454982609,-0.2436210513,0.0642767921,0.2809624672,0.200641349,-0.4732762277,-0.0961568207,-0.0593900643,0.1325763613,-0.1590367705,0.3321847022,-0.320988059,-0.6278653741,-0.0989341289,0.0389646478,-0.1498236507,-0.2108939886,0.2689701319,0.1829663962,-0.1170474216,-0.0242071711,0.1855788827,0.3524076045,-0.2660956383,-0.0416394435,-0.1509880871,-0.3491880298,-0.7120478749,0.1918044239,0.0732700676,0.0796468109,-0.0410744399,-0.2622717917,0.0435617678,-0.3028933108,-0.3058649004,0.0577230304,-0.1697437465,-0.0810702741,0.0272014178,-0.1541400999,-0.3781108856,-0.2736579776,-0.0023806791,-0.1636536419,-0.1409661919,0.118033357,-0.2873587608,-0.4336165488,-0.4267374873,-0.4536304772,-0.0530870706,-0.1416699886,0.3145368099,0.2562345862,0.0632836595,0.3353785574,-0.3295733035,0.0930118039,0.2780516744,0.4002052546,-0.3519272208,-0.4099279046,0.358445555,-0.1365947574,-0.4528926313,-0.0235643201,0.1484946758,0.0834878907,-0.0692305788,-0.4910207987,0.2069643885,-0.2885848284,0.0113306036,0.1198522747,0.0267633591,0.2422291338,0.1011978611,-0.2581208944,-0.0639290363,-0.0957722068,0.2960177362,0.2582881749,0.2977918386,-0.2564225197,-0.2517824769,-0.0318347737,-0.0260694809,-0.1021022573,-0.1441984624,0.4561549127,-0.044237338,0.1877220869,-0.1615409851,-0.0391880684,0.0314300433,-0.3032161891,-0.0150900073,0.7271082997,0.0212649219,-0.4289526343,-0.0332350843,-0.3933043182,-0.5735100508,-0.4311955571,-0.1001486406,-0.0432981625,0.0939050391,0.1236358956,-0.0178424045,0.1207751781,-0.4445358813,0.0327545628,0.1532097906,-0.1836449504,0.0478671417,-0.3420101404,0.1687647551,-0.3717239797,0.2553083897,0.0052625178,-0.0013446155,-0.0451658145,0.0586261153,0.2602805197,-0.2168097198,0.1064090803,0.0102944728,0.0653243586,0.2743197083,-0.3003895879,-0.0027092644,-0.1228680387,-0.2420933694,0.1531206071,0.3913672566,0.1850697398,-0.3122660518,0.0835666955,-0.3329971731,0.12830019,-0.1680285931,0.3180052638,-0.0592937469,-0.2124680877,-0.3637363315,0.0656473115,0.0530694015,0.427385658,-0.1115669459,-0.045570489,0.0330482088,0.2324436754,0.2790347338,0.228036508,0.4632433057,-0.271353215,0.0324243791,-0.1344939321,0.1827018857,0.1138167307,0.4114049971,0.0695169196,0.0734396651,-0.0852294937,-0.4019813836,-0.2859327793,0.3970335126,0.1608572453,0.2917511761,0.2020860612,-0.0327832513,0.0800197572,0.132382825,-0.0058360333,-0.0144794993,-0.0985424221,-0.0809307098,0.1879972965,0.0858461633,-0.0652606413,-0.013061638,0.1561303884,-0.3943966329,0.4994822741,0.0091067534,1.2245506048,0.1914113611,0.247314617,0.1475096196,0.0828830525,0.1094572023,-0.5010140538,-0.0462772883,-0.3080055118,0.1485587209,-0.0597315393,0.3675073683,-0.0961659104,0.1177002341,-0.3358370364,0.1413776129,-0.2380871028,0.0940168127,0.0226754937,-0.0210822131,0.1031466052,-0.1923015416,0.108588025,0.2364027053,-0.0654326007,0.1071671918,0.0198925901,-0.2630675733,-0.2315975875,-0.1430666298,-0.290202558,0.2394070923,0.0503569469,0.0529960096,0.2848292589,-0.0000994055,-0.1121583879,0.1071237773,0.306851685,0.249031648,-0.3617753386,0.2733414471,0.2050228268,-0.106580548,0.2486872524,0.1970612407,0.3249379992,-0.139112249,-0.1275558025,-0.1675458997,-0.0458946899,-0.1557647884,-0.0518639758,0.0214324854,0.0463715047,-0.333611697,-0.0767352059,0.2401323467,-0.3926161826,-0.2628386021,0.1982800364,-0.1990271509,-0.1592067778,-0.0691775158,0.1285282224,-0.3441812992,-0.1976015419,0.4220589995,0.2594234645,-0.1140966937,0.2870840728,-0.0380649716,-0.2450443655,-0.3479488492,-0.1590729803,0.2692504823,-0.4661871195,0.1822966784,-0.0721394271,-0.3561222553,0.2949458957,0.6725558043,-0.0804816037,0.3593213856,-0.237511903,-0.3537041247,0.0164922327,-0.0583337918,-0.0230597761,0.1235371307,0.3605514765,0.228778407,-0.0449989662,-0.123118639,-0.4092462957,0.3093141615,-0.4030047953,0.2577026784,-0.1985124648,-0.0855369419,-0.1081541181,-0.0206874311,0.0510033369,0.3924660981,-0.054748036,-0.3172295988,-0.3506034315,0.0725567639,0.2348154187,-0.1416841894,-0.1793186069,0.0117172441,0.0772335604,-0.1079203039,0.0482684225,0.0659533888,0.0420017242,0.3306063414,0.0579712018,0.1590735614,-0.3384088576,-0.0257624332,-0.0776409954,0.0079833185,-0.146288231,0.0645454451,-0.211655587,-0.0769070014,-0.1116123572,0.1302298307,0.1451040059,0.4586117864,0.2783311903,0.1129997373,0.1690621227,0.1006908342,0.1460395902,0.2584141195,-0.2015935779,-0.4826548696,0.2902131975,0.2997664809,-0.0947847217,-0.3748229742,-0.0598176941,-0.4852978587,-0.0221097767,0.0629761964,0.4774008691,-0.2948589921,0.2269394249,0.0935661122,0.3392340243,-0.0036567936,0.2050056905,0.3165465593,-0.0829434544,0.0184575282,0.189301312,-0.0224325247,-0.1535359174,0.4744415283,0.1242653728,0.4253406227,0.5266216993,0.2677662969,-0.0496025681,0.2188428789,0.3550836742,0.0118997404,0.1806090474,0.3160149455,0.0988196656,0.2178559303,-0.270113945,-0.3222949803,-0.1784775108,0.179180786,-0.3917804062,-0.04410211,-0.1466632932,-0.1504088491,-0.1310459673,-0.0305565707,-0.2464739978,0.1594862938,0.0352632701,-0.0778793618,-0.1439423412,-0.3957456946,-0.1519313157,-0.195406422,0.3066391945,-0.1740107238,0.2878361344,0.0850268677,-0.0048501892,0.0827036574,0.2119253874,0.1977032125,0.2964372933,-0.1173013374,-0.0388398916,-0.037428122,-0.0987749547,0.1580717862,0.2782498598,0.1883050501,-0.0295790862,0.1161465347,0.1148043945,-0.2072363049,0.2041778117,-0.1279338151,0.241641596,0.1335805804,0.1190790236,0.0195859093,-0.1260525584,-0.3083533943,-0.2294475883,-0.2746400535,0.067080006,0.2077347487,-0.0519191138,0.1856831759,-0.287563026,0.1513199508,0.1411016732,0.6888765693,0.2781017125,-0.0805138126,-0.0842655003,-0.1521545947,-0.3064154685,0.3324434161,0.1166924462,-0.2755005658,-0.0252467301,-0.0877102464,-0.0442933068,0.0128355715,-0.3822121322,0.2271617949,-0.1277780831,-0.2840722799,-0.3597864211,0.3236964047,0.1303897798,-0.11333327,0.0209485926,-0.1534525901,0.1544737667,-0.0766033754,0.1796651036,-0.1036647856,-0.5949347615,-0.1030548811,0.1579025239,0.1936388016,0.2918349802,0.0025181267,-0.2615239322,0.0434960723,0.0204941295,-0.2243358195,-0.0426711366,0.0278207343,0.1692958474,0.1975400448,0.1363653392,0.3651080728,-0.0719057322,0.2666305304,0.2732154131,0.1929305643,-0.1848855913,-0.0891332552,-0.1658139974,-0.0525440909,-0.0295560267,0.5959393382,0.0899257511,-0.0072155339,-0.2557160556,-0.4805971682,0.5174429417,-0.257045567,-0.1837040484,-0.029739188,0.1526863575,0.477586478,-0.0879563615,-0.6576775312,-0.1006937325,0.1419468671,-0.0659026578,-0.5773202181,0.2950589955,-0.1587462723,-0.0939943716,-0.0757436752,0.2928362191,0.3342243135,-0.0027443802,0.1261534542,-0.1822180003]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/486","title":"Bookcorpus data contains pretokenized text","comments":"Yes actually the BookCorpus on hugginface is based on [this](https:\/\/github.com\/soskek\/bookcorpus\/issues\/24#issuecomment-643933352). And I kind of regret naming it as \"BookCorpus\" instead of something like \"BookCorpusLike\".\r\n\r\nBut there is a good news ! @shawwn has replicated BookCorpus in his way, and also provided a link to download the plain text files. see [here](https:\/\/github.com\/soskek\/bookcorpus\/issues\/27). There is chance we can have a \"OpenBookCorpus\" !","body":"It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575","comment_length":60,"text":"Bookcorpus data contains pretokenized text\nIt seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, \"didn't\" becomes \"did\" + \"n't\", and double quotes are changed to `` and '' for start and end quotes, respectively.\r\n\r\nOn my own projects, I just run the data through NLTK's TreebankWordDetokenizer to reverse the tokenization (as best as possible). I think it would be beneficial to apply this transformation directly on your remote cached copy of the dataset. If you choose to do so, I would also suggest to use my fork of NLTK that fixes several bugs in their detokenizer (I've opened a pull-request, but they've yet to respond): https:\/\/github.com\/nltk\/nltk\/pull\/2575\nYes actually the BookCorpus on hugginface is based on [this](https:\/\/github.com\/soskek\/bookcorpus\/issues\/24#issuecomment-643933352). And I kind of regret naming it as \"BookCorpus\" instead of something like \"BookCorpusLike\".\r\n\r\nBut there is a good news ! @shawwn has replicated BookCorpus in his way, and also provided a link to download the plain text files. see [here](https:\/\/github.com\/soskek\/bookcorpus\/issues\/27). There is chance we can have a \"OpenBookCorpus\" !","embeddings":[0.093844004,-0.1207267866,0.0057472717,0.3477259874,-0.0450499319,-0.2574845254,0.0231197923,0.1923967004,-0.3466641903,0.1204395294,-0.2318461388,0.0498438291,0.2508267164,0.2811335027,0.0633646697,0.1325377971,0.287984848,0.1430951655,-0.0784864053,-0.1598722935,-0.2173705101,0.4336135089,0.116646789,0.1677406728,-0.0388899483,-0.1268423349,-0.2003344297,0.4001177847,-0.207802996,-0.2791159749,0.152703926,0.2251888514,-0.1080720127,0.4011824429,-0.0001066821,-0.1593561471,-0.1229970977,-0.0155404378,-0.3229634464,0.0267891679,0.3888083696,-0.1011466682,0.1833069026,-0.3243453205,-0.3408608139,0.1802245229,-0.1350855082,0.0620260052,0.1249145865,0.103968285,0.2136241943,0.3086166382,0.0371994488,0.2225579172,0.0696240216,-0.0375635587,-0.352791667,0.2101964653,0.491291523,0.2205249071,-0.0633485019,0.2900327742,-0.1276962608,-0.1118726209,0.0393990688,-0.0221189708,0.2333828956,-0.115067035,0.2039379179,0.3186214566,-0.0194089804,-0.2599667609,-0.1064576581,-0.1943802088,-0.0761001483,-0.0726003647,0.3364339769,0.2846116722,0.2253692299,0.096883148,-0.3741988838,-0.2277469039,0.0811936334,0.1313805431,-0.0405469351,0.0361633077,-0.1846025884,-0.0237984136,0.0744446889,0.0134654371,-0.4897994399,0.0269408822,-0.3042084873,0.0350655206,-0.0974979848,-0.149834767,-0.0538868718,0.7010756135,0.0746983662,0.0693677664,0.0530724972,0.1078742296,-0.3359383345,-0.0265541635,0.0820744038,0.1007655188,0.1369506568,0.0156086097,0.0678852722,0.1843396723,-0.0453089848,0.0362810791,0.0438416339,0.1020604074,-0.1639277041,-0.0209339894,-0.0724795759,-0.3996303976,0.126874879,0.1415815949,-0.1334972382,0.364220053,0.0879585668,0.0076306001,-0.2582374215,-0.0210713036,0.296479553,-0.0102810683,-0.2136100382,-0.2217784077,-0.16154778,-0.0958715677,-0.3191138506,0.2264874578,0.146472156,-0.3952271342,0.1858628094,0.0444760211,-0.3081521988,-0.0185635202,-0.0056978543,-0.0902869701,0.1193651035,-0.1410991549,-0.0466800146,0.1980329752,0.0333113484,0.0604279637,-0.2304292917,-0.0702006742,-0.2735208571,0.012078396,-0.3084296882,0.2019418627,-0.3340657055,-0.1973949969,-0.2077858448,0.0061911861,0.1389762759,-0.2477938682,0.3023016453,0.1350104809,-0.1767831594,0.0876918957,-0.140010938,0.2940986454,-0.0604807362,-0.2447988987,0.2254447043,0.031728033,0.1518551409,0.1032098159,-0.1967705488,-0.0480698682,0.0697814971,0.6613706946,-0.1448672712,-0.1817839146,-0.4148927331,0.1682779044,-0.112037316,0.1598674804,-0.1114613265,-0.1683021486,-0.0464653559,0.0063901581,-0.2677539289,0.1824931204,0.5105293393,0.2303016931,-0.5244213343,-0.3914472461,0.2984268665,0.1271823496,-0.2067021132,-0.0531212427,-0.0020292336,0.0140398871,0.1554815918,-0.2145636827,0.1634925157,0.3532806039,0.2083569169,0.3966579437,0.2448130995,-0.0061980812,0.1457246989,-0.0089488942,-0.6330733299,0.5566461682,0.1423982382,-0.373445034,-0.1040412486,-0.347904712,-0.091561757,-0.1056637168,0.2224725783,0.1875253022,0.0559776649,0.2985978723,0.1524914354,0.2028595805,0.0043207542,0.1445838511,-0.6349841952,-0.034098044,-0.1526804864,-0.0318962373,-0.1222349033,0.3136814237,0.1959505826,0.0005569774,-0.0889044479,0.5558328629,-0.5106656551,0.2036436498,-0.191286698,0.2681001127,0.2760772407,0.1986155957,0.0812509358,0.4749922156,0.3740749657,0.0067572808,0.1219974533,0.3797285259,0.4839081764,-0.0598702766,0.1076591164,-0.15554142,0.035182979,-0.2817848325,0.0004006945,0.0369275361,0.2631747127,-0.1114642695,0.2040321678,0.1861056387,-0.3016117215,0.257679075,0.4086394906,-0.1320691109,0.0894851312,0.1542828232,-0.3398966193,-0.1776928604,0.0346417055,-0.126543656,0.0189012662,0.0660448223,-0.0793131217,0.1490279734,0.1661371887,-0.23138313,0.3121629059,0.1072962955,-0.2484957576,-0.033841297,0.2163281888,0.1573525667,-0.4162901044,-0.069895938,-0.1470281631,-0.0663297027,-0.1102841347,0.2880488634,-0.2646134496,-0.5740508437,-0.1406423002,-0.001930643,-0.1594645232,-0.1174939051,0.3127172291,0.3862805068,-0.0017168057,-0.1491294056,0.1377091557,0.4958688617,-0.3587916195,-0.0933328494,-0.1212109774,-0.2203113735,-0.6400431991,0.1287747324,0.0795243084,-0.1360808611,0.0171533003,-0.2981890738,0.0103077944,-0.3268692791,-0.3456943035,0.1376442462,-0.2445939332,-0.1988650262,0.0822715163,-0.1902072579,-0.4083915949,-0.1923135668,0.0181793198,-0.1432582587,-0.2758737803,0.1623868495,-0.1835000366,-0.3752998114,-0.3884890378,-0.3572832644,-0.0505762696,-0.1736682206,0.5705422163,0.191896379,-0.0065136272,0.5147127509,-0.3683890104,0.1496383548,0.2725290954,0.3947570026,-0.3479954302,-0.4685353041,0.3225314319,-0.0649405494,-0.4029202163,-0.0109294038,0.0989057571,0.0454924107,-0.092030108,-0.4073903263,0.0923260376,-0.2119834721,-0.015826799,0.1361187845,0.1005617604,0.4459448159,0.1266246885,-0.2117586136,-0.0010535609,-0.2075610161,0.3247995973,0.3419704139,0.3122250736,-0.1932136416,-0.3087399304,0.0377808772,-0.1208563894,-0.0683812499,-0.0469391122,0.4255401492,-0.0183441024,0.2324013263,-0.2110125571,-0.0943398401,-0.0144865038,-0.3801892102,0.0313815996,0.6702115536,0.1985557079,-0.1369229555,-0.0599776879,-0.3689665198,-0.6417499185,-0.4424596727,-0.2246522307,0.1710258871,0.0653162524,0.059396144,0.0653525442,0.1679581106,-0.4523164928,0.2160944641,0.1880343854,-0.1169765145,0.1536767781,-0.1043583006,0.1665079147,-0.3501736522,0.3285301626,-0.1217252761,-0.1349209696,-0.0923758596,-0.082918793,0.2527399659,-0.251593262,0.3255562484,0.1049826071,0.0539811552,0.1783373356,-0.4626269341,-0.0315129682,-0.1001412943,-0.1385671198,0.1359021515,0.4657759964,0.2852057517,-0.4652677476,0.0572000481,-0.2035703063,0.2359097451,-0.1319594979,0.2894384563,-0.076541394,-0.2617914677,-0.3590090573,0.0714079365,-0.002796513,0.4658588767,-0.061463099,0.0881539211,0.0423531383,0.075993903,0.3209174573,0.1585340947,0.3812205791,-0.1816037744,0.1032799557,-0.0077860518,0.2582409382,-0.0194939766,0.4296961129,0.2879306674,0.033430025,-0.08159291,-0.2895697355,-0.3119293451,0.4915589392,0.1787161231,0.1890119016,0.3776551485,-0.0836758837,-0.0504905879,0.1121816263,-0.0673044771,0.0549028181,-0.1972082704,0.0284637846,0.0690328628,0.1082099229,-0.0777719021,0.0953316465,0.3547429442,-0.3693214357,0.3874281347,-0.1326045245,1.2466686964,0.0610324293,0.3273184299,0.2324031442,0.2237144858,0.0517900027,-0.4737737477,0.0959865674,-0.3381618261,0.2183776796,-0.0965205655,0.4169149995,-0.1062445715,0.0406304412,-0.2432555258,0.1963570118,-0.0930291116,0.1904976815,-0.0916992724,-0.0072291284,-0.1202440709,-0.3448228538,0.0097190943,0.1911434233,0.0145948268,0.320369184,0.145957917,-0.3317462504,-0.2183902711,-0.0407759808,-0.1546416879,0.3159371912,-0.0854476616,-0.0817432702,0.4184537828,0.0431342125,-0.1074430943,0.1573020369,0.4088445902,0.2863362432,-0.4062108099,0.065909408,0.0239760038,-0.19142735,0.2301289737,0.0792364776,0.360321641,-0.0542956293,-0.0340519547,-0.1631662399,0.0012091695,-0.1712646633,0.0285216141,0.0273656733,0.1606477946,-0.397877425,-0.025506286,0.2487515509,-0.3316328526,-0.2336686701,0.1452807337,-0.2110293657,-0.1632934809,-0.1514056325,0.1720249653,-0.2713449001,-0.1732176393,0.4485718012,0.1804521531,-0.1976035386,0.2521137297,-0.0242790561,-0.2300775647,-0.2597089112,-0.1339485645,0.3980849981,-0.3519406915,0.0863053054,0.0551347323,-0.3697426617,0.1973038316,0.7515901327,-0.0260745008,0.1744614244,-0.2122738808,-0.2387893945,0.0217944477,0.0409233458,0.0566874221,0.1364224106,0.6430277824,0.2666280866,-0.039521724,-0.21417059,-0.3800364733,0.2778612375,-0.3200713396,0.1124018133,0.0029334701,-0.0667312816,0.0916443765,-0.0896682367,0.0401452221,0.3479632735,-0.0852565169,-0.2581972182,-0.4077641964,0.0772904232,0.2306943238,-0.2337165177,-0.0889581218,0.0437885262,0.0584510937,-0.1084640548,0.0704260394,0.1037857682,0.1214168966,0.1209269464,-0.1024183109,0.2657802403,-0.4172893465,-0.1044715717,0.0169202723,0.0542703532,-0.0603360571,-0.0149169071,-0.2595127821,-0.0674868524,-0.0660633519,0.2184064686,0.3321194351,0.436609,0.3683241904,0.136967048,0.0183150079,0.1096078008,0.2874644101,0.1851907372,-0.1328526735,-0.3547146618,0.3443646133,0.2502982616,0.0044988487,-0.2905494869,0.1117965281,-0.5064710379,0.0177645367,0.1481598914,0.4995015562,-0.070021376,0.3091269732,0.0297119077,0.2923957109,0.1236858293,0.1615945846,0.2556518614,-0.1314310282,0.068718195,0.270629108,-0.0009666967,-0.0322648473,0.2306300253,0.0727450028,0.3118824959,0.354347378,0.3712405562,-0.0912480801,0.1493474245,0.3343860209,-0.1993266344,0.2292354107,0.2483122349,-0.0521306768,0.4463845789,-0.4679372311,-0.3230538666,-0.2296720147,0.2725282907,-0.3189399838,-0.0621684305,-0.1676694751,-0.0924040377,-0.142683357,0.0767872334,-0.2698045075,-0.0065294732,0.189479813,-0.1842357069,-0.0856516063,-0.4537175,-0.3113759756,-0.1693000644,0.3823489249,-0.1074153259,0.2846629024,0.0380349122,0.0034094728,-0.0238400996,0.3346515894,0.0962040424,0.1980905235,-0.0604442656,0.0217161905,-0.1130579859,-0.0072518983,0.0356865302,0.2513354421,0.1394228488,0.004351439,0.0718948245,0.0843121186,-0.1742600948,0.2088598162,-0.2105252445,0.2519859374,0.0795351192,-0.0590310842,0.0023219078,-0.0617113635,-0.4243811965,-0.2515658736,-0.2580960989,0.1377370358,0.2446865439,0.067351982,0.1199315637,-0.3152993917,0.1366770566,0.2973252237,0.490390718,0.261811316,-0.0695900097,-0.0861519501,0.0290581863,-0.1447816044,0.2138178647,0.1083106995,-0.3495375812,-0.1055348441,-0.077762492,0.074202612,0.1447277963,-0.4900467992,0.0882464573,0.0159529597,-0.2991987467,-0.1523761898,0.27576226,0.2243058383,-0.1155274436,0.1021832675,-0.1675844193,0.2022671103,-0.0899589434,0.0990945697,-0.0947201476,-0.5073533058,-0.2690565288,0.0019004121,0.1711944789,0.232400924,-0.229864791,-0.2758561671,-0.0460463837,0.0296141692,-0.0929939225,-0.1754731387,-0.0921415389,0.0415048823,0.1299182922,0.1022346094,0.0525227822,-0.0641561449,0.3429868221,0.3477360606,0.19087708,-0.1066462547,0.0014483362,-0.2416457236,-0.1074558198,0.0510744266,0.7210329175,-0.0177647285,-0.0027940497,-0.2597800493,-0.3887057304,0.5166321993,-0.2793808281,-0.0613865741,0.002797507,0.1745122373,0.6265729666,-0.1433508992,-0.6584721208,0.015463952,-0.0379904248,0.0367274098,-0.3834742904,0.3062702119,-0.2717520893,-0.0689118207,-0.0705546066,0.2927133739,0.3333768249,-0.1025821492,0.1612814367,-0.1487855464]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/483","title":"rotten tomatoes movie review dataset taken down","comments":"found a mirror: https:\/\/storage.googleapis.com\/seldon-datasets\/sentence_polarity_v1\/rt-polaritydata.tar.gz","body":"In an interesting twist of events, the individual who created the movie review seems to have left Cornell, and their webpage has been removed, along with the movie review dataset (http:\/\/www.cs.cornell.edu\/people\/pabo\/movie-review-data\/rt-polaritydata.tar.gz). It's not downloadable anymore.","comment_length":4,"text":"rotten tomatoes movie review dataset taken down\nIn an interesting twist of events, the individual who created the movie review seems to have left Cornell, and their webpage has been removed, along with the movie review dataset (http:\/\/www.cs.cornell.edu\/people\/pabo\/movie-review-data\/rt-polaritydata.tar.gz). It's not downloadable anymore.\nfound a mirror: https:\/\/storage.googleapis.com\/seldon-datasets\/sentence_polarity_v1\/rt-polaritydata.tar.gz","embeddings":[0.0940540954,0.4555675983,-0.1925284714,-0.0446762145,0.123290211,0.0647398829,0.056007918,0.2448810488,-0.2137286812,0.0006343164,-0.3237887919,-0.0520209409,0.0047657439,-0.0996730775,0.1637380719,-0.0255666692,-0.2327698022,0.0319654308,-0.0766340494,0.0278008506,0.1223046556,0.3403700888,-0.0673023313,-0.1344223022,-0.3998959959,0.0771637931,0.1165543795,0.1774563938,0.0875604004,-0.3275774419,0.2422607392,0.1632737368,-0.0485329032,0.4094734192,-0.0001110224,-0.1601299047,0.3215585351,-0.1606253982,-0.1866290569,-0.1719692945,0.0704915449,0.1957471073,0.098994866,0.0760917738,-0.0043797898,-0.2227912992,-0.0328451209,-0.2059768289,-0.1390450001,0.2013765126,0.2405616343,0.176354602,-0.0493229553,-0.1542336941,0.4417915642,0.5694614649,-0.2073954195,0.3289612234,0.374771446,0.1450482458,-0.0732908025,-0.2259387672,-0.1041841507,-0.3226753175,-0.2406382561,-0.0803922787,0.094251506,-0.2807690501,0.2221280336,0.4855825007,0.4475929439,-0.1409691274,-0.1519214362,0.3069649339,0.1159941033,0.0105749555,-0.0934712589,0.387421906,-0.1597946137,0.1039969102,-0.3373256326,-0.01776262,-0.1479109228,0.3047358692,0.2583153248,0.532348156,-0.1543065012,0.0273748245,-0.0785035491,0.1599720865,0.2658187151,0.2231928408,-0.0135134039,-0.0287173744,0.0628177896,-0.0908658653,-0.0627594888,0.0692009628,0.1191623509,0.3358511925,0.2474713624,0.0767029151,0.090149574,-0.0281955749,0.2605776489,0.2335509807,0.2119104564,0.2185748369,0.316783905,0.3308602273,0.1646823436,0.1393978298,0.318436712,0.0872360244,0.0413686186,-0.1328180134,0.3198427558,-0.5254862309,-0.4966595173,0.0666769743,0.0012720502,0.0408399142,-0.0479248054,0.2428197712,0.0042080432,0.2102183849,-0.0233539585,-0.1217117608,0.1493306309,-0.2877396047,-0.2407813221,0.0221564416,-0.2007027119,0.0774774849,0.3926480114,0.049744878,0.4328015745,0.0174110122,0.1938224733,-0.2071909606,-0.1890285015,-0.0467716195,0.068620123,0.3564685285,0.1679783463,0.1984087378,-0.0752494857,-0.0350712538,-0.1567734778,0.1336804032,-0.0904504433,-0.0653847158,-0.1074029431,0.1539177299,-0.205506146,-0.2669534683,-0.0114457998,0.0553992502,-0.0968753397,-0.1292734891,0.138723582,0.0081420932,0.0199098885,-0.204520911,0.0685279667,0.3852422833,-0.6462727785,-0.1429380029,-0.4373919368,-0.1484640688,0.094347395,0.4501593411,-0.1872880906,0.1566172689,-0.2223725915,-0.0928399265,0.1148130521,-0.010372906,-0.3982384205,0.3331846893,-0.2671540976,-0.07354711,-0.1468459666,0.444817692,0.1754748821,-0.0660165176,-0.1338246167,0.191841349,0.058539819,-0.1454662532,-0.2378510982,-0.0212140139,0.1169760227,0.1178371906,0.087631233,0.0189589169,-0.0493781045,-0.3929144144,0.5155304074,0.224304229,0.2515768409,0.2920586765,0.2888417542,0.6333485246,0.0220696889,-0.1924763173,-0.1694410592,-0.0204621367,-0.2440411896,-0.0320210829,0.0841482207,-0.2959602177,-0.5710161924,-0.1525357962,0.0096026622,-0.2535406053,0.1734365821,0.0307454988,0.0272590369,0.2730031013,-0.1525999606,-0.0160354245,-0.3008124828,0.0429173857,0.2631682158,0.2422620207,-0.0244537424,-0.0199986268,-0.0400701612,0.0406226963,0.1049476787,0.0776636973,0.0666889623,0.4315847456,-0.1769899577,0.2213604748,0.4216573536,-0.1912097633,0.0708011314,-0.5606147051,0.2678551972,0.3351642489,0.0088552358,-0.0095648514,-0.0917360261,0.3935841024,-0.3568823338,-0.0384046696,-0.1997295171,0.337611258,0.3043012917,-0.0861738846,-0.148616299,-0.328395009,0.308730036,0.0279157348,0.13022241,-0.0723603219,0.027502358,0.1032536924,0.1874690652,-0.1126042381,0.2001607567,0.1510922611,-0.3021748364,-0.0869999006,0.0543249473,0.2666806877,0.2831827402,0.159857899,-0.030241793,-0.0504479259,0.0160902776,-0.3615487814,0.0805819035,0.351831466,0.0157139003,0.2835303843,0.143966198,0.2155357003,-0.3603385389,-0.1969078481,-0.220617041,-0.1712505966,-0.0924401134,-0.0691270381,-0.3581632972,-0.0808029994,0.035019502,0.0346807428,-0.0065634609,-0.3625300527,0.0834013596,-0.2614160478,-0.1812210977,0.0478658192,0.0567235723,0.4065265357,-0.1806188375,-0.0762769356,0.0671649948,-0.4359031022,0.0085841427,0.1705049872,-0.0344226696,-0.1586312652,0.2676543295,-0.2769360542,-0.2669092417,-0.5582646728,-0.251609832,-0.0444747321,-0.1144501194,0.0808844045,-0.0362869278,0.274903506,0.0050978935,0.0423110202,-0.178629905,-0.2406110913,0.0634810627,0.1085472107,-0.0248392597,-0.2097473294,-0.0809650123,-0.2972662449,-0.0727893338,-0.032947883,0.4604971409,0.1655857861,-0.1622152925,-0.4276564121,0.0613846444,-0.0143365012,-0.2781353593,0.3352944255,-0.2274308205,-0.0732620433,0.4022460878,-0.2346164733,-0.5364183187,0.2289667726,0.354026407,0.2667643726,-0.0956747681,-0.6627873778,0.5682843328,-0.2134162188,0.1523216814,0.1598393321,-0.3807866275,0.4891273081,-0.2603232563,-0.1297950894,0.1433164328,-0.0196987074,-0.1455938518,-0.1142857894,0.1982171685,-0.0998679399,0.1507519782,0.1270790994,0.4196516573,0.1266522706,-0.0369915813,0.0092923436,0.0675952137,0.5280759931,0.2062376291,-0.1369837373,-0.0234205015,-0.1828273982,0.0650143698,0.1369013786,0.1611305475,-0.4437626004,-0.3157825172,-0.0158817489,-0.4479649961,-0.1289087981,-0.0853440985,0.1801709235,0.2442516685,0.1312528253,-0.0966498256,0.1640009731,-0.4331972003,0.0262719076,0.3447406888,0.1906983107,-0.2435461134,-0.0798121095,0.090342395,-0.0886077508,0.2377224267,0.0371615626,0.023133954,-0.2210369855,-0.0974368677,0.2439657301,0.1940432638,0.1727726907,0.2408530414,0.1433678567,0.1213245168,-0.1177679524,-0.3854034543,-0.1316297799,-0.5065336227,0.2708575428,0.4020624757,-0.2309347093,-0.0771477148,0.0728881508,0.3521401286,-0.0018691343,-0.2231574208,0.0135067208,-0.0549913123,-0.0284601115,0.0679131076,-0.0822014362,0.1385559589,-0.2637467384,-0.3058560491,-0.1809122413,0.2325206101,0.0910278037,-0.0511917174,-0.1252362132,0.2131093591,-0.1399732083,-0.2201273888,-0.1394578516,-0.3411504328,0.1387062818,0.2241586,0.0654819459,-0.0554979965,0.3310298026,-0.2199854702,0.1385105252,0.5659031868,0.0852977112,-0.2383994311,0.1570749283,0.0080003506,-0.1740477979,-0.2082505077,0.4484551549,-0.2764577568,0.0766038299,-0.2481069416,0.1622960567,0.0210186802,-0.1254492253,-0.0795992836,0.3963926733,0.0148343891,0.2837372124,0.2261485904,1.1940380335,0.050347738,0.3148739934,-0.1955645829,0.1220615208,0.3619166613,-0.1202946156,-0.1026569009,-0.101084277,0.1447667331,-0.1747340709,0.0220299866,-0.0816470012,0.027267741,-0.2091069371,0.2223455459,-0.1841825992,0.2955374122,0.3719804585,0.0480270311,-0.1790689826,0.1280254275,0.112728633,0.2833483219,-0.1367085725,0.1649043858,-0.1536378413,0.0376406573,-0.3256208301,0.35081622,-0.7079859376,0.0064529842,0.0184029751,0.3670134544,-0.0697303712,-0.1226571724,0.087309733,0.3724011481,0.1474944651,0.4030211568,-0.2574906647,0.2756005526,0.1932769716,-0.2305887192,0.0918465704,0.1356651336,0.0667254776,-0.0331059992,-0.4333132803,0.153099969,0.2694683671,-0.1632262766,-0.4295158386,-0.0327398144,0.0818024054,-0.4394486547,0.0631834641,-0.0427750982,-0.1052614748,-0.4180672765,0.1570329368,0.2290703654,0.1182655618,-0.2130827606,0.5677035451,-0.1935963035,-0.2349116951,0.5121383667,0.0507252738,-0.0899918526,-0.0658322871,0.4183778167,-0.2554006577,-0.3486641645,0.1413064599,-0.5356818438,-0.9375551939,0.238822788,-0.3766698539,0.1107908487,-0.171042487,0.1011813134,0.3020537794,0.2883433104,-0.3774154782,-0.2976205945,0.0953374878,0.3074828684,0.1782675087,0.098431237,-0.1052828357,-0.0896246731,-0.0036510767,-0.1309582293,-0.3134093583,0.1143770814,-0.0257338155,-0.1895766556,0.1728840768,-0.2369530201,0.2333766818,0.0063554435,0.1487109512,0.1349516213,0.2518803179,-0.2201871872,-0.11685168,0.1491662562,-0.0746890679,0.0191750191,-0.350217104,-0.0969824716,-0.0586344227,0.0326283164,0.0976009965,0.2719036341,0.3852096796,0.3732555509,0.2502100766,-0.2781343758,-0.0840357691,0.2413684875,-0.3474979997,0.0684797242,0.2824778557,0.0050452044,0.1358584911,-0.1114670411,-0.3027317822,0.1682533473,0.0321402475,0.1586141139,0.3610900342,-0.0070564784,0.1099820063,-0.0282749292,0.4179670513,0.2786771059,-0.1858602762,-0.3494634032,-0.0692413524,0.1869714409,-0.1948360205,-0.0564129539,-0.1440417618,0.3578687608,0.0010458912,-0.0211771633,-0.1670697778,-0.1268314868,0.2107464373,0.2056733072,0.5767796636,-0.1506790072,0.1188329458,-0.1312868148,0.305264771,0.0820804313,-0.1250300407,0.0138886739,-0.0998872668,-0.1088150814,-0.0532037877,0.0029388391,-0.2203211635,0.2685793936,0.1366353333,0.0604541413,0.1717728823,0.3435311913,-0.5017744303,-0.0991309658,-0.0215336028,0.728415668,-0.0347214043,0.1781880111,-0.317352742,0.1873170137,-0.1872605532,0.0443220958,-0.39474684,-0.1843997836,-0.2075345814,-0.0543727502,0.0567477942,0.0723546818,-0.0717474669,0.2988768816,-0.2646150887,-0.4923280478,0.3523026705,-0.2333865315,0.2904116213,0.0644598082,-0.1110250428,0.367792964,0.3126113117,0.0228474848,0.2799554765,0.1359490454,-0.1592780054,-0.0173198469,0.2239377052,-0.0278091729,-0.3411594629,0.1192718595,0.2320408076,0.2695895433,0.0378967337,-0.0586835295,0.3102678359,-0.1842352003,0.2413053811,0.1346258521,0.5890652537,-0.1337419599,0.3602931201,-0.1799692065,-0.1612443924,-0.2493399233,-0.1744726896,-0.2741250992,-0.6247327328,-0.1271552145,-0.2858250439,-0.0239451844,-0.0994527489,0.0861564726,0.0016694483,0.2423235923,0.3868529201,-0.2181400955,-0.13947469,-0.3728772402,-0.396196574,0.1299952269,0.2793727219,-0.1440758407,-0.2180984616,0.1645123363,-0.2008341551,0.1906246841,0.2340087742,0.2141223848,-0.1493299305,-0.1899821758,-0.3438534737,0.0683100298,-0.0073822732,0.2053282559,0.0246750489,0.0654749349,-0.2218527794,0.0004238092,0.0834513903,-0.1311770231,-0.6050914526,0.0143213784,0.616417408,-0.0825968608,-0.1465164423,0.2006687224,-0.0599720255,-0.2555076778,-0.2868143618,-0.2821010351,0.1883329004,-0.0799630359,0.0214171987,0.2705541551,-0.1324932724,-0.16997917,0.1600939929,0.5236707926,-0.0699198991,-0.1889999807,-0.4578751028,-0.1237701774,-0.179148972,-0.0054741316,0.1437795758,0.0539907739,-0.0346337967,0.0447724313,-0.0880721435,-0.3539851904,0.6424487233,0.0869730115,0.2317032516,-0.1527982503,0.1674140245,0.0955496654,-0.0649281442,0.057508003,0.0652301759,0.1336941123,0.1767800748,-0.2080178261,0.4886719584,-0.6226924658,-0.034884233,-0.0593651682,0.2971400321,0.0604520179,-0.4060911834,-0.0470215194,-0.2059119195]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/483","title":"rotten tomatoes movie review dataset taken down","comments":"fixed in #484 ","body":"In an interesting twist of events, the individual who created the movie review seems to have left Cornell, and their webpage has been removed, along with the movie review dataset (http:\/\/www.cs.cornell.edu\/people\/pabo\/movie-review-data\/rt-polaritydata.tar.gz). It's not downloadable anymore.","comment_length":3,"text":"rotten tomatoes movie review dataset taken down\nIn an interesting twist of events, the individual who created the movie review seems to have left Cornell, and their webpage has been removed, along with the movie review dataset (http:\/\/www.cs.cornell.edu\/people\/pabo\/movie-review-data\/rt-polaritydata.tar.gz). It's not downloadable anymore.\nfixed in #484 ","embeddings":[-0.0425177105,0.4346787035,-0.2001430392,-0.0240862258,0.1695964336,0.0062655173,0.0608291626,0.2478503287,-0.143117398,-0.0300086234,-0.2994853854,-0.0205635075,0.0943494886,-0.0123527357,0.1843734533,0.0148354806,-0.2772168815,0.0558446981,-0.0992571265,0.0433850586,0.0410708375,0.2735620141,-0.0664727017,-0.186864838,-0.3904908001,0.0617674105,0.1458568573,0.1576227844,0.0013437525,-0.3609345853,0.2465199381,0.0694457665,-0.0664036646,0.3829580843,-0.0001062083,-0.1119772643,0.414272964,-0.1334920377,-0.1781807095,-0.1665208489,0.0212568939,0.1716620773,0.0790254995,0.0573734455,0.0295980461,-0.1847918183,-0.1111538038,-0.1819520891,-0.1957961023,0.224344492,0.2899101079,0.1693576425,-0.0677215829,-0.1663078219,0.4000733197,0.5402953625,-0.2552702725,0.3032431304,0.3221278787,0.0581582114,-0.0588981472,-0.2542583048,-0.036473047,-0.2770753205,-0.0911006033,-0.0213097595,0.2304927409,-0.2793714702,0.2037946135,0.4458133578,0.3564076722,-0.1189332232,-0.0664007589,0.3131410182,0.0948068276,-0.0580062494,-0.0561563,0.3654698431,-0.1676198691,0.0345616341,-0.4095421731,-0.0361781605,-0.1133342981,0.3079210222,0.2773034573,0.3394732475,-0.13812235,0.0324712954,-0.080508709,0.1882621944,0.3298357129,0.1566830128,-0.0184359346,-0.0310217552,0.0910966769,-0.053829167,-0.0376176909,0.1933215708,0.0910597146,0.4261573553,0.2448746264,0.0122014889,0.1058240086,-0.0171215441,0.3109273314,0.2003108412,0.2103099823,0.2286485583,0.2947466373,0.4243248403,0.1809052229,0.1215302125,0.4473460019,0.0426494665,0.027752921,-0.1092764363,0.3925181925,-0.5196741223,-0.4537936151,0.0238518119,0.0959167033,0.0138405347,0.0150575815,0.2352998555,-0.029295383,0.1440760493,-0.0264459588,-0.0545453727,0.1373672485,-0.2714265585,-0.2753900886,0.0227675736,-0.2249884754,0.0708259046,0.370842129,0.0080192806,0.4234260321,0.1166623086,0.292260021,-0.2304561585,-0.1831,0.0102213072,0.1211543605,0.3974051774,0.2149576694,0.1524143219,0.0169341583,-0.1154069453,-0.1893757433,0.1620360911,-0.1198738068,-0.0453335606,-0.0785018802,0.1994000226,-0.129990086,-0.2274788618,-0.1304431409,-0.0580128543,-0.0739096329,-0.1184447184,0.1511688381,-0.0287546851,0.0540309362,-0.2003847361,0.086181052,0.2543032467,-0.5500470996,-0.1202052087,-0.4857307076,-0.1800608486,0.1223926619,0.39572528,-0.2374000549,0.0829950422,-0.2673150003,-0.0457124077,0.1395882666,-0.0424012765,-0.3844663501,0.3787454665,-0.2163757384,-0.0970471799,-0.2568163872,0.3944224119,0.1314385533,-0.0844076648,-0.1584301144,0.2240536064,0.0345205627,-0.1488685012,-0.3276514113,-0.0202520043,0.172339052,0.1039419547,0.0731207579,0.1068923771,-0.0696667433,-0.4124529362,0.4310690165,0.2566543221,0.2242834419,0.3019278049,0.4085743427,0.5462064743,-0.0135489292,-0.1349063367,-0.2543875873,-0.0033694543,-0.270578593,-0.1060613617,0.1357240379,-0.3536294699,-0.5882774591,-0.0949773341,-0.0448892526,-0.2575763166,0.2322235107,0.0560700409,0.029373331,0.2036349773,-0.189851746,-0.0066062673,-0.2871998847,-0.0303050242,0.2176257968,0.3131098151,-0.1108722687,0.0249817893,-0.0142334998,-0.0364905596,0.2318820208,0.0382913202,0.0588419884,0.4436085522,-0.1367713362,0.0787403882,0.3824531734,-0.1866581291,0.1096099615,-0.4445562661,0.2320247889,0.3345392942,-0.038070932,-0.0416640639,-0.0328852423,0.3432286382,-0.4450238049,0.0261426251,-0.1495265216,0.3595175147,0.3035085797,-0.1501086652,-0.0800596252,-0.3423942327,0.2037432939,0.0186753701,0.1776353717,-0.1219388917,0.0518959314,0.191124782,0.1945070624,-0.087395899,0.0859948471,0.1815176606,-0.212667942,-0.1060632169,0.0900483429,0.3242691159,0.2910448909,0.1476190686,-0.0019168714,0.0105125979,0.0391529612,-0.3629914224,0.173508659,0.3270922899,-0.0359209143,0.2397377789,0.03675301,0.110227704,-0.3180398047,-0.1231074035,-0.1504159868,-0.1154993027,-0.1346254349,-0.1055756584,-0.3872300684,-0.0553183742,0.0775241479,-0.0059701656,-0.0695519522,-0.3501300514,0.0659217164,-0.2155702561,-0.1944916993,0.0700005218,0.080712229,0.3925210834,-0.1906186789,-0.0485890843,0.0221393071,-0.3374360204,0.0042886948,0.1845878661,0.0767211467,-0.2191099972,0.217491582,-0.202806145,-0.2318126112,-0.5963714719,-0.1703891456,-0.0452792905,-0.1051889807,0.1297582239,-0.0571920052,0.2892864645,-0.0345070101,0.0640534386,-0.1233652607,-0.2200827003,0.0573314093,0.1081344187,-0.0145330224,-0.2240079641,-0.1249017119,-0.3192481995,0.0002285714,-0.0392880067,0.3625760078,0.14786084,-0.2119968981,-0.3666348755,0.0741685033,-0.0613784418,-0.2107428312,0.1944462955,-0.268681407,-0.1186245158,0.3863391578,-0.3013879061,-0.569570303,0.2782772481,0.3523666263,0.2719342411,-0.0598820783,-0.6587876081,0.505425334,-0.1803620309,0.2098187506,0.1368752271,-0.3023393154,0.4541866481,-0.2040985823,-0.1703178287,0.1336490512,-0.0337418281,-0.1638405025,-0.0827978998,0.2085521221,-0.1151697785,0.2178942412,0.1372561604,0.5060987473,0.18984586,0.0153649012,0.045500204,0.0096650701,0.531404078,0.1946870238,-0.1534656435,-0.0450501628,-0.2327311188,0.0367132239,0.1051944271,0.1086657196,-0.4172174633,-0.3038907945,0.0844528228,-0.4155604541,-0.0983314514,-0.140613839,0.2304974347,0.2431427687,0.1856350154,-0.0550593957,0.1669914126,-0.3893910944,0.085456036,0.3329642415,0.1714319736,-0.2430906594,-0.180283621,0.1069075018,-0.0576144941,0.235836342,-0.0196398534,0.0113239074,-0.1629834473,-0.0835314691,0.2419518083,0.1536414623,0.1642980278,0.1887830049,0.1629068702,0.1525154263,-0.1749305725,-0.315966785,-0.1406672895,-0.3722186089,0.3619772494,0.3826577961,-0.2560275793,-0.0488899648,0.032582663,0.2819558382,-0.0959507525,-0.1856266707,0.0241530277,-0.1020044312,-0.0460778624,0.0800114796,-0.1623713672,0.2211076468,-0.2170538306,-0.3347088695,-0.1619302034,0.2200987041,0.0760787427,-0.1274351776,-0.1892721951,0.2314534634,-0.0947152674,-0.1249885038,-0.1866184175,-0.2746611536,0.104915075,0.2858870029,-0.0315358043,-0.1499806195,0.2939164937,-0.2226632982,0.1173172072,0.494045198,0.015567502,-0.3028881848,0.1049653739,0.0171447154,-0.1631424874,-0.1694465727,0.4149158597,-0.4274544716,0.113393873,-0.2342576683,0.291686058,-0.0495322645,-0.0665088817,0.1024800166,0.3948923945,0.0451772921,0.1369609684,0.2463107109,1.076125145,0.0439967141,0.227079615,-0.1277648062,0.0205886792,0.2399429083,-0.0945207328,-0.0481831357,-0.1798062027,0.0237978417,-0.1184894815,0.0011581276,-0.1061653942,0.0703007728,-0.1755012572,0.1717706472,-0.2949557304,0.2743045092,0.3793833256,0.1075799763,-0.059462022,0.1360493451,0.104062587,0.3518154323,-0.1614543498,0.1994654387,-0.1945292503,-0.0234437417,-0.3429661989,0.3184152246,-0.5974160433,0.0913101435,0.0278481878,0.3982093632,-0.0116677983,-0.1360946894,0.0503366254,0.2900057137,0.1039235741,0.2951078415,-0.2940893173,0.3322527409,0.1938341558,-0.2425386757,0.0788229033,0.1412498355,-0.0535425507,-0.0632306188,-0.4027332962,0.1010028943,0.2538533211,-0.219342351,-0.3932823539,0.0868903995,0.1671736985,-0.4511369467,0.0307115,-0.1121113449,-0.06019862,-0.4033514857,0.1997456551,0.2462611645,0.1638522148,-0.223736316,0.5695924163,-0.133738175,-0.2025265843,0.506814599,-0.0183149818,-0.1476729959,-0.1487909853,0.3178934455,-0.3098047078,-0.3480024636,0.2033554465,-0.4452024698,-0.8656628132,0.2687439024,-0.289893657,0.195566982,-0.1627144068,0.1928774863,0.3080750108,0.3250721097,-0.299852103,-0.2426016033,0.1157866269,0.2056571245,0.1725203246,0.0775786489,-0.1493403018,-0.0990167335,0.0676428899,-0.1065699831,-0.3541980982,0.126627326,-0.0328225456,-0.2137897015,0.1207705215,-0.2574236393,0.200666815,0.0510309711,0.1834376305,0.0873812065,0.1554478556,-0.243173629,-0.0891108811,0.116089426,-0.0289184526,-0.0718589798,-0.406129688,-0.1107081398,-0.1467868686,0.0557551123,0.0848791078,0.2412121296,0.3097571135,0.2837934196,0.1994093806,-0.3732251227,-0.1144229546,0.2308653295,-0.2764767408,0.1165286377,0.2797234356,-0.0157739427,0.1170718446,-0.0671715885,-0.2622081935,0.1092008427,-0.0282032806,0.2041098326,0.4132621884,-0.0969247669,0.1460663527,-0.0031455613,0.3676292896,0.3374555707,-0.220771119,-0.2729562819,-0.0852099732,0.2120525986,-0.1884618849,-0.0462723486,-0.0872905999,0.3794158399,-0.0299748909,-0.015602015,-0.1304840595,-0.1060692593,0.2614061236,0.2381585985,0.5742771626,-0.1976741999,0.1529000551,-0.175391838,0.1746341437,0.0480525233,-0.0998480693,0.0290441159,-0.135741666,-0.0367135592,-0.0666037872,-0.0159275997,-0.2073397487,0.2319063246,0.0771688595,0.0023960797,0.1754984409,0.2943331003,-0.4259499609,-0.1250205338,0.0210265964,0.6062902808,-0.1273917258,0.10940703,-0.2782261968,0.1542990953,-0.1734231412,0.1325293183,-0.3427048624,-0.1454211324,-0.2080765814,-0.0189840402,0.0513188206,0.0529097952,-0.0334809721,0.3218859732,-0.3103375137,-0.3582104146,0.2752542496,-0.1386135519,0.2494836599,0.0412021279,-0.13279818,0.3140447438,0.3320236802,-0.026737038,0.1786135286,0.1486883759,-0.0337449908,-0.0533670336,0.1763946414,0.0634451807,-0.345197171,0.1292018443,0.2051163316,0.2231862545,0.0805968121,-0.0116251968,0.3359437883,-0.1965648085,0.1868705302,0.1322940886,0.6247689128,-0.112264663,0.4068168998,-0.2491350472,-0.1285525411,-0.2920045853,-0.2385378331,-0.2222699374,-0.6665386558,-0.0676465556,-0.2732073963,-0.019753661,-0.134072274,0.123805441,-0.0374659225,0.2619590461,0.2605614364,-0.2457686067,-0.1704645753,-0.4622826874,-0.4255818725,0.1256656498,0.2218436599,-0.1716667116,-0.2225747854,0.224270463,-0.2819304466,0.1576017588,0.2894673049,0.0922379345,-0.0485864542,-0.1678014845,-0.3159099221,0.1078540012,-0.0339997113,0.152555272,0.1145836636,0.0375181884,-0.1723562777,0.0100018205,0.1506011039,-0.098590523,-0.5656619072,0.0143627478,0.6272609234,-0.0259617418,-0.1116242483,0.2080225199,0.0101612303,-0.2729607224,-0.2767693698,-0.301441282,0.1784823239,-0.000061383,-0.0193524994,0.2006396949,-0.0721862987,-0.1681746989,-0.0159409828,0.5814570785,-0.0214620568,-0.1328322887,-0.3982604444,-0.1396933496,-0.1026489884,0.0262259841,0.1676099747,0.000086801,-0.0600685105,0.0421081036,-0.116929844,-0.3400464952,0.7199472785,0.1161214486,0.1673715115,-0.1570130736,0.2671844065,0.0438959338,-0.1353106201,0.054086294,-0.0306417663,0.1364461333,0.1390104145,-0.1864757538,0.4933222532,-0.6410788894,0.0058132457,-0.0500834696,0.3512328863,0.0967449844,-0.3907137513,0.0452415757,-0.1817022115]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/483","title":"rotten tomatoes movie review dataset taken down","comments":"Closing this one. Thanks again @jxmorris12 for taking care of this :)","body":"In an interesting twist of events, the individual who created the movie review seems to have left Cornell, and their webpage has been removed, along with the movie review dataset (http:\/\/www.cs.cornell.edu\/people\/pabo\/movie-review-data\/rt-polaritydata.tar.gz). It's not downloadable anymore.","comment_length":12,"text":"rotten tomatoes movie review dataset taken down\nIn an interesting twist of events, the individual who created the movie review seems to have left Cornell, and their webpage has been removed, along with the movie review dataset (http:\/\/www.cs.cornell.edu\/people\/pabo\/movie-review-data\/rt-polaritydata.tar.gz). It's not downloadable anymore.\nClosing this one. Thanks again @jxmorris12 for taking care of this :)","embeddings":[-0.0065911743,0.4376698732,-0.2165104002,-0.0086744213,0.1876952499,0.0448232442,0.118142657,0.2147626132,-0.1493462473,0.0232173353,-0.2698146701,-0.0061037922,-0.0091157993,0.0222848579,0.2036159486,0.0000030083,-0.2352455258,-0.0090849437,-0.1023562402,0.0237661283,0.042926129,0.241335243,-0.0023685419,-0.0987604782,-0.357585907,0.0769985244,0.1378089637,0.1850385517,0.0626495779,-0.3585107625,0.247290805,0.025454279,-0.115621455,0.4012389481,-0.0001056437,-0.1799888313,0.4209690988,-0.0974084586,-0.2103877813,-0.0918653607,-0.0741261542,0.1679461896,-0.0053974027,0.0704704449,0.0371151492,-0.2479953468,-0.1039154306,-0.2580784559,-0.2070547193,0.1964354366,0.2981702983,0.1534164697,0.0139345257,-0.212188527,0.3829687536,0.5618550181,-0.2078839242,0.2671195567,0.3738715053,0.0268775765,0.0098544583,-0.2530751526,-0.1247496456,-0.3226585388,-0.070865944,-0.0332183838,0.132447511,-0.3221164048,0.1826387197,0.4543637335,0.3458320796,-0.1193780452,-0.0403307863,0.2553261518,0.0723031312,-0.0229790136,-0.0582210906,0.3687548339,-0.1132051572,0.0598489679,-0.4200324416,0.036362756,-0.1778634638,0.3410974145,0.3124570549,0.3328334391,-0.1589151174,0.0284387469,-0.1084669828,0.1300759614,0.2836641371,0.1041067913,-0.0488378629,-0.0434333868,0.0748403221,-0.1588034183,-0.06477689,0.2465389371,0.1432458758,0.3947619498,0.2449945807,0.0210920218,0.0989675522,0.0136807263,0.2877883017,0.2334885895,0.2129514813,0.1784964651,0.2633239329,0.3567848504,0.211352393,0.1506782472,0.4886956811,0.0932058394,0.0044757235,-0.0680958703,0.3900867403,-0.5632160902,-0.4454048574,0.066030018,0.0462380946,-0.0046013361,-0.1020336077,0.1919036359,-0.0766805783,0.172970742,-0.1112978384,-0.0675990433,0.1794827729,-0.291315645,-0.2379930466,0.0597203858,-0.2197054327,0.063409999,0.3678866923,0.0677784532,0.3497131765,0.1425009072,0.2561374605,-0.1671324223,-0.1364707947,-0.0520279109,0.1695349962,0.3227740526,0.1660673618,0.0951249301,-0.0011324582,-0.133616209,-0.2206154019,0.1516243815,-0.1154286489,-0.1362822652,-0.1222884133,0.2169479132,-0.114830777,-0.2340394258,-0.1459544003,0.027179189,-0.1597411335,-0.0443798602,0.1957339197,0.0507835001,0.0532822683,-0.1794371456,0.0432288721,0.249582231,-0.5570643544,-0.1752274036,-0.5179738402,-0.1333630681,0.1067644283,0.4184054732,-0.3076773584,0.0741191804,-0.2766211033,-0.0867779404,0.1109354943,-0.0205748081,-0.3217777014,0.3450165093,-0.2729656994,-0.1053742617,-0.2218796015,0.3316842914,0.1635392755,-0.1418839544,-0.1455579698,0.2107923329,-0.0298467465,-0.1702639014,-0.3384591937,-0.090645723,0.2250682712,0.1253034323,0.1322069019,0.0892665982,0.0019045477,-0.3978773654,0.4333865047,0.2159040421,0.1760693789,0.2955708504,0.3349010646,0.5573027134,-0.0264115911,-0.1800320446,-0.2361311316,0.0360340364,-0.2118023783,-0.0378275439,0.1299254,-0.335945189,-0.5320709944,-0.1202111244,-0.0256965943,-0.1892568171,0.2300427705,0.0226965789,-0.020523563,0.253899157,-0.2197715193,0.0211226493,-0.3581687212,-0.0342078991,0.189236939,0.2799786031,-0.0503524952,0.0513696074,-0.1187346056,-0.0140695395,0.1979029924,0.0198590718,0.0892316401,0.3717536032,-0.2177155316,0.1344038993,0.4738063514,-0.1350747794,0.1542545408,-0.5148200393,0.2570678592,0.3062506318,-0.0392744802,0.0105183693,-0.0204097256,0.3734461665,-0.4798629284,-0.0213457737,-0.1717686057,0.3040292859,0.2129464149,-0.2179093212,-0.0884834379,-0.3202958703,0.2316357195,0.0666420758,0.0832174644,-0.0905950889,0.0536588654,0.2040904909,0.1840410084,-0.1423592269,0.0499196984,0.1673934907,-0.2075475752,-0.0275274999,0.14903678,0.3429442644,0.3149110675,0.1710702479,-0.0255391747,0.0047455882,0.0408645831,-0.3615794182,0.1636439264,0.3489165306,-0.0683705583,0.2553495467,0.0913520083,0.1867010444,-0.2801232934,-0.1629509926,-0.1195291057,-0.1866759658,-0.1380892843,-0.0345573835,-0.320395261,-0.1245862693,0.076410085,-0.0081774695,-0.0835399777,-0.3668524027,0.0300026461,-0.2269569039,-0.2150265127,0.077906616,0.0750158951,0.502830863,-0.2521342039,0.0424344577,0.0245197527,-0.329147011,0.0508351624,0.2299488485,0.0353367776,-0.2121757716,0.232399404,-0.1971635073,-0.2633458674,-0.5804544687,-0.1604035795,-0.0499560572,-0.1216268018,0.11094971,-0.1140733585,0.2434616089,-0.0514518619,0.0128352409,-0.1934855729,-0.1407234669,0.0784756392,0.0592518225,0.000141597,-0.2030197382,-0.1372383982,-0.2533162832,-0.0022761845,-0.0482274592,0.421100527,0.0900177807,-0.2193158716,-0.3628781736,-0.0298824944,-0.0711666197,-0.2148460299,0.1795571148,-0.3120980859,-0.1467647552,0.3454419076,-0.2620328665,-0.4810020328,0.2820281088,0.4178220332,0.2381892502,-0.0949915797,-0.6851104498,0.4739792645,-0.1148335114,0.1936091483,0.0619737916,-0.380022794,0.43835935,-0.1399410218,-0.1936519146,0.1873914897,0.0381003506,-0.2635920942,-0.0712363347,0.1940242052,-0.1167188734,0.185107857,0.1497163177,0.4770096242,0.0967363715,0.0071536787,0.0204873588,0.0527887717,0.5278574228,0.2184114456,-0.2154169381,-0.0107192071,-0.2063162774,0.0529220924,0.0257887878,0.0750602111,-0.3656348884,-0.2982551754,0.0848203227,-0.3621595502,-0.0845510513,-0.1357536763,0.2167340815,0.183927983,0.082237348,-0.1070183665,0.21000655,-0.4201793671,0.0789490491,0.3562915623,0.1692280173,-0.3017568588,-0.188927114,0.2025193721,-0.0993814394,0.2517163754,-0.0474369265,-0.0496832207,-0.2566456497,-0.0990754515,0.1999181807,0.2286241949,0.1451073885,0.2024359703,0.1692650765,0.1916670352,-0.153679654,-0.2595545352,-0.115335919,-0.4261903167,0.2931166887,0.3433510959,-0.1978345215,0.0335910581,0.0659800917,0.3324710131,-0.0792083144,-0.159600988,0.0674257129,-0.019970648,-0.0311827064,0.0129383747,-0.1016175151,0.2349246889,-0.2616199255,-0.3889122605,-0.1717571318,0.1995103508,0.1157748923,-0.041614484,-0.2489613891,0.1928519309,-0.0186863318,-0.2006255686,-0.1392802149,-0.2295650393,0.1212795749,0.2264294028,0.0500709824,-0.1839081645,0.2353452742,-0.2564819753,0.177224651,0.5713159442,0.1166540533,-0.1481983066,0.0858194605,0.0598053224,-0.1646331549,-0.1650073528,0.3897823691,-0.3494186997,0.1100600287,-0.2151382715,0.2329928428,-0.0078114518,-0.1278064102,0.1208455339,0.307598561,-0.0426214375,0.1474293023,0.2558953464,1.1909804344,0.1178209484,0.3273025453,-0.2274943441,0.0266737044,0.2753382623,-0.0749207884,-0.0407933891,-0.1117856801,0.0255696587,-0.1464927346,0.0059628519,-0.1308624148,0.0755315572,-0.184648335,0.1911976188,-0.3390553892,0.2494006157,0.3255573809,0.0733421147,-0.0794907287,0.106415011,0.0202581119,0.3605284691,-0.1490548551,0.1680771708,-0.1922412068,-0.0147678899,-0.3164013624,0.3768140376,-0.5935995579,0.0351278223,0.0595360436,0.3839949369,-0.0421651229,-0.1968352646,0.0714803338,0.3134936392,0.1501886696,0.311649859,-0.2441347837,0.2912779748,0.20758605,-0.1924942583,0.0720685795,0.0619443804,-0.0128130801,-0.0661608651,-0.4203489125,0.157970041,0.2363018543,-0.1509855986,-0.4925658405,0.0572381951,0.1436656266,-0.5287760496,0.0653272346,-0.1139057651,-0.0652329177,-0.3817667961,0.1903948635,0.2281348109,0.1679211259,-0.2210443318,0.5258335471,-0.15733473,-0.2270589173,0.4944015145,0.0024011319,-0.1889605522,-0.1242371425,0.3271078169,-0.2150552422,-0.3468016088,0.2059346139,-0.4634534717,-0.8311295509,0.2975915372,-0.3175131083,0.1889468729,-0.1570872068,0.1376854479,0.3245356977,0.4119129479,-0.2766850293,-0.2544139326,0.1164969951,0.2647309899,0.1863293648,0.0969992951,-0.1721424758,-0.1559516042,-0.0090106782,-0.1454637349,-0.3733613789,0.1627287418,-0.0788349882,-0.2131676227,0.1579469293,-0.2636973858,0.1873701215,0.0047371704,0.1871719062,0.0585480481,0.1728425473,-0.2468866408,-0.0770410448,0.129121542,-0.0134969316,-0.1341114193,-0.4133185446,-0.1239009276,-0.0587225445,0.0770249665,0.1004281566,0.271489352,0.3458616436,0.3331889808,0.2338912785,-0.4049432278,-0.1210312843,0.2269293964,-0.2285544276,0.1916351169,0.2739704251,-0.0384069681,0.1455686986,-0.0831250176,-0.2643877268,0.1444404721,0.0119784176,0.1859038621,0.4576817453,-0.0537987165,0.1240945831,0.0344022177,0.3135491312,0.2750907838,-0.2784972191,-0.2991606593,-0.0774141923,0.2365876436,-0.0961159766,-0.0921173319,-0.066585049,0.3976240754,-0.0634990409,-0.0255785342,-0.156109646,-0.1015250906,0.2699349225,0.27839306,0.5647424459,-0.2034147084,0.1967786998,-0.2277129143,0.1552578211,0.0464361645,-0.0952933729,-0.0045646024,-0.0587946437,-0.0423063338,0.0254534539,-0.008480696,-0.1617060453,0.244291082,0.1305864602,-0.0025708019,0.2172806412,0.2921613455,-0.4761501551,-0.0558175296,0.1157492474,0.6016530395,-0.1363413185,0.0148226218,-0.2025043815,0.1773383021,-0.1756211817,0.1842684746,-0.3128439784,-0.1812725216,-0.2207216471,-0.0247324705,-0.0025370519,0.044614315,-0.0171790905,0.3439127207,-0.3537909687,-0.3459858,0.3179205656,-0.0404572785,0.3235952854,0.0470111296,-0.1782194525,0.40327245,0.3126113415,-0.0415635891,0.258497715,0.1995006204,-0.1233037859,-0.0589465126,0.3310800791,0.0877224654,-0.3250439167,0.188880071,0.1833362877,0.2271639407,-0.0095165018,-0.005371972,0.3470929861,-0.167306155,0.1826497018,0.109051533,0.5974191427,-0.1837455183,0.3798433542,-0.188058272,-0.1875886619,-0.3722034395,-0.1618041098,-0.2209064215,-0.6824771166,0.026506288,-0.2371114939,0.0262493994,-0.1025190949,0.122604765,-0.0362072848,0.2926594317,0.3413610458,-0.2167255133,-0.1419109702,-0.4419909716,-0.3752931654,0.1873951554,0.2249195129,-0.1004029214,-0.1883077025,0.0669887885,-0.2300526202,0.1999649405,0.1749706864,0.0964575633,-0.0837200806,-0.1836559623,-0.3857408464,0.1063879877,0.0220570602,0.1995271742,0.0591421835,0.0730377734,-0.0834136456,0.0173457302,0.1534024477,-0.0871588811,-0.5477195978,-0.0042141536,0.6030766964,-0.0240917392,-0.0948803499,0.1995172799,-0.0009196308,-0.2796429396,-0.1526220739,-0.323857367,0.1679123938,0.0224531516,0.0313119777,0.1744945645,-0.0706663877,-0.1688027084,0.0640224367,0.5950391889,-0.074740231,-0.1182378978,-0.4729025364,-0.1951780468,-0.1153134629,-0.0190024804,0.1611736417,0.0248339958,-0.062065877,0.070062533,-0.1184021905,-0.298776865,0.6744161248,0.1912864447,0.0897693187,-0.2340462655,0.2363526821,-0.0559668541,-0.1234914958,0.0744884163,-0.1248336136,0.1742937416,0.1025775,-0.1898147315,0.4904205203,-0.5886353254,-0.0626317635,-0.0653836206,0.2551460862,0.05910144,-0.358631134,0.0503698885,-0.153701514]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/482","title":"Bugs : dataset.map() is frozen on ELI5","comments":"This comes from an overflow in pyarrow's array.\r\nIt is stuck inside the loop that reduces the batch size to avoid the overflow.\r\nI'll take a look","body":"Hi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?","comment_length":27,"text":"Bugs : dataset.map() is frozen on ELI5\nHi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?\nThis comes from an overflow in pyarrow's array.\r\nIt is stuck inside the loop that reduces the batch size to avoid the overflow.\r\nI'll take a look","embeddings":[0.0050863987,-0.2545411885,-0.0975884423,0.1058631837,0.1502091438,-0.1698086411,0.4342986345,0.1180540547,-0.2856822014,0.2499660552,-0.1675368547,0.4118994176,0.1468531787,0.1485168487,-0.0885317475,0.1992187798,-0.1245994642,-0.0692450553,-0.1046517417,-0.0854028761,0.0994062051,0.1942095608,-0.2547451556,0.0946521685,-0.1753012687,0.1913972795,-0.0475018173,0.0138271702,-0.1394347847,-0.3484272361,0.1026879475,-0.001340957,-0.0601915382,0.0998518914,-0.0001287397,0.066696398,0.0852747262,0.0151902912,-0.0321877785,0.1148892269,0.3419377804,-0.0924857259,0.2941782475,-0.2267786264,0.2994029224,0.2010616064,0.3933571279,0.2928218246,0.5463498235,-0.1131911725,0.0042867605,0.2285585999,0.0229935329,0.513486743,0.3447038531,-0.1565420777,-0.2626554668,0.0360518061,0.4549467862,-0.2447688878,0.0202714056,-0.1735136062,-0.1445726901,-0.1309605837,0.1657719761,0.3925587535,0.0644609034,-0.62484622,0.0050570308,0.1813290417,0.3918077946,-0.219531998,-0.0531340241,-0.2484401911,0.0206694249,-0.3290591836,0.4384991527,-0.0801222995,-0.1662214547,0.0215378106,-0.2199908644,-0.0387936346,-0.0763660818,0.3075369895,-0.0092367847,0.0896168426,0.1044290662,0.139065057,0.0363292657,0.0282410532,-0.1358946413,-0.182715103,0.0109648211,0.3332873583,-0.5986675024,-0.163416326,-0.0271216985,-0.069117099,0.2069696337,-0.0307871327,0.1467565149,0.1431549042,0.1338284761,0.0912330374,0.4436792433,0.0077233384,0.160488829,0.0440581217,0.2785817385,-0.1372031122,-0.3210589886,0.1931507587,-0.0700885803,-0.2468800843,0.0818447024,-0.2111313492,0.0528816432,-0.2320253253,0.202260375,0.1252010316,-0.6203747392,0.0176222157,0.0182809364,0.2910580337,-0.0792117566,-0.0218280535,-0.0857140869,0.0450759456,-0.4482832849,-0.1256741732,0.0942524001,-0.387989521,-0.1894799024,-0.0087158941,0.2495262027,-0.2611215413,0.1432356536,-0.1522934586,0.322812438,0.1161507741,0.162984699,-0.0682689473,0.2398326546,0.0830465108,0.1794318259,0.1534902602,-0.2191044092,-0.0566177852,-0.0779641569,0.4114333093,0.0235470198,-0.0727850944,0.0412019044,-0.0273830257,-0.2359686941,-0.0660005808,-0.0888220817,0.1308397055,0.4844260514,-0.3778196275,0.0910024941,-0.2624390721,-0.0215132516,-0.0950281173,0.338176012,0.194452703,-0.1940383017,-0.0274670385,-0.026571244,-0.0410611741,-0.1195073947,0.4236699045,0.0039515602,0.1033091024,-0.0318871029,0.2316198349,0.3955104053,-0.2058109194,-0.7310912609,0.0396428034,-0.2763487399,-0.1362841278,-0.252252996,0.0629158169,0.2470750809,-0.0630355403,0.1072485298,0.4468324184,-0.2058178186,0.0618838295,-0.4313601255,-0.2683362067,-0.0633401647,0.086264208,-0.1417798549,-0.34730497,-0.1305567175,0.1829256564,-0.0755519345,0.0349516943,0.2099795938,0.2865862846,0.2029321343,0.0456465259,-0.0300458632,-0.1439756453,-0.4601614177,0.0268316828,-0.0586411394,0.2615787387,-0.1202884167,0.0648181513,0.2349448502,0.0113875018,-0.212957561,0.0243550446,-0.0730482116,-0.2284717709,0.0837188587,0.2112513632,-0.0019162963,-0.2065973282,-0.0059790402,0.1431186646,-0.4388467073,0.2543537915,-0.1280645281,-0.4770114422,-0.2971019447,-0.0658227503,-0.388165921,-0.1865554303,-0.1196711808,0.1069658697,-0.349888742,-0.1667898148,-0.160900712,0.4271230102,0.2303902209,-0.3507650793,0.1343254447,-0.2931654155,-0.0293530375,-0.0842273384,-0.0780694559,0.228428036,-0.1594274044,0.1833505183,0.0703863725,-0.0162780341,-0.0938535333,0.0467910692,-0.1287907511,0.1275016516,0.1001781374,0.315977484,0.3469285965,-0.0476956703,-0.0876378044,0.3985711634,0.8972319961,0.1478902996,0.0494407117,0.1409237832,-0.0217285715,-0.1202460155,-0.3746708333,-0.2797940075,0.2434234321,0.0051102475,0.025267208,-0.0428303815,-0.0932139233,-0.0164670441,0.1810831726,0.2945491076,0.0239809863,0.1551367044,0.4825344086,0.0677755848,-0.1131497025,-0.0568579584,0.1012780666,0.1829892248,-0.1776940972,0.1054924577,0.0243339818,-0.117035985,0.1892231405,-0.0143280104,-0.308470428,-0.3321069181,0.2306244224,0.1722477823,-0.0961345211,0.2171561718,0.3602740467,-0.0538765267,0.1465309411,-0.2379711419,-0.01161825,-0.3989604414,-0.4639269114,-0.0646316707,0.5152745247,-0.0921800733,-0.152976796,0.1284042448,-0.4225045741,0.0034484377,-0.5699742436,0.2885662913,-0.2367332727,0.3893535733,-0.0867004767,0.2780524194,-0.4414614141,-0.1928228289,0.274579972,-0.3448044956,-0.3394559026,0.1445184648,0.0448548347,0.0084365793,-0.2066278756,-0.5001220107,-0.0840337425,-0.1841686964,0.3770566285,-0.074672468,0.0755597278,0.0847931802,0.0967242941,0.343975395,-0.0327094458,-0.3106962442,0.0395304039,-0.1168307737,0.3748587668,-0.208447367,-0.3211794198,-0.1357550621,0.130845502,-0.1553771496,0.0252420902,-0.4053165317,-0.2954768538,-0.3072522283,0.5448829532,-0.1725354791,0.038863685,0.3364681005,0.2721450031,0.109076798,0.0870999172,-0.2881274819,0.3796558678,0.6361579895,0.0603710562,-0.114497602,0.5326623917,0.1268235296,0.8650525212,0.0501326509,-0.0962904021,0.1846015006,0.0508210883,-0.0524213836,-0.4184018672,0.1097523496,0.1593179554,-0.0444966443,-0.0510262772,0.2817857265,-0.0252391864,-0.272364825,-0.0099888584,0.1290397346,-0.2740392387,-0.3053347468,0.2833794951,-0.1501799822,-0.184742853,0.2023890316,-0.1284491122,-0.2675516307,-0.3865621388,-0.2657364309,-0.2626971602,0.3266680837,-0.031001715,-0.5268408656,-0.1693628579,-0.4538227022,0.3003254831,0.2017198056,0.1829954088,-0.3149499297,-0.0796466097,0.2595863938,-0.3560302258,0.5623479486,-0.2501103282,-0.0077964161,0.0664591715,0.1660588831,-0.3364859521,-0.1052750275,-0.1529739201,0.1791736186,0.565525353,-0.2064905763,-0.3848067522,-0.1788005531,0.38433218,0.0949750617,-0.1083512828,0.0949029401,-0.3024912477,-0.1956835985,-0.3143540621,-0.0694737211,0.3912868202,-0.0088346973,-0.1087371334,0.008465399,0.2198887616,0.1827401221,-0.0794844106,0.2370210588,-0.1183428168,-0.1753094196,0.3294097781,0.2528044283,-0.1059926972,-0.1877209395,0.2943144739,-0.0627307072,-0.2948416173,0.4838995636,-0.153640151,0.3558014631,0.3829171956,0.2371374965,-0.0817955583,-0.0177955087,0.1487634033,0.3896481395,-0.0745605305,0.4932511747,0.3155474961,-0.0177192837,-0.162680462,0.1486384869,-0.1034399271,0.0189362951,0.4798079133,-0.1455781311,-0.0643191338,0.6060297489,-0.0395094603,0.9872861505,-0.4438498318,0.3768614531,0.2483164817,0.1166119128,0.3669240475,-0.2621617913,0.286393404,-0.3973955512,-0.2996115088,0.239746362,-0.2731795609,-0.2487073243,0.3163546026,-0.114467375,0.0510349162,-0.0501104072,0.2295383215,0.0140193207,0.2362154126,0.1502758563,-0.1015114635,-0.4429033697,-0.0289850179,0.0388889126,0.0606170744,-0.0724192113,-0.1615415961,-0.2318305373,-0.131878823,-0.2155957222,0.217111513,-0.1324204654,0.0900543705,0.3183254004,0.0041818125,0.2054664046,0.6948333979,0.0254576001,-0.0103642382,0.0692700297,0.1160050631,0.0554092526,0.0455110148,0.3793322444,0.0686988682,0.2638886273,-0.1732295901,-0.2370496392,0.1212744266,0.1692429483,-0.245124042,-0.1821863204,0.0905484855,0.2008384764,-0.1988248229,0.0886796638,-0.0004925982,-0.5018952489,0.0365957692,-0.0460097529,-0.0520533584,-0.0365064479,0.3843833804,-0.0595831089,-0.1122862473,0.0794157311,0.6717551351,-0.1674194783,-0.1439260244,0.7501493096,0.2646194696,-0.0452823155,-0.0137438672,0.1758431345,-0.2018995732,-0.5593947172,0.0608635806,-0.3067263067,-0.110931769,0.148719728,0.2183097303,-0.2155748457,-0.224221468,-0.2134022564,-0.3311450183,-0.3069344163,0.2494446784,0.0512155034,-0.042965699,0.0193407331,0.3812385201,0.1494000107,-0.0937482491,-0.109399803,-0.0954764262,-0.2970127761,0.246008113,0.0471738987,-0.4202589989,0.3659621477,-0.1423820853,-0.0453598164,0.2743968666,-0.22311984,-0.0985880569,0.1066613123,0.2159394324,-0.081830278,-0.0772705227,-0.2459409684,0.0146403257,-0.0089586889,-0.2189126611,0.2781186998,-0.2800689042,-0.0316737518,-0.0328374431,-0.0027318953,0.2612271905,0.0019770598,0.0846836194,-0.0318552405,0.0304130651,0.3323302567,0.3044151366,0.0692910105,-0.0151135083,-0.2690472603,-0.0072411881,0.100159362,0.0877484307,0.2753633559,-0.4385765493,-0.1707716435,-0.2107210159,-0.1101357192,0.592623353,-0.1895660907,-0.3875776827,0.2923759818,0.0179400332,0.0572986901,0.1216674373,0.5638595819,0.047848817,-0.0091147237,-0.0947179198,-0.0207892936,-0.0896747559,0.0290757231,0.2730593979,0.3346316814,-0.5037521124,-0.0086339852,0.3587886393,-0.0602080934,0.0653520897,0.0825372115,-0.0608053319,0.178490296,0.3039407134,0.1708104163,-0.2196980864,-0.5023559928,0.450533092,0.2428024262,-0.2482268512,-0.3020808101,0.5325145721,0.3196109831,0.5286595821,0.284661442,0.3522031009,0.1508636624,-0.2730666399,-0.2414687127,0.3165407777,-0.0552545264,-0.0754422545,-0.0673507452,-0.052594278,0.0400403701,0.3657518923,-0.2580023706,0.0849262103,0.4335435033,0.4367712736,0.0632361844,-0.2560559511,-0.1521411687,-0.073275879,0.3150064051,-0.0367239825,0.3979531527,-0.0304544177,-0.1338930875,-0.3677637875,0.2811619043,0.2063678801,0.3232463896,-0.560747087,0.4288626909,0.0544943884,0.0278191306,-0.0197821856,0.1850552261,0.2880257666,0.0756831765,0.0303501803,-0.0751667172,0.0033643651,-0.2747893631,0.187106207,0.3222306073,0.1376528591,-0.2230949253,0.2568222284,-0.3423522413,0.1055420861,-0.1402105689,-0.4819663465,0.2016893476,-0.0350150131,0.0202044118,0.0368413776,-0.0110754417,-0.0054087634,0.0955958962,0.6279258728,0.3679856062,0.2386387885,-0.0641040951,0.0885594562,-0.4655943811,0.2648676932,-0.0297681279,0.1389427483,0.3166323304,0.3369347453,0.1583118439,0.3255509734,0.2213921249,-0.2028590739,-0.4033801556,0.5688691735,-0.5948470235,-0.1591483802,0.0863332748,0.1069520563,-0.1165293157,-0.2735643387,0.0781192929,-0.2249492407,-0.1005941257,-0.213075608,-0.0973997191,-0.0582082942,0.1613886207,0.2476559281,0.2895669937,0.1646666825,0.0070549496,-0.1617268026,-0.1248115748,-0.1267134249,-0.0711519644,0.0957815498,0.0410696305,0.0500342511,-0.254912138,0.2613507807,-0.2735891044,0.0168298949,-0.0059973109,0.2123406827,0.0228472091,-0.3081659377,0.4737376273,0.2139340341,-0.0859781578,0.5660753846,-0.0658093765,-0.0802732036,-0.6256957054,0.1396764219,0.2787373066,-0.409179002,-0.1944171339,0.1413305104,0.1041905954,0.2522991896,-0.047465127,-0.4054484963,-0.0626422539,0.1113284379,-0.1385063976,-0.6580840349,0.1909585595,0.3545120656,0.13917014,-0.2182251215,0.0990579054,-0.0152473394,-0.1090777516,-0.1203650013,-0.2823943198]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/482","title":"Bugs : dataset.map() is frozen on ELI5","comments":"I created a PR to fix the issue.\r\nIt was due to an overflow check that handled badly an empty list.\r\n\r\nYou can try the changes by using \r\n```\r\n!pip install git+https:\/\/github.com\/huggingface\/nlp.git@fix-bad-type-in-overflow-check\r\n```\r\n\r\nAlso I noticed that the first 1000 examples have an empty list in the `title_urls` field. The feature type inference in `.map` will consider it `null` because of that, and it will crash when it encounter the next example with a `title_urls` that is not empty.\r\n\r\nTherefore to fix that, what you can do for now is increase the writer batch size so that the feature inference will take into account at least one example with a non-empty `title_urls`:\r\n\r\n```python\r\n# default batch size is 1_000 and it's not enough for feature type inference because of empty lists\r\nvalid_dataset = valid_dataset.map(make_input_target, writer_batch_size=3_000) \r\n```\r\n\r\nI was able to run the frozen cell with these changes.","body":"Hi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?","comment_length":147,"text":"Bugs : dataset.map() is frozen on ELI5\nHi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?\nI created a PR to fix the issue.\r\nIt was due to an overflow check that handled badly an empty list.\r\n\r\nYou can try the changes by using \r\n```\r\n!pip install git+https:\/\/github.com\/huggingface\/nlp.git@fix-bad-type-in-overflow-check\r\n```\r\n\r\nAlso I noticed that the first 1000 examples have an empty list in the `title_urls` field. The feature type inference in `.map` will consider it `null` because of that, and it will crash when it encounter the next example with a `title_urls` that is not empty.\r\n\r\nTherefore to fix that, what you can do for now is increase the writer batch size so that the feature inference will take into account at least one example with a non-empty `title_urls`:\r\n\r\n```python\r\n# default batch size is 1_000 and it's not enough for feature type inference because of empty lists\r\nvalid_dataset = valid_dataset.map(make_input_target, writer_batch_size=3_000) \r\n```\r\n\r\nI was able to run the frozen cell with these changes.","embeddings":[0.1289257258,-0.3091129959,-0.06825611,0.1572482884,0.2135083973,-0.1734551787,0.3533759713,0.1509891152,-0.1582410634,0.1141942814,-0.2389454395,0.4564895332,0.108321473,0.1284253299,-0.0192696843,0.1580439359,-0.1359721124,0.0216666386,-0.138399452,-0.121157378,0.128272742,0.3092678487,-0.2475365996,0.0375251397,-0.1744519174,0.2080098689,-0.1016796157,0.0072080214,-0.1468411386,-0.3405945599,0.1364496052,0.0454997532,-0.1865528375,0.1471100897,-0.0001257587,0.0547152981,0.0463493541,-0.0335171595,-0.0842954516,-0.0720032007,0.3044876158,-0.1794094592,0.3073302507,-0.1446628273,0.127319634,0.3064395189,0.3526678681,0.2476817071,0.6475158334,-0.1010444462,0.0140077453,0.341314137,0.0202773102,0.5116378069,0.243788138,-0.0982789919,-0.3499835432,0.0335915461,0.3750883937,-0.1832286865,0.1367816925,-0.0608269759,-0.0739217028,-0.0898051485,0.19255732,0.4039262831,-0.0741314888,-0.6528962255,-0.0021793693,0.2572253644,0.2794762552,-0.1696970314,-0.1200202331,-0.3368326128,0.0148051614,-0.4079931676,0.4944007397,-0.0794702917,-0.1530280858,0.0317088887,-0.1824898273,-0.0652211085,-0.0135252317,0.2793029249,0.0590097718,0.1251696646,-0.0725423694,0.0867164731,0.0587756559,0.0040247319,-0.379796654,-0.181068778,-0.0119346613,0.3528054059,-0.5798324943,-0.2715927958,-0.0327341259,0.0594736859,0.1253583431,0.0762948468,0.1430134922,0.1827675253,0.0188215673,0.041459322,0.4148799777,0.0017276356,0.2859444022,0.0397364311,0.2681476772,-0.0332963541,-0.2699636221,0.1644416004,-0.0023881095,-0.143230781,-0.1291372329,-0.2096811086,0.1443166882,-0.3338890374,0.2238596082,0.1400536895,-0.5295664072,-0.0235963576,0.0451130867,0.348033309,-0.1645060629,-0.0118470388,-0.0154613648,0.028652804,-0.4626110792,-0.0920969024,0.0960532278,-0.400587678,-0.1462285817,0.0067687454,0.2733539939,-0.3486697674,0.2713479698,-0.1475230157,0.2598335445,0.1013623103,0.1670903414,-0.0870442986,0.2051347792,0.0524544194,0.1002826616,0.1181941032,-0.133488521,-0.1894541234,-0.0926409513,0.2097845674,0.1490687281,0.015464914,0.0886512548,-0.0408379659,-0.2455912977,-0.0535609424,-0.0409294963,0.2703981996,0.3567351997,-0.3351740539,0.0605804026,-0.1100196838,-0.16547966,0.0336645097,0.3799240887,0.2518183589,0.0483144373,-0.241093874,0.0075256964,-0.0359109528,-0.2282191813,0.4138222039,0.0667631179,-0.0232566744,0.0048591988,0.1988445371,0.4201336801,-0.2138728797,-0.5843521357,0.1227919087,-0.290155828,-0.0611700825,-0.3185210228,0.0572850741,0.1865839958,-0.1132518649,0.108619906,0.3744200468,-0.152241379,0.0520576946,-0.4770565927,-0.2232019603,-0.0265080687,0.0327686444,-0.1929167211,-0.2336608022,-0.1907966137,0.2977527082,-0.0531564094,0.0488934442,0.1820814759,0.3569732606,0.1773028821,0.1362027526,-0.1015440002,-0.1310838908,-0.5059298277,0.0301138759,-0.082680501,0.436884135,-0.0742785484,0.0415672958,0.216215238,-0.0862872154,-0.2092489004,-0.0907132924,-0.0478548557,-0.1511384547,0.0789053962,0.2883453667,-0.0765572265,-0.146270141,-0.0684965029,0.2294391543,-0.5454393029,0.2303234339,-0.1101198122,-0.3616029024,-0.2829392552,0.0522339493,-0.3480919898,-0.1953464448,-0.1199693307,0.0457643829,-0.4057110548,-0.1182210371,-0.0787373632,0.39961344,0.3000273108,-0.2792613208,0.0287999511,-0.1769136488,0.0589298122,-0.0994167253,-0.1693899333,0.3000504673,-0.1911527514,0.1421568394,0.0937402993,0.0221831705,-0.0751492158,0.0131150745,-0.1991497427,0.0291856732,0.2271129638,0.3052901626,0.3777759671,0.0039162342,-0.2030637562,0.2866193652,0.8647702932,0.2128255963,0.0915230587,0.1407784224,0.0420252047,-0.1311241686,-0.3347011805,-0.2349749207,0.2353041619,-0.0368026346,0.0103973439,0.0471230038,-0.0723187029,0.0022590877,0.2268696427,0.2180667669,-0.0744219571,0.1118378416,0.4056487083,0.1075129434,-0.1748203337,0.0449038595,0.0736119226,0.1916543245,-0.2182614058,0.0064427429,-0.0735023543,-0.1446316391,0.1790862083,0.0557615347,-0.4943912029,-0.3087976277,0.1794397235,0.086846754,-0.0174654014,0.2599861324,0.3050743043,-0.0461216941,0.0266661495,-0.1626860201,-0.0823711902,-0.371319592,-0.3768123388,-0.0614233613,0.4895643294,0.0645242706,-0.108421877,0.0187370367,-0.5041586757,-0.0094818696,-0.6358896494,0.3481926322,-0.3971453011,0.4128521681,-0.0781828463,0.2758353055,-0.4864445031,-0.0928508863,0.347694844,-0.3571275473,-0.3248539567,0.0642937124,0.084577702,0.0274519268,-0.2390641421,-0.5075418949,-0.0715254396,-0.200479269,0.6052663922,0.0293335542,0.0404657349,0.2019646466,-0.0366263017,0.4214391708,-0.1367387027,-0.1317187399,-0.028817758,-0.1558273286,0.3360876739,-0.1849846542,-0.3257573545,-0.1265583932,0.1665827483,-0.1240958273,-0.1564632952,-0.3962708116,-0.408788681,-0.3515097499,0.419516623,-0.0856776983,0.0416094027,0.3636344075,0.2145564407,0.0893893614,0.0769142732,-0.3594811559,0.4156855643,0.5649589896,0.0811531171,-0.0618204251,0.5869220495,0.1258977056,0.76325351,0.164718166,-0.0777330026,0.2583013773,0.0622731037,0.0986941308,-0.3876313865,-0.0207066331,0.1824162602,0.0286726989,0.1055803671,0.2620311975,0.049495589,-0.1202005446,-0.1215321273,0.1444302648,-0.2589597106,-0.3341295719,0.1795340031,-0.1171993241,-0.1604577601,0.2035465837,-0.0725462735,-0.212299794,-0.4061678648,-0.2326294035,-0.0760389268,0.34263587,-0.1049862653,-0.6462792158,-0.0349220596,-0.5429958701,0.2988605201,0.1454876512,0.1586165577,-0.3372978866,-0.0721188858,0.2224736065,-0.2879295051,0.4114738107,-0.1915258914,-0.0516271181,0.0511150062,-0.0398447253,-0.284080416,0.1139924377,-0.1935252398,0.1324242651,0.6303717494,-0.0946355984,-0.5235413313,-0.3121537268,0.3164809644,0.0571205467,-0.0688864142,0.1559969634,-0.3152288795,-0.1021759957,-0.4176949263,-0.0969286785,0.3561074138,-0.0781287774,-0.0542628728,0.1125816852,0.1985895187,0.1470995843,0.0097382627,0.2579273283,-0.1269590706,-0.0956372991,0.3598088324,0.2257257998,-0.0573016517,-0.3894451261,0.4335392416,-0.1299568266,-0.3766605556,0.3981061876,-0.1200799197,0.2812707424,0.469620496,0.2516357899,0.0401980542,-0.0422499441,0.2157252431,0.3636834919,-0.0476307832,0.519500494,0.3840844631,0.0357618965,-0.1599829644,0.1930097789,-0.1579409689,0.0557067133,0.4740781188,-0.0046681841,0.0101704961,0.4886056483,-0.0979200006,1.0059657097,-0.3302904665,0.4059495926,0.2713979185,0.0539783277,0.5820281506,-0.1612088233,0.2256225646,-0.418875277,-0.3026839197,0.2059723288,-0.144808203,-0.1869802624,0.2278217375,-0.1691356003,0.1528988779,0.0752880424,0.1367150694,-0.0368947536,0.2807765603,0.0383461863,-0.197774753,-0.6501989961,0.0040083537,-0.0306060575,0.2507350147,-0.0437817946,-0.2037801594,-0.2344392091,-0.1280349195,-0.1990260631,0.2368147075,-0.1423741281,0.035892047,0.3933475018,-0.0302773248,0.2490167916,0.7023185492,0.1513835043,0.0587398671,-0.0325748771,0.1326153874,-0.1110471115,0.0730695054,0.3623226881,0.092264995,0.2442237586,-0.1902023554,-0.1516869962,0.0013156292,0.106145151,-0.213843748,-0.1322256476,0.0358188301,0.0391894318,-0.1144882739,-0.051801987,0.1280816495,-0.4593143761,0.0096015781,-0.0241174735,-0.0343447961,-0.0522572063,0.1823451519,0.1109322682,-0.1212472245,0.0807991102,0.6947203279,-0.367164284,-0.189338401,0.7166757584,0.2581754327,-0.0477392077,-0.0592699684,0.1794999093,-0.2438335121,-0.6025191545,0.078659676,-0.1984284073,-0.2134572119,0.1613490283,0.3437258005,-0.1560999006,-0.1773862988,-0.0807451606,-0.2184283435,-0.311740309,0.2542218566,-0.0011970504,-0.0305785947,-0.0276501756,0.3899829984,0.2204924524,-0.0599082075,-0.1270696819,-0.0305316858,-0.4184255004,0.2184303701,0.210466817,-0.5042176843,0.4190069735,-0.1174020991,-0.0263556875,0.2935333252,-0.2947068214,-0.0972054303,0.0434844866,0.199643746,-0.0416945331,-0.0970214233,-0.2538351119,0.092587024,-0.065358974,-0.293756783,0.2045025378,-0.2466244102,0.0194893833,0.0361423977,-0.0797324181,0.3411549032,0.0247637238,0.0372031368,-0.0388082415,-0.0269216597,0.3201968968,0.1649168879,-0.0070517068,0.0222638454,-0.2711969614,0.0372970626,0.074558489,0.0469816774,0.2867676914,-0.328499645,-0.1481306851,-0.1287609786,-0.06387306,0.5036643147,-0.0933879465,-0.2509553134,0.2977611423,0.0191158596,-0.0197152663,0.1968261749,0.5825093389,0.1026283056,0.0625448525,-0.0672754496,-0.01541162,-0.0110469246,0.0916270763,0.2495031655,0.2788556218,-0.6175307035,0.0525375493,0.2572178245,-0.0248939432,0.0649422705,0.0557320192,-0.0471888892,0.1620463878,0.1522075832,0.1121230125,-0.1864190549,-0.565767765,0.5901626945,0.3083994985,-0.2395474613,-0.2778965831,0.4732435346,0.3027635217,0.5327272415,0.2283580154,0.4041005373,-0.0285833217,-0.3256970048,-0.2354796231,0.4408734441,-0.0501162447,-0.1448832005,-0.0015083717,-0.0594526418,-0.0572994463,0.2851233482,-0.2409449518,0.0650757998,0.331847012,0.4306073487,0.0960694402,-0.3528822064,-0.2499443889,-0.0605903864,0.4490511119,0.0333409719,0.4102720022,-0.0495201685,-0.0978206918,-0.3445423841,0.3547847569,0.1537969857,0.2586193681,-0.4484901726,0.442822665,0.0477019995,0.0321000628,0.046962712,0.2841471434,0.2792511582,0.0861321688,0.0426205844,-0.0812563673,-0.02646612,-0.2729158998,0.0337475799,0.2761287093,0.0765255764,-0.0008550167,0.1916518807,-0.2293901592,-0.0552538782,-0.1104443669,-0.5082321167,0.2749866247,-0.1005485281,0.0665364191,0.033957541,-0.0932467952,0.0029775314,0.1635857671,0.4926814735,0.2839574814,0.1227544025,0.0100098373,0.0288747046,-0.6039531231,0.1588700712,0.1284501404,0.1115101948,0.2243480235,0.2920475304,0.2114553899,0.3512560129,0.0846398026,-0.2181586027,-0.2167148441,0.3210610747,-0.5578278899,-0.075772956,0.0069502792,0.0788913444,-0.0357011817,-0.1501208097,-0.0647814125,-0.0872539952,-0.135329321,-0.2084426433,-0.0388614126,-0.2230318785,0.1890411824,0.0953545347,0.3319296837,0.0939929783,-0.0453674868,-0.3141005635,-0.0758104548,-0.1758027971,-0.0934366807,0.1746332943,0.0095013175,0.1047799811,-0.2335823476,0.2436619401,-0.3363449574,0.028993072,-0.0339803696,0.2507981658,-0.0553374961,-0.2667056024,0.4273792207,0.2657974362,-0.0795620382,0.6648865938,-0.0860898495,-0.088924028,-0.5790915489,0.1271741986,0.3910949528,-0.4176441431,-0.0960395262,0.1227006242,0.1189786419,0.3112986982,-0.1813847572,-0.5510279536,-0.1144214272,0.0826354697,-0.2009817064,-0.5723954439,0.1652618349,0.265208602,0.095553644,-0.2220436335,0.0918626785,0.0356683098,-0.0469325334,0.0217827111,-0.2249054015]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/482","title":"Bugs : dataset.map() is frozen on ELI5","comments":"@lhoestq Perfect and thank you very much!!\r\nClose the issue.","body":"Hi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?","comment_length":10,"text":"Bugs : dataset.map() is frozen on ELI5\nHi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?\n@lhoestq Perfect and thank you very much!!\r\nClose the issue.","embeddings":[0.0647895113,-0.173011899,-0.0817397609,0.1211161911,0.1627188325,-0.0771627277,0.4821651578,0.1855669022,-0.2774398327,0.1364147961,-0.2553304136,0.4861959517,0.1580599546,0.1096802354,-0.0706778765,0.1690307111,-0.1220334619,0.0595788062,-0.1237451807,-0.1030779257,0.1141718775,0.3070580065,-0.3172127008,0.1394386888,-0.2726353705,0.2050769478,-0.1123085767,-0.039906662,-0.1454892457,-0.3151308,0.1348776966,0.0785015821,-0.0801061317,0.1616897881,-0.0001256356,0.0496525243,0.0536882542,-0.0137582021,-0.0559958667,-0.0137462504,0.2480821311,0.0092047323,0.3296349049,-0.2114895284,0.0999002382,0.3304212093,0.3992601633,0.1511498839,0.5939770937,-0.0785380006,0.0311110169,0.2395109683,-0.0647640377,0.4713525772,0.210368678,-0.123488836,-0.3593181074,-0.0376917683,0.3329018652,-0.1528518349,0.1615100652,-0.1082911938,-0.044905521,-0.0412390009,0.1687927395,0.45301947,-0.0056887516,-0.7097072005,0.1015597284,0.2047515661,0.5010082126,-0.1742959321,-0.0517076552,-0.1665729135,-0.0166490451,-0.3319564462,0.4755107164,-0.0695050433,-0.1038495004,0.0606414713,-0.1105995923,-0.0676183403,-0.0533171408,0.2591084242,0.030245252,0.1018640622,0.003690256,0.127839312,-0.0220134836,0.006130931,-0.225243926,-0.1831675619,-0.0317406952,0.3219645321,-0.632621944,-0.2711752653,-0.1669471264,-0.0238344856,0.0112676844,0.0372054353,0.1489889324,0.2178275287,0.0589177348,0.0882150084,0.4190204442,-0.0617587119,0.3736900091,0.1421436071,0.3016980588,-0.1421146095,-0.4343596697,0.1785981953,-0.0290746689,-0.209880963,-0.0166979916,-0.1732291579,0.1165097803,-0.3335776031,0.2571482956,0.0439050235,-0.6268817782,-0.0282512829,-0.0206615906,0.3144974411,-0.1129823253,-0.0328187458,-0.0258015003,0.0420748033,-0.42752105,-0.1421096027,0.1008487195,-0.4003097713,-0.152134791,0.0298774205,0.2806836367,-0.3173550665,0.1933730096,-0.1846406162,0.201773867,0.1231736913,0.2553052604,-0.0696561113,0.1761265099,0.0431733318,0.1199890599,0.1690486819,-0.2359258085,-0.1752225906,-0.0744362548,0.2776333988,0.1189088523,0.0229654275,0.0979841202,-0.0434363671,-0.1479084641,-0.1084014922,-0.0188479125,0.2368407547,0.369995445,-0.3624402285,0.0361361057,-0.2284469903,-0.1808694303,-0.0208540149,0.4125130177,0.2094541788,0.0303877499,-0.1720772833,-0.0203779154,-0.0888805538,-0.2421760112,0.3590633869,0.0616175607,-0.0099705951,0.0457806997,0.2785947621,0.4723248184,-0.2181180865,-0.606289804,0.1080849692,-0.2854952812,-0.1977238804,-0.3837628365,0.1058895811,0.2444388717,-0.0908616781,0.1327079535,0.4148559868,-0.1081599146,0.0688263178,-0.4001692235,-0.240368098,-0.0179042593,0.06686984,-0.1434009671,-0.2344407141,-0.1162204593,0.3027274609,-0.0760185644,0.0830142424,0.219532147,0.2547824681,0.2192598879,0.043040812,-0.0743466616,-0.1187845469,-0.5218832493,0.0506925695,0.0360550694,0.3404132724,-0.0434801318,0.0364340842,0.2350542992,-0.0917787999,-0.2366800159,-0.111007221,-0.0501383208,-0.2062427253,0.031626597,0.2744505703,-0.0604036152,-0.1600596756,0.0358162038,0.1693468392,-0.4103100598,0.1659270227,-0.1335074306,-0.3778078258,-0.2480151057,-0.0534937158,-0.3758623004,-0.1997511089,-0.1556916386,0.0605525598,-0.3317590952,-0.1207706481,-0.0331278741,0.4199013412,0.2798830867,-0.3444800675,0.0499310717,-0.2141993344,0.0504247285,-0.1267780662,-0.1444129497,0.1809860468,-0.1761683822,0.1320520937,0.0679671541,0.0004616249,-0.1102860942,0.0452118628,-0.1405006796,0.0244867895,0.0633718967,0.2958435118,0.3389877081,0.0319924131,-0.152564317,0.331420064,0.8619858623,0.2192951292,0.0638285726,0.1063724086,0.0172010772,-0.0975246876,-0.376627475,-0.2491333187,0.2449072301,-0.0903790593,0.0601984337,0.0819583237,-0.068051599,-0.0011818585,0.2013930827,0.2562544048,0.036816828,0.0669005066,0.3896027803,0.1103613451,-0.0564303547,-0.0276436247,0.1348282695,0.1927434057,-0.1702748686,0.01850296,-0.1360882521,-0.0169784762,0.1592863202,0.0748676062,-0.3196864724,-0.2844397426,0.1848942786,0.0812684074,0.0308196526,0.2818038166,0.2689598799,-0.1023148,0.0445499755,-0.10513179,-0.0254669003,-0.4365401566,-0.4058618844,-0.0432668962,0.4654131532,0.0913573056,-0.1379401982,-0.0319873802,-0.4085533023,-0.0535789058,-0.587557435,0.3380306363,-0.3336392939,0.3934470117,-0.1703498662,0.2605761588,-0.4770998657,-0.1013378873,0.333307445,-0.3909343779,-0.3081052601,0.0676885918,0.0706788152,-0.0285960361,-0.1521244645,-0.643114388,-0.1454400718,-0.1130352244,0.5127945542,-0.0230632648,0.0439634435,0.100990586,0.0203469023,0.366615057,-0.0211732481,-0.1715557277,0.0296993684,-0.1672464758,0.3338733912,-0.2140241116,-0.3161227405,-0.0709592029,0.2362394333,-0.1419963986,-0.1034279764,-0.3461093009,-0.2796738148,-0.308873266,0.4435450733,-0.0818647593,-0.045088511,0.3253461123,0.266726464,0.1014784873,0.1028714254,-0.3977921307,0.4441005886,0.6768741608,0.0631201491,-0.0157427173,0.5888246298,0.0965670273,0.8148408532,0.1171244979,-0.175482288,0.1841166019,0.0623632781,0.0060605998,-0.4500698745,0.052101817,0.1957686096,-0.0568853915,0.0473348424,0.2628878057,0.0092566796,-0.2587121725,-0.1614920199,0.1249829456,-0.292612344,-0.3404255211,0.1165014952,-0.1965701878,-0.1414628178,0.2032937258,-0.0997273028,-0.1749408245,-0.3566929698,-0.222650826,-0.1064785346,0.3043692708,-0.0557888113,-0.6453237534,-0.0892901272,-0.4531890154,0.2834005654,0.1368109435,0.2320193946,-0.3722704649,-0.0356581062,0.2764821649,-0.3001386225,0.3358543813,-0.2948171496,-0.0237749591,0.1376668811,0.0065499172,-0.2241128683,-0.0104611516,-0.1335494369,0.189456448,0.581515193,-0.1898037195,-0.3735813797,-0.3179166317,0.2875760198,0.0656929538,-0.0608229749,0.1156893373,-0.303109169,-0.0445196405,-0.4108330905,-0.15025042,0.326638639,-0.1697849631,-0.0374012962,0.0869700611,0.1986177415,0.1630946845,-0.0235505067,0.303881675,-0.19169572,-0.1042421013,0.4330860674,0.3149941564,-0.0817289054,-0.331582725,0.3272157907,-0.1122588366,-0.3391591609,0.4549048543,-0.218903169,0.3318097889,0.3542166054,0.2309877872,-0.0543758534,-0.0785610527,0.1127687246,0.3962699473,-0.0525329821,0.6168530583,0.3538946211,-0.054971464,-0.2043630183,0.1401194036,-0.1361789703,-0.0155432029,0.45514974,-0.1913293153,0.0508523807,0.5189101696,-0.0562699884,0.8905590773,-0.3646228015,0.350518316,0.1276928335,0.0369834229,0.4381889105,-0.2318872511,0.2485722601,-0.4810245037,-0.3056742847,0.2131951749,-0.1938583851,-0.2380657047,0.2701554,-0.2047816366,0.0509282798,0.0059076035,0.2016426176,-0.0299681649,0.2732048631,0.1549254209,-0.1661948264,-0.423995167,0.0276977774,-0.0202025734,0.2359690815,-0.0524771251,-0.1644947976,-0.2971839607,-0.0521979816,-0.1803970039,0.2786610425,-0.091106616,0.1482724547,0.2813367248,0.0018735073,0.249832049,0.7739331722,0.1720790267,-0.0073527554,-0.0054469886,0.1788099706,-0.0699318051,0.1068213284,0.3839206696,0.0487877131,0.2639487386,-0.1892240196,-0.248266533,0.0687965155,0.1605066657,-0.2335122973,-0.1946115196,-0.0026980925,0.1169682741,-0.0688996464,-0.028853232,0.1105969101,-0.5054994822,0.0321163796,-0.0027888427,-0.0743766278,0.0614819303,0.3265758157,0.0634981766,-0.1325893402,0.1632101834,0.7412158251,-0.3187795281,-0.1244408488,0.7333645225,0.1989643723,-0.0412968844,-0.1006274894,0.2538057864,-0.2954460382,-0.4708253145,0.0251792409,-0.2031374276,-0.1593414694,0.2190265208,0.2876329124,-0.2373984456,-0.2778618336,-0.1006580442,-0.2375116199,-0.3071902096,0.1617861986,-0.0685218424,-0.0035278404,-0.0093416777,0.385712862,0.2148113698,-0.0260419566,-0.1231606379,0.0261315014,-0.3172776699,0.2509204447,0.1912160069,-0.4269502163,0.3378434479,-0.1263154298,-0.0447890423,0.286195755,-0.2474250197,-0.1002230719,0.0583098568,0.2054062635,-0.0341771729,-0.0331456512,-0.2703037858,0.0402327292,-0.1269408464,-0.2487495244,0.2444673181,-0.2166634947,-0.0177593082,0.0652064458,-0.0907161161,0.3224015832,-0.0082563832,0.0682642162,-0.0399585068,-0.0218016077,0.3183631003,0.1429695338,0.0065496247,0.0027199695,-0.3114817441,0.0087505393,0.0012783558,0.0464311056,0.289644599,-0.3544057012,-0.0238438677,-0.1889787912,-0.1702939272,0.4972061515,-0.1020302102,-0.279586792,0.2727624774,0.031239545,0.0473909266,0.1871631891,0.6086966991,0.0533728339,-0.0057866266,-0.1576520652,-0.0102780415,-0.0175590795,0.068779394,0.2266520113,0.3267390132,-0.6635729671,0.0494537018,0.2751318514,-0.0182143822,0.1104681045,0.048488494,-0.0284509696,0.1382744759,0.2329019308,0.0957745537,-0.2375262976,-0.474160552,0.5363413095,0.3546724916,-0.2007630765,-0.2777655125,0.4372714162,0.2957822978,0.4722868204,0.2749831975,0.3902372122,-0.0140074128,-0.2845619619,-0.2800025642,0.3880082071,-0.0328187086,-0.1322212517,0.0138723347,-0.0937428474,-0.0541224703,0.301384896,-0.3461699486,0.1539036334,0.3050162494,0.4902631938,0.0701044947,-0.2755191028,-0.2115294933,-0.113106668,0.408308655,0.0620917827,0.3270112276,-0.1692481339,-0.0725143403,-0.3060619235,0.2772125006,0.1341760904,0.2838143408,-0.5813289285,0.4366323352,0.035039939,0.032458026,0.1908602118,0.2264077812,0.2808856368,-0.0535594262,0.070926182,-0.0768623576,0.0080942353,-0.3019630015,0.0858262479,0.3608510196,0.1052091494,-0.10578271,0.2339499593,-0.2540466785,0.0493936092,-0.0505580045,-0.4578291476,0.2440789789,-0.1505814344,-0.013630555,0.0055315048,-0.1130852401,0.0043372884,0.1483773887,0.5443822742,0.1709697843,0.1173828915,0.0396515913,0.049514696,-0.6042340398,0.1744638681,0.1763097048,0.1684488952,0.215056628,0.2396904677,0.2768250704,0.3330728412,0.1439622343,-0.1883303672,-0.2445592582,0.4069647789,-0.5611205101,-0.0134857828,-0.0002525445,0.1678295135,-0.0847159401,-0.1811950803,-0.0357256196,-0.1168138385,-0.1134594679,-0.2840046287,-0.1568840146,-0.0880801454,0.1268437058,0.2297258824,0.2466938496,0.1413016617,-0.0139254825,-0.2078951895,-0.0933402628,-0.1555504948,-0.0369180404,0.0596179739,0.0051927017,0.0271138065,-0.1540622562,0.3392396569,-0.335292697,-0.006432598,-0.1097192541,0.3243733644,-0.0500965491,-0.3320064545,0.4791527689,0.26813519,-0.0775370374,0.6273575425,-0.1061454117,-0.1247078776,-0.626046598,0.1551803946,0.3220805824,-0.3160372078,-0.162028864,0.0788756385,0.0975084826,0.2674430609,0.0345104747,-0.4515981674,-0.0553846732,0.1076415777,-0.2335265428,-0.5928457975,0.1637630463,0.4404097795,0.1512483656,-0.2065153122,0.1067060083,-0.0526169688,0.0278435666,-0.1637637466,-0.2105637491]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/482","title":"Bugs : dataset.map() is frozen on ELI5","comments":"@lhoestq mapping the function `make_input_target` was passed by your fixing.\r\n\r\nHowever, there is another error in the final step of `valid_dataset.map(convert_to_features, batched=True)`\r\n\r\n`ArrowInvalid: Could not convert Thepiratebay.vg with type str: converting to null type`\r\n(The [same colab notebook above with new error message](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing#scrollTo=5sRrJ3_C8rLt))\r\n\r\nDo you have some ideas? (I am really sorry I could not debug it by myself since I never used `pyarrow` before) \r\nNote that `train_dataset.map(convert_to_features, batched=True)` can be run successfully even though train_dataset is 27x bigger than `valid_dataset` so I believe the problem lies in some field of `valid_dataset` again .","body":"Hi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?","comment_length":94,"text":"Bugs : dataset.map() is frozen on ELI5\nHi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?\n@lhoestq mapping the function `make_input_target` was passed by your fixing.\r\n\r\nHowever, there is another error in the final step of `valid_dataset.map(convert_to_features, batched=True)`\r\n\r\n`ArrowInvalid: Could not convert Thepiratebay.vg with type str: converting to null type`\r\n(The [same colab notebook above with new error message](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing#scrollTo=5sRrJ3_C8rLt))\r\n\r\nDo you have some ideas? (I am really sorry I could not debug it by myself since I never used `pyarrow` before) \r\nNote that `train_dataset.map(convert_to_features, batched=True)` can be run successfully even though train_dataset is 27x bigger than `valid_dataset` so I believe the problem lies in some field of `valid_dataset` again .","embeddings":[0.0539374612,-0.1870064586,-0.074278377,0.0692923591,0.1881322116,-0.0918401927,0.4278537929,0.2430600524,-0.387001276,0.1208235621,-0.2542549968,0.472434938,0.1759656668,0.2033965439,-0.1053897738,0.1071930751,-0.1148367003,0.0579198562,-0.095841758,-0.0491202138,0.1031022519,0.3184986413,-0.2580778003,0.1650636345,-0.310503304,0.1500938833,-0.0909850746,-0.063239105,-0.2213093042,-0.3299462497,0.0668836012,-0.0400810428,-0.0191349313,0.1387688369,-0.0001242538,0.0448895171,0.0489350632,-0.0396052748,-0.1172182783,0.0515075326,0.2815468907,-0.092722103,0.3024719059,-0.1993808448,0.121528171,0.1908759773,0.3395656347,0.1250128746,0.610242784,0.0073208311,0.055975467,0.2696267664,-0.033320576,0.4009646177,0.2310598642,-0.1057361588,-0.4094343781,-0.0279104002,0.3057385683,-0.073764801,0.0937425643,-0.0345544443,-0.0047439155,-0.0890352428,0.2041344047,0.447301507,0.0543917529,-0.7332022786,0.0954589695,0.1459839493,0.4292942584,-0.163677901,-0.0888444856,-0.1996159106,0.0285850577,-0.4379118681,0.433860302,-0.0437152497,-0.1576816589,0.0147970412,-0.0896566436,-0.0634034351,-0.0821782574,0.1596626043,0.0451035388,0.082153365,-0.0101994351,0.0811636001,-0.000775388,-0.0306908581,-0.2821880579,-0.1224424839,-0.0369157791,0.3129152358,-0.6525633335,-0.2602997124,-0.2111930996,-0.0127820289,0.0090761315,0.0486088358,0.1738576442,0.1454139352,0.0982822105,0.037865784,0.3953722715,-0.0234734677,0.2744033635,0.1286569089,0.2145038545,-0.1013746262,-0.3396109343,0.1972129792,-0.0063052555,-0.2773442566,-0.0041833417,-0.1296566874,0.213401258,-0.3412664831,0.2097338438,0.0644885823,-0.6048839688,0.0543595366,-0.0302813929,0.290712148,-0.0818567723,-0.0012985746,0.0003661499,0.1126073301,-0.4264134765,-0.1001875922,0.0938718542,-0.3189900815,-0.1659364551,0.0253956765,0.2953777015,-0.3251473308,0.1266686767,-0.1675279588,0.241194278,0.0676400959,0.2005026042,-0.0495285466,0.1837721467,0.0253388211,0.0428988747,0.1298454702,-0.2254213095,-0.1423178613,-0.0521086678,0.3205448985,0.1021712124,0.0004592855,0.0579670258,-0.0064169588,-0.115836516,-0.1142978668,-0.0920281857,0.2776235044,0.3569951952,-0.3189867139,0.005636408,-0.3019961715,-0.097862497,0.0077759926,0.4229875803,0.2221725732,-0.0393861979,-0.1806021333,0.009934105,-0.0670612156,-0.1639227271,0.3881370723,0.0671417788,-0.068432875,0.0487903021,0.2698665857,0.5018539429,-0.2360262871,-0.5605533719,0.0842080191,-0.3151651323,-0.2124067694,-0.3700900674,0.0882243887,0.2761770785,-0.1064587682,0.0501356497,0.3990454078,-0.1892985553,0.1165638417,-0.4404366612,-0.2546896636,0.0071925665,0.0387986004,-0.0335728563,-0.2508604825,-0.145751074,0.3366514742,-0.1375248879,0.0508824363,0.2727854848,0.2217576206,0.1935016662,-0.0702337101,-0.01252052,-0.0654193461,-0.5470225811,0.0346957445,-0.0950001851,0.4290871024,-0.0072943452,0.0913429484,0.1463080198,-0.0574684739,-0.2194213569,-0.1102265045,0.001585059,-0.2296857834,-0.0181350112,0.1705188304,-0.0965174288,-0.1747611016,0.0136870602,0.2303231955,-0.4198413789,0.2105484158,-0.2102124691,-0.4316301942,-0.2027973533,-0.1207083538,-0.2658187449,-0.136696592,-0.2471276373,0.0317470096,-0.2937158048,-0.0242926031,-0.0219215434,0.4465959072,0.3535358012,-0.4185455739,0.0960636437,-0.2039043456,0.0543001778,-0.0944244862,-0.1779820919,0.193772018,-0.2655295432,0.2218671441,0.0622492656,0.0845488533,-0.0982313752,0.0846856982,-0.2007941753,0.0403160304,0.0268419627,0.2826346755,0.2617367208,-0.030728396,-0.1640805304,0.3843754232,0.7712831497,0.2200623602,0.0041627479,0.1855040193,0.0684932172,-0.1383883208,-0.2919831574,-0.2560809851,0.1829895675,-0.0792281553,0.0277565308,0.0691761002,-0.111623086,0.0367188156,0.2313179225,0.2731529176,-0.0039716396,0.0878642797,0.3924007416,0.1269481331,-0.0596859753,-0.0752699748,0.1465491652,0.1641421467,-0.2193275094,0.0125429546,-0.1379264891,-0.0650568902,0.0902582482,0.0596543662,-0.3185710311,-0.3496057093,0.222376436,0.1179311275,-0.0026539993,0.3269678652,0.1722249091,-0.0225334708,0.0529188626,-0.1150834188,-0.0832120776,-0.3923435509,-0.4076896906,-0.0095232241,0.4277973175,0.1665219367,-0.0966349766,-0.0519466847,-0.2906079888,-0.1139912382,-0.6097022891,0.3551804125,-0.3738603294,0.4278289974,-0.1368179321,0.2860459089,-0.4543642402,-0.0607181378,0.3598079383,-0.319421947,-0.2685777545,0.0863336995,-0.0198582541,-0.0580552779,-0.156555444,-0.6044725776,-0.1301190853,-0.1716372073,0.3883109093,0.0377664454,0.0577210151,0.1623965949,0.1000991017,0.4368853867,0.0836393908,-0.1327850074,0.0724737868,-0.0684207231,0.3780962229,-0.2309399247,-0.3176396489,-0.099513635,0.2158452421,-0.1500914246,-0.1090742052,-0.3428517878,-0.3040424883,-0.3614273071,0.3931966722,-0.0801544338,-0.0459433161,0.3395408094,0.2874933481,0.0575974956,0.1020124033,-0.4317061007,0.457695365,0.7315322757,0.0766995922,-0.0183507688,0.5675571561,0.099723421,0.881677568,0.1115860194,-0.1668649614,0.2578700781,0.0367083512,0.0419228077,-0.483291924,0.0303248689,0.2113536298,-0.0642779842,0.0273593813,0.1935285628,-0.0152920876,-0.1945049316,-0.1208990589,0.1814959198,-0.228867352,-0.3178036213,0.1604900211,-0.2234326601,-0.1628768593,0.1666614413,-0.0731886923,-0.1899709255,-0.3545717597,-0.1877993941,-0.0863399655,0.3585194945,-0.0291259978,-0.6207333207,-0.1096363962,-0.4323072731,0.3310400844,0.1397511512,0.2593171895,-0.3684085906,-0.0728357434,0.2251539975,-0.2911626697,0.2268965244,-0.2127636671,-0.0970881954,0.1606950462,0.0468074977,-0.2511030138,-0.0750726908,-0.143516019,0.1912960559,0.4949720204,-0.1217662394,-0.3668164313,-0.2991461456,0.3354327977,0.1143845022,-0.0547587499,0.2381698787,-0.2625123262,-0.0675505921,-0.4722187519,-0.0732920021,0.4041702151,-0.1168118119,-0.0586255342,0.0688989609,0.1878005564,0.1327370405,0.006150519,0.3224442005,-0.0995450839,-0.0768676698,0.4044603705,0.2092096508,-0.0673904419,-0.3526376188,0.3883754611,-0.0610332489,-0.4078305662,0.499401927,-0.14459306,0.3027717769,0.4482409954,0.1930813342,-0.140414387,-0.0068502864,0.1819932759,0.3776974976,0.0128546329,0.5262821913,0.2901239097,-0.0197927728,-0.1219247729,0.1650863737,-0.1874382347,0.0183325876,0.4783314764,-0.0286264867,0.0383423567,0.5677272081,0.0381432176,0.9758557081,-0.2699212432,0.274384737,0.1987369359,0.0093007991,0.4218445122,-0.1443619579,0.2698315382,-0.5266176462,-0.2899628878,0.2433763891,-0.2169338465,-0.2555153072,0.1828205585,-0.2217987478,0.0740425885,0.0077109351,0.1622265875,-0.0115707265,0.2984126508,0.1152763218,-0.2819029987,-0.4721390605,0.0487281904,-0.1049386188,0.2490475476,-0.0030684969,-0.1782489866,-0.3353842795,-0.0918077379,-0.2292079329,0.2933659852,-0.1197624356,0.104752712,0.2387638986,-0.0602765717,0.1600346565,0.6947133541,0.1500712037,-0.0886845887,0.0165481083,0.1643887609,-0.0444903821,0.0688967034,0.4028805494,0.037327759,0.2904087901,-0.2089942098,-0.2103515565,0.0769350231,0.1517149061,-0.1790326238,-0.1565603763,-0.1059655845,0.0931830332,-0.0638352484,0.0012963371,0.0256593004,-0.5122880936,0.0354584232,0.0154430103,-0.1162627637,0.0667452663,0.2826854885,0.1746970713,-0.1020618975,0.1229657084,0.6355888247,-0.3503260016,-0.112467058,0.7513549328,0.2271392941,-0.0329823978,-0.1053858846,0.2558537126,-0.3011709452,-0.5768549442,0.0415050499,-0.3038912416,-0.2306888849,0.1752122939,0.3404957652,-0.1392192245,-0.2399087548,-0.0772819519,-0.2000790834,-0.3080741763,0.2129828334,-0.0899992436,0.0486358814,-0.1046149731,0.3674404323,0.18637003,-0.0412438661,-0.1533664316,0.012668848,-0.3003509939,0.279807508,0.175885573,-0.4567356706,0.2779487371,-0.0599133857,-0.0336163864,0.3118651211,-0.2465885133,-0.1095213071,0.0179146212,0.1941848844,-0.0406498201,-0.0990184918,-0.3305047452,0.0980903134,-0.1401507258,-0.2644595802,0.2234963924,-0.2495934218,-0.0252179839,0.1089908332,-0.1077735052,0.2229652852,0.0980560333,0.024152657,0.0800200477,0.0913212299,0.3303565383,0.2263109237,0.0073866076,-0.0227951035,-0.3700707257,0.0525562167,0.0209172741,0.0719647855,0.2483473569,-0.3310933709,-0.0387911722,-0.178529501,-0.1730979681,0.5225913525,-0.1272381544,-0.2101666927,0.2399602979,0.0557606705,0.0225931779,0.1883437932,0.5434411764,-0.0400637873,-0.0050351247,-0.089243643,0.0082850382,0.0305793453,0.1435264349,0.2447312027,0.3117624223,-0.6137400866,0.0980035365,0.217349261,-0.06617973,0.1176310182,0.0674415827,-0.0732716918,0.1785868406,0.1344177574,-0.0379489958,-0.2305824012,-0.4712120295,0.5559195876,0.4206781983,-0.1522450596,-0.3123300672,0.4449594021,0.2855593264,0.4789653122,0.3186174333,0.3901514411,-0.0537828021,-0.357695967,-0.3769514561,0.3257186711,-0.099756442,-0.0764376968,-0.0174459349,-0.1103936657,-0.1150767133,0.2794765234,-0.3362813592,0.1715204269,0.3045938313,0.5075873733,0.0787336081,-0.2455798686,-0.237263605,-0.070352383,0.4208482206,0.038633924,0.4022470415,-0.1057444885,-0.1016260982,-0.3066695929,0.3072878718,0.1976912022,0.2935115397,-0.6400924325,0.3875577748,0.1321139336,-0.0059111528,0.2638387382,0.2280207872,0.2538414896,-0.0595619008,0.1545855701,-0.0592816062,-0.0248896759,-0.3027148545,-0.0009209187,0.2783112526,0.0832827836,-0.0609323792,0.2151129544,-0.2164380997,0.0376051329,-0.1263158619,-0.3947073221,0.1769414544,-0.1372810453,-0.0508239605,-0.0380623564,-0.0798351914,0.0224105865,0.1355582476,0.5751467347,0.1674951315,0.1027368978,-0.0704302415,0.1621203423,-0.6096419096,0.2065858394,0.2451083958,0.1028089896,0.2863226831,0.3193433881,0.2587864101,0.4257708788,0.2457889915,-0.2119677812,-0.1558288187,0.4385854006,-0.5146776438,-0.0335086845,0.0834737048,0.1615607888,-0.0992897972,-0.2035086155,-0.0254346505,-0.1245819703,-0.0927737877,-0.2628906667,-0.2807004154,-0.0632415563,0.1414949447,0.227792576,0.2979299426,0.1224624738,-0.0314636454,-0.2810340524,-0.1568282098,-0.2138567269,-0.0536394306,0.0513221882,-0.0312870666,0.124536559,-0.0799799487,0.360101521,-0.4278178513,-0.0401998237,-0.0839043856,0.2810070217,-0.0773704126,-0.2369511425,0.3989081383,0.2445235997,-0.0396410972,0.5937414169,-0.0521083288,-0.0306180362,-0.675906539,0.0881917179,0.3299511671,-0.3246211708,-0.1571730077,0.1668061614,0.0793770254,0.3383032978,0.0851813778,-0.4567795694,-0.0500467047,0.0678004101,-0.2592575848,-0.5594134331,0.1471493691,0.4032057822,0.168986842,-0.2204862833,0.1316612065,-0.0463543274,0.0522674657,-0.2600908577,-0.218925342]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/482","title":"Bugs : dataset.map() is frozen on ELI5","comments":"I got this issue too and fixed it by specifying `writer_batch_size=3_000` in `.map`.\r\nThis is because Arrow didn't expect `Thepiratebay.vg` in `title_urls `, as all previous examples have empty lists in `title_urls `","body":"Hi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?","comment_length":33,"text":"Bugs : dataset.map() is frozen on ELI5\nHi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?\nI got this issue too and fixed it by specifying `writer_batch_size=3_000` in `.map`.\r\nThis is because Arrow didn't expect `Thepiratebay.vg` in `title_urls `, as all previous examples have empty lists in `title_urls `","embeddings":[0.0882880241,-0.0822167769,-0.0899109319,0.1087930799,0.1185982525,-0.0889242217,0.4757874906,0.2081893831,-0.2753664255,0.1408063024,-0.2147414088,0.4715832472,0.1646788418,0.0616819225,-0.1162421107,0.1537863612,-0.158514291,0.0361106656,-0.0435180888,-0.0942484066,0.0881526694,0.3064975441,-0.3244808018,0.1371067315,-0.2413132936,0.2019753009,-0.0675988644,-0.049233485,-0.1338675022,-0.3987390995,0.0902898759,0.0110534541,-0.0539940186,0.1521504819,-0.0001268104,0.0641141385,0.1259067506,-0.0447385274,-0.0788739249,0.0518748946,0.2480291873,-0.0301098023,0.3839221597,-0.2232055962,0.1997665316,0.2042776346,0.3820057213,0.0985115841,0.5644937158,-0.0939926431,0.0145405876,0.2681625783,-0.0877727866,0.4349065721,0.2906715572,-0.1680425107,-0.3519189656,-0.0400066823,0.3742506206,-0.0923342779,0.0832195282,-0.0479501262,-0.0200902931,-0.0459484421,0.1264987141,0.477640003,0.0331442691,-0.7292939425,0.1101480797,0.2397381663,0.4590658844,-0.1601786017,-0.0912276506,-0.1838967204,0.0088336263,-0.3579076231,0.4505927861,-0.0299419817,-0.0901348293,0.0185388252,-0.0743217617,-0.040401265,-0.0883394331,0.2276088297,0.1097409278,0.1179815084,0.020053992,0.1539659053,-0.0590441562,0.0156816021,-0.1864434332,-0.2057789415,-0.1096618548,0.3239534199,-0.6241595745,-0.2234710604,-0.1515565813,-0.0737281516,0.0315627158,0.0660682321,0.2326868922,0.2130044103,0.1060308442,0.0576225072,0.4097149968,-0.05941578,0.3594403565,0.152526021,0.2901147604,-0.0899540484,-0.3181607127,0.1908805221,-0.0522587821,-0.2366308719,0.0019143455,-0.1777714938,0.1045408696,-0.2911323011,0.2693873048,0.0756654963,-0.6122701764,-0.0127222324,-0.035284657,0.3106817305,-0.1387453377,-0.019065652,-0.0420422032,0.0441930778,-0.3719218969,-0.05985808,0.0867926851,-0.4458230138,-0.1268520951,0.0292038359,0.2845354378,-0.2773237526,0.2217059582,-0.2348583341,0.2051868737,0.0649385452,0.2884622514,-0.0199454855,0.189987272,0.0648311228,0.1371371001,0.204111293,-0.279910177,-0.1149778813,-0.0912585706,0.3285954893,0.0530043356,-0.0008449478,0.0881713405,-0.0615696721,-0.1474289894,-0.0972121879,-0.0326791704,0.1581484973,0.3567199707,-0.299053371,0.0931835771,-0.1606671363,-0.074257426,-0.0157865789,0.3695381284,0.2457429469,-0.011288641,-0.1803428829,-0.0644403547,-0.0077457121,-0.2145560086,0.3933416605,0.017316198,-0.051456891,0.0583467707,0.2661234438,0.4861608148,-0.220908463,-0.5594331622,0.1241096482,-0.2801663876,-0.2539252639,-0.327547282,0.0745495856,0.2965097725,-0.1045871228,0.0722666606,0.3634411097,-0.0934674516,0.0332766175,-0.4079682231,-0.3176830113,-0.073690787,0.0167573653,-0.1066107079,-0.334757477,-0.146824494,0.215567708,-0.0763266757,0.0471528545,0.2931105196,0.3203018606,0.1911579967,-0.0128341764,0.0090577397,-0.0042152465,-0.5717056394,0.0901390538,-0.0780278146,0.3007609248,-0.0624999627,0.0468623526,0.2347545475,-0.0451570153,-0.2232609391,-0.1361989379,-0.0694333762,-0.2399135232,-0.0109200589,0.2301578224,-0.0736903101,-0.3139733076,-0.008088002,0.1840165406,-0.3748850524,0.1614419222,-0.1260195673,-0.3616601527,-0.2581923306,-0.0989926234,-0.3698384464,-0.1656944454,-0.1613461524,0.0632629097,-0.3258742392,-0.0153220026,0.0269875415,0.416449964,0.2937585711,-0.4194709957,0.1238848865,-0.1380817145,0.097223781,-0.0955071077,-0.1869752407,0.1891407818,-0.2533555031,0.2792419791,0.0949694365,0.0388385057,-0.0884369239,0.0600587912,-0.1594147831,-0.0119249616,0.0588181466,0.3035000861,0.3837769628,0.000923311,-0.0879838765,0.3538539708,0.8066471815,0.2237540185,0.031028809,0.1539036632,-0.0020598087,-0.1163451374,-0.3227665424,-0.2487761676,0.1705665141,-0.0870793313,0.0229926798,-0.0028441863,-0.0616028048,-0.0714263096,0.2523103654,0.3050374389,0.0414583199,0.071284987,0.3709096313,0.1204286441,-0.0578712933,-0.1169271842,0.1672040671,0.1258959174,-0.1920571178,0.0119043533,-0.0655374378,0.0032772161,0.1812362224,0.0310357623,-0.3219733238,-0.2996349931,0.2012772709,0.0419376157,-0.0665793046,0.2993482351,0.1596488506,-0.0770488828,0.0552628525,-0.1453842968,-0.0574788824,-0.490311712,-0.3879302442,-0.0461065769,0.4349363744,0.069507964,-0.154833734,-0.0031807069,-0.3482359648,-0.0557775386,-0.6003726721,0.345295608,-0.3194232583,0.38280496,-0.1645485461,0.2790251076,-0.4631020129,-0.0367032327,0.310987711,-0.3287284672,-0.2486947328,0.065076448,0.0176021606,-0.0721572936,-0.184745416,-0.5815076828,-0.0476904474,-0.1539790779,0.4527025223,-0.0041602803,0.0596409813,0.1243086755,0.0747015327,0.4305662811,0.0298446659,-0.1916427165,0.1210582182,-0.1054960713,0.400839746,-0.2193355262,-0.3028165698,-0.0781511813,0.1628539413,-0.1272924691,-0.0984951407,-0.3735532761,-0.1985849291,-0.3790394366,0.3669737279,-0.0718740225,-0.0167547651,0.379386276,0.3064507544,0.0722278506,0.0580987036,-0.3891451955,0.5050684214,0.675767839,0.0092982249,-0.0089655314,0.5026328564,0.070911184,0.9094951749,0.1167577133,-0.0880283639,0.2245072275,0.0675015375,0.0284250975,-0.3903768361,0.1601770818,0.1316403598,-0.0401244275,0.0388072915,0.2364289761,0.0380493291,-0.2484052181,-0.1227400377,0.1951144934,-0.2875275612,-0.3168263137,0.1847686917,-0.2551832795,-0.168825224,0.1740369499,-0.0765980482,-0.1780084819,-0.375181973,-0.1998485327,-0.1906105131,0.3779657781,-0.1055293158,-0.6005120873,-0.0463043191,-0.4620148242,0.3174316585,0.1200261638,0.2238008976,-0.2955452204,-0.0564065985,0.2446452975,-0.3228532672,0.2590301037,-0.2851579487,-0.0492566526,0.1612728238,0.0685211495,-0.3726627529,-0.0046681575,-0.1773078591,0.1747203618,0.4599044025,-0.2281535566,-0.3975753486,-0.3105400801,0.2359529287,0.0375094078,-0.0658997744,0.1733648032,-0.2594775856,-0.0183648188,-0.4175831676,-0.126186505,0.3429453969,-0.1650981605,-0.0610799752,0.0112269111,0.1740145683,0.1327437907,-0.059905652,0.2988522947,-0.1146555543,-0.1276051849,0.4424629211,0.3040943444,-0.061884474,-0.3205316663,0.3552519083,-0.0954778045,-0.3660571277,0.4882130623,-0.2189038992,0.3159546554,0.4747321308,0.2133302987,-0.0357875302,-0.0234066248,0.1447728723,0.4186307192,-0.0603995956,0.5418787599,0.2795414031,-0.0161217116,-0.1286270022,0.1725663841,-0.1708380431,-0.035518609,0.6318179369,-0.1137912944,0.0038764495,0.5671449304,0.0636702999,0.9373657703,-0.2821104228,0.3562494516,0.1291278303,-0.0394516326,0.4074738026,-0.1294268221,0.2650200427,-0.4918097556,-0.2174855024,0.219911322,-0.2149684876,-0.264709264,0.2747201324,-0.1567385197,0.0506980009,-0.0032421562,0.1992225349,-0.0261162389,0.292262435,0.1135441214,-0.2582213283,-0.3926380575,-0.010196914,-0.0051297317,0.1710567772,-0.0618745573,-0.170561105,-0.3496136665,-0.0965967178,-0.1793381721,0.2796680033,-0.1814926863,0.1445914507,0.2455635667,-0.0341222025,0.1703039855,0.71786201,0.0609153509,-0.0505996384,-0.0307883229,0.254899472,-0.0124060865,0.0887362882,0.3964911103,0.0344577916,0.2211810052,-0.1801509112,-0.239880681,0.0477488823,0.1024102122,-0.2337890416,-0.1609705091,0.0478168167,0.1398224384,0.0227074511,-0.0623249859,0.0740206018,-0.5765266418,0.0539221317,-0.0186329745,-0.0547051914,0.0997099802,0.2908628583,0.1346251965,-0.170737043,0.1136572137,0.6775932908,-0.3244546652,-0.1086257994,0.7458800077,0.1795631945,0.0014097891,-0.0949165672,0.2476831079,-0.2478708774,-0.6283718348,0.0337576829,-0.2483889759,-0.2120059878,0.2343789339,0.3080371916,-0.1999984533,-0.1975630075,-0.1469415277,-0.2412347198,-0.2423961759,0.2498526871,-0.0583459958,0.0068158698,-0.1725541055,0.2741122842,0.2016077936,-0.0881668255,-0.1053355113,0.0764524117,-0.2692142427,0.2743497491,0.1036965996,-0.3691770434,0.4151694477,-0.1232331395,-0.0757552087,0.2767160237,-0.2741017342,-0.0830468461,-0.016344659,0.2227385193,-0.0710496083,-0.1054349393,-0.3280333281,0.0408844054,-0.142818898,-0.2872326672,0.2128674984,-0.2900471389,-0.044814799,0.0042102532,-0.0880877301,0.2888958156,-0.0618680194,0.1017446741,-0.0217916407,0.076853469,0.3140352368,0.1966611296,0.0544643216,-0.0076148058,-0.2682597339,0.0105351964,-0.0505149104,0.0935515389,0.2986558378,-0.3337076604,-0.0411009192,-0.1884239167,-0.1477077901,0.595375061,-0.0976812765,-0.3353036642,0.2371452302,0.0260588471,0.082323879,0.2001847029,0.520804286,0.0638967305,0.0121354591,-0.1359037608,-0.0063772821,-0.0324636027,0.1020547599,0.3046663702,0.2812542319,-0.6254974604,0.081424728,0.2558406591,-0.0214992501,0.1413773298,0.0449637882,-0.0664561912,0.2502112389,0.1468690187,0.0969790965,-0.2232927233,-0.3549111485,0.5749953985,0.3272977769,-0.2109616697,-0.2955943644,0.3983950317,0.2951464653,0.5204439163,0.3410260975,0.4236570597,0.0308460891,-0.2934636772,-0.3252934813,0.2823751271,-0.0197914857,-0.1260680407,-0.0459985472,-0.1114296168,-0.0530716516,0.2995021343,-0.3030313551,0.16529347,0.3224520385,0.4959098995,0.0424022861,-0.3634294271,-0.1948479414,-0.0675644428,0.4407273829,0.0173916109,0.3731265068,-0.1021297649,-0.1070425063,-0.3444982767,0.2899489105,0.1152360514,0.2705461383,-0.5767050982,0.4279710352,0.1324668527,0.0330879018,0.1897371858,0.206623733,0.3099295199,-0.1102646068,0.0974732712,-0.0849167183,-0.0020610869,-0.2866070271,0.0705495253,0.3329071701,0.1678402126,-0.1049033031,0.2067901641,-0.2799986303,0.0762466863,-0.0978531837,-0.4134726524,0.2258626819,-0.1182277575,-0.111310944,-0.0520900637,-0.132493481,-0.0013125126,0.1276620775,0.6129342318,0.1930652559,0.1246221736,-0.0267079044,0.0825605094,-0.5853529572,0.1925190836,0.1387237906,0.2084907442,0.279799968,0.313542366,0.2333824784,0.3741554916,0.2446382493,-0.177049607,-0.246221602,0.4191214144,-0.6102465987,0.0101738414,0.0743525326,0.1879317611,-0.1064255387,-0.2250774503,-0.0230039191,-0.1562495083,-0.1390533745,-0.2743950486,-0.2682373524,-0.0231824964,0.1495163143,0.3450475633,0.2363648564,0.0682358965,-0.0300127082,-0.2721800208,-0.1444950849,-0.2129716575,0.0134778488,0.0994295627,0.0685308799,0.0575660691,-0.1969742775,0.2913994193,-0.3510700762,-0.0179674923,-0.1465967447,0.3190267384,-0.0336717218,-0.3093468249,0.4059657454,0.2705375552,-0.0823320225,0.5921954513,-0.1156926006,-0.0927955359,-0.6072288752,0.1054434776,0.3132190108,-0.3774883151,-0.1397368163,0.1671625525,0.1499504596,0.3385232091,0.0498351008,-0.4270492196,-0.0275150016,0.1106718928,-0.2297035456,-0.6133560538,0.06107568,0.4153716564,0.1881698221,-0.2335268259,0.1861012727,-0.0459268205,-0.0301697627,-0.2157278806,-0.2437679023]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/482","title":"Bugs : dataset.map() is frozen on ELI5","comments":"I am clear now . Thank so much again Quentin!","body":"Hi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?","comment_length":10,"text":"Bugs : dataset.map() is frozen on ELI5\nHi Huggingface Team!\r\n\r\nThank you guys once again for this amazing repo.\r\n\r\nI have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https:\/\/github.com\/patil-suraj\/exploring-T5\/blob\/master\/T5_on_TPU.ipynb) \r\n\r\nHowever, when I run `dataset.map()` on ELI5 to prepare `input_text, target_text`, `dataset.map` is **frozen** in the first hundreds examples. On the contrary, this works totally fine on SQUAD (80,000 examples). Both `nlp` version 0.3.0 and 0.4.0 cause frozen process . Also try various `pyarrow` versions from 0.16.0 \/ 0.17.0 \/ 1.0.0 also have the same frozen process.\r\n\r\nReproducible code can be found on [this colab notebook ](https:\/\/colab.research.google.com\/drive\/14wttOTv3ky74B_c0kv5WrbgQjCF2fYQk?usp=sharing), where I also show that the same mapping function works fine on SQUAD, so the problem is likely due to ELI5 somehow.\r\n\r\n----------------------------------------\r\n**More Info :** instead of `map`, if I run `for` loop and apply function by myself, there's no error and can finish within 10 seconds. However, `nlp dataset` is immutable (I couldn't manually assign a new key-value to `dataset `object)\r\n\r\nI also notice that SQUAD texts are quite clean while ELI5 texts contain many special characters, not sure if this is the cause ?\nI am clear now . Thank so much again Quentin!","embeddings":[0.0927398801,-0.1724505574,-0.0837821886,0.126543358,0.1572223157,-0.0878264904,0.4809217155,0.1667148769,-0.3014018834,0.1273972392,-0.2538481951,0.4689527154,0.1571187675,0.1127332076,-0.0675934926,0.1715164334,-0.1096805185,0.0569481887,-0.1190284491,-0.1204550415,0.1157760844,0.3011617661,-0.3016949892,0.1331679374,-0.2765979171,0.1960762292,-0.1445055008,-0.0179514103,-0.1540339887,-0.3054856956,0.1335908175,0.0858429894,-0.068041712,0.169829309,-0.0001254676,0.0396989733,0.0423562154,-0.0217832737,-0.0542984344,-0.0088435635,0.2520165145,-0.0001790701,0.3377073407,-0.2234581113,0.1119849384,0.330824554,0.3864018023,0.1551940888,0.6030445099,-0.0808215216,0.0321681164,0.2367797345,-0.0693953037,0.4634512663,0.2230401635,-0.1207901388,-0.3545206189,-0.0415746607,0.35208565,-0.1585686654,0.1522245705,-0.1058277935,-0.0331678092,-0.0502621792,0.1973086149,0.4603486657,-0.0073906155,-0.7290485501,0.1032778248,0.2099193782,0.5209403634,-0.1727609187,-0.069342427,-0.1723633558,-0.0218306836,-0.3329451084,0.4987781048,-0.0743261352,-0.1084862351,0.0464429855,-0.0966849253,-0.0696154162,-0.0658011809,0.2492281497,0.0089089526,0.0976632088,0.0200863425,0.1107941866,-0.023439277,-0.0029966417,-0.2222809792,-0.1780501604,-0.0237921327,0.3443883955,-0.6150419116,-0.2848104835,-0.1505582631,-0.0326450951,-0.0032508308,0.0481887385,0.1383505613,0.2145933956,0.0427623317,0.0685819909,0.4313432574,-0.0673467964,0.3751855493,0.1481904238,0.3016126752,-0.161643371,-0.4337692559,0.1705387086,-0.0393143371,-0.1945792139,-0.0240551177,-0.1761869788,0.1111069992,-0.3554839194,0.2675681412,0.0423038453,-0.6287733912,-0.0361227244,-0.0259100441,0.3057716489,-0.1041754484,-0.0325499326,-0.0261283964,0.0434504636,-0.4372896254,-0.137221992,0.0996928588,-0.3960222304,-0.1561185271,0.0293634012,0.2888244689,-0.2990849912,0.1840995103,-0.1912102848,0.2119461,0.135945946,0.2368682623,-0.0708237141,0.1714931875,0.0339738652,0.1100289673,0.1735947132,-0.2418480068,-0.1891831607,-0.0728380233,0.287409395,0.1354237497,0.026927948,0.0864135772,-0.0403154716,-0.1738557667,-0.1172936261,0.0068861847,0.2312331796,0.3661238849,-0.3538580239,0.0258290656,-0.2211257517,-0.152589798,-0.0248941667,0.4054943323,0.2065108418,0.0147409253,-0.1704816669,-0.014816355,-0.0843681917,-0.2420021594,0.3547145426,0.0694318935,-0.0217103492,0.0548819117,0.2808822989,0.4739400744,-0.2307098657,-0.6012734771,0.0950043052,-0.2642265558,-0.1853082776,-0.3754788637,0.0947054103,0.2535015643,-0.1064757258,0.1233710498,0.4234822989,-0.1081763506,0.0567909479,-0.4152727425,-0.2325610071,-0.0170735586,0.0648910031,-0.1183186918,-0.2380985767,-0.1232725903,0.2835352719,-0.0865383148,0.0752072856,0.2152938992,0.241863057,0.2169068605,0.0623394176,-0.0628758222,-0.1013099924,-0.5085815191,0.0394191407,0.028634673,0.3676029146,-0.0471097641,0.0439477004,0.2469002157,-0.1047189534,-0.2225979418,-0.1152988747,-0.0465168916,-0.2080016583,0.0339868106,0.2678874731,-0.0800447315,-0.162893489,0.0302972253,0.1708200723,-0.3860572278,0.174606517,-0.141842559,-0.3942222595,-0.2390128076,-0.0315945931,-0.3715678453,-0.1894314736,-0.1556600928,0.0615842938,-0.3262865841,-0.1295667291,-0.018050408,0.4424030185,0.3042119741,-0.3458757699,0.0595416166,-0.2154818922,0.042738162,-0.130989626,-0.1706084907,0.1832452714,-0.187360242,0.1254710108,0.0789805055,-0.0027811097,-0.1191304699,0.0368349142,-0.1459919363,0.0359485485,0.0778753459,0.2965763509,0.3371768296,0.0256925169,-0.1617322117,0.3484908044,0.843675673,0.2165461034,0.0563409589,0.0981767774,0.0150642609,-0.1024603769,-0.3751531541,-0.2478437275,0.2379531264,-0.0854845494,0.0561088324,0.072308369,-0.0772714242,0.0048563904,0.1957648844,0.2498804927,0.0215007327,0.0425558276,0.3785427511,0.1243986413,-0.0465026982,-0.0227752458,0.1214474887,0.1887886971,-0.1844374537,0.0142784761,-0.1256844401,-0.0463814326,0.162556529,0.0539215021,-0.3288117647,-0.3003205359,0.1932557374,0.0604248121,0.0394952781,0.2817931473,0.2645995617,-0.0857881904,0.0370729528,-0.0735380203,-0.0225856323,-0.4422388971,-0.4114542902,-0.0368375145,0.4833407998,0.0949172974,-0.1252068281,-0.032306686,-0.3935490251,-0.0279508606,-0.6040877104,0.3295165896,-0.333136797,0.4074757099,-0.1720993519,0.2469400465,-0.4784019589,-0.1110970378,0.3333790004,-0.3845154941,-0.3017195761,0.0676881522,0.0710870326,-0.0466582663,-0.1489862353,-0.6470291615,-0.1473304182,-0.1091402397,0.4985009432,-0.0058405274,0.0458777025,0.1098115295,0.0164884422,0.3667317629,-0.0059525901,-0.1577126831,0.0318522938,-0.1640760601,0.3424066603,-0.2221754491,-0.3280686438,-0.0690231994,0.2337097973,-0.1552115828,-0.1099913195,-0.3532603979,-0.2852552831,-0.2987848818,0.434972167,-0.0611681677,-0.0350607634,0.3117545843,0.275563091,0.0976741165,0.0934046656,-0.4001431465,0.4344343841,0.6834816933,0.0833120421,-0.0342245102,0.5533647537,0.0899443552,0.8097878098,0.1031533405,-0.1656104326,0.1677048951,0.0778776035,0.005261431,-0.4429137707,0.0429212749,0.2017114013,-0.0632162169,0.0438807979,0.2678488791,0.0202184822,-0.2609099746,-0.1716641188,0.1436476707,-0.2923775911,-0.3414456248,0.1241761819,-0.2106292993,-0.1484820545,0.1941257119,-0.0987070277,-0.1822680682,-0.3483558595,-0.2254256308,-0.1001440734,0.3196951151,-0.05339019,-0.6479715705,-0.0875155032,-0.4460581243,0.2668369114,0.1279461384,0.212784037,-0.3735551536,-0.0544394925,0.2839043736,-0.30193156,0.3205850124,-0.2891272604,-0.0480998494,0.1519804299,0.0159742236,-0.2380598634,-0.0223413613,-0.1441534162,0.1940582395,0.5879587531,-0.1936554611,-0.3545030057,-0.3136759996,0.2816260755,0.0463079326,-0.0623368509,0.1042388901,-0.2907800972,-0.0549556948,-0.4001672268,-0.1576462239,0.3127060831,-0.1740195304,-0.0534380563,0.1099234968,0.2132883668,0.1688006073,-0.0124370763,0.3079572022,-0.1972868443,-0.1283271313,0.4286997914,0.3257335722,-0.0797898322,-0.3301987052,0.3213172853,-0.1193918735,-0.3438797891,0.4488614798,-0.2451579124,0.330567956,0.3700147867,0.2473347187,-0.065853484,-0.0752681941,0.0922386497,0.3903131783,-0.0741988569,0.5959640145,0.3472822905,-0.0595443808,-0.2042069733,0.1596005857,-0.1428693086,-0.0234225336,0.4482417405,-0.1692369729,0.041779004,0.5369915962,-0.0495180972,0.9247649908,-0.3499218524,0.3505095541,0.1357945949,0.0182194281,0.4365855157,-0.2344040126,0.240011096,-0.4798047543,-0.2921574414,0.208843857,-0.1886552125,-0.2514409423,0.2594407499,-0.2159516066,0.0457162336,0.0013122101,0.1847947836,-0.0227824934,0.2629057169,0.1552161127,-0.1638199091,-0.4315559864,0.0302754231,0.0067291418,0.241835013,-0.062391717,-0.1761704683,-0.3166872859,-0.0520762317,-0.1919191778,0.2700083554,-0.0803584456,0.1262137145,0.2949758768,-0.0069244448,0.2545636594,0.7602497935,0.1860231906,-0.0185454115,-0.0112047484,0.1779197305,-0.0544717796,0.1112193689,0.4020736217,0.0590772219,0.2902452052,-0.1869535595,-0.2488675267,0.0755033121,0.1626166552,-0.2225751132,-0.2151713967,-0.0122397747,0.1055054739,-0.0755755156,0.001919309,0.1198566183,-0.5025991201,0.038489636,0.0036411064,-0.0727237388,0.0649872348,0.3388033807,0.0638106689,-0.1075140983,0.1561478674,0.7182194591,-0.3120693862,-0.1387109458,0.7221978903,0.2079787552,-0.0418201759,-0.0995992646,0.2471390516,-0.304558605,-0.4656019807,0.0179535914,-0.2125183195,-0.1623988599,0.2119241953,0.2940163314,-0.220177263,-0.272539854,-0.0979628712,-0.2364477515,-0.3015934825,0.1787427217,-0.0578235798,0.0050509446,-0.0420723706,0.3791486621,0.2284284234,-0.0167641025,-0.1284270287,0.0341603011,-0.3115146756,0.2559322119,0.1985469609,-0.4328902066,0.3220158517,-0.1326296479,-0.0461922735,0.262499094,-0.2474624664,-0.1005740911,0.0702863112,0.205632627,-0.0237762723,-0.0521885902,-0.2659519911,0.0523419566,-0.0974233076,-0.2412008047,0.231006071,-0.2239622921,-0.0207442213,0.0724927783,-0.0768057853,0.3257153034,-0.0045717256,0.0822957307,-0.0244880095,-0.0008573722,0.3325154185,0.1481817961,-0.0034487802,0.000260327,-0.2897581458,0.0241578799,-0.0271782558,0.049903065,0.2957170308,-0.3581775129,-0.0035801344,-0.1861483008,-0.1735498309,0.5093322396,-0.1074940562,-0.2644350529,0.2671612799,0.0344574414,0.0598587058,0.195624128,0.6023293138,0.0553345196,-0.0162441377,-0.1583835036,-0.0095602451,-0.0368193239,0.0736599043,0.2247649282,0.3374265134,-0.6618201137,0.0494201332,0.2687804401,-0.0405998714,0.1027490869,0.0354493298,-0.040319033,0.1243018731,0.2339228243,0.0949204043,-0.2511602342,-0.4647243023,0.5415120721,0.3499782681,-0.1950553656,-0.2824059129,0.4390031397,0.3190989196,0.467707485,0.2687078416,0.393148005,-0.0098133702,-0.2834889591,-0.2741727531,0.3842073083,-0.0253637284,-0.1247905418,-0.0068920995,-0.1023710147,-0.0516265705,0.2966699302,-0.3550789356,0.1451126486,0.3079491556,0.493080169,0.0731965527,-0.2727909386,-0.2053485215,-0.1373152882,0.4207772613,0.0681494325,0.3508649468,-0.1475654989,-0.0496549793,-0.2837129831,0.2799006402,0.1476244926,0.3031749725,-0.595653832,0.4350794554,0.0269202143,0.0228235908,0.1944430172,0.2336976379,0.2844567895,-0.0502659045,0.0580865592,-0.0733015761,0.0107049569,-0.2811210155,0.0781213492,0.3538541496,0.1165746972,-0.1111379713,0.244793117,-0.2593701184,0.0568449832,-0.0591134988,-0.4560569525,0.250371784,-0.1553744972,-0.0021282854,0.010184831,-0.0923608169,0.0050078905,0.1619819105,0.5490398407,0.1979664564,0.1381159723,0.0393289924,0.0509526543,-0.5822478533,0.1632904857,0.1882907897,0.1630127877,0.2088681757,0.2540136278,0.2885809541,0.3300600052,0.1523543745,-0.1867273897,-0.2593983114,0.4002560377,-0.5717066526,-0.0073625739,0.0127686309,0.1711681336,-0.099544093,-0.1628735811,-0.0443682633,-0.1043577269,-0.1096518561,-0.2734408677,-0.1667292714,-0.0709542483,0.1438029259,0.2195367515,0.2537907064,0.1641733497,0.0044323751,-0.2144688368,-0.1059022397,-0.1761987507,-0.0437833034,0.0550512709,0.0046705739,0.0140324319,-0.1635570675,0.3399705291,-0.3152092099,0.0040162094,-0.1181774884,0.3310361207,-0.045544941,-0.3487639427,0.4909962118,0.2506178617,-0.0844623372,0.6284148097,-0.1146673188,-0.1135939062,-0.6335580945,0.1557327211,0.3288764954,-0.3157316744,-0.1450184584,0.0803747922,0.0883033946,0.268824935,0.056056805,-0.4645878375,-0.0650695264,0.1141358763,-0.2358741015,-0.5998781323,0.1608873159,0.4694691598,0.1428550035,-0.2050244063,0.0963355303,-0.0697270259,0.035414964,-0.1779156178,-0.2029335797]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/478","title":"Export TFRecord to GCP bucket","comments":"Nevermind, I restarted my python session and it worked fine...\r\n\r\n---\r\n\r\nI had an authentification error, and I authenticated from another terminal. After that, no more error but it was not working. Restarting the sessions makes it work :)","body":"Previously, I was writing TFRecords manually to GCP bucket with : `with tf.io.TFRecordWriter('gs:\/\/my_bucket\/x.tfrecord')`\r\n\r\nSince `0.4.0` is out with the `export()` function, I tried it. But it seems TFRecords cannot be directly written to GCP bucket.\r\n\r\n`dataset.export('local.tfrecord')` works fine, \r\nbut `dataset.export('gs:\/\/my_bucket\/x.tfrecord')` does not work. \r\n\r\nThere is no error message, I just can't find the file on my bucket...\r\n\r\n---\r\n\r\nLooking at the code, `nlp` is using `tf.data.experimental.TFRecordWriter`, while I was using `tf.io.TFRecordWriter`. \r\n\r\n**What's the difference between those 2 ? How can I write TFRecords files directly to GCP bucket ?**\r\n\r\n@jarednielsen @lhoestq ","comment_length":39,"text":"Export TFRecord to GCP bucket\nPreviously, I was writing TFRecords manually to GCP bucket with : `with tf.io.TFRecordWriter('gs:\/\/my_bucket\/x.tfrecord')`\r\n\r\nSince `0.4.0` is out with the `export()` function, I tried it. But it seems TFRecords cannot be directly written to GCP bucket.\r\n\r\n`dataset.export('local.tfrecord')` works fine, \r\nbut `dataset.export('gs:\/\/my_bucket\/x.tfrecord')` does not work. \r\n\r\nThere is no error message, I just can't find the file on my bucket...\r\n\r\n---\r\n\r\nLooking at the code, `nlp` is using `tf.data.experimental.TFRecordWriter`, while I was using `tf.io.TFRecordWriter`. \r\n\r\n**What's the difference between those 2 ? How can I write TFRecords files directly to GCP bucket ?**\r\n\r\n@jarednielsen @lhoestq \nNevermind, I restarted my python session and it worked fine...\r\n\r\n---\r\n\r\nI had an authentification error, and I authenticated from another terminal. After that, no more error but it was not working. Restarting the sessions makes it work :)","embeddings":[0.1043086722,-0.0196554437,0.0418730006,0.1138932854,0.0320830382,-0.1475580037,0.1983275712,0.1378264129,-0.4294385016,0.1086212397,-0.2599080205,0.236711055,-0.2770297229,0.1902204603,0.4431501031,0.0458829962,-0.1624467522,-0.0884751529,-0.061361704,-0.2464275807,-0.2078003883,0.3399341702,-0.1633239537,0.1298568994,-0.1636379361,-0.3291425407,-0.139205575,0.0100800823,0.2158965915,-0.0817750022,0.1531289816,-0.1572728902,0.2936231494,0.3548467457,-0.0001241777,0.0557242855,0.0362486765,-0.2642901838,-0.100471437,-0.1585778296,-0.3611402512,0.0675945058,0.0743403882,-0.2352459431,0.0370032154,0.1353317052,0.5893600583,0.2575761676,-0.1871163547,0.4628814161,0.1116086021,0.4620780349,0.1533130854,0.171333313,-0.1146622375,-0.240670979,-0.3874956667,0.056643907,-0.1590555906,0.1580326706,0.143365398,-0.1098379567,0.0033528022,-0.2057417184,0.316311717,0.1110550016,-0.1072750613,-0.3120269477,-0.0031496282,0.1946424842,0.6180038452,-0.5000046492,-0.5036103725,0.0647539794,-0.0492016487,-0.2039890885,0.3115359247,0.4222564101,0.3510525525,0.1510547101,-0.2972370982,-0.1687379032,-0.305837661,0.1791097075,-0.0949265212,0.6592948437,-0.063075982,0.0978821591,-0.2444780767,0.0606238991,0.1662686914,-0.0485495552,-0.1336772144,0.3695758283,-0.1432873607,-0.3499161899,0.2528569102,0.4646779299,-0.551558435,0.3166106343,0.3559399545,-0.194340691,-0.0089944676,0.0084281713,-0.008708938,0.0439142585,0.3559091091,0.4459598958,0.0651987717,-0.3567683697,0.0706164166,-0.1742616594,-0.0649130419,-0.2846621871,-0.1426874548,0.0446223691,0.0379393287,-0.1374833286,0.0949023291,-0.1490453333,-0.0987590104,0.1095617786,-0.3324379027,0.0881567821,0.078425169,-0.2568153739,0.2772000432,-0.0266955011,-0.6436520815,0.0527903698,-0.0495500788,0.1958491653,0.1571635455,-0.0411067493,-0.0443773828,0.0338041112,0.0224749073,-0.2291047275,0.050985489,0.2616701722,0.3601639271,-0.0454379916,-0.1735641062,0.2035631239,0.039008785,0.2462253273,0.0227716509,-0.1490443349,-0.2866272926,0.1152702644,-0.2529262006,-0.3067851067,-0.0610647313,0.0415606424,-0.1484234333,-0.3995211124,-0.2166468948,0.001940498,0.12320555,-0.3684224784,0.2590978742,-0.3070381284,-0.0856938884,-0.3298635483,-0.0941886902,0.1332684308,-0.4248131216,-0.0791197941,-0.0175602064,0.0383081697,0.2924700975,0.2817420959,0.1365849376,0.183299154,0.0643903017,0.3293117881,0.3842251599,-0.5604013801,0.03586917,0.3170734644,-0.3729988337,-0.1465917081,-0.2245298326,-0.0009617406,0.3030170798,-0.2277836204,-0.036367517,0.3661177754,0.0217803996,0.173636511,-0.2771191001,-0.1050426215,0.0246056728,-0.277939409,-0.0159468167,-0.131043762,0.4008926451,0.1757673174,0.1823630035,-0.0711171404,0.2193006724,0.4072651267,0.3658869863,0.0230924152,-0.0836257339,0.0767629668,-0.531768918,0.0073541328,-0.2281797528,0.183049798,0.3466967344,0.0587602891,0.1974603981,-0.2921520174,0.1712455302,-0.1641822904,-0.0951042697,0.1215174198,0.1402537674,0.3574543297,-0.0441086814,0.2640912533,-0.1312345117,-0.1229351237,-0.2632279694,0.3617303073,0.1286933124,-0.3282949626,0.0945415348,-0.1186136305,0.0846155584,-0.2326744199,-0.2019333243,0.3773418665,-0.2473888844,0.5208188891,0.0495079793,0.6487165093,0.1313399225,-0.2497121394,-0.1034749895,0.3585964739,-0.0531645492,0.2756613791,-0.1737519354,0.5877634287,0.1556873918,-0.1525974125,0.386910677,0.0487791412,-0.2978127301,-0.1340873986,-0.0127204712,0.1818125993,0.1251281351,-0.1221780106,0.0884689987,-0.1293487549,-0.3697161674,0.3899160624,0.6853585243,-0.1292418689,0.0035886897,-0.1097488627,0.0905126184,-0.0675529614,0.512488842,0.0625751838,0.2000595033,0.0357431956,0.1346185952,-0.1173051447,-0.2365724593,-0.1088180393,0.3869111836,-0.1922742873,-0.2225830555,-0.0102991899,-0.0086948993,0.256480515,-0.1329466552,-0.0602575056,0.0856533647,0.3868994415,-0.4979802072,0.070247151,-0.4411150217,-0.1988404244,-0.107116513,0.3426580131,-0.3205321133,-0.327308774,0.0365762599,0.0401102044,-0.1647715867,-0.0049517136,-0.115125224,0.0103284158,0.0407161601,-0.1432955116,-0.1063830405,0.0423582457,-0.2163069546,-0.0144215478,0.0300439857,0.1656529754,0.288261354,-0.0380332991,-0.1832355559,-0.0969579294,-0.3768307269,0.0041226777,-0.1586945951,0.1078701615,0.0106572863,0.0450568758,-0.5808387399,-0.1346590817,-0.1810050756,-0.4385964274,-0.2040456384,0.0246810187,-0.0896000415,-0.0534392856,-0.25321576,-0.1335134059,-0.237013936,-0.3757142723,0.2796498239,0.2792012393,0.3979219794,0.4036666453,-0.464941889,0.2192778587,0.1281318218,-0.2627945244,0.271951437,-0.1234670654,0.4003706574,-0.5539553165,-0.3774174154,0.1360996515,0.1978217512,-0.0095765833,0.0663508475,-0.2033832073,0.0864781067,-0.3496908844,-0.0653625652,0.1325469315,0.2524751425,0.4491384029,-0.2184735686,0.161797449,-0.1543953419,-0.1096281409,0.0963791758,0.9362360239,0.023415098,-0.0219114739,0.115382202,0.058070343,0.3341713846,0.0435016714,0.0677085593,0.0357965454,-0.3759589493,0.19030267,-0.2492443174,-0.1469091624,0.6196631193,0.4616970718,-0.1018358991,0.3610274196,0.0375996083,0.2365515381,0.1718462408,0.0989680439,-0.0671919063,-0.2950975895,-0.0060201604,0.2981249392,0.3115376532,-0.0430108234,-0.0540911071,-0.4550203085,-0.3560125232,-0.3157883286,0.3434743881,0.5764551163,0.4297193587,0.1184362918,0.2611116767,-0.5747964382,0.2416063249,0.0991277918,-0.0526561998,-0.1630297899,-0.164091453,0.197307989,-0.0980384424,0.0605771728,-0.3111293018,-0.0378794335,0.2995841205,-0.1518817544,-0.5790774822,0.0650467053,-0.1185839921,0.0833327845,0.3748459518,0.1614403427,-0.0566270798,-0.1530476958,-0.0001290093,-0.1339464933,-0.2941281796,0.0789089203,-0.129574284,-0.3237827122,-0.0456325784,-0.2065203339,-0.0002370517,0.1084332019,0.0292207338,0.3305555284,0.3428139091,0.0191643108,0.2844891548,-0.092566736,0.0921057984,0.2597309649,-0.0445696674,0.1416483372,-0.1034709662,0.0686074421,0.0790599734,0.1394546777,0.2667695582,0.3338950872,-0.015051824,0.1412465572,0.3176981807,0.1359122694,0.1435529143,-0.3187103868,0.2108194679,0.0707107559,-0.0898435861,-0.1369028687,-0.2198533565,-0.5876247287,-0.1534554511,0.2062444389,-0.1491347551,0.0905934423,0.4807756543,0.244858861,-0.4734282196,0.0507445857,0.0618905164,0.9134733081,-0.3971925378,0.3144845963,0.459332943,-0.0659245998,0.6889105439,-0.2493766099,0.2878426313,-0.2602439225,0.0969928876,-0.004633483,0.0240999386,0.1859912872,0.2977377474,-0.3869919479,0.243687287,0.0800648183,0.078329131,-0.1010479182,0.1207978204,0.0740610361,-0.0537050106,-0.0044608247,0.0387087241,0.1480157226,0.1679361314,-0.1008350328,-0.2626407743,-0.150668785,0.097585097,-0.0364320204,0.107765913,-0.0858727843,0.0125954272,0.1153871119,-0.2285217494,0.1785608083,0.0447077006,0.2295649648,0.174263224,-0.1165140942,-0.1177859977,0.1121453345,0.0951123014,0.0477783196,0.2640109658,0.0588142052,-0.1901765615,0.0238818415,-0.0143232942,-0.0229323637,-0.6582657695,-0.2426000834,-0.019506488,-0.3383099437,0.040693745,-0.1690116525,0.1314940155,-0.3500055373,0.0776489004,0.0422845073,-0.0317068845,-0.0073299357,0.1334018856,-0.3764158785,-0.1055553555,-0.0696593374,0.3178859353,-0.1067608222,0.0260589123,0.2930838466,0.1390781701,-0.1154675707,-0.1924812943,0.1168998629,0.2639497221,-0.34426862,0.3508827984,-0.091165632,0.389770329,0.2459413111,0.4746273458,-0.1922618002,0.116606839,-0.2697854936,0.1351474524,-0.2463658452,0.1444520056,0.1061556265,-0.0853376985,-0.2414738983,0.8087238669,0.2888928354,-0.1660509706,-0.210435763,0.1199412495,-0.2968470156,0.2891426384,0.4651018083,0.0027299952,0.1073265299,0.1166567355,-0.0226150379,-0.0354599804,-0.1430861056,-0.0093736853,-0.4408384264,0.2326544225,0.3971709609,-0.2790672481,0.0164313354,-0.2374552488,-0.307277739,-0.0352573507,0.1781384796,0.3160158396,-0.1167777032,0.2806138098,0.3236216009,0.042404294,-0.1924270242,-0.0534701124,0.2104379833,0.3997379839,-0.0866997838,0.2149668187,-0.0377441421,0.0833413005,-0.1107140705,0.2121828943,-0.181283161,0.2066606283,0.1498606056,-0.2446236461,0.0587378144,-0.2719559073,0.0709206983,-0.0187377129,-0.1596562713,-0.110574238,0.2401583791,-0.0146167073,0.2646248043,0.1981234103,0.1586775184,0.1647420526,-0.1007743701,0.2378634661,-0.1556211561,-0.1299574226,0.0340242311,0.1847242564,0.1766388267,-0.3030108511,0.043707408,0.2135511339,-0.0956344828,0.0900674015,-0.0498673134,-0.3752327859,0.0936056972,-0.141403541,0.0692962632,-0.0426435918,-0.018485805,-0.0263287406,0.2260893285,-0.7964652777,-0.4338706434,-0.1880561411,0.2785821557,0.4275818169,0.4440070391,-0.1418630481,-0.234551698,-0.1608277112,-0.1247887462,0.5020839572,-0.1063784361,0.0345153734,-0.307882309,-0.0675121397,0.1722891927,0.2360776961,-0.1693919003,0.0225640237,-0.0608972795,0.4165675044,0.2164749503,-0.0459020436,-0.2489138693,-0.1122570708,0.2209977508,-0.0683110282,0.4019662738,-0.4644498825,-0.1036423668,-0.3935958147,0.1805748791,0.1125632524,0.0332113802,0.2970167398,0.5219265819,0.1952507943,0.1929665357,0.2672564089,0.2638758123,0.026286535,0.3219049275,0.1198598072,-0.0960358903,-0.0178402103,0.0783526823,0.174567759,-0.0004817065,-0.1113834828,0.5997070074,0.0313789509,-0.0041406257,-0.1581807882,0.1655310392,-0.3940217495,0.1993963718,0.1589346677,0.0326426849,0.0328828394,-0.0332193114,0.0036345283,0.0754840448,0.2761546075,0.4486453235,0.2513770461,-0.249519527,-0.20274131,-0.258102566,0.3710936904,-0.0538533553,0.1372666508,-0.0615194514,0.3472933769,-0.0496514998,-0.2888678908,0.1055768952,-0.0285973195,-0.0261401813,0.0440636463,-0.0828725323,-0.0690063313,-0.3028970361,0.0072684935,-0.1658412814,-0.428278029,0.0905362517,-0.0716690198,-0.0630941093,-0.2350928485,0.0287554134,0.2469121516,0.1086403206,0.2696193457,0.1132545248,0.0510669015,-0.08165209,-0.0887761638,-0.3101071119,-0.1805019677,-0.3347492218,-0.168242231,-0.2967460155,0.5181001425,-0.0839154422,-0.2681004405,-0.3095509112,0.3083866239,-0.3641299009,0.5591872334,0.3405432999,0.1070056334,-0.4990778863,-0.1324353367,0.0893776715,0.3456855416,0.0699133277,0.0313380659,-0.4642445743,-0.1810526997,0.4082534313,-0.8163272738,-0.3400095999,0.2773923576,0.2063267082,-0.0303844456,-0.2536889613,-0.288924098,-0.107532464,0.339209646,-0.0721909553,-0.081497781,0.0767053291,-0.042154178,0.1234855056,0.2067712992,0.1660784483,0.3275052905,-0.2120841593,-0.0851779133,-0.1489029825]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/477","title":"Overview.ipynb throws exceptions with nlp 0.4.0","comments":"Thanks for reporting this issue\r\n\r\nThere was a bug where numpy arrays would get returned instead of tensorflow tensors.\r\nThis is fixed on master.\r\n\r\nI tried to re-run the colab and encountered this error instead:\r\n\r\n```\r\nAttributeError: 'tensorflow.python.framework.ops.EagerTensor' object has no attribute 'to_tensor'\r\n```\r\n\r\nThis is because the dataset returns a Tensor and not a RaggedTensor.\r\nBut I think we should always return a RaggedTensor unless the length of the sequence is fixed (it that case they can be stack into a Tensor).","body":"with nlp 0.4.0, the TensorFlow example in Overview.ipynb throws the following exceptions:\r\n\r\n\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-48907f2ad433> in <module>\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\n<ipython-input-5-48907f2ad433> in <dictcomp>(.0)\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\nAttributeError: 'numpy.ndarray' object has no attribute 'to_tensor'","comment_length":83,"text":"Overview.ipynb throws exceptions with nlp 0.4.0\nwith nlp 0.4.0, the TensorFlow example in Overview.ipynb throws the following exceptions:\r\n\r\n\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-48907f2ad433> in <module>\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\n<ipython-input-5-48907f2ad433> in <dictcomp>(.0)\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\nAttributeError: 'numpy.ndarray' object has no attribute 'to_tensor'\nThanks for reporting this issue\r\n\r\nThere was a bug where numpy arrays would get returned instead of tensorflow tensors.\r\nThis is fixed on master.\r\n\r\nI tried to re-run the colab and encountered this error instead:\r\n\r\n```\r\nAttributeError: 'tensorflow.python.framework.ops.EagerTensor' object has no attribute 'to_tensor'\r\n```\r\n\r\nThis is because the dataset returns a Tensor and not a RaggedTensor.\r\nBut I think we should always return a RaggedTensor unless the length of the sequence is fixed (it that case they can be stack into a Tensor).","embeddings":[-0.0391482376,0.082511723,-0.0073065483,-0.0530798919,0.0318035781,-0.0482144579,0.8157610297,0.2792399526,-0.0684225112,0.1789912581,0.0316702574,0.2038193345,-0.297031194,-0.0530733988,0.0210223142,-0.4359816313,-0.1079105809,0.286254257,0.0336867906,-0.0560540147,-0.1683885306,0.2695775628,-0.4814034402,0.0699896812,-0.3009614944,-0.2224038243,-0.003463051,-0.1184036061,0.1805594563,-0.3218075335,0.2983432412,-0.0853027105,-0.0729661137,0.2157041878,-0.0001094074,0.1312537491,0.2247508913,0.0399746299,-0.1744725704,-0.3294422925,-0.1835713238,-0.0175784472,-0.0488839783,-0.3077403009,0.1411332339,0.0695705861,0.2543091178,0.2293205559,0.187586844,0.4259967804,0.2343905419,0.3789539039,0.0684154481,0.0336801223,0.0507045202,-0.0231374409,-0.0912747681,0.1140883863,-0.0621383898,-0.1997990906,-0.2145836651,0.2809059322,-0.0087522771,0.0915340558,-0.158366695,0.0241951719,0.0742403716,-0.2711643577,-0.0711959004,0.4140101969,-0.0402942188,-0.0426515564,-0.392829448,-0.3703237176,0.0652718395,-0.5704241991,-0.060038127,0.1048349366,-0.1118555367,-0.016561389,-0.1545907557,0.095754236,-0.5827885866,0.3474161923,-0.1908776015,0.1448999345,0.0082724849,0.081057407,-0.0094359042,-0.0986803472,-0.0254501868,-0.0089679696,0.1772603244,0.0017488351,-0.2649015784,-0.220381543,0.1818951219,-0.206580326,-0.1672124565,-0.0775770694,0.0233217161,0.0268009454,0.0007514657,0.3499961197,-0.122096628,0.2610676587,0.1183250472,0.1576666385,0.3323271275,-0.0056893472,-0.1858868748,0.0312410183,0.0496797599,-0.1437185258,-0.1408727169,0.017316794,0.3466324508,0.1334186941,-0.5548805594,0.0387058556,-0.3659700453,0.1409137547,0.192010656,0.2450779825,-0.0872803405,0.2586591542,0.2062356174,0.1746394187,-0.4506019652,-0.0207848474,-0.2161477357,0.3215353191,-0.054345686,-0.2500003576,0.2410529703,0.32728827,0.3226420283,-0.0850195363,0.0863246098,0.0834727958,0.3203797638,-0.2739877701,0.1113804728,0.2461777329,-0.0800185576,-0.1207126752,0.046420671,-0.1517851949,-0.0383200198,-0.0575329512,-0.2080945373,-0.3066107631,-0.0356256999,0.2675876915,0.1201595962,-0.0608261302,-0.2594278157,0.1365095973,0.0316395611,-0.1947879344,-0.0350830965,-0.3620331287,-0.1259934753,-0.0612786673,0.0401064605,0.190676704,0.1341350079,-0.1566854864,0.1043589339,0.1802005768,0.2523798943,0.2696486115,-0.2044544071,0.2307111025,-0.036194697,0.3748850822,0.3027423024,-0.2861011326,-0.1879547834,0.1153768301,-0.0510694645,0.0571820475,-0.0915255472,0.019735707,0.4229823351,-0.038066525,0.3864845932,0.2939309478,-0.2305163592,0.098540768,-0.0844678134,0.1140950769,0.292221874,-0.0645060986,0.1169267669,0.0630712658,-0.1999797374,0.5250784159,0.0386198647,-0.1206296608,-0.0273567997,0.0281608887,0.0251637865,-0.3499792218,-0.1735033095,-0.1349679232,-0.3476230204,0.0723298341,0.0827099904,0.2156667709,0.3347974122,-0.0584712513,0.0152782639,0.1186521649,-0.077429764,-0.2848463356,0.2511830032,0.1563904881,-0.1726750731,0.0747464448,-0.1370531023,0.0778900459,-0.3810476959,0.1017818525,-0.1399620175,0.0689370707,-0.0984523445,-0.1941973716,-0.2610077858,0.3692515492,-0.0739928558,-0.1785955727,-0.0611250252,0.2158330977,0.1814824939,0.2297656089,-0.2306142896,0.1413954794,0.3241619766,-0.0578077212,-0.1601210982,0.2085707486,0.0083810706,-0.1782839298,0.1594430953,0.4819951057,0.2851334512,0.1752204299,0.4278370738,0.1367619187,0.2068368047,0.0023742397,-0.1972270608,0.08879821,0.3635754287,-0.0712224767,0.1973559558,-0.2939554155,-0.7215968966,0.1118019372,0.549682498,0.1343672872,0.0659737885,0.1962781847,-0.0909676552,0.1029396802,0.3837305307,-0.2072698623,0.1110235453,0.0864473283,0.2527721822,0.0789593309,-0.3717411757,-0.2521877289,0.0808835775,0.1414926648,-0.1022182554,-0.0228342731,0.2730284631,0.2171438932,-0.2615646124,-0.3487101197,-0.1701575369,0.2663602531,-0.3354310095,0.3268645406,-0.1205521524,-0.2741838694,-0.0820815712,-0.5632296801,-0.1426299363,-0.3243647516,-0.0196133722,0.626950562,-0.3355834782,0.2322810441,0.1679948866,0.0385323465,0.0734659433,0.2469550222,0.1575049609,0.0089548556,-0.3124403656,0.1394797713,0.1999257207,-0.1503335834,0.2847865522,-0.0808443576,-0.3065349758,0.2594654858,-0.5440360904,0.1175272763,-0.1258309186,-0.0007881688,0.2060860097,0.2676906288,-0.2906229794,-0.5785192251,0.0555495806,-0.1116683707,-0.1575769037,0.3648533523,-0.0376024731,-0.0444794111,-0.3675981462,-0.5435934067,-0.1716129035,-0.4446734786,0.361694634,-0.0806766227,0.307082206,0.7853786349,-0.0730392784,0.5275722146,0.2695943415,0.1862934232,0.2173419595,-0.0682444572,-0.00816104,-0.2857410312,-0.3695432544,-0.106303297,-0.2888970375,0.1831954867,0.1218610108,-0.329808116,-0.0585052744,-0.0348194987,0.134925127,0.1361687481,-0.1742237657,0.3653981388,-0.0563014857,-0.0302218366,-0.0342406295,0.0573622622,0.3721608818,0.2451589406,-0.0622616112,-0.0394694619,0.2059445679,0.1754864305,0.3692196012,0.0424718671,-0.2425438613,0.2184517384,0.0004712994,0.0349336825,-0.1755885184,-0.276870966,0.2389113009,-0.0171982739,-0.0787892938,0.0796449482,-0.1992204487,-0.294082135,-0.1044809669,0.0583732612,-0.2043749243,-0.3995546997,0.1599550992,-0.3206953704,0.1283507496,-0.0597034842,0.0156895258,-0.215094164,0.0657821,-0.1475070715,0.1715197265,0.0861308575,-0.2596295774,-0.0938589796,-0.5862797499,-0.6067792773,0.2046329379,0.2464426905,0.5263921618,-0.0870106295,-0.018824283,-0.1141606569,-0.0416838527,0.4461074471,0.1069852561,-0.1425531656,0.2608311474,-0.1329194158,-0.3578276038,-0.1220484376,-0.3692784905,0.2719855607,0.5299979448,0.3015776575,-0.2063939273,0.0024211989,0.3558468223,-0.0466872342,0.1243460029,0.1750926822,0.0414709896,-0.0300873742,-0.1414518505,0.0908922628,0.1586921364,-0.0416010357,-0.3086980581,-0.0585056618,0.0238241851,-0.074673824,0.1684126258,0.2123264968,-0.1910576522,-0.0090058353,0.0770116299,-0.2531859875,-0.1229064316,0.4033557475,-0.0854448378,-0.1612051129,-0.2986449897,-0.0300266352,0.1772330552,0.0864015371,0.1985335499,-0.026353687,0.0448872186,-0.2302386761,-0.0440556444,-0.0860254243,0.4295450151,0.5223546624,0.0664971247,-0.1000387594,-0.1619127095,0.2714983523,-0.3283502162,0.071457386,0.3990154266,-0.0608640239,-0.3527769744,0.2800023556,-0.2473477125,0.475312084,0.4118616879,0.2183436602,0.6721669436,0.3472309411,0.6203202605,0.1277313232,0.475355804,-0.1620083749,-0.017395379,-0.1005233079,-0.0788868144,0.0276142135,0.1096018106,-0.1319761723,0.0889560133,0.1767239571,0.0960331336,0.0463646725,0.0985127762,0.0611856617,-0.2404057235,-0.2900931239,0.1539499611,-0.1581514478,0.1574359983,-0.0482134856,-0.1438087523,-0.0824364349,-0.326462388,-0.2867511511,0.2754968703,-0.141487211,0.1955625713,0.2045818567,-0.1971868873,-0.0391067825,-0.028757615,0.2409318835,-0.0701698363,-0.1455816776,-0.2068726122,-0.0692644343,0.0280247983,0.1032120883,-0.1363179386,0.5273317695,-0.1486338824,-0.1375990957,-0.0097949114,-0.0955040231,-0.217352435,0.2621585131,0.1152564734,0.0907626301,0.1473237872,0.0249874983,0.1188034415,-0.4523480833,-0.2321523577,0.1703427881,-0.1349198818,-0.3381385505,0.2134271711,-0.1455636322,-0.0226632133,0.0562408082,0.2497803569,-0.0592680126,0.2341690958,0.6033888459,-0.0211929455,-0.0915532932,-0.143573314,-0.2165425569,0.0541517623,-0.1071855724,0.1656975746,-0.1493014842,0.0013679207,0.0525460094,0.1575044692,0.0081935311,0.2858624756,-0.1662966609,-0.3224808276,-0.3871929049,0.2180494815,-0.1401991397,0.2134718597,0.1630478501,0.617010653,0.0856665149,-0.1259983033,-0.32188797,0.0321272388,-0.2677123547,0.324180305,0.0389137156,-0.2092727721,0.1503251642,-0.0048052659,0.1132221594,0.1361236721,-0.1097702086,-0.258374095,-0.191740647,0.0714030862,0.1857145876,-0.3465097547,0.1661629826,0.0046596178,-0.1557182819,-0.1230325326,0.117200613,-0.056309741,-0.2534221411,0.5090382695,0.1321459264,0.6595722437,0.2523018718,0.0457537584,-0.1782053113,0.1686657667,-0.0889488161,0.1971289515,-0.0100230779,-0.0413222201,-0.3386868536,-0.0331776142,-0.4179776013,-0.1148272082,0.2204306722,-0.3328475952,-0.0377382897,-0.2488700449,0.0429368578,0.1350106597,-0.1134634167,0.026608292,-0.0060438812,0.2527722418,-0.0741036981,0.1623068899,0.2498591393,-0.0320506953,0.2161780149,0.1043511257,0.1092673615,-0.0902838409,0.1097071767,0.1391165406,-0.1979448646,-0.311792165,0.2412803024,0.3260473609,-0.3616290689,0.1162670925,-0.1829433441,0.0854910091,0.4222062826,-0.0515633076,0.0166013688,0.2456998378,0.0994251668,-0.1708704084,0.2237033695,-0.3996602893,-0.2821905315,-0.0317088962,0.1926876158,0.2779779732,0.1008707657,0.0297555309,-0.2888753414,-0.1592281461,-0.0015650663,0.2095531225,-0.397054702,0.0147343781,-0.3463863134,0.1573568583,0.0574848093,-0.3433611095,-0.037702471,0.0445325114,0.1723369956,0.1850241721,-0.2280970365,-0.0942424387,-0.1356519014,0.0157811623,0.2179116905,-0.061141856,0.3229675293,0.367595464,-0.0189267956,-0.2765446603,0.5054782033,0.4357511699,0.2400355935,-0.3785278499,0.0104650129,0.1237666383,-0.1097797677,-0.0272483379,0.0968386903,0.1076465994,0.2009624839,0.225017637,0.1305900365,-0.1786044687,0.3606996238,0.1361265182,-0.0453327298,-0.2492054254,0.3173814714,0.1219383851,-0.1387658566,-0.1831054091,-0.0519299507,-0.3085249066,0.1512973756,0.3084787726,-0.1038086489,0.1319825351,0.1568076015,0.0841484368,-0.0936999395,0.3824259341,0.2785471082,0.3412989974,-0.2391836047,-0.0532665066,-0.1395322829,0.1518156528,-0.1091642231,0.3494437933,0.1335536242,0.1528654844,-0.2013279647,-0.055491142,-0.1758026183,-0.4866576195,0.0382897556,0.0140972175,-0.5309222341,-0.0677651465,0.2167239338,-0.2995285392,0.0809302628,-0.2919172645,0.1038764492,-0.0216722004,0.0582708493,-0.0243856199,0.2547587454,0.0580757819,0.4695756733,-0.1598595679,0.2173871994,0.3523491621,-0.0399275683,-0.2264608741,0.0155344419,-0.1288682073,-0.3512552083,0.0623174794,0.0021564274,0.6310626864,0.0308014378,-0.3426463008,-0.0471834131,0.2966679037,0.0744638145,-0.2258533686,-0.2897467613,0.515901804,-0.3268099427,0.1209048629,-0.0355342515,0.2257708907,0.0469698273,0.1271169484,-0.4208705127,-0.5104087591,0.4589865506,-0.6723730564,-0.2989540994,-0.150011301,0.3177681267,-0.1784148961,0.1541518718,-0.7014427781,-0.1308532357,0.3214225769,-0.3068626225,-0.3737615347,0.1170787886,0.2110434473,0.016910512,0.0359381065,0.5025029182,0.0519079305,-0.3045867383,0.0904337913,-0.3307105601]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/477","title":"Overview.ipynb throws exceptions with nlp 0.4.0","comments":"Hi, I got another error (on Colab):\r\n\r\n```python\r\n# You can read a few attributes of the datasets before loading them (they are python dataclasses)\r\nfrom dataclasses import asdict\r\n\r\nfor key, value in asdict(datasets[6]).items():\r\n print('\ud83d\udc49 ' + key + ': ' + str(value))\r\n\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-6-b8ace6c227a2> in <module>()\r\n 2 from dataclasses import asdict\r\n 3 \r\n----> 4 for key, value in asdict(datasets[6]).items():\r\n 5 print('\ud83d\udc49 ' + key + ': ' + str(value))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/dataclasses.py in asdict(obj, dict_factory)\r\n 1008 \"\"\"\r\n 1009 if not _is_dataclass_instance(obj):\r\n-> 1010 raise TypeError(\"asdict() should be called on dataclass instances\")\r\n 1011 return _asdict_inner(obj, dict_factory)\r\n 1012 \r\n\r\nTypeError: asdict() should be called on dataclass instances\r\n```","body":"with nlp 0.4.0, the TensorFlow example in Overview.ipynb throws the following exceptions:\r\n\r\n\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-48907f2ad433> in <module>\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\n<ipython-input-5-48907f2ad433> in <dictcomp>(.0)\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\nAttributeError: 'numpy.ndarray' object has no attribute 'to_tensor'","comment_length":110,"text":"Overview.ipynb throws exceptions with nlp 0.4.0\nwith nlp 0.4.0, the TensorFlow example in Overview.ipynb throws the following exceptions:\r\n\r\n\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-48907f2ad433> in <module>\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\n<ipython-input-5-48907f2ad433> in <dictcomp>(.0)\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\nAttributeError: 'numpy.ndarray' object has no attribute 'to_tensor'\nHi, I got another error (on Colab):\r\n\r\n```python\r\n# You can read a few attributes of the datasets before loading them (they are python dataclasses)\r\nfrom dataclasses import asdict\r\n\r\nfor key, value in asdict(datasets[6]).items():\r\n print('\ud83d\udc49 ' + key + ': ' + str(value))\r\n\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-6-b8ace6c227a2> in <module>()\r\n 2 from dataclasses import asdict\r\n 3 \r\n----> 4 for key, value in asdict(datasets[6]).items():\r\n 5 print('\ud83d\udc49 ' + key + ': ' + str(value))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/dataclasses.py in asdict(obj, dict_factory)\r\n 1008 \"\"\"\r\n 1009 if not _is_dataclass_instance(obj):\r\n-> 1010 raise TypeError(\"asdict() should be called on dataclass instances\")\r\n 1011 return _asdict_inner(obj, dict_factory)\r\n 1012 \r\n\r\nTypeError: asdict() should be called on dataclass instances\r\n```","embeddings":[-0.0391482376,0.082511723,-0.0073065483,-0.0530798919,0.0318035781,-0.0482144579,0.8157610297,0.2792399526,-0.0684225112,0.1789912581,0.0316702574,0.2038193345,-0.297031194,-0.0530733988,0.0210223142,-0.4359816313,-0.1079105809,0.286254257,0.0336867906,-0.0560540147,-0.1683885306,0.2695775628,-0.4814034402,0.0699896812,-0.3009614944,-0.2224038243,-0.003463051,-0.1184036061,0.1805594563,-0.3218075335,0.2983432412,-0.0853027105,-0.0729661137,0.2157041878,-0.0001094074,0.1312537491,0.2247508913,0.0399746299,-0.1744725704,-0.3294422925,-0.1835713238,-0.0175784472,-0.0488839783,-0.3077403009,0.1411332339,0.0695705861,0.2543091178,0.2293205559,0.187586844,0.4259967804,0.2343905419,0.3789539039,0.0684154481,0.0336801223,0.0507045202,-0.0231374409,-0.0912747681,0.1140883863,-0.0621383898,-0.1997990906,-0.2145836651,0.2809059322,-0.0087522771,0.0915340558,-0.158366695,0.0241951719,0.0742403716,-0.2711643577,-0.0711959004,0.4140101969,-0.0402942188,-0.0426515564,-0.392829448,-0.3703237176,0.0652718395,-0.5704241991,-0.060038127,0.1048349366,-0.1118555367,-0.016561389,-0.1545907557,0.095754236,-0.5827885866,0.3474161923,-0.1908776015,0.1448999345,0.0082724849,0.081057407,-0.0094359042,-0.0986803472,-0.0254501868,-0.0089679696,0.1772603244,0.0017488351,-0.2649015784,-0.220381543,0.1818951219,-0.206580326,-0.1672124565,-0.0775770694,0.0233217161,0.0268009454,0.0007514657,0.3499961197,-0.122096628,0.2610676587,0.1183250472,0.1576666385,0.3323271275,-0.0056893472,-0.1858868748,0.0312410183,0.0496797599,-0.1437185258,-0.1408727169,0.017316794,0.3466324508,0.1334186941,-0.5548805594,0.0387058556,-0.3659700453,0.1409137547,0.192010656,0.2450779825,-0.0872803405,0.2586591542,0.2062356174,0.1746394187,-0.4506019652,-0.0207848474,-0.2161477357,0.3215353191,-0.054345686,-0.2500003576,0.2410529703,0.32728827,0.3226420283,-0.0850195363,0.0863246098,0.0834727958,0.3203797638,-0.2739877701,0.1113804728,0.2461777329,-0.0800185576,-0.1207126752,0.046420671,-0.1517851949,-0.0383200198,-0.0575329512,-0.2080945373,-0.3066107631,-0.0356256999,0.2675876915,0.1201595962,-0.0608261302,-0.2594278157,0.1365095973,0.0316395611,-0.1947879344,-0.0350830965,-0.3620331287,-0.1259934753,-0.0612786673,0.0401064605,0.190676704,0.1341350079,-0.1566854864,0.1043589339,0.1802005768,0.2523798943,0.2696486115,-0.2044544071,0.2307111025,-0.036194697,0.3748850822,0.3027423024,-0.2861011326,-0.1879547834,0.1153768301,-0.0510694645,0.0571820475,-0.0915255472,0.019735707,0.4229823351,-0.038066525,0.3864845932,0.2939309478,-0.2305163592,0.098540768,-0.0844678134,0.1140950769,0.292221874,-0.0645060986,0.1169267669,0.0630712658,-0.1999797374,0.5250784159,0.0386198647,-0.1206296608,-0.0273567997,0.0281608887,0.0251637865,-0.3499792218,-0.1735033095,-0.1349679232,-0.3476230204,0.0723298341,0.0827099904,0.2156667709,0.3347974122,-0.0584712513,0.0152782639,0.1186521649,-0.077429764,-0.2848463356,0.2511830032,0.1563904881,-0.1726750731,0.0747464448,-0.1370531023,0.0778900459,-0.3810476959,0.1017818525,-0.1399620175,0.0689370707,-0.0984523445,-0.1941973716,-0.2610077858,0.3692515492,-0.0739928558,-0.1785955727,-0.0611250252,0.2158330977,0.1814824939,0.2297656089,-0.2306142896,0.1413954794,0.3241619766,-0.0578077212,-0.1601210982,0.2085707486,0.0083810706,-0.1782839298,0.1594430953,0.4819951057,0.2851334512,0.1752204299,0.4278370738,0.1367619187,0.2068368047,0.0023742397,-0.1972270608,0.08879821,0.3635754287,-0.0712224767,0.1973559558,-0.2939554155,-0.7215968966,0.1118019372,0.549682498,0.1343672872,0.0659737885,0.1962781847,-0.0909676552,0.1029396802,0.3837305307,-0.2072698623,0.1110235453,0.0864473283,0.2527721822,0.0789593309,-0.3717411757,-0.2521877289,0.0808835775,0.1414926648,-0.1022182554,-0.0228342731,0.2730284631,0.2171438932,-0.2615646124,-0.3487101197,-0.1701575369,0.2663602531,-0.3354310095,0.3268645406,-0.1205521524,-0.2741838694,-0.0820815712,-0.5632296801,-0.1426299363,-0.3243647516,-0.0196133722,0.626950562,-0.3355834782,0.2322810441,0.1679948866,0.0385323465,0.0734659433,0.2469550222,0.1575049609,0.0089548556,-0.3124403656,0.1394797713,0.1999257207,-0.1503335834,0.2847865522,-0.0808443576,-0.3065349758,0.2594654858,-0.5440360904,0.1175272763,-0.1258309186,-0.0007881688,0.2060860097,0.2676906288,-0.2906229794,-0.5785192251,0.0555495806,-0.1116683707,-0.1575769037,0.3648533523,-0.0376024731,-0.0444794111,-0.3675981462,-0.5435934067,-0.1716129035,-0.4446734786,0.361694634,-0.0806766227,0.307082206,0.7853786349,-0.0730392784,0.5275722146,0.2695943415,0.1862934232,0.2173419595,-0.0682444572,-0.00816104,-0.2857410312,-0.3695432544,-0.106303297,-0.2888970375,0.1831954867,0.1218610108,-0.329808116,-0.0585052744,-0.0348194987,0.134925127,0.1361687481,-0.1742237657,0.3653981388,-0.0563014857,-0.0302218366,-0.0342406295,0.0573622622,0.3721608818,0.2451589406,-0.0622616112,-0.0394694619,0.2059445679,0.1754864305,0.3692196012,0.0424718671,-0.2425438613,0.2184517384,0.0004712994,0.0349336825,-0.1755885184,-0.276870966,0.2389113009,-0.0171982739,-0.0787892938,0.0796449482,-0.1992204487,-0.294082135,-0.1044809669,0.0583732612,-0.2043749243,-0.3995546997,0.1599550992,-0.3206953704,0.1283507496,-0.0597034842,0.0156895258,-0.215094164,0.0657821,-0.1475070715,0.1715197265,0.0861308575,-0.2596295774,-0.0938589796,-0.5862797499,-0.6067792773,0.2046329379,0.2464426905,0.5263921618,-0.0870106295,-0.018824283,-0.1141606569,-0.0416838527,0.4461074471,0.1069852561,-0.1425531656,0.2608311474,-0.1329194158,-0.3578276038,-0.1220484376,-0.3692784905,0.2719855607,0.5299979448,0.3015776575,-0.2063939273,0.0024211989,0.3558468223,-0.0466872342,0.1243460029,0.1750926822,0.0414709896,-0.0300873742,-0.1414518505,0.0908922628,0.1586921364,-0.0416010357,-0.3086980581,-0.0585056618,0.0238241851,-0.074673824,0.1684126258,0.2123264968,-0.1910576522,-0.0090058353,0.0770116299,-0.2531859875,-0.1229064316,0.4033557475,-0.0854448378,-0.1612051129,-0.2986449897,-0.0300266352,0.1772330552,0.0864015371,0.1985335499,-0.026353687,0.0448872186,-0.2302386761,-0.0440556444,-0.0860254243,0.4295450151,0.5223546624,0.0664971247,-0.1000387594,-0.1619127095,0.2714983523,-0.3283502162,0.071457386,0.3990154266,-0.0608640239,-0.3527769744,0.2800023556,-0.2473477125,0.475312084,0.4118616879,0.2183436602,0.6721669436,0.3472309411,0.6203202605,0.1277313232,0.475355804,-0.1620083749,-0.017395379,-0.1005233079,-0.0788868144,0.0276142135,0.1096018106,-0.1319761723,0.0889560133,0.1767239571,0.0960331336,0.0463646725,0.0985127762,0.0611856617,-0.2404057235,-0.2900931239,0.1539499611,-0.1581514478,0.1574359983,-0.0482134856,-0.1438087523,-0.0824364349,-0.326462388,-0.2867511511,0.2754968703,-0.141487211,0.1955625713,0.2045818567,-0.1971868873,-0.0391067825,-0.028757615,0.2409318835,-0.0701698363,-0.1455816776,-0.2068726122,-0.0692644343,0.0280247983,0.1032120883,-0.1363179386,0.5273317695,-0.1486338824,-0.1375990957,-0.0097949114,-0.0955040231,-0.217352435,0.2621585131,0.1152564734,0.0907626301,0.1473237872,0.0249874983,0.1188034415,-0.4523480833,-0.2321523577,0.1703427881,-0.1349198818,-0.3381385505,0.2134271711,-0.1455636322,-0.0226632133,0.0562408082,0.2497803569,-0.0592680126,0.2341690958,0.6033888459,-0.0211929455,-0.0915532932,-0.143573314,-0.2165425569,0.0541517623,-0.1071855724,0.1656975746,-0.1493014842,0.0013679207,0.0525460094,0.1575044692,0.0081935311,0.2858624756,-0.1662966609,-0.3224808276,-0.3871929049,0.2180494815,-0.1401991397,0.2134718597,0.1630478501,0.617010653,0.0856665149,-0.1259983033,-0.32188797,0.0321272388,-0.2677123547,0.324180305,0.0389137156,-0.2092727721,0.1503251642,-0.0048052659,0.1132221594,0.1361236721,-0.1097702086,-0.258374095,-0.191740647,0.0714030862,0.1857145876,-0.3465097547,0.1661629826,0.0046596178,-0.1557182819,-0.1230325326,0.117200613,-0.056309741,-0.2534221411,0.5090382695,0.1321459264,0.6595722437,0.2523018718,0.0457537584,-0.1782053113,0.1686657667,-0.0889488161,0.1971289515,-0.0100230779,-0.0413222201,-0.3386868536,-0.0331776142,-0.4179776013,-0.1148272082,0.2204306722,-0.3328475952,-0.0377382897,-0.2488700449,0.0429368578,0.1350106597,-0.1134634167,0.026608292,-0.0060438812,0.2527722418,-0.0741036981,0.1623068899,0.2498591393,-0.0320506953,0.2161780149,0.1043511257,0.1092673615,-0.0902838409,0.1097071767,0.1391165406,-0.1979448646,-0.311792165,0.2412803024,0.3260473609,-0.3616290689,0.1162670925,-0.1829433441,0.0854910091,0.4222062826,-0.0515633076,0.0166013688,0.2456998378,0.0994251668,-0.1708704084,0.2237033695,-0.3996602893,-0.2821905315,-0.0317088962,0.1926876158,0.2779779732,0.1008707657,0.0297555309,-0.2888753414,-0.1592281461,-0.0015650663,0.2095531225,-0.397054702,0.0147343781,-0.3463863134,0.1573568583,0.0574848093,-0.3433611095,-0.037702471,0.0445325114,0.1723369956,0.1850241721,-0.2280970365,-0.0942424387,-0.1356519014,0.0157811623,0.2179116905,-0.061141856,0.3229675293,0.367595464,-0.0189267956,-0.2765446603,0.5054782033,0.4357511699,0.2400355935,-0.3785278499,0.0104650129,0.1237666383,-0.1097797677,-0.0272483379,0.0968386903,0.1076465994,0.2009624839,0.225017637,0.1305900365,-0.1786044687,0.3606996238,0.1361265182,-0.0453327298,-0.2492054254,0.3173814714,0.1219383851,-0.1387658566,-0.1831054091,-0.0519299507,-0.3085249066,0.1512973756,0.3084787726,-0.1038086489,0.1319825351,0.1568076015,0.0841484368,-0.0936999395,0.3824259341,0.2785471082,0.3412989974,-0.2391836047,-0.0532665066,-0.1395322829,0.1518156528,-0.1091642231,0.3494437933,0.1335536242,0.1528654844,-0.2013279647,-0.055491142,-0.1758026183,-0.4866576195,0.0382897556,0.0140972175,-0.5309222341,-0.0677651465,0.2167239338,-0.2995285392,0.0809302628,-0.2919172645,0.1038764492,-0.0216722004,0.0582708493,-0.0243856199,0.2547587454,0.0580757819,0.4695756733,-0.1598595679,0.2173871994,0.3523491621,-0.0399275683,-0.2264608741,0.0155344419,-0.1288682073,-0.3512552083,0.0623174794,0.0021564274,0.6310626864,0.0308014378,-0.3426463008,-0.0471834131,0.2966679037,0.0744638145,-0.2258533686,-0.2897467613,0.515901804,-0.3268099427,0.1209048629,-0.0355342515,0.2257708907,0.0469698273,0.1271169484,-0.4208705127,-0.5104087591,0.4589865506,-0.6723730564,-0.2989540994,-0.150011301,0.3177681267,-0.1784148961,0.1541518718,-0.7014427781,-0.1308532357,0.3214225769,-0.3068626225,-0.3737615347,0.1170787886,0.2110434473,0.016910512,0.0359381065,0.5025029182,0.0519079305,-0.3045867383,0.0904337913,-0.3307105601]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/477","title":"Overview.ipynb throws exceptions with nlp 0.4.0","comments":"Indeed we'll update the cola with the new release coming up this week.","body":"with nlp 0.4.0, the TensorFlow example in Overview.ipynb throws the following exceptions:\r\n\r\n\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-48907f2ad433> in <module>\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\n<ipython-input-5-48907f2ad433> in <dictcomp>(.0)\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\nAttributeError: 'numpy.ndarray' object has no attribute 'to_tensor'","comment_length":13,"text":"Overview.ipynb throws exceptions with nlp 0.4.0\nwith nlp 0.4.0, the TensorFlow example in Overview.ipynb throws the following exceptions:\r\n\r\n\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-5-48907f2ad433> in <module>\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\n<ipython-input-5-48907f2ad433> in <dictcomp>(.0)\r\n----> 1 features = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]}\r\n 2 labels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\n 3 labels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n 4 tfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n\r\nAttributeError: 'numpy.ndarray' object has no attribute 'to_tensor'\nIndeed we'll update the cola with the new release coming up this week.","embeddings":[-0.0391482376,0.082511723,-0.0073065483,-0.0530798919,0.0318035781,-0.0482144579,0.8157610297,0.2792399526,-0.0684225112,0.1789912581,0.0316702574,0.2038193345,-0.297031194,-0.0530733988,0.0210223142,-0.4359816313,-0.1079105809,0.286254257,0.0336867906,-0.0560540147,-0.1683885306,0.2695775628,-0.4814034402,0.0699896812,-0.3009614944,-0.2224038243,-0.003463051,-0.1184036061,0.1805594563,-0.3218075335,0.2983432412,-0.0853027105,-0.0729661137,0.2157041878,-0.0001094074,0.1312537491,0.2247508913,0.0399746299,-0.1744725704,-0.3294422925,-0.1835713238,-0.0175784472,-0.0488839783,-0.3077403009,0.1411332339,0.0695705861,0.2543091178,0.2293205559,0.187586844,0.4259967804,0.2343905419,0.3789539039,0.0684154481,0.0336801223,0.0507045202,-0.0231374409,-0.0912747681,0.1140883863,-0.0621383898,-0.1997990906,-0.2145836651,0.2809059322,-0.0087522771,0.0915340558,-0.158366695,0.0241951719,0.0742403716,-0.2711643577,-0.0711959004,0.4140101969,-0.0402942188,-0.0426515564,-0.392829448,-0.3703237176,0.0652718395,-0.5704241991,-0.060038127,0.1048349366,-0.1118555367,-0.016561389,-0.1545907557,0.095754236,-0.5827885866,0.3474161923,-0.1908776015,0.1448999345,0.0082724849,0.081057407,-0.0094359042,-0.0986803472,-0.0254501868,-0.0089679696,0.1772603244,0.0017488351,-0.2649015784,-0.220381543,0.1818951219,-0.206580326,-0.1672124565,-0.0775770694,0.0233217161,0.0268009454,0.0007514657,0.3499961197,-0.122096628,0.2610676587,0.1183250472,0.1576666385,0.3323271275,-0.0056893472,-0.1858868748,0.0312410183,0.0496797599,-0.1437185258,-0.1408727169,0.017316794,0.3466324508,0.1334186941,-0.5548805594,0.0387058556,-0.3659700453,0.1409137547,0.192010656,0.2450779825,-0.0872803405,0.2586591542,0.2062356174,0.1746394187,-0.4506019652,-0.0207848474,-0.2161477357,0.3215353191,-0.054345686,-0.2500003576,0.2410529703,0.32728827,0.3226420283,-0.0850195363,0.0863246098,0.0834727958,0.3203797638,-0.2739877701,0.1113804728,0.2461777329,-0.0800185576,-0.1207126752,0.046420671,-0.1517851949,-0.0383200198,-0.0575329512,-0.2080945373,-0.3066107631,-0.0356256999,0.2675876915,0.1201595962,-0.0608261302,-0.2594278157,0.1365095973,0.0316395611,-0.1947879344,-0.0350830965,-0.3620331287,-0.1259934753,-0.0612786673,0.0401064605,0.190676704,0.1341350079,-0.1566854864,0.1043589339,0.1802005768,0.2523798943,0.2696486115,-0.2044544071,0.2307111025,-0.036194697,0.3748850822,0.3027423024,-0.2861011326,-0.1879547834,0.1153768301,-0.0510694645,0.0571820475,-0.0915255472,0.019735707,0.4229823351,-0.038066525,0.3864845932,0.2939309478,-0.2305163592,0.098540768,-0.0844678134,0.1140950769,0.292221874,-0.0645060986,0.1169267669,0.0630712658,-0.1999797374,0.5250784159,0.0386198647,-0.1206296608,-0.0273567997,0.0281608887,0.0251637865,-0.3499792218,-0.1735033095,-0.1349679232,-0.3476230204,0.0723298341,0.0827099904,0.2156667709,0.3347974122,-0.0584712513,0.0152782639,0.1186521649,-0.077429764,-0.2848463356,0.2511830032,0.1563904881,-0.1726750731,0.0747464448,-0.1370531023,0.0778900459,-0.3810476959,0.1017818525,-0.1399620175,0.0689370707,-0.0984523445,-0.1941973716,-0.2610077858,0.3692515492,-0.0739928558,-0.1785955727,-0.0611250252,0.2158330977,0.1814824939,0.2297656089,-0.2306142896,0.1413954794,0.3241619766,-0.0578077212,-0.1601210982,0.2085707486,0.0083810706,-0.1782839298,0.1594430953,0.4819951057,0.2851334512,0.1752204299,0.4278370738,0.1367619187,0.2068368047,0.0023742397,-0.1972270608,0.08879821,0.3635754287,-0.0712224767,0.1973559558,-0.2939554155,-0.7215968966,0.1118019372,0.549682498,0.1343672872,0.0659737885,0.1962781847,-0.0909676552,0.1029396802,0.3837305307,-0.2072698623,0.1110235453,0.0864473283,0.2527721822,0.0789593309,-0.3717411757,-0.2521877289,0.0808835775,0.1414926648,-0.1022182554,-0.0228342731,0.2730284631,0.2171438932,-0.2615646124,-0.3487101197,-0.1701575369,0.2663602531,-0.3354310095,0.3268645406,-0.1205521524,-0.2741838694,-0.0820815712,-0.5632296801,-0.1426299363,-0.3243647516,-0.0196133722,0.626950562,-0.3355834782,0.2322810441,0.1679948866,0.0385323465,0.0734659433,0.2469550222,0.1575049609,0.0089548556,-0.3124403656,0.1394797713,0.1999257207,-0.1503335834,0.2847865522,-0.0808443576,-0.3065349758,0.2594654858,-0.5440360904,0.1175272763,-0.1258309186,-0.0007881688,0.2060860097,0.2676906288,-0.2906229794,-0.5785192251,0.0555495806,-0.1116683707,-0.1575769037,0.3648533523,-0.0376024731,-0.0444794111,-0.3675981462,-0.5435934067,-0.1716129035,-0.4446734786,0.361694634,-0.0806766227,0.307082206,0.7853786349,-0.0730392784,0.5275722146,0.2695943415,0.1862934232,0.2173419595,-0.0682444572,-0.00816104,-0.2857410312,-0.3695432544,-0.106303297,-0.2888970375,0.1831954867,0.1218610108,-0.329808116,-0.0585052744,-0.0348194987,0.134925127,0.1361687481,-0.1742237657,0.3653981388,-0.0563014857,-0.0302218366,-0.0342406295,0.0573622622,0.3721608818,0.2451589406,-0.0622616112,-0.0394694619,0.2059445679,0.1754864305,0.3692196012,0.0424718671,-0.2425438613,0.2184517384,0.0004712994,0.0349336825,-0.1755885184,-0.276870966,0.2389113009,-0.0171982739,-0.0787892938,0.0796449482,-0.1992204487,-0.294082135,-0.1044809669,0.0583732612,-0.2043749243,-0.3995546997,0.1599550992,-0.3206953704,0.1283507496,-0.0597034842,0.0156895258,-0.215094164,0.0657821,-0.1475070715,0.1715197265,0.0861308575,-0.2596295774,-0.0938589796,-0.5862797499,-0.6067792773,0.2046329379,0.2464426905,0.5263921618,-0.0870106295,-0.018824283,-0.1141606569,-0.0416838527,0.4461074471,0.1069852561,-0.1425531656,0.2608311474,-0.1329194158,-0.3578276038,-0.1220484376,-0.3692784905,0.2719855607,0.5299979448,0.3015776575,-0.2063939273,0.0024211989,0.3558468223,-0.0466872342,0.1243460029,0.1750926822,0.0414709896,-0.0300873742,-0.1414518505,0.0908922628,0.1586921364,-0.0416010357,-0.3086980581,-0.0585056618,0.0238241851,-0.074673824,0.1684126258,0.2123264968,-0.1910576522,-0.0090058353,0.0770116299,-0.2531859875,-0.1229064316,0.4033557475,-0.0854448378,-0.1612051129,-0.2986449897,-0.0300266352,0.1772330552,0.0864015371,0.1985335499,-0.026353687,0.0448872186,-0.2302386761,-0.0440556444,-0.0860254243,0.4295450151,0.5223546624,0.0664971247,-0.1000387594,-0.1619127095,0.2714983523,-0.3283502162,0.071457386,0.3990154266,-0.0608640239,-0.3527769744,0.2800023556,-0.2473477125,0.475312084,0.4118616879,0.2183436602,0.6721669436,0.3472309411,0.6203202605,0.1277313232,0.475355804,-0.1620083749,-0.017395379,-0.1005233079,-0.0788868144,0.0276142135,0.1096018106,-0.1319761723,0.0889560133,0.1767239571,0.0960331336,0.0463646725,0.0985127762,0.0611856617,-0.2404057235,-0.2900931239,0.1539499611,-0.1581514478,0.1574359983,-0.0482134856,-0.1438087523,-0.0824364349,-0.326462388,-0.2867511511,0.2754968703,-0.141487211,0.1955625713,0.2045818567,-0.1971868873,-0.0391067825,-0.028757615,0.2409318835,-0.0701698363,-0.1455816776,-0.2068726122,-0.0692644343,0.0280247983,0.1032120883,-0.1363179386,0.5273317695,-0.1486338824,-0.1375990957,-0.0097949114,-0.0955040231,-0.217352435,0.2621585131,0.1152564734,0.0907626301,0.1473237872,0.0249874983,0.1188034415,-0.4523480833,-0.2321523577,0.1703427881,-0.1349198818,-0.3381385505,0.2134271711,-0.1455636322,-0.0226632133,0.0562408082,0.2497803569,-0.0592680126,0.2341690958,0.6033888459,-0.0211929455,-0.0915532932,-0.143573314,-0.2165425569,0.0541517623,-0.1071855724,0.1656975746,-0.1493014842,0.0013679207,0.0525460094,0.1575044692,0.0081935311,0.2858624756,-0.1662966609,-0.3224808276,-0.3871929049,0.2180494815,-0.1401991397,0.2134718597,0.1630478501,0.617010653,0.0856665149,-0.1259983033,-0.32188797,0.0321272388,-0.2677123547,0.324180305,0.0389137156,-0.2092727721,0.1503251642,-0.0048052659,0.1132221594,0.1361236721,-0.1097702086,-0.258374095,-0.191740647,0.0714030862,0.1857145876,-0.3465097547,0.1661629826,0.0046596178,-0.1557182819,-0.1230325326,0.117200613,-0.056309741,-0.2534221411,0.5090382695,0.1321459264,0.6595722437,0.2523018718,0.0457537584,-0.1782053113,0.1686657667,-0.0889488161,0.1971289515,-0.0100230779,-0.0413222201,-0.3386868536,-0.0331776142,-0.4179776013,-0.1148272082,0.2204306722,-0.3328475952,-0.0377382897,-0.2488700449,0.0429368578,0.1350106597,-0.1134634167,0.026608292,-0.0060438812,0.2527722418,-0.0741036981,0.1623068899,0.2498591393,-0.0320506953,0.2161780149,0.1043511257,0.1092673615,-0.0902838409,0.1097071767,0.1391165406,-0.1979448646,-0.311792165,0.2412803024,0.3260473609,-0.3616290689,0.1162670925,-0.1829433441,0.0854910091,0.4222062826,-0.0515633076,0.0166013688,0.2456998378,0.0994251668,-0.1708704084,0.2237033695,-0.3996602893,-0.2821905315,-0.0317088962,0.1926876158,0.2779779732,0.1008707657,0.0297555309,-0.2888753414,-0.1592281461,-0.0015650663,0.2095531225,-0.397054702,0.0147343781,-0.3463863134,0.1573568583,0.0574848093,-0.3433611095,-0.037702471,0.0445325114,0.1723369956,0.1850241721,-0.2280970365,-0.0942424387,-0.1356519014,0.0157811623,0.2179116905,-0.061141856,0.3229675293,0.367595464,-0.0189267956,-0.2765446603,0.5054782033,0.4357511699,0.2400355935,-0.3785278499,0.0104650129,0.1237666383,-0.1097797677,-0.0272483379,0.0968386903,0.1076465994,0.2009624839,0.225017637,0.1305900365,-0.1786044687,0.3606996238,0.1361265182,-0.0453327298,-0.2492054254,0.3173814714,0.1219383851,-0.1387658566,-0.1831054091,-0.0519299507,-0.3085249066,0.1512973756,0.3084787726,-0.1038086489,0.1319825351,0.1568076015,0.0841484368,-0.0936999395,0.3824259341,0.2785471082,0.3412989974,-0.2391836047,-0.0532665066,-0.1395322829,0.1518156528,-0.1091642231,0.3494437933,0.1335536242,0.1528654844,-0.2013279647,-0.055491142,-0.1758026183,-0.4866576195,0.0382897556,0.0140972175,-0.5309222341,-0.0677651465,0.2167239338,-0.2995285392,0.0809302628,-0.2919172645,0.1038764492,-0.0216722004,0.0582708493,-0.0243856199,0.2547587454,0.0580757819,0.4695756733,-0.1598595679,0.2173871994,0.3523491621,-0.0399275683,-0.2264608741,0.0155344419,-0.1288682073,-0.3512552083,0.0623174794,0.0021564274,0.6310626864,0.0308014378,-0.3426463008,-0.0471834131,0.2966679037,0.0744638145,-0.2258533686,-0.2897467613,0.515901804,-0.3268099427,0.1209048629,-0.0355342515,0.2257708907,0.0469698273,0.1271169484,-0.4208705127,-0.5104087591,0.4589865506,-0.6723730564,-0.2989540994,-0.150011301,0.3177681267,-0.1784148961,0.1541518718,-0.7014427781,-0.1308532357,0.3214225769,-0.3068626225,-0.3737615347,0.1170787886,0.2110434473,0.016910512,0.0359381065,0.5025029182,0.0519079305,-0.3045867383,0.0904337913,-0.3307105601]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/474","title":"test_load_real_dataset when config has BUILDER_CONFIGS that matter","comments":"The `data_dir` parameter has been removed. Now the error is `ValueError: Config name is missing`\r\n\r\nAs mentioned in #470 I think we can have one test with the first config of BUILDER_CONFIGS, and another test that runs all of the configs in BUILDER_CONFIGS","body":"It a dataset has custom `BUILDER_CONFIGS` with non-keyword arguments (or keyword arguments with non default values), the config is not loaded during the test and causes an error.\r\nI think the problem is that `test_load_real_dataset` calls `load_dataset` with `data_dir=temp_data_dir` ([here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/tests\/test_dataset_common.py#L200)). This causes [this line](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/builder.py#L201) to always be false because `config_kwargs` is not `None`. [This line](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/builder.py#L222) will be run instead, which doesn't use `BUILDER_CONFIGS`.\r\n\r\nFor an example, you can try running the test for lince:\r\n` RUN_SLOW=1 pytest tests\/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_lince`\r\nwhich yields\r\n> E TypeError: __init__() missing 3 required positional arguments: 'colnames', 'classes', and 'label_column'","comment_length":43,"text":"test_load_real_dataset when config has BUILDER_CONFIGS that matter\nIt a dataset has custom `BUILDER_CONFIGS` with non-keyword arguments (or keyword arguments with non default values), the config is not loaded during the test and causes an error.\r\nI think the problem is that `test_load_real_dataset` calls `load_dataset` with `data_dir=temp_data_dir` ([here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/tests\/test_dataset_common.py#L200)). This causes [this line](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/builder.py#L201) to always be false because `config_kwargs` is not `None`. [This line](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/builder.py#L222) will be run instead, which doesn't use `BUILDER_CONFIGS`.\r\n\r\nFor an example, you can try running the test for lince:\r\n` RUN_SLOW=1 pytest tests\/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_lince`\r\nwhich yields\r\n> E TypeError: __init__() missing 3 required positional arguments: 'colnames', 'classes', and 'label_column'\nThe `data_dir` parameter has been removed. Now the error is `ValueError: Config name is missing`\r\n\r\nAs mentioned in #470 I think we can have one test with the first config of BUILDER_CONFIGS, and another test that runs all of the configs in BUILDER_CONFIGS","embeddings":[-0.2808128297,0.0386497863,0.0163441859,0.1777213067,0.058332853,0.2016587704,0.0901036114,0.3195054829,-0.1677363366,0.0523618609,0.042073112,0.357843399,-0.1656038463,-0.0654388294,0.1720553786,0.0222495254,-0.3066465855,0.354832232,0.1001126543,0.1028401852,-0.1480746418,0.1693325341,0.0237269141,0.0275578331,-0.1753005832,-0.0250488333,-0.2061964273,0.3140675724,0.2690215111,-0.5922114253,0.425737828,0.1297422051,-0.1612460613,0.1231278181,-0.0001130642,0.1107812375,0.2472267151,-0.1157406643,-0.5981776118,-0.1378033608,0.0640317202,-0.2330078781,0.2686037719,-0.3578184545,-0.2629135251,-0.2691299319,-0.0671395734,0.0914066806,0.1089665443,0.3557496965,0.2239871174,0.2617229521,-0.4300852418,-0.113718845,-0.1765928119,-0.0511278436,0.0575864986,0.3311009109,0.0536821112,0.0526558831,0.1169041097,-0.1423465014,-0.0944502577,0.1072169021,0.0086366739,-0.0255859867,0.0688417554,-0.0536702201,-0.0848884284,0.2826161385,0.1915073693,-0.4512461424,-0.0090727797,-0.3596467078,0.2316243201,-0.0907495618,0.3416333497,-0.1556084007,-0.4267599285,0.0131679038,-0.4288163185,-0.2629853487,0.1304967254,0.0926602334,0.1009298712,0.0209593885,0.0291006658,0.2418982983,-0.0891791135,0.1615595967,-0.1071039736,-0.3283272684,-0.1013362929,0.0169518851,-0.0456780195,-0.2923914194,0.2084433436,0.0993104652,0.2808901966,-0.0533519275,-0.0385394134,-0.1233905256,0.2437750101,0.1774764955,-0.1193953082,0.3732145131,0.4342042804,0.059252888,0.1818019152,-0.0008896501,-0.1197538599,0.0659153834,0.1539162099,-0.3203549087,-0.1054672748,-0.0126288449,0.3360986114,-0.3401197791,-0.3468776345,-0.0612612963,-0.062299829,0.1099945158,0.3181942701,0.3967021704,-0.0057034586,-0.1252384633,-0.0159600694,0.2539609969,-0.0674445704,-0.0837189704,-0.3294445574,-0.0484220423,-0.0665614083,0.1366144568,0.4345739186,-0.3286035955,0.6601085067,0.2206451893,0.0609017126,0.0424178317,0.0439316519,-0.0425020121,-0.0450807549,0.3762796819,-0.137886405,0.0346907042,0.0766118392,-0.1368466765,-0.4026349187,0.2292724848,-0.1443717033,-0.52518785,0.3482974768,0.179026559,-0.3709740043,0.0275205392,0.0139297089,0.0734271854,0.0692850649,-0.2042810321,-0.1741871983,-0.3507989347,0.0237939265,-0.2072624117,0.2525056899,0.3057611883,-0.1314186901,-0.2009166777,0.1866978854,-0.0990063772,0.117111221,-0.172343567,-0.3042074442,0.2274881303,-0.2789945006,0.0282068755,0.6426609755,-0.3164250255,-0.0749777108,0.3298943937,-0.0031589051,0.164363265,0.0045769601,-0.0373653956,-0.2150313556,-0.1312565506,0.0781562403,0.2976283431,-0.0622226521,-0.0454830453,-0.3610289991,-0.1531265676,0.1661152244,0.0698524788,-0.2492715269,-0.034018252,-0.2962556481,-0.0673040375,0.2288819402,-0.0425872952,-0.2239330113,0.0862083733,-0.0886953026,-0.1204980314,-0.006259515,0.1013942957,-0.611117959,0.461247623,0.0510441214,0.1673367172,0.1060817167,-0.0808744878,-0.221636951,-0.0650239959,-0.2754856646,-0.3786343932,0.1259016395,0.4341852665,0.1042221561,-0.2089665234,-0.1609920412,-0.0981572941,-0.3174370229,-0.037504483,-0.0407108702,0.0693355501,-0.0057057119,-0.025555905,-0.1366919875,0.1385820359,0.1388397366,-0.2931470573,-0.0865144283,0.5086601973,-0.1005271673,-0.1135223508,0.067047976,-0.0492178053,0.0144278808,-0.0268902183,-0.0401120745,0.1391652077,-0.0228069499,-0.1525020152,0.0116877975,0.4082338512,0.1521519572,0.2619357109,-0.0039576483,-0.0580006763,0.2521714568,-0.2402492762,-0.0883734822,-0.1512288153,0.439422369,0.1059501097,0.4570346475,0.0323099867,-0.3409210145,-0.0486268103,0.2273120731,-0.0086172642,-0.1118612066,0.1671793163,-0.1314900815,0.0118462201,0.0743678957,-0.0071038129,0.6731563807,0.1936108768,0.0327809304,0.1103147343,-0.5834096074,-0.2602223158,0.3016805351,-0.035758961,0.2742649019,0.1225185916,0.2030345947,-0.0732069761,-0.4160089493,-0.3327857852,0.2195360065,0.2261593491,-0.3869549334,0.1408006698,-0.080101788,-0.0013913615,-0.2797278464,-0.3139405549,-0.0723810717,-0.3478244245,0.2141195536,-0.089001216,-0.213189736,0.2064433396,-0.0489376001,0.2271481603,-0.2309503704,-0.5881280303,-0.3020557165,-0.1458092183,-0.4142384231,0.0414332487,0.1540944874,-0.0204575807,0.2949227393,-0.1228463799,-0.5448807478,0.103290759,-0.3132393956,0.209128201,-0.0427241735,0.7572254539,0.3080534041,0.1065723971,0.4649317265,-0.2761743665,0.4519259036,0.1219889447,-0.1166918278,0.2048782259,0.107731916,-0.0862267613,-0.3488543034,-0.4813729227,-0.0115978047,-0.2266955972,-0.1176695898,0.1248537153,0.1642356068,0.4464901984,0.1673844457,0.2023621053,-0.1938327998,0.4660231173,-0.0798346773,-0.1949100792,-0.1284496784,-0.0745197013,-0.1809367537,-0.3876855075,-0.1372406334,0.3511223793,-0.1121397689,-0.4986842275,-0.2527987957,-0.1208135262,0.4919119775,-0.2137345076,-0.2025675923,0.2597002685,0.3100275099,-0.0842043161,-0.1884583235,0.0413901173,0.2889184654,0.111712195,0.2339570671,0.0027118395,0.2264725417,-0.1682234854,0.3377180099,0.3033349514,-0.2950260937,0.1454984695,-0.2538925409,0.2515290678,-0.182641685,-0.3044295907,0.1686677635,0.1307938695,-0.1615768224,0.3201492727,-0.1414804906,-0.044294063,-0.1323329806,0.0162521433,0.0394280478,-0.1494450718,0.2298580855,-0.19080697,0.0086010732,0.0478129163,0.0411837362,-0.0809965283,0.0466443412,0.042565681,0.2682785988,-0.0543797426,-0.0927187055,-0.2088309079,-0.3003647923,-0.4212277532,0.1644621789,-0.0991439298,0.3014188707,-0.046743203,-0.0265601147,0.1795356274,0.000533039,0.4785411358,-0.0032723229,0.1700183451,0.4871151149,0.0659948289,-0.3755955696,-0.1330588907,-0.295486033,0.1635190845,0.378821671,0.6604258418,-0.3550482094,-0.4132323861,0.0995786414,0.4210105538,-0.1064963788,-0.3035877049,-0.2665137947,-0.3496364057,-0.1467337012,0.0701794922,-0.3176826239,0.3553071916,-0.3038996756,0.0568311289,-0.0653927103,-0.0897990465,0.0434089191,0.3618538678,0.2549633682,-0.1290694773,0.2132919878,0.0077910186,0.1998325139,0.0466913655,0.4059337974,-0.0683811605,0.283271879,-0.0779445767,0.1865792125,0.5016235709,0.277615279,-0.1241765916,-0.0561816953,-0.3990889788,0.1736404151,-0.2655557394,0.2824919224,0.3087879419,0.0439320058,0.0081730345,-0.2562392354,0.5897767544,-0.026384674,-0.2228999287,0.4522735178,-0.093740277,-0.4636976123,0.3889221847,-0.0300232284,0.760294795,0.3173417449,0.2274958491,0.6086381674,0.2118727565,0.1707531363,0.1247127354,0.1548669189,-0.2730989754,-0.0845781639,0.2797228396,-0.1793148965,0.1964721233,0.2276956141,-0.2754840553,0.2837625444,-0.0638538525,-0.0449198484,0.1381046623,0.2204245925,-0.0480701476,-0.0373605415,-0.4018090069,0.1581183374,0.0321190245,0.2868534029,0.0300637949,0.053051576,0.0524244159,-0.255284667,0.0507867262,0.1388699412,-0.3315854371,-0.0008888579,0.1234661043,-0.1947832406,0.0131060863,0.1538493037,0.1285512298,0.2766211629,-0.1290299892,0.1358105242,0.0137005793,0.1033774018,0.1839089543,-0.0896471888,0.2951235175,-0.1704012901,-0.2964981496,0.0602017082,-0.3801342249,-0.0585321561,0.2786425054,-0.16151838,-0.0661644563,-0.3292330205,-0.3309015632,0.0869205073,0.1180754304,-0.2144934088,0.1361999512,-0.0282535143,-0.3646183312,0.1135281101,-0.0767538771,-0.3893271685,-0.0756363794,-0.106943883,0.0940754563,0.2095009834,0.6270083189,-0.3351728618,0.2477501184,-0.1618227661,0.0379427671,0.1511998177,-0.1700678617,0.1599756032,0.2201096416,-0.2147943974,0.0113862269,0.4493312836,0.1310528219,0.1979993284,-0.1519994885,-0.2812598348,-0.1871857494,0.0357021913,0.2232614607,0.31444332,-0.0407342762,0.5240074992,-0.0977954641,0.2897000313,-0.3693549931,-0.0361141302,-0.4362352192,0.3607875109,-0.0845117345,0.2169929743,0.2612893879,0.141424194,0.1556659788,-0.2430790067,-0.0575820953,-0.2900200486,-0.1918907613,0.1262518317,0.0445535518,0.1527551264,0.0073162247,-0.1333147734,0.1083943695,-0.1033813506,0.1381946951,-0.3329194784,-0.006425539,0.2893799543,0.2476063818,0.1590152681,-0.0521149859,0.2165991366,-0.2068298161,-0.0528036542,-0.0547473654,0.2293614447,-0.4055795074,-0.1391334087,-0.0639213249,0.0480666086,-0.2943244874,0.1606639326,-0.1572329253,0.0150282783,0.0451961681,0.0703844726,0.2515389323,0.5024479628,-0.0748998597,0.056421984,0.1374621093,0.2301279306,-0.3603684306,-0.050132148,0.0565462112,-0.1226828545,0.3516516685,0.1919915974,0.1136415377,-0.0524165481,0.211417228,-0.1112443134,0.117349878,-0.1513699442,-0.0625124723,0.3782608211,-0.1077270508,0.1377432793,0.281750381,-0.0336467475,0.1510791779,0.071100533,-0.2178883553,0.3636238575,0.0021124484,-0.1638921648,0.1992930174,-0.1242934391,0.447861135,0.0578688644,0.0850253925,-0.1152183115,-0.166311875,0.5778751969,-0.5218051672,-0.2537317276,-0.0446054824,0.1963032484,-0.150412336,-0.2721489668,-0.4087147415,-0.1099317074,0.0089789266,-0.139512077,-0.0893709883,-0.2711934745,0.0749697089,-0.0390519612,-0.1797074378,-0.4126122594,-0.0326081961,0.3341915607,-0.0930392668,-0.1596977115,0.4812886119,0.2137865722,-0.0185972545,0.0896015093,0.4621696174,0.5086627603,0.1307404339,-0.078073673,-0.188678056,-0.0380982049,0.0525451601,-0.2050904632,0.1065383852,0.0240305737,-0.2988906205,0.4634021223,0.2007218301,-0.1836306006,-0.1176309735,0.1081554294,-0.1675929129,-0.1707926095,0.5300661922,0.1958314031,-0.097841382,-0.0381044634,0.1672277451,-0.2096356153,0.0716116652,0.4497458935,0.1866221726,0.1008023545,0.1261885911,0.1134095266,-0.1322555393,0.5245953202,0.0897293612,-0.0436541513,-0.5897715688,-0.1635481119,-0.6522300243,0.1591015309,-0.0294817351,-0.2246663272,0.2934619188,0.0248232167,0.2001622766,-0.0379630663,-0.158332184,0.2733855844,0.3098800182,-0.1415099651,-0.3142865896,-0.2356010675,-0.264275372,-0.0255268458,0.0640600398,-0.2706092894,0.0211015306,-0.1963492334,0.0494559929,0.1104649827,0.1145595759,0.5461165309,0.1187748462,0.2451619655,-0.0055927103,0.5098612905,0.0337561145,-0.1479347944,-0.0083978651,-0.1826662868,-0.1600972116,-0.103449598,-0.093094565,0.0018282627,-0.1612288356,0.1169153899,-0.2332955003,0.0622347109,0.1686074287,-0.2084342092,0.1038310379,0.108446382,-0.1600910425,0.2272068709,0.225565508,0.4649651349,-0.106659323,0.1099360585,-0.3157237768,-0.3565710783,0.7502676249,-0.1635053754,-0.2016156018,-0.0326655395,0.1670453548,-0.1902435273,0.0705953166,-0.3250292242,-0.2004292905,0.4116537273,-0.1168966144,-0.2372823507,0.3358038068,0.1896924973,0.082022652,0.0655163005,0.3776127696,0.0757652447,-0.2050289959,-0.1730704457,-0.3017949164]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/474","title":"test_load_real_dataset when config has BUILDER_CONFIGS that matter","comments":"This was fixed in #527 \r\n\r\nClosing this one, but feel free to re-open if you have other questions","body":"It a dataset has custom `BUILDER_CONFIGS` with non-keyword arguments (or keyword arguments with non default values), the config is not loaded during the test and causes an error.\r\nI think the problem is that `test_load_real_dataset` calls `load_dataset` with `data_dir=temp_data_dir` ([here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/tests\/test_dataset_common.py#L200)). This causes [this line](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/builder.py#L201) to always be false because `config_kwargs` is not `None`. [This line](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/builder.py#L222) will be run instead, which doesn't use `BUILDER_CONFIGS`.\r\n\r\nFor an example, you can try running the test for lince:\r\n` RUN_SLOW=1 pytest tests\/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_lince`\r\nwhich yields\r\n> E TypeError: __init__() missing 3 required positional arguments: 'colnames', 'classes', and 'label_column'","comment_length":18,"text":"test_load_real_dataset when config has BUILDER_CONFIGS that matter\nIt a dataset has custom `BUILDER_CONFIGS` with non-keyword arguments (or keyword arguments with non default values), the config is not loaded during the test and causes an error.\r\nI think the problem is that `test_load_real_dataset` calls `load_dataset` with `data_dir=temp_data_dir` ([here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/tests\/test_dataset_common.py#L200)). This causes [this line](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/builder.py#L201) to always be false because `config_kwargs` is not `None`. [This line](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/builder.py#L222) will be run instead, which doesn't use `BUILDER_CONFIGS`.\r\n\r\nFor an example, you can try running the test for lince:\r\n` RUN_SLOW=1 pytest tests\/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_lince`\r\nwhich yields\r\n> E TypeError: __init__() missing 3 required positional arguments: 'colnames', 'classes', and 'label_column'\nThis was fixed in #527 \r\n\r\nClosing this one, but feel free to re-open if you have other questions","embeddings":[-0.3385565877,-0.0536466315,0.0424275808,0.1108628884,0.1634410322,0.0783360898,0.1228836551,0.3306728005,-0.0973314494,-0.0040300302,0.0537734479,0.4402484596,-0.1571183205,0.0414613187,0.1643651128,0.0443918817,-0.2682584524,0.2769702077,0.0980256796,-0.0319090597,-0.1879016161,0.1579577476,-0.0973070934,0.0949115977,-0.2986759841,0.0340311453,-0.1379248351,0.2986700535,0.1990152597,-0.5106943846,0.4461774826,0.0938677639,-0.0859578028,0.1431889981,-0.0001083159,0.0801491439,0.3250587583,-0.0839207619,-0.637114048,-0.0625694469,0.1405450553,-0.2929130793,0.2855110765,-0.3282714188,-0.3212667704,-0.2411285639,-0.1270383745,0.0736482516,0.0725026578,0.3807226121,0.2747091055,0.3636160195,-0.2297531366,-0.1485881507,-0.1620922387,-0.0258874465,-0.0437101349,0.3394280672,0.1688621491,0.0183412638,0.0527107343,-0.1217424348,-0.0641052276,0.0916210115,0.0495265983,-0.1177207232,-0.0773677304,-0.0686179549,-0.0201732405,0.2787656188,0.2818546593,-0.4329428077,-0.0211793836,-0.2513436973,0.1684996784,-0.1498256773,0.3519796133,-0.081268847,-0.4076192379,0.0426622964,-0.4676843584,-0.1900425106,0.0658134222,0.1104304567,0.0950841829,-0.0414986461,-0.0108464593,0.2007044852,0.0036649953,0.1042253673,-0.1386699528,-0.176100865,-0.1310182065,0.127907902,-0.1391436011,-0.3100568354,0.2468237728,0.2167610675,0.2853586078,0.0043460596,-0.0273894947,0.0024011645,0.2310737222,0.2241033465,-0.1374670118,0.3698545098,0.3622436821,-0.0116292955,0.1981758475,0.010260568,-0.1906317323,0.0484408103,0.1153854132,-0.3490093052,-0.0083672144,-0.0628785267,0.2651869655,-0.4105986357,-0.3491712511,0.0293950774,-0.055791311,0.165577814,0.3375837207,0.4199337661,-0.0324665643,-0.1970600188,0.0237554982,0.3201874793,-0.1912802011,-0.1085969433,-0.3206711113,-0.0218131747,-0.1324210614,0.0523711406,0.4688149393,-0.2727923095,0.6226427555,0.1460936219,0.0216841716,0.0165438391,0.0746406168,-0.0024734484,-0.0568611361,0.3785873055,-0.1605929583,-0.0764332563,0.2329165339,-0.1536602527,-0.3517065644,0.2044932842,-0.0884962678,-0.5394568443,0.2401414067,0.2285760939,-0.3029271066,-0.0736225322,0.0094516249,0.0187685583,0.0577064641,-0.1259180307,-0.1104718149,-0.3919752538,0.0619846098,-0.1528332233,0.2685030401,0.2812782228,-0.1850993037,-0.1322146803,0.1188690141,-0.143509388,0.1952687651,-0.0970532298,-0.2381485105,0.1500054151,-0.1737826616,0.1072408408,0.5219545364,-0.2305840254,-0.1180210561,0.2901421785,-0.0084151104,0.0573636927,-0.1372818202,0.0773709044,-0.1516706347,-0.037383344,0.0596372448,0.3409302235,0.0408540182,0.0341271013,-0.4409683347,-0.1354342848,0.1570859104,0.1454819739,-0.346373409,-0.0475006253,-0.146941334,-0.0601771884,0.199085027,-0.0826975927,-0.2510528862,0.0619646609,0.0419992507,-0.0967449993,0.042194359,0.0955780298,-0.556489706,0.4191554189,0.0833571851,0.2934090197,0.158155486,-0.0835636482,-0.2718794048,0.0456508473,-0.2793190479,-0.3855575025,0.1636399031,0.2197535187,0.1555460095,-0.1151784882,-0.1425247043,-0.0953576714,-0.3342646956,0.0108676804,-0.0560782999,0.0678906813,-0.0168093368,-0.0625691488,-0.074856855,0.2662242055,0.1372029185,-0.2954993546,-0.2036425918,0.4747344553,0.0138577111,-0.0961078107,0.107180208,-0.0413774624,-0.0719467476,-0.0976847112,0.0078976294,0.1678729057,-0.0173534788,-0.1123137698,-0.046283327,0.3852535784,0.1947698593,0.17417337,-0.081750229,0.0258382596,0.341634661,-0.1674983799,-0.0014194042,-0.1094363108,0.495426178,0.1282527,0.4595437348,0.0573717766,-0.3472210169,0.0108899977,0.279953748,0.0147187952,-0.0306053627,0.3178577125,-0.1561895907,-0.0745468959,0.1391989291,-0.0617589951,0.5611637235,0.1805538386,-0.0291944649,0.1363343,-0.6024135947,-0.2829419971,0.3375745118,-0.1507919133,0.2357178032,0.0174277071,0.1795159876,-0.1059577316,-0.5612171292,-0.2691088319,0.0600530989,0.3444687724,-0.31918329,0.104244478,-0.1219017431,0.0684502944,-0.1120392606,-0.2230800092,-0.1525081098,-0.3797399998,0.2065999806,0.019258704,-0.0720106736,0.2248980701,-0.064521417,0.2702748775,-0.1633848846,-0.5609422326,-0.3093764484,-0.2441608906,-0.5122612119,0.0824075118,0.1711145639,-0.0119708125,0.3132204115,-0.0825492069,-0.3603592515,0.1028654799,-0.329537183,0.1290487647,-0.0900543705,0.7290681601,0.2706765831,0.1940115541,0.3838451207,-0.1142290682,0.5275809169,0.0100331483,-0.1275102943,0.281848222,0.1550060958,-0.1615920663,-0.3950868845,-0.4603402317,-0.0449561626,-0.251778543,0.0392451547,0.1456494033,0.0847620443,0.4176315665,0.2565802038,0.2545304894,-0.1059367731,0.3779492676,-0.1628271192,-0.2203240842,-0.0969339535,-0.0557349101,-0.3735413253,-0.3066041768,-0.099149026,0.3467929065,-0.1125307754,-0.4803193212,-0.2475796938,-0.2052383423,0.4086900651,-0.0434773453,-0.2014959753,0.2554351985,0.2472823262,-0.1151552126,-0.1670496315,0.0405262187,0.2664709091,0.0364575721,0.1248781085,-0.0117564164,0.3133134544,-0.185517028,0.2949364185,0.2942859232,-0.3394878507,0.2867256701,-0.2529634535,0.327739805,-0.2344418019,-0.322924614,0.2463134229,0.0871208012,-0.1973433644,0.2904693782,-0.0703137591,-0.1744456142,-0.1309032142,0.0823346227,0.0625959933,-0.1458005011,0.1759445667,-0.0543449149,0.0276567824,0.0180719327,0.1418788731,-0.0979135334,-0.0989708602,0.0257286057,0.2842672765,-0.0789645463,-0.1229357049,-0.256139487,-0.3128241003,-0.4667945802,0.2327456176,-0.1251157522,0.3888898194,-0.1696741581,-0.0531123355,0.1999863088,-0.0648229048,0.4661886394,-0.0625413433,0.175284043,0.4553387463,0.0648050532,-0.4149452746,-0.0980449319,-0.3053014874,0.1492465734,0.3097378314,0.6342876554,-0.2931162119,-0.3505549133,0.087817803,0.386518836,-0.0853345022,-0.2622833848,-0.2818942964,-0.3366252184,-0.1886927187,0.1518344879,-0.3122467697,0.392604053,-0.3093537092,0.0417596661,-0.0965659022,-0.076850839,0.0746121779,0.4339574873,0.213289395,-0.1293460578,0.1532240361,-0.0502420701,0.1660062075,0.0519781373,0.3848619461,-0.0900924206,0.2654480338,-0.1243425161,0.0878723934,0.4876841009,0.2203260511,-0.1677978635,-0.110847719,-0.3851255178,0.2015215755,-0.3323498964,0.2147592157,0.339014709,0.1325270683,0.0496402644,-0.2370405793,0.5995282531,-0.0255442597,-0.1007100567,0.375469476,-0.0649465993,-0.3394945264,0.3482230604,0.0802409872,0.7891100049,0.2262751162,0.2127138823,0.6046993136,0.2290666848,0.1493351609,0.0640855804,0.1478568166,-0.3400282264,-0.0823621973,0.229773581,-0.1954124719,0.2035363913,0.185327217,-0.2958571911,0.1574375778,-0.115439415,-0.1364897639,0.132516697,0.2557715476,-0.0615234934,-0.0725658163,-0.3598587215,0.2513634861,0.0307876188,0.3629925549,0.0453267656,0.064253822,0.0008849986,-0.2537640333,-0.0929714292,0.1888664663,-0.2171536386,-0.0174509827,0.1472246796,-0.1967605501,0.0516488515,0.0782263801,0.1751749367,0.2904871106,-0.1754975319,0.0524091721,0.0459554382,0.0707390681,0.282741189,0.0073892996,0.3172931969,-0.1767940819,-0.2712947428,0.052704256,-0.3210828602,-0.0509176701,0.2848851085,-0.2373489588,-0.036734622,-0.3911944032,-0.3904461861,-0.0422729068,-0.0068267523,-0.1889247745,0.1713588387,0.0257089492,-0.24600555,0.1252579093,0.0336853862,-0.3654953837,-0.0639909133,-0.0301253647,0.0257311817,0.262596041,0.6146235466,-0.2057127506,0.1340368539,-0.2302792072,0.024984736,0.180556044,-0.2168991864,0.1104892939,0.258572638,-0.2201317102,0.0646209344,0.4208984375,0.0896682739,0.1069263443,-0.1108952984,-0.3891231716,-0.2301580906,0.0464073047,0.1363064945,0.3031693101,-0.0548467897,0.5074740648,0.0490083657,0.155522272,-0.4011349082,-0.0861839801,-0.4393151402,0.2175886035,-0.0180111155,0.1239899918,0.2580363154,0.0669384822,0.1721222401,-0.1851353645,-0.1733320653,-0.3309585154,-0.216824621,0.0949628502,0.109759748,0.0917482749,-0.0415183604,-0.1717210412,-0.0249398407,-0.184597522,-0.0612857193,-0.3775550127,-0.0767610669,0.2772757411,0.2682214677,0.1808112413,-0.0327042043,0.1179186702,-0.1616005749,-0.0552701578,-0.1441792846,0.0326439478,-0.314424485,-0.1329372972,-0.1057775766,0.0905716792,-0.2027169913,0.1687099189,-0.0648824871,0.0053157373,0.0613972023,0.0235369205,0.2797965109,0.4092744589,-0.0631125346,-0.0312698148,0.2344020456,0.2600903511,-0.4162306786,-0.0071469373,0.1063606143,-0.0779239461,0.3318062723,0.1322983354,0.0769869164,-0.0115646217,0.1614895463,-0.0746058002,0.1683527976,-0.2032975256,-0.1042971835,0.3018105626,-0.1390677392,0.1760900319,0.3138335049,-0.0177445672,0.1356622875,0.1255775094,-0.119480215,0.351634562,0.0769038722,-0.0118937325,0.0888273716,-0.1562257111,0.4014158547,0.0090341242,0.1210858822,-0.0861207843,-0.1869275421,0.6330305934,-0.5700560212,-0.2774113119,-0.1221755743,0.1877730936,-0.2250175476,-0.1661674678,-0.3115476668,-0.0710883662,-0.0374412611,-0.1971277297,-0.0877318159,-0.1883586794,0.0391740985,-0.0581153966,-0.0882006809,-0.4211507142,-0.1079904512,0.3032979071,-0.0499657393,-0.0625258908,0.4666863382,0.0907623395,0.06655287,0.0129099395,0.4504221976,0.5012670159,0.2217959911,-0.1015418246,-0.1400892287,-0.158622548,0.0515245907,-0.1498552114,0.2292057425,0.0469560698,-0.2733751535,0.4103941619,0.2153731585,-0.2012120634,-0.1290589422,0.0607400611,-0.088523306,-0.120994471,0.4698430598,0.2086050659,-0.1601248831,-0.1643498838,0.0871341303,-0.204184711,0.1490507424,0.4095197916,0.1968949437,0.1260686815,-0.0518322587,0.1399005204,-0.0844634548,0.6105009317,0.1193031669,-0.0350407436,-0.5040110946,-0.1553121507,-0.7108892798,0.1991672516,-0.1437133551,-0.3385183215,0.2648856342,-0.0615065843,0.2010038048,-0.0105591752,-0.1847457141,0.2963285148,0.2760757804,-0.1557746083,-0.2949592769,-0.1401153654,-0.2625838816,-0.0780635849,0.031190468,-0.3317849636,0.0586188547,-0.2389077097,0.0939721093,0.0648314506,0.0764858276,0.421223551,0.214362517,0.1641712487,-0.0429045931,0.4837388992,0.0438734666,-0.0726623684,-0.0829004198,-0.2332598418,-0.1337160766,-0.0350031219,-0.1012427509,-0.0198267773,-0.0553851128,0.0709223449,-0.3084037602,0.1950899661,0.1493379176,-0.206478104,0.0872708112,0.1054441407,-0.1367052346,0.1767999828,0.2128749788,0.4557380676,-0.1234279498,0.1840361506,-0.2646231055,-0.4019247591,0.7486212254,-0.2990206182,-0.1954526901,0.0654030517,0.139028728,-0.1361296922,-0.0232064947,-0.231041044,-0.1156558916,0.3482373953,-0.1259844899,-0.1509616226,0.3325615823,0.1050182879,0.0815673023,0.0588990077,0.4525281191,0.0271408577,-0.0806582645,-0.1114689484,-0.2522569895]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/469","title":"invalid data type 'str' at _convert_outputs in arrow_dataset.py","comments":"Hi ! Did you try to set the output format to pytorch ? (or tensorflow if you're using tensorflow)\r\nIt can be done with `dataset.set_format(\"torch\", columns=columns)` (or \"tensorflow\").\r\n\r\nNote that for pytorch, string columns can't be converted to `torch.Tensor`, so you have to specify in `columns=` the list of columns you want to keep (`input_ids` for example)","body":"I trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n","comment_length":57,"text":"invalid data type 'str' at _convert_outputs in arrow_dataset.py\nI trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n\nHi ! Did you try to set the output format to pytorch ? (or tensorflow if you're using tensorflow)\r\nIt can be done with `dataset.set_format(\"torch\", columns=columns)` (or \"tensorflow\").\r\n\r\nNote that for pytorch, string columns can't be converted to `torch.Tensor`, so you have to specify in `columns=` the list of columns you want to keep (`input_ids` for example)","embeddings":[-0.2228239626,0.0242501777,0.1073713675,0.2303070426,0.5109525919,-0.1072595492,0.5323296189,0.0775902644,-0.3410620093,-0.1247765422,0.036657881,0.3864071369,-0.3696954846,0.0422176793,0.0443617031,-0.2024641186,-0.0880346373,0.0099904612,-0.0232526027,-0.0536389053,0.0482440926,0.1829803139,-0.1568261683,0.4144774675,-0.6451317668,-0.1835785806,0.1746639758,-0.0963928327,-0.0992212296,-0.4487785995,0.2202744484,-0.1020444036,0.4976367652,0.3472318351,-0.0001241712,0.1380359232,0.1092858762,0.1016949043,-0.2908720374,-0.3016375303,0.1341145188,-0.0593848564,0.4149942696,-0.2015384734,-0.0571896322,-0.2600911558,-0.0606259108,0.2617689073,0.3837757409,0.5016983747,0.1173220798,0.1685167998,0.2464857996,0.0738129616,0.2595641315,0.3155063391,0.0754768178,0.2071238756,0.0168337896,0.0266056098,-0.0856947526,0.0748076811,-0.0701284856,-0.1997739971,0.2166583389,0.227282986,0.17457892,-0.1476221532,-0.2287852317,0.1198840439,0.4332152605,-0.3214256167,-0.1559445858,0.0079212757,-0.0111622717,-0.2727991641,-0.0061669713,0.2483308762,-0.4277769327,0.1667014062,-0.1622606069,-0.0254668966,-0.3792498708,0.199556306,-0.0925177336,0.5032640696,0.1183697209,0.1568167061,0.1842403114,-0.0144359078,0.371710062,0.0876382813,0.1397552341,0.0621479377,-0.3116942644,-0.030463824,-0.4463334084,-0.648362577,-0.2685149014,-0.2125236392,0.1655772626,-0.0711523443,0.2047266066,0.202950567,0.233322382,0.1388006508,-0.2101232409,0.5648387074,0.1000927314,-0.1552505791,-0.1335455626,-0.0250890013,-0.3282387555,-0.0922897011,0.5121243596,0.1628403068,0.1257427186,-0.0050322958,-0.0550226681,-0.1488328129,-0.2688644528,-0.0882172287,0.2847151458,0.1850690395,0.0958356038,0.3832220435,0.2154195458,0.111135751,0.0493784584,-0.1352442801,0.0353224054,0.0855564028,-0.2179709226,-0.1198172048,0.1821132153,0.3281980753,-0.2699041963,0.1215015873,-0.1425367743,-0.2085365504,0.001571518,-0.293589741,-0.0053074467,0.0134973433,-0.2574837208,0.3097872436,0.0935748518,-0.3035934269,0.0196346212,0.4677716196,-0.4638129771,-0.4341678917,-0.4358201325,0.0899604112,0.3708474934,-0.1129878536,-0.0923099965,0.059735287,0.4520861804,-0.1888256073,0.1029469818,-0.5797597766,0.3521030843,-0.0152924247,-0.0832994059,0.0905463472,-0.4035288692,-0.1816472858,0.1696708053,-0.1533035785,0.3111130297,0.4423996806,-0.3827602863,0.2947919369,0.0795271397,0.0166932028,0.8023683429,-0.3571899831,-0.2179583609,0.132254988,-0.1704488099,-0.3259194195,-0.0770724937,0.2736980319,0.0906305313,0.0717987269,-0.0094030369,0.3914262056,-0.0933154151,0.2531608641,-0.3342471123,-0.0749925673,0.3700163364,0.3320881724,0.2507386208,-0.0335372835,0.0665313303,0.3422166705,0.2596169412,-0.1454966366,0.2384007722,0.1618343294,0.0183554627,-0.1759249866,-0.1052252948,-0.1606465131,-0.3050886989,-0.2020210922,-0.3353252709,0.484374553,-0.2394975275,0.08094953,-0.1629112065,-0.0729379505,-0.1123220474,0.2390538901,0.1080784798,0.0814386383,-0.1436548978,-0.0689880103,-0.1000490189,-0.0618887804,-0.0523724556,0.1881001443,0.025137445,0.2732270956,-0.1691009253,-0.1986376792,-0.2601747513,0.0564328432,0.2558159232,0.0960626528,-0.2582112551,0.2561086416,-0.1447918415,-0.2581233084,-0.308811754,0.3103091717,0.155409649,-0.5250985622,0.0369421802,0.1614515334,0.1663212776,-0.0896117315,0.1419673562,0.1363632828,0.0686425567,0.1380071789,-0.0040352675,0.3693377972,0.2155335993,-0.0409564897,0.2696602643,0.0495535694,-0.1201176718,-0.1636262536,0.1193024367,0.0317241512,-0.2905445397,-0.0196277313,0.4356467724,-0.1525559574,0.241422236,0.4414304197,-0.1187741309,-0.0253542997,-0.1871619076,0.21855928,0.140799433,-0.0268120375,-0.0851378366,-0.0053834487,-0.253885746,-0.0652514324,0.3296323121,-0.1268236339,0.4315139949,0.1532039493,0.2892072797,-0.0184702352,-0.0514069125,-0.2919022739,0.2905476987,0.2543340623,-0.2341018468,0.4157405496,-0.1025249436,-0.2135272771,-0.6377745271,-0.3706792295,-0.0837398767,-0.1330630034,-0.0448598526,-0.0630305111,0.1643538177,0.0614361018,0.0771597028,-0.0667615905,0.2787643373,-0.1845620722,0.1062294468,-0.3318827152,-0.3990285695,-0.019308228,0.238119483,0.0182728078,-0.1148811653,0.0174304564,-0.0464675352,-0.0756031647,-0.1490138918,0.1128379405,-0.0141116986,-0.1526148766,0.1377332509,-0.0253324211,-0.2624071538,-0.227836132,0.3238838017,-0.3288731277,0.0045541548,0.4567353427,-0.1896839291,-0.0553343557,-0.0512933806,-0.3140721619,-0.5814594626,-0.0539097823,0.0172906443,0.0343012847,0.3296330273,0.3480425179,0.0307737906,0.1842597723,0.0340055116,-0.0597997718,0.0854178891,0.1192031726,0.3146250844,-0.0384516716,-0.288662076,-0.3471459448,-0.1107914522,0.0709793866,0.0762151107,-0.319314748,0.1262018532,-0.1260373443,0.0512030981,-0.2906250954,0.0220351126,0.4480994344,0.1102448702,-0.0087918788,0.2124655098,0.0139716482,0.3060884774,0.0416650251,0.2144845873,-0.034694355,0.1787259728,-0.04564083,0.6495577097,-0.1418004334,-0.3859343827,0.1767015159,-0.1262178123,-0.1793362498,-0.0094065778,-0.3002758324,-0.0031256527,-0.2587314844,-0.035383977,0.0021558774,-0.2854906023,-0.3254140913,-0.1811380535,0.1342670619,-0.1632559448,-0.3281493783,0.3672778308,-0.2507418394,0.2550690472,-0.1565136909,-0.1447717696,-0.1669237763,-0.4364982843,-0.030233575,0.1978235394,-0.299941659,0.0271620415,-0.5152040124,-0.2360682935,-0.1595627517,0.0845983028,0.3666800857,0.5396204591,0.0303010792,0.0349531136,0.0805857182,0.1010705158,0.4106414914,0.1839824766,-0.0750322565,0.1855862737,0.1655876786,-0.3579363525,-0.2835209072,-0.5400992036,-0.0822041705,0.2528803945,0.0967269838,-0.2896073461,-0.2426508218,0.2225344628,0.0414290503,-0.2890586257,-0.0435758941,0.0172090866,-0.2311727703,-0.2042599469,0.1062254459,0.2091031224,0.3887903094,-0.2306026816,-0.108919628,0.018064294,0.206056118,0.1758808345,-0.0173781198,0.1210725829,-0.1272633523,-0.2719097137,-0.2843087614,0.2212008238,0.1246747598,0.4509957135,0.1175308526,-0.3715092242,0.1879755855,-0.0572563186,0.3409126699,0.0074067381,-0.1104089469,0.3472835124,0.1715258509,0.3176763058,0.0993583575,-0.0289534982,0.2658733726,-0.077132225,-0.4949772954,-0.0418625735,0.2745336592,-0.073024258,-0.0733684525,-0.0241515674,-0.209478572,-0.2615410686,0.4185969532,0.3421198726,1.0206490755,-0.2603802681,0.3446718752,0.1092102751,-0.0606011935,0.4950573146,-0.1539972872,0.1934902817,-0.5524522066,0.0472402014,-0.1116915494,-0.1971725821,-0.1783649027,0.3437050283,-0.3830962181,0.1329992712,-0.0964831635,0.196059972,0.1422273517,0.3196293116,0.2093944252,-0.2090153843,-0.236824289,-0.0035606665,-0.4814991951,-0.0990892276,-0.0379736423,-0.2874800265,-0.0001370848,-0.3918527365,-0.2000364512,0.1904892027,-0.292277813,0.3384348452,0.5059814453,-0.1064435542,0.2228132337,0.1987271458,0.3068524003,-0.0456333794,0.061189007,-0.2703238726,-0.0626038909,0.1166037172,-0.0572850145,-0.0534290858,0.4100271761,-0.0476828143,-0.3138557673,0.3178364336,0.014437031,-0.1757712215,0.1572690457,-0.0672014877,0.0875045732,-0.4117056429,0.2714482546,-0.2188573778,-0.3185107708,-0.2686835229,0.00391638,0.0189208221,-0.1227038875,0.1043698266,0.3181701899,0.0148003632,0.1429253668,0.1666362286,0.2979536057,0.2228274792,0.5636560321,0.3159725964,0.011753371,-0.2083150148,0.1311721206,0.0069096126,-0.8838168979,0.1976176947,-0.0313806608,-0.2637524307,0.2123913914,0.3753276467,0.127044633,0.4987849891,-0.1740798354,-0.4247024059,-0.3973774314,0.0065994682,0.1742637008,0.0011970872,0.0514225103,0.5306829214,0.1982193142,0.154717207,-0.2049757689,0.1831800342,-0.3399907053,0.2554392815,-0.1108253449,-0.4195622504,-0.0381523035,0.1254320294,0.0438735113,0.001752787,0.0914920568,-0.054966893,-0.182664901,0.1793142408,0.2735623419,-0.2096082419,-0.2047402263,-0.0046507875,-0.0219300948,0.1029740423,-0.0729965493,0.007896659,-0.1585482657,0.5496968627,-0.0968374833,-0.0782949477,-0.3212496936,0.0320710763,-0.2329460979,-0.0552968457,0.1789356768,0.2310499996,0.1414344162,-0.1671888977,-0.7330248952,0.1469236463,0.2807041705,-0.0789484009,0.2824497819,-0.6361711621,-0.2515559494,-0.1240685359,0.0589370988,0.0178229157,-0.2906434834,-0.1190241799,0.1500851214,0.0792080164,-0.2228344381,-0.2199819088,-0.0071267015,-0.1568069011,-0.1895316839,0.4388341308,0.0504113249,-0.1781734973,-0.0951463059,0.0208672546,0.1502609104,-0.1083433032,0.0020464617,0.5980404615,-0.1331814378,-0.0291011278,0.4560702145,-0.0017088738,-0.0808275715,0.0111113079,-0.2403192073,0.1064087152,-0.1864718348,0.0284161344,0.2730591297,0.0053153406,-0.161443457,0.1116662025,0.0180342142,-0.1406107545,-0.2085789144,0.5285694599,-0.1294042617,-0.1111316606,-0.2098496258,0.2522187829,-0.2701119184,0.0049775327,-0.1391522586,-0.0752415732,-0.1743354499,-0.2161694765,-0.1959977895,0.1043635309,0.267588973,0.267826885,0.1731937677,-0.2923143804,0.0107252067,0.4745835066,0.0857413039,-0.2254720032,0.4008435607,0.3285622001,-0.0989184082,0.0243499521,0.3410256207,0.5116843581,0.5268767476,-0.2755404711,0.0296224151,0.0998523682,-0.1723219901,0.3009316325,0.1220268235,0.5902811885,0.1589998305,0.1322440356,-0.0237083342,-0.025082143,0.2330643088,-0.0365252122,0.1909649968,0.080381535,0.6249300241,0.2398195416,-0.3591584265,-0.1490707994,0.0843185484,-0.4341334403,0.0055399798,0.4604842663,-0.0493658446,0.2880000174,-0.0646258593,-0.0263440218,0.1581723094,0.6270230412,0.2333035618,0.2689677775,-0.3417343795,0.1156057864,-0.3256517053,0.3007265031,-0.001474376,-0.08987315,0.0791968703,0.0009102058,0.0593974143,0.1766050905,0.1285300255,0.0768041238,-0.1470921636,0.2389507443,-0.1691421568,0.0240432117,0.0808030888,0.227913022,-0.0594038852,-0.3299914598,0.2175324857,0.1071319431,-0.0194472913,-0.0732561797,0.0540491007,0.1230195463,0.1730821282,0.103905499,0.0345203616,0.5826874971,0.1241591573,-0.0638274997,0.3305566311,-0.1148174033,-0.1746624708,0.1645623893,0.2476192266,0.6241380572,-0.0343466848,0.2556886077,-0.4483470321,-0.2127869725,-0.0261634514,0.0727965012,-0.2580463588,0.2388151288,-0.2617353499,-0.0323142111,0.320294708,-0.0063781231,0.2982015908,0.0939638168,-0.4200736582,-0.4641219974,0.3670325279,-0.2881059051,-0.2306285352,-0.1451697499,0.112281464,0.0307800062,0.0831518546,-0.6610728502,-0.0098561831,0.3323136866,-0.0263023917,-0.221364066,-0.1205213144,0.0122635132,0.1851020753,-0.0976042598,0.2706556916,-0.0960784331,-0.0646676719,-0.454110235,-0.1759285182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/469","title":"invalid data type 'str' at _convert_outputs in arrow_dataset.py","comments":"Hello . Yes, I did set the output format as below for the two columns \r\n\r\n `train_dataset.set_format('torch',columns=['Text','Label'])`\r\n ","body":"I trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n","comment_length":16,"text":"invalid data type 'str' at _convert_outputs in arrow_dataset.py\nI trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n\nHello . Yes, I did set the output format as below for the two columns \r\n\r\n `train_dataset.set_format('torch',columns=['Text','Label'])`\r\n ","embeddings":[-0.2228239626,0.0242501777,0.1073713675,0.2303070426,0.5109525919,-0.1072595492,0.5323296189,0.0775902644,-0.3410620093,-0.1247765422,0.036657881,0.3864071369,-0.3696954846,0.0422176793,0.0443617031,-0.2024641186,-0.0880346373,0.0099904612,-0.0232526027,-0.0536389053,0.0482440926,0.1829803139,-0.1568261683,0.4144774675,-0.6451317668,-0.1835785806,0.1746639758,-0.0963928327,-0.0992212296,-0.4487785995,0.2202744484,-0.1020444036,0.4976367652,0.3472318351,-0.0001241712,0.1380359232,0.1092858762,0.1016949043,-0.2908720374,-0.3016375303,0.1341145188,-0.0593848564,0.4149942696,-0.2015384734,-0.0571896322,-0.2600911558,-0.0606259108,0.2617689073,0.3837757409,0.5016983747,0.1173220798,0.1685167998,0.2464857996,0.0738129616,0.2595641315,0.3155063391,0.0754768178,0.2071238756,0.0168337896,0.0266056098,-0.0856947526,0.0748076811,-0.0701284856,-0.1997739971,0.2166583389,0.227282986,0.17457892,-0.1476221532,-0.2287852317,0.1198840439,0.4332152605,-0.3214256167,-0.1559445858,0.0079212757,-0.0111622717,-0.2727991641,-0.0061669713,0.2483308762,-0.4277769327,0.1667014062,-0.1622606069,-0.0254668966,-0.3792498708,0.199556306,-0.0925177336,0.5032640696,0.1183697209,0.1568167061,0.1842403114,-0.0144359078,0.371710062,0.0876382813,0.1397552341,0.0621479377,-0.3116942644,-0.030463824,-0.4463334084,-0.648362577,-0.2685149014,-0.2125236392,0.1655772626,-0.0711523443,0.2047266066,0.202950567,0.233322382,0.1388006508,-0.2101232409,0.5648387074,0.1000927314,-0.1552505791,-0.1335455626,-0.0250890013,-0.3282387555,-0.0922897011,0.5121243596,0.1628403068,0.1257427186,-0.0050322958,-0.0550226681,-0.1488328129,-0.2688644528,-0.0882172287,0.2847151458,0.1850690395,0.0958356038,0.3832220435,0.2154195458,0.111135751,0.0493784584,-0.1352442801,0.0353224054,0.0855564028,-0.2179709226,-0.1198172048,0.1821132153,0.3281980753,-0.2699041963,0.1215015873,-0.1425367743,-0.2085365504,0.001571518,-0.293589741,-0.0053074467,0.0134973433,-0.2574837208,0.3097872436,0.0935748518,-0.3035934269,0.0196346212,0.4677716196,-0.4638129771,-0.4341678917,-0.4358201325,0.0899604112,0.3708474934,-0.1129878536,-0.0923099965,0.059735287,0.4520861804,-0.1888256073,0.1029469818,-0.5797597766,0.3521030843,-0.0152924247,-0.0832994059,0.0905463472,-0.4035288692,-0.1816472858,0.1696708053,-0.1533035785,0.3111130297,0.4423996806,-0.3827602863,0.2947919369,0.0795271397,0.0166932028,0.8023683429,-0.3571899831,-0.2179583609,0.132254988,-0.1704488099,-0.3259194195,-0.0770724937,0.2736980319,0.0906305313,0.0717987269,-0.0094030369,0.3914262056,-0.0933154151,0.2531608641,-0.3342471123,-0.0749925673,0.3700163364,0.3320881724,0.2507386208,-0.0335372835,0.0665313303,0.3422166705,0.2596169412,-0.1454966366,0.2384007722,0.1618343294,0.0183554627,-0.1759249866,-0.1052252948,-0.1606465131,-0.3050886989,-0.2020210922,-0.3353252709,0.484374553,-0.2394975275,0.08094953,-0.1629112065,-0.0729379505,-0.1123220474,0.2390538901,0.1080784798,0.0814386383,-0.1436548978,-0.0689880103,-0.1000490189,-0.0618887804,-0.0523724556,0.1881001443,0.025137445,0.2732270956,-0.1691009253,-0.1986376792,-0.2601747513,0.0564328432,0.2558159232,0.0960626528,-0.2582112551,0.2561086416,-0.1447918415,-0.2581233084,-0.308811754,0.3103091717,0.155409649,-0.5250985622,0.0369421802,0.1614515334,0.1663212776,-0.0896117315,0.1419673562,0.1363632828,0.0686425567,0.1380071789,-0.0040352675,0.3693377972,0.2155335993,-0.0409564897,0.2696602643,0.0495535694,-0.1201176718,-0.1636262536,0.1193024367,0.0317241512,-0.2905445397,-0.0196277313,0.4356467724,-0.1525559574,0.241422236,0.4414304197,-0.1187741309,-0.0253542997,-0.1871619076,0.21855928,0.140799433,-0.0268120375,-0.0851378366,-0.0053834487,-0.253885746,-0.0652514324,0.3296323121,-0.1268236339,0.4315139949,0.1532039493,0.2892072797,-0.0184702352,-0.0514069125,-0.2919022739,0.2905476987,0.2543340623,-0.2341018468,0.4157405496,-0.1025249436,-0.2135272771,-0.6377745271,-0.3706792295,-0.0837398767,-0.1330630034,-0.0448598526,-0.0630305111,0.1643538177,0.0614361018,0.0771597028,-0.0667615905,0.2787643373,-0.1845620722,0.1062294468,-0.3318827152,-0.3990285695,-0.019308228,0.238119483,0.0182728078,-0.1148811653,0.0174304564,-0.0464675352,-0.0756031647,-0.1490138918,0.1128379405,-0.0141116986,-0.1526148766,0.1377332509,-0.0253324211,-0.2624071538,-0.227836132,0.3238838017,-0.3288731277,0.0045541548,0.4567353427,-0.1896839291,-0.0553343557,-0.0512933806,-0.3140721619,-0.5814594626,-0.0539097823,0.0172906443,0.0343012847,0.3296330273,0.3480425179,0.0307737906,0.1842597723,0.0340055116,-0.0597997718,0.0854178891,0.1192031726,0.3146250844,-0.0384516716,-0.288662076,-0.3471459448,-0.1107914522,0.0709793866,0.0762151107,-0.319314748,0.1262018532,-0.1260373443,0.0512030981,-0.2906250954,0.0220351126,0.4480994344,0.1102448702,-0.0087918788,0.2124655098,0.0139716482,0.3060884774,0.0416650251,0.2144845873,-0.034694355,0.1787259728,-0.04564083,0.6495577097,-0.1418004334,-0.3859343827,0.1767015159,-0.1262178123,-0.1793362498,-0.0094065778,-0.3002758324,-0.0031256527,-0.2587314844,-0.035383977,0.0021558774,-0.2854906023,-0.3254140913,-0.1811380535,0.1342670619,-0.1632559448,-0.3281493783,0.3672778308,-0.2507418394,0.2550690472,-0.1565136909,-0.1447717696,-0.1669237763,-0.4364982843,-0.030233575,0.1978235394,-0.299941659,0.0271620415,-0.5152040124,-0.2360682935,-0.1595627517,0.0845983028,0.3666800857,0.5396204591,0.0303010792,0.0349531136,0.0805857182,0.1010705158,0.4106414914,0.1839824766,-0.0750322565,0.1855862737,0.1655876786,-0.3579363525,-0.2835209072,-0.5400992036,-0.0822041705,0.2528803945,0.0967269838,-0.2896073461,-0.2426508218,0.2225344628,0.0414290503,-0.2890586257,-0.0435758941,0.0172090866,-0.2311727703,-0.2042599469,0.1062254459,0.2091031224,0.3887903094,-0.2306026816,-0.108919628,0.018064294,0.206056118,0.1758808345,-0.0173781198,0.1210725829,-0.1272633523,-0.2719097137,-0.2843087614,0.2212008238,0.1246747598,0.4509957135,0.1175308526,-0.3715092242,0.1879755855,-0.0572563186,0.3409126699,0.0074067381,-0.1104089469,0.3472835124,0.1715258509,0.3176763058,0.0993583575,-0.0289534982,0.2658733726,-0.077132225,-0.4949772954,-0.0418625735,0.2745336592,-0.073024258,-0.0733684525,-0.0241515674,-0.209478572,-0.2615410686,0.4185969532,0.3421198726,1.0206490755,-0.2603802681,0.3446718752,0.1092102751,-0.0606011935,0.4950573146,-0.1539972872,0.1934902817,-0.5524522066,0.0472402014,-0.1116915494,-0.1971725821,-0.1783649027,0.3437050283,-0.3830962181,0.1329992712,-0.0964831635,0.196059972,0.1422273517,0.3196293116,0.2093944252,-0.2090153843,-0.236824289,-0.0035606665,-0.4814991951,-0.0990892276,-0.0379736423,-0.2874800265,-0.0001370848,-0.3918527365,-0.2000364512,0.1904892027,-0.292277813,0.3384348452,0.5059814453,-0.1064435542,0.2228132337,0.1987271458,0.3068524003,-0.0456333794,0.061189007,-0.2703238726,-0.0626038909,0.1166037172,-0.0572850145,-0.0534290858,0.4100271761,-0.0476828143,-0.3138557673,0.3178364336,0.014437031,-0.1757712215,0.1572690457,-0.0672014877,0.0875045732,-0.4117056429,0.2714482546,-0.2188573778,-0.3185107708,-0.2686835229,0.00391638,0.0189208221,-0.1227038875,0.1043698266,0.3181701899,0.0148003632,0.1429253668,0.1666362286,0.2979536057,0.2228274792,0.5636560321,0.3159725964,0.011753371,-0.2083150148,0.1311721206,0.0069096126,-0.8838168979,0.1976176947,-0.0313806608,-0.2637524307,0.2123913914,0.3753276467,0.127044633,0.4987849891,-0.1740798354,-0.4247024059,-0.3973774314,0.0065994682,0.1742637008,0.0011970872,0.0514225103,0.5306829214,0.1982193142,0.154717207,-0.2049757689,0.1831800342,-0.3399907053,0.2554392815,-0.1108253449,-0.4195622504,-0.0381523035,0.1254320294,0.0438735113,0.001752787,0.0914920568,-0.054966893,-0.182664901,0.1793142408,0.2735623419,-0.2096082419,-0.2047402263,-0.0046507875,-0.0219300948,0.1029740423,-0.0729965493,0.007896659,-0.1585482657,0.5496968627,-0.0968374833,-0.0782949477,-0.3212496936,0.0320710763,-0.2329460979,-0.0552968457,0.1789356768,0.2310499996,0.1414344162,-0.1671888977,-0.7330248952,0.1469236463,0.2807041705,-0.0789484009,0.2824497819,-0.6361711621,-0.2515559494,-0.1240685359,0.0589370988,0.0178229157,-0.2906434834,-0.1190241799,0.1500851214,0.0792080164,-0.2228344381,-0.2199819088,-0.0071267015,-0.1568069011,-0.1895316839,0.4388341308,0.0504113249,-0.1781734973,-0.0951463059,0.0208672546,0.1502609104,-0.1083433032,0.0020464617,0.5980404615,-0.1331814378,-0.0291011278,0.4560702145,-0.0017088738,-0.0808275715,0.0111113079,-0.2403192073,0.1064087152,-0.1864718348,0.0284161344,0.2730591297,0.0053153406,-0.161443457,0.1116662025,0.0180342142,-0.1406107545,-0.2085789144,0.5285694599,-0.1294042617,-0.1111316606,-0.2098496258,0.2522187829,-0.2701119184,0.0049775327,-0.1391522586,-0.0752415732,-0.1743354499,-0.2161694765,-0.1959977895,0.1043635309,0.267588973,0.267826885,0.1731937677,-0.2923143804,0.0107252067,0.4745835066,0.0857413039,-0.2254720032,0.4008435607,0.3285622001,-0.0989184082,0.0243499521,0.3410256207,0.5116843581,0.5268767476,-0.2755404711,0.0296224151,0.0998523682,-0.1723219901,0.3009316325,0.1220268235,0.5902811885,0.1589998305,0.1322440356,-0.0237083342,-0.025082143,0.2330643088,-0.0365252122,0.1909649968,0.080381535,0.6249300241,0.2398195416,-0.3591584265,-0.1490707994,0.0843185484,-0.4341334403,0.0055399798,0.4604842663,-0.0493658446,0.2880000174,-0.0646258593,-0.0263440218,0.1581723094,0.6270230412,0.2333035618,0.2689677775,-0.3417343795,0.1156057864,-0.3256517053,0.3007265031,-0.001474376,-0.08987315,0.0791968703,0.0009102058,0.0593974143,0.1766050905,0.1285300255,0.0768041238,-0.1470921636,0.2389507443,-0.1691421568,0.0240432117,0.0808030888,0.227913022,-0.0594038852,-0.3299914598,0.2175324857,0.1071319431,-0.0194472913,-0.0732561797,0.0540491007,0.1230195463,0.1730821282,0.103905499,0.0345203616,0.5826874971,0.1241591573,-0.0638274997,0.3305566311,-0.1148174033,-0.1746624708,0.1645623893,0.2476192266,0.6241380572,-0.0343466848,0.2556886077,-0.4483470321,-0.2127869725,-0.0261634514,0.0727965012,-0.2580463588,0.2388151288,-0.2617353499,-0.0323142111,0.320294708,-0.0063781231,0.2982015908,0.0939638168,-0.4200736582,-0.4641219974,0.3670325279,-0.2881059051,-0.2306285352,-0.1451697499,0.112281464,0.0307800062,0.0831518546,-0.6610728502,-0.0098561831,0.3323136866,-0.0263023917,-0.221364066,-0.1205213144,0.0122635132,0.1851020753,-0.0976042598,0.2706556916,-0.0960784331,-0.0646676719,-0.454110235,-0.1759285182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/469","title":"invalid data type 'str' at _convert_outputs in arrow_dataset.py","comments":"I think you're having this issue because you try to format strings as pytorch tensors, which is not possible.\r\nIndeed by having \"Text\" in `columns=['Text','Label']`, you try to convert the text values to pytorch tensors.\r\n\r\nInstead I recommend you to first tokenize your dataset using a tokenizer from transformers. For example\r\n\r\n```python\r\nfrom transformers import BertTokenizer\r\ntokenizer = BertTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\ntrain_dataset.map(lambda x: tokenizer(x[\"Text\"]), batched=True)\r\ntrain_dataset.set_format(\"torch\", column=[\"input_ids\"])\r\n```\r\n\r\nAnother way to fix your issue would be to not set the format to pytorch, and leave the dataset as it is by default. In that case, the strings are returned normally when you get examples from your dataloader. It means that you would have to tokenize the examples in the training loop (or using a data collator) though.\r\n\r\nLet me know if you have other questions","body":"I trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n","comment_length":133,"text":"invalid data type 'str' at _convert_outputs in arrow_dataset.py\nI trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n\nI think you're having this issue because you try to format strings as pytorch tensors, which is not possible.\r\nIndeed by having \"Text\" in `columns=['Text','Label']`, you try to convert the text values to pytorch tensors.\r\n\r\nInstead I recommend you to first tokenize your dataset using a tokenizer from transformers. For example\r\n\r\n```python\r\nfrom transformers import BertTokenizer\r\ntokenizer = BertTokenizer.from_pretrained(\"bert-base-uncased\")\r\n\r\ntrain_dataset.map(lambda x: tokenizer(x[\"Text\"]), batched=True)\r\ntrain_dataset.set_format(\"torch\", column=[\"input_ids\"])\r\n```\r\n\r\nAnother way to fix your issue would be to not set the format to pytorch, and leave the dataset as it is by default. In that case, the strings are returned normally when you get examples from your dataloader. It means that you would have to tokenize the examples in the training loop (or using a data collator) though.\r\n\r\nLet me know if you have other questions","embeddings":[-0.2228239626,0.0242501777,0.1073713675,0.2303070426,0.5109525919,-0.1072595492,0.5323296189,0.0775902644,-0.3410620093,-0.1247765422,0.036657881,0.3864071369,-0.3696954846,0.0422176793,0.0443617031,-0.2024641186,-0.0880346373,0.0099904612,-0.0232526027,-0.0536389053,0.0482440926,0.1829803139,-0.1568261683,0.4144774675,-0.6451317668,-0.1835785806,0.1746639758,-0.0963928327,-0.0992212296,-0.4487785995,0.2202744484,-0.1020444036,0.4976367652,0.3472318351,-0.0001241712,0.1380359232,0.1092858762,0.1016949043,-0.2908720374,-0.3016375303,0.1341145188,-0.0593848564,0.4149942696,-0.2015384734,-0.0571896322,-0.2600911558,-0.0606259108,0.2617689073,0.3837757409,0.5016983747,0.1173220798,0.1685167998,0.2464857996,0.0738129616,0.2595641315,0.3155063391,0.0754768178,0.2071238756,0.0168337896,0.0266056098,-0.0856947526,0.0748076811,-0.0701284856,-0.1997739971,0.2166583389,0.227282986,0.17457892,-0.1476221532,-0.2287852317,0.1198840439,0.4332152605,-0.3214256167,-0.1559445858,0.0079212757,-0.0111622717,-0.2727991641,-0.0061669713,0.2483308762,-0.4277769327,0.1667014062,-0.1622606069,-0.0254668966,-0.3792498708,0.199556306,-0.0925177336,0.5032640696,0.1183697209,0.1568167061,0.1842403114,-0.0144359078,0.371710062,0.0876382813,0.1397552341,0.0621479377,-0.3116942644,-0.030463824,-0.4463334084,-0.648362577,-0.2685149014,-0.2125236392,0.1655772626,-0.0711523443,0.2047266066,0.202950567,0.233322382,0.1388006508,-0.2101232409,0.5648387074,0.1000927314,-0.1552505791,-0.1335455626,-0.0250890013,-0.3282387555,-0.0922897011,0.5121243596,0.1628403068,0.1257427186,-0.0050322958,-0.0550226681,-0.1488328129,-0.2688644528,-0.0882172287,0.2847151458,0.1850690395,0.0958356038,0.3832220435,0.2154195458,0.111135751,0.0493784584,-0.1352442801,0.0353224054,0.0855564028,-0.2179709226,-0.1198172048,0.1821132153,0.3281980753,-0.2699041963,0.1215015873,-0.1425367743,-0.2085365504,0.001571518,-0.293589741,-0.0053074467,0.0134973433,-0.2574837208,0.3097872436,0.0935748518,-0.3035934269,0.0196346212,0.4677716196,-0.4638129771,-0.4341678917,-0.4358201325,0.0899604112,0.3708474934,-0.1129878536,-0.0923099965,0.059735287,0.4520861804,-0.1888256073,0.1029469818,-0.5797597766,0.3521030843,-0.0152924247,-0.0832994059,0.0905463472,-0.4035288692,-0.1816472858,0.1696708053,-0.1533035785,0.3111130297,0.4423996806,-0.3827602863,0.2947919369,0.0795271397,0.0166932028,0.8023683429,-0.3571899831,-0.2179583609,0.132254988,-0.1704488099,-0.3259194195,-0.0770724937,0.2736980319,0.0906305313,0.0717987269,-0.0094030369,0.3914262056,-0.0933154151,0.2531608641,-0.3342471123,-0.0749925673,0.3700163364,0.3320881724,0.2507386208,-0.0335372835,0.0665313303,0.3422166705,0.2596169412,-0.1454966366,0.2384007722,0.1618343294,0.0183554627,-0.1759249866,-0.1052252948,-0.1606465131,-0.3050886989,-0.2020210922,-0.3353252709,0.484374553,-0.2394975275,0.08094953,-0.1629112065,-0.0729379505,-0.1123220474,0.2390538901,0.1080784798,0.0814386383,-0.1436548978,-0.0689880103,-0.1000490189,-0.0618887804,-0.0523724556,0.1881001443,0.025137445,0.2732270956,-0.1691009253,-0.1986376792,-0.2601747513,0.0564328432,0.2558159232,0.0960626528,-0.2582112551,0.2561086416,-0.1447918415,-0.2581233084,-0.308811754,0.3103091717,0.155409649,-0.5250985622,0.0369421802,0.1614515334,0.1663212776,-0.0896117315,0.1419673562,0.1363632828,0.0686425567,0.1380071789,-0.0040352675,0.3693377972,0.2155335993,-0.0409564897,0.2696602643,0.0495535694,-0.1201176718,-0.1636262536,0.1193024367,0.0317241512,-0.2905445397,-0.0196277313,0.4356467724,-0.1525559574,0.241422236,0.4414304197,-0.1187741309,-0.0253542997,-0.1871619076,0.21855928,0.140799433,-0.0268120375,-0.0851378366,-0.0053834487,-0.253885746,-0.0652514324,0.3296323121,-0.1268236339,0.4315139949,0.1532039493,0.2892072797,-0.0184702352,-0.0514069125,-0.2919022739,0.2905476987,0.2543340623,-0.2341018468,0.4157405496,-0.1025249436,-0.2135272771,-0.6377745271,-0.3706792295,-0.0837398767,-0.1330630034,-0.0448598526,-0.0630305111,0.1643538177,0.0614361018,0.0771597028,-0.0667615905,0.2787643373,-0.1845620722,0.1062294468,-0.3318827152,-0.3990285695,-0.019308228,0.238119483,0.0182728078,-0.1148811653,0.0174304564,-0.0464675352,-0.0756031647,-0.1490138918,0.1128379405,-0.0141116986,-0.1526148766,0.1377332509,-0.0253324211,-0.2624071538,-0.227836132,0.3238838017,-0.3288731277,0.0045541548,0.4567353427,-0.1896839291,-0.0553343557,-0.0512933806,-0.3140721619,-0.5814594626,-0.0539097823,0.0172906443,0.0343012847,0.3296330273,0.3480425179,0.0307737906,0.1842597723,0.0340055116,-0.0597997718,0.0854178891,0.1192031726,0.3146250844,-0.0384516716,-0.288662076,-0.3471459448,-0.1107914522,0.0709793866,0.0762151107,-0.319314748,0.1262018532,-0.1260373443,0.0512030981,-0.2906250954,0.0220351126,0.4480994344,0.1102448702,-0.0087918788,0.2124655098,0.0139716482,0.3060884774,0.0416650251,0.2144845873,-0.034694355,0.1787259728,-0.04564083,0.6495577097,-0.1418004334,-0.3859343827,0.1767015159,-0.1262178123,-0.1793362498,-0.0094065778,-0.3002758324,-0.0031256527,-0.2587314844,-0.035383977,0.0021558774,-0.2854906023,-0.3254140913,-0.1811380535,0.1342670619,-0.1632559448,-0.3281493783,0.3672778308,-0.2507418394,0.2550690472,-0.1565136909,-0.1447717696,-0.1669237763,-0.4364982843,-0.030233575,0.1978235394,-0.299941659,0.0271620415,-0.5152040124,-0.2360682935,-0.1595627517,0.0845983028,0.3666800857,0.5396204591,0.0303010792,0.0349531136,0.0805857182,0.1010705158,0.4106414914,0.1839824766,-0.0750322565,0.1855862737,0.1655876786,-0.3579363525,-0.2835209072,-0.5400992036,-0.0822041705,0.2528803945,0.0967269838,-0.2896073461,-0.2426508218,0.2225344628,0.0414290503,-0.2890586257,-0.0435758941,0.0172090866,-0.2311727703,-0.2042599469,0.1062254459,0.2091031224,0.3887903094,-0.2306026816,-0.108919628,0.018064294,0.206056118,0.1758808345,-0.0173781198,0.1210725829,-0.1272633523,-0.2719097137,-0.2843087614,0.2212008238,0.1246747598,0.4509957135,0.1175308526,-0.3715092242,0.1879755855,-0.0572563186,0.3409126699,0.0074067381,-0.1104089469,0.3472835124,0.1715258509,0.3176763058,0.0993583575,-0.0289534982,0.2658733726,-0.077132225,-0.4949772954,-0.0418625735,0.2745336592,-0.073024258,-0.0733684525,-0.0241515674,-0.209478572,-0.2615410686,0.4185969532,0.3421198726,1.0206490755,-0.2603802681,0.3446718752,0.1092102751,-0.0606011935,0.4950573146,-0.1539972872,0.1934902817,-0.5524522066,0.0472402014,-0.1116915494,-0.1971725821,-0.1783649027,0.3437050283,-0.3830962181,0.1329992712,-0.0964831635,0.196059972,0.1422273517,0.3196293116,0.2093944252,-0.2090153843,-0.236824289,-0.0035606665,-0.4814991951,-0.0990892276,-0.0379736423,-0.2874800265,-0.0001370848,-0.3918527365,-0.2000364512,0.1904892027,-0.292277813,0.3384348452,0.5059814453,-0.1064435542,0.2228132337,0.1987271458,0.3068524003,-0.0456333794,0.061189007,-0.2703238726,-0.0626038909,0.1166037172,-0.0572850145,-0.0534290858,0.4100271761,-0.0476828143,-0.3138557673,0.3178364336,0.014437031,-0.1757712215,0.1572690457,-0.0672014877,0.0875045732,-0.4117056429,0.2714482546,-0.2188573778,-0.3185107708,-0.2686835229,0.00391638,0.0189208221,-0.1227038875,0.1043698266,0.3181701899,0.0148003632,0.1429253668,0.1666362286,0.2979536057,0.2228274792,0.5636560321,0.3159725964,0.011753371,-0.2083150148,0.1311721206,0.0069096126,-0.8838168979,0.1976176947,-0.0313806608,-0.2637524307,0.2123913914,0.3753276467,0.127044633,0.4987849891,-0.1740798354,-0.4247024059,-0.3973774314,0.0065994682,0.1742637008,0.0011970872,0.0514225103,0.5306829214,0.1982193142,0.154717207,-0.2049757689,0.1831800342,-0.3399907053,0.2554392815,-0.1108253449,-0.4195622504,-0.0381523035,0.1254320294,0.0438735113,0.001752787,0.0914920568,-0.054966893,-0.182664901,0.1793142408,0.2735623419,-0.2096082419,-0.2047402263,-0.0046507875,-0.0219300948,0.1029740423,-0.0729965493,0.007896659,-0.1585482657,0.5496968627,-0.0968374833,-0.0782949477,-0.3212496936,0.0320710763,-0.2329460979,-0.0552968457,0.1789356768,0.2310499996,0.1414344162,-0.1671888977,-0.7330248952,0.1469236463,0.2807041705,-0.0789484009,0.2824497819,-0.6361711621,-0.2515559494,-0.1240685359,0.0589370988,0.0178229157,-0.2906434834,-0.1190241799,0.1500851214,0.0792080164,-0.2228344381,-0.2199819088,-0.0071267015,-0.1568069011,-0.1895316839,0.4388341308,0.0504113249,-0.1781734973,-0.0951463059,0.0208672546,0.1502609104,-0.1083433032,0.0020464617,0.5980404615,-0.1331814378,-0.0291011278,0.4560702145,-0.0017088738,-0.0808275715,0.0111113079,-0.2403192073,0.1064087152,-0.1864718348,0.0284161344,0.2730591297,0.0053153406,-0.161443457,0.1116662025,0.0180342142,-0.1406107545,-0.2085789144,0.5285694599,-0.1294042617,-0.1111316606,-0.2098496258,0.2522187829,-0.2701119184,0.0049775327,-0.1391522586,-0.0752415732,-0.1743354499,-0.2161694765,-0.1959977895,0.1043635309,0.267588973,0.267826885,0.1731937677,-0.2923143804,0.0107252067,0.4745835066,0.0857413039,-0.2254720032,0.4008435607,0.3285622001,-0.0989184082,0.0243499521,0.3410256207,0.5116843581,0.5268767476,-0.2755404711,0.0296224151,0.0998523682,-0.1723219901,0.3009316325,0.1220268235,0.5902811885,0.1589998305,0.1322440356,-0.0237083342,-0.025082143,0.2330643088,-0.0365252122,0.1909649968,0.080381535,0.6249300241,0.2398195416,-0.3591584265,-0.1490707994,0.0843185484,-0.4341334403,0.0055399798,0.4604842663,-0.0493658446,0.2880000174,-0.0646258593,-0.0263440218,0.1581723094,0.6270230412,0.2333035618,0.2689677775,-0.3417343795,0.1156057864,-0.3256517053,0.3007265031,-0.001474376,-0.08987315,0.0791968703,0.0009102058,0.0593974143,0.1766050905,0.1285300255,0.0768041238,-0.1470921636,0.2389507443,-0.1691421568,0.0240432117,0.0808030888,0.227913022,-0.0594038852,-0.3299914598,0.2175324857,0.1071319431,-0.0194472913,-0.0732561797,0.0540491007,0.1230195463,0.1730821282,0.103905499,0.0345203616,0.5826874971,0.1241591573,-0.0638274997,0.3305566311,-0.1148174033,-0.1746624708,0.1645623893,0.2476192266,0.6241380572,-0.0343466848,0.2556886077,-0.4483470321,-0.2127869725,-0.0261634514,0.0727965012,-0.2580463588,0.2388151288,-0.2617353499,-0.0323142111,0.320294708,-0.0063781231,0.2982015908,0.0939638168,-0.4200736582,-0.4641219974,0.3670325279,-0.2881059051,-0.2306285352,-0.1451697499,0.112281464,0.0307800062,0.0831518546,-0.6610728502,-0.0098561831,0.3323136866,-0.0263023917,-0.221364066,-0.1205213144,0.0122635132,0.1851020753,-0.0976042598,0.2706556916,-0.0960784331,-0.0646676719,-0.454110235,-0.1759285182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/469","title":"invalid data type 'str' at _convert_outputs in arrow_dataset.py","comments":"Hi, actually the thing is I am getting the same error and even after tokenizing them I am passing them through batch_encode_plus.\r\nI dont know what seems to be the problem is. I even converted it into 'pt' while passing them through batch_encode_plus but when I am evaluating my model , i am getting this error\r\n\r\n\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-145-ca218223c9fc> in <module>()\r\n----> 1 val_loss, predictions, true_val = evaluate(dataloader_validation)\r\n 2 val_f1 = f1_score_func(predictions, true_val)\r\n 3 tqdm.write(f'Validation loss: {val_loss}')\r\n 4 tqdm.write(f'F1 Score (Weighted): {val_f1}')\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataset.py in <genexpr>(.0)\r\n 160 \r\n 161 def __getitem__(self, index):\r\n--> 162 return tuple(tensor[index] for tensor in self.tensors)\r\n 163 \r\n 164 def __len__(self):\r\n\r\nTypeError: new(): invalid data type 'str' ","body":"I trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n","comment_length":115,"text":"invalid data type 'str' at _convert_outputs in arrow_dataset.py\nI trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n\nHi, actually the thing is I am getting the same error and even after tokenizing them I am passing them through batch_encode_plus.\r\nI dont know what seems to be the problem is. I even converted it into 'pt' while passing them through batch_encode_plus but when I am evaluating my model , i am getting this error\r\n\r\n\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-145-ca218223c9fc> in <module>()\r\n----> 1 val_loss, predictions, true_val = evaluate(dataloader_validation)\r\n 2 val_f1 = f1_score_func(predictions, true_val)\r\n 3 tqdm.write(f'Validation loss: {val_loss}')\r\n 4 tqdm.write(f'F1 Score (Weighted): {val_f1}')\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataset.py in <genexpr>(.0)\r\n 160 \r\n 161 def __getitem__(self, index):\r\n--> 162 return tuple(tensor[index] for tensor in self.tensors)\r\n 163 \r\n 164 def __len__(self):\r\n\r\nTypeError: new(): invalid data type 'str' ","embeddings":[-0.2228239626,0.0242501777,0.1073713675,0.2303070426,0.5109525919,-0.1072595492,0.5323296189,0.0775902644,-0.3410620093,-0.1247765422,0.036657881,0.3864071369,-0.3696954846,0.0422176793,0.0443617031,-0.2024641186,-0.0880346373,0.0099904612,-0.0232526027,-0.0536389053,0.0482440926,0.1829803139,-0.1568261683,0.4144774675,-0.6451317668,-0.1835785806,0.1746639758,-0.0963928327,-0.0992212296,-0.4487785995,0.2202744484,-0.1020444036,0.4976367652,0.3472318351,-0.0001241712,0.1380359232,0.1092858762,0.1016949043,-0.2908720374,-0.3016375303,0.1341145188,-0.0593848564,0.4149942696,-0.2015384734,-0.0571896322,-0.2600911558,-0.0606259108,0.2617689073,0.3837757409,0.5016983747,0.1173220798,0.1685167998,0.2464857996,0.0738129616,0.2595641315,0.3155063391,0.0754768178,0.2071238756,0.0168337896,0.0266056098,-0.0856947526,0.0748076811,-0.0701284856,-0.1997739971,0.2166583389,0.227282986,0.17457892,-0.1476221532,-0.2287852317,0.1198840439,0.4332152605,-0.3214256167,-0.1559445858,0.0079212757,-0.0111622717,-0.2727991641,-0.0061669713,0.2483308762,-0.4277769327,0.1667014062,-0.1622606069,-0.0254668966,-0.3792498708,0.199556306,-0.0925177336,0.5032640696,0.1183697209,0.1568167061,0.1842403114,-0.0144359078,0.371710062,0.0876382813,0.1397552341,0.0621479377,-0.3116942644,-0.030463824,-0.4463334084,-0.648362577,-0.2685149014,-0.2125236392,0.1655772626,-0.0711523443,0.2047266066,0.202950567,0.233322382,0.1388006508,-0.2101232409,0.5648387074,0.1000927314,-0.1552505791,-0.1335455626,-0.0250890013,-0.3282387555,-0.0922897011,0.5121243596,0.1628403068,0.1257427186,-0.0050322958,-0.0550226681,-0.1488328129,-0.2688644528,-0.0882172287,0.2847151458,0.1850690395,0.0958356038,0.3832220435,0.2154195458,0.111135751,0.0493784584,-0.1352442801,0.0353224054,0.0855564028,-0.2179709226,-0.1198172048,0.1821132153,0.3281980753,-0.2699041963,0.1215015873,-0.1425367743,-0.2085365504,0.001571518,-0.293589741,-0.0053074467,0.0134973433,-0.2574837208,0.3097872436,0.0935748518,-0.3035934269,0.0196346212,0.4677716196,-0.4638129771,-0.4341678917,-0.4358201325,0.0899604112,0.3708474934,-0.1129878536,-0.0923099965,0.059735287,0.4520861804,-0.1888256073,0.1029469818,-0.5797597766,0.3521030843,-0.0152924247,-0.0832994059,0.0905463472,-0.4035288692,-0.1816472858,0.1696708053,-0.1533035785,0.3111130297,0.4423996806,-0.3827602863,0.2947919369,0.0795271397,0.0166932028,0.8023683429,-0.3571899831,-0.2179583609,0.132254988,-0.1704488099,-0.3259194195,-0.0770724937,0.2736980319,0.0906305313,0.0717987269,-0.0094030369,0.3914262056,-0.0933154151,0.2531608641,-0.3342471123,-0.0749925673,0.3700163364,0.3320881724,0.2507386208,-0.0335372835,0.0665313303,0.3422166705,0.2596169412,-0.1454966366,0.2384007722,0.1618343294,0.0183554627,-0.1759249866,-0.1052252948,-0.1606465131,-0.3050886989,-0.2020210922,-0.3353252709,0.484374553,-0.2394975275,0.08094953,-0.1629112065,-0.0729379505,-0.1123220474,0.2390538901,0.1080784798,0.0814386383,-0.1436548978,-0.0689880103,-0.1000490189,-0.0618887804,-0.0523724556,0.1881001443,0.025137445,0.2732270956,-0.1691009253,-0.1986376792,-0.2601747513,0.0564328432,0.2558159232,0.0960626528,-0.2582112551,0.2561086416,-0.1447918415,-0.2581233084,-0.308811754,0.3103091717,0.155409649,-0.5250985622,0.0369421802,0.1614515334,0.1663212776,-0.0896117315,0.1419673562,0.1363632828,0.0686425567,0.1380071789,-0.0040352675,0.3693377972,0.2155335993,-0.0409564897,0.2696602643,0.0495535694,-0.1201176718,-0.1636262536,0.1193024367,0.0317241512,-0.2905445397,-0.0196277313,0.4356467724,-0.1525559574,0.241422236,0.4414304197,-0.1187741309,-0.0253542997,-0.1871619076,0.21855928,0.140799433,-0.0268120375,-0.0851378366,-0.0053834487,-0.253885746,-0.0652514324,0.3296323121,-0.1268236339,0.4315139949,0.1532039493,0.2892072797,-0.0184702352,-0.0514069125,-0.2919022739,0.2905476987,0.2543340623,-0.2341018468,0.4157405496,-0.1025249436,-0.2135272771,-0.6377745271,-0.3706792295,-0.0837398767,-0.1330630034,-0.0448598526,-0.0630305111,0.1643538177,0.0614361018,0.0771597028,-0.0667615905,0.2787643373,-0.1845620722,0.1062294468,-0.3318827152,-0.3990285695,-0.019308228,0.238119483,0.0182728078,-0.1148811653,0.0174304564,-0.0464675352,-0.0756031647,-0.1490138918,0.1128379405,-0.0141116986,-0.1526148766,0.1377332509,-0.0253324211,-0.2624071538,-0.227836132,0.3238838017,-0.3288731277,0.0045541548,0.4567353427,-0.1896839291,-0.0553343557,-0.0512933806,-0.3140721619,-0.5814594626,-0.0539097823,0.0172906443,0.0343012847,0.3296330273,0.3480425179,0.0307737906,0.1842597723,0.0340055116,-0.0597997718,0.0854178891,0.1192031726,0.3146250844,-0.0384516716,-0.288662076,-0.3471459448,-0.1107914522,0.0709793866,0.0762151107,-0.319314748,0.1262018532,-0.1260373443,0.0512030981,-0.2906250954,0.0220351126,0.4480994344,0.1102448702,-0.0087918788,0.2124655098,0.0139716482,0.3060884774,0.0416650251,0.2144845873,-0.034694355,0.1787259728,-0.04564083,0.6495577097,-0.1418004334,-0.3859343827,0.1767015159,-0.1262178123,-0.1793362498,-0.0094065778,-0.3002758324,-0.0031256527,-0.2587314844,-0.035383977,0.0021558774,-0.2854906023,-0.3254140913,-0.1811380535,0.1342670619,-0.1632559448,-0.3281493783,0.3672778308,-0.2507418394,0.2550690472,-0.1565136909,-0.1447717696,-0.1669237763,-0.4364982843,-0.030233575,0.1978235394,-0.299941659,0.0271620415,-0.5152040124,-0.2360682935,-0.1595627517,0.0845983028,0.3666800857,0.5396204591,0.0303010792,0.0349531136,0.0805857182,0.1010705158,0.4106414914,0.1839824766,-0.0750322565,0.1855862737,0.1655876786,-0.3579363525,-0.2835209072,-0.5400992036,-0.0822041705,0.2528803945,0.0967269838,-0.2896073461,-0.2426508218,0.2225344628,0.0414290503,-0.2890586257,-0.0435758941,0.0172090866,-0.2311727703,-0.2042599469,0.1062254459,0.2091031224,0.3887903094,-0.2306026816,-0.108919628,0.018064294,0.206056118,0.1758808345,-0.0173781198,0.1210725829,-0.1272633523,-0.2719097137,-0.2843087614,0.2212008238,0.1246747598,0.4509957135,0.1175308526,-0.3715092242,0.1879755855,-0.0572563186,0.3409126699,0.0074067381,-0.1104089469,0.3472835124,0.1715258509,0.3176763058,0.0993583575,-0.0289534982,0.2658733726,-0.077132225,-0.4949772954,-0.0418625735,0.2745336592,-0.073024258,-0.0733684525,-0.0241515674,-0.209478572,-0.2615410686,0.4185969532,0.3421198726,1.0206490755,-0.2603802681,0.3446718752,0.1092102751,-0.0606011935,0.4950573146,-0.1539972872,0.1934902817,-0.5524522066,0.0472402014,-0.1116915494,-0.1971725821,-0.1783649027,0.3437050283,-0.3830962181,0.1329992712,-0.0964831635,0.196059972,0.1422273517,0.3196293116,0.2093944252,-0.2090153843,-0.236824289,-0.0035606665,-0.4814991951,-0.0990892276,-0.0379736423,-0.2874800265,-0.0001370848,-0.3918527365,-0.2000364512,0.1904892027,-0.292277813,0.3384348452,0.5059814453,-0.1064435542,0.2228132337,0.1987271458,0.3068524003,-0.0456333794,0.061189007,-0.2703238726,-0.0626038909,0.1166037172,-0.0572850145,-0.0534290858,0.4100271761,-0.0476828143,-0.3138557673,0.3178364336,0.014437031,-0.1757712215,0.1572690457,-0.0672014877,0.0875045732,-0.4117056429,0.2714482546,-0.2188573778,-0.3185107708,-0.2686835229,0.00391638,0.0189208221,-0.1227038875,0.1043698266,0.3181701899,0.0148003632,0.1429253668,0.1666362286,0.2979536057,0.2228274792,0.5636560321,0.3159725964,0.011753371,-0.2083150148,0.1311721206,0.0069096126,-0.8838168979,0.1976176947,-0.0313806608,-0.2637524307,0.2123913914,0.3753276467,0.127044633,0.4987849891,-0.1740798354,-0.4247024059,-0.3973774314,0.0065994682,0.1742637008,0.0011970872,0.0514225103,0.5306829214,0.1982193142,0.154717207,-0.2049757689,0.1831800342,-0.3399907053,0.2554392815,-0.1108253449,-0.4195622504,-0.0381523035,0.1254320294,0.0438735113,0.001752787,0.0914920568,-0.054966893,-0.182664901,0.1793142408,0.2735623419,-0.2096082419,-0.2047402263,-0.0046507875,-0.0219300948,0.1029740423,-0.0729965493,0.007896659,-0.1585482657,0.5496968627,-0.0968374833,-0.0782949477,-0.3212496936,0.0320710763,-0.2329460979,-0.0552968457,0.1789356768,0.2310499996,0.1414344162,-0.1671888977,-0.7330248952,0.1469236463,0.2807041705,-0.0789484009,0.2824497819,-0.6361711621,-0.2515559494,-0.1240685359,0.0589370988,0.0178229157,-0.2906434834,-0.1190241799,0.1500851214,0.0792080164,-0.2228344381,-0.2199819088,-0.0071267015,-0.1568069011,-0.1895316839,0.4388341308,0.0504113249,-0.1781734973,-0.0951463059,0.0208672546,0.1502609104,-0.1083433032,0.0020464617,0.5980404615,-0.1331814378,-0.0291011278,0.4560702145,-0.0017088738,-0.0808275715,0.0111113079,-0.2403192073,0.1064087152,-0.1864718348,0.0284161344,0.2730591297,0.0053153406,-0.161443457,0.1116662025,0.0180342142,-0.1406107545,-0.2085789144,0.5285694599,-0.1294042617,-0.1111316606,-0.2098496258,0.2522187829,-0.2701119184,0.0049775327,-0.1391522586,-0.0752415732,-0.1743354499,-0.2161694765,-0.1959977895,0.1043635309,0.267588973,0.267826885,0.1731937677,-0.2923143804,0.0107252067,0.4745835066,0.0857413039,-0.2254720032,0.4008435607,0.3285622001,-0.0989184082,0.0243499521,0.3410256207,0.5116843581,0.5268767476,-0.2755404711,0.0296224151,0.0998523682,-0.1723219901,0.3009316325,0.1220268235,0.5902811885,0.1589998305,0.1322440356,-0.0237083342,-0.025082143,0.2330643088,-0.0365252122,0.1909649968,0.080381535,0.6249300241,0.2398195416,-0.3591584265,-0.1490707994,0.0843185484,-0.4341334403,0.0055399798,0.4604842663,-0.0493658446,0.2880000174,-0.0646258593,-0.0263440218,0.1581723094,0.6270230412,0.2333035618,0.2689677775,-0.3417343795,0.1156057864,-0.3256517053,0.3007265031,-0.001474376,-0.08987315,0.0791968703,0.0009102058,0.0593974143,0.1766050905,0.1285300255,0.0768041238,-0.1470921636,0.2389507443,-0.1691421568,0.0240432117,0.0808030888,0.227913022,-0.0594038852,-0.3299914598,0.2175324857,0.1071319431,-0.0194472913,-0.0732561797,0.0540491007,0.1230195463,0.1730821282,0.103905499,0.0345203616,0.5826874971,0.1241591573,-0.0638274997,0.3305566311,-0.1148174033,-0.1746624708,0.1645623893,0.2476192266,0.6241380572,-0.0343466848,0.2556886077,-0.4483470321,-0.2127869725,-0.0261634514,0.0727965012,-0.2580463588,0.2388151288,-0.2617353499,-0.0323142111,0.320294708,-0.0063781231,0.2982015908,0.0939638168,-0.4200736582,-0.4641219974,0.3670325279,-0.2881059051,-0.2306285352,-0.1451697499,0.112281464,0.0307800062,0.0831518546,-0.6610728502,-0.0098561831,0.3323136866,-0.0263023917,-0.221364066,-0.1205213144,0.0122635132,0.1851020753,-0.0976042598,0.2706556916,-0.0960784331,-0.0646676719,-0.454110235,-0.1759285182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/469","title":"invalid data type 'str' at _convert_outputs in arrow_dataset.py","comments":"> Hi, actually the thing is I am getting the same error and even after tokenizing them I am passing them through batch_encode_plus.\r\n> I dont know what seems to be the problem is. I even converted it into 'pt' while passing them through batch_encode_plus but when I am evaluating my model , i am getting this error\r\n> \r\n> TypeError Traceback (most recent call last)\r\n> in ()\r\n> ----> 1 val_loss, predictions, true_val = evaluate(dataloader_validation)\r\n> 2 val_f1 = f1_score_func(predictions, true_val)\r\n> 3 tqdm.write(f'Validation loss: {val_loss}')\r\n> 4 tqdm.write(f'F1 Score (Weighted): {val_f1}')\r\n> \r\n> 6 frames\r\n> \/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataset.py in (.0)\r\n> 160\r\n> 161 def **getitem**(self, index):\r\n> --> 162 return tuple(tensor[index] for tensor in self.tensors)\r\n> 163\r\n> 164 def **len**(self):\r\n> \r\n> TypeError: new(): invalid data type 'str'\r\n\r\nI got the same error and fix it .\r\nyou can check your input where there may be string contained.\r\nsuch as\r\n```\r\na = [1,2,3,4,'<unk>']\r\ntorch.tensor(a)\r\n```","body":"I trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n","comment_length":160,"text":"invalid data type 'str' at _convert_outputs in arrow_dataset.py\nI trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n\n> Hi, actually the thing is I am getting the same error and even after tokenizing them I am passing them through batch_encode_plus.\r\n> I dont know what seems to be the problem is. I even converted it into 'pt' while passing them through batch_encode_plus but when I am evaluating my model , i am getting this error\r\n> \r\n> TypeError Traceback (most recent call last)\r\n> in ()\r\n> ----> 1 val_loss, predictions, true_val = evaluate(dataloader_validation)\r\n> 2 val_f1 = f1_score_func(predictions, true_val)\r\n> 3 tqdm.write(f'Validation loss: {val_loss}')\r\n> 4 tqdm.write(f'F1 Score (Weighted): {val_f1}')\r\n> \r\n> 6 frames\r\n> \/usr\/local\/lib\/python3.6\/dist-packages\/torch\/utils\/data\/dataset.py in (.0)\r\n> 160\r\n> 161 def **getitem**(self, index):\r\n> --> 162 return tuple(tensor[index] for tensor in self.tensors)\r\n> 163\r\n> 164 def **len**(self):\r\n> \r\n> TypeError: new(): invalid data type 'str'\r\n\r\nI got the same error and fix it .\r\nyou can check your input where there may be string contained.\r\nsuch as\r\n```\r\na = [1,2,3,4,'<unk>']\r\ntorch.tensor(a)\r\n```","embeddings":[-0.2228239626,0.0242501777,0.1073713675,0.2303070426,0.5109525919,-0.1072595492,0.5323296189,0.0775902644,-0.3410620093,-0.1247765422,0.036657881,0.3864071369,-0.3696954846,0.0422176793,0.0443617031,-0.2024641186,-0.0880346373,0.0099904612,-0.0232526027,-0.0536389053,0.0482440926,0.1829803139,-0.1568261683,0.4144774675,-0.6451317668,-0.1835785806,0.1746639758,-0.0963928327,-0.0992212296,-0.4487785995,0.2202744484,-0.1020444036,0.4976367652,0.3472318351,-0.0001241712,0.1380359232,0.1092858762,0.1016949043,-0.2908720374,-0.3016375303,0.1341145188,-0.0593848564,0.4149942696,-0.2015384734,-0.0571896322,-0.2600911558,-0.0606259108,0.2617689073,0.3837757409,0.5016983747,0.1173220798,0.1685167998,0.2464857996,0.0738129616,0.2595641315,0.3155063391,0.0754768178,0.2071238756,0.0168337896,0.0266056098,-0.0856947526,0.0748076811,-0.0701284856,-0.1997739971,0.2166583389,0.227282986,0.17457892,-0.1476221532,-0.2287852317,0.1198840439,0.4332152605,-0.3214256167,-0.1559445858,0.0079212757,-0.0111622717,-0.2727991641,-0.0061669713,0.2483308762,-0.4277769327,0.1667014062,-0.1622606069,-0.0254668966,-0.3792498708,0.199556306,-0.0925177336,0.5032640696,0.1183697209,0.1568167061,0.1842403114,-0.0144359078,0.371710062,0.0876382813,0.1397552341,0.0621479377,-0.3116942644,-0.030463824,-0.4463334084,-0.648362577,-0.2685149014,-0.2125236392,0.1655772626,-0.0711523443,0.2047266066,0.202950567,0.233322382,0.1388006508,-0.2101232409,0.5648387074,0.1000927314,-0.1552505791,-0.1335455626,-0.0250890013,-0.3282387555,-0.0922897011,0.5121243596,0.1628403068,0.1257427186,-0.0050322958,-0.0550226681,-0.1488328129,-0.2688644528,-0.0882172287,0.2847151458,0.1850690395,0.0958356038,0.3832220435,0.2154195458,0.111135751,0.0493784584,-0.1352442801,0.0353224054,0.0855564028,-0.2179709226,-0.1198172048,0.1821132153,0.3281980753,-0.2699041963,0.1215015873,-0.1425367743,-0.2085365504,0.001571518,-0.293589741,-0.0053074467,0.0134973433,-0.2574837208,0.3097872436,0.0935748518,-0.3035934269,0.0196346212,0.4677716196,-0.4638129771,-0.4341678917,-0.4358201325,0.0899604112,0.3708474934,-0.1129878536,-0.0923099965,0.059735287,0.4520861804,-0.1888256073,0.1029469818,-0.5797597766,0.3521030843,-0.0152924247,-0.0832994059,0.0905463472,-0.4035288692,-0.1816472858,0.1696708053,-0.1533035785,0.3111130297,0.4423996806,-0.3827602863,0.2947919369,0.0795271397,0.0166932028,0.8023683429,-0.3571899831,-0.2179583609,0.132254988,-0.1704488099,-0.3259194195,-0.0770724937,0.2736980319,0.0906305313,0.0717987269,-0.0094030369,0.3914262056,-0.0933154151,0.2531608641,-0.3342471123,-0.0749925673,0.3700163364,0.3320881724,0.2507386208,-0.0335372835,0.0665313303,0.3422166705,0.2596169412,-0.1454966366,0.2384007722,0.1618343294,0.0183554627,-0.1759249866,-0.1052252948,-0.1606465131,-0.3050886989,-0.2020210922,-0.3353252709,0.484374553,-0.2394975275,0.08094953,-0.1629112065,-0.0729379505,-0.1123220474,0.2390538901,0.1080784798,0.0814386383,-0.1436548978,-0.0689880103,-0.1000490189,-0.0618887804,-0.0523724556,0.1881001443,0.025137445,0.2732270956,-0.1691009253,-0.1986376792,-0.2601747513,0.0564328432,0.2558159232,0.0960626528,-0.2582112551,0.2561086416,-0.1447918415,-0.2581233084,-0.308811754,0.3103091717,0.155409649,-0.5250985622,0.0369421802,0.1614515334,0.1663212776,-0.0896117315,0.1419673562,0.1363632828,0.0686425567,0.1380071789,-0.0040352675,0.3693377972,0.2155335993,-0.0409564897,0.2696602643,0.0495535694,-0.1201176718,-0.1636262536,0.1193024367,0.0317241512,-0.2905445397,-0.0196277313,0.4356467724,-0.1525559574,0.241422236,0.4414304197,-0.1187741309,-0.0253542997,-0.1871619076,0.21855928,0.140799433,-0.0268120375,-0.0851378366,-0.0053834487,-0.253885746,-0.0652514324,0.3296323121,-0.1268236339,0.4315139949,0.1532039493,0.2892072797,-0.0184702352,-0.0514069125,-0.2919022739,0.2905476987,0.2543340623,-0.2341018468,0.4157405496,-0.1025249436,-0.2135272771,-0.6377745271,-0.3706792295,-0.0837398767,-0.1330630034,-0.0448598526,-0.0630305111,0.1643538177,0.0614361018,0.0771597028,-0.0667615905,0.2787643373,-0.1845620722,0.1062294468,-0.3318827152,-0.3990285695,-0.019308228,0.238119483,0.0182728078,-0.1148811653,0.0174304564,-0.0464675352,-0.0756031647,-0.1490138918,0.1128379405,-0.0141116986,-0.1526148766,0.1377332509,-0.0253324211,-0.2624071538,-0.227836132,0.3238838017,-0.3288731277,0.0045541548,0.4567353427,-0.1896839291,-0.0553343557,-0.0512933806,-0.3140721619,-0.5814594626,-0.0539097823,0.0172906443,0.0343012847,0.3296330273,0.3480425179,0.0307737906,0.1842597723,0.0340055116,-0.0597997718,0.0854178891,0.1192031726,0.3146250844,-0.0384516716,-0.288662076,-0.3471459448,-0.1107914522,0.0709793866,0.0762151107,-0.319314748,0.1262018532,-0.1260373443,0.0512030981,-0.2906250954,0.0220351126,0.4480994344,0.1102448702,-0.0087918788,0.2124655098,0.0139716482,0.3060884774,0.0416650251,0.2144845873,-0.034694355,0.1787259728,-0.04564083,0.6495577097,-0.1418004334,-0.3859343827,0.1767015159,-0.1262178123,-0.1793362498,-0.0094065778,-0.3002758324,-0.0031256527,-0.2587314844,-0.035383977,0.0021558774,-0.2854906023,-0.3254140913,-0.1811380535,0.1342670619,-0.1632559448,-0.3281493783,0.3672778308,-0.2507418394,0.2550690472,-0.1565136909,-0.1447717696,-0.1669237763,-0.4364982843,-0.030233575,0.1978235394,-0.299941659,0.0271620415,-0.5152040124,-0.2360682935,-0.1595627517,0.0845983028,0.3666800857,0.5396204591,0.0303010792,0.0349531136,0.0805857182,0.1010705158,0.4106414914,0.1839824766,-0.0750322565,0.1855862737,0.1655876786,-0.3579363525,-0.2835209072,-0.5400992036,-0.0822041705,0.2528803945,0.0967269838,-0.2896073461,-0.2426508218,0.2225344628,0.0414290503,-0.2890586257,-0.0435758941,0.0172090866,-0.2311727703,-0.2042599469,0.1062254459,0.2091031224,0.3887903094,-0.2306026816,-0.108919628,0.018064294,0.206056118,0.1758808345,-0.0173781198,0.1210725829,-0.1272633523,-0.2719097137,-0.2843087614,0.2212008238,0.1246747598,0.4509957135,0.1175308526,-0.3715092242,0.1879755855,-0.0572563186,0.3409126699,0.0074067381,-0.1104089469,0.3472835124,0.1715258509,0.3176763058,0.0993583575,-0.0289534982,0.2658733726,-0.077132225,-0.4949772954,-0.0418625735,0.2745336592,-0.073024258,-0.0733684525,-0.0241515674,-0.209478572,-0.2615410686,0.4185969532,0.3421198726,1.0206490755,-0.2603802681,0.3446718752,0.1092102751,-0.0606011935,0.4950573146,-0.1539972872,0.1934902817,-0.5524522066,0.0472402014,-0.1116915494,-0.1971725821,-0.1783649027,0.3437050283,-0.3830962181,0.1329992712,-0.0964831635,0.196059972,0.1422273517,0.3196293116,0.2093944252,-0.2090153843,-0.236824289,-0.0035606665,-0.4814991951,-0.0990892276,-0.0379736423,-0.2874800265,-0.0001370848,-0.3918527365,-0.2000364512,0.1904892027,-0.292277813,0.3384348452,0.5059814453,-0.1064435542,0.2228132337,0.1987271458,0.3068524003,-0.0456333794,0.061189007,-0.2703238726,-0.0626038909,0.1166037172,-0.0572850145,-0.0534290858,0.4100271761,-0.0476828143,-0.3138557673,0.3178364336,0.014437031,-0.1757712215,0.1572690457,-0.0672014877,0.0875045732,-0.4117056429,0.2714482546,-0.2188573778,-0.3185107708,-0.2686835229,0.00391638,0.0189208221,-0.1227038875,0.1043698266,0.3181701899,0.0148003632,0.1429253668,0.1666362286,0.2979536057,0.2228274792,0.5636560321,0.3159725964,0.011753371,-0.2083150148,0.1311721206,0.0069096126,-0.8838168979,0.1976176947,-0.0313806608,-0.2637524307,0.2123913914,0.3753276467,0.127044633,0.4987849891,-0.1740798354,-0.4247024059,-0.3973774314,0.0065994682,0.1742637008,0.0011970872,0.0514225103,0.5306829214,0.1982193142,0.154717207,-0.2049757689,0.1831800342,-0.3399907053,0.2554392815,-0.1108253449,-0.4195622504,-0.0381523035,0.1254320294,0.0438735113,0.001752787,0.0914920568,-0.054966893,-0.182664901,0.1793142408,0.2735623419,-0.2096082419,-0.2047402263,-0.0046507875,-0.0219300948,0.1029740423,-0.0729965493,0.007896659,-0.1585482657,0.5496968627,-0.0968374833,-0.0782949477,-0.3212496936,0.0320710763,-0.2329460979,-0.0552968457,0.1789356768,0.2310499996,0.1414344162,-0.1671888977,-0.7330248952,0.1469236463,0.2807041705,-0.0789484009,0.2824497819,-0.6361711621,-0.2515559494,-0.1240685359,0.0589370988,0.0178229157,-0.2906434834,-0.1190241799,0.1500851214,0.0792080164,-0.2228344381,-0.2199819088,-0.0071267015,-0.1568069011,-0.1895316839,0.4388341308,0.0504113249,-0.1781734973,-0.0951463059,0.0208672546,0.1502609104,-0.1083433032,0.0020464617,0.5980404615,-0.1331814378,-0.0291011278,0.4560702145,-0.0017088738,-0.0808275715,0.0111113079,-0.2403192073,0.1064087152,-0.1864718348,0.0284161344,0.2730591297,0.0053153406,-0.161443457,0.1116662025,0.0180342142,-0.1406107545,-0.2085789144,0.5285694599,-0.1294042617,-0.1111316606,-0.2098496258,0.2522187829,-0.2701119184,0.0049775327,-0.1391522586,-0.0752415732,-0.1743354499,-0.2161694765,-0.1959977895,0.1043635309,0.267588973,0.267826885,0.1731937677,-0.2923143804,0.0107252067,0.4745835066,0.0857413039,-0.2254720032,0.4008435607,0.3285622001,-0.0989184082,0.0243499521,0.3410256207,0.5116843581,0.5268767476,-0.2755404711,0.0296224151,0.0998523682,-0.1723219901,0.3009316325,0.1220268235,0.5902811885,0.1589998305,0.1322440356,-0.0237083342,-0.025082143,0.2330643088,-0.0365252122,0.1909649968,0.080381535,0.6249300241,0.2398195416,-0.3591584265,-0.1490707994,0.0843185484,-0.4341334403,0.0055399798,0.4604842663,-0.0493658446,0.2880000174,-0.0646258593,-0.0263440218,0.1581723094,0.6270230412,0.2333035618,0.2689677775,-0.3417343795,0.1156057864,-0.3256517053,0.3007265031,-0.001474376,-0.08987315,0.0791968703,0.0009102058,0.0593974143,0.1766050905,0.1285300255,0.0768041238,-0.1470921636,0.2389507443,-0.1691421568,0.0240432117,0.0808030888,0.227913022,-0.0594038852,-0.3299914598,0.2175324857,0.1071319431,-0.0194472913,-0.0732561797,0.0540491007,0.1230195463,0.1730821282,0.103905499,0.0345203616,0.5826874971,0.1241591573,-0.0638274997,0.3305566311,-0.1148174033,-0.1746624708,0.1645623893,0.2476192266,0.6241380572,-0.0343466848,0.2556886077,-0.4483470321,-0.2127869725,-0.0261634514,0.0727965012,-0.2580463588,0.2388151288,-0.2617353499,-0.0323142111,0.320294708,-0.0063781231,0.2982015908,0.0939638168,-0.4200736582,-0.4641219974,0.3670325279,-0.2881059051,-0.2306285352,-0.1451697499,0.112281464,0.0307800062,0.0831518546,-0.6610728502,-0.0098561831,0.3323136866,-0.0263023917,-0.221364066,-0.1205213144,0.0122635132,0.1851020753,-0.0976042598,0.2706556916,-0.0960784331,-0.0646676719,-0.454110235,-0.1759285182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/469","title":"invalid data type 'str' at _convert_outputs in arrow_dataset.py","comments":"I didn't know tokenizers could return strings in the token ids. Which tokenizer are you using to get this @Doragd ?","body":"I trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n","comment_length":21,"text":"invalid data type 'str' at _convert_outputs in arrow_dataset.py\nI trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n\nI didn't know tokenizers could return strings in the token ids. Which tokenizer are you using to get this @Doragd ?","embeddings":[-0.2228239626,0.0242501777,0.1073713675,0.2303070426,0.5109525919,-0.1072595492,0.5323296189,0.0775902644,-0.3410620093,-0.1247765422,0.036657881,0.3864071369,-0.3696954846,0.0422176793,0.0443617031,-0.2024641186,-0.0880346373,0.0099904612,-0.0232526027,-0.0536389053,0.0482440926,0.1829803139,-0.1568261683,0.4144774675,-0.6451317668,-0.1835785806,0.1746639758,-0.0963928327,-0.0992212296,-0.4487785995,0.2202744484,-0.1020444036,0.4976367652,0.3472318351,-0.0001241712,0.1380359232,0.1092858762,0.1016949043,-0.2908720374,-0.3016375303,0.1341145188,-0.0593848564,0.4149942696,-0.2015384734,-0.0571896322,-0.2600911558,-0.0606259108,0.2617689073,0.3837757409,0.5016983747,0.1173220798,0.1685167998,0.2464857996,0.0738129616,0.2595641315,0.3155063391,0.0754768178,0.2071238756,0.0168337896,0.0266056098,-0.0856947526,0.0748076811,-0.0701284856,-0.1997739971,0.2166583389,0.227282986,0.17457892,-0.1476221532,-0.2287852317,0.1198840439,0.4332152605,-0.3214256167,-0.1559445858,0.0079212757,-0.0111622717,-0.2727991641,-0.0061669713,0.2483308762,-0.4277769327,0.1667014062,-0.1622606069,-0.0254668966,-0.3792498708,0.199556306,-0.0925177336,0.5032640696,0.1183697209,0.1568167061,0.1842403114,-0.0144359078,0.371710062,0.0876382813,0.1397552341,0.0621479377,-0.3116942644,-0.030463824,-0.4463334084,-0.648362577,-0.2685149014,-0.2125236392,0.1655772626,-0.0711523443,0.2047266066,0.202950567,0.233322382,0.1388006508,-0.2101232409,0.5648387074,0.1000927314,-0.1552505791,-0.1335455626,-0.0250890013,-0.3282387555,-0.0922897011,0.5121243596,0.1628403068,0.1257427186,-0.0050322958,-0.0550226681,-0.1488328129,-0.2688644528,-0.0882172287,0.2847151458,0.1850690395,0.0958356038,0.3832220435,0.2154195458,0.111135751,0.0493784584,-0.1352442801,0.0353224054,0.0855564028,-0.2179709226,-0.1198172048,0.1821132153,0.3281980753,-0.2699041963,0.1215015873,-0.1425367743,-0.2085365504,0.001571518,-0.293589741,-0.0053074467,0.0134973433,-0.2574837208,0.3097872436,0.0935748518,-0.3035934269,0.0196346212,0.4677716196,-0.4638129771,-0.4341678917,-0.4358201325,0.0899604112,0.3708474934,-0.1129878536,-0.0923099965,0.059735287,0.4520861804,-0.1888256073,0.1029469818,-0.5797597766,0.3521030843,-0.0152924247,-0.0832994059,0.0905463472,-0.4035288692,-0.1816472858,0.1696708053,-0.1533035785,0.3111130297,0.4423996806,-0.3827602863,0.2947919369,0.0795271397,0.0166932028,0.8023683429,-0.3571899831,-0.2179583609,0.132254988,-0.1704488099,-0.3259194195,-0.0770724937,0.2736980319,0.0906305313,0.0717987269,-0.0094030369,0.3914262056,-0.0933154151,0.2531608641,-0.3342471123,-0.0749925673,0.3700163364,0.3320881724,0.2507386208,-0.0335372835,0.0665313303,0.3422166705,0.2596169412,-0.1454966366,0.2384007722,0.1618343294,0.0183554627,-0.1759249866,-0.1052252948,-0.1606465131,-0.3050886989,-0.2020210922,-0.3353252709,0.484374553,-0.2394975275,0.08094953,-0.1629112065,-0.0729379505,-0.1123220474,0.2390538901,0.1080784798,0.0814386383,-0.1436548978,-0.0689880103,-0.1000490189,-0.0618887804,-0.0523724556,0.1881001443,0.025137445,0.2732270956,-0.1691009253,-0.1986376792,-0.2601747513,0.0564328432,0.2558159232,0.0960626528,-0.2582112551,0.2561086416,-0.1447918415,-0.2581233084,-0.308811754,0.3103091717,0.155409649,-0.5250985622,0.0369421802,0.1614515334,0.1663212776,-0.0896117315,0.1419673562,0.1363632828,0.0686425567,0.1380071789,-0.0040352675,0.3693377972,0.2155335993,-0.0409564897,0.2696602643,0.0495535694,-0.1201176718,-0.1636262536,0.1193024367,0.0317241512,-0.2905445397,-0.0196277313,0.4356467724,-0.1525559574,0.241422236,0.4414304197,-0.1187741309,-0.0253542997,-0.1871619076,0.21855928,0.140799433,-0.0268120375,-0.0851378366,-0.0053834487,-0.253885746,-0.0652514324,0.3296323121,-0.1268236339,0.4315139949,0.1532039493,0.2892072797,-0.0184702352,-0.0514069125,-0.2919022739,0.2905476987,0.2543340623,-0.2341018468,0.4157405496,-0.1025249436,-0.2135272771,-0.6377745271,-0.3706792295,-0.0837398767,-0.1330630034,-0.0448598526,-0.0630305111,0.1643538177,0.0614361018,0.0771597028,-0.0667615905,0.2787643373,-0.1845620722,0.1062294468,-0.3318827152,-0.3990285695,-0.019308228,0.238119483,0.0182728078,-0.1148811653,0.0174304564,-0.0464675352,-0.0756031647,-0.1490138918,0.1128379405,-0.0141116986,-0.1526148766,0.1377332509,-0.0253324211,-0.2624071538,-0.227836132,0.3238838017,-0.3288731277,0.0045541548,0.4567353427,-0.1896839291,-0.0553343557,-0.0512933806,-0.3140721619,-0.5814594626,-0.0539097823,0.0172906443,0.0343012847,0.3296330273,0.3480425179,0.0307737906,0.1842597723,0.0340055116,-0.0597997718,0.0854178891,0.1192031726,0.3146250844,-0.0384516716,-0.288662076,-0.3471459448,-0.1107914522,0.0709793866,0.0762151107,-0.319314748,0.1262018532,-0.1260373443,0.0512030981,-0.2906250954,0.0220351126,0.4480994344,0.1102448702,-0.0087918788,0.2124655098,0.0139716482,0.3060884774,0.0416650251,0.2144845873,-0.034694355,0.1787259728,-0.04564083,0.6495577097,-0.1418004334,-0.3859343827,0.1767015159,-0.1262178123,-0.1793362498,-0.0094065778,-0.3002758324,-0.0031256527,-0.2587314844,-0.035383977,0.0021558774,-0.2854906023,-0.3254140913,-0.1811380535,0.1342670619,-0.1632559448,-0.3281493783,0.3672778308,-0.2507418394,0.2550690472,-0.1565136909,-0.1447717696,-0.1669237763,-0.4364982843,-0.030233575,0.1978235394,-0.299941659,0.0271620415,-0.5152040124,-0.2360682935,-0.1595627517,0.0845983028,0.3666800857,0.5396204591,0.0303010792,0.0349531136,0.0805857182,0.1010705158,0.4106414914,0.1839824766,-0.0750322565,0.1855862737,0.1655876786,-0.3579363525,-0.2835209072,-0.5400992036,-0.0822041705,0.2528803945,0.0967269838,-0.2896073461,-0.2426508218,0.2225344628,0.0414290503,-0.2890586257,-0.0435758941,0.0172090866,-0.2311727703,-0.2042599469,0.1062254459,0.2091031224,0.3887903094,-0.2306026816,-0.108919628,0.018064294,0.206056118,0.1758808345,-0.0173781198,0.1210725829,-0.1272633523,-0.2719097137,-0.2843087614,0.2212008238,0.1246747598,0.4509957135,0.1175308526,-0.3715092242,0.1879755855,-0.0572563186,0.3409126699,0.0074067381,-0.1104089469,0.3472835124,0.1715258509,0.3176763058,0.0993583575,-0.0289534982,0.2658733726,-0.077132225,-0.4949772954,-0.0418625735,0.2745336592,-0.073024258,-0.0733684525,-0.0241515674,-0.209478572,-0.2615410686,0.4185969532,0.3421198726,1.0206490755,-0.2603802681,0.3446718752,0.1092102751,-0.0606011935,0.4950573146,-0.1539972872,0.1934902817,-0.5524522066,0.0472402014,-0.1116915494,-0.1971725821,-0.1783649027,0.3437050283,-0.3830962181,0.1329992712,-0.0964831635,0.196059972,0.1422273517,0.3196293116,0.2093944252,-0.2090153843,-0.236824289,-0.0035606665,-0.4814991951,-0.0990892276,-0.0379736423,-0.2874800265,-0.0001370848,-0.3918527365,-0.2000364512,0.1904892027,-0.292277813,0.3384348452,0.5059814453,-0.1064435542,0.2228132337,0.1987271458,0.3068524003,-0.0456333794,0.061189007,-0.2703238726,-0.0626038909,0.1166037172,-0.0572850145,-0.0534290858,0.4100271761,-0.0476828143,-0.3138557673,0.3178364336,0.014437031,-0.1757712215,0.1572690457,-0.0672014877,0.0875045732,-0.4117056429,0.2714482546,-0.2188573778,-0.3185107708,-0.2686835229,0.00391638,0.0189208221,-0.1227038875,0.1043698266,0.3181701899,0.0148003632,0.1429253668,0.1666362286,0.2979536057,0.2228274792,0.5636560321,0.3159725964,0.011753371,-0.2083150148,0.1311721206,0.0069096126,-0.8838168979,0.1976176947,-0.0313806608,-0.2637524307,0.2123913914,0.3753276467,0.127044633,0.4987849891,-0.1740798354,-0.4247024059,-0.3973774314,0.0065994682,0.1742637008,0.0011970872,0.0514225103,0.5306829214,0.1982193142,0.154717207,-0.2049757689,0.1831800342,-0.3399907053,0.2554392815,-0.1108253449,-0.4195622504,-0.0381523035,0.1254320294,0.0438735113,0.001752787,0.0914920568,-0.054966893,-0.182664901,0.1793142408,0.2735623419,-0.2096082419,-0.2047402263,-0.0046507875,-0.0219300948,0.1029740423,-0.0729965493,0.007896659,-0.1585482657,0.5496968627,-0.0968374833,-0.0782949477,-0.3212496936,0.0320710763,-0.2329460979,-0.0552968457,0.1789356768,0.2310499996,0.1414344162,-0.1671888977,-0.7330248952,0.1469236463,0.2807041705,-0.0789484009,0.2824497819,-0.6361711621,-0.2515559494,-0.1240685359,0.0589370988,0.0178229157,-0.2906434834,-0.1190241799,0.1500851214,0.0792080164,-0.2228344381,-0.2199819088,-0.0071267015,-0.1568069011,-0.1895316839,0.4388341308,0.0504113249,-0.1781734973,-0.0951463059,0.0208672546,0.1502609104,-0.1083433032,0.0020464617,0.5980404615,-0.1331814378,-0.0291011278,0.4560702145,-0.0017088738,-0.0808275715,0.0111113079,-0.2403192073,0.1064087152,-0.1864718348,0.0284161344,0.2730591297,0.0053153406,-0.161443457,0.1116662025,0.0180342142,-0.1406107545,-0.2085789144,0.5285694599,-0.1294042617,-0.1111316606,-0.2098496258,0.2522187829,-0.2701119184,0.0049775327,-0.1391522586,-0.0752415732,-0.1743354499,-0.2161694765,-0.1959977895,0.1043635309,0.267588973,0.267826885,0.1731937677,-0.2923143804,0.0107252067,0.4745835066,0.0857413039,-0.2254720032,0.4008435607,0.3285622001,-0.0989184082,0.0243499521,0.3410256207,0.5116843581,0.5268767476,-0.2755404711,0.0296224151,0.0998523682,-0.1723219901,0.3009316325,0.1220268235,0.5902811885,0.1589998305,0.1322440356,-0.0237083342,-0.025082143,0.2330643088,-0.0365252122,0.1909649968,0.080381535,0.6249300241,0.2398195416,-0.3591584265,-0.1490707994,0.0843185484,-0.4341334403,0.0055399798,0.4604842663,-0.0493658446,0.2880000174,-0.0646258593,-0.0263440218,0.1581723094,0.6270230412,0.2333035618,0.2689677775,-0.3417343795,0.1156057864,-0.3256517053,0.3007265031,-0.001474376,-0.08987315,0.0791968703,0.0009102058,0.0593974143,0.1766050905,0.1285300255,0.0768041238,-0.1470921636,0.2389507443,-0.1691421568,0.0240432117,0.0808030888,0.227913022,-0.0594038852,-0.3299914598,0.2175324857,0.1071319431,-0.0194472913,-0.0732561797,0.0540491007,0.1230195463,0.1730821282,0.103905499,0.0345203616,0.5826874971,0.1241591573,-0.0638274997,0.3305566311,-0.1148174033,-0.1746624708,0.1645623893,0.2476192266,0.6241380572,-0.0343466848,0.2556886077,-0.4483470321,-0.2127869725,-0.0261634514,0.0727965012,-0.2580463588,0.2388151288,-0.2617353499,-0.0323142111,0.320294708,-0.0063781231,0.2982015908,0.0939638168,-0.4200736582,-0.4641219974,0.3670325279,-0.2881059051,-0.2306285352,-0.1451697499,0.112281464,0.0307800062,0.0831518546,-0.6610728502,-0.0098561831,0.3323136866,-0.0263023917,-0.221364066,-0.1205213144,0.0122635132,0.1851020753,-0.0976042598,0.2706556916,-0.0960784331,-0.0646676719,-0.454110235,-0.1759285182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/469","title":"invalid data type 'str' at _convert_outputs in arrow_dataset.py","comments":"> I didn't know tokenizers could return strings in the token ids. Which tokenizer are you using to get this @Doragd ?\r\n\r\ni'm sorry that i met this issue in another place (not in huggingface repo). ","body":"I trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n","comment_length":36,"text":"invalid data type 'str' at _convert_outputs in arrow_dataset.py\nI trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n\n> I didn't know tokenizers could return strings in the token ids. Which tokenizer are you using to get this @Doragd ?\r\n\r\ni'm sorry that i met this issue in another place (not in huggingface repo). ","embeddings":[-0.2228239626,0.0242501777,0.1073713675,0.2303070426,0.5109525919,-0.1072595492,0.5323296189,0.0775902644,-0.3410620093,-0.1247765422,0.036657881,0.3864071369,-0.3696954846,0.0422176793,0.0443617031,-0.2024641186,-0.0880346373,0.0099904612,-0.0232526027,-0.0536389053,0.0482440926,0.1829803139,-0.1568261683,0.4144774675,-0.6451317668,-0.1835785806,0.1746639758,-0.0963928327,-0.0992212296,-0.4487785995,0.2202744484,-0.1020444036,0.4976367652,0.3472318351,-0.0001241712,0.1380359232,0.1092858762,0.1016949043,-0.2908720374,-0.3016375303,0.1341145188,-0.0593848564,0.4149942696,-0.2015384734,-0.0571896322,-0.2600911558,-0.0606259108,0.2617689073,0.3837757409,0.5016983747,0.1173220798,0.1685167998,0.2464857996,0.0738129616,0.2595641315,0.3155063391,0.0754768178,0.2071238756,0.0168337896,0.0266056098,-0.0856947526,0.0748076811,-0.0701284856,-0.1997739971,0.2166583389,0.227282986,0.17457892,-0.1476221532,-0.2287852317,0.1198840439,0.4332152605,-0.3214256167,-0.1559445858,0.0079212757,-0.0111622717,-0.2727991641,-0.0061669713,0.2483308762,-0.4277769327,0.1667014062,-0.1622606069,-0.0254668966,-0.3792498708,0.199556306,-0.0925177336,0.5032640696,0.1183697209,0.1568167061,0.1842403114,-0.0144359078,0.371710062,0.0876382813,0.1397552341,0.0621479377,-0.3116942644,-0.030463824,-0.4463334084,-0.648362577,-0.2685149014,-0.2125236392,0.1655772626,-0.0711523443,0.2047266066,0.202950567,0.233322382,0.1388006508,-0.2101232409,0.5648387074,0.1000927314,-0.1552505791,-0.1335455626,-0.0250890013,-0.3282387555,-0.0922897011,0.5121243596,0.1628403068,0.1257427186,-0.0050322958,-0.0550226681,-0.1488328129,-0.2688644528,-0.0882172287,0.2847151458,0.1850690395,0.0958356038,0.3832220435,0.2154195458,0.111135751,0.0493784584,-0.1352442801,0.0353224054,0.0855564028,-0.2179709226,-0.1198172048,0.1821132153,0.3281980753,-0.2699041963,0.1215015873,-0.1425367743,-0.2085365504,0.001571518,-0.293589741,-0.0053074467,0.0134973433,-0.2574837208,0.3097872436,0.0935748518,-0.3035934269,0.0196346212,0.4677716196,-0.4638129771,-0.4341678917,-0.4358201325,0.0899604112,0.3708474934,-0.1129878536,-0.0923099965,0.059735287,0.4520861804,-0.1888256073,0.1029469818,-0.5797597766,0.3521030843,-0.0152924247,-0.0832994059,0.0905463472,-0.4035288692,-0.1816472858,0.1696708053,-0.1533035785,0.3111130297,0.4423996806,-0.3827602863,0.2947919369,0.0795271397,0.0166932028,0.8023683429,-0.3571899831,-0.2179583609,0.132254988,-0.1704488099,-0.3259194195,-0.0770724937,0.2736980319,0.0906305313,0.0717987269,-0.0094030369,0.3914262056,-0.0933154151,0.2531608641,-0.3342471123,-0.0749925673,0.3700163364,0.3320881724,0.2507386208,-0.0335372835,0.0665313303,0.3422166705,0.2596169412,-0.1454966366,0.2384007722,0.1618343294,0.0183554627,-0.1759249866,-0.1052252948,-0.1606465131,-0.3050886989,-0.2020210922,-0.3353252709,0.484374553,-0.2394975275,0.08094953,-0.1629112065,-0.0729379505,-0.1123220474,0.2390538901,0.1080784798,0.0814386383,-0.1436548978,-0.0689880103,-0.1000490189,-0.0618887804,-0.0523724556,0.1881001443,0.025137445,0.2732270956,-0.1691009253,-0.1986376792,-0.2601747513,0.0564328432,0.2558159232,0.0960626528,-0.2582112551,0.2561086416,-0.1447918415,-0.2581233084,-0.308811754,0.3103091717,0.155409649,-0.5250985622,0.0369421802,0.1614515334,0.1663212776,-0.0896117315,0.1419673562,0.1363632828,0.0686425567,0.1380071789,-0.0040352675,0.3693377972,0.2155335993,-0.0409564897,0.2696602643,0.0495535694,-0.1201176718,-0.1636262536,0.1193024367,0.0317241512,-0.2905445397,-0.0196277313,0.4356467724,-0.1525559574,0.241422236,0.4414304197,-0.1187741309,-0.0253542997,-0.1871619076,0.21855928,0.140799433,-0.0268120375,-0.0851378366,-0.0053834487,-0.253885746,-0.0652514324,0.3296323121,-0.1268236339,0.4315139949,0.1532039493,0.2892072797,-0.0184702352,-0.0514069125,-0.2919022739,0.2905476987,0.2543340623,-0.2341018468,0.4157405496,-0.1025249436,-0.2135272771,-0.6377745271,-0.3706792295,-0.0837398767,-0.1330630034,-0.0448598526,-0.0630305111,0.1643538177,0.0614361018,0.0771597028,-0.0667615905,0.2787643373,-0.1845620722,0.1062294468,-0.3318827152,-0.3990285695,-0.019308228,0.238119483,0.0182728078,-0.1148811653,0.0174304564,-0.0464675352,-0.0756031647,-0.1490138918,0.1128379405,-0.0141116986,-0.1526148766,0.1377332509,-0.0253324211,-0.2624071538,-0.227836132,0.3238838017,-0.3288731277,0.0045541548,0.4567353427,-0.1896839291,-0.0553343557,-0.0512933806,-0.3140721619,-0.5814594626,-0.0539097823,0.0172906443,0.0343012847,0.3296330273,0.3480425179,0.0307737906,0.1842597723,0.0340055116,-0.0597997718,0.0854178891,0.1192031726,0.3146250844,-0.0384516716,-0.288662076,-0.3471459448,-0.1107914522,0.0709793866,0.0762151107,-0.319314748,0.1262018532,-0.1260373443,0.0512030981,-0.2906250954,0.0220351126,0.4480994344,0.1102448702,-0.0087918788,0.2124655098,0.0139716482,0.3060884774,0.0416650251,0.2144845873,-0.034694355,0.1787259728,-0.04564083,0.6495577097,-0.1418004334,-0.3859343827,0.1767015159,-0.1262178123,-0.1793362498,-0.0094065778,-0.3002758324,-0.0031256527,-0.2587314844,-0.035383977,0.0021558774,-0.2854906023,-0.3254140913,-0.1811380535,0.1342670619,-0.1632559448,-0.3281493783,0.3672778308,-0.2507418394,0.2550690472,-0.1565136909,-0.1447717696,-0.1669237763,-0.4364982843,-0.030233575,0.1978235394,-0.299941659,0.0271620415,-0.5152040124,-0.2360682935,-0.1595627517,0.0845983028,0.3666800857,0.5396204591,0.0303010792,0.0349531136,0.0805857182,0.1010705158,0.4106414914,0.1839824766,-0.0750322565,0.1855862737,0.1655876786,-0.3579363525,-0.2835209072,-0.5400992036,-0.0822041705,0.2528803945,0.0967269838,-0.2896073461,-0.2426508218,0.2225344628,0.0414290503,-0.2890586257,-0.0435758941,0.0172090866,-0.2311727703,-0.2042599469,0.1062254459,0.2091031224,0.3887903094,-0.2306026816,-0.108919628,0.018064294,0.206056118,0.1758808345,-0.0173781198,0.1210725829,-0.1272633523,-0.2719097137,-0.2843087614,0.2212008238,0.1246747598,0.4509957135,0.1175308526,-0.3715092242,0.1879755855,-0.0572563186,0.3409126699,0.0074067381,-0.1104089469,0.3472835124,0.1715258509,0.3176763058,0.0993583575,-0.0289534982,0.2658733726,-0.077132225,-0.4949772954,-0.0418625735,0.2745336592,-0.073024258,-0.0733684525,-0.0241515674,-0.209478572,-0.2615410686,0.4185969532,0.3421198726,1.0206490755,-0.2603802681,0.3446718752,0.1092102751,-0.0606011935,0.4950573146,-0.1539972872,0.1934902817,-0.5524522066,0.0472402014,-0.1116915494,-0.1971725821,-0.1783649027,0.3437050283,-0.3830962181,0.1329992712,-0.0964831635,0.196059972,0.1422273517,0.3196293116,0.2093944252,-0.2090153843,-0.236824289,-0.0035606665,-0.4814991951,-0.0990892276,-0.0379736423,-0.2874800265,-0.0001370848,-0.3918527365,-0.2000364512,0.1904892027,-0.292277813,0.3384348452,0.5059814453,-0.1064435542,0.2228132337,0.1987271458,0.3068524003,-0.0456333794,0.061189007,-0.2703238726,-0.0626038909,0.1166037172,-0.0572850145,-0.0534290858,0.4100271761,-0.0476828143,-0.3138557673,0.3178364336,0.014437031,-0.1757712215,0.1572690457,-0.0672014877,0.0875045732,-0.4117056429,0.2714482546,-0.2188573778,-0.3185107708,-0.2686835229,0.00391638,0.0189208221,-0.1227038875,0.1043698266,0.3181701899,0.0148003632,0.1429253668,0.1666362286,0.2979536057,0.2228274792,0.5636560321,0.3159725964,0.011753371,-0.2083150148,0.1311721206,0.0069096126,-0.8838168979,0.1976176947,-0.0313806608,-0.2637524307,0.2123913914,0.3753276467,0.127044633,0.4987849891,-0.1740798354,-0.4247024059,-0.3973774314,0.0065994682,0.1742637008,0.0011970872,0.0514225103,0.5306829214,0.1982193142,0.154717207,-0.2049757689,0.1831800342,-0.3399907053,0.2554392815,-0.1108253449,-0.4195622504,-0.0381523035,0.1254320294,0.0438735113,0.001752787,0.0914920568,-0.054966893,-0.182664901,0.1793142408,0.2735623419,-0.2096082419,-0.2047402263,-0.0046507875,-0.0219300948,0.1029740423,-0.0729965493,0.007896659,-0.1585482657,0.5496968627,-0.0968374833,-0.0782949477,-0.3212496936,0.0320710763,-0.2329460979,-0.0552968457,0.1789356768,0.2310499996,0.1414344162,-0.1671888977,-0.7330248952,0.1469236463,0.2807041705,-0.0789484009,0.2824497819,-0.6361711621,-0.2515559494,-0.1240685359,0.0589370988,0.0178229157,-0.2906434834,-0.1190241799,0.1500851214,0.0792080164,-0.2228344381,-0.2199819088,-0.0071267015,-0.1568069011,-0.1895316839,0.4388341308,0.0504113249,-0.1781734973,-0.0951463059,0.0208672546,0.1502609104,-0.1083433032,0.0020464617,0.5980404615,-0.1331814378,-0.0291011278,0.4560702145,-0.0017088738,-0.0808275715,0.0111113079,-0.2403192073,0.1064087152,-0.1864718348,0.0284161344,0.2730591297,0.0053153406,-0.161443457,0.1116662025,0.0180342142,-0.1406107545,-0.2085789144,0.5285694599,-0.1294042617,-0.1111316606,-0.2098496258,0.2522187829,-0.2701119184,0.0049775327,-0.1391522586,-0.0752415732,-0.1743354499,-0.2161694765,-0.1959977895,0.1043635309,0.267588973,0.267826885,0.1731937677,-0.2923143804,0.0107252067,0.4745835066,0.0857413039,-0.2254720032,0.4008435607,0.3285622001,-0.0989184082,0.0243499521,0.3410256207,0.5116843581,0.5268767476,-0.2755404711,0.0296224151,0.0998523682,-0.1723219901,0.3009316325,0.1220268235,0.5902811885,0.1589998305,0.1322440356,-0.0237083342,-0.025082143,0.2330643088,-0.0365252122,0.1909649968,0.080381535,0.6249300241,0.2398195416,-0.3591584265,-0.1490707994,0.0843185484,-0.4341334403,0.0055399798,0.4604842663,-0.0493658446,0.2880000174,-0.0646258593,-0.0263440218,0.1581723094,0.6270230412,0.2333035618,0.2689677775,-0.3417343795,0.1156057864,-0.3256517053,0.3007265031,-0.001474376,-0.08987315,0.0791968703,0.0009102058,0.0593974143,0.1766050905,0.1285300255,0.0768041238,-0.1470921636,0.2389507443,-0.1691421568,0.0240432117,0.0808030888,0.227913022,-0.0594038852,-0.3299914598,0.2175324857,0.1071319431,-0.0194472913,-0.0732561797,0.0540491007,0.1230195463,0.1730821282,0.103905499,0.0345203616,0.5826874971,0.1241591573,-0.0638274997,0.3305566311,-0.1148174033,-0.1746624708,0.1645623893,0.2476192266,0.6241380572,-0.0343466848,0.2556886077,-0.4483470321,-0.2127869725,-0.0261634514,0.0727965012,-0.2580463588,0.2388151288,-0.2617353499,-0.0323142111,0.320294708,-0.0063781231,0.2982015908,0.0939638168,-0.4200736582,-0.4641219974,0.3670325279,-0.2881059051,-0.2306285352,-0.1451697499,0.112281464,0.0307800062,0.0831518546,-0.6610728502,-0.0098561831,0.3323136866,-0.0263023917,-0.221364066,-0.1205213144,0.0122635132,0.1851020753,-0.0976042598,0.2706556916,-0.0960784331,-0.0646676719,-0.454110235,-0.1759285182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/469","title":"invalid data type 'str' at _convert_outputs in arrow_dataset.py","comments":"@akhilkapil do you have strings in your dataset ? When you set the dataset format to \"pytorch\" you should exclude columns with strings as pytorch can't make tensors out of strings","body":"I trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n","comment_length":31,"text":"invalid data type 'str' at _convert_outputs in arrow_dataset.py\nI trying to build multi label text classifier model using Transformers lib. \r\n\r\nI'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error \r\n\r\nFile \"C:\\***\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n\r\nI'm using pyarrow 1.0.0. And I have simple custom data set with Text and Integer Label. \r\nEx: Data\r\n Text , Label #Column Header\r\n I'm facing an Network issue, 1\r\n I forgot my password, 2\r\n\r\nError StackTrace:\r\n\r\nFile \"C:\\**\\transformers\\trainer.py\", line 492, in train\r\n for step, inputs in enumerate(epoch_iterator):\r\n File \"C:\\**\\tqdm\\std.py\", line 1104, in __iter__\r\n for obj in iterable:\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"C:\\**\\torch\\utils\\data\\dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\torch\\utils\\data\\_utils\\fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 414, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 403, in _getitem\r\n outputs, format_type=format_type, format_columns=format_columns, output_all_columns=output_all_columns\r\n File \"C:\\**\\nlp\\arrow_dataset.py\", line 343, in _convert_outputs\r\n v = command(v)\r\nTypeError: new(): invalid data type 'str'\r\n \r\n\n@akhilkapil do you have strings in your dataset ? When you set the dataset format to \"pytorch\" you should exclude columns with strings as pytorch can't make tensors out of strings","embeddings":[-0.2228239626,0.0242501777,0.1073713675,0.2303070426,0.5109525919,-0.1072595492,0.5323296189,0.0775902644,-0.3410620093,-0.1247765422,0.036657881,0.3864071369,-0.3696954846,0.0422176793,0.0443617031,-0.2024641186,-0.0880346373,0.0099904612,-0.0232526027,-0.0536389053,0.0482440926,0.1829803139,-0.1568261683,0.4144774675,-0.6451317668,-0.1835785806,0.1746639758,-0.0963928327,-0.0992212296,-0.4487785995,0.2202744484,-0.1020444036,0.4976367652,0.3472318351,-0.0001241712,0.1380359232,0.1092858762,0.1016949043,-0.2908720374,-0.3016375303,0.1341145188,-0.0593848564,0.4149942696,-0.2015384734,-0.0571896322,-0.2600911558,-0.0606259108,0.2617689073,0.3837757409,0.5016983747,0.1173220798,0.1685167998,0.2464857996,0.0738129616,0.2595641315,0.3155063391,0.0754768178,0.2071238756,0.0168337896,0.0266056098,-0.0856947526,0.0748076811,-0.0701284856,-0.1997739971,0.2166583389,0.227282986,0.17457892,-0.1476221532,-0.2287852317,0.1198840439,0.4332152605,-0.3214256167,-0.1559445858,0.0079212757,-0.0111622717,-0.2727991641,-0.0061669713,0.2483308762,-0.4277769327,0.1667014062,-0.1622606069,-0.0254668966,-0.3792498708,0.199556306,-0.0925177336,0.5032640696,0.1183697209,0.1568167061,0.1842403114,-0.0144359078,0.371710062,0.0876382813,0.1397552341,0.0621479377,-0.3116942644,-0.030463824,-0.4463334084,-0.648362577,-0.2685149014,-0.2125236392,0.1655772626,-0.0711523443,0.2047266066,0.202950567,0.233322382,0.1388006508,-0.2101232409,0.5648387074,0.1000927314,-0.1552505791,-0.1335455626,-0.0250890013,-0.3282387555,-0.0922897011,0.5121243596,0.1628403068,0.1257427186,-0.0050322958,-0.0550226681,-0.1488328129,-0.2688644528,-0.0882172287,0.2847151458,0.1850690395,0.0958356038,0.3832220435,0.2154195458,0.111135751,0.0493784584,-0.1352442801,0.0353224054,0.0855564028,-0.2179709226,-0.1198172048,0.1821132153,0.3281980753,-0.2699041963,0.1215015873,-0.1425367743,-0.2085365504,0.001571518,-0.293589741,-0.0053074467,0.0134973433,-0.2574837208,0.3097872436,0.0935748518,-0.3035934269,0.0196346212,0.4677716196,-0.4638129771,-0.4341678917,-0.4358201325,0.0899604112,0.3708474934,-0.1129878536,-0.0923099965,0.059735287,0.4520861804,-0.1888256073,0.1029469818,-0.5797597766,0.3521030843,-0.0152924247,-0.0832994059,0.0905463472,-0.4035288692,-0.1816472858,0.1696708053,-0.1533035785,0.3111130297,0.4423996806,-0.3827602863,0.2947919369,0.0795271397,0.0166932028,0.8023683429,-0.3571899831,-0.2179583609,0.132254988,-0.1704488099,-0.3259194195,-0.0770724937,0.2736980319,0.0906305313,0.0717987269,-0.0094030369,0.3914262056,-0.0933154151,0.2531608641,-0.3342471123,-0.0749925673,0.3700163364,0.3320881724,0.2507386208,-0.0335372835,0.0665313303,0.3422166705,0.2596169412,-0.1454966366,0.2384007722,0.1618343294,0.0183554627,-0.1759249866,-0.1052252948,-0.1606465131,-0.3050886989,-0.2020210922,-0.3353252709,0.484374553,-0.2394975275,0.08094953,-0.1629112065,-0.0729379505,-0.1123220474,0.2390538901,0.1080784798,0.0814386383,-0.1436548978,-0.0689880103,-0.1000490189,-0.0618887804,-0.0523724556,0.1881001443,0.025137445,0.2732270956,-0.1691009253,-0.1986376792,-0.2601747513,0.0564328432,0.2558159232,0.0960626528,-0.2582112551,0.2561086416,-0.1447918415,-0.2581233084,-0.308811754,0.3103091717,0.155409649,-0.5250985622,0.0369421802,0.1614515334,0.1663212776,-0.0896117315,0.1419673562,0.1363632828,0.0686425567,0.1380071789,-0.0040352675,0.3693377972,0.2155335993,-0.0409564897,0.2696602643,0.0495535694,-0.1201176718,-0.1636262536,0.1193024367,0.0317241512,-0.2905445397,-0.0196277313,0.4356467724,-0.1525559574,0.241422236,0.4414304197,-0.1187741309,-0.0253542997,-0.1871619076,0.21855928,0.140799433,-0.0268120375,-0.0851378366,-0.0053834487,-0.253885746,-0.0652514324,0.3296323121,-0.1268236339,0.4315139949,0.1532039493,0.2892072797,-0.0184702352,-0.0514069125,-0.2919022739,0.2905476987,0.2543340623,-0.2341018468,0.4157405496,-0.1025249436,-0.2135272771,-0.6377745271,-0.3706792295,-0.0837398767,-0.1330630034,-0.0448598526,-0.0630305111,0.1643538177,0.0614361018,0.0771597028,-0.0667615905,0.2787643373,-0.1845620722,0.1062294468,-0.3318827152,-0.3990285695,-0.019308228,0.238119483,0.0182728078,-0.1148811653,0.0174304564,-0.0464675352,-0.0756031647,-0.1490138918,0.1128379405,-0.0141116986,-0.1526148766,0.1377332509,-0.0253324211,-0.2624071538,-0.227836132,0.3238838017,-0.3288731277,0.0045541548,0.4567353427,-0.1896839291,-0.0553343557,-0.0512933806,-0.3140721619,-0.5814594626,-0.0539097823,0.0172906443,0.0343012847,0.3296330273,0.3480425179,0.0307737906,0.1842597723,0.0340055116,-0.0597997718,0.0854178891,0.1192031726,0.3146250844,-0.0384516716,-0.288662076,-0.3471459448,-0.1107914522,0.0709793866,0.0762151107,-0.319314748,0.1262018532,-0.1260373443,0.0512030981,-0.2906250954,0.0220351126,0.4480994344,0.1102448702,-0.0087918788,0.2124655098,0.0139716482,0.3060884774,0.0416650251,0.2144845873,-0.034694355,0.1787259728,-0.04564083,0.6495577097,-0.1418004334,-0.3859343827,0.1767015159,-0.1262178123,-0.1793362498,-0.0094065778,-0.3002758324,-0.0031256527,-0.2587314844,-0.035383977,0.0021558774,-0.2854906023,-0.3254140913,-0.1811380535,0.1342670619,-0.1632559448,-0.3281493783,0.3672778308,-0.2507418394,0.2550690472,-0.1565136909,-0.1447717696,-0.1669237763,-0.4364982843,-0.030233575,0.1978235394,-0.299941659,0.0271620415,-0.5152040124,-0.2360682935,-0.1595627517,0.0845983028,0.3666800857,0.5396204591,0.0303010792,0.0349531136,0.0805857182,0.1010705158,0.4106414914,0.1839824766,-0.0750322565,0.1855862737,0.1655876786,-0.3579363525,-0.2835209072,-0.5400992036,-0.0822041705,0.2528803945,0.0967269838,-0.2896073461,-0.2426508218,0.2225344628,0.0414290503,-0.2890586257,-0.0435758941,0.0172090866,-0.2311727703,-0.2042599469,0.1062254459,0.2091031224,0.3887903094,-0.2306026816,-0.108919628,0.018064294,0.206056118,0.1758808345,-0.0173781198,0.1210725829,-0.1272633523,-0.2719097137,-0.2843087614,0.2212008238,0.1246747598,0.4509957135,0.1175308526,-0.3715092242,0.1879755855,-0.0572563186,0.3409126699,0.0074067381,-0.1104089469,0.3472835124,0.1715258509,0.3176763058,0.0993583575,-0.0289534982,0.2658733726,-0.077132225,-0.4949772954,-0.0418625735,0.2745336592,-0.073024258,-0.0733684525,-0.0241515674,-0.209478572,-0.2615410686,0.4185969532,0.3421198726,1.0206490755,-0.2603802681,0.3446718752,0.1092102751,-0.0606011935,0.4950573146,-0.1539972872,0.1934902817,-0.5524522066,0.0472402014,-0.1116915494,-0.1971725821,-0.1783649027,0.3437050283,-0.3830962181,0.1329992712,-0.0964831635,0.196059972,0.1422273517,0.3196293116,0.2093944252,-0.2090153843,-0.236824289,-0.0035606665,-0.4814991951,-0.0990892276,-0.0379736423,-0.2874800265,-0.0001370848,-0.3918527365,-0.2000364512,0.1904892027,-0.292277813,0.3384348452,0.5059814453,-0.1064435542,0.2228132337,0.1987271458,0.3068524003,-0.0456333794,0.061189007,-0.2703238726,-0.0626038909,0.1166037172,-0.0572850145,-0.0534290858,0.4100271761,-0.0476828143,-0.3138557673,0.3178364336,0.014437031,-0.1757712215,0.1572690457,-0.0672014877,0.0875045732,-0.4117056429,0.2714482546,-0.2188573778,-0.3185107708,-0.2686835229,0.00391638,0.0189208221,-0.1227038875,0.1043698266,0.3181701899,0.0148003632,0.1429253668,0.1666362286,0.2979536057,0.2228274792,0.5636560321,0.3159725964,0.011753371,-0.2083150148,0.1311721206,0.0069096126,-0.8838168979,0.1976176947,-0.0313806608,-0.2637524307,0.2123913914,0.3753276467,0.127044633,0.4987849891,-0.1740798354,-0.4247024059,-0.3973774314,0.0065994682,0.1742637008,0.0011970872,0.0514225103,0.5306829214,0.1982193142,0.154717207,-0.2049757689,0.1831800342,-0.3399907053,0.2554392815,-0.1108253449,-0.4195622504,-0.0381523035,0.1254320294,0.0438735113,0.001752787,0.0914920568,-0.054966893,-0.182664901,0.1793142408,0.2735623419,-0.2096082419,-0.2047402263,-0.0046507875,-0.0219300948,0.1029740423,-0.0729965493,0.007896659,-0.1585482657,0.5496968627,-0.0968374833,-0.0782949477,-0.3212496936,0.0320710763,-0.2329460979,-0.0552968457,0.1789356768,0.2310499996,0.1414344162,-0.1671888977,-0.7330248952,0.1469236463,0.2807041705,-0.0789484009,0.2824497819,-0.6361711621,-0.2515559494,-0.1240685359,0.0589370988,0.0178229157,-0.2906434834,-0.1190241799,0.1500851214,0.0792080164,-0.2228344381,-0.2199819088,-0.0071267015,-0.1568069011,-0.1895316839,0.4388341308,0.0504113249,-0.1781734973,-0.0951463059,0.0208672546,0.1502609104,-0.1083433032,0.0020464617,0.5980404615,-0.1331814378,-0.0291011278,0.4560702145,-0.0017088738,-0.0808275715,0.0111113079,-0.2403192073,0.1064087152,-0.1864718348,0.0284161344,0.2730591297,0.0053153406,-0.161443457,0.1116662025,0.0180342142,-0.1406107545,-0.2085789144,0.5285694599,-0.1294042617,-0.1111316606,-0.2098496258,0.2522187829,-0.2701119184,0.0049775327,-0.1391522586,-0.0752415732,-0.1743354499,-0.2161694765,-0.1959977895,0.1043635309,0.267588973,0.267826885,0.1731937677,-0.2923143804,0.0107252067,0.4745835066,0.0857413039,-0.2254720032,0.4008435607,0.3285622001,-0.0989184082,0.0243499521,0.3410256207,0.5116843581,0.5268767476,-0.2755404711,0.0296224151,0.0998523682,-0.1723219901,0.3009316325,0.1220268235,0.5902811885,0.1589998305,0.1322440356,-0.0237083342,-0.025082143,0.2330643088,-0.0365252122,0.1909649968,0.080381535,0.6249300241,0.2398195416,-0.3591584265,-0.1490707994,0.0843185484,-0.4341334403,0.0055399798,0.4604842663,-0.0493658446,0.2880000174,-0.0646258593,-0.0263440218,0.1581723094,0.6270230412,0.2333035618,0.2689677775,-0.3417343795,0.1156057864,-0.3256517053,0.3007265031,-0.001474376,-0.08987315,0.0791968703,0.0009102058,0.0593974143,0.1766050905,0.1285300255,0.0768041238,-0.1470921636,0.2389507443,-0.1691421568,0.0240432117,0.0808030888,0.227913022,-0.0594038852,-0.3299914598,0.2175324857,0.1071319431,-0.0194472913,-0.0732561797,0.0540491007,0.1230195463,0.1730821282,0.103905499,0.0345203616,0.5826874971,0.1241591573,-0.0638274997,0.3305566311,-0.1148174033,-0.1746624708,0.1645623893,0.2476192266,0.6241380572,-0.0343466848,0.2556886077,-0.4483470321,-0.2127869725,-0.0261634514,0.0727965012,-0.2580463588,0.2388151288,-0.2617353499,-0.0323142111,0.320294708,-0.0063781231,0.2982015908,0.0939638168,-0.4200736582,-0.4641219974,0.3670325279,-0.2881059051,-0.2306285352,-0.1451697499,0.112281464,0.0307800062,0.0831518546,-0.6610728502,-0.0098561831,0.3323136866,-0.0263023917,-0.221364066,-0.1205213144,0.0122635132,0.1851020753,-0.0976042598,0.2706556916,-0.0960784331,-0.0646676719,-0.454110235,-0.1759285182]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/468","title":"UnicodeDecodeError while loading PAN-X task of XTREME dataset","comments":"Indeed. Solution 1 is the simplest.\r\n\r\nThis is actually a recurring problem.\r\nI think we should scan all the datasets with regexpr to fix the use of `open()` without encodings.\r\nAnd probably add a test in the CI to forbid using this in the future.","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!","comment_length":45,"text":"UnicodeDecodeError while loading PAN-X task of XTREME dataset\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!\nIndeed. Solution 1 is the simplest.\r\n\r\nThis is actually a recurring problem.\r\nI think we should scan all the datasets with regexpr to fix the use of `open()` without encodings.\r\nAnd probably add a test in the CI to forbid using this in the future.","embeddings":[-0.3326011896,-0.0930106342,-0.1336479485,0.2554301322,0.3521757722,-0.0277493317,0.2249831408,0.3496463001,-0.0286989305,0.038228903,-0.0904103443,0.1502787173,-0.0761749223,-0.1970372051,0.1891056299,-0.2467030734,-0.0738269314,0.1824576855,0.1078069955,0.0069923094,-0.3558878899,0.1144126877,-0.1869958341,0.2637093067,-0.3771478236,0.0810821429,0.2978654206,0.3204651177,-0.193280682,-0.6116119623,0.4130367041,-0.0536639951,0.4325025082,0.3163448274,-0.0001086041,0.0825855061,0.3929110467,-0.1319698691,-0.3346782327,-0.1939976513,-0.5915244818,-0.1812518835,-0.0098235868,-0.2595872581,0.1902403831,0.0562078469,-0.0642362684,-0.3830473125,0.1811341196,0.2071898282,0.24163872,0.3705638647,0.0213383194,0.0912284628,-0.045574829,0.1074261367,-0.1542558074,0.249418959,0.4346242249,0.0627749935,-0.019149052,0.2926935852,-0.0486163385,0.248120591,-0.1023555249,0.1306457818,0.1399910003,-0.4660047889,0.1134835854,0.0036376556,0.2718741894,-0.3664452136,-0.2865583897,-0.2252654582,-0.1094700545,-0.5443437099,0.2644951046,0.0905644149,-0.159659639,-0.0211321656,0.113159202,-0.0423871912,-0.1735193729,0.0391214527,0.0543593504,0.3878184855,-0.0600738861,-0.0513223708,-0.0941757336,-0.1786022931,0.4082849622,-0.1458789855,-0.1436942965,0.324629128,-0.2863637805,0.072422564,-0.0550461523,-0.3371772468,0.0668165982,-0.0463103987,-0.078823179,0.2744791806,-0.0495554432,0.053195674,0.3093065917,0.1556752175,0.2095072418,0.2245565802,0.1834983528,0.1982241273,0.0829440206,-0.0012048965,-0.1758536994,-0.1771315336,0.3654833138,0.1934714913,0.5531736016,-0.3191607296,-0.1221157163,0.0084841102,0.1728852391,-0.1787237525,0.0444031768,0.0144082429,0.2223311216,0.5070413947,-0.2224069536,0.2185738385,-0.4622749686,-0.1283478737,-0.1358893216,0.1796104163,-0.1460784674,-0.0783130527,0.0963894874,0.1252889782,0.163143605,-0.1422412395,0.3579080105,-0.0693807527,0.0396495759,-0.2039281875,-0.2201946378,0.0137812095,0.1307834387,0.1039626971,0.0551503152,-0.1753356308,-0.0369680114,0.2171099186,-0.1871833205,-0.2070101798,-0.0715540051,0.2629823387,0.0267733559,0.025153663,-0.4134668708,-0.0887362659,0.3430744112,0.062088266,0.1139909923,-0.2957776785,0.0404423103,-0.0858764797,0.2037117779,0.3430534005,-0.1321642697,0.0464633144,0.1779045761,-0.1550109833,0.2246282697,0.296744436,-0.1922280043,0.0937530622,-0.0989214331,0.5539204478,0.5879890919,-0.4102394879,-0.4520799518,0.1696793735,-0.2588975728,0.1952017099,-0.1164151728,0.1070558354,0.326749444,-0.117432341,0.306620568,0.6193355918,-0.1105521396,0.1110311449,-0.215863362,0.0245816521,0.0014836154,0.2785247266,0.2359109074,-0.1938206851,0.1145048365,0.2641795576,0.243651405,-0.3050079346,0.1434532404,0.2776193917,0.0995111689,0.0506758131,-0.1596455127,-0.3252392113,-0.2702022195,0.2318239212,-0.1383464932,0.1085467264,-0.3714389801,0.0401465409,-0.2963201106,0.0447943099,-0.5638593435,-0.1213942021,0.1630701125,0.2012814134,0.0795095712,0.0613085926,-0.3085088432,0.2793070674,-0.1050261855,-0.0565539338,-0.5994650126,0.2591193616,-0.0674927309,-0.3134038746,-0.0747863129,0.3225338161,0.0828683749,-0.13501966,-0.3367395401,0.2787976265,-0.0855251029,0.0143846814,0.0664433986,-0.1704661995,0.2425206453,-0.4210937619,0.1616301388,0.0719261542,0.12785393,0.0151306158,0.1468042135,0.3288503289,0.0096400762,0.0002207728,-0.0022295045,-0.0426436365,0.4274256229,-0.0020315761,0.0864704177,-0.257260561,0.2339752018,-0.0238232464,0.1626571417,-0.0563809276,-0.1947318316,-0.0884142816,0.5688619018,-0.1365721375,0.1961637586,0.1271996349,-0.0223116353,0.0912192017,-0.0803106651,0.0161401406,0.2614569962,0.0181957334,0.0555701181,0.0068808333,0.1075893268,0.02909513,0.1412385255,-0.084366791,-0.0864135996,-0.0710775107,-0.0867036805,-0.0300667249,-0.3335590065,-0.1923588961,-0.1055369303,0.3316015005,-0.1660034209,-0.0535194837,-0.1744429022,-0.1915307194,0.037373513,-0.1773097664,-0.1640250832,-0.2362444401,-0.2318590879,0.0255354829,0.1254676878,0.322096318,-0.4353750944,-0.35697487,0.1276556998,-0.3397228718,0.0316216648,-0.2119199336,-0.2546372116,0.1059964895,0.3112894893,0.1780466586,0.0755409226,-0.389688462,0.1155692562,-0.2473884374,-0.0755902529,0.0593283139,-0.3163183928,0.1726506948,-0.0393578149,0.0839216486,-0.0562964082,-0.181186527,0.2916715443,-0.2333991379,-0.1716851741,0.206258446,0.0019390341,-0.19655931,0.0024222732,-0.1196481287,-0.3280853927,-0.3669332564,0.0012898719,-0.0777989402,0.1852790117,0.1442967504,-0.2602292299,-0.1182346195,0.1143380255,-0.1334733516,-0.1536588073,-0.4095715284,0.4874741733,-0.372284323,-0.2143343836,-0.0219009388,-0.1756924838,0.4781981111,0.1308036745,-0.22841084,-0.0028280204,-0.3269365728,0.2939243317,0.0070022531,0.0189219732,0.1398832947,0.0004282062,0.0058593466,0.0415178165,-0.0540438183,-0.1960995942,0.2766460776,0.1708433926,0.1009547412,0.383543551,-0.1219677925,0.1801291257,-0.0492355041,0.3242917955,0.5478320718,0.0955017433,0.0651276484,-0.0233803149,-0.4221803546,-0.0489595979,-0.0551072024,0.0486402884,0.0079956595,-0.0325141177,-0.105156444,-0.0710553825,-0.1858091354,-0.1542174369,-0.2681567669,0.1497886032,-0.2403441966,0.2449389249,-0.1164389551,0.0076551237,-0.3299103379,0.0237852167,-0.0713196546,0.1350256205,0.2146928608,0.0793461651,-0.1952005923,0.3681942523,-0.4347116053,0.0998158306,0.3249466121,0.703830719,0.0326968022,0.0539382137,0.0414398424,-0.0472928733,-0.0577151068,-0.1116312221,0.1558757871,0.1666961759,-0.2209492028,-0.2402218878,-0.0395986289,-0.1052687243,0.3003042936,0.2227607071,-0.0354341567,-0.3422704339,-0.1095726341,0.3165017962,0.1825249046,-0.1780320257,-0.0400575511,-0.4110560715,-0.5928244591,-0.4134850502,0.0432143509,0.2406070083,0.4273461401,0.3378159106,0.1650990695,-0.2048251033,-0.2078794092,0.2217248082,0.3031234145,0.0947903246,0.0685112104,0.0718200728,-0.2338235378,-0.3942333758,0.1363064498,0.5306170583,0.1089847833,-0.7092872262,-0.078453362,-0.2032940984,0.2511305511,0.059877906,-0.0541696139,0.1121880934,-0.1826997846,0.0938625634,0.043718785,0.1091841459,0.1448744535,0.2128843516,-0.6021432281,-0.4522859156,0.1971878111,0.0163849033,0.2952745855,0.3392183483,-0.1642782241,-0.3164373934,-0.0895475596,0.1628365666,0.5518426895,0.0499165542,0.1146935076,0.3466674387,0.0413424149,0.058239501,0.0459595174,0.1835665703,-0.2096627951,-0.2382662147,-0.0543798544,-0.025275711,0.1658082753,-0.0543145016,0.0872022435,0.2775225341,0.0734330863,0.0351325609,0.2459097058,0.198462829,-0.2746990919,-0.4539572895,-0.5819117427,0.1532522589,-0.3220199943,-0.0309987292,-0.196727097,-0.2522437572,0.0267227255,-0.2227253616,0.0441954397,0.224859193,-0.1557139754,0.2739735544,-0.1650171429,-0.5444099903,0.2062320113,0.0880706459,0.3015840352,-0.0706007481,0.1350811571,0.1731298119,0.1769678742,-0.0216342527,-0.0828447193,-0.0253757741,0.0173967034,-0.0310315695,-0.2499279827,0.0698867887,-0.0188626833,-0.3566619158,-0.2213180214,0.2003680915,-0.0024183826,-0.0873170272,-0.2053951919,-0.1993931383,0.0175070502,-0.1896483749,0.1594203115,0.0621682294,-0.2342941463,0.0830627158,0.0320613347,-0.3447245657,0.0183398556,0.0949900448,0.1793593764,-0.4032029808,0.5781168342,0.1079861,-0.0876559317,-0.123005636,0.0509770848,0.188012287,-0.5171360373,0.0444077961,-0.0069698733,0.0980357304,-0.1275515854,0.0985043943,0.0108745405,-0.096129775,0.1677367389,-0.4163189828,-0.5901713967,0.1163460761,-0.1201160848,0.2132864892,0.0890787989,0.1483061165,-0.1083181128,-0.0477421358,-0.304250598,0.1132213548,-0.3167765141,0.2198823541,0.2084016055,-0.0789639801,0.0466105267,0.0550280698,0.2011434734,0.3281204402,-0.0766396374,-0.1902381778,0.0340617821,0.08731208,-0.0867043063,-0.4576300979,-0.01823573,0.2222194225,-0.0597430989,-0.1353600472,0.1167752296,0.1854208857,-0.0626491234,0.0720704943,0.1372987181,0.1072310358,-0.1261375993,0.298120141,-0.209562853,0.4348088503,0.0386724807,0.2651923597,0.0883063301,0.0721179917,-0.2311838567,-0.1048307195,0.1180974394,0.1814716011,0.14326258,-0.5272974372,0.1788813025,0.2601992488,-0.0461180806,0.3760097027,-0.1835526228,-0.0165387448,0.120889388,0.2417025566,-0.1411454827,-0.0686583668,0.3440247476,-0.2144441307,0.0001063908,0.4460248053,0.1517319828,0.1151359379,0.2358342707,0.1073320955,0.2825607061,-0.0068990048,0.2959387302,0.475163877,-0.1015865877,-0.0476528928,0.2322836965,0.1343567818,0.0184182413,0.6424046159,-0.442773819,0.5353500843,-0.0505037382,0.2583877444,-0.3224719167,-0.3795407712,-0.2869464159,-0.1631638855,-0.0148015292,0.2366979718,0.1603486538,-0.3238866031,-0.1891293228,0.0048719519,-0.0439247601,0.1654580981,-0.2276591212,0.1310626715,-0.0501659214,-0.0186570399,-0.1377204359,-0.0187568478,0.0113090724,0.1884762943,-0.0382181816,0.2333977371,-0.1012568995,-0.0860914066,0.159076497,0.2796638608,-0.0650407374,-0.2643450499,0.4634735584,0.5010718703,-0.0114834644,-0.0209639426,0.1890849322,0.3574219346,0.4347184002,-0.2235045582,0.1034589782,0.000924604,0.0185824297,0.0500134677,0.1436081976,0.0140555101,0.6835596561,0.1786809713,0.2524062097,-0.1252028644,-0.0167536419,0.0970192775,-0.2486756146,0.0330268592,0.2661145031,0.1370356828,-0.0323732123,-0.2701379657,-0.0949243903,-0.2792062163,0.006541695,0.384693414,0.0213529747,0.106054008,-0.1315313131,0.0763784647,0.0020454954,0.4745551646,0.3403455615,-0.0144478558,-0.3034825623,-0.2072483152,-0.7285958529,0.194361046,-0.1984153986,-0.3285027146,0.2497081608,-0.0827106461,-0.0733957067,0.3360482752,0.2181287706,0.398761332,0.0736603737,0.1313437372,-0.3173839748,-0.0350323953,0.4439384341,0.0782283843,-0.0843572021,-0.3402358592,0.1098290831,0.0742676035,0.1089789867,-0.1935682148,-0.0207891837,-0.0690447688,-0.13243635,0.3674142957,0.2805423439,0.393017292,-0.1928328872,-0.1226181611,-0.17145513,-0.3176394999,-0.081048727,0.3091652095,0.0883627608,0.5725002289,0.0104075894,-0.087408416,-0.6893698573,0.3223952949,-0.1301128119,0.0926097482,0.0033725582,0.1471588314,-0.074550949,0.0625717491,0.006926755,0.1489863098,0.0457483046,0.4057685733,-0.1709182858,-0.5643859506,0.3966764808,-0.0855885968,-0.685715735,-0.110391736,0.0107939122,-0.0288262423,-0.0517250597,-0.3539994061,0.3004591465,0.2369799316,-0.0079769436,-0.1723502874,0.2072672099,0.0321722776,0.0402625687,-0.0043208124,0.4464668632,-0.1106186807,-0.0629551485,0.1378686875,-0.2023232877]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/468","title":"UnicodeDecodeError while loading PAN-X task of XTREME dataset","comments":"I'm happy to tackle the broader problem - will open a PR when it's ready!","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!","comment_length":15,"text":"UnicodeDecodeError while loading PAN-X task of XTREME dataset\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!\nI'm happy to tackle the broader problem - will open a PR when it's ready!","embeddings":[-0.3326011896,-0.0930106342,-0.1336479485,0.2554301322,0.3521757722,-0.0277493317,0.2249831408,0.3496463001,-0.0286989305,0.038228903,-0.0904103443,0.1502787173,-0.0761749223,-0.1970372051,0.1891056299,-0.2467030734,-0.0738269314,0.1824576855,0.1078069955,0.0069923094,-0.3558878899,0.1144126877,-0.1869958341,0.2637093067,-0.3771478236,0.0810821429,0.2978654206,0.3204651177,-0.193280682,-0.6116119623,0.4130367041,-0.0536639951,0.4325025082,0.3163448274,-0.0001086041,0.0825855061,0.3929110467,-0.1319698691,-0.3346782327,-0.1939976513,-0.5915244818,-0.1812518835,-0.0098235868,-0.2595872581,0.1902403831,0.0562078469,-0.0642362684,-0.3830473125,0.1811341196,0.2071898282,0.24163872,0.3705638647,0.0213383194,0.0912284628,-0.045574829,0.1074261367,-0.1542558074,0.249418959,0.4346242249,0.0627749935,-0.019149052,0.2926935852,-0.0486163385,0.248120591,-0.1023555249,0.1306457818,0.1399910003,-0.4660047889,0.1134835854,0.0036376556,0.2718741894,-0.3664452136,-0.2865583897,-0.2252654582,-0.1094700545,-0.5443437099,0.2644951046,0.0905644149,-0.159659639,-0.0211321656,0.113159202,-0.0423871912,-0.1735193729,0.0391214527,0.0543593504,0.3878184855,-0.0600738861,-0.0513223708,-0.0941757336,-0.1786022931,0.4082849622,-0.1458789855,-0.1436942965,0.324629128,-0.2863637805,0.072422564,-0.0550461523,-0.3371772468,0.0668165982,-0.0463103987,-0.078823179,0.2744791806,-0.0495554432,0.053195674,0.3093065917,0.1556752175,0.2095072418,0.2245565802,0.1834983528,0.1982241273,0.0829440206,-0.0012048965,-0.1758536994,-0.1771315336,0.3654833138,0.1934714913,0.5531736016,-0.3191607296,-0.1221157163,0.0084841102,0.1728852391,-0.1787237525,0.0444031768,0.0144082429,0.2223311216,0.5070413947,-0.2224069536,0.2185738385,-0.4622749686,-0.1283478737,-0.1358893216,0.1796104163,-0.1460784674,-0.0783130527,0.0963894874,0.1252889782,0.163143605,-0.1422412395,0.3579080105,-0.0693807527,0.0396495759,-0.2039281875,-0.2201946378,0.0137812095,0.1307834387,0.1039626971,0.0551503152,-0.1753356308,-0.0369680114,0.2171099186,-0.1871833205,-0.2070101798,-0.0715540051,0.2629823387,0.0267733559,0.025153663,-0.4134668708,-0.0887362659,0.3430744112,0.062088266,0.1139909923,-0.2957776785,0.0404423103,-0.0858764797,0.2037117779,0.3430534005,-0.1321642697,0.0464633144,0.1779045761,-0.1550109833,0.2246282697,0.296744436,-0.1922280043,0.0937530622,-0.0989214331,0.5539204478,0.5879890919,-0.4102394879,-0.4520799518,0.1696793735,-0.2588975728,0.1952017099,-0.1164151728,0.1070558354,0.326749444,-0.117432341,0.306620568,0.6193355918,-0.1105521396,0.1110311449,-0.215863362,0.0245816521,0.0014836154,0.2785247266,0.2359109074,-0.1938206851,0.1145048365,0.2641795576,0.243651405,-0.3050079346,0.1434532404,0.2776193917,0.0995111689,0.0506758131,-0.1596455127,-0.3252392113,-0.2702022195,0.2318239212,-0.1383464932,0.1085467264,-0.3714389801,0.0401465409,-0.2963201106,0.0447943099,-0.5638593435,-0.1213942021,0.1630701125,0.2012814134,0.0795095712,0.0613085926,-0.3085088432,0.2793070674,-0.1050261855,-0.0565539338,-0.5994650126,0.2591193616,-0.0674927309,-0.3134038746,-0.0747863129,0.3225338161,0.0828683749,-0.13501966,-0.3367395401,0.2787976265,-0.0855251029,0.0143846814,0.0664433986,-0.1704661995,0.2425206453,-0.4210937619,0.1616301388,0.0719261542,0.12785393,0.0151306158,0.1468042135,0.3288503289,0.0096400762,0.0002207728,-0.0022295045,-0.0426436365,0.4274256229,-0.0020315761,0.0864704177,-0.257260561,0.2339752018,-0.0238232464,0.1626571417,-0.0563809276,-0.1947318316,-0.0884142816,0.5688619018,-0.1365721375,0.1961637586,0.1271996349,-0.0223116353,0.0912192017,-0.0803106651,0.0161401406,0.2614569962,0.0181957334,0.0555701181,0.0068808333,0.1075893268,0.02909513,0.1412385255,-0.084366791,-0.0864135996,-0.0710775107,-0.0867036805,-0.0300667249,-0.3335590065,-0.1923588961,-0.1055369303,0.3316015005,-0.1660034209,-0.0535194837,-0.1744429022,-0.1915307194,0.037373513,-0.1773097664,-0.1640250832,-0.2362444401,-0.2318590879,0.0255354829,0.1254676878,0.322096318,-0.4353750944,-0.35697487,0.1276556998,-0.3397228718,0.0316216648,-0.2119199336,-0.2546372116,0.1059964895,0.3112894893,0.1780466586,0.0755409226,-0.389688462,0.1155692562,-0.2473884374,-0.0755902529,0.0593283139,-0.3163183928,0.1726506948,-0.0393578149,0.0839216486,-0.0562964082,-0.181186527,0.2916715443,-0.2333991379,-0.1716851741,0.206258446,0.0019390341,-0.19655931,0.0024222732,-0.1196481287,-0.3280853927,-0.3669332564,0.0012898719,-0.0777989402,0.1852790117,0.1442967504,-0.2602292299,-0.1182346195,0.1143380255,-0.1334733516,-0.1536588073,-0.4095715284,0.4874741733,-0.372284323,-0.2143343836,-0.0219009388,-0.1756924838,0.4781981111,0.1308036745,-0.22841084,-0.0028280204,-0.3269365728,0.2939243317,0.0070022531,0.0189219732,0.1398832947,0.0004282062,0.0058593466,0.0415178165,-0.0540438183,-0.1960995942,0.2766460776,0.1708433926,0.1009547412,0.383543551,-0.1219677925,0.1801291257,-0.0492355041,0.3242917955,0.5478320718,0.0955017433,0.0651276484,-0.0233803149,-0.4221803546,-0.0489595979,-0.0551072024,0.0486402884,0.0079956595,-0.0325141177,-0.105156444,-0.0710553825,-0.1858091354,-0.1542174369,-0.2681567669,0.1497886032,-0.2403441966,0.2449389249,-0.1164389551,0.0076551237,-0.3299103379,0.0237852167,-0.0713196546,0.1350256205,0.2146928608,0.0793461651,-0.1952005923,0.3681942523,-0.4347116053,0.0998158306,0.3249466121,0.703830719,0.0326968022,0.0539382137,0.0414398424,-0.0472928733,-0.0577151068,-0.1116312221,0.1558757871,0.1666961759,-0.2209492028,-0.2402218878,-0.0395986289,-0.1052687243,0.3003042936,0.2227607071,-0.0354341567,-0.3422704339,-0.1095726341,0.3165017962,0.1825249046,-0.1780320257,-0.0400575511,-0.4110560715,-0.5928244591,-0.4134850502,0.0432143509,0.2406070083,0.4273461401,0.3378159106,0.1650990695,-0.2048251033,-0.2078794092,0.2217248082,0.3031234145,0.0947903246,0.0685112104,0.0718200728,-0.2338235378,-0.3942333758,0.1363064498,0.5306170583,0.1089847833,-0.7092872262,-0.078453362,-0.2032940984,0.2511305511,0.059877906,-0.0541696139,0.1121880934,-0.1826997846,0.0938625634,0.043718785,0.1091841459,0.1448744535,0.2128843516,-0.6021432281,-0.4522859156,0.1971878111,0.0163849033,0.2952745855,0.3392183483,-0.1642782241,-0.3164373934,-0.0895475596,0.1628365666,0.5518426895,0.0499165542,0.1146935076,0.3466674387,0.0413424149,0.058239501,0.0459595174,0.1835665703,-0.2096627951,-0.2382662147,-0.0543798544,-0.025275711,0.1658082753,-0.0543145016,0.0872022435,0.2775225341,0.0734330863,0.0351325609,0.2459097058,0.198462829,-0.2746990919,-0.4539572895,-0.5819117427,0.1532522589,-0.3220199943,-0.0309987292,-0.196727097,-0.2522437572,0.0267227255,-0.2227253616,0.0441954397,0.224859193,-0.1557139754,0.2739735544,-0.1650171429,-0.5444099903,0.2062320113,0.0880706459,0.3015840352,-0.0706007481,0.1350811571,0.1731298119,0.1769678742,-0.0216342527,-0.0828447193,-0.0253757741,0.0173967034,-0.0310315695,-0.2499279827,0.0698867887,-0.0188626833,-0.3566619158,-0.2213180214,0.2003680915,-0.0024183826,-0.0873170272,-0.2053951919,-0.1993931383,0.0175070502,-0.1896483749,0.1594203115,0.0621682294,-0.2342941463,0.0830627158,0.0320613347,-0.3447245657,0.0183398556,0.0949900448,0.1793593764,-0.4032029808,0.5781168342,0.1079861,-0.0876559317,-0.123005636,0.0509770848,0.188012287,-0.5171360373,0.0444077961,-0.0069698733,0.0980357304,-0.1275515854,0.0985043943,0.0108745405,-0.096129775,0.1677367389,-0.4163189828,-0.5901713967,0.1163460761,-0.1201160848,0.2132864892,0.0890787989,0.1483061165,-0.1083181128,-0.0477421358,-0.304250598,0.1132213548,-0.3167765141,0.2198823541,0.2084016055,-0.0789639801,0.0466105267,0.0550280698,0.2011434734,0.3281204402,-0.0766396374,-0.1902381778,0.0340617821,0.08731208,-0.0867043063,-0.4576300979,-0.01823573,0.2222194225,-0.0597430989,-0.1353600472,0.1167752296,0.1854208857,-0.0626491234,0.0720704943,0.1372987181,0.1072310358,-0.1261375993,0.298120141,-0.209562853,0.4348088503,0.0386724807,0.2651923597,0.0883063301,0.0721179917,-0.2311838567,-0.1048307195,0.1180974394,0.1814716011,0.14326258,-0.5272974372,0.1788813025,0.2601992488,-0.0461180806,0.3760097027,-0.1835526228,-0.0165387448,0.120889388,0.2417025566,-0.1411454827,-0.0686583668,0.3440247476,-0.2144441307,0.0001063908,0.4460248053,0.1517319828,0.1151359379,0.2358342707,0.1073320955,0.2825607061,-0.0068990048,0.2959387302,0.475163877,-0.1015865877,-0.0476528928,0.2322836965,0.1343567818,0.0184182413,0.6424046159,-0.442773819,0.5353500843,-0.0505037382,0.2583877444,-0.3224719167,-0.3795407712,-0.2869464159,-0.1631638855,-0.0148015292,0.2366979718,0.1603486538,-0.3238866031,-0.1891293228,0.0048719519,-0.0439247601,0.1654580981,-0.2276591212,0.1310626715,-0.0501659214,-0.0186570399,-0.1377204359,-0.0187568478,0.0113090724,0.1884762943,-0.0382181816,0.2333977371,-0.1012568995,-0.0860914066,0.159076497,0.2796638608,-0.0650407374,-0.2643450499,0.4634735584,0.5010718703,-0.0114834644,-0.0209639426,0.1890849322,0.3574219346,0.4347184002,-0.2235045582,0.1034589782,0.000924604,0.0185824297,0.0500134677,0.1436081976,0.0140555101,0.6835596561,0.1786809713,0.2524062097,-0.1252028644,-0.0167536419,0.0970192775,-0.2486756146,0.0330268592,0.2661145031,0.1370356828,-0.0323732123,-0.2701379657,-0.0949243903,-0.2792062163,0.006541695,0.384693414,0.0213529747,0.106054008,-0.1315313131,0.0763784647,0.0020454954,0.4745551646,0.3403455615,-0.0144478558,-0.3034825623,-0.2072483152,-0.7285958529,0.194361046,-0.1984153986,-0.3285027146,0.2497081608,-0.0827106461,-0.0733957067,0.3360482752,0.2181287706,0.398761332,0.0736603737,0.1313437372,-0.3173839748,-0.0350323953,0.4439384341,0.0782283843,-0.0843572021,-0.3402358592,0.1098290831,0.0742676035,0.1089789867,-0.1935682148,-0.0207891837,-0.0690447688,-0.13243635,0.3674142957,0.2805423439,0.393017292,-0.1928328872,-0.1226181611,-0.17145513,-0.3176394999,-0.081048727,0.3091652095,0.0883627608,0.5725002289,0.0104075894,-0.087408416,-0.6893698573,0.3223952949,-0.1301128119,0.0926097482,0.0033725582,0.1471588314,-0.074550949,0.0625717491,0.006926755,0.1489863098,0.0457483046,0.4057685733,-0.1709182858,-0.5643859506,0.3966764808,-0.0855885968,-0.685715735,-0.110391736,0.0107939122,-0.0288262423,-0.0517250597,-0.3539994061,0.3004591465,0.2369799316,-0.0079769436,-0.1723502874,0.2072672099,0.0321722776,0.0402625687,-0.0043208124,0.4464668632,-0.1106186807,-0.0629551485,0.1378686875,-0.2023232877]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/468","title":"UnicodeDecodeError while loading PAN-X task of XTREME dataset","comments":"That would be awesome!","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!","comment_length":4,"text":"UnicodeDecodeError while loading PAN-X task of XTREME dataset\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!\nThat would be awesome!","embeddings":[-0.3326011896,-0.0930106342,-0.1336479485,0.2554301322,0.3521757722,-0.0277493317,0.2249831408,0.3496463001,-0.0286989305,0.038228903,-0.0904103443,0.1502787173,-0.0761749223,-0.1970372051,0.1891056299,-0.2467030734,-0.0738269314,0.1824576855,0.1078069955,0.0069923094,-0.3558878899,0.1144126877,-0.1869958341,0.2637093067,-0.3771478236,0.0810821429,0.2978654206,0.3204651177,-0.193280682,-0.6116119623,0.4130367041,-0.0536639951,0.4325025082,0.3163448274,-0.0001086041,0.0825855061,0.3929110467,-0.1319698691,-0.3346782327,-0.1939976513,-0.5915244818,-0.1812518835,-0.0098235868,-0.2595872581,0.1902403831,0.0562078469,-0.0642362684,-0.3830473125,0.1811341196,0.2071898282,0.24163872,0.3705638647,0.0213383194,0.0912284628,-0.045574829,0.1074261367,-0.1542558074,0.249418959,0.4346242249,0.0627749935,-0.019149052,0.2926935852,-0.0486163385,0.248120591,-0.1023555249,0.1306457818,0.1399910003,-0.4660047889,0.1134835854,0.0036376556,0.2718741894,-0.3664452136,-0.2865583897,-0.2252654582,-0.1094700545,-0.5443437099,0.2644951046,0.0905644149,-0.159659639,-0.0211321656,0.113159202,-0.0423871912,-0.1735193729,0.0391214527,0.0543593504,0.3878184855,-0.0600738861,-0.0513223708,-0.0941757336,-0.1786022931,0.4082849622,-0.1458789855,-0.1436942965,0.324629128,-0.2863637805,0.072422564,-0.0550461523,-0.3371772468,0.0668165982,-0.0463103987,-0.078823179,0.2744791806,-0.0495554432,0.053195674,0.3093065917,0.1556752175,0.2095072418,0.2245565802,0.1834983528,0.1982241273,0.0829440206,-0.0012048965,-0.1758536994,-0.1771315336,0.3654833138,0.1934714913,0.5531736016,-0.3191607296,-0.1221157163,0.0084841102,0.1728852391,-0.1787237525,0.0444031768,0.0144082429,0.2223311216,0.5070413947,-0.2224069536,0.2185738385,-0.4622749686,-0.1283478737,-0.1358893216,0.1796104163,-0.1460784674,-0.0783130527,0.0963894874,0.1252889782,0.163143605,-0.1422412395,0.3579080105,-0.0693807527,0.0396495759,-0.2039281875,-0.2201946378,0.0137812095,0.1307834387,0.1039626971,0.0551503152,-0.1753356308,-0.0369680114,0.2171099186,-0.1871833205,-0.2070101798,-0.0715540051,0.2629823387,0.0267733559,0.025153663,-0.4134668708,-0.0887362659,0.3430744112,0.062088266,0.1139909923,-0.2957776785,0.0404423103,-0.0858764797,0.2037117779,0.3430534005,-0.1321642697,0.0464633144,0.1779045761,-0.1550109833,0.2246282697,0.296744436,-0.1922280043,0.0937530622,-0.0989214331,0.5539204478,0.5879890919,-0.4102394879,-0.4520799518,0.1696793735,-0.2588975728,0.1952017099,-0.1164151728,0.1070558354,0.326749444,-0.117432341,0.306620568,0.6193355918,-0.1105521396,0.1110311449,-0.215863362,0.0245816521,0.0014836154,0.2785247266,0.2359109074,-0.1938206851,0.1145048365,0.2641795576,0.243651405,-0.3050079346,0.1434532404,0.2776193917,0.0995111689,0.0506758131,-0.1596455127,-0.3252392113,-0.2702022195,0.2318239212,-0.1383464932,0.1085467264,-0.3714389801,0.0401465409,-0.2963201106,0.0447943099,-0.5638593435,-0.1213942021,0.1630701125,0.2012814134,0.0795095712,0.0613085926,-0.3085088432,0.2793070674,-0.1050261855,-0.0565539338,-0.5994650126,0.2591193616,-0.0674927309,-0.3134038746,-0.0747863129,0.3225338161,0.0828683749,-0.13501966,-0.3367395401,0.2787976265,-0.0855251029,0.0143846814,0.0664433986,-0.1704661995,0.2425206453,-0.4210937619,0.1616301388,0.0719261542,0.12785393,0.0151306158,0.1468042135,0.3288503289,0.0096400762,0.0002207728,-0.0022295045,-0.0426436365,0.4274256229,-0.0020315761,0.0864704177,-0.257260561,0.2339752018,-0.0238232464,0.1626571417,-0.0563809276,-0.1947318316,-0.0884142816,0.5688619018,-0.1365721375,0.1961637586,0.1271996349,-0.0223116353,0.0912192017,-0.0803106651,0.0161401406,0.2614569962,0.0181957334,0.0555701181,0.0068808333,0.1075893268,0.02909513,0.1412385255,-0.084366791,-0.0864135996,-0.0710775107,-0.0867036805,-0.0300667249,-0.3335590065,-0.1923588961,-0.1055369303,0.3316015005,-0.1660034209,-0.0535194837,-0.1744429022,-0.1915307194,0.037373513,-0.1773097664,-0.1640250832,-0.2362444401,-0.2318590879,0.0255354829,0.1254676878,0.322096318,-0.4353750944,-0.35697487,0.1276556998,-0.3397228718,0.0316216648,-0.2119199336,-0.2546372116,0.1059964895,0.3112894893,0.1780466586,0.0755409226,-0.389688462,0.1155692562,-0.2473884374,-0.0755902529,0.0593283139,-0.3163183928,0.1726506948,-0.0393578149,0.0839216486,-0.0562964082,-0.181186527,0.2916715443,-0.2333991379,-0.1716851741,0.206258446,0.0019390341,-0.19655931,0.0024222732,-0.1196481287,-0.3280853927,-0.3669332564,0.0012898719,-0.0777989402,0.1852790117,0.1442967504,-0.2602292299,-0.1182346195,0.1143380255,-0.1334733516,-0.1536588073,-0.4095715284,0.4874741733,-0.372284323,-0.2143343836,-0.0219009388,-0.1756924838,0.4781981111,0.1308036745,-0.22841084,-0.0028280204,-0.3269365728,0.2939243317,0.0070022531,0.0189219732,0.1398832947,0.0004282062,0.0058593466,0.0415178165,-0.0540438183,-0.1960995942,0.2766460776,0.1708433926,0.1009547412,0.383543551,-0.1219677925,0.1801291257,-0.0492355041,0.3242917955,0.5478320718,0.0955017433,0.0651276484,-0.0233803149,-0.4221803546,-0.0489595979,-0.0551072024,0.0486402884,0.0079956595,-0.0325141177,-0.105156444,-0.0710553825,-0.1858091354,-0.1542174369,-0.2681567669,0.1497886032,-0.2403441966,0.2449389249,-0.1164389551,0.0076551237,-0.3299103379,0.0237852167,-0.0713196546,0.1350256205,0.2146928608,0.0793461651,-0.1952005923,0.3681942523,-0.4347116053,0.0998158306,0.3249466121,0.703830719,0.0326968022,0.0539382137,0.0414398424,-0.0472928733,-0.0577151068,-0.1116312221,0.1558757871,0.1666961759,-0.2209492028,-0.2402218878,-0.0395986289,-0.1052687243,0.3003042936,0.2227607071,-0.0354341567,-0.3422704339,-0.1095726341,0.3165017962,0.1825249046,-0.1780320257,-0.0400575511,-0.4110560715,-0.5928244591,-0.4134850502,0.0432143509,0.2406070083,0.4273461401,0.3378159106,0.1650990695,-0.2048251033,-0.2078794092,0.2217248082,0.3031234145,0.0947903246,0.0685112104,0.0718200728,-0.2338235378,-0.3942333758,0.1363064498,0.5306170583,0.1089847833,-0.7092872262,-0.078453362,-0.2032940984,0.2511305511,0.059877906,-0.0541696139,0.1121880934,-0.1826997846,0.0938625634,0.043718785,0.1091841459,0.1448744535,0.2128843516,-0.6021432281,-0.4522859156,0.1971878111,0.0163849033,0.2952745855,0.3392183483,-0.1642782241,-0.3164373934,-0.0895475596,0.1628365666,0.5518426895,0.0499165542,0.1146935076,0.3466674387,0.0413424149,0.058239501,0.0459595174,0.1835665703,-0.2096627951,-0.2382662147,-0.0543798544,-0.025275711,0.1658082753,-0.0543145016,0.0872022435,0.2775225341,0.0734330863,0.0351325609,0.2459097058,0.198462829,-0.2746990919,-0.4539572895,-0.5819117427,0.1532522589,-0.3220199943,-0.0309987292,-0.196727097,-0.2522437572,0.0267227255,-0.2227253616,0.0441954397,0.224859193,-0.1557139754,0.2739735544,-0.1650171429,-0.5444099903,0.2062320113,0.0880706459,0.3015840352,-0.0706007481,0.1350811571,0.1731298119,0.1769678742,-0.0216342527,-0.0828447193,-0.0253757741,0.0173967034,-0.0310315695,-0.2499279827,0.0698867887,-0.0188626833,-0.3566619158,-0.2213180214,0.2003680915,-0.0024183826,-0.0873170272,-0.2053951919,-0.1993931383,0.0175070502,-0.1896483749,0.1594203115,0.0621682294,-0.2342941463,0.0830627158,0.0320613347,-0.3447245657,0.0183398556,0.0949900448,0.1793593764,-0.4032029808,0.5781168342,0.1079861,-0.0876559317,-0.123005636,0.0509770848,0.188012287,-0.5171360373,0.0444077961,-0.0069698733,0.0980357304,-0.1275515854,0.0985043943,0.0108745405,-0.096129775,0.1677367389,-0.4163189828,-0.5901713967,0.1163460761,-0.1201160848,0.2132864892,0.0890787989,0.1483061165,-0.1083181128,-0.0477421358,-0.304250598,0.1132213548,-0.3167765141,0.2198823541,0.2084016055,-0.0789639801,0.0466105267,0.0550280698,0.2011434734,0.3281204402,-0.0766396374,-0.1902381778,0.0340617821,0.08731208,-0.0867043063,-0.4576300979,-0.01823573,0.2222194225,-0.0597430989,-0.1353600472,0.1167752296,0.1854208857,-0.0626491234,0.0720704943,0.1372987181,0.1072310358,-0.1261375993,0.298120141,-0.209562853,0.4348088503,0.0386724807,0.2651923597,0.0883063301,0.0721179917,-0.2311838567,-0.1048307195,0.1180974394,0.1814716011,0.14326258,-0.5272974372,0.1788813025,0.2601992488,-0.0461180806,0.3760097027,-0.1835526228,-0.0165387448,0.120889388,0.2417025566,-0.1411454827,-0.0686583668,0.3440247476,-0.2144441307,0.0001063908,0.4460248053,0.1517319828,0.1151359379,0.2358342707,0.1073320955,0.2825607061,-0.0068990048,0.2959387302,0.475163877,-0.1015865877,-0.0476528928,0.2322836965,0.1343567818,0.0184182413,0.6424046159,-0.442773819,0.5353500843,-0.0505037382,0.2583877444,-0.3224719167,-0.3795407712,-0.2869464159,-0.1631638855,-0.0148015292,0.2366979718,0.1603486538,-0.3238866031,-0.1891293228,0.0048719519,-0.0439247601,0.1654580981,-0.2276591212,0.1310626715,-0.0501659214,-0.0186570399,-0.1377204359,-0.0187568478,0.0113090724,0.1884762943,-0.0382181816,0.2333977371,-0.1012568995,-0.0860914066,0.159076497,0.2796638608,-0.0650407374,-0.2643450499,0.4634735584,0.5010718703,-0.0114834644,-0.0209639426,0.1890849322,0.3574219346,0.4347184002,-0.2235045582,0.1034589782,0.000924604,0.0185824297,0.0500134677,0.1436081976,0.0140555101,0.6835596561,0.1786809713,0.2524062097,-0.1252028644,-0.0167536419,0.0970192775,-0.2486756146,0.0330268592,0.2661145031,0.1370356828,-0.0323732123,-0.2701379657,-0.0949243903,-0.2792062163,0.006541695,0.384693414,0.0213529747,0.106054008,-0.1315313131,0.0763784647,0.0020454954,0.4745551646,0.3403455615,-0.0144478558,-0.3034825623,-0.2072483152,-0.7285958529,0.194361046,-0.1984153986,-0.3285027146,0.2497081608,-0.0827106461,-0.0733957067,0.3360482752,0.2181287706,0.398761332,0.0736603737,0.1313437372,-0.3173839748,-0.0350323953,0.4439384341,0.0782283843,-0.0843572021,-0.3402358592,0.1098290831,0.0742676035,0.1089789867,-0.1935682148,-0.0207891837,-0.0690447688,-0.13243635,0.3674142957,0.2805423439,0.393017292,-0.1928328872,-0.1226181611,-0.17145513,-0.3176394999,-0.081048727,0.3091652095,0.0883627608,0.5725002289,0.0104075894,-0.087408416,-0.6893698573,0.3223952949,-0.1301128119,0.0926097482,0.0033725582,0.1471588314,-0.074550949,0.0625717491,0.006926755,0.1489863098,0.0457483046,0.4057685733,-0.1709182858,-0.5643859506,0.3966764808,-0.0855885968,-0.685715735,-0.110391736,0.0107939122,-0.0288262423,-0.0517250597,-0.3539994061,0.3004591465,0.2369799316,-0.0079769436,-0.1723502874,0.2072672099,0.0321722776,0.0402625687,-0.0043208124,0.4464668632,-0.1106186807,-0.0629551485,0.1378686875,-0.2023232877]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/468","title":"UnicodeDecodeError while loading PAN-X task of XTREME dataset","comments":"I've created a simple function that seems to do the trick:\r\n\r\n```python\r\ndef apply_encoding_on_file_open(filepath: str):\r\n \"\"\"Apply UTF-8 encoding for all instances where a non-binary file is opened.\"\"\"\r\n \r\n with open(filepath, 'r', encoding='utf-8') as input_file:\r\n regexp = re.compile(r\"\"\"\r\n (?!.*\\b(?:encoding|rb|wb|wb+|ab|ab+)\\b)\r\n (open)\r\n \\((.*)\\)\r\n \"\"\")\r\n input_text = input_file.read()\r\n match = regexp.search(input_text)\r\n \r\n if match:\r\n print('Found match!', match.group())\r\n # append utf-8 encoding to matching groups in-place\r\n output = regexp.sub(lambda m: m.group()[:-1]+', encoding=\"utf-8\")', input_text)\r\n with open(filepath, 'w', encoding='utf-8') as output_file:\r\n output_file.write(output)\r\n else:\r\n print(\"No match found!\")\r\n```\r\n\r\nThe regexp does a negative lookahead to avoid matching on cases where the encoding is already specified or when binary files are involved.\r\n\r\nFrom an implementation perspective:\r\n\r\n* Would it make sense to include this function in `nlp-cli` so that we can run something like\r\n```\r\nnlp-cli fix_encoding path\/to\/folder\r\n```\r\nand the command recursively fixes all files in the target?\r\n* What is the desired behaviour in the CI test? Here we could either have a simple script that we run as a `job` in the CI and raises an error if a missing encoding is detected. Alternatively we could incorporate this behaviour into the CLI and run that in the CI.\r\n\r\nPlease let me know what you prefer among the alternatives.\r\n","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!","comment_length":200,"text":"UnicodeDecodeError while loading PAN-X task of XTREME dataset\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!\nI've created a simple function that seems to do the trick:\r\n\r\n```python\r\ndef apply_encoding_on_file_open(filepath: str):\r\n \"\"\"Apply UTF-8 encoding for all instances where a non-binary file is opened.\"\"\"\r\n \r\n with open(filepath, 'r', encoding='utf-8') as input_file:\r\n regexp = re.compile(r\"\"\"\r\n (?!.*\\b(?:encoding|rb|wb|wb+|ab|ab+)\\b)\r\n (open)\r\n \\((.*)\\)\r\n \"\"\")\r\n input_text = input_file.read()\r\n match = regexp.search(input_text)\r\n \r\n if match:\r\n print('Found match!', match.group())\r\n # append utf-8 encoding to matching groups in-place\r\n output = regexp.sub(lambda m: m.group()[:-1]+', encoding=\"utf-8\")', input_text)\r\n with open(filepath, 'w', encoding='utf-8') as output_file:\r\n output_file.write(output)\r\n else:\r\n print(\"No match found!\")\r\n```\r\n\r\nThe regexp does a negative lookahead to avoid matching on cases where the encoding is already specified or when binary files are involved.\r\n\r\nFrom an implementation perspective:\r\n\r\n* Would it make sense to include this function in `nlp-cli` so that we can run something like\r\n```\r\nnlp-cli fix_encoding path\/to\/folder\r\n```\r\nand the command recursively fixes all files in the target?\r\n* What is the desired behaviour in the CI test? Here we could either have a simple script that we run as a `job` in the CI and raises an error if a missing encoding is detected. Alternatively we could incorporate this behaviour into the CLI and run that in the CI.\r\n\r\nPlease let me know what you prefer among the alternatives.\r\n","embeddings":[-0.3326011896,-0.0930106342,-0.1336479485,0.2554301322,0.3521757722,-0.0277493317,0.2249831408,0.3496463001,-0.0286989305,0.038228903,-0.0904103443,0.1502787173,-0.0761749223,-0.1970372051,0.1891056299,-0.2467030734,-0.0738269314,0.1824576855,0.1078069955,0.0069923094,-0.3558878899,0.1144126877,-0.1869958341,0.2637093067,-0.3771478236,0.0810821429,0.2978654206,0.3204651177,-0.193280682,-0.6116119623,0.4130367041,-0.0536639951,0.4325025082,0.3163448274,-0.0001086041,0.0825855061,0.3929110467,-0.1319698691,-0.3346782327,-0.1939976513,-0.5915244818,-0.1812518835,-0.0098235868,-0.2595872581,0.1902403831,0.0562078469,-0.0642362684,-0.3830473125,0.1811341196,0.2071898282,0.24163872,0.3705638647,0.0213383194,0.0912284628,-0.045574829,0.1074261367,-0.1542558074,0.249418959,0.4346242249,0.0627749935,-0.019149052,0.2926935852,-0.0486163385,0.248120591,-0.1023555249,0.1306457818,0.1399910003,-0.4660047889,0.1134835854,0.0036376556,0.2718741894,-0.3664452136,-0.2865583897,-0.2252654582,-0.1094700545,-0.5443437099,0.2644951046,0.0905644149,-0.159659639,-0.0211321656,0.113159202,-0.0423871912,-0.1735193729,0.0391214527,0.0543593504,0.3878184855,-0.0600738861,-0.0513223708,-0.0941757336,-0.1786022931,0.4082849622,-0.1458789855,-0.1436942965,0.324629128,-0.2863637805,0.072422564,-0.0550461523,-0.3371772468,0.0668165982,-0.0463103987,-0.078823179,0.2744791806,-0.0495554432,0.053195674,0.3093065917,0.1556752175,0.2095072418,0.2245565802,0.1834983528,0.1982241273,0.0829440206,-0.0012048965,-0.1758536994,-0.1771315336,0.3654833138,0.1934714913,0.5531736016,-0.3191607296,-0.1221157163,0.0084841102,0.1728852391,-0.1787237525,0.0444031768,0.0144082429,0.2223311216,0.5070413947,-0.2224069536,0.2185738385,-0.4622749686,-0.1283478737,-0.1358893216,0.1796104163,-0.1460784674,-0.0783130527,0.0963894874,0.1252889782,0.163143605,-0.1422412395,0.3579080105,-0.0693807527,0.0396495759,-0.2039281875,-0.2201946378,0.0137812095,0.1307834387,0.1039626971,0.0551503152,-0.1753356308,-0.0369680114,0.2171099186,-0.1871833205,-0.2070101798,-0.0715540051,0.2629823387,0.0267733559,0.025153663,-0.4134668708,-0.0887362659,0.3430744112,0.062088266,0.1139909923,-0.2957776785,0.0404423103,-0.0858764797,0.2037117779,0.3430534005,-0.1321642697,0.0464633144,0.1779045761,-0.1550109833,0.2246282697,0.296744436,-0.1922280043,0.0937530622,-0.0989214331,0.5539204478,0.5879890919,-0.4102394879,-0.4520799518,0.1696793735,-0.2588975728,0.1952017099,-0.1164151728,0.1070558354,0.326749444,-0.117432341,0.306620568,0.6193355918,-0.1105521396,0.1110311449,-0.215863362,0.0245816521,0.0014836154,0.2785247266,0.2359109074,-0.1938206851,0.1145048365,0.2641795576,0.243651405,-0.3050079346,0.1434532404,0.2776193917,0.0995111689,0.0506758131,-0.1596455127,-0.3252392113,-0.2702022195,0.2318239212,-0.1383464932,0.1085467264,-0.3714389801,0.0401465409,-0.2963201106,0.0447943099,-0.5638593435,-0.1213942021,0.1630701125,0.2012814134,0.0795095712,0.0613085926,-0.3085088432,0.2793070674,-0.1050261855,-0.0565539338,-0.5994650126,0.2591193616,-0.0674927309,-0.3134038746,-0.0747863129,0.3225338161,0.0828683749,-0.13501966,-0.3367395401,0.2787976265,-0.0855251029,0.0143846814,0.0664433986,-0.1704661995,0.2425206453,-0.4210937619,0.1616301388,0.0719261542,0.12785393,0.0151306158,0.1468042135,0.3288503289,0.0096400762,0.0002207728,-0.0022295045,-0.0426436365,0.4274256229,-0.0020315761,0.0864704177,-0.257260561,0.2339752018,-0.0238232464,0.1626571417,-0.0563809276,-0.1947318316,-0.0884142816,0.5688619018,-0.1365721375,0.1961637586,0.1271996349,-0.0223116353,0.0912192017,-0.0803106651,0.0161401406,0.2614569962,0.0181957334,0.0555701181,0.0068808333,0.1075893268,0.02909513,0.1412385255,-0.084366791,-0.0864135996,-0.0710775107,-0.0867036805,-0.0300667249,-0.3335590065,-0.1923588961,-0.1055369303,0.3316015005,-0.1660034209,-0.0535194837,-0.1744429022,-0.1915307194,0.037373513,-0.1773097664,-0.1640250832,-0.2362444401,-0.2318590879,0.0255354829,0.1254676878,0.322096318,-0.4353750944,-0.35697487,0.1276556998,-0.3397228718,0.0316216648,-0.2119199336,-0.2546372116,0.1059964895,0.3112894893,0.1780466586,0.0755409226,-0.389688462,0.1155692562,-0.2473884374,-0.0755902529,0.0593283139,-0.3163183928,0.1726506948,-0.0393578149,0.0839216486,-0.0562964082,-0.181186527,0.2916715443,-0.2333991379,-0.1716851741,0.206258446,0.0019390341,-0.19655931,0.0024222732,-0.1196481287,-0.3280853927,-0.3669332564,0.0012898719,-0.0777989402,0.1852790117,0.1442967504,-0.2602292299,-0.1182346195,0.1143380255,-0.1334733516,-0.1536588073,-0.4095715284,0.4874741733,-0.372284323,-0.2143343836,-0.0219009388,-0.1756924838,0.4781981111,0.1308036745,-0.22841084,-0.0028280204,-0.3269365728,0.2939243317,0.0070022531,0.0189219732,0.1398832947,0.0004282062,0.0058593466,0.0415178165,-0.0540438183,-0.1960995942,0.2766460776,0.1708433926,0.1009547412,0.383543551,-0.1219677925,0.1801291257,-0.0492355041,0.3242917955,0.5478320718,0.0955017433,0.0651276484,-0.0233803149,-0.4221803546,-0.0489595979,-0.0551072024,0.0486402884,0.0079956595,-0.0325141177,-0.105156444,-0.0710553825,-0.1858091354,-0.1542174369,-0.2681567669,0.1497886032,-0.2403441966,0.2449389249,-0.1164389551,0.0076551237,-0.3299103379,0.0237852167,-0.0713196546,0.1350256205,0.2146928608,0.0793461651,-0.1952005923,0.3681942523,-0.4347116053,0.0998158306,0.3249466121,0.703830719,0.0326968022,0.0539382137,0.0414398424,-0.0472928733,-0.0577151068,-0.1116312221,0.1558757871,0.1666961759,-0.2209492028,-0.2402218878,-0.0395986289,-0.1052687243,0.3003042936,0.2227607071,-0.0354341567,-0.3422704339,-0.1095726341,0.3165017962,0.1825249046,-0.1780320257,-0.0400575511,-0.4110560715,-0.5928244591,-0.4134850502,0.0432143509,0.2406070083,0.4273461401,0.3378159106,0.1650990695,-0.2048251033,-0.2078794092,0.2217248082,0.3031234145,0.0947903246,0.0685112104,0.0718200728,-0.2338235378,-0.3942333758,0.1363064498,0.5306170583,0.1089847833,-0.7092872262,-0.078453362,-0.2032940984,0.2511305511,0.059877906,-0.0541696139,0.1121880934,-0.1826997846,0.0938625634,0.043718785,0.1091841459,0.1448744535,0.2128843516,-0.6021432281,-0.4522859156,0.1971878111,0.0163849033,0.2952745855,0.3392183483,-0.1642782241,-0.3164373934,-0.0895475596,0.1628365666,0.5518426895,0.0499165542,0.1146935076,0.3466674387,0.0413424149,0.058239501,0.0459595174,0.1835665703,-0.2096627951,-0.2382662147,-0.0543798544,-0.025275711,0.1658082753,-0.0543145016,0.0872022435,0.2775225341,0.0734330863,0.0351325609,0.2459097058,0.198462829,-0.2746990919,-0.4539572895,-0.5819117427,0.1532522589,-0.3220199943,-0.0309987292,-0.196727097,-0.2522437572,0.0267227255,-0.2227253616,0.0441954397,0.224859193,-0.1557139754,0.2739735544,-0.1650171429,-0.5444099903,0.2062320113,0.0880706459,0.3015840352,-0.0706007481,0.1350811571,0.1731298119,0.1769678742,-0.0216342527,-0.0828447193,-0.0253757741,0.0173967034,-0.0310315695,-0.2499279827,0.0698867887,-0.0188626833,-0.3566619158,-0.2213180214,0.2003680915,-0.0024183826,-0.0873170272,-0.2053951919,-0.1993931383,0.0175070502,-0.1896483749,0.1594203115,0.0621682294,-0.2342941463,0.0830627158,0.0320613347,-0.3447245657,0.0183398556,0.0949900448,0.1793593764,-0.4032029808,0.5781168342,0.1079861,-0.0876559317,-0.123005636,0.0509770848,0.188012287,-0.5171360373,0.0444077961,-0.0069698733,0.0980357304,-0.1275515854,0.0985043943,0.0108745405,-0.096129775,0.1677367389,-0.4163189828,-0.5901713967,0.1163460761,-0.1201160848,0.2132864892,0.0890787989,0.1483061165,-0.1083181128,-0.0477421358,-0.304250598,0.1132213548,-0.3167765141,0.2198823541,0.2084016055,-0.0789639801,0.0466105267,0.0550280698,0.2011434734,0.3281204402,-0.0766396374,-0.1902381778,0.0340617821,0.08731208,-0.0867043063,-0.4576300979,-0.01823573,0.2222194225,-0.0597430989,-0.1353600472,0.1167752296,0.1854208857,-0.0626491234,0.0720704943,0.1372987181,0.1072310358,-0.1261375993,0.298120141,-0.209562853,0.4348088503,0.0386724807,0.2651923597,0.0883063301,0.0721179917,-0.2311838567,-0.1048307195,0.1180974394,0.1814716011,0.14326258,-0.5272974372,0.1788813025,0.2601992488,-0.0461180806,0.3760097027,-0.1835526228,-0.0165387448,0.120889388,0.2417025566,-0.1411454827,-0.0686583668,0.3440247476,-0.2144441307,0.0001063908,0.4460248053,0.1517319828,0.1151359379,0.2358342707,0.1073320955,0.2825607061,-0.0068990048,0.2959387302,0.475163877,-0.1015865877,-0.0476528928,0.2322836965,0.1343567818,0.0184182413,0.6424046159,-0.442773819,0.5353500843,-0.0505037382,0.2583877444,-0.3224719167,-0.3795407712,-0.2869464159,-0.1631638855,-0.0148015292,0.2366979718,0.1603486538,-0.3238866031,-0.1891293228,0.0048719519,-0.0439247601,0.1654580981,-0.2276591212,0.1310626715,-0.0501659214,-0.0186570399,-0.1377204359,-0.0187568478,0.0113090724,0.1884762943,-0.0382181816,0.2333977371,-0.1012568995,-0.0860914066,0.159076497,0.2796638608,-0.0650407374,-0.2643450499,0.4634735584,0.5010718703,-0.0114834644,-0.0209639426,0.1890849322,0.3574219346,0.4347184002,-0.2235045582,0.1034589782,0.000924604,0.0185824297,0.0500134677,0.1436081976,0.0140555101,0.6835596561,0.1786809713,0.2524062097,-0.1252028644,-0.0167536419,0.0970192775,-0.2486756146,0.0330268592,0.2661145031,0.1370356828,-0.0323732123,-0.2701379657,-0.0949243903,-0.2792062163,0.006541695,0.384693414,0.0213529747,0.106054008,-0.1315313131,0.0763784647,0.0020454954,0.4745551646,0.3403455615,-0.0144478558,-0.3034825623,-0.2072483152,-0.7285958529,0.194361046,-0.1984153986,-0.3285027146,0.2497081608,-0.0827106461,-0.0733957067,0.3360482752,0.2181287706,0.398761332,0.0736603737,0.1313437372,-0.3173839748,-0.0350323953,0.4439384341,0.0782283843,-0.0843572021,-0.3402358592,0.1098290831,0.0742676035,0.1089789867,-0.1935682148,-0.0207891837,-0.0690447688,-0.13243635,0.3674142957,0.2805423439,0.393017292,-0.1928328872,-0.1226181611,-0.17145513,-0.3176394999,-0.081048727,0.3091652095,0.0883627608,0.5725002289,0.0104075894,-0.087408416,-0.6893698573,0.3223952949,-0.1301128119,0.0926097482,0.0033725582,0.1471588314,-0.074550949,0.0625717491,0.006926755,0.1489863098,0.0457483046,0.4057685733,-0.1709182858,-0.5643859506,0.3966764808,-0.0855885968,-0.685715735,-0.110391736,0.0107939122,-0.0288262423,-0.0517250597,-0.3539994061,0.3004591465,0.2369799316,-0.0079769436,-0.1723502874,0.2072672099,0.0321722776,0.0402625687,-0.0043208124,0.4464668632,-0.1106186807,-0.0629551485,0.1378686875,-0.2023232877]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/468","title":"UnicodeDecodeError while loading PAN-X task of XTREME dataset","comments":"I realised I was overthinking the problem, so decided to just run the regexp over the codebase and make the PR. In other words, we can ignore my comments about using the CLI \ud83d\ude38 ","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!","comment_length":34,"text":"UnicodeDecodeError while loading PAN-X task of XTREME dataset\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nI'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: \r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnicodeDecodeError Traceback (most recent call last)\r\n<ipython-input-5-1d61f439b843> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 528 ignore_verifications = ignore_verifications or save_infos\r\n 529 # Download\/copy dataset processing script\r\n--> 530 module_path, hash = prepare_module(path, download_config=download_config, dataset=True)\r\n 531 \r\n 532 # Get dataset builder class from the processing script\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in prepare_module(path, download_config, dataset, force_local_path, **download_kwargs)\r\n 265 \r\n 266 # Download external imports if needed\r\n--> 267 imports = get_imports(local_path)\r\n 268 local_imports = []\r\n 269 library_imports = []\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in get_imports(file_path)\r\n 156 lines = []\r\n 157 with open(file_path, mode=\"r\") as f:\r\n--> 158 lines.extend(f.readlines())\r\n 159 \r\n 160 logger.info(\"Checking %s for additional imports.\", file_path)\r\n\r\n\/usr\/lib\/python3.6\/encodings\/ascii.py in decode(self, input, final)\r\n 24 class IncrementalDecoder(codecs.IncrementalDecoder):\r\n 25 def decode(self, input, final=False):\r\n---> 26 return codecs.ascii_decode(input, self.errors)[0]\r\n 27 \r\n 28 class StreamWriter(Codec,codecs.StreamWriter):\r\n\r\nUnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 111: ordinal not in range(128)\r\n```\r\n\r\n## Steps to reproduce\r\nInstall from nlp's master branch\r\n```python\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```\r\nthen run\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\n```\r\n\r\n## OS \/ platform details\r\n\r\n- `nlp` version: latest from master\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: True\r\n- Using distributed or parallel set-up in script?: False\r\n\r\n## Proposed solution\r\nEither change [line 762](https:\/\/github.com\/huggingface\/nlp\/blob\/7ada00b1d62f94eee22a7df38c6b01e3f27194b7\/datasets\/xtreme\/xtreme.py#L762) in `xtreme.py` to include UTF-8 encoding:\r\n\r\n```\r\n# old\r\nwith open(filepath) as f\r\n# new\r\nwith open(filepath, encoding='utf-8') as f\r\n```\r\n\r\nor raise a warning that suggests setting the locale explicitly, e.g.\r\n```python\r\nimport locale\r\nlocale.setlocale(locale.LC_ALL, 'C.UTF-8')\r\n```\r\nI have a preference for the first solution. Let me know if you agree and I'll be happy to implement the simple fix!\nI realised I was overthinking the problem, so decided to just run the regexp over the codebase and make the PR. In other words, we can ignore my comments about using the CLI \ud83d\ude38 ","embeddings":[-0.3326011896,-0.0930106342,-0.1336479485,0.2554301322,0.3521757722,-0.0277493317,0.2249831408,0.3496463001,-0.0286989305,0.038228903,-0.0904103443,0.1502787173,-0.0761749223,-0.1970372051,0.1891056299,-0.2467030734,-0.0738269314,0.1824576855,0.1078069955,0.0069923094,-0.3558878899,0.1144126877,-0.1869958341,0.2637093067,-0.3771478236,0.0810821429,0.2978654206,0.3204651177,-0.193280682,-0.6116119623,0.4130367041,-0.0536639951,0.4325025082,0.3163448274,-0.0001086041,0.0825855061,0.3929110467,-0.1319698691,-0.3346782327,-0.1939976513,-0.5915244818,-0.1812518835,-0.0098235868,-0.2595872581,0.1902403831,0.0562078469,-0.0642362684,-0.3830473125,0.1811341196,0.2071898282,0.24163872,0.3705638647,0.0213383194,0.0912284628,-0.045574829,0.1074261367,-0.1542558074,0.249418959,0.4346242249,0.0627749935,-0.019149052,0.2926935852,-0.0486163385,0.248120591,-0.1023555249,0.1306457818,0.1399910003,-0.4660047889,0.1134835854,0.0036376556,0.2718741894,-0.3664452136,-0.2865583897,-0.2252654582,-0.1094700545,-0.5443437099,0.2644951046,0.0905644149,-0.159659639,-0.0211321656,0.113159202,-0.0423871912,-0.1735193729,0.0391214527,0.0543593504,0.3878184855,-0.0600738861,-0.0513223708,-0.0941757336,-0.1786022931,0.4082849622,-0.1458789855,-0.1436942965,0.324629128,-0.2863637805,0.072422564,-0.0550461523,-0.3371772468,0.0668165982,-0.0463103987,-0.078823179,0.2744791806,-0.0495554432,0.053195674,0.3093065917,0.1556752175,0.2095072418,0.2245565802,0.1834983528,0.1982241273,0.0829440206,-0.0012048965,-0.1758536994,-0.1771315336,0.3654833138,0.1934714913,0.5531736016,-0.3191607296,-0.1221157163,0.0084841102,0.1728852391,-0.1787237525,0.0444031768,0.0144082429,0.2223311216,0.5070413947,-0.2224069536,0.2185738385,-0.4622749686,-0.1283478737,-0.1358893216,0.1796104163,-0.1460784674,-0.0783130527,0.0963894874,0.1252889782,0.163143605,-0.1422412395,0.3579080105,-0.0693807527,0.0396495759,-0.2039281875,-0.2201946378,0.0137812095,0.1307834387,0.1039626971,0.0551503152,-0.1753356308,-0.0369680114,0.2171099186,-0.1871833205,-0.2070101798,-0.0715540051,0.2629823387,0.0267733559,0.025153663,-0.4134668708,-0.0887362659,0.3430744112,0.062088266,0.1139909923,-0.2957776785,0.0404423103,-0.0858764797,0.2037117779,0.3430534005,-0.1321642697,0.0464633144,0.1779045761,-0.1550109833,0.2246282697,0.296744436,-0.1922280043,0.0937530622,-0.0989214331,0.5539204478,0.5879890919,-0.4102394879,-0.4520799518,0.1696793735,-0.2588975728,0.1952017099,-0.1164151728,0.1070558354,0.326749444,-0.117432341,0.306620568,0.6193355918,-0.1105521396,0.1110311449,-0.215863362,0.0245816521,0.0014836154,0.2785247266,0.2359109074,-0.1938206851,0.1145048365,0.2641795576,0.243651405,-0.3050079346,0.1434532404,0.2776193917,0.0995111689,0.0506758131,-0.1596455127,-0.3252392113,-0.2702022195,0.2318239212,-0.1383464932,0.1085467264,-0.3714389801,0.0401465409,-0.2963201106,0.0447943099,-0.5638593435,-0.1213942021,0.1630701125,0.2012814134,0.0795095712,0.0613085926,-0.3085088432,0.2793070674,-0.1050261855,-0.0565539338,-0.5994650126,0.2591193616,-0.0674927309,-0.3134038746,-0.0747863129,0.3225338161,0.0828683749,-0.13501966,-0.3367395401,0.2787976265,-0.0855251029,0.0143846814,0.0664433986,-0.1704661995,0.2425206453,-0.4210937619,0.1616301388,0.0719261542,0.12785393,0.0151306158,0.1468042135,0.3288503289,0.0096400762,0.0002207728,-0.0022295045,-0.0426436365,0.4274256229,-0.0020315761,0.0864704177,-0.257260561,0.2339752018,-0.0238232464,0.1626571417,-0.0563809276,-0.1947318316,-0.0884142816,0.5688619018,-0.1365721375,0.1961637586,0.1271996349,-0.0223116353,0.0912192017,-0.0803106651,0.0161401406,0.2614569962,0.0181957334,0.0555701181,0.0068808333,0.1075893268,0.02909513,0.1412385255,-0.084366791,-0.0864135996,-0.0710775107,-0.0867036805,-0.0300667249,-0.3335590065,-0.1923588961,-0.1055369303,0.3316015005,-0.1660034209,-0.0535194837,-0.1744429022,-0.1915307194,0.037373513,-0.1773097664,-0.1640250832,-0.2362444401,-0.2318590879,0.0255354829,0.1254676878,0.322096318,-0.4353750944,-0.35697487,0.1276556998,-0.3397228718,0.0316216648,-0.2119199336,-0.2546372116,0.1059964895,0.3112894893,0.1780466586,0.0755409226,-0.389688462,0.1155692562,-0.2473884374,-0.0755902529,0.0593283139,-0.3163183928,0.1726506948,-0.0393578149,0.0839216486,-0.0562964082,-0.181186527,0.2916715443,-0.2333991379,-0.1716851741,0.206258446,0.0019390341,-0.19655931,0.0024222732,-0.1196481287,-0.3280853927,-0.3669332564,0.0012898719,-0.0777989402,0.1852790117,0.1442967504,-0.2602292299,-0.1182346195,0.1143380255,-0.1334733516,-0.1536588073,-0.4095715284,0.4874741733,-0.372284323,-0.2143343836,-0.0219009388,-0.1756924838,0.4781981111,0.1308036745,-0.22841084,-0.0028280204,-0.3269365728,0.2939243317,0.0070022531,0.0189219732,0.1398832947,0.0004282062,0.0058593466,0.0415178165,-0.0540438183,-0.1960995942,0.2766460776,0.1708433926,0.1009547412,0.383543551,-0.1219677925,0.1801291257,-0.0492355041,0.3242917955,0.5478320718,0.0955017433,0.0651276484,-0.0233803149,-0.4221803546,-0.0489595979,-0.0551072024,0.0486402884,0.0079956595,-0.0325141177,-0.105156444,-0.0710553825,-0.1858091354,-0.1542174369,-0.2681567669,0.1497886032,-0.2403441966,0.2449389249,-0.1164389551,0.0076551237,-0.3299103379,0.0237852167,-0.0713196546,0.1350256205,0.2146928608,0.0793461651,-0.1952005923,0.3681942523,-0.4347116053,0.0998158306,0.3249466121,0.703830719,0.0326968022,0.0539382137,0.0414398424,-0.0472928733,-0.0577151068,-0.1116312221,0.1558757871,0.1666961759,-0.2209492028,-0.2402218878,-0.0395986289,-0.1052687243,0.3003042936,0.2227607071,-0.0354341567,-0.3422704339,-0.1095726341,0.3165017962,0.1825249046,-0.1780320257,-0.0400575511,-0.4110560715,-0.5928244591,-0.4134850502,0.0432143509,0.2406070083,0.4273461401,0.3378159106,0.1650990695,-0.2048251033,-0.2078794092,0.2217248082,0.3031234145,0.0947903246,0.0685112104,0.0718200728,-0.2338235378,-0.3942333758,0.1363064498,0.5306170583,0.1089847833,-0.7092872262,-0.078453362,-0.2032940984,0.2511305511,0.059877906,-0.0541696139,0.1121880934,-0.1826997846,0.0938625634,0.043718785,0.1091841459,0.1448744535,0.2128843516,-0.6021432281,-0.4522859156,0.1971878111,0.0163849033,0.2952745855,0.3392183483,-0.1642782241,-0.3164373934,-0.0895475596,0.1628365666,0.5518426895,0.0499165542,0.1146935076,0.3466674387,0.0413424149,0.058239501,0.0459595174,0.1835665703,-0.2096627951,-0.2382662147,-0.0543798544,-0.025275711,0.1658082753,-0.0543145016,0.0872022435,0.2775225341,0.0734330863,0.0351325609,0.2459097058,0.198462829,-0.2746990919,-0.4539572895,-0.5819117427,0.1532522589,-0.3220199943,-0.0309987292,-0.196727097,-0.2522437572,0.0267227255,-0.2227253616,0.0441954397,0.224859193,-0.1557139754,0.2739735544,-0.1650171429,-0.5444099903,0.2062320113,0.0880706459,0.3015840352,-0.0706007481,0.1350811571,0.1731298119,0.1769678742,-0.0216342527,-0.0828447193,-0.0253757741,0.0173967034,-0.0310315695,-0.2499279827,0.0698867887,-0.0188626833,-0.3566619158,-0.2213180214,0.2003680915,-0.0024183826,-0.0873170272,-0.2053951919,-0.1993931383,0.0175070502,-0.1896483749,0.1594203115,0.0621682294,-0.2342941463,0.0830627158,0.0320613347,-0.3447245657,0.0183398556,0.0949900448,0.1793593764,-0.4032029808,0.5781168342,0.1079861,-0.0876559317,-0.123005636,0.0509770848,0.188012287,-0.5171360373,0.0444077961,-0.0069698733,0.0980357304,-0.1275515854,0.0985043943,0.0108745405,-0.096129775,0.1677367389,-0.4163189828,-0.5901713967,0.1163460761,-0.1201160848,0.2132864892,0.0890787989,0.1483061165,-0.1083181128,-0.0477421358,-0.304250598,0.1132213548,-0.3167765141,0.2198823541,0.2084016055,-0.0789639801,0.0466105267,0.0550280698,0.2011434734,0.3281204402,-0.0766396374,-0.1902381778,0.0340617821,0.08731208,-0.0867043063,-0.4576300979,-0.01823573,0.2222194225,-0.0597430989,-0.1353600472,0.1167752296,0.1854208857,-0.0626491234,0.0720704943,0.1372987181,0.1072310358,-0.1261375993,0.298120141,-0.209562853,0.4348088503,0.0386724807,0.2651923597,0.0883063301,0.0721179917,-0.2311838567,-0.1048307195,0.1180974394,0.1814716011,0.14326258,-0.5272974372,0.1788813025,0.2601992488,-0.0461180806,0.3760097027,-0.1835526228,-0.0165387448,0.120889388,0.2417025566,-0.1411454827,-0.0686583668,0.3440247476,-0.2144441307,0.0001063908,0.4460248053,0.1517319828,0.1151359379,0.2358342707,0.1073320955,0.2825607061,-0.0068990048,0.2959387302,0.475163877,-0.1015865877,-0.0476528928,0.2322836965,0.1343567818,0.0184182413,0.6424046159,-0.442773819,0.5353500843,-0.0505037382,0.2583877444,-0.3224719167,-0.3795407712,-0.2869464159,-0.1631638855,-0.0148015292,0.2366979718,0.1603486538,-0.3238866031,-0.1891293228,0.0048719519,-0.0439247601,0.1654580981,-0.2276591212,0.1310626715,-0.0501659214,-0.0186570399,-0.1377204359,-0.0187568478,0.0113090724,0.1884762943,-0.0382181816,0.2333977371,-0.1012568995,-0.0860914066,0.159076497,0.2796638608,-0.0650407374,-0.2643450499,0.4634735584,0.5010718703,-0.0114834644,-0.0209639426,0.1890849322,0.3574219346,0.4347184002,-0.2235045582,0.1034589782,0.000924604,0.0185824297,0.0500134677,0.1436081976,0.0140555101,0.6835596561,0.1786809713,0.2524062097,-0.1252028644,-0.0167536419,0.0970192775,-0.2486756146,0.0330268592,0.2661145031,0.1370356828,-0.0323732123,-0.2701379657,-0.0949243903,-0.2792062163,0.006541695,0.384693414,0.0213529747,0.106054008,-0.1315313131,0.0763784647,0.0020454954,0.4745551646,0.3403455615,-0.0144478558,-0.3034825623,-0.2072483152,-0.7285958529,0.194361046,-0.1984153986,-0.3285027146,0.2497081608,-0.0827106461,-0.0733957067,0.3360482752,0.2181287706,0.398761332,0.0736603737,0.1313437372,-0.3173839748,-0.0350323953,0.4439384341,0.0782283843,-0.0843572021,-0.3402358592,0.1098290831,0.0742676035,0.1089789867,-0.1935682148,-0.0207891837,-0.0690447688,-0.13243635,0.3674142957,0.2805423439,0.393017292,-0.1928328872,-0.1226181611,-0.17145513,-0.3176394999,-0.081048727,0.3091652095,0.0883627608,0.5725002289,0.0104075894,-0.087408416,-0.6893698573,0.3223952949,-0.1301128119,0.0926097482,0.0033725582,0.1471588314,-0.074550949,0.0625717491,0.006926755,0.1489863098,0.0457483046,0.4057685733,-0.1709182858,-0.5643859506,0.3966764808,-0.0855885968,-0.685715735,-0.110391736,0.0107939122,-0.0288262423,-0.0517250597,-0.3539994061,0.3004591465,0.2369799316,-0.0079769436,-0.1723502874,0.2072672099,0.0321722776,0.0402625687,-0.0043208124,0.4464668632,-0.1106186807,-0.0629551485,0.1378686875,-0.2023232877]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/445","title":"DEFAULT_TOKENIZER import error in sacrebleu","comments":"This issue was resolved by #447 ","body":"Latest Version 0.3.0\r\n\r\nWhen loading the metric \"sacrebleu\" there is an import error due to the wrong path\r\n![image](https:\/\/user-images.githubusercontent.com\/5303103\/88633063-2c5e5f00-d0bd-11ea-8ca8-4704dc975433.png)\r\n","comment_length":6,"text":"DEFAULT_TOKENIZER import error in sacrebleu\nLatest Version 0.3.0\r\n\r\nWhen loading the metric \"sacrebleu\" there is an import error due to the wrong path\r\n![image](https:\/\/user-images.githubusercontent.com\/5303103\/88633063-2c5e5f00-d0bd-11ea-8ca8-4704dc975433.png)\r\n\nThis issue was resolved by #447 ","embeddings":[-0.4125730991,-0.0699953437,0.0429188125,0.1354275197,0.2939696014,-0.2114159018,0.043850977,0.280447185,-0.2095745206,0.2177171558,0.1345955133,0.2010787576,-0.232255578,-0.0183717348,-0.1110823452,-0.0975368246,0.0793196931,-0.0302112103,0.157021746,-0.0811711252,-0.360172689,0.3452454209,-0.1542122513,0.1644795686,-0.3349221647,0.2308065891,0.057583902,0.1387517899,-0.5949498415,-0.5236250758,-0.071520105,0.1469681859,0.0067394171,0.4471901953,-0.0001126273,-0.0756276771,0.3987275362,-0.0858111531,-0.0512011647,-0.7973020077,-0.3311479092,-0.2482031584,0.0273233075,0.0524585955,0.0831126943,-0.2415523827,0.1402032971,0.0640299246,0.3024266958,0.2500537038,0.2064070255,0.3299705088,-0.1662169695,-0.2671813965,-0.0661868677,0.1536598504,0.1179943532,0.280584693,-0.3428792357,0.2021982372,-0.0342235789,0.0623295121,0.0332953706,0.1431213766,0.244626686,-0.2276738733,-0.0459414795,-0.151922375,0.2110078633,0.0617843084,0.2683828771,0.0334023573,0.014271074,0.3470304608,-0.0629438907,-0.5239708424,0.3966137767,-0.0021090801,0.0269983392,-0.1778625697,-0.2122482955,-0.2711434662,-0.2409944832,0.1097337455,-0.5039749742,-0.001399328,-0.2307457775,0.0489590392,0.1088490263,-0.1549095362,-0.4511323273,0.0732322112,-0.2464932501,0.0286836941,0.0379618593,0.2107929289,-0.1182248741,-0.5402991772,-0.1938422024,0.4378305972,-0.1969600171,0.3586630523,0.0602053739,0.2446505278,-0.0595950931,0.435908258,0.2807990313,0.1907105297,0.1889999062,0.10185574,0.0801721737,-0.0561110899,-0.2251584083,-0.3853940368,-0.1538377851,-0.0069749844,0.0071721748,-0.1190709099,-0.1292404085,0.0186739974,0.5230765939,0.0892335102,-0.0228879396,0.3327384889,-0.0185852107,0.1755400449,-0.2292932272,0.3051975369,-0.0542149693,0.027838178,-0.1423380524,0.1282697767,-0.1821418107,-0.1235274598,-0.1057477295,0.1927042007,0.2347793877,-0.2147533894,0.2812688351,-0.346911341,0.2365703285,-0.0607214235,-0.0305998232,0.1530026495,0.0487942882,0.0408831388,0.2105650604,-0.5193985701,-0.1042810306,-0.147598505,-0.7275771499,-0.3960540891,0.0403186157,0.2448393553,-0.1063863486,-0.0252684932,-0.5425841212,-0.0102008237,-0.3208436668,0.1456013471,-0.3610146046,0.0137292435,-0.256052047,-0.1155799404,0.1902577728,0.0498513654,-0.0073749437,-0.3944700062,0.2273274958,-0.0484531187,0.4621033669,0.0317466483,-0.03308082,0.1538286358,-0.1788678169,0.4428036511,0.4685977399,-0.4466293752,-0.0844063386,0.034480013,-0.1381628364,-0.4021086097,-0.0291158129,-0.208496049,0.6687312126,0.168782413,0.0610159039,0.1760977358,0.2275056392,-0.1142690033,-0.1474516392,-0.3586615622,-0.1456478089,0.0064011435,0.5267480016,0.000135601,0.1413280815,0.3352245092,-0.1067742854,0.0509419888,-0.1586083323,0.3908594847,0.5844450593,-0.2794567645,0.0921436101,-0.0118894055,0.5020543933,0.1438550502,-0.0533121638,0.0984106511,0.0024516871,0.0060555954,-0.560094893,-0.0266908258,-0.1835996509,0.2152489573,0.1773850471,0.1908313334,-0.2387720495,0.4909118414,-0.2172756642,-0.1795569658,-0.0123371845,0.3249999881,0.2611448467,0.1017114744,-0.0552220605,0.0028529596,0.0990743712,0.0957082734,0.3900083899,-0.3196001053,-0.0217567831,0.3941647112,0.2479362041,0.3725735843,-0.0490691327,0.2150389254,0.3278882802,-0.3150615692,-0.3130422831,0.0036541056,-0.1108112112,0.0647302717,0.2578459978,0.3309282959,0.2777808011,-0.1442936361,0.0122487387,0.234312579,0.1438721716,0.041050002,-0.0097971745,-0.5364122987,0.0365688428,-0.0781340078,-0.0895922706,-0.1889180243,0.4686041176,-0.2840056717,0.0940222815,0.2148622572,-0.0348228998,0.1646105796,-0.0781411231,-0.0392251424,-0.1275898665,0.1556358188,0.14092803,0.039250236,0.2635742128,0.1010544077,0.0575909913,-0.0952434763,0.3230043352,0.0661630332,-0.0097553814,0.1367435902,-0.0039320332,0.0228811949,-0.4077219367,0.1036828831,0.1029135734,0.4168943465,-0.0968361869,0.0952314734,-0.3627411425,0.1068911403,-0.0216717925,0.0200916026,-0.0395632423,-0.2447437346,0.18332389,0.1254788786,0.0014630584,0.3860371411,0.2588261366,0.4560129642,-0.0409355648,0.0400445238,-0.0536252782,0.0326574259,-0.0599717423,-0.0008812005,0.1576633155,-0.2970727384,0.1009296924,-0.4718016982,0.0809790865,-0.2696547508,-0.5217746496,0.2253808081,0.3393008113,0.1815092564,0.2567728758,-0.138029024,-0.0151601741,0.0567828901,0.2516124249,-0.0135222077,-0.1272214055,0.3323451877,-0.2856973708,-0.2746245563,-0.0241161194,-0.0442649312,0.1100915074,-0.1533799917,-0.2041442394,0.1435320377,-0.0277631618,-0.1314182878,0.2919033468,0.1099715158,-0.0486999862,0.0063002771,-0.0132760294,-0.3140916228,0.2325320244,-0.2457997203,-0.4061046243,0.2254401296,0.0589180663,0.4801911414,-0.0088010691,-0.3636650443,-0.4508512616,0.1236513853,0.1257379651,0.2320866138,-0.3062933087,-0.0891636908,0.0634818897,-0.2460864931,-0.0312483031,0.1387604475,0.076910004,-0.1069568992,0.2210989147,-0.077742219,-0.0126053896,0.5285239816,0.8168204427,0.030533446,-0.0792684034,0.1653148383,-0.1387237459,0.3202609718,0.1540472209,-0.1788098812,0.1737684011,-0.0135755343,0.0440947451,0.2292462438,0.0533752143,-0.0083351703,0.0241002757,0.143770799,0.0113228718,-0.0442584008,-0.2406439185,-0.0880899355,0.191952467,0.0743046552,-0.2100147009,-0.2089798152,-0.0582827181,0.3071797788,0.5154348016,-0.2305250913,0.151811108,-0.3834403455,0.3136160076,-0.2664975822,0.2836049497,-0.1832558811,0.3542811275,-0.1398519427,0.0774932653,-0.055992391,-0.1808894128,0.3043343425,0.1227063388,-0.1989566535,-0.0795590207,-0.1122450531,-0.3071695566,-0.0905008838,0.1656330377,0.0901979059,0.5106942058,0.5559943318,-0.0941991583,-0.024993863,-0.1366117597,0.0049507273,-0.2010946125,0.1753307134,-0.3789049089,-0.3285961449,-0.2082452774,-0.1322428733,-0.1122445092,0.3324175775,0.2260736823,0.2527216971,-0.5233534575,-0.3367653787,0.0395866297,0.0098105278,0.2100080103,-0.2481686026,-0.021238761,0.1185524687,0.2434491664,0.4131887555,-0.049499806,0.1213256419,-0.7862876058,0.1682285964,-0.0341751501,0.0270168222,0.2245048285,-0.322058022,0.1210909262,0.1327786893,0.2632660866,-0.2670386732,0.761333704,0.162446022,0.2125279903,0.1378480643,0.1453982741,-0.1238711327,0.0840023085,-0.0182089955,0.1271652877,0.0013299721,-0.1341308057,0.2155655026,0.0938770473,0.8032046556,0.1306119114,-0.1137161106,0.367284447,-0.303242445,-0.069577612,-0.3037224412,0.1341400594,-0.2066722214,-0.0147353243,-0.1108378544,-0.2117111087,0.2705465555,-0.2589665055,-0.006685229,0.0114501081,-0.0079256669,0.3128745556,-0.0695656687,0.1052605063,0.1196944341,-0.0565739796,0.7305675149,0.1251752228,-0.1245544776,0.1956546456,-0.2058725953,0.0866416693,0.3636448383,-0.1448428631,-0.0565091558,-0.2062043995,-0.228182584,0.2518517077,0.0961373374,0.0705247074,-0.2343803346,0.2379118651,0.5934066176,0.0374131128,-0.197210744,0.2593189776,-0.0851148218,-0.1345479637,-0.0693983212,-0.1479307264,0.0896722302,-0.2028011084,0.1172882542,0.1149019226,0.0494460054,0.053993836,-0.0322048254,-0.068527244,0.4683007002,0.1660255939,-0.0345971696,-0.0091920216,0.1028945968,-0.3790332079,0.1616051495,-0.1761184335,-0.0258166399,-0.0181978215,0.0519874357,-0.3057853878,0.0429637581,0.3041619062,-0.0146552911,-0.5754373074,0.328055948,-0.5961173177,-0.0512391366,-0.1878667176,0.2439985126,-0.0789447501,-0.3026345968,0.1911863685,-0.2577893436,-0.1445793062,0.0604118295,0.2919855118,0.4515883625,-0.3056320846,0.0098380605,-0.0051668519,-0.2250802666,0.0559735522,-0.3911862671,0.1812527925,-0.1793317795,-0.0456344113,-0.1576474905,-0.0035733203,-0.2696669996,0.432561934,-0.2981638014,-0.1371848434,-0.2143269628,-0.0616397038,-0.0660234466,-0.0978693962,0.0219255034,0.0339999162,-0.4398143589,-0.1430416256,-0.4225861132,0.0781408846,0.1690921187,-0.0592119955,0.1990292668,0.1686392426,-0.0935933515,0.078221269,-0.0688559562,0.1672129929,-0.0741972625,-0.1214887351,-0.1611390412,0.1164756268,0.0265764501,0.0421533771,0.4171618819,0.6010628939,0.1405715197,0.0629772469,-0.2203056514,-0.0162909739,-0.1597318947,0.1021550223,0.3021764457,-0.1296460479,0.0107992152,-0.2853483558,0.0532410517,-0.0760122985,-0.0123347761,-0.0969736055,-0.3296803534,-0.1686045676,0.4428855777,0.237291038,-0.0597443096,-0.2549481094,-0.3528338671,0.0937162936,-0.0282052718,0.0400422513,0.4351038635,-0.027850436,0.0970887616,0.189464286,0.3714250326,-0.1507577747,0.4795370102,-0.5654307604,-0.2948408425,0.1858148575,0.1943482608,0.1958592981,-0.1123495549,0.2312654853,-0.114695549,0.6388356686,0.1006007642,0.2134967893,-0.0944553167,-0.2042217404,0.0237975083,-0.1145593897,-0.0459019206,0.1157013103,-0.0060090008,0.2278102636,-0.0820896477,-0.3152232468,-0.3132202625,-0.0261412673,-0.1934112161,-0.1940247416,-0.3504871428,-0.2813907862,-0.1516655087,0.0963669568,0.1884887218,0.0734171569,-0.3783566356,0.038189996,-0.0488278419,-0.0149303852,-0.1611582488,0.1456154734,0.016544085,-0.2191004604,0.2508390546,-0.1218049005,0.0110879382,0.2106258273,0.3949640393,0.4081993699,0.294865042,0.0669165254,-0.4295985103,-0.0419691838,0.1308070719,-0.0468783639,0.2055462152,0.1768131703,-0.1480388343,0.1438212395,0.1889212728,-0.208901599,0.1430654526,0.1388206482,-0.0666351393,-0.044677373,0.2674077451,-0.0022230055,0.2428067476,-0.0238504764,-0.0876253471,-0.3921578228,0.2170976251,0.4422888756,0.0815343484,0.2721637487,-0.0184597671,0.0694886446,0.2008445859,0.0936984569,0.2038761526,-0.1767193526,-0.3890990317,-0.3485777974,-0.4551485181,0.4580906034,0.1478644907,0.0792569816,0.1122244522,0.171792239,-0.0611631274,-0.0077387076,-0.0274981409,0.2046683431,-0.115262486,-0.3387774229,-0.5690388083,0.2779392004,0.1576402634,0.0806853175,-0.1933607012,-0.252245903,0.0257319529,-0.2436074466,0.1078005433,0.108908616,-0.0743730217,-0.2990549505,-0.0862083808,0.3093788624,-0.0034825669,-0.1818174273,0.0528329574,0.0504653342,0.09807235,-0.203598842,-0.2036300749,0.0651658699,-0.0129474699,0.1797752529,-0.0634974986,0.1282417923,-0.4604611695,0.0099871429,0.3323314786,-0.2473523617,-0.2624680102,0.3200780153,-0.0102204606,0.1433022022,0.0956648886,0.3827979863,0.2215045244,0.2174238712,-0.0865352601,-0.28554371,0.5398635864,-0.3380818069,-0.2910089195,-0.0522173382,0.3276410997,0.4446735978,0.1580759585,-0.254514724,-0.1097493172,0.2365010828,-0.0031154994,-0.133142516,0.120987162,-0.2321683913,0.0999410152,-0.0967512056,0.230046466,-0.0541444793,-0.3525137901,0.0008787864,0.0252503473]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/444","title":"Keep loading old file even I specify a new file in load_dataset","comments":"Same here !","body":"I used load a file called 'a.csv' by \r\n```\r\ndataset = load_dataset('csv', data_file='.\/a.csv')\r\n```\r\nAnd after a while, I tried to load another csv called 'b.csv'\r\n```\r\ndataset = load_dataset('csv', data_file='.\/b.csv')\r\n```\r\nHowever, the new dataset seems to remain the old 'a.csv' and not loading new csv file.\r\n\r\nEven worse, after I load a.csv, the load_dataset function keeps loading the 'a.csv' afterward. \r\n\r\nIs this a cache problem?\r\n","comment_length":3,"text":"Keep loading old file even I specify a new file in load_dataset\nI used load a file called 'a.csv' by \r\n```\r\ndataset = load_dataset('csv', data_file='.\/a.csv')\r\n```\r\nAnd after a while, I tried to load another csv called 'b.csv'\r\n```\r\ndataset = load_dataset('csv', data_file='.\/b.csv')\r\n```\r\nHowever, the new dataset seems to remain the old 'a.csv' and not loading new csv file.\r\n\r\nEven worse, after I load a.csv, the load_dataset function keeps loading the 'a.csv' afterward. \r\n\r\nIs this a cache problem?\r\n\nSame here !","embeddings":[-0.1793705374,-0.212596491,-0.1329220086,0.2831755877,-0.0223734658,0.1146043018,0.2654767334,0.227098763,0.5123314261,-0.0145185646,0.0791702792,0.0408157073,0.3277437389,-0.0370748416,0.0149394888,0.3241684735,-0.0145201851,0.2800486386,0.0009385077,-0.0751570538,-0.3693520129,0.0884306878,-0.1289694309,-0.2522624731,-0.1957994103,0.0972640812,-0.0283622127,-0.06998007,-0.0967131779,-0.2095914781,0.341746062,0.3172363639,0.2069521546,0.6984717846,-0.0001106282,-0.0534330755,0.2307249755,-0.0314042382,-0.3263750076,-0.3549534082,-0.0944163203,-0.2298747748,0.1471746862,-0.1890066564,-0.1302862465,-0.0648193806,-0.0166918542,-0.5459915996,0.5006978512,0.1247868314,0.2540838122,-0.290786922,-0.1084601134,-0.18488051,0.1766285002,-0.0817195922,0.1415402293,0.2376575023,0.0883381292,-0.0751221627,-0.1478142589,0.068114616,-0.1948313713,0.228068769,0.2757626772,0.0220453124,-0.0242371615,-0.1163938046,0.2650873959,0.0288163014,0.9887946844,-0.0786269009,-0.238377288,-0.0205829665,-0.0615437366,-0.3468887508,0.1321720183,0.0195248351,0.1573821753,0.0678090677,-0.0254929252,-0.337470293,-0.0751891062,0.1651506126,0.0202255845,-0.2145985961,-0.1684981287,0.1988210529,0.2712448239,0.1297944486,0.364716351,-0.0720537901,-0.1310831755,0.1948052943,-0.5518507957,0.2412742823,-0.1613600403,0.1856755912,0.0256019421,0.0695969686,0.199672401,0.1414411664,-0.0985298157,0.0549721979,0.2573888004,0.192358613,0.3341951668,-0.2608846128,0.188805908,0.0057026767,-0.3730758429,-0.2134575993,-0.0214100163,-0.3076661229,0.3631495237,-0.1151365116,0.2451097518,-0.224628672,-0.0310415998,0.0601802431,0.3276462257,-0.3735513985,-0.1586505324,0.6081885695,-0.1204604805,0.2812158465,-0.0114241876,-0.0087981364,-0.4259694815,-0.031685777,-0.1028014645,-0.1833682209,-0.2406169474,0.1192866787,0.2794251144,-0.161702469,0.2243276983,0.3318047523,-0.1611839384,-0.3115132451,0.2007944137,-0.0963461697,0.1494003087,0.11825528,0.0017515152,0.3521647453,-0.0148449345,-0.245272696,0.0324644409,0.2085768133,-0.6505453587,-0.2651840746,0.1694330275,0.2236861736,0.0689277872,-0.0307433456,-0.2257876098,-0.0404672548,0.1033012718,-0.3534373343,-0.2199032158,-0.1278515309,-0.3009276688,-0.1972607374,0.0229762178,0.6279928088,-0.3274149001,0.0752538368,-0.1144307628,-0.0202842467,-0.266469568,0.0226798225,-0.1796180904,-0.1033601612,-0.2829681039,-0.3777399063,0.351188302,-0.2816778123,-0.4409584701,0.1861273199,-0.0308586601,0.0635635257,0.3096032739,0.2467355579,-0.0204792544,-0.0202055406,-0.122364752,0.0200056657,0.2439660728,-0.064510107,0.0020645333,-0.0475942716,0.1189759299,-0.0135577554,-0.0814068392,0.2503289878,0.2594768703,-0.0323894806,0.2531305254,-0.0059299986,0.0966469869,0.1348436177,0.0900452062,0.1119768471,0.1603908092,0.1263709664,-0.6268913746,0.1083820462,-0.0116276806,-0.516923666,0.1451545507,-0.1914075315,-0.311652422,-0.1958757192,-0.3166609406,-0.1761780977,-0.01444133,0.3911630511,0.0243787654,-0.1369299293,-0.2092683613,0.5341483355,-0.0354862809,-0.1037525758,-0.5580470562,0.0455156453,0.1465581656,0.0266388357,0.041559197,0.0264118314,0.4037293494,-0.1596506536,-0.1407378763,0.3044916689,0.3001393676,0.2979159951,0.0021678905,-0.2528364062,0.0168266222,0.279976666,0.0486450382,-0.1471562535,0.0416384786,-0.1111591533,-0.0674700812,0.2195150256,-0.1230041012,0.2478276789,0.0841375962,-0.3578334749,0.1527132094,-0.261151135,0.3408361077,-0.1783485264,0.1500562429,0.1991871148,0.3342577815,0.2937663496,-0.0938585103,0.0247230642,0.3670998514,0.0316156112,-0.2908042967,0.1147695631,-0.0876018554,-0.0220618993,0.0012844882,0.3062876463,0.3354882896,0.244150728,-0.1981672645,-0.0502814464,0.3757955134,-0.1774961799,0.0324246734,-0.333653003,0.2905320525,0.3118716478,-0.0032714403,-0.1338929832,-0.2291819304,0.315580368,0.1727921516,0.0126362648,-0.172918722,0.1143425629,-0.4884524941,0.149282679,-0.3961778581,0.1384337395,-0.0863995701,-0.0391212255,-0.2855003476,0.6716721058,0.3850256503,0.1235830933,-0.0436288901,0.2157427222,0.0122764828,-0.1798709631,-0.0668858588,-0.1808187217,-0.1703210473,0.0246709418,0.1004073396,-0.0360536166,0.1064485982,-0.0248600841,-0.0441025868,-0.2994722128,-0.036613401,-0.0662790835,0.1577985585,0.2734795511,-0.2116252482,0.401773572,-0.0186835378,-0.1998295486,0.0197594408,-0.4444266856,-0.0657257065,-0.025984725,-0.0201737359,0.1166673452,0.0618403219,-0.6702834368,-0.2097662985,-0.1848932207,0.1760805547,-0.1176161915,-0.0350660197,0.1523957103,-0.0442169905,-0.1741328537,-0.0543771312,0.1851309389,-0.397988081,-0.4589161575,0.3544836044,-0.2061129063,-0.2866055369,0.2730823457,0.0680947453,0.1151947007,0.1106178239,-0.5341038704,-0.1697049737,0.0739730224,0.0643966869,-0.0939659998,-0.0147300716,0.1292114705,-0.1518419385,-0.0748167858,-0.1520914584,-0.1224661022,0.0221890677,0.0605891161,0.3269994557,0.2431199402,0.186248824,-0.4225288332,0.448970288,0.1435923129,-0.1329981983,0.4277791679,-0.1255834997,0.4784880877,-0.3759140372,-0.2353604138,-0.2663930655,-0.2624609172,-0.1712465882,0.1872315556,-0.0229727197,-0.0408082679,-0.269181639,-0.1216465682,-0.1894623488,-0.160956502,-0.0284313839,0.0083869556,0.0411950164,0.3266090453,0.0177001394,-0.1272762418,-0.2020316273,-0.1691849083,0.286295563,0.147511974,-0.07052771,-0.1430947632,0.2092337161,-0.230105862,0.2956239581,0.0350497738,0.0934033543,0.0187298432,0.0240360964,0.0509370416,-0.0122336531,0.9025893211,-0.1311173588,0.0479931794,0.0572382584,-0.0299337916,-0.0908031091,0.1264212132,0.3492653668,0.281450808,0.0121570947,0.0474836789,0.1007569432,-0.275886327,-0.0354216471,0.3341523409,-0.2793852687,-0.4187465906,-0.4742383957,0.0432801247,-0.141318202,-0.3419416249,-0.40737468,-0.0424796976,-0.2451018691,-0.2007101327,-0.1898954362,-0.1993730068,0.0605442524,0.0030431545,0.4117276073,0.3489277363,0.0466124676,0.372361064,0.1550730318,0.3547272682,0.4302995503,-0.2895727158,-0.2089224756,0.1645972133,-0.0981628969,0.2718771398,0.0652754307,-0.1151823327,0.2499686927,0.0048884549,-0.1591000855,-0.0737478212,-0.0479357503,0.1298581362,0.1357139945,-0.362149924,-0.7170673013,0.3703247011,-0.2081222087,-0.1120935827,0.2488385588,-0.4429593384,0.0186635535,0.1479031146,-0.4158057868,0.6920958161,-0.0688928366,0.0610381849,-0.0760192797,0.2097901255,0.3302989602,-0.3021903336,0.1176046506,-0.2044519931,-0.3457727432,-0.0737909377,-0.0548050962,0.1464096755,0.0981911942,-0.2157445401,0.4484638572,-0.0325946249,0.4676469266,-0.1233666539,0.422226429,-0.0209280793,0.0016173715,-0.2271106243,0.1401975751,-0.1062836051,0.1559076458,-0.1311830282,-0.0308973417,0.1637915075,-0.0860407278,0.3028700054,0.090454258,-0.1060748473,0.231130302,-0.340403825,-0.4572926462,-0.1256596446,0.3217206895,0.0986931175,-0.0880957842,-0.1332485676,0.3109615147,0.1383289099,0.1099969894,-0.5000371933,-0.1992897689,0.1221272573,0.0826345906,-0.4267536402,0.1809234172,-0.0885249376,-0.4097435176,0.0901117399,0.3486741781,0.0317329355,-0.3590865433,-0.1414691806,0.1816044748,0.084556371,-0.0705401599,0.1555901766,-0.0939009935,0.0901481658,0.1180701777,0.066804491,-0.47893852,-0.1411468685,0.5442479849,0.2113405615,-0.0978319645,0.5324065685,-0.0851291195,-0.0368154384,-0.2046940029,-0.0777027756,0.2030727863,-0.2233815938,0.0353601128,0.1813427061,0.3800317645,0.3016940355,0.0077700978,-0.1380994618,0.0180242397,0.0668695346,-0.4014623165,-0.1409123987,-0.0797497705,-0.0737801865,0.2326296866,0.2620303333,-0.274679929,0.1767782569,0.3519402146,-0.2701324821,0.1900593042,0.3447101414,0.1813850254,0.2694730759,-0.0677616075,-0.0740870014,-0.2956148684,0.0838871747,-0.050124187,-0.2922175527,-0.2230026424,-0.13644813,0.132920146,-0.1904990971,0.3343132436,-0.3471012712,-0.4828694463,0.1654252112,-0.2810676992,0.2141737193,0.0200010203,0.0612600259,-0.1258493215,0.1787165999,0.0683328658,-0.2495972216,0.149791643,-0.2021393627,0.1734170318,0.1629653573,-0.0620190687,-0.1005951688,0.1676657647,-0.3639301062,-0.1335030198,0.2451533228,-0.087218374,0.0835857093,-0.2801530063,-0.0151271708,0.1933308244,0.3297201097,0.5284406543,0.0120123224,-0.0513794757,0.1551302969,0.2258527428,-0.2194711566,0.1396252811,0.2990403175,0.0525384881,-0.0644639134,0.0880090073,0.3090031743,-0.1998404115,0.2310764641,0.1847522706,0.4833546281,-0.4226143658,0.462013036,0.0486954823,0.2644936442,0.010107927,0.4359619915,0.3735286295,-0.1206561103,0.4654329717,-0.0912106559,0.2112414241,0.1567749381,0.2424978018,-0.4265504777,-0.5828881264,0.1600686908,0.4189822376,-0.3661132157,0.1507469118,0.223598659,0.4335501492,0.0330448635,-0.1679552943,-0.4363111258,0.2211206555,-0.1017370224,-0.274169445,0.119333677,-0.1527307928,0.1651411355,0.1658450663,0.0732964277,0.0540880747,0.2448260635,0.0965289697,-0.1556398124,-0.3576267064,0.1355167925,-0.0195693579,0.165021196,-0.0427978449,0.3513882458,-0.0863206908,-0.0530883409,0.3356804848,0.0136941662,0.0973419473,0.2324913144,0.1008963361,0.1409772784,0.136036694,-0.0787351504,-0.1513858587,0.1279562265,0.0728248358,-0.0032966444,0.1440272927,0.2404723018,-0.1450527161,0.3502406776,0.2253700793,0.1617251188,-0.4983971715,-0.161028564,0.3959455192,-0.0073658554,-0.176418528,0.0359230414,-0.4234867096,0.1281300336,0.3441075683,-0.1938452721,0.1704003811,-0.0549796782,0.0840602219,-0.1097605079,0.3934999406,-0.1047987193,0.0669821054,-0.1977508962,-0.116119653,-0.9518576264,0.1674037874,-0.1299151927,0.2967853546,-0.083235167,-0.0499753281,0.0485114902,-0.1846391857,0.2020845115,-0.2408901006,0.2140237093,0.0128381131,-0.3316783607,-0.1308004856,0.0949476212,0.2393248826,-0.0610905029,-0.2121007293,0.1382859051,-0.1813638806,0.0173474681,0.0151836732,0.0367296897,0.1785529703,0.1447087973,0.6353874207,-0.1135048643,0.3890916407,-0.1927773058,0.2146458775,-0.209312588,-0.3009111583,0.1047234386,-0.0949699581,0.2217062116,0.0689846128,-0.2728091776,0.105206348,-0.2654649317,0.2129324079,-0.0815869272,0.0417031534,0.03066664,-0.1473419666,0.1160811335,-0.0123094479,-0.0254353806,0.3953112662,-0.0373723768,0.28881827,-0.0955616012,-0.0387775674,0.4426199794,-0.3672364652,-0.1745893359,0.0668390989,0.2382866144,-0.0569839589,0.0261656363,-0.1694209576,0.2544820905,0.3998332024,0.0772546977,-0.1040174589,0.3419972956,-0.1977607161,0.0524201766,-0.0202434808,0.3538422585,0.3231263459,-0.3701592386,-0.0068647023,-0.195986107]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/444","title":"Keep loading old file even I specify a new file in load_dataset","comments":"This is the only fix I could come up with without touching the repo's code.\r\n```python\r\nfrom nlp.builder import FORCE_REDOWNLOAD\r\ndataset = load_dataset('csv', data_file='.\/a.csv', download_mode=FORCE_REDOWNLOAD, version='0.0.1')\r\n```\r\nYou'll have to change the version each time you want to load a different csv file.\r\nIf you're willing to add a ```print```, you can go to ```nlp.load``` and add ```print(builder_instance.cache_dir)``` right before the ```return ds``` in the ```load_dataset``` method. It'll print the cache folder, and you'll just have to erase it (and then you won't need the change here above).","body":"I used load a file called 'a.csv' by \r\n```\r\ndataset = load_dataset('csv', data_file='.\/a.csv')\r\n```\r\nAnd after a while, I tried to load another csv called 'b.csv'\r\n```\r\ndataset = load_dataset('csv', data_file='.\/b.csv')\r\n```\r\nHowever, the new dataset seems to remain the old 'a.csv' and not loading new csv file.\r\n\r\nEven worse, after I load a.csv, the load_dataset function keeps loading the 'a.csv' afterward. \r\n\r\nIs this a cache problem?\r\n","comment_length":88,"text":"Keep loading old file even I specify a new file in load_dataset\nI used load a file called 'a.csv' by \r\n```\r\ndataset = load_dataset('csv', data_file='.\/a.csv')\r\n```\r\nAnd after a while, I tried to load another csv called 'b.csv'\r\n```\r\ndataset = load_dataset('csv', data_file='.\/b.csv')\r\n```\r\nHowever, the new dataset seems to remain the old 'a.csv' and not loading new csv file.\r\n\r\nEven worse, after I load a.csv, the load_dataset function keeps loading the 'a.csv' afterward. \r\n\r\nIs this a cache problem?\r\n\nThis is the only fix I could come up with without touching the repo's code.\r\n```python\r\nfrom nlp.builder import FORCE_REDOWNLOAD\r\ndataset = load_dataset('csv', data_file='.\/a.csv', download_mode=FORCE_REDOWNLOAD, version='0.0.1')\r\n```\r\nYou'll have to change the version each time you want to load a different csv file.\r\nIf you're willing to add a ```print```, you can go to ```nlp.load``` and add ```print(builder_instance.cache_dir)``` right before the ```return ds``` in the ```load_dataset``` method. It'll print the cache folder, and you'll just have to erase it (and then you won't need the change here above).","embeddings":[-0.0888758972,0.2823802829,-0.0012339302,0.1619006693,-0.1625473797,0.0777331367,0.1432165802,0.2737467587,0.3626849055,-0.113328442,0.1163817868,0.189830035,0.1802767515,-0.0821299329,0.1950766742,0.3331559002,-0.0630543754,0.5072548389,0.1366073638,-0.0146824615,-0.3488639891,0.1107328013,-0.0943152234,-0.07250195,-0.1926436722,0.1469480544,-0.0438434035,0.13965334,-0.2056970298,-0.4888148606,0.3809181452,0.3573814631,0.2581088245,0.4660643637,-0.0001152982,-0.0119408052,0.2476485223,0.0682446957,-0.5148969293,-0.3869971335,0.0314668678,-0.5372769833,0.3891557753,-0.1841786206,-0.1016150042,-0.0859777629,0.1840104908,-0.0501311496,0.5789194703,0.2138358355,0.195167318,-0.1492288709,0.0171368178,-0.0130087351,0.1113954931,-0.2091115862,0.1625352651,0.353564024,0.1116896793,-0.2694231272,-0.2668689489,0.0342900194,-0.2663342357,0.2355277091,0.2771705687,0.0081724627,0.2694102824,-0.0516900495,0.0297865998,-0.0389136113,0.66036129,-0.0337102041,-0.1231487319,-0.2626151741,0.0286018085,-0.5128679276,0.1346926093,0.1874417514,-0.0809473768,-0.0461172797,-0.2042680085,-0.4057559073,-0.0188115314,0.2534582317,0.1112423763,0.123455584,-0.0647201687,0.228922829,0.2850901783,0.1747080088,0.2771388292,-0.1166692227,-0.2123604566,0.2447838634,-0.4885699153,-0.0201186202,0.1123057678,0.1282074004,-0.1136159897,0.1372920871,0.2320876122,0.0109714754,-0.1496887505,0.1623385996,0.3279623091,0.3042567372,0.2717228532,-0.1469151378,0.1787361205,0.158106178,-0.3821918964,-0.094241932,-0.0421042815,-0.2711846828,0.2378876805,-0.1810322255,0.3370941877,-0.1668266654,-0.0508520752,0.0400000699,0.1158589423,-0.2263805419,0.0172547791,0.4845026135,-0.1868886054,0.1356288791,0.1583126336,0.0463617519,-0.5059089065,-0.1612841636,-0.0898496956,-0.1798771769,-0.3385742307,0.0726501048,0.4552691579,-0.2072428316,0.3022795916,0.3358885646,-0.2005399913,-0.3571424782,0.074267894,0.1200318113,-0.0250850767,0.0515104644,0.0902633667,0.3180943727,0.024142446,-0.5446966887,-0.173861593,0.1558584124,-0.5722796321,-0.3847288191,0.0376878008,0.177387774,-0.150834173,-0.1242483631,-0.0813765153,-0.0549606383,0.152668938,-0.4405647218,-0.2240802795,-0.1603152603,-0.230242759,-0.2056117654,-0.0603124611,0.5919473171,-0.107986778,-0.0069066458,-0.1732042134,0.1078955978,-0.1307269037,0.1100469753,-0.1966307312,-0.0026769985,-0.2292711586,-0.5279220343,0.6706701517,-0.0909134001,-0.482563138,0.2108750939,-0.1627443135,0.1470751166,0.1583846807,0.2092542499,-0.166003868,-0.1586555839,0.0442557707,0.2686133683,0.1136607304,0.0111965248,-0.1344201863,-0.2561934888,0.2314466238,0.0314552635,-0.1690050215,0.2235095501,-0.024848463,0.13542521,0.273614049,0.0281254537,0.037286479,-0.0211752299,-0.0188249201,0.0938779712,-0.0387785695,0.2074213624,-0.6352767944,0.2108055055,-0.1857466847,-0.3761024773,0.1938500106,-0.0648839474,-0.2942895591,-0.2255121917,-0.4506825805,-0.2055748105,0.018885212,0.3500680923,0.0924992934,-0.1076359674,-0.1527515501,0.3722836077,-0.07438232,-0.0678525418,-0.5102145672,0.1139070317,0.0197452512,-0.0636630729,0.113454327,0.2977822125,0.3451760709,-0.0230163988,-0.1121129692,0.2423195541,-0.0802451372,0.3043351471,0.0062468676,-0.3474616408,0.0004544011,0.2994796336,0.1332806349,0.1017207205,-0.0083979983,-0.0168770626,-0.0799977407,0.3322361112,-0.0560341217,0.3085152805,0.0633942559,-0.1673322618,0.2275238931,-0.3928000331,0.2067195475,-0.1777364165,0.2628163099,0.2479507774,0.1953124255,0.2646502554,-0.0829107538,0.0650189519,0.4671657681,-0.0469818301,-0.2923628986,0.1642831862,-0.0841177851,-0.1841216385,-0.1236988753,0.3632374108,0.4298818707,0.1477506459,-0.1818337142,0.0822671801,-0.0667831078,-0.2304429114,0.1101651788,-0.2013234943,0.2853393853,0.2399768084,0.2579445541,-0.0999965593,-0.3501327932,0.3381066918,0.2240700126,0.110327065,-0.0796889514,0.2068432868,-0.5125057697,-0.1935468465,-0.5543268919,0.0061852699,-0.2469133139,-0.0969223008,-0.0754553229,0.6249861717,0.2246181369,0.1039894149,0.0526425838,0.3406598568,-0.1474728882,-0.3065598607,0.0291031431,-0.3341251016,-0.3812094033,-0.0456452146,0.2931637168,-0.0742159411,0.2642597258,0.0402296893,-0.1407391429,-0.2185639888,-0.1337854117,-0.0987230465,0.1271894276,0.2915703356,-0.1833294779,0.1628341079,0.138350606,-0.2592160702,0.1522710025,-0.4241508245,-0.0251136199,0.172117725,-0.0859804153,0.1537035853,-0.0118244858,-0.422603339,-0.2225028872,-0.2736027241,0.1390750706,0.1597259045,0.151983887,0.3598029017,-0.108785063,-0.1524826139,-0.1071985587,0.2338398695,-0.3367555141,-0.2792285085,0.2858326435,-0.1888992041,-0.2545824349,0.250490427,-0.0153042125,0.2039731592,-0.1013497114,-0.6121990085,-0.07070259,-0.009309873,-0.0277735572,0.0958586261,0.0553385541,0.1928489059,-0.1464696229,-0.0504265502,-0.039213635,-0.1802838445,-0.0088396231,0.2555666864,0.4061007798,0.2051307261,0.1547252089,-0.2963939011,0.5076721311,0.399600178,-0.1259077191,0.410513401,-0.0076688351,0.4529300928,-0.3399046659,-0.1743122488,-0.1087409109,-0.1506913602,-0.0775153115,0.1882668585,0.0459606387,-0.1770215333,-0.2123094052,-0.0854604393,-0.0428779796,-0.3324961662,0.1021435261,-0.1324265152,0.0285492539,0.4036848545,-0.2121608257,-0.2638756037,-0.4108496308,-0.3645507395,0.3081257641,0.0857929811,0.0269671772,-0.064481169,-0.0304787401,-0.3124123216,0.210033834,-0.0334742554,0.1108683646,0.1170168743,0.0738254339,0.2352378666,0.0246181954,0.60765481,-0.0361419134,0.1442110986,0.1510512233,0.056335628,-0.1386025548,-0.0542592369,0.2957320511,0.4776977003,0.2010528445,0.0684841648,-0.0150319878,-0.4613938332,0.1272092909,0.3126029968,-0.3689418137,-0.4041787088,-0.4548129141,0.1523551047,-0.3105368316,-0.2186832726,-0.453633368,0.1892790198,-0.2344124168,-0.1143863797,-0.1408667564,0.0339928754,-0.0269311611,0.0028769223,0.3091115654,0.4205068052,-0.0521419421,0.1925815791,0.2409769595,-0.1799023151,0.4959931374,-0.0364597179,0.0043456103,0.0109083252,0.0607550442,0.2748132348,0.2771266401,0.0029778304,0.153257668,-0.0888366029,-0.0954787508,-0.1706480384,-0.1272304207,0.1746071726,-0.0192987844,-0.2430473417,-0.624779284,0.5867572427,-0.228198722,-0.0958015025,0.0771201849,-0.429784447,0.0043202909,0.3284889162,-0.3328558207,0.7670230865,0.0136230066,0.2584234476,0.068875879,0.085794352,0.7262017727,-0.3534884453,0.1585288048,-0.2406385839,0.0268778522,0.0671018735,-0.0156931728,-0.0373694561,0.0611010268,-0.2364809811,0.5385795832,0.1941304505,0.2164651304,0.1895298213,0.5119624138,-0.0139746116,-0.1716061234,-0.388681531,0.0655357838,-0.1131620482,0.2626148462,-0.1508743465,-0.1621663719,0.321356684,-0.078524448,0.1230568215,0.2460944951,0.0687521696,0.2204843909,-0.1085436717,-0.3991333842,-0.0335055105,0.1864880025,-0.0318296328,-0.0040573035,-0.2104968578,0.2432615459,0.0758459195,0.1048815325,-0.2682752013,-0.0988501385,0.3014588058,0.0377515554,-0.3364536762,0.111760743,-0.1059244201,-0.4096359313,0.1102480367,0.3318060637,0.0765402317,-0.1532813162,-0.2351976782,0.1930825859,0.0814829022,-0.0987384617,0.124410212,0.0387887582,-0.0530340783,-0.04135805,0.0260921083,-0.4763903618,0.0609044991,0.3872442245,0.2643859684,-0.0910675079,0.6199370623,-0.0544082075,0.1009747982,-0.1384754926,-0.1553155482,0.0039826552,-0.3479216099,0.073509708,0.3034491539,0.2219399959,0.3384997547,0.0701493025,-0.1725638509,0.1892220676,0.0229783095,-0.4424035847,-0.2097513527,-0.18008627,-0.2489678413,0.1425334662,0.3069705367,-0.0182875898,0.1766197085,0.422170341,-0.2422644496,0.2179239541,0.036387641,0.2837193906,0.1794681847,-0.1455035508,-0.2760009766,-0.223844707,-0.0397574417,-0.1276075691,-0.3365162611,-0.2114665806,-0.2694696188,0.1738516092,0.0560821369,0.2405853122,-0.348625809,-0.3961256444,0.2473444343,-0.1670863181,0.1138704196,-0.255615741,0.051993724,0.0567375422,0.0273377206,0.1373144388,-0.1126865447,0.1204753667,-0.2819398344,0.3032253385,0.1721111089,0.0267839711,-0.2301598936,0.2022953779,-0.462767005,-0.1068911105,-0.0496713594,0.0468472354,0.1288542897,-0.3395739198,-0.0416400209,0.0133723617,0.1993759573,0.5579795241,0.0093691256,-0.0509724133,0.2949975133,0.1351966113,-0.2785677314,0.1379586905,0.4374043941,0.0619372353,-0.048947867,0.1274027079,0.2259240597,-0.2998074889,0.2056145966,0.2648714781,0.3571155965,-0.4329585135,0.3450414836,0.1510655284,0.2942136228,-0.0452230759,0.4577881396,0.2673683763,-0.0011886609,0.3178556263,-0.115100801,-0.0553442128,0.0897959992,0.2091727257,-0.2968596518,-0.3625848889,0.1096713394,0.5631207228,-0.0496521108,0.1745132208,0.1760571599,0.5881574154,-0.0891447514,-0.1031950638,-0.4032124877,0.4081043303,-0.1066976115,-0.2956745028,-0.1100805327,-0.0177412722,-0.1267818511,0.2605084777,-0.0965156406,0.0007644144,0.054717496,0.1165261343,-0.2532088161,-0.6731388569,0.1490396112,0.0911904648,0.0718467683,-0.1221193969,0.3966829181,-0.1008651778,0.0015583939,0.1063883901,0.1515479088,0.2985385358,0.2616482377,-0.0261190105,0.2150194347,0.0984299704,-0.0990653783,-0.2725051939,0.1522102058,0.1588630974,0.0735249445,0.073724702,0.127131179,-0.1456317604,0.1910413951,0.2808283865,0.0177976172,-0.4239731729,0.2723133266,0.4910109639,-0.1004145592,-0.1347147226,0.1501859576,-0.5083786845,0.0870216712,0.2666011751,-0.0274718776,0.1571957916,-0.2614791393,0.0647617429,-0.0993392989,0.4398812354,-0.0613911673,0.0301488657,-0.2412988544,-0.21176669,-1.0339400768,0.1108426973,-0.0859863907,0.0675170794,0.0298559498,0.1638640016,-0.1324921101,-0.0199047979,-0.0333389863,-0.0068990323,0.2087882906,-0.0633899793,-0.2428737581,-0.1244172677,0.1882929951,0.1633123159,-0.1582173854,-0.1272715032,0.0362091586,-0.2769181728,-0.0169722792,0.0682015046,0.0450558364,0.2270138711,0.3052161038,0.4869482517,-0.0479958095,0.1529428512,-0.0526885353,0.1403115243,-0.401856035,-0.1542679965,-0.1185613349,-0.2395340651,0.0146923484,0.209272787,-0.3926819563,0.3989510536,-0.3408300877,0.0581692383,-0.1670539081,-0.1050884351,0.0883125514,-0.1183825508,0.0023446565,0.3344597816,0.0544686541,0.5390545726,-0.0718650371,0.2023704052,-0.3011860251,0.0192575287,0.3384816647,-0.4451626241,-0.2249624729,0.1939318925,0.3209458888,-0.0809384361,-0.0111825466,-0.2096023262,0.1600855142,0.2623906732,0.0497747958,-0.1716387421,0.3985253572,-0.1940024644,0.0339898318,-0.0756102353,0.2051556259,0.4144532084,-0.4878883362,0.0202204511,-0.2616923153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/443","title":"Cannot unpickle saved .pt dataset with torch.save()\/load()","comments":"This seems to be fixed in a non-released version. \r\n\r\nInstalling nlp from source\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp\r\ncd nlp\r\npip install .\r\n```\r\nsolves the issue. ","body":"Saving a formatted torch dataset to file using `torch.save()`. Loading the same file fails during unpickling:\r\n\r\n```python\r\n>>> import torch\r\n>>> import nlp\r\n\r\n>>> squad = nlp.load_dataset(\"squad.py\", split=\"train\")\r\n>>> squad\r\nDataset(features: {'source_text': Value(dtype='string', id=None), 'target_text': Value(dtype='string', id=None)}, num_rows: 87599)\r\n>>> squad = squad.map(create_features, batched=True)\r\n>>> squad.set_format(type=\"torch\", columns=[\"source_ids\", \"target_ids\", \"attention_mask\"])\r\n>>> torch.save(squad, \"squad.pt\")\r\n\r\n>>> squad_pt = torch.load(\"squad.pt\")\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/torch\/serialization.py\", line 593, in load\r\n return _legacy_load(opened_file, map_location, pickle_module, **pickle_load_args)\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/torch\/serialization.py\", line 773, in _legacy_load\r\n result = unpickler.load()\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/splits.py\", line 493, in __setitem__\r\n raise ValueError(\"Cannot add elem. Use .add() instead.\")\r\nValueError: Cannot add elem. Use .add() instead.\r\n```\r\nwhere `create_features` is a function that tokenizes the data using `batch_encode_plus` and returns a Dict with `input_ids`, `target_ids` and `attention_mask`. \r\n```python\r\ndef create_features(batch):\r\n source_text_encoding = tokenizer.batch_encode_plus(\r\n batch[\"source_text\"],\r\n max_length=max_source_length,\r\n pad_to_max_length=True,\r\n truncation=True)\r\n\r\n target_text_encoding = tokenizer.batch_encode_plus(\r\n batch[\"target_text\"],\r\n max_length=max_target_length,\r\n pad_to_max_length=True,\r\n truncation=True)\r\n\r\n features = {\r\n \"source_ids\": source_text_encoding[\"input_ids\"],\r\n \"target_ids\": target_text_encoding[\"input_ids\"],\r\n \"attention_mask\": source_text_encoding[\"attention_mask\"]\r\n }\r\n\r\n return features\r\n```\r\n\r\nI found a similar issue in [issue 5267 in the huggingface\/transformers repo](https:\/\/github.com\/huggingface\/transformers\/issues\/5267) which was solved by downgrading to `nlp==0.2.0`. That did not solve this problem, however. ","comment_length":26,"text":"Cannot unpickle saved .pt dataset with torch.save()\/load()\nSaving a formatted torch dataset to file using `torch.save()`. Loading the same file fails during unpickling:\r\n\r\n```python\r\n>>> import torch\r\n>>> import nlp\r\n\r\n>>> squad = nlp.load_dataset(\"squad.py\", split=\"train\")\r\n>>> squad\r\nDataset(features: {'source_text': Value(dtype='string', id=None), 'target_text': Value(dtype='string', id=None)}, num_rows: 87599)\r\n>>> squad = squad.map(create_features, batched=True)\r\n>>> squad.set_format(type=\"torch\", columns=[\"source_ids\", \"target_ids\", \"attention_mask\"])\r\n>>> torch.save(squad, \"squad.pt\")\r\n\r\n>>> squad_pt = torch.load(\"squad.pt\")\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/torch\/serialization.py\", line 593, in load\r\n return _legacy_load(opened_file, map_location, pickle_module, **pickle_load_args)\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/torch\/serialization.py\", line 773, in _legacy_load\r\n result = unpickler.load()\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/splits.py\", line 493, in __setitem__\r\n raise ValueError(\"Cannot add elem. Use .add() instead.\")\r\nValueError: Cannot add elem. Use .add() instead.\r\n```\r\nwhere `create_features` is a function that tokenizes the data using `batch_encode_plus` and returns a Dict with `input_ids`, `target_ids` and `attention_mask`. \r\n```python\r\ndef create_features(batch):\r\n source_text_encoding = tokenizer.batch_encode_plus(\r\n batch[\"source_text\"],\r\n max_length=max_source_length,\r\n pad_to_max_length=True,\r\n truncation=True)\r\n\r\n target_text_encoding = tokenizer.batch_encode_plus(\r\n batch[\"target_text\"],\r\n max_length=max_target_length,\r\n pad_to_max_length=True,\r\n truncation=True)\r\n\r\n features = {\r\n \"source_ids\": source_text_encoding[\"input_ids\"],\r\n \"target_ids\": target_text_encoding[\"input_ids\"],\r\n \"attention_mask\": source_text_encoding[\"attention_mask\"]\r\n }\r\n\r\n return features\r\n```\r\n\r\nI found a similar issue in [issue 5267 in the huggingface\/transformers repo](https:\/\/github.com\/huggingface\/transformers\/issues\/5267) which was solved by downgrading to `nlp==0.2.0`. That did not solve this problem, however. \nThis seems to be fixed in a non-released version. \r\n\r\nInstalling nlp from source\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp\r\ncd nlp\r\npip install .\r\n```\r\nsolves the issue. ","embeddings":[-0.138673082,-0.184037298,0.1423232853,0.2854889631,0.3872589171,0.073986955,0.4242343307,0.1544946581,-0.1818623096,0.0088987788,-0.2541096807,0.7546944022,-0.3429293633,-0.1627208889,-0.0087755695,-0.2707043588,-0.0224570408,-0.0999822319,0.1635401398,0.0048668031,-0.1239810959,0.171049282,-0.2978944182,-0.0645304099,-0.2080973536,-0.2048983574,-0.00159157,0.3980633318,0.0651103556,-0.3406357765,0.1945503354,-0.1563013941,0.0915548205,0.4689872861,-0.0001217698,-0.0382243693,0.1266155988,-0.2294037938,-0.4221751392,-0.1044335216,0.4167321622,-0.1648723036,0.2794097066,-0.3884229064,-0.0467770696,0.1305866987,0.0142891519,-0.3021172285,0.1780150235,0.0991781801,0.1452196389,0.6736469269,0.0289200787,-0.0280884672,0.2441470474,0.1641828418,-0.2486750931,0.2910862863,0.1041378975,0.0843061432,-0.1918402463,0.0632601455,-0.3168004453,-0.1475128382,0.1134289578,-0.0622912869,-0.675268054,-0.2057644874,0.2055133134,0.1767506003,0.270327121,-0.3025207818,-0.3350048363,-0.1799490452,0.1947433501,-0.3114592433,0.1561120898,0.3004496992,-0.1158854887,-0.0864392594,-0.1456921399,-0.0299980491,0.0258899014,0.2111272216,0.4304872155,-0.0353298672,-0.0998982415,0.3880563378,0.0998012945,-0.0481595919,0.0078423256,-0.1047171652,0.1041183323,0.0092728334,-0.2710994482,0.0964374021,-0.036419794,-0.069630228,-0.0793859959,0.3811096847,0.5383771062,0.1145914942,0.2311057746,0.2259292006,0.2127767354,0.0571255423,0.2554359734,0.345038414,0.0177330505,-0.3164455295,-0.0845254287,0.0598807707,0.1468535215,-0.2113463879,0.1873695552,0.0921167806,0.0115662497,-0.0737667009,-0.26115641,-0.0732759833,-0.1164809167,-0.2230283469,-0.0387124568,0.1521583796,-0.0074682552,0.0885713175,0.0037150709,0.3209428191,-0.3440079391,-0.0227742791,-0.0524146669,0.0857812166,-0.1722242236,-0.053886909,0.3327815235,-0.4329432845,0.2967667282,-0.0294859093,-0.1854732782,-0.0706741363,-0.0606143512,-0.2570724189,0.1994284689,-0.1585468948,0.2861680984,0.1599562466,0.107398048,0.0266728699,-0.0570053495,0.127826333,-0.2086867541,-0.2174158841,-0.2217054367,0.0641504154,-0.1422485262,-0.1611874253,-0.1856589466,-0.0741371512,0.1521243453,-0.1553201973,0.1956332773,-0.1929909736,-0.1605002582,-0.3936680555,0.0562315956,0.386901021,-0.2387972623,-0.1330624223,0.4711282253,-0.1688480377,0.182765767,0.2557338178,-0.2038462907,0.2076244503,-0.2920520306,0.1300198436,0.5466675162,-0.4901282489,-0.2935793996,0.1681383252,0.0104873283,0.0890493691,-0.0081727197,0.1268035918,0.0117182015,-0.030325504,0.1082632244,0.3592544198,-0.0240153167,-0.1920452118,-0.3910906911,-0.0672138035,0.5623107553,0.1039690971,-0.0450830124,0.0051229019,0.1489918679,0.1643618792,0.2806615531,-0.1335194856,0.0790414959,0.2275406122,0.0479265004,-0.0479298718,-0.0614272654,-0.1117487922,-0.7583842874,0.1425738186,-0.4217596352,-0.2041576058,0.0460046083,0.0194343757,-0.009685711,-0.2699629962,-0.4672890007,-0.2535904944,-0.0701944456,0.0350564234,-0.0027187457,-0.1018663868,-0.288267374,-0.058699578,0.017073296,0.0949506313,-0.4358216524,0.0351113565,-0.1008365378,-0.6016008854,0.0788973644,0.3346041143,0.1992215812,-0.213537246,-0.1547768265,0.5112801194,-0.18784751,0.0458433293,-0.0911479741,-0.3962245286,-0.0257686339,-0.3873724043,-0.1387775093,0.3057427704,0.1651026756,-0.1439479887,-0.4406921268,0.4739771783,-0.1370309144,-0.0694071278,0.1428314,-0.1771287918,-0.0402116552,-0.1298947334,-0.0850231126,-0.007480979,0.0441753715,0.2500829995,0.4493282437,-0.0599434823,-0.4387768209,-0.0666086078,0.2690041065,0.1827788055,0.1510511935,0.0907959491,-0.086550504,-0.1462569535,-0.0994369462,-0.5965250134,0.4699759185,0.0079564676,-0.0964354649,-0.0773334578,0.1617991924,-0.1623529196,0.0246821791,0.0292417072,0.0215613749,0.5249430537,0.3105398715,0.1777420938,-0.3436542451,0.0938974395,-0.0082977777,-0.0154359266,-0.2287285775,0.2008437216,-0.2161109895,-0.0077682058,-0.2932492495,0.0225316659,-0.413454473,-0.2659108341,-0.0204125382,0.4188207984,0.1691660434,0.2131918222,0.0327040926,0.0886699483,0.1502593905,-0.2873945534,0.0878560841,-0.1656799167,-0.1369889975,-0.0366177447,0.0988967717,-0.1391302794,0.542593658,0.2488512248,0.0985452309,-0.2540007532,0.1322928369,0.198636964,0.0405147187,0.0283316672,0.158302933,0.0773255229,-0.2481870353,-0.2664538622,0.2911869287,-0.1191455722,-0.1861178726,0.0801180676,0.0109007526,-0.022456577,-0.0050233896,-0.6097726822,-0.224684,-0.2241446525,0.2270322591,-0.0491173156,0.0815528184,0.1741752177,0.004866139,0.2601941228,0.0624346733,-0.1001800075,-0.0126522649,-0.2533332407,0.2158102095,-0.1460841894,-0.4376562238,0.0273987018,0.0929994136,-0.1016625315,0.114223443,-0.0406717472,0.0233942363,0.1410516948,0.2924295068,-0.2655847073,0.1149229184,0.3443350196,0.0430167839,0.0778845921,-0.0306632072,-0.2457076609,0.0596966222,0.1554536372,0.3513274193,-0.0206241515,0.0392360799,0.1629590541,0.861915648,0.1855268627,-0.2794865966,0.3941228688,-0.0292289648,0.2752987444,-0.2377856523,-0.4469310939,0.4460179508,-0.1733710915,0.3204520941,0.0607319698,-0.2573391199,-0.0113226445,-0.0635862499,0.2717902958,-0.149232462,-0.2496081144,0.2806574404,-0.2385175377,0.1049523726,-0.1814576238,0.2275058478,-0.0711084604,-0.3509435952,-0.0032614963,0.1643670052,0.4109746814,0.1616709679,-0.4156162441,0.0040468061,-0.4042151868,0.3957034349,0.1615813524,0.4010522366,-0.0990785956,-0.2193687409,0.1332556605,0.118099466,0.4137420356,0.0096635763,-0.0607225001,0.3061230481,-0.1628252119,-0.5174528956,0.0127819031,0.0220160261,0.4153403342,0.1618519425,0.2548152506,-0.1388515085,-0.1510863304,0.1635999978,0.3699416518,-0.1934055537,-0.1140023693,-0.0574990474,-0.142900452,-0.4427398443,-0.061674431,-0.0853107497,0.2670448124,-0.0017044708,0.2492301315,-0.167213127,-0.2109501213,0.0936309174,-0.3807398677,0.2851109803,-0.0029859515,0.0652357563,0.0234203711,0.4080882668,0.0373503603,0.2515792549,-0.2945907116,0.2742486596,0.1988430172,-0.0704244301,0.1691569537,0.2372219115,0.0521402396,0.1120989099,-0.3705205321,-0.0667927936,0.1262665689,-0.0717885792,0.0833370984,-0.3304269314,-0.2719463706,-0.0292310659,0.232157886,0.0300724432,0.2362706512,0.1166316792,-0.5146976709,-0.0924147293,0.5315946341,0.0563528575,0.9179935455,-0.1772387028,0.3260867894,-0.0321265645,0.1401221454,0.2217609137,0.0599846058,0.1831368506,-0.5430380702,0.1856843084,0.0739597753,-0.0524221361,0.0743605867,0.1941403896,-0.1064019874,0.3549281657,-0.21080634,0.0285267122,-0.1503780931,0.329659909,0.0767176822,-0.235246852,-0.2108256072,0.0077417726,-0.2141858041,0.4785354733,0.109730497,-0.0900828689,0.0994526148,-0.1806674302,-0.0863746703,0.1381822079,-0.1973132342,0.5574913621,0.1291615367,-0.8780997992,-0.0591266416,0.6718172431,-0.0118320268,-0.2320770323,0.0549747683,0.1600875407,0.2959391773,-0.1229823679,-0.2097715288,0.119007051,0.4859683514,-0.0989251882,0.0646486357,0.177004829,-0.1137921289,-0.3283302784,-0.1556128711,0.2993595004,0.3328757584,-0.4707868397,-0.4273250997,-0.220424667,-0.2364569157,-0.151019901,-0.0215645973,0.0950371251,0.1787849516,0.1700586826,0.1484197825,-0.1462188065,0.0156763773,0.8382951021,-0.234483242,-0.3432264626,0.5526185632,0.2660033107,-0.136707142,-0.2085972279,0.1926613599,0.0007457353,-0.3280690908,0.1250885427,-0.0769784898,0.3351719379,0.1462395042,0.2212705314,-0.0975729898,-0.0387975909,0.0120826075,-0.0563085824,-0.4843292236,-0.184666127,0.1859751791,0.0648798347,0.0486844815,0.4098499715,0.3081647456,0.0342079252,-0.1843404025,0.3117743134,-0.0307664555,0.2121686041,0.4773930311,-0.135863632,-0.0903611109,-0.0063377069,0.0305294693,0.0770376399,-0.1119336039,-0.1447269022,-0.2219722569,0.1734952927,0.1860111207,-0.2156836092,-0.2647044063,-0.0095412731,0.1922622323,-0.2260737121,0.1810259819,0.3889671862,0.0224126503,0.1593828946,-0.2167812735,0.2523555458,-0.0527840108,-0.0752296895,-0.1053138301,0.1062319949,0.1237959936,-0.0736687109,0.028034227,-0.0351495743,-0.2667632401,-0.1414224505,0.4792152345,0.2084566355,0.0186066218,-0.383841455,-0.0946761668,-0.3431570828,0.0114340959,0.1630901098,-0.2923072875,-0.1363119632,0.1053908989,0.0154874483,0.0025632419,0.0211565308,0.2502375841,0.0474839285,-0.0992149711,0.090593785,-0.0050882013,0.090545401,-0.2394490689,0.0969281793,0.2744243145,-0.3339277208,0.1398563236,0.1069499031,0.086537458,-0.0021410773,0.1351284832,-0.1257360429,0.2360789776,0.1482811123,-0.0117649511,0.678045094,0.1148011014,0.0790791065,0.2390661836,-0.4852781892,-0.2917052209,-0.067266874,0.2548889518,0.3690403104,0.0166792665,0.8710457683,-0.4937476218,-0.4484514892,0.2343614399,0.5710667372,-0.1630049795,-0.2334986478,-0.1191929355,-0.056972418,0.0392917842,0.2150210738,-0.4183959961,0.1941185296,0.204061836,0.231021449,-0.2608084679,-0.1571810395,0.0455937944,0.0524479076,0.0539755896,0.0109282546,-0.074902758,-0.0152895041,-0.003812019,-0.042199742,0.2800779939,0.2520420551,0.2627357841,-0.2105847299,0.0034328122,0.0593666248,0.1137889326,0.0121133905,-0.0418465473,-0.0237418748,-0.2986751199,0.3994519413,0.0088036871,-0.0923211277,0.0307796076,0.2637254596,0.1043353602,-0.1144671291,-0.0061398167,-0.0696770027,-0.3914432526,-0.2446207702,-0.0146359541,-0.0980908573,0.0952305347,0.2605697513,0.4064261317,0.1591134667,-0.0513001531,0.0466476269,0.0826473981,0.5026730299,0.1973201931,-0.435013622,-0.2800347805,0.1274704188,-0.4370206296,0.0792033374,0.1688423157,0.1076085269,-0.0158124641,0.2462878972,-0.0497196838,0.1373989284,-0.0568726771,-0.0229412131,0.4012254179,0.3032230139,-0.3561688066,-0.0688746497,-0.1765711606,0.1089239568,0.1225910559,-0.3322435617,-0.0499976315,0.1990323514,0.0644451827,-0.3312883973,-0.5067588687,0.0539671853,0.2614976764,0.2660625279,-0.0659694299,0.3709565401,-0.043962542,-0.326020956,0.0345068686,-0.1666181833,0.1727516651,0.1132428795,0.1573904157,0.4533771276,0.0161142256,0.1369044185,-0.4455273747,0.5350661874,-0.051947996,0.1240373626,-0.1002100259,0.1667048782,0.077030167,0.2301463932,0.0832030848,0.5846363306,-0.1828308553,0.3271812201,-0.6916027665,-0.322006166,0.5005269647,-0.2176014632,-0.1631496251,0.0246391091,0.1323576123,0.2069170922,0.1832203716,-0.6202585697,-0.0762675181,0.2132002264,-0.2426442802,-0.1637170166,-0.0185505655,-0.3184702396,-0.0223405343,0.0580590852,0.3536970615,0.2758470774,-0.0369552597,0.0476075299,-0.1522230655]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/439","title":"Issues: Adding a FAISS or Elastic Search index to a Dataset","comments":"`DPRContextEncoder` and `DPRContextEncoderTokenizer` will be available in the next release of `transformers`.\r\n\r\nRight now you can experiment with it by installing `transformers` from the master branch.\r\nYou can also check the docs of DPR [here](https:\/\/huggingface.co\/transformers\/master\/model_doc\/dpr.html).\r\n\r\nMoreover all the indexing features will also be available in the next release of `nlp`.","body":"It seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?","comment_length":50,"text":"Issues: Adding a FAISS or Elastic Search index to a Dataset\nIt seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?\n`DPRContextEncoder` and `DPRContextEncoderTokenizer` will be available in the next release of `transformers`.\r\n\r\nRight now you can experiment with it by installing `transformers` from the master branch.\r\nYou can also check the docs of DPR [here](https:\/\/huggingface.co\/transformers\/master\/model_doc\/dpr.html).\r\n\r\nMoreover all the indexing features will also be available in the next release of `nlp`.","embeddings":[0.0365026444,-0.1409246475,-0.0793723911,-0.1369627267,-0.1559900939,-0.3242330253,-0.1591320634,0.182186991,-0.3461941779,0.015834054,-0.0663877428,0.3206751347,-0.1125187874,-0.0442969948,0.0447908007,-0.2458745092,0.2475958169,0.41393134,0.1482364088,-0.1551828384,-0.0532758646,0.2191117406,-0.2554685771,0.0467051938,-0.4506125152,0.1217017248,-0.4123180807,-0.2625599802,-0.153648138,-0.5197570324,0.4782420397,0.1243595406,0.267839849,0.0743239671,-0.0001179493,-0.0224405769,0.5727151632,0.0642087758,-0.109931156,0.1077400967,0.0450584255,-0.0126223117,0.5089781284,-0.0682091266,-0.0491528623,-0.3134188652,0.2057967186,0.506159544,0.3378701508,0.0675188676,0.1206033751,-0.0178223103,0.3498007655,-0.0210936479,0.4944708645,0.1088032499,-0.1879488379,-0.060293179,0.4111689627,0.0164823085,-0.0369286165,0.0635463297,0.1623755395,-0.2319170237,0.1169097349,0.2194074392,0.1412903219,-0.3700337708,-0.0103806946,0.1250045151,0.2487776577,-0.1683013588,-0.4690059423,-0.1358831078,0.0131439865,-0.3554195464,0.3728207052,-0.2848743498,-0.1211849079,0.1636741906,-0.0299982075,-0.2509265244,-0.1611312628,0.3247069716,-0.1813227534,0.5299859047,0.0605144016,-0.2540729642,0.2310189754,-0.2415739745,0.1807152778,0.1000634655,0.0172251817,0.3674331605,-0.1381929964,-0.2774769664,0.2277394235,0.1133941337,-0.1607001722,0.0322502255,-0.40910694,0.1459733844,0.1459764689,0.0879990384,-0.1691190153,0.2571031451,0.108583197,-0.0899450853,0.2458999306,-0.4051271975,0.1125740632,0.1378308535,0.0466918498,0.162292853,-0.3634460568,-0.3564530611,-0.1548929811,-0.1499220878,0.0551367179,-0.2625108063,-0.2247933596,-0.3351933956,-0.1180677488,0.297842443,0.1253295094,-0.3970762193,0.0724128932,0.1841661036,-0.1341165602,0.0026780257,-0.0640867129,-0.1569979489,-0.1755444109,0.264549017,0.2724645436,-0.3028456867,0.1628210098,-0.1974896938,-0.2274571806,0.2945098579,-0.2795906067,0.12166325,-0.0080272518,0.2258044034,-0.1684639603,-0.1023586318,-0.4655607045,-0.1620360315,-0.2403615564,0.0633199513,-0.0646843016,-0.2811254859,-0.3323854208,0.1153732687,-0.2723724544,-0.2752392888,0.3653652966,0.6273083091,-0.1054331884,-0.1463750899,0.037082769,0.1117174402,0.2649072707,-0.0908317789,0.1279012114,-0.0669749528,-0.2392733842,-0.3873827755,-0.1192891672,0.1379834414,0.0270555951,0.2212779969,-0.2794346213,0.2451289445,0.0291392673,0.5474945307,0.6518902183,-0.2099924684,-0.1275744587,0.0607614219,0.1220447719,-0.2443282157,-0.0022637483,-0.1838946193,0.135699302,-0.0307523031,-0.0200368818,0.2575927973,0.0576005317,-0.302549839,-0.2730573714,-0.3634371459,0.3274671137,0.3770821989,0.1571439058,-0.0795033798,0.2327446043,-0.3032660186,-0.0990575701,-0.0700983182,-0.0342253484,0.1696012467,0.1677809805,0.0799389184,0.326069504,-0.2400519401,-0.1048390195,-0.0388669893,-0.6121016145,0.5534069538,-0.1013173312,0.0261404924,0.0357758291,0.0137112308,0.0058050686,-0.3206295073,0.0161508936,-0.1123400554,0.2814399302,0.1460608691,-0.302093327,-0.0639567226,0.3813228309,0.0859741345,-0.2923620939,0.4071855247,-0.2209590226,-0.5338458419,0.0497613251,0.260435015,0.0968717858,-0.146481961,-0.0170062333,0.1078794524,-0.3001843989,-0.1210719272,0.254668802,0.185167402,0.2065512985,-0.4720928073,0.0112438379,0.0818925053,-0.0036514818,-0.0161530916,-0.0921355337,0.2416178137,0.484639585,0.0955204144,0.2377486825,-0.1293536872,0.2788258791,-0.1929997653,0.0459284931,0.0898495391,0.1539402604,-0.0809658393,0.2004221827,0.048523102,-0.2047701925,0.1411314607,0.2866167724,-0.2145594358,0.0336788446,0.5055497289,-0.2039590478,-0.2953153849,0.0027814482,-0.4623137116,0.2144079655,0.3736828566,-0.0918909013,-0.1253003329,-0.2086354643,-0.2770307362,0.2586474717,0.0922297165,-0.0739523172,-0.0727834627,0.3840747476,0.1688706279,-0.243040517,-0.4180779755,-0.076842986,-0.0994065329,-0.027301304,-0.0034203569,-0.2512782514,-0.4178160429,0.069103308,-0.1639865041,-0.4351183772,-0.1990760267,0.4360974431,-0.2216433734,-0.0673548952,0.4732997715,0.1288897693,0.2370726615,-0.1431431621,0.007104923,-0.4106118679,-0.2103329301,-0.5618729591,0.0192084201,-0.1001333073,0.2254370749,0.0012152165,0.0646459237,0.2162142843,0.2121963054,-0.7023997903,0.3988601863,-0.0864363834,0.2627839744,-0.1488780379,-0.5324491262,-0.226772964,-0.4946579635,0.0077943332,-0.1952170283,-0.1020655558,0.0882287621,-0.0602550656,-0.2269314528,-0.1090350673,-0.3814074993,-0.5280771852,-0.1570465714,0.204012841,-0.067910254,0.1073895395,0.318112731,-0.0506506078,0.1100220159,-0.212894097,0.004577552,-0.0235255323,0.3745417893,0.6374518871,-0.0600388274,-0.318636924,-0.2763063908,-0.3181281686,-0.0593611188,0.1465306282,-0.1140685603,0.1144101694,-0.2133482397,0.0933070555,0.2897822559,0.2270315737,0.5762727857,0.1408128142,-0.11085926,0.0571631454,-0.2429828346,0.1431851685,0.1934714764,0.3914706111,-0.1920492202,-0.0581867434,-0.2578208148,0.752178967,-0.1656473577,-0.3004691899,0.3863733709,0.0796524957,0.1235759407,0.0249821786,-0.1162035987,0.2011026591,0.0311027616,0.0922005028,0.4274145067,0.0600241311,-0.7173547745,-0.2091349512,-0.0713737383,-0.360771209,-0.0738186315,0.0101721277,0.2364364564,0.3325144649,-0.0094520794,0.0543699712,-0.1251545548,-0.2496002167,0.0913428739,0.015642561,0.0475948043,-0.1001980379,0.070395872,-0.3965108097,-0.2799776793,0.3130528033,0.2775229812,0.1295528114,-0.0985505134,0.0140917301,0.3031609952,0.0464336574,0.4732941091,-0.0971062481,-0.2817866504,0.303475827,0.1876395494,-0.3634348512,-0.0693610609,-0.4106271565,0.1294392943,0.458759129,0.22167027,-0.2273666114,-0.0255948231,0.1146298647,-0.3202495575,-0.1248117834,-0.3508260548,-0.1368908137,-0.1614022404,-0.048022639,0.0086910678,-0.0388684608,0.122348845,-0.1172252446,-0.1230458319,-0.1439417005,-0.0151909105,-0.0530881062,0.3789561689,-0.0107016768,-0.2664443552,0.0614037514,0.1220488697,-0.0638591871,0.0691426918,0.066753529,0.1088246554,0.0550577827,0.2126855999,0.0544390641,0.4107648134,0.2898604274,0.0406301618,0.3420206606,-0.1964225173,0.0381487012,-0.0950250253,-0.1867629141,0.3242314458,-0.0163463429,-0.0734036937,0.1530244499,0.6108981967,0.0410074443,-0.2492134571,-0.2941947281,0.494892329,-0.2021753639,0.9087722898,0.017033089,1.133374095,0.1396925598,0.0611042194,0.2462553382,-0.0842124447,0.6033166647,-0.4079711735,0.3965848684,-0.1873841584,0.147798717,0.0454423949,-0.1510640532,0.0959808007,0.1272281855,-0.4280603528,0.2706246078,0.5491973758,-0.3083282709,0.193508327,0.4062750936,0.2713347375,-0.1410001814,-0.3117100894,0.042742487,0.1393652558,0.3951956332,0.0157008693,-0.4653589427,-0.181456089,-0.1694209576,-0.1713799387,0.2771310508,0.1936981976,0.1036171466,0.6412373781,-0.0107709114,0.3417340219,-0.1403710842,0.146108076,-0.2261621058,-0.3170706332,0.3154677153,-0.1712604612,-0.0721534491,0.1493685693,0.0974320918,0.4282272756,-0.089545913,-0.2788538635,0.1490695328,0.0220954381,-0.1760278046,-0.2314785719,-0.2273624986,-0.4244578481,-0.0597065352,0.3894451261,0.1086545736,-0.0448740274,-0.0818957165,0.1337123364,0.0657279417,-0.0608475581,0.3853732646,0.1624206603,0.1091585159,-0.1178532764,0.4234651923,0.1116663516,-0.0448558778,0.2996018231,0.0848898515,-0.3486418724,0.0073482003,-0.5456367135,-0.0168501996,-0.4922614396,-0.055906333,0.0285593458,-0.2865239382,0.0879115537,0.2721900046,0.1671655774,-0.050202366,-0.4117514789,0.0614252798,-0.1304931939,0.4549987316,0.0094224196,0.1388649642,-0.0578336418,0.2327504158,0.2530173957,0.4221578538,-0.2839446366,-0.1441483498,-0.1475256681,0.2386001647,-0.2370141298,-0.4023139775,0.1479158551,0.0238201208,0.0114108073,0.0819426998,-0.0960620269,-0.0742328838,-0.0392615497,0.1349999011,0.2621522844,-0.140286535,-0.326072365,-0.0002276463,0.1000276506,-0.0954461023,-0.0983676985,-0.2853772044,0.0039758598,0.2688608468,0.3101648986,-0.1968741417,0.2346119434,0.1285112053,0.1994066685,0.4068383873,0.2551558018,0.2564770281,-0.1416209191,-0.0747372359,0.1387975663,0.1374055594,0.1261006445,0.0589178838,0.2528766692,-0.3256140947,-0.2401995063,-0.2625833452,0.1354920864,0.2462040931,-0.1196286976,-0.1188750342,0.4286911786,0.0911407992,-0.3163745701,-0.0043920204,0.3697552085,0.2083333433,-0.0783289894,-0.0964377001,0.5143988132,0.0273677632,0.1132033616,0.1560065001,0.1627745032,0.3030823469,-0.3191551566,-0.0194678269,0.0985656381,0.0524445809,0.4392335415,0.0536861457,0.1793846041,0.1957720071,-0.1282253414,0.2207372487,-0.2564578354,0.260524869,0.2463692874,0.0980716944,0.029285064,0.0417117216,0.6925807595,0.0133149587,0.1029677466,0.3768158257,0.4046405852,-0.1469326168,-0.1315660924,0.3730296791,0.0391933955,-0.3156326115,0.3761726022,-0.0364095122,-0.2104445696,-0.025673138,-0.0667571947,-0.1785059273,0.0451071523,0.0448324606,-0.1364885569,-0.2802773416,0.1971090287,0.1187314019,0.3354721963,0.1892700344,0.3666147292,-0.2245603651,-0.2313597351,0.0424578898,0.4983382225,0.2246090919,0.1280283928,-0.6307685971,0.060575109,-0.0917689949,-0.1626633257,-0.0357746258,0.3476501405,0.1403847486,0.2971367836,-0.1498179734,-0.009017189,-0.0943763182,-0.1213889718,0.0095826052,0.2688102424,-0.0016039768,0.1707886457,0.234953776,-0.2277476192,0.0978339761,-0.1615010947,-0.4631528556,0.2085963339,0.2134367973,-0.1320759058,-0.059651427,0.0094561772,0.0610692427,0.1110362038,0.4639387429,0.5046027303,0.4634210765,0.0534520932,-0.1097840592,-0.0632171035,0.0661505163,-0.0042868196,0.0065690381,0.2461321354,0.1695417613,-0.0130824912,0.0458689034,-0.0164378397,0.2242896557,0.0097208861,-0.0279510003,-0.334608227,0.2038651258,-0.3020474017,-0.0616787151,-0.1221168637,0.2529063225,0.04355409,-0.1262925267,-0.0740373433,-0.2473900467,-0.006357478,0.1367205083,0.0700307339,0.4863363206,0.2643500566,0.2858442068,-0.1685404778,0.357231468,-0.1491307765,-0.2684081495,-0.0369234383,-0.0944071934,-0.0409678705,-0.0880129114,-0.2358717024,-0.2033033371,-0.1965103447,0.0558082797,-0.2785726488,-0.0634656921,-0.1141147763,-0.0547995493,-0.2686131299,0.0321201906,-0.0386712328,0.0790919885,-0.0798291564,0.1704727113,-0.2691622078,-0.2769066989,0.4201645851,-0.4225324392,0.0081148827,0.0434491262,0.0265987813,0.6671044827,0.1105668023,-0.5503877997,-0.2402530462,0.0807975903,0.0239000674,-0.1151850522,0.0876497775,0.3364045322,-0.196549654,-0.2017681748,0.0055699931,0.2425362021,0.0938605368,-0.3759280741,-0.2207045704]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/439","title":"Issues: Adding a FAISS or Elastic Search index to a Dataset","comments":"@lhoestq Thanks for the info ","body":"It seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?","comment_length":5,"text":"Issues: Adding a FAISS or Elastic Search index to a Dataset\nIt seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?\n@lhoestq Thanks for the info ","embeddings":[-0.0005728491,-0.0433662347,-0.1281870306,-0.1616427302,-0.2700934708,-0.3180909753,-0.1881952435,0.1559949666,-0.3328993022,0.0118547576,0.0719050393,0.3624118865,-0.08804252,0.0433548652,0.0530386828,-0.0919608027,0.301271677,0.432298094,0.2515711486,-0.1437955201,-0.0847475082,0.2545364201,-0.356554836,0.0831305236,-0.3310432732,0.1096082628,-0.4480222762,-0.2163606435,-0.2125193328,-0.5021004081,0.5132933855,0.1627689749,0.1703145653,-0.0171015505,-0.0001210224,-0.0814294964,0.5105143189,0.0475339554,0.0084816385,0.1269827783,0.0151676377,-0.0857815295,0.4660691321,-0.0448492654,-0.0209225696,-0.3444065452,0.2104459554,0.4107153118,0.2123342603,0.1024275124,0.0936245322,-0.1826345772,0.4026413262,0.0032056307,0.5839272141,-0.0687126443,-0.1450759768,-0.0447300188,0.5091927052,-0.022024747,-0.0118925842,-0.0454017706,0.1281314343,-0.2776149809,0.0675706193,0.1675360948,0.2793584168,-0.3763757348,-0.0264344513,0.0621410161,0.2271897793,-0.1119756475,-0.3535893261,-0.0635611191,0.0946680829,-0.4815112352,0.4189603627,-0.2536614239,-0.2412798107,0.1353937387,-0.0518073551,-0.3914181888,-0.208028391,0.3075883687,-0.2682688236,0.4162752926,0.0920159519,-0.2326638848,0.223287642,-0.2366577685,0.2544148266,0.1317359507,0.0120399222,0.3666706383,-0.0958687961,-0.1903039068,0.2251513302,0.2826619446,-0.1030693054,0.0983411521,-0.4622764587,0.1143693104,0.0786511824,0.1229480207,-0.2486568987,0.1716210395,0.1866899282,-0.2615396976,0.2325842381,-0.3880774081,0.1210515648,0.2110713869,-0.0274019819,0.0395407639,-0.4353526235,-0.3386831284,-0.1613726169,-0.2092517316,-0.1173680276,-0.1245806664,-0.3585345149,-0.2656780779,-0.1117475927,0.1710568368,0.147720173,-0.3749229908,0.1254147142,0.213864848,-0.1601815671,-0.1054482758,-0.0092510227,-0.0435682349,-0.1591959596,0.2443881929,0.2431576252,-0.3554673791,0.1403519511,-0.1643582135,-0.2190283239,0.3712398708,-0.3190948367,0.0782722831,-0.0778693929,0.1890890151,-0.1035197154,-0.1496707052,-0.5063604116,-0.1911386251,-0.2230169326,0.0864123702,-0.0679445863,-0.1594855189,-0.3750544786,0.1020258069,-0.203993991,-0.4183212221,0.4412987828,0.6470413804,-0.1090507284,-0.113340877,0.0306162313,0.0278218854,0.2621298134,-0.1777557284,0.122113809,-0.1409997791,-0.4273988307,-0.2961872816,-0.0426413119,0.0703025535,0.0665771738,0.1928218156,-0.1833486557,0.2139972895,0.0496526174,0.6125769019,0.6431605816,-0.2586215138,-0.0531090051,0.0253765937,0.1977274418,-0.2291392088,0.0299107283,-0.2245107591,0.1057475582,-0.039227441,0.0813605338,0.2696784735,0.0941357836,-0.4307818711,-0.2127266228,-0.4451807439,0.3611807525,0.370876193,0.2295947075,-0.1612865627,0.2078066468,-0.529669404,-0.0903146937,-0.0968640372,0.0356851742,0.1031555459,0.2056934237,0.1315399706,0.2401844114,-0.1997672766,0.0334603637,-0.0348021165,-0.5445060134,0.4802722037,-0.0401154757,0.0382721983,0.0969657153,-0.0113452757,-0.035505753,-0.2973996401,-0.0072447094,-0.11475043,0.3257943988,0.1981728226,-0.2675499916,-0.0712465644,0.2859100699,0.0169022679,-0.1034923419,0.3592851162,-0.2310105562,-0.6332054734,0.105603382,0.3280320466,0.1051890999,-0.0546365045,0.0808176026,0.138623178,-0.2897596657,-0.1691949517,0.1570290029,0.0192565881,0.1734499484,-0.3641731739,0.0632477477,-0.1188295707,-0.0520270243,0.0026047195,0.0911231935,0.1944022328,0.6303340793,-0.0034640862,0.290576309,-0.0540452562,0.2482799739,-0.1245462745,0.0775974616,0.1407359838,0.1255188286,-0.1108344346,0.2337611169,-0.0695797503,-0.2598693669,0.2119114399,0.2872964144,-0.2260112911,-0.0116194645,0.4041003287,-0.1453705281,-0.2832556367,0.0662483275,-0.6590614915,0.2720279098,0.3696914911,-0.104044877,-0.0899872109,-0.3034691513,-0.3238960505,0.2808303833,-0.0227197371,-0.0132606709,-0.1467975676,0.340880245,0.0982987359,-0.3085029125,-0.3743172586,-0.0580965206,-0.1111056879,-0.0687320679,-0.0696312487,-0.3782636821,-0.4554360211,0.099328585,-0.1611821055,-0.3582412899,-0.2265446484,0.4438015819,-0.1675900966,0.0636575818,0.4592600763,0.0237844493,0.3094112575,-0.1227205992,0.0512732156,-0.4008319676,-0.2389051914,-0.594587326,0.0274636839,-0.063890256,0.1326955855,0.0100343265,0.0597099438,0.1802032739,0.1357838959,-0.6611139774,0.379339695,-0.1610596031,0.2635665834,-0.0747438073,-0.5950272679,-0.2692507803,-0.4336057901,-0.0665487945,-0.1717321873,-0.1779719591,-0.0298748165,-0.1064487696,-0.3292008042,-0.1147249117,-0.2937689126,-0.4854054153,-0.024461478,0.1834976524,0.0139020728,0.112794973,0.1666338891,-0.0159015767,0.1987327337,-0.2398367077,0.0312836282,-0.0594377741,0.3014599681,0.5734540224,-0.057866659,-0.323277384,-0.1101765186,-0.181163609,0.0796942711,0.2396766543,-0.1948302686,0.105269447,-0.1000655964,0.3256410956,0.2328191251,0.2140440345,0.5745401978,0.0611406825,-0.0844819173,0.0128852185,-0.2243565768,0.2245058566,0.0779540613,0.380766809,-0.0963220298,-0.0537642576,-0.2270361036,0.7084752917,-0.1731277108,-0.2797248662,0.4416053593,0.0124095846,0.1612989604,0.0545468181,-0.0128555084,0.3029320836,0.0717057437,0.1331864595,0.3710061312,-0.0032483167,-0.699377954,-0.2548531592,-0.0719630569,-0.3824655712,0.0250202697,-0.197716862,0.1365480423,0.3465587795,-0.0671175867,0.0703860521,0.0149619812,-0.2225720733,0.1063174829,0.0423419625,0.1041924208,-0.0690944269,0.0400882587,-0.4239802957,-0.2635501325,0.3537887931,0.208396256,0.1246383563,-0.1172901243,0.0412163995,0.2901940644,-0.075032711,0.4178807139,-0.1020243391,-0.356526345,0.3757524192,0.4183238447,-0.3125242591,-0.1247012988,-0.3311385512,0.2096091658,0.4780567288,0.1792363375,-0.1488509923,-0.0567471832,0.163748458,-0.3992879391,-0.0182657335,-0.2881739438,-0.1396178901,-0.2477751076,0.0280786231,0.040390715,-0.0347870775,0.1256387085,-0.1834365278,-0.0897646397,-0.1394307911,-0.0347776748,-0.0394252129,0.3023409247,-0.093630068,-0.2966933548,0.0214887168,0.0430032238,-0.0127333496,0.1519946903,0.0310891606,0.1700540185,0.0561541356,0.2796247303,-0.0625447705,0.5810094476,0.2999942601,-0.0065857209,0.3328880668,-0.3060152829,-0.0654444546,-0.047005605,-0.2103426009,0.4316086471,-0.0763610378,-0.0598368496,0.1652100384,0.4756461382,0.1570772976,-0.2271935493,-0.2929911613,0.5624752641,-0.2469537258,0.7161688209,0.0434141271,1.0953243971,0.1645942032,0.0685948431,0.3798027337,-0.0594525039,0.5251387954,-0.3620476723,0.3734549582,-0.1578535736,0.3070719242,-0.0048591914,-0.1340152621,0.2832717597,0.0910792649,-0.4176329076,0.2621389031,0.5808618069,-0.3950846195,0.249052763,0.242130518,0.2326229513,-0.0454164818,-0.2785351872,0.0478122234,0.1754039675,0.2904745638,-0.018601466,-0.3395021856,-0.1997811198,-0.1640990973,-0.171155408,0.3058553636,0.1708750576,0.0314888246,0.5451120138,-0.0741674453,0.3046736419,-0.2451993078,0.1402771771,-0.2595630288,-0.2944765389,0.2446025759,-0.0240335017,-0.128427431,0.2582857013,0.0714765564,0.4354897439,-0.1957001239,-0.2571916282,0.3085407019,0.0091008618,-0.1405538619,-0.3686766028,-0.2192860991,-0.3482624292,0.0288123339,0.3270379603,0.1356400251,0.1862197816,-0.0407048464,0.1049313247,0.0233955774,0.065530628,0.408167541,-0.0250311401,0.1413914561,-0.1253872663,0.3788621426,0.1392504126,-0.0684032962,0.3608025014,0.0146901663,-0.3420086801,0.0081265429,-0.5571652055,-0.0007600208,-0.3385653794,-0.2149381042,0.0667276531,-0.1890444905,0.0577430055,0.3220976889,0.0132664051,-0.0613811985,-0.2655226886,0.0617342331,-0.0157810617,0.3862275779,-0.057313215,0.0820755884,0.0446915962,0.2713628411,0.2820868194,0.3221976757,-0.2354462743,-0.0719084963,-0.0071464991,0.1542820185,-0.247023657,-0.2650471926,0.0695205852,0.0415076055,0.0285177007,0.0453798883,0.0034316618,-0.029467633,-0.0254142843,0.1294003129,0.332924962,-0.15781717,-0.2091625929,-0.1172402799,0.0473072901,0.0962078869,-0.0709193572,-0.2457402647,-0.12818712,0.23424761,0.2236884683,-0.067372568,0.365314126,0.1676186621,0.1918429136,0.4821090102,0.2374751866,0.2886714935,-0.057086356,-0.0407600142,0.2595747411,0.1770250499,0.0828019902,0.0507070422,0.210346669,-0.3670747876,-0.2685824037,-0.3706044555,0.0784988105,0.3737337589,-0.1518747807,-0.1209523082,0.5318419337,0.0664168671,-0.2887474,0.0187011361,0.3953999579,0.1528473198,-0.1921505481,-0.1247996017,0.5388392806,-0.0348843522,0.0510082021,0.1180833802,0.0435648821,0.3344770968,-0.4072998464,-0.120131582,0.1487698853,0.005603157,0.3903916478,0.0997948796,0.1041930616,0.211667046,-0.097834371,0.1136659458,-0.1209189743,0.2271120101,0.1821363866,0.0575892851,-0.0283687543,0.0228954777,0.600998342,-0.0435791649,0.1480521262,0.5219131112,0.4575629234,-0.1598664075,-0.1405060291,0.4186472893,0.0181958433,-0.3480014503,0.2105666101,-0.0826009139,-0.165915966,-0.0364275686,-0.0810532868,-0.2138483226,0.1201909259,0.0745613649,-0.1184830144,-0.3211233914,0.125852868,0.03136703,0.3500638008,0.2297074497,0.4192808568,-0.2847613394,-0.2527825534,0.1371263862,0.4931594729,0.3373316824,0.2364542782,-0.6053808928,-0.1237224638,-0.1312019229,-0.128166914,-0.0984350815,0.357082963,0.243700251,0.3288280666,-0.1008030325,-0.0106119988,-0.035604164,0.0141731966,0.1402931064,0.1959873587,-0.152818203,0.0375471935,0.1629976034,-0.1935437918,0.0905948952,-0.1109645814,-0.4367716014,0.1798135489,0.1720016748,0.0004894907,-0.0805028453,-0.0493447632,0.0444958881,0.120950684,0.4053967893,0.5207721591,0.4194135964,0.1269430816,-0.224865213,-0.0031426421,0.1187664568,-0.1110726744,-0.133467108,0.1113169342,0.2358048409,0.0270754546,0.0179983042,-0.0015810149,0.2204099745,0.0736160204,-0.0124862297,-0.3912204802,0.2468479574,-0.1551332623,-0.0659504458,-0.1609391272,0.2730821073,0.1157847419,-0.2447595894,-0.066490896,-0.1956599802,0.043030221,0.1430738866,-0.0098609338,0.6027161479,0.2695738077,0.3618194759,-0.12204694,0.2949353755,-0.2124774158,-0.1721458882,-0.1379444748,0.0099391649,-0.0666697621,-0.2283872515,-0.2023525685,-0.1530683339,-0.0475709289,0.141737625,-0.2686432004,0.0048682052,-0.0860575512,-0.0256003905,-0.2326705456,0.0506920442,-0.0962750614,0.0453339741,-0.1238682345,0.2467664182,-0.3570521772,-0.1700277776,0.3391656876,-0.4801633656,0.0902188867,-0.028678149,0.0429007858,0.6544452906,0.1042732447,-0.4291671813,-0.2854018211,0.1277927458,0.0383634828,-0.1158087105,0.1485042423,0.2893218994,-0.2017847151,-0.1505577266,-0.0856366307,0.1439736336,0.0577405244,-0.3361805677,-0.2742719352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/439","title":"Issues: Adding a FAISS or Elastic Search index to a Dataset","comments":"@lhoestq I tried installing transformer from the master branch. Python imports for DPR again didnt' work. Anyways, Looking forward to trying it in the next release of nlp ","body":"It seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?","comment_length":28,"text":"Issues: Adding a FAISS or Elastic Search index to a Dataset\nIt seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?\n@lhoestq I tried installing transformer from the master branch. Python imports for DPR again didnt' work. Anyways, Looking forward to trying it in the next release of nlp ","embeddings":[0.0592440069,-0.0141797466,-0.1230471805,-0.17468445,-0.2136998028,-0.3094342649,-0.2024402469,0.1878231615,-0.2731280327,-0.0180026889,0.1258614212,0.4488860071,-0.1200058386,0.0128093157,0.027212875,-0.201290831,0.2756510079,0.4141331017,0.18807742,-0.1802969575,-0.0639467686,0.2660615146,-0.3757490814,0.1084293425,-0.373339057,0.1195128113,-0.4001512825,-0.1794672757,-0.2072295845,-0.5200274587,0.4982812703,0.1200130805,0.1677700281,-0.0054156356,-0.0001168645,-0.083438307,0.5949627757,0.0467752628,-0.0684185997,0.0520079508,0.0476001501,-0.1152839735,0.5010398626,-0.0478227511,0.0164522193,-0.3274890482,0.202020362,0.4445816875,0.2767577767,0.1424150616,0.1396361738,-0.1405017525,0.3247395158,-0.0397446156,0.4015784264,-0.0442504436,-0.1242005676,-0.0819095224,0.4457002282,-0.0774324238,0.0063584549,-0.0440451503,0.0988059491,-0.1819062978,0.106622003,0.1319274306,0.2320646197,-0.3828480244,-0.022904532,0.0497767664,0.1373972446,-0.0845372304,-0.3730430305,-0.1281932592,0.107231304,-0.4572899044,0.4009217322,-0.2419735342,-0.2164808661,0.1059383303,-0.0273044035,-0.3213636875,-0.2079934627,0.3174877763,-0.2134216279,0.4697518945,0.0514706261,-0.2200693488,0.2339069843,-0.2200532258,0.2091199905,0.1614217758,0.0046576653,0.352492094,-0.0736411884,-0.2080271542,0.2646898031,0.2331193388,-0.1770539433,0.0817951933,-0.4286388457,0.1595923752,0.0953153446,0.081504412,-0.1969233453,0.2000460476,0.0850958079,-0.1773308069,0.2752023041,-0.3191640973,0.0964028835,0.2266253084,-0.0213946383,0.0757582933,-0.4990968406,-0.3258011639,-0.1640281379,-0.1835734099,-0.0718299076,-0.169581309,-0.3082044721,-0.27475214,-0.1742535233,0.2096710205,0.0724695474,-0.2116130888,0.2005261332,0.2056858689,-0.1665838808,-0.0948199704,-0.0224240031,-0.0403507724,-0.2784218788,0.2005558014,0.2567535937,-0.2662537098,0.223569259,-0.1683184952,-0.2347901314,0.3373529315,-0.271402061,0.1439445168,-0.0348651819,0.1348387897,-0.1238651574,-0.1286252737,-0.4871392548,-0.209195435,-0.1740940064,0.049970828,-0.0126799364,-0.1684695929,-0.3270560205,0.1455114484,-0.2288087308,-0.3878189027,0.4384027123,0.6058789492,-0.1549185663,-0.0916505978,0.0036915618,0.0008417973,0.3148565292,-0.1032304689,0.1083210781,-0.1296107024,-0.3205526471,-0.3506694734,-0.1114011481,0.1302195936,0.0836698115,0.2001714855,-0.2039765567,0.208803013,0.0842564851,0.5598309636,0.713757515,-0.2494044751,-0.0825687945,0.0901599079,0.139941439,-0.2311058044,0.0015102483,-0.1959902644,0.1299800873,-0.034863811,0.0693843663,0.3618213832,0.0850708634,-0.3851507008,-0.2908707261,-0.4074507356,0.3496367037,0.3368124068,0.2358149737,-0.0990298837,0.1416032314,-0.2938734591,-0.0359723344,-0.1276394129,0.0053384253,0.0921117142,0.093486242,0.1497267783,0.2479845136,-0.2244759053,0.0610379986,-0.0596712418,-0.5969663858,0.5358674526,0.0315454192,0.1064156517,0.0217187665,-0.0304881409,-0.0421889871,-0.3593740463,0.0465151928,-0.0584433936,0.3487761319,0.2340877205,-0.2853091955,-0.0835982934,0.2847379744,0.0492272377,-0.2090052515,0.3189229667,-0.2638424933,-0.5850793123,0.1359120309,0.3734184206,0.0751914605,-0.0719692111,0.0612218715,0.0764651969,-0.360091269,-0.1545645148,0.1580303311,0.0245531667,0.2197674215,-0.3405786455,0.0494574197,0.036509674,-0.0102006774,0.0282032751,0.0433063731,0.2037339509,0.5271430016,0.0429546535,0.2765484452,-0.0150383543,0.3046869934,-0.2103610784,0.0834927857,0.1226692572,0.1765886694,-0.1501626223,0.1475792229,-0.066856876,-0.3327257633,0.2287943065,0.2919632196,-0.1852939576,0.0069529135,0.4050530791,-0.168533206,-0.3128302991,0.0323313214,-0.5013378263,0.3017611206,0.3670205474,-0.1285232157,-0.0740580112,-0.3326497972,-0.3564534783,0.2985062003,0.0053439741,-0.103591077,-0.1053145006,0.3690223098,0.1527138799,-0.2460769266,-0.3854289651,-0.0478543416,-0.0822337419,-0.0234088711,-0.0061432971,-0.4174491763,-0.5681114793,0.0978690609,-0.1454336494,-0.3099070489,-0.274780035,0.4315383434,-0.1552889496,-0.0024639731,0.4569200873,0.0884085447,0.273609668,-0.151521638,0.10129302,-0.4035412073,-0.2560894787,-0.5996538401,0.0461750254,-0.0070877578,0.27017501,0.0883503035,0.027943125,0.1230091974,0.1659335941,-0.6194009781,0.3482745886,-0.148736909,0.285646081,-0.094053477,-0.551143229,-0.2057674527,-0.4268226922,-0.0547473095,-0.2271092087,-0.1468821913,0.0188630633,-0.0572920479,-0.3229562342,-0.1480235606,-0.4084331095,-0.4351570606,-0.1401527673,0.2142420709,0.039539028,0.0588513575,0.1975746006,-0.0757111982,0.187077418,-0.1790194809,0.0774518028,-0.0484722108,0.4085975885,0.5637871623,-0.0856353119,-0.4115509689,-0.1669464409,-0.2422399521,0.0285181347,0.1471623182,-0.1828513294,0.0580337495,-0.1403615177,0.1632647067,0.2589230835,0.2589215636,0.5880922079,0.0734345764,-0.1134550944,0.0305126105,-0.2120328695,0.1438655108,0.0883485004,0.3923083246,-0.1175510585,-0.0221724696,-0.2029241323,0.6386987567,-0.1549409479,-0.3184780777,0.3930821717,0.0697892755,0.1298812181,0.0054898551,-0.083811827,0.2943471074,0.0505582765,0.206607312,0.3836007118,-0.0216292888,-0.6913452148,-0.2100384682,-0.0116654644,-0.3025423288,0.0077629285,-0.1039873511,0.061493881,0.3284472823,0.0121214446,0.0533964634,0.007418565,-0.2160621881,0.0984765813,-0.0119185485,0.0943688303,-0.0891462564,0.0429699048,-0.4413600862,-0.2340092957,0.3478421867,0.1716234982,0.1526611745,-0.1299131215,-0.0009567663,0.2992293835,-0.024338888,0.3316754997,-0.1249462962,-0.3391498029,0.4021455944,0.3033942878,-0.3724469543,-0.1119722873,-0.39904055,0.1747431457,0.4907310009,0.2035344839,-0.1172671691,-0.0463108681,0.1898667067,-0.3391607702,-0.057077758,-0.246708706,-0.1130192354,-0.2649118602,-0.084726423,0.0074499589,-0.0721193328,0.1336932927,-0.1844217628,-0.0550441369,-0.1349455863,-0.015063799,-0.0477612913,0.3356503248,-0.1110273749,-0.2225155085,0.0734509602,0.0124915717,0.0551683903,0.0320753902,0.0116988849,0.1928334832,0.1264008284,0.2378136516,-0.0401193611,0.4711703658,0.3158352077,0.0417376906,0.3017108738,-0.3949320316,0.0374774523,-0.074853465,-0.144320488,0.3766115606,-0.0541478544,-0.0129250763,0.1214611754,0.4406951964,0.1119288653,-0.2288442552,-0.2961524725,0.4306162596,-0.2071007192,0.8095923066,0.0416992642,1.0631238222,0.2075321376,0.1021282598,0.32149297,-0.077637881,0.6415960789,-0.3527663946,0.3848264813,-0.2034813315,0.3640380204,0.050268162,-0.1063916981,0.2313315421,0.0651066229,-0.4607389569,0.3050265312,0.6316831708,-0.5515389442,0.2387447506,0.2986892164,0.2288756222,-0.1000685766,-0.3262163103,0.0651117787,0.1458827406,0.3255530298,0.0299458168,-0.3902041912,-0.2425750792,-0.1544918865,-0.1828269362,0.3070864677,0.2162060142,0.0256640259,0.5504886508,-0.0480522513,0.3429053426,-0.1830756068,0.2019944191,-0.2214521766,-0.2795245349,0.2059276104,-0.1806640476,-0.1863909811,0.2789824605,0.1108559221,0.4783039093,-0.1785264909,-0.2471667081,0.2273511738,-0.0119851027,-0.102162011,-0.4108915031,-0.1723776162,-0.3801229894,0.0314521976,0.2445616275,0.1389167607,0.0763692185,-0.0390642732,0.147603184,0.0275731105,-0.0304056369,0.4520755708,0.097859256,0.1462634057,-0.079644464,0.3522555232,0.0828048959,-0.1016990989,0.336019665,0.0391187966,-0.3150661588,-0.0359505676,-0.5736920834,-0.059138909,-0.3388136923,-0.1081736237,0.0787259862,-0.2951435745,0.0935455635,0.2771802247,0.054238867,-0.0378457531,-0.2932760119,0.0016484924,-0.0085250214,0.4036152661,-0.0916769803,0.0483510643,-0.0371697545,0.3229559958,0.3199605644,0.3426150084,-0.2777295709,-0.073433727,-0.1636998653,0.2091830075,-0.1636304706,-0.3471163511,-0.0160349421,0.0460452847,0.0319350399,0.0929893926,-0.0288794227,-0.0872123167,-0.0063764174,0.1165388525,0.3504098058,-0.1507745981,-0.182880193,-0.0751937851,0.0393702202,0.0581018068,-0.1120650992,-0.3036646247,-0.1069651619,0.3208753169,0.2055813372,-0.0668429956,0.3638146818,0.1654615998,0.1632012725,0.4954358339,0.1743815541,0.2557825744,-0.0062730811,-0.0377170369,0.1178360283,0.1068407893,-0.0002796347,0.1137574762,0.2845740914,-0.3544378877,-0.2378747761,-0.3208035827,0.1212215051,0.371085912,-0.1773017943,-0.1021061167,0.5498787761,0.1040200368,-0.3391272426,0.0448367596,0.3776488304,0.1836223304,-0.1782503575,-0.161305666,0.5033266544,-0.1237191558,0.0655734316,0.1693846136,0.0281030759,0.192570135,-0.320222795,-0.1176666468,0.128037855,-0.0256602131,0.2917709351,0.0684878528,0.1458537579,0.2214780152,-0.0340515226,0.1699712873,-0.1314250231,0.2671310306,0.2083176076,0.1047535315,-0.0451948643,0.0180856921,0.6436565518,-0.0035916748,0.0994978249,0.4499218166,0.3544284105,-0.191055432,-0.0970866755,0.4414842427,0.0177575033,-0.3282758892,0.2047424465,-0.0465886481,-0.2082740217,-0.0006446209,-0.0472014546,-0.1011464149,0.1203374863,0.084299095,-0.1994640827,-0.348992914,0.143758446,0.0285492372,0.348212719,0.2354936898,0.3813447952,-0.2906370163,-0.2081051916,0.0629163831,0.5100591183,0.3198822141,0.2481946945,-0.6800207496,-0.0642883256,-0.1547395289,-0.1307306737,-0.1188483462,0.4470883608,0.2539405525,0.3018109202,-0.100831762,0.000934527,-0.0910386145,0.0122664422,0.0730973035,0.1565634757,-0.102514632,0.1844874918,0.2533718944,-0.1798577011,0.073838979,-0.1097613052,-0.5304480791,0.2042800486,0.1903686076,-0.0047531333,-0.0675146803,-0.1293386817,0.0647716001,0.0852474198,0.4504702091,0.5197365284,0.3963978887,0.1408574283,-0.2211326957,-0.0473365895,0.1154525727,-0.0504339933,-0.114269577,0.1308113188,0.248954013,0.0097400993,0.0486930199,-0.0511353984,0.2500558794,0.030925015,-0.0090078656,-0.3527624309,0.2127010226,-0.1986591816,-0.0564961322,-0.1171879172,0.237661466,0.0571975224,-0.2142342329,-0.0326868184,-0.1692966074,0.0233711842,0.1450867951,0.2061759979,0.4772190452,0.3472390473,0.3181161284,-0.0800693408,0.2827456295,-0.2654673457,-0.2303610742,-0.1008697674,-0.0156668946,-0.0727607235,-0.1788671464,-0.190988183,-0.1551169157,-0.1508610249,0.080104202,-0.3075985014,0.0304202009,-0.1123346314,-0.0618696958,-0.2379007488,0.1116109043,-0.1198265702,0.1063623577,-0.1496864408,0.2067828923,-0.3719473779,-0.2287930846,0.3346479535,-0.4912219644,-0.0291734748,0.061373964,0.0329634771,0.524091661,0.0522615425,-0.4273997247,-0.3150725067,0.0883797333,0.0220305547,-0.1135816649,0.1176481321,0.2750373185,-0.2203362137,-0.1301643848,-0.0367375501,0.187740311,0.0569212064,-0.2648131847,-0.2374388278]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/439","title":"Issues: Adding a FAISS or Elastic Search index to a Dataset","comments":"@nsankar have you tried with the latest version of the library?","body":"It seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?","comment_length":11,"text":"Issues: Adding a FAISS or Elastic Search index to a Dataset\nIt seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?\n@nsankar have you tried with the latest version of the library?","embeddings":[-0.0245542824,-0.001539452,-0.1135864332,-0.1547237635,-0.2073552459,-0.3561769128,-0.1621095389,0.2129932493,-0.341753453,0.0482128002,0.0232495107,0.4107867777,-0.055466868,0.0625387877,0.0505981334,-0.1286363751,0.2190342396,0.4451150298,0.2279504091,-0.1211856902,-0.0436922163,0.2544832528,-0.3549373746,0.0833179057,-0.3355126679,0.0906089395,-0.3965982497,-0.1685612351,-0.2141688764,-0.4963502884,0.4322224557,0.1086656004,0.149423942,0.0034425692,-0.0001163825,-0.0812009349,0.4697083533,0.0273580216,0.0112125725,0.0543754958,0.1218591556,-0.0488333814,0.4858407378,-0.0708017424,-0.0586872436,-0.3351112008,0.1556550413,0.4582878351,0.2428094149,0.1769543588,0.1463176012,-0.1389537007,0.3781413436,0.0038164663,0.6323518753,-0.103785187,-0.1885399222,0.0021566998,0.4955141246,-0.0459520184,-0.0230656937,-0.0301571153,0.1063705683,-0.2256179601,0.057413891,0.2061547041,0.2377672791,-0.3189767301,0.018343525,0.0608662702,0.2374035716,-0.1684053838,-0.2930535078,-0.0340865031,0.1050058082,-0.4781893492,0.3844186962,-0.2260549814,-0.237477988,0.1263524741,-0.0090060402,-0.3211363852,-0.2138632089,0.3241856694,-0.1863450855,0.3751094341,0.069911316,-0.198029995,0.2903054357,-0.246318832,0.2173662484,0.2013404071,0.0135873463,0.3341629207,-0.1196700633,-0.133411631,0.2427708358,0.282931447,-0.0674477741,0.0906932577,-0.3819538653,0.1589216292,0.1155930087,0.1230606809,-0.2819137573,0.1891221702,0.164245531,-0.2678282559,0.2361722291,-0.331445843,0.0556127653,0.1557309926,-0.0006977412,0.0181977078,-0.4011991918,-0.3446329832,-0.1749732494,-0.2186411619,-0.1246767715,-0.1435180604,-0.363979131,-0.2336608917,-0.0891214833,0.264016062,0.1101448983,-0.4358128011,0.1709771603,0.2475895584,-0.2119646966,-0.068589665,-0.0376697071,0.0101482337,-0.2292437404,0.2551927269,0.2866322398,-0.3470789194,0.1691747606,-0.1097721756,-0.2355819046,0.2972417474,-0.2644736767,0.1105662882,-0.0816435292,0.1572897285,-0.047102917,-0.1501174271,-0.5101784468,-0.2175646871,-0.2129098177,0.1105603427,-0.0656304657,-0.1741945148,-0.4215660691,0.1337893754,-0.1908548474,-0.3714081049,0.429561764,0.6174010634,-0.1368160546,-0.1429593265,0.0116541805,0.0587527566,0.2552918494,-0.1824948043,0.1812925935,-0.1717953682,-0.3895470202,-0.2612292469,-0.0475365631,0.1000768989,0.1032883897,0.2410578281,-0.2066971511,0.1903051436,0.0195271038,0.6808969378,0.5908865333,-0.2290640324,-0.0884272084,0.0088063581,0.1393360347,-0.2605522275,-0.0260926746,-0.1982150823,0.0968435183,-0.0129652638,0.090909116,0.2969644964,0.1020522788,-0.3559378088,-0.2748663425,-0.3628128767,0.3624451756,0.3944834173,0.1616647542,-0.1731985658,0.184239164,-0.5021722317,-0.0694904178,-0.178805545,0.0313791782,0.0821012408,0.25947541,0.1459674984,0.2278744429,-0.1688607782,0.0328534208,-0.0500723273,-0.5197003484,0.5464216471,-0.0937126651,0.0243815687,0.0475618578,-0.0377072282,-0.0271134693,-0.2913638353,0.0446329005,-0.1168467924,0.2825915813,0.2327965051,-0.1940346956,-0.0985611007,0.2767080069,0.0481332839,-0.1108254939,0.37957865,-0.2813224494,-0.6578404307,0.0900058448,0.3189540803,0.1399358958,-0.0519903116,0.0362279601,0.1742653847,-0.3103432655,-0.1560119987,0.1497020274,0.0028498166,0.1464009881,-0.3952440321,0.0604016446,-0.0784906968,-0.0419441238,-0.0045683947,0.1147574186,0.2029353231,0.5944253802,-0.0100242244,0.2390759289,-0.0498533435,0.2664126456,-0.1845467836,0.1179322004,0.1645567715,0.1987147629,-0.1031218693,0.2498749197,-0.0507521443,-0.3252240717,0.2205350846,0.2797906995,-0.2160922885,0.0061271316,0.3959937692,-0.1731856167,-0.2580266893,0.0559636429,-0.6606028676,0.2298502922,0.38136971,-0.1281891167,-0.0539371334,-0.3359933794,-0.3465453982,0.2609963119,-0.0486200862,0.0347128436,-0.1267963648,0.3832954764,0.0735900551,-0.3701353371,-0.3501959443,-0.0800268427,-0.0543674827,-0.0849949941,-0.0503417216,-0.3570199609,-0.454700768,0.10718216,-0.2213718295,-0.3483038843,-0.2082534283,0.4506312013,-0.1121803224,0.0079387762,0.4468497336,0.0885843039,0.2356881797,-0.0797634646,0.000908976,-0.3771668077,-0.2224158347,-0.6241559982,0.0541109554,-0.0307403654,0.2025180757,0.0586835667,0.0413737744,0.178267017,0.0987932906,-0.6772302389,0.3362223208,-0.1977782547,0.3124272227,-0.0837309882,-0.5017846227,-0.2867713869,-0.4798568785,0.0245130025,-0.1897824705,-0.1408585757,0.0167504866,-0.1205248684,-0.305989325,-0.1767312139,-0.3450546861,-0.4211092591,-0.1257430911,0.2005823851,0.0112439003,0.100688532,0.1745295078,0.0181956328,0.2174475789,-0.2053563744,0.0426334254,-0.0921289623,0.2866069078,0.5858498216,-0.0872455835,-0.3678989112,-0.1044264361,-0.1553801,0.0955195799,0.1893403828,-0.1741998196,0.0512190461,-0.1230263337,0.3127374351,0.1949191988,0.2377985865,0.514631629,0.0397939831,-0.114112936,0.0265128314,-0.2888178825,0.2853465378,0.097780183,0.3881807327,-0.1305207461,-0.0572338104,-0.2289932668,0.630841434,-0.2222570777,-0.2938040495,0.5121849775,-0.0103013916,0.1506168991,0.0204717629,0.0036350635,0.330809921,0.0477855615,0.1332917809,0.4073446393,-0.0052241138,-0.6652308702,-0.2513327003,-0.1133026034,-0.4060069323,-0.0669377521,-0.1066581756,0.1968105882,0.2855465114,-0.0483643375,0.0599352866,0.0149880191,-0.2396517098,0.12597543,-0.020684503,0.0251263455,-0.0685396641,0.0395013839,-0.4200662076,-0.3132324815,0.3914328814,0.2157233506,0.0699346513,-0.0974062383,0.0207147337,0.3107304275,-0.0663854256,0.4426119328,-0.0736439899,-0.3462972939,0.3362149,0.3745627999,-0.3392273188,-0.1189585105,-0.3517142236,0.1679641008,0.4553541541,0.1396360099,-0.2109535187,-0.0072822492,0.1350600123,-0.3735393286,0.0325521305,-0.3196778893,-0.1168586388,-0.3129466176,-0.0379076526,0.0476282611,0.0006742069,0.2067198753,-0.1405279189,-0.1254246831,-0.0729585513,-0.0073018181,-0.0187406726,0.3143306077,-0.00416771,-0.2822290659,0.0688885972,0.0100479629,-0.0042281155,0.1479924619,0.0554466173,0.1649875641,0.0577155426,0.1918496639,-0.0658826157,0.5083691478,0.2503637969,0.0082238242,0.3645422757,-0.3428517878,0.0033523883,-0.0000318472,-0.1078316569,0.4210521281,-0.0540322512,-0.0278083049,0.1735319048,0.4858513474,0.0964448154,-0.1879675537,-0.2603549063,0.539373517,-0.240486607,0.7565554976,0.0336291529,1.0368853807,0.1285073161,0.0739260763,0.3879051507,-0.0467452258,0.5392677188,-0.3371371627,0.3418580592,-0.2295720279,0.25623101,0.0517058261,-0.1231684312,0.175148502,0.0312651098,-0.4429558516,0.2221535891,0.5489951968,-0.3778914809,0.2065080851,0.2508445978,0.2443801761,-0.0862433836,-0.3492280543,0.0812318027,0.1319749206,0.2694880068,0.0246226285,-0.3637912571,-0.2320937961,-0.1835349202,-0.1481844634,0.3221752942,0.1022071242,-0.0027124272,0.5628889203,-0.0349770971,0.3150503039,-0.2726286054,0.1060230359,-0.2169120908,-0.3056964278,0.2744453549,-0.049224589,-0.1710905135,0.2303186953,0.039415352,0.4336561561,-0.1907735169,-0.2672872245,0.2930591702,0.0173649043,-0.1330909431,-0.2989194393,-0.2186817974,-0.3516080678,0.0097831683,0.2741017938,0.12183889,0.1482135803,-0.0432327762,0.1410051435,0.0716711059,0.0055406126,0.3907176852,0.0319006592,0.1133241951,-0.1166644022,0.3384457529,0.0776373819,-0.045689065,0.3464992344,0.098527424,-0.3217143714,-0.0135607095,-0.6182681322,0.0029025371,-0.3527041972,-0.1847097278,0.1175085381,-0.2084526122,0.1200897247,0.3552391231,0.0200609975,-0.0648360103,-0.203477785,0.0670322031,-0.0339510329,0.3760292232,-0.0507968515,0.1180885881,0.0740568489,0.324624598,0.2776540816,0.3139818609,-0.2881685495,-0.063528873,-0.0819867328,0.181825757,-0.2414872795,-0.3168199658,0.0804058611,0.0699317902,0.0709952265,0.1415076852,-0.0450670235,-0.085935168,-0.0393305495,0.1042379364,0.2981131673,-0.1542753428,-0.1950400323,-0.0997908413,0.0540086515,0.0330132283,-0.080132395,-0.2702392042,-0.137443617,0.2198089957,0.1893294603,-0.1226808205,0.3479219377,0.1114983261,0.1725762039,0.4372424185,0.2149777114,0.2440991104,-0.1201478541,-0.0641302988,0.2804994285,0.1338222772,0.0947945118,0.0835155174,0.1238925159,-0.3146496415,-0.3241154253,-0.3869099617,0.1391511112,0.3405974209,-0.1528180838,-0.1354507059,0.5297644138,0.1281735897,-0.3264799416,0.0520902947,0.3494630456,0.1343473494,-0.1152145416,-0.1343857795,0.5030190945,-0.0319355242,0.0267265923,0.1226019412,0.0279014651,0.3149045408,-0.4194056094,-0.1346681416,0.0443822555,-0.0137823708,0.3750376403,0.0707632825,0.1517921984,0.2457557023,-0.1494507939,0.1189690754,-0.190524891,0.261033833,0.1113180816,0.0964486524,-0.0542383268,0.0538661517,0.5981782079,-0.0631164834,0.1290321499,0.5424049497,0.4653044939,-0.2081383467,-0.1048274934,0.452950865,-0.0291998107,-0.27583161,0.1875370443,-0.0979625955,-0.216697216,-0.0197389852,-0.0893647894,-0.1860192567,0.1326860338,0.0915560424,-0.1170289516,-0.33747226,0.0893151611,0.0795543492,0.2801910937,0.2148476988,0.4425668418,-0.2643951178,-0.2752006054,0.049881041,0.4122597575,0.3516042233,0.2450339347,-0.5735533834,-0.1130961552,-0.1293108016,-0.0929493681,-0.0861060992,0.3042808473,0.3008931577,0.3391284049,-0.0446497425,0.0181150977,-0.1110125482,-0.0065739108,0.0680006593,0.224329263,-0.1019364148,-0.034573853,0.1782685518,-0.1791155189,-0.0313906409,-0.193154797,-0.4550761878,0.2213860005,0.1394250244,-0.0186335724,-0.0708231479,-0.0866178498,0.0701076761,0.1118378118,0.419023335,0.4936645627,0.5085160136,0.0471466258,-0.2621205747,-0.0812241808,0.1601284891,-0.1109049246,-0.1135913357,0.1105054915,0.2255972326,0.0541197024,0.0250309482,0.0457284264,0.2603518665,0.0993642807,-0.0152138611,-0.3969551325,0.2714510262,-0.1794723123,-0.1348454505,-0.1022088453,0.2446039766,0.104849793,-0.2526442409,-0.003584425,-0.2452488691,0.0581054576,0.1228583828,0.0308433715,0.590149641,0.2420895398,0.3154292107,-0.1343026161,0.3056674004,-0.1619220078,-0.162504226,-0.107773751,-0.0488301963,-0.0342202187,-0.1493588984,-0.1283126026,-0.1207967848,-0.1024136394,0.1830076575,-0.2118986696,-0.0001050065,-0.0456096455,-0.0090395724,-0.2015458792,0.0329622552,-0.1393264532,0.063324824,-0.1157144383,0.2165559381,-0.3857641816,-0.2978006899,0.410890758,-0.4538801312,0.0223631281,0.0489040054,0.0807991624,0.6253060699,0.0748334303,-0.4578447342,-0.2437365949,0.0803133398,0.044340793,-0.1628679633,0.1821096241,0.2252066582,-0.1510492712,-0.1208052263,-0.0701349825,0.2437193245,0.0600931756,-0.2978931367,-0.2895350158]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/439","title":"Issues: Adding a FAISS or Elastic Search index to a Dataset","comments":"@yjernite it worked. Thanks","body":"It seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?","comment_length":4,"text":"Issues: Adding a FAISS or Elastic Search index to a Dataset\nIt seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https:\/\/huggingface.co\/nlp\/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on the latest PyArrow 1.0.0 ? Is it yet to be made generally available ?\n@yjernite it worked. Thanks","embeddings":[0.0094540464,-0.0357360281,-0.1238542646,-0.141221866,-0.252651006,-0.3397864699,-0.1636346728,0.1495333016,-0.3262387812,0.0181055777,0.0415333807,0.3825190365,-0.0817255527,0.0544614941,0.0559259616,-0.1090934873,0.2738979459,0.4375941753,0.2490928024,-0.1449594647,-0.0744017139,0.2579011321,-0.3638204038,0.058104571,-0.3384922445,0.1530575752,-0.4624394178,-0.2193035185,-0.2018120438,-0.4924741983,0.5068749189,0.1259762943,0.1717402339,-0.001196351,-0.0001207345,-0.0477548987,0.5175023079,0.0354391038,-0.0001129354,0.1045169607,0.0149445143,-0.087510325,0.4474404156,-0.052098617,-0.0394278578,-0.352848798,0.2128932476,0.3923454285,0.2174171805,0.1261616349,0.0961306542,-0.1726703495,0.4118255377,0.0213084556,0.5804804564,-0.0874430612,-0.1515645832,-0.0596912429,0.5103073716,-0.0102582118,0.0223495662,-0.0488061309,0.1310038567,-0.2571552992,0.0754978955,0.1813103408,0.2464198619,-0.3702451587,0.0037748553,0.0538277924,0.231638208,-0.1140418425,-0.3326314688,-0.0577313192,0.0801454857,-0.4548309743,0.4176031053,-0.256898731,-0.2396246642,0.1296420246,-0.0479804948,-0.3586061895,-0.2126643062,0.322165221,-0.2594604194,0.4101572633,0.0757976919,-0.2204825729,0.2287064046,-0.2311636508,0.2493600845,0.1337126344,0.0257871393,0.3835374117,-0.0989242345,-0.1754938513,0.2196744382,0.2538594902,-0.1152216941,0.0893336162,-0.4314972758,0.1191103756,0.0881666839,0.121681191,-0.2918832898,0.1978039443,0.2091540843,-0.2358865738,0.2142101079,-0.4075627327,0.070785448,0.1973724216,-0.0279993843,0.0243256819,-0.4213965237,-0.3498814404,-0.1591719538,-0.1937052459,-0.08449357,-0.1367829889,-0.3770424724,-0.2674830258,-0.1326702386,0.2222733051,0.1666626781,-0.4432296455,0.1358647645,0.2470799983,-0.1519603878,-0.0886324048,-0.0173136145,-0.0298220906,-0.1918947995,0.2491955012,0.2758085728,-0.3644800782,0.1527435482,-0.1873045415,-0.2413317114,0.3224204779,-0.2984886467,0.0806957781,-0.030052999,0.1779915243,-0.0872668624,-0.1643619686,-0.4862891138,-0.2114792019,-0.2095933706,0.1126573384,-0.0813167319,-0.1547572762,-0.3744917214,0.0954604894,-0.2119273543,-0.413033396,0.4329143465,0.6652031541,-0.0920538977,-0.1333378702,0.0283828583,0.0355295129,0.2518826127,-0.1607886404,0.153186053,-0.1489996612,-0.437515825,-0.2894413471,-0.0502865501,0.1068531126,0.100968644,0.1789331138,-0.1903114766,0.2146993279,0.018387273,0.6726570725,0.6390295625,-0.2512992918,-0.060709279,0.0200447347,0.170649901,-0.2480745465,0.0184249505,-0.2352078408,0.1263448298,-0.0084553212,0.0841671675,0.2493190169,0.100753814,-0.4055460691,-0.2286963165,-0.4024018049,0.3759069443,0.3796854019,0.2361094505,-0.1529919952,0.2378743589,-0.5355479717,-0.0854745135,-0.0957759321,0.0525387749,0.108568415,0.2182436734,0.1280775815,0.2489366531,-0.1877454072,0.0467771105,-0.0318783782,-0.5495729446,0.518471837,-0.0689589307,0.0533798859,0.0765874237,-0.0301047564,-0.0551944636,-0.3173451424,-0.0040898952,-0.0983164459,0.3196188211,0.1768294722,-0.2591540515,-0.0639098063,0.2906500697,0.0231505912,-0.1152781099,0.3826515377,-0.2359859794,-0.6359205246,0.073677212,0.317394197,0.1167758331,-0.0833316669,0.0675584525,0.1315572113,-0.3005569279,-0.171963349,0.1843318045,0.0022155447,0.1555220634,-0.3953784406,0.0509598516,-0.1170498133,-0.0262738224,-0.0085314782,0.11387752,0.1941253692,0.6254419684,0.0101306429,0.2651257217,-0.0676420182,0.245365724,-0.136508733,0.0643566102,0.1455031484,0.1392696351,-0.1057647765,0.2094829082,-0.0718789101,-0.2601416707,0.2151393741,0.2767885327,-0.2183095962,-0.0112022217,0.3847700655,-0.1775731742,-0.266913712,0.0542543568,-0.6434066892,0.2922174037,0.3664958477,-0.0997827873,-0.066759482,-0.2859955132,-0.3235893548,0.2918123603,-0.0309875552,-0.0048910575,-0.1214433238,0.3569918871,0.0823461115,-0.3211316466,-0.3698528409,-0.0237917993,-0.0973535031,-0.1062873602,-0.0498959459,-0.3664730191,-0.4447219968,0.0950667188,-0.1723025441,-0.3544752002,-0.2305057794,0.4335576594,-0.1731306911,0.0556508973,0.4875670373,0.0492941476,0.2975474894,-0.1190383881,0.0074383169,-0.3970850706,-0.2244241685,-0.5870612264,0.0269960202,-0.0693240464,0.1385253072,-0.0022984422,0.0608213618,0.169567138,0.1294540018,-0.6695142984,0.3712559938,-0.1642206758,0.2957879901,-0.1018170118,-0.56773597,-0.2995181084,-0.4233169556,-0.0373924151,-0.1472709179,-0.1745836735,-0.0085128089,-0.0961393863,-0.3208441138,-0.1269982159,-0.3161683977,-0.4873342514,-0.0241051745,0.1660468429,-0.0129044233,0.0946638361,0.1640307158,-0.0234022066,0.2087618411,-0.2304756939,0.0134399915,-0.0739939585,0.3137950897,0.5852748156,-0.0842980593,-0.3261758685,-0.1005242318,-0.1723805964,0.0688279048,0.2353311777,-0.158431381,0.1035600081,-0.1083878055,0.3117881417,0.2462663949,0.2317254543,0.5563802123,0.0575905405,-0.081900835,0.015590528,-0.2418079823,0.2563230693,0.0889969245,0.3943057954,-0.1022317708,-0.0623141639,-0.2305398732,0.7096121311,-0.1902943254,-0.2886635363,0.4703139067,-0.0145178512,0.1617920697,0.037676353,-0.0141086318,0.3262636364,0.0659870058,0.1254571825,0.3558137119,-0.0079645058,-0.7082495093,-0.2514848709,-0.0777228102,-0.4113292396,0.0195844844,-0.1779597402,0.1615950018,0.3431048393,-0.0370675921,0.0598328039,0.0143465977,-0.2306232899,0.0660361871,0.0371151045,0.0922557116,-0.0683311224,0.0494372547,-0.4403883517,-0.2833693624,0.3807015121,0.1890991181,0.121462971,-0.1231392547,0.0354212709,0.2960581481,-0.0701611862,0.4518096447,-0.1011815295,-0.3471783102,0.3522139192,0.4122709632,-0.3149521947,-0.1321873069,-0.3187739849,0.2206818312,0.4653257132,0.179840818,-0.1744995564,0.0024681301,0.1434897929,-0.3995710313,0.0068900944,-0.3155227602,-0.1254719049,-0.2615566552,0.003031536,0.0239027459,-0.0206904002,0.1630274206,-0.1560318172,-0.1153289378,-0.135500282,-0.0337413922,-0.0292833317,0.3134919405,-0.0738589242,-0.2856421173,0.0301283281,0.01910864,0.0133607201,0.1401610672,0.0343609676,0.1715648621,0.06951572,0.2827900052,-0.0915002078,0.5716290474,0.2984599471,-0.0019675253,0.3190789521,-0.2862324119,-0.0619528629,-0.0257025585,-0.2031553537,0.3976537287,-0.0473299064,-0.0604059994,0.1525593698,0.4669006765,0.1213827357,-0.2072561085,-0.2915594876,0.5784497261,-0.2638864219,0.725282371,0.0079049859,1.1063711643,0.1629609317,0.055424463,0.365807116,-0.0448570177,0.5364899635,-0.3555871248,0.384663254,-0.1751705855,0.2802248597,0.0003773367,-0.1451737583,0.2540291548,0.0836213976,-0.420381546,0.246266976,0.5734055638,-0.4077023268,0.2317418307,0.2174749672,0.2549011111,-0.0392473266,-0.287053138,0.0505179092,0.1741058528,0.3077270389,0.0007070872,-0.3423781693,-0.2374493182,-0.1784695834,-0.1668026596,0.3176146448,0.1454059333,0.026063649,0.5601052642,-0.0785060897,0.3061084747,-0.2421694249,0.1422657967,-0.2753143907,-0.2936785519,0.2545573115,-0.0110435393,-0.1419198215,0.2696094811,0.0478260331,0.4201253653,-0.1894711852,-0.2699014544,0.3071289062,0.009287891,-0.1603162587,-0.3481035829,-0.2058342844,-0.3809318542,0.0221928488,0.2785112262,0.1413228363,0.1761488616,-0.0428071171,0.1006007418,0.0362668298,0.0256444942,0.4086310267,-0.0089545799,0.1190171838,-0.1109970286,0.372410655,0.1041464061,-0.0591251254,0.3861734867,0.0333031528,-0.3309830129,0.0132293347,-0.5650982261,0.0020986991,-0.3378307521,-0.2144046575,0.070663102,-0.1718198061,0.0703093559,0.3340591192,0.0101974849,-0.0273138657,-0.2607450783,0.0477122068,-0.0138434405,0.3439603448,-0.0640141442,0.0959987789,0.047229901,0.285168469,0.3007401526,0.3032785952,-0.2378156036,-0.0768664703,-0.0306100771,0.1635163128,-0.2443610579,-0.2515267134,0.0740565807,0.0298723616,0.0226464383,0.086944595,-0.003207627,-0.0364200398,-0.0179299023,0.1338108033,0.3432252109,-0.1632845402,-0.2322294563,-0.1209359542,0.0354637951,0.0717884675,-0.0894821137,-0.2504986525,-0.1474997252,0.2257659733,0.2257313728,-0.1086261943,0.3735572398,0.1665724069,0.2130981088,0.4776049852,0.2144716382,0.2837162912,-0.0699151605,-0.0226004943,0.2662695944,0.160824731,0.0560228378,0.0854570195,0.2179456651,-0.3816607296,-0.2682374418,-0.394107908,0.0898921415,0.3642427325,-0.1702048331,-0.1230206564,0.5061643124,0.0650639161,-0.2903262377,0.0527098849,0.4087695181,0.1642670929,-0.1725904047,-0.1199088395,0.5224514008,-0.0056985589,0.0369980522,0.1292365044,0.0638167113,0.327801019,-0.4083188772,-0.1425917596,0.122637406,0.0267031007,0.3764810562,0.0901254341,0.0961341485,0.2424373478,-0.1090464443,0.1251366735,-0.1270441711,0.242990151,0.1406780332,0.0771874934,0.0001606749,0.0408702269,0.5934143066,-0.0298555586,0.1586448252,0.5116259456,0.4464848638,-0.1494604945,-0.1219684556,0.4462280869,0.008115082,-0.3330695629,0.197189942,-0.0960268974,-0.174850136,-0.0440182537,-0.0892777219,-0.1938221902,0.1084480509,0.0832841173,-0.1413889527,-0.3212343752,0.1375267059,0.0552320741,0.3154754639,0.2193086594,0.4170950353,-0.3079690039,-0.2486338317,0.138189137,0.4834546149,0.3351843059,0.2390382141,-0.5847216249,-0.0866157562,-0.1440473646,-0.1208994612,-0.1125885397,0.3522328436,0.2775637805,0.2787273228,-0.0793421268,-0.0131693967,-0.0416172445,0.0160586555,0.1482792795,0.2299523503,-0.1228288412,0.0161812995,0.1542787105,-0.1790492088,0.04880514,-0.127372995,-0.4308683276,0.2104864717,0.1671837717,-0.0023771008,-0.0752239302,-0.0504712164,0.0448085479,0.1018786952,0.3937651813,0.5339479446,0.4348209798,0.1060228646,-0.2260167599,-0.0380295925,0.1216482744,-0.1025554761,-0.1244020015,0.1059517041,0.2350880206,0.0524591021,-0.0089170272,0.0236673113,0.2468495667,0.0865888968,-0.0019189393,-0.37347278,0.2257677764,-0.1666469872,-0.0539833643,-0.1564720124,0.2760296762,0.115881674,-0.2301570624,-0.0623382889,-0.201801002,0.0397446528,0.1161467433,-0.0022313036,0.6120072603,0.2661681175,0.3707815409,-0.1386486292,0.3351712227,-0.2113618851,-0.1573517919,-0.1428247839,0.0036233754,-0.0362552218,-0.2114487439,-0.193731606,-0.1611197293,-0.0761065856,0.1260576993,-0.2550311983,-0.0014812085,-0.0748073533,-0.0288860593,-0.2321487814,0.0476294905,-0.0913524777,0.0419857055,-0.1208751276,0.2437853068,-0.3549458981,-0.191865176,0.3539789915,-0.4621332288,0.0474082753,-0.0089702914,0.0447670929,0.6412859559,0.1040497124,-0.4430038035,-0.2918545902,0.1360300183,0.0433669984,-0.1149946377,0.1376481801,0.2691596746,-0.1750718206,-0.1604895294,-0.0877808705,0.1713685542,0.0684798658,-0.3574677706,-0.2736963034]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/438","title":"New Datasets: IWSLT15+, ITTB","comments":"Thanks Sam, we now have a very detailed tutorial and template on how to add a new dataset to the library. It typically take 1-2 hours to add one. Do you want to give it a try ?\r\nThe tutorial on writing a new dataset loading script is here: https:\/\/huggingface.co\/nlp\/add_dataset.html\r\nAnd the part on how to share a new dataset is here: https:\/\/huggingface.co\/nlp\/share_dataset.html","body":"**Links:**\r\n[iwslt](https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/_modules\/torchnlp\/datasets\/iwslt.html)\r\nDon't know if that link is up to date.\r\n\r\n[ittb](http:\/\/www.cfilt.iitb.ac.in\/iitb_parallel\/)\r\n**Motivation**: replicate mbart finetuning results (table below)\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/88490093-0c1c8c00-cf67-11ea-960d-8dcaad2aa8eb.png)\r\n\r\n\r\nFor future readers, we already have the following language pairs in the wmt namespaces:\r\n\r\n```\r\nwmt14: ['cs-en', 'de-en', 'fr-en', 'hi-en', 'ru-en']\r\nwmt15: ['cs-en', 'de-en', 'fi-en', 'fr-en', 'ru-en']\r\nwmt16: ['cs-en', 'de-en', 'fi-en', 'ro-en', 'ru-en', 'tr-en']\r\nwmt17: ['cs-en', 'de-en', 'fi-en', 'lv-en', 'ru-en', 'tr-en', 'zh-en']\r\nwmt18: ['cs-en', 'de-en', 'et-en', 'fi-en', 'kk-en', 'ru-en', 'tr-en', 'zh-en']\r\nwmt19: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']\r\n```","comment_length":63,"text":"New Datasets: IWSLT15+, ITTB\n**Links:**\r\n[iwslt](https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/_modules\/torchnlp\/datasets\/iwslt.html)\r\nDon't know if that link is up to date.\r\n\r\n[ittb](http:\/\/www.cfilt.iitb.ac.in\/iitb_parallel\/)\r\n**Motivation**: replicate mbart finetuning results (table below)\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/88490093-0c1c8c00-cf67-11ea-960d-8dcaad2aa8eb.png)\r\n\r\n\r\nFor future readers, we already have the following language pairs in the wmt namespaces:\r\n\r\n```\r\nwmt14: ['cs-en', 'de-en', 'fr-en', 'hi-en', 'ru-en']\r\nwmt15: ['cs-en', 'de-en', 'fi-en', 'fr-en', 'ru-en']\r\nwmt16: ['cs-en', 'de-en', 'fi-en', 'ro-en', 'ru-en', 'tr-en']\r\nwmt17: ['cs-en', 'de-en', 'fi-en', 'lv-en', 'ru-en', 'tr-en', 'zh-en']\r\nwmt18: ['cs-en', 'de-en', 'et-en', 'fi-en', 'kk-en', 'ru-en', 'tr-en', 'zh-en']\r\nwmt19: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']\r\n```\nThanks Sam, we now have a very detailed tutorial and template on how to add a new dataset to the library. It typically take 1-2 hours to add one. Do you want to give it a try ?\r\nThe tutorial on writing a new dataset loading script is here: https:\/\/huggingface.co\/nlp\/add_dataset.html\r\nAnd the part on how to share a new dataset is here: https:\/\/huggingface.co\/nlp\/share_dataset.html","embeddings":[-0.3334828615,-0.3722026646,-0.0924769789,0.0178870913,-0.182699725,0.1093524843,0.3688517809,0.2643735707,-0.125605315,0.3148258328,-0.4613344967,-0.0776399374,-0.064803049,0.1805540472,0.1016528159,0.1559136212,0.0136737432,-0.0235788152,-0.1810164452,-0.0543645583,-0.2652773261,0.2664156854,-0.1272369921,-0.1815755814,-0.3280387521,0.0483214222,-0.3616868556,0.148147583,-0.1781795323,-0.3179267943,-0.0962832645,0.2018675804,0.0043573729,0.2368448973,-0.0001070355,-0.1905525923,-0.0762819201,-0.0908374637,-0.3510018885,0.1006621495,-0.2646728456,-0.2651143074,-0.1263943166,-0.2893665731,-0.3872865438,-0.0671802238,-0.1151145101,-0.3258391023,0.0424567536,0.2000764608,0.2592927814,0.2519880533,-0.363966167,-0.2319181859,0.0204405412,-0.172586754,-0.1983187199,-0.0694193318,0.5608642101,0.2720471323,-0.2547256052,0.222103551,-0.0003467133,0.1921161562,-0.0048931451,0.121178329,0.3246239126,-0.006517875,-0.1593531072,0.4086940289,0.3740608096,-0.1164143011,-0.2830058634,-0.0593642667,-0.1900770962,-0.2303505838,-0.210508734,0.0816255063,-0.1196680292,0.124930203,-0.0415161848,0.0868017897,-0.2472940832,0.1161827669,-0.0923119634,0.9784516096,-0.0090050101,-0.0894223601,-0.0342593007,-0.2545594573,-0.045578938,-0.1130447984,0.0165110826,-0.0601291843,-0.0861638039,-0.2744801641,0.2924725115,0.0434693918,-0.0622249246,0.0590407737,0.1029751599,0.071152173,0.3325048089,-0.0934899375,-0.2517021,0.0551238433,-0.1344221234,-0.4063022733,0.092986919,-0.1309958398,0.049163416,0.1383208483,0.1221535429,-0.1984278262,-0.5016508698,0.0056258766,-0.1008068398,0.0083084553,-0.1851601899,-0.0633359104,0.1519845128,-0.217383787,-0.0983585417,0.1566028446,0.0331035592,0.0121343229,0.141125977,-0.0101370718,-0.2406610996,-0.2322077155,-0.1520594209,0.1009305716,-0.2026448548,0.0283981077,0.3800966144,0.5238032937,-0.041899018,-0.018971961,0.2195511907,0.40004915,-0.069653362,0.1028330252,0.7088277936,0.1578485519,-0.0999300256,0.0830124691,-0.0932005867,0.1064212769,-0.2886227369,-0.0697039813,0.2261247188,-0.1155597046,-0.0921943486,0.2713848352,-0.1784690917,-0.2367864996,0.5732960701,0.4183843136,-0.0948930457,-0.0091865566,-0.05276094,0.0809449703,-0.3100984395,-0.1258616447,0.0837474838,0.2288593352,-0.5910980701,-0.1100829914,-0.1314503253,0.2189666182,0.2230558246,0.1125309691,-0.0430241302,0.1650476307,-0.2222633064,0.3110174239,-0.0164760873,0.0106547289,0.1134816855,-0.0392748713,-0.1310180575,-0.1334378123,0.4240181148,0.191364333,0.2322646677,-0.1136516184,-0.1647869945,0.2035278976,-0.2517412901,0.0421081819,-0.2451383621,-0.2300683558,0.5268051028,0.1110182777,0.1924537271,-0.3725347519,0.0999875665,0.1659070253,0.4689549506,-0.2773914337,0.1345429271,-0.1531642228,0.3069565296,-0.0678336099,0.1636535823,-0.2083519697,-0.0154855736,0.1424131989,-0.1754942983,0.6150257587,0.2148750275,0.3419876397,0.0176722854,-0.0650309548,-0.2481763512,-0.2017679214,0.2806344032,0.2754361033,-0.1205326095,-0.283652246,-0.1200032383,-0.0141040767,-0.079423286,-0.1126092747,-0.3751195669,0.0413340479,-0.1244366914,-0.0230781771,0.11000631,0.4384022355,0.0560680926,-0.1176028624,-0.0888798907,0.0916532204,-0.2073235661,0.280642122,0.259760052,0.5066600442,0.1799169332,-0.3302954435,0.2640108168,-0.0370446555,0.0020613095,-0.1470851004,0.2458806783,0.4308546185,-0.1414052993,0.2891124785,0.0322199203,-0.1204729751,0.1530533135,-0.140721187,-0.1774052829,-0.2646365762,0.2876604795,-0.0283265524,0.030748507,-0.037250746,0.3311652243,0.398981601,0.4106965065,0.1012398601,0.0374423377,-0.1233828664,-0.0856721327,-0.1140096337,0.1963426024,-0.1481398344,0.1982419342,0.3941271305,0.2930689156,-0.0063033886,-0.0081451247,-0.2674535811,0.368232429,0.0831596777,-0.0806127489,0.3035694659,-0.1025728881,-0.0797005966,-0.3303620815,0.0464310311,0.0004443525,-0.0467909686,0.0490889437,0.0173673406,-0.234914422,-0.1608103514,-0.2334724814,-0.3551646471,-0.0800617859,-0.3991764784,0.2940744758,-0.0091903163,-0.2574594021,0.1491001695,0.0934518054,0.5956983566,-0.1647156179,0.3882248104,-0.2032001019,-0.2412425578,-0.2354210913,0.3034731448,0.1065895557,0.0333327726,0.5829252005,-0.287176162,0.1692098826,-0.0815070644,-0.4728700519,0.1743195802,0.0151837757,0.0840854123,0.2246305645,0.0063932128,0.1907295883,-0.2680400908,0.1473170817,-0.1281913519,-0.0734211504,-0.0660985336,-0.2781065702,-0.3567057848,-0.0101308562,-0.8072680831,-0.4668091536,-0.2272167355,-0.001244503,-0.274687618,0.077118367,0.2422076464,-0.1127375364,0.1935001612,-0.1483844221,0.1213304102,-0.2372917533,-0.0725687891,0.2338352799,-0.3333695829,-0.1917576641,0.0768791437,-0.0045270985,0.3215009868,-0.1392311901,-0.0256989654,0.4002657533,0.0484406389,0.2887951732,0.3326581419,0.4452306926,0.2124417275,0.2143184096,-0.278519243,-0.3447280824,-0.0780801773,0.0277503394,0.1879918724,0.2707400024,0.1232825741,-0.0448735654,0.1365181059,0.4299829304,0.1162668988,-0.1589788347,0.2109692693,0.0902185962,-0.017117288,0.0547474697,-0.4609020054,0.4322128594,-0.0753142014,-0.0908729509,0.6659477353,-0.0355999582,-0.0778805688,-0.1606562436,0.4709334075,-0.0669513494,-0.2566491067,0.0866394043,-0.0473858938,0.2723076642,0.1539360732,-0.3449326754,-0.3575872481,-0.2558023334,0.0479978025,0.055306565,-0.1811759919,-0.1124144942,0.0172268059,0.0508789085,-0.4486725032,0.1748123169,0.241456598,0.3168322742,-0.2165375054,-0.1181848943,-0.0517671183,0.1543852389,0.1541744322,-0.3376966417,-0.0922998115,0.1863248199,-0.1468840241,-0.055648312,0.0642167777,-0.4884222448,0.016567219,0.406889379,0.002917479,-0.0957784057,-0.2474063188,0.0839542449,0.3582585752,-0.2509702444,-0.0592717342,0.0600356832,-0.016442636,-0.0580673926,0.086789079,0.0004030965,0.127349928,0.0102264015,0.0821225792,0.1796138436,-0.0833232254,0.2549474239,0.1805141121,-0.14192307,0.2616842091,-0.0194865689,0.0551654547,0.3926153779,0.1041793376,0.2982866764,-0.237762332,-0.0168035813,0.1231902465,-0.2735191584,0.3650423288,0.5031778216,0.0542167649,0.0683701262,0.1348582655,0.0356107391,-0.3472727239,0.1898887306,-0.1537944227,0.2292508334,-0.1234714016,0.1608449817,0.0043979678,-0.1318328828,-0.1393216103,0.1764625162,0.0369965918,-0.4177050591,0.0813199952,0.2697201967,0.9081616402,0.1166390479,0.1585991681,0.2479411513,-0.1069554314,-0.1235681847,-0.0871870965,0.0164082069,-0.2227161527,0.0723864213,-0.0025101851,0.0884088203,-0.0037574368,0.4009153247,-0.0551584847,0.1393928081,0.5569096804,0.1016310528,0.1585954577,0.2449044585,0.1326387078,-0.4203194678,-0.0103844004,0.2484358698,-0.0707783699,-0.155305773,-0.1357741356,-0.1143147647,0.0853030682,0.0539998747,-0.2807552814,-0.0407026298,-0.5479024649,0.1362305433,-0.4296941161,-0.1737632006,0.1124768928,-0.2767390609,0.3626979887,-0.0050135865,-0.1961693466,-0.0884709209,-0.1599104255,0.1444286108,-0.2347540706,-0.0022887855,0.3436667025,-0.2297107875,-0.2189768106,-0.1852139533,0.154443875,-0.1659740657,0.0369705521,0.0560201071,0.4264519811,-0.1247679889,0.2401207387,0.3004263043,-0.1864312887,-0.2889264822,0.1699923277,0.0530195273,-0.4662047327,0.4557079077,0.0103968438,-0.1018569395,-0.2058898658,0.1422105879,0.1639945656,0.1020242795,0.0889318511,0.4726662338,-0.2444574237,-0.3337054551,0.1085772812,0.2642264068,-0.2122599781,0.0464463793,0.2941990495,-0.2451594323,0.0327763706,0.5464956164,0.4080560505,0.0927013382,-0.2307901531,-0.0631742179,-0.1403226405,0.1981207132,0.011873832,0.511702776,-0.0265912507,0.200482592,-0.1185657978,-0.0341289192,-0.4652319252,-0.0401999615,0.0808378533,0.0937294886,-0.1019302681,-0.2751204669,-0.1194088012,0.173966527,0.1288014799,-0.1284701377,-0.257157594,-0.2986705005,-0.2605069578,0.0846715719,0.0810297281,-0.2016494423,0.0910657495,0.0473579355,-0.0333055183,-0.1721798331,0.1191200763,-0.0243670102,-0.0794990659,0.3310456574,0.3911221325,-0.4266450107,-0.0646516308,0.0187307,0.3707017601,0.0060192882,0.1250882894,0.2350393385,0.1173316985,-0.1971157491,0.3461292982,0.2029721886,0.1662789136,-0.0980643705,0.1162739173,-0.0107697267,-0.1371082366,0.2711567283,0.2495612651,0.0644428059,0.0991470963,-0.3754108548,0.0503889546,0.287440449,-0.4011575282,-0.0783396214,-0.0078469487,-0.017826641,0.3304078877,0.2903861105,0.4299240708,-0.0786552429,0.1788142323,-0.0167873539,-0.2648407519,-0.124405697,0.0836110115,0.1068064719,-0.0598888993,0.1551773995,0.1306484044,-0.1934787929,0.0083457725,0.0659056529,-0.0772136822,0.9309542775,-0.0250179712,0.0329416022,0.2934620678,-0.090506576,0.136129424,-0.3587466776,0.1030771956,0.229292348,0.13882339,0.2720466852,0.0696391836,-0.3048446774,-0.2312242985,0.2099305093,-0.3689034581,-0.0648866966,0.0471069552,-0.2788591981,-0.2459999174,-0.1239402592,-0.0792432427,0.0644879267,0.3866289854,0.0630617514,0.1302510798,-0.1684917212,-0.0924484655,-0.0843956247,0.0218426753,-0.1803234071,0.3002913892,0.4029868543,-0.2464878708,-0.0792395622,0.4237110913,0.0625783578,-0.0248855017,-0.1714933962,0.1750069857,-0.2262555808,-0.1402631104,-0.1907266378,0.2400518209,-0.018945517,-0.2605830729,0.2505828738,0.1317292601,-0.1262590736,-0.1668637246,0.1634866446,-0.5733739734,-0.0259807464,0.033754725,0.1563693434,-0.0879802927,-0.1432314664,0.28407076,-0.3045538664,0.0279310513,0.2656625211,-0.3911350667,0.0452726893,0.0023919637,0.1117327288,-0.0416817479,0.7474104762,0.1418002546,-0.1485965848,-0.3499348164,-0.2589431703,-0.3720820844,-0.081516698,0.1000944898,0.3815597594,-0.3538078666,0.1064775512,0.2908460498,0.1203454137,0.1783047915,0.1155849174,-0.1099238619,0.1122639701,-0.5181726217,-0.0041556647,0.3043670058,0.2814522684,-0.0202700626,0.077609621,0.0000653169,-0.073958084,0.1314824373,-0.1722527891,-0.1767570823,0.3135387003,0.2274091542,-0.0641395077,0.2742759287,0.353069514,-0.3639172614,-0.176512152,-0.2551377714,-0.1928963065,-0.0312148575,0.292240411,0.0906791687,0.3055387139,0.0111887557,-0.1774166673,-0.046042081,-0.2982518673,-0.0074244561,-0.0276958663,-0.2427572012,0.2052917033,-0.174741596,0.3201977909,0.0931885615,0.5315694213,-0.0653217211,0.2351937294,-0.1621197164,-0.3991539478,0.3055489063,0.0906884298,-0.3430291414,-0.272295028,0.1945737153,-0.0463958196,0.2940964997,-0.3802622855,-0.1947213709,0.1568619758,0.0015752137,0.1056381837,0.5221001506,0.3171329498,-0.15150626,-0.1247108504,-0.0763132423,0.432012856,-0.2402193993,-0.3114913404,-0.2136413753]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/438","title":"New Datasets: IWSLT15+, ITTB","comments":"Hi @sshleifer, I'm trying to add IWSLT using the link you provided but the download urls are not working. Only `[en, de]` pair is working. For others language pairs it throws a `404` error.\r\n\r\n","body":"**Links:**\r\n[iwslt](https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/_modules\/torchnlp\/datasets\/iwslt.html)\r\nDon't know if that link is up to date.\r\n\r\n[ittb](http:\/\/www.cfilt.iitb.ac.in\/iitb_parallel\/)\r\n**Motivation**: replicate mbart finetuning results (table below)\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/88490093-0c1c8c00-cf67-11ea-960d-8dcaad2aa8eb.png)\r\n\r\n\r\nFor future readers, we already have the following language pairs in the wmt namespaces:\r\n\r\n```\r\nwmt14: ['cs-en', 'de-en', 'fr-en', 'hi-en', 'ru-en']\r\nwmt15: ['cs-en', 'de-en', 'fi-en', 'fr-en', 'ru-en']\r\nwmt16: ['cs-en', 'de-en', 'fi-en', 'ro-en', 'ru-en', 'tr-en']\r\nwmt17: ['cs-en', 'de-en', 'fi-en', 'lv-en', 'ru-en', 'tr-en', 'zh-en']\r\nwmt18: ['cs-en', 'de-en', 'et-en', 'fi-en', 'kk-en', 'ru-en', 'tr-en', 'zh-en']\r\nwmt19: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']\r\n```","comment_length":34,"text":"New Datasets: IWSLT15+, ITTB\n**Links:**\r\n[iwslt](https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/_modules\/torchnlp\/datasets\/iwslt.html)\r\nDon't know if that link is up to date.\r\n\r\n[ittb](http:\/\/www.cfilt.iitb.ac.in\/iitb_parallel\/)\r\n**Motivation**: replicate mbart finetuning results (table below)\r\n![image](https:\/\/user-images.githubusercontent.com\/6045025\/88490093-0c1c8c00-cf67-11ea-960d-8dcaad2aa8eb.png)\r\n\r\n\r\nFor future readers, we already have the following language pairs in the wmt namespaces:\r\n\r\n```\r\nwmt14: ['cs-en', 'de-en', 'fr-en', 'hi-en', 'ru-en']\r\nwmt15: ['cs-en', 'de-en', 'fi-en', 'fr-en', 'ru-en']\r\nwmt16: ['cs-en', 'de-en', 'fi-en', 'ro-en', 'ru-en', 'tr-en']\r\nwmt17: ['cs-en', 'de-en', 'fi-en', 'lv-en', 'ru-en', 'tr-en', 'zh-en']\r\nwmt18: ['cs-en', 'de-en', 'et-en', 'fi-en', 'kk-en', 'ru-en', 'tr-en', 'zh-en']\r\nwmt19: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']\r\n```\nHi @sshleifer, I'm trying to add IWSLT using the link you provided but the download urls are not working. Only `[en, de]` pair is working. For others language pairs it throws a `404` error.\r\n\r\n","embeddings":[-0.3510955274,-0.4201570749,-0.0554755107,0.0014696205,-0.2052135766,0.1677963436,0.3096143603,0.2312010378,-0.1078790054,0.326869607,-0.474678725,-0.0717082173,-0.0530310422,0.0761437491,0.112648055,0.203145802,-0.04159775,0.0062062764,-0.205002591,-0.0367566273,-0.2811895311,0.2966844738,-0.1565782279,-0.1628477275,-0.3049374223,0.1075593084,-0.323240906,0.1289975792,-0.0864088759,-0.2628643215,-0.0710038841,0.1811725646,-0.0486795306,0.0939843208,-0.0001127129,-0.1136948019,0.011659598,-0.1016759723,-0.3240738809,0.0458127372,-0.366542697,-0.2639515102,-0.2142889798,-0.3062394559,-0.4367866516,0.0272238739,-0.1831997633,-0.3564963043,-0.0052962503,0.1945134848,0.2235888541,0.3457286954,-0.314399749,-0.282266587,0.0567384325,-0.2643663585,-0.1838622093,-0.0266631953,0.476251781,0.3321308494,-0.0531216003,0.2584946156,0.0270025842,0.1334303468,-0.0349096283,0.1428125501,0.3649306893,0.0173942931,-0.0727352574,0.3325250447,0.4082353711,-0.0555605181,-0.2953056693,0.0243605524,-0.1866715103,-0.1836407036,-0.16567716,-0.0150444638,-0.0690909922,0.1800312996,-0.0824468732,0.0707493275,-0.2317941189,0.2495971769,-0.0581960455,0.9896836281,0.0268575624,-0.027306173,0.0627432838,-0.2822177708,-0.1278498471,-0.1018943191,0.0349727683,-0.053546384,-0.0472452939,-0.1655717939,0.3346602917,0.1215329766,-0.1151128337,0.0373837128,0.0262791626,0.018120233,0.3821308017,-0.1026368216,-0.3399943113,0.1468216777,-0.1371391863,-0.3576861024,0.067668654,-0.0888855606,0.0849977955,0.1274543256,0.0989745855,-0.2352352142,-0.5490626693,-0.0274979714,-0.1431760043,-0.0249166712,-0.3145843446,-0.0480165891,0.0230618082,-0.2022242695,-0.0835162848,0.2277546972,0.0744655356,-0.0106371902,0.295471251,0.0556020476,-0.2634866834,-0.1860876977,-0.156822443,0.2214243561,-0.1861123443,0.0067126662,0.3576880693,0.5407779813,-0.0904791281,-0.0771081075,0.2304045409,0.3313212097,-0.2134877145,0.0881097764,0.688242197,0.1761722863,-0.0530854613,0.1961775124,-0.127132833,0.1290379614,-0.2528000772,0.035009563,0.1868709922,-0.1165912598,-0.0831105784,0.2308927774,-0.1885018349,-0.2557547688,0.6013925672,0.2973755598,0.0287985578,-0.0596831441,-0.0814999938,0.0976899937,-0.3347741663,-0.0301047657,0.1800950915,0.2333528847,-0.6572201252,-0.166527912,-0.2068193406,0.2950187325,0.3693426549,0.1796917021,-0.045350723,0.2924351096,-0.258497715,0.4198895097,0.01014918,-0.004906219,-0.0069708456,-0.1160336435,-0.1494865566,-0.0744260177,0.4314526916,0.2410665005,0.1457803398,-0.0348599926,-0.1443373263,0.176410988,-0.1311493665,-0.0318324305,-0.2274167985,-0.1784517914,0.5984969139,0.1068547294,0.1780119687,-0.307426095,0.1291626096,0.2362651676,0.4651766121,-0.3246689737,0.2200807035,-0.1310893893,0.3670401573,-0.0452757701,0.1571962982,-0.1688234806,0.0765762031,0.1846998334,-0.1489457935,0.6234954,0.1539902389,0.3873569369,-0.0878154188,-0.0449774489,-0.3361417353,-0.1787450463,0.208551541,0.3206881285,-0.1827824414,-0.3383514583,-0.0287235677,-0.0390562378,-0.0480930693,-0.2065082639,-0.4026569426,0.1354035735,-0.2129510343,-0.0521700233,0.1030203104,0.5125248432,0.1033758,-0.0699645653,-0.1244540438,0.0962446481,-0.3203604519,0.303958267,0.2060095519,0.3942100406,0.1310862303,-0.3835148811,0.1998650879,0.0237835627,0.0238178968,-0.1549521089,0.3204905093,0.3441520035,-0.0772997662,0.3010514975,-0.0337892212,-0.1454748809,0.1417863369,-0.1534954756,-0.1454747617,-0.3616620302,0.3138562441,-0.0337949321,0.0471854769,-0.1138969809,0.3828154504,0.4016873837,0.3852693737,0.1135991439,0.1079316214,-0.0844484568,-0.0436761789,-0.0492627881,0.228262201,-0.2101130486,0.2338314205,0.3430885971,0.2765452266,0.0183295012,-0.0082963184,-0.2246361673,0.3973411024,0.0774704963,-0.1221395805,0.2896922827,-0.1076157615,-0.0456981845,-0.2772288918,0.1048429385,0.0116622923,-0.0693208501,-0.1093093231,-0.0416320115,-0.3411887586,-0.2628680468,-0.238880977,-0.3346974552,-0.0919712856,-0.403072983,0.2417892963,-0.0480080359,-0.2613851428,0.2518025041,0.1039010063,0.541272819,-0.1396017075,0.2570070028,-0.2098888457,-0.1634466499,-0.3350367844,0.2615450919,0.0714806691,-0.0096898237,0.6337125301,-0.2877264917,0.2285160273,-0.0371143296,-0.4671827555,0.113104485,0.0834075585,0.0288717747,0.2216667682,0.0138424914,0.2017314732,-0.3470591903,0.221379295,-0.0603058264,-0.0653206185,-0.1738288552,-0.2651665211,-0.4449757338,-0.0321839564,-0.7677303553,-0.4418911934,-0.265173316,-0.086501509,-0.3642118871,0.0767292082,0.2934867144,-0.1501422673,0.219566226,-0.0729876533,0.1199759915,-0.2860831618,0.0994729996,0.3871655166,-0.2667453885,-0.2149743736,0.1425022781,0.010607346,0.3104967177,-0.1521057487,0.0015578999,0.4561625421,0.08439935,0.2617935538,0.3264421523,0.5436169505,0.1978022754,0.1510165781,-0.2413122505,-0.3730677664,-0.184239924,0.1567745954,0.161958158,0.4169943631,0.1445748657,0.064528212,0.1511705369,0.4219455421,0.11032819,-0.1442943364,0.2931169569,0.0483067818,-0.0265367664,-0.0388945378,-0.5767920017,0.4208676219,-0.0256164633,-0.1082330048,0.6724064946,-0.0665040165,-0.0511077419,-0.2036521584,0.461772114,-0.0509654582,-0.2799990177,-0.0089730965,0.0442184471,0.2523649931,0.1876971126,-0.3371657133,-0.3310939968,-0.347961247,0.1593678445,0.0800979361,-0.2242390066,-0.084935993,0.0612809099,-0.071092017,-0.4463780224,0.2790630758,0.336253345,0.3429732323,-0.2130683362,-0.1520628333,-0.0080816559,0.1516465545,0.2631430924,-0.3619675338,0.0387280323,0.1935981214,-0.0585504696,-0.035601411,-0.0033464055,-0.4417619705,0.0097291777,0.5186340213,-0.0517321751,-0.1301388741,-0.0895861015,0.1040929928,0.2793541551,-0.2680600882,-0.1181624308,0.035395626,-0.0534730703,-0.0841062516,-0.010659379,0.0163344797,0.2835980058,0.0665003359,0.1180791631,0.2026327699,-0.2388780564,0.2810306549,0.0962100998,-0.0650148541,0.3816203177,-0.0399195626,0.0655696541,0.3526310027,0.0206252821,0.2904703617,-0.3022008836,0.0148082227,0.1298612058,-0.3256891668,0.3235577643,0.5054846406,-0.0188406501,0.0247140694,0.138424322,-0.0317915678,-0.2004708946,0.2371165901,-0.1519352496,0.2129073441,-0.0947300345,0.1840767711,0.0550777204,-0.1166441664,-0.1040548831,0.1748669297,0.1363048851,-0.4941784441,-0.0732403621,0.2296174318,0.9664259553,0.1904885024,0.1384125799,0.2758404613,-0.0335018262,-0.0657847002,0.0482109562,0.0508120619,-0.1869111061,0.1847465038,-0.0306970812,0.0752531216,-0.0217729229,0.431455791,-0.0369337574,0.135479793,0.5156269073,0.1383627802,0.2337651551,0.2205385119,0.2501472235,-0.4087945819,-0.0756470785,0.1737967581,-0.0941095352,-0.0286718886,-0.146177873,-0.1536179483,0.0346104689,-0.0042648432,-0.2722617388,0.0789710209,-0.5890011787,0.2749351561,-0.4028263092,-0.1198007762,0.1747227311,-0.3142213523,0.3008863628,0.0565457419,-0.2269769609,-0.0626105964,-0.1554569006,0.0963045508,-0.2385855168,0.0207430534,0.2923785746,-0.3063317835,-0.1959640831,-0.1612965316,0.164928779,-0.2451204807,0.0903508291,0.0248100255,0.3714930415,-0.0985419825,0.1155399159,0.3394878805,-0.0872314274,-0.2886785865,0.1070468128,-0.0192578305,-0.5331572294,0.3946993053,-0.0765464827,-0.0747835115,-0.1277844608,0.273950696,0.0700789094,0.1094877273,0.2058130652,0.4449132383,-0.2676798105,-0.2178170383,0.0435208306,0.2048487514,-0.2141147256,0.0088804001,0.3181270063,-0.2386740893,-0.0540528819,0.6365104914,0.4068299532,0.1738205701,-0.1967407167,-0.15196307,-0.1225923151,0.134927839,0.0622538552,0.4929794967,-0.0479075909,0.2813062966,-0.0494195446,-0.1467207819,-0.3966434598,-0.1032468528,0.0390230753,0.1037643701,-0.0148998527,-0.2560119927,-0.0667769089,0.2031358778,0.0904709473,-0.155072093,-0.2829782367,-0.220004037,-0.2144641578,0.0886420757,0.1174132079,-0.2440492511,0.1041123196,-0.0271591786,0.0535267778,-0.1022974104,0.1413944364,-0.1261367202,-0.1579963267,0.2246879637,0.3801629543,-0.5120620131,-0.1245449632,0.0705391914,0.3344875276,-0.0049788654,0.1500234902,0.2852220833,0.1169431284,-0.1642131805,0.383592397,0.1137949228,0.1523494273,-0.0438060686,0.1136318445,-0.1248553693,-0.1711814851,0.2598101199,0.3023602664,0.0808533952,0.0230444875,-0.4058658183,-0.035205625,0.2159339041,-0.4422450662,-0.0763634369,-0.0454624332,-0.0431356393,0.2890428901,0.2946321368,0.3833940029,-0.0936121196,0.2146470696,0.0091505386,-0.2317997664,-0.106312044,-0.0498088486,0.1242767125,-0.1273806393,0.2068740427,0.0705615059,-0.2522337139,-0.0125626735,0.035778854,-0.0393282883,0.9443347454,0.0554240197,0.1277288646,0.2607842684,-0.1525976211,0.1444082409,-0.4156581163,0.0907362178,0.2228733301,0.0885283127,0.1659009606,0.1051399633,-0.3098652959,-0.1888139248,0.3234905899,-0.3527396619,-0.1041432992,0.0336164944,-0.2721531391,-0.2560215592,-0.1753171384,-0.0485176556,0.0755503923,0.3166670501,0.0322546735,0.1280558407,-0.1600305885,-0.1547187418,-0.1163784862,0.0096212756,-0.2166687399,0.2953330874,0.3576920331,-0.2117865384,-0.0517920107,0.3719338775,0.1156517491,0.0277739689,-0.050783705,0.1282894462,-0.2201647311,-0.1960467994,-0.1785313189,0.3015795648,0.0281735789,-0.2969713807,0.2875295877,0.0589090772,-0.1092372611,-0.1599850655,0.1798593253,-0.4980974197,-0.0589873567,0.083505556,0.0308133643,-0.1172932088,-0.196173653,0.2696996033,-0.234976396,0.052969154,0.2362524271,-0.4396250546,0.0057427119,0.0263645425,0.0801034942,-0.0749606192,0.7365747094,0.0980662853,-0.1362930089,-0.3667143881,-0.2292541564,-0.4123720229,-0.1163257062,0.1767959744,0.3161047101,-0.3612354398,0.079573147,0.2017934322,-0.0329302922,0.2464620769,0.141008392,-0.0573460385,0.0953298211,-0.4217137992,-0.0286547393,0.2718576193,0.2104886919,-0.0182135701,0.0703177229,-0.0346258841,-0.0834260955,0.0849488527,-0.1449537575,-0.2116883546,0.2417709827,0.2613486946,-0.0441509783,0.2784499228,0.492744565,-0.3866874278,-0.0587782711,-0.2552832365,-0.2277171463,0.0211428143,0.2991320491,0.0188781451,0.3166448474,0.0603140518,-0.1228146479,-0.0644247532,-0.3232993186,0.0511759184,-0.0216291416,-0.2860192955,0.2448616624,-0.295673877,0.3996625841,0.1211718544,0.4754552245,-0.0391974151,0.255738765,-0.2701255381,-0.4410145581,0.3719326854,0.0036646416,-0.3797165155,-0.3095819056,0.2952937782,-0.1029625386,0.2130259573,-0.4063831866,-0.3061914742,0.2061426193,0.0328865908,0.0601146333,0.5993667841,0.3119817376,-0.0803834349,-0.0860492736,-0.139454931,0.4770441353,-0.2324855179,-0.2941697538,-0.2447911352]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"Indeed, we\u2019ll make a new PyPi release next week to solve this. Cc @lhoestq ","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":14,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\nIndeed, we\u2019ll make a new PyPi release next week to solve this. Cc @lhoestq ","embeddings":[-0.3392369747,0.2934693098,0.0317576043,-0.0220402982,-0.0496871509,-0.2233234495,0.262786448,0.1757927984,-0.2286839038,0.1325761229,0.00798593,0.6241756678,-0.2351939827,0.0988904461,0.0926541761,-0.0990001857,0.1063360125,0.3986949027,0.0128295356,0.1297717839,-0.229357183,0.2964934409,-0.3884560168,0.2502502203,-0.3117334545,-0.0392987207,-0.0200500265,0.0962794125,-0.2842769623,-0.6543644667,0.4784843922,-0.0572060198,0.2400705367,0.2173406929,-0.000111599,-0.0091814334,0.4241196215,0.0578185394,-0.2090955079,-0.4088561833,0.0362439454,-0.4664761722,0.42355147,-0.0825851336,0.0226546805,-0.0233201943,0.2677412331,0.6741569638,0.3069875836,0.2092727274,0.2148227394,0.2245769799,0.2967117131,-0.0144301755,0.4070102274,-0.1744731963,-0.0917883068,0.7653144002,0.144439742,-0.3450081646,-0.0039798003,0.0935420617,-0.2853981256,0.046531409,0.0053580175,-0.004025267,-0.1573207527,-0.302296102,-0.0368055589,0.2728122175,0.2513110042,-0.3649214208,-0.1440426558,-0.0104160774,0.1506655514,-0.5133193731,0.1511305124,0.248223722,-0.2821825147,0.2090216279,0.1314978302,-0.0723930746,-0.1946214437,0.1805589348,-0.112667948,0.7240642309,0.1325311214,0.110041678,0.291788429,-0.0528314374,0.5350092053,0.3152382076,-0.0698299631,0.1235053763,-0.0370288417,0.0411692485,0.0409464315,0.3268656731,0.0255139191,-0.2360066473,0.0249065142,0.0191480927,0.1348026395,0.470793575,0.0330586769,-0.0363314077,-0.0503133908,-0.0162047092,0.4206344783,0.0041476702,0.1295876354,0.0729732513,-0.0021716065,-0.2614982128,-0.0080086188,-0.0807637572,0.27564767,-0.1611829102,-0.4174940586,-0.2886713147,-0.4326880276,0.0397633277,0.0547927879,0.1962444037,-0.1212167069,-0.1084928066,0.0032814262,0.1652504057,-0.259469986,-0.0702780262,-0.0530278459,0.1969365627,-0.2244419456,0.053575445,0.2422944754,-0.1523776501,0.2560264468,0.1866278648,-0.112609826,0.213655144,-0.0186242778,-0.1483319402,-0.1473385841,0.3004699051,0.162759468,-0.0739635229,-0.1616992056,-0.3537233174,-0.2537117302,0.3430134952,-0.2767560184,-0.369631052,-0.3131356537,0.1827379018,-0.3541996479,-0.2238612175,-0.2265501171,0.1951098293,-0.1071543396,-0.4970433712,-0.1171649769,-0.3591523767,-0.0752383918,-0.2197665274,-0.0094393147,0.2177894264,-0.3511114419,-0.01363004,-0.1251176596,-0.137627095,0.3017826378,0.1964330077,-0.2715476751,0.0173929576,-0.1085005775,0.2879131138,0.2780938148,-0.0720730126,-0.3690602183,-0.0179263055,-0.0876758173,-0.1168197319,-0.0605678633,-0.0096796183,-0.0777804926,-0.0847178474,0.1970696747,0.2553397119,-0.0268606637,0.0927154198,-0.3226326108,-0.3252992928,0.2697103024,0.1829030514,0.0601787902,-0.1777040213,0.084073633,-0.0476221517,0.1352734417,-0.1670370996,-0.1797887385,0.0181705821,0.4019990563,-0.1601289362,-0.12204469,0.0229338389,-0.3845100403,0.073037602,-0.2471447736,0.3366112709,-0.1973266155,-0.1378156096,-0.122428298,0.1753091812,-0.0096553713,0.0350711346,0.124436371,-0.0545007326,0.0928732753,0.232864812,0.0415566787,0.0646025389,0.0973148271,0.0991351083,0.0707452074,0.3095397651,-0.3489869237,-0.513366878,0.1627689004,0.1900219768,0.1063151583,-0.01488858,0.0516965911,0.094851844,-0.0356682315,0.1084047183,-0.3037870228,-0.1864468753,0.214681536,-0.5018404126,0.085302785,0.0245988108,-0.0621625893,0.1214073971,0.380961448,0.2975330055,0.1774479747,-0.0549827442,-0.0175655242,0.0762096867,0.1711201072,-0.025827175,0.0849068612,0.0254151355,0.1273914874,0.307010591,0.1355317235,0.0026644112,-0.3554394841,0.1885105222,0.4818565249,0.0675521195,0.2604633868,0.1988289207,-0.1024804786,-0.2092631757,0.291215539,-0.218974039,0.2444397509,0.2590412796,-0.0612598285,0.2338923961,-0.3261166811,-0.3109157979,0.0834610611,0.034276586,0.1741087735,0.0050375536,0.3240847588,0.0142593337,-0.3909910321,-0.0716073141,-0.0112094944,0.3002662063,-0.1184788421,0.2415478081,-0.3960611522,-0.5724668503,-0.2022744417,-0.366464287,-0.3680895865,-0.0790619105,0.0748339966,0.3199467659,0.0530616418,0.2419890314,0.4113979638,0.0666190162,-0.2123963535,-0.2475361675,-0.2058165371,-0.4139110744,-0.5503935814,0.0644567683,0.2293907255,0.164899081,0.2071464062,-0.1668998897,0.0270595383,0.1094113588,-0.5953454971,0.0757646114,-0.2145796716,0.3488024771,0.1133668423,-0.0554782599,-0.3919770718,-0.3934222758,0.1048762202,-0.2510055304,-0.161050722,0.0409226492,-0.3410570323,0.0822747499,-0.2105450332,-0.1059995815,-0.6377460361,-0.2796353102,0.1449969262,0.1106170639,0.1182082146,0.3559190333,0.2504656315,0.1862797886,0.0552418977,-0.1549960822,-0.0134091033,0.0352934822,0.2056339383,-0.2131726295,-0.3103862703,0.153420046,-0.1589957774,0.0860274583,-0.0440390483,-0.4285992682,-0.437898159,0.0462589785,0.3645993471,0.1589928567,-0.0086495578,0.3375492692,0.0859184191,-0.0340553075,0.0581345223,-0.3418219388,-0.0897673368,0.1160088927,0.1328777969,0.0401616544,0.1724189669,-0.0903072059,0.6186881065,-0.2025326043,-0.1125110462,0.4701718688,-0.0987037569,0.3531838059,-0.0949959755,-0.078729026,0.0545735732,0.0583333336,-0.0729781091,0.2084480524,-0.2046915591,-0.5404559374,-0.1865817457,-0.2287668437,-0.1760933548,-0.2621060312,0.0930795595,0.4260255992,0.1568375975,0.1325678229,-0.1181971878,-0.0257542934,-0.3427359462,-0.1349135637,0.0268242825,-0.1255804747,0.0619319119,0.2764796913,-0.5300635695,-0.50499928,0.2608523965,0.3789630532,0.0215159543,-0.018556172,-0.0651370659,0.0933162645,-0.0853278562,0.0702372566,0.086839281,-0.0814478695,0.2208240032,0.2145804167,-0.2459621578,-0.1886479259,-0.1993587315,0.1642290354,0.3190975189,0.010098964,-0.3280435205,0.0879507884,0.5190327764,0.3675465882,0.0606097206,0.0656757653,-0.3867916763,-0.5745697618,-0.0558340289,0.0707352683,-0.0450844765,0.2848134935,-0.146700263,0.1442228556,0.1028063223,-0.1488038749,-0.1108399704,0.5146559477,0.0878384113,0.012136736,-0.0095371604,-0.1339746416,-0.2777296305,0.3942574561,0.4588159025,0.1384555697,-0.0476366915,-0.0537068695,0.2175900489,0.2126573175,0.0482199676,-0.066593796,0.1199905351,-0.2188140452,-0.1566564441,0.1731608361,0.0947425142,0.0901845619,-0.0417652354,-0.2154070735,0.0552402437,0.6632024646,0.1704327017,-0.0672640726,-0.0134828323,-0.0569247194,-0.2005129904,0.333983928,0.0857330486,0.8003559709,-0.0473045073,0.128250882,0.7686565518,-0.040770039,0.6616322994,0.1576844007,0.3007634878,-0.36757797,0.044546742,0.083835721,-0.0679553375,0.0557264052,-0.1542322934,-0.3966910243,0.2608518004,0.2773330808,-0.0209185556,0.3089393377,0.1316370368,-0.0367564671,-0.0607144646,-0.3428018391,0.1206031293,-0.1904006153,0.1996344477,-0.3532149494,-0.1206987128,0.0884729698,-0.3082356751,-0.3451384604,0.2689445019,-0.2955235839,0.0522629097,0.1779969633,-0.457439661,0.0107655162,-0.0744332373,-0.0632106066,-0.0269485526,-0.1484275162,0.1690430045,-0.0034527031,-0.322381258,-0.0679296702,0.2244364768,0.2334852368,-0.2004403323,-0.1860544533,0.1886053383,-0.0494428463,-0.1744143069,0.1130274162,0.1881930977,-0.1406112611,0.1809178591,0.0878951922,0.0666580051,0.2039437741,-0.1604315042,0.1449853182,0.0397650078,0.0724370405,0.1510445327,-0.0511770248,-0.2007578313,-0.1594549417,0.2025973797,0.0837348998,0.1778749228,0.4595393836,0.1866169721,-0.0709845722,-0.1493940949,-0.1532202065,-0.3281040788,-0.3581804633,0.0910356194,0.2186585069,0.1175531521,0.2290717214,0.387516439,0.1127440333,0.000554153,-0.0403282307,-0.3668115437,-0.2208737582,0.3274893463,0.0520301461,-0.0686656833,0.2957417667,0.4177664518,0.1823618561,-0.2387156487,-0.2955943644,0.1944308728,-0.2003257424,0.1930877119,-0.1089406833,-0.1404058486,0.0708929002,0.0985514,0.0940825939,0.0955042988,-0.0438842252,-0.2379386276,-0.1199545339,0.0856584981,0.1457294822,-0.1974217445,-0.1321279258,-0.1186948493,0.0592231974,0.118349418,0.0853426456,-0.0598725155,-0.2294171453,0.1627371609,0.13114582,0.2349924892,0.0058821877,0.128124997,-0.1958207786,0.3858366311,0.2257558554,0.3212281168,0.0988749191,0.0809754655,0.3035736084,-0.1269137114,-0.1301867813,-0.1394865215,0.0338316783,-0.3402389884,-0.0136811277,-0.3949730098,0.1536194831,0.322670579,-0.0862005204,-0.1765196621,0.5400688052,0.1687319279,-0.4595645964,0.0540804155,0.1055909619,0.0695563331,-0.0328467786,0.2252644449,0.2458655685,-0.029831795,0.1952607483,0.1830272824,0.2687755823,0.2683974802,-0.0169690456,0.2411066741,-0.1381444633,0.1102428734,0.2526797652,0.2032853067,0.3938998878,0.0573224276,-0.2947685719,0.2101459354,-0.2681460977,0.0502208211,0.044683598,-0.4375732243,-0.133096233,0.2036598325,0.2805158496,-0.1908807307,0.1574424356,0.4274638295,-0.1384090632,0.0074763051,-0.1625699252,0.4161293805,-0.1453242898,-0.0255786851,0.0394336134,0.0239699781,-0.0745191053,0.2211205363,-0.1117377654,-0.094936125,0.2273573726,0.1917529553,-0.2000326216,-0.1608785838,0.0660189241,0.2553951144,-0.1169338822,-0.0224327873,0.2725971937,-0.0847876742,-0.3400856256,-0.2642115057,0.4986222088,0.5584606528,0.2903787792,-0.3203565776,-0.212390244,-0.255795151,0.0720279515,-0.1442080587,0.0707688779,0.444601953,0.4450095296,0.1891911626,0.1048671603,-0.1902817339,0.2347327024,0.1954122782,-0.0340795889,-0.0244053639,0.038913276,0.0871930197,-0.1198214293,-0.1599206328,0.0115789454,-0.2816613317,0.4089615941,0.2135634422,0.2120060325,-0.1024674252,-0.1490730643,0.0796958208,0.1085501611,0.6651382446,0.4189825356,0.5218883157,-0.2273147851,-0.2548103034,-0.5802953243,0.214461118,-0.4799848497,0.087062858,0.0440145284,-0.0460729264,-0.1898985207,0.3722726107,-0.2117864639,0.3542897701,-0.1106803715,-0.2885082662,-0.3285348713,-0.0417517014,0.0190956071,-0.3645994961,-0.0941203088,-0.2341274172,0.1322393119,-0.6707667708,0.0693676993,-0.3256232142,-0.2097930759,0.0571125187,0.1455117464,0.5608908534,0.1404225826,0.4892860651,0.1072753295,0.0252774972,-0.1350228488,-0.3623390496,-0.2255033106,0.1945313364,0.1277589798,0.143209666,-0.0736518353,0.0947615951,-0.2397429496,0.4307916164,-0.1301139593,0.1364756376,0.2110990584,0.1063375622,-0.1092028692,-0.0875111818,-0.1858198494,0.0474834405,-0.0038236165,-0.0294104479,-0.4920468926,-0.3538238406,0.3496243954,-0.4890987873,0.000989357,-0.1628233194,0.0697047412,0.070963636,-0.2418992668,-0.3712394536,0.1719385684,0.2556075454,0.0615517087,-0.3274610937,0.1555868387,-0.0002126237,-0.0849915445,-0.0972529352,0.2275911421,0.0445728004,-0.2989744246,-0.1009897962,-0.1762904376]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"+1! this is the reason our tests are failing at [TextAttack](https:\/\/github.com\/QData\/TextAttack) \r\n\r\n(Though it's worth noting if we fixed the version number of pyarrow to 0.16.0 that would fix our problem too. But in this case we'll just wait for you all to update)","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":43,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\n+1! this is the reason our tests are failing at [TextAttack](https:\/\/github.com\/QData\/TextAttack) \r\n\r\n(Though it's worth noting if we fixed the version number of pyarrow to 0.16.0 that would fix our problem too. But in this case we'll just wait for you all to update)","embeddings":[-0.3629812002,0.2522223592,0.0292827375,0.0668664575,-0.0656062812,-0.3179534078,0.368732363,0.1455648243,-0.137786448,0.1537568867,-0.0934416428,0.4229706526,-0.1183706224,0.0239222087,0.1916114539,-0.0460920483,0.0454057381,0.3213040233,-0.1009332687,0.2128739059,-0.1327811182,0.3483653367,-0.5873017311,0.2542974353,-0.2375742793,-0.0698926151,-0.0117018726,-0.0221687034,-0.1265799105,-0.7411338687,0.3176329434,-0.0789566487,0.0764134601,0.3993419707,-0.0001173493,0.0144254118,0.4822439849,0.0728600919,-0.2069341391,-0.3872157335,0.038675338,-0.2277514338,0.3138642013,0.0373717546,0.0358879901,0.1512118876,0.2212150395,0.4453008771,0.3842617869,0.4977120161,0.1408520192,0.1463243812,0.2370873541,0.0650931373,0.5099895,-0.2810863853,-0.1513428986,0.495528698,0.3179702461,-0.1587514877,-0.0368670933,0.103491798,-0.1534598768,0.0865598246,0.0030837287,0.1041215062,-0.2000458539,-0.125892058,-0.0360218585,0.2756792307,0.3768588006,-0.450868994,-0.2491981983,-0.01317563,0.1185371056,-0.6331674457,0.2870849073,0.116099149,-0.2995830178,0.2648494542,-0.1115242243,0.0107542686,-0.1450327188,0.1135522872,-0.0236420818,0.6451193094,0.012620586,0.1286133379,0.1633942574,-0.0405531228,0.4524587691,0.1952841729,-0.0274376236,0.1291082948,-0.0782346576,0.0992477909,-0.1888786703,0.0984959528,0.1803100705,-0.1165357903,0.0249171834,0.0484945662,-0.0032617196,0.354835093,-0.0418050028,0.19820939,0.1764465868,-0.075462766,0.4000109434,0.2195628434,-0.156437695,0.1064228714,0.0805545896,-0.3143377602,0.011677769,-0.092380479,0.2431232333,-0.334220916,-0.4480620027,-0.2376210988,-0.5925034285,-0.058828339,-0.0057860017,0.3084498942,-0.1026286483,-0.2543334961,0.0440540761,0.3627564907,-0.3065457642,-0.103482984,-0.0444494896,0.1358826905,-0.1913894862,0.0998171195,0.3135398924,-0.1987226009,0.2418410033,0.0600405857,-0.06561023,0.1117260531,-0.1427793056,-0.1940472722,-0.0809840113,0.2763290703,0.1291200817,-0.1630798131,-0.177103579,-0.4345960021,-0.1928695738,0.341162473,-0.1402751505,-0.221914947,0.0338350162,0.123443462,-0.3155338764,-0.2954392433,-0.2915062904,0.242105931,-0.1154171899,-0.3325147629,0.0051547447,-0.338324517,-0.2549366951,-0.0052289949,0.1019866019,0.317040205,-0.338414371,-0.0378375091,-0.0716530606,-0.1312973946,0.1759172082,0.1195302606,-0.1356283724,0.0528887697,-0.0810474902,0.3707363605,0.1033161879,-0.0800416991,-0.3151873052,-0.1095573083,-0.0785582811,-0.1270705312,-0.0561670698,-0.2193423212,-0.2982594669,-0.0450606719,0.2091311812,0.2197275311,0.0764507055,0.0485545918,-0.3125258684,-0.2832559049,0.277035594,0.1366285682,0.0436743051,-0.2376753539,-0.0044845329,-0.3707582653,0.0189564694,-0.1722556204,-0.2514601946,-0.0375154167,0.3037078083,-0.267026782,-0.0709287301,-0.0049370858,-0.251334697,0.0999912769,-0.198675558,0.3306252658,-0.2112036496,-0.101684548,-0.3360035121,0.2426897883,0.1568356603,-0.0757179931,0.0537428781,0.0139934346,0.1294900775,0.4455618858,-0.0342876799,0.1879517436,-0.0059918989,0.1484079063,0.2585426569,0.1822604388,-0.3008645475,-0.5154079199,0.2087287307,0.1776776761,0.0894742981,0.0034555825,-0.0724015683,0.1765432656,-0.009275347,0.0763942003,-0.2671247125,-0.2846678197,0.1067418158,-0.4681561887,-0.0044742413,0.0258461498,-0.0397189781,0.0368184112,0.3439415097,0.1939021349,0.086098358,-0.0027802992,0.1075394079,-0.0836933181,0.1214539632,-0.1025411636,0.1862477362,0.1555067599,0.1888844669,0.1804465204,0.3516589105,-0.081736885,-0.1929059774,0.0209464189,0.5004723668,-0.0104153566,0.3968636394,0.2594711483,-0.1391992867,-0.2249691039,0.1984295547,-0.2625118792,0.3533537388,0.1530496925,-0.0697853416,0.3267526031,-0.2691057026,-0.3553437293,0.1939434111,0.0401574075,0.2229404449,-0.0164994374,0.38502267,-0.000071901,-0.4463835061,-0.2158685774,-0.0241146572,0.2902513146,-0.195781976,0.2333467603,-0.2788990736,-0.4786580503,-0.3216550648,-0.3746164441,-0.2644890845,-0.1139181554,0.1861757487,0.0875777602,0.0209138468,0.2171658278,0.3065685928,0.253418535,-0.2405567765,-0.2795983553,-0.2933273613,-0.2630257607,-0.6036442518,0.0401455499,0.3904749453,0.1700224131,0.162003234,-0.168516472,-0.0609380081,0.4120291173,-0.5072560906,0.0628771558,-0.1883171499,0.3759168983,0.0254065637,-0.1615848392,-0.495667547,-0.4668341577,0.0776250064,-0.3616183698,-0.271908015,0.0947322994,-0.0823626518,-0.0533253923,-0.2126678675,-0.4288894832,-0.4334754944,-0.176218003,0.2912440598,0.1114313304,0.0875864848,0.4321451485,0.1826329231,0.3838966787,-0.0788138732,-0.0156474859,0.0568739362,0.157979548,0.1709013134,-0.1130349115,-0.2863180339,0.0840016007,-0.0555577464,0.1250379384,-0.1318465769,-0.5317515731,-0.2815971673,-0.0561024584,0.1225522831,0.2538729012,0.023475131,0.3396735489,0.0629492849,-0.0319777355,0.0732188746,-0.4039273262,0.033862669,-0.0514901988,-0.0284922495,0.0802505016,0.1063992977,-0.0677924603,0.7185683846,-0.130193755,-0.109185569,0.3972301483,-0.103979826,0.1648678333,-0.2360560596,-0.0234847609,0.1025646329,-0.0177805927,-0.1606339812,0.2749057412,-0.1622764468,-0.4740835726,-0.190534696,-0.1989635229,-0.2541110814,-0.2911300659,0.1263343543,0.3540947437,0.0426923856,0.0501411371,-0.1241014898,0.1096645296,-0.3338997662,-0.0762941316,-0.0461393669,-0.1030333266,-0.0380759351,0.0496417023,-0.4655984938,-0.456582427,0.3199999332,0.4237254858,0.2264307439,-0.0685278252,0.0034591206,-0.0761084333,-0.2204844803,-0.0725475252,0.0178992432,-0.0144482991,0.3299291134,0.1291982085,-0.0789631829,-0.1727335304,-0.328953445,-0.0892146379,0.3047033846,-0.0836103335,-0.3240818083,0.1537425667,0.4683853686,0.3481386602,0.2129237056,0.198785767,-0.3215835094,-0.3988630176,-0.1383432597,0.2397527248,0.0547616109,0.3466660082,-0.1757502556,0.2279440314,0.0948214382,-0.1361166388,0.0270532798,0.4200960398,-0.0073742648,-0.0413128957,-0.0127852065,-0.1898573488,-0.4006470442,0.123042658,0.6168137789,0.1352256835,-0.2412740141,-0.2904062569,0.1564190835,0.4171556234,0.1968608946,-0.0951201692,0.0853090137,-0.290874958,-0.1272576451,0.1797147095,0.0900782943,0.1424271315,0.134699896,-0.0143321985,0.13342987,0.6629952788,0.060598027,-0.0195607301,0.1840840131,-0.0039892681,-0.3596172929,0.494567126,-0.0165605601,0.9023059011,0.0052780369,0.1025133058,0.5774897933,-0.0107110897,0.3412154317,0.0877749398,0.2345594019,-0.4027553499,-0.098108463,0.0909472555,-0.1632783115,0.1461850256,-0.1311172992,-0.5301916599,0.2345527709,0.3508245945,0.0767901316,0.3780787289,0.1693213135,-0.1686767936,-0.008611735,-0.4974405169,0.0724884495,-0.0794616193,0.185856849,-0.2010641992,0.025602622,0.0126186553,-0.2772426903,-0.4413924217,0.3447722495,-0.4248201549,-0.1189629734,0.4366097152,-0.2345286906,0.2927358449,0.1134288982,0.2366983443,0.1430516094,-0.1989377588,0.2038395703,-0.1790079176,-0.2164798081,0.1933829039,0.1081499457,0.2353775054,-0.1960169822,-0.2463551611,0.1433787495,0.0282349661,-0.2993826568,0.2750101984,0.0317361318,-0.0479600467,-0.0497145168,0.0818249658,0.1291484982,0.1013898477,-0.1726247221,0.0994525254,-0.014350662,-0.0968640298,0.1849939674,0.0256235022,-0.2739040256,-0.085663408,0.3191238642,-0.1160248742,0.1227009296,0.6109801531,0.358225733,-0.0135364477,-0.1439747512,-0.1954408437,-0.1976335645,-0.3472584784,0.1349243224,-0.0893481299,-0.0829553679,0.2065331638,0.6358377337,0.0321331881,0.034259323,0.0322360024,-0.4378579855,-0.1001101956,0.2935982943,0.3192722797,0.0036186913,0.3327151239,0.2987152338,0.2375250459,-0.0031664311,-0.2429557294,0.1284068674,-0.2802470624,0.1713701934,-0.2991990447,-0.1453392953,0.2783918679,0.0976639464,0.0292557981,0.2056999058,-0.0773144215,-0.2213221788,-0.1646217853,0.1092192605,0.1179897785,-0.2578455806,-0.098903656,-0.066543974,0.2309484333,0.0404099524,0.0834897459,-0.1175380647,-0.2224590331,0.2460855991,0.0818295851,0.362999469,-0.1728776991,-0.035626743,-0.409432292,0.378341198,0.2551013231,0.2278321236,0.1126706079,-0.0049376879,0.1982796341,-0.0532479696,-0.1598259211,0.058852829,0.0549331158,-0.4088422954,-0.0978532583,-0.3386643529,0.1693819761,0.3265035748,-0.1732739806,-0.1791542917,0.4539534748,0.1184345931,-0.4062667489,-0.0042899908,0.1541658491,0.1295667738,-0.0078924587,0.2604048252,0.2204558998,-0.1430953741,0.0729009956,0.2042341381,0.2934803665,0.071089752,-0.1808232516,0.2278164625,-0.2106996924,0.0991223231,0.187372908,0.0473175682,0.4222980142,0.1560611874,-0.2856242359,0.209425807,-0.2741923332,0.2412849814,0.1257189214,-0.1985055059,0.0964618474,0.4508843124,0.290731132,-0.1011620387,0.2107511163,0.5228333473,-0.1316392422,0.0980874673,-0.185736537,0.2679846585,-0.0171091035,-0.1236460954,-0.0688001439,0.035530705,-0.1854599714,0.1501879096,-0.1470804662,-0.0095154298,0.3960517049,0.2103192508,-0.2445914,-0.0783877596,0.0146137411,0.1155375615,0.0676930994,-0.1753289998,0.3981393576,0.0947015956,-0.1871603429,-0.0908760875,0.6913028359,0.5602606535,0.1460423172,-0.3398172855,-0.1216662079,-0.3152953982,0.0107180187,-0.1049774066,0.2884894013,0.3598891497,0.4698311388,0.2505150437,0.0743763223,-0.0902576298,0.3026503921,0.1644702256,-0.1791979522,-0.0782844126,0.0311258715,0.1242984161,-0.1412472427,-0.3082651198,0.0276749469,-0.1040876955,0.3828437924,0.1828320771,-0.0276300665,-0.0220519099,-0.1087611765,0.042198807,0.2338199615,0.596929431,0.4660426676,0.5083932877,-0.3644014895,-0.1765926331,-0.5498260856,0.1105112806,-0.3375613093,-0.0191143081,0.0430709533,-0.1309378594,-0.1224307194,0.3655650318,-0.2373734117,0.432290405,-0.3409906626,-0.1812257767,-0.4165719151,0.0285931099,0.0152045824,-0.2957243919,-0.1329243481,-0.2455765903,0.2400437593,-0.6261115074,-0.0680845529,-0.112146236,-0.222458899,0.1780542135,0.110880062,0.441093117,0.1155429333,0.528813839,0.0060330592,0.0991595089,-0.1156327203,-0.3268683553,-0.1695924848,0.1565978825,0.0898521245,0.1825833023,0.002996421,0.2266043723,-0.0841833428,0.5815643072,-0.0408004485,0.242108047,0.2631011307,0.0312380008,-0.0804893225,-0.082791172,-0.0333433561,0.200013116,-0.0544318743,-0.109260112,-0.5326638818,-0.4132597148,0.3147661686,-0.5150575638,-0.0720607787,-0.205005914,0.0867945626,0.2675552666,-0.3100770414,-0.6111453772,0.0503031947,0.2234050632,0.0954609737,-0.4149084091,0.1335827708,0.289944917,0.0462747104,-0.0948521644,0.4303206205,0.0408821739,-0.0606977008,-0.234391585,-0.2921995521]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"Came to raise this issue, great to see other already have and it's being fixed so soon!\r\n\r\nAs an aside, since no one wrote this already, it seems like the version check only looks at the second part of the version number making sure it is >16, but pyarrow newest version is 1.0.0 so the second past is 0!","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":59,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\nCame to raise this issue, great to see other already have and it's being fixed so soon!\r\n\r\nAs an aside, since no one wrote this already, it seems like the version check only looks at the second part of the version number making sure it is >16, but pyarrow newest version is 1.0.0 so the second past is 0!","embeddings":[-0.3563594818,0.269282788,0.0403541103,0.0255090278,-0.1082258523,-0.2124165297,0.1698410064,0.2209273875,-0.14622657,0.1146730036,0.0956548974,0.5587813854,-0.2984774113,0.0380698033,0.0623061098,-0.0809185207,0.1304244846,0.3858693838,0.0610733591,0.1696226895,-0.2576373518,0.3070941865,-0.4419914484,0.2494819164,-0.2984107435,0.0139223505,0.0193617586,0.0773457959,-0.3677995801,-0.6848931909,0.4637121558,-0.1372096092,0.2255123407,0.3252450824,-0.000117499,-0.0126847932,0.4980806112,0.1090355963,-0.1948447227,-0.4525268376,0.0962289944,-0.4826610386,0.3875705004,0.0097497031,0.0457869023,-0.0264520142,0.3028815985,0.6757784486,0.2657588124,0.2176844776,0.1667813957,0.2213878483,0.4098852873,0.0257353429,0.4948376417,-0.2362098843,-0.1034501716,0.762858808,0.2007291913,-0.3635478616,0.0591918826,0.1609631181,-0.2842781842,0.0079968292,0.0868148655,-0.0093613975,-0.0282807872,-0.2803566754,-0.070423916,0.2334549874,0.2786686122,-0.3727610111,-0.2139852345,-0.0369421542,0.1685265005,-0.5201522708,0.2031876445,0.2061320841,-0.2479265481,0.2166700363,0.0423115343,-0.0363257118,-0.1677532941,0.1888074577,-0.1594609469,0.7550790906,0.0910478309,0.1258877069,0.2515863478,-0.0595551692,0.5646770597,0.2610051334,-0.0171777867,0.1957466453,-0.0181440879,0.0440122299,0.0082142614,0.5163198113,0.0293623358,-0.3448951542,0.0096516013,-0.0534603298,0.1265423,0.4474957287,0.0249658301,-0.063434042,0.0032090126,-0.125906378,0.3919048011,0.0218870789,0.1578956246,0.1030367613,0.0269265324,-0.3716893792,-0.0051645674,-0.0394150317,0.3415209949,-0.2093355209,-0.3028344214,-0.256716758,-0.5051163435,-0.0087984167,0.0819451138,0.1959880888,-0.1884271652,-0.0456597507,-0.013504128,0.1332431287,-0.2568027377,-0.1132037044,-0.0069235414,0.1749416292,-0.2337508351,0.0082895365,0.2488785535,-0.1568795741,0.2385914922,0.1839190722,-0.1583384871,0.1740048379,-0.0879792348,-0.1543875486,-0.1690748334,0.3463720977,0.0948102027,-0.0689558387,-0.154970929,-0.4418059886,-0.2765375674,0.3989022672,-0.2178521305,-0.3779511452,-0.2776358426,0.1390897632,-0.41422382,-0.1977523565,-0.2286471874,0.1975177526,-0.0626465976,-0.5185403824,-0.1106970683,-0.4185709655,-0.2588008642,-0.2167459875,-0.0355000496,0.2313151956,-0.4213174284,-0.0013216604,-0.205294922,-0.1577323377,0.2029190063,0.1241731644,-0.2396297157,0.0035150398,-0.0710430518,0.200630635,0.3293647766,-0.1011555046,-0.3482894599,0.0037729219,-0.124355115,-0.0796842873,-0.0503579341,-0.0146262608,-0.2075435221,-0.1026687846,0.111505717,0.2191842645,-0.0628921464,0.0518534742,-0.3201494515,-0.282230407,0.2409398407,0.1795413792,0.029073203,-0.1322246045,0.1251459122,-0.051167652,0.150691092,-0.1088017076,-0.1956821382,-0.0068311817,0.3765248358,-0.1861535609,-0.1213011965,-0.0252113603,-0.424626112,0.1397025734,-0.2081705332,0.2105352879,-0.181357637,-0.1387983263,-0.1585410833,0.2024338245,0.0917202979,0.0864932016,0.0574882589,-0.0538417473,0.1482982486,0.2236875743,0.0374069326,0.0488238707,0.0423657559,0.1162125692,-0.0075122435,0.2853103876,-0.3939154744,-0.5148022175,0.110420078,0.1565466523,0.1446731836,0.0030891907,0.0473880097,0.0985333472,0.0302920323,0.1519058049,-0.4095864296,-0.1629502177,0.1721079648,-0.436049968,0.0903435722,-0.0005215072,-0.0892995968,0.1381736994,0.3890042007,0.2938117683,0.1787322462,-0.0771885663,-0.0478576161,0.0068728905,0.0108244224,-0.0219762344,0.0616079159,0.0079064276,0.1270696223,0.3067643642,0.116550684,0.0130092064,-0.3182851076,0.1458008438,0.4736977518,0.0787649006,0.1612203717,0.1202903986,-0.0150473509,-0.2019604892,0.2824196815,-0.2004913539,0.236426726,0.2488380522,-0.0791958645,0.2862409353,-0.3474994898,-0.2463964373,0.066219762,0.0795558318,0.2478338629,0.0545605719,0.280064851,0.0440858603,-0.3395344317,-0.0405010656,-0.0013776888,0.3130808175,-0.1538745612,0.2280437797,-0.3657150269,-0.5152053833,-0.2934384942,-0.3485786617,-0.4190300107,-0.0455743335,0.0744508505,0.3852240443,0.0903392434,0.226042062,0.3749234974,0.0267463271,-0.2834899724,-0.2804327011,-0.1622307599,-0.3187162578,-0.5653555393,0.0051587559,0.2741716206,0.1464026272,0.1834054291,-0.1747495085,-0.0374336317,0.1914411783,-0.6024135351,0.0411879085,-0.2708110511,0.3899646997,0.1654329151,-0.0756173134,-0.4261088669,-0.3029899299,0.0499657393,-0.2672175169,-0.2311026752,-0.0775561035,-0.3112016916,0.1056579724,-0.2025433332,-0.0183191933,-0.5836764574,-0.2559036613,0.0728108585,0.0824737847,0.1966979802,0.3112359941,0.2778764069,0.204835847,0.0999180824,-0.1780395508,0.0744474828,0.0728753358,0.2003842145,-0.1788429618,-0.2410305887,0.2172353715,-0.1202084944,0.021112103,-0.0540615655,-0.4464507699,-0.4577817917,0.0401118957,0.3177597523,0.176334843,-0.0380546562,0.4169327021,0.0920343176,0.0213226434,0.0110069215,-0.4104071558,-0.2225304395,0.0307433084,0.195093289,0.1226700991,0.2151464373,-0.0781218708,0.6302813888,-0.2324964255,-0.1335863471,0.4108282328,-0.1130619198,0.2641091943,-0.0445308462,-0.0441889837,0.0038075019,0.0769339874,-0.0762546882,0.2126753181,-0.1427685022,-0.538187027,-0.2042012811,-0.2447575778,-0.2156720161,-0.1774834543,0.0836955458,0.4492432177,0.1635320038,0.1125485301,-0.1237434521,0.0148286475,-0.3518214226,-0.1838137805,0.0659230202,-0.1047117412,0.0882325172,0.3130837083,-0.5570732951,-0.3975996673,0.3536090553,0.4260613918,0.0035991524,0.0674597025,-0.058137536,0.0563075691,-0.12319576,0.0949182361,-0.0509752184,-0.0039711976,0.2375788689,0.1718689799,-0.2469349056,-0.1955090314,-0.1156431884,0.0992516056,0.2949614823,-0.0017342225,-0.26601547,0.0775737464,0.5640846491,0.4030723274,0.0865184739,0.0418523103,-0.4508181214,-0.4398842454,0.0007967193,0.0757178664,-0.0469195135,0.280264765,-0.1493751407,0.1977668256,0.1838614494,-0.1324869841,-0.1131210551,0.4778265357,0.0719999894,0.0505934954,-0.0745617598,-0.1453448981,-0.2305831462,0.3679276705,0.493519485,0.1361130923,-0.0544998161,-0.1077054739,0.3420845568,0.2212444246,0.1268036366,-0.051787328,0.0296374708,-0.1030545682,-0.2111224979,0.2338956743,-0.0035016048,0.0897081718,-0.0347037837,-0.2627060115,0.129774332,0.666051805,0.2021786869,-0.088226825,0.0257416982,-0.011450801,-0.1990587711,0.3481175601,0.0743565336,0.8326461315,-0.0258939564,0.0767496377,0.6967654228,-0.1195190176,0.6577323675,0.2338549048,0.3234875798,-0.351778537,0.0525329746,0.0392548516,-0.0763085037,0.1141756251,-0.1606550962,-0.4357738793,0.2853183448,0.2705795467,0.0429765396,0.3506116271,0.1500795186,-0.064796485,0.0263774544,-0.3548265398,0.0793812126,-0.1717133671,0.2368308008,-0.3579492867,-0.0989944711,0.0726469755,-0.3737101853,-0.4061584473,0.1980192661,-0.31195122,-0.085991323,0.2118489891,-0.4263009429,0.0714577436,0.0048069241,-0.0199049786,0.0301708002,-0.1736846566,0.2135380805,0.0277872533,-0.2682162523,-0.0950965732,0.1593033522,0.1995506436,-0.1809952557,-0.1806913912,0.1141969636,-0.0766462386,-0.1602272093,0.1103007719,0.2179468572,-0.1576022059,0.259026885,0.0707157701,0.0208603106,0.2043578774,-0.1548306644,0.1161689833,-0.0316252783,0.0968471318,0.1447460651,-0.1322196573,-0.2569980621,-0.1296866536,0.2917925715,0.0648808032,0.1338762939,0.4178504646,0.2300008684,-0.0808648318,-0.0924579576,-0.190721646,-0.2852126062,-0.3821772635,0.0474342816,0.1602156013,0.1208622456,0.2363419682,0.4325951636,0.1759969443,0.0349692255,-0.0027753459,-0.3610956967,-0.1583451629,0.3478938043,0.111884743,-0.0918346345,0.2412369549,0.4302352071,0.2563734949,-0.2707996964,-0.2438819557,0.2088712752,-0.0774837062,0.1995982677,-0.1697113365,-0.0384780429,0.0325778984,0.0321033821,0.0391288213,0.1034985408,-0.0770381391,-0.198640734,-0.1042039022,0.1216510534,0.1287732124,-0.2363008261,-0.1117767617,-0.1011826098,0.0765416473,0.1567929685,0.0715915039,0.0031398151,-0.1735082716,0.1497431099,0.1444462985,0.2842686176,0.0129265692,0.1131559312,-0.2149542719,0.4630050659,0.2806185782,0.3607988656,0.1898021847,0.0785512105,0.2958897948,-0.150378108,-0.2741986811,-0.1127204075,0.0156949926,-0.3167326152,-0.0584833473,-0.3318295181,0.2149787545,0.3215718567,-0.0776570365,-0.187449351,0.5202891827,0.1215229034,-0.4359510541,0.0913804099,0.19158566,0.1452506036,-0.0492452569,0.2615137696,0.2311929762,-0.0020171981,0.2134942114,0.2170177698,0.2639148533,0.2480226159,-0.0180776827,0.253719449,-0.1021510661,0.1275254786,0.2133232206,0.190853402,0.3731150627,0.0213820152,-0.2727660239,0.1267822832,-0.2584782541,0.0701356381,0.015274018,-0.4455841482,-0.0612625778,0.329721272,0.2798146904,-0.2126589715,0.2182136029,0.3516299427,-0.0680863559,-0.0510080196,-0.1703102887,0.3314247429,-0.1020059288,-0.0153345242,-0.0075727324,0.0484619476,-0.1511905044,0.2374203354,-0.063115567,-0.0850854814,0.1830561012,0.2144213915,-0.2929424047,-0.1625159383,0.0525893942,0.2145334929,-0.0869381428,-0.0650166944,0.3325576484,-0.130939126,-0.3034469485,-0.2252887785,0.4857273996,0.5330047011,0.3011166453,-0.2612904906,-0.2222591937,-0.2724920511,0.0934620202,-0.1609717906,0.0910235122,0.401037842,0.4848851562,0.2285537422,0.0634376928,-0.1281144321,0.3109259903,0.1588606536,-0.0454526246,-0.0872867107,0.0489241444,0.0879374817,-0.0846334472,-0.1388156414,0.1435378939,-0.2690296173,0.4802122712,0.1894149482,0.2815273106,-0.0888153315,-0.1190420166,0.0334636122,0.120550774,0.5594960451,0.493165493,0.570435822,-0.1983357221,-0.2719709277,-0.617634654,0.1832291186,-0.5359067917,0.217970863,0.0753013939,-0.0515671223,-0.2211536169,0.3695656657,-0.2084883749,0.1653892845,-0.1704200059,-0.1918792576,-0.2623123229,-0.0385404564,0.0660195127,-0.3587390482,-0.0989603028,-0.2771787643,0.1449618489,-0.760283649,-0.0124444226,-0.3348179162,-0.2329118997,0.0799503475,0.1844256371,0.5588228703,0.1272923201,0.4683969319,0.1397472918,0.016538525,-0.2916190624,-0.2619532049,-0.3168052137,0.2344325185,0.093042776,0.2211835086,-0.1172395274,0.1349559426,-0.2433464974,0.47950688,-0.1296809912,0.1860696077,0.2657738626,0.0335793234,-0.1438556463,-0.0505993702,-0.2359084338,0.0007025003,0.0278323926,0.0293656308,-0.560693562,-0.3480979502,0.3855327964,-0.4710200727,0.0295958575,-0.2068922222,0.1031734571,0.1791056991,-0.3120688796,-0.3964773118,0.1284188628,0.280356288,0.1081899852,-0.3001506627,0.1338803023,0.0365003608,-0.0632793903,-0.0691587627,0.2066374421,-0.0287695266,-0.2944736183,-0.1139807478,-0.2467910945]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"> Indeed, we\u2019ll make a new PyPi release next week to solve this. Cc @lhoestq\r\n\r\nYes definitely","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":17,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\n> Indeed, we\u2019ll make a new PyPi release next week to solve this. Cc @lhoestq\r\n\r\nYes definitely","embeddings":[-0.3427734375,0.2993348837,0.0278658252,0.0008248843,-0.0533736497,-0.2322781533,0.2582436204,0.1790954322,-0.2328974456,0.1098069698,-0.0035005987,0.6347985268,-0.259023875,0.1106953695,0.1111482233,-0.0966253504,0.1224771291,0.3842332661,0.0320785642,0.1150280535,-0.2321145833,0.3013066947,-0.4014840126,0.2736143768,-0.3001113236,-0.0178907402,-0.0336268581,0.0969627202,-0.2867270708,-0.663888514,0.4799574316,-0.0461135171,0.2491188794,0.2177296728,-0.0001120963,0.0018571639,0.4446177185,0.0588436872,-0.1980582774,-0.4029640257,0.0341191813,-0.4738245308,0.4145335853,-0.0984018818,0.009839911,-0.0233778544,0.2621846795,0.6624282002,0.3113102913,0.202608943,0.2112783194,0.2363433391,0.3234765232,0.0074101943,0.4227241278,-0.1575131118,-0.0963935852,0.7648112178,0.1429357082,-0.3419243395,0.0089764548,0.1031950638,-0.2788573802,0.0289603621,0.021922132,0.0070117442,-0.1472040117,-0.307143867,-0.063910909,0.2735510468,0.2462306172,-0.369807303,-0.1704452038,-0.0098759793,0.1577402055,-0.5250217319,0.1430397928,0.2469917387,-0.3028105199,0.2058746964,0.1286191195,-0.1188331693,-0.1860572994,0.1715101898,-0.1171362326,0.7375593781,0.1335549504,0.115255326,0.2805883288,-0.0567426905,0.540766418,0.3088673055,-0.0792701244,0.1202857196,-0.0427252576,0.0211688671,0.0225457773,0.3409097195,0.0249087419,-0.2473079115,0.0255866759,0.0225669984,0.1256219298,0.4503282309,0.0599614382,-0.0484906733,-0.0453351624,-0.0046340348,0.4267911315,-0.0066436813,0.1369570494,0.0808409229,-0.0122739682,-0.2847607136,-0.0170374289,-0.0676540509,0.2775869071,-0.1611774117,-0.3963501453,-0.2821988463,-0.435975343,0.0238579642,0.0372934192,0.1950220466,-0.1200190112,-0.1023528725,0.0121809551,0.1517194808,-0.2696354389,-0.0800236538,-0.0334360711,0.1952785403,-0.2237413228,0.0621654652,0.2317143083,-0.1511159539,0.2136739492,0.1638038158,-0.1062263176,0.2310115993,-0.0294301063,-0.1341701299,-0.1386497468,0.2887904346,0.1331017762,-0.0839099661,-0.1724686921,-0.3421243429,-0.26027897,0.3834969401,-0.281029433,-0.34865731,-0.3057534695,0.1715805233,-0.3623221517,-0.2298499197,-0.2415509671,0.1925079226,-0.1000900492,-0.5023195744,-0.103967838,-0.352139771,-0.0717574656,-0.201839149,-0.0292974189,0.203484118,-0.3529725671,-0.0353136845,-0.1182195693,-0.1545044184,0.2850580513,0.2002337426,-0.2665144205,0.0042171464,-0.113277331,0.2782341242,0.2827667296,-0.078201808,-0.3504206538,-0.0246291589,-0.0884435028,-0.1190676168,-0.056546621,-0.0112745231,-0.0793270618,-0.0798390061,0.1883729994,0.228994295,-0.0206914451,0.0899544433,-0.3444251418,-0.3358956873,0.2805465162,0.1920298785,0.0722375363,-0.1789543182,0.0699494854,-0.0459714234,0.1574608237,-0.1699227095,-0.1814156622,0.0086773662,0.418684721,-0.1727122962,-0.1212951094,0.0149611123,-0.363623023,0.0653887987,-0.2499880642,0.3446653783,-0.168976739,-0.1383545101,-0.0968618467,0.1540222764,-0.0032698438,0.044844158,0.1141757965,-0.0692190751,0.0930160359,0.230772078,0.0139852418,0.0700308532,0.0974470749,0.0902948454,0.0755294412,0.3200631142,-0.3432150185,-0.5013923645,0.1558911055,0.1919502318,0.0954931602,-0.015271429,0.0595762432,0.0701608732,-0.0414922535,0.1012136936,-0.3247621655,-0.1853398979,0.2462165505,-0.4993384182,0.0803937092,0.0131683946,-0.0752409548,0.1265430897,0.4028690457,0.3070367277,0.1803319901,-0.0566328913,-0.0034645754,0.060174834,0.15966627,-0.0279672854,0.0723464265,0.0341917388,0.1038167626,0.3129440248,0.1196284741,-0.0223808829,-0.3551903069,0.1838106662,0.4711066186,0.0623613261,0.2583274245,0.2070450187,-0.0968046412,-0.1911536008,0.2963224947,-0.2238921374,0.2534521222,0.2617909908,-0.0768608302,0.2503487766,-0.3105472326,-0.3091457188,0.0943920091,0.0421279073,0.1733682454,0.0248517804,0.3159976006,0.0110517154,-0.3823555112,-0.0541484468,0.0020076446,0.2897807658,-0.1607694328,0.2292197645,-0.3812175989,-0.5844123363,-0.2033057958,-0.3761793077,-0.3669314086,-0.0743166432,0.0838027894,0.310374558,0.0543873236,0.2557633817,0.4026352465,0.0697187334,-0.2327948362,-0.2434329242,-0.2093127072,-0.4130916893,-0.5413900018,0.0675750971,0.2533719242,0.1499053538,0.2001347244,-0.1714408398,0.0169509649,0.1043809354,-0.589812994,0.1002970487,-0.2247340679,0.3380070925,0.1060066372,-0.0794158205,-0.3944554627,-0.3828268647,0.1079265326,-0.2567720413,-0.1813481152,0.0085919993,-0.3333748877,0.0928751379,-0.2160847336,-0.1032108814,-0.6534888744,-0.2656674087,0.1510529071,0.1169588268,0.1299255788,0.3539073765,0.2291953713,0.1824577004,0.0490744039,-0.1751547903,-0.0114472955,0.0290607493,0.2000274956,-0.2042187601,-0.2945664227,0.1472581327,-0.1673745513,0.0931809545,-0.0429615006,-0.4215467572,-0.4417282939,0.085579209,0.3840212226,0.1543041021,-0.0067445892,0.3419800699,0.0956139043,-0.0341994241,0.0839712992,-0.3151020706,-0.0966725573,0.1171037853,0.1417523026,0.0434586443,0.1518180668,-0.1093009561,0.6300576925,-0.2139459401,-0.1077207401,0.4509525597,-0.1072049215,0.3438112736,-0.0944144279,-0.0752923638,0.048007112,0.0440239683,-0.0963850319,0.1973455548,-0.2002141178,-0.5494316816,-0.1968811005,-0.2261254638,-0.179066509,-0.2719097733,0.0905909017,0.4456666112,0.1552732587,0.1197678447,-0.1132985726,-0.0142350718,-0.3394672275,-0.1343715489,0.0364560075,-0.101661399,0.0346686468,0.2814468741,-0.5303834081,-0.5247498751,0.250290513,0.3806584477,0.0215020124,0.0081431679,-0.0626109391,0.0872104093,-0.1062912419,0.0684470907,0.0922199562,-0.0971359089,0.2445694506,0.2222717106,-0.2340100706,-0.1920470148,-0.1934928745,0.1639661789,0.3089673221,0.0047291126,-0.3216294348,0.0847145319,0.5238486528,0.3645200729,0.0546268411,0.0535175353,-0.3809590638,-0.5595187545,-0.0602215864,0.0733549297,-0.0392752513,0.2775353491,-0.1529998928,0.1538328081,0.1033133492,-0.1528432816,-0.104369916,0.4947182238,0.0936842188,-0.0129777007,-0.0004545523,-0.138958171,-0.288795501,0.39499861,0.4543976784,0.1464978307,-0.0440171845,-0.0468860604,0.2194838077,0.2195369154,0.0662064701,-0.0709814653,0.1114895493,-0.2176203877,-0.1598270535,0.1562422812,0.0562020876,0.0816565678,-0.0320807397,-0.2292525321,0.0558033548,0.6698864102,0.1846267134,-0.0693131164,-0.0279192291,-0.0430279374,-0.2264950573,0.3169512451,0.1045504063,0.8228833079,-0.0489419959,0.1441760957,0.7622259259,-0.0296684075,0.6805464029,0.1497793645,0.3072378337,-0.3835254908,0.0588020682,0.072825484,-0.060990002,0.0731468871,-0.1455001831,-0.3930557072,0.2606292963,0.2867897451,-0.026475789,0.2877852917,0.1419994384,-0.0269435979,-0.0531502515,-0.3452907503,0.1084247306,-0.1777224243,0.2079711407,-0.3530268371,-0.1281893551,0.0871096477,-0.3031850755,-0.3591161668,0.2581678331,-0.2969035208,0.0468782671,0.1742727011,-0.4596208036,0.0165950488,-0.0784514248,-0.0620592535,-0.0371708982,-0.1562341899,0.1665721238,0.0004108604,-0.3037151098,-0.0537541285,0.2256669253,0.2189044058,-0.2000871748,-0.1893266886,0.1877377033,-0.0495269746,-0.1875086874,0.1061261445,0.2045487165,-0.1142987534,0.1869370341,0.0794345587,0.0730263367,0.1906563342,-0.1356146336,0.1421960592,0.0207579006,0.077823855,0.1563089341,-0.0535741039,-0.1859851331,-0.164383322,0.2252757996,0.0967029706,0.1861312985,0.4558210373,0.1741258055,-0.083238773,-0.1470932662,-0.1463324428,-0.3217686415,-0.3653511703,0.0773964524,0.2588824928,0.1288864166,0.2183810025,0.3914041519,0.1139288172,0.0233532619,-0.0448608845,-0.3423527777,-0.219738543,0.3276678622,0.0574603267,-0.0691528916,0.2971717715,0.424074918,0.1978619844,-0.263489306,-0.2902501822,0.1804887354,-0.1675817668,0.16671139,-0.11062067,-0.1134628505,0.0569197237,0.0713071302,0.0857152268,0.1060400233,-0.0339439437,-0.2348019779,-0.0859385952,0.0918136463,0.1614659131,-0.2099704593,-0.1105784774,-0.1162215099,0.0552263185,0.1327283829,0.0840237811,-0.0290759932,-0.2390405536,0.1647332311,0.1377286315,0.2250225991,-0.0005176949,0.1514251232,-0.1787545234,0.4222078025,0.2305238545,0.3370660543,0.1093783975,0.098259367,0.3143340945,-0.1214888319,-0.157128334,-0.1349676698,0.0401114412,-0.3493875861,-0.0280441884,-0.3854213357,0.1530694962,0.3556878269,-0.0881109014,-0.1875224113,0.5486344695,0.161155045,-0.4660819471,0.0523325019,0.0850412026,0.0544780493,-0.0468836538,0.2423156649,0.2471879423,-0.0384107493,0.2116467804,0.1933163106,0.2679802179,0.261572212,0.0040839906,0.227621153,-0.1358465254,0.1243592873,0.2575100064,0.1837988943,0.3727381825,0.0371414125,-0.3018246591,0.2114650756,-0.2427956462,0.0761604533,0.0475604162,-0.4293700755,-0.1298656613,0.2108361721,0.2840987444,-0.1893675774,0.1755264699,0.4341075122,-0.1155914962,0.003204158,-0.1722661406,0.4161432385,-0.1437201053,-0.0262854509,0.057835076,0.0214267317,-0.06301561,0.2276252359,-0.1368113905,-0.1164047867,0.2230812013,0.1955062151,-0.1940458864,-0.1434903145,0.0630196556,0.2389404476,-0.1176962033,-0.0176228434,0.2848542035,-0.0744715706,-0.3305772841,-0.2518534362,0.5178362727,0.557028532,0.2944157422,-0.3138181865,-0.2281095237,-0.2676800489,0.0981898457,-0.1367557198,0.0734225139,0.4513367414,0.4381735027,0.18123658,0.1011427715,-0.1797259301,0.2612932324,0.2013241202,-0.0368178971,-0.0303330496,0.0452713296,0.0853581056,-0.1082077473,-0.1608337909,0.009868254,-0.2683596611,0.4186890721,0.210688144,0.216628015,-0.1077706441,-0.1446011215,0.0757286847,0.1124672294,0.6587206721,0.4152352214,0.5267051458,-0.2332031131,-0.2329095155,-0.5888140798,0.2068114132,-0.4867604673,0.0917184502,0.0359305739,-0.0490406565,-0.1855459064,0.3980052769,-0.2160431594,0.3255217075,-0.1239702851,-0.2867915332,-0.3391119242,-0.027101811,0.0189101715,-0.3432717323,-0.0960095972,-0.2429418862,0.1460620165,-0.6712964177,0.0536420643,-0.3101531267,-0.1983036399,0.0495859534,0.1595621854,0.5663316846,0.1240320355,0.5083733797,0.1158211753,0.0103747854,-0.1272680014,-0.3805966973,-0.2422290593,0.2114135921,0.1376100481,0.1287819445,-0.0762550011,0.0794652626,-0.2455943078,0.4317283928,-0.1461814195,0.1392713785,0.2135460675,0.1006486714,-0.0860767514,-0.0981786475,-0.2094151378,0.0486903936,-0.0006399192,-0.0257482976,-0.5031926036,-0.347979188,0.357809484,-0.4954043031,0.0160116833,-0.1722183824,0.0751730353,0.075884372,-0.2522554398,-0.3977367282,0.16558294,0.2635034919,0.0735246539,-0.2914824486,0.1351567209,0.0072117564,-0.0872236341,-0.1138577238,0.2276299,0.0252945293,-0.285498172,-0.1105776429,-0.1900275946]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"please fix this on pypi! @lhoestq ","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":6,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\nplease fix this on pypi! @lhoestq ","embeddings":[-0.3320481479,0.2817124128,0.0301262271,0.0110935755,-0.0485273451,-0.1780888438,0.2379115224,0.1709772795,-0.2266886234,0.1266423613,0.0336528942,0.6072169542,-0.2527963817,0.0827801824,0.092621766,-0.0979538932,0.0890565068,0.403360635,0.0383878909,0.1441511661,-0.2198723108,0.2873290479,-0.3952676356,0.2613079548,-0.3176207244,-0.0286881104,-0.0266382806,0.0707770884,-0.3138790429,-0.6541877389,0.4775688946,-0.0730054379,0.2595072091,0.2346879691,-0.0001144796,-0.0287169218,0.4415752292,0.0874760896,-0.2154026628,-0.4391254485,0.0427005962,-0.4218397439,0.4071692526,-0.0633337051,0.0217973944,-0.0450599492,0.2617107034,0.6591467857,0.2778630555,0.2163621485,0.1942873448,0.1619849205,0.3389591575,0.0137910135,0.4062121212,-0.1871497929,-0.0639893636,0.8014297485,0.1382646263,-0.3499383628,0.0282458942,0.1028868034,-0.2970845699,0.0087817861,0.0500458479,-0.033588998,-0.1125421077,-0.3072715998,-0.0108610718,0.2341384441,0.2729250193,-0.3839715719,-0.1289320886,0.0120251523,0.1903516501,-0.5446801782,0.1637243032,0.2571639419,-0.2577157319,0.2086044699,0.0944617316,-0.1295763552,-0.2045145184,0.1871485114,-0.1379263401,0.7242237926,0.1424137652,0.1155010983,0.2904059291,-0.0444803052,0.5801090598,0.300458014,-0.0671084151,0.1509382576,-0.0421698354,0.0529290624,0.0135063985,0.3853174448,0.0302327052,-0.285558641,-0.0112725077,-0.0340483785,0.1237836108,0.4569376111,0.0424689315,-0.0169415865,-0.0229323152,-0.0266365465,0.3927304149,0.0067411913,0.1196044385,0.0808126479,-0.0315130651,-0.2890917361,0.0065975185,-0.0342234373,0.2834529281,-0.183740288,-0.3864302039,-0.2849137485,-0.4678107202,0.0160089359,0.0706105754,0.1921148598,-0.1460038126,-0.1224298775,0.0076408358,0.1712342352,-0.2409031093,-0.0904111415,-0.0309131369,0.2000102848,-0.2328871191,0.0608400851,0.2374371439,-0.1791876107,0.2433292717,0.1934903413,-0.1447771341,0.2000276893,-0.0480369925,-0.161394313,-0.2028790414,0.2996637523,0.1580535024,-0.0522723049,-0.1558363736,-0.3435918391,-0.2444770932,0.4212977886,-0.3076256812,-0.390491277,-0.3110674918,0.1631534249,-0.3572823107,-0.2305000424,-0.2651433945,0.2193196267,-0.1011203527,-0.5057119727,-0.1292100996,-0.3628820181,-0.1246776879,-0.2241811603,0.0047467216,0.2289558649,-0.3822821081,-0.0152183855,-0.14727588,-0.1484618932,0.2979254127,0.1767735332,-0.2861637175,0.0288878959,-0.1104167998,0.2749851048,0.2838401496,-0.064680241,-0.3805441856,-0.0178082604,-0.1133292094,-0.1332639456,-0.0561575294,-0.0371404067,-0.1043555662,-0.0735014156,0.1902284473,0.2329483181,-0.0419718474,0.0789868087,-0.3438206017,-0.3392307162,0.3109975457,0.1732291281,0.0692551136,-0.1387917548,0.0794301778,-0.0623416789,0.1550213099,-0.1574360877,-0.1803420931,0.0281575229,0.4310985804,-0.1568671912,-0.1277092248,0.0185697004,-0.3928892314,0.0703802258,-0.2444662005,0.3250611126,-0.2168994993,-0.1164141819,-0.1501968503,0.1525542438,0.0073091132,0.0603860095,0.0961880833,-0.0827642083,0.0905088633,0.2343099266,0.0600393303,0.0631409809,0.0743818209,0.0905898064,0.0654616877,0.3358879089,-0.3752911985,-0.5196611285,0.155921936,0.1774039418,0.1624408662,0.0044348212,0.0312973745,0.078834638,-0.0209331103,0.1231635883,-0.325115025,-0.1590761542,0.2109868228,-0.5183627605,0.0904685035,0.0000332055,-0.0959229097,0.1484887898,0.4200341105,0.2888303995,0.2021558732,-0.0538087003,-0.0491396859,0.0692025423,0.1188502982,-0.0105703771,0.1027378812,0.0345456898,0.1430624574,0.3152592182,0.1653927416,0.0120029086,-0.3239170015,0.1777620614,0.4529214203,0.0772031546,0.2329765856,0.1938358098,-0.0884844884,-0.1967743337,0.2992303371,-0.2514206171,0.2378437966,0.2462808341,-0.0999266729,0.2681447268,-0.2982399762,-0.2903712988,0.0896327496,0.0163387172,0.2314651906,0.0201823954,0.3028454483,0.0333276428,-0.3581202626,-0.0848190933,-0.001651174,0.3490800858,-0.1380659044,0.2383041084,-0.396974802,-0.545869112,-0.2252631783,-0.3626281321,-0.3782647848,-0.0382973924,0.0585088693,0.3481204808,0.1099945977,0.2627291381,0.44581756,0.0722018108,-0.2083622515,-0.2242494226,-0.2016362995,-0.4081453681,-0.5487499833,0.0330487564,0.2287818491,0.1527715027,0.1825540811,-0.1448489726,0.0206589829,0.1273085177,-0.5970567465,0.0528707914,-0.2211499661,0.3400672674,0.1100757718,-0.0285058282,-0.4174223542,-0.3705472946,0.0898988917,-0.2516132891,-0.1684273779,0.0181697626,-0.3617891669,0.1139134467,-0.1770281494,-0.0696516931,-0.6192202568,-0.2564225495,0.1124249175,0.0770190731,0.1585927457,0.3335022926,0.2571108937,0.1815475523,0.05538252,-0.1599035561,0.0056870407,0.0279747918,0.1779528409,-0.1909044236,-0.2726352215,0.1748112291,-0.1310334057,0.0650406852,-0.031600982,-0.447923094,-0.4450224936,0.0544336401,0.3355758786,0.1147353128,-0.0258071851,0.344833672,0.0883175656,0.0068650735,0.0748575255,-0.3745406568,-0.1536622196,0.1035745442,0.1535336375,0.0729036033,0.1991791129,-0.0859895498,0.6228697896,-0.1892671138,-0.1086880267,0.4445578754,-0.1038327441,0.2960375249,-0.0503575504,-0.1091089547,0.0407840572,0.1045360565,-0.0821955651,0.2080596238,-0.2098469287,-0.5351960659,-0.1836679727,-0.2625721097,-0.1920429915,-0.2417090386,0.0876422897,0.4049643278,0.1881648302,0.1062601879,-0.1401178539,-0.0498407967,-0.3529387414,-0.1381676346,0.0670473278,-0.1152457595,0.0828247741,0.2905066609,-0.5542588234,-0.4815495908,0.2567614615,0.359621495,0.0328165181,-0.0069608693,-0.0074993856,0.1057544574,-0.077580817,0.1286641657,0.0616826341,-0.0679086968,0.1925225705,0.24382101,-0.2166615874,-0.1853104383,-0.1532868445,0.1841037422,0.288924098,0.0177684855,-0.320263207,0.0848520473,0.5252376199,0.3679187596,0.0509608239,0.0223020632,-0.4036962688,-0.5420247316,-0.0369572751,0.0622095503,-0.0918972716,0.3151996434,-0.1380312294,0.1663335711,0.13536641,-0.1780836135,-0.1175388172,0.4792303443,0.1053834781,0.032434646,-0.0333531648,-0.123543717,-0.2767365873,0.4001649618,0.4419287443,0.1498600692,-0.0601862073,-0.0434998907,0.2538067997,0.2101402283,0.0473470502,-0.0876811221,0.1246773601,-0.2067323625,-0.1567877978,0.1996298134,0.0434380732,0.0942326039,-0.0556794666,-0.2498163581,0.0662200227,0.6576656103,0.1666472107,-0.0610767044,0.0014521186,-0.0257755108,-0.215504542,0.3052996695,0.0667100847,0.8355257511,-0.0562452003,0.087450318,0.7444522977,-0.0531047992,0.6570123434,0.175763607,0.3068867922,-0.3783561289,0.0366470739,0.0759513825,-0.0900053829,0.0821720213,-0.1295401007,-0.4056269228,0.2547568977,0.2515688241,-0.017349273,0.3358162045,0.1586815715,-0.0253298003,-0.0119890645,-0.3468250334,0.094624266,-0.2169257551,0.2438051254,-0.3667936921,-0.0848563761,0.1031809598,-0.2905519009,-0.3783318698,0.2467745394,-0.3241496086,0.0013965456,0.1664021909,-0.4569426477,0.0523822792,-0.0512468107,-0.0351974666,-0.007183325,-0.1540660858,0.1728886962,0.0232144203,-0.3533211648,-0.0747759417,0.234646976,0.186414659,-0.197587207,-0.2079911977,0.1978513896,-0.0289138593,-0.1766922325,0.1067812443,0.1493922323,-0.1682521254,0.1867703497,0.0801767334,0.0428590998,0.2330792993,-0.1496205926,0.1304801852,0.0305979829,0.0701765344,0.1393112689,-0.0699929371,-0.2023855448,-0.1440339983,0.2309901714,0.0933561996,0.160588026,0.4622556567,0.2105862051,-0.0471155643,-0.1312143505,-0.1612441242,-0.3102467656,-0.3686894178,0.0672772452,0.2076790184,0.1575208753,0.226973772,0.3839343488,0.1332731098,0.045235455,-0.0067147245,-0.3664602935,-0.2168043703,0.3325587809,0.0537737086,-0.0714059472,0.3005398214,0.4507534504,0.1827981174,-0.2251107395,-0.2731246054,0.2209883928,-0.1596541703,0.1991370618,-0.1161480471,-0.1185980886,0.0387605131,0.083762072,0.0766762346,0.0734541565,-0.0404772423,-0.2169432491,-0.0952584371,0.1005292758,0.15716061,-0.1980766058,-0.1473148465,-0.1564154178,0.0511295684,0.159144327,0.0821215734,-0.0323901996,-0.2101290226,0.1511922777,0.1731474251,0.2180822343,-0.0458510406,0.1305869371,-0.196436286,0.3801576793,0.2682334483,0.333037138,0.1261876523,0.0831816569,0.2735144496,-0.125233084,-0.155693844,-0.1139890403,0.0073995129,-0.3610028625,-0.0025461961,-0.357742101,0.1578023583,0.3090903461,-0.0843329132,-0.1668236256,0.544510603,0.1481405348,-0.4392239153,0.048808787,0.1346322447,0.0613383763,-0.0532224663,0.2433885336,0.2521725893,0.0159259103,0.1941553354,0.1918105781,0.2741142511,0.301427871,-0.0098928316,0.2519435585,-0.1200071871,0.0960267037,0.2599243522,0.2003871351,0.3566608429,0.0252908152,-0.2993742526,0.1748059541,-0.2758904696,0.0322961956,0.0400511399,-0.4683256149,-0.1265962273,0.2341782898,0.2764541507,-0.2540013492,0.1420532465,0.4402348995,-0.1194262132,-0.0299873035,-0.1287078261,0.4038154483,-0.152230382,-0.0061698169,0.0138520543,0.0310498793,-0.0961542353,0.2181879133,-0.1273263246,-0.1211310998,0.1941441447,0.1976831406,-0.2479868978,-0.1705807596,0.0463035107,0.302598387,-0.1250217557,-0.0070109065,0.2981676757,-0.1368466914,-0.3505035937,-0.2511793673,0.4918399155,0.5318293571,0.3120690882,-0.3051469326,-0.2263713777,-0.2126285136,0.0677789152,-0.1019600332,0.0400037095,0.4733408689,0.4592013359,0.1514759809,0.0856445655,-0.1670408249,0.2590636015,0.2280768603,-0.0312270094,-0.0272292569,0.00973251,0.1108684316,-0.1122852266,-0.1559614539,0.022951968,-0.2900555432,0.422748059,0.2252870649,0.2223844677,-0.1171848327,-0.1264101863,0.0638455376,0.1103141084,0.6077866554,0.452267915,0.5414161086,-0.2855110466,-0.2989674807,-0.5766744018,0.2322793603,-0.4798608422,0.112442866,0.0725866333,-0.0638449192,-0.2267210484,0.3638674319,-0.218202427,0.3325029314,-0.1215904206,-0.286121428,-0.3231861293,-0.0379456915,0.0180741623,-0.338488251,-0.1183649451,-0.2133441865,0.1482466459,-0.6671681404,0.034820009,-0.3417227864,-0.198007673,0.0317274854,0.1643997133,0.5547747612,0.1122186705,0.5083244443,0.1475777775,0.0409411974,-0.1581236869,-0.335891366,-0.2705181241,0.2115857154,0.1332966387,0.1497546732,-0.079511568,0.0995357335,-0.2784188688,0.4654169977,-0.1238263473,0.1497159153,0.1955623627,0.1179170907,-0.0865691677,-0.0899244919,-0.1830035597,0.0222242568,0.0276754554,-0.0111777615,-0.5230873227,-0.3275964856,0.3504337966,-0.4665641785,0.0144600449,-0.1743173748,0.1007467434,0.0894898027,-0.2533599436,-0.3832131922,0.1760013849,0.2804470956,0.0730417743,-0.3110176325,0.1674868912,-0.0015841686,-0.0431945585,-0.0762822628,0.2048652768,0.0373925418,-0.2844824791,-0.1205012798,-0.1761244088]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"Is this issue fixed ?","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":5,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\nIs this issue fixed ?","embeddings":[-0.3575034738,0.2892061174,0.0456804708,0.064275682,-0.0771372244,-0.1734333485,0.197161302,0.1413282752,-0.228644982,0.0768760741,0.1149672046,0.5862120986,-0.3371917903,-0.0051665124,0.0758495852,-0.0694565773,0.0969063714,0.4063760936,0.1008527651,0.1609080583,-0.2169827968,0.3168114424,-0.4926137626,0.241909042,-0.2846109867,0.0501382574,0.0090115648,0.027887769,-0.2952517271,-0.6667721272,0.5000610948,-0.068113625,0.2599525154,0.2559458613,-0.0001221248,-0.0284017846,0.5456049442,0.0937014818,-0.2268278897,-0.4487872124,0.0485353768,-0.4237070084,0.3674147725,0.005163237,-0.0125052454,-0.0316512175,0.3543642461,0.6228750348,0.1965484321,0.1795178056,0.1283401847,0.113047637,0.453461051,0.1078887656,0.4608507752,-0.1679528505,-0.0800262541,0.813341856,0.1815494597,-0.3897451162,0.0436508022,0.1132529974,-0.2651978135,-0.0337815955,0.1009841636,-0.0432643518,-0.0041639684,-0.2128920853,-0.0349184424,0.231760487,0.2926416099,-0.3439249098,-0.2487223744,-0.0406729616,0.2007116079,-0.5399296284,0.1903978884,0.2616497278,-0.2301025391,0.2439850122,0.0619698614,-0.2065881193,-0.1891460568,0.2146665603,-0.204426825,0.7672876716,0.1160664111,0.1670651287,0.2166068107,-0.0135655943,0.5948577523,0.2714275122,-0.0455371663,0.1692885906,-0.044147633,0.0649486035,-0.0130217951,0.5255144238,0.0390612222,-0.3334518373,-0.0510594137,-0.0791908428,0.1168845072,0.4121738672,0.1165451482,-0.071450755,0.033269681,-0.0730038434,0.4419531226,0.0267174933,0.1170964837,0.1134194285,-0.0553824976,-0.3535195589,0.0441618338,-0.0012506653,0.338738054,-0.150590837,-0.3162331581,-0.2969117463,-0.5610883236,0.0354145989,0.1109729409,0.0769252032,-0.2063972205,-0.0765422955,0.0325446799,0.1556462497,-0.2044659704,-0.1337816268,0.0294571128,0.1680212617,-0.2341176718,0.0363918729,0.2357255965,-0.2001557201,0.1674283743,0.1840349287,-0.1942940503,0.1128039137,-0.100288935,-0.1611268222,-0.2255152315,0.3519966602,0.1430627257,-0.0593459606,-0.1862950325,-0.4136186838,-0.2287944704,0.5124770999,-0.3206869066,-0.3754195869,-0.244556129,0.0786125362,-0.3883818686,-0.1862037778,-0.3287690282,0.2510109544,-0.0393284224,-0.5098038912,-0.1493531764,-0.4156335294,-0.2565268874,-0.2000540197,-0.051229503,0.2309639007,-0.3808170557,0.0061399257,-0.1954063773,-0.1390626878,0.2360926569,0.1372223347,-0.3120277226,0.0472986996,-0.1490423977,0.2054808736,0.3002551496,-0.0629010946,-0.3941163719,-0.0504084341,-0.097788319,-0.0877091065,-0.0379557386,-0.0514398851,-0.2243130505,-0.0908661112,0.161266312,0.1932044625,-0.081068404,0.0886701047,-0.3115685284,-0.3598886132,0.3557012081,0.1630169451,0.067070663,-0.0776222497,0.0859782994,-0.0980778337,0.2169847935,-0.1388312578,-0.1596058905,0.0454791039,0.4673460722,-0.2092601061,-0.1000456586,0.0152848409,-0.4038907588,0.1075213253,-0.2279120833,0.2124351561,-0.1815494746,-0.1396891922,-0.1447748542,0.1303115934,0.1029608846,0.1421173066,-0.0029944424,-0.0913659036,0.0570315272,0.2547571957,0.0634429231,0.0823936239,0.0543814078,0.0983150005,0.0363243148,0.338307023,-0.3303178251,-0.5239762664,0.1456165314,0.1461755633,0.1693951041,-0.0051058475,0.0574969202,0.0682088956,0.0715708435,0.1704884917,-0.4348712564,-0.1450346708,0.1928227544,-0.483019352,0.0335157812,-0.0502487756,-0.1301379502,0.161994651,0.4932235479,0.3439658582,0.2142984569,-0.0634261668,-0.0262597837,0.0432463586,0.0430728607,-0.0239672307,0.0032598397,0.0689102486,0.1253630519,0.3056398332,0.1963308603,0.0028555179,-0.3140211105,0.1387964785,0.4472028613,0.1393594593,0.212503314,0.136262998,-0.0715110973,-0.1618019044,0.2675538361,-0.2446559072,0.2550039291,0.214611873,-0.087952897,0.3010336757,-0.2936895788,-0.2716570497,0.0913611278,0.0676496997,0.2671325505,0.0778898746,0.2691613436,0.0112295253,-0.3219194412,-0.0014552311,0.0435953625,0.364189893,-0.1933288425,0.2375709713,-0.4003891945,-0.4925537705,-0.3020910323,-0.3256745934,-0.4260711074,-0.0267012324,0.0004716117,0.4218554199,0.1290074587,0.2647658885,0.4485256076,0.0410506576,-0.2630127072,-0.2133401185,-0.1491859406,-0.356030792,-0.5901415348,-0.0550509803,0.2354554981,0.0247176383,0.1158504486,-0.1725975424,-0.0159074403,0.1693310291,-0.6060656309,0.0438021645,-0.1981709898,0.3902420104,0.1650788337,-0.0682044923,-0.4676985741,-0.2652578652,0.0693664178,-0.2510098517,-0.1902030706,-0.0777376294,-0.3408636451,0.1573370546,-0.160547033,0.0584651344,-0.5486683846,-0.1798711419,0.0759309605,-0.0197855458,0.2007305324,0.3501478136,0.2462227792,0.1525546163,0.0597291589,-0.2772377729,0.0489227623,0.0759284347,0.1792278439,-0.1600553989,-0.2186676562,0.2411717772,-0.1012460887,0.0286456048,0.0302117504,-0.3914150298,-0.4176371694,0.1224974468,0.2533755898,0.111381419,-0.0526968203,0.3645016849,0.0483023375,0.1041247249,0.0521648899,-0.3844707012,-0.222507,0.1326135397,0.2014407367,0.1643473059,0.1951599419,-0.0425117612,0.6236301064,-0.2003175169,-0.0814630538,0.3995680511,-0.164977923,0.2369250506,0.0610084124,-0.0792022049,-0.0306156334,0.1708610654,-0.098029092,0.1847988814,-0.1580742002,-0.5035770535,-0.1711388081,-0.2859545648,-0.2396086454,-0.2830763757,0.0812684372,0.3607314527,0.2461465597,0.1205044687,-0.1421339512,-0.0035370863,-0.3313276172,-0.188148886,0.1266995519,-0.0172821246,0.0801827833,0.3385299444,-0.6055163145,-0.464789927,0.2710725665,0.4054426253,0.1104435027,0.0886853039,0.0483302735,0.0636160076,-0.1098980084,0.1858922988,-0.0156436656,-0.0181909967,0.1939270198,0.2568919659,-0.2476299852,-0.2221847773,-0.0012900917,0.1623787135,0.2810534537,-0.0348749459,-0.2546073794,0.0861273408,0.548101306,0.3830603361,0.0328108035,0.0157841817,-0.4715173244,-0.4916702509,0.0727161318,0.0863104165,-0.0700458214,0.2802959085,-0.0845083296,0.1770837307,0.1313180923,-0.2539827228,-0.0914512128,0.3505503237,0.1222682819,0.0350404531,-0.0676379427,-0.1327982843,-0.2915899754,0.3810206652,0.4118835628,0.1647192389,-0.0634159669,-0.0275375731,0.2902988791,0.2413718849,0.0840421766,-0.1185430512,0.0737967193,-0.157504946,-0.2397676706,0.2299426496,-0.0937560797,0.0673803538,-0.0595247,-0.2887191772,0.1323292255,0.6328302622,0.184954226,-0.0747765973,-0.0003745893,0.0496741533,-0.2676107287,0.2414990366,0.0057510301,0.8142017722,-0.0296281315,0.0515657514,0.6661900282,-0.0761903077,0.7036731839,0.2661636472,0.3528727591,-0.3629660904,0.0752827078,0.0127586834,-0.1304983199,0.1840191036,-0.0909759253,-0.3536271751,0.2753056884,0.2211853415,0.0363310799,0.3177632987,0.1732017696,-0.0024978693,0.0787056088,-0.3247781396,0.0041711219,-0.2018572092,0.2967710197,-0.3591496646,-0.0805774778,0.0738127083,-0.2858181298,-0.4786446989,0.1992829889,-0.3487883806,-0.1221350506,0.2465490699,-0.4966215193,0.0953278989,-0.0164618958,0.0199613739,0.019620223,-0.201690197,0.2468871474,-0.0255462788,-0.2979871929,-0.0959393159,0.1842326075,0.0951421186,-0.137866497,-0.199685216,0.0924672186,-0.0935827792,-0.192083016,0.1219532564,0.1810778528,-0.1708514243,0.2578439415,0.0723887607,0.0352376737,0.2820960581,-0.1410411447,0.0814673156,-0.0013515748,0.1241636798,0.1338847578,-0.1384643018,-0.2203139812,-0.1010913998,0.3434837162,0.118501462,0.1323244125,0.4635111094,0.1885223836,-0.0222501419,-0.0515637919,-0.1698374748,-0.2832883,-0.3888899684,0.0557544157,0.1640552729,0.2647970021,0.24753353,0.3323554993,0.1171675026,0.0634387732,-0.012718807,-0.3526568115,-0.1620147973,0.3513338268,0.133786425,-0.1118116677,0.2708748877,0.4370234907,0.2325568199,-0.2713712454,-0.1812610328,0.2420377582,-0.0177196376,0.174693644,-0.1222125441,-0.0377500877,-0.0122402301,-0.0180035308,0.0242513958,0.0836162791,-0.0145133017,-0.1368601471,-0.1140373573,0.15488635,0.1769232005,-0.2196128368,-0.1032012776,-0.2271044254,-0.0057965741,0.176925838,0.032126978,0.038334582,-0.2101633996,0.097586751,0.1662620604,0.3129869401,-0.0714405552,0.1428781748,-0.24012582,0.4941482246,0.3211106956,0.3765466511,0.2219655961,0.1127870232,0.2302962244,-0.1440642923,-0.2548358142,-0.1213790625,0.0101210903,-0.3708667159,-0.0447587296,-0.3500881195,0.1484654546,0.3040749431,-0.069891803,-0.1775495559,0.4846126437,0.0617816783,-0.413808763,0.0382892564,0.2101428956,0.156310454,-0.1152122393,0.3074785471,0.3127394319,0.0474895574,0.1905768365,0.17607297,0.2926397622,0.3245047331,0.0117984451,0.2205801159,-0.0789973512,0.1357562542,0.2459178418,0.1815381199,0.3703653812,0.0102032274,-0.2524585426,0.0794761479,-0.22381787,0.0705206394,0.0082657645,-0.5372055769,-0.1439762563,0.3426620662,0.2833074927,-0.2511577308,0.1802402735,0.3687352538,-0.1258741617,-0.0277711898,-0.1038320884,0.3749889135,-0.1519203782,-0.0687548295,0.0088015366,0.0625821501,-0.110702984,0.2042470127,-0.1038184166,-0.1219884083,0.107633017,0.1789976507,-0.2664551437,-0.1714940518,-0.0030568258,0.3085206449,-0.0797654018,-0.0363798402,0.3245770931,-0.228065148,-0.3384031057,-0.1858619004,0.5080713034,0.4384659529,0.3087807298,-0.2624335289,-0.2667557597,-0.2319175005,0.125834778,-0.1042219624,-0.0079501532,0.4623076022,0.5319444537,0.1178497523,0.0225403495,-0.0614130273,0.3108057678,0.2099615484,-0.0557058007,-0.105698213,0.0231035855,0.0443166532,-0.0973172262,-0.1338405907,0.1326925755,-0.1819348186,0.4299164712,0.2465865016,0.2611832917,-0.1198614463,-0.0703891814,0.0120201819,0.1010550931,0.4806208909,0.4830845892,0.5739287138,-0.2637377083,-0.3367985189,-0.5740039349,0.2288509756,-0.5123119354,0.212241441,0.1462234557,-0.025787184,-0.3095862269,0.3150336742,-0.195591256,0.2499122173,-0.1147128493,-0.2726498842,-0.241667822,-0.1123864278,0.0365628935,-0.3866930604,-0.1546099335,-0.2600916624,0.1850961,-0.7096677423,-0.0817284286,-0.3806694448,-0.1794926375,0.0289867204,0.1839584112,0.6221885085,0.0742927194,0.4580357373,0.187094301,0.0280406568,-0.2896276712,-0.2676192224,-0.3311424851,0.2662706077,0.0583254993,0.1828648895,-0.1187847629,0.1551634967,-0.2721810937,0.4418957531,-0.1129724458,0.2406209558,0.1937556267,0.1133278683,-0.1347301602,-0.0170922223,-0.2136979103,-0.0027930029,0.0329815149,0.0204159562,-0.5701084137,-0.2316344976,0.3195487261,-0.4612442851,0.0216740575,-0.2186681479,0.152680397,0.1570495367,-0.2903895378,-0.3826716542,0.0830446705,0.3648610115,0.1277815104,-0.2624890804,0.1156125963,0.0398568064,0.0131215891,-0.0611205436,0.2065990567,-0.0190719403,-0.265186578,-0.1481828839,-0.188028574]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"We\u2019ll release the new version later today. Apologies for the delay.","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":11,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\nWe\u2019ll release the new version later today. Apologies for the delay.","embeddings":[-0.3928619623,0.265586853,0.0211132672,0.0056217709,-0.0660371706,-0.1835848242,0.1908972412,0.1489988416,-0.1800424308,0.1524226815,0.0167764965,0.627473712,-0.3186945617,0.0652184114,0.1079571322,-0.1102266684,0.0916200429,0.38757658,0.0339737684,0.1212250069,-0.2369133681,0.3257184029,-0.4104473293,0.2572149336,-0.2873312235,-0.0298990738,0.029598752,0.0856958181,-0.3213042021,-0.6685456634,0.4273485541,-0.0884734839,0.2396136373,0.2405719608,-0.0001127618,-0.0310443919,0.4426565766,0.0976706818,-0.1840823591,-0.4008617401,0.0666888803,-0.449442029,0.3975343108,-0.0412606634,0.0215265211,-0.0388950333,0.3337551653,0.6269931197,0.3148453236,0.2499588579,0.2013298869,0.1818131059,0.3285518885,-0.0142899621,0.4713775218,-0.2373925447,-0.0796111524,0.7525509596,0.1613449156,-0.3280172348,-0.0022737267,0.1282790452,-0.3156880438,0.0439055786,0.0566129461,-0.016002113,-0.109558709,-0.3133020401,-0.0079940371,0.2396161705,0.3033410907,-0.3365846276,-0.1225629449,0.0018697697,0.1720004678,-0.5164616108,0.1400690973,0.2636066377,-0.235295549,0.202057451,0.0792927742,-0.1226466671,-0.1810997576,0.2043073177,-0.1081981733,0.7457428575,0.0957630873,0.1287005395,0.2587382197,-0.0362515152,0.5486931801,0.2890076935,-0.0388724059,0.1275891513,-0.0010833718,0.0967619494,0.0477116033,0.4738041461,0.0040356372,-0.305475086,0.0149123073,-0.0029531303,0.1174461469,0.4877696633,0.0358190387,-0.0865610763,-0.0210245587,-0.0513033271,0.4425576329,0.0244457535,0.0901426822,0.086167261,-0.0139351394,-0.3052896261,-0.0051765083,-0.0598389544,0.2932459712,-0.138645187,-0.3549229205,-0.2837913334,-0.4308178127,0.0175992958,0.0724459812,0.2271878123,-0.1944729239,-0.1200819239,0.0100565264,0.1263552755,-0.221121937,-0.134115532,-0.042403847,0.2070760429,-0.2470369488,-0.0038082772,0.2412455976,-0.139056161,0.2508159876,0.1866459996,-0.1579906493,0.2046823651,-0.0628639609,-0.1300960183,-0.1716933697,0.2880769074,0.1920234263,-0.0759080648,-0.1550746113,-0.4207522571,-0.2595423758,0.3946075737,-0.290395081,-0.378242135,-0.2829343081,0.1668417156,-0.3350865543,-0.2427256107,-0.2525768578,0.2214252204,-0.1293059289,-0.5388994217,-0.1313453615,-0.3878832161,-0.196976006,-0.237493813,-0.046870932,0.1865907907,-0.3759917319,-0.0184231382,-0.2060789615,-0.1389924735,0.3028576672,0.1974721104,-0.2917577624,0.0350199565,-0.1182103977,0.2496536225,0.3652318418,-0.0948117375,-0.331268549,-0.0300675835,-0.1113214791,-0.1302596778,-0.0781752616,-0.0132622505,-0.1569687128,-0.0745518953,0.1850048453,0.2689126432,-0.0472206585,0.0847991556,-0.3150583208,-0.2877329588,0.2538984716,0.1932625324,0.0573303364,-0.1571135223,0.1134762913,-0.0735147446,0.1251297891,-0.1627202034,-0.1881479472,0.065356195,0.3895967305,-0.1715402156,-0.1037907153,-0.0232728384,-0.3672347963,0.0667932481,-0.2590959966,0.2921105623,-0.1629046053,-0.1447914988,-0.192370683,0.157345131,0.0686914325,0.0639811754,0.1100213602,-0.0469240211,0.1205509827,0.2525580525,0.0518575348,0.0673388615,0.0625526682,0.0715923086,0.0132963741,0.2738650143,-0.3720898926,-0.5238743424,0.166758582,0.1546794474,0.1118968204,0.0182356499,0.0726778433,0.1047119051,-0.0231769513,0.1344881058,-0.3520150781,-0.1431106627,0.1909599751,-0.4850099087,0.0860637948,-0.0137114236,-0.0787493736,0.1476390064,0.3847483397,0.270660758,0.1247335523,-0.1009699181,0.0103488555,0.066141583,0.129523918,-0.032990694,0.0906166062,0.0347043648,0.1660395414,0.2870663702,0.1378797144,-0.0145480661,-0.34561342,0.1462796628,0.5138734579,0.1207722798,0.2526911795,0.1631111354,-0.0941804051,-0.2132610977,0.2606529593,-0.1559281796,0.271772325,0.2789547145,-0.0289303679,0.2941123545,-0.3221452832,-0.3071650565,0.0956746489,0.0145622492,0.211391598,0.0576824769,0.282738924,0.0271889772,-0.3873322904,-0.0689890832,-0.0346229412,0.3347578049,-0.1296111792,0.2140217274,-0.4033831954,-0.5679747462,-0.2339728624,-0.3338429332,-0.3989624977,-0.083727248,0.0832416788,0.3483988643,0.0618157908,0.2609183192,0.442468375,0.0578226186,-0.2587819695,-0.1995637566,-0.1992365718,-0.3707919717,-0.5364171863,0.0402366407,0.2339836955,0.1774205863,0.2168437093,-0.1574206352,0.0521249808,0.1270951331,-0.5863828063,0.0644469187,-0.2213591784,0.3972024918,0.157667011,-0.0715162233,-0.4580964446,-0.3727555871,0.088134706,-0.3110891879,-0.1539316177,-0.0220779199,-0.323330462,0.047329966,-0.1956515908,-0.0625344515,-0.6100064516,-0.2911258638,0.1276990026,0.1112787947,0.1501705348,0.3212847412,0.2065825611,0.2087185234,0.0808460265,-0.1015523449,-0.0231536217,0.0404158086,0.213080436,-0.2417276055,-0.2967261076,0.2184891254,-0.1478996426,0.1215623617,-0.044603724,-0.4358358085,-0.4339602888,0.0357816257,0.2800705135,0.1476392597,0.0363053381,0.3327871263,0.1040136591,-0.0340008736,0.0319600329,-0.3564096391,-0.1760418862,0.1002814546,0.1435606629,0.0344900526,0.1369282454,-0.0892772824,0.5816776752,-0.2194190919,-0.0745385066,0.4348016381,-0.1277737767,0.3395959139,-0.0658831149,-0.0715913624,0.0070649707,0.0679211989,-0.0961051509,0.2410869896,-0.1990359724,-0.5346443653,-0.1951682568,-0.2407983243,-0.1949211061,-0.2278166562,0.0922109932,0.4309327006,0.151700303,0.1210397258,-0.1928579509,0.0042145797,-0.3483369648,-0.1603388339,0.0468729548,-0.1062666923,0.0434887372,0.2516714036,-0.53021878,-0.4617776275,0.2958478332,0.3734951317,0.0007775106,-0.0312046669,-0.0027495485,0.0615601204,-0.0962887183,0.0923673734,-0.0424539633,-0.073545441,0.2140326053,0.2376083136,-0.2790341377,-0.1610026062,-0.1906701624,0.1468968391,0.3291750252,-0.0474603884,-0.2876831889,0.098327741,0.5470016599,0.3685385287,0.0773132071,0.090121232,-0.4112402797,-0.518627286,-0.0413766131,0.0451740995,-0.0389924683,0.2847243845,-0.0993348807,0.1791635752,0.1237559095,-0.1389996409,-0.085192427,0.4364383519,0.0793822408,0.0621300079,-0.0816457048,-0.1714352667,-0.2476165742,0.324277401,0.4695363641,0.1136844009,-0.0762639344,-0.090823181,0.2723959684,0.2496158034,0.068659924,-0.0731220618,0.0821309611,-0.1656980813,-0.1484423727,0.2138954848,0.0689057931,0.0768660754,-0.044273179,-0.2072188854,0.0792867318,0.6434319615,0.2255206108,-0.0712764636,0.029266078,-0.0485713184,-0.2335541844,0.3367161453,0.0523756817,0.7968222499,-0.0345991403,0.1007860228,0.6916451454,-0.0777896568,0.6967650056,0.1905237436,0.3257184923,-0.3575980961,0.0490527228,0.0645754039,-0.0782653466,0.0954186022,-0.1248991191,-0.4398855269,0.2828764915,0.3114647865,-0.0150725786,0.275403291,0.1808250397,-0.0453200266,-0.0169200432,-0.3691766262,0.0983172804,-0.2441322803,0.2114052176,-0.3643822968,-0.1010386497,0.0827938393,-0.3065524399,-0.3683643043,0.2382064164,-0.3133550286,0.0138000213,0.1920966655,-0.4213193953,0.0738419592,0.0337376334,-0.0347984806,0.0115730939,-0.1766508967,0.2105634511,-0.0471551865,-0.36366871,-0.0864127278,0.2284861952,0.1986487508,-0.2101545632,-0.1932811886,0.1587896496,-0.065551497,-0.1856131107,0.1651477665,0.2079090625,-0.1298205554,0.1972217262,0.0294411462,0.0632802472,0.1807620227,-0.1414452493,0.1396474838,-0.0107529555,0.0822593048,0.1684626788,-0.0332645997,-0.2133683413,-0.153075248,0.2591201961,0.065104112,0.1200298518,0.4494114518,0.1948517859,-0.0882457346,-0.1357953697,-0.1923191994,-0.3059974313,-0.3530040085,0.1035138369,0.2113858461,0.1510873735,0.2619870603,0.3921471834,0.1396550387,0.0288474336,0.0000165963,-0.3490558863,-0.1802673936,0.3559380472,0.1000560001,-0.081247285,0.2733875513,0.3584241867,0.2282194346,-0.2103022337,-0.2887899876,0.241688624,-0.1613867432,0.1962512434,-0.1001624614,-0.1173657104,0.0465751328,0.0924050435,0.076794818,0.0632328987,-0.0771086887,-0.2317153066,-0.1138321981,0.1014485508,0.1388917118,-0.2169037759,-0.1055976152,-0.1374408752,0.0479081869,0.1290694624,0.0904925615,-0.0331519693,-0.2315212488,0.1693896055,0.130553335,0.2612940669,-0.0014290113,0.0921957567,-0.2195491046,0.4294117093,0.253605932,0.3178195655,0.1630413234,0.0555252768,0.2987040579,-0.1593558341,-0.1329533458,-0.1236068606,0.0335830487,-0.3481328785,-0.038431298,-0.3484321833,0.1772597283,0.3195530772,-0.0951220989,-0.1516462862,0.5109324455,0.1528750807,-0.4550715983,0.0303542912,0.1836602837,0.1057432815,-0.0332463756,0.2189600766,0.2794484496,-0.0082536265,0.2108598799,0.1860845834,0.2683892548,0.1992966831,-0.0058661047,0.2043770105,-0.1277206689,0.1093823612,0.2290905565,0.2563354075,0.3824566901,0.0612812936,-0.2906113565,0.1553618759,-0.2481351644,0.0501481444,0.0442735478,-0.4540173411,-0.0705521107,0.2817179263,0.2916861176,-0.1900586486,0.1660884619,0.404861927,-0.09360414,0.0269910451,-0.1287951767,0.4079390168,-0.1080033183,-0.0419694483,0.0109196575,0.0528541431,-0.1340346783,0.2470913678,-0.0856806785,-0.0842565149,0.1365499496,0.2071661502,-0.2033037245,-0.1510156989,0.0613526776,0.2240278274,-0.100082323,-0.0397591926,0.260109067,-0.1178049669,-0.3100940883,-0.2224158645,0.505964458,0.5416741967,0.2697434127,-0.3149753213,-0.2180320472,-0.2548582256,0.0601380169,-0.1561240107,0.0862512067,0.4850741625,0.484765321,0.1764291525,0.0849464238,-0.1714150012,0.2670813203,0.1862159371,-0.0442179739,-0.0228175987,0.004703857,0.0986026153,-0.1165620834,-0.1846796572,0.0333109051,-0.280171603,0.4427824914,0.2251491994,0.1951107383,-0.0996483266,-0.1489378661,0.0658098906,0.116931282,0.6242977977,0.4585308433,0.6078999639,-0.2093818635,-0.3121260703,-0.5995371342,0.1858277619,-0.4579473734,0.1184997186,0.0570898578,-0.041615922,-0.2388275415,0.3895166516,-0.2350209355,0.2411487848,-0.1484231651,-0.2592346072,-0.2939122617,-0.0164430309,0.0577656589,-0.3580429554,-0.0965199992,-0.2263828367,0.1279900074,-0.6864494681,0.025855694,-0.3320232034,-0.2160741687,0.0017189574,0.1813963354,0.5521525741,0.1236259192,0.4865070283,0.06666141,0.0407598987,-0.1676058024,-0.3055434823,-0.2400092781,0.2331112623,0.1490565687,0.1931441873,-0.0936118513,0.142068401,-0.2290839255,0.4464408457,-0.1146419346,0.21884121,0.2379254103,0.0936499313,-0.0818440616,-0.0497968346,-0.229406774,0.0240148157,0.0167923532,0.0029449277,-0.525873363,-0.3466750979,0.3878202438,-0.432007134,-0.0032191747,-0.175593093,0.0914948508,0.1152140498,-0.2977773845,-0.4016132653,0.1321647316,0.2825413048,0.0709348321,-0.3026912808,0.1634206176,-0.007244477,-0.0871322826,-0.0679637119,0.2167313248,0.0365789644,-0.2977956235,-0.136519134,-0.206905514]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"I just pushed the new version on pypi :)","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":9,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\nI just pushed the new version on pypi :)","embeddings":[-0.3508317471,0.2680680156,0.0205311198,0.0258314051,-0.0463165157,-0.1936451495,0.2140858471,0.1566912681,-0.1997242272,0.1394782513,0.0081854248,0.6095814705,-0.3005533516,0.0775661543,0.0857656598,-0.0840008631,0.0875910074,0.374792546,0.0421937108,0.1191469133,-0.2289435118,0.3022626042,-0.4048751593,0.2636783421,-0.3155056238,-0.0249946844,0.0099598486,0.0742144436,-0.3029278517,-0.6610277891,0.4454162717,-0.09296152,0.2283584774,0.2456614226,-0.0001125968,-0.015523091,0.4485095143,0.0955779627,-0.1846574098,-0.4083963633,0.0630607903,-0.4497354031,0.4080223739,-0.061790999,0.0125369718,-0.0333287604,0.3030797541,0.6422285438,0.2961751223,0.2403484136,0.2058142722,0.1891800761,0.3061630726,0.0108857313,0.4494490921,-0.2107879519,-0.0752184764,0.7553877831,0.1514164805,-0.3376544118,0.0025336824,0.1148026139,-0.3060088158,0.0342179798,0.0597930811,-0.0186910909,-0.1161401495,-0.3164305389,-0.018508479,0.2547187209,0.2715970278,-0.3531976044,-0.1172072738,-0.0001199143,0.176698938,-0.5225009322,0.1400215477,0.260571748,-0.2504837811,0.2041671574,0.0751607493,-0.1047217473,-0.1947127432,0.194644019,-0.1216449961,0.7545172572,0.1195234209,0.111895062,0.28012532,-0.0447816066,0.5560870767,0.2888222635,-0.0509282425,0.1351216435,-0.0268970411,0.0736348331,0.0543642789,0.4330784976,0.0158588663,-0.2998568714,0.0204232335,0.0058629927,0.1170117185,0.4832438529,0.0263542123,-0.0698473454,-0.0210274179,-0.0505614653,0.4367246628,0.0162569527,0.1042483822,0.08364968,-0.0090736048,-0.2959830463,-0.0077734562,-0.0523797236,0.295699954,-0.1682008505,-0.3692882359,-0.2851962149,-0.4297475517,0.0158126354,0.0781400055,0.2185093164,-0.159133479,-0.1375026554,-0.0029398166,0.1548661143,-0.2231743634,-0.1123214811,-0.0462503955,0.2019455731,-0.25110659,0.0460540764,0.2399810553,-0.1284029633,0.2488639355,0.1977431178,-0.1398638785,0.2075790316,-0.0428296998,-0.1469058245,-0.1729917228,0.2837948203,0.1532133967,-0.0819794387,-0.1595107466,-0.3885949552,-0.2485820651,0.3779113591,-0.2807967961,-0.3738734424,-0.2937239707,0.1760843992,-0.3473351002,-0.232780695,-0.2419377714,0.2364782989,-0.0930132344,-0.5240267515,-0.1218581274,-0.3579743505,-0.1629974693,-0.2291339934,-0.0218190793,0.2090779692,-0.368760258,-0.0079216966,-0.1747625023,-0.1327167153,0.3013397157,0.2171193957,-0.2832858562,0.0311607253,-0.1179042235,0.2696638107,0.32238397,-0.0648647845,-0.3416082561,-0.0223647393,-0.1042122766,-0.1219468042,-0.0709008723,-0.0176672749,-0.1465487927,-0.0668857619,0.1824769974,0.2682054639,-0.0408015661,0.0899329931,-0.3331260383,-0.3010832965,0.2976618707,0.186291337,0.0526121557,-0.1548533142,0.1059245467,-0.0877991021,0.1402735859,-0.1752076,-0.1816137582,0.04095193,0.3906413317,-0.1846007854,-0.1238253936,-0.0019310232,-0.3878824413,0.0720554143,-0.2667293847,0.3224515319,-0.1965634376,-0.1404677778,-0.1497124285,0.1448054314,0.0244884305,0.0609855466,0.1161839142,-0.0601558797,0.097715959,0.2466173619,0.0565318577,0.0426316932,0.0590017736,0.0769157037,0.0375921726,0.3014974892,-0.3623258471,-0.5295037627,0.1712924391,0.1539871991,0.1320308447,0.0012002969,0.0570869856,0.0875017345,-0.0206149053,0.1339949667,-0.3447114825,-0.1507551223,0.2092625201,-0.4879759848,0.0728071928,-0.0109914048,-0.0823476091,0.1329093724,0.3704680204,0.2872930169,0.1557238698,-0.065716967,-0.0226339307,0.064592205,0.1270708889,-0.0232581366,0.1000705436,0.031766776,0.1422871798,0.3035101593,0.1392792165,0.0120530175,-0.3407210112,0.1667722166,0.4997148514,0.1016581953,0.2521179914,0.1808045059,-0.0835794359,-0.2297865003,0.2936893404,-0.2062995881,0.2527575195,0.2649404109,-0.05747547,0.2782749832,-0.3262929618,-0.3145926595,0.0806694627,0.0155337732,0.2224550396,0.040407192,0.2984961569,0.0111547606,-0.3924562633,-0.076640971,-0.0177873112,0.3254519701,-0.1324270517,0.2285059392,-0.4054168463,-0.5517749786,-0.2141981572,-0.3383580744,-0.4067614377,-0.0820841119,0.0691582039,0.3345482349,0.0547733381,0.2530862093,0.4430470467,0.0683467686,-0.2397996634,-0.2030531913,-0.2151540369,-0.4141944051,-0.5352424979,0.046566058,0.2358040363,0.1862943769,0.1936306357,-0.1542484462,0.0403742231,0.1286516041,-0.5887747407,0.0591379181,-0.2204429209,0.3582536876,0.1162846237,-0.0604575612,-0.4352612495,-0.3753906488,0.0922010243,-0.2715719044,-0.1533640325,0.0054572849,-0.3358978927,0.0657819137,-0.2037088573,-0.0538089983,-0.6270270348,-0.2828673124,0.1212908328,0.0937232971,0.1462503225,0.337831527,0.237892434,0.2074893266,0.0615851507,-0.1193652004,-0.0135453437,0.0263047945,0.2071847171,-0.2235632241,-0.2928881347,0.1812321693,-0.140472427,0.1129658446,-0.0419401713,-0.4483516812,-0.4349688292,0.027935192,0.3105922937,0.1572688371,-0.0107121216,0.3352378607,0.0965864211,-0.0314516388,0.0537536666,-0.3609650731,-0.1555559635,0.0800094455,0.1371096075,0.0409699306,0.1537256986,-0.0847323835,0.5848183632,-0.232278809,-0.0925081894,0.4489090145,-0.121630922,0.3357083499,-0.0650290921,-0.0909219459,0.0398237146,0.0817664564,-0.0930499882,0.2133185863,-0.2200874984,-0.5515707731,-0.1862342209,-0.2560796738,-0.1672228128,-0.2428095192,0.0850506425,0.4319095612,0.1526491791,0.115137741,-0.1596841067,-0.0092586754,-0.3582448661,-0.1405462027,0.0613634996,-0.1180137396,0.0529061295,0.2564632893,-0.5393281579,-0.4648052752,0.2872029245,0.3699335158,0.0003676745,-0.0274502095,-0.0189910363,0.080043532,-0.0892629772,0.0913243145,0.0136730745,-0.0794008672,0.2217096984,0.2294642627,-0.2496531606,-0.1744817942,-0.1790077239,0.1580491364,0.3080521226,-0.0077327853,-0.2995739281,0.0947905481,0.5314478874,0.3702380061,0.0666676387,0.0693754703,-0.3998682797,-0.5215491056,-0.0495596342,0.0626437366,-0.0618435591,0.299271524,-0.1100355238,0.1719460934,0.125792101,-0.1467431337,-0.0907668397,0.480913192,0.0899810418,0.0497900844,-0.0585267246,-0.1561190337,-0.2520170808,0.3527752757,0.4585244656,0.1407930553,-0.0616885833,-0.0775765181,0.2561168969,0.2330732495,0.0578689426,-0.063120164,0.1111309603,-0.1993943304,-0.1501573324,0.2165771723,0.0867098272,0.0799155012,-0.0376532488,-0.2184195817,0.0731517226,0.6497707963,0.1885305494,-0.0705180988,0.0204320811,-0.0472486727,-0.2290129215,0.3272971213,0.0784899369,0.8129765391,-0.026593402,0.111354433,0.7389096022,-0.073122941,0.6670873165,0.183878541,0.3112436533,-0.3537389636,0.0383008495,0.0816776976,-0.0787640661,0.0785440058,-0.1207272038,-0.4248327315,0.2662982941,0.2790385485,-0.0188185032,0.2760612369,0.1741544604,-0.0159593802,-0.0278906431,-0.3542720675,0.1085103452,-0.2299512625,0.2101995051,-0.3604547381,-0.1097938567,0.106569849,-0.2808859944,-0.3530727923,0.2623608708,-0.2899084687,0.0330133028,0.1609227359,-0.4377331436,0.0599913634,-0.0190445352,-0.036546018,-0.0027035512,-0.1679457426,0.1983912885,-0.026846543,-0.3416581154,-0.1067510098,0.2149324864,0.1956563592,-0.1956285685,-0.2044248581,0.1850815266,-0.0439812765,-0.1674169451,0.1140672043,0.1872764826,-0.1691515595,0.1968254447,0.0508868098,0.0648491383,0.1945194453,-0.1405551881,0.1440162063,0.0189442001,0.0717753246,0.1535475701,-0.0430387855,-0.2262385935,-0.1592779905,0.2359356135,0.0781898722,0.1398868263,0.461994648,0.2069515735,-0.0688759461,-0.1421791315,-0.1581386477,-0.3005910516,-0.3551547825,0.094354488,0.2207161188,0.1289371848,0.2434998602,0.3849505186,0.1317562312,0.0341523439,-0.0156892873,-0.3489277363,-0.2098043412,0.3328018188,0.0846288577,-0.0654998645,0.261765331,0.3994864821,0.2028087527,-0.2261392772,-0.2915697694,0.2311779112,-0.1757757366,0.1963832676,-0.1140250564,-0.1323813498,0.0377489775,0.0771232694,0.0777873546,0.077666454,-0.053366594,-0.2337750047,-0.1155537441,0.0976739526,0.1460529715,-0.2119998783,-0.1213732064,-0.1289277673,0.0602952056,0.1234336719,0.0929978713,-0.0412143469,-0.2179548144,0.1743669063,0.145253107,0.2261411548,-0.0048174532,0.0954275578,-0.2127473056,0.4004851878,0.2564915717,0.3038153052,0.1411653906,0.0670235232,0.3138961494,-0.1415722668,-0.1511455923,-0.1219386309,0.0271940343,-0.3435825109,-0.0256765261,-0.3655988872,0.1542771161,0.313734293,-0.0881988555,-0.173738271,0.5237312317,0.1602634043,-0.4514446259,0.0263267141,0.1579510719,0.0814097002,-0.0331360251,0.21822685,0.2439642698,0.0116534308,0.1881196797,0.1928498298,0.2750020623,0.2436883301,-0.0117724072,0.1956076175,-0.1486485004,0.0957422182,0.2504188716,0.2345005423,0.3840842843,0.0555958338,-0.3154777288,0.1728827059,-0.270290792,0.0445564315,0.0517620109,-0.4520296454,-0.0963128582,0.2709935308,0.2819722891,-0.2072890699,0.1658545285,0.3883560896,-0.1002450809,-0.0224137288,-0.1260544807,0.4290037155,-0.1332758218,-0.0157774035,0.0120420214,0.0430322289,-0.1166837513,0.2361524105,-0.0875495523,-0.086923331,0.1597953439,0.1881573945,-0.2232466042,-0.1377759576,0.0747179687,0.2673287392,-0.1068638787,-0.0297015216,0.2720684111,-0.1011703908,-0.3231380582,-0.2353357077,0.4880987406,0.543679595,0.3035125434,-0.3099532723,-0.2103082687,-0.233050704,0.062199235,-0.1506360471,0.0806674883,0.4701154828,0.4717576802,0.1692983806,0.0898793191,-0.180474773,0.2701489329,0.1947871745,-0.0336162858,-0.035408996,0.0104147103,0.1076169983,-0.1280627996,-0.1849357039,0.018341884,-0.2833152711,0.4278299212,0.2336269021,0.2066345513,-0.0956826732,-0.1477377713,0.0702075213,0.1268964708,0.6447007656,0.4475016594,0.5687713027,-0.2551181018,-0.293201834,-0.6103157401,0.2011010349,-0.4556846917,0.1247732937,0.0537884161,-0.041017618,-0.2316500396,0.3788407147,-0.2278790921,0.3014771938,-0.1492731869,-0.2921201587,-0.3191390038,-0.0097543579,0.0320640244,-0.354552269,-0.0970342755,-0.2215728313,0.1455064565,-0.683201015,0.046984721,-0.3261548579,-0.2045505494,0.0101966904,0.1842989475,0.5388609767,0.1344024539,0.4962949157,0.0943756476,0.0514835119,-0.1630333364,-0.3330749273,-0.2357866466,0.2388114482,0.1391340643,0.176915735,-0.081567511,0.1169387326,-0.2519991398,0.4683250785,-0.1146225408,0.1887094975,0.2378264666,0.1072779,-0.0857378542,-0.0641710311,-0.2179020643,0.0180179793,0.0251598153,-0.0143923126,-0.5092054009,-0.3507784605,0.3877584636,-0.4590578973,-0.0007756741,-0.1702089012,0.0923937261,0.102423653,-0.287989378,-0.3933350146,0.1464700252,0.2848354876,0.0582953244,-0.3034025133,0.1534351259,0.0076052509,-0.0792022645,-0.0874844119,0.1841225624,0.0496209562,-0.3062545955,-0.1210152283,-0.1891924441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/436","title":"Google Colab - load_dataset - PyArrow exception","comments":"Thanks for the update.","body":"With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0","comment_length":4,"text":"Google Colab - load_dataset - PyArrow exception\nWith latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue\r\n\r\nImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.\r\n\r\nThe error goes only when I install version 0.16.0 \r\ni.e. !pip install pyarrow==0.16.0\nThanks for the update.","embeddings":[-0.3700290322,0.2840512991,0.0222639367,0.0192585308,-0.0522176176,-0.182397902,0.2077063769,0.152354762,-0.1941755712,0.1381382048,0.0105304401,0.6366501451,-0.3332311809,0.0655692294,0.1191536188,-0.1080089286,0.107176587,0.3795983791,0.0438284576,0.1206514686,-0.2381965816,0.3165137172,-0.4334976971,0.2426273525,-0.2950979769,-0.0079284944,0.0174420159,0.0742982253,-0.3078432381,-0.6941232085,0.4479759634,-0.0852703229,0.2505176663,0.2462004721,-0.0001138767,-0.0281738434,0.4584410787,0.0927537903,-0.1957945675,-0.4097465575,0.0665480345,-0.4644884169,0.4069104493,-0.040326111,0.0136713097,-0.0515956432,0.3226002753,0.6516059041,0.28871876,0.252068013,0.1947896034,0.1585375667,0.3575797379,0.0151442513,0.4671547115,-0.2299829721,-0.0610377602,0.7538289428,0.1681219786,-0.3452362418,-0.0130081866,0.1387402117,-0.3258865178,0.0161175057,0.0760749131,-0.0394247882,-0.0995059162,-0.3018703759,-0.0061576515,0.2241140753,0.3004875779,-0.3343207538,-0.1276616603,0.0131297149,0.1700264513,-0.5232781172,0.1516243964,0.2655702829,-0.2348738909,0.203404814,0.0699692592,-0.1474315971,-0.1796646267,0.2029129118,-0.101253137,0.7523468137,0.0986592621,0.1241826192,0.2469930202,-0.0391096063,0.5691335797,0.2933160961,-0.0353692435,0.1421473175,-0.0218790229,0.0734252185,0.0501943715,0.4685126543,-0.0082600899,-0.2977666855,-0.0075882617,-0.0050423136,0.1118072644,0.4714275897,0.037944179,-0.075791575,0.0058996715,-0.0401602574,0.4317034781,0.0377348736,0.0906941667,0.0914467946,-0.0251119491,-0.3021586835,-0.0023825706,-0.0391830616,0.3091134727,-0.13907592,-0.3429655433,-0.2883269787,-0.436347723,0.0039638537,0.0696396753,0.2112331986,-0.2033345997,-0.1321368814,0.0254963469,0.1201685667,-0.2066525072,-0.1307749748,-0.0416623093,0.2133820802,-0.2499329448,0.0065149586,0.2435425222,-0.1443676651,0.2438098639,0.1949630827,-0.1865650117,0.1773011088,-0.0848200172,-0.1293644458,-0.175347358,0.2803961635,0.1791609526,-0.1114689335,-0.1507591605,-0.4090836048,-0.2467686236,0.4302259684,-0.312378794,-0.3880565763,-0.3009550273,0.1536988318,-0.3322655261,-0.249591738,-0.2656819522,0.2390916348,-0.1157117337,-0.5281743407,-0.1144579202,-0.3723929822,-0.2197040319,-0.2300867438,-0.0440458469,0.2067791075,-0.3696523011,-0.0110391732,-0.2000601888,-0.1306314915,0.2974668741,0.16928339,-0.3053983748,0.0335046798,-0.1145975739,0.2207538784,0.3646850884,-0.0875782073,-0.3346005082,-0.0373876393,-0.1149881929,-0.1378725022,-0.0789133534,-0.0298200995,-0.169530049,-0.0642969906,0.189226523,0.2415283173,-0.0452870205,0.0973526388,-0.321573168,-0.3054945469,0.2698526978,0.1863614023,0.0607118048,-0.1336658448,0.1170416474,-0.0974609479,0.1645849794,-0.1564503312,-0.1905155331,0.0692585632,0.426581502,-0.175214082,-0.0941230878,-0.0113650989,-0.386395663,0.0698508546,-0.2540844977,0.2935895324,-0.18059282,-0.1445747316,-0.1863031387,0.1335254312,0.0611987039,0.0655757934,0.0959282294,-0.0759568661,0.1123685837,0.2552945614,0.0541278906,0.0681048334,0.057681337,0.0800917968,0.0325179137,0.3079106808,-0.3524124026,-0.5338782072,0.1741391569,0.145492807,0.1370512694,0.0145998774,0.047264263,0.0909008235,-0.0150376251,0.1523732692,-0.3614982069,-0.1471060812,0.1865389347,-0.4879683256,0.0919890702,-0.0306111947,-0.0862735361,0.1501856446,0.4018920958,0.2764922082,0.1307992339,-0.1080840454,-0.0060766251,0.064668335,0.1137345284,-0.0372238122,0.0980120674,0.0400146507,0.1436400712,0.3087717891,0.1511649489,-0.0104919523,-0.3240111768,0.1536962688,0.5069296956,0.1350986362,0.2373052835,0.1752336174,-0.0929955989,-0.2176992148,0.2781469524,-0.1642112285,0.2732792497,0.270833075,-0.0673082322,0.3168970048,-0.2849110365,-0.3056811094,0.1062174439,0.0266135372,0.2197972983,0.0510205291,0.2954320908,0.0253633931,-0.3914650679,-0.0531647727,-0.0279471725,0.3525187373,-0.1308194548,0.2072227895,-0.4006085694,-0.5605683327,-0.2454788685,-0.3246372938,-0.4062162638,-0.0719597116,0.0624903999,0.3618970811,0.0775534883,0.2808375955,0.4655168056,0.0614531785,-0.2440776825,-0.190956369,-0.2034802288,-0.368547529,-0.5317576528,0.0263882149,0.2266501486,0.1610089988,0.2032186985,-0.161667183,0.0670768842,0.139808625,-0.5977400541,0.0471301451,-0.2274673283,0.3786275685,0.1575967968,-0.0773672685,-0.4734034836,-0.3577050269,0.0709544495,-0.3036281466,-0.1628609896,-0.040263053,-0.3232260942,0.0848465562,-0.1828766614,-0.0509306826,-0.6270453334,-0.2715550959,0.1086626872,0.0821890086,0.173591882,0.3132401407,0.2070240378,0.2102760226,0.0654639378,-0.1441255063,-0.0101382099,0.0379471518,0.2041191012,-0.228804633,-0.2698405087,0.2032827437,-0.1490236819,0.1077937484,-0.0593183078,-0.4351278543,-0.4412122667,0.0377408117,0.2760820985,0.1456205547,0.0202634819,0.3509619236,0.1122421026,-0.0215458702,0.0533538722,-0.3539419472,-0.1866569668,0.1007749289,0.142073065,0.0562054701,0.1382920891,-0.0745905936,0.5948349833,-0.2159234881,-0.0861143619,0.4337234795,-0.1369396299,0.323569715,-0.0347241051,-0.0982911512,-0.0060101603,0.0897766426,-0.0840321034,0.2170196325,-0.2187387198,-0.5324994922,-0.1777584553,-0.2702568173,-0.1832966954,-0.2413723767,0.0908494294,0.4238536656,0.1662437022,0.1190338284,-0.1888583004,-0.0174940005,-0.3656499684,-0.1616077125,0.0702909902,-0.1040442362,0.0405855514,0.2731027901,-0.5517754555,-0.4767205715,0.2953789532,0.3700573444,0.0053603435,0.0036937031,0.0058293184,0.0761286095,-0.0801953524,0.1045669466,-0.036233142,-0.0727966279,0.1967004687,0.2469771802,-0.2584191561,-0.161976859,-0.1618344486,0.1597431749,0.2995797098,-0.024793677,-0.2769728005,0.0791423768,0.5696098804,0.3569338322,0.0709130615,0.0638196245,-0.4105973244,-0.504512012,-0.030914925,0.07117562,-0.0506874882,0.303678453,-0.1135904193,0.1822764874,0.1359026581,-0.1421931386,-0.0861045942,0.4161031246,0.0782108754,0.0726145431,-0.0738029256,-0.1640729308,-0.2666336298,0.3438106775,0.4442815185,0.0978054181,-0.0854071677,-0.1004226357,0.2789185345,0.2504006326,0.0661804155,-0.0766461864,0.0944670811,-0.1765929312,-0.160272643,0.219996959,0.0438935459,0.0482877977,-0.0397792794,-0.1942496598,0.079940021,0.6396145821,0.2114661336,-0.0699696541,0.0256247371,-0.0427390337,-0.2379076183,0.3273634613,0.0462510101,0.8020504117,-0.0094160084,0.0821512192,0.7128692865,-0.0771973729,0.6813219786,0.203420803,0.3321917355,-0.3643740714,0.0400435515,0.0693602711,-0.0796195716,0.1098848581,-0.1216770336,-0.4330507815,0.2848908901,0.2810418904,-0.022674866,0.2676724792,0.1821307242,-0.0236018971,-0.0026583092,-0.3529869914,0.0825249329,-0.2372426242,0.2259264439,-0.3567779064,-0.1060468778,0.1030965149,-0.306656152,-0.3630219996,0.2496110499,-0.3088864386,0.0032499041,0.2051853091,-0.4312529266,0.0823467299,0.0343581885,-0.0174330063,-0.0021139118,-0.1755562574,0.2154920399,-0.0319200978,-0.3461460173,-0.1005423665,0.2220951319,0.1724616289,-0.1887849718,-0.2181536406,0.1505133957,-0.0621668026,-0.1843065172,0.1683656573,0.1875918359,-0.1500006616,0.1938426793,0.0290614385,0.0691112056,0.1871576905,-0.1454884559,0.1320530921,-0.0029435281,0.0845854655,0.1635867506,-0.0324503556,-0.2219582945,-0.1388868392,0.2935275435,0.0780521855,0.1142999008,0.4699003994,0.2018894404,-0.0671819821,-0.1303754896,-0.1785734743,-0.2985887229,-0.3703556061,0.1008595526,0.2145346254,0.1936017275,0.2598288655,0.3755586743,0.1097033098,0.0624251924,0.0125177205,-0.3339145482,-0.1918672472,0.358574003,0.0847774819,-0.0881755948,0.2707226872,0.3660224974,0.2319250405,-0.2067666501,-0.276489377,0.2609707117,-0.1469493657,0.1941198111,-0.1105744392,-0.1230229959,0.0348643623,0.0809668601,0.0615914166,0.0697648376,-0.0667475984,-0.2193890959,-0.1012234464,0.1080154926,0.150924027,-0.2101520151,-0.1320088059,-0.1538246125,0.0378321037,0.1495899111,0.0887706429,-0.0376434177,-0.2327208519,0.1793506294,0.1280825436,0.2494903803,-0.0059657963,0.0889460742,-0.2237322181,0.4293895662,0.2562603652,0.3174159825,0.1706792414,0.0563195571,0.3018046319,-0.1460427791,-0.1440250129,-0.1139339283,0.0258795973,-0.3434234262,-0.0660953894,-0.3595168889,0.1336220801,0.3118779957,-0.0975824744,-0.1731549203,0.5025810599,0.1450115889,-0.453148067,0.0248710793,0.194314152,0.1112560257,-0.0380896367,0.2233567983,0.279635191,0.0105241472,0.2125751227,0.1988218129,0.2984337509,0.2038481385,0.0003728197,0.2004333138,-0.1294393986,0.0988543853,0.2445765734,0.2491762936,0.3722976744,0.0510329604,-0.2876084149,0.1541699469,-0.2396284789,0.0483386554,0.0450484157,-0.470000118,-0.0693738014,0.293530792,0.2588562071,-0.2270623147,0.1640987694,0.4007689357,-0.0752252489,0.0162233338,-0.1122616529,0.4111482799,-0.1277391315,-0.0370320193,0.0089251827,0.0591116659,-0.1397221982,0.2419923693,-0.0883162916,-0.08203803,0.1480266452,0.1870425493,-0.2156959772,-0.1489097774,0.0460955054,0.2676198483,-0.0991446674,-0.024262473,0.2578191459,-0.1219584197,-0.3165866435,-0.2253909409,0.5133313537,0.5202716589,0.2693199813,-0.281622231,-0.2294066101,-0.2469027489,0.0673217624,-0.1268483996,0.0662369579,0.4688662291,0.4884332418,0.1340845525,0.0775634125,-0.1560686082,0.2816014588,0.2100670487,-0.0396898389,-0.0202400945,-0.0021825477,0.1215313002,-0.11585024,-0.1827841401,0.0475051291,-0.2579496503,0.4321660995,0.2590052485,0.2160276622,-0.1047228277,-0.1575865299,0.0615907423,0.1124741808,0.606103003,0.4710084498,0.5872038007,-0.2290047109,-0.3212719858,-0.5952891111,0.1831261367,-0.4418338835,0.139734596,0.0698052943,-0.0520950258,-0.2592192888,0.3727819324,-0.239553079,0.2304561585,-0.1474730372,-0.2785251141,-0.3075968325,-0.0007977721,0.0526354946,-0.3403731585,-0.0960055739,-0.2205194086,0.1419960111,-0.6766186953,0.0067319507,-0.3497225046,-0.2160630822,-0.0031101038,0.1843311489,0.5605757236,0.1048572958,0.4742826223,0.0823713392,0.0405960344,-0.1629899293,-0.2933665216,-0.2526433468,0.2444384396,0.1447579861,0.1966755688,-0.0955010802,0.1217415184,-0.2354782671,0.4573546648,-0.1319747716,0.2060926855,0.2391500324,0.099831067,-0.0740243122,-0.0384351686,-0.2348379344,0.0115595218,0.0272347424,0.0130684953,-0.5246798396,-0.3445242047,0.386816442,-0.432939142,-0.0020716148,-0.1949372888,0.1177558526,0.1000631526,-0.3089579046,-0.4167046547,0.1296416223,0.2989931703,0.084109202,-0.2802976966,0.172423169,-0.0011794192,-0.055927787,-0.0758604854,0.1903989613,0.0263923407,-0.2838076651,-0.1469335705,-0.209394291]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/435","title":"ImportWarning for pyarrow 1.0.0","comments":"This was fixed in #434 \r\nWe'll do a release later this week to include this fix.\r\nThanks for reporting","body":"The following PR raised ImportWarning at `pyarrow ==1.0.0` https:\/\/github.com\/huggingface\/nlp\/pull\/265\/files","comment_length":19,"text":"ImportWarning for pyarrow 1.0.0\nThe following PR raised ImportWarning at `pyarrow ==1.0.0` https:\/\/github.com\/huggingface\/nlp\/pull\/265\/files\nThis was fixed in #434 \r\nWe'll do a release later this week to include this fix.\r\nThanks for reporting","embeddings":[-0.0830754414,-0.1102307364,-0.0867729485,-0.139640525,0.2431191951,-0.2674666047,0.3366219699,0.3281983733,-0.1044746488,0.1609233618,-0.0459907986,0.2184556425,-0.1111827567,0.0628343746,0.027580386,-0.2568036914,0.097771287,0.2359910011,0.0056410851,-0.0169456638,0.0004630512,0.2912732363,-0.1947281063,0.3622003198,-0.2395525277,0.0270071961,0.0885716602,0.1242647395,-0.2790331244,-0.5286542773,0.1933578402,-0.1631902903,-0.0732050613,-0.0040341383,-0.0001010639,0.1438983232,0.4179382026,0.1709832847,0.1610470265,-0.0188637301,0.4765137434,-0.3481638432,0.3162206411,0.0011413468,-0.2127271444,-0.3874779344,0.2012822777,0.6739887595,0.2996217012,0.0531941764,0.352093637,0.5651536584,0.3067946732,0.2126060575,0.3988427818,-0.1721241176,-0.0575685389,0.2334617823,0.0991275385,-0.4453291595,-0.2211328596,0.106356211,-0.0244036019,-0.0982079208,0.0148001928,-0.0480629094,0.0487468578,-0.1012411714,-0.2533308566,0.03156152,-0.1268713623,-0.2548316419,-0.1118742824,-0.0782755986,0.0606308952,-0.4917188287,0.2599356771,0.2389288992,-0.0639472902,0.2154643387,0.0959483609,0.0326453596,-0.1715609282,0.1390209645,0.0803394243,0.5061997771,0.0320007168,-0.0365809537,0.1415236145,-0.1213111877,-0.0105357459,0.1806747019,-0.065680109,-0.0150712943,-0.0555665717,-0.051771313,0.2417271435,0.2598069906,0.126937747,-0.1598719954,-0.053952124,0.1668393463,0.1309002042,0.2279284447,0.0629451945,0.0282733031,0.1008226648,-0.3034647107,0.3983899057,0.0791475028,0.1498911083,0.0987928882,0.15991956,-0.3288703859,0.0572980009,-0.0595705882,0.232298106,-0.1589361131,-0.1830835789,-0.1484579891,-0.2489197403,0.0096771494,-0.099664323,0.269536823,-0.0973090604,-0.0196081437,0.0986448377,0.262240082,-0.3504638672,-0.2332783192,-0.2548336983,0.0658039004,-0.3251521587,-0.0997655764,0.1824346632,-0.0633449033,0.3893645108,0.299908042,-0.0197889134,0.1312524229,-0.0293195471,0.2258119583,-0.00648562,0.2267420739,0.0049192179,-0.0279853512,-0.0467880853,-0.1139084473,-0.0660101324,0.0741590858,0.0568603761,-0.1301459819,-0.3839268684,0.3116227388,-0.2090824097,-0.104519397,0.0282637663,0.0513054468,0.1286733598,-0.0548700988,-0.0117081562,-0.1456808895,0.0393042564,-0.1020550281,-0.0698948428,0.0730507001,-0.1338871121,-0.1944649667,0.2042917013,-0.4161705375,0.2749917507,0.3692949414,-0.2970252335,-0.0789089277,-0.2009683996,0.3247884512,-0.0235642958,-0.0966951475,-0.1888201982,0.0235766321,-0.266833812,-0.0348267853,0.0754944086,-0.1217662469,-0.2040691674,0.0721507594,0.0707328022,0.3919305801,0.1285731643,0.2569903433,-0.4870685339,-0.2503657341,0.1534605175,0.0033284535,-0.0707069114,-0.2464265823,-0.0404124409,-0.1909304559,0.3194949925,-0.1372527778,-0.0349238999,0.0468536839,0.0672133341,0.0690261126,-0.0961697549,0.0366544388,-0.1779856831,0.021493515,-0.3616137803,0.3930594623,-0.1452844292,-0.3226929009,-0.07861007,0.0359093808,-0.1544843316,-0.1556718946,0.3084644675,-0.0589612313,-0.021510534,0.2313106507,0.0603005849,-0.1698938459,-0.0058377567,0.1254486144,0.0518530421,0.1411392987,-0.3814891279,-0.391199708,0.1973904222,-0.0112614939,0.1413132995,0.0031072774,0.1131301299,0.0895699263,-0.1998307854,0.3266761899,-0.318072319,-0.0501882657,0.134268716,-0.3729337454,0.0234289225,0.0651890188,-0.1765625924,-0.0148426471,0.2517152131,0.4727603793,0.2811139822,-0.1402591318,-0.0036962843,-0.1384207755,0.1402438134,-0.1233885586,0.0942482874,-0.0179691501,0.0567531511,0.1691510975,0.061151512,-0.1107048243,-0.3336109221,0.0328680091,0.3761782646,0.0379981548,0.1958342642,0.0069674528,-0.0281108972,-0.1388168633,0.2378657162,-0.4669159949,-0.0157597661,0.2753902078,0.0200040098,0.2059604824,-0.2243602127,-0.4125036895,0.048190143,0.0607327484,-0.0026752418,0.1401378214,0.4111925662,-0.2279912084,-0.7777081132,0.2353706658,-0.081293717,0.0095219081,-0.2907804549,0.0664653704,-0.29391554,-0.3391764164,-0.0755752698,-0.2050065994,-0.3589579761,-0.0981444642,0.3100822568,0.1755020916,-0.2631780505,0.3151647747,0.3475986719,-0.0642475411,-0.0756793693,-0.1202402487,-0.1154793203,-0.2051045746,-0.2663134038,0.1959459931,0.0624703765,0.0458264984,0.1699760705,-0.1139839962,-0.0664578676,0.0129902996,-0.5971788168,0.0656021759,-0.2121935189,0.1207150519,0.193306759,0.1167021841,-0.0155746909,-0.2416051924,0.229631871,-0.3196700215,-0.4370693564,-0.0598572269,-0.159314692,0.1028265432,-0.3726234436,-0.1548138708,-0.1904674768,-0.3148061037,0.1847970039,0.0559411347,0.0161652267,0.310484916,0.1640842408,0.0933733359,-0.317596823,0.0469761863,-0.0244123619,-0.1549781263,0.16934219,-0.2249408513,-0.4008027315,0.0334066935,0.0106638968,0.0975719392,-0.3147257268,-0.4104326069,-0.3181371987,-0.0403931998,0.5853024721,0.3685545027,-0.0770060644,0.289948076,0.1316108555,-0.2970892787,-0.052383028,-0.3068110049,0.0330360904,-0.1884214878,0.0231340695,-0.169214949,0.0882597715,-0.0709627643,0.6686916947,-0.3755657971,-0.1222009882,0.5347418785,-0.0492617823,0.3575075269,0.0169340614,-0.0162534211,0.0615126342,0.0671216622,0.2661387622,0.1894720197,-0.1099519879,-0.4191838503,-0.045481503,-0.4166472256,-0.0361576453,-0.2257389128,-0.0053758682,0.5628577471,-0.0140687665,0.1647192687,0.0120681114,0.0173970293,-0.269516021,0.2594310045,0.0008520787,-0.1660767794,-0.0606491305,-0.1289081872,-0.0638436005,-0.3101232648,0.0570143387,0.2350649089,-0.0570891164,-0.0245720632,-0.258379966,0.1342057437,-0.1056805402,0.4005034566,0.2934625149,0.1665647477,0.2135399282,0.0701527968,-0.2036423236,-0.1704911739,-0.1556848884,-0.0211958215,0.3835706413,0.0528605282,-0.438958168,-0.1754422188,0.300475806,0.282010287,0.0027459546,-0.0153147755,-0.5284619331,-0.2185235769,-0.1631642282,0.1246434897,0.0985189825,0.2308525592,-0.1832678765,0.062419679,0.0380949713,-0.1195346713,-0.1824250072,0.371053189,0.0309958439,-0.1367566586,-0.007447754,-0.0881538391,-0.0357663259,-0.0579177365,0.3514465094,0.0464074761,-0.2605164945,-0.122331813,0.2231657803,0.0523231439,0.1710094661,0.0057619666,0.1916316152,-0.1243805289,0.0281496719,0.1198759601,0.1390536129,0.349753052,-0.0350488797,0.2092650533,0.2415168881,0.3253231347,-0.0578267165,0.0641008243,-0.0339834951,0.230105117,-0.1437182128,0.220985055,0.0451653078,0.820797801,0.0330351256,0.2702249289,0.6261534691,-0.2362233847,0.3250900209,0.1172436774,0.2000279725,-0.3581175208,0.0033055148,0.2600872815,-0.0228623878,0.0847516358,-0.1178092659,0.018182192,0.0287804361,0.1112652197,-0.0092795733,0.0248556715,0.1781781912,0.1207873076,-0.0127579952,-0.3865138888,0.1847656369,-0.009845878,0.2434541732,-0.1725934297,-0.301813215,0.2809734643,-0.3456578851,0.0044910279,0.3744012713,0.0178187303,0.1027278677,0.2606951594,-0.2044039071,-0.1182845756,0.0858620554,-0.0166333485,-0.0952289924,-0.1101046726,0.3050880432,0.0727003366,-0.2583871484,-0.1442807764,0.3166054487,0.0560269691,-0.2428621501,-0.3549968004,0.2014134824,0.0469358452,-0.0237467103,-0.1675528586,0.0315039717,-0.0152524495,-0.0413667299,0.1499981731,-0.0467620417,0.1018386856,-0.2062679678,0.2574655414,0.0557672195,-0.0219871104,-0.0316199958,0.2130368501,-0.1793032885,-0.2644487917,0.4202330709,-0.1042945683,0.0694811493,0.3385528624,0.3837928474,-0.3116826117,-0.212202847,-0.2277960628,0.061923217,-0.570969224,-0.007539222,0.2905113101,-0.1169249266,0.0515387319,0.3112246692,-0.0189898163,-0.0013939731,0.0555657335,-0.0304722879,-0.1131909639,0.1001548842,-0.0573813319,-0.0165950116,0.4399378002,0.4364564717,0.1472838819,-0.1039575711,-0.4113297164,0.0673326254,-0.1125503406,0.138101235,-0.2909931839,-0.2318249047,0.2360016704,0.1290106475,0.2740167677,0.0868183374,-0.1110764146,-0.3384961188,-0.1322936267,0.0018455217,-0.0048814458,-0.0717094764,0.1792272925,0.2255895585,0.0820273682,0.1244025603,0.309689492,-0.0295081083,-0.150388822,0.2898205221,0.1425187439,0.035309609,0.146583274,-0.2073167413,-0.1482466012,0.2773084342,0.1130541489,-0.0286237523,-0.1440726817,0.0268369764,0.2978827059,-0.0621425025,-0.1109967604,0.1789117455,0.1344165802,-0.2298986167,-0.3243172169,-0.2257203311,0.2501014173,0.1445132345,-0.2257966399,-0.1727882773,0.3819535077,0.3430961967,-0.4943426549,-0.0907828435,-0.0316528603,0.1363242716,0.0802178606,0.0350175016,-0.0199275706,-0.0952978283,0.196745649,0.3701377511,0.1514445096,0.0223533157,-0.2284718305,0.0745001659,-0.2980050147,-0.1211713701,0.45496732,0.1497988254,0.2490103245,0.0247083027,-0.1215824708,0.0598121025,-0.2983395159,0.1750699133,-0.1613067389,-0.1363921911,-0.1171570718,0.3793145716,0.2954331338,-0.1782367676,0.1823404878,0.4003160894,0.2745031118,-0.1360537708,-0.0515779406,0.605923593,-0.1846618503,-0.0462746173,-0.0622539222,0.124379769,0.0385912545,0.2332673669,-0.0319228545,0.0491470993,0.2032893598,0.0685300454,-0.1821649671,-0.3582143188,0.0103542805,0.0803233236,0.2148051858,-0.0225482415,0.2984034121,0.2142087817,-0.1925636381,-0.1964099705,0.2984684706,0.4337421656,0.1835556328,-0.054359287,-0.1605568975,-0.2027054131,0.0328567475,-0.2084278911,0.1702623665,0.2215856612,0.5038293004,0.1546589881,0.1907963753,-0.370556891,0.0332400501,0.1691541225,-0.0247028004,0.0629143193,0.0638836026,-0.0297709666,-0.2653668225,-0.3680840433,-0.1542306989,-0.3825883567,0.2647891641,0.098038584,0.2649374604,-0.0150410859,-0.4861194193,0.1417194009,0.1846014261,0.6637275219,0.2069656104,0.268461436,-0.1088390201,-0.0998880565,-0.4346327484,0.3114922941,-0.3065361083,0.0421797223,-0.0522691235,0.0214559324,-0.1013638675,0.2282588333,0.0045616012,0.0757816955,-0.0998053253,-0.2323069721,-0.3131775856,0.3172436357,0.0876188576,-0.2602440715,0.1048862487,-0.1456599832,0.2333258986,-0.4985124767,0.2598397732,-0.2432782203,-0.0090720914,-0.1888140142,0.3120508492,0.2228735834,0.1276967674,0.1513318419,-0.0133229652,-0.0628322959,-0.141747877,-0.3774041533,-0.2469451278,0.1444800347,-0.0085265208,0.1654462516,0.0690770522,-0.1187744141,-0.204598859,0.4295692444,-0.0040786006,0.078015238,0.1664197445,0.2578724623,0.1175777689,-0.070656307,-0.2239874452,0.0518489964,0.0200977698,-0.2055508047,-0.4341712296,-0.5926018357,0.347640574,-0.4550389647,0.0570193082,0.0646058694,0.148239851,0.1230561286,-0.394347012,-0.4716579616,0.3042547107,0.1011684537,0.1949093938,-0.2898897827,0.1886667609,0.0005848549,-0.0815228671,-0.0591255315,0.1133671552,0.2314946502,-0.2264800519,0.089315854,-0.1714505404]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/435","title":"ImportWarning for pyarrow 1.0.0","comments":"I dont know if the fix was made but the problem is still present : \r\nInstaled with pip : NLP 0.3.0 \/\/ pyarrow 1.0.0 \r\nOS : archlinux with kernel zen 5.8.5","body":"The following PR raised ImportWarning at `pyarrow ==1.0.0` https:\/\/github.com\/huggingface\/nlp\/pull\/265\/files","comment_length":31,"text":"ImportWarning for pyarrow 1.0.0\nThe following PR raised ImportWarning at `pyarrow ==1.0.0` https:\/\/github.com\/huggingface\/nlp\/pull\/265\/files\nI dont know if the fix was made but the problem is still present : \r\nInstaled with pip : NLP 0.3.0 \/\/ pyarrow 1.0.0 \r\nOS : archlinux with kernel zen 5.8.5","embeddings":[-0.1463128328,-0.0825235769,-0.0497649387,-0.0821150839,0.1566722393,-0.2191777378,0.1178104952,0.4189058244,-0.2502535582,0.1290670931,0.0764926895,0.2807854414,-0.2030645758,-0.1170991436,0.0609498955,-0.1767264009,0.1294214725,0.2613931894,-0.0432577878,-0.0118179601,-0.1181317866,0.379407227,-0.3579395711,0.4410783052,-0.1741458625,0.0080258381,0.1847311258,0.1068225801,-0.274545908,-0.7176750302,0.3703352213,-0.255653888,0.1187354326,-0.1453176588,-0.0001122298,0.1098703891,0.4844831526,0.1264279336,0.2048352659,-0.2869176269,0.3934236765,-0.7273046374,0.3218928874,-0.0929811299,-0.0015563931,-0.3426614404,0.2967853546,0.7822851539,0.3996650279,0.1758708507,0.2156397253,0.5306107402,0.3595175743,0.3860240281,0.3381447494,-0.231363371,-0.0678984895,0.2951617837,0.053193137,-0.5167246461,-0.0443867557,0.150931865,-0.1811769754,-0.1192705855,0.0169097781,-0.1557048559,0.2961479425,-0.2603829205,-0.3814111352,0.086864017,-0.2023055851,-0.2443889529,-0.1900428832,-0.0965145975,0.2009543478,-0.4492753148,0.3753640056,0.2514262199,-0.0926634148,0.2638771534,0.1020497903,0.1437640041,-0.2070808858,0.2501846254,0.1616183519,0.7215542793,0.1772344261,0.1273999959,0.1793180257,-0.0906528607,0.2533272505,0.1564489454,-0.1677651107,0.241388306,-0.0913616642,-0.1201786697,0.1022301018,0.3962029219,0.0492236875,-0.1726308763,-0.1069478914,-0.033830151,0.2005792707,0.1612658203,0.1280685365,-0.0260971393,0.2030237317,-0.2324667126,0.2462064177,0.1590737998,0.1445098966,0.0729045793,-0.0251889694,-0.3879301548,0.1451650858,-0.0661641955,0.3707583845,-0.1980346739,-0.2538054585,-0.2639258802,-0.3639071584,0.0195967183,-0.0219472293,0.1275433749,-0.1162193418,0.1894862354,0.0578527153,0.193916589,-0.3054687679,-0.1874005944,-0.0425399952,0.1129766926,-0.3520491123,-0.0135481702,0.0832812861,-0.1224773899,0.4979560673,0.22464104,-0.1518140584,0.0297393724,0.0791248679,0.1071715876,0.0368607044,0.2614901364,-0.0684240013,-0.0179729182,0.0009729172,-0.0820256397,-0.2350455076,0.2638747096,-0.0133769978,-0.2735418081,-0.4603399038,0.2024455816,-0.3389339149,-0.127925694,0.1861551106,0.0615641959,0.1776736379,-0.3524859548,0.0419456065,-0.3899518549,0.1260893345,0.0136418939,-0.1244788542,-0.0413493179,-0.0070408615,-0.1972961873,-0.032097429,-0.4223334491,0.3923664093,0.3698949814,-0.3251792192,-0.0223598611,-0.097861886,0.2279029638,0.2245870978,-0.3489607275,-0.3137609363,0.0563647039,-0.1993921846,-0.1143462732,0.2462938875,-0.0272548515,-0.1828125864,-0.0191403497,-0.0409662649,0.3157671094,0.0823566169,0.1319957525,-0.4862550795,-0.2567749023,0.1407622397,-0.022390347,0.0237175096,-0.1570117325,-0.0441814922,-0.0445421115,0.3431508839,-0.0166793838,0.0339643471,-0.0141071463,0.0175170042,0.1112060845,-0.1198917255,0.0779933929,-0.0134687936,-0.022679586,-0.5015802383,0.3683543801,-0.191938594,-0.3551448882,-0.1008395255,-0.007090318,-0.0447308198,-0.1006916314,0.1612117141,-0.1429770291,0.0715140477,0.2534148395,-0.0067576789,-0.3259921968,-0.0255912319,0.1006076783,0.0014618938,0.1108679995,-0.3691414595,-0.507037282,0.028635703,0.2058620304,0.2018126994,0.1119769737,0.0981957391,-0.0303253774,-0.3082160056,0.2859077156,-0.4791698456,-0.1952307373,0.1172275543,-0.4769582152,0.0436512791,-0.0142154777,-0.168717891,0.1005290821,0.4360937476,0.4499976933,0.1965037286,-0.0531026796,-0.1498875022,-0.0545801818,0.1406523734,-0.1640532166,0.0259623844,-0.0036244569,0.1470006406,0.1947016716,0.1001415178,-0.0109529803,-0.1298136264,-0.1207543835,0.3967891037,0.0622103475,0.0553406514,0.0301933233,0.0124041168,-0.1178832129,0.3387461901,-0.3547405303,0.0835553184,0.3039248288,-0.0718043596,0.3368053138,-0.1803065091,-0.290581435,0.1546374857,0.1677264273,0.0824358612,0.2134037614,0.4092847705,-0.1627025008,-0.6508201957,0.1368492395,-0.0687040836,0.1575957686,-0.1914118975,0.0635445192,-0.2776795626,-0.4018160403,-0.1863512397,-0.2930063903,-0.3364791572,-0.1061582565,0.1684437394,0.3411311805,-0.0711148754,0.3280318975,0.3384560943,-0.175470531,-0.2444175482,-0.0481759757,-0.1508713067,-0.2504651248,-0.3752191663,0.0670826584,0.2000453919,0.1900669634,0.1615685672,-0.14503178,-0.2201214731,0.08467336,-0.4907344282,0.0075798118,-0.2482957393,0.1981133223,0.1954679489,-0.0094684558,-0.1259321868,-0.2910656035,0.1533661038,-0.3169594407,-0.3713766932,-0.145677343,-0.3450291455,0.2003509104,-0.3035596907,-0.0370836556,-0.4336920381,-0.2729843855,0.184934184,0.0883676559,0.0782099664,0.3221369386,0.0923085436,0.1517430842,-0.27811414,-0.1642434001,0.125639379,-0.0698492751,0.1968343407,-0.0362192355,-0.2200060338,0.1404721141,-0.0261148494,0.1026242673,-0.2304704189,-0.4257997274,-0.2739188373,-0.0281991493,0.6165572405,0.2406507879,-0.1069806814,0.4314074516,0.07985311,-0.1554450989,0.0156342946,-0.1081204563,-0.0195611268,-0.2196579427,0.1145343706,-0.0134998597,0.1643051952,0.0284330267,0.7524285913,-0.2021986991,-0.0865355507,0.5082700253,-0.0427492261,0.2506545186,0.1585089266,0.0116206193,-0.0751903579,0.2429630309,0.155157119,0.0164738894,-0.1186723486,-0.316737771,0.1006586999,-0.4192097783,0.0081104077,-0.0788079873,-0.0037511277,0.4223539531,0.1017467231,0.2647047341,-0.1091939285,-0.0516460352,-0.4151854217,0.182240501,0.133654356,-0.1440276951,-0.0040657809,-0.0660975277,-0.3085956573,-0.438986212,0.0151167251,0.3585349321,-0.0007147406,0.0156408325,-0.1333144605,0.284825474,-0.0649421737,0.2460818887,0.144722417,0.0827081054,0.2315773815,0.1617176235,-0.249564752,-0.1933167577,-0.1135947779,-0.0307921469,0.4878688157,0.0340217277,-0.3158238828,-0.1914644837,0.6278867126,0.2776323855,0.007831241,-0.1595433056,-0.5531588793,-0.3403336704,-0.0945027992,0.1357285678,0.0595889017,0.2282238603,-0.0856160671,0.1425572485,-0.0233295001,0.1138454378,-0.2402732372,0.2719141841,0.0305835102,-0.1417432576,-0.0802137777,-0.1551506817,-0.220905304,0.0045707528,0.2699087858,0.2285877466,-0.2923704386,-0.2940796018,0.3749374151,0.0400975123,0.0891682655,0.0762604102,0.1906944811,-0.1389034837,0.0279945619,0.1909882128,0.0107700238,0.3302236795,0.0095628444,0.205278635,0.2049923688,0.3583036959,0.075907208,0.0203047413,0.1160156727,0.0630372167,-0.3131756186,0.1893791407,0.1192527488,0.7727373838,0.0468878187,0.2768118978,0.780810535,0.0381207168,0.3821668327,0.2557758391,0.1406747997,-0.3187006414,0.0656814277,0.1776507348,-0.0057661296,0.1608769596,-0.1035435945,-0.0871679708,0.1506629735,0.3170361817,-0.1086102799,0.1454940587,0.1781952977,-0.120065704,0.0131124835,-0.4551359117,0.0432029888,-0.0728662089,0.3411375284,-0.1738136858,-0.3098094463,0.2617656291,-0.2420981675,-0.0631484017,0.2948921919,0.0729022622,0.18008174,0.1867466569,-0.097966902,-0.0169631299,0.2230789363,-0.0047040582,-0.1608380675,-0.0243574269,0.2252328694,0.0777618513,-0.3879622519,-0.0755177811,0.3331819773,0.0539720133,-0.2730444074,-0.384221375,0.1162783131,0.0047676922,0.0226523932,-0.0530640371,-0.0094610872,0.0467624255,0.0904525518,0.0458837748,-0.0603076927,0.0528452098,-0.1239376068,0.1723189652,-0.0085582593,-0.0709716901,-0.0210493002,0.0537556931,-0.3723710179,-0.1013084054,0.3956664503,0.017361261,0.0825557336,0.5533752441,0.1342066377,-0.1725985557,-0.1685627252,0.0888208896,-0.1204932928,-0.5595656633,-0.0649083033,0.3915897906,0.0053681191,0.0524303205,0.3142743409,0.1729961187,0.0728257224,-0.140964061,-0.1577625424,-0.1599019021,0.2311978042,-0.0303481054,-0.2218043059,0.3306155205,0.4356178939,0.1888486445,-0.2454167753,-0.2971188426,0.0243450496,-0.1334160119,0.2212602049,-0.2471286803,-0.10504134,0.0576607622,0.0880895257,0.116614528,0.0358034782,-0.055749286,-0.2365225255,-0.1890358925,0.0993035361,0.0874451101,-0.1711909622,0.0329070501,0.1443337798,0.1743557453,0.1258205771,-0.0252644997,0.0263153873,-0.0780581683,0.3505366445,0.1515222788,0.1835598201,-0.0883225054,-0.0433657914,-0.3393918276,0.2155554146,0.0755553022,0.1742344648,0.0559472479,0.0806269273,0.1038271859,-0.2235939056,-0.1362440586,0.0465387814,0.0684805736,-0.2091190368,-0.3051457107,-0.1213176325,0.1599951088,0.2268363982,-0.2400157452,-0.2193303704,0.395960331,0.2351083457,-0.3919550478,0.0256279167,-0.0590791665,0.0732116252,-0.1173321605,-0.0312863998,0.044725243,-0.2196570039,0.2954949737,0.2880702019,0.1427744925,-0.1183972061,-0.1173934042,0.2002957314,-0.0627440363,-0.0369548649,0.4063513279,0.0467905514,0.3146163821,-0.1146048903,-0.1296189576,0.2657221854,-0.2911003232,0.294375807,-0.3425343335,-0.1696187854,-0.2282963842,0.368937701,0.2435631156,-0.2419520617,0.2854421139,0.3730116189,0.2542885542,-0.1198335439,0.0358052701,0.6180189252,-0.148903951,-0.121696353,-0.2572516203,0.2055373341,0.0619052537,0.2556100488,0.0212983731,0.1901595742,0.2496865541,0.2022270858,-0.2579716444,-0.3524663448,-0.0519879498,0.3320541978,0.2005601227,-0.1307684481,0.2552770674,0.1283192635,-0.1924657673,-0.2567723393,0.4530804455,0.6163247824,0.2657215297,0.1032017767,-0.220965609,-0.2673633993,0.0752575919,-0.1873039603,0.2475881726,0.2774343491,0.774667263,-0.0107691558,0.0944807976,-0.281193018,-0.0117216967,0.1550169736,-0.2454235405,-0.0142716058,-0.0159770623,0.055208534,-0.3199709058,-0.2273922712,0.1110943481,-0.3247716129,0.3930579722,0.0914084539,0.333445251,-0.0036915613,-0.446783632,0.0685936734,0.0960259885,0.4824586213,0.2509271204,0.1322217137,-0.2348388284,-0.1713162214,-0.5863732696,0.2388712913,-0.4293252528,-0.0694376379,0.1120382696,-0.1098492444,-0.3579770029,0.2536035478,-0.1213655546,0.1574552357,-0.1195030361,-0.2146402597,-0.1870663166,0.3027914166,0.2191271335,-0.2342185825,-0.0289156623,-0.2543509007,0.2548310757,-0.5901117325,0.0814794526,-0.2319281995,-0.0342657566,-0.0690837726,0.5230580568,0.1959623992,0.1945030838,0.3007864356,0.0861575305,0.1274977475,-0.1984216869,-0.3919112384,-0.3851025403,0.2205200344,-0.0796589702,0.1576739252,-0.1454688758,-0.0824467987,-0.2672022283,0.4386237264,-0.1976332814,0.0326679572,0.2748295665,0.1524995118,0.0448475853,0.0765509531,-0.1438394636,-0.0142237917,0.0946588665,-0.197076112,-0.5314705968,-0.4598428905,0.2082968205,-0.5019981265,0.0528521352,0.020666495,0.2318798751,0.1214507744,-0.4393998981,-0.4035781622,0.1622241884,0.2716173828,0.2200743854,-0.1676883101,0.2431374341,0.1953085214,0.0246983729,-0.0125452252,-0.1446715891,0.1201960966,-0.3154450059,0.1479803771,-0.2379944324]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/435","title":"ImportWarning for pyarrow 1.0.0","comments":"Yes it was fixed in `nlp>=0.4.0`\r\nYou can update with pip","body":"The following PR raised ImportWarning at `pyarrow ==1.0.0` https:\/\/github.com\/huggingface\/nlp\/pull\/265\/files","comment_length":11,"text":"ImportWarning for pyarrow 1.0.0\nThe following PR raised ImportWarning at `pyarrow ==1.0.0` https:\/\/github.com\/huggingface\/nlp\/pull\/265\/files\nYes it was fixed in `nlp>=0.4.0`\r\nYou can update with pip","embeddings":[-0.0887958035,-0.033828184,-0.0959169269,-0.1748253107,0.1985275149,-0.2785079479,0.2463124543,0.2891695797,-0.1062946692,0.1037860215,-0.0099450247,0.2737372518,-0.1364479959,-0.0154105769,0.0474855714,-0.2346006483,0.0826648101,0.2507703602,-0.0461506844,-0.0209529027,-0.0540119559,0.3300443888,-0.2781843841,0.3739852905,-0.201450631,0.005122575,0.1078910753,0.0617049858,-0.2494084388,-0.6246519089,0.2546360195,-0.1434793323,-0.0067362804,-0.1019117385,-0.0001038836,0.1097955629,0.4532860816,0.1570789516,0.1588119119,-0.1238538548,0.4919243753,-0.5085181594,0.4108836651,-0.1071738303,-0.1413876265,-0.3619322479,0.2613588572,0.7227515578,0.3328154385,0.1153785959,0.3080121577,0.5798754096,0.3169381022,0.3164127469,0.3628939688,-0.2040487677,-0.048843585,0.2716396451,0.1007202342,-0.4919560552,-0.2283340544,0.0974417999,-0.0895865336,-0.0340818688,0.0080652544,-0.0732715875,0.0937181041,-0.1534430087,-0.3086854219,0.0513302237,-0.1180155948,-0.2796166837,-0.1179684177,-0.0572624877,0.1298791766,-0.5174029469,0.2554970086,0.256825,-0.0865328759,0.1703187078,0.0824096128,0.0139359459,-0.2192866355,0.2140568495,0.1625310183,0.6374462247,0.0839986801,0.0052108467,0.162882477,-0.1013770476,0.0313673578,0.215983808,-0.0534923337,0.0714349449,-0.0672209188,-0.1003553942,0.1538566649,0.3231523037,0.0739101619,-0.2159889489,-0.0708229914,0.0871387869,0.1247187331,0.2291898131,0.0316386223,0.0198885016,0.1405788213,-0.2801316977,0.3816335499,0.074399896,0.1447949558,0.0859094188,0.1416918486,-0.3412117958,0.0797833949,-0.0144033665,0.2743059099,-0.1328958273,-0.1920820475,-0.1728955358,-0.2420265079,-0.0267332643,-0.0703213587,0.1926949173,-0.1459538192,-0.015501353,0.1114854664,0.2463323772,-0.3668373823,-0.2467447072,-0.1676756442,0.1046226472,-0.3267345428,-0.1436270326,0.1788096577,-0.0441754945,0.4089523852,0.2682935894,-0.0797365457,0.0969834551,0.0069762208,0.1813012511,0.0156936571,0.1844247133,-0.0109243728,-0.0486034788,-0.0341097079,-0.116628401,-0.157015875,0.1149572283,-0.0248533934,-0.1184800789,-0.474414885,0.269089222,-0.2345084846,-0.122646451,0.0351175666,0.0177429654,0.1258657575,-0.1594532281,0.000473761,-0.2204062939,0.0678814426,-0.0744194984,-0.1213380173,0.022787407,-0.0634302795,-0.1726652235,0.0638429001,-0.4047810733,0.3786534369,0.3395776749,-0.3153258264,-0.0598730519,-0.1434033066,0.2855083644,0.1449688077,-0.1852440685,-0.1968856752,0.0803112984,-0.2074363977,-0.0865986422,0.0604753308,-0.0772421956,-0.2617935538,0.0512018204,-0.0024306946,0.4208135307,0.1926575154,0.246824339,-0.5017081499,-0.2710357904,0.2263769507,-0.015137217,-0.0279199798,-0.2329066545,-0.0734960437,-0.0961408094,0.2907433808,-0.1050810069,-0.0716601387,-0.0318016186,-0.0122185135,0.1021426618,-0.1031897739,0.0709745735,-0.1702965945,-0.0597828962,-0.4313760996,0.4168561995,-0.1554896384,-0.3134540915,-0.0623953342,0.027059257,-0.1133041754,-0.1369260252,0.2785752416,-0.0826175436,0.012674069,0.220685944,0.0637833029,-0.2395492196,-0.0639537424,0.1126127094,0.012170353,0.0988189653,-0.4061357677,-0.3946220279,0.2050596029,0.0634046718,0.1821092963,0.0400718264,0.090933539,0.0632697716,-0.2519937754,0.3157958984,-0.3908134401,-0.1246315315,0.1425686926,-0.3782959282,0.0801908672,0.1184554249,-0.1808637381,0.0652330145,0.3329956532,0.4405824244,0.1996188313,-0.1649824977,-0.0435097851,-0.0909840316,0.1523578018,-0.1179003492,0.0950278938,0.0005113626,0.1053317413,0.1821143627,0.0787797868,-0.1125839725,-0.2973203957,-0.0453743376,0.3904305995,0.0656034574,0.2314381599,0.0290878769,0.0160617735,-0.1738462597,0.2585172951,-0.4787358046,0.0138339065,0.2942165136,-0.0037985221,0.2227261961,-0.2454726696,-0.4008094966,0.057306163,0.0990479663,-0.0212279782,0.118632488,0.4009287357,-0.2036633193,-0.737125814,0.2283800989,-0.1093450561,0.0960761532,-0.2462783456,0.1037544608,-0.2999864221,-0.4105066657,-0.1095280796,-0.202008456,-0.3600592315,-0.0815815106,0.2955844998,0.1405380964,-0.2027724832,0.3078516126,0.3354777098,-0.1210756153,-0.1472278088,-0.1063657776,-0.1282149553,-0.244213447,-0.2923542261,0.171086207,0.1037904993,0.1531175524,0.1545741558,-0.1537151784,-0.1063714847,0.0557721555,-0.5341647863,0.015955558,-0.1964331418,0.1464742869,0.2183893919,0.0120724207,-0.0941516683,-0.2997791767,0.1967719495,-0.3384428322,-0.3528697491,-0.076646246,-0.2080633342,0.1417075694,-0.3213407993,-0.1484097242,-0.3369484246,-0.3015047312,0.1760526597,0.1099090353,0.089679569,0.2720889449,0.0613812096,0.0643591359,-0.2931687534,-0.0121604623,0.0092619313,-0.1049911305,0.1688026488,-0.1689461619,-0.3760809004,0.0689203441,-0.0060051857,0.139388293,-0.2798192501,-0.4034070373,-0.3512242138,-0.0132177602,0.544228971,0.308791548,-0.0740899295,0.3209848106,0.0875963792,-0.2548921704,0.0095199402,-0.1651950181,-0.0062509417,-0.1632042378,0.0494531505,-0.1068131775,0.0991689414,-0.0618377142,0.6610437036,-0.3498865068,-0.1755620241,0.5596804619,-0.0615607239,0.3484246731,0.0905523524,-0.0398639813,0.0314371362,0.1105945036,0.2748850286,0.1395077705,-0.1948826611,-0.3998075426,-0.0045302939,-0.5025673509,-0.0760498121,-0.216926679,0.0199100152,0.5369777083,0.043847423,0.1720302552,-0.0112507548,0.0134638185,-0.3344857991,0.2067508101,0.0369349867,-0.1254958063,-0.0336057507,-0.127305001,-0.2096169442,-0.3258736134,0.0802407637,0.2575268149,0.0038429336,0.033326,-0.1345870048,0.1728848964,-0.0596345291,0.3139970601,0.2209982425,0.1605253071,0.2510507107,0.1907233298,-0.2223203331,-0.1647136062,-0.1513599008,0.049523972,0.4444863796,0.0100375963,-0.402992487,-0.2009108067,0.3995757699,0.2508807182,-0.0103584137,-0.0483647957,-0.5392416716,-0.2108495831,-0.1294262707,0.1735776216,0.0813347474,0.2417630404,-0.1480500698,0.1532674283,0.0235070083,0.0018348917,-0.1979600936,0.3541355431,0.0139611606,-0.0930596814,-0.0876269192,-0.1431504041,-0.0988347307,-0.1376075298,0.2377427667,0.1169218794,-0.1657608002,-0.1627341956,0.3146299422,0.11347159,0.0886990204,0.0382511131,0.2011462003,-0.1718612164,0.0327323154,0.1273730993,0.1884102523,0.3172276318,-0.0490300953,0.1876971573,0.1819545925,0.3705884516,-0.0436971001,0.0446972139,-0.0367103107,0.1464150399,-0.1925344318,0.2139190137,0.1096570492,0.7044849992,0.0436847284,0.263923198,0.6396080852,-0.1463734359,0.4524229169,0.119359538,0.1906842291,-0.3160387278,0.0719647482,0.2422892004,-0.0107613001,0.1316708773,-0.0974083245,-0.0521023981,0.1517879367,0.2346236408,-0.1254941672,0.0098520508,0.2626195252,0.0594848134,-0.001691633,-0.4000330269,0.1342759132,-0.1303358376,0.2979826629,-0.1523356736,-0.3068301678,0.3502237499,-0.3495553136,0.021292733,0.3407769799,0.0265244227,0.1108118668,0.2339736223,-0.2024665326,-0.021858355,0.1302127242,-0.0724269301,-0.0447398163,-0.0594626069,0.2558226287,-0.0220249631,-0.3223077357,-0.1142201424,0.3820916116,0.0296753533,-0.2768679857,-0.3608280122,0.1879999787,0.0838174596,-0.0254551079,-0.094484143,0.0449162088,0.0354232751,0.018657947,0.0888779536,-0.0092061851,0.0273816492,-0.1256231219,0.2403964102,0.0644640326,-0.0368595235,0.0308924448,0.1492221206,-0.2228762954,-0.220315665,0.3755120933,-0.0300446358,0.045293536,0.4057326019,0.3363674283,-0.297267139,-0.2187666148,-0.2060902715,-0.0650430247,-0.5592403412,-0.0451232344,0.3825663924,-0.0450207479,0.0414667353,0.2925806642,-0.0376158245,0.0801966861,0.0066383379,-0.0901105329,-0.1114200577,0.1357477307,-0.0195620526,-0.1129109561,0.4986001551,0.4153761566,0.195714727,-0.1281870157,-0.3787007332,0.0649056137,-0.1119400188,0.1795546263,-0.3081860244,-0.212040782,0.168155387,0.1340577751,0.2248778343,0.0424478315,-0.0839260072,-0.3171558976,-0.1468663961,0.0248478018,-0.0227415431,-0.1219540834,0.1007592976,0.1775208563,0.1515349895,0.1344310939,0.1563372314,-0.0953195393,-0.1343671829,0.4299877584,0.1485817879,0.0468787998,0.1175609529,-0.1560703516,-0.1879857332,0.2485077232,0.1177902967,0.0646709576,-0.0269705988,0.0632264912,0.2306174785,-0.155865401,-0.1318875253,0.1468409449,0.1028045937,-0.1885950416,-0.340675503,-0.2639268637,0.1945462525,0.2066741437,-0.207138285,-0.2225935757,0.4253627956,0.3048168123,-0.501957953,-0.0970576406,-0.011337975,0.1640899628,0.0355834626,-0.0061703986,0.0133638531,-0.1162653118,0.2041275054,0.3471823633,0.1982111037,-0.0630232692,-0.1836340576,0.1257595867,-0.2058205605,-0.1198819578,0.437000066,0.1402089,0.2653725743,-0.0498106964,-0.1007947773,0.1234958395,-0.2221024036,0.1853147447,-0.223006472,-0.0841729566,-0.2048629969,0.4546136558,0.3041120172,-0.1314212978,0.2157051712,0.3124600649,0.3318946362,-0.0882577524,-0.0538406409,0.6328897476,-0.1457945257,-0.0728464425,-0.0737451091,0.1714680344,0.0515932217,0.2414163202,-0.0385062508,0.0368261524,0.1298440248,0.0850457624,-0.1839848459,-0.3040242791,0.0687053576,0.1659716219,0.1913112998,-0.0476175845,0.2911190391,0.1902391165,-0.133468926,-0.2524310648,0.3705947399,0.4932948947,0.240015924,-0.0329827107,-0.1852702647,-0.2066573948,0.0533370301,-0.2133054137,0.2229898572,0.232072711,0.584820807,0.0763693154,0.1521359533,-0.3414497077,0.0667180941,0.1790717095,-0.1014621183,0.0553839244,0.1197327375,0.0384089686,-0.2816334665,-0.3221763372,-0.0498939492,-0.3993813992,0.3576456308,0.1122682542,0.3250600696,-0.0203411542,-0.5498232245,0.1243097633,0.1672413945,0.6408046484,0.216611743,0.3072350621,-0.1578784138,-0.1438770592,-0.4985899031,0.2821663916,-0.333560735,0.0290715434,0.0356683247,-0.0260779858,-0.2131027281,0.288305819,-0.1103286669,0.122194007,-0.16921404,-0.2213828266,-0.2803322971,0.3392255604,0.1577763706,-0.2250754684,0.0798185691,-0.1714814454,0.2073958218,-0.5460155606,0.2042113543,-0.2770421207,0.0201664604,-0.095716022,0.4812790155,0.2085442841,0.1629531831,0.196469605,0.0077084997,0.0258558709,-0.1600299329,-0.3940214515,-0.2556814253,0.1512456238,0.0018327811,0.1260339767,-0.0023688674,-0.0494333655,-0.2213460952,0.3893404305,-0.1268590689,0.060103707,0.2717795968,0.1988350004,0.0839133039,-0.0696113557,-0.2612526119,-0.0038107221,0.0697052702,-0.1930490136,-0.464037329,-0.5378295183,0.2817750275,-0.5429243445,0.0519982465,0.0217066705,0.1777061224,0.1440891773,-0.356808424,-0.4508270919,0.2507298589,0.1916571707,0.2472473681,-0.1959851682,0.2298592627,0.057105545,-0.0906434953,-0.0354756191,0.032716997,0.1949905157,-0.2999774516,0.0673695952,-0.2475644499]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/435","title":"ImportWarning for pyarrow 1.0.0","comments":"Sorry, I didn't got the updated version, all is now working perfectly thanks","body":"The following PR raised ImportWarning at `pyarrow ==1.0.0` https:\/\/github.com\/huggingface\/nlp\/pull\/265\/files","comment_length":13,"text":"ImportWarning for pyarrow 1.0.0\nThe following PR raised ImportWarning at `pyarrow ==1.0.0` https:\/\/github.com\/huggingface\/nlp\/pull\/265\/files\nSorry, I didn't got the updated version, all is now working perfectly thanks","embeddings":[-0.0635246038,-0.0892708451,-0.1202820316,-0.1328860074,0.1825052351,-0.2968225181,0.3555518985,0.2942008078,-0.1116905063,0.1452423632,-0.0886163935,0.2498994023,-0.0540748574,0.0476816744,0.0109027456,-0.2792163789,0.0340285264,0.2409600765,-0.0198775716,-0.0205866508,-0.010750277,0.2926754951,-0.2988884151,0.3311219811,-0.215702638,0.0154130459,0.0760242268,0.1156876236,-0.2709846795,-0.4835907817,0.1956357658,-0.0492185876,-0.0084486026,0.0882154107,-0.0001012587,0.1353725642,0.4145276546,0.1116081104,0.1066273674,-0.0684448853,0.4700236619,-0.3405478299,0.311549902,-0.098624438,-0.2819465697,-0.4081077874,0.2505971789,0.6711853147,0.3359561563,0.1211105585,0.3480755389,0.5001601577,0.3954427242,0.1658053696,0.3523463607,-0.1978536248,-0.0361132957,0.2677507997,0.1481625736,-0.4075258076,-0.2368760854,0.07096228,-0.0093408721,-0.1127516031,0.0301447492,0.047961086,0.0806766078,-0.0997200385,-0.2059053928,0.0161200631,-0.0574331619,-0.3205562234,-0.0708786324,-0.0156158544,-0.0208418034,-0.5089048147,0.2064497471,0.2006173432,-0.0542783327,0.2096865773,0.1003701985,0.0446934253,-0.2387151271,0.2258310318,0.1216246635,0.5240388513,0.0379293375,-0.0238940232,0.105646275,-0.1272696853,-0.0082167964,0.1798074394,-0.0201324895,0.0425075702,-0.0552778393,-0.037365742,0.2063015401,0.2702012956,0.105284445,-0.1081267074,-0.0625578761,0.2314663678,0.1569803208,0.2653484046,-0.0428863615,0.0489424802,0.1638955921,-0.2536016703,0.4046475887,0.1588349789,0.0745864436,0.0340363681,0.0635632947,-0.2887741625,0.0585681237,-0.017119877,0.2171913385,-0.0965818763,-0.0833475515,-0.1682157964,-0.2672124803,-0.0539838448,-0.0712618232,0.3109348714,-0.1543706954,-0.1074745581,0.1612372249,0.3042466342,-0.3067513108,-0.1645632833,-0.2093216181,0.1111552641,-0.3626244664,-0.0949544087,0.1783848405,-0.0661119372,0.4190838933,0.265878737,-0.0402030312,0.1578613371,-0.0750475079,0.1995469332,0.0013793501,0.1259538978,0.0846485868,-0.0778634548,-0.0280964915,-0.1836329401,-0.0369263068,0.106199272,0.0347326361,-0.1273873895,-0.3307666481,0.318235606,-0.1651190072,-0.1347769201,0.039372053,0.0851858631,0.1165906638,-0.1288776994,-0.0398765877,-0.0937708318,0.092606768,-0.1198615432,0.004330907,0.0212782491,-0.1535105258,-0.1694055349,0.2005732805,-0.3917555809,0.2791665494,0.2945415676,-0.2822350562,-0.0447247215,-0.2081811428,0.3181878626,0.0813365281,-0.1044838056,-0.1863571405,-0.0764140412,-0.2252242863,-0.0851439238,0.0592235737,-0.1372092813,-0.1898913085,0.1120365635,0.0164406747,0.3423857987,0.182410568,0.2250766754,-0.4763455391,-0.2586126626,0.1468507648,0.0664588287,0.0230237897,-0.1705940813,0.012482984,-0.177541256,0.3005571067,-0.158482179,-0.0189026594,0.0076063834,0.0855460763,0.0611027777,-0.1115498543,0.0254185796,-0.1709395349,-0.0026160947,-0.3724623919,0.4320304096,-0.1290363967,-0.2306715548,-0.1294745803,-0.0606168397,-0.1809107512,-0.1842549592,0.336932838,-0.0137804532,0.1039847657,0.2353380322,0.0041840654,-0.2178957462,0.0144015551,0.1284302026,0.0288766827,0.2037082464,-0.4045517147,-0.4417032599,0.1709386855,-0.0461350568,0.2153577358,-0.024928879,0.128935203,0.0630316809,-0.243855685,0.3281761408,-0.3387069404,-0.1174917743,0.1344677359,-0.3917201161,-0.0060217432,0.0530459695,-0.073123917,0.0491109341,0.2116378993,0.4641542733,0.2355956435,-0.1295155734,0.0799313635,-0.1752558649,0.0940429196,-0.2181628346,0.1063482836,0.0639543533,0.051428169,0.1446449906,0.0831859633,-0.1385643035,-0.3161577582,0.0152327679,0.4189319909,0.0980302393,0.1676376462,-0.0125149395,-0.0490160845,-0.1178159267,0.2009183615,-0.5675620437,-0.0400182158,0.3262566626,0.0204320364,0.2574639618,-0.1666729301,-0.3516590595,0.0428374745,-0.0033390347,0.0281960126,0.1978377998,0.3531061709,-0.2695583403,-0.7390102148,0.1887566298,-0.0535354614,0.0520773791,-0.2663395405,-0.0003774771,-0.2753043771,-0.3825582266,-0.0464937761,-0.2470650971,-0.3113899231,-0.0502916612,0.2770524621,0.2221802622,-0.1764128208,0.3310760856,0.3593808711,-0.0941310227,-0.0549927056,-0.1581531167,-0.1620900631,-0.1792615652,-0.2970066369,0.2292392403,0.0523059219,0.0860825181,0.1486210078,-0.1228255928,-0.078695789,0.0013653582,-0.5312141776,0.0580219328,-0.1799623221,0.0904038101,0.1199069172,0.0441794433,-0.0703162923,-0.2853721976,0.3016642332,-0.3679868579,-0.385569334,-0.0633039847,-0.0918101966,0.0772606581,-0.3338805437,-0.1429333836,-0.2459408492,-0.3253278434,0.1830741763,0.1489655823,0.0596888475,0.2086916417,0.1391779631,0.1245823056,-0.2607911527,0.0355235822,0.0099630486,-0.1550075859,0.180803895,-0.2440594137,-0.4273223281,0.0036520972,-0.0250468887,0.1437660754,-0.2992115617,-0.3097371161,-0.3619363308,0.0171230286,0.6045031548,0.3813969493,-0.0246997364,0.3332929611,0.052149564,-0.2780450284,-0.051090762,-0.3169169426,-0.0000747388,-0.1725507528,0.0154258925,-0.0898745432,0.0951257944,-0.1031514779,0.7162159681,-0.3443971872,-0.1640798897,0.5709380507,-0.1614411026,0.2671315968,0.0111629842,-0.1037029326,0.035782028,0.0782717094,0.2829317749,0.1763738543,-0.1691609919,-0.3252235055,-0.1719830036,-0.3856511116,-0.1695446521,-0.2349778116,-0.0284416843,0.5755477548,-0.0170576777,0.1480150521,-0.0271566492,0.0595643781,-0.1974754333,0.2500547171,0.0569541641,-0.1684383154,-0.0497869439,-0.2076005489,-0.1054569855,-0.2911826074,0.1644016951,0.1411540657,-0.134992674,-0.0076699611,-0.2763393819,0.1494256854,-0.1652489901,0.389547199,0.2480855435,0.1742343754,0.2051043361,0.0717517808,-0.1589295864,-0.1130563766,-0.1424104571,-0.0518174805,0.318526119,0.0641785562,-0.4388537109,-0.1291134506,0.2902989686,0.1671341807,0.0330564007,-0.0816078037,-0.5188971162,-0.2308009863,-0.1938465685,0.0818187147,0.0479568243,0.2659160793,-0.1101712659,0.0723032206,0.0573873594,-0.0972492844,-0.2095820457,0.3896647394,0.032702297,-0.1427304596,-0.0913315639,-0.07490848,-0.0184262823,-0.0161842704,0.3112937808,0.1312457472,-0.221511364,-0.1273939013,0.1709264517,0.1626739949,0.1168815792,0.0409196764,0.2281319648,-0.0984501019,0.1062303036,0.1769817322,0.1393135935,0.3002857268,-0.0544730537,0.1809251308,0.2272676826,0.2877385914,-0.0484498665,0.094732143,-0.0752651691,0.2149498016,-0.2053005844,0.1492197365,0.007487312,0.7680068016,0.0761309564,0.2104943842,0.6422811151,-0.2271831781,0.3503385782,0.0819659084,0.1829580516,-0.4050073922,0.0429866128,0.2244747728,-0.0363075584,0.0671994835,-0.1467119008,-0.1254956573,0.0650454387,0.1090779528,-0.0742074773,0.0069554788,0.2265466601,0.1556535065,0.0204281304,-0.398524642,0.1946176738,0.018936906,0.3295451701,-0.1877236217,-0.2866224051,0.2308532,-0.3202812374,0.0625260398,0.3509094119,0.0014304215,0.0088611571,0.2283275574,-0.2104381919,-0.1209769696,0.0812266618,-0.0035937028,-0.0729902089,-0.1060247496,0.2286629677,0.0417102538,-0.2833078802,-0.2292841077,0.2728439271,0.0326541737,-0.2719822526,-0.3732590079,0.238841638,0.0440947972,-0.0237306748,-0.1905411482,0.0188753027,0.0217463337,-0.0245165769,0.1378337145,-0.027288653,-0.0140619474,-0.2212867886,0.2729762793,0.0151979774,-0.0782758966,-0.0624720827,0.2111312598,-0.1381330788,-0.2341832817,0.3990569115,-0.1236465499,0.0743977129,0.3554134369,0.4130141735,-0.2508852184,-0.2214540839,-0.2503725588,0.0494532809,-0.5689085722,-0.0581767671,0.3290356696,-0.0277754012,0.0499628559,0.3349453509,-0.0103448946,-0.0020608366,0.086575523,-0.0616175607,-0.1658620685,0.0801384002,0.0050323592,-0.0811093226,0.4251028001,0.4409349561,0.2606270611,-0.1629951447,-0.4242507219,0.0885329545,-0.1109518036,0.1491156071,-0.2684071064,-0.3252624273,0.1424764395,0.1223281249,0.2818827331,0.0855144262,-0.0990869775,-0.3659952283,-0.1273598075,0.0164749101,0.0426101573,-0.1351796538,0.0855447203,0.1667261869,0.092936933,0.1755120605,0.2093611062,-0.0683120117,-0.1360498965,0.2656249106,0.1472411603,-0.0094863875,0.0548738055,-0.2414258569,-0.0865174234,0.2173107564,0.0913912803,-0.0121694598,-0.1168199256,0.0218576491,0.2830783129,-0.0460751168,-0.0533515699,0.1344892532,0.1874646395,-0.2675083578,-0.2990043759,-0.3259352148,0.2857321501,0.1367183179,-0.1718844175,-0.0815339237,0.4485119879,0.355173558,-0.493281424,-0.0007997304,0.0230366085,0.0422270335,0.1361253262,0.0369145051,-0.058583267,-0.0270045772,0.1628056765,0.3364424706,0.1721842438,-0.0182645041,-0.2412798703,0.0883366987,-0.2695306838,-0.0853940398,0.4863008857,0.2177912891,0.1966185272,-0.0695477203,-0.1658845395,0.0675874576,-0.2817701995,0.1904411614,-0.139228195,-0.0781760961,-0.1481595486,0.3664598465,0.2366108447,-0.2096528709,0.1218015552,0.4050146341,0.3279471695,-0.1017601341,-0.0352139808,0.5940719247,-0.1501726806,-0.0372414254,-0.1205552891,0.0247092731,0.0186923388,0.1990353465,-0.0100840926,0.0371301472,0.1817221045,0.0449161977,-0.1317514628,-0.3449285626,0.0278904364,0.13352184,0.1944080144,-0.0109306751,0.3146857321,0.1818221807,-0.1768801361,-0.2067466229,0.3157662451,0.4289801419,0.190857321,-0.0834466442,-0.0905012861,-0.1761847734,0.0130693791,-0.2651736736,0.1998058856,0.3569592237,0.469602108,0.1155807078,0.2181480825,-0.3972553015,0.1083578765,0.1431080252,-0.051491946,0.0501389951,0.045056574,-0.0111389738,-0.3065873086,-0.2804445624,-0.2022271454,-0.4173827171,0.2953651249,0.0527491309,0.2489847988,0.0291254241,-0.5049664974,0.1465916932,0.1816422343,0.6292338967,0.2281741947,0.3879196644,-0.1220241413,-0.1330974102,-0.4203570783,0.329138279,-0.2190537602,0.0000092769,-0.0777836964,0.0103322221,-0.0749561489,0.2517430484,0.0125342039,0.0436716713,-0.1176454723,-0.1502147764,-0.3011311293,0.4134970605,0.0924495831,-0.1343783289,0.0727386996,-0.1670098752,0.2411808074,-0.4894816279,0.2721237838,-0.2460789233,0.0679926574,-0.115235433,0.3288315237,0.2586695552,0.0902900025,0.2239749581,-0.0821103379,0.0896252766,-0.1233866587,-0.3137269914,-0.2358466685,0.1291939467,0.0301580764,0.1460093856,0.1008004621,-0.1483061314,-0.2044864744,0.3544847965,0.0142885726,0.0659903511,0.2446502596,0.2006126344,0.1369782239,-0.1885241717,-0.1799708158,0.0794759393,0.010028054,-0.223279804,-0.483081907,-0.4808937609,0.3390308619,-0.4343389273,-0.0705741346,0.034397997,0.1019957438,0.1384659559,-0.3384727538,-0.4281983972,0.3069325686,0.1789505929,0.1956395507,-0.2019683421,0.2111749649,-0.0276824459,-0.0544211119,-0.0343261808,0.0695876926,0.2646894753,-0.2263114452,-0.0114331143,-0.2381405979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/433","title":"How to reuse functionality of a (generic) dataset?","comments":"Hi @ArneBinder, we have a few \"generic\" datasets which are intended to load data files with a predefined format:\r\n- csv: https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/csv\r\n- json: https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/json\r\n- text: https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text\r\n\r\nYou can find more details about this way to load datasets here in the documentation: https:\/\/huggingface.co\/nlp\/loading_datasets.html#from-local-files\r\n\r\nMaybe your brat loading script could be shared in a similar fashion?","body":"I have written a generic dataset for corpora created with the Brat annotation tool ([specification](https:\/\/brat.nlplab.org\/standoff.html), [dataset code](https:\/\/github.com\/ArneBinder\/nlp\/blob\/brat\/datasets\/brat\/brat.py)). Now I wonder how to use that to create specific dataset instances. What's the recommended way to reuse formats and loading functionality for datasets with a common format?\r\n\r\nIn my case, it took a bit of time to create the Brat dataset and I think others would appreciate to not have to think about that again. Also, I assume there are other formats (e.g. conll) that are widely used, so having this would really ease dataset onboarding and adoption of the library.","comment_length":56,"text":"How to reuse functionality of a (generic) dataset?\nI have written a generic dataset for corpora created with the Brat annotation tool ([specification](https:\/\/brat.nlplab.org\/standoff.html), [dataset code](https:\/\/github.com\/ArneBinder\/nlp\/blob\/brat\/datasets\/brat\/brat.py)). Now I wonder how to use that to create specific dataset instances. What's the recommended way to reuse formats and loading functionality for datasets with a common format?\r\n\r\nIn my case, it took a bit of time to create the Brat dataset and I think others would appreciate to not have to think about that again. Also, I assume there are other formats (e.g. conll) that are widely used, so having this would really ease dataset onboarding and adoption of the library.\nHi @ArneBinder, we have a few \"generic\" datasets which are intended to load data files with a predefined format:\r\n- csv: https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/csv\r\n- json: https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/json\r\n- text: https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text\r\n\r\nYou can find more details about this way to load datasets here in the documentation: https:\/\/huggingface.co\/nlp\/loading_datasets.html#from-local-files\r\n\r\nMaybe your brat loading script could be shared in a similar fashion?","embeddings":[-0.1659951061,-0.1904758215,0.1139429435,0.3808882535,0.2183723897,-0.0047787465,0.1885102093,0.112266995,-0.1110894829,-0.2132682204,-0.195565477,0.2885515392,-0.0397951044,0.5394513607,0.1926315129,-0.1346117556,-0.211735636,0.2186616659,-0.1162287667,-0.0262852851,0.0394088477,-0.1943162233,0.0686826631,-0.1159505472,-0.3702635169,0.1318924874,-0.2384282053,-0.0179268848,0.2592282891,-0.2346274108,0.5442937613,0.3445178866,0.45950225,0.2523490191,-0.0001251962,0.0330158919,-0.0391854458,-0.1448282897,-0.2924483716,-0.1255922019,-0.2134353817,-0.223388806,0.0898034647,-0.2437822074,-0.3222340643,-0.3877206147,0.1589093655,-0.5658196211,0.4574590027,-0.0978449285,0.0306623075,-0.0806691349,-0.2010449767,-0.1027776524,-0.1274460554,0.3995322883,-0.0048828083,0.4021582305,0.6236608028,0.3071166873,0.1142585874,0.0186164398,0.1892261654,0.092304334,0.2961039841,-0.1304038018,-0.0321157947,0.0373831466,0.0351924412,0.3731410205,0.9288996458,-0.3871008158,-0.057318639,-0.1039505824,-0.0022320303,-0.2916268706,-0.2740898132,0.1782106757,0.1706531495,0.1169880182,-0.2853214741,-0.3464595079,-0.2974860668,0.1213214993,0.2955724597,0.0247342754,-0.0860804692,0.2227665782,-0.202981934,-0.087689206,0.0028414812,-0.2409488559,-0.1546615213,0.0481973104,0.1250663996,-0.1266933084,-0.4318877161,0.1242781803,0.3226211965,0.4385862648,0.3070451021,0.2488357425,-0.3475722671,0.0543904081,0.4337318242,0.0727873668,0.3920842409,0.275516659,0.0898746774,0.2174168527,-0.4256957173,-0.1605033726,-0.4732805192,0.4096366763,0.2828017175,-0.0478396602,0.0258606561,-0.2413187325,-0.0057717771,-0.0553799048,-0.3146702945,-0.1710859239,0.220073089,0.0787808299,-0.1268196851,0.1943231225,-0.076090306,0.2540785372,-0.0605482236,-0.3372296393,0.0072380351,0.0385019518,-0.3324481845,0.1706179529,0.3277175426,-0.4035727084,0.2055005133,0.3074685931,-0.149845466,-0.1152982339,0.3684354126,0.1104314327,-0.2815550566,-0.5091332197,-0.1593072563,0.0936785638,0.2187217474,0.1501332074,-0.3380337358,0.0335500129,-0.1240274757,-0.3322888911,-0.1301243901,0.0158546586,-0.2962245345,0.0313484743,-0.1064650267,0.4295431674,-0.0012109492,-0.3559826612,-0.2109797001,-0.1043991074,-0.2914378643,-0.0680791661,0.1657022238,0.2590589225,-0.5212173462,-0.1286891103,-0.1151430011,0.000508684,0.414929688,0.2499226183,-0.4169629514,0.21721223,-0.1238055825,-0.1836066842,0.455332756,-0.0797947124,-0.0108388131,-0.1229192987,0.2009911388,0.1629832983,0.0826169178,0.0667771026,0.1893800944,-0.2476023138,-0.3442697525,0.5017463565,0.1047518551,-0.2457328141,-0.0767719299,-0.212257579,0.1183497384,0.0398364216,-0.2308100015,0.1301293969,0.0850554705,0.2706214488,0.0201237518,-0.2445560247,0.1147236824,-0.1340281516,0.0210207812,0.3941365778,0.1071013063,-0.2427259684,-0.501625061,0.1725225747,-0.3441224694,-0.2779965997,0.5025932193,-0.5051969886,-0.3756371439,-0.24389714,-0.1494172662,-0.0231227893,0.0117689408,0.5268944502,0.1075812802,-0.0793845057,-0.1046641544,0.1127329916,-0.1162446439,-0.0206026565,-0.0946501344,0.2827685475,0.1337838322,0.0327418298,0.0808496028,0.5369929671,0.0391520001,-0.0522328988,0.1695102006,0.2841656208,0.1377085,0.1723063439,0.0890184417,0.1135612205,0.1253343076,0.1243243963,0.4157356918,-0.0623390041,0.3338619769,-0.1436809897,-0.4933606386,0.2344053239,0.0853987038,0.0032412079,0.0942281261,-0.0990343839,0.1269897819,-0.1194883958,-0.163816914,-0.1160496622,-0.1618515253,0.166863054,-0.0490618013,0.3370729685,-0.3186077476,-0.0587684885,0.2334949374,-0.1676668525,0.2127047628,-0.1248301268,-0.1789677143,-0.1150687933,0.3755524158,-0.169630155,0.0814922675,0.051055152,0.2505188286,-0.009340059,-0.0731735155,0.0596740022,0.2454495728,-0.1996546388,0.1939686537,-0.2038097978,0.1229524687,0.072569862,0.0402918831,-0.0532920286,0.2882683873,0.0353524126,-0.1095572338,0.1736063063,-0.7263829708,-0.2955489755,-0.3641794622,0.0308190566,0.0220962502,0.0343479775,-0.2832489014,0.052305799,0.1293509156,0.1510513425,0.1038258076,0.5219677687,-0.3055048883,-0.0818183273,0.0544117503,-0.3273842633,-0.3867360055,0.0133745177,0.4593727887,0.3856511116,0.4869647026,-0.0203824341,0.0797765553,-0.4213525951,-0.2247262448,0.2008585185,0.2226727307,-0.0748157427,0.1745197177,0.0166827347,-0.3718517721,-0.2864752412,0.1982575506,0.1177977994,0.2160826027,-0.1023503989,-0.4804204106,0.16403763,-0.1767154932,-0.0417492054,-0.0790914074,-0.2007599622,0.2851938009,-0.0058270465,-0.0302391835,0.370518744,-0.2798815668,0.2028849274,-0.110438019,0.2872950435,-0.2397326082,-0.2474448681,0.2628939748,-0.1447509378,-0.149028644,-0.1419074386,-0.3319850862,0.1567607224,0.5356497169,-0.2468817383,-0.3932585418,-0.1005750448,-0.2145656049,0.2922896743,0.2270694226,0.0891076401,0.0088778539,0.1633804739,-0.0381374285,-0.1159494296,0.2437331229,-0.0833875611,0.1644108295,0.4508982301,0.0424259454,0.2239995003,0.3212416768,0.2181174904,-0.4716588557,0.4767279625,-0.1040036231,0.4264326394,0.0322025195,0.0494236611,0.24937509,-0.0998954177,0.1182431653,0.2240378261,0.0198332388,0.1488362253,-0.1424521208,0.0832243562,-0.0674774647,-0.4802441299,0.1267703623,-0.2630298436,0.506565094,0.2004096359,-0.16664204,-0.2821270525,0.0118128499,0.2048148364,0.1115639284,0.1518702954,0.111413829,-0.3922570348,0.1145912409,0.1744352281,0.5108853579,0.090602532,0.0278528668,0.2118860483,0.0001372858,-0.0055027385,0.0644413009,-0.0107165156,-0.2585104704,0.0617623292,0.0170505792,-0.2340942919,-0.3129954934,0.0051870849,-0.1482111514,0.2156682611,0.0613821521,0.0937313586,-0.5858938694,-0.1134346202,0.305680722,0.4081916213,-0.1027640402,0.1559307873,0.2447598726,-0.0840716437,-0.3423586488,-0.1207061261,-0.2220969051,0.1091594994,0.077484712,0.1068611294,-0.4060516655,0.1789985001,0.3230231106,0.143601492,0.5515426397,0.1673408002,0.102503702,0.1422673464,0.0629676878,0.1254569739,0.5182183385,-0.2743422985,-0.4875465035,-0.1123322174,0.0527324863,0.355232209,0.1335095316,0.268671453,-0.2644345164,-0.2075336277,-0.3050481081,-0.2465967387,0.2815004885,-0.0214466788,-0.0089538442,-0.4367788434,-1.003655076,0.1773266494,0.2341749221,-0.0221639741,0.1960072815,0.0442957506,-0.0561630577,0.0839707255,0.1482159942,0.8185894489,-0.234224245,0.2623119354,0.2913030982,0.2999873459,0.672930181,-0.1592321396,-0.2336211503,-0.3155064583,0.0776641518,0.0124218604,-0.0548908673,0.4027072191,0.2620368302,-0.1770489812,0.3238634467,0.2328755409,0.414917171,0.0472156145,0.2062345147,-0.4215626717,-0.2640066743,0.0007392883,-0.0266651791,-0.3107623756,-0.0222587846,-0.2223700136,0.1319068223,0.176897198,0.1092705429,-0.2893408835,-0.0679774657,-0.2111096382,0.0153954905,-0.1433986425,-0.4598444998,0.4378589392,0.4143341184,0.2606966197,0.2978545725,-0.1250857115,0.0351935513,-0.1367508918,-0.0574591421,-0.2723386884,0.0598763376,0.2151693404,0.1179764047,-0.0462159067,0.1231957152,0.3675069511,0.150108248,-0.193992883,-0.2078894228,0.3192488551,-0.3471474648,-0.2080489695,0.1136159748,0.0116935931,-0.0000477247,-0.0354513302,0.2170712799,-0.0585962757,0.0365926176,-0.0799802393,0.0506308787,-0.0118119149,-0.2164823711,0.2862481773,-0.244828552,0.3801532388,0.0534391291,0.096025981,-0.12397632,0.1542889476,0.2004524916,-0.3535531163,-0.0472874083,0.0481462255,-0.2010813653,0.1257612407,0.2455195785,-0.1104493812,0.0416256599,0.1067188904,-0.2709896564,-0.062190868,0.2552555799,0.0847928002,0.3913347721,0.0335180908,-0.0114197601,-0.2339341938,-0.076387465,-0.1716609597,0.2520009577,0.2532122135,0.2020986229,-0.0755690485,0.0658908337,-0.0873563215,-0.043743398,-0.0359815471,0.1191643476,-0.1045867577,-0.0319880955,-0.5068326592,0.2202413678,-0.0208259095,-0.1808736622,-0.0664783046,-0.320027709,0.0748323053,-0.4403782189,0.1769111603,0.0493836068,-0.0720403865,0.2194672078,0.5647692084,0.0570192933,-0.2097332925,0.0016651475,0.1266692281,0.410019815,0.3731326461,0.0852252543,-0.1460329145,-0.1167678759,0.0035766077,0.2336939424,0.8808066249,-0.0425570756,0.2894618213,0.137909025,0.0667545423,0.1670592427,0.3605265021,0.4735856652,-0.090605244,0.2389072329,0.3097982407,0.0023917221,-0.0827837959,0.0948000029,0.1694345623,-0.0890068337,-0.1327607334,0.3536591232,0.1840624064,0.2066236138,-0.1526349634,-0.3317937553,0.171079725,0.0503299646,0.4426064491,0.2776347697,0.0129948696,-0.3771143258,-0.0332405157,0.3894673288,0.2786750495,-0.4481979609,0.2374075204,0.2064731568,0.0597752891,0.254324615,0.3961917758,-0.5807134509,-0.0108073922,0.3840749562,0.2369926572,0.3525371552,-0.0982664973,0.2982102334,0.1065689921,0.0288063362,-0.0740947649,0.1353553385,-0.0243054889,-0.1885050386,-0.2255789936,-0.3066186309,-0.2714518905,0.0051427763,-0.2844661474,-0.0392599143,-0.2430738062,-0.0055701775,0.0913800001,-0.4289422929,0.088523224,0.1282114536,0.1109811664,-0.0272958521,0.1762390435,0.2643396556,-0.2772353888,0.3040177822,0.3469225168,0.1574381888,0.0013450078,-0.0784207731,0.2037943453,-0.2044151574,0.1393670142,-0.1719984859,0.084212102,0.1782952249,-0.3458361626,0.2469351739,-0.017918475,0.059367314,-0.0808811113,-0.0942523777,-0.2252890021,-0.5750337243,0.4838283956,0.2467943579,-0.1766696125,0.1953533441,0.1366724372,-0.2376323938,-0.4150587916,0.3314507604,0.3181432784,-0.0127968993,-0.1789317429,-0.0029143055,0.166766271,0.2657192051,0.1069095507,-0.1033225134,-0.0076270727,-0.3208794892,-0.7505283356,0.2008836269,-0.1550965309,-0.5894007087,0.2148181498,-0.1606934369,0.2775504887,0.1601593345,-0.2683603168,-0.0050947028,0.1115243807,-0.1754319966,-0.1335077435,-0.5121636987,-0.0151768383,-0.036365822,-0.1568967402,-0.3666356206,-0.0744640753,0.0146499434,-0.1392393261,-0.0253695063,-0.1672609597,0.1073861048,-0.411196053,0.0796096772,0.081120871,0.3132669926,0.212406382,-0.1055359989,-0.2097654045,0.0714218169,0.0201069769,0.3408660293,-0.1046823934,0.0942917913,-0.5658497214,-0.0598277971,-0.2669111788,-0.1238186285,0.0861118883,0.1740021259,-0.1259275526,-0.0274488609,-0.3547935784,-0.104649201,0.3244485557,0.6672959924,-0.2203667909,0.0389131941,-0.0882799104,0.2342381924,0.3661120236,-0.2147353739,-0.3655220866,-0.19023785,0.0704674646,-0.2707994282,0.047245197,-0.3562135696,0.1222978085,0.2829151154,-0.0311342422,-0.1060800925,-0.0718564838,0.1953959316,-0.1844020039,-0.1983388513,0.4595988393,0.1070045233,-0.1265895367,-0.0130088041,-0.1345304698]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/433","title":"How to reuse functionality of a (generic) dataset?","comments":"> Maybe your brat loading script could be shared in a similar fashion?\r\n\r\n@thomwolf that was also my first idea and I think I will tackle that in the next days. I separated the code and created a real abstract class `AbstractBrat` to allow to inherit from that (I've just seen that the dataset_loader loads the first non abstract class), now `Brat` is very similar in its functionality to https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text but inherits from `AbstractBrat`.\r\n\r\nHowever, it is still not clear to me how to add a specific dataset (as explained in https:\/\/huggingface.co\/nlp\/add_dataset.html) to your repo that uses this format\/abstract class, i.e. re-using the `features` entry of the `DatasetInfo` object and `_generate_examples()`. Again, by doing so, the only remaining entries\/functions to define would be `_DESCRIPTION`, `_CITATION`, `homepage` and `_URL` (which is all copy-paste stuff) and `_split_generators()`.\r\n \r\nIn a lack of better ideas, I tried sth like below, but of course it does not work outside `nlp` (`AbstractBrat` is currently defined in [datasets\/brat.py](https:\/\/github.com\/ArneBinder\/nlp\/blob\/5e81fb8710546ee7be3353a7f02a3045e9a8351e\/datasets\/brat\/brat.py)):\r\n```python\r\nfrom __future__ import absolute_import, division, print_function\r\n\r\nimport os\r\n\r\nimport nlp\r\n\r\nfrom datasets.brat.brat import AbstractBrat\r\n\r\n_CITATION = \"\"\"\r\n@inproceedings{lauscher2018b,\r\n title = {An argument-annotated corpus of scientific publications},\r\n booktitle = {Proceedings of the 5th Workshop on Mining Argumentation},\r\n publisher = {Association for Computational Linguistics},\r\n author = {Lauscher, Anne and Glava\\v{s}, Goran and Ponzetto, Simone Paolo},\r\n address = {Brussels, Belgium},\r\n year = {2018},\r\n pages = {40\u201346}\r\n}\r\n\"\"\"\r\n\r\n_DESCRIPTION = \"\"\"\\\r\nThis dataset is an extension of the Dr. Inventor corpus (Fisas et al., 2015, 2016) with an annotation layer containing \r\nfine-grained argumentative components and relations. It is the first argument-annotated corpus of scientific \r\npublications (in English), which allows for joint analyses of argumentation and other rhetorical dimensions of \r\nscientific writing.\r\n\"\"\"\r\n\r\n_URL = \"http:\/\/data.dws.informatik.uni-mannheim.de\/sci-arg\/compiled_corpus.zip\"\r\n\r\n\r\nclass Sciarg(AbstractBrat):\r\n\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def _info(self):\r\n\r\n brat_features = super()._info().features\r\n return nlp.DatasetInfo(\r\n # This is the description that will appear on the datasets page.\r\n description=_DESCRIPTION,\r\n # nlp.features.FeatureConnectors\r\n features=brat_features,\r\n # If there's a common (input, target) tuple from the features,\r\n # specify them here. They'll be used if as_supervised=True in\r\n # builder.as_dataset.\r\n #supervised_keys=None,\r\n # Homepage of the dataset for documentation\r\n homepage=\"https:\/\/github.com\/anlausch\/ArguminSci\",\r\n citation=_CITATION,\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n # TODO: Downloads the data and defines the splits\r\n # dl_manager is a nlp.download.DownloadManager that can be used to\r\n # download and extract URLs\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n data_dir = os.path.join(dl_dir, \"compiled_corpus\")\r\n print(f'data_dir: {data_dir}')\r\n return [\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TRAIN,\r\n # These kwargs will be passed to _generate_examples\r\n gen_kwargs={\r\n \"directory\": data_dir,\r\n },\r\n ),\r\n ]\r\n``` \r\n\r\nNevertheless, many thanks for tackling the dataset accessibility problem with this great library!","body":"I have written a generic dataset for corpora created with the Brat annotation tool ([specification](https:\/\/brat.nlplab.org\/standoff.html), [dataset code](https:\/\/github.com\/ArneBinder\/nlp\/blob\/brat\/datasets\/brat\/brat.py)). Now I wonder how to use that to create specific dataset instances. What's the recommended way to reuse formats and loading functionality for datasets with a common format?\r\n\r\nIn my case, it took a bit of time to create the Brat dataset and I think others would appreciate to not have to think about that again. Also, I assume there are other formats (e.g. conll) that are widely used, so having this would really ease dataset onboarding and adoption of the library.","comment_length":416,"text":"How to reuse functionality of a (generic) dataset?\nI have written a generic dataset for corpora created with the Brat annotation tool ([specification](https:\/\/brat.nlplab.org\/standoff.html), [dataset code](https:\/\/github.com\/ArneBinder\/nlp\/blob\/brat\/datasets\/brat\/brat.py)). Now I wonder how to use that to create specific dataset instances. What's the recommended way to reuse formats and loading functionality for datasets with a common format?\r\n\r\nIn my case, it took a bit of time to create the Brat dataset and I think others would appreciate to not have to think about that again. Also, I assume there are other formats (e.g. conll) that are widely used, so having this would really ease dataset onboarding and adoption of the library.\n> Maybe your brat loading script could be shared in a similar fashion?\r\n\r\n@thomwolf that was also my first idea and I think I will tackle that in the next days. I separated the code and created a real abstract class `AbstractBrat` to allow to inherit from that (I've just seen that the dataset_loader loads the first non abstract class), now `Brat` is very similar in its functionality to https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets\/text but inherits from `AbstractBrat`.\r\n\r\nHowever, it is still not clear to me how to add a specific dataset (as explained in https:\/\/huggingface.co\/nlp\/add_dataset.html) to your repo that uses this format\/abstract class, i.e. re-using the `features` entry of the `DatasetInfo` object and `_generate_examples()`. Again, by doing so, the only remaining entries\/functions to define would be `_DESCRIPTION`, `_CITATION`, `homepage` and `_URL` (which is all copy-paste stuff) and `_split_generators()`.\r\n \r\nIn a lack of better ideas, I tried sth like below, but of course it does not work outside `nlp` (`AbstractBrat` is currently defined in [datasets\/brat.py](https:\/\/github.com\/ArneBinder\/nlp\/blob\/5e81fb8710546ee7be3353a7f02a3045e9a8351e\/datasets\/brat\/brat.py)):\r\n```python\r\nfrom __future__ import absolute_import, division, print_function\r\n\r\nimport os\r\n\r\nimport nlp\r\n\r\nfrom datasets.brat.brat import AbstractBrat\r\n\r\n_CITATION = \"\"\"\r\n@inproceedings{lauscher2018b,\r\n title = {An argument-annotated corpus of scientific publications},\r\n booktitle = {Proceedings of the 5th Workshop on Mining Argumentation},\r\n publisher = {Association for Computational Linguistics},\r\n author = {Lauscher, Anne and Glava\\v{s}, Goran and Ponzetto, Simone Paolo},\r\n address = {Brussels, Belgium},\r\n year = {2018},\r\n pages = {40\u201346}\r\n}\r\n\"\"\"\r\n\r\n_DESCRIPTION = \"\"\"\\\r\nThis dataset is an extension of the Dr. Inventor corpus (Fisas et al., 2015, 2016) with an annotation layer containing \r\nfine-grained argumentative components and relations. It is the first argument-annotated corpus of scientific \r\npublications (in English), which allows for joint analyses of argumentation and other rhetorical dimensions of \r\nscientific writing.\r\n\"\"\"\r\n\r\n_URL = \"http:\/\/data.dws.informatik.uni-mannheim.de\/sci-arg\/compiled_corpus.zip\"\r\n\r\n\r\nclass Sciarg(AbstractBrat):\r\n\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def _info(self):\r\n\r\n brat_features = super()._info().features\r\n return nlp.DatasetInfo(\r\n # This is the description that will appear on the datasets page.\r\n description=_DESCRIPTION,\r\n # nlp.features.FeatureConnectors\r\n features=brat_features,\r\n # If there's a common (input, target) tuple from the features,\r\n # specify them here. They'll be used if as_supervised=True in\r\n # builder.as_dataset.\r\n #supervised_keys=None,\r\n # Homepage of the dataset for documentation\r\n homepage=\"https:\/\/github.com\/anlausch\/ArguminSci\",\r\n citation=_CITATION,\r\n )\r\n\r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n # TODO: Downloads the data and defines the splits\r\n # dl_manager is a nlp.download.DownloadManager that can be used to\r\n # download and extract URLs\r\n dl_dir = dl_manager.download_and_extract(_URL)\r\n data_dir = os.path.join(dl_dir, \"compiled_corpus\")\r\n print(f'data_dir: {data_dir}')\r\n return [\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TRAIN,\r\n # These kwargs will be passed to _generate_examples\r\n gen_kwargs={\r\n \"directory\": data_dir,\r\n },\r\n ),\r\n ]\r\n``` \r\n\r\nNevertheless, many thanks for tackling the dataset accessibility problem with this great library!","embeddings":[-0.0527758598,-0.0183876511,0.1475319415,0.4042401016,0.2143378109,-0.0373157039,0.1315058917,0.1071476415,-0.2023022622,-0.2372688204,-0.1918668449,0.4055892527,-0.0635214299,0.5268573761,0.4249489903,-0.2587217391,-0.1975932866,0.1804039478,0.0004199797,-0.0935178176,-0.0021080349,-0.1374595165,0.1369740367,-0.0939151198,-0.2226285785,0.1669851691,-0.2874785364,-0.0323904529,0.1107129976,-0.2528283596,0.5313698053,0.3521875441,0.468341887,0.2782885134,-0.0001250765,0.0179283749,-0.0833464041,-0.1071614251,-0.2715194821,-0.166204229,-0.2018947899,-0.2514975965,0.0697826967,-0.2622407079,-0.333869338,-0.2891654074,0.0832434818,-0.420181632,0.4280590713,0.0142240403,0.012890866,-0.1331839859,-0.1333251446,-0.1284095943,0.028225448,0.4498718977,-0.1006873623,0.3456860483,0.5528991818,0.29368788,0.1292904764,0.028397087,0.1609407812,0.1065396816,0.3837427795,-0.0584125482,-0.0705620497,-0.0389486663,0.018995218,0.2638921142,0.8025806546,-0.4108220041,-0.0872110203,-0.1729798019,0.024767505,-0.4295556545,-0.2573292553,0.0599815175,0.1624982953,0.0851936564,-0.3099957705,-0.3607076705,-0.3834712803,0.2709828615,0.3352554142,0.0396562219,-0.0566550158,0.2215604335,-0.1515273005,-0.0298044477,-0.0301856287,-0.1971120536,-0.1586646438,0.1109791696,0.1712026298,-0.1918716282,-0.3047834337,0.0230149236,0.4192993343,0.4918847084,0.1451487094,0.20817779,-0.3522256017,-0.0213294961,0.4007333517,0.1259273291,0.2779958248,0.3652256131,0.0427371822,0.1957389116,-0.3644953966,-0.1703820825,-0.4490543008,0.5557587743,0.2351506203,-0.1451735795,0.0688069761,-0.2394081652,0.1407324374,-0.1055003926,-0.3880719841,-0.1388137192,0.2200806141,0.0473230407,0.0224381071,0.0921761915,0.0555789769,0.2978412211,-0.0114642493,-0.2726375759,-0.0049478202,0.0923360065,-0.4189287424,0.2664704025,0.3285477161,-0.310305357,0.1651708037,0.3441130519,-0.0810408816,-0.170816958,0.3620076478,0.1349384189,-0.3155711889,-0.4831865728,-0.233671546,0.2006834298,0.2484113574,0.2031096816,-0.2982755601,-0.0437861159,-0.0155711658,-0.395182848,-0.1822104603,0.0121255508,-0.371001482,-0.0800734833,-0.2152220756,0.4285044372,-0.0043485137,-0.33845824,-0.1976387054,-0.0232472923,-0.3284014165,-0.0783782676,0.1933834553,0.2733840644,-0.3810253441,-0.171832487,-0.2532711029,0.0059723281,0.435426414,0.1614467502,-0.4003018141,0.2875467539,-0.1048991606,-0.1292135268,0.4984529018,-0.0334244557,-0.0209876448,-0.0589336753,0.1858076751,0.1797676235,0.0416482836,0.0760742053,0.2536152005,-0.1743207127,-0.3745343983,0.5141628981,0.0862224251,-0.2672062814,-0.0373789333,-0.1952516288,0.1279145032,0.0897817835,-0.1702144742,0.1609873921,-0.03143242,0.2894899249,-0.017752042,-0.2148414999,0.1200026497,-0.1749357283,0.1473779976,0.3941198289,0.1085904539,-0.3249780834,-0.4225756228,0.1913678646,-0.2947507203,-0.0480473042,0.4472073019,-0.4088856578,-0.4264878631,-0.184352234,-0.1645276546,-0.1849027872,0.0076964754,0.5292384624,0.0993560106,-0.1695847213,-0.1643251032,0.1189443842,-0.1181549653,-0.0168597493,-0.1055908874,0.2538265586,0.0643545538,0.0435818844,0.0140847452,0.5441199541,0.1163565516,-0.075459525,0.1564706415,0.1206884012,0.037100032,0.0826740861,0.1455654502,0.1017806157,0.1178467348,0.1147307307,0.4078248143,0.0088951029,0.3467353284,-0.1688862741,-0.4644254446,0.134048894,0.2337526232,0.0326775424,0.061245203,-0.1139331609,0.1701093763,-0.1843965948,-0.0612344779,-0.1180025041,-0.0761754885,0.2510493696,-0.0876895115,0.3345283568,-0.2914963067,0.1417697072,0.141065836,-0.2346038222,0.1524762511,0.0045161257,-0.2224809974,-0.1769359112,0.3832339346,0.0758576915,0.1644231379,0.037848182,0.2431350946,0.0694824308,-0.1100002229,0.0444386192,0.2112405896,-0.1673858762,0.1849935353,-0.237921685,0.1198144183,0.1696004868,0.0800332278,-0.0412702784,0.3497784734,0.1414006054,-0.1712508053,0.2018589973,-0.6481920481,-0.3009235859,-0.3514772654,-0.0437039435,0.0139782606,0.0545269027,-0.1946784705,0.0467866361,0.1307185292,0.134296909,0.1131130457,0.4815168083,-0.3766599298,-0.1216218546,0.0801850632,-0.299516499,-0.3330663443,0.0063181017,0.4122456312,0.3769203722,0.554540813,0.0870391503,0.029055303,-0.4986234903,-0.2839494646,0.2014662921,0.0905627161,-0.03976468,0.1327048838,-0.0485130958,-0.3301447928,-0.3042160869,0.1590521336,0.0961806029,0.1809403598,-0.0600807853,-0.4992723167,0.1211428046,-0.1876276433,-0.1224418804,-0.1508996189,-0.19926247,0.283375591,0.0598632134,0.0701045096,0.3311741054,-0.3769756258,0.2394531071,-0.0450571552,0.2134193629,-0.3368493617,-0.1046841592,0.2901270688,-0.1467796415,-0.185217306,-0.0646872669,-0.3328315914,0.1002845466,0.5816040039,-0.1899225861,-0.4333093166,-0.0709262416,-0.193177253,0.335211575,0.2023753673,0.1774219871,-0.0302753989,0.1482476145,-0.0246563498,-0.2292751521,0.3347116113,-0.0698234886,0.2151486725,0.3179880083,0.136447832,0.1477787793,0.4958414137,0.1994640976,-0.3975553811,0.4418984354,-0.0755604357,0.2939326167,0.091992937,-0.0275791306,0.2489297539,-0.1203065887,0.0949374884,0.1931623071,0.0448031351,0.0250683911,-0.0283336118,0.0092360545,-0.003529883,-0.4363783002,0.1996441931,-0.1517345607,0.5237365961,0.1381768137,-0.2355799377,-0.3791066706,0.0496141165,0.1934675574,0.1337411702,0.0738770515,0.164358452,-0.4970503151,0.0758114308,0.1406286359,0.4142273366,0.085603483,-0.0225954652,0.1782038212,0.0315712206,0.0914835781,0.0926709026,-0.0481353514,-0.2960968614,0.007522766,0.1244783849,-0.3474850059,-0.2607431114,0.0564990379,-0.1145010665,0.181301102,0.12085329,-0.0085927639,-0.6265869737,-0.085976921,0.4911956489,0.3378349245,-0.117398493,0.0225376952,0.3983537257,-0.1619605273,-0.4426513016,-0.1478760391,-0.1674231142,0.0900754407,0.1807972044,0.1200106442,-0.3284374774,0.2297539413,0.3346863091,0.122901015,0.42855528,0.2178947181,0.0859570131,0.0995591283,0.0065946802,0.0419074744,0.4521704316,-0.2863284051,-0.4647592902,-0.1230604053,0.0439880677,0.2652990818,0.1608941108,0.3431626558,-0.0893259197,-0.2227253467,-0.5019604564,-0.2145175636,0.2109344602,-0.0418740287,-0.0549856126,-0.3572918177,-1.0777642727,0.3356913924,0.180964455,-0.0971047059,0.176104188,0.1328935921,-0.1349574029,0.1247949377,0.0295259468,0.9321939945,-0.1772612035,0.2380483449,0.3661992252,0.2965842485,0.7135195136,-0.1419455558,-0.2076855749,-0.3434434831,0.2142758071,0.0496629514,-0.0595827922,0.2961932421,0.2053502798,-0.1560040116,0.2518339753,0.2036560327,0.4291244149,0.1284229755,0.2230033875,-0.2504243255,-0.3115473688,-0.1322673261,0.011568374,-0.3206801116,0.0376142636,-0.1738489121,-0.0039345194,0.0170445275,0.1563149393,-0.4413566291,0.013680581,-0.2030917108,0.0191740841,-0.0973052308,-0.4997603893,0.4716842175,0.1864792705,0.3057287037,0.2079543322,-0.1851714551,0.0395433493,-0.2062273771,-0.0801260099,-0.200369224,0.0007408783,0.2439535409,0.0106173595,-0.02981041,0.0981591716,0.3809088171,-0.0212844405,-0.1528524756,-0.2496741116,0.3948430419,-0.3585500419,-0.2738022208,0.2733775377,0.0186455511,-0.0811736807,-0.0242215544,0.3156312704,-0.1783572286,-0.0747335479,-0.199306339,0.1323680729,-0.0118841073,-0.3328549266,0.3187772632,-0.2915685475,0.3401264846,0.1281153113,0.112278156,-0.0593238398,0.0284579974,0.2757189572,-0.4038208425,-0.1267390847,0.108664155,-0.2364013493,0.0819725394,0.1765654087,-0.1015785858,0.0840267688,0.0360073894,-0.262513876,-0.1277973801,0.1553677619,0.0030010391,0.3532754779,0.0457549766,-0.01007911,-0.1774008423,0.0435951203,-0.173563078,0.1767266095,0.1601530761,0.1843805462,0.2037207335,0.0576778539,-0.1015830711,-0.1093999445,-0.0787587985,0.0701249987,-0.1479987055,-0.0097603034,-0.5596266389,0.2132185996,0.057190001,-0.1296141893,-0.0371959209,-0.2645624876,0.1283727586,-0.3482802212,0.224641636,-0.0293610059,-0.0460956432,0.1597710997,0.5605437756,-0.0127614439,-0.2915018201,0.0097551402,0.1320857406,0.4077638686,0.4412096739,0.0612201355,-0.1038744971,-0.124504149,-0.0271415021,0.2084719092,0.7594000101,0.0702517182,0.321388334,0.1391651183,0.1181168109,0.2112986892,0.1848992705,0.4557944238,-0.091823712,0.2779638171,0.2086735964,-0.0071698423,-0.089571856,0.1609559953,0.2544286549,-0.1813842207,-0.1500155181,0.3750498891,0.1749408543,0.2845790088,-0.1489906907,-0.2645767033,0.2097462416,0.0274868384,0.3904090822,0.1939252913,-0.0120404856,-0.2825765312,-0.0412243381,0.400023371,0.2880563736,-0.4512830079,0.2234057039,0.2511084676,0.1872936934,0.1809878051,0.5107917786,-0.4333376288,-0.0405470617,0.4104088545,0.3606260121,0.2895861268,-0.0849083811,0.2871323824,0.1232731864,0.0652575269,0.0793494359,0.2252430618,0.0982777029,-0.0702172965,-0.2390363365,-0.2490397692,-0.3212158382,-0.0558160432,-0.3482563794,-0.166655913,-0.1457806677,-0.0525197908,-0.0341498405,-0.4228678942,0.0582166426,0.1503516138,-0.0318880305,-0.0337354429,0.1851816922,0.1267452836,-0.3633873165,0.2255139053,0.3972007334,0.109761402,0.0152250426,-0.0066316151,0.2950597107,-0.1489103138,0.1632122695,-0.137565732,0.0216868594,0.1722297668,-0.2595814764,0.1358036101,-0.0716032982,0.0557165816,-0.1894650012,-0.0126526691,-0.1854285449,-0.4377538562,0.6592111588,0.2342060059,-0.210686475,0.1065575182,0.213805005,-0.15989317,-0.34995386,0.4518506527,0.3653381765,-0.0347416475,-0.2207174301,-0.0010922833,0.1904411018,0.2138692886,0.0775146186,-0.1051168963,-0.0865746364,-0.262545675,-0.7192497849,0.2276403904,-0.1733564883,-0.6922211051,0.2622715831,-0.1413422525,0.3419551551,0.1014136001,-0.2939868569,-0.0348933637,0.0552139208,-0.2313892543,-0.173784256,-0.4438749552,0.0258593597,0.051515229,-0.1029371843,-0.2470999211,-0.0737904906,0.0714299902,-0.1363351047,0.0353096686,-0.198196426,0.1184593067,-0.4753552675,0.1128882542,0.1332099736,0.2694587708,0.188393116,-0.0659010634,-0.2441362441,0.0553755797,-0.1352278739,0.2859680355,-0.1012534797,0.0479674079,-0.5305737853,0.006874328,-0.292224735,-0.1701012403,-0.0079450775,0.0704059377,-0.1326496601,0.1036821082,-0.3454536796,-0.0599828139,0.3603956699,0.5799030066,-0.2321650386,0.022356391,-0.15164873,0.1939751208,0.3261487484,-0.2340167314,-0.4163009524,-0.062206354,0.0432803966,-0.2508497536,0.0315812342,-0.5333890915,0.0338165611,0.263151437,-0.02095294,-0.1001891419,-0.0299169142,0.2262379229,-0.2082659751,-0.161411494,0.4423395097,0.079369992,-0.0737436116,0.0107915131,-0.1659831405]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/433","title":"How to reuse functionality of a (generic) dataset?","comments":"As temporary fix I've created [ArneBinder\/nlp-formats](https:\/\/github.com\/ArneBinder\/nlp-formats) (contributions welcome).","body":"I have written a generic dataset for corpora created with the Brat annotation tool ([specification](https:\/\/brat.nlplab.org\/standoff.html), [dataset code](https:\/\/github.com\/ArneBinder\/nlp\/blob\/brat\/datasets\/brat\/brat.py)). Now I wonder how to use that to create specific dataset instances. What's the recommended way to reuse formats and loading functionality for datasets with a common format?\r\n\r\nIn my case, it took a bit of time to create the Brat dataset and I think others would appreciate to not have to think about that again. Also, I assume there are other formats (e.g. conll) that are widely used, so having this would really ease dataset onboarding and adoption of the library.","comment_length":8,"text":"How to reuse functionality of a (generic) dataset?\nI have written a generic dataset for corpora created with the Brat annotation tool ([specification](https:\/\/brat.nlplab.org\/standoff.html), [dataset code](https:\/\/github.com\/ArneBinder\/nlp\/blob\/brat\/datasets\/brat\/brat.py)). Now I wonder how to use that to create specific dataset instances. What's the recommended way to reuse formats and loading functionality for datasets with a common format?\r\n\r\nIn my case, it took a bit of time to create the Brat dataset and I think others would appreciate to not have to think about that again. Also, I assume there are other formats (e.g. conll) that are widely used, so having this would really ease dataset onboarding and adoption of the library.\nAs temporary fix I've created [ArneBinder\/nlp-formats](https:\/\/github.com\/ArneBinder\/nlp-formats) (contributions welcome).","embeddings":[-0.1799507439,-0.1630488932,0.1238646433,0.301163137,0.164542526,-0.0197150595,0.2357494235,0.1255231351,-0.23343952,-0.2266042978,-0.2068018764,0.4248532653,-0.0154535882,0.4941488802,0.1642580032,-0.1521085203,-0.191160813,0.2969347239,-0.0964312106,0.0303155705,0.0905729011,-0.2147969306,0.0142098125,-0.0728892311,-0.3415699005,0.0581459105,-0.2382072061,-0.1610350013,0.2872367501,-0.2076722533,0.4816417396,0.3233340979,0.4522971511,0.1928937286,-0.0001256535,0.0015537674,-0.0469903126,-0.0958321318,-0.2618281841,-0.123023212,-0.177174136,-0.2196676731,0.0758252293,-0.2772657871,-0.312620759,-0.4054501355,0.1707173139,-0.5472615361,0.4351249635,-0.0527805835,0.034754701,-0.0491020046,-0.2048210651,-0.0957496688,-0.1829075664,0.3975821137,-0.0351432338,0.363265276,0.7149806619,0.3077362776,0.079074204,0.0520960614,0.1537605077,0.0743611604,0.2044145763,-0.1829632074,-0.027270345,0.0770743713,0.1335144043,0.3593483865,0.9471532702,-0.377556026,-0.026282575,-0.0663213357,0.0251012538,-0.2421931624,-0.3016163111,0.1777876765,0.2174054235,0.1083229557,-0.2695991397,-0.3563554883,-0.3021561503,0.153106451,0.3087294698,0.1036858931,-0.0550544783,0.2560859621,-0.1795107722,-0.0511036292,0.0324239619,-0.1912173927,-0.1762356013,0.0214986764,0.129818365,-0.1600202173,-0.3937460482,0.1345305294,0.3059707284,0.400801599,0.2349507511,0.2427453846,-0.3303095698,0.0705791712,0.3837603033,0.030420471,0.395814836,0.3331991732,0.1108265072,0.1565252095,-0.4895265102,-0.1400072277,-0.4615721405,0.42481336,0.244687289,-0.0966380537,0.0143073853,-0.2565381527,0.0539826229,-0.0677935183,-0.3839553595,-0.1775020659,0.1688036174,0.0274427552,-0.1083542779,0.1866581738,-0.1149033979,0.2996137142,-0.0519872941,-0.3926970363,0.0414590575,0.0660666898,-0.3845517933,0.103150636,0.3347450793,-0.3624421954,0.2111147642,0.290615648,-0.1975218058,-0.1624846607,0.4515014291,0.1128519773,-0.289604038,-0.5491024256,-0.1631789953,0.0887266025,0.2056605667,0.152061224,-0.3457585573,0.0904614478,-0.0715162307,-0.3300113976,-0.2041600645,-0.0063873501,-0.278200835,0.0263878256,-0.0269864406,0.4530582428,0.0225729533,-0.4089550674,-0.1972793937,-0.1249730065,-0.2701570988,-0.0629684925,0.0844088197,0.1705598384,-0.5173358917,-0.1218344569,-0.1707620025,0.0204946827,0.4264614582,0.1836935729,-0.4590799212,0.2215601355,-0.1088499948,-0.1348463595,0.5498046279,-0.0342615545,-0.0288706031,-0.1476153135,0.1924620569,0.049434524,0.0477854945,0.1223483831,0.158333689,-0.2082876414,-0.3598217368,0.462233901,0.0804830268,-0.2411234975,-0.0571154393,-0.1795306653,0.1685245633,-0.014055917,-0.1792445928,0.1057764441,0.0891060829,0.3231929541,-0.0114591094,-0.2514335811,0.1159164459,-0.2199731916,0.0127400355,0.2945499718,0.0900653452,-0.2662425041,-0.4334064722,0.1091357246,-0.3481083512,-0.2516926229,0.6138749123,-0.471957624,-0.3473468423,-0.2377176732,-0.0981097817,-0.0745937154,0.0130996248,0.550140202,0.1129533127,-0.0623092242,-0.1060123816,0.0118912859,-0.1050231233,-0.0751531124,-0.0482923388,0.2318774313,0.0882845521,0.0292534623,0.1339523345,0.5698366165,0.0205962472,0.0123466682,0.1560823768,0.1789671332,0.0864400789,0.2007521093,0.0795055404,0.1081714854,0.1390642822,0.1161047369,0.4375793338,-0.070473358,0.325162679,-0.189734742,-0.4226646721,0.196849972,0.1529302448,-0.0267467089,0.0817814991,-0.0742196888,0.1503663957,-0.1342151165,-0.1533535719,-0.0914232358,-0.140857473,0.1334082633,-0.0155804045,0.3934088647,-0.3671460152,-0.0135990549,0.2255435139,-0.1655054688,0.2397111356,-0.0996555462,-0.2096623033,-0.1188334748,0.3966802955,-0.2012539059,0.0796134099,0.0709175915,0.2562955916,-0.0254556332,-0.1348943859,0.0339530669,0.2755458951,-0.2206451148,0.1729284227,-0.2147314996,0.1619266123,0.0858185217,0.0458368175,-0.0484194942,0.2283241749,0.0365096927,-0.0396042466,0.2050334215,-0.7187316418,-0.3843326569,-0.3618913889,0.0462508127,0.0426990837,0.0434616283,-0.217721194,0.0436682403,0.1747508049,0.1910142899,0.0073162029,0.5471546054,-0.3260409236,0.0058482583,0.126207605,-0.3436647952,-0.454536885,0.0316418037,0.4644823372,0.3778604269,0.5153445005,0.0651541874,0.0951043218,-0.4385221899,-0.3185591698,0.2125447989,0.2142169029,-0.089466013,0.1643580645,-0.047359582,-0.3831494749,-0.2738624215,0.1293023974,0.1294405013,0.247442022,-0.061496418,-0.5198377967,0.1319721937,-0.168028906,-0.0559551716,-0.0413239896,-0.2022127211,0.2463888377,-0.0680216178,0.0100987647,0.3911086917,-0.3180051446,0.2625139058,-0.1282362342,0.3288136125,-0.2411362529,-0.1063248292,0.2670603693,-0.0947854668,-0.1483853012,-0.164719522,-0.3556164205,0.2208846509,0.5239163637,-0.2483144552,-0.3911144435,-0.0807031915,-0.2820880413,0.2931441367,0.2598619759,0.112661317,0.03359548,0.1531504542,-0.0284575745,-0.0957124829,0.2486357689,-0.0772412643,0.1728299409,0.4786517322,0.0289229956,0.2554177642,0.2370346338,0.2628549337,-0.4994132817,0.4781120718,-0.064456448,0.3653496206,0.0124837635,0.0601559393,0.3162155449,-0.0574135184,0.1303061396,0.2904903889,0.014816355,0.0195799787,-0.1448473036,0.0978016481,-0.0818407461,-0.451908946,0.1720603108,-0.3112580478,0.5501087904,0.2566040158,-0.1879154742,-0.2956259847,0.0017339475,0.2033753395,0.0054501556,0.1340853274,0.1413051635,-0.4077976942,0.063593477,0.1543861628,0.5220628977,0.0613631792,0.1057687476,0.1982023418,0.0131581631,0.0105360802,0.0878077745,-0.1165407822,-0.2600839436,0.049999509,0.057019569,-0.1825242937,-0.2446062118,0.0306922793,-0.1698570251,0.2496942133,0.062387038,0.096273981,-0.5350401402,-0.0634318441,0.2878100872,0.4257284999,-0.0501010045,0.1610472053,0.2833544016,-0.0864791572,-0.3640463948,-0.0622083172,-0.2904027998,0.0615098998,0.145731166,0.1102215201,-0.4019153118,0.0985125676,0.3380228579,0.2135923356,0.5034178495,0.130032897,0.0848659873,0.0585207008,0.0266865622,0.0596142262,0.3944862783,-0.174719587,-0.3653547168,-0.0960907564,-0.0204855688,0.4007440507,0.1435943246,0.2843815088,-0.1893704981,-0.2462013513,-0.3840133846,-0.2043184638,0.2717859447,-0.0535224788,0.0147570167,-0.4809937477,-0.8999791145,0.0984022394,0.2943085134,-0.0439557843,0.2397073358,0.0567827187,-0.0435920767,0.108782582,0.1602571011,0.8434744477,-0.1961313933,0.2386291921,0.3054979742,0.274053067,0.6816400886,-0.0988371,-0.1968141198,-0.318421334,0.0468251854,0.0029655092,-0.103858687,0.4689374566,0.2823358774,-0.1846817583,0.305568099,0.2822430432,0.4324681163,0.1244658381,0.1705072373,-0.3846611679,-0.2404798716,-0.0167596638,-0.0361401848,-0.4089757204,-0.0838619396,-0.2355683148,0.1563217491,0.1682070196,0.1049347296,-0.3196618557,-0.0686812252,-0.1502848268,0.003634403,-0.0997600406,-0.3818408847,0.4917150736,0.3532697856,0.2701352537,0.3388346434,-0.1113311872,0.0024089708,-0.0978394672,-0.0559127741,-0.2117899358,0.0061994442,0.2473798543,0.0928796008,-0.0526176952,0.1945834309,0.4021859765,0.1136531308,-0.1892237663,-0.205874756,0.331474036,-0.2589922249,-0.2896231711,0.1539376825,-0.0430790745,0.0000661884,-0.0287665725,0.2122863978,-0.064898625,0.1003776193,-0.0761051551,0.1135409996,0.0044749957,-0.290958941,0.3216520548,-0.259632051,0.3395314217,0.0322425067,0.090078935,-0.1112048402,0.1200570986,0.0695297644,-0.309892565,-0.0903278589,-0.0064887377,-0.2308074981,0.0986790061,0.2386226207,-0.1534247845,0.1253493577,0.0982585922,-0.2655629814,-0.0656959638,0.2454054803,0.044494424,0.3656274676,0.0755943507,0.0346213505,-0.266426146,-0.008973984,-0.1597735733,0.2221191227,0.2604975402,0.2968042195,-0.1277662069,0.0556478538,-0.1947854459,-0.055256851,-0.0472681187,0.1619271487,-0.0808453113,-0.0423704311,-0.4844783843,0.2209660709,0.018322276,-0.2221826613,0.0013390851,-0.3314947188,0.0516983345,-0.3858225942,0.1199370697,0.0038313987,-0.1374461353,0.179216519,0.4918342829,0.0160547011,-0.1408525705,-0.021074919,0.0875252038,0.3569458127,0.4175534844,0.0975294486,-0.0776773244,-0.1272099465,-0.0602807589,0.2523084581,0.8456018567,-0.0513595827,0.2690608501,0.1653821617,0.0364645571,0.1891157329,0.3202407956,0.4004516304,-0.0580671579,0.23731637,0.3621992767,-0.0176123679,-0.1529698521,0.0938370451,0.2652910054,-0.118705295,-0.1269673556,0.3678472042,0.2173218876,0.2349041849,-0.1331760734,-0.3314281106,0.0801103041,-0.0461395979,0.4241732061,0.2546234727,-0.0003765837,-0.371214509,-0.0409177244,0.3536787629,0.3149034381,-0.5061255097,0.2637831867,0.2403619587,0.1511033773,0.1997778416,0.4512698948,-0.4899795353,-0.0190248843,0.3411895931,0.2814656794,0.3299975395,-0.0871854722,0.243400842,0.1488704234,0.0203441419,-0.0591660775,0.133140713,-0.0439845398,-0.1711413711,-0.2579449117,-0.2916388214,-0.313174665,0.0163487494,-0.3172549605,0.0200875551,-0.1961405724,0.0309492704,0.0887907743,-0.4552840889,0.1200851202,0.13582775,0.0891097561,-0.0086699659,0.2019193619,0.2070171684,-0.2760693431,0.2678615153,0.3765708804,0.1550698876,-0.0136616025,-0.0835415423,0.1810963005,-0.2534424663,0.1592851728,-0.1912187338,0.1101081744,0.2435279191,-0.361166209,0.2401907444,-0.037936952,0.0864960179,-0.0184717197,-0.0719639212,-0.3018695414,-0.5542805195,0.4207468629,0.2541737854,-0.2173120677,0.2080229074,0.144292295,-0.1960360855,-0.3841302097,0.3648413718,0.3382545412,-0.0232230723,-0.2342289239,-0.006267596,0.1784893125,0.1886086911,0.1004008576,-0.081240721,-0.009005704,-0.3424267173,-0.7371789813,0.2264852524,-0.1631511599,-0.5314537287,0.1868384033,-0.2040281445,0.30616045,0.1617087871,-0.2202134579,0.0304538943,0.1450514048,-0.244666934,-0.1265503764,-0.5337688327,-0.0108779361,-0.0607891381,-0.19209297,-0.3520793319,-0.0833519027,-0.0041289572,-0.1287162155,-0.0258605704,-0.2178068012,0.1144210845,-0.2999043465,0.053316839,0.1052670628,0.3086135387,0.2020618021,-0.0635632351,-0.188479811,0.1390470266,0.0592366979,0.2836537361,-0.0810500979,0.0792640448,-0.5827376842,0.0150615508,-0.2413267791,-0.1473367363,0.0590785779,0.147760883,-0.1732148826,-0.0053082118,-0.4080114365,-0.0860990658,0.306384027,0.6834857464,-0.1929555237,0.0505989157,-0.0933306962,0.2378285825,0.3226784766,-0.2557693124,-0.4162865877,-0.1523157954,0.0503258891,-0.2661451995,0.0756995901,-0.342038244,0.0695347786,0.2692900002,-0.0591107793,-0.0666951835,-0.08387357,0.1948474944,-0.2001720816,-0.1896668524,0.5265093446,0.0731472224,-0.1157029122,-0.0304374844,-0.1289470196]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/426","title":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter","comments":"Yes that's definitely something we plan to add ^^","body":"It would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?","comment_length":9,"text":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter\nIt would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?\nYes that's definitely something we plan to add ^^","embeddings":[-0.2617352903,-0.2195705324,-0.1818095297,-0.1607926488,-0.3494630158,-0.0038602545,0.1658258736,0.2341453433,0.0505689904,0.0404876024,-0.0466238558,0.3939706385,-0.105338335,0.4011116326,0.0562107712,-0.2518656552,0.160697341,0.0987035781,-0.091493018,0.0988577604,-0.0390915945,0.0042517176,-0.2140696496,-0.0931789279,-0.2396363318,-0.1281022578,-0.1953098178,-0.0855928957,0.0326323919,-0.2810249627,0.0276083481,0.4175644815,-0.184824422,0.1111745387,-0.000105187,-0.246946618,0.2412562817,0.237780109,-0.0795097575,0.3257490396,-0.3090558052,-0.4262056947,0.082673654,-0.2000194043,0.1416012645,0.193863526,0.2608658671,-0.3042229116,0.0708668157,-0.211009562,0.1632284969,0.2195428312,-0.1628806293,0.249911353,-0.1671710163,0.3047628701,-0.1301045269,-0.293125838,0.4885960519,-0.194544062,-0.2686460912,0.5029209852,-0.1830316931,0.106073007,0.1273188442,-0.1033673435,0.0816237405,-0.3953154385,-0.1418834031,0.2762221396,-0.0261006504,-0.1562751532,-0.4176866114,-0.2619740367,-0.1124360859,-0.4772993326,-0.2075065225,0.1108126789,-0.2143891305,-0.0648312643,-0.457949549,-0.3942385018,0.2657066584,-0.1148887277,0.306813091,0.488260746,0.2550447881,0.0819491148,0.2128991932,-0.2288335264,-0.0910366103,-0.2749279737,0.1047925651,-0.0093001481,-0.6403989792,-0.1199571937,0.0353421122,0.0530753024,0.0213019587,0.0032417662,0.0579099543,0.2978697717,0.0877984688,0.2180091441,0.1339171678,-0.381370604,0.2395961881,-0.0260447171,0.1107545421,-0.3521470428,-0.1759427339,0.2688884735,0.2513138354,-0.0352005661,-0.1973207742,0.0522279292,-0.0930328667,0.099741213,0.1296702027,-0.361405462,0.1447912604,-0.2812577188,0.1345796287,0.0106473202,0.0473445654,0.3528572023,-0.2890741825,-0.1322047859,-0.1614204943,-0.2068188637,-0.0931813493,-0.0245167855,-0.0773898512,0.1914330274,0.1902996153,0.2086874396,-0.0327614546,0.1095308363,0.1047180817,0.3275120854,0.296012342,-0.0620720163,0.063872993,0.1353190541,-0.1383395791,-0.0830934793,-0.1680490524,0.058375854,-0.3463800251,-0.0371033289,-0.2376165986,-0.1380416006,-0.0901978388,0.1483397484,0.1272954941,-0.1049658805,-0.238432914,0.7012827992,0.0445413142,-0.0805307031,-0.2671084404,-0.0806902424,-0.4311804771,-0.1875489056,0.2031621933,0.0868012235,0.0089974795,-0.0356398225,-0.3112593591,-0.0764401555,0.315071255,0.3028850555,-0.180105418,0.0649364442,-0.069495514,0.4847205877,0.5901512504,-0.1190975085,0.1322699934,0.4102869928,-0.1377651244,0.0189298224,0.3589554727,0.2611442804,0.1668178588,-0.0542060658,0.0957471728,0.5074102879,-0.3133209646,0.2812963724,-0.0351647176,-0.143623054,0.2057074904,0.1075434014,0.035253752,0.0639513358,0.0391287468,-0.2810564935,0.4377708137,-0.2622989118,0.1216640323,-0.0300683193,-0.0936163291,-0.2019037306,-0.105148688,-0.276561141,-0.2871222198,0.040180631,-0.1283135265,-0.0331614576,0.206087023,-0.3655872345,0.4393782616,0.123242341,0.0408344269,0.0492414646,0.1726255864,0.2608201504,0.1187828481,-0.2526117861,-0.2238705903,0.1696892083,0.0293370876,-0.1621979326,-0.1024246216,0.2271467,0.1546097994,-0.10709811,0.1043164954,0.152935341,-0.0245427042,-0.0110007012,0.1566526592,-0.0342428796,-0.0093519585,0.2919555306,0.7340606451,0.3043384254,0.1640231311,0.0447369516,0.0788839683,-0.1628268957,-0.0638409182,-0.2050166279,-0.1693061143,0.5050578117,-0.0367118455,0.4310558438,0.1020874903,0.0489810109,0.2703766227,0.0383765101,-0.108386375,0.1587675363,0.1279766113,-0.0182232223,-0.0908584818,0.205960989,-0.4538829029,0.3682850301,0.5166467428,0.099110797,-0.1298432499,0.0945721418,0.201985836,-0.1492881328,0.2353907675,0.0890965536,0.2551118731,0.2462469041,0.138733834,-0.125264138,-0.021624146,-0.0668671876,0.2055539638,0.1741821468,0.276021868,0.0210296456,0.1306229979,-0.3170219362,-0.2349536717,-0.3844031096,0.191036433,-0.0418449938,0.1250573695,-0.1098199189,-0.2192609012,-0.2143878639,-0.0420594066,0.1544019729,-0.2337571383,-0.3690338433,0.0619613938,0.1290933341,-0.1726433337,0.1488343477,0.0424021184,0.2124100327,-0.2506121397,0.071595937,-0.2478254735,-0.1324509084,0.0292471703,0.0791570246,0.3312518597,0.2530089021,0.4469469786,0.2666684389,0.127472356,-0.4003023505,-0.2225798517,-0.0944647938,-0.1224615648,0.0700338036,-0.0046379729,-0.0890238807,-0.1499924362,-0.4411938787,0.1273988932,0.1312247366,-0.1568423063,-0.2021129578,0.0056981728,-0.0066268654,-0.0154504674,-0.2118165791,-0.5005150437,-0.3928304315,0.3950667679,-0.2679482698,0.2918401659,-0.3237549067,-0.1066805422,-0.0920976996,-0.3090139031,0.0362032279,0.0994861349,-0.3239731789,0.0426702686,-0.0940995589,-0.1030081213,-0.3543889225,-0.0531857237,0.0353923142,0.3441718817,-0.1000063717,-0.0277133584,-0.2145379633,0.1507392079,0.2943863571,-0.0543749221,0.1268445849,0.0955102518,-0.1584196985,0.1816501766,-0.2594280839,0.032629285,-0.0433519818,-0.1534583122,0.1882921904,0.2696032226,-0.0375697538,0.6174558401,0.1310176402,-0.076590091,0.0830881149,0.1688289195,-0.212561667,0.1252278984,-0.494939059,-0.0352453887,-0.1261063963,0.0479500256,0.0911003873,-0.0790479854,-0.5478128791,-0.0140881408,0.1155878305,-0.2131924033,-0.1070595235,0.1566868126,-0.31992957,0.2999628484,0.0718381554,-0.3348490894,-0.3902180791,0.0691337287,-0.2567119002,-0.04477828,0.2427627891,-0.2159781009,-0.4863491654,-0.2375615388,-0.5574262142,0.3290404081,0.1885097623,0.1531513333,-0.0587623529,0.0213283133,-0.0155326175,-0.1224692836,0.5170815587,-0.3571929038,-0.2769497931,0.3018264472,-0.3743883371,0.1077692881,0.2187685072,0.0014378192,0.4862264395,0.4413681924,0.3387447,-0.2018859386,-0.4363304675,0.2601941228,0.0599320158,-0.058370173,-0.0239480864,-0.0313337892,-0.001910647,-0.1285357028,0.0959984139,-0.1231934056,0.0843047276,-0.0232807826,-0.1316317916,-0.1405375451,-0.0016061529,0.2590397298,0.0328073129,-0.1195596233,0.282468468,0.2779839039,-0.3713709414,0.1900435835,-0.171062991,-0.0182706118,-0.2466256618,0.0315770023,-0.0565547161,-0.0962794572,0.7045500875,0.2217977196,0.2387915701,0.4274035692,-0.1055192277,0.4070324004,-0.1982840002,-0.0384919606,0.3695097566,0.192134887,-0.2623895407,-0.2453113496,0.3198356926,0.2984897494,-0.0926896036,0.4692162275,-0.2159597576,-0.0383242257,0.2659738362,0.0672718585,0.7050275207,-0.3232125342,0.0545202605,-0.4205074012,0.0885138512,0.3833313584,-0.3013653159,0.2035217434,-0.014224112,-0.1887605339,0.1163568795,0.0467877612,0.2426073253,0.4464696646,-0.2428376824,0.2396405488,0.3270199001,-0.1963644922,-0.0939351916,0.5294165611,0.2031310946,-0.3943642378,-0.1168270931,0.0744182318,-0.0218047556,-0.3506497443,0.0499708802,-0.2884840369,0.169198975,0.1916107833,-0.3431823254,-0.0090010781,0.1352891177,0.3652839959,-0.2401524782,-0.0595469512,-0.085685581,0.1052790061,-0.1467464864,-0.0057677529,-0.0194244664,0.0635028929,-0.1405880749,0.3117502332,-0.1910571158,-0.2647877932,0.1761603206,-0.2329791784,-0.2224758714,-0.0230849497,0.0673585981,-0.5983371139,-0.2510301173,0.1942293197,0.1157130972,0.4248101115,0.0336652324,0.3952545524,-0.3785684109,-0.1652209014,0.1307968348,0.295347929,-0.1821744293,0.5997629166,0.1771358997,-0.1005156413,-0.0133960759,0.0729740337,0.3349862099,-0.0406703129,0.2903628051,0.0741348416,-0.1516385525,-0.3148339093,-0.204742074,-0.1987797469,0.0951880664,0.2580714524,-0.1274250597,-0.2094696611,0.0898716003,-0.1760489345,-0.1158707142,0.1424324214,-0.2121586204,-0.155414775,-0.1448025107,0.0507943816,-0.2550164163,0.2020139992,-0.0197243635,0.0182066113,0.0462087654,0.5038390756,-0.3390859962,-0.1369871348,-0.2471359521,0.2038360834,0.3705517054,0.0226604883,-0.1867373735,0.2715517581,0.1832449436,0.0818043575,-0.1311897337,-0.283500731,0.2333596051,0.0708861873,-0.1028851867,0.0684438571,0.132027328,0.1717061251,-0.2786633074,-0.2922906578,-0.1809658408,-0.1769933552,-0.0511378497,0.129317835,0.2683158219,-0.3185954094,0.3554596603,0.0735987797,0.1230268478,-0.0522479713,-0.1522011608,0.4201330543,0.1730191708,-0.1878517717,0.2539328337,-0.0427384339,0.0180727709,-0.0902765468,0.2540989816,0.073631309,-0.2764262855,-0.0432521738,0.2830457687,0.3570231795,0.2422501743,0.2028211653,0.0889726877,0.1605675519,-0.2871599495,-0.2128375471,0.5780342817,-0.0067022666,0.2350682914,-0.1084105521,0.2504038513,0.3535024226,0.1114020646,-0.0339892507,-0.15066275,-0.3303405643,0.0841902494,-0.1373848468,0.3092551529,-0.0800801888,0.0939318389,0.0739258975,0.1034121811,0.1020893604,-0.0002473979,0.8153515458,-0.0234121345,0.263309598,0.2494825572,-0.1002207026,0.1916939467,0.1539123207,0.1319073737,0.2172074914,0.1924083084,0.1732734591,0.074837327,-0.4974792004,-0.2026384771,0.6765787005,-0.0537955724,-0.3482597172,-0.0641894564,-0.1043571606,-0.1039176062,0.1079872176,0.1161049679,0.1609156132,0.7483167648,0.0161568504,0.367441386,-0.046418149,0.2087814808,-0.1618713588,0.338860482,-0.1455581486,-0.2339701802,0.1241128892,-0.1832483411,-0.0769716352,0.3968537152,0.2147031426,0.3262762427,-0.1318256557,0.2185249478,-0.303501904,-0.0866683349,-0.2816131413,0.2763218284,0.0288546719,-0.0008052557,0.4190347195,0.1016726643,-0.1193161234,-0.2150267214,0.0209271461,0.0147487931,-0.2529478669,0.2356297374,0.1978082061,-0.2425162494,-0.039006874,0.1106572524,-0.4973834455,-0.2906754911,0.6427894235,-0.4903197885,-0.0661622211,-0.1436402798,0.1133292392,-0.1507408768,0.6730071306,0.0872429833,0.3700416088,-0.0540036485,-0.2665868402,-0.4954598844,-0.3581890166,-0.1060644314,0.0626946166,0.1423410624,0.547365129,0.1634825617,0.3249936402,-0.0221232083,-0.2920533121,0.0108882496,0.0886215717,-0.4467107952,0.2216050327,-0.0609962009,-0.0582752936,0.1035423353,-0.2717359066,0.3114349842,0.0843442753,0.1136701405,-0.189051643,0.0960753262,0.4578845799,0.183459118,-0.0830554441,0.1532821506,0.3724195957,-0.3071487248,-0.0396270305,-0.1480249316,0.0478914566,-0.1642689109,-0.0959906802,-0.021602191,0.3425107896,-0.2072344571,0.1825242043,-0.0455318131,0.2161662579,-0.2770949602,-0.0427026749,0.1575825065,0.0591826364,0.1133983731,0.1789693981,-0.0898631588,0.2413684279,0.0164330583,0.0754101053,-0.1364383698,-0.1177848056,0.1823690385,-0.3877607286,-0.2469100803,-0.34068501,0.1267187744,0.0098047983,0.3991346955,-0.3438272476,-0.3688398898,0.2398637235,-0.0462038703,0.0236853771,0.5605224371,-0.084506385,-0.2702028751,-0.1965469122,-0.3071301579,-0.0112254489,-0.2867701948,-0.207918331,-0.2644223869]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/426","title":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter","comments":"Yes, that would be nice. We could take a look at what tensorflow `tf.data` does under the hood for instance.","body":"It would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?","comment_length":20,"text":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter\nIt would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?\nYes, that would be nice. We could take a look at what tensorflow `tf.data` does under the hood for instance.","embeddings":[-0.2759443223,-0.2842460275,-0.2022300363,-0.1166619658,-0.2691851258,0.1010134444,0.2325940132,0.3550632596,-0.0165046155,0.0616694838,-0.1445138901,0.3324525654,-0.1353501827,0.4115578234,0.0130814621,-0.2661807835,0.0175814442,0.1032483652,-0.1281686127,0.1098897457,0.0762666091,-0.0655459911,-0.1153079122,-0.0926844701,-0.132297352,-0.1024550274,-0.097571224,-0.202066645,0.1825311631,-0.059196949,0.0819187984,0.4362562299,-0.1955319494,0.3209602237,-0.0001035186,-0.0330525786,0.0230353232,0.1411820501,0.0554775074,0.1950994432,-0.2407266498,-0.3496336043,0.0154762631,-0.2446846366,0.114108704,0.1449297965,0.2296795696,-0.2520439625,0.1510989964,-0.1482693851,0.1650044471,0.3244563043,-0.1013417765,0.3549084663,-0.1160900891,0.288120389,-0.1332563013,-0.2739548385,0.4300302565,-0.165862903,-0.1933044195,0.4834124148,-0.204895094,0.1841047704,0.1868391186,-0.0659237429,0.0773275569,-0.4097806513,-0.1818062216,0.4087259471,-0.0675883219,-0.2307182699,-0.4504499733,-0.278904289,-0.1264075339,-0.4070104361,-0.2271302044,0.0858728066,-0.2099242806,-0.0233104695,-0.4120799899,-0.3114105463,0.1997829229,-0.0812529847,0.1616560817,0.3613924086,0.2724672854,0.1311332434,0.1919373572,-0.1969308853,-0.0564663969,-0.2537738979,0.3110478818,0.0066752806,-0.6924254894,-0.1842935532,0.0167129505,-0.2867717147,-0.0392810479,-0.0706893578,0.1534596086,0.3635826707,0.0018450117,0.2756719887,0.0096383132,-0.3759275675,0.0630588233,-0.0125332382,0.1332935989,-0.4187017977,-0.2308539599,0.2932199836,0.126777783,-0.0834778473,-0.2717575133,0.0474008434,-0.0980958119,0.2626233101,0.0492146499,-0.4000425637,0.0890791491,-0.1915702969,0.1827962846,-0.0364476778,-0.0701015964,0.3581661284,-0.1625254452,-0.2483626902,-0.1925841868,-0.1561348289,-0.0763624981,-0.006903613,0.0626296625,0.1361531317,0.1571865976,0.1219406947,-0.1967089027,0.0479605384,0.1393325925,0.3047914207,0.3225729167,-0.1254656911,0.0392487682,0.1707835346,-0.1373326778,-0.1019913107,-0.1454034448,0.085971579,-0.3219798803,-0.0447810702,-0.3592560291,-0.2022037357,-0.0783007964,0.1522820741,0.1580559462,-0.083456777,-0.2392239273,0.727088213,0.1221747771,-0.013926208,-0.3829063475,-0.1173573509,-0.5056692362,-0.2212937474,0.3403012455,-0.007171046,-0.0925393403,-0.0244391449,-0.2103015184,-0.0364561193,0.2022960037,0.3431307077,-0.1650477052,0.2029755563,-0.0161112938,0.4196084142,0.5856342316,-0.1346532851,0.1360911876,0.3772717118,-0.1753327996,0.0519081727,0.2391431928,0.1840669513,0.1506536305,-0.1263084859,0.1601605415,0.4892570078,-0.3131014705,0.258397609,0.0568521433,-0.221428588,0.1675420702,0.163731873,0.0658375323,0.0036606842,-0.1034322381,-0.3204090595,0.3190239668,-0.3128894567,0.0569208525,0.0483365171,-0.1053899452,-0.0348551497,-0.1259045154,-0.3479111493,-0.3173058927,0.1093390733,0.0865781754,-0.0670085549,0.1533927917,-0.3714462221,0.4874791801,0.1221940294,0.1755738258,0.2396434397,0.180188939,0.2371768355,0.0878022239,-0.2353841513,-0.2851777971,0.1088109389,-0.1383740157,-0.1427969933,-0.1061125845,0.246758908,0.2857477069,-0.1014980897,-0.021430755,0.133170262,-0.1547442973,-0.1268596351,0.1578128189,-0.0227402691,-0.0729566664,0.3669102192,0.495400548,0.5253913403,0.1801846474,0.1083614379,-0.1301073432,-0.2087201476,-0.1430886388,-0.205413565,-0.1446287185,0.466601342,-0.0668567717,0.3619411588,0.1625888497,-0.0146639319,0.238034457,0.0549203418,-0.0770052522,0.2333343476,0.0734031051,-0.1057646275,-0.0397085734,0.1877189279,-0.4425652027,0.3263513744,0.5805601478,0.1085751727,-0.0362419263,0.1326859742,0.2328303456,-0.010291948,0.118101716,0.134821251,0.1647725701,0.1919315308,0.1233764365,-0.0926784351,-0.014028457,0.0075354669,0.1509769112,0.2267988175,0.297591567,0.0732766688,0.0710349008,-0.191386506,-0.1167141795,-0.3844871223,0.242283985,0.0336244367,0.022234071,-0.075333789,-0.2187243402,-0.0839655325,0.0200216454,0.1564753354,-0.108487092,-0.279620707,0.0375791527,0.2110372335,-0.1723946929,0.1639958471,0.1517231613,0.1216598004,-0.0795259029,-0.0444308482,-0.1731158495,-0.112800248,0.115402922,0.0780870467,0.3604660928,0.2093340456,0.4150375724,0.2297144085,0.101619646,-0.3085978627,-0.3013170362,-0.060573034,-0.2326071113,0.0706917867,-0.0197261609,0.0140115079,-0.2078688145,-0.424069345,0.1349754184,0.1233988032,-0.1107923985,-0.1494641453,0.0558032654,0.0630634502,-0.0742373988,-0.2053208798,-0.51438272,-0.4376353025,0.3669537008,-0.3244447112,0.349766463,-0.1991966963,0.0240951423,-0.0096491147,-0.174112469,0.0837804303,0.1551626623,-0.4178397059,0.1165356115,-0.0766956657,-0.1169132814,-0.3409132957,-0.0333880819,0.0715974271,0.4641480744,-0.2137114704,-0.0585741065,-0.2273505628,0.1493424028,0.1650560796,-0.0818075538,0.1073483229,-0.0158897005,-0.1121382341,0.0774197653,-0.1644622833,0.0516895056,-0.0614806153,-0.2959742248,0.2579024136,0.3612661958,0.1235006452,0.7081286907,0.1143643185,-0.2318828553,-0.0408289395,0.072738111,-0.2688436806,0.2038408667,-0.4948345423,-0.0708803236,-0.1688321829,0.1085114405,0.0931626335,-0.1084734425,-0.419706434,-0.0343335532,0.2127263099,-0.1820169389,-0.1730897278,0.2890797257,-0.30162552,0.3059793115,0.0404462665,-0.2821751833,-0.4290571511,0.0758815706,-0.1535704732,0.0001312793,0.2750862241,-0.1826774031,-0.3942542672,-0.2149473727,-0.4903289974,0.333840996,0.1873455197,0.1052177548,-0.017547071,0.0272646658,-0.0534178577,-0.1533872932,0.5795544982,-0.3685047925,-0.2850637436,0.2537964284,-0.4896868467,-0.0196860638,0.2228054106,-0.0259724092,0.4166381955,0.3337048888,0.4109072089,-0.2376558632,-0.4871218204,0.1558928639,0.021949267,-0.0720125586,0.0266816262,-0.0253530443,-0.0449184813,-0.1176822409,-0.0097690905,-0.0658997446,-0.0280598197,-0.0187596623,-0.1868542582,-0.16463238,0.0536768809,0.3366567492,0.0215823781,-0.1456777006,0.3455347717,0.2921149135,-0.2915527523,0.114172928,-0.0874810591,-0.1014223546,-0.185351789,-0.1297592968,-0.0313217118,-0.0368647389,0.5222084522,0.0930930004,0.1535838544,0.39201051,-0.1096459851,0.4256867766,-0.2488377988,0.1663354039,0.3916359246,0.2579731643,-0.2799405754,-0.3690032363,0.2246362269,0.2787813246,0.0205803476,0.5443371534,-0.2962469161,-0.0203338638,0.2075714469,0.0824856386,0.5658167005,-0.3810481429,0.0208447482,-0.2761297822,0.1997947842,0.4320572615,-0.4796617925,0.1967347562,0.0666372702,-0.3153504133,0.0637310594,-0.0058261212,0.1173130795,0.426453799,-0.1951825172,0.0736682042,0.3178193867,-0.0364035144,-0.0825515091,0.4622345269,0.2771174014,-0.3780421317,-0.09199965,0.1073800772,-0.0745103657,-0.3487041295,0.0642073601,-0.2750181854,0.1603492796,0.3369334042,-0.3123576641,-0.1006103829,-0.0329274312,0.4128578901,-0.2716251314,-0.0356819071,-0.0980781093,0.1445015669,-0.0621918775,-0.0515224561,-0.0887068361,0.0366987437,-0.1406469494,0.3195926249,-0.2093704343,-0.3096981645,0.2303267419,-0.1214029044,-0.1893571466,0.0026277285,0.1380224079,-0.5908483267,-0.13972646,0.0934133455,0.0945591703,0.4548108578,0.1122353449,0.3758336604,-0.3117886484,-0.0755104125,0.1418021023,0.2096755654,-0.1527104825,0.5664060116,0.0669871569,-0.1376767904,0.0543280765,0.0388535522,0.318723321,0.0386620313,0.2706732452,0.1541334689,-0.137809217,-0.2728895843,-0.1065818742,-0.2155002952,0.0848815367,0.2780231833,-0.1531941146,-0.1919222474,0.1020032614,-0.1520504206,-0.1595102698,0.116439335,-0.2900215685,-0.1523669958,-0.1031883657,0.0998684317,-0.1188057587,0.2155810744,-0.0794777051,0.1791784465,-0.0537462384,0.5050323009,-0.3481152058,-0.1349555552,-0.1647222191,0.1981509924,0.3397384286,0.0221185926,-0.0613251179,0.2210672647,0.1781978458,0.1315311939,-0.1826761663,-0.3112728596,0.1795540452,0.0597392023,-0.1320134699,0.0578906462,0.1754211932,0.1522132456,-0.3791618347,-0.3611781001,0.0145178633,-0.090442054,-0.028688794,0.0797054097,0.1471130103,-0.2112824619,0.2349593043,-0.0554059073,0.131070435,-0.0193425175,-0.138555482,0.3766928017,0.1650500447,-0.1514664441,0.152201429,-0.0971617252,0.1188129857,-0.1742065698,0.2069409192,0.0635359138,-0.2430644631,-0.0912083313,0.2783658504,0.3514907062,0.1390746087,0.2302773148,0.0571813546,0.1735034138,-0.1607595235,-0.150310725,0.5617012978,0.0002524624,0.3441422582,-0.0869367346,0.1112551764,0.4852898419,0.1047268137,-0.0686949417,-0.123232305,-0.3425774574,0.0975733325,0.0001649504,0.2361600101,-0.0399695113,-0.1033607572,0.0276408233,0.1377209127,0.1111187711,0.0776181519,0.7436546683,-0.1332692355,0.1466140002,0.2358623892,-0.2185104489,0.1554222107,0.1508917063,0.070659481,0.2849662304,0.1042458266,0.0836238116,0.115759328,-0.5258510709,-0.1622172445,0.6554818153,-0.0517862104,-0.335093081,-0.0595480353,-0.0637189075,-0.0553979725,0.0772393122,0.1890801638,0.1039631292,0.5785467625,0.0631640106,0.423332274,0.021710977,0.1380326897,-0.1756118387,0.3721530735,-0.0519323833,-0.16001755,0.1439989209,-0.2131197006,-0.2313869298,0.3223556578,0.1182821169,0.2578311563,-0.060610842,0.3421157598,-0.2579092681,-0.036384277,-0.2379796952,0.2340898514,-0.0216608383,0.0199266691,0.4125625193,0.1230833307,-0.1453164518,-0.2635038495,-0.0451858863,0.0962260813,-0.4112082422,0.0877639055,0.1337747127,-0.1602257341,0.0140001327,0.0020579726,-0.3977244794,-0.2798145711,0.7091524005,-0.5511259437,-0.0335208885,0.0124388151,0.1236992031,-0.2153753787,0.6456469297,0.0827541649,0.3928978741,0.0056269467,-0.164499104,-0.4179255068,-0.2629401684,-0.0663555488,0.1159659773,0.1231002063,0.4917398393,0.2183323056,0.3176274002,0.0032264588,-0.4619162381,0.0243652258,0.0811447427,-0.4527287781,0.1186487451,-0.3704134822,-0.0622871667,0.1752673984,-0.3342029452,0.2675736547,0.1106448919,0.1272596419,-0.2127434909,0.1328650117,0.4938168228,0.1521319002,0.0125986142,0.0098248115,0.3086353242,-0.3062476516,0.0071036979,-0.1148119792,0.0962170139,-0.1907643974,0.0032972959,-0.04330194,0.3226057589,-0.2178197503,0.0597901605,-0.0545968786,0.1038937494,-0.1082255542,-0.1784220487,0.1569593549,0.0113971569,0.1666012853,0.1770830005,-0.0151793063,0.3487411737,-0.0062885722,0.0953622013,-0.1045332402,-0.0823695138,0.2420099676,-0.4555224478,-0.2231687158,-0.2831244767,0.0669715628,-0.0349163935,0.4219544828,-0.3687782586,-0.3382135332,0.257460624,-0.0874077678,-0.0455556773,0.5557900071,0.0016172295,-0.2452807277,-0.1554080546,-0.2612406909,-0.0094331745,-0.3534398377,-0.1483712643,-0.2303111702]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/426","title":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter","comments":"So `tf.data.Dataset.map()` returns a `ParallelMapDataset` if `num_parallel_calls is not None` [link](https:\/\/github.com\/tensorflow\/tensorflow\/blob\/2b96f3662bd776e277f86997659e61046b56c315\/tensorflow\/python\/data\/ops\/dataset_ops.py#L1623).\r\n\r\nThere, `num_parallel_calls` is turned into a tensor and and fed to `gen_dataset_ops.parallel_map_dataset` where it looks like tensorflow takes over.\r\n\r\nWe could start with something simple like a thread or process pool that `imap`s over some shards.\r\n ","body":"It would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?","comment_length":47,"text":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter\nIt would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?\nSo `tf.data.Dataset.map()` returns a `ParallelMapDataset` if `num_parallel_calls is not None` [link](https:\/\/github.com\/tensorflow\/tensorflow\/blob\/2b96f3662bd776e277f86997659e61046b56c315\/tensorflow\/python\/data\/ops\/dataset_ops.py#L1623).\r\n\r\nThere, `num_parallel_calls` is turned into a tensor and and fed to `gen_dataset_ops.parallel_map_dataset` where it looks like tensorflow takes over.\r\n\r\nWe could start with something simple like a thread or process pool that `imap`s over some shards.\r\n ","embeddings":[-0.3959482312,-0.3589043319,-0.1633280963,-0.1043198109,-0.2803543806,0.1315265745,0.2860785127,0.2982311547,0.0233868193,0.0550605059,-0.1639014632,0.3279609382,-0.0940098166,0.3442825079,-0.1141018569,-0.1108684912,-0.0506920554,0.0559121408,-0.0223453958,0.069045648,-0.0700980723,-0.0351481959,-0.1102686077,-0.2142758369,-0.1099678352,-0.1081891507,-0.0303025674,-0.1212914214,0.232478708,-0.02456652,0.1026776507,0.4613172412,-0.2688800693,0.3424686491,-0.0001032941,0.0418287069,-0.0300085545,0.0868738294,0.0992644206,0.1611691415,-0.2120831609,-0.2810766399,-0.0511213876,-0.2975957096,0.0698213577,0.1315265,0.203703776,-0.2799983621,0.2039439827,-0.1203097701,0.1440793872,0.4225911498,-0.0239606388,0.2986965775,-0.2356478721,0.1569117755,-0.1393171698,-0.1624263227,0.3427735269,-0.2062585652,-0.1268125921,0.4884384871,-0.259321034,0.2763180733,0.123053506,-0.0430888794,0.0226265695,-0.4618017375,-0.1920103133,0.3221647143,-0.1607878059,-0.1975753158,-0.4645855427,-0.3551502824,-0.1647726744,-0.3611134291,-0.1993352026,0.0582988001,-0.2181826681,-0.0167478118,-0.3532605469,-0.2185585499,0.1069770381,-0.1390846819,0.0176804438,0.3166686594,0.3593294621,0.1542654634,0.2229836285,-0.1610965878,-0.1056815088,-0.1764618158,0.3239942789,0.1019217223,-0.7124086022,-0.2153791189,0.0315666348,-0.3484629393,-0.0871070772,-0.0917281434,0.0713716969,0.3659612238,0.0206590053,0.2957697511,-0.0312648565,-0.4160400629,0.0468206592,0.0443537198,0.1093435064,-0.4134148061,-0.1735185087,0.2978520095,0.0914853588,-0.0863815397,-0.2096258402,0.046808634,-0.0589461885,0.2689777017,-0.0146428319,-0.3458239734,0.0569184199,-0.11129006,0.1320361346,-0.0379306227,-0.0666669011,0.293163985,-0.0731025636,-0.2985054255,-0.3626380265,-0.0711502433,-0.1218412369,-0.018844638,-0.0068022716,0.1827823073,0.17651245,0.1731300056,-0.2708663046,0.014701183,0.1303204,0.309343785,0.3430289626,-0.188289091,-0.036269255,0.2379075736,-0.1522753835,-0.1348415762,-0.0187121928,0.1012221724,-0.2487326264,-0.0216226503,-0.368262738,-0.2509817183,0.0032871452,0.1618471444,0.0316624194,-0.1213393211,-0.247723043,0.6351502538,0.1733308434,0.053981252,-0.4437565506,-0.1052965075,-0.6214120984,-0.2067533433,0.34303689,0.0838774443,-0.1477437019,0.0056999214,-0.1197174639,-0.0573705994,0.1628466696,0.3418937922,-0.0894847363,0.0796350613,0.0038429236,0.4780191481,0.3685803711,-0.1520615667,0.0209941082,0.3077547848,-0.2115902454,0.0637930483,0.2098005414,0.1910682321,0.2824667096,-0.205237478,0.1804116964,0.5076910853,-0.2538856268,0.278078258,-0.0517066307,-0.2002614141,-0.024469085,0.3239072561,0.1461585313,-0.0261237565,-0.1242436171,-0.3164177239,0.3325299621,-0.180084914,0.0445960239,0.0815765411,0.0334726349,0.0663525984,-0.0954743475,-0.3216052353,-0.3459466994,0.242272824,0.1438238174,0.0825455338,0.0948243365,-0.3457756042,0.4237269461,0.1098069176,0.1453585923,0.2596757412,0.2118769139,0.1120656729,-0.0073613925,-0.262969166,-0.282695353,0.2408760786,-0.2067618966,-0.1024324,-0.1578761935,0.332416296,0.2523002028,-0.1547695994,-0.1125105619,0.1080850288,-0.0961473659,-0.1481085867,0.1552660167,0.0421345122,0.007130268,0.4032177329,0.3985290229,0.5866167545,0.2142303139,0.1764787287,-0.118756935,-0.2487889677,-0.1447888315,-0.1825559437,-0.0806524828,0.4537775815,0.0177679453,0.427516669,0.1542966962,0.003184967,0.1733975857,0.010593608,-0.0275998954,0.2379271537,0.1078168228,-0.1230462119,-0.0627012849,0.192746073,-0.2868266404,0.3883141875,0.5061119199,0.1174508259,-0.1138854921,0.0805556849,0.2276956588,0.024080649,0.0678068921,0.2511113882,0.2452669889,0.1791780591,0.1480734795,-0.0011040824,-0.0717214048,0.0195030607,0.1289645135,0.2748678923,0.3190974593,-0.0208060239,0.06974262,-0.1868710667,-0.1235901266,-0.4200427234,0.2997809052,0.1124174222,0.0483122058,-0.0575497299,-0.1074941009,0.1269658506,0.0162731484,0.1375830472,-0.0727732629,-0.3041386604,0.0181808211,0.3667221069,-0.0663115978,0.1369047612,0.1925064772,0.0463772044,-0.0803009272,-0.1121405214,-0.1769384146,-0.2440679818,0.2080888748,0.0885316879,0.3739584982,0.1900398135,0.4317305088,0.174772501,-0.0021741062,-0.1536821872,-0.2592749894,-0.0543094091,-0.2863566279,-0.0191690344,-0.0316722021,0.1141080707,-0.222876668,-0.305883348,0.1724511981,0.004610864,-0.0688407272,-0.1139108315,-0.0127072101,0.0893527269,-0.0477353595,-0.250335604,-0.568400383,-0.4092233777,0.3266500235,-0.2324374616,0.3775835633,-0.0547006205,0.0361393988,-0.0018444136,-0.1080201417,-0.0059895418,0.084172383,-0.2796820998,0.1081388742,-0.0342624746,-0.1321190149,-0.2544576824,-0.0246623587,0.0796695128,0.5135455728,-0.1619199365,-0.1321908534,-0.2244547904,0.1193795204,0.1274669319,-0.0757752806,0.1937647313,0.0238464158,-0.1134343892,0.0619438291,-0.0049564792,0.1277588308,-0.0798965469,-0.3843893707,0.2287478149,0.3610131145,0.1952300668,0.7414991856,0.189039588,-0.317825973,-0.0958654135,0.0290842112,-0.2857871652,0.1605235636,-0.4328313172,-0.0814658254,-0.0768269673,0.0474997833,0.046065744,-0.1417331994,-0.4039613605,-0.0671354458,0.1964143813,-0.2161610425,-0.2535507679,0.2536222637,-0.1575149894,0.2934401333,0.0309125036,-0.2456306964,-0.4624833167,0.0370785892,-0.161060065,0.0008600882,0.3101627529,-0.2374068797,-0.303689003,-0.1044949368,-0.5077471137,0.3355196714,0.0779227167,0.1264084131,0.041226469,-0.0216404889,0.0204034634,-0.1583373249,0.6184366345,-0.3220214546,-0.2553034723,0.2162010372,-0.4241093695,-0.1754305363,0.1810816377,-0.0449716002,0.3480903506,0.3188766241,0.4349924624,-0.3314713836,-0.4567175508,0.1255398244,-0.0280895736,-0.0977558643,-0.0610612966,-0.1144474372,-0.0743116736,-0.1976044476,0.0487014912,-0.0044869944,-0.0453450121,-0.0092505431,-0.1380379647,-0.1529382169,0.0189390797,0.2857368588,0.1053525507,-0.1077662855,0.2979786098,0.3880722821,-0.1602469683,-0.0859046429,0.0324031077,-0.0230688602,-0.0851471424,-0.1626372188,-0.05772743,0.0524093509,0.4032557309,0.0532753654,0.1544159949,0.2574467957,-0.1386050731,0.4407037795,-0.2465183288,0.1414550245,0.2674773037,0.2806440294,-0.2167596817,-0.4176130295,0.2283666879,0.2824575901,0.0356752947,0.5328615904,-0.2536222637,0.006878261,0.3022963703,0.1098844707,0.5154480934,-0.485750109,-0.0372979306,-0.2026857585,0.3226899505,0.3930733502,-0.5588086247,0.2324300706,-0.0293800924,-0.3676802814,0.0822344869,-0.0416127294,0.1130400375,0.4261130095,-0.1032335311,0.0577917583,0.3466845155,-0.0036739169,-0.0781667903,0.3455252051,0.2027903646,-0.3903526962,-0.0352292024,0.1311016083,-0.0206062198,-0.2782941759,0.0426179133,-0.1541851163,0.1644086093,0.313302815,-0.3648082614,-0.1199443117,-0.1013828069,0.3564145267,-0.2605820894,0.0154460575,-0.1193085313,0.1340242475,-0.0776553154,-0.0452137813,-0.0682605654,0.0828441232,-0.0709707811,0.2556941211,-0.0639365017,-0.3214081526,0.2498639375,-0.1251499504,-0.2517109811,-0.0653515682,0.1504544914,-0.4629350603,-0.157273531,0.0920547023,0.1018479541,0.4866551757,0.1113803163,0.4011411667,-0.3141964078,-0.0329846852,0.1462135166,0.1290736943,-0.1824346185,0.6450843215,-0.1432973593,-0.1451507807,0.0570268966,0.1164400205,0.3842929304,-0.0462187678,0.2132510692,0.1516088396,-0.1438690126,-0.2672668397,-0.140086934,-0.1376986057,0.071660459,0.213972196,-0.115877755,-0.1828244776,0.1562100649,-0.2269119769,-0.1832347214,0.1615784019,-0.3105659187,-0.2790353596,-0.1183370724,0.1062592864,-0.0522751287,0.2375908196,-0.0872713029,0.1804929376,-0.0868488923,0.4261001647,-0.3705820441,-0.0214438885,-0.1634512693,0.165109843,0.3253673315,-0.024627056,0.0152844228,0.14704597,0.1758812368,0.1975680739,-0.1980699748,-0.3065065444,0.1562500298,0.047756955,-0.122880511,0.0148316165,0.180935666,0.0875284523,-0.44227916,-0.3774246871,0.1015636176,-0.0086371358,-0.0991132036,0.0038007915,0.3627699912,-0.0334007107,0.1455197781,-0.0058677495,0.0634088442,-0.0765857249,-0.1244179234,0.3410062194,0.1938910037,-0.1175997406,0.1448974013,-0.1172141135,0.160001874,-0.1159836575,0.2434903681,0.0448683053,-0.2600496709,-0.09802939,0.4227248728,0.3529421389,0.2267611176,0.0967410356,0.0720928237,0.2289396822,-0.1738938838,-0.1163844392,0.6512266397,0.0166976582,0.4703913331,-0.0157639347,0.08511623,0.5125299692,-0.0365244932,0.0038920776,0.0008237657,-0.329749465,0.0768737867,-0.0688797757,0.2154845893,-0.1073746383,-0.0540698208,-0.0920072794,0.1457467526,0.1421990246,-0.0178029053,0.845295608,-0.1318363547,0.05420449,0.1747168154,-0.2661342323,0.1045222208,0.124260135,0.0783950686,0.3264674544,0.0545977913,0.0771180913,0.0885591879,-0.4675466716,-0.1652884334,0.6315925121,-0.0764263049,-0.2411170006,-0.0466094315,-0.0262712296,-0.0856410563,-0.0103626559,0.2047714442,0.0620877258,0.5979956985,0.0851118043,0.4313570559,0.0676743016,-0.0179103725,-0.0940605104,0.3815837502,-0.0917866901,-0.1464930624,0.1808088571,-0.2079187632,-0.2164738029,0.2328568995,0.1020289287,0.1816219985,-0.0667531639,0.453627497,-0.2325475067,-0.0095973248,-0.2429827154,0.2401319891,-0.1555421352,0.119610779,0.3448847532,0.1126712039,-0.1828779429,-0.3477285206,-0.0873991102,0.0183531884,-0.2619009912,0.1596794128,0.0909976512,-0.2678846121,0.0224009845,0.0659282655,-0.3229214251,-0.2946972251,0.5998606086,-0.5497958064,-0.0602735914,0.0782752261,0.1313943118,-0.2361160666,0.6309201717,0.0459252298,0.3957271874,-0.0325090773,-0.0984342843,-0.4964119494,-0.141385287,-0.1641002744,0.0614777282,0.0978117734,0.4516782761,0.2257592082,0.3760368228,0.0037352459,-0.3737081289,0.0093107121,0.1776870191,-0.4579306245,0.1084488556,-0.4192533195,-0.2052963376,0.1998188645,-0.3315633833,0.2484562397,0.0671998709,0.1189687327,-0.1650651097,0.1183110327,0.5072969794,0.007178165,0.0685008094,0.0069124326,0.3343527317,-0.249589324,0.0270455088,-0.1677881926,0.1200655401,-0.2708984911,0.011997886,-0.0357053392,0.3253619373,-0.1558250189,-0.1059399545,-0.0696618855,-0.0041537327,-0.1198502108,-0.2138369679,0.0823436007,0.0179493967,0.1702774763,0.276676327,-0.1018977612,0.4429580271,0.0640806779,0.1211883128,-0.1146948636,-0.1287365258,0.3002875149,-0.4690983593,-0.2409884185,-0.3412491977,0.0718250498,-0.1741064191,0.3793014586,-0.3779056966,-0.2622273266,0.2299674749,-0.0829022601,-0.0809898823,0.5007374883,0.1311314106,-0.3234911859,-0.137752369,-0.2630752027,-0.0466941744,-0.3111960888,-0.087440744,-0.1657190919]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/426","title":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter","comments":"Multiprocessing was added in #552 . You can set the number of processes with `.map(..., num_proc=...)`. It also works for `filter`\r\n\r\nClosing this one, but feel free to reo-open if you have other questions","body":"It would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?","comment_length":34,"text":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter\nIt would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?\nMultiprocessing was added in #552 . You can set the number of processes with `.map(..., num_proc=...)`. It also works for `filter`\r\n\r\nClosing this one, but feel free to reo-open if you have other questions","embeddings":[-0.3118088543,-0.2080799341,-0.1779322773,-0.173046574,-0.343650192,-0.1113172323,0.197154507,0.2835273743,0.039257545,0.0792978406,-0.0767444968,0.3925559521,-0.0509052128,0.3508171141,-0.0408180468,-0.1657482535,0.0791809261,0.1307990849,-0.1149892434,0.0648014918,-0.0683189854,0.0558378361,-0.1673122197,-0.0710989088,-0.1849099994,-0.1321129352,-0.1406542808,-0.0166990142,0.0662526339,-0.241936028,0.0318653993,0.4328874052,-0.1755050123,0.0703963786,-0.0000980988,-0.2013363689,0.2006985694,0.1778134704,-0.1014168561,0.3299798667,-0.2782140374,-0.4332962632,0.0383698978,-0.2476124614,0.1491229236,0.204867363,0.2674977779,-0.2454677522,0.1922703683,-0.1812831014,0.2317156792,0.2043158412,-0.1151811853,0.2320506722,-0.1050489247,0.1807965189,-0.0922977626,-0.2746455073,0.3728884757,-0.1693915129,-0.3322257996,0.5420419574,-0.2429718375,0.15918006,0.1380308121,-0.0897840783,0.1310691237,-0.3727107644,-0.1179900616,0.259978354,-0.053914126,-0.1353920847,-0.3391513526,-0.2734780014,-0.2056402862,-0.4567804039,-0.2121410221,0.0896802992,-0.2025435865,-0.1186684445,-0.3784823716,-0.2798651457,0.2643017173,-0.1941073388,0.2318017036,0.3839837313,0.2291881889,0.0688471273,0.1787993908,-0.2928026617,-0.13450858,-0.2245154828,0.0903647915,0.0094030937,-0.5751150846,-0.1305833161,0.0920881927,0.0742391646,0.0772548765,0.0394538976,0.0862812772,0.341940254,0.1756995767,0.2525441945,0.0341194235,-0.342007488,0.2699638009,-0.0482739359,0.0975597575,-0.267749846,-0.2028669715,0.3006490469,0.1756908298,-0.1166880876,-0.1895580888,0.0690390393,-0.0817728937,0.116187349,0.0531156473,-0.380705148,0.1933645904,-0.1785689592,0.1173214838,0.1244455352,0.0175260175,0.2864505649,-0.2257845402,-0.2121169418,-0.2677025497,-0.202941522,-0.1525212973,0.0074563441,-0.0609151907,0.0846334323,0.1458331794,0.2318814993,-0.0329232365,0.0872344673,0.068052955,0.3604520261,0.2550831437,-0.0851559415,0.0856151953,0.1747232378,-0.0497545153,-0.1048840806,-0.1860293001,0.005004378,-0.3372870088,-0.0475950055,-0.2184208333,-0.0954538286,-0.0260660686,0.2172264457,0.1290778071,-0.0312367529,-0.1767545342,0.5951275826,0.0113264984,-0.1451157928,-0.3826845586,-0.0607046299,-0.3916879296,-0.2338740677,0.2800220251,0.1081847921,0.0472647436,0.0406413898,-0.2878512144,-0.028379865,0.3236938417,0.2599790394,-0.1431951076,-0.0149886422,-0.0436384119,0.5232228041,0.516982317,-0.0268271007,0.0530014634,0.4043429792,-0.1324664205,-0.0553305671,0.2862443626,0.2277933508,0.2258700132,-0.0635008514,0.0482455343,0.5034412742,-0.2482093573,0.2910609841,-0.0909451097,-0.1039722562,0.0981245264,0.1033186689,0.0272240117,0.0312673934,0.027979387,-0.2089791298,0.4157721698,-0.2304131687,0.1709154844,0.0214936901,-0.0577129014,-0.1529524177,-0.1221983582,-0.301199168,-0.2843013406,0.0986698642,-0.1794967651,-0.0079391506,0.1824814975,-0.372618705,0.3392186165,0.1165099666,0.0980499312,0.1306410283,0.2733260691,0.3272752166,0.0731168762,-0.3091557324,-0.1573940367,0.1696812212,0.0174951237,-0.1916555762,-0.1207503155,0.2065017968,0.0937884152,-0.1517454386,0.1674118936,0.1059089452,0.0496891364,-0.0346130058,0.1924587339,0.0169885047,-0.0459609814,0.3286037445,0.6206994653,0.2523434162,0.1050925478,0.1744488627,-0.0091989832,-0.1313982755,-0.0103494348,-0.1294962019,-0.1501306593,0.4955796003,0.0422686674,0.3763100803,0.2261741906,0.077959992,0.2708156109,-0.0474646911,-0.0896143615,0.1630669683,0.1636178493,-0.0212114248,-0.1145576835,0.2652320862,-0.4278348684,0.3981330097,0.4963065982,0.1293104142,-0.042589087,0.0880843028,0.2510956526,-0.1494485438,0.188910231,0.1442229748,0.2122191191,0.296241492,0.1868578345,-0.1414690614,-0.1081625968,-0.0351196714,0.1874668002,0.1344648153,0.2369910032,0.0746929944,0.1096189916,-0.2849855423,-0.2190054357,-0.3890277743,0.165653348,-0.0617084019,0.0815394148,-0.0809906572,-0.2183191329,-0.1557605267,-0.0316579677,0.2191955447,-0.2191598564,-0.3372049332,0.1016316488,0.1640125066,-0.2236168087,0.1987195611,0.1181520671,0.1623241007,-0.18387945,0.1328248978,-0.1803711057,-0.117669858,-0.0076289084,0.1361314505,0.3037855923,0.378546983,0.4664989412,0.1868599951,0.0707414523,-0.3563061059,-0.1830935925,-0.059596628,-0.152449578,0.0921869054,-0.0250975303,-0.0386617333,-0.2330762595,-0.3906508684,0.1592654288,0.0492815264,-0.1521445066,-0.1326260418,0.0165969823,-0.0352673121,-0.0512725487,-0.2436923236,-0.4269717038,-0.4275345802,0.3586696684,-0.1808321327,0.360144496,-0.3318041861,-0.074373059,-0.0303600132,-0.2477468997,0.0795424581,0.0881257281,-0.2829639018,0.0857495293,-0.0766464919,-0.1964254677,-0.292612195,-0.0988129899,0.0426921137,0.3463363647,-0.084776625,-0.002807508,-0.2195988744,0.153880164,0.2612432241,-0.0588142313,0.1657884419,0.0510349087,-0.2215289921,0.1123751104,-0.2412109524,0.1503892243,-0.1320764273,-0.1409629136,0.1440397799,0.303394109,-0.0152792158,0.531575501,0.1151196361,-0.1121071875,0.0215538722,0.1098569855,-0.1857664138,0.1743285954,-0.4403367937,0.0002036659,-0.1181213185,0.0327636972,0.1311004758,-0.1172965243,-0.5028033257,-0.0272687674,0.0582268201,-0.198155567,-0.17052266,0.201812312,-0.1989325136,0.2783462107,0.0801807344,-0.3016012311,-0.3575441539,0.0743290558,-0.268096447,-0.0270331912,0.2139933556,-0.1887508482,-0.4772756398,-0.1842854321,-0.5468443632,0.3180502951,0.1374520659,0.1132429466,-0.0543640144,-0.0037843988,-0.0591738448,-0.1751533151,0.4736835957,-0.3019973934,-0.1694518477,0.2956008911,-0.3368089795,0.0315157026,0.2125362456,-0.0706273541,0.4595478475,0.4683873057,0.2617396712,-0.332392782,-0.3635452688,0.1668351442,0.0684038326,-0.0942510366,0.0202641804,-0.1293440163,-0.0694131628,-0.2078354508,0.1128198728,-0.0865995437,0.1333026886,-0.036421001,-0.1267885864,-0.1307300627,0.0079304511,0.3029812872,0.0826520845,-0.0328120813,0.2658084929,0.2377577722,-0.3488393128,0.2270439565,-0.1522962749,-0.0199716333,-0.1843560487,0.1344465911,-0.0626533255,-0.0692750216,0.5562533736,0.2502156496,0.1834810674,0.3105381727,-0.0753910169,0.4139234722,-0.1532603502,0.0114478823,0.3346478939,0.1599075794,-0.2039086223,-0.2229419798,0.2769015729,0.2694518566,-0.1253663152,0.3823762834,-0.2421091199,-0.0711113736,0.2561421692,0.1250007004,0.6941605806,-0.2547995448,0.0451462641,-0.2514950037,0.1451072097,0.3248785436,-0.2804261148,0.219497323,-0.0586193204,-0.1863626093,0.1432299316,0.0192233473,0.190998897,0.4685852528,-0.2656114399,0.1729698926,0.2839288116,-0.1742915958,-0.078282848,0.4526757598,0.1473712772,-0.4094931781,-0.1683617085,0.1609100401,-0.0478139222,-0.3570655286,0.0380352512,-0.2977586985,0.1869595498,0.1704109311,-0.3053102791,-0.0162862614,0.1349386424,0.3702848256,-0.2635102868,-0.010328521,-0.2282410115,0.0804496855,-0.1640238911,0.0282278601,0.0138420733,0.1121842787,-0.1329812109,0.2483469248,-0.1728421748,-0.2532656491,0.1958269328,-0.2532755435,-0.1808156818,-0.0334777236,0.0669211373,-0.52875036,-0.1446629912,0.1538774222,0.0937529504,0.3899872303,0.0303077791,0.3620783389,-0.4090334773,-0.1828260571,0.1932815015,0.2313537747,-0.2200433314,0.541182816,0.1586213559,-0.1013461649,-0.0034013209,0.083224535,0.289155215,0.0061691906,0.2480873615,0.0655295998,-0.206912756,-0.3257484138,-0.2069140226,-0.2269089818,0.0533736944,0.3017608225,-0.1378846616,-0.2142485678,0.0612096414,-0.1285809129,-0.1259346455,0.1365891397,-0.2172750831,-0.1887904853,-0.1422332227,0.0968129411,-0.2331408262,0.1975513101,-0.0019942932,0.0519076996,0.0366596207,0.4807837903,-0.434433192,-0.1275336891,-0.3187756836,0.2330362946,0.27375108,-0.0472055525,-0.1861439496,0.2489390671,0.2355581969,0.1225078106,-0.1975656152,-0.352914691,0.2140961438,0.0386997387,-0.1522496492,0.068836242,0.0761160627,0.1387800276,-0.3181966543,-0.3029017746,-0.1315650344,-0.1898017973,-0.0796306804,0.0764660463,0.2217507064,-0.2593083084,0.335077703,0.0214413907,0.084064886,-0.1367718279,-0.2212947607,0.3479637206,0.1634105593,-0.2223724276,0.2342487872,-0.0871866941,0.0017208443,-0.1312724799,0.2663226426,0.0407943316,-0.2608653009,-0.0193500891,0.3714659512,0.3568589687,0.2390728891,0.1567325294,0.0942208692,0.2489311695,-0.3249039352,-0.1693696827,0.5596210361,-0.0339076668,0.3383587897,-0.1567271799,0.2189144045,0.2402933687,0.0832248628,0.0472283959,-0.0850991979,-0.3423672318,-0.0223668367,-0.0906853601,0.3015961051,-0.0777262673,0.1149758026,0.122780107,0.1877332777,0.1768874675,-0.016962314,0.7731490135,-0.053665299,0.2881838083,0.1529935747,-0.129353568,0.1669362932,0.1924277246,0.0711640939,0.1993346363,0.1327951699,0.1543232948,0.0372755304,-0.4504743218,-0.2149816304,0.682502985,-0.1505435407,-0.3235414326,-0.0915462598,-0.1169138402,-0.1602390707,0.0944652036,0.1278132349,0.1765397042,0.6568735242,0.015062361,0.3660641015,-0.0714492947,0.1660767496,-0.1578214318,0.3040832579,-0.2052638233,-0.2164439559,0.1420260519,-0.1892103255,-0.065516457,0.3563248813,0.2488278151,0.284822762,-0.1441094428,0.2726025283,-0.2956683934,-0.0594364516,-0.378901422,0.248614043,0.0145971095,0.0457059331,0.4520916641,0.1537104547,-0.2073938996,-0.2555698156,0.0529309139,0.0108491573,-0.2308036238,0.1999276727,0.1541866511,-0.1907088012,-0.061173033,0.0579442568,-0.4901557863,-0.2592764497,0.549092412,-0.4322645366,-0.0196171701,-0.1755620092,0.1528597623,-0.2001218349,0.6736976504,0.0027888345,0.3433564007,-0.0264597349,-0.2550089359,-0.5719377995,-0.3019400835,-0.090059109,-0.056381695,0.1824566871,0.5611234307,0.1624627858,0.2724913657,-0.1371912956,-0.2531253099,-0.0721479356,0.0583383627,-0.4192561507,0.2377564311,-0.1238056719,-0.0789516941,0.1540430933,-0.3220898509,0.3157340586,0.0160882864,0.2028989196,-0.0945673659,0.0261441413,0.4906813502,0.1940528303,-0.1569813788,0.14602606,0.3112100661,-0.3373529315,-0.0026023695,-0.1494635195,0.1065343767,-0.166220963,-0.0949635804,0.0439932793,0.3311820626,-0.1619597077,0.2568047047,-0.0441048853,0.2164625674,-0.2096709162,-0.0430671908,0.1807123572,0.0299117081,0.1488728076,0.2463068068,-0.1401857138,0.2737384737,-0.0078091007,0.0337414481,-0.1430571377,-0.2125964463,0.2409758717,-0.3951701522,-0.2428909689,-0.3473039567,0.219029665,0.0408531949,0.3273250759,-0.2465286106,-0.3446032703,0.2122892439,-0.0348665677,-0.019113807,0.4801186621,-0.1346850395,-0.2861377299,-0.1634634435,-0.2962388694,0.0462740511,-0.2897840142,-0.1670204997,-0.2816137969]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/426","title":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter","comments":"@lhoestq Great feature implemented! Do you have plans to add it to official tutorials [Processing data in a Dataset](https:\/\/huggingface.co\/docs\/datasets\/processing.html?highlight=save#augmenting-the-dataset)? It took me sometime to find this parallel processing api.","body":"It would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?","comment_length":29,"text":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter\nIt would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?\n@lhoestq Great feature implemented! Do you have plans to add it to official tutorials [Processing data in a Dataset](https:\/\/huggingface.co\/docs\/datasets\/processing.html?highlight=save#augmenting-the-dataset)? It took me sometime to find this parallel processing api.","embeddings":[-0.2608761489,-0.4052204192,-0.1671193838,0.0242641233,-0.347699672,0.1083791703,0.0958692953,0.1001096219,0.009775402,0.0779817924,-0.1297877133,0.2687852681,0.0288041253,0.5679566264,0.0352190323,-0.3084892929,0.0757969543,0.1260470152,-0.0477574356,0.197243616,0.0197631307,0.1598396897,-0.2788158059,-0.1252332777,-0.2452908158,-0.1634081453,-0.1911026686,-0.0196516346,-0.0594976693,-0.1503470093,0.0461669154,0.4871737361,-0.231605649,0.2361281961,-0.0001079078,-0.2029749304,0.1503316909,0.229871884,-0.0537671708,0.2007020563,-0.1934657097,-0.4777649045,-0.0147053944,-0.0763369575,0.0592207387,0.1195816919,0.1778105795,-0.2691320777,0.1904294044,-0.3726128042,0.1244619191,0.3582191467,-0.054527048,0.2704071105,-0.2098025978,0.4645996392,-0.0739304945,-0.1920111775,0.4698049128,-0.1441965103,-0.1906628162,0.5513617992,-0.2167661786,0.1467214376,0.2011539787,-0.0467660762,-0.1824139059,-0.4495758414,-0.1506305039,0.3538194895,-0.0240541082,-0.1382976919,-0.5037174225,-0.4508947432,-0.1339050084,-0.4086400867,-0.1706858724,0.1109529361,-0.1954035014,0.0309994668,-0.4884880483,-0.3645783663,0.2795680463,-0.1878647953,0.2982876897,0.2111876905,0.0041947966,0.0964184031,0.2584792674,-0.193264544,-0.2565283775,-0.2772119343,0.1194253042,0.0337730274,-0.6784651875,-0.0603897683,-0.0065606362,0.1300600022,0.1923829019,0.1290129721,0.1012567431,0.3147673905,-0.0229618046,0.2253417224,0.1074023172,-0.3512957096,0.1750337929,-0.1152893603,0.1311005801,-0.190440014,-0.0298142545,0.2025753856,0.2138280272,-0.0518173426,-0.4135623574,0.0292811263,0.0277363043,0.1004153416,0.0746325105,-0.3755646944,0.185649246,-0.2991416752,0.1410411149,0.0419949554,-0.0095862448,0.3397317529,-0.1418078989,-0.1332836449,-0.2340057939,-0.2382114083,-0.0971704721,-0.051624734,-0.0331980363,0.2781025767,0.1205001622,0.0442820638,-0.0506823249,0.1477277428,0.1439828128,0.2672173381,0.2075041085,-0.1242554262,0.0695268214,0.1833892465,-0.0726915076,-0.0383959524,-0.1169660687,0.222246632,-0.2571871281,-0.0959708989,-0.2205808759,-0.2075158358,-0.1869576871,0.0773487091,0.1048262939,-0.1092598736,-0.478769958,0.7779146433,0.0713087916,-0.0115013355,-0.2945772409,0.0227482058,-0.5353251696,-0.1817254275,0.3416647911,0.1284206361,-0.0485664867,-0.0365401208,-0.2228354365,-0.1078438088,0.240354076,0.4414246082,-0.2017016411,0.129213199,-0.1785436422,0.4836609364,0.5142174363,-0.1295250207,0.0830005407,0.374253124,-0.1477963775,0.0708050653,0.3142757714,0.1900290549,0.2082633078,0.0276135541,0.1444729567,0.5401500463,-0.198023513,0.2230969518,0.0059556887,-0.1897812635,0.2261707634,0.1623000354,-0.1524645388,0.0858475268,-0.0761108026,-0.5189324617,0.4866414666,-0.2974569798,0.1562602669,0.0589025021,-0.1313957125,0.0926597267,-0.0669747666,-0.3271737397,-0.3598528206,0.0688545182,-0.0417775176,-0.0491082408,0.0830473527,-0.3939555883,0.4155661464,0.098172076,0.0268570874,0.0086587686,0.1069580168,0.1812459826,0.1519801766,-0.2360940427,-0.3061080873,0.3358493447,-0.0799676403,-0.0123659791,-0.3526895046,0.2138335556,0.2508554459,-0.0553257056,0.0419051088,0.1212062687,-0.0514611639,-0.1264780462,0.1427656263,-0.0031845146,-0.0493048243,0.255017966,0.6550810337,0.2654030621,0.2627613246,0.0785564929,0.0230953749,-0.2753276527,0.0382473506,-0.2153047323,-0.1683977544,0.554084897,-0.0410236605,0.4023038745,0.0917556733,-0.0131629966,0.2349601239,0.073301129,-0.0701760054,0.2105982751,0.1526693106,-0.0811327845,0.0270831119,0.1431891918,-0.4629768431,0.3415274024,0.4402229786,0.005871688,-0.2078667134,0.0840375572,0.1872197688,-0.1110775024,0.1967876256,0.023821231,0.1881490797,0.1884752512,0.0503348112,-0.0454424955,0.1028473601,-0.0663725734,0.1176738515,0.3205150068,0.2168473303,-0.0265257265,0.1670744419,-0.2363259196,-0.282782346,-0.2901486754,0.1736860722,-0.1077105999,0.0196602978,-0.0615283288,-0.1852201074,-0.1912414283,-0.0461944677,0.0478870608,-0.3056801856,-0.304176867,-0.0812803283,0.3092374504,-0.0518721119,0.0761979371,0.2212892175,0.2130075097,-0.1976543516,-0.0081336992,-0.2818010747,0.012489561,0.1285527945,0.031392388,0.3022121191,0.3320901394,0.4555438161,0.1232950687,0.0910328701,-0.4455685318,-0.2040769607,-0.0179644022,-0.2767621279,0.1366851181,0.1226217151,0.1201799288,-0.1607810706,-0.2790879011,0.1972416043,0.1088804454,-0.212207824,-0.2680939436,0.0608802177,-0.0249814037,-0.0241125487,-0.1490212828,-0.4795736969,-0.3343916237,0.5942107439,-0.2876599729,0.2276882678,-0.3308070898,-0.1193307415,-0.0027465667,-0.350494504,0.0665277094,0.0190508571,-0.4718952477,0.020882234,0.0118575813,-0.1137490943,-0.249085173,-0.0198605601,-0.0645403117,0.2378332466,-0.0836322829,-0.1006002724,-0.358882457,0.2384422123,0.212108314,0.0981801376,0.2778333127,-0.0796130672,-0.0705577731,0.1338597387,-0.4800581038,0.0397935398,-0.0009143428,-0.1605335027,0.2632178068,0.3609154224,0.0008576288,0.7097066045,0.2304415405,-0.047403343,0.1350751817,0.1280041486,-0.1666100621,0.0568133518,-0.5510403514,-0.0351173393,-0.1622912437,0.0703056157,0.0248753652,-0.0346966907,-0.2492029369,-0.0900886059,0.0162702352,-0.3376843333,-0.0910864845,0.0867475867,-0.047154434,0.2308442444,0.0760264248,-0.3100879788,-0.2607222497,0.0541560985,-0.2018708885,0.1385470629,0.3485731781,-0.1705937535,-0.5018056631,-0.1323041469,-0.6374371648,0.4433451295,0.2480612844,0.0398766063,0.0086298501,0.0663429126,0.1311942339,-0.1964478791,0.7243824601,-0.2557786107,-0.3600910008,0.2271154523,-0.5849372149,0.087229833,0.197490111,0.0706140399,0.4089244008,0.4323523641,0.4398611188,-0.3128746748,-0.4502788484,0.2078810632,0.1022503749,-0.0641266629,-0.0108836591,-0.1586316973,-0.0743902549,-0.2673548758,0.1051842421,-0.042994231,0.0770496354,0.0950401649,-0.0455875136,-0.0707234144,0.0183444582,0.1913141161,-0.0161735471,-0.0698352829,0.2421740592,0.2998712063,-0.0922838748,0.094618924,-0.0691815466,0.1599670947,-0.1237234473,-0.1960050762,-0.0725320727,-0.0338151082,0.6675156355,0.2453345954,0.2298676074,0.3670384288,0.0978974029,0.3856647313,-0.4025575221,0.0253545083,0.436771363,0.2292354703,-0.3920322061,-0.3054986596,0.3532458246,0.305529654,-0.0169571955,0.5018374324,-0.0144236321,0.0041289781,0.1329773217,-0.0142768268,0.7471269369,-0.3258095086,0.1332370937,-0.3017090261,0.0463985167,0.3842152655,-0.3119762242,0.2014098465,-0.1073825136,-0.2371866405,0.0442393534,0.0127994251,0.3001088202,0.264727205,-0.1261518449,0.3128532469,0.266926229,-0.0907428414,-0.1210808158,0.4641560316,0.1661953926,-0.440836221,-0.2453465611,0.0730640367,-0.0293262862,-0.1736722589,0.0485472158,-0.2938822508,0.2022453696,0.1478758305,-0.3816892803,-0.0909404531,-0.0811213553,0.2681492269,-0.1340636909,-0.1242271587,-0.0441485085,0.0612482131,-0.0484496206,-0.0556651093,-0.056890931,0.0340337753,-0.247452572,0.2347683311,-0.2475908697,-0.256188482,0.1087354049,-0.1369177252,-0.1850027442,-0.1227851585,0.1235465556,-0.5898906589,-0.2854607105,0.2059534341,0.0049389866,0.2578456998,-0.0189393722,0.3219513297,-0.1504007727,-0.1649798602,0.0669941902,0.3953622282,-0.1820902377,0.421240449,0.1083273441,-0.1037901491,-0.046499934,0.1154733077,0.1658522785,-0.1708625406,0.182062909,0.1965959966,-0.1834616065,-0.2009715736,-0.1537409872,-0.1337397844,0.0466590635,0.2491383553,-0.0046175867,-0.0908899978,0.00936729,0.0312433578,-0.1258030832,-0.0645911843,-0.1487535536,-0.3159017563,-0.1274642199,0.1000357345,-0.2690615654,0.1380461007,0.0921475217,0.0871023983,0.0886880159,0.5282360315,-0.2911816835,0.0323012471,-0.0600161292,0.1311891973,0.4815721512,-0.0207298473,-0.028610982,0.2252863348,0.1107013449,0.1718196273,-0.2150518447,-0.2530894279,0.2104914933,0.0922301337,-0.0867487714,0.0380463302,0.0912616476,0.1559923887,-0.2978252172,-0.356559962,0.0225431751,-0.0821360573,0.0193822794,0.0203433223,0.1683955193,-0.2722152174,0.1822599173,0.032092981,0.1487873644,-0.0859450325,-0.1620506048,0.250608474,0.0997243524,-0.1601608992,0.1833009571,-0.0429135635,0.3144593835,-0.2085178643,0.2027050406,-0.0107516944,-0.2946521938,-0.0723331198,0.4625986814,0.337759614,0.2870204151,0.2620433569,0.077076152,0.1131526977,-0.2261070758,-0.1419248432,0.6985279918,0.0277234074,0.2345365733,-0.028705148,0.2569651008,0.5741378665,0.0210928153,-0.0488772988,-0.0279527474,-0.2640927136,0.0817643553,-0.2153850049,0.2547769845,-0.0206003301,0.2107731551,0.2571157515,0.1407174617,0.133398205,-0.0400635451,0.7664589882,-0.1230493486,0.2914153636,0.1756437272,-0.3377158046,0.23824431,0.1596973538,0.1308978945,0.1412181109,0.0603822693,0.3576036692,-0.0241964422,-0.5762885809,-0.2208371311,0.7770311236,-0.0992091671,-0.2864766121,0.1479812562,-0.0764127299,-0.0093325134,0.1394117624,0.1587350965,0.0367024839,0.8322872519,-0.1424617767,0.3206216991,-0.1860070378,0.0500833169,-0.0696892962,0.3375319839,-0.0912108645,-0.1996356547,0.0603376739,-0.1561439931,-0.1357198954,0.4773379266,0.2478214055,0.1339818984,-0.0934017524,0.2575248182,-0.4307624102,0.0304976106,-0.1887446791,0.3448911905,-0.1604806781,0.0238776729,0.4507909119,0.0561622046,-0.1108668968,-0.2868130505,-0.0931768268,0.1409197897,-0.4509787261,0.1636558622,-0.1376704276,-0.106495738,-0.1212155446,0.0233027022,-0.425539583,-0.17504251,0.6109216809,-0.5048497915,-0.1179581136,-0.0576430969,0.0909234062,-0.1598988324,0.5806363225,0.3098267019,0.2996732295,-0.0502881967,-0.1927755475,-0.5379379988,-0.4539735317,-0.0166788567,0.0847620219,0.1283781528,0.4404751956,0.2364767939,0.3570866883,0.0451309681,-0.3217486441,0.0870716646,0.0873020664,-0.3887600899,0.2369842976,-0.1745332927,-0.1542673856,0.1626605242,-0.2101679742,0.331918776,0.0708286315,0.0122049665,-0.1752081811,0.2532280982,0.2549224198,-0.0906015262,-0.1170517653,0.2109996229,0.3174999058,-0.2899769843,-0.0238898713,-0.0814292058,0.0624877997,-0.1937701553,0.0704773813,-0.0215172227,0.4209704995,-0.2175205499,0.1094511151,-0.0596671142,0.2623104453,-0.1140568554,-0.092353873,0.0659225807,0.1022785529,0.0451195538,0.128989175,-0.0045115841,0.3254570663,0.0358528011,0.0831916034,-0.1544280052,-0.0935174674,0.2818125784,-0.4838246703,-0.2623060048,-0.3023208976,0.0809223801,0.1815183759,0.2234048694,-0.5047796965,-0.2906189859,0.1853735149,0.0037674408,-0.0025371495,0.6374949217,-0.1342746019,-0.3422393203,-0.1719223112,-0.1800536066,0.0303835385,-0.309104532,-0.0335459374,-0.1872974485]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/426","title":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter","comments":"Thanks for the heads up !\r\n\r\nI just added a paragraph about multiprocessing:\r\nhttps:\/\/huggingface.co\/docs\/datasets\/master\/processing.html#multiprocessing","body":"It would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?","comment_length":14,"text":"[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter\nIt would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process\/thread\/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?\nThanks for the heads up !\r\n\r\nI just added a paragraph about multiprocessing:\r\nhttps:\/\/huggingface.co\/docs\/datasets\/master\/processing.html#multiprocessing","embeddings":[-0.2417271137,-0.5337124467,-0.0713957474,0.0201531947,-0.3602772355,0.0835091174,0.1235489175,0.115074195,0.0433269702,0.0259437133,-0.1680210084,0.1926009208,-0.0400226749,0.5575112104,0.0937455148,-0.3077604473,0.1391180456,0.0928630233,-0.0549302734,0.1721002907,0.0445952192,0.1536127925,-0.1835022569,-0.0515167937,-0.2325306237,-0.0616339855,-0.1917110831,-0.0409953073,-0.0183580555,-0.1604605168,0.0502604805,0.4818174541,-0.2731352448,0.1951869428,-0.0001166622,-0.1930906773,0.1066058576,0.2355331779,-0.0236794315,0.2996551096,-0.1795283258,-0.4571345448,0.0714368299,-0.0588608086,0.0939774811,0.2544654012,0.2036321163,-0.2047711462,0.1895712018,-0.3378889859,0.0481142998,0.3983403146,-0.128033787,0.2935442328,-0.1875948608,0.4972862303,-0.0639680102,-0.1736515462,0.4167754352,-0.2029202431,-0.2421845198,0.5340806246,-0.2394892275,0.1422058642,0.2418271303,-0.0321054868,-0.0709465072,-0.3804225922,-0.1455935687,0.3719437122,-0.1421710253,-0.1632490456,-0.5388797522,-0.483674258,-0.0996649638,-0.4346542358,-0.1147468239,0.0572927147,-0.195689097,-0.0003932698,-0.5786070824,-0.3901755214,0.3777685463,-0.1943767965,0.216315344,0.2805151045,0.0817480311,0.1199067235,0.3782761693,-0.2899057269,-0.3368574977,-0.2810436487,0.1576242596,0.036093127,-0.5981835127,-0.0697461888,0.0463497937,0.2444093674,0.1914456785,0.0541019179,-0.0022180905,0.2460804433,-0.0006709382,0.157562077,0.0815409049,-0.3408043087,0.2125879228,-0.1072133854,0.1513805538,-0.0976504385,-0.1244120225,0.2404485792,0.2550765276,-0.0951479822,-0.3295258284,0.043772921,0.0132017378,0.0739045292,0.0947799012,-0.4609274566,0.2215867788,-0.3017231524,0.1947869807,0.0818676725,-0.0033535631,0.3359076679,-0.1164365336,-0.187954694,-0.2826501131,-0.2342957407,-0.0471112095,-0.0703744143,-0.0118753752,0.327337712,0.059190955,-0.0226682574,-0.1344109178,0.173205018,0.1917792708,0.2761690617,0.1326986998,-0.1111938059,-0.0540863015,0.1674630344,-0.1418985426,0.0504370183,-0.1022382751,0.1748782247,-0.3800939322,-0.1850723773,-0.2209319025,-0.1899910122,-0.0593667552,0.0056179422,0.0851600319,-0.0492064022,-0.4295628667,0.7414498925,0.0367129631,-0.0033610959,-0.3613651991,0.0072339717,-0.519233346,-0.1364691406,0.3436326087,0.2566226721,0.0873611793,-0.1851003319,-0.1780127883,-0.1202442944,0.231401965,0.4888186455,-0.1457407475,0.072288841,-0.1638385057,0.4451909065,0.5148231983,-0.2013085634,0.1564145982,0.3828032017,-0.1579393744,0.1176427826,0.3666434288,0.1787613183,0.0984643102,-0.0287645757,0.1494462788,0.5212036967,-0.2195901126,0.3170514107,0.0510395654,-0.205040887,0.1839181185,0.076694876,-0.1174869165,0.0914216489,-0.1269441247,-0.4108044803,0.4736970663,-0.3134751916,0.1255268008,0.032088723,-0.1045800075,0.0155138187,-0.1124412268,-0.2692781985,-0.4281441867,0.0863264278,0.0058847941,-0.04596195,0.0627174303,-0.4250349104,0.4678710103,0.0931651965,0.0272108298,0.10780406,0.0036231098,0.2771939039,0.2118200809,-0.2439618111,-0.311239332,0.2733721137,-0.0149419941,0.0196806695,-0.2052478492,0.2725364268,0.2738957703,-0.0472204983,-0.0130598005,0.121975638,-0.0163769946,-0.1525379866,0.1854173988,-0.0161909964,-0.1658261418,0.2898708582,0.6852639318,0.313913852,0.2161806822,0.1365796328,0.0343915895,-0.2727378309,-0.0744914338,-0.184491843,-0.1683408022,0.5998206735,-0.0140424725,0.4637779593,0.12450248,0.0161402766,0.2041460574,0.0519173145,-0.0634287596,0.2228632271,0.3018843234,-0.0633598641,0.0389580429,0.1984835714,-0.5174875259,0.3662194014,0.5618277192,0.0204133354,-0.2030178607,0.1690208316,0.2101719528,-0.1038046479,0.1419193298,-0.0415256582,0.2087391317,0.1585720181,0.1179284826,-0.0678606927,0.0542456694,-0.0497793555,0.1372822076,0.1381000876,0.2204448879,0.0018661851,0.1506847888,-0.283646822,-0.2685750723,-0.2917861044,0.177754581,-0.0718961135,0.0101161404,-0.0704019889,-0.178822726,-0.1828874648,-0.0714192614,0.1259657443,-0.3630818129,-0.2696949244,-0.0128562143,0.2425182015,-0.0344563387,0.1265592575,0.3251162171,0.2205762863,-0.2097807378,-0.0292370804,-0.3193391562,0.0127628986,0.1298243254,-0.0781447515,0.2891869545,0.2108855993,0.3619936705,0.0631850138,0.0033280742,-0.4760684371,-0.1293813288,-0.0054734368,-0.2573159039,0.149151668,0.1148338467,0.1059922203,-0.2058687508,-0.2748154104,0.2950798571,0.1609950811,-0.2612763941,-0.2160706222,0.082881324,0.0717442632,0.0171804149,-0.0126680499,-0.4556742907,-0.3141210675,0.6144338846,-0.3349191546,0.1995882541,-0.3116193414,-0.1188745424,-0.0672000721,-0.375888139,0.1564348638,0.123886399,-0.5098872185,0.0115591241,0.0765492469,-0.0719269067,-0.3055360615,0.0593451336,-0.1284348369,0.3425966799,-0.1451997608,-0.1818318814,-0.3531471491,0.2227023393,0.2806251049,0.0808547139,0.2303673923,-0.0066967253,-0.0460875183,0.2179149091,-0.5478337407,0.0514658615,-0.0275190342,-0.225131914,0.2059124261,0.3937220573,-0.0657221749,0.698502183,0.2327252179,-0.0546037145,0.079177916,0.1058045551,-0.1620604843,0.0792652443,-0.5742771626,-0.0749844983,-0.085230656,0.1314907223,-0.0142814312,-0.0564299747,-0.2098163664,-0.0834245756,0.1092985943,-0.2157377154,-0.1491218507,0.0940395668,-0.0764598399,0.2606415749,0.0131567214,-0.333643049,-0.2862174809,0.0621946827,-0.2053037137,0.2579520941,0.3568188846,-0.17282103,-0.4844960272,-0.2232121974,-0.635753274,0.4355053306,0.1921811551,-0.009120414,-0.0816364586,0.1054301113,0.1647631079,-0.2243766189,0.6910618544,-0.2037560493,-0.2637610734,0.2256608903,-0.6402680874,0.0864888951,0.1671721488,0.1187863126,0.3876622319,0.4509106278,0.5235896111,-0.3956730068,-0.5414092541,0.1828792989,0.1317850649,-0.0529208779,0.0193395335,-0.1313537955,-0.0753118396,-0.1941631734,0.0835533142,-0.129420191,0.110629864,0.1027878374,-0.0370904952,-0.0578343198,0.0091064181,0.2668739259,-0.0322134122,-0.0902836844,0.3156908154,0.3008129597,-0.1654060185,0.2038614899,-0.2387535274,0.1147100478,-0.1745379567,-0.2069048733,-0.2162050605,0.0620183013,0.6206789613,0.3551598191,0.2062289268,0.4559169412,0.0465967394,0.4762690663,-0.3307176828,0.0582879893,0.5117327571,0.2649245262,-0.3802994788,-0.2002625912,0.3450416625,0.3504221141,-0.0121146869,0.5086841583,0.0423605703,-0.060446132,0.0444564484,-0.0583310276,0.7593258619,-0.4249180555,0.1466023475,-0.338953048,0.1035973951,0.4250380397,-0.2568286359,0.2103053033,-0.0430509895,-0.1907902062,0.0493200384,-0.0096456539,0.2014254034,0.2783035934,-0.1936883926,0.319020927,0.2305432558,-0.0606961325,-0.194079116,0.4177840948,0.1798280478,-0.4362957478,-0.3066778481,-0.0150526604,-0.0649351627,-0.1510459334,0.0750262588,-0.2331499755,0.2507272065,0.1179224774,-0.3976918757,-0.1488889158,-0.0366206393,0.237542659,-0.0615056418,-0.1523457468,-0.106715098,0.1302315891,-0.0566658117,-0.1137297079,-0.0271515939,-0.0089071793,-0.2871978879,0.2094552517,-0.2869935036,-0.2622612417,0.0655259266,-0.0927570462,-0.1397621781,-0.1045940146,0.1187636852,-0.6062473655,-0.1703051776,0.1514945477,-0.0287242122,0.2963555455,0.0383203402,0.3484083414,-0.142250061,-0.1584411114,-0.00508695,0.3774757981,-0.1763743013,0.4044017196,0.1233305037,-0.1359499097,-0.0294333901,0.0941171274,0.109527275,-0.1464933455,0.2336682081,0.1438096762,-0.1953278333,-0.1703956127,-0.1697768718,-0.1085418388,-0.0375288762,0.3576135635,-0.053379897,-0.1699129343,-0.0069105015,0.0630111322,-0.084813036,-0.0033894738,-0.1503293067,-0.2641432881,-0.1541773975,0.124679625,-0.1960364878,0.1195263192,0.0146565922,0.1598179191,0.0996453315,0.5086374283,-0.2206014544,-0.0527483374,-0.1507127732,0.082596831,0.5204796791,0.0385434479,0.021119928,0.294254005,0.0798988342,0.0336829536,-0.1852442771,-0.1719635576,0.1514390111,0.1333415508,-0.0991497412,-0.0125978459,0.1749425232,0.1742756218,-0.2076808661,-0.4301811755,-0.0268772952,-0.083162941,0.0276098419,-0.0333142206,0.2164851576,-0.2499692142,0.2745042443,0.0109730577,0.1896639019,-0.0638648421,-0.1929313838,0.2683271766,0.0884772241,-0.1640546918,0.2256069779,-0.040109504,0.2410572767,-0.1838134825,0.1901144236,0.0607354902,-0.2847617269,-0.0582740046,0.4337834418,0.3455114067,0.2503737807,0.3162878156,0.112263456,0.0182970017,-0.243984282,-0.1557812393,0.6245524883,0.0473308787,0.2133215666,-0.0203064624,0.1649937928,0.5159298778,0.0690936968,-0.0537639074,-0.0778062418,-0.2117260695,0.0363140777,-0.1234947443,0.3134807944,0.0214873869,0.2551138997,0.2318671644,0.1389112622,0.0086516915,-0.0456012487,0.6921669841,-0.2063691914,0.350348264,0.2106399536,-0.2947515845,0.2167814523,0.1577153355,0.2116494924,0.0827709734,0.0134332692,0.3568864167,-0.1093297154,-0.5609958172,-0.1765211523,0.9012926817,-0.0376706384,-0.3799575865,0.1442844123,-0.0615610406,-0.0592277907,0.2246249169,0.159296602,0.0516081601,0.8774401546,-0.1056218371,0.4569301903,-0.137046054,0.0782549977,-0.092511557,0.3780355752,-0.0728996098,-0.2244943231,0.0641603172,-0.0923504308,-0.0718253851,0.4819128811,0.2236920893,0.1731495261,-0.0984741673,0.3460917473,-0.4285177588,0.0545228533,-0.1794224828,0.2929781973,-0.1623265594,-0.0109837828,0.3843776286,-0.0161977056,-0.0858764127,-0.3041474819,-0.1996902823,0.0512237847,-0.3601373434,0.2270432413,-0.0726895183,-0.1172734946,-0.130866617,0.0098167453,-0.4142777026,-0.1568533629,0.5696909428,-0.416139245,-0.0873731151,-0.0183593556,0.0547336675,-0.0591137633,0.5470656753,0.2235113531,0.3829602599,-0.0993445292,-0.2579564154,-0.5372639298,-0.5388491154,-0.0264556836,-0.0320624337,0.115336284,0.4471927285,0.193145588,0.2433094233,0.0486228839,-0.3383333087,0.1635773182,0.0548706576,-0.3029490113,0.1746530235,-0.1389034986,-0.0983852968,0.1345933527,-0.1988972425,0.40515095,0.1380420625,-0.0710135251,-0.180071637,0.2221333385,0.2324645817,-0.0348790735,-0.1650140435,0.1895060986,0.3696228564,-0.2539097071,-0.0384013988,-0.0649534166,0.0466290899,-0.1877788603,0.0080218418,-0.126412496,0.3596163094,-0.2276293337,0.1917669326,-0.1608671695,0.2650917768,-0.1194320545,-0.084133327,0.1314115077,0.0772344172,0.090367727,0.0938851759,0.0139723876,0.3151397407,-0.0007177517,0.1136693507,-0.2220336348,-0.052893918,0.3545237184,-0.4865018427,-0.2119890749,-0.2927509248,0.1320073754,0.1737545133,0.1891282201,-0.5084637403,-0.2953651249,0.1800438613,0.0227168091,-0.0319271237,0.5896507502,-0.1400583684,-0.3424473107,-0.1829026639,-0.1905979365,0.0651327744,-0.3358170986,-0.0759488121,-0.1746187806]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/425","title":"Correct data structure for PAN-X task in XTREME dataset?","comments":"Thanks for noticing ! This looks more reasonable indeed.\r\nFeel free to open a PR","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.","comment_length":15,"text":"Correct data structure for PAN-X task in XTREME dataset?\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.\nThanks for noticing ! This looks more reasonable indeed.\r\nFeel free to open a PR","embeddings":[-0.0446428172,-0.1960519552,-0.0213917121,0.3716609478,0.0390401855,-0.2763307691,-0.0206372105,0.1579513103,-0.0717805475,-0.0662426949,-0.1613530368,0.2627067268,0.0014278574,0.1680819094,0.4264350533,-0.1467018872,0.0484421924,-0.0699055418,0.2786860168,-0.0526146963,-0.109093152,0.1506099999,-0.3253746927,0.0544689558,-0.3497874737,0.0837271884,-0.0678211153,0.3462400138,-0.2758393288,-0.493368417,0.5901547074,0.0610429049,0.3375472426,0.3253481388,-0.0001185823,0.007394406,0.0491500795,-0.2976795733,0.063000001,-0.5153079629,-0.21160613,-0.6882328987,-0.0707507059,-0.1863514483,-0.249424383,-0.0691500232,0.2359091789,0.1613155603,0.3286878169,0.404129982,0.0667554885,0.1186724603,-0.0112289721,0.1646236479,0.4069048762,0.3912606835,-0.2984111607,-0.0554718524,0.5836415291,0.1500933468,-0.2007237524,0.226136744,0.0665774941,-0.0957404003,0.2168319821,0.311748594,0.1106207892,-0.4583888352,-0.2030344903,0.2828918993,0.2304347605,-0.2553840578,-0.2035078704,-0.2779316902,-0.0572837256,-0.2900469899,-0.0280274134,0.3609626591,-0.0808850378,-0.0090993838,0.2543638647,-0.0029822402,-0.1114807874,0.2031368911,-0.0105202226,0.4096223116,-0.0134635018,-0.1075180918,-0.1396067441,-0.1162076294,0.7018569112,-0.0104258293,0.1576858014,0.1197108179,-0.6041291952,0.0803796276,-0.1363217384,-0.2818925083,0.1663054824,-0.0484371334,0.2098659426,0.1589148641,0.148524031,0.0145868212,0.3213757873,0.2003956288,0.2381373495,0.0310105663,-0.1307128817,0.0734806508,0.0858036652,-0.0963578075,-0.3183752894,0.0653057396,0.2368329167,-0.0228066426,0.2889071107,-0.1962373108,-0.3680241704,-0.0199918374,-0.1284303069,0.072580345,0.1414791495,0.061180193,0.0080608483,0.5874376297,-0.0742886215,0.1787090451,-0.2538342476,0.0291408673,-0.1472996622,0.1694096178,-0.6243599057,-0.1061867177,0.2582582831,0.091962114,0.1970394701,-0.0064032744,0.121820122,-0.3322490454,0.1758813262,-0.2791434526,0.2651101649,-0.0158670694,0.2457386106,-0.1295719892,-0.3093241453,-0.0095198667,-0.1647403091,0.2194372267,-0.7123686075,-0.0245730337,0.0450958833,0.0754593015,-0.214194268,0.1435741037,-0.504322648,0.0322593153,0.0630441457,0.2181688845,0.1069046035,-0.3306136727,-0.2142900974,-0.0720484778,0.0830698535,0.1326568276,-0.1291505694,0.0973811597,0.0630992949,0.0002566593,0.1041970924,0.2162794471,-0.0109667862,0.284181118,-0.1886345744,0.5723384023,0.4195818007,-0.8037276268,-0.3378547132,0.0875090212,-0.1656147987,0.0018321252,-0.0777145699,0.2999708056,0.2168296725,-0.1097091362,0.2860379517,0.4717787802,-0.0601560883,0.1129643992,-0.1055826619,0.2662775815,-0.1180936024,0.3215996325,0.0420975052,-0.2685292661,-0.1594627202,0.0751415715,0.1948614568,-0.1158283576,0.0870857239,0.4314538538,0.089273192,-0.0513841324,-0.0506237373,0.0383254662,-0.7624905109,0.1141011417,-0.4289596081,0.3793443143,0.0106568281,-0.0671950504,-0.1728422344,-0.0510757342,-0.296262145,-0.1906543225,0.107161127,-0.1209991425,-0.1237812713,-0.3034522235,-0.0762852654,0.3974147141,-0.2260980457,0.0737625808,-0.5780182481,0.340782702,0.1180048063,-0.0664847717,-0.0123105403,0.4472563863,0.1922498047,-0.0482966453,0.1684597582,0.2209219038,0.2148898542,-0.1571745276,-0.0612342209,-0.0020257682,0.1334930658,-0.5777854919,0.0659277737,0.2230994701,-0.001422982,-0.2143050879,-0.1243546456,0.5517559648,0.1847706735,0.2072845101,0.0289730113,0.0707501099,0.2122232467,-0.1713617146,-0.2344333678,-0.1039047167,0.1545492858,0.1233171746,0.0305474959,0.0995655283,-0.5023413897,0.0200132038,0.3077193201,0.0274304561,0.1898184419,0.2047269344,-0.1531089544,-0.0791647956,0.0374575406,0.3701202571,0.3590866625,0.0914273858,0.1098857149,0.1304076463,-0.0329413153,0.0724881813,-0.0358842276,0.0617540143,0.1510176659,0.2119652778,0.1785159111,-0.0841950774,-0.0858964771,-0.2191504538,0.0321776755,0.152163282,-0.3294998705,0.0780693293,0.0368220694,-0.450265348,-0.1473740041,0.0441721193,-0.1361576915,-0.3193361759,-0.0379218347,-0.0571559481,0.2387743443,0.084067151,-0.4746881425,-0.2150861621,-0.1986020505,-0.3406288028,-0.0583148114,0.0742932335,-0.1382389069,-0.0267540924,0.2119731903,0.5459074378,-0.031271968,-0.4353145659,0.3901607394,-0.3872416317,-0.2023112923,0.301656425,-0.2638221681,0.3121575713,0.217154935,-0.1040859297,-0.4046024084,0.195917204,0.1874918491,-0.321480006,-0.3141257763,0.3168615997,-0.1772629917,-0.0916748643,-0.0264440607,-0.40634498,-0.5237715244,-0.2054876089,0.0119940396,0.2230312675,0.3162436187,0.2151381969,-0.2596668601,-0.1451510787,-0.1321804971,-0.1955112368,0.0069593326,-0.5726137161,0.4568421841,-0.24419415,-0.2051525563,-0.0859806389,-0.2658783793,0.4713196158,0.3780293763,-0.0684937015,-0.0375404693,0.1413383931,0.0351820067,0.177708894,-0.0622194856,0.2964116931,-0.2141356468,0.0457051024,-0.0277902856,-0.0081237135,0.0785258338,0.3201446831,0.2304375917,-0.0265186951,-0.2571677864,0.110347338,0.0635766909,0.1552903205,0.0463344343,0.4367932975,0.005378745,0.1145690829,0.0847834796,0.0224085972,-0.164030239,0.0601466,0.233269453,-0.0009895281,0.2357880324,-0.1547724605,-0.072097823,0.1263714284,-0.1329991817,-0.215120554,0.5398620367,-0.1397541612,0.4105234146,-0.1709616035,0.0268179663,-0.0614328459,-0.4580695629,-0.257130444,0.086149618,0.1797157824,0.155405134,-0.6861747503,0.3985427618,-0.3888285756,-0.0206326935,0.1728511304,0.5378324389,0.2120253146,0.0804060176,0.1571722478,-0.0373992324,0.3112249672,-0.2897289991,-0.1462424099,0.1092613339,0.1001149788,-0.3250236809,-0.0430728868,-0.2508806586,0.0265304223,0.4807166457,0.4030665755,-0.3020348847,-0.4488573372,-0.0142154954,0.0474353395,-0.2095984071,-0.1977830976,-0.2662701309,-0.4505195916,-0.0922009498,-0.0910787657,0.5540920496,-0.0119683752,0.285989821,0.0289197415,-0.2357486933,0.2654157579,0.3301908374,-0.0371142514,0.5644020438,-0.2395819277,0.0057377298,-0.0657557547,-0.0843305588,0.2901473939,0.4803256691,-0.18594037,-0.1504533291,0.1800045818,-0.2238452286,0.4223891199,0.1663793623,0.1885036528,0.0188958589,-0.2950081825,-0.0011859048,-0.1358927488,0.1362023652,0.3885795772,-0.1082977429,-0.4188079834,-0.3653247952,0.1768360138,0.3088775873,0.1055980548,-0.2310434431,-0.1721609235,-0.2617760301,-0.1415586323,0.4663684666,1.016680479,0.030185407,0.433570534,0.1726903319,0.0897267163,0.0607269071,-0.444486618,0.2049940228,-0.3482038677,-0.3285882473,-0.2036454231,-0.2574140429,-0.0620909221,0.2026154697,0.0319126844,0.6626021862,0.2256908417,-0.0447899438,0.0229511857,0.2159696668,0.0093683628,-0.3867395818,-0.3745247722,0.0040317033,-0.2863345742,0.0149859404,-0.1535770446,-0.2083832622,-0.0591487885,0.1994757652,0.1399660408,0.2819466293,0.0934466422,-0.1829066277,0.4239276946,-0.5605942607,-0.0533471443,0.2428724915,0.077303797,-0.322191,0.0031517914,0.2361234426,0.4050419629,0.1009970233,0.3913243711,-0.3321993649,0.2601027191,0.0498278886,0.1221913695,-0.2428604513,0.1581598967,-0.0372625589,-0.4174329937,0.0479725488,-0.1750261933,-0.2749729753,-0.2087987065,0.0269379523,-0.1443690658,-0.049738355,0.0549408123,0.2594358027,-0.0902868882,0.148101449,0.1132918224,-0.2592061758,0.1990920603,0.0831901953,0.2976744771,-0.3125315309,0.0376100764,0.2482126951,0.0151349036,-0.0992361903,0.180581525,0.4821507633,-0.617361784,0.0337477662,-0.2434498519,-0.0137992604,-0.0163004398,0.1851167381,0.0020547709,-0.2624997497,0.1627749503,-0.2828570008,-0.3409923017,0.3751991391,0.1163776442,0.2782913744,-0.1979854256,0.0098260185,0.1509936601,-0.2200442106,-0.2390500307,0.1753156036,-0.1794706732,0.1961244643,-0.1157648787,-0.096291475,0.080403842,-0.0942012146,-0.0249693338,0.2664697766,-0.1458446831,-0.0578539148,0.0179341715,0.1243914366,0.1327522844,-0.1157702655,0.1248916909,0.1200044379,-0.0515412576,-0.1373377591,-0.1135678813,0.1800017059,0.0228337441,0.2303531468,-0.2976282835,0.0392743871,-0.1231651306,0.2310590148,0.189538762,0.1202059165,0.0657504871,0.2072514594,-0.0511192568,0.0248664636,-0.5182061791,0.2064265907,0.1967262477,0.3688436747,0.2151999176,-0.1553563476,0.1056381837,-0.032278128,0.1498689055,0.3083644211,-0.2699720263,-0.0058445805,0.241904065,0.1483671665,-0.1588442177,-0.0049726604,0.3671055734,0.0078911902,0.008068773,0.4790737331,0.3525026143,0.136815384,-0.1463459283,0.0633270293,0.2355753332,-0.2535267174,0.0676125661,0.4052120447,-0.005103806,0.1394571364,0.5995871425,0.0332877487,-0.1143290102,0.6000959277,0.3011852205,0.7809863091,0.199559778,0.1787547022,-0.0658113658,-0.1044648662,-0.4129894078,0.2548288405,0.1273963898,0.1769531816,0.0383527577,0.143287167,0.1453569233,-0.1860472709,-0.0964768156,0.113500081,-0.3404924273,0.0229304079,-0.331787169,-0.2342431843,-0.1217740625,-0.1177005768,-0.1722740084,0.0367383994,0.061489068,0.1339388788,-0.0074618193,-0.0612117462,0.0823539197,0.2852355242,0.0478237383,-0.1977075338,0.5288455486,0.0339015722,-0.0469572246,-0.1587117612,0.2742477059,0.2531481087,0.2315032333,-0.0519730113,-0.1070040986,0.3387274742,-0.1102280021,-0.02969338,0.1004743874,0.2005059421,0.2549166381,0.1804338545,0.0651821718,-0.0043521184,0.3010003567,0.0695356131,-0.1928669661,0.2435857952,0.2801603675,0.1861452311,-0.2890454829,-0.1294614524,-0.2241025269,-0.3388958275,-0.1642102152,0.314212203,-0.1428406239,0.2404812127,-0.130814895,0.0508457981,0.042049367,0.1968667656,0.1187227666,0.1140162945,-0.4945984483,-0.0885237828,-0.7457126975,0.2005533725,-0.3096918166,-0.2826579809,0.2355704159,0.0266045816,-0.0474609956,0.3022997379,-0.23467426,0.2289671898,0.0685480013,0.334798485,-0.6064161658,0.0685795322,0.2317826003,0.35899809,0.0531915426,-0.124981299,-0.276468724,0.2186534852,-0.0702891573,-0.0727587193,-0.0685441419,-0.1896738708,-0.0558257475,0.2673411965,-0.212423116,0.1567766815,-0.0746062472,-0.1744633466,-0.2747464776,-0.0597104505,-0.1644858569,0.0210557934,0.1005385742,0.3167755306,0.0033465335,0.1547871828,-0.3915068209,-0.2984790206,0.0132773137,0.0827491954,-0.3641844988,0.4328106642,-0.2321629971,0.1437955201,0.0377221107,0.2259716541,0.074632287,-0.0033203885,-0.1593918055,-0.4455848634,0.2613032758,0.2699011862,-0.1986567527,-0.1195013747,0.0043353084,0.1073929965,0.1258694679,-0.4479656518,0.1364092082,0.1799704432,-0.0451218039,-0.0945755392,0.0839961171,-0.0237168744,-0.0781790763,-0.1301056743,0.1238042712,-0.0262836535,-0.1145851761,-0.2040123045,-0.1585304439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/425","title":"Correct data structure for PAN-X task in XTREME dataset?","comments":"Hi @lhoestq \r\nI made the proposed changes to the `xtreme.py` script. I noticed that I also need to change the schema in the `dataset_infos.json` file. More specifically the `\"features\"` part of the PAN-X.LANG dataset:\r\n\r\n```json\r\n\"features\":{\r\n \"word\":{\r\n \"dtype\":\"string\",\r\n \"id\":null,\r\n \"_type\":\"Value\"\r\n },\r\n \"ner_tag\":{\r\n \"dtype\":\"string\",\r\n \"id\":null,\r\n \"_type\":\"Value\"\r\n },\r\n \"lang\":{\r\n \"dtype\":\"string\",\r\n \"id\":null,\r\n \"_type\":\"Value\"\r\n }\r\n}\r\n```\r\nTo fit the code above the fields `\"word\"`, `\"ner_tag\"`, and `\"lang\"` would become `\"words\"`, `ner_tags\"` and `\"langs\"`. In addition the `dtype` should be changed from `\"string\"` to `\"list\"`.\r\n\r\n I made this changes but when trying to test this locally with `dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')` I face the issue that the `dataset_info.json` file is always overwritten by a downloaded version with the old settings, which then throws an error because the schema does not match. This makes it hard to test the changes locally. Do you have any suggestions on how to deal with that?\r\n","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.","comment_length":148,"text":"Correct data structure for PAN-X task in XTREME dataset?\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.\nHi @lhoestq \r\nI made the proposed changes to the `xtreme.py` script. I noticed that I also need to change the schema in the `dataset_infos.json` file. More specifically the `\"features\"` part of the PAN-X.LANG dataset:\r\n\r\n```json\r\n\"features\":{\r\n \"word\":{\r\n \"dtype\":\"string\",\r\n \"id\":null,\r\n \"_type\":\"Value\"\r\n },\r\n \"ner_tag\":{\r\n \"dtype\":\"string\",\r\n \"id\":null,\r\n \"_type\":\"Value\"\r\n },\r\n \"lang\":{\r\n \"dtype\":\"string\",\r\n \"id\":null,\r\n \"_type\":\"Value\"\r\n }\r\n}\r\n```\r\nTo fit the code above the fields `\"word\"`, `\"ner_tag\"`, and `\"lang\"` would become `\"words\"`, `ner_tags\"` and `\"langs\"`. In addition the `dtype` should be changed from `\"string\"` to `\"list\"`.\r\n\r\n I made this changes but when trying to test this locally with `dataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')` I face the issue that the `dataset_info.json` file is always overwritten by a downloaded version with the old settings, which then throws an error because the schema does not match. This makes it hard to test the changes locally. Do you have any suggestions on how to deal with that?\r\n","embeddings":[-0.0446428172,-0.1960519552,-0.0213917121,0.3716609478,0.0390401855,-0.2763307691,-0.0206372105,0.1579513103,-0.0717805475,-0.0662426949,-0.1613530368,0.2627067268,0.0014278574,0.1680819094,0.4264350533,-0.1467018872,0.0484421924,-0.0699055418,0.2786860168,-0.0526146963,-0.109093152,0.1506099999,-0.3253746927,0.0544689558,-0.3497874737,0.0837271884,-0.0678211153,0.3462400138,-0.2758393288,-0.493368417,0.5901547074,0.0610429049,0.3375472426,0.3253481388,-0.0001185823,0.007394406,0.0491500795,-0.2976795733,0.063000001,-0.5153079629,-0.21160613,-0.6882328987,-0.0707507059,-0.1863514483,-0.249424383,-0.0691500232,0.2359091789,0.1613155603,0.3286878169,0.404129982,0.0667554885,0.1186724603,-0.0112289721,0.1646236479,0.4069048762,0.3912606835,-0.2984111607,-0.0554718524,0.5836415291,0.1500933468,-0.2007237524,0.226136744,0.0665774941,-0.0957404003,0.2168319821,0.311748594,0.1106207892,-0.4583888352,-0.2030344903,0.2828918993,0.2304347605,-0.2553840578,-0.2035078704,-0.2779316902,-0.0572837256,-0.2900469899,-0.0280274134,0.3609626591,-0.0808850378,-0.0090993838,0.2543638647,-0.0029822402,-0.1114807874,0.2031368911,-0.0105202226,0.4096223116,-0.0134635018,-0.1075180918,-0.1396067441,-0.1162076294,0.7018569112,-0.0104258293,0.1576858014,0.1197108179,-0.6041291952,0.0803796276,-0.1363217384,-0.2818925083,0.1663054824,-0.0484371334,0.2098659426,0.1589148641,0.148524031,0.0145868212,0.3213757873,0.2003956288,0.2381373495,0.0310105663,-0.1307128817,0.0734806508,0.0858036652,-0.0963578075,-0.3183752894,0.0653057396,0.2368329167,-0.0228066426,0.2889071107,-0.1962373108,-0.3680241704,-0.0199918374,-0.1284303069,0.072580345,0.1414791495,0.061180193,0.0080608483,0.5874376297,-0.0742886215,0.1787090451,-0.2538342476,0.0291408673,-0.1472996622,0.1694096178,-0.6243599057,-0.1061867177,0.2582582831,0.091962114,0.1970394701,-0.0064032744,0.121820122,-0.3322490454,0.1758813262,-0.2791434526,0.2651101649,-0.0158670694,0.2457386106,-0.1295719892,-0.3093241453,-0.0095198667,-0.1647403091,0.2194372267,-0.7123686075,-0.0245730337,0.0450958833,0.0754593015,-0.214194268,0.1435741037,-0.504322648,0.0322593153,0.0630441457,0.2181688845,0.1069046035,-0.3306136727,-0.2142900974,-0.0720484778,0.0830698535,0.1326568276,-0.1291505694,0.0973811597,0.0630992949,0.0002566593,0.1041970924,0.2162794471,-0.0109667862,0.284181118,-0.1886345744,0.5723384023,0.4195818007,-0.8037276268,-0.3378547132,0.0875090212,-0.1656147987,0.0018321252,-0.0777145699,0.2999708056,0.2168296725,-0.1097091362,0.2860379517,0.4717787802,-0.0601560883,0.1129643992,-0.1055826619,0.2662775815,-0.1180936024,0.3215996325,0.0420975052,-0.2685292661,-0.1594627202,0.0751415715,0.1948614568,-0.1158283576,0.0870857239,0.4314538538,0.089273192,-0.0513841324,-0.0506237373,0.0383254662,-0.7624905109,0.1141011417,-0.4289596081,0.3793443143,0.0106568281,-0.0671950504,-0.1728422344,-0.0510757342,-0.296262145,-0.1906543225,0.107161127,-0.1209991425,-0.1237812713,-0.3034522235,-0.0762852654,0.3974147141,-0.2260980457,0.0737625808,-0.5780182481,0.340782702,0.1180048063,-0.0664847717,-0.0123105403,0.4472563863,0.1922498047,-0.0482966453,0.1684597582,0.2209219038,0.2148898542,-0.1571745276,-0.0612342209,-0.0020257682,0.1334930658,-0.5777854919,0.0659277737,0.2230994701,-0.001422982,-0.2143050879,-0.1243546456,0.5517559648,0.1847706735,0.2072845101,0.0289730113,0.0707501099,0.2122232467,-0.1713617146,-0.2344333678,-0.1039047167,0.1545492858,0.1233171746,0.0305474959,0.0995655283,-0.5023413897,0.0200132038,0.3077193201,0.0274304561,0.1898184419,0.2047269344,-0.1531089544,-0.0791647956,0.0374575406,0.3701202571,0.3590866625,0.0914273858,0.1098857149,0.1304076463,-0.0329413153,0.0724881813,-0.0358842276,0.0617540143,0.1510176659,0.2119652778,0.1785159111,-0.0841950774,-0.0858964771,-0.2191504538,0.0321776755,0.152163282,-0.3294998705,0.0780693293,0.0368220694,-0.450265348,-0.1473740041,0.0441721193,-0.1361576915,-0.3193361759,-0.0379218347,-0.0571559481,0.2387743443,0.084067151,-0.4746881425,-0.2150861621,-0.1986020505,-0.3406288028,-0.0583148114,0.0742932335,-0.1382389069,-0.0267540924,0.2119731903,0.5459074378,-0.031271968,-0.4353145659,0.3901607394,-0.3872416317,-0.2023112923,0.301656425,-0.2638221681,0.3121575713,0.217154935,-0.1040859297,-0.4046024084,0.195917204,0.1874918491,-0.321480006,-0.3141257763,0.3168615997,-0.1772629917,-0.0916748643,-0.0264440607,-0.40634498,-0.5237715244,-0.2054876089,0.0119940396,0.2230312675,0.3162436187,0.2151381969,-0.2596668601,-0.1451510787,-0.1321804971,-0.1955112368,0.0069593326,-0.5726137161,0.4568421841,-0.24419415,-0.2051525563,-0.0859806389,-0.2658783793,0.4713196158,0.3780293763,-0.0684937015,-0.0375404693,0.1413383931,0.0351820067,0.177708894,-0.0622194856,0.2964116931,-0.2141356468,0.0457051024,-0.0277902856,-0.0081237135,0.0785258338,0.3201446831,0.2304375917,-0.0265186951,-0.2571677864,0.110347338,0.0635766909,0.1552903205,0.0463344343,0.4367932975,0.005378745,0.1145690829,0.0847834796,0.0224085972,-0.164030239,0.0601466,0.233269453,-0.0009895281,0.2357880324,-0.1547724605,-0.072097823,0.1263714284,-0.1329991817,-0.215120554,0.5398620367,-0.1397541612,0.4105234146,-0.1709616035,0.0268179663,-0.0614328459,-0.4580695629,-0.257130444,0.086149618,0.1797157824,0.155405134,-0.6861747503,0.3985427618,-0.3888285756,-0.0206326935,0.1728511304,0.5378324389,0.2120253146,0.0804060176,0.1571722478,-0.0373992324,0.3112249672,-0.2897289991,-0.1462424099,0.1092613339,0.1001149788,-0.3250236809,-0.0430728868,-0.2508806586,0.0265304223,0.4807166457,0.4030665755,-0.3020348847,-0.4488573372,-0.0142154954,0.0474353395,-0.2095984071,-0.1977830976,-0.2662701309,-0.4505195916,-0.0922009498,-0.0910787657,0.5540920496,-0.0119683752,0.285989821,0.0289197415,-0.2357486933,0.2654157579,0.3301908374,-0.0371142514,0.5644020438,-0.2395819277,0.0057377298,-0.0657557547,-0.0843305588,0.2901473939,0.4803256691,-0.18594037,-0.1504533291,0.1800045818,-0.2238452286,0.4223891199,0.1663793623,0.1885036528,0.0188958589,-0.2950081825,-0.0011859048,-0.1358927488,0.1362023652,0.3885795772,-0.1082977429,-0.4188079834,-0.3653247952,0.1768360138,0.3088775873,0.1055980548,-0.2310434431,-0.1721609235,-0.2617760301,-0.1415586323,0.4663684666,1.016680479,0.030185407,0.433570534,0.1726903319,0.0897267163,0.0607269071,-0.444486618,0.2049940228,-0.3482038677,-0.3285882473,-0.2036454231,-0.2574140429,-0.0620909221,0.2026154697,0.0319126844,0.6626021862,0.2256908417,-0.0447899438,0.0229511857,0.2159696668,0.0093683628,-0.3867395818,-0.3745247722,0.0040317033,-0.2863345742,0.0149859404,-0.1535770446,-0.2083832622,-0.0591487885,0.1994757652,0.1399660408,0.2819466293,0.0934466422,-0.1829066277,0.4239276946,-0.5605942607,-0.0533471443,0.2428724915,0.077303797,-0.322191,0.0031517914,0.2361234426,0.4050419629,0.1009970233,0.3913243711,-0.3321993649,0.2601027191,0.0498278886,0.1221913695,-0.2428604513,0.1581598967,-0.0372625589,-0.4174329937,0.0479725488,-0.1750261933,-0.2749729753,-0.2087987065,0.0269379523,-0.1443690658,-0.049738355,0.0549408123,0.2594358027,-0.0902868882,0.148101449,0.1132918224,-0.2592061758,0.1990920603,0.0831901953,0.2976744771,-0.3125315309,0.0376100764,0.2482126951,0.0151349036,-0.0992361903,0.180581525,0.4821507633,-0.617361784,0.0337477662,-0.2434498519,-0.0137992604,-0.0163004398,0.1851167381,0.0020547709,-0.2624997497,0.1627749503,-0.2828570008,-0.3409923017,0.3751991391,0.1163776442,0.2782913744,-0.1979854256,0.0098260185,0.1509936601,-0.2200442106,-0.2390500307,0.1753156036,-0.1794706732,0.1961244643,-0.1157648787,-0.096291475,0.080403842,-0.0942012146,-0.0249693338,0.2664697766,-0.1458446831,-0.0578539148,0.0179341715,0.1243914366,0.1327522844,-0.1157702655,0.1248916909,0.1200044379,-0.0515412576,-0.1373377591,-0.1135678813,0.1800017059,0.0228337441,0.2303531468,-0.2976282835,0.0392743871,-0.1231651306,0.2310590148,0.189538762,0.1202059165,0.0657504871,0.2072514594,-0.0511192568,0.0248664636,-0.5182061791,0.2064265907,0.1967262477,0.3688436747,0.2151999176,-0.1553563476,0.1056381837,-0.032278128,0.1498689055,0.3083644211,-0.2699720263,-0.0058445805,0.241904065,0.1483671665,-0.1588442177,-0.0049726604,0.3671055734,0.0078911902,0.008068773,0.4790737331,0.3525026143,0.136815384,-0.1463459283,0.0633270293,0.2355753332,-0.2535267174,0.0676125661,0.4052120447,-0.005103806,0.1394571364,0.5995871425,0.0332877487,-0.1143290102,0.6000959277,0.3011852205,0.7809863091,0.199559778,0.1787547022,-0.0658113658,-0.1044648662,-0.4129894078,0.2548288405,0.1273963898,0.1769531816,0.0383527577,0.143287167,0.1453569233,-0.1860472709,-0.0964768156,0.113500081,-0.3404924273,0.0229304079,-0.331787169,-0.2342431843,-0.1217740625,-0.1177005768,-0.1722740084,0.0367383994,0.061489068,0.1339388788,-0.0074618193,-0.0612117462,0.0823539197,0.2852355242,0.0478237383,-0.1977075338,0.5288455486,0.0339015722,-0.0469572246,-0.1587117612,0.2742477059,0.2531481087,0.2315032333,-0.0519730113,-0.1070040986,0.3387274742,-0.1102280021,-0.02969338,0.1004743874,0.2005059421,0.2549166381,0.1804338545,0.0651821718,-0.0043521184,0.3010003567,0.0695356131,-0.1928669661,0.2435857952,0.2801603675,0.1861452311,-0.2890454829,-0.1294614524,-0.2241025269,-0.3388958275,-0.1642102152,0.314212203,-0.1428406239,0.2404812127,-0.130814895,0.0508457981,0.042049367,0.1968667656,0.1187227666,0.1140162945,-0.4945984483,-0.0885237828,-0.7457126975,0.2005533725,-0.3096918166,-0.2826579809,0.2355704159,0.0266045816,-0.0474609956,0.3022997379,-0.23467426,0.2289671898,0.0685480013,0.334798485,-0.6064161658,0.0685795322,0.2317826003,0.35899809,0.0531915426,-0.124981299,-0.276468724,0.2186534852,-0.0702891573,-0.0727587193,-0.0685441419,-0.1896738708,-0.0558257475,0.2673411965,-0.212423116,0.1567766815,-0.0746062472,-0.1744633466,-0.2747464776,-0.0597104505,-0.1644858569,0.0210557934,0.1005385742,0.3167755306,0.0033465335,0.1547871828,-0.3915068209,-0.2984790206,0.0132773137,0.0827491954,-0.3641844988,0.4328106642,-0.2321629971,0.1437955201,0.0377221107,0.2259716541,0.074632287,-0.0033203885,-0.1593918055,-0.4455848634,0.2613032758,0.2699011862,-0.1986567527,-0.1195013747,0.0043353084,0.1073929965,0.1258694679,-0.4479656518,0.1364092082,0.1799704432,-0.0451218039,-0.0945755392,0.0839961171,-0.0237168744,-0.0781790763,-0.1301056743,0.1238042712,-0.0262836535,-0.1145851761,-0.2040123045,-0.1585304439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/425","title":"Correct data structure for PAN-X task in XTREME dataset?","comments":"Hi !\r\n\r\nYou have to point to your local script.\r\nFirst clone the repo and then:\r\n\r\n```python\r\ndataset = load_dataset(\".\/datasets\/xtreme\", \"PAN-X.en\")\r\n```\r\nThe \"xtreme\" directory contains \"xtreme.py\".\r\n\r\nYou also have to change the features definition in the `_info` method. You could use:\r\n\r\n```python\r\nfeatures = nlp.Features({\r\n \"words\": [nlp.Value(\"string\")],\r\n \"ner_tags\": [nlp.Value(\"string\")],\r\n \"langs\": [nlp.Value(\"string\")],\r\n})\r\n```\r\n\r\nHope this helps !\r\nLet me know if you have other questions.","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.","comment_length":66,"text":"Correct data structure for PAN-X task in XTREME dataset?\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.\nHi !\r\n\r\nYou have to point to your local script.\r\nFirst clone the repo and then:\r\n\r\n```python\r\ndataset = load_dataset(\".\/datasets\/xtreme\", \"PAN-X.en\")\r\n```\r\nThe \"xtreme\" directory contains \"xtreme.py\".\r\n\r\nYou also have to change the features definition in the `_info` method. You could use:\r\n\r\n```python\r\nfeatures = nlp.Features({\r\n \"words\": [nlp.Value(\"string\")],\r\n \"ner_tags\": [nlp.Value(\"string\")],\r\n \"langs\": [nlp.Value(\"string\")],\r\n})\r\n```\r\n\r\nHope this helps !\r\nLet me know if you have other questions.","embeddings":[-0.0446428172,-0.1960519552,-0.0213917121,0.3716609478,0.0390401855,-0.2763307691,-0.0206372105,0.1579513103,-0.0717805475,-0.0662426949,-0.1613530368,0.2627067268,0.0014278574,0.1680819094,0.4264350533,-0.1467018872,0.0484421924,-0.0699055418,0.2786860168,-0.0526146963,-0.109093152,0.1506099999,-0.3253746927,0.0544689558,-0.3497874737,0.0837271884,-0.0678211153,0.3462400138,-0.2758393288,-0.493368417,0.5901547074,0.0610429049,0.3375472426,0.3253481388,-0.0001185823,0.007394406,0.0491500795,-0.2976795733,0.063000001,-0.5153079629,-0.21160613,-0.6882328987,-0.0707507059,-0.1863514483,-0.249424383,-0.0691500232,0.2359091789,0.1613155603,0.3286878169,0.404129982,0.0667554885,0.1186724603,-0.0112289721,0.1646236479,0.4069048762,0.3912606835,-0.2984111607,-0.0554718524,0.5836415291,0.1500933468,-0.2007237524,0.226136744,0.0665774941,-0.0957404003,0.2168319821,0.311748594,0.1106207892,-0.4583888352,-0.2030344903,0.2828918993,0.2304347605,-0.2553840578,-0.2035078704,-0.2779316902,-0.0572837256,-0.2900469899,-0.0280274134,0.3609626591,-0.0808850378,-0.0090993838,0.2543638647,-0.0029822402,-0.1114807874,0.2031368911,-0.0105202226,0.4096223116,-0.0134635018,-0.1075180918,-0.1396067441,-0.1162076294,0.7018569112,-0.0104258293,0.1576858014,0.1197108179,-0.6041291952,0.0803796276,-0.1363217384,-0.2818925083,0.1663054824,-0.0484371334,0.2098659426,0.1589148641,0.148524031,0.0145868212,0.3213757873,0.2003956288,0.2381373495,0.0310105663,-0.1307128817,0.0734806508,0.0858036652,-0.0963578075,-0.3183752894,0.0653057396,0.2368329167,-0.0228066426,0.2889071107,-0.1962373108,-0.3680241704,-0.0199918374,-0.1284303069,0.072580345,0.1414791495,0.061180193,0.0080608483,0.5874376297,-0.0742886215,0.1787090451,-0.2538342476,0.0291408673,-0.1472996622,0.1694096178,-0.6243599057,-0.1061867177,0.2582582831,0.091962114,0.1970394701,-0.0064032744,0.121820122,-0.3322490454,0.1758813262,-0.2791434526,0.2651101649,-0.0158670694,0.2457386106,-0.1295719892,-0.3093241453,-0.0095198667,-0.1647403091,0.2194372267,-0.7123686075,-0.0245730337,0.0450958833,0.0754593015,-0.214194268,0.1435741037,-0.504322648,0.0322593153,0.0630441457,0.2181688845,0.1069046035,-0.3306136727,-0.2142900974,-0.0720484778,0.0830698535,0.1326568276,-0.1291505694,0.0973811597,0.0630992949,0.0002566593,0.1041970924,0.2162794471,-0.0109667862,0.284181118,-0.1886345744,0.5723384023,0.4195818007,-0.8037276268,-0.3378547132,0.0875090212,-0.1656147987,0.0018321252,-0.0777145699,0.2999708056,0.2168296725,-0.1097091362,0.2860379517,0.4717787802,-0.0601560883,0.1129643992,-0.1055826619,0.2662775815,-0.1180936024,0.3215996325,0.0420975052,-0.2685292661,-0.1594627202,0.0751415715,0.1948614568,-0.1158283576,0.0870857239,0.4314538538,0.089273192,-0.0513841324,-0.0506237373,0.0383254662,-0.7624905109,0.1141011417,-0.4289596081,0.3793443143,0.0106568281,-0.0671950504,-0.1728422344,-0.0510757342,-0.296262145,-0.1906543225,0.107161127,-0.1209991425,-0.1237812713,-0.3034522235,-0.0762852654,0.3974147141,-0.2260980457,0.0737625808,-0.5780182481,0.340782702,0.1180048063,-0.0664847717,-0.0123105403,0.4472563863,0.1922498047,-0.0482966453,0.1684597582,0.2209219038,0.2148898542,-0.1571745276,-0.0612342209,-0.0020257682,0.1334930658,-0.5777854919,0.0659277737,0.2230994701,-0.001422982,-0.2143050879,-0.1243546456,0.5517559648,0.1847706735,0.2072845101,0.0289730113,0.0707501099,0.2122232467,-0.1713617146,-0.2344333678,-0.1039047167,0.1545492858,0.1233171746,0.0305474959,0.0995655283,-0.5023413897,0.0200132038,0.3077193201,0.0274304561,0.1898184419,0.2047269344,-0.1531089544,-0.0791647956,0.0374575406,0.3701202571,0.3590866625,0.0914273858,0.1098857149,0.1304076463,-0.0329413153,0.0724881813,-0.0358842276,0.0617540143,0.1510176659,0.2119652778,0.1785159111,-0.0841950774,-0.0858964771,-0.2191504538,0.0321776755,0.152163282,-0.3294998705,0.0780693293,0.0368220694,-0.450265348,-0.1473740041,0.0441721193,-0.1361576915,-0.3193361759,-0.0379218347,-0.0571559481,0.2387743443,0.084067151,-0.4746881425,-0.2150861621,-0.1986020505,-0.3406288028,-0.0583148114,0.0742932335,-0.1382389069,-0.0267540924,0.2119731903,0.5459074378,-0.031271968,-0.4353145659,0.3901607394,-0.3872416317,-0.2023112923,0.301656425,-0.2638221681,0.3121575713,0.217154935,-0.1040859297,-0.4046024084,0.195917204,0.1874918491,-0.321480006,-0.3141257763,0.3168615997,-0.1772629917,-0.0916748643,-0.0264440607,-0.40634498,-0.5237715244,-0.2054876089,0.0119940396,0.2230312675,0.3162436187,0.2151381969,-0.2596668601,-0.1451510787,-0.1321804971,-0.1955112368,0.0069593326,-0.5726137161,0.4568421841,-0.24419415,-0.2051525563,-0.0859806389,-0.2658783793,0.4713196158,0.3780293763,-0.0684937015,-0.0375404693,0.1413383931,0.0351820067,0.177708894,-0.0622194856,0.2964116931,-0.2141356468,0.0457051024,-0.0277902856,-0.0081237135,0.0785258338,0.3201446831,0.2304375917,-0.0265186951,-0.2571677864,0.110347338,0.0635766909,0.1552903205,0.0463344343,0.4367932975,0.005378745,0.1145690829,0.0847834796,0.0224085972,-0.164030239,0.0601466,0.233269453,-0.0009895281,0.2357880324,-0.1547724605,-0.072097823,0.1263714284,-0.1329991817,-0.215120554,0.5398620367,-0.1397541612,0.4105234146,-0.1709616035,0.0268179663,-0.0614328459,-0.4580695629,-0.257130444,0.086149618,0.1797157824,0.155405134,-0.6861747503,0.3985427618,-0.3888285756,-0.0206326935,0.1728511304,0.5378324389,0.2120253146,0.0804060176,0.1571722478,-0.0373992324,0.3112249672,-0.2897289991,-0.1462424099,0.1092613339,0.1001149788,-0.3250236809,-0.0430728868,-0.2508806586,0.0265304223,0.4807166457,0.4030665755,-0.3020348847,-0.4488573372,-0.0142154954,0.0474353395,-0.2095984071,-0.1977830976,-0.2662701309,-0.4505195916,-0.0922009498,-0.0910787657,0.5540920496,-0.0119683752,0.285989821,0.0289197415,-0.2357486933,0.2654157579,0.3301908374,-0.0371142514,0.5644020438,-0.2395819277,0.0057377298,-0.0657557547,-0.0843305588,0.2901473939,0.4803256691,-0.18594037,-0.1504533291,0.1800045818,-0.2238452286,0.4223891199,0.1663793623,0.1885036528,0.0188958589,-0.2950081825,-0.0011859048,-0.1358927488,0.1362023652,0.3885795772,-0.1082977429,-0.4188079834,-0.3653247952,0.1768360138,0.3088775873,0.1055980548,-0.2310434431,-0.1721609235,-0.2617760301,-0.1415586323,0.4663684666,1.016680479,0.030185407,0.433570534,0.1726903319,0.0897267163,0.0607269071,-0.444486618,0.2049940228,-0.3482038677,-0.3285882473,-0.2036454231,-0.2574140429,-0.0620909221,0.2026154697,0.0319126844,0.6626021862,0.2256908417,-0.0447899438,0.0229511857,0.2159696668,0.0093683628,-0.3867395818,-0.3745247722,0.0040317033,-0.2863345742,0.0149859404,-0.1535770446,-0.2083832622,-0.0591487885,0.1994757652,0.1399660408,0.2819466293,0.0934466422,-0.1829066277,0.4239276946,-0.5605942607,-0.0533471443,0.2428724915,0.077303797,-0.322191,0.0031517914,0.2361234426,0.4050419629,0.1009970233,0.3913243711,-0.3321993649,0.2601027191,0.0498278886,0.1221913695,-0.2428604513,0.1581598967,-0.0372625589,-0.4174329937,0.0479725488,-0.1750261933,-0.2749729753,-0.2087987065,0.0269379523,-0.1443690658,-0.049738355,0.0549408123,0.2594358027,-0.0902868882,0.148101449,0.1132918224,-0.2592061758,0.1990920603,0.0831901953,0.2976744771,-0.3125315309,0.0376100764,0.2482126951,0.0151349036,-0.0992361903,0.180581525,0.4821507633,-0.617361784,0.0337477662,-0.2434498519,-0.0137992604,-0.0163004398,0.1851167381,0.0020547709,-0.2624997497,0.1627749503,-0.2828570008,-0.3409923017,0.3751991391,0.1163776442,0.2782913744,-0.1979854256,0.0098260185,0.1509936601,-0.2200442106,-0.2390500307,0.1753156036,-0.1794706732,0.1961244643,-0.1157648787,-0.096291475,0.080403842,-0.0942012146,-0.0249693338,0.2664697766,-0.1458446831,-0.0578539148,0.0179341715,0.1243914366,0.1327522844,-0.1157702655,0.1248916909,0.1200044379,-0.0515412576,-0.1373377591,-0.1135678813,0.1800017059,0.0228337441,0.2303531468,-0.2976282835,0.0392743871,-0.1231651306,0.2310590148,0.189538762,0.1202059165,0.0657504871,0.2072514594,-0.0511192568,0.0248664636,-0.5182061791,0.2064265907,0.1967262477,0.3688436747,0.2151999176,-0.1553563476,0.1056381837,-0.032278128,0.1498689055,0.3083644211,-0.2699720263,-0.0058445805,0.241904065,0.1483671665,-0.1588442177,-0.0049726604,0.3671055734,0.0078911902,0.008068773,0.4790737331,0.3525026143,0.136815384,-0.1463459283,0.0633270293,0.2355753332,-0.2535267174,0.0676125661,0.4052120447,-0.005103806,0.1394571364,0.5995871425,0.0332877487,-0.1143290102,0.6000959277,0.3011852205,0.7809863091,0.199559778,0.1787547022,-0.0658113658,-0.1044648662,-0.4129894078,0.2548288405,0.1273963898,0.1769531816,0.0383527577,0.143287167,0.1453569233,-0.1860472709,-0.0964768156,0.113500081,-0.3404924273,0.0229304079,-0.331787169,-0.2342431843,-0.1217740625,-0.1177005768,-0.1722740084,0.0367383994,0.061489068,0.1339388788,-0.0074618193,-0.0612117462,0.0823539197,0.2852355242,0.0478237383,-0.1977075338,0.5288455486,0.0339015722,-0.0469572246,-0.1587117612,0.2742477059,0.2531481087,0.2315032333,-0.0519730113,-0.1070040986,0.3387274742,-0.1102280021,-0.02969338,0.1004743874,0.2005059421,0.2549166381,0.1804338545,0.0651821718,-0.0043521184,0.3010003567,0.0695356131,-0.1928669661,0.2435857952,0.2801603675,0.1861452311,-0.2890454829,-0.1294614524,-0.2241025269,-0.3388958275,-0.1642102152,0.314212203,-0.1428406239,0.2404812127,-0.130814895,0.0508457981,0.042049367,0.1968667656,0.1187227666,0.1140162945,-0.4945984483,-0.0885237828,-0.7457126975,0.2005533725,-0.3096918166,-0.2826579809,0.2355704159,0.0266045816,-0.0474609956,0.3022997379,-0.23467426,0.2289671898,0.0685480013,0.334798485,-0.6064161658,0.0685795322,0.2317826003,0.35899809,0.0531915426,-0.124981299,-0.276468724,0.2186534852,-0.0702891573,-0.0727587193,-0.0685441419,-0.1896738708,-0.0558257475,0.2673411965,-0.212423116,0.1567766815,-0.0746062472,-0.1744633466,-0.2747464776,-0.0597104505,-0.1644858569,0.0210557934,0.1005385742,0.3167755306,0.0033465335,0.1547871828,-0.3915068209,-0.2984790206,0.0132773137,0.0827491954,-0.3641844988,0.4328106642,-0.2321629971,0.1437955201,0.0377221107,0.2259716541,0.074632287,-0.0033203885,-0.1593918055,-0.4455848634,0.2613032758,0.2699011862,-0.1986567527,-0.1195013747,0.0043353084,0.1073929965,0.1258694679,-0.4479656518,0.1364092082,0.1799704432,-0.0451218039,-0.0945755392,0.0839961171,-0.0237168744,-0.0781790763,-0.1301056743,0.1238042712,-0.0262836535,-0.1145851761,-0.2040123045,-0.1585304439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/425","title":"Correct data structure for PAN-X task in XTREME dataset?","comments":"Thanks, I am making progress. I got a new error `NonMatchingSplitsSizesError ` (see traceback below), which I suspect is due to the fact that number of rows in the dataset changed (one row per word --> one row per sentence) as well as the number of bytes due to the slightly updated data structure. \r\n\r\n```python\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=1756492, num_examples=80536, dataset_name='xtreme'), 'recorded': SplitInfo(name='validation', num_bytes=1837109, num_examples=10000, dataset_name='xtreme')}, {'expected': SplitInfo(name='test', num_bytes=1752572, num_examples=80326, dataset_name='xtreme'), 'recorded': SplitInfo(name='test', num_bytes=1833214, num_examples=10000, dataset_name='xtreme')}, {'expected': SplitInfo(name='train', num_bytes=3496832, num_examples=160394, dataset_name='xtreme'), 'recorded': SplitInfo(name='train', num_bytes=3658428, num_examples=20000, dataset_name='xtreme')}]\r\n```\r\nI can fix the error by replacing the values in the `datasets_infos.json` file, which I tested for English. However, to update this for all 40 datasets manually is slightly painful. Is there a better way to update the expected values for all datasets?","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.","comment_length":130,"text":"Correct data structure for PAN-X task in XTREME dataset?\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.\nThanks, I am making progress. I got a new error `NonMatchingSplitsSizesError ` (see traceback below), which I suspect is due to the fact that number of rows in the dataset changed (one row per word --> one row per sentence) as well as the number of bytes due to the slightly updated data structure. \r\n\r\n```python\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='validation', num_bytes=1756492, num_examples=80536, dataset_name='xtreme'), 'recorded': SplitInfo(name='validation', num_bytes=1837109, num_examples=10000, dataset_name='xtreme')}, {'expected': SplitInfo(name='test', num_bytes=1752572, num_examples=80326, dataset_name='xtreme'), 'recorded': SplitInfo(name='test', num_bytes=1833214, num_examples=10000, dataset_name='xtreme')}, {'expected': SplitInfo(name='train', num_bytes=3496832, num_examples=160394, dataset_name='xtreme'), 'recorded': SplitInfo(name='train', num_bytes=3658428, num_examples=20000, dataset_name='xtreme')}]\r\n```\r\nI can fix the error by replacing the values in the `datasets_infos.json` file, which I tested for English. However, to update this for all 40 datasets manually is slightly painful. Is there a better way to update the expected values for all datasets?","embeddings":[-0.0446428172,-0.1960519552,-0.0213917121,0.3716609478,0.0390401855,-0.2763307691,-0.0206372105,0.1579513103,-0.0717805475,-0.0662426949,-0.1613530368,0.2627067268,0.0014278574,0.1680819094,0.4264350533,-0.1467018872,0.0484421924,-0.0699055418,0.2786860168,-0.0526146963,-0.109093152,0.1506099999,-0.3253746927,0.0544689558,-0.3497874737,0.0837271884,-0.0678211153,0.3462400138,-0.2758393288,-0.493368417,0.5901547074,0.0610429049,0.3375472426,0.3253481388,-0.0001185823,0.007394406,0.0491500795,-0.2976795733,0.063000001,-0.5153079629,-0.21160613,-0.6882328987,-0.0707507059,-0.1863514483,-0.249424383,-0.0691500232,0.2359091789,0.1613155603,0.3286878169,0.404129982,0.0667554885,0.1186724603,-0.0112289721,0.1646236479,0.4069048762,0.3912606835,-0.2984111607,-0.0554718524,0.5836415291,0.1500933468,-0.2007237524,0.226136744,0.0665774941,-0.0957404003,0.2168319821,0.311748594,0.1106207892,-0.4583888352,-0.2030344903,0.2828918993,0.2304347605,-0.2553840578,-0.2035078704,-0.2779316902,-0.0572837256,-0.2900469899,-0.0280274134,0.3609626591,-0.0808850378,-0.0090993838,0.2543638647,-0.0029822402,-0.1114807874,0.2031368911,-0.0105202226,0.4096223116,-0.0134635018,-0.1075180918,-0.1396067441,-0.1162076294,0.7018569112,-0.0104258293,0.1576858014,0.1197108179,-0.6041291952,0.0803796276,-0.1363217384,-0.2818925083,0.1663054824,-0.0484371334,0.2098659426,0.1589148641,0.148524031,0.0145868212,0.3213757873,0.2003956288,0.2381373495,0.0310105663,-0.1307128817,0.0734806508,0.0858036652,-0.0963578075,-0.3183752894,0.0653057396,0.2368329167,-0.0228066426,0.2889071107,-0.1962373108,-0.3680241704,-0.0199918374,-0.1284303069,0.072580345,0.1414791495,0.061180193,0.0080608483,0.5874376297,-0.0742886215,0.1787090451,-0.2538342476,0.0291408673,-0.1472996622,0.1694096178,-0.6243599057,-0.1061867177,0.2582582831,0.091962114,0.1970394701,-0.0064032744,0.121820122,-0.3322490454,0.1758813262,-0.2791434526,0.2651101649,-0.0158670694,0.2457386106,-0.1295719892,-0.3093241453,-0.0095198667,-0.1647403091,0.2194372267,-0.7123686075,-0.0245730337,0.0450958833,0.0754593015,-0.214194268,0.1435741037,-0.504322648,0.0322593153,0.0630441457,0.2181688845,0.1069046035,-0.3306136727,-0.2142900974,-0.0720484778,0.0830698535,0.1326568276,-0.1291505694,0.0973811597,0.0630992949,0.0002566593,0.1041970924,0.2162794471,-0.0109667862,0.284181118,-0.1886345744,0.5723384023,0.4195818007,-0.8037276268,-0.3378547132,0.0875090212,-0.1656147987,0.0018321252,-0.0777145699,0.2999708056,0.2168296725,-0.1097091362,0.2860379517,0.4717787802,-0.0601560883,0.1129643992,-0.1055826619,0.2662775815,-0.1180936024,0.3215996325,0.0420975052,-0.2685292661,-0.1594627202,0.0751415715,0.1948614568,-0.1158283576,0.0870857239,0.4314538538,0.089273192,-0.0513841324,-0.0506237373,0.0383254662,-0.7624905109,0.1141011417,-0.4289596081,0.3793443143,0.0106568281,-0.0671950504,-0.1728422344,-0.0510757342,-0.296262145,-0.1906543225,0.107161127,-0.1209991425,-0.1237812713,-0.3034522235,-0.0762852654,0.3974147141,-0.2260980457,0.0737625808,-0.5780182481,0.340782702,0.1180048063,-0.0664847717,-0.0123105403,0.4472563863,0.1922498047,-0.0482966453,0.1684597582,0.2209219038,0.2148898542,-0.1571745276,-0.0612342209,-0.0020257682,0.1334930658,-0.5777854919,0.0659277737,0.2230994701,-0.001422982,-0.2143050879,-0.1243546456,0.5517559648,0.1847706735,0.2072845101,0.0289730113,0.0707501099,0.2122232467,-0.1713617146,-0.2344333678,-0.1039047167,0.1545492858,0.1233171746,0.0305474959,0.0995655283,-0.5023413897,0.0200132038,0.3077193201,0.0274304561,0.1898184419,0.2047269344,-0.1531089544,-0.0791647956,0.0374575406,0.3701202571,0.3590866625,0.0914273858,0.1098857149,0.1304076463,-0.0329413153,0.0724881813,-0.0358842276,0.0617540143,0.1510176659,0.2119652778,0.1785159111,-0.0841950774,-0.0858964771,-0.2191504538,0.0321776755,0.152163282,-0.3294998705,0.0780693293,0.0368220694,-0.450265348,-0.1473740041,0.0441721193,-0.1361576915,-0.3193361759,-0.0379218347,-0.0571559481,0.2387743443,0.084067151,-0.4746881425,-0.2150861621,-0.1986020505,-0.3406288028,-0.0583148114,0.0742932335,-0.1382389069,-0.0267540924,0.2119731903,0.5459074378,-0.031271968,-0.4353145659,0.3901607394,-0.3872416317,-0.2023112923,0.301656425,-0.2638221681,0.3121575713,0.217154935,-0.1040859297,-0.4046024084,0.195917204,0.1874918491,-0.321480006,-0.3141257763,0.3168615997,-0.1772629917,-0.0916748643,-0.0264440607,-0.40634498,-0.5237715244,-0.2054876089,0.0119940396,0.2230312675,0.3162436187,0.2151381969,-0.2596668601,-0.1451510787,-0.1321804971,-0.1955112368,0.0069593326,-0.5726137161,0.4568421841,-0.24419415,-0.2051525563,-0.0859806389,-0.2658783793,0.4713196158,0.3780293763,-0.0684937015,-0.0375404693,0.1413383931,0.0351820067,0.177708894,-0.0622194856,0.2964116931,-0.2141356468,0.0457051024,-0.0277902856,-0.0081237135,0.0785258338,0.3201446831,0.2304375917,-0.0265186951,-0.2571677864,0.110347338,0.0635766909,0.1552903205,0.0463344343,0.4367932975,0.005378745,0.1145690829,0.0847834796,0.0224085972,-0.164030239,0.0601466,0.233269453,-0.0009895281,0.2357880324,-0.1547724605,-0.072097823,0.1263714284,-0.1329991817,-0.215120554,0.5398620367,-0.1397541612,0.4105234146,-0.1709616035,0.0268179663,-0.0614328459,-0.4580695629,-0.257130444,0.086149618,0.1797157824,0.155405134,-0.6861747503,0.3985427618,-0.3888285756,-0.0206326935,0.1728511304,0.5378324389,0.2120253146,0.0804060176,0.1571722478,-0.0373992324,0.3112249672,-0.2897289991,-0.1462424099,0.1092613339,0.1001149788,-0.3250236809,-0.0430728868,-0.2508806586,0.0265304223,0.4807166457,0.4030665755,-0.3020348847,-0.4488573372,-0.0142154954,0.0474353395,-0.2095984071,-0.1977830976,-0.2662701309,-0.4505195916,-0.0922009498,-0.0910787657,0.5540920496,-0.0119683752,0.285989821,0.0289197415,-0.2357486933,0.2654157579,0.3301908374,-0.0371142514,0.5644020438,-0.2395819277,0.0057377298,-0.0657557547,-0.0843305588,0.2901473939,0.4803256691,-0.18594037,-0.1504533291,0.1800045818,-0.2238452286,0.4223891199,0.1663793623,0.1885036528,0.0188958589,-0.2950081825,-0.0011859048,-0.1358927488,0.1362023652,0.3885795772,-0.1082977429,-0.4188079834,-0.3653247952,0.1768360138,0.3088775873,0.1055980548,-0.2310434431,-0.1721609235,-0.2617760301,-0.1415586323,0.4663684666,1.016680479,0.030185407,0.433570534,0.1726903319,0.0897267163,0.0607269071,-0.444486618,0.2049940228,-0.3482038677,-0.3285882473,-0.2036454231,-0.2574140429,-0.0620909221,0.2026154697,0.0319126844,0.6626021862,0.2256908417,-0.0447899438,0.0229511857,0.2159696668,0.0093683628,-0.3867395818,-0.3745247722,0.0040317033,-0.2863345742,0.0149859404,-0.1535770446,-0.2083832622,-0.0591487885,0.1994757652,0.1399660408,0.2819466293,0.0934466422,-0.1829066277,0.4239276946,-0.5605942607,-0.0533471443,0.2428724915,0.077303797,-0.322191,0.0031517914,0.2361234426,0.4050419629,0.1009970233,0.3913243711,-0.3321993649,0.2601027191,0.0498278886,0.1221913695,-0.2428604513,0.1581598967,-0.0372625589,-0.4174329937,0.0479725488,-0.1750261933,-0.2749729753,-0.2087987065,0.0269379523,-0.1443690658,-0.049738355,0.0549408123,0.2594358027,-0.0902868882,0.148101449,0.1132918224,-0.2592061758,0.1990920603,0.0831901953,0.2976744771,-0.3125315309,0.0376100764,0.2482126951,0.0151349036,-0.0992361903,0.180581525,0.4821507633,-0.617361784,0.0337477662,-0.2434498519,-0.0137992604,-0.0163004398,0.1851167381,0.0020547709,-0.2624997497,0.1627749503,-0.2828570008,-0.3409923017,0.3751991391,0.1163776442,0.2782913744,-0.1979854256,0.0098260185,0.1509936601,-0.2200442106,-0.2390500307,0.1753156036,-0.1794706732,0.1961244643,-0.1157648787,-0.096291475,0.080403842,-0.0942012146,-0.0249693338,0.2664697766,-0.1458446831,-0.0578539148,0.0179341715,0.1243914366,0.1327522844,-0.1157702655,0.1248916909,0.1200044379,-0.0515412576,-0.1373377591,-0.1135678813,0.1800017059,0.0228337441,0.2303531468,-0.2976282835,0.0392743871,-0.1231651306,0.2310590148,0.189538762,0.1202059165,0.0657504871,0.2072514594,-0.0511192568,0.0248664636,-0.5182061791,0.2064265907,0.1967262477,0.3688436747,0.2151999176,-0.1553563476,0.1056381837,-0.032278128,0.1498689055,0.3083644211,-0.2699720263,-0.0058445805,0.241904065,0.1483671665,-0.1588442177,-0.0049726604,0.3671055734,0.0078911902,0.008068773,0.4790737331,0.3525026143,0.136815384,-0.1463459283,0.0633270293,0.2355753332,-0.2535267174,0.0676125661,0.4052120447,-0.005103806,0.1394571364,0.5995871425,0.0332877487,-0.1143290102,0.6000959277,0.3011852205,0.7809863091,0.199559778,0.1787547022,-0.0658113658,-0.1044648662,-0.4129894078,0.2548288405,0.1273963898,0.1769531816,0.0383527577,0.143287167,0.1453569233,-0.1860472709,-0.0964768156,0.113500081,-0.3404924273,0.0229304079,-0.331787169,-0.2342431843,-0.1217740625,-0.1177005768,-0.1722740084,0.0367383994,0.061489068,0.1339388788,-0.0074618193,-0.0612117462,0.0823539197,0.2852355242,0.0478237383,-0.1977075338,0.5288455486,0.0339015722,-0.0469572246,-0.1587117612,0.2742477059,0.2531481087,0.2315032333,-0.0519730113,-0.1070040986,0.3387274742,-0.1102280021,-0.02969338,0.1004743874,0.2005059421,0.2549166381,0.1804338545,0.0651821718,-0.0043521184,0.3010003567,0.0695356131,-0.1928669661,0.2435857952,0.2801603675,0.1861452311,-0.2890454829,-0.1294614524,-0.2241025269,-0.3388958275,-0.1642102152,0.314212203,-0.1428406239,0.2404812127,-0.130814895,0.0508457981,0.042049367,0.1968667656,0.1187227666,0.1140162945,-0.4945984483,-0.0885237828,-0.7457126975,0.2005533725,-0.3096918166,-0.2826579809,0.2355704159,0.0266045816,-0.0474609956,0.3022997379,-0.23467426,0.2289671898,0.0685480013,0.334798485,-0.6064161658,0.0685795322,0.2317826003,0.35899809,0.0531915426,-0.124981299,-0.276468724,0.2186534852,-0.0702891573,-0.0727587193,-0.0685441419,-0.1896738708,-0.0558257475,0.2673411965,-0.212423116,0.1567766815,-0.0746062472,-0.1744633466,-0.2747464776,-0.0597104505,-0.1644858569,0.0210557934,0.1005385742,0.3167755306,0.0033465335,0.1547871828,-0.3915068209,-0.2984790206,0.0132773137,0.0827491954,-0.3641844988,0.4328106642,-0.2321629971,0.1437955201,0.0377221107,0.2259716541,0.074632287,-0.0033203885,-0.1593918055,-0.4455848634,0.2613032758,0.2699011862,-0.1986567527,-0.1195013747,0.0043353084,0.1073929965,0.1258694679,-0.4479656518,0.1364092082,0.1799704432,-0.0451218039,-0.0945755392,0.0839961171,-0.0237168744,-0.0781790763,-0.1301056743,0.1238042712,-0.0262836535,-0.1145851761,-0.2040123045,-0.1585304439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/425","title":"Correct data structure for PAN-X task in XTREME dataset?","comments":"You can update the json file by calling\r\n```\r\nnlp-cli test .\/datasets\/xtreme --save_infos --all_configs\r\n```","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.","comment_length":15,"text":"Correct data structure for PAN-X task in XTREME dataset?\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.\nYou can update the json file by calling\r\n```\r\nnlp-cli test .\/datasets\/xtreme --save_infos --all_configs\r\n```","embeddings":[-0.0446428172,-0.1960519552,-0.0213917121,0.3716609478,0.0390401855,-0.2763307691,-0.0206372105,0.1579513103,-0.0717805475,-0.0662426949,-0.1613530368,0.2627067268,0.0014278574,0.1680819094,0.4264350533,-0.1467018872,0.0484421924,-0.0699055418,0.2786860168,-0.0526146963,-0.109093152,0.1506099999,-0.3253746927,0.0544689558,-0.3497874737,0.0837271884,-0.0678211153,0.3462400138,-0.2758393288,-0.493368417,0.5901547074,0.0610429049,0.3375472426,0.3253481388,-0.0001185823,0.007394406,0.0491500795,-0.2976795733,0.063000001,-0.5153079629,-0.21160613,-0.6882328987,-0.0707507059,-0.1863514483,-0.249424383,-0.0691500232,0.2359091789,0.1613155603,0.3286878169,0.404129982,0.0667554885,0.1186724603,-0.0112289721,0.1646236479,0.4069048762,0.3912606835,-0.2984111607,-0.0554718524,0.5836415291,0.1500933468,-0.2007237524,0.226136744,0.0665774941,-0.0957404003,0.2168319821,0.311748594,0.1106207892,-0.4583888352,-0.2030344903,0.2828918993,0.2304347605,-0.2553840578,-0.2035078704,-0.2779316902,-0.0572837256,-0.2900469899,-0.0280274134,0.3609626591,-0.0808850378,-0.0090993838,0.2543638647,-0.0029822402,-0.1114807874,0.2031368911,-0.0105202226,0.4096223116,-0.0134635018,-0.1075180918,-0.1396067441,-0.1162076294,0.7018569112,-0.0104258293,0.1576858014,0.1197108179,-0.6041291952,0.0803796276,-0.1363217384,-0.2818925083,0.1663054824,-0.0484371334,0.2098659426,0.1589148641,0.148524031,0.0145868212,0.3213757873,0.2003956288,0.2381373495,0.0310105663,-0.1307128817,0.0734806508,0.0858036652,-0.0963578075,-0.3183752894,0.0653057396,0.2368329167,-0.0228066426,0.2889071107,-0.1962373108,-0.3680241704,-0.0199918374,-0.1284303069,0.072580345,0.1414791495,0.061180193,0.0080608483,0.5874376297,-0.0742886215,0.1787090451,-0.2538342476,0.0291408673,-0.1472996622,0.1694096178,-0.6243599057,-0.1061867177,0.2582582831,0.091962114,0.1970394701,-0.0064032744,0.121820122,-0.3322490454,0.1758813262,-0.2791434526,0.2651101649,-0.0158670694,0.2457386106,-0.1295719892,-0.3093241453,-0.0095198667,-0.1647403091,0.2194372267,-0.7123686075,-0.0245730337,0.0450958833,0.0754593015,-0.214194268,0.1435741037,-0.504322648,0.0322593153,0.0630441457,0.2181688845,0.1069046035,-0.3306136727,-0.2142900974,-0.0720484778,0.0830698535,0.1326568276,-0.1291505694,0.0973811597,0.0630992949,0.0002566593,0.1041970924,0.2162794471,-0.0109667862,0.284181118,-0.1886345744,0.5723384023,0.4195818007,-0.8037276268,-0.3378547132,0.0875090212,-0.1656147987,0.0018321252,-0.0777145699,0.2999708056,0.2168296725,-0.1097091362,0.2860379517,0.4717787802,-0.0601560883,0.1129643992,-0.1055826619,0.2662775815,-0.1180936024,0.3215996325,0.0420975052,-0.2685292661,-0.1594627202,0.0751415715,0.1948614568,-0.1158283576,0.0870857239,0.4314538538,0.089273192,-0.0513841324,-0.0506237373,0.0383254662,-0.7624905109,0.1141011417,-0.4289596081,0.3793443143,0.0106568281,-0.0671950504,-0.1728422344,-0.0510757342,-0.296262145,-0.1906543225,0.107161127,-0.1209991425,-0.1237812713,-0.3034522235,-0.0762852654,0.3974147141,-0.2260980457,0.0737625808,-0.5780182481,0.340782702,0.1180048063,-0.0664847717,-0.0123105403,0.4472563863,0.1922498047,-0.0482966453,0.1684597582,0.2209219038,0.2148898542,-0.1571745276,-0.0612342209,-0.0020257682,0.1334930658,-0.5777854919,0.0659277737,0.2230994701,-0.001422982,-0.2143050879,-0.1243546456,0.5517559648,0.1847706735,0.2072845101,0.0289730113,0.0707501099,0.2122232467,-0.1713617146,-0.2344333678,-0.1039047167,0.1545492858,0.1233171746,0.0305474959,0.0995655283,-0.5023413897,0.0200132038,0.3077193201,0.0274304561,0.1898184419,0.2047269344,-0.1531089544,-0.0791647956,0.0374575406,0.3701202571,0.3590866625,0.0914273858,0.1098857149,0.1304076463,-0.0329413153,0.0724881813,-0.0358842276,0.0617540143,0.1510176659,0.2119652778,0.1785159111,-0.0841950774,-0.0858964771,-0.2191504538,0.0321776755,0.152163282,-0.3294998705,0.0780693293,0.0368220694,-0.450265348,-0.1473740041,0.0441721193,-0.1361576915,-0.3193361759,-0.0379218347,-0.0571559481,0.2387743443,0.084067151,-0.4746881425,-0.2150861621,-0.1986020505,-0.3406288028,-0.0583148114,0.0742932335,-0.1382389069,-0.0267540924,0.2119731903,0.5459074378,-0.031271968,-0.4353145659,0.3901607394,-0.3872416317,-0.2023112923,0.301656425,-0.2638221681,0.3121575713,0.217154935,-0.1040859297,-0.4046024084,0.195917204,0.1874918491,-0.321480006,-0.3141257763,0.3168615997,-0.1772629917,-0.0916748643,-0.0264440607,-0.40634498,-0.5237715244,-0.2054876089,0.0119940396,0.2230312675,0.3162436187,0.2151381969,-0.2596668601,-0.1451510787,-0.1321804971,-0.1955112368,0.0069593326,-0.5726137161,0.4568421841,-0.24419415,-0.2051525563,-0.0859806389,-0.2658783793,0.4713196158,0.3780293763,-0.0684937015,-0.0375404693,0.1413383931,0.0351820067,0.177708894,-0.0622194856,0.2964116931,-0.2141356468,0.0457051024,-0.0277902856,-0.0081237135,0.0785258338,0.3201446831,0.2304375917,-0.0265186951,-0.2571677864,0.110347338,0.0635766909,0.1552903205,0.0463344343,0.4367932975,0.005378745,0.1145690829,0.0847834796,0.0224085972,-0.164030239,0.0601466,0.233269453,-0.0009895281,0.2357880324,-0.1547724605,-0.072097823,0.1263714284,-0.1329991817,-0.215120554,0.5398620367,-0.1397541612,0.4105234146,-0.1709616035,0.0268179663,-0.0614328459,-0.4580695629,-0.257130444,0.086149618,0.1797157824,0.155405134,-0.6861747503,0.3985427618,-0.3888285756,-0.0206326935,0.1728511304,0.5378324389,0.2120253146,0.0804060176,0.1571722478,-0.0373992324,0.3112249672,-0.2897289991,-0.1462424099,0.1092613339,0.1001149788,-0.3250236809,-0.0430728868,-0.2508806586,0.0265304223,0.4807166457,0.4030665755,-0.3020348847,-0.4488573372,-0.0142154954,0.0474353395,-0.2095984071,-0.1977830976,-0.2662701309,-0.4505195916,-0.0922009498,-0.0910787657,0.5540920496,-0.0119683752,0.285989821,0.0289197415,-0.2357486933,0.2654157579,0.3301908374,-0.0371142514,0.5644020438,-0.2395819277,0.0057377298,-0.0657557547,-0.0843305588,0.2901473939,0.4803256691,-0.18594037,-0.1504533291,0.1800045818,-0.2238452286,0.4223891199,0.1663793623,0.1885036528,0.0188958589,-0.2950081825,-0.0011859048,-0.1358927488,0.1362023652,0.3885795772,-0.1082977429,-0.4188079834,-0.3653247952,0.1768360138,0.3088775873,0.1055980548,-0.2310434431,-0.1721609235,-0.2617760301,-0.1415586323,0.4663684666,1.016680479,0.030185407,0.433570534,0.1726903319,0.0897267163,0.0607269071,-0.444486618,0.2049940228,-0.3482038677,-0.3285882473,-0.2036454231,-0.2574140429,-0.0620909221,0.2026154697,0.0319126844,0.6626021862,0.2256908417,-0.0447899438,0.0229511857,0.2159696668,0.0093683628,-0.3867395818,-0.3745247722,0.0040317033,-0.2863345742,0.0149859404,-0.1535770446,-0.2083832622,-0.0591487885,0.1994757652,0.1399660408,0.2819466293,0.0934466422,-0.1829066277,0.4239276946,-0.5605942607,-0.0533471443,0.2428724915,0.077303797,-0.322191,0.0031517914,0.2361234426,0.4050419629,0.1009970233,0.3913243711,-0.3321993649,0.2601027191,0.0498278886,0.1221913695,-0.2428604513,0.1581598967,-0.0372625589,-0.4174329937,0.0479725488,-0.1750261933,-0.2749729753,-0.2087987065,0.0269379523,-0.1443690658,-0.049738355,0.0549408123,0.2594358027,-0.0902868882,0.148101449,0.1132918224,-0.2592061758,0.1990920603,0.0831901953,0.2976744771,-0.3125315309,0.0376100764,0.2482126951,0.0151349036,-0.0992361903,0.180581525,0.4821507633,-0.617361784,0.0337477662,-0.2434498519,-0.0137992604,-0.0163004398,0.1851167381,0.0020547709,-0.2624997497,0.1627749503,-0.2828570008,-0.3409923017,0.3751991391,0.1163776442,0.2782913744,-0.1979854256,0.0098260185,0.1509936601,-0.2200442106,-0.2390500307,0.1753156036,-0.1794706732,0.1961244643,-0.1157648787,-0.096291475,0.080403842,-0.0942012146,-0.0249693338,0.2664697766,-0.1458446831,-0.0578539148,0.0179341715,0.1243914366,0.1327522844,-0.1157702655,0.1248916909,0.1200044379,-0.0515412576,-0.1373377591,-0.1135678813,0.1800017059,0.0228337441,0.2303531468,-0.2976282835,0.0392743871,-0.1231651306,0.2310590148,0.189538762,0.1202059165,0.0657504871,0.2072514594,-0.0511192568,0.0248664636,-0.5182061791,0.2064265907,0.1967262477,0.3688436747,0.2151999176,-0.1553563476,0.1056381837,-0.032278128,0.1498689055,0.3083644211,-0.2699720263,-0.0058445805,0.241904065,0.1483671665,-0.1588442177,-0.0049726604,0.3671055734,0.0078911902,0.008068773,0.4790737331,0.3525026143,0.136815384,-0.1463459283,0.0633270293,0.2355753332,-0.2535267174,0.0676125661,0.4052120447,-0.005103806,0.1394571364,0.5995871425,0.0332877487,-0.1143290102,0.6000959277,0.3011852205,0.7809863091,0.199559778,0.1787547022,-0.0658113658,-0.1044648662,-0.4129894078,0.2548288405,0.1273963898,0.1769531816,0.0383527577,0.143287167,0.1453569233,-0.1860472709,-0.0964768156,0.113500081,-0.3404924273,0.0229304079,-0.331787169,-0.2342431843,-0.1217740625,-0.1177005768,-0.1722740084,0.0367383994,0.061489068,0.1339388788,-0.0074618193,-0.0612117462,0.0823539197,0.2852355242,0.0478237383,-0.1977075338,0.5288455486,0.0339015722,-0.0469572246,-0.1587117612,0.2742477059,0.2531481087,0.2315032333,-0.0519730113,-0.1070040986,0.3387274742,-0.1102280021,-0.02969338,0.1004743874,0.2005059421,0.2549166381,0.1804338545,0.0651821718,-0.0043521184,0.3010003567,0.0695356131,-0.1928669661,0.2435857952,0.2801603675,0.1861452311,-0.2890454829,-0.1294614524,-0.2241025269,-0.3388958275,-0.1642102152,0.314212203,-0.1428406239,0.2404812127,-0.130814895,0.0508457981,0.042049367,0.1968667656,0.1187227666,0.1140162945,-0.4945984483,-0.0885237828,-0.7457126975,0.2005533725,-0.3096918166,-0.2826579809,0.2355704159,0.0266045816,-0.0474609956,0.3022997379,-0.23467426,0.2289671898,0.0685480013,0.334798485,-0.6064161658,0.0685795322,0.2317826003,0.35899809,0.0531915426,-0.124981299,-0.276468724,0.2186534852,-0.0702891573,-0.0727587193,-0.0685441419,-0.1896738708,-0.0558257475,0.2673411965,-0.212423116,0.1567766815,-0.0746062472,-0.1744633466,-0.2747464776,-0.0597104505,-0.1644858569,0.0210557934,0.1005385742,0.3167755306,0.0033465335,0.1547871828,-0.3915068209,-0.2984790206,0.0132773137,0.0827491954,-0.3641844988,0.4328106642,-0.2321629971,0.1437955201,0.0377221107,0.2259716541,0.074632287,-0.0033203885,-0.1593918055,-0.4455848634,0.2613032758,0.2699011862,-0.1986567527,-0.1195013747,0.0043353084,0.1073929965,0.1258694679,-0.4479656518,0.1364092082,0.1799704432,-0.0451218039,-0.0945755392,0.0839961171,-0.0237168744,-0.0781790763,-0.1301056743,0.1238042712,-0.0262836535,-0.1145851761,-0.2040123045,-0.1585304439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/425","title":"Correct data structure for PAN-X task in XTREME dataset?","comments":"One more thing about features. I mentioned\r\n\r\n```python\r\nfeatures = nlp.Features({\r\n \"words\": [nlp.Value(\"string\")],\r\n \"ner_tags\": [nlp.Value(\"string\")],\r\n \"langs\": [nlp.Value(\"string\")],\r\n})\r\n```\r\n\r\nbut it's actually not consistent with the way we write datasets. Something like this is simpler to read and more consistent with the way we define datasets:\r\n\r\n```python\r\nfeatures = nlp.Features({\r\n \"words\": nlp.Sequence(nlp.Value(\"string\")),\r\n \"ner_tags\": nlp.Sequence(nlp.Value(\"string\")),\r\n \"langs\": nlp.Sequence(nlp.Value(\"string\")),\r\n})\r\n```\r\n\r\nSorry about that","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.","comment_length":61,"text":"Correct data structure for PAN-X task in XTREME dataset?\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.\nOne more thing about features. I mentioned\r\n\r\n```python\r\nfeatures = nlp.Features({\r\n \"words\": [nlp.Value(\"string\")],\r\n \"ner_tags\": [nlp.Value(\"string\")],\r\n \"langs\": [nlp.Value(\"string\")],\r\n})\r\n```\r\n\r\nbut it's actually not consistent with the way we write datasets. Something like this is simpler to read and more consistent with the way we define datasets:\r\n\r\n```python\r\nfeatures = nlp.Features({\r\n \"words\": nlp.Sequence(nlp.Value(\"string\")),\r\n \"ner_tags\": nlp.Sequence(nlp.Value(\"string\")),\r\n \"langs\": nlp.Sequence(nlp.Value(\"string\")),\r\n})\r\n```\r\n\r\nSorry about that","embeddings":[-0.0446428172,-0.1960519552,-0.0213917121,0.3716609478,0.0390401855,-0.2763307691,-0.0206372105,0.1579513103,-0.0717805475,-0.0662426949,-0.1613530368,0.2627067268,0.0014278574,0.1680819094,0.4264350533,-0.1467018872,0.0484421924,-0.0699055418,0.2786860168,-0.0526146963,-0.109093152,0.1506099999,-0.3253746927,0.0544689558,-0.3497874737,0.0837271884,-0.0678211153,0.3462400138,-0.2758393288,-0.493368417,0.5901547074,0.0610429049,0.3375472426,0.3253481388,-0.0001185823,0.007394406,0.0491500795,-0.2976795733,0.063000001,-0.5153079629,-0.21160613,-0.6882328987,-0.0707507059,-0.1863514483,-0.249424383,-0.0691500232,0.2359091789,0.1613155603,0.3286878169,0.404129982,0.0667554885,0.1186724603,-0.0112289721,0.1646236479,0.4069048762,0.3912606835,-0.2984111607,-0.0554718524,0.5836415291,0.1500933468,-0.2007237524,0.226136744,0.0665774941,-0.0957404003,0.2168319821,0.311748594,0.1106207892,-0.4583888352,-0.2030344903,0.2828918993,0.2304347605,-0.2553840578,-0.2035078704,-0.2779316902,-0.0572837256,-0.2900469899,-0.0280274134,0.3609626591,-0.0808850378,-0.0090993838,0.2543638647,-0.0029822402,-0.1114807874,0.2031368911,-0.0105202226,0.4096223116,-0.0134635018,-0.1075180918,-0.1396067441,-0.1162076294,0.7018569112,-0.0104258293,0.1576858014,0.1197108179,-0.6041291952,0.0803796276,-0.1363217384,-0.2818925083,0.1663054824,-0.0484371334,0.2098659426,0.1589148641,0.148524031,0.0145868212,0.3213757873,0.2003956288,0.2381373495,0.0310105663,-0.1307128817,0.0734806508,0.0858036652,-0.0963578075,-0.3183752894,0.0653057396,0.2368329167,-0.0228066426,0.2889071107,-0.1962373108,-0.3680241704,-0.0199918374,-0.1284303069,0.072580345,0.1414791495,0.061180193,0.0080608483,0.5874376297,-0.0742886215,0.1787090451,-0.2538342476,0.0291408673,-0.1472996622,0.1694096178,-0.6243599057,-0.1061867177,0.2582582831,0.091962114,0.1970394701,-0.0064032744,0.121820122,-0.3322490454,0.1758813262,-0.2791434526,0.2651101649,-0.0158670694,0.2457386106,-0.1295719892,-0.3093241453,-0.0095198667,-0.1647403091,0.2194372267,-0.7123686075,-0.0245730337,0.0450958833,0.0754593015,-0.214194268,0.1435741037,-0.504322648,0.0322593153,0.0630441457,0.2181688845,0.1069046035,-0.3306136727,-0.2142900974,-0.0720484778,0.0830698535,0.1326568276,-0.1291505694,0.0973811597,0.0630992949,0.0002566593,0.1041970924,0.2162794471,-0.0109667862,0.284181118,-0.1886345744,0.5723384023,0.4195818007,-0.8037276268,-0.3378547132,0.0875090212,-0.1656147987,0.0018321252,-0.0777145699,0.2999708056,0.2168296725,-0.1097091362,0.2860379517,0.4717787802,-0.0601560883,0.1129643992,-0.1055826619,0.2662775815,-0.1180936024,0.3215996325,0.0420975052,-0.2685292661,-0.1594627202,0.0751415715,0.1948614568,-0.1158283576,0.0870857239,0.4314538538,0.089273192,-0.0513841324,-0.0506237373,0.0383254662,-0.7624905109,0.1141011417,-0.4289596081,0.3793443143,0.0106568281,-0.0671950504,-0.1728422344,-0.0510757342,-0.296262145,-0.1906543225,0.107161127,-0.1209991425,-0.1237812713,-0.3034522235,-0.0762852654,0.3974147141,-0.2260980457,0.0737625808,-0.5780182481,0.340782702,0.1180048063,-0.0664847717,-0.0123105403,0.4472563863,0.1922498047,-0.0482966453,0.1684597582,0.2209219038,0.2148898542,-0.1571745276,-0.0612342209,-0.0020257682,0.1334930658,-0.5777854919,0.0659277737,0.2230994701,-0.001422982,-0.2143050879,-0.1243546456,0.5517559648,0.1847706735,0.2072845101,0.0289730113,0.0707501099,0.2122232467,-0.1713617146,-0.2344333678,-0.1039047167,0.1545492858,0.1233171746,0.0305474959,0.0995655283,-0.5023413897,0.0200132038,0.3077193201,0.0274304561,0.1898184419,0.2047269344,-0.1531089544,-0.0791647956,0.0374575406,0.3701202571,0.3590866625,0.0914273858,0.1098857149,0.1304076463,-0.0329413153,0.0724881813,-0.0358842276,0.0617540143,0.1510176659,0.2119652778,0.1785159111,-0.0841950774,-0.0858964771,-0.2191504538,0.0321776755,0.152163282,-0.3294998705,0.0780693293,0.0368220694,-0.450265348,-0.1473740041,0.0441721193,-0.1361576915,-0.3193361759,-0.0379218347,-0.0571559481,0.2387743443,0.084067151,-0.4746881425,-0.2150861621,-0.1986020505,-0.3406288028,-0.0583148114,0.0742932335,-0.1382389069,-0.0267540924,0.2119731903,0.5459074378,-0.031271968,-0.4353145659,0.3901607394,-0.3872416317,-0.2023112923,0.301656425,-0.2638221681,0.3121575713,0.217154935,-0.1040859297,-0.4046024084,0.195917204,0.1874918491,-0.321480006,-0.3141257763,0.3168615997,-0.1772629917,-0.0916748643,-0.0264440607,-0.40634498,-0.5237715244,-0.2054876089,0.0119940396,0.2230312675,0.3162436187,0.2151381969,-0.2596668601,-0.1451510787,-0.1321804971,-0.1955112368,0.0069593326,-0.5726137161,0.4568421841,-0.24419415,-0.2051525563,-0.0859806389,-0.2658783793,0.4713196158,0.3780293763,-0.0684937015,-0.0375404693,0.1413383931,0.0351820067,0.177708894,-0.0622194856,0.2964116931,-0.2141356468,0.0457051024,-0.0277902856,-0.0081237135,0.0785258338,0.3201446831,0.2304375917,-0.0265186951,-0.2571677864,0.110347338,0.0635766909,0.1552903205,0.0463344343,0.4367932975,0.005378745,0.1145690829,0.0847834796,0.0224085972,-0.164030239,0.0601466,0.233269453,-0.0009895281,0.2357880324,-0.1547724605,-0.072097823,0.1263714284,-0.1329991817,-0.215120554,0.5398620367,-0.1397541612,0.4105234146,-0.1709616035,0.0268179663,-0.0614328459,-0.4580695629,-0.257130444,0.086149618,0.1797157824,0.155405134,-0.6861747503,0.3985427618,-0.3888285756,-0.0206326935,0.1728511304,0.5378324389,0.2120253146,0.0804060176,0.1571722478,-0.0373992324,0.3112249672,-0.2897289991,-0.1462424099,0.1092613339,0.1001149788,-0.3250236809,-0.0430728868,-0.2508806586,0.0265304223,0.4807166457,0.4030665755,-0.3020348847,-0.4488573372,-0.0142154954,0.0474353395,-0.2095984071,-0.1977830976,-0.2662701309,-0.4505195916,-0.0922009498,-0.0910787657,0.5540920496,-0.0119683752,0.285989821,0.0289197415,-0.2357486933,0.2654157579,0.3301908374,-0.0371142514,0.5644020438,-0.2395819277,0.0057377298,-0.0657557547,-0.0843305588,0.2901473939,0.4803256691,-0.18594037,-0.1504533291,0.1800045818,-0.2238452286,0.4223891199,0.1663793623,0.1885036528,0.0188958589,-0.2950081825,-0.0011859048,-0.1358927488,0.1362023652,0.3885795772,-0.1082977429,-0.4188079834,-0.3653247952,0.1768360138,0.3088775873,0.1055980548,-0.2310434431,-0.1721609235,-0.2617760301,-0.1415586323,0.4663684666,1.016680479,0.030185407,0.433570534,0.1726903319,0.0897267163,0.0607269071,-0.444486618,0.2049940228,-0.3482038677,-0.3285882473,-0.2036454231,-0.2574140429,-0.0620909221,0.2026154697,0.0319126844,0.6626021862,0.2256908417,-0.0447899438,0.0229511857,0.2159696668,0.0093683628,-0.3867395818,-0.3745247722,0.0040317033,-0.2863345742,0.0149859404,-0.1535770446,-0.2083832622,-0.0591487885,0.1994757652,0.1399660408,0.2819466293,0.0934466422,-0.1829066277,0.4239276946,-0.5605942607,-0.0533471443,0.2428724915,0.077303797,-0.322191,0.0031517914,0.2361234426,0.4050419629,0.1009970233,0.3913243711,-0.3321993649,0.2601027191,0.0498278886,0.1221913695,-0.2428604513,0.1581598967,-0.0372625589,-0.4174329937,0.0479725488,-0.1750261933,-0.2749729753,-0.2087987065,0.0269379523,-0.1443690658,-0.049738355,0.0549408123,0.2594358027,-0.0902868882,0.148101449,0.1132918224,-0.2592061758,0.1990920603,0.0831901953,0.2976744771,-0.3125315309,0.0376100764,0.2482126951,0.0151349036,-0.0992361903,0.180581525,0.4821507633,-0.617361784,0.0337477662,-0.2434498519,-0.0137992604,-0.0163004398,0.1851167381,0.0020547709,-0.2624997497,0.1627749503,-0.2828570008,-0.3409923017,0.3751991391,0.1163776442,0.2782913744,-0.1979854256,0.0098260185,0.1509936601,-0.2200442106,-0.2390500307,0.1753156036,-0.1794706732,0.1961244643,-0.1157648787,-0.096291475,0.080403842,-0.0942012146,-0.0249693338,0.2664697766,-0.1458446831,-0.0578539148,0.0179341715,0.1243914366,0.1327522844,-0.1157702655,0.1248916909,0.1200044379,-0.0515412576,-0.1373377591,-0.1135678813,0.1800017059,0.0228337441,0.2303531468,-0.2976282835,0.0392743871,-0.1231651306,0.2310590148,0.189538762,0.1202059165,0.0657504871,0.2072514594,-0.0511192568,0.0248664636,-0.5182061791,0.2064265907,0.1967262477,0.3688436747,0.2151999176,-0.1553563476,0.1056381837,-0.032278128,0.1498689055,0.3083644211,-0.2699720263,-0.0058445805,0.241904065,0.1483671665,-0.1588442177,-0.0049726604,0.3671055734,0.0078911902,0.008068773,0.4790737331,0.3525026143,0.136815384,-0.1463459283,0.0633270293,0.2355753332,-0.2535267174,0.0676125661,0.4052120447,-0.005103806,0.1394571364,0.5995871425,0.0332877487,-0.1143290102,0.6000959277,0.3011852205,0.7809863091,0.199559778,0.1787547022,-0.0658113658,-0.1044648662,-0.4129894078,0.2548288405,0.1273963898,0.1769531816,0.0383527577,0.143287167,0.1453569233,-0.1860472709,-0.0964768156,0.113500081,-0.3404924273,0.0229304079,-0.331787169,-0.2342431843,-0.1217740625,-0.1177005768,-0.1722740084,0.0367383994,0.061489068,0.1339388788,-0.0074618193,-0.0612117462,0.0823539197,0.2852355242,0.0478237383,-0.1977075338,0.5288455486,0.0339015722,-0.0469572246,-0.1587117612,0.2742477059,0.2531481087,0.2315032333,-0.0519730113,-0.1070040986,0.3387274742,-0.1102280021,-0.02969338,0.1004743874,0.2005059421,0.2549166381,0.1804338545,0.0651821718,-0.0043521184,0.3010003567,0.0695356131,-0.1928669661,0.2435857952,0.2801603675,0.1861452311,-0.2890454829,-0.1294614524,-0.2241025269,-0.3388958275,-0.1642102152,0.314212203,-0.1428406239,0.2404812127,-0.130814895,0.0508457981,0.042049367,0.1968667656,0.1187227666,0.1140162945,-0.4945984483,-0.0885237828,-0.7457126975,0.2005533725,-0.3096918166,-0.2826579809,0.2355704159,0.0266045816,-0.0474609956,0.3022997379,-0.23467426,0.2289671898,0.0685480013,0.334798485,-0.6064161658,0.0685795322,0.2317826003,0.35899809,0.0531915426,-0.124981299,-0.276468724,0.2186534852,-0.0702891573,-0.0727587193,-0.0685441419,-0.1896738708,-0.0558257475,0.2673411965,-0.212423116,0.1567766815,-0.0746062472,-0.1744633466,-0.2747464776,-0.0597104505,-0.1644858569,0.0210557934,0.1005385742,0.3167755306,0.0033465335,0.1547871828,-0.3915068209,-0.2984790206,0.0132773137,0.0827491954,-0.3641844988,0.4328106642,-0.2321629971,0.1437955201,0.0377221107,0.2259716541,0.074632287,-0.0033203885,-0.1593918055,-0.4455848634,0.2613032758,0.2699011862,-0.1986567527,-0.1195013747,0.0043353084,0.1073929965,0.1258694679,-0.4479656518,0.1364092082,0.1799704432,-0.0451218039,-0.0945755392,0.0839961171,-0.0237168744,-0.0781790763,-0.1301056743,0.1238042712,-0.0262836535,-0.1145851761,-0.2040123045,-0.1585304439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/425","title":"Correct data structure for PAN-X task in XTREME dataset?","comments":"Closing this since PR #437 fixed the problem and has been merged to `master`. ","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.","comment_length":14,"text":"Correct data structure for PAN-X task in XTREME dataset?\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nThanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows:\r\n\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is located in data\/\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.en\", data_dir='.\/data')\r\ndataset_train = dataset['train']\r\n```\r\n\r\nHowever, I am not sure that `load_dataset()` is returning the correct data structure for NER. \r\n\r\nCurrently, every row in `dataset_train` is of the form\r\n```python\r\n{'word': str, 'ner_tag': str, 'lang': str}\r\n```\r\nbut I think we actually want something like\r\n```python\r\n{'words': List[str], 'ner_tags': List[str], 'langs': List[str]}\r\n```\r\nso that each row corresponds to a _sequence_ of words associated with each example. With the current data structure I do not think it is possible to transform `dataset_train` into a form suitable for training because we do not know the boundaries between examples.\r\n\r\nIndeed, [this line](https:\/\/github.com\/google-research\/xtreme\/blob\/522434d1aece34131d997a97ce7e9242a51a688a\/third_party\/utils_tag.py#L58) in the XTREME repo, processes the texts as lists of sentences, tags, and languages.\r\n\r\n## Proposed solution\r\nReplace\r\n```python\r\nwith open(filepath) as f:\r\n data = csv.reader(f, delimiter=\"\\t\", quoting=csv.QUOTE_NONE)\r\n for id_, row in enumerate(data):\r\n if row:\r\n lang, word = row[0].split(\":\")[0], row[0].split(\":\")[1]\r\n tag = row[1]\r\n yield id_, {\"word\": word, \"ner_tag\": tag, \"lang\": lang}\r\n```\r\nfrom [these lines](https:\/\/github.com\/huggingface\/nlp\/blob\/ce7d3a1d630b78fe27188d1706f3ea980e8eec43\/datasets\/xtreme\/xtreme.py#L881-L887) of the `_generate_examples()` function with something like\r\n\r\n```python\r\nguid_index = 1\r\nwith open(filepath, encoding=\"utf-8\") as f:\r\n words = []\r\n ner_tags = []\r\n langs = []\r\n for line in f:\r\n if line.startswith(\"-DOCSTART-\") or line == \"\" or line == \"\\n\":\r\n if words:\r\n yield guid_index, {\"words\": words, \"ner_tags\": ner_tags, \"langs\": langs}\r\n guid_index += 1\r\n words = []\r\n ner_tags = []\r\n else:\r\n # pan-x data is tab separated\r\n splits = line.split(\"\\t\")\r\n # strip out en: prefix\r\n langs.append(splits[0][:2])\r\n words.append(splits[0][3:])\r\n if len(splits) > 1:\r\n labels.append(splits[-1].replace(\"\\n\", \"\"))\r\n else:\r\n # examples have no label in test set\r\n labels.append(\"O\")\r\n```\r\nIf you agree, me or @lvwerra would be happy to implement this and create a PR.\nClosing this since PR #437 fixed the problem and has been merged to `master`. ","embeddings":[-0.0446428172,-0.1960519552,-0.0213917121,0.3716609478,0.0390401855,-0.2763307691,-0.0206372105,0.1579513103,-0.0717805475,-0.0662426949,-0.1613530368,0.2627067268,0.0014278574,0.1680819094,0.4264350533,-0.1467018872,0.0484421924,-0.0699055418,0.2786860168,-0.0526146963,-0.109093152,0.1506099999,-0.3253746927,0.0544689558,-0.3497874737,0.0837271884,-0.0678211153,0.3462400138,-0.2758393288,-0.493368417,0.5901547074,0.0610429049,0.3375472426,0.3253481388,-0.0001185823,0.007394406,0.0491500795,-0.2976795733,0.063000001,-0.5153079629,-0.21160613,-0.6882328987,-0.0707507059,-0.1863514483,-0.249424383,-0.0691500232,0.2359091789,0.1613155603,0.3286878169,0.404129982,0.0667554885,0.1186724603,-0.0112289721,0.1646236479,0.4069048762,0.3912606835,-0.2984111607,-0.0554718524,0.5836415291,0.1500933468,-0.2007237524,0.226136744,0.0665774941,-0.0957404003,0.2168319821,0.311748594,0.1106207892,-0.4583888352,-0.2030344903,0.2828918993,0.2304347605,-0.2553840578,-0.2035078704,-0.2779316902,-0.0572837256,-0.2900469899,-0.0280274134,0.3609626591,-0.0808850378,-0.0090993838,0.2543638647,-0.0029822402,-0.1114807874,0.2031368911,-0.0105202226,0.4096223116,-0.0134635018,-0.1075180918,-0.1396067441,-0.1162076294,0.7018569112,-0.0104258293,0.1576858014,0.1197108179,-0.6041291952,0.0803796276,-0.1363217384,-0.2818925083,0.1663054824,-0.0484371334,0.2098659426,0.1589148641,0.148524031,0.0145868212,0.3213757873,0.2003956288,0.2381373495,0.0310105663,-0.1307128817,0.0734806508,0.0858036652,-0.0963578075,-0.3183752894,0.0653057396,0.2368329167,-0.0228066426,0.2889071107,-0.1962373108,-0.3680241704,-0.0199918374,-0.1284303069,0.072580345,0.1414791495,0.061180193,0.0080608483,0.5874376297,-0.0742886215,0.1787090451,-0.2538342476,0.0291408673,-0.1472996622,0.1694096178,-0.6243599057,-0.1061867177,0.2582582831,0.091962114,0.1970394701,-0.0064032744,0.121820122,-0.3322490454,0.1758813262,-0.2791434526,0.2651101649,-0.0158670694,0.2457386106,-0.1295719892,-0.3093241453,-0.0095198667,-0.1647403091,0.2194372267,-0.7123686075,-0.0245730337,0.0450958833,0.0754593015,-0.214194268,0.1435741037,-0.504322648,0.0322593153,0.0630441457,0.2181688845,0.1069046035,-0.3306136727,-0.2142900974,-0.0720484778,0.0830698535,0.1326568276,-0.1291505694,0.0973811597,0.0630992949,0.0002566593,0.1041970924,0.2162794471,-0.0109667862,0.284181118,-0.1886345744,0.5723384023,0.4195818007,-0.8037276268,-0.3378547132,0.0875090212,-0.1656147987,0.0018321252,-0.0777145699,0.2999708056,0.2168296725,-0.1097091362,0.2860379517,0.4717787802,-0.0601560883,0.1129643992,-0.1055826619,0.2662775815,-0.1180936024,0.3215996325,0.0420975052,-0.2685292661,-0.1594627202,0.0751415715,0.1948614568,-0.1158283576,0.0870857239,0.4314538538,0.089273192,-0.0513841324,-0.0506237373,0.0383254662,-0.7624905109,0.1141011417,-0.4289596081,0.3793443143,0.0106568281,-0.0671950504,-0.1728422344,-0.0510757342,-0.296262145,-0.1906543225,0.107161127,-0.1209991425,-0.1237812713,-0.3034522235,-0.0762852654,0.3974147141,-0.2260980457,0.0737625808,-0.5780182481,0.340782702,0.1180048063,-0.0664847717,-0.0123105403,0.4472563863,0.1922498047,-0.0482966453,0.1684597582,0.2209219038,0.2148898542,-0.1571745276,-0.0612342209,-0.0020257682,0.1334930658,-0.5777854919,0.0659277737,0.2230994701,-0.001422982,-0.2143050879,-0.1243546456,0.5517559648,0.1847706735,0.2072845101,0.0289730113,0.0707501099,0.2122232467,-0.1713617146,-0.2344333678,-0.1039047167,0.1545492858,0.1233171746,0.0305474959,0.0995655283,-0.5023413897,0.0200132038,0.3077193201,0.0274304561,0.1898184419,0.2047269344,-0.1531089544,-0.0791647956,0.0374575406,0.3701202571,0.3590866625,0.0914273858,0.1098857149,0.1304076463,-0.0329413153,0.0724881813,-0.0358842276,0.0617540143,0.1510176659,0.2119652778,0.1785159111,-0.0841950774,-0.0858964771,-0.2191504538,0.0321776755,0.152163282,-0.3294998705,0.0780693293,0.0368220694,-0.450265348,-0.1473740041,0.0441721193,-0.1361576915,-0.3193361759,-0.0379218347,-0.0571559481,0.2387743443,0.084067151,-0.4746881425,-0.2150861621,-0.1986020505,-0.3406288028,-0.0583148114,0.0742932335,-0.1382389069,-0.0267540924,0.2119731903,0.5459074378,-0.031271968,-0.4353145659,0.3901607394,-0.3872416317,-0.2023112923,0.301656425,-0.2638221681,0.3121575713,0.217154935,-0.1040859297,-0.4046024084,0.195917204,0.1874918491,-0.321480006,-0.3141257763,0.3168615997,-0.1772629917,-0.0916748643,-0.0264440607,-0.40634498,-0.5237715244,-0.2054876089,0.0119940396,0.2230312675,0.3162436187,0.2151381969,-0.2596668601,-0.1451510787,-0.1321804971,-0.1955112368,0.0069593326,-0.5726137161,0.4568421841,-0.24419415,-0.2051525563,-0.0859806389,-0.2658783793,0.4713196158,0.3780293763,-0.0684937015,-0.0375404693,0.1413383931,0.0351820067,0.177708894,-0.0622194856,0.2964116931,-0.2141356468,0.0457051024,-0.0277902856,-0.0081237135,0.0785258338,0.3201446831,0.2304375917,-0.0265186951,-0.2571677864,0.110347338,0.0635766909,0.1552903205,0.0463344343,0.4367932975,0.005378745,0.1145690829,0.0847834796,0.0224085972,-0.164030239,0.0601466,0.233269453,-0.0009895281,0.2357880324,-0.1547724605,-0.072097823,0.1263714284,-0.1329991817,-0.215120554,0.5398620367,-0.1397541612,0.4105234146,-0.1709616035,0.0268179663,-0.0614328459,-0.4580695629,-0.257130444,0.086149618,0.1797157824,0.155405134,-0.6861747503,0.3985427618,-0.3888285756,-0.0206326935,0.1728511304,0.5378324389,0.2120253146,0.0804060176,0.1571722478,-0.0373992324,0.3112249672,-0.2897289991,-0.1462424099,0.1092613339,0.1001149788,-0.3250236809,-0.0430728868,-0.2508806586,0.0265304223,0.4807166457,0.4030665755,-0.3020348847,-0.4488573372,-0.0142154954,0.0474353395,-0.2095984071,-0.1977830976,-0.2662701309,-0.4505195916,-0.0922009498,-0.0910787657,0.5540920496,-0.0119683752,0.285989821,0.0289197415,-0.2357486933,0.2654157579,0.3301908374,-0.0371142514,0.5644020438,-0.2395819277,0.0057377298,-0.0657557547,-0.0843305588,0.2901473939,0.4803256691,-0.18594037,-0.1504533291,0.1800045818,-0.2238452286,0.4223891199,0.1663793623,0.1885036528,0.0188958589,-0.2950081825,-0.0011859048,-0.1358927488,0.1362023652,0.3885795772,-0.1082977429,-0.4188079834,-0.3653247952,0.1768360138,0.3088775873,0.1055980548,-0.2310434431,-0.1721609235,-0.2617760301,-0.1415586323,0.4663684666,1.016680479,0.030185407,0.433570534,0.1726903319,0.0897267163,0.0607269071,-0.444486618,0.2049940228,-0.3482038677,-0.3285882473,-0.2036454231,-0.2574140429,-0.0620909221,0.2026154697,0.0319126844,0.6626021862,0.2256908417,-0.0447899438,0.0229511857,0.2159696668,0.0093683628,-0.3867395818,-0.3745247722,0.0040317033,-0.2863345742,0.0149859404,-0.1535770446,-0.2083832622,-0.0591487885,0.1994757652,0.1399660408,0.2819466293,0.0934466422,-0.1829066277,0.4239276946,-0.5605942607,-0.0533471443,0.2428724915,0.077303797,-0.322191,0.0031517914,0.2361234426,0.4050419629,0.1009970233,0.3913243711,-0.3321993649,0.2601027191,0.0498278886,0.1221913695,-0.2428604513,0.1581598967,-0.0372625589,-0.4174329937,0.0479725488,-0.1750261933,-0.2749729753,-0.2087987065,0.0269379523,-0.1443690658,-0.049738355,0.0549408123,0.2594358027,-0.0902868882,0.148101449,0.1132918224,-0.2592061758,0.1990920603,0.0831901953,0.2976744771,-0.3125315309,0.0376100764,0.2482126951,0.0151349036,-0.0992361903,0.180581525,0.4821507633,-0.617361784,0.0337477662,-0.2434498519,-0.0137992604,-0.0163004398,0.1851167381,0.0020547709,-0.2624997497,0.1627749503,-0.2828570008,-0.3409923017,0.3751991391,0.1163776442,0.2782913744,-0.1979854256,0.0098260185,0.1509936601,-0.2200442106,-0.2390500307,0.1753156036,-0.1794706732,0.1961244643,-0.1157648787,-0.096291475,0.080403842,-0.0942012146,-0.0249693338,0.2664697766,-0.1458446831,-0.0578539148,0.0179341715,0.1243914366,0.1327522844,-0.1157702655,0.1248916909,0.1200044379,-0.0515412576,-0.1373377591,-0.1135678813,0.1800017059,0.0228337441,0.2303531468,-0.2976282835,0.0392743871,-0.1231651306,0.2310590148,0.189538762,0.1202059165,0.0657504871,0.2072514594,-0.0511192568,0.0248664636,-0.5182061791,0.2064265907,0.1967262477,0.3688436747,0.2151999176,-0.1553563476,0.1056381837,-0.032278128,0.1498689055,0.3083644211,-0.2699720263,-0.0058445805,0.241904065,0.1483671665,-0.1588442177,-0.0049726604,0.3671055734,0.0078911902,0.008068773,0.4790737331,0.3525026143,0.136815384,-0.1463459283,0.0633270293,0.2355753332,-0.2535267174,0.0676125661,0.4052120447,-0.005103806,0.1394571364,0.5995871425,0.0332877487,-0.1143290102,0.6000959277,0.3011852205,0.7809863091,0.199559778,0.1787547022,-0.0658113658,-0.1044648662,-0.4129894078,0.2548288405,0.1273963898,0.1769531816,0.0383527577,0.143287167,0.1453569233,-0.1860472709,-0.0964768156,0.113500081,-0.3404924273,0.0229304079,-0.331787169,-0.2342431843,-0.1217740625,-0.1177005768,-0.1722740084,0.0367383994,0.061489068,0.1339388788,-0.0074618193,-0.0612117462,0.0823539197,0.2852355242,0.0478237383,-0.1977075338,0.5288455486,0.0339015722,-0.0469572246,-0.1587117612,0.2742477059,0.2531481087,0.2315032333,-0.0519730113,-0.1070040986,0.3387274742,-0.1102280021,-0.02969338,0.1004743874,0.2005059421,0.2549166381,0.1804338545,0.0651821718,-0.0043521184,0.3010003567,0.0695356131,-0.1928669661,0.2435857952,0.2801603675,0.1861452311,-0.2890454829,-0.1294614524,-0.2241025269,-0.3388958275,-0.1642102152,0.314212203,-0.1428406239,0.2404812127,-0.130814895,0.0508457981,0.042049367,0.1968667656,0.1187227666,0.1140162945,-0.4945984483,-0.0885237828,-0.7457126975,0.2005533725,-0.3096918166,-0.2826579809,0.2355704159,0.0266045816,-0.0474609956,0.3022997379,-0.23467426,0.2289671898,0.0685480013,0.334798485,-0.6064161658,0.0685795322,0.2317826003,0.35899809,0.0531915426,-0.124981299,-0.276468724,0.2186534852,-0.0702891573,-0.0727587193,-0.0685441419,-0.1896738708,-0.0558257475,0.2673411965,-0.212423116,0.1567766815,-0.0746062472,-0.1744633466,-0.2747464776,-0.0597104505,-0.1644858569,0.0210557934,0.1005385742,0.3167755306,0.0033465335,0.1547871828,-0.3915068209,-0.2984790206,0.0132773137,0.0827491954,-0.3641844988,0.4328106642,-0.2321629971,0.1437955201,0.0377221107,0.2259716541,0.074632287,-0.0033203885,-0.1593918055,-0.4455848634,0.2613032758,0.2699011862,-0.1986567527,-0.1195013747,0.0043353084,0.1073929965,0.1258694679,-0.4479656518,0.1364092082,0.1799704432,-0.0451218039,-0.0945755392,0.0839961171,-0.0237168744,-0.0781790763,-0.1301056743,0.1238042712,-0.0262836535,-0.1145851761,-0.2040123045,-0.1585304439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/418","title":"Addition of google drive links to dl_manager","comments":"I think the problem is the way you wrote your urls. Try the following structure to see `https:\/\/drive.google.com\/uc?export=download&id=your_file_id` . \r\n\r\n@lhoestq ","body":"Hello there, I followed the template to create a download script of my own, which works fine for me, although I had to shun the dl_manager because it was downloading nothing from the drive links and instead use gdown.\r\n\r\nThis is the script for me:\r\n\r\n```python\r\nclass EmoConfig(nlp.BuilderConfig):\r\n \"\"\"BuilderConfig for SQUAD.\"\"\"\r\n\r\n def __init__(self, **kwargs):\r\n \"\"\"BuilderConfig for EmoContext.\r\n Args:\r\n **kwargs: keyword arguments forwarded to super.\r\n \"\"\"\r\n super(EmoConfig, self).__init__(**kwargs)\r\n\r\n_TEST_URL = \"https:\/\/drive.google.com\/file\/d\/1Hn5ytHSSoGOC4sjm3wYy0Dh0oY_oXBbb\/view?usp=sharing\"\r\n_TRAIN_URL = \"https:\/\/drive.google.com\/file\/d\/12Uz59TYg_NtxOy7SXraYeXPMRT7oaO7X\/view?usp=sharing\"\r\n\r\nclass EmoDataset(nlp.GeneratorBasedBuilder):\r\n \"\"\" SemEval-2019 Task 3: EmoContext Contextual Emotion Detection in Text. Version 1.0.0 \"\"\"\r\n\r\n VERSION = nlp.Version(\"1.0.0\")\r\n force = False\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(\r\n description=_DESCRIPTION,\r\n features=nlp.Features(\r\n {\r\n \"text\": nlp.Value(\"string\"),\r\n \"label\": nlp.features.ClassLabel(names=[\"others\", \"happy\", \"sad\", \"angry\"]),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/www.aclweb.org\/anthology\/S19-2005\/\",\r\n citation=_CITATION,\r\n )\r\n \r\n def _get_drive_url(self, url):\r\n base_url = 'https:\/\/drive.google.com\/uc?id='\r\n split_url = url.split('\/')\r\n return base_url + split_url[5]\r\n \r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n if(not os.path.exists(\"emo-train.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TRAIN_URL), \"emo-train.json\", quiet = True)\r\n if(not os.path.exists(\"emo-test.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TEST_URL), \"emo-test.json\", quiet = True)\r\n return [\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TRAIN,\r\n gen_kwargs={\r\n \"filepath\": \"emo-train.json\",\r\n \"split\": \"train\",\r\n },\r\n ),\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TEST,\r\n gen_kwargs={\"filepath\": \"emo-test.json\", \"split\": \"test\"},\r\n ),\r\n ]\r\n\r\n def _generate_examples(self, filepath, split):\r\n \"\"\" Yields examples. \"\"\"\r\n with open(filepath, 'rb') as f:\r\n data = json.load(f)\r\n for id_, text, label in zip(data[\"text\"].keys(), data[\"text\"].values(), data[\"Label\"].values()):\r\n yield id_, {\r\n \"text\": text,\r\n \"label\": label,\r\n }\r\n```\r\n\r\nCan someone help me in adding gdrive links to be used with default dl_manager or adding gdown as another dl_manager, because I'd like to add this dataset to nlp's official database.","comment_length":20,"text":"Addition of google drive links to dl_manager\nHello there, I followed the template to create a download script of my own, which works fine for me, although I had to shun the dl_manager because it was downloading nothing from the drive links and instead use gdown.\r\n\r\nThis is the script for me:\r\n\r\n```python\r\nclass EmoConfig(nlp.BuilderConfig):\r\n \"\"\"BuilderConfig for SQUAD.\"\"\"\r\n\r\n def __init__(self, **kwargs):\r\n \"\"\"BuilderConfig for EmoContext.\r\n Args:\r\n **kwargs: keyword arguments forwarded to super.\r\n \"\"\"\r\n super(EmoConfig, self).__init__(**kwargs)\r\n\r\n_TEST_URL = \"https:\/\/drive.google.com\/file\/d\/1Hn5ytHSSoGOC4sjm3wYy0Dh0oY_oXBbb\/view?usp=sharing\"\r\n_TRAIN_URL = \"https:\/\/drive.google.com\/file\/d\/12Uz59TYg_NtxOy7SXraYeXPMRT7oaO7X\/view?usp=sharing\"\r\n\r\nclass EmoDataset(nlp.GeneratorBasedBuilder):\r\n \"\"\" SemEval-2019 Task 3: EmoContext Contextual Emotion Detection in Text. Version 1.0.0 \"\"\"\r\n\r\n VERSION = nlp.Version(\"1.0.0\")\r\n force = False\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(\r\n description=_DESCRIPTION,\r\n features=nlp.Features(\r\n {\r\n \"text\": nlp.Value(\"string\"),\r\n \"label\": nlp.features.ClassLabel(names=[\"others\", \"happy\", \"sad\", \"angry\"]),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/www.aclweb.org\/anthology\/S19-2005\/\",\r\n citation=_CITATION,\r\n )\r\n \r\n def _get_drive_url(self, url):\r\n base_url = 'https:\/\/drive.google.com\/uc?id='\r\n split_url = url.split('\/')\r\n return base_url + split_url[5]\r\n \r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n if(not os.path.exists(\"emo-train.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TRAIN_URL), \"emo-train.json\", quiet = True)\r\n if(not os.path.exists(\"emo-test.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TEST_URL), \"emo-test.json\", quiet = True)\r\n return [\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TRAIN,\r\n gen_kwargs={\r\n \"filepath\": \"emo-train.json\",\r\n \"split\": \"train\",\r\n },\r\n ),\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TEST,\r\n gen_kwargs={\"filepath\": \"emo-test.json\", \"split\": \"test\"},\r\n ),\r\n ]\r\n\r\n def _generate_examples(self, filepath, split):\r\n \"\"\" Yields examples. \"\"\"\r\n with open(filepath, 'rb') as f:\r\n data = json.load(f)\r\n for id_, text, label in zip(data[\"text\"].keys(), data[\"text\"].values(), data[\"Label\"].values()):\r\n yield id_, {\r\n \"text\": text,\r\n \"label\": label,\r\n }\r\n```\r\n\r\nCan someone help me in adding gdrive links to be used with default dl_manager or adding gdown as another dl_manager, because I'd like to add this dataset to nlp's official database.\nI think the problem is the way you wrote your urls. Try the following structure to see `https:\/\/drive.google.com\/uc?export=download&id=your_file_id` . \r\n\r\n@lhoestq ","embeddings":[-0.0986939371,0.0516564436,-0.1309203506,-0.0558965728,0.0639402568,-0.0953544378,0.3679338396,0.2208102643,0.2062921822,0.2249778062,-0.0750645027,0.3655760586,-0.2819617987,0.1416775286,0.1477239877,0.1173321381,-0.224351123,0.2025327682,0.2756347656,-0.1639436185,-0.2751327753,0.3278068602,-0.0318780653,-0.0081285173,0.0502305627,0.0505593009,-0.1391222626,0.198068127,-0.1188958138,-0.2999538481,-0.4197380543,0.1671042591,0.160240218,-0.0974567235,-0.0001042657,-0.1480781883,0.2225690186,-0.2084534466,-0.4087229073,-0.293666929,0.2981956303,-0.1724279672,0.1073081344,-0.2651377916,-0.0045758276,-0.0983251333,0.2041182965,0.1354039758,0.2655147016,0.3032342196,0.2717006207,0.0790492222,0.113470979,0.0298931692,0.3312549591,-0.0354893319,-0.0682809278,0.35271433,0.3344673514,-0.3054541349,-0.1618669927,0.2883867025,-0.005473515,-0.0109913573,0.2195129991,0.1113411263,0.5780613422,-0.3311682045,0.1066150144,0.0327493623,0.0236196127,-0.0413232967,-0.0850505233,-0.2842149436,0.08381401,-0.2874571979,0.2766533494,0.1495649666,-0.2334840745,-0.0543575138,-0.1294789314,-0.0255045816,-0.015468711,0.0540597402,-0.0889275596,0.4138535857,-0.0389394574,-0.2830852568,0.3861259818,-0.0381522439,-0.0971575156,-0.1532618254,-0.0394701101,0.0766005889,-0.0374772809,-0.2887773812,0.1567475349,0.0466397926,0.093797259,0.2489909828,0.0460647494,0.0328915268,0.2356236577,0.0752923787,0.1630280316,0.4459595382,-0.0003144037,0.0400670469,0.3538968563,0.4176515937,0.1609001309,0.1501174271,-0.1202337146,-0.1861605793,-0.1416518539,0.1408617496,-0.1575787514,0.0906012058,-0.4310445487,-0.1330971122,-0.3732341826,0.0073386524,-0.0343621001,0.4378283918,-0.0487489253,-0.1396611929,0.2441136092,0.2374013662,-0.0423724875,0.0900178328,-0.1254783124,0.255615443,-0.2168087512,-0.0520229526,0.3310387731,0.0770800337,0.421222806,-0.2571605742,-0.1213400289,-0.224132359,0.1388938576,0.0859965533,-0.0139247123,-0.064111203,0.4167032838,0.0155231385,-0.1533248723,-0.4158217013,-0.071225442,0.0388117135,0.0347075351,-0.037107002,-0.1740760654,0.3355267346,-0.0862290189,-0.0942458659,-0.0037766567,0.2271474302,-0.1502159536,0.0436013825,0.0858766139,0.3182860315,-0.0827771053,-0.000988554,0.1685232818,0.2986229062,-0.0949452594,-0.2877131999,-0.1839938015,0.1673677117,-0.1132550612,-0.0632707402,-0.044929985,0.2455672026,-0.3794006705,0.2419207394,0.304346174,-0.2896392047,-0.2050281167,0.2148452997,-0.3752191663,-0.1749029905,0.030269891,0.0142999822,0.4108042121,-0.0042553372,-0.1432295889,0.4827906489,0.0771704242,0.242865935,-0.3040136099,-0.466017276,-0.1829135418,0.2075032294,0.1516643167,-0.1184460148,0.1088404134,-0.0877448469,0.3002710938,-0.1562739909,0.1865597665,0.1028919965,0.082430616,0.0962019637,0.0902650356,0.1277572811,-0.4553069472,-0.0245841052,-0.1491093785,0.2400395274,-0.3181627095,-0.1654351503,-0.2285122573,-0.1823611856,-0.2087038606,-0.3032354414,0.2976228893,0.2181787938,0.0589148887,0.2489134073,0.0791672543,-0.078336969,0.1662929952,0.1352360249,0.0483056195,0.3308407664,-0.3822062016,-0.1154452637,0.1777767688,0.0533847101,0.1627697051,-0.0097971642,0.0177794714,0.2022307515,-0.2338354439,0.4560253918,0.0342990011,-0.018487528,0.2643815875,-0.001894454,0.3286045492,0.3785975277,0.1253685653,-0.042322088,-0.1335391253,0.3993208706,0.2978732884,0.1315351278,0.1984390169,0.1492633671,-0.0345823541,-0.3092212379,-0.0366149396,-0.175003618,0.1060321406,0.4917028844,0.3812161088,-0.1171898544,-0.1798555851,0.2268595248,0.3522838056,-0.0431040488,0.2884167135,0.0797794685,-0.3493873775,-0.1552882195,-0.1172491759,0.3525727689,0.1629645526,0.2543209791,0.1712719202,0.0493353643,-0.0080083162,-0.3064471781,0.399212122,0.0655613244,0.1062253714,0.3554275632,0.112254627,-0.1241622567,-0.3668965697,-0.0078185881,-0.0030904301,-0.051934097,-0.2809714377,-0.0359772854,-0.2021116912,-0.1673544347,-0.040416833,0.1189079285,-0.0453860685,-0.1994874179,0.241571486,0.2851106822,-0.2777339518,0.127188921,0.1675787568,0.3435669839,-0.0704192296,-0.051740665,-0.1093210727,0.0156291276,-0.331397295,0.1907478869,0.2792063951,0.0847769007,0.1668625176,0.035354279,0.0824102759,-0.4212260544,-0.1966773272,0.2405076027,-0.1467298567,0.0396607593,0.1082567051,-0.0911808386,-0.0478258096,-0.3472113609,0.0503748134,-0.2066100538,-0.2028837651,-0.3251749873,0.1594581753,0.0629494265,-0.488184154,-0.3987681568,-0.2369147837,-0.380084008,0.1677398533,0.1598065197,0.1822991818,0.1246550754,-0.0548261032,0.1800568253,0.2231410295,-0.1606066525,-0.147671774,0.2775504291,0.4198610485,-0.4095939994,-0.3190635145,0.0439115465,-0.3232084513,0.0133835878,-0.2990259528,-0.2501083314,-0.4383622408,0.1537317634,-0.177799508,-0.0845339298,-0.0026808882,0.2986175418,0.1873589605,-0.2003104836,0.026588954,-0.1110703126,0.2070545703,0.0882992595,0.3847719133,-0.1615374535,0.4067229629,0.1174867451,0.2566284835,-0.26918751,0.0335369334,0.3619191945,-0.0506728664,0.1701949239,0.1592020392,-0.0863516107,0.1363561302,0.0409470275,-0.0944347158,0.5336619616,-0.0854697302,-0.4864979982,-0.1826670021,0.0532185845,-0.273686707,-0.2679212093,-0.1089556739,0.2660419047,-0.1956836432,0.0115898009,0.2149553001,0.0052229138,-0.0608394817,0.2059178799,0.2196903825,-0.0772171617,-0.1206152216,0.1185845733,0.1042682603,-0.3951476812,0.1117486209,0.1043822393,-0.4041220844,0.1074390486,-0.2256135195,0.2036410272,-0.3195444345,0.358741492,-0.1209610775,0.2662653029,-0.0143701937,0.216438055,0.10930226,-0.099338375,-0.2988061607,-0.3211817145,0.450747788,-0.0478757992,-0.265604794,-0.1441638917,0.3384916186,-0.2589463294,0.0323872305,-0.0962956622,-0.0191143472,-0.5297087431,-0.5422529578,0.0772396997,0.1260938197,0.0979745761,-0.1870060712,-0.163955152,0.2704376578,-0.2099725604,-0.1122278869,0.0596685223,0.1827882677,0.141125977,-0.04690478,0.2014990896,0.4859793782,-0.0863522738,0.0502427071,0.0887132287,0.1350985318,0.3584620059,-0.2915494144,-0.0120516224,0.4303926826,0.0423833728,-0.1449910998,-0.337811172,0.1311997622,-0.0050207195,0.0262090676,0.3523302376,-0.0076964782,0.3189740777,-0.1285665333,0.3718144,-0.114938952,-0.1495705545,-0.1133223847,0.137941882,-0.4891560674,0.2023630291,0.249479115,1.0095961094,0.1828901321,0.1841591895,0.3770165145,-0.1828773022,0.2244256288,-0.2048053294,0.2420448959,-0.275912106,0.2594016492,0.0270561446,0.123555094,-0.2554804087,0.2857078016,-0.3485202491,0.2226643264,0.1918338388,-0.0550935976,0.0397399962,0.5254421234,-0.414765358,-0.3550721705,-0.0840060115,0.2253390402,0.2074783742,0.3336937129,-0.0224050712,-0.3262494206,0.0230254158,-0.3689983487,-0.1200434417,0.222809732,-0.6562026739,0.0811413154,-0.4205805063,-0.2054655552,0.0714982525,0.1216394231,0.1825933158,0.1002613306,-0.0640582293,0.2806988955,0.0451345742,-0.0464380011,-0.0050444896,-0.117036812,0.2287954986,-0.254381597,-0.2008146793,-0.052955918,0.150778234,-0.2844575346,0.0494143516,-0.0253137629,-0.2893356681,0.1632657945,0.0008363276,0.0888023153,-0.4070391357,-0.1701178551,0.2157660872,-0.2228650451,-0.0063023539,-0.0420259535,0.0753558651,-0.1657579988,-0.1717545539,0.4114311039,-0.1920544058,-0.3257522881,0.5842943788,0.0643735826,-0.1504262239,-0.3654353917,-0.2140043378,0.0150121469,-0.1508513838,0.1102009565,-0.1128665805,-0.256216377,-0.0146313105,0.5000475645,-0.0012302455,0.2679022551,0.0265189763,-0.0871425867,-0.0991812721,0.2623232007,-0.1346050203,0.2985242903,-0.0751187578,0.0820538327,0.1869878322,-0.1457145363,-0.4634034336,0.0790768564,-0.1712537557,-0.1405716389,0.1250740886,-0.2987323999,0.3095208406,0.1704262793,0.2733668387,0.2205571383,-0.077256754,-0.2663613558,-0.0304335151,0.07006412,0.0473003574,-0.1006334201,0.0123664159,-0.1115101725,-0.218509391,0.1263172477,0.2010792345,-0.1831626147,0.0833370537,-0.1032224596,-0.2481856495,0.0976704136,0.2554208934,-0.1850833893,-0.0976004004,0.4207085371,0.0166867264,0.1075952277,0.0078057367,0.0033925159,0.0431230143,0.1411653012,0.0564506911,0.3387710452,0.2962433994,0.0672249943,-0.5515539646,0.0619442984,0.2637961805,0.078537114,-0.2515684962,-0.1156085432,0.2228315324,0.3556102216,-0.259417206,0.1732178032,-0.1623107642,-0.4482299685,0.3617484272,0.2996283472,0.1987316757,0.0672524571,0.0475771576,0.1538341343,0.404625982,-0.0747884288,-0.092839621,-0.0525272973,-0.1224941462,0.1339276731,0.185081616,0.067402795,0.3257806003,-0.1616298556,-0.1871491969,0.008799796,0.0235794131,0.2640464604,0.2166270763,0.0519343317,0.03836805,0.081463933,-0.2451706231,0.0961864218,-0.0673369393,0.2200666219,-0.3234996498,0.0240175724,-0.3409782052,0.4161132872,-0.0618106611,-0.0085473182,-0.2516661286,-0.3260459602,-0.6693037152,0.1929431558,0.0450478755,-0.1420245618,-0.4067165256,0.1091223657,-0.0436599553,0.0488749109,-0.2354439199,-0.09591306,0.3324644864,-0.4127458632,0.1486727297,0.0149962129,0.1223974302,-0.1983048767,0.3227460682,0.2554567456,0.2159596384,-0.3102447391,-0.2035475373,-0.079424113,-0.0601548404,-0.144217059,0.037109971,0.3265762031,-0.1694569141,0.1500537395,0.2160780877,-0.3853534162,-0.0577884056,0.2834987938,-0.2225657403,0.2526809275,0.1624028534,0.2718719542,0.0135586848,-0.3872140646,-0.3066829741,-0.5732123256,-0.1348943859,-0.2507286668,-0.1008889601,0.2407356501,-0.3101221323,0.1533886641,0.1174057275,0.180290997,0.3067573309,-0.0013956358,-0.1100461632,-0.2357510924,-0.439438194,0.0553815812,0.2680011988,-0.1324961334,-0.219313547,0.4046549201,-0.0119787026,-0.0210185908,0.0093867835,0.5449488163,0.2535102069,0.0710766315,-0.1868591011,0.1759403348,0.1250489652,0.1145300791,0.0818908438,-0.1827889383,-0.1159717515,-0.2869517803,0.214859724,-0.1342401356,0.1613276899,-0.3132973611,0.0037505045,0.4126843512,-0.2459646463,0.2662286758,-0.286221087,-0.1534481794,-0.1860763282,0.2085224092,-0.295242399,-0.3550876975,0.2407322079,0.1949659288,0.0700791925,0.0669670627,-0.2691867948,0.2615852654,0.0445120037,0.0049234945,-0.1911266148,0.11490722,-0.0623385981,0.0978202522,0.0315187462,0.205317378,-0.1249731705,0.0482877232,-0.495287329,-0.5152050853,0.1610213369,-0.0524530187,-0.086882256,-0.0676097795,0.3968174756,0.2529715598,0.1593210995,-0.2931996584,0.125608176,-0.0083493926,0.1823005974,-0.1703491807,0.1373269558,0.138203606,-0.1351919025,-0.0715939626,-0.2012002319,0.3138321936,0.0080849808,-0.206582889,-0.3480609059]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/418","title":"Addition of google drive links to dl_manager","comments":"Oh sorry, I think `_get_drive_url` is doing that. \r\n\r\nHave you tried to use `dl_manager.download_and_extract(_get_drive_url(_TRAIN_URL)`? it should work with google drive links.\r\n","body":"Hello there, I followed the template to create a download script of my own, which works fine for me, although I had to shun the dl_manager because it was downloading nothing from the drive links and instead use gdown.\r\n\r\nThis is the script for me:\r\n\r\n```python\r\nclass EmoConfig(nlp.BuilderConfig):\r\n \"\"\"BuilderConfig for SQUAD.\"\"\"\r\n\r\n def __init__(self, **kwargs):\r\n \"\"\"BuilderConfig for EmoContext.\r\n Args:\r\n **kwargs: keyword arguments forwarded to super.\r\n \"\"\"\r\n super(EmoConfig, self).__init__(**kwargs)\r\n\r\n_TEST_URL = \"https:\/\/drive.google.com\/file\/d\/1Hn5ytHSSoGOC4sjm3wYy0Dh0oY_oXBbb\/view?usp=sharing\"\r\n_TRAIN_URL = \"https:\/\/drive.google.com\/file\/d\/12Uz59TYg_NtxOy7SXraYeXPMRT7oaO7X\/view?usp=sharing\"\r\n\r\nclass EmoDataset(nlp.GeneratorBasedBuilder):\r\n \"\"\" SemEval-2019 Task 3: EmoContext Contextual Emotion Detection in Text. Version 1.0.0 \"\"\"\r\n\r\n VERSION = nlp.Version(\"1.0.0\")\r\n force = False\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(\r\n description=_DESCRIPTION,\r\n features=nlp.Features(\r\n {\r\n \"text\": nlp.Value(\"string\"),\r\n \"label\": nlp.features.ClassLabel(names=[\"others\", \"happy\", \"sad\", \"angry\"]),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/www.aclweb.org\/anthology\/S19-2005\/\",\r\n citation=_CITATION,\r\n )\r\n \r\n def _get_drive_url(self, url):\r\n base_url = 'https:\/\/drive.google.com\/uc?id='\r\n split_url = url.split('\/')\r\n return base_url + split_url[5]\r\n \r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n if(not os.path.exists(\"emo-train.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TRAIN_URL), \"emo-train.json\", quiet = True)\r\n if(not os.path.exists(\"emo-test.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TEST_URL), \"emo-test.json\", quiet = True)\r\n return [\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TRAIN,\r\n gen_kwargs={\r\n \"filepath\": \"emo-train.json\",\r\n \"split\": \"train\",\r\n },\r\n ),\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TEST,\r\n gen_kwargs={\"filepath\": \"emo-test.json\", \"split\": \"test\"},\r\n ),\r\n ]\r\n\r\n def _generate_examples(self, filepath, split):\r\n \"\"\" Yields examples. \"\"\"\r\n with open(filepath, 'rb') as f:\r\n data = json.load(f)\r\n for id_, text, label in zip(data[\"text\"].keys(), data[\"text\"].values(), data[\"Label\"].values()):\r\n yield id_, {\r\n \"text\": text,\r\n \"label\": label,\r\n }\r\n```\r\n\r\nCan someone help me in adding gdrive links to be used with default dl_manager or adding gdown as another dl_manager, because I'd like to add this dataset to nlp's official database.","comment_length":21,"text":"Addition of google drive links to dl_manager\nHello there, I followed the template to create a download script of my own, which works fine for me, although I had to shun the dl_manager because it was downloading nothing from the drive links and instead use gdown.\r\n\r\nThis is the script for me:\r\n\r\n```python\r\nclass EmoConfig(nlp.BuilderConfig):\r\n \"\"\"BuilderConfig for SQUAD.\"\"\"\r\n\r\n def __init__(self, **kwargs):\r\n \"\"\"BuilderConfig for EmoContext.\r\n Args:\r\n **kwargs: keyword arguments forwarded to super.\r\n \"\"\"\r\n super(EmoConfig, self).__init__(**kwargs)\r\n\r\n_TEST_URL = \"https:\/\/drive.google.com\/file\/d\/1Hn5ytHSSoGOC4sjm3wYy0Dh0oY_oXBbb\/view?usp=sharing\"\r\n_TRAIN_URL = \"https:\/\/drive.google.com\/file\/d\/12Uz59TYg_NtxOy7SXraYeXPMRT7oaO7X\/view?usp=sharing\"\r\n\r\nclass EmoDataset(nlp.GeneratorBasedBuilder):\r\n \"\"\" SemEval-2019 Task 3: EmoContext Contextual Emotion Detection in Text. Version 1.0.0 \"\"\"\r\n\r\n VERSION = nlp.Version(\"1.0.0\")\r\n force = False\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(\r\n description=_DESCRIPTION,\r\n features=nlp.Features(\r\n {\r\n \"text\": nlp.Value(\"string\"),\r\n \"label\": nlp.features.ClassLabel(names=[\"others\", \"happy\", \"sad\", \"angry\"]),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/www.aclweb.org\/anthology\/S19-2005\/\",\r\n citation=_CITATION,\r\n )\r\n \r\n def _get_drive_url(self, url):\r\n base_url = 'https:\/\/drive.google.com\/uc?id='\r\n split_url = url.split('\/')\r\n return base_url + split_url[5]\r\n \r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n if(not os.path.exists(\"emo-train.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TRAIN_URL), \"emo-train.json\", quiet = True)\r\n if(not os.path.exists(\"emo-test.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TEST_URL), \"emo-test.json\", quiet = True)\r\n return [\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TRAIN,\r\n gen_kwargs={\r\n \"filepath\": \"emo-train.json\",\r\n \"split\": \"train\",\r\n },\r\n ),\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TEST,\r\n gen_kwargs={\"filepath\": \"emo-test.json\", \"split\": \"test\"},\r\n ),\r\n ]\r\n\r\n def _generate_examples(self, filepath, split):\r\n \"\"\" Yields examples. \"\"\"\r\n with open(filepath, 'rb') as f:\r\n data = json.load(f)\r\n for id_, text, label in zip(data[\"text\"].keys(), data[\"text\"].values(), data[\"Label\"].values()):\r\n yield id_, {\r\n \"text\": text,\r\n \"label\": label,\r\n }\r\n```\r\n\r\nCan someone help me in adding gdrive links to be used with default dl_manager or adding gdown as another dl_manager, because I'd like to add this dataset to nlp's official database.\nOh sorry, I think `_get_drive_url` is doing that. \r\n\r\nHave you tried to use `dl_manager.download_and_extract(_get_drive_url(_TRAIN_URL)`? it should work with google drive links.\r\n","embeddings":[-0.0986939371,0.0516564436,-0.1309203506,-0.0558965728,0.0639402568,-0.0953544378,0.3679338396,0.2208102643,0.2062921822,0.2249778062,-0.0750645027,0.3655760586,-0.2819617987,0.1416775286,0.1477239877,0.1173321381,-0.224351123,0.2025327682,0.2756347656,-0.1639436185,-0.2751327753,0.3278068602,-0.0318780653,-0.0081285173,0.0502305627,0.0505593009,-0.1391222626,0.198068127,-0.1188958138,-0.2999538481,-0.4197380543,0.1671042591,0.160240218,-0.0974567235,-0.0001042657,-0.1480781883,0.2225690186,-0.2084534466,-0.4087229073,-0.293666929,0.2981956303,-0.1724279672,0.1073081344,-0.2651377916,-0.0045758276,-0.0983251333,0.2041182965,0.1354039758,0.2655147016,0.3032342196,0.2717006207,0.0790492222,0.113470979,0.0298931692,0.3312549591,-0.0354893319,-0.0682809278,0.35271433,0.3344673514,-0.3054541349,-0.1618669927,0.2883867025,-0.005473515,-0.0109913573,0.2195129991,0.1113411263,0.5780613422,-0.3311682045,0.1066150144,0.0327493623,0.0236196127,-0.0413232967,-0.0850505233,-0.2842149436,0.08381401,-0.2874571979,0.2766533494,0.1495649666,-0.2334840745,-0.0543575138,-0.1294789314,-0.0255045816,-0.015468711,0.0540597402,-0.0889275596,0.4138535857,-0.0389394574,-0.2830852568,0.3861259818,-0.0381522439,-0.0971575156,-0.1532618254,-0.0394701101,0.0766005889,-0.0374772809,-0.2887773812,0.1567475349,0.0466397926,0.093797259,0.2489909828,0.0460647494,0.0328915268,0.2356236577,0.0752923787,0.1630280316,0.4459595382,-0.0003144037,0.0400670469,0.3538968563,0.4176515937,0.1609001309,0.1501174271,-0.1202337146,-0.1861605793,-0.1416518539,0.1408617496,-0.1575787514,0.0906012058,-0.4310445487,-0.1330971122,-0.3732341826,0.0073386524,-0.0343621001,0.4378283918,-0.0487489253,-0.1396611929,0.2441136092,0.2374013662,-0.0423724875,0.0900178328,-0.1254783124,0.255615443,-0.2168087512,-0.0520229526,0.3310387731,0.0770800337,0.421222806,-0.2571605742,-0.1213400289,-0.224132359,0.1388938576,0.0859965533,-0.0139247123,-0.064111203,0.4167032838,0.0155231385,-0.1533248723,-0.4158217013,-0.071225442,0.0388117135,0.0347075351,-0.037107002,-0.1740760654,0.3355267346,-0.0862290189,-0.0942458659,-0.0037766567,0.2271474302,-0.1502159536,0.0436013825,0.0858766139,0.3182860315,-0.0827771053,-0.000988554,0.1685232818,0.2986229062,-0.0949452594,-0.2877131999,-0.1839938015,0.1673677117,-0.1132550612,-0.0632707402,-0.044929985,0.2455672026,-0.3794006705,0.2419207394,0.304346174,-0.2896392047,-0.2050281167,0.2148452997,-0.3752191663,-0.1749029905,0.030269891,0.0142999822,0.4108042121,-0.0042553372,-0.1432295889,0.4827906489,0.0771704242,0.242865935,-0.3040136099,-0.466017276,-0.1829135418,0.2075032294,0.1516643167,-0.1184460148,0.1088404134,-0.0877448469,0.3002710938,-0.1562739909,0.1865597665,0.1028919965,0.082430616,0.0962019637,0.0902650356,0.1277572811,-0.4553069472,-0.0245841052,-0.1491093785,0.2400395274,-0.3181627095,-0.1654351503,-0.2285122573,-0.1823611856,-0.2087038606,-0.3032354414,0.2976228893,0.2181787938,0.0589148887,0.2489134073,0.0791672543,-0.078336969,0.1662929952,0.1352360249,0.0483056195,0.3308407664,-0.3822062016,-0.1154452637,0.1777767688,0.0533847101,0.1627697051,-0.0097971642,0.0177794714,0.2022307515,-0.2338354439,0.4560253918,0.0342990011,-0.018487528,0.2643815875,-0.001894454,0.3286045492,0.3785975277,0.1253685653,-0.042322088,-0.1335391253,0.3993208706,0.2978732884,0.1315351278,0.1984390169,0.1492633671,-0.0345823541,-0.3092212379,-0.0366149396,-0.175003618,0.1060321406,0.4917028844,0.3812161088,-0.1171898544,-0.1798555851,0.2268595248,0.3522838056,-0.0431040488,0.2884167135,0.0797794685,-0.3493873775,-0.1552882195,-0.1172491759,0.3525727689,0.1629645526,0.2543209791,0.1712719202,0.0493353643,-0.0080083162,-0.3064471781,0.399212122,0.0655613244,0.1062253714,0.3554275632,0.112254627,-0.1241622567,-0.3668965697,-0.0078185881,-0.0030904301,-0.051934097,-0.2809714377,-0.0359772854,-0.2021116912,-0.1673544347,-0.040416833,0.1189079285,-0.0453860685,-0.1994874179,0.241571486,0.2851106822,-0.2777339518,0.127188921,0.1675787568,0.3435669839,-0.0704192296,-0.051740665,-0.1093210727,0.0156291276,-0.331397295,0.1907478869,0.2792063951,0.0847769007,0.1668625176,0.035354279,0.0824102759,-0.4212260544,-0.1966773272,0.2405076027,-0.1467298567,0.0396607593,0.1082567051,-0.0911808386,-0.0478258096,-0.3472113609,0.0503748134,-0.2066100538,-0.2028837651,-0.3251749873,0.1594581753,0.0629494265,-0.488184154,-0.3987681568,-0.2369147837,-0.380084008,0.1677398533,0.1598065197,0.1822991818,0.1246550754,-0.0548261032,0.1800568253,0.2231410295,-0.1606066525,-0.147671774,0.2775504291,0.4198610485,-0.4095939994,-0.3190635145,0.0439115465,-0.3232084513,0.0133835878,-0.2990259528,-0.2501083314,-0.4383622408,0.1537317634,-0.177799508,-0.0845339298,-0.0026808882,0.2986175418,0.1873589605,-0.2003104836,0.026588954,-0.1110703126,0.2070545703,0.0882992595,0.3847719133,-0.1615374535,0.4067229629,0.1174867451,0.2566284835,-0.26918751,0.0335369334,0.3619191945,-0.0506728664,0.1701949239,0.1592020392,-0.0863516107,0.1363561302,0.0409470275,-0.0944347158,0.5336619616,-0.0854697302,-0.4864979982,-0.1826670021,0.0532185845,-0.273686707,-0.2679212093,-0.1089556739,0.2660419047,-0.1956836432,0.0115898009,0.2149553001,0.0052229138,-0.0608394817,0.2059178799,0.2196903825,-0.0772171617,-0.1206152216,0.1185845733,0.1042682603,-0.3951476812,0.1117486209,0.1043822393,-0.4041220844,0.1074390486,-0.2256135195,0.2036410272,-0.3195444345,0.358741492,-0.1209610775,0.2662653029,-0.0143701937,0.216438055,0.10930226,-0.099338375,-0.2988061607,-0.3211817145,0.450747788,-0.0478757992,-0.265604794,-0.1441638917,0.3384916186,-0.2589463294,0.0323872305,-0.0962956622,-0.0191143472,-0.5297087431,-0.5422529578,0.0772396997,0.1260938197,0.0979745761,-0.1870060712,-0.163955152,0.2704376578,-0.2099725604,-0.1122278869,0.0596685223,0.1827882677,0.141125977,-0.04690478,0.2014990896,0.4859793782,-0.0863522738,0.0502427071,0.0887132287,0.1350985318,0.3584620059,-0.2915494144,-0.0120516224,0.4303926826,0.0423833728,-0.1449910998,-0.337811172,0.1311997622,-0.0050207195,0.0262090676,0.3523302376,-0.0076964782,0.3189740777,-0.1285665333,0.3718144,-0.114938952,-0.1495705545,-0.1133223847,0.137941882,-0.4891560674,0.2023630291,0.249479115,1.0095961094,0.1828901321,0.1841591895,0.3770165145,-0.1828773022,0.2244256288,-0.2048053294,0.2420448959,-0.275912106,0.2594016492,0.0270561446,0.123555094,-0.2554804087,0.2857078016,-0.3485202491,0.2226643264,0.1918338388,-0.0550935976,0.0397399962,0.5254421234,-0.414765358,-0.3550721705,-0.0840060115,0.2253390402,0.2074783742,0.3336937129,-0.0224050712,-0.3262494206,0.0230254158,-0.3689983487,-0.1200434417,0.222809732,-0.6562026739,0.0811413154,-0.4205805063,-0.2054655552,0.0714982525,0.1216394231,0.1825933158,0.1002613306,-0.0640582293,0.2806988955,0.0451345742,-0.0464380011,-0.0050444896,-0.117036812,0.2287954986,-0.254381597,-0.2008146793,-0.052955918,0.150778234,-0.2844575346,0.0494143516,-0.0253137629,-0.2893356681,0.1632657945,0.0008363276,0.0888023153,-0.4070391357,-0.1701178551,0.2157660872,-0.2228650451,-0.0063023539,-0.0420259535,0.0753558651,-0.1657579988,-0.1717545539,0.4114311039,-0.1920544058,-0.3257522881,0.5842943788,0.0643735826,-0.1504262239,-0.3654353917,-0.2140043378,0.0150121469,-0.1508513838,0.1102009565,-0.1128665805,-0.256216377,-0.0146313105,0.5000475645,-0.0012302455,0.2679022551,0.0265189763,-0.0871425867,-0.0991812721,0.2623232007,-0.1346050203,0.2985242903,-0.0751187578,0.0820538327,0.1869878322,-0.1457145363,-0.4634034336,0.0790768564,-0.1712537557,-0.1405716389,0.1250740886,-0.2987323999,0.3095208406,0.1704262793,0.2733668387,0.2205571383,-0.077256754,-0.2663613558,-0.0304335151,0.07006412,0.0473003574,-0.1006334201,0.0123664159,-0.1115101725,-0.218509391,0.1263172477,0.2010792345,-0.1831626147,0.0833370537,-0.1032224596,-0.2481856495,0.0976704136,0.2554208934,-0.1850833893,-0.0976004004,0.4207085371,0.0166867264,0.1075952277,0.0078057367,0.0033925159,0.0431230143,0.1411653012,0.0564506911,0.3387710452,0.2962433994,0.0672249943,-0.5515539646,0.0619442984,0.2637961805,0.078537114,-0.2515684962,-0.1156085432,0.2228315324,0.3556102216,-0.259417206,0.1732178032,-0.1623107642,-0.4482299685,0.3617484272,0.2996283472,0.1987316757,0.0672524571,0.0475771576,0.1538341343,0.404625982,-0.0747884288,-0.092839621,-0.0525272973,-0.1224941462,0.1339276731,0.185081616,0.067402795,0.3257806003,-0.1616298556,-0.1871491969,0.008799796,0.0235794131,0.2640464604,0.2166270763,0.0519343317,0.03836805,0.081463933,-0.2451706231,0.0961864218,-0.0673369393,0.2200666219,-0.3234996498,0.0240175724,-0.3409782052,0.4161132872,-0.0618106611,-0.0085473182,-0.2516661286,-0.3260459602,-0.6693037152,0.1929431558,0.0450478755,-0.1420245618,-0.4067165256,0.1091223657,-0.0436599553,0.0488749109,-0.2354439199,-0.09591306,0.3324644864,-0.4127458632,0.1486727297,0.0149962129,0.1223974302,-0.1983048767,0.3227460682,0.2554567456,0.2159596384,-0.3102447391,-0.2035475373,-0.079424113,-0.0601548404,-0.144217059,0.037109971,0.3265762031,-0.1694569141,0.1500537395,0.2160780877,-0.3853534162,-0.0577884056,0.2834987938,-0.2225657403,0.2526809275,0.1624028534,0.2718719542,0.0135586848,-0.3872140646,-0.3066829741,-0.5732123256,-0.1348943859,-0.2507286668,-0.1008889601,0.2407356501,-0.3101221323,0.1533886641,0.1174057275,0.180290997,0.3067573309,-0.0013956358,-0.1100461632,-0.2357510924,-0.439438194,0.0553815812,0.2680011988,-0.1324961334,-0.219313547,0.4046549201,-0.0119787026,-0.0210185908,0.0093867835,0.5449488163,0.2535102069,0.0710766315,-0.1868591011,0.1759403348,0.1250489652,0.1145300791,0.0818908438,-0.1827889383,-0.1159717515,-0.2869517803,0.214859724,-0.1342401356,0.1613276899,-0.3132973611,0.0037505045,0.4126843512,-0.2459646463,0.2662286758,-0.286221087,-0.1534481794,-0.1860763282,0.2085224092,-0.295242399,-0.3550876975,0.2407322079,0.1949659288,0.0700791925,0.0669670627,-0.2691867948,0.2615852654,0.0445120037,0.0049234945,-0.1911266148,0.11490722,-0.0623385981,0.0978202522,0.0315187462,0.205317378,-0.1249731705,0.0482877232,-0.495287329,-0.5152050853,0.1610213369,-0.0524530187,-0.086882256,-0.0676097795,0.3968174756,0.2529715598,0.1593210995,-0.2931996584,0.125608176,-0.0083493926,0.1823005974,-0.1703491807,0.1373269558,0.138203606,-0.1351919025,-0.0715939626,-0.2012002319,0.3138321936,0.0080849808,-0.206582889,-0.3480609059]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/418","title":"Addition of google drive links to dl_manager","comments":"Yes it worked, thank you!","body":"Hello there, I followed the template to create a download script of my own, which works fine for me, although I had to shun the dl_manager because it was downloading nothing from the drive links and instead use gdown.\r\n\r\nThis is the script for me:\r\n\r\n```python\r\nclass EmoConfig(nlp.BuilderConfig):\r\n \"\"\"BuilderConfig for SQUAD.\"\"\"\r\n\r\n def __init__(self, **kwargs):\r\n \"\"\"BuilderConfig for EmoContext.\r\n Args:\r\n **kwargs: keyword arguments forwarded to super.\r\n \"\"\"\r\n super(EmoConfig, self).__init__(**kwargs)\r\n\r\n_TEST_URL = \"https:\/\/drive.google.com\/file\/d\/1Hn5ytHSSoGOC4sjm3wYy0Dh0oY_oXBbb\/view?usp=sharing\"\r\n_TRAIN_URL = \"https:\/\/drive.google.com\/file\/d\/12Uz59TYg_NtxOy7SXraYeXPMRT7oaO7X\/view?usp=sharing\"\r\n\r\nclass EmoDataset(nlp.GeneratorBasedBuilder):\r\n \"\"\" SemEval-2019 Task 3: EmoContext Contextual Emotion Detection in Text. Version 1.0.0 \"\"\"\r\n\r\n VERSION = nlp.Version(\"1.0.0\")\r\n force = False\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(\r\n description=_DESCRIPTION,\r\n features=nlp.Features(\r\n {\r\n \"text\": nlp.Value(\"string\"),\r\n \"label\": nlp.features.ClassLabel(names=[\"others\", \"happy\", \"sad\", \"angry\"]),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/www.aclweb.org\/anthology\/S19-2005\/\",\r\n citation=_CITATION,\r\n )\r\n \r\n def _get_drive_url(self, url):\r\n base_url = 'https:\/\/drive.google.com\/uc?id='\r\n split_url = url.split('\/')\r\n return base_url + split_url[5]\r\n \r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n if(not os.path.exists(\"emo-train.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TRAIN_URL), \"emo-train.json\", quiet = True)\r\n if(not os.path.exists(\"emo-test.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TEST_URL), \"emo-test.json\", quiet = True)\r\n return [\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TRAIN,\r\n gen_kwargs={\r\n \"filepath\": \"emo-train.json\",\r\n \"split\": \"train\",\r\n },\r\n ),\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TEST,\r\n gen_kwargs={\"filepath\": \"emo-test.json\", \"split\": \"test\"},\r\n ),\r\n ]\r\n\r\n def _generate_examples(self, filepath, split):\r\n \"\"\" Yields examples. \"\"\"\r\n with open(filepath, 'rb') as f:\r\n data = json.load(f)\r\n for id_, text, label in zip(data[\"text\"].keys(), data[\"text\"].values(), data[\"Label\"].values()):\r\n yield id_, {\r\n \"text\": text,\r\n \"label\": label,\r\n }\r\n```\r\n\r\nCan someone help me in adding gdrive links to be used with default dl_manager or adding gdown as another dl_manager, because I'd like to add this dataset to nlp's official database.","comment_length":5,"text":"Addition of google drive links to dl_manager\nHello there, I followed the template to create a download script of my own, which works fine for me, although I had to shun the dl_manager because it was downloading nothing from the drive links and instead use gdown.\r\n\r\nThis is the script for me:\r\n\r\n```python\r\nclass EmoConfig(nlp.BuilderConfig):\r\n \"\"\"BuilderConfig for SQUAD.\"\"\"\r\n\r\n def __init__(self, **kwargs):\r\n \"\"\"BuilderConfig for EmoContext.\r\n Args:\r\n **kwargs: keyword arguments forwarded to super.\r\n \"\"\"\r\n super(EmoConfig, self).__init__(**kwargs)\r\n\r\n_TEST_URL = \"https:\/\/drive.google.com\/file\/d\/1Hn5ytHSSoGOC4sjm3wYy0Dh0oY_oXBbb\/view?usp=sharing\"\r\n_TRAIN_URL = \"https:\/\/drive.google.com\/file\/d\/12Uz59TYg_NtxOy7SXraYeXPMRT7oaO7X\/view?usp=sharing\"\r\n\r\nclass EmoDataset(nlp.GeneratorBasedBuilder):\r\n \"\"\" SemEval-2019 Task 3: EmoContext Contextual Emotion Detection in Text. Version 1.0.0 \"\"\"\r\n\r\n VERSION = nlp.Version(\"1.0.0\")\r\n force = False\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(\r\n description=_DESCRIPTION,\r\n features=nlp.Features(\r\n {\r\n \"text\": nlp.Value(\"string\"),\r\n \"label\": nlp.features.ClassLabel(names=[\"others\", \"happy\", \"sad\", \"angry\"]),\r\n }\r\n ),\r\n supervised_keys=None,\r\n homepage=\"https:\/\/www.aclweb.org\/anthology\/S19-2005\/\",\r\n citation=_CITATION,\r\n )\r\n \r\n def _get_drive_url(self, url):\r\n base_url = 'https:\/\/drive.google.com\/uc?id='\r\n split_url = url.split('\/')\r\n return base_url + split_url[5]\r\n \r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n if(not os.path.exists(\"emo-train.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TRAIN_URL), \"emo-train.json\", quiet = True)\r\n if(not os.path.exists(\"emo-test.json\") or self.force):\r\n gdown.download(self._get_drive_url(_TEST_URL), \"emo-test.json\", quiet = True)\r\n return [\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TRAIN,\r\n gen_kwargs={\r\n \"filepath\": \"emo-train.json\",\r\n \"split\": \"train\",\r\n },\r\n ),\r\n nlp.SplitGenerator(\r\n name=nlp.Split.TEST,\r\n gen_kwargs={\"filepath\": \"emo-test.json\", \"split\": \"test\"},\r\n ),\r\n ]\r\n\r\n def _generate_examples(self, filepath, split):\r\n \"\"\" Yields examples. \"\"\"\r\n with open(filepath, 'rb') as f:\r\n data = json.load(f)\r\n for id_, text, label in zip(data[\"text\"].keys(), data[\"text\"].values(), data[\"Label\"].values()):\r\n yield id_, {\r\n \"text\": text,\r\n \"label\": label,\r\n }\r\n```\r\n\r\nCan someone help me in adding gdrive links to be used with default dl_manager or adding gdown as another dl_manager, because I'd like to add this dataset to nlp's official database.\nYes it worked, thank you!","embeddings":[-0.0986939371,0.0516564436,-0.1309203506,-0.0558965728,0.0639402568,-0.0953544378,0.3679338396,0.2208102643,0.2062921822,0.2249778062,-0.0750645027,0.3655760586,-0.2819617987,0.1416775286,0.1477239877,0.1173321381,-0.224351123,0.2025327682,0.2756347656,-0.1639436185,-0.2751327753,0.3278068602,-0.0318780653,-0.0081285173,0.0502305627,0.0505593009,-0.1391222626,0.198068127,-0.1188958138,-0.2999538481,-0.4197380543,0.1671042591,0.160240218,-0.0974567235,-0.0001042657,-0.1480781883,0.2225690186,-0.2084534466,-0.4087229073,-0.293666929,0.2981956303,-0.1724279672,0.1073081344,-0.2651377916,-0.0045758276,-0.0983251333,0.2041182965,0.1354039758,0.2655147016,0.3032342196,0.2717006207,0.0790492222,0.113470979,0.0298931692,0.3312549591,-0.0354893319,-0.0682809278,0.35271433,0.3344673514,-0.3054541349,-0.1618669927,0.2883867025,-0.005473515,-0.0109913573,0.2195129991,0.1113411263,0.5780613422,-0.3311682045,0.1066150144,0.0327493623,0.0236196127,-0.0413232967,-0.0850505233,-0.2842149436,0.08381401,-0.2874571979,0.2766533494,0.1495649666,-0.2334840745,-0.0543575138,-0.1294789314,-0.0255045816,-0.015468711,0.0540597402,-0.0889275596,0.4138535857,-0.0389394574,-0.2830852568,0.3861259818,-0.0381522439,-0.0971575156,-0.1532618254,-0.0394701101,0.0766005889,-0.0374772809,-0.2887773812,0.1567475349,0.0466397926,0.093797259,0.2489909828,0.0460647494,0.0328915268,0.2356236577,0.0752923787,0.1630280316,0.4459595382,-0.0003144037,0.0400670469,0.3538968563,0.4176515937,0.1609001309,0.1501174271,-0.1202337146,-0.1861605793,-0.1416518539,0.1408617496,-0.1575787514,0.0906012058,-0.4310445487,-0.1330971122,-0.3732341826,0.0073386524,-0.0343621001,0.4378283918,-0.0487489253,-0.1396611929,0.2441136092,0.2374013662,-0.0423724875,0.0900178328,-0.1254783124,0.255615443,-0.2168087512,-0.0520229526,0.3310387731,0.0770800337,0.421222806,-0.2571605742,-0.1213400289,-0.224132359,0.1388938576,0.0859965533,-0.0139247123,-0.064111203,0.4167032838,0.0155231385,-0.1533248723,-0.4158217013,-0.071225442,0.0388117135,0.0347075351,-0.037107002,-0.1740760654,0.3355267346,-0.0862290189,-0.0942458659,-0.0037766567,0.2271474302,-0.1502159536,0.0436013825,0.0858766139,0.3182860315,-0.0827771053,-0.000988554,0.1685232818,0.2986229062,-0.0949452594,-0.2877131999,-0.1839938015,0.1673677117,-0.1132550612,-0.0632707402,-0.044929985,0.2455672026,-0.3794006705,0.2419207394,0.304346174,-0.2896392047,-0.2050281167,0.2148452997,-0.3752191663,-0.1749029905,0.030269891,0.0142999822,0.4108042121,-0.0042553372,-0.1432295889,0.4827906489,0.0771704242,0.242865935,-0.3040136099,-0.466017276,-0.1829135418,0.2075032294,0.1516643167,-0.1184460148,0.1088404134,-0.0877448469,0.3002710938,-0.1562739909,0.1865597665,0.1028919965,0.082430616,0.0962019637,0.0902650356,0.1277572811,-0.4553069472,-0.0245841052,-0.1491093785,0.2400395274,-0.3181627095,-0.1654351503,-0.2285122573,-0.1823611856,-0.2087038606,-0.3032354414,0.2976228893,0.2181787938,0.0589148887,0.2489134073,0.0791672543,-0.078336969,0.1662929952,0.1352360249,0.0483056195,0.3308407664,-0.3822062016,-0.1154452637,0.1777767688,0.0533847101,0.1627697051,-0.0097971642,0.0177794714,0.2022307515,-0.2338354439,0.4560253918,0.0342990011,-0.018487528,0.2643815875,-0.001894454,0.3286045492,0.3785975277,0.1253685653,-0.042322088,-0.1335391253,0.3993208706,0.2978732884,0.1315351278,0.1984390169,0.1492633671,-0.0345823541,-0.3092212379,-0.0366149396,-0.175003618,0.1060321406,0.4917028844,0.3812161088,-0.1171898544,-0.1798555851,0.2268595248,0.3522838056,-0.0431040488,0.2884167135,0.0797794685,-0.3493873775,-0.1552882195,-0.1172491759,0.3525727689,0.1629645526,0.2543209791,0.1712719202,0.0493353643,-0.0080083162,-0.3064471781,0.399212122,0.0655613244,0.1062253714,0.3554275632,0.112254627,-0.1241622567,-0.3668965697,-0.0078185881,-0.0030904301,-0.051934097,-0.2809714377,-0.0359772854,-0.2021116912,-0.1673544347,-0.040416833,0.1189079285,-0.0453860685,-0.1994874179,0.241571486,0.2851106822,-0.2777339518,0.127188921,0.1675787568,0.3435669839,-0.0704192296,-0.051740665,-0.1093210727,0.0156291276,-0.331397295,0.1907478869,0.2792063951,0.0847769007,0.1668625176,0.035354279,0.0824102759,-0.4212260544,-0.1966773272,0.2405076027,-0.1467298567,0.0396607593,0.1082567051,-0.0911808386,-0.0478258096,-0.3472113609,0.0503748134,-0.2066100538,-0.2028837651,-0.3251749873,0.1594581753,0.0629494265,-0.488184154,-0.3987681568,-0.2369147837,-0.380084008,0.1677398533,0.1598065197,0.1822991818,0.1246550754,-0.0548261032,0.1800568253,0.2231410295,-0.1606066525,-0.147671774,0.2775504291,0.4198610485,-0.4095939994,-0.3190635145,0.0439115465,-0.3232084513,0.0133835878,-0.2990259528,-0.2501083314,-0.4383622408,0.1537317634,-0.177799508,-0.0845339298,-0.0026808882,0.2986175418,0.1873589605,-0.2003104836,0.026588954,-0.1110703126,0.2070545703,0.0882992595,0.3847719133,-0.1615374535,0.4067229629,0.1174867451,0.2566284835,-0.26918751,0.0335369334,0.3619191945,-0.0506728664,0.1701949239,0.1592020392,-0.0863516107,0.1363561302,0.0409470275,-0.0944347158,0.5336619616,-0.0854697302,-0.4864979982,-0.1826670021,0.0532185845,-0.273686707,-0.2679212093,-0.1089556739,0.2660419047,-0.1956836432,0.0115898009,0.2149553001,0.0052229138,-0.0608394817,0.2059178799,0.2196903825,-0.0772171617,-0.1206152216,0.1185845733,0.1042682603,-0.3951476812,0.1117486209,0.1043822393,-0.4041220844,0.1074390486,-0.2256135195,0.2036410272,-0.3195444345,0.358741492,-0.1209610775,0.2662653029,-0.0143701937,0.216438055,0.10930226,-0.099338375,-0.2988061607,-0.3211817145,0.450747788,-0.0478757992,-0.265604794,-0.1441638917,0.3384916186,-0.2589463294,0.0323872305,-0.0962956622,-0.0191143472,-0.5297087431,-0.5422529578,0.0772396997,0.1260938197,0.0979745761,-0.1870060712,-0.163955152,0.2704376578,-0.2099725604,-0.1122278869,0.0596685223,0.1827882677,0.141125977,-0.04690478,0.2014990896,0.4859793782,-0.0863522738,0.0502427071,0.0887132287,0.1350985318,0.3584620059,-0.2915494144,-0.0120516224,0.4303926826,0.0423833728,-0.1449910998,-0.337811172,0.1311997622,-0.0050207195,0.0262090676,0.3523302376,-0.0076964782,0.3189740777,-0.1285665333,0.3718144,-0.114938952,-0.1495705545,-0.1133223847,0.137941882,-0.4891560674,0.2023630291,0.249479115,1.0095961094,0.1828901321,0.1841591895,0.3770165145,-0.1828773022,0.2244256288,-0.2048053294,0.2420448959,-0.275912106,0.2594016492,0.0270561446,0.123555094,-0.2554804087,0.2857078016,-0.3485202491,0.2226643264,0.1918338388,-0.0550935976,0.0397399962,0.5254421234,-0.414765358,-0.3550721705,-0.0840060115,0.2253390402,0.2074783742,0.3336937129,-0.0224050712,-0.3262494206,0.0230254158,-0.3689983487,-0.1200434417,0.222809732,-0.6562026739,0.0811413154,-0.4205805063,-0.2054655552,0.0714982525,0.1216394231,0.1825933158,0.1002613306,-0.0640582293,0.2806988955,0.0451345742,-0.0464380011,-0.0050444896,-0.117036812,0.2287954986,-0.254381597,-0.2008146793,-0.052955918,0.150778234,-0.2844575346,0.0494143516,-0.0253137629,-0.2893356681,0.1632657945,0.0008363276,0.0888023153,-0.4070391357,-0.1701178551,0.2157660872,-0.2228650451,-0.0063023539,-0.0420259535,0.0753558651,-0.1657579988,-0.1717545539,0.4114311039,-0.1920544058,-0.3257522881,0.5842943788,0.0643735826,-0.1504262239,-0.3654353917,-0.2140043378,0.0150121469,-0.1508513838,0.1102009565,-0.1128665805,-0.256216377,-0.0146313105,0.5000475645,-0.0012302455,0.2679022551,0.0265189763,-0.0871425867,-0.0991812721,0.2623232007,-0.1346050203,0.2985242903,-0.0751187578,0.0820538327,0.1869878322,-0.1457145363,-0.4634034336,0.0790768564,-0.1712537557,-0.1405716389,0.1250740886,-0.2987323999,0.3095208406,0.1704262793,0.2733668387,0.2205571383,-0.077256754,-0.2663613558,-0.0304335151,0.07006412,0.0473003574,-0.1006334201,0.0123664159,-0.1115101725,-0.218509391,0.1263172477,0.2010792345,-0.1831626147,0.0833370537,-0.1032224596,-0.2481856495,0.0976704136,0.2554208934,-0.1850833893,-0.0976004004,0.4207085371,0.0166867264,0.1075952277,0.0078057367,0.0033925159,0.0431230143,0.1411653012,0.0564506911,0.3387710452,0.2962433994,0.0672249943,-0.5515539646,0.0619442984,0.2637961805,0.078537114,-0.2515684962,-0.1156085432,0.2228315324,0.3556102216,-0.259417206,0.1732178032,-0.1623107642,-0.4482299685,0.3617484272,0.2996283472,0.1987316757,0.0672524571,0.0475771576,0.1538341343,0.404625982,-0.0747884288,-0.092839621,-0.0525272973,-0.1224941462,0.1339276731,0.185081616,0.067402795,0.3257806003,-0.1616298556,-0.1871491969,0.008799796,0.0235794131,0.2640464604,0.2166270763,0.0519343317,0.03836805,0.081463933,-0.2451706231,0.0961864218,-0.0673369393,0.2200666219,-0.3234996498,0.0240175724,-0.3409782052,0.4161132872,-0.0618106611,-0.0085473182,-0.2516661286,-0.3260459602,-0.6693037152,0.1929431558,0.0450478755,-0.1420245618,-0.4067165256,0.1091223657,-0.0436599553,0.0488749109,-0.2354439199,-0.09591306,0.3324644864,-0.4127458632,0.1486727297,0.0149962129,0.1223974302,-0.1983048767,0.3227460682,0.2554567456,0.2159596384,-0.3102447391,-0.2035475373,-0.079424113,-0.0601548404,-0.144217059,0.037109971,0.3265762031,-0.1694569141,0.1500537395,0.2160780877,-0.3853534162,-0.0577884056,0.2834987938,-0.2225657403,0.2526809275,0.1624028534,0.2718719542,0.0135586848,-0.3872140646,-0.3066829741,-0.5732123256,-0.1348943859,-0.2507286668,-0.1008889601,0.2407356501,-0.3101221323,0.1533886641,0.1174057275,0.180290997,0.3067573309,-0.0013956358,-0.1100461632,-0.2357510924,-0.439438194,0.0553815812,0.2680011988,-0.1324961334,-0.219313547,0.4046549201,-0.0119787026,-0.0210185908,0.0093867835,0.5449488163,0.2535102069,0.0710766315,-0.1868591011,0.1759403348,0.1250489652,0.1145300791,0.0818908438,-0.1827889383,-0.1159717515,-0.2869517803,0.214859724,-0.1342401356,0.1613276899,-0.3132973611,0.0037505045,0.4126843512,-0.2459646463,0.2662286758,-0.286221087,-0.1534481794,-0.1860763282,0.2085224092,-0.295242399,-0.3550876975,0.2407322079,0.1949659288,0.0700791925,0.0669670627,-0.2691867948,0.2615852654,0.0445120037,0.0049234945,-0.1911266148,0.11490722,-0.0623385981,0.0978202522,0.0315187462,0.205317378,-0.1249731705,0.0482877232,-0.495287329,-0.5152050853,0.1610213369,-0.0524530187,-0.086882256,-0.0676097795,0.3968174756,0.2529715598,0.1593210995,-0.2931996584,0.125608176,-0.0083493926,0.1823005974,-0.1703491807,0.1373269558,0.138203606,-0.1351919025,-0.0715939626,-0.2012002319,0.3138321936,0.0080849808,-0.206582889,-0.3480609059]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/414","title":"from_dict delete?","comments":"`from_dict` was added in #350 that was unfortunately not included in the 0.3.0 release. It's going to be included in the next release that will be out pretty soon though.\r\nRight now if you want to use `from_dict` you have to install the package from the master branch\r\n```\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```","body":"AttributeError: type object 'Dataset' has no attribute 'from_dict'","comment_length":53,"text":"from_dict delete?\nAttributeError: type object 'Dataset' has no attribute 'from_dict'\n`from_dict` was added in #350 that was unfortunately not included in the 0.3.0 release. It's going to be included in the next release that will be out pretty soon though.\r\nRight now if you want to use `from_dict` you have to install the package from the master branch\r\n```\r\npip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n```","embeddings":[-0.1168889999,-0.2096914053,-0.1167027056,0.0465560853,0.1620283723,-0.1888360828,0.094357267,0.3141218424,0.0377569273,0.1403821409,0.0539773069,0.5944421887,-0.1069423631,0.2920179665,-0.0426585861,-0.1649783999,-0.0688440055,0.3200041056,-0.3128439188,-0.0461874977,0.0110787461,0.3731954098,-0.1320789903,0.2549414635,0.1205590218,-0.1711758375,0.0303646121,0.1430951506,-0.1418486983,-0.331137538,0.2542820275,0.0454760641,-0.1363679469,0.1577937752,-0.000099203,-0.0431271009,0.3241799176,-0.0708885342,-0.2661366463,-0.3808972239,-0.1292472035,-0.1784549057,0.3302332461,-0.1057158113,-0.2622191906,0.148300916,0.0286048111,-0.0083772056,0.3838491142,0.3884246945,0.3685189188,0.1577926576,0.1676738858,-0.1256329417,0.1861926913,0.1869291961,-0.0660707951,-0.0968847573,0.0539165996,-0.1805331856,0.3063396215,0.2829711437,-0.1644661874,-0.0961181298,0.1874166429,-0.0243609287,0.0242319051,-0.1416930109,0.0300105345,0.1506774873,0.2641409039,-0.3806945384,-0.1140294224,-0.301546067,-0.0242865924,-0.3830109835,0.0237073582,-0.0531720333,-0.1575684696,0.1301054358,-0.0527608097,-0.4518622458,-0.1853362769,0.3525791764,0.1932297796,0.0689729527,-0.1127059087,0.0655093715,0.2613548338,-0.0540376678,-0.3767986,0.2988931835,-0.1951348484,0.1416613609,-0.0280430634,-0.2018370926,0.3356459141,-0.0524854325,-0.0726941824,0.1180981249,-0.0129491165,-0.0188899003,-0.0099218143,0.2746320367,0.4173329771,0.1929970533,0.3750910759,0.2258402407,0.1976595968,0.0005200541,-0.0670649111,-0.0705474317,0.1084683463,0.0079595437,-0.1542845666,-0.0837850049,0.2740116119,-0.1018361449,-0.2801439464,-0.2536678016,0.1547212452,0.1114788353,0.0496223792,0.3161873817,0.0285122562,0.2045546919,0.0234101508,0.1625921279,-0.3559416831,-0.0318716839,-0.2442645878,0.3273057044,-0.1398757249,-0.47908324,0.0662470534,0.1737167984,0.3600257933,-0.1125761643,0.0398115031,0.2609203756,0.0090481127,0.0365596674,0.0756462812,0.2162600309,-0.0976463109,-0.1582449973,0.2198809236,-0.535482049,-0.2740392387,0.1050391942,-0.2255008072,-0.2852151394,-0.450623244,0.3046717346,0.0804250538,-0.2931632996,-0.3293567002,0.1842859983,0.1359829754,-0.216448158,-0.1059303656,-0.1513110548,-0.0192243401,-0.1611665487,-0.0074461503,0.1157196164,-0.2353723943,-0.209140107,-0.2025369406,-0.3065539002,0.2619755566,0.3139546514,-0.2446835637,0.2455075383,-0.1226886734,0.1057924107,0.5160627961,-0.5724243522,-0.2312455475,0.2054415643,-0.0490397513,-0.1142272428,-0.1684024781,-0.017951373,0.1819056571,-0.0890898108,-0.0099350531,0.3474216759,0.1860726625,0.1431891173,-0.1495317519,-0.1222073957,-0.0283667054,-0.0966636017,-0.0759283155,0.0703809261,0.1684971303,0.1468969434,0.2025403529,-0.094522506,0.009312707,0.1962150484,0.5006925464,0.0556181557,-0.093551971,-0.4382086992,-0.0323105566,-0.139447704,-0.1978676319,0.3533257842,-0.318826437,-0.1176964566,-0.2525280714,0.0997808427,-0.2223487645,-0.1572247148,0.3052579463,0.0586955361,0.143273145,0.0759015009,-0.3448251784,0.1163838208,0.0605933741,0.0451134145,-0.2440586537,-0.0218843371,-0.3494740725,-0.2440223992,0.0595742241,0.4624376595,0.0109009612,-0.0529735908,0.0195062347,0.3652351201,-0.0927099735,0.0546632409,0.1078714877,-0.1840557158,0.0560944751,-0.2318473458,-0.0272628814,0.0637400672,0.0834897533,0.1336135119,-0.10646566,0.1729995161,0.2040770501,-0.0279307161,0.2542048097,0.1592370421,0.3644242585,-0.2514116168,0.1046667099,-0.3587047458,0.2191609293,-0.0493634529,-0.0715416819,-0.1774632037,-0.423636198,0.0452174321,0.6067508459,0.155034408,0.213289544,0.095278956,-0.0068472968,-0.1788431257,-0.05146119,0.0949727669,0.1725858301,0.2499797642,-0.1400128007,0.0440866053,-0.0233692378,-0.0711451247,0.3133590221,-0.1102237701,0.1291939616,0.1159423962,0.1464599818,-0.1220927611,-0.4732962847,-0.0990835652,-0.2719185948,0.31240803,-0.1202232242,0.2324871868,-0.3833058178,-0.511946559,0.029831389,-0.2020338774,-0.3217611313,-0.3416158855,0.2300844043,0.1038933098,-0.0862878338,0.1916826814,0.1746332943,0.0412034765,0.0341077857,-0.1091085002,-0.0448255539,-0.3886269331,-0.2937385142,0.2078532726,-0.0464740433,0.1856131107,0.34539029,-0.0927943438,-0.1346259564,-0.0653238818,-0.4512675405,0.1339701116,-0.1982107759,0.4171160161,0.4988261759,0.0458737314,-0.1586198509,-0.1184889674,0.2855276465,-0.5542433262,-0.2418099046,0.1625321209,-0.0430538021,0.0001885036,-0.2799878418,-0.2784693837,-0.2698887289,-0.4429297745,0.3062042296,0.2609966099,0.2338288426,0.3336763084,-0.0887839645,0.1830664426,-0.1976733357,0.3888370395,-0.1767815799,-0.1744702905,-0.0336391032,-0.2309451103,-0.4337511659,0.0063298405,0.1090219766,0.3247979581,-0.2439881861,-0.3170567751,-0.2650544345,0.0487910323,0.1092738509,0.0224366449,0.1758804917,0.2571715117,0.1897458881,-0.2251132876,-0.2148870528,-0.0785518885,0.2249791026,0.201737389,0.111191377,-0.0416105129,0.2193104327,0.1520111561,0.362911284,-0.0490334257,0.1733387113,0.328886658,-0.123388432,0.4986999035,-0.1633700728,-0.3937039673,-0.0244874656,0.2137499005,0.1091391668,0.1704998314,-0.0209091175,-0.2336777449,-0.2458532006,0.0355988704,-0.210914284,-0.216230765,-0.0307531543,-0.2094967067,-0.0089890584,0.0670689866,-0.0869473815,-0.236055553,-0.1214585826,-0.0317260176,-0.0994309187,-0.1177403107,-0.1608858556,-0.5274441838,-0.3798387051,-0.4278229177,0.2465217859,-0.0692678317,0.1195460483,-0.1371690333,-0.1768676788,0.0999169126,-0.128868252,0.3627329469,-0.004335789,-0.1355555058,0.1412733048,0.0477272868,-0.3704724908,0.2148815989,-0.2964704335,0.2145401835,0.4119858146,0.0834664032,-0.4349457026,-0.2798308432,0.3200950623,-0.0390959196,-0.1202356741,0.1618261188,-0.2939400673,-0.3636198342,-0.544957757,-0.1299726069,0.0731195137,0.3367328346,0.1882373095,0.1335387379,0.0055836467,-0.1185669005,0.0229387004,0.0187420864,0.2522697151,-0.2153329998,-0.3697969913,-0.2291405052,0.2577160597,0.0217592623,0.4180098176,-0.1041851118,-0.2362633646,-0.2990650237,-0.0036223955,0.4200933576,-0.0112966644,0.0881404057,0.2431387603,-0.1229887083,0.2615291178,0.1734337807,0.0778044015,0.2646639943,-0.0300039034,0.0469723791,-0.2418219149,0.1561151296,-0.0562052727,0.0065016346,0.0302877221,0.0922401994,-0.2980342507,0.3624907136,-0.1737013906,0.8276715875,0.1541241556,0.149521485,0.1627134979,0.0200281236,0.6298738122,0.1459995359,0.3014276624,-0.2305775732,0.2545523047,0.0371038392,-0.048094511,0.2337826043,0.0967631117,-0.2803462148,0.3622639179,0.0007974776,0.0417945832,-0.1365338862,0.1986435205,0.0912431031,-0.0252849404,-0.3467203677,0.2415170074,-0.1234219074,0.1706300378,-0.035270229,-0.1865023077,0.1307333708,-0.2843438089,-0.1883136779,0.1650152653,0.0355023369,0.3320157528,0.2341720313,-0.3574246466,-0.1196799055,0.2072726637,0.1231377944,0.0464342497,-0.1416504532,0.1934479624,-0.150472641,-0.1314358413,0.0581149459,0.3791010678,0.4059224725,-0.3037533462,-0.3127824366,0.1355921328,0.0106381308,-0.4071266949,0.2054925561,0.0124373157,-0.0358399898,0.0005648024,-0.0626730621,0.1061061993,0.0027788281,-0.2611286938,0.2479954809,0.2681899667,-0.2971423268,0.1301719248,0.0277893115,-0.133281976,-0.0597800277,-0.0008432104,0.1286654919,0.0012186307,0.4648866951,0.1023398414,-0.1767463982,-0.4118673503,-0.2713217139,-0.1673767865,-0.2263266146,0.3819690645,0.2274310142,-0.180946961,0.1564970613,0.4324983656,-0.1519449502,0.0465012975,-0.2163886428,-0.2470472455,-0.1826239079,0.1593076587,-0.047622405,0.086260587,0.1756960899,0.1856440753,0.11674501,0.0309038255,-0.4574420154,0.0554917827,-0.3008749187,0.1754256934,0.431032896,-0.2218666375,0.1155421585,-0.0534149073,0.2771427035,-0.1712550819,-0.1608266979,-0.3855388165,-0.0767776892,0.0738792121,-0.0763095394,-0.0384725519,0.0168642793,-0.0655055493,-0.1414912194,-0.0536488891,0.1224134862,-0.0986469314,-0.044729542,0.5866776705,0.0955242738,-0.0281800125,0.3729470372,-0.0531996824,0.0246782862,0.1688961536,-0.1014152765,0.1502834111,-0.0921994373,-0.1458046436,-0.0007061277,0.0461677574,0.0317956395,0.1970802099,0.2288706303,-0.290632695,-0.092196703,0.0392900258,0.4682260454,-0.0393222421,-0.0785331503,0.0475337021,0.3697397113,0.3186566234,-0.4926092327,0.0549898073,0.2415742725,-0.0794779211,0.0962627977,0.1426690221,0.0707423314,0.0915965661,0.0073609003,0.11958161,0.2814496458,-0.3491170406,0.0857382044,0.2375695258,-0.2161261737,-0.1672631055,0.1247375533,0.2451853305,0.3118290603,0.6150824428,-0.0412337445,0.1317596287,0.1235011816,0.0718664676,0.1753038019,-0.0961659923,0.2723702788,0.1497651339,0.1668110788,0.0526756793,0.143591404,0.3294469416,-0.2138656676,-0.0179159194,-0.0795616433,0.6706982255,-0.1344926953,-0.1535613984,-0.2235077769,0.0610391237,0.1749209911,-0.1438238174,0.1336372197,-0.0305898748,0.0905559808,0.1358239651,0.0513514094,-0.4324580133,0.316131562,-0.1809588522,0.1757671684,-0.0775606409,0.0436395295,-0.0343453214,-0.0201536492,-0.1533806473,0.3481298387,0.4525027275,0.1537193656,-0.2976202369,-0.0402155295,-0.1711034775,-0.2158661932,-0.1685933769,0.3800278604,0.1822301149,0.4804610908,0.1164074913,0.2183573246,-0.2005445361,-0.1231264472,0.1783151925,0.0904934779,-0.1555392742,0.575304389,0.0531927384,-0.3424189985,-0.1728662252,0.03833564,-0.4150214791,0.0661642775,0.294295162,0.0953267142,0.1888163686,-0.3335586488,0.1745780855,0.1563628912,0.3946991265,0.3342249095,0.2382938713,0.1870157868,-0.1841732562,-0.4572910368,0.078884989,-0.0789052323,-0.2765892148,-0.0074919821,0.3077087402,-0.0091798333,0.2107315809,-0.0831295103,-0.0004255717,-0.0617790371,-0.4295384288,-0.2338014245,0.0362691768,-0.100089781,-0.0556546822,0.1671201736,-0.2100697756,-0.1162021011,0.0270134285,0.176995337,-0.1185852364,0.1766653955,-0.0098192338,-0.0512385853,-0.1382741481,0.298473835,0.1597542912,-0.1168615296,-0.1258471459,0.187639907,-0.360920608,-0.3744492829,0.1019270271,-0.0730052367,0.3420498669,-0.1621476263,0.1303639859,-0.1603464484,0.1503836513,0.0480322391,0.0211187843,-0.2224087119,0.1399756521,0.1745988578,-0.1314189732,-0.1432362944,0.1130637154,-0.1071955487,0.056935966,-0.2670480609,-0.3180978596,0.461589694,-0.3228894472,-0.4482115507,-0.0576568469,0.1911208332,-0.0651581213,-0.3452365696,-0.4551792443,0.1387112737,0.2597662508,0.070532307,-0.0792654678,0.4142343402,-0.2969504297,0.0366846509,0.1778421551,0.4717923105,0.2032424212,-0.1371378303,0.0796471536,-0.1601916105]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/414","title":"from_dict delete?","comments":"> `from_dict` was added in #350 that was unfortunately not included in the 0.3.0 release. It's going to be included in the next release that will be out pretty soon though.\r\n> Right now if you want to use `from_dict` you have to install the package from the master branch\r\n> \r\n> ```\r\n> pip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n> ```\r\nOK, thank you.\r\n","body":"AttributeError: type object 'Dataset' has no attribute 'from_dict'","comment_length":62,"text":"from_dict delete?\nAttributeError: type object 'Dataset' has no attribute 'from_dict'\n> `from_dict` was added in #350 that was unfortunately not included in the 0.3.0 release. It's going to be included in the next release that will be out pretty soon though.\r\n> Right now if you want to use `from_dict` you have to install the package from the master branch\r\n> \r\n> ```\r\n> pip install git+https:\/\/github.com\/huggingface\/nlp.git\r\n> ```\r\nOK, thank you.\r\n","embeddings":[-0.0747189745,-0.2516213059,-0.1155344173,0.0879213288,0.1837774664,-0.1787118465,0.0908716917,0.2998166084,0.0316059254,0.1319086105,0.051009573,0.5649474859,-0.1044338271,0.2966352105,-0.0471764207,-0.2244530171,-0.0426336601,0.2933095992,-0.3304531276,-0.0517251752,0.0404684059,0.3629773259,-0.1296854913,0.2741340995,0.1252758354,-0.1901750714,0.001523576,0.1526876837,-0.1340594888,-0.3176116943,0.2797694802,0.0268054567,-0.1014018655,0.1761731356,-0.0000985522,-0.0737006441,0.3171120882,-0.0745516494,-0.2624133527,-0.3476303816,-0.1504907906,-0.1589511782,0.2975644171,-0.1199103072,-0.2388195693,0.1387144476,0.0045639747,0.0186313447,0.3920916617,0.3837104142,0.3749192655,0.1249384284,0.1929481477,-0.121417217,0.1639735997,0.1866833717,-0.0807384998,-0.0862299129,0.0554489754,-0.1900994629,0.3049733937,0.2913117707,-0.1488623619,-0.1117041856,0.1986547112,-0.0462510139,0.0224084966,-0.159161672,0.0389379784,0.1428239197,0.3123072684,-0.3866829574,-0.1058308482,-0.2720894814,-0.0205371641,-0.3580051363,0.0111028478,-0.0354096703,-0.127446726,0.144500792,-0.0395705551,-0.4792864919,-0.2171789408,0.3515553772,0.2264175266,0.0382249355,-0.1236405745,0.0532368906,0.2304850519,-0.0637611374,-0.3869415224,0.2951688766,-0.1951296329,0.1586609036,-0.0446749255,-0.2178733796,0.3263928592,-0.0638526082,-0.0525567643,0.0825085267,0.0046483949,-0.0221263003,-0.0281217378,0.2544954121,0.4578906298,0.2475712299,0.3936284482,0.2212034017,0.1548516899,0.0250813756,-0.0860896185,-0.0489335395,0.0843025744,0.0313623026,-0.1318344325,-0.0685866177,0.2815732956,-0.1133547574,-0.2768691778,-0.206428647,0.1209661365,0.1106280833,0.0282629132,0.3377598524,0.0086521134,0.2231969088,0.0360624008,0.1516708881,-0.3684809208,-0.0670754686,-0.2348340005,0.3588404357,-0.1695245504,-0.4798001647,0.0434884764,0.175326407,0.3514825106,-0.1070454344,0.0640288368,0.2143966556,-0.005742793,0.0531018078,0.0964144394,0.1829386801,-0.0929335877,-0.156086877,0.210033372,-0.5214315653,-0.252196312,0.1298751682,-0.2238720655,-0.2987514734,-0.4562395811,0.2990592718,0.0899032652,-0.2795068324,-0.3280282915,0.1972243041,0.1388174295,-0.1668095142,-0.0827485323,-0.1410974413,-0.0044679153,-0.1480266899,-0.024340041,0.1140932292,-0.2422045469,-0.1710827649,-0.1728948206,-0.3147648573,0.2596482635,0.3290223181,-0.262842983,0.2250309587,-0.1182504892,0.1163873598,0.479701817,-0.5727295876,-0.1921099126,0.2145445198,-0.0597661547,-0.102557525,-0.1759753078,-0.0275619011,0.1942097992,-0.109426111,-0.0382939875,0.3194501102,0.1817690879,0.1489367485,-0.2012803257,-0.0979351029,-0.0093584759,-0.1098808944,-0.0755400434,0.0522700548,0.1559681892,0.1709434241,0.2374147177,-0.0858651698,0.0225113519,0.2546881735,0.4777030051,0.0747559741,-0.0862904862,-0.4368614554,-0.0294936486,-0.1539904773,-0.1910116524,0.3839293718,-0.2584607303,-0.1255776882,-0.2810029984,0.1069899797,-0.1987516731,-0.1712756753,0.3102854788,0.0364366509,0.128769666,0.0446912907,-0.3484444618,0.1158475727,0.029240394,0.0292404294,-0.2396176904,-0.0224088896,-0.366972357,-0.2318977118,0.101094231,0.4514656663,0.0169409215,-0.0570092387,0.029873278,0.3622139096,-0.107615076,0.0602518618,0.1294952929,-0.1758580655,0.0612258874,-0.2530985177,-0.0331598967,0.0573220812,0.0873133913,0.1328568459,-0.0833986476,0.1473474205,0.1943663806,-0.0475816242,0.2373598367,0.145862475,0.3561370671,-0.2580803931,0.0934287906,-0.3171673715,0.2263148725,-0.0559443198,-0.1203266159,-0.1707588434,-0.4570818841,0.072836414,0.6142229438,0.1181451082,0.2027907521,0.1100267321,-0.0104902601,-0.2033447474,-0.0465196967,0.1275081933,0.1598938704,0.2393365949,-0.1442097872,0.040248815,-0.0160735231,-0.0424356349,0.3344411254,-0.0953002125,0.1490649432,0.0981971845,0.1295661926,-0.1051477864,-0.4526612759,-0.1029855534,-0.2659780681,0.3086747527,-0.1300005466,0.2137014419,-0.3413686454,-0.5405299067,0.0080291582,-0.2147608399,-0.3553180695,-0.3191390634,0.2329045534,0.0794749558,-0.1158465967,0.1832075566,0.1557338089,0.0698950738,0.0368973091,-0.0881305784,-0.0506448783,-0.3942842782,-0.3033125699,0.2088561654,-0.030275939,0.21846506,0.3349041045,-0.092314288,-0.1171327457,-0.0730040818,-0.427192837,0.1169190109,-0.2132866383,0.4308361709,0.4775032103,0.1004523262,-0.1677221805,-0.1249687374,0.274867624,-0.5327470899,-0.2540537715,0.1580503881,-0.0154889198,-0.0048853084,-0.27590397,-0.2948980033,-0.2445005029,-0.4642463923,0.3006665707,0.2462503165,0.2239492089,0.3138276041,-0.0987153798,0.1614603102,-0.1947257966,0.4215764701,-0.1935763955,-0.1916532964,-0.0404682308,-0.2186050117,-0.4552693665,0.0204544384,0.1377618909,0.3500982821,-0.2896838188,-0.3144922554,-0.2251933664,0.058029402,0.0870286822,-0.0012439908,0.1882114261,0.2421658337,0.2206172049,-0.2163992375,-0.2033874989,-0.1030128598,0.2228536159,0.2019103914,0.1072848216,-0.0429824851,0.2035516948,0.142375052,0.3852213323,-0.0343018472,0.1985272169,0.3196974397,-0.0675486848,0.4661508203,-0.1550482363,-0.4070017934,-0.0091275973,0.1755029112,0.1023463309,0.1916301847,-0.0077863792,-0.2419212312,-0.2517975867,-0.0047007701,-0.2284081131,-0.2309152931,0.0008257007,-0.2384738624,0.0005100101,0.0652895123,-0.1143400669,-0.2259874642,-0.1049547344,-0.0050723068,-0.1000623479,-0.0912701562,-0.1490985602,-0.5165436268,-0.3555939794,-0.421189636,0.210348174,-0.0652920827,0.1419126987,-0.1439503133,-0.17875126,0.1019724533,-0.1125372499,0.3513404131,0.0308236834,-0.1319409013,0.1510486603,0.0532494001,-0.3147056699,0.2232507318,-0.2976679504,0.1993822306,0.3850404918,0.0479152538,-0.4532953501,-0.2799518108,0.2857570052,-0.0285098795,-0.1267178208,0.1414327919,-0.2920064628,-0.3294558525,-0.5601935387,-0.1146838591,0.1040034294,0.3459015489,0.1609196216,0.1414265931,0.0273951385,-0.1401552707,0.037916556,-0.0090295281,0.2666153014,-0.2080396265,-0.3748337924,-0.2492122054,0.2636886835,0.0184715688,0.4048213959,-0.1258100867,-0.2229819596,-0.3117398322,-0.030758366,0.4273918867,-0.012826005,0.1002398655,0.2510443032,-0.1236005649,0.2783620358,0.1314157546,0.0324288085,0.2652900815,-0.0184586756,0.0434190892,-0.2153191268,0.1376204491,-0.079584524,0.0065778317,0.0342803709,0.1068119332,-0.3075437844,0.344707638,-0.1606524885,0.8757866025,0.1566372812,0.1482198834,0.1443024278,0.0060097179,0.623742044,0.1574923843,0.3093409836,-0.2359822989,0.2198898643,0.0453437418,-0.0233531874,0.2113473117,0.0939024314,-0.2781503201,0.341594398,-0.011055164,0.0125486609,-0.1442002505,0.2061146647,0.0848263875,-0.0195671916,-0.4315616488,0.2368656844,-0.1003966182,0.1428439766,-0.066538848,-0.2186398506,0.1212835386,-0.272521764,-0.1870218217,0.1411041319,0.0477349758,0.2928524911,0.2638768554,-0.3232089877,-0.1018036678,0.229443267,0.1322084963,0.0215004273,-0.1622178406,0.1684937626,-0.1676187664,-0.1405907273,0.0638556033,0.3672299683,0.4578964412,-0.3215275407,-0.3099904954,0.1329866648,-0.0128236348,-0.377673924,0.1706566066,0.0170445647,-0.047989592,-0.017922366,-0.0911345705,0.1027874872,-0.0172750615,-0.2291879505,0.2480984479,0.2937904298,-0.3036979139,0.1194648743,0.0586557239,-0.1238421053,-0.0439162739,-0.009820722,0.1302970648,0.0020776906,0.47631073,0.1353935599,-0.20083417,-0.4135890901,-0.2553914189,-0.1405200809,-0.2420743704,0.3765794337,0.2040150166,-0.1696826369,0.1431548893,0.4442288578,-0.11122033,0.0574417561,-0.1993821859,-0.2484329939,-0.2025046051,0.1892824024,-0.0655541644,0.0782250166,0.142544046,0.1986285895,0.1594912857,0.0600586943,-0.4667700231,0.0484206341,-0.3432683945,0.1815657765,0.4437235296,-0.2099007368,0.1012222469,-0.0235764422,0.2671224177,-0.1746080667,-0.153897807,-0.3755474091,-0.0732358247,0.0731340125,-0.0796449184,-0.0282618999,0.0293466784,-0.0664709732,-0.1379357725,-0.0232475586,0.0933142826,-0.1077141315,-0.0335229039,0.5934742689,0.0909658894,-0.0071243821,0.3224605918,-0.0606763437,0.0415875427,0.2016901374,-0.0930852592,0.1157102883,-0.0881643742,-0.1559306681,-0.002357542,0.0622478127,0.0199803915,0.1833561063,0.2383450121,-0.287448436,-0.0894558504,0.0469670445,0.4964928627,-0.059453018,-0.100984998,0.0709990934,0.38264364,0.3303325772,-0.5057243109,0.0831441954,0.2515092492,-0.118150048,0.0589179397,0.1303807646,0.0547285452,0.0908322483,0.0286451075,0.1321124882,0.2479617596,-0.3675575852,0.071939759,0.2021697164,-0.2696197331,-0.1480326056,0.1263964325,0.2348931432,0.3072808087,0.6015740037,-0.0306528844,0.1460761428,0.136285767,0.0902565941,0.1895992011,-0.0361280367,0.2644059658,0.1784321368,0.1773348749,0.0430850387,0.117386125,0.3615167737,-0.1947198659,-0.0503715463,-0.0662263334,0.6730319858,-0.1513729244,-0.13842462,-0.2244410813,0.0297664609,0.1638586223,-0.1666562706,0.1155286729,-0.0265722312,0.1167303398,0.1477523297,0.0404452421,-0.4394561648,0.287722379,-0.1617629081,0.1960939765,-0.0942888856,0.0443686545,-0.0271909032,-0.0181137603,-0.1653067023,0.35921821,0.4551663995,0.1674995422,-0.2915835977,-0.0534514301,-0.1397529244,-0.2116463482,-0.1342041194,0.4058215022,0.2126342803,0.461016655,0.1347284019,0.2065756023,-0.1926115453,-0.1079936996,0.1983430684,0.0853297636,-0.1661788821,0.5535987616,0.0306686591,-0.3605210185,-0.2123486549,0.0388970301,-0.4257965684,0.0837273821,0.2968178391,0.0673176348,0.2045307457,-0.356780678,0.1773058176,0.1780985892,0.4015876353,0.3509917557,0.2550213635,0.1661327034,-0.1810497791,-0.4496583641,0.0841700807,-0.050788898,-0.241899386,-0.0206428729,0.3006988764,0.000259471,0.2212192863,-0.0865427554,-0.008487219,-0.0446576551,-0.3914658725,-0.248638615,0.0468301028,-0.0883680657,-0.0308397003,0.1717141122,-0.1824656129,-0.1235304475,0.0278107952,0.1662775129,-0.1078459322,0.1905750334,-0.0070866453,-0.0322084054,-0.1730866581,0.2991682291,0.1324914694,-0.11775776,-0.163381353,0.1962734312,-0.3696710169,-0.3476009071,0.1061205938,-0.0462089963,0.3358204365,-0.1556408554,0.1364307255,-0.1630212516,0.1335616112,0.0356467962,0.0332460925,-0.2534837127,0.1212149858,0.1745646894,-0.1633093804,-0.1208329052,0.1273242533,-0.1131867841,0.0501785539,-0.290261656,-0.3075346053,0.4495240748,-0.333569169,-0.459261328,-0.0600973293,0.2076717913,-0.0630892813,-0.3752245605,-0.4853552282,0.1419934332,0.2410847843,0.0844379067,-0.076874882,0.4214234352,-0.3127444088,0.0632280335,0.1830020696,0.4975370169,0.222159341,-0.0971205831,0.0885697976,-0.163336128]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/413","title":"Is there a way to download only NQ dev?","comments":"Unfortunately it's not possible to download only the dev set of NQ.\r\n\r\nI think we could add a way to download only the test set by adding a custom configuration to the processing script though.","body":"Maybe I missed that in the docs, but is there a way to only download the dev set of natural questions (~1 GB)? \r\nAs we want to benchmark QA models on different datasets, I would like to avoid downloading the 41GB of training data. \r\n\r\nI tried\r\n```\r\ndataset = nlp.load_dataset('natural_questions', split=\"validation\", beam_runner=\"DirectRunner\")\r\n```\r\nBut this still triggered a big download of presumably the whole dataset. Is there any way of doing this or are splits \/ slicing options only available after downloading?\r\n\r\nThanks!","comment_length":35,"text":"Is there a way to download only NQ dev?\nMaybe I missed that in the docs, but is there a way to only download the dev set of natural questions (~1 GB)? \r\nAs we want to benchmark QA models on different datasets, I would like to avoid downloading the 41GB of training data. \r\n\r\nI tried\r\n```\r\ndataset = nlp.load_dataset('natural_questions', split=\"validation\", beam_runner=\"DirectRunner\")\r\n```\r\nBut this still triggered a big download of presumably the whole dataset. Is there any way of doing this or are splits \/ slicing options only available after downloading?\r\n\r\nThanks!\nUnfortunately it's not possible to download only the dev set of NQ.\r\n\r\nI think we could add a way to download only the test set by adding a custom configuration to the processing script though.","embeddings":[-0.4314169586,0.1772087067,-0.1129720211,0.1146225333,-0.1070222557,-0.034727104,-0.2576389313,0.7160260677,-0.0329578072,0.3145226538,-0.2831300795,0.0151575794,-0.2205640823,0.6092382073,0.4744931757,-0.0987759307,-0.1157774776,0.4419937134,-0.0799669698,0.0518985651,-0.1738967299,0.2375749052,-0.0511896051,0.1008467898,0.3044531047,-0.2877178788,0.1591368765,-0.0306114946,-0.3407664597,-0.2889768183,0.022300493,0.2152756155,0.174672693,-0.3234871924,-0.0001247323,-0.0638319403,-0.0178546533,-0.1316623539,-0.1318190396,0.1239356697,-0.1069383696,-0.133865431,0.0931724682,0.011332579,-0.0697069094,-0.2781607509,0.2135231346,-0.0861073732,0.3582045138,0.4454734623,0.0204185676,0.1405423582,-0.2529484332,0.1526878625,-0.0644539297,-0.282286644,-0.0892292708,0.0156455152,0.1498271525,0.135431245,0.1219885722,0.1407377571,0.2157315165,0.2066383064,0.0054100887,-0.0171899367,0.3492211103,-0.7237771153,-0.0792933777,0.7125211954,0.1285497844,0.1623681486,-0.1726223826,-0.1767225862,0.1247088239,-0.4916466177,-0.0423816964,0.5084100366,-0.565615952,-0.0027759231,-0.6391010284,-0.5820971131,-0.1727650017,0.1160465032,-0.0537559353,-0.0337662585,0.0801800266,0.0538342148,0.2881833017,0.1955110282,0.0965853035,0.1249824688,-0.1501685828,0.1346161813,-0.2485608608,-0.4248794615,-0.2950527072,-0.1588326544,0.2999621332,0.0453286581,-0.0710311085,0.0286634341,0.072369352,-0.1054556295,0.408921808,0.1076537222,0.284154892,0.231772691,0.0504684858,0.021156406,0.0802846327,0.1652096063,-0.0076865098,-0.0758912042,-0.6195682287,0.0124149127,-0.0180410407,-0.431645453,0.0045180675,-0.1753554791,-0.159151867,-0.1118600443,-0.1868516058,0.1990063787,0.0409796089,0.0766252801,-0.297126174,0.2975178957,-0.2782457471,-0.8463834524,-0.1588612795,0.0120070577,0.0453217141,0.1391306669,0.3189242482,0.1616570503,0.211309135,-0.2638457716,0.001660227,-0.0189123191,0.398670435,-0.0074116085,0.0302929301,0.3964280188,0.099882625,0.1514957845,-0.1389228553,0.3128338754,-0.4387754202,0.2237685323,-0.0319213159,-0.3537499309,0.2378725559,0.0444737375,-0.0815992504,0.0902851298,-0.0688426122,0.2779449522,-0.0511246175,-0.1465737671,-0.095301941,-0.184546411,-0.2540001869,-0.0510652103,0.3700125515,-0.0547664836,-0.317294836,-0.2180195451,-0.0503768437,-0.1873558611,0.2057102621,0.3892886043,-0.1953092515,0.2749291956,-0.1282560378,0.0419043265,0.6039460897,0.063419573,-0.1607677937,0.4452090263,-0.0958242044,-0.3545050323,-0.0886768773,0.2540343702,0.1982444972,-0.0106193749,-0.3353327513,0.7256408334,-0.1832430065,-0.2423432767,-0.0794680268,-0.5672937632,0.294388175,0.1072142869,0.2334671617,0.1547670364,-0.1318540126,-0.16444543,0.228008762,0.0481451899,-0.0218945872,0.0195631459,-0.0771089494,-0.4434758723,-0.1948779076,-0.1264057606,-0.4273959398,0.2192077935,-0.6686814427,-0.2727111876,0.433701098,-0.0741638169,-0.0856534392,-0.2160948962,0.2276449651,-0.3765787184,-0.0897453427,-0.0897621512,0.232812196,0.1345527172,-0.3531804979,-0.0775736645,-0.0844805837,0.1200894713,-0.1485579014,-0.1885522455,0.194761619,0.1635546535,0.3136901557,-0.0682017878,-0.1422620267,0.0293508675,0.1174350977,0.1840603054,-0.0653511137,0.0679311156,0.1901082098,-0.0035382891,0.0262003094,-0.0636295378,-0.0851446092,0.0705611259,-0.2034726292,0.0601695739,-0.1732619405,0.4890032113,-0.2501839697,0.238762185,-0.0243577268,0.0712083802,0.1150856763,-0.1953156888,0.0422033593,-0.1012481377,0.262136519,0.079492934,0.0874665529,-0.0781058222,-0.1392308772,0.0596786924,0.3397408128,-0.1995896846,-0.129084155,-0.0971145108,-0.0654710755,-0.1062469259,0.2073575854,0.3562300205,0.2953766584,0.1393584013,0.5149038434,0.0771054104,0.14949736,-0.251996845,0.1438812613,0.0286945105,-0.203351289,-0.061133597,-0.0969168097,-0.1329973936,-0.2285000235,0.5298206806,0.1884369105,-0.0437838472,-0.0721282363,-0.1056225598,-0.4070443809,-0.136988923,-0.0033696876,-0.0215591509,-0.1591154933,-0.1997402012,0.3900647759,-0.1750443578,-0.2246314883,0.1717157811,0.1362401545,0.4625746608,-0.1568747312,-0.2329501212,0.1639208943,0.0432777442,-0.0902830958,0.0886400566,0.085990347,0.3979047537,0.4500082433,-0.100373134,-0.1956561357,0.0405306928,0.0862389207,0.0650635436,0.2300628126,0.4814370871,-0.0224951878,0.4113826454,-0.0127600012,0.1447871774,-0.3928005695,-0.1107226685,-0.2695804536,-0.1097872332,0.2999337912,0.3856029212,-0.2025088966,-0.7505491376,-0.2044207156,-0.2655576468,0.1911294907,0.0078091859,0.2178954482,-0.1302042305,-0.2237562239,-0.0723755732,-0.419121325,-0.0299893711,-0.0546580702,-0.0272561871,0.2050245255,-0.1796647459,-0.1089262292,0.1190538779,-0.0054975273,0.3167388141,0.2234031409,-0.658018887,0.3405663371,-0.0411544591,0.173910901,-0.0416291095,-0.1339777559,0.4145452976,-0.4879589677,0.034420561,0.2119375765,0.2159923613,0.0893984661,0.0918622985,0.153855443,-0.0929381475,0.32253474,0.0503838137,0.9196417928,0.0844696537,0.2862994671,-0.0774068832,0.0287648663,0.2585729957,-0.1073842943,0.3119176328,0.2791598439,0.197838068,-0.2146597654,0.2588627636,0.0241551753,-0.1635827571,-0.1606380194,-0.0481733643,-0.085032545,-0.1581908464,0.1432489008,-0.0431427322,0.137149781,0.2363631278,0.1082578972,0.2843376696,-0.1966169775,-0.0721735954,-0.0845047086,0.2468004972,0.0981274694,-0.5428899527,-0.0851744264,-0.2477727383,0.0057320818,-0.1068633944,0.1800314337,-0.1452999711,-0.0585731566,0.1785815656,0.1014825031,0.6295846701,0.0048300806,-0.1356538981,-0.0292936508,-0.3415820301,-0.1198016107,-0.0477270447,-0.0507908799,-0.2233602107,0.1544108093,-0.1649305969,-0.2447297275,0.0283706188,0.0734497085,0.4258039892,-0.2620765865,0.1635083705,0.0626244769,0.0756008551,-0.1320589483,0.3050738573,-0.0568813123,-0.0151061136,-0.2590179741,-0.1273944676,-0.2318470925,-0.1011641771,0.2034730166,0.1448333114,0.1429016441,0.0512661673,0.2887452245,0.0686542317,-0.2530813813,-0.3048666716,0.1492017359,0.0673601478,0.0474096052,-0.1193512902,0.0078325085,0.2520499229,0.3751187027,-0.1827657819,-0.2367277741,0.027421765,0.0550519712,-0.0090607526,0.4180619419,-0.0186042693,-0.3834147751,-0.1481364667,-0.1736949384,0.3704393506,0.2203922123,-0.2018370181,0.2406020463,0.0264361762,0.0278726481,-0.1864353716,0.2639498711,0.8607467413,-0.2654821575,0.2149347812,0.1646295935,-0.3632864058,0.8256191611,-0.334928453,0.1162393913,-0.1972406507,-0.1531572044,-0.1065631434,-0.2051377147,-0.1102228388,0.2792856395,0.0164199546,0.4865499139,0.117609784,-0.071860373,0.0547379814,0.0452391505,0.1444048584,0.0758988932,0.0172455627,-0.0416715406,-0.3121592104,0.1052947491,-0.0429833159,-0.0302201845,0.1688419729,0.0086281337,-0.0195344407,0.0852716863,-0.1231602058,-0.2328203171,-0.1784796119,-0.1600672305,-0.0935574025,0.634536922,0.150361076,0.2026652098,-0.3431531489,0.2239542007,-0.1167136058,0.0065891077,0.1394179165,0.1315451711,-0.1487892568,-0.2536192536,0.0774620697,0.1407611221,-0.1852730811,0.138182506,-0.3526147902,0.2445802689,0.1456682384,-0.1788820773,-0.1581056565,-0.0500044487,0.2716680467,-0.0119502749,0.0293890126,0.1175784618,-0.2636385858,0.0658612251,0.1705257446,-0.3241087496,-0.1111442298,0.1945013702,0.4288095236,-0.1700004637,0.332762599,-0.1409085393,0.0136464275,-0.306353569,-0.0172153953,0.0022659684,-0.3533315361,0.207145229,-0.1468132436,-0.1726896465,-0.0795624331,0.2791498005,-0.0207251422,-0.1307428926,0.0704071745,-0.1994920671,-0.0329446532,0.2163630426,-0.3628818095,0.3788318336,0.1354702264,-0.299767822,-0.1634667665,0.6286057234,-0.1036683396,0.032201048,-0.0944364592,0.251124382,-0.3191685975,0.251447171,0.0328676961,0.1078878716,-0.1501199007,-0.054022532,-0.3074723184,-0.073682569,-0.0204964653,0.1999012232,-0.0562040806,0.1895475388,0.3652932346,0.0124282306,-0.1579039097,0.0946147442,0.0829909071,-0.4526723623,0.2468918264,0.2111490369,0.5735405087,0.1258860081,-0.5383149385,0.4397482872,-0.178057462,0.4297509789,-0.0310197286,0.2363819927,0.071154438,-0.2082405537,-0.4042149484,-0.148828432,0.2778960764,-0.0559302494,-0.1306239814,0.1122292727,-0.0010715552,-0.3463710248,0.1391667277,0.7270395756,-0.1468434334,0.086114116,0.3228086233,0.0782287791,-0.0484977923,-0.0131457113,0.2004357725,-0.0356866159,-0.249909699,0.3352946043,0.0642417669,-0.2685917616,0.4162910879,-0.0460429974,-0.153701365,-0.1983644813,0.0161412656,0.5277784467,-0.2210380137,-0.2766153514,-0.1627861261,0.1033722758,0.0404639058,0.2024612576,-0.285554409,0.2845876515,0.4057112336,0.1913775206,0.0336736776,-0.1473397911,-0.3288441896,0.44915995,0.1161428615,0.0889782533,-0.02072387,0.2700832784,0.1414738446,-0.2981549501,-0.3012731373,0.2144737691,0.2034718096,-0.028561132,-0.2557725608,-0.0071108975,-0.1221089736,0.3441336751,-0.0058770124,-0.4594337642,-0.08456067,0.2982451022,-0.1821533591,-0.0693450496,0.3256308734,-0.0325854495,-0.128443867,-0.2529567182,0.1698726416,-0.1750003994,0.1404659301,0.4270246625,0.0154744759,-0.2307575792,-0.1206708699,-0.4186610579,-0.1668029428,0.3144170642,0.1680991203,-0.1549815238,0.1549421847,0.2139951736,0.54662925,0.404859066,-0.0147506353,0.0428885259,0.1710371524,0.0891628191,0.1824764758,-0.4049273133,0.4496495128,0.2269563824,0.207274735,-0.1210863516,-0.0541495942,-0.2880183458,0.0611567646,-0.3377894163,-0.2357199937,0.040467035,0.1949313432,-0.0027504237,-0.5021932721,0.503677845,0.0489068702,0.2614566088,-0.312384069,-0.1567742974,-0.2423513532,-0.0338422433,-0.0923614129,0.1479975432,0.0521389842,0.257334888,0.0968045592,0.5066362023,-0.4359943867,0.2668399811,-0.3112769425,0.2950918078,-0.2862005532,-0.1277657747,-0.3360529542,0.4263557792,0.0000956876,-0.2176023424,0.4132010937,-0.3228461444,-0.2708884478,0.6675411463,-0.1453432739,0.3749167621,0.2538483441,0.2959195971,0.2553870082,0.1831922084,0.0493685342,-0.1062306464,-0.0990056098,0.0672192723,-0.3648922741,-0.0909853801,-0.0493930243,0.1306518912,-0.2727119327,0.5207656026,0.0134379799,0.1782701164,-0.0694670305,0.3151212335,0.1213391796,-0.2395362556,0.0684812069,0.3522231579,0.4145163,0.337056607,-0.2478993982,-0.2640332282,-0.2773193121,0.0865659267,0.4922342896,-0.0528935865,0.0804802328,0.0453804694,0.2191266119,0.3004128933,-0.1477041543,-0.346588105,-0.0207600687,0.2700821161,-0.3474008739,0.0100170635,0.329218179,0.1928596646,-0.2000578642,-0.123465158,0.2232589573,0.2461208552,-0.2174123675,-0.2893902361,-0.3939068019]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/413","title":"Is there a way to download only NQ dev?","comments":"Ok, got it. I think this could be a valuable feature - especially for large datasets like NQ, but potentially also others. \r\nFor us, it will in this case make the difference of using the library or keeping the old downloads of the raw dev datasets. \r\nHowever, I don't know if that fits into your plans with the library and can also understand if you don't want to support this.","body":"Maybe I missed that in the docs, but is there a way to only download the dev set of natural questions (~1 GB)? \r\nAs we want to benchmark QA models on different datasets, I would like to avoid downloading the 41GB of training data. \r\n\r\nI tried\r\n```\r\ndataset = nlp.load_dataset('natural_questions', split=\"validation\", beam_runner=\"DirectRunner\")\r\n```\r\nBut this still triggered a big download of presumably the whole dataset. Is there any way of doing this or are splits \/ slicing options only available after downloading?\r\n\r\nThanks!","comment_length":70,"text":"Is there a way to download only NQ dev?\nMaybe I missed that in the docs, but is there a way to only download the dev set of natural questions (~1 GB)? \r\nAs we want to benchmark QA models on different datasets, I would like to avoid downloading the 41GB of training data. \r\n\r\nI tried\r\n```\r\ndataset = nlp.load_dataset('natural_questions', split=\"validation\", beam_runner=\"DirectRunner\")\r\n```\r\nBut this still triggered a big download of presumably the whole dataset. Is there any way of doing this or are splits \/ slicing options only available after downloading?\r\n\r\nThanks!\nOk, got it. I think this could be a valuable feature - especially for large datasets like NQ, but potentially also others. \r\nFor us, it will in this case make the difference of using the library or keeping the old downloads of the raw dev datasets. \r\nHowever, I don't know if that fits into your plans with the library and can also understand if you don't want to support this.","embeddings":[-0.3837682307,0.2108084261,-0.0951718166,0.1044244468,-0.0920639411,-0.0907708108,-0.2257674634,0.6605112553,-0.0418677889,0.2795556784,-0.3325553834,0.0646842942,-0.2607530355,0.5805532932,0.4200989902,-0.1612845212,-0.1143595204,0.4430704117,-0.0725371018,0.0320941284,-0.1458450556,0.2355462015,-0.0852555707,0.0845535174,0.3158827424,-0.2681672573,0.1613009423,0.0003769562,-0.4130051434,-0.2256505787,0.1046490893,0.2318772078,0.2737070918,-0.315141201,-0.0001247978,-0.1767543703,-0.0492420606,-0.1076974943,-0.1048340127,0.0290610269,-0.0316629075,-0.1345097423,0.144841224,-0.0146086644,-0.0661740303,-0.2787023783,0.1802223027,-0.0877976641,0.2828414738,0.3137025237,0.0347506143,0.0575815029,-0.178581506,0.1744213551,0.0454849042,-0.1572285444,-0.1496081054,0.0710525289,0.2409635633,0.173973918,0.1371459216,0.1628846824,0.1711564213,0.1461558193,0.1041327715,-0.0084158126,0.2902481854,-0.6779151559,-0.1346675307,0.7349892259,0.2752510607,0.1195652485,-0.1453244388,-0.1751475334,0.1211699694,-0.4858598709,-0.1015630513,0.502063036,-0.5056434274,0.0058689145,-0.5925880671,-0.5954379439,-0.2289485931,0.1964994371,-0.0154539598,-0.0786429271,0.0763812885,0.080804728,0.343953222,0.1932834536,0.1554152519,0.1268605143,-0.1199788153,0.1305152327,-0.2635022998,-0.4542864263,-0.3151841164,-0.107878089,0.2948121428,0.1053656712,-0.1306050718,0.0311698355,0.0303931069,-0.1063595265,0.5197607279,0.0478787497,0.245787248,0.2374639064,0.1265533864,0.0372288898,0.0975349173,0.1258990169,-0.0222979411,-0.0219954774,-0.6431497335,-0.0611066706,-0.05613897,-0.4703986049,0.102476716,-0.2125980556,-0.1848107725,-0.0741305798,-0.0765821636,0.2119109333,0.0375511758,0.0674566627,-0.2454402,0.2834025621,-0.2661134005,-0.7744125724,-0.1252468824,-0.0079068132,-0.0679525137,0.061968524,0.3375356495,0.0844027996,0.1490256041,-0.1850019246,0.0073810844,0.0263682678,0.3335590363,-0.0166183654,-0.0097264368,0.3612703979,0.0033703598,0.1638413966,-0.1632596552,0.2585628331,-0.4758756161,0.2490354329,-0.0942887217,-0.3748087287,0.0877634361,0.0242971443,-0.0430559218,0.055895485,-0.0828236714,0.2786794305,-0.0765085891,-0.2345643044,-0.0444730856,-0.1747214794,-0.223338455,-0.0668757558,0.285582304,-0.0780791566,-0.2522703409,-0.1871508658,-0.0746145099,-0.1785927415,0.2108701766,0.3680124879,-0.2529716194,0.1583964825,-0.1155179739,0.0218368247,0.6580260396,0.0790001824,-0.2337393165,0.4908336997,-0.0526507497,-0.3138030469,-0.1015424505,0.3563975692,0.1839399785,-0.0589393936,-0.3391218185,0.7834393978,-0.137980625,-0.2503324449,-0.0784127042,-0.5916569233,0.3074536622,0.1783097833,0.2205729932,0.1342015266,-0.049559094,-0.1202784181,0.2245157361,0.0300307367,0.0112632439,-0.0088108769,0.0123789692,-0.3818090558,-0.1711748689,-0.1530351341,-0.4459687769,0.1846539676,-0.6536505222,-0.1945620179,0.4189762473,-0.0865709931,0.0073981299,-0.2341000736,0.2203123569,-0.4121864736,-0.1198642105,-0.1790969968,0.1835347563,0.1684643179,-0.3527587056,-0.1195261106,-0.1414876729,0.1672486067,-0.181732744,-0.121665515,0.1836248934,0.2101252377,0.2229205817,-0.0013606852,-0.1546615213,0.060196545,0.1264700443,0.1464357674,-0.1569622457,0.0748496726,0.1380552202,0.0602238365,0.1468370855,-0.0621340722,-0.010638955,0.0278234184,-0.1594226509,0.0359759629,-0.1841621548,0.467012316,-0.2941001356,0.2273011208,-0.0432110429,0.0427761637,0.0677795038,-0.1987802386,0.0702115521,-0.0314566493,0.2993671,0.0751121044,0.0893668532,-0.0684214234,-0.1516072303,0.0916004479,0.3092695177,-0.1839173287,-0.1773821414,-0.0365724377,-0.0742196068,-0.0707327425,0.2380770743,0.4090535045,0.3239270449,0.1457184553,0.5053703785,0.0464160293,0.1550115496,-0.2461326569,0.1222533956,-0.0016359024,-0.2593738437,-0.0784983635,-0.1056861132,-0.0683617219,-0.2644273341,0.4836115837,0.1211462021,-0.0516694672,-0.010891648,-0.0918668211,-0.4039867222,-0.1954006702,-0.0059841196,-0.0668179095,-0.2175918669,-0.1385278851,0.3773465157,-0.1401188821,-0.1859869957,0.1419488192,0.2079991102,0.4611381292,-0.2285566479,-0.3333145082,0.1294851452,0.064887926,-0.0823554248,0.0876156241,0.1210082322,0.3232124448,0.4368119538,-0.0148643134,-0.1111296266,-0.0220286027,-0.031248292,0.0310642365,0.2105919272,0.4511539042,0.0031831211,0.3574376702,-0.0207836907,0.1039652675,-0.3731699288,-0.2048400789,-0.2044688314,-0.1332655698,0.3195647001,0.3701706827,-0.2776895761,-0.709852159,-0.2263504416,-0.2926571965,0.2092171013,0.0054044379,0.1874392331,-0.0258354172,-0.2257738262,-0.0712637082,-0.3463261724,-0.0302542821,-0.0538278595,-0.0661431774,0.2306489795,-0.1736421436,-0.077406019,0.1423440576,0.0093302578,0.2786389291,0.2440709174,-0.7312856317,0.3878257275,-0.0149218971,0.179317981,0.012404128,-0.0316703245,0.3906842172,-0.4956606627,0.0461903438,0.2226561159,0.1801547706,0.0473610051,0.1463140994,0.2642141581,-0.0505623966,0.2984301746,0.0042140135,0.9026047587,-0.021901818,0.2770243287,-0.0279032495,0.1159623042,0.3132763207,-0.1179407462,0.2958511412,0.3645724654,0.159880802,-0.2335767299,0.2932390869,0.1239419281,-0.2589546442,-0.2144537866,-0.0696256012,-0.0567028932,-0.1273163855,0.1541204304,-0.0097610662,0.0421919152,0.2184979022,0.0720128566,0.3293571174,-0.1914976984,-0.0961925834,0.0322565399,0.2623871267,0.0922153741,-0.554985106,-0.069148995,-0.2553257644,-0.0045989109,-0.0989544392,0.0317617543,-0.1204397082,-0.0082693407,0.2318597138,0.1683475524,0.5909474492,0.0614553466,-0.2426662743,-0.0883286372,-0.3295909464,-0.0227915905,-0.0294515602,0.0524412617,-0.1857880056,0.0847243741,-0.162515223,-0.277608037,-0.0038188952,0.1067299023,0.3743975163,-0.1717571765,0.0683405921,0.1319336891,0.0820216388,-0.1553023905,0.2775474489,-0.07118655,0.0161296129,-0.2766104639,-0.1321997792,-0.1908853203,-0.052037511,0.2010483593,0.0611883029,0.1420811266,0.0450431816,0.25174281,0.0456528179,-0.1850001216,-0.2596506774,0.2126801163,0.0181566756,0.051940009,-0.0969629213,0.0326504782,0.2234618664,0.385471195,-0.159027487,-0.1818449795,-0.0089871082,0.0372518077,-0.0170681868,0.3170141578,-0.0053068311,-0.3997896314,-0.2267181724,-0.2269950062,0.4494838715,0.1634717733,-0.1413418204,0.3220737576,0.065571256,0.0499741584,-0.2421557158,0.268582046,0.9881724119,-0.2815471292,0.254835546,0.1280252039,-0.4641484916,0.9363817573,-0.3721264601,0.093471311,-0.2408130318,-0.12615031,-0.1237549186,-0.185152024,-0.0779531524,0.2559374273,0.030219363,0.44264853,0.1611507237,-0.0513372645,0.0561254546,0.0560338162,0.1520567238,0.0439574867,-0.0281068143,-0.0443492569,-0.274613291,0.1111410409,-0.0541203879,-0.1022004858,0.1529758573,-0.0006277625,-0.0187220685,0.0444305874,-0.012726387,-0.2662277818,-0.0623796657,-0.2185972929,0.0143777356,0.6153013706,0.1269195229,0.1762401611,-0.3800661862,0.2230011225,-0.1217524856,0.0153685128,0.1590144187,0.1044135168,-0.140165329,-0.2097868323,0.1296275407,0.0871399119,-0.2226068228,0.0513727516,-0.3873911202,0.2717068493,0.2484145015,-0.1653220356,-0.1095157191,-0.0347577482,0.2807543278,-0.0071190558,0.0276636388,0.170744881,-0.2003898919,0.0062110387,0.1212964058,-0.2966047823,-0.1784132719,0.2170600146,0.441069454,-0.2956876457,0.2311336845,-0.1447519362,-0.0406469367,-0.2397845387,0.0104832025,-0.0299254749,-0.4008364379,0.1358234733,-0.110670127,-0.1830233932,-0.0245854557,0.2750994861,-0.0994584039,-0.094178766,0.0773350447,-0.2116999477,-0.0717723295,0.2115344107,-0.4958394468,0.3922190666,0.1611790508,-0.3440656662,-0.0800407231,0.5380374193,-0.0875144377,0.1143823713,-0.0327161774,0.2392184436,-0.3301357925,0.2376050502,-0.0335436985,0.0713960305,-0.1780042797,-0.0215448234,-0.318970561,-0.044727914,-0.083052896,0.220854193,-0.0200914387,0.1732337475,0.2979392409,-0.0526754521,-0.1839698851,0.1012202799,0.103920199,-0.3889257908,0.2401883006,0.1574955434,0.5783489347,0.1097845957,-0.5913115144,0.5145001411,-0.167704761,0.4666441977,0.0472618006,0.1512610018,-0.0127697298,-0.2139864713,-0.4592818022,-0.1276166886,0.2890667319,-0.0764441192,-0.0014027942,0.1141810715,0.0653450862,-0.278064549,0.2255567312,0.7534227967,-0.1339771748,0.0620856248,0.3673748374,0.0649304464,-0.0093423938,-0.0571749136,0.0890804827,-0.0138782188,-0.2894073129,0.3129281402,0.0575322583,-0.3034372628,0.4116078317,-0.1116233543,-0.0943765193,-0.1530000269,0.0425379798,0.5735353231,-0.2732243836,-0.3262380064,-0.1394720227,0.1508860588,0.0221971851,0.2271665186,-0.3691071868,0.2177572101,0.4459453225,0.2370329797,0.0058743469,-0.1344820559,-0.3252525926,0.5056000352,0.2182758749,0.0406728089,-0.0696755871,0.3011369109,0.1572867036,-0.2912782431,-0.1980879456,0.2419241667,0.1974517703,0.0166680403,-0.2028537542,-0.0496811494,-0.1826837659,0.3123771846,-0.1010722443,-0.4574673176,-0.1391304135,0.3240175545,-0.1236159727,-0.1569908857,0.3307210803,-0.0419630669,-0.1021295637,-0.255685389,0.1795627177,-0.1770020574,0.0958305895,0.4640857279,-0.0459361374,-0.2596224844,-0.0257901698,-0.3653516769,-0.2095395178,0.2498323768,0.2034852207,-0.1096158326,0.1557316035,0.2807765305,0.5663048625,0.3942402899,-0.0472349897,0.03246985,0.2881226242,0.0807332322,0.2034658939,-0.4621665478,0.4283205867,0.2240886241,0.214914307,-0.0970672667,-0.0631548315,-0.2014005929,-0.0163571816,-0.374848634,-0.2838506103,0.0123094842,0.1459267586,-0.0112008201,-0.477258116,0.5015835762,0.1012591049,0.2848774493,-0.3360569477,-0.1893147677,-0.2721227705,-0.0261168536,-0.0678438395,0.2251155078,0.0602589771,0.2284604907,0.1323888451,0.5546506047,-0.4826958477,0.2411868721,-0.2426615357,0.2789440453,-0.2555556893,-0.1002761424,-0.3028171062,0.4092620909,-0.0069009308,-0.1917091459,0.3745424151,-0.3308826983,-0.2955681384,0.6019467115,-0.1075994074,0.3654634356,0.2640299499,0.3025884032,0.2152695954,0.1133119836,0.1121458709,-0.1622615755,-0.0897890478,0.0291143991,-0.3479005694,-0.1157476157,-0.0441379659,0.1511312574,-0.2918075621,0.5623233318,-0.0307909828,0.1898498833,-0.0388919003,0.3846497238,0.030290056,-0.2880816758,0.0488305129,0.3543291986,0.3404435217,0.35595873,-0.2089968324,-0.2918730378,-0.2775607705,0.104759939,0.4886984527,-0.0419333056,0.1118400916,0.145421356,0.1829407066,0.3227471113,-0.1110480055,-0.4189252555,-0.028390618,0.2660476863,-0.3359093964,0.0608075969,0.3284502625,0.1726679802,-0.2116775066,-0.1320112348,0.1879139096,0.2639768124,-0.1589772403,-0.1981959492,-0.4046045244]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/413","title":"Is there a way to download only NQ dev?","comments":"I don't think we could force this behavior generally since the dataset script authors are free to organize the file download as they want (sometimes the mapping between split and files can be very much nontrivial) but we can add an additional configuration for Natural Question indeed as @lhoestq indicate.","body":"Maybe I missed that in the docs, but is there a way to only download the dev set of natural questions (~1 GB)? \r\nAs we want to benchmark QA models on different datasets, I would like to avoid downloading the 41GB of training data. \r\n\r\nI tried\r\n```\r\ndataset = nlp.load_dataset('natural_questions', split=\"validation\", beam_runner=\"DirectRunner\")\r\n```\r\nBut this still triggered a big download of presumably the whole dataset. Is there any way of doing this or are splits \/ slicing options only available after downloading?\r\n\r\nThanks!","comment_length":50,"text":"Is there a way to download only NQ dev?\nMaybe I missed that in the docs, but is there a way to only download the dev set of natural questions (~1 GB)? \r\nAs we want to benchmark QA models on different datasets, I would like to avoid downloading the 41GB of training data. \r\n\r\nI tried\r\n```\r\ndataset = nlp.load_dataset('natural_questions', split=\"validation\", beam_runner=\"DirectRunner\")\r\n```\r\nBut this still triggered a big download of presumably the whole dataset. Is there any way of doing this or are splits \/ slicing options only available after downloading?\r\n\r\nThanks!\nI don't think we could force this behavior generally since the dataset script authors are free to organize the file download as they want (sometimes the mapping between split and files can be very much nontrivial) but we can add an additional configuration for Natural Question indeed as @lhoestq indicate.","embeddings":[-0.4887266457,0.2052318454,-0.0939359516,0.1260942966,-0.0418884084,-0.1087556854,-0.2285951823,0.6617497206,0.0076493579,0.2669901252,-0.2714673281,0.0445381142,-0.2082810551,0.5945587158,0.4444682896,-0.0646661669,-0.127142325,0.4709189236,-0.1361080408,-0.0069543622,-0.17444098,0.1564280093,-0.0427492335,0.0399172641,0.2861556411,-0.2435347885,0.1814636737,0.0041279709,-0.3567369282,-0.2835049629,0.0866329819,0.275703162,0.216697678,-0.3432943523,-0.0001268621,-0.1054657102,0.0004297808,-0.1328522861,-0.2147211879,0.0927825794,-0.0724057853,-0.1583822072,0.1609330475,0.0109025743,-0.0843144357,-0.3348846138,0.2662496567,-0.1632023156,0.377425164,0.4012565017,-0.0041164611,0.0548433065,-0.2272785604,0.1745741814,0.0175522603,-0.1025494188,-0.1552434415,0.0558356084,0.190091759,0.0763694495,0.0393826738,0.0963236168,0.2445637882,0.2181500047,0.1171687841,-0.0001179587,0.3439489305,-0.7146770358,-0.1467894167,0.7247014046,0.2575412095,0.0835120454,-0.122038886,-0.2418496609,0.1380847991,-0.4262854755,-0.0425271839,0.4883491695,-0.6017991304,0.0279397331,-0.5949159861,-0.5436315536,-0.1921218932,0.2189568728,-0.0272634011,-0.0606630482,0.0457835793,0.0554199666,0.2680762708,0.1433386952,0.1559547931,0.0705497861,-0.0610100403,0.1489268839,-0.1999875754,-0.4450471997,-0.2982934415,-0.0994832441,0.3308860064,0.1128259301,-0.0295024868,0.0462544113,0.0011221379,-0.087605603,0.496910125,0.0745897293,0.2353423387,0.2745423615,0.1146288663,0.1112658232,0.0898417383,0.1389095038,0.0243525449,-0.0688010678,-0.6210039258,0.0036772368,0.0770883262,-0.3594807088,-0.0306123234,-0.218272835,-0.0773649141,-0.1010038108,-0.1445732713,0.1883690059,0.0624618568,0.0796580985,-0.3489792049,0.3237820268,-0.2918675244,-0.7594263554,-0.141937539,-0.0309817586,-0.0322257578,0.130548954,0.4035463929,0.1777429879,0.1623430699,-0.1546227336,-0.0081777824,0.0228776559,0.3501376808,-0.0447858796,0.0521627255,0.3954898417,0.0708256289,0.1198439822,-0.0991976485,0.2762240171,-0.4767386317,0.2121436447,-0.0814641565,-0.303979218,0.2000146061,0.0312736481,-0.0611732416,0.0619083792,-0.0824498087,0.295879513,-0.0218027122,-0.261738658,-0.0780139342,-0.1849835515,-0.2660306096,-0.0194636844,0.281535089,0.0521525703,-0.2876964211,-0.1943801641,-0.0426557697,-0.195380345,0.2409430295,0.3353285789,-0.2221540213,0.2208238691,-0.1777199805,0.0727050379,0.6215856075,0.0408351049,-0.2153276354,0.5189719796,-0.125259012,-0.2580466866,0.0099432487,0.2915630341,0.2015536427,-0.0975813791,-0.3081535697,0.7577164173,-0.138188526,-0.2085577399,-0.0468732975,-0.5798709989,0.2178352922,0.1311499923,0.1677621305,0.1886194348,-0.0218464416,-0.177037105,0.230263114,0.0760411546,0.0210764278,-0.0027037351,-0.0927264541,-0.4067582488,-0.2314436287,-0.1624530256,-0.5866490602,0.1941296905,-0.5578173399,-0.2835443318,0.3848071396,-0.1371848881,-0.0203147884,-0.2599238455,0.1540474147,-0.3446823955,-0.1026777774,-0.0458075665,0.1526650041,0.1315482408,-0.3660292029,-0.0622352958,-0.0604103506,0.2207046449,-0.1313726455,-0.1769902259,0.2260778099,0.1880297214,0.2558657229,-0.0725524202,-0.1099506393,0.046778962,0.1513633579,0.1734099388,-0.0262590759,0.0936715454,0.208229512,-0.0001844334,0.0592573173,-0.019603692,-0.0299162604,0.0918958932,-0.2208607346,0.0049648001,-0.2566196918,0.4621081352,-0.2703951299,0.2382683903,-0.0310361367,0.0220022183,0.1374265552,-0.2340873629,0.0234775841,-0.1472925842,0.2413588911,0.1238056868,0.1553203613,-0.0022999132,-0.1333712041,0.0855032951,0.414788723,-0.2111367136,-0.1127043888,-0.0137233594,-0.055762101,-0.098478049,0.1594362855,0.4292439818,0.3710593879,0.1735153049,0.5662353039,0.048326917,0.1430452913,-0.3168494105,0.1112156808,-0.0117881577,-0.2532935441,0.0380241685,-0.1188154966,-0.1213262826,-0.2932133079,0.4736105204,0.1529059112,-0.0682681575,-0.0972216949,-0.1410442442,-0.4016961157,-0.15072909,-0.071524255,0.0242202077,-0.1566215456,-0.2005376518,0.3691398203,-0.1295188069,-0.2382731438,0.1708692461,0.1290085763,0.4911543131,-0.2071625888,-0.321387887,0.1032970324,0.087634027,-0.0775730237,0.0606639124,0.1547174305,0.3283340931,0.4007332921,-0.1120109856,-0.1641483605,0.0141161876,0.1065445766,0.0587569587,0.2443142682,0.4849342704,-0.0601027235,0.4556983113,0.0552116148,0.1288237125,-0.4062317014,-0.0473387539,-0.2428159565,-0.1148092747,0.2974045575,0.3979501128,-0.2779551744,-0.7279382944,-0.1925116926,-0.3107815087,0.19715482,0.0362021253,0.1660703123,-0.020988876,-0.1801543087,-0.1689496636,-0.4117167294,0.030272264,-0.0424679704,-0.0558419898,0.2495708019,-0.1656604111,-0.0818884,0.098591134,-0.0002208454,0.1828828752,0.3064815402,-0.7142486572,0.3078974783,-0.0837082341,0.1705634296,-0.0621972457,-0.0692710057,0.4284794033,-0.4704176784,0.0272308886,0.2099205852,0.2742372751,0.1153624803,0.0321403928,0.134550482,-0.0440304019,0.309697926,0.0385462567,0.8778749108,0.0845527649,0.2853832543,0.0053808554,0.0823159739,0.3701732159,-0.1114830747,0.3447764814,0.3482345343,0.1952168494,-0.1944239885,0.3263472915,0.077482529,-0.1878171414,-0.1526004672,-0.0101807564,-0.0675125197,-0.1857351065,0.1985675991,-0.0588744432,0.0681098327,0.2273648232,0.1045218259,0.2960306704,-0.176845476,-0.1366890222,0.0278593674,0.2456605732,0.1218137443,-0.5731310248,-0.0082164509,-0.2568287551,-0.0231647473,-0.1184411496,0.0914421305,-0.2068186849,-0.0856493339,0.2234115303,0.2078452855,0.6067394018,0.0300897118,-0.1100326478,-0.1166502014,-0.2653922141,-0.0457315557,-0.0447135754,0.0103014084,-0.1716824174,0.1064627767,-0.047043331,-0.288446933,-0.0024439422,0.0786771476,0.3771257699,-0.1641007513,0.0801431239,0.1451834589,0.0664271489,-0.1432053,0.2743828297,0.0222200081,-0.0582528785,-0.2841442525,-0.1746897846,-0.2178646326,-0.0764283761,0.186862573,0.0768213347,0.0833543167,0.0766530707,0.225919053,0.078620851,-0.2055819035,-0.2046727538,0.2767520547,-0.0017925592,-0.0984973013,-0.1395348012,-0.0281245131,0.153605938,0.4546102285,-0.1922066063,-0.2828111053,-0.0167307854,0.0893433392,-0.0223377757,0.4292801321,0.0398144051,-0.4462184906,-0.2101210356,-0.258949399,0.373242259,0.194597736,-0.2055404931,0.2144986242,0.0013383226,-0.0282608122,-0.2472882867,0.2591445744,0.9118409157,-0.2326833308,0.2671704888,0.1472861916,-0.3779021502,0.8221436739,-0.4244488478,0.0780253261,-0.1748118103,-0.1513875723,-0.1384528875,-0.2045548409,-0.1035988405,0.3686384857,0.0392800346,0.4770531654,0.0891523734,-0.0144779319,0.0720808432,0.0688485056,0.0940318033,0.0331120566,0.0610829964,-0.0643871874,-0.3221387267,0.1072013602,-0.0213538241,-0.074822247,0.1851852983,-0.0554802492,-0.0290843379,0.0574598238,-0.157543093,-0.2827277184,-0.2014380693,-0.238813132,-0.0415569097,0.7085371017,0.1310724765,0.1403218806,-0.3597532213,0.2176436633,-0.1332637966,0.0653968826,0.0704182312,0.0525597259,-0.1054477543,-0.1683650315,0.140669331,0.012032086,-0.1634980887,0.137887761,-0.3886717558,0.2521441579,0.1531344801,-0.219817251,-0.1275345683,-0.1228999496,0.3295395672,-0.0046904478,0.0095537854,0.0961323604,-0.1693182588,0.0130121987,0.1482700855,-0.3236745,-0.1380087584,0.2522522211,0.3861665726,-0.2282228917,0.3095296621,-0.1906152219,-0.0504054986,-0.2935269177,0.032229472,-0.0071937405,-0.3725096583,0.1485250145,-0.0461407714,-0.1984892488,-0.0560622364,0.2099430412,-0.077358529,-0.1213219538,0.0820888281,-0.1997180581,-0.0244314261,0.1956809312,-0.4095665514,0.4032945037,0.1959793866,-0.3230592608,-0.1145842969,0.5872229934,-0.085320957,0.0332796425,0.0169557109,0.2204615474,-0.2670638561,0.2823041975,0.0293559916,0.1078803837,-0.1735433042,-0.0970687792,-0.2927987576,-0.0551036559,-0.0235795956,0.2111917734,-0.0421671718,0.1556611061,0.2824379802,-0.0781599283,-0.1579772085,0.0044939416,0.1034578234,-0.4241890609,0.244860217,0.1957980096,0.596870482,0.2148340344,-0.4498147368,0.4909949899,-0.2002983391,0.4467799962,-0.0122619178,0.1721559763,0.0115224617,-0.1901782453,-0.4182386398,-0.104304567,0.398624897,0.0369059667,-0.1261127144,0.0897479728,0.0416334458,-0.3117083609,0.1528309733,0.720867455,-0.1670783758,0.0277369097,0.3030209839,0.0733283982,-0.011628475,-0.0867304429,0.2075522095,-0.0328283496,-0.217563957,0.3294558525,-0.0068045771,-0.2957662046,0.4166152179,-0.0636664778,-0.0969546363,-0.1166885719,0.0263307597,0.5689668059,-0.2224676311,-0.2802537084,-0.1525290161,0.1486296505,0.0431913361,0.3451730311,-0.3122249842,0.2479107678,0.4005114138,0.2087184936,-0.0269348323,-0.2441418618,-0.2748731375,0.5230000615,0.1104312837,0.0336665474,-0.0003106605,0.2438363433,0.1401624829,-0.2416859418,-0.3140130341,0.1384702325,0.1926989853,0.0025465912,-0.2052243799,-0.0280448869,-0.0956045017,0.3212982416,-0.1187282428,-0.4251153767,-0.1308763772,0.302572161,-0.1611563712,-0.0812013596,0.3739449382,-0.0712851211,-0.1348882169,-0.2602090836,0.1898048073,-0.1848386973,0.1720391661,0.4114941955,-0.0462577716,-0.2266837806,-0.0455702953,-0.362421453,-0.2169677466,0.2848589718,0.1728212386,-0.2372649759,0.1584549099,0.2072387338,0.5742630363,0.4248839617,-0.0192973483,0.0464588068,0.2571292222,0.1105091795,0.1029170901,-0.4459004998,0.4196874499,0.2157385945,0.2236372381,-0.0775015205,-0.0779944584,-0.261613071,-0.0133577716,-0.3012205064,-0.2348341048,0.0605699793,0.1134694517,-0.0178120304,-0.4746333957,0.4637052119,0.0473077744,0.1785967946,-0.3177027404,-0.152558431,-0.2793531716,-0.0210688375,-0.1326093078,0.1412640363,-0.0114480397,0.2095387429,0.124026306,0.5167966485,-0.4997382164,0.2142004222,-0.2274483591,0.3636538982,-0.2333955169,-0.1672560424,-0.3186056912,0.4481709003,0.0079243165,-0.2858369946,0.3640494943,-0.2809878588,-0.2817651331,0.6151070595,-0.0763221085,0.3583169281,0.2011262774,0.2634149194,0.2481324971,0.1929303408,0.0672078431,-0.1127575561,-0.0985040218,0.1025803164,-0.4266868234,-0.1189944968,-0.0869919509,0.1052410379,-0.3675815165,0.5479363799,0.0155672645,0.1099902242,-0.0931463763,0.4020737112,0.0751655772,-0.2500087321,0.0830965489,0.3318130672,0.3681016266,0.333971709,-0.1986511052,-0.3193082809,-0.2046660036,0.030167019,0.4539538622,0.0250947941,0.0854401067,0.069846943,0.2511938214,0.2723048627,-0.0705248713,-0.2877126038,0.0154354572,0.2493832707,-0.3803718686,0.0441295132,0.3894332051,0.1632055044,-0.2141721249,-0.1160927117,0.2265221775,0.3127984107,-0.2167773992,-0.2701512873,-0.4153765738]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/412","title":"Unable to load XTREME dataset from disk","comments":"Hi @lewtun, you have to provide the full path to the downloaded file for example `\/home\/lewtum\/..`","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nFollowing the [docs](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https:\/\/github.com\/google-research\/xtreme) benchmark.\r\n\r\nI have manually downloaded the `AmazonPhotos.zip` file from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1) and am running into a `FileNotFoundError` when I point to the location of the dataset.\r\n\r\nAs far as I can tell, the problem is that `AmazonPhotos.zip` decompresses to `panx_dataset` and `load_dataset()` is not looking in the correct path:\r\n\r\n```\r\n# path where load_dataset is looking for fr.tar.gz\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/\r\n# path where it actually exists\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/panx_dataset\/\r\n```\r\n\r\n## Steps to reproduce the problem\r\n\r\n1. Manually download the XTREME benchmark from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1)\r\n\r\n2. Run the following code snippet\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is in the root of the folder\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n```\r\n\r\n3. Here is the stack trace\r\n```\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n<ipython-input-4-26786bb5fa93> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 464 split_dict = SplitDict(dataset_name=self.name)\r\n 465 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 466 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 467 # Checksums verification\r\n 468 if verify_infos:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/datasets\/xtreme\/b8c2ed3583a7a7ac60b503576dfed3271ac86757628897e945bd329c43b8a746\/xtreme.py in _split_generators(self, dl_manager)\r\n 725 panx_dl_dir = dl_manager.extract(panx_path)\r\n 726 lang = self.config.name.split(\".\")[1]\r\n--> 727 lang_folder = dl_manager.extract(os.path.join(panx_dl_dir, lang + \".tar.gz\"))\r\n 728 return [\r\n 729 nlp.SplitGenerator(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in extract(self, path_or_paths)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 170 return tuple(mapped)\r\n 171 # Singleton\r\n--> 172 return function(data_struct)\r\n 173 \r\n 174 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in <lambda>(path)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 203 elif urlparse(url_or_filename).scheme == \"\":\r\n 204 # File, but it doesn't exist.\r\n--> 205 raise FileNotFoundError(\"Local file {} doesn't exist\".format(url_or_filename))\r\n 206 else:\r\n 207 # Something unknown\r\n\r\nFileNotFoundError: Local file \/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/fr.tar.gz doesn't exist\r\n```\r\n\r\n## OS and hardware\r\n```\r\n- `nlp` version: 0.3.0\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: <fill in>\r\n- Using distributed or parallel set-up in script?: <fill in>\r\n```","comment_length":16,"text":"Unable to load XTREME dataset from disk\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nFollowing the [docs](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https:\/\/github.com\/google-research\/xtreme) benchmark.\r\n\r\nI have manually downloaded the `AmazonPhotos.zip` file from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1) and am running into a `FileNotFoundError` when I point to the location of the dataset.\r\n\r\nAs far as I can tell, the problem is that `AmazonPhotos.zip` decompresses to `panx_dataset` and `load_dataset()` is not looking in the correct path:\r\n\r\n```\r\n# path where load_dataset is looking for fr.tar.gz\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/\r\n# path where it actually exists\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/panx_dataset\/\r\n```\r\n\r\n## Steps to reproduce the problem\r\n\r\n1. Manually download the XTREME benchmark from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1)\r\n\r\n2. Run the following code snippet\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is in the root of the folder\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n```\r\n\r\n3. Here is the stack trace\r\n```\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n<ipython-input-4-26786bb5fa93> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 464 split_dict = SplitDict(dataset_name=self.name)\r\n 465 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 466 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 467 # Checksums verification\r\n 468 if verify_infos:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/datasets\/xtreme\/b8c2ed3583a7a7ac60b503576dfed3271ac86757628897e945bd329c43b8a746\/xtreme.py in _split_generators(self, dl_manager)\r\n 725 panx_dl_dir = dl_manager.extract(panx_path)\r\n 726 lang = self.config.name.split(\".\")[1]\r\n--> 727 lang_folder = dl_manager.extract(os.path.join(panx_dl_dir, lang + \".tar.gz\"))\r\n 728 return [\r\n 729 nlp.SplitGenerator(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in extract(self, path_or_paths)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 170 return tuple(mapped)\r\n 171 # Singleton\r\n--> 172 return function(data_struct)\r\n 173 \r\n 174 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in <lambda>(path)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 203 elif urlparse(url_or_filename).scheme == \"\":\r\n 204 # File, but it doesn't exist.\r\n--> 205 raise FileNotFoundError(\"Local file {} doesn't exist\".format(url_or_filename))\r\n 206 else:\r\n 207 # Something unknown\r\n\r\nFileNotFoundError: Local file \/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/fr.tar.gz doesn't exist\r\n```\r\n\r\n## OS and hardware\r\n```\r\n- `nlp` version: 0.3.0\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: <fill in>\r\n- Using distributed or parallel set-up in script?: <fill in>\r\n```\nHi @lewtun, you have to provide the full path to the downloaded file for example `\/home\/lewtum\/..`","embeddings":[-0.2940948308,-0.3835189939,-0.0021737483,0.6607134342,0.1483453512,0.123879388,-0.1730880439,0.1085548103,0.1807117611,0.0440969653,-0.1427524537,-0.000344655,0.1383353919,-0.0291913021,0.1631223559,-0.1389353275,-0.0050913743,-0.109264195,-0.0455735289,0.0021922421,-0.3334750533,0.3826159835,-0.1232326701,-0.0158359203,-0.3343698978,0.0152691752,0.0525251888,0.6103966832,-0.2204438448,-0.0490750596,0.4548045695,-0.1380422413,0.4193027318,0.7166676521,-0.0001206809,0.036272157,0.1372581273,-0.1804513633,-0.1188215613,-0.525824964,-0.3079940677,-0.1304575354,-0.0578953996,-0.127268672,-0.072370559,0.0013968141,0.1307007074,-0.5407083035,0.4081201255,0.1079233736,0.1058065891,0.2785436809,0.0526131652,-0.0774010569,0.2205448151,0.1630187631,-0.3261559606,0.0098360153,0.3490776718,0.1568444073,0.0890085846,0.3036373258,0.0028810215,0.0416494124,0.4682084322,-0.0255379956,0.1632110029,-0.2858410776,0.0740050152,0.3027319312,0.5413934588,-0.1766107529,-0.4551265836,-0.1600851715,-0.0877793655,-0.2420943677,0.3571464717,0.28176108,-0.1429187506,-0.0044436865,-0.2909394205,-0.1955347657,-0.165093556,0.161051318,-0.0667801872,0.1623710394,-0.216212377,0.0020341689,0.3741413057,0.0450923182,0.2139801234,-0.0631333888,-0.1090405881,0.1156973541,-0.2318337858,0.237139225,-0.1404143274,-0.0313966423,0.237890929,0.197227791,0.02315435,0.1833641529,-0.0200086348,0.1590581387,0.2242522091,0.2844460011,0.2408591509,-0.0752442777,0.0127866007,0.1807836294,0.0062315604,-0.1993156075,-0.1559436321,-0.2016490996,-0.1380642653,-0.0938202664,0.117636323,-0.3426390588,-0.5313278437,-0.0701550394,-0.0944774523,0.0107492469,0.2356742918,0.2389828861,0.0993805081,0.0076575205,0.0823181123,0.3665469289,-0.318562001,-0.0045816679,-0.2031152993,0.2044391781,-0.3886571229,0.2901443839,0.1603095978,-0.0814564601,0.3730145693,-0.1143650338,0.3260141015,-0.3670960069,0.0424612015,-0.0647062659,-0.0382351913,0.0822852254,0.2362206876,0.1614243388,0.1086064428,-0.0067187431,-0.1187783182,0.1436628252,-0.4055755734,-0.1271426678,-0.0795442015,0.0333358757,-0.4040099978,0.01824327,-0.6510798931,-0.1683329791,-0.270290643,0.033224877,-0.0252170842,0.0418415554,-0.1289660037,-0.0762089863,0.2826340199,0.487043649,-0.1808573157,-0.035213463,-0.2090551108,-0.1099215448,0.0813191235,0.1574320048,-0.2371750027,0.0877061337,-0.4471827447,0.3432358503,0.2603805363,-0.6826903224,-0.2903496921,0.3264546096,-0.2940928042,0.2417879999,0.0807776526,0.0621065237,-0.2141914964,-0.0856752247,0.1738442183,0.1770916879,0.1379901171,-0.0998015851,-0.2131122798,-0.1681564748,-0.19047831,0.3321940601,-0.013252819,-0.089894034,0.1114539504,-0.0186509453,0.1093090326,-0.1374859065,-0.1267775446,0.5887778997,0.074264124,-0.0401530415,-0.188909471,0.0607483536,-0.6912902594,0.3157786131,-0.2630388141,-0.2366461307,-0.0897766799,0.1072436571,-0.3324464858,-0.090750128,-0.3421750963,-0.1773630977,-0.0320939831,0.1423979104,-0.0039518424,-0.1507627666,-0.237391457,0.3739419878,0.0338048264,0.1004236937,-0.5817263126,0.266469717,0.1891721934,-0.0366892628,-0.0080807889,0.0295071863,0.244940415,-0.301872015,0.0110715451,0.5150936246,0.0157895256,0.3034413755,0.3414364457,0.0678832307,0.2183351815,-0.4746906459,0.1706380397,0.1261678338,-0.0502253696,-0.0008387527,-0.3532588184,0.4718583226,-0.1958650202,-0.0360133201,-0.0729345977,-0.1564880461,0.1472139955,-0.1102163047,-0.1181403995,-0.102371864,0.3790089786,0.1971785873,0.2383924276,-0.1636102349,-0.1151506752,-0.1355713755,-0.078775309,0.0191710368,0.0012407303,0.0319226608,-0.2628386021,0.1123144925,-0.0349363983,0.4033972025,0.4151169658,0.1524639279,0.0233504567,-0.0273290258,-0.0414358824,-0.0764893144,0.0318663605,0.0110485563,0.1094328389,0.1649197638,-0.3268106282,-0.0543202497,-0.3668630719,-0.1084010601,0.1083225384,0.0949982926,-0.3544782996,0.0258565135,-0.1741150916,-0.3895713091,-0.1025009453,0.1597319096,-0.2119823694,-0.2974932194,-0.1224297583,0.5138393641,0.2439617366,-0.033237014,-0.3228708804,0.0551576391,-0.1608925462,-0.5349494219,-0.2296060771,0.2254982889,-0.0298479758,-0.0379258394,0.3940607607,0.1775508374,0.06346903,-0.3751940429,0.4378796816,-0.5723701119,-0.1013892964,0.1536201239,-0.1776455343,0.4700734317,-0.1234334558,0.1739168316,-0.4125457108,0.0737455785,0.2622705102,-0.1435983479,-0.0725756884,-0.0108244531,0.0582668334,-0.1717712879,0.1029991955,-0.1550430059,-0.0834874436,-0.2285580337,-0.0605197586,0.0269689281,0.0039468836,0.1430416107,0.0121737709,0.0526414923,0.0021120338,-0.2169024199,-0.2150437683,-0.9191765785,0.482129097,-0.1535041332,-0.2848579884,0.371545285,0.1730707735,0.2463556677,0.2914629281,-0.4455307424,-0.1748443544,0.1003422588,0.1766921878,0.079654336,0.0285435226,0.2198371738,-0.2039557248,0.0781378672,-0.1340860575,-0.3067451119,-0.0713055357,0.1857059449,0.6754620075,-0.0910091251,0.0155380322,0.0305088181,0.48123914,0.0822562873,0.2168375403,0.4140934944,-0.105378665,0.344689697,-0.084782213,-0.1683692187,-0.2016595155,-0.1253787726,0.0345995463,0.0388450995,0.3197815716,0.1089996696,-0.2826604843,0.2524104714,-0.3590715826,-0.2499759197,0.0791105777,0.1507035643,0.1439125091,-0.0271688625,0.2944025099,-0.1495239437,-0.1423307806,0.0472172946,0.3464936018,0.4332347214,0.201624468,-0.3357114196,0.2821694314,-0.2268442959,0.4784918725,0.0736236945,0.171867758,-0.0929099917,0.0414229184,0.1955893338,-0.0877942741,0.3791092932,-0.0812442377,0.1548496783,0.0090517486,-0.373888284,-0.3268097043,0.023585448,0.0402890965,0.1563166231,-0.015382153,0.5480804443,-0.2565806508,-0.3150651753,-0.2097482979,0.1390483826,-0.0742261186,-0.134029299,-0.2735158205,-0.5143994093,-0.0718383864,-0.2955914736,0.1968692243,0.2673868239,0.3172526658,0.2283362299,-0.104235813,-0.0428409837,0.3793025911,0.0042143934,0.6529353261,-0.1743396372,0.0354927815,0.4252307117,-0.1178241447,0.2531483471,0.6561619043,0.1108206064,-0.1297160238,0.2663336098,-0.0517165177,-0.1295134425,0.3273470402,0.020796543,-0.3362400234,-0.0526033379,-0.0299214311,-0.1372974515,0.0104177892,0.3201183081,-0.0367610715,-0.1670839339,-0.0699462742,0.5326874852,0.2755137384,0.1686859876,0.3096197546,0.0979354084,-0.1516371071,-0.2330201864,0.1875227839,1.036370635,-0.1817591041,0.134875834,0.236576587,0.1435467601,-0.029546693,-0.2570065558,0.3308091462,-0.259376049,-0.3965887427,-0.2539665997,-0.1413376182,0.2486439943,0.0921125934,0.318685174,0.2940897942,0.1441646814,0.0109183667,0.0423499309,-0.0192194879,-0.2817929685,-0.3110000491,-0.5007690191,0.0626073852,-0.0410945155,0.5804379582,-0.1612380892,-0.1279451996,-0.1890367568,-0.0302863419,-0.0370579548,0.3009278476,-0.2347045839,0.1349930614,-0.0442406163,-0.4934231341,-0.1455714852,0.4456167817,0.0857500285,-0.2637138665,-0.2606283426,0.3505084515,-0.0731190369,-0.1928687692,0.0179756023,-0.1091079339,-0.1411272585,-0.1512780339,0.1349853575,-0.0841426849,-0.1423194855,-0.3568322659,-0.2701237798,-0.0669967681,-0.2799673975,-0.0454852395,-0.2646587491,0.0523822755,0.0430450402,-0.0202295631,0.043968454,0.1505742818,-0.118280299,0.0264785066,0.1194932088,-0.4309362769,-0.0279015079,0.5984717011,-0.3182874918,-0.2164822072,0.2697497904,0.1841867119,0.0473578349,-0.1390714347,0.3956476748,0.3742910922,-0.5853356123,0.0172140505,-0.1310363412,0.2062303424,0.0936489403,0.286962539,0.4050015807,-0.4055064619,0.2294516265,-0.4906133711,-0.0221327357,0.3766581118,-0.0306952801,0.2828480601,-0.0257449728,-0.2261558175,0.1318860054,-0.3666865528,-0.2243918777,0.3964022696,-0.0309916735,-0.0513414443,0.2765412629,0.1092196479,0.5288570523,0.0609976947,0.0577792339,0.2188753486,-0.26735273,-0.0422088914,0.0702691451,0.1751425117,0.0311533306,-0.1865858436,0.0188074168,-0.1455456465,-0.086516045,-0.0247351229,0.1848336607,0.3872470856,0.0307885446,-0.1474805921,-0.1098290011,0.1663480699,-0.4709260762,0.10619317,0.1898027956,0.4924295247,0.0212684292,-0.0236135721,-0.095186986,-0.0887796804,-0.3238078654,-0.0843955502,0.2894161344,0.2882059216,0.3363096118,-0.3605225086,0.3339119256,0.1849844456,0.2267098576,0.3662360311,-0.2820647657,0.1108194366,0.0315920524,0.1250391454,-0.0269203577,-0.1127346456,0.1083496362,-0.0816022754,-0.1020234451,0.3889040649,0.1993317008,0.2852793336,0.2068143487,-0.0222183745,0.700296402,-0.0479071997,0.2296517789,0.4015395939,-0.0840390474,0.1403926313,0.2540922761,0.3396313488,0.1165546924,0.4072642624,-0.2131436914,0.509999156,-0.0704315901,0.4083868861,-0.3891384006,-0.4278742671,-0.2259721011,0.2799412906,0.1763959825,0.1806118041,-0.2744500637,0.2971872985,-0.4949175119,-0.014748767,-0.4205287099,0.1546151489,-0.0668893605,0.018383773,-0.2635987997,-0.3721785545,-0.148756355,0.0487965196,-0.1414866298,-0.1530177146,-0.0398862623,0.0794364959,-0.0456199683,0.0826277435,-0.1365236938,0.2178891748,-0.0494195446,-0.1700529605,0.4847475886,0.07236249,-0.0143226581,0.2167026997,0.2018682808,0.1844723523,0.0698768646,-0.1004493013,-0.1165881082,0.2017828524,-0.04673535,0.1467270255,0.0984191075,-0.0198986046,0.0405208319,0.2966165543,0.161962077,-0.0573879555,0.197358489,-0.0739957243,0.3093760312,-0.0460786484,0.0277054235,-0.4094930887,0.1580156386,-0.151088804,-0.2176139355,-0.1051064804,-0.1655072719,0.0924878642,-0.1326874495,0.0062669455,0.2456581444,0.0179532673,0.0025493947,0.2727975547,0.4918943346,-0.2644323409,-0.5227890611,-0.2552607656,-0.7857858539,0.1571288705,-0.0327609889,-0.2347466797,0.0456146449,0.1628720462,-0.0656049103,0.2150097638,0.3023923934,-0.0805267319,0.333007127,0.302087605,-0.2846804261,-0.0318464749,0.2175108343,0.0961918607,-0.0552845821,-0.2977227867,-0.0097985798,0.1965633631,-0.072098203,-0.0004320726,0.0005193675,-0.3038447499,-0.2595980763,0.3466371596,-0.1283624768,0.2355115414,0.0069287908,-0.3553111553,-0.2619484365,-0.2585630119,0.1081844568,0.081758298,0.0971930698,0.4771674871,-0.0876284242,0.1216642633,-0.5112359524,0.2954388261,0.1425307244,-0.0016410062,-0.3504886031,0.100261718,-0.2921203673,0.0513204746,0.2885354459,0.3758064508,0.0382824279,0.3379191458,-0.141376093,-0.1874237657,0.471930474,0.0057807025,-0.1403841078,0.2902044058,0.0600744113,0.2342756391,-0.143301025,-0.342662096,0.2560198307,0.0822849944,0.0243478362,-0.0055715484,-0.0157550145,-0.2069303691,-0.1993103772,0.1498200297,0.2356415242,0.1080833972,-0.1697506309,0.1547189057,0.048286207]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/412","title":"Unable to load XTREME dataset from disk","comments":"I was able to repro. Opening a PR to fix that.\r\nThanks for reporting this issue !","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nFollowing the [docs](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https:\/\/github.com\/google-research\/xtreme) benchmark.\r\n\r\nI have manually downloaded the `AmazonPhotos.zip` file from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1) and am running into a `FileNotFoundError` when I point to the location of the dataset.\r\n\r\nAs far as I can tell, the problem is that `AmazonPhotos.zip` decompresses to `panx_dataset` and `load_dataset()` is not looking in the correct path:\r\n\r\n```\r\n# path where load_dataset is looking for fr.tar.gz\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/\r\n# path where it actually exists\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/panx_dataset\/\r\n```\r\n\r\n## Steps to reproduce the problem\r\n\r\n1. Manually download the XTREME benchmark from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1)\r\n\r\n2. Run the following code snippet\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is in the root of the folder\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n```\r\n\r\n3. Here is the stack trace\r\n```\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n<ipython-input-4-26786bb5fa93> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 464 split_dict = SplitDict(dataset_name=self.name)\r\n 465 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 466 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 467 # Checksums verification\r\n 468 if verify_infos:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/datasets\/xtreme\/b8c2ed3583a7a7ac60b503576dfed3271ac86757628897e945bd329c43b8a746\/xtreme.py in _split_generators(self, dl_manager)\r\n 725 panx_dl_dir = dl_manager.extract(panx_path)\r\n 726 lang = self.config.name.split(\".\")[1]\r\n--> 727 lang_folder = dl_manager.extract(os.path.join(panx_dl_dir, lang + \".tar.gz\"))\r\n 728 return [\r\n 729 nlp.SplitGenerator(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in extract(self, path_or_paths)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 170 return tuple(mapped)\r\n 171 # Singleton\r\n--> 172 return function(data_struct)\r\n 173 \r\n 174 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in <lambda>(path)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 203 elif urlparse(url_or_filename).scheme == \"\":\r\n 204 # File, but it doesn't exist.\r\n--> 205 raise FileNotFoundError(\"Local file {} doesn't exist\".format(url_or_filename))\r\n 206 else:\r\n 207 # Something unknown\r\n\r\nFileNotFoundError: Local file \/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/fr.tar.gz doesn't exist\r\n```\r\n\r\n## OS and hardware\r\n```\r\n- `nlp` version: 0.3.0\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: <fill in>\r\n- Using distributed or parallel set-up in script?: <fill in>\r\n```","comment_length":17,"text":"Unable to load XTREME dataset from disk\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nFollowing the [docs](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https:\/\/github.com\/google-research\/xtreme) benchmark.\r\n\r\nI have manually downloaded the `AmazonPhotos.zip` file from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1) and am running into a `FileNotFoundError` when I point to the location of the dataset.\r\n\r\nAs far as I can tell, the problem is that `AmazonPhotos.zip` decompresses to `panx_dataset` and `load_dataset()` is not looking in the correct path:\r\n\r\n```\r\n# path where load_dataset is looking for fr.tar.gz\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/\r\n# path where it actually exists\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/panx_dataset\/\r\n```\r\n\r\n## Steps to reproduce the problem\r\n\r\n1. Manually download the XTREME benchmark from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1)\r\n\r\n2. Run the following code snippet\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is in the root of the folder\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n```\r\n\r\n3. Here is the stack trace\r\n```\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n<ipython-input-4-26786bb5fa93> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 464 split_dict = SplitDict(dataset_name=self.name)\r\n 465 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 466 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 467 # Checksums verification\r\n 468 if verify_infos:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/datasets\/xtreme\/b8c2ed3583a7a7ac60b503576dfed3271ac86757628897e945bd329c43b8a746\/xtreme.py in _split_generators(self, dl_manager)\r\n 725 panx_dl_dir = dl_manager.extract(panx_path)\r\n 726 lang = self.config.name.split(\".\")[1]\r\n--> 727 lang_folder = dl_manager.extract(os.path.join(panx_dl_dir, lang + \".tar.gz\"))\r\n 728 return [\r\n 729 nlp.SplitGenerator(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in extract(self, path_or_paths)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 170 return tuple(mapped)\r\n 171 # Singleton\r\n--> 172 return function(data_struct)\r\n 173 \r\n 174 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in <lambda>(path)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 203 elif urlparse(url_or_filename).scheme == \"\":\r\n 204 # File, but it doesn't exist.\r\n--> 205 raise FileNotFoundError(\"Local file {} doesn't exist\".format(url_or_filename))\r\n 206 else:\r\n 207 # Something unknown\r\n\r\nFileNotFoundError: Local file \/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/fr.tar.gz doesn't exist\r\n```\r\n\r\n## OS and hardware\r\n```\r\n- `nlp` version: 0.3.0\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: <fill in>\r\n- Using distributed or parallel set-up in script?: <fill in>\r\n```\nI was able to repro. Opening a PR to fix that.\r\nThanks for reporting this issue !","embeddings":[-0.2940948308,-0.3835189939,-0.0021737483,0.6607134342,0.1483453512,0.123879388,-0.1730880439,0.1085548103,0.1807117611,0.0440969653,-0.1427524537,-0.000344655,0.1383353919,-0.0291913021,0.1631223559,-0.1389353275,-0.0050913743,-0.109264195,-0.0455735289,0.0021922421,-0.3334750533,0.3826159835,-0.1232326701,-0.0158359203,-0.3343698978,0.0152691752,0.0525251888,0.6103966832,-0.2204438448,-0.0490750596,0.4548045695,-0.1380422413,0.4193027318,0.7166676521,-0.0001206809,0.036272157,0.1372581273,-0.1804513633,-0.1188215613,-0.525824964,-0.3079940677,-0.1304575354,-0.0578953996,-0.127268672,-0.072370559,0.0013968141,0.1307007074,-0.5407083035,0.4081201255,0.1079233736,0.1058065891,0.2785436809,0.0526131652,-0.0774010569,0.2205448151,0.1630187631,-0.3261559606,0.0098360153,0.3490776718,0.1568444073,0.0890085846,0.3036373258,0.0028810215,0.0416494124,0.4682084322,-0.0255379956,0.1632110029,-0.2858410776,0.0740050152,0.3027319312,0.5413934588,-0.1766107529,-0.4551265836,-0.1600851715,-0.0877793655,-0.2420943677,0.3571464717,0.28176108,-0.1429187506,-0.0044436865,-0.2909394205,-0.1955347657,-0.165093556,0.161051318,-0.0667801872,0.1623710394,-0.216212377,0.0020341689,0.3741413057,0.0450923182,0.2139801234,-0.0631333888,-0.1090405881,0.1156973541,-0.2318337858,0.237139225,-0.1404143274,-0.0313966423,0.237890929,0.197227791,0.02315435,0.1833641529,-0.0200086348,0.1590581387,0.2242522091,0.2844460011,0.2408591509,-0.0752442777,0.0127866007,0.1807836294,0.0062315604,-0.1993156075,-0.1559436321,-0.2016490996,-0.1380642653,-0.0938202664,0.117636323,-0.3426390588,-0.5313278437,-0.0701550394,-0.0944774523,0.0107492469,0.2356742918,0.2389828861,0.0993805081,0.0076575205,0.0823181123,0.3665469289,-0.318562001,-0.0045816679,-0.2031152993,0.2044391781,-0.3886571229,0.2901443839,0.1603095978,-0.0814564601,0.3730145693,-0.1143650338,0.3260141015,-0.3670960069,0.0424612015,-0.0647062659,-0.0382351913,0.0822852254,0.2362206876,0.1614243388,0.1086064428,-0.0067187431,-0.1187783182,0.1436628252,-0.4055755734,-0.1271426678,-0.0795442015,0.0333358757,-0.4040099978,0.01824327,-0.6510798931,-0.1683329791,-0.270290643,0.033224877,-0.0252170842,0.0418415554,-0.1289660037,-0.0762089863,0.2826340199,0.487043649,-0.1808573157,-0.035213463,-0.2090551108,-0.1099215448,0.0813191235,0.1574320048,-0.2371750027,0.0877061337,-0.4471827447,0.3432358503,0.2603805363,-0.6826903224,-0.2903496921,0.3264546096,-0.2940928042,0.2417879999,0.0807776526,0.0621065237,-0.2141914964,-0.0856752247,0.1738442183,0.1770916879,0.1379901171,-0.0998015851,-0.2131122798,-0.1681564748,-0.19047831,0.3321940601,-0.013252819,-0.089894034,0.1114539504,-0.0186509453,0.1093090326,-0.1374859065,-0.1267775446,0.5887778997,0.074264124,-0.0401530415,-0.188909471,0.0607483536,-0.6912902594,0.3157786131,-0.2630388141,-0.2366461307,-0.0897766799,0.1072436571,-0.3324464858,-0.090750128,-0.3421750963,-0.1773630977,-0.0320939831,0.1423979104,-0.0039518424,-0.1507627666,-0.237391457,0.3739419878,0.0338048264,0.1004236937,-0.5817263126,0.266469717,0.1891721934,-0.0366892628,-0.0080807889,0.0295071863,0.244940415,-0.301872015,0.0110715451,0.5150936246,0.0157895256,0.3034413755,0.3414364457,0.0678832307,0.2183351815,-0.4746906459,0.1706380397,0.1261678338,-0.0502253696,-0.0008387527,-0.3532588184,0.4718583226,-0.1958650202,-0.0360133201,-0.0729345977,-0.1564880461,0.1472139955,-0.1102163047,-0.1181403995,-0.102371864,0.3790089786,0.1971785873,0.2383924276,-0.1636102349,-0.1151506752,-0.1355713755,-0.078775309,0.0191710368,0.0012407303,0.0319226608,-0.2628386021,0.1123144925,-0.0349363983,0.4033972025,0.4151169658,0.1524639279,0.0233504567,-0.0273290258,-0.0414358824,-0.0764893144,0.0318663605,0.0110485563,0.1094328389,0.1649197638,-0.3268106282,-0.0543202497,-0.3668630719,-0.1084010601,0.1083225384,0.0949982926,-0.3544782996,0.0258565135,-0.1741150916,-0.3895713091,-0.1025009453,0.1597319096,-0.2119823694,-0.2974932194,-0.1224297583,0.5138393641,0.2439617366,-0.033237014,-0.3228708804,0.0551576391,-0.1608925462,-0.5349494219,-0.2296060771,0.2254982889,-0.0298479758,-0.0379258394,0.3940607607,0.1775508374,0.06346903,-0.3751940429,0.4378796816,-0.5723701119,-0.1013892964,0.1536201239,-0.1776455343,0.4700734317,-0.1234334558,0.1739168316,-0.4125457108,0.0737455785,0.2622705102,-0.1435983479,-0.0725756884,-0.0108244531,0.0582668334,-0.1717712879,0.1029991955,-0.1550430059,-0.0834874436,-0.2285580337,-0.0605197586,0.0269689281,0.0039468836,0.1430416107,0.0121737709,0.0526414923,0.0021120338,-0.2169024199,-0.2150437683,-0.9191765785,0.482129097,-0.1535041332,-0.2848579884,0.371545285,0.1730707735,0.2463556677,0.2914629281,-0.4455307424,-0.1748443544,0.1003422588,0.1766921878,0.079654336,0.0285435226,0.2198371738,-0.2039557248,0.0781378672,-0.1340860575,-0.3067451119,-0.0713055357,0.1857059449,0.6754620075,-0.0910091251,0.0155380322,0.0305088181,0.48123914,0.0822562873,0.2168375403,0.4140934944,-0.105378665,0.344689697,-0.084782213,-0.1683692187,-0.2016595155,-0.1253787726,0.0345995463,0.0388450995,0.3197815716,0.1089996696,-0.2826604843,0.2524104714,-0.3590715826,-0.2499759197,0.0791105777,0.1507035643,0.1439125091,-0.0271688625,0.2944025099,-0.1495239437,-0.1423307806,0.0472172946,0.3464936018,0.4332347214,0.201624468,-0.3357114196,0.2821694314,-0.2268442959,0.4784918725,0.0736236945,0.171867758,-0.0929099917,0.0414229184,0.1955893338,-0.0877942741,0.3791092932,-0.0812442377,0.1548496783,0.0090517486,-0.373888284,-0.3268097043,0.023585448,0.0402890965,0.1563166231,-0.015382153,0.5480804443,-0.2565806508,-0.3150651753,-0.2097482979,0.1390483826,-0.0742261186,-0.134029299,-0.2735158205,-0.5143994093,-0.0718383864,-0.2955914736,0.1968692243,0.2673868239,0.3172526658,0.2283362299,-0.104235813,-0.0428409837,0.3793025911,0.0042143934,0.6529353261,-0.1743396372,0.0354927815,0.4252307117,-0.1178241447,0.2531483471,0.6561619043,0.1108206064,-0.1297160238,0.2663336098,-0.0517165177,-0.1295134425,0.3273470402,0.020796543,-0.3362400234,-0.0526033379,-0.0299214311,-0.1372974515,0.0104177892,0.3201183081,-0.0367610715,-0.1670839339,-0.0699462742,0.5326874852,0.2755137384,0.1686859876,0.3096197546,0.0979354084,-0.1516371071,-0.2330201864,0.1875227839,1.036370635,-0.1817591041,0.134875834,0.236576587,0.1435467601,-0.029546693,-0.2570065558,0.3308091462,-0.259376049,-0.3965887427,-0.2539665997,-0.1413376182,0.2486439943,0.0921125934,0.318685174,0.2940897942,0.1441646814,0.0109183667,0.0423499309,-0.0192194879,-0.2817929685,-0.3110000491,-0.5007690191,0.0626073852,-0.0410945155,0.5804379582,-0.1612380892,-0.1279451996,-0.1890367568,-0.0302863419,-0.0370579548,0.3009278476,-0.2347045839,0.1349930614,-0.0442406163,-0.4934231341,-0.1455714852,0.4456167817,0.0857500285,-0.2637138665,-0.2606283426,0.3505084515,-0.0731190369,-0.1928687692,0.0179756023,-0.1091079339,-0.1411272585,-0.1512780339,0.1349853575,-0.0841426849,-0.1423194855,-0.3568322659,-0.2701237798,-0.0669967681,-0.2799673975,-0.0454852395,-0.2646587491,0.0523822755,0.0430450402,-0.0202295631,0.043968454,0.1505742818,-0.118280299,0.0264785066,0.1194932088,-0.4309362769,-0.0279015079,0.5984717011,-0.3182874918,-0.2164822072,0.2697497904,0.1841867119,0.0473578349,-0.1390714347,0.3956476748,0.3742910922,-0.5853356123,0.0172140505,-0.1310363412,0.2062303424,0.0936489403,0.286962539,0.4050015807,-0.4055064619,0.2294516265,-0.4906133711,-0.0221327357,0.3766581118,-0.0306952801,0.2828480601,-0.0257449728,-0.2261558175,0.1318860054,-0.3666865528,-0.2243918777,0.3964022696,-0.0309916735,-0.0513414443,0.2765412629,0.1092196479,0.5288570523,0.0609976947,0.0577792339,0.2188753486,-0.26735273,-0.0422088914,0.0702691451,0.1751425117,0.0311533306,-0.1865858436,0.0188074168,-0.1455456465,-0.086516045,-0.0247351229,0.1848336607,0.3872470856,0.0307885446,-0.1474805921,-0.1098290011,0.1663480699,-0.4709260762,0.10619317,0.1898027956,0.4924295247,0.0212684292,-0.0236135721,-0.095186986,-0.0887796804,-0.3238078654,-0.0843955502,0.2894161344,0.2882059216,0.3363096118,-0.3605225086,0.3339119256,0.1849844456,0.2267098576,0.3662360311,-0.2820647657,0.1108194366,0.0315920524,0.1250391454,-0.0269203577,-0.1127346456,0.1083496362,-0.0816022754,-0.1020234451,0.3889040649,0.1993317008,0.2852793336,0.2068143487,-0.0222183745,0.700296402,-0.0479071997,0.2296517789,0.4015395939,-0.0840390474,0.1403926313,0.2540922761,0.3396313488,0.1165546924,0.4072642624,-0.2131436914,0.509999156,-0.0704315901,0.4083868861,-0.3891384006,-0.4278742671,-0.2259721011,0.2799412906,0.1763959825,0.1806118041,-0.2744500637,0.2971872985,-0.4949175119,-0.014748767,-0.4205287099,0.1546151489,-0.0668893605,0.018383773,-0.2635987997,-0.3721785545,-0.148756355,0.0487965196,-0.1414866298,-0.1530177146,-0.0398862623,0.0794364959,-0.0456199683,0.0826277435,-0.1365236938,0.2178891748,-0.0494195446,-0.1700529605,0.4847475886,0.07236249,-0.0143226581,0.2167026997,0.2018682808,0.1844723523,0.0698768646,-0.1004493013,-0.1165881082,0.2017828524,-0.04673535,0.1467270255,0.0984191075,-0.0198986046,0.0405208319,0.2966165543,0.161962077,-0.0573879555,0.197358489,-0.0739957243,0.3093760312,-0.0460786484,0.0277054235,-0.4094930887,0.1580156386,-0.151088804,-0.2176139355,-0.1051064804,-0.1655072719,0.0924878642,-0.1326874495,0.0062669455,0.2456581444,0.0179532673,0.0025493947,0.2727975547,0.4918943346,-0.2644323409,-0.5227890611,-0.2552607656,-0.7857858539,0.1571288705,-0.0327609889,-0.2347466797,0.0456146449,0.1628720462,-0.0656049103,0.2150097638,0.3023923934,-0.0805267319,0.333007127,0.302087605,-0.2846804261,-0.0318464749,0.2175108343,0.0961918607,-0.0552845821,-0.2977227867,-0.0097985798,0.1965633631,-0.072098203,-0.0004320726,0.0005193675,-0.3038447499,-0.2595980763,0.3466371596,-0.1283624768,0.2355115414,0.0069287908,-0.3553111553,-0.2619484365,-0.2585630119,0.1081844568,0.081758298,0.0971930698,0.4771674871,-0.0876284242,0.1216642633,-0.5112359524,0.2954388261,0.1425307244,-0.0016410062,-0.3504886031,0.100261718,-0.2921203673,0.0513204746,0.2885354459,0.3758064508,0.0382824279,0.3379191458,-0.141376093,-0.1874237657,0.471930474,0.0057807025,-0.1403841078,0.2902044058,0.0600744113,0.2342756391,-0.143301025,-0.342662096,0.2560198307,0.0822849944,0.0243478362,-0.0055715484,-0.0157550145,-0.2069303691,-0.1993103772,0.1498200297,0.2356415242,0.1080833972,-0.1697506309,0.1547189057,0.048286207]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/412","title":"Unable to load XTREME dataset from disk","comments":"Thanks for the rapid fix @lhoestq!","body":"Hi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nFollowing the [docs](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https:\/\/github.com\/google-research\/xtreme) benchmark.\r\n\r\nI have manually downloaded the `AmazonPhotos.zip` file from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1) and am running into a `FileNotFoundError` when I point to the location of the dataset.\r\n\r\nAs far as I can tell, the problem is that `AmazonPhotos.zip` decompresses to `panx_dataset` and `load_dataset()` is not looking in the correct path:\r\n\r\n```\r\n# path where load_dataset is looking for fr.tar.gz\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/\r\n# path where it actually exists\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/panx_dataset\/\r\n```\r\n\r\n## Steps to reproduce the problem\r\n\r\n1. Manually download the XTREME benchmark from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1)\r\n\r\n2. Run the following code snippet\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is in the root of the folder\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n```\r\n\r\n3. Here is the stack trace\r\n```\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n<ipython-input-4-26786bb5fa93> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 464 split_dict = SplitDict(dataset_name=self.name)\r\n 465 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 466 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 467 # Checksums verification\r\n 468 if verify_infos:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/datasets\/xtreme\/b8c2ed3583a7a7ac60b503576dfed3271ac86757628897e945bd329c43b8a746\/xtreme.py in _split_generators(self, dl_manager)\r\n 725 panx_dl_dir = dl_manager.extract(panx_path)\r\n 726 lang = self.config.name.split(\".\")[1]\r\n--> 727 lang_folder = dl_manager.extract(os.path.join(panx_dl_dir, lang + \".tar.gz\"))\r\n 728 return [\r\n 729 nlp.SplitGenerator(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in extract(self, path_or_paths)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 170 return tuple(mapped)\r\n 171 # Singleton\r\n--> 172 return function(data_struct)\r\n 173 \r\n 174 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in <lambda>(path)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 203 elif urlparse(url_or_filename).scheme == \"\":\r\n 204 # File, but it doesn't exist.\r\n--> 205 raise FileNotFoundError(\"Local file {} doesn't exist\".format(url_or_filename))\r\n 206 else:\r\n 207 # Something unknown\r\n\r\nFileNotFoundError: Local file \/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/fr.tar.gz doesn't exist\r\n```\r\n\r\n## OS and hardware\r\n```\r\n- `nlp` version: 0.3.0\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: <fill in>\r\n- Using distributed or parallel set-up in script?: <fill in>\r\n```","comment_length":6,"text":"Unable to load XTREME dataset from disk\nHi \ud83e\udd17 team!\r\n\r\n## Description of the problem\r\nFollowing the [docs](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https:\/\/github.com\/google-research\/xtreme) benchmark.\r\n\r\nI have manually downloaded the `AmazonPhotos.zip` file from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1) and am running into a `FileNotFoundError` when I point to the location of the dataset.\r\n\r\nAs far as I can tell, the problem is that `AmazonPhotos.zip` decompresses to `panx_dataset` and `load_dataset()` is not looking in the correct path:\r\n\r\n```\r\n# path where load_dataset is looking for fr.tar.gz\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/\r\n# path where it actually exists\r\n\/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/panx_dataset\/\r\n```\r\n\r\n## Steps to reproduce the problem\r\n\r\n1. Manually download the XTREME benchmark from [here](https:\/\/www.amazon.com\/clouddrive\/share\/d3KGCRCIYwhKJF0H3eWA26hjg2ZCRhjpEQtDL70FSBN?_encoding=UTF8&%2AVersion%2A=1&%2Aentries%2A=0&mgh=1)\r\n\r\n2. Run the following code snippet\r\n```python\r\nfrom nlp import load_dataset\r\n# AmazonPhotos.zip is in the root of the folder\r\ndataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n```\r\n\r\n3. Here is the stack trace\r\n```\r\n---------------------------------------------------------------------------\r\nFileNotFoundError Traceback (most recent call last)\r\n<ipython-input-4-26786bb5fa93> in <module>\r\n----> 1 dataset = load_dataset(\"xtreme\", \"PAN-X.fr\", data_dir='.\/')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 464 split_dict = SplitDict(dataset_name=self.name)\r\n 465 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 466 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 467 # Checksums verification\r\n 468 if verify_infos:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/datasets\/xtreme\/b8c2ed3583a7a7ac60b503576dfed3271ac86757628897e945bd329c43b8a746\/xtreme.py in _split_generators(self, dl_manager)\r\n 725 panx_dl_dir = dl_manager.extract(panx_path)\r\n 726 lang = self.config.name.split(\".\")[1]\r\n--> 727 lang_folder = dl_manager.extract(os.path.join(panx_dl_dir, lang + \".tar.gz\"))\r\n 728 return [\r\n 729 nlp.SplitGenerator(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in extract(self, path_or_paths)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 170 return tuple(mapped)\r\n 171 # Singleton\r\n--> 172 return function(data_struct)\r\n 173 \r\n 174 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/download_manager.py in <lambda>(path)\r\n 196 \"\"\"\r\n 197 return map_nested(\r\n--> 198 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 199 )\r\n 200 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 203 elif urlparse(url_or_filename).scheme == \"\":\r\n 204 # File, but it doesn't exist.\r\n--> 205 raise FileNotFoundError(\"Local file {} doesn't exist\".format(url_or_filename))\r\n 206 else:\r\n 207 # Something unknown\r\n\r\nFileNotFoundError: Local file \/root\/.cache\/huggingface\/datasets\/9b8c4f1578e45cb2539332c79738beb3b54afbcd842b079cabfd79e3ed6704f6\/fr.tar.gz doesn't exist\r\n```\r\n\r\n## OS and hardware\r\n```\r\n- `nlp` version: 0.3.0\r\n- Platform: Linux-4.15.0-72-generic-x86_64-with-Ubuntu-18.04-bionic\r\n- Python version: 3.6.9\r\n- PyTorch version (GPU?): 1.4.0 (True)\r\n- Tensorflow version (GPU?): 2.1.0 (True)\r\n- Using GPU in script?: <fill in>\r\n- Using distributed or parallel set-up in script?: <fill in>\r\n```\nThanks for the rapid fix @lhoestq!","embeddings":[-0.2940948308,-0.3835189939,-0.0021737483,0.6607134342,0.1483453512,0.123879388,-0.1730880439,0.1085548103,0.1807117611,0.0440969653,-0.1427524537,-0.000344655,0.1383353919,-0.0291913021,0.1631223559,-0.1389353275,-0.0050913743,-0.109264195,-0.0455735289,0.0021922421,-0.3334750533,0.3826159835,-0.1232326701,-0.0158359203,-0.3343698978,0.0152691752,0.0525251888,0.6103966832,-0.2204438448,-0.0490750596,0.4548045695,-0.1380422413,0.4193027318,0.7166676521,-0.0001206809,0.036272157,0.1372581273,-0.1804513633,-0.1188215613,-0.525824964,-0.3079940677,-0.1304575354,-0.0578953996,-0.127268672,-0.072370559,0.0013968141,0.1307007074,-0.5407083035,0.4081201255,0.1079233736,0.1058065891,0.2785436809,0.0526131652,-0.0774010569,0.2205448151,0.1630187631,-0.3261559606,0.0098360153,0.3490776718,0.1568444073,0.0890085846,0.3036373258,0.0028810215,0.0416494124,0.4682084322,-0.0255379956,0.1632110029,-0.2858410776,0.0740050152,0.3027319312,0.5413934588,-0.1766107529,-0.4551265836,-0.1600851715,-0.0877793655,-0.2420943677,0.3571464717,0.28176108,-0.1429187506,-0.0044436865,-0.2909394205,-0.1955347657,-0.165093556,0.161051318,-0.0667801872,0.1623710394,-0.216212377,0.0020341689,0.3741413057,0.0450923182,0.2139801234,-0.0631333888,-0.1090405881,0.1156973541,-0.2318337858,0.237139225,-0.1404143274,-0.0313966423,0.237890929,0.197227791,0.02315435,0.1833641529,-0.0200086348,0.1590581387,0.2242522091,0.2844460011,0.2408591509,-0.0752442777,0.0127866007,0.1807836294,0.0062315604,-0.1993156075,-0.1559436321,-0.2016490996,-0.1380642653,-0.0938202664,0.117636323,-0.3426390588,-0.5313278437,-0.0701550394,-0.0944774523,0.0107492469,0.2356742918,0.2389828861,0.0993805081,0.0076575205,0.0823181123,0.3665469289,-0.318562001,-0.0045816679,-0.2031152993,0.2044391781,-0.3886571229,0.2901443839,0.1603095978,-0.0814564601,0.3730145693,-0.1143650338,0.3260141015,-0.3670960069,0.0424612015,-0.0647062659,-0.0382351913,0.0822852254,0.2362206876,0.1614243388,0.1086064428,-0.0067187431,-0.1187783182,0.1436628252,-0.4055755734,-0.1271426678,-0.0795442015,0.0333358757,-0.4040099978,0.01824327,-0.6510798931,-0.1683329791,-0.270290643,0.033224877,-0.0252170842,0.0418415554,-0.1289660037,-0.0762089863,0.2826340199,0.487043649,-0.1808573157,-0.035213463,-0.2090551108,-0.1099215448,0.0813191235,0.1574320048,-0.2371750027,0.0877061337,-0.4471827447,0.3432358503,0.2603805363,-0.6826903224,-0.2903496921,0.3264546096,-0.2940928042,0.2417879999,0.0807776526,0.0621065237,-0.2141914964,-0.0856752247,0.1738442183,0.1770916879,0.1379901171,-0.0998015851,-0.2131122798,-0.1681564748,-0.19047831,0.3321940601,-0.013252819,-0.089894034,0.1114539504,-0.0186509453,0.1093090326,-0.1374859065,-0.1267775446,0.5887778997,0.074264124,-0.0401530415,-0.188909471,0.0607483536,-0.6912902594,0.3157786131,-0.2630388141,-0.2366461307,-0.0897766799,0.1072436571,-0.3324464858,-0.090750128,-0.3421750963,-0.1773630977,-0.0320939831,0.1423979104,-0.0039518424,-0.1507627666,-0.237391457,0.3739419878,0.0338048264,0.1004236937,-0.5817263126,0.266469717,0.1891721934,-0.0366892628,-0.0080807889,0.0295071863,0.244940415,-0.301872015,0.0110715451,0.5150936246,0.0157895256,0.3034413755,0.3414364457,0.0678832307,0.2183351815,-0.4746906459,0.1706380397,0.1261678338,-0.0502253696,-0.0008387527,-0.3532588184,0.4718583226,-0.1958650202,-0.0360133201,-0.0729345977,-0.1564880461,0.1472139955,-0.1102163047,-0.1181403995,-0.102371864,0.3790089786,0.1971785873,0.2383924276,-0.1636102349,-0.1151506752,-0.1355713755,-0.078775309,0.0191710368,0.0012407303,0.0319226608,-0.2628386021,0.1123144925,-0.0349363983,0.4033972025,0.4151169658,0.1524639279,0.0233504567,-0.0273290258,-0.0414358824,-0.0764893144,0.0318663605,0.0110485563,0.1094328389,0.1649197638,-0.3268106282,-0.0543202497,-0.3668630719,-0.1084010601,0.1083225384,0.0949982926,-0.3544782996,0.0258565135,-0.1741150916,-0.3895713091,-0.1025009453,0.1597319096,-0.2119823694,-0.2974932194,-0.1224297583,0.5138393641,0.2439617366,-0.033237014,-0.3228708804,0.0551576391,-0.1608925462,-0.5349494219,-0.2296060771,0.2254982889,-0.0298479758,-0.0379258394,0.3940607607,0.1775508374,0.06346903,-0.3751940429,0.4378796816,-0.5723701119,-0.1013892964,0.1536201239,-0.1776455343,0.4700734317,-0.1234334558,0.1739168316,-0.4125457108,0.0737455785,0.2622705102,-0.1435983479,-0.0725756884,-0.0108244531,0.0582668334,-0.1717712879,0.1029991955,-0.1550430059,-0.0834874436,-0.2285580337,-0.0605197586,0.0269689281,0.0039468836,0.1430416107,0.0121737709,0.0526414923,0.0021120338,-0.2169024199,-0.2150437683,-0.9191765785,0.482129097,-0.1535041332,-0.2848579884,0.371545285,0.1730707735,0.2463556677,0.2914629281,-0.4455307424,-0.1748443544,0.1003422588,0.1766921878,0.079654336,0.0285435226,0.2198371738,-0.2039557248,0.0781378672,-0.1340860575,-0.3067451119,-0.0713055357,0.1857059449,0.6754620075,-0.0910091251,0.0155380322,0.0305088181,0.48123914,0.0822562873,0.2168375403,0.4140934944,-0.105378665,0.344689697,-0.084782213,-0.1683692187,-0.2016595155,-0.1253787726,0.0345995463,0.0388450995,0.3197815716,0.1089996696,-0.2826604843,0.2524104714,-0.3590715826,-0.2499759197,0.0791105777,0.1507035643,0.1439125091,-0.0271688625,0.2944025099,-0.1495239437,-0.1423307806,0.0472172946,0.3464936018,0.4332347214,0.201624468,-0.3357114196,0.2821694314,-0.2268442959,0.4784918725,0.0736236945,0.171867758,-0.0929099917,0.0414229184,0.1955893338,-0.0877942741,0.3791092932,-0.0812442377,0.1548496783,0.0090517486,-0.373888284,-0.3268097043,0.023585448,0.0402890965,0.1563166231,-0.015382153,0.5480804443,-0.2565806508,-0.3150651753,-0.2097482979,0.1390483826,-0.0742261186,-0.134029299,-0.2735158205,-0.5143994093,-0.0718383864,-0.2955914736,0.1968692243,0.2673868239,0.3172526658,0.2283362299,-0.104235813,-0.0428409837,0.3793025911,0.0042143934,0.6529353261,-0.1743396372,0.0354927815,0.4252307117,-0.1178241447,0.2531483471,0.6561619043,0.1108206064,-0.1297160238,0.2663336098,-0.0517165177,-0.1295134425,0.3273470402,0.020796543,-0.3362400234,-0.0526033379,-0.0299214311,-0.1372974515,0.0104177892,0.3201183081,-0.0367610715,-0.1670839339,-0.0699462742,0.5326874852,0.2755137384,0.1686859876,0.3096197546,0.0979354084,-0.1516371071,-0.2330201864,0.1875227839,1.036370635,-0.1817591041,0.134875834,0.236576587,0.1435467601,-0.029546693,-0.2570065558,0.3308091462,-0.259376049,-0.3965887427,-0.2539665997,-0.1413376182,0.2486439943,0.0921125934,0.318685174,0.2940897942,0.1441646814,0.0109183667,0.0423499309,-0.0192194879,-0.2817929685,-0.3110000491,-0.5007690191,0.0626073852,-0.0410945155,0.5804379582,-0.1612380892,-0.1279451996,-0.1890367568,-0.0302863419,-0.0370579548,0.3009278476,-0.2347045839,0.1349930614,-0.0442406163,-0.4934231341,-0.1455714852,0.4456167817,0.0857500285,-0.2637138665,-0.2606283426,0.3505084515,-0.0731190369,-0.1928687692,0.0179756023,-0.1091079339,-0.1411272585,-0.1512780339,0.1349853575,-0.0841426849,-0.1423194855,-0.3568322659,-0.2701237798,-0.0669967681,-0.2799673975,-0.0454852395,-0.2646587491,0.0523822755,0.0430450402,-0.0202295631,0.043968454,0.1505742818,-0.118280299,0.0264785066,0.1194932088,-0.4309362769,-0.0279015079,0.5984717011,-0.3182874918,-0.2164822072,0.2697497904,0.1841867119,0.0473578349,-0.1390714347,0.3956476748,0.3742910922,-0.5853356123,0.0172140505,-0.1310363412,0.2062303424,0.0936489403,0.286962539,0.4050015807,-0.4055064619,0.2294516265,-0.4906133711,-0.0221327357,0.3766581118,-0.0306952801,0.2828480601,-0.0257449728,-0.2261558175,0.1318860054,-0.3666865528,-0.2243918777,0.3964022696,-0.0309916735,-0.0513414443,0.2765412629,0.1092196479,0.5288570523,0.0609976947,0.0577792339,0.2188753486,-0.26735273,-0.0422088914,0.0702691451,0.1751425117,0.0311533306,-0.1865858436,0.0188074168,-0.1455456465,-0.086516045,-0.0247351229,0.1848336607,0.3872470856,0.0307885446,-0.1474805921,-0.1098290011,0.1663480699,-0.4709260762,0.10619317,0.1898027956,0.4924295247,0.0212684292,-0.0236135721,-0.095186986,-0.0887796804,-0.3238078654,-0.0843955502,0.2894161344,0.2882059216,0.3363096118,-0.3605225086,0.3339119256,0.1849844456,0.2267098576,0.3662360311,-0.2820647657,0.1108194366,0.0315920524,0.1250391454,-0.0269203577,-0.1127346456,0.1083496362,-0.0816022754,-0.1020234451,0.3889040649,0.1993317008,0.2852793336,0.2068143487,-0.0222183745,0.700296402,-0.0479071997,0.2296517789,0.4015395939,-0.0840390474,0.1403926313,0.2540922761,0.3396313488,0.1165546924,0.4072642624,-0.2131436914,0.509999156,-0.0704315901,0.4083868861,-0.3891384006,-0.4278742671,-0.2259721011,0.2799412906,0.1763959825,0.1806118041,-0.2744500637,0.2971872985,-0.4949175119,-0.014748767,-0.4205287099,0.1546151489,-0.0668893605,0.018383773,-0.2635987997,-0.3721785545,-0.148756355,0.0487965196,-0.1414866298,-0.1530177146,-0.0398862623,0.0794364959,-0.0456199683,0.0826277435,-0.1365236938,0.2178891748,-0.0494195446,-0.1700529605,0.4847475886,0.07236249,-0.0143226581,0.2167026997,0.2018682808,0.1844723523,0.0698768646,-0.1004493013,-0.1165881082,0.2017828524,-0.04673535,0.1467270255,0.0984191075,-0.0198986046,0.0405208319,0.2966165543,0.161962077,-0.0573879555,0.197358489,-0.0739957243,0.3093760312,-0.0460786484,0.0277054235,-0.4094930887,0.1580156386,-0.151088804,-0.2176139355,-0.1051064804,-0.1655072719,0.0924878642,-0.1326874495,0.0062669455,0.2456581444,0.0179532673,0.0025493947,0.2727975547,0.4918943346,-0.2644323409,-0.5227890611,-0.2552607656,-0.7857858539,0.1571288705,-0.0327609889,-0.2347466797,0.0456146449,0.1628720462,-0.0656049103,0.2150097638,0.3023923934,-0.0805267319,0.333007127,0.302087605,-0.2846804261,-0.0318464749,0.2175108343,0.0961918607,-0.0552845821,-0.2977227867,-0.0097985798,0.1965633631,-0.072098203,-0.0004320726,0.0005193675,-0.3038447499,-0.2595980763,0.3466371596,-0.1283624768,0.2355115414,0.0069287908,-0.3553111553,-0.2619484365,-0.2585630119,0.1081844568,0.081758298,0.0971930698,0.4771674871,-0.0876284242,0.1216642633,-0.5112359524,0.2954388261,0.1425307244,-0.0016410062,-0.3504886031,0.100261718,-0.2921203673,0.0513204746,0.2885354459,0.3758064508,0.0382824279,0.3379191458,-0.141376093,-0.1874237657,0.471930474,0.0057807025,-0.1403841078,0.2902044058,0.0600744113,0.2342756391,-0.143301025,-0.342662096,0.2560198307,0.0822849944,0.0243478362,-0.0055715484,-0.0157550145,-0.2069303691,-0.1993103772,0.1498200297,0.2356415242,0.1080833972,-0.1697506309,0.1547189057,0.048286207]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/409","title":"train_test_split error: 'dict' object has no attribute 'deepcopy'","comments":"It was fixed in 2ddd18d139d3047c9c3abe96e1e7d05bb360132c.\r\nCould you pull the latest changes from master @morganmcg1 ?","body":"`train_test_split` is giving me an error when I try and call it:\r\n\r\n`'dict' object has no attribute 'deepcopy'`\r\n\r\n## To reproduce\r\n\r\n```\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.train_test_split(test_size=0.2)\r\n```\r\n\r\n## Full Stacktrace\r\n```\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-12-feb740dbec9a> in <module>\r\n 1 dataset = load_dataset('glue', 'mrpc', split='train')\r\n----> 2 dataset = dataset.train_test_split(test_size=0.2)\r\n\r\n~\/anaconda3\/envs\/fastai2_me\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in train_test_split(self, test_size, train_size, shuffle, seed, generator, keep_in_memory, load_from_cache_file, train_cache_file_name, test_cache_file_name, writer_batch_size)\r\n 1032 \"writer_batch_size\": writer_batch_size,\r\n 1033 }\r\n-> 1034 train_kwargs = cache_kwargs.deepcopy()\r\n 1035 train_kwargs[\"split\"] = \"train\"\r\n 1036 test_kwargs = cache_kwargs.deepcopy()\r\n\r\nAttributeError: 'dict' object has no attribute 'deepcopy'\r\n```","comment_length":15,"text":"train_test_split error: 'dict' object has no attribute 'deepcopy'\n`train_test_split` is giving me an error when I try and call it:\r\n\r\n`'dict' object has no attribute 'deepcopy'`\r\n\r\n## To reproduce\r\n\r\n```\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.train_test_split(test_size=0.2)\r\n```\r\n\r\n## Full Stacktrace\r\n```\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-12-feb740dbec9a> in <module>\r\n 1 dataset = load_dataset('glue', 'mrpc', split='train')\r\n----> 2 dataset = dataset.train_test_split(test_size=0.2)\r\n\r\n~\/anaconda3\/envs\/fastai2_me\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in train_test_split(self, test_size, train_size, shuffle, seed, generator, keep_in_memory, load_from_cache_file, train_cache_file_name, test_cache_file_name, writer_batch_size)\r\n 1032 \"writer_batch_size\": writer_batch_size,\r\n 1033 }\r\n-> 1034 train_kwargs = cache_kwargs.deepcopy()\r\n 1035 train_kwargs[\"split\"] = \"train\"\r\n 1036 test_kwargs = cache_kwargs.deepcopy()\r\n\r\nAttributeError: 'dict' object has no attribute 'deepcopy'\r\n```\nIt was fixed in 2ddd18d139d3047c9c3abe96e1e7d05bb360132c.\r\nCould you pull the latest changes from master @morganmcg1 ?","embeddings":[-0.2430309951,-0.0904732272,0.0003229392,0.0525522828,0.1288731694,0.007450358,0.3769094348,0.5188900232,0.0007268622,0.0506978817,0.0096392091,0.5001062751,-0.1641973257,0.1419215649,-0.0337450877,-0.3376310468,0.080365397,0.0954004005,-0.0612636507,0.0528249368,-0.252566874,0.208755374,-0.3073851764,0.0289700888,-0.3526663184,-0.0945937186,-0.217893824,0.0508309901,0.0251380503,-0.4331841469,0.4715957344,-0.1846660823,0.0829429626,0.4275281727,-0.0001179102,0.0744146705,0.5266975164,-0.025440326,-0.4324480891,-0.1489746571,-0.4580597579,-0.0282302089,0.3633123636,-0.4677169621,-0.0468950197,0.0502689332,-0.1839870512,-0.1044113562,0.6483406425,0.4795620441,0.1849884987,0.089347668,0.2606548071,-0.3767324984,0.2739802599,0.2085764557,-0.0843559727,0.235619083,-0.2282457352,-0.1526263356,0.1714741588,0.1839513928,-0.0872317255,0.1130343005,-0.1432219595,0.0785952359,0.059234459,-0.2913966179,-0.0330439061,0.243483156,-0.0170205161,-0.2833271027,-0.2864155769,-0.1533154547,0.0804037526,-0.6803515553,-0.1535033286,0.0325166546,-0.0882179439,0.0968551114,0.0330240317,-0.0685511827,-0.0503846668,0.3335687518,-0.1478834152,0.0223131198,-0.050156787,0.0933349952,0.3979634643,-0.0162373763,-0.0234931242,0.2073200196,-0.2703033388,0.0181810148,-0.1764831841,-0.3912528157,-0.0002549044,-0.3191053271,-0.1341014951,0.3878155649,0.0176855065,-0.064566955,0.2577806115,0.2188484222,0.0624829382,0.4496837556,0.0481811985,0.8710276484,0.199090004,-0.0053697377,-0.0558384359,-0.1215042248,-0.2192576528,-0.141293928,0.2043744475,0.1457729638,0.3434510529,-0.1309026033,-0.2778675854,-0.3583551645,-0.232035324,-0.0977196023,0.0650434569,0.342302084,0.0432159603,0.0757902488,-0.0884849355,0.3248477876,-0.1886294335,-0.1350573301,-0.228871733,0.3745124042,-0.1781211048,-0.0354130641,0.2919825912,0.2547863424,0.186687991,-0.1641830504,0.0306921676,-0.2236554176,-0.0524061397,-0.2777130902,0.1439236104,0.2463751733,-0.1293718815,-0.1671503037,0.3436351717,-0.0531790219,-0.1898308545,0.0276490208,-0.5408576727,-0.486041069,-0.1288615912,0.1614986807,-0.023589151,-0.2450150251,-0.1439890414,-0.1419112235,0.2486996055,-0.1216597408,0.1532095075,-0.3974026442,-0.1791483313,-0.2362658232,0.2453333884,0.39122051,-0.1007293835,-0.2666241229,-0.4364306331,0.192726329,0.3978490531,0.1276355386,-0.1570869535,0.2521066666,0.0090545835,0.0406650305,0.6577590108,-0.3584024012,-0.3280189931,0.1417558938,0.035572309,-0.1657350212,0.0221515633,-0.1195431128,0.4087431729,-0.0667686611,0.2881222963,0.2207947224,-0.0207881629,-0.1682307273,-0.1439072192,-0.1442129612,0.1819557399,0.1402908117,0.184629187,0.0644631311,-0.0677518323,0.0510462224,0.0985479876,0.1502958834,0.0739889443,-0.0268388055,0.4450594187,-0.0969080031,-0.0317307115,-0.0950663462,-0.0206925105,0.1291278899,0.0852244273,0.0106802545,-0.2283006608,0.0447362848,-0.4131640494,0.096860081,-0.4629735947,-0.3493302763,0.0987880155,-0.1426107734,0.2399077863,0.1296448112,-0.1714816242,0.2778931856,-0.1175075397,0.2327461094,-0.1257647425,0.3992980123,-0.0940904617,-0.3108988106,-0.1139127016,0.310736388,0.1771961898,-0.1352144629,-0.2276296914,0.3353103697,0.0962043926,-0.2103099525,-0.2375216931,0.0954912156,0.174211055,-0.4172563851,0.2095571756,0.1258116513,0.1424463093,-0.1108325124,-0.0999884158,0.5204695463,0.1227547675,0.3066320717,-0.0020527241,0.1651537269,-0.1521608979,-0.2325354218,0.0223215166,-0.2066972554,-0.0578454696,-0.1942804009,0.2980448604,-0.2669658065,0.038303718,-0.0685528964,0.5440787673,0.1653800458,0.168625474,0.2008247077,-0.157583043,0.1391693652,-0.1670501381,0.6118075252,0.4507924914,0.2039536089,-0.0245837905,0.027771879,-0.0653045326,0.1087196469,0.2719053924,-0.0632432029,0.5329877138,0.0212543756,0.1450446993,0.0463764817,-0.047746297,-0.3302198052,0.1369984895,0.2423622459,-0.1322706938,0.1333626956,-0.2110342383,-0.2954769731,-0.2166609019,-0.3252261579,0.1741653979,-0.293805927,-0.1491171569,-0.0915659443,-0.0014658833,0.1959457397,-0.3049390912,-0.0276186839,0.156764105,-0.3659560084,0.0486784577,-0.3093336523,-0.1551052034,0.0091024982,0.2254818082,-0.2070954889,0.1018692404,0.1044479609,-0.2352760136,0.2676257789,-0.1643413156,0.0974268243,-0.2017759681,-0.0806771889,0.3033491373,0.0051997085,0.1187690049,-0.2822883725,0.3416102231,-0.1665514112,-0.081181258,0.1451903582,0.1337814927,-0.0885310024,-0.1705868393,-0.3728453517,-0.3125472069,-0.224580735,0.3296000659,-0.0662402436,0.1503891945,0.1305451095,0.1821425557,0.1488186419,0.0865965933,-0.1012863368,0.1088703945,-0.0681262389,0.2061406523,-0.3407063186,-0.2190373242,-0.2149083614,-0.1311086267,0.1939715594,-0.0656707883,-0.3104209304,-0.0575122982,-0.40399459,0.3987495005,-0.0498114154,-0.1162562221,0.4287350178,0.0294659249,-0.0422694199,0.0137770846,-0.0411637686,0.2903231978,0.3932226598,0.1544591188,-0.0068705431,0.4821072817,-0.1035226807,0.6277320981,0.1472429186,-0.1187792346,0.1651859134,-0.2604223192,-0.0103919124,-0.2778404653,-0.4354406893,-0.2193963528,-0.249126181,-0.1881687343,-0.0942807272,-0.1138285473,-0.3708038032,-0.2214561105,-0.1073781401,-0.0773165748,-0.2619494498,0.0571713336,-0.4139858186,0.0110577829,-0.2170363218,0.1705489904,-0.2722671926,0.050975807,-0.0055782823,-0.173553735,-0.1463996619,-0.0605672114,-0.0891205817,-0.4248179495,-0.5043855906,0.2438041568,0.0684798583,0.5037469864,-0.0588183887,-0.1827223599,-0.053491801,0.0437431261,0.4503590763,0.081840679,0.0767365023,0.3834943771,-0.0271159373,-0.1729907244,-0.2014370114,-0.1577868909,0.3268741369,0.2122015804,0.0278680604,-0.1312448978,-0.0968372524,0.6783878803,-0.2013685852,-0.1134384498,0.2472415715,-0.1104991585,-0.2218187749,-0.0550959632,0.0236959718,0.26150316,0.1157831475,-0.1673085839,0.1173743159,-0.2727948129,-0.3854475319,0.0342627801,0.1199691072,0.1211441532,-0.2953989506,0.0236995853,-0.0741612986,0.2911489606,0.135410741,0.4349352419,-0.1963123232,-0.2571867704,-0.0590305664,-0.0588326231,0.569011569,0.0825147107,0.0205395147,0.1691699028,-0.2912230194,0.1662970483,-0.0807007626,-0.0752918944,0.4156784713,-0.0272651277,-0.3784964979,-0.2828960121,0.6297958493,0.1839873344,0.1465428174,0.4473782182,-0.1260866523,-0.239024803,0.4521989226,-0.2326593995,0.8379457593,0.3543151319,0.0345559679,0.3221448064,-0.069835566,0.2265108228,-0.080510892,0.2254851162,-0.1784041226,0.0949576944,0.0306908358,-0.1240235716,0.1027829573,-0.2143863142,-0.3066275716,0.3105449677,-0.0997912958,-0.1830910295,-0.016052356,0.0808085725,0.1892041266,0.0173187181,0.0460605249,0.1089305207,0.1697837561,0.3077163994,0.1177056059,-0.0760183781,-0.3178927004,-0.2105564028,-0.3294270933,0.1939419061,-0.1657883525,0.4575289488,0.1740498692,-0.5109893084,0.241724506,-0.0237299819,0.1241833866,-0.1209780425,-0.0642707422,0.2385888547,0.2298075706,-0.0145975659,0.0247175526,-0.0187843982,0.4653916061,-0.0687313676,-0.2951426804,0.1553077102,-0.1590778083,-0.0627102554,0.1183353886,0.1692362279,0.506981194,-0.4509812891,-0.1137202382,0.0539807193,-0.0592885688,-0.2198138684,0.1706585437,-0.1915046275,-0.2810474932,0.2122241706,-0.1172079667,-0.2693044543,0.0231593475,0.2857912183,0.1320576221,0.3819643557,0.7637184858,-0.1087173,0.1665912569,-0.3172612786,0.0459272191,0.343903929,-0.4955285192,0.4000358284,-0.1736086011,0.1612942815,0.1612255871,0.0364253148,-0.0120910564,0.2065973133,-0.116706565,-0.4244697988,-0.0669972673,0.1406808645,-0.2026600242,0.1764729768,-0.3554813564,0.5062906146,0.0710919052,0.1680439264,-0.2597596347,0.1511378884,-0.1518443525,0.0694120005,0.1029792204,-0.1967865229,0.2967450321,0.0523257926,0.1411973983,0.131764248,0.109526962,-0.2064211071,-0.1140705496,0.1071471721,-0.2135015875,0.0280899666,-0.029248016,0.0243348461,-0.227237016,0.0022383516,-0.0647328869,-0.0633015782,-0.059692245,0.5836427212,0.332142055,0.1007112712,-0.0343455561,0.2466231436,0.0156435929,0.084494099,-0.0593365021,0.3571885228,0.0753834769,0.051588051,-0.3514332771,-0.1166024953,-0.2117427886,-0.3910141885,0.4337399304,-0.3484520614,-0.0929033458,-0.3318519592,0.1221028417,0.4578700662,-0.1105462164,0.1637624353,0.2220796049,0.1883313656,-0.3358791769,0.1960199475,0.2058407068,0.0405330025,-0.0001227967,0.1974725723,-0.0650933012,0.1559744477,-0.4457393885,0.2461038679,0.0787578151,-0.218693018,0.5223844051,0.4790277779,0.2139100581,-0.1687202603,0.2140117139,-0.0118133053,0.2746187449,0.3674997687,-0.1348162889,0.5656125546,0.1922338158,0.0937947929,0.3579902053,-0.1663382053,0.1045740545,0.1018048227,-0.1694857925,-0.084567219,-0.1508052349,0.745862782,-0.2528184056,-0.0816629604,-0.211300388,0.1827624887,-0.0460881293,-0.1327363104,-0.0287312623,-0.1834551245,-0.3686963022,-0.083012946,-0.0011285432,-0.031198727,0.3039413393,0.1701010466,-0.0186068248,-0.1554469466,-0.0170353968,0.185970068,0.3770774305,-0.3964030445,-0.0265030395,0.23808451,-0.3629777133,-0.0248213522,0.2360650003,0.5855591297,0.1126054302,-0.2880745828,-0.1827214062,0.0787260011,-0.3212321699,0.2761600912,0.0825777352,0.0580376498,0.0320269577,0.0299263485,0.1374816895,-0.0962639973,0.0584992431,-0.052360855,0.0778181106,-0.1484448612,0.8196856976,0.0112562599,-0.3083073199,-0.1919409186,0.0731802285,-0.2717074156,-0.1773722023,0.3026743531,-0.0234757457,-0.0379697047,0.0162493847,0.0411055833,-0.2221356332,0.4902878106,0.3286442161,0.086545907,-0.3322231472,-0.2518153191,-0.0412197076,-0.1410901397,-0.3503188789,-0.2384199649,0.0799844041,0.1760599315,-0.1894828528,-0.0119381575,-0.0451911166,0.0146008618,-0.0090696802,0.1191632152,-0.5761774778,-0.0756114945,-0.1900818646,0.0003165458,0.2170701325,-0.2311822772,-0.2273955941,-0.1525646895,-0.0609026961,0.0760396048,-0.166669488,0.3527786136,0.1116655096,0.0748518184,0.0908468813,0.4753064215,0.1108889282,-0.055627428,0.0375955738,-0.3761961162,-0.28769961,0.0924587995,0.0837069303,0.3483585417,-0.0560296662,-0.0436204784,-0.3642679751,0.1659378111,-0.0382245667,-0.3190695643,-0.0965507105,0.0530077107,-0.0609477088,0.1682776064,0.3351201713,0.2957863212,-0.1564223617,0.1397153735,-0.2916174233,-0.3726358414,0.393309921,-0.4297352433,-0.1567338854,-0.1844888777,0.2845656872,-0.0461202115,0.1168633625,-0.4802741706,-0.0392120332,0.3435289562,-0.0686688647,-0.1883485168,0.4739549756,0.1762921214,0.2088628113,0.0578347668,0.6851693988,-0.0463124141,-0.0454227105,0.0294385478,-0.1360788196]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/409","title":"train_test_split error: 'dict' object has no attribute 'deepcopy'","comments":"Thanks @lhoestq, works fine now!","body":"`train_test_split` is giving me an error when I try and call it:\r\n\r\n`'dict' object has no attribute 'deepcopy'`\r\n\r\n## To reproduce\r\n\r\n```\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.train_test_split(test_size=0.2)\r\n```\r\n\r\n## Full Stacktrace\r\n```\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-12-feb740dbec9a> in <module>\r\n 1 dataset = load_dataset('glue', 'mrpc', split='train')\r\n----> 2 dataset = dataset.train_test_split(test_size=0.2)\r\n\r\n~\/anaconda3\/envs\/fastai2_me\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in train_test_split(self, test_size, train_size, shuffle, seed, generator, keep_in_memory, load_from_cache_file, train_cache_file_name, test_cache_file_name, writer_batch_size)\r\n 1032 \"writer_batch_size\": writer_batch_size,\r\n 1033 }\r\n-> 1034 train_kwargs = cache_kwargs.deepcopy()\r\n 1035 train_kwargs[\"split\"] = \"train\"\r\n 1036 test_kwargs = cache_kwargs.deepcopy()\r\n\r\nAttributeError: 'dict' object has no attribute 'deepcopy'\r\n```","comment_length":5,"text":"train_test_split error: 'dict' object has no attribute 'deepcopy'\n`train_test_split` is giving me an error when I try and call it:\r\n\r\n`'dict' object has no attribute 'deepcopy'`\r\n\r\n## To reproduce\r\n\r\n```\r\ndataset = load_dataset('glue', 'mrpc', split='train')\r\ndataset = dataset.train_test_split(test_size=0.2)\r\n```\r\n\r\n## Full Stacktrace\r\n```\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-12-feb740dbec9a> in <module>\r\n 1 dataset = load_dataset('glue', 'mrpc', split='train')\r\n----> 2 dataset = dataset.train_test_split(test_size=0.2)\r\n\r\n~\/anaconda3\/envs\/fastai2_me\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py in train_test_split(self, test_size, train_size, shuffle, seed, generator, keep_in_memory, load_from_cache_file, train_cache_file_name, test_cache_file_name, writer_batch_size)\r\n 1032 \"writer_batch_size\": writer_batch_size,\r\n 1033 }\r\n-> 1034 train_kwargs = cache_kwargs.deepcopy()\r\n 1035 train_kwargs[\"split\"] = \"train\"\r\n 1036 test_kwargs = cache_kwargs.deepcopy()\r\n\r\nAttributeError: 'dict' object has no attribute 'deepcopy'\r\n```\nThanks @lhoestq, works fine now!","embeddings":[-0.2237376273,-0.1666347086,0.0247758254,0.1338874251,0.1938315183,0.0552755967,0.4678992331,0.4575436115,0.0768526345,0.0668793321,0.012559779,0.4325339496,-0.1328224838,0.1475853473,-0.0321431756,-0.3626011908,0.045861721,0.1101353839,-0.0206512138,0.0540111996,-0.2322868109,0.1536903083,-0.2952710986,0.1216394454,-0.3847426176,-0.1308465451,-0.1611794978,0.0507109873,0.0260705072,-0.3410872817,0.4863745868,-0.2768262327,0.1210490242,0.4387665987,-0.0001167313,0.0897380039,0.4525579512,-0.0031594248,-0.4532915056,-0.1906421781,-0.4979248941,0.0043902402,0.3513647914,-0.5441683531,-0.042019438,0.0753790438,-0.2040430009,-0.1419107765,0.6363607645,0.485032022,0.2005514503,0.0791381225,0.1345571578,-0.3784969151,0.2663288116,0.2557847798,-0.1203861088,0.2592293322,-0.2897607982,-0.1369455159,0.1920061558,0.1560190469,-0.075342685,0.1783074588,-0.1611261666,0.1872690618,0.0479254462,-0.2695594728,0.018743027,0.256893903,-0.030249713,-0.3257185221,-0.2560101151,-0.1671967506,0.0584573485,-0.6988596916,-0.1965681762,0.0415045246,-0.0502516367,0.1211426407,0.0609849878,-0.0087242918,-0.046128083,0.318919301,-0.0811031461,-0.0589174666,-0.0699505582,0.1485081762,0.3835340142,-0.0294236466,-0.0449157469,0.1503052711,-0.2205448449,0.032056082,-0.1745676249,-0.3116437793,-0.0170895532,-0.2529496253,-0.0491032749,0.3588103652,0.0292291939,-0.100331746,0.2252433896,0.2923784256,0.1448195279,0.3921255469,-0.01603337,0.8706507683,0.1663673222,-0.0608362891,-0.1277957559,-0.2231529504,-0.2105125338,-0.1275177002,0.2760258019,0.1031029373,0.374302417,-0.1266551912,-0.3111316562,-0.3017165363,-0.2804971933,-0.050852064,0.1098144799,0.3778692186,0.0716284737,0.048358757,-0.1034670621,0.3984641731,-0.2277881652,-0.1210849956,-0.2687649429,0.3585449755,-0.1936870664,-0.0207399838,0.2717041373,0.2584705055,0.2085186392,-0.165875122,0.0903264433,-0.2323851436,0.080977574,-0.3380019665,0.135920465,0.2045733333,-0.0628970787,-0.1018280387,0.4085504413,-0.1032658145,-0.1837846041,-0.0074481596,-0.4779140055,-0.4838939905,-0.0802173987,0.1735893488,-0.022120038,-0.1415839344,-0.1402661055,-0.2207283974,0.2963829339,-0.1341463327,0.1054113284,-0.3912715316,-0.1401649565,-0.2248690128,0.27096048,0.3944924474,-0.0726637915,-0.2247758508,-0.4151659608,0.1801807731,0.4060516059,0.1581671238,-0.2056785375,0.3241185844,-0.0558091216,0.1272213757,0.6393452287,-0.3196689188,-0.2742792964,0.1991488636,-0.0163456034,-0.113717556,-0.0419758148,-0.07332436,0.3727302551,-0.0315633751,0.3255939782,0.2582602203,-0.0902778953,-0.1404075176,-0.1173788458,-0.0823601335,0.2345965058,0.100754343,0.1495717764,0.0654751062,-0.0670903474,0.0532700792,0.0841749758,0.1167438552,0.0966118947,-0.0726689547,0.3912801445,-0.0251208488,-0.0648770407,-0.1470053792,0.014197804,0.146988228,0.1346563995,-0.0084777446,-0.1956412196,0.0093174167,-0.4295108318,0.0916422158,-0.4627188146,-0.3100115657,0.1086392403,-0.0640692934,0.1953755468,0.1495377123,-0.1632281393,0.2488712221,-0.0610646196,0.2153518647,-0.1540290564,0.3656193614,-0.1435180604,-0.2986617386,-0.1965119094,0.2870727181,0.2094086111,-0.0837987438,-0.1738697886,0.3216924667,0.1016737893,-0.2982302606,-0.2233414203,0.0615716167,0.1286276728,-0.4170238972,0.1327955574,0.0653676093,0.1673054695,-0.2156138718,-0.0699708834,0.5265491009,0.1284036934,0.3690820634,-0.0253790803,0.2108912021,-0.1056792662,-0.228606984,0.0115543557,-0.1745117307,0.0195378587,-0.2493645698,0.2560393214,-0.3104578257,-0.0174126662,-0.142977491,0.4851713479,0.1946912855,0.18440862,0.0988660753,-0.2053889185,0.229709506,-0.1396404803,0.5375404954,0.5083037615,0.1830110699,-0.0150990617,-0.0095827905,-0.1032946333,0.1383354664,0.1664730161,-0.092275925,0.5140886903,-0.0123880515,0.0872322172,-0.0145534743,-0.0518664531,-0.3417927623,0.144631207,0.1980321854,-0.1573687494,0.2429445684,-0.172168836,-0.2699964345,-0.2016959935,-0.35322088,0.2080784738,-0.2580690384,-0.1814071238,-0.0056798281,0.0083085513,0.2116187215,-0.3330461085,-0.1042994186,0.1641721725,-0.3991348147,0.0728817731,-0.3214917481,-0.1422349513,0.0163805895,0.2154191881,-0.1200496778,0.1293571889,0.1474347115,-0.2544160187,0.2251218706,-0.0909813493,0.1288577616,-0.2310344726,-0.0631724969,0.2680512667,0.0531200729,0.0790477097,-0.3030265868,0.4282800853,-0.0889731199,-0.0469894931,0.2166225761,0.1975534409,-0.1441864967,-0.2191480547,-0.3383539617,-0.2687028646,-0.1957515627,0.3236369193,-0.0482008122,0.0851142928,0.1264806092,0.1895348728,0.1782114208,0.0968200117,-0.149808079,0.0326396078,-0.0455043837,0.1785485595,-0.3783836365,-0.2663778067,-0.1468061805,-0.1395378262,0.1973773241,-0.0368887261,-0.2783103287,-0.0273898467,-0.3523778915,0.3562728465,-0.103067182,-0.0962287635,0.3658639193,0.0069519826,-0.0250953548,-0.0079763224,-0.0624667071,0.3951959014,0.3504797518,0.202726841,-0.0432925299,0.4581619501,-0.1360234916,0.6123536229,0.0799684823,-0.1230117977,0.1780771613,-0.3193722069,0.0196066014,-0.2842378318,-0.471114248,-0.2398414165,-0.2657487988,-0.2621047199,-0.1168338135,-0.1218260676,-0.3318482339,-0.2924024463,-0.0215867404,-0.036727082,-0.2976679504,0.0932395235,-0.4374738038,0.0107705472,-0.1897160262,0.1613941789,-0.2866241336,0.1247691289,-0.0473862626,-0.2133651376,-0.1062941775,-0.0577633493,-0.1250962913,-0.437458545,-0.4455318153,0.2263303548,0.061428342,0.4274280369,-0.1200011075,-0.1477918923,-0.0315134414,0.0708722025,0.4627011418,0.0566523708,0.1193297282,0.380680114,-0.0381255969,-0.2209846824,-0.1836214364,-0.2217107117,0.3221032321,0.1584822685,0.0080786971,-0.2095447928,0.0233612955,0.5947523713,-0.1785821319,-0.119705148,0.23236157,-0.1419417709,-0.2405887693,-0.0742908269,0.0065638092,0.2878642678,0.1110838875,-0.1562038362,0.1698751599,-0.3290117383,-0.3212184608,0.1009476036,0.1781554967,0.1496281177,-0.3580590189,0.0242589563,-0.0932701901,0.299815625,0.1090810597,0.4996471703,-0.1641165763,-0.2857057452,-0.068247892,-0.0741917118,0.5495584011,0.0141590042,-0.010248404,0.1310247034,-0.2833680511,0.1499368399,-0.0965087488,0.0009938417,0.449331224,-0.0151711907,-0.3932029903,-0.2522929013,0.5908629894,0.1517899632,0.1618752033,0.4275559187,-0.1383496672,-0.2505961359,0.5000470877,-0.293428272,0.8349853754,0.2999040186,0.0505525135,0.2965297699,-0.0002367011,0.2485640943,-0.0283368863,0.2276501805,-0.1769082993,0.0550281107,0.0238447022,-0.1489483416,0.0376148447,-0.131754756,-0.2972019911,0.2631440163,-0.148425132,-0.1274537593,-0.0524927638,0.0494528227,0.2259441912,0.024115514,0.0853059366,0.1387590468,0.1443499923,0.3057374358,0.0772451609,-0.0508450456,-0.361284107,-0.2348779887,-0.3025633097,0.1967677623,-0.1834298521,0.4872477949,0.0794636235,-0.4917917848,0.1998867393,0.0297536589,0.1086687967,-0.1284021288,-0.0854862779,0.1848743856,0.1851145625,-0.0955866948,-0.0435394421,-0.0329799093,0.447902441,-0.0478501879,-0.2817271352,0.2107583433,-0.1496100277,-0.0194021072,0.1122102365,0.2437108606,0.463398248,-0.5626877546,-0.1339274347,0.0563785248,-0.0746651515,-0.1677895784,0.1753444672,-0.1752448231,-0.3147115707,0.1600302905,-0.0888042077,-0.2396059632,0.0224979445,0.2590876222,0.1644251943,0.3953074217,0.7222936749,-0.1156617254,0.1707747132,-0.3192651272,0.0566484667,0.3310590386,-0.4414576292,0.3575661778,-0.10436932,0.1438838094,0.0832014233,-0.0297634695,0.0033382855,0.1745625585,-0.0955035537,-0.4136312902,-0.0579436533,0.0914967433,-0.1849781424,0.1534187049,-0.323292762,0.546536386,0.0759999081,0.1524567157,-0.2664961219,0.1259563863,-0.2464435846,0.135237366,0.0624086745,-0.1872020811,0.2935563326,0.072297588,0.1714638472,0.1783162653,0.105761081,-0.2152785659,-0.0842148736,0.0942484289,-0.1125969589,0.0284871049,-0.0355671719,-0.0128739299,-0.181404978,-0.0491164513,-0.0535689853,0.0045855879,-0.0787488893,0.5128445625,0.3473595977,0.1809879988,-0.0977537781,0.20044972,-0.0579688884,0.1056374609,-0.0133420182,0.2592493892,0.0667259917,0.0471073836,-0.3908107877,-0.1220597848,-0.1007888466,-0.3801797032,0.4201513231,-0.3966470361,-0.0848821774,-0.3308744431,0.1671884209,0.4310074449,-0.1385966837,0.2755485177,0.1971280426,0.2149016559,-0.3132866919,0.2336025536,0.2339524478,-0.0251024459,0.0311051793,0.2292168587,-0.1634272933,0.184696123,-0.505566597,0.1287947893,0.0925190896,-0.1827364713,0.4289779961,0.4428696334,0.1460971534,-0.1539914757,0.1702006906,0.0314226002,0.3131949008,0.3895846605,-0.1682211608,0.5445082784,0.1239097416,0.0577314906,0.3448649347,-0.1697044522,0.0531908311,0.1213947609,-0.1203369647,-0.0670556128,-0.2571202815,0.6664947271,-0.278023243,-0.0926993415,-0.1738493294,0.1601096392,-0.092634581,-0.0383522771,-0.0059487023,-0.1424242407,-0.3612774909,-0.1593709737,0.0375932604,-0.07043235,0.3379016221,0.168769896,0.00000262,-0.1049968824,-0.030643804,0.2285993695,0.3036784828,-0.3980864882,-0.0220757164,0.2579899132,-0.3460808396,-0.0085535152,0.1870642602,0.5588656664,0.1306631714,-0.289583236,-0.1534369737,0.0709279478,-0.2915350795,0.2486316264,0.0390293114,0.0457911715,0.0722931102,0.1546709538,0.1482148468,-0.1306961179,0.0283210278,-0.0264124628,0.0822282881,-0.1758103222,0.770688951,-0.087865144,-0.2668913007,-0.267591387,0.0274267495,-0.3244686425,-0.1616200805,0.2886746824,-0.0531021133,0.0337102525,0.0299364589,0.0448307991,-0.2757689059,0.4800130129,0.2786825299,0.1475610137,-0.4035181999,-0.2230845988,-0.087453194,-0.0376963876,-0.4874516726,-0.2377933264,0.0411876366,0.17192249,-0.1560558379,0.0514282919,-0.055530224,0.0911747888,-0.0414078683,0.1393089145,-0.5039960146,-0.2127579153,-0.2002324164,-0.0354891568,0.2438368201,-0.2087217122,-0.2184689641,-0.1449593306,-0.0314826146,0.1131527126,-0.1717859805,0.3044438958,0.1203171909,0.1211145744,0.1430058628,0.5090491176,0.135265097,0.0113679869,0.1078686789,-0.3867365718,-0.2545471489,0.0760189295,0.1491142958,0.3104173839,-0.1040477082,-0.0129911946,-0.4120004475,0.18505463,0.0356744453,-0.3153178394,-0.0668743774,0.0481930301,-0.0884909108,0.1987333298,0.309056282,0.3111336827,-0.1802778989,0.1830619574,-0.2077221572,-0.3747706711,0.4423579872,-0.3777779341,-0.2091790736,-0.0873713195,0.3015685081,-0.0454323255,0.0623608008,-0.4476456046,-0.0365835093,0.3393935859,-0.0771086812,-0.2200890034,0.342856288,0.1607447118,0.194856137,0.1043939218,0.6668985486,-0.028542744,-0.056552995,0.1470709294,-0.1179357693]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/407","title":"MissingBeamOptions for Wikipedia 20200501.en","comments":"Fixed. Could you try again @mitchellgordon95 ?\r\nIt was due a file not being updated on S3.\r\n\r\nWe need to make sure all the datasets scripts get updated properly @julien-c ","body":"There may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available):\r\n\r\n```\r\nnlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n```\r\n\r\nAnd now, having pulled master, I get:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, total: 34.06 GiB) to \/home\/hltcoe\/mgordon\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/76b0b2747b679bb0ee7a1621e50e5a6378477add0c662668a324a5bc07d516dd...\r\nTraceback (most recent call last):\r\n File \"scripts\/download.py\", line 11, in <module>\r\n fire.Fire(download_pretrain)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 138, in Fire\r\n component_trace = _Fire(component, args, parsed_flag_args, context, name)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 468, in _Fire\r\n target=component.__name__)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 672, in _CallAndUpdateTrace\r\n component = fn(*varargs, **kwargs)\r\n File \"scripts\/download.py\", line 6, in download_pretrain\r\n nlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/load.py\", line 534, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 460, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 870, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, S\r\npark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory).\r\nExample of usage:\r\n `load_dataset('wikipedia', '20200501.en', beam_runner='DirectRunner')`\r\n```","comment_length":30,"text":"MissingBeamOptions for Wikipedia 20200501.en\nThere may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available):\r\n\r\n```\r\nnlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n```\r\n\r\nAnd now, having pulled master, I get:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, total: 34.06 GiB) to \/home\/hltcoe\/mgordon\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/76b0b2747b679bb0ee7a1621e50e5a6378477add0c662668a324a5bc07d516dd...\r\nTraceback (most recent call last):\r\n File \"scripts\/download.py\", line 11, in <module>\r\n fire.Fire(download_pretrain)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 138, in Fire\r\n component_trace = _Fire(component, args, parsed_flag_args, context, name)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 468, in _Fire\r\n target=component.__name__)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 672, in _CallAndUpdateTrace\r\n component = fn(*varargs, **kwargs)\r\n File \"scripts\/download.py\", line 6, in download_pretrain\r\n nlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/load.py\", line 534, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 460, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 870, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, S\r\npark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory).\r\nExample of usage:\r\n `load_dataset('wikipedia', '20200501.en', beam_runner='DirectRunner')`\r\n```\nFixed. Could you try again @mitchellgordon95 ?\r\nIt was due a file not being updated on S3.\r\n\r\nWe need to make sure all the datasets scripts get updated properly @julien-c ","embeddings":[0.0666168258,0.1266757995,-0.0205123089,-0.0542329103,0.1996194869,0.1526870281,0.1799744368,0.3656727076,-0.022208821,-0.081425719,0.1489358097,0.2753814459,-0.0917265415,-0.168423295,0.2434722483,-0.3058640659,0.1455037892,0.1239019409,-0.3501991034,-0.1286828965,-0.0025927718,0.4125896096,-0.2114285231,-0.1782554686,-0.5298822522,0.0631536841,-0.1315196306,0.1428049058,-0.0437919199,-0.2499871403,0.2779204845,0.0593028367,-0.0827241987,0.122238256,-0.0001143081,0.0776108727,0.62224859,-0.2239052504,-0.5686735511,0.0119778551,0.1300001889,-0.3336535096,0.384344399,-0.2550982535,-0.2759362757,-0.1181294397,0.1141192019,-0.0585302375,0.3938059509,-0.0109363226,0.2136250138,0.2360561192,0.297298044,-0.1291413456,0.6020246148,0.0674149916,-0.1662565321,0.0793358013,-0.1638745517,-0.0564770475,-0.2710405588,0.4797945023,0.1387171596,-0.1370008588,0.5526376367,-0.2816212773,0.0226213243,-0.2574208677,0.0174538549,0.1826507896,0.4940163493,-0.0229834914,-0.1063677594,-0.3746719062,0.0446043313,0.1787300706,0.2480826825,0.0172076635,-0.1532892138,-0.0672086403,0.0512106009,-0.2295793146,-0.0332925208,0.28875193,-0.0673940703,0.4202920794,-0.0289063603,0.0214702915,0.166854009,0.0496226139,-0.313372463,0.0365183763,0.0982150361,0.2350160927,-0.2165132016,0.0636979118,0.2868846357,0.2705270946,0.0620143712,0.1695944816,-0.2794000506,0.1912726462,0.0204293542,-0.0470910259,0.2198940963,0.2408223003,-0.0418657698,-0.353320688,0.251672864,0.3364218175,-0.0628019497,0.1384649873,0.1992915124,-0.1387363076,-0.2713562548,0.1334020048,0.0091168443,-0.1717023998,0.1072276235,0.0534083545,-0.2777944505,-0.1483329535,-0.2378234863,0.31988886,-0.0257740859,0.4086242616,0.2018820941,0.1185922623,-0.2835624218,-0.2737653553,-0.1957246959,0.0212342087,-0.4779030085,0.0258821566,0.3871365488,-0.2097051591,0.5165669322,0.0758559778,-0.0362861082,0.0592919849,-0.0301977824,0.2644242346,-0.0725837499,0.181024909,0.1212264374,0.2927809656,0.2036839426,0.0199421402,-0.0530857034,-0.329444617,0.0704844519,-0.2763873637,-0.2597137988,0.1677166075,-0.3119882941,0.1547896415,0.047507599,0.1872284114,-0.0253873691,-0.1822030544,-0.0967266262,-0.1229704767,-0.0582913347,-0.1482739151,0.5051084161,0.4033136666,0.0298880525,-0.3304650486,-0.0621768199,0.1041162759,-0.1011983678,0.0117212692,-0.1564818919,0.5247724652,-0.2857876718,-0.1501885206,0.224312827,0.0519686528,-0.2112267464,0.1486301273,-0.0126577159,0.2708986998,0.0975890458,-0.0528358966,-0.0672804639,0.0518001579,0.0919757187,0.2184601426,0.1131943539,0.0529397614,-0.3270510733,-0.0877883509,-0.0076871128,0.0795486197,0.3505393863,-0.0505552888,0.0635950118,0.5883682966,0.2980857491,-0.1387826651,0.2244143188,0.4500527084,-0.2889430523,0.1368469149,0.2453726381,-0.0598621815,-0.4651874602,0.1208006591,-0.1401829123,0.4367903173,-0.1831934154,-0.00221489,-0.3467137516,-0.1367142648,-0.1968161315,-0.5165299773,0.2065130919,0.2034472376,-0.018576948,0.3046480119,0.0254303049,0.1162029579,-0.040442083,0.0568883121,-0.7796523571,0.214526847,-0.3115582466,-0.1411182284,-0.0192293879,0.3081658483,0.1150461584,-0.1935398728,-0.1284949929,0.2234227657,-0.0789242238,0.1341899037,0.0148870628,-0.0550866164,0.2404268086,-0.3211013675,0.0984831378,0.3122585118,-0.0463057272,-0.1272407025,-0.3034783602,0.2799139619,0.0182238296,0.3226444721,0.033403419,0.0919322968,0.1548368037,0.0725830048,-0.2331330776,-0.202653572,0.3785173595,0.2176063359,0.1366422474,-0.0235250685,0.114962697,0.0045698718,0.1719020158,0.0660793856,0.0112598883,-0.0809159651,-0.4510022402,-0.2482421845,0.3509304225,-0.0857894942,0.0816844404,0.2645476758,0.0524951369,-0.0169572085,-0.1228682548,-0.2095779032,0.1974878311,0.1140880585,0.3243004978,0.0876459852,0.1405431777,0.1975445449,-0.3815246522,0.1906591803,-0.1774683148,0.0871339515,-0.2838542759,-0.0979878008,-0.3540599644,-0.4514617026,-0.1052412391,0.2488120049,-0.3974134922,-0.5874935985,-0.0569580458,0.1890996397,-0.1633748561,0.3057290614,-0.0293643046,-0.1303642839,0.0741929859,0.019791821,-0.3342637122,-0.3922384083,-0.200067386,0.0861517861,0.1043603644,-0.0021498958,0.2901448607,-0.0029742981,-0.0271919109,-0.4731692374,-0.575897634,0.3161930442,-0.2838526368,0.190267697,0.1779287755,0.0881575495,-0.2256871164,-0.058733575,0.4069474936,-0.1701157838,-0.1504672468,-0.0161600616,-0.1277897358,0.101005815,-0.1923410892,-0.2217373699,0.0953757316,-0.3692909479,0.0153438859,0.0921726599,-0.1812787056,0.4691084921,-0.1002497077,-0.0062876246,0.1005272493,0.1722671986,-0.2448800653,0.0835745335,0.3416597843,-0.3003979027,-0.4568088353,0.0295674503,-0.1735530943,0.0571707934,0.318931818,-0.4997883439,-0.0496766046,0.0941225365,0.0024681096,-0.1354268342,0.0942459553,0.5044302344,0.0114218183,-0.0168616381,-0.2326325625,-0.1519111693,-0.1495186388,-0.3148490787,0.4269850552,-0.192369014,0.32359007,-0.0036709276,1.0987776518,-0.0326621234,0.0656037927,0.4224383831,0.093303889,0.4521008432,-0.2759513259,-0.296728462,0.2323563248,-0.1486648619,0.301333636,0.39190045,-0.0147325573,-0.2050670981,-0.365370661,0.009143726,-0.2090857476,-0.3228644729,0.0901104659,-0.3193496764,0.291395694,0.1660908908,0.2335969061,-0.001936554,-0.1577165425,0.2441106439,0.1709194183,0.2070368826,0.0695521533,0.1379282326,-0.056091655,-0.498675853,0.0409604572,0.0189635176,-0.003006133,0.0389499553,-0.1371269971,0.2020363808,0.045385763,0.568950057,-0.1977424026,-0.0263437368,0.2490573078,0.0347809158,-0.7107806206,-0.0275662523,-0.0550747104,0.1947839111,0.2862170339,0.1784339249,-0.3835534155,-0.2234925479,0.3125829697,0.0610993989,-0.0469833761,-0.0261743944,-0.2916185558,0.0066375742,-0.420763731,-0.1309033781,-0.1549676657,0.145596087,0.0774429739,0.168731302,0.0894142538,0.0287991762,-0.1335077137,0.1053509489,0.2781824172,0.2502248585,-0.1114856303,0.1409600228,0.6523587108,-0.3377575278,0.2628805041,0.198228538,0.0455500036,0.2173913866,-0.0400636755,-0.0590368211,0.3987356126,0.0289372355,-0.048762694,-0.203052327,0.0094223302,0.094053559,0.229137376,0.3506721556,-0.2690441608,-0.1362774521,-0.2736847401,0.5236557126,-0.0176693387,-0.0113714179,0.0692942441,0.067014277,-0.3365369737,0.346108377,0.2415182889,0.7413404584,0.0380274691,0.3953185976,0.2077925205,0.3027880788,0.8601722717,-0.3698050082,0.189751476,-0.1742741317,0.2201997638,0.0416868031,0.035432715,-0.1920433491,-0.1310457438,-0.0968420133,0.3844055831,0.3848448396,0.4188909829,0.0760006681,0.2999126911,0.0761646107,-0.3209706545,-0.1950532347,0.1271081567,-0.2076167464,0.5865810513,-0.3765703142,-0.1308689266,-0.2108011395,-0.1166528091,-0.2505488098,0.188552022,-0.2747045159,0.5069428682,-0.1782887429,-0.5959970355,0.0212483089,0.112775512,0.1474469155,0.0883108452,-0.4008170366,0.3684974611,-0.194089815,-0.2362576425,-0.1085617468,0.2097142488,0.4952664971,-0.1708497852,-0.0616780668,-0.0610723421,-0.2416674942,-0.2755625546,-0.2949698865,-0.0613051355,0.0028359615,0.1048691124,-0.0824528262,0.1037253663,-0.0140288854,-0.2572488785,0.1487520188,-0.1144913137,0.0374176167,-0.1460709274,0.102444768,-0.1043076217,-0.0703977495,0.4993968308,-0.0020985689,-0.1289848834,0.5953381062,0.3426305056,-0.1198802143,-0.11467278,-0.0902574286,-0.1393836737,-0.472225517,-0.1296564788,0.0124801882,-0.0530011132,-0.0521329977,0.2447199076,0.1844117939,-0.1626237184,0.0780859217,-0.1943453699,-0.2954605818,0.2547383606,-0.3494720161,-0.0420393012,0.1390184164,0.1013662666,0.3333048522,0.0148150027,-0.2964564562,-0.2154649049,0.0540286303,0.2008091211,0.3083446026,-0.1186889783,-0.0500540733,-0.1276490241,0.1212444231,0.0212709345,-0.1894858032,-0.2302402109,-0.0315677673,0.1042667329,0.0467173308,0.0865679979,-0.078176856,0.1114092842,-0.0371292271,-0.2535363138,0.0528716072,-0.2281834632,0.1240534335,0.1823162884,0.1191580296,0.1763534993,-0.1209528968,0.0150982216,0.0482740849,0.1423756182,0.0532632731,0.0129568297,0.2215794772,-0.0117944498,-0.1690639108,0.1042752936,-0.1854999661,0.2854600251,0.2926937342,-0.0773175657,-0.2990315855,-0.0270367842,0.2942747474,-0.0326080546,-0.0528618991,-0.0485746935,0.0103593031,0.190612644,-0.2196088284,0.267460674,0.5279888511,0.1143720597,0.055378437,-0.0702123344,0.374345392,-0.0246604662,0.3387497365,0.1726608127,0.1217527017,-0.1183099598,0.242642194,0.2166147381,0.3510863185,0.0566907004,0.0239127222,-0.0307099391,0.2304615378,0.2397233546,0.1288086921,0.2786290646,-0.0882287398,0.0975492597,0.0309134126,-0.0629837662,0.1639131606,0.3340352774,0.180387482,0.1440526098,-0.1291939914,0.5687990189,0.1001943797,-0.3893401027,-0.2345822453,0.2402310967,-0.1169962138,0.0740318149,-0.3172836304,-0.1861781478,-0.2391943634,0.1574652493,-0.1074403599,-0.1232673377,-0.0504299775,0.2156560123,-0.1900828779,-0.4732025564,0.2101701349,-0.0363380834,0.3338373005,-0.0730825141,0.1706403196,-0.0616370812,-0.2775771618,-0.0798462331,0.2877354026,0.2786316872,0.1054166332,-0.3909924626,-0.1218862981,0.1556931436,-0.0640750453,-0.1527651697,0.0842841491,0.2606301904,-0.101303719,0.2616636157,0.0972530246,-0.2740935087,-0.1220369935,0.0980754793,-0.0550222024,-0.1236003935,0.0130101694,-0.3181710243,-0.19825764,-0.1200942472,0.0065360139,-0.4909358025,0.1610680819,-0.0082066758,0.1992535591,-0.0541196167,-0.2968486249,0.0856656879,-0.1368458569,0.59737885,0.3424452245,0.1117996126,-0.146999687,-0.5365690589,-0.5853412151,0.2090298384,-0.062993072,-0.0486578308,0.0740430281,0.1942421943,0.2390605956,0.1642809063,0.0021003089,-0.1792952865,0.0889608636,-0.0662597269,-0.4647616744,-0.0645498931,-0.1904877871,-0.2589036822,-0.1246322542,0.0241009742,-0.0880486891,0.105777964,0.0488002226,-0.0199225526,-0.2640572786,0.0616629459,-0.0682748631,0.2982192934,-0.0519034304,0.3681990504,-0.1012711897,-0.220356375,-0.2901730239,-0.1561024785,-0.0462986268,0.3026449978,-0.1934378147,0.521027565,-0.0083604716,-0.100206092,-0.3288210034,0.6905267239,0.1733905077,0.1175569594,-0.3923233449,0.0239075776,-0.3116728961,0.1622987688,0.091017738,0.3411852121,-0.2714780271,0.1741360724,-0.4668044746,-0.4499904215,0.2498712838,-0.4550741613,-0.092697911,-0.0041839303,0.1153395548,-0.004739088,-0.0571005195,-0.5571916699,0.0335970558,0.1637207419,0.054309655,-0.0907495096,0.1095089465,-0.0455884449,-0.1208940893,-0.104185313,-0.1784140915,0.1156888828,-0.1695554703,-0.089875564,-0.2151401788]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/407","title":"MissingBeamOptions for Wikipedia 20200501.en","comments":"Works for me! Thanks.","body":"There may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available):\r\n\r\n```\r\nnlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n```\r\n\r\nAnd now, having pulled master, I get:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, total: 34.06 GiB) to \/home\/hltcoe\/mgordon\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/76b0b2747b679bb0ee7a1621e50e5a6378477add0c662668a324a5bc07d516dd...\r\nTraceback (most recent call last):\r\n File \"scripts\/download.py\", line 11, in <module>\r\n fire.Fire(download_pretrain)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 138, in Fire\r\n component_trace = _Fire(component, args, parsed_flag_args, context, name)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 468, in _Fire\r\n target=component.__name__)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 672, in _CallAndUpdateTrace\r\n component = fn(*varargs, **kwargs)\r\n File \"scripts\/download.py\", line 6, in download_pretrain\r\n nlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/load.py\", line 534, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 460, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 870, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, S\r\npark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory).\r\nExample of usage:\r\n `load_dataset('wikipedia', '20200501.en', beam_runner='DirectRunner')`\r\n```","comment_length":4,"text":"MissingBeamOptions for Wikipedia 20200501.en\nThere may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available):\r\n\r\n```\r\nnlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n```\r\n\r\nAnd now, having pulled master, I get:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, total: 34.06 GiB) to \/home\/hltcoe\/mgordon\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/76b0b2747b679bb0ee7a1621e50e5a6378477add0c662668a324a5bc07d516dd...\r\nTraceback (most recent call last):\r\n File \"scripts\/download.py\", line 11, in <module>\r\n fire.Fire(download_pretrain)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 138, in Fire\r\n component_trace = _Fire(component, args, parsed_flag_args, context, name)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 468, in _Fire\r\n target=component.__name__)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 672, in _CallAndUpdateTrace\r\n component = fn(*varargs, **kwargs)\r\n File \"scripts\/download.py\", line 6, in download_pretrain\r\n nlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/load.py\", line 534, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 460, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 870, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, S\r\npark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory).\r\nExample of usage:\r\n `load_dataset('wikipedia', '20200501.en', beam_runner='DirectRunner')`\r\n```\nWorks for me! Thanks.","embeddings":[0.0666168258,0.1266757995,-0.0205123089,-0.0542329103,0.1996194869,0.1526870281,0.1799744368,0.3656727076,-0.022208821,-0.081425719,0.1489358097,0.2753814459,-0.0917265415,-0.168423295,0.2434722483,-0.3058640659,0.1455037892,0.1239019409,-0.3501991034,-0.1286828965,-0.0025927718,0.4125896096,-0.2114285231,-0.1782554686,-0.5298822522,0.0631536841,-0.1315196306,0.1428049058,-0.0437919199,-0.2499871403,0.2779204845,0.0593028367,-0.0827241987,0.122238256,-0.0001143081,0.0776108727,0.62224859,-0.2239052504,-0.5686735511,0.0119778551,0.1300001889,-0.3336535096,0.384344399,-0.2550982535,-0.2759362757,-0.1181294397,0.1141192019,-0.0585302375,0.3938059509,-0.0109363226,0.2136250138,0.2360561192,0.297298044,-0.1291413456,0.6020246148,0.0674149916,-0.1662565321,0.0793358013,-0.1638745517,-0.0564770475,-0.2710405588,0.4797945023,0.1387171596,-0.1370008588,0.5526376367,-0.2816212773,0.0226213243,-0.2574208677,0.0174538549,0.1826507896,0.4940163493,-0.0229834914,-0.1063677594,-0.3746719062,0.0446043313,0.1787300706,0.2480826825,0.0172076635,-0.1532892138,-0.0672086403,0.0512106009,-0.2295793146,-0.0332925208,0.28875193,-0.0673940703,0.4202920794,-0.0289063603,0.0214702915,0.166854009,0.0496226139,-0.313372463,0.0365183763,0.0982150361,0.2350160927,-0.2165132016,0.0636979118,0.2868846357,0.2705270946,0.0620143712,0.1695944816,-0.2794000506,0.1912726462,0.0204293542,-0.0470910259,0.2198940963,0.2408223003,-0.0418657698,-0.353320688,0.251672864,0.3364218175,-0.0628019497,0.1384649873,0.1992915124,-0.1387363076,-0.2713562548,0.1334020048,0.0091168443,-0.1717023998,0.1072276235,0.0534083545,-0.2777944505,-0.1483329535,-0.2378234863,0.31988886,-0.0257740859,0.4086242616,0.2018820941,0.1185922623,-0.2835624218,-0.2737653553,-0.1957246959,0.0212342087,-0.4779030085,0.0258821566,0.3871365488,-0.2097051591,0.5165669322,0.0758559778,-0.0362861082,0.0592919849,-0.0301977824,0.2644242346,-0.0725837499,0.181024909,0.1212264374,0.2927809656,0.2036839426,0.0199421402,-0.0530857034,-0.329444617,0.0704844519,-0.2763873637,-0.2597137988,0.1677166075,-0.3119882941,0.1547896415,0.047507599,0.1872284114,-0.0253873691,-0.1822030544,-0.0967266262,-0.1229704767,-0.0582913347,-0.1482739151,0.5051084161,0.4033136666,0.0298880525,-0.3304650486,-0.0621768199,0.1041162759,-0.1011983678,0.0117212692,-0.1564818919,0.5247724652,-0.2857876718,-0.1501885206,0.224312827,0.0519686528,-0.2112267464,0.1486301273,-0.0126577159,0.2708986998,0.0975890458,-0.0528358966,-0.0672804639,0.0518001579,0.0919757187,0.2184601426,0.1131943539,0.0529397614,-0.3270510733,-0.0877883509,-0.0076871128,0.0795486197,0.3505393863,-0.0505552888,0.0635950118,0.5883682966,0.2980857491,-0.1387826651,0.2244143188,0.4500527084,-0.2889430523,0.1368469149,0.2453726381,-0.0598621815,-0.4651874602,0.1208006591,-0.1401829123,0.4367903173,-0.1831934154,-0.00221489,-0.3467137516,-0.1367142648,-0.1968161315,-0.5165299773,0.2065130919,0.2034472376,-0.018576948,0.3046480119,0.0254303049,0.1162029579,-0.040442083,0.0568883121,-0.7796523571,0.214526847,-0.3115582466,-0.1411182284,-0.0192293879,0.3081658483,0.1150461584,-0.1935398728,-0.1284949929,0.2234227657,-0.0789242238,0.1341899037,0.0148870628,-0.0550866164,0.2404268086,-0.3211013675,0.0984831378,0.3122585118,-0.0463057272,-0.1272407025,-0.3034783602,0.2799139619,0.0182238296,0.3226444721,0.033403419,0.0919322968,0.1548368037,0.0725830048,-0.2331330776,-0.202653572,0.3785173595,0.2176063359,0.1366422474,-0.0235250685,0.114962697,0.0045698718,0.1719020158,0.0660793856,0.0112598883,-0.0809159651,-0.4510022402,-0.2482421845,0.3509304225,-0.0857894942,0.0816844404,0.2645476758,0.0524951369,-0.0169572085,-0.1228682548,-0.2095779032,0.1974878311,0.1140880585,0.3243004978,0.0876459852,0.1405431777,0.1975445449,-0.3815246522,0.1906591803,-0.1774683148,0.0871339515,-0.2838542759,-0.0979878008,-0.3540599644,-0.4514617026,-0.1052412391,0.2488120049,-0.3974134922,-0.5874935985,-0.0569580458,0.1890996397,-0.1633748561,0.3057290614,-0.0293643046,-0.1303642839,0.0741929859,0.019791821,-0.3342637122,-0.3922384083,-0.200067386,0.0861517861,0.1043603644,-0.0021498958,0.2901448607,-0.0029742981,-0.0271919109,-0.4731692374,-0.575897634,0.3161930442,-0.2838526368,0.190267697,0.1779287755,0.0881575495,-0.2256871164,-0.058733575,0.4069474936,-0.1701157838,-0.1504672468,-0.0161600616,-0.1277897358,0.101005815,-0.1923410892,-0.2217373699,0.0953757316,-0.3692909479,0.0153438859,0.0921726599,-0.1812787056,0.4691084921,-0.1002497077,-0.0062876246,0.1005272493,0.1722671986,-0.2448800653,0.0835745335,0.3416597843,-0.3003979027,-0.4568088353,0.0295674503,-0.1735530943,0.0571707934,0.318931818,-0.4997883439,-0.0496766046,0.0941225365,0.0024681096,-0.1354268342,0.0942459553,0.5044302344,0.0114218183,-0.0168616381,-0.2326325625,-0.1519111693,-0.1495186388,-0.3148490787,0.4269850552,-0.192369014,0.32359007,-0.0036709276,1.0987776518,-0.0326621234,0.0656037927,0.4224383831,0.093303889,0.4521008432,-0.2759513259,-0.296728462,0.2323563248,-0.1486648619,0.301333636,0.39190045,-0.0147325573,-0.2050670981,-0.365370661,0.009143726,-0.2090857476,-0.3228644729,0.0901104659,-0.3193496764,0.291395694,0.1660908908,0.2335969061,-0.001936554,-0.1577165425,0.2441106439,0.1709194183,0.2070368826,0.0695521533,0.1379282326,-0.056091655,-0.498675853,0.0409604572,0.0189635176,-0.003006133,0.0389499553,-0.1371269971,0.2020363808,0.045385763,0.568950057,-0.1977424026,-0.0263437368,0.2490573078,0.0347809158,-0.7107806206,-0.0275662523,-0.0550747104,0.1947839111,0.2862170339,0.1784339249,-0.3835534155,-0.2234925479,0.3125829697,0.0610993989,-0.0469833761,-0.0261743944,-0.2916185558,0.0066375742,-0.420763731,-0.1309033781,-0.1549676657,0.145596087,0.0774429739,0.168731302,0.0894142538,0.0287991762,-0.1335077137,0.1053509489,0.2781824172,0.2502248585,-0.1114856303,0.1409600228,0.6523587108,-0.3377575278,0.2628805041,0.198228538,0.0455500036,0.2173913866,-0.0400636755,-0.0590368211,0.3987356126,0.0289372355,-0.048762694,-0.203052327,0.0094223302,0.094053559,0.229137376,0.3506721556,-0.2690441608,-0.1362774521,-0.2736847401,0.5236557126,-0.0176693387,-0.0113714179,0.0692942441,0.067014277,-0.3365369737,0.346108377,0.2415182889,0.7413404584,0.0380274691,0.3953185976,0.2077925205,0.3027880788,0.8601722717,-0.3698050082,0.189751476,-0.1742741317,0.2201997638,0.0416868031,0.035432715,-0.1920433491,-0.1310457438,-0.0968420133,0.3844055831,0.3848448396,0.4188909829,0.0760006681,0.2999126911,0.0761646107,-0.3209706545,-0.1950532347,0.1271081567,-0.2076167464,0.5865810513,-0.3765703142,-0.1308689266,-0.2108011395,-0.1166528091,-0.2505488098,0.188552022,-0.2747045159,0.5069428682,-0.1782887429,-0.5959970355,0.0212483089,0.112775512,0.1474469155,0.0883108452,-0.4008170366,0.3684974611,-0.194089815,-0.2362576425,-0.1085617468,0.2097142488,0.4952664971,-0.1708497852,-0.0616780668,-0.0610723421,-0.2416674942,-0.2755625546,-0.2949698865,-0.0613051355,0.0028359615,0.1048691124,-0.0824528262,0.1037253663,-0.0140288854,-0.2572488785,0.1487520188,-0.1144913137,0.0374176167,-0.1460709274,0.102444768,-0.1043076217,-0.0703977495,0.4993968308,-0.0020985689,-0.1289848834,0.5953381062,0.3426305056,-0.1198802143,-0.11467278,-0.0902574286,-0.1393836737,-0.472225517,-0.1296564788,0.0124801882,-0.0530011132,-0.0521329977,0.2447199076,0.1844117939,-0.1626237184,0.0780859217,-0.1943453699,-0.2954605818,0.2547383606,-0.3494720161,-0.0420393012,0.1390184164,0.1013662666,0.3333048522,0.0148150027,-0.2964564562,-0.2154649049,0.0540286303,0.2008091211,0.3083446026,-0.1186889783,-0.0500540733,-0.1276490241,0.1212444231,0.0212709345,-0.1894858032,-0.2302402109,-0.0315677673,0.1042667329,0.0467173308,0.0865679979,-0.078176856,0.1114092842,-0.0371292271,-0.2535363138,0.0528716072,-0.2281834632,0.1240534335,0.1823162884,0.1191580296,0.1763534993,-0.1209528968,0.0150982216,0.0482740849,0.1423756182,0.0532632731,0.0129568297,0.2215794772,-0.0117944498,-0.1690639108,0.1042752936,-0.1854999661,0.2854600251,0.2926937342,-0.0773175657,-0.2990315855,-0.0270367842,0.2942747474,-0.0326080546,-0.0528618991,-0.0485746935,0.0103593031,0.190612644,-0.2196088284,0.267460674,0.5279888511,0.1143720597,0.055378437,-0.0702123344,0.374345392,-0.0246604662,0.3387497365,0.1726608127,0.1217527017,-0.1183099598,0.242642194,0.2166147381,0.3510863185,0.0566907004,0.0239127222,-0.0307099391,0.2304615378,0.2397233546,0.1288086921,0.2786290646,-0.0882287398,0.0975492597,0.0309134126,-0.0629837662,0.1639131606,0.3340352774,0.180387482,0.1440526098,-0.1291939914,0.5687990189,0.1001943797,-0.3893401027,-0.2345822453,0.2402310967,-0.1169962138,0.0740318149,-0.3172836304,-0.1861781478,-0.2391943634,0.1574652493,-0.1074403599,-0.1232673377,-0.0504299775,0.2156560123,-0.1900828779,-0.4732025564,0.2101701349,-0.0363380834,0.3338373005,-0.0730825141,0.1706403196,-0.0616370812,-0.2775771618,-0.0798462331,0.2877354026,0.2786316872,0.1054166332,-0.3909924626,-0.1218862981,0.1556931436,-0.0640750453,-0.1527651697,0.0842841491,0.2606301904,-0.101303719,0.2616636157,0.0972530246,-0.2740935087,-0.1220369935,0.0980754793,-0.0550222024,-0.1236003935,0.0130101694,-0.3181710243,-0.19825764,-0.1200942472,0.0065360139,-0.4909358025,0.1610680819,-0.0082066758,0.1992535591,-0.0541196167,-0.2968486249,0.0856656879,-0.1368458569,0.59737885,0.3424452245,0.1117996126,-0.146999687,-0.5365690589,-0.5853412151,0.2090298384,-0.062993072,-0.0486578308,0.0740430281,0.1942421943,0.2390605956,0.1642809063,0.0021003089,-0.1792952865,0.0889608636,-0.0662597269,-0.4647616744,-0.0645498931,-0.1904877871,-0.2589036822,-0.1246322542,0.0241009742,-0.0880486891,0.105777964,0.0488002226,-0.0199225526,-0.2640572786,0.0616629459,-0.0682748631,0.2982192934,-0.0519034304,0.3681990504,-0.1012711897,-0.220356375,-0.2901730239,-0.1561024785,-0.0462986268,0.3026449978,-0.1934378147,0.521027565,-0.0083604716,-0.100206092,-0.3288210034,0.6905267239,0.1733905077,0.1175569594,-0.3923233449,0.0239075776,-0.3116728961,0.1622987688,0.091017738,0.3411852121,-0.2714780271,0.1741360724,-0.4668044746,-0.4499904215,0.2498712838,-0.4550741613,-0.092697911,-0.0041839303,0.1153395548,-0.004739088,-0.0571005195,-0.5571916699,0.0335970558,0.1637207419,0.054309655,-0.0907495096,0.1095089465,-0.0455884449,-0.1208940893,-0.104185313,-0.1784140915,0.1156888828,-0.1695554703,-0.089875564,-0.2151401788]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/407","title":"MissingBeamOptions for Wikipedia 20200501.en","comments":"I found the same issue with almost any language other than English. (For English, it works). Will someone need to update the file on S3 again?","body":"There may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available):\r\n\r\n```\r\nnlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n```\r\n\r\nAnd now, having pulled master, I get:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, total: 34.06 GiB) to \/home\/hltcoe\/mgordon\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/76b0b2747b679bb0ee7a1621e50e5a6378477add0c662668a324a5bc07d516dd...\r\nTraceback (most recent call last):\r\n File \"scripts\/download.py\", line 11, in <module>\r\n fire.Fire(download_pretrain)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 138, in Fire\r\n component_trace = _Fire(component, args, parsed_flag_args, context, name)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 468, in _Fire\r\n target=component.__name__)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 672, in _CallAndUpdateTrace\r\n component = fn(*varargs, **kwargs)\r\n File \"scripts\/download.py\", line 6, in download_pretrain\r\n nlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/load.py\", line 534, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 460, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 870, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, S\r\npark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory).\r\nExample of usage:\r\n `load_dataset('wikipedia', '20200501.en', beam_runner='DirectRunner')`\r\n```","comment_length":26,"text":"MissingBeamOptions for Wikipedia 20200501.en\nThere may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available):\r\n\r\n```\r\nnlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n```\r\n\r\nAnd now, having pulled master, I get:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, total: 34.06 GiB) to \/home\/hltcoe\/mgordon\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/76b0b2747b679bb0ee7a1621e50e5a6378477add0c662668a324a5bc07d516dd...\r\nTraceback (most recent call last):\r\n File \"scripts\/download.py\", line 11, in <module>\r\n fire.Fire(download_pretrain)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 138, in Fire\r\n component_trace = _Fire(component, args, parsed_flag_args, context, name)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 468, in _Fire\r\n target=component.__name__)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 672, in _CallAndUpdateTrace\r\n component = fn(*varargs, **kwargs)\r\n File \"scripts\/download.py\", line 6, in download_pretrain\r\n nlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/load.py\", line 534, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 460, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 870, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, S\r\npark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory).\r\nExample of usage:\r\n `load_dataset('wikipedia', '20200501.en', beam_runner='DirectRunner')`\r\n```\nI found the same issue with almost any language other than English. (For English, it works). Will someone need to update the file on S3 again?","embeddings":[0.0666168258,0.1266757995,-0.0205123089,-0.0542329103,0.1996194869,0.1526870281,0.1799744368,0.3656727076,-0.022208821,-0.081425719,0.1489358097,0.2753814459,-0.0917265415,-0.168423295,0.2434722483,-0.3058640659,0.1455037892,0.1239019409,-0.3501991034,-0.1286828965,-0.0025927718,0.4125896096,-0.2114285231,-0.1782554686,-0.5298822522,0.0631536841,-0.1315196306,0.1428049058,-0.0437919199,-0.2499871403,0.2779204845,0.0593028367,-0.0827241987,0.122238256,-0.0001143081,0.0776108727,0.62224859,-0.2239052504,-0.5686735511,0.0119778551,0.1300001889,-0.3336535096,0.384344399,-0.2550982535,-0.2759362757,-0.1181294397,0.1141192019,-0.0585302375,0.3938059509,-0.0109363226,0.2136250138,0.2360561192,0.297298044,-0.1291413456,0.6020246148,0.0674149916,-0.1662565321,0.0793358013,-0.1638745517,-0.0564770475,-0.2710405588,0.4797945023,0.1387171596,-0.1370008588,0.5526376367,-0.2816212773,0.0226213243,-0.2574208677,0.0174538549,0.1826507896,0.4940163493,-0.0229834914,-0.1063677594,-0.3746719062,0.0446043313,0.1787300706,0.2480826825,0.0172076635,-0.1532892138,-0.0672086403,0.0512106009,-0.2295793146,-0.0332925208,0.28875193,-0.0673940703,0.4202920794,-0.0289063603,0.0214702915,0.166854009,0.0496226139,-0.313372463,0.0365183763,0.0982150361,0.2350160927,-0.2165132016,0.0636979118,0.2868846357,0.2705270946,0.0620143712,0.1695944816,-0.2794000506,0.1912726462,0.0204293542,-0.0470910259,0.2198940963,0.2408223003,-0.0418657698,-0.353320688,0.251672864,0.3364218175,-0.0628019497,0.1384649873,0.1992915124,-0.1387363076,-0.2713562548,0.1334020048,0.0091168443,-0.1717023998,0.1072276235,0.0534083545,-0.2777944505,-0.1483329535,-0.2378234863,0.31988886,-0.0257740859,0.4086242616,0.2018820941,0.1185922623,-0.2835624218,-0.2737653553,-0.1957246959,0.0212342087,-0.4779030085,0.0258821566,0.3871365488,-0.2097051591,0.5165669322,0.0758559778,-0.0362861082,0.0592919849,-0.0301977824,0.2644242346,-0.0725837499,0.181024909,0.1212264374,0.2927809656,0.2036839426,0.0199421402,-0.0530857034,-0.329444617,0.0704844519,-0.2763873637,-0.2597137988,0.1677166075,-0.3119882941,0.1547896415,0.047507599,0.1872284114,-0.0253873691,-0.1822030544,-0.0967266262,-0.1229704767,-0.0582913347,-0.1482739151,0.5051084161,0.4033136666,0.0298880525,-0.3304650486,-0.0621768199,0.1041162759,-0.1011983678,0.0117212692,-0.1564818919,0.5247724652,-0.2857876718,-0.1501885206,0.224312827,0.0519686528,-0.2112267464,0.1486301273,-0.0126577159,0.2708986998,0.0975890458,-0.0528358966,-0.0672804639,0.0518001579,0.0919757187,0.2184601426,0.1131943539,0.0529397614,-0.3270510733,-0.0877883509,-0.0076871128,0.0795486197,0.3505393863,-0.0505552888,0.0635950118,0.5883682966,0.2980857491,-0.1387826651,0.2244143188,0.4500527084,-0.2889430523,0.1368469149,0.2453726381,-0.0598621815,-0.4651874602,0.1208006591,-0.1401829123,0.4367903173,-0.1831934154,-0.00221489,-0.3467137516,-0.1367142648,-0.1968161315,-0.5165299773,0.2065130919,0.2034472376,-0.018576948,0.3046480119,0.0254303049,0.1162029579,-0.040442083,0.0568883121,-0.7796523571,0.214526847,-0.3115582466,-0.1411182284,-0.0192293879,0.3081658483,0.1150461584,-0.1935398728,-0.1284949929,0.2234227657,-0.0789242238,0.1341899037,0.0148870628,-0.0550866164,0.2404268086,-0.3211013675,0.0984831378,0.3122585118,-0.0463057272,-0.1272407025,-0.3034783602,0.2799139619,0.0182238296,0.3226444721,0.033403419,0.0919322968,0.1548368037,0.0725830048,-0.2331330776,-0.202653572,0.3785173595,0.2176063359,0.1366422474,-0.0235250685,0.114962697,0.0045698718,0.1719020158,0.0660793856,0.0112598883,-0.0809159651,-0.4510022402,-0.2482421845,0.3509304225,-0.0857894942,0.0816844404,0.2645476758,0.0524951369,-0.0169572085,-0.1228682548,-0.2095779032,0.1974878311,0.1140880585,0.3243004978,0.0876459852,0.1405431777,0.1975445449,-0.3815246522,0.1906591803,-0.1774683148,0.0871339515,-0.2838542759,-0.0979878008,-0.3540599644,-0.4514617026,-0.1052412391,0.2488120049,-0.3974134922,-0.5874935985,-0.0569580458,0.1890996397,-0.1633748561,0.3057290614,-0.0293643046,-0.1303642839,0.0741929859,0.019791821,-0.3342637122,-0.3922384083,-0.200067386,0.0861517861,0.1043603644,-0.0021498958,0.2901448607,-0.0029742981,-0.0271919109,-0.4731692374,-0.575897634,0.3161930442,-0.2838526368,0.190267697,0.1779287755,0.0881575495,-0.2256871164,-0.058733575,0.4069474936,-0.1701157838,-0.1504672468,-0.0161600616,-0.1277897358,0.101005815,-0.1923410892,-0.2217373699,0.0953757316,-0.3692909479,0.0153438859,0.0921726599,-0.1812787056,0.4691084921,-0.1002497077,-0.0062876246,0.1005272493,0.1722671986,-0.2448800653,0.0835745335,0.3416597843,-0.3003979027,-0.4568088353,0.0295674503,-0.1735530943,0.0571707934,0.318931818,-0.4997883439,-0.0496766046,0.0941225365,0.0024681096,-0.1354268342,0.0942459553,0.5044302344,0.0114218183,-0.0168616381,-0.2326325625,-0.1519111693,-0.1495186388,-0.3148490787,0.4269850552,-0.192369014,0.32359007,-0.0036709276,1.0987776518,-0.0326621234,0.0656037927,0.4224383831,0.093303889,0.4521008432,-0.2759513259,-0.296728462,0.2323563248,-0.1486648619,0.301333636,0.39190045,-0.0147325573,-0.2050670981,-0.365370661,0.009143726,-0.2090857476,-0.3228644729,0.0901104659,-0.3193496764,0.291395694,0.1660908908,0.2335969061,-0.001936554,-0.1577165425,0.2441106439,0.1709194183,0.2070368826,0.0695521533,0.1379282326,-0.056091655,-0.498675853,0.0409604572,0.0189635176,-0.003006133,0.0389499553,-0.1371269971,0.2020363808,0.045385763,0.568950057,-0.1977424026,-0.0263437368,0.2490573078,0.0347809158,-0.7107806206,-0.0275662523,-0.0550747104,0.1947839111,0.2862170339,0.1784339249,-0.3835534155,-0.2234925479,0.3125829697,0.0610993989,-0.0469833761,-0.0261743944,-0.2916185558,0.0066375742,-0.420763731,-0.1309033781,-0.1549676657,0.145596087,0.0774429739,0.168731302,0.0894142538,0.0287991762,-0.1335077137,0.1053509489,0.2781824172,0.2502248585,-0.1114856303,0.1409600228,0.6523587108,-0.3377575278,0.2628805041,0.198228538,0.0455500036,0.2173913866,-0.0400636755,-0.0590368211,0.3987356126,0.0289372355,-0.048762694,-0.203052327,0.0094223302,0.094053559,0.229137376,0.3506721556,-0.2690441608,-0.1362774521,-0.2736847401,0.5236557126,-0.0176693387,-0.0113714179,0.0692942441,0.067014277,-0.3365369737,0.346108377,0.2415182889,0.7413404584,0.0380274691,0.3953185976,0.2077925205,0.3027880788,0.8601722717,-0.3698050082,0.189751476,-0.1742741317,0.2201997638,0.0416868031,0.035432715,-0.1920433491,-0.1310457438,-0.0968420133,0.3844055831,0.3848448396,0.4188909829,0.0760006681,0.2999126911,0.0761646107,-0.3209706545,-0.1950532347,0.1271081567,-0.2076167464,0.5865810513,-0.3765703142,-0.1308689266,-0.2108011395,-0.1166528091,-0.2505488098,0.188552022,-0.2747045159,0.5069428682,-0.1782887429,-0.5959970355,0.0212483089,0.112775512,0.1474469155,0.0883108452,-0.4008170366,0.3684974611,-0.194089815,-0.2362576425,-0.1085617468,0.2097142488,0.4952664971,-0.1708497852,-0.0616780668,-0.0610723421,-0.2416674942,-0.2755625546,-0.2949698865,-0.0613051355,0.0028359615,0.1048691124,-0.0824528262,0.1037253663,-0.0140288854,-0.2572488785,0.1487520188,-0.1144913137,0.0374176167,-0.1460709274,0.102444768,-0.1043076217,-0.0703977495,0.4993968308,-0.0020985689,-0.1289848834,0.5953381062,0.3426305056,-0.1198802143,-0.11467278,-0.0902574286,-0.1393836737,-0.472225517,-0.1296564788,0.0124801882,-0.0530011132,-0.0521329977,0.2447199076,0.1844117939,-0.1626237184,0.0780859217,-0.1943453699,-0.2954605818,0.2547383606,-0.3494720161,-0.0420393012,0.1390184164,0.1013662666,0.3333048522,0.0148150027,-0.2964564562,-0.2154649049,0.0540286303,0.2008091211,0.3083446026,-0.1186889783,-0.0500540733,-0.1276490241,0.1212444231,0.0212709345,-0.1894858032,-0.2302402109,-0.0315677673,0.1042667329,0.0467173308,0.0865679979,-0.078176856,0.1114092842,-0.0371292271,-0.2535363138,0.0528716072,-0.2281834632,0.1240534335,0.1823162884,0.1191580296,0.1763534993,-0.1209528968,0.0150982216,0.0482740849,0.1423756182,0.0532632731,0.0129568297,0.2215794772,-0.0117944498,-0.1690639108,0.1042752936,-0.1854999661,0.2854600251,0.2926937342,-0.0773175657,-0.2990315855,-0.0270367842,0.2942747474,-0.0326080546,-0.0528618991,-0.0485746935,0.0103593031,0.190612644,-0.2196088284,0.267460674,0.5279888511,0.1143720597,0.055378437,-0.0702123344,0.374345392,-0.0246604662,0.3387497365,0.1726608127,0.1217527017,-0.1183099598,0.242642194,0.2166147381,0.3510863185,0.0566907004,0.0239127222,-0.0307099391,0.2304615378,0.2397233546,0.1288086921,0.2786290646,-0.0882287398,0.0975492597,0.0309134126,-0.0629837662,0.1639131606,0.3340352774,0.180387482,0.1440526098,-0.1291939914,0.5687990189,0.1001943797,-0.3893401027,-0.2345822453,0.2402310967,-0.1169962138,0.0740318149,-0.3172836304,-0.1861781478,-0.2391943634,0.1574652493,-0.1074403599,-0.1232673377,-0.0504299775,0.2156560123,-0.1900828779,-0.4732025564,0.2101701349,-0.0363380834,0.3338373005,-0.0730825141,0.1706403196,-0.0616370812,-0.2775771618,-0.0798462331,0.2877354026,0.2786316872,0.1054166332,-0.3909924626,-0.1218862981,0.1556931436,-0.0640750453,-0.1527651697,0.0842841491,0.2606301904,-0.101303719,0.2616636157,0.0972530246,-0.2740935087,-0.1220369935,0.0980754793,-0.0550222024,-0.1236003935,0.0130101694,-0.3181710243,-0.19825764,-0.1200942472,0.0065360139,-0.4909358025,0.1610680819,-0.0082066758,0.1992535591,-0.0541196167,-0.2968486249,0.0856656879,-0.1368458569,0.59737885,0.3424452245,0.1117996126,-0.146999687,-0.5365690589,-0.5853412151,0.2090298384,-0.062993072,-0.0486578308,0.0740430281,0.1942421943,0.2390605956,0.1642809063,0.0021003089,-0.1792952865,0.0889608636,-0.0662597269,-0.4647616744,-0.0645498931,-0.1904877871,-0.2589036822,-0.1246322542,0.0241009742,-0.0880486891,0.105777964,0.0488002226,-0.0199225526,-0.2640572786,0.0616629459,-0.0682748631,0.2982192934,-0.0519034304,0.3681990504,-0.1012711897,-0.220356375,-0.2901730239,-0.1561024785,-0.0462986268,0.3026449978,-0.1934378147,0.521027565,-0.0083604716,-0.100206092,-0.3288210034,0.6905267239,0.1733905077,0.1175569594,-0.3923233449,0.0239075776,-0.3116728961,0.1622987688,0.091017738,0.3411852121,-0.2714780271,0.1741360724,-0.4668044746,-0.4499904215,0.2498712838,-0.4550741613,-0.092697911,-0.0041839303,0.1153395548,-0.004739088,-0.0571005195,-0.5571916699,0.0335970558,0.1637207419,0.054309655,-0.0907495096,0.1095089465,-0.0455884449,-0.1208940893,-0.104185313,-0.1784140915,0.1156888828,-0.1695554703,-0.089875564,-0.2151401788]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/407","title":"MissingBeamOptions for Wikipedia 20200501.en","comments":"This is because only some languages are already preprocessed (en, de, fr, it) and stored on our google storage.\r\nWe plan to have a systematic way to preprocess more wikipedia languages in the future.\r\n\r\nFor the other languages you have to process them on your side using apache beam. That's why the lib asks for a Beam runner.","body":"There may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available):\r\n\r\n```\r\nnlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n```\r\n\r\nAnd now, having pulled master, I get:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, total: 34.06 GiB) to \/home\/hltcoe\/mgordon\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/76b0b2747b679bb0ee7a1621e50e5a6378477add0c662668a324a5bc07d516dd...\r\nTraceback (most recent call last):\r\n File \"scripts\/download.py\", line 11, in <module>\r\n fire.Fire(download_pretrain)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 138, in Fire\r\n component_trace = _Fire(component, args, parsed_flag_args, context, name)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 468, in _Fire\r\n target=component.__name__)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 672, in _CallAndUpdateTrace\r\n component = fn(*varargs, **kwargs)\r\n File \"scripts\/download.py\", line 6, in download_pretrain\r\n nlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/load.py\", line 534, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 460, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 870, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, S\r\npark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory).\r\nExample of usage:\r\n `load_dataset('wikipedia', '20200501.en', beam_runner='DirectRunner')`\r\n```","comment_length":58,"text":"MissingBeamOptions for Wikipedia 20200501.en\nThere may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available):\r\n\r\n```\r\nnlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n```\r\n\r\nAnd now, having pulled master, I get:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, total: 34.06 GiB) to \/home\/hltcoe\/mgordon\/.cache\/huggingface\/datasets\/wikipedia\/20200501.en\/1.0.0\/76b0b2747b679bb0ee7a1621e50e5a6378477add0c662668a324a5bc07d516dd...\r\nTraceback (most recent call last):\r\n File \"scripts\/download.py\", line 11, in <module>\r\n fire.Fire(download_pretrain)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 138, in Fire\r\n component_trace = _Fire(component, args, parsed_flag_args, context, name)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 468, in _Fire\r\n target=component.__name__)\r\n File \"\/home\/hltcoe\/mgordon\/.conda\/envs\/huggingface\/lib\/python3.6\/site-packages\/fire\/core.py\", line 672, in _CallAndUpdateTrace\r\n component = fn(*varargs, **kwargs)\r\n File \"scripts\/download.py\", line 6, in download_pretrain\r\n nlp.load_dataset('wikipedia', \"20200501.en\", split='train')\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/load.py\", line 534, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 460, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/exp\/mgordon\/nlp\/src\/nlp\/builder.py\", line 870, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, S\r\npark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory).\r\nExample of usage:\r\n `load_dataset('wikipedia', '20200501.en', beam_runner='DirectRunner')`\r\n```\nThis is because only some languages are already preprocessed (en, de, fr, it) and stored on our google storage.\r\nWe plan to have a systematic way to preprocess more wikipedia languages in the future.\r\n\r\nFor the other languages you have to process them on your side using apache beam. That's why the lib asks for a Beam runner.","embeddings":[0.0666168258,0.1266757995,-0.0205123089,-0.0542329103,0.1996194869,0.1526870281,0.1799744368,0.3656727076,-0.022208821,-0.081425719,0.1489358097,0.2753814459,-0.0917265415,-0.168423295,0.2434722483,-0.3058640659,0.1455037892,0.1239019409,-0.3501991034,-0.1286828965,-0.0025927718,0.4125896096,-0.2114285231,-0.1782554686,-0.5298822522,0.0631536841,-0.1315196306,0.1428049058,-0.0437919199,-0.2499871403,0.2779204845,0.0593028367,-0.0827241987,0.122238256,-0.0001143081,0.0776108727,0.62224859,-0.2239052504,-0.5686735511,0.0119778551,0.1300001889,-0.3336535096,0.384344399,-0.2550982535,-0.2759362757,-0.1181294397,0.1141192019,-0.0585302375,0.3938059509,-0.0109363226,0.2136250138,0.2360561192,0.297298044,-0.1291413456,0.6020246148,0.0674149916,-0.1662565321,0.0793358013,-0.1638745517,-0.0564770475,-0.2710405588,0.4797945023,0.1387171596,-0.1370008588,0.5526376367,-0.2816212773,0.0226213243,-0.2574208677,0.0174538549,0.1826507896,0.4940163493,-0.0229834914,-0.1063677594,-0.3746719062,0.0446043313,0.1787300706,0.2480826825,0.0172076635,-0.1532892138,-0.0672086403,0.0512106009,-0.2295793146,-0.0332925208,0.28875193,-0.0673940703,0.4202920794,-0.0289063603,0.0214702915,0.166854009,0.0496226139,-0.313372463,0.0365183763,0.0982150361,0.2350160927,-0.2165132016,0.0636979118,0.2868846357,0.2705270946,0.0620143712,0.1695944816,-0.2794000506,0.1912726462,0.0204293542,-0.0470910259,0.2198940963,0.2408223003,-0.0418657698,-0.353320688,0.251672864,0.3364218175,-0.0628019497,0.1384649873,0.1992915124,-0.1387363076,-0.2713562548,0.1334020048,0.0091168443,-0.1717023998,0.1072276235,0.0534083545,-0.2777944505,-0.1483329535,-0.2378234863,0.31988886,-0.0257740859,0.4086242616,0.2018820941,0.1185922623,-0.2835624218,-0.2737653553,-0.1957246959,0.0212342087,-0.4779030085,0.0258821566,0.3871365488,-0.2097051591,0.5165669322,0.0758559778,-0.0362861082,0.0592919849,-0.0301977824,0.2644242346,-0.0725837499,0.181024909,0.1212264374,0.2927809656,0.2036839426,0.0199421402,-0.0530857034,-0.329444617,0.0704844519,-0.2763873637,-0.2597137988,0.1677166075,-0.3119882941,0.1547896415,0.047507599,0.1872284114,-0.0253873691,-0.1822030544,-0.0967266262,-0.1229704767,-0.0582913347,-0.1482739151,0.5051084161,0.4033136666,0.0298880525,-0.3304650486,-0.0621768199,0.1041162759,-0.1011983678,0.0117212692,-0.1564818919,0.5247724652,-0.2857876718,-0.1501885206,0.224312827,0.0519686528,-0.2112267464,0.1486301273,-0.0126577159,0.2708986998,0.0975890458,-0.0528358966,-0.0672804639,0.0518001579,0.0919757187,0.2184601426,0.1131943539,0.0529397614,-0.3270510733,-0.0877883509,-0.0076871128,0.0795486197,0.3505393863,-0.0505552888,0.0635950118,0.5883682966,0.2980857491,-0.1387826651,0.2244143188,0.4500527084,-0.2889430523,0.1368469149,0.2453726381,-0.0598621815,-0.4651874602,0.1208006591,-0.1401829123,0.4367903173,-0.1831934154,-0.00221489,-0.3467137516,-0.1367142648,-0.1968161315,-0.5165299773,0.2065130919,0.2034472376,-0.018576948,0.3046480119,0.0254303049,0.1162029579,-0.040442083,0.0568883121,-0.7796523571,0.214526847,-0.3115582466,-0.1411182284,-0.0192293879,0.3081658483,0.1150461584,-0.1935398728,-0.1284949929,0.2234227657,-0.0789242238,0.1341899037,0.0148870628,-0.0550866164,0.2404268086,-0.3211013675,0.0984831378,0.3122585118,-0.0463057272,-0.1272407025,-0.3034783602,0.2799139619,0.0182238296,0.3226444721,0.033403419,0.0919322968,0.1548368037,0.0725830048,-0.2331330776,-0.202653572,0.3785173595,0.2176063359,0.1366422474,-0.0235250685,0.114962697,0.0045698718,0.1719020158,0.0660793856,0.0112598883,-0.0809159651,-0.4510022402,-0.2482421845,0.3509304225,-0.0857894942,0.0816844404,0.2645476758,0.0524951369,-0.0169572085,-0.1228682548,-0.2095779032,0.1974878311,0.1140880585,0.3243004978,0.0876459852,0.1405431777,0.1975445449,-0.3815246522,0.1906591803,-0.1774683148,0.0871339515,-0.2838542759,-0.0979878008,-0.3540599644,-0.4514617026,-0.1052412391,0.2488120049,-0.3974134922,-0.5874935985,-0.0569580458,0.1890996397,-0.1633748561,0.3057290614,-0.0293643046,-0.1303642839,0.0741929859,0.019791821,-0.3342637122,-0.3922384083,-0.200067386,0.0861517861,0.1043603644,-0.0021498958,0.2901448607,-0.0029742981,-0.0271919109,-0.4731692374,-0.575897634,0.3161930442,-0.2838526368,0.190267697,0.1779287755,0.0881575495,-0.2256871164,-0.058733575,0.4069474936,-0.1701157838,-0.1504672468,-0.0161600616,-0.1277897358,0.101005815,-0.1923410892,-0.2217373699,0.0953757316,-0.3692909479,0.0153438859,0.0921726599,-0.1812787056,0.4691084921,-0.1002497077,-0.0062876246,0.1005272493,0.1722671986,-0.2448800653,0.0835745335,0.3416597843,-0.3003979027,-0.4568088353,0.0295674503,-0.1735530943,0.0571707934,0.318931818,-0.4997883439,-0.0496766046,0.0941225365,0.0024681096,-0.1354268342,0.0942459553,0.5044302344,0.0114218183,-0.0168616381,-0.2326325625,-0.1519111693,-0.1495186388,-0.3148490787,0.4269850552,-0.192369014,0.32359007,-0.0036709276,1.0987776518,-0.0326621234,0.0656037927,0.4224383831,0.093303889,0.4521008432,-0.2759513259,-0.296728462,0.2323563248,-0.1486648619,0.301333636,0.39190045,-0.0147325573,-0.2050670981,-0.365370661,0.009143726,-0.2090857476,-0.3228644729,0.0901104659,-0.3193496764,0.291395694,0.1660908908,0.2335969061,-0.001936554,-0.1577165425,0.2441106439,0.1709194183,0.2070368826,0.0695521533,0.1379282326,-0.056091655,-0.498675853,0.0409604572,0.0189635176,-0.003006133,0.0389499553,-0.1371269971,0.2020363808,0.045385763,0.568950057,-0.1977424026,-0.0263437368,0.2490573078,0.0347809158,-0.7107806206,-0.0275662523,-0.0550747104,0.1947839111,0.2862170339,0.1784339249,-0.3835534155,-0.2234925479,0.3125829697,0.0610993989,-0.0469833761,-0.0261743944,-0.2916185558,0.0066375742,-0.420763731,-0.1309033781,-0.1549676657,0.145596087,0.0774429739,0.168731302,0.0894142538,0.0287991762,-0.1335077137,0.1053509489,0.2781824172,0.2502248585,-0.1114856303,0.1409600228,0.6523587108,-0.3377575278,0.2628805041,0.198228538,0.0455500036,0.2173913866,-0.0400636755,-0.0590368211,0.3987356126,0.0289372355,-0.048762694,-0.203052327,0.0094223302,0.094053559,0.229137376,0.3506721556,-0.2690441608,-0.1362774521,-0.2736847401,0.5236557126,-0.0176693387,-0.0113714179,0.0692942441,0.067014277,-0.3365369737,0.346108377,0.2415182889,0.7413404584,0.0380274691,0.3953185976,0.2077925205,0.3027880788,0.8601722717,-0.3698050082,0.189751476,-0.1742741317,0.2201997638,0.0416868031,0.035432715,-0.1920433491,-0.1310457438,-0.0968420133,0.3844055831,0.3848448396,0.4188909829,0.0760006681,0.2999126911,0.0761646107,-0.3209706545,-0.1950532347,0.1271081567,-0.2076167464,0.5865810513,-0.3765703142,-0.1308689266,-0.2108011395,-0.1166528091,-0.2505488098,0.188552022,-0.2747045159,0.5069428682,-0.1782887429,-0.5959970355,0.0212483089,0.112775512,0.1474469155,0.0883108452,-0.4008170366,0.3684974611,-0.194089815,-0.2362576425,-0.1085617468,0.2097142488,0.4952664971,-0.1708497852,-0.0616780668,-0.0610723421,-0.2416674942,-0.2755625546,-0.2949698865,-0.0613051355,0.0028359615,0.1048691124,-0.0824528262,0.1037253663,-0.0140288854,-0.2572488785,0.1487520188,-0.1144913137,0.0374176167,-0.1460709274,0.102444768,-0.1043076217,-0.0703977495,0.4993968308,-0.0020985689,-0.1289848834,0.5953381062,0.3426305056,-0.1198802143,-0.11467278,-0.0902574286,-0.1393836737,-0.472225517,-0.1296564788,0.0124801882,-0.0530011132,-0.0521329977,0.2447199076,0.1844117939,-0.1626237184,0.0780859217,-0.1943453699,-0.2954605818,0.2547383606,-0.3494720161,-0.0420393012,0.1390184164,0.1013662666,0.3333048522,0.0148150027,-0.2964564562,-0.2154649049,0.0540286303,0.2008091211,0.3083446026,-0.1186889783,-0.0500540733,-0.1276490241,0.1212444231,0.0212709345,-0.1894858032,-0.2302402109,-0.0315677673,0.1042667329,0.0467173308,0.0865679979,-0.078176856,0.1114092842,-0.0371292271,-0.2535363138,0.0528716072,-0.2281834632,0.1240534335,0.1823162884,0.1191580296,0.1763534993,-0.1209528968,0.0150982216,0.0482740849,0.1423756182,0.0532632731,0.0129568297,0.2215794772,-0.0117944498,-0.1690639108,0.1042752936,-0.1854999661,0.2854600251,0.2926937342,-0.0773175657,-0.2990315855,-0.0270367842,0.2942747474,-0.0326080546,-0.0528618991,-0.0485746935,0.0103593031,0.190612644,-0.2196088284,0.267460674,0.5279888511,0.1143720597,0.055378437,-0.0702123344,0.374345392,-0.0246604662,0.3387497365,0.1726608127,0.1217527017,-0.1183099598,0.242642194,0.2166147381,0.3510863185,0.0566907004,0.0239127222,-0.0307099391,0.2304615378,0.2397233546,0.1288086921,0.2786290646,-0.0882287398,0.0975492597,0.0309134126,-0.0629837662,0.1639131606,0.3340352774,0.180387482,0.1440526098,-0.1291939914,0.5687990189,0.1001943797,-0.3893401027,-0.2345822453,0.2402310967,-0.1169962138,0.0740318149,-0.3172836304,-0.1861781478,-0.2391943634,0.1574652493,-0.1074403599,-0.1232673377,-0.0504299775,0.2156560123,-0.1900828779,-0.4732025564,0.2101701349,-0.0363380834,0.3338373005,-0.0730825141,0.1706403196,-0.0616370812,-0.2775771618,-0.0798462331,0.2877354026,0.2786316872,0.1054166332,-0.3909924626,-0.1218862981,0.1556931436,-0.0640750453,-0.1527651697,0.0842841491,0.2606301904,-0.101303719,0.2616636157,0.0972530246,-0.2740935087,-0.1220369935,0.0980754793,-0.0550222024,-0.1236003935,0.0130101694,-0.3181710243,-0.19825764,-0.1200942472,0.0065360139,-0.4909358025,0.1610680819,-0.0082066758,0.1992535591,-0.0541196167,-0.2968486249,0.0856656879,-0.1368458569,0.59737885,0.3424452245,0.1117996126,-0.146999687,-0.5365690589,-0.5853412151,0.2090298384,-0.062993072,-0.0486578308,0.0740430281,0.1942421943,0.2390605956,0.1642809063,0.0021003089,-0.1792952865,0.0889608636,-0.0662597269,-0.4647616744,-0.0645498931,-0.1904877871,-0.2589036822,-0.1246322542,0.0241009742,-0.0880486891,0.105777964,0.0488002226,-0.0199225526,-0.2640572786,0.0616629459,-0.0682748631,0.2982192934,-0.0519034304,0.3681990504,-0.1012711897,-0.220356375,-0.2901730239,-0.1561024785,-0.0462986268,0.3026449978,-0.1934378147,0.521027565,-0.0083604716,-0.100206092,-0.3288210034,0.6905267239,0.1733905077,0.1175569594,-0.3923233449,0.0239075776,-0.3116728961,0.1622987688,0.091017738,0.3411852121,-0.2714780271,0.1741360724,-0.4668044746,-0.4499904215,0.2498712838,-0.4550741613,-0.092697911,-0.0041839303,0.1153395548,-0.004739088,-0.0571005195,-0.5571916699,0.0335970558,0.1637207419,0.054309655,-0.0907495096,0.1095089465,-0.0455884449,-0.1208940893,-0.104185313,-0.1784140915,0.1156888828,-0.1695554703,-0.089875564,-0.2151401788]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/406","title":"Faster Shuffling?","comments":"I think the slowness here probably come from the fact that we are copying from and to python.\r\n\r\n@lhoestq for all the `select`-based methods I think we should stay in Arrow format and update the writer so that it can accept Arrow tables or batches as well. What do you think?","body":"Consider shuffling bookcorpus:\r\n\r\n```\r\ndataset = nlp.load_dataset('bookcorpus', split='train')\r\ndataset.shuffle()\r\n```\r\nAccording to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`.\r\n\r\nBut I can also just write the lines to a text file:\r\n\r\n```\r\nbatch_size = 100000\r\nwith open('tmp.txt', 'w+') as out_f:\r\n for i in tqdm(range(0, len(dataset), batch_size)):\r\n batch = dataset[i:i+batch_size]['text']\r\n print(\"\\n\".join(batch), file=out_f)\r\n```\r\n\r\nWhich completes in a couple minutes, followed by `shuf tmp.txt > tmp2.txt` which completes in under a minute. And finally,\r\n\r\n```\r\ndataset = nlp.load_dataset('text', data_files='tmp2.txt')\r\n```\r\n\r\nWhich completes in under 10 minutes. I read up on Apache Arrow this morning, and it seems like the columnar data format is not especially well-suited to shuffling rows, since moving items around requires a lot of book-keeping. \r\n\r\nIs shuffle inherently slow, or am I just using it wrong? And if it is slow, would it make sense to try converting the data to a row-based format on disk and then shuffling? (Instead of calling select with a random permutation, as is currently done.)","comment_length":51,"text":"Faster Shuffling?\nConsider shuffling bookcorpus:\r\n\r\n```\r\ndataset = nlp.load_dataset('bookcorpus', split='train')\r\ndataset.shuffle()\r\n```\r\nAccording to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`.\r\n\r\nBut I can also just write the lines to a text file:\r\n\r\n```\r\nbatch_size = 100000\r\nwith open('tmp.txt', 'w+') as out_f:\r\n for i in tqdm(range(0, len(dataset), batch_size)):\r\n batch = dataset[i:i+batch_size]['text']\r\n print(\"\\n\".join(batch), file=out_f)\r\n```\r\n\r\nWhich completes in a couple minutes, followed by `shuf tmp.txt > tmp2.txt` which completes in under a minute. And finally,\r\n\r\n```\r\ndataset = nlp.load_dataset('text', data_files='tmp2.txt')\r\n```\r\n\r\nWhich completes in under 10 minutes. I read up on Apache Arrow this morning, and it seems like the columnar data format is not especially well-suited to shuffling rows, since moving items around requires a lot of book-keeping. \r\n\r\nIs shuffle inherently slow, or am I just using it wrong? And if it is slow, would it make sense to try converting the data to a row-based format on disk and then shuffling? (Instead of calling select with a random permutation, as is currently done.)\nI think the slowness here probably come from the fact that we are copying from and to python.\r\n\r\n@lhoestq for all the `select`-based methods I think we should stay in Arrow format and update the writer so that it can accept Arrow tables or batches as well. What do you think?","embeddings":[-0.1038477495,0.0485302918,-0.0324605815,-0.1238688901,-0.078709498,-0.1393203437,0.1168373376,0.5494186878,-0.2944583595,0.0831830278,-0.0154477637,0.530700922,-0.1336493641,-0.261156857,0.1678622067,-0.0926039442,0.0606321879,-0.0731456354,-0.2566205859,-0.0911279842,-0.1002621353,-0.3960591853,-0.2107811421,-0.2046841085,-0.0074308575,-0.2436168194,-0.2369048744,0.1180206612,-0.1844848245,0.0845983326,-0.1934022754,0.2316126525,-0.1237936988,0.2456135005,-0.0001148454,-0.1855444014,0.0773180798,0.1337084025,-0.3280711174,0.1574876904,0.4777601063,-0.0623344183,-0.1090694144,-0.2044636607,0.1624595076,-0.3435192406,0.1161517948,-0.0758223012,-0.0899132639,-0.0375098065,0.1314271837,0.3801160455,-0.1988957077,0.220427081,0.4817522466,0.2134924829,-0.35807693,0.3437612355,0.3956694305,-0.0749704242,-0.1464883834,0.2646939158,-0.1878203601,-0.0394971371,0.207262978,0.134258002,0.1285514385,0.3230400085,0.1072722375,0.3726660013,0.3013318181,-0.2000201046,-0.3548533916,-0.2874730527,0.203556478,-0.030812392,0.0791790113,0.2857201099,-0.2808455527,-0.1876985878,-0.1240445226,0.104160592,0.2330206931,-0.2302716225,0.3201219738,0.417814225,0.2373136878,-0.0998314694,0.6690923572,-0.0953145623,0.5340384245,-0.1793558896,0.070103839,0.4054663777,-0.287830919,-0.2892364562,-0.1990911961,-0.0882906094,-0.1359459311,0.1803567261,0.669680953,0.2038075179,0.2009965777,-0.0543011837,0.082189694,-0.0159299355,-0.3168695271,0.1569919735,-0.0950163826,0.08423765,0.257474333,0.0918948352,-0.0717737079,-0.3007595837,-0.0448231511,-0.2145741284,-0.4414107203,0.1036592349,0.0323504582,-0.2918610275,-0.0960134044,-0.0679451972,0.0125763277,0.2145961076,-0.1745274365,0.2548381984,-0.0650107861,-0.1067523137,-0.1531518549,0.0528996103,-0.0397013053,0.1025035307,-0.0759030655,0.001008702,0.0193844587,0.3855785131,-0.2566693723,0.0715126395,0.198361218,0.3333364427,0.0390317068,-0.1371054351,0.0811998993,0.1164640933,-0.1295931041,-0.1626015753,-0.1090415865,0.056350112,-0.3282532692,0.3956876993,-0.0282882564,-0.1497776657,0.1397400051,0.1568017751,0.2815747559,-0.2844125032,-0.4706712365,-0.0742615908,0.0032804434,-0.0329053663,0.0783743635,-0.1077171117,0.1294034868,-0.1461091936,0.0069920802,-0.2068298161,-0.0313831307,0.1186726764,0.0181323458,0.5336561799,0.4336855114,0.6766946316,-0.083565034,0.0888084099,0.1313965321,-0.0325662717,0.7245671749,0.175294742,-0.1763671339,0.0303561576,-0.0777412355,0.0723708272,0.4168796837,0.1878635138,0.0499962233,-0.14467071,0.4556990862,0.2858871818,0.0005487598,0.2795182467,-0.5462657809,-0.1266948283,0.2373047322,0.4326201081,-0.499161005,-0.3179177642,-0.0106280642,-0.3204751015,0.3234396875,-0.2063714862,-0.1109222397,0.0519927666,-0.0280231498,0.0888931453,-0.1518577039,0.0710438192,-0.1989396811,-0.0571107827,-0.0291539915,-0.0216976348,0.2426874191,-0.3557789922,0.1991183609,-0.1732147187,0.119469434,0.1699107736,-0.0024224098,-0.2069241703,0.3986496329,-0.1000551656,-0.0798934922,-0.0160999931,-0.2569810152,0.0842135325,-0.1689046919,-0.1116963401,0.0451900475,-0.2143689543,0.1580178738,0.3718034625,-0.2997277379,0.0496766642,0.1027156636,0.214271009,-0.0766529739,-0.1232783645,0.0910351723,0.1982082427,-0.0637715831,0.257558763,0.1842211038,-0.1660939157,0.1514940262,-0.0345027223,-0.3289418519,0.534730494,-0.4930259287,0.3141444921,0.1985875219,-0.2543620169,-0.140823707,0.2990074456,-0.0973584726,0.4259059429,0.6466958523,0.1116910577,0.1135969535,0.4000864327,-0.3382675946,0.0998384133,0.5146846175,-0.0320696644,-0.1510275751,0.1014408544,0.0868231952,-0.1578638852,-0.1339073926,0.0724576041,0.2643631101,0.2578005791,0.1068165824,-0.0220652316,-0.0651489496,-0.2313472778,0.2150713354,-0.0015312025,0.0961966291,-0.1439817995,0.1075777113,-0.0137522351,-0.1625159234,-0.1773954779,0.0197073761,0.163955912,0.0767694637,0.0736185014,-0.0629530102,-0.1240025759,-0.1818030775,-0.0365182571,-0.0409475677,-0.1341001391,0.1759616435,0.2394411713,-0.229558602,-0.2709872723,-0.1023018509,0.2744619548,-0.1486407816,-0.1649528146,-0.1873629093,0.0778129622,-0.1182080805,0.0518480204,0.2363512069,0.2894334495,0.3592737317,0.0621871911,-0.2367981374,0.0518280044,0.0315158628,-0.2365831137,0.024727948,-0.2823906243,-0.0512630306,-0.1843814403,-0.3590669632,-0.3232360184,0.0791548342,-0.1600939333,-0.100713633,0.3606928885,0.2180092782,-0.2246398479,-0.1231466159,-0.070720382,-0.1353553385,-0.3641655445,0.2947729826,-0.0735338554,0.1561614126,-0.0280660987,-0.1540985107,-0.0747499615,-0.4670978189,-0.0291595124,0.1779473573,-0.1603415906,0.2665257752,0.0757241547,-0.1876378804,-0.4554571509,-0.0309031475,-0.0455955602,0.4398033917,-0.0504291095,0.1585071087,-0.4090271294,-0.0185445659,-0.0691038892,0.2417527437,0.4783862233,-0.1012297198,-0.113726072,0.3343468904,0.190007925,-0.0072763637,0.281583935,0.0939925388,0.0873351619,0.1453133076,0.1917298436,0.5485806465,-0.0089885378,0.4024249911,-0.1678857952,0.1251711547,-0.0263603814,-0.3092458546,0.0191647112,0.154581815,-0.0148901921,-0.2202124,0.2860586643,-0.1322850436,-0.216705814,-0.1457933187,0.442979604,0.0978032202,-0.3071257174,0.5698950887,-0.2439899892,0.011495227,0.0262768287,0.242693156,-0.1146337539,-0.1829121709,0.0393732414,-0.0735249221,-0.0736508071,-0.0306380913,-0.6124544144,-0.1522423327,-0.8053997159,0.2165995538,0.2030180991,0.3697021902,0.26963678,-0.320687741,0.4373373091,-0.1314526647,0.2683980465,0.1939319819,-0.3236965239,0.2640224993,-0.1401326209,-0.4568900168,-0.0840428248,-0.3802316785,-0.4031210542,0.1451105475,0.1813238412,-0.3581776321,-0.0427722,-0.0153322555,-0.0565103516,-0.0170648601,0.0129579678,-0.3901495934,-0.2344471812,0.0323277116,0.2596327066,-0.0461357683,0.0254317485,-0.0565665662,0.2224269658,0.15895468,0.0007253136,0.1433291137,-0.1042865813,-0.0540571809,-0.0540811494,0.158406198,0.1658077687,0.0621211417,-0.0909933969,0.0022126515,0.1231649965,0.0416770875,-0.1576361358,-0.1893505454,0.4593550265,0.3867465556,0.2079528123,0.2129657716,-0.0241839197,0.3218667507,-0.0059941658,-0.5208618641,0.1308053285,-0.2954317629,-0.1013388187,-0.7717430592,0.2592333555,0.254368335,0.1491476446,0.2281627655,-0.1963314563,0.2470769435,0.4275337458,0.1410053372,1.0107350349,-0.3745579123,0.4948277771,0.3026477695,-0.2127191871,0.0632101074,-0.1288544089,0.3773986697,-0.1711562723,-0.1408917606,0.2182426751,-0.1835123748,0.0028821838,0.0469279699,-0.0376596376,0.2010120898,-0.0023710043,-0.3239124119,-0.3013598323,-0.0370964482,-0.4331661463,-0.3865849078,-0.1620003879,0.0026144644,-0.0321729071,-0.1597066969,0.1417668164,0.081244953,-0.4642912745,0.1138687581,-0.1251077652,-0.012902528,0.0488075726,-0.0064567206,0.2049918473,-0.2861238718,-0.0617328174,-0.1386677772,-0.1246639192,0.1100128219,0.1806344539,0.1125110537,0.2573269904,0.3945131004,0.3727914095,-0.2835698426,0.287743181,-0.063258864,0.250033915,0.0561996177,-0.0084810769,-0.7355349064,-0.3018519282,0.0860433877,0.5681176782,-0.0286947712,-0.140967086,0.2391030937,-0.3686720729,-0.0958732739,0.1014309302,0.057384342,0.1053926796,0.4705395699,0.1320168525,-0.188612029,-0.0110093188,0.2801900208,0.1060088128,0.139658168,0.3078068495,0.0161026884,-0.4802675843,-0.2216252834,0.181569308,0.0415754952,-0.34996894,0.2590771914,0.0156889539,-0.0785714835,0.1282119453,0.2876173854,-0.0295777675,-0.0591709167,-0.3001861274,-0.014993214,0.1073984206,0.4669550657,0.5261853337,0.0099294046,0.0050599789,0.3379234076,-0.3852916956,-0.150651142,-0.2425634563,-0.1853693128,0.0913322344,0.1223526224,-0.1389974654,-0.2877808511,-0.1926407516,0.1505212486,0.1307436824,0.1484171152,0.0329903439,-0.1907644272,0.0702230856,0.0341222472,0.0276608076,-0.0140173575,0.0477781706,-0.3899277747,-0.3449161649,-0.2392159402,-0.2736591995,0.0012196295,-0.0083562499,-0.2712943852,0.2088960707,0.1324304342,-0.0776069388,0.5249646902,-0.4731840789,-0.2732441425,-0.220849365,-0.006398737,-0.0190975238,-0.0999407694,-0.1110328436,0.1236861572,0.0560273826,0.0054031769,0.2663468122,-0.1399292797,0.0574536361,-0.1029636264,0.1943242848,0.635150969,0.2898965776,-0.4409421086,0.4520812929,0.1157830656,-0.1129201874,-0.2333065867,-0.354419589,0.214579165,-0.0045310077,0.102565296,-0.2718020976,-0.1124611795,-0.3286367059,-0.0594900995,-0.1369457692,-0.4206710458,-0.2316709161,0.3065519333,0.3698605299,-0.2523021996,0.1801477373,-0.0040579252,-0.0467773005,0.1294298768,0.2270887941,0.4077351391,0.6244723797,0.3014619648,-0.3874140084,-0.1106976047,-0.0040118839,0.4378830791,0.1801593602,0.3086165488,0.3406611383,0.1345147341,0.006603586,0.0796901807,-0.0219778884,-0.0928382948,-0.0370790735,0.0731732398,0.0725146458,-0.0947212353,-0.2181289345,0.2042514384,-0.3194910884,0.3458896279,0.2264336348,0.1036254391,0.4627437592,-0.3930014372,0.1452660263,-0.2250460386,0.4474727511,-0.3134241998,-0.1612286717,-0.037501283,0.1031903774,0.2538506091,0.0662699789,-0.0725488067,0.272868067,-0.079048112,0.0622294396,-0.5087035894,-0.0149668874,-0.1064291,0.2163717002,0.4450425804,-0.0854283497,0.279548198,0.0122327469,-0.1925521791,-0.1656651348,-0.0980604365,-0.0690087974,-0.1779288948,0.1612359434,0.2084825933,-0.4335984588,-0.0445253477,0.1502262503,-0.2526210546,-0.0961487591,-0.5036643147,-0.2149572074,-0.1091824397,0.0238641426,0.0551286079,0.190128535,0.6890671253,0.4244920313,-0.0805729255,0.1032524556,-0.2726463377,-0.2619601786,-0.3324976861,-0.2798487544,-0.0097650504,0.2360472381,0.4841252565,0.2455261201,0.2714175284,-0.0475976057,0.151284501,-0.3728962243,0.3568670452,-0.1409450322,0.207420975,0.1653525829,-0.0630150139,-0.2124655098,-0.6004893184,0.2104823291,0.1409494728,-0.002743091,-0.0707928687,0.0825923905,0.2068638504,0.3266091347,-0.169467181,-0.0374941304,-0.1736766696,0.2170685977,-0.2059338093,-0.1906139553,0.1858600676,-0.022793036,0.0193882063,0.0742552578,0.1204371378,0.0634020343,0.0078346701,0.1306756288,0.5502994061,-0.2406554073,0.1733045876,-0.2645807266,0.0541518368,0.2109876722,0.2993127704,-0.1580787897,0.142650187,-0.0488657169,-0.1984006017,-0.2610489726,0.0039233067,0.3909659088,-0.1595080346,-0.2572924197,-0.1266072094,0.1781403422,0.2004371434,0.3141005635,0.0848385692,-0.2719454169,-0.042041786,-0.0028391564,0.3295638859,0.269949615,-0.1165978536,-0.0616496615,-0.1546361446,0.3200790882,-0.4524775445,0.0876661092,-0.2333867252,-0.3020479977]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/406","title":"Faster Shuffling?","comments":"> @lhoestq for all the `select`-based methods I think we should stay in Arrow format and update the writer so that it can accept Arrow tables or batches as well. What do you think?\r\n\r\nI just tried with `writer.write_table` with tables of 1000 elements and it's slower that the solution in #405 \r\n\r\nOn my side (select 10 000 examples):\r\n- Original implementation: 12s\r\n- Batched solution: 100ms\r\n- solution using arrow tables: 350ms\r\n\r\nI'll try with arrays and record batches to see if we can make it work.","body":"Consider shuffling bookcorpus:\r\n\r\n```\r\ndataset = nlp.load_dataset('bookcorpus', split='train')\r\ndataset.shuffle()\r\n```\r\nAccording to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`.\r\n\r\nBut I can also just write the lines to a text file:\r\n\r\n```\r\nbatch_size = 100000\r\nwith open('tmp.txt', 'w+') as out_f:\r\n for i in tqdm(range(0, len(dataset), batch_size)):\r\n batch = dataset[i:i+batch_size]['text']\r\n print(\"\\n\".join(batch), file=out_f)\r\n```\r\n\r\nWhich completes in a couple minutes, followed by `shuf tmp.txt > tmp2.txt` which completes in under a minute. And finally,\r\n\r\n```\r\ndataset = nlp.load_dataset('text', data_files='tmp2.txt')\r\n```\r\n\r\nWhich completes in under 10 minutes. I read up on Apache Arrow this morning, and it seems like the columnar data format is not especially well-suited to shuffling rows, since moving items around requires a lot of book-keeping. \r\n\r\nIs shuffle inherently slow, or am I just using it wrong? And if it is slow, would it make sense to try converting the data to a row-based format on disk and then shuffling? (Instead of calling select with a random permutation, as is currently done.)","comment_length":88,"text":"Faster Shuffling?\nConsider shuffling bookcorpus:\r\n\r\n```\r\ndataset = nlp.load_dataset('bookcorpus', split='train')\r\ndataset.shuffle()\r\n```\r\nAccording to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`.\r\n\r\nBut I can also just write the lines to a text file:\r\n\r\n```\r\nbatch_size = 100000\r\nwith open('tmp.txt', 'w+') as out_f:\r\n for i in tqdm(range(0, len(dataset), batch_size)):\r\n batch = dataset[i:i+batch_size]['text']\r\n print(\"\\n\".join(batch), file=out_f)\r\n```\r\n\r\nWhich completes in a couple minutes, followed by `shuf tmp.txt > tmp2.txt` which completes in under a minute. And finally,\r\n\r\n```\r\ndataset = nlp.load_dataset('text', data_files='tmp2.txt')\r\n```\r\n\r\nWhich completes in under 10 minutes. I read up on Apache Arrow this morning, and it seems like the columnar data format is not especially well-suited to shuffling rows, since moving items around requires a lot of book-keeping. \r\n\r\nIs shuffle inherently slow, or am I just using it wrong? And if it is slow, would it make sense to try converting the data to a row-based format on disk and then shuffling? (Instead of calling select with a random permutation, as is currently done.)\n> @lhoestq for all the `select`-based methods I think we should stay in Arrow format and update the writer so that it can accept Arrow tables or batches as well. What do you think?\r\n\r\nI just tried with `writer.write_table` with tables of 1000 elements and it's slower that the solution in #405 \r\n\r\nOn my side (select 10 000 examples):\r\n- Original implementation: 12s\r\n- Batched solution: 100ms\r\n- solution using arrow tables: 350ms\r\n\r\nI'll try with arrays and record batches to see if we can make it work.","embeddings":[-0.1645014137,0.1181401834,-0.0444608256,-0.1049747691,-0.0814409629,-0.1731703728,0.147723943,0.5459221005,-0.2346828729,0.0624632165,-0.000081832,0.5998945832,-0.1947995573,-0.2284020036,0.2151886672,-0.0612115674,0.0476633497,-0.0043822541,-0.2822352946,-0.0438992195,-0.0562193953,-0.4137574136,-0.1994884014,-0.199706912,-0.0257708244,-0.2938672304,-0.1878563613,0.0697672665,-0.1976449341,0.0757235065,-0.217338413,0.281186074,-0.0521544069,0.2584172785,-0.0001122204,-0.1701223403,0.0425045677,0.1037315354,-0.3138458133,0.1286937594,0.3800330162,-0.0521176793,-0.1779631376,-0.2150924802,0.1908933669,-0.3175557554,0.1569269449,-0.078463316,-0.1330254078,-0.0641253069,0.1537112445,0.2656764984,-0.158904925,0.2145538479,0.5277472734,0.2444077432,-0.3638020456,0.2725051939,0.3779156208,-0.1275762767,-0.2558321059,0.2600165606,-0.1540908068,-0.05056981,0.2140832543,0.1419372708,0.0656322241,0.3851957321,0.0977453068,0.3810899258,0.2966608405,-0.1960287392,-0.2890027761,-0.2553756237,0.1719614565,-0.036944937,0.0981168896,0.2776352167,-0.276034385,-0.2166483849,-0.1713003367,0.0930572525,0.2021472156,-0.2257492393,0.3682128191,0.3595198393,0.1895429045,-0.0979154855,0.6719359756,-0.1242221296,0.5510306358,-0.1262992322,-0.0047993273,0.4045140147,-0.3214240074,-0.318256259,-0.2337398827,-0.1205970347,-0.0468584299,0.1339881122,0.6908426881,0.2351148576,0.244086206,-0.0330901146,0.029474631,0.0050701234,-0.1881677061,0.0786928162,-0.1245698184,0.1490849257,0.254429251,0.1171050519,-0.0766196996,-0.3060936332,-0.1184603646,-0.2209762186,-0.5335650444,0.1360857934,0.0855467096,-0.2906821966,-0.0308560114,-0.0727870837,0.0191447586,0.217607528,-0.1134597436,0.2636504173,-0.0390845872,-0.1840486228,-0.0767866448,-0.0296166018,-0.0213162433,0.039533224,-0.1393529177,0.0402012654,0.117330268,0.4005283415,-0.2564073205,0.0705069825,0.1031442136,0.3123986125,0.023626605,-0.1334683448,0.0955936089,0.0091467304,-0.1462748051,-0.2022948563,-0.1624473184,0.0748468265,-0.2686469853,0.4047932327,0.008407644,-0.1314480156,0.1649600267,0.1682451367,0.2920650542,-0.2565487027,-0.5357545614,0.028208904,0.0150799761,-0.0602184683,0.0654673576,-0.0238683186,0.067362085,-0.1401836425,0.0244531073,-0.2616640925,-0.0000500487,0.150228098,-0.0067202686,0.551787138,0.4012793303,0.6005730629,-0.0661774054,0.0918936878,0.1528079659,-0.0147358095,0.65400213,0.2467100769,-0.2185637951,0.0435642451,-0.0482172333,-0.0262146741,0.3894706368,0.2075691968,0.1455650479,-0.1148245186,0.4333437085,0.236869961,-0.002023858,0.3048751354,-0.5437975526,-0.1804259866,0.2548130453,0.4310247898,-0.5073711276,-0.3955105245,0.0435986966,-0.3157792091,0.3072642088,-0.1929805428,-0.0006519615,0.048137743,-0.0866354853,0.0819223076,-0.180737868,0.0500819385,-0.2162789851,-0.0749926195,-0.0966446176,-0.0121546257,0.2877600491,-0.342728734,0.2107428014,-0.2337404788,0.2078987062,0.190705061,0.0213446133,-0.2241520733,0.3385126889,-0.1749186516,-0.1115132198,-0.0718107522,-0.2885776162,0.0425322279,-0.1818866283,-0.1339042038,0.0493438505,-0.2068353295,0.1409365982,0.3582656384,-0.3306203187,0.0571808144,0.1285651922,0.2276866734,-0.1215711236,-0.2004279792,0.1078855172,0.1838665158,-0.072261624,0.2219088227,0.1093335375,-0.2368921489,0.1598075479,-0.0240699071,-0.3439081609,0.5287792683,-0.4670099616,0.3060247004,0.1508340091,-0.2817956209,-0.1983257383,0.3005359471,-0.0364965759,0.4161673188,0.6197937131,0.0703198686,-0.0085375905,0.382840991,-0.2867331803,0.1110055223,0.4442731142,0.0354825109,-0.0874339268,0.0763645917,0.1151951253,-0.1434492767,-0.1449111551,0.0289835874,0.220110178,0.2666477263,0.0663305745,-0.0658477992,-0.0587906055,-0.2190838456,0.1854216009,-0.0540741794,0.1081660613,-0.116148755,0.1254767329,-0.0036939338,-0.1517492533,-0.1532523036,0.1124854162,0.1517265737,0.1317405552,0.0266395733,-0.0931461453,-0.1624273658,-0.1187378243,-0.0148237161,-0.047975637,-0.0085713789,0.2399174869,0.239157632,-0.2472816855,-0.2346794307,-0.036851123,0.3515058458,-0.1599138081,-0.0430689976,-0.1856266111,0.0427749306,-0.0941626951,0.0918409526,0.2291274518,0.2741205692,0.3628977239,0.098799035,-0.2361453772,0.0463990048,0.0200848281,-0.2806491256,0.0375690013,-0.2801628113,-0.0506352596,-0.1814999431,-0.4017320573,-0.2940499485,0.0484575406,-0.2136878073,-0.1209500507,0.3594197929,0.2213585973,-0.1924252957,-0.1647051722,-0.1412424743,-0.0271904748,-0.4196079373,0.3166359365,-0.1018242091,0.2229262888,-0.0784572884,-0.1216172352,-0.0703331009,-0.5059905052,-0.0254836939,0.1991568059,-0.1796475351,0.2476859689,0.0956990272,-0.1881326139,-0.4313289523,-0.0367180184,-0.0820129737,0.4609701633,-0.0559894554,0.2647229135,-0.3275972903,-0.0486322194,-0.0881516859,0.164072901,0.4675890207,-0.123099193,-0.1321263015,0.2906356752,0.1878260374,0.0932345688,0.2486408055,0.148440972,0.107899718,0.1583170742,0.2577732503,0.4758410454,-0.0132597741,0.3537451029,-0.1806486249,0.1811614782,-0.0629272386,-0.3044174314,0.0902216434,0.1761184484,-0.0104922382,-0.1328268498,0.3145889342,-0.1303447336,-0.2358288169,-0.151966542,0.4087905586,0.1406176537,-0.2767400742,0.6161751151,-0.2175594568,0.0669737384,0.0130955642,0.1989501119,-0.0729394928,-0.2581873536,0.0346684307,-0.0771946162,-0.0035631838,0.0157671049,-0.6201167703,-0.1037961766,-0.7348018289,0.1945824176,0.2546136975,0.4203133285,0.2630220056,-0.2731857598,0.4100108147,-0.1316811293,0.2454742938,0.1847005486,-0.3379400074,0.2078168094,-0.0334034637,-0.3991730511,-0.0382954292,-0.3553336263,-0.3880005479,0.0934058726,0.2187519968,-0.3698848188,-0.0151682589,-0.0648521632,-0.0963634029,-0.0208882075,-0.0213152077,-0.3431763351,-0.1563177705,0.0247229803,0.3189601898,-0.0419459343,0.0636190102,-0.036534097,0.2075858116,0.1854451448,0.0052047325,0.0925563872,-0.1126060188,-0.0189581849,-0.0445405841,0.1609745771,0.1449077725,0.0884859785,-0.0016811267,-0.0676052347,0.1565766633,0.1307452619,-0.077515915,-0.2168564498,0.4103648067,0.3544317782,0.213284418,0.2864861488,0.0318825357,0.2609132528,-0.004494627,-0.4609469771,0.0524824671,-0.3368801177,-0.1111986488,-0.7022079825,0.22072725,0.2098456919,0.1291486621,0.2435236424,-0.1922961473,0.1943425387,0.399099648,0.1793601811,0.9543699622,-0.3425573409,0.439509511,0.200553298,-0.2498934418,0.0421645492,-0.1718843728,0.4339771271,-0.1869428158,-0.1428172886,0.2209912241,-0.1602591276,-0.0038971228,0.1749671698,0.0103839319,0.1408389062,0.0545829684,-0.2724593282,-0.2982908487,-0.0769248977,-0.3973248601,-0.4000498354,-0.0812608972,0.007990622,-0.0912701115,-0.2616395354,0.1299060285,0.0397277512,-0.4066441357,0.1374417692,-0.1283323169,-0.0339099467,0.02526352,-0.0187811032,0.2313003093,-0.2472966909,-0.0118320631,-0.138942495,-0.1654712111,0.1175363883,0.1782703102,0.0876710638,0.2732319832,0.3476555347,0.3208275437,-0.2797983587,0.2450403422,-0.0386296809,0.2470979393,0.0269273277,-0.0094231339,-0.7398144603,-0.2770707905,0.0814099312,0.5765282512,-0.0462440848,-0.0412510894,0.2288649678,-0.3677733243,-0.0874123052,0.1226013601,0.0234579053,0.1255804002,0.4317756295,0.1474375427,-0.1711010784,-0.0032447018,0.3074453771,0.1008310392,0.1374729723,0.2142409235,0.0512026809,-0.4982791543,-0.2763272822,0.1729940027,0.0560969338,-0.4236968458,0.2555827796,0.0346920267,-0.0998304859,0.198374629,0.280202359,-0.0444325991,-0.0701618642,-0.2347720563,-0.0422129072,0.0778445154,0.4823131859,0.437608093,-0.0092766155,0.0434517562,0.2773646414,-0.4098943174,-0.1012692004,-0.2635428607,-0.1706923693,0.103113912,0.19171606,-0.1570941061,-0.2485818565,-0.2314547747,0.1217744797,0.1095124483,0.1856368482,0.0099595217,-0.2204296887,0.0652233809,0.0268537328,0.0692745671,-0.0124053657,-0.0442201123,-0.4056577384,-0.4026626945,-0.2240932882,-0.2556392252,-0.0581012294,-0.0186286401,-0.2951931357,0.1893857419,0.1526644081,-0.145420894,0.5044122338,-0.440568924,-0.2880323827,-0.2536023259,-0.0390556492,-0.0204363931,-0.1000388265,-0.1237268001,0.1322373152,0.0891379118,-0.009800585,0.2713639438,-0.1450659782,0.032684885,-0.0542566627,0.1430614591,0.5390290022,0.2918041646,-0.4942596257,0.4230967462,0.1387192309,-0.068557471,-0.2051184624,-0.4018735886,0.2666890323,0.0158820525,0.0736496896,-0.2112586647,-0.228680566,-0.3007588387,-0.0467096493,-0.1985903382,-0.4065413475,-0.2695232928,0.3000281751,0.3499613702,-0.281037271,0.1548964679,-0.020922618,-0.0263043027,0.1267722547,0.2693783641,0.392806381,0.7060084939,0.3252286315,-0.4333920479,-0.0891521126,-0.0549950935,0.4373283386,0.2419597507,0.2988152206,0.3137167096,0.173214525,0.0510274507,0.0750978217,0.0485931672,-0.1860374361,-0.0548712313,0.0830108374,-0.0129107442,-0.1432317793,-0.2376493961,0.184269011,-0.3152509928,0.3856135309,0.186242491,0.0794779509,0.4294978082,-0.442810595,0.1762872338,-0.0807410628,0.4567211568,-0.2809702158,-0.0646833256,-0.0432205647,0.105342634,0.3080555797,0.090157412,-0.1798043996,0.2704426944,-0.0519987643,0.0788251609,-0.4874316156,0.0217130836,-0.1057703868,0.2321112156,0.4933184087,-0.1453858167,0.257178396,0.0264486112,-0.1935745478,-0.1281039119,-0.06671004,-0.0381668992,-0.0670603365,0.0632705763,0.2352298796,-0.4257184267,0.0281729829,0.145007655,-0.2164347321,-0.0908324495,-0.4721774757,-0.2413542867,-0.0959173217,-0.0337368622,0.0694231391,0.1823056936,0.6906647682,0.3652186692,-0.1364850402,0.1033619344,-0.3405781984,-0.1763215214,-0.3086301088,-0.2202321738,0.0691790432,0.2312633395,0.4550159276,0.2721434236,0.2278193235,-0.0682937652,0.1393437982,-0.3666184545,0.3984475136,-0.1806874126,0.2856487632,0.2190860659,-0.0290677194,-0.1811083555,-0.6279549003,0.2281725109,0.1717589051,0.0131299291,-0.0495901294,0.1054144725,0.1774181724,0.3426143527,-0.2193612456,-0.0282706562,-0.2209174186,0.2068751156,-0.1463597417,-0.2218937129,0.1480949819,-0.0441628173,-0.0158977844,0.163041696,0.1358191073,0.0746878237,0.0321721844,0.1745644957,0.4784035981,-0.2635795772,0.280218184,-0.2502786219,0.0017437503,0.1643892527,0.2768270969,-0.1753014326,0.1120071188,-0.0119547257,-0.3055001795,-0.2477165461,-0.0073188781,0.3155240417,-0.1459379494,-0.2534733415,-0.0785351843,0.1924701035,0.2831954062,0.3042110801,0.0231905542,-0.3225080371,-0.053712368,0.0298332796,0.3315554261,0.2489203066,-0.1983092427,-0.0402146094,-0.1452148855,0.3150457442,-0.420828253,0.1110670343,-0.289252162,-0.3250941038]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/406","title":"Faster Shuffling?","comments":"I tried using `.take` from pyarrow recordbatches but it doesn't improve the speed that much:\r\n```python\r\nimport nlp\r\nimport numpy as np\r\n\r\ndset = nlp.Dataset.from_file(\"dummy_test_select.arrow\") # dummy dataset with 100000 examples like {\"a\": \"h\"*512}\r\nindices = np.random.randint(0, 100_000, 1000_000)\r\n```\r\n\r\n```python\r\n%%time\r\nbatch_size = 10_000\r\nwriter = ArrowWriter(schema=dset.schema, path=\"dummy_path\",\r\n writer_batch_size=1000, disable_nullable=False)\r\nfor i in tqdm(range(0, len(indices), batch_size)):\r\n table = pa.concat_tables(dset._data.slice(int(i), 1) for i in indices[i : min(len(indices), i + batch_size)])\r\n batch = table.to_pydict()\r\n writer.write_batch(batch)\r\nwriter.finalize()\r\n# 9.12s\r\n```\r\n\r\n\r\n```python\r\n%%time\r\nbatch_size = 10_000\r\nwriter = ArrowWriter(schema=dset.schema, path=\"dummy_path\", \r\n writer_batch_size=1000, disable_nullable=False)\r\nfor i in tqdm(range(0, len(indices), batch_size)):\r\n batch_indices = indices[i : min(len(indices), i + batch_size)]\r\n # First, extract only the indices that we need with a mask\r\n mask = [False] * len(dset)\r\n for k in batch_indices:\r\n mask[k] = True\r\n t_batch = dset._data.filter(pa.array(mask))\r\n # Second, build the list of indices for the filtered table, and taking care of duplicates\r\n rev_positions = {}\r\n duplicates = 0\r\n for i, j in enumerate(sorted(batch_indices)):\r\n if j in rev_positions:\r\n duplicates += 1\r\n else:\r\n rev_positions[j] = i - duplicates\r\n rev_map = [rev_positions[j] for j in batch_indices]\r\n # Third, use `.take` from the combined recordbatch\r\n t_combined = t_batch.combine_chunks() # load in memory\r\n recordbatch = t_combined.to_batches()[0]\r\n table = pa.Table.from_arrays(\r\n [recordbatch[c].take(pa.array(rev_map)) for c in range(len(dset._data.column_names))],\r\n schema=writer.schema\r\n )\r\n writer.write_table(table)\r\nwriter.finalize()\r\n# 3.2s\r\n```\r\n","body":"Consider shuffling bookcorpus:\r\n\r\n```\r\ndataset = nlp.load_dataset('bookcorpus', split='train')\r\ndataset.shuffle()\r\n```\r\nAccording to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`.\r\n\r\nBut I can also just write the lines to a text file:\r\n\r\n```\r\nbatch_size = 100000\r\nwith open('tmp.txt', 'w+') as out_f:\r\n for i in tqdm(range(0, len(dataset), batch_size)):\r\n batch = dataset[i:i+batch_size]['text']\r\n print(\"\\n\".join(batch), file=out_f)\r\n```\r\n\r\nWhich completes in a couple minutes, followed by `shuf tmp.txt > tmp2.txt` which completes in under a minute. And finally,\r\n\r\n```\r\ndataset = nlp.load_dataset('text', data_files='tmp2.txt')\r\n```\r\n\r\nWhich completes in under 10 minutes. I read up on Apache Arrow this morning, and it seems like the columnar data format is not especially well-suited to shuffling rows, since moving items around requires a lot of book-keeping. \r\n\r\nIs shuffle inherently slow, or am I just using it wrong? And if it is slow, would it make sense to try converting the data to a row-based format on disk and then shuffling? (Instead of calling select with a random permutation, as is currently done.)","comment_length":210,"text":"Faster Shuffling?\nConsider shuffling bookcorpus:\r\n\r\n```\r\ndataset = nlp.load_dataset('bookcorpus', split='train')\r\ndataset.shuffle()\r\n```\r\nAccording to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`.\r\n\r\nBut I can also just write the lines to a text file:\r\n\r\n```\r\nbatch_size = 100000\r\nwith open('tmp.txt', 'w+') as out_f:\r\n for i in tqdm(range(0, len(dataset), batch_size)):\r\n batch = dataset[i:i+batch_size]['text']\r\n print(\"\\n\".join(batch), file=out_f)\r\n```\r\n\r\nWhich completes in a couple minutes, followed by `shuf tmp.txt > tmp2.txt` which completes in under a minute. And finally,\r\n\r\n```\r\ndataset = nlp.load_dataset('text', data_files='tmp2.txt')\r\n```\r\n\r\nWhich completes in under 10 minutes. I read up on Apache Arrow this morning, and it seems like the columnar data format is not especially well-suited to shuffling rows, since moving items around requires a lot of book-keeping. \r\n\r\nIs shuffle inherently slow, or am I just using it wrong? And if it is slow, would it make sense to try converting the data to a row-based format on disk and then shuffling? (Instead of calling select with a random permutation, as is currently done.)\nI tried using `.take` from pyarrow recordbatches but it doesn't improve the speed that much:\r\n```python\r\nimport nlp\r\nimport numpy as np\r\n\r\ndset = nlp.Dataset.from_file(\"dummy_test_select.arrow\") # dummy dataset with 100000 examples like {\"a\": \"h\"*512}\r\nindices = np.random.randint(0, 100_000, 1000_000)\r\n```\r\n\r\n```python\r\n%%time\r\nbatch_size = 10_000\r\nwriter = ArrowWriter(schema=dset.schema, path=\"dummy_path\",\r\n writer_batch_size=1000, disable_nullable=False)\r\nfor i in tqdm(range(0, len(indices), batch_size)):\r\n table = pa.concat_tables(dset._data.slice(int(i), 1) for i in indices[i : min(len(indices), i + batch_size)])\r\n batch = table.to_pydict()\r\n writer.write_batch(batch)\r\nwriter.finalize()\r\n# 9.12s\r\n```\r\n\r\n\r\n```python\r\n%%time\r\nbatch_size = 10_000\r\nwriter = ArrowWriter(schema=dset.schema, path=\"dummy_path\", \r\n writer_batch_size=1000, disable_nullable=False)\r\nfor i in tqdm(range(0, len(indices), batch_size)):\r\n batch_indices = indices[i : min(len(indices), i + batch_size)]\r\n # First, extract only the indices that we need with a mask\r\n mask = [False] * len(dset)\r\n for k in batch_indices:\r\n mask[k] = True\r\n t_batch = dset._data.filter(pa.array(mask))\r\n # Second, build the list of indices for the filtered table, and taking care of duplicates\r\n rev_positions = {}\r\n duplicates = 0\r\n for i, j in enumerate(sorted(batch_indices)):\r\n if j in rev_positions:\r\n duplicates += 1\r\n else:\r\n rev_positions[j] = i - duplicates\r\n rev_map = [rev_positions[j] for j in batch_indices]\r\n # Third, use `.take` from the combined recordbatch\r\n t_combined = t_batch.combine_chunks() # load in memory\r\n recordbatch = t_combined.to_batches()[0]\r\n table = pa.Table.from_arrays(\r\n [recordbatch[c].take(pa.array(rev_map)) for c in range(len(dset._data.column_names))],\r\n schema=writer.schema\r\n )\r\n writer.write_table(table)\r\nwriter.finalize()\r\n# 3.2s\r\n```\r\n","embeddings":[-0.2073036879,0.1424287409,-0.025603326,-0.1453503072,-0.1025906205,-0.1113140285,0.1211357713,0.5408861041,-0.2365823239,0.0190715939,-0.0381175689,0.5862330198,-0.1683035642,-0.276460439,0.2058225423,-0.0411366634,0.0331520326,-0.0362084135,-0.2147115171,-0.0471256189,-0.0500182919,-0.4499439895,-0.238439858,-0.2234770507,-0.0126475478,-0.225724414,-0.1654447317,0.0786909088,-0.165457949,0.0551608689,-0.1150205806,0.2012873143,-0.108007662,0.1667573005,-0.0001167829,-0.1386466324,0.099314861,0.1305755377,-0.3299337327,0.1783211082,0.451499939,-0.0504897796,-0.1290057003,-0.190895468,0.1902823597,-0.414375037,0.086571835,-0.1146503985,-0.1485535353,-0.0229181349,0.1140140891,0.3141115904,-0.1818512678,0.2556697726,0.4748253226,0.2703826129,-0.3750654459,0.3879562318,0.2586728334,-0.021415716,-0.1545352042,0.2556572556,-0.1949238777,0.0181237869,0.1430734247,0.1754679829,0.1023451537,0.4606055915,0.1073968634,0.4037255645,0.2905994356,-0.2549730241,-0.2978284657,-0.2941875756,0.2330021709,-0.0797465667,0.0570254363,0.3016262352,-0.2735319436,-0.2154012322,-0.0974206701,0.1640346795,0.2480155677,-0.2000269145,0.3221797049,0.3900923729,0.2021176666,-0.0607311912,0.7018341422,-0.105781056,0.4943529963,-0.1712976098,0.0219092872,0.411686033,-0.3191821575,-0.2636018097,-0.3010390103,-0.1255467832,-0.0711555183,0.1603339463,0.6939632297,0.2144636214,0.1904992461,-0.0473995619,0.0462994874,0.0036276807,-0.300027132,0.071423158,-0.1324055791,0.0648491308,0.3124971092,0.1253009737,-0.0777109712,-0.3089742362,-0.0426212512,-0.2280753404,-0.4801682532,0.1160597578,0.0091327066,-0.3454588354,-0.1375005543,-0.0533413328,0.0066610733,0.1520721167,-0.1462170929,0.2534679472,-0.0821314976,-0.1687271297,-0.1135871559,0.0287617557,-0.0324031971,0.0895730853,-0.1062909514,0.0718561411,0.0299970079,0.4717019498,-0.2529067397,0.0169554781,0.0952395871,0.3210919499,0.0433104672,-0.1136180982,0.083029367,0.1123675853,-0.0979958698,-0.149040252,-0.1326774359,0.0459325425,-0.339520812,0.3854333162,0.0259141456,-0.1179540604,0.0696305931,0.1244064644,0.2436410189,-0.288731873,-0.5230558515,-0.0648371875,-0.0311097726,-0.1096803695,0.0887992233,-0.0550133102,0.0886194408,-0.1288460195,0.0646334291,-0.2630294859,-0.0793608427,0.1027542576,-0.0587744936,0.5555560589,0.4522926211,0.6798374057,-0.071005933,0.0568028465,0.1080939099,0.0309065524,0.6843819618,0.1945423037,-0.253213495,-0.0193489902,-0.1120895743,0.0253338367,0.3429556489,0.2405741066,0.0689611956,-0.0779747441,0.3746800721,0.2942918837,0.0483492427,0.2648988962,-0.4871974587,-0.1362498254,0.2310036123,0.431266427,-0.6047837138,-0.3673205674,0.0409888513,-0.2627362311,0.3437721729,-0.1634198278,-0.0241476707,-0.002875691,-0.0728363693,0.0040968494,-0.1673831344,0.1009118035,-0.2770056427,-0.0851025879,-0.0123086534,-0.0934438035,0.2220733315,-0.3609888852,0.1847952455,-0.1915204227,0.1916127056,0.207287401,-0.010497516,-0.1980127543,0.4299571514,-0.1024159417,-0.0807139352,-0.031038601,-0.2975625396,0.0371089727,-0.1596858054,-0.1389845014,0.1140311584,-0.2169331312,0.135617584,0.3588954806,-0.2914271355,0.0721107796,0.0999829918,0.2370676994,-0.0914403051,-0.1885744333,0.0344020613,0.2401288152,-0.0574304909,0.2263465077,0.118693836,-0.2045537233,0.1560783982,-0.0530229695,-0.4021678865,0.5432929397,-0.4847533405,0.3439686,0.1877929568,-0.2420200109,-0.1256657988,0.2605871856,-0.0751947165,0.4508640468,0.59654212,0.0635082498,0.0748018399,0.4007713795,-0.2726922035,0.0491742678,0.4478168488,-0.0474845879,-0.0537610278,0.0778427869,-0.0244941916,-0.2125136256,-0.0856432244,0.0269686021,0.2142338306,0.2641347349,0.1126311794,-0.0479865111,-0.0872229114,-0.2304076701,0.1614469439,-0.0465985313,0.0837044716,-0.0975083485,0.1120569482,-0.0303853583,-0.2148164213,-0.1830169559,0.0966954678,0.1835551709,0.1354514509,0.0691546574,-0.1409353316,-0.1086602136,-0.1263204813,-0.0809588283,0.0280734841,-0.1020173505,0.1147613153,0.2489230931,-0.2036471665,-0.2030912042,-0.1168933585,0.335203737,-0.1054509804,-0.1940307021,-0.182344228,-0.0091365818,-0.1343149394,0.0485685654,0.197874859,0.3558698595,0.3271757662,0.1494715214,-0.220095247,0.0215608403,0.0938112438,-0.2209096402,0.0969474986,-0.3778560162,-0.0812894404,-0.1770183295,-0.4263556898,-0.2799151838,0.0594400652,-0.1173179075,-0.0908795819,0.3733411133,0.1376002133,-0.2064498812,-0.1201335639,-0.0493216701,-0.0749145225,-0.3606601655,0.3090262711,-0.1253742874,0.1422898024,-0.0030990706,-0.1096464843,-0.0249080025,-0.5705391169,-0.055480022,0.1581888199,-0.1193132997,0.2787475884,0.1174005941,-0.1336577386,-0.451035738,-0.0693096444,-0.0896496326,0.4983684719,-0.077488102,0.221291855,-0.3810477257,-0.0344085172,-0.1012929603,0.140446201,0.4684356451,-0.187535882,-0.1023728997,0.3078249693,0.1736066341,0.0701604113,0.2383754402,0.1195240542,0.1213327423,0.1363343149,0.2258581817,0.4733502865,-0.0470225215,0.2490298152,-0.2466053814,0.0461117513,-0.030770259,-0.3000233769,0.1443730891,0.1948063076,0.0175135043,-0.1204537004,0.2602733374,-0.1087932959,-0.2526511848,-0.0579940528,0.4636721909,0.1634529978,-0.2731415927,0.5276740193,-0.2075201571,0.0069012763,-0.0099865785,0.2193984836,-0.0758057386,-0.2332177907,0.0865597874,-0.0811879113,-0.0393811017,-0.0054829991,-0.591270566,-0.0939203575,-0.6903495193,0.2243586034,0.2033419162,0.4551722109,0.2934636474,-0.3032829165,0.4281890094,-0.1267066747,0.2624970973,0.1844921261,-0.2910933495,0.2060330063,0.006221327,-0.4882081151,-0.074489519,-0.3385715187,-0.3702441752,0.09648031,0.2184390426,-0.3918660879,0.0253927764,0.0135671292,-0.0772423446,-0.0157019086,0.0119210128,-0.3758602142,-0.246671021,0.0783744976,0.2998023927,-0.053579133,0.0080257803,0.1029722691,0.2369657308,0.126701206,-0.0103700822,0.0648303553,-0.0899246037,-0.0197218005,-0.00009425,0.1376554072,0.1903957278,0.0832222849,-0.1275289059,-0.0508880392,0.2487169355,0.0248962976,-0.1076378599,-0.1637818962,0.3971323073,0.3331096172,0.2223040015,0.2512410581,0.0435301773,0.2704720795,0.0193484239,-0.44960922,0.090274848,-0.3458971083,-0.077921927,-0.6981349587,0.3353745639,0.303420186,0.1467077434,0.2382647842,-0.1312589049,0.2584004104,0.3906689584,0.186815992,0.9388628006,-0.3655802906,0.4324247241,0.3174656332,-0.2735365629,0.0770978332,-0.0952979773,0.3620699942,-0.1664985716,-0.061501082,0.2290952355,-0.2028226852,-0.0359408446,0.0739555582,0.0364176407,0.1265502274,-0.0108730439,-0.2703076303,-0.279145211,-0.0251851324,-0.3461185098,-0.4222583473,-0.0649609119,0.009078389,-0.1261395663,-0.1887585372,0.1764241606,0.1053680554,-0.3683370352,0.1437904984,-0.1949707121,-0.0230597276,0.1270229816,-0.0127135003,0.2234876156,-0.2524955869,-0.1077337638,-0.1966214031,-0.186824441,0.0435277447,0.1743912399,0.1419889778,0.2909935415,0.3207314909,0.3545377851,-0.3061875403,0.2015719861,0.0204854496,0.2604016662,0.0838741437,0.0118568987,-0.7292296886,-0.2502437532,0.0936560482,0.6161046028,0.0301326979,-0.0538853221,0.178406015,-0.3653874695,-0.0951613113,0.0847700536,0.0038558878,0.1426538676,0.3974073231,0.216878593,-0.1787638962,-0.0021242311,0.3385836184,0.0916307792,0.1737975478,0.2557826042,0.0044427477,-0.4787675142,-0.2208833098,0.119575344,0.0927251503,-0.3995622098,0.2578277886,0.0304933097,-0.0647594333,0.1437389702,0.2030364871,-0.0384067558,-0.0316661634,-0.3079482019,-0.0839979649,0.1671343893,0.4880536497,0.4494493306,0.0407273211,0.0059980326,0.2657169104,-0.4514318109,-0.1856047809,-0.2125832886,-0.1643106341,0.1183736995,0.1486976594,-0.1962727904,-0.2467669398,-0.2014065534,0.1216192469,0.1083658114,0.2122836113,0.0690098479,-0.1696735173,0.0784123614,0.056526728,0.0680002421,-0.0296172164,0.0130018573,-0.4090905488,-0.3787909746,-0.2591435611,-0.2203105539,-0.0359104015,-0.0143615436,-0.3235987723,0.1757287085,0.137086153,-0.1018621624,0.5023982525,-0.4558425844,-0.2822435796,-0.2124385983,-0.0117253121,0.0002811252,-0.0607919544,-0.1332336217,0.0808449388,0.1220135987,-0.066112943,0.232734248,-0.1760277599,0.0310681332,-0.1178009957,0.2123600692,0.5724335909,0.2858697176,-0.5049549937,0.413299948,0.0761313736,-0.1094661951,-0.2035363019,-0.3783605397,0.2629653513,0.0307829492,0.0544832088,-0.2800425589,-0.1498418748,-0.4016994834,-0.1037763879,-0.1208236217,-0.4394886792,-0.3144517541,0.344271481,0.3613233268,-0.2875060141,0.1918438226,-0.0268156193,0.0144084906,0.1336645186,0.212620914,0.3895176351,0.7033469081,0.298170954,-0.4683302045,-0.0908866152,-0.0182529539,0.5345419645,0.2181617469,0.315285027,0.3625561595,0.1196441129,0.060079582,0.1762859374,-0.0175822228,-0.1426124275,0.0021404983,0.0642027706,-0.0033137943,-0.0859617144,-0.2722110152,0.181903854,-0.3435494602,0.3356219232,0.2366926223,0.0677237585,0.4312195778,-0.4143007994,0.1808579415,-0.1028547883,0.4148302674,-0.2638714314,-0.1417444199,-0.0495339483,0.1888810098,0.2471141666,0.0517047606,-0.13872765,0.2436002642,-0.0785278976,0.0411729217,-0.5003128648,-0.0093053924,-0.0582355782,0.2283153236,0.4879380763,-0.1585579664,0.2480301559,0.0245643593,-0.1947028488,-0.1007224992,-0.0872062594,-0.0288396198,-0.1066076458,0.1006258577,0.1622860581,-0.37289235,-0.0478509516,0.1560518891,-0.1136872396,-0.1456872821,-0.5345342159,-0.1929797679,-0.1256036162,0.0831799358,0.042838309,0.0916244015,0.7034350038,0.4074105322,-0.1263747513,0.1671237648,-0.2396189421,-0.2513005435,-0.2989163399,-0.3349252045,-0.1049491912,0.2417289317,0.5042904615,0.2844422162,0.3205120862,-0.0640754476,0.1768755019,-0.3291893601,0.3407906592,-0.1731744409,0.2252847999,0.252581954,-0.0485382825,-0.2170930207,-0.6656162143,0.1990628839,0.2431149781,-0.0058195135,0.0005772477,0.0327274352,0.1565547884,0.302046001,-0.1743313521,-0.0529877059,-0.166540727,0.226595208,-0.1683638841,-0.2666847706,0.2494230568,0.0240982026,0.0117007997,0.0665719658,0.1200326756,0.0621138774,0.0830299333,0.1514196247,0.5137016773,-0.2719222307,0.2024232447,-0.2667109966,0.0697059706,0.1493946165,0.3052288592,-0.2260488421,0.0448298939,-0.0256067291,-0.2617835104,-0.1980850548,0.0609435514,0.4382947683,-0.1793805063,-0.2132152319,-0.0514393151,0.2328603566,0.2575269639,0.3161094189,0.1360980123,-0.2586580515,-0.0449450985,-0.0011650252,0.3642781377,0.1605045199,-0.1450717896,-0.0107186791,-0.1453756094,0.3728183806,-0.4335768521,0.0835622177,-0.3130494952,-0.3294507861]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/406","title":"Faster Shuffling?","comments":"Shuffling is now significantly faster thanks to #513 \r\nFeel free to play with it now :)\r\n\r\nClosing this one, but feel free to re-open if you have other questions","body":"Consider shuffling bookcorpus:\r\n\r\n```\r\ndataset = nlp.load_dataset('bookcorpus', split='train')\r\ndataset.shuffle()\r\n```\r\nAccording to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`.\r\n\r\nBut I can also just write the lines to a text file:\r\n\r\n```\r\nbatch_size = 100000\r\nwith open('tmp.txt', 'w+') as out_f:\r\n for i in tqdm(range(0, len(dataset), batch_size)):\r\n batch = dataset[i:i+batch_size]['text']\r\n print(\"\\n\".join(batch), file=out_f)\r\n```\r\n\r\nWhich completes in a couple minutes, followed by `shuf tmp.txt > tmp2.txt` which completes in under a minute. And finally,\r\n\r\n```\r\ndataset = nlp.load_dataset('text', data_files='tmp2.txt')\r\n```\r\n\r\nWhich completes in under 10 minutes. I read up on Apache Arrow this morning, and it seems like the columnar data format is not especially well-suited to shuffling rows, since moving items around requires a lot of book-keeping. \r\n\r\nIs shuffle inherently slow, or am I just using it wrong? And if it is slow, would it make sense to try converting the data to a row-based format on disk and then shuffling? (Instead of calling select with a random permutation, as is currently done.)","comment_length":29,"text":"Faster Shuffling?\nConsider shuffling bookcorpus:\r\n\r\n```\r\ndataset = nlp.load_dataset('bookcorpus', split='train')\r\ndataset.shuffle()\r\n```\r\nAccording to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`.\r\n\r\nBut I can also just write the lines to a text file:\r\n\r\n```\r\nbatch_size = 100000\r\nwith open('tmp.txt', 'w+') as out_f:\r\n for i in tqdm(range(0, len(dataset), batch_size)):\r\n batch = dataset[i:i+batch_size]['text']\r\n print(\"\\n\".join(batch), file=out_f)\r\n```\r\n\r\nWhich completes in a couple minutes, followed by `shuf tmp.txt > tmp2.txt` which completes in under a minute. And finally,\r\n\r\n```\r\ndataset = nlp.load_dataset('text', data_files='tmp2.txt')\r\n```\r\n\r\nWhich completes in under 10 minutes. I read up on Apache Arrow this morning, and it seems like the columnar data format is not especially well-suited to shuffling rows, since moving items around requires a lot of book-keeping. \r\n\r\nIs shuffle inherently slow, or am I just using it wrong? And if it is slow, would it make sense to try converting the data to a row-based format on disk and then shuffling? (Instead of calling select with a random permutation, as is currently done.)\nShuffling is now significantly faster thanks to #513 \r\nFeel free to play with it now :)\r\n\r\nClosing this one, but feel free to re-open if you have other questions","embeddings":[-0.1669747978,0.0758930519,-0.0756695494,-0.1069152504,-0.0451760292,-0.1096984372,0.1049450263,0.545971632,-0.2224266231,0.1107433885,-0.0171533898,0.5978963375,-0.1472128481,-0.2247790396,0.1525786072,-0.0917216018,0.0197470859,-0.0066342889,-0.2984012663,-0.0681981742,-0.1076799184,-0.3407295346,-0.1966788322,-0.2414292395,-0.0211049747,-0.2671126127,-0.1641610265,0.1271981448,-0.1086063236,0.035609737,-0.1816464514,0.2036722153,-0.1019436568,0.2518034875,-0.0001064402,-0.1878170967,-0.0013785269,0.1245126128,-0.3185117841,0.1700074822,0.4203979969,-0.0366483703,-0.1810671836,-0.2470211387,0.1335794181,-0.2439090461,0.1669650376,-0.0882611051,-0.0787344053,-0.0348335542,0.1946194321,0.3067126572,-0.1423862129,0.1317119598,0.5130022168,0.0983550847,-0.333455056,0.2333613783,0.2945879102,-0.0863215774,-0.1773552746,0.2387263626,-0.0959547982,-0.0672887415,0.1371427625,0.1144751385,0.0959913209,0.3252631128,0.0591608323,0.4156249166,0.2406332344,-0.2018209398,-0.2674714327,-0.2720342875,0.2022290379,-0.0449447893,0.1043059304,0.270442605,-0.2393940538,-0.1595278084,-0.0910153463,0.1850961596,0.2104175687,-0.1917834729,0.2724059224,0.4423377514,0.2230263501,-0.1347956955,0.6850678325,-0.1429264992,0.3533965349,-0.0897843167,0.0236880276,0.4135656357,-0.2834292352,-0.3294468224,-0.2499971837,-0.1825529486,-0.1107685864,0.183896631,0.7327424288,0.245347321,0.2195363343,-0.0025334849,0.0044257445,0.0172558874,-0.3033528924,0.0676690936,-0.083319135,0.1741636992,0.1895967871,0.1265306473,-0.0875518471,-0.3122540712,-0.1035777926,-0.167716518,-0.4825846553,0.1387092918,0.0155337201,-0.2844125628,-0.0667908415,-0.0099283764,0.0018037392,0.2151009291,-0.1538619548,0.2751540542,-0.03062962,-0.2389086932,-0.1474363059,-0.0921697393,-0.0841054097,0.0927617177,-0.1854152381,0.0111724278,0.1070576981,0.4471132457,-0.1813143045,0.0750126243,0.0932677612,0.3140338957,0.0747543871,-0.1450024843,0.052906286,0.0301001482,-0.1538800597,-0.1871067733,-0.163969323,0.1298393905,-0.2964134812,0.3371615708,0.0634578317,-0.1148025468,0.1358181983,0.241995126,0.2536460161,-0.2783489227,-0.3240192831,0.0042138305,-0.0300784502,-0.0588060208,0.0889037922,0.0079699345,0.0558057651,-0.1506493837,0.0332125574,-0.2832565904,-0.030102957,0.1387460679,-0.0630680472,0.4716358483,0.3784511387,0.5905828476,-0.0927200541,0.0524727404,0.1518076509,0.1000737399,0.6754481792,0.1847668141,-0.222253859,0.0069721234,-0.1234954968,0.0189208183,0.329983443,0.2190211713,0.1064449102,-0.0470619909,0.4243724942,0.3086519241,0.057607308,0.267239809,-0.5271468163,-0.1680125743,0.2397631556,0.4268015623,-0.4836195111,-0.3931616545,0.0215596184,-0.2457869351,0.3352416754,-0.2223267257,0.0048054908,0.0780970454,-0.0988785103,0.0684452802,-0.1722119153,-0.0461623035,-0.2284312844,-0.0307575669,-0.051114507,-0.0219243765,0.2823062837,-0.3760942221,0.138916105,-0.2351230532,0.2231093794,0.1523119509,0.1031891629,-0.1222540736,0.3160507083,-0.1441675276,-0.1249852479,-0.0329151638,-0.2858459651,0.0602336489,-0.1428825557,-0.1032397375,0.0414223596,-0.1743188947,0.1801927984,0.4312620759,-0.3207770288,0.0851427093,0.1532296389,0.2604211271,-0.1293295026,-0.1546777636,0.1511733979,0.177882418,-0.0839824229,0.2608782649,0.0671195239,-0.098170504,0.1191634312,0.0282842238,-0.3607949317,0.576117754,-0.4362382293,0.2577603459,0.1505965441,-0.2896023989,-0.1527445763,0.192486763,-0.0689205825,0.3874967694,0.6559903026,0.0933465883,0.0408643782,0.4048003852,-0.3637098074,0.1366246641,0.505181849,0.0175943319,-0.0612066314,0.0454889685,0.0998163521,-0.0828498527,-0.1516596079,0.0592621751,0.1986665726,0.3086874187,0.1105102152,-0.0429085679,-0.152237311,-0.2253775895,0.1467131227,-0.0810682774,0.1443347335,-0.0461751521,0.1731486171,0.0282911379,-0.246881023,-0.1958767325,0.1010336876,0.1578456461,0.1300912499,0.0456534065,-0.1185931787,-0.1928806603,-0.1469243616,-0.0087559484,-0.0410588533,-0.0655926466,0.220376268,0.2674484551,-0.2967941463,-0.2441508919,-0.0966950953,0.3949586451,-0.0981088355,-0.0882318169,-0.2022629529,0.0434115082,-0.1129604802,0.1577924937,0.2093465328,0.278398037,0.4065812826,0.1497819871,-0.2056581974,0.1125960425,-0.0035842732,-0.2856873572,0.0597447045,-0.2784018219,-0.0796291307,-0.1288377196,-0.3703228831,-0.3678686917,0.0662479699,-0.1745668799,-0.1471477002,0.2800329924,0.2361859679,-0.1627960354,-0.2388626486,-0.2241316885,-0.0004679693,-0.4718467295,0.331335783,-0.0771482959,0.2015341073,-0.0852287114,-0.1230721995,-0.0027067505,-0.4613552392,-0.0153979519,0.1507292241,-0.1685830355,0.2793680131,0.0150951482,-0.1759365797,-0.4089842737,-0.0797424465,-0.0427275747,0.4191710949,-0.1128403246,0.2114582658,-0.354473412,-0.0188404694,-0.1047074422,0.0991762578,0.4139265716,-0.1405698955,-0.2141785324,0.2670760453,0.1871360689,0.056598898,0.2416381836,0.1690844893,0.0107140765,0.1163991541,0.2601451874,0.4409126639,-0.0831339657,0.3029327691,-0.1711633205,0.1169440746,-0.0787562653,-0.2479393631,0.0979947448,0.1595518291,0.0225039534,-0.1101275235,0.3593632579,-0.0794383511,-0.288523972,-0.0989053771,0.373028636,0.1068892702,-0.2913453579,0.5682029724,-0.1953889728,0.0134630566,-0.0165437032,0.1917994618,-0.0891628563,-0.2027381361,0.0412720069,-0.1133088693,-0.0761409476,0.0043632886,-0.5783767104,-0.1198523864,-0.7263136506,0.1553502232,0.2316436321,0.3683967292,0.2024151087,-0.3260190487,0.3546539843,-0.1572314799,0.3486216366,0.1606164873,-0.3252027333,0.2688458264,-0.0398482978,-0.4078156948,-0.0648059323,-0.3783377707,-0.3934673667,0.1685018539,0.2311153412,-0.3417012095,-0.032316763,-0.0146008488,-0.1042356938,-0.0056248005,-0.0057705073,-0.3624520898,-0.2114628106,-0.057143975,0.3083366454,-0.0509892739,0.112847887,-0.0503511541,0.1073432937,0.1902977973,-0.0195575561,0.1275953948,-0.0592936128,0.0056244261,0.0301178191,0.1298806071,0.1527166516,0.1490024626,-0.1030580923,-0.0561410785,0.1591078043,0.1278185099,-0.136247918,-0.2232064903,0.4086349905,0.3234094977,0.2304273397,0.270647347,0.0015274328,0.3054816723,0.0099506481,-0.3429881632,0.1012383625,-0.3108819425,-0.1403184384,-0.6326962709,0.2537314892,0.2337563336,0.1193951219,0.2409456819,-0.2747868299,0.1304340214,0.3508600891,0.1714952737,0.9784497619,-0.3513526917,0.4275633097,0.2920373976,-0.1599371582,0.1142089143,-0.2114717364,0.4349349737,-0.1626183391,-0.0779476836,0.2206713706,-0.1078561321,-0.0328954309,0.1291023791,-0.0351381376,0.1118033156,0.0240857694,-0.2238420099,-0.3050850034,-0.0248682443,-0.3679124713,-0.4044308066,-0.1523007005,0.0756226704,-0.03530984,-0.2183144838,0.0845697746,0.0569175966,-0.3565792143,0.1671098322,-0.1411593109,-0.0466478653,0.04243223,-0.0049425685,0.1969114542,-0.2198662907,-0.0890046954,-0.1474564821,-0.1571686268,0.1294109374,0.1665559262,0.1377653331,0.2241070718,0.317035377,0.2754783928,-0.2594660521,0.2806793451,-0.0948649943,0.1715536118,0.0587383807,0.0107995709,-0.7354450226,-0.2590070665,0.062895529,0.4252827764,-0.0435124524,0.0112038776,0.2331183404,-0.4184881449,-0.162594378,0.1803778261,-0.0564525872,0.0677822083,0.4045177996,0.259169817,-0.2443016618,-0.0556296669,0.3341696262,0.0881197527,0.1921718419,0.2500740588,0.0357504711,-0.4441968203,-0.3494407535,0.0937160552,0.0310786478,-0.285161227,0.259940654,0.0369748771,-0.1273554862,0.1988562644,0.2838091552,-0.0391919985,-0.0405680723,-0.2472267598,-0.0956170112,0.0702369511,0.4469086528,0.4409604371,0.0791365281,-0.0008864227,0.3681529462,-0.3891652226,-0.1299374849,-0.333552897,-0.1338435709,-0.0133705596,0.1615582258,-0.1246140674,-0.2722464502,-0.1843118072,0.2149248868,0.1929067522,0.2269631773,-0.0056598396,-0.2980475128,0.0635205731,-0.0011383753,-0.0246388912,0.0445705466,0.0134925675,-0.3567683697,-0.3619440496,-0.2215157002,-0.2338103205,-0.1148808673,-0.0143768946,-0.1896773577,0.2071962655,0.1564918458,-0.0573200509,0.4489630759,-0.4027828276,-0.3101201355,-0.2046224475,-0.010111901,-0.0348246656,-0.0576540604,-0.0612218864,0.0715658069,0.1007815748,-0.0398184173,0.2357549071,-0.102798,-0.0396680869,-0.0157075655,0.1746306866,0.532772541,0.2393196076,-0.4209706485,0.3680213392,0.2163214386,-0.1668007225,-0.1690333039,-0.3094874322,0.1626503617,0.0659176484,0.0296354052,-0.2250951529,-0.1986941099,-0.3012602329,-0.0637249649,-0.1909381896,-0.3947408199,-0.1797793359,0.3211910725,0.3621634841,-0.3003880084,0.1553273648,-0.031640742,0.0585025512,0.1827101707,0.2390761971,0.3683657646,0.6607504487,0.3610320091,-0.4724350274,-0.1147835925,-0.0151023082,0.5250042677,0.2303861529,0.2822842598,0.3461492956,0.1851512641,0.0884339213,0.0645432621,-0.0455438495,-0.0914310664,-0.0940985754,0.0419210568,-0.0113568138,-0.1790161878,-0.2583903372,0.1940493286,-0.2067514658,0.3013904989,0.256706059,0.0706088021,0.4215821624,-0.4581938088,0.2092623413,-0.1324180961,0.466248244,-0.3132483363,-0.0837444514,0.0033856817,0.0879184157,0.2972169518,0.0233023055,-0.1268870234,0.2161459327,-0.1597123295,0.0817432851,-0.5459910035,0.027018873,-0.1925859153,0.2017210722,0.5362648964,-0.1407138556,0.3526444137,0.0941163301,-0.255708307,-0.1196413785,-0.0964167491,-0.0493319668,-0.1389132142,0.1064432785,0.232694909,-0.3602101803,-0.0536473617,0.1026574895,-0.2391812801,-0.1113286614,-0.4861654937,-0.2253012955,-0.1179696098,0.0232421514,0.1024109349,0.1478252858,0.7567028999,0.3429252803,-0.1528652012,0.166173026,-0.3669343293,-0.3012743592,-0.2292391658,-0.2744218409,0.034788616,0.1922419518,0.4784707129,0.2030562311,0.1948627234,-0.0333128013,0.097869806,-0.3102423251,0.3355526328,-0.2313392013,0.3123430908,0.1673236787,-0.0707292482,-0.1520499885,-0.6242113709,0.1967942417,0.1044086814,0.0960789323,0.0145560047,0.0407964475,0.1834350973,0.3243216872,-0.1760913879,-0.07344836,-0.2031284273,0.1617668569,-0.1379484981,-0.2202107906,0.1556535363,-0.012902773,-0.0332173966,0.1238514185,0.1579966545,0.0604770258,0.1138710529,0.1917484552,0.4745639563,-0.1944722235,0.241783008,-0.2040428221,0.008096097,0.1869803667,0.2431642711,-0.2253530622,0.2110218406,0.0181420688,-0.2957251966,-0.249424383,-0.0688087642,0.3604046702,-0.1840324849,-0.2736568451,-0.1092581972,0.1635110378,0.2349599153,0.3031378984,0.0220221132,-0.2748109996,-0.0233339462,0.0324436799,0.2525326014,0.2655858696,-0.1720371097,-0.0808958188,-0.134687975,0.2549236119,-0.4111463428,0.0948213711,-0.2713313699,-0.3557106853]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/388","title":"\ud83d\udc1b [Dataset] Cannot download wmt14, wmt15 and wmt17","comments":"similar slow download speed here for nlp.load_dataset('wmt14', 'fr-en')\r\n`\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 658M\/658M [1:00:42<00:00, 181kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 918M\/918M [1:39:38<00:00, 154kB\/s]\r\nDownloading: 2%|\u2589 | 40.9M\/2.37G [04:48<5:03:06, 128kB\/s]\r\n`\r\nCould we just download a specific subdataset in 'wmt14', such as 'newstest14'? ","body":"1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code:\r\n```\r\nnlp.load_dataset('wmt14','de-en')\r\nnlp.load_dataset('wmt15','de-en')\r\nnlp.load_dataset('wmt17','de-en')\r\nnlp.load_dataset('wmt19','de-en')\r\n```\r\nThe code runs but the download speed is **extremely slow**, the same behaviour is not observed on `wmt16` and `wmt18`\r\n\r\n2. When trying to download `wmt17 zh-en`, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz","comment_length":38,"text":"\ud83d\udc1b [Dataset] Cannot download wmt14, wmt15 and wmt17\n1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code:\r\n```\r\nnlp.load_dataset('wmt14','de-en')\r\nnlp.load_dataset('wmt15','de-en')\r\nnlp.load_dataset('wmt17','de-en')\r\nnlp.load_dataset('wmt19','de-en')\r\n```\r\nThe code runs but the download speed is **extremely slow**, the same behaviour is not observed on `wmt16` and `wmt18`\r\n\r\n2. When trying to download `wmt17 zh-en`, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz\nsimilar slow download speed here for nlp.load_dataset('wmt14', 'fr-en')\r\n`\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 658M\/658M [1:00:42<00:00, 181kB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 918M\/918M [1:39:38<00:00, 154kB\/s]\r\nDownloading: 2%|\u2589 | 40.9M\/2.37G [04:48<5:03:06, 128kB\/s]\r\n`\r\nCould we just download a specific subdataset in 'wmt14', such as 'newstest14'? ","embeddings":[-0.2754454911,-0.1981760114,-0.0195056945,0.4694994688,0.0656870753,0.1942017078,0.0270849541,0.2684627771,0.086118564,-0.031171849,-0.2506634593,-0.0679134354,0.2125768811,0.0307872146,-0.062822625,-0.1890521795,-0.1460027099,0.1014119461,-0.2369921207,-0.0269567817,-0.2904830277,0.2036181539,-0.090696156,0.127126202,0.2401779741,0.0780032203,-0.0501666181,-0.1968962699,-0.1596407294,-0.2748236656,0.5160508156,0.3601162136,0.0586538464,0.1450440586,-0.0001219386,0.0825733617,0.1835669577,-0.1369367689,-0.363604337,-0.138734743,-0.5023856759,-0.3797949255,-0.0570703857,-0.045134332,-0.0286948029,-0.4139443934,0.294800967,-0.1618006527,0.2084503323,0.3149540126,0.1214062124,0.191299662,-0.1842961311,0.0588121191,0.2260121107,-0.2973448932,-0.15996328,0.1628722101,0.0911368802,-0.0998708978,0.1882558018,0.0394495241,0.105443202,0.4685547948,0.0881750509,0.1837196797,0.0343793817,-0.3379814327,0.1303819716,0.4677025378,0.7240515351,0.0134222656,-0.5606609583,0.1094950885,-0.1864440739,-0.1793483198,0.1590444297,0.2385326922,-0.3889275491,-0.0097060362,-0.2580251098,0.1541268826,-0.2150075138,0.5520458817,-0.1538664401,0.6958835721,0.2310094684,0.1035004184,0.2030011714,-0.0791506842,-0.0816324279,-0.1097037196,0.0649881288,0.1140191108,-0.4582254291,-0.1419677138,-0.2928094864,-0.0565237701,0.0961091667,0.037047755,0.2588021159,0.3004488647,0.0732925534,-0.0901720747,0.2149017006,-0.0372556001,-0.1363024712,-0.1615839452,0.3071226478,0.0234277286,0.1086892188,0.1278395057,-0.0146291135,-0.0734967291,-0.3333629966,0.0429438651,-0.0641035438,0.0900537521,-0.4714164734,-0.1830093265,-0.172557205,-0.3095319271,-0.263366878,0.2749702632,-0.147243157,0.2098654658,0.0350211672,0.1641625017,-0.0373844989,-0.2366257906,-0.0851307809,-0.0776000842,-0.1194321811,0.0852759331,0.3873597682,0.2008733898,0.1021284536,-0.1979434341,-0.4624218345,0.1757552177,0.0462515987,-0.1504184306,-0.3103292286,0.2144239396,0.022912899,0.5044224858,0.0043044649,0.267393291,-0.2623713315,-0.0306956433,0.0474637002,0.1335317343,0.1591810733,0.0375398397,-0.1781826764,-0.0758744255,-0.411654979,0.1583167166,-0.1979249269,0.0384485349,-0.3133223057,-0.0542283878,-0.4436590672,-0.0637915805,0.2325396687,0.3726833463,-0.2943882048,-0.1404788494,-0.3939944208,-0.0772324055,0.2761011422,0.4165425003,-0.0116093047,0.0389969349,-0.1322283447,0.0135205025,0.612870872,-0.0660050809,-0.5310536623,0.4080584943,-0.3529969454,0.0949895233,0.2438896894,0.2227171361,0.0412154607,-0.0381213464,0.1125601828,0.5967074037,-0.0655452013,0.0432799943,0.033988364,-0.391399622,0.4888458848,0.1582465619,-0.0716146007,0.1904319674,0.0406712964,0.2375050634,0.4658760428,0.0016447369,0.1256796867,-0.0367695279,-0.076817669,-0.1869158,0.0371690392,0.0805691332,-0.8039488196,0.3596593142,-0.042655237,-0.0757531524,0.0591385812,0.0498416498,-0.16872181,-0.3384788334,0.1157402471,0.1946621388,0.0570870303,0.4249056578,-0.0243317038,-0.0735513195,0.0306157544,0.2424025834,0.1400818974,0.0745666996,0.0931107998,0.014225984,0.0210619085,0.1914711744,0.5240718126,-0.232347995,0.1202669069,-0.1981445998,-0.0477989577,0.0404798798,-0.1348661929,0.3278014064,0.0472766086,-0.1049361378,0.0062304628,-0.3015390337,-0.0232193898,0.5032980442,0.0925630406,-0.2129791379,-0.0748015568,0.2086768299,0.3825483024,0.3309425116,0.3173202872,0.1293129772,0.3910722435,-0.1069186926,0.1999790072,-0.173958376,0.2182213813,0.1208617911,-0.0325229466,-0.0482301228,-0.1440731436,0.4479544163,0.4699537456,-0.0467506498,0.035925664,0.299310565,0.0086816559,-0.0588148348,-0.2114624977,0.4395283461,0.4815129042,0.0023181334,0.33396855,-0.0354880765,0.2888124585,-0.4605615735,0.2112386674,0.1993149221,-0.3576081693,0.1509036571,0.0544255674,-0.0624086,-0.1587812304,0.0544615164,0.3649361134,0.337857753,-0.1209440306,0.1075226367,-0.6640239358,-0.5608876944,-0.1469938904,-0.145015344,-0.1277522445,-0.2961667478,-0.0341589265,0.3210790455,-0.1442078799,0.0491754673,-0.0051479996,0.1736545414,-0.0359969102,0.0824435428,-0.3095356822,-0.0627546608,-0.2631098926,0.0413994379,0.3632288277,0.2588679492,0.3428502977,-0.3035113811,-0.2165745795,-0.2507110834,-0.087889187,-0.162899971,-0.1206063554,0.044017747,-0.0509420671,0.6214644909,0.0594322719,-0.2801193595,0.0967657641,0.0080840215,-0.1996394396,-0.0294790789,-0.0728666931,0.0812951103,-0.0819146112,-0.4950372279,-0.5639942288,-0.4488963485,0.3042041361,-0.0260782875,0.4307365417,0.180479601,-0.4492824674,-0.0245239902,0.2001229227,0.1783764362,-0.0993988961,0.211172998,0.5299037099,0.0487380959,-0.3680869341,0.1324007064,0.184414804,0.1965640336,0.5387356877,-0.5226668119,0.1722503304,-0.1968161315,-0.3020075858,0.2634525299,-0.2586895823,0.1561907977,-0.2869668007,-0.1237122044,-0.0530794226,-0.1602719724,0.0036501093,0.1923757643,0.264264226,0.1058426574,0.2024157941,0.0867900923,0.3931768835,0.2769284248,-0.0128255757,0.4106549025,0.0292589981,-0.1475881338,0.1673332006,-0.2676241696,0.0524177402,-0.1229718179,0.3658302426,0.4481184781,-0.2541319728,-0.1831549853,-0.3033004999,0.0940262377,-0.0100965556,-0.3930917978,0.0041977428,0.0122559192,0.0525957644,0.214644134,-0.0400631391,-0.0332001895,-0.31058833,0.0108706811,0.2935233414,-0.0451610759,0.1850490272,-0.0280553047,-0.0978086963,-0.549272716,0.2894978523,0.0763230994,0.7073537111,-0.1326804459,0.1520878226,0.0723669529,-0.0145369954,0.3030672967,-0.1729931682,0.4643155634,0.1467847228,-0.1720147282,-0.3408595324,-0.1419171095,-0.1690539271,-0.1852245182,0.3035874069,0.3525131345,-0.5763117671,-0.3394087255,0.0019276398,0.1817183942,-0.1563268751,0.0998178646,-0.114884682,-0.4248751402,-0.0797263831,-0.1973153353,-0.1418590546,0.1513551772,-0.0258544981,-0.1020517349,0.1112805605,-0.1125811189,0.1008933708,0.096813865,0.0111806393,0.410004884,0.2851949334,0.1245302334,0.1230281517,0.1208341941,0.4883982241,0.2290777713,0.0528079942,0.2032563984,-0.0611119159,-0.1043682992,0.579274714,0.0183543433,0.2489416152,-0.0697890744,0.1338157356,-0.0625192523,0.1462149769,0.2862613201,-0.3440874517,-0.2097570002,-0.326593101,0.4035044312,0.0848550349,-0.104962036,0.3273226023,-0.0833917558,-0.2306993455,-0.0298580844,0.1237998903,0.9365190864,-0.1632585526,0.1926649511,-0.0968296155,0.0193079319,0.3620097935,-0.0062121549,0.1284442097,-0.0610414445,-0.3517689705,-0.066842787,0.1129168496,-0.0246465597,0.334556371,-0.2419375926,0.4412826002,0.1158205867,0.3251469135,0.4289510846,0.1513634324,0.0347838141,-0.0403638929,-0.2547502816,0.0458967164,-0.1982335001,0.4878913462,-0.0211248379,-0.3443233967,-0.1988645494,-0.0988534167,-0.4196380079,0.0415865928,-0.3956823647,0.0358497128,-0.3102172017,-0.0945119187,0.1591805816,0.0962362364,0.2075096816,0.0768227428,-0.456687212,0.0620992035,-0.6321641207,-0.0321627557,0.0685845241,-0.1318035722,-0.1424509734,-0.196486339,-0.114329949,-0.1696117371,0.1090627611,-0.2224170715,-0.07567624,-0.102371417,-0.4478806257,0.2547486722,-0.1786811948,-0.1608711183,-0.2535403669,-0.1225440577,0.0466517322,-0.1645885855,-0.0397009626,0.2860139608,0.1273254007,-0.1963328719,-0.0019032944,0.6198271513,0.002909488,0.1054777876,0.3953410983,0.0506549887,-0.1604822278,-0.2704815865,-0.3811042309,-0.1060793921,-0.1625458449,-0.0781910121,-0.2101413161,0.0140310461,-0.006038019,0.4558546841,-0.0392552651,0.168202579,0.0280108619,-0.6202651262,-0.1609710008,0.1318268031,-0.1765398979,0.1199000701,-0.5219163895,0.1192381904,0.1135824919,0.4160388708,-0.2092128992,0.056569647,-0.3388041258,0.0471051522,0.2880390286,-0.070682615,0.2289940417,0.1886253357,-0.0341754667,0.0251633674,-0.3593530655,-0.054240685,-0.0218978263,0.2447007596,0.1939218938,-0.3757940531,0.2078461647,-0.2565253377,-0.0859248489,-0.0188445859,0.0220741741,-0.111956656,0.0593064837,0.0382560603,0.2219023407,0.1416695863,-0.2665583491,0.1743684262,0.0305271391,0.3759011924,0.1019108221,0.2991654873,0.0468899496,-0.0848335996,-0.4554961026,-0.2001157701,0.1027134508,0.1104805321,0.3243230283,0.0250203125,-0.0200866535,0.0019965477,0.1979020834,0.5765169263,0.01475533,-0.0014471045,0.1004846022,0.0253083091,-0.2922025323,-0.1287103891,0.4113741815,-0.020107165,-0.0409840643,0.0618020967,0.1289478093,-0.2663394213,-0.0094034411,-0.2022638619,0.4215307236,-0.0747956038,0.0920509845,0.374851644,0.1693941504,0.0775663033,-0.0218249112,0.0485169254,0.1603427529,0.2448494583,-0.2588815689,0.2282999903,-0.1019001678,0.0528459698,-0.100081332,-0.3135136366,-0.1546791047,0.6282542944,-0.0257671382,0.0091527831,0.2310025841,0.0964127108,-0.0531193167,-0.2443241924,-0.3528076112,0.1790960878,0.2587441504,-0.1399997175,-0.227678135,-0.0766251534,-0.4428304136,0.1516772956,-0.145724535,-0.4017068446,0.2690563202,0.2768545747,0.0600415505,-0.282273829,-0.3852895796,-0.1095987111,-0.0400090404,-0.3327306211,0.1910598427,0.0114703737,-0.1317567825,0.190714702,0.1674356163,-0.0386508219,0.1761793494,0.0461453311,0.1767426729,-0.0575563796,-0.0859023407,-0.0975413695,0.2834539711,0.2556301653,-0.1936446875,0.4711665213,-0.0174480956,-0.0536398701,0.1863858849,0.1968060136,-0.2136816382,-0.1679443568,0.475125581,-0.1047998667,0.0741505027,-0.0653725192,-0.1176452711,-0.2411349863,-0.1365991533,-0.0488958396,-0.107115224,0.3333055973,0.0484650955,0.0064600315,-0.1535184383,0.9997263551,-0.0436173081,0.2473966032,-0.2425633371,-0.3746013641,-0.624622643,0.0966064855,-0.232196942,0.089879483,0.2122482955,0.1397720724,0.1016815081,0.2786849737,0.0979442447,0.5578716397,-0.2036718279,0.2489025742,-0.4591287076,-0.3036428988,0.0157663431,0.4082073569,-0.2625909448,-0.4670158029,0.4813464284,-0.3863253593,-0.0498682633,-0.0533837005,-0.0902344063,0.2436767071,0.3555136919,0.3386852145,0.0117222629,0.6352945566,0.0285194218,-0.194727391,-0.3011881113,-0.0414866135,-0.1197318137,-0.0060221185,0.1648761034,0.1999556422,0.0487308986,0.4505467415,-0.2935015857,0.2260513902,-0.2680616677,0.2887162268,0.1474458128,0.1920249909,-0.1904368997,0.0886309892,0.1265806705,-0.0310887173,-0.0599888377,-0.1564525962,-0.5516497493,-0.2328882515,0.5788881183,-0.2664411366,-0.2547086775,-0.1597991139,0.3558369577,-0.0425936095,0.3084096313,-0.0320449919,0.2063061148,0.1983070523,-0.0943489224,-0.3142537177,0.2210626006,0.0923448205,-0.3264966905,0.0395807065,0.0355664752,0.1923960149,-0.4283826947,-0.2856146991,-0.2166663706]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/388","title":"\ud83d\udc1b [Dataset] Cannot download wmt14, wmt15 and wmt17","comments":"> The code runs but the download speed is extremely slow, the same behaviour is not observed on wmt16 and wmt18\r\n\r\nThe original source for the files may provide slow download speeds.\r\nWe can probably host these files ourselves.\r\n\r\n> When trying to download wmt17 zh-en, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz\r\n\r\nLooks like the file`UNv1.0.en-zh.tar.gz` is missing, or the url changed. We need to fix that\r\n\r\n> Could we just download a specific subdataset in 'wmt14', such as 'newstest14'?\r\n\r\nRight now I don't think it's possible. Maybe @patrickvonplaten knows more about it\r\n","body":"1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code:\r\n```\r\nnlp.load_dataset('wmt14','de-en')\r\nnlp.load_dataset('wmt15','de-en')\r\nnlp.load_dataset('wmt17','de-en')\r\nnlp.load_dataset('wmt19','de-en')\r\n```\r\nThe code runs but the download speed is **extremely slow**, the same behaviour is not observed on `wmt16` and `wmt18`\r\n\r\n2. When trying to download `wmt17 zh-en`, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz","comment_length":97,"text":"\ud83d\udc1b [Dataset] Cannot download wmt14, wmt15 and wmt17\n1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code:\r\n```\r\nnlp.load_dataset('wmt14','de-en')\r\nnlp.load_dataset('wmt15','de-en')\r\nnlp.load_dataset('wmt17','de-en')\r\nnlp.load_dataset('wmt19','de-en')\r\n```\r\nThe code runs but the download speed is **extremely slow**, the same behaviour is not observed on `wmt16` and `wmt18`\r\n\r\n2. When trying to download `wmt17 zh-en`, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz\n> The code runs but the download speed is extremely slow, the same behaviour is not observed on wmt16 and wmt18\r\n\r\nThe original source for the files may provide slow download speeds.\r\nWe can probably host these files ourselves.\r\n\r\n> When trying to download wmt17 zh-en, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz\r\n\r\nLooks like the file`UNv1.0.en-zh.tar.gz` is missing, or the url changed. We need to fix that\r\n\r\n> Could we just download a specific subdataset in 'wmt14', such as 'newstest14'?\r\n\r\nRight now I don't think it's possible. Maybe @patrickvonplaten knows more about it\r\n","embeddings":[-0.268609345,-0.1404747516,-0.0071865544,0.5020000935,-0.0168406237,0.2251569629,-0.0530777313,0.2993670106,0.0911353305,-0.0368140861,-0.2252270728,-0.0799275488,0.1161911115,0.0058771251,0.034304779,-0.2081655562,-0.1193282083,0.0341802686,-0.3183064461,0.0080130184,-0.3116195202,0.1468855292,-0.0868213475,0.0733703375,0.2240645885,-0.0069648223,-0.2052784413,-0.0254717804,-0.1000093371,-0.3394281864,0.4278948009,0.2947577834,0.084403567,0.2282148302,-0.0001149252,0.0709944889,0.1664440781,-0.1815565974,-0.3427716792,-0.0916524976,-0.509786129,-0.4445599318,-0.1338346601,-0.0916885734,-0.0498927794,-0.4900558889,0.2561153769,-0.1568859816,0.3314899802,0.3418523967,0.1644345075,0.2418355197,-0.1462133229,-0.0085078701,0.2681643069,-0.241802901,-0.1191282272,0.1768724173,0.0858175084,-0.0647064373,0.2019879371,0.0539438315,0.0507229492,0.4143848717,0.0648013875,0.114726074,-0.0066977637,-0.2883557379,0.0657300279,0.4004980624,0.707028687,0.0280685835,-0.5350027084,0.094114162,-0.0312393866,-0.124926351,0.1994022578,0.2483088821,-0.2839077115,0.023875311,-0.2142187357,0.0485210232,-0.1946800798,0.5462664962,-0.0335705765,0.6662647128,0.2300116271,0.0512684174,0.0850208998,-0.0345224477,0.0830066949,-0.1274167895,0.0934071094,0.0012820574,-0.3166339695,-0.0942261145,-0.2909419537,-0.0173591822,0.040531341,0.0083770044,0.2951494753,0.2358808815,0.0183149688,-0.0712846443,0.1781585515,-0.1028664187,-0.1670800298,-0.2394867986,0.3152788579,0.1925608367,0.1304365695,0.1407724321,-0.0634528548,-0.1405701041,-0.4265267253,0.1794076413,0.0678066015,0.0819200277,-0.3032383025,-0.2359872609,-0.0952850729,-0.2585288584,-0.2333587557,0.2315392792,-0.2232052982,0.3033592701,-0.01407394,0.0879307836,-0.0782802776,-0.3754115403,-0.1115830392,-0.0619362295,-0.0947364196,0.1375948638,0.4606338143,0.2327832133,0.2125827819,-0.1855624467,-0.324811846,0.1220188588,0.0827296302,-0.0089063961,-0.2034354061,0.2094967663,0.0389030725,0.5146356821,-0.0152603425,0.2459966391,-0.2595425546,-0.0184405837,-0.0653082058,-0.0278567933,0.1655278355,0.0922636688,-0.1611418277,-0.1466825753,-0.3369729519,0.0296972711,-0.1942420602,-0.0796273872,-0.3151355982,-0.0935668498,-0.334267199,-0.1286198795,0.2433601469,0.4242477119,-0.3487033546,-0.0085138381,-0.3759478927,-0.0351442434,0.2923084199,0.3570439219,-0.0795979276,0.0716927201,-0.1589744389,-0.1769257188,0.5281008482,-0.1343882829,-0.4725863039,0.270416677,-0.3284674287,0.1723009497,0.246023193,0.1870347112,0.0267481897,-0.083505325,0.1047134697,0.6350846291,0.0012760299,0.0515114032,-0.0706807077,-0.3561582863,0.3798070848,0.1910879463,-0.0478560701,0.1234174594,-0.0262463279,0.2235040963,0.4299486279,-0.006066205,0.2387932241,-0.0190680958,-0.1557723433,-0.1321374476,0.0291817784,0.0744898766,-0.7336922884,0.3186699152,-0.1053289399,-0.165256232,0.0810761154,-0.0034324117,-0.2751204967,-0.3440809846,0.0790603012,0.097276628,0.1050883904,0.5249676108,-0.041287411,-0.0831199959,0.0002133563,0.1862969249,0.1373163313,0.0402249508,-0.0418928415,0.0532663576,-0.0445458777,0.171317786,0.5440144539,-0.2567563951,0.1717381179,-0.2130444944,-0.0371350423,0.1358809024,-0.1739643812,0.3611807823,0.1119710281,-0.1792988628,0.0926525816,-0.3793899119,0.0370959714,0.5412628055,0.0641174167,-0.1362142265,-0.072694622,0.1897853315,0.14487046,0.2603428662,0.292127192,0.1436413825,0.4012096822,-0.141046375,0.1060652286,-0.0886812508,0.1613156497,0.1243789494,-0.1367560625,-0.0958056077,-0.1503305286,0.4106875062,0.4148617685,-0.0372205786,0.0187883023,0.2357937843,-0.0668034405,-0.0749669373,-0.224123776,0.5326936841,0.4717298448,0.0562893786,0.3155216277,0.027367577,0.2627882063,-0.4930698276,0.2585000098,0.2068519592,-0.344170332,0.2035879642,0.0771230608,-0.1378411949,-0.1374786347,0.071620442,0.3822630644,0.2573137581,-0.1986294985,0.0536603667,-0.6793519258,-0.5351989865,-0.3434390128,-0.2511228621,-0.1482522488,-0.305246979,0.1282158345,0.2785882354,-0.2804341018,0.1202675849,0.0317885987,0.2259991914,-0.1738285571,0.1984794587,-0.2139069885,-0.1356356591,-0.2548263371,0.0976746306,0.358640343,0.337623179,0.2934270799,-0.3645837009,-0.1957132667,-0.3108948171,-0.0325513594,-0.1599157453,-0.0036146466,0.1372742951,-0.0390132852,0.593483448,0.1324960589,-0.3987885416,0.1563427597,-0.0093738837,-0.0448995493,-0.0634622052,-0.096499823,0.0993852839,-0.0341018587,-0.623754859,-0.590438664,-0.523463726,0.1860347688,-0.0353248976,0.4430011809,0.1396968961,-0.3087504506,-0.0415654331,0.1024816632,0.1844176352,-0.0566680357,-0.0280259289,0.5375439525,-0.1619543135,-0.3553983867,0.1727953702,0.2059616596,0.200956434,0.4632624984,-0.5046476126,0.0646782294,-0.1197670549,-0.2729349136,0.1870685965,-0.1981201172,0.1420211941,-0.1899036169,-0.1756812483,-0.0921619833,-0.110205248,-0.0474175699,0.1542581618,0.3609259427,0.1070156842,0.0580693409,0.0540272072,0.5281319618,0.2144051343,0.0288450476,0.382537514,0.0698562115,-0.0967571065,0.2006230056,-0.2211503983,0.055933401,-0.1493368596,0.3521571457,0.4451558888,-0.177445814,-0.1599802673,-0.3324055374,0.0135215819,-0.1174127609,-0.412101388,0.0676093325,0.0765770897,0.0848159045,0.2502114177,-0.0907302871,-0.104770638,-0.2486215532,0.1401176751,0.2866019905,-0.0554878153,0.2225299031,-0.1112919599,-0.0839064345,-0.4219074249,0.3272365332,0.0696538538,0.6666907668,-0.0727573559,0.0959879085,0.0811181217,-0.0427734181,0.3269474506,-0.2589514554,0.3664771318,0.1476517916,-0.0270378646,-0.3202750087,-0.1805524081,-0.155113861,-0.1363304555,0.2854500115,0.241364181,-0.5075770617,-0.2647195458,-0.0120119005,0.3046756387,-0.2263693362,0.1271538138,-0.0926806629,-0.4546771646,-0.1686352491,-0.2483867109,-0.1952324957,0.0476712361,-0.0636231378,-0.1080398336,0.1206831187,0.0130324783,-0.0089067919,0.1007082239,0.1134610549,0.4704127312,0.2650997043,0.1576219648,0.2695895135,0.2070246637,0.4462994337,0.1590937674,0.0876426399,0.1112572402,-0.0822960436,-0.0150394971,0.5899693966,0.0758285448,0.1944023669,0.0797456577,0.1139244512,-0.1158549935,0.1670591384,0.3086415827,-0.456677407,-0.1548491418,-0.3943791091,0.3080306649,0.0518216081,-0.1139824241,0.3615948558,-0.0987034589,-0.2175687551,0.0566775762,0.1134927049,1.0516214371,-0.0584438592,0.1815832108,-0.1388884485,-0.0930822343,0.4372522235,0.0056297383,0.2301346511,-0.040001478,-0.3949959874,-0.0398700237,0.1387973428,0.0006312727,0.3304264247,-0.309235394,0.4790844321,0.1480701119,0.2751792967,0.2732686698,0.0889242068,0.0763859823,-0.0982351378,-0.3572638929,0.0540135466,-0.2071112245,0.4592020214,-0.1164929569,-0.3337394595,-0.1740609407,-0.0624478012,-0.3034552932,0.0511226952,-0.2851750851,0.0560327731,-0.3349586725,-0.0503586531,0.1038979888,0.0765501708,0.2055120766,0.1284273714,-0.4939323962,0.1827997416,-0.5684784651,-0.0973134115,0.0150151169,-0.0919860676,-0.152451843,-0.2483957857,-0.1820304394,-0.1257848293,0.0838256776,-0.1685776263,-0.0482179895,-0.1796245128,-0.2898651659,0.2355901152,-0.140868783,-0.1040883809,-0.1262224168,-0.2186392993,0.1043441743,-0.2395486087,-0.1040244624,0.2727040052,0.1063077077,-0.174978286,-0.0320207775,0.6141952872,-0.0027181543,0.1257404834,0.5313014984,0.182005018,-0.1084022373,-0.2623317242,-0.320967555,-0.2857589722,-0.3247822225,-0.0723350793,-0.1397088021,0.0443736576,-0.0148887755,0.4144520462,0.0939488858,0.1857426167,-0.0060853837,-0.5259037018,-0.1334527433,0.1294777244,0.0421939604,0.0920006186,-0.4193423986,0.0345064141,0.1674450338,0.386277467,-0.2889705896,0.0907671154,-0.3084711134,0.1050012484,0.1838264316,-0.0147718284,0.1369226873,0.2186166346,-0.0207849555,-0.0205849968,-0.2606444955,-0.0791365802,-0.0208676457,0.2387755215,0.0886692554,-0.3916057944,0.1831959486,-0.2471767664,-0.0604224615,0.052607175,-0.0663637519,-0.1141838506,0.0583758764,0.0471803844,0.1464033723,0.0883586854,-0.255584985,0.2022084147,0.0496769957,0.3802536726,0.0688227341,0.265239954,0.0698115379,-0.1944235563,-0.3725170791,-0.2481484264,0.0905400366,0.2064654231,0.3934916556,0.0517015681,-0.0551675595,0.1035649478,0.2066735625,0.6060753465,0.0949704573,-0.0037310906,0.1826515496,0.0850056186,-0.3482971489,-0.1051263958,0.4268743992,-0.0493981428,-0.1335227638,-0.0182770453,0.1604173183,-0.2951154411,-0.0482874699,-0.1061281487,0.3336077034,-0.0446987525,0.1348261088,0.3830808997,0.1318676323,-0.0223079342,-0.1458292454,-0.011410933,0.1830568314,0.2075378001,-0.2937128544,0.2960051298,0.0022070375,0.0512290709,-0.0633131191,-0.324136436,-0.0649504736,0.523096323,0.1908030212,0.0568315163,0.2189496756,0.1626998633,-0.0995308459,-0.1670859307,-0.4335703552,0.3511622548,0.3017348051,-0.1393084228,-0.3634730875,-0.1048655808,-0.4123687744,0.189158082,-0.1128331125,-0.3035281301,0.3537469506,0.2533999085,0.0291214306,-0.4037004113,-0.32204023,-0.1665655673,0.010588605,-0.3628243804,0.1755483747,0.2881410718,-0.169610858,0.2712331116,0.27342996,0.0342609324,0.1940802932,0.0514925569,0.2307003289,-0.0529654287,-0.0396345668,-0.1296419948,0.3255562484,0.2726576924,-0.1567445099,0.4806818962,0.0442807153,-0.1215065122,0.1598025113,0.0724946931,-0.2559205592,-0.174317956,0.4113011658,-0.0269315429,0.0457281433,-0.0965430364,-0.1521426737,-0.2968739569,-0.1195234731,0.0169865787,-0.0918754861,0.2861560285,-0.0004839713,0.0248235762,-0.1885652244,1.0825707912,0.0145787392,0.1532885581,-0.2088084072,-0.4053679109,-0.696079731,0.143344,-0.2470806539,0.092423752,0.1040166244,0.1266988516,0.0964725912,0.2451936156,0.0724123567,0.4204584956,-0.1497842968,0.3062482178,-0.5024481416,-0.2226482779,0.0942267179,0.4402625561,-0.2074291855,-0.425110966,0.4701156914,-0.2363034636,-0.0323649645,0.0643197671,-0.1571582556,0.2465579212,0.2980615795,0.3182753325,0.0408845544,0.5406653285,-0.054589197,-0.2572309375,-0.4410834908,-0.1929121315,-0.0798035413,0.0345673375,0.1539821625,0.2596019506,0.0195875298,0.4092366695,-0.3668284118,0.2018212974,-0.2354795039,0.341622591,0.0719434172,0.0303546004,-0.2026437521,0.0776838884,0.2446888536,-0.0529658049,-0.1159354001,-0.1798537374,-0.5251284242,-0.2719207704,0.6197147369,-0.2890968323,-0.22164464,-0.2276801169,0.3758016825,-0.0810559317,0.3304470778,0.0215584636,0.230120495,0.2397793829,0.0174535569,-0.190307498,0.2094176561,-0.0343326628,-0.2718366683,0.029139502,0.0931023359,0.1694599092,-0.4575901031,-0.2863952219,-0.104282923]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/388","title":"\ud83d\udc1b [Dataset] Cannot download wmt14, wmt15 and wmt17","comments":"Yeah, the download speed is sadly always extremely slow :-\/. \r\nI will try to check out the `wmt17 zh-en` bug :-) ","body":"1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code:\r\n```\r\nnlp.load_dataset('wmt14','de-en')\r\nnlp.load_dataset('wmt15','de-en')\r\nnlp.load_dataset('wmt17','de-en')\r\nnlp.load_dataset('wmt19','de-en')\r\n```\r\nThe code runs but the download speed is **extremely slow**, the same behaviour is not observed on `wmt16` and `wmt18`\r\n\r\n2. When trying to download `wmt17 zh-en`, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz","comment_length":21,"text":"\ud83d\udc1b [Dataset] Cannot download wmt14, wmt15 and wmt17\n1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code:\r\n```\r\nnlp.load_dataset('wmt14','de-en')\r\nnlp.load_dataset('wmt15','de-en')\r\nnlp.load_dataset('wmt17','de-en')\r\nnlp.load_dataset('wmt19','de-en')\r\n```\r\nThe code runs but the download speed is **extremely slow**, the same behaviour is not observed on `wmt16` and `wmt18`\r\n\r\n2. When trying to download `wmt17 zh-en`, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz\nYeah, the download speed is sadly always extremely slow :-\/. \r\nI will try to check out the `wmt17 zh-en` bug :-) ","embeddings":[-0.3693395257,-0.2487271726,0.0216565207,0.3820067346,0.1480184346,0.0315845087,0.0206154753,0.1097690016,0.1508650482,-0.0224935636,-0.119229719,0.0813111216,0.2570133209,0.0221288633,-0.0504060537,-0.1498070061,-0.0636185184,-0.0127320979,-0.3244119287,-0.0993769243,-0.3538653851,0.2009574771,-0.2064668089,0.0905934945,0.2752312422,0.0832998753,-0.1313607991,-0.1531077623,-0.2954229116,-0.2548477352,0.3610233665,0.3276956975,-0.0559857078,0.2809772789,-0.0001131215,-0.0560203865,0.2967323363,-0.0383981802,-0.2685244679,-0.2018249631,-0.5110461116,-0.3942409158,-0.0794065818,0.0854354054,-0.0184098836,-0.2966769338,0.249693349,-0.0439682901,0.0649197176,0.2494037896,0.1683285236,0.2879309058,-0.06867522,-0.1455195248,0.2407011241,-0.2033242583,-0.3107456267,0.0399787426,0.0785966814,-0.179604575,0.1557213366,0.0584835187,0.0875241607,0.3362604082,0.1760083884,0.094273597,0.0510465316,-0.2017598003,0.0827950239,0.4504013658,0.7103342414,0.0851512849,-0.4957500994,0.2819173038,-0.0918947011,-0.2350519001,0.3443357646,0.3135761321,-0.2158723027,-0.0734717548,-0.4669174552,0.2572805882,-0.138332665,0.4892605841,-0.0957623422,0.7104789019,0.1332416981,0.1434031427,0.3129828274,-0.1106801853,-0.0816215351,-0.2723140717,0.0001883647,0.0334910639,-0.4814959466,-0.0322296843,-0.0658684671,0.0954648629,0.0754289627,0.1016091108,0.2850106359,0.2121489793,0.0210256781,-0.0517260022,0.2097490281,-0.021904029,-0.2130883783,-0.1640496552,0.4323970079,0.2450239211,0.1523038894,0.1677443534,0.1467083544,-0.0488005243,-0.2299163342,0.0242733099,-0.0505109094,0.0019589318,-0.4001110196,-0.1665736735,-0.1238353029,-0.2453398854,-0.164540872,0.3225750029,-0.2840417027,0.1698942333,-0.1212052405,0.1779886484,-0.0842853412,-0.2417668253,-0.1613672078,-0.1574996859,-0.2656621635,0.0487123169,0.5434920788,0.2756901085,0.1349683702,-0.0276945438,-0.4596246779,0.1299184412,0.1053118333,-0.1755293161,-0.1506625712,0.3081341982,0.0414034352,0.4511938691,-0.118147999,0.2141903192,-0.218013972,-0.0147893755,-0.1912533939,0.068316333,0.1833389252,0.1696573198,-0.1521883905,-0.1954449713,-0.1842024326,0.0742724016,-0.2544246316,0.0659201667,-0.2866135538,-0.1090282798,-0.3650815785,-0.1189389229,0.1813006848,0.3788802028,-0.3131741285,-0.045764599,-0.3016707897,-0.1056640968,0.3926259875,0.4808875918,-0.0214605518,-0.0690824315,-0.2611771822,-0.1133297682,0.4793394804,-0.0569392294,-0.5938494802,0.4285428524,-0.1727694869,0.0856167078,0.1779770851,0.1772713661,-0.0855822489,-0.0825825557,0.1270345747,0.7016557455,0.0059168912,0.0621071346,-0.0823496282,-0.3496221304,0.4960255623,0.0793752894,-0.1310658306,0.0629866794,0.077469945,0.2308818102,0.4280626178,-0.0492956378,0.12070667,0.2119649202,0.0061041862,-0.2326266766,0.1675234735,0.2613880336,-0.8036822081,0.3271984756,-0.1731810719,0.0420759991,0.0533733033,0.0986738876,-0.0930985138,-0.2325115949,-0.1148577854,0.1196191534,0.0992339775,0.4008775949,-0.0808944702,0.0385469273,0.2440878004,0.1637265384,0.1860145032,0.0721243173,0.0375178754,0.089004606,-0.1497063041,0.1078927368,0.4190086126,-0.2034531236,0.0905700102,-0.2255449295,-0.1061334312,0.0301307943,-0.118764393,0.3457844853,0.0342662819,0.0129582426,0.1846446246,-0.3644969165,0.0430598632,0.5446993709,0.1592684686,-0.2901196182,-0.2831042409,0.1604015976,0.2294861972,0.2556654811,0.2638523579,-0.028795328,0.3943089545,-0.0271758307,0.069723703,-0.0704719946,0.3118414581,0.1041933075,-0.1764555275,-0.104705222,-0.1786276996,0.383715421,0.7334485054,-0.0273740515,0.0445228145,0.2151580453,-0.009701008,-0.2793851197,-0.1021256074,0.2419556975,0.5307390094,-0.018456554,0.3182319105,-0.0075913663,0.0528219528,-0.4657371938,0.2658325434,0.0323507264,-0.2632800937,0.208351925,0.146182701,-0.1172621399,-0.2702292204,-0.0391779318,0.372279346,0.3183618188,-0.1137540638,0.0215991866,-0.5688726306,-0.4337557256,-0.0417039953,-0.0972280428,-0.232840389,-0.4373254478,-0.010616377,0.359289974,-0.1175308824,0.0834662691,0.0112375906,0.0518173613,0.0321936123,0.1122513264,-0.2232310176,-0.0508841649,-0.2936744094,0.0506423265,0.2501958609,-0.0048260638,0.3465156853,-0.2549163997,0.0043228078,-0.2484766245,-0.1828152537,-0.0734227747,-0.1328422874,-0.0097922077,0.0210055634,0.3756832182,-0.0484725758,0.0197904035,0.267724216,-0.1306687891,-0.0933632925,0.1500162184,-0.1475165784,-0.1163983941,-0.145531565,-0.4618680477,-0.4196478128,-0.3909377456,0.2283409983,0.0177119784,0.2647218704,0.2665109634,-0.2433954328,-0.0090891244,0.1340952963,0.1358187795,-0.2116517127,0.1096784174,0.5519052148,-0.0122113675,-0.4962178767,0.2889939845,0.3083995283,0.1282163411,0.4666349888,-0.5050539374,0.0814793184,-0.2245031744,-0.0868187249,0.2105652094,-0.0806289166,0.1803443134,-0.2014332861,-0.1320724189,-0.0635461286,-0.2266353667,0.0003429247,0.3354968131,0.2945871353,0.0575160049,0.0751543716,0.1130234003,0.2551252842,0.322781235,0.0355762355,0.450271219,0.0570785068,-0.0395572037,0.0555571578,-0.1616863161,0.0864985436,-0.150836125,0.4195181131,0.3306363225,-0.1776616573,-0.2543928921,-0.2211967409,0.2076511234,0.1560290903,-0.3032636046,-0.0362837315,0.0926279351,-0.0247281324,0.2582338154,0.0545312949,0.0281000584,-0.4963302314,0.1535634547,0.2811174393,-0.1159652844,0.1709709167,-0.0542916693,-0.0251252297,-0.5100089908,0.2622943223,-0.086937733,0.6925250888,-0.2452953607,0.1321940422,0.1541800797,-0.1334348172,0.3795636594,-0.3355473876,0.4085904062,0.2441133857,-0.0998255461,-0.3103894293,-0.2699892819,-0.3074089885,-0.2978402078,0.2070447057,0.4671503305,-0.403686136,-0.2595311701,-0.0028819384,0.2527556419,-0.0834332258,0.0329959579,-0.1655886769,-0.4258913696,-0.0759302527,-0.1960380524,-0.2923614383,0.079040423,-0.2044104487,-0.0717161745,0.1685197055,0.0046440326,-0.0387744606,0.0195355639,0.0290737245,0.2013516873,0.3527790904,0.0411571078,0.208811447,-0.0041826358,0.4881168902,0.1826081425,-0.0191791095,0.2027254105,-0.1937299371,-0.1419256479,0.4482376873,0.0616980419,0.1161122471,-0.0197453834,0.1926821172,-0.1204143688,-0.0646362826,0.4672897458,-0.3162603378,-0.2791849971,-0.22826235,0.2917951047,0.0922988579,-0.06323006,0.3406104445,-0.0752303302,-0.100624077,0.0542610772,-0.0727052316,0.8992088437,-0.0750673711,0.1676117778,0.0284452066,-0.0120264059,0.3487119675,-0.0787988603,0.2581682503,-0.1769597828,-0.2249809951,-0.003369336,0.0311147626,0.0274606533,0.210125953,-0.4092715979,0.5043522716,0.2017362118,0.2596512735,0.2110849023,0.0916044265,-0.0027352108,-0.0822545588,-0.0691424087,0.1338377148,-0.1257581562,0.5930079818,-0.0115368422,-0.3399750292,-0.0633043051,-0.0504248813,-0.3375661671,0.2289010882,-0.2211631536,0.1064875945,-0.2780903578,-0.0289234295,0.1129968315,-0.1082199812,0.116898492,0.1473509669,-0.4352402687,0.1044977531,-0.5686905384,0.1226675287,0.0455063507,-0.0656875595,-0.0766174197,-0.2632240057,-0.1547881514,-0.1998171359,-0.0494098216,-0.0731024295,-0.0469887629,-0.1141859144,-0.292728126,0.2307781577,-0.0337505154,-0.1279283613,-0.254648447,-0.095203191,0.1239215359,-0.0279910807,0.0628987849,0.4031952322,0.1070542634,-0.2656989396,-0.1343066096,0.61309129,-0.1519094408,0.1501207054,0.3498308361,0.1883085519,-0.2084124833,-0.3143332601,-0.3719800711,-0.0897525921,-0.4002477527,-0.0692304298,-0.1440845579,0.1001696885,0.1170555949,0.4755736589,-0.049025476,0.0558669567,0.0562261716,-0.5779898763,-0.0762098655,-0.0173782632,-0.0752179697,0.0603905357,-0.3815562725,-0.0613974147,0.3596575856,0.0912115276,-0.2972774506,0.0841393024,-0.2831428051,-0.0205845535,0.2991089225,-0.1970580369,0.139849633,0.1217768192,0.0317455009,-0.1153990254,-0.5138576627,-0.1349563301,-0.0126885241,0.1968260258,0.231201753,-0.3998675346,0.101338394,-0.2693153322,-0.1387002766,-0.0299089011,0.0238059908,-0.0697159469,-0.1134992316,0.0155442748,0.0159867853,0.1784756035,-0.0522683896,0.1789715737,0.0810708255,0.3983921707,-0.0322970785,0.2166847438,0.1420236081,-0.1306734979,-0.4640181065,-0.1250662655,0.004228509,0.1668862402,0.4557456076,-0.150338158,0.114596881,-0.1903169602,0.2894704342,0.5473506451,-0.0197061058,-0.1659722924,0.0818419233,0.1417401284,-0.2528591454,-0.3240988851,0.2707701921,0.2660083175,-0.0896599069,-0.1165638193,0.2668006122,-0.2275668085,0.0049268859,-0.1171082631,0.5208179355,-0.1662727743,-0.07325425,0.325098455,0.0912305713,0.1914716065,-0.045719631,-0.0515027046,0.0135421911,0.146106407,-0.1071897075,0.1941192299,0.1337234527,0.1880331486,-0.0064205951,-0.4491778612,-0.134198308,0.4470541179,-0.1082417071,0.0837220773,0.2773938477,0.0468393005,-0.186020121,-0.2336350083,-0.521222055,0.2492410392,0.1441927403,-0.2176211029,-0.1578579843,-0.0406396464,-0.383920759,0.3018143773,-0.2011140436,-0.1044839844,0.3324412107,0.1468765885,0.0592007078,-0.4044609666,-0.2451430261,-0.1018609107,-0.0117119122,-0.2720139027,0.1310622096,0.0434969552,-0.1083547324,0.0509570912,0.1166175455,0.1323673278,0.2466509491,0.044190079,0.1514762491,-0.0550393537,-0.0769833624,-0.1267118007,0.5055840611,0.3176569641,-0.3132899702,0.3788610995,0.0105594471,-0.1253325492,0.1614066809,0.0476467386,-0.1173021272,-0.0278121252,0.4096532464,-0.1731015742,0.0070706587,-0.0808529854,-0.0291830972,-0.2272978872,-0.0489756763,0.2235088199,-0.0314474888,0.3116567731,-0.1972492635,0.06267602,-0.1403964162,0.9632143974,0.1747051626,0.0081980284,-0.1843709052,-0.2348640561,-0.6164389253,0.1030555218,-0.2522733212,0.1988199204,0.0616462193,0.2748517096,-0.0309101529,0.2128812969,0.179409191,0.4960567057,-0.0012606987,0.3387595415,-0.3486464024,-0.1576447785,0.0671236292,0.2313790023,-0.3162714541,-0.5026994944,0.5336249471,-0.4530445933,0.0299827009,-0.1813534349,-0.0418832824,0.144813329,0.3595470786,0.3521796167,0.1409448832,0.5347114205,0.1001047567,-0.2366069704,-0.405269295,-0.0648823977,0.0872770026,-0.0066201109,0.062853016,0.0696176663,0.1808972955,0.4258285761,-0.3366212249,0.2417767793,-0.3410274684,0.3019993007,0.1489571333,0.2037635148,-0.1593160927,0.0825898349,0.1561184675,-0.0960129648,-0.0475367755,-0.2808299959,-0.466897428,-0.2708584368,0.3842181861,-0.1362174451,-0.2359120995,-0.3162315786,0.2579324841,-0.1017074436,0.2079438716,-0.0433062054,0.2353961766,0.179402262,-0.1781789064,-0.2937729061,0.2973264754,-0.0095803551,-0.1704115421,0.0426895395,0.1026007086,0.1412753761,-0.4580461979,-0.1470395774,-0.0394971296]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/388","title":"\ud83d\udc1b [Dataset] Cannot download wmt14, wmt15 and wmt17","comments":"Maybe this can be used - https:\/\/stuncorpusprod.blob.core.windows.net\/corpusfiles\/UNv1.0.en-zh.tar.gz.00 ","body":"1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code:\r\n```\r\nnlp.load_dataset('wmt14','de-en')\r\nnlp.load_dataset('wmt15','de-en')\r\nnlp.load_dataset('wmt17','de-en')\r\nnlp.load_dataset('wmt19','de-en')\r\n```\r\nThe code runs but the download speed is **extremely slow**, the same behaviour is not observed on `wmt16` and `wmt18`\r\n\r\n2. When trying to download `wmt17 zh-en`, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz","comment_length":7,"text":"\ud83d\udc1b [Dataset] Cannot download wmt14, wmt15 and wmt17\n1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code:\r\n```\r\nnlp.load_dataset('wmt14','de-en')\r\nnlp.load_dataset('wmt15','de-en')\r\nnlp.load_dataset('wmt17','de-en')\r\nnlp.load_dataset('wmt19','de-en')\r\n```\r\nThe code runs but the download speed is **extremely slow**, the same behaviour is not observed on `wmt16` and `wmt18`\r\n\r\n2. When trying to download `wmt17 zh-en`, I got the following error:\r\n> ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/tfdataset-data\/downloadataset\/uncorpus\/UNv1.0.en-zh.tar.gz\nMaybe this can be used - https:\/\/stuncorpusprod.blob.core.windows.net\/corpusfiles\/UNv1.0.en-zh.tar.gz.00 ","embeddings":[-0.4666007161,-0.0613560677,0.0023681885,0.4255926013,0.1058961526,0.133838743,-0.0265779458,0.1100055873,0.0580001138,-0.0522339679,-0.169115752,-0.0292323157,0.2012997866,0.0357220434,-0.0244321432,-0.2486982495,-0.0328836106,0.0769558772,-0.217603907,-0.0220446289,-0.4356454015,0.2392529249,-0.2933826745,0.0945538431,0.240624994,-0.0319010206,-0.1926941872,-0.0864934027,-0.3402494788,-0.2953885198,0.3519860208,0.3146460652,0.0374115035,0.1983935088,-0.0001145176,-0.0429031402,0.2685977519,-0.04811465,-0.2277691811,-0.079460457,-0.4012271166,-0.3013664484,-0.1191498116,0.0467054062,-0.0131716076,-0.2537424862,0.2601158023,-0.0001000612,0.0355314389,0.3332296014,0.1611624807,0.2702199817,-0.0221949294,-0.117894426,0.2502974272,-0.2581876516,-0.2501865625,0.0643365681,0.179655835,-0.1400509775,0.1953553259,0.0005709926,-0.017755691,0.3899977803,0.080531314,0.063500151,-0.068867214,-0.2570843697,0.147614032,0.3344986141,0.764074266,0.0066799019,-0.4397119582,0.2990726829,-0.1124202162,-0.2320476472,0.3541903496,0.2407447696,-0.2296599299,-0.1351990551,-0.4725728333,0.2694756985,-0.2086679637,0.5205147266,-0.0445180312,0.5666949749,0.0938964337,0.1219365224,0.2959192991,-0.1112491563,-0.0812748,-0.2708155513,0.0594923124,-0.0092367725,-0.4464229345,-0.037219625,-0.2082065493,0.189777866,0.020389054,0.0854922533,0.2617764473,0.1755915433,0.1349416524,-0.1173703298,0.2032254189,-0.0194081236,-0.2278195024,-0.2299954891,0.4067288935,0.1974112689,0.170475319,0.1864100546,-0.039568767,-0.1772484481,-0.1098113433,0.0406732373,-0.0129887257,-0.0093733091,-0.4950144589,-0.2188034654,-0.1693468988,-0.2173136026,-0.1086784825,0.3028250337,-0.3126793504,0.1243705377,0.0040061828,0.1276586205,-0.1338665485,-0.2005779445,-0.1162716672,-0.1359542459,-0.1946766078,0.0091701373,0.6756917834,-0.0172886681,0.1485700458,-0.1273227036,-0.4028507769,0.1848618537,0.1063684225,-0.2742691934,-0.1779565662,0.2314437926,0.1238908991,0.4180956185,-0.0910945833,0.1528509259,-0.2064498812,-0.0777916461,-0.1035776511,0.1094210073,0.2027736455,0.1692954898,-0.0990190729,-0.2027347833,-0.2961498201,0.0439949743,-0.2672415376,0.0396854356,-0.2928656042,-0.1664354354,-0.4583885074,-0.1830777079,0.1547240168,0.4288588762,-0.3360509574,-0.0657329783,-0.1612694561,-0.2079932988,0.3316022456,0.4356063008,0.063566722,-0.0053470992,-0.3079103231,0.0296895504,0.3882247508,-0.1383147836,-0.5457426906,0.4495818615,-0.1542133093,-0.0101283221,0.1319589913,0.1761517227,0.239777118,-0.1085446402,0.2533723116,0.5714196563,0.0101019014,0.0304253362,0.0400365293,-0.3045075834,0.3632073402,0.1846137494,-0.1487131417,0.0693118274,0.1320024282,0.3089618683,0.314409554,0.0396297313,0.1628246158,0.2686784267,-0.1065037325,-0.2389067411,0.1001253352,0.0598803908,-0.8175389767,0.3065367937,-0.1576181352,0.0675251782,0.1331083775,0.0999251232,-0.1627536118,-0.2339722216,-0.1265041679,0.2137860656,0.1036847681,0.3931070864,0.1441348344,-0.0512548387,0.2752964497,0.1206082702,0.1050050035,0.0326467156,0.1331348568,0.1215276942,-0.1902012378,0.0204206444,0.4915887713,-0.181082204,0.1363354921,-0.1901906431,-0.0213839728,0.0325609371,-0.2342304438,0.2999821305,0.0553995222,0.1566597074,0.1822638959,-0.4134092033,0.082453303,0.5565155745,0.1757244319,-0.3349374533,-0.1443365961,0.0920455381,0.1424973011,0.2936713994,0.2617965341,0.0039287717,0.3575453162,-0.0429026224,0.059315484,-0.0653856471,0.3796262145,0.1431232691,-0.2010332793,-0.074128367,-0.1900877953,0.279961437,0.6111729741,-0.0732423216,-0.0264868326,0.184697032,-0.0857961848,-0.1542007476,-0.13632752,0.1732329577,0.526487112,0.0387819745,0.3562061787,0.0598637648,0.0457015783,-0.4712703526,0.302927047,0.0726208687,-0.2301009595,0.2235997915,0.0895609111,-0.0534994677,-0.2557745278,-0.0365489721,0.3899645209,0.2839386761,-0.1171491444,-0.0262920242,-0.4886805415,-0.629472971,-0.1720690876,-0.0745235682,-0.238415584,-0.2411802411,-0.0399925336,0.3652430177,-0.0300402436,0.0339388326,-0.1037243679,0.1031991616,0.0280215684,-0.0424675681,-0.1451074481,0.1279887855,-0.3604713678,0.0705997273,0.3344529569,0.0067410148,0.4462576807,-0.2747804821,0.0830640346,-0.3141248226,-0.0735383108,-0.0553230084,-0.0525957532,0.0219690651,0.0413315371,0.3803696334,-0.0178158116,-0.1309403479,0.2207988054,-0.1874160618,-0.1490741074,0.0157591347,-0.1586076915,-0.184755221,-0.0594357848,-0.4955085218,-0.438559562,-0.409178853,0.1519368887,0.0612728558,0.2962222695,0.1990162581,-0.3098751009,0.0670580342,0.2222115099,0.2446000129,-0.1341544539,0.1136305183,0.4582787454,0.0566735826,-0.4710108042,0.3163897693,0.3371765316,0.1441592276,0.5189720392,-0.51771456,0.1336816549,-0.18561773,-0.1059645042,0.3072526455,-0.0534154885,0.2130165547,-0.1570073366,-0.1598539054,-0.018902,-0.3163732588,-0.0550406761,0.2644567192,0.3524834514,0.1583206803,0.0535951927,0.059332408,0.3443747163,0.4277781546,-0.0257617347,0.4741599858,0.0145105906,-0.0987564549,0.1632784009,-0.2226633877,0.2058523744,-0.2298802286,0.4414470196,0.4476431608,-0.1520387381,-0.2186822146,-0.2386313826,0.140114069,0.0020816219,-0.2633108497,0.003622706,0.0923362076,0.1753289402,0.1463951766,-0.0493734442,0.0873506293,-0.4806034565,0.163287282,0.2913583517,-0.1385707557,0.1525577605,-0.0734654963,0.0225849096,-0.5621097088,0.2744087875,0.0444196872,0.7035296559,-0.2736819685,0.2382984757,0.0849390626,-0.1048118323,0.5278481841,-0.3615759611,0.3902440071,0.2764299214,-0.0402739272,-0.2225923687,-0.1659364402,-0.377423048,-0.162786901,0.334584862,0.4514190853,-0.3873175085,-0.2304129899,-0.0300839357,0.2824923098,-0.0977773592,0.0551307537,-0.1315275878,-0.4986647367,-0.1390952319,-0.1682984382,-0.2623143494,0.1745508313,-0.1291919649,-0.0758251026,0.1085601747,-0.0288300235,0.0603536367,-0.0565669388,0.0538741387,0.3381936252,0.328915894,0.0625045002,0.3259230554,0.0484659225,0.490265727,0.0506290533,0.0383595861,0.2268975228,-0.1783871353,-0.0750123784,0.4279144108,0.0235303268,0.0692642182,-0.0017453897,0.1042837724,-0.2650980055,0.0351510569,0.4809899926,-0.2582345009,-0.2828359008,-0.2392141819,0.2677893043,0.2261020541,-0.0525793172,0.2820131779,0.0294661764,-0.1538190544,-0.0434579998,0.0781157687,0.8634960055,-0.180053547,0.1421076655,-0.110574156,-0.1038483083,0.3253965676,0.0819702372,0.273062706,-0.081037581,-0.2923512757,-0.0483249947,0.0486743897,0.0102099739,0.3253326416,-0.5037299395,0.4735594094,0.307885468,0.2132321745,0.2085909098,0.1502442062,-0.0031120859,-0.1221671626,-0.1153078675,0.1216704324,-0.1279052496,0.5467486978,-0.130095005,-0.3056496978,0.0675363466,-0.0720773786,-0.3883793652,0.2326243222,-0.3751081526,0.0267789811,-0.2497193962,0.0638916418,0.2462059259,-0.0315251835,0.1500549614,0.0388397239,-0.4534516931,-0.0286192372,-0.4751510024,-0.0155142304,0.0770030767,-0.0851046592,-0.0438640974,-0.2844151556,-0.0837757066,-0.1429452002,0.0077879946,-0.1373119354,-0.0131706037,-0.1223895326,-0.3630295694,0.1073791683,-0.0712114722,-0.1527114511,-0.2790620029,-0.0664930865,0.1036381647,-0.1187373996,0.0717411861,0.386192888,0.1634884328,-0.2094940245,-0.0731083378,0.6155091524,-0.1546706706,-0.0361317843,0.4306474924,0.2267646939,-0.218045041,-0.3049408197,-0.4710821211,-0.2946903408,-0.2718067467,-0.0492691249,-0.0283425543,0.1675621271,0.1458733827,0.5644654632,-0.0137320235,0.0831334293,0.1004635394,-0.5186864734,-0.0880046189,0.0210318565,0.0337810144,0.125545159,-0.3910222948,-0.0159874875,0.3957294822,0.0156520698,-0.289016217,0.0750275329,-0.3587237,-0.0312111378,0.2326279134,-0.1354850233,0.1271000504,0.2095196247,0.002544377,-0.1377437562,-0.4032511115,-0.1117336079,0.0600842908,0.2018311322,0.1848031133,-0.4610686004,0.0989811495,-0.3225543797,-0.103336595,0.0123262545,-0.0191835444,-0.0371627584,-0.1343818754,0.0202263929,0.0156652983,0.1556862295,-0.0553453602,0.1889215559,0.0663505197,0.454133153,0.0070444383,0.2712911069,0.1489361376,-0.1690037698,-0.4802488983,-0.0793563277,0.0873715281,0.0383815877,0.3267272711,-0.1208321676,-0.0047568893,0.0569125786,0.2977977991,0.5481067896,0.001830377,-0.0544270352,0.0461375751,0.1286758929,-0.2309959978,-0.1828455925,0.1455365717,0.1395457089,-0.078193903,-0.0596849397,0.2257874757,-0.2729021013,-0.0132355932,-0.1360345185,0.531797111,-0.1282051206,-0.095912151,0.3437308967,0.1919166297,0.1439755857,0.004623408,0.0291340966,-0.0421081297,0.0837981999,-0.0746286586,0.2262224406,0.0050438568,0.2671247721,-0.0603700243,-0.328199625,-0.1146647111,0.4739073515,-0.0054962998,0.1060701981,0.2176535428,0.0589424744,-0.1985302418,-0.2444370836,-0.4860773981,0.2050886899,0.1518384516,-0.266895473,-0.3174819052,-0.0984217599,-0.3014617562,0.3047569096,-0.1459598094,-0.2385136783,0.3212547302,0.0957228318,-0.0346895978,-0.4882730544,-0.244638592,-0.0321246721,-0.0550636873,-0.2307218164,0.109587796,0.0756054893,-0.0115583269,0.1405265331,0.3029154539,0.1966401935,0.2305775434,-0.0160138458,0.1773269922,-0.1596584767,-0.1018638462,-0.1503671557,0.4437405169,0.3445143104,-0.2712525725,0.4085499048,0.0120973019,-0.094181858,0.1403252929,0.0490347594,-0.2444388419,0.0344616175,0.4765079916,-0.1876094341,0.0643342212,-0.1110891327,-0.0428417958,-0.2532640398,-0.0266383942,0.2420438379,-0.1905178279,0.3474434018,-0.1663864106,0.0505164117,-0.0618080199,0.892185688,0.2175399363,0.285348326,-0.2153155655,-0.3089250922,-0.5948107243,0.0221262351,-0.1380529553,0.1228102148,0.0901532546,0.2581024766,-0.1211829633,0.1904658079,0.0350764431,0.4303418994,0.0055368529,0.2804634273,-0.3552380204,-0.1496738493,0.1436246485,0.2648054361,-0.2855787277,-0.5215306282,0.4857106507,-0.3605967462,-0.0116998395,-0.2700028419,0.0085837441,0.0157868378,0.3362442851,0.2387595773,0.1370200962,0.5750887394,0.0202181656,-0.1045267656,-0.3998678923,0.020465564,-0.0519523807,-0.0593110733,0.1064574048,0.0089080669,0.1259327233,0.4813604951,-0.2127563506,0.0239031259,-0.2664682567,0.4105964899,0.1236577481,0.0974123776,-0.1524470001,0.0064910199,0.1195622459,-0.0281883422,-0.0381923877,-0.3053188622,-0.5110045671,-0.3026279807,0.3595905304,-0.1777718365,-0.163205266,-0.2878600061,0.244081676,-0.2060917765,0.2657542825,-0.1180186421,0.2053561956,0.2602130175,-0.2088374645,-0.3659065366,0.3373946846,-0.1006749496,-0.1585582644,0.0977754816,0.1110186726,0.1371422112,-0.4317889512,-0.1539323181,-0.1013736501]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/387","title":"Conversion through to_pandas output numpy arrays for lists instead of python objects","comments":"To convert from arrow type we have three options: to_numpy, to_pandas and to_pydict\/to_pylist.\r\n\r\n- to_numpy and to_pandas return numpy arrays instead of lists but are very fast.\r\n- to_pydict\/to_pylist can be 100x slower and become the bottleneck for reading data, but at least they return lists.\r\n\r\nMaybe we can have to_pydict\/to_pylist as the default and use to_numpy or to_pandas when the format (set by `set_format`) is 'numpy' or 'pandas'","body":"In a related question, the conversion through to_pandas output numpy arrays for the lists instead of python objects.\r\n\r\nHere is an example:\r\n```python\r\n>>> dataset._data.slice(key, 1).to_pandas().to_dict(\"list\")\r\n{'sentence1': ['Amrozi accused his brother , whom he called \" the witness \" , of deliberately distorting his evidence .'], 'sentence2': ['Referring to him as only \" the witness \" , Amrozi accused his brother of deliberately distorting his evidence .'], 'label': [1], 'idx': [0], 'input_ids': [array([ 101, 7277, 2180, 5303, 4806, 1117, 1711, 117, 2292,\r\n 1119, 1270, 107, 1103, 7737, 107, 117, 1104, 9938,\r\n 4267, 12223, 21811, 1117, 2554, 119, 102])], 'token_type_ids': [array([0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0])], 'attention_mask': [array([1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,\r\n 1, 1, 1])]}\r\n>>> type(dataset._data.slice(key, 1).to_pandas().to_dict(\"list\")['input_ids'][0])\r\n<class 'numpy.ndarray'>\r\n>>> dataset._data.slice(key, 1).to_pydict()\r\n{'sentence1': ['Amrozi accused his brother , whom he called \" the witness \" , of deliberately distorting his evidence .'], 'sentence2': ['Referring to him as only \" the witness \" , Amrozi accused his brother of deliberately distorting his evidence .'], 'label': [1], 'idx': [0], 'input_ids': [[101, 7277, 2180, 5303, 4806, 1117, 1711, 117, 2292, 1119, 1270, 107, 1103, 7737, 107, 117, 1104, 9938, 4267, 12223, 21811, 1117, 2554, 119, 102]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]}\r\n```","comment_length":69,"text":"Conversion through to_pandas output numpy arrays for lists instead of python objects\nIn a related question, the conversion through to_pandas output numpy arrays for the lists instead of python objects.\r\n\r\nHere is an example:\r\n```python\r\n>>> dataset._data.slice(key, 1).to_pandas().to_dict(\"list\")\r\n{'sentence1': ['Amrozi accused his brother , whom he called \" the witness \" , of deliberately distorting his evidence .'], 'sentence2': ['Referring to him as only \" the witness \" , Amrozi accused his brother of deliberately distorting his evidence .'], 'label': [1], 'idx': [0], 'input_ids': [array([ 101, 7277, 2180, 5303, 4806, 1117, 1711, 117, 2292,\r\n 1119, 1270, 107, 1103, 7737, 107, 117, 1104, 9938,\r\n 4267, 12223, 21811, 1117, 2554, 119, 102])], 'token_type_ids': [array([0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0])], 'attention_mask': [array([1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,\r\n 1, 1, 1])]}\r\n>>> type(dataset._data.slice(key, 1).to_pandas().to_dict(\"list\")['input_ids'][0])\r\n<class 'numpy.ndarray'>\r\n>>> dataset._data.slice(key, 1).to_pydict()\r\n{'sentence1': ['Amrozi accused his brother , whom he called \" the witness \" , of deliberately distorting his evidence .'], 'sentence2': ['Referring to him as only \" the witness \" , Amrozi accused his brother of deliberately distorting his evidence .'], 'label': [1], 'idx': [0], 'input_ids': [[101, 7277, 2180, 5303, 4806, 1117, 1711, 117, 2292, 1119, 1270, 107, 1103, 7737, 107, 117, 1104, 9938, 4267, 12223, 21811, 1117, 2554, 119, 102]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]}\r\n```\nTo convert from arrow type we have three options: to_numpy, to_pandas and to_pydict\/to_pylist.\r\n\r\n- to_numpy and to_pandas return numpy arrays instead of lists but are very fast.\r\n- to_pydict\/to_pylist can be 100x slower and become the bottleneck for reading data, but at least they return lists.\r\n\r\nMaybe we can have to_pydict\/to_pylist as the default and use to_numpy or to_pandas when the format (set by `set_format`) is 'numpy' or 'pandas'","embeddings":[0.2560274899,-0.0724572092,-0.2367637008,-0.0092357192,0.2625888288,0.0842923298,0.6678315401,0.3210130334,0.0850154981,0.1096882522,-0.2230844349,0.631875813,0.046179045,-0.0179825127,0.0116416439,-0.1747506559,0.0956746191,0.2547401786,-0.2903570533,-0.0689310506,-0.283067435,0.0156398285,-0.0143754175,0.0470164716,-0.0316810086,-0.1229436621,-0.243011266,-0.066965647,-0.0099635795,-0.2817721665,-0.1102737114,-0.0522085726,-0.1565473825,0.1077477857,-0.0000965855,0.0117588332,0.1007464156,0.1284487993,-0.2778981626,0.0988120884,0.2208989561,-0.4964868724,0.128321141,-0.3898863494,0.0194244478,-0.3709512949,-0.0011586872,-0.2352722436,0.2233627439,0.2378044575,0.3248372078,0.0824997649,0.0406746008,-0.0091811223,-0.1530109197,-0.0536514074,-0.111673519,-0.096805796,0.0740232095,-0.000344872,0.0224829968,0.1968286335,-0.2312167436,-0.0540923066,-0.1202843636,0.0783430561,0.1750875115,-0.2016627789,-0.0591094345,0.1322260797,0.0710335746,-0.2181085497,-0.2725657225,-0.3579931557,0.0110659432,-0.1044410616,-0.2819867134,0.1497367173,-0.0419386439,0.304756701,-0.0567268282,0.141591832,-0.1975571215,0.1329679042,-0.2234634757,0.1334752887,-0.00387744,-0.0269974023,0.1181257516,-0.2326640189,0.2993658781,-0.1540308446,0.2643842101,0.0086971587,-0.0715571567,-0.133841604,0.1959290653,-0.3873053789,-0.0322724395,-0.3329618573,0.183025524,-0.0538950451,-0.0235340986,0.1319846958,0.1248031706,0.2315275073,-0.0103136888,0.1086452603,-0.0331914015,-0.0222173836,0.1428206414,0.1067358926,0.0208146535,0.047676295,0.1039289162,0.0543530323,0.2067093253,0.0948465019,-0.3995526731,0.0214920491,-0.4321007133,0.3727926016,-0.1416668147,0.0871335492,-0.0056861765,0.5033516288,0.05709479,0.078409791,0.0120771769,0.3128516376,-0.2222591639,0.2659431696,0.0089575863,-0.2215346247,-0.0210270155,0.3906449676,0.0046083741,0.0056062392,0.2724509835,0.052951891,0.0090363808,-0.0570253506,0.398491621,0.1581862122,-0.0183216892,0.1029567644,0.3305059373,-0.273571074,-0.073408559,0.3235862255,0.0897550732,0.09800338,-0.0730490088,0.3501078188,0.0267446786,-0.1353610754,0.1204932183,0.196056366,0.0979915559,0.0391967669,-0.0627012551,-0.3927683234,0.1285284311,-0.2342426628,-0.0258016828,0.0015649783,-0.4275617898,0.1042091325,0.0402691923,-0.0326838046,0.2190034539,0.3847593069,-0.0153800594,-0.0311420728,-0.1035962626,0.1959885359,-0.0700909197,-0.2117768377,-0.1431392729,0.025964573,0.1356393546,0.3213233054,0.2309161872,0.1676853299,0.1878184676,0.0333486423,0.1708783209,0.0854336321,0.0253945105,0.1654042155,-0.2166432291,-0.200807184,0.217568621,-0.0501669012,-0.0277145114,0.3013141155,-0.1478921324,-0.075238511,0.2646343112,-0.3522586226,-0.0863965824,-0.1085294634,0.1263146996,-0.1856984198,-0.0706584454,-0.0009579193,0.0929225385,0.0091781514,-0.1311029196,0.3257482946,-0.2217811942,-0.4243864715,-0.3495079875,0.0457587838,-0.0132237077,0.02768231,0.4024395049,0.0568417311,0.098162964,-0.082914494,0.1711259037,0.0485419556,-0.1781453192,-0.0365344621,-0.2771191597,0.0510211736,-0.1610230505,-0.0616619512,-0.0102373334,0.2658472657,0.0410777479,0.1367579401,0.2536705434,0.1568620354,0.2253478616,-0.1363914758,0.0789509043,0.1731404364,0.1421353817,-0.0107404422,0.3149241209,0.5039960146,0.1378080547,0.0506746434,-0.0394132398,0.5265206099,0.1004746482,0.0673764423,0.1498181969,0.2438424081,0.2283335477,0.2267895043,-0.022183869,-0.1647676975,-0.2045824379,0.0883951709,-0.0372102074,0.0783450082,-0.5608398914,0.3296601772,0.3871723115,-0.1254360825,0.2982020974,-0.042947948,-0.1170924529,-0.1296837926,0.0092824018,-0.0854976699,0.0647060201,0.3047472239,0.1346231252,-0.1011198908,-0.246352464,-0.1573565155,0.171603322,0.1403886676,0.1497249454,-0.0519284531,0.2285126746,-0.212198332,-0.2140843123,0.0572399236,-0.2752458155,0.1016200408,-0.2288697362,0.2389459163,-0.3486661315,-0.2187672853,-0.0047962326,-0.164686799,0.2367773652,-0.4387485087,0.1186256036,0.0831531063,-0.0962845609,0.027575884,0.3894482851,-0.0102326907,0.3071250319,0.112402685,-0.1375089139,-0.2112655789,0.0176344551,0.3510561585,-0.0325058028,0.1101428121,0.1001262739,-0.0634950548,-0.0570035428,-0.1335255355,-0.2751377225,0.0031088397,-0.2275829762,0.1885088682,0.2113188505,0.2826877832,-0.2487779707,-0.0493518785,-0.071722433,-0.0535933971,-0.0864409059,0.1011063978,0.0296371207,-0.0127767259,-0.1391559243,-0.273807615,-0.3708685338,-0.2461612523,0.2457753569,0.19725734,0.1448776424,0.04192717,0.37204656,0.3241763413,0.2355500162,0.2128996551,-0.125249356,0.1298516542,0.2862127423,-0.1785424948,-0.3247752786,-0.3495488167,-0.0403357595,-0.0918325037,-0.2265648842,-0.0471110865,-0.4609470665,-0.1354733258,-0.0125896325,0.1935326755,-0.032664001,0.1490452886,0.2680903077,-0.3067965209,-0.0502039045,0.0971967503,-0.0428387299,0.0432678424,-0.0840690583,-0.0630861968,0.0501085632,0.1526055336,0.2194743305,0.0977593884,-0.1959163845,-0.1312121451,-0.0152707053,0.3280456066,-0.1606306136,-0.1010646895,-0.2410317212,-0.1682165265,-0.0981954187,0.1248301044,-0.1472046673,-0.3451806903,0.0037209918,0.237668708,-0.3825841546,-0.0968593806,0.113186419,-0.2335439175,-0.0478622131,0.0416291803,0.2179597169,-0.2125125229,0.0127611812,-0.1765966564,-0.036039874,0.2504291832,-0.1270272285,-0.372066766,-0.2742795646,-0.3555977345,0.3058631718,0.2006749511,0.1969309002,0.1245522052,-0.4118640125,-0.0682600588,0.0190742593,0.4703346193,0.0821754187,-0.3371404409,0.364630878,0.0205653626,-0.0961635113,0.1042672619,-0.563534677,0.0093350159,0.1641109139,0.1090123579,-0.2452834845,-0.0595848784,0.081137374,0.2119965553,0.0118747381,0.1769613922,-0.1113035679,-0.3024594486,-0.261457175,0.0198083911,0.1983464509,0.1258126348,-0.4265234172,-0.2748203576,-0.1618731916,-0.2216855139,0.3759464622,0.1774932444,0.0346638374,-0.040534202,-0.1933378279,-0.2015725225,0.2310631722,0.156595394,-0.0567222647,-0.2254025936,-0.1200210676,-0.0981596932,-0.0551369637,0.1993043274,0.2418480963,-0.0896878242,-0.0564583056,-0.2413827777,-0.0663809404,0.1501811892,-0.1460464895,0.2211691588,0.0700615793,-0.0276516508,-0.3291964531,0.1209775805,0.0463762842,-0.1051248536,-0.0222852752,-0.1152670309,-0.2081453502,0.2638191581,0.0325118266,0.8158717155,-0.1056176797,0.3005525172,0.4592902362,-0.197716549,0.4725379646,0.2285956144,0.1159961447,-0.2518949211,-0.0576714613,-0.0594796166,-0.0100465035,0.2043711096,0.2532420754,-0.2561308444,0.1157730743,-0.2948709428,0.1316436976,0.0482266843,-0.0555580519,0.0385022312,-0.1773585975,-0.5897296667,0.3352954388,0.0547802262,-0.4174698293,-0.0300410073,-0.1306803524,-0.3652647436,0.1664578319,0.1722037047,0.0149755022,0.2371016145,0.2857072949,0.1130525693,-0.3973706067,-0.3194504976,0.2990199029,0.0782362744,0.1430559903,0.0597052835,-0.1594385356,0.1249621883,0.109646365,0.2293771207,-0.0577199385,0.4965383708,0.1031076387,-0.0500581786,0.1760503054,0.0092719067,-0.2180434018,-0.1885130405,-0.0271958765,0.0076899366,-0.3465358913,-0.1482536942,-0.0516144335,-0.4196206629,-0.1843434125,0.2770851851,-0.0314754397,-0.0995997414,0.3277488947,0.0246171486,-0.0796904042,-0.0120317796,0.1132595986,0.4137775898,0.0345138125,0.0591794066,0.0701439381,-0.3847639561,-0.2387193143,0.0825428367,-0.1526095122,0.0703601539,0.4293036461,-0.1121812686,-0.1417976469,-0.1393560171,-0.1004749611,-0.1253191382,0.3728622496,-0.4571977556,-0.204630807,-0.2589762211,0.1818985343,0.0911662728,0.3353270292,0.238316983,0.1342772543,-0.2093559951,-0.2403194159,-0.5028528571,-0.1735637784,-0.3184006512,0.1316963136,0.0348731764,-0.2332936525,-0.1063106358,0.0622402765,0.3776398003,0.1625671685,-0.0079252115,-0.4519948363,0.204461664,0.0235300846,-0.0102157798,-0.1632574946,0.2116341591,-0.0201328266,-0.2478336245,-0.2464574426,0.0121443942,0.0365635045,-0.0795863122,0.2235228866,0.0653869212,0.3287332654,0.3674655259,-0.1215731949,-0.2407366484,-0.0627778992,-0.0520085357,0.1585796028,-0.0030759752,-0.0591703691,0.0468243361,0.2976479232,-0.0031876278,0.0774445832,0.4809102416,-0.0406188443,-0.1691295207,-0.0404672325,0.3440822661,0.0845605507,-0.2110838443,-0.0958569422,0.0081233773,0.4007232189,-0.3073301911,-0.1081484258,0.157155633,-0.3483892679,0.2309493721,0.4312076867,0.0835086703,-0.0116918823,0.1095966548,-0.0501530021,0.0815821812,-0.2662101388,-0.1164510548,0.582059741,0.006645292,-0.1017887145,0.0711218193,-0.1162680313,0.1665239483,0.491871655,0.0209923927,0.4034590721,0.1135592237,0.0254196431,0.2546972334,0.3929942548,0.0771278515,0.0758874267,-0.1733088493,0.2135417461,0.2939058244,0.069291234,-0.2093723416,0.1442483515,-0.0291089099,0.3133695722,-0.2688260376,-0.129201144,-0.1011277661,-0.0936127678,0.014550765,-0.3355529606,-0.0351876616,-0.2210866809,0.2059135288,-0.1674287468,0.04022865,-0.0685614049,0.2408025116,-0.1409084797,0.3462824821,-0.0516201667,-0.2799859345,0.2119690776,0.2059757859,-0.0634924322,0.2458054423,0.4413476884,0.0201792084,-0.2248957902,-0.003479253,-0.2018452138,-0.4503282607,-0.0309195332,-0.026562538,0.0975788757,0.1408422738,0.1832570136,0.3068815768,-0.2546873093,-0.0303884596,0.152246058,-0.1255204827,-0.0353383459,0.3983753324,-0.0006993397,-0.4177678823,-0.0795597807,0.0275207013,-0.4694454968,-0.3235495985,0.2452063262,0.0198499486,0.2174038142,0.0547990762,0.2119052708,-0.0290133152,0.2661048174,-0.0585661046,0.1766676605,0.0587322973,0.1991707534,-0.1691031158,0.1327243745,0.2316051126,-0.1686920524,-0.148295179,0.0618930161,0.1804656088,0.2957518995,-0.319784075,0.0363380946,-0.0032569016,0.0950066075,-0.1532213837,0.0503708385,-0.082219474,-0.0826199725,0.0740317702,-0.1644277722,0.0799148679,-0.1453085542,0.3205461502,-0.1464013159,0.0949231312,-0.0931243598,0.4012173116,-0.129218474,0.3858845234,0.2196268141,-0.181461677,-0.0560146831,0.2221431881,-0.0839765295,-0.1339957267,0.5022016168,-0.0601774007,0.0746876374,-0.0050452249,-0.1248677149,0.2421612293,0.2362212241,0.2632470727,-0.3227277994,-0.2985230684,0.2755728066,-0.2742536366,0.111961022,-0.1134265959,0.2221007347,0.1778130233,0.0509324521,-0.1403018683,-0.4273499846,0.1870024651,-0.096857138,-0.4538885355,-0.0789670795,-0.0686431751,-0.1896493435,0.0528111458,-0.3690462708,-0.1206938624,0.2917222977,-0.2359176278,-0.0586578585,0.133746624,-0.0355449617,0.1161901057,-0.2565558553,-0.1335965097,0.0595044196,0.0598586537,-0.224807784,-0.3436225057]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/378","title":"[dataset] Structure of MLQA seems unecessary nested","comments":"Same for the RACE dataset: https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/race\/race.py\r\n\r\nShould we scan all the datasets to remove this pattern of un-necessary nesting?","body":"The features of the MLQA dataset comprise several nested dictionaries with a single element inside (for `questions` and `ids`): https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/mlqa\/mlqa.py#L90-L97\r\n\r\nShould we keep this @mariamabarham @patrickvonplaten? Was this added for compatibility with tfds?\r\n\r\n```python\r\n features=nlp.Features(\r\n {\r\n \"context\": nlp.Value(\"string\"),\r\n \"questions\": nlp.features.Sequence({\"question\": nlp.Value(\"string\")}),\r\n \"answers\": nlp.features.Sequence(\r\n {\"text\": nlp.Value(\"string\"), \"answer_start\": nlp.Value(\"int32\"),}\r\n ),\r\n \"ids\": nlp.features.Sequence({\"idx\": nlp.Value(\"string\")})\r\n```","comment_length":19,"text":"[dataset] Structure of MLQA seems unecessary nested\nThe features of the MLQA dataset comprise several nested dictionaries with a single element inside (for `questions` and `ids`): https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/mlqa\/mlqa.py#L90-L97\r\n\r\nShould we keep this @mariamabarham @patrickvonplaten? Was this added for compatibility with tfds?\r\n\r\n```python\r\n features=nlp.Features(\r\n {\r\n \"context\": nlp.Value(\"string\"),\r\n \"questions\": nlp.features.Sequence({\"question\": nlp.Value(\"string\")}),\r\n \"answers\": nlp.features.Sequence(\r\n {\"text\": nlp.Value(\"string\"), \"answer_start\": nlp.Value(\"int32\"),}\r\n ),\r\n \"ids\": nlp.features.Sequence({\"idx\": nlp.Value(\"string\")})\r\n```\nSame for the RACE dataset: https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/race\/race.py\r\n\r\nShould we scan all the datasets to remove this pattern of un-necessary nesting?","embeddings":[-0.1161423251,-0.0593214743,-0.0939640701,0.3199597001,-0.0490284264,0.2495961785,0.087612316,0.2466984987,-0.1752906442,-0.1441369653,0.0130557083,0.197159186,0.058465369,0.6652584672,-0.037603993,-0.0648569986,0.2087631822,-0.0377437025,-0.056361977,0.1263812631,-0.2246391177,0.1446332335,-0.2206718177,0.0209032241,-0.0983920097,-0.1649568379,-0.3148606718,0.1034778133,0.052318912,-0.2634041607,0.1510075182,0.240365535,-0.1195746213,-0.0443268009,-0.0001109656,-0.0709206089,0.1456373781,-0.0216371175,-0.0548776016,0.0413485505,0.0320792012,-0.1099435091,0.0695012212,-0.2508447468,-0.0989763662,-0.2140037864,0.0711577013,-0.3372648358,0.2594273388,-0.0382847153,0.2054678649,-0.0469221212,-0.1306341141,0.3519122005,-0.1220628396,0.1027288139,0.0172574706,-0.0262117665,0.2594551742,-0.1247475371,-0.0430970453,0.415356338,-0.1193582788,-0.1493011266,0.2905679941,0.0684557483,0.1848945171,-0.596270442,-0.0432923436,0.4557463527,0.2367852926,-0.2617447674,-0.1849277169,-0.4822607934,-0.0259912424,-0.2272670269,0.1633026153,0.1675083935,0.0069016796,0.2347602099,0.1263682395,-0.1401169151,0.0433410294,0.0321894847,-0.2222628295,0.4312478602,0.1123895496,0.0557990298,-0.1916797012,-0.1940625161,0.3292733133,-0.2776329219,-0.1152416542,-0.0648504049,-0.2658900917,-0.2681444287,0.145891428,0.0540675484,0.0904277116,-0.0091617191,-0.1017989889,-0.1273285747,0.0415140092,-0.1156784967,0.4433220327,0.0349118933,0.0427645966,-0.3553722501,-0.0186075065,-0.11106354,0.0128494827,0.0441047587,0.1379095465,0.1818000823,-0.5502961874,0.0230321009,0.4006924629,0.1182039306,-0.359192878,-0.1037032604,-0.6785646677,-0.1690202504,0.0719344467,0.137252897,0.1910883784,0.2438910604,-0.0796461776,0.090106979,-0.1095337421,-0.1471690536,-0.1280268729,0.0261902623,-0.0045824051,-0.1100144163,-0.1038460433,0.2253595591,0.0583133064,0.0152876982,0.0062835799,0.2510578036,0.1159466356,-0.1665658504,0.2065866441,0.2225798517,-0.0297140665,0.0133785838,-0.0154654486,-0.0512965992,-0.2058114111,0.2050140053,-0.0776299611,-0.2186259031,0.0134882564,0.2043150067,-0.1696726829,-0.1015770286,0.1816952676,0.7562408447,0.3734673858,-0.0927801728,0.0911263749,0.009046712,-0.4748316705,-0.2003763467,-0.031352248,0.3942876458,-0.3361498117,0.0484468639,-0.3053285182,0.0333992802,-0.0127937766,0.3472161591,-0.0665526241,0.339394927,-0.2069299817,0.2875927687,-0.0089720543,-0.3737544715,-0.242327258,0.2417011708,0.0202056449,0.2550999224,0.2007426918,0.0926874056,-0.13739869,-0.1837859005,0.1889773756,0.3555290997,-0.1046040282,-0.0923256651,-0.0993929431,-0.1669281274,0.6348803639,-0.0364637971,-0.4140782654,-0.2391839176,-0.343290031,-0.0578938536,0.3980113268,-0.0094857067,-0.1960019618,-0.1509506404,0.1513853073,-0.1476749629,-0.0829645693,-0.1501871347,-0.5748469234,0.0441733375,-0.2565105557,0.1166458204,-0.1293932199,-0.4134356081,0.310305357,-0.0182228591,-0.3776371181,-0.1487938017,0.223158747,-0.2980284393,-0.1173890755,0.0568259954,-0.2501930296,0.0233592167,0.0178599674,0.0614118539,-0.3382733166,0.2788203061,-0.1605370939,0.308401227,0.3650185764,0.69320786,-0.0535367467,-0.1911392063,0.2040143311,0.1118071303,-0.0075117047,-0.1424809545,0.0454091579,0.2737965584,0.3954080641,-0.2320894599,0.0044339946,0.3511698544,0.0159224402,-0.2095693797,-0.2848925889,0.4127983451,0.0557215177,0.349927187,0.0185294058,0.2035997957,-0.0105989622,-0.1132051125,-0.0705682337,-0.3927462697,0.1088526025,0.1086640731,0.0308570322,0.426107645,-0.2589320838,0.8457382321,0.465623796,-0.0576035343,0.0363926813,0.0054506445,0.1375107765,-0.4387932718,0.2027046084,0.1513318121,0.282648325,0.1742970794,0.0417098217,0.2460756153,0.2267715335,-0.2295657694,0.3983789682,0.0990012512,-0.0684945509,0.1255247891,0.3096333146,0.1429218799,-0.4245000482,-0.2349423021,-0.0185276307,0.0328291021,-0.2017996162,0.2663801014,-0.3401319385,-0.8277870417,0.0951259658,-0.1595639139,-0.3452222645,-0.414337039,0.1041768119,-0.1909134388,-0.1400655657,0.2615066469,0.3718318343,0.3114873469,0.0528241023,0.1535454839,-0.0749019086,-0.0934202448,-0.1512100995,0.1434291303,0.1969043463,0.180300355,0.2717166543,0.2582640648,-0.0109019019,-0.1724998802,-0.6402465105,0.2058470547,-0.4038881958,0.3266704679,0.1748285741,-0.0207620133,-0.2076142728,-0.2606464028,-0.0391047262,0.6991737485,-0.2577943802,-0.0002876561,-0.0311224088,-0.1765300333,-0.1504601687,-0.3982402086,-0.4945943356,-0.2323882133,0.2652852535,-0.200856179,0.4566139579,0.2072576582,-0.3831195831,-0.2453195155,0.0288470536,-0.134417966,-0.1962102652,0.1302180588,0.1351571381,-0.0603557937,-0.1612295508,-0.3660825789,-0.2903484106,0.3013529181,-0.2624475658,-0.438018173,-0.2253845781,-0.1168083847,0.1511100382,0.226504758,-0.0880563706,0.2082458586,0.3800465167,0.0148182008,0.2296254635,-0.0620111972,0.1058310494,0.1495172232,0.3227373064,0.0125862993,0.437213093,0.1064922437,0.2702871263,0.6254138947,0.196948722,0.0795066059,0.1210876182,0.2560326755,-0.2426040024,-0.2051140815,0.4185655415,0.1404145807,-0.0588461868,0.3224636912,-0.1500507444,-0.0727324858,0.1247329786,0.1506258845,0.1947640777,-0.2797316611,0.0122810956,-0.5949560404,0.2844963372,0.0216043852,-0.0077404245,0.1875219941,-0.2130334675,-0.0691158697,0.1293254346,0.156049937,-0.0014966285,-0.7965765595,-0.6033447385,-0.5060996413,0.0577493571,-0.0378534123,0.1418628842,-0.0846328735,0.065848276,0.0711490586,-0.0676314756,0.5189670324,-0.2795136273,-0.2345725596,0.2563893795,0.0427499972,-0.157079339,0.0150379594,-0.012921582,0.4084734023,0.2966897488,-0.0276692566,-0.3415409029,-0.2420085371,0.1911635101,0.128614068,-0.4076527059,-0.3557876945,-0.1688462049,-0.2850779593,-0.0315501988,-0.1503407657,0.228567481,0.1309167743,0.0849868432,-0.2012487501,-0.0813322812,-0.3202268481,0.4161017537,0.2294744104,0.1594106555,0.0948702693,0.2396360189,-0.1131370962,0.2765878737,-0.083134152,0.4926637709,0.1150816604,-0.0733320937,-0.2357814759,-0.1626849025,0.4998654723,0.2208077461,-0.1301795542,0.2457042634,-0.1459175646,-0.1141961291,-0.4050400257,-0.110821642,0.1224457398,0.0407714695,-0.069275327,-0.3382659554,0.3429746628,0.0266004372,-0.241807133,-0.1507674903,0.027509274,-0.3063729405,0.4966258109,0.0495938696,0.7502151728,0.0802693218,0.2257988155,0.4995680451,0.1529822201,0.4479938447,-0.0196744017,0.1059955508,-0.2297638208,-0.0201859456,-0.0801836699,-0.077208057,-0.032640554,0.2269545346,-0.2821175456,0.3359482884,-0.0033662017,-0.3440558314,-0.0055838763,0.2839399576,0.2938865125,-0.1255471259,-0.2925033867,0.0554176979,-0.0217533112,0.0426061228,-0.123104617,-0.10986913,0.2133318335,0.0194102433,-0.1174660251,-0.1151945442,-0.3224740922,0.2670014799,0.1172659621,-0.16850546,0.4472988844,0.2850080729,0.1239574254,0.0077961907,-0.0711963177,0.0781344846,0.208024472,-0.0823081061,0.1668940932,-0.0428409614,0.1570523828,-0.0013513472,-0.0025432699,-0.0050385557,0.2122475505,0.1522619128,-0.4720571041,-0.0966365561,-0.0523436219,-0.1472113281,-0.2569829822,0.2628075778,-0.0819587111,0.0326912329,0.0766985565,0.0239622407,-0.0052396986,0.188545391,-0.1153898314,-0.0173722506,-0.025045434,0.1201421991,0.402279079,0.2349775583,0.3080202937,0.3002696931,-0.0681713969,-0.1803524643,0.1713416874,-0.1220774874,0.2909734845,0.1756087989,0.1464813352,-0.2375948876,-0.3044601083,0.1180512011,0.0134803792,0.0669712424,-0.0260012746,-0.183790952,-0.235414803,0.442573458,-0.1116050854,-0.0637499467,-0.1207009852,0.3430419862,-0.0111435689,0.2345221341,-0.3226072192,-0.0901742727,-0.0342494473,-0.0006589094,-0.0721487105,-0.3950374424,0.0485444404,0.0562660731,0.1170466766,-0.3386164904,-0.1039807945,-0.201116696,0.0163675342,0.0918511897,0.1450993568,-0.0100537175,0.0832845122,-0.0046016574,-0.1219006106,-0.1263104528,0.0334950201,-0.2133260667,0.2303094119,0.2241814733,-0.2433054,-0.3525441289,-0.144126907,0.0464698672,0.1212242767,-0.1529877335,0.1883778274,0.1076349542,-0.1870577782,-0.0602206066,0.0611152314,0.0929599255,-0.2350378931,0.0220601857,0.260961473,0.2135649174,-0.2092922926,0.1950672567,0.0953018442,0.5737602711,-0.3621697426,-0.1237952933,-0.1961314231,0.1801472902,-0.3975524902,0.1067782938,-0.0357708298,0.2692307234,0.1508831233,0.324377358,0.3548027277,0.1190224215,-0.2247096449,-0.0045980257,0.2842541635,0.0008053809,0.0646130368,0.0652469769,-0.0418842323,0.095235832,0.4795926511,0.0568917505,0.2737270594,0.5982534289,0.1881877035,0.133950904,0.5289391875,0.015870098,0.4654017687,0.0724635869,-0.0744092986,0.1654357016,0.1665534973,-0.1509913802,0.2589283288,-0.0840415359,-0.0470688865,-0.5038694739,0.061820291,0.1850280762,-0.1930899769,-0.3244959712,0.1354408562,0.0958599746,-0.0190268084,-0.1349738836,-0.2654228508,-0.3414900601,0.1372236162,-0.1111051664,-0.0186736472,-0.2364375442,0.3297955692,0.1804779917,0.2503113449,0.0762813389,0.3437651992,0.3048168123,-0.1078618616,-0.0460155755,0.4059788287,0.2508131862,0.1334399432,-0.7279350162,0.0768476799,-0.3143244088,-0.1537243575,-0.1712713093,0.0198038146,-0.1110552996,0.2067063153,0.4055154622,0.0355420113,0.0332212634,0.0554631315,0.4479646087,0.033194229,-0.2274219841,0.3062977493,-0.1743718386,0.0557491481,0.1576089412,0.1194696426,-0.4857297838,-0.2708645761,0.6146504879,-0.023026187,0.1486743987,-0.1452292353,0.0934865847,0.0028621738,0.6333819032,0.1780219525,0.0357498303,-0.130356133,0.0039926744,-0.5874857306,-0.2055065632,-0.0397071429,-0.0746472776,0.1376367062,0.286462456,-0.0014385763,0.2262895554,-0.4077381492,0.1404839456,-0.3960298002,-0.2935956419,-0.6176458001,0.0392358229,-0.2511718273,0.2211206406,-0.190226838,0.0159047227,0.2776587009,0.1377040297,-0.0955407321,-0.0318648852,-0.0550692528,-0.0857351497,0.1362803876,-0.0336258598,0.4873383641,-0.0316753387,0.0028377497,0.0123232054,-0.1042631939,0.0031745976,-0.1777554899,0.5362423658,0.106227763,0.1782298833,-0.1864857078,0.1219007969,0.14419204,0.118016161,0.1954147369,-0.264931649,-0.4106279314,0.3911326528,0.1150140911,0.4158944488,-0.0828976408,0.5141797066,-0.1452318281,0.1062401012,-0.1097330898,-0.369781822,0.4228875637,-0.1451584548,0.0434376523,0.1298264563,0.0635659918,0.1955458671,0.0920954198,-0.4175347686,-0.3561795354,0.3726508915,-0.4296780825,-0.3153370917,0.2559309602,0.025458632,-0.4772297442,-0.2210823298,-0.0440044068,-0.0213682223,-0.4149132371,0.0735889673,-0.0863084421]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/378","title":"[dataset] Structure of MLQA seems unecessary nested","comments":"You're right, I think we don't need to use the nested dictionary. \r\n","body":"The features of the MLQA dataset comprise several nested dictionaries with a single element inside (for `questions` and `ids`): https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/mlqa\/mlqa.py#L90-L97\r\n\r\nShould we keep this @mariamabarham @patrickvonplaten? Was this added for compatibility with tfds?\r\n\r\n```python\r\n features=nlp.Features(\r\n {\r\n \"context\": nlp.Value(\"string\"),\r\n \"questions\": nlp.features.Sequence({\"question\": nlp.Value(\"string\")}),\r\n \"answers\": nlp.features.Sequence(\r\n {\"text\": nlp.Value(\"string\"), \"answer_start\": nlp.Value(\"int32\"),}\r\n ),\r\n \"ids\": nlp.features.Sequence({\"idx\": nlp.Value(\"string\")})\r\n```","comment_length":12,"text":"[dataset] Structure of MLQA seems unecessary nested\nThe features of the MLQA dataset comprise several nested dictionaries with a single element inside (for `questions` and `ids`): https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/mlqa\/mlqa.py#L90-L97\r\n\r\nShould we keep this @mariamabarham @patrickvonplaten? Was this added for compatibility with tfds?\r\n\r\n```python\r\n features=nlp.Features(\r\n {\r\n \"context\": nlp.Value(\"string\"),\r\n \"questions\": nlp.features.Sequence({\"question\": nlp.Value(\"string\")}),\r\n \"answers\": nlp.features.Sequence(\r\n {\"text\": nlp.Value(\"string\"), \"answer_start\": nlp.Value(\"int32\"),}\r\n ),\r\n \"ids\": nlp.features.Sequence({\"idx\": nlp.Value(\"string\")})\r\n```\nYou're right, I think we don't need to use the nested dictionary. \r\n","embeddings":[0.1129020005,-0.1617100984,-0.167859748,0.2404197454,-0.0401410758,0.2366998941,0.1821682304,0.0967536718,-0.0977478474,-0.0532935299,-0.030526476,0.2341212183,0.0398347527,0.6980101466,0.006767984,-0.1038919315,0.2103238106,0.0516499542,0.0336686224,0.1154036522,-0.1862295419,0.235056892,-0.2071382999,0.0493236184,-0.1934910417,-0.1076018736,-0.3099075854,0.1635463983,0.1095410213,-0.2849237025,0.0694364682,0.2586242259,-0.1236065775,-0.098145403,-0.0001037803,-0.151943326,0.1003324091,-0.0448807552,-0.1809391528,0.0262464453,-0.0163421798,-0.2460024804,0.0328003764,-0.2757477164,-0.1210982725,-0.3237063289,0.0683294684,-0.1519547105,0.1994572878,-0.0963046178,0.2519131899,-0.141916886,-0.0095080705,0.3049568534,-0.1109694168,0.0482278802,-0.0766587034,0.0580967665,0.2490698397,-0.1324665248,-0.0093806675,0.4154791236,-0.0664092973,-0.1307559013,0.3685471117,0.1091893241,0.0984507874,-0.4826312661,-0.0523626916,0.3358978629,0.2993191183,-0.0833727196,-0.2085277289,-0.4850482047,-0.1252269298,-0.1550242752,0.072080493,0.0554729551,0.0644014999,0.1788849384,0.1113890931,-0.2303030789,-0.0355064049,0.1313561946,-0.1554445475,0.3227158487,0.0224589147,0.0464161038,-0.1952154934,-0.2284358889,0.15930067,-0.1304008961,-0.0228491873,-0.0518554337,-0.2370217592,-0.3065786362,0.2232555151,0.0464704894,0.0211894605,0.018860396,-0.0581494458,-0.1293378174,0.0380772762,-0.0764679536,0.4048275352,0.0635048226,0.0901597738,-0.4005055726,0.1018212214,-0.1415857822,0.0117721325,-0.0336772352,0.1742837429,0.2469700128,-0.4084916115,-0.0516822934,0.462410748,0.1416374445,-0.3127142489,-0.0585301965,-0.6937094927,-0.1335221529,-0.0069971271,0.12000563,0.0909528211,0.3462826908,-0.0967018381,0.0115604559,-0.08920867,-0.2102695405,-0.1442048103,0.0344182141,-0.0142138144,-0.04410927,-0.0187293533,0.1961064637,0.1412345618,-0.1010593399,0.0937044322,0.2730742693,0.1542865485,-0.0527665243,0.2073159814,0.1248637289,-0.0198622383,-0.0161044933,-0.0127200829,-0.10801889,-0.1402182877,0.151706785,-0.0628697798,-0.326800555,-0.0256044939,0.2760942578,-0.050262481,-0.159104988,0.2505046129,0.7449442148,0.3390759826,-0.0937882215,-0.0532538928,0.1369133294,-0.4243008196,-0.24531129,-0.1778023094,0.1627647877,-0.3163942993,0.0890415013,-0.3317714632,0.0544681363,-0.1758932918,0.3180467188,-0.0852564797,0.3821187913,-0.1409472525,0.3326865733,-0.0553517379,-0.3694323003,-0.330930531,0.3071120381,0.1079646796,0.1123393923,0.1568608433,0.2183415592,-0.0539327189,-0.273699075,0.209211275,0.3001350462,-0.1597666144,-0.0403801948,-0.0751994327,-0.1852402389,0.5864815712,-0.0483210012,-0.5093967319,-0.1421280205,-0.2229768932,0.0541032813,0.4089272022,0.0049241018,-0.1345012933,-0.1675753593,0.2658639848,-0.1055073738,0.0364505872,-0.1756792516,-0.6048766375,-0.0071260012,-0.2411969155,0.2515308857,-0.0051752632,-0.4437871575,0.1569332182,-0.0223366525,-0.2293015122,-0.3155376911,0.2783632278,-0.2526366413,-0.0722583979,-0.1017593592,-0.2528606653,-0.0233209152,0.0824717134,-0.1230715141,-0.3573718369,0.2604019642,-0.174679935,0.271520704,0.3366699219,0.7286416888,-0.0086491713,-0.0615337603,0.2035207599,0.0684749559,-0.1053514853,-0.1406834871,-0.0394276381,0.3033707738,0.2617571354,-0.1393074095,0.049588684,0.3156534135,0.122085467,-0.1809040308,-0.1447414905,0.3175444007,0.1459709704,0.3444746733,0.0309294984,0.1970330626,-0.0040721027,-0.0845440775,-0.094563894,-0.4414532483,-0.0073600798,-0.0208915714,-0.049007725,0.2241958827,-0.2816946507,0.9435124397,0.4639377594,-0.0763172209,-0.017328158,-0.0472510718,0.0843229517,-0.4183889925,0.2593080699,0.1079505458,0.3038903773,0.2880138457,-0.0286567416,0.1069050059,0.1907873452,-0.2352232337,0.3048237562,0.1228475571,-0.1637366414,0.0097383754,0.3019728363,0.1143318191,-0.2902640402,-0.1373989135,-0.0619137809,0.0390687548,-0.2153292745,0.3280590177,-0.2392202914,-0.8135586381,0.2610372007,-0.260456562,-0.2558636069,-0.4941615164,0.13213557,-0.1085252762,-0.1259079725,0.1809988767,0.5080777407,0.3379573226,0.031767875,0.2027642876,-0.0664124563,-0.1208943352,-0.1904046386,0.2245268524,0.1256752908,0.1062041521,0.3246563971,0.2873571813,-0.0815038234,-0.2235117555,-0.5870536566,0.1988425851,-0.4439298213,0.4579761624,0.2996492386,0.062648043,-0.153355822,-0.2206124365,-0.0137534533,0.4634609818,-0.2787651122,-0.1096927524,0.0243734773,-0.1566776484,-0.1109969988,-0.4471596181,-0.3833472729,-0.3712235987,0.3126412928,-0.1145146415,0.4288516045,0.2440855503,-0.3785023093,-0.1579925269,0.1056689024,-0.109749347,-0.3104499876,0.202114746,0.1548136175,-0.1600515991,-0.2850977182,-0.2337066233,-0.3168998957,0.3472671211,-0.4200587869,-0.3249661326,-0.224615261,-0.1171856076,0.2301759273,0.2324376851,-0.0139555372,0.2350477427,0.2340883315,-0.0722072124,0.1699731201,-0.1119702086,0.2346432954,0.2768822908,0.2484181076,0.0951461494,0.4282280803,0.1437743157,0.2843646109,0.4552217424,0.2544784546,0.1283341795,0.1038583145,0.1218162701,-0.2509588897,-0.22855331,0.3702272475,0.1049727574,-0.089247942,0.3201301992,-0.0490745232,-0.1487412453,-0.0147938579,0.0118303839,0.1878573298,-0.2189705521,0.0214740764,-0.4769682884,0.2643146813,0.0303412545,-0.0625369474,0.1666706055,-0.1673036516,-0.1088831499,0.0052486458,0.1892498583,0.0357557461,-0.6304648519,-0.541649282,-0.5336409211,0.0441330895,0.0814368501,0.086105451,-0.1003911868,0.0173898656,0.0507921502,-0.036495056,0.4747058153,-0.124531135,-0.2620954216,0.1803464293,-0.0207852833,-0.0867446512,0.0876494125,0.0357614234,0.4004414678,0.3183524609,0.0965726152,-0.4269354045,-0.2156915218,0.1586455107,0.0782047138,-0.4169171453,-0.2913177013,-0.1593929976,-0.2128798068,-0.189455539,-0.3084094822,0.1786957532,0.1116585284,0.1977036595,-0.1256081611,-0.0539850518,-0.3882591426,0.2501245141,0.2429426908,0.0652820021,0.0744458586,0.1773074269,-0.1590687782,0.3366489112,-0.3008479774,0.4584333897,0.1718747467,0.0478321463,-0.1033725664,-0.1364687979,0.5586607456,0.1710759103,-0.1290926635,0.2173723727,-0.1385501623,-0.1196343452,-0.339431107,-0.1072708145,0.0472745746,0.0083648516,-0.0299887937,-0.388215512,0.3647526205,-0.1503587961,-0.2111834288,-0.2090357393,0.1368315816,-0.237335369,0.3776637316,0.0793629289,0.7489511371,0.1905665398,0.3354167044,0.6157534122,0.1892852634,0.4925862253,0.0336025767,0.0885625109,-0.215381518,0.0762709826,-0.0777514651,0.023643231,-0.075484395,0.2479281425,-0.2922059596,0.2984279394,0.1687969714,-0.3733332753,-0.0002632,0.3776885867,0.2274922132,-0.1349539608,-0.3983166516,0.1271006316,-0.0046206922,-0.0796640813,-0.1238239855,-0.1910806,0.0903546587,0.0384335108,0.0309089385,-0.1697992235,-0.2229822129,0.2094692588,0.1412904561,-0.2046161443,0.2362853885,0.2417664975,0.0612378716,-0.0648436844,-0.1546104997,-0.0518071838,0.070988901,-0.0922299102,0.163755849,-0.0993134007,0.2988150716,-0.1372038275,-0.0039242236,-0.0079810601,0.1889076084,0.052395124,-0.4686602652,0.0045563225,-0.1273737997,-0.2211480588,-0.2748422325,0.2127213925,-0.1839918047,0.0095682237,0.1523882449,0.0705917776,-0.0126765687,0.0705136135,0.0202751942,0.1225966886,-0.0672075525,0.1221648529,0.4017608762,0.2644040287,0.2323828787,0.1801781058,-0.0769839063,-0.186989665,0.0950600654,-0.2155737579,0.2258879095,0.2643170059,0.2353719473,-0.2203288227,-0.2991108894,0.2437725663,-0.0519993603,0.0592384227,-0.0244756248,-0.1704368442,-0.3056199551,0.4568190873,-0.1557535231,-0.0059025888,-0.0306368805,0.2961098254,0.0129512763,0.2671975195,-0.3968257308,-0.2288678139,0.0107596545,0.008010312,0.0499890782,-0.4013254642,0.0643666014,0.1428083181,0.1734694839,-0.3227255344,-0.1179991513,-0.2722670734,-0.0083436547,0.0627276897,0.2726044953,-0.0023454083,0.0726562291,-0.020613322,-0.2437339574,-0.0593498833,0.0896929055,-0.1883234531,0.1175707057,0.3125926554,-0.2990819812,-0.2371905297,-0.0777700543,0.0921856463,0.1153039038,-0.1531674862,0.2535198629,-0.0332739986,-0.2003663331,-0.0664106905,-0.0297622699,0.1334127933,-0.2478591949,-0.0794343427,0.3052472174,0.1521270275,-0.2686014175,0.1618153155,0.0505605265,0.3549330831,-0.3027730286,-0.0370302498,-0.1076642126,0.2536513507,-0.3878917098,0.1311376691,-0.0856904387,0.1022232473,0.1631417274,0.3351960182,0.3117071688,-0.0328799635,-0.1872356236,0.0308749005,0.1563267708,0.0274452288,-0.1310404539,-0.0265433434,-0.098830089,0.1378926784,0.4203836918,0.0227577649,0.2420419604,0.5524483919,0.1693445146,0.1877411157,0.5425141454,0.0918226391,0.4639032781,0.2223411202,-0.0268049743,0.0000928676,0.1564619392,-0.2104990482,0.2695952952,-0.0767137259,-0.0229586028,-0.3566917777,0.1188201159,0.2178607732,-0.1931703091,-0.2449112982,0.1683302522,0.1598316878,-0.0526047833,-0.1920372695,-0.3005848229,-0.4236874878,0.074796319,-0.1451200396,-0.0013721206,-0.2763203084,0.3352847993,0.1713446528,0.3503618538,0.085881643,0.3254060447,0.1853034049,-0.0780063942,0.0269446187,0.3468596637,0.1773794293,0.2120513618,-0.5851086974,0.0667626858,-0.3293509185,-0.2294770032,-0.1097431704,0.0835072175,-0.1073881313,0.3252482712,0.3541865349,0.0751024485,-0.0608544014,0.0528980009,0.4397214949,-0.0558877625,-0.2631515563,0.2683419883,-0.1674149334,-0.0388687328,0.1450010836,0.1932974309,-0.4761151671,-0.2676160634,0.3897163868,-0.0286991,0.2202823013,-0.2255214155,0.129200533,-0.0778609365,0.6228441596,0.1661511511,0.1411572546,-0.0791186765,0.0460130461,-0.5353443623,-0.1607857794,0.1029037386,-0.032448858,0.07490246,0.3083255887,-0.0055165729,0.3119633496,-0.4592508376,0.1442514062,-0.2767985165,-0.2115733474,-0.5024560094,-0.066538021,-0.1726712734,0.2299446464,-0.1261191219,0.1045277789,0.1657797396,0.0975304544,-0.009228494,-0.093525067,-0.1333955973,-0.0879659057,0.254324764,-0.0912788957,0.4587984085,-0.0462844446,0.0085446797,-0.0540599972,-0.0598353818,0.0519994423,-0.2089088559,0.3134673834,0.1234054789,0.23612234,-0.1274418235,0.2072522044,0.1282597482,0.1050034314,0.1671151221,-0.1495591998,-0.3601041734,0.470013082,0.0957548171,0.3577435911,-0.1892931312,0.4751830399,-0.1370369494,0.0027937395,-0.1187129244,-0.3126731515,0.44253093,-0.1485387534,0.0309076104,0.2669891715,0.1244590506,0.1636341363,-0.0541520454,-0.3827115893,-0.3505646884,0.3655843139,-0.3482284248,-0.2495884746,0.3308201432,0.0486287251,-0.4145289958,-0.2007243037,0.09856686,0.0044851117,-0.2582331896,0.1204661876,-0.1327682287]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/376","title":"to_pandas conversion doesn't always work","comments":"**Edit**: other topic previously in this message moved to a new issue: https:\/\/github.com\/huggingface\/nlp\/issues\/387","body":"For some complex nested types, the conversion from Arrow to python dict through pandas doesn't seem to be possible.\r\n\r\nHere is an example using the official SQUAD v2 JSON file.\r\n\r\nThis example was found while investigating #373.\r\n\r\n```python\r\n>>> squad = load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/train-v2.0.json\"]}, download_mode=nlp.GenerateMode.FORCE_REDOWNLOAD, version=\"1.0.0\", field='data')\r\n>>> squad['train']\r\nDataset(schema: {'title': 'string', 'paragraphs': 'list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>'}, num_rows: 442)\r\n>>> squad['train'][0]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/nlp\/arrow_dataset.py\", line 589, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/nlp\/arrow_dataset.py\", line 529, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pandas().to_dict(\"list\"))\r\n File \"pyarrow\/array.pxi\", line 559, in pyarrow.lib._PandasConvertible.to_pandas\r\n File \"pyarrow\/table.pxi\", line 1367, in pyarrow.lib.Table._to_pandas\r\n File \"\/Users\/thomwolf\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 766, in table_to_blockmanager\r\n blocks = _table_to_blocks(options, table, categories, ext_columns_dtypes)\r\n File \"\/Users\/thomwolf\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 1101, in _table_to_blocks\r\n list(extension_columns.keys()))\r\n File \"pyarrow\/table.pxi\", line 881, in pyarrow.lib.table_to_blocks\r\n File \"pyarrow\/error.pxi\", line 105, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowNotImplementedError: Not implemented type for Arrow list to pandas: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n```\r\n\r\ncc @lhoestq would we have a way to detect this from the schema maybe?\r\n\r\nHere is the schema for this pretty complex JSON:\r\n```python\r\n>>> squad['train'].schema\r\ntitle: string\r\nparagraphs: list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>\r\n child 0, item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n child 0, qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>\r\n child 0, item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>\r\n child 0, question: string\r\n child 1, id: string\r\n child 2, answers: list<item: struct<text: string, answer_start: int64>>\r\n child 0, item: struct<text: string, answer_start: int64>\r\n child 0, text: string\r\n child 1, answer_start: int64\r\n child 3, is_impossible: bool\r\n child 4, plausible_answers: list<item: struct<text: string, answer_start: int64>>\r\n child 0, item: struct<text: string, answer_start: int64>\r\n child 0, text: string\r\n child 1, answer_start: int64\r\n child 1, context: string\r\n```","comment_length":13,"text":"to_pandas conversion doesn't always work\nFor some complex nested types, the conversion from Arrow to python dict through pandas doesn't seem to be possible.\r\n\r\nHere is an example using the official SQUAD v2 JSON file.\r\n\r\nThis example was found while investigating #373.\r\n\r\n```python\r\n>>> squad = load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/train-v2.0.json\"]}, download_mode=nlp.GenerateMode.FORCE_REDOWNLOAD, version=\"1.0.0\", field='data')\r\n>>> squad['train']\r\nDataset(schema: {'title': 'string', 'paragraphs': 'list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>'}, num_rows: 442)\r\n>>> squad['train'][0]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/nlp\/arrow_dataset.py\", line 589, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/nlp\/arrow_dataset.py\", line 529, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pandas().to_dict(\"list\"))\r\n File \"pyarrow\/array.pxi\", line 559, in pyarrow.lib._PandasConvertible.to_pandas\r\n File \"pyarrow\/table.pxi\", line 1367, in pyarrow.lib.Table._to_pandas\r\n File \"\/Users\/thomwolf\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 766, in table_to_blockmanager\r\n blocks = _table_to_blocks(options, table, categories, ext_columns_dtypes)\r\n File \"\/Users\/thomwolf\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 1101, in _table_to_blocks\r\n list(extension_columns.keys()))\r\n File \"pyarrow\/table.pxi\", line 881, in pyarrow.lib.table_to_blocks\r\n File \"pyarrow\/error.pxi\", line 105, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowNotImplementedError: Not implemented type for Arrow list to pandas: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n```\r\n\r\ncc @lhoestq would we have a way to detect this from the schema maybe?\r\n\r\nHere is the schema for this pretty complex JSON:\r\n```python\r\n>>> squad['train'].schema\r\ntitle: string\r\nparagraphs: list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>\r\n child 0, item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n child 0, qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>\r\n child 0, item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>\r\n child 0, question: string\r\n child 1, id: string\r\n child 2, answers: list<item: struct<text: string, answer_start: int64>>\r\n child 0, item: struct<text: string, answer_start: int64>\r\n child 0, text: string\r\n child 1, answer_start: int64\r\n child 3, is_impossible: bool\r\n child 4, plausible_answers: list<item: struct<text: string, answer_start: int64>>\r\n child 0, item: struct<text: string, answer_start: int64>\r\n child 0, text: string\r\n child 1, answer_start: int64\r\n child 1, context: string\r\n```\n**Edit**: other topic previously in this message moved to a new issue: https:\/\/github.com\/huggingface\/nlp\/issues\/387","embeddings":[0.0346822031,0.1334801316,-0.1031353548,0.0542142317,0.2966339886,-0.0923537612,0.3753246367,0.4698339701,-0.1327777356,-0.2164529711,-0.2198514938,0.8954614401,0.1729376763,0.0701537579,-0.0751237944,-0.0060988646,0.1450803429,0.1565450579,0.0088705253,-0.0708985552,-0.0599108003,0.2531312108,-0.3004131019,0.1826229095,-0.0905479863,-0.2454887033,-0.0693996027,0.2031732351,-0.0372755565,-0.4164663553,0.3161965609,-0.2416543663,-0.0060523916,0.3274362385,-0.0001048727,0.1640702635,0.2705137134,-0.0000030442,-0.3173806369,0.0224589016,0.211726591,-0.2963275909,0.3813583255,-0.3474819362,0.086946182,-0.3566198349,0.01714704,-0.0835760757,0.7314293385,0.4084213972,0.2785182595,0.2500542998,0.2212230712,0.1054728404,0.0283313319,0.0175163429,-0.1275167167,0.090666689,0.0750308335,-0.110254325,-0.0946216285,-0.0678505898,0.0189643037,-0.1256456077,-0.0272465795,0.0404620282,-0.0931733847,-0.2096050382,-0.0769636706,0.272613436,0.0939596146,-0.3873830736,-0.3478513062,-0.3279161751,0.0629389137,-0.0009637792,-0.0402867384,0.0938734487,-0.1286568493,0.115122132,0.3740177155,0.212165907,-0.1918032914,0.0442614295,-0.1055234522,0.2287072986,-0.1356019676,0.0541744567,-0.0578796342,-0.3256850243,0.0986953899,-0.0277544223,0.1278437823,0.1253151596,-0.3464330137,-0.2328877002,-0.0223375186,-0.5411040187,0.0283725094,-0.0601906963,0.4258277118,0.0907199085,0.1717899889,0.2311955243,0.366830945,-0.0149523327,0.0346243419,0.1606027633,0.112255618,0.0811905414,-0.1864200234,0.1348263174,-0.1662887484,-0.2421691865,-0.0859451592,-0.1153013557,0.3311738074,0.0254375171,-0.4399408996,-0.0537530594,-0.5590521097,0.2023044229,0.0710934773,0.1338669956,-0.0746624321,0.3300741613,0.0325634964,0.329606086,-0.1141772419,0.2582934797,-0.1887711883,0.1703869402,-0.0874276236,-0.2365327626,0.098211132,0.3039860427,0.0466019623,0.0504975505,0.0918959305,-0.0839026645,-0.1096951291,-0.0015048975,0.1392647624,0.068577148,0.0357962586,-0.0865211561,0.1393049657,-0.5467719436,-0.1163137704,0.0084967315,-0.1911562383,0.0046371268,-0.1746991277,0.2353428602,-0.2089509219,-0.2094300687,-0.2487303913,0.3075787723,0.2559840083,-0.1344774067,0.0554898828,-0.213483125,0.1080434397,-0.3103169799,0.0428105704,0.0812463909,-0.5389331579,-0.0203246493,0.0307998136,0.0184504427,0.085801281,0.3102589548,-0.1158483103,0.1331138611,-0.0083087208,0.3600106537,0.4705029726,-0.5010068417,-0.2272088826,0.1233954057,-0.1028219387,0.0469573326,-0.0094133411,-0.0619990788,0.1139433756,-0.0847894177,0.0009975655,0.4146360457,0.0844547227,-0.0295385756,-0.3178605139,-0.0425884239,0.1657475382,-0.0479368903,-0.20704934,-0.0420340598,0.1097064242,0.1532749087,0.2502881289,-0.0173268691,0.1595972627,0.053288836,0.1093756333,-0.1778445691,0.1243778244,-0.1607412398,-0.4700013101,0.0486028008,-0.4537335336,0.0608935803,-0.4144336879,-0.22325477,-0.2128350884,0.1473476887,-0.1878427714,0.0255582426,0.3575513363,-0.1679501832,-0.0512810908,-0.096446313,-0.1266192496,-0.1882031858,-0.2063402385,0.1719735116,-0.1708500385,0.2271744609,-0.1129681841,-0.1993496716,0.2624519467,0.4142284393,0.108648628,0.0614258423,0.0796472654,0.2213223726,-0.0544628128,-0.0317883864,0.3201678097,-0.290997088,-0.0313173868,-0.2659501731,0.014143005,0.3197631836,0.1506410837,-0.1804138571,-0.3506352603,0.511505425,-0.0059452746,0.3427329957,-0.0489868112,0.2725570798,0.209335655,0.0373880304,-0.2711518705,-0.0717702508,-0.4835414886,0.3345214725,-0.0204405095,-0.0447024144,-0.2443451136,0.3678493798,0.5505483747,-0.0689330548,0.1616614759,0.0775086731,0.0161989927,-0.2619673908,0.0648875162,0.0690608919,0.2639737427,0.2140438706,0.2403579056,-0.1615493745,-0.2850795984,-0.0713730976,0.2082889676,0.097463429,0.5278190374,0.0792274028,0.5001674294,-0.0341707654,-0.3575006723,-0.199280113,-0.0516332835,0.0686855614,-0.2854181528,0.2564881742,-0.3057742119,0.047743652,-0.0250567626,-0.1595958322,-0.018488232,-0.3592126667,-0.0316123068,-0.1378120631,-0.1378915906,0.2406532764,0.0266213566,0.0844946653,0.2327074111,-0.2775380611,-0.1527980268,-0.2352082133,-0.2616933882,0.2177621275,0.3379372656,0.041135069,-0.0485288985,0.2794268429,-0.0604334585,-0.0615333468,-0.3725876808,-0.0082610091,-0.335912466,0.0715323091,0.2027655989,0.1623568982,-0.2620196342,-0.165569827,0.2315650582,-0.0164052621,-0.2091492265,0.0947052687,-0.1275328845,-0.0346242264,-0.1428414434,-0.2469440997,-0.226098612,-0.3014287055,0.5213175416,0.1887671053,0.221952498,0.1249702498,0.3326109946,0.3914610147,0.1052844748,0.0423734561,-0.0694774911,0.0976865962,0.2977075577,-0.1696940958,-0.3797700405,-0.0698821843,-0.1876223981,0.2407642007,-0.331091404,-0.1938592494,-0.1503854394,-0.0856653973,0.2066676021,-0.0267159585,-0.0419333242,0.2467585206,0.2092577815,-0.2028105557,-0.2000724673,-0.0342848822,0.1739806235,0.1710574031,0.037709944,-0.1712387204,0.3040522933,0.2585916817,0.438880384,0.2099828124,-0.1460916698,0.0409785584,-0.0747577697,0.1780075431,-0.0446518026,-0.1240197867,0.0583537109,-0.2827706337,0.2170437872,0.0208704174,-0.04804511,-0.4201204777,0.0662787408,0.2097399384,-0.2522659302,-0.2328830659,0.1130290851,-0.2635749876,-0.2586224973,0.0600233674,0.0072320648,-0.20386675,0.0052686059,-0.1134463027,-0.1282087862,-0.0791153312,0.0545390956,-0.427932173,-0.2682369351,-0.2955734134,0.1940231025,-0.027475344,0.3233992457,0.238930136,-0.322911799,0.1085109115,-0.2627046406,0.3716490567,0.0132563552,-0.0600725934,0.362927407,0.3301502168,-0.7625823617,-0.0459959805,-0.3256264329,0.4142325819,0.2585711181,0.1136583909,-0.373893708,-0.2970243096,0.1337301433,-0.0611591451,0.0308413804,0.1280531287,-0.2378212065,-0.3379614949,-0.2001910806,0.2259794474,0.3510623872,0.2922064066,-0.2115909904,-0.4218184948,-0.1945056021,-0.193995744,0.1717282683,0.218531549,0.1352571398,0.0284803491,-0.0527354442,-0.1477459818,0.2658387423,0.0832413137,0.4682664275,-0.0706397519,-0.1046361029,0.1021997929,0.0393761508,0.3186959624,0.2960724235,-0.18906793,0.0780218169,-0.148100242,0.011173225,0.2016936243,-0.2384642959,0.3859431744,-0.0168808661,0.0025899843,-0.151629582,0.4655601084,-0.020357227,-0.0327580348,0.0383430868,0.1135957241,-0.1131090596,0.4639159739,0.2065855414,0.8790787458,0.0262420699,0.3171678185,0.5428175926,-0.1710009724,0.5304216146,-0.0539954454,0.1478681862,-0.2938592434,0.5065795183,0.0645175874,-0.1425771415,-0.0863090008,0.2681821287,-0.2327599376,0.025835868,-0.1960622072,0.2170095593,0.2107066065,0.2751063704,-0.0305988174,-0.2755971253,-0.6548104286,0.2362291515,0.1165919676,-0.2514944077,0.1502362937,-0.4565244615,-0.4458034337,-0.1074710786,-0.0148462281,0.2622072399,-0.1795082837,0.3122819662,0.3167698383,-0.3436285853,0.197101891,0.5936039686,-0.1254546791,-0.2281091809,-0.0550539345,0.0861301646,0.0346791111,-0.0109290527,0.1019862443,-0.085417971,0.3436592221,-0.061799515,-0.1343171746,0.0810630471,-0.0348944142,-0.1304962933,0.0797432214,0.4892712533,-0.1618458778,-0.3484072089,-0.0983697921,-0.0384707935,-0.3954730332,-0.2417124212,0.2116261721,-0.0092253108,-0.1125635207,0.4255144,0.2137937844,-0.083978571,0.1484847665,0.2626099885,0.0930737779,0.3781287968,0.5176310539,0.0884295255,-0.1763887107,-0.2753964961,0.1769768298,0.1379491389,-0.1889899224,0.2807774246,-0.1882597059,-0.3576730788,0.0459904037,0.0207597371,-0.0589365773,0.1955636144,-0.259755373,-0.4056970775,-0.2798852921,0.2288470268,-0.0408817828,-0.0069536217,-0.04864024,0.3339018524,-0.0145018864,-0.0963488221,-0.3843574524,0.2725990415,-0.3716552556,0.0740748867,-0.3814620078,-0.3459908068,0.2668173015,0.0308048669,0.2160767615,0.1098613292,0.0036861273,-0.3170911372,-0.0461049266,0.0781307891,0.1460115761,0.1028930545,-0.015235885,0.0452596396,-0.3004338741,-0.1509357691,-0.1212889999,-0.055539526,-0.0096376119,0.1606694907,0.0099008707,0.0740238652,0.20944947,-0.1554741412,-0.0692407489,0.0523920283,-0.073882468,0.2532299161,-0.0490227118,0.0039508045,0.1122217774,0.2262326032,-0.0058491929,0.1865203679,0.2958050668,-0.162464872,-0.4413197041,-0.2185005248,0.3224136531,0.5740982294,-0.1763646603,0.05707068,-0.2007500678,0.2156663388,-0.4009667635,0.0981801525,0.2420495152,-0.1563884616,0.1629412621,0.1809740514,0.0274456404,-0.1156514958,-0.0125562027,0.0659356713,0.2097218782,-0.0765962973,-0.0596351214,0.2138530463,-0.016143905,0.0193220153,0.407071352,-0.0857663229,0.2633347809,0.3329678476,0.0312269777,0.0426279791,0.2017087191,0.0390785821,0.3710512221,0.1623593867,-0.0544953756,0.173710078,0.2876287699,-0.0220837481,0.1096094251,0.74064821,-0.0993698314,-0.1496194601,-0.0029150986,0.0964953899,-0.1920408905,-0.2609267235,-0.3624968529,-0.1506780684,-0.1066598147,-0.3490464091,-0.0708513036,-0.2164484411,-0.0891565904,-0.0799689069,0.1176134571,-0.1278524995,0.1915337145,0.0721953586,0.25208956,-0.0163424741,0.1791752428,0.191896528,-0.0074003758,-0.1781455725,0.1820492744,0.5045114756,0.1560384631,-0.4481222332,-0.0124125304,0.0752601326,-0.3575073183,0.0184038635,0.106576845,0.1137405708,-0.2378424704,0.3356722295,0.2290237993,-0.1850778013,-0.2165261954,0.222421512,-0.0761555955,0.1127024069,0.2996042371,-0.3052110076,-0.3296503425,-0.094468601,-0.0046699205,-0.2677340209,-0.2765890062,0.1584847271,0.0071122129,0.2048698217,-0.2126351893,0.1546743214,0.0693259239,0.5761985183,-0.0310886372,0.0948051661,0.0794112533,0.2102885246,-0.4790940881,0.1391242743,-0.1064219251,-0.1181201637,0.0687155053,0.4990251958,0.0630581155,0.1504907161,0.1508840472,0.0104325032,-0.1435123533,0.0968663096,-0.3981228769,0.0104692169,0.1829186082,0.0597024709,0.162882179,-0.2981536686,0.0647494644,-0.1292551905,0.1347406656,-0.0217997041,-0.2169077098,-0.027020989,-0.1069159508,0.2247478515,0.1091793403,0.1795613319,-0.0586220548,-0.2362982482,-0.1898047179,-0.1666766107,-0.2485951334,0.4792469144,-0.1454416364,0.2362560928,-0.0857738853,0.0941277295,0.1562325209,-0.0268115476,0.114797838,-0.134017095,-0.326020658,0.335285455,-0.0969309732,0.2485348582,-0.2594831586,0.3689167798,-0.1254217476,0.2203360498,-0.3546590805,-0.5474947691,0.5134512186,-0.3442061543,-0.3539212644,0.258849889,0.1708127558,0.2654415369,0.1027440578,-0.1676395684,-0.1582760662,0.2111672759,-0.1707343012,-0.3202186525,-0.0052573131,-0.1648396105,0.1009314954,-0.2636694014,-0.0430937521,0.005074298,0.1069141254,-0.2034539729,-0.3928065896]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/376","title":"to_pandas conversion doesn't always work","comments":"Could you try to update pyarrow to >=0.17.0 ? It should fix the `to_pandas` bug\r\n\r\nAlso I'm not sure that structures like list<struct> are fully supported in the lib (none of the datasets use that).\r\nIt can cause issues when using dataset transforms like `filter` for example","body":"For some complex nested types, the conversion from Arrow to python dict through pandas doesn't seem to be possible.\r\n\r\nHere is an example using the official SQUAD v2 JSON file.\r\n\r\nThis example was found while investigating #373.\r\n\r\n```python\r\n>>> squad = load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/train-v2.0.json\"]}, download_mode=nlp.GenerateMode.FORCE_REDOWNLOAD, version=\"1.0.0\", field='data')\r\n>>> squad['train']\r\nDataset(schema: {'title': 'string', 'paragraphs': 'list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>'}, num_rows: 442)\r\n>>> squad['train'][0]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/nlp\/arrow_dataset.py\", line 589, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/nlp\/arrow_dataset.py\", line 529, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pandas().to_dict(\"list\"))\r\n File \"pyarrow\/array.pxi\", line 559, in pyarrow.lib._PandasConvertible.to_pandas\r\n File \"pyarrow\/table.pxi\", line 1367, in pyarrow.lib.Table._to_pandas\r\n File \"\/Users\/thomwolf\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 766, in table_to_blockmanager\r\n blocks = _table_to_blocks(options, table, categories, ext_columns_dtypes)\r\n File \"\/Users\/thomwolf\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 1101, in _table_to_blocks\r\n list(extension_columns.keys()))\r\n File \"pyarrow\/table.pxi\", line 881, in pyarrow.lib.table_to_blocks\r\n File \"pyarrow\/error.pxi\", line 105, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowNotImplementedError: Not implemented type for Arrow list to pandas: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n```\r\n\r\ncc @lhoestq would we have a way to detect this from the schema maybe?\r\n\r\nHere is the schema for this pretty complex JSON:\r\n```python\r\n>>> squad['train'].schema\r\ntitle: string\r\nparagraphs: list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>\r\n child 0, item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n child 0, qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>\r\n child 0, item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>\r\n child 0, question: string\r\n child 1, id: string\r\n child 2, answers: list<item: struct<text: string, answer_start: int64>>\r\n child 0, item: struct<text: string, answer_start: int64>\r\n child 0, text: string\r\n child 1, answer_start: int64\r\n child 3, is_impossible: bool\r\n child 4, plausible_answers: list<item: struct<text: string, answer_start: int64>>\r\n child 0, item: struct<text: string, answer_start: int64>\r\n child 0, text: string\r\n child 1, answer_start: int64\r\n child 1, context: string\r\n```","comment_length":47,"text":"to_pandas conversion doesn't always work\nFor some complex nested types, the conversion from Arrow to python dict through pandas doesn't seem to be possible.\r\n\r\nHere is an example using the official SQUAD v2 JSON file.\r\n\r\nThis example was found while investigating #373.\r\n\r\n```python\r\n>>> squad = load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/train-v2.0.json\"]}, download_mode=nlp.GenerateMode.FORCE_REDOWNLOAD, version=\"1.0.0\", field='data')\r\n>>> squad['train']\r\nDataset(schema: {'title': 'string', 'paragraphs': 'list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>'}, num_rows: 442)\r\n>>> squad['train'][0]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/nlp\/arrow_dataset.py\", line 589, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/Users\/thomwolf\/Documents\/GitHub\/datasets\/src\/nlp\/arrow_dataset.py\", line 529, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pandas().to_dict(\"list\"))\r\n File \"pyarrow\/array.pxi\", line 559, in pyarrow.lib._PandasConvertible.to_pandas\r\n File \"pyarrow\/table.pxi\", line 1367, in pyarrow.lib.Table._to_pandas\r\n File \"\/Users\/thomwolf\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 766, in table_to_blockmanager\r\n blocks = _table_to_blocks(options, table, categories, ext_columns_dtypes)\r\n File \"\/Users\/thomwolf\/miniconda2\/envs\/datasets\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 1101, in _table_to_blocks\r\n list(extension_columns.keys()))\r\n File \"pyarrow\/table.pxi\", line 881, in pyarrow.lib.table_to_blocks\r\n File \"pyarrow\/error.pxi\", line 105, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowNotImplementedError: Not implemented type for Arrow list to pandas: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n```\r\n\r\ncc @lhoestq would we have a way to detect this from the schema maybe?\r\n\r\nHere is the schema for this pretty complex JSON:\r\n```python\r\n>>> squad['train'].schema\r\ntitle: string\r\nparagraphs: list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>\r\n child 0, item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n child 0, qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>\r\n child 0, item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>\r\n child 0, question: string\r\n child 1, id: string\r\n child 2, answers: list<item: struct<text: string, answer_start: int64>>\r\n child 0, item: struct<text: string, answer_start: int64>\r\n child 0, text: string\r\n child 1, answer_start: int64\r\n child 3, is_impossible: bool\r\n child 4, plausible_answers: list<item: struct<text: string, answer_start: int64>>\r\n child 0, item: struct<text: string, answer_start: int64>\r\n child 0, text: string\r\n child 1, answer_start: int64\r\n child 1, context: string\r\n```\nCould you try to update pyarrow to >=0.17.0 ? It should fix the `to_pandas` bug\r\n\r\nAlso I'm not sure that structures like list<struct> are fully supported in the lib (none of the datasets use that).\r\nIt can cause issues when using dataset transforms like `filter` for example","embeddings":[0.0346822031,0.1334801316,-0.1031353548,0.0542142317,0.2966339886,-0.0923537612,0.3753246367,0.4698339701,-0.1327777356,-0.2164529711,-0.2198514938,0.8954614401,0.1729376763,0.0701537579,-0.0751237944,-0.0060988646,0.1450803429,0.1565450579,0.0088705253,-0.0708985552,-0.0599108003,0.2531312108,-0.3004131019,0.1826229095,-0.0905479863,-0.2454887033,-0.0693996027,0.2031732351,-0.0372755565,-0.4164663553,0.3161965609,-0.2416543663,-0.0060523916,0.3274362385,-0.0001048727,0.1640702635,0.2705137134,-0.0000030442,-0.3173806369,0.0224589016,0.211726591,-0.2963275909,0.3813583255,-0.3474819362,0.086946182,-0.3566198349,0.01714704,-0.0835760757,0.7314293385,0.4084213972,0.2785182595,0.2500542998,0.2212230712,0.1054728404,0.0283313319,0.0175163429,-0.1275167167,0.090666689,0.0750308335,-0.110254325,-0.0946216285,-0.0678505898,0.0189643037,-0.1256456077,-0.0272465795,0.0404620282,-0.0931733847,-0.2096050382,-0.0769636706,0.272613436,0.0939596146,-0.3873830736,-0.3478513062,-0.3279161751,0.0629389137,-0.0009637792,-0.0402867384,0.0938734487,-0.1286568493,0.115122132,0.3740177155,0.212165907,-0.1918032914,0.0442614295,-0.1055234522,0.2287072986,-0.1356019676,0.0541744567,-0.0578796342,-0.3256850243,0.0986953899,-0.0277544223,0.1278437823,0.1253151596,-0.3464330137,-0.2328877002,-0.0223375186,-0.5411040187,0.0283725094,-0.0601906963,0.4258277118,0.0907199085,0.1717899889,0.2311955243,0.366830945,-0.0149523327,0.0346243419,0.1606027633,0.112255618,0.0811905414,-0.1864200234,0.1348263174,-0.1662887484,-0.2421691865,-0.0859451592,-0.1153013557,0.3311738074,0.0254375171,-0.4399408996,-0.0537530594,-0.5590521097,0.2023044229,0.0710934773,0.1338669956,-0.0746624321,0.3300741613,0.0325634964,0.329606086,-0.1141772419,0.2582934797,-0.1887711883,0.1703869402,-0.0874276236,-0.2365327626,0.098211132,0.3039860427,0.0466019623,0.0504975505,0.0918959305,-0.0839026645,-0.1096951291,-0.0015048975,0.1392647624,0.068577148,0.0357962586,-0.0865211561,0.1393049657,-0.5467719436,-0.1163137704,0.0084967315,-0.1911562383,0.0046371268,-0.1746991277,0.2353428602,-0.2089509219,-0.2094300687,-0.2487303913,0.3075787723,0.2559840083,-0.1344774067,0.0554898828,-0.213483125,0.1080434397,-0.3103169799,0.0428105704,0.0812463909,-0.5389331579,-0.0203246493,0.0307998136,0.0184504427,0.085801281,0.3102589548,-0.1158483103,0.1331138611,-0.0083087208,0.3600106537,0.4705029726,-0.5010068417,-0.2272088826,0.1233954057,-0.1028219387,0.0469573326,-0.0094133411,-0.0619990788,0.1139433756,-0.0847894177,0.0009975655,0.4146360457,0.0844547227,-0.0295385756,-0.3178605139,-0.0425884239,0.1657475382,-0.0479368903,-0.20704934,-0.0420340598,0.1097064242,0.1532749087,0.2502881289,-0.0173268691,0.1595972627,0.053288836,0.1093756333,-0.1778445691,0.1243778244,-0.1607412398,-0.4700013101,0.0486028008,-0.4537335336,0.0608935803,-0.4144336879,-0.22325477,-0.2128350884,0.1473476887,-0.1878427714,0.0255582426,0.3575513363,-0.1679501832,-0.0512810908,-0.096446313,-0.1266192496,-0.1882031858,-0.2063402385,0.1719735116,-0.1708500385,0.2271744609,-0.1129681841,-0.1993496716,0.2624519467,0.4142284393,0.108648628,0.0614258423,0.0796472654,0.2213223726,-0.0544628128,-0.0317883864,0.3201678097,-0.290997088,-0.0313173868,-0.2659501731,0.014143005,0.3197631836,0.1506410837,-0.1804138571,-0.3506352603,0.511505425,-0.0059452746,0.3427329957,-0.0489868112,0.2725570798,0.209335655,0.0373880304,-0.2711518705,-0.0717702508,-0.4835414886,0.3345214725,-0.0204405095,-0.0447024144,-0.2443451136,0.3678493798,0.5505483747,-0.0689330548,0.1616614759,0.0775086731,0.0161989927,-0.2619673908,0.0648875162,0.0690608919,0.2639737427,0.2140438706,0.2403579056,-0.1615493745,-0.2850795984,-0.0713730976,0.2082889676,0.097463429,0.5278190374,0.0792274028,0.5001674294,-0.0341707654,-0.3575006723,-0.199280113,-0.0516332835,0.0686855614,-0.2854181528,0.2564881742,-0.3057742119,0.047743652,-0.0250567626,-0.1595958322,-0.018488232,-0.3592126667,-0.0316123068,-0.1378120631,-0.1378915906,0.2406532764,0.0266213566,0.0844946653,0.2327074111,-0.2775380611,-0.1527980268,-0.2352082133,-0.2616933882,0.2177621275,0.3379372656,0.041135069,-0.0485288985,0.2794268429,-0.0604334585,-0.0615333468,-0.3725876808,-0.0082610091,-0.335912466,0.0715323091,0.2027655989,0.1623568982,-0.2620196342,-0.165569827,0.2315650582,-0.0164052621,-0.2091492265,0.0947052687,-0.1275328845,-0.0346242264,-0.1428414434,-0.2469440997,-0.226098612,-0.3014287055,0.5213175416,0.1887671053,0.221952498,0.1249702498,0.3326109946,0.3914610147,0.1052844748,0.0423734561,-0.0694774911,0.0976865962,0.2977075577,-0.1696940958,-0.3797700405,-0.0698821843,-0.1876223981,0.2407642007,-0.331091404,-0.1938592494,-0.1503854394,-0.0856653973,0.2066676021,-0.0267159585,-0.0419333242,0.2467585206,0.2092577815,-0.2028105557,-0.2000724673,-0.0342848822,0.1739806235,0.1710574031,0.037709944,-0.1712387204,0.3040522933,0.2585916817,0.438880384,0.2099828124,-0.1460916698,0.0409785584,-0.0747577697,0.1780075431,-0.0446518026,-0.1240197867,0.0583537109,-0.2827706337,0.2170437872,0.0208704174,-0.04804511,-0.4201204777,0.0662787408,0.2097399384,-0.2522659302,-0.2328830659,0.1130290851,-0.2635749876,-0.2586224973,0.0600233674,0.0072320648,-0.20386675,0.0052686059,-0.1134463027,-0.1282087862,-0.0791153312,0.0545390956,-0.427932173,-0.2682369351,-0.2955734134,0.1940231025,-0.027475344,0.3233992457,0.238930136,-0.322911799,0.1085109115,-0.2627046406,0.3716490567,0.0132563552,-0.0600725934,0.362927407,0.3301502168,-0.7625823617,-0.0459959805,-0.3256264329,0.4142325819,0.2585711181,0.1136583909,-0.373893708,-0.2970243096,0.1337301433,-0.0611591451,0.0308413804,0.1280531287,-0.2378212065,-0.3379614949,-0.2001910806,0.2259794474,0.3510623872,0.2922064066,-0.2115909904,-0.4218184948,-0.1945056021,-0.193995744,0.1717282683,0.218531549,0.1352571398,0.0284803491,-0.0527354442,-0.1477459818,0.2658387423,0.0832413137,0.4682664275,-0.0706397519,-0.1046361029,0.1021997929,0.0393761508,0.3186959624,0.2960724235,-0.18906793,0.0780218169,-0.148100242,0.011173225,0.2016936243,-0.2384642959,0.3859431744,-0.0168808661,0.0025899843,-0.151629582,0.4655601084,-0.020357227,-0.0327580348,0.0383430868,0.1135957241,-0.1131090596,0.4639159739,0.2065855414,0.8790787458,0.0262420699,0.3171678185,0.5428175926,-0.1710009724,0.5304216146,-0.0539954454,0.1478681862,-0.2938592434,0.5065795183,0.0645175874,-0.1425771415,-0.0863090008,0.2681821287,-0.2327599376,0.025835868,-0.1960622072,0.2170095593,0.2107066065,0.2751063704,-0.0305988174,-0.2755971253,-0.6548104286,0.2362291515,0.1165919676,-0.2514944077,0.1502362937,-0.4565244615,-0.4458034337,-0.1074710786,-0.0148462281,0.2622072399,-0.1795082837,0.3122819662,0.3167698383,-0.3436285853,0.197101891,0.5936039686,-0.1254546791,-0.2281091809,-0.0550539345,0.0861301646,0.0346791111,-0.0109290527,0.1019862443,-0.085417971,0.3436592221,-0.061799515,-0.1343171746,0.0810630471,-0.0348944142,-0.1304962933,0.0797432214,0.4892712533,-0.1618458778,-0.3484072089,-0.0983697921,-0.0384707935,-0.3954730332,-0.2417124212,0.2116261721,-0.0092253108,-0.1125635207,0.4255144,0.2137937844,-0.083978571,0.1484847665,0.2626099885,0.0930737779,0.3781287968,0.5176310539,0.0884295255,-0.1763887107,-0.2753964961,0.1769768298,0.1379491389,-0.1889899224,0.2807774246,-0.1882597059,-0.3576730788,0.0459904037,0.0207597371,-0.0589365773,0.1955636144,-0.259755373,-0.4056970775,-0.2798852921,0.2288470268,-0.0408817828,-0.0069536217,-0.04864024,0.3339018524,-0.0145018864,-0.0963488221,-0.3843574524,0.2725990415,-0.3716552556,0.0740748867,-0.3814620078,-0.3459908068,0.2668173015,0.0308048669,0.2160767615,0.1098613292,0.0036861273,-0.3170911372,-0.0461049266,0.0781307891,0.1460115761,0.1028930545,-0.015235885,0.0452596396,-0.3004338741,-0.1509357691,-0.1212889999,-0.055539526,-0.0096376119,0.1606694907,0.0099008707,0.0740238652,0.20944947,-0.1554741412,-0.0692407489,0.0523920283,-0.073882468,0.2532299161,-0.0490227118,0.0039508045,0.1122217774,0.2262326032,-0.0058491929,0.1865203679,0.2958050668,-0.162464872,-0.4413197041,-0.2185005248,0.3224136531,0.5740982294,-0.1763646603,0.05707068,-0.2007500678,0.2156663388,-0.4009667635,0.0981801525,0.2420495152,-0.1563884616,0.1629412621,0.1809740514,0.0274456404,-0.1156514958,-0.0125562027,0.0659356713,0.2097218782,-0.0765962973,-0.0596351214,0.2138530463,-0.016143905,0.0193220153,0.407071352,-0.0857663229,0.2633347809,0.3329678476,0.0312269777,0.0426279791,0.2017087191,0.0390785821,0.3710512221,0.1623593867,-0.0544953756,0.173710078,0.2876287699,-0.0220837481,0.1096094251,0.74064821,-0.0993698314,-0.1496194601,-0.0029150986,0.0964953899,-0.1920408905,-0.2609267235,-0.3624968529,-0.1506780684,-0.1066598147,-0.3490464091,-0.0708513036,-0.2164484411,-0.0891565904,-0.0799689069,0.1176134571,-0.1278524995,0.1915337145,0.0721953586,0.25208956,-0.0163424741,0.1791752428,0.191896528,-0.0074003758,-0.1781455725,0.1820492744,0.5045114756,0.1560384631,-0.4481222332,-0.0124125304,0.0752601326,-0.3575073183,0.0184038635,0.106576845,0.1137405708,-0.2378424704,0.3356722295,0.2290237993,-0.1850778013,-0.2165261954,0.222421512,-0.0761555955,0.1127024069,0.2996042371,-0.3052110076,-0.3296503425,-0.094468601,-0.0046699205,-0.2677340209,-0.2765890062,0.1584847271,0.0071122129,0.2048698217,-0.2126351893,0.1546743214,0.0693259239,0.5761985183,-0.0310886372,0.0948051661,0.0794112533,0.2102885246,-0.4790940881,0.1391242743,-0.1064219251,-0.1181201637,0.0687155053,0.4990251958,0.0630581155,0.1504907161,0.1508840472,0.0104325032,-0.1435123533,0.0968663096,-0.3981228769,0.0104692169,0.1829186082,0.0597024709,0.162882179,-0.2981536686,0.0647494644,-0.1292551905,0.1347406656,-0.0217997041,-0.2169077098,-0.027020989,-0.1069159508,0.2247478515,0.1091793403,0.1795613319,-0.0586220548,-0.2362982482,-0.1898047179,-0.1666766107,-0.2485951334,0.4792469144,-0.1454416364,0.2362560928,-0.0857738853,0.0941277295,0.1562325209,-0.0268115476,0.114797838,-0.134017095,-0.326020658,0.335285455,-0.0969309732,0.2485348582,-0.2594831586,0.3689167798,-0.1254217476,0.2203360498,-0.3546590805,-0.5474947691,0.5134512186,-0.3442061543,-0.3539212644,0.258849889,0.1708127558,0.2654415369,0.1027440578,-0.1676395684,-0.1582760662,0.2111672759,-0.1707343012,-0.3202186525,-0.0052573131,-0.1648396105,0.1009314954,-0.2636694014,-0.0430937521,0.005074298,0.1069141254,-0.2034539729,-0.3928065896]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/375","title":"TypeError when computing bertscore","comments":"I am not able to reproduce this issue on my side.\r\nCould you give us more details about the inputs you used ?\r\n\r\nI do get another error though:\r\n```\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/bert_score\/utils.py in bert_cos_score_idf(model, refs, hyps, tokenizer, idf_dict, verbose, batch_size, device, all_layers)\r\n 371 return sorted(list(set(l)), key=lambda x: len(x.split(\" \")))\r\n 372 \r\n--> 373 sentences = dedup_and_sort(refs + hyps)\r\n 374 embs = []\r\n 375 iter_range = range(0, len(sentences), batch_size)\r\n\r\nValueError: operands could not be broadcast together with shapes (0,) (2,)\r\n```\r\nThat's because it gets numpy arrays as input and not lists. See #387 ","body":"Hi, \r\n\r\nI installed nlp 0.3.0 via pip, and my python version is 3.7.\r\nWhen I tried to compute bertscore with the code:\r\n```\r\nimport nlp \r\nbertscore = nlp.load_metric('bertscore') \r\n# load hyps and refs \r\n...\r\nprint (bertscore.compute(hyps, refs, lang='en'))\r\n```\r\n\r\nI got the following error.\r\n```\r\nTraceback (most recent call last):\r\n File \"bert_score_evaluate.py\", line 16, in <module>\r\n print (bertscore.compute(hyps, refs, lang='en'))\r\n File \"\/home\/willywsm\/anaconda3\/envs\/torcher\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 200, in compute\r\n output = self._compute(predictions=predictions, references=references, **metrics_kwargs)\r\n File \"\/home\/willywsm\/anaconda3\/envs\/torcher\/lib\/python3.7\/site-packages\/nlp\/metrics\/bertscore\/fb176889831bf0ce995ed197edc94b2e9a83f647a869bb8c9477dbb2d04d0f08\/bertscore.py\", line 105, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() takes 3 positional arguments but 4 were given\r\n```\r\n\r\nIt seems like there is something wrong with get_hash() function?","comment_length":91,"text":"TypeError when computing bertscore\nHi, \r\n\r\nI installed nlp 0.3.0 via pip, and my python version is 3.7.\r\nWhen I tried to compute bertscore with the code:\r\n```\r\nimport nlp \r\nbertscore = nlp.load_metric('bertscore') \r\n# load hyps and refs \r\n...\r\nprint (bertscore.compute(hyps, refs, lang='en'))\r\n```\r\n\r\nI got the following error.\r\n```\r\nTraceback (most recent call last):\r\n File \"bert_score_evaluate.py\", line 16, in <module>\r\n print (bertscore.compute(hyps, refs, lang='en'))\r\n File \"\/home\/willywsm\/anaconda3\/envs\/torcher\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 200, in compute\r\n output = self._compute(predictions=predictions, references=references, **metrics_kwargs)\r\n File \"\/home\/willywsm\/anaconda3\/envs\/torcher\/lib\/python3.7\/site-packages\/nlp\/metrics\/bertscore\/fb176889831bf0ce995ed197edc94b2e9a83f647a869bb8c9477dbb2d04d0f08\/bertscore.py\", line 105, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() takes 3 positional arguments but 4 were given\r\n```\r\n\r\nIt seems like there is something wrong with get_hash() function?\nI am not able to reproduce this issue on my side.\r\nCould you give us more details about the inputs you used ?\r\n\r\nI do get another error though:\r\n```\r\n~\/.virtualenvs\/hf-datasets\/lib\/python3.7\/site-packages\/bert_score\/utils.py in bert_cos_score_idf(model, refs, hyps, tokenizer, idf_dict, verbose, batch_size, device, all_layers)\r\n 371 return sorted(list(set(l)), key=lambda x: len(x.split(\" \")))\r\n 372 \r\n--> 373 sentences = dedup_and_sort(refs + hyps)\r\n 374 embs = []\r\n 375 iter_range = range(0, len(sentences), batch_size)\r\n\r\nValueError: operands could not be broadcast together with shapes (0,) (2,)\r\n```\r\nThat's because it gets numpy arrays as input and not lists. See #387 ","embeddings":[0.2761255801,-0.1318785548,0.0756907165,0.5165124536,0.0988566577,-0.0892142057,0.1624636799,0.4776298404,-0.0689703971,0.1780642122,0.1973382682,0.4109018445,-0.2279064208,-0.2859780788,-0.0107829962,-0.4891375601,-0.0108120758,0.2576285899,0.0752307326,-0.1934443265,-0.4337232411,0.0454928912,-0.305316776,0.2282249331,0.2554677129,0.0886286646,-0.2604391873,0.1348609179,-0.2278603017,-0.4158537388,0.3323247433,-0.1785545796,0.0079615479,0.2735729516,-0.0001196758,-0.164226532,0.3422837853,-0.0503761508,0.1353789568,-0.296705693,-0.5725338459,-0.3105911314,-0.0319616012,-0.3177138269,0.2614753842,0.1489257812,-0.1354343146,-0.0481011495,0.0123424977,0.1812463403,0.1339588463,0.0254176091,-0.2118637115,0.0177610274,0.2604244053,0.0071910596,0.018549338,0.3635784686,-0.4025587738,-0.3008162379,0.0149126882,0.3288400471,0.0135539277,0.0823643878,0.2077835202,0.2061754912,0.0641864613,0.1880900711,-0.099902384,0.0908538997,-0.3608181179,0.0836356282,-0.3250648081,0.2343546301,0.0563024618,-0.4785050154,0.2022828162,-0.1153278351,-0.0014472344,-0.5615645051,-0.2783876956,-0.1005957797,-0.3167917728,0.1922527552,0.335963428,-0.1287992448,-0.1018865183,0.1736433208,0.4988453686,-0.1260612458,-0.1148243397,-0.1357129812,0.1316629201,0.4445529878,-0.0911608711,0.1092811972,0.1363413036,0.1133350655,-0.0303328838,-0.5219548345,0.0719118044,-0.0270394236,-0.3086431623,0.2734445333,0.0143340947,0.4041199982,0.0988553092,0.2138057202,-0.1712890267,0.0942789614,0.1294122785,-0.1019419357,0.0056878072,-0.1329765767,0.2726546824,0.41658476,-0.2363314182,-0.1184179634,0.1208985001,-0.0305107329,-0.5161265135,-0.0156838633,0.2908795774,0.3036076128,-0.031806089,0.375775218,0.0542506538,0.1252973229,-0.2185616344,0.0710509196,-0.1651147604,0.4736360908,-0.4104004204,-0.096329473,0.2049661279,-0.2787081301,0.2501209974,0.0212142337,0.1306795478,0.0555527583,-0.232017979,-0.2523785233,0.1263786256,-0.1410631984,0.0272896793,-0.0139504671,0.2995972931,0.2706362307,-0.1369116604,0.0181666724,-0.4476976693,-0.3143444657,0.1559612304,0.09911374,0.1796132475,0.1577753723,-0.2132614106,-0.0983943343,0.2718791962,-0.2792686522,-0.2460840195,-0.3767074347,-0.1823873967,-0.2621821165,0.1168278083,0.2059371173,-0.1212450862,0.0793723911,-0.334472537,0.3835711479,0.4181898832,0.25747177,-0.2473279387,0.0452045463,-0.0492055863,0.3692497611,0.5125999451,-0.5257790089,-0.0867503434,-0.1490114927,0.3625636697,-0.3087703288,-0.0399736166,-0.0240799952,-0.3384366333,0.0245214477,0.0581522286,-0.1452075392,-0.1038011536,0.0133661851,-0.207016781,0.0359151326,0.2369994968,0.0171717443,0.2854468524,-0.0981566757,-0.0882567093,0.3454071879,-0.1192731261,-0.1363748908,0.0345535725,0.050476633,0.3094350994,-0.2021875232,0.2588392794,-0.1655931473,0.109282732,0.0673473254,-0.4491257966,0.2856458724,-0.031146612,0.2148447782,-0.1701603532,-0.1016705185,-0.0813260823,0.00450397,0.0332289487,0.0805325657,0.2835669518,0.0595616512,-0.1374049336,-0.3083928823,-0.0790306404,-0.2511029243,-0.2863682508,0.039926257,-0.357680887,-0.1113914028,0.2529581189,0.4223967195,0.4608678222,0.2282650322,-0.1752714217,0.192355141,0.0128553184,-0.2530879378,0.2528431416,0.0289127473,-0.0318939537,-0.2653280497,0.1053089499,-0.0803646669,0.3789343536,-0.0778161585,0.1700160503,0.3666506708,0.0215997268,0.3864731193,0.058427494,0.1290577799,0.1458586156,0.1078662798,0.1775075793,-0.1183460057,0.3007713854,0.0703478456,-0.2067730576,-0.2661764324,0.2244000733,-0.3974009454,0.5296509862,0.1097301766,-0.0443881415,0.3107477725,0.0447515436,-0.0797535256,-0.3063564599,-0.5760126114,0.4189803898,-0.1433951408,-0.0812470615,-0.0095235929,-0.0576681644,-0.0341157988,0.1319587827,0.1455872804,-0.151384145,-0.1614838094,0.0986194164,-0.0865082294,0.0416444801,-0.2019615769,0.0317007899,0.5163311958,-0.3323081136,0.4592030942,-0.2828938663,-0.211226508,-0.271343112,-0.6009112,-0.3930343986,-0.4392984807,-0.061124675,-0.0487508886,0.1180296987,0.211656481,0.1126753166,0.3243050575,0.213519901,0.1973182112,-0.1149031147,-0.226452142,-0.4213554561,-0.0705223903,-0.1354673356,0.0677993298,0.1510768384,-0.1498572677,-0.1466472298,-0.0514496788,-0.2478653193,0.0892843381,-0.3712442219,0.0114818271,0.4287257791,-0.2520729899,-0.5686548948,-0.1865833253,0.411308229,-0.2112159282,-0.136223346,0.6053295732,-0.104811132,0.0049573975,-0.1419307888,-0.2325833738,0.1631800234,-0.0266358219,0.2447889596,0.1147983596,0.3527388871,0.0608623438,-0.0426308624,0.3670552373,-0.2433395982,-0.0545858815,-0.2454433143,-0.2338038385,0.1739595383,-0.1633548141,-0.3638004065,-0.2329206914,-0.0839452669,0.2205966562,0.0326967537,-0.0719918832,-0.6988813281,0.1609139889,-0.087573573,-0.113075532,0.3430009484,0.2070798278,-0.0264376942,-0.0185622331,-0.2006419897,-0.0735509619,0.2997160256,0.1567745954,0.2876432836,0.0552392825,0.2681938112,0.210577786,0.7750529647,0.41103971,-0.1427959055,0.4108362496,0.0980710983,-0.0665929839,0.1418499202,-0.4568258524,0.0531860776,0.0878207758,0.2652163804,0.1736506671,-0.1130234897,-0.0135813169,-0.1446095556,-0.1400755793,-0.4488294423,-0.2586409152,0.1118821874,-0.2477547377,0.501953125,0.0393708423,0.0282560196,-0.401119411,-0.2987891436,0.0437128805,-0.1346630603,0.1600886881,-0.0850046426,0.319912225,-0.4315064251,-0.2609436214,0.4325511158,0.5315396786,0.4602389634,-0.0440842025,0.2140751481,0.0286206212,0.0435654633,0.391659528,-0.2287540585,0.1957283765,0.1939029396,0.0125674317,-0.2409177721,0.0994615182,-0.4506720304,0.0259025525,0.2522263527,0.196654588,-0.1824195981,0.0829214677,0.2392496914,-0.1913156956,-0.2168228179,-0.1717543304,-0.3922537863,-0.1034556478,0.1337929964,0.0708656386,-0.1848283112,0.3278142512,0.1164414436,0.342713654,0.0810552239,-0.0545408539,0.160566479,0.0525477193,-0.015751224,0.0007018874,0.0075159906,0.0105600357,-0.0969670564,0.1803432703,0.0753133073,-0.0390518494,-0.3406938612,-0.0114943041,0.1553639621,0.5570816994,0.0420730747,-0.025938645,0.0074095093,0.0348688997,0.0451917537,-0.097172834,0.1141790971,0.65434587,-0.1623400897,-0.2744380832,-0.3511208296,0.5985896587,0.1391275227,-0.0441228785,-0.1602846682,0.3554715216,-0.109679319,0.2573424876,0.407900095,0.9862613082,0.0076943128,0.0063673132,0.39518857,0.3146586418,0.6596641541,0.1578849256,0.243186906,-0.4803039432,0.0930513069,-0.0470333174,-0.1843595654,-0.0433024988,-0.143394053,-0.1519216299,0.5336700678,-0.1822413504,-0.0090792608,0.0595018603,-0.0926761329,0.1497077942,-0.3648533225,-0.2230275124,0.092977114,-0.2057846487,0.1314042658,-0.1327215284,-0.1358576566,0.0986191481,-0.3095805943,-0.3487919271,0.2200475931,0.3045174778,0.0868370757,-0.0660902485,-0.0008576434,0.1731225252,-0.0144815734,0.3831582963,-0.0302862506,-0.1866534352,-0.0229079891,0.0708984137,-0.0881785825,0.1204803139,0.2667514384,-0.1431594938,-0.127994597,-0.033028625,0.117567867,-0.0159538891,-0.2441244125,-0.0770286992,0.0197639428,0.1707286835,-0.0075826221,0.0523209907,-0.034711998,0.0013457417,-0.0325207449,0.0771799237,0.0615729429,-0.3198444247,0.2046662718,0.0610937998,-0.1008357033,0.2633227408,0.3545152545,0.1891111284,0.0707342103,0.433087945,0.0131017137,-0.24831976,-0.014343923,-0.2957587242,0.0910893306,-0.3135705292,0.0125327865,0.1495845914,0.257735312,0.0460842848,0.3942522109,0.0963944197,0.0542419218,0.0115344385,-0.3625415266,0.0454562493,0.0628629625,-0.2470956147,-0.2727504075,0.0383835658,0.1021152213,-0.1322076172,0.3776407838,-0.2192404121,0.0390757881,-0.3728845119,0.3584414423,-0.192639336,-0.0827742666,0.2540491521,0.1947726458,0.0388453379,0.3014284074,-0.1767047793,-0.0731121078,-0.2100672424,0.1635114998,0.131047368,-0.1000208035,-0.143018499,-0.3009257913,-0.2370301485,-0.0528404899,-0.5451624393,0.0936053395,-0.0505118035,0.5203589797,-0.0175810829,0.209358722,0.2586188316,0.1882200539,-0.0817689225,0.3411987126,-0.050276842,0.1644684523,0.0733920112,-0.1656045616,-0.1188354567,0.0411758311,0.143288821,-0.0187748112,0.2660484612,-0.0018674174,0.3188217282,-0.2743104398,0.3269651532,0.140278101,-0.0391174182,-0.056872379,0.3071949184,0.0340563729,-0.1577091366,-0.016701417,-0.3050073683,0.164696455,-0.2107364535,0.1912527233,-0.1915954947,-0.0024428389,0.0365468487,-0.2161626071,-0.0961601213,-0.2499932498,-0.105124034,0.240621537,-0.1562424004,0.0950379595,0.3902393878,0.1090631634,0.1705679744,0.2774473727,-0.1277304292,0.2209457159,-0.368193835,-0.0341564454,-0.111566782,-0.0962208062,-0.4266493917,0.3289455175,0.3853102326,0.2180593908,-0.2089059949,0.619712472,-0.2494833469,0.0003625874,-0.4004915357,0.4326594472,0.1480334401,-0.0509177484,-0.228865236,-0.12701796,-0.2422426641,0.024107093,-0.0477326512,0.1367578357,0.491625309,0.4510845244,-0.0505795442,-0.2604432404,-0.3250429034,0.5483903289,0.2560437918,-0.3059969842,-0.0033517773,0.1129801795,0.0017465089,-0.0900233537,0.3348165452,0.2629611492,0.405659765,-0.2561149001,-0.2389256507,0.0008821391,0.0525389612,0.0479513705,0.2962498069,0.3155489862,0.3683305681,0.1411130875,-0.0089415619,-0.0464185178,0.1642612517,-0.0995212197,0.5115295053,0.0894530267,0.8424601555,-0.0701090917,0.2200593352,-0.0643104687,0.0433491208,-0.4138903618,0.2294117063,-0.0224551577,0.4532651007,0.1220006123,-0.0497545898,0.0311040431,-0.0103857825,0.5032489896,0.3842631876,-0.1603479087,-0.6217654943,-0.0705332831,-0.1833683848,0.082519941,-0.0731541663,0.1935972124,-0.0136903189,0.0601201579,-0.1586876959,0.1457274705,-0.1382827908,-0.1218959168,-0.2110373229,0.2093118727,-0.2460129857,-0.2201060355,-0.2454258204,-0.4443536699,0.0382015444,-0.0276737977,-0.0358009711,-0.2291761041,0.0267046951,-0.1150477827,-0.2235094905,0.0858788192,0.5075319409,0.2599181831,0.3871493638,0.3245076835,0.1401470304,-0.2034894526,0.0518977568,-0.1327571571,0.0878442004,0.0638767779,-0.3981203735,0.1798437089,-0.1646803021,0.0677561238,-0.5098586679,0.3263379931,-0.163345933,-0.1433932781,-0.3805131614,0.1197051853,0.0851338282,-0.0169912763,-0.1053539813,0.0401982479,-0.0319031514,0.3425037861,-0.4504338205,-0.0074978955,0.6227499843,-0.5898154378,-0.3655669689,0.1650718451,0.2566818893,0.0099781556,-0.1730729192,-0.8751227856,-0.1224634871,0.3599118292,0.1091388837,-0.1743117124,-0.1304674298,-0.2308664024,-0.030322941,0.0319113731,-0.0329074226,-0.0787345544,-0.147003442,0.314570725,-0.0969066769]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/375","title":"TypeError when computing bertscore","comments":"The other issue was fixed by #403 \r\n\r\nDo you still get this issue @willywsm1013 ?\r\n","body":"Hi, \r\n\r\nI installed nlp 0.3.0 via pip, and my python version is 3.7.\r\nWhen I tried to compute bertscore with the code:\r\n```\r\nimport nlp \r\nbertscore = nlp.load_metric('bertscore') \r\n# load hyps and refs \r\n...\r\nprint (bertscore.compute(hyps, refs, lang='en'))\r\n```\r\n\r\nI got the following error.\r\n```\r\nTraceback (most recent call last):\r\n File \"bert_score_evaluate.py\", line 16, in <module>\r\n print (bertscore.compute(hyps, refs, lang='en'))\r\n File \"\/home\/willywsm\/anaconda3\/envs\/torcher\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 200, in compute\r\n output = self._compute(predictions=predictions, references=references, **metrics_kwargs)\r\n File \"\/home\/willywsm\/anaconda3\/envs\/torcher\/lib\/python3.7\/site-packages\/nlp\/metrics\/bertscore\/fb176889831bf0ce995ed197edc94b2e9a83f647a869bb8c9477dbb2d04d0f08\/bertscore.py\", line 105, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() takes 3 positional arguments but 4 were given\r\n```\r\n\r\nIt seems like there is something wrong with get_hash() function?","comment_length":15,"text":"TypeError when computing bertscore\nHi, \r\n\r\nI installed nlp 0.3.0 via pip, and my python version is 3.7.\r\nWhen I tried to compute bertscore with the code:\r\n```\r\nimport nlp \r\nbertscore = nlp.load_metric('bertscore') \r\n# load hyps and refs \r\n...\r\nprint (bertscore.compute(hyps, refs, lang='en'))\r\n```\r\n\r\nI got the following error.\r\n```\r\nTraceback (most recent call last):\r\n File \"bert_score_evaluate.py\", line 16, in <module>\r\n print (bertscore.compute(hyps, refs, lang='en'))\r\n File \"\/home\/willywsm\/anaconda3\/envs\/torcher\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 200, in compute\r\n output = self._compute(predictions=predictions, references=references, **metrics_kwargs)\r\n File \"\/home\/willywsm\/anaconda3\/envs\/torcher\/lib\/python3.7\/site-packages\/nlp\/metrics\/bertscore\/fb176889831bf0ce995ed197edc94b2e9a83f647a869bb8c9477dbb2d04d0f08\/bertscore.py\", line 105, in _compute\r\n hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)\r\nTypeError: get_hash() takes 3 positional arguments but 4 were given\r\n```\r\n\r\nIt seems like there is something wrong with get_hash() function?\nThe other issue was fixed by #403 \r\n\r\nDo you still get this issue @willywsm1013 ?\r\n","embeddings":[0.2904174626,-0.0922274441,0.0491558611,0.5237545371,0.1049919277,-0.0615841709,0.1349087059,0.499720186,-0.0832763538,0.1573421359,0.2311001867,0.4060015678,-0.2866458297,-0.301553905,0.0494342037,-0.5268012881,-0.0282650422,0.27926898,0.0992682874,-0.1411128789,-0.4561849236,0.0702061653,-0.3064157665,0.2097167224,0.2271009684,0.0845934302,-0.2209957838,0.1212881207,-0.2142444998,-0.4208644927,0.3191001713,-0.194087714,-0.0712139383,0.2431201041,-0.0001182109,-0.1601535976,0.3708927035,-0.0338775143,0.1975390166,-0.3093294501,-0.5463821292,-0.3163008392,0.0074842069,-0.3180601895,0.2296390086,0.1989477277,-0.112964727,-0.0866039842,-0.0095640626,0.1640895307,0.1599534005,0.0035791383,-0.2369049042,0.0091188094,0.2670991421,-0.11018271,0.0136352638,0.3905308247,-0.3097732961,-0.2876586914,0.011288045,0.324162513,0.0116905933,0.1009009555,0.2086823285,0.1528432518,0.1729765087,0.1698229611,-0.0829740167,0.0949732512,-0.3707225919,0.0938167647,-0.2856841683,0.2020276785,0.0801035315,-0.4766536653,0.2819564044,-0.0696065128,0.009785804,-0.5507907867,-0.3322661817,-0.1652544737,-0.2755975723,0.1779367328,0.3377323151,-0.0968437791,-0.09147273,0.1587177962,0.4954340458,-0.1088161692,-0.0970662609,-0.0586758815,0.165763244,0.3854787946,0.0168094542,0.1307317764,0.1688496917,0.1689506322,-0.0501428694,-0.4577011466,0.11792133,0.0090298289,-0.2220349461,0.279273361,-0.0190768838,0.4005976617,0.0737718642,0.1264384538,-0.1431999356,0.136749059,0.0988060161,-0.0848337635,0.0501186624,-0.1476552337,0.2357533425,0.422344923,-0.1919434667,-0.0683202446,0.1046824083,-0.0409332886,-0.5537745357,-0.0424256325,0.301558435,0.2358010709,-0.0160895195,0.3854562342,0.0532154851,0.1406478882,-0.2008283436,0.0827843398,-0.1688016802,0.4461547434,-0.370349139,-0.1081425399,0.1836819351,-0.2481914014,0.2876333892,-0.0100123789,0.1286651343,0.0591965877,-0.191159606,-0.1948984414,0.084547095,-0.1183156222,0.0110450936,-0.0944863856,0.2513407767,0.2534931898,-0.1359507442,0.0155865336,-0.4096572399,-0.3133074343,0.1361920983,0.1212773398,0.1210442334,0.1549419165,-0.243530184,-0.0692962855,0.2792483866,-0.2679221034,-0.2240694761,-0.3416415453,-0.225735724,-0.2566771507,0.088434197,0.1327305883,-0.0511343554,0.0854187459,-0.3229417801,0.347678721,0.3498413265,0.2215134054,-0.2589527369,0.0764771104,-0.0406301655,0.341697067,0.5219309926,-0.5296539068,-0.1204511374,-0.2014780641,0.2880837023,-0.275811255,-0.0678644702,0.0041945791,-0.3556537628,0.0032597047,0.1058284417,-0.1367780566,-0.0745484903,0.0147093041,-0.2032391578,0.0608301833,0.3075386584,-0.0227830857,0.2435222566,-0.0948715582,-0.1435382962,0.3874900639,-0.1213771403,-0.1253415048,0.0303636417,0.0464051403,0.2779480219,-0.219672516,0.218858704,-0.2309199125,0.168104738,0.0656901971,-0.5022539496,0.2760654688,-0.0608187802,0.2195712477,-0.1498693377,-0.1183573008,-0.0835081935,0.0130383708,0.0618601553,0.0954452306,0.3256824911,0.1070378944,-0.1430049539,-0.3868094385,-0.0603471212,-0.2667038143,-0.2506922781,-0.0039754082,-0.3185167015,-0.113461189,0.2765190303,0.4354084134,0.4271657467,0.2068095803,-0.1607598513,0.1878409833,-0.0003167735,-0.2021586299,0.2429588139,-0.00015972,-0.0145691996,-0.3201680779,0.0966452733,-0.0420160294,0.3492535949,-0.0786416903,0.2048713565,0.3339096308,0.0532233678,0.287530452,0.0433361381,0.1540598422,0.1711477637,0.093458131,0.1405216157,-0.1626294404,0.31131199,0.0719347149,-0.2037606984,-0.2712045014,0.2676585913,-0.4321385026,0.5544569492,0.1619329453,-0.0469177701,0.2461349517,0.0476793349,-0.0994152799,-0.3072475195,-0.5435811877,0.400203377,-0.1430899948,-0.0402532518,0.043080084,-0.0578273907,-0.0765387565,0.1403026283,0.1780820638,-0.1604284495,-0.2206779867,0.1154613569,-0.1079413742,0.0269161798,-0.1568041742,0.0804962069,0.5080613494,-0.2958153188,0.4236321747,-0.2551444769,-0.1962884068,-0.2406973541,-0.5807244778,-0.3932805359,-0.4310674369,-0.0113548916,-0.0045826975,0.099211432,0.1984135807,0.0511892028,0.2784143686,0.1785910875,0.1507310271,-0.0842516422,-0.2052287459,-0.3607750833,-0.0485028476,-0.0848740861,0.053147424,0.1771374941,-0.1938165575,-0.1417104304,-0.0696578771,-0.2714033723,0.0386705101,-0.3793825507,-0.0139888357,0.4401356578,-0.2940277159,-0.5850625038,-0.1560530066,0.4484012127,-0.2225213498,-0.1439946443,0.6108608842,-0.0215119403,0.0106681343,-0.1866715848,-0.2387656122,0.2211546749,-0.0234717634,0.3290148079,0.1312876791,0.341485709,-0.0180408228,-0.0968374908,0.3998994529,-0.2825091481,-0.0100696124,-0.2524374723,-0.2528705597,0.1450577527,-0.1340660155,-0.3728225231,-0.1677078605,-0.0841820166,0.2155767381,0.0468060188,-0.0700210929,-0.7259473801,0.2088816166,-0.1023848429,-0.0793779492,0.2624851167,0.2453421503,0.0138137545,-0.035339646,-0.2105401456,-0.0568328314,0.2768721282,0.1620008051,0.3039762676,0.0552911013,0.213666454,0.2272766978,0.7062087059,0.4302788079,-0.0370992199,0.4145525098,0.0823720992,-0.0708407313,0.1631493419,-0.4111403227,0.0344700702,0.1248781607,0.2191670984,0.1446522623,-0.0755788237,0.0047113025,-0.1769392639,-0.1084539741,-0.4404537082,-0.2590355277,0.064926967,-0.3910019398,0.489174813,0.0428851694,0.059238445,-0.3648044169,-0.2747917771,0.0441347621,-0.1960714906,0.1525276601,-0.1175427511,0.2175270468,-0.4458345473,-0.2697966695,0.4120499492,0.5424811244,0.4436338842,-0.0267844591,0.2036388367,0.025666412,0.0461499281,0.3376170695,-0.2594157755,0.2039955109,0.2068466097,-0.0059946738,-0.2708322406,0.1021305323,-0.4152575731,0.0651294887,0.2587801516,0.2313803285,-0.1404029429,0.1396229416,0.1579517722,-0.1975784451,-0.2088772058,-0.1249715611,-0.4230959713,-0.1232615262,0.1620269269,0.0695640221,-0.1536773443,0.3219766915,0.1799845099,0.2988864183,0.1042262986,-0.0675371513,0.1358772665,0.0433299579,-0.0086181629,-0.0333843045,-0.0146427816,0.038694106,-0.1042148024,0.1601289064,0.0755576119,-0.0623714961,-0.3306590319,-0.0381685607,0.1952429712,0.5770746469,0.014597889,-0.0324257389,-0.0306300018,0.090681985,0.0097707501,-0.1054687127,0.0821258426,0.6564193964,-0.2147291005,-0.1859663725,-0.3386553228,0.6040155888,0.1938414723,-0.0250487905,-0.1772575229,0.295137316,-0.0963149816,0.159410134,0.3811167181,0.9107512236,-0.0567409955,-0.0188192595,0.3573856354,0.2828600109,0.6702302098,0.2404674739,0.2942899466,-0.4324535728,0.1500746012,-0.0468968414,-0.2103658468,-0.0601382628,-0.2091364563,-0.162483409,0.4873063862,-0.2053415924,-0.0277978778,0.0194977261,-0.0904640779,0.0549731553,-0.2829090655,-0.2512855232,0.1124667376,-0.2009394914,0.1265697628,-0.1388177872,-0.1479282975,0.1068110913,-0.2878459394,-0.3446026146,0.2231067419,0.2574718297,0.0734877735,-0.0972902998,-0.0625399575,0.1099501327,0.0299456231,0.2814872265,-0.0679924712,-0.23595801,-0.0104844747,0.0310497917,-0.1146194786,0.1915427744,0.2904157043,-0.1366613209,-0.1527976245,-0.0159626137,0.1245563403,-0.0534598008,-0.2983971536,-0.0273456499,0.046265021,0.1480565071,0.0486822687,0.0407931991,-0.0539817326,0.0236259606,-0.0713661686,0.1004901752,0.0896388069,-0.3393472135,0.1612416655,0.0679840073,-0.1096671596,0.3061047792,0.3814484179,0.1982176453,0.0351191722,0.3918955624,0.0097205089,-0.2221055329,-0.0135293026,-0.3323475122,-0.0119090807,-0.2608850896,0.0879888758,0.1544625908,0.2707075477,0.0250552408,0.4130760431,0.0934293494,0.0615525395,0.0124953594,-0.3384604752,-0.0131100444,0.0914130509,-0.2517139018,-0.2668474317,0.0026075318,0.1392206252,-0.0580950826,0.314291805,-0.2284548134,0.075353682,-0.421094507,0.3296049237,-0.2035216242,-0.0962959602,0.2431237102,0.2017073929,0.0531821325,0.3213964999,-0.2035358846,-0.0880871788,-0.1818605214,0.1454526335,0.1216410846,-0.1318777949,-0.0963667408,-0.3229756057,-0.2213948518,-0.0467833392,-0.5821325779,0.1147771776,-0.0581640266,0.438652128,-0.0182059724,0.213606596,0.2899671197,0.0935867801,-0.1665210724,0.3169226348,-0.102834411,0.1370310783,0.1014946923,-0.1448224336,-0.1064706594,0.0104751224,0.124786064,0.0098675219,0.2633093297,0.0152458213,0.279025197,-0.179580912,0.3222173452,0.1817023754,-0.0434454903,-0.0210500713,0.2634158134,0.0567513257,-0.162904337,-0.0152792335,-0.2783769071,0.2440703511,-0.2480600029,0.1466545612,-0.1684125364,-0.0403548069,-0.0134856766,-0.1936637908,-0.1232427806,-0.309635669,-0.076390259,0.2226924896,-0.1957081258,0.0773908123,0.4012205601,0.1408319175,0.1669455618,0.2967816293,-0.0915007368,0.2331999391,-0.3449114561,-0.0331933238,-0.126328662,0.00857771,-0.4746382236,0.3226793706,0.4144371748,0.2290356606,-0.2073529661,0.674331069,-0.3006381691,0.0727448761,-0.4104418755,0.4658389986,0.1382856071,-0.0586888231,-0.1774603277,-0.1243086234,-0.1171503067,0.0316428244,-0.0297730807,0.1733529419,0.561475575,0.4466930032,-0.0728768483,-0.2498514503,-0.3451193571,0.5180963874,0.24690409,-0.3214112818,0.0132671529,0.1317588389,-0.0143523905,-0.0793295503,0.370849669,0.2399627715,0.3532238305,-0.3193243146,-0.2778848708,-0.0294322819,0.0793329775,0.0295432359,0.3861568868,0.3933221996,0.4080020487,0.1068928912,0.0070009446,-0.0600250959,0.2305085361,-0.1380633712,0.5424201488,0.0750358999,0.845422864,-0.1254322678,0.1957302839,-0.1161361486,0.0179647468,-0.3747187853,0.1940066367,-0.0453919992,0.4868942797,0.1023416519,-0.075276792,0.0428331085,0.0224077888,0.4549942613,0.3861740828,-0.1753269583,-0.5789880157,-0.0966042578,-0.204747349,0.097757794,-0.1072829962,0.2111728787,-0.0452060588,0.0657955036,-0.2001926601,0.1688110828,-0.122718662,-0.1054828241,-0.2191365361,0.1316706985,-0.2063664198,-0.1440442502,-0.2046044618,-0.472034663,0.0835019052,-0.0070053176,-0.0418244563,-0.2637794614,0.0543336235,-0.1033514515,-0.2240255028,0.0139091117,0.5695141554,0.3293266892,0.4149047732,0.2977357805,0.1599027663,-0.217714414,-0.0053970804,-0.1010567918,0.1191025525,0.0614116117,-0.4052550495,0.1595237404,-0.1900942922,0.1848226339,-0.4926558137,0.385419935,-0.1808723211,-0.1175920963,-0.3977457285,0.0771403983,0.0841121301,-0.0042854962,-0.0959376246,0.0069856523,-0.0434136353,0.3032416701,-0.4454147816,-0.0250569992,0.6198911071,-0.5832110643,-0.410474956,0.1816491783,0.2914614081,-0.0032573589,-0.1798079908,-0.909907639,-0.153538242,0.3586063683,0.1226617172,-0.2015678436,-0.1121116355,-0.2358475178,-0.0396234319,0.0523301587,-0.0054568318,-0.1325996667,-0.1339805275,0.2951880693,-0.0739116445]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"I've seen this sort of thing before -- it might help to delete the directory -- I've also noticed that there is an error with the json Dataloader for any data I've tried to load. I've replaced it with this, which skips over the data feature population step:\r\n\r\n\r\n```python\r\nimport os\r\n\r\nimport pyarrow.json as paj\r\n\r\nimport nlp as hf_nlp\r\n\r\nfrom nlp import DatasetInfo, BuilderConfig, SplitGenerator, Split, utils\r\nfrom nlp.arrow_writer import ArrowWriter\r\n\r\n\r\nclass JSONDatasetBuilder(hf_nlp.ArrowBasedBuilder):\r\n BUILDER_CONFIG_CLASS = BuilderConfig\r\n\r\n def _info(self):\r\n return DatasetInfo()\r\n\r\n def _split_generators(self, dl_manager):\r\n \"\"\" We handle string, list and dicts in datafiles\r\n \"\"\"\r\n if isinstance(self.config.data_files, (str, list, tuple)):\r\n files = self.config.data_files\r\n if isinstance(files, str):\r\n files = [files]\r\n return [SplitGenerator(name=Split.TRAIN, gen_kwargs={\"files\": files})]\r\n splits = []\r\n for split_name in [Split.TRAIN, Split.VALIDATION, Split.TEST]:\r\n if split_name in self.config.data_files:\r\n files = self.config.data_files[split_name]\r\n if isinstance(files, str):\r\n files = [files]\r\n splits.append(SplitGenerator(name=split_name, gen_kwargs={\"files\": files}))\r\n return splits\r\n\r\n def _prepare_split(self, split_generator):\r\n fname = \"{}-{}.arrow\".format(self.name, split_generator.name)\r\n fpath = os.path.join(self._cache_dir, fname)\r\n\r\n writer = ArrowWriter(path=fpath)\r\n\r\n generator = self._generate_tables(**split_generator.gen_kwargs)\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n writer.write_table(table)\r\n num_examples, num_bytes = writer.finalize()\r\n\r\n split_generator.split_info.num_examples = num_examples\r\n split_generator.split_info.num_bytes = num_bytes\r\n\r\n def _generate_tables(self, files):\r\n for i, file in enumerate(files):\r\n pa_table = paj.read_json(\r\n file\r\n )\r\n yield i, pa_table\r\n\r\n```","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":191,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nI've seen this sort of thing before -- it might help to delete the directory -- I've also noticed that there is an error with the json Dataloader for any data I've tried to load. I've replaced it with this, which skips over the data feature population step:\r\n\r\n\r\n```python\r\nimport os\r\n\r\nimport pyarrow.json as paj\r\n\r\nimport nlp as hf_nlp\r\n\r\nfrom nlp import DatasetInfo, BuilderConfig, SplitGenerator, Split, utils\r\nfrom nlp.arrow_writer import ArrowWriter\r\n\r\n\r\nclass JSONDatasetBuilder(hf_nlp.ArrowBasedBuilder):\r\n BUILDER_CONFIG_CLASS = BuilderConfig\r\n\r\n def _info(self):\r\n return DatasetInfo()\r\n\r\n def _split_generators(self, dl_manager):\r\n \"\"\" We handle string, list and dicts in datafiles\r\n \"\"\"\r\n if isinstance(self.config.data_files, (str, list, tuple)):\r\n files = self.config.data_files\r\n if isinstance(files, str):\r\n files = [files]\r\n return [SplitGenerator(name=Split.TRAIN, gen_kwargs={\"files\": files})]\r\n splits = []\r\n for split_name in [Split.TRAIN, Split.VALIDATION, Split.TEST]:\r\n if split_name in self.config.data_files:\r\n files = self.config.data_files[split_name]\r\n if isinstance(files, str):\r\n files = [files]\r\n splits.append(SplitGenerator(name=split_name, gen_kwargs={\"files\": files}))\r\n return splits\r\n\r\n def _prepare_split(self, split_generator):\r\n fname = \"{}-{}.arrow\".format(self.name, split_generator.name)\r\n fpath = os.path.join(self._cache_dir, fname)\r\n\r\n writer = ArrowWriter(path=fpath)\r\n\r\n generator = self._generate_tables(**split_generator.gen_kwargs)\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n writer.write_table(table)\r\n num_examples, num_bytes = writer.finalize()\r\n\r\n split_generator.split_info.num_examples = num_examples\r\n split_generator.split_info.num_bytes = num_bytes\r\n\r\n def _generate_tables(self, files):\r\n for i, file in enumerate(files):\r\n pa_table = paj.read_json(\r\n file\r\n )\r\n yield i, pa_table\r\n\r\n```","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"Yes, deleting the directory solves the error whenever I try to rerun.\r\n\r\nBy replacing the json-loader, you mean the cached file in my `site-packages` directory? e.g. `\/home\/XXX\/.cache\/lib\/python3.7\/site-packages\/nlp\/datasets\/json\/(...)\/json.py` \r\n\r\nWhen I was testing this out before the #372 PR was merged I had issues installing it properly locally. Since the `json.py` script was downloaded instead of actually using the one provided in the local install. Manually updating that file seemed to solve it, but it didn't seem like a proper solution. Especially when having to run this on a remote compute cluster with no access to that directory.","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":96,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nYes, deleting the directory solves the error whenever I try to rerun.\r\n\r\nBy replacing the json-loader, you mean the cached file in my `site-packages` directory? e.g. `\/home\/XXX\/.cache\/lib\/python3.7\/site-packages\/nlp\/datasets\/json\/(...)\/json.py` \r\n\r\nWhen I was testing this out before the #372 PR was merged I had issues installing it properly locally. Since the `json.py` script was downloaded instead of actually using the one provided in the local install. Manually updating that file seemed to solve it, but it didn't seem like a proper solution. Especially when having to run this on a remote compute cluster with no access to that directory.","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"I see, diving in the JSON file for SQuAD it's a pretty complex structure.\r\n\r\nThe best solution for you, if you have a dataset really similar to SQuAD would be to copy and modify the SQuAD data processing script. We will probably add soon an option to be able to specify file path to use instead of the automatic URL encoded in the script but in the meantime you can:\r\n- copy the [squad script](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/squad\/squad.py) in a new script for your dataset\r\n- in the new script replace [these `urls_to_download `](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/squad\/squad.py#L99-L102) by `urls_to_download=self.config.data_files`\r\n- load the dataset with `dataset = load_dataset('path\/to\/your\/new\/script', data_files={nlp.Split.TRAIN: \".\/datasets\/train-v2.0.json\"})`\r\n\r\nThis way you can reuse all the processing logic of the SQuAD loading script.","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":117,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nI see, diving in the JSON file for SQuAD it's a pretty complex structure.\r\n\r\nThe best solution for you, if you have a dataset really similar to SQuAD would be to copy and modify the SQuAD data processing script. We will probably add soon an option to be able to specify file path to use instead of the automatic URL encoded in the script but in the meantime you can:\r\n- copy the [squad script](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/squad\/squad.py) in a new script for your dataset\r\n- in the new script replace [these `urls_to_download `](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/squad\/squad.py#L99-L102) by `urls_to_download=self.config.data_files`\r\n- load the dataset with `dataset = load_dataset('path\/to\/your\/new\/script', data_files={nlp.Split.TRAIN: \".\/datasets\/train-v2.0.json\"})`\r\n\r\nThis way you can reuse all the processing logic of the SQuAD loading script.","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"This seems like a more sensible solution! Thanks, @thomwolf. It's been a little daunting to understand what these scripts actually do, due to the level of abstraction and central documentation.\r\n\r\nAm I correct in assuming that the `_generate_examples()` function is the actual procedure for how the data is loaded from file? Meaning that essentially with a file containing another format, that is the only function that requires re-implementation? I'm working with a lot of datasets that, due to licensing and privacy, cannot be published. As this library is so neatly integrated with the transformers library and gives easy access to public sets such as SQUAD and increased performance, it is very neat to be able to load my private sets as well. As of now, I have just been working on scripts for translating all my data into the SQUAD-format before using the json script, but I see that it might not be necessary after all. ","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":156,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nThis seems like a more sensible solution! Thanks, @thomwolf. It's been a little daunting to understand what these scripts actually do, due to the level of abstraction and central documentation.\r\n\r\nAm I correct in assuming that the `_generate_examples()` function is the actual procedure for how the data is loaded from file? Meaning that essentially with a file containing another format, that is the only function that requires re-implementation? I'm working with a lot of datasets that, due to licensing and privacy, cannot be published. As this library is so neatly integrated with the transformers library and gives easy access to public sets such as SQUAD and increased performance, it is very neat to be able to load my private sets as well. As of now, I have just been working on scripts for translating all my data into the SQUAD-format before using the json script, but I see that it might not be necessary after all. ","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"Yes `_generate_examples()` is the main entry point. If you change the shape of the returned dictionary you also need to update the `features` in the `_info`.\r\n\r\nI'm currently writing the doc so it should be easier soon to use the library and know how to add your datasets.\r\n","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":48,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nYes `_generate_examples()` is the main entry point. If you change the shape of the returned dictionary you also need to update the `features` in the `_info`.\r\n\r\nI'm currently writing the doc so it should be easier soon to use the library and know how to add your datasets.\r\n","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"Could you try to update pyarrow to >=0.17.0 @vegarab ?\r\nI don't have any segmentation fault with my version of pyarrow (0.17.1)\r\n\r\nI tested with\r\n```python\r\nimport nlp\r\ns = nlp.load_dataset(\"json\", data_files=\"train-v2.0.json\", field=\"data\", split=\"train\")\r\ns[0]\r\n# {'title': 'Normans', 'paragraphs': [{'qas': [{'question': 'In what country is Normandy located?', 'id':...\r\n```","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":49,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nCould you try to update pyarrow to >=0.17.0 @vegarab ?\r\nI don't have any segmentation fault with my version of pyarrow (0.17.1)\r\n\r\nI tested with\r\n```python\r\nimport nlp\r\ns = nlp.load_dataset(\"json\", data_files=\"train-v2.0.json\", field=\"data\", split=\"train\")\r\ns[0]\r\n# {'title': 'Normans', 'paragraphs': [{'qas': [{'question': 'In what country is Normandy located?', 'id':...\r\n```","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"Also if you want to have your own dataset script, we now have a new documentation !\r\nSee here:\r\nhttps:\/\/huggingface.co\/nlp\/add_dataset.html","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":20,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nAlso if you want to have your own dataset script, we now have a new documentation !\r\nSee here:\r\nhttps:\/\/huggingface.co\/nlp\/add_dataset.html","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"@lhoestq \r\nFor some reason, I am not able to reproduce the segmentation fault, on pyarrow==0.16.0. Using the exact same environment and file.\r\n\r\nAnyhow, I discovered that pyarrow>=0.17.0 is required to read in a JSON file where the pandas structs contain lists. Otherwise, pyarrow complains when attempting to cast the struct:\r\n```py\r\nimport nlp\r\n>>> s = nlp.load_dataset(\"json\", data_files=\"datasets\/train-v2.0.json\", field=\"data\", split=\"train\")\r\nUsing custom data configuration default\r\n>>> s[0]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 558, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 498, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pandas().to_dict(\"list\"))\r\n File \"pyarrow\/array.pxi\", line 559, in pyarrow.lib._PandasConvertible.to_pandas\r\n File \"pyarrow\/table.pxi\", line 1367, in pyarrow.lib.Table._to_pandas\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 766, in table_to_blockmanager\r\n blocks = _table_to_blocks(options, table, categories, ext_columns_dtypes)\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 1101, in _table_to_blocks\r\n list(extension_columns.keys()))\r\n File \"pyarrow\/table.pxi\", line 881, in pyarrow.lib.table_to_blocks\r\n File \"pyarrow\/error.pxi\", line 105, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowNotImplementedError: Not implemented type for Arrow list to pandas: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n>>> s\r\nDataset(schema: {'title': 'string', 'paragraphs': 'list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>'}, num_rows: 35)\r\n```\r\n\r\nUpgrading to >=0.17.0 provides the same dataset structure, but accessing the records is possible without the same exception. \r\n\r\n","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":219,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\n@lhoestq \r\nFor some reason, I am not able to reproduce the segmentation fault, on pyarrow==0.16.0. Using the exact same environment and file.\r\n\r\nAnyhow, I discovered that pyarrow>=0.17.0 is required to read in a JSON file where the pandas structs contain lists. Otherwise, pyarrow complains when attempting to cast the struct:\r\n```py\r\nimport nlp\r\n>>> s = nlp.load_dataset(\"json\", data_files=\"datasets\/train-v2.0.json\", field=\"data\", split=\"train\")\r\nUsing custom data configuration default\r\n>>> s[0]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 558, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 498, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pandas().to_dict(\"list\"))\r\n File \"pyarrow\/array.pxi\", line 559, in pyarrow.lib._PandasConvertible.to_pandas\r\n File \"pyarrow\/table.pxi\", line 1367, in pyarrow.lib.Table._to_pandas\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 766, in table_to_blockmanager\r\n blocks = _table_to_blocks(options, table, categories, ext_columns_dtypes)\r\n File \"\/home\/vegarab\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/pyarrow\/pandas_compat.py\", line 1101, in _table_to_blocks\r\n list(extension_columns.keys()))\r\n File \"pyarrow\/table.pxi\", line 881, in pyarrow.lib.table_to_blocks\r\n File \"pyarrow\/error.pxi\", line 105, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowNotImplementedError: Not implemented type for Arrow list to pandas: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>\r\n>>> s\r\nDataset(schema: {'title': 'string', 'paragraphs': 'list<item: struct<qas: list<item: struct<question: string, id: string, answers: list<item: struct<text: string, answer_start: int64>>, is_impossible: bool, plausible_answers: list<item: struct<text: string, answer_start: int64>>>>, context: string>>'}, num_rows: 35)\r\n```\r\n\r\nUpgrading to >=0.17.0 provides the same dataset structure, but accessing the records is possible without the same exception. \r\n\r\n","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"Very happy to see some extended documentation! ","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":7,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nVery happy to see some extended documentation! ","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"#376 seems to be reporting the same issue as mentioned above. ","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":11,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\n#376 seems to be reporting the same issue as mentioned above. ","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/373","title":"Segmentation fault when loading local JSON dataset as of #372","comments":"This issue helped me a lot, thanks.\r\nHope this issue will be fixed soon.","body":"The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)","comment_length":14,"text":"Segmentation fault when loading local JSON dataset as of #372\nThe last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. \r\n\r\n```\r\ndataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n```\r\ncauses\r\n```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset json\/default (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0...\r\n0 tables [00:00, ? tables\/s]Segmentation fault (core dumped)\r\n```\r\nwhere `.\/datasets\/train-v2.0.json` is downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/.\r\nThis is consistent with other SQuAD-formatted JSON files.\r\n\r\nWhen attempting to load the dataset again, I get the following:\r\n```\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 6, in <module>\r\n 'json', data_files={nlp.Split.TRAIN: [\".\/datasets\/train-v2.0.json\"]}, field='data')\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 382, in download_and_prepare\r\n with incomplete_dir(self._cache_dir) as tmp_data_dir:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/contextlib.py\", line 112, in __enter__\r\n return next(self.gen)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 368, in incomplete_dir\r\n os.makedirs(tmp_dir)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/os.py\", line 223, in makedirs\r\n mkdir(name, mode)\r\nFileExistsError: [Errno 17] File exists: '\/home\/XXX\/.cache\/huggingface\/datasets\/json\/default\/0.0.0.incomplete'\r\n```\r\n\r\n(Not sure if you wanted this in the previous issue #369 or not as it was closed.)\nThis issue helped me a lot, thanks.\r\nHope this issue will be fixed soon.","embeddings":[-0.0172924567,0.2893957794,0.0609415546,0.0578706525,0.2236097753,-0.0165483262,0.3200395405,0.6101837754,-0.4010095596,-0.2130427957,-0.2422538549,0.771407783,0.1327875555,-0.4265963435,0.1489231437,-0.0715281218,0.0462259129,0.1267296821,0.2239412665,0.0536700115,-0.1744497716,0.3416509628,-0.2299684733,0.1320418268,-0.1667520404,0.2660649717,0.0065363278,0.5419237018,-0.0745723546,-0.7302021384,0.2977763414,-0.0093303947,0.1517449468,0.3450249732,-0.0001186113,0.1845317483,0.3318485618,-0.1096370071,-0.3517018259,0.0521292798,0.2805030644,-0.2414649427,0.3380265236,-0.1453841031,0.0627313927,-0.3119795918,0.0318316966,0.0812750906,0.6866552234,0.1553962231,0.0921009928,0.1377252787,0.0738054588,0.0523657314,0.0207765196,0.0678778663,0.0166039914,0.7200931311,0.5202925205,-0.1260128319,-0.3388995528,-0.0720849857,0.121271871,-0.0382684954,0.20485726,-0.0635562018,-0.0138488207,-0.1832702905,0.0524652265,0.1170204207,0.2416470051,-0.2380599082,-0.2328487635,-0.2157492042,0.0747866482,-0.2181084901,0.3773207963,0.2772295773,-0.13986063,0.1523420066,-0.0569971204,0.0344605073,-0.1918932199,0.0293105822,-0.2142735124,0.5192520618,0.1215123013,0.0416844822,0.1051236168,-0.1629457474,-0.1519261301,0.3172393739,-0.2625278533,0.3998276293,-0.3908263743,-0.1686170995,0.1120524853,-0.3626522124,0.2163050026,0.0901769698,0.1800890863,0.167204693,0.2139120996,0.1203705147,0.4853105545,0.0546770915,0.1718356609,0.2025324851,0.1526307762,-0.0196298789,-0.0259742644,0.2397748679,-0.1398817003,-0.2250073105,-0.2135159373,0.0280673709,-0.0268218834,-0.10946697,0.0148830935,0.0695176572,-0.3147706687,-0.0120260334,0.0509856232,0.0848620832,0.1312868148,0.1926079243,-0.250028193,0.204761222,-0.101454854,-0.1726957709,-0.0175752416,-0.026819814,-0.3066568077,-0.0206658803,0.4054740667,-0.1225604787,0.1185211167,-0.0812497288,-0.1116154194,-0.2269219011,-0.4334718287,-0.0438977554,0.0076662428,0.1496739089,0.0316801295,-0.0704863369,-0.036723651,-0.2581102848,-0.1803209335,0.212189436,-0.351188004,-0.2851627171,-0.5924198031,0.1183918193,-0.3227140605,-0.0053090458,0.1444022059,0.2050788999,0.0566594712,-0.4543268085,-0.137789458,-0.0893592536,-0.0031521304,-0.2295428067,0.1466828585,0.2523785532,-0.8450297117,-0.2359818965,0.1668373644,-0.0015070704,0.0830982402,0.5275310278,-0.351418823,0.2732034326,-0.1530969292,0.5261092186,0.4013668597,-0.2779783309,-0.3369089961,0.2427911609,-0.3994488418,-0.0413578488,-0.2372673005,-0.0655715317,0.2967054844,0.0823332146,0.3412532806,0.44938609,-0.1155516654,0.1483249068,-0.3325651586,-0.0804989934,0.1726427674,0.3601403534,-0.067415975,-0.1608242691,-0.0298587419,0.1449324936,0.4040019214,-0.1719319224,-0.0207536705,0.1606730372,0.1372898668,-0.0173203666,0.1661421508,-0.1080801859,-0.775840342,0.102051191,-0.2894922793,0.0807941109,-0.0411318801,0.006053275,-0.2767830789,-0.1135980338,-0.2061314285,-0.0540232696,0.0321170986,-0.160494104,-0.0348596834,-0.0682966858,-0.4652936161,-0.1567875296,-0.154756695,0.378049463,-0.1153785959,0.4039317071,-0.1302649975,-0.4008334279,0.0927924365,0.3672265112,0.1282616705,-0.2672620118,-0.0068155113,0.3295680285,-0.0804471746,0.1338739544,-0.1708227694,-0.2979542315,0.2476089597,-0.0857613161,-0.1292840242,-0.0635809377,-0.1022319719,-0.0883648023,-0.3246678412,0.3362679482,0.0197260007,-0.0326367095,-0.0567735732,-0.2471363395,0.2981203198,-0.0848109946,-0.1274032891,-0.0282969773,0.0172862969,0.1671073884,-0.0124944821,0.1766584069,-0.1765216589,-0.0144745773,0.369494915,-0.0651190057,0.0355697162,0.1873179078,0.1008718833,-0.2918989658,-0.024752669,-0.0487226285,0.3634798825,-0.007696704,-0.1558919847,-0.0039152685,0.0336674862,-0.0184020884,-0.011545361,0.0482457764,0.4226091802,0.5049403906,0.3639849424,0.1684012264,-0.2733654976,-0.1548733562,0.1276226789,0.0563049763,-0.3743780255,0.0602172576,-0.2678267956,-0.144593209,-0.2930882573,-0.1426913887,-0.3419747949,-0.1941553801,0.1088883728,-0.0382642075,-0.086782597,0.2005051821,0.1665448993,0.4000805914,0.1002783403,-0.1863676459,0.0381456204,-0.1772579998,-0.2182420641,-0.0453671627,0.3783227801,-0.0644285679,-0.2067784965,-0.1894416362,-0.2709756494,0.1097929254,-0.4360963702,0.1680292189,-0.0091555053,0.1945903599,0.0222651493,0.2368633002,-0.1479672343,-0.3669347167,0.2814826965,-0.0462120995,-0.2511283457,0.0004014499,-0.1034957469,0.1337920129,-0.2878107727,-0.3897340596,-0.2453112602,-0.2271138728,0.4359025359,0.0752960369,0.1372660846,0.2004070431,-0.202068001,0.0147434054,-0.2370598018,-0.2110955119,0.0494783483,-0.1857471168,0.2052339017,-0.2988070846,-0.1417259425,-0.0219020918,0.0942525193,0.1670048088,-0.2243585736,-0.1604168266,0.1161189228,0.1713583618,0.2460084856,-0.3234409392,-0.2597648203,0.2524788082,-0.2297296822,0.0524282791,0.1972107291,-0.3602097034,-0.0937124863,0.518563509,0.1982491612,-0.2657060623,0.2363282442,0.0597211421,0.8267533183,-0.1374402046,-0.162537694,0.1258324534,0.0303803496,0.1219486594,-0.1489123851,-0.0352488011,0.3921883106,-0.0397980995,-0.1129414514,0.2475955635,0.1624519825,-0.5424818397,-0.0059676873,0.1580976844,-0.5473148823,-0.1683234572,0.2632973492,0.5551204085,-0.085087575,-0.0429348014,0.0294524468,0.0029247708,-0.231551066,-0.0598361567,0.2242513895,0.2707134187,0.093233414,-0.5434257984,-0.3422010243,-0.0447595082,-0.0204031896,0.1243701354,0.4645565748,-0.0012731466,-0.2062437534,0.0272208229,-0.264184773,0.3476617634,0.2620742619,0.3518283069,0.3491418064,0.3293307722,-0.0743013844,-0.3046782911,0.145189181,0.4771124721,0.6556548476,0.4198312759,-0.1016693041,-0.1317106336,0.2848747075,0.0859613344,-0.1447812468,-0.073383823,-0.2265140861,-0.07947772,0.0030262177,0.0393503457,-0.0417612009,0.3973942995,0.0603224225,0.1337610185,0.42367661,-0.2695536017,-0.2426751405,0.3894139826,0.1091234982,0.0767814219,0.0468042009,0.2670255005,-0.0012031003,0.3386415839,0.4461458921,0.289424628,-0.2029463798,0.2748586237,-0.0069873552,0.2213015556,0.4691963792,0.2012590766,-0.1830905974,-0.1404711902,-0.1271867305,0.023022363,-0.4001841843,0.1378407776,-0.307553798,-0.2267783582,0.2155743688,0.2009165585,-0.095583871,-0.063385807,-0.0289977491,-0.0604570806,-0.0157040246,0.5079244971,0.1642167866,0.7734225988,-0.0452003293,0.1062064245,0.2713397741,-0.1177693158,0.401232779,-0.6511294246,0.3604255617,-0.4377222955,0.2696182728,0.1214250103,-0.1291204244,0.168038398,0.1842432022,0.0067205373,-0.0547789708,-0.147625342,0.29451105,-0.0257627424,0.3388976157,0.1799478382,-0.3059536815,-0.5036039948,0.0607923009,-0.3701075613,0.0345579162,-0.1490779966,-0.169402957,-0.0852131099,-0.1265627742,-0.2838003337,0.2769851685,0.1298918426,0.0143147791,0.5128045082,-0.2600142658,0.1598159075,0.3644202054,-0.1501441002,0.0535293594,-0.0029322898,0.361552,-0.1234305203,0.1117698103,0.2297596037,0.019884022,0.2879821062,-0.1601614803,-0.040681988,-0.0503899902,-0.0745755062,-0.2796954215,-0.0691923723,0.3941940665,0.1375084072,-0.1656415462,0.0129157789,-0.1245392114,0.1229193956,-0.0493018106,0.0542031564,0.0106456531,-0.1491626054,-0.0271749925,0.1586455107,-0.0698154643,0.0864368379,0.608911097,-0.1084498614,-0.319620043,0.3064624667,0.3026753664,0.0052801622,-0.2204806805,0.22126472,0.0504690558,-0.3886024654,-0.1376272291,-0.1989231855,-0.0684243813,0.0346806385,0.102228038,0.217286557,-0.1017905772,0.0036822893,-0.5508986712,-0.303109169,0.191187039,-0.0767692327,-0.181370005,0.1192315444,0.3389428258,0.0155134331,0.2547386885,-0.2034851313,0.3259330094,-0.3733081222,0.1441611499,-0.5635966659,-0.237191543,-0.0035354267,-0.2305135131,0.0180746578,0.2454891652,0.0734469891,-0.1074404418,-0.0155515783,0.1593936682,0.1383399069,-0.0180853177,-0.1694668531,-0.0002068877,0.1075090393,0.1735597849,0.1035003662,-0.1377709955,0.146763742,0.1609328389,0.2953796685,0.10649167,0.1960280687,0.0935889035,0.029574303,0.3905389011,0.0588820763,0.2412942648,-0.2039603144,0.033432588,-0.3554688394,0.1890608072,0.0100139622,0.2804839313,0.0838274881,-0.57982862,-0.0544024706,-0.0720525235,0.301358223,0.5721237659,-0.366515249,0.1124070734,0.1887703538,0.104174085,-0.0099351732,-0.2594088018,0.7478207946,-0.0356641449,-0.30246225,0.018036209,0.1428461671,-0.0983976647,0.1315189898,0.1389717311,0.5670432448,0.1919361502,0.1983341277,0.2170209289,-0.225434497,-0.0237511899,0.139786467,-0.0790000558,0.1816965044,0.254132539,-0.1939036101,0.0633177236,0.1178167611,0.1037081182,0.2329499274,-0.3413440585,-0.355749011,0.0997506082,0.1379030049,0.0106006861,0.1785005927,0.8641580343,-0.1908433586,-0.4217528105,-0.0853035226,0.0879023224,-0.3081797957,0.0873617306,-0.4969729483,-0.1789304167,-0.0405430794,0.1543643922,-0.30171296,-0.032923758,0.0855990499,0.0390480533,-0.2745316327,-0.1928051412,-0.0414176881,-0.2142160535,0.0656099617,-0.0752989128,0.3355386555,-0.1362095475,0.0428803414,0.2439544052,-0.0208171811,0.3282360733,0.3369998038,-0.1919885725,-0.126455605,-0.2578289807,0.0216807593,0.074430272,0.1027912199,0.214665249,-0.1302479059,0.2761435509,0.1168200597,-0.0468781255,0.1349721849,0.2768838704,0.1607480645,0.097309269,-0.0871051773,-0.1949210763,-0.2481790036,-0.0677022934,-0.1263933629,0.0338315964,-0.1226417944,-0.1555690169,0.5012949109,0.1177110523,-0.0630636439,0.0850425586,0.1083315387,0.5546950102,0.2226144671,-0.3615115285,-0.0978454202,-0.2026806325,-0.2905749083,0.2430060059,-0.0448605269,0.0714031458,0.3016998768,0.4403110147,-0.1458824575,0.2847689092,-0.2370343208,-0.0482020974,0.0044927313,0.245632574,-0.4291105866,0.179716602,0.0566488616,0.1454111487,-0.1059210822,-0.1340216696,0.2665147781,-0.2023610175,-0.0858606696,0.0350698419,-0.4277082086,-0.0206916574,0.0424795561,0.2701973617,-0.082025148,0.3326258659,0.1676799059,-0.3741495013,-0.416782409,-0.4191499352,-0.1761026829,0.1557717323,-0.127624318,0.4238348901,-0.0271939393,0.2040146142,-0.3149231672,-0.0030210139,-0.3587971926,0.3453436196,-0.3175049126,0.08617495,0.0861119926,0.1618037969,0.0088501237,0.3273944855,-0.0774755999,0.2494784147,-0.5126638412,-0.2689225972,0.4422735274,-0.6223313212,-0.2002592087,0.0687686652,0.0811640993,0.1981111467,0.0035751106,-0.2775660753,-0.0704600587,0.1616212279,-0.1745931953,-0.2950669229,-0.1357098073,-0.2625449002,-0.2551771402,-0.0769581124,-0.3599270582,0.1837785244,-0.0453099422,-0.2211906612,-0.2330328971]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/369","title":"can't load local dataset: pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries","comments":"I am able to reproduce this with the official SQuAD `train-v2.0.json` file downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/","body":"Trying to load a local SQuAD-formatted dataset (from a JSON file, about 60MB):\r\n```\r\ndataset = nlp.load_dataset(path='json', data_files={nlp.Split.TRAIN: [\".\/path\/to\/file.json\"]})\r\n```\r\ncauses\r\n```\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 9, in <module>\r\n [\".\/path\/to\/file.json\"]})\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 483, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 719, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/datasets\/json\/88c1bc5c68489f7eda549ed05a5a738527c613b3e7a4ee3524d9d233353a949b\/json.py\", line 53, in _generate_tables\r\n file, read_options=self.config.pa_read_options, parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 191, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\n\r\nI haven't been able to find any reports of this specific pyarrow error here or elsewhere. ","comment_length":16,"text":"can't load local dataset: pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries\nTrying to load a local SQuAD-formatted dataset (from a JSON file, about 60MB):\r\n```\r\ndataset = nlp.load_dataset(path='json', data_files={nlp.Split.TRAIN: [\".\/path\/to\/file.json\"]})\r\n```\r\ncauses\r\n```\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 9, in <module>\r\n [\".\/path\/to\/file.json\"]})\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 483, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 719, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/datasets\/json\/88c1bc5c68489f7eda549ed05a5a738527c613b3e7a4ee3524d9d233353a949b\/json.py\", line 53, in _generate_tables\r\n file, read_options=self.config.pa_read_options, parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 191, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\n\r\nI haven't been able to find any reports of this specific pyarrow error here or elsewhere. \nI am able to reproduce this with the official SQuAD `train-v2.0.json` file downloaded directly from https:\/\/rajpurkar.github.io\/SQuAD-explorer\/","embeddings":[-0.2248245478,0.2079820931,0.0130987065,0.1377785653,0.278103739,-0.2597190738,0.3169232011,0.4687684178,-0.1691406667,-0.149010852,0.0023774887,0.4203414321,0.0690719336,-0.017227646,0.1122415587,-0.1478196979,-0.2049249411,0.2665922344,-0.0051441933,-0.0658807755,0.0004076251,0.2411607355,-0.1398990303,0.2639468908,-0.1967165917,-0.0133888656,0.1871291995,0.489185065,-0.2591634989,-0.5631475449,0.3153218329,-0.296150893,0.0954629779,0.1106008142,-0.0001166556,0.3696361482,0.1890549362,-0.0164393801,-0.3757407367,-0.3551534712,0.2414884269,-0.490455389,0.4686816335,-0.1142355204,-0.0764247999,-0.3508143425,-0.2223610729,-0.0738732889,0.4803349078,0.1195221841,0.1514048278,0.4681633711,-0.1832385957,0.1065228358,0.1193087772,0.1260323972,-0.1624573022,0.4610108435,0.3759771287,-0.3450180292,-0.3874123096,-0.1435601264,-0.1003543884,0.3243615329,0.1929460168,0.0093747759,-0.0391238257,0.1259256303,0.1254607588,0.1766230315,0.3256267607,-0.3496564627,-0.0632260889,-0.0745448321,0.0622528456,-0.0295382254,0.1810942143,0.3203273416,-0.3467211425,0.1248235554,-0.306050688,0.1964052171,-0.1067041978,0.3012481332,-0.1913444847,0.3296273649,0.0628052801,0.297368139,0.1061380431,0.1481057107,0.1631402075,0.2127095461,-0.0343891941,0.2124329805,-0.2666904628,0.2280091494,0.1013871729,-0.1723196059,0.2379253805,0.1013728306,0.2416795492,0.042447485,0.2140950859,0.2787180543,0.3740185499,-0.0143485079,0.1454510838,0.076501362,-0.017595483,0.0274803136,-0.1041604877,0.1215523779,-0.1991192549,-0.3780224919,0.0324992426,-0.1924571395,0.2749885619,-0.187133342,-0.2933093011,0.0940950513,-0.4935642183,-0.0608110391,0.1227081344,0.276789993,0.2118470371,0.2285734266,-0.1052113771,0.2164437771,-0.0928959176,-0.1712169647,-0.2024862021,0.0592293888,-0.319909662,-0.1551751643,0.4050982296,-0.2470349222,0.2493386269,-0.0130813839,0.1818499863,0.009779647,0.1716861129,0.0250100493,0.2165830731,0.1760776043,0.0181108508,0.0982851386,0.171315372,-0.2304755002,-0.3545801938,0.3259263337,0.0253611729,-0.320799768,-0.315030843,0.0986372679,-0.1578409076,0.0982589722,0.1755226851,0.0540475734,0.1067288667,-0.106240876,0.0962656811,-0.1180267259,-0.0711897016,-0.2982014418,0.3827488422,0.2899155319,-0.953222096,-0.1176798344,0.1975107193,-0.0724640563,0.0969277993,0.5727161169,-0.5708695054,0.1833847016,-0.0560126379,0.2390382886,0.5799345374,-0.3536818922,-0.5187262297,0.4949888289,-0.3832627237,-0.1650160998,-0.1425090581,-0.0413008891,-0.0210614819,0.282671243,0.094546847,0.683675468,-0.1377247125,0.1846747398,-0.3309652209,-0.1655487716,0.2316760421,0.4529773295,0.0024090782,-0.355371207,0.0153144868,0.1048190445,0.0743056834,-0.0041734935,0.2683136165,0.2235309184,-0.1914619803,-0.2516146898,0.0803214014,-0.0126006631,-0.7056607008,0.0893205032,-0.0640129596,-0.0399806537,-0.2193442732,0.0276732985,-0.389359206,0.2925418317,-0.1446148455,-0.0062229387,0.0418271422,-0.0105411699,0.0670851246,-0.0903656334,-0.3402930796,-0.149033159,0.0774711519,-0.0015936849,0.0514434464,0.4117664993,-0.0129660284,-0.278139919,0.0033560798,0.3550055921,-0.0173281319,-0.1272228658,-0.1079458073,0.2835788131,-0.1293059438,0.1203492433,-0.0040456974,-0.3351640403,0.2203945816,-0.3095673323,0.1677345783,-0.0205300692,-0.0447345078,-0.1287044138,-0.220154658,0.403267175,-0.190984115,0.1956115216,-0.0854582712,-0.1480924189,0.2147214115,-0.1240264177,-0.0586371161,0.0097111287,0.1784366816,0.2656929195,0.5207040906,0.2706751823,-0.2427762002,0.0891448408,0.3627480268,-0.0371227339,-0.04015835,0.0438290164,-0.2223280221,-0.1289470792,0.1081068367,-0.106606029,0.1868596524,0.1889305115,0.0756322891,-0.1392150819,-0.2522187829,-0.2151123434,-0.0319145508,-0.0436208732,0.1300845146,0.4038740695,0.448564589,-0.138234511,-0.4100547135,-0.0738447234,0.23459059,0.2450695783,-0.2443244755,0.0373928584,-0.3147729039,-0.0499378555,-0.1750921458,-0.4401914775,-0.264534831,-0.199967891,-0.0551090539,0.2310739011,0.0131704817,0.1373821497,-0.0116336057,0.2456823736,0.100873366,-0.4298436642,-0.0119966045,-0.5101564527,-0.3695273399,0.0328288302,0.4673339128,0.438559413,0.1786762923,0.1814610958,-0.0701135993,0.0090797385,-0.0991811231,0.2115611881,0.0304584187,0.1755426973,0.0036325869,0.3344269693,-0.4671179652,-0.3113957047,0.6026245952,-0.0913662165,-0.2721755505,0.1993504167,-0.3250698447,-0.1124728322,-0.2069631219,-0.1498916447,-0.2904261947,-0.3914505243,0.3365614712,-0.0109002227,0.0356000885,-0.1289472133,0.1338442415,0.2838368416,0.0949847326,-0.2116737813,-0.0383752808,-0.4520311952,0.2738104463,-0.115242824,-0.2694824636,-0.071028322,0.0409649499,0.2816606462,0.062776126,-0.3481173217,0.0024619957,0.0731174126,0.5344305038,-0.1907640994,-0.1876478046,0.3679308295,-0.0579305328,0.0226893835,0.0716024786,-0.2526082098,0.1822912842,0.0439497717,0.0337406024,-0.175504297,0.4175980687,0.1020742804,0.9232445955,-0.2643375397,-0.1505668312,0.3155281246,-0.0246923175,0.0868570358,-0.1268964261,-0.3084769845,0.1649303287,0.0774956495,-0.0769030675,0.1023014933,0.0603703596,-0.3845144212,-0.0449143164,-0.0214512665,-0.2746353447,-0.099416025,0.3585598767,0.1034401506,-0.108390227,-0.010700753,-0.0802660361,-0.1179673746,-0.1918515861,-0.1134530008,0.0333905295,-0.0317989998,0.1877583414,-0.4592273831,-0.1170401797,-0.0875759348,-0.0683355033,0.3628225327,0.315877974,-0.0147822415,-0.1853874922,0.3408860564,-0.3085046113,0.5129215717,0.2342293561,0.192559883,0.2869602442,0.0705933198,-0.6148582101,-0.1541522592,0.1200243011,0.4119726419,0.4449229836,0.3248276114,-0.1001518816,-0.3170759082,0.12564978,0.1735467613,-0.0331000462,-0.2027922273,-0.3746901155,-0.3073224425,-0.4436307549,0.0817539543,0.0238223206,0.2499192953,0.4211018682,-0.0807263851,0.2893456221,-0.1060653105,-0.272272557,0.247982502,0.1461409628,-0.0633544698,0.0726812407,0.2435635477,0.3455856144,0.0976910666,0.6214662194,0.1869533211,-0.0988838747,0.1482156068,0.2310016453,0.3008661866,0.2950569391,0.1793146729,-0.0573054291,-0.1176065579,-0.0166513901,0.1620851904,0.1781002283,0.0401495062,-0.2675119936,0.1096058413,0.1865427941,0.2891855538,-0.2739403546,0.1674655527,0.5299274921,-0.0523521714,0.069254525,0.4074668288,0.1166573986,0.7771912813,-0.264470011,0.1479298323,0.5714970827,-0.2305463403,0.3164277375,-0.2754616439,0.0059591779,-0.5188856125,0.2529405057,0.0849900842,-0.239861384,-0.0096132262,0.3080280423,0.095092006,-0.1409421861,-0.1081847623,0.3453782499,-0.0026210255,0.220746085,0.0508506522,-0.2714333534,-0.2926777899,0.0353431068,-0.3116565943,-0.0429676063,-0.064911738,-0.1619161963,-0.096907042,-0.1001790538,-0.1564715803,0.4074653983,-0.0556710511,0.2410966903,0.3928505778,-0.4113485515,-0.0838277936,0.0472864322,-0.02236389,-0.1843993068,-0.1851106584,0.2453989238,0.0351632722,-0.0871742442,-0.0076094461,0.1267258078,0.2111590356,-0.1905830503,-0.3152864575,-0.0097105782,-0.110760957,-0.3773855567,0.0098416358,0.2698491514,0.1093739048,-0.0706365108,-0.2126791924,-0.1799738109,0.1643077284,0.0283878334,0.0598794259,-0.0270294957,0.1562407613,-0.2532791495,0.3292662203,-0.3607409,0.0462623872,0.4963206947,-0.101433821,0.164053604,0.5222313404,0.395812422,-0.1078451872,-0.2344770283,-0.0417825356,0.5676164031,-0.4813598096,-0.0808526054,-0.0654442012,-0.0923317298,0.1245621368,0.2647167146,0.3303147852,0.0091077928,-0.0937877223,-0.5132707953,-0.3523992002,0.2194441855,-0.1140954867,-0.0998067334,0.2764377892,0.5523682833,0.1198208034,0.3457675874,-0.2247915864,0.1964669228,-0.0767386258,0.1733497381,-0.29373312,-0.2949861586,0.0617094859,-0.0491167195,0.0436315909,0.0755816251,-0.1114885211,-0.2160479426,-0.0542553216,0.1099572033,0.1158914566,-0.1538788229,-0.1924220771,-0.1486531943,0.033608675,-0.0341766663,-0.0114123672,-0.0316841267,-0.0636330768,0.0092392042,0.1375732422,-0.0495583229,-0.1432165056,0.0238068402,-0.0251240805,0.0513093583,0.0436231419,0.2476351559,0.0279661119,-0.0172809381,-0.2315071225,0.0654815733,0.2410795689,0.1861478686,0.2242752016,-0.2875194252,-0.1363506019,-0.1797350198,0.1812481582,0.3907132447,-0.1477690637,-0.0057631144,0.0170549676,0.1338837296,-0.1501574665,-0.1573177129,0.262271136,0.0530038476,-0.0168350432,-0.0100825345,0.1297941357,-0.2084887028,-0.08111763,0.1575099975,0.3992906511,0.0401901715,-0.4732250869,0.3453002572,-0.2074123472,-0.1616266519,0.1562598497,0.0847399905,0.3768481314,-0.0452592634,-0.1026147753,0.1214384288,-0.1278783083,0.2560073733,-0.0429029129,-0.4255738556,-0.1164732426,0.2421922982,0.1526723951,-0.3106538951,0.2070242316,0.7946990132,0.2270723581,-0.3669965863,0.0756444409,0.1639969647,-0.1651429385,0.0323552154,-0.7381464243,-0.1738948971,-0.1725956202,0.180132702,-0.1772890836,-0.0201254096,0.3450933695,0.1026235744,-0.0690098554,-0.4073937833,0.0018845239,0.1002660021,-0.3239278793,-0.237759307,0.4012735784,0.1469526887,-0.1587483436,-0.0346477181,-0.128130272,0.3428938985,0.2950404584,-0.3094574511,-0.1638810784,-0.0877760798,-0.1351440102,-0.2076944262,0.1110962331,0.1466931105,0.1099853516,0.3763891459,0.1300737709,-0.1291117966,0.0981806517,0.1858906001,-0.1291131377,0.0844404921,-0.0947917476,-0.0002223049,-0.2281888723,-0.0361342542,-0.2055173963,-0.1976293474,-0.2189730108,0.1318597496,0.173056826,0.1635760814,-0.2263656259,0.0502876006,0.0127595123,0.7020566463,0.3510788977,0.1076621711,-0.2142923772,-0.3235682249,-0.5967857838,0.3562062681,-0.3159343898,0.1146410704,0.3648397624,0.4121897817,0.148203209,0.427143544,-0.1436979771,0.4120664895,-0.0402688608,-0.1203148663,-0.4084435701,-0.0316564813,0.0676868632,-0.0288704447,-0.1615073234,-0.4222982824,0.2162654996,-0.1609279513,0.0273401141,-0.1858121008,-0.3727512658,0.0986800417,0.0545374863,0.1093927696,-0.0871365517,0.3187762201,0.0858085155,0.0406495146,-0.2039973289,-0.2124969065,-0.2477271408,0.3365734518,0.0230686758,0.3301460147,-0.1964949071,0.1363684684,-0.1094119996,0.19717969,-0.338567704,0.1788088232,0.0149049293,-0.1495160908,-0.1281839609,0.2249827534,-0.0945451409,0.0898000896,0.0397565961,0.2163556665,-0.4016822278,-0.2031950355,0.4354741275,-0.5085446835,-0.198722437,0.2857094705,-0.0270965286,0.4125850201,-0.1164046377,-0.4313945472,-0.0754463971,0.2239834964,-0.234625414,-0.1203470752,0.0563717447,-0.0145737138,0.083092086,-0.1923094243,-0.0018164626,0.0800635144,-0.2020296305,-0.2691349685,-0.3540152907]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/369","title":"can't load local dataset: pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries","comments":"I am facing this issue in transformers library 3.0.2 while reading a csv using datasets.\r\nIs this fixed in latest version? \r\nI updated the latest version 4.0.1 but still getting this error. What could cause this error?","body":"Trying to load a local SQuAD-formatted dataset (from a JSON file, about 60MB):\r\n```\r\ndataset = nlp.load_dataset(path='json', data_files={nlp.Split.TRAIN: [\".\/path\/to\/file.json\"]})\r\n```\r\ncauses\r\n```\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 9, in <module>\r\n [\".\/path\/to\/file.json\"]})\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 483, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 719, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/datasets\/json\/88c1bc5c68489f7eda549ed05a5a738527c613b3e7a4ee3524d9d233353a949b\/json.py\", line 53, in _generate_tables\r\n file, read_options=self.config.pa_read_options, parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 191, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\n\r\nI haven't been able to find any reports of this specific pyarrow error here or elsewhere. ","comment_length":37,"text":"can't load local dataset: pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries\nTrying to load a local SQuAD-formatted dataset (from a JSON file, about 60MB):\r\n```\r\ndataset = nlp.load_dataset(path='json', data_files={nlp.Split.TRAIN: [\".\/path\/to\/file.json\"]})\r\n```\r\ncauses\r\n```\r\nTraceback (most recent call last):\r\n File \"dataloader.py\", line 9, in <module>\r\n [\".\/path\/to\/file.json\"]})\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 524, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 483, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 719, in _prepare_split\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/tqdm\/std.py\", line 1129, in __iter__\r\n for obj in iterable:\r\n File \"\/home\/XXX\/.conda\/envs\/torch\/lib\/python3.7\/site-packages\/nlp\/datasets\/json\/88c1bc5c68489f7eda549ed05a5a738527c613b3e7a4ee3524d9d233353a949b\/json.py\", line 53, in _generate_tables\r\n file, read_options=self.config.pa_read_options, parse_options=self.config.pa_parse_options,\r\n File \"pyarrow\/_json.pyx\", line 191, in pyarrow._json.read_json\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)\r\n```\r\n\r\nI haven't been able to find any reports of this specific pyarrow error here or elsewhere. \nI am facing this issue in transformers library 3.0.2 while reading a csv using datasets.\r\nIs this fixed in latest version? \r\nI updated the latest version 4.0.1 but still getting this error. What could cause this error?","embeddings":[-0.2248245478,0.2079820931,0.0130987065,0.1377785653,0.278103739,-0.2597190738,0.3169232011,0.4687684178,-0.1691406667,-0.149010852,0.0023774887,0.4203414321,0.0690719336,-0.017227646,0.1122415587,-0.1478196979,-0.2049249411,0.2665922344,-0.0051441933,-0.0658807755,0.0004076251,0.2411607355,-0.1398990303,0.2639468908,-0.1967165917,-0.0133888656,0.1871291995,0.489185065,-0.2591634989,-0.5631475449,0.3153218329,-0.296150893,0.0954629779,0.1106008142,-0.0001166556,0.3696361482,0.1890549362,-0.0164393801,-0.3757407367,-0.3551534712,0.2414884269,-0.490455389,0.4686816335,-0.1142355204,-0.0764247999,-0.3508143425,-0.2223610729,-0.0738732889,0.4803349078,0.1195221841,0.1514048278,0.4681633711,-0.1832385957,0.1065228358,0.1193087772,0.1260323972,-0.1624573022,0.4610108435,0.3759771287,-0.3450180292,-0.3874123096,-0.1435601264,-0.1003543884,0.3243615329,0.1929460168,0.0093747759,-0.0391238257,0.1259256303,0.1254607588,0.1766230315,0.3256267607,-0.3496564627,-0.0632260889,-0.0745448321,0.0622528456,-0.0295382254,0.1810942143,0.3203273416,-0.3467211425,0.1248235554,-0.306050688,0.1964052171,-0.1067041978,0.3012481332,-0.1913444847,0.3296273649,0.0628052801,0.297368139,0.1061380431,0.1481057107,0.1631402075,0.2127095461,-0.0343891941,0.2124329805,-0.2666904628,0.2280091494,0.1013871729,-0.1723196059,0.2379253805,0.1013728306,0.2416795492,0.042447485,0.2140950859,0.2787180543,0.3740185499,-0.0143485079,0.1454510838,0.076501362,-0.017595483,0.0274803136,-0.1041604877,0.1215523779,-0.1991192549,-0.3780224919,0.0324992426,-0.1924571395,0.2749885619,-0.187133342,-0.2933093011,0.0940950513,-0.4935642183,-0.0608110391,0.1227081344,0.276789993,0.2118470371,0.2285734266,-0.1052113771,0.2164437771,-0.0928959176,-0.1712169647,-0.2024862021,0.0592293888,-0.319909662,-0.1551751643,0.4050982296,-0.2470349222,0.2493386269,-0.0130813839,0.1818499863,0.009779647,0.1716861129,0.0250100493,0.2165830731,0.1760776043,0.0181108508,0.0982851386,0.171315372,-0.2304755002,-0.3545801938,0.3259263337,0.0253611729,-0.320799768,-0.315030843,0.0986372679,-0.1578409076,0.0982589722,0.1755226851,0.0540475734,0.1067288667,-0.106240876,0.0962656811,-0.1180267259,-0.0711897016,-0.2982014418,0.3827488422,0.2899155319,-0.953222096,-0.1176798344,0.1975107193,-0.0724640563,0.0969277993,0.5727161169,-0.5708695054,0.1833847016,-0.0560126379,0.2390382886,0.5799345374,-0.3536818922,-0.5187262297,0.4949888289,-0.3832627237,-0.1650160998,-0.1425090581,-0.0413008891,-0.0210614819,0.282671243,0.094546847,0.683675468,-0.1377247125,0.1846747398,-0.3309652209,-0.1655487716,0.2316760421,0.4529773295,0.0024090782,-0.355371207,0.0153144868,0.1048190445,0.0743056834,-0.0041734935,0.2683136165,0.2235309184,-0.1914619803,-0.2516146898,0.0803214014,-0.0126006631,-0.7056607008,0.0893205032,-0.0640129596,-0.0399806537,-0.2193442732,0.0276732985,-0.389359206,0.2925418317,-0.1446148455,-0.0062229387,0.0418271422,-0.0105411699,0.0670851246,-0.0903656334,-0.3402930796,-0.149033159,0.0774711519,-0.0015936849,0.0514434464,0.4117664993,-0.0129660284,-0.278139919,0.0033560798,0.3550055921,-0.0173281319,-0.1272228658,-0.1079458073,0.2835788131,-0.1293059438,0.1203492433,-0.0040456974,-0.3351640403,0.2203945816,-0.3095673323,0.1677345783,-0.0205300692,-0.0447345078,-0.1287044138,-0.220154658,0.403267175,-0.190984115,0.1956115216,-0.0854582712,-0.1480924189,0.2147214115,-0.1240264177,-0.0586371161,0.0097111287,0.1784366816,0.2656929195,0.5207040906,0.2706751823,-0.2427762002,0.0891448408,0.3627480268,-0.0371227339,-0.04015835,0.0438290164,-0.2223280221,-0.1289470792,0.1081068367,-0.106606029,0.1868596524,0.1889305115,0.0756322891,-0.1392150819,-0.2522187829,-0.2151123434,-0.0319145508,-0.0436208732,0.1300845146,0.4038740695,0.448564589,-0.138234511,-0.4100547135,-0.0738447234,0.23459059,0.2450695783,-0.2443244755,0.0373928584,-0.3147729039,-0.0499378555,-0.1750921458,-0.4401914775,-0.264534831,-0.199967891,-0.0551090539,0.2310739011,0.0131704817,0.1373821497,-0.0116336057,0.2456823736,0.100873366,-0.4298436642,-0.0119966045,-0.5101564527,-0.3695273399,0.0328288302,0.4673339128,0.438559413,0.1786762923,0.1814610958,-0.0701135993,0.0090797385,-0.0991811231,0.2115611881,0.0304584187,0.1755426973,0.0036325869,0.3344269693,-0.4671179652,-0.3113957047,0.6026245952,-0.0913662165,-0.2721755505,0.1993504167,-0.3250698447,-0.1124728322,-0.2069631219,-0.1498916447,-0.2904261947,-0.3914505243,0.3365614712,-0.0109002227,0.0356000885,-0.1289472133,0.1338442415,0.2838368416,0.0949847326,-0.2116737813,-0.0383752808,-0.4520311952,0.2738104463,-0.115242824,-0.2694824636,-0.071028322,0.0409649499,0.2816606462,0.062776126,-0.3481173217,0.0024619957,0.0731174126,0.5344305038,-0.1907640994,-0.1876478046,0.3679308295,-0.0579305328,0.0226893835,0.0716024786,-0.2526082098,0.1822912842,0.0439497717,0.0337406024,-0.175504297,0.4175980687,0.1020742804,0.9232445955,-0.2643375397,-0.1505668312,0.3155281246,-0.0246923175,0.0868570358,-0.1268964261,-0.3084769845,0.1649303287,0.0774956495,-0.0769030675,0.1023014933,0.0603703596,-0.3845144212,-0.0449143164,-0.0214512665,-0.2746353447,-0.099416025,0.3585598767,0.1034401506,-0.108390227,-0.010700753,-0.0802660361,-0.1179673746,-0.1918515861,-0.1134530008,0.0333905295,-0.0317989998,0.1877583414,-0.4592273831,-0.1170401797,-0.0875759348,-0.0683355033,0.3628225327,0.315877974,-0.0147822415,-0.1853874922,0.3408860564,-0.3085046113,0.5129215717,0.2342293561,0.192559883,0.2869602442,0.0705933198,-0.6148582101,-0.1541522592,0.1200243011,0.4119726419,0.4449229836,0.3248276114,-0.1001518816,-0.3170759082,0.12564978,0.1735467613,-0.0331000462,-0.2027922273,-0.3746901155,-0.3073224425,-0.4436307549,0.0817539543,0.0238223206,0.2499192953,0.4211018682,-0.0807263851,0.2893456221,-0.1060653105,-0.272272557,0.247982502,0.1461409628,-0.0633544698,0.0726812407,0.2435635477,0.3455856144,0.0976910666,0.6214662194,0.1869533211,-0.0988838747,0.1482156068,0.2310016453,0.3008661866,0.2950569391,0.1793146729,-0.0573054291,-0.1176065579,-0.0166513901,0.1620851904,0.1781002283,0.0401495062,-0.2675119936,0.1096058413,0.1865427941,0.2891855538,-0.2739403546,0.1674655527,0.5299274921,-0.0523521714,0.069254525,0.4074668288,0.1166573986,0.7771912813,-0.264470011,0.1479298323,0.5714970827,-0.2305463403,0.3164277375,-0.2754616439,0.0059591779,-0.5188856125,0.2529405057,0.0849900842,-0.239861384,-0.0096132262,0.3080280423,0.095092006,-0.1409421861,-0.1081847623,0.3453782499,-0.0026210255,0.220746085,0.0508506522,-0.2714333534,-0.2926777899,0.0353431068,-0.3116565943,-0.0429676063,-0.064911738,-0.1619161963,-0.096907042,-0.1001790538,-0.1564715803,0.4074653983,-0.0556710511,0.2410966903,0.3928505778,-0.4113485515,-0.0838277936,0.0472864322,-0.02236389,-0.1843993068,-0.1851106584,0.2453989238,0.0351632722,-0.0871742442,-0.0076094461,0.1267258078,0.2111590356,-0.1905830503,-0.3152864575,-0.0097105782,-0.110760957,-0.3773855567,0.0098416358,0.2698491514,0.1093739048,-0.0706365108,-0.2126791924,-0.1799738109,0.1643077284,0.0283878334,0.0598794259,-0.0270294957,0.1562407613,-0.2532791495,0.3292662203,-0.3607409,0.0462623872,0.4963206947,-0.101433821,0.164053604,0.5222313404,0.395812422,-0.1078451872,-0.2344770283,-0.0417825356,0.5676164031,-0.4813598096,-0.0808526054,-0.0654442012,-0.0923317298,0.1245621368,0.2647167146,0.3303147852,0.0091077928,-0.0937877223,-0.5132707953,-0.3523992002,0.2194441855,-0.1140954867,-0.0998067334,0.2764377892,0.5523682833,0.1198208034,0.3457675874,-0.2247915864,0.1964669228,-0.0767386258,0.1733497381,-0.29373312,-0.2949861586,0.0617094859,-0.0491167195,0.0436315909,0.0755816251,-0.1114885211,-0.2160479426,-0.0542553216,0.1099572033,0.1158914566,-0.1538788229,-0.1924220771,-0.1486531943,0.033608675,-0.0341766663,-0.0114123672,-0.0316841267,-0.0636330768,0.0092392042,0.1375732422,-0.0495583229,-0.1432165056,0.0238068402,-0.0251240805,0.0513093583,0.0436231419,0.2476351559,0.0279661119,-0.0172809381,-0.2315071225,0.0654815733,0.2410795689,0.1861478686,0.2242752016,-0.2875194252,-0.1363506019,-0.1797350198,0.1812481582,0.3907132447,-0.1477690637,-0.0057631144,0.0170549676,0.1338837296,-0.1501574665,-0.1573177129,0.262271136,0.0530038476,-0.0168350432,-0.0100825345,0.1297941357,-0.2084887028,-0.08111763,0.1575099975,0.3992906511,0.0401901715,-0.4732250869,0.3453002572,-0.2074123472,-0.1616266519,0.1562598497,0.0847399905,0.3768481314,-0.0452592634,-0.1026147753,0.1214384288,-0.1278783083,0.2560073733,-0.0429029129,-0.4255738556,-0.1164732426,0.2421922982,0.1526723951,-0.3106538951,0.2070242316,0.7946990132,0.2270723581,-0.3669965863,0.0756444409,0.1639969647,-0.1651429385,0.0323552154,-0.7381464243,-0.1738948971,-0.1725956202,0.180132702,-0.1772890836,-0.0201254096,0.3450933695,0.1026235744,-0.0690098554,-0.4073937833,0.0018845239,0.1002660021,-0.3239278793,-0.237759307,0.4012735784,0.1469526887,-0.1587483436,-0.0346477181,-0.128130272,0.3428938985,0.2950404584,-0.3094574511,-0.1638810784,-0.0877760798,-0.1351440102,-0.2076944262,0.1110962331,0.1466931105,0.1099853516,0.3763891459,0.1300737709,-0.1291117966,0.0981806517,0.1858906001,-0.1291131377,0.0844404921,-0.0947917476,-0.0002223049,-0.2281888723,-0.0361342542,-0.2055173963,-0.1976293474,-0.2189730108,0.1318597496,0.173056826,0.1635760814,-0.2263656259,0.0502876006,0.0127595123,0.7020566463,0.3510788977,0.1076621711,-0.2142923772,-0.3235682249,-0.5967857838,0.3562062681,-0.3159343898,0.1146410704,0.3648397624,0.4121897817,0.148203209,0.427143544,-0.1436979771,0.4120664895,-0.0402688608,-0.1203148663,-0.4084435701,-0.0316564813,0.0676868632,-0.0288704447,-0.1615073234,-0.4222982824,0.2162654996,-0.1609279513,0.0273401141,-0.1858121008,-0.3727512658,0.0986800417,0.0545374863,0.1093927696,-0.0871365517,0.3187762201,0.0858085155,0.0406495146,-0.2039973289,-0.2124969065,-0.2477271408,0.3365734518,0.0230686758,0.3301460147,-0.1964949071,0.1363684684,-0.1094119996,0.19717969,-0.338567704,0.1788088232,0.0149049293,-0.1495160908,-0.1281839609,0.2249827534,-0.0945451409,0.0898000896,0.0397565961,0.2163556665,-0.4016822278,-0.2031950355,0.4354741275,-0.5085446835,-0.198722437,0.2857094705,-0.0270965286,0.4125850201,-0.1164046377,-0.4313945472,-0.0754463971,0.2239834964,-0.234625414,-0.1203470752,0.0563717447,-0.0145737138,0.083092086,-0.1923094243,-0.0018164626,0.0800635144,-0.2020296305,-0.2691349685,-0.3540152907]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/368","title":"load_metric can't acquire lock anymore","comments":"I found that, in the same process (or the same interactive session), if I do\r\n\r\nimport nlp\r\n\r\nm1 = nlp.load_metric('glue', 'mrpc')\r\nm2 = nlp.load_metric('glue', 'sst2')\r\n\r\nI will get the same error `ValueError: Cannot acquire lock, caching file might be used by another process, you should setup a unique 'experiment_id'`.","body":"I can't load metric (glue) anymore after an error in a previous run. I even removed the whole cache folder `\/home\/XXX\/.cache\/huggingface\/`, and the issue persisted. What are the steps to fix this?\r\n\r\n Traceback (most recent call last):\r\n File \"\/home\/XXX\/miniconda3\/envs\/ML-DL-py-3.7\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 101, in __init__\r\n self.filelock.acquire(timeout=1)\r\n File \"\/home\/XXX\/miniconda3\/envs\/ML-DL-py-3.7\/lib\/python3.7\/site-packages\/filelock.py\", line 278, in acquire\r\n raise Timeout(self._lock_file)\r\n filelock.Timeout: The file lock '\/home\/XXX\/.cache\/huggingface\/metrics\/glue\/1.0.0\/1-glue-0.arrow.lock' could not be acquired.\r\n\r\n During handling of the above exception, another exception occurred:\r\n\r\n Traceback (most recent call last):\r\n File \"examples_huggingface_nlp.py\", line 268, in <module>\r\n main()\r\n File \"examples_huggingface_nlp.py\", line 242, in main\r\n dataset, metric = get_dataset_metric(glue_task)\r\n File \"examples_huggingface_nlp.py\", line 77, in get_dataset_metric\r\n metric = nlp.load_metric('glue', glue_config, experiment_id=1)\r\n File \"\/home\/XXX\/miniconda3\/envs\/ML-DL-py-3.7\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 440, in load_metric\r\n **metric_init_kwargs,\r\n File \"\/home\/XXX\/miniconda3\/envs\/ML-DL-py-3.7\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 104, in __init__\r\n \"Cannot acquire lock, caching file might be used by another process, \"\r\n ValueError: Cannot acquire lock, caching file might be used by another process, you should setup a unique 'experiment_id' for this run.\r\n I0709 15:54:41.008838 139854118430464 filelock.py:318] Lock 139852058030936 released on \/home\/XXX\/.cache\/huggingface\/metrics\/glue\/1.0.0\/1-glue-0.arrow.lock\r\n","comment_length":49,"text":"load_metric can't acquire lock anymore\nI can't load metric (glue) anymore after an error in a previous run. I even removed the whole cache folder `\/home\/XXX\/.cache\/huggingface\/`, and the issue persisted. What are the steps to fix this?\r\n\r\n Traceback (most recent call last):\r\n File \"\/home\/XXX\/miniconda3\/envs\/ML-DL-py-3.7\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 101, in __init__\r\n self.filelock.acquire(timeout=1)\r\n File \"\/home\/XXX\/miniconda3\/envs\/ML-DL-py-3.7\/lib\/python3.7\/site-packages\/filelock.py\", line 278, in acquire\r\n raise Timeout(self._lock_file)\r\n filelock.Timeout: The file lock '\/home\/XXX\/.cache\/huggingface\/metrics\/glue\/1.0.0\/1-glue-0.arrow.lock' could not be acquired.\r\n\r\n During handling of the above exception, another exception occurred:\r\n\r\n Traceback (most recent call last):\r\n File \"examples_huggingface_nlp.py\", line 268, in <module>\r\n main()\r\n File \"examples_huggingface_nlp.py\", line 242, in main\r\n dataset, metric = get_dataset_metric(glue_task)\r\n File \"examples_huggingface_nlp.py\", line 77, in get_dataset_metric\r\n metric = nlp.load_metric('glue', glue_config, experiment_id=1)\r\n File \"\/home\/XXX\/miniconda3\/envs\/ML-DL-py-3.7\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 440, in load_metric\r\n **metric_init_kwargs,\r\n File \"\/home\/XXX\/miniconda3\/envs\/ML-DL-py-3.7\/lib\/python3.7\/site-packages\/nlp\/metric.py\", line 104, in __init__\r\n \"Cannot acquire lock, caching file might be used by another process, \"\r\n ValueError: Cannot acquire lock, caching file might be used by another process, you should setup a unique 'experiment_id' for this run.\r\n I0709 15:54:41.008838 139854118430464 filelock.py:318] Lock 139852058030936 released on \/home\/XXX\/.cache\/huggingface\/metrics\/glue\/1.0.0\/1-glue-0.arrow.lock\r\n\nI found that, in the same process (or the same interactive session), if I do\r\n\r\nimport nlp\r\n\r\nm1 = nlp.load_metric('glue', 'mrpc')\r\nm2 = nlp.load_metric('glue', 'sst2')\r\n\r\nI will get the same error `ValueError: Cannot acquire lock, caching file might be used by another process, you should setup a unique 'experiment_id'`.","embeddings":[0.0387074426,-0.1158736274,0.0723508522,0.31246683,0.3204596639,0.0294491537,0.0900948569,0.0702063814,0.5023596287,-0.0785918608,-0.1850662529,0.030607095,-0.0480661876,-0.2614514828,-0.031660337,-0.1943873614,-0.2007998526,-0.0895176679,-0.0916291848,0.1228041351,-0.2072485536,0.3985546231,-0.1751362681,-0.1407353282,-0.3555804789,-0.1709385812,-0.0894152075,0.3625312448,-0.0941761583,-0.4639343917,0.1580357552,0.0359998159,0.0657720491,0.0958728641,-0.0001253988,0.0097096125,0.4328870475,0.0207843147,0.0158178359,-0.2742264271,-0.0472081974,-0.538641572,0.2313128412,-0.0381141342,0.0553530268,0.1542789787,0.0128486948,-0.3735764623,0.3929035664,0.2302975506,0.117739819,0.4218791127,0.24505198,-0.3009939492,-0.2283041924,-0.5178474188,0.1226883009,0.9323651195,0.1983928382,-0.4340361357,-0.1681259274,0.1428887099,-0.0306593888,0.1357378513,0.4364429712,-0.1017124504,0.3628014028,-0.1146173105,-0.0268294904,0.0082821026,0.033585906,-0.1131085306,-0.266302824,-0.1967049837,0.3940870166,-0.3095428348,0.3262719214,-0.2071050406,0.1107339635,0.0614548922,0.022858521,-0.3017476499,0.0311336033,0.1103480235,-0.0704878271,0.1314450353,-0.0304556023,0.1059605926,0.4824415147,0.1158646271,-0.1343362033,0.4646417797,-0.2977823317,0.252591908,-0.223045066,0.3128059208,0.0215294249,0.4838051498,0.0795658678,0.047243502,0.0058021699,0.0168824252,0.3405639827,0.1478574723,-0.0365649909,0.5148493648,0.406049639,-0.0803215504,0.1938289553,0.2729445696,0.007888251,-0.0202901494,-0.0450518727,-0.3363509476,0.1489802897,0.1632087529,0.0452220961,-0.088920109,0.0154884411,0.0259291045,0.0870660245,-0.1591012478,0.1520417929,0.1608738452,-0.0276133586,0.2953260541,0.1074499413,0.2743976414,-0.0106981592,-0.1547306031,-0.1928951293,0.3185302019,-0.3265420198,0.2753768265,0.2608951926,-0.2996115088,0.3550312221,0.0881137773,0.088500753,-0.4025236368,-0.0697804168,0.0475794412,-0.2875433266,0.2150298953,-0.1425008029,0.0949410796,0.1955335587,-0.2912884653,-0.0110831,-0.3652180433,-0.4574064314,-0.3263042271,-0.0511174686,0.0123522189,-0.4049867094,0.059775997,0.0596663393,0.1430255324,0.0267869532,0.0218423177,-0.0301715992,-0.1087390333,-0.1520867944,-0.2062555403,0.2813636065,0.5562860966,0.0976210237,-0.3056697249,-0.1878911257,-0.1079912558,-0.1382289678,0.234931007,0.0458945334,0.0970936641,-0.2242059708,0.0844082087,0.2618544698,-0.7882769108,-0.6193390489,-0.0095672943,-0.1618588418,-0.0446815379,0.0548863038,0.0888190717,0.1180931777,0.0993980765,0.3330300748,0.0924305469,0.0943365321,-0.4169382155,-0.2193696946,0.0052568289,0.0715348944,0.0779734626,0.2852445543,0.2650149465,0.1815573722,0.3520883322,0.0231217835,0.3938896656,-0.2276394665,0.4092956483,-0.0556347631,-0.2220461667,0.0867227688,-0.1873000264,-0.3539688587,0.4585732222,-0.5009264946,-0.0561315306,0.0323615,0.0708050653,-0.4591314793,0.0309203062,0.1142947748,-0.0322566815,0.0307951979,0.0454827324,0.0259703789,0.0601222664,-0.1053756922,0.5104598403,-0.0319066048,0.2045419663,-0.2858424187,-0.0922260359,0.0968826637,-0.0596505105,-0.0074587725,0.3105420172,0.494872421,-0.0591302067,0.046972122,0.5281563401,0.0510515682,0.260019958,-0.0119120143,-0.014790481,0.2664884925,0.0134368101,-0.0115792537,-0.5587484837,-0.0614249967,-0.061917197,0.3316039443,0.0798467323,0.1930182427,0.1027684882,-0.1634512842,0.1938889027,0.0462909266,-0.2189996392,-0.1978612393,-0.2039311379,0.5161855221,-0.1291542202,0.0738871396,0.0157022104,0.2341696322,-0.0599894449,0.2643677294,0.0904286429,0.1763944626,-0.1143529415,-0.1747926176,-0.0878666565,-0.2505739927,0.3127426505,0.3478238881,-0.0408706702,0.0977798402,0.1051948369,0.0304459818,-0.1265745908,0.0144020161,0.0762419403,-0.0217216946,0.3686549664,0.1944015324,0.0989873633,-0.0384875908,0.1585069001,0.1161005422,0.2183551937,-0.2039139122,0.0098381992,-0.3964014351,0.0102863135,0.2288608551,-0.0039137155,-0.4821487069,-0.4297889173,-0.0750375018,0.1887993068,0.3223307729,0.0609571673,0.1792288721,0.229039982,-0.1921593994,0.2414934486,-0.1553550065,-0.4484227598,-0.0379551537,-0.16159527,0.1437360942,-0.095669426,0.114574559,-0.1248923391,-0.1733731776,-0.1402734071,-0.2574286163,0.0239719469,0.0843010917,0.7145802975,0.2129424065,0.0088944845,-0.001066129,0.1230750382,0.3546610475,-0.1275537908,0.0524185337,0.2911889255,-0.3696325719,0.1463911831,-0.2685171366,0.1616153717,-0.5034708977,-0.4056679904,0.2910679579,-0.1105204225,-0.2216467559,0.1172694862,-0.2792262137,0.3122066855,-0.2726821899,0.1251493245,-0.2774215341,-0.1923308969,0.2103399932,0.0043342048,-0.0900195614,-0.2496180683,0.1020823419,0.1823785007,-0.1399124414,-0.392642647,-0.7514959574,-0.1141082719,0.0987052619,-0.2474500388,-0.1125537828,0.1439746022,-0.3512949944,0.0414206386,-0.1349384189,-0.0026854076,0.1042374671,0.0757852346,0.2012827694,0.0657725707,0.2217899412,0.1661201119,0.5310268402,-0.0442974865,-0.1021614,0.3465194404,0.0595608316,0.4767459035,0.1927142441,-0.3890644908,-0.1930055022,0.1417024732,0.2469887286,-0.2544842958,0.0348412432,0.242321372,-0.1883638799,-0.3141641021,-0.2212717533,-0.093635276,-0.1036054343,-0.1123079658,0.1638350636,0.1402924359,0.2603793442,-0.148579672,-0.0792707875,0.2771490216,0.3417000473,0.0367456079,0.1605739743,0.0545909628,-0.1954059005,-0.4282639921,0.3973340392,-0.0049797916,0.5216323733,-0.2394588143,0.0594329983,0.104599461,-0.1041496545,0.5359894037,0.0970067754,0.1271158159,0.1798020005,0.0382610857,-0.3157488704,-0.0304222666,0.0299389604,0.1489219517,0.4948722124,0.7226865888,-0.2659997344,-0.1192564592,-0.260363698,0.1650894135,-0.3048899174,-0.2024811804,-0.2063498646,-0.3236882091,-0.1710140556,0.0123416819,-0.1153358594,0.3378127813,0.4723193049,0.2161138952,-0.1360187531,-0.4786417186,-0.0407852493,-0.072795108,0.4126513302,-0.0869294778,0.0880513936,0.4231230021,-0.1487879604,-0.2982723117,0.1877476126,-0.1136941314,0.1213305816,0.2333737314,0.1737683415,0.1053901166,0.4253507257,-0.1597548127,0.2259803712,0.1264660209,0.0955732763,-0.0979735851,-0.0315058976,0.1203087941,-0.0238322522,0.0295598432,0.2193056941,0.3600981534,0.2023499608,-0.1878729463,0.418087244,-0.0431181192,-0.0350698307,-0.0972783417,-0.1000874862,0.8022153974,-0.1595791727,-0.2619310319,0.0298735965,-0.1457985342,0.6428802609,-0.1042103767,0.1761111617,-0.0742923245,-0.3932453096,0.0441793054,-0.2175304592,0.0657586604,-0.4121157229,0.0514316969,0.2708448172,-0.4369366169,-0.0335604399,-0.0650877059,0.1836887747,-0.6214601994,0.0531327799,0.0301300343,0.0401384309,-0.3123498559,0.3805063665,-0.2036480308,-0.0550628379,0.2770389915,-0.0553837903,-0.3994785547,-0.0654048175,-0.2844845653,0.0708888918,0.2390864789,-0.0704395249,-0.1759535223,-0.0610572174,0.3143347502,0.139804408,-0.2994338572,0.3438142538,-0.0917133242,-0.295432359,-0.0030769217,0.2781415284,0.1166203693,0.0029364107,-0.0925313905,0.0463676602,-0.1733463109,-0.110754244,-0.114487052,0.1817105263,0.0533906259,0.3183444142,-0.1433248371,-0.0191320311,0.2342263609,-0.1494246572,0.0253103655,0.2552128136,-0.0040863091,-0.1048381031,-0.2064364851,-0.2550046742,0.1637461782,0.5607220531,-0.0581932627,-0.507388711,0.2165059447,0.0164635982,-0.0639621839,-0.1016346216,-0.3467877805,-0.1690581441,-0.4205929339,0.2429348081,0.1547178775,-0.0193132311,0.0040054456,-0.1858248413,0.4843347371,-0.2830844223,-0.0019895388,-0.7558655739,-0.2973373234,0.2834347785,-0.2853268385,-0.0136601031,-0.1148889735,0.0471758172,-0.2104178071,0.4024417996,-0.1725854874,0.1829887778,-0.2871595025,-0.2451511472,0.5845168233,-0.0654659942,0.1676660776,-0.0475941338,0.0198076591,0.4047594666,-0.1240420714,-0.0497582592,-0.1973866969,0.1797023565,-0.085015215,0.1407679319,0.0183720086,-0.2761431038,0.2459150851,0.0895486623,-0.0811777785,0.2046791613,0.0613295808,0.0384051986,0.0426196158,0.5235905051,0.2501638532,0.2672936916,0.0090645002,-0.0032560914,0.006251011,0.0792045072,-0.0596478842,0.1828986108,-0.4656667709,-0.1377230138,-0.0974091142,-0.1361397952,0.1559678614,0.1431010067,-0.1016526967,0.0190405119,0.1252435595,-0.1216821522,0.066697374,0.0342366025,-0.0300025847,0.0869543105,-0.2467589527,-0.0668205544,-0.0870528519,0.3180530071,-0.1310955435,0.0735199004,0.3424254358,-0.0961303413,0.1432819515,0.111812368,0.4970526397,0.0075979293,0.0205601845,-0.2067308426,0.1344556659,-0.2691765726,0.4959367216,-0.0428581424,0.1166442484,-0.0367383026,-0.3253391087,0.1450218856,-0.3323923051,0.2503056824,-0.0408238173,-0.3007710278,-0.1008448005,0.2648639977,0.2209362835,0.1186063141,-0.4627085924,0.691084981,-0.2642379105,-0.3084988892,-0.0429287441,0.3635565341,-0.0644735843,-0.095384635,-0.006921432,-0.2487750351,-0.1431326568,-0.0225575119,0.184918642,0.085372515,0.4013891816,0.1992626935,-0.1833755076,-0.3830614388,-0.484972626,0.3002397716,0.1777810156,-0.1937877685,0.358045578,0.1135041416,-0.0232857354,0.1853680611,0.2641861141,0.3333707452,0.1051836833,-0.1506636143,0.0953453407,-0.0113157034,0.010702461,-0.1575168371,0.5133901238,0.3697788119,-0.3136532605,-0.0351257771,0.0404248461,-0.1790904254,0.2119154036,-0.0549708791,0.2851828039,-0.0977840349,0.5591534972,-0.0657936409,0.0186750442,-0.1547776759,0.1321267486,-0.336193949,-0.0434215739,-0.0058514457,0.1318584532,0.0049857446,-0.3068116009,0.041179616,0.0552182905,0.2113997936,0.5045401454,0.1183800399,-0.4218707383,-0.1334498525,-0.7453029752,0.113484107,-0.0222577676,0.0924694166,0.2634624541,-0.1080278382,-0.1685672402,-0.0461022668,0.2452772409,0.3517722785,0.1647494137,-0.454644829,-0.4675461054,0.1131402329,0.0678355992,-0.2684638202,0.0692890435,-0.0464069955,-0.0191464238,-0.3807815611,-0.0374480709,-0.0435910933,-0.3628323972,0.1442728341,0.5466169119,0.2543390393,0.1203696132,0.2188725919,0.0913777575,-0.0422803834,-0.3698080182,-0.4085541964,0.1035869345,-0.335095644,-0.0380124822,0.0683520958,-0.368853271,-0.1574761719,-0.2832649052,0.5227215886,-0.2089237571,-0.0723497272,-0.4062183499,0.0933076218,-0.1109778658,0.0674845576,0.2345815003,0.4868639708,0.2205927223,0.3542439342,-0.2795716524,-0.0774608701,0.5171990395,-0.4738144577,-0.1686322093,0.2175594121,0.3969512284,0.173976928,-0.0374174304,-0.4593833685,-0.118847318,0.2529200912,0.0813983828,-0.1838924289,0.2594431341,-0.2825557292,-0.0089298617,0.0027261649,-0.353985697,0.3217030168,-0.2165519893,-0.0081745256,0.1252815425]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/365","title":"How to augment data ?","comments":"Using batched map is probably the easiest way at the moment.\r\nWhat kind of augmentation would you like to do ?","body":"Is there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```","comment_length":21,"text":"How to augment data ?\nIs there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```\nUsing batched map is probably the easiest way at the moment.\r\nWhat kind of augmentation would you like to do ?","embeddings":[-0.0076108566,-0.0786108673,-0.3029457033,-0.0933210775,0.0747116208,0.2788416147,-0.104579322,0.2814402282,0.1223137155,0.1161646768,-0.122473754,0.0561201237,-0.0021822758,0.1659331471,0.1615757942,-0.1614081562,0.1548456997,0.1445470452,-0.1470573843,0.1611988991,-0.1049567237,-0.4687489569,0.0024022388,-0.5504875779,-0.113978371,-0.1055827886,-0.2983653247,-0.1648691446,0.0454924963,-0.377905339,-0.2231483459,0.0820774585,0.1151698157,0.2421452105,-0.0000998411,-0.208939597,0.1646328717,0.0892125666,-0.3107390702,0.4721699059,0.1105210781,-0.8047289252,-0.0186209511,-0.0786827132,-0.0834561139,-0.4446455538,0.0461451188,-0.0636526868,0.3388094604,-0.1767680943,0.2676832378,-0.0098133795,0.374391377,0.0117657464,0.1348076314,0.3799756467,0.1615886688,0.0862629563,0.7172884345,0.143665731,0.336512953,0.1444842219,-0.1315525174,-0.0973731354,0.2357285768,0.0122268666,0.1705203503,-0.3990068138,-0.015450797,0.1021066234,0.3820779622,0.2447232902,-0.5512055755,-0.3171167672,0.2154936045,-0.0941848382,-0.3701936007,-0.1095228866,0.154990375,-0.0199620537,-0.2288735956,-0.2409577668,-0.0837807953,0.3638654649,0.0529173575,0.4135732949,0.0324129425,-0.0221883953,-0.1271591485,-0.279292196,-0.1728684753,0.1948121637,-0.15500696,0.4723395407,-0.1161077693,-0.1414955705,-0.2426722497,-0.2189955115,0.0774522275,0.10739398,0.1769826859,0.178008154,-0.1855886132,0.1188087761,0.1672571748,-0.300796181,-0.0579392277,0.1513618678,0.0100308871,-0.1358963102,0.2673924267,-0.0651103631,-0.104189299,-0.1596966088,0.0384277478,-0.1040328294,-0.1227104515,0.0058560395,-0.0125691555,0.0119503718,-0.2453552037,0.1206428185,-0.0216629002,0.0626219884,-0.2390582412,0.0745224953,-0.1906955987,0.2592640519,0.2707093954,-0.2050131112,-0.1140430719,0.1182821542,-0.0716915056,-0.1564167589,0.2377939522,0.117393069,0.2470694333,0.0270550232,0.0919327438,-0.036212869,0.0832180902,0.0104128635,0.3068697453,0.2046578526,0.1408862025,0.0554040857,-0.185100466,0.1971400678,-0.0127218086,0.0888806656,-0.3098223209,-0.0053880825,-0.3795255423,0.2868727744,0.0608546808,-0.1946115643,-0.1409720331,0.3655321002,0.1183648705,-0.2584454715,0.0642438978,0.0939885601,0.1048165783,-0.3767409027,0.4527572691,0.0382832251,-0.3822378814,0.0499950238,-0.1435533464,0.3040354252,0.1318435967,0.0393009409,-0.0303404424,0.1447838247,-0.1284517348,-0.1212250367,0.2840679884,-0.1609270573,-0.333278656,-0.0140593601,0.1716776937,0.1381157041,-0.1504609734,0.32857728,0.2457571626,-0.0475737303,0.1889441162,0.3833818734,-0.1481109262,-0.1654292792,-0.0901371315,-0.1093993634,0.3156548142,0.1865181774,-0.061891932,-0.1337626874,0.2854762673,-0.4684000611,0.1513703465,-0.1575251222,0.318633914,0.1901426464,0.1881601959,0.3953309953,-0.0614690818,-0.1812356263,-0.1726555377,-0.0147862732,-0.1361605525,-0.148172766,0.1438458115,-0.1893931925,-0.0441929176,-0.1230905354,0.1288403124,-0.3399832845,0.2959398031,-0.1768559813,0.0442280509,-0.4881070256,-0.003710927,-0.1384122372,0.2111998796,0.079161264,-0.200127244,0.0091773821,0.269851625,-0.3361878097,-0.2222453654,0.335079819,-0.1025168821,-0.1454635412,0.1096941531,0.2788130343,0.0695466027,0.210286662,0.2375261784,0.2521240711,0.0980793834,0.1130802631,0.2049984187,0.2804630697,0.0073606591,-0.1436427087,-0.0969061852,0.2299025208,0.0896204039,0.1053560823,-0.2545942366,0.2065225691,0.265427798,0.0409889631,-0.0918569341,-0.0353268236,-0.2036500126,-0.0124819223,0.2784784436,-0.2759414911,-0.1348335743,0.0308585111,-0.4751145542,-0.1810441017,-0.1317112744,-0.175999701,-0.4588342309,-0.1853166819,0.1180563569,-0.0558093674,-0.179898262,0.4024444222,0.1857137829,-0.0418794826,0.255289793,0.1530427337,0.0145458812,0.1135466024,-0.0431540236,-0.0487925969,0.3766633868,0.2131991088,0.1516609043,0.064551115,0.0053588343,0.1715642661,-0.0934923217,-0.2121171206,-0.0636444911,-0.189534232,0.0392831191,-0.6104679108,-0.2298272103,-0.3691900074,-0.0815134421,0.2527185977,-0.2411987484,0.0109083708,0.1769155115,0.1015365049,0.0514890067,-0.2087112665,0.0035752738,0.2888846099,0.002571587,0.2455696315,-0.0415379591,0.0583165772,0.5223010182,0.158653304,0.2197445631,-0.1253729761,-0.5068557262,-0.2053861767,0.0732154995,0.212038964,0.1939320415,-0.0102066258,-0.0845808759,-0.1673970968,-0.0973307937,-0.1113427505,0.1161694825,-0.1131237,-0.1461567432,-0.0387624428,-0.2443744093,0.1633575261,-0.4032546282,-0.1918665469,0.1039541438,0.2002733201,0.0443589427,-0.226942718,0.2483760267,-0.0357570499,0.1118209958,-0.0389118828,-0.1680570245,-0.2406021357,0.5701763034,-0.2656957507,0.0013287901,-0.0552209988,-0.3044517338,0.1406930983,0.1969478428,-0.040401794,-0.20166637,-0.0954496562,0.0220215674,-0.2020997852,0.1835167259,0.1146987304,-0.1391179115,-0.1791120172,0.058222495,-0.0121351136,-0.0202245116,0.3734327853,0.1653358638,-0.1720096022,-0.1517802924,-0.1649215519,0.0782194883,0.0789995417,-0.0428789146,0.1349996328,-0.1324146539,0.0905666873,0.0054748622,-0.0927537233,-0.0154377744,-0.1174418554,-0.2254609615,-0.0095396359,0.2916847765,-0.1485463679,0.0818371773,-0.1999635398,-0.3598059714,-0.0773131773,0.195136562,-0.4344590008,0.2437220514,-0.2577315867,-0.1267358065,-0.4629677236,-0.2162117362,-0.0094657168,-0.1793209612,0.2181589007,-0.0718271583,-0.3334373236,-0.0339215435,-0.5081927776,0.6014177203,0.2712253034,0.0943000838,0.4166786671,-0.0557099655,0.0767642781,-0.0162957609,0.3793668747,-0.2902058959,-0.3733693957,-0.1220445037,0.0605355762,-0.2919299901,-0.0664405376,-0.3477963507,0.1122818738,0.3791012168,0.4248925447,0.0626002923,0.0361148715,0.3265299201,0.5222181082,-0.1211359948,0.0620600507,0.1321882606,-0.11373768,-0.0604071058,0.0975675881,0.0429013744,-0.0264291614,0.4464077353,-0.1634343266,-0.1746935993,0.1322166473,0.3371099234,0.1987120956,0.2410274446,0.33975932,-0.0092987204,0.0632389486,0.0371348374,0.3818003535,0.2620842457,-0.121124953,-0.309084028,-0.268571496,-0.4397201836,0.7446144819,-0.2391800582,0.3427326381,0.2730872929,-0.2558057308,-0.034062136,-0.3682528734,0.1775990129,0.1166257188,0.1629988551,-0.3898420632,-0.6086577773,0.0950630829,-0.0550279915,-0.0754461363,0.2048726529,-0.2119989246,-0.0661271587,0.331089437,0.4750807881,0.6668503284,0.3043497205,-0.0304865465,-0.0017351764,-0.1681497097,0.2637513578,-0.1249459311,0.1636634171,-0.0347782634,-0.1753425151,0.0363152251,0.0919379964,0.0353778861,0.2566660345,-0.0770539269,0.2041571289,-0.1932994872,-0.0081464862,-0.0521626808,0.0789847299,0.2488762885,-0.224346295,-0.0252265278,0.2686210871,-0.0155818276,-0.306399703,-0.1489024758,-0.0248524882,-0.1278847754,0.1565118283,-0.1564068198,-0.0651301369,-0.1435315311,-0.0372006074,0.2617553174,-0.2358973175,-0.0895469934,0.0098614823,0.0609896332,0.4246042669,0.0444212705,0.1089750305,0.160765633,0.2134145647,-0.0879860818,0.0537441745,0.2833583057,0.0617836304,-0.219744876,0.2065999955,0.0877044275,-0.4093133211,-0.303245455,0.0217586588,0.00976492,-0.3889300525,-0.0714752376,0.4038933516,0.0414333194,-0.1766953617,0.2063099146,0.2602943182,-0.4080252647,0.3073631227,0.1646901965,-0.1629061997,-0.0814710706,0.0624558553,0.0894903839,-0.3430860341,0.4695490301,-0.0964902714,0.0148587991,-0.2405357957,-0.090734005,0.0660513937,0.144494459,-0.1780950725,0.0590978488,0.1074360609,0.1204388142,-0.0869834274,0.2047009617,0.0359196924,-0.1315042973,-0.3764164746,-0.565154314,0.2206884623,0.4015713036,0.2265172899,-0.0090960497,0.1098169088,-0.2741596997,0.5316253304,-0.4084593952,0.1374558508,0.3545907438,0.229019776,-0.0154557619,-0.0441209599,-0.0938913226,-0.1243731454,0.2687463164,-0.1383428276,0.0089734746,-0.2991035283,-0.1092933044,0.0146182673,-0.0773288906,-0.0449198708,-0.0239887256,0.0309058614,0.0723413378,-0.4569346607,0.0233955625,-0.0968520716,-0.0409793854,0.2198071629,0.3289917707,0.2789762318,-0.1022356302,0.1524555385,-0.107658498,-0.03753791,0.1082071662,0.024584556,0.229359746,-0.0667486489,0.2334054708,0.1745454967,0.1658675373,-0.139394477,0.0500465222,0.1432812214,0.0567190945,-0.0536567457,0.216628477,0.2777225077,0.1126280352,0.2396587133,0.3434223533,0.4050264657,-0.2293674499,0.058635205,0.270295471,0.1469022483,0.1718360633,0.184688881,0.2326921225,0.0667620748,-0.0691915229,-0.0993691161,0.0514886826,-0.0264223516,0.002059099,0.41939044,-0.1256745309,-0.3333927989,-0.0997269452,-0.0049824635,-0.020089332,0.0695248246,0.064614661,0.4033404887,0.4989077747,-0.2140619606,-0.0579282939,-0.2152508199,0.1457214057,0.3286412954,0.1389950067,0.4616528153,0.4834704399,0.3421583772,0.0305833407,-0.4088159204,0.0132857896,0.0493590236,0.0966358706,0.0203998536,0.0104433857,-0.1942884624,-0.2485844195,-0.1224425808,0.2361288965,-0.3936660886,0.1462329328,-0.188844502,-0.1361034662,-0.1875850707,0.1707567424,0.1503903568,0.2097183168,-0.2779640853,-0.0743929148,0.0780007839,-0.1526313722,0.2677415311,0.1799112409,0.2255945206,-0.0255021546,-0.1455079466,0.372020334,-0.2667145729,-0.0936895609,-0.1716820151,-0.0359138511,0.0097585088,-0.0681817532,0.0413307436,0.2832562029,-0.2193230391,0.2327999473,-0.0459244661,0.1422993839,-0.3630723059,0.026436042,0.2322731018,-0.5020996928,-0.0782477856,-0.321580112,-0.1785210222,-0.4239553511,0.1658907533,-0.3847845793,-0.1192836538,0.1992290318,0.1566546857,-0.1398010105,0.530213654,0.6634432673,0.1763309389,0.0529038012,-0.3323199749,-0.3900742233,0.17309165,-0.0707127377,-0.1365450174,-0.064116776,-0.1198808849,0.2669940293,-0.0046594054,0.3345374167,-0.0598358326,-0.2998943031,0.1108061448,-0.1535372287,0.2455341071,-0.0954857543,0.3103294373,0.1946123093,0.0610554405,-0.0867738128,-0.069465369,0.1932781935,-0.0968898907,0.0020185073,0.0599211715,0.0911883041,0.4429362714,0.2636013627,0.0797936395,-0.4356223941,-0.0531482957,-0.351200074,0.0163178425,-0.3566222191,0.330992341,-0.0566236116,0.2636432946,0.0097009959,-0.1251349449,0.2435498834,0.2723661363,-0.0551581271,-0.0806190744,-0.2442428321,0.1907430738,-0.1638905853,-0.2496409714,-0.0956591144,0.1525158882,0.0598496459,-0.0583802387,0.0142232636,0.0174985882,0.0475417711,-0.1556226164,-0.1403152943,-0.5458517075,-0.0969870016,0.0516062267,0.1564487964,-0.3696105778,-0.2171123177,0.3101464212,0.0661328882,-0.0796067864,0.442563653,-0.1411465555,-0.3818380535,-0.1352281719,0.2422277629,0.064267613,0.0296866,-0.3522314131,-0.3934127986]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/365","title":"How to augment data ?","comments":"Some samples in the dataset are too long, I want to divide them in several samples.","body":"Is there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```","comment_length":16,"text":"How to augment data ?\nIs there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```\nSome samples in the dataset are too long, I want to divide them in several samples.","embeddings":[-0.0921547189,-0.0665563941,-0.2976478934,-0.0226983447,0.0231515002,0.3546432853,-0.0144629246,0.2735868096,0.1672643125,0.1517361403,-0.1143039241,0.0416504815,-0.012731811,0.1264989078,0.1626523435,-0.2058319449,0.1696426421,0.0873438567,-0.1048345342,0.2275170088,-0.0604989119,-0.4782770574,-0.0764304549,-0.6003987789,-0.116329439,-0.0985122621,-0.3571471274,-0.3164554238,0.1215020195,-0.3110496104,-0.2560763955,0.0157930832,0.0884928107,0.280728966,-0.0001033333,-0.1614461541,0.1503436416,0.0704381019,-0.3406285942,0.4995984435,0.0792680457,-0.744867444,-0.0671173632,-0.0829781145,-0.1195215806,-0.4099473357,0.0361109711,-0.1217740104,0.3469309509,-0.143333286,0.2301471531,-0.0034851111,0.3508443832,-0.0468548723,0.1203243881,0.4339418113,0.2089641988,0.103696689,0.7045372128,0.1954564601,0.3622108996,0.1173312813,-0.1749290675,-0.0523294769,0.2249269783,-0.0106468778,0.2237958312,-0.4108218253,-0.0086791795,0.1689893305,0.3199174106,0.2096618265,-0.5844902992,-0.343937248,0.1641148627,-0.0818067342,-0.4392800629,0.0093651451,0.1522530168,-0.025808597,-0.2340262532,-0.2493508905,-0.0669920817,0.4183287024,0.0526309125,0.4480652213,0.0393823311,0.0323318653,-0.1009379029,-0.2489603311,-0.1554080993,0.1496042907,-0.1278482676,0.3905822039,-0.0650754347,-0.137221545,-0.3636052608,-0.2775261402,0.1508690566,0.0819587559,0.0989653319,0.1121674106,-0.1534661502,0.1329180747,0.2125650793,-0.342487067,-0.0359569117,0.2162266821,-0.0182291195,-0.1561191231,0.2932980955,-0.0941528976,-0.1405782849,-0.1551611722,0.0157136731,-0.0816375241,-0.1643199474,0.1094416156,-0.1225902736,-0.0545510948,-0.2011449635,0.0755982399,0.0647476166,0.1065370217,-0.2623960078,0.0036335993,-0.3086779118,0.2466654927,0.1757230908,-0.2556721866,-0.0760074407,0.1257371902,-0.1042709127,-0.1410946399,0.189350307,0.1175352484,0.2761706114,-0.0135685485,0.091220513,-0.1359864771,0.0667712688,-0.045589257,0.2609443665,0.13358365,0.1653188914,0.0235141478,-0.195586741,0.1977247447,-0.0709342733,0.0804622695,-0.3996521533,-0.018064918,-0.388820529,0.2422772646,0.1782967746,-0.1079659238,-0.1250014752,0.3729400039,0.2526080012,-0.3006931245,0.0890919492,0.0653070882,0.1029087678,-0.335172683,0.4814350307,0.0424609743,-0.3115996718,0.0160700977,-0.2066366225,0.2904301882,0.1537154317,0.0912341774,0.022525901,0.0954157561,-0.2393282652,-0.1379316151,0.3075343966,-0.1951460093,-0.3103208542,0.0397401154,0.1528491229,0.0272366367,-0.1793166995,0.3627087772,0.3158904016,-0.0820764452,0.2134750783,0.4815993309,-0.2080274224,-0.1461940557,-0.0089435065,-0.1502866894,0.3748991191,0.2271438837,-0.0541485064,-0.1308891028,0.2716344893,-0.4427704513,0.186020419,-0.1181798428,0.3794408739,0.1347667724,0.1710024625,0.3944880664,-0.0669768602,-0.2257087529,-0.2261144519,-0.0571709014,-0.0621210113,-0.2353710532,0.2126369625,-0.2676018178,0.0265949033,-0.1191705763,0.1559116095,-0.3202858865,0.2384455353,-0.1932311803,0.1870041043,-0.5061725974,0.0610345751,-0.2233324051,0.2332093269,0.1024845913,-0.2418869585,0.0311354287,0.2829587758,-0.2612969875,-0.2415936291,0.3624904156,-0.0874168128,-0.1497420222,0.1179000661,0.2847110629,0.1307742,0.1272636205,0.1616806686,0.1158380732,0.1229902804,0.0553345867,0.1622658372,0.2773104906,-0.0494484901,-0.1482307464,-0.0923692584,0.1870813966,0.0217762981,0.2239459604,-0.2352086902,0.244343549,0.2990865409,0.0249035638,-0.1108221039,-0.1024338529,-0.2274216712,-0.0303910505,0.3708465099,-0.2931017578,-0.0476089679,0.0122405682,-0.5297372937,-0.1937215775,-0.1063223928,-0.1373767853,-0.4725438654,-0.1186695844,0.1379899234,0.0225921012,-0.0970028266,0.3941380978,0.2377459854,-0.0841580778,0.2825246155,0.157409668,-0.0124919526,0.1354325414,-0.1045706049,0.0229452122,0.3718268573,0.1324880272,0.2272262275,0.0688881949,0.1317285001,0.1229524538,-0.0665533468,-0.1939289123,-0.1079429239,-0.2196011096,-0.0119364224,-0.6965720057,-0.1861798912,-0.3246959746,-0.1221535653,0.2187304348,-0.212937057,0.0391223133,0.2452377826,0.1445958167,0.083821021,-0.1548655033,0.090640299,0.2803698778,0.0687174201,0.203671217,0.0550056994,0.1548541039,0.5739241838,0.1061321124,0.1109563857,-0.2117847949,-0.3798002899,-0.221218884,0.0574127473,0.2341428548,0.2103155702,-0.0212551169,-0.0346369445,-0.1980115324,-0.0595406964,-0.1020124182,0.1978409588,-0.1250301749,-0.0831870735,-0.0428713299,-0.2294331342,0.1627705246,-0.4700725675,-0.2381106168,0.0521292984,0.1220853254,0.0903787166,-0.2794121504,0.3012722135,-0.0222951937,0.0641671419,-0.1506201327,-0.157328099,-0.179742679,0.4674947858,-0.3396631777,0.1202916577,-0.0520101637,-0.3217193186,0.1637375653,0.1866900027,-0.0275577847,-0.1436758637,-0.0969828665,0.0596022531,-0.1858639866,0.1629549712,0.0828569829,-0.2485521734,-0.0959439576,0.1055774838,-0.0112335328,-0.0025405842,0.3178333938,0.1828472167,-0.1401714236,-0.0772917643,-0.2146828622,0.1893817037,0.1611502469,-0.0674246475,0.0197200067,-0.2104617208,0.0398922637,0.0396238193,-0.096558854,-0.0287131257,-0.0833248198,-0.2234878093,-0.0048728199,0.2702593207,-0.121339865,0.1273220032,-0.2472563386,-0.3907498121,-0.1469990015,0.1526195556,-0.4917109311,0.2618219256,-0.2806887329,-0.1275280714,-0.4692977369,-0.1887200326,-0.1047920063,-0.1614059657,0.2652887404,-0.0583375655,-0.3561531603,-0.0844509378,-0.5110245347,0.6042351723,0.3101934791,0.1707242578,0.4370257556,-0.0654230565,0.0764412805,0.0232641548,0.4769046903,-0.2746005654,-0.3861761987,-0.1249652654,0.1016202271,-0.2826370001,-0.0520046577,-0.2705363631,0.1159948111,0.3876579106,0.3787555695,0.0095866229,0.0693712682,0.442481935,0.5657035112,-0.0990029648,0.026834229,0.1507284939,-0.0923482329,-0.0636569783,0.1204255447,0.0913113952,-0.0522900559,0.4356538057,-0.2565535009,-0.1560580134,0.1430496573,0.3302808702,0.1137514263,0.2782663107,0.2620791495,0.0637969524,-0.0065192198,-0.0017001667,0.4113222957,0.288048476,-0.2564451993,-0.3336547315,-0.2030014098,-0.4443608522,0.7946298122,-0.2818864882,0.3249369264,0.346630156,-0.2545506954,0.1173107997,-0.3671302199,0.2343787849,0.0904839933,0.1857845932,-0.485591203,-0.6768802404,0.0268063881,0.0139641957,0.0423151329,0.2104673684,-0.2805089355,-0.0812294111,0.325107038,0.418784529,0.6144989729,0.2699094415,-0.0058696372,-0.0597639233,-0.044176165,0.3404585719,-0.1314438432,0.1807764173,0.0006917327,-0.1594684869,0.0249673985,0.0587784462,-0.0081837261,0.2708622217,-0.120160155,0.2207409292,-0.2152338624,-0.0003138345,-0.2047333717,0.1338318139,0.1247476265,-0.2191448659,0.0907958522,0.2244816422,-0.0194555521,-0.3223442435,-0.1234588921,0.0400969945,-0.034627676,0.1531695276,-0.1436036974,-0.1229502708,-0.0875681117,-0.0178649556,0.268042922,-0.2091583163,-0.0579117164,-0.0595553368,0.0469833203,0.4749448597,0.0830102488,0.0931570679,0.1042758226,0.1381230205,-0.1402864754,-0.0227407217,0.2226509601,0.146052286,-0.1895871162,0.2490584999,0.1716060787,-0.4353044331,-0.3188324571,0.045046486,0.0474669486,-0.4183588624,-0.1262007505,0.339096427,-0.0147893541,-0.1913797557,0.1618330628,0.2115653604,-0.4530205131,0.3214583993,0.116969347,-0.1542836279,-0.1028691009,0.0575272702,0.1098165959,-0.3559036851,0.460819453,-0.028756598,0.0284118727,-0.2682008147,-0.0778703392,0.0540493838,0.097590901,-0.1610608995,-0.0160109028,0.1642109007,0.1389945894,-0.1316550821,0.1196246222,0.1152604967,-0.1619481444,-0.4005547762,-0.5174028873,0.2700829506,0.3813764751,0.2557670474,0.0352604054,0.1718021631,-0.3337787688,0.6286413074,-0.3503203094,0.1693044156,0.3655645251,0.2435310036,-0.069683753,-0.000521925,-0.1333026439,-0.0096566863,0.2015912533,-0.0205810033,0.0806734189,-0.2476274669,-0.0447631478,0.0393288247,-0.0765382648,-0.0372777432,-0.0468263254,0.0439796522,0.0571599714,-0.4754128754,0.0588158555,-0.0878189877,-0.1041302085,0.0902895927,0.3367287815,0.2322868407,-0.1297996044,0.2429975569,-0.09648031,-0.0747139901,0.0860500261,0.078189902,0.2276204079,-0.0717220902,0.1356787086,0.1772405803,0.3311461508,-0.2294764668,0.0507052355,0.1601716578,0.0505796932,-0.0301739238,0.2075314373,0.3017405868,0.1239082217,0.3368938565,0.3831687272,0.3606934249,-0.2259492725,0.0515181683,0.2434756607,0.1107781902,0.2150777727,0.2451016009,0.1342371553,0.1161525995,-0.1113182381,-0.0948397368,0.0579565391,0.0464523956,0.0027825187,0.5716448426,-0.076829657,-0.3242096603,-0.0987121984,-0.0120593682,0.0106788836,0.060458079,0.0611865707,0.5287553668,0.4854890704,-0.1693527102,-0.1019106358,-0.2474679947,0.146853894,0.3622612059,0.1380723268,0.4326592386,0.3992360234,0.4062812328,-0.0924077928,-0.4007273316,-0.0485337526,0.1135685816,0.1369554102,0.013221439,0.0804360509,-0.1333819926,-0.2592554092,-0.1468067318,0.2556401491,-0.4342423975,0.0638422891,-0.1879029572,-0.10465426,-0.164686352,0.2017745823,0.1639823318,0.1576624215,-0.338678807,-0.0578478128,-0.0050496738,-0.0924032927,0.2078321874,0.1546870619,0.2194679826,-0.0045988443,-0.0708321035,0.2528020144,-0.2889788747,-0.0531213842,-0.2084926814,-0.0132862097,0.0091538839,-0.1011135057,0.1081356406,0.2520718873,-0.1908330768,0.2433756143,-0.0326409526,0.1004198566,-0.4774592221,0.0416662879,0.1873503178,-0.4557678103,-0.1458471417,-0.3373722136,-0.2430200279,-0.3728682995,0.1628329158,-0.5795015097,-0.0825817287,0.2132816464,0.1386141628,-0.2262695581,0.4907077849,0.6586553454,0.2515902519,-0.0103413854,-0.3757664561,-0.363825649,0.1219911575,-0.2262852341,-0.062199682,-0.0669344887,-0.1357917637,0.2469606847,0.0311392806,0.2954438627,0.0071910089,-0.2777396441,0.1308324784,-0.120397009,0.1989426613,-0.2289441377,0.2858726978,0.2291433811,-0.0513811409,-0.073120378,-0.0155642917,0.0748966709,-0.0118626012,0.0380502306,0.1570263058,0.0344089344,0.3442349732,0.2281091064,0.108291015,-0.3414267898,0.0163098015,-0.3145482838,0.0357470848,-0.2887768745,0.297352314,-0.0171134137,0.1958338469,-0.060481593,-0.0329032131,0.2337293774,0.2723932564,-0.0157758333,-0.0742740706,-0.2161072791,0.2192266136,-0.243401587,-0.1523364186,0.0390131585,0.1653618217,0.0350599214,-0.038534198,0.1303575486,0.0786518902,-0.0162457768,-0.1023337469,-0.149935618,-0.5930963755,-0.0154275475,0.122527495,0.0666931048,-0.3561076522,-0.2237806916,0.339669615,0.1048019081,-0.0266910568,0.3971044123,-0.0637746006,-0.3384094834,-0.1645894349,0.2110570073,0.0602418333,-0.0124022029,-0.2504859567,-0.4126064181]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/365","title":"How to augment data ?","comments":"Using batched map is the way to go then.\r\nWe'll make it clearer in the docs that map could be used for augmentation.\r\n\r\nLet me know if you think there should be another way to do it. Or feel free to close the issue otherwise.","body":"Is there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```","comment_length":45,"text":"How to augment data ?\nIs there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```\nUsing batched map is the way to go then.\r\nWe'll make it clearer in the docs that map could be used for augmentation.\r\n\r\nLet me know if you think there should be another way to do it. Or feel free to close the issue otherwise.","embeddings":[-0.0169448406,-0.1455653608,-0.3001067638,-0.085400261,0.1244004071,0.1755867749,-0.0629433766,0.3091185093,0.1504426748,0.1681790799,-0.1306253672,0.187817201,0.0309323147,0.1995354891,0.1029892415,-0.0984935388,0.0637115464,0.1679257452,-0.2380012125,0.1587176472,-0.0724246055,-0.3081819117,-0.0647776127,-0.5189904571,-0.088035509,-0.0856838748,-0.2475226521,-0.0627598315,0.016347168,-0.3594219685,-0.2194650471,0.0810869411,0.0797761083,0.1623182744,-0.0000954823,-0.167870909,0.1755304486,0.0509370416,-0.2073566914,0.4534386992,0.0881541222,-0.6956601739,-0.0384642966,-0.1407383531,-0.1430088729,-0.4254902005,0.0193878524,-0.0791336894,0.393286854,-0.1298876554,0.3211502135,0.0678816885,0.3687788844,-0.0091608884,0.1637964994,0.2675371766,0.0934257507,0.0432249308,0.6970156431,0.1130867824,0.1968534589,0.275015533,-0.055956807,-0.0547927804,0.2242706716,-0.0399329886,0.1791331172,-0.4076254964,-0.0293971486,0.0814065933,0.361102283,0.2461704761,-0.4855210483,-0.2073439807,0.1587932259,-0.1088655069,-0.3031988144,-0.1451983899,0.077515237,-0.0084158154,-0.17961739,-0.2245142162,-0.1105176359,0.2781727016,0.0317730308,0.4702418745,-0.0542672537,-0.0470597856,-0.0925379768,-0.2123641819,-0.203853786,0.2598105669,-0.1692332923,0.3838725686,-0.1215264276,-0.1353017837,-0.1618470252,-0.2341883332,0.1250404119,0.0737611726,0.2169513404,0.2232525498,-0.126928404,0.1107358187,0.2387806475,-0.2189143598,-0.0347953998,0.1099753901,0.026437778,-0.0756686404,0.1884724945,-0.0773251802,-0.0992698893,-0.2176902741,-0.0355629064,-0.0619829111,-0.0680253729,-0.0185658894,-0.100947246,0.0252062753,-0.1755307615,0.1599729508,-0.0470886491,0.178590253,-0.1924699694,0.089439936,-0.1249763742,0.2691834867,0.1609803438,-0.1669647098,-0.179566592,0.14058204,-0.1030169353,-0.2629325688,0.1747775376,0.1583021581,0.3005460501,0.0060822605,0.2012646794,-0.0010647125,0.0921211541,0.0558235385,0.3494169414,0.2594133615,0.2155177891,0.0042818142,-0.2074762583,0.1923594624,0.0386597067,0.0949691683,-0.3523138762,0.0304879546,-0.3897178173,0.3262459338,0.1028606221,-0.1323000491,-0.0619796664,0.3417808712,0.1576268822,-0.2634567916,-0.051137533,0.0761926398,0.0479429476,-0.4188563824,0.4802416861,0.0124730654,-0.3166131675,0.0820771679,-0.0641261488,0.2990228236,0.156714946,0.0775551572,-0.1071477756,0.0946886539,-0.2257410139,-0.0112321898,0.3483181,-0.1518348157,-0.2772152722,0.0411917642,0.1846949011,0.070836246,-0.0940926746,0.2117481381,0.2242880613,-0.088944003,0.230582729,0.3381913304,-0.1150586158,-0.0823141411,-0.166919142,-0.103600435,0.2578800321,0.1245642677,-0.0462001152,-0.1519450694,0.2800815701,-0.3861431181,0.2009379119,-0.2013785094,0.3344907165,0.2194981426,0.263907522,0.2855584919,-0.0182075873,-0.1426215321,-0.1639259756,-0.0198118072,-0.1903655976,-0.0073357313,0.1801940948,-0.2361218929,-0.1505279839,-0.1746350527,0.0857909694,-0.3688027263,0.3643602431,-0.143016845,0.0284695011,-0.4876209199,-0.0113339555,-0.145951733,0.1322830915,0.0187699851,-0.1852544546,0.0301892981,0.1393851638,-0.2906778157,-0.1264595985,0.2679728866,-0.0552551448,-0.1496492177,0.0936707705,0.3210107982,0.0813130364,0.1584233195,0.1777062267,0.1764170676,0.1058721989,0.076997228,0.1516087651,0.2718897164,0.0201323647,-0.0853935108,-0.0976314321,0.28532511,0.209307164,0.0269206576,-0.1385434568,0.1989011019,0.2945254445,-0.0012597077,-0.1319825202,-0.0646301806,-0.2141540647,0.0002325894,0.2770747244,-0.312436074,-0.1753575802,0.1000304297,-0.3331007063,-0.0305125508,-0.0974095464,-0.2062395662,-0.3846090436,-0.1452928036,0.0785186142,-0.0137724206,-0.1569926143,0.4487093389,0.1871991456,-0.046175357,0.2047528327,0.1581123471,0.0465566106,0.0721344501,0.0315404572,0.0949202478,0.3714199662,0.0627289414,-0.0108150085,0.101358667,0.0127511006,0.1902557015,-0.1662208587,-0.2549054325,-0.0093709631,-0.1310875416,0.0247405376,-0.4721281826,-0.1677758098,-0.417199254,0.0260447096,0.2850016952,-0.3464559913,0.1147303507,0.1940826029,0.0446637273,0.0734466985,-0.013753878,-0.0212275814,0.2723361254,0.0149814077,0.2983080745,-0.04408326,0.0612418056,0.4653105736,0.1847611368,0.1090756357,-0.1462473273,-0.6015242934,-0.0726515353,0.0547881983,0.2365457863,0.2434007227,0.0315064788,-0.0687150732,-0.1536568403,-0.0579480641,-0.2477636337,-0.0023935768,-0.1426816136,-0.1327076107,-0.0629425645,-0.3509025574,-0.060577523,-0.2882076502,-0.2773609459,0.0394844338,0.1879211217,0.0891084373,-0.178179428,0.2530728281,-0.0417772494,0.1533772051,-0.0730848312,-0.3125744164,-0.2820256352,0.5944551826,-0.3446516395,-0.1548205763,-0.0636408851,-0.3578438461,0.2325512171,0.1982691437,-0.0591976233,-0.3129131794,-0.0276914071,-0.0003507619,-0.1766397357,0.1462289691,0.1281085014,-0.1459636241,-0.2287901491,-0.0527053252,0.0190038048,0.0899171531,0.3732918203,0.1933889538,-0.1765831411,-0.1203416884,-0.0765810534,0.0706177652,-0.0274457764,0.0174370408,0.2258876711,-0.1589358449,0.1872397214,0.0216044635,-0.1034894064,0.0051153959,-0.1158913374,-0.2093849331,0.115248166,0.2734782696,-0.2112435997,0.0363749005,-0.1313429624,-0.3586927652,-0.0923408791,0.2182246894,-0.2953594327,0.1778493226,-0.1908435374,-0.0640851557,-0.4218356609,-0.1551457196,0.0004917794,-0.1836362183,0.183932066,-0.1080469638,-0.378593266,-0.0123381456,-0.541719377,0.5136914253,0.2506775558,0.0765951797,0.3777765334,-0.1029010564,0.0396285132,-0.1556667686,0.3985165954,-0.3002894223,-0.274071008,-0.102881141,0.0311428383,-0.2915138006,-0.0026046115,-0.409483701,0.111860171,0.4298164546,0.3906328678,-0.0598621666,-0.003145142,0.3289982677,0.4012627304,-0.1535186768,0.1334208846,0.0586460568,-0.2008094192,-0.1939124912,0.0616844818,0.1312017888,-0.0089591807,0.4233728647,-0.2347906381,-0.0938647687,0.0015045687,0.2801461816,0.2089467198,0.3529753685,0.3294098973,-0.0130727319,-0.038533479,0.0447142683,0.3722147644,0.3186182082,-0.0576771051,-0.2116483301,-0.1892740726,-0.4861545563,0.6014270782,-0.0915881395,0.2584628761,0.2055285573,-0.3042526245,0.0586816445,-0.2436978072,0.2452334017,0.1928693652,0.0815605223,-0.2515930235,-0.5197947025,0.0003864556,-0.1165876612,-0.0839469358,0.2205493748,-0.1855596602,-0.1190514937,0.339748472,0.3805867136,0.7213774323,0.3727225661,-0.0444136262,0.0649095401,-0.2082360238,0.2102980763,-0.1514592767,0.1676413715,-0.1084763855,-0.149298504,0.0927325189,0.1388881356,0.0394611433,0.2289902866,-0.1073729321,0.182310164,-0.1157019213,-0.0196163487,-0.0945398584,0.0810038894,0.204097122,-0.1747233868,-0.0526404269,0.2750528157,-0.0100245886,-0.2535112798,-0.1283598989,-0.1450096369,-0.0953414738,0.0102564767,-0.0497433916,0.0456539728,-0.1479800642,0.0118607888,0.2016211897,-0.2020267844,-0.1585796773,-0.0038156509,0.0804932192,0.384767592,-0.0004489952,0.2140657604,0.2011197954,0.2012854218,-0.0893240049,0.0679803863,0.2940028012,-0.0240132436,-0.221397832,0.1572824717,0.0887473524,-0.47080037,-0.2747162879,-0.0228284746,0.061899256,-0.3910628557,-0.0462359972,0.4183835685,-0.0696107373,-0.2254306674,0.2535315156,0.2210981846,-0.440109849,0.3310343921,0.2801369131,-0.1314958483,-0.0531033017,0.0921126306,0.1217627451,-0.2552761436,0.4135271311,-0.0591217875,-0.0225003827,-0.2673148811,-0.1771594137,0.1137655899,0.0449483283,-0.0836923793,0.0587839298,0.0113007184,0.0981936082,0.029810762,0.1216481104,0.0794437975,-0.1278665662,-0.3241097927,-0.6035918593,0.1919022202,0.3020663857,0.2361082435,-0.018614376,0.1828116328,-0.242685318,0.5098890662,-0.4830343425,0.1032637581,0.2558164299,0.2338089347,-0.0276298039,-0.0949623734,-0.002896955,-0.1241882294,0.2989677191,-0.036212042,-0.0985418484,-0.3450575173,-0.0860115588,0.0079727545,-0.0643564537,-0.0786527097,0.0277407486,0.0084652314,-0.0038844266,-0.3635752201,0.0797458217,-0.1006625146,-0.0201908033,0.2206115574,0.3439671695,0.2404594719,0.0768495128,0.1137532666,-0.0868780538,-0.0303420555,0.1041279882,0.0236726645,0.1436489671,-0.1040573418,0.3017784953,0.1183745712,0.1194551587,-0.0834000111,0.1052599773,0.0853780657,0.0126013327,-0.1063187048,0.3554545343,0.2193899304,0.084182404,0.2463787198,0.2900204659,0.4652894437,-0.3328395784,0.0970400497,0.1872601509,0.1010139138,0.2055881768,0.1568875909,0.2433466017,0.0172064342,-0.007142147,0.0407140851,0.0353426002,-0.0917153955,-0.0693186224,0.3280282021,-0.1989980191,-0.2737838626,-0.0883590579,-0.093399018,0.0289199594,0.1247581616,0.0135724153,0.3981097341,0.4890503883,-0.1650624424,-0.0689255446,-0.1213798076,0.2238374054,0.2790647447,0.1664381176,0.4490498006,0.4822932184,0.3401313722,0.0608148947,-0.4302314222,-0.0536186919,0.1077175587,0.0273704268,-0.050312046,-0.0741057992,-0.2121245116,-0.22331734,-0.084065713,0.2226672173,-0.3029943705,0.1754538119,-0.2292858511,-0.1337379068,-0.1996999979,0.0953209996,0.0933138505,0.234337166,-0.27042225,-0.0077713248,0.1015135795,-0.1385533065,0.2698786855,0.2487731576,0.2883717716,0.0173134208,-0.1522371024,0.2424096465,-0.1098287404,-0.0940553024,-0.2680336833,-0.0830814391,0.0670163333,-0.0245250389,0.1339440197,0.2917024791,-0.3056378663,0.2154829353,0.033137653,0.0881298631,-0.2701213062,-0.0004437366,0.2386180162,-0.4933004081,-0.1219578236,-0.4256601036,-0.253367424,-0.428517282,0.0797475874,-0.3791305423,-0.0102188531,0.0892607868,0.1759544462,-0.1422221959,0.5551962256,0.4812188745,0.0739833713,-0.0124305068,-0.3123758435,-0.4836330414,0.2194251269,-0.0562758557,-0.1390255392,-0.0185142532,0.0015062917,0.2872326672,-0.0130773811,0.2859382033,-0.0585887954,-0.2068437636,0.0713019669,-0.3207525909,0.2485234737,-0.1291376352,0.3100477755,0.2314136177,-0.0674786568,-0.0694197118,-0.1177903712,0.2514120638,-0.0413070805,-0.0110294726,0.0740177184,0.1275631636,0.4170662463,0.2210706472,0.0901792794,-0.4682819247,-0.0659741089,-0.3132880926,-0.0571288243,-0.3910633326,0.2766557038,0.0515570343,0.3043657243,0.0733211339,-0.2173111588,0.1498270035,0.2771361172,0.0013820529,-0.0360093303,-0.1950936168,0.32855919,-0.0998544767,-0.1976162493,-0.1854780614,0.1860242933,0.0083920108,-0.1333105862,-0.0704701021,-0.2347144932,0.2004674524,-0.1623768657,-0.1646199226,-0.4515866339,-0.0501348414,0.0992007703,0.0184291378,-0.4311226308,-0.1648070067,0.2243675888,0.0660405308,-0.1228901669,0.4751557112,-0.1930029541,-0.3418256342,-0.0845567808,0.1866379827,0.1729302853,-0.0744506195,-0.3311717212,-0.3802111447]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/365","title":"How to augment data ?","comments":"It just feels awkward to use map to augment data. Also it means it's not possible to augment data in a non-batched way.\r\n\r\nBut to be honest I have no idea of a good API...","body":"Is there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```","comment_length":35,"text":"How to augment data ?\nIs there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```\nIt just feels awkward to use map to augment data. Also it means it's not possible to augment data in a non-batched way.\r\n\r\nBut to be honest I have no idea of a good API...","embeddings":[-0.0686048344,-0.030538993,-0.2990654707,-0.1101847813,0.0624446608,0.335452497,-0.1749581844,0.2741323113,0.1564610451,0.132221967,-0.044367671,0.0903131962,-0.017896561,0.2145764679,0.1571808308,-0.1104836017,0.1782806069,0.1195115671,-0.0964070708,0.2644732594,-0.068970643,-0.4659138918,-0.0256337281,-0.5230019093,-0.0577692166,-0.1176650152,-0.3457253575,-0.2689724863,0.0295523014,-0.3138853014,-0.1750326008,0.0457123145,0.1227496937,0.2257133126,-0.0001022616,-0.1665669084,0.2202338725,0.0800639912,-0.3130410612,0.3663995564,0.0928266197,-0.7476427555,0.0595020093,-0.0558866635,-0.120866172,-0.4913313389,0.1032007635,-0.1203033775,0.3686804175,-0.1997427195,0.2372495234,0.0310002714,0.3845404088,0.006108203,0.0283367317,0.4011588693,0.1855530292,0.0620645024,0.7107824683,0.1219980195,0.361277163,0.0551538989,-0.2042975426,-0.0643414855,0.242851451,0.0068094409,0.1272957474,-0.4292892516,0.0064514223,0.1234805584,0.4534501433,0.2251906097,-0.5979700089,-0.2979202271,0.1898910105,-0.0986823291,-0.4096084833,-0.0384184308,0.1686005443,0.0455293134,-0.2218215913,-0.2193209827,-0.0807439536,0.3997989297,0.1386635005,0.3918269277,0.0202561822,-0.0249934979,-0.1155003384,-0.2687825561,-0.1262351424,0.1236284077,-0.2063962519,0.452824831,-0.1286777258,-0.1373749077,-0.2147859484,-0.2292485088,0.1077811792,0.1561010182,0.1647679955,0.1895729601,-0.2065299004,0.1317075193,0.2149588764,-0.3526519239,0.0053517423,0.1251104176,-0.0039195661,-0.1864408851,0.3491332531,-0.0952046737,-0.0962555408,-0.0527517274,-0.0202529766,-0.1264339536,-0.1354442388,0.048149541,-0.0846445262,-0.0587637015,-0.2746983469,0.1306734085,-0.0200754125,0.0558104403,-0.2272862196,0.0899396613,-0.1803676784,0.2548744678,0.3009600639,-0.2191844583,-0.0647228807,0.1165493801,-0.0582758598,-0.0933070481,0.1988828778,0.1463552713,0.2114189863,0.0358114168,0.0836604089,-0.1050397009,0.0959555879,0.0261888225,0.2474202663,0.1416900605,0.2138693482,0.0465458594,-0.1783533096,0.1664629728,-0.0362404473,0.122170113,-0.4160352647,0.0128166247,-0.389531672,0.2586481869,0.1247188225,-0.2470235825,-0.2244285196,0.3998591006,0.0999816284,-0.294292748,0.0754973292,0.1829731762,0.1183260307,-0.3427895308,0.480463326,0.0219706092,-0.3363804221,0.0242480338,-0.1549204141,0.3433930576,0.1077004746,-0.0038099964,0.0000533018,0.198673889,-0.1652544141,-0.158856675,0.3236880004,-0.2099808455,-0.2859991491,-0.0238593593,0.199710682,0.0716654137,-0.1360217929,0.3386332691,0.2250989527,-0.0124285584,0.2704179883,0.4746152759,-0.1091526896,-0.1832757294,0.0216448363,-0.1854127795,0.2906926274,0.1925904304,-0.1357132345,-0.1341420859,0.3120118082,-0.510428369,0.1714842469,-0.2035457343,0.3540636897,0.1331249177,0.2125593871,0.395483762,-0.0858008265,-0.2248328626,-0.2453152537,-0.0238486901,-0.0882091299,-0.1936911196,0.1417811513,-0.2581710815,0.0926206112,-0.1290224344,0.1425096691,-0.3532492816,0.2598412633,-0.2204664499,0.0691283643,-0.5130602121,-0.0564298145,-0.1649242789,0.2361110747,0.066379644,-0.1898200512,-0.0410508253,0.3220966458,-0.3198963106,-0.2195447236,0.3375580609,-0.1587689668,-0.149061799,0.1133331582,0.2951254845,0.0626491159,0.2453920841,0.263964355,0.3063476682,0.0540742464,0.1400565803,0.0641869828,0.3234606385,-0.0238998681,-0.1454495341,-0.082139045,0.1868114322,0.1285758764,0.1723138243,-0.2582692802,0.2777769268,0.3268328905,0.030629823,-0.123232834,0.0002799981,-0.2246222049,-0.0325745121,0.3339380622,-0.2960822284,-0.0950020179,0.0237667747,-0.5188763738,-0.1817835569,-0.0920926556,-0.1316090822,-0.4617499113,-0.1724644899,0.0933030322,-0.0977834538,-0.1816693693,0.3911105096,0.1194313243,-0.0705558211,0.2989989221,0.1878963858,0.0435693376,0.1824368387,-0.0562934503,-0.0972770974,0.3971741796,0.222453028,0.2089003623,0.0221775714,0.0124482121,0.1462172121,-0.1039224342,-0.1517767012,-0.1404375583,-0.2071709782,0.0644374415,-0.6927348971,-0.1695429385,-0.4011876881,-0.0822533667,0.2475290596,-0.2263851762,0.0246047787,0.1577132791,0.1307263076,0.0276360642,-0.2404650599,-0.0117862448,0.237435326,0.0463207103,0.2038218826,0.00970532,0.0305844638,0.5850212574,0.1105990484,0.2222079486,-0.1531336606,-0.4871169627,-0.1613034755,0.122722283,0.2028142959,0.1914292276,-0.0455105864,-0.0616297498,-0.1874715239,-0.1129961312,-0.1646448076,0.1367574632,-0.156336844,-0.1317685843,-0.0678249002,-0.2224403918,0.1538297385,-0.4205823541,-0.232258141,0.0845218375,0.1622949392,0.0418925881,-0.2198492438,0.2272851616,0.0155684538,0.0743990764,-0.0559060834,-0.1184881479,-0.1790708005,0.5194659829,-0.263531357,0.0046393126,-0.0111848051,-0.3212347329,0.1550671607,0.1527661383,-0.0102336062,-0.2706620693,-0.0787213594,0.0689047128,-0.1759991795,0.159502551,0.082227923,-0.1964635849,-0.1351796985,0.0817197412,-0.0439850651,-0.0049758386,0.3328840137,0.1656613797,-0.0508646294,-0.1112074777,-0.1176544949,0.0994257778,0.1196711957,-0.1047912538,0.1222100109,-0.2279505879,0.1350808293,-0.0081044687,-0.070765458,-0.1330641508,-0.1515099853,-0.1862459034,-0.0248101167,0.2846742868,-0.132504195,0.0450830758,-0.1982472837,-0.4388124943,-0.0720951855,0.0953867733,-0.4609332681,0.2318614423,-0.2455287278,-0.1252286285,-0.4214309454,-0.2600910664,0.0442185253,-0.1415375918,0.2868053615,-0.0417852029,-0.3027790487,-0.0875807777,-0.5137757063,0.6783239841,0.2343788743,0.1515173614,0.466348201,-0.0487828143,0.1367245913,-0.0199855696,0.3645620942,-0.2393355072,-0.4498530626,-0.1944304258,0.0865531266,-0.2527430058,-0.044155743,-0.3235747814,0.1160063893,0.369918406,0.3898539245,0.0368691869,0.0640150309,0.3046593964,0.4886601567,-0.1422290951,0.0118847247,0.1554080695,-0.1388306022,-0.0950961784,0.0397458486,0.0608253404,-0.0051520155,0.4857851863,-0.2444624007,-0.1724437177,0.1839718521,0.342212677,0.082583122,0.2498890162,0.3170359135,0.0022915024,-0.0581324324,0.075914517,0.4048277736,0.255546391,-0.1385137439,-0.329158783,-0.2433080226,-0.4462713301,0.7755662203,-0.3907968998,0.3191578686,0.3719860017,-0.2240998894,-0.0369268171,-0.4037377238,0.1591902524,0.1295942366,0.1980758756,-0.3826711178,-0.6550890207,0.1126242727,0.0173241924,-0.0222812388,0.2155147493,-0.2285441905,-0.014586526,0.3182497621,0.4359499812,0.6240648031,0.3476094604,0.0405567177,-0.0611746497,-0.2139322013,0.3333134651,-0.111597456,0.2151458561,-0.0151036251,-0.133129105,-0.0182747431,0.095151186,-0.0188474655,0.2144787163,-0.0471988469,0.2760601938,-0.1295548081,0.0813131705,-0.0696477666,0.0608043931,0.2009563893,-0.2505064309,-0.0092637744,0.2317185402,0.0416478924,-0.304641813,-0.1959142834,-0.0213053394,-0.0609897412,0.1346796602,-0.1553375274,-0.0658601895,-0.0760360211,-0.0299906489,0.303812772,-0.2288011312,-0.1412012279,-0.0339293666,-0.0257022846,0.437728852,0.0033879343,0.0600702688,0.1074737608,0.1966538429,-0.1127087101,-0.0034191848,0.2407535464,0.1189403981,-0.1846676469,0.2388106436,0.145596087,-0.4697574079,-0.276824683,0.0003467883,0.0107705388,-0.3005549312,-0.1240957305,0.4121633172,0.0356791355,-0.1533350199,0.1705083847,0.2669606507,-0.370700866,0.3103790581,0.1299424469,-0.187943697,-0.0595060252,0.067980431,0.0540506914,-0.3736447692,0.489371568,-0.0558591932,0.0053876871,-0.1790041775,-0.1076268032,0.0915155485,0.211471647,-0.2339433432,0.0985867456,0.1644648761,0.1982328147,-0.1207278967,0.1492046565,0.0705930069,-0.1806441694,-0.388605386,-0.5093191266,0.2766167223,0.3227663636,0.2374886274,0.115533255,0.0207692459,-0.2241155207,0.5227136612,-0.3601153493,0.0942513943,0.3992094994,0.1778465062,0.0366686918,0.0398631692,-0.1118739694,-0.0668565556,0.206897378,-0.1160575375,0.0647535175,-0.2455611378,-0.0262573883,0.0453972742,-0.0402676538,0.0513609946,-0.0631357878,-0.0252775289,0.0155973341,-0.52346766,0.0147896726,-0.1247652099,-0.1039089784,0.1025296077,0.2808092535,0.2882390916,-0.1843795031,0.2178685516,-0.1010931954,-0.0590312928,0.0583570264,0.0271070469,0.2443356961,-0.0623425245,0.2635540664,0.1539420635,0.3184332848,-0.2002068907,0.0232464485,0.1509010643,0.0137541834,-0.0927897766,0.2279708833,0.2638510466,0.1314636767,0.2674955428,0.3891095817,0.3625017405,-0.2104796469,0.0550903343,0.2448414415,0.1726584882,0.1529679745,0.1774129421,0.2189472914,0.0491280742,-0.0819100961,-0.0799718872,0.0684910566,-0.0467879139,-0.0280635618,0.3850809932,-0.1099335328,-0.3174657226,-0.0624713376,-0.0103150476,-0.0300272945,0.0398962013,0.0332576893,0.3949257433,0.5789949894,-0.2148539275,-0.0597625785,-0.2156558335,0.1446567923,0.2955485582,0.2111576796,0.4773837626,0.5241519213,0.334281534,0.0388525128,-0.3386428356,-0.0191397443,0.0598636791,0.1324778497,0.0130388243,-0.0174242742,-0.1390550882,-0.2302507609,-0.1648145318,0.2870804667,-0.395457536,0.0660755411,-0.197888121,-0.0959370658,-0.1431284845,0.1666704565,0.2247415036,0.1581085771,-0.3085287213,-0.076001212,0.02086013,-0.0995984972,0.1876104176,0.2136706561,0.2148960233,-0.0233282186,-0.1248137802,0.3304813206,-0.268557936,-0.0725425929,-0.1951676607,-0.0014982627,-0.0101772193,-0.1138462648,0.0031902422,0.2585468292,-0.1697767079,0.2449406832,-0.1241001189,0.2383805513,-0.4589588642,0.0838287324,0.267906487,-0.4813739955,-0.0507754833,-0.2813665271,-0.1576600075,-0.3973490298,0.1848388165,-0.430730015,-0.1620962173,0.1841868609,0.1441444457,-0.1859395951,0.5007240176,0.7532907128,0.214844659,0.1294790506,-0.2797742486,-0.3995542228,0.1798867881,-0.0666404963,-0.1061073244,-0.0972601697,-0.1658447832,0.2976010144,0.0156700052,0.3528470695,-0.0596828312,-0.2333123982,0.0900555477,-0.1315961331,0.2456474602,-0.1332493722,0.304369837,0.18254821,0.0187437758,-0.1065884233,-0.0690148994,0.1248448342,-0.1145304516,0.0239362717,0.138708204,-0.0179992672,0.4528111815,0.2902180851,0.092264019,-0.4518582523,0.0232942402,-0.3329681456,0.0347301364,-0.3404895067,0.3179118633,-0.1197086275,0.1195727438,-0.0525700115,-0.1263474226,0.2401863486,0.2379948348,-0.0941953585,-0.0937096849,-0.2329601198,0.1638816595,-0.2874905467,-0.2557799816,-0.0693471879,0.122328639,0.0813199803,-0.0710071921,0.0695215017,0.0525320768,-0.0644512922,-0.1906768084,-0.1212419569,-0.5665037632,-0.0719178841,0.0426800624,0.1491261125,-0.3013896346,-0.2243121862,0.342287451,0.0864360556,-0.0063539427,0.4844034314,-0.1697268933,-0.2994285524,-0.1539603323,0.235177964,0.1029732674,0.0103906104,-0.339487642,-0.3807590902]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/365","title":"How to augment data ?","comments":"Or for non-batched samples, how about returning a tuple ?\r\n\r\n```python\r\ndef aug(sample):\r\n # Simply copy the existing data to have x2 amount of data\r\n return sample, sample\r\n\r\ndataset = dataset.map(aug)\r\n```\r\n\r\nIt feels really natural and easy, but :\r\n\r\n* it means the behavior with batched data is different\r\n* I don't know how doable it is backend-wise\r\n\r\n@lhoestq ","body":"Is there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```","comment_length":60,"text":"How to augment data ?\nIs there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```\nOr for non-batched samples, how about returning a tuple ?\r\n\r\n```python\r\ndef aug(sample):\r\n # Simply copy the existing data to have x2 amount of data\r\n return sample, sample\r\n\r\ndataset = dataset.map(aug)\r\n```\r\n\r\nIt feels really natural and easy, but :\r\n\r\n* it means the behavior with batched data is different\r\n* I don't know how doable it is backend-wise\r\n\r\n@lhoestq ","embeddings":[-0.0632994846,-0.0556469187,-0.2771684527,-0.0368639417,0.0604924336,0.2008812279,-0.0867594108,0.3342858255,0.2316298783,0.1316544712,-0.0439282358,0.1579211652,-0.089111492,0.2297118306,0.1699337959,-0.0823392421,0.1174822375,0.105264537,-0.2327551842,0.1378301978,-0.1590786278,-0.4287635088,-0.0023527686,-0.5560159683,-0.0530026592,-0.1400103718,-0.2512088418,-0.1560887098,-0.0858923569,-0.2709840238,-0.1612508297,-0.0034372846,0.0691071302,0.2275177091,-0.0001008931,-0.1184825152,0.1640579551,-0.009550795,-0.3493941128,0.329985857,0.1156730875,-0.6480942965,0.0277305916,-0.189093098,-0.079566434,-0.3808273077,0.0249030348,-0.1372267753,0.3346520066,-0.2232288718,0.2483513355,0.0744488463,0.278870672,-0.0202090424,0.0942939371,0.4097540379,0.1738606691,0.0109542906,0.7498846054,0.1245981306,0.2768816054,0.1257326603,-0.1657391042,-0.0216296017,0.1608050019,0.0382642336,0.2337287664,-0.3121768534,0.0138915488,0.0777553543,0.2978186905,0.1781383604,-0.5725196004,-0.2732112408,0.1189942807,-0.1241749749,-0.4445827901,-0.0784591064,0.0991723612,0.0009637591,-0.2592446804,-0.1105473191,-0.0711180493,0.4402382374,0.0891344324,0.4257209003,0.0323687866,-0.0025471232,-0.1321633756,-0.2531953752,-0.1832222044,0.097515516,-0.1246486157,0.4209003747,-0.0579503365,-0.1266138703,-0.0677434877,-0.3055943549,0.1858882457,0.1099734455,0.192939043,0.2106933743,-0.2272464782,0.154216826,0.2658590376,-0.2425051183,-0.0022894356,0.2041574121,-0.0758706927,-0.1693224609,0.2239778042,-0.0477011092,-0.0655767545,-0.0060860398,-0.026939854,-0.0382164568,-0.0631440207,0.0974364653,-0.1065984815,0.0104910042,-0.178797096,0.0322605483,-0.041901134,0.1070414633,-0.1549945325,0.0590064488,-0.1547713727,0.2281045169,0.2352566272,-0.2266714126,-0.1633024961,0.0801987126,-0.0895977616,-0.2012700886,0.218915388,0.2489342391,0.2741272748,-0.0490846969,0.14571172,-0.0827633962,0.1349953413,0.0380844586,0.3588130772,0.1116046831,0.242141813,0.0051657469,-0.1018314958,0.1701470166,-0.0962676108,0.0845347568,-0.3631028831,-0.0655138791,-0.3702999949,0.2830742002,0.0265305657,-0.1139076576,-0.2542050183,0.3372006714,0.1885350347,-0.1818661243,0.0896435231,0.059528444,0.1152425334,-0.3669933379,0.4468903244,0.1134103462,-0.373280555,0.0227278117,-0.0959550664,0.2931321859,0.0677033886,0.0707121119,-0.0800485909,0.1440456808,-0.2160286903,-0.0971201062,0.4118097425,-0.1938976645,-0.3195028603,-0.017013805,0.1635004729,0.1572091877,-0.0828367248,0.257635951,0.3237344325,-0.0105778491,0.3484742939,0.4347503185,-0.1468312293,-0.1505580992,-0.0593645759,-0.1252626479,0.2915830612,0.1138485521,-0.1347663254,-0.0784999728,0.3138869703,-0.4948126674,0.2585176229,-0.205004558,0.3731128871,0.0273981169,0.2850507498,0.3254345059,-0.086201936,-0.1755846888,-0.3358184099,-0.038416978,-0.1873887926,-0.2033323646,0.1539624482,-0.2810046077,0.0203976836,-0.0487804785,0.093945168,-0.3281690478,0.2975734472,-0.2360204905,0.0740346089,-0.4809199274,-0.04141986,-0.2468388379,0.1729852259,0.0421877652,-0.225071013,-0.0013361446,0.2824294865,-0.3155222535,-0.2743340135,0.3438553214,-0.1667285115,-0.1599190533,0.1341206431,0.2437431812,0.0334378406,0.1577346772,0.2464082986,0.1587113589,-0.0070624952,0.2250783443,0.1397742033,0.2573010921,-0.1308217794,-0.1930284053,-0.0539353155,0.2313882411,0.1118431613,0.1938218623,-0.2354970723,0.2226424962,0.2522436678,-0.0014300012,-0.1625969559,-0.1034443527,-0.2591074705,0.0364792347,0.386831671,-0.2824320197,-0.0578859448,0.0966756046,-0.3447910547,-0.1486644894,-0.0739679113,-0.2248086482,-0.4727678597,-0.2004031539,0.1041828692,0.0174954701,-0.0523001701,0.4373734891,0.1997584552,-0.0958631858,0.3652170599,0.1758301556,0.0071631689,0.1341395676,-0.0781101808,0.0565464646,0.4229617715,0.1454011053,0.1783400625,0.0495876186,0.0405426472,0.0778422058,-0.1019479856,-0.2347960472,-0.0607714057,-0.1969971657,0.0359817743,-0.6457634568,-0.1388046443,-0.3514402211,-0.0789289325,0.2720778883,-0.3246179223,-0.0186638944,0.0961515903,0.0940212235,0.0681630895,-0.2781265676,0.0735720843,0.2336939126,0.0297429841,0.2181044668,-0.0871243477,0.0034386315,0.5335658789,0.1826736927,0.1518421024,-0.1639770418,-0.4550438523,-0.0837381631,0.0782779604,0.2234655321,0.3338837922,-0.0333237424,-0.0305688921,-0.2349003404,0.0122920778,-0.0477756783,0.0319390148,-0.1580991298,-0.0603867434,-0.0506211929,-0.2878417671,0.0573045686,-0.4049879313,-0.2539071441,0.0681859553,0.219306156,0.1133335084,-0.1974880099,0.2885578871,-0.022100646,0.1497111768,-0.2128760517,-0.2604113817,-0.213245064,0.5349240303,-0.3195921183,0.0504669659,-0.1254554689,-0.3893169165,0.2763389647,0.2912285924,0.017255336,-0.2947852612,0.0436653607,0.1051568761,-0.1098918691,0.2192497551,0.2075728029,-0.1153014526,-0.1686352044,-0.0432938263,0.0579324365,0.0500366166,0.3731495142,0.2342387438,-0.1522739083,-0.1461705416,-0.0776285306,0.1971176565,0.0783795342,-0.0774530247,0.1273072511,-0.1476609856,0.1118197367,-0.0258344691,-0.0768003613,-0.1838399917,-0.1954331696,-0.2768783271,-0.0051435977,0.1864017248,-0.0620648526,0.0897915959,-0.0901137292,-0.4138380885,-0.1433576643,0.2293684334,-0.4528870881,0.1833265275,-0.1798184961,-0.0539080426,-0.5003442764,-0.0958311334,-0.0321056657,-0.279514581,0.3314752281,-0.0826104432,-0.4040456712,0.0106442422,-0.4854611456,0.6610722542,0.2718071938,0.0265723281,0.5393341184,-0.1853725463,0.0325600207,0.0023206954,0.4078962505,-0.2625854313,-0.347853452,-0.1125153676,0.0016585444,-0.2887254655,-0.0932437405,-0.3349478841,0.1229617,0.3708439469,0.3897208571,-0.023838954,0.1330492049,0.2525302768,0.4104085267,-0.1656053662,0.0772895813,0.1077203676,-0.0894365758,-0.0334252678,0.0357292667,0.0311306119,-0.0031652609,0.447553426,-0.2872542441,-0.1344518512,0.1289482117,0.2831683755,0.1002388671,0.3788496852,0.2231844366,0.0161690973,-0.1264451295,0.0592145026,0.3904907107,0.3411543965,-0.2531392574,-0.2315784544,-0.205869481,-0.512368679,0.6896954775,-0.2535876036,0.2955570817,0.3373090029,-0.3510529399,0.0188250057,-0.340611428,0.1701037139,0.1007559597,0.0842803568,-0.3459071219,-0.590956986,0.0531342253,-0.1161700711,-0.0468521193,0.1868210137,-0.3011867404,-0.1200276762,0.4250341654,0.4480885863,0.5949056745,0.3401128948,0.0870731026,-0.0557735302,-0.0333279334,0.2986180186,-0.0762587264,0.1441027969,-0.0258995667,-0.0750056282,0.0151064256,0.0921250731,-0.0761178881,0.3233535588,-0.0840922073,0.1574980915,-0.1557340771,0.0077821971,-0.11164736,0.0539859273,0.2207533419,-0.1985207051,-0.0210072435,0.2352899611,0.0335564241,-0.2614860237,-0.1723607481,-0.0704653561,-0.1075227484,0.0303695817,-0.0626307353,-0.0471483245,-0.1693163365,0.0062612398,0.3074277937,-0.2784874439,-0.2213053256,-0.1035398245,-0.0684420317,0.4430475533,0.0471433178,0.0354176648,0.2193325758,0.3094555438,-0.0641488805,-0.0689174682,0.269954145,0.151663512,-0.1144075394,0.2225407362,0.1302426308,-0.474619031,-0.3539478481,0.0114165097,-0.001012176,-0.4632718563,-0.1342921108,0.4919712543,-0.0073934011,-0.1561767608,0.1699604988,0.2109733969,-0.414381057,0.3574540019,0.1502197385,-0.1569999307,-0.0096387649,0.0831009895,0.1553248912,-0.2740044594,0.4114089012,-0.0407930464,0.0334749818,-0.2200433016,-0.224653542,0.1850557625,0.0983281136,-0.1128589734,0.0988502949,0.1010632366,0.115933165,-0.0018398365,0.1043530554,0.2066406012,-0.1912954152,-0.2985829413,-0.5659019947,0.3158526719,0.3864011467,0.2673149407,0.0548994057,0.1596603245,-0.2889725864,0.5095253587,-0.3953597844,0.0165907592,0.3821215332,0.2297488302,0.0415968411,-0.0119093899,-0.0195809621,-0.0496753789,0.2320823669,-0.0433302671,-0.0343031585,-0.2969153821,-0.0451008119,0.0337536149,-0.0689404458,0.1060498282,-0.0016058943,0.1331327856,0.0911701918,-0.4856988788,0.0786508024,-0.0235218797,-0.0783713162,0.1418719143,0.3187411129,0.3381460011,-0.1551587731,0.2361300141,-0.0927872211,-0.0269700084,0.0284729563,0.0829201564,0.1781572104,-0.089605324,0.1978159398,0.1371515393,0.2371111363,-0.0130118178,0.1360730827,0.1671831906,0.0137833357,-0.1930579692,0.2749155164,0.2787074745,0.0896210521,0.2042456269,0.2675207257,0.4182379246,-0.2221463472,0.1189200953,0.1931221336,0.0985130146,0.1722989529,0.2065149695,0.1616923511,0.0920293406,0.0197030436,-0.05659898,0.1204199046,-0.0441843159,-0.0036639534,0.4070051312,-0.1291939765,-0.2746009529,-0.1428507417,-0.1386274099,0.011245423,0.0462130569,0.050577715,0.4357422292,0.502040267,-0.2215156257,-0.0381134674,-0.182212308,0.1833444834,0.3443455994,0.2033536732,0.6369608641,0.4767535031,0.3960195482,0.0190374069,-0.3341273665,-0.0228690766,0.0455947109,0.0643717796,0.0233126711,-0.0558776855,-0.1345075369,-0.2604801357,-0.1258268654,0.1825301498,-0.317045629,0.093240574,-0.2400182933,-0.0976924449,-0.2019862235,0.135967195,0.0750067681,0.1225425079,-0.3271100223,-0.0051008938,0.0715208575,-0.0644684732,0.1521445364,0.2337120324,0.1777095944,0.034566287,-0.0313329659,0.2604059577,-0.2468584627,-0.0641295239,-0.3413462937,-0.1032400802,-0.0292910784,-0.1420269907,0.1172940359,0.2845404148,-0.2158595026,0.2540856004,-0.0543985739,0.0278828721,-0.3220424354,0.1123365164,0.3016870618,-0.5159484148,-0.1086741835,-0.290923655,-0.2737604082,-0.4126060009,0.098928906,-0.3885514736,-0.043059241,0.1713320017,0.1535686255,-0.1431219429,0.5102573633,0.6492153406,0.1804955006,-0.0139034269,-0.2436192483,-0.3587182462,0.2236052901,-0.0885429233,-0.1090300009,-0.1453375071,-0.0088960305,0.3029865026,0.0262528639,0.2671393156,-0.0504154824,-0.3545337617,0.1882059574,-0.1807930022,0.1738956273,-0.1874708533,0.270785898,0.250341475,-0.0993700102,-0.1384289265,0.0170486141,0.1661048532,-0.1552551836,0.0161291864,0.0909141377,0.1035415307,0.482532084,0.3226444125,0.0563975573,-0.4423956275,-0.050533995,-0.3334934711,0.0036295746,-0.2995579839,0.2854177952,-0.0547495075,0.1813041121,0.0267792344,-0.1795478016,0.1609043479,0.2777335346,-0.0023224384,-0.129756555,-0.2278165966,0.2474313527,-0.2702032626,-0.2676488459,-0.1234763637,0.1761462688,0.0489290133,-0.0092597399,0.0110429926,-0.0646982715,0.0136621511,-0.1957703084,-0.1220627055,-0.5525932908,0.0290753935,0.0659212247,-0.0016190328,-0.380012244,-0.2454479784,0.368458271,0.102154918,-0.0117967008,0.4022865295,-0.1692766994,-0.2776994407,-0.1890702844,0.2403185517,0.0850915164,-0.0821243599,-0.256981492,-0.4197179973]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/365","title":"How to augment data ?","comments":"As we're working with arrow's columnar format we prefer to play with batches that are dictionaries instead of tuples.\r\nIf we have tuple it implies to re-format the data each time we want to write to arrow, which can lower the speed of map for example.\r\n\r\nIt's also a matter of coherence, as we don't want users to be confused whether they have to return dictionaries for some functions and tuples for others when they're doing batches.","body":"Is there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```","comment_length":77,"text":"How to augment data ?\nIs there any clean way to augment data ?\r\n\r\nFor now my work-around is to use batched map, like this :\r\n\r\n```python\r\ndef aug(samples):\r\n # Simply copy the existing data to have x2 amount of data\r\n for k, v in samples.items():\r\n samples[k].extend(v)\r\n return samples\r\n\r\ndataset = dataset.map(aug, batched=True)\r\n```\nAs we're working with arrow's columnar format we prefer to play with batches that are dictionaries instead of tuples.\r\nIf we have tuple it implies to re-format the data each time we want to write to arrow, which can lower the speed of map for example.\r\n\r\nIt's also a matter of coherence, as we don't want users to be confused whether they have to return dictionaries for some functions and tuples for others when they're doing batches.","embeddings":[-0.0449631922,0.0315054543,-0.2623604536,-0.0620328896,0.0712592602,0.1740825176,-0.0435555987,0.2856449783,0.2927056849,0.0688027442,-0.0452285558,0.2766702175,-0.0615255609,0.0991373286,0.0134356795,-0.2535323203,0.1473680437,0.1497779638,-0.3047791719,0.1889687628,-0.1046198532,-0.381380111,-0.0792351738,-0.4454675913,-0.1265061647,-0.2560842037,-0.1376860291,-0.2588270307,-0.0291532669,-0.4197904766,-0.1189498752,-0.1064741984,0.0321356878,0.2190885246,-0.0001070394,-0.0777587593,0.3182469606,0.0758765265,-0.3291758299,0.2881889045,0.0744246617,-0.5883304477,0.1895926893,-0.1483372748,0.0005680347,-0.566341579,-0.0038558543,-0.2213453054,0.4242578149,-0.1464163959,0.2143065035,-0.0009614629,0.3605853021,0.1151259467,0.2499141544,0.3350800276,-0.0361365899,0.052888982,0.6350506544,0.0298000295,0.1589223295,0.2770664394,-0.0203181915,-0.0861726552,0.1291488558,0.1219946519,0.1845043749,-0.2338151336,0.0329687484,0.1512702405,0.4620940983,0.1598397195,-0.5492655039,-0.1295826137,0.0587237403,-0.0840115994,-0.282930553,-0.168948248,0.1015856639,0.0041412404,-0.0054871086,-0.16555278,-0.0878733993,0.4407989681,0.1486133486,0.4273617268,0.0645060316,0.0699813664,-0.0995412394,-0.2775317729,-0.0916133299,0.1496660262,-0.1602429003,0.410500139,-0.0713777617,-0.012885455,-0.2568856776,-0.3334091008,0.0220548455,0.1018839553,0.3761999905,0.1902011037,-0.1630193144,0.0982387736,0.2746863663,-0.2005943656,0.0142447203,0.2476577461,-0.1037527099,-0.2508893311,0.2756493986,-0.0562036149,-0.083276473,-0.0784710497,0.070635438,-0.1746521145,-0.2213665694,0.0792391673,0.004904232,0.0197239909,-0.0175558124,0.0744814202,-0.083247833,0.1453243792,-0.0534715354,0.2024701536,-0.1878627539,0.2492563576,0.303598851,-0.0639066547,-0.0625420734,0.0752420053,-0.2145076841,-0.2098854333,0.0772959292,0.3709008098,0.1855743974,-0.0100021968,0.0702611804,-0.0770587474,0.1091511697,-0.1297928244,0.4284990132,0.0761953667,0.0568466485,0.0317525901,-0.1038799137,0.0815597102,-0.0219879746,0.1219736412,-0.3230205178,-0.0418441221,-0.4821155965,0.1992630512,0.1621777713,-0.1413287818,-0.2607134283,0.2704459131,0.1766418368,-0.3363068998,0.1516435593,0.077542305,0.2024887651,-0.499091953,0.3427675366,0.0446180478,-0.417014271,0.0328524783,-0.1063544825,0.4234547913,0.2125740647,0.1393312961,-0.1623224169,0.0343467146,-0.0578219667,-0.0136518758,0.5873201489,-0.250216186,-0.274402678,-0.0086100046,0.1392573714,0.1536213756,-0.0692370757,0.2453819811,0.156303212,0.0249975994,0.2388848662,0.407787025,-0.0065613319,-0.1438524425,0.0002215149,-0.0887074545,0.1144869179,0.1512439698,-0.2663123906,-0.3306487501,0.304225564,-0.641069293,0.2132044137,-0.2068176419,0.4185775518,0.0880037025,0.2264771312,0.368098855,0.0227470975,-0.0148942368,-0.3838593066,-0.0206501614,-0.0569170974,-0.2303418219,-0.0467187762,-0.2595000863,0.005536763,0.0464920029,0.1018628851,-0.1722168475,0.2130829245,-0.2088260949,0.1028579921,-0.443461448,-0.0221212357,-0.2194159776,0.0620040558,0.0842285678,-0.0593720265,-0.0488721803,0.2604750991,-0.3671447635,-0.2766019702,0.3287090063,-0.2676295042,-0.0874426514,0.0066956072,0.3033412993,0.215750277,0.1685023308,0.1334202737,0.1584985852,-0.0353387333,0.0663862601,0.1497315764,0.2639790773,0.0453885049,-0.1659338921,-0.1654551923,0.2021573782,0.0454257466,0.3024059236,-0.2025537342,0.2945614457,0.353086412,0.0562304556,-0.1930909306,-0.0953233168,-0.2077766806,-0.0979158953,0.3189956248,-0.2392760664,-0.0645780861,0.0109414458,-0.2084864527,0.0258278344,0.0223982967,-0.192753911,-0.5307372212,-0.1057418287,0.186987862,0.0848148316,-0.1719257683,0.3399724662,0.1259012371,-0.2647234797,0.2518343925,0.0908784866,0.0777939558,0.0999108925,0.1562826037,0.0097636748,0.3862086535,0.0888933167,0.0986336917,0.0392765552,0.0426651128,0.0370285958,-0.0083182584,-0.1605688184,0.0001100852,-0.1803583801,0.0516251922,-0.6276811361,-0.0828135237,-0.3152197003,-0.111690335,0.1336806864,-0.3548227251,0.0682942495,0.0428554304,0.0047916509,0.0162837543,-0.3871847689,-0.0655821785,0.0567806736,0.0432574376,0.1481971294,-0.0240253117,-0.0951555669,0.4585327506,0.3018179238,0.276140213,-0.0924377665,-0.4389209747,-0.0779127553,-0.0016337711,-0.0425758101,0.3088806272,-0.0789653137,-0.2523444295,-0.1831150502,-0.01051745,-0.1569566578,0.0649125353,-0.1248631701,-0.0938959643,-0.0743303224,-0.2183416337,-0.0330914892,-0.3234651089,-0.3384042978,0.2372867614,0.2095003873,0.0771850571,-0.3131403029,0.3482193053,0.0356433578,-0.0219961144,-0.127111882,-0.1488686949,0.0672438145,0.5090066195,-0.2755546868,-0.0019017439,-0.0643548071,-0.424572736,0.2866921425,0.3556601405,-0.0719246715,-0.0999897122,-0.1816033125,0.079202868,-0.2332166433,0.2640167773,0.2581160963,0.0931416005,-0.175895229,-0.019752793,0.0001382893,0.1556055844,0.337949723,0.1088867709,-0.0048563862,-0.0421648324,-0.1415073425,0.2414645404,0.0386009589,-0.0504027344,0.0848740339,-0.2094838023,0.0364745297,-0.0294405092,0.0693500862,-0.1476153284,-0.1958067268,-0.1011284217,-0.0545562617,0.1600608379,-0.2452844977,0.0490127839,-0.0047936821,-0.4095490575,-0.0989037827,0.43972978,-0.588421464,0.2396432459,-0.2576619685,-0.1125564277,-0.4354302287,-0.1512355506,0.0298305526,-0.4478643537,0.2770803273,-0.1528020203,-0.1997331083,-0.0024504915,-0.5094037652,0.6173356771,0.3407272398,0.1416914016,0.5109146833,-0.2703467906,0.0388940498,-0.0224732868,0.1445499659,-0.3867683411,-0.4822348654,-0.1489629,0.0565244444,-0.5446115732,0.0453961045,-0.4296467006,0.0789373145,0.1753780544,0.3766127527,-0.1923593581,0.0562489517,0.1564025879,0.2604751289,-0.1554168463,0.1205745488,0.1243771911,0.0046462878,-0.1029449999,0.0679115877,0.1699332595,-0.0474389829,0.5780174136,-0.2204655707,-0.2604865432,0.1671956778,0.1181892455,0.120951429,0.4411543012,0.1597186923,-0.1065880284,-0.0680274069,0.1769419909,0.2869966924,0.4402089417,-0.1363719702,-0.1719540209,-0.2713153362,-0.4779589474,0.7552831769,-0.1796367615,0.249085933,0.502630055,-0.4150941372,-0.0819967538,-0.1827347875,0.0427164175,0.2052186877,0.1293452978,-0.2690243721,-0.5675098896,0.1869134009,0.0669860914,-0.111222297,0.4496652484,0.0426340587,-0.0253161304,0.5850726366,0.4780203402,0.7165122032,0.3173341751,0.1846252978,0.0463653728,-0.3272405267,0.4004529119,0.0701652467,0.0808078572,-0.0547946952,0.1025410071,0.0009730373,0.0179703813,-0.1384094357,0.2876406908,-0.1518889964,0.15002729,-0.2367848009,0.0100839939,-0.1330455393,0.0580248013,0.1733932048,-0.2823072374,0.0491811559,0.1456318796,0.0023430106,-0.4167021215,-0.1246458218,-0.0805624649,-0.2426794022,-0.0545363389,-0.1380874068,-0.0321417116,-0.2830956876,0.0215363372,0.3141749799,-0.3613505065,-0.0451421626,-0.0145278201,-0.1150659993,0.3307726681,0.0643189326,0.0906220451,0.2942405641,0.1756629199,-0.1249152049,-0.0758656859,0.2506099641,0.0968923569,-0.0730134472,0.2433031946,-0.0332548879,-0.5641134977,-0.2155822068,0.0455955788,0.0802175328,-0.2566101849,-0.1502013206,0.4297662973,-0.3009395301,-0.0832195655,0.1620203257,0.1880558431,-0.2602071464,0.3336820006,0.2605100572,-0.2356591672,0.0156477243,-0.0450060517,0.1454173774,-0.2462287396,0.5600543022,0.0048747319,-0.0744182095,-0.2285591066,-0.3298814297,0.3779739738,-0.1600486487,-0.1020270213,0.0367767625,-0.028991539,0.1916473061,0.0871227384,0.2979385853,0.3023321331,-0.3340016007,-0.2269848585,-0.553945601,0.4804137647,0.2971708477,0.212228328,-0.0700751767,-0.1323900074,-0.2415797412,0.5198348165,-0.3649608493,0.0541797504,0.3694392741,0.2712548971,-0.0078905802,0.0338670611,0.1295351982,0.0011188347,0.2268512398,-0.0118248528,0.0886302441,-0.2576602697,-0.0377601162,0.0666367039,-0.0706667677,-0.0237444658,-0.1256183684,-0.0510209799,0.0505984724,-0.4686259031,-0.0284208059,-0.1368898004,-0.23456949,0.0967760831,0.2378017157,0.2897175252,-0.1185136512,0.2620232105,-0.1800413132,-0.0217493083,0.0069461023,0.0462488979,0.286562711,-0.180545181,0.3799668252,0.1336243451,0.1907670498,-0.0546638817,0.0108199064,0.0726343393,-0.0574465916,-0.2792015076,0.2909563482,0.2643674612,0.1183542833,0.10864795,0.2221930474,0.3442614079,-0.2363184243,-0.0017850385,-0.1496377289,0.1602110565,0.232027933,0.2347549796,0.1550633162,-0.0138556892,-0.0915753692,-0.051003661,-0.2042002082,-0.259450078,-0.1191178635,0.4184331894,-0.0561250374,-0.3389959633,-0.1105778143,-0.0296589751,0.0458420888,-0.007733651,0.0287055448,0.4061870277,0.6985001564,-0.1252538413,-0.1514202058,-0.0195997097,0.062420927,0.3148455024,0.1865274459,0.5614483356,0.5762034059,0.3474791944,0.2870393395,-0.3146401644,-0.0132955331,-0.1193246245,0.0299683288,-0.0644759014,-0.11116153,-0.0645016953,-0.3055882156,-0.1894479096,0.1012379006,-0.1942662597,0.3417070508,-0.1086990386,0.0547950566,-0.2436426431,0.2484984994,0.1756338626,0.3531354666,-0.3195340037,0.0233139824,0.0130953137,-0.0288181845,0.0325077251,0.3075966239,0.178349942,0.0284825545,-0.1392765194,0.2700147629,-0.1918343306,-0.1115117744,-0.1392233968,0.1388920546,0.1343429089,-0.2207656503,0.2025390118,0.2486941963,-0.1949062794,0.2103411704,-0.1208629832,0.0523634627,-0.196449101,0.2150077671,0.2053636163,-0.5519217849,-0.1238853484,-0.1983913928,-0.2035439909,-0.3652025163,0.2021933347,-0.4419761896,-0.1564689279,0.0877138525,0.1127539203,-0.0868226662,0.6534002423,0.6501076221,0.4077478945,-0.0029830255,-0.4474177957,-0.260594368,0.1022815481,-0.3121052086,-0.147126317,-0.0733870864,0.0649353564,0.3865477443,0.1107751504,0.385918051,0.07033889,-0.3434292972,0.2263901085,-0.3226324618,0.2810468674,0.1677543819,0.2737787068,0.1945687681,-0.1446954906,-0.1030599624,0.1704747826,0.1538925469,-0.0958024114,0.0370049365,0.1521031559,0.1813178062,0.5217484236,0.2050193697,0.0396761671,-0.3355919719,-0.0860084221,-0.3407739103,-0.1025016606,-0.2674382925,0.3767380416,0.0196099468,0.2899918258,-0.0036170052,-0.0766386539,0.2338164151,0.0461143777,-0.1184457988,-0.0311349612,-0.1199419051,0.1658708304,-0.1964733899,-0.2484062761,-0.1983761936,0.2559405565,0.0955991298,-0.0960787758,0.0053158281,-0.1152263358,0.0523725115,-0.0773533732,-0.1800580174,-0.2604097426,0.0013356969,0.2386951149,0.0844915733,-0.3445119262,-0.2247289121,0.2123771012,0.1343297511,-0.1316025853,0.2490274459,-0.1184342131,-0.1462399065,-0.1720625311,0.4722698629,0.0154967839,0.0196419749,-0.4694660306,-0.4412471056]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/362","title":"[dateset subset missing] xtreme paws-x","comments":"You're right, thanks for pointing it out. We will update it ","body":"I tried nlp.load_dataset('xtreme', 'PAWS-X.es') but get the value error\r\nIt turns out that the subset for Spanish is missing\r\nhttps:\/\/github.com\/google-research-datasets\/paws\/tree\/master\/pawsx","comment_length":11,"text":"[dateset subset missing] xtreme paws-x\nI tried nlp.load_dataset('xtreme', 'PAWS-X.es') but get the value error\r\nIt turns out that the subset for Spanish is missing\r\nhttps:\/\/github.com\/google-research-datasets\/paws\/tree\/master\/pawsx\nYou're right, thanks for pointing it out. We will update it ","embeddings":[-0.2723863423,0.0006952372,-0.0914846584,0.0489885025,0.1709388047,0.0924368724,0.0160519686,0.3171224892,0.152491793,0.1875384003,0.1667599976,0.2799213231,0.0049498309,0.123891443,0.0886844769,-0.3561535478,0.1757402867,0.2258462161,-0.2555732727,-0.2096619457,-0.2876420021,0.2521874011,-0.2408380657,0.035684295,-0.2994790077,-0.2202248871,-0.2171874046,0.2178255618,-0.2156783938,-0.4028117657,0.1486869901,-0.1411253959,0.2588767707,0.3559325635,-0.0001057746,-0.1943027228,0.3609939516,-0.0793708563,0.0186835714,-0.3070233762,-0.1794595718,-0.3881790042,-0.0674504787,-0.2644178271,-0.347207725,-0.3875034451,0.1691016853,-0.1587573886,0.1464087963,0.35717237,0.2652309537,0.4011282921,-0.0850666687,-0.3507474065,0.2515887618,-0.0938238949,-0.1439416409,0.1036625355,0.1457812935,-0.2500474155,0.2107236981,0.1748126149,0.3343430161,-0.2375557274,-0.0707044005,0.0697241649,0.3023766875,-0.1671793163,0.0662805662,0.0147637408,0.4198351502,0.0151392352,-0.1716321707,-0.0555245876,-0.040338818,-0.4136557877,0.2165343314,0.2597543597,0.1935590953,-0.0416914821,0.1158329844,0.0237940829,-0.0614887923,0.3454704583,-0.0012849572,0.8340978622,-0.0160531793,0.0465177819,0.0134799834,0.0393687934,0.4465917051,0.3594827056,-0.1187719703,0.2788745463,-0.2489586622,-0.1005548015,0.1722963601,0.3614239693,0.0867909566,0.2612136602,0.0452767946,0.0714267567,0.0965017751,0.0992661119,0.3548004329,0.1074724942,0.2129763067,0.2638505101,0.0047723721,-0.1295006722,0.0519993827,0.0678893253,-0.0429938845,-0.3853466213,-0.1893384308,0.2093623579,0.1108416095,-0.3272904158,-0.0915303528,-0.0615351349,-0.0134326508,-0.2365424037,-0.145479992,0.371691823,-0.0059782006,0.2843196988,-0.1703707427,0.3329637945,-0.141915381,-0.2906172276,-0.4568061233,0.2650963664,-0.3631557822,-0.3445568383,-0.0637470856,-0.1220751554,0.5182842016,-0.0537651666,-0.0423845351,-0.0695587173,-0.0256239101,-0.1389463693,-0.2085812539,0.0398057587,0.3005495369,0.031883873,0.1171852499,0.0337006077,-0.2295834422,0.0568278655,-0.1184246987,-0.0004304358,-0.199227497,0.2759284973,-0.225796625,-0.2709468007,-0.4000444412,0.2640072703,0.0144178485,0.1035412252,-0.0383229516,-0.0752925798,-0.4801713228,-0.0176260546,0.0231898874,0.0518201627,-0.0222174097,-0.2858150899,0.1711542159,-0.089998357,0.1610951722,0.2961949706,-0.0898867026,0.1831859052,-0.224966988,0.4528551698,0.2652120292,-0.1563598514,-0.0726549849,0.07778579,-0.0525892712,-0.4404550791,-0.0760117322,-0.0331357755,0.1407153904,-0.0431301296,-0.1479643285,0.3599586785,0.0035373424,0.0438944288,-0.3143992722,-0.0816966593,0.4235168099,0.1608241946,0.3229604661,-0.0041045537,0.273494035,-0.0237542335,0.1801989377,-0.0450649187,-0.0044750548,0.489679873,0.0051980563,0.0479346327,-0.0042402754,-0.1776257902,0.0711791962,0.1290178448,-0.2990307808,0.2884264588,0.3171682358,-0.046030391,-0.4836602211,-0.0592057966,-0.1138190478,-0.0294685978,0.1814825088,0.1707495004,0.2096120268,0.3443465233,-0.0054981024,0.1254927069,-0.2069069445,-0.0992371291,-0.2536619306,0.4569380879,-0.1901187152,-0.1612103879,0.2241331041,0.267823875,0.2943041623,-0.1026264802,-0.098049745,0.1237218082,-0.1481331438,0.3015097082,0.2574957013,-0.2728233933,-0.0452421121,-0.3814653754,-0.0127096809,0.1394991577,0.0775379986,0.0548887961,-0.0383100025,0.2478731126,0.3042419255,-0.2222872525,-0.0910751298,0.0563354604,0.2430194914,-0.212976113,-0.0298911408,-0.1627577543,0.4264274538,0.1706855893,-0.0138769839,-0.0213246252,-0.4169247746,-0.0370632932,0.6056559682,-0.1193823069,0.078567706,0.1529786289,0.0715116784,-0.3066865802,-0.0058337557,0.2398807406,0.3572744727,0.3439065218,0.2029372305,0.0787528679,-0.2068259567,-0.0654425472,0.1056444719,0.0519415364,-0.0732401907,-0.1383994371,0.1999257952,-0.0638126433,-0.3856398165,0.1349423975,0.2144734561,0.3548604846,-0.2816807628,-0.0975524262,-0.2479280829,-0.2998086512,-0.0016009553,0.0826931,-0.2456462085,-0.3243044019,0.15763928,-0.0365763716,0.0417454652,0.3652068079,-0.1367705911,0.0056618615,-0.0049396721,0.0401026011,-0.0350030363,-0.2318172008,-0.3527548313,0.1305429339,0.0905396193,0.3170442879,0.4059418738,-0.368400991,0.002693407,-0.3611170053,-0.2993424237,0.1045932323,-0.2507089376,0.2684367895,0.0087427739,0.1266117692,-0.3201704919,0.2132326216,0.2397200018,-0.2458877563,-0.1926632077,-0.1085881963,-0.0671733692,-0.0918410346,-0.0606154837,-0.5278220177,-0.37084952,-0.367000699,-0.065754585,0.0985037833,0.1006365046,0.0856583118,-0.0600687899,-0.0719454512,-0.1658993065,-0.1063372642,-0.4470635056,-0.32631585,0.2928777039,-0.2266394794,-0.581495285,0.1143617108,-0.0506657436,0.4476826191,0.0005441874,-0.3492148519,-0.0603530705,-0.0028152422,-0.0314194895,-0.1303566992,0.1001973301,0.5534501672,-0.2979641557,-0.1712639332,-0.1251854002,-0.2177530974,0.1510343552,0.2973302603,0.3211088479,-0.1410774738,0.1151762605,-0.0631577373,0.2846742868,-0.0316478312,0.053291738,0.6840032339,-0.244512856,0.3231371939,-0.0995489806,-0.0442074463,0.101444602,0.1382006258,0.1466751248,0.3989432752,0.0411096886,-0.4312523007,-0.2597790658,-0.5884571671,-0.1512645632,-0.0286318623,0.0602216609,-0.0078265574,0.4727508724,0.1639181674,0.0782341436,-0.1649162918,-0.1441238225,-0.074662827,0.0168302134,-0.0861935616,0.3197941482,-0.6754413843,0.0850926265,-0.3543172777,0.215866372,0.0555521064,0.3087714314,-0.1962283552,0.1455997527,0.0288745258,-0.112495698,0.1755921245,-0.1998397112,0.1459698975,0.1682378352,-0.0794092268,-0.1936047971,0.0825845152,-0.1728276014,-0.0158680659,0.0505344272,0.2659788728,-0.1948968917,-0.3792351186,-0.0886875764,0.312982142,-0.1598508805,-0.0770651177,-0.3286899328,-0.4161555767,-0.1963049918,0.1169769466,0.3321980834,0.1837708056,0.1748734117,0.0634647533,-0.1247034371,-0.1186424717,0.0842481032,0.2441332042,0.4113840163,0.0400702655,0.0839463547,-0.0310169086,-0.0121624917,-0.0585346185,0.5683514476,-0.0980489179,0.0977978781,0.2395068705,-0.2971134186,0.3363630474,0.1620427519,-0.0051534344,0.1323476732,0.0393813998,0.0826630816,-0.0266342964,0.3308772147,0.3699972928,0.2269187421,-0.2418851405,-0.4228973091,0.252505213,0.1919614673,0.0552855432,0.2002763748,0.1063241884,-0.206462115,0.2370548099,-0.0657967031,0.7613944411,0.1651649475,0.1864792407,0.4167962968,-0.0718524605,0.2779851854,0.2717631459,-0.0631320104,-0.2475988418,-0.044186987,-0.1071983129,0.0442002229,0.1917046905,0.2141421586,-0.2015154511,0.6288459897,0.204745844,-0.1525346786,0.225449726,0.1988147348,-0.0142929768,-0.1343834549,-0.5944886208,0.2069788873,-0.1462178826,0.0633764043,-0.0769869015,-0.146280542,-0.2195197046,-0.4036499262,-0.0737463385,0.1529299021,-0.0580739006,0.0418180525,0.0385475121,-0.3990470171,0.191177845,0.0397943147,0.2170634419,-0.0582335852,-0.1818310469,0.304746151,-0.0075841625,-0.1739197224,0.1500828713,0.1198992208,0.3171972334,-0.2586765587,-0.1875372827,0.0114493296,-0.2032788396,-0.0345216841,-0.2465624958,0.1595947444,-0.3737510443,-0.1959358603,-0.4727874696,0.0860461891,0.0474974588,-0.3153258264,0.2017521262,0.0689901933,0.020989297,-0.0169924349,0.3045969307,-0.2383452952,-0.138642326,0.2091613263,0.0371555015,0.0471446849,0.4808010161,0.2420149446,-0.0281382743,-0.2966547012,-0.1257820278,0.1354923546,-0.5430535674,-0.0528952181,-0.05477589,-0.0961024687,0.0428737551,0.3098929524,-0.1373122782,-0.1078467518,0.2535555661,-0.4714662433,-0.2278929204,0.1159134954,-0.1796811819,-0.028868841,0.201036647,0.4700260162,0.1158104315,-0.2557653487,-0.381455034,0.0190775655,-0.4054698348,0.3051436245,0.1177060455,-0.0948497131,0.2681597769,-0.0620010048,0.2017674893,0.1235170811,-0.0662649423,-0.284871012,-0.2188810557,0.0538617,0.009668733,0.1855555177,0.2400459051,-0.0620525442,-0.313400656,0.026548326,-0.2233751565,0.1273055822,-0.0521082804,0.2594484091,-0.0408790596,-0.1122523695,0.1122565642,0.0057252422,-0.0920172408,-0.0333662964,-0.1260653287,0.2747962475,0.0690020472,-0.0233717635,-0.3808797598,-0.0505193844,-0.0284922998,-0.0545280613,0.0226160791,-0.198433131,0.0622427687,0.208148554,0.3568871915,-0.1056776494,-0.2642483115,0.0998723358,0.350250423,0.2651237547,-0.6771472692,0.036079254,-0.0227375347,0.0212935582,0.2305302471,0.2213765383,0.3519206345,0.1105236858,-0.0410683379,0.0564380251,0.3963054419,-0.2074804455,0.0175850671,0.195000574,-0.0854920521,0.1242547482,0.1230957359,0.2380729765,-0.0440188088,0.4058528244,0.1298917085,0.4283076823,0.2533702552,0.2649691999,-0.1244138628,-0.2897444963,-0.0661484301,0.3276028335,0.3186393678,0.0472314283,0.0619567744,0.703791976,0.014499859,-0.1319578886,-0.0655257329,0.3119128048,-0.1082535461,0.072444208,-0.0833041295,0.011717286,-0.3387099802,-0.0635717735,0.0005893185,0.0293800049,-0.0602076091,0.0804828182,0.0039379429,-0.284201026,0.0698169917,0.1259331852,-0.1314823478,-0.3188655972,0.5945231318,-0.1218502447,-0.4201043546,0.1584955007,0.2582613826,0.5188597441,0.1071155444,0.0488681793,-0.0437292941,0.0018212389,-0.13461712,0.0806843638,0.218795985,0.2845212221,0.3679182827,0.3577747941,0.1634309292,-0.2513649166,0.2575922906,0.0484451428,0.226793468,-0.0682457536,0.2583428621,-0.0113768457,0.10259974,-0.3965281546,-0.1172337383,-0.5254933834,-0.0007473679,0.077161774,0.3422869146,0.1111289337,-0.4354730844,0.1094709933,-0.0964957848,0.3991636634,0.2502362728,0.3670509756,-0.2507531047,-0.4560555816,-0.7210580707,0.0890895724,-0.3774915338,-0.2002441138,0.0423678346,-0.1226584539,0.2631544173,-0.0724227205,0.3123472035,0.3638033867,-0.0798251927,-0.1145439669,-0.3845595121,-0.0688848644,-0.2045748681,0.366907835,0.1376626194,-0.1243307889,-0.0605729707,0.0328891799,0.1181330234,-0.3173041046,-0.2135602087,-0.1606429219,0.175759986,0.1687215865,0.075053595,0.4927922189,-0.2915099263,0.1042633131,-0.3180469573,-0.4224328995,-0.3445198834,0.2017444074,0.2859131992,0.329772681,-0.1275156587,-0.0638239086,-0.1427688599,0.1203714684,-0.0973110497,-0.1763219982,-0.0813697055,0.3280978203,0.0590958409,-0.0428595357,0.0983443856,0.1540744156,-0.022706436,-0.1645748913,-0.2568528354,-0.3355124891,0.4338591993,-0.2117649019,-0.2895965576,-0.0873526782,-0.1773522496,0.4782169163,-0.298173219,-0.3877659142,0.1563051939,0.1619605422,0.2428533435,-0.0613118298,0.1735099405,-0.098410137,-0.0226342063,0.0941088423,0.0798415616,0.1280599087,-0.2436033487,0.2776788175,-0.2082701474]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/361","title":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic","comments":"Hi, can you give a full self-contained example to reproduce this behavior?","body":"If I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?","comment_length":12,"text":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic\nIf I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?\nHi, can you give a full self-contained example to reproduce this behavior?","embeddings":[-0.0278753433,-0.4438011944,-0.0672562122,0.2349775285,0.1492943913,-0.247438997,0.0243676938,-0.3512237668,-0.0580411628,0.4765961468,0.0292001087,0.2727572024,0.0450993367,0.0568931773,-0.2647830844,0.0540780462,0.0710916668,0.0142234117,0.073901616,-0.2816468179,0.0538524799,0.0789456144,-0.2899128795,-0.1901162267,-0.1556191742,0.0510731377,-0.1686619669,-0.0937194005,-0.3361530304,-0.0801444873,0.0429904051,0.2987160981,-0.0704508275,0.5046932101,-0.0001208238,-0.4116142094,-0.0394376852,-0.182890147,0.3130660653,-0.5566269159,-0.2189091295,0.3871706426,0.0896749496,-0.1333379596,-0.1067069545,0.0927268639,-0.3590557575,0.0175039694,0.5081048012,0.1280786842,0.1317320168,-0.1578069031,-0.3599253893,-0.3842913508,-0.0521167628,0.0400119834,-0.1251857579,0.4521890581,-0.0263366606,0.1206827685,-0.031067241,0.0540378056,0.143203035,0.3617203236,0.4338986278,-0.1531135589,0.2439230382,-0.0465910211,0.0188784879,0.3572002947,0.2742506862,-0.1333668232,-0.0665942803,0.2570202947,0.2235364318,-0.3642040491,-0.0886247829,-0.0532382503,0.1581895947,-0.1119932532,-0.0107477903,0.3495724499,-0.0731922835,-0.1750465035,0.2426553518,-0.0778461099,-0.0084260134,0.0701486766,0.1234633029,0.4425570667,-0.0974235609,0.117695123,-0.4019389749,-0.1495050341,-0.5115230083,-0.0696679354,-0.02171297,0.397213608,0.0236001182,0.3255233467,0.5235888362,-0.1826651692,-0.1412472427,0.0391218029,-0.0449648313,0.3117986619,-0.1604887694,0.0199935585,0.0635393709,-0.0372557901,-0.1526964456,0.0043648127,0.3584748805,-0.4927637279,0.2992377877,0.2712788284,-0.1535445154,-0.3051620424,-0.6451265812,-0.0614442378,-0.4595505595,-0.2221877277,-0.0229110904,0.2110589594,-0.2730186284,0.0450726971,-0.2882178724,0.2446187437,-0.1312377602,-0.0893824399,-0.3016633391,0.0918281525,-0.536886394,-0.034471944,-0.3452130258,0.1876329482,0.1931496114,0.4384371936,-0.0053053112,-0.3554160595,0.0522271805,-0.4527687132,0.0392010361,-0.052713871,-0.0540801659,0.0354993306,0.1066779718,-0.2032135725,-0.1234376281,0.0289496519,-0.1562673748,0.2646735907,0.4462097287,0.0187637545,-0.1036880314,0.1235219985,-0.1334896237,0.167426154,-0.3082198203,-0.0053208852,0.2123286426,-0.1708208621,-0.3235839605,-0.3063796759,-0.0104165841,0.2050487697,-0.077727221,0.1582280248,0.6404648423,0.1646484137,0.2139376402,0.5318229795,0.0763246045,0.1109487265,-0.2119579166,-0.5803598166,0.4420110881,-0.4384528697,0.1029299349,-0.0256284717,-0.0985038057,0.0085482048,0.1591248512,0.1901966184,-0.2429219931,-0.0182144325,0.1403371841,-0.2131678015,0.0257534087,-0.2438972443,0.0117116561,0.0236346032,0.3204472363,-0.356628567,0.1335382313,0.2732741237,-0.009469863,-0.1824552864,0.3077844381,0.2059743106,-0.3304471672,0.0381290875,0.2256657481,0.3567247093,-0.0505654737,-0.176347971,0.147156179,0.1658817977,-0.4844454229,-0.2280698568,0.5902767777,-0.2040706426,-0.2241917402,-0.1361300498,0.0409147292,-0.1908532232,0.0793899745,0.3256103694,0.2410688996,0.288743794,0.0404367037,0.2066932172,-0.1851624697,-0.1461693794,-0.1823547333,-0.095960483,0.0368529558,-0.1462370604,0.0473125987,0.1920402646,0.2985054851,0.0113861971,-0.1329713017,0.3146069646,0.5357373357,-0.0223019272,0.1094002873,0.2514232695,0.1362713873,0.1997799575,-0.0800659731,0.4268021286,0.1377360076,-0.118948929,0.073786892,0.2941905558,-0.4763468504,0.1139522716,-0.0925862417,-0.0099028051,0.0672638118,0.0893095061,-0.153548643,-0.0787230283,0.4158908725,0.0142983198,0.296697408,0.0357226357,-0.4013650417,-0.1745181978,0.0853948593,0.250187248,0.0222196393,-0.1845514923,0.1102074087,0.0607792176,-0.0608187355,0.1667534709,0.2515928745,0.0216481909,-0.0909739956,0.0154052759,-0.3701698482,-0.3327532113,-0.2194212973,-0.0903499722,0.2797894478,0.3556670249,0.2816880047,0.262473911,-0.2135902047,-0.4265472591,-0.0865504816,0.0459716879,0.0435453765,0.315795213,-0.399671644,0.3920629919,-0.0689584687,-0.3222067356,-0.0597521216,-0.0599656589,-0.073042199,0.3514169157,0.3280570805,-0.1244822145,0.5166563988,0.1138474122,-0.0399118625,0.2945661247,-0.0485734269,-0.1258661151,0.0386052169,0.0257935897,-0.1465658098,0.0065416307,0.2259476781,-0.2792923152,-0.0432695448,0.0098606898,-0.2080435455,-0.0780397728,0.1157752275,0.2343508899,0.0549341813,-0.4795146585,-0.2191681415,0.4642523825,0.3518336713,-0.1074974984,-0.1560204029,0.10075479,-0.160857752,0.1215035617,0.0654905438,0.132737726,0.0725313872,0.0395128988,0.4967183471,0.0078664003,-0.0281502046,0.3500496149,0.1243591681,-0.0917818025,0.165153712,0.322995007,-0.3300295174,-0.1880920082,0.1278525144,-0.3182552457,-0.3155886829,0.0329898782,0.26559183,0.0528682806,0.0723185614,-0.1860859245,-0.5460278392,-0.0096511347,-0.158655405,-0.2991524339,-0.2407806069,0.5065968633,-0.2663873732,-0.1301956028,-0.0749636441,0.1618080437,0.1346945167,0.0740804598,0.3048796058,-0.2567648888,0.1781851947,-0.0916606411,0.0357001051,0.4478002191,-0.0805075467,0.0761221796,-0.0278219748,0.2398611903,-0.4224629104,0.0898834541,0.2956893742,0.0149230817,-0.2667980492,0.3048688471,-0.1308186799,0.642741859,-0.0945243984,0.1017924473,0.1823842376,-0.0325168371,-0.2033628076,-0.398306191,-0.1589248031,0.0286576506,0.1381332129,0.1641624719,0.0750601292,0.0777146965,0.1084052622,0.2182200104,0.0602447614,0.0526296198,0.0202185251,-0.2460401356,-0.0002200854,-0.1759421676,0.0635043681,-0.1841225475,0.1807703078,0.2337255627,-0.05302459,0.567276597,-0.3618338704,0.2181439847,0.0172554888,-0.5400359035,-0.2354649305,-0.2948061228,-0.3621315062,0.0859689713,0.1945393384,0.1172304228,0.1302436739,-0.1136598364,-0.100186117,0.0303575825,-0.0727081001,-0.1632770449,-0.5953788757,0.1679608375,-0.0046692463,0.1351539791,0.0893698335,0.1177455932,-0.1546327472,-0.1843206286,-0.3244767189,-0.039995499,0.4181532562,0.1380161196,0.1079841778,-0.0182667971,-0.1873029023,0.0368927605,-0.1127714813,-0.2453712225,-0.0265838038,0.2155593187,-0.5710961819,0.1864033192,-0.1305015385,-0.0199342594,0.2239757925,-0.0803633407,-0.0517323986,-0.6463707685,0.2968768775,-0.4156374037,-0.141059041,0.4882903099,0.0009970749,-0.0269100592,-0.2525211275,0.1150972247,0.1468362361,-0.4094544947,0.2718222439,-0.2731720209,0.070834212,0.4075294733,0.3400958478,0.6679777503,-0.1354468018,-0.085116379,0.1301268339,-0.1540392339,0.2074337006,-0.1922009438,-0.0416799262,-0.2609692514,-0.3509344161,-0.0768148154,-0.0713159218,0.1558678001,-0.0447768457,-0.0581761412,0.3362097442,-0.1677242517,0.2320861071,-0.0925908238,-0.1791812629,0.3554528952,0.2699898183,0.4027711749,0.0864810646,0.1959271729,0.3023408353,-0.2253949791,0.1571039706,-0.2491638362,0.112015903,-0.2702568173,-0.1889420897,-0.3117049634,0.1350363046,-0.0258314405,-0.4638655782,-0.0761787072,0.8267238736,0.6647475958,0.2031361312,-0.0141458577,0.2898457348,0.0751218125,0.0856389552,-0.3115056455,-0.2719309926,-0.1711216122,0.0774807259,-0.300006032,0.0007177666,0.2084453106,-0.2381766886,-0.3873626888,0.2169073075,-0.2879522443,0.2068847269,0.0182511397,0.014254922,-0.0132383741,-0.1319406927,0.0562449209,0.0705324262,0.0748895854,0.2811674178,0.0331885368,-0.4102781415,-0.1308422983,0.3591259718,-0.0451259091,0.1156591251,-0.2565909922,-0.3242981136,0.1171321645,-0.2658929527,-0.1963142753,-0.2819597423,-0.1215733141,0.1786253303,-0.0699002072,-0.0425633565,0.2719117105,0.5846372843,0.1869547218,0.2781018615,0.2172236741,-0.3259078562,0.182635814,-0.043282453,0.2731642425,-0.0178167447,-0.2893386781,0.1878705472,0.0562373325,0.4470780194,-0.2470200658,-0.1740149856,-0.0701262206,0.0517432801,-0.590626955,-0.2060349435,-0.204322204,0.2285388857,-0.0372772887,0.1530370861,-0.259976238,-0.0955786109,-0.276168108,0.1688641906,-0.0331848077,0.540381968,-0.1583297253,0.0963397771,-0.0552125983,-0.2159932256,0.1517035365,0.3743793666,0.2593429685,0.04505191,-0.0362332538,0.3332433701,-0.093290925,-0.1041305214,0.1177133918,0.0635356307,-0.128960222,-0.2341127545,-0.0919272378,-0.0826780349,-0.6418210268,-0.1670224518,-0.0421159528,-0.1729538143,-0.2072566599,0.1563063115,0.4606195092,-0.23602359,0.3007846475,0.1306436211,0.0047323448,-0.1959394515,-0.0255996641,0.080881156,-0.2319694906,0.2392216623,-0.2360356897,0.3392282724,0.1007023081,0.3833017349,0.4501611888,-0.1195375398,-0.047797814,-0.1847408861,-0.0234115366,-0.2794835269,0.197025165,-0.2066393197,0.0740245953,0.1018618196,0.6492866874,0.0169332922,0.0056448262,0.3732789159,0.163542971,0.1898635775,-0.2665962577,0.1109392643,-0.1217236668,0.0596726388,-0.0810902938,0.4697394073,0.0759622231,-0.1805046499,-0.1550807506,0.1215929464,-0.3016766012,-0.2199709564,-0.421296984,-0.093772836,-0.0313563123,-0.2432367653,-0.2117754966,-0.1190543547,-0.1191827729,0.0405582003,0.0614275075,0.4128224552,0.3896722198,0.0722330511,0.1582189649,-0.1866866201,-0.1302544326,-0.0842262879,0.4027122557,-0.1399715841,0.0459308252,0.0421299264,0.3508031368,0.0345716178,0.2600538433,0.1648775786,0.3020413518,-0.2351072133,0.1116927788,-0.2768465877,-0.0101024946,0.1959635913,0.247657463,0.1638347656,-0.3899145424,0.3288639188,0.0017222378,-0.1187012494,0.5564575791,0.1312116534,0.4048690796,-0.2384514064,0.4729175866,0.1366615444,0.3663713932,0.0491846874,0.3365818262,-0.3654001355,0.1031586006,-0.6393126845,-0.1756731272,-0.0554862656,-0.1889964938,0.0285119712,-0.0734442025,0.1124227867,-0.0101579269,0.299850136,-0.1565606296,-0.2420536876,-0.4881208241,0.2110588402,-0.2928185165,0.3253662288,-0.1236195564,0.3953752816,-0.4626609385,0.2079264671,0.2127076983,0.465498805,-0.1471383125,0.4687775671,-0.4542673826,0.2756391466,-0.3342673481,-0.1997590512,0.2328943759,0.1019798592,-0.0023281113,-0.0525343418,-0.0550017804,0.1114678904,-0.041842211,-0.0362814292,0.6551299095,0.5831336379,0.081098184,0.1635158658,-0.048700586,-0.207491979,0.0536227748,-0.1026380137,0.0999093801,0.2811009586,-0.0376623385,-0.1727637798,-0.1568658054,0.1608632356,0.1417271793,0.6789883971,0.1824771762,-0.0864614919,-0.6283499002,0.048129335,-0.0047420603,0.3107945919,-0.3685335815,0.4526985586,0.0406331904,0.3038793206,-0.1289441735,-0.0993000865,0.6079416871,0.0320950225,0.0379605331,-0.1789093614,0.3654030263,0.4084919393,-0.0614529699,-0.1274050474,0.1235738546,0.0763524249,0.2688087523,0.0873670653,0.2134679705,-0.3736586869,0.0807372406,-0.06516359,0.1801507324,-0.0593731143,-0.0318181515,0.3373695612,-0.2821269333]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/361","title":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic","comments":"> Hi, can you give a full self-contained example to reproduce this behavior?\r\n\r\nThere is a notebook in the post ;)","body":"If I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?","comment_length":21,"text":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic\nIf I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?\n> Hi, can you give a full self-contained example to reproduce this behavior?\r\n\r\nThere is a notebook in the post ;)","embeddings":[-0.0729421377,-0.4319974482,-0.0746613219,0.2472379804,0.1424586624,-0.2991015911,-0.0346640162,-0.3622231781,-0.0423693508,0.4704315066,0.0269525275,0.304450959,0.0302148387,0.0872765556,-0.2722607553,0.0752440989,0.0376889817,0.013303184,0.0488777757,-0.2914239764,0.0159806181,0.099529922,-0.2607537806,-0.1770699918,-0.1143433228,0.0565365963,-0.1556732804,-0.0681025684,-0.3433382809,-0.0870002061,0.0321296379,0.3313088417,-0.1052891612,0.493986696,-0.000117025,-0.4163560271,-0.0219289437,-0.1781865507,0.3075726628,-0.5136871338,-0.2497319132,0.3652491868,0.0798635185,-0.127443403,-0.1293844134,0.1144191846,-0.3657512963,0.01464028,0.4840198755,0.1275172383,0.1668632179,-0.1628429443,-0.3187301457,-0.3868788481,0.0139950886,0.059260942,-0.1181055233,0.4923393726,-0.0330632441,0.1233153045,-0.0663546845,0.0778925791,0.174616307,0.3484489322,0.4598203599,-0.1589015424,0.2443424612,-0.062895298,-0.0225948077,0.3563770354,0.2740356624,-0.1160635427,-0.1012605578,0.2822776139,0.2155551463,-0.3651972115,-0.0931866467,-0.073911868,0.1439158469,-0.1461345404,-0.0108550582,0.3827982545,-0.0761184171,-0.178085044,0.2129086852,-0.0751299411,-0.0184812434,0.0697376803,0.1373713762,0.3848738372,-0.1237272248,0.1314071715,-0.4110630751,-0.152568385,-0.5036755204,-0.0680446923,-0.0418667905,0.4301230311,0.0287697762,0.3245633841,0.5289825797,-0.1465084553,-0.1225256547,0.0420772284,-0.0674977675,0.3009240329,-0.1637224704,0.0034658327,0.0948855877,-0.0217084475,-0.161670506,0.0056091971,0.3751334846,-0.5000551343,0.2849849463,0.2503116727,-0.1803478599,-0.3356368542,-0.6347414851,-0.0664335564,-0.4162109792,-0.2036540061,-0.0150857586,0.2462238371,-0.271269381,0.0718201175,-0.2622617781,0.2142347842,-0.1250938922,-0.0586309284,-0.2955050468,0.0800914764,-0.5461654067,-0.0126219122,-0.3351665735,0.1911013871,0.1781329662,0.4183347821,0.0020132307,-0.3315204084,0.0598278269,-0.4487573206,0.072365433,-0.0718562976,-0.0791169256,0.0141067971,0.0821315646,-0.1971664429,-0.1339937598,-0.0057607917,-0.1412930787,0.2725314498,0.4450811148,0.0656629279,-0.0854326189,0.1210341156,-0.1398105472,0.1766982973,-0.317086935,-0.0083278306,0.2046132982,-0.1841770709,-0.3165988624,-0.3004755378,-0.001530298,0.2072269619,-0.0665686652,0.1484946162,0.6132826209,0.1436503977,0.2077416033,0.5073764324,0.0661428347,0.0879924595,-0.2220107317,-0.571229279,0.4453023374,-0.4446240366,0.0846231282,-0.0251812525,-0.0576002188,0.0062043541,0.1305315346,0.1785040796,-0.2394026369,-0.0235974994,0.1534713209,-0.2068719119,0.0482669435,-0.2135521472,-0.0194636304,0.0297448244,0.3167249858,-0.3030980229,0.1502765715,0.2283768803,-0.0114703728,-0.1718247235,0.3134987354,0.2069143802,-0.3694640398,0.0516028255,0.1938422918,0.3044385016,-0.0575993247,-0.2001896203,0.1571059376,0.1491965353,-0.5066497922,-0.1957545131,0.5925680995,-0.2029515654,-0.2326534986,-0.1298025697,0.0167076644,-0.2094645947,0.1213448197,0.3131524324,0.244681716,0.2755913138,0.0435582101,0.2283722609,-0.178737551,-0.1511544585,-0.1725236475,-0.1051983461,0.0018531756,-0.1402863562,0.0833922029,0.2186939716,0.3086818457,0.0139424242,-0.1252151877,0.3100949526,0.5252616405,-0.0263115745,0.1189501062,0.2512631714,0.1599371433,0.2055658996,-0.115612857,0.4238214493,0.1592493653,-0.1148899123,0.0598702095,0.3031254411,-0.4730485082,0.0715667158,-0.0767103732,-0.0230605267,0.0747991577,0.100366883,-0.1731087863,-0.0771300048,0.3988668025,0.0202211291,0.3070412874,0.0335895941,-0.4082301259,-0.1613553315,0.1065685079,0.2213362008,0.03647089,-0.1656696498,0.1028542891,0.0469974875,-0.0295289867,0.174823001,0.2265623212,0.0465116277,-0.0839708596,-0.0033859268,-0.3627881706,-0.3238306344,-0.2000466585,-0.1123357266,0.2540789247,0.356498301,0.2895348668,0.2467264384,-0.2504338324,-0.4256802499,-0.089273572,0.0223558098,0.0306770355,0.320001781,-0.3778673708,0.3741837442,-0.0508995801,-0.3276847005,-0.0824532956,-0.0896395445,-0.0475592837,0.330388993,0.3002762794,-0.1265237629,0.5133865476,0.1238656566,-0.0614105873,0.2800595164,-0.0947363824,-0.1490516961,0.0418749265,0.0614753775,-0.1346155852,0.0218229443,0.2588360906,-0.281209141,-0.0097511169,0.0195135772,-0.2017592639,-0.0430031568,0.1012569368,0.2257147729,0.0627168268,-0.4950441718,-0.2061870396,0.4547218382,0.3494848311,-0.1335418522,-0.1433508545,0.0927427411,-0.1561851501,0.1136120483,0.064965032,0.1115123779,0.0507349037,0.0077125579,0.4770751894,0.0406520516,-0.0323164463,0.3786587715,0.0874840021,-0.0692563355,0.1745936424,0.3019635379,-0.3619597554,-0.1703048944,0.1534349918,-0.3342146575,-0.3372672498,-0.0037971735,0.2548951507,0.0573144667,0.0719000846,-0.1600867659,-0.561626792,-0.0084271198,-0.1259204745,-0.2945589721,-0.2496599108,0.5192809105,-0.2608339489,-0.1672192514,-0.0561702922,0.1818184108,0.1151712909,0.0744687691,0.3241333365,-0.2724435329,0.14256455,-0.1437831819,0.0398101844,0.4241265953,-0.1063037962,0.0679801106,-0.0478787459,0.2415934354,-0.4286982417,0.0777276382,0.335377574,0.0047692019,-0.2590504587,0.2845087647,-0.1497368813,0.6038037539,-0.1270698756,0.0989090279,0.1798587739,-0.0424424782,-0.2053054422,-0.3841387928,-0.1888152212,0.0148056848,0.144737184,0.1506263018,0.0689899549,0.1150014773,0.1120623797,0.2014232725,0.0154255573,0.0449073054,0.0535325594,-0.2599620223,0.0311485529,-0.1688402295,0.0394633412,-0.1582559049,0.1542323977,0.1969505399,-0.1055848673,0.5462046266,-0.3483902812,0.2001517415,0.0214270279,-0.5407803655,-0.2420251966,-0.2875945568,-0.3703911006,0.0820829123,0.201350376,0.117059648,0.111678116,-0.102787964,-0.1129119545,0.0102802897,-0.0713436976,-0.1393746585,-0.5917127728,0.1660339385,0.0199531317,0.1845137328,0.0796697512,0.123135753,-0.1292865425,-0.1786277443,-0.3222105503,-0.0605731383,0.3907668591,0.2007409632,0.1498928219,-0.0240551606,-0.1971601099,0.0285025034,-0.1098369882,-0.2267870009,0.003175701,0.1989545524,-0.5485672355,0.2127310485,-0.1606103778,-0.0469370596,0.2214839906,-0.0933404714,-0.0975081846,-0.6738174558,0.3121762872,-0.4528850615,-0.1113971174,0.4746368229,0.0105750263,-0.0218153484,-0.2490186542,0.1142335087,0.1261955351,-0.4261642396,0.2862332463,-0.2654322684,0.0386849269,0.4411566556,0.3684294224,0.6875545382,-0.1333996505,-0.0636217892,0.1431773901,-0.1540894955,0.2247236967,-0.2082998157,-0.0265798364,-0.2640281618,-0.3276492953,-0.0482112467,-0.037744578,0.1220279336,-0.0377887189,-0.0722165257,0.3350603878,-0.1171616688,0.2474176735,-0.1141145453,-0.1422788501,0.3587920964,0.2268887013,0.3435049057,0.1264494956,0.1702564657,0.2904884517,-0.2271944135,0.1309096813,-0.2318707258,0.1126542687,-0.2502038777,-0.1527740359,-0.2988250256,0.1434079409,-0.0602940731,-0.4588409364,-0.107003808,0.796884656,0.6291093826,0.2359060794,-0.0231773183,0.2885487974,0.0672376826,0.1290567815,-0.2900576591,-0.2482478768,-0.1415725797,0.0503365807,-0.2829748094,0.0069707558,0.209493354,-0.2150886059,-0.3606831431,0.2045874149,-0.3005197942,0.1903464943,0.0454042256,0.0197649896,-0.0329486504,-0.1293003261,0.0893403664,0.0341837257,0.039887134,0.3000204861,0.070244588,-0.3948745728,-0.1391726285,0.3577367663,-0.066495806,0.1350867301,-0.2667106986,-0.3262555599,0.1047233045,-0.301920265,-0.1799211353,-0.2614293098,-0.1382792741,0.1824704409,-0.0564287007,-0.067134209,0.2789308131,0.6154510379,0.1856501698,0.2741474509,0.2410749197,-0.3023461401,0.1398621202,-0.0658835843,0.2619777322,0.0108386707,-0.2457009554,0.1602035761,0.0775496662,0.431709677,-0.2948765159,-0.1789622307,-0.1434402913,0.0270354729,-0.6010736227,-0.2214311361,-0.1617846191,0.2057250589,0.0024004981,0.1748354137,-0.2614100873,-0.128441751,-0.2560678124,0.1408542246,-0.0465693735,0.5226805806,-0.1612805277,0.0954839662,-0.0471086241,-0.2260114998,0.1570693254,0.3563844264,0.2539891601,0.0663863942,-0.0365470462,0.3187324107,-0.0808604807,-0.0889675245,0.1430629939,0.0676660836,-0.1192598194,-0.2296673656,-0.0938554257,-0.0814584568,-0.6070907116,-0.1350967437,-0.076391086,-0.1567072421,-0.1704012454,0.1404921561,0.4126685858,-0.2225686163,0.3149325848,0.1584102809,0.0117467772,-0.2432762533,-0.0636963695,0.1181384549,-0.2605796754,0.2265068591,-0.2292990685,0.3252803981,0.120074451,0.378754288,0.4270202219,-0.135965839,-0.0122592989,-0.1740570813,-0.0483452715,-0.2579518855,0.1864400655,-0.2596788406,0.0819971114,0.1029334217,0.6461299062,0.0085079726,0.0106310975,0.385386914,0.1674394011,0.2143237144,-0.2646903694,0.1232413501,-0.1209506914,0.0653526857,-0.0654734299,0.4651512504,0.0773720145,-0.1564274877,-0.1243256107,0.1073848084,-0.2876572609,-0.2792736292,-0.4090044498,-0.0871842057,-0.0265638269,-0.231995061,-0.2225252986,-0.1401963979,-0.1384878159,0.0466161147,0.0626090765,0.4171894491,0.382437408,0.0754114687,0.1739461124,-0.1623557061,-0.1418605298,-0.0882278904,0.4310386777,-0.1680330634,0.0816467777,0.0948937684,0.3489446044,0.0578178726,0.2674446404,0.1616434902,0.2786949277,-0.2399479598,0.1121961251,-0.2485887855,-0.0032199207,0.1796891242,0.2443201095,0.2036582232,-0.4013041854,0.3890455961,0.0393904597,-0.1451783627,0.5418245196,0.1110337302,0.3921962976,-0.2113577127,0.4320445359,0.1735882759,0.3634889126,-0.0001421461,0.3181552589,-0.3570294082,0.1055056825,-0.6689849496,-0.171386525,-0.0461662188,-0.2048518658,0.0514836945,-0.0484858453,0.1875864267,-0.0160366725,0.2722626626,-0.1575938165,-0.240581423,-0.4908547997,0.1989634037,-0.240904659,0.329377681,-0.1476686001,0.3971633613,-0.4126535356,0.2178213596,0.2038401663,0.4500792325,-0.126835376,0.4430060983,-0.4771740735,0.2912019193,-0.3749237657,-0.2154468,0.2228862941,0.0904399306,-0.0142699573,-0.0556677878,-0.0188208465,0.1260990649,-0.0845933855,-0.0322501287,0.6356549859,0.5892747641,0.0898814872,0.1651600301,-0.0616423339,-0.2201287299,0.0725355595,-0.1212240309,0.0976991281,0.2772809863,-0.0232315715,-0.1421556622,-0.1076292694,0.1572176218,0.0837742761,0.6822968125,0.1531055868,-0.0711385012,-0.5956967473,0.0260296389,0.0386705659,0.2575575411,-0.3959360123,0.4797008038,0.0283339377,0.3287315965,-0.1378914118,-0.1104477569,0.6011202931,0.0285401419,0.0503990799,-0.1647542715,0.3484908938,0.3952538371,-0.0488929972,-0.1494700462,0.1269877553,0.0697486848,0.2653723061,0.0776030943,0.1911769658,-0.3626302779,0.0843726173,-0.1014965698,0.1728831828,-0.0594970211,-0.0220189169,0.3068386018,-0.2685052752]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/361","title":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic","comments":"> If I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n> \r\n> Refer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n> \r\n> Example of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n> \r\n> > ['0.3350', '0.1470', '0.2329']\r\n> > ['0.3358', '0.1451', '0.2332']\r\n> \r\n> Why ROUGE is not deterministic ?\r\n\r\nThis is because of rouge's `BootstrapAggregator` that uses sampling to get confidence intervals (low, mid, high).\r\nYou can get deterministic scores per sentence pair by using\r\n```python\r\nscore = rouge.compute(rouge_types=[\"rouge1\", \"rouge2\", \"rougeL\"], use_agregator=False)\r\n```\r\nOr you can set numpy's random seed if you still want to use the aggregator.","body":"If I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?","comment_length":112,"text":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic\nIf I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?\n> If I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n> \r\n> Refer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n> \r\n> Example of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n> \r\n> > ['0.3350', '0.1470', '0.2329']\r\n> > ['0.3358', '0.1451', '0.2332']\r\n> \r\n> Why ROUGE is not deterministic ?\r\n\r\nThis is because of rouge's `BootstrapAggregator` that uses sampling to get confidence intervals (low, mid, high).\r\nYou can get deterministic scores per sentence pair by using\r\n```python\r\nscore = rouge.compute(rouge_types=[\"rouge1\", \"rouge2\", \"rougeL\"], use_agregator=False)\r\n```\r\nOr you can set numpy's random seed if you still want to use the aggregator.","embeddings":[-0.0629726052,-0.1342073232,-0.0796853825,0.0105740586,0.1754827946,-0.3666586876,0.0312740766,-0.2816157341,-0.3356750607,0.4649936557,-0.1453202069,0.3407977819,-0.0238616895,0.1076028943,-0.289493829,-0.0058361944,-0.1367871761,0.0883074626,0.0505806021,-0.2794827819,-0.0264891218,0.2689063847,-0.1446817368,-0.112200141,-0.2301799059,0.0439043455,-0.1309628934,0.1760581583,-0.4420460463,-0.1366114914,0.0006488205,0.5135985017,-0.1335638762,0.3453351557,-0.0001090062,-0.3026773632,-0.0053829737,-0.2721399665,0.0684326664,-0.5490511656,-0.0850553513,0.1678696126,-0.0363746472,-0.1538967341,-0.230387792,0.1567343622,-0.132400021,-0.2148977518,0.6561849713,0.3610711992,0.2037741095,-0.0275056381,-0.1488342732,-0.211245507,0.1747769415,-0.0872877389,-0.0166684855,0.5573641658,0.0314750075,0.218893528,-0.2464755923,0.1858105808,0.152736187,0.0676839948,0.3688113093,-0.1152091995,0.1754817069,-0.3695171475,-0.1286011636,0.3732657731,0.2760186791,-0.0519507863,-0.4068128765,0.1725014895,0.194771871,-0.2736438513,-0.1944601536,-0.0175777674,-0.0913497135,-0.0652742907,-0.0967804939,0.4438829124,-0.1332540959,-0.1383667886,0.1420723051,0.0766993016,-0.0088885538,-0.0005147148,0.3281321228,0.1229601204,-0.2245833427,0.2116307616,-0.209591195,0.0695717409,-0.4204727709,-0.1357257813,0.1251077354,0.2849909067,0.1111524031,0.4854750931,0.3948625326,-0.1147538722,-0.069045119,0.0454537049,0.0252647009,0.1961714029,0.0500495322,0.0383876227,0.141609937,-0.0029745968,-0.0259299502,0.0993311703,0.3126792312,-0.315680027,0.0414018482,0.2174888849,-0.1523083597,-0.1720807254,-0.5653681159,-0.1114857048,-0.478197515,-0.1255738437,-0.1535402685,0.3803393543,-0.2059172988,0.1545051336,-0.1938217133,0.1907310188,-0.1400554031,-0.1622923762,-0.2599925399,0.1628428847,-0.2660717368,-0.0745896026,-0.1121097133,0.0864350796,0.1294143945,0.1545472443,0.1475837678,-0.2458141446,0.0382168405,-0.4583488107,-0.082585834,-0.1328946948,-0.0913568065,-0.0753936544,0.0561134219,-0.2839280069,-0.1321660876,-0.12134251,-0.1262935251,0.1846183389,0.3253518939,0.1625862718,0.0098106805,-0.0861550793,0.0743729025,0.2732861042,-0.3147985637,0.1554822028,0.1502420306,-0.025994584,-0.3226451874,-0.2363419086,0.0831082612,0.2124174386,-0.002733697,-0.1042993292,0.583800137,0.1370000392,0.1727111787,0.5475632548,0.1629826874,0.1839769036,-0.0027143301,-0.3590961993,0.4775166214,-0.4854388237,0.0683306158,0.0602491684,-0.0288142804,0.1495360583,0.2133648396,0.1030385718,-0.0766566545,-0.0438036285,0.2398133576,0.1870018542,0.1334643811,-0.2636555731,-0.0209828485,0.0908618793,0.2314932048,-0.1147617996,0.0626321137,0.046259813,-0.2069050819,-0.3015894294,0.3498267829,-0.0088291103,-0.3083656728,0.2478799373,0.0876596794,0.1918273717,0.0869458988,-0.1816856414,0.2812576592,0.1035984084,-0.8122193217,0.2431687415,0.2177292556,-0.2853504419,-0.451205045,-0.130316183,0.1445597708,-0.3410880268,0.1980885714,0.1725724339,0.2733030021,0.2643563747,-0.0477089435,0.3236653805,-0.1935337782,-0.0180978086,-0.2536085844,-0.198676005,-0.050671231,-0.0811391324,0.2211104631,0.44662112,0.0897334665,0.0987754241,-0.0810163692,0.4132423699,0.4271882772,-0.0247583576,-0.0330941342,-0.0327458382,0.0773893893,0.1873423755,-0.218270883,0.5565975904,0.225764662,-0.0682012364,-0.1123133674,0.3855130076,-0.4762592614,0.0049159378,0.0653227419,0.061057847,0.1700087488,0.1440412253,-0.321056217,-0.0902716517,0.3082022369,-0.0289418381,0.6414523721,0.2357598692,-0.2955445051,-0.1263951808,0.0865039602,0.1140249968,0.1410741061,-0.0957917199,0.0867771804,-0.0590308756,-0.0750466809,0.0270542186,-0.0322183296,0.211177811,-0.0056918808,-0.0626981854,-0.2837332189,-0.2822376192,-0.0775493383,-0.138440147,0.288238734,0.1391821504,0.3164981306,0.1880495399,-0.2739830017,-0.5686125755,-0.3180910647,-0.036398001,-0.1358696967,0.1271980405,-0.2475135028,-0.0630130917,0.0696974322,-0.2970941067,-0.0642220899,-0.2096978575,0.1557428986,0.0802705437,0.0693588182,-0.0306591969,0.5392485261,0.2734598219,0.011913632,0.2975368202,-0.1017268375,-0.2644327879,0.0604698844,0.2002763003,-0.1333313137,-0.0249790568,0.1201741472,-0.4248000383,-0.0719186813,0.1428778619,-0.1266681552,0.0757164732,-0.0981830135,0.2909098566,0.1341505945,-0.3271768987,-0.3285121918,0.3417824507,0.2961697876,-0.0150603671,-0.1789551675,0.1072572097,0.0355433114,0.0203547925,0.0356846638,-0.119449228,-0.2463433295,-0.1701472104,0.5418916345,0.242584601,-0.0166938938,0.4519591928,0.0108007621,-0.0595282167,0.102975361,0.4948757589,-0.2613433599,-0.2291278839,0.3377356231,-0.2354305536,-0.4837260246,-0.2980839908,0.088767685,-0.0603663549,0.0093466165,0.1168930605,-0.3563518822,-0.0826548263,-0.0415779985,-0.2018310726,-0.0710330829,0.4534035325,-0.172801733,-0.2120850384,-0.0440846831,0.3471273482,0.1045517921,0.1094752923,0.1872497201,-0.2009674907,0.0722362697,0.0548028834,0.1274549663,0.4970481992,0.0237850584,0.0555972345,-0.0792819187,0.2268170714,-0.2797750235,0.052875489,0.2462792993,0.0764745176,-0.2304570973,0.2309364825,-0.1129720733,0.4178794026,-0.2574656308,0.0388289429,0.0287062153,-0.0909330323,-0.0192771889,-0.3598273695,-0.1379480213,-0.0087427525,0.1893775463,0.1435693353,0.0545647368,0.1428648531,-0.0493662804,0.3422606885,0.0329733938,0.0365742669,0.2057769001,-0.3593283594,0.1581130028,0.057816457,-0.0561056882,-0.0303860791,0.0016268623,0.0125824781,-0.0291041844,0.3670475483,-0.3468731344,-0.1255812198,0.1208589822,-0.2924034595,-0.4044805169,-0.0286532808,-0.5490256548,-0.1081628427,0.169148311,-0.1260567456,-0.1499746889,0.034565337,0.1518105417,-0.1971428543,-0.0772140771,-0.0292112287,-0.4185020328,0.0718787462,-0.0767299458,0.2225262225,0.0704554319,0.2563611865,-0.0424101911,-0.1049350277,-0.1720269918,-0.0251968689,0.4814112186,0.2461594492,0.1562116146,-0.0903977975,-0.2662247121,-0.1463945359,-0.1505852193,-0.199605152,0.1895070374,0.0152619826,-0.6571384668,-0.0384223498,-0.1666606069,0.1685894579,0.1548557431,-0.1976621449,-0.1159150675,-0.574603796,0.2282319516,-0.4392106533,-0.036352288,0.3599653244,0.0648710579,0.1356104463,-0.2745454609,0.1723674238,0.1712717861,-0.3440077305,-0.122380808,-0.2971184254,-0.0006398571,0.6353337765,0.2008113712,0.7802702188,0.0081792064,0.0361435972,0.1069833413,0.0827194825,0.3759540915,-0.3120654523,0.0499523841,-0.0926077515,-0.1861398667,-0.1178764105,0.1165541708,0.0963097289,0.2510374188,-0.1902992576,0.4400809705,-0.0062454762,0.0803439096,0.0197490733,-0.0302470699,-0.001280684,-0.0020521996,-0.0266558863,0.1806299686,0.1246197596,0.1315397024,-0.100544095,0.101019755,-0.3447440565,0.0227893069,-0.1532878578,-0.107746534,-0.3275676966,0.1785721928,0.0154510187,-0.3179487586,-0.321108371,0.6958513856,0.3500355184,0.2806239128,-0.0372616947,0.17396833,0.0262522697,0.1239001602,0.0625125989,-0.0658480972,0.1633388847,0.0229656585,-0.199406147,0.0879433677,0.1445674747,-0.259915024,-0.0262720641,-0.0026541948,-0.3019395173,0.1306394935,-0.0025994915,0.0868672654,-0.1519218981,-0.1816911697,0.1379817277,0.0430669636,-0.068843849,0.223366037,0.0550942346,-0.2668168843,-0.172237888,0.2863359451,0.0703241378,-0.1017246917,-0.2562138736,-0.1745773703,-0.0864023417,-0.2994310856,-0.0894373506,-0.2981826663,-0.1044154465,0.377630651,0.0189895015,-0.1154141426,0.0512116216,0.6299238801,0.2141922861,0.3644081056,0.1524917334,-0.3336916566,-0.1081482247,0.1698487401,0.3736403584,-0.0335493498,-0.0641351417,0.0487426519,0.2199577242,0.4095642865,-0.4029997587,-0.3656337261,-0.3964770734,-0.08165925,-0.4356089532,-0.2211939394,-0.1281733215,0.0936490744,0.0950353593,0.2374957353,-0.228664428,-0.2447851598,-0.1591931134,0.0821181983,-0.1456952542,0.4021069109,0.0413876586,0.1853424311,-0.1141620576,-0.3821869493,0.0746247098,0.3171326816,0.2551471293,0.0177339055,-0.0402403437,0.2804537714,-0.1135493368,-0.1370173097,0.0539059229,-0.0051347069,-0.0932940915,-0.0665030628,-0.0675680414,-0.0987473354,-0.2189815044,-0.0269001983,0.1849087775,-0.1214119717,-0.1053618267,0.0837505609,0.0456283689,-0.3730870485,0.4326426983,0.2211647034,-0.0085237874,-0.1985666156,-0.1335076392,0.2194379419,-0.458932668,0.2654249966,-0.2891064882,0.1563417614,0.3173478544,0.4556881785,0.2233205438,-0.0649050102,-0.0679805204,-0.0987004712,-0.1383628547,-0.1002944857,0.062933147,-0.3096041381,0.2589257061,-0.0202683285,0.4849733412,-0.1121174693,0.2652066946,0.1309785694,-0.0034165408,0.31945315,-0.1190062836,0.1960355192,0.0193226915,0.1845236868,0.1401707679,0.4659497738,0.1110079139,0.0070335236,0.0470275506,0.2889150083,-0.1265300214,-0.1758518517,-0.2906351984,0.0854660273,0.0524398312,-0.3401103914,-0.2391038239,-0.2276740968,-0.4451782107,-0.1638911217,0.0658188462,0.1539925188,0.3426798582,0.0251932032,0.1911427826,-0.2424500436,-0.2109352052,-0.2723113596,0.5643007159,-0.2484370172,0.1921278387,0.1572182029,0.3751194477,0.1882608682,0.4240934551,0.2116145343,0.1786782146,-0.2010879666,0.2564738095,-0.3552730978,-0.1552319974,0.0189652536,0.2312572449,0.1610189527,-0.2643475831,0.4957166612,0.1359371543,-0.2322770804,0.4143300354,0.0861595869,0.4273025692,-0.2493314743,0.3634840846,0.2400841862,0.2938927412,-0.0414929055,0.1830854565,-0.4155995548,0.0779677555,-0.6234217882,-0.1174249873,-0.1011873037,-0.2356607765,0.1084447652,0.0924106166,0.205197528,0.1181329116,0.3476846218,-0.0357577614,-0.1855497658,-0.4065920115,-0.042920582,-0.1452928782,-0.0774550661,-0.2424650788,0.3812710047,-0.0584685802,0.555036068,0.0424526259,0.5779340863,-0.2886163592,0.2897717059,-0.5447959304,0.3275338709,-0.3254658878,-0.0551965535,0.1837163866,0.0911929831,-0.0703221932,-0.1465926617,0.0699736029,0.2719673812,-0.3368913531,0.2480897009,0.3208194077,0.3344044089,0.1007984728,0.1854650527,-0.0619373135,-0.1647574008,0.0199882947,-0.1395956278,-0.0411484055,0.2344657928,0.2039119452,0.1925023347,-0.0568336099,0.0826726332,0.1938952804,0.7380520701,0.1352984458,-0.1584627926,-0.3204585314,0.0522537529,0.0395437106,0.0217203535,-0.5980477929,0.3872767091,-0.0852566436,0.2723891735,-0.033462014,-0.3992609978,0.4943167865,0.0310679153,0.0911919698,-0.2405092567,0.2732808888,0.2411660701,-0.0841868818,-0.3257416189,0.0797034875,0.0479561538,0.3362016082,0.0674583316,0.1727021784,-0.1792183667,0.0896348804,-0.3200744689,0.0919146091,-0.1244571134,0.1173276678,0.0839771628,-0.4272243679]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/361","title":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic","comments":"Maybe we can set all the random seeds of numpy\/torch etc. while running `metric.compute` ?","body":"If I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?","comment_length":15,"text":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic\nIf I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?\nMaybe we can set all the random seeds of numpy\/torch etc. while running `metric.compute` ?","embeddings":[-0.3071688116,-0.3486666679,-0.1056947112,0.1829099357,0.2425306588,-0.2551713586,-0.0964718312,-0.169013381,0.0383498073,0.5325791836,-0.000080875,0.338652432,0.0471655987,0.0112029035,-0.2225715816,0.0628180727,-0.0477628224,-0.0120155774,-0.0196598638,-0.2982631624,-0.2247394621,0.0446664169,-0.2020913512,-0.1766828746,-0.1038242355,-0.0176766273,0.0080622118,0.02762614,-0.4444930851,-0.0773079172,-0.0258409902,0.3403029442,-0.1220054179,0.4497037828,-0.0001130536,-0.2740183771,0.0494141765,-0.1270534396,0.209228754,-0.4266759753,-0.0062103053,0.160988301,0.0120171793,-0.1981359124,-0.106811434,0.1036126614,-0.2087495178,-0.1786199808,0.3288964629,0.2066942453,0.1814853847,-0.0189275574,-0.1831047982,-0.2920492887,-0.0098734116,0.0039267396,-0.1489508599,0.5082100034,0.0208433699,0.2129074633,-0.1407389194,0.0463533401,0.2043962926,0.2223544121,0.5045521855,-0.1945196986,0.4104897082,-0.1617786884,-0.1390412301,0.3539476395,0.2505722642,-0.1146517247,-0.0902426243,0.4209291935,0.0737080127,-0.4961270392,-0.0759759769,0.0164898001,-0.0335100517,-0.1680622697,-0.0147617832,0.3003780246,-0.1617907435,-0.0923710763,0.1779441237,-0.0077436147,0.0227024481,0.0921171084,0.1928072721,0.4297475815,-0.0924040079,0.1293128282,-0.3246498108,-0.076334089,-0.4897935987,-0.0357861295,-0.0211488418,0.3562854528,0.1053723246,0.3411879539,0.4873179793,0.0171684232,-0.0515537113,0.1216150373,-0.170541048,0.3494719863,-0.2301095277,0.0440974385,0.2458483875,0.0572571754,0.030761309,0.0150410701,0.4267674387,-0.632773757,0.2098538429,0.3960829973,-0.1820952594,-0.1387283504,-0.567238152,-0.157331571,-0.2693554461,-0.1768655479,-0.0624911077,0.2180206031,-0.1925328374,0.0238599144,-0.1196752265,0.1626657993,-0.116223745,-0.0008022381,-0.2876538932,0.133716464,-0.426769793,-0.0739220604,-0.2942961454,0.0471362844,0.1568453908,0.312092483,0.1133018583,-0.1673799753,0.0835047141,-0.4124015272,0.0088950861,0.0103664836,-0.1911911517,-0.1352604777,0.1450576782,-0.091338098,-0.136943534,-0.1211650372,-0.1292176396,0.2166012526,0.3786970079,0.1052620038,0.0243496783,0.2134269327,-0.1133527756,0.2101715654,-0.3564652205,0.0809220076,0.1408926398,-0.1336670518,-0.2788926959,-0.3087985814,0.1350933909,0.1376220286,-0.0333044864,0.1684652567,0.5491127968,0.0027146109,0.2577679455,0.4878573418,0.1275294274,0.0972004011,-0.2567596436,-0.6177899241,0.4872510433,-0.4754666686,0.1378493905,-0.0649493709,-0.1007417962,-0.1021167487,0.1059131399,0.1896212697,-0.2873932123,0.083040297,0.2668371797,-0.0867927969,-0.0168861058,-0.3102291226,-0.1248051077,-0.0147506194,0.3205568492,-0.2120385915,0.299441576,0.2642761469,0.0291451961,-0.3353279531,0.2161999494,0.1145992056,-0.3973164558,0.1029447913,0.0894590244,0.1706748754,-0.0352104045,-0.3080494702,0.1051071659,0.1075172052,-0.5630425811,0.0013302976,0.4416306019,-0.1414522231,-0.2385753244,-0.0711005181,0.0838703737,-0.176228255,0.1716247499,0.2116722614,0.2480884343,0.1928101629,0.0070125009,0.1662473828,-0.3233706355,-0.1128162518,-0.2407324016,-0.1684615314,-0.0673749819,-0.0242157951,0.0733070299,0.26931867,0.233274743,-0.0584415905,-0.1535453051,0.3406137824,0.4947661161,-0.1076860279,0.2116789818,0.2146015316,0.0716172159,0.1242035702,-0.1898176521,0.5052815676,0.1439654827,-0.1473278701,0.1658174396,0.2924778461,-0.5452288389,0.0770556107,-0.1236602366,0.0403558165,0.096547097,0.2339107841,-0.141443342,0.0098272385,0.4592808187,0.1075910851,0.3435699046,0.0538212024,-0.3303044438,-0.3162704706,0.1928669065,0.1553273052,0.0830648765,-0.118459262,0.2096378356,0.0151117081,-0.1182823777,-0.1098705381,0.1401434392,0.0760636106,-0.061039418,0.0378812887,-0.2896448374,-0.3430507183,-0.1591786444,0.0150692165,0.1496402025,0.2640662193,0.1969182938,0.1654215604,-0.2817268074,-0.2757239938,-0.1611284018,0.0220738072,0.0554314479,0.2452594936,-0.2741921544,0.4351193905,0.0350728184,-0.1873741597,-0.1272006035,-0.171759963,0.0059523149,0.2068740726,0.2122708559,0.0220586322,0.546092093,0.1960154474,0.0319115371,0.0859679058,-0.2156645209,-0.1760206968,0.0516794808,0.0849605203,-0.0954848453,0.0681101605,0.2836454213,-0.3714491129,0.0690492615,-0.0917345881,-0.1623264402,-0.1710133702,0.0321215838,0.3118729293,0.1836816818,-0.4161848426,-0.2150340974,0.4828081131,0.351801455,-0.1875349432,-0.155581072,0.0929151326,-0.1402235925,0.1010479107,0.0360231735,-0.0352836847,-0.1007601023,-0.1873556525,0.5571433306,0.1030348092,-0.0582870394,0.2828419805,0.0613535419,-0.1063121557,0.2731641829,0.2738355696,-0.3725709021,-0.2650184035,0.2681061029,-0.2202876061,-0.4329479039,-0.0847421139,0.1684612334,0.0451311544,0.2519676685,-0.1250103712,-0.4636136293,0.0301546697,0.1120977998,-0.2353634685,-0.1392844468,0.5804662108,-0.2581083179,-0.1832357645,-0.0265334621,0.1679528058,0.0592262633,0.0197962709,0.2008321732,-0.1686944664,0.1858615279,-0.1309723556,0.1605399996,0.3325456381,-0.0934568197,0.1199183911,0.0063947896,0.3827969134,-0.3928347528,-0.0137320468,0.4141015708,0.1525927931,-0.238516584,0.1391624808,-0.2140777409,0.4970116615,-0.2016540766,0.1340014637,0.2673155367,0.0050467718,-0.1174543276,-0.4213687778,-0.0472094268,0.017551668,0.2809440792,-0.0007405393,0.0678394884,0.2605513632,0.1210130081,0.2892855108,0.0376408212,0.0436864384,0.0328321718,-0.3341841102,0.1276302934,-0.0955976993,0.0783540532,-0.1535052657,0.2046763897,0.1275328398,-0.0518307909,0.546202898,-0.4143949151,0.0306808818,0.1701609641,-0.5283370614,-0.2362229675,-0.1307359338,-0.2590715587,0.0251738746,0.2200779766,0.2195806056,-0.0930067226,-0.1684991419,-0.1384697706,0.1125211492,-0.1266055703,-0.1265139282,-0.6075134277,0.0908089057,0.0752537325,0.2020387203,0.1465095133,0.1806215495,-0.0057920516,-0.1022854522,-0.4052717686,0.0619356222,0.3059812188,0.1109411567,0.0979376286,-0.1026023999,-0.2043650597,0.06653063,-0.0830461904,-0.1002259776,-0.0168157499,0.105938822,-0.5876333714,0.1893910915,-0.1321341246,-0.0093811853,0.0999739394,-0.1244236231,-0.2285663784,-0.6708893776,0.3494000137,-0.505875349,-0.0028587263,0.4361331165,0.0832642987,0.0358709358,-0.2945687771,0.1238795072,0.1399357021,-0.3176773489,0.2769515812,-0.3334616423,0.0064909793,0.3103612363,0.4838728607,0.6186304688,-0.131367892,-0.1229100823,0.1712699234,-0.110100694,0.2557354271,-0.4233346879,0.0631129593,-0.1806609184,-0.421415776,-0.1173332632,-0.0493816547,0.1314555407,-0.1009342223,0.0112151168,0.456248045,-0.0918744579,0.1954208016,-0.0243170243,-0.0050087264,0.1888028085,0.0461776145,0.4135700464,0.1332813352,0.0880834758,0.267401129,-0.2226205468,0.0722086057,-0.152937144,0.1379254013,-0.1754379719,-0.2988605499,-0.3430783451,0.2633574903,-0.1394699663,-0.3520556688,-0.1556844115,0.7142415643,0.6851271987,0.2379565686,0.0003305153,0.257599324,0.0338763632,0.0226886533,-0.1466402262,-0.1359835863,-0.030524157,0.0079674274,-0.19826442,0.0444883667,0.0649788976,-0.3481219411,-0.3373194039,0.0639777407,-0.3041005135,0.2912574708,0.1368459463,-0.0909281895,-0.0984914675,-0.1478771865,0.1316215247,0.0523379967,0.0138756158,0.3708333969,0.0611125007,-0.4661808312,-0.148219496,0.2801054418,-0.1327505261,0.0138123007,-0.2592051625,-0.3332419097,-0.0243096389,-0.3508079946,-0.1390509754,-0.186406374,-0.0772726834,0.3172185421,0.0359580852,-0.0677031055,0.0935520902,0.5891634822,0.2121540755,0.1018737182,0.1656658798,-0.3521412909,-0.0969288722,-0.0128946584,0.2019720376,0.048567649,-0.1371462941,0.1326278001,0.0583772697,0.406293571,-0.3389250934,-0.2799271643,-0.2222305685,0.0255046356,-0.6056932211,-0.1955828965,-0.1056989431,0.2600302696,0.0325350612,0.1825203001,-0.4058769941,-0.1745394766,-0.1069540754,0.1081600413,-0.0721203387,0.4262012541,-0.0278858859,-0.0159609616,-0.1543115973,-0.2530748844,-0.0280598663,0.2532704473,0.2134234309,0.0413817726,-0.0267624985,0.4862571657,0.0107632559,-0.0374889895,0.1666043401,0.1046523228,-0.3167167902,-0.2367882431,-0.1614117622,-0.1250146925,-0.5374773741,-0.1490602344,0.0246208794,-0.091899693,-0.0836190358,0.1871673167,0.2700785697,-0.2876213193,0.2627876401,0.2160527259,0.0100747878,-0.2202557474,-0.04242431,0.1668915153,-0.2805265486,0.2872318327,-0.2737960517,0.2964896262,0.2723712027,0.3313200176,0.3885004222,-0.1131901667,-0.0280461628,-0.2074231803,-0.0915593803,-0.2870478034,0.1546484232,-0.1724530905,0.1523491293,0.1200982407,0.6779108047,0.1597706527,-0.0000259336,0.465988934,0.1357311606,0.3235052526,-0.2731000483,0.0455513857,-0.1796839684,0.1208146736,-0.0952622965,0.388181895,0.1472611874,-0.1698942631,-0.0417659581,0.0561330877,-0.3907006979,-0.2497467548,-0.3996177614,-0.1114466563,0.0191172138,-0.1666596085,-0.227426067,-0.0805464908,-0.0253840871,-0.0536296405,0.2295603007,0.3170382679,0.2686028481,0.0951962173,0.168624416,-0.0225057025,-0.1267735213,-0.0302702542,0.4462606907,-0.1537112296,-0.0241245013,0.0299264416,0.4090354145,-0.0257155113,0.2647469938,0.0528689437,0.3578320444,-0.4224910736,0.0984164104,-0.2416149825,-0.067215845,-0.016755322,0.3047628701,0.135499537,-0.3248305321,0.4916397333,0.0768827647,-0.2175402343,0.4296916425,0.0366579294,0.4775718451,-0.1160358042,0.4312669337,0.1630425751,0.2634854317,0.0123963254,0.1673470885,-0.2784704268,0.208575055,-0.7018001676,-0.1481464356,-0.017298419,-0.1818991005,0.1003841832,-0.0810424238,0.22454229,0.0587725863,0.2002773732,-0.236643523,-0.213076666,-0.4754006267,0.1335634142,-0.0765098259,0.174855575,-0.1184963882,0.3406633139,-0.2593601942,0.2279869318,0.0793741941,0.4671773016,0.0149136009,0.420553118,-0.5850593448,0.2641522586,-0.4944480956,-0.2148056477,0.2384154648,-0.0002051011,0.0275110044,-0.0555003211,0.0641499758,0.0847276151,-0.0006650062,0.0369222313,0.8078070283,0.5269052386,0.0892181993,0.0759112611,-0.0593464412,-0.1299541593,-0.001089685,-0.0989929214,0.0586380698,0.2863797545,-0.0097497338,-0.1764185429,-0.1141716018,-0.0045527206,-0.0819385946,0.8358350992,0.0853826031,-0.1034378186,-0.3838441968,0.0785609037,0.0145507641,0.1683319062,-0.2815384865,0.3795695305,0.0461364724,0.2292894572,-0.1356183141,-0.1763154715,0.5751307011,0.1377317905,-0.0448851176,-0.1534895152,0.2500424087,0.5654434562,-0.0947335362,-0.1499228328,0.1542792469,0.0438741148,0.2151224613,0.0596779808,0.1480820328,-0.3632470071,0.1589556783,-0.1535050571,0.1022541821,-0.1779581904,0.0291427113,0.2101557255,-0.250438869]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/361","title":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic","comments":"We should probably indeed!","body":"If I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?","comment_length":4,"text":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic\nIf I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?\nWe should probably indeed!","embeddings":[-0.1868329942,-0.3820251226,-0.1310739517,0.1244154349,0.2069661617,-0.3292270303,-0.1116598621,-0.2533222437,-0.0107660452,0.4409889579,-0.0467224047,0.3216395974,0.0021065935,0.076502949,-0.2607601881,0.0633751974,-0.0309194867,-0.0124057746,-0.0021001149,-0.2893410921,-0.1370517462,0.0542556569,-0.1057893187,-0.0912712291,-0.0381632671,0.0718705654,-0.0698888525,-0.0022065814,-0.4454659522,-0.0896837711,-0.0421111472,0.3296396732,-0.111952439,0.3652858436,-0.0001073887,-0.3276283443,0.0722343922,-0.1527214497,0.2462241352,-0.4164839387,-0.2068138868,0.2621767819,-0.0462455563,-0.1297953874,-0.1151640862,0.0981125385,-0.3086803257,-0.0344686769,0.3919472694,0.2149597406,0.2519872487,-0.0983963162,-0.2328746468,-0.3281814754,0.1156718805,0.0502459668,-0.0683090538,0.4572075009,0.0168437604,0.1273802966,-0.1959655285,0.121299617,0.268409878,0.1646289378,0.4498093426,-0.2232506424,0.2914678752,-0.062441729,-0.1490819603,0.3279915452,0.2620075345,-0.1708195955,-0.198752284,0.3283813298,0.1602395922,-0.3644971848,-0.1137759313,-0.122522153,0.0849900916,-0.1680197567,0.0338201895,0.3756248057,-0.1387617141,-0.1636543572,0.1885714978,0.0083586527,-0.0254405923,0.0585503876,0.1343435496,0.3746121228,-0.2405302674,0.1093296036,-0.3410531282,-0.1588601023,-0.4331399202,-0.1215356365,-0.0250670388,0.4194001555,0.128801927,0.4114228487,0.4400978982,-0.1143802777,-0.0019414999,0.0064484947,-0.0664896518,0.3083770573,-0.2310779691,-0.0397523232,0.1631080806,0.0029179973,-0.0619718917,0.0431696177,0.4392738044,-0.529129684,0.1862962246,0.2351814359,-0.2170609385,-0.2781304121,-0.6306470037,-0.0559014119,-0.2847448289,-0.1583527625,-0.0030170523,0.2668909132,-0.2578425705,0.1180985793,-0.2248953879,0.1503446847,-0.0539545529,0.0026644962,-0.3166173697,0.1566881984,-0.5278322101,-0.1583383679,-0.3407934904,0.1886647642,0.1244528517,0.2786102593,0.0894527659,-0.1547610164,0.0431925431,-0.4620055556,0.0759060904,-0.0167020429,-0.2497481704,-0.0850927308,0.0805103704,-0.162200138,-0.1889112741,-0.1284428537,-0.1059843898,0.2045177668,0.3953988552,0.1704618335,-0.0054314607,0.1091010273,-0.1333515197,0.1974597275,-0.4238055348,0.0326236039,0.1411114037,-0.1281948686,-0.2086164206,-0.2856645286,0.0290023834,0.1727999449,-0.0609467179,0.0895453617,0.6058724523,-0.0227913763,0.2231531143,0.4621888697,0.1332287341,0.0120253852,-0.2227273285,-0.5630184412,0.4791244268,-0.4543258846,0.0706592575,0.0008875453,-0.0580071434,-0.0511478819,0.113210313,0.1480252743,-0.2623061538,-0.0118829627,0.1606170684,-0.1420883536,-0.0209134892,-0.2035630196,-0.0934491009,-0.0452615097,0.310700655,-0.1541433483,0.2595015764,0.1284834445,-0.0648917034,-0.1846206188,0.2526140213,0.1536877453,-0.4057266414,0.0503889062,0.1929364502,0.2532126904,-0.0316944458,-0.2557194531,0.1050211489,0.0212716479,-0.5074988008,-0.0482115187,0.5442959666,-0.195097059,-0.2462213039,-0.0591188632,0.0863797069,-0.1759476811,0.2410847694,0.2646804452,0.2092234194,0.2077679038,-0.0604934469,0.1965868026,-0.3090080023,-0.1450550556,-0.1224610656,-0.1396504939,-0.1331427991,0.0308183543,0.1519915015,0.3575056493,0.1927637458,0.0393880419,-0.0977614075,0.2407502681,0.4523765743,-0.0975507423,0.1662641615,0.2524042428,0.0847834349,0.1802632362,-0.230038926,0.4459053278,0.183458373,-0.0937266499,0.0947056636,0.2832410038,-0.487839371,0.0424129814,-0.0594317429,0.0133812772,0.0825587139,0.1308110654,-0.1715982854,-0.0804411396,0.4021075368,0.039682392,0.3941456079,0.0892556086,-0.3898450434,-0.1670433134,0.2051979154,0.1269613057,0.0997831225,-0.0652534366,0.1807874292,-0.0137179708,0.023517685,0.079248935,0.1147440895,0.1567984968,-0.1227531955,-0.014293842,-0.3157766461,-0.2952973545,-0.1908819675,-0.0748753771,0.1894635111,0.3616863489,0.152854532,0.2688712776,-0.3094057441,-0.4841583669,-0.2119410336,-0.0021198466,-0.0044140974,0.2395001799,-0.2562286258,0.3229572475,0.070366621,-0.330227077,-0.0854355097,-0.1935395002,0.0267447066,0.1496684402,0.1471778303,-0.0583159886,0.4256923497,0.2270112783,-0.0810159072,0.2247377038,-0.2250508517,-0.1500669122,0.0913242027,0.164119795,-0.1050740257,-0.0125942957,0.2878491879,-0.3048312068,0.086433351,0.0246081017,-0.249012351,-0.056558203,-0.0465912856,0.1931528449,0.1511023939,-0.4815579653,-0.2109377831,0.430878818,0.2370175868,-0.1798562557,-0.185460642,0.0637653396,-0.1404277086,0.11321114,0.0716556236,-0.0124138109,-0.0856313407,-0.098973617,0.4414330125,0.1091010273,-0.0882896706,0.3782384694,0.0315805674,-0.0433119498,0.1527813375,0.2788930535,-0.3980428576,-0.1827379465,0.1893917024,-0.265280962,-0.4594363868,-0.1289179623,0.2124416977,0.074731335,0.0775164738,-0.0798738897,-0.5114732385,0.047983896,-0.0288457964,-0.2504256964,-0.1616290957,0.5564633012,-0.2390878052,-0.2710494101,-0.0393289737,0.2414560467,0.0130619081,0.0143916225,0.222901389,-0.3350796402,0.1768470556,-0.1116986498,0.0602959096,0.4011085927,-0.0991942808,0.0786763057,0.0169755202,0.3171259463,-0.413353622,0.077020362,0.4289723933,0.0296788979,-0.2087238431,0.2207171619,-0.1750689149,0.536162138,-0.1319532692,0.1569310278,0.2260276526,-0.0137190036,-0.1295332313,-0.3061078191,-0.2139220834,0.0124139329,0.1819624454,0.1053853929,0.1157303229,0.2425478697,0.1364931166,0.2551181912,-0.0019893609,-0.0496809408,0.1443843395,-0.3079297245,0.0567067489,-0.093847461,0.0019720243,-0.127687335,0.0763895139,0.0641262531,-0.1464233547,0.4547672868,-0.3206009865,0.0943099335,0.0516567975,-0.5329067111,-0.2380025685,-0.1708258986,-0.3044445515,0.0405090936,0.2040159553,0.1656197459,-0.0058376044,-0.1328471601,-0.0788594857,0.0033503431,-0.0326584131,-0.1239487007,-0.4979712069,0.1436924189,0.031315919,0.235402897,0.0993255749,0.0988123715,-0.0488161743,-0.0564344712,-0.3426407576,-0.0401220433,0.3448575139,0.1968804151,0.0649350956,-0.0355781354,-0.2415314317,-0.0216468051,-0.0136760445,-0.1704569161,0.060934376,0.1657238007,-0.5762451291,0.191777885,-0.120046407,-0.012120191,0.1723191589,-0.1021506488,-0.1820546091,-0.6633113623,0.3249073625,-0.551594913,-0.0822806358,0.3909974992,0.0186353475,0.0013933826,-0.2367587388,0.1313038766,0.0618440136,-0.4200308323,0.1867961437,-0.1606332958,-0.0146466279,0.5139412284,0.4458146989,0.6930046678,-0.0824716985,0.0266116485,0.2220344543,-0.122328721,0.2426048815,-0.3213399947,-0.0292466786,-0.2003032118,-0.3337451816,-0.0550667532,0.0150734307,0.1167001277,-0.0351288207,-0.01900452,0.3259264231,0.0109587647,0.2303990573,-0.0273515861,-0.0783647895,0.274184823,0.0889264047,0.2454777658,0.2292731404,0.1043791249,0.2073681653,-0.1696545035,0.1122837365,-0.2621998489,0.113086924,-0.119213298,-0.1674020886,-0.2684091032,0.2034647167,-0.1593233198,-0.3865301609,-0.123227343,0.7034853101,0.5501500368,0.2816726863,-0.0109240832,0.2591167986,0.0692513809,0.1147686839,-0.142273128,-0.1808340251,-0.0066018817,-0.0386494249,-0.1876217127,0.1022412628,0.1504538059,-0.3386609554,-0.326582402,0.1033858433,-0.2998981178,0.1481349021,0.0196983926,0.0098153623,-0.1344832927,-0.0993228108,0.1724595129,0.0470774621,-0.0243955962,0.3497768939,0.1163060516,-0.3641503155,-0.2651942074,0.3361677825,-0.1473871022,0.1385841668,-0.268771112,-0.4065554142,-0.0104706865,-0.379445523,-0.1314314008,-0.098863557,-0.1002213284,0.1431701481,0.0147107365,-0.1246058494,0.1420059949,0.5597622991,0.2457712889,0.1383330673,0.2938136756,-0.2751525342,-0.1378108114,0.0353996865,0.2056105584,0.0298151821,-0.1081992313,0.0805380195,0.0683650523,0.3579197824,-0.4156235754,-0.2534103096,-0.2248643935,-0.0159831773,-0.6290693283,-0.2835083008,-0.061800532,0.1642417759,0.1042355523,0.1338131428,-0.3285191655,-0.2339075506,-0.156902656,0.0778210014,0.0200490095,0.4673627317,-0.1615550816,0.1490562111,-0.1246951148,-0.1728420109,0.0787181705,0.3387693465,0.244059965,0.1053498387,-0.1221901923,0.2874769568,-0.0097845895,-0.0731559172,0.1820210069,0.0520639867,-0.2058625519,-0.2252456397,-0.1227797419,-0.1152989343,-0.5732499957,-0.0340265818,-0.0441862382,-0.0589883327,-0.0846716464,0.0986899436,0.3384975195,-0.1436430663,0.3120319247,0.169009909,-0.0267651491,-0.2816347182,-0.030752318,0.2268004715,-0.3374214172,0.2939179838,-0.2462858856,0.2929472327,0.2691016197,0.3198329806,0.3724926114,-0.1238178834,0.0396431945,-0.1327880174,-0.1504703015,-0.1779999435,0.1202859953,-0.2243051976,0.1776306629,0.0990675464,0.606374383,0.1057454571,0.0473776609,0.3948785961,0.1279547364,0.3096334338,-0.301327765,0.1161972433,-0.1192884669,0.0990935639,-0.0507525429,0.3946661055,0.1129396409,-0.1318799406,0.0050234315,0.0250629634,-0.1863994896,-0.3559059203,-0.3109351099,-0.1889096349,-0.0506624542,-0.182748422,-0.2635511458,-0.0625972599,-0.114883177,-0.0045101913,0.0609246232,0.3308899105,0.2831991911,0.0831110999,0.166564405,0.0014432039,-0.1495685279,-0.0686459541,0.4919020832,-0.1465205401,0.0363426171,0.0692355037,0.4255665541,0.1091473028,0.2582312524,0.0629801676,0.2564480305,-0.2849952281,0.0742319077,-0.2006184906,-0.0766502619,0.0621676035,0.2413691729,0.161971122,-0.3022025526,0.5529306531,0.1459435672,-0.2253106087,0.5158164501,0.0749709383,0.3790692091,-0.0778473839,0.3850711584,0.2512733936,0.2868478596,-0.0081422692,0.273450315,-0.2960493863,0.0832054019,-0.6911401153,-0.1326416582,-0.0157416742,-0.2729343772,0.1254582554,0.0634471849,0.213082552,0.0133143393,0.2478012145,-0.1625012308,-0.1691331416,-0.4491275847,0.1497945637,-0.0505092815,0.2903128862,-0.1079273969,0.3591668904,-0.2611941397,0.2754891813,0.1142485589,0.4547592998,-0.0936250612,0.3841608167,-0.5362656116,0.3478264809,-0.3615690172,-0.1538399756,0.2169893235,0.035860531,-0.0038767545,-0.023589164,0.1051125526,0.0930930525,-0.0424508639,-0.03983403,0.6680932641,0.5583518147,0.070913516,0.1401723921,-0.1297888011,-0.2338669002,0.0166305173,-0.1369310468,0.0235497244,0.2530646622,0.0065585524,-0.1075058281,-0.0192332435,0.0739432201,0.0813416392,0.6987603307,0.1129723266,-0.0405220836,-0.4773541987,0.0965273976,0.1063421145,0.1118107513,-0.4171556234,0.4573455751,0.0212970879,0.362636447,-0.1284968555,-0.1360679865,0.5653137565,0.0854171589,0.107810393,-0.0875257328,0.253467083,0.4937237501,-0.1238370761,-0.1778147817,0.1186614484,0.0343672968,0.2563946247,0.0650013164,0.1439231038,-0.3326025307,0.0948733911,-0.1617919207,0.1858672649,-0.1549553573,0.0351855494,0.2672810256,-0.2662825286]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/361","title":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic","comments":"Now if you re-run the notebook, the two printed results are the same @colanim\r\n```\r\n['0.3356', '0.1466', '0.2318']\r\n['0.3356', '0.1466', '0.2318']\r\n```\r\nHowever across sessions, the results may change (as numpy's random seed can be different). You can prevent that by setting your seed:\r\n```python\r\nrouge = nlp.load_metric('rouge', seed=42)\r\n```","body":"If I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?","comment_length":50,"text":"\ud83d\udc1b [Metrics] ROUGE is non-deterministic\nIf I run the ROUGE metric 2 times, with same predictions \/ references, the scores are slightly different.\r\n\r\nRefer to [this Colab notebook](https:\/\/colab.research.google.com\/drive\/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem.\r\n\r\nExample of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differents run :\r\n\r\n> ['0.3350', '0.1470', '0.2329']\r\n['0.3358', '0.1451', '0.2332']\r\n\r\n---\r\n\r\nWhy ROUGE is not deterministic ?\nNow if you re-run the notebook, the two printed results are the same @colanim\r\n```\r\n['0.3356', '0.1466', '0.2318']\r\n['0.3356', '0.1466', '0.2318']\r\n```\r\nHowever across sessions, the results may change (as numpy's random seed can be different). You can prevent that by setting your seed:\r\n```python\r\nrouge = nlp.load_metric('rouge', seed=42)\r\n```","embeddings":[-0.1073278859,-0.2159305811,-0.0668013096,0.1431299895,0.182839334,-0.3188183606,-0.136663571,-0.1566527337,-0.0810187757,0.394208163,-0.0958697796,0.57323879,0.0757611692,-0.0353061259,-0.004969799,0.0457197651,-0.082976751,0.1978804171,-0.0148286335,-0.3589980304,-0.2548868954,0.1192239076,-0.1512854099,0.034798637,-0.1317696273,-0.0622886382,0.0263473354,0.1554979384,-0.4091148674,-0.2266361415,0.0821629763,0.2336032242,-0.1760028303,0.2849853337,-0.0001076949,-0.2813829482,0.1078296527,-0.116991885,0.0769580379,-0.4047334492,-0.0405928195,-0.0956251547,0.1767806113,-0.2945886552,-0.097563915,0.136918813,-0.0897162333,0.0338977352,0.4158981442,0.2269689739,0.2249114364,0.0455720685,-0.2391411811,-0.1621117443,0.0426181145,-0.0929484665,-0.0321005546,0.5454743505,-0.1822020411,-0.0067190747,-0.1354897916,0.1270417869,0.1383430809,0.2235369384,0.4007498324,0.0411152132,0.3929459453,-0.135845989,-0.2343785465,0.2525841892,-0.0129311113,-0.0997708142,-0.1324526966,0.0841452777,0.1587257683,-0.6156039834,-0.1732088774,0.044013951,-0.1350554824,-0.2012695372,-0.0508786775,0.3788500428,-0.0968494341,0.1117188707,-0.0043131849,0.1680879146,0.027076317,0.1550635397,0.1478287429,0.2116188109,-0.1319663823,0.1222030297,-0.2151378393,-0.0598680042,-0.3910151124,-0.0947208554,0.1377711445,0.351981163,0.0234441087,0.1232677326,0.5981099606,0.0189317055,-0.1684436947,0.1034671888,-0.1521719992,0.3907270133,-0.0975617915,0.020345198,0.2817369103,0.037295226,-0.040691454,0.0754893124,0.5044953227,-0.4349454939,0.205779776,0.2119951993,-0.1122406051,-0.2276438624,-0.5611451864,-0.0430729054,-0.2027511895,-0.019362092,-0.1004149541,0.331977874,-0.3697909117,0.1567452401,-0.0824483782,0.1941112578,-0.2658010125,-0.0755543262,-0.3253019154,0.0633458197,-0.5507808924,-0.0057863784,-0.1304560006,0.2926286161,0.2511183321,0.2495184094,0.0323722661,-0.12549375,0.07507395,-0.3315087259,0.0869627148,-0.0499799624,-0.1087681577,-0.0286275763,0.1192320287,-0.2189495564,-0.196575284,-0.0809318274,-0.05959085,0.2353450209,0.3346344829,0.1933609396,-0.1109464914,0.0923801437,-0.0715812221,0.2088449001,-0.3361066282,-0.0010506114,0.2094528973,-0.2351088524,-0.3856660128,-0.2787617743,0.0916537717,0.1444386244,0.1349799186,-0.0671066642,0.4620255828,0.0558448583,0.144136712,0.4726661146,0.1237938255,0.0034985109,-0.060925696,-0.4403909147,0.6245052218,-0.414652586,-0.041249264,-0.1303750575,-0.0703398138,0.0291820616,-0.0193940718,0.2022103518,-0.1669088006,0.0568864979,0.4013474584,0.1551738977,0.1170737743,-0.1946797669,-0.2683934867,0.0136913834,0.2815442681,-0.0744164065,0.1742078066,0.0707921907,-0.0301680341,-0.1171849221,0.371190846,0.1120216623,-0.4049769938,0.0647225007,0.0489668772,0.1148314252,-0.0608992241,-0.2330337018,0.136604473,0.097282283,-0.6623622179,0.105829224,0.402593255,-0.1487614512,-0.3113524616,-0.0709548816,0.0653429404,-0.1784998327,0.2433368862,0.1851489395,0.2427431047,0.3561433554,0.0626441985,0.3224667907,-0.2490847111,-0.0847934559,-0.381008178,-0.0408868678,-0.0289997775,-0.1580854654,0.1339620054,0.4530848563,0.3085568547,0.0720693171,-0.1417773813,0.3126994073,0.3389642537,-0.0263898242,-0.0527405776,0.0680244789,0.1982621104,0.2358064055,-0.0403654985,0.5818987489,0.1609146297,-0.0876307338,0.0840421468,0.4946383536,-0.4532775283,-0.013118078,-0.0674491301,-0.0155320885,0.1388317198,0.0436688177,-0.1320204288,-0.1433892548,0.4456670284,0.1765436679,0.1647564918,0.0733368993,-0.3982080817,-0.073458977,0.2575798333,0.0217039362,0.1302789897,-0.0767992288,0.0549408756,-0.1026630625,0.0229178928,0.0528387204,0.3336082101,0.1653774828,-0.0692492425,0.1218868271,-0.4862632453,-0.4011192918,-0.1301844567,-0.0406682976,0.1257009953,0.2686919868,0.39283517,0.1780779362,-0.3108200431,-0.1725024879,-0.1780834198,0.0028736275,-0.0867993236,0.262774229,-0.2829509079,0.0638921037,-0.1240324751,-0.2499917597,-0.2210532576,-0.3301238716,0.1189463511,0.2607831359,0.1801377535,-0.0213392973,0.4165262878,0.1643574536,-0.1106036603,0.177942425,-0.1152252406,-0.2789510787,-0.0980238542,0.1053920835,-0.2187730819,0.2054293007,0.2601990104,-0.3080379665,0.0203066058,0.1796343774,-0.2534666061,-0.0686162412,-0.0020722789,0.4098419547,0.1068234742,-0.335804671,-0.2656064034,0.3347665071,0.3430611193,-0.2291763425,-0.1546251178,0.1012243256,-0.0897234976,0.0792596042,-0.0167217255,-0.0430627316,-0.2317102551,-0.2127954215,0.4316241145,0.2908910215,0.0083824014,0.5002599955,-0.0831679404,-0.0073336782,0.137941733,0.2961850166,-0.467445761,-0.0331860483,0.2939333618,-0.3487680256,-0.4366020858,-0.2240982056,0.050590992,0.0724688843,0.0108167669,-0.2337208688,-0.57028234,0.0090026958,-0.0026481207,-0.19519642,-0.1662589312,0.4672875404,-0.2651472092,-0.2121557742,0.0052440828,0.0653482154,-0.0029160671,0.2327388972,0.3163963258,-0.3067874312,0.2171624005,-0.1260423213,0.0643994361,0.3522106111,-0.1832674742,0.1160768121,-0.0247520711,0.3117873371,-0.3168749213,0.0364080109,0.2948443294,0.0857056826,-0.1949567348,0.1709591448,-0.1188688874,0.2867122591,-0.1773141325,-0.0613702387,0.1188722029,-0.1592527479,-0.0122515401,-0.2380353957,-0.220440343,0.0803628266,0.2464231402,-0.0869599432,-0.2009221464,-0.0135544436,0.0842464343,0.0701996684,0.0027534876,-0.0103817564,-0.0720058382,-0.4640107453,0.1893832088,-0.100658983,0.100913845,-0.0456420183,0.0669939071,0.1691402197,-0.0566844903,0.3666310608,-0.2106122077,0.0921635702,0.1756336987,-0.2941549718,-0.4604099691,-0.1159779653,-0.3113110363,0.0444862098,0.2178642005,0.2070884705,-0.1083559766,-0.0799207687,-0.0648076013,0.0359759815,-0.1981047988,-0.0415209197,-0.6293358207,-0.0046169246,0.0181646626,0.1143142954,-0.0526495427,0.2312363535,0.0274748262,-0.0094722463,-0.2461321801,-0.0133181233,0.3121687472,0.4288457036,0.2147683948,0.0606809258,-0.272800833,-0.0105386488,-0.0225608852,-0.3033393323,0.0782152042,0.1013693586,-0.2962467074,0.0424454696,-0.0132967466,0.0357423536,0.2210521251,0.013816095,-0.1438022554,-0.7564556599,0.3295621872,-0.3697873354,0.1255402565,0.3464003503,-0.0126134893,0.0129843475,-0.32890746,0.351390034,0.1373826712,-0.4095774889,0.278074801,-0.3876506686,-0.1274400055,0.4977778494,0.4410946965,0.8089743257,-0.0554032177,0.2344987839,0.3494738936,-0.1319746226,0.5014903545,-0.2265448719,0.1540482938,-0.2458849698,-0.1971984953,0.0298117027,0.0289453957,-0.0411168858,-0.0968406424,-0.1022951007,0.4249106646,0.1097568348,0.0189149659,-0.0572139509,0.0078086467,0.2016165853,-0.0173373502,-0.1541358531,0.2030908167,-0.0818943977,0.1908839494,-0.1752750576,0.0122954315,-0.2148535848,0.0576887988,0.0145755671,-0.1486009508,-0.2444498092,0.2113227397,-0.0724924281,-0.4159750044,-0.3774949908,0.558713913,0.4174341857,0.1943918914,0.0290064719,0.201682657,0.0190362483,0.1178581268,-0.0588446148,-0.0745102912,0.1307758391,-0.0487648062,-0.2427684814,0.0947327837,0.1781242043,-0.2563756108,-0.2358670533,0.164246425,-0.393520087,0.1428208053,0.0131322565,-0.0823320672,-0.0574942157,-0.0706345439,0.1708704382,0.0125719709,-0.0627123266,0.3557574451,0.1462882459,-0.5041589141,-0.1616048962,0.3129502237,-0.0394810885,0.0052437829,-0.1337226331,-0.3825682998,0.04567324,-0.3021409512,-0.1049314141,-0.3436256647,-0.0163989738,0.2843949199,0.0778475478,-0.0667212158,0.2067267299,0.6172465086,0.2040031999,0.2291588038,0.2028275728,-0.3756342828,-0.1011359394,0.026163945,0.2002658397,0.0270954296,0.0410112292,0.2591670454,0.0441957898,0.3328755796,-0.4154178798,-0.2112275064,-0.2728557885,0.0771254525,-0.4033289254,-0.2485918999,-0.1983134747,0.1788430363,0.106834732,0.2859473825,-0.3695852757,-0.2480739653,-0.1829580367,0.0778274387,-0.0146715716,0.256880343,-0.1227301806,0.0897889137,0.0280077681,-0.2528812587,0.0257890224,0.2819306254,0.1865692288,0.1521971524,-0.0954382792,0.4118034542,0.189834848,-0.0141284475,0.064536646,0.030332692,-0.0923135132,-0.0846865401,-0.1575947553,0.0199735705,-0.4911462963,-0.1587667316,-0.1551728696,-0.1030333638,0.037730705,0.0317360535,0.1262182891,-0.2419977337,0.2969405055,0.2501825094,-0.0586201549,-0.3473286033,-0.0301831886,0.2194411308,-0.3312292993,0.2263871431,-0.0857336298,0.2460583001,0.2790098488,0.4026723802,0.2872608006,-0.1761205494,-0.0394820906,-0.1349145621,-0.0936425552,-0.2637962997,0.1667924374,-0.1747069061,0.1136259362,-0.0128397606,0.6192075014,0.0229768306,0.1489230543,0.3894222379,0.1835260242,0.3473848104,-0.3537896872,0.0347736701,-0.1572966129,0.1302668005,-0.0962465554,0.4951853752,-0.0047521233,0.0571498685,0.1610793024,0.2370565087,-0.3910468221,-0.262357831,-0.3800973892,0.0447983555,-0.0752819106,-0.1696825773,-0.2992704809,-0.1049141437,-0.2750798464,-0.0137257613,0.0598349348,0.1909666061,0.2889841795,0.1396646947,-0.0436059758,-0.1941465586,-0.1264562905,0.0102858962,0.4056713283,-0.1645210236,0.0863614753,0.0578821935,0.2172217965,0.0415737629,0.2197797745,0.2097709328,0.3153032064,-0.3565009236,0.1037838608,-0.1926604658,-0.0260035191,-0.0689903274,0.231944859,0.3385539949,-0.1831485331,0.4716710448,0.1637120396,-0.2706426084,0.4996530712,0.1417888403,0.3519830108,-0.2614817619,0.5389919877,0.2331406772,0.231144622,-0.0883341581,0.0677476376,-0.4462493062,0.1519169658,-0.5843315125,-0.0335709676,0.0321124643,-0.3731010556,0.1177594587,-0.0533859245,0.39911744,-0.1048742905,0.1799817681,-0.1166928336,-0.1933872402,-0.6247979403,0.1171365306,-0.119276017,0.011770295,-0.1134250015,0.2710291445,-0.3030306697,0.3159366846,-0.0864443034,0.5827918649,-0.0794930831,0.2144659758,-0.5019690394,0.2297379971,-0.4718154371,-0.1888470352,0.1791081131,-0.0297099203,-0.0602140054,-0.2444480509,0.1195142493,0.0902114213,-0.0805593058,0.0156909563,0.8752211928,0.4947991371,0.1753017604,0.1983550191,-0.1136667654,-0.1404851824,-0.0375549644,-0.1914144754,-0.0399724543,0.1490693539,0.0210790634,0.1228794679,-0.0095520299,0.1311477572,0.0100224027,0.7477493882,0.1075146198,-0.28836447,-0.1580875367,0.1112050861,0.0338586904,0.0951979831,-0.3561339974,0.5702241659,-0.0062336791,0.3806046546,-0.1571126133,-0.3438039124,0.5270280838,-0.1536008269,-0.0558105744,-0.0444962792,0.2838690877,0.19970119,-0.0955431387,-0.3783438206,0.0703424737,0.1069687307,0.0896181315,0.0172652286,0.2902804017,-0.2487468272,-0.0887805372,-0.1904656589,0.1322317272,-0.0104810176,-0.0895540044,0.3145931959,-0.3432698548]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/360","title":"[Feature request] Add dataset.ragged_map() function for many-to-many transformations","comments":"Actually `map(batched=True)` can already change the size of the dataset.\r\nIt can accept examples of length `N` and returns a batch of length `M` (can be null or greater than `N`).\r\n\r\nI'll make that explicit in the doc that I'm currently writing.","body":"`dataset.map()` enables one-to-one transformations. Input one example and output one example. This is helpful for tokenizing and cleaning individual lines.\r\n`dataset.filter()` enables one-to-(one-or-none) transformations. Input one example and output either zero\/one example. This is helpful for removing portions from the dataset.\r\nHowever, some dataset transformations are many-to-many. Consider constructing BERT training examples from a dataset of sentences, where you map `[\"a\", \"b\", \"c\"] -> [\"a[SEP]b\", \"a[SEP]c\", \"b[SEP]c\", \"c[SEP]b\", ...]`\r\n\r\nI propose a more general `ragged_map()` method that takes in a batch of examples of length `N` and return a batch of examples `M`. This is different from the `map(batched=True)` method, which takes examples of length `N` and returns a batch of length `N`, processing individual examples in parallel. I don't have a clear vision of how this would be implemented efficiently and lazily, but would love to hear the community's feedback on this.\r\n\r\nMy specific use case is creating an end-to-end ELECTRA data pipeline. I would like to take the raw WikiText data and generate training examples from this using the `ragged_map()` method, then export to TFRecords and train quickly. This would be a reproducible pipeline with no bash scripts. Currently I'm relying on scripts like https:\/\/github.com\/google-research\/electra\/blob\/master\/build_pretraining_dataset.py, which are less general.\r\n\r\n","comment_length":42,"text":"[Feature request] Add dataset.ragged_map() function for many-to-many transformations\n`dataset.map()` enables one-to-one transformations. Input one example and output one example. This is helpful for tokenizing and cleaning individual lines.\r\n`dataset.filter()` enables one-to-(one-or-none) transformations. Input one example and output either zero\/one example. This is helpful for removing portions from the dataset.\r\nHowever, some dataset transformations are many-to-many. Consider constructing BERT training examples from a dataset of sentences, where you map `[\"a\", \"b\", \"c\"] -> [\"a[SEP]b\", \"a[SEP]c\", \"b[SEP]c\", \"c[SEP]b\", ...]`\r\n\r\nI propose a more general `ragged_map()` method that takes in a batch of examples of length `N` and return a batch of examples `M`. This is different from the `map(batched=True)` method, which takes examples of length `N` and returns a batch of length `N`, processing individual examples in parallel. I don't have a clear vision of how this would be implemented efficiently and lazily, but would love to hear the community's feedback on this.\r\n\r\nMy specific use case is creating an end-to-end ELECTRA data pipeline. I would like to take the raw WikiText data and generate training examples from this using the `ragged_map()` method, then export to TFRecords and train quickly. This would be a reproducible pipeline with no bash scripts. Currently I'm relying on scripts like https:\/\/github.com\/google-research\/electra\/blob\/master\/build_pretraining_dataset.py, which are less general.\r\n\r\n\nActually `map(batched=True)` can already change the size of the dataset.\r\nIt can accept examples of length `N` and returns a batch of length `M` (can be null or greater than `N`).\r\n\r\nI'll make that explicit in the doc that I'm currently writing.","embeddings":[-0.5117926598,-0.064943336,0.0467378795,-0.2127768099,-0.0525975712,-0.0972506702,0.2488025129,0.3504051864,-0.2492751777,0.0586383492,0.193172425,0.4199866951,-0.4208355546,-0.1185231209,0.3404656947,-0.2591928542,0.1204847023,-0.071868211,-0.0588385314,-0.0770459324,-0.3946816921,-0.0967775807,-0.251825273,-0.0019553835,-0.0170704108,-0.0298995767,-0.1776050031,-0.3150987327,0.1913707405,-0.026635265,0.1828670651,0.3647984862,-0.0834427327,0.198416695,-0.0001292811,-0.190408498,0.0921218768,-0.0385715701,-0.224856481,-0.1235299334,-0.0980342552,-0.1194367781,-0.1217392385,-0.1869545132,0.1676254869,0.0402919054,0.2579863667,-0.3276602328,0.4382550716,0.1333797723,-0.058807455,0.0316222496,-0.0949261263,0.0337903313,0.509775579,0.4657484889,-0.0328059047,-0.285799861,0.5987883806,-0.2376432717,-0.0481912158,0.3115279675,-0.0568308569,-0.0095256111,0.2963942289,-0.0102193793,-0.1644727737,-0.2860151827,-0.1781579256,0.6859874129,-0.1074637547,-0.3305649459,-0.5636526942,-0.3081480265,0.1056257188,-0.2365754098,-0.4840701222,0.031682048,-0.1258307993,-0.04669315,-0.8150773048,-0.1063044965,-0.3973284662,0.3978858888,-0.162531063,0.1373672634,0.2568991184,0.2220887244,0.1073892564,-0.2615977824,-0.1022015661,-0.1448865384,0.2572753727,0.4562511742,-0.0917671844,-0.3255052567,0.0890367627,-0.0486535728,0.550316155,0.0820575207,-0.00353126,0.1841469854,-0.2206216902,-0.0259346366,-0.028433146,-0.005363218,0.0318878628,0.5232826471,0.1720310748,-0.0608002618,-0.0620665215,0.3738218546,-0.0533123016,-0.1968531162,-0.1862699091,0.2287461907,-0.1539468169,0.1261386871,-0.0783340186,-0.2251610458,-0.2609084845,-0.0259206407,0.0960605145,0.1137732565,-0.0592957027,-0.0202015359,0.1481970996,0.118082881,0.0345744863,-0.0761324316,0.0366137736,-0.1324026436,-0.3662855625,0.1333995163,0.3161870241,0.1634283215,0.035779465,-0.0900668651,0.0493343472,0.3958720863,0.0317810066,-0.1015291959,0.4242266417,-0.1063067615,-0.3029841185,0.0902887806,0.0502145141,0.2103304714,-0.3275679946,0.1787117124,-0.5758726597,-0.5673735738,0.2367747873,0.0075080106,0.088310495,-0.0259559434,-0.1396830231,0.4832127392,0.3983263075,-0.3995802104,-0.0187349375,-0.4631467462,-0.3660453558,-0.0446320809,0.1977967173,0.3899159133,-0.0647996515,-0.1040581763,-0.1227666587,0.2528773546,0.2617071569,0.2549603283,-0.3634381294,0.3214164972,-0.0104309842,0.0024093937,0.5380627513,-0.1634080559,-0.0462382026,0.4684036672,-0.0806689039,0.1257233024,0.0206975732,-0.0552747622,0.630387485,-0.2714833617,0.5482940674,0.495585382,-0.2449007332,0.2318511903,0.1040629372,-0.2603802383,0.1129255816,0.1038386598,0.1581514329,-0.157986775,-0.0919636264,0.1619398892,0.2925562561,-0.3035011292,0.1494297534,-0.2504131198,-0.2802056968,0.0837863535,-0.1690109819,-0.2650311589,-0.4043860137,-0.0532115959,-0.1862901747,-0.1181837693,0.3396584988,-0.3409480751,0.3310184479,-0.0361087322,0.1190072149,0.0143996514,-0.0951026827,0.0769227296,0.0848621503,0.0141277006,-0.143578425,-0.3389590383,-0.0538143329,0.1589299142,0.0072393846,0.0650214627,0.3833913803,0.0107543766,-0.2258309573,0.2299621552,-0.0129864849,-0.2821635306,0.2307552844,0.2818350494,0.1516554207,-0.0679395497,-0.1761771142,0.3548897505,0.4281879961,0.1057404578,0.1138768941,0.2418976873,0.1026602164,-0.2012156546,-0.5654881597,0.3278254271,0.0159291048,0.5668686628,0.2172817886,0.0367184393,-0.2614708543,0.0843571872,-0.39247787,-0.1233014762,0.0642563999,-0.0924340263,0.0619214326,-0.1779865474,-0.1098280549,0.1385016441,0.3341235816,-0.0623130761,0.0160424057,0.4599877596,-0.0083584832,-0.1453412324,-0.1426701397,0.074156642,0.2954389751,0.1929785013,0.0457330942,-0.1567011923,0.0330740884,0.0837342963,0.215299055,0.1570747346,0.0650497451,0.125674516,0.1853105426,0.1826387197,0.0090827541,-0.5682793856,0.1882546246,-0.0079583153,-0.2539595962,0.131059587,-0.0517422855,-0.074928619,-0.3282676935,-0.380648464,0.2121582627,-0.2187453359,0.2884810269,-0.1282152534,-0.296951592,0.2772268057,0.4369253218,0.4191134572,0.0573175102,-0.2148327678,-0.0691943541,-0.2195558697,-0.0100067463,-0.1027998999,0.0953414962,-0.2313009351,0.31728369,0.0528557301,-0.0300484616,-0.403062433,-0.3596819043,-0.0422480777,-0.1454541534,-0.0949648321,0.4797690809,0.0653494895,-0.0845193639,-0.2374870777,0.0989018232,-0.3198269904,0.0672237352,-0.0130907139,0.0672867745,0.0276809167,-0.0035394682,-0.1749517918,-0.0653904378,-0.1445049196,0.285490036,-0.2513404787,0.1826523244,0.0410832167,0.2352617234,-0.0113291265,-0.0986149311,-0.1638326943,0.3204434216,-0.0722717419,0.1216054708,-0.2969004214,-0.0577400401,-0.3848017454,0.0147551904,-0.4168524742,0.8833577037,-0.2570570409,0.06533847,-0.010964782,0.114447847,-0.0286804121,0.1144721881,0.459435761,0.1734045744,0.0842865109,-0.0753118172,0.0018325129,0.0628118664,0.0764713213,-0.0132882753,0.1676447541,0.3107929826,0.1382342875,1.0362778902,0.4065803289,-0.1977274418,-0.1237772778,-0.0955225676,-0.1306502968,-0.0942156538,-0.0773937255,0.2283999175,-0.1027041525,0.0158002656,0.264061898,0.0193353035,-0.2913355529,0.0927309021,0.2628012598,-0.062993139,-0.3514671326,0.4311914742,-0.2534629703,0.3162635267,-0.0628938004,-0.0218245182,-0.4590005279,-0.0450870469,0.1393495649,0.0284789652,0.4992492199,0.0089775408,-0.1856807321,-0.1942503899,-0.1743834466,0.2592366338,0.3169304132,-0.1696095467,-0.0471475683,-0.270478189,-0.1804647446,0.0408409461,0.6683760881,-0.4042588174,-0.0944759846,-0.0640063584,-0.1411257833,-0.3563019037,0.0312466938,-0.2481983751,0.0697445869,-0.0198519211,0.49482131,-0.3880448937,-0.1199446321,0.5387397408,0.2025344968,0.0707616135,0.2278278023,0.3494002819,0.0512491353,-0.2930174768,0.0461546071,0.0083861556,-0.1381513327,-0.0573400743,-0.1563096493,-0.2667965889,0.1851998121,0.1792492419,0.0469414257,0.0365433507,0.1741307527,0.2474831641,-0.1948188245,-0.0388820022,0.5333442688,-0.0232817028,-0.5298101306,-0.2901954949,-0.2957293987,0.093487151,0.4838528633,0.3759437799,0.1037957072,-0.2002668679,0.0990797728,0.344466269,-0.1938913316,0.3590343893,0.2385072857,-0.0046016001,-0.2231168002,-0.6674753428,0.288518548,0.3396402597,-0.1539263576,0.4999477565,-0.1128187925,-0.4338577986,0.4376761317,0.2138666809,0.860599041,-0.0930364653,0.4288926721,0.0714304894,0.2922262549,0.035597533,-0.3562489152,0.1745380014,-0.0388236083,0.0377970487,-0.0339293554,0.0631567538,-0.065637514,0.3533826768,-0.3637475967,0.1785397381,0.16646415,0.0536408871,-0.1254020482,-0.1496284455,0.1231345087,-0.4545201659,0.1960609853,-0.0737136826,0.1805967987,-0.0163821559,-0.0644321516,-0.2170467973,0.0175487734,-0.1197939292,-0.4037570357,-0.1684028953,-0.4634839296,0.1180835441,-0.1645263135,-0.3696554303,0.1360815018,0.3140815496,-0.0894626006,0.2606812418,-0.0109714046,-0.0849465951,0.053175509,0.1234388798,-0.1126028076,-0.4361713827,0.4604460895,0.0241079126,0.0635959804,0.1180090904,-0.0135911442,0.0203636102,-0.2320082337,-0.1347922087,0.6238181591,-0.1697426289,0.0811618418,0.4407758117,-0.2101944089,-0.1649959087,-0.0818400383,-0.0821775496,-0.0403970145,0.4395231009,-0.177764833,-0.0196162593,0.0887481645,0.1774638891,-0.143002674,-0.2058619708,0.2194245458,0.0347097963,0.0017068953,-0.2143192291,0.4186762869,0.34126243,-0.5785907507,0.2002499402,-0.0755247921,0.1383098364,0.1926778406,0.4940732419,0.0020638062,0.2198422104,-0.105129905,-0.1493909359,-0.0877563432,0.0988139808,0.0980292186,0.3579158783,-0.2643957138,0.0938266143,0.1101917848,0.2262541205,-0.1430218965,0.1302990466,0.1500340402,0.3245951831,-0.1469014734,0.1860022098,-0.3450327516,0.1204134822,-0.145729959,0.0248902328,-0.0870718732,-0.0339143127,-0.1341600716,0.2138228416,0.0136191845,-0.0126302922,0.266287595,-0.006508057,-0.2293902636,-0.4863310158,0.4170761406,-0.0988717601,-0.1932519525,0.0323353857,0.2872149944,0.1487859488,0.0434741154,0.2304131538,0.1464752257,0.1685269475,-0.1405680478,0.1705088764,0.0516397431,-0.3247896731,0.0742109641,0.2445800304,0.4982964098,0.1670572907,0.3701964319,-0.0455544367,-0.2654011846,0.2559414506,0.0607979111,0.3003189266,-0.2145725638,0.089118734,0.2506799996,0.0432809517,-0.1449576616,0.0008068595,0.1159375533,-0.1271288246,0.140929535,0.1531627327,-0.3267061412,-0.1371109039,-0.0099196834,-0.1005174667,-0.0054874234,-0.2498066574,0.1589382142,0.5841220617,-0.0964713022,0.0054836809,-0.3491958976,0.2093930393,0.307967037,-0.1342761219,0.062469095,0.7259107828,0.1867705882,0.0445107333,0.3982729316,-0.2389344573,0.0379587226,0.1182677075,0.0814309791,0.4266923666,0.2245128453,0.0289221816,-0.0516347811,-0.2223332524,-0.1780666411,0.3308289349,0.2070915103,-0.2560718656,-0.1284089535,-0.1608002782,0.1912703365,-0.0682940334,-0.2282555699,-0.0121855466,0.6396490932,0.1894582212,0.4283140898,0.0356008969,-0.2567301393,-0.2338404655,0.4714787304,-0.2450587749,-0.118325606,0.1864346862,-0.1090712994,-0.0430827737,0.3037449419,0.373842001,0.2554517984,-0.2062002867,0.3856816292,-0.3402126431,0.1566354185,0.0453633107,0.1711402237,0.0649492219,-0.2425818145,0.2798639238,-0.1120095626,-0.0136456471,0.1677275598,-0.1319573969,-0.0159598142,-0.2207806706,0.091425702,0.2969560921,-0.2220663875,-0.0637007952,0.1455152184,-0.0826904103,0.1635764986,0.2586451471,-0.3232423365,0.0181079786,0.2209970951,-0.0079622883,0.078527838,0.4866625667,0.0731781349,0.0823518112,-0.3261429071,-0.3485638499,-0.0445375219,-0.2237143964,-0.1492238492,0.0782305449,-0.0020910054,0.0520903878,0.1542301476,0.3146726489,-0.0017990893,-0.7042708993,-0.1145151258,0.5776193142,-0.1748529077,-0.1160906255,0.1270684153,0.1364876181,0.1195360497,-0.3580099642,0.126615271,-0.2097089738,-0.1929229647,-0.0605751574,0.157268092,0.3012132943,0.0756626502,0.1441108137,0.1144430786,0.110820137,0.0371249691,0.2231857628,-0.112659961,0.1721082032,-0.0361610986,0.2040284127,0.1006088629,0.0013342027,-0.2895727158,-0.3310934305,-0.0419561937,-0.383117795,0.2911996841,0.0190231949,-0.1299679726,-0.0623690933,-0.0453004912,-0.0445536599,0.1269276291,0.4295780361,-0.1205238178,-0.2004196793,0.0462114438,-0.2601973712,0.3955245316,-0.4662588537,-0.3204602897,-0.4795957804,0.1678976119,-0.4146011472,0.2287065238,-0.4381655455,-0.1076838672,0.4040952623,-0.003203626,-0.408266902,0.2361128926,0.1241551638,-0.2943826616,-0.2554002702,0.3656482995,0.0122955963,-0.2283165902,-0.1586784124,-0.6390990019]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/360","title":"[Feature request] Add dataset.ragged_map() function for many-to-many transformations","comments":"You're two steps ahead of me :) In my testing, it also works if `M` < `N`.\r\n\r\nA batched map of different length seems to work if you directly overwrite all of the original keys, but fails if any of the original keys are preserved.\r\n\r\nFor example,\r\n```python\r\n# Create a dummy dataset\r\ndset = load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")[\"test\"]\r\ndset = dset.map(lambda ex: {\"length\": len(ex[\"text\"]), \"foo\": 1})\r\n\r\n# Do an allreduce on each batch, overwriting both keys\r\ndset.map(lambda batch: {\"length\": [sum(batch[\"length\"])], \"foo\": [1]})\r\n# Dataset(schema: {'length': 'int64', 'foo': 'int64'}, num_rows: 5)\r\n\r\n# Now attempt an allreduce without touching the `foo` key\r\ndset.map(lambda batch: {\"length\": [sum(batch[\"length\"])]})\r\n# This fails with the error message below\r\n```\r\n\r\n```bash\r\n File \"\/path\/to\/nlp\/src\/nlp\/arrow_dataset.py\", line 728, in map\r\n arrow_schema = pa.Table.from_pydict(test_output).schema\r\n File \"pyarrow\/io.pxi\", line 1532, in pyarrow.lib.Codec.detect\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named foo expected length 1 but got length 2\r\n```\r\n\r\nAdding the `remove_columns=[\"length\", \"foo\"]` argument to `map()` solves the issue. Leaving the above error for future visitors. Perfect, thank you!","body":"`dataset.map()` enables one-to-one transformations. Input one example and output one example. This is helpful for tokenizing and cleaning individual lines.\r\n`dataset.filter()` enables one-to-(one-or-none) transformations. Input one example and output either zero\/one example. This is helpful for removing portions from the dataset.\r\nHowever, some dataset transformations are many-to-many. Consider constructing BERT training examples from a dataset of sentences, where you map `[\"a\", \"b\", \"c\"] -> [\"a[SEP]b\", \"a[SEP]c\", \"b[SEP]c\", \"c[SEP]b\", ...]`\r\n\r\nI propose a more general `ragged_map()` method that takes in a batch of examples of length `N` and return a batch of examples `M`. This is different from the `map(batched=True)` method, which takes examples of length `N` and returns a batch of length `N`, processing individual examples in parallel. I don't have a clear vision of how this would be implemented efficiently and lazily, but would love to hear the community's feedback on this.\r\n\r\nMy specific use case is creating an end-to-end ELECTRA data pipeline. I would like to take the raw WikiText data and generate training examples from this using the `ragged_map()` method, then export to TFRecords and train quickly. This would be a reproducible pipeline with no bash scripts. Currently I'm relying on scripts like https:\/\/github.com\/google-research\/electra\/blob\/master\/build_pretraining_dataset.py, which are less general.\r\n\r\n","comment_length":179,"text":"[Feature request] Add dataset.ragged_map() function for many-to-many transformations\n`dataset.map()` enables one-to-one transformations. Input one example and output one example. This is helpful for tokenizing and cleaning individual lines.\r\n`dataset.filter()` enables one-to-(one-or-none) transformations. Input one example and output either zero\/one example. This is helpful for removing portions from the dataset.\r\nHowever, some dataset transformations are many-to-many. Consider constructing BERT training examples from a dataset of sentences, where you map `[\"a\", \"b\", \"c\"] -> [\"a[SEP]b\", \"a[SEP]c\", \"b[SEP]c\", \"c[SEP]b\", ...]`\r\n\r\nI propose a more general `ragged_map()` method that takes in a batch of examples of length `N` and return a batch of examples `M`. This is different from the `map(batched=True)` method, which takes examples of length `N` and returns a batch of length `N`, processing individual examples in parallel. I don't have a clear vision of how this would be implemented efficiently and lazily, but would love to hear the community's feedback on this.\r\n\r\nMy specific use case is creating an end-to-end ELECTRA data pipeline. I would like to take the raw WikiText data and generate training examples from this using the `ragged_map()` method, then export to TFRecords and train quickly. This would be a reproducible pipeline with no bash scripts. Currently I'm relying on scripts like https:\/\/github.com\/google-research\/electra\/blob\/master\/build_pretraining_dataset.py, which are less general.\r\n\r\n\nYou're two steps ahead of me :) In my testing, it also works if `M` < `N`.\r\n\r\nA batched map of different length seems to work if you directly overwrite all of the original keys, but fails if any of the original keys are preserved.\r\n\r\nFor example,\r\n```python\r\n# Create a dummy dataset\r\ndset = load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")[\"test\"]\r\ndset = dset.map(lambda ex: {\"length\": len(ex[\"text\"]), \"foo\": 1})\r\n\r\n# Do an allreduce on each batch, overwriting both keys\r\ndset.map(lambda batch: {\"length\": [sum(batch[\"length\"])], \"foo\": [1]})\r\n# Dataset(schema: {'length': 'int64', 'foo': 'int64'}, num_rows: 5)\r\n\r\n# Now attempt an allreduce without touching the `foo` key\r\ndset.map(lambda batch: {\"length\": [sum(batch[\"length\"])]})\r\n# This fails with the error message below\r\n```\r\n\r\n```bash\r\n File \"\/path\/to\/nlp\/src\/nlp\/arrow_dataset.py\", line 728, in map\r\n arrow_schema = pa.Table.from_pydict(test_output).schema\r\n File \"pyarrow\/io.pxi\", line 1532, in pyarrow.lib.Codec.detect\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named foo expected length 1 but got length 2\r\n```\r\n\r\nAdding the `remove_columns=[\"length\", \"foo\"]` argument to `map()` solves the issue. Leaving the above error for future visitors. Perfect, thank you!","embeddings":[-0.5140849352,-0.0638845786,0.0467357934,-0.2271822691,-0.0690640658,-0.0551058464,0.2775110304,0.3193071485,-0.2335919738,0.0517762713,0.1853372306,0.433014214,-0.3891707063,-0.156760633,0.3206688166,-0.2123526037,0.1380255073,-0.0705325082,-0.0471089743,-0.0676657781,-0.4131712615,-0.0813822672,-0.3076458871,-0.0193038564,0.0158393364,-0.0517234616,-0.1688470989,-0.3216543794,0.1985131055,-0.0572382286,0.1546299011,0.3466903567,-0.0777745917,0.2024631649,-0.0001282075,-0.183033362,0.0852343738,-0.0660537779,-0.2778266966,-0.1043357328,-0.1222465485,-0.0921766236,-0.1874970943,-0.1921699494,0.1732089221,0.0139108766,0.2294358611,-0.3430962861,0.4886731803,0.1127876192,-0.0509738103,0.0658145025,-0.0822368413,0.0404130407,0.4733759463,0.4220072329,-0.0335137136,-0.256085813,0.5853219628,-0.223612532,-0.0561208613,0.3167016804,-0.0640675277,-0.0125534339,0.2690361738,0.0069327224,-0.1276357174,-0.278216213,-0.1596962959,0.668779254,-0.1796213984,-0.3169804215,-0.5546558499,-0.3592832386,0.0945755467,-0.2510692477,-0.4691944718,0.0308720991,-0.1524645239,-0.0594953187,-0.848056078,-0.0725239664,-0.3502874374,0.3833610415,-0.1656908244,0.157326296,0.2363447398,0.2193863541,0.1254206747,-0.2635751665,-0.1169262007,-0.1893372387,0.2012728453,0.4865551293,-0.0621958971,-0.3287470937,0.0657368004,-0.0454892628,0.5274835825,0.0178330969,0.0014931947,0.1767750829,-0.1448729336,-0.0563268773,-0.02902597,0.0129228448,0.0860021412,0.5244885683,0.183690235,-0.0710849985,-0.0734400898,0.3761557639,0.001037673,-0.1939128339,-0.1375450641,0.2832209468,-0.1539544463,0.1606362313,-0.0712326467,-0.2117683589,-0.2780438662,-0.0161033422,0.0704892129,0.1073955446,-0.0445260406,-0.1205081493,0.1225223318,0.1043268889,0.0342733189,-0.0758775398,0.0335377976,-0.1791582555,-0.3669870794,0.112219952,0.3516450524,0.1253017336,0.0558127873,-0.0914677531,0.0226247273,0.3612656593,-0.0332889147,-0.0695660934,0.4820078313,-0.1084303185,-0.2738676667,0.1136994064,0.065422304,0.1628779471,-0.3124417365,0.1807008833,-0.5805314183,-0.4811038375,0.3143530786,0.024119895,0.0556711778,-0.0127374465,-0.2111569047,0.4394823909,0.3932487965,-0.3599758148,0.0097447149,-0.4304398894,-0.3189178705,-0.0402075797,0.1983999163,0.4012480378,-0.0779255331,-0.139183864,-0.0653007403,0.2619438171,0.2537104487,0.272292763,-0.3692961335,0.329579711,-0.0330144539,0.014164594,0.5044759512,-0.1357091367,-0.0761539638,0.4683811665,-0.1021965593,0.1658272296,-0.0335429013,-0.0839251727,0.6721920371,-0.3023916185,0.5716998577,0.4612060487,-0.2615791559,0.2165637463,0.0959388763,-0.2506913841,0.057170406,0.0597375743,0.17809847,-0.1654473394,-0.1107605547,0.1469749212,0.2859001458,-0.2964674532,0.1687088162,-0.2055943012,-0.3021275103,0.0792605057,-0.1763586402,-0.2184639126,-0.4042701423,-0.0012166777,-0.2266483456,-0.1566907614,0.2583780885,-0.3116954267,0.347312361,-0.02537046,0.0528371818,-0.0055875718,-0.0706905425,0.0335752405,0.122445479,0.0577409007,-0.1151438132,-0.3456280529,0.0082599223,0.1561136544,0.0033235655,0.050349161,0.3568118215,0.0058248593,-0.2318897247,0.2177124023,0.0298403669,-0.3160478473,0.2320103347,0.2861753106,0.1013383418,-0.0644035041,-0.2576484382,0.3316794634,0.4543642104,0.1182983294,0.0973233506,0.2350211143,0.1191035733,-0.214691028,-0.5587961078,0.3671116829,0.0026506779,0.5796784759,0.1737111509,0.0429742821,-0.2740180194,0.0838133618,-0.4269851744,-0.1542911083,0.0699479356,-0.1255001128,0.0179762356,-0.1776826978,-0.0578427464,0.1229690686,0.2958962619,-0.086525254,0.0253415611,0.3899826109,-0.0557599403,-0.1243669018,-0.162128076,0.0794301555,0.2992251515,0.1775403321,0.0780121312,-0.170812428,0.0283878185,0.0630374551,0.2288245857,0.1442825347,0.0305295512,0.156362623,0.1670929193,0.17812711,-0.0295912512,-0.507416904,0.2409530729,-0.0365322381,-0.2441274673,0.1146547943,-0.0737815127,-0.0787590295,-0.3204546273,-0.3741789162,0.2363279462,-0.2455091327,0.2917232513,-0.0937588289,-0.2794338763,0.3288775086,0.4729547799,0.373539567,0.0670037195,-0.2365320325,-0.0940204114,-0.2538489699,-0.0075246468,-0.0984296054,0.0669919476,-0.2963275015,0.3102849424,0.0491142683,-0.0590434112,-0.4437866807,-0.3047181368,-0.031240372,-0.1708153635,-0.0816821158,0.4561350942,0.0946931466,-0.1012774557,-0.2439613044,0.1165831238,-0.2921503186,0.033107996,0.0019354714,0.0883976817,-0.0165858939,-0.0104008699,-0.1024866402,-0.0310778525,-0.1095343307,0.2773324549,-0.3038092554,0.1510036737,0.1006516367,0.2372594923,-0.040225368,-0.1226488426,-0.217658788,0.2773085833,-0.0570498444,0.1644926518,-0.3148436844,-0.0804291517,-0.3558786511,0.0080240862,-0.4721953869,0.8751971126,-0.2329043746,0.0979556218,0.0061029769,0.1651857793,0.0104060406,0.1134646833,0.4450578392,0.1565902978,0.0744688809,-0.0869632214,0.0531915091,0.0861222893,0.0463439375,0.0251018163,0.1552979201,0.3127292395,0.1483407021,1.0749411583,0.4066047072,-0.2169663608,-0.1586939394,-0.1032632589,-0.1172735468,-0.1119469106,-0.1229794025,0.1862343848,-0.1086738035,0.0446408167,0.2721088529,-0.0108469343,-0.3270231783,0.1042990908,0.3022099733,-0.0696531087,-0.3931072652,0.4112078547,-0.2923818231,0.3231608868,-0.0093752295,0.01823592,-0.4317817688,-0.0527660586,0.17727907,-0.0218271855,0.5039941669,-0.0004381533,-0.217068702,-0.1989746541,-0.2317492813,0.2421082556,0.3034308851,-0.1031006351,-0.0184445698,-0.2629833519,-0.1854943484,0.044402428,0.6822598577,-0.4207910299,-0.0235353094,-0.0587054268,-0.1007908285,-0.3538382351,-0.0056988262,-0.2146229893,0.0614530556,-0.0185948797,0.5420695543,-0.3696386218,-0.1054530069,0.4996542931,0.2015981674,0.0292028785,0.2182332873,0.3442276716,0.0626822859,-0.2690508962,0.0428277329,0.0618405789,-0.1242794693,-0.0376532599,-0.125259161,-0.2955954671,0.1893558055,0.1836909801,0.0336560607,0.0484157801,0.2226525694,0.3056958914,-0.1597019732,-0.0187753215,0.5131608844,-0.0489232354,-0.5048871636,-0.2913348079,-0.3084584177,0.0604871586,0.430306524,0.3761101663,0.0962697119,-0.1676415503,0.1000352055,0.3084876835,-0.2345796674,0.2947369516,0.1828976572,0.0257245265,-0.2415442765,-0.6020063758,0.2622419298,0.3142386675,-0.1863298267,0.4721348584,-0.1496134549,-0.4798114002,0.4604497552,0.2156080157,0.7913640738,-0.0752829537,0.4272469878,0.0521432646,0.321987927,0.0058775726,-0.3096720278,0.1915813684,-0.0336151533,0.0491670482,-0.0364149623,0.0795576349,-0.0201458707,0.3469963372,-0.3440819681,0.1437401175,0.1949300617,0.0461235754,-0.140245989,-0.1863615215,0.1465869993,-0.4494156837,0.2331820875,-0.0641707927,0.2282533497,0.0323419347,-0.0666129813,-0.1915911138,-0.0098837167,-0.157910049,-0.4218663871,-0.1853283495,-0.4894335568,0.1597470939,-0.1500019878,-0.3506145179,0.1016644388,0.3187216222,-0.0347137973,0.2399649024,-0.0003744903,-0.0902286842,0.1099749058,0.1146919876,-0.1349407732,-0.4417134225,0.3896439373,0.0380558632,0.1116053089,0.1235311925,-0.0229668636,0.0129075302,-0.251973629,-0.0973585695,0.5527216196,-0.1518054307,0.0738602355,0.4275228083,-0.2010427713,-0.1576443017,-0.0879363492,-0.0958958343,-0.0440952182,0.4851010144,-0.1949032992,-0.0248712394,0.0959987938,0.2647047043,-0.1766144931,-0.2013468742,0.2617540061,0.0356639102,-0.0086548114,-0.2109425813,0.4334567487,0.3232331872,-0.5678303838,0.1925442368,-0.1228358373,0.1622717381,0.214834854,0.55618608,0.0189171024,0.1946834326,-0.0793145746,-0.1670255214,-0.064718537,0.0995929763,0.1461919099,0.3351038396,-0.2218913436,0.1050136015,0.1111062467,0.2339682877,-0.1464440823,0.1537613273,0.1523100883,0.3649273217,-0.1695054322,0.1577084363,-0.3202649057,0.0994031876,-0.1308975816,0.0763147399,-0.0747133791,-0.0521823503,-0.1478397846,0.1980329752,-0.0132215628,-0.0067333216,0.2309432775,-0.0070654303,-0.1956614256,-0.4530231655,0.4037022889,-0.0532202609,-0.1667290777,0.0213420969,0.3260833621,0.1886412501,0.0462094881,0.2146919519,0.0915464759,0.1807409972,-0.1452656239,0.1620860249,0.046580039,-0.2829186916,0.0381367691,0.2388537973,0.449277848,0.1272949725,0.4130896628,-0.1120687872,-0.2644169331,0.2797034979,0.0649406239,0.3199543059,-0.2407997847,0.0406464301,0.2759259939,0.0551873147,-0.1661801189,-0.0041981325,0.1194326878,-0.1092994958,0.1878078282,0.1554257423,-0.3166798651,-0.1522838026,-0.035422124,-0.116430454,0.0390457548,-0.2255422175,0.142763555,0.5783985853,-0.0256984141,0.0035863365,-0.3377917111,0.2078521848,0.3080685437,-0.151379779,0.0317840427,0.6774339676,0.2314914912,0.0745761469,0.3520646095,-0.2667876184,0.0436642952,0.1050195098,0.0488186441,0.4545572102,0.2442039549,0.0040772269,-0.0732301548,-0.2370629758,-0.1540704519,0.3648970425,0.1767337918,-0.3124366105,-0.1710822284,-0.1617449075,0.2002430111,-0.0611118749,-0.2140655518,-0.0100788372,0.6305127144,0.1756513566,0.4212464988,0.0368881375,-0.2943111956,-0.2349816263,0.4713162184,-0.2673817575,-0.1262574345,0.1309455037,-0.0781082585,-0.0403154604,0.318654716,0.3682189286,0.2398973405,-0.197065115,0.4103486836,-0.362278372,0.1330965906,0.0303984024,0.2129995078,0.0660050064,-0.2545666397,0.3269317448,-0.0984086096,-0.0080633033,0.1757559329,-0.0643038526,-0.032188464,-0.1354912668,0.0698088109,0.2563369572,-0.2120691687,-0.0113364412,0.1224327311,-0.0456161126,0.1632272899,0.2427403033,-0.3365352154,0.050639268,0.2127824724,-0.0005314453,0.0612336881,0.5081930757,0.1196383685,0.0469738953,-0.3565030694,-0.3246795237,-0.0701638684,-0.281175524,-0.2140506059,0.0895973817,-0.0327570662,0.0900260955,0.1760786027,0.3353978992,-0.0176699962,-0.650146246,-0.1359869093,0.6674085259,-0.153907001,-0.0999753401,0.1910440028,0.1651743054,0.0985708982,-0.3816342652,0.1492694765,-0.184280768,-0.1660046726,-0.0539837256,0.143378526,0.285128504,0.0482207239,0.1643342972,0.1412635893,0.0911542401,0.0603119582,0.2097205371,-0.1240128428,0.1957070529,0.0016568518,0.1437959969,0.1497963071,0.0080006309,-0.305138737,-0.3161023259,-0.0174160264,-0.3985348642,0.3030865788,0.0144698806,-0.1911913604,-0.0300499015,-0.038428165,-0.0068732831,0.1197865978,0.4581579566,-0.1326551884,-0.1447971612,0.0447073095,-0.2986310124,0.3907758594,-0.5670872331,-0.3139787316,-0.4490803182,0.1355700195,-0.3679909706,0.2406044304,-0.3750258982,-0.1169377342,0.4249212444,-0.0524976142,-0.3907633126,0.196900934,0.1074797437,-0.2897548974,-0.2804971039,0.3630032241,0.0619708486,-0.2455327064,-0.1476063281,-0.6292204261]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/359","title":"ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures","comments":"Hi, it depends on what it is in your `dataset_builder.py` file. Can you share it?\r\n\r\nIf you are just loading `json` files, you can also directly use the `json` script (which will find the schema\/features from your JSON structure):\r\n\r\n```python\r\nfrom nlp import load_dataset\r\nds = load_dataset(\"json\", data_files=rel_datafiles)\r\n```","body":"I tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-23-9aecfbee53bd> in <module>\r\n 55 from nlp import load_dataset\r\n 56 \r\n---> 57 ds = load_dataset(\"..\/text2struct\/model\/dataset_builder.py\", data_files=rel_datafiles)\r\n 58 \r\n 59 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: string> nor list<item: string>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\r\n```\r\n\r\nIf I create the dataset imperatively, using a pyarrow table, the dataset is created correctly. If I override the `_prepare_split` method to avoid calling the validate schema, the dataset can load as well. ","comment_length":49,"text":"ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures\nI tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-23-9aecfbee53bd> in <module>\r\n 55 from nlp import load_dataset\r\n 56 \r\n---> 57 ds = load_dataset(\"..\/text2struct\/model\/dataset_builder.py\", data_files=rel_datafiles)\r\n 58 \r\n 59 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: string> nor list<item: string>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\r\n```\r\n\r\nIf I create the dataset imperatively, using a pyarrow table, the dataset is created correctly. If I override the `_prepare_split` method to avoid calling the validate schema, the dataset can load as well. \nHi, it depends on what it is in your `dataset_builder.py` file. Can you share it?\r\n\r\nIf you are just loading `json` files, you can also directly use the `json` script (which will find the schema\/features from your JSON structure):\r\n\r\n```python\r\nfrom nlp import load_dataset\r\nds = load_dataset(\"json\", data_files=rel_datafiles)\r\n```","embeddings":[-0.3260453045,0.3610112369,-0.0600554347,0.5486999154,0.0407031141,-0.1400738508,0.3431124687,0.4284070134,0.2143322676,-0.1139701083,-0.0436313152,0.3332704902,0.2364399433,0.0587036684,-0.2658620179,-0.1656162292,-0.0196319576,0.2937225401,-0.124623999,0.3482965231,-0.222588554,0.4523977935,-0.1000257656,0.1312664449,0.0139916008,-0.0840229094,0.0551255234,0.3630198836,-0.1661049724,-0.5221460462,0.261677146,-0.297565192,-0.1071128175,0.4265525639,-0.0001121765,0.283685714,0.2796024978,-0.1050383896,-0.367891252,-0.0563196056,0.0219089277,-0.1897962093,0.080716528,-0.1802360266,0.2029777616,-0.2733659148,-0.3524167538,-0.1409696192,0.5280256271,0.103697896,0.1615075171,-0.1449666619,0.3078524768,-0.0087297661,0.0976146981,-0.0128852651,-0.1341842711,0.3496532142,-0.2746606171,0.1606747657,-0.1773140728,0.0339708216,-0.1473898888,-0.0363861844,0.0670264214,0.0353453904,0.1572595835,0.0242215134,0.0062473495,-0.0083963685,0.1443327069,-0.1276887953,-0.0759726912,-0.2398124933,-0.2881832421,-0.309875071,0.2149091959,0.1042353511,-0.0598927923,0.1017578244,0.3165929914,-0.2579512298,-0.1195938438,0.0087799039,0.0931574926,0.3294774592,-0.0843710005,0.0652875751,0.2218031287,-0.0210187193,0.2137233913,-0.0134952012,-0.5275073051,0.069325693,-0.0279790387,0.0091241542,-0.1933581531,-0.626809001,-0.0088782348,0.1031659096,0.5481147766,-0.0142838014,0.1519878954,0.4092028141,0.2554835379,-0.1661252826,0.264388293,-0.0303027369,0.0417524204,0.2514635026,0.1111461669,-0.1434712708,-0.0249572955,-0.1038496122,-0.3425197601,-0.0681292191,0.1332846284,0.0181018002,-0.2350534201,0.0447584912,-0.4381090999,-0.103877835,0.2934699059,0.2456234843,0.2248572111,0.2331757993,0.1868967414,0.2919615805,0.0202706456,-0.2026331127,-0.1242306456,0.105250217,-0.1044747904,0.1108412072,0.2131985873,0.1113535017,0.2996945977,0.128821224,-0.3420161307,-0.3404206634,0.1220161244,-0.2721666992,0.0283037871,0.2411942035,0.0798239335,-0.1810231358,0.1847556829,-0.2773363292,-0.0575832799,0.0720744729,-0.3603481948,-0.354390502,-0.1715970337,0.1504694819,0.0036926903,0.1574962288,-0.4513111711,-0.0426098816,0.2559912503,-0.3391976357,0.1138054729,-0.1639285982,-0.1804626137,-0.1675998271,0.1240322888,0.0814560801,-0.1796029061,-0.0717220083,-0.2841574252,-0.0085382825,0.2089584321,0.1984399259,-0.1502616256,0.5191944242,-0.2032293975,0.5580466986,0.5535882711,-0.1875405759,-0.2054157257,0.5373868942,0.1265777946,-0.035294205,0.1371379495,-0.4054441452,0.2556640804,-0.1492992491,0.0239111912,0.4544432163,0.1228784025,-0.0366672389,-0.0533076935,0.0591707565,0.094570972,-0.0533153489,-0.4166417718,-0.2638634443,-0.1177741587,0.3790189922,0.2506234944,-0.0887018591,-0.0387289897,0.1767893434,0.0014025921,0.0067821331,-0.0009072539,-0.0842046812,-0.5144618154,0.025190603,-0.3117554784,-0.0911139995,-0.3747565448,0.1402051151,-0.1032433212,0.3250381052,-0.0609508604,-0.2853431404,0.2218149751,0.0576820746,-0.2218000144,-0.317121923,-0.0351766907,0.0971565247,-0.0542958602,0.3626121581,-0.099317044,0.4648391902,0.0869142637,-0.2966175377,0.0466390364,0.1641552597,0.0355114825,-0.2589477897,-0.180991292,0.4646167755,0.4503356814,-0.1478167474,-0.3648447692,-0.3587809801,0.0069056614,0.1058970094,0.0639766827,0.0289983246,0.1435213387,-0.3214043975,0.0080417814,0.3127244115,-0.0123363407,0.4519161284,0.0811970308,-0.0262396168,0.1977178007,-0.0217952374,0.1612604111,-0.3399371505,-0.1020602882,-0.0194319971,-0.0510117151,0.1063097194,0.0067979475,0.0667161867,0.3158483207,0.1552186608,0.1025185212,-0.0363587663,-0.0600491017,-0.1964571029,0.1785123348,0.08895538,0.4353640378,0.0364248864,-0.0735301599,0.042567309,0.1256317198,-0.091405414,0.354873836,0.0301692206,0.4449598491,0.3242442012,0.2033975869,-0.0742958412,-0.3939174116,-0.383282125,0.0276446324,0.118215248,-0.2580774128,0.2396309376,-0.2732917964,0.1311361641,-0.0723114461,-0.031521067,-0.1027374715,-0.3991473615,-0.2967226505,0.3699502349,-0.2913790047,0.0173462927,-0.1271987855,0.0827429891,0.1297951043,-0.1101027355,-0.3194954991,-0.1922740489,-0.1088828892,0.0139107835,0.3567505777,0.26393646,-0.1311806589,0.2827697992,-0.0451867245,-0.048971083,-0.3095544875,0.3275258541,-0.2431077212,-0.2856105268,0.2199111581,0.1815671623,0.1619457006,-0.6107788682,0.1853190362,0.0088885864,-0.2933406532,0.2972552478,0.223001197,-0.0029811368,-0.1525841057,-0.1341142058,-0.1398987919,-0.5762497187,0.4581654668,0.1357269585,0.2581652999,-0.0825648159,-0.0690140352,0.0106100226,-0.2777140737,-0.0837056562,0.0277054086,0.1249312013,0.100296624,-0.3619613945,0.0575923212,0.0255032443,0.0281640198,0.5084940195,-0.1329767108,-0.4221979976,0.4498475492,0.1196181849,-0.0777374431,-0.1762279719,-0.3853220046,0.0696588457,0.396043241,-0.0083335303,-0.1137979776,0.0838596746,0.2298501283,-0.0317384899,0.1729091257,-0.1163412631,0.5597530007,-0.0800676197,0.2650151253,0.0567664243,0.0075933975,0.2237437665,0.0908614248,-0.0370974168,-0.2848674953,-0.1019059196,-0.1746464074,-0.2492518425,0.0721041784,0.1653672457,-0.0519329235,-0.2720279992,-0.072112225,-0.1612115949,-0.1755963713,-0.3624178469,0.3011761606,-0.3059144318,0.0879911706,-0.3111079931,-0.0841155499,0.0438527651,0.0580729954,-0.0536612757,-0.2777520716,0.065991126,-0.1160641313,0.0500696078,0.1661209166,-0.2029540837,0.1418597847,-0.0379835851,0.2989365458,0.0233653262,-0.171015054,-0.0642995536,-0.1698566228,0.1479855776,0.0392503366,0.2059403062,0.2348193973,-0.1164165884,-0.2022816986,-0.0036498476,-0.2349404544,0.440018177,0.5935145617,0.3597708642,-0.3514997661,-0.1708199084,0.3545523286,-0.1483627856,-0.0955266133,-0.1216840819,0.0697270185,-0.2168030739,-0.2290590703,0.0502099358,0.2537956536,0.3231954575,0.1532005668,-0.1422909796,-0.2310124189,-0.447760731,-0.0806013793,0.1426682919,0.1965821832,-0.5051634312,0.1015682593,-0.1520391405,0.2174491286,0.4984903336,0.7618778944,0.2852302492,-0.2908283472,0.2480547726,-0.0173353758,0.4242439568,0.1194377765,-0.1531616449,0.1303167492,-0.2092983574,0.0522411428,0.1517799646,-0.2608374357,0.3953027427,-0.168643415,-0.2760580778,-0.2417637855,0.4991045296,-0.0435558371,-0.012494226,-0.0739121661,-0.2426343709,-0.3518019915,0.6364927888,0.0889950842,0.5242301822,0.4776637256,0.0604504272,0.5830122232,-0.1552409679,0.2822134793,-0.7763892412,-0.0235230736,-0.3103724122,0.1324641109,-0.021711072,-0.2295863181,0.2328011096,0.3454679549,-0.2449895442,0.0065117166,0.0266148224,0.0639140978,-0.0325961113,0.3628292084,-0.1441193074,-0.2786220908,-0.4063245654,0.1032003239,0.0568545312,-0.3126535714,0.0303328279,-0.3509737849,-0.0865488276,-0.2357477397,-0.2639683783,0.4338127375,-0.0716566816,0.3886187971,0.22395657,-0.0562004,0.3615510166,-0.099153474,-0.1680011004,-0.0753625408,0.0291987136,0.1506631374,-0.0272615943,-0.2743914723,-0.0218238253,0.1191586554,0.1346259564,-0.1488002837,0.0384193622,-0.0248725805,-0.1394096613,-0.3718870878,0.2763091624,0.2076297551,-0.4314110279,-0.3709971309,-0.1943300515,0.1156334281,-0.149670139,-0.0934759825,0.1180916354,-0.0015497981,-0.2765603065,-0.3599806428,-0.0708585009,-0.2900587916,0.0538640916,0.354416281,0.2216332406,0.0431616269,0.6239032149,0.2414469421,-0.0021351206,-0.3211764991,0.0817211047,0.3431159854,-0.4442771077,0.3989853263,-0.20757626,-0.1646204591,0.122054033,0.091997847,0.2259000689,0.1671583802,-0.2459678501,-0.5066788793,-0.0920283124,0.3302319348,-0.2621886134,-0.0244904067,-0.1378588229,0.2208992988,-0.3556537032,0.2872291505,-0.243415907,0.2291268557,-0.0172618553,-0.0485382937,-0.0581631362,-0.1316035539,0.1055705175,0.0473976433,0.1401485652,0.1752716601,0.0176036768,-0.1623847783,-0.1138048619,0.1333757192,0.1678049713,0.1552291363,-0.2986271381,0.1163604409,-0.1770166606,0.0377225056,0.1328800321,-0.3939383328,-0.0549995415,0.1374362558,0.4441269934,0.2775044739,-0.0967658013,0.1493074596,-0.1239722371,0.4012761116,-0.0929718241,0.2098116577,-0.1530317217,0.0011428164,0.0350741036,0.1986967027,-0.0442558415,0.0952140391,0.5684457421,-0.1944071203,-0.0607279614,-0.5084893703,0.2955584824,0.5219070911,-0.0011157548,-0.042270191,-0.2038498819,0.1671247333,-0.1614583582,-0.0393575616,0.3196537793,0.0164184812,-0.0983282998,0.2596956789,0.0950077251,-0.0378495455,-0.2987880409,0.1304668188,0.1833295673,0.072718136,0.2018181533,0.6401519775,-0.3257715106,0.0498724915,0.3796495199,0.1183043867,0.606377542,0.2897162437,0.1244285703,0.0713098943,0.420783788,0.0520862304,0.0774490163,-0.20151712,-0.4815178812,0.0902980268,0.2912359536,-0.1424287856,-0.111909695,0.3422737718,0.1441443115,-0.1292780638,-0.2035949975,0.3980275393,-0.1603071243,-0.2040880024,-0.2676356733,-0.1337142885,-0.6330999732,-0.1815863252,-0.0812522322,-0.0243735518,-0.0915029943,0.1115613207,0.0836272165,-0.2546156943,0.1198743135,0.0494740419,-0.0678948909,-0.2596521974,0.4679685235,0.0913939625,0.0372751132,0.0409069173,0.1487784088,0.4960689545,0.2502405345,-0.2635149658,-0.2182382792,-0.1987072527,-0.1470847279,-0.0179430582,0.2585283816,0.0494269542,0.1489408314,0.3585207164,0.114110738,-0.1599735469,0.0438396372,0.1822644323,0.0814697891,0.1857505143,0.6384834647,-0.0477938801,-0.1213485524,0.2243060172,-0.1319791675,-0.2510057688,-0.6355654001,0.2918816805,0.0026242989,0.1969861239,0.0760706961,0.0750238448,0.1019888222,0.638327539,0.0586278252,0.0899371728,-0.103629984,-0.1408049464,-0.0365859717,-0.0300198365,-0.3730284274,0.0104072904,0.1738446802,0.0966818631,0.167988807,0.2986499667,-0.0251649953,0.2913760543,-0.2197563946,-0.1188327298,-0.484356761,0.0687586665,0.2037036866,0.0928060189,-0.1202129647,-0.1476213634,0.0684486479,0.2533197403,0.0239625182,0.09037669,-0.2485013604,0.1168634072,-0.3223131895,0.1750455648,0.0504131168,0.1140335202,-0.0494049378,-0.1896672249,-0.1887531877,-0.5733943582,-0.1219350994,0.4686674476,0.1284101605,0.2680280805,-0.1597730964,0.1188096032,-0.294305563,0.1210640818,-0.4128634334,0.0388882309,-0.345837146,0.2615273893,-0.1557042599,0.2354152501,-0.1095620096,-0.0678942725,-0.2019300908,0.0950917751,-0.207477808,-0.3111889362,0.376058042,0.108802557,-0.232853353,0.464329958,0.1188012734,0.116584979,0.2316484749,-0.3983220756,-0.030120343,0.254344672,-0.1892004758,-0.3222723901,-0.0306952801,0.1550977826,-0.138454631,-0.0962281823,0.1503652036,0.0437868014,-0.1663366854,-0.175719738,0.0138161369]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/359","title":"ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures","comments":"The behavior I'm seeing is from the `json` script. \r\nI hacked this together to overcome the error with the `JSON` dataloader\r\n\r\n```\r\nclass DatasetBuilder(hf_nlp.ArrowBasedBuilder):\r\n BUILDER_CONFIG_CLASS = BuilderConfig\r\n\r\n def _info(self):\r\n return DatasetInfo()\r\n\r\n def _split_generators(self, dl_manager):\r\n \"\"\" We handle string, list and dicts in datafiles\r\n \"\"\"\r\n if isinstance(self.config.data_files, (str, list, tuple)):\r\n files = self.config.data_files\r\n if isinstance(files, str):\r\n files = [files]\r\n return [SplitGenerator(name=Split.TRAIN, gen_kwargs={\"files\": files})]\r\n splits = []\r\n for split_name in [Split.TRAIN, Split.VALIDATION, Split.TEST]:\r\n if split_name in self.config.data_files:\r\n files = self.config.data_files[split_name]\r\n if isinstance(files, str):\r\n files = [files]\r\n splits.append(SplitGenerator(name=split_name, gen_kwargs={\"files\": files}))\r\n return splits\r\n\r\n def _prepare_split(self, split_generator):\r\n fname = \"{}-{}.arrow\".format(self.name, split_generator.name)\r\n fpath = os.path.join(self._cache_dir, fname)\r\n\r\n writer = ArrowWriter(path=fpath)\r\n\r\n generator = self._generate_tables(**split_generator.gen_kwargs)\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n writer.write_table(table)\r\n num_examples, num_bytes = writer.finalize()\r\n\r\n split_generator.split_info.num_examples = num_examples\r\n split_generator.split_info.num_bytes = num_bytes\r\n # this is where the error is coming from\r\n # def parse_schema(schema, schema_dict):\r\n # for field in schema:\r\n # if pa.types.is_struct(field.type):\r\n # schema_dict[field.name] = {}\r\n # parse_schema(field.type, schema_dict[field.name])\r\n # elif pa.types.is_list(field.type) and pa.types.is_struct(field.type.value_type):\r\n # schema_dict[field.name] = {}\r\n # parse_schema(field.type.value_type, schema_dict[field.name])\r\n # else:\r\n # schema_dict[field.name] = Value(str(field.type))\r\n # \r\n # parse_schema(writer.schema, features)\r\n # self.info.features = Features(features)\r\n\r\n def _generate_tables(self, files):\r\n for i, file in enumerate(files):\r\n pa_table = paj.read_json(\r\n file\r\n )\r\n yield i, pa_table\r\n```\r\n\r\nSo I basically just don't populate the `self.info.features` though this doesn't seem to cause any problems in my downstream applications. \r\n\r\nThe other workaround I was doing was to just use pyarrow.json to build a table and then to create the Dataset with its constructor or from_table methods. `load_dataset` has nice split logic, so I'd prefer to use that.\r\n\r\n","body":"I tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-23-9aecfbee53bd> in <module>\r\n 55 from nlp import load_dataset\r\n 56 \r\n---> 57 ds = load_dataset(\"..\/text2struct\/model\/dataset_builder.py\", data_files=rel_datafiles)\r\n 58 \r\n 59 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: string> nor list<item: string>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\r\n```\r\n\r\nIf I create the dataset imperatively, using a pyarrow table, the dataset is created correctly. If I override the `_prepare_split` method to avoid calling the validate schema, the dataset can load as well. ","comment_length":254,"text":"ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures\nI tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-23-9aecfbee53bd> in <module>\r\n 55 from nlp import load_dataset\r\n 56 \r\n---> 57 ds = load_dataset(\"..\/text2struct\/model\/dataset_builder.py\", data_files=rel_datafiles)\r\n 58 \r\n 59 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: string> nor list<item: string>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\r\n```\r\n\r\nIf I create the dataset imperatively, using a pyarrow table, the dataset is created correctly. If I override the `_prepare_split` method to avoid calling the validate schema, the dataset can load as well. \nThe behavior I'm seeing is from the `json` script. \r\nI hacked this together to overcome the error with the `JSON` dataloader\r\n\r\n```\r\nclass DatasetBuilder(hf_nlp.ArrowBasedBuilder):\r\n BUILDER_CONFIG_CLASS = BuilderConfig\r\n\r\n def _info(self):\r\n return DatasetInfo()\r\n\r\n def _split_generators(self, dl_manager):\r\n \"\"\" We handle string, list and dicts in datafiles\r\n \"\"\"\r\n if isinstance(self.config.data_files, (str, list, tuple)):\r\n files = self.config.data_files\r\n if isinstance(files, str):\r\n files = [files]\r\n return [SplitGenerator(name=Split.TRAIN, gen_kwargs={\"files\": files})]\r\n splits = []\r\n for split_name in [Split.TRAIN, Split.VALIDATION, Split.TEST]:\r\n if split_name in self.config.data_files:\r\n files = self.config.data_files[split_name]\r\n if isinstance(files, str):\r\n files = [files]\r\n splits.append(SplitGenerator(name=split_name, gen_kwargs={\"files\": files}))\r\n return splits\r\n\r\n def _prepare_split(self, split_generator):\r\n fname = \"{}-{}.arrow\".format(self.name, split_generator.name)\r\n fpath = os.path.join(self._cache_dir, fname)\r\n\r\n writer = ArrowWriter(path=fpath)\r\n\r\n generator = self._generate_tables(**split_generator.gen_kwargs)\r\n for key, table in utils.tqdm(generator, unit=\" tables\", leave=False):\r\n writer.write_table(table)\r\n num_examples, num_bytes = writer.finalize()\r\n\r\n split_generator.split_info.num_examples = num_examples\r\n split_generator.split_info.num_bytes = num_bytes\r\n # this is where the error is coming from\r\n # def parse_schema(schema, schema_dict):\r\n # for field in schema:\r\n # if pa.types.is_struct(field.type):\r\n # schema_dict[field.name] = {}\r\n # parse_schema(field.type, schema_dict[field.name])\r\n # elif pa.types.is_list(field.type) and pa.types.is_struct(field.type.value_type):\r\n # schema_dict[field.name] = {}\r\n # parse_schema(field.type.value_type, schema_dict[field.name])\r\n # else:\r\n # schema_dict[field.name] = Value(str(field.type))\r\n # \r\n # parse_schema(writer.schema, features)\r\n # self.info.features = Features(features)\r\n\r\n def _generate_tables(self, files):\r\n for i, file in enumerate(files):\r\n pa_table = paj.read_json(\r\n file\r\n )\r\n yield i, pa_table\r\n```\r\n\r\nSo I basically just don't populate the `self.info.features` though this doesn't seem to cause any problems in my downstream applications. \r\n\r\nThe other workaround I was doing was to just use pyarrow.json to build a table and then to create the Dataset with its constructor or from_table methods. `load_dataset` has nice split logic, so I'd prefer to use that.\r\n\r\n","embeddings":[-0.3260453045,0.3610112369,-0.0600554347,0.5486999154,0.0407031141,-0.1400738508,0.3431124687,0.4284070134,0.2143322676,-0.1139701083,-0.0436313152,0.3332704902,0.2364399433,0.0587036684,-0.2658620179,-0.1656162292,-0.0196319576,0.2937225401,-0.124623999,0.3482965231,-0.222588554,0.4523977935,-0.1000257656,0.1312664449,0.0139916008,-0.0840229094,0.0551255234,0.3630198836,-0.1661049724,-0.5221460462,0.261677146,-0.297565192,-0.1071128175,0.4265525639,-0.0001121765,0.283685714,0.2796024978,-0.1050383896,-0.367891252,-0.0563196056,0.0219089277,-0.1897962093,0.080716528,-0.1802360266,0.2029777616,-0.2733659148,-0.3524167538,-0.1409696192,0.5280256271,0.103697896,0.1615075171,-0.1449666619,0.3078524768,-0.0087297661,0.0976146981,-0.0128852651,-0.1341842711,0.3496532142,-0.2746606171,0.1606747657,-0.1773140728,0.0339708216,-0.1473898888,-0.0363861844,0.0670264214,0.0353453904,0.1572595835,0.0242215134,0.0062473495,-0.0083963685,0.1443327069,-0.1276887953,-0.0759726912,-0.2398124933,-0.2881832421,-0.309875071,0.2149091959,0.1042353511,-0.0598927923,0.1017578244,0.3165929914,-0.2579512298,-0.1195938438,0.0087799039,0.0931574926,0.3294774592,-0.0843710005,0.0652875751,0.2218031287,-0.0210187193,0.2137233913,-0.0134952012,-0.5275073051,0.069325693,-0.0279790387,0.0091241542,-0.1933581531,-0.626809001,-0.0088782348,0.1031659096,0.5481147766,-0.0142838014,0.1519878954,0.4092028141,0.2554835379,-0.1661252826,0.264388293,-0.0303027369,0.0417524204,0.2514635026,0.1111461669,-0.1434712708,-0.0249572955,-0.1038496122,-0.3425197601,-0.0681292191,0.1332846284,0.0181018002,-0.2350534201,0.0447584912,-0.4381090999,-0.103877835,0.2934699059,0.2456234843,0.2248572111,0.2331757993,0.1868967414,0.2919615805,0.0202706456,-0.2026331127,-0.1242306456,0.105250217,-0.1044747904,0.1108412072,0.2131985873,0.1113535017,0.2996945977,0.128821224,-0.3420161307,-0.3404206634,0.1220161244,-0.2721666992,0.0283037871,0.2411942035,0.0798239335,-0.1810231358,0.1847556829,-0.2773363292,-0.0575832799,0.0720744729,-0.3603481948,-0.354390502,-0.1715970337,0.1504694819,0.0036926903,0.1574962288,-0.4513111711,-0.0426098816,0.2559912503,-0.3391976357,0.1138054729,-0.1639285982,-0.1804626137,-0.1675998271,0.1240322888,0.0814560801,-0.1796029061,-0.0717220083,-0.2841574252,-0.0085382825,0.2089584321,0.1984399259,-0.1502616256,0.5191944242,-0.2032293975,0.5580466986,0.5535882711,-0.1875405759,-0.2054157257,0.5373868942,0.1265777946,-0.035294205,0.1371379495,-0.4054441452,0.2556640804,-0.1492992491,0.0239111912,0.4544432163,0.1228784025,-0.0366672389,-0.0533076935,0.0591707565,0.094570972,-0.0533153489,-0.4166417718,-0.2638634443,-0.1177741587,0.3790189922,0.2506234944,-0.0887018591,-0.0387289897,0.1767893434,0.0014025921,0.0067821331,-0.0009072539,-0.0842046812,-0.5144618154,0.025190603,-0.3117554784,-0.0911139995,-0.3747565448,0.1402051151,-0.1032433212,0.3250381052,-0.0609508604,-0.2853431404,0.2218149751,0.0576820746,-0.2218000144,-0.317121923,-0.0351766907,0.0971565247,-0.0542958602,0.3626121581,-0.099317044,0.4648391902,0.0869142637,-0.2966175377,0.0466390364,0.1641552597,0.0355114825,-0.2589477897,-0.180991292,0.4646167755,0.4503356814,-0.1478167474,-0.3648447692,-0.3587809801,0.0069056614,0.1058970094,0.0639766827,0.0289983246,0.1435213387,-0.3214043975,0.0080417814,0.3127244115,-0.0123363407,0.4519161284,0.0811970308,-0.0262396168,0.1977178007,-0.0217952374,0.1612604111,-0.3399371505,-0.1020602882,-0.0194319971,-0.0510117151,0.1063097194,0.0067979475,0.0667161867,0.3158483207,0.1552186608,0.1025185212,-0.0363587663,-0.0600491017,-0.1964571029,0.1785123348,0.08895538,0.4353640378,0.0364248864,-0.0735301599,0.042567309,0.1256317198,-0.091405414,0.354873836,0.0301692206,0.4449598491,0.3242442012,0.2033975869,-0.0742958412,-0.3939174116,-0.383282125,0.0276446324,0.118215248,-0.2580774128,0.2396309376,-0.2732917964,0.1311361641,-0.0723114461,-0.031521067,-0.1027374715,-0.3991473615,-0.2967226505,0.3699502349,-0.2913790047,0.0173462927,-0.1271987855,0.0827429891,0.1297951043,-0.1101027355,-0.3194954991,-0.1922740489,-0.1088828892,0.0139107835,0.3567505777,0.26393646,-0.1311806589,0.2827697992,-0.0451867245,-0.048971083,-0.3095544875,0.3275258541,-0.2431077212,-0.2856105268,0.2199111581,0.1815671623,0.1619457006,-0.6107788682,0.1853190362,0.0088885864,-0.2933406532,0.2972552478,0.223001197,-0.0029811368,-0.1525841057,-0.1341142058,-0.1398987919,-0.5762497187,0.4581654668,0.1357269585,0.2581652999,-0.0825648159,-0.0690140352,0.0106100226,-0.2777140737,-0.0837056562,0.0277054086,0.1249312013,0.100296624,-0.3619613945,0.0575923212,0.0255032443,0.0281640198,0.5084940195,-0.1329767108,-0.4221979976,0.4498475492,0.1196181849,-0.0777374431,-0.1762279719,-0.3853220046,0.0696588457,0.396043241,-0.0083335303,-0.1137979776,0.0838596746,0.2298501283,-0.0317384899,0.1729091257,-0.1163412631,0.5597530007,-0.0800676197,0.2650151253,0.0567664243,0.0075933975,0.2237437665,0.0908614248,-0.0370974168,-0.2848674953,-0.1019059196,-0.1746464074,-0.2492518425,0.0721041784,0.1653672457,-0.0519329235,-0.2720279992,-0.072112225,-0.1612115949,-0.1755963713,-0.3624178469,0.3011761606,-0.3059144318,0.0879911706,-0.3111079931,-0.0841155499,0.0438527651,0.0580729954,-0.0536612757,-0.2777520716,0.065991126,-0.1160641313,0.0500696078,0.1661209166,-0.2029540837,0.1418597847,-0.0379835851,0.2989365458,0.0233653262,-0.171015054,-0.0642995536,-0.1698566228,0.1479855776,0.0392503366,0.2059403062,0.2348193973,-0.1164165884,-0.2022816986,-0.0036498476,-0.2349404544,0.440018177,0.5935145617,0.3597708642,-0.3514997661,-0.1708199084,0.3545523286,-0.1483627856,-0.0955266133,-0.1216840819,0.0697270185,-0.2168030739,-0.2290590703,0.0502099358,0.2537956536,0.3231954575,0.1532005668,-0.1422909796,-0.2310124189,-0.447760731,-0.0806013793,0.1426682919,0.1965821832,-0.5051634312,0.1015682593,-0.1520391405,0.2174491286,0.4984903336,0.7618778944,0.2852302492,-0.2908283472,0.2480547726,-0.0173353758,0.4242439568,0.1194377765,-0.1531616449,0.1303167492,-0.2092983574,0.0522411428,0.1517799646,-0.2608374357,0.3953027427,-0.168643415,-0.2760580778,-0.2417637855,0.4991045296,-0.0435558371,-0.012494226,-0.0739121661,-0.2426343709,-0.3518019915,0.6364927888,0.0889950842,0.5242301822,0.4776637256,0.0604504272,0.5830122232,-0.1552409679,0.2822134793,-0.7763892412,-0.0235230736,-0.3103724122,0.1324641109,-0.021711072,-0.2295863181,0.2328011096,0.3454679549,-0.2449895442,0.0065117166,0.0266148224,0.0639140978,-0.0325961113,0.3628292084,-0.1441193074,-0.2786220908,-0.4063245654,0.1032003239,0.0568545312,-0.3126535714,0.0303328279,-0.3509737849,-0.0865488276,-0.2357477397,-0.2639683783,0.4338127375,-0.0716566816,0.3886187971,0.22395657,-0.0562004,0.3615510166,-0.099153474,-0.1680011004,-0.0753625408,0.0291987136,0.1506631374,-0.0272615943,-0.2743914723,-0.0218238253,0.1191586554,0.1346259564,-0.1488002837,0.0384193622,-0.0248725805,-0.1394096613,-0.3718870878,0.2763091624,0.2076297551,-0.4314110279,-0.3709971309,-0.1943300515,0.1156334281,-0.149670139,-0.0934759825,0.1180916354,-0.0015497981,-0.2765603065,-0.3599806428,-0.0708585009,-0.2900587916,0.0538640916,0.354416281,0.2216332406,0.0431616269,0.6239032149,0.2414469421,-0.0021351206,-0.3211764991,0.0817211047,0.3431159854,-0.4442771077,0.3989853263,-0.20757626,-0.1646204591,0.122054033,0.091997847,0.2259000689,0.1671583802,-0.2459678501,-0.5066788793,-0.0920283124,0.3302319348,-0.2621886134,-0.0244904067,-0.1378588229,0.2208992988,-0.3556537032,0.2872291505,-0.243415907,0.2291268557,-0.0172618553,-0.0485382937,-0.0581631362,-0.1316035539,0.1055705175,0.0473976433,0.1401485652,0.1752716601,0.0176036768,-0.1623847783,-0.1138048619,0.1333757192,0.1678049713,0.1552291363,-0.2986271381,0.1163604409,-0.1770166606,0.0377225056,0.1328800321,-0.3939383328,-0.0549995415,0.1374362558,0.4441269934,0.2775044739,-0.0967658013,0.1493074596,-0.1239722371,0.4012761116,-0.0929718241,0.2098116577,-0.1530317217,0.0011428164,0.0350741036,0.1986967027,-0.0442558415,0.0952140391,0.5684457421,-0.1944071203,-0.0607279614,-0.5084893703,0.2955584824,0.5219070911,-0.0011157548,-0.042270191,-0.2038498819,0.1671247333,-0.1614583582,-0.0393575616,0.3196537793,0.0164184812,-0.0983282998,0.2596956789,0.0950077251,-0.0378495455,-0.2987880409,0.1304668188,0.1833295673,0.072718136,0.2018181533,0.6401519775,-0.3257715106,0.0498724915,0.3796495199,0.1183043867,0.606377542,0.2897162437,0.1244285703,0.0713098943,0.420783788,0.0520862304,0.0774490163,-0.20151712,-0.4815178812,0.0902980268,0.2912359536,-0.1424287856,-0.111909695,0.3422737718,0.1441443115,-0.1292780638,-0.2035949975,0.3980275393,-0.1603071243,-0.2040880024,-0.2676356733,-0.1337142885,-0.6330999732,-0.1815863252,-0.0812522322,-0.0243735518,-0.0915029943,0.1115613207,0.0836272165,-0.2546156943,0.1198743135,0.0494740419,-0.0678948909,-0.2596521974,0.4679685235,0.0913939625,0.0372751132,0.0409069173,0.1487784088,0.4960689545,0.2502405345,-0.2635149658,-0.2182382792,-0.1987072527,-0.1470847279,-0.0179430582,0.2585283816,0.0494269542,0.1489408314,0.3585207164,0.114110738,-0.1599735469,0.0438396372,0.1822644323,0.0814697891,0.1857505143,0.6384834647,-0.0477938801,-0.1213485524,0.2243060172,-0.1319791675,-0.2510057688,-0.6355654001,0.2918816805,0.0026242989,0.1969861239,0.0760706961,0.0750238448,0.1019888222,0.638327539,0.0586278252,0.0899371728,-0.103629984,-0.1408049464,-0.0365859717,-0.0300198365,-0.3730284274,0.0104072904,0.1738446802,0.0966818631,0.167988807,0.2986499667,-0.0251649953,0.2913760543,-0.2197563946,-0.1188327298,-0.484356761,0.0687586665,0.2037036866,0.0928060189,-0.1202129647,-0.1476213634,0.0684486479,0.2533197403,0.0239625182,0.09037669,-0.2485013604,0.1168634072,-0.3223131895,0.1750455648,0.0504131168,0.1140335202,-0.0494049378,-0.1896672249,-0.1887531877,-0.5733943582,-0.1219350994,0.4686674476,0.1284101605,0.2680280805,-0.1597730964,0.1188096032,-0.294305563,0.1210640818,-0.4128634334,0.0388882309,-0.345837146,0.2615273893,-0.1557042599,0.2354152501,-0.1095620096,-0.0678942725,-0.2019300908,0.0950917751,-0.207477808,-0.3111889362,0.376058042,0.108802557,-0.232853353,0.464329958,0.1188012734,0.116584979,0.2316484749,-0.3983220756,-0.030120343,0.254344672,-0.1892004758,-0.3222723901,-0.0306952801,0.1550977826,-0.138454631,-0.0962281823,0.1503652036,0.0437868014,-0.1663366854,-0.175719738,0.0138161369]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/359","title":"ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures","comments":"Also noticed that if you for example in a loader script\r\n\r\n```\r\nfrom nlp import ArrowBasedBuilder\r\n\r\nclass MyBuilder(ArrowBasedBuilder):\r\n...\r\n\r\n```\r\nand use that in the subclass, it will be on the module's __dict__ and will be selected before the `MyBuilder` subclass, and it will raise `NotImplementedError` on its `_generate_examples` method... In the code it check for abstract classes but Builder and ArrowBasedBuilder aren't abstract classes, they're regular classes with `@abstract_methods`.","body":"I tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-23-9aecfbee53bd> in <module>\r\n 55 from nlp import load_dataset\r\n 56 \r\n---> 57 ds = load_dataset(\"..\/text2struct\/model\/dataset_builder.py\", data_files=rel_datafiles)\r\n 58 \r\n 59 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: string> nor list<item: string>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\r\n```\r\n\r\nIf I create the dataset imperatively, using a pyarrow table, the dataset is created correctly. If I override the `_prepare_split` method to avoid calling the validate schema, the dataset can load as well. ","comment_length":70,"text":"ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures\nI tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-23-9aecfbee53bd> in <module>\r\n 55 from nlp import load_dataset\r\n 56 \r\n---> 57 ds = load_dataset(\"..\/text2struct\/model\/dataset_builder.py\", data_files=rel_datafiles)\r\n 58 \r\n 59 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: string> nor list<item: string>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\r\n```\r\n\r\nIf I create the dataset imperatively, using a pyarrow table, the dataset is created correctly. If I override the `_prepare_split` method to avoid calling the validate schema, the dataset can load as well. \nAlso noticed that if you for example in a loader script\r\n\r\n```\r\nfrom nlp import ArrowBasedBuilder\r\n\r\nclass MyBuilder(ArrowBasedBuilder):\r\n...\r\n\r\n```\r\nand use that in the subclass, it will be on the module's __dict__ and will be selected before the `MyBuilder` subclass, and it will raise `NotImplementedError` on its `_generate_examples` method... In the code it check for abstract classes but Builder and ArrowBasedBuilder aren't abstract classes, they're regular classes with `@abstract_methods`.","embeddings":[-0.3260453045,0.3610112369,-0.0600554347,0.5486999154,0.0407031141,-0.1400738508,0.3431124687,0.4284070134,0.2143322676,-0.1139701083,-0.0436313152,0.3332704902,0.2364399433,0.0587036684,-0.2658620179,-0.1656162292,-0.0196319576,0.2937225401,-0.124623999,0.3482965231,-0.222588554,0.4523977935,-0.1000257656,0.1312664449,0.0139916008,-0.0840229094,0.0551255234,0.3630198836,-0.1661049724,-0.5221460462,0.261677146,-0.297565192,-0.1071128175,0.4265525639,-0.0001121765,0.283685714,0.2796024978,-0.1050383896,-0.367891252,-0.0563196056,0.0219089277,-0.1897962093,0.080716528,-0.1802360266,0.2029777616,-0.2733659148,-0.3524167538,-0.1409696192,0.5280256271,0.103697896,0.1615075171,-0.1449666619,0.3078524768,-0.0087297661,0.0976146981,-0.0128852651,-0.1341842711,0.3496532142,-0.2746606171,0.1606747657,-0.1773140728,0.0339708216,-0.1473898888,-0.0363861844,0.0670264214,0.0353453904,0.1572595835,0.0242215134,0.0062473495,-0.0083963685,0.1443327069,-0.1276887953,-0.0759726912,-0.2398124933,-0.2881832421,-0.309875071,0.2149091959,0.1042353511,-0.0598927923,0.1017578244,0.3165929914,-0.2579512298,-0.1195938438,0.0087799039,0.0931574926,0.3294774592,-0.0843710005,0.0652875751,0.2218031287,-0.0210187193,0.2137233913,-0.0134952012,-0.5275073051,0.069325693,-0.0279790387,0.0091241542,-0.1933581531,-0.626809001,-0.0088782348,0.1031659096,0.5481147766,-0.0142838014,0.1519878954,0.4092028141,0.2554835379,-0.1661252826,0.264388293,-0.0303027369,0.0417524204,0.2514635026,0.1111461669,-0.1434712708,-0.0249572955,-0.1038496122,-0.3425197601,-0.0681292191,0.1332846284,0.0181018002,-0.2350534201,0.0447584912,-0.4381090999,-0.103877835,0.2934699059,0.2456234843,0.2248572111,0.2331757993,0.1868967414,0.2919615805,0.0202706456,-0.2026331127,-0.1242306456,0.105250217,-0.1044747904,0.1108412072,0.2131985873,0.1113535017,0.2996945977,0.128821224,-0.3420161307,-0.3404206634,0.1220161244,-0.2721666992,0.0283037871,0.2411942035,0.0798239335,-0.1810231358,0.1847556829,-0.2773363292,-0.0575832799,0.0720744729,-0.3603481948,-0.354390502,-0.1715970337,0.1504694819,0.0036926903,0.1574962288,-0.4513111711,-0.0426098816,0.2559912503,-0.3391976357,0.1138054729,-0.1639285982,-0.1804626137,-0.1675998271,0.1240322888,0.0814560801,-0.1796029061,-0.0717220083,-0.2841574252,-0.0085382825,0.2089584321,0.1984399259,-0.1502616256,0.5191944242,-0.2032293975,0.5580466986,0.5535882711,-0.1875405759,-0.2054157257,0.5373868942,0.1265777946,-0.035294205,0.1371379495,-0.4054441452,0.2556640804,-0.1492992491,0.0239111912,0.4544432163,0.1228784025,-0.0366672389,-0.0533076935,0.0591707565,0.094570972,-0.0533153489,-0.4166417718,-0.2638634443,-0.1177741587,0.3790189922,0.2506234944,-0.0887018591,-0.0387289897,0.1767893434,0.0014025921,0.0067821331,-0.0009072539,-0.0842046812,-0.5144618154,0.025190603,-0.3117554784,-0.0911139995,-0.3747565448,0.1402051151,-0.1032433212,0.3250381052,-0.0609508604,-0.2853431404,0.2218149751,0.0576820746,-0.2218000144,-0.317121923,-0.0351766907,0.0971565247,-0.0542958602,0.3626121581,-0.099317044,0.4648391902,0.0869142637,-0.2966175377,0.0466390364,0.1641552597,0.0355114825,-0.2589477897,-0.180991292,0.4646167755,0.4503356814,-0.1478167474,-0.3648447692,-0.3587809801,0.0069056614,0.1058970094,0.0639766827,0.0289983246,0.1435213387,-0.3214043975,0.0080417814,0.3127244115,-0.0123363407,0.4519161284,0.0811970308,-0.0262396168,0.1977178007,-0.0217952374,0.1612604111,-0.3399371505,-0.1020602882,-0.0194319971,-0.0510117151,0.1063097194,0.0067979475,0.0667161867,0.3158483207,0.1552186608,0.1025185212,-0.0363587663,-0.0600491017,-0.1964571029,0.1785123348,0.08895538,0.4353640378,0.0364248864,-0.0735301599,0.042567309,0.1256317198,-0.091405414,0.354873836,0.0301692206,0.4449598491,0.3242442012,0.2033975869,-0.0742958412,-0.3939174116,-0.383282125,0.0276446324,0.118215248,-0.2580774128,0.2396309376,-0.2732917964,0.1311361641,-0.0723114461,-0.031521067,-0.1027374715,-0.3991473615,-0.2967226505,0.3699502349,-0.2913790047,0.0173462927,-0.1271987855,0.0827429891,0.1297951043,-0.1101027355,-0.3194954991,-0.1922740489,-0.1088828892,0.0139107835,0.3567505777,0.26393646,-0.1311806589,0.2827697992,-0.0451867245,-0.048971083,-0.3095544875,0.3275258541,-0.2431077212,-0.2856105268,0.2199111581,0.1815671623,0.1619457006,-0.6107788682,0.1853190362,0.0088885864,-0.2933406532,0.2972552478,0.223001197,-0.0029811368,-0.1525841057,-0.1341142058,-0.1398987919,-0.5762497187,0.4581654668,0.1357269585,0.2581652999,-0.0825648159,-0.0690140352,0.0106100226,-0.2777140737,-0.0837056562,0.0277054086,0.1249312013,0.100296624,-0.3619613945,0.0575923212,0.0255032443,0.0281640198,0.5084940195,-0.1329767108,-0.4221979976,0.4498475492,0.1196181849,-0.0777374431,-0.1762279719,-0.3853220046,0.0696588457,0.396043241,-0.0083335303,-0.1137979776,0.0838596746,0.2298501283,-0.0317384899,0.1729091257,-0.1163412631,0.5597530007,-0.0800676197,0.2650151253,0.0567664243,0.0075933975,0.2237437665,0.0908614248,-0.0370974168,-0.2848674953,-0.1019059196,-0.1746464074,-0.2492518425,0.0721041784,0.1653672457,-0.0519329235,-0.2720279992,-0.072112225,-0.1612115949,-0.1755963713,-0.3624178469,0.3011761606,-0.3059144318,0.0879911706,-0.3111079931,-0.0841155499,0.0438527651,0.0580729954,-0.0536612757,-0.2777520716,0.065991126,-0.1160641313,0.0500696078,0.1661209166,-0.2029540837,0.1418597847,-0.0379835851,0.2989365458,0.0233653262,-0.171015054,-0.0642995536,-0.1698566228,0.1479855776,0.0392503366,0.2059403062,0.2348193973,-0.1164165884,-0.2022816986,-0.0036498476,-0.2349404544,0.440018177,0.5935145617,0.3597708642,-0.3514997661,-0.1708199084,0.3545523286,-0.1483627856,-0.0955266133,-0.1216840819,0.0697270185,-0.2168030739,-0.2290590703,0.0502099358,0.2537956536,0.3231954575,0.1532005668,-0.1422909796,-0.2310124189,-0.447760731,-0.0806013793,0.1426682919,0.1965821832,-0.5051634312,0.1015682593,-0.1520391405,0.2174491286,0.4984903336,0.7618778944,0.2852302492,-0.2908283472,0.2480547726,-0.0173353758,0.4242439568,0.1194377765,-0.1531616449,0.1303167492,-0.2092983574,0.0522411428,0.1517799646,-0.2608374357,0.3953027427,-0.168643415,-0.2760580778,-0.2417637855,0.4991045296,-0.0435558371,-0.012494226,-0.0739121661,-0.2426343709,-0.3518019915,0.6364927888,0.0889950842,0.5242301822,0.4776637256,0.0604504272,0.5830122232,-0.1552409679,0.2822134793,-0.7763892412,-0.0235230736,-0.3103724122,0.1324641109,-0.021711072,-0.2295863181,0.2328011096,0.3454679549,-0.2449895442,0.0065117166,0.0266148224,0.0639140978,-0.0325961113,0.3628292084,-0.1441193074,-0.2786220908,-0.4063245654,0.1032003239,0.0568545312,-0.3126535714,0.0303328279,-0.3509737849,-0.0865488276,-0.2357477397,-0.2639683783,0.4338127375,-0.0716566816,0.3886187971,0.22395657,-0.0562004,0.3615510166,-0.099153474,-0.1680011004,-0.0753625408,0.0291987136,0.1506631374,-0.0272615943,-0.2743914723,-0.0218238253,0.1191586554,0.1346259564,-0.1488002837,0.0384193622,-0.0248725805,-0.1394096613,-0.3718870878,0.2763091624,0.2076297551,-0.4314110279,-0.3709971309,-0.1943300515,0.1156334281,-0.149670139,-0.0934759825,0.1180916354,-0.0015497981,-0.2765603065,-0.3599806428,-0.0708585009,-0.2900587916,0.0538640916,0.354416281,0.2216332406,0.0431616269,0.6239032149,0.2414469421,-0.0021351206,-0.3211764991,0.0817211047,0.3431159854,-0.4442771077,0.3989853263,-0.20757626,-0.1646204591,0.122054033,0.091997847,0.2259000689,0.1671583802,-0.2459678501,-0.5066788793,-0.0920283124,0.3302319348,-0.2621886134,-0.0244904067,-0.1378588229,0.2208992988,-0.3556537032,0.2872291505,-0.243415907,0.2291268557,-0.0172618553,-0.0485382937,-0.0581631362,-0.1316035539,0.1055705175,0.0473976433,0.1401485652,0.1752716601,0.0176036768,-0.1623847783,-0.1138048619,0.1333757192,0.1678049713,0.1552291363,-0.2986271381,0.1163604409,-0.1770166606,0.0377225056,0.1328800321,-0.3939383328,-0.0549995415,0.1374362558,0.4441269934,0.2775044739,-0.0967658013,0.1493074596,-0.1239722371,0.4012761116,-0.0929718241,0.2098116577,-0.1530317217,0.0011428164,0.0350741036,0.1986967027,-0.0442558415,0.0952140391,0.5684457421,-0.1944071203,-0.0607279614,-0.5084893703,0.2955584824,0.5219070911,-0.0011157548,-0.042270191,-0.2038498819,0.1671247333,-0.1614583582,-0.0393575616,0.3196537793,0.0164184812,-0.0983282998,0.2596956789,0.0950077251,-0.0378495455,-0.2987880409,0.1304668188,0.1833295673,0.072718136,0.2018181533,0.6401519775,-0.3257715106,0.0498724915,0.3796495199,0.1183043867,0.606377542,0.2897162437,0.1244285703,0.0713098943,0.420783788,0.0520862304,0.0774490163,-0.20151712,-0.4815178812,0.0902980268,0.2912359536,-0.1424287856,-0.111909695,0.3422737718,0.1441443115,-0.1292780638,-0.2035949975,0.3980275393,-0.1603071243,-0.2040880024,-0.2676356733,-0.1337142885,-0.6330999732,-0.1815863252,-0.0812522322,-0.0243735518,-0.0915029943,0.1115613207,0.0836272165,-0.2546156943,0.1198743135,0.0494740419,-0.0678948909,-0.2596521974,0.4679685235,0.0913939625,0.0372751132,0.0409069173,0.1487784088,0.4960689545,0.2502405345,-0.2635149658,-0.2182382792,-0.1987072527,-0.1470847279,-0.0179430582,0.2585283816,0.0494269542,0.1489408314,0.3585207164,0.114110738,-0.1599735469,0.0438396372,0.1822644323,0.0814697891,0.1857505143,0.6384834647,-0.0477938801,-0.1213485524,0.2243060172,-0.1319791675,-0.2510057688,-0.6355654001,0.2918816805,0.0026242989,0.1969861239,0.0760706961,0.0750238448,0.1019888222,0.638327539,0.0586278252,0.0899371728,-0.103629984,-0.1408049464,-0.0365859717,-0.0300198365,-0.3730284274,0.0104072904,0.1738446802,0.0966818631,0.167988807,0.2986499667,-0.0251649953,0.2913760543,-0.2197563946,-0.1188327298,-0.484356761,0.0687586665,0.2037036866,0.0928060189,-0.1202129647,-0.1476213634,0.0684486479,0.2533197403,0.0239625182,0.09037669,-0.2485013604,0.1168634072,-0.3223131895,0.1750455648,0.0504131168,0.1140335202,-0.0494049378,-0.1896672249,-0.1887531877,-0.5733943582,-0.1219350994,0.4686674476,0.1284101605,0.2680280805,-0.1597730964,0.1188096032,-0.294305563,0.1210640818,-0.4128634334,0.0388882309,-0.345837146,0.2615273893,-0.1557042599,0.2354152501,-0.1095620096,-0.0678942725,-0.2019300908,0.0950917751,-0.207477808,-0.3111889362,0.376058042,0.108802557,-0.232853353,0.464329958,0.1188012734,0.116584979,0.2316484749,-0.3983220756,-0.030120343,0.254344672,-0.1892004758,-0.3222723901,-0.0306952801,0.1550977826,-0.138454631,-0.0962281823,0.1503652036,0.0437868014,-0.1663366854,-0.175719738,0.0138161369]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/359","title":"ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures","comments":"Indeed this is part of a more general limitation which is the fact that we should generate and update the `features` from the auto-inferred Arrow schema when they are not provided (also happen when a user change the schema using `map()`, the features should be auto-generated and guessed as much as possible to keep the `features` synced with the underlying Arrow table schema).\r\n\r\nWe will try to solve this soon.","body":"I tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-23-9aecfbee53bd> in <module>\r\n 55 from nlp import load_dataset\r\n 56 \r\n---> 57 ds = load_dataset(\"..\/text2struct\/model\/dataset_builder.py\", data_files=rel_datafiles)\r\n 58 \r\n 59 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: string> nor list<item: string>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\r\n```\r\n\r\nIf I create the dataset imperatively, using a pyarrow table, the dataset is created correctly. If I override the `_prepare_split` method to avoid calling the validate schema, the dataset can load as well. ","comment_length":70,"text":"ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures\nI tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-23-9aecfbee53bd> in <module>\r\n 55 from nlp import load_dataset\r\n 56 \r\n---> 57 ds = load_dataset(\"..\/text2struct\/model\/dataset_builder.py\", data_files=rel_datafiles)\r\n 58 \r\n 59 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: string> nor list<item: string>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\r\n```\r\n\r\nIf I create the dataset imperatively, using a pyarrow table, the dataset is created correctly. If I override the `_prepare_split` method to avoid calling the validate schema, the dataset can load as well. \nIndeed this is part of a more general limitation which is the fact that we should generate and update the `features` from the auto-inferred Arrow schema when they are not provided (also happen when a user change the schema using `map()`, the features should be auto-generated and guessed as much as possible to keep the `features` synced with the underlying Arrow table schema).\r\n\r\nWe will try to solve this soon.","embeddings":[-0.3260453045,0.3610112369,-0.0600554347,0.5486999154,0.0407031141,-0.1400738508,0.3431124687,0.4284070134,0.2143322676,-0.1139701083,-0.0436313152,0.3332704902,0.2364399433,0.0587036684,-0.2658620179,-0.1656162292,-0.0196319576,0.2937225401,-0.124623999,0.3482965231,-0.222588554,0.4523977935,-0.1000257656,0.1312664449,0.0139916008,-0.0840229094,0.0551255234,0.3630198836,-0.1661049724,-0.5221460462,0.261677146,-0.297565192,-0.1071128175,0.4265525639,-0.0001121765,0.283685714,0.2796024978,-0.1050383896,-0.367891252,-0.0563196056,0.0219089277,-0.1897962093,0.080716528,-0.1802360266,0.2029777616,-0.2733659148,-0.3524167538,-0.1409696192,0.5280256271,0.103697896,0.1615075171,-0.1449666619,0.3078524768,-0.0087297661,0.0976146981,-0.0128852651,-0.1341842711,0.3496532142,-0.2746606171,0.1606747657,-0.1773140728,0.0339708216,-0.1473898888,-0.0363861844,0.0670264214,0.0353453904,0.1572595835,0.0242215134,0.0062473495,-0.0083963685,0.1443327069,-0.1276887953,-0.0759726912,-0.2398124933,-0.2881832421,-0.309875071,0.2149091959,0.1042353511,-0.0598927923,0.1017578244,0.3165929914,-0.2579512298,-0.1195938438,0.0087799039,0.0931574926,0.3294774592,-0.0843710005,0.0652875751,0.2218031287,-0.0210187193,0.2137233913,-0.0134952012,-0.5275073051,0.069325693,-0.0279790387,0.0091241542,-0.1933581531,-0.626809001,-0.0088782348,0.1031659096,0.5481147766,-0.0142838014,0.1519878954,0.4092028141,0.2554835379,-0.1661252826,0.264388293,-0.0303027369,0.0417524204,0.2514635026,0.1111461669,-0.1434712708,-0.0249572955,-0.1038496122,-0.3425197601,-0.0681292191,0.1332846284,0.0181018002,-0.2350534201,0.0447584912,-0.4381090999,-0.103877835,0.2934699059,0.2456234843,0.2248572111,0.2331757993,0.1868967414,0.2919615805,0.0202706456,-0.2026331127,-0.1242306456,0.105250217,-0.1044747904,0.1108412072,0.2131985873,0.1113535017,0.2996945977,0.128821224,-0.3420161307,-0.3404206634,0.1220161244,-0.2721666992,0.0283037871,0.2411942035,0.0798239335,-0.1810231358,0.1847556829,-0.2773363292,-0.0575832799,0.0720744729,-0.3603481948,-0.354390502,-0.1715970337,0.1504694819,0.0036926903,0.1574962288,-0.4513111711,-0.0426098816,0.2559912503,-0.3391976357,0.1138054729,-0.1639285982,-0.1804626137,-0.1675998271,0.1240322888,0.0814560801,-0.1796029061,-0.0717220083,-0.2841574252,-0.0085382825,0.2089584321,0.1984399259,-0.1502616256,0.5191944242,-0.2032293975,0.5580466986,0.5535882711,-0.1875405759,-0.2054157257,0.5373868942,0.1265777946,-0.035294205,0.1371379495,-0.4054441452,0.2556640804,-0.1492992491,0.0239111912,0.4544432163,0.1228784025,-0.0366672389,-0.0533076935,0.0591707565,0.094570972,-0.0533153489,-0.4166417718,-0.2638634443,-0.1177741587,0.3790189922,0.2506234944,-0.0887018591,-0.0387289897,0.1767893434,0.0014025921,0.0067821331,-0.0009072539,-0.0842046812,-0.5144618154,0.025190603,-0.3117554784,-0.0911139995,-0.3747565448,0.1402051151,-0.1032433212,0.3250381052,-0.0609508604,-0.2853431404,0.2218149751,0.0576820746,-0.2218000144,-0.317121923,-0.0351766907,0.0971565247,-0.0542958602,0.3626121581,-0.099317044,0.4648391902,0.0869142637,-0.2966175377,0.0466390364,0.1641552597,0.0355114825,-0.2589477897,-0.180991292,0.4646167755,0.4503356814,-0.1478167474,-0.3648447692,-0.3587809801,0.0069056614,0.1058970094,0.0639766827,0.0289983246,0.1435213387,-0.3214043975,0.0080417814,0.3127244115,-0.0123363407,0.4519161284,0.0811970308,-0.0262396168,0.1977178007,-0.0217952374,0.1612604111,-0.3399371505,-0.1020602882,-0.0194319971,-0.0510117151,0.1063097194,0.0067979475,0.0667161867,0.3158483207,0.1552186608,0.1025185212,-0.0363587663,-0.0600491017,-0.1964571029,0.1785123348,0.08895538,0.4353640378,0.0364248864,-0.0735301599,0.042567309,0.1256317198,-0.091405414,0.354873836,0.0301692206,0.4449598491,0.3242442012,0.2033975869,-0.0742958412,-0.3939174116,-0.383282125,0.0276446324,0.118215248,-0.2580774128,0.2396309376,-0.2732917964,0.1311361641,-0.0723114461,-0.031521067,-0.1027374715,-0.3991473615,-0.2967226505,0.3699502349,-0.2913790047,0.0173462927,-0.1271987855,0.0827429891,0.1297951043,-0.1101027355,-0.3194954991,-0.1922740489,-0.1088828892,0.0139107835,0.3567505777,0.26393646,-0.1311806589,0.2827697992,-0.0451867245,-0.048971083,-0.3095544875,0.3275258541,-0.2431077212,-0.2856105268,0.2199111581,0.1815671623,0.1619457006,-0.6107788682,0.1853190362,0.0088885864,-0.2933406532,0.2972552478,0.223001197,-0.0029811368,-0.1525841057,-0.1341142058,-0.1398987919,-0.5762497187,0.4581654668,0.1357269585,0.2581652999,-0.0825648159,-0.0690140352,0.0106100226,-0.2777140737,-0.0837056562,0.0277054086,0.1249312013,0.100296624,-0.3619613945,0.0575923212,0.0255032443,0.0281640198,0.5084940195,-0.1329767108,-0.4221979976,0.4498475492,0.1196181849,-0.0777374431,-0.1762279719,-0.3853220046,0.0696588457,0.396043241,-0.0083335303,-0.1137979776,0.0838596746,0.2298501283,-0.0317384899,0.1729091257,-0.1163412631,0.5597530007,-0.0800676197,0.2650151253,0.0567664243,0.0075933975,0.2237437665,0.0908614248,-0.0370974168,-0.2848674953,-0.1019059196,-0.1746464074,-0.2492518425,0.0721041784,0.1653672457,-0.0519329235,-0.2720279992,-0.072112225,-0.1612115949,-0.1755963713,-0.3624178469,0.3011761606,-0.3059144318,0.0879911706,-0.3111079931,-0.0841155499,0.0438527651,0.0580729954,-0.0536612757,-0.2777520716,0.065991126,-0.1160641313,0.0500696078,0.1661209166,-0.2029540837,0.1418597847,-0.0379835851,0.2989365458,0.0233653262,-0.171015054,-0.0642995536,-0.1698566228,0.1479855776,0.0392503366,0.2059403062,0.2348193973,-0.1164165884,-0.2022816986,-0.0036498476,-0.2349404544,0.440018177,0.5935145617,0.3597708642,-0.3514997661,-0.1708199084,0.3545523286,-0.1483627856,-0.0955266133,-0.1216840819,0.0697270185,-0.2168030739,-0.2290590703,0.0502099358,0.2537956536,0.3231954575,0.1532005668,-0.1422909796,-0.2310124189,-0.447760731,-0.0806013793,0.1426682919,0.1965821832,-0.5051634312,0.1015682593,-0.1520391405,0.2174491286,0.4984903336,0.7618778944,0.2852302492,-0.2908283472,0.2480547726,-0.0173353758,0.4242439568,0.1194377765,-0.1531616449,0.1303167492,-0.2092983574,0.0522411428,0.1517799646,-0.2608374357,0.3953027427,-0.168643415,-0.2760580778,-0.2417637855,0.4991045296,-0.0435558371,-0.012494226,-0.0739121661,-0.2426343709,-0.3518019915,0.6364927888,0.0889950842,0.5242301822,0.4776637256,0.0604504272,0.5830122232,-0.1552409679,0.2822134793,-0.7763892412,-0.0235230736,-0.3103724122,0.1324641109,-0.021711072,-0.2295863181,0.2328011096,0.3454679549,-0.2449895442,0.0065117166,0.0266148224,0.0639140978,-0.0325961113,0.3628292084,-0.1441193074,-0.2786220908,-0.4063245654,0.1032003239,0.0568545312,-0.3126535714,0.0303328279,-0.3509737849,-0.0865488276,-0.2357477397,-0.2639683783,0.4338127375,-0.0716566816,0.3886187971,0.22395657,-0.0562004,0.3615510166,-0.099153474,-0.1680011004,-0.0753625408,0.0291987136,0.1506631374,-0.0272615943,-0.2743914723,-0.0218238253,0.1191586554,0.1346259564,-0.1488002837,0.0384193622,-0.0248725805,-0.1394096613,-0.3718870878,0.2763091624,0.2076297551,-0.4314110279,-0.3709971309,-0.1943300515,0.1156334281,-0.149670139,-0.0934759825,0.1180916354,-0.0015497981,-0.2765603065,-0.3599806428,-0.0708585009,-0.2900587916,0.0538640916,0.354416281,0.2216332406,0.0431616269,0.6239032149,0.2414469421,-0.0021351206,-0.3211764991,0.0817211047,0.3431159854,-0.4442771077,0.3989853263,-0.20757626,-0.1646204591,0.122054033,0.091997847,0.2259000689,0.1671583802,-0.2459678501,-0.5066788793,-0.0920283124,0.3302319348,-0.2621886134,-0.0244904067,-0.1378588229,0.2208992988,-0.3556537032,0.2872291505,-0.243415907,0.2291268557,-0.0172618553,-0.0485382937,-0.0581631362,-0.1316035539,0.1055705175,0.0473976433,0.1401485652,0.1752716601,0.0176036768,-0.1623847783,-0.1138048619,0.1333757192,0.1678049713,0.1552291363,-0.2986271381,0.1163604409,-0.1770166606,0.0377225056,0.1328800321,-0.3939383328,-0.0549995415,0.1374362558,0.4441269934,0.2775044739,-0.0967658013,0.1493074596,-0.1239722371,0.4012761116,-0.0929718241,0.2098116577,-0.1530317217,0.0011428164,0.0350741036,0.1986967027,-0.0442558415,0.0952140391,0.5684457421,-0.1944071203,-0.0607279614,-0.5084893703,0.2955584824,0.5219070911,-0.0011157548,-0.042270191,-0.2038498819,0.1671247333,-0.1614583582,-0.0393575616,0.3196537793,0.0164184812,-0.0983282998,0.2596956789,0.0950077251,-0.0378495455,-0.2987880409,0.1304668188,0.1833295673,0.072718136,0.2018181533,0.6401519775,-0.3257715106,0.0498724915,0.3796495199,0.1183043867,0.606377542,0.2897162437,0.1244285703,0.0713098943,0.420783788,0.0520862304,0.0774490163,-0.20151712,-0.4815178812,0.0902980268,0.2912359536,-0.1424287856,-0.111909695,0.3422737718,0.1441443115,-0.1292780638,-0.2035949975,0.3980275393,-0.1603071243,-0.2040880024,-0.2676356733,-0.1337142885,-0.6330999732,-0.1815863252,-0.0812522322,-0.0243735518,-0.0915029943,0.1115613207,0.0836272165,-0.2546156943,0.1198743135,0.0494740419,-0.0678948909,-0.2596521974,0.4679685235,0.0913939625,0.0372751132,0.0409069173,0.1487784088,0.4960689545,0.2502405345,-0.2635149658,-0.2182382792,-0.1987072527,-0.1470847279,-0.0179430582,0.2585283816,0.0494269542,0.1489408314,0.3585207164,0.114110738,-0.1599735469,0.0438396372,0.1822644323,0.0814697891,0.1857505143,0.6384834647,-0.0477938801,-0.1213485524,0.2243060172,-0.1319791675,-0.2510057688,-0.6355654001,0.2918816805,0.0026242989,0.1969861239,0.0760706961,0.0750238448,0.1019888222,0.638327539,0.0586278252,0.0899371728,-0.103629984,-0.1408049464,-0.0365859717,-0.0300198365,-0.3730284274,0.0104072904,0.1738446802,0.0966818631,0.167988807,0.2986499667,-0.0251649953,0.2913760543,-0.2197563946,-0.1188327298,-0.484356761,0.0687586665,0.2037036866,0.0928060189,-0.1202129647,-0.1476213634,0.0684486479,0.2533197403,0.0239625182,0.09037669,-0.2485013604,0.1168634072,-0.3223131895,0.1750455648,0.0504131168,0.1140335202,-0.0494049378,-0.1896672249,-0.1887531877,-0.5733943582,-0.1219350994,0.4686674476,0.1284101605,0.2680280805,-0.1597730964,0.1188096032,-0.294305563,0.1210640818,-0.4128634334,0.0388882309,-0.345837146,0.2615273893,-0.1557042599,0.2354152501,-0.1095620096,-0.0678942725,-0.2019300908,0.0950917751,-0.207477808,-0.3111889362,0.376058042,0.108802557,-0.232853353,0.464329958,0.1188012734,0.116584979,0.2316484749,-0.3983220756,-0.030120343,0.254344672,-0.1892004758,-0.3222723901,-0.0306952801,0.1550977826,-0.138454631,-0.0962281823,0.1503652036,0.0437868014,-0.1663366854,-0.175719738,0.0138161369]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/355","title":"can't load SNLI dataset","comments":"I just added the processed files of `snli` on our google storage, so that when you do `load_dataset` it can download the processed files from there :)\r\n\r\nWe are thinking about having available those processed files for more datasets in the future, because sometimes files aren't available (like for `snli`), or the download speed is too slow, or sometimes the files take time to be processed.","body":"`nlp` seems to load `snli` from some URL based on nlp.stanford.edu. This subdomain is frequently down -- including right now, when I'd like to load `snli` in a Colab notebook, but can't.\r\n\r\nIs there a plan to move these datasets to huggingface servers for a more stable solution?\r\n\r\nBtw, here's the stack trace:\r\n\r\n```\r\nFile \"\/content\/nlp\/src\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/content\/nlp\/src\/nlp\/builder.py\", line 466, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/content\/nlp\/src\/nlp\/datasets\/snli\/e417f6f2e16254938d977a17ed32f3998f5b23e4fcab0f6eb1d28784f23ea60d\/snli.py\", line 76, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 217, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in download\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/py_utils.py\", line 190, in map_nested\r\n return function(data_struct)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in <lambda>\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/nlp.stanford.edu\/projects\/snli\/snli_1.0.zip\r\n```","comment_length":66,"text":"can't load SNLI dataset\n`nlp` seems to load `snli` from some URL based on nlp.stanford.edu. This subdomain is frequently down -- including right now, when I'd like to load `snli` in a Colab notebook, but can't.\r\n\r\nIs there a plan to move these datasets to huggingface servers for a more stable solution?\r\n\r\nBtw, here's the stack trace:\r\n\r\n```\r\nFile \"\/content\/nlp\/src\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/content\/nlp\/src\/nlp\/builder.py\", line 466, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/content\/nlp\/src\/nlp\/datasets\/snli\/e417f6f2e16254938d977a17ed32f3998f5b23e4fcab0f6eb1d28784f23ea60d\/snli.py\", line 76, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 217, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in download\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/py_utils.py\", line 190, in map_nested\r\n return function(data_struct)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in <lambda>\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/nlp.stanford.edu\/projects\/snli\/snli_1.0.zip\r\n```\nI just added the processed files of `snli` on our google storage, so that when you do `load_dataset` it can download the processed files from there :)\r\n\r\nWe are thinking about having available those processed files for more datasets in the future, because sometimes files aren't available (like for `snli`), or the download speed is too slow, or sometimes the files take time to be processed.","embeddings":[0.1304482073,-0.0483656712,0.059793774,0.4217781425,0.1687606275,-0.2157386839,0.2670191824,0.0161337201,-0.0244341921,-0.1066373214,-0.3873247504,-0.0008527985,0.1453942955,0.4104743004,0.3283267915,-0.0848518759,0.1194742993,-0.0357170701,0.0139390798,-0.1574850529,-0.348336339,0.3196726143,-0.375736624,0.0969095305,-0.0297649503,0.0196623597,-0.2143009752,0.2622815967,-0.1611686647,-0.1393920928,0.4381097257,0.164498955,0.3262892067,0.0996377394,-0.000117559,0.0295298286,0.1056620181,-0.0870951042,-0.5640883446,0.0936509073,-0.324842751,-0.2369283587,-0.0412270017,-0.0553937703,-0.2117430717,-0.0341277607,0.0878446028,-0.1338448077,0.2620665431,0.2927174568,0.1312162429,0.457020551,-0.5215488076,0.0837669373,0.0226054713,-0.1075958535,-0.1660803854,0.2811186016,0.0594205968,-0.2348731458,0.1584093273,0.1544827074,-0.3054082096,-0.391025722,0.1858433038,0.0251871701,-0.0524713881,-0.0781482533,0.1044324785,0.2801420391,0.069898501,0.1527647376,-0.2861181498,-0.3133838475,0.0973004028,-0.0441063717,0.1898174584,0.1714379638,0.1277593672,0.3028598428,-0.0501641259,-0.2644038498,-0.0007003463,0.2948257923,-0.1543981135,0.6570668221,-0.1417783648,0.1334464103,0.343195796,-0.0286403541,0.0854430571,-0.0099403653,0.0727775246,0.1083649546,-0.3882722855,0.0951081514,-0.1398562491,0.3578723669,0.021769397,0.326841563,0.2466684282,-0.0482897162,0.0075823609,0.1901656985,0.0771290883,0.1908450723,0.0269667581,-0.2320130914,0.5072234869,0.2394815236,0.2719387114,0.0962707847,-0.2195151895,-0.2679294348,-0.5367240906,-0.3278502524,0.0782074779,-0.2093335092,-0.1450171471,-0.1577027738,-0.0719775483,0.1155304983,0.1822850853,0.2834401429,-0.0627139285,0.5076025724,0.1111344695,0.1902718544,-0.2350015044,-0.2010626197,-0.1571724117,0.0150193674,-0.3410459161,0.3710554838,0.1181917563,-0.3645623624,0.1672880203,-0.0471499227,-0.1467894614,0.1659791023,-0.1471507847,0.3010153174,-0.03390342,0.1608228981,-0.0406754129,0.155169934,-0.028171353,-0.1401581466,-0.1304618865,-0.1663203984,-0.1513635218,-0.3299483657,-0.0857278407,0.0310140047,-0.3338066041,-0.2495877594,-0.4519361556,0.0037020426,0.0267700646,-0.1050219312,-0.0672885627,-0.0789620429,-0.238785848,-0.1213366464,0.3064864874,0.5235179067,0.2027140111,-0.3623385131,-0.2355785221,0.0376874134,0.1617934704,0.6364001036,-0.1188412756,-0.4001249969,-0.1824987233,0.1775024086,0.2002283484,-0.1794870049,-0.3256947696,0.2752823234,-0.3613670468,-0.1439986974,0.2223538607,0.1931883544,-0.1639113724,-0.0499501638,0.1892511249,0.6330448389,0.1260435879,-0.0856131613,-0.4507155418,-0.3640879691,0.0047560032,0.2229622453,0.1132772043,-0.0371184796,-0.1291671842,0.230927825,0.2957028747,-0.1445367038,0.0546408296,-0.0628841147,0.228770718,0.2753065228,-0.0180238392,0.1458810717,-0.3736081421,0.1601790786,-0.0156620312,0.4346553385,0.0774695277,0.0404837169,-0.1789983958,-0.0094729289,-0.1319319606,-0.1548575908,-0.0453127883,0.3207741976,0.2329060882,0.1456717849,-0.2976048589,0.2277678549,-0.2371035963,0.2496478856,-0.5744193792,0.5656714439,0.1177590191,0.2276495844,-0.011413021,0.319873482,0.0307091698,-0.1356557757,0.1645552218,0.0994904861,-0.3558295667,0.2359803766,0.4676783979,0.087553747,0.2638019621,-0.2606596351,0.4226095676,-0.0008728617,-0.056670703,0.0165115315,0.137795046,0.5027769804,-0.1210881695,0.4202392697,0.2470627725,0.2384273708,0.4183702469,-0.0620677955,-0.0541550741,-0.0759209841,0.20869039,0.243683219,0.4183528721,0.4291169643,-0.4238772094,0.1012035012,0.0573867261,-0.2955073416,0.0375080407,0.0305925068,-0.3085520267,-0.2286909372,0.3273907304,0.4721651673,0.2052293718,0.0524367541,-0.0489891656,-0.0664653108,-0.1507889926,-0.5673172474,0.3456598818,-0.0326571651,-0.0688784719,-0.0018092771,0.2520260513,-0.0300832093,-0.6293240786,-0.0304291733,-0.0839683637,-0.0329370573,-0.2263214886,0.197623089,-0.3592180908,-0.6340147257,-0.5378745794,0.0821900517,-0.3298075497,-0.5116081834,0.0921253562,0.1828790009,0.0226485208,-0.063473478,-0.0067485287,0.6645886302,-0.4301497042,-0.0604205877,-0.1470168829,-0.5190829635,-0.3770621419,0.0634477362,0.3587060571,0.200772509,0.1039768681,-0.26919505,-0.1576095521,-0.3166330755,-0.4544120729,0.0933533683,-0.0087030036,0.0978781506,-0.0349688865,0.3597619832,0.1173564121,-0.1316226274,0.2207645923,-0.1457942873,0.1616683453,0.3093687594,-0.1411330402,-0.0629481077,0.0991666466,-0.1920919567,-0.3986883461,-0.2858089209,0.158243984,-0.4708935916,-0.0865265802,0.4178456068,-0.1209960356,0.1282298565,-0.0140495216,-0.1653679162,-0.0069052121,-0.1221962199,0.3401512504,-0.0845435411,-0.1688057035,-0.0179712996,0.1193795726,0.2667138278,-0.0978106707,-0.5858650208,-0.0070258765,-0.0959324166,0.0875686035,0.3945233822,-0.062265709,-0.0200650394,-0.1766350716,0.0149214873,0.113628909,-0.1519600898,-0.0434901454,0.1634542197,0.3132613897,0.1599602699,0.2865488827,-0.0055425023,0.4817097783,0.1003508791,-0.1031534821,0.4657756984,0.1412285864,0.1298272312,0.0549871661,-0.4065544307,0.1902965307,0.0333689824,-0.3419718444,0.4272550642,0.1819275767,0.1205514818,-0.3684725463,-0.2022230029,0.1697643846,-0.2676969469,0.0530468076,0.2985784411,0.0603098162,0.0972270295,-0.1055049673,-0.0377928801,-0.1585014462,-0.0520994365,0.4101315737,0.0977639407,0.0250490569,-0.2611637115,-0.0920041576,-0.5408343673,0.3158062696,0.0527636856,0.2028032094,-0.081078127,-0.0301086549,0.2543749213,0.0679179206,0.2663500309,-0.1048046798,0.1423966587,0.0410783403,-0.0973922536,-0.4929026067,0.0652067438,0.1237916499,-0.0717306361,0.4792842269,0.3520555198,-0.0394820161,-0.2428181022,0.1453234404,0.23052001,-0.243587032,0.0669356212,-0.1930082887,-0.5233623981,-0.1967772543,-0.1565243006,-0.1726042926,0.0737290382,-0.0955599621,0.0560566895,0.0631453618,-0.1865683347,0.1151105016,0.3434083462,0.0233111586,0.2734473348,-0.1275288016,0.0795082822,0.3538227379,0.1347263604,0.7317372561,-0.1766162068,-0.3419310749,0.142473489,0.3091111183,0.4259915352,0.551543653,0.0122105302,-0.0027419277,0.2469418049,0.169174403,-0.0710203499,0.0248558484,0.1896543056,-0.0271573644,-0.2438091934,-0.3387823403,0.6757808328,0.1753596961,-0.0686709732,-0.0196814202,0.1879802197,-0.1997472495,0.3174859285,0.196384415,1.112601757,-0.3865783811,0.2527160347,0.4754591286,0.1300272048,0.7861910462,-0.0476480052,0.0393128358,-0.175775826,-0.0386876054,-0.0158695821,0.1026389748,-0.2545431554,0.0931004807,0.1193133816,0.3765717745,0.1462594569,0.1244838163,0.17723234,0.4552799165,-0.0149794305,-0.1715274304,-0.9147063494,0.0899304599,-0.0443625525,0.1974340379,-0.1149675623,-0.1363938898,-0.1437603682,0.2531207204,-0.064076595,0.3095707595,0.0926276669,0.216854617,-0.3098124862,-0.0930678099,-0.1108041182,0.2111047804,0.4911542535,-0.1463082433,-0.2870134413,0.0100247413,-0.2619129717,-0.1846578121,-0.0040820618,0.1146591678,0.2568469346,0.0299446266,-0.0322138518,-0.0952626467,0.2902306318,-0.2916546464,-0.3434294164,-0.1409877241,-0.3479572833,-0.0593659244,0.1842235178,0.1704667509,0.0988735482,-0.1592891961,0.0438034944,0.0201071743,0.1397152543,-0.3530656993,0.1558365375,-0.2608713806,-0.0902478248,0.422477901,-0.1290141791,0.2367495596,0.2137132287,0.2999237478,0.041608043,-0.1443562359,-0.2040252537,-0.1250776201,-0.3493469954,0.2267754525,0.026972441,-0.2679871917,-0.1560677141,0.4593490362,0.2066574097,-0.3265994191,-0.1299116611,-0.4308762848,0.0066580661,0.3610094488,-0.0963620618,0.0311282147,0.328096807,0.0480952784,-0.0867451876,-0.1157348827,-0.2161809951,0.0298732258,-0.2376494855,-0.0058725709,0.3596525788,-0.1470711231,0.196218729,0.1221146062,-0.0142046446,-0.2798121274,0.0494470485,-0.0729348287,-0.1754334569,0.170799762,0.258508265,-0.2010315508,-0.1681411415,-0.2821453214,0.064847149,-0.288736403,0.0232700724,0.0604452528,0.0497078821,0.1475997567,-0.1680620313,-0.2170601338,-0.032462921,0.4655557871,0.0711520612,-0.0806486979,0.2655051649,0.1415583938,0.2142727077,-0.0884355903,-0.1847543418,0.047533188,0.0450642444,0.005697865,0.2424156815,-0.2242477536,-0.180459708,0.2852202654,0.0655076727,0.6470024586,0.2247956246,0.0137375575,-0.0031777136,0.0066327401,-0.1387801021,-0.0193507839,0.3327872157,0.0699873418,-0.2775742412,0.2848497033,0.3158890307,-0.2858683467,0.0928034857,-0.0361051485,0.2029639035,0.1325009912,0.2233767509,0.2039144933,0.2522875071,0.1599001437,0.0533145629,-0.0269811824,0.3574564457,-0.0929626599,-0.2084430903,0.2846877277,-0.4164149165,0.3620056212,0.1814380139,-0.0925890058,-0.1739553064,0.2266274542,-0.1083585769,-0.023672374,0.0493286811,0.8438782692,-0.3339235485,-0.2409933954,-0.3582416773,0.6556246281,-0.2454179227,-0.2910168171,-0.501608789,-0.1364713907,-0.179775849,0.2041600496,-0.3215113878,-0.2917788327,0.288403064,0.2090781331,-0.0305051748,-0.6123155355,-0.1765327156,0.3434591889,0.1124517098,-0.1138048545,0.3136467636,0.0580066144,-0.20412606,0.11886096,0.0968801975,0.3500439823,0.2010899931,-0.0146963773,-0.0477133207,0.1254771203,-0.0012670903,-0.2866496146,0.1265265197,-0.0653265864,0.2059822679,0.2354422808,0.0774263591,-0.0531416796,0.0560033172,-0.0051431498,0.1339714825,-0.239379555,0.2333390713,0.0041830079,-0.037076775,-0.3344923258,0.0534916259,-0.3265328705,0.1424583644,0.1018857211,-0.2222626209,0.0846439674,-0.1736570299,0.0138783967,0.1851200461,0.7972928286,0.3956439197,0.0028446128,-0.2536246479,-0.3362244964,-0.7480276823,-0.0303410869,-0.0455267169,0.0314102806,-0.0989700183,0.0264583584,0.0147554036,0.0864960402,-0.5139561892,0.3119202256,0.0454297662,-0.1438587606,-0.2488614917,-0.0460019484,0.0006246938,-0.1093419716,-0.0308333486,-0.1851891428,0.1970299631,-0.0995515063,-0.1593070328,0.0405900516,0.1372340322,-0.1498680115,-0.1960590482,0.3116006553,0.2188539952,0.2902201712,0.0941513926,-0.153239727,-0.3999745548,-0.3762974441,-0.2154217362,-0.0351465382,-0.1620055437,-0.0105525842,-0.0265235361,0.3910281956,-0.269816637,0.1867967993,0.1291775107,0.1626279354,-0.2214720845,-0.1415459961,-0.282215625,0.0769395605,0.1828295141,0.3219568133,-0.1019818559,-0.1648567468,-0.0388853252,-0.2984727025,0.3104025126,-0.667491734,-0.0051825256,0.2318101823,0.1003702134,0.19054313,-0.0044605597,-0.280321151,0.0216780696,-0.040231023,0.0536148027,-0.0572229102,0.2559002638,0.2245052308,-0.2706518173,-0.2282407582,-0.2176375389,0.2312410623,-0.3037879765,0.1339839399,-0.3156906068]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/355","title":"can't load SNLI dataset","comments":"Closing this one. Feel free to re-open if you have other questions :)","body":"`nlp` seems to load `snli` from some URL based on nlp.stanford.edu. This subdomain is frequently down -- including right now, when I'd like to load `snli` in a Colab notebook, but can't.\r\n\r\nIs there a plan to move these datasets to huggingface servers for a more stable solution?\r\n\r\nBtw, here's the stack trace:\r\n\r\n```\r\nFile \"\/content\/nlp\/src\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/content\/nlp\/src\/nlp\/builder.py\", line 466, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/content\/nlp\/src\/nlp\/datasets\/snli\/e417f6f2e16254938d977a17ed32f3998f5b23e4fcab0f6eb1d28784f23ea60d\/snli.py\", line 76, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 217, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in download\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/py_utils.py\", line 190, in map_nested\r\n return function(data_struct)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in <lambda>\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/nlp.stanford.edu\/projects\/snli\/snli_1.0.zip\r\n```","comment_length":13,"text":"can't load SNLI dataset\n`nlp` seems to load `snli` from some URL based on nlp.stanford.edu. This subdomain is frequently down -- including right now, when I'd like to load `snli` in a Colab notebook, but can't.\r\n\r\nIs there a plan to move these datasets to huggingface servers for a more stable solution?\r\n\r\nBtw, here's the stack trace:\r\n\r\n```\r\nFile \"\/content\/nlp\/src\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/content\/nlp\/src\/nlp\/builder.py\", line 466, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/content\/nlp\/src\/nlp\/datasets\/snli\/e417f6f2e16254938d977a17ed32f3998f5b23e4fcab0f6eb1d28784f23ea60d\/snli.py\", line 76, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 217, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in download\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/py_utils.py\", line 190, in map_nested\r\n return function(data_struct)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in <lambda>\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/nlp.stanford.edu\/projects\/snli\/snli_1.0.zip\r\n```\nClosing this one. Feel free to re-open if you have other questions :)","embeddings":[0.1304482073,-0.0483656712,0.059793774,0.4217781425,0.1687606275,-0.2157386839,0.2670191824,0.0161337201,-0.0244341921,-0.1066373214,-0.3873247504,-0.0008527985,0.1453942955,0.4104743004,0.3283267915,-0.0848518759,0.1194742993,-0.0357170701,0.0139390798,-0.1574850529,-0.348336339,0.3196726143,-0.375736624,0.0969095305,-0.0297649503,0.0196623597,-0.2143009752,0.2622815967,-0.1611686647,-0.1393920928,0.4381097257,0.164498955,0.3262892067,0.0996377394,-0.000117559,0.0295298286,0.1056620181,-0.0870951042,-0.5640883446,0.0936509073,-0.324842751,-0.2369283587,-0.0412270017,-0.0553937703,-0.2117430717,-0.0341277607,0.0878446028,-0.1338448077,0.2620665431,0.2927174568,0.1312162429,0.457020551,-0.5215488076,0.0837669373,0.0226054713,-0.1075958535,-0.1660803854,0.2811186016,0.0594205968,-0.2348731458,0.1584093273,0.1544827074,-0.3054082096,-0.391025722,0.1858433038,0.0251871701,-0.0524713881,-0.0781482533,0.1044324785,0.2801420391,0.069898501,0.1527647376,-0.2861181498,-0.3133838475,0.0973004028,-0.0441063717,0.1898174584,0.1714379638,0.1277593672,0.3028598428,-0.0501641259,-0.2644038498,-0.0007003463,0.2948257923,-0.1543981135,0.6570668221,-0.1417783648,0.1334464103,0.343195796,-0.0286403541,0.0854430571,-0.0099403653,0.0727775246,0.1083649546,-0.3882722855,0.0951081514,-0.1398562491,0.3578723669,0.021769397,0.326841563,0.2466684282,-0.0482897162,0.0075823609,0.1901656985,0.0771290883,0.1908450723,0.0269667581,-0.2320130914,0.5072234869,0.2394815236,0.2719387114,0.0962707847,-0.2195151895,-0.2679294348,-0.5367240906,-0.3278502524,0.0782074779,-0.2093335092,-0.1450171471,-0.1577027738,-0.0719775483,0.1155304983,0.1822850853,0.2834401429,-0.0627139285,0.5076025724,0.1111344695,0.1902718544,-0.2350015044,-0.2010626197,-0.1571724117,0.0150193674,-0.3410459161,0.3710554838,0.1181917563,-0.3645623624,0.1672880203,-0.0471499227,-0.1467894614,0.1659791023,-0.1471507847,0.3010153174,-0.03390342,0.1608228981,-0.0406754129,0.155169934,-0.028171353,-0.1401581466,-0.1304618865,-0.1663203984,-0.1513635218,-0.3299483657,-0.0857278407,0.0310140047,-0.3338066041,-0.2495877594,-0.4519361556,0.0037020426,0.0267700646,-0.1050219312,-0.0672885627,-0.0789620429,-0.238785848,-0.1213366464,0.3064864874,0.5235179067,0.2027140111,-0.3623385131,-0.2355785221,0.0376874134,0.1617934704,0.6364001036,-0.1188412756,-0.4001249969,-0.1824987233,0.1775024086,0.2002283484,-0.1794870049,-0.3256947696,0.2752823234,-0.3613670468,-0.1439986974,0.2223538607,0.1931883544,-0.1639113724,-0.0499501638,0.1892511249,0.6330448389,0.1260435879,-0.0856131613,-0.4507155418,-0.3640879691,0.0047560032,0.2229622453,0.1132772043,-0.0371184796,-0.1291671842,0.230927825,0.2957028747,-0.1445367038,0.0546408296,-0.0628841147,0.228770718,0.2753065228,-0.0180238392,0.1458810717,-0.3736081421,0.1601790786,-0.0156620312,0.4346553385,0.0774695277,0.0404837169,-0.1789983958,-0.0094729289,-0.1319319606,-0.1548575908,-0.0453127883,0.3207741976,0.2329060882,0.1456717849,-0.2976048589,0.2277678549,-0.2371035963,0.2496478856,-0.5744193792,0.5656714439,0.1177590191,0.2276495844,-0.011413021,0.319873482,0.0307091698,-0.1356557757,0.1645552218,0.0994904861,-0.3558295667,0.2359803766,0.4676783979,0.087553747,0.2638019621,-0.2606596351,0.4226095676,-0.0008728617,-0.056670703,0.0165115315,0.137795046,0.5027769804,-0.1210881695,0.4202392697,0.2470627725,0.2384273708,0.4183702469,-0.0620677955,-0.0541550741,-0.0759209841,0.20869039,0.243683219,0.4183528721,0.4291169643,-0.4238772094,0.1012035012,0.0573867261,-0.2955073416,0.0375080407,0.0305925068,-0.3085520267,-0.2286909372,0.3273907304,0.4721651673,0.2052293718,0.0524367541,-0.0489891656,-0.0664653108,-0.1507889926,-0.5673172474,0.3456598818,-0.0326571651,-0.0688784719,-0.0018092771,0.2520260513,-0.0300832093,-0.6293240786,-0.0304291733,-0.0839683637,-0.0329370573,-0.2263214886,0.197623089,-0.3592180908,-0.6340147257,-0.5378745794,0.0821900517,-0.3298075497,-0.5116081834,0.0921253562,0.1828790009,0.0226485208,-0.063473478,-0.0067485287,0.6645886302,-0.4301497042,-0.0604205877,-0.1470168829,-0.5190829635,-0.3770621419,0.0634477362,0.3587060571,0.200772509,0.1039768681,-0.26919505,-0.1576095521,-0.3166330755,-0.4544120729,0.0933533683,-0.0087030036,0.0978781506,-0.0349688865,0.3597619832,0.1173564121,-0.1316226274,0.2207645923,-0.1457942873,0.1616683453,0.3093687594,-0.1411330402,-0.0629481077,0.0991666466,-0.1920919567,-0.3986883461,-0.2858089209,0.158243984,-0.4708935916,-0.0865265802,0.4178456068,-0.1209960356,0.1282298565,-0.0140495216,-0.1653679162,-0.0069052121,-0.1221962199,0.3401512504,-0.0845435411,-0.1688057035,-0.0179712996,0.1193795726,0.2667138278,-0.0978106707,-0.5858650208,-0.0070258765,-0.0959324166,0.0875686035,0.3945233822,-0.062265709,-0.0200650394,-0.1766350716,0.0149214873,0.113628909,-0.1519600898,-0.0434901454,0.1634542197,0.3132613897,0.1599602699,0.2865488827,-0.0055425023,0.4817097783,0.1003508791,-0.1031534821,0.4657756984,0.1412285864,0.1298272312,0.0549871661,-0.4065544307,0.1902965307,0.0333689824,-0.3419718444,0.4272550642,0.1819275767,0.1205514818,-0.3684725463,-0.2022230029,0.1697643846,-0.2676969469,0.0530468076,0.2985784411,0.0603098162,0.0972270295,-0.1055049673,-0.0377928801,-0.1585014462,-0.0520994365,0.4101315737,0.0977639407,0.0250490569,-0.2611637115,-0.0920041576,-0.5408343673,0.3158062696,0.0527636856,0.2028032094,-0.081078127,-0.0301086549,0.2543749213,0.0679179206,0.2663500309,-0.1048046798,0.1423966587,0.0410783403,-0.0973922536,-0.4929026067,0.0652067438,0.1237916499,-0.0717306361,0.4792842269,0.3520555198,-0.0394820161,-0.2428181022,0.1453234404,0.23052001,-0.243587032,0.0669356212,-0.1930082887,-0.5233623981,-0.1967772543,-0.1565243006,-0.1726042926,0.0737290382,-0.0955599621,0.0560566895,0.0631453618,-0.1865683347,0.1151105016,0.3434083462,0.0233111586,0.2734473348,-0.1275288016,0.0795082822,0.3538227379,0.1347263604,0.7317372561,-0.1766162068,-0.3419310749,0.142473489,0.3091111183,0.4259915352,0.551543653,0.0122105302,-0.0027419277,0.2469418049,0.169174403,-0.0710203499,0.0248558484,0.1896543056,-0.0271573644,-0.2438091934,-0.3387823403,0.6757808328,0.1753596961,-0.0686709732,-0.0196814202,0.1879802197,-0.1997472495,0.3174859285,0.196384415,1.112601757,-0.3865783811,0.2527160347,0.4754591286,0.1300272048,0.7861910462,-0.0476480052,0.0393128358,-0.175775826,-0.0386876054,-0.0158695821,0.1026389748,-0.2545431554,0.0931004807,0.1193133816,0.3765717745,0.1462594569,0.1244838163,0.17723234,0.4552799165,-0.0149794305,-0.1715274304,-0.9147063494,0.0899304599,-0.0443625525,0.1974340379,-0.1149675623,-0.1363938898,-0.1437603682,0.2531207204,-0.064076595,0.3095707595,0.0926276669,0.216854617,-0.3098124862,-0.0930678099,-0.1108041182,0.2111047804,0.4911542535,-0.1463082433,-0.2870134413,0.0100247413,-0.2619129717,-0.1846578121,-0.0040820618,0.1146591678,0.2568469346,0.0299446266,-0.0322138518,-0.0952626467,0.2902306318,-0.2916546464,-0.3434294164,-0.1409877241,-0.3479572833,-0.0593659244,0.1842235178,0.1704667509,0.0988735482,-0.1592891961,0.0438034944,0.0201071743,0.1397152543,-0.3530656993,0.1558365375,-0.2608713806,-0.0902478248,0.422477901,-0.1290141791,0.2367495596,0.2137132287,0.2999237478,0.041608043,-0.1443562359,-0.2040252537,-0.1250776201,-0.3493469954,0.2267754525,0.026972441,-0.2679871917,-0.1560677141,0.4593490362,0.2066574097,-0.3265994191,-0.1299116611,-0.4308762848,0.0066580661,0.3610094488,-0.0963620618,0.0311282147,0.328096807,0.0480952784,-0.0867451876,-0.1157348827,-0.2161809951,0.0298732258,-0.2376494855,-0.0058725709,0.3596525788,-0.1470711231,0.196218729,0.1221146062,-0.0142046446,-0.2798121274,0.0494470485,-0.0729348287,-0.1754334569,0.170799762,0.258508265,-0.2010315508,-0.1681411415,-0.2821453214,0.064847149,-0.288736403,0.0232700724,0.0604452528,0.0497078821,0.1475997567,-0.1680620313,-0.2170601338,-0.032462921,0.4655557871,0.0711520612,-0.0806486979,0.2655051649,0.1415583938,0.2142727077,-0.0884355903,-0.1847543418,0.047533188,0.0450642444,0.005697865,0.2424156815,-0.2242477536,-0.180459708,0.2852202654,0.0655076727,0.6470024586,0.2247956246,0.0137375575,-0.0031777136,0.0066327401,-0.1387801021,-0.0193507839,0.3327872157,0.0699873418,-0.2775742412,0.2848497033,0.3158890307,-0.2858683467,0.0928034857,-0.0361051485,0.2029639035,0.1325009912,0.2233767509,0.2039144933,0.2522875071,0.1599001437,0.0533145629,-0.0269811824,0.3574564457,-0.0929626599,-0.2084430903,0.2846877277,-0.4164149165,0.3620056212,0.1814380139,-0.0925890058,-0.1739553064,0.2266274542,-0.1083585769,-0.023672374,0.0493286811,0.8438782692,-0.3339235485,-0.2409933954,-0.3582416773,0.6556246281,-0.2454179227,-0.2910168171,-0.501608789,-0.1364713907,-0.179775849,0.2041600496,-0.3215113878,-0.2917788327,0.288403064,0.2090781331,-0.0305051748,-0.6123155355,-0.1765327156,0.3434591889,0.1124517098,-0.1138048545,0.3136467636,0.0580066144,-0.20412606,0.11886096,0.0968801975,0.3500439823,0.2010899931,-0.0146963773,-0.0477133207,0.1254771203,-0.0012670903,-0.2866496146,0.1265265197,-0.0653265864,0.2059822679,0.2354422808,0.0774263591,-0.0531416796,0.0560033172,-0.0051431498,0.1339714825,-0.239379555,0.2333390713,0.0041830079,-0.037076775,-0.3344923258,0.0534916259,-0.3265328705,0.1424583644,0.1018857211,-0.2222626209,0.0846439674,-0.1736570299,0.0138783967,0.1851200461,0.7972928286,0.3956439197,0.0028446128,-0.2536246479,-0.3362244964,-0.7480276823,-0.0303410869,-0.0455267169,0.0314102806,-0.0989700183,0.0264583584,0.0147554036,0.0864960402,-0.5139561892,0.3119202256,0.0454297662,-0.1438587606,-0.2488614917,-0.0460019484,0.0006246938,-0.1093419716,-0.0308333486,-0.1851891428,0.1970299631,-0.0995515063,-0.1593070328,0.0405900516,0.1372340322,-0.1498680115,-0.1960590482,0.3116006553,0.2188539952,0.2902201712,0.0941513926,-0.153239727,-0.3999745548,-0.3762974441,-0.2154217362,-0.0351465382,-0.1620055437,-0.0105525842,-0.0265235361,0.3910281956,-0.269816637,0.1867967993,0.1291775107,0.1626279354,-0.2214720845,-0.1415459961,-0.282215625,0.0769395605,0.1828295141,0.3219568133,-0.1019818559,-0.1648567468,-0.0388853252,-0.2984727025,0.3104025126,-0.667491734,-0.0051825256,0.2318101823,0.1003702134,0.19054313,-0.0044605597,-0.280321151,0.0216780696,-0.040231023,0.0536148027,-0.0572229102,0.2559002638,0.2245052308,-0.2706518173,-0.2282407582,-0.2176375389,0.2312410623,-0.3037879765,0.1339839399,-0.3156906068]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/355","title":"can't load SNLI dataset","comments":"Thank you!","body":"`nlp` seems to load `snli` from some URL based on nlp.stanford.edu. This subdomain is frequently down -- including right now, when I'd like to load `snli` in a Colab notebook, but can't.\r\n\r\nIs there a plan to move these datasets to huggingface servers for a more stable solution?\r\n\r\nBtw, here's the stack trace:\r\n\r\n```\r\nFile \"\/content\/nlp\/src\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/content\/nlp\/src\/nlp\/builder.py\", line 466, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/content\/nlp\/src\/nlp\/datasets\/snli\/e417f6f2e16254938d977a17ed32f3998f5b23e4fcab0f6eb1d28784f23ea60d\/snli.py\", line 76, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 217, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in download\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/py_utils.py\", line 190, in map_nested\r\n return function(data_struct)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in <lambda>\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/nlp.stanford.edu\/projects\/snli\/snli_1.0.zip\r\n```","comment_length":2,"text":"can't load SNLI dataset\n`nlp` seems to load `snli` from some URL based on nlp.stanford.edu. This subdomain is frequently down -- including right now, when I'd like to load `snli` in a Colab notebook, but can't.\r\n\r\nIs there a plan to move these datasets to huggingface servers for a more stable solution?\r\n\r\nBtw, here's the stack trace:\r\n\r\n```\r\nFile \"\/content\/nlp\/src\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/content\/nlp\/src\/nlp\/builder.py\", line 466, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/content\/nlp\/src\/nlp\/datasets\/snli\/e417f6f2e16254938d977a17ed32f3998f5b23e4fcab0f6eb1d28784f23ea60d\/snli.py\", line 76, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 217, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in download\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/py_utils.py\", line 190, in map_nested\r\n return function(data_struct)\r\n File \"\/content\/nlp\/src\/nlp\/utils\/download_manager.py\", line 156, in <lambda>\r\n lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 198, in cached_path\r\n local_files_only=download_config.local_files_only,\r\n File \"\/content\/nlp\/src\/nlp\/utils\/file_utils.py\", line 356, in get_from_cache\r\n raise ConnectionError(\"Couldn't reach {}\".format(url))\r\nConnectionError: Couldn't reach https:\/\/nlp.stanford.edu\/projects\/snli\/snli_1.0.zip\r\n```\nThank you!","embeddings":[0.1304482073,-0.0483656712,0.059793774,0.4217781425,0.1687606275,-0.2157386839,0.2670191824,0.0161337201,-0.0244341921,-0.1066373214,-0.3873247504,-0.0008527985,0.1453942955,0.4104743004,0.3283267915,-0.0848518759,0.1194742993,-0.0357170701,0.0139390798,-0.1574850529,-0.348336339,0.3196726143,-0.375736624,0.0969095305,-0.0297649503,0.0196623597,-0.2143009752,0.2622815967,-0.1611686647,-0.1393920928,0.4381097257,0.164498955,0.3262892067,0.0996377394,-0.000117559,0.0295298286,0.1056620181,-0.0870951042,-0.5640883446,0.0936509073,-0.324842751,-0.2369283587,-0.0412270017,-0.0553937703,-0.2117430717,-0.0341277607,0.0878446028,-0.1338448077,0.2620665431,0.2927174568,0.1312162429,0.457020551,-0.5215488076,0.0837669373,0.0226054713,-0.1075958535,-0.1660803854,0.2811186016,0.0594205968,-0.2348731458,0.1584093273,0.1544827074,-0.3054082096,-0.391025722,0.1858433038,0.0251871701,-0.0524713881,-0.0781482533,0.1044324785,0.2801420391,0.069898501,0.1527647376,-0.2861181498,-0.3133838475,0.0973004028,-0.0441063717,0.1898174584,0.1714379638,0.1277593672,0.3028598428,-0.0501641259,-0.2644038498,-0.0007003463,0.2948257923,-0.1543981135,0.6570668221,-0.1417783648,0.1334464103,0.343195796,-0.0286403541,0.0854430571,-0.0099403653,0.0727775246,0.1083649546,-0.3882722855,0.0951081514,-0.1398562491,0.3578723669,0.021769397,0.326841563,0.2466684282,-0.0482897162,0.0075823609,0.1901656985,0.0771290883,0.1908450723,0.0269667581,-0.2320130914,0.5072234869,0.2394815236,0.2719387114,0.0962707847,-0.2195151895,-0.2679294348,-0.5367240906,-0.3278502524,0.0782074779,-0.2093335092,-0.1450171471,-0.1577027738,-0.0719775483,0.1155304983,0.1822850853,0.2834401429,-0.0627139285,0.5076025724,0.1111344695,0.1902718544,-0.2350015044,-0.2010626197,-0.1571724117,0.0150193674,-0.3410459161,0.3710554838,0.1181917563,-0.3645623624,0.1672880203,-0.0471499227,-0.1467894614,0.1659791023,-0.1471507847,0.3010153174,-0.03390342,0.1608228981,-0.0406754129,0.155169934,-0.028171353,-0.1401581466,-0.1304618865,-0.1663203984,-0.1513635218,-0.3299483657,-0.0857278407,0.0310140047,-0.3338066041,-0.2495877594,-0.4519361556,0.0037020426,0.0267700646,-0.1050219312,-0.0672885627,-0.0789620429,-0.238785848,-0.1213366464,0.3064864874,0.5235179067,0.2027140111,-0.3623385131,-0.2355785221,0.0376874134,0.1617934704,0.6364001036,-0.1188412756,-0.4001249969,-0.1824987233,0.1775024086,0.2002283484,-0.1794870049,-0.3256947696,0.2752823234,-0.3613670468,-0.1439986974,0.2223538607,0.1931883544,-0.1639113724,-0.0499501638,0.1892511249,0.6330448389,0.1260435879,-0.0856131613,-0.4507155418,-0.3640879691,0.0047560032,0.2229622453,0.1132772043,-0.0371184796,-0.1291671842,0.230927825,0.2957028747,-0.1445367038,0.0546408296,-0.0628841147,0.228770718,0.2753065228,-0.0180238392,0.1458810717,-0.3736081421,0.1601790786,-0.0156620312,0.4346553385,0.0774695277,0.0404837169,-0.1789983958,-0.0094729289,-0.1319319606,-0.1548575908,-0.0453127883,0.3207741976,0.2329060882,0.1456717849,-0.2976048589,0.2277678549,-0.2371035963,0.2496478856,-0.5744193792,0.5656714439,0.1177590191,0.2276495844,-0.011413021,0.319873482,0.0307091698,-0.1356557757,0.1645552218,0.0994904861,-0.3558295667,0.2359803766,0.4676783979,0.087553747,0.2638019621,-0.2606596351,0.4226095676,-0.0008728617,-0.056670703,0.0165115315,0.137795046,0.5027769804,-0.1210881695,0.4202392697,0.2470627725,0.2384273708,0.4183702469,-0.0620677955,-0.0541550741,-0.0759209841,0.20869039,0.243683219,0.4183528721,0.4291169643,-0.4238772094,0.1012035012,0.0573867261,-0.2955073416,0.0375080407,0.0305925068,-0.3085520267,-0.2286909372,0.3273907304,0.4721651673,0.2052293718,0.0524367541,-0.0489891656,-0.0664653108,-0.1507889926,-0.5673172474,0.3456598818,-0.0326571651,-0.0688784719,-0.0018092771,0.2520260513,-0.0300832093,-0.6293240786,-0.0304291733,-0.0839683637,-0.0329370573,-0.2263214886,0.197623089,-0.3592180908,-0.6340147257,-0.5378745794,0.0821900517,-0.3298075497,-0.5116081834,0.0921253562,0.1828790009,0.0226485208,-0.063473478,-0.0067485287,0.6645886302,-0.4301497042,-0.0604205877,-0.1470168829,-0.5190829635,-0.3770621419,0.0634477362,0.3587060571,0.200772509,0.1039768681,-0.26919505,-0.1576095521,-0.3166330755,-0.4544120729,0.0933533683,-0.0087030036,0.0978781506,-0.0349688865,0.3597619832,0.1173564121,-0.1316226274,0.2207645923,-0.1457942873,0.1616683453,0.3093687594,-0.1411330402,-0.0629481077,0.0991666466,-0.1920919567,-0.3986883461,-0.2858089209,0.158243984,-0.4708935916,-0.0865265802,0.4178456068,-0.1209960356,0.1282298565,-0.0140495216,-0.1653679162,-0.0069052121,-0.1221962199,0.3401512504,-0.0845435411,-0.1688057035,-0.0179712996,0.1193795726,0.2667138278,-0.0978106707,-0.5858650208,-0.0070258765,-0.0959324166,0.0875686035,0.3945233822,-0.062265709,-0.0200650394,-0.1766350716,0.0149214873,0.113628909,-0.1519600898,-0.0434901454,0.1634542197,0.3132613897,0.1599602699,0.2865488827,-0.0055425023,0.4817097783,0.1003508791,-0.1031534821,0.4657756984,0.1412285864,0.1298272312,0.0549871661,-0.4065544307,0.1902965307,0.0333689824,-0.3419718444,0.4272550642,0.1819275767,0.1205514818,-0.3684725463,-0.2022230029,0.1697643846,-0.2676969469,0.0530468076,0.2985784411,0.0603098162,0.0972270295,-0.1055049673,-0.0377928801,-0.1585014462,-0.0520994365,0.4101315737,0.0977639407,0.0250490569,-0.2611637115,-0.0920041576,-0.5408343673,0.3158062696,0.0527636856,0.2028032094,-0.081078127,-0.0301086549,0.2543749213,0.0679179206,0.2663500309,-0.1048046798,0.1423966587,0.0410783403,-0.0973922536,-0.4929026067,0.0652067438,0.1237916499,-0.0717306361,0.4792842269,0.3520555198,-0.0394820161,-0.2428181022,0.1453234404,0.23052001,-0.243587032,0.0669356212,-0.1930082887,-0.5233623981,-0.1967772543,-0.1565243006,-0.1726042926,0.0737290382,-0.0955599621,0.0560566895,0.0631453618,-0.1865683347,0.1151105016,0.3434083462,0.0233111586,0.2734473348,-0.1275288016,0.0795082822,0.3538227379,0.1347263604,0.7317372561,-0.1766162068,-0.3419310749,0.142473489,0.3091111183,0.4259915352,0.551543653,0.0122105302,-0.0027419277,0.2469418049,0.169174403,-0.0710203499,0.0248558484,0.1896543056,-0.0271573644,-0.2438091934,-0.3387823403,0.6757808328,0.1753596961,-0.0686709732,-0.0196814202,0.1879802197,-0.1997472495,0.3174859285,0.196384415,1.112601757,-0.3865783811,0.2527160347,0.4754591286,0.1300272048,0.7861910462,-0.0476480052,0.0393128358,-0.175775826,-0.0386876054,-0.0158695821,0.1026389748,-0.2545431554,0.0931004807,0.1193133816,0.3765717745,0.1462594569,0.1244838163,0.17723234,0.4552799165,-0.0149794305,-0.1715274304,-0.9147063494,0.0899304599,-0.0443625525,0.1974340379,-0.1149675623,-0.1363938898,-0.1437603682,0.2531207204,-0.064076595,0.3095707595,0.0926276669,0.216854617,-0.3098124862,-0.0930678099,-0.1108041182,0.2111047804,0.4911542535,-0.1463082433,-0.2870134413,0.0100247413,-0.2619129717,-0.1846578121,-0.0040820618,0.1146591678,0.2568469346,0.0299446266,-0.0322138518,-0.0952626467,0.2902306318,-0.2916546464,-0.3434294164,-0.1409877241,-0.3479572833,-0.0593659244,0.1842235178,0.1704667509,0.0988735482,-0.1592891961,0.0438034944,0.0201071743,0.1397152543,-0.3530656993,0.1558365375,-0.2608713806,-0.0902478248,0.422477901,-0.1290141791,0.2367495596,0.2137132287,0.2999237478,0.041608043,-0.1443562359,-0.2040252537,-0.1250776201,-0.3493469954,0.2267754525,0.026972441,-0.2679871917,-0.1560677141,0.4593490362,0.2066574097,-0.3265994191,-0.1299116611,-0.4308762848,0.0066580661,0.3610094488,-0.0963620618,0.0311282147,0.328096807,0.0480952784,-0.0867451876,-0.1157348827,-0.2161809951,0.0298732258,-0.2376494855,-0.0058725709,0.3596525788,-0.1470711231,0.196218729,0.1221146062,-0.0142046446,-0.2798121274,0.0494470485,-0.0729348287,-0.1754334569,0.170799762,0.258508265,-0.2010315508,-0.1681411415,-0.2821453214,0.064847149,-0.288736403,0.0232700724,0.0604452528,0.0497078821,0.1475997567,-0.1680620313,-0.2170601338,-0.032462921,0.4655557871,0.0711520612,-0.0806486979,0.2655051649,0.1415583938,0.2142727077,-0.0884355903,-0.1847543418,0.047533188,0.0450642444,0.005697865,0.2424156815,-0.2242477536,-0.180459708,0.2852202654,0.0655076727,0.6470024586,0.2247956246,0.0137375575,-0.0031777136,0.0066327401,-0.1387801021,-0.0193507839,0.3327872157,0.0699873418,-0.2775742412,0.2848497033,0.3158890307,-0.2858683467,0.0928034857,-0.0361051485,0.2029639035,0.1325009912,0.2233767509,0.2039144933,0.2522875071,0.1599001437,0.0533145629,-0.0269811824,0.3574564457,-0.0929626599,-0.2084430903,0.2846877277,-0.4164149165,0.3620056212,0.1814380139,-0.0925890058,-0.1739553064,0.2266274542,-0.1083585769,-0.023672374,0.0493286811,0.8438782692,-0.3339235485,-0.2409933954,-0.3582416773,0.6556246281,-0.2454179227,-0.2910168171,-0.501608789,-0.1364713907,-0.179775849,0.2041600496,-0.3215113878,-0.2917788327,0.288403064,0.2090781331,-0.0305051748,-0.6123155355,-0.1765327156,0.3434591889,0.1124517098,-0.1138048545,0.3136467636,0.0580066144,-0.20412606,0.11886096,0.0968801975,0.3500439823,0.2010899931,-0.0146963773,-0.0477133207,0.1254771203,-0.0012670903,-0.2866496146,0.1265265197,-0.0653265864,0.2059822679,0.2354422808,0.0774263591,-0.0531416796,0.0560033172,-0.0051431498,0.1339714825,-0.239379555,0.2333390713,0.0041830079,-0.037076775,-0.3344923258,0.0534916259,-0.3265328705,0.1424583644,0.1018857211,-0.2222626209,0.0846439674,-0.1736570299,0.0138783967,0.1851200461,0.7972928286,0.3956439197,0.0028446128,-0.2536246479,-0.3362244964,-0.7480276823,-0.0303410869,-0.0455267169,0.0314102806,-0.0989700183,0.0264583584,0.0147554036,0.0864960402,-0.5139561892,0.3119202256,0.0454297662,-0.1438587606,-0.2488614917,-0.0460019484,0.0006246938,-0.1093419716,-0.0308333486,-0.1851891428,0.1970299631,-0.0995515063,-0.1593070328,0.0405900516,0.1372340322,-0.1498680115,-0.1960590482,0.3116006553,0.2188539952,0.2902201712,0.0941513926,-0.153239727,-0.3999745548,-0.3762974441,-0.2154217362,-0.0351465382,-0.1620055437,-0.0105525842,-0.0265235361,0.3910281956,-0.269816637,0.1867967993,0.1291775107,0.1626279354,-0.2214720845,-0.1415459961,-0.282215625,0.0769395605,0.1828295141,0.3219568133,-0.1019818559,-0.1648567468,-0.0388853252,-0.2984727025,0.3104025126,-0.667491734,-0.0051825256,0.2318101823,0.1003702134,0.19054313,-0.0044605597,-0.280321151,0.0216780696,-0.040231023,0.0536148027,-0.0572229102,0.2559002638,0.2245052308,-0.2706518173,-0.2282407582,-0.2176375389,0.2312410623,-0.3037879765,0.1339839399,-0.3156906068]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/353","title":"[Dataset requests] New datasets for Text Classification","comments":"Pinging @mariamabarham as well","body":"We are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705","comment_length":4,"text":"[Dataset requests] New datasets for Text Classification\nWe are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705\nPinging @mariamabarham as well","embeddings":[0.0115056699,0.1141199917,-0.1910357177,0.1999402046,0.2285178006,0.2350338399,0.2402176559,0.1117574349,-0.1602311581,-0.0280329138,-0.1551205218,0.0445593297,-0.1305259615,0.1583450437,0.069891341,-0.2651985884,0.0052187382,-0.0229230169,-0.1080695614,-0.2169982642,-0.1513809115,0.2123754621,-0.0966075435,-0.2254793793,-0.3578056097,-0.0854782015,-0.1268238574,0.1236310005,-0.2721914947,-0.4812310338,0.047330685,0.1087944806,0.1437276006,0.3386830389,-0.000097488,-0.2481977195,0.0241904911,-0.0448231697,-0.2087298632,-0.2027125508,-0.1579287797,-0.390339613,-0.0785961598,-0.1112326756,-0.140840441,0.0393986814,0.0493505448,-0.2036977857,0.0069073024,0.3420635164,0.3144716918,-0.0620939732,-0.0480513535,-0.2143129855,0.2946068048,0.4708190262,-0.3388802707,-0.0756582618,0.5982760787,0.1412106454,-0.063729234,0.3361603916,-0.2350931317,-0.0875835791,0.1505088657,0.1412210315,-0.1960145682,-0.3765784502,0.0965151638,0.3389234543,0.7362016439,-0.2900658846,-0.3954140842,-0.2450648993,0.1491811872,-0.2527163327,-0.1407944262,0.2473171204,0.0088601224,0.1370868683,-0.1545364559,-0.1961169839,-0.2148229927,0.2066241205,0.2771199644,0.095560655,-0.0320370533,-0.2588176131,-0.1127177849,0.0984570608,0.2913126647,-0.3016297817,0.1847318113,-0.0358893387,-0.1780061871,-0.1977525651,0.0115345214,-0.029011637,-0.033587724,0.0829198286,-0.1426823586,0.1380389184,-0.1912590414,-0.0828061625,-0.0449430682,-0.1345386952,0.190645963,0.2754262984,0.0573465228,-0.1730305701,0.1004187316,0.058006946,-0.1120452434,0.1842932254,-0.060857214,0.0282844417,-0.1517892629,-0.2317615747,-0.0123852314,-0.0423625559,0.022614995,-0.0816590488,-0.1827678829,0.0554809868,0.0146047585,0.1039846838,-0.1844376624,0.2539661527,-0.0284761488,-0.4196957052,-0.1564173549,-0.0676477849,-0.1178204045,0.1097512618,0.5927137136,-0.1116399169,0.1885489523,-0.1598035991,-0.146594733,0.1446469277,-0.0544628128,-0.0461895913,0.0836385116,0.1596108824,-0.1112475768,0.0814120695,0.1146212593,0.0983470604,-0.2123074234,0.2376877964,-0.0195777062,-0.2408113182,-0.2047690749,0.3242529035,0.0424854383,-0.5377767086,-0.0951574221,0.5570358634,-0.0501327924,-0.0863951445,0.2146683633,-0.0019230194,-0.086511679,-0.1443656534,-0.0135624427,0.3017602861,-0.4125982821,-0.1894026548,-0.3059391975,0.0992337316,0.1612060666,0.176317215,-0.1617304534,0.0431238599,0.0914160237,0.0577124022,0.2252629548,-0.019579269,-0.1858144552,-0.1420925856,0.0132617531,-0.1618792266,0.1215679646,0.2680240273,0.181807071,-0.0753816292,-0.0608308502,0.1424499899,-0.2852880955,-0.0225921292,-0.1568504423,-0.0732284635,0.1066582501,0.4050813317,0.4763915837,-0.1783029586,0.0946775675,-0.2082434744,0.0946054459,-0.0293262322,0.1915779561,0.1453698426,0.1471173614,0.1864406615,0.0770133063,-0.3266232014,-0.158846572,-0.0108559588,0.2721295655,0.3499366045,0.0267435554,-0.1459004581,-0.1543150991,-0.1198792607,0.0542729385,-0.0506836884,0.2275629193,-0.1495402008,0.0822818726,0.15164195,-0.2803983688,0.1808022261,-0.3585906327,0.3558246791,0.1900745481,0.2324546874,-0.0960913748,-0.0348112807,0.1921270043,0.3428022861,-0.0890412331,0.0091743069,0.0324893072,-0.0255943052,-0.1477643698,-0.0346031748,0.3598506153,0.351739049,0.332051307,-0.4617651105,0.2306239605,-0.0573246703,0.0656241477,-0.1321124434,-0.1142356321,0.3267025352,0.0096176425,0.0831934139,0.058290083,0.0101706032,0.0313378796,-0.0517116114,0.0640912056,0.1205002815,0.3407564461,-0.165516451,0.1347858608,-0.0252278037,-0.1775493771,0.2269497365,0.1990721822,-0.0687390715,0.2423356175,0.2289719731,-0.4121674001,-0.0672468692,-0.0072309356,0.050802093,0.2556385994,0.5346294641,0.0067240591,-0.1331732124,0.0687307715,-0.4023368955,0.0487922058,0.3493286073,0.0094599789,0.056209974,0.3007188737,0.2081483603,-0.2643583715,-0.1432059556,0.1386125237,-0.0828647614,0.018983854,-0.0945418477,-0.040345788,-0.3425836265,0.056237638,-0.0095325699,-0.0856492519,-0.2653041184,0.26863873,-0.180892542,-0.00406481,0.1040750891,-0.3026538789,0.5136149526,-0.2396945953,0.0453658029,-0.0892228112,-0.1974702477,-0.2336581796,0.3242678046,0.0556746535,-0.0336402766,0.4914983809,-0.126964584,0.3003948331,-0.2555092871,-0.4959823191,0.0194611009,-0.1580129713,-0.0883357227,-0.1623128951,-0.1557155252,0.0105015459,-0.1467826068,-0.0181859173,-0.1523259431,-0.1381135136,-0.0365689248,-0.1617233008,-0.2195718586,-0.083289586,-0.643140018,-0.3587786257,-0.1055326611,0.1074944213,0.0078625204,0.0076195677,0.2462770045,-0.1540872753,-0.0381553099,-0.1915987581,0.1405401528,-0.0922925845,-0.2436422408,0.1145701483,-0.1726814359,-0.2257030606,-0.0952559263,-0.0554934479,-0.0416854285,-0.0782246515,-0.4911924601,0.0904504508,-0.2233254164,0.1647497863,0.1317361891,-0.0821944997,0.439866066,0.292396158,-0.2351806909,0.0444362648,-0.0804076344,-0.0441675857,0.021742424,0.2323170006,-0.344332546,0.0380357839,0.2400835603,0.5381413698,0.0242296886,-0.0433549769,0.1139984056,0.2319396436,0.3206492662,-0.0414169058,-0.3844069839,0.2464596331,-0.1686198562,0.1346433163,0.3499066532,0.2373391986,-0.3531303108,-0.1954343617,0.0454356484,-0.3195349574,-0.1315580308,0.2513367236,-0.0587046035,0.2812214494,0.0578674898,-0.1232305989,-0.2127778679,-0.4395703971,-0.0973218307,0.2314856052,0.1575784683,-0.2253839225,-0.5546702743,0.1771295369,-0.4631044865,0.1621356755,0.1762663573,-0.0997435153,-0.2095448673,0.133727327,-0.1117511541,0.4036679864,0.0565158837,-0.4062527418,-0.3062141836,0.2297712713,0.0554065928,-0.0879548043,0.1636382639,-0.4536347985,-0.0348044373,0.2541902065,0.0007181012,0.0950239152,-0.0031525511,0.4790267944,-0.0106948512,0.0339715965,-0.2741353214,0.3109200895,0.0346146598,-0.236336872,-0.1013030335,0.0518303886,0.086812973,-0.178248316,-0.0499405377,0.1828619689,-0.0146579947,0.3595075607,-0.268546313,-0.2930607498,0.1969796866,0.1851142496,0.0568559989,-0.039029751,0.3249427378,0.4713756442,-0.1882206798,-0.4825723171,-0.0129677691,-0.1407786012,0.4667708576,0.2821000218,0.3350655138,0.2231024057,-0.1471354663,-0.109519504,-0.1143772379,0.183997795,0.0937585682,-0.058834359,-0.5086569786,-0.27512604,0.3035342395,-0.0956605077,-0.0691357404,-0.0126173319,0.2311586291,-0.1887378395,0.1523099989,-0.0883503929,0.9418475628,-0.0399520807,0.1525636762,-0.3047965169,-0.0438409708,0.4606690109,-0.2707617581,0.2201190889,-0.0330193415,-0.2689504325,-0.0397800989,-0.022806624,-0.027538849,0.375759393,0.0985619128,0.2493750453,0.200110957,0.1490142196,0.1231564581,0.1373014152,0.065306589,-0.4220643044,0.0640243962,0.2993256748,0.0503371544,0.2603137791,0.0221134443,0.0624105558,-0.19458583,0.1232325733,-0.6553910971,-0.1473780721,-0.1441681087,0.0582578219,-0.0580828264,-0.3651029766,0.4303863645,0.2534003258,0.4850494564,0.1122877002,-0.2133798301,0.0610214435,-0.2218079567,-0.1399634331,0.1646072268,-0.1364346296,0.2821217775,-0.0208878424,-0.1830467731,0.1853070259,0.0450453237,-0.1831037402,-0.4328872561,-0.3138218522,0.2152281553,-0.3266335726,0.3012922704,0.1818543077,-0.0090872068,-0.3207397759,0.2589060962,0.2243557572,0.302136451,0.1158616245,0.1909088939,-0.0692899451,-0.4000338912,0.1969360262,0.217586726,-0.2736433148,0.36149773,0.4455492496,-0.1760672778,-0.316273272,0.1811475009,0.2185919136,-0.2710334063,-0.0129803428,-0.1331326067,-0.0545309857,0.1289009601,0.345297724,0.3843571544,-0.1990380883,0.038181413,-0.1764728576,-0.0658639669,0.2015692294,-0.1117138788,0.0468328856,-0.205348894,0.1325788498,0.0727350637,0.2005597949,-0.4465903938,0.1302275509,0.0864095688,0.0344077684,0.2051211149,-0.0507551692,0.1437214166,-0.2319756448,0.120965533,0.0016399496,-0.0082401764,-0.3414768577,0.0258509908,0.0957223102,0.0111133121,0.0332378186,-0.0100884484,0.0921355337,-0.115540497,-0.0787198171,-0.1026405692,0.114878051,0.0812097937,0.4215379357,-0.199158445,-0.0188025143,0.0966107696,0.0687433258,-0.3199037313,0.3952269256,0.2568592429,0.1080314219,0.389626801,0.0841682926,-0.1330156922,0.1774923205,0.1712321639,-0.1225181818,0.5723749399,-0.1631145924,0.2380948514,0.4811245501,0.1205977499,-0.2279640436,-0.1590022147,0.0786492452,0.0186565313,0.35968104,-0.3527772427,-0.1727444828,-0.0533398464,0.196343407,0.0938071236,0.217200771,0.2536992431,0.0418162905,0.2540469468,0.1036455706,0.4516834617,-0.1103591323,0.3339632154,-0.0832420737,0.1814350933,-0.0663422868,0.2517234385,0.0960171595,0.0277651679,0.2480907589,0.0606913529,0.4861084819,-0.1226680204,0.1759922504,0.3457559943,-0.1636079997,0.3771340251,0.2297273129,-0.0509718768,0.160755232,-0.0374336913,0.2271354645,-0.2059803605,0.0352880135,-0.1516370624,-0.0312363915,-0.0775090754,0.0893533155,0.0470939763,-0.3408595324,0.0189700127,-0.1612757295,-0.2855446041,-0.2837046981,0.2115943283,-0.082435891,-0.2054938376,-0.58830446,0.1507870555,0.1198262349,0.4034147263,-0.1578462869,-0.0454424769,0.1752370596,-0.2147419006,0.2348469645,0.5224359632,0.1853471398,-0.2136633396,-0.0325220414,0.0592227727,-0.27726987,-0.2126296163,0.1420697272,0.2576033175,0.3825170398,0.1176289469,0.0105061578,0.2426230013,-0.1270331889,0.1129163653,0.1579982638,-0.0155920591,-0.2176746577,0.1071733013,0.1432962567,-0.2351398021,-0.1857891679,-0.0307006948,-0.2580518723,-0.1278777868,0.4986974001,-0.1330064982,0.0473270863,-0.0925036669,0.1640952528,-0.0917944089,0.2472555339,0.3494556248,-0.1745448261,-0.180388853,-0.1445050836,-0.0239237901,-0.1650690883,0.1160951853,0.2155016959,-0.1991193295,0.0153644998,0.2713679075,0.2230369747,-0.2817997336,-0.1233010367,-0.084184736,0.0300512332,-0.5286594629,-0.0608589873,0.2197885215,0.2063670903,-0.0001244494,0.1220524758,-0.0686254352,-0.0748889297,0.1018296629,-0.3818541169,0.1581818461,0.2227345556,0.1121602133,0.1397106946,0.0595166497,0.3723462224,-0.0595563538,-0.1610460877,-0.0471332446,-0.3470813632,-0.0396227315,0.3635964394,0.0858837515,0.3046097457,-0.2339481562,-0.2928065658,0.3136937022,0.09695746,0.0451126546,-0.0744308308,-0.5242666006,0.0282731019,0.0390702263,-0.0929825231,0.1469487846,0.0000834718,0.0435257964,-0.1514509916,-0.1311495155,-0.4674848616,0.0763723701,-0.0579646789,0.0765870661,-0.2586548328,0.1542360336,0.3562430143,0.2648355067,-0.6497910023,0.025612371,0.4844953418,-0.1495489478,-0.0165976919,0.4185627699,0.1375549585,-0.0623944737,-0.1561486721,0.324105978,-0.2164700627,-0.2630463541,-0.1732639372,-0.2265219986]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/353","title":"[Dataset requests] New datasets for Text Classification","comments":"- `nlp` has MR! It's called `rotten_tomatoes`\r\n- SST is part of GLUE, or is that just SST-2?\r\n- `nlp` also has `ag_news`, a popular news classification dataset\r\n\r\nI'd also like to see:\r\n- the Yahoo Answers topic classification dataset\r\n- the Kaggle Fake News classification dataset","body":"We are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705","comment_length":47,"text":"[Dataset requests] New datasets for Text Classification\nWe are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705\n- `nlp` has MR! It's called `rotten_tomatoes`\r\n- SST is part of GLUE, or is that just SST-2?\r\n- `nlp` also has `ag_news`, a popular news classification dataset\r\n\r\nI'd also like to see:\r\n- the Yahoo Answers topic classification dataset\r\n- the Kaggle Fake News classification dataset","embeddings":[0.0527632833,0.1770247221,-0.1492281705,0.2923184335,0.1770986617,0.2543239892,0.2225594372,0.1188267395,-0.14430812,-0.1214152873,-0.1536755115,0.0564632602,-0.1662919223,0.1704207063,0.0837513357,-0.2406843454,0.051472567,-0.0587837324,-0.0951605663,-0.2292533368,-0.1327498257,0.2470685542,-0.1144927219,-0.2490836084,-0.3917073607,-0.0300808158,-0.1971095949,0.0825114176,-0.233087644,-0.5466544032,0.1298060268,0.1312667429,0.2001403868,0.3561389148,-0.0001037212,-0.2407378256,0.0268519837,-0.0144627271,-0.1714262664,-0.165544644,-0.2061451823,-0.5361247659,-0.0227512456,-0.0801116824,-0.1275690049,0.0410604812,0.0543383174,-0.2447412461,-0.0130655728,0.3233343363,0.2447050363,-0.0661868975,-0.0584240593,-0.1412222981,0.2988788188,0.551233232,-0.31063205,-0.0630631968,0.532618165,0.1365834326,-0.0013358365,0.3090865016,-0.2009138614,-0.0890441313,0.160744831,0.2195890993,-0.2908791602,-0.38539204,0.0384694748,0.3302031755,0.7269296646,-0.2991444767,-0.435644418,-0.2929211855,0.1790418923,-0.2091692686,-0.1762985885,0.2850813866,-0.0127935493,0.1451199055,-0.1474387795,-0.2335708737,-0.2125960141,0.2384119332,0.2997066081,0.0945870131,0.0134552671,-0.1983968467,-0.0950295106,0.128020227,0.3503784537,-0.380717665,0.2048949003,-0.0010360541,-0.1995961964,-0.2089159638,-0.0466030464,0.0258889645,-0.0650751293,0.0100435726,-0.2320323586,0.0853959918,-0.1917665154,-0.08208514,0.0305850226,-0.1031994075,0.2131817639,0.3429120779,0.1036159694,-0.2787934244,0.1565501988,0.0335344262,-0.1759021431,0.2272807062,-0.0201550461,0.0545121059,-0.1334668249,-0.2294958681,0.0464145094,-0.1172540039,0.0202058312,-0.1066801175,-0.1993536204,-0.0666343272,-0.0499973409,0.1499421448,-0.1912824959,0.2406200469,0.0035727895,-0.4072596729,-0.1258370876,-0.0520173386,-0.1064224318,0.1718494594,0.6023263931,-0.1670332551,0.1791901886,-0.1379181594,-0.1681382656,0.1534096748,-0.076660715,0.0061889212,0.0636818781,0.1282594204,-0.1471486986,0.1247724742,0.1550988555,0.1334684193,-0.2032557428,0.2613936961,-0.0514775105,-0.2116403878,-0.2426079661,0.2461383641,-0.0127631761,-0.5387881994,-0.1642753035,0.6117169857,-0.034047205,-0.1151471287,0.1698567867,-0.0192075875,-0.0849988833,-0.1658167541,-0.0492004789,0.4430484474,-0.4810695648,-0.2499638051,-0.3836504817,0.1358901262,0.1784741431,0.1647046208,-0.1955979913,0.0349018574,0.0822529271,0.0167149641,0.2063207477,-0.0245061144,-0.2242669463,-0.091634959,0.0819611922,-0.1140097901,0.1038505808,0.3170546293,0.1133601367,-0.0756833702,-0.1596257389,0.1264414042,-0.2982211113,-0.0488292575,-0.1245086491,-0.0446394011,0.2099894434,0.4207590222,0.4951865375,-0.0852148756,0.0735037327,-0.2081886679,0.0925953314,0.0692988038,0.2265700251,0.0930776745,0.1255875081,0.2216483355,0.0558079332,-0.3216334283,-0.2000431269,0.0503138192,0.3117822111,0.345841825,0.0033857967,-0.1304207295,-0.1460157186,-0.1206993684,0.0435250513,-0.0300355572,0.1596376151,-0.1221902966,0.118841514,0.1875600666,-0.300601542,0.2069494873,-0.3214774132,0.3590740263,0.1875853091,0.3283416033,-0.0352597944,-0.0059967902,0.1668463796,0.3152220845,-0.031121999,0.0017882282,0.011521183,-0.0421964489,-0.1521205753,-0.03780457,0.3529798388,0.3763906658,0.3767844141,-0.52286762,0.2861688733,-0.0594802573,0.0689433068,-0.1887095422,-0.0736831203,0.3114968836,-0.0839342028,0.0817986578,-0.0040461672,0.0514792502,-0.0256481674,-0.0040906458,0.0368916728,0.1111270636,0.2754785419,-0.1613609493,0.1710827351,0.0034509918,-0.1263172925,0.2079485804,0.2214201242,-0.0540879555,0.1790332198,0.2693894804,-0.4363262653,-0.0446643047,-0.0360016599,0.1678316742,0.3374681473,0.4586303532,-0.0075571775,-0.0921161547,0.1000887826,-0.3318985701,0.0539445169,0.3946705163,0.0238215551,0.0016801551,0.2778208256,0.243368268,-0.2611657083,-0.1199082807,0.1832582355,-0.0790406242,0.0415701494,-0.0527037233,-0.0734100342,-0.3831567168,-0.0118776262,-0.0608394183,-0.091112569,-0.2869122624,0.2335093468,-0.28126508,0.1054873466,0.0937801749,-0.3011012077,0.474956423,-0.2583185732,0.0935938582,-0.0809932426,-0.2227909565,-0.2262812704,0.264413029,0.0643127933,-0.0026947286,0.4211891592,-0.0927983522,0.3203366995,-0.2678386569,-0.4691827595,-0.0161218587,-0.2101109028,-0.0101745343,-0.1509811729,-0.1415689439,0.014812394,-0.1314619929,0.0011892322,-0.1384777576,-0.1376017928,-0.0223242939,-0.2006468624,-0.1917314082,-0.0129956193,-0.6513702869,-0.5205703378,-0.0659397766,0.0964546576,0.0098314267,-0.0208403338,0.3068898916,-0.1584125012,-0.0948942974,-0.1482901573,0.0830030441,-0.0726875439,-0.3189515769,0.1442463994,-0.1456786394,-0.1967961937,-0.0786523148,-0.041298423,-0.1070435047,-0.0746239796,-0.5309435725,0.1441844702,-0.268520236,0.181995362,0.1227803752,-0.1149686724,0.4368275702,0.2920848727,-0.120861344,0.0626671314,-0.1701074243,-0.1198526695,0.0555914938,0.1654337794,-0.2778554261,0.1061424837,0.2385752797,0.5572987199,0.0877767429,-0.1116384119,0.0903944299,0.2851227522,0.3834935725,0.0125235869,-0.4623270333,0.2260669619,-0.2044538409,0.1779331267,0.2970435917,0.2645272017,-0.3142875731,-0.2004949152,0.1137503684,-0.2948045135,-0.0459498726,0.1783654541,-0.1388091296,0.2812418342,0.057397794,-0.1314416379,-0.2332429439,-0.4535067379,-0.1118651927,0.3031828105,0.2348776758,-0.1552293748,-0.646165669,0.0382987186,-0.4787890911,0.089390941,0.1747152656,-0.0781371668,-0.1947266906,0.2404093742,-0.1146248206,0.4444725513,0.0193658788,-0.4512856007,-0.3338707685,0.2409869879,0.0585489869,-0.0943010449,0.1333487779,-0.3841445148,-0.0655911118,0.2734688818,-0.0731278434,0.0967717841,-0.012211645,0.4771545231,-0.0117730144,-0.000174585,-0.3162128031,0.3753697574,0.1242550388,-0.1765608937,-0.2105774432,0.0750048682,0.0364503711,-0.237223208,-0.0196513496,0.1807062179,0.0154314572,0.3284829855,-0.2906268835,-0.2960877717,0.2343979031,0.2130177915,0.0700454265,-0.0419205576,0.28078264,0.5084705949,-0.2091922462,-0.5425741076,-0.0160921942,-0.1136777326,0.5346791148,0.2124382257,0.2784397006,0.2342030853,-0.1407708675,-0.1889597178,-0.1410025507,0.2169254571,0.1151650175,-0.0961548463,-0.4878028929,-0.3194537163,0.3196599185,-0.0789463446,-0.1006030291,-0.0245126598,0.2379218042,-0.1844978034,0.1345093995,-0.1880151778,0.8308113217,-0.0646939948,0.122452572,-0.3462646902,0.0034010382,0.5688192844,-0.250418663,0.1834023893,-0.001331883,-0.2832993865,-0.0686313808,-0.0549051575,0.009967966,0.4351372421,0.0973288789,0.2558712363,0.1897925884,0.2015055418,0.0624811538,0.1596852988,0.0927691609,-0.3814063668,0.0425379053,0.2434984297,0.0195031427,0.3009161949,-0.0174933244,0.088347517,-0.2239394039,0.1797337681,-0.7378605008,-0.1611166,-0.1183720157,0.0995341092,-0.0112907514,-0.3718676269,0.4561332762,0.3514236808,0.568269372,0.0649139732,-0.2017302662,0.0309522506,-0.2211615592,-0.1586835831,0.15189448,-0.1231954396,0.2417224795,0.0153145343,-0.2222799957,0.1778397411,0.0042779781,-0.2087237388,-0.4498830438,-0.3305296004,0.2800061405,-0.3033126891,0.3471176028,0.1718563884,0.0626682714,-0.3116731942,0.1967435628,0.2942277491,0.3189515173,0.1205535382,0.2112025917,-0.0315547138,-0.4240737259,0.1737750918,0.2421079725,-0.2807995677,0.4127334356,0.4719037712,-0.1487902701,-0.2214155644,0.2455909103,0.1872136295,-0.2705706954,-0.074196361,-0.1488347501,-0.03312869,0.1928812563,0.304118067,0.3911813498,-0.2794886231,0.018892685,-0.2333868891,-0.0565885305,0.1763997823,-0.137332499,0.0951542035,-0.2373582572,0.1373422146,0.0938088223,0.2622424662,-0.3498083353,0.1299263388,0.1491361707,0.0551438183,0.2374936044,0.0251934379,0.1273240596,-0.2558550239,0.0349806249,0.0103954561,0.039561525,-0.2490551472,0.0759653822,0.131013751,0.0141825164,0.0220444333,-0.0599603876,0.1174069643,-0.1217669249,-0.0285258088,-0.1022374332,0.1260844916,0.057990279,0.5187693834,-0.1874863654,-0.039988663,0.1310334802,0.0602216609,-0.278336674,0.4006113708,0.3284617066,0.1202046275,0.4760238528,0.0949270278,-0.2307254076,0.1763989478,0.0927723348,-0.087839812,0.6063158512,-0.2409617454,0.2653624117,0.4911391735,0.0821076408,-0.260422647,-0.0842221901,0.1215368137,-0.0134149762,0.2712398469,-0.3505735993,-0.204254806,0.0193941221,0.2798811495,0.0451724529,0.1955763549,0.2910692692,0.0367107913,0.2092922777,0.1125170365,0.4954883158,-0.0687464848,0.3932679892,-0.0535323061,0.1863026917,-0.0916723981,0.2409771979,0.1088311821,-0.0160726663,0.2082688212,0.0718134344,0.4569135308,-0.1139340624,0.1205675527,0.3959608078,-0.1470081657,0.3860013783,0.2909430265,-0.1010904312,0.1179048121,-0.0907841325,0.2581425905,-0.2204218805,0.0017508995,-0.155417785,-0.077387251,-0.050407622,0.085224539,0.1088257581,-0.3221072555,0.0587657206,-0.1575824171,-0.2597844601,-0.3441855013,0.239599064,-0.1194178835,-0.2106220126,-0.6339172721,0.1954016834,0.1309528202,0.3912188113,-0.1357330978,-0.0747176185,0.0918052867,-0.2612673938,0.2775428593,0.5547748804,0.1446849257,-0.1552878022,0.0057850783,0.0479366854,-0.2726131678,-0.2029710859,0.1911283731,0.2833507955,0.3340112865,0.1013445482,-0.0617237389,0.1865949333,-0.0430021696,0.0830180645,0.088481456,0.0046229283,-0.3006363511,0.1651795357,0.0795321614,-0.2540403903,-0.1077440977,-0.0441864692,-0.2126710564,-0.202478081,0.5217410922,-0.1249362826,0.0119945966,-0.1216107085,0.1276407391,-0.1510784179,0.1781790406,0.3320963979,-0.2001376599,-0.2637599111,-0.1479518712,-0.0320388749,-0.1874038279,0.1445921808,0.1973628551,-0.2097578347,0.0329247341,0.2578451335,0.246088177,-0.3451769948,-0.0985843465,-0.0981000811,-0.023528222,-0.5865010619,-0.1341862828,0.154818669,0.2323828638,-0.0127064371,0.1974993795,-0.0983118117,-0.017211942,0.0223830603,-0.3528050482,0.1500775218,0.2342923284,0.1128525883,0.1448772699,0.1082502976,0.4487566948,0.0312771164,-0.1258467287,-0.0953545719,-0.4393796623,-0.0311774537,0.3536724746,-0.0399169028,0.345067203,-0.3038221002,-0.2439208925,0.3059532344,0.0560059398,0.0972972065,-0.1216288358,-0.52627033,0.0174826663,-0.015953308,-0.0718268752,0.2164773643,-0.0236603282,0.0369966999,-0.1227627844,-0.0397366546,-0.3579162955,0.0555739962,-0.0923437178,0.1185166463,-0.2445089519,0.062012326,0.3544395566,0.2664332986,-0.624558568,-0.0098280208,0.5481746793,-0.1670346558,0.031550575,0.4038210213,0.1149759293,-0.0515665486,-0.0985536128,0.3903352916,-0.2673252523,-0.2338880301,-0.1389212757,-0.1567140073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/353","title":"[Dataset requests] New datasets for Text Classification","comments":"Thanks @jxmorris12 for pointing this out. \r\n\r\nIn glue we only have SST-2 maybe we can add separately SST-1.\r\n","body":"We are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705","comment_length":18,"text":"[Dataset requests] New datasets for Text Classification\nWe are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705\nThanks @jxmorris12 for pointing this out. \r\n\r\nIn glue we only have SST-2 maybe we can add separately SST-1.\r\n","embeddings":[0.0815415308,0.0239060819,-0.1260785609,0.2537373304,0.2022433132,0.1909445673,0.2313451171,0.1288252771,-0.0196939688,-0.0524679348,-0.2371735871,0.0281864014,-0.0793611482,0.2284713984,0.1158258021,-0.3347227573,0.0327517204,-0.0278218016,-0.2127438188,-0.1855586022,-0.1148792878,0.3360117972,-0.0453988202,-0.2311266661,-0.4425388873,-0.0575625375,-0.2063391656,0.0668693855,-0.2878912389,-0.4556583166,0.0428925753,0.1286984086,0.2520892322,0.2492392361,-0.0001037809,-0.2161204368,0.0535999425,-0.091308713,-0.0742770955,-0.0727216601,-0.1487201601,-0.4849986732,-0.0241378415,-0.1006161571,-0.2236712724,0.2096851468,0.0609343462,-0.2331152707,0.0375484861,0.2961035371,0.2434536368,-0.0123440856,-0.0020559046,-0.1590737998,0.4260741174,0.3211229444,-0.3257283568,0.0370839611,0.6237726212,0.0663693473,-0.0005364151,0.3617648482,-0.2438280433,-0.0247037914,0.1756167412,0.2001561224,-0.2193982452,-0.4464477003,0.0715357512,0.3973942101,0.6543508768,-0.2877071798,-0.3910399973,-0.3127829134,0.1366227865,-0.0791728422,-0.0994484797,0.1954574585,0.0398405604,0.2265112102,-0.0492472313,-0.245719716,-0.1470465362,0.2701330185,0.3945715725,0.0568461344,-0.0370777883,-0.1775099933,-0.0003612236,0.179885,0.1766577512,-0.2836074233,0.1341858357,-0.0177958272,-0.2663389742,-0.177357927,-0.0541212335,-0.0193151273,-0.08409632,0.0844806358,-0.245130375,0.103039436,-0.0586198866,-0.0488160476,-0.017619174,-0.0382055715,0.2564458847,0.3307938576,0.0510782972,-0.3528569639,0.1201025695,0.0050097569,-0.1631537974,0.1350767314,-0.0659490302,0.0418742001,-0.2124118209,-0.1809259951,0.0381796286,-0.1414325237,-0.0032424561,-0.055586163,-0.2538170218,-0.0165537633,0.0661199987,0.1161912978,-0.1412743628,0.2230422795,-0.0193490442,-0.4236039817,-0.1252016127,0.0979864895,-0.1602702737,0.152887091,0.548071146,-0.2299818397,0.1155242324,-0.2185691297,-0.2966979146,0.1594123244,-0.1556195617,-0.0610738061,0.0852686614,0.2059712857,-0.1045297235,0.1281835884,0.1077968329,0.0212762039,-0.2046411932,0.1557881236,-0.038562499,-0.3110814989,-0.3726719618,0.2138151973,-0.1005914137,-0.4864225388,-0.1874146163,0.5722736716,-0.0236890614,-0.0310174394,0.2205074131,-0.0103252493,-0.1518809348,-0.1902374476,0.0116472179,0.4308137894,-0.4707907736,-0.2307440192,-0.3453292251,0.0235168077,0.0336860716,0.241004467,-0.0922274441,0.0245775469,0.0140255112,0.1602306068,0.2162276506,-0.068709366,-0.2146432847,-0.075349085,0.0846202299,-0.1966258287,0.1634407341,0.3142953217,0.2429367155,0.0132585289,-0.1556187272,0.1721234322,-0.2283507884,-0.0336175412,-0.1966410279,-0.1364067346,0.1465033889,0.3985637128,0.4805242121,-0.1305693984,0.068939738,-0.3062627017,-0.0355404355,0.1058141366,0.2744992673,0.1641733348,0.2053435147,0.2714706659,0.1024119854,-0.4187617302,-0.2145657688,0.0550114289,0.3147020042,0.4139363766,0.0726572648,-0.1219885647,-0.1317267716,-0.1913783103,0.0834059268,-0.0476080626,0.137248978,-0.1788128614,0.1703273058,0.1093635783,-0.2658056021,0.2994686365,-0.2152725458,0.4264630675,0.2125779539,0.3759650588,-0.0918564424,-0.1134921685,0.1790644974,0.3466808498,-0.0683494136,-0.0561639331,0.0313503332,0.140604347,-0.213462621,-0.0584961362,0.2669733465,0.3660677373,0.3261840641,-0.513733685,0.1703124791,-0.0794895589,0.0716542527,-0.1760181785,0.0449972637,0.189144358,0.0367133468,0.1446832567,0.005856161,0.1148122996,-0.1064943671,-0.0820765421,-0.0027086877,0.1102708802,0.304279685,-0.2344645709,0.3341112733,0.0095608598,-0.0775271282,0.2798493505,0.1818888038,-0.0192011576,0.2518855631,0.2147261202,-0.5472897291,-0.0057125818,0.0321520157,0.2584816217,0.3475606441,0.5090733171,-0.0490460396,-0.147824809,0.0817446262,-0.3104186952,0.0761174336,0.3464794755,-0.0318370871,-0.0022071903,0.2686341405,0.1926966906,-0.2491729409,0.0267807748,0.0885097533,-0.0470763668,-0.0210079327,-0.1603132933,-0.0493501872,-0.4877552688,0.012579483,0.0661214441,-0.1052722484,-0.2437970638,0.2872220278,-0.3143340945,0.0283844899,0.0965959057,-0.1657503843,0.4271807373,-0.2522673309,0.0750089884,-0.0557899326,-0.1706417203,-0.2362039387,0.2360295057,0.0335204862,0.0894505903,0.4961579144,-0.0915073827,0.3037363291,-0.2775154114,-0.5079554319,0.0609963723,-0.2328941077,-0.0041528451,-0.0171501767,-0.1594959348,0.1356656998,-0.2140608281,0.0266100951,-0.2283368558,-0.0827754438,-0.0041486304,-0.2452598065,-0.2524430156,-0.1303008348,-0.657120347,-0.5468477607,-0.1074697003,0.1612108797,-0.0874642953,-0.0005496375,0.1316719502,-0.1874059439,0.0237714704,-0.28291291,0.0646608174,-0.0692201927,-0.2697477043,0.1288060993,-0.1303965449,-0.2666116357,-0.0889950544,-0.0791966319,-0.1633375883,-0.1946291476,-0.456413269,-0.0327097625,-0.3555341065,0.1456098557,0.132913515,0.0244399905,0.4918686152,0.2438689172,-0.1384389251,0.0291558672,-0.2249152958,0.0785069689,0.0840614885,0.2012487799,-0.2755500376,0.1582756042,0.2967652678,0.4924660623,0.0400808565,-0.101574935,0.1274571568,0.1824348122,0.3775154054,-0.0146982595,-0.4498075545,0.2699258626,-0.2119074613,0.2726047635,0.3029667735,0.3241715729,-0.2568630278,-0.205516085,0.1334165931,-0.2536545694,-0.1615064889,0.1577555537,-0.0627446175,0.2349365801,0.0184645597,-0.2019221932,-0.2235529274,-0.440880537,-0.0064141629,0.2707579136,0.209339723,-0.0464640036,-0.6493407488,0.043453943,-0.4677713811,0.1304818243,0.2208282202,-0.1084982082,-0.1914462894,0.1372621506,-0.0909463391,0.4632479548,0.1011894196,-0.5415429473,-0.3169794679,0.2783178091,0.0771071091,-0.205855757,0.0643496141,-0.431116581,-0.131876722,0.4279152453,0.0110961488,-0.0116736693,-0.0630823001,0.5058277249,-0.0420790575,0.0455502681,-0.3159418702,0.3045090437,0.1805040687,-0.2231691182,-0.1821767241,0.1421315819,0.0341132767,-0.0992967263,-0.0157810152,0.2213718891,-0.1403499693,0.3071540594,-0.3063746095,-0.2161121368,0.1517287642,0.2439229488,0.1034974903,0.0296992362,0.0617583543,0.5007281899,-0.285571903,-0.3766000569,0.0341982171,-0.1609439701,0.5179302692,0.1911662072,0.2609319091,0.2677951455,-0.2097599953,-0.1387262642,-0.090520516,0.2293335199,0.0915510952,0.0061898106,-0.4028666914,-0.1746915579,0.3463866413,-0.0790270045,-0.0670975447,-0.0885521099,0.3795876205,-0.2153231949,0.1644384414,-0.2758260667,0.9699501395,-0.1210296303,0.0756631568,-0.3877993524,-0.1119085327,0.590333581,-0.2048414797,0.0601712354,-0.0335945748,-0.3451654911,-0.0255095121,-0.0025068086,0.0242286678,0.3962245882,0.0290653929,0.3269210756,0.2837750018,0.2704972029,0.0883139446,0.2414004505,0.1021610275,-0.3811486661,0.13286452,0.232842505,0.000943529,0.3418913782,0.0296669062,0.06484209,-0.1240692437,0.1012328491,-0.7577839494,-0.2139420807,-0.3507162929,0.0436202139,0.0502852425,-0.2678447664,0.4461229742,0.2037588805,0.5095028281,0.1625567228,-0.2427159697,-0.0011158952,-0.0816875547,-0.2073991895,0.2016285807,-0.0513993762,0.2647672296,0.0047987215,-0.1863396913,0.1751133502,0.0388051234,-0.3161408007,-0.4412825108,-0.3561668694,0.2335520536,-0.3172682822,0.2348845899,0.2059262395,0.0166881233,-0.3014271557,0.1950974166,0.2835125923,0.2865530849,0.1362299919,0.0921962485,-0.057964921,-0.3824459016,0.2452124208,0.0822057948,-0.1998982131,0.2826634347,0.4651038945,-0.1968359947,-0.226662606,0.1099112928,0.1775452197,-0.2984182835,-0.0382850617,-0.1456108838,-0.2014144957,0.118660517,0.2720454335,0.4747985899,-0.2090720981,-0.0033030135,-0.1260496676,-0.0483529493,0.2455235273,-0.0699733496,0.1024826616,-0.2855119407,0.211336121,0.1137873083,0.3128367662,-0.3668057323,0.1842241287,0.1234311312,0.0576565303,0.4841401875,-0.043131154,0.118682459,-0.2519409955,0.0317794941,0.0443167761,0.0200937185,-0.2575843632,0.029140003,0.1194110066,-0.0032426196,0.0531210601,0.0714929104,0.031892024,-0.1085828245,-0.0181500539,-0.0095392885,0.0798342004,-0.0294696148,0.462904036,-0.1698637754,-0.1433872581,0.1254790574,0.1680780649,-0.3947422504,0.3452174664,0.2613274753,0.0968923718,0.4265040755,0.0820404962,-0.1390731782,0.1200965568,0.1277493089,-0.0818490312,0.4949890375,-0.1090949327,0.1605901271,0.4223472178,0.1340919733,-0.3808861971,-0.1047189459,0.1454264671,0.0053880559,0.2655996978,-0.3391369879,-0.1223411933,-0.0475140028,0.1903776824,0.156934455,0.1819957346,0.30013448,0.1519844085,0.2274168879,0.116702266,0.5631713271,-0.0346525423,0.283672303,-0.0869765356,0.2064458877,-0.1356343776,0.2207376808,-0.0661067665,-0.0283751916,0.0988647193,0.0660903454,0.4115754962,0.0296082832,0.1415602267,0.4603031576,-0.0387735255,0.3335841596,0.2953754663,-0.0643002391,0.1179010272,-0.0899088532,0.4574866295,-0.2523855865,0.0063297306,-0.0677308366,0.0525082201,-0.0469506197,-0.0037048929,0.1858673096,-0.2191386819,-0.0752633587,-0.1628289968,-0.2939468324,-0.3484770358,0.4031432569,-0.0179534182,-0.145498082,-0.5478562117,0.0214937832,0.0940941796,0.3775739968,-0.1377046257,0.0390170589,0.00650584,-0.2297263443,0.2765110731,0.6441014409,0.1741458327,-0.2020914406,-0.055770956,-0.0002708347,-0.2083180547,-0.1925583333,0.1811538637,0.2652167082,0.2910472751,0.0529157035,0.0059721177,0.1664626002,-0.0909043923,-0.0215365272,0.0012131939,-0.0027710018,-0.3021706939,0.2089512497,0.0820143074,-0.2682292461,-0.1788060367,0.0007872992,-0.3088904619,-0.185646832,0.4646795392,-0.1991598904,0.0024087441,-0.1707443744,0.1247432977,-0.1156489775,0.2756515145,0.3615760803,-0.1205013841,-0.2004090548,-0.1713348776,-0.0457235426,-0.24279809,0.1364397854,0.1231557503,-0.2247141302,0.043340072,0.262994796,0.2088835388,-0.2679113746,0.0609809868,-0.120219402,-0.1037462652,-0.5572156906,-0.0213995744,0.2087282091,0.1916583329,0.0820417851,0.0517556183,-0.0353427529,0.0049040508,0.0223936047,-0.2996783853,0.1943834871,0.2767232955,0.0230524223,0.2054880559,0.0334662125,0.4271264374,-0.0015180822,-0.0768611655,-0.0870772824,-0.4335414767,-0.1251790673,0.2828869522,0.0304236021,0.3439725637,-0.3293581009,-0.2667044103,0.2854602039,0.0275790915,0.0840165764,0.0823161304,-0.5688415766,-0.0122401118,0.0258424692,-0.0636239946,0.1642417312,-0.0145257376,0.0040574134,-0.1538853496,-0.1045840532,-0.5019505024,0.0492037125,-0.0985216647,0.134104684,-0.2396983802,0.1752555221,0.5417258739,0.2522426248,-0.6720554233,-0.1178060845,0.4783791602,-0.0441265889,-0.0649471879,0.5287404656,0.206123054,-0.011946884,-0.0930266306,0.2641491294,-0.1595571041,-0.2162479013,-0.1624890864,-0.2623987496]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/353","title":"[Dataset requests] New datasets for Text Classification","comments":"This is the homepage for the Amazon dataset: https:\/\/www.kaggle.com\/datafiniti\/consumer-reviews-of-amazon-products\r\n\r\nIs there an easy way to download kaggle datasets programmatically? If so, I can add this one!","body":"We are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705","comment_length":26,"text":"[Dataset requests] New datasets for Text Classification\nWe are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705\nThis is the homepage for the Amazon dataset: https:\/\/www.kaggle.com\/datafiniti\/consumer-reviews-of-amazon-products\r\n\r\nIs there an easy way to download kaggle datasets programmatically? If so, I can add this one!","embeddings":[-0.1458348036,0.0658523887,-0.1481411606,0.190245375,0.3451894522,0.3538247645,0.1370032579,0.073779203,-0.1220624968,-0.0394350663,-0.2502006292,0.2587281764,-0.1631076485,0.4211382866,0.0899194926,-0.1647979021,-0.0227949545,-0.0886518732,0.1359842271,-0.1430172473,-0.1249985769,0.1028074697,-0.09122812,-0.2270246893,-0.1792382449,-0.3646232486,-0.1602799743,-0.1030361876,-0.3790989816,-0.3690577149,0.3359175324,0.1774003953,0.2373439074,0.1500301808,-0.0001078193,-0.3248600364,0.0003606891,-0.1197961271,-0.3088076413,-0.1295684427,-0.1039344147,-0.5073528886,-0.0890804902,-0.2598694265,0.0910557508,-0.0354643874,0.0495942235,-0.2475443631,-0.0602004416,0.2949465811,0.204582572,0.0185297076,0.0701869354,-0.1878490746,0.334087044,0.4829710722,-0.2553501427,0.0124023221,0.46330899,0.2070602924,0.1856566668,0.2027293295,-0.3064487576,-0.1065172404,0.2755382359,0.112584509,-0.3106137812,-0.4856223166,0.1785595268,0.4579693079,0.970321238,-0.4533036947,-0.5060064197,-0.3497488201,0.2375041395,-0.152540639,-0.3129490316,0.5556718707,-0.0784912631,0.0007059887,-0.1585119367,-0.4362519979,-0.1533342749,0.1665377468,0.2362736017,0.3287392855,0.070403479,-0.2436540127,0.0811757594,-0.0162672531,0.5947266221,-0.3934555352,0.130332619,0.0336790383,-0.2174791098,-0.2010198236,-0.1059347987,-0.0421087109,0.0622147433,0.0863996148,0.0007821672,0.1940434873,-0.1398465484,-0.1695522964,0.0122073432,0.0201016217,0.2767164409,0.3368701637,0.0530988425,-0.0473928787,0.2718046904,0.0226452984,-0.3008702993,0.1779805571,-0.0927955508,0.0095043881,-0.2158112824,-0.1667107344,-0.1336227059,-0.226462096,-0.1438032538,-0.0931730792,-0.2892290354,0.1214969605,-0.1194738746,0.0921548381,-0.087097995,0.2747785747,-0.1070149317,-0.3732473552,-0.0636869371,0.101596415,0.2123429179,0.0125101032,0.3886572123,-0.3078111112,0.2903837562,-0.2664626539,-0.0588190183,0.0681074262,0.0345649868,0.0559180491,-0.331359446,0.1262033135,0.089690797,0.0198177435,-0.0611537881,0.0645950586,-0.2371143401,0.2297895849,-0.0074012163,-0.2494551241,-0.1083722338,0.1898384243,-0.0466048308,-0.6931470037,-0.344630897,0.4325671792,-0.0215864442,-0.0997880623,0.1143115535,0.0151210381,-0.0687841699,-0.2183511853,0.0221006982,0.4878204167,-0.4603776932,-0.2505888939,-0.2430824935,0.0543338768,0.0991812646,0.0070484811,-0.0971060023,0.3354073763,0.0367300101,0.0289713331,0.4266469479,-0.1561398059,-0.2334921807,0.0047492487,-0.0274881925,-0.1572909802,0.1542441398,0.4053142369,0.0540761799,0.0954790413,-0.1891800463,0.2154071778,-0.147028327,-0.0636374205,-0.0311037321,-0.2104397118,0.1887797117,0.453360647,0.5115642548,-0.141790837,0.1126803458,-0.0789700672,0.204727754,-0.0264047273,0.1127469763,0.1084835976,0.147639811,0.2481903732,0.0519464202,-0.1844959855,-0.3535889089,0.1882290095,0.1195325404,0.1798555851,-0.1976394951,-0.2929627001,-0.1732512861,-0.0184195079,-0.0598591976,0.0343751498,0.0665083006,-0.1729312241,0.245930627,0.2049800158,-0.0696805269,0.2434435338,-0.2629662156,0.3557042181,-0.019129198,0.1765373647,0.0324240662,0.1675954014,0.1569022685,0.1811335087,0.0304620825,0.0196961537,0.0671713352,0.0794843137,-0.0288240444,0.2057452798,0.3982343674,0.2409583777,0.3269950449,-0.3789923787,0.3036179245,0.0010745006,0.1522791535,-0.135686934,-0.2688925266,0.3572216928,0.0751191676,0.0856947824,0.0338318124,0.1737848371,0.0815224648,0.1841052026,0.2052222788,0.1305986792,0.2918301523,-0.2030846328,0.3116686642,-0.2133226693,-0.2172529995,0.3556894362,0.1185146421,-0.1999642253,0.196610406,0.1007307768,-0.3290809393,0.0510095693,0.1607095897,0.1727669239,0.1420184821,0.4380430877,-0.0073880027,-0.0723125786,0.1096052527,-0.2723276615,0.0723109022,0.3957447708,0.114891991,0.0230934341,0.2149344236,0.1084581465,-0.2109977156,-0.0760432109,0.1719285697,-0.0094922613,-0.0051254821,-0.1017907709,-0.0998298675,-0.3285647333,-0.1100289747,0.152197957,0.0331786238,-0.3579837084,0.1127663478,-0.1880217344,0.0209956411,-0.0725448057,-0.3703675568,0.5109134316,-0.2676707506,-0.165388763,-0.0368975513,-0.309289664,-0.1858089268,0.1964372247,0.2492019534,-0.0065035434,0.4684090912,-0.1205239296,0.4063918889,-0.4003465176,-0.4549329877,-0.0711317286,-0.0402050354,-0.1495891511,-0.0947711021,0.0039905128,-0.0539056435,-0.18788369,-0.1421690881,-0.0577224419,-0.195838213,-0.0704630613,-0.0550234467,-0.3058831692,0.0188405868,-0.7203142047,-0.5714181066,-0.0859818906,0.1243749335,0.0709805116,0.0216838457,0.0199498422,-0.0857780129,0.0222123154,-0.1208952665,0.1738438904,0.1174019277,-0.0295295194,0.2599438131,-0.0651600882,-0.3358579576,0.0561509468,-0.0589567013,-0.1528324634,0.1433831751,-0.6533876657,0.2787922025,-0.1593821496,0.2507083118,0.0915816873,0.0094199954,0.3900659382,0.3259739876,-0.0046840375,0.0481488816,-0.2004248649,-0.0664402321,0.0186327696,0.1881430894,-0.2580963671,0.147907868,0.2330837548,0.7644212246,0.1004631072,-0.0396000445,0.004106604,0.1000653878,0.2880669832,0.1741345376,-0.3623054922,0.0765023604,-0.2538989484,0.194787696,0.2485779673,0.2329382896,-0.4174831212,-0.3761481047,0.0852972493,-0.3985083401,-0.0593182184,0.0200704932,-0.1571809053,0.250010103,0.1569011658,0.1980464011,-0.3067079186,-0.3829735518,-0.007611277,0.0950429216,0.0403235629,-0.0855691954,-0.6600493789,0.0852116868,-0.4546732903,0.3500308394,0.2858995497,-0.0139333839,0.0228560716,0.0968736112,-0.0129582603,0.1042202637,0.3377256989,-0.4176976383,-0.1463476717,0.1085479558,0.1544068009,-0.2899884284,0.0166510213,-0.421819061,0.0120441373,0.1763529032,-0.1067231596,-0.136977151,0.1208311915,0.1174421832,0.0953431055,-0.1275434196,-0.1310062408,0.260982126,-0.1921807826,-0.2475022972,-0.1067176685,-0.1915423721,0.1351096183,-0.2206446081,0.1119871736,0.087113753,0.1021275669,0.2077430934,-0.388297379,-0.1546428651,0.338870734,0.2146318406,-0.0203534383,-0.0908816084,0.2794800997,0.7352371216,-0.2495378554,-0.6877534389,-0.2033219039,-0.0593380593,0.371330142,0.2866702676,0.3106376231,0.1781737655,0.045267731,-0.1820601672,-0.0866233185,0.2527271509,0.1468854994,-0.2068834603,-0.4491500258,-0.4511360228,0.4768800437,-0.0468299687,-0.0066018491,-0.0308256708,0.0884351209,-0.1857964694,0.079673253,0.0334792472,0.8832516074,-0.0965520069,0.1206953228,-0.2108294964,-0.2397917956,0.6575296521,-0.3693648577,0.1451699585,-0.0319654755,-0.1078589857,-0.1583987474,-0.0057650972,-0.066047214,0.2764970064,0.0986420661,0.2545510828,0.1750445068,0.2646405101,0.1889281422,0.2676707506,-0.2133773118,-0.3211888075,-0.0217805747,0.2072969079,-0.0450460166,0.3466682732,-0.0601663738,-0.0404977575,-0.2185213417,0.0740632936,-0.6114406586,0.0503235199,-0.111356087,0.2592781484,-0.1681750864,-0.3132202327,0.3115784824,0.3959937394,0.4740001261,0.1545553058,-0.2781150639,0.0338784941,-0.1677888632,-0.2845949829,0.2685397565,-0.1171770468,0.2070041746,-0.0916294158,-0.1466668248,0.1773543656,0.0102858702,-0.093631722,-0.2707796097,-0.2812490463,0.0278405268,-0.2544699013,0.0718578473,0.2199272811,0.060745094,-0.2442300022,0.1158013716,0.2720436752,0.3126243651,0.1492731273,0.1311455369,0.042663414,-0.4036286771,0.3365068436,0.0537920408,-0.2013221681,0.3659550846,0.3573597968,-0.2291146368,-0.1734786183,0.2591618598,-0.0151893487,-0.2552028596,-0.0624840111,-0.1510736644,0.1275444627,-0.0006344505,0.295442611,0.326721251,-0.2449730784,0.0683485717,-0.26706177,-0.0456879996,0.3528864384,-0.0263609812,0.1615918875,-0.0568490736,-0.0070580533,0.0637388453,0.1684003621,-0.2951265872,0.2014482319,0.0457097739,-0.1295149028,0.1684550494,0.1904905438,-0.0736700594,-0.1628782898,-0.0309058968,0.0137712546,-0.1415182948,-0.1543598473,0.0864760131,0.1561498642,0.00046323,0.0615994036,-0.0054318109,-0.1991835982,-0.1987439096,-0.0280100312,-0.0401108004,0.0799321011,-0.0388810113,0.2800905406,-0.051855702,0.2226857096,-0.1991787851,0.1217425764,-0.213632077,0.1685869992,0.3860096037,0.1973818392,0.4179407358,0.1422142833,0.0235009026,0.097543478,0.2849408984,-0.2350788116,0.6867839694,-0.2889358997,0.259245038,0.3534525633,0.0812408701,0.0852923095,-0.1868261099,0.2281637192,0.234226808,0.1856685728,-0.5083346963,-0.3514140248,-0.0117295077,0.2154450119,-0.0015338295,0.1440443546,0.3221586347,0.0514773168,0.1890985221,0.187958762,0.4547486305,-0.0768007785,0.4784501493,0.2017870843,0.1357188523,-0.0542968437,0.1197222173,0.2849607766,-0.019388441,0.2118199617,-0.0686850399,0.539617002,-0.1689623594,0.1156086922,0.3625087738,-0.4359865785,0.3034897745,0.2183664441,-0.0631244332,0.1779616177,0.0134653682,0.3344654143,-0.2260023654,0.0742387697,-0.2706797421,0.0230934676,-0.0754788816,0.1115901619,-0.0258196332,-0.4800797999,0.1415528208,0.0542163253,-0.0710732564,-0.6137166619,0.0098598236,-0.1722025722,-0.1956641525,-0.4460481107,0.1265332252,0.0475831665,0.4512725472,-0.2456586063,-0.2158625722,0.2288338542,-0.2907611728,0.0318141133,0.4591094553,0.2172703743,-0.0518815666,-0.1536590606,0.0496250801,-0.0833993778,-0.1502198279,0.2596854866,0.3501443565,0.3649348021,-0.0309637971,0.2239759862,0.1064252034,-0.0772897676,0.2208933532,0.2014784664,0.0658394918,-0.2684996724,0.1456182152,-0.0097033149,-0.2128193229,-0.1392196566,0.0354055986,-0.2620571256,-0.1173944399,0.3134700358,-0.0482907109,0.0353490077,-0.1118346453,0.1048950851,-0.0433282629,0.3734110594,0.3794985414,0.0498736054,-0.192576915,-0.0144012105,-0.1483947784,-0.2592768669,0.102871798,0.0508901887,-0.1302158386,-0.0016712105,0.269788444,0.2477408201,-0.353395611,0.0241540819,0.2377682179,0.1335591823,-0.531668961,-0.156987235,0.3332626522,0.0851756036,0.0234206486,-0.0445228517,0.0223556403,-0.2352219671,-0.0853207484,-0.290889591,0.1221309975,0.1611092091,-0.0532051586,0.3564682901,-0.2169418633,0.5409767628,0.018423032,-0.0958150104,-0.1442964226,-0.2210029662,0.0111723086,0.2358661592,0.1549036801,0.2718773782,-0.323260963,-0.1143455952,0.35959059,-0.0222015213,0.1469313502,0.0603667609,-0.4647145569,-0.1330141276,-0.257071197,-0.1377783716,0.2574695647,0.003706241,-0.045511242,-0.2148298621,-0.1155497283,-0.3864067495,0.1926054955,0.0864056945,0.12661843,-0.2853858173,-0.0096458187,0.3591628969,0.2550026476,-0.4740018547,0.220404923,0.4707938135,-0.1923101246,0.1209472567,0.4455162287,-0.0076179537,-0.112205945,-0.1521120667,0.5035298467,-0.2934396267,-0.2272112221,-0.2597520053,-0.248032853]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/353","title":"[Dataset requests] New datasets for Text Classification","comments":"Hi @jxmorris12 for now I think our `dl_manager` does not download from Kaggle.\r\n@thomwolf , @lhoestq","body":"We are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705","comment_length":16,"text":"[Dataset requests] New datasets for Text Classification\nWe are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705\nHi @jxmorris12 for now I think our `dl_manager` does not download from Kaggle.\r\n@thomwolf , @lhoestq","embeddings":[0.0363270715,0.1208182648,-0.1671025902,0.2490486056,0.2784865797,0.2174416929,0.2918265462,0.058676593,-0.0694200024,0.0894200951,-0.0749871731,0.2429377139,-0.2216520309,0.1872980148,0.1075434536,-0.2674973309,-0.0024418817,-0.0645568967,-0.05856473,-0.1949672848,-0.1161485165,0.2323865891,-0.1021067202,-0.2988000214,-0.3894096315,-0.1188440919,-0.1741247177,0.203107357,-0.3696795106,-0.5515012741,0.0710659698,0.1108713523,0.2446143627,0.3044234514,-0.000100496,-0.2866234183,0.1290518939,-0.0962061509,-0.2718328536,-0.154832527,-0.1398425251,-0.5224165916,-0.0034975831,-0.1702831686,-0.0646550804,0.0960617885,-0.010226341,-0.1524552703,0.0233407225,0.3746043742,0.284389317,0.0045372597,-0.0091257114,-0.3068839014,0.398807466,0.3493154943,-0.2872608602,0.0362687595,0.6251108646,0.1298020631,-0.0595832877,0.3732942045,-0.2840688527,-0.0756148323,0.1847102493,0.1710983813,-0.0633509457,-0.4465399981,0.1536754221,0.3627845943,0.8537475467,-0.3193746805,-0.2965325415,-0.3729111254,0.1825645268,-0.2185397595,-0.0184796993,0.1393242925,-0.0681619048,0.0716142356,-0.0426279232,-0.244466722,-0.1498306096,0.2160182297,0.0402437672,0.3208631873,-0.0304148775,-0.3652468026,0.0909142271,0.1313055605,0.427668035,-0.2541504204,0.0599310137,-0.0015200948,-0.2455904633,-0.1704480052,-0.002382966,-0.1185175255,-0.0255971979,0.1150279269,-0.1519063264,0.1014916748,-0.1292260587,-0.1060267687,-0.0125242872,0.1141977087,0.1958247125,0.3053243756,0.1475717723,-0.0988136381,0.1847091764,0.045794826,-0.2336816341,0.13585715,-0.0357335284,0.0952232704,-0.2831825316,-0.0848114043,-0.0323271267,-0.0517270938,-0.049588751,-0.1634258926,-0.2022054493,0.1117959842,-0.0234020054,0.1728325188,-0.1526251286,0.2224625796,0.0433585718,-0.3606481254,-0.1383419335,-0.0274214,-0.0917269439,-0.0014240913,0.5291514993,-0.1442968696,0.2326451689,-0.2287693918,-0.1358251274,0.1641445458,-0.0226340704,0.0378942229,-0.0145485923,0.2572909892,-0.0061306641,0.1331600547,0.03365447,-0.0673391968,-0.2038238794,0.2830096483,0.0436704792,-0.3164167702,-0.2934389114,0.2884786725,0.0269857459,-0.5546091795,-0.1763255596,0.5332195163,-0.0657942817,-0.0989340022,0.2336559594,0.0619950481,0.0190879423,-0.1440949589,0.0625898838,0.3794981241,-0.4466327429,-0.1556040496,-0.3843356669,0.0678201392,0.1779544353,0.1370632648,-0.1412973702,0.103801243,0.0061652353,0.0018525793,0.3576285243,-0.2106905133,-0.2597446144,-0.0973088369,-0.032531511,-0.1417059153,0.2275728881,0.2446389496,0.2304923683,0.0032837908,-0.2420407385,0.1514585763,-0.2314142436,0.0088747144,-0.1517031193,-0.1266579181,0.1738824397,0.4673376083,0.5211064816,-0.2039029002,0.0964089707,-0.102705583,0.1300780773,0.06458202,0.181998238,0.1078508198,0.0502419733,0.2105703354,0.1069571823,-0.252481252,-0.3206230104,0.0419504344,0.2923546135,0.4220080674,0.0574039109,-0.1772577018,-0.2055796236,-0.0860821232,-0.0594474226,-0.1151983812,0.1802355498,-0.1373231858,0.0746803433,0.2228439152,-0.2554362416,0.1196095049,-0.2633062005,0.3440307677,0.1384392679,0.1995771229,-0.1859367192,0.0385998711,0.1123191118,0.3069336414,-0.0818982124,0.0136688957,0.0554979034,-0.0457963422,-0.1088721529,-0.0359763168,0.3803277016,0.3463966846,0.443066895,-0.3643418252,0.256474942,0.0166180059,0.1205729395,-0.1962995529,-0.1523725539,0.1953105032,0.1451340467,0.1179495901,0.0073442226,0.0682806447,0.0312686414,-0.0739850402,0.1388143301,0.0883899033,0.3093252778,-0.1635049582,0.2068687677,-0.0220921505,-0.1827761531,0.2687956095,0.2085866928,-0.114101164,0.28835392,0.1803227514,-0.4136097133,0.0028619606,0.0334143117,0.1817580909,0.1602043211,0.4741146266,0.0715278685,-0.1295448095,0.1909532696,-0.4396854043,0.0570931099,0.3365569711,0.0779422298,0.1949090809,0.288970381,0.1609755307,-0.252912432,-0.0739216432,0.1869309545,-0.0621049628,0.0964879394,-0.1228912994,-0.1473363042,-0.4734416306,0.009839247,-0.0316348821,-0.2634119987,-0.3572675884,0.2397592068,-0.1863238961,-0.0097318674,0.1090915799,-0.2038612962,0.4440691173,-0.3063391149,0.0034324632,-0.1450700462,-0.2510160804,-0.2677837014,0.2888260186,0.1671507657,-0.0761385188,0.4527226686,-0.139581725,0.3039369583,-0.2922314703,-0.5026999116,0.0818642601,-0.1006454229,-0.1514420956,-0.0796726793,-0.2038016468,0.0558108762,-0.132137239,-0.0117888311,-0.1040636003,-0.1461094916,-0.13238509,-0.1588013619,-0.1699044108,-0.0905093327,-0.6321281791,-0.4981149435,-0.1765124798,-0.0205679294,0.0031175094,0.0394998901,0.0527027212,-0.2293874621,-0.0312613882,-0.156179145,0.1062041372,-0.0379838757,-0.1608324349,0.1060853451,-0.1881787926,-0.2769782841,-0.0139446566,-0.0938116834,-0.0592466965,-0.2440544963,-0.5775509477,0.011682177,-0.1460658312,0.135050863,0.0823446438,0.014955977,0.4460029602,0.3186326623,-0.1670176983,0.0728072003,-0.1841626465,-0.0408815593,0.0871326849,0.3037729561,-0.3289527893,0.1120520905,0.1990931034,0.6252944469,-0.0762832314,0.0286285691,0.1277460605,0.1458060294,0.3261945248,0.0009282017,-0.3885497153,0.2392318547,-0.1540058702,0.1448472142,0.3402160406,0.244049564,-0.4113268554,-0.3100475073,0.0058306102,-0.3088421822,-0.1392024606,0.2457527518,-0.121912986,0.1020714194,0.0751668066,0.0301390737,-0.1549415737,-0.4110212028,-0.0093675535,0.3312243223,0.0705697611,-0.1585753411,-0.5272670984,0.0967199206,-0.5372110009,0.1099708229,0.2433333695,-0.1414558887,-0.1414815634,0.0674248785,-0.0124097103,0.2497187257,0.2316854596,-0.3759273887,-0.3057524562,0.2944363952,0.0295410119,-0.0160272066,0.0980092809,-0.419567436,-0.1425141245,0.350779891,-0.2209923714,0.049856551,-0.0149006844,0.3752944171,-0.0834888816,-0.0421894714,-0.2808251679,0.2156101167,-0.0121061858,-0.3490279913,-0.09257029,-0.0554535091,0.1127369255,-0.2458267361,-0.0381891578,0.2279602289,-0.0590561293,0.257951498,-0.216237396,-0.2635127306,0.2018695474,0.1389270723,0.1079142243,0.1332260519,0.3016775846,0.5075218081,-0.2157354504,-0.3669584394,0.0352633931,-0.1764235198,0.442897886,0.2902781963,0.3332870007,0.2164183408,-0.1570171118,-0.1707198471,-0.1184295267,0.1177883968,0.1813747138,-0.1047724262,-0.4715121388,-0.2676584721,0.4235581756,-0.0750060529,-0.0745217204,-0.0366296694,0.1526153088,-0.2258865535,0.0945143476,-0.0756776929,0.9857670069,-0.0546133369,0.1973148733,-0.2663496733,-0.0489397645,0.535959661,-0.3048813343,0.2788438797,-0.1173877791,-0.1341159493,-0.0623155907,0.0381366611,-0.0863123387,0.2807739079,0.0571569502,0.2046216279,0.1323180348,0.2306620777,0.0560077578,0.2629805803,-0.1214092746,-0.4137179852,-0.0063284473,0.2581975162,0.1662629843,0.3738595247,-0.0688864663,0.0217226874,-0.2274913043,0.0718559772,-0.6199696064,-0.0948185474,-0.2102095783,0.0355348922,-0.0633566454,-0.3951873481,0.3602654636,0.2881467938,0.4655901194,0.0671775937,-0.1785434335,0.1044784561,-0.1090021208,-0.1447925419,0.2486453801,-0.0874030516,0.2993197739,-0.1255787909,-0.1634186357,0.2403898239,0.0138765108,-0.2040067017,-0.3762980103,-0.1987235993,0.1692372561,-0.1861861944,0.3645673096,0.1886492819,0.0109080421,-0.2813081741,0.224824816,0.2207089216,0.2425478846,0.1298389733,0.0734993666,-0.0061556716,-0.4221653044,0.2767052948,0.1592311859,-0.2599048018,0.3653331995,0.4951324761,-0.1639520675,-0.3070437908,0.1767657846,0.1588031054,-0.1803541034,0.0462118648,-0.1370422393,-0.1898177415,0.1539030373,0.4831224382,0.4475258887,-0.2244649082,-0.0222328361,-0.1541671455,-0.1365885884,0.2057547569,-0.1310404837,0.064239189,-0.3104792535,0.1135818884,0.1106724367,0.2138182968,-0.4183434844,0.1057856381,-0.0026916282,-0.0010271918,0.2698207796,0.0371616893,0.1687712669,-0.1148047969,0.0779277161,-0.0281155333,-0.0769268796,-0.2974217832,0.0503066219,0.0981894732,0.0278768744,0.0624202043,0.0130275786,-0.0557430647,-0.1584130973,0.0195385981,0.0429111458,0.0156413764,0.0653073564,0.3806163371,-0.176500544,-0.0038350357,0.076705277,0.1506212205,-0.2631678283,0.3517302275,0.2140648365,0.1563253552,0.4049961269,0.1145723313,0.1018358022,0.1170222461,0.0556378625,-0.1151006073,0.5720162392,-0.1836516708,0.2140289694,0.5334522128,0.1297416538,-0.1990553588,-0.1963352114,0.087670587,0.1470390856,0.3044627011,-0.3589924574,-0.2522806227,-0.075388141,0.2254796922,0.0856962055,0.1171586812,0.3760237992,0.0447909869,0.1790889949,0.2037140876,0.4474936426,-0.0895675123,0.3262723088,-0.0159410387,0.0945716053,-0.01618414,0.2691628337,0.080552116,0.02957456,0.2775474489,0.0727246031,0.533880949,-0.1040646583,0.1475655138,0.40432477,-0.2420401275,0.3320989609,0.2094635665,-0.0561638921,0.117655918,-0.0231846999,0.3225650489,-0.2496720999,0.0185770672,-0.2316113263,0.0487239584,-0.1139774024,0.1500729024,-0.0096637728,-0.3767280877,0.0208674911,-0.0294172838,-0.2372244149,-0.3354690969,0.1512869,-0.075034745,-0.2642959058,-0.537863493,-0.0051315203,0.0171665251,0.3859813213,-0.2130107433,-0.0802979022,0.2222642601,-0.2817739844,0.2070705146,0.5161467195,0.2406761795,-0.06502112,-0.1272567809,-0.1005326957,-0.2148464918,-0.1743795872,0.1005886346,0.3649821281,0.4146254957,0.0134759257,0.0421180315,0.1669245511,-0.1491863132,0.0155162429,0.242221877,-0.0947978348,-0.0975517184,0.1830942631,0.1417445093,-0.2188388705,-0.2323552966,0.0394623093,-0.3943842053,-0.0144924177,0.423203826,-0.0835352167,0.0879913941,-0.196663633,0.1408361197,-0.0525404997,0.3778327703,0.3114274144,-0.0616152994,-0.1288599223,-0.1837223023,-0.1686691642,-0.1467707157,0.1684227139,0.1683295816,-0.174813211,-0.0063794116,0.2265821546,0.2520141304,-0.2523545623,0.0136280376,0.1262555271,0.0143537829,-0.5593928099,-0.0895801708,0.2098473459,0.0673221871,-0.0537847951,0.0548974834,0.0260824058,-0.2426677048,0.043141149,-0.3713621795,0.2435296625,0.2112380564,0.1630855501,0.2539019883,0.0116871456,0.4366324246,-0.0334259421,-0.1314099431,-0.0681050345,-0.233093679,-0.1569173485,0.1507695019,0.1977438033,0.2544566691,-0.2499749511,-0.101615943,0.2324191183,0.2099459618,0.0980769023,0.0916349366,-0.6198494434,-0.0849127695,0.0384933315,-0.1675667316,0.287828356,0.0637846962,0.0505585931,-0.231416598,-0.2736485898,-0.4273025692,0.06004823,-0.054565452,0.1208738014,-0.4299781919,0.0572426915,0.4172289968,0.3035710454,-0.6652132869,0.0371349603,0.3935830593,-0.1419793069,0.0052053058,0.3796028793,0.1705808342,-0.0874247551,-0.1887239963,0.2291713953,-0.1860613525,-0.2594029605,-0.2517861426,-0.223628059]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/353","title":"[Dataset requests] New datasets for Text Classification","comments":"Pretty sure the quora dataset is the same one I implemented here: https:\/\/github.com\/huggingface\/nlp\/pull\/366","body":"We are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705","comment_length":13,"text":"[Dataset requests] New datasets for Text Classification\nWe are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705\nPretty sure the quora dataset is the same one I implemented here: https:\/\/github.com\/huggingface\/nlp\/pull\/366","embeddings":[0.0566885769,0.1276321411,-0.1692308486,0.196194306,0.1949201524,0.2260396034,0.1887369603,0.1263084263,-0.145042792,-0.0797479972,-0.1454940289,0.1288715601,-0.130829528,0.1290705651,0.1513905972,-0.2922786474,-0.0173573159,0.010623279,-0.0858447701,-0.2069904655,-0.1079873741,0.2775412202,-0.1154190823,-0.2038578093,-0.3473118544,-0.0519542173,-0.1968381256,0.1525246799,-0.2313973904,-0.5007256866,0.1085518897,0.1020250618,0.1363148391,0.2952294648,-0.0000993312,-0.2425292283,0.0392379873,-0.0552243814,-0.1680204719,-0.2713805139,-0.1050099358,-0.4583977759,-0.0546974726,-0.1187517121,-0.1353191435,0.012787451,0.0710248053,-0.19393152,0.0423695408,0.3523743749,0.3028116524,-0.0372743644,-0.0394755863,-0.1225800663,0.3064972162,0.4627343714,-0.2892799675,-0.0700257048,0.5567073822,0.0870810822,-0.0829056799,0.3072106242,-0.2367987484,-0.0959107727,0.1755213141,0.1687210351,-0.2400723547,-0.3746621311,0.0080022989,0.3169445395,0.724105835,-0.2802686989,-0.3347759545,-0.2969860435,0.1678986698,-0.259961009,-0.1142954081,0.2169654369,-0.0249432735,0.0996665955,-0.1884857416,-0.2008580863,-0.2064570934,0.243205443,0.3086127043,0.1064747795,-0.0290261675,-0.2586435676,-0.1240092367,0.1151477247,0.2530893087,-0.3190906644,0.2195166051,-0.0062396051,-0.1840710044,-0.213337481,0.0533437803,-0.0433519222,-0.0071359072,0.0766493827,-0.1416177452,0.1000142917,-0.2270857245,-0.0650446787,-0.0627850741,-0.0825469121,0.1981089562,0.2509924173,0.072413601,-0.1877155006,0.1094491631,0.0574723706,-0.1182158291,0.1656758934,-0.0850731134,0.0659916177,-0.0967150107,-0.2286707163,-0.0140280444,-0.0984279513,-0.0394150652,-0.1082354411,-0.1851137131,0.0442170091,-0.0002593925,0.1392648518,-0.1629822999,0.2366105169,-0.122479178,-0.4066718817,-0.1655023992,-0.0459691398,-0.1297450215,0.103172183,0.6136122346,-0.1069877297,0.2366069406,-0.1730463356,-0.1652943343,0.1489962041,-0.0684234723,-0.0138433296,0.0432741307,0.1350877285,-0.1175006032,0.1105284691,0.1036159396,0.1324439645,-0.2042533308,0.2282742858,-0.040227998,-0.2480108291,-0.1846319586,0.3113862574,-0.002523537,-0.5981891751,-0.1284911036,0.5943570137,-0.0074594216,-0.1413792223,0.177859515,-0.0080517679,-0.1221423447,-0.1345061809,0.0074051595,0.380038023,-0.3987297118,-0.2305645049,-0.2739988565,0.1073465273,0.171921581,0.1894022524,-0.1653505266,0.0906360075,0.0684951097,0.130011335,0.235757038,-0.0038216268,-0.2056534588,-0.1653785259,0.0125846276,-0.1436440349,0.1121663526,0.2281587273,0.1650720537,-0.070502542,-0.0665749088,0.2074082196,-0.2749081254,0.0149288271,-0.1888124049,-0.0811936334,0.1578860432,0.3825259805,0.4316782951,-0.1883490384,0.0578639619,-0.1855609566,0.1248506829,-0.0308266748,0.1929667592,0.1637401581,0.1451422572,0.1593023092,0.039491117,-0.3435085714,-0.1554725319,-0.0115389349,0.2477961481,0.3604110479,0.0262485687,-0.1111834943,-0.180613026,-0.167441681,0.0416526422,-0.1014013886,0.2275230438,-0.1091317385,0.1404931247,0.1475247443,-0.2935887277,0.211031273,-0.3204366267,0.3320719302,0.0897272006,0.2733629942,-0.1356969029,0.0071234889,0.238120243,0.4077989459,-0.0327600166,-0.0259159356,0.0487101227,-0.0135566294,-0.220540002,-0.018131882,0.3435701132,0.2997593582,0.3508870006,-0.4878053367,0.2263718396,-0.0637268052,0.086687386,-0.1745656282,-0.086808905,0.3148507476,0.0222974718,0.0837244615,0.0605034642,0.0229849089,0.0508848839,-0.0748004392,0.091767095,0.1362914145,0.3747400343,-0.105576098,0.1675774157,-0.0063940897,-0.1870827228,0.2635902464,0.2286502719,-0.0632698238,0.2088905424,0.2689296603,-0.3897541165,-0.1135973707,-0.0336717442,0.0816278309,0.3191332221,0.5359097123,-0.0040912041,-0.0959875211,0.0097796107,-0.3985590041,0.0624519363,0.3202102184,0.0407854281,0.0594218895,0.334349215,0.2404499948,-0.2901435196,-0.1746359617,0.1445587873,-0.0615298152,-0.0035620367,-0.0580132753,-0.1040819138,-0.3936361969,0.0261012018,-0.0152848894,-0.1177802235,-0.3084943891,0.2878182232,-0.2400050461,0.0164696649,0.1174916923,-0.2951138318,0.5113760829,-0.2848060131,0.0570852347,-0.1215823293,-0.2165778577,-0.2627609074,0.316193372,0.0455674529,-0.0169072058,0.4866949916,-0.1036536172,0.2629494369,-0.2394257933,-0.5118152499,0.0301073082,-0.2122667134,-0.0772042498,-0.1596073061,-0.1599481255,0.0497714803,-0.1247152835,0.009976075,-0.1663482785,-0.1935194731,-0.0200915374,-0.1706539094,-0.1929354668,-0.1064104512,-0.6765928268,-0.4070600271,-0.1236726791,0.1352696717,0.0459915884,0.0606621653,0.2996735871,-0.1676939279,0.0063125165,-0.2143511027,0.1296746582,-0.0751152113,-0.2595367432,0.1465547532,-0.1889096349,-0.2506606877,-0.1184833422,-0.056388922,-0.0236005634,-0.1015982553,-0.537786305,0.0727282912,-0.249334842,0.1581681967,0.1387419999,-0.0447424836,0.4599045813,0.262439847,-0.2180176079,0.0146510862,-0.1256978959,-0.0505199805,0.0395123027,0.277728945,-0.3181885779,0.081449464,0.2324476093,0.5726889372,0.0491083041,-0.050136067,0.161346212,0.2307810783,0.3370778859,-0.0160695482,-0.3762290776,0.2725857496,-0.1478512138,0.143118903,0.321231097,0.2405260503,-0.3563091159,-0.2082580775,0.0456567779,-0.3162105083,-0.1143864989,0.2321375012,-0.0665700808,0.2561680973,0.0919159949,-0.1518717408,-0.2438979,-0.4505446255,-0.1096240878,0.2330022603,0.1787249595,-0.1543555558,-0.5836022496,0.201749891,-0.5448508263,0.1821356416,0.1576204747,-0.0692609847,-0.1941453069,0.1321438849,-0.1189996302,0.3481927514,0.061504744,-0.37629655,-0.3204020858,0.2345654666,0.0772866011,-0.1078448519,0.1639022231,-0.4134991169,-0.0069971364,0.2695349157,-0.0338973366,0.0586231612,-0.0357066393,0.4773080647,-0.0109822964,0.0300932806,-0.284185946,0.2958496809,-0.0230785944,-0.1973949671,-0.1384482384,0.0517243482,0.1067133397,-0.1979449987,0.0097573986,0.2032691687,-0.0353980586,0.3483685851,-0.2567231357,-0.278450489,0.1593092978,0.1666059941,0.0557359308,-0.0492439866,0.2516543269,0.5195673704,-0.1753330976,-0.4649529159,-0.0042892531,-0.1196892858,0.5018873811,0.2621311843,0.3538145721,0.21812208,-0.1593071818,-0.1449495554,-0.1002187878,0.1635531336,0.0865008608,-0.0292103067,-0.4421637058,-0.2974823117,0.3116586506,-0.107786119,-0.0696586668,0.0021976077,0.215551421,-0.2193934023,0.1646095365,-0.0952301398,0.9357751608,-0.0339918435,0.1758350432,-0.247703135,-0.0414298698,0.5485967994,-0.290488869,0.23985897,-0.0338846669,-0.2250061333,-0.0501299463,-0.0373613909,-0.0392283984,0.3679526448,0.0946825519,0.2773178816,0.2715792656,0.1054661348,0.1360050291,0.1615828276,0.078473337,-0.4171737731,-0.0262289215,0.2877217233,-0.0241436735,0.2933602631,0.0064183464,0.0491978899,-0.2058167309,0.1116337627,-0.6511619091,-0.0962242633,-0.16924797,0.0895168707,-0.0122665511,-0.3669750392,0.425136745,0.2910899818,0.4683715105,0.0655864403,-0.2364071757,0.0497492813,-0.2297098935,-0.1208475977,0.1571404636,-0.1393771619,0.2988036573,-0.0128204273,-0.1662674546,0.163367033,0.0222675744,-0.2130790651,-0.4184070826,-0.29148826,0.1901490241,-0.2781490684,0.2767622173,0.2203001082,-0.0108667826,-0.3241910636,0.238634333,0.2488185465,0.2798657119,0.0688006133,0.1917796135,-0.0646993294,-0.3642306626,0.1882019937,0.2167962343,-0.2216316164,0.4334492385,0.4509546459,-0.1628475934,-0.2691655755,0.158796221,0.1911682487,-0.3155724704,-0.0387084447,-0.1220817566,-0.037579637,0.1578077674,0.3629391193,0.3372302949,-0.1929927468,0.0548638888,-0.2440165728,-0.09742181,0.1963335723,-0.1342002749,0.0479698554,-0.1524371654,0.1658715308,0.1127586812,0.2109071463,-0.4276207685,0.1180061549,0.061469648,0.0460518636,0.2023475468,-0.0635984316,0.1617216021,-0.2232262641,0.1054634228,0.044124838,-0.027034672,-0.3256803751,0.0056834193,0.0987438262,0.0305090696,0.0269424878,-0.0335607752,0.1142819002,-0.1009441987,-0.050422702,-0.1250747591,0.0935462043,0.0776980668,0.4519529641,-0.1866676956,-0.0644847378,0.095596768,0.0493523404,-0.3074078858,0.356852442,0.2733794749,0.1151651815,0.367881,0.0919522569,-0.1764284521,0.1478414237,0.1486049891,-0.088684313,0.5652942657,-0.1879415214,0.2640521526,0.4511682391,0.1291651875,-0.2227452546,-0.1402050853,0.1075098589,0.0267667416,0.3362828493,-0.3700738847,-0.1413968951,0.023285972,0.2312207371,0.0784939378,0.1846355647,0.2664148211,0.0420541801,0.2439119518,0.0946001261,0.4632264078,-0.1192564666,0.2975398004,-0.1237983555,0.146475926,-0.0713052824,0.2710522711,0.0934846699,0.0426063202,0.2147865295,0.0913834497,0.4831658602,-0.1218613312,0.160753265,0.3389582038,-0.1755820662,0.3521150053,0.2216697484,-0.0010304585,0.1522942185,-0.0508621261,0.2544858754,-0.1999412477,0.0448518619,-0.1155168861,0.027466502,-0.0726248398,0.0548442006,0.0339571089,-0.3070440292,0.0241530184,-0.163664341,-0.2811687887,-0.3018662333,0.2150625587,-0.0720668882,-0.2239278555,-0.6502150893,0.1655270308,0.107879959,0.4071906507,-0.1555340141,0.0067360452,0.1586491466,-0.2434702516,0.2061605155,0.5628172159,0.2194208801,-0.1717648655,-0.0309628937,0.0419717543,-0.2220696509,-0.2008972019,0.1349166334,0.2947669625,0.416389972,0.0993275121,0.02014683,0.2230652124,-0.1166496426,0.1280083954,0.2059661448,-0.0310540311,-0.2148459256,0.145220682,0.1248303056,-0.2700580955,-0.1848663241,-0.0170572866,-0.2985635996,-0.1524666101,0.4838336706,-0.1163666174,0.0508341938,-0.128919214,0.1563222557,-0.1049951538,0.2574489713,0.3890725374,-0.1857486516,-0.1872067004,-0.1299801469,-0.1149913073,-0.1888962686,0.1060451418,0.1402379572,-0.2245564014,0.0288288873,0.2767218053,0.1857094914,-0.3433025777,-0.0912553519,-0.0669854805,-0.0089247636,-0.5570239425,-0.0362445638,0.1676517576,0.173308,-0.0048596393,0.1440144628,-0.0922470316,-0.0713135377,0.1000457183,-0.349606961,0.1652121097,0.2111988664,0.123915799,0.0756642967,0.1126959696,0.3952917457,-0.0541045777,-0.1699752063,-0.0871959925,-0.3560568392,-0.0945791155,0.3671696484,0.0355123505,0.3546119928,-0.2441942096,-0.2483315766,0.2620430887,0.0979209542,0.0525791869,-0.0932520702,-0.5240798593,0.0526789464,-0.0009105002,-0.0635932758,0.1885968894,0.0482219085,0.0377901644,-0.1342862099,-0.1518946737,-0.461925298,0.0754498839,-0.122618109,0.1052092835,-0.2300112993,0.148701027,0.3777005672,0.2649074495,-0.6730523705,-0.0173679534,0.4908881783,-0.1797864586,0.0112607852,0.4316365719,0.0993543863,-0.0443056561,-0.1481413245,0.3596580923,-0.2104138583,-0.2349853963,-0.1386819035,-0.2268693298]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/353","title":"[Dataset requests] New datasets for Text Classification","comments":"Great list. Any idea if Amazon Reviews has been added?\r\n\r\n- ~40 GB of text (sadly no emoji)\r\n- popular MLM pre-training dataset before bigger datasets like WebText https:\/\/arxiv.org\/abs\/1808.01371\r\n- turns out that binarizing the 1-5 star rating leads to great Pos\/Neg\/Neutral dataset, T5 paper claims to get very high accuracy (98%!) on this with small amount of finetuning https:\/\/arxiv.org\/abs\/2004.14546\r\n\r\nApologies if it's been included (great to see where) and if not, it's one of the better medium\/large NLP dataset for semi-supervised learning, albeit a bit out of date. \r\n\r\nThanks!! \r\n\r\ncc @sshleifer ","body":"We are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705","comment_length":92,"text":"[Dataset requests] New datasets for Text Classification\nWe are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705\nGreat list. Any idea if Amazon Reviews has been added?\r\n\r\n- ~40 GB of text (sadly no emoji)\r\n- popular MLM pre-training dataset before bigger datasets like WebText https:\/\/arxiv.org\/abs\/1808.01371\r\n- turns out that binarizing the 1-5 star rating leads to great Pos\/Neg\/Neutral dataset, T5 paper claims to get very high accuracy (98%!) on this with small amount of finetuning https:\/\/arxiv.org\/abs\/2004.14546\r\n\r\nApologies if it's been included (great to see where) and if not, it's one of the better medium\/large NLP dataset for semi-supervised learning, albeit a bit out of date. \r\n\r\nThanks!! \r\n\r\ncc @sshleifer ","embeddings":[0.0099564521,0.1201289371,-0.1890824586,0.2179352045,0.212858811,0.2180952728,0.2061844319,0.1419543177,-0.186056301,-0.1030766219,-0.1535795331,-0.0204460192,-0.1454027146,0.2192624658,0.0605063699,-0.1807077974,0.0125586893,-0.0416300111,-0.0774128139,-0.2203581184,-0.1637407839,0.1607874334,-0.0990080684,-0.187086165,-0.3995814919,-0.1154435873,-0.1499630511,0.0827758685,-0.274766326,-0.4816746712,0.032899864,0.0766901076,0.1454466879,0.348202467,-0.0000977496,-0.2388429195,0.0384176113,-0.0287116487,-0.2368616611,-0.2248953581,-0.05126068,-0.4016395807,-0.0896616057,-0.1259749681,-0.1336456239,0.0372294113,0.0565926582,-0.1836146414,0.0036868402,0.3463908136,0.302976042,-0.0283137504,-0.0866739526,-0.1233578399,0.2595639825,0.5216811299,-0.3470135927,-0.1082492024,0.5796987414,0.1141715199,-0.0773810744,0.3088273704,-0.2269610912,-0.1091128886,0.1848832667,0.1643171161,-0.144945845,-0.3419852555,0.0569649972,0.3712534308,0.6867119074,-0.2845940292,-0.4477046132,-0.3079707921,0.160413444,-0.2913414836,-0.1641082615,0.2534989715,-0.0334708802,0.1641044319,-0.1707695574,-0.2218768001,-0.1847559214,0.194541499,0.2576349676,0.0969118178,-0.0571926571,-0.265884608,-0.0637662709,0.1105750948,0.2781270742,-0.3594803214,0.1945616901,0.0019449674,-0.1968546808,-0.3035461903,0.0432977565,-0.0020648381,-0.0313554294,0.0415115394,-0.1057237834,0.1028996855,-0.1994989365,-0.1225185767,0.0027354322,-0.1741217524,0.2056171149,0.2162726969,0.0856299326,-0.1431404501,0.1288760453,0.0778543279,-0.099279575,0.1666503102,-0.1030005291,-0.0183247197,-0.1320113093,-0.2542120218,-0.0467144512,-0.0375110134,-0.0053702989,-0.0627389401,-0.1830398887,0.0345913023,-0.0360739008,0.1088376641,-0.1634377241,0.2397159636,-0.0484842807,-0.4128642678,-0.1474092454,-0.0148842735,-0.1394736916,0.0897452757,0.6010738611,-0.0843565464,0.1415133625,-0.1657430977,-0.1212156042,0.1633103341,-0.0195770059,-0.1055756882,0.0884209424,0.1080307961,-0.1405914724,0.0719513223,0.080187127,0.1892982274,-0.2042859644,0.2494806051,0.0147895273,-0.2428173721,-0.1787462533,0.3136505485,-0.0418704785,-0.5465002656,-0.0762110278,0.6240939498,-0.0148734422,-0.0354613252,0.1351902187,-0.0337279476,-0.1267390549,-0.1285327673,-0.0070464928,0.3449832797,-0.4318048358,-0.1866945773,-0.3239611983,0.1120487973,0.1348615587,0.2293367237,-0.1754046082,0.0672391281,0.1419585496,0.0284113158,0.2154877186,0.0268196668,-0.2289338112,-0.0001927779,-0.0042736144,-0.1134461239,0.0808871612,0.2780292332,0.1063636839,-0.0669074282,-0.1142901704,0.1946752369,-0.2612954974,-0.0444032289,-0.2252386063,-0.1810869575,0.1740026176,0.4263752699,0.4575031996,-0.2117215842,0.0510214977,-0.1381463706,0.089514114,-0.0357856192,0.219382152,0.1367531419,0.1556399167,0.1583779156,0.0518692285,-0.2809582055,-0.1513505429,-0.0058419961,0.2436332405,0.3482487798,-0.0044400301,-0.0971555412,-0.0817783177,-0.1140727475,0.0308290329,-0.1058610976,0.2170848846,-0.1768062711,0.0776404962,0.1714372486,-0.2573434711,0.2146844119,-0.3896972537,0.3276173472,0.1842177361,0.2867545784,-0.0657102019,-0.0608383194,0.1710824966,0.3616926074,-0.1500442773,0.011408533,0.0230691377,-0.0033252565,-0.2483469695,-0.0362363942,0.2523275912,0.3923212886,0.3120626211,-0.4238718152,0.2985501587,-0.0272493307,0.0875369012,-0.1284648031,-0.1620542109,0.4090004265,-0.0216230936,0.0647427812,0.0060243458,-0.051721137,-0.0408727266,-0.0209088661,0.058990851,0.109634608,0.3607464731,-0.1972751021,0.2059567571,0.0017746818,-0.1617477685,0.1815605164,0.185368672,-0.0686452761,0.2047098428,0.2270733267,-0.3626619279,-0.1416245848,-0.0496312864,0.0105183348,0.2799790204,0.5136606693,-0.0318662785,-0.1047907025,0.055083409,-0.3672748804,0.0804235339,0.3299196661,0.0532738455,0.0742286444,0.2775609493,0.216109395,-0.2416408956,-0.1270565689,0.1679585874,-0.0353841558,-0.0167417303,-0.1019518748,-0.0277997404,-0.3347744644,0.0467789993,0.0481056012,-0.0874387771,-0.2642134726,0.2252803296,-0.2067918479,0.0417878851,0.0604249723,-0.3150878847,0.5391520858,-0.2429105192,0.0756278634,-0.0991195589,-0.2081801891,-0.2126780301,0.3141339421,0.0941990092,-0.0847781971,0.4556823075,-0.0665524378,0.2534090877,-0.191388607,-0.5189396739,-0.0148550747,-0.166545555,-0.043255534,-0.1463810056,-0.1647109538,0.0170951001,-0.1085614264,-0.055175934,-0.1354310364,-0.172918275,-0.0820486546,-0.1282363832,-0.2330127954,-0.0639267191,-0.6488698721,-0.3751605749,-0.0966444016,0.1201147959,0.0243501179,-0.0177140962,0.2662031054,-0.0593328327,0.0211563539,-0.1397211105,0.1283466667,-0.0776318833,-0.269048661,0.1993799359,-0.105558306,-0.2119161338,-0.1357294023,-0.0527598076,-0.0643496141,-0.0682598352,-0.5273972154,0.1693695039,-0.1972025782,0.1870899796,0.118149586,-0.0900440067,0.3759340644,0.2685990334,-0.1958193034,0.0663096979,-0.0772469938,-0.0818636045,0.0512133539,0.2883867323,-0.3087562323,0.1018721983,0.285759449,0.5518405437,0.0370039679,-0.1277289689,0.0581873246,0.2767521739,0.331504643,-0.0084972307,-0.3893318176,0.2839775681,-0.1347843558,0.1779325306,0.3239388168,0.187380597,-0.3543049693,-0.2062395364,0.0417388454,-0.2724870145,-0.0839297473,0.2441463917,-0.176972881,0.2626500428,0.0470926911,-0.0892001465,-0.2200296521,-0.4361051917,-0.1354859173,0.1896086782,0.1607812494,-0.1956532896,-0.6076632142,0.2148035467,-0.4611176848,0.1065192297,0.1472973973,-0.0778759867,-0.1626666486,0.1421458125,-0.1019606069,0.3258804679,0.1027676165,-0.4636113942,-0.3247773349,0.2446568906,0.029968407,-0.0822091699,0.1484399289,-0.4467480183,0.0217912626,0.2362397611,0.0361279771,0.1049464941,-0.0590054095,0.4637440443,-0.027275268,0.0316438898,-0.2878127694,0.2964057922,0.0648155361,-0.159772262,-0.0287914891,0.0629830286,0.0637437031,-0.1971363276,0.0318073481,0.1918150336,0.0667750835,0.3971684873,-0.2791351378,-0.3446131051,0.1628166586,0.2187755555,0.0670477375,0.0086488957,0.3463251591,0.4330371618,-0.1477717459,-0.5273458362,-0.0494022816,-0.1409042627,0.4833728373,0.3429302573,0.3617251217,0.2523905039,-0.1165616065,-0.1228625551,-0.0720191374,0.2508350015,0.1270611137,-0.0357369818,-0.5171096325,-0.3428186476,0.3212039471,-0.0887966976,-0.052865997,-0.0371814556,0.1858061254,-0.1782037318,0.1393808275,-0.0054333946,0.9393482804,-0.0167497601,0.217758745,-0.2964960933,-0.1543392688,0.4936573803,-0.2867141962,0.1672136337,-0.0426430069,-0.1965599805,-0.0001622751,-0.0425067432,0.010546729,0.388258189,0.105950959,0.2221493572,0.2518794537,0.1419556588,0.0797021836,0.1705516726,0.0826379284,-0.4323704243,0.0679334626,0.2891987264,0.0313089527,0.2347820699,-0.0014483785,0.0132837864,-0.1919443607,0.1294080168,-0.6016510129,-0.0803753883,-0.1888521463,0.067407541,0.0226119757,-0.4113134742,0.5069013238,0.2674866021,0.4851576686,0.1461804956,-0.2111534327,0.0121581517,-0.2298677117,-0.0999070182,0.2072235644,-0.1449709386,0.324198246,-0.0038156684,-0.2325271517,0.1667747498,0.0671089813,-0.1346790791,-0.4355966151,-0.3422700465,0.2398144901,-0.3440447748,0.2265376151,0.2108393461,-0.0073695183,-0.2803477943,0.2544867396,0.2258088142,0.248579517,0.1272220612,0.1870500743,-0.0966895074,-0.3992341459,0.2455760539,0.1889496893,-0.2624406815,0.3753581941,0.4487677217,-0.22259745,-0.3211379349,0.2563750148,0.2719466686,-0.23475191,-0.0509394519,-0.1291604936,-0.0731910542,0.1580411196,0.2855263948,0.3492252827,-0.2070112228,0.0567534938,-0.1386402398,-0.0904974565,0.2378675491,-0.0784673467,0.0557264313,-0.1956433058,0.1426939964,0.0733028576,0.1949750334,-0.4472538829,0.1631421447,0.1436947286,0.0508331954,0.2079174817,-0.0913311914,0.1202331558,-0.2043098956,0.1457258016,0.0205735452,-0.0671782717,-0.3357700109,-0.0075485907,0.1071916223,0.0622836687,-0.0503342375,0.0278255139,0.1178087667,-0.169655025,-0.1041068435,-0.1119066551,0.0892320052,0.1002388,0.392852962,-0.2086254209,0.0107552921,0.0919550434,0.0781638697,-0.349018991,0.3264140487,0.2692592442,0.1559086442,0.4530671239,0.0773401111,-0.1119287536,0.1587596685,0.1464133859,-0.0920319259,0.6272504926,-0.2160811424,0.2297983915,0.4485687315,0.135135591,-0.1592944115,-0.1255429536,-0.0036731493,0.0795669556,0.341350466,-0.3197437823,-0.1428836137,-0.0598406233,0.1881302446,0.075958088,0.2066175938,0.2478871793,0.0535622165,0.2319911122,0.082610108,0.5046768785,-0.1222791225,0.2788366079,-0.012822371,0.1634682119,-0.1063464582,0.2626691461,0.1599369943,0.0391112082,0.266923964,0.0883874372,0.4581475854,-0.1392396986,0.2407558411,0.2345954031,-0.1436107308,0.340421766,0.2780047357,-0.1257243752,0.2205054611,-0.0631456524,0.2299706042,-0.1750208139,0.049261719,-0.1364424974,-0.0477988832,-0.1134296432,0.0578445829,-0.0351770483,-0.3497605324,-0.0415518321,-0.1292924583,-0.2964368165,-0.3324360847,0.1411336362,-0.0628620163,-0.1891950667,-0.573241055,0.2355664223,0.1159089431,0.4328241348,-0.2088290602,-0.0516188294,0.1438401788,-0.1780401766,0.1907114983,0.4758204222,0.2004232109,-0.1719058305,-0.1040591821,0.1188338399,-0.2416636944,-0.1992031783,0.1770108044,0.2362865508,0.3595470786,0.1639284492,0.005833996,0.2157749683,-0.1195164472,0.1666228473,0.1374994516,-0.0744859502,-0.2650489509,0.080821611,0.1331862509,-0.2794405818,-0.1732406467,-0.0223399848,-0.2270413339,-0.2063815296,0.4948813021,-0.140098393,0.0367836878,-0.0952353701,0.1659815758,-0.1450923383,0.2787550688,0.3749412,-0.2043953389,-0.1946482956,-0.0743005574,-0.0456002131,-0.1837632805,0.1418144256,0.2079697847,-0.2435238957,0.0404481329,0.2586272061,0.2225130945,-0.3685129881,-0.074093692,-0.1136296093,0.037768092,-0.5360791683,-0.0562280715,0.1630550027,0.255902648,-0.0167484172,0.1442202777,-0.1042819917,-0.0879720077,0.0968448967,-0.4222563505,0.106496729,0.1884108186,0.1729870588,0.1173982695,0.0587842464,0.3530173302,-0.0459753722,-0.1179470643,-0.0879017711,-0.2885902524,-0.0794680491,0.3862180412,0.0675957426,0.3366408944,-0.2556755245,-0.3060522676,0.2978149652,0.0483824573,0.0209983606,-0.1510129124,-0.4657732844,-0.0059739975,0.0238205772,-0.0556560047,0.1521779895,0.0178643093,0.0192331262,-0.1478468776,-0.1118666157,-0.3519074619,0.0722220093,-0.0022423333,0.128591001,-0.3013212383,0.1177308634,0.3662571907,0.2721978724,-0.5800654292,0.0121660177,0.4983743429,-0.1462717354,-0.0521159023,0.4400008023,0.1853115708,-0.0742066875,-0.1611198336,0.2752127945,-0.2585220337,-0.224185884,-0.1401050389,-0.2096580118]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/353","title":"[Dataset requests] New datasets for Text Classification","comments":"On the Amazon Reviews dataset, the original UCSD website has noted these are now updated to include product reviews through 2018 -- actually quite recent compared to many other datasets. Almost certainly the largest NLP dataset out there with labels!\r\nhttps:\/\/jmcauley.ucsd.edu\/data\/amazon\/ \r\n\r\nAny chance someone has time to onboard this dataset in a HF way?\r\n\r\ncc @sshleifer ","body":"We are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705","comment_length":56,"text":"[Dataset requests] New datasets for Text Classification\nWe are missing a few datasets for Text Classification which is an important field.\r\n\r\nNamely, it would be really nice to add:\r\n- TREC-6 dataset (see here for instance: https:\/\/pytorchnlp.readthedocs.io\/en\/latest\/source\/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]**\r\n- Yelp-5\r\n- Movie review (Movie Review (MR) dataset [156]) **[done (same as rotten_tomatoes)]**\r\n- SST (Stanford Sentiment Treebank) **[include in glue]**\r\n- Multi-Perspective Question Answering (MPQA) dataset **[require authentication (indeed manual download)]**\r\n- Amazon. This is a popular corpus of product reviews collected from the Amazon website [159]. It contains labels for both binary classification and multi-class (5-class) classification\r\n- 20 Newsgroups. The 20 Newsgroups dataset **[done]**\r\n- Sogou News dataset **[done]**\r\n- Reuters news. The Reuters-21578 dataset [165] **[done]**\r\n- DBpedia. The DBpedia dataset [170]\r\n- Ohsumed. The Ohsumed collection [171] is a subset of the MEDLINE database\r\n- EUR-Lex. The EUR-Lex dataset\r\n- WOS. The Web Of Science (WOS) dataset **[done]**\r\n- PubMed. PubMed [173]\r\n- TREC-QA. TREC-QA\r\n- Quora. The Quora dataset [180]\r\n\r\nAll these datasets are cited in https:\/\/arxiv.org\/abs\/2004.03705\nOn the Amazon Reviews dataset, the original UCSD website has noted these are now updated to include product reviews through 2018 -- actually quite recent compared to many other datasets. Almost certainly the largest NLP dataset out there with labels!\r\nhttps:\/\/jmcauley.ucsd.edu\/data\/amazon\/ \r\n\r\nAny chance someone has time to onboard this dataset in a HF way?\r\n\r\ncc @sshleifer ","embeddings":[-0.0198358893,0.1645356119,-0.1285864413,0.1226923466,0.139930889,0.2225347757,0.1768303663,0.0815881714,0.0018178036,-0.1834153235,-0.2986063063,-0.0017472015,-0.2310984731,0.2553616464,0.1400381625,-0.2117386609,-0.0201150216,-0.0759108588,-0.1246431395,-0.216335699,-0.1675653309,0.2093166113,-0.0892350897,-0.2444806993,-0.3734586537,-0.1297220141,-0.18040739,0.1245356202,-0.2303669751,-0.5034515262,0.1528969705,0.2555025518,0.2345860153,0.3584259748,-0.0001047258,-0.3163300753,0.0129639581,-0.0361251011,-0.2524424791,-0.2285141349,0.0480050482,-0.4326225221,-0.1093345955,-0.0115563245,-0.0414830036,-0.0352776609,0.0633286834,-0.0734608024,-0.0728205293,0.2706288099,0.2498831302,-0.0968179628,-0.0141124018,-0.1115363166,0.4056470096,0.5342847109,-0.2694160342,-0.0557010733,0.5810772777,-0.0698628947,-0.1388198137,0.280479759,-0.2075151205,-0.152629301,0.3274661005,0.0953846574,-0.1789127141,-0.3495911062,-0.0063551958,0.3909525871,0.7599865198,-0.2851526737,-0.4934873283,-0.3570120037,0.1204494983,-0.244896546,-0.165517211,0.2134292126,-0.0270258021,0.0201331563,-0.174770847,-0.26001665,-0.2148726732,0.2961455286,0.3274449408,0.303799212,-0.0029598372,-0.226894781,0.0062388442,0.080479376,0.3732764423,-0.2820252776,0.1854224205,0.0842638984,-0.2543910444,-0.2538296282,0.0419816077,-0.0419491827,0.0082056439,0.0299477,-0.2754578888,0.0522342511,-0.2359441668,-0.0820202082,-0.0114379106,-0.104495652,0.1859629452,0.2776843011,-0.0064376257,-0.1476683617,0.1175474674,0.0522241853,-0.1956689805,0.1799935997,-0.1484035701,-0.1301525831,-0.3069113493,-0.3268981874,-0.0546195693,-0.0930696949,-0.0276658256,-0.0372009389,-0.1610669941,0.1006105542,-0.0305315424,0.1423638612,-0.0749849156,0.2043559253,-0.0643115491,-0.2759672999,-0.1067782044,-0.0132692596,-0.0978111774,0.1325537115,0.5589262247,0.0023388695,0.1266877204,-0.2371996194,-0.0930320919,0.1769449264,-0.1804188937,-0.0499144085,0.0144962082,0.1237010285,-0.0877870023,0.0635389313,0.0813061371,0.1484273523,-0.218797192,0.1358947605,0.012635963,-0.1809809655,-0.2063608319,0.2547236979,0.0151912319,-0.5899583697,-0.1809228063,0.564458847,-0.0478290059,-0.1022184938,0.1412489116,-0.0144039914,-0.1507670283,-0.0756463781,-0.1472384781,0.3579092622,-0.4638137221,-0.2868506014,-0.3036377728,0.1789581031,0.1972148418,0.1546200812,-0.1420154274,0.1414174885,0.0884599388,0.0173453223,0.1978496909,-0.0281157568,-0.2213232815,0.0716080293,-0.0373298861,-0.2037146091,0.1091809869,0.3472159505,0.1259894669,-0.1429607421,0.0136763956,0.1798772812,-0.1910176575,-0.0351377018,-0.223478213,-0.1538220942,0.1330546886,0.4792046547,0.5008529425,-0.1846703738,0.0178316738,-0.0566968396,0.1833344847,0.0148106888,0.1735473722,0.224555105,0.1589919776,0.3652983308,0.0200952068,-0.3260862827,-0.1582815647,-0.0368316062,0.1321993917,0.3238742054,0.0837286413,-0.1299978495,-0.1620239019,-0.0727357492,0.0639772788,-0.0884794369,0.1685641706,-0.182387054,0.0802148432,0.0695329905,-0.2511688173,0.2445216626,-0.3218910694,0.348934412,0.2296547443,0.3098436594,-0.0638302192,0.0823661089,0.2734518349,0.4851833582,-0.0561845377,-0.0047940561,0.0865122676,-0.050734818,-0.331142664,-0.0520032682,0.4012410045,0.340618819,0.2171980143,-0.4308609962,0.2395130098,-0.0908190012,0.0702354014,-0.2257074565,-0.1770806164,0.3826260567,-0.0643694401,0.1074867174,0.0163397603,0.0191110782,-0.0575710796,-0.0584995933,0.1567184329,0.1226610392,0.477445215,-0.2402418554,0.2156090438,-0.0560744368,-0.1531550139,0.310239166,0.2410908937,-0.0261276029,0.1495759785,0.3112009168,-0.3767238557,-0.1238452345,-0.0405461304,0.0679410473,0.2094611824,0.5444110632,-0.0162582882,-0.0429848023,-0.0633231923,-0.2928969264,0.0039459919,0.2457319349,0.1017832384,0.0940496102,0.3382118046,0.2782356739,-0.3023326993,-0.2575507164,0.1384813488,-0.0207988545,0.0421207435,-0.0394282117,-0.0457469262,-0.4623198807,-0.0133940717,-0.0140032191,-0.1270578057,-0.3637327254,0.3001460135,-0.0999475718,0.0934824646,0.0719564259,-0.3957877159,0.5337898731,-0.3382836282,-0.0312035251,-0.0873488486,-0.2671612501,-0.2508142889,0.2313467115,0.1402768046,-0.0088543231,0.4727497399,0.0387484021,0.23265405,-0.1457149535,-0.5765917897,-0.0559280142,-0.1253763139,-0.1209645122,-0.1876256019,-0.2154998481,0.047962781,-0.1673706174,-0.139753595,-0.120382674,-0.1475423276,-0.032984864,-0.2493180633,-0.1951202452,-0.0326809548,-0.5112283826,-0.3697476983,-0.1461063176,0.1025940999,0.0358285457,0.0933143869,0.2685100138,-0.2643089592,0.0148301283,-0.124264583,0.0666403845,-0.0374447927,-0.1886550784,0.1413713247,-0.1202241778,-0.2281951606,-0.1240064204,-0.0967828631,-0.0756764933,-0.1028150022,-0.6674728394,0.2082010508,-0.1985080093,0.1226012185,0.1556165367,-0.140558973,0.4066683352,0.2476305664,-0.1814626157,-0.0203112289,-0.1293238103,-0.0327682607,0.0750853196,0.3712266684,-0.3001052141,0.0803213045,0.2399805188,0.5578199029,0.0438777916,-0.035122633,0.1014531702,0.3303006589,0.2996345162,0.0231633168,-0.4512149692,0.286590904,-0.1416729689,0.2059862465,0.3342998624,0.2041311115,-0.4348141253,-0.1071692854,0.1427900344,-0.2731654644,-0.0399806313,0.2332309037,-0.2175328881,0.2240020633,0.1042257324,-0.1819122136,-0.3055625856,-0.5037221313,-0.1082317382,0.2369356751,0.173688814,-0.0743048415,-0.6273703575,0.2053293139,-0.3945928216,0.140062809,0.146317035,0.0674419701,-0.2749199867,0.2281356603,-0.082759425,0.2678698301,0.1284728944,-0.4005752504,-0.2270065248,0.295706898,0.1561955512,-0.0684424937,0.1542138457,-0.3768279552,0.0599434189,0.2813207209,0.0173195861,0.0865545496,-0.0921225622,0.4317032099,-0.0573148951,0.0095430138,-0.2893531919,0.3814591467,0.0226725359,-0.1425521523,-0.0904996619,-0.0130276717,0.1536349505,-0.2646354437,0.0852354392,0.1945631057,0.0428825244,0.3212280273,-0.416534096,-0.3929335773,0.2106627971,0.1838892549,0.0113947568,0.0309399087,0.2688947022,0.5306916833,-0.1769513786,-0.4539092779,0.036710836,-0.0552644916,0.4870544374,0.3452856243,0.4270687401,0.3030504882,-0.0884711593,-0.0477688015,-0.1185523495,0.1891394705,0.0697000548,-0.033224985,-0.5455816984,-0.2410629243,0.4245165586,-0.0938082486,-0.052428294,-0.0012594773,0.2747274935,-0.1869429648,0.2298601568,-0.0704729334,0.989127636,-0.0320348702,0.2314077765,-0.3351996243,-0.1499940455,0.5273062587,-0.3390073478,0.1651449651,-0.1366246343,-0.0622489862,-0.0773578882,0.0059612594,-0.1355396807,0.4320210218,0.2012410462,0.2437459528,0.3407815695,0.0423562713,0.1266358942,0.2039076835,0.0389912575,-0.506049037,0.0213233177,0.2347002923,-0.0364863835,0.366746366,-0.0041774474,-0.0057187025,-0.1408800334,0.1594284475,-0.7354986668,-0.0673653707,-0.1063631251,0.1266802102,0.0681761205,-0.3045451641,0.524651289,0.1131881922,0.4435447156,0.0865085423,-0.2198759913,-0.0272047892,-0.2468966842,-0.1631071866,0.2093583047,-0.1171966493,0.3968867958,-0.0255735368,-0.1978737265,0.0845589936,0.0637759566,-0.2301936001,-0.3948069811,-0.3200411797,0.1908973753,-0.2362110466,0.2588210702,0.3226072192,-0.0108893644,-0.3071137667,0.1961591393,0.1634349674,0.2605302036,0.0353921056,0.2692892849,-0.0035228578,-0.3461027145,0.219470486,0.1676374525,-0.1958144903,0.3862884939,0.5314847231,-0.2508363128,-0.2816473544,0.1999554187,0.1823787987,-0.3813153803,-0.1412218362,-0.0811863616,-0.060969051,0.2064898312,0.3244880736,0.384754777,-0.258020848,0.0251595955,-0.3177651763,-0.120753251,0.3003714085,-0.2548667789,0.054679729,-0.1722333133,0.2259543091,0.0648069531,0.3074397147,-0.3720640242,0.1942458153,0.0546494685,0.0094557758,0.1304036379,-0.0732256249,0.1413398832,-0.2162869573,0.0447176509,0.022611592,-0.0884638578,-0.2534921765,-0.094813928,0.1510846764,0.10359326,0.0487661883,-0.0773820207,0.0600363798,-0.1753800511,0.0145290131,-0.0607483722,0.09624888,0.12250714,0.4448112845,-0.145194754,0.0026813261,0.0493395291,0.0958275571,-0.1530480683,0.3397254646,0.2522175908,0.1069644541,0.3758492172,0.1319261491,-0.2536731362,0.2059230953,0.1825124025,-0.2143227309,0.6236365438,-0.3106362224,0.3434371054,0.4052039683,0.181194827,-0.1900020093,-0.038114775,0.0210202001,0.2060129344,0.2469580919,-0.3663802743,-0.2355708778,-0.0632307455,0.1820346415,0.0496905781,0.1599102467,0.325137496,0.0974356234,0.3078377545,0.1186231449,0.450309217,-0.1487881988,0.1708055586,-0.1072042286,0.1341331154,-0.0072139972,0.284710288,0.2316281646,0.0425444916,0.2432035059,0.1988579631,0.4407488704,-0.1563293487,0.1625998467,0.2973189056,-0.261161834,0.2581979632,0.2804408669,0.0506993309,0.186468035,-0.1658572704,0.341514498,-0.2284858078,-0.0036577419,-0.0194140561,0.0028682898,-0.0493478663,-0.0062726554,0.0534831285,-0.2535124421,-0.1139497384,-0.1281630546,-0.2739522457,-0.417527914,0.1384992748,-0.105855301,-0.1620160639,-0.7215718627,0.1422510743,0.1325696409,0.4418779314,-0.1659159958,-0.0574466251,0.0515807979,-0.1537546217,0.1801449955,0.4200522602,0.1736367345,-0.2005614936,-0.1233220547,0.0051050954,-0.2862403393,-0.1789830625,0.1752093285,0.3023533523,0.4547441602,0.0118456557,0.1499176025,0.1575321257,-0.0640638396,0.2059066445,0.1363329291,-0.0830258951,-0.1092215851,0.2585862577,0.1475930512,-0.2569752634,-0.1977559179,-0.012341694,-0.2303877324,-0.2367638499,0.4216784537,-0.1355620325,0.085024789,-0.1042156592,0.1350457221,-0.0826170817,0.3175186813,0.4432278574,-0.144182831,-0.116592072,-0.0728257969,-0.0408158973,-0.2165642977,0.0911086202,0.1053294837,-0.1427567899,0.0306990687,0.2910807729,0.1708277613,-0.3768725693,-0.1190773025,-0.0612900928,0.0696207285,-0.4864189327,0.010039472,0.1854389459,0.258515358,-0.0504090004,0.1169508323,-0.1581368148,-0.0232279189,0.0155783072,-0.3896496296,0.1958055496,0.2118862867,0.1357304305,0.0224187654,-0.0431522317,0.4692188203,0.012455578,-0.0873516947,-0.1607501954,-0.3508628607,-0.0878445581,0.2033870518,-0.0014765754,0.4487056732,-0.2618574202,-0.296900481,0.3471090496,-0.0083512543,-0.0122514917,-0.0826241374,-0.4484746158,0.0859550238,-0.1286575645,-0.0025240497,0.2685819864,0.1318809986,0.1355550736,-0.1389167458,-0.0911830664,-0.3851820529,0.1371689439,-0.0214328431,0.110930495,-0.137824595,0.1309030205,0.4573967159,0.1223885566,-0.6342869401,-0.0298267473,0.429725498,-0.1390914619,0.0752901733,0.4732365608,0.0454103351,-0.0811371356,-0.1597827673,0.2771237493,-0.3164700568,-0.2770523429,-0.2406642735,-0.1742233038]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"This is probably a Windows issue, we need to specify the encoding when `load_dataset()` reads the original CSV file.\r\nTry to find the `open()` statement called by `load_dataset()` and add an `encoding='utf-8'` parameter.\r\nSee issues #242 and #307 ","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":38,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\nThis is probably a Windows issue, we need to specify the encoding when `load_dataset()` reads the original CSV file.\r\nTry to find the `open()` statement called by `load_dataset()` and add an `encoding='utf-8'` parameter.\r\nSee issues #242 and #307 ","embeddings":[-0.2610889971,-0.0766245872,-0.0532211661,0.1464464962,0.3858148754,0.0102546979,0.0254466422,0.1763569266,-0.1791240722,0.0838485435,0.1790162325,0.4352126122,0.0588609129,0.3545265794,0.2403868884,0.0416257381,0.0944179073,0.3228261769,0.0393954292,0.0472039506,-0.3697523773,0.1955367774,-0.103214398,0.0522328466,-0.2589356601,-0.0607879423,0.0105512701,0.2736924589,-0.2823358774,-0.5056849122,0.2017236203,-0.0907642022,0.2071509212,0.2249621302,-0.0001033256,-0.0700257793,0.2572823763,-0.2035377771,-0.0729974657,-0.3186079264,-0.1721763462,-0.3807736039,0.0590986423,-0.2513725162,0.0154255684,0.2983249426,-0.0178563762,-0.069501169,0.4933331013,0.3671520352,0.3225337267,0.1151859388,-0.120438993,0.1490378827,-0.0000738069,0.0481821895,-0.0301628429,0.4893274605,0.1714179367,-0.2337498516,0.0854414403,0.15753223,-0.2183983922,-0.1051619127,0.0754632354,0.1099671572,-0.1662448347,-0.3456841409,0.1217203811,0.0935375169,0.2599755526,-0.2827702761,-0.0374360569,0.1049030572,0.1600743979,-0.6936118007,0.2831137776,0.3683131933,-0.2686055005,0.134733662,0.013303794,0.1663535088,0.0272235144,-0.0592778325,0.0735416785,-0.0005281873,-0.1928128749,-0.0872904509,0.2068818808,0.0006389527,0.129370451,0.0012708894,-0.1845574528,0.2111000866,-0.2259348184,-0.0672173575,-0.1538407356,0.0041102082,-0.1760222912,-0.1566052586,0.5614205003,0.1108473465,0.047019057,0.3558768332,0.0998752788,0.1248829663,-0.2047813535,0.129919678,0.253900528,-0.0032394202,0.2112276554,0.0027607603,-0.1976271719,-0.3183389008,-0.2264608443,0.3368732929,0.1330850273,-0.167297408,-0.4322226942,0.1432407051,0.1057569832,0.1413569152,0.0213067234,0.2131028175,0.1955601424,0.2922651768,0.0155842183,0.2012343407,-0.349791944,-0.2578395903,-0.046728801,0.2218290865,-0.3768264055,-0.1314049214,0.2358993292,-0.1387218237,0.1109037548,0.2852515578,0.141297698,-0.183926031,-0.0799183846,-0.1549814343,0.0537384972,0.1382840127,0.0759721398,0.1385126859,0.1670151353,-0.3159835935,-0.0821788833,0.1015482843,-0.4818613827,-0.0225306526,-0.058851216,0.2757769227,-0.0652818531,-0.0694927648,-0.5188581347,0.1689957231,0.0898341611,-0.3107695878,0.0260849893,-0.2124463171,-0.1478360891,-0.1852401197,0.1550226808,0.5013543963,-0.558544457,0.1694433838,0.2123724222,0.0155501338,0.3184648752,0.1244617924,-0.0968644843,-0.1430572569,-0.1716031879,-0.0217708685,0.2493201792,-0.2651790679,-0.222624585,0.3089697361,-0.0901828632,-0.058612667,0.0460300893,-0.1065662131,0.1471290737,-0.1886292547,0.1144837588,0.3725188673,-0.0196821876,0.2809734643,-0.219192341,-0.1849874258,0.2239281535,-0.0044941646,-0.2701480985,-0.1170027182,0.2115180343,0.0669501349,0.1973384023,-0.1416933239,-0.2056812495,0.1531885713,0.2411946952,-0.2049695402,-0.1022406146,0.2412946075,0.0358477309,-0.0403329097,-0.256057173,0.163705647,-0.3069273829,-0.1502541304,-0.1704839915,0.0351581611,-0.3680139184,-0.2693879902,0.2428944856,-0.0371168219,-0.0683988929,-0.0123575265,-0.0955900922,0.2467948198,-0.013406449,-0.0126051661,-0.227121979,0.1789416075,-0.2131493539,-0.1868838519,0.3183470666,0.295512557,0.1668151915,-0.1177361012,-0.2245182097,0.2406990677,0.0372983851,0.363714993,-0.2209492177,-0.2943725288,0.4554791451,-0.4603476822,0.0675635487,0.4531856179,0.2957178056,-0.0873812363,0.22945638,0.3592467606,-0.1594866365,-0.0167384148,-0.0725952908,-0.0614305548,0.257991612,0.0848891437,0.1084826887,-0.1574765146,0.3592934906,0.6213696003,0.2019231766,-0.0257057976,-0.1232932657,0.0141802179,0.4981156886,-0.0518534034,-0.1169799641,0.0897622854,0.0535767749,-0.1114095151,0.0894828662,0.0712491646,0.227318868,-0.1056673229,-0.3125161827,0.2457751483,0.0206145719,-0.2307857871,0.1905181408,-0.1344037801,-0.0368113667,-0.0156292282,0.0608470812,0.1106140465,-0.2907760441,-0.128715992,0.1040016934,0.177634269,-0.3210408688,0.3218948245,-0.4479729831,-0.378326416,-0.3087700605,0.2326155007,-0.143322289,0.0219256971,-0.1595176905,0.0173547585,0.0631201416,0.1683906466,-0.2043867111,0.0400572196,-0.0683101192,-0.1754354537,0.1225209385,-0.2182811052,-0.2620295882,0.1559137553,0.428724438,-0.0599661134,0.0835664794,-0.3489724398,0.1415617019,0.2766743898,-0.146533072,0.1601233035,-0.2612069845,0.1343978047,-0.0371876061,0.0921340734,-0.061995592,-0.4000653028,0.1540004611,-0.0559684895,-0.0773286819,0.1851528734,0.0075843544,-0.3202289343,-0.2613763809,-0.1141118556,-0.4982915223,-0.3325833976,-0.1811806411,0.0014529042,0.1016743407,0.3449436724,0.136863023,0.0497554876,0.2142150104,-0.1142487302,-0.0944051892,-0.2219955176,0.3507091105,-0.3976939619,-0.5242287517,0.0896824673,-0.0443613939,0.4002031386,-0.2082488239,-0.185305953,-0.0273388159,-0.1131972522,0.1191595569,0.0449815802,0.1042199135,-0.0041818526,-0.0227979105,-0.0459144637,-0.1186295375,-0.1014546603,-0.1292091757,0.09624511,0.3029592037,0.0399042889,0.2478981167,-0.2661068738,0.09984795,0.2606476843,0.0474691428,0.5908392668,-0.1422195286,0.3743231297,-0.0816609934,-0.1248044372,0.1246356294,-0.2733381391,-0.0562205017,0.2765930295,-0.1166036204,-0.2867608964,-0.3135625422,0.0396993831,-0.3513276279,-0.2961669266,0.188545838,0.0594999418,0.0969607681,0.024850348,0.1353806257,0.0701625347,-0.3088001609,-0.1110788882,0.2989730537,-0.1462356746,0.1520843804,-0.1521614939,-0.4839813113,-0.2414459437,0.1134493351,0.1711219996,0.4359165728,-0.0367386751,0.1763122082,0.079497084,-0.1191367805,0.21652098,0.1351017803,0.043507725,0.1071514636,0.1585657895,0.1158135459,-0.1062518582,-0.2047526687,0.1765428334,0.1213096604,-0.2626975477,-0.0457274914,0.0551604405,0.0229958426,0.3702946305,-0.1329976767,0.1060533524,-0.4633558095,-0.5534041524,-0.024614118,-0.1616851538,0.1898548901,0.293400526,-0.2198519707,0.0093207536,0.1202004477,-0.2522866726,0.2437074631,0.4230842888,0.374091506,0.1527099311,-0.0220923573,-0.1854198426,-0.0722190514,0.18729496,0.7113944292,0.0337085947,-0.3642597198,0.1105943322,-0.2761214375,0.316373229,0.1974186599,-0.1727222353,-0.0785625353,-0.128143847,-0.0657642856,-0.2221494019,0.0830025449,0.2763551474,0.1747066379,-0.5603710413,-0.4825350642,0.4688642621,-0.0914611146,0.1933649033,0.0365706868,-0.0631642416,-0.3073396087,0.2559743226,-0.0114889704,0.8470351696,0.2099158913,0.1276678443,0.2932154834,0.0173725318,0.290672332,-0.0832253247,0.203803122,-0.2967996299,-0.1601109505,-0.0126132714,0.1855671406,0.240424335,0.0636227876,-0.3559832275,0.2242915928,0.0648612157,-0.1683230698,0.1615427434,0.1692884564,-0.1898385286,-0.2559829354,-0.5498685837,0.2420231253,-0.074174583,0.0597805083,-0.0901403055,-0.2030752152,0.0201646928,-0.3251199424,-0.2977225184,0.2888044417,-0.4078468382,0.166845575,-0.2021711767,-0.6580182314,0.2853273153,0.1242898032,0.198546499,0.2002895772,-0.0216458663,0.1231300011,0.2777646184,-0.0227611214,-0.0133756464,0.119110696,0.0687000975,-0.164490357,-0.0619397685,0.1669941992,0.108495906,-0.0052970308,0.4046760201,0.0390396751,-0.2194951475,-0.1117109656,-0.2296898216,-0.2004286051,0.1841035038,-0.0592126921,0.1926203668,-0.0688742697,-0.1279043257,0.0036830842,0.0680626109,-0.3047739267,-0.1466597468,0.0651419386,-0.0677976683,-0.1524512768,0.4950759113,0.1653570682,-0.2378005236,-0.1186074764,0.2235974818,-0.1604558527,-0.4836042821,-0.0424497463,0.1999964714,-0.0120628439,0.0410431288,0.63182652,-0.0175027661,0.0172414742,0.1021786109,-0.4966746271,-0.2612887919,-0.068494685,-0.1012652144,-0.0446722582,0.2102424204,0.3430010974,0.1056419462,-0.1439816803,-0.3483979106,0.0469593145,-0.0635712966,0.0744443461,0.2907143533,-0.0819433182,0.1172337681,0.0607281923,0.1660142094,0.3002019823,-0.0275974199,-0.2418089509,-0.1697691977,0.0561316311,-0.0356497951,-0.3160914779,-0.0820622221,0.0505351573,-0.0692952573,-0.0936086699,0.1222847402,0.0730746388,0.0730864704,0.1747053415,-0.3870049119,0.1315183938,-0.0197950564,0.2080585957,-0.2156207412,0.5619537234,0.1797867268,0.1604511738,0.0498262309,0.0527412742,-0.1532720029,-0.0252486002,-0.3163553178,0.0578348637,0.1277320981,-0.3855437636,0.3905307055,0.1264442205,-0.008264157,0.5075657964,-0.1379269809,-0.0211760383,0.4258249998,0.2020479441,-0.3272441626,0.105744943,-0.3296947181,-0.0640655234,-0.0720487759,0.2804112434,0.2723197937,-0.1586751491,0.2167009711,-0.1451956332,0.4394041896,0.1370041519,0.3914925456,0.4240065217,0.1557471752,0.2737905681,-0.0957097337,0.1282919496,-0.04705685,0.0887832567,-0.5585888028,0.0297569372,0.0904516503,0.177356109,-0.1411350518,-0.3880759776,-0.0429100096,-0.1291218102,-0.198298052,0.3351866901,0.1471820623,0.5732957125,-0.4751986861,-0.053410437,-0.1215258241,0.1643474847,-0.1561279297,0.059683647,0.077623412,0.0138266142,-0.2852063477,0.1848021299,-0.0392283387,0.1242767498,-0.0761307031,0.3221030831,-0.1820150912,-0.0505408756,-0.0110672507,-0.0040788697,0.0636589453,-0.0879494995,0.4701707363,0.2787474394,0.1963555813,0.239633292,0.0966278687,0.6247934699,0.4012522399,-0.1397327483,-0.0370995738,-0.1724214703,0.0626664758,0.1204679757,0.0972917601,0.3171797395,0.4335707128,0.1459421664,0.2074578255,-0.135806486,-0.0845307335,0.2828327417,-0.0246578269,0.0637780726,0.0951789618,-0.0357815176,0.2239776105,-0.5465280414,-0.0746000782,-0.3710962534,0.285451889,0.334453553,0.0766331404,0.0316530205,-0.1154678836,0.1219319031,0.1146439165,0.4839324653,0.2948755622,0.0673658475,-0.1934751123,-0.0274676811,-0.7056211233,0.2235512435,0.0908994079,-0.3779625893,-0.2408954203,-0.2864883542,0.0177811272,0.3394007385,-0.0403151996,0.2757698596,0.0841895193,0.0604777858,-0.0808862448,0.0141544426,-0.1719048023,0.0116177415,-0.2442377508,-0.2519827783,0.0690807104,-0.2601279318,0.1166035011,-0.2869591415,-0.2666415274,-0.4350647926,0.0025400259,0.277205348,0.1945083886,0.4332052469,-0.2569182515,-0.1383413821,-0.1196714714,-0.1758017093,-0.0627516881,0.3558363318,0.2394197881,0.0617621653,0.0547001921,0.1469726861,-0.3888507485,0.3169057071,-0.0380781032,-0.1657897234,-0.2426064014,0.0363164768,-0.0004940127,-0.0887324214,0.203325972,0.3809025586,-0.1002787575,0.1084835827,-0.1204256266,-0.4743161201,0.1404096186,-0.5154396892,-0.1619222313,0.1350958347,0.237483725,-0.177898407,0.1382843405,-0.4251002371,0.2577645183,-0.0297689792,-0.1882052571,-0.281365186,0.2817209363,-0.3067510426,-0.1085519642,-0.0804569051,0.4227133989,0.2139987797,-0.1878929436,0.1064991578,-0.3179031312]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"It should be in `xtreme.py:L755`:\r\n```python\r\n if self.config.name == \"tydiqa\" or self.config.name.startswith(\"MLQA\") or self.config.name == \"SQuAD\":\r\n with open(filepath) as f:\r\n data = json.load(f)\r\n```\r\n\r\nCould you try to add the encoding parameter:\r\n```python\r\nopen(filepath, encoding='utf-8')\r\n```","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":36,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\nIt should be in `xtreme.py:L755`:\r\n```python\r\n if self.config.name == \"tydiqa\" or self.config.name.startswith(\"MLQA\") or self.config.name == \"SQuAD\":\r\n with open(filepath) as f:\r\n data = json.load(f)\r\n```\r\n\r\nCould you try to add the encoding parameter:\r\n```python\r\nopen(filepath, encoding='utf-8')\r\n```","embeddings":[-0.269592315,-0.0288400799,-0.020871589,0.2028595954,0.4484202564,-0.1219135448,-0.1402623653,0.2741995752,-0.3068618476,0.1308181882,0.1182492003,0.5972364545,0.0107578505,0.3725733459,0.1493888646,0.0074850386,-0.0501957126,0.2807819247,0.1142940968,0.0160581674,-0.3124450147,0.1804717332,-0.1091933325,-0.0322549529,-0.3568943143,-0.0902507678,0.1276563406,0.278445214,-0.4236949384,-0.5119258165,0.1833940446,-0.0129814195,0.2230469882,0.2653470933,-0.0001058442,-0.0454376265,0.3183930814,-0.2982024848,-0.2216451466,-0.3611499667,-0.2516929507,-0.315095216,0.1989216208,-0.2637838125,-0.1006912366,0.2404016256,0.0199444219,-0.0952382758,0.5233158469,0.2909530699,0.2963303328,0.073832728,-0.0744115859,0.1486849338,-0.0282303728,0.1482578516,-0.1475578398,0.4451701045,0.2602996826,-0.0911355689,0.1426788718,0.2766979337,-0.1515211314,-0.0569933876,0.0407141112,0.0499193668,-0.151903525,-0.4266541302,0.1209061518,0.1055289283,0.1793570071,-0.3225930333,-0.0961947814,0.0870558769,0.1963281035,-0.5246263146,0.2534211278,0.3471793532,-0.2969875336,0.097451359,-0.0773089752,0.1050549448,0.0375204943,-0.0349980071,0.0779615492,0.210955441,-0.1810681522,-0.0667629689,0.1364755481,-0.0218385831,0.2110842615,0.0128953513,-0.2235882878,0.179078266,-0.1587533057,-0.1336388737,-0.1817007214,-0.0875805095,-0.087361142,-0.1382653415,0.5145932436,0.1206673607,-0.0586921424,0.3560672104,0.1107803285,0.1049097627,-0.0984211862,0.2221195102,0.1833802611,0.1279567331,0.09569408,0.0296438895,-0.2141443342,-0.1327173412,-0.3393443823,0.2891263664,0.1441908479,-0.1674303114,-0.3648921549,0.0851829275,-0.0286666937,0.1087728441,0.1422024965,0.2070315331,0.1510671526,0.2321667224,0.0610188432,0.2638306618,-0.3025706112,-0.2582405806,-0.0983013734,0.2584832609,-0.3669599593,-0.1151665226,0.1970708221,-0.0907112062,0.1982972622,0.1363954842,0.1843987405,-0.2621587217,0.0463314913,-0.1218954474,0.0610194989,0.0838399231,0.0885400698,0.035077408,0.1756173223,-0.3857584,-0.095351465,0.1032874957,-0.4930075705,0.0364094265,-0.1123503819,0.2696914375,-0.1408186704,-0.0545914955,-0.5896954536,0.1581823975,0.0557486117,-0.2681747675,0.0798110068,-0.1571032405,-0.0741815716,-0.1766550988,0.2183379233,0.4230662286,-0.4035671949,0.0333889574,0.2349449694,0.0550833866,0.3388772011,0.0612149462,-0.2030248791,-0.1410146654,-0.1047271565,-0.0357994325,0.3881499767,-0.4512214661,-0.2139129043,0.2725942433,-0.0509270504,-0.0216157157,-0.0301322732,-0.131116271,0.0554815792,-0.1026321203,0.1379778534,0.4275479913,0.0005008268,0.3331618607,-0.2610883713,-0.2735882103,0.1134488732,0.1089548767,-0.2255675197,-0.1487599313,0.1748792529,0.3316148221,0.2449594587,-0.1301338673,-0.245599404,0.1132457927,0.2360798419,-0.3353266418,-0.2025874555,0.230512172,0.0102625033,-0.0392397977,-0.2957889438,0.2397301346,-0.3415217996,-0.0709602088,-0.135705635,0.0705726892,-0.4842479229,-0.2870798409,0.2603209317,0.0040702247,-0.1470945925,-0.0489315167,-0.1445099413,0.3022420108,-0.047584746,0.0106714582,-0.2947007418,0.2131462991,-0.1876621097,-0.281137526,0.3004021347,0.2748621702,0.1912814528,-0.1671635062,-0.2761870623,0.172657907,0.0572981164,0.3439613581,-0.1737114638,-0.2655003071,0.4478878975,-0.4799709916,0.05142143,0.4397865236,0.2158050537,-0.1009260118,0.2564740777,0.3604639471,-0.0498562455,0.0057903202,-0.0060213162,-0.1606789827,0.1911538541,0.0216744412,0.0419140495,-0.141872257,0.2428831607,0.6975977421,0.2468798906,-0.0303255673,-0.1135328561,0.0522243008,0.5576232076,0.0134540731,-0.0485008955,0.0479931496,0.1208856702,-0.1739585549,0.123891376,0.0689123943,0.1667799205,-0.0529841036,-0.2737358212,0.2346289754,0.0185396355,-0.1578697264,0.1978085488,-0.0990896896,0.0686605126,0.0721914247,0.1992562711,0.0100442003,-0.3035995662,-0.2533633411,0.042458754,0.1445677429,-0.2297809273,0.3444653749,-0.4669023752,-0.402338773,-0.2357724309,0.1430747062,-0.1867650598,-0.0578529201,-0.120883368,-0.0568394773,0.15535146,0.2601627409,-0.1332838088,-0.0113901822,-0.1319015324,-0.206618309,0.0643789545,-0.2219242901,-0.2030799836,0.1332081258,0.4137871265,-0.0475224443,0.0261726025,-0.3408566713,0.0964660197,0.3336159289,-0.2299674153,0.3083393872,-0.2643299699,0.1596699506,-0.0279841069,0.1091463119,-0.0951455683,-0.3874913454,0.2449334264,-0.0245360136,-0.1276833713,0.0561904758,-0.1040130556,-0.2081224173,-0.2407755554,0.0273787212,-0.484936595,-0.3486636579,-0.2008547336,0.1295459569,0.1333154887,0.3928813636,0.2247519791,0.1689454764,0.2177811414,-0.0570540391,-0.1314014345,-0.381948173,0.2595172524,-0.3200514317,-0.335532397,0.0427410342,-0.0658447221,0.5374211073,-0.2214785963,-0.1797289848,-0.1533397585,-0.0088255834,0.1046936288,-0.0650105104,0.1340196282,0.0049341419,-0.0113704633,-0.0147210769,-0.1544592083,-0.139175117,-0.0744661093,0.1369945258,0.4062225819,-0.0833959356,0.2225062847,-0.226818487,0.0787741393,0.2252549678,0.0042761029,0.6793308854,-0.0974558964,0.3227324486,-0.1281467974,-0.0877529681,0.1625579149,-0.1000215188,-0.0727064759,0.1974796653,-0.0449011624,-0.2861955464,-0.3524032831,0.0150639219,-0.340998888,-0.2340870351,0.090501003,0.2625516653,-0.039737016,-0.0230138265,0.1832149625,0.0406130292,-0.1779970229,-0.123878099,0.2872675955,-0.1525914222,0.2216819674,-0.2870139182,-0.5363566875,-0.1644163877,0.1338132173,0.2102372348,0.2769465148,0.044546131,0.1823502183,0.1378421038,-0.1647197157,0.171429947,0.0734708756,0.0071996991,0.0798160359,0.1017655879,0.0262048803,-0.1008871123,-0.1671182215,0.1911314875,0.1716031879,-0.188401565,-0.1574798524,-0.0023053086,0.0705848485,0.2752684951,-0.0750973374,0.1587522328,-0.4937387705,-0.6795377135,-0.1820587367,-0.2335925549,0.2208952755,0.3266308606,-0.0434838049,-0.0135291703,0.1251825243,-0.2238433361,0.2270705253,0.5767014027,0.4269607663,0.2597108781,-0.0815525055,-0.1813802123,-0.085073784,0.2476549894,0.7725036144,0.0393482521,-0.2483740598,0.1788683087,-0.0974324569,0.288115561,0.2831268609,-0.0874727741,-0.1002378017,-0.1817708164,-0.0565207303,-0.1019537896,0.2562603056,0.1494083852,0.0205734223,-0.5256541371,-0.4884115756,0.560664475,-0.054915864,0.1898398548,0.1073905155,-0.0492880978,-0.3160758317,0.2909820676,0.0824171752,0.9118980169,0.2074112743,0.1326513439,0.3785893023,-0.0078509748,0.2663113773,-0.1112296283,0.2233990878,-0.2986947,-0.0614563264,0.050710801,0.1321149766,0.1290602982,0.0739663914,-0.3411415517,0.1588944197,-0.0305064954,-0.3309117258,0.12979725,0.3077484965,-0.2023123801,-0.2501312494,-0.7101199031,0.225706771,-0.2092167586,0.0350707658,-0.0871648565,-0.2560060322,-0.0722238421,-0.2950129807,-0.207840547,0.376031965,-0.5171034336,0.1067209691,-0.062963821,-0.5599789619,0.2495436221,0.1234673932,0.104408294,0.2184451222,-0.0160450302,0.2406456172,0.2833400369,-0.0059492975,0.0110538425,0.1758532822,0.0805312842,-0.2479119897,-0.0770086646,0.2005444467,0.1386253089,-0.0912182555,0.3390100598,0.0428849235,-0.2078883499,-0.1063856706,-0.3097418249,-0.130573824,0.1471561641,-0.0394900329,0.1691490561,-0.0256108176,-0.2011843175,-0.0777621791,0.2628116608,-0.3546227515,-0.1116666794,0.1045368388,-0.0236604828,-0.0551133379,0.4700589776,0.1100138128,-0.2028773725,-0.1128229722,0.2733691633,-0.1000829414,-0.4406729341,-0.0156564694,0.1384795457,-0.1954837143,0.0086420951,0.7061377168,-0.0132929003,0.0516712144,0.2228101343,-0.3965664506,-0.3372613192,0.0339794643,-0.144214794,-0.0090895547,0.1709855795,0.4224148691,0.0119778821,-0.0714295432,-0.3345586061,0.0772370473,-0.1851819456,0.0922182202,0.0920181721,-0.1470727175,0.1361893415,0.1042295173,0.1643655002,0.3269123733,-0.1279890239,-0.221996665,-0.1826176494,0.0694143027,-0.0000199425,-0.3460960388,-0.2182556838,0.0524591468,-0.0732843131,-0.0780802891,0.1275210381,0.0299434271,0.1278790832,0.1545260847,-0.2598352432,0.0735444129,0.0710724369,0.1250919849,-0.0021256786,0.576377809,0.2222473919,0.1196842566,0.0221084114,0.0370835811,-0.0727366358,-0.0311005544,-0.2551571727,0.2031828016,0.086775355,-0.3685713708,0.3687850237,0.0811621919,-0.0743523017,0.4766412973,-0.1920268387,-0.0511043034,0.2678045332,0.1945385188,-0.2686937749,0.0565888658,-0.182387352,-0.0601923987,-0.0239729676,0.4054143727,0.1480820626,-0.0305430423,0.2371189594,-0.0708750784,0.436930716,0.2265251428,0.3792157769,0.2138878107,-0.0676764324,0.1707872599,-0.0843179449,0.110603027,-0.0269387271,-0.076130867,-0.5743137002,0.1385021061,0.075140886,0.0983586088,0.0161980912,-0.1613383889,-0.1337962896,-0.0820500478,-0.0703252181,0.3402709961,0.0914386511,0.4065175056,-0.4054417312,-0.0660917535,-0.008206564,0.1307263374,-0.1202393323,0.048768688,0.0025693439,-0.0446656644,-0.386964798,0.1005105153,-0.1038219556,0.0603701621,-0.1292275339,0.346491605,-0.1597101539,0.0560174175,0.029041959,0.0874943882,0.0793066323,-0.1478644758,0.6503372192,0.2597216368,0.1448372155,0.195076555,0.1092399284,0.5741664171,0.3779676855,-0.182083562,-0.0452345386,-0.0910710767,0.012746851,0.0713319257,0.0215482954,0.4151414633,0.2848604321,0.2318213731,0.2036958039,-0.163689509,-0.1569397151,0.2618773282,-0.0566187799,0.1092533916,0.2053094506,-0.0973205045,0.2734782398,-0.4711074233,-0.0992673561,-0.4386304021,0.3083442152,0.2360984683,-0.0013732816,0.0319079049,-0.2150591165,0.1161063239,-0.0403568186,0.5322924256,0.2266666293,-0.0033688187,-0.156912595,-0.0840735286,-0.5758488774,0.2599777579,0.0649033487,-0.446488291,-0.1634184718,-0.1827462316,0.0211066045,0.4211796522,-0.0096592717,0.3094619215,0.1508059055,0.0013071881,-0.1990279257,-0.0340944901,-0.1309002787,0.0468435474,-0.1482460946,-0.2397922426,-0.0290272161,-0.3242877722,0.1147612408,-0.2961993217,-0.4041757584,-0.4243032038,0.0271788985,0.2851333618,0.1705962718,0.3588993251,-0.3229241073,-0.234201625,-0.0174814742,-0.2376617938,-0.1061405241,0.3435553908,0.2130287439,0.3197680712,0.1029945984,0.2732108235,-0.4937636256,0.3357635736,-0.0837132037,-0.2716353834,-0.1880092174,0.0018076045,-0.0851684511,-0.0208526794,0.0997845307,0.3712540865,-0.0338048488,0.1574396193,-0.1849863231,-0.4357159138,0.3454655409,-0.4225984514,-0.1698672324,0.2028681785,0.1223302186,-0.1553902477,0.0797984526,-0.5110164285,0.2554682195,0.0197685584,-0.122792691,-0.278162986,0.295969069,-0.2903804481,-0.0786039084,-0.0585518926,0.3128242791,0.239962697,-0.1206459105,-0.0135893524,-0.2334342748]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"Hello @jerryIsHere :) Did it work ?\r\nIf so we may change the dataset script to force the utf-8 encoding","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":20,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\nHello @jerryIsHere :) Did it work ?\r\nIf so we may change the dataset script to force the utf-8 encoding","embeddings":[-0.3731933236,-0.0007190405,-0.0570683368,0.1422875971,0.4632261395,-0.0583365299,0.0105651617,0.2293293774,-0.2868659198,0.0433460809,0.1535942852,0.4569696486,-0.0138362274,0.3835915327,0.266928196,0.0252128504,0.0164343826,0.2814874053,-0.1030595526,-0.0219441876,-0.2794499397,0.1553264111,-0.0629759505,0.0027551295,-0.3882107735,-0.0570778549,0.0714768097,0.2772664726,-0.3569349647,-0.5312875509,0.2094703764,0.0535364896,0.1863969564,0.2738594711,-0.0001022789,-0.0029409204,0.329123199,-0.2302227765,-0.1595393419,-0.2611712813,-0.2483115345,-0.3530781269,0.1106154472,-0.2077700198,-0.0237200428,0.1612143517,0.0368489735,-0.1297168732,0.4614186883,0.2716950476,0.3374944031,0.1603048146,-0.0636833683,0.1194714382,-0.0223568417,0.1525055915,-0.0523618981,0.34658283,0.2546724677,-0.1929524541,0.0855830535,0.2532845736,-0.1368109435,-0.0948411673,0.0521471277,0.0395273119,-0.1605703384,-0.3848971725,0.1053259075,0.1600777805,0.2978846431,-0.3879599571,-0.1183259711,0.0882200897,0.1210183427,-0.4723842442,0.2392136306,0.3041261435,-0.2488500476,0.1334116906,-0.0709755272,0.1129439101,0.029688431,-0.0367267355,0.0147210117,0.0923566073,-0.1295583695,-0.0976859555,0.124841176,-0.049868688,0.1877581924,0.050129585,-0.2015790194,0.1952289492,-0.1748973727,-0.090685524,-0.1389023364,0.0374198519,-0.0177331455,-0.1268707961,0.4794391692,0.1832569838,-0.0675333962,0.3135921359,0.1453900337,0.0608680323,-0.1526663303,0.2708898485,0.2193612009,0.0321703851,0.1343903393,0.0710698888,-0.169864729,-0.2104030102,-0.2478932291,0.2668915987,0.1573391259,-0.1860109717,-0.411269635,0.0974360853,0.1417272538,0.0561990514,0.0682383701,0.2181813866,0.2298815995,0.3139407039,-0.029175777,0.2334385216,-0.355664432,-0.3158363998,-0.1523568034,0.2118054628,-0.3692724109,-0.1307553053,0.1776291281,-0.0606790558,0.1488868892,0.1713613123,0.1450042725,-0.114101775,0.0414618813,-0.1093736142,0.1005498767,0.1388428956,0.0439043827,0.0962276682,0.1858623326,-0.2839111686,-0.0842195228,0.065156728,-0.4052479267,-0.0274737738,-0.1326574981,0.3021100163,-0.0976724103,-0.0723736212,-0.5015494823,0.1596663594,0.1064520031,-0.2595107853,0.0770499334,-0.1923130304,-0.0640770346,-0.1431927979,0.2095197439,0.4781813323,-0.4952636957,0.0663281828,0.1981609017,-0.0513669401,0.3271408975,0.0646894053,-0.1527037621,-0.0950650647,-0.1846716851,0.0148836588,0.347771585,-0.3756852448,-0.2681895792,0.3038841784,-0.0430336483,-0.0552403368,0.0085136248,-0.0133375116,0.0574759953,-0.1772714555,0.0595130399,0.4527754188,-0.0160307027,0.2917616069,-0.2091446519,-0.2561106682,0.192169711,0.1395067424,-0.2073940039,-0.152781263,0.2245241106,0.2628910244,0.2572357059,-0.0965588316,-0.2111480832,0.2039387524,0.1723736227,-0.2741017938,-0.164906621,0.1530279815,-0.0532297455,-0.0128101874,-0.2910028696,0.221300751,-0.3516929746,-0.1563200057,-0.1589294374,0.0381986052,-0.4941663444,-0.2556962371,0.2810660601,-0.0190961864,-0.1264655739,0.0135039287,-0.15965271,0.2963830829,0.0199439097,-0.0196307618,-0.3017684817,0.2150720656,-0.1520977616,-0.2504400015,0.3062167466,0.2735082209,0.0892287642,-0.1297211349,-0.1804229915,0.2587802112,0.0460719243,0.3964953423,-0.2296985537,-0.2436213046,0.4404168129,-0.5327318907,0.0991478413,0.3941129148,0.2343506515,-0.15958637,0.2340748012,0.2926603854,-0.1530442536,-0.0467362441,0.0212418362,-0.0999416783,0.2121109366,0.0230979156,0.0503354706,-0.2419478297,0.3193282187,0.5829188228,0.3353452086,-0.0260665752,-0.1082596779,0.0674530193,0.515915215,-0.03269976,-0.0043930314,0.0391603783,0.0645848066,-0.1366206855,0.0882593542,0.1023404524,0.193241477,-0.0185558517,-0.2185662687,0.2224379927,-0.0169931725,-0.1787496507,0.1433495283,-0.1276643872,0.0318578966,0.034728758,0.1080371886,-0.007887152,-0.3578631878,-0.1810830534,0.0537049361,0.1668539941,-0.2967945635,0.2733753622,-0.511202693,-0.4090254903,-0.186880514,0.1079191864,-0.1811732352,-0.0947801545,-0.076821126,-0.0853769407,0.1153330207,0.2908653617,-0.1969730258,-0.061572358,-0.0830110088,-0.159324795,-0.020455217,-0.1876849979,-0.2183427066,0.1528368294,0.4614593685,0.0059018852,0.1112049073,-0.3916853368,0.07324972,0.194764629,-0.2173335999,0.2696352005,-0.2560012639,0.1230480969,0.0718326122,0.1163184494,0.0312645026,-0.2786531746,0.2485301793,-0.0708810836,-0.1150765345,0.1409763098,-0.0682281181,-0.2105594575,-0.2368371934,0.0126492986,-0.5228254795,-0.3507963717,-0.2645406127,0.1185957789,0.1428292692,0.3499618769,0.1977186352,0.0584549792,0.1639474779,-0.1200960875,-0.1110019684,-0.3482830524,0.3166835308,-0.3320203125,-0.4280412197,0.0193366781,-0.0038295251,0.4887190759,-0.1189449355,-0.2341216505,-0.092367582,-0.1230493709,0.1430981457,-0.0310135782,0.147395432,-0.0255248211,0.0119263092,-0.0425533466,-0.1464657933,0.0079843262,-0.0698680952,0.0022528658,0.2842376232,-0.0301314592,0.2277228534,-0.2086023688,0.0189590212,0.2235666364,-0.022464877,0.7035977244,-0.1742078811,0.3842787147,-0.1911888719,-0.0875153914,0.1026371196,-0.0767122954,-0.0156148849,0.1611462981,-0.0796989799,-0.2145682126,-0.273891747,0.0530565754,-0.2512015402,-0.2426974773,0.1096379384,0.115219973,0.0127795609,-0.0002069459,0.1531667411,0.088237755,-0.2258845419,-0.0663614869,0.1648574769,-0.159569636,0.2148155719,-0.2964556515,-0.2983582616,-0.2577537298,0.1107346639,0.1485255361,0.335616678,-0.0790966377,0.153194651,0.116830267,-0.1291458905,0.2168400437,0.0291321911,0.1142597273,0.087906003,0.0901457369,-0.018889267,-0.0758940205,-0.1978045255,0.1719638556,0.0955977291,-0.1960682869,-0.1699277759,-0.0016315884,0.0589859076,0.3772258162,-0.0978037044,0.1263487935,-0.4287360013,-0.6874220967,-0.1177395806,-0.2442572117,0.2344142795,0.2473376393,-0.0949963406,-0.0444278419,0.0066079097,-0.2473978549,0.2189928889,0.4687686861,0.4177284241,0.1295853704,0.0223175772,-0.1359939575,-0.1060856357,0.2562142909,0.7766752243,0.0220115408,-0.280893892,0.0966924801,-0.2284713238,0.2586398721,0.2258514613,-0.0765503868,-0.1581783891,-0.0957818106,-0.0640959963,-0.2174804807,0.2045222223,0.2207326889,0.1283140332,-0.6074514389,-0.5158417821,0.4817233086,-0.0257370099,0.2049150169,0.123177886,-0.0749501884,-0.3552894592,0.2622034848,0.0351322442,0.8382989764,0.1588293016,0.22848171,0.3226373792,0.1386994421,0.1874375492,-0.0317839049,0.1091980487,-0.3560633063,-0.1851942688,-0.0195192192,0.1509481519,0.1931738704,0.1352781206,-0.3118436038,0.1862616986,-0.0538868681,-0.2737147808,0.151650995,0.269806087,-0.2737750113,-0.2056007087,-0.6117542386,0.2700067461,-0.1186788231,0.0272534769,-0.1151670292,-0.2042444944,0.0142220827,-0.3317554593,-0.2252712995,0.334813565,-0.4587022364,0.2300195843,-0.1353653967,-0.6156854033,0.1628254652,0.1720420122,0.1507769227,0.1937804818,-0.0101661477,0.2571097314,0.255905807,0.0317181088,0.0555355027,0.1400502473,0.1517919302,-0.1858286262,-0.0883148983,0.1521257311,0.1377554387,-0.0620557107,0.2877087593,-0.006257067,-0.1587270498,-0.1283807307,-0.2757997513,-0.1705020964,0.1848901212,-0.0551957041,0.2012414634,-0.0062445672,-0.1744205356,-0.0054243044,0.1873109788,-0.3181982934,-0.1325163096,0.0828512013,-0.0555685312,-0.0788573846,0.5007184148,0.2100293189,-0.2192791253,-0.1669026315,0.2493761927,-0.001798215,-0.5435650945,-0.0312808454,0.1454030424,-0.0885960981,-0.0008253477,0.6154595613,-0.0781905055,-0.0378502347,0.1249645725,-0.4164334834,-0.3383079767,-0.0199523512,-0.1058946848,0.0769013613,0.1353712231,0.3030590415,0.0179048441,-0.0751930699,-0.3758338988,0.057842534,-0.1599833667,0.0512880795,0.1627022624,-0.0597326048,0.1141208112,0.0571744181,0.2042425424,0.2957517505,-0.0587748699,-0.2655416131,-0.1633339524,0.0448211394,-0.0447309539,-0.3438979685,-0.0863279104,0.0497450382,-0.1354933232,-0.0814273432,0.1621017009,0.1329279095,0.0965037122,0.1838114113,-0.229092598,0.0540944524,0.032515917,0.1618752182,-0.1049631536,0.4907989502,0.1859702468,0.1443757117,0.0478410386,-0.0010487342,-0.0478375666,0.023732638,-0.2163306475,0.0968612283,0.0982679501,-0.3974511027,0.4032672644,0.1052452177,0.0381632969,0.482917577,-0.2567420304,-0.1113973334,0.3376916647,0.2441076785,-0.3323169053,0.022885194,-0.2274404317,-0.0486905314,-0.027072547,0.3549534082,0.175457418,-0.0744633302,0.2037746608,-0.0859549046,0.4425880313,0.100548774,0.366566062,0.3080146611,-0.049370449,0.1541154534,-0.1038606316,0.1590823084,-0.0920451805,0.1089132354,-0.5868619084,0.1726316661,0.0507146567,0.1186177954,-0.0504694246,-0.2833313942,-0.0698191971,-0.0359478407,-0.1264662296,0.3150508404,0.1728378236,0.4308012724,-0.339723289,-0.0370755605,-0.0746368021,0.1877142936,-0.1154494286,0.0538828857,0.0222675689,-0.0218787845,-0.3208213449,0.1186904907,-0.120840922,0.0519776568,-0.0135728978,0.3341334462,-0.1179917902,-0.0927926451,0.0595224351,0.005821527,-0.0208273511,-0.1202606484,0.5656138659,0.3377194703,0.1767883152,0.3259630203,0.1647764593,0.6169279218,0.4158925116,-0.1283584088,0.0329091772,-0.1991829127,0.0448215343,0.0093143424,0.0949316472,0.2698283494,0.3797510564,0.2434946001,0.23330459,-0.1668172777,-0.197440505,0.2925789952,0.0156132067,0.087832734,0.2161625028,-0.0928954855,0.2102701068,-0.5015550256,-0.0489588343,-0.434252888,0.2403140068,0.4158198535,0.020653028,0.0559104346,-0.2033473849,0.1347759068,0.0010748045,0.5677646995,0.2840283811,0.0830740705,-0.0933835357,-0.0814676285,-0.7156070471,0.2470430732,0.0399615094,-0.4754844904,-0.1407996565,-0.2607809007,0.1011030525,0.3434419036,0.0972545221,0.3271721303,0.0704674497,0.0694175065,-0.1378182769,-0.0190064553,-0.1851951331,0.1184407696,-0.2628828287,-0.3089424968,0.0403945073,-0.2737837136,0.157704711,-0.3079627752,-0.2911491394,-0.4004510641,0.0246143602,0.2664486468,0.2022701502,0.408361733,-0.2999082506,-0.204060629,-0.1089320183,-0.2557532489,-0.1500159502,0.3890317678,0.2026680857,0.1393884867,0.0956534445,0.1532764882,-0.4704892635,0.3255023658,-0.0776246488,-0.0981988683,-0.2244172841,0.0946472585,0.0099899536,-0.0435834937,0.1534400284,0.3403096795,0.0001789649,0.1427140683,-0.1330498159,-0.482403785,0.2574142516,-0.4593118429,-0.2252649069,0.1388773769,0.1024020836,-0.1715398729,0.0529758707,-0.502163887,0.2745053768,0.0210056826,-0.1494607627,-0.2724819779,0.3049333692,-0.2336805314,-0.0641323105,-0.0872094929,0.3601798117,0.178939268,-0.173619926,-0.0067087403,-0.2403552979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"@lhoestq sorry for being that late, I found 4 copy of xtreme.py. I did the changes as what has been told to all of them.\r\nThe problem is not solved","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":30,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\n@lhoestq sorry for being that late, I found 4 copy of xtreme.py. I did the changes as what has been told to all of them.\r\nThe problem is not solved","embeddings":[-0.2841983736,-0.1710122824,-0.0102975909,0.2463469803,0.4728092849,-0.0588771589,-0.1426550001,0.2119477093,-0.2602452338,0.1263464093,0.089154087,0.5027424693,-0.0454213433,0.429033041,0.2850128412,0.013322426,0.0035831209,0.2751812339,-0.0077735465,0.0021374247,-0.2719275057,0.1288376153,-0.1685154438,0.0082540326,-0.4515329897,0.0142073883,0.0879136324,0.2322887033,-0.3210153282,-0.6467185616,0.2900666296,-0.0919905528,0.2375106514,0.3235117793,-0.0001081838,-0.0482052565,0.3104091287,-0.2778413296,-0.0658099726,-0.2856352329,-0.2337793857,-0.3430391252,0.101397112,-0.2250879407,-0.0894931257,0.2258188576,-0.0545104966,-0.1139665097,0.4601177573,0.2725270689,0.2595206499,0.1287803799,-0.0217991527,0.0607361123,0.0604668036,0.1154146492,-0.090522632,0.4867531657,0.3114893436,-0.1195533574,0.1933287829,0.2543049455,-0.1568315923,-0.133475557,0.125910148,-0.0169644523,0.0053897742,-0.4357089698,0.0417733826,0.2204738408,0.1514525414,-0.3889326453,-0.0276959762,0.1341760457,0.1199851707,-0.5919212699,0.2534968257,0.3236535192,-0.2409827858,0.122527197,-0.0020337789,0.1651620567,0.0704384893,-0.1112985164,0.0444017649,0.1325530857,-0.2128743827,-0.0954208598,0.1789368838,0.0546290763,0.2867621779,0.0904347524,-0.2125891447,0.1940508038,-0.3190436959,-0.0655719638,-0.1915603131,-0.0587911159,-0.0675602034,-0.1182293817,0.4792559147,0.0995845497,0.0326140933,0.2687477767,0.1060628891,0.0777690187,-0.1224479824,0.2007592916,0.1557708085,0.0766420811,0.1295703799,0.0229571834,-0.2260320783,-0.1936638504,-0.2553225458,0.3502626717,0.1382745057,-0.2279368937,-0.435672462,0.0755720437,0.0437256806,0.1526127011,0.1308053136,0.2008614242,0.2930817604,0.3658074737,0.0010363258,0.3005890846,-0.319924891,-0.3942802846,-0.1051794812,0.3204549253,-0.4212428033,-0.1009720042,0.135539785,-0.1225451827,0.1391242296,0.2384171933,0.2263340205,-0.3725573719,0.0250783376,-0.1242956519,-0.0242590159,0.1046297476,0.0223297384,0.0954796895,0.1946078539,-0.2149739414,-0.095124267,0.0950684175,-0.5300650597,-0.0877708048,0.0242652074,0.2465445399,-0.0689616352,0.0110309822,-0.6024036407,0.0946132466,0.1463538259,-0.2219099253,0.1035979092,-0.2148195356,0.0085423989,-0.1859185398,0.2275683284,0.432982713,-0.4388881326,0.061635498,0.1866447628,0.0247416794,0.3251600862,-0.0308978539,-0.2035872191,-0.0952485949,-0.1479416788,-0.0822999328,0.336143285,-0.4542464614,-0.2818059623,0.2109288871,-0.0842543989,-0.006146031,0.0338435583,-0.0713115484,0.0243238676,-0.2501995564,0.0208876785,0.3124954104,0.0346341506,0.2238523364,-0.2625687122,-0.255695343,0.1888916343,0.022490548,-0.2124091983,-0.2279863209,0.1698408872,0.2444564998,0.2650094926,-0.0716834217,-0.2540249825,0.2266713232,0.1636253744,-0.2654485106,-0.1530476213,0.1478769779,0.0975921303,-0.0085667716,-0.4412302375,0.1540108025,-0.1883801818,-0.1166941449,-0.1519980431,0.0756862387,-0.5332611799,-0.2902416587,0.225946486,-0.0512886345,-0.1036307216,-0.0128428033,-0.1843506545,0.2484761029,-0.0960384905,-0.0238135438,-0.322747618,0.1692887843,-0.1479879618,-0.2545021176,0.3124353588,0.2807561159,0.2038686723,-0.188059926,-0.3029399812,0.2458579242,0.0449902639,0.3558740914,-0.3245793879,-0.2304282486,0.4206211567,-0.6723706722,0.146245122,0.4327425957,0.2300490141,-0.0863217786,0.3771478534,0.2802092135,0.0099499812,-0.0249159969,-0.0998435691,-0.1123463511,0.135006249,0.0347357765,0.149747178,-0.1532124579,0.3822408915,0.677051127,0.2885441482,-0.0659379065,-0.0031925647,0.0499922782,0.4282445014,-0.0313407183,0.009942499,0.0277372506,0.011520464,-0.1405825019,0.0482859537,0.0674979016,0.157757923,-0.0938323066,-0.3045616746,0.27883178,0.0654535219,-0.1355648935,0.1910290122,-0.1055879071,-0.0699119642,0.0142461183,0.0924220234,0.0405950993,-0.1860477328,-0.1129704267,0.0957059041,0.148297146,-0.3401213288,0.2751016915,-0.4626467824,-0.3754618168,-0.1783538014,0.2071150392,-0.1286837161,-0.0076481444,-0.1607840657,-0.1041119099,0.1516872495,0.284309268,-0.1858735085,-0.0620627627,-0.1134632751,-0.1641200632,0.1164784878,-0.1277104765,-0.2553988695,0.0823456049,0.4200480878,0.0251667332,0.0639339313,-0.3856783211,0.1357679069,0.3003232479,-0.1755381525,0.2783754766,-0.2725768387,0.156186372,0.0260643121,0.0388391837,-0.0439628772,-0.221233353,0.2425484657,-0.1499065608,-0.004079158,0.140884012,-0.0309933666,-0.2017132193,-0.1660741568,-0.0085946694,-0.514534831,-0.3367132545,-0.3198547065,0.0043239407,0.1315589249,0.2935812771,0.1360127479,0.1235325933,0.145314455,-0.1120187864,-0.0669114739,-0.4351952672,0.2787834704,-0.2855579555,-0.382735014,0.0567738861,-0.0542963892,0.5393147469,-0.1494328082,-0.1855553985,-0.0595885925,-0.0793232843,0.0459489711,-0.003210966,0.1624264121,-0.0027852738,-0.0708893687,0.0175639,-0.1337160766,0.0254941545,-0.0617818907,0.0362773314,0.3804641068,0.0331328586,0.1845371872,-0.2000236213,0.0022228998,0.2838160992,0.075142853,0.7266833186,-0.1042949632,0.3343746066,-0.1572365761,-0.0879113749,0.031177057,-0.1539879143,0.0081962729,0.0513677746,-0.0408372171,-0.1235364974,-0.3065727353,-0.0719730258,-0.175637573,-0.2540784776,0.0641207024,0.1964412183,0.1248775795,-0.0756762102,0.1900192946,0.1153945252,-0.2332873195,0.0341154188,0.3092287183,-0.1363596469,0.2283605188,-0.3142040968,-0.4231190085,-0.1252245307,0.1707861573,0.216989547,0.3219676912,0.0325643308,0.1843996644,0.1240589097,-0.1191463768,0.110594824,0.1513098329,0.1104796156,0.1094779223,0.0363932177,0.0851013735,-0.1229368076,-0.1389254183,0.0926551223,0.0919786915,-0.30649212,-0.1430082023,-0.0295896009,0.0537065454,0.4073210359,-0.0948460624,0.1092241183,-0.4352787137,-0.6628590822,-0.1123477519,-0.2616870403,0.2751927972,0.2804517448,-0.093049325,-0.0211579856,0.0791032016,-0.250554502,0.3101434112,0.427491188,0.5339711905,0.0862649456,-0.0194543526,-0.1762205809,-0.2426553369,0.2761381865,0.6962288022,-0.0227933545,-0.2288333178,0.1570194662,-0.1308867186,0.2720611393,0.2541864216,-0.1872227341,-0.0783882365,-0.2135688514,-0.1060027257,-0.1555956304,0.2363244742,0.2061840892,0.0780973732,-0.6163372993,-0.5784109235,0.6196987033,0.0039472622,0.1832074076,0.1434502155,-0.0224706084,-0.3126707673,0.1387595981,0.0617132075,0.9569510221,0.2100054771,0.1338701844,0.3063666224,0.093584761,0.140322879,-0.1279684305,0.2171009779,-0.3326469362,-0.2194079757,0.0028503982,0.0785792768,0.148081407,-0.0010918227,-0.2661407888,0.2339378446,-0.0966951251,-0.3270173967,0.1215312183,0.1934319288,-0.2453078777,-0.1436849236,-0.607085526,0.214099288,-0.1180020794,0.0181742702,-0.1240861341,-0.2341250628,-0.0462035313,-0.3025847673,-0.1978775561,0.3659680784,-0.4864293933,0.1834530979,-0.1319803298,-0.6856116056,0.2966009974,-0.0068874941,0.0750277191,0.2151509076,0.0087230876,0.2325280905,0.3472604454,-0.0084367897,0.1008885577,0.1394307464,0.0247577671,-0.1841704845,-0.0433435403,0.2549691498,0.116174005,-0.038815394,0.2637769282,-0.0586362556,-0.1236658543,-0.1759617329,-0.2906427085,-0.2101520598,0.2039057016,-0.0643695965,0.158063814,0.0173814166,-0.2333697528,-0.0397105515,0.1144282296,-0.318439573,-0.1402510107,0.0468606576,0.0103987623,-0.0828837752,0.4898578227,0.2013557553,-0.149177134,-0.0912685245,0.3218546808,0.0427366868,-0.5850658417,0.0545306988,0.1139190421,-0.1064218879,0.0597924776,0.7707697749,-0.0121221151,-0.0816623867,0.1693049222,-0.3554033637,-0.3297325373,-0.0021987639,-0.1268319339,0.0535094664,0.1177991703,0.5053521991,0.0303915273,-0.0944211259,-0.3083256483,0.0938748717,-0.0797493458,-0.0014513729,0.1456511319,-0.0619983673,0.1321708262,0.1597167701,0.1371439248,0.3783659041,-0.1036975607,-0.1838033497,-0.1208277643,0.0895378888,-0.0762759373,-0.3509170115,-0.1385745555,0.0530447364,-0.0840035006,-0.0165410135,0.1091456041,0.0537907742,0.0979353115,0.1658830047,-0.3130356669,-0.0078537753,-0.017828295,0.2861248851,-0.0905454531,0.4674896896,0.1526813805,0.0884235725,0.0615560301,0.0038008254,-0.192420736,-0.0162332635,-0.2317231148,0.0429790914,0.1086519882,-0.4080076814,0.3844197989,0.166202873,-0.0784979239,0.511318326,-0.2503395677,-0.0002710425,0.2907701731,0.1652014256,-0.3032809794,0.098556228,-0.2944170833,0.010941281,-0.1226773113,0.428916961,0.2969359457,-0.1355918497,0.2463413775,-0.0419785567,0.465747118,0.1641955078,0.4294066727,0.3339241445,0.1176095754,0.1835429668,-0.1204459742,0.1563709229,-0.042437952,-0.0288286563,-0.6253312826,0.1836440414,0.0830867365,0.0889217257,-0.0404972062,-0.282809943,-0.1842344701,-0.0523002371,-0.1952906847,0.3339410424,0.0554026105,0.4945180416,-0.4045983851,-0.0543086715,-0.0903536528,0.1913446933,-0.1062500924,0.1158473566,0.1751669496,-0.0295301005,-0.3728174567,0.0510005206,-0.0478915647,0.1516232491,-0.098498404,0.3364896774,-0.1681473702,0.0062477239,0.0209681168,0.1837598234,0.0745998472,-0.1302243471,0.5655001402,0.38682881,0.1632657945,0.265109688,0.1187305376,0.5911539197,0.424575597,-0.1783138067,-0.0057317526,-0.0481445678,0.070298329,0.198710531,0.0354103446,0.2802945673,0.4603325427,0.18720074,0.1704154611,-0.1349004209,-0.2581371665,0.1844439805,0.024944853,-0.0015242482,0.2883338034,-0.0720066428,0.2568310499,-0.5061605573,-0.1153296083,-0.4890584946,0.2965939343,0.4165177643,0.0372217298,-0.0408974476,-0.2159250379,0.0996102169,-0.0284941252,0.4844785929,0.247202903,0.0303105377,-0.2621785402,-0.1003774926,-0.6561942101,0.2546974421,0.0283436142,-0.4872553051,-0.1307214946,-0.3052963018,-0.0057944581,0.3557794988,0.1714761555,0.3289460242,0.0851049945,0.0911223441,-0.120682463,-0.0237271804,-0.2309011668,0.0887989402,-0.15668571,-0.2004085034,0.0185226705,-0.2647876143,0.0986330509,-0.2584551871,-0.3411832154,-0.2736749649,0.0481207669,0.3000489175,0.135330528,0.3911925554,-0.183463037,-0.2070967108,-0.0924072042,-0.17412135,-0.1101753861,0.2745246291,0.2171794921,0.1259855777,0.0533538423,0.1425209045,-0.5296491385,0.3337309659,-0.0269119199,-0.2261540145,-0.2503288388,0.042420797,-0.0791867226,-0.0027168246,0.2878602445,0.3772039711,0.0333362482,0.2308592945,-0.197436288,-0.4907915592,0.2464108765,-0.418490231,-0.1275347918,0.0717024505,0.2167004198,-0.1613464952,0.0805051252,-0.5184621215,0.2307827175,0.032079827,-0.0563874282,-0.210431546,0.3002867699,-0.2732470036,-0.0560418442,-0.0295712631,0.3224092126,0.1971317977,-0.1095011756,0.0736241341,-0.241511479]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"Could you provide a better error message so that we can make sure it comes from the opening of the `tydiqa`'s json files ?\r\n","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":24,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\nCould you provide a better error message so that we can make sure it comes from the opening of the `tydiqa`'s json files ?\r\n","embeddings":[-0.2817687988,-0.0771224797,-0.0259544104,0.1864886135,0.431878835,-0.1063414365,-0.0490796752,0.2102413774,-0.2888760865,0.0789318606,0.169837907,0.5185136795,-0.0157341827,0.4039361477,0.2165159136,-0.0277956389,-0.0087143779,0.3307390511,-0.0386377312,0.0391564257,-0.3182015419,0.180805847,-0.062818177,0.0498370789,-0.3340267837,-0.1292742044,0.1371211559,0.3347364366,-0.3895091414,-0.5701818466,0.2702987194,-0.0080769332,0.170627743,0.3130139709,-0.0001040535,0.0268084146,0.3452884555,-0.3027345538,-0.104040131,-0.2532044947,-0.2504182756,-0.3094664514,0.1723178774,-0.1982563883,-0.0262208078,0.2419230789,-0.0710227042,-0.1036925316,0.4995187521,0.2284307033,0.2823840976,0.1334480047,0.0270902142,0.1191968992,0.0230809283,0.2401787043,-0.0793570653,0.4386374652,0.2769225538,-0.1497905105,0.2019462287,0.2887742221,-0.1177519113,-0.0590951592,0.0971384794,0.0272609908,-0.1220089495,-0.386395514,0.0478322245,0.1814053655,0.3376056552,-0.3324699104,-0.0737317801,0.1352531314,0.1053181142,-0.4322821498,0.3052654266,0.3276412189,-0.3073068261,0.0905286595,-0.0683731213,0.138113156,0.0039155181,-0.0165487751,0.036680311,0.0251126923,-0.1426296532,-0.0910675451,0.171558246,-0.1108821109,0.1577507257,0.0657360405,-0.2696868479,0.2028368711,-0.135870412,-0.1643541753,-0.1352392137,-0.1784209907,-0.0712781101,-0.1799986809,0.6284134984,0.0887388811,-0.0453395173,0.3436106443,0.2137916833,0.077939406,-0.1613688618,0.2298027128,0.1764535606,0.1067804992,0.135150075,0.0063298065,-0.1947845817,-0.2007736117,-0.3806664348,0.261467278,0.2055311203,-0.1674138606,-0.3905965984,0.1570671499,-0.0251287259,0.1251740307,0.1215743199,0.2305425704,0.2509425282,0.2337392271,0.1034999639,0.301407367,-0.2691760659,-0.2696082592,-0.0902691707,0.2240213007,-0.3366367221,-0.1306651384,0.1875256002,0.0248261392,0.1507218182,0.1681637764,0.1875312179,-0.1822350174,0.0831004009,-0.0185790062,0.0666431636,0.1819321811,0.01842333,0.0592943951,0.1957460344,-0.3177241683,-0.1496910155,0.0478967987,-0.4498771727,-0.1029236764,-0.0498548374,0.2764468193,-0.0709629059,0.0148943821,-0.5774091482,0.103324227,0.1036705449,-0.3065210879,0.1104103252,-0.1724876761,0.0652475953,-0.1789132655,0.1632585973,0.4239064157,-0.5196294785,0.061409194,0.2629779577,-0.008728506,0.3078123629,0.0395209603,-0.2188350558,0.001408648,-0.1103211194,-0.0673660189,0.3427746892,-0.3594678044,-0.2083869427,0.2648138702,-0.0599900931,-0.0064671785,-0.0183021985,-0.0459113009,-0.0296961311,-0.1674282551,0.1635740548,0.4073846638,0.0208942853,0.331974268,-0.1896002144,-0.2001384199,0.1352940947,0.1335037798,-0.3661079109,-0.1996991634,0.1727053821,0.3472834229,0.2326287776,-0.1098583341,-0.2995851934,0.1958723962,0.1872804165,-0.2911566794,-0.1688216776,0.2727039456,-0.0299199149,-0.0008161278,-0.46689412,0.1937803179,-0.3488879502,-0.125282824,-0.101116389,0.1086002886,-0.4706578851,-0.2607012987,0.2740674317,-0.0427687578,-0.0989649072,0.0541113913,-0.1565451622,0.1738281846,0.0019520853,-0.0128229782,-0.3405911624,0.1497582346,-0.136209771,-0.2563149035,0.3186994195,0.2369572818,0.1342215836,-0.1979389042,-0.2061989456,0.2130203396,0.0390104018,0.3725278676,-0.2939714491,-0.249104172,0.4123367071,-0.5082938671,0.0771567151,0.4808596075,0.2183993459,-0.1461538523,0.2769643962,0.2817266583,-0.0891663954,0.0151476813,-0.0292358566,-0.1517893523,0.1802475899,0.0448530503,0.0785501152,-0.1457226872,0.3540159166,0.6760111451,0.2317332625,-0.0618653074,-0.0612974912,0.063147597,0.4979129434,-0.0228534881,0.0224453695,0.0306606852,0.0818222836,-0.1498587877,0.0854414925,0.0991884395,0.1967508644,-0.0726002753,-0.2659802437,0.1621042192,-0.0040254625,-0.176297009,0.1893495917,-0.133513093,-0.0095683141,0.0479486361,0.1350293905,-0.0119345337,-0.2644740641,-0.2243680209,0.0472704321,0.1501175314,-0.2828509212,0.2840691507,-0.469281286,-0.3395822942,-0.1189216748,0.0577480309,-0.1735459566,-0.0477437265,-0.0287636388,-0.0875506178,0.02008464,0.2474021167,-0.1192872673,-0.0076294728,-0.0873216093,-0.1683951616,-0.0243215002,-0.2341928482,-0.2052877247,0.1376561821,0.4320602119,-0.0060566664,0.0468511246,-0.4066323936,0.1306010187,0.3167106807,-0.188171193,0.2786333859,-0.2370859981,0.083637476,0.0733186156,0.1783911884,0.0843783543,-0.3423646688,0.2534838915,0.0106194774,-0.1843847781,0.1638585776,-0.0660434067,-0.1703947484,-0.2623527944,0.0312266294,-0.4781665206,-0.3827649355,-0.1456872225,0.1230577156,0.1545903385,0.2949495018,0.2861062586,0.118511036,0.1604100317,-0.0210369192,-0.0711759627,-0.3474283814,0.2403917313,-0.3694214523,-0.3852407336,0.0362640545,-0.0017279785,0.5735808611,-0.2455367446,-0.2102106065,-0.012437663,-0.097946465,0.0303754266,-0.1026786491,0.0817887112,0.0242636185,0.0206769407,-0.0392615274,-0.1288281977,-0.0239885319,-0.0752912387,-0.028744407,0.3422817886,-0.0258510169,0.2903343439,-0.2431191504,0.0379476249,0.2327430695,-0.0770645812,0.656467855,-0.0829391703,0.3366741538,-0.2007137984,-0.0529074147,0.083165288,-0.1412670314,-0.0648583025,0.1083849818,-0.0754474327,-0.2820371091,-0.3335742652,0.0537846424,-0.2663478255,-0.2430244684,0.0889939889,0.1951134205,-0.0806271136,-0.0360126868,0.1178681999,0.0675627515,-0.1302367598,-0.0453021899,0.2454382479,-0.1497395039,0.2119579017,-0.2705279589,-0.4818193018,-0.1639059484,0.1904546469,0.1681249142,0.2538117766,-0.026094282,0.0711108297,0.1143475473,-0.1675317734,0.2137321234,0.1218517572,0.0823778138,0.125548318,0.0904084593,0.0639796108,-0.0746584907,-0.2222863734,0.1211321875,0.1885791868,-0.2162660956,-0.2339185476,0.004644366,-0.0255858861,0.3330483139,-0.0405117385,0.1393743306,-0.4960223734,-0.6708529592,-0.2037331462,-0.2221938968,0.2060863823,0.2809029222,-0.0897955447,-0.0868966058,0.0504809059,-0.2398332059,0.188161239,0.5552309752,0.4350998998,0.1240930557,-0.0356483161,-0.2201029956,-0.101619117,0.3710555732,0.8140745163,0.1218465343,-0.2139300406,0.1597454101,-0.106150791,0.1949416101,0.2183979899,-0.0612570457,-0.1229310408,-0.2011158168,-0.0471099056,-0.1970367879,0.2360365093,0.2261994779,0.0052687125,-0.5391126871,-0.4374729991,0.5951070786,-0.1041430905,0.1791022569,0.1939283907,-0.0497260131,-0.2902439535,0.2526358962,0.0939044207,0.8820834756,0.2124817222,0.1971960664,0.3946861327,0.1023965031,0.2337594628,-0.1697785258,0.1595804691,-0.294819802,-0.056724783,-0.0166515615,0.071566999,0.1229254529,0.1062177047,-0.3144182861,0.0424727164,-0.0723961368,-0.3561918437,0.059395086,0.2124865651,-0.2269318998,-0.2082309723,-0.6637165546,0.2391317636,-0.1073242798,0.0407227948,-0.1362830102,-0.2645180523,-0.0510923304,-0.3845811784,-0.1963122189,0.3531443179,-0.4750059843,0.1848690808,-0.0700014234,-0.6309326887,0.185944438,0.0950208828,0.1329978555,0.1886107028,-0.0468606204,0.2326347083,0.2573446333,-0.0418463759,-0.0256652404,0.1446772814,0.0865650997,-0.2290065587,-0.046131514,0.2251648754,0.0787734389,-0.0925494805,0.3337953985,0.0486081205,-0.2959442735,-0.150269568,-0.3046675324,-0.1649678499,0.1636933833,-0.0122374985,0.1836998314,0.0649413913,-0.2007926255,-0.1235959679,0.1584541202,-0.3354058862,-0.1280152649,-0.0002243076,-0.0564101562,-0.0518178158,0.469266057,0.1955725551,-0.1823511273,-0.1286049187,0.3528130054,0.0209728871,-0.5445841551,0.0216621235,0.2507894039,-0.2140463293,0.0216922574,0.7245564461,-0.0386081338,-0.0383135267,0.197974354,-0.3346596658,-0.4100528359,0.1211121529,-0.176541999,0.1115836054,0.1292643547,0.3700279891,-0.0438421629,-0.0467106141,-0.340513587,0.0951456428,-0.1919039935,-0.0264839362,0.118927449,-0.0848551095,0.1138947234,0.1564535648,0.155579865,0.3289302886,-0.1472267658,-0.2335826606,-0.1009584665,0.0682492331,-0.0222878307,-0.2691427767,-0.1357531846,0.0717206597,-0.0849232897,-0.0229936447,0.1489899158,-0.0507919788,0.1142263934,0.1481943429,-0.228745237,0.0628196374,0.0075677405,0.1854738295,-0.1000136882,0.5562202334,0.1064177454,0.0211145282,-0.0779824704,-0.0006730952,-0.1246824935,0.0693604425,-0.232816264,0.109967716,0.1433181167,-0.3258744776,0.3833560944,-0.0472308397,-0.0225433037,0.482129544,-0.2561000288,-0.0375840031,0.2865125537,0.2163231075,-0.255805701,-0.0219416693,-0.2864889503,-0.0266394224,-0.0596191175,0.3174259365,0.0948875993,-0.0968195796,0.2679116428,-0.0562099367,0.4560812116,0.1287513822,0.3024080694,0.3055929244,-0.1382286549,0.1300306767,-0.1319562495,0.1824097186,0.0500269383,-0.0265692491,-0.6331772208,0.0941193774,0.0548142344,0.0723895654,-0.0161439572,-0.2497695982,-0.2015666068,-0.0490618423,-0.0829798058,0.3896677494,0.1720814854,0.3763697445,-0.4089490473,-0.0280420743,-0.1358761191,0.1379606873,-0.0701369941,0.0554983616,-0.0409296416,-0.0583390556,-0.2832136452,0.085040018,-0.0863703266,0.1195398644,-0.0830630139,0.2690292299,-0.1116135716,-0.0280173253,0.0860885978,0.0169620421,-0.0092138713,-0.1316576004,0.5656224489,0.3985664845,0.1152795404,0.2964440882,0.1866954267,0.6348474622,0.4427382648,-0.2009814531,0.0176079888,-0.2134805918,0.0193561018,0.0494363904,0.1225811839,0.3237188458,0.2292325199,0.2328833789,0.2080927789,-0.1593469083,-0.2050779015,0.2338088006,0.0916536525,0.1269600093,0.2348296046,-0.2353520989,0.2718701363,-0.4492451549,-0.087882705,-0.4653194547,0.2442933321,0.2965211272,0.0449356288,0.0403482839,-0.2311996222,0.1190267354,-0.0631499961,0.6006459594,0.1987868845,0.1034514755,-0.1107354537,-0.0290537532,-0.6104630232,0.2797623873,-0.0068355072,-0.4517880678,-0.1234366968,-0.2283339947,0.131938681,0.4181392491,0.0715558007,0.3477075398,0.0696175247,0.0989482254,-0.163103357,-0.0020972488,-0.2136660218,0.0870614424,-0.1896215379,-0.2382355183,0.025643073,-0.3435182273,0.1415512264,-0.2849485874,-0.2998342514,-0.3997810483,0.1126060933,0.3344656229,0.2275514901,0.3273837864,-0.2716763914,-0.2994087934,-0.0551989004,-0.285425365,-0.0677866638,0.3536012769,0.209764272,0.2202964872,0.0422313102,0.2010894418,-0.4806059301,0.2990707457,-0.1135342792,-0.1927038282,-0.2164245695,0.0568376891,-0.0278045535,-0.0177574325,0.1971734017,0.3319278359,0.0158205237,0.15810965,-0.173114568,-0.446483314,0.3486944735,-0.4384197593,-0.1625081301,0.2954719365,0.0785425678,-0.2457388043,0.1090268642,-0.5252718925,0.271455884,0.0139727741,-0.0921182483,-0.2960338593,0.2218853086,-0.2321225703,-0.1918736845,-0.0756713897,0.4190500081,0.1878013313,-0.078854993,-0.0127318939,-0.2181255966]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"@lhoestq \r\nThe error message is same as before:\r\nException has occurred: UnicodeDecodeError\r\n'cp950' codec can't decode byte 0xe2 in position 111: illegal multibyte sequence\r\n File \"D:\\python\\test\\test.py\", line 3, in <module>\r\n dataset = load_dataset('xtreme', 'tydiqa')\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/87748794-7c216880-c829-11ea-94f0-7caeacb4d865.png)\r\n\r\nI said that I found 4 copy of xtreme.py and add the \u300c, encoding='utf-8'\u300d parameter to the open() function\r\nthese python script was found under this directory\r\nC:\\Users\\USER\\AppData\\Local\\Programs\\Python\\Python37\\Lib\\site-packages\\nlp\\datasets\\xtreme\r\n","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":63,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\n@lhoestq \r\nThe error message is same as before:\r\nException has occurred: UnicodeDecodeError\r\n'cp950' codec can't decode byte 0xe2 in position 111: illegal multibyte sequence\r\n File \"D:\\python\\test\\test.py\", line 3, in <module>\r\n dataset = load_dataset('xtreme', 'tydiqa')\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/87748794-7c216880-c829-11ea-94f0-7caeacb4d865.png)\r\n\r\nI said that I found 4 copy of xtreme.py and add the \u300c, encoding='utf-8'\u300d parameter to the open() function\r\nthese python script was found under this directory\r\nC:\\Users\\USER\\AppData\\Local\\Programs\\Python\\Python37\\Lib\\site-packages\\nlp\\datasets\\xtreme\r\n","embeddings":[-0.2993037999,-0.0795088634,-0.0254610032,0.2176792026,0.3856869936,-0.0703267455,-0.1006556004,0.1843703985,-0.2766493559,0.0560930595,0.1345290989,0.4247641265,-0.0865387693,0.2728081942,0.2161013633,-0.0068473215,-0.0502494387,0.2241975665,0.1564086825,0.0707991794,-0.4088910818,0.1733012795,-0.1744892299,0.1244190782,-0.4139832854,-0.059594024,0.0413277484,0.3134814799,-0.2633447051,-0.6551192999,0.1879901588,-0.0460015163,0.143569544,0.3817632496,-0.000106594,-0.0980057493,0.2307551205,-0.2170889676,-0.1518953592,-0.2767798305,-0.3016500175,-0.3093874454,0.1195951626,-0.3082416654,0.081188336,0.1373253018,0.0110626118,-0.3305060565,0.4484871328,0.3554725945,0.2603322268,0.1282583177,0.021756649,0.1405823976,0.02450905,0.0218689628,-0.0638803914,0.4434989989,0.2685891986,-0.0473258868,0.1204436198,0.1458221078,-0.1865746528,-0.048009187,0.0152620394,0.0447202437,0.0667636618,-0.4407677948,0.1138863415,0.118936263,0.080988206,-0.3776184916,-0.1859687865,0.0942310467,0.1264690012,-0.6105197072,0.2371102422,0.4000205994,-0.3269696534,0.1599569172,-0.067077361,0.0638428926,0.0541390106,-0.0907466486,-0.0465912186,0.1650673151,-0.1873859465,-0.0774285272,0.1572607011,0.0634271652,0.1613905579,0.0171967782,-0.1779077351,0.181212917,-0.2354558259,0.0056721289,-0.1401464492,-0.0122230006,-0.0569794104,-0.070378527,0.4700904489,0.1602924466,-0.0891896114,0.2671580017,0.0374502428,0.1032643393,-0.2167339176,0.100880295,0.1906232685,0.122556828,0.1977968961,0.0752735138,-0.2027528584,-0.2452730983,-0.1989915967,0.4515694082,0.3463511169,-0.1463848501,-0.4804702997,0.0671454966,0.015336344,0.0724580362,0.0974962115,0.1500560194,0.1725723296,0.3843930066,-0.0482332893,0.2677053809,-0.3312215507,-0.2914781868,-0.0979884863,0.2515812218,-0.3490155041,-0.0358149298,0.1118646935,-0.1107490659,0.1102609709,0.2399997562,0.1703832746,-0.3305484354,0.0894226581,-0.171946466,-0.0011075635,0.026131954,0.0958600864,0.155036062,0.1559718102,-0.3678236604,-0.0557275154,0.0636262074,-0.5577607155,-0.1297597438,-0.1729639024,0.2627521753,-0.0124775432,-0.0123801986,-0.5066611767,0.0706070215,0.1725929826,-0.2303977162,0.0631193146,-0.1705555022,-0.1377055347,-0.2525263429,0.175331369,0.4754906595,-0.2941102684,0.0538599826,0.1609711498,0.0338650122,0.3461773694,0.0864404142,-0.2136350572,-0.1357930154,-0.2194357067,-0.0354538262,0.3955557644,-0.4324065447,-0.2422579378,0.3587153852,-0.243787989,-0.008346376,0.0660074726,-0.131799683,0.0356077738,-0.2514543235,0.1312204003,0.4155409336,0.0266562011,0.2746024728,-0.1828944981,-0.3092409074,0.2006269693,0.0163544845,-0.2178280652,-0.220205456,0.2515701354,0.3486474752,0.2898349464,-0.1736857295,-0.1671789885,0.2128673196,0.027012622,-0.2167555988,-0.2222600579,0.0945966989,0.0223033037,-0.0189579371,-0.3267049193,0.0898943469,-0.3444603086,-0.1623487771,-0.1593017578,0.037658222,-0.6050456762,-0.2833513021,0.2350183278,0.2086819708,-0.0824370235,0.0819688886,-0.2075414658,0.2771047354,0.0008027555,-0.0396683104,-0.3657068312,0.0936851799,-0.1336218268,-0.1018197462,0.2282004058,0.3284243047,0.1678859293,-0.1420123279,-0.2412918061,0.3130163848,0.0796097592,0.3886031508,-0.2427394837,-0.1854430884,0.4853195548,-0.6272907257,0.1525293738,0.480600059,0.2121469826,-0.1064735949,0.3261395991,0.3260163665,-0.0221676007,-0.087656416,0.0208773259,-0.1277964264,0.1704484522,-0.0059429887,0.1041062623,-0.2580585182,0.3132897615,0.6806163788,0.3463375866,-0.0056937668,-0.0863397047,-0.0046069031,0.5590988994,-0.0820821971,0.0221946873,0.0778365806,0.12169002,-0.1580358744,-0.0382321887,0.0860441029,0.2751679122,-0.0825376585,-0.1489648372,0.2428869456,0.0354851931,-0.2573782802,0.152756542,-0.0567202866,0.0502872542,0.0647864863,0.1666754782,0.0959484652,-0.257853955,-0.3063188493,0.0870975703,0.1154112071,-0.3568193913,0.314188689,-0.5408015251,-0.3548946381,-0.271730274,0.1222766638,-0.1007573232,-0.0870124325,-0.1659294218,-0.0778104141,0.1684072912,0.1757585406,-0.2078566998,0.0004747527,-0.0593759939,-0.0908879787,0.1813361049,-0.1437168568,-0.2546927035,0.0867510512,0.4046459496,-0.0034862498,0.0353804938,-0.39510566,0.0685483217,0.1820477396,-0.1195266172,0.2201739848,-0.2149030715,0.1363780946,-0.0043688198,0.006755732,0.0071471189,-0.2906437218,0.1497115046,-0.023972474,-0.0502182283,0.1258330345,-0.0785772577,-0.1874869466,-0.2346599698,0.0157003179,-0.5259251595,-0.4137800336,-0.2053127885,0.0765401572,0.2234555483,0.2994021177,-0.0135563686,0.2073113024,0.2510852516,-0.0007844482,-0.121512264,-0.3657850921,0.3067182302,-0.4246502519,-0.4335766435,0.0322742648,-0.0820378885,0.4345403314,-0.0832580104,-0.1792014241,-0.0581072532,-0.1336797178,0.0765964463,-0.0586904474,0.1000950485,-0.0111700762,0.0214985237,-0.0183023475,-0.0689403266,0.080017589,-0.1598852575,0.0732782185,0.3267397285,0.0562297292,0.2255901694,-0.1048821285,-0.0407772698,0.3091386557,0.132510379,0.6449682117,-0.1720910966,0.2896558642,-0.0610420629,-0.059587162,-0.0101828519,-0.1836554706,-0.0560677201,0.1665945053,-0.0855595917,-0.16607669,-0.257684201,0.0474131703,-0.2950181067,-0.2238310277,0.0523650981,0.1114468351,0.1309532374,-0.1122145802,0.1643209457,0.033924412,-0.2491917759,-0.022250006,0.3163499236,-0.1691477895,0.1459115595,-0.222098738,-0.343735069,-0.2073237449,0.1932173818,0.3202220798,0.366300106,0.0591940209,0.2135512084,0.1143074185,-0.1909123659,0.1653681546,0.0742493868,0.0477262735,0.1528622955,0.1147703752,0.0004943624,-0.0868354663,-0.2772709727,0.2468750626,0.071014218,-0.2654931247,-0.1259057671,-0.0268591903,0.1045967042,0.3931480646,-0.1350217611,0.2678692937,-0.4257536232,-0.7568022013,-0.1499162018,-0.2945359647,0.2177583277,0.330001533,-0.0727688521,0.0765950978,0.0575870015,-0.253415674,0.2762099206,0.3764479458,0.3708622456,0.2668077648,-0.1186253428,-0.14264597,-0.2091729045,0.211629197,0.6942180395,0.1674204469,-0.2489526421,0.1772017181,-0.1624294221,0.2050908655,0.2624578476,-0.1384047419,-0.0221755859,-0.1742610186,-0.0046035517,-0.1279199123,0.1411250085,0.2708049715,0.0400169753,-0.5778186917,-0.5629856586,0.5285752416,0.0018852967,0.1206891686,0.1148115173,-0.0753367916,-0.2992390692,0.1864057183,0.1224783957,0.8694338799,0.2407933623,0.1287990808,0.4219368696,0.0710680112,0.1791025996,-0.0133958729,0.2008657306,-0.2237320989,-0.1441680044,-0.0482103005,0.1069190502,0.1695763916,-0.0610566065,-0.3289296329,0.248092711,-0.0318161063,-0.3604763746,0.1572130769,0.2678994536,-0.2972082198,-0.1847035289,-0.5931873918,0.1891243607,-0.1117908433,0.0649575219,-0.1340326816,-0.1960577667,-0.0200170279,-0.2767279744,-0.2806451023,0.3329108953,-0.473508656,0.3480068743,-0.2584133744,-0.5290863514,0.2686245739,0.1564138234,0.2244635075,0.0829588547,0.0273848046,0.1534836739,0.2622379959,0.0240559708,0.0552040339,0.0767488927,0.1274213344,-0.1599383503,-0.0795191377,0.3281883597,0.1406796277,-0.0697795525,0.2809198797,-0.0097971186,-0.1568863839,-0.0493089855,-0.2253405452,-0.2602359653,0.1359807849,-0.0604952574,0.1317283362,-0.0802658871,-0.1057086512,-0.1073965728,0.0571067408,-0.3944144249,-0.0997213945,0.1367491782,0.0153043438,-0.0763811618,0.5025988817,0.1954506785,-0.1406813413,-0.1048961803,0.3054180741,-0.1084147543,-0.5312290192,-0.0495971553,0.0164839197,-0.0170059856,0.0806869939,0.6536359787,-0.0109750638,0.068804346,0.1723845452,-0.5049066544,-0.2466585934,0.0166138001,-0.1233027056,0.0090255952,0.1266393811,0.4019545913,0.0770141184,-0.0840230584,-0.3070727587,0.1770121753,-0.1739885211,-0.0136988126,0.2161898315,-0.0522598214,0.1325297654,0.1732369065,0.1574268639,0.3606143892,-0.0934450552,-0.1727006584,-0.1912987679,0.0932845324,-0.0934174582,-0.3760542572,-0.1042425185,0.0646767616,-0.0465692133,-0.0101372404,0.1986922771,0.0797905475,0.1309215873,0.2023891211,-0.2446155548,0.0136060594,-0.0848861858,0.1965000778,-0.0774524435,0.5980936885,0.164655596,0.2236826569,0.017284723,-0.0131133581,-0.3025083244,0.0197588746,-0.1615501046,0.0791383684,0.0524100959,-0.3789255023,0.3416692615,0.1473592818,-0.1021378636,0.4704594016,-0.3041957915,-0.033303827,0.2028287202,0.1615918726,-0.2177616507,0.0460013971,-0.2467534542,-0.1125781462,-0.0840742588,0.4584672451,0.232008338,-0.08067967,0.1945513338,-0.0239909459,0.3604413569,0.2142913789,0.4479630291,0.313416034,0.0202893112,0.1337037235,-0.1337766498,0.1738455892,0.0654901639,0.1365724355,-0.5954394341,0.2841957211,0.0992020667,0.1946317703,-0.0844506472,-0.3110916018,-0.0892352611,-0.0549091138,-0.2381763756,0.3179439306,0.1723587364,0.3926716447,-0.4416765273,-0.0227639675,-0.1770831794,0.2958931625,-0.0724510252,0.0080596088,0.0955119804,-0.0349784605,-0.4344461858,0.0805488825,-0.0886061564,0.1283377558,-0.0331964493,0.3416313529,-0.1514340639,-0.0637246445,0.0173055306,0.1080752686,0.080984503,-0.2101391107,0.5812780857,0.4417741001,0.14717035,0.2122598737,0.1768318564,0.5937770009,0.4060044587,-0.2153435051,0.0126566328,-0.0964889079,0.0308235269,0.1480909735,0.0102318302,0.2071189284,0.4157857001,0.2502596974,0.1861782372,-0.1576273441,-0.134620145,0.310470134,0.0544281155,-0.0621566512,0.3012603819,-0.071673058,0.2231062055,-0.3759078085,-0.0721329972,-0.4998143613,0.233231768,0.3207105398,0.0169819351,0.0273134243,-0.1229553819,0.1045160666,0.0163004342,0.4972697496,0.2941488624,0.0312457699,-0.1358421743,-0.2252735645,-0.769690156,0.2917549908,0.0701573715,-0.4844248593,-0.1144130304,-0.2295144051,-0.1175687462,0.3979943693,0.049812559,0.2959581017,0.1677707881,0.0109778102,-0.180351451,0.0211547967,-0.0218765177,0.0617012009,-0.2551729381,-0.3233397007,0.1144113615,-0.3997050226,0.1131575033,-0.3319924176,-0.2369958758,-0.3685522377,0.0032364447,0.2550115883,0.2010789812,0.3229779303,-0.3041164577,-0.2139414996,-0.1175049841,-0.1041313261,-0.0419158265,0.3619714379,0.1920211017,0.2813420594,0.0825634748,0.1538796425,-0.5075464845,0.418389976,-0.0698080584,-0.1183640435,-0.2643514872,0.1811924279,-0.0990329012,-0.0012427669,0.2663118839,0.3861822784,0.0138145536,0.2909946442,-0.0476445146,-0.5047470927,0.2918061018,-0.4830423594,-0.1419981569,0.1137423739,0.282933563,-0.206161052,0.0223946162,-0.4061278105,0.2821900249,0.1225741059,-0.0387648717,-0.3610475659,0.3057990968,-0.315697968,-0.1741196662,-0.0470320582,0.3284199238,0.1485316008,-0.1605902612,0.1165081337,-0.2194468826]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"Hi there !\r\nI encountered the same issue with the IMDB dataset on windows. It threw an error about charmap not being able to decode a symbol during the first time I tried to download it. I checked on a remote linux machine I have, and it can't be reproduced.\r\nI added ```encoding='UTF-8'``` to both lines that have ```open``` in ```imdb.py``` (108 and 114) and it worked for me.\r\nThank you !","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":72,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\nHi there !\r\nI encountered the same issue with the IMDB dataset on windows. It threw an error about charmap not being able to decode a symbol during the first time I tried to download it. I checked on a remote linux machine I have, and it can't be reproduced.\r\nI added ```encoding='UTF-8'``` to both lines that have ```open``` in ```imdb.py``` (108 and 114) and it worked for me.\r\nThank you !","embeddings":[-0.3488940001,0.0580565482,-0.0053174524,0.2148150355,0.2911395133,0.0693918839,-0.0231634378,0.1682781577,-0.1171187684,-0.008803363,0.0351622328,0.3814254105,0.0592853576,0.2256459594,0.2042711079,0.137657851,0.0813008025,0.227961272,0.0762692764,-0.0321259983,-0.2550890446,0.2474260926,-0.1208336279,0.1178522632,-0.3460617363,-0.1771093756,0.1752639413,0.2684678733,-0.2989810705,-0.5578821898,0.3264290392,-0.0420311838,0.1637300253,0.3156361282,-0.000107127,-0.1720654815,0.4879135489,-0.2730462551,-0.0977526829,-0.2996372283,-0.0662683621,-0.1792032421,0.208705768,-0.1951277703,0.1642090976,0.2361993343,0.0254221167,-0.2193976045,0.2429872006,0.2548004389,0.2512708008,0.0758318156,-0.0124273971,0.0530121475,0.1059629247,0.0892353132,-0.0343464613,0.4207041264,0.4167104363,-0.0661733001,0.1576461345,0.1413359195,-0.2976844311,-0.1031862423,0.1930611432,0.1217105463,-0.1926509887,-0.5360171199,0.0280171651,0.237992093,0.4053965211,-0.4262819886,-0.2080039084,0.0869111568,0.2992160618,-0.4392408431,0.2078835219,0.4138452113,-0.3085672259,0.1394884735,-0.0505688712,0.1756735742,0.1524308771,-0.0138395345,0.0176998414,-0.0796573311,-0.0850763693,-0.0789386407,0.107614696,-0.0334568508,0.2237876803,0.0531084575,-0.1419592947,0.0521738902,-0.0420180485,-0.0374315307,-0.1653201431,0.0665675104,-0.0967051834,0.0523701906,0.3962653875,0.1933194697,-0.1017117128,0.3944354057,0.2274758965,0.0416718125,-0.2219593525,0.2559438944,0.3583293557,0.1660337746,0.0576864257,-0.0020175995,-0.1636369377,-0.1274368614,-0.1428083181,0.2887331247,0.1514272094,-0.1898154914,-0.4294265211,0.1617305577,0.1705228984,0.2196198106,0.1940185279,0.1838280559,0.1750237495,0.3112075627,0.0333496518,0.3036215901,-0.2544294596,-0.2565978169,-0.0809628963,0.1975775361,-0.2479849905,-0.1120387986,0.2367900759,-0.0987468883,0.1741169244,0.2018454969,0.2456615269,-0.1505301893,0.0676360577,-0.1074694172,0.2129864246,0.2414590269,0.1607815474,0.2428202629,0.1067941934,-0.2313483059,-0.1709671021,0.0372146033,-0.421536386,-0.1010330245,-0.0597743765,0.207286194,-0.081762813,-0.1171906739,-0.6567404866,0.0730897933,0.009207366,-0.3679394126,0.069724001,-0.1777156293,-0.0362018384,-0.0768480822,0.1931199431,0.5341475606,-0.4658977091,0.1469022632,0.0349595398,0.0121597387,0.224407956,0.1452992857,-0.138791889,-0.1420612782,-0.3158390522,0.1119498387,0.3089248538,-0.5038155913,-0.3365866244,0.3986084163,-0.0998629704,-0.0152446581,-0.0035208648,0.0348181501,0.1104175672,-0.2140943259,0.0104302736,0.3849824369,-0.0740710571,0.1820012629,-0.1900574714,-0.1728015393,0.1346137375,-0.0327699557,-0.171205312,-0.1084196121,0.1116732508,0.1699962765,0.230095759,0.0933081061,-0.1538109034,0.1082648337,0.2403595299,-0.008167116,-0.0733689293,0.1044959798,0.0796272904,0.0673356056,-0.2456248999,0.106788218,-0.4861810505,-0.2660392523,-0.1546813101,0.0328306705,-0.3553241491,-0.0886504427,0.2059870809,-0.0387346335,-0.2071613967,0.0492851771,-0.3249136806,0.1765441746,0.1097027436,0.0658427849,-0.2932023108,0.1220303252,-0.185179919,-0.2235825807,0.2431552261,0.1649621874,0.1026879102,-0.0626986623,-0.1499416083,0.2574966848,0.0730739236,0.3449715972,-0.2452047169,-0.2048249543,0.5607063174,-0.548206687,0.113046661,0.4191701412,0.2602438331,-0.1504663527,0.2045992911,0.2479991317,-0.1719045043,0.0086719245,0.0719211251,-0.0847103596,0.2972041368,0.1503051519,0.0177873746,-0.3461951315,0.2551012039,0.5953899026,0.2146585137,-0.0987160578,-0.1114511937,0.1296662688,0.4439419806,-0.0543929003,-0.0165188089,0.0682232007,0.0208389536,-0.0622440167,-0.0035003296,0.0270835478,0.0906314179,-0.0888238624,-0.2187934071,0.212004438,0.0657975301,-0.2460459769,0.1466578692,-0.0243357345,-0.1499833316,0.0709517151,0.0682789981,0.1316547096,-0.2474688441,-0.2776138484,0.0352979638,0.0612460449,-0.3537080884,0.2199955136,-0.4938854873,-0.3810700774,-0.1177900359,0.069818452,-0.0420004651,-0.1311085373,-0.0912832394,0.1379749924,0.1113818809,0.1218913347,-0.3576264679,-0.0745340884,-0.1458536834,-0.1930759996,0.03047801,-0.2158692032,-0.1704759747,0.083177276,0.4746145606,0.0928667188,0.0708577931,-0.4188654423,0.201561749,0.0766297132,-0.1373047084,0.2315514088,-0.2333486974,0.1168904975,0.0479770824,0.0582223013,-0.0004436653,-0.4387207329,0.1450605541,-0.1077361926,-0.0211324245,0.2000148147,-0.0105557768,-0.3811513782,-0.1138438806,0.0026009511,-0.4156636298,-0.2895945609,-0.1918073744,0.0485388711,0.0491561033,0.1971783787,0.1036536843,0.1325834692,0.2278992087,-0.0076735439,-0.0012111965,-0.3418912888,0.4118304849,-0.3309630156,-0.4681514204,0.2416665554,0.0385172293,0.370306462,-0.236439243,-0.2635747492,-0.0220803767,-0.1116258428,0.1798568368,0.1413113028,0.0253992938,-0.0990375355,-0.0507818423,-0.0319888592,-0.0363839045,-0.08936093,-0.1141126975,-0.0868043303,0.1754243821,0.0025579187,0.2680507004,-0.3292317688,-0.0546297096,0.2706374228,0.0768693388,0.7014579177,-0.1841569096,0.38472718,-0.2749556601,-0.1120501235,-0.1152989566,-0.1782622933,-0.0452872217,0.132973507,-0.0346847475,-0.3643234074,-0.3445565701,-0.0151835652,-0.385396719,-0.3656519949,0.0515643284,-0.0386593528,-0.0795245469,-0.0491099954,0.1858221591,0.1852455586,-0.3853033483,0.059519358,0.3000605404,-0.0337144211,0.1691735089,-0.2003348321,-0.4145610631,-0.0238720607,0.0646648854,0.1260301322,0.3964134157,-0.0625854358,0.1580934376,0.0996640101,-0.0700343847,0.2098917663,0.0011674645,0.1445217878,0.1332652718,0.0414980911,-0.0035135527,-0.0736076534,-0.1917610317,0.1426026076,0.075923577,-0.232016921,-0.1212175637,0.0906411484,-0.0179130863,0.4438256025,-0.0154714156,0.1246917099,-0.3756381571,-0.6495364904,-0.1162413061,-0.217860207,0.2925518453,0.1491621435,-0.0827389956,-0.0954445675,-0.0313663334,-0.2161992192,0.2599011958,0.3667046726,0.4251414537,0.0928976387,0.1322709471,-0.1587828696,-0.2337957621,0.4214046299,0.7679156661,-0.0410797372,-0.3446137607,0.0866518691,-0.2720607221,0.1166204363,0.2882134318,-0.1064957753,-0.3059370816,-0.0120732421,-0.149490416,-0.0843588635,-0.0203369372,0.2839379311,-0.0101470994,-0.5025901198,-0.5565865636,0.5252873898,0.0544942208,0.1484496593,0.1141889468,0.141029641,-0.3060490489,0.3216570318,0.0084046377,0.8980358243,0.1697081923,0.0520554148,0.1316681355,0.0856742114,0.2320006788,-0.1181476265,0.1271194369,-0.2963447869,-0.2047519982,-0.0603118427,0.1422052383,0.026593687,0.1766857654,-0.3149734437,0.205743745,-0.1948060393,-0.2809530795,0.0025012782,0.4298159778,-0.4101660848,-0.196333617,-0.3811074495,0.2214437872,-0.0037671316,0.059215717,-0.016595073,-0.1233695894,-0.0648109838,-0.1962016374,-0.5394253135,0.2126027346,-0.4355987608,0.2661518157,-0.1694538295,-0.6015484929,0.2076185495,0.0584996417,0.2319207042,0.2045704573,0.0033769005,0.1583482772,0.4542040825,-0.1000343114,0.1493114084,0.0307283662,0.2625710368,-0.2266551405,-0.109912321,0.179532215,0.1999657303,0.0505417585,0.4349759221,0.0638069957,-0.1949104369,-0.0802446902,-0.4043856561,-0.3073031604,0.1276115477,0.0048043262,0.1415297836,0.0007690262,-0.0947991163,-0.0846794769,0.1823080331,-0.2930835187,-0.1149969697,0.1405196041,-0.1353619695,-0.2118996233,0.440168649,0.2668436766,-0.2931907177,-0.0937638357,0.252941072,0.080999881,-0.6456253529,-0.0327518061,0.1414754093,-0.0271258857,0.0768993646,0.5570614338,-0.0146498252,-0.0392854586,-0.0212190207,-0.4366118908,-0.1500803977,0.1035267711,-0.0610397644,0.0435239375,0.0346659347,0.2332764864,0.1965366453,-0.3172602355,-0.2861512601,0.1628330946,-0.1895154268,-0.0744392052,0.1498775184,0.0548622087,0.0498869643,0.0334464908,0.0986448973,0.3002924621,0.0585887879,-0.1640835106,-0.0407448895,0.0971774086,-0.0833944082,-0.3801465333,-0.0983820185,0.0328741632,-0.1393338293,-0.0724401996,0.1072917134,0.249310866,0.1552940309,0.1637902558,-0.2214391679,-0.0776809156,-0.0907815993,0.2290740013,-0.088227652,0.6197983623,0.1204080135,0.0937776715,0.0755984783,0.0117450198,-0.0024525407,0.0852478743,-0.3687499762,0.0874380246,0.2601130605,-0.2806626558,0.4758726656,0.3395780325,0.0342617221,0.6537329555,-0.1924583912,-0.1549039334,0.2746970057,0.1757349521,-0.2116664946,0.0447173268,-0.1870619357,-0.018688865,-0.1748131216,0.2415549457,0.2729088068,-0.070191659,0.1137248501,-0.1272247285,0.679684937,0.0675558671,0.445405513,0.1634871513,0.0051600812,0.2355174422,-0.1625999659,0.2476620227,0.0433170088,0.1495422125,-0.6107838154,0.0075278827,0.1179182827,0.1389143169,-0.0233899467,-0.3540790081,-0.1168867797,-0.189713195,-0.0814373791,0.2204178572,0.0778034851,0.4481519759,-0.3725140691,0.0363464467,-0.0800350457,0.2290075421,-0.0346462876,0.0572868027,0.1025852114,-0.060394913,-0.3064063191,0.0626410916,-0.1271173954,0.1142823249,-0.1865484118,0.4548628926,-0.0956261158,-0.1367363483,0.0715411529,-0.1400915235,0.1747966558,-0.1992776394,0.3945745528,0.4614478052,0.265183121,0.3899483085,0.1301807612,0.7003698349,0.4630869031,-0.2856807113,0.1554400027,-0.1311315596,-0.0159366168,0.1274091601,0.0577639826,0.2666423917,0.3672748506,0.1465813518,0.1811800301,-0.1162868515,-0.1475501955,0.1802453697,0.2748171389,0.0868372768,0.2116278112,-0.215183422,0.2226527929,-0.6005729437,-0.1891816407,-0.362719655,0.1600746065,0.522095561,-0.122646004,-0.1263445616,-0.1507479846,0.1021173075,0.0950000212,0.4163969755,0.3352278471,0.1280604303,-0.1833306104,0.0290790573,-0.7274269462,0.2501837313,0.1206396669,-0.4472252429,-0.0691539645,-0.0969258994,-0.0752484575,0.317058146,0.0283107739,0.3503979146,-0.0214043166,0.2206030339,0.0137888119,0.0328275822,-0.2921746373,0.0086514223,-0.240058586,-0.1825530827,0.0216026977,-0.5079545379,0.0847153813,-0.2318248898,-0.474378109,-0.3261176646,0.2696191669,0.3063947558,0.1306885779,0.5721482635,-0.1645707488,-0.2164476216,-0.2139157653,-0.2999965549,-0.1513897628,0.5063722134,0.1918576509,0.015399849,0.062090151,0.1250704676,-0.4375211298,0.3358762264,-0.1156736389,-0.1946085095,-0.381129086,-0.0392204374,-0.0960305333,-0.0140085453,0.1611045301,0.3276620805,0.0059702643,0.1686802953,-0.1238929704,-0.6365171075,0.3105712235,-0.3692490757,-0.0626580566,0.051092647,0.2514259815,-0.150035426,0.1229878366,-0.510400176,0.2605609596,-0.0195220206,-0.1634387821,-0.3511119187,0.1121345982,-0.212051779,-0.0391761847,-0.1606920809,0.3674101532,0.220997259,-0.2025869191,-0.0319335125,-0.1216388717]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"> Hi there !\r\n> I encountered the same issue with the IMDB dataset on windows. It threw an error about charmap not being able to decode a symbol during the first time I tried to download it. I checked on a remote linux machine I have, and it can't be reproduced.\r\n> I added `encoding='UTF-8'` to both lines that have `open` in `imdb.py` (108 and 114) and it worked for me.\r\n> Thank you !\r\n\r\nHello !\r\nGlad you managed to fix this issue on your side.\r\nDo you mind opening a PR for IMDB ?","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":97,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\n> Hi there !\r\n> I encountered the same issue with the IMDB dataset on windows. It threw an error about charmap not being able to decode a symbol during the first time I tried to download it. I checked on a remote linux machine I have, and it can't be reproduced.\r\n> I added `encoding='UTF-8'` to both lines that have `open` in `imdb.py` (108 and 114) and it worked for me.\r\n> Thank you !\r\n\r\nHello !\r\nGlad you managed to fix this issue on your side.\r\nDo you mind opening a PR for IMDB ?","embeddings":[-0.3819619417,0.0328492075,-0.0154810492,0.2028752118,0.3120900989,0.0689587519,0.0058169118,0.1825693846,-0.083670944,0.0293346681,0.014070197,0.4008607566,0.0577816293,0.2399227768,0.1625403166,0.14097929,0.0723040178,0.2220351994,0.1050566211,-0.0275190789,-0.2586425841,0.2288965434,-0.1308452934,0.1349900067,-0.3728035688,-0.1832351387,0.1941950768,0.2883248329,-0.2690778673,-0.5402297378,0.3226154447,-0.0409633145,0.1525747478,0.3042993248,-0.0001053745,-0.1787819266,0.5192816854,-0.2602533102,-0.1324340999,-0.2859140933,-0.0765627101,-0.1620351523,0.2123644501,-0.1953220963,0.1557098031,0.2527001202,0.0229268353,-0.2271054238,0.2151814699,0.2319758087,0.267660141,0.0608118139,-0.0186496228,0.045802366,0.0669800267,0.1003600582,-0.0261969399,0.429251641,0.4187467098,-0.0728279501,0.1366805434,0.1390876472,-0.2907060683,-0.117742613,0.2308941931,0.1356046498,-0.1527103037,-0.539737761,0.0303562395,0.2396232039,0.4043304026,-0.4341339767,-0.2198884487,0.0727621093,0.2851529717,-0.4451404214,0.2295992374,0.3813069165,-0.3105047345,0.1312863827,-0.0726721138,0.1851212233,0.1598808467,-0.0166890323,0.0241958275,-0.1213527247,-0.0719853118,-0.0958218426,0.0644970015,-0.0755638033,0.2218587697,0.0408137366,-0.1298449337,0.031149948,-0.0231444854,-0.0654445216,-0.1572759002,0.0888461322,-0.1031697839,0.0526708998,0.3892244995,0.1904302835,-0.1060628444,0.3926351666,0.2187597603,0.0764838234,-0.1863272786,0.2649510503,0.3690944612,0.1792969108,0.0352390744,-0.007588035,-0.1356833875,-0.1409618109,-0.1185562164,0.2891425192,0.1462384462,-0.1817835271,-0.4506699145,0.1488844603,0.1540304571,0.2377270162,0.1990226209,0.1871261448,0.1694632322,0.3224521875,0.0406058207,0.2857062519,-0.2505064309,-0.2612249553,-0.0903877988,0.1797525436,-0.2592477798,-0.1103608608,0.218302846,-0.101804696,0.1758921295,0.2183071077,0.2862075567,-0.1571925879,0.05349309,-0.0925322846,0.2198350728,0.2458468229,0.1864881665,0.2278587371,0.1167296171,-0.2549138665,-0.1482722759,0.0122956978,-0.4237723351,-0.1114712805,-0.0551036336,0.2240451723,-0.0579555444,-0.0987666398,-0.6468676925,0.06157168,0.0012063287,-0.36565727,0.075383015,-0.1864982992,-0.0492687449,-0.0849167258,0.1986242384,0.5331299901,-0.4609873891,0.1437164843,0.0010171983,0.0304283388,0.2228224427,0.1494708657,-0.1381026655,-0.1621319205,-0.3192888498,0.1065663695,0.2765079141,-0.5096930861,-0.2977873385,0.4063526988,-0.093693465,-0.0049822489,-0.0029611564,0.0064352555,0.1321465522,-0.2374905348,0.0350820944,0.3566595018,-0.0839654431,0.1826423705,-0.212310642,-0.1224609464,0.0972468033,-0.0655148029,-0.1674876511,-0.1106645837,0.13592498,0.1589934826,0.2215439826,0.1049713492,-0.1610754728,0.10810332,0.2616814077,0.0201453511,-0.0755081847,0.1233400404,0.0588506982,0.0779778659,-0.2107319385,0.0863468647,-0.4398860335,-0.2682012618,-0.1456341296,0.0335430428,-0.3326475322,-0.0839553401,0.2256686985,-0.0241779946,-0.2100736797,0.0258989856,-0.3468070328,0.1851453334,0.088781476,0.0475903675,-0.2662093937,0.1227076501,-0.166383177,-0.2624225616,0.2336537838,0.1494119316,0.1016341522,-0.0474017262,-0.1530420631,0.282733947,0.0698587149,0.3505370021,-0.2592075169,-0.1781454682,0.5491707921,-0.5299206972,0.1276390105,0.3686937094,0.246872142,-0.1486336738,0.201056689,0.2885649502,-0.1976458579,0.0161664598,0.0763596967,-0.0940714255,0.2952249348,0.1128005981,0.0039010101,-0.3647915423,0.2632720768,0.5909789801,0.2062430233,-0.1096372753,-0.1240621433,0.1479021758,0.449782759,-0.0390714072,-0.0305704251,0.0685048848,0.0158061367,-0.0551207773,-0.0131513281,-0.0006509282,0.078051284,-0.0851645991,-0.2253017575,0.2005329728,0.0598540753,-0.2409396917,0.1696550995,-0.0282123424,-0.1508947313,0.1010195836,0.0679655373,0.117464222,-0.2235124409,-0.2603949308,0.0180319659,0.0348623209,-0.3721189499,0.1928663701,-0.462643981,-0.3444407284,-0.0923713073,0.0627866834,-0.0048273318,-0.1374607533,-0.0746689439,0.1510362029,0.1117161438,0.1418263763,-0.328617245,-0.0649257377,-0.1468442678,-0.1589344442,0.0047206846,-0.2141066492,-0.159800753,0.098413527,0.4974024892,0.0695274174,0.0424909256,-0.4195205271,0.1661721468,0.0735995099,-0.1208945736,0.2249321342,-0.2262669951,0.1265570521,0.042630665,0.0500894226,0.0114448238,-0.4530969262,0.14853549,-0.123546958,-0.020363126,0.2205782533,-0.0266792905,-0.3890245259,-0.1062777266,-0.0042591956,-0.3799142241,-0.2897257507,-0.1875100136,0.0466420129,0.0419532172,0.2030296475,0.1531606913,0.1282235682,0.2437355965,-0.0116908848,-0.011725178,-0.3491702974,0.3963178694,-0.3242890537,-0.4361437857,0.2484275997,0.0181289911,0.3712930679,-0.2505528033,-0.2293863297,-0.0445735864,-0.11632967,0.2337530404,0.1334402561,0.0063670604,-0.1004881933,-0.0388952345,-0.0566139854,-0.0311206598,-0.0822284594,-0.1118074134,-0.0823013261,0.1642276645,-0.0008164028,0.3257593513,-0.344471097,-0.0641670078,0.2906715572,0.06649369,0.6802045703,-0.1825071126,0.3822169006,-0.2870160043,-0.106695205,-0.1105034053,-0.1770844311,-0.0840747654,0.141868934,-0.0455126464,-0.3872093856,-0.3534776568,-0.0205649044,-0.3872495592,-0.3599244356,0.0646854118,-0.0233648717,-0.0833874717,-0.0514740385,0.1716666669,0.1632600129,-0.37394014,0.0709823072,0.2905241847,0.0032344093,0.139034912,-0.145648241,-0.4058155119,0.0373891219,0.0740528181,0.1374882311,0.3621405661,-0.0664484799,0.1329337209,0.1096546352,-0.0678949133,0.2257568687,-0.0245674383,0.1511810124,0.1217980757,0.032578215,0.0282563902,-0.0670831352,-0.2033531219,0.1408226043,0.0898728818,-0.2273913473,-0.1217870191,0.0814794227,-0.0189183913,0.4337356687,-0.0289860461,0.1312062442,-0.3555260301,-0.60489434,-0.1163032204,-0.2135594338,0.305626452,0.131364733,-0.0855510533,-0.0886630267,-0.0354815125,-0.2256162763,0.2622626424,0.3639516532,0.4189310074,0.0952569917,0.1512110978,-0.1506405622,-0.2962839603,0.475600034,0.7579312921,-0.0628825426,-0.3353407085,0.0696815476,-0.2625261545,0.1225967556,0.2908003926,-0.0995926261,-0.3634563088,-0.0428484455,-0.1352626979,-0.090411447,-0.0461262055,0.2596377134,0.0115136458,-0.4596985877,-0.5531419516,0.4936639369,0.0687866807,0.1376950592,0.1065335572,0.1512177438,-0.3186249435,0.3118928969,0.049996499,0.8768302798,0.1690900028,0.017832635,0.1353872865,0.1137554646,0.1773058623,-0.1257158667,0.1167740375,-0.3097364902,-0.2357067168,-0.0424279012,0.1386725008,0.0176188853,0.2051659971,-0.3177805543,0.1894604564,-0.223888427,-0.3268328011,0.0049137045,0.4409636855,-0.4308351874,-0.1852421165,-0.3423239887,0.2386997789,0.0068175299,0.0410209261,0.0084728645,-0.1343964189,-0.0586546138,-0.1826479435,-0.5604509711,0.2045677751,-0.4502363503,0.2335237414,-0.1552653015,-0.5828344822,0.2049240768,0.0524209179,0.2246316224,0.2009427547,0.0352862254,0.1392655373,0.4998711348,-0.0876825824,0.1361323297,0.0344668925,0.302097261,-0.2286529839,-0.0887025446,0.1868336946,0.2094143927,0.0480540991,0.4210775197,0.0568980686,-0.1814818233,-0.0926158726,-0.3883619905,-0.3068867922,0.1135516539,-0.0009188732,0.1626931578,-0.0125491563,-0.0806987584,-0.0822210684,0.1979566365,-0.2848179638,-0.1028458774,0.1209294274,-0.1530727446,-0.1807934344,0.4431288242,0.2550024688,-0.3089768291,-0.097634539,0.2654067874,0.0895829126,-0.6359009147,0.010724498,0.1750412732,-0.0135917645,0.0732014477,0.5347981453,-0.0093496982,-0.0093095414,-0.0575879104,-0.4145116508,-0.1540647298,0.1111314148,-0.0529177524,0.0361155644,0.0036041439,0.2317657173,0.1910300106,-0.3191353977,-0.3135760128,0.1277838945,-0.1809907854,-0.0826671049,0.1218072101,0.0374046639,0.0613201521,0.045812238,0.1207921579,0.2752498984,0.0704007819,-0.1856065989,-0.0310355369,0.0935355574,-0.094586961,-0.3848304451,-0.1018102244,0.0509496555,-0.1382485926,-0.0940426961,0.1085347012,0.2424838245,0.1397394985,0.1503269523,-0.2062701583,-0.073832877,-0.0967533737,0.2157619298,-0.0656300932,0.6222367287,0.0835404024,0.0899163187,0.0687403977,0.0016751818,-0.0090172756,0.0916678309,-0.3974384665,0.1120015681,0.26126647,-0.2647684216,0.4513229728,0.3277536035,0.0248488858,0.6391988993,-0.1918762475,-0.1447709203,0.248061344,0.1942305416,-0.1797551811,0.0820862949,-0.191009149,-0.0413830206,-0.1458640099,0.2278404236,0.2588407397,-0.0676348731,0.1220639497,-0.1367561072,0.6583895087,0.0565245487,0.4018707275,0.1692891717,0.0072391331,0.2384666204,-0.1481671482,0.2285065949,0.0441475995,0.1129239276,-0.5978399515,0.0120287994,0.0928180888,0.1366985142,-0.0247608852,-0.3500495851,-0.0962140411,-0.2052513063,-0.1049077883,0.2066025734,0.0769884586,0.4065267742,-0.3523640335,0.046308428,-0.072803013,0.2305919677,-0.0307384916,0.0494089574,0.079704769,-0.0599665344,-0.3355242014,0.0768496022,-0.14995794,0.1301865727,-0.1789731979,0.4968090057,-0.0891897082,-0.0880018175,0.0696658641,-0.1445417553,0.2056099623,-0.208762899,0.4046178758,0.4655504823,0.2585849166,0.3722355068,0.1180809736,0.7204274535,0.4647769928,-0.2893509865,0.1401359439,-0.1179881096,-0.0371787772,0.1352969557,0.0572050437,0.2359024286,0.3432407081,0.1024343148,0.2056760788,-0.1445978731,-0.1523905098,0.171033591,0.2750335634,0.0960647315,0.1789361537,-0.1939712465,0.2216333151,-0.6160324812,-0.1946111172,-0.327604562,0.1413036734,0.5070897341,-0.142665863,-0.1561326534,-0.1630194932,0.11896687,0.0971945822,0.3863836229,0.3257065117,0.1111597866,-0.181224227,0.0590067506,-0.6790173054,0.2511717975,0.1505207568,-0.4328428507,-0.062822096,-0.0845034048,-0.070796594,0.2998313904,0.0380311236,0.3017372489,-0.0169891678,0.237749055,0.0184392165,0.0462243892,-0.2748252153,-0.0114823598,-0.2171442807,-0.1663309038,0.0131078484,-0.5385060906,0.1001397893,-0.2253272384,-0.4575330317,-0.2884438634,0.3197563291,0.2966777682,0.1021384448,0.5777387619,-0.1906375885,-0.2266547531,-0.1735040247,-0.3024668396,-0.1558376402,0.5097603202,0.1807519644,0.0282093976,0.0652351901,0.0682393238,-0.4261283576,0.3391428888,-0.1371008456,-0.2331565768,-0.3533230424,-0.0364744887,-0.1109627709,-0.0156721137,0.1409453601,0.3411542177,0.0085548265,0.1746233702,-0.1367999017,-0.6540452242,0.3415573239,-0.3543437719,-0.0692147166,0.0435866006,0.2550292015,-0.1510982215,0.115041703,-0.4863814116,0.2621539533,-0.0149342008,-0.1655167192,-0.3533979356,0.1019227803,-0.1916577369,-0.0216968153,-0.1677824706,0.3394935131,0.214702636,-0.1901502311,-0.0554054491,-0.1051219925]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"> This is probably a Windows issue, we need to specify the encoding when `load_dataset()` reads the original CSV file.\r\n> Try to find the `open()` statement called by `load_dataset()` and add an `encoding='utf-8'` parameter.\r\n> See issues #242 and #307\r\n\r\nSorry for not responding for about a month.\r\nI have just found that it is necessary to change \/ add the environment variable as what was told in #242.\r\nEverything works after I add the new environment variable and restart my PC.\r\n\r\nI think the encoding issue for windows isn't limited to the open() function call specific to few dataset, but actually in the entire library, depends on the machine \/ os you use.","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":115,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\n> This is probably a Windows issue, we need to specify the encoding when `load_dataset()` reads the original CSV file.\r\n> Try to find the `open()` statement called by `load_dataset()` and add an `encoding='utf-8'` parameter.\r\n> See issues #242 and #307\r\n\r\nSorry for not responding for about a month.\r\nI have just found that it is necessary to change \/ add the environment variable as what was told in #242.\r\nEverything works after I add the new environment variable and restart my PC.\r\n\r\nI think the encoding issue for windows isn't limited to the open() function call specific to few dataset, but actually in the entire library, depends on the machine \/ os you use.","embeddings":[-0.3218461871,0.0410403796,-0.0172628704,0.1921315044,0.384137392,-0.0418330804,0.0533850081,0.1839184314,-0.2438150942,0.07465633,0.1122436076,0.4671255648,0.1107988954,0.3263352215,0.2156340927,0.0793417469,0.121593371,0.2935247719,0.0367938206,0.0253933985,-0.4600175023,0.1801846474,-0.1028250009,0.1062120944,-0.2276374996,-0.1009323597,0.0553654954,0.2734963,-0.2023732513,-0.5459724069,0.1896327585,-0.0750284195,0.2605702877,0.2277506739,-0.0001059823,-0.0440538898,0.2804324925,-0.1803229153,-0.1569127738,-0.2889727354,-0.1975437403,-0.3678892255,0.1078184769,-0.2368742228,0.0635781735,0.2137286961,-0.002422865,-0.2136175185,0.4395005107,0.3770242333,0.2794305682,0.135423705,-0.1147260815,0.1099153012,0.0541248806,0.0573215336,-0.0186663661,0.5010844469,0.2131777406,-0.2339810729,0.0298318323,0.1721233726,-0.2423514128,-0.0580992885,0.0973245502,0.1594382823,-0.2101520449,-0.4088995457,0.1295930147,0.1351365596,0.3100127876,-0.32497558,-0.1276594549,0.0083150454,0.1066843569,-0.7205759287,0.3261366487,0.4268108606,-0.2456021011,0.1342461705,0.0609913804,0.2269624919,0.0461214818,-0.0214395002,0.0695758462,-0.0316437259,-0.2375248969,-0.0469956733,0.2229003608,-0.0380632356,0.0924569294,0.0012147638,-0.2106275111,0.0815219954,-0.2083330303,0.0376022011,-0.1750782877,0.0315689147,-0.1538994163,-0.0620154887,0.5168815851,0.1546454132,0.013629592,0.3213183582,0.1049254388,0.0797767639,-0.1220801398,0.1708958149,0.3792809248,-0.0342668928,0.1415133774,-0.0152551662,-0.2251727432,-0.347530961,-0.2094828337,0.3959975243,0.216550678,-0.1523399949,-0.4361654222,0.1212785318,0.0980656147,0.0702150315,0.0240607169,0.2243415564,0.2047697306,0.3466665447,0.0168047529,0.2039867938,-0.3592662215,-0.2429448366,-0.017309187,0.1819052845,-0.368889004,-0.069752641,0.2735366821,-0.1578790694,0.0549857467,0.2533173859,0.168840006,-0.2118099332,-0.0451447181,-0.2011538744,0.1154903769,0.1348486096,0.0499350391,0.2183388621,0.1298353076,-0.3230085373,-0.0904251114,0.0744729415,-0.5151638985,-0.0060920529,-0.0799663141,0.2387244105,-0.1374502778,-0.0513865463,-0.4799760282,0.1780171543,0.0198981147,-0.4306436777,0.0101035116,-0.2054034472,-0.1354725808,-0.2135009319,0.187598452,0.5726647377,-0.5493261814,0.2175498307,0.2308416069,0.0270515978,0.2820723355,0.1310224682,-0.1339088529,-0.1769125313,-0.1706348807,-0.0320758484,0.2616347671,-0.2815617621,-0.273617506,0.3522999585,-0.0072575039,-0.0457922742,0.0350057259,-0.1634368151,0.1933966726,-0.2519917786,0.0321125425,0.3793457448,0.0374853723,0.3035112619,-0.201108709,-0.1544061601,0.2732438147,0.0107407169,-0.2936501801,-0.2032541931,0.2821953595,0.1416630596,0.1744512916,-0.1511706412,-0.1337528527,0.1587605625,0.2327430993,-0.2096036524,-0.0254872553,0.1805101931,-0.0249594674,-0.0260686409,-0.2298215181,0.0581737272,-0.2890240252,-0.1495994925,-0.0929415002,0.0466047227,-0.324644804,-0.2399466932,0.2032574117,0.036951106,-0.1426317543,-0.0014226306,-0.1411515474,0.2768875062,0.0107317977,-0.0139016593,-0.1724307984,0.186179474,-0.0900348499,-0.2032659352,0.2726712525,0.3069351315,0.1852348894,-0.1314292997,-0.2873487175,0.291446209,0.0348336473,0.3578834236,-0.2649577558,-0.2401684523,0.4578382671,-0.4840072691,0.0709460452,0.4611472189,0.3018136621,-0.0979863778,0.1545099765,0.3615797758,-0.1767522395,0.0195953175,-0.0165709723,-0.1244196668,0.2131323516,0.030701099,0.0570870601,-0.1943311393,0.3480555415,0.5848902464,0.2559981644,0.0374703594,-0.1499472708,-0.0009214273,0.5480136275,-0.0926613361,-0.0730473921,0.1219992116,0.0281347297,-0.1021295711,0.0819737464,0.0354270525,0.3410062194,-0.118401669,-0.2693512142,0.2631337941,0.0719557628,-0.2493242174,0.1785944253,-0.0663466156,-0.0398111865,0.001624636,0.109225072,0.2062709332,-0.2464142591,-0.0896531641,0.1819762886,0.1307782084,-0.3706768751,0.3590084016,-0.5268562436,-0.4222705364,-0.2778965831,0.2370275557,-0.1506250948,0.0565890856,-0.2060832381,-0.0146384789,0.0048544742,0.1809751391,-0.1959636807,0.0591054447,-0.1677372307,-0.1239162162,0.1611836404,-0.2020172626,-0.3256495297,0.1091831252,0.4991661012,-0.067203477,0.0567062274,-0.3631934226,0.1340762377,0.1923754066,-0.1332850009,0.1386752278,-0.1958273351,0.217488721,-0.0461457185,0.0197295472,-0.0940226093,-0.4398115575,0.1757384837,-0.0620789118,0.0180765465,0.1688304543,-0.05194363,-0.2897469103,-0.2744269073,-0.1558913887,-0.5099146366,-0.3886059225,-0.1136434451,-0.1147130132,0.0890478119,0.3229112625,0.0256176572,0.0290212259,0.1873557121,-0.1173400953,-0.0495327041,-0.1487429291,0.3419561982,-0.4083913565,-0.5525043011,0.1008481905,-0.0198440161,0.319747597,-0.1563616544,-0.189694345,-0.006978326,-0.1507255435,0.1050705761,0.0035450684,0.1002129838,-0.0547260791,-0.047074724,-0.0254627932,-0.1998526454,-0.19223997,-0.0613547042,0.0895951837,0.2490141839,0.0475215241,0.2492836118,-0.2908984721,0.0163077638,0.3181402981,0.0356075317,0.5225813985,-0.136158064,0.4234559834,-0.0275136884,-0.0585542843,0.1832724214,-0.3034586012,-0.0593872592,0.3025193512,-0.0573699512,-0.2925336659,-0.3560016751,-0.0560589321,-0.3991360068,-0.4300525188,0.2192284018,0.0196995214,0.0863977373,-0.0032785621,0.1869965345,0.0527210459,-0.3269921839,-0.0897169411,0.3263348043,-0.0792777687,0.1276283264,-0.2166727036,-0.5177024007,-0.1570398062,0.1196628883,0.1495017409,0.4917266667,-0.0036261335,0.1317594349,0.0595846921,-0.1126442179,0.1731643528,0.1403090954,0.0347486883,0.2260690928,0.1334953159,0.0010703381,-0.1015533581,-0.2331471592,0.1813055128,0.0255562756,-0.2500877678,-0.1180429384,-0.0594409928,0.0258445945,0.4494216442,-0.1025727168,0.1163970679,-0.4000313282,-0.5393142104,-0.0066741668,-0.2123168111,0.1392350644,0.2193727046,-0.2538374066,0.0415286683,0.0574586503,-0.2547736168,0.3427104056,0.4044591188,0.4600491524,0.1762464046,0.0257118642,-0.2020936906,-0.0036790203,0.2242475152,0.6949768662,-0.0148349497,-0.3524723947,0.1786700338,-0.226141721,0.3051573932,0.2745775878,-0.230590567,-0.0252358019,-0.1198883206,-0.0643103868,-0.2265751362,0.13734442,0.2278237045,0.1492484212,-0.6115511656,-0.5212317705,0.4964287579,-0.0708808675,0.1042995974,0.1622760743,-0.0925315693,-0.2440129369,0.226763472,0.0110185053,0.8180406094,0.1871762276,0.1551414132,0.2317282557,0.0124782827,0.2823655009,-0.0824645236,0.1657141596,-0.2303946763,-0.1943763047,0.0515924543,0.1619336754,0.2197645456,0.0098032048,-0.3544755876,0.1651540101,0.1574190259,-0.1693667918,0.1786859185,0.2002147585,-0.2210113108,-0.3044354618,-0.546343267,0.1954889596,-0.0740817934,0.0194921792,-0.0677196905,-0.2836328149,0.0822208226,-0.2928809226,-0.2578263581,0.3182097077,-0.5277633071,0.1614257395,-0.1627800465,-0.6413753033,0.3334997892,0.1623921841,0.174607113,0.2281083018,-0.0163869765,0.1101090387,0.2558592558,0.0816322118,0.0255466737,0.1103220358,0.1319285929,-0.172083497,-0.0298330896,0.2104112357,0.049755279,0.1005036831,0.3739160895,0.0870173424,-0.1857058257,-0.1484855264,-0.1668726653,-0.2551545501,0.1653783023,-0.0172036979,0.1532731205,-0.0348779857,-0.0256393645,0.0324689448,-0.0089336662,-0.4061691761,-0.1536923647,0.0468813442,-0.1016245335,-0.1706160307,0.4922056496,0.1520532519,-0.2427227795,-0.089872852,0.2722656429,-0.1683488339,-0.4991505742,-0.0479399003,0.2028663605,0.0255379174,0.0953286886,0.6194525957,-0.0614955872,0.0107835559,0.0822400376,-0.5254033804,-0.2215742469,-0.0073884814,-0.076701276,-0.0994525179,0.1991027445,0.2663198113,0.0984671786,-0.1928877681,-0.3095079958,0.0479178317,-0.017729694,0.0585790426,0.2214011699,0.0117230108,0.1854635179,0.0903156251,0.1096259281,0.3042016923,-0.0431339182,-0.1653181911,-0.196587041,0.1016260684,-0.0849458277,-0.335591346,-0.0742775202,0.0007749144,0.0120751513,-0.1772140265,0.1821675152,0.0850233063,0.0071399249,0.1622144133,-0.3617420793,0.1223220974,-0.0545535982,0.2537569106,-0.1959923655,0.545429647,0.1013526544,0.1406719983,0.0746451542,0.0525413789,-0.1080557853,-0.0751104429,-0.3167439103,0.0266650803,0.1425265819,-0.3232636452,0.3874883056,0.1754031032,-0.04124685,0.5807672739,-0.1149188876,0.0385227427,0.4276025891,0.1528723538,-0.2509970963,0.1376321912,-0.3420262933,-0.0307525136,-0.1033081412,0.2464760393,0.2873434424,-0.2221657932,0.2221741229,-0.1821422726,0.4188605249,0.1687107235,0.4179499447,0.5055607557,0.2075406164,0.2703203857,-0.1195038408,0.1081862748,-0.0125814946,0.0693636984,-0.5291959047,0.0862778947,0.1331267655,0.2308525443,-0.156981945,-0.4490400553,-0.0620261766,-0.1416445524,-0.1820909083,0.3372346759,0.0895039961,0.5444588661,-0.5150149465,-0.0241567567,-0.0906512216,0.171039328,-0.0851531252,0.018697707,0.1001987159,0.0202853847,-0.2237954736,0.1610750109,-0.1375225186,0.127242133,-0.0745180696,0.3670581579,-0.1495955139,-0.031094322,0.0380569659,0.0136259831,0.0813056752,-0.1314342916,0.4910912514,0.26424703,0.1730858535,0.2993219197,0.0991365165,0.6880450845,0.4429658651,-0.0856582895,0.0140996436,-0.1428178549,0.084791787,0.127745375,0.1018152013,0.3066107929,0.4459532201,0.1761083007,0.1773955375,-0.1048400998,-0.0791594461,0.171827808,0.026785057,-0.0118570514,0.0672976747,-0.0916434675,0.2169593424,-0.4925687313,-0.1169272512,-0.3307475448,0.2565571666,0.4066555798,0.1152113527,0.0020711564,-0.1124933958,0.1001630872,0.1549316645,0.4857947528,0.1887947321,0.1078948379,-0.138364017,-0.0711545795,-0.700378418,0.2771182954,0.1238695905,-0.3950075209,-0.1486751139,-0.3675146401,0.0559088774,0.296987921,-0.091839768,0.2815228701,0.0999480262,0.0551989488,-0.1643672138,-0.00532752,-0.1478890628,0.0649378076,-0.2525982261,-0.2197632194,0.0949417353,-0.2618749738,0.0492761545,-0.2654634416,-0.2644872367,-0.3507200778,0.0578473471,0.3458444476,0.2307631522,0.4740209877,-0.2685818374,-0.0703146458,-0.1581528485,-0.1835353225,0.0014247,0.3380472064,0.1791274995,0.0468649864,0.0372321531,0.1575435847,-0.4053705633,0.1725748181,-0.0981007591,-0.0803118274,-0.1913542151,0.0461889617,0.0541714057,-0.1207838506,0.297914356,0.4182400405,-0.079811208,0.0535574481,-0.1110068411,-0.5214460492,0.2054070532,-0.6085336804,-0.1537293494,0.1054372787,0.3048875034,-0.2272161394,0.1380761266,-0.4638049304,0.2531892657,0.0248383805,-0.1234541833,-0.3444531858,0.1683741957,-0.2985502779,-0.1377820671,-0.0892174542,0.4464707971,0.1969401538,-0.3050907552,0.1170786098,-0.2842414379]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/347","title":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')","comments":"Since #481 we shouldn't have other issues with encodings as they need to be set to \"utf-8\" be default.\r\n\r\nClosing this one, but feel free to re-open if you gave other questions","body":"![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n","comment_length":32,"text":"'cp950' codec error from load_dataset('xtreme', 'tydiqa')\n![image](https:\/\/user-images.githubusercontent.com\/50871412\/86744744-67481680-c06c-11ea-8612-b77eba92a392.png)\r\n\r\nI guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps :\r\nhttps:\/\/www.python.org\/dev\/peps\/pep-0263\/\r\n\r\nI guess the error was triggered by the code \" module = importlib.import_module(module_path)\" at line 57 in the source code: nlp\/src\/nlp\/load.py \/ (https:\/\/github.com\/huggingface\/nlp\/blob\/911d5596f9b500e39af8642fe3d1b891758999c7\/src\/nlp\/load.py#L51)\r\n\r\nAny ideas?\r\n\r\np.s. tried the same code on colab, that runs perfectly\r\n\nSince #481 we shouldn't have other issues with encodings as they need to be set to \"utf-8\" be default.\r\n\r\nClosing this one, but feel free to re-open if you gave other questions","embeddings":[-0.2920763791,-0.11546655,-0.0485051274,0.1429698169,0.3976854682,-0.1386851072,-0.0299375225,0.207310617,-0.3486039042,0.1107885614,0.1741642654,0.4716804028,0.0220731571,0.3367837369,0.2393419445,-0.022137329,0.034644302,0.2812201679,-0.0362177938,-0.0421609916,-0.270514071,0.1711585224,-0.0659664944,-0.0170620605,-0.3997241855,-0.012458141,0.0486944318,0.2974319458,-0.3648321629,-0.475156337,0.1951902062,0.0119117526,0.135913834,0.259455055,-0.0001011886,-0.0771028772,0.3384810984,-0.2374668419,-0.0726838037,-0.2214623988,-0.216215387,-0.3395412564,0.0997417122,-0.1581599265,-0.0460250676,0.2871535122,0.0547698215,-0.0264533162,0.4907978475,0.2225217819,0.3361176252,0.1082952097,-0.0871384889,0.0970920026,0.0383858159,0.1536255032,-0.0810595378,0.40453583,0.2657889128,-0.2489144504,0.130001545,0.3025204837,-0.1547402442,-0.1583583951,0.0455346406,0.0050726337,-0.07711339,-0.4147141278,0.0786356255,0.1539196968,0.1933080405,-0.3020585477,-0.0559194386,0.0919888839,0.1005073786,-0.5315515399,0.2414298356,0.2892900407,-0.2620902658,0.0932654068,-0.0564309619,0.1054813787,0.0520006791,-0.043190334,0.0325556062,0.1327358186,-0.1192581356,-0.124322392,0.1317437887,0.0007684184,0.1545886397,0.0858947411,-0.20291394,0.210920468,-0.1349487007,-0.1016724706,-0.119392328,-0.0037293723,-0.079894498,-0.1457085758,0.5618266463,0.1537452787,-0.0523674786,0.3236987889,0.1299596429,0.0667191371,-0.1919053942,0.1937870681,0.231200248,0.0484264903,0.1330812573,0.0626800954,-0.1734261662,-0.2025741041,-0.3286463916,0.2584735751,0.1405035704,-0.1802671403,-0.387039572,0.0865402147,0.0774649903,0.1144331023,0.0800361261,0.1966690421,0.2150232792,0.3200768828,-0.0310161225,0.2661576271,-0.3405717909,-0.3002048433,-0.1536055356,0.2318949401,-0.426307559,-0.1337781698,0.1555860192,-0.0103649162,0.1558938175,0.2073730826,0.1698080152,-0.124043718,0.0667798519,-0.0724978596,0.0879873931,0.1300014406,-0.0004067399,0.0970363915,0.1606966257,-0.2576760948,-0.0956009999,0.0696871206,-0.406974107,-0.0201010816,-0.052260939,0.3179693818,-0.0753247589,-0.046668414,-0.4660201669,0.1913840771,0.1194681749,-0.3060785532,0.1238185167,-0.1691370606,-0.0626114383,-0.1708720624,0.1583511978,0.4247146547,-0.4670240283,0.0496596918,0.2211828232,-0.0012428117,0.3695967495,0.0587585829,-0.1565062255,-0.1428921074,-0.1330500394,-0.0047206599,0.3198845088,-0.3632705808,-0.2036395967,0.2700951695,-0.0475943014,-0.0306829866,-0.0173880551,-0.0106357597,0.0530152656,-0.1864280701,0.0972895399,0.4720681906,0.0049763909,0.2972848713,-0.2908039093,-0.2138978094,0.1408780962,0.091919668,-0.2211559862,-0.205946371,0.1534100026,0.3134850562,0.2427745014,-0.0770800635,-0.2891998589,0.1630904824,0.2696890235,-0.2634699345,-0.1851076335,0.2320987582,0.0784944147,-0.0356946923,-0.3901937306,0.2268241793,-0.2439683229,-0.1308945715,-0.0936571956,0.0231815316,-0.4545505345,-0.262798816,0.2949182689,-0.0049352744,-0.1328277886,0.0112286191,-0.1847508252,0.2601161003,0.0009451048,-0.0335316882,-0.2378202379,0.1849799454,-0.1797088087,-0.2675313652,0.2869929969,0.2790139914,0.1291242242,-0.1447089165,-0.157281667,0.2860490382,0.01248312,0.41312778,-0.2609884441,-0.2436141074,0.5184406638,-0.5575979352,0.0700042397,0.4544993341,0.1927979589,-0.1603338122,0.243435055,0.2702005506,-0.0523837693,-0.054996945,0.0412183218,-0.1327365488,0.1504889429,0.0241405908,0.0469062366,-0.1603646725,0.3626900911,0.6080385447,0.2547626793,-0.0487532355,-0.1286948621,0.1122962534,0.4738336802,-0.010348795,-0.0013956904,0.0326432958,0.0848798305,-0.1663371325,0.1038350016,0.050344225,0.1257792264,-0.0344345793,-0.2082009315,0.2009344399,-0.0656058043,-0.2105448842,0.1600576788,-0.1011009738,-0.0115903672,0.0224148184,0.1100702882,0.0249632355,-0.3192924559,-0.1397740245,0.0235098451,0.0939735919,-0.2500429451,0.2658503652,-0.4543836415,-0.4244672954,-0.1767519563,0.0882067382,-0.2056944966,-0.0796077102,0.0103617562,-0.0777699351,0.0670166165,0.2921089828,-0.1647594422,-0.0042428644,-0.1235999987,-0.0989780948,-0.0117906826,-0.1825900376,-0.2122241855,0.1677803546,0.4413799942,-0.0398410149,0.1024440229,-0.3816798031,0.0646965057,0.2964733243,-0.3084105253,0.2883993983,-0.2581631541,0.1106206998,-0.0010713122,0.0948402807,-0.0560660399,-0.3148552179,0.2171664089,-0.1198956147,-0.0971306637,0.1169732958,-0.0459102392,-0.1961919665,-0.2785631418,0.0642169341,-0.441055119,-0.3425576687,-0.2023861259,0.1259260923,0.0929157734,0.415661633,0.2045500875,0.0987194106,0.1759499758,-0.0851396099,-0.1331690401,-0.3334023356,0.2693322003,-0.3517205119,-0.4344560206,-0.0000725311,-0.0121584991,0.5023617148,-0.1869822741,-0.2606030405,-0.1026283726,-0.0670835301,0.0555351451,-0.0144215785,0.1379838437,-0.0324569158,-0.0005478442,-0.0640170276,-0.1284192353,0.0312203113,-0.0455351882,0.0314098448,0.3433048427,-0.0890802518,0.1688875556,-0.2043135017,0.0703647956,0.169530645,0.0827557668,0.6934217811,-0.0739770532,0.4205706418,-0.1995118409,-0.0481585786,0.1415821463,-0.12069235,-0.0635125339,0.1851137131,-0.0492119044,-0.2864255607,-0.333673358,0.0551530719,-0.2285636216,-0.217108801,0.0999125764,0.1652230769,-0.0143466648,0.0129885841,0.1171453819,0.0883735642,-0.2286766022,-0.0144519089,0.1805531681,-0.1751569062,0.1710319221,-0.2869604826,-0.3914885223,-0.191530928,0.0800630525,0.174543947,0.2313801795,-0.080242455,0.0730666518,0.1344737411,-0.1544647962,0.1718646437,0.1257071346,0.0307810958,0.0954240784,0.109988302,0.0595229305,-0.0545947365,-0.2330917269,0.124798581,0.1994679421,-0.2205104679,-0.1688950062,0.0259046834,0.0339105763,0.3165860176,-0.0589165464,0.1562267095,-0.4766590893,-0.6225417256,-0.1087163165,-0.1997411996,0.2200851291,0.2296116799,-0.1289592236,-0.0488642119,0.043868959,-0.2837002277,0.2512353361,0.5140666366,0.4233029783,0.1211608127,-0.0226252209,-0.1903211623,-0.1330409348,0.1896959543,0.7450622916,0.0363705382,-0.2097330093,0.1391443908,-0.1770808697,0.2507608831,0.2999685705,-0.0407048538,-0.1753133237,-0.1670929641,-0.0752964914,-0.1916790903,0.2307228148,0.2340694517,0.0813337564,-0.5356986523,-0.448048681,0.5573652387,-0.050416857,0.1726907194,0.1419838071,-0.0903694704,-0.3141259551,0.2411754876,0.0783977732,0.9852373004,0.2112350315,0.2190216035,0.3602869511,0.1238727346,0.1683176756,-0.0752917975,0.1940020174,-0.3183617592,-0.147491321,0.0280084684,0.150052309,0.1256140769,0.0791484043,-0.3388420045,0.1087125018,-0.0423833765,-0.2682104409,0.10928718,0.2032420337,-0.2257400006,-0.1824253201,-0.6316775084,0.2712887526,-0.0699543655,0.0368465744,-0.1795902848,-0.2181740105,-0.0113351848,-0.3394553661,-0.1856632382,0.3587397635,-0.4648076594,0.1656633019,-0.1256934106,-0.660346508,0.0589871034,0.1018798277,0.1422085464,0.1705820858,-0.031631425,0.2886002362,0.3289836049,-0.0333039276,0.0538527481,0.1803847849,0.1164435223,-0.2379673719,-0.0408893488,0.2168306112,0.1136522889,-0.0507298261,0.274654597,-0.0139086675,-0.1687868834,-0.042547565,-0.2854605913,-0.1196893007,0.1908654273,-0.0586594082,0.2239396125,-0.0279350169,-0.2230914533,-0.0733639821,0.1914937496,-0.2924984097,-0.164585948,0.0671759918,-0.0564565919,-0.1051125526,0.4461648762,0.1842793375,-0.2075339258,-0.1513286531,0.2436147034,-0.0754422694,-0.538685441,-0.0066284998,0.1410817057,-0.1534482688,0.0034851271,0.6803991199,-0.0269048493,-0.0568693131,0.1441957802,-0.3272307515,-0.3443568945,0.015019821,-0.1162098348,0.074912332,0.0879411921,0.3469924033,0.0539475158,-0.0673552081,-0.3985676169,0.06626793,-0.142666921,0.0488451086,0.1289657801,-0.1215144917,0.1071466506,0.1047938913,0.2029962987,0.3219655752,-0.0979110971,-0.2832179368,-0.1359742433,0.0565714948,0.0017054655,-0.3379487395,-0.0889106765,0.0667092279,-0.08586099,0.0071735587,0.1423536539,0.0524303354,0.0931169391,0.1324237436,-0.2461743355,0.1586414129,0.1053987518,0.1628994942,-0.0696339607,0.5597633123,0.1636885256,0.062875472,0.005984996,-0.0253742598,-0.0447147936,0.0505549759,-0.2376691699,0.106227845,0.1044628248,-0.359746784,0.4122332036,0.0589301027,-0.0163107459,0.4570335746,-0.2496934682,-0.0978301913,0.3505036533,0.2655173838,-0.3189133704,0.0738120526,-0.2843265831,-0.0188764408,-0.04210658,0.3487904072,0.2089370489,-0.1145657152,0.2774545252,-0.0404063836,0.3768453002,0.1393831521,0.3384260833,0.3199067712,-0.0633219555,0.1767777503,-0.1449131519,0.1443767399,-0.0619793423,-0.006590677,-0.6218464375,0.156592384,0.0202316791,0.0615851134,0.0130636469,-0.2305778265,-0.1371741742,-0.0970709622,-0.107970655,0.3458487093,0.1533379257,0.4684987962,-0.3881831169,-0.0618439727,-0.0999525487,0.2083484381,-0.1023447961,0.0380977914,0.0167677943,-0.0260502789,-0.3448731899,0.1241511628,-0.1106198356,0.1190149486,-0.0481748283,0.3128437698,-0.1080418453,-0.0022145307,0.0105920602,-0.0133055458,0.0652079955,-0.0903548375,0.5817042589,0.3365058899,0.1659681052,0.2955405414,0.1637489647,0.5992535353,0.4194280505,-0.198505193,0.0197415296,-0.1594896168,0.0386608839,0.0385331698,0.119136706,0.3507337868,0.3548156321,0.2214703262,0.2291055769,-0.2119884044,-0.1589124501,0.2516649663,0.0198069699,0.1084337458,0.1820885539,-0.1276624799,0.2285721004,-0.5063562393,-0.1043687314,-0.4751032293,0.2843633294,0.3184300959,0.0251297336,0.0145884352,-0.1881729364,0.1490680575,0.0117615405,0.545556128,0.2461996973,0.0812008008,-0.1053365096,-0.094171308,-0.7023863196,0.2900265157,0.0358975828,-0.4220153391,-0.2040278167,-0.214459911,0.0586932898,0.3786677122,0.0576029681,0.2639373839,0.0873138085,0.0427205563,-0.1206456497,0.0599486865,-0.1676315665,0.0728152916,-0.2522137761,-0.2338365316,0.025285542,-0.3026720881,0.1665266305,-0.2468509078,-0.3811714351,-0.4446495473,0.0948226899,0.2352760732,0.1897628009,0.3141135275,-0.2809827924,-0.2891317904,-0.0594042055,-0.2559399903,-0.0849945769,0.3175151944,0.1799068451,0.1395936608,0.1158799008,0.1496804357,-0.453995645,0.369181633,-0.0407513529,-0.150304243,-0.2544985116,0.0400625132,0.0016038356,-0.0154122207,0.1373398453,0.370809108,-0.043660678,0.1041941568,-0.1964035481,-0.4943504333,0.2880335152,-0.4873786569,-0.1555866152,0.146310702,0.1133800521,-0.1505159438,0.0910948291,-0.5280274749,0.236716345,-0.0112893702,-0.148549214,-0.2645401955,0.2816722095,-0.2627762854,-0.0958281755,-0.0795658827,0.323029846,0.1931323707,-0.1097927168,0.0396768674,-0.2406678796]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/345","title":"Supporting documents in ELI5","comments":"Hi @saverymax ! For licensing reasons, the original team was unable to release pre-processed CommonCrawl documents. Instead, they provided a script to re-create them from a CommonCrawl dump, but it unfortunately requires access to a medium-large size cluster:\r\nhttps:\/\/github.com\/facebookresearch\/ELI5#downloading-support-documents-from-the-commoncrawl\r\n\r\nIn order to make the task accessible to people who may not have access to this kind of infrastructure, we suggest to use Wikipedia as a knowledge source rather than the full CommonCrawl. The following blog post shows how you can create Wikipedia support documents and get a performance that is on par with a system that uses CommonCrawl pages.\r\nhttps:\/\/yjernite.github.io\/lfqa.html#task_description\r\n\r\nHope that helps, using ElasticSearch to index Wiki40b and create the documents should take about 4 hours. Let us know if you have any trouble with the blog post though!","body":"I was attempting to use the ELI5 dataset, when I realized that huggingface does not provide the supporting documents (the source documents from the common crawl). Without the supporting documents, this makes the dataset about as useful for my project as a block of cheese, or some other more apt metaphor. According to facebook, the entire document collection is quite large. However, it would still be helpful to at least include a subset of the supporting documents i.e., having some data is better than having a block of cheese, in my case at least.\r\n\r\nIf you choose not to include them, it would be helpful to have documentation mentioning this specifically. It is especially confusing because the hf nlp ELI5 dataset has the key `'document'` but there are no documents to be found :(","comment_length":130,"text":"Supporting documents in ELI5\nI was attempting to use the ELI5 dataset, when I realized that huggingface does not provide the supporting documents (the source documents from the common crawl). Without the supporting documents, this makes the dataset about as useful for my project as a block of cheese, or some other more apt metaphor. According to facebook, the entire document collection is quite large. However, it would still be helpful to at least include a subset of the supporting documents i.e., having some data is better than having a block of cheese, in my case at least.\r\n\r\nIf you choose not to include them, it would be helpful to have documentation mentioning this specifically. It is especially confusing because the hf nlp ELI5 dataset has the key `'document'` but there are no documents to be found :(\nHi @saverymax ! For licensing reasons, the original team was unable to release pre-processed CommonCrawl documents. Instead, they provided a script to re-create them from a CommonCrawl dump, but it unfortunately requires access to a medium-large size cluster:\r\nhttps:\/\/github.com\/facebookresearch\/ELI5#downloading-support-documents-from-the-commoncrawl\r\n\r\nIn order to make the task accessible to people who may not have access to this kind of infrastructure, we suggest to use Wikipedia as a knowledge source rather than the full CommonCrawl. The following blog post shows how you can create Wikipedia support documents and get a performance that is on par with a system that uses CommonCrawl pages.\r\nhttps:\/\/yjernite.github.io\/lfqa.html#task_description\r\n\r\nHope that helps, using ElasticSearch to index Wiki40b and create the documents should take about 4 hours. Let us know if you have any trouble with the blog post though!","embeddings":[0.0806761757,-0.1414157599,-0.1308178753,-0.09870141,-0.3395778239,-0.0216143336,-0.2173523605,0.2103035599,0.050301034,-0.0259131584,0.1280906349,-0.2411950082,0.1008691564,0.1427407563,-0.2445807904,0.1461946964,0.0276544765,0.2660853565,0.2416867018,-0.1947776228,-0.0178429596,0.1976368874,0.1432036906,-0.1716149896,-0.3346624076,0.2507556081,-0.2730874717,0.1042412817,-0.237261638,-0.0662403479,0.1023939028,0.2378306091,-0.036256399,-0.0565391257,-0.0001082366,0.0151389698,0.0998261422,-0.1303936988,-0.0717165768,0.0352354124,0.3620178401,-0.4477023184,0.1384598911,-0.1620794684,-0.3059388995,-0.1202642322,0.2798631489,0.3556277752,0.3528578579,-0.004643877,0.1948395669,0.013586794,0.1496860087,0.0355122834,0.3426299393,0.3699345589,-0.2522276342,-0.262568593,0.2649269104,-0.2322939336,0.0697334111,0.3647731245,0.0875473768,-0.1129977629,0.3730433881,0.1021438241,0.0469903685,-0.613278985,0.2595945895,0.251003474,0.3036833107,0.2703653574,-0.1207334548,-0.3047217429,-0.2468187362,0.285892874,0.195596233,0.2177018672,-0.1502867788,0.15116404,0.0314800292,-0.1184746474,-0.018271666,0.3093078136,0.1452437937,0.1628545076,-0.0881980807,-0.1460662037,0.2186290026,-0.1449057013,-0.6590011716,0.1220768467,0.2056598514,0.134045437,0.048492562,-0.1340939105,0.1747570187,0.1377401203,0.5864340663,0.0596163012,-0.0866412669,0.127357617,-0.0404516757,0.0465862304,0.088900201,0.0602401681,0.0087468158,-0.2209196389,0.2579364777,0.1123382151,-0.0838561356,0.0046388875,0.0627129897,0.3078454733,-0.3241544962,-0.5395190716,0.120175004,-0.1676558703,-0.1507179439,0.1569003314,-0.1685593724,0.0254629962,-0.223440364,0.5335275531,0.0401802175,-0.0055817738,-0.0136975134,0.038695626,-0.1921836734,-0.1128755286,-0.0731188506,-0.1688325703,-0.1097167656,0.1998763829,0.1995472163,-0.3310860693,0.5541297197,-0.2196207196,0.2041632682,0.0977548063,0.1932164282,0.0301174037,0.253544122,0.3538282216,0.1545122564,0.0394503847,-0.0487941206,-0.1594578773,-0.2483039796,-0.247927621,0.069725357,-0.3629739881,-0.4516446888,0.2046125382,0.1505705267,0.2050661743,0.0231000446,0.5333048105,-0.3146410584,-0.10384579,0.1686959267,0.3648088276,-0.3038465977,-0.1020546108,0.4092980027,0.2519221604,0.4009886086,-0.3729685843,-0.0606633127,0.3335617185,-0.0977839157,0.3365208507,-0.0658703297,-0.2021239847,-0.033272773,0.3529548943,0.2694994807,-0.242778033,-0.5641846061,0.2591613829,-0.002454015,0.2508822978,0.0845526829,0.0684894547,0.0826148391,-0.2189555317,0.0249457024,0.2849489152,0.0519593544,0.0720635951,-0.3615594208,-0.4118731618,0.0270244945,0.143393442,-0.0667524934,-0.2504672706,0.0029691877,-0.0645665601,-0.0254445206,-0.0772941783,0.152787447,0.3830137551,0.1348627508,0.1990221888,0.1052660868,-0.0190326814,-0.3713848591,0.2328276932,-0.0967208967,0.4336290061,0.1736372411,-0.3850110471,-0.0803899765,-0.0744089335,-0.105005689,-0.4436501265,0.2301823944,-0.1027259529,0.2857616842,0.0918073878,-0.0282305367,-0.286475122,-0.2602204978,0.0710940734,-0.1651402712,0.3315623999,-0.3213315308,0.0342530198,0.0782489479,0.1614308506,-0.124255687,0.1105072349,0.1942817867,-0.0071293996,-0.3551361263,-0.0379281119,0.4002793133,0.7665842175,0.2776437402,-0.3818296492,0.1005618647,0.1995650083,-0.0353210419,-0.1732577682,-0.3145941496,0.0889273435,0.278570503,0.3200398982,0.1890916228,-0.0571337007,0.0382560976,0.0903183743,-0.1557933241,-0.1499132663,0.3771613836,-0.1340329051,0.0639237911,0.0019980865,-0.148224771,0.4138783514,0.1723026484,0.1056227162,0.1920792609,0.5266599655,-0.1382295191,-0.4718621373,-0.1715259701,-0.3971883059,0.0164452065,0.2203441113,0.1913430542,0.0278093237,0.1325486749,0.0655028746,0.1050011143,0.0083910571,-0.0028981415,0.0728535429,0.2263055742,0.2209694684,-0.2642669976,-0.1024458483,-0.3160074651,-0.079657793,0.1014831662,-0.0092377923,0.2213156968,-0.2507482767,0.2634441555,-0.0781887695,-0.414742142,-0.3153249025,0.2948235869,0.113560386,-0.1521184295,0.0933333561,-0.0359079204,0.1366750598,-0.1347661465,0.2000321746,-0.4448846579,-0.1261646748,-0.1927219033,0.0864221007,-0.0453944877,0.4896547794,0.1996760964,-0.191867426,-0.0891262367,-0.6268857718,-0.6305283904,0.2847003639,-0.1327684671,0.1055973545,0.2470073849,0.3481204212,-0.3925293088,-0.4493710101,0.2627792656,-0.4444777668,-0.1999738365,-0.2748990357,-0.1903129518,0.022515811,0.0493832305,-0.1188695207,-0.2579317689,-0.2666805983,0.3291163146,0.2762427926,-0.0609605238,-0.0458911248,0.0205528494,0.1251767427,0.2258081734,0.2626574636,-0.2936744094,-0.0474499576,0.2918271422,-0.3029043972,-0.2832897305,-0.037007954,-0.1125949994,-0.2324836552,-0.18669568,-0.3569921851,-0.3837314844,-0.1370197833,0.054883413,-0.0264961943,0.0882039368,0.1968396157,-0.0107379574,-0.2033339292,0.1613602638,-0.0292122066,-0.0383645259,0.0866158679,0.0075273709,-0.2022761106,0.2060182989,0.0629343763,0.7775626183,0.2560121715,0.0957140848,0.3554350734,0.0965010971,0.3023887575,-0.3307038844,-0.0481676236,0.0820795968,0.0261234082,0.1000516564,0.5637340546,0.2327352613,-0.1557636708,-0.1026142463,-0.52257514,-0.3188152313,-0.2693008482,0.2323869765,0.2323057055,0.2514648139,-0.0427640676,0.0313500091,0.0448824763,-0.4328540266,0.1710645556,0.1494749635,0.0656210557,0.067215696,-0.2274291068,0.0286750291,-0.9157403708,0.2903389633,0.1272510737,-0.2261869758,-0.2038739026,0.0086204028,0.167307362,-0.1656234711,0.3456693888,-0.2442613393,-0.2296356112,-0.2217074335,-0.0772225186,-0.1184782535,0.0776623264,0.0778403804,0.0282881893,0.3759855628,-0.0090696504,-0.6138387918,-0.167189911,0.2821722627,-0.0146173388,-0.0445770845,0.0969725847,-0.3596684337,-0.4528429508,-0.5167990327,-0.1217091009,0.2898416221,-0.1836678386,-0.1271049976,0.1153054386,0.1621761173,0.1249370575,0.1095976308,0.1497636139,0.0782529861,0.0392141864,0.0177037399,0.4854711592,0.162313059,0.1095933616,0.4987751842,0.0102823833,-0.0290291887,0.0863296017,0.0926338136,0.0507366024,0.3944362998,0.4012108445,0.0613245852,0.2153005898,0.1702982336,-0.1811871529,-0.0273428317,0.4937495887,-0.1580030322,-0.1015143692,0.0634479523,0.4403325021,-0.0228714366,0.0520248041,-0.1205277517,0.8181396127,0.0034936273,0.6203841567,0.02450886,1.0834466219,0.0712132007,0.2355078161,-0.0662083551,0.1022235528,0.4590815902,-0.5924046636,-0.0719935969,-0.1716339141,-0.3579326868,-0.0282045212,0.2058293223,-0.2424095124,0.0943545848,-0.2666331828,0.1641512662,0.1665958315,0.0792628899,-0.1323837042,0.5353415608,-0.2092642635,-0.2426697165,-0.2931995094,0.1606056541,0.1370087564,0.4707176685,-0.0683386847,-0.0239954609,-0.0310567133,-0.1576667428,-0.4364046156,-0.0868464038,-0.0571709238,0.0721726939,0.2667095661,-0.0620512702,0.0479139164,0.1726251096,0.1590501517,-0.1449038535,-0.4947530329,0.1609764248,-0.0852394551,-0.2354751825,0.221770823,-0.3119463325,0.3180812001,-0.157039836,-0.2050399184,0.242334336,-0.1365594566,-0.0473492742,-0.1830668151,-0.266245544,-0.283218056,-0.1319664717,-0.0575913638,0.2548996508,-0.5066122413,-0.1073727086,0.1360843629,0.0619507954,0.1497884244,-0.1424946934,0.3879220486,-0.0254303943,-0.164731279,0.4760460556,-0.0733071417,-0.1303655505,0.3072896302,0.0134335691,-0.2064886391,-0.0190460514,-0.1087079495,0.2752626836,-0.282153368,-0.1286257803,0.1442892104,-0.1705996692,0.0885861814,0.7039009333,0.1237256378,-0.1236470491,-0.1176973209,-0.0933519974,0.1011834592,0.4985333681,0.0505791157,0.0725107193,0.0642409101,-0.1272994429,0.2589456141,0.2433390021,-0.3592049479,-0.0717600435,-0.4390507936,0.0184076838,-0.1205719262,-0.1613070965,0.4346778989,0.2047705352,0.0707754567,0.0910657793,-0.0933141857,-0.1808376014,0.0803188533,0.1305054426,0.1066498756,0.020297138,-0.1118559688,0.0898529887,-0.1579394937,-0.3185772896,0.0888937265,-0.1715652347,0.0360235348,-0.0528546348,-0.1689074486,0.0717952326,-0.0415923595,-0.0718804598,0.1163448021,0.4744982123,0.0726780891,-0.0939534381,-0.2063888758,-0.1261648089,0.261061281,0.2690448761,0.2563708723,-0.2336864173,0.2742327154,0.1976024806,-0.2272613049,0.017748069,0.0360476486,0.1653280109,-0.1373502463,0.1518940926,0.1155493334,0.2441235185,-0.2058787346,-0.091357097,0.3557405174,-0.2692943215,0.3280307055,0.1972301602,0.6575436592,-0.1029172242,0.2686088383,0.1029348597,0.1431605518,-0.3198325038,-0.2167585194,-0.3446303904,0.064380832,-0.0358222835,-0.1502176076,0.0252593886,0.6202681661,-0.1843103617,0.0242386609,0.2300519645,-0.2149218023,0.4202449322,0.2570254505,-0.0216611084,0.1826197058,0.1664552689,0.2891645432,0.3017503321,-0.0999804065,0.0564863943,0.0933226719,-0.0102173276,-0.4102499783,0.3412421346,0.0440335348,-0.0449781418,0.2121960223,-0.297798872,-0.1784314215,0.1179586276,-0.0404759385,-0.1816560626,-0.248512283,-0.1287418455,0.0907335058,-0.0210818779,0.2186365575,-0.12473315,0.6175799966,-0.0710847154,0.1010066643,0.1910881996,-0.1985791177,-0.0609299392,0.3612090349,0.2990721464,-0.0895740315,-0.442096442,0.1821316332,0.063787207,-0.0583394803,-0.1847285628,0.0843779296,0.2508333921,-0.3955073059,0.0231666453,0.0982852355,-0.3194877207,-0.5002756715,-0.2329954952,0.3162249029,0.062484134,-0.4794466197,-0.1802380383,-0.029405836,-0.0606039762,0.0291258488,-0.3866060078,0.1661674231,-0.159111768,0.0206479542,0.0503566638,-0.2859903872,0.0850376338,0.0807042792,0.4946759939,0.3730095327,0.1127117798,0.0598622635,-0.431278795,-0.1302331835,0.1595833153,-0.2466768026,-0.0424954444,0.4238420725,0.3498511612,0.2614299655,0.3045891523,0.029717546,-0.2874191403,-0.2152039856,-0.0699589923,-0.0957851931,0.084046796,-0.2433263361,0.1507636756,0.0208424591,0.0374238007,0.0005401623,-0.2379784733,0.0414123721,-0.2940924764,-0.0217357464,-0.1969223171,0.3018136919,0.0099941967,0.2058706582,-0.1764591932,-0.2507236302,-0.4384669363,0.0337839872,0.1012849957,0.0593898892,0.13748537,0.0188350752,-0.0253789779,-0.0652534068,-0.1125294268,-0.0281356294,0.1453640759,0.242572993,0.1292925477,-0.2725088298,-0.0987302214,-0.1670645177,0.1415548474,0.0640585795,0.4082987309,-0.1655519456,-0.3016099036,-0.2824160457,-0.1516109854,0.2724377513,-0.1437879205,0.1419830471,-0.1527414769,0.1226041764,0.4253350496,0.097856991,-0.3977632821,-0.0714544803,0.1101638824,0.3151708543,-0.262743026,-0.2627634406,0.1294181496,-0.169670552,-0.2428094745,-0.0516266599,-0.063453868,-0.0654650182,-0.2057819217,-0.1455370635]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/345","title":"Supporting documents in ELI5","comments":"Hi, thanks for the quick response. The blog post is quite an interesting working example, thanks for sharing it.\r\nTwo follow-up points\/questions about my original question:\r\n\r\n1. Yes, I read that the facebook team could not share the CommonCrawl b\/c of licensing reasons. They state \"No, we are not allowed to host processed Reddit or CommonCrawl data,\" which indicates they could also not share the Reddit data for licensing reasons. But it seems that HuggingFace is able to share the Reddit data, so why not a subset of CommonCrawl?\r\n\r\n2. Thanks for the suggestion about ElasticSearch and Wiki40b. This is good to know about performance. I definitely could do the indexing and querying myself. What I like about the ELI5 dataset though, at least what is suggested by the paper, is that to create the dataset they had already selected the top 100 web sources and made a single support document from those. Though it doesn't appear to be too sophisticated an approach, having a single support document pre-computed (without having to run the facebook code or a replacement with another dataset) is super useful for my work, especially since I'm not working on developing the latest and greatest retrieval model. Of course, I don't expect HF NLP datasets to be perfectly tailored to my use-case. I know there is overhead to any project, I'm just illustrating a use-case of ELI5 which is not possible with the data provided as-is. If it's for licensing reasons, that is perfectly acceptable a reason, and I appreciate your response.","body":"I was attempting to use the ELI5 dataset, when I realized that huggingface does not provide the supporting documents (the source documents from the common crawl). Without the supporting documents, this makes the dataset about as useful for my project as a block of cheese, or some other more apt metaphor. According to facebook, the entire document collection is quite large. However, it would still be helpful to at least include a subset of the supporting documents i.e., having some data is better than having a block of cheese, in my case at least.\r\n\r\nIf you choose not to include them, it would be helpful to have documentation mentioning this specifically. It is especially confusing because the hf nlp ELI5 dataset has the key `'document'` but there are no documents to be found :(","comment_length":256,"text":"Supporting documents in ELI5\nI was attempting to use the ELI5 dataset, when I realized that huggingface does not provide the supporting documents (the source documents from the common crawl). Without the supporting documents, this makes the dataset about as useful for my project as a block of cheese, or some other more apt metaphor. According to facebook, the entire document collection is quite large. However, it would still be helpful to at least include a subset of the supporting documents i.e., having some data is better than having a block of cheese, in my case at least.\r\n\r\nIf you choose not to include them, it would be helpful to have documentation mentioning this specifically. It is especially confusing because the hf nlp ELI5 dataset has the key `'document'` but there are no documents to be found :(\nHi, thanks for the quick response. The blog post is quite an interesting working example, thanks for sharing it.\r\nTwo follow-up points\/questions about my original question:\r\n\r\n1. Yes, I read that the facebook team could not share the CommonCrawl b\/c of licensing reasons. They state \"No, we are not allowed to host processed Reddit or CommonCrawl data,\" which indicates they could also not share the Reddit data for licensing reasons. But it seems that HuggingFace is able to share the Reddit data, so why not a subset of CommonCrawl?\r\n\r\n2. Thanks for the suggestion about ElasticSearch and Wiki40b. This is good to know about performance. I definitely could do the indexing and querying myself. What I like about the ELI5 dataset though, at least what is suggested by the paper, is that to create the dataset they had already selected the top 100 web sources and made a single support document from those. Though it doesn't appear to be too sophisticated an approach, having a single support document pre-computed (without having to run the facebook code or a replacement with another dataset) is super useful for my work, especially since I'm not working on developing the latest and greatest retrieval model. Of course, I don't expect HF NLP datasets to be perfectly tailored to my use-case. I know there is overhead to any project, I'm just illustrating a use-case of ELI5 which is not possible with the data provided as-is. If it's for licensing reasons, that is perfectly acceptable a reason, and I appreciate your response.","embeddings":[0.1618807167,-0.0452031456,-0.0387861505,-0.0477261692,-0.3124100566,0.0002395466,-0.1353973299,0.1302300096,-0.0219204035,-0.0501846857,0.0941168517,-0.2651159167,0.1130612791,0.1443804055,-0.1669064164,0.1842308939,0.0264950655,0.2427613288,0.3465193808,-0.3121981323,0.0172460712,0.2343271077,0.1443893909,-0.1357314438,-0.3666524291,0.28333202,-0.29987818,0.1768079549,-0.2238978297,-0.0061549218,0.0662747547,0.2734149992,0.0293229446,-0.2047086656,-0.000114399,0.0644626245,-0.0929929838,-0.1426993608,-0.1507161111,-0.1806523949,0.2431360483,-0.3142079413,-0.0273024756,-0.0974219441,-0.1486665159,-0.202641353,0.2346536964,0.2599248886,0.3535690904,-0.0374876298,0.1049089432,0.0820606202,0.0161260869,0.1020364538,0.1941461414,0.3405364156,-0.2443336695,-0.0740674809,0.3869939446,-0.202209726,0.0654283911,0.2211506218,0.0566235259,-0.1823208183,0.2999623716,0.0382442065,0.0203543436,-0.6953431368,0.2635998726,0.3578397334,0.4532786012,0.283123523,-0.2213172764,-0.4320504665,-0.1177456304,0.2731931806,-0.0302476678,0.3697188497,-0.0638160259,0.3528544009,0.0016578792,-0.2023566216,-0.0750342309,0.3632193506,0.2741084099,0.056665618,-0.1507907361,-0.0997989476,0.1249804646,0.0599622242,-0.4961788058,-0.0230370481,0.1168772578,-0.0398264416,0.1816927195,-0.1658809334,0.0937453657,0.3632128835,0.6906947494,0.2466049641,0.0316821747,-0.0119366283,-0.2202199697,0.0572159067,0.2629973888,-0.0168864857,-0.0620113946,-0.1783059835,0.1461059302,-0.0312419031,0.1357302517,-0.0091118775,-0.0422690026,0.5626630187,-0.5317935944,-0.5932882428,0.2085096389,-0.2363729179,-0.1962751895,0.1573073119,-0.2473305762,-0.0459486768,-0.1888718009,0.43571347,0.1879499108,0.0037178257,-0.3072887957,0.1680948287,0.0074788956,-0.3712285161,-0.0267664976,-0.1037182957,-0.1844659448,0.1345674396,0.1396351904,-0.3869445026,0.6252250075,-0.2365910709,0.347804606,0.1796341985,0.1624160111,-0.0185363404,0.3222464919,0.3538758755,0.1069955528,0.0527894758,0.0617658943,-0.0122318016,-0.2694570422,-0.2196465433,-0.0066613117,-0.456005007,-0.4140790105,0.0596289001,0.0276856106,0.2341445684,0.1031522974,0.513455689,-0.4354706407,-0.1399757564,0.2109805793,0.2206294537,-0.3961333632,-0.2374761701,0.3468028307,0.2940281332,0.3394902349,-0.4009534419,-0.0533855036,0.2656393647,-0.1998755634,0.3898680806,-0.2220280468,-0.2204244137,-0.0046833977,0.3456551135,0.3490616083,-0.1972573102,-0.7282791138,0.2450958639,0.1184997559,0.3259021938,0.1589938551,-0.0551896207,0.0623803996,-0.3640515804,-0.043246977,0.1965417415,-0.04358707,-0.1194300279,-0.2560020685,-0.4202414453,-0.0762731656,0.1884012222,-0.1058001444,-0.1561291069,0.0355004296,-0.0632957593,0.008365415,-0.0580607727,0.1475289613,0.3281921446,0.1072339192,0.159090206,0.0168096703,0.0722490102,-0.3731556535,0.3071223497,-0.0141973868,0.3480271399,0.3044205904,-0.3181044757,0.0587491319,-0.0931872576,-0.0007662947,-0.6210433841,0.1355814934,-0.0142499907,0.1797038913,0.0173147321,-0.2134844363,-0.0974055231,-0.1941886991,0.1273107678,-0.1348647773,0.3530066609,-0.2602706552,0.0966936871,0.0792050138,0.4140844643,-0.1390437633,0.0689707175,0.2654939592,-0.0479334146,-0.3764588535,-0.0900990143,0.2554413974,0.8737131953,0.3631456494,-0.3448274136,0.2278608829,0.066680409,-0.0577497259,-0.1940209866,-0.212673828,0.0391613171,0.1341947019,0.3453151286,0.3507530987,0.0651238263,-0.0663319826,0.0633670911,-0.2704552412,-0.214571476,0.4714906514,-0.102643691,0.2174859494,0.0538595393,-0.2446395755,0.4113785625,0.2338645756,0.165873453,-0.0037701109,0.4241563678,-0.2534435689,-0.3952467442,0.0571383052,-0.183164537,0.2414886951,0.2258030474,0.2277128696,0.0635995716,0.1560093015,0.0568045117,0.017448796,0.1503199488,-0.0517089963,0.0024007715,0.1140671596,0.3764139116,-0.2117667347,-0.2193944007,-0.3091649413,-0.278567493,0.0724299774,0.0705558062,0.1820521951,-0.3416146636,0.1765885949,-0.0956139043,-0.4134987891,-0.2661287189,0.2347912937,0.1461952627,-0.0589662455,-0.007164279,-0.2509242594,0.2849396169,-0.480771035,0.3026829064,-0.3664853275,-0.0576579049,-0.2014650553,0.0328033753,0.0083475029,0.3727059066,0.2562305927,-0.0921007767,-0.1207940727,-0.5289492607,-0.4521756768,0.3688605726,-0.2268683314,0.1133744866,0.1305222362,0.1967614889,-0.3303303123,-0.264618814,0.1088884845,-0.3639263213,-0.0934845358,-0.3351455033,-0.0973831713,-0.0986822918,-0.0030945179,-0.0411960892,-0.3535021544,-0.2808041573,0.3430304825,0.2254042476,-0.1302045584,0.0172077008,-0.0800358802,0.1123522669,0.1259935349,0.3323071599,-0.2970348895,-0.1613458693,0.1960982829,-0.2074230164,-0.2651272416,-0.0386919864,-0.0539244488,-0.0615293831,-0.100594312,-0.425477922,-0.3604826629,0.0556921661,-0.0251169037,-0.0703638494,0.0615719445,0.0977726057,0.0125412596,-0.0685769394,0.0065078842,0.00510677,-0.1429712772,0.1419217736,0.0686362609,-0.0926286355,0.1250013858,0.2023942322,0.7989224195,0.4217329025,0.1286455244,0.4140589237,0.2201206833,0.4891663492,-0.2331579179,-0.0684255064,0.1044610292,-0.0167270079,0.287329495,0.5585052967,0.3046330512,0.0850377679,-0.0615950897,-0.475753963,-0.2863051295,-0.1630813181,0.2455817163,0.0341197588,0.2572931051,0.0019510248,-0.022763297,0.1661116779,-0.3102349341,0.0538970977,0.1662834585,0.2355100811,0.1411301941,-0.3051784337,0.1529219896,-0.8590841889,0.2311124057,-0.0189008955,-0.187437579,-0.2980825007,0.0848464742,0.1466863602,0.0557167083,0.323502928,-0.3109868169,-0.3885932863,-0.2066270709,-0.0179771166,-0.0430691727,0.0947535336,0.1059954762,0.0098154768,0.408177048,-0.1639486849,-0.4936892092,-0.2149727941,0.2640773654,-0.0316419639,0.0092504937,0.1360560507,-0.1757677346,-0.2938088775,-0.5504943132,-0.4285238981,0.2547289133,-0.3438403606,-0.219306156,0.2031396925,0.2557872534,0.0077915792,0.2721949518,0.1839651763,-0.0555270165,-0.0856643468,-0.0926734358,0.5099778175,0.3596833944,0.2354166508,0.7077587247,0.0179830547,-0.0893132985,0.1664328426,0.052218467,0.2376980484,0.4825991988,0.3366220891,0.0189697798,0.0892785639,-0.007108428,-0.1003565937,-0.0050263796,0.4554409087,-0.3331573308,-0.1282137632,0.0333348177,0.3052026033,-0.1238710657,0.0464240313,-0.0847176015,0.9499078393,0.1301208586,0.5122741461,-0.0499970093,1.1526865959,0.2842029333,0.387003988,-0.1136421487,-0.0532836616,0.5818097591,-0.502538085,-0.194201678,-0.0220448691,-0.38685745,-0.1671909243,0.0516408719,-0.2265332788,0.2156896442,-0.0801803693,0.2316945493,0.3886117935,0.0542905144,0.0149285579,0.4949814975,-0.2575530708,-0.1681659967,-0.1999782324,0.0432274379,0.1722919494,0.4088801444,-0.1498887837,-0.1088568345,-0.2044103444,-0.1663981229,-0.5499116182,-0.049784068,0.060107626,-0.061446134,0.3552017808,-0.0797137022,0.3253508806,0.1505557001,0.3010341823,-0.3107363582,-0.5965957642,0.0113912923,-0.0683422759,-0.2154919207,0.2730764151,-0.193450436,0.4823544323,-0.1064095572,-0.1110984758,0.3303718865,-0.0982631072,-0.0512322336,-0.4185140729,-0.2416194081,-0.1881910264,-0.1779618561,-0.0705719218,0.3471911252,-0.4256944358,-0.1267219633,0.0480546765,0.1066824049,0.3316638172,-0.2672721148,0.3429391384,0.0106951604,-0.2360139787,0.3876951933,-0.1587111354,-0.0637343079,0.2717557549,-0.2133662701,-0.1275206208,0.0727553293,-0.106447801,0.2839754522,-0.4165966213,-0.2355314046,0.1129472554,-0.1551657766,-0.1096847877,0.4567147195,0.1899131984,-0.0448394641,-0.204405576,-0.1590964645,0.1070047319,0.5737264752,-0.1819009334,0.0827108547,-0.0194352251,-0.0116757387,0.2233771384,0.1554015428,-0.2522448003,-0.0365175195,-0.3129373193,-0.0768672302,0.0278435517,-0.0764382482,0.3367640078,0.1138090417,-0.0922478214,-0.0516176112,-0.1766045243,-0.0428036079,0.0224406812,0.1930937916,0.3122955263,-0.0885729268,-0.1383562684,0.0868222415,-0.1138550267,-0.3022900522,0.1337784678,-0.2211059928,0.070797354,0.0809220299,-0.1911887527,0.194678545,-0.1025141329,0.0423028544,0.1174937859,0.5127567053,0.2672607899,-0.0269920267,-0.2109585106,-0.1394847333,-0.0087477546,0.2265075743,0.3295251429,-0.2503394485,0.3612289727,0.3472689092,0.0261633582,-0.1014584452,-0.0064436067,0.1815853864,-0.2297015041,0.2761463523,0.2086239457,0.1057318375,-0.1489964128,-0.0837416053,0.3395874798,-0.1748004705,0.2245009392,0.1914620697,0.8282392621,-0.1262876242,0.3334259987,-0.1276555955,0.1897799671,-0.2044483125,-0.1817617565,-0.2316094786,0.0415635817,0.1843354851,-0.1679465473,0.0841623843,0.5764597058,-0.3147503436,-0.0840197802,0.1850130111,-0.2706155479,0.2517425716,0.4075971246,-0.0716698021,0.109772563,0.1970787793,0.4905525744,0.2935710549,-0.1452923268,-0.0370134749,0.1738227457,0.152501896,-0.2693368495,0.163602069,0.1845989823,-0.0508921184,0.1807234883,-0.1903873682,-0.2371996492,0.1965666562,-0.2412137538,-0.2911819816,-0.2178951651,-0.1432462037,-0.0616890192,-0.1196151599,0.2244772166,-0.0884875804,0.3507426083,-0.0996967033,0.1117534116,0.0399553813,-0.1257642061,-0.0004623462,0.3130822778,0.3190271556,-0.1744496673,-0.1170387641,-0.0126744108,0.2129801363,0.0158330128,-0.2076572925,0.1143768504,0.2840118408,-0.2482049316,0.1243272498,-0.0195968729,-0.1539877057,-0.3735148907,-0.1696233302,0.2303642035,-0.1834688634,-0.3113382161,-0.1564998031,-0.1659861505,0.1074691415,0.0604187585,-0.2830290794,0.2151736319,-0.107143946,0.0541044846,0.0130196288,-0.2539302707,0.0379404835,0.1356281042,0.3063026071,0.5627692938,0.109566696,-0.0617809631,-0.398617357,-0.117264621,0.1899101436,-0.2081127018,-0.0411836542,0.3487135768,0.1832495034,0.3635133505,0.3037836254,-0.0953089818,-0.2481217533,-0.2477929592,-0.1816629618,-0.0233041104,-0.0725573525,-0.1085301489,0.2357600182,-0.0107638715,0.2202775478,0.0317005962,-0.1569307894,-0.1226209775,-0.3057942688,0.0225638617,-0.2268680483,0.3308991194,-0.0405952819,0.1599254757,-0.1917667985,-0.2094217986,-0.4069240689,-0.0480840914,0.1296694428,0.0477347858,0.1616849303,0.0094907079,-0.0973561704,-0.275991708,-0.0923778266,0.0149700055,-0.029608231,0.129630506,0.1409839541,-0.3074487448,-0.0239940658,-0.189832136,0.1414301246,0.0891478136,0.3798620105,-0.2511128187,-0.3843094409,-0.260412097,0.0579466261,0.2941581905,-0.1104182303,0.2896015346,-0.0296144299,0.1552997082,0.4543701708,0.0183811951,-0.5039560199,-0.2176716179,0.2137532383,0.1811883003,-0.375446558,-0.2439012676,0.1902193576,-0.1772358567,-0.2124372721,-0.0106056929,-0.1518122405,-0.175859496,0.1100060791,-0.1202374548]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/342","title":"Features should be updated when `map()` changes schema","comments":"`dataset.column_names` are being updated but `dataset.features` aren't indeed...","body":"`dataset.map()` can change the schema and column names.\r\n\r\nWe should update the features in this case (with what is possible to infer).","comment_length":8,"text":"Features should be updated when `map()` changes schema\n`dataset.map()` can change the schema and column names.\r\n\r\nWe should update the features in this case (with what is possible to infer).\n`dataset.column_names` are being updated but `dataset.features` aren't indeed...","embeddings":[-0.0770480782,-0.2360989451,-0.2290511727,0.0304846205,0.0986381397,0.0106040826,0.1780465543,0.1911529005,0.2378466725,0.0060715177,0.1091589183,0.4632224441,0.1449858546,0.2659249008,-0.2699092627,0.0670404509,0.3282714486,0.1196199432,-0.283057183,-0.1341726333,-0.420710057,0.1632417738,-0.2707624733,0.0773729905,-0.1219684333,0.1088311449,-0.1673126966,0.1026935875,-0.1446959972,-0.2429869473,-0.1625219584,0.2959296405,-0.1170351505,0.2372929007,-0.0000943661,0.0167355854,0.3437040746,-0.0619912967,0.0193842016,0.1400664896,-0.4105840027,-0.388384521,-0.1399131566,-0.5432164073,0.092930913,-0.1304043233,0.0381660275,-0.0409838147,0.0052767596,0.0894874036,0.3619425595,0.0741195753,0.1895551831,-0.0582917146,0.2447872758,0.0940426439,-0.0048809978,-0.1493291408,0.0076345052,0.0569487549,-0.0702113956,0.5356238484,-0.0886384994,-0.0882788301,0.3261481225,-0.0158675797,0.1278010607,-0.1783806682,0.2885854244,-0.0791405961,0.5193584561,-0.0156794544,-0.0841236711,0.0038413596,-0.2190068811,-0.0744576752,0.2596774101,-0.3194361031,0.3000381589,0.0411508307,-0.0977554992,-0.2640226781,0.2576452792,0.0535469241,-0.2294223756,0.2294690609,-0.1108021811,0.2212145478,-0.4059657454,-0.3200114071,-0.035173852,-0.2772271335,-0.0941277146,0.069832541,-0.1185719743,-0.1230096444,0.3384835124,-0.043624673,-0.0824966505,-0.1706089824,0.0294015184,0.2242168188,-0.0296533518,0.3963266909,0.2467082739,-0.1610192508,0.2868922055,0.1249845624,0.2491247803,0.093024686,-0.3912891448,-0.0837612301,0.2288098484,0.0574582666,0.1418807805,-0.0732529461,0.4270917773,0.0022735507,0.1255754381,0.0468718708,0.0099787088,-0.1697904617,-0.0735022277,0.153484568,0.1072481349,0.0714612082,-0.1001231745,-0.1181545779,0.0725070983,-0.0148997242,-0.3722537756,-0.1996854693,-0.1115218699,-0.0685570464,-0.1816036105,0.2986560762,0.3043671846,-0.1130127534,-0.3977388144,-0.1195311695,-0.1265436858,-0.1012265012,0.212752685,0.2251388729,-0.3288916945,-0.0666533262,-0.099937439,0.1977325231,-0.0683364421,0.1704471558,-0.2152537256,-0.0565963276,0.182293877,0.3828795552,-0.0004262943,-0.0350002907,0.0793014243,0.2381261289,0.208860904,-0.4478503466,0.1761466712,-0.0307885297,-0.0338070765,-0.170668751,-0.0491490141,0.2712569833,-0.0170965716,-0.0349644609,0.2725909054,-0.0170861222,-0.1412463635,-0.1433736235,-0.1650387794,-0.0821358785,0.0275112968,0.1752649397,0.3578175604,-0.0537007786,-0.0417983867,-0.0090743229,-0.0430976003,-0.1127298027,0.0152471289,0.0193226151,0.0108625274,0.07014478,-0.0485199764,-0.1021741405,-0.0444485806,0.1036682129,-0.2320960462,0.0179353971,0.1130176932,-0.2760978937,0.0754230246,0.2432638705,0.111817494,-0.0410150327,0.1670847833,0.1681001186,0.0239627361,0.2198026031,0.1528150737,0.1464825124,0.1358543485,0.0091582518,-0.4808494449,0.0913112909,0.0479004756,-0.0793273225,-0.0279060341,-0.1454638094,-0.1970927864,-0.0329259522,0.0378784463,0.007116599,0.3582265377,0.2660820782,-0.1405778825,-0.1274635196,-0.1965752691,-0.0748312175,-0.1412631422,0.0607527122,-0.2914860249,0.0941183344,-0.0707261935,0.035837464,-0.2132992595,-0.1354904473,0.1820635945,-0.1846716851,-0.2197284251,0.3758216798,0.1069802642,0.2766827941,0.0178595763,0.2367111295,-0.0244073998,0.2136280388,0.152164042,-0.089203909,0.220142588,-0.0280104298,-0.0989144444,0.1374459863,0.1680658311,0.1611059755,0.0747559518,-0.0418280251,0.200458467,-0.0786003396,-0.1062582955,-0.2781051993,-0.6941021681,0.1848222315,-0.114340663,0.234574154,-0.2496956885,0.1328315437,0.7775987983,0.0964572579,-0.0644418299,-0.1530303061,-0.1139147133,-0.195862636,0.2093822807,-0.0605277792,0.1629344374,0.1792620718,0.1482581645,0.0487456135,0.1896802634,0.0050663562,0.0112030637,0.1875862479,0.1302541196,0.0681225285,0.2982029915,-0.0145162642,-0.2825856507,0.0844384208,0.4057835639,0.0744942501,-0.1669382006,-0.114807494,-0.0274791513,0.0305747222,0.0054266811,0.006861378,-0.0955061242,-0.4356955886,0.1250905544,0.2341469079,-0.4163664877,0.3326435387,-0.0722600818,-0.1643588096,0.0049153673,-0.0324528106,-0.1588961631,-0.1072515398,-0.0156877059,0.2056005001,-0.0995983109,-0.0037558912,0.0271735881,-0.1192050502,0.1286043823,-0.4730828702,-0.5799900889,0.0736573488,-0.3627660871,-0.2398155332,0.1660139412,-0.0849088207,0.0135641443,-0.0959141254,0.2853991687,-0.2305099517,-0.2514098287,-0.0738202706,0.2073796839,-0.1735912412,-0.1366205215,-0.557320416,0.0548544563,-0.436563313,0.101188682,0.018891573,0.1017327383,0.0375571363,-0.0820854977,-0.1690710038,-0.1362932622,-0.0514218844,-0.293867141,-0.0489822477,0.1600505561,-0.0954016894,-0.1362264305,0.1065514386,-0.0223394427,0.1718405634,0.0870222747,-0.1610367745,-0.2975949049,-0.0730601102,0.5418993831,0.2266817093,-0.0641282722,0.3999685049,0.2694306076,-0.3475996256,-0.1914690882,-0.3262618482,0.1282105744,0.4744692445,0.0033857294,0.0033432383,0.3399909735,-0.2511548698,0.0921167806,-0.2350277603,-0.051174514,0.3001089394,-0.1415145248,0.2849588394,-0.1428914219,-0.0071283537,-0.0974474847,-0.1634303182,-0.1340658516,0.2539280653,-0.1737800688,-0.3170131743,-0.1334287226,-0.0476195328,-0.1256586909,-0.1669546217,-0.0212014038,-0.0950190574,0.346999228,0.1685113907,0.0485512763,-0.3341020048,0.084361352,0.1554456353,0.0494349003,0.2000489533,-0.1551099271,-0.7678235173,0.1532075554,-0.0706857741,0.3174476027,-0.1691426784,0.2255896032,-0.0195240881,-0.4396429956,-0.0972897857,0.0526980199,0.2252807468,-0.0602381341,-0.0601037703,0.1844872683,-0.03928205,0.1776095778,-0.214052096,-0.161970377,0.2218600363,0.1676322967,0.5520951748,-0.089400813,-0.2688205838,0.2293926924,-0.0158827137,-0.1981133968,0.1438087523,-0.1839707345,-0.1986193061,-0.1286721528,-0.0333907641,0.0879759043,-0.2221798748,0.0281767379,-0.2384418696,-0.0474877283,-0.1088155434,-0.1946702302,0.4526596665,0.0595416017,0.1362853944,-0.0471700281,-0.0558626279,0.2471228093,-0.0569285229,0.3623647988,-0.1530197263,-0.0913657621,0.0098151024,-0.2691310346,0.2775522172,0.1233643442,0.0262373835,0.2892575264,-0.1537756771,-0.0037390348,-0.1444713324,-0.081129685,0.4851858914,0.1619958431,0.0137931202,-0.3295781016,0.4239297509,-0.0664293244,-0.3327226043,-0.002479393,-0.0652283803,-0.3013924658,0.3489468098,0.101089932,0.6996944547,0.1155826226,-0.2431416959,0.0281724986,-0.1690158397,0.2073853016,-0.1701450944,0.1015933678,-0.2778677046,-0.3681688011,-0.0361555666,0.1232699603,0.2986629605,0.2746829391,-0.2960993648,0.1071839929,0.0425277129,0.4803037941,-0.1217010915,-0.0271597933,0.207688421,0.1009652391,-0.1617412716,0.2736833394,0.2048407644,0.1690930128,0.0473605134,-0.0433176272,0.1017142981,-0.1981489211,0.3188779652,-0.1081362516,-0.1012215242,0.1909431964,0.0040910351,0.017159963,-0.3878792226,0.0789596364,-0.0423652232,0.0823902637,0.1051879525,0.1162331402,0.5273289084,0.2515758574,-0.1477859318,0.1848680377,0.0944557488,-0.0451368839,-0.2678123415,0.0427888036,-0.243457213,-0.2682476342,-0.4397322536,0.0593077801,0.2657950521,-0.3779340386,-0.1201066002,0.2143042386,-0.2247184962,-0.279542774,0.2999521494,-0.1862876713,-0.0054431786,0.3359343112,0.1411119848,-0.2399690002,-0.0877956152,0.3488672674,0.1124981567,0.2704488933,0.291092068,-0.2685190141,-0.1219459847,-0.2861854136,-0.0532765463,0.1302606463,-0.0723645985,-0.068696484,0.0694574267,-0.1763560027,0.037554685,0.3567217886,0.0087376134,0.1834573448,-0.2202643901,-0.0734001622,-0.4034048021,0.3001064062,0.2302738428,0.1446648538,-0.0149796018,-0.1447079927,0.0598672703,0.2085016817,-0.5262194872,-0.0356625654,-0.111458011,0.3520565629,0.3536806703,0.061529994,-0.1901154071,-0.2057293504,0.3601993918,-0.1235353053,-0.1746911407,-0.3682048321,0.1318941861,0.0474722795,0.057220716,0.2752078474,-0.4085027277,-0.0449355207,-0.0394587964,-0.1263583899,0.0814679638,-0.0507949777,-0.0611363836,0.3497933447,-0.0232724305,0.0698971152,0.0947789028,0.1430911273,0.1448373049,-0.1588804573,-0.1876920015,0.0660460219,-0.1108857021,0.011590845,0.1014957204,0.0254563037,-0.3572868705,0.1902519614,0.4930791557,0.1030752063,0.2870853841,0.1493749022,0.0570147336,0.2747018039,0.020835463,-0.3025456667,0.0566907153,0.4529055357,-0.5251702666,-0.1691331416,0.2796272635,-0.088461861,0.3018027842,0.023140179,0.3356195092,0.2046800554,0.006468487,0.2736975849,0.1341715008,-0.0857229829,0.1367304176,0.5069727898,-0.0234641675,0.3237041831,0.4167168736,0.2611184716,-0.0403494872,0.3395244181,0.465842098,0.3872600794,0.3250512481,0.229268536,0.2035392374,-0.2617809474,0.3029993474,0.0796358734,0.013896456,0.2118854076,0.1724850833,-0.1661159396,-0.1193988025,-0.2296013683,-0.1163912416,0.0818713605,-0.2179923207,-0.2001150846,0.1175502539,-0.3299767375,-0.1569164544,-0.232613191,-0.0484570079,0.2548478842,0.6760661602,-0.3203324676,0.0132699097,-0.0347862989,0.0143543351,-0.2662617862,0.3216330707,-0.1386518925,0.1675922722,0.0863247812,-0.1352593601,0.2571660578,0.1303147972,-0.2115638703,0.1769381016,0.1280970871,0.1109262183,0.0744271055,-0.1741025448,-0.2813105285,0.1408310235,-0.1326126903,-0.0360920988,0.0768747553,0.3049050272,-0.2858789265,0.2086988091,0.062698327,0.1158384904,-0.0276488569,0.3197962642,0.1736640632,-0.4356670082,0.0200859923,-0.0994988531,-0.067002669,-0.4087151289,0.1540423334,-0.1878576279,0.2500641048,0.0377100632,0.2094711512,-0.0030194966,0.4899560511,0.1143553704,-0.2780036628,-0.0428390168,-0.0849010646,-0.4443606436,0.1755844653,0.0577759035,-0.1593482047,0.2530558705,-0.0116042756,0.2613921762,-0.1348225474,-0.0223281384,-0.4609578848,-0.2282187045,-0.0186677817,-0.3219127953,-0.142153576,-0.2430179417,0.0004842409,0.1979290992,-0.096862264,0.3306086957,-0.0018812761,0.218646422,-0.2042109966,-0.1000876874,0.1139481813,0.0034749541,0.4234070778,-0.0168442521,0.1501135975,-0.2632865608,-0.0052291611,-0.0154262828,-0.2999805808,-0.333892405,0.2420206964,-0.1267617643,0.1842337847,0.2645210326,0.1361540407,-0.0314150751,0.057400655,-0.0445044897,-0.1537882835,-0.1114982814,0.090954043,0.1614604145,-0.0613172427,0.1073868647,0.2878067493,0.0884871855,-0.0117161321,-0.1513342559,-0.3755518198,0.4516317844,0.0104454206,-0.3504465222,-0.2358574122,0.1048553139,0.1376280785,-0.1870743334,-0.4265524447,-0.146454215,0.3762686849,-0.209205091,-0.1155866012,0.146756649,0.0016336729,-0.164762482,-0.0864933729,0.3128185272,-0.1695459634,-0.2043665797,-0.2141511589,-0.011708064]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/331","title":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`","comments":"I couldn't reproduce on my side.\r\nIt looks like you were not able to generate all the examples, and you have the problem for each split train-test-validation.\r\nCould you try to enable logging, try again and send the logs ?\r\n```python\r\nimport logging\r\nlogging.basicConfig(level=logging.INFO)\r\n```","body":"```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```","comment_length":45,"text":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`\n```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```\nI couldn't reproduce on my side.\r\nIt looks like you were not able to generate all the examples, and you have the problem for each split train-test-validation.\r\nCould you try to enable logging, try again and send the logs ?\r\n```python\r\nimport logging\r\nlogging.basicConfig(level=logging.INFO)\r\n```","embeddings":[-0.1410695314,-0.0569875538,-0.008213372,0.2086568773,0.0776493773,0.0907960758,0.1833027154,0.5251758695,0.0987313315,0.1542323083,0.0162980705,0.0769967064,-0.3141199946,-0.0437742993,-0.0335896686,0.0359017923,-0.1871679723,0.251511842,0.044887729,-0.0229966342,-0.2874323428,0.2817907035,-0.2087394744,-0.0366006866,-0.4835281372,-0.2941971123,0.0143964859,0.3698371649,0.037647862,-0.3093646169,-0.054115586,-0.1348574609,-0.1841386408,0.1029714793,-0.0001130321,0.1136059314,0.3911184371,-0.1433287412,-0.4867380261,0.0093758097,-0.210433796,-0.4906257391,0.0880899355,-0.354680568,-0.1552342176,0.0253197029,0.0945640802,-0.2195387632,0.4969272316,0.3904682994,0.2402192354,0.5636107922,0.0662313849,0.1633443981,-0.0887694806,-0.1809548885,-0.1458708197,-0.0048658135,-0.3178360164,-0.1510139853,-0.0242430959,0.4289468825,-0.0926206186,0.1253475994,-0.0872375593,0.1102501974,0.0889446437,-0.1658464372,0.093881309,0.1360183805,-0.0329074785,-0.0759069175,-0.2113918364,-0.4385608137,-0.0160641279,-0.3980173469,0.2251276523,0.2127174139,-0.1745027453,0.0653799996,-0.3212834597,0.0384899378,-0.0294721033,0.2034061104,0.0819332451,0.4111941159,-0.1002976,0.0258119442,0.0555451661,-0.0901165307,0.1567516327,-0.1769279093,0.1543263644,0.077836372,-0.1457687616,-0.045277901,0.1644648165,-0.1484151632,0.0430351757,-0.0675480068,0.1789103448,-0.0621050708,0.1744745076,0.0868742242,0.047065638,0.3125295639,-0.0163507275,0.268633306,0.1245981306,0.2852095664,0.0025192983,0.0181032717,0.1172021255,-0.0053176638,0.1188476905,0.2207996994,0.3734848201,-0.0592549704,-0.3276322186,0.1337225586,-0.2627150416,-0.2590237856,-0.0015641216,0.2562178373,-0.0583013371,0.3132656813,-0.0608430654,-0.0814521238,-0.3839427233,-0.1547316015,-0.3804471493,0.170080021,-0.1844006926,0.080809094,0.2795961797,-0.0796258152,0.457221657,-0.249753952,-0.0141821839,-0.14726381,0.2093333155,-0.039835576,0.0250497442,0.28318578,-0.130973205,0.1706077605,0.2590435743,0.1690691859,-0.2127455175,-0.0859967843,-0.0040091961,-0.605062902,0.1887341738,0.2487290353,0.059362527,-0.0686998665,0.6144146919,0.0758781433,0.2487533242,0.035048902,-0.1620905846,-0.0698391646,-0.0113207363,-0.1684927046,0.0756478012,0.4380223155,-0.2325228155,-0.2369064242,-0.0135412402,-0.1419267803,0.5154548883,0.4340136647,-0.1053208038,0.0439397804,-0.1265488267,0.1538269967,0.2003797293,-0.1545396298,-0.3955631256,0.1898384094,-0.0884127989,0.2039062083,0.2070534676,0.1967763603,-0.0294462275,0.1016341671,0.0539142787,0.5395390391,-0.2290878296,0.1055220217,-0.434442699,-0.4346444309,0.3582523465,0.0564582348,0.2241702229,-0.0408378653,-0.0627056584,0.4361121953,0.2115100026,-0.1110629141,0.0374838039,0.0333377346,-0.1516636014,-0.1143535376,0.0171922985,-0.0431710444,-0.4885517359,0.1375229061,-0.1571971476,0.3005177379,-0.050177224,-0.1014922038,-0.2627061307,-0.194733575,-0.1750023663,-0.2838497162,0.2178567648,0.2274239957,0.209719494,0.0044355784,-0.278423816,0.3022136688,-0.0314067528,0.1214026138,-0.5230283737,0.2627040744,-0.2279074043,-0.1810560226,0.2639639974,0.4298867285,0.2641758621,-0.1530768126,-0.2832179368,0.2833616138,-0.0861629993,0.231596604,-0.1480443329,-0.1591009498,0.1137663871,-0.0814554542,0.1697773635,0.4025975764,-0.03424301,-0.1033013612,-0.1191942096,0.4826929867,0.0484582894,0.0898778141,0.1499334574,-0.1141223013,0.0772690475,-0.1309580058,0.1130278036,-0.5237566233,0.3187466562,-0.169630602,0.2132307738,-0.0878212526,-0.2500077784,0.0530378111,0.6815172434,-0.0259576812,0.0975138694,0.0207601693,-0.1026206613,-0.2020810992,-0.1837860048,0.159725666,0.3386713266,0.2413775921,0.1866382957,0.2474578917,-0.1729585528,-0.286001265,0.27237764,0.095468618,0.2538036108,0.385949403,0.0060878471,-0.0024015717,-0.5314667225,-0.0248315446,0.013506908,0.3741949201,-0.2236362398,-0.201827243,-0.4316599369,-0.4015900493,-0.1702875495,-0.2987873554,-0.1874077171,-0.5853300691,0.0800147429,0.030056335,-0.1252746582,0.1624230742,-0.1581038833,0.231892392,0.2769706845,-0.0165657867,-0.192729637,-0.0813186616,-0.1340953261,0.0754104182,0.1756269485,0.038011238,0.3282631934,-0.1323553175,-0.1795409769,-0.0246300977,-0.4010378718,0.0776721239,0.0673702881,0.1397411227,0.304790169,0.1800093949,0.3312228322,-0.3353323936,0.1972529143,-0.2418714613,-0.2770200372,-0.0081244279,0.0024520406,0.1431996375,-0.2382315993,-0.4810171425,-0.4326365292,-0.5125101209,0.1009555608,0.0488469638,0.4066635966,0.2704299092,-0.2863166034,0.1968383491,-0.121017009,0.1822268218,-0.2584162951,-0.1776084006,0.250297606,-0.2715799809,-0.2704790533,-0.1265957206,-0.148699671,0.2048910111,-0.2855255306,-0.4239244759,-0.3295074403,-0.1458872259,-0.0527010113,0.1072003022,-0.0558419265,0.2296750695,-0.0121127972,-0.1654176414,-0.033512149,-0.3873255849,0.0534179732,0.1250744611,0.0702088997,-0.0927717984,0.6258087158,0.0622662865,0.5375280976,0.3718982339,0.0934443921,0.2314644009,0.1095010936,0.1484319568,-0.0122379875,-0.3427927792,0.2506149709,0.1518257856,0.0592511334,0.3488012552,-0.0516511016,-0.3857150078,-0.0800836086,-0.3046748042,-0.465890646,-0.2806211114,0.1811255217,0.1248168126,0.257915318,0.2710350752,0.1946982741,-0.1767562777,-0.1700816005,-0.1458094865,0.1727721691,0.0059876768,-0.0366931371,-0.1083944216,0.0141898142,-0.610968709,0.2600288689,0.0066701602,0.6111478806,-0.1570565253,-0.0673669726,-0.0168537963,-0.2866101563,0.3602798283,-0.3092956245,-0.0543416403,0.1828872859,-0.1095798239,-0.3564962447,-0.1205786094,-0.1527364999,0.0463261046,0.5042914748,0.5133962035,-0.2715702355,-0.2822863758,0.4173494279,0.1056818143,-0.1925458461,-0.2405329794,-0.2602197528,-0.2365468591,-0.3589510024,-0.0029185996,-0.3549587131,0.2806672752,0.0347830579,-0.0494445525,-0.0367941447,-0.2479297519,0.1712224931,0.0238339435,-0.0075592832,0.2060532123,0.1902969331,0.2586178482,0.3282211721,-0.2777053118,0.321195066,-0.0436938889,0.0135160172,-0.086697489,0.1257402003,0.0457252376,0.4861247241,0.0170813538,-0.0189167988,0.2016194761,0.3254525959,-0.0044429717,-0.0331222415,0.4496917725,-0.0619700886,-0.411162287,-0.1648898721,0.2997184098,0.1231901348,0.1361619532,-0.1300019771,-0.1973913908,-0.2437216341,0.1276027113,0.1129900441,0.7832134366,0.2591175437,0.0648233965,0.3565958738,-0.154997915,0.6783713102,-0.2112062573,0.2569588423,-0.3706091642,-0.3312232494,-0.0379182659,-0.110877268,0.3169820905,0.1003670096,-0.1770121157,0.4902181625,-0.1291067749,-0.0994330347,-0.0617544949,0.4637355208,-0.2605726421,-0.1433636397,-0.4001678526,0.1066973135,-0.1958624274,0.3526180387,-0.1023912877,-0.1949197054,0.2577715814,-0.2590811551,-0.0749825463,0.2773513794,-0.2684628367,0.1398487538,0.0061621419,0.2401976585,-0.2210501879,-0.3827387393,0.3137292862,-0.1355180293,-0.0385885052,-0.0691945553,-0.2621516585,-0.024181718,-0.1012062579,0.1060120016,0.4224326015,-0.2187235355,-0.1145567298,0.0203035623,-0.1885794401,-0.1630533636,-0.0189780127,-0.0354848094,0.1039875001,-0.1709197462,-0.1970958263,0.0954251364,-0.2310493588,-0.3154529631,0.146178931,-0.057743106,-0.2906800508,0.3077005148,-0.0858525261,-0.280965358,-0.2579652071,0.2445214242,0.1696913093,0.0508790649,0.4922745228,0.0225029569,-0.289804548,-0.2790828943,-0.0300642699,0.0263601001,-0.3613367677,0.1998145133,-0.0870138481,-0.0155756744,0.2755271792,0.2051900923,0.1997067034,0.2912752032,-0.2323927581,-0.5686837435,-0.1438727677,0.0629848167,-0.0053662611,0.2893784344,0.0466341376,0.5640015602,0.0183579158,0.3612833321,-0.3558589816,0.0370596088,-0.2532141507,0.1774706542,0.1644686908,-0.2871336341,0.4796305895,0.3361745477,0.1896099001,0.184483543,-0.3276686668,-0.2398814857,-0.1398786306,0.1459911615,0.0528673306,-0.1507482231,0.1091264933,0.0210518595,-0.1626609713,-0.0761597157,-0.0131698949,-0.1183911264,-0.0879904702,0.246785149,-0.0422548801,-0.0603146441,0.1202732995,0.0563640222,-0.1231211573,0.3887268901,-0.011862115,0.230624184,-0.2371812314,0.0459101312,-0.1055980399,-0.1075530127,0.0492279939,0.0140714366,0.2833747864,-0.4350683391,-0.2895481288,0.1623046845,0.231419459,0.2564552128,-0.0285925828,-0.1506037265,0.1654750556,0.2218014747,-0.3195352852,0.0245437622,0.3671759367,-0.0035081224,0.3218832314,0.3530758023,0.1187401861,0.0195527542,-0.0508777685,0.0890130997,0.2165187448,-0.2659980059,0.0368568152,0.2631331086,-0.1103861332,-0.0150807723,0.3015813529,-0.1218026355,0.3009927273,0.237793833,-0.0317757912,0.4294939041,-0.0397178307,0.3304133117,-0.0069701187,-0.2167673558,0.1839850694,0.4842499197,-0.1119999141,0.3832813501,0.0147898477,0.4105642736,-0.2803665698,0.0144452732,-0.177463308,0.4721761346,-0.1400122941,0.0913876519,-0.0336264521,-0.0987380594,-0.2867724299,-0.0064010476,0.078561306,0.0901599079,0.1506229341,0.1159298345,-0.0376419351,-0.3303664923,-0.1878765672,0.1589913517,0.2382988781,-0.3991445005,0.3074398935,0.2775505185,-0.0423321314,-0.1595377773,0.6101233363,0.3431236744,0.3848828971,-0.1388491094,-0.0203314312,0.1635613889,-0.0777899623,-0.0890924409,0.3411318362,0.1897569448,-0.0746393427,0.3646362722,0.1043184698,-0.2784147859,-0.046337042,0.0937767401,-0.2416163087,-0.1392454207,0.2932831347,-0.0494255424,0.2100278288,-0.222061947,0.0304331221,-0.6043111682,-0.0503048562,0.1934356242,-0.0807917118,0.2918081284,0.0341566429,0.1002931669,-0.1915131211,0.4911065996,0.0694363788,-0.0748937726,-0.3047780395,-0.2246827185,-0.5798493028,0.1475732327,-0.1549478322,-0.0564664453,0.0097489506,0.2813649476,-0.2015340179,0.049443271,0.0776405334,0.3546506464,-0.0404426381,-0.1575725079,-0.3835550249,0.0163158383,0.0054045385,-0.2268491983,0.0736885443,-0.4133264124,0.2439867556,-0.2011682093,0.0368155167,-0.1237300634,0.0597289316,-0.0689730644,0.4797645807,0.0129686212,0.5052636266,0.28163445,-0.2362820506,-0.3487403989,-0.1656108499,-0.1421875954,-0.2967919409,0.1579185128,0.3191114068,0.5874743462,0.017294988,0.0549576357,-0.3607902825,0.3801414073,0.0239438303,-0.1877516508,0.0241975896,0.3261172771,-0.2411894798,0.2674721181,0.2473868579,0.3860733807,0.1126281247,0.1490353644,-0.2446985543,-0.3713934422,0.5164570212,-0.1831858456,-0.2234716415,-0.0118958345,0.1257257313,0.0264652986,-0.1279568076,-0.4188547432,0.0567013286,0.3580688238,-0.1017922908,-0.0891665518,0.2000727803,0.2182815969,-0.2056010067,-0.0198608879,-0.0954742506,0.3953032792,-0.431453079,0.2040592879,-0.0773261264]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/331","title":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`","comments":"here's the log\r\n```\r\n>>> import nlp\r\nimport logging\r\nlogging.basicConfig(level=logging.INFO)\r\nnlp.load_dataset('cnn_dailymail', '3.0.0')\r\n>>> import logging\r\n>>> logging.basicConfig(level=logging.INFO)\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nINFO:nlp.load:Checking \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py for additional imports.\r\nINFO:filelock:Lock 140443095301136 acquired on \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py.lock\r\nINFO:nlp.load:Found main folder for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\r\nINFO:nlp.load:Found specific version folder for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\r\nINFO:nlp.load:Found script file from https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py to \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/cnn_dailymail.py\r\nINFO:nlp.load:Updating dataset infos file from https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/dataset_infos.json to \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/cnn_dailymail.json\r\nINFO:filelock:Lock 140443095301136 released on \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py.lock\r\nINFO:nlp.info:Loading Dataset Infos from \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\r\nINFO:nlp.builder:Generating dataset cnn_dailymail (\/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0)\r\nINFO:nlp.builder:Dataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\nINFO:nlp.utils.info_utils:All the checksums matched successfully.\r\nINFO:nlp.builder:Generating split train\r\nINFO:nlp.arrow_writer:Done writing 285161 examples in 1240618482 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-train.arrow.\r\nINFO:nlp.builder:Generating split validation\r\nINFO:nlp.arrow_writer:Done writing 13255 examples in 56637485 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-validation.arrow.\r\nINFO:nlp.builder:Generating split test\r\nINFO:nlp.arrow_writer:Done writing 11379 examples in 48931393 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-test.arrow.\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```","body":"```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```","comment_length":223,"text":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`\n```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```\nhere's the log\r\n```\r\n>>> import nlp\r\nimport logging\r\nlogging.basicConfig(level=logging.INFO)\r\nnlp.load_dataset('cnn_dailymail', '3.0.0')\r\n>>> import logging\r\n>>> logging.basicConfig(level=logging.INFO)\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nINFO:nlp.load:Checking \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py for additional imports.\r\nINFO:filelock:Lock 140443095301136 acquired on \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py.lock\r\nINFO:nlp.load:Found main folder for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\r\nINFO:nlp.load:Found specific version folder for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\r\nINFO:nlp.load:Found script file from https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py to \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/cnn_dailymail.py\r\nINFO:nlp.load:Updating dataset infos file from https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/dataset_infos.json to \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/cnn_dailymail.json\r\nINFO:filelock:Lock 140443095301136 released on \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py.lock\r\nINFO:nlp.info:Loading Dataset Infos from \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\r\nINFO:nlp.builder:Generating dataset cnn_dailymail (\/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0)\r\nINFO:nlp.builder:Dataset not on Hf google storage. Downloading and preparing it from source\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\nINFO:nlp.utils.info_utils:All the checksums matched successfully.\r\nINFO:nlp.builder:Generating split train\r\nINFO:nlp.arrow_writer:Done writing 285161 examples in 1240618482 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-train.arrow.\r\nINFO:nlp.builder:Generating split validation\r\nINFO:nlp.arrow_writer:Done writing 13255 examples in 56637485 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-validation.arrow.\r\nINFO:nlp.builder:Generating split test\r\nINFO:nlp.arrow_writer:Done writing 11379 examples in 48931393 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-test.arrow.\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```","embeddings":[-0.1410695314,-0.0569875538,-0.008213372,0.2086568773,0.0776493773,0.0907960758,0.1833027154,0.5251758695,0.0987313315,0.1542323083,0.0162980705,0.0769967064,-0.3141199946,-0.0437742993,-0.0335896686,0.0359017923,-0.1871679723,0.251511842,0.044887729,-0.0229966342,-0.2874323428,0.2817907035,-0.2087394744,-0.0366006866,-0.4835281372,-0.2941971123,0.0143964859,0.3698371649,0.037647862,-0.3093646169,-0.054115586,-0.1348574609,-0.1841386408,0.1029714793,-0.0001130321,0.1136059314,0.3911184371,-0.1433287412,-0.4867380261,0.0093758097,-0.210433796,-0.4906257391,0.0880899355,-0.354680568,-0.1552342176,0.0253197029,0.0945640802,-0.2195387632,0.4969272316,0.3904682994,0.2402192354,0.5636107922,0.0662313849,0.1633443981,-0.0887694806,-0.1809548885,-0.1458708197,-0.0048658135,-0.3178360164,-0.1510139853,-0.0242430959,0.4289468825,-0.0926206186,0.1253475994,-0.0872375593,0.1102501974,0.0889446437,-0.1658464372,0.093881309,0.1360183805,-0.0329074785,-0.0759069175,-0.2113918364,-0.4385608137,-0.0160641279,-0.3980173469,0.2251276523,0.2127174139,-0.1745027453,0.0653799996,-0.3212834597,0.0384899378,-0.0294721033,0.2034061104,0.0819332451,0.4111941159,-0.1002976,0.0258119442,0.0555451661,-0.0901165307,0.1567516327,-0.1769279093,0.1543263644,0.077836372,-0.1457687616,-0.045277901,0.1644648165,-0.1484151632,0.0430351757,-0.0675480068,0.1789103448,-0.0621050708,0.1744745076,0.0868742242,0.047065638,0.3125295639,-0.0163507275,0.268633306,0.1245981306,0.2852095664,0.0025192983,0.0181032717,0.1172021255,-0.0053176638,0.1188476905,0.2207996994,0.3734848201,-0.0592549704,-0.3276322186,0.1337225586,-0.2627150416,-0.2590237856,-0.0015641216,0.2562178373,-0.0583013371,0.3132656813,-0.0608430654,-0.0814521238,-0.3839427233,-0.1547316015,-0.3804471493,0.170080021,-0.1844006926,0.080809094,0.2795961797,-0.0796258152,0.457221657,-0.249753952,-0.0141821839,-0.14726381,0.2093333155,-0.039835576,0.0250497442,0.28318578,-0.130973205,0.1706077605,0.2590435743,0.1690691859,-0.2127455175,-0.0859967843,-0.0040091961,-0.605062902,0.1887341738,0.2487290353,0.059362527,-0.0686998665,0.6144146919,0.0758781433,0.2487533242,0.035048902,-0.1620905846,-0.0698391646,-0.0113207363,-0.1684927046,0.0756478012,0.4380223155,-0.2325228155,-0.2369064242,-0.0135412402,-0.1419267803,0.5154548883,0.4340136647,-0.1053208038,0.0439397804,-0.1265488267,0.1538269967,0.2003797293,-0.1545396298,-0.3955631256,0.1898384094,-0.0884127989,0.2039062083,0.2070534676,0.1967763603,-0.0294462275,0.1016341671,0.0539142787,0.5395390391,-0.2290878296,0.1055220217,-0.434442699,-0.4346444309,0.3582523465,0.0564582348,0.2241702229,-0.0408378653,-0.0627056584,0.4361121953,0.2115100026,-0.1110629141,0.0374838039,0.0333377346,-0.1516636014,-0.1143535376,0.0171922985,-0.0431710444,-0.4885517359,0.1375229061,-0.1571971476,0.3005177379,-0.050177224,-0.1014922038,-0.2627061307,-0.194733575,-0.1750023663,-0.2838497162,0.2178567648,0.2274239957,0.209719494,0.0044355784,-0.278423816,0.3022136688,-0.0314067528,0.1214026138,-0.5230283737,0.2627040744,-0.2279074043,-0.1810560226,0.2639639974,0.4298867285,0.2641758621,-0.1530768126,-0.2832179368,0.2833616138,-0.0861629993,0.231596604,-0.1480443329,-0.1591009498,0.1137663871,-0.0814554542,0.1697773635,0.4025975764,-0.03424301,-0.1033013612,-0.1191942096,0.4826929867,0.0484582894,0.0898778141,0.1499334574,-0.1141223013,0.0772690475,-0.1309580058,0.1130278036,-0.5237566233,0.3187466562,-0.169630602,0.2132307738,-0.0878212526,-0.2500077784,0.0530378111,0.6815172434,-0.0259576812,0.0975138694,0.0207601693,-0.1026206613,-0.2020810992,-0.1837860048,0.159725666,0.3386713266,0.2413775921,0.1866382957,0.2474578917,-0.1729585528,-0.286001265,0.27237764,0.095468618,0.2538036108,0.385949403,0.0060878471,-0.0024015717,-0.5314667225,-0.0248315446,0.013506908,0.3741949201,-0.2236362398,-0.201827243,-0.4316599369,-0.4015900493,-0.1702875495,-0.2987873554,-0.1874077171,-0.5853300691,0.0800147429,0.030056335,-0.1252746582,0.1624230742,-0.1581038833,0.231892392,0.2769706845,-0.0165657867,-0.192729637,-0.0813186616,-0.1340953261,0.0754104182,0.1756269485,0.038011238,0.3282631934,-0.1323553175,-0.1795409769,-0.0246300977,-0.4010378718,0.0776721239,0.0673702881,0.1397411227,0.304790169,0.1800093949,0.3312228322,-0.3353323936,0.1972529143,-0.2418714613,-0.2770200372,-0.0081244279,0.0024520406,0.1431996375,-0.2382315993,-0.4810171425,-0.4326365292,-0.5125101209,0.1009555608,0.0488469638,0.4066635966,0.2704299092,-0.2863166034,0.1968383491,-0.121017009,0.1822268218,-0.2584162951,-0.1776084006,0.250297606,-0.2715799809,-0.2704790533,-0.1265957206,-0.148699671,0.2048910111,-0.2855255306,-0.4239244759,-0.3295074403,-0.1458872259,-0.0527010113,0.1072003022,-0.0558419265,0.2296750695,-0.0121127972,-0.1654176414,-0.033512149,-0.3873255849,0.0534179732,0.1250744611,0.0702088997,-0.0927717984,0.6258087158,0.0622662865,0.5375280976,0.3718982339,0.0934443921,0.2314644009,0.1095010936,0.1484319568,-0.0122379875,-0.3427927792,0.2506149709,0.1518257856,0.0592511334,0.3488012552,-0.0516511016,-0.3857150078,-0.0800836086,-0.3046748042,-0.465890646,-0.2806211114,0.1811255217,0.1248168126,0.257915318,0.2710350752,0.1946982741,-0.1767562777,-0.1700816005,-0.1458094865,0.1727721691,0.0059876768,-0.0366931371,-0.1083944216,0.0141898142,-0.610968709,0.2600288689,0.0066701602,0.6111478806,-0.1570565253,-0.0673669726,-0.0168537963,-0.2866101563,0.3602798283,-0.3092956245,-0.0543416403,0.1828872859,-0.1095798239,-0.3564962447,-0.1205786094,-0.1527364999,0.0463261046,0.5042914748,0.5133962035,-0.2715702355,-0.2822863758,0.4173494279,0.1056818143,-0.1925458461,-0.2405329794,-0.2602197528,-0.2365468591,-0.3589510024,-0.0029185996,-0.3549587131,0.2806672752,0.0347830579,-0.0494445525,-0.0367941447,-0.2479297519,0.1712224931,0.0238339435,-0.0075592832,0.2060532123,0.1902969331,0.2586178482,0.3282211721,-0.2777053118,0.321195066,-0.0436938889,0.0135160172,-0.086697489,0.1257402003,0.0457252376,0.4861247241,0.0170813538,-0.0189167988,0.2016194761,0.3254525959,-0.0044429717,-0.0331222415,0.4496917725,-0.0619700886,-0.411162287,-0.1648898721,0.2997184098,0.1231901348,0.1361619532,-0.1300019771,-0.1973913908,-0.2437216341,0.1276027113,0.1129900441,0.7832134366,0.2591175437,0.0648233965,0.3565958738,-0.154997915,0.6783713102,-0.2112062573,0.2569588423,-0.3706091642,-0.3312232494,-0.0379182659,-0.110877268,0.3169820905,0.1003670096,-0.1770121157,0.4902181625,-0.1291067749,-0.0994330347,-0.0617544949,0.4637355208,-0.2605726421,-0.1433636397,-0.4001678526,0.1066973135,-0.1958624274,0.3526180387,-0.1023912877,-0.1949197054,0.2577715814,-0.2590811551,-0.0749825463,0.2773513794,-0.2684628367,0.1398487538,0.0061621419,0.2401976585,-0.2210501879,-0.3827387393,0.3137292862,-0.1355180293,-0.0385885052,-0.0691945553,-0.2621516585,-0.024181718,-0.1012062579,0.1060120016,0.4224326015,-0.2187235355,-0.1145567298,0.0203035623,-0.1885794401,-0.1630533636,-0.0189780127,-0.0354848094,0.1039875001,-0.1709197462,-0.1970958263,0.0954251364,-0.2310493588,-0.3154529631,0.146178931,-0.057743106,-0.2906800508,0.3077005148,-0.0858525261,-0.280965358,-0.2579652071,0.2445214242,0.1696913093,0.0508790649,0.4922745228,0.0225029569,-0.289804548,-0.2790828943,-0.0300642699,0.0263601001,-0.3613367677,0.1998145133,-0.0870138481,-0.0155756744,0.2755271792,0.2051900923,0.1997067034,0.2912752032,-0.2323927581,-0.5686837435,-0.1438727677,0.0629848167,-0.0053662611,0.2893784344,0.0466341376,0.5640015602,0.0183579158,0.3612833321,-0.3558589816,0.0370596088,-0.2532141507,0.1774706542,0.1644686908,-0.2871336341,0.4796305895,0.3361745477,0.1896099001,0.184483543,-0.3276686668,-0.2398814857,-0.1398786306,0.1459911615,0.0528673306,-0.1507482231,0.1091264933,0.0210518595,-0.1626609713,-0.0761597157,-0.0131698949,-0.1183911264,-0.0879904702,0.246785149,-0.0422548801,-0.0603146441,0.1202732995,0.0563640222,-0.1231211573,0.3887268901,-0.011862115,0.230624184,-0.2371812314,0.0459101312,-0.1055980399,-0.1075530127,0.0492279939,0.0140714366,0.2833747864,-0.4350683391,-0.2895481288,0.1623046845,0.231419459,0.2564552128,-0.0285925828,-0.1506037265,0.1654750556,0.2218014747,-0.3195352852,0.0245437622,0.3671759367,-0.0035081224,0.3218832314,0.3530758023,0.1187401861,0.0195527542,-0.0508777685,0.0890130997,0.2165187448,-0.2659980059,0.0368568152,0.2631331086,-0.1103861332,-0.0150807723,0.3015813529,-0.1218026355,0.3009927273,0.237793833,-0.0317757912,0.4294939041,-0.0397178307,0.3304133117,-0.0069701187,-0.2167673558,0.1839850694,0.4842499197,-0.1119999141,0.3832813501,0.0147898477,0.4105642736,-0.2803665698,0.0144452732,-0.177463308,0.4721761346,-0.1400122941,0.0913876519,-0.0336264521,-0.0987380594,-0.2867724299,-0.0064010476,0.078561306,0.0901599079,0.1506229341,0.1159298345,-0.0376419351,-0.3303664923,-0.1878765672,0.1589913517,0.2382988781,-0.3991445005,0.3074398935,0.2775505185,-0.0423321314,-0.1595377773,0.6101233363,0.3431236744,0.3848828971,-0.1388491094,-0.0203314312,0.1635613889,-0.0777899623,-0.0890924409,0.3411318362,0.1897569448,-0.0746393427,0.3646362722,0.1043184698,-0.2784147859,-0.046337042,0.0937767401,-0.2416163087,-0.1392454207,0.2932831347,-0.0494255424,0.2100278288,-0.222061947,0.0304331221,-0.6043111682,-0.0503048562,0.1934356242,-0.0807917118,0.2918081284,0.0341566429,0.1002931669,-0.1915131211,0.4911065996,0.0694363788,-0.0748937726,-0.3047780395,-0.2246827185,-0.5798493028,0.1475732327,-0.1549478322,-0.0564664453,0.0097489506,0.2813649476,-0.2015340179,0.049443271,0.0776405334,0.3546506464,-0.0404426381,-0.1575725079,-0.3835550249,0.0163158383,0.0054045385,-0.2268491983,0.0736885443,-0.4133264124,0.2439867556,-0.2011682093,0.0368155167,-0.1237300634,0.0597289316,-0.0689730644,0.4797645807,0.0129686212,0.5052636266,0.28163445,-0.2362820506,-0.3487403989,-0.1656108499,-0.1421875954,-0.2967919409,0.1579185128,0.3191114068,0.5874743462,0.017294988,0.0549576357,-0.3607902825,0.3801414073,0.0239438303,-0.1877516508,0.0241975896,0.3261172771,-0.2411894798,0.2674721181,0.2473868579,0.3860733807,0.1126281247,0.1490353644,-0.2446985543,-0.3713934422,0.5164570212,-0.1831858456,-0.2234716415,-0.0118958345,0.1257257313,0.0264652986,-0.1279568076,-0.4188547432,0.0567013286,0.3580688238,-0.1017922908,-0.0891665518,0.2000727803,0.2182815969,-0.2056010067,-0.0198608879,-0.0954742506,0.3953032792,-0.431453079,0.2040592879,-0.0773261264]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/331","title":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`","comments":"> here's the log\r\n> \r\n> ```\r\n> >>> import nlp\r\n> import logging\r\n> logging.basicConfig(level=logging.INFO)\r\n> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\n> >>> import logging\r\n> >>> logging.basicConfig(level=logging.INFO)\r\n> >>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\n> INFO:nlp.load:Checking \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py for additional imports.\r\n> INFO:filelock:Lock 140443095301136 acquired on \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py.lock\r\n> INFO:nlp.load:Found main folder for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\r\n> INFO:nlp.load:Found specific version folder for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\r\n> INFO:nlp.load:Found script file from https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py to \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/cnn_dailymail.py\r\n> INFO:nlp.load:Updating dataset infos file from https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/dataset_infos.json to \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/dataset_infos.json\r\n> INFO:nlp.load:Found metadata file for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/cnn_dailymail.json\r\n> INFO:filelock:Lock 140443095301136 released on \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py.lock\r\n> INFO:nlp.info:Loading Dataset Infos from \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\r\n> INFO:nlp.builder:Generating dataset cnn_dailymail (\/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0)\r\n> INFO:nlp.builder:Dataset not on Hf google storage. Downloading and preparing it from source\r\n> Downloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n> INFO:nlp.utils.info_utils:All the checksums matched successfully.\r\n> INFO:nlp.builder:Generating split train\r\n> INFO:nlp.arrow_writer:Done writing 285161 examples in 1240618482 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-train.arrow.\r\n> INFO:nlp.builder:Generating split validation\r\n> INFO:nlp.arrow_writer:Done writing 13255 examples in 56637485 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-validation.arrow.\r\n> INFO:nlp.builder:Generating split test\r\n> INFO:nlp.arrow_writer:Done writing 11379 examples in 48931393 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-test.arrow.\r\n> Traceback (most recent call last):\r\n> File \"<stdin>\", line 1, in <module>\r\n> File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n> builder_instance.download_and_prepare(\r\n> File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n> self._download_and_prepare(\r\n> File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n> verify_splits(self.info.splits, split_dict)\r\n> File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n> raise NonMatchingSplitsSizesError(str(bad_splits))\r\n> nlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n> ```\r\n\r\nWith `nlp == 0.3.0` version, I'm not able to reproduce this error on my side.\r\nWhich version are you using for reproducing your bug?\r\n\r\n```\r\n>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n8.90k\/8.90k [00:18<00:00, 486B\/s]\r\n\r\nDownloading: 100%\r\n9.37k\/9.37k [00:00<00:00, 234kB\/s]\r\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\nDownloading:\r\n159M\/? [00:09<00:00, 16.7MB\/s]\r\n\r\nDownloading:\r\n376M\/? [00:06<00:00, 62.6MB\/s]\r\n\r\nDownloading:\r\n2.11M\/? [00:06<00:00, 333kB\/s]\r\n\r\nDownloading:\r\n46.4M\/? [00:02<00:00, 18.4MB\/s]\r\n\r\nDownloading:\r\n2.43M\/? [00:00<00:00, 2.62MB\/s]\r\n\r\nDataset cnn_dailymail downloaded and prepared to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0. Subsequent calls will reuse this data.\r\n{'test': Dataset(schema: {'article': 'string', 'highlights': 'string'}, num_rows: 11490),\r\n 'train': Dataset(schema: {'article': 'string', 'highlights': 'string'}, num_rows: 287113),\r\n 'validation': Dataset(schema: {'article': 'string', 'highlights': 'string'}, num_rows: 13368)}\r\n\r\n>> ...\r\n\r\n```","body":"```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```","comment_length":376,"text":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`\n```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```\n> here's the log\r\n> \r\n> ```\r\n> >>> import nlp\r\n> import logging\r\n> logging.basicConfig(level=logging.INFO)\r\n> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\n> >>> import logging\r\n> >>> logging.basicConfig(level=logging.INFO)\r\n> >>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\n> INFO:nlp.load:Checking \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py for additional imports.\r\n> INFO:filelock:Lock 140443095301136 acquired on \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py.lock\r\n> INFO:nlp.load:Found main folder for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\r\n> INFO:nlp.load:Found specific version folder for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\r\n> INFO:nlp.load:Found script file from https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py to \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/cnn_dailymail.py\r\n> INFO:nlp.load:Updating dataset infos file from https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/dataset_infos.json to \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/dataset_infos.json\r\n> INFO:nlp.load:Found metadata file for dataset https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/datasets\/cnn_dailymail\/cnn_dailymail.py at \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\/cnn_dailymail.json\r\n> INFO:filelock:Lock 140443095301136 released on \/u\/jm8wx\/.cache\/huggingface\/datasets\/720d2e20d8dc6d98f21195a39cc934bb41dd0a40b57ea3d323661a7c5d70522c.d44c2417f4e0fe938ede0a684dcbb1fa9b4789de22e8a99c43103d4b4c374b3b.py.lock\r\n> INFO:nlp.info:Loading Dataset Infos from \/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/datasets\/cnn_dailymail\/9645e0bc96f647decf46541f6f4bef6936ee82ace653ac362bab03309a46d4ad\r\n> INFO:nlp.builder:Generating dataset cnn_dailymail (\/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0)\r\n> INFO:nlp.builder:Dataset not on Hf google storage. Downloading and preparing it from source\r\n> Downloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n> INFO:nlp.utils.info_utils:All the checksums matched successfully.\r\n> INFO:nlp.builder:Generating split train\r\n> INFO:nlp.arrow_writer:Done writing 285161 examples in 1240618482 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-train.arrow.\r\n> INFO:nlp.builder:Generating split validation\r\n> INFO:nlp.arrow_writer:Done writing 13255 examples in 56637485 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-validation.arrow.\r\n> INFO:nlp.builder:Generating split test\r\n> INFO:nlp.arrow_writer:Done writing 11379 examples in 48931393 bytes \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0.incomplete\/cnn_dailymail-test.arrow.\r\n> Traceback (most recent call last):\r\n> File \"<stdin>\", line 1, in <module>\r\n> File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n> builder_instance.download_and_prepare(\r\n> File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n> self._download_and_prepare(\r\n> File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n> verify_splits(self.info.splits, split_dict)\r\n> File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n> raise NonMatchingSplitsSizesError(str(bad_splits))\r\n> nlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n> ```\r\n\r\nWith `nlp == 0.3.0` version, I'm not able to reproduce this error on my side.\r\nWhich version are you using for reproducing your bug?\r\n\r\n```\r\n>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\n\r\n8.90k\/8.90k [00:18<00:00, 486B\/s]\r\n\r\nDownloading: 100%\r\n9.37k\/9.37k [00:00<00:00, 234kB\/s]\r\n\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\nDownloading:\r\n159M\/? [00:09<00:00, 16.7MB\/s]\r\n\r\nDownloading:\r\n376M\/? [00:06<00:00, 62.6MB\/s]\r\n\r\nDownloading:\r\n2.11M\/? [00:06<00:00, 333kB\/s]\r\n\r\nDownloading:\r\n46.4M\/? [00:02<00:00, 18.4MB\/s]\r\n\r\nDownloading:\r\n2.43M\/? [00:00<00:00, 2.62MB\/s]\r\n\r\nDataset cnn_dailymail downloaded and prepared to \/root\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0. Subsequent calls will reuse this data.\r\n{'test': Dataset(schema: {'article': 'string', 'highlights': 'string'}, num_rows: 11490),\r\n 'train': Dataset(schema: {'article': 'string', 'highlights': 'string'}, num_rows: 287113),\r\n 'validation': Dataset(schema: {'article': 'string', 'highlights': 'string'}, num_rows: 13368)}\r\n\r\n>> ...\r\n\r\n```","embeddings":[-0.1410695314,-0.0569875538,-0.008213372,0.2086568773,0.0776493773,0.0907960758,0.1833027154,0.5251758695,0.0987313315,0.1542323083,0.0162980705,0.0769967064,-0.3141199946,-0.0437742993,-0.0335896686,0.0359017923,-0.1871679723,0.251511842,0.044887729,-0.0229966342,-0.2874323428,0.2817907035,-0.2087394744,-0.0366006866,-0.4835281372,-0.2941971123,0.0143964859,0.3698371649,0.037647862,-0.3093646169,-0.054115586,-0.1348574609,-0.1841386408,0.1029714793,-0.0001130321,0.1136059314,0.3911184371,-0.1433287412,-0.4867380261,0.0093758097,-0.210433796,-0.4906257391,0.0880899355,-0.354680568,-0.1552342176,0.0253197029,0.0945640802,-0.2195387632,0.4969272316,0.3904682994,0.2402192354,0.5636107922,0.0662313849,0.1633443981,-0.0887694806,-0.1809548885,-0.1458708197,-0.0048658135,-0.3178360164,-0.1510139853,-0.0242430959,0.4289468825,-0.0926206186,0.1253475994,-0.0872375593,0.1102501974,0.0889446437,-0.1658464372,0.093881309,0.1360183805,-0.0329074785,-0.0759069175,-0.2113918364,-0.4385608137,-0.0160641279,-0.3980173469,0.2251276523,0.2127174139,-0.1745027453,0.0653799996,-0.3212834597,0.0384899378,-0.0294721033,0.2034061104,0.0819332451,0.4111941159,-0.1002976,0.0258119442,0.0555451661,-0.0901165307,0.1567516327,-0.1769279093,0.1543263644,0.077836372,-0.1457687616,-0.045277901,0.1644648165,-0.1484151632,0.0430351757,-0.0675480068,0.1789103448,-0.0621050708,0.1744745076,0.0868742242,0.047065638,0.3125295639,-0.0163507275,0.268633306,0.1245981306,0.2852095664,0.0025192983,0.0181032717,0.1172021255,-0.0053176638,0.1188476905,0.2207996994,0.3734848201,-0.0592549704,-0.3276322186,0.1337225586,-0.2627150416,-0.2590237856,-0.0015641216,0.2562178373,-0.0583013371,0.3132656813,-0.0608430654,-0.0814521238,-0.3839427233,-0.1547316015,-0.3804471493,0.170080021,-0.1844006926,0.080809094,0.2795961797,-0.0796258152,0.457221657,-0.249753952,-0.0141821839,-0.14726381,0.2093333155,-0.039835576,0.0250497442,0.28318578,-0.130973205,0.1706077605,0.2590435743,0.1690691859,-0.2127455175,-0.0859967843,-0.0040091961,-0.605062902,0.1887341738,0.2487290353,0.059362527,-0.0686998665,0.6144146919,0.0758781433,0.2487533242,0.035048902,-0.1620905846,-0.0698391646,-0.0113207363,-0.1684927046,0.0756478012,0.4380223155,-0.2325228155,-0.2369064242,-0.0135412402,-0.1419267803,0.5154548883,0.4340136647,-0.1053208038,0.0439397804,-0.1265488267,0.1538269967,0.2003797293,-0.1545396298,-0.3955631256,0.1898384094,-0.0884127989,0.2039062083,0.2070534676,0.1967763603,-0.0294462275,0.1016341671,0.0539142787,0.5395390391,-0.2290878296,0.1055220217,-0.434442699,-0.4346444309,0.3582523465,0.0564582348,0.2241702229,-0.0408378653,-0.0627056584,0.4361121953,0.2115100026,-0.1110629141,0.0374838039,0.0333377346,-0.1516636014,-0.1143535376,0.0171922985,-0.0431710444,-0.4885517359,0.1375229061,-0.1571971476,0.3005177379,-0.050177224,-0.1014922038,-0.2627061307,-0.194733575,-0.1750023663,-0.2838497162,0.2178567648,0.2274239957,0.209719494,0.0044355784,-0.278423816,0.3022136688,-0.0314067528,0.1214026138,-0.5230283737,0.2627040744,-0.2279074043,-0.1810560226,0.2639639974,0.4298867285,0.2641758621,-0.1530768126,-0.2832179368,0.2833616138,-0.0861629993,0.231596604,-0.1480443329,-0.1591009498,0.1137663871,-0.0814554542,0.1697773635,0.4025975764,-0.03424301,-0.1033013612,-0.1191942096,0.4826929867,0.0484582894,0.0898778141,0.1499334574,-0.1141223013,0.0772690475,-0.1309580058,0.1130278036,-0.5237566233,0.3187466562,-0.169630602,0.2132307738,-0.0878212526,-0.2500077784,0.0530378111,0.6815172434,-0.0259576812,0.0975138694,0.0207601693,-0.1026206613,-0.2020810992,-0.1837860048,0.159725666,0.3386713266,0.2413775921,0.1866382957,0.2474578917,-0.1729585528,-0.286001265,0.27237764,0.095468618,0.2538036108,0.385949403,0.0060878471,-0.0024015717,-0.5314667225,-0.0248315446,0.013506908,0.3741949201,-0.2236362398,-0.201827243,-0.4316599369,-0.4015900493,-0.1702875495,-0.2987873554,-0.1874077171,-0.5853300691,0.0800147429,0.030056335,-0.1252746582,0.1624230742,-0.1581038833,0.231892392,0.2769706845,-0.0165657867,-0.192729637,-0.0813186616,-0.1340953261,0.0754104182,0.1756269485,0.038011238,0.3282631934,-0.1323553175,-0.1795409769,-0.0246300977,-0.4010378718,0.0776721239,0.0673702881,0.1397411227,0.304790169,0.1800093949,0.3312228322,-0.3353323936,0.1972529143,-0.2418714613,-0.2770200372,-0.0081244279,0.0024520406,0.1431996375,-0.2382315993,-0.4810171425,-0.4326365292,-0.5125101209,0.1009555608,0.0488469638,0.4066635966,0.2704299092,-0.2863166034,0.1968383491,-0.121017009,0.1822268218,-0.2584162951,-0.1776084006,0.250297606,-0.2715799809,-0.2704790533,-0.1265957206,-0.148699671,0.2048910111,-0.2855255306,-0.4239244759,-0.3295074403,-0.1458872259,-0.0527010113,0.1072003022,-0.0558419265,0.2296750695,-0.0121127972,-0.1654176414,-0.033512149,-0.3873255849,0.0534179732,0.1250744611,0.0702088997,-0.0927717984,0.6258087158,0.0622662865,0.5375280976,0.3718982339,0.0934443921,0.2314644009,0.1095010936,0.1484319568,-0.0122379875,-0.3427927792,0.2506149709,0.1518257856,0.0592511334,0.3488012552,-0.0516511016,-0.3857150078,-0.0800836086,-0.3046748042,-0.465890646,-0.2806211114,0.1811255217,0.1248168126,0.257915318,0.2710350752,0.1946982741,-0.1767562777,-0.1700816005,-0.1458094865,0.1727721691,0.0059876768,-0.0366931371,-0.1083944216,0.0141898142,-0.610968709,0.2600288689,0.0066701602,0.6111478806,-0.1570565253,-0.0673669726,-0.0168537963,-0.2866101563,0.3602798283,-0.3092956245,-0.0543416403,0.1828872859,-0.1095798239,-0.3564962447,-0.1205786094,-0.1527364999,0.0463261046,0.5042914748,0.5133962035,-0.2715702355,-0.2822863758,0.4173494279,0.1056818143,-0.1925458461,-0.2405329794,-0.2602197528,-0.2365468591,-0.3589510024,-0.0029185996,-0.3549587131,0.2806672752,0.0347830579,-0.0494445525,-0.0367941447,-0.2479297519,0.1712224931,0.0238339435,-0.0075592832,0.2060532123,0.1902969331,0.2586178482,0.3282211721,-0.2777053118,0.321195066,-0.0436938889,0.0135160172,-0.086697489,0.1257402003,0.0457252376,0.4861247241,0.0170813538,-0.0189167988,0.2016194761,0.3254525959,-0.0044429717,-0.0331222415,0.4496917725,-0.0619700886,-0.411162287,-0.1648898721,0.2997184098,0.1231901348,0.1361619532,-0.1300019771,-0.1973913908,-0.2437216341,0.1276027113,0.1129900441,0.7832134366,0.2591175437,0.0648233965,0.3565958738,-0.154997915,0.6783713102,-0.2112062573,0.2569588423,-0.3706091642,-0.3312232494,-0.0379182659,-0.110877268,0.3169820905,0.1003670096,-0.1770121157,0.4902181625,-0.1291067749,-0.0994330347,-0.0617544949,0.4637355208,-0.2605726421,-0.1433636397,-0.4001678526,0.1066973135,-0.1958624274,0.3526180387,-0.1023912877,-0.1949197054,0.2577715814,-0.2590811551,-0.0749825463,0.2773513794,-0.2684628367,0.1398487538,0.0061621419,0.2401976585,-0.2210501879,-0.3827387393,0.3137292862,-0.1355180293,-0.0385885052,-0.0691945553,-0.2621516585,-0.024181718,-0.1012062579,0.1060120016,0.4224326015,-0.2187235355,-0.1145567298,0.0203035623,-0.1885794401,-0.1630533636,-0.0189780127,-0.0354848094,0.1039875001,-0.1709197462,-0.1970958263,0.0954251364,-0.2310493588,-0.3154529631,0.146178931,-0.057743106,-0.2906800508,0.3077005148,-0.0858525261,-0.280965358,-0.2579652071,0.2445214242,0.1696913093,0.0508790649,0.4922745228,0.0225029569,-0.289804548,-0.2790828943,-0.0300642699,0.0263601001,-0.3613367677,0.1998145133,-0.0870138481,-0.0155756744,0.2755271792,0.2051900923,0.1997067034,0.2912752032,-0.2323927581,-0.5686837435,-0.1438727677,0.0629848167,-0.0053662611,0.2893784344,0.0466341376,0.5640015602,0.0183579158,0.3612833321,-0.3558589816,0.0370596088,-0.2532141507,0.1774706542,0.1644686908,-0.2871336341,0.4796305895,0.3361745477,0.1896099001,0.184483543,-0.3276686668,-0.2398814857,-0.1398786306,0.1459911615,0.0528673306,-0.1507482231,0.1091264933,0.0210518595,-0.1626609713,-0.0761597157,-0.0131698949,-0.1183911264,-0.0879904702,0.246785149,-0.0422548801,-0.0603146441,0.1202732995,0.0563640222,-0.1231211573,0.3887268901,-0.011862115,0.230624184,-0.2371812314,0.0459101312,-0.1055980399,-0.1075530127,0.0492279939,0.0140714366,0.2833747864,-0.4350683391,-0.2895481288,0.1623046845,0.231419459,0.2564552128,-0.0285925828,-0.1506037265,0.1654750556,0.2218014747,-0.3195352852,0.0245437622,0.3671759367,-0.0035081224,0.3218832314,0.3530758023,0.1187401861,0.0195527542,-0.0508777685,0.0890130997,0.2165187448,-0.2659980059,0.0368568152,0.2631331086,-0.1103861332,-0.0150807723,0.3015813529,-0.1218026355,0.3009927273,0.237793833,-0.0317757912,0.4294939041,-0.0397178307,0.3304133117,-0.0069701187,-0.2167673558,0.1839850694,0.4842499197,-0.1119999141,0.3832813501,0.0147898477,0.4105642736,-0.2803665698,0.0144452732,-0.177463308,0.4721761346,-0.1400122941,0.0913876519,-0.0336264521,-0.0987380594,-0.2867724299,-0.0064010476,0.078561306,0.0901599079,0.1506229341,0.1159298345,-0.0376419351,-0.3303664923,-0.1878765672,0.1589913517,0.2382988781,-0.3991445005,0.3074398935,0.2775505185,-0.0423321314,-0.1595377773,0.6101233363,0.3431236744,0.3848828971,-0.1388491094,-0.0203314312,0.1635613889,-0.0777899623,-0.0890924409,0.3411318362,0.1897569448,-0.0746393427,0.3646362722,0.1043184698,-0.2784147859,-0.046337042,0.0937767401,-0.2416163087,-0.1392454207,0.2932831347,-0.0494255424,0.2100278288,-0.222061947,0.0304331221,-0.6043111682,-0.0503048562,0.1934356242,-0.0807917118,0.2918081284,0.0341566429,0.1002931669,-0.1915131211,0.4911065996,0.0694363788,-0.0748937726,-0.3047780395,-0.2246827185,-0.5798493028,0.1475732327,-0.1549478322,-0.0564664453,0.0097489506,0.2813649476,-0.2015340179,0.049443271,0.0776405334,0.3546506464,-0.0404426381,-0.1575725079,-0.3835550249,0.0163158383,0.0054045385,-0.2268491983,0.0736885443,-0.4133264124,0.2439867556,-0.2011682093,0.0368155167,-0.1237300634,0.0597289316,-0.0689730644,0.4797645807,0.0129686212,0.5052636266,0.28163445,-0.2362820506,-0.3487403989,-0.1656108499,-0.1421875954,-0.2967919409,0.1579185128,0.3191114068,0.5874743462,0.017294988,0.0549576357,-0.3607902825,0.3801414073,0.0239438303,-0.1877516508,0.0241975896,0.3261172771,-0.2411894798,0.2674721181,0.2473868579,0.3860733807,0.1126281247,0.1490353644,-0.2446985543,-0.3713934422,0.5164570212,-0.1831858456,-0.2234716415,-0.0118958345,0.1257257313,0.0264652986,-0.1279568076,-0.4188547432,0.0567013286,0.3580688238,-0.1017922908,-0.0891665518,0.2000727803,0.2182815969,-0.2056010067,-0.0198608879,-0.0954742506,0.3953032792,-0.431453079,0.2040592879,-0.0773261264]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/331","title":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`","comments":"In general if some examples are missing after processing (hence causing the `NonMatchingSplitsSizesError `), it is often due to either\r\n1) corrupted cached files\r\n2) decoding errors\r\n\r\nI just checked the dataset script for code that could lead to decoding errors but I couldn't find any. Before we try to dive more into the processing of the dataset, could you try to clear your cache ? Just to make sure that it isn't 1)","body":"```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```","comment_length":74,"text":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`\n```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```\nIn general if some examples are missing after processing (hence causing the `NonMatchingSplitsSizesError `), it is often due to either\r\n1) corrupted cached files\r\n2) decoding errors\r\n\r\nI just checked the dataset script for code that could lead to decoding errors but I couldn't find any. Before we try to dive more into the processing of the dataset, could you try to clear your cache ? Just to make sure that it isn't 1)","embeddings":[-0.1410695314,-0.0569875538,-0.008213372,0.2086568773,0.0776493773,0.0907960758,0.1833027154,0.5251758695,0.0987313315,0.1542323083,0.0162980705,0.0769967064,-0.3141199946,-0.0437742993,-0.0335896686,0.0359017923,-0.1871679723,0.251511842,0.044887729,-0.0229966342,-0.2874323428,0.2817907035,-0.2087394744,-0.0366006866,-0.4835281372,-0.2941971123,0.0143964859,0.3698371649,0.037647862,-0.3093646169,-0.054115586,-0.1348574609,-0.1841386408,0.1029714793,-0.0001130321,0.1136059314,0.3911184371,-0.1433287412,-0.4867380261,0.0093758097,-0.210433796,-0.4906257391,0.0880899355,-0.354680568,-0.1552342176,0.0253197029,0.0945640802,-0.2195387632,0.4969272316,0.3904682994,0.2402192354,0.5636107922,0.0662313849,0.1633443981,-0.0887694806,-0.1809548885,-0.1458708197,-0.0048658135,-0.3178360164,-0.1510139853,-0.0242430959,0.4289468825,-0.0926206186,0.1253475994,-0.0872375593,0.1102501974,0.0889446437,-0.1658464372,0.093881309,0.1360183805,-0.0329074785,-0.0759069175,-0.2113918364,-0.4385608137,-0.0160641279,-0.3980173469,0.2251276523,0.2127174139,-0.1745027453,0.0653799996,-0.3212834597,0.0384899378,-0.0294721033,0.2034061104,0.0819332451,0.4111941159,-0.1002976,0.0258119442,0.0555451661,-0.0901165307,0.1567516327,-0.1769279093,0.1543263644,0.077836372,-0.1457687616,-0.045277901,0.1644648165,-0.1484151632,0.0430351757,-0.0675480068,0.1789103448,-0.0621050708,0.1744745076,0.0868742242,0.047065638,0.3125295639,-0.0163507275,0.268633306,0.1245981306,0.2852095664,0.0025192983,0.0181032717,0.1172021255,-0.0053176638,0.1188476905,0.2207996994,0.3734848201,-0.0592549704,-0.3276322186,0.1337225586,-0.2627150416,-0.2590237856,-0.0015641216,0.2562178373,-0.0583013371,0.3132656813,-0.0608430654,-0.0814521238,-0.3839427233,-0.1547316015,-0.3804471493,0.170080021,-0.1844006926,0.080809094,0.2795961797,-0.0796258152,0.457221657,-0.249753952,-0.0141821839,-0.14726381,0.2093333155,-0.039835576,0.0250497442,0.28318578,-0.130973205,0.1706077605,0.2590435743,0.1690691859,-0.2127455175,-0.0859967843,-0.0040091961,-0.605062902,0.1887341738,0.2487290353,0.059362527,-0.0686998665,0.6144146919,0.0758781433,0.2487533242,0.035048902,-0.1620905846,-0.0698391646,-0.0113207363,-0.1684927046,0.0756478012,0.4380223155,-0.2325228155,-0.2369064242,-0.0135412402,-0.1419267803,0.5154548883,0.4340136647,-0.1053208038,0.0439397804,-0.1265488267,0.1538269967,0.2003797293,-0.1545396298,-0.3955631256,0.1898384094,-0.0884127989,0.2039062083,0.2070534676,0.1967763603,-0.0294462275,0.1016341671,0.0539142787,0.5395390391,-0.2290878296,0.1055220217,-0.434442699,-0.4346444309,0.3582523465,0.0564582348,0.2241702229,-0.0408378653,-0.0627056584,0.4361121953,0.2115100026,-0.1110629141,0.0374838039,0.0333377346,-0.1516636014,-0.1143535376,0.0171922985,-0.0431710444,-0.4885517359,0.1375229061,-0.1571971476,0.3005177379,-0.050177224,-0.1014922038,-0.2627061307,-0.194733575,-0.1750023663,-0.2838497162,0.2178567648,0.2274239957,0.209719494,0.0044355784,-0.278423816,0.3022136688,-0.0314067528,0.1214026138,-0.5230283737,0.2627040744,-0.2279074043,-0.1810560226,0.2639639974,0.4298867285,0.2641758621,-0.1530768126,-0.2832179368,0.2833616138,-0.0861629993,0.231596604,-0.1480443329,-0.1591009498,0.1137663871,-0.0814554542,0.1697773635,0.4025975764,-0.03424301,-0.1033013612,-0.1191942096,0.4826929867,0.0484582894,0.0898778141,0.1499334574,-0.1141223013,0.0772690475,-0.1309580058,0.1130278036,-0.5237566233,0.3187466562,-0.169630602,0.2132307738,-0.0878212526,-0.2500077784,0.0530378111,0.6815172434,-0.0259576812,0.0975138694,0.0207601693,-0.1026206613,-0.2020810992,-0.1837860048,0.159725666,0.3386713266,0.2413775921,0.1866382957,0.2474578917,-0.1729585528,-0.286001265,0.27237764,0.095468618,0.2538036108,0.385949403,0.0060878471,-0.0024015717,-0.5314667225,-0.0248315446,0.013506908,0.3741949201,-0.2236362398,-0.201827243,-0.4316599369,-0.4015900493,-0.1702875495,-0.2987873554,-0.1874077171,-0.5853300691,0.0800147429,0.030056335,-0.1252746582,0.1624230742,-0.1581038833,0.231892392,0.2769706845,-0.0165657867,-0.192729637,-0.0813186616,-0.1340953261,0.0754104182,0.1756269485,0.038011238,0.3282631934,-0.1323553175,-0.1795409769,-0.0246300977,-0.4010378718,0.0776721239,0.0673702881,0.1397411227,0.304790169,0.1800093949,0.3312228322,-0.3353323936,0.1972529143,-0.2418714613,-0.2770200372,-0.0081244279,0.0024520406,0.1431996375,-0.2382315993,-0.4810171425,-0.4326365292,-0.5125101209,0.1009555608,0.0488469638,0.4066635966,0.2704299092,-0.2863166034,0.1968383491,-0.121017009,0.1822268218,-0.2584162951,-0.1776084006,0.250297606,-0.2715799809,-0.2704790533,-0.1265957206,-0.148699671,0.2048910111,-0.2855255306,-0.4239244759,-0.3295074403,-0.1458872259,-0.0527010113,0.1072003022,-0.0558419265,0.2296750695,-0.0121127972,-0.1654176414,-0.033512149,-0.3873255849,0.0534179732,0.1250744611,0.0702088997,-0.0927717984,0.6258087158,0.0622662865,0.5375280976,0.3718982339,0.0934443921,0.2314644009,0.1095010936,0.1484319568,-0.0122379875,-0.3427927792,0.2506149709,0.1518257856,0.0592511334,0.3488012552,-0.0516511016,-0.3857150078,-0.0800836086,-0.3046748042,-0.465890646,-0.2806211114,0.1811255217,0.1248168126,0.257915318,0.2710350752,0.1946982741,-0.1767562777,-0.1700816005,-0.1458094865,0.1727721691,0.0059876768,-0.0366931371,-0.1083944216,0.0141898142,-0.610968709,0.2600288689,0.0066701602,0.6111478806,-0.1570565253,-0.0673669726,-0.0168537963,-0.2866101563,0.3602798283,-0.3092956245,-0.0543416403,0.1828872859,-0.1095798239,-0.3564962447,-0.1205786094,-0.1527364999,0.0463261046,0.5042914748,0.5133962035,-0.2715702355,-0.2822863758,0.4173494279,0.1056818143,-0.1925458461,-0.2405329794,-0.2602197528,-0.2365468591,-0.3589510024,-0.0029185996,-0.3549587131,0.2806672752,0.0347830579,-0.0494445525,-0.0367941447,-0.2479297519,0.1712224931,0.0238339435,-0.0075592832,0.2060532123,0.1902969331,0.2586178482,0.3282211721,-0.2777053118,0.321195066,-0.0436938889,0.0135160172,-0.086697489,0.1257402003,0.0457252376,0.4861247241,0.0170813538,-0.0189167988,0.2016194761,0.3254525959,-0.0044429717,-0.0331222415,0.4496917725,-0.0619700886,-0.411162287,-0.1648898721,0.2997184098,0.1231901348,0.1361619532,-0.1300019771,-0.1973913908,-0.2437216341,0.1276027113,0.1129900441,0.7832134366,0.2591175437,0.0648233965,0.3565958738,-0.154997915,0.6783713102,-0.2112062573,0.2569588423,-0.3706091642,-0.3312232494,-0.0379182659,-0.110877268,0.3169820905,0.1003670096,-0.1770121157,0.4902181625,-0.1291067749,-0.0994330347,-0.0617544949,0.4637355208,-0.2605726421,-0.1433636397,-0.4001678526,0.1066973135,-0.1958624274,0.3526180387,-0.1023912877,-0.1949197054,0.2577715814,-0.2590811551,-0.0749825463,0.2773513794,-0.2684628367,0.1398487538,0.0061621419,0.2401976585,-0.2210501879,-0.3827387393,0.3137292862,-0.1355180293,-0.0385885052,-0.0691945553,-0.2621516585,-0.024181718,-0.1012062579,0.1060120016,0.4224326015,-0.2187235355,-0.1145567298,0.0203035623,-0.1885794401,-0.1630533636,-0.0189780127,-0.0354848094,0.1039875001,-0.1709197462,-0.1970958263,0.0954251364,-0.2310493588,-0.3154529631,0.146178931,-0.057743106,-0.2906800508,0.3077005148,-0.0858525261,-0.280965358,-0.2579652071,0.2445214242,0.1696913093,0.0508790649,0.4922745228,0.0225029569,-0.289804548,-0.2790828943,-0.0300642699,0.0263601001,-0.3613367677,0.1998145133,-0.0870138481,-0.0155756744,0.2755271792,0.2051900923,0.1997067034,0.2912752032,-0.2323927581,-0.5686837435,-0.1438727677,0.0629848167,-0.0053662611,0.2893784344,0.0466341376,0.5640015602,0.0183579158,0.3612833321,-0.3558589816,0.0370596088,-0.2532141507,0.1774706542,0.1644686908,-0.2871336341,0.4796305895,0.3361745477,0.1896099001,0.184483543,-0.3276686668,-0.2398814857,-0.1398786306,0.1459911615,0.0528673306,-0.1507482231,0.1091264933,0.0210518595,-0.1626609713,-0.0761597157,-0.0131698949,-0.1183911264,-0.0879904702,0.246785149,-0.0422548801,-0.0603146441,0.1202732995,0.0563640222,-0.1231211573,0.3887268901,-0.011862115,0.230624184,-0.2371812314,0.0459101312,-0.1055980399,-0.1075530127,0.0492279939,0.0140714366,0.2833747864,-0.4350683391,-0.2895481288,0.1623046845,0.231419459,0.2564552128,-0.0285925828,-0.1506037265,0.1654750556,0.2218014747,-0.3195352852,0.0245437622,0.3671759367,-0.0035081224,0.3218832314,0.3530758023,0.1187401861,0.0195527542,-0.0508777685,0.0890130997,0.2165187448,-0.2659980059,0.0368568152,0.2631331086,-0.1103861332,-0.0150807723,0.3015813529,-0.1218026355,0.3009927273,0.237793833,-0.0317757912,0.4294939041,-0.0397178307,0.3304133117,-0.0069701187,-0.2167673558,0.1839850694,0.4842499197,-0.1119999141,0.3832813501,0.0147898477,0.4105642736,-0.2803665698,0.0144452732,-0.177463308,0.4721761346,-0.1400122941,0.0913876519,-0.0336264521,-0.0987380594,-0.2867724299,-0.0064010476,0.078561306,0.0901599079,0.1506229341,0.1159298345,-0.0376419351,-0.3303664923,-0.1878765672,0.1589913517,0.2382988781,-0.3991445005,0.3074398935,0.2775505185,-0.0423321314,-0.1595377773,0.6101233363,0.3431236744,0.3848828971,-0.1388491094,-0.0203314312,0.1635613889,-0.0777899623,-0.0890924409,0.3411318362,0.1897569448,-0.0746393427,0.3646362722,0.1043184698,-0.2784147859,-0.046337042,0.0937767401,-0.2416163087,-0.1392454207,0.2932831347,-0.0494255424,0.2100278288,-0.222061947,0.0304331221,-0.6043111682,-0.0503048562,0.1934356242,-0.0807917118,0.2918081284,0.0341566429,0.1002931669,-0.1915131211,0.4911065996,0.0694363788,-0.0748937726,-0.3047780395,-0.2246827185,-0.5798493028,0.1475732327,-0.1549478322,-0.0564664453,0.0097489506,0.2813649476,-0.2015340179,0.049443271,0.0776405334,0.3546506464,-0.0404426381,-0.1575725079,-0.3835550249,0.0163158383,0.0054045385,-0.2268491983,0.0736885443,-0.4133264124,0.2439867556,-0.2011682093,0.0368155167,-0.1237300634,0.0597289316,-0.0689730644,0.4797645807,0.0129686212,0.5052636266,0.28163445,-0.2362820506,-0.3487403989,-0.1656108499,-0.1421875954,-0.2967919409,0.1579185128,0.3191114068,0.5874743462,0.017294988,0.0549576357,-0.3607902825,0.3801414073,0.0239438303,-0.1877516508,0.0241975896,0.3261172771,-0.2411894798,0.2674721181,0.2473868579,0.3860733807,0.1126281247,0.1490353644,-0.2446985543,-0.3713934422,0.5164570212,-0.1831858456,-0.2234716415,-0.0118958345,0.1257257313,0.0264652986,-0.1279568076,-0.4188547432,0.0567013286,0.3580688238,-0.1017922908,-0.0891665518,0.2000727803,0.2182815969,-0.2056010067,-0.0198608879,-0.0954742506,0.3953032792,-0.431453079,0.2040592879,-0.0773261264]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/331","title":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`","comments":"Yes thanks for the support! I cleared out my cache folder and everything works fine now","body":"```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```","comment_length":16,"text":"Loading CNN\/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`\n```\r\n>>> import nlp\r\n>>> nlp.load_dataset('cnn_dailymail', '3.0.0')\r\nDownloading and preparing dataset cnn_dailymail\/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to \/u\/jm8wx\/.cache\/huggingface\/datasets\/cnn_dailymail\/3.0.0\/3.0.0...\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/load.py\", line 520, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 431, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/p\/qdata\/jm8wx\/datasets\/nlp\/src\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=49424491, num_examples=11490, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='test', num_bytes=48931393, num_examples=11379, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='train', num_bytes=1249178681, num_examples=287113, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='train', num_bytes=1240618482, num_examples=285161, dataset_name='cnn_dailymail')}, {'expected': SplitInfo(name='validation', num_bytes=57149241, num_examples=13368, dataset_name='cnn_dailymail'), 'recorded': SplitInfo(name='validation', num_bytes=56637485, num_examples=13255, dataset_name='cnn_dailymail')}]\r\n```\nYes thanks for the support! I cleared out my cache folder and everything works fine now","embeddings":[-0.1410695314,-0.0569875538,-0.008213372,0.2086568773,0.0776493773,0.0907960758,0.1833027154,0.5251758695,0.0987313315,0.1542323083,0.0162980705,0.0769967064,-0.3141199946,-0.0437742993,-0.0335896686,0.0359017923,-0.1871679723,0.251511842,0.044887729,-0.0229966342,-0.2874323428,0.2817907035,-0.2087394744,-0.0366006866,-0.4835281372,-0.2941971123,0.0143964859,0.3698371649,0.037647862,-0.3093646169,-0.054115586,-0.1348574609,-0.1841386408,0.1029714793,-0.0001130321,0.1136059314,0.3911184371,-0.1433287412,-0.4867380261,0.0093758097,-0.210433796,-0.4906257391,0.0880899355,-0.354680568,-0.1552342176,0.0253197029,0.0945640802,-0.2195387632,0.4969272316,0.3904682994,0.2402192354,0.5636107922,0.0662313849,0.1633443981,-0.0887694806,-0.1809548885,-0.1458708197,-0.0048658135,-0.3178360164,-0.1510139853,-0.0242430959,0.4289468825,-0.0926206186,0.1253475994,-0.0872375593,0.1102501974,0.0889446437,-0.1658464372,0.093881309,0.1360183805,-0.0329074785,-0.0759069175,-0.2113918364,-0.4385608137,-0.0160641279,-0.3980173469,0.2251276523,0.2127174139,-0.1745027453,0.0653799996,-0.3212834597,0.0384899378,-0.0294721033,0.2034061104,0.0819332451,0.4111941159,-0.1002976,0.0258119442,0.0555451661,-0.0901165307,0.1567516327,-0.1769279093,0.1543263644,0.077836372,-0.1457687616,-0.045277901,0.1644648165,-0.1484151632,0.0430351757,-0.0675480068,0.1789103448,-0.0621050708,0.1744745076,0.0868742242,0.047065638,0.3125295639,-0.0163507275,0.268633306,0.1245981306,0.2852095664,0.0025192983,0.0181032717,0.1172021255,-0.0053176638,0.1188476905,0.2207996994,0.3734848201,-0.0592549704,-0.3276322186,0.1337225586,-0.2627150416,-0.2590237856,-0.0015641216,0.2562178373,-0.0583013371,0.3132656813,-0.0608430654,-0.0814521238,-0.3839427233,-0.1547316015,-0.3804471493,0.170080021,-0.1844006926,0.080809094,0.2795961797,-0.0796258152,0.457221657,-0.249753952,-0.0141821839,-0.14726381,0.2093333155,-0.039835576,0.0250497442,0.28318578,-0.130973205,0.1706077605,0.2590435743,0.1690691859,-0.2127455175,-0.0859967843,-0.0040091961,-0.605062902,0.1887341738,0.2487290353,0.059362527,-0.0686998665,0.6144146919,0.0758781433,0.2487533242,0.035048902,-0.1620905846,-0.0698391646,-0.0113207363,-0.1684927046,0.0756478012,0.4380223155,-0.2325228155,-0.2369064242,-0.0135412402,-0.1419267803,0.5154548883,0.4340136647,-0.1053208038,0.0439397804,-0.1265488267,0.1538269967,0.2003797293,-0.1545396298,-0.3955631256,0.1898384094,-0.0884127989,0.2039062083,0.2070534676,0.1967763603,-0.0294462275,0.1016341671,0.0539142787,0.5395390391,-0.2290878296,0.1055220217,-0.434442699,-0.4346444309,0.3582523465,0.0564582348,0.2241702229,-0.0408378653,-0.0627056584,0.4361121953,0.2115100026,-0.1110629141,0.0374838039,0.0333377346,-0.1516636014,-0.1143535376,0.0171922985,-0.0431710444,-0.4885517359,0.1375229061,-0.1571971476,0.3005177379,-0.050177224,-0.1014922038,-0.2627061307,-0.194733575,-0.1750023663,-0.2838497162,0.2178567648,0.2274239957,0.209719494,0.0044355784,-0.278423816,0.3022136688,-0.0314067528,0.1214026138,-0.5230283737,0.2627040744,-0.2279074043,-0.1810560226,0.2639639974,0.4298867285,0.2641758621,-0.1530768126,-0.2832179368,0.2833616138,-0.0861629993,0.231596604,-0.1480443329,-0.1591009498,0.1137663871,-0.0814554542,0.1697773635,0.4025975764,-0.03424301,-0.1033013612,-0.1191942096,0.4826929867,0.0484582894,0.0898778141,0.1499334574,-0.1141223013,0.0772690475,-0.1309580058,0.1130278036,-0.5237566233,0.3187466562,-0.169630602,0.2132307738,-0.0878212526,-0.2500077784,0.0530378111,0.6815172434,-0.0259576812,0.0975138694,0.0207601693,-0.1026206613,-0.2020810992,-0.1837860048,0.159725666,0.3386713266,0.2413775921,0.1866382957,0.2474578917,-0.1729585528,-0.286001265,0.27237764,0.095468618,0.2538036108,0.385949403,0.0060878471,-0.0024015717,-0.5314667225,-0.0248315446,0.013506908,0.3741949201,-0.2236362398,-0.201827243,-0.4316599369,-0.4015900493,-0.1702875495,-0.2987873554,-0.1874077171,-0.5853300691,0.0800147429,0.030056335,-0.1252746582,0.1624230742,-0.1581038833,0.231892392,0.2769706845,-0.0165657867,-0.192729637,-0.0813186616,-0.1340953261,0.0754104182,0.1756269485,0.038011238,0.3282631934,-0.1323553175,-0.1795409769,-0.0246300977,-0.4010378718,0.0776721239,0.0673702881,0.1397411227,0.304790169,0.1800093949,0.3312228322,-0.3353323936,0.1972529143,-0.2418714613,-0.2770200372,-0.0081244279,0.0024520406,0.1431996375,-0.2382315993,-0.4810171425,-0.4326365292,-0.5125101209,0.1009555608,0.0488469638,0.4066635966,0.2704299092,-0.2863166034,0.1968383491,-0.121017009,0.1822268218,-0.2584162951,-0.1776084006,0.250297606,-0.2715799809,-0.2704790533,-0.1265957206,-0.148699671,0.2048910111,-0.2855255306,-0.4239244759,-0.3295074403,-0.1458872259,-0.0527010113,0.1072003022,-0.0558419265,0.2296750695,-0.0121127972,-0.1654176414,-0.033512149,-0.3873255849,0.0534179732,0.1250744611,0.0702088997,-0.0927717984,0.6258087158,0.0622662865,0.5375280976,0.3718982339,0.0934443921,0.2314644009,0.1095010936,0.1484319568,-0.0122379875,-0.3427927792,0.2506149709,0.1518257856,0.0592511334,0.3488012552,-0.0516511016,-0.3857150078,-0.0800836086,-0.3046748042,-0.465890646,-0.2806211114,0.1811255217,0.1248168126,0.257915318,0.2710350752,0.1946982741,-0.1767562777,-0.1700816005,-0.1458094865,0.1727721691,0.0059876768,-0.0366931371,-0.1083944216,0.0141898142,-0.610968709,0.2600288689,0.0066701602,0.6111478806,-0.1570565253,-0.0673669726,-0.0168537963,-0.2866101563,0.3602798283,-0.3092956245,-0.0543416403,0.1828872859,-0.1095798239,-0.3564962447,-0.1205786094,-0.1527364999,0.0463261046,0.5042914748,0.5133962035,-0.2715702355,-0.2822863758,0.4173494279,0.1056818143,-0.1925458461,-0.2405329794,-0.2602197528,-0.2365468591,-0.3589510024,-0.0029185996,-0.3549587131,0.2806672752,0.0347830579,-0.0494445525,-0.0367941447,-0.2479297519,0.1712224931,0.0238339435,-0.0075592832,0.2060532123,0.1902969331,0.2586178482,0.3282211721,-0.2777053118,0.321195066,-0.0436938889,0.0135160172,-0.086697489,0.1257402003,0.0457252376,0.4861247241,0.0170813538,-0.0189167988,0.2016194761,0.3254525959,-0.0044429717,-0.0331222415,0.4496917725,-0.0619700886,-0.411162287,-0.1648898721,0.2997184098,0.1231901348,0.1361619532,-0.1300019771,-0.1973913908,-0.2437216341,0.1276027113,0.1129900441,0.7832134366,0.2591175437,0.0648233965,0.3565958738,-0.154997915,0.6783713102,-0.2112062573,0.2569588423,-0.3706091642,-0.3312232494,-0.0379182659,-0.110877268,0.3169820905,0.1003670096,-0.1770121157,0.4902181625,-0.1291067749,-0.0994330347,-0.0617544949,0.4637355208,-0.2605726421,-0.1433636397,-0.4001678526,0.1066973135,-0.1958624274,0.3526180387,-0.1023912877,-0.1949197054,0.2577715814,-0.2590811551,-0.0749825463,0.2773513794,-0.2684628367,0.1398487538,0.0061621419,0.2401976585,-0.2210501879,-0.3827387393,0.3137292862,-0.1355180293,-0.0385885052,-0.0691945553,-0.2621516585,-0.024181718,-0.1012062579,0.1060120016,0.4224326015,-0.2187235355,-0.1145567298,0.0203035623,-0.1885794401,-0.1630533636,-0.0189780127,-0.0354848094,0.1039875001,-0.1709197462,-0.1970958263,0.0954251364,-0.2310493588,-0.3154529631,0.146178931,-0.057743106,-0.2906800508,0.3077005148,-0.0858525261,-0.280965358,-0.2579652071,0.2445214242,0.1696913093,0.0508790649,0.4922745228,0.0225029569,-0.289804548,-0.2790828943,-0.0300642699,0.0263601001,-0.3613367677,0.1998145133,-0.0870138481,-0.0155756744,0.2755271792,0.2051900923,0.1997067034,0.2912752032,-0.2323927581,-0.5686837435,-0.1438727677,0.0629848167,-0.0053662611,0.2893784344,0.0466341376,0.5640015602,0.0183579158,0.3612833321,-0.3558589816,0.0370596088,-0.2532141507,0.1774706542,0.1644686908,-0.2871336341,0.4796305895,0.3361745477,0.1896099001,0.184483543,-0.3276686668,-0.2398814857,-0.1398786306,0.1459911615,0.0528673306,-0.1507482231,0.1091264933,0.0210518595,-0.1626609713,-0.0761597157,-0.0131698949,-0.1183911264,-0.0879904702,0.246785149,-0.0422548801,-0.0603146441,0.1202732995,0.0563640222,-0.1231211573,0.3887268901,-0.011862115,0.230624184,-0.2371812314,0.0459101312,-0.1055980399,-0.1075530127,0.0492279939,0.0140714366,0.2833747864,-0.4350683391,-0.2895481288,0.1623046845,0.231419459,0.2564552128,-0.0285925828,-0.1506037265,0.1654750556,0.2218014747,-0.3195352852,0.0245437622,0.3671759367,-0.0035081224,0.3218832314,0.3530758023,0.1187401861,0.0195527542,-0.0508777685,0.0890130997,0.2165187448,-0.2659980059,0.0368568152,0.2631331086,-0.1103861332,-0.0150807723,0.3015813529,-0.1218026355,0.3009927273,0.237793833,-0.0317757912,0.4294939041,-0.0397178307,0.3304133117,-0.0069701187,-0.2167673558,0.1839850694,0.4842499197,-0.1119999141,0.3832813501,0.0147898477,0.4105642736,-0.2803665698,0.0144452732,-0.177463308,0.4721761346,-0.1400122941,0.0913876519,-0.0336264521,-0.0987380594,-0.2867724299,-0.0064010476,0.078561306,0.0901599079,0.1506229341,0.1159298345,-0.0376419351,-0.3303664923,-0.1878765672,0.1589913517,0.2382988781,-0.3991445005,0.3074398935,0.2775505185,-0.0423321314,-0.1595377773,0.6101233363,0.3431236744,0.3848828971,-0.1388491094,-0.0203314312,0.1635613889,-0.0777899623,-0.0890924409,0.3411318362,0.1897569448,-0.0746393427,0.3646362722,0.1043184698,-0.2784147859,-0.046337042,0.0937767401,-0.2416163087,-0.1392454207,0.2932831347,-0.0494255424,0.2100278288,-0.222061947,0.0304331221,-0.6043111682,-0.0503048562,0.1934356242,-0.0807917118,0.2918081284,0.0341566429,0.1002931669,-0.1915131211,0.4911065996,0.0694363788,-0.0748937726,-0.3047780395,-0.2246827185,-0.5798493028,0.1475732327,-0.1549478322,-0.0564664453,0.0097489506,0.2813649476,-0.2015340179,0.049443271,0.0776405334,0.3546506464,-0.0404426381,-0.1575725079,-0.3835550249,0.0163158383,0.0054045385,-0.2268491983,0.0736885443,-0.4133264124,0.2439867556,-0.2011682093,0.0368155167,-0.1237300634,0.0597289316,-0.0689730644,0.4797645807,0.0129686212,0.5052636266,0.28163445,-0.2362820506,-0.3487403989,-0.1656108499,-0.1421875954,-0.2967919409,0.1579185128,0.3191114068,0.5874743462,0.017294988,0.0549576357,-0.3607902825,0.3801414073,0.0239438303,-0.1877516508,0.0241975896,0.3261172771,-0.2411894798,0.2674721181,0.2473868579,0.3860733807,0.1126281247,0.1490353644,-0.2446985543,-0.3713934422,0.5164570212,-0.1831858456,-0.2234716415,-0.0118958345,0.1257257313,0.0264652986,-0.1279568076,-0.4188547432,0.0567013286,0.3580688238,-0.1017922908,-0.0891665518,0.2000727803,0.2182815969,-0.2056010067,-0.0198608879,-0.0954742506,0.3953032792,-0.431453079,0.2040592879,-0.0773261264]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/329","title":"[Bug] FileLock dependency incompatible with filesystem","comments":"Hi, can you give details on your environment\/os\/packages versions\/etc?","body":"I'm downloading a dataset successfully with\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")`\r\n\r\nBut when I attempt to cache it on an external volume, it hangs indefinitely:\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", cache_dir=\"\/fsx\") # \/fsx is an external volume mount`\r\n\r\nThe filesystem when hanging looks like this:\r\n```bash\r\n\/fsx\r\n----downloads\r\n ----94be...73.lock\r\n----wikitext\r\n ----wikitext-2-raw\r\n ----wikitext-2-raw-1.0.0.incomplete\r\n```\r\n\r\nIt appears that on this filesystem, the FileLock object is forever stuck in its \"acquire\" stage. I have verified that the issue lies specifically with the `filelock` dependency:\r\n```python\r\nopen(\"\/fsx\/hello.txt\").write(\"hello\") # succeeds\r\n\r\nfrom filelock import FileLock\r\nwith FileLock(\"\/fsx\/hello.lock\"):\r\n open(\"\/fsx\/hello.txt\").write(\"hello\") # hangs indefinitely\r\n```\r\n\r\nHas anyone else run into this issue? I'd raise it directly on the FileLock repo, but that project appears abandoned with the last update over a year ago. Or if there's a solution that would remove the FileLock dependency from the project, I would appreciate that.","comment_length":9,"text":"[Bug] FileLock dependency incompatible with filesystem\nI'm downloading a dataset successfully with\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")`\r\n\r\nBut when I attempt to cache it on an external volume, it hangs indefinitely:\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", cache_dir=\"\/fsx\") # \/fsx is an external volume mount`\r\n\r\nThe filesystem when hanging looks like this:\r\n```bash\r\n\/fsx\r\n----downloads\r\n ----94be...73.lock\r\n----wikitext\r\n ----wikitext-2-raw\r\n ----wikitext-2-raw-1.0.0.incomplete\r\n```\r\n\r\nIt appears that on this filesystem, the FileLock object is forever stuck in its \"acquire\" stage. I have verified that the issue lies specifically with the `filelock` dependency:\r\n```python\r\nopen(\"\/fsx\/hello.txt\").write(\"hello\") # succeeds\r\n\r\nfrom filelock import FileLock\r\nwith FileLock(\"\/fsx\/hello.lock\"):\r\n open(\"\/fsx\/hello.txt\").write(\"hello\") # hangs indefinitely\r\n```\r\n\r\nHas anyone else run into this issue? I'd raise it directly on the FileLock repo, but that project appears abandoned with the last update over a year ago. Or if there's a solution that would remove the FileLock dependency from the project, I would appreciate that.\nHi, can you give details on your environment\/os\/packages versions\/etc?","embeddings":[-0.0051458799,-0.006434917,-0.023097042,0.0569061227,0.1029243693,0.1768240333,0.3489715755,0.0676314756,0.7186329961,-0.0481542423,0.0739284456,-0.0319546536,-0.0145170791,-0.2596742511,-0.2090379,0.2100241184,0.2105367631,0.103807807,-0.108027406,-0.0017610808,-0.142353639,0.3784559071,-0.2225159705,-0.2870797515,-0.7657347918,0.1493160725,-0.0558577999,0.0275828298,-0.2707179487,-0.0553404726,0.4443973601,0.5364134312,-0.0772386342,0.5501334071,-0.0001238893,-0.2436811328,0.2469676584,-0.0220041536,-0.2320673317,-0.2228654027,-0.367403239,-0.3008428514,0.0605557039,-0.0754374638,0.228399545,-0.148032546,0.0687833875,-0.5430995822,0.2996369302,-0.1221754849,0.0619475059,0.1405128986,0.2092956901,-0.1468916088,0.3337991834,-0.1586626321,-0.1703431159,0.5384657383,0.3551771045,-0.1736550331,-0.0764872804,-0.0105772149,0.0050884341,-0.1075644195,0.5793908834,-0.2247950882,0.3235572577,-0.2335931957,-0.0509173237,0.3053613901,0.7499597073,-0.0396958701,-0.5516858101,-0.1913976073,0.3568972647,0.0653154403,0.5682668686,0.0095890136,-0.072962828,0.4147792161,-0.0323998369,-0.1251830459,0.0757900253,0.0197569933,-0.198469013,0.0624974743,0.1741357744,0.0138848033,0.2375921905,-0.0876510814,0.2463193983,-0.2200170457,-0.0333422013,0.1242173389,-0.4771464169,0.1241765395,-0.0029685565,0.1694346219,-0.0638041422,-0.253171891,-0.564522624,0.1579178423,0.0351250842,0.3261380196,0.2397553176,0.1052264869,0.0779714137,-0.0196355544,0.2429417819,-0.0233739987,-0.2562952638,0.0824310705,-0.0872494355,-0.1705443114,0.1274344176,-0.0615641996,0.0697957203,-0.4429310858,0.1263566166,0.216888994,0.0912978649,0.3309181631,0.1446394473,0.0585981682,0.0316762105,0.4786789417,0.1361235976,0.0084523018,-0.1935048103,0.3044184446,-0.0461260453,-0.1455347091,0.024029227,0.1204938293,0.0739762858,0.2295536995,0.1277501732,0.211555168,0.3256307244,0.1497673541,0.1154222712,-0.1844460815,-0.1837358773,0.3704171479,-0.1545003057,0.1282206178,0.0033130115,0.1097828671,-0.0836314261,0.0502104498,-0.2080678642,0.0759325549,-0.0412138589,0.0290105641,-0.4034745395,0.0482066944,0.1677766591,-0.192570433,-0.1786362082,-0.3774959147,-0.0274713933,-0.4284633994,-0.0491789244,-0.0054230951,-0.0073231016,0.3914173543,-0.2924706042,-0.0012966282,0.2290174365,-0.0420361869,-0.1474086046,0.2481583059,-0.0894012973,0.0644752979,-0.218806386,0.0837035701,0.0046501332,-0.4911375344,-0.8228986263,0.2386667728,-0.2548437119,0.2746059895,0.0682282522,0.3686966598,0.099637337,-0.2135257125,0.1105177402,0.4422155619,0.0885902047,-0.1854247749,-0.0665686652,-0.3366092443,-0.1368258595,0.1575407535,0.0612637736,0.2221339196,0.0013373959,0.6355970502,0.2140434384,0.0461026914,-0.0336862244,0.5561134815,0.6533918381,0.1290879846,0.0063192146,-0.1375551075,-0.6586831212,0.437874496,-0.3373485506,-0.1379074603,-0.1197265908,-0.1258791089,0.0303371698,-0.0273667742,-0.0870487839,0.262860924,0.0348383412,-0.0839493796,0.0873987451,0.220301345,0.0988999084,0.4394233227,0.1315366775,-0.0392653793,-0.0225532539,-0.0464556329,0.1017223895,-0.1632863134,0.0850967094,-0.0974453315,0.1395215839,-0.2140325308,-0.2081384361,0.2323244959,0.4467125833,0.2572504878,-0.2211449444,0.2811659276,0.0519132167,0.3171530962,-0.1442207843,-0.0972058326,0.0948994607,-0.2504143715,0.0447855517,0.0680080876,-0.2800655663,0.0154604008,-0.2316338122,0.0692373589,0.1111964285,0.0240206104,-0.1297218204,-0.0283195954,0.2801696956,-0.0740204304,0.3053816557,-0.1333345026,0.574808538,0.0418593809,0.4432496428,0.1388012618,0.0671314672,0.0436834879,0.0342043079,-0.098319903,0.0116606597,0.219045192,0.5284807682,0.1247460619,0.288451612,-0.2549062669,0.1788540781,-0.263068825,0.0904921964,-0.1602436304,-0.2676975429,0.3271216154,0.0570269227,-0.1118504927,0.1407175362,-0.244668588,0.1115102321,0.0475446694,0.1434467286,0.0014795386,-0.4006223083,0.5140924454,0.2210192829,0.084671244,-0.2062919885,-0.426869303,0.0848683938,0.6912555695,0.2113793939,0.1028392762,-0.4054599702,0.0459265932,0.0550285764,-0.330584228,-0.0822195336,-0.3372192383,-0.0655180886,-0.1375742257,0.5970637798,0.0396894142,0.1679532081,0.0591252856,-0.2448493987,-0.3836550713,-0.1776441932,0.1402826607,-0.0584251732,0.3586268425,0.0804321095,0.2585320771,-0.1020003706,-0.0026291024,0.0215785615,-0.3474665284,-0.0158022828,-0.0079600047,0.0100870477,-0.0750296265,-0.2225087881,0.2019271404,-0.1715212613,-0.3563611209,0.2733250558,-0.418936938,-0.2606914043,0.2763243318,0.0989041775,0.2519560754,0.0050617489,-0.1673525572,-0.0602216423,-0.4327980876,0.2576298416,0.0568029918,-0.4156255424,-0.115618892,0.1915606558,-0.4260771871,0.2648839056,-0.5505449176,-0.1198190525,-0.3470570743,0.3366925418,-0.0490712486,-0.0845613405,0.4450289607,-0.2618167102,0.0631129518,-0.4211976826,0.2156774551,0.1754195243,0.0210794844,0.1116155013,0.0362349562,-0.1280689389,0.0882440954,0.4544968605,-0.0162998959,-0.0220635161,0.2254147828,0.3973492682,0.4397190809,-0.3661363423,-0.214165777,-0.0751933008,-0.1020333916,-0.1796655357,0.0045937914,-0.1574067622,-0.1499622762,-0.2094130814,-0.0976835191,-0.1089550406,-0.2470623255,0.0152617116,0.1828758717,0.1043389812,0.2475499213,-0.0341986194,-0.1322229207,-0.1541196704,0.2271750122,0.2479459196,0.4775480032,0.0589688607,-0.0801916271,0.2092338502,-0.1176044568,0.1986075938,0.0666328445,0.3798116148,-0.1008370146,-0.0011332444,-0.0798099935,-0.3585930765,0.3408061266,-0.0020136621,0.1401734799,0.2244683206,0.1709966213,-0.2604631484,-0.2146520615,-0.1219788492,0.192137897,0.3844896853,0.1594565958,-0.1860814095,-0.1979216486,-0.5487331152,0.4058345854,-0.1047522873,-0.586991787,-0.1629507244,0.2751255631,-0.1847477108,-0.0205727536,-0.1309809238,0.0432014763,0.0658026263,0.0696237832,-0.2767861187,-0.1149477884,-0.1248469427,-0.0448609814,0.0713566616,-0.1492038369,0.3688107729,0.098371692,-0.2098211944,0.2243334353,0.2174407691,0.1344234049,0.1086927578,0.3440109491,-0.2486539483,-0.2269686759,0.4995743632,-0.0188218709,0.0450478792,-0.0981452018,-0.0734648332,-0.0187166054,0.0456379876,0.1966689229,-0.0173364356,0.3008484542,0.0771007016,0.0905547291,0.1251281351,-0.2163981348,0.5388244987,0.4112693965,-0.0100507615,-0.0234427415,-0.0487752669,0.9047334194,0.0975800827,0.1217891425,0.002096985,-0.021683082,0.1804639697,-0.4295860529,0.0216312446,-0.0679853484,-0.4163114429,-0.0074268654,-0.3239558935,0.3726127446,-0.1291779578,-0.0628537983,0.1282804757,-0.3083564937,-0.0427160189,-0.1931148767,0.3133723736,-0.355394274,-0.2900305688,-0.1255848259,0.1579557359,0.2095347196,0.1897312254,-0.1858265102,-0.1235070601,-0.2229889184,-0.0310182124,-0.2559341788,0.0443811156,-0.0539264306,0.4061546326,-0.436812669,0.1668939292,-0.092166096,0.1023743823,0.1704030633,0.3882990777,-0.3845307231,0.1415048242,-0.115243569,-0.1511906683,0.1479870528,-0.1086193472,0.2150575519,-0.1137710065,-0.1414888501,-0.2190354764,-0.1096993834,-0.2332288772,0.1650853157,-0.1878620237,0.1101803854,0.1652031541,-0.8146268725,0.0595019013,-0.0490199141,0.0316501111,0.0586470217,0.1954089403,-0.1033608764,0.2391622066,-0.3918167055,-0.0928435996,-0.0407262817,0.6245103478,-0.2116547972,-0.4304094017,0.3698435724,0.036612168,-0.1283735335,-0.1848147362,0.4040245116,-0.1360109597,-0.2614206672,-0.1762212366,-0.1590020806,-0.0040113106,-0.0712005571,0.0201557726,0.2618107498,-0.1608266234,0.083612524,-0.408015281,-0.3315266073,0.0646355301,-0.3832529783,0.1397418976,-0.2481609136,-0.3263668716,-0.1347552687,0.2027281374,-0.1970701367,0.1680952162,-0.1432994604,-0.1642900109,0.3199674189,-0.1951482892,0.1908509284,-0.29429093,-0.0325954445,0.3334890306,-0.1849635094,-0.0396010093,-0.3254976571,0.1779907942,0.0813188404,0.4000708163,0.1777343601,-0.3661835194,-0.0861163661,0.003160096,0.0644557253,0.2120690942,0.127300933,-0.0967149809,0.0929268226,0.7710031867,-0.1794283837,0.1686402559,0.1889043748,-0.119961828,0.0007713275,0.0101408726,0.045477733,0.2089020461,-0.5654246211,0.1081190109,0.0579755493,0.1496831626,0.3705292046,-0.1538679153,0.1742726415,0.3599531651,0.2344636023,0.3700695038,0.1100522727,-0.2130719125,0.3592972457,0.1100753695,-0.169143185,-0.239653945,0.4028107226,0.0339758284,-0.1490621716,-0.1615044028,0.2053804845,-0.2558317482,-0.0890677944,0.3253211677,0.2903513908,0.1143730655,-0.1366235167,0.1232044548,0.1320763826,0.080273658,0.1035882309,0.1146725044,0.0707950294,0.3778044879,-0.5270092487,-0.0711613968,-0.1651747376,0.0998580083,0.01929285,-0.1437128484,0.1127315015,0.3018265069,0.3394128978,0.1030857638,-0.2475929856,-0.0392081514,-0.3277648687,-0.1317759901,-0.1296868473,0.1246639267,-0.1395261139,-0.0441211015,0.5525745749,-0.0573639795,-0.0531904362,-0.2205085903,0.1057151854,0.194195509,0.1610441506,0.4133290052,0.2486473024,-0.2230841219,-0.311917156,0.1616059989,0.2348461002,-0.2760861218,-0.0564147644,0.1692354083,0.1134895384,0.2103697211,0.1023247465,0.161009267,0.3943727612,-0.3485002518,-0.0526026264,-0.1509160697,-0.0933264717,-0.0540424474,0.3424715102,-0.0070361025,0.0169246327,0.1850292087,0.0140317613,-0.1549442112,0.2025801241,0.1005687937,0.2384246439,-0.2351112515,0.2352739125,0.2691606283,0.3267879188,0.1476660818,0.3302741349,-0.2302066684,-0.2148969173,0.013554831,0.1732248366,0.0886356384,0.0325368345,0.0527398735,-0.223371461,0.2898449302,0.1701950729,0.0218015723,-0.2609313428,0.1822646111,-0.6573223472,0.1790771484,-0.1552881151,-0.3742406964,-0.0896737128,0.2225123197,0.2393975258,0.2310740352,0.005774701,0.1412276626,-0.0417147018,0.1046397164,-0.3945060372,0.3040625155,0.2928874195,-0.1630122364,-0.1672852337,-0.4805484116,-0.0928176939,-0.0784245059,-0.0235398281,0.0088572763,-0.2540885806,0.1129637137,0.0890843421,0.3233125806,0.1530522704,0.4667845964,-0.0049052685,-0.4120467901,-0.5003908277,-0.0939524844,-0.1169576943,-0.3604584336,0.0564964712,0.414283514,-0.5631532669,-0.2819898427,-0.2900083959,0.4696748257,-0.1870671213,-0.1862505376,-0.3628036082,-0.0004369509,-0.1462663859,0.3556334972,0.2224698961,0.4271714091,0.097051613,0.2765199244,-0.4321490526,0.0308896359,0.3314535916,-0.4461791217,-0.3098087311,0.1002459154,0.2184598148,0.0352312736,-0.2208418697,-0.1718598306,0.3122579455,-0.0098214857,0.0389543734,-0.072228618,0.1307385415,0.1337008327,0.0645970255,-0.196634993,0.1101096198,0.0645981878,-0.0803332254,-0.1872682273,-0.1392814964]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/329","title":"[Bug] FileLock dependency incompatible with filesystem","comments":"Environment is Ubuntu 18.04, Python 3.7.5, nlp==0.3.0, filelock=3.0.12.\r\n\r\nThe external volume is Amazon FSx for Lustre, and it by default creates files with limited permissions. My working theory is that FileLock creates a lockfile that isn't writable, and thus there's no way to acquire it by removing the .lock file. But Python is able to create new files and write to them outside of the FileLock package.\r\n\r\nWhen I attempt to use FileLock within a Docker container by writing to `\/root\/.cache\/hello.txt`, it succeeds. So there's some permissions issue. But it's not a Docker configuration issue; I've replicated it without Docker.\r\n```bash\r\necho \"hello world\" >> hello.txt\r\nls -l\r\n\r\n-rw-rw-r-- 1 ubuntu ubuntu 10 Jun 30 19:52 hello.txt\r\n```","body":"I'm downloading a dataset successfully with\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")`\r\n\r\nBut when I attempt to cache it on an external volume, it hangs indefinitely:\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", cache_dir=\"\/fsx\") # \/fsx is an external volume mount`\r\n\r\nThe filesystem when hanging looks like this:\r\n```bash\r\n\/fsx\r\n----downloads\r\n ----94be...73.lock\r\n----wikitext\r\n ----wikitext-2-raw\r\n ----wikitext-2-raw-1.0.0.incomplete\r\n```\r\n\r\nIt appears that on this filesystem, the FileLock object is forever stuck in its \"acquire\" stage. I have verified that the issue lies specifically with the `filelock` dependency:\r\n```python\r\nopen(\"\/fsx\/hello.txt\").write(\"hello\") # succeeds\r\n\r\nfrom filelock import FileLock\r\nwith FileLock(\"\/fsx\/hello.lock\"):\r\n open(\"\/fsx\/hello.txt\").write(\"hello\") # hangs indefinitely\r\n```\r\n\r\nHas anyone else run into this issue? I'd raise it directly on the FileLock repo, but that project appears abandoned with the last update over a year ago. Or if there's a solution that would remove the FileLock dependency from the project, I would appreciate that.","comment_length":118,"text":"[Bug] FileLock dependency incompatible with filesystem\nI'm downloading a dataset successfully with\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")`\r\n\r\nBut when I attempt to cache it on an external volume, it hangs indefinitely:\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", cache_dir=\"\/fsx\") # \/fsx is an external volume mount`\r\n\r\nThe filesystem when hanging looks like this:\r\n```bash\r\n\/fsx\r\n----downloads\r\n ----94be...73.lock\r\n----wikitext\r\n ----wikitext-2-raw\r\n ----wikitext-2-raw-1.0.0.incomplete\r\n```\r\n\r\nIt appears that on this filesystem, the FileLock object is forever stuck in its \"acquire\" stage. I have verified that the issue lies specifically with the `filelock` dependency:\r\n```python\r\nopen(\"\/fsx\/hello.txt\").write(\"hello\") # succeeds\r\n\r\nfrom filelock import FileLock\r\nwith FileLock(\"\/fsx\/hello.lock\"):\r\n open(\"\/fsx\/hello.txt\").write(\"hello\") # hangs indefinitely\r\n```\r\n\r\nHas anyone else run into this issue? I'd raise it directly on the FileLock repo, but that project appears abandoned with the last update over a year ago. Or if there's a solution that would remove the FileLock dependency from the project, I would appreciate that.\nEnvironment is Ubuntu 18.04, Python 3.7.5, nlp==0.3.0, filelock=3.0.12.\r\n\r\nThe external volume is Amazon FSx for Lustre, and it by default creates files with limited permissions. My working theory is that FileLock creates a lockfile that isn't writable, and thus there's no way to acquire it by removing the .lock file. But Python is able to create new files and write to them outside of the FileLock package.\r\n\r\nWhen I attempt to use FileLock within a Docker container by writing to `\/root\/.cache\/hello.txt`, it succeeds. So there's some permissions issue. But it's not a Docker configuration issue; I've replicated it without Docker.\r\n```bash\r\necho \"hello world\" >> hello.txt\r\nls -l\r\n\r\n-rw-rw-r-- 1 ubuntu ubuntu 10 Jun 30 19:52 hello.txt\r\n```","embeddings":[0.0329926312,0.0771608278,-0.0250304174,0.027964592,0.0889836028,0.1513945013,0.364697963,0.0651528686,0.7094057202,-0.0978959948,0.090967983,-0.0051194588,0.0214458238,-0.3159707189,-0.2095369399,0.2686368525,0.1863837689,0.0667651743,0.005568597,0.0047016451,-0.0693236962,0.3521534801,-0.1633292437,-0.2751578987,-0.7754433155,0.120644629,0.0365542248,0.0431686118,-0.2784680426,0.0024243908,0.4160512686,0.5284094214,-0.0816129148,0.496588856,-0.0001226855,-0.1948298067,0.2004580349,-0.0361196883,-0.1746515036,-0.2741996944,-0.2164897621,-0.3265877962,0.1097059995,-0.1128876805,0.1944538206,-0.2241576165,0.0919875652,-0.4995852709,0.3337172568,-0.0410731472,0.0680703148,0.2360705435,0.254280597,-0.0269481838,0.3872797191,-0.2296983749,-0.2215995938,0.5681561828,0.456705898,-0.2117682695,-0.140966773,-0.036995776,-0.0129712755,-0.0679331571,0.5948626995,-0.1976968348,0.3488734663,-0.2678504586,-0.1012830809,0.3191520572,0.6145647764,-0.0441019721,-0.5926792026,-0.2471349835,0.327287972,0.1455295384,0.5233204365,0.0666401461,-0.0742337778,0.3606702387,0.0522742458,-0.1590826958,0.0457444265,0.0224095583,-0.1197813973,0.2216324359,0.1453604996,0.0084855128,0.1708891541,-0.0806063488,0.3088315129,-0.1967713535,0.0066907033,0.0987603217,-0.454965502,0.1360242963,0.0213125143,0.1014321446,-0.140349254,-0.2220833302,-0.4776568115,0.1874590665,0.0509255603,0.3599926233,0.1120271906,0.1694357842,0.0524430089,0.0851673633,0.2234933525,-0.0380137265,-0.2711602747,0.0675251037,-0.0540335141,-0.1094276682,0.1089459434,-0.0756204799,0.026497189,-0.3769022226,0.171045959,0.1699299812,0.0443213135,0.3801297247,0.1577482671,0.0316457227,0.1108215377,0.4331046641,0.1974970698,0.0049862051,-0.1965510547,0.3437236845,-0.0323614851,-0.098464556,0.0690493658,0.0930997208,0.0324408524,0.2763173282,0.1785526872,0.156566456,0.3272791803,0.1014767811,0.2352108508,-0.1581271142,-0.127683118,0.393193692,-0.1377974898,0.0754465386,0.0243199468,-0.0310192034,-0.0276711918,0.0367911831,-0.1925075799,0.0691401735,-0.0996998847,0.0263191275,-0.3808284998,0.1100183353,0.1620588303,-0.2464067787,-0.1027880833,-0.4254748821,0.0668773428,-0.4721822739,-0.0609576814,-0.0015061117,-0.0517792217,0.3292973042,-0.2778441012,-0.0546500087,0.2590722144,-0.0380633622,-0.1223073527,0.2285180241,-0.0811449364,0.1059946492,-0.2078044564,0.0991857424,-0.0261375289,-0.4818366766,-0.7829118371,0.2422858626,-0.2895569205,0.2557770312,0.1629284471,0.3792766035,0.1021165773,-0.2274315655,0.0992249176,0.3591865301,0.1214847341,-0.217090264,-0.0638454482,-0.2845760286,-0.2017313987,0.0688606873,0.0111601846,0.2223489285,0.0180596188,0.6689196825,0.2116287053,0.0055397418,0.0050580725,0.5441519618,0.6210247278,0.0836993009,0.0081204129,-0.1322745979,-0.6575937271,0.5051038265,-0.4428118467,-0.0797979981,-0.2008097768,-0.0727460012,0.1174281761,-0.0122002792,-0.0691100806,0.1472261846,0.0574351884,-0.0790043771,0.0755307823,0.1750714034,0.0899828151,0.3439259827,0.0125084687,-0.0200022738,-0.0590691678,-0.0895981789,0.1630969197,-0.1867703646,0.053861659,-0.0638033301,0.1063678265,-0.2141563445,-0.1605974883,0.2369826883,0.3473391831,0.2953688502,-0.2741957605,0.3801431656,0.0418263227,0.4804558158,-0.1965429932,-0.0898995325,0.1089190617,-0.2300198227,0.0361249559,0.0947789326,-0.2191524804,0.0928045437,-0.2077879012,0.0561608113,0.0911774412,0.0453527644,-0.1688566357,-0.0269355029,0.2806842327,-0.0091582621,0.2706624269,-0.10584113,0.6271888614,0.0670904517,0.366843611,0.1496047974,0.0516429357,0.0494002216,0.0035033473,-0.1143359765,0.0166757964,0.3775188327,0.498829186,0.1017267331,0.288933903,-0.2164707333,0.2117234468,-0.2904098332,0.0543171652,-0.1254851967,-0.2110745907,0.3450728357,0.134191215,-0.1068298295,0.1315164119,-0.2868654728,0.1256860495,0.0205750726,0.1268946826,0.1132571846,-0.3515273929,0.5667666793,0.2086833566,0.1933236867,-0.2534320056,-0.4644887447,0.1059635729,0.7339765429,0.1522963494,0.0692449436,-0.4219465256,-0.0498631708,0.0390344374,-0.3337166607,-0.0399523787,-0.3362483978,-0.0026227303,-0.1162400544,0.553121686,0.0384058096,0.1448250413,0.1003429815,-0.3246200979,-0.3931279778,-0.18387191,0.1499125063,-0.0926098526,0.4356099367,0.0584887452,0.1634122133,-0.1018638685,0.0385645032,-0.0192833506,-0.3886204362,-0.0294425618,-0.0975789279,-0.0303108245,-0.1097448319,-0.2369118631,0.171840176,-0.1660530269,-0.3670009971,0.3042767346,-0.3411258161,-0.2106982768,0.2431464344,0.0185857695,0.287735641,-0.0033969062,-0.0847150981,0.0122465109,-0.459141165,0.2137269825,0.1240149066,-0.4351314008,-0.1326473653,0.1863021255,-0.4067225158,0.2642199993,-0.4266271293,-0.2256042361,-0.32558617,0.2897674143,-0.0849460885,-0.0755462423,0.4451619387,-0.2783597112,0.0496294312,-0.4187117219,0.3278357983,0.1877295375,0.0333668292,-0.0206506439,0.0646261796,-0.20487158,0.1377081126,0.513119638,-0.0706710443,0.0123607032,0.1415621191,0.3056558967,0.438752979,-0.3285657167,-0.1682540476,-0.1418815255,-0.0881157145,-0.1464124471,-0.028260814,-0.1522278786,-0.063107796,-0.1003580615,-0.0561155565,-0.0604511984,-0.300537914,0.0356877632,0.1938851476,0.1703025848,0.2378978282,-0.0101714488,-0.0816976428,-0.1531502604,0.125331074,0.2319324315,0.4605024457,0.0418451838,-0.0895562395,0.2507178187,-0.1388715804,0.1664326191,0.0581349358,0.3817953467,-0.0692597926,-0.0027474093,-0.0256756973,-0.3396189809,0.314963758,0.0228527877,0.1409370303,0.2082722038,0.1402716488,-0.2288290858,-0.1793694496,-0.1924062222,0.2797608972,0.4190057516,0.1958255768,-0.1657379866,-0.165255487,-0.6652338505,0.3763881028,-0.0641951412,-0.5701953769,-0.120379366,0.2926162779,-0.2946732342,-0.00751406,-0.0453766584,0.0121914307,0.0891532078,0.1238368377,-0.252624929,-0.1297720373,-0.1409834623,-0.1335089952,0.0248142947,-0.1612214297,0.2471484095,0.0566672869,-0.26813972,0.1358754486,0.1964807659,0.194649443,0.1310192198,0.3067689836,-0.2398160696,-0.2409114987,0.472989589,-0.0340853781,0.1218698174,0.0318526551,-0.0542271994,0.0342299789,0.0109433308,0.1316773593,-0.0267074984,0.4426619411,0.068115674,0.0203917194,0.1507995576,-0.1900191158,0.4326274693,0.3856013715,-0.0337894782,0.0248092189,-0.0407207124,0.8245965838,0.0921501741,0.1442941874,-0.0362538956,-0.0155593576,0.2752341032,-0.4026061893,0.0966877043,-0.1102206483,-0.4013601542,-0.0234980155,-0.2942135036,0.3266855478,-0.183565259,0.0453545712,0.1467427909,-0.1992115825,-0.1096248776,-0.0855044872,0.3566527367,-0.4257876277,-0.3449666798,-0.1093275473,0.1533336043,0.2002201974,0.1391858011,-0.0824589953,-0.1181075126,-0.2436950654,-0.0548408255,-0.2625560164,0.0846152157,-0.0476019569,0.4155933261,-0.3947254717,0.1000957116,-0.0801592395,0.044527214,0.0227881782,0.3616425991,-0.3698623478,0.1039457694,-0.0302532353,-0.1415355206,0.1491737962,-0.1640695632,0.2833949327,-0.0795299634,-0.1332876682,-0.2885744572,-0.1448111832,-0.2349168807,0.2240302861,-0.089076817,0.1189407855,0.2076825052,-0.8349124789,0.0561655834,-0.1746170968,-0.0080148391,0.0594589971,0.1896305531,-0.1696871668,0.1729542613,-0.3796271682,-0.1412031502,-0.0095264632,0.6327989101,-0.1575657427,-0.3251600862,0.4551859796,0.1251761168,-0.0939947516,-0.1562014371,0.359762311,-0.1781922281,-0.2810073197,-0.1570887715,-0.1280915886,0.021751048,-0.1071635634,-0.0376140587,0.2150369436,-0.1415710747,0.0009853785,-0.4553619921,-0.2899051309,-0.0132948793,-0.3651156723,0.07086806,-0.2607094646,-0.354385823,-0.2182277292,0.2064197361,-0.2022307962,0.1624437869,-0.2028410286,-0.1324501932,0.3135014176,-0.1432558298,0.2176746279,-0.3177880347,-0.0388179906,0.4003426731,-0.1684274673,-0.0534727722,-0.3270488381,0.1815370023,0.1397498548,0.3898147345,0.1737300009,-0.3615197241,-0.0878404528,0.0177420843,0.0914114267,0.2003791779,0.1327970624,-0.1036611348,-0.007462617,0.8132432103,-0.1819959134,0.1114888489,0.079298459,-0.1420636177,0.0595960468,0.0092012351,0.098899655,0.2614958584,-0.5912240148,0.0871781558,0.0094878497,0.1891735196,0.3638569117,-0.1222799569,0.0873582512,0.306261152,0.2429109216,0.3095578551,0.1071130186,-0.1949279606,0.3150210679,0.0827363059,-0.1386357695,-0.1861426979,0.3364927769,0.00009193,-0.0383739993,-0.1212081462,0.2178881168,-0.2060879022,-0.128636539,0.3628096581,0.2267933041,0.0419038311,-0.1312205046,-0.0241871979,0.1435373127,0.1016445756,0.1517173201,0.0697907433,0.0558384024,0.3068059683,-0.3838959038,-0.0301961843,-0.1403522044,0.1081604064,0.0049633728,-0.1297217906,0.122539483,0.1950872093,0.4629550874,0.1371627003,-0.2884195447,-0.0618603714,-0.3252459764,-0.0417996831,-0.0186615046,0.1944165826,-0.1048465371,-0.1307231635,0.543027401,-0.0533543304,-0.0094788186,-0.2325727344,0.1250622571,0.1852504015,-0.0379816741,0.3821392953,0.2015940845,-0.1882234067,-0.3295243084,0.0993864462,0.2410270423,-0.2834415734,-0.1395076364,0.1950459927,0.1730169505,0.1247149557,0.0686639473,0.210226953,0.3694635034,-0.3128955662,0.0123259686,-0.1889050156,-0.0972298086,-0.0744773,0.3880660236,-0.0245932136,-0.049476359,0.1560754329,0.018243311,-0.2037349194,0.163720578,0.1346607059,0.2287757248,-0.1798717827,0.3555248082,0.3213663101,0.2588977516,0.114177309,0.3011619747,-0.2230374813,-0.1569641232,0.0158119947,0.1607815474,0.0891635269,-0.0217632689,0.0613140389,-0.2351773679,0.2948059738,0.2314912379,-0.0113277109,-0.2304638773,0.338453114,-0.639773488,0.1641094238,-0.2192554474,-0.4225708842,-0.0913716033,0.2267868966,0.2722913027,0.2294969112,-0.0486783423,0.0103650196,-0.0766913146,0.0255014021,-0.2985700369,0.266882807,0.2643013895,-0.1601854712,-0.1551843882,-0.4845214784,-0.0849695951,-0.1766770929,0.0021155213,0.0124988565,-0.298918426,0.1190501824,0.1107706949,0.3183737099,0.20132792,0.3909295499,-0.0760905519,-0.4142149091,-0.5229349136,-0.0357376672,-0.1828191429,-0.3264659345,0.024637457,0.4542936385,-0.6023714542,-0.3832444251,-0.3101326227,0.4385449588,-0.264303416,-0.2452713251,-0.328189075,0.1080601141,-0.2162434757,0.3898150325,0.1635169238,0.3700908422,0.1109334379,0.2532015443,-0.4738084078,0.0268698987,0.3253875375,-0.4806062281,-0.2813548148,0.1587108672,0.1845037043,0.0396693423,-0.2161522359,-0.1581626981,0.3095881045,0.0067396681,-0.02763067,-0.0821889564,0.0420230962,0.055139415,0.0974704996,-0.1874362677,0.0811233744,0.0602797717,-0.1005465165,-0.2201518565,-0.0833189413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/329","title":"[Bug] FileLock dependency incompatible with filesystem","comments":"Looks like the `flock` syscall does not work on Lustre filesystems by default: https:\/\/github.com\/benediktschmitt\/py-filelock\/issues\/67.\r\n\r\nI added the `-o flock` option when mounting the filesystem, as [described here](https:\/\/docs.aws.amazon.com\/fsx\/latest\/LustreGuide\/getting-started-step2.html), which fixed the issue.","body":"I'm downloading a dataset successfully with\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")`\r\n\r\nBut when I attempt to cache it on an external volume, it hangs indefinitely:\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", cache_dir=\"\/fsx\") # \/fsx is an external volume mount`\r\n\r\nThe filesystem when hanging looks like this:\r\n```bash\r\n\/fsx\r\n----downloads\r\n ----94be...73.lock\r\n----wikitext\r\n ----wikitext-2-raw\r\n ----wikitext-2-raw-1.0.0.incomplete\r\n```\r\n\r\nIt appears that on this filesystem, the FileLock object is forever stuck in its \"acquire\" stage. I have verified that the issue lies specifically with the `filelock` dependency:\r\n```python\r\nopen(\"\/fsx\/hello.txt\").write(\"hello\") # succeeds\r\n\r\nfrom filelock import FileLock\r\nwith FileLock(\"\/fsx\/hello.lock\"):\r\n open(\"\/fsx\/hello.txt\").write(\"hello\") # hangs indefinitely\r\n```\r\n\r\nHas anyone else run into this issue? I'd raise it directly on the FileLock repo, but that project appears abandoned with the last update over a year ago. Or if there's a solution that would remove the FileLock dependency from the project, I would appreciate that.","comment_length":31,"text":"[Bug] FileLock dependency incompatible with filesystem\nI'm downloading a dataset successfully with\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")`\r\n\r\nBut when I attempt to cache it on an external volume, it hangs indefinitely:\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", cache_dir=\"\/fsx\") # \/fsx is an external volume mount`\r\n\r\nThe filesystem when hanging looks like this:\r\n```bash\r\n\/fsx\r\n----downloads\r\n ----94be...73.lock\r\n----wikitext\r\n ----wikitext-2-raw\r\n ----wikitext-2-raw-1.0.0.incomplete\r\n```\r\n\r\nIt appears that on this filesystem, the FileLock object is forever stuck in its \"acquire\" stage. I have verified that the issue lies specifically with the `filelock` dependency:\r\n```python\r\nopen(\"\/fsx\/hello.txt\").write(\"hello\") # succeeds\r\n\r\nfrom filelock import FileLock\r\nwith FileLock(\"\/fsx\/hello.lock\"):\r\n open(\"\/fsx\/hello.txt\").write(\"hello\") # hangs indefinitely\r\n```\r\n\r\nHas anyone else run into this issue? I'd raise it directly on the FileLock repo, but that project appears abandoned with the last update over a year ago. Or if there's a solution that would remove the FileLock dependency from the project, I would appreciate that.\nLooks like the `flock` syscall does not work on Lustre filesystems by default: https:\/\/github.com\/benediktschmitt\/py-filelock\/issues\/67.\r\n\r\nI added the `-o flock` option when mounting the filesystem, as [described here](https:\/\/docs.aws.amazon.com\/fsx\/latest\/LustreGuide\/getting-started-step2.html), which fixed the issue.","embeddings":[-0.0220689438,-0.0703312829,-0.0241179243,0.0477825627,0.0648071468,0.0538179018,0.2719034553,-0.0904818401,0.7102741599,-0.0634486228,-0.0230357796,-0.0301015638,0.1403870285,-0.4040715694,-0.283573091,0.2419641763,0.2077274472,0.0305831209,-0.2452179492,-0.0770897567,-0.1613958776,0.4342415631,-0.0815225467,-0.2347530723,-0.5598556995,0.1654607058,-0.0396309122,0.1403411925,-0.3457119465,0.18981947,0.3417885005,0.6025447249,-0.0452574156,0.5372428298,-0.0001287701,-0.2627755105,0.2452052087,-0.0852422863,-0.2170967311,-0.1017389745,-0.0870472118,-0.1840791106,0.2249264866,-0.1094866619,0.160850212,0.1228189543,-0.0150860017,-0.5285058022,0.265088439,-0.1673846096,-0.0002370196,0.1131123155,0.167238161,-0.1510144174,0.3145914972,-0.2097328752,-0.1848717779,0.4976549149,0.4717492759,-0.0901188776,-0.2479738295,0.0887352005,0.025788812,0.0382361636,0.6046129465,-0.2951402366,0.4669860005,-0.2977612913,-0.3035379052,0.2270716578,0.5503914952,-0.0176632497,-0.5384922028,-0.1820754558,0.358096391,-0.0000865813,0.4926611483,0.0541597866,-0.1355169564,0.3503001034,0.015288325,-0.1867275983,0.0734807923,0.0515650138,-0.1668386459,0.0551695526,0.1540466845,0.0693725049,0.2360328436,-0.141763553,0.0047063935,-0.2013449967,-0.1389338672,0.0655190945,-0.5184341669,0.0440693386,-0.0237710048,0.0005390666,0.0014865845,-0.1775720417,-0.4711024463,0.1959103942,-0.1072873995,0.4131366014,0.1072750688,0.1250377297,0.1560991257,-0.2706227601,0.3129884303,-0.0366698503,-0.2005043477,0.1234974861,-0.0997190326,-0.2006221265,-0.0706194639,-0.0085441861,0.0244507249,-0.4772162735,0.1940928698,0.2955318987,0.1394579113,0.4006765485,0.1581330299,0.1941503435,0.1847222,0.4262399673,0.2075468004,0.0793263689,-0.1944486201,0.1226030812,0.0305183455,-0.1448899508,0.1050764844,0.1360038221,0.0371015035,0.2895999551,0.1767875552,0.1284878403,0.2166540474,0.1174142882,-0.0327689983,-0.0519724824,-0.1933226138,0.4518730342,-0.141878292,0.1657255739,0.0007304286,0.1726854295,-0.0379014574,0.0773888379,-0.3118938506,-0.0237897579,-0.1736692488,-0.0390766449,-0.4359885156,0.0211994778,-0.1416533291,-0.2198771387,-0.2442763001,-0.3579334319,0.0514369868,-0.504481256,-0.0349159278,0.0546138324,-0.0038452642,0.3324382603,-0.4034097791,-0.0246994253,0.2663726509,-0.1798488051,0.0174748898,0.3031401038,-0.0489033759,-0.0927232653,-0.1638783067,-0.009584276,0.1985514015,-0.4244553745,-0.6493101716,0.1907721907,-0.334951371,0.1472037286,0.1931843311,0.2882729173,-0.0813930184,-0.1296011358,0.0087062344,0.3986360133,0.1607896686,-0.198757872,-0.1371515691,-0.2944769859,-0.1860226989,0.1477249265,-0.0508104973,0.152619198,0.0321165994,0.5327267051,0.2337102592,-0.0062223668,-0.0406285077,0.6131659746,0.677834034,0.1461275667,0.1264191866,-0.1563685983,-0.5479946136,0.385871917,-0.3822480738,0.0200355537,-0.1693601161,-0.0390051603,0.1234154478,-0.0224512219,-0.0536840633,0.161467284,-0.018963689,-0.1154923663,0.0692482442,0.0914661661,0.1012218744,0.4054683447,-0.1135651767,0.0921659842,-0.1364718378,-0.0361966118,0.1138330176,-0.2384926677,0.0204266049,-0.0355385467,0.0971088484,-0.2887970805,-0.2271137089,0.4242851436,0.2979375124,0.1792170554,-0.0711193234,0.3551245034,0.0821090788,0.5420109034,-0.2560358644,-0.1445958018,0.0648711547,-0.1487627327,-0.1179759428,0.0460430309,-0.1618244499,0.1645622849,-0.2425416857,-0.0333045833,0.1862594187,0.1448422223,-0.0942100361,0.1040548384,0.3544766605,0.0335772596,-0.103646636,-0.1695373505,0.4346260428,0.0931999087,0.2976520658,0.0829860792,0.0781313181,0.107116282,-0.0289707314,-0.091502741,0.1539365351,0.2255529016,0.4854830503,0.0972242504,0.1201269925,-0.3374759555,0.1207467094,-0.2458670586,0.0951208025,-0.0483834073,-0.2100300491,0.2208928168,0.2606562376,-0.1177728176,0.0648326278,-0.4731079936,0.1089856178,0.0579156019,0.0545388721,-0.0150621124,-0.2951617241,0.4948147535,0.1755051762,0.3769619763,-0.2835226357,-0.4526466131,0.1798097193,0.8311291337,0.1327476948,0.0647038743,-0.4011010528,0.100080736,0.1955150813,-0.3513923585,-0.2648263276,-0.2396947294,0.0074710893,-0.1677084416,0.4072366357,0.1740886271,0.1044062749,0.1826917827,-0.1740866899,-0.4815850854,-0.1713532358,0.2042211145,-0.0078629497,0.4539900124,0.0329961255,0.1318009645,-0.1141724586,-0.0199321788,0.1900190413,-0.3241443634,-0.035949707,-0.0779606476,-0.0392701887,-0.1526338756,-0.280564189,0.2152979225,-0.1210353002,-0.3584309816,0.2134422213,-0.2896826267,-0.2595065534,0.0396323167,0.076623328,0.2536055148,-0.1284696013,-0.1014168411,0.0218411945,-0.4564404488,0.2862502635,0.0493835211,-0.5032615066,-0.0686936527,0.290350765,-0.4942887127,0.1319266856,-0.4954322875,-0.317903012,-0.2460888028,0.1960786581,0.035322126,-0.0818380564,0.4039615095,-0.2405238897,0.0513252616,-0.3743172586,0.2550759614,0.158996731,0.1650806814,0.0968525857,-0.0252178665,-0.1344376951,-0.0114859687,0.6712635159,-0.162784934,-0.1171938926,0.1715481579,0.2996037602,0.2783558369,-0.3603678346,-0.1101806983,-0.0983461738,-0.0916987881,0.009312422,0.0218713433,-0.0106868353,-0.1097532585,-0.1348102093,-0.1368597746,0.0612050034,-0.3520058393,0.0050268113,0.266382128,-0.1018702239,0.2015020102,-0.1082897261,-0.0589301623,-0.1559250653,0.2489109635,0.401643157,0.4164842069,0.0854802057,-0.0278370306,0.0806599334,-0.1156342626,0.2246972919,-0.076371558,0.3127216697,-0.0328803547,0.0389622711,0.01067275,-0.3546395004,0.4242370725,-0.0576989204,0.1282720268,0.1406544,0.0565012172,-0.2863332331,-0.1601760834,-0.2828541398,-0.0436638445,0.3871494532,0.2369051576,-0.1635142416,-0.2345156223,-0.4966443181,0.3092567921,-0.1080526114,-0.5445774198,-0.0965401158,0.3303972483,-0.3534878492,0.0477932543,-0.1198704913,0.156253919,-0.0131975971,0.1064740494,-0.2379442304,-0.0273867734,-0.0754133686,-0.0080913436,0.0163863804,-0.0439742841,0.2349879146,0.0385372378,-0.0726483315,0.1333613396,0.2118765563,0.4685258865,0.2167075723,0.4319480658,-0.2559484243,-0.1288783699,0.470571965,-0.0090790624,-0.042952165,-0.1804578304,0.081756793,0.0764367655,0.1618201584,0.1221978366,0.0046963268,0.3768506348,0.1502006203,0.1201419532,0.0793380961,-0.2263431698,0.5570188165,0.4297812879,0.1192735881,0.1248109341,-0.0013749221,1.009852767,0.1102246046,0.0891092792,0.1218445674,0.033814583,0.2011273801,-0.6567310691,0.0791682601,0.0160584971,-0.2834217846,-0.0612606257,-0.2211582363,0.3837388754,-0.2212117463,0.1036453992,0.2203228772,0.1064143479,-0.1884318888,-0.1864147633,0.3203713298,-0.4331746399,-0.4605294466,0.0465037823,0.0883577019,0.2271878123,0.2807311118,-0.0668821409,-0.1008887962,-0.1867897213,-0.1367330402,-0.2401668429,0.0603207536,-0.057121899,0.2376734465,-0.6428996921,0.0503373891,-0.1325326413,0.1312490553,-0.0716646537,0.2758539319,-0.3406359255,0.1956918538,0.0455400869,-0.0971593931,0.1098515764,-0.0986965671,0.2272893041,-0.1277881116,-0.0428792872,-0.267903626,-0.1020253748,-0.1175598875,0.1526528299,-0.1130960286,0.0399345234,0.1098317429,-0.751541853,0.1185879484,-0.1799274087,0.0076037003,0.0078583965,0.2270798981,-0.211128518,0.0875641555,-0.4490461051,-0.1456823051,-0.096945554,0.6907750368,-0.1882463545,-0.4291864038,0.3878200054,0.1820187271,-0.0248332005,-0.1662536711,0.2785246372,0.1498802602,-0.5977125764,-0.1147994995,-0.0817321464,-0.1916285008,0.0737218857,0.1710986793,0.1712403148,-0.2321943343,-0.0244575106,-0.4262325466,-0.1898557842,0.0371150412,-0.3765431643,0.1015165448,-0.1818583459,-0.453702569,-0.1810139865,0.0585197434,-0.1408689767,0.2131169289,-0.1011467129,-0.0680404231,0.2884111106,-0.0186859928,0.2443310618,-0.1946051121,-0.0744401664,0.3325659335,-0.1290651411,0.0094961151,-0.300260514,0.1881702095,0.1192868873,0.4957099855,0.1154989228,-0.3775988519,-0.0708090588,-0.0054678251,0.0890505984,0.1156487837,0.0807780474,-0.1039889753,0.3058717251,0.6972977519,-0.0860889852,0.1983677,0.2485138327,-0.0549638607,-0.1351300329,-0.101970166,-0.0427587032,0.1243599802,-0.5576992631,0.0499534011,0.0696817115,0.3294982612,0.3463843167,-0.0850841329,0.0267595239,0.2640788853,0.437523365,0.3466055691,0.1543820947,-0.1451997459,0.3441730738,0.0757539421,-0.1017270461,-0.2094870061,0.4380947649,0.0302215703,-0.0104039125,-0.007636467,0.1041410938,-0.1241200566,0.1275944263,0.3202594519,0.4867460132,-0.0664863363,-0.169422105,0.1280142516,0.1750625074,0.0314672738,0.2002261281,0.100430049,0.1505336314,0.4474329948,-0.3802274764,0.0218057651,0.2399681956,0.0390008017,0.0401868485,-0.0057542589,-0.0691166446,0.1953559071,0.5384966135,0.152158916,-0.2523939312,0.0875382349,-0.3108756542,-0.2210139632,-0.1560127884,0.2961390018,-0.075629577,-0.0382717103,0.3463694453,-0.1114040613,0.0399672464,-0.1754452139,0.0525841936,0.1362793893,0.1210569367,0.3864115179,0.3674781024,-0.3956397176,-0.3698976338,0.0938782468,0.2836430669,-0.2083596587,-0.0337962098,0.2968219221,0.0612662174,0.340702951,0.0579257533,0.2708117068,0.2780500054,-0.2221732438,0.0769170672,-0.1248655841,-0.0818737596,-0.0950929224,0.3482890129,0.0078955106,-0.0478937328,0.22525464,0.0159999393,-0.1737379879,0.1647745818,0.0418886282,0.2347173244,-0.1700987071,0.4057605863,0.2302639037,0.3443794549,0.1322940439,0.1089438573,-0.2730688453,-0.0650887787,-0.093683131,0.087460734,0.0296581946,-0.036687579,0.031654764,-0.2727530003,0.227741465,0.2071667314,-0.0866365433,-0.2552128434,0.3333531916,-0.6523105502,0.1505028605,-0.1023142561,-0.4395861924,-0.222251907,0.3195895851,0.2867953777,0.2792307734,-0.0373553298,-0.1852324158,-0.1396140456,0.2462437302,-0.324627161,0.3226268291,0.1380733103,-0.0423329175,-0.1567665339,-0.4385940433,-0.0080303205,-0.1300816089,-0.0691764876,0.1633478552,-0.3762868941,0.1774116457,0.0282086246,0.255361706,0.1002020985,0.3817992508,0.0496052131,-0.3303436339,-0.5639330149,-0.2779224515,-0.0836468339,-0.3346694708,-0.0548274554,0.4474784136,-0.5062488914,-0.2603892386,-0.3515311182,0.2791649103,-0.2844417095,-0.1108864024,-0.3698858619,0.1871896535,-0.2657650709,0.405244261,0.1279306114,0.3573997915,0.1689555198,0.3010959923,-0.4526656866,-0.0991738588,0.4318039417,-0.4769830108,-0.2671391666,0.1060707048,0.2468598187,0.2311085016,-0.1564673185,-0.212507531,0.2675844431,-0.0525471605,0.1363869905,-0.0372591503,0.0257630609,0.1732923388,-0.0254708473,-0.2425495833,-0.0126669109,0.0613084696,-0.0972380936,-0.1867421716,-0.1274042875]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/329","title":"[Bug] FileLock dependency incompatible with filesystem","comments":"Awesome, thanks a lot for sharing your fix!","body":"I'm downloading a dataset successfully with\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")`\r\n\r\nBut when I attempt to cache it on an external volume, it hangs indefinitely:\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", cache_dir=\"\/fsx\") # \/fsx is an external volume mount`\r\n\r\nThe filesystem when hanging looks like this:\r\n```bash\r\n\/fsx\r\n----downloads\r\n ----94be...73.lock\r\n----wikitext\r\n ----wikitext-2-raw\r\n ----wikitext-2-raw-1.0.0.incomplete\r\n```\r\n\r\nIt appears that on this filesystem, the FileLock object is forever stuck in its \"acquire\" stage. I have verified that the issue lies specifically with the `filelock` dependency:\r\n```python\r\nopen(\"\/fsx\/hello.txt\").write(\"hello\") # succeeds\r\n\r\nfrom filelock import FileLock\r\nwith FileLock(\"\/fsx\/hello.lock\"):\r\n open(\"\/fsx\/hello.txt\").write(\"hello\") # hangs indefinitely\r\n```\r\n\r\nHas anyone else run into this issue? I'd raise it directly on the FileLock repo, but that project appears abandoned with the last update over a year ago. Or if there's a solution that would remove the FileLock dependency from the project, I would appreciate that.","comment_length":8,"text":"[Bug] FileLock dependency incompatible with filesystem\nI'm downloading a dataset successfully with\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\")`\r\n\r\nBut when I attempt to cache it on an external volume, it hangs indefinitely:\r\n`load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", cache_dir=\"\/fsx\") # \/fsx is an external volume mount`\r\n\r\nThe filesystem when hanging looks like this:\r\n```bash\r\n\/fsx\r\n----downloads\r\n ----94be...73.lock\r\n----wikitext\r\n ----wikitext-2-raw\r\n ----wikitext-2-raw-1.0.0.incomplete\r\n```\r\n\r\nIt appears that on this filesystem, the FileLock object is forever stuck in its \"acquire\" stage. I have verified that the issue lies specifically with the `filelock` dependency:\r\n```python\r\nopen(\"\/fsx\/hello.txt\").write(\"hello\") # succeeds\r\n\r\nfrom filelock import FileLock\r\nwith FileLock(\"\/fsx\/hello.lock\"):\r\n open(\"\/fsx\/hello.txt\").write(\"hello\") # hangs indefinitely\r\n```\r\n\r\nHas anyone else run into this issue? I'd raise it directly on the FileLock repo, but that project appears abandoned with the last update over a year ago. Or if there's a solution that would remove the FileLock dependency from the project, I would appreciate that.\nAwesome, thanks a lot for sharing your fix!","embeddings":[-0.0076734023,0.0097793378,-0.0231539961,0.0512804948,0.1272634864,0.1981419325,0.3464359641,0.0612710454,0.7116608024,-0.0512954108,0.0536316298,-0.0363602452,-0.0010532561,-0.2431037128,-0.1808877885,0.2046639174,0.2025125772,0.0913902819,-0.1005518287,0.0103591597,-0.1431073993,0.3850389421,-0.235606879,-0.3015982211,-0.7572391033,0.167845428,-0.0751160681,0.0293418542,-0.2802717388,-0.0570871569,0.4525908232,0.5540817976,-0.0797485858,0.5339693427,-0.0001240431,-0.2386542559,0.2603720427,-0.0317375921,-0.2373848408,-0.2076942325,-0.3478657305,-0.3055988252,0.0552087612,-0.0624305345,0.2051943839,-0.1450191289,0.0732339919,-0.536667645,0.3131193221,-0.1347166896,0.0592134893,0.1399396509,0.2160496563,-0.1378368288,0.3231158555,-0.1427930444,-0.173737675,0.5222958326,0.3845928013,-0.1842415631,-0.0626856983,-0.0096265255,-0.0007444928,-0.1118206754,0.5916334987,-0.2255762368,0.3271414638,-0.2409713566,-0.0664751381,0.2954209745,0.7672129273,-0.0502438359,-0.5510516167,-0.2003196627,0.3568514585,0.0395565033,0.5681681037,0.0117432075,-0.0880448371,0.414393723,-0.040440619,-0.1379661262,0.0753381401,0.0062384773,-0.2048979998,0.0812284872,0.1882241815,0.0189136658,0.2504156828,-0.0676767305,0.2466800362,-0.2071983069,-0.0507258214,0.1283876598,-0.4815732539,0.1213441566,-0.0102177616,0.1880559623,-0.061717838,-0.2394709885,-0.55559659,0.1632439941,0.030707432,0.3138603866,0.2670823932,0.1148299202,0.0810744315,-0.0198566169,0.2377533019,-0.0368026681,-0.2533181012,0.0868660361,-0.0831217915,-0.1697322875,0.1287813336,-0.04589203,0.0608145967,-0.4425078034,0.1447642893,0.2048992962,0.0932758898,0.3311029077,0.1311912835,0.0613971874,0.0399512984,0.4603050649,0.1423786432,0.0216152947,-0.2110380828,0.2906770408,-0.0365277342,-0.1608535051,0.0162770655,0.1051706448,0.0949816555,0.2238990515,0.136681661,0.2061617374,0.3350233436,0.1395793706,0.1185082346,-0.1534569412,-0.1825641245,0.3688208163,-0.1610945612,0.1281691343,0.0126123065,0.0723481029,-0.0664691776,0.0675922409,-0.2014655918,0.0713401809,-0.0394129641,0.0222837124,-0.4207431376,0.0400810726,0.1353872716,-0.1953596622,-0.1580325663,-0.3746499419,-0.0314002819,-0.4413694739,-0.050514169,0.0022448276,-0.0004920681,0.3877295256,-0.3087955117,-0.0047300621,0.2046559453,-0.026928639,-0.1337799132,0.2412903607,-0.0946394429,0.0631173253,-0.221809715,0.0906384289,-0.0157570951,-0.4783608317,-0.8166896105,0.2185086906,-0.2551716566,0.2880479395,0.0520536229,0.3806164861,0.114413403,-0.2074390799,0.1083046347,0.4441544116,0.0816690102,-0.1895727515,-0.0730709881,-0.3189725578,-0.1269501448,0.1493979245,0.051158417,0.2346167713,-0.0254798681,0.6282660365,0.2463653982,0.0461098067,-0.0216615256,0.537116766,0.6547981501,0.1269494593,0.0144094909,-0.1228342727,-0.658965528,0.457562834,-0.3105434179,-0.1017337367,-0.1310077161,-0.1303044111,0.0241684113,-0.010069727,-0.1131296754,0.2328106165,0.0295095984,-0.0806661248,0.090608947,0.2175965756,0.0895942524,0.4313810766,0.1279778183,-0.0303446557,-0.0406932682,-0.0467075333,0.1075434089,-0.152465716,0.0942100361,-0.093396984,0.1498004943,-0.2161425799,-0.2101459205,0.2246413678,0.4520836174,0.2556502223,-0.2244581431,0.2682109773,0.0570094846,0.3097125292,-0.1456186324,-0.0822074041,0.0968153328,-0.2709512711,0.049606245,0.0746188685,-0.2838700116,0.0358654484,-0.252672106,0.094568558,0.1213984191,0.0336439945,-0.1478250474,-0.0165757574,0.2747130692,-0.0802892894,0.2926558554,-0.1469325572,0.5887752175,0.075634338,0.4343735874,0.1442782134,0.0386882834,0.0362953953,0.0328157209,-0.0982692167,0.0221132915,0.2159369439,0.5186313391,0.1227994114,0.2945775986,-0.24349612,0.1765810847,-0.2725465894,0.1117469594,-0.1673385799,-0.2924365997,0.3052218854,0.0642081127,-0.1374743879,0.1255759597,-0.2315091342,0.1015428379,0.0534950532,0.1240617484,-0.0011082038,-0.4022593498,0.5199513435,0.2194752842,0.0484596975,-0.2234219015,-0.4599366188,0.0954560563,0.6891955733,0.211067602,0.0918406397,-0.3930459917,0.0459607802,0.042729672,-0.3322361112,-0.0930675045,-0.3673326671,-0.0590910278,-0.1374720335,0.6056536436,0.0173801929,0.1705505252,0.0478076339,-0.2598804533,-0.3689542413,-0.186478138,0.1371017545,-0.082670331,0.3622115254,0.0784875602,0.2764585018,-0.0838743374,0.006524676,0.0195895601,-0.3498882651,-0.0207851566,-0.0098250806,0.0358610563,-0.0645903423,-0.2351195067,0.1936939061,-0.1638864726,-0.3428010345,0.284696579,-0.4136819243,-0.2635734975,0.2818170488,0.0978647396,0.2374093682,0.0065453276,-0.212362662,-0.0676521882,-0.4288432896,0.2715653777,0.0431251824,-0.4024288952,-0.0937344655,0.1991229951,-0.4189532697,0.247228682,-0.5584520102,-0.1095134541,-0.3461774886,0.3311525583,-0.049026642,-0.0865046456,0.4286730587,-0.2382939309,0.0725857764,-0.4164084494,0.2112763077,0.174492985,0.0292624775,0.1258136183,0.0223830566,-0.1327381283,0.0725301504,0.4828671515,-0.0249986164,-0.0059119323,0.2292601913,0.3903651237,0.4262257218,-0.3840201497,-0.2227059901,-0.080941081,-0.1233711019,-0.1908313185,0.0058592162,-0.1408091784,-0.1493293345,-0.2064545304,-0.0898623988,-0.0852783918,-0.2584879696,0.0225740504,0.1879353225,0.0751585588,0.2670049667,-0.0221316461,-0.1375739872,-0.1790686548,0.2276012301,0.2348705828,0.4818471968,0.0687994063,-0.0728833228,0.2068317831,-0.142385155,0.204488501,0.0579376183,0.3608255684,-0.0877024382,0.0018677088,-0.0625026152,-0.3581751883,0.3650507629,0.0000281289,0.1447613239,0.2312398702,0.1738123447,-0.2483163178,-0.2246966511,-0.1057691202,0.2057512403,0.3876726031,0.1502351761,-0.1794173568,-0.1815498918,-0.5600143671,0.42503649,-0.1131398603,-0.5719246268,-0.1554051936,0.2630062997,-0.1834967732,-0.0243509319,-0.1066206396,0.0402692519,0.0427497961,0.0672246367,-0.2629019916,-0.1096929014,-0.1256440133,-0.0542692617,0.0851575956,-0.1368595362,0.3810681701,0.0916618928,-0.201879546,0.2106716633,0.2421762794,0.1579863131,0.1203433126,0.3241555989,-0.2673327327,-0.2261418253,0.5029512048,-0.0074764956,0.0274147596,-0.1036341116,-0.0861972198,-0.0248279646,0.042432826,0.1839613169,-0.002146899,0.2962367535,0.0446315072,0.0962966084,0.1029028594,-0.2100288272,0.5521950722,0.4296231866,-0.0076394756,-0.0116651375,-0.0714474171,0.9117828608,0.0856594369,0.1239151359,-0.0022946692,-0.0114613697,0.1820434034,-0.4523610473,0.0345638357,-0.0690904111,-0.3942634761,-0.0073110731,-0.3197074831,0.3574051857,-0.1237180158,-0.0663599968,0.1295291781,-0.3046334982,-0.041608423,-0.2000088841,0.2936785817,-0.3514897227,-0.2827788889,-0.1272845417,0.1543829441,0.2087812722,0.196007818,-0.1738840491,-0.1202638522,-0.2310530841,-0.0200870968,-0.2846262157,0.0340463333,-0.0473916307,0.3848363459,-0.4509198368,0.1687270701,-0.0667281747,0.1006800756,0.1818889081,0.3895353675,-0.407635808,0.1601204723,-0.0964261517,-0.1486187279,0.1568411589,-0.1128967851,0.2066560388,-0.1192236915,-0.1538930684,-0.2238929123,-0.1098076105,-0.252299875,0.1514246613,-0.1751302928,0.1147944853,0.1661736667,-0.8016607165,0.0738790408,-0.0309026316,0.0092868255,0.0563134588,0.1971632391,-0.1179840863,0.2444865257,-0.3924649656,-0.0779843032,-0.02690983,0.6260701418,-0.2152326554,-0.4358640611,0.3918934762,0.0665365979,-0.1089328229,-0.1784688383,0.3926523626,-0.1328975856,-0.2645893693,-0.1765755266,-0.1570106,-0.0002545246,-0.0735336393,0.0471902452,0.2500626147,-0.1660383344,0.0856894702,-0.4069825709,-0.3362805545,0.0363487378,-0.3742535114,0.1453418434,-0.2492161691,-0.3096071184,-0.1210337579,0.1965864301,-0.1971909106,0.1630774438,-0.1387273371,-0.1570649147,0.3110893369,-0.2056910843,0.1764559746,-0.3125461936,-0.0361618064,0.3433775902,-0.1994083971,-0.03273274,-0.3071316481,0.1769879609,0.0653260127,0.3910630941,0.1879487336,-0.3575829864,-0.0919539407,0.0039843293,0.0577348359,0.2131229937,0.1284113973,-0.106053777,0.1012610421,0.7775290012,-0.1632616967,0.150733009,0.1906184405,-0.1009053439,0.0200923365,0.0011495433,0.0407238007,0.2071180791,-0.5389282107,0.118194893,0.0387223512,0.1698509306,0.3854085505,-0.1609247774,0.1659054309,0.338886261,0.265313983,0.3840015233,0.1110684425,-0.2121677101,0.3436865509,0.1088913456,-0.1733881235,-0.2338851243,0.4099583924,0.0471352339,-0.1620152146,-0.1447361112,0.2118820846,-0.2656679749,-0.1080628186,0.3212931454,0.3122648895,0.0997727439,-0.1332080364,0.109573096,0.1197794452,0.0739810616,0.0810854733,0.1038896069,0.0652775243,0.386290431,-0.5393441319,-0.0703964159,-0.1578869969,0.0889269114,0.010188139,-0.1463050544,0.1170940995,0.2795007527,0.3532570302,0.1238376796,-0.2203018814,-0.0205392875,-0.3257108331,-0.1512861252,-0.1297336072,0.1390152872,-0.1299288422,-0.053966783,0.5443514585,-0.0722043738,-0.0584989749,-0.2154573649,0.1023476869,0.1888383925,0.1588277817,0.4020618498,0.2513804734,-0.2303389013,-0.3110261858,0.1388103366,0.2425661981,-0.2804124355,-0.0637176111,0.1764204204,0.1103199944,0.2126127183,0.0880895332,0.1779111475,0.38775599,-0.3604749143,-0.0337256044,-0.1572816968,-0.0856673419,-0.0655072108,0.3441523612,-0.0111078164,0.0160755739,0.2019850761,0.010938718,-0.147414431,0.1814322919,0.1011064574,0.2597264647,-0.2345615029,0.2407217324,0.2634601295,0.3226812184,0.1312966645,0.314343363,-0.2312020808,-0.217950955,0.0139670074,0.1707110852,0.0836682543,0.0201343782,0.0531918108,-0.2358940691,0.2890699506,0.1749785095,0.0124646053,-0.2690522671,0.1864296198,-0.6664788723,0.1770676821,-0.1591749042,-0.3804576993,-0.0957731307,0.2334534079,0.2743958235,0.2281421125,-0.0008551345,0.1368872672,-0.038788721,0.1089202315,-0.3986485302,0.3031091988,0.278752178,-0.1512165517,-0.169605732,-0.4554014802,-0.0870154724,-0.0802731067,-0.0219096132,0.0148759168,-0.2547795177,0.1321618706,0.1007721871,0.3122495115,0.1472733617,0.4615896046,0.0072003677,-0.4050822258,-0.5162525177,-0.0843256265,-0.100243032,-0.3539431393,0.0499194041,0.4165735245,-0.5587602258,-0.278847903,-0.3024604917,0.4544564784,-0.1803448796,-0.167874366,-0.3694775701,0.0038285365,-0.1557595432,0.3605530858,0.2073545307,0.4394350052,0.0975987539,0.2849321365,-0.4373247325,0.0176638365,0.3319953084,-0.4447521865,-0.3238803744,0.1166257113,0.2005633116,0.0451543331,-0.224660188,-0.1920218468,0.2997680902,-0.0204949155,0.0291678682,-0.0655682683,0.1496957541,0.1357000768,0.0835990161,-0.1955286115,0.1202590689,0.0841853395,-0.0720624626,-0.2006072849,-0.137906298]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/328","title":"Fork dataset","comments":"To be able to generate the Arrow dataset you need to either use our csv or json utilities `load_dataset(\"json\", data_files=my_json_files)` OR write your own custom dataset script (you can find some inspiration from the [squad](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/squad\/squad.py) script for example). Custom dataset scripts can be called locally with `nlp.load_dataset(path_to_my_script_directory)`.\r\n\r\nThis should help you get what you call \"Dataset1\".\r\n\r\nThen using some dataset transforms like `.map` for example you can get to \"DatasetNER\" and \"DatasetREL\".\r\n","body":"We have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n","comment_length":72,"text":"Fork dataset\nWe have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n\nTo be able to generate the Arrow dataset you need to either use our csv or json utilities `load_dataset(\"json\", data_files=my_json_files)` OR write your own custom dataset script (you can find some inspiration from the [squad](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/squad\/squad.py) script for example). Custom dataset scripts can be called locally with `nlp.load_dataset(path_to_my_script_directory)`.\r\n\r\nThis should help you get what you call \"Dataset1\".\r\n\r\nThen using some dataset transforms like `.map` for example you can get to \"DatasetNER\" and \"DatasetREL\".\r\n","embeddings":[-0.2403900176,-0.1519307494,-0.0046681575,0.1388017535,-0.1991832703,0.1661750227,0.0855344608,0.3236584067,0.1992565691,-0.1408088207,0.0522016436,0.676864922,-0.4008255899,0.1611249298,0.2155926079,-0.2229901701,0.1336966902,0.0569895543,0.1327417791,0.1123379022,0.0333925709,-0.1810865551,0.1336497664,0.0569470935,-0.0264412761,-0.228029862,-0.1203991249,0.1118857339,0.0269776415,0.0009552161,0.4616040885,0.1885261834,-0.0171257891,0.3208827376,-0.000117806,0.1305478066,0.0574894063,0.0988214687,-0.2189414799,-0.2057286054,-0.1271968186,-0.1774255037,0.1727922112,-0.1873126775,0.0866854414,-0.3770356774,0.1852132231,-0.2320502251,0.7697014809,0.1443224102,0.1043770611,0.0922689065,-0.124608703,0.2189230025,-0.2451384217,0.3413786292,0.0497781746,0.1736006588,-0.0169577524,-0.20677948,-0.0755041018,0.0774039701,-0.1792523563,0.0272165816,0.2267156541,-0.0575824305,-0.0144064045,-0.1606176794,0.0055113602,0.3372310698,-0.0058291848,-0.2154867202,-0.2079762518,-0.6004967093,0.1435125619,-0.1713543832,-0.1086236909,0.2095341831,-0.1596323252,0.2616223693,0.0679493621,-0.5672591925,-0.0091920318,0.0720446706,0.0448700786,0.2262707204,0.1848957539,0.2127395421,0.2443888336,0.0795683265,0.3197899163,-0.2550052404,-0.0211977717,0.0234170444,-0.1553244144,-0.2513640225,-0.1570304781,-0.5725252032,0.0786036775,0.297617346,0.3420595825,0.0134496707,-0.3720589876,0.1997283399,0.1516202688,-0.2357715368,-0.1162772998,0.3510949016,-0.0139152203,-0.2985035181,0.0886309892,0.1362757087,-0.4068128765,-0.0246962197,-0.275595963,-0.0017347616,0.303938508,0.24162817,-0.1230674088,-0.4214791059,-0.2392184436,-0.1486857384,0.1763511151,-0.0079226987,0.0141327465,0.5878418088,-0.0423438363,0.4042405486,-0.1854728758,-0.2160092145,0.0273767319,-0.0208279565,-0.1518221945,0.0427034162,0.1486830264,0.4183577597,0.1775204837,-0.0463534743,-0.2537013888,-0.0799321458,0.46836555,0.00269732,-0.0315183774,-0.0669803172,-0.0444803685,0.0338094793,0.0545952395,-0.0170926508,-0.2666883767,0.147213608,-0.1560296565,-0.1347330958,-0.0913876295,0.0619671568,-0.0013741046,-0.2966220379,-0.6265000105,0.6202912927,0.2018990964,-0.0099184895,0.0208921041,-0.0113555975,-0.3717825711,-0.2209550738,0.0867062584,0.1400279701,-0.3647662103,-0.3302684128,-0.2051743865,-0.0137742171,-0.1466054618,0.4300253093,-0.2711233497,0.1972029805,0.0315717794,-0.0230285693,0.6090839505,-0.2180972397,-0.036523357,0.1634851992,-0.1810098141,-0.102796562,0.1132742837,0.1215592399,0.0628547147,-0.1734144092,-0.0528556556,0.7855044603,-0.3208650947,0.0275785476,0.1742578149,-0.288834244,0.2234943807,0.2195893079,-0.0326541848,0.0045393747,-0.2716989517,-0.0584679432,0.2219848484,-0.1346555054,0.3524713516,-0.0911045223,-0.1485607177,-0.0693360269,-0.1277203113,-0.2041727602,-0.5392062664,-0.0095585017,0.0517538153,-0.0288864318,0.1372354478,-0.13561441,0.052892141,-0.0021244539,0.0300062131,0.0323185436,-0.0159958825,0.0022268153,0.0753961504,-0.3097703755,-0.3593778014,0.0514098927,0.0316320695,0.1768800467,-0.3762131035,0.5385991335,0.2853182554,0.0463969484,-0.0958447531,0.2762735486,-0.1842873245,0.019609997,0.1722978055,0.0894773006,-0.2095708102,0.3988023102,0.3215484619,0.2255489975,0.2627266347,-0.4603494704,0.350099653,0.0316814519,-0.0697623342,-0.1383180022,-0.2709311843,0.3280028403,0.1001761854,0.4635065794,-0.0369856022,0.2488966584,0.1319746375,-0.138108477,-0.2372635305,0.1167031154,0.011380407,-0.0981424898,0.2888348401,-0.0873024389,-0.4452435076,0.3614445031,0.3100599945,-0.0912935808,0.013084664,0.0618504137,-0.2570584416,-0.0112765171,-0.1076851338,0.5400252938,0.08790306,0.2115104347,0.2821084857,0.0279927403,-0.2887629271,0.0348274782,0.2299976349,0.1382597983,0.5379901528,-0.0090583256,0.1656315774,-0.2080513835,0.0035571624,-0.2566023171,0.3904086947,0.0426651984,-0.048581481,0.4440466166,-0.3302446306,-0.4054999352,-0.6255347729,-0.14469558,-0.0908595398,-0.2803320289,-0.0969087034,0.2085824758,-0.0711979792,0.2793180346,-0.0553824082,0.2379255444,-0.2599774897,-0.0371599421,0.0208123028,-0.515835762,-0.0122365383,0.0440675281,0.524874866,0.2014617622,0.228610158,0.4384919703,-0.2095719725,-0.1107245311,-0.1922592819,0.1964453459,-0.1147720665,-0.0068913992,0.1081255451,-0.1384060532,0.1701038778,-0.2603875101,-0.126014322,-0.0806494951,-0.1264219135,-0.1522290558,-0.1260615587,0.272267729,-0.0650102049,-0.1946340948,-0.3711027205,-0.2666157484,0.3674730062,0.019024875,0.2201631665,0.3498848379,0.1230061427,0.0886919498,0.0806673169,-0.0853569657,0.1626693457,-0.0287244804,-0.0575835221,-0.279826045,0.1778043211,-0.2847963572,-0.3303154409,0.356744945,0.1205294803,-0.2028897554,0.3326491416,-0.0591552928,-0.0081521729,-0.1078237444,-0.2849696875,0.073661685,0.1092720851,0.0739220679,-0.0420350507,0.192833662,0.1713174582,0.0816804022,0.1657755673,0.2389998138,0.2205332369,0.1532522291,0.8823812008,0.0961877257,-0.2332434058,-0.2094873488,0.1341323555,-0.3067849874,0.1412734985,-0.0705723837,-0.141685307,-0.0307902005,0.0928693637,0.0504512861,0.0544679575,-0.3637413979,0.0279164072,0.0164344143,-0.0976040587,-0.3104010522,0.344476819,-0.6596311927,0.0089335861,-0.0463718139,-0.3165359497,-0.34312585,-0.0066951285,-0.3269066811,-0.0686535537,0.1676107198,-0.0292697549,-0.507126987,0.2579825819,-0.4315718114,0.1202673838,-0.0834002346,0.0603631362,0.2149150372,-0.160599798,-0.0080814399,-0.1677450836,0.7057886124,-0.0820003152,0.0358330682,-0.0929186195,0.2638991177,-0.5462855101,0.1581639498,-0.1408661157,0.4263220429,0.1450157613,0.1092651635,-0.3436118364,-0.3307368457,0.1954764426,-0.0901959464,-0.2377159297,0.091670461,0.1204438433,-0.0763523877,-0.2651775777,-0.0840471312,0.1225115508,-0.1332663,-0.1493365914,-0.0854697526,-0.230469957,-0.0773836896,0.1313549429,-0.0339747965,-0.1385278404,0.0584504493,0.2399113774,-0.0263402183,0.1379597485,0.1820327044,0.4787066281,0.0689756721,-0.4266797304,-0.0599450693,0.180578053,0.3776087761,-0.0757100433,0.3494696915,0.4628926516,0.5009769797,-0.1913427263,-0.2140131742,0.1475140005,-0.1411326975,0.1573563069,-0.4000528455,-0.5321150422,0.456771493,0.2663406134,0.0034412523,0.385111481,-0.2737953663,-0.3635835946,0.3545489609,0.1856127232,0.7174018621,-0.1879119724,0.2493954003,0.2364908457,-0.4876545668,0.8625789881,-0.4363308549,0.0064868894,-0.1172526851,0.0915810093,-0.0858598053,-0.1608453393,-0.1302591115,0.6075696349,-0.0718342587,0.175315395,0.2298946828,-0.3705943525,-0.0043971147,0.4035183191,-0.0108016217,-0.3210259974,-0.0747728348,-0.002570776,-0.3699755669,-0.2056083381,0.0852034017,-0.2210645378,-0.109965764,0.0532785952,-0.1914250851,0.1074028909,0.0554041676,0.0998196378,0.1233324781,-0.4270588756,0.2436643988,-0.0507095754,0.1934253722,-0.309584111,-0.1150324717,0.0610688366,-0.1171867847,0.0524721853,-0.0921593904,-0.2186591774,0.3866884112,0.1606971473,-0.2823694646,-0.1151412874,0.0793830156,-0.2370330542,0.0405811742,-0.1036423966,0.1019340083,0.3212151825,-0.131190598,0.4480590522,-0.3300965726,-0.1316241324,0.0037202211,0.0989346132,-0.1084956601,-0.1570864916,-0.043849811,-0.0517287552,0.1102961525,-0.0456445925,0.189800337,0.0328045599,0.4517897069,0.438262403,0.1849964112,-0.1834537387,-0.1205483302,0.3161776066,-0.5272353292,0.034859959,-0.2859779298,-0.1961443722,0.1881087422,-0.1107748821,0.025411034,0.3236614764,-0.1766671538,-0.3987576365,-0.2347109765,0.5112709403,-0.1878656745,0.087379843,-0.0445172414,0.2155918032,-0.0062348461,0.3952154815,-0.18375431,0.2079981267,0.1332274824,0.3373390138,0.0425122045,-0.0526206493,-0.0974916816,0.0780471191,-0.0534388348,-0.0628123954,0.0414903238,-0.1153306291,-0.219652012,0.1184237823,-0.0277115759,-0.1209554672,0.0800826997,-0.2183678597,-0.0677797347,-0.0934069604,-0.1152935326,-0.288554728,0.0195917487,0.2287470549,0.0717817768,0.0809646547,-0.2614687383,0.1132050827,0.1173835322,-0.0293145329,-0.1184697822,0.4501012564,0.2463177145,-0.1423181593,-0.5291055441,-0.036175102,0.2934220433,0.0178206526,0.0412505493,-0.0635458082,-0.093713142,-0.0815743357,0.2306848019,0.4306113124,0.0399840586,0.2269215733,0.2728826106,0.1234994754,-0.0761569291,0.1316493899,0.4431096613,-0.2541219592,0.1279861182,0.3206240535,-0.1318927705,0.1309421957,-0.2874296606,0.0343132392,-0.0642108992,-0.202507481,0.4111101627,0.2690484226,-0.1702833474,-0.1183940768,0.1516787708,-0.0685793832,0.3975032866,0.1099756286,0.211563468,0.4511907399,0.2925952077,-0.0665415078,0.5318660736,-0.1450625807,-0.3541168869,-0.1324116141,0.1264010668,0.264467001,0.2252118289,0.6926088333,0.2272455245,-0.282800138,0.0556712337,0.4929951131,0.0644734129,-0.3106920421,-0.6167650223,0.145427987,-0.1103949174,-0.0357380696,0.014125417,-0.2954674363,0.2550462186,0.1459860951,0.1808397025,-0.3622286022,0.2633672357,0.1927806735,0.0450936221,-0.0546778925,0.1777577996,0.1717291623,-0.1702226251,0.0209486131,0.0810638517,0.097705178,-0.0676141158,-0.4525377154,0.0962839052,0.1943967044,-0.0693977177,-0.1371865124,0.1222014502,0.2356624752,0.008483761,0.2597692907,0.0195141789,-0.1069608703,-0.0792252198,0.3964887559,-0.0399052575,-0.3999809623,0.4468268752,0.2478086352,-0.2203109264,0.2062216252,0.2295757085,-0.3740428686,-0.1238937005,0.3689208627,-0.239084661,-0.0505560376,0.095680818,0.0219324827,-0.0627149269,0.4967689514,0.1408808231,0.4004292786,0.0384270512,0.05476515,-0.3039270341,-0.1496280581,-0.1999980509,-0.0380620845,-0.3141908944,0.1837069541,0.2592154741,0.3747394383,-0.1363188475,0.0206894632,0.287596792,0.3725083768,-0.0866807029,-0.2635836303,0.0417837575,0.3251542747,0.2149627358,-0.3820336163,-0.1086415574,0.4458052814,-0.2035887688,0.0533745624,-0.0341527052,0.3256237209,-0.1564896405,0.0859090164,0.0061817989,0.3960815072,-0.0536642335,-0.3540638685,-0.3446359336,-0.1459411085,-0.5333598852,0.2980462909,-0.0745125487,0.5710756779,-0.582434237,0.1992218196,-0.054803025,-0.5520603657,-0.0332470164,-0.0165009182,-0.4366427362,0.2478283495,-0.2463943511,0.1747565717,0.0671191514,0.4830775857,0.0326565728,0.0731242299,-0.1083625406,0.0193063188,0.2219753116,-0.1630266011,-0.0337870345,0.0701634884,-0.0563481674,0.1557708234,0.1068743765,-0.5881656408,-0.0457236245,0.2437045276,-0.2993692458,-0.0371276401,0.1695758104,0.220313862,-0.0677686036,-0.1151805967,0.3601847887,0.0467816666,-0.3041184843,0.0224007145,-0.3639565408]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/328","title":"Fork dataset","comments":"Thanks for the helpful advice, @lhoestq -- I wasn't quite able to get the json recipe working - \r\n\r\n```\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/pyarrow\/ipc.py in __init__(self, source)\r\n 60 \r\n 61 def __init__(self, source):\r\n---> 62 self._open(source)\r\n 63 \r\n 64 \r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/pyarrow\/ipc.pxi in pyarrow.lib._RecordBatchStreamReader._open()\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\nArrowInvalid: Tried reading schema message, was null or length 0\r\n```\r\n\r\nBut I'm going to give the generator_dataset_builder a try.\r\n\r\n1 more quick question -- can .map be used to output different length mappings -- could I skip one, or yield 2, can you map_batch ","body":"We have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n","comment_length":87,"text":"Fork dataset\nWe have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n\nThanks for the helpful advice, @lhoestq -- I wasn't quite able to get the json recipe working - \r\n\r\n```\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/pyarrow\/ipc.py in __init__(self, source)\r\n 60 \r\n 61 def __init__(self, source):\r\n---> 62 self._open(source)\r\n 63 \r\n 64 \r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/pyarrow\/ipc.pxi in pyarrow.lib._RecordBatchStreamReader._open()\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.pyarrow_internal_check_status()\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\nArrowInvalid: Tried reading schema message, was null or length 0\r\n```\r\n\r\nBut I'm going to give the generator_dataset_builder a try.\r\n\r\n1 more quick question -- can .map be used to output different length mappings -- could I skip one, or yield 2, can you map_batch ","embeddings":[-0.2709982991,-0.1708691269,-0.0142526412,0.2424219102,-0.0901527107,0.037377201,0.0722795576,0.3559105396,0.0219828505,-0.1569152027,0.0568319112,0.6875869036,-0.3158581555,0.0238236431,0.1384809613,-0.1774831414,0.1697377712,0.0175792221,0.1954474896,0.131913662,0.0567809828,-0.1072145179,0.0791314393,0.1026265845,-0.1071948856,-0.251937747,-0.0987824127,0.0902697444,-0.0763715804,-0.141680643,0.3796569705,0.1386334747,-0.055186294,0.3206908703,-0.000122671,0.2056131512,0.1715611666,0.1351349652,-0.200448066,-0.1448926032,-0.1280783415,-0.1294977069,0.188286677,-0.1889585853,0.1550608724,-0.3981489241,0.0495883562,-0.1102559045,0.7451307774,0.2432459891,0.0788233802,0.0991889387,-0.0036854877,0.2266288549,-0.118631199,0.3455074728,0.0820627958,0.2443725169,0.0720906332,-0.215568915,-0.1497940123,0.1466435045,-0.1729086936,0.0248665381,0.2092568129,-0.0760230646,0.1784182489,-0.0695761219,0.0831777304,0.3103863895,0.0022011038,-0.2151121795,-0.2066604644,-0.5397039652,0.1270538718,-0.2709568143,-0.070603773,0.2371488363,-0.2252679169,0.2454847097,-0.058990173,-0.4718372822,-0.0915905312,0.0282218084,-0.0222262386,0.2925259769,0.2098951042,0.3022347391,0.1680295467,0.0337278321,0.3152107894,-0.2900062203,-0.0619789362,0.0518531501,-0.1652068943,-0.2472952306,-0.0889400765,-0.5191529989,0.0715018362,0.2056325674,0.3251101971,-0.0732594505,-0.2554788888,0.1540989876,0.1518896073,-0.2257803231,-0.1276627034,0.3733853996,-0.0609765612,-0.2684909403,0.0765160844,0.1259830594,-0.2788668871,-0.1188951209,-0.1605695635,0.0677713603,0.2837854922,0.2333727032,-0.1346600056,-0.3933128417,-0.386433959,-0.2138633728,0.1876254976,0.03080694,0.0291465819,0.5331355929,-0.0671073273,0.4122139513,-0.2591350377,-0.2082477361,0.0152530279,-0.0591532625,-0.2480623424,-0.00271524,0.1099961847,0.4786125422,0.1313544661,-0.0131072924,-0.2265704423,-0.0390045419,0.3589449823,-0.1173250899,0.0950445235,-0.0254520886,-0.1783608049,0.0185399521,0.0100759463,-0.0709420368,-0.3327418864,0.2551265955,-0.1009522676,-0.1926943362,0.0021885815,0.0228179675,-0.0413328893,-0.1952394694,-0.4399907291,0.616286099,0.3059460521,-0.0188291837,0.0528093912,-0.1076049581,-0.2903373539,-0.2586704791,0.1246924475,0.1132117063,-0.4203320146,-0.2970167994,-0.0107072741,-0.0253951382,-0.1112256795,0.421446085,-0.2157961279,0.1598107368,0.0871579275,0.0753271282,0.6000877023,-0.24979873,-0.0632772669,0.202613309,-0.1168496683,-0.156263575,0.1286632717,0.0331943929,0.0206194595,-0.0945704654,0.0209348779,0.6081792116,-0.3583090007,0.0410197787,0.009521381,-0.3130632639,0.2251249403,0.2373591065,0.0278315172,-0.1905456632,-0.3208423853,-0.1956507117,0.2701670825,-0.0993392467,0.3689213991,-0.0595755614,-0.0842377618,-0.1394963413,-0.0287111495,-0.1532613635,-0.5317848921,0.0004241855,-0.0153504675,0.0928583294,-0.044291202,-0.1126272008,0.0699822158,0.0900176242,-0.0436857715,-0.0526153371,-0.0137467533,-0.0741436258,0.1348852664,-0.2574138343,-0.3342540264,-0.1103570908,0.0086052101,0.2037959695,-0.3359615207,0.5445890427,0.2347190678,-0.0095283492,-0.1654924452,0.251721859,-0.0922930166,0.0961074159,0.1214930266,0.1124425307,-0.1708140671,0.3681159914,0.1516937912,0.180601269,0.3069016635,-0.4986450672,0.4170792699,0.0051066941,-0.1247829795,-0.1463017613,-0.2852309644,0.2827926874,0.0896427408,0.5726414323,0.0074494923,0.1161555573,0.0105549926,-0.1423983723,-0.2524617314,0.153743431,-0.0168345403,-0.00683589,0.330814451,-0.0498399325,-0.4534574747,0.3651574552,0.3189692497,-0.0790392011,-0.1559526175,0.0444515981,-0.2688310444,-0.0187270585,0.0327847786,0.4359911084,0.1295998991,0.2646863759,0.3439447284,-0.023430381,-0.2978405356,-0.05219483,0.2612974346,0.1168888882,0.5971403718,0.0559234805,0.2753997445,-0.2705343962,-0.0698106661,-0.2716559768,0.4424142838,0.0593959391,-0.1237623617,0.4224994481,-0.2999272943,-0.4046645761,-0.6216878891,-0.226263538,-0.1811684668,-0.3333445787,0.0100121684,0.1288983673,-0.1645735949,0.3371834159,0.0312465783,0.0554596409,-0.1915539801,0.0031767224,-0.073634468,-0.4876655042,-0.055017814,0.0224791169,0.5342305303,0.1249972209,0.2216801047,0.5519185662,-0.3891839683,-0.037762776,-0.3224725425,0.2598617971,-0.1278274208,-0.0251384564,0.1154051274,-0.141209051,0.251417011,-0.210085392,0.045722913,-0.0451223627,-0.1329952478,-0.1487724185,-0.1853579581,0.2294775397,-0.0772327781,-0.1576831639,-0.3379792273,-0.2774730921,0.427798003,0.0270631481,0.1580067128,0.3351547122,0.1768142134,0.0498505309,0.0437799431,-0.0661600977,0.1313047558,0.0498726815,-0.0275705121,-0.1940673739,0.2004646808,-0.3057989776,-0.3318732679,0.4547982514,0.215854302,-0.2086229324,0.3465017974,-0.0335923731,0.1218090877,-0.0845845342,-0.2690059543,0.1025605574,0.1928120255,0.089964211,-0.0179420784,0.2616141737,0.1740378588,0.0247754809,0.1520762444,0.2242884338,0.2025141567,0.1355823576,0.961515069,0.0917422175,-0.1690108478,-0.259436816,0.0101259807,-0.371111542,0.1074114069,0.0459491871,-0.1450534016,-0.0656828582,-0.0021031871,-0.0002098807,0.0568529591,-0.4240120649,0.0956174731,0.0687125698,-0.1075339392,-0.3062421381,0.3730202317,-0.5912077427,0.118031539,-0.0506455936,-0.2795487642,-0.3265397251,0.0445426926,-0.2696339786,-0.1826967597,0.1530706882,-0.1265435964,-0.451241672,0.0924337059,-0.5370225906,0.1233645827,-0.0071201944,-0.0968020409,0.226779595,-0.2825099528,0.0064118793,-0.2530957758,0.729845047,-0.0977628753,0.0527748615,0.0396025218,0.2298903763,-0.5488437414,0.1013626754,-0.1378295422,0.3780576587,0.2070141286,0.1954917759,-0.4321470261,-0.3591599762,0.2859161794,-0.1534584463,-0.2045990378,0.0034125636,0.1774920225,-0.0738245919,-0.2873137295,-0.0029661814,0.1983371377,-0.0111111263,-0.15937078,-0.0543320589,-0.1382427812,-0.0134045407,0.0286665,0.0071774665,-0.0436875448,-0.1088057905,0.2084569633,-0.0966156051,0.0845700279,0.2789638638,0.5220506191,0.0554297939,-0.3628227413,-0.0905189291,0.1511208713,0.3655584455,0.1464164704,0.3258492053,0.4495049715,0.3898549676,-0.0241511576,-0.1499517858,0.0184981804,-0.1162972227,0.1519062817,-0.3429191113,-0.3364130855,0.5278544426,0.244213745,-0.0316028632,0.428801626,-0.2757220566,-0.3967437446,0.2894907594,0.2824402153,0.8820127249,-0.0412443429,0.2451674193,0.4288580716,-0.4744338691,0.689979732,-0.5011779666,-0.0081260167,-0.2250358909,0.1768207103,-0.0109702395,-0.1766536832,-0.1450345665,0.6129155159,-0.058989808,0.0663664937,0.3291284144,-0.3208192587,0.0056246021,0.3306918144,-0.057552129,-0.1969763786,-0.1106470525,-0.0546386987,-0.2695824504,-0.2840951979,0.0886819214,-0.3098029494,-0.1838231236,-0.0833676383,-0.1295934618,0.0842533559,-0.0685071945,0.0788909122,0.2631238401,-0.4813998044,0.236975342,-0.0980193466,0.0542123057,-0.4273022711,-0.1175628379,0.1090059057,-0.0264748018,0.0396040231,-0.1083836406,-0.2043939233,0.3602646291,0.0853202417,-0.2321543097,-0.1213420033,-0.0468100719,-0.0623497367,0.001722108,-0.1963724494,0.1519373357,0.2448466271,-0.0199869853,0.4476201832,-0.3700377345,-0.086066924,-0.0133537333,0.0342977867,-0.1728719026,-0.0901799053,-0.0315409116,-0.0473661572,0.0888057351,-0.0653011724,0.3045382202,0.0887990966,0.4743157327,0.4823849201,0.1281764656,-0.1675911248,-0.1615802944,0.3642773628,-0.6409146786,0.0868002996,-0.2537645102,-0.15237239,0.1737757176,-0.0462127887,0.0672982186,0.4354813695,-0.1906712949,-0.2262094617,-0.2581787109,0.5244814157,-0.145292446,0.1204876006,-0.1918353736,0.3252471983,0.0749841034,0.4396362901,-0.1820240021,0.2253145427,0.0979340002,0.3882689476,-0.0723962486,-0.1090587452,-0.1132069528,0.070518434,-0.0724810138,-0.0643317848,0.0123351542,-0.0894461423,-0.1491754055,0.1177711785,0.0603620075,-0.1018851101,0.1024392843,-0.1873316318,-0.009459625,-0.036971435,-0.076271683,-0.349060595,-0.0733242705,0.1595446169,0.1517009288,0.1069021225,-0.3149268925,0.1749271005,0.1201311946,0.0713576972,-0.0852652416,0.4262804985,0.230299741,-0.1162020862,-0.4974854887,-0.0258680414,0.1137558743,-0.0623666868,0.1227828488,-0.1025782377,-0.1341922283,-0.2581683397,0.2884768844,0.4120496809,0.0171677768,0.1973018646,0.1009962186,0.1181034371,-0.1254897565,0.15020518,0.3381069899,-0.3637137115,0.1236136183,0.2436734587,-0.166686669,0.1396011263,-0.2355642766,0.1679411829,-0.0064404849,-0.2678118646,0.3055499196,0.3528040349,-0.1886676401,-0.0293505434,0.2828600407,-0.0730519444,0.3902784288,0.0133101586,0.1955373734,0.4305582941,0.2297839969,-0.0683105811,0.4417047799,-0.056118764,-0.4187366664,-0.0846249759,0.120066084,0.1204311922,0.2948899269,0.7293596864,0.3205150068,-0.3478203416,0.1853823364,0.5375406146,-0.0106578479,-0.2062982917,-0.7132182717,0.0688483268,-0.167631641,-0.0260273013,-0.1149510443,-0.2241872996,0.3993258476,0.1802768558,0.2019369751,-0.420106411,0.2862582803,0.1332343817,0.1023653001,-0.1227316484,0.234530285,0.2515219748,-0.1935739219,0.0052805613,0.1113868728,0.237328738,0.0086043673,-0.4945700467,0.0074294787,0.2081177086,-0.1191313714,-0.1330265105,0.1433831155,0.3661993146,0.0611306056,0.3061699569,-0.0200905334,-0.0907648578,-0.0039543766,0.3932209611,-0.0377735309,-0.2537739873,0.5006620884,0.1897699386,-0.3410308063,0.2525041699,0.1708341986,-0.4054368138,-0.0208982546,0.3292309642,-0.2298856527,-0.0052026445,0.0529686101,0.0116043454,-0.0510987118,0.5025145411,0.2102500349,0.435362339,-0.1251078546,-0.0355421826,-0.3355429471,-0.1686947197,-0.3205323219,0.0130562661,-0.2437047362,0.2071515769,0.2902574837,0.3711934984,-0.0773659199,0.083468318,0.2264590859,0.4366154969,-0.1140753105,-0.2237603962,0.1184246391,0.2578309476,0.137567997,-0.3139264584,-0.0115698511,0.3504678607,-0.2153019458,0.0389457904,-0.0080328723,0.4428017437,-0.0256609805,0.0372274369,0.0445508994,0.4479853809,-0.0112552373,-0.2736093104,-0.1870454848,-0.1563470364,-0.5374302864,0.3463817537,-0.0637223795,0.6216843128,-0.5220506191,0.1028234437,-0.0278854221,-0.5713891983,-0.1363963038,0.0437540635,-0.3469476402,0.1673676819,-0.1485082358,0.2270120531,-0.0076643201,0.4074018598,0.1073877513,0.054612197,-0.2766765058,-0.0632088184,0.2735039592,-0.1833561659,-0.0337594151,0.0563396178,-0.1401645094,0.1763148457,0.1132691354,-0.6911987662,-0.0561757237,0.271166414,-0.3308924437,-0.0930013582,0.1532911062,0.3074660003,-0.0901591778,-0.1591700464,0.2922626734,0.0803831965,-0.3020390868,-0.111715883,-0.4423952103]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/328","title":"Fork dataset","comments":"You can use `.map(my_func, batched=True)` and return less examples, or more examples if you want","body":"We have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n","comment_length":15,"text":"Fork dataset\nWe have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n\nYou can use `.map(my_func, batched=True)` and return less examples, or more examples if you want","embeddings":[-0.3346369863,-0.2905991673,0.0045895372,0.1299723089,-0.2365575433,0.1289829761,0.0239409488,0.3387122452,0.1523958296,-0.1180353686,0.0310889184,0.6847208738,-0.3909477592,0.1242850646,0.1493852288,-0.2603683472,0.1640078574,-0.0049415352,0.1889290363,0.1098544672,0.0590775684,-0.1472997367,0.0898680463,0.0470929444,-0.0557252355,-0.3060141206,-0.1320464164,0.096111916,0.0956597775,0.0586656407,0.4413862228,0.1155743375,-0.0290669985,0.3029562235,-0.0001199958,0.1147022769,0.0752085522,0.1012525856,-0.138329193,-0.1019314751,-0.1672243029,-0.0701655224,0.1556930542,-0.1435968429,0.0580889434,-0.2994049489,0.188038677,-0.1922176331,0.7377747297,0.1823512018,0.0922068134,0.0907790735,-0.1599233598,0.2076610327,-0.2389333546,0.2795394361,0.1008362323,0.1607557535,0.0535767153,-0.2243674994,-0.1141725406,0.0857426673,-0.2030309737,0.0272745024,0.0708339512,-0.097776629,0.0713677332,-0.118151404,0.0560670644,0.3613093495,-0.0763622001,-0.1443130523,-0.1623984128,-0.5952412486,0.0264292583,-0.1816348881,-0.169340089,0.1712503135,-0.1901026964,0.259278208,0.0675268993,-0.5283116102,0.0577740185,0.0710999072,0.0034218046,0.3245171309,0.2226323485,0.2377549857,0.2797088921,0.0631054491,0.2424338162,-0.1992803663,-0.0559087247,0.0265773013,-0.1351418346,-0.2755252719,-0.0949984044,-0.5591034889,0.0522331297,0.2348721772,0.2602282465,0.0478156954,-0.34194085,0.2133614123,0.1584059894,-0.2696143091,-0.1579033881,0.4144700766,-0.0090351868,-0.3119778037,0.0037919381,0.1628254503,-0.3416497409,0.0068364781,-0.2426389605,0.0194476675,0.3017448187,0.225370422,-0.1404337436,-0.4087178111,-0.233212322,-0.1508304626,0.1488438696,0.073557511,0.0544238389,0.5004848242,-0.1485155672,0.3721607924,-0.2033411711,-0.2171865702,-0.0101306029,-0.0262930039,-0.1842501014,-0.021260852,0.0799654573,0.4704395235,0.1748025715,-0.0281776506,-0.3014272153,-0.0704004019,0.4852121472,-0.0472328588,0.0310234949,-0.059087377,-0.0905534849,0.0989337265,0.1160620302,-0.0145644657,-0.2945983708,0.1980253905,-0.163014695,-0.1834073216,0.0143209994,0.0353404917,0.0041248016,-0.2675306797,-0.5819891095,0.7204762101,0.2163040936,-0.0420115925,0.0052470472,0.0037708755,-0.3689194322,-0.2281747758,0.0503614396,0.0411945581,-0.2873538435,-0.3595952988,-0.165065974,-0.006132538,-0.0445507318,0.4503319263,-0.2192365378,0.1978186816,0.0164650027,-0.0081617665,0.6175189018,-0.2688552737,-0.0317374952,0.1720315069,-0.1587014049,-0.051949434,0.0124366879,0.1322464198,0.0338332877,-0.134811461,0.0251711588,0.7220831513,-0.3678635657,0.0825074762,0.1195987985,-0.2593340576,0.2777491808,0.1504963934,0.0076287156,0.0270513017,-0.3369589746,-0.1431936473,0.2638602853,-0.0755939037,0.3065678179,-0.0940674916,-0.0834679976,-0.1285141855,-0.1350618899,-0.2905486524,-0.4702569544,-0.0423907153,0.1101495549,-0.0005797385,0.2169109881,-0.1439381689,0.1243568212,0.0096357455,0.0751200765,-0.0362162367,-0.0211712401,0.0122632105,0.0971044451,-0.2936196923,-0.3370069861,0.0133792916,0.0615259781,0.1845170557,-0.3425635993,0.4843218029,0.2870934904,-0.0144353975,-0.1654703468,0.2139431238,-0.1530406326,0.0163274836,0.1811429858,0.0715775937,-0.1133946702,0.3805439472,0.2105073333,0.1802950352,0.1951273084,-0.4252529442,0.2449197322,0.0787186474,-0.1373574287,-0.1239128858,-0.3280932009,0.2802168727,0.1698680818,0.423856467,0.010775445,0.2063915133,0.1395849884,-0.1485157609,-0.2127303183,0.1293695122,0.0775984228,-0.1525904983,0.2756582797,-0.1003767923,-0.4989168048,0.3072444797,0.2525891066,-0.0560956486,0.0312047582,0.0081936121,-0.268137008,-0.0020507977,-0.0218377784,0.5145918727,0.0544344597,0.218205452,0.3417751193,-0.0168789718,-0.287727803,0.0795073509,0.2356286794,0.1735762209,0.5633508563,-0.0529241078,0.2058375627,-0.2192979604,-0.0141504183,-0.2521547377,0.3778597713,0.0883413702,-0.0387282558,0.4170477092,-0.34023121,-0.4615244269,-0.6157364845,-0.1191680953,-0.0881848335,-0.2857559919,-0.078039661,0.2602833211,-0.1589985937,0.322326988,0.0051134927,0.2372372746,-0.2958425283,0.0131766172,0.0198555905,-0.4852550626,-0.0349821113,0.0445157178,0.4686111808,0.1691176593,0.2544096708,0.5151556134,-0.2722876072,-0.0475558974,-0.2421540022,0.2172394246,-0.0526786149,-0.0144642415,0.1212807447,-0.1397141218,0.1484634876,-0.2834120393,-0.1225570589,-0.0914620534,-0.1284601539,-0.1867054552,-0.0826589763,0.2708037496,-0.0466291457,-0.0882745013,-0.2832022607,-0.2763592899,0.3321955502,-0.0378945395,0.1954707354,0.3487655818,0.1022650599,0.0639542118,0.0836357921,-0.1003656536,0.1840342283,-0.0381663814,-0.1818484813,-0.2585034668,0.2495633513,-0.3190028965,-0.3268358409,0.362960279,0.2119031996,-0.1588648409,0.340098083,-0.0857388303,-0.0215222817,-0.1402032375,-0.3361780047,0.0830794498,0.1209460124,0.0884164795,-0.056830734,0.2171324193,0.2067977488,0.0994640738,0.1461563557,0.1654238552,0.1888929904,0.2213220894,0.8788926005,0.0933685824,-0.2181233615,-0.2121333629,0.0699073821,-0.3979445398,0.1082922518,-0.0342064165,-0.1685487479,0.0095562451,0.0296418238,0.0982939005,0.0613760129,-0.4171889424,0.0606728718,-0.0278420355,-0.0531350374,-0.3124678135,0.3622234166,-0.6418172121,0.0531562828,-0.0153342774,-0.3515082002,-0.3239828348,0.0373007245,-0.3267600834,-0.1460815072,0.1628180891,-0.0390239693,-0.5418767929,0.2421475798,-0.4707146883,0.117870301,-0.0457928367,0.0181628671,0.1634895504,-0.2028915286,-0.0358227082,-0.1978738904,0.7761897445,-0.1486954689,0.1045594215,-0.086294055,0.2548668981,-0.539268136,0.1201647744,-0.213332504,0.5042408109,0.2546928525,0.1899404228,-0.3215149939,-0.2905110419,0.1733634323,-0.1758829951,-0.1944424212,0.1281394958,0.1168283597,-0.0208331943,-0.2756370306,0.0388242714,0.1981861293,-0.1611319333,-0.1113397852,-0.0951767266,-0.2301094085,-0.0364995077,0.1228010878,-0.0269768853,-0.1568614542,0.0397269875,0.1984696984,-0.1559872478,0.1784731001,0.1417881995,0.4578437209,0.0180491004,-0.4073977172,-0.0875062421,0.2106755674,0.4006624818,-0.0709261671,0.3939453363,0.510964632,0.4926321208,-0.1470881402,-0.1777945608,0.1908736676,-0.1252265722,0.2534103394,-0.504591167,-0.4291280806,0.4009571671,0.2849857807,-0.0533991717,0.4638133943,-0.2737786174,-0.3485088646,0.3723090291,0.1661353856,0.7055780888,-0.1466169357,0.2389609218,0.2009732723,-0.4366592765,0.868771255,-0.4210555851,0.0766770765,-0.040186476,0.032464996,-0.0487447977,-0.144396916,-0.1404660493,0.5983471274,-0.0745264962,0.1203843802,0.2419548333,-0.345407933,-0.0123911481,0.2898671329,0.0883792862,-0.2093202323,-0.0661762655,-0.0238353591,-0.3497836888,-0.1946277767,0.0620665029,-0.1764033586,-0.1096644849,0.026102351,-0.2287573069,0.0677429289,0.0600383282,0.082600221,0.1503495723,-0.4121502936,0.1749402285,-0.1097854748,0.14498806,-0.3075174391,-0.0634466186,0.0554296263,-0.1189656407,0.0748734698,-0.1034143418,-0.2404307127,0.3782981038,0.1453205943,-0.2098640501,-0.0862830281,0.0511557907,-0.1705933958,-0.0227760486,-0.0659108162,0.0965203643,0.305904001,-0.1571924835,0.5255258679,-0.3442078829,-0.122175999,-0.0184162315,0.0533687957,-0.1173292845,-0.1567319036,-0.0338432603,-0.0739033446,0.1219682246,-0.0319728814,0.2571457326,0.0202132575,0.3703294992,0.4124453068,0.2566840053,-0.1946304291,-0.1607700586,0.3580013216,-0.4678655267,0.10800381,-0.3504838645,-0.2385917157,0.2203981131,-0.0914210603,0.0170231927,0.3432518244,-0.2137129158,-0.3144957423,-0.233439967,0.4745205939,-0.1862145811,0.1094229072,-0.1052319854,0.2606518567,-0.0156351347,0.4432114959,-0.1687540561,0.2332804948,0.1350781322,0.4413145781,0.0387718827,0.0243234523,-0.1203202084,0.1120130643,-0.0812513083,-0.0335139968,0.1308237612,-0.1032976657,-0.2069464773,0.1124954447,-0.0037095586,-0.1680113673,0.1606001854,-0.1850098372,-0.0153193939,-0.1005848944,-0.093040958,-0.3235063851,-0.0079568019,0.2469779551,0.111027278,0.0821275935,-0.2267105132,0.131045118,0.087246947,-0.0692992508,-0.1236708239,0.3961805105,0.3209515214,-0.1519153416,-0.4546282589,-0.1106420904,0.3199375868,0.0121508306,-0.0697688833,0.0001697985,-0.1329859346,-0.1511508375,0.231732294,0.3268661797,0.0378122404,0.2754825056,0.2903771102,0.1155448481,-0.0783660784,0.075875856,0.4812789261,-0.3554058373,0.1485852003,0.2967045605,-0.2090078741,0.1524200439,-0.2145496458,-0.0183464885,-0.0550994836,-0.2257976681,0.4465484023,0.2472873628,-0.1883016825,-0.1019298062,0.0968005955,-0.1081869006,0.4057563245,0.0910513401,0.296821028,0.471306473,0.3385431767,-0.1478305757,0.5746856332,-0.2064492851,-0.3939862549,-0.1088997275,0.1100518182,0.3292792439,0.2560751736,0.5604239106,0.2669286728,-0.3317898512,0.1080756709,0.4791411757,-0.0390843414,-0.2980708182,-0.5172961354,0.1627462208,-0.0722387508,0.0130223706,-0.0139457425,-0.2846928239,0.2932442427,0.1464134604,0.2300208956,-0.3004267812,0.2164546102,0.1449513733,0.0987057462,-0.0331092328,0.2175317854,0.1192751899,-0.1970123947,-0.0081386231,0.076489307,0.1427544504,-0.0646154061,-0.4738244116,0.1213101894,0.1629222929,-0.0534312464,-0.2285665572,0.0937255621,0.28203547,0.0138193667,0.2678010762,-0.0056557297,-0.0916739404,-0.0102476403,0.3961052299,-0.066469714,-0.4489337504,0.4304958284,0.2328756452,-0.2346701026,0.2692180574,0.219907999,-0.3354950249,-0.1185060814,0.3095113933,-0.3295990229,-0.0383849815,0.1485318691,0.0151727302,-0.0925756842,0.4863446355,0.1421741247,0.483845979,0.0333012976,-0.0390868708,-0.2861442268,-0.1194182187,-0.3012876213,0.0027667873,-0.3642533123,0.143086791,0.2512269616,0.3455854058,-0.1302718371,0.0054842248,0.2898448408,0.4213133156,-0.0745828301,-0.2825375795,0.0021368372,0.2758943439,0.2426659763,-0.3355152011,-0.0582231656,0.4975121617,-0.2096102685,0.0825032741,-0.0641311482,0.2991265357,-0.1624661684,0.0360358395,0.0767469853,0.4100485742,-0.1069861948,-0.3428489864,-0.2828327417,-0.1087342575,-0.5412107706,0.2263096571,-0.079279989,0.5459894538,-0.5765148997,0.2272714376,0.0203590095,-0.6251535416,0.0093320999,0.1200962588,-0.4240585268,0.2451799512,-0.1713968962,0.1579105407,0.0632452667,0.5244281888,0.046682477,-0.020811215,-0.1485336125,0.0630470812,0.1833838969,-0.2271651775,-0.0664416775,-0.0360068679,-0.0639730617,0.1696992368,0.13969329,-0.5836237669,-0.1437496543,0.2549634278,-0.2982183397,-0.0766463652,0.2082031071,0.263471961,-0.1472626328,-0.0759599358,0.3760736883,0.1212364435,-0.3561181426,0.1097381189,-0.4027212858]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/328","title":"Fork dataset","comments":"Thanks this answers my question. I think the issue I was having using the json loader were due to using gzipped jsonl files.\r\n\r\nThe error I get now is :\r\n\r\n```\r\n\r\nUsing custom data configuration test\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-38-29082a31e5b2> in <module>\r\n 5 print(ner_datafiles)\r\n 6 \r\n----> 7 ds = nlp.load_dataset(\"json\", \"test\", data_files=ner_datafiles[0])\r\n 8 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: int64> nor list<item: int64>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions.\r\n```\r\n\r\nIf I just create a pa- table manually like is done in the jsonloader -- it seems to work fine. Ths JSON I'm trying to load isn't overly complex - 1 integer field, the rest text fields with a nested list of objects with text fields .","body":"We have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n","comment_length":324,"text":"Fork dataset\nWe have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n\nThanks this answers my question. I think the issue I was having using the json loader were due to using gzipped jsonl files.\r\n\r\nThe error I get now is :\r\n\r\n```\r\n\r\nUsing custom data configuration test\r\n---------------------------------------------------------------------------\r\n\r\nValueError Traceback (most recent call last)\r\n\r\n<ipython-input-38-29082a31e5b2> in <module>\r\n 5 print(ner_datafiles)\r\n 6 \r\n----> 7 ds = nlp.load_dataset(\"json\", \"test\", data_files=ner_datafiles[0])\r\n 8 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n--> 738 parse_schema(writer.schema, features)\r\n 739 self.info.features = Features(features)\r\n 740 \r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/builder.py in parse_schema(schema, schema_dict)\r\n 734 parse_schema(field.type.value_type, schema_dict[field.name])\r\n 735 else:\r\n--> 736 schema_dict[field.name] = Value(str(field.type))\r\n 737 \r\n 738 parse_schema(writer.schema, features)\r\n\r\n<string> in __init__(self, dtype, id, _type)\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in __post_init__(self)\r\n 55 \r\n 56 def __post_init__(self):\r\n---> 57 self.pa_type = string_to_arrow(self.dtype)\r\n 58 \r\n 59 def __call__(self):\r\n\r\n~\/.virtualenvs\/inv-text2struct\/lib\/python3.6\/site-packages\/nlp\/features.py in string_to_arrow(type_str)\r\n 32 if str(type_str + \"_\") not in pa.__dict__:\r\n 33 raise ValueError(\r\n---> 34 f\"Neither {type_str} nor {type_str + '_'} seems to be a pyarrow data type. \"\r\n 35 f\"Please make sure to use a correct data type, see: \"\r\n 36 f\"https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions\"\r\n\r\nValueError: Neither list<item: int64> nor list<item: int64>_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https:\/\/arrow.apache.org\/docs\/python\/api\/datatypes.html#factory-functions.\r\n```\r\n\r\nIf I just create a pa- table manually like is done in the jsonloader -- it seems to work fine. Ths JSON I'm trying to load isn't overly complex - 1 integer field, the rest text fields with a nested list of objects with text fields .","embeddings":[-0.2616365254,-0.2061418295,0.0375017449,0.3549158573,-0.1682421863,0.1174596846,0.1433506906,0.4742084146,0.3013393879,-0.0251564812,-0.0413274691,0.7794690728,-0.4380642772,-0.0006891052,-0.0343453512,-0.1329026371,0.078056626,0.0321564749,0.1479655653,0.1702057421,-0.0766780823,0.0815187022,0.1405754238,0.0701387823,-0.0335839391,-0.23559241,-0.0052827508,0.322276175,-0.0223012716,-0.0998795033,0.3949323893,0.025425557,0.0121661732,0.370148629,-0.0001212367,0.2414514273,0.1444253325,0.0434316434,-0.2983271182,-0.1597706378,-0.1191835627,-0.2082093805,0.2016608864,-0.1714549512,0.0930440053,-0.3221847415,0.010363779,-0.2175858319,0.8227084279,0.1959593892,0.085380286,0.094117865,-0.0375261307,0.1864823252,-0.1528877467,0.2486982346,0.0706446469,0.4197851419,-0.0719188601,-0.158483699,-0.1350350529,0.1528037339,-0.21148929,0.1532353312,0.1155335903,-0.0662333667,0.1061297208,-0.0129734129,-0.0425176546,0.2376263887,-0.0137331095,-0.1154446304,-0.2180214226,-0.5963210464,0.0687730238,-0.3153635263,0.1451151222,0.1890842319,-0.2590257525,0.2256511152,0.1236033216,-0.5307229757,-0.0557474494,0.1267889142,-0.0098405648,0.3766208887,0.1091585681,0.2603290081,0.2779296935,0.0567415208,0.1978452653,-0.1859574169,-0.1330088377,-0.0677112937,-0.1957906485,-0.272372067,-0.1936009526,-0.7956878543,-0.0301025473,0.2747699022,0.4105606079,-0.0649349988,-0.2221665382,0.2716937959,0.2230012715,-0.1622830629,-0.1264720708,0.3251129389,0.0797926933,-0.1740464419,0.111032851,0.1423989087,-0.3329451382,-0.1393825412,-0.2668761909,0.035817977,0.3838409483,0.1243890375,-0.1721629798,-0.3282208741,-0.4179406464,-0.1695621014,0.1243307143,0.0437096506,0.0265701655,0.4992043972,0.0284126885,0.3907045424,-0.1886146218,-0.2341752499,0.0649520531,-0.0155802146,-0.115491353,0.0258706603,0.2306987345,0.393681556,0.2521832585,0.0243458468,-0.2440880388,-0.084308289,0.3822432756,-0.0215821322,-0.0870607421,0.0885294601,-0.0290046353,0.0562349223,0.1478366107,-0.16957739,-0.328660965,0.2814548016,-0.2943262756,-0.1961506754,-0.0213513691,0.0327280276,-0.0374715142,-0.2179765999,-0.5268861651,0.3935555816,0.347150743,-0.1026957855,-0.0189551022,-0.1631554067,-0.3421741128,-0.2664163411,-0.0560863055,0.1445371211,-0.4778347909,-0.3250251114,-0.2092002928,0.0039791819,0.0233761054,0.5663744807,-0.2850210667,0.2381998748,-0.0070042107,0.0657345131,0.5314764977,-0.2201392949,-0.108461909,0.3468538821,-0.1492508948,0.0518519506,0.1439966857,0.039588891,0.0526149273,-0.1308285594,0.1937210113,0.6393967867,-0.2344771177,-0.0323568322,-0.0442832783,-0.2477199435,0.4179079533,0.2243718803,-0.0949275196,0.0199594758,-0.3340435326,-0.0949614346,0.1253610998,-0.1228609085,0.2665160596,0.0090401275,0.0167940482,-0.0821830109,-0.0557357036,-0.0020878376,-0.5716664195,0.0258828644,-0.1563588679,0.0483758412,0.0367052443,0.0840058327,0.0246533602,0.0242607631,-0.1075538546,-0.1872881055,-0.0309743583,0.132142216,0.0364425257,-0.3403074741,-0.3291967213,0.2118789703,-0.0324937887,0.2425681204,-0.4329964519,0.54313761,0.2083956897,-0.1171969622,-0.0873500779,0.0489445068,-0.0986335203,-0.0070063639,0.001949327,0.1942148209,-0.0757218078,0.2442175895,0.1699133962,0.065474458,0.1595394015,-0.3383493125,0.3486060202,0.2228185982,-0.0673015639,-0.1175182685,-0.2634740174,0.453484565,0.0424078889,0.4329811037,-0.1104202271,0.141562596,0.2285179049,-0.1195356101,-0.2843779027,0.0755167827,0.038110964,-0.198992908,0.2602618635,-0.0758707374,-0.5680077672,0.1407360286,0.2386331111,-0.0199066997,-0.0573452674,0.1785087138,-0.1155240834,-0.0487361103,-0.0507612117,0.5043224096,0.3902941644,0.2023362517,0.1330361664,0.0527726822,-0.333237946,-0.0390539505,0.2645007372,0.0731956884,0.572296381,0.1111420766,0.1884936392,-0.2298469096,-0.0260418691,-0.225959748,0.375713408,0.196081847,-0.2328561097,0.3923481405,-0.3859460056,-0.3946233392,-0.6372320652,-0.126499474,-0.2228471041,-0.4076824188,-0.0596623421,0.2941827476,-0.2008382082,0.3044977188,-0.0083969347,0.2533296943,-0.0969766378,-0.0633145422,-0.0650340095,-0.3226173222,-0.0971216485,0.0095733339,0.4663543105,0.0486909263,0.1779366434,0.4626578093,-0.3111791313,0.0040370659,-0.2330508381,0.2138879597,-0.1255089641,0.0886386111,0.1230219081,-0.0334311761,0.3305614293,-0.3943111002,0.0101411119,-0.0132068647,-0.2062790096,-0.0949026421,0.0286064222,0.3486715555,-0.1070190743,-0.309383601,-0.3159711063,-0.3946548104,0.2259314805,0.0516273938,0.2161054462,0.4295203686,0.1762686521,0.0463485681,0.1555684656,-0.0262172334,0.0147960875,-0.0455983505,0.0338850096,-0.3673450053,0.189736262,-0.2843198776,-0.2059624046,0.4252792597,0.1152253449,-0.3176317811,0.3440875411,-0.0434317552,-0.0167123545,-0.2224956304,-0.3622526824,0.0244202372,0.1780630201,0.0452707447,-0.0338765793,0.1468674242,0.226079613,0.1121597514,0.23243922,0.2633699179,0.3287668526,0.1926621348,0.8260392547,0.0851388574,-0.1176849231,-0.0616794378,0.0974354222,-0.272415936,0.0375084952,-0.0940982327,-0.2391368002,-0.0824233815,0.069993794,-0.0274069346,0.0607344955,-0.4748288095,-0.0164991431,0.038460698,-0.0694508702,-0.3480145037,0.4364901781,-0.5182201266,0.0621522963,-0.1140318885,-0.3119466007,-0.2559911907,0.0514924824,-0.2783671319,-0.0520955473,0.1209167242,-0.0720344335,-0.4131587744,0.1993457973,-0.364397049,0.2030837685,-0.1106469855,0.2612341046,0.0975530446,-0.2445000708,-0.0281413179,-0.2486938834,0.7580448985,-0.0748394728,0.1026746035,0.0970802605,0.1042810604,-0.5244817138,0.0219222046,-0.2186834961,0.4839394093,0.3086551428,0.3903082609,-0.3387852907,-0.2800493836,0.1992188245,-0.1463635713,-0.2613398433,-0.0154221896,0.0275570564,-0.1254511923,-0.4156197906,-0.0843992233,0.0651170239,0.0581829324,-0.1375562549,0.0462692752,-0.1992900819,-0.1609227955,0.0917416066,-0.015353323,-0.1788317412,0.0135653485,0.2081344128,-0.1128666103,0.221439898,0.1589300781,0.600450933,0.2492440492,-0.3162969947,-0.0304803289,0.1683489531,0.2941430807,0.0430310406,0.1952896118,0.3621751964,0.3023662567,-0.0650164858,-0.1220981106,0.1000132188,-0.0910298526,0.083595328,-0.4143509269,-0.3978821039,0.4033086598,0.2466863543,-0.0018950383,0.4153799117,-0.3548695147,-0.3739944696,0.2413821071,0.2081928849,0.8665277362,-0.0116072642,0.1625175029,0.4749668837,-0.5066372752,0.7132511735,-0.4881280363,0.0508987829,-0.1740199029,0.2263568044,-0.038675148,-0.1600791365,0.0653526038,0.5145103335,-0.1249236539,0.1910086721,0.211553216,-0.4104621112,-0.0267951805,0.4065257609,-0.00349478,-0.263939172,-0.2312739938,-0.0425962731,-0.4342306852,-0.0986037329,0.1534000933,-0.3186725378,-0.2109117657,-0.1321133673,-0.1447932422,0.1191678271,0.1216173768,0.2122847736,0.2192089409,-0.4170315862,0.2105443329,-0.1337071359,0.0080593107,-0.2745546401,-0.0473288633,-0.0342733674,-0.2372593582,0.0514215566,-0.0789100826,-0.1207421646,0.3871484399,0.0204416849,-0.1895026863,-0.0873415098,0.0075708129,-0.3587487638,0.1281233579,-0.0432041548,-0.0922154039,0.0538496785,-0.2129109353,0.3441676795,-0.3149490654,-0.1751491576,-0.0147760157,0.06271559,-0.1868644953,-0.1603610963,-0.0179924183,-0.1548309028,0.0377542078,0.1175945252,0.2087223232,0.118431665,0.5312075615,0.3709816337,0.215075314,-0.1885426491,-0.0057036127,0.2521185577,-0.4778160751,0.1372616738,-0.2956440747,-0.168781504,0.057558503,-0.0778491572,0.1326597631,0.4425214231,-0.046945598,-0.398165524,-0.2747646272,0.4557522535,-0.1103341728,0.0998242721,-0.1173505262,0.2304960191,-0.0533957742,0.3978671134,-0.1770787537,0.2526502013,-0.0048270612,0.3383489251,0.0565916337,-0.0688425303,-0.0623678565,0.1294478327,-0.0358810276,-0.0560731143,0.0516231246,-0.0848975554,-0.1986640692,0.1135849655,0.0098100072,-0.140191406,0.0715758428,-0.2780890167,-0.0606758334,0.0399383307,-0.1343806833,-0.3121055067,-0.0911653712,0.3229662478,0.0029080191,0.0850504041,-0.1496038586,0.2255050838,0.0032475574,0.117382057,-0.0654402152,0.3353913426,0.1935218871,-0.0999348313,-0.6856938601,-0.0532719865,0.0523254462,0.003072002,0.1092383564,-0.2000803202,-0.1315521151,-0.2898883224,0.2536652088,0.4463513196,-0.073939167,0.1958033442,0.2406324893,0.0633559152,-0.0634961575,0.0882200971,0.4835499525,-0.3166556656,0.0711585358,0.3000304103,-0.1604160219,0.1473093331,-0.3009879887,0.05271836,0.0822551474,-0.1254575551,0.3439559042,0.3214915693,-0.3618203402,-0.040101327,0.1614882797,-0.1193910539,0.4025032222,0.0484440997,0.2046440095,0.3204606771,0.2750099003,-0.0557660423,0.4221299589,-0.3098084033,-0.3592763543,-0.0007552673,0.0358414128,0.2483557165,0.2836802602,0.592163384,0.1722432822,-0.2773671448,0.0916704535,0.5221629739,-0.0121864658,-0.2795131505,-0.6138976812,0.0627858043,-0.2234600782,0.0579088554,-0.0424851067,-0.2430527508,0.2019950598,0.1961255521,0.046090316,-0.4166122079,0.303342253,0.1623995751,-0.0245267898,-0.2010998577,0.2490916699,0.2544416785,-0.128400594,0.070522055,0.034829136,0.3172624707,0.0864619911,-0.3684066534,-0.0089643188,0.13968876,-0.0774103925,-0.1733276546,0.1853989065,0.265421927,0.0629155263,0.3606309891,0.0076661156,-0.0907965824,0.0281162299,0.4000339508,0.0957026556,-0.3185197413,0.5138184428,0.1729287207,-0.2144089341,0.3287236392,0.0866153836,-0.4180997312,-0.166408971,0.2632783055,-0.0345898345,0.0029009422,0.1458986849,0.0150297461,-0.0731466189,0.6327582598,0.1750956029,0.2684523463,-0.0302807931,0.0487759858,-0.3237715662,-0.0966863781,-0.2670504451,-0.0193461664,-0.2650863528,0.2112194449,0.3434881866,0.483907938,-0.1652050316,0.0860879123,0.245912686,0.2894406617,-0.1610227376,-0.3069123328,0.2202851027,0.3486307859,0.1889543533,-0.3502087891,-0.0686325952,0.4095515907,-0.1618711054,0.0464933403,-0.0674897581,0.2403895855,0.0328013599,0.0490160733,0.1321929097,0.4877574146,-0.03355553,-0.3763448596,-0.2149562985,-0.1724827737,-0.4667954445,0.2378898114,-0.018556783,0.7020082474,-0.5233489871,0.1436809301,-0.2510558069,-0.3602839708,-0.1742968112,0.0690402463,-0.3635821939,0.2612660527,-0.1166180894,0.1927618533,-0.0447853729,0.4278230071,0.0420891568,0.1251770407,-0.3780818284,0.0384442173,0.3709776402,-0.2145706117,-0.1265936196,0.2050530165,-0.1225412786,0.1562440246,0.1938429326,-0.558031857,-0.0503837243,0.2729419768,-0.2999858856,-0.1132793128,0.1660781205,0.2374661714,-0.1107979715,-0.0310584046,0.2576287389,0.1307673156,-0.2719537318,0.0382674485,-0.3147016168]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/328","title":"Fork dataset","comments":"I'll close this -- It's still unclear how to go about troubleshooting the json example as I mentioned above. If I decide it's worth the trouble, I'll create another issue, or wait for a better support for using nlp for making custom data-loaders.","body":"We have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n","comment_length":43,"text":"Fork dataset\nWe have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. \r\n\r\nWe're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers.\r\n\r\nOur preprocessing flow parses raw text and json with Entity and Relations annotations and creates 2 datasets for training a NER and Relations prediction heads.\r\n\r\nIs there some good way to \"fork\" dataset-\r\n\r\nEG\r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 -> DatasetREL\r\n\r\nor \r\n\r\n1. text + json -> Dataset1\r\n1. Dataset1 -> DatasetNER\r\n1. Dataset1 + DatasetNER -> DatasetREL\r\n\r\n\nI'll close this -- It's still unclear how to go about troubleshooting the json example as I mentioned above. If I decide it's worth the trouble, I'll create another issue, or wait for a better support for using nlp for making custom data-loaders.","embeddings":[-0.225592494,-0.2139752954,0.0137120206,0.1874609441,-0.2855693996,0.1027285755,0.0859784484,0.3090703785,0.2629738152,-0.1347025335,0.0771383941,0.675709486,-0.3767235875,0.130015865,0.1040436774,-0.2021784484,0.1567045003,0.0305683278,0.2432409227,0.1246745661,-0.0030573662,-0.1404227614,0.1567752212,0.0151140513,-0.0611150563,-0.2410472333,-0.1119404659,0.1831781864,0.0436238237,0.0080808066,0.4242900312,0.1682476401,-0.0357528664,0.331484884,-0.0001182203,0.1242470443,0.1082486734,0.1458239555,-0.2584878206,-0.155663982,-0.1127671227,-0.1371055692,0.2131112218,-0.1448480487,0.1367443353,-0.3278983831,0.170842737,-0.1297238171,0.7699853778,0.1334997565,0.1088852584,0.0454335846,-0.1247143596,0.2250990868,-0.2478457242,0.2831077874,0.0789029673,0.2020106614,0.0115350084,-0.1621371806,-0.163832292,0.0203957241,-0.1525958925,0.1096807346,0.1250389367,-0.1221601069,0.0543632098,-0.0273976233,-0.0277142264,0.3535755873,-0.0114999125,-0.1214190722,-0.1666322649,-0.6308935881,0.0714958981,-0.1956183165,-0.0327244774,0.1070844233,-0.1702801138,0.2213978022,0.1279788017,-0.5801433325,0.0357806198,0.1103352532,-0.0623915866,0.3296544254,0.1729614139,0.197650671,0.2552981377,0.0594155826,0.3220237494,-0.2034488916,-0.0187667608,-0.0308435988,-0.1546315551,-0.2607319951,-0.1478810906,-0.626164794,-0.0011969999,0.2987721562,0.3444442451,0.0295497198,-0.3762353063,0.2349631339,0.1856475919,-0.2687038183,-0.1834815145,0.2808970213,0.0147724431,-0.2175779343,0.0500678793,0.1455634981,-0.3928111792,0.0109011577,-0.3534698486,-0.063665919,0.3451400697,0.1989928335,-0.1900560558,-0.402746737,-0.2324145585,-0.1443460584,0.1434203684,0.0017943544,0.0272262767,0.5649724007,-0.0731390789,0.3858673573,-0.213207379,-0.2311062813,0.0232887547,-0.0543792807,-0.1361839622,0.0107970713,0.107192874,0.5068042278,0.151108399,-0.0194066018,-0.3058052361,-0.0769732967,0.4386520684,-0.0546703786,-0.0773444548,-0.0458450578,-0.0622554049,0.0818981826,0.0954199135,-0.0816645473,-0.2606015503,0.21036686,-0.1766934097,-0.1315820366,0.0219493564,0.0671400279,0.0027995305,-0.3125676215,-0.5295146704,0.6395683289,0.1960008442,-0.0690217316,0.0101074744,-0.0039348067,-0.3481339216,-0.2317089736,-0.0104253301,0.0529447421,-0.3890049756,-0.3615546823,-0.2279181182,0.0117453504,-0.0676146746,0.4565347433,-0.2434231192,0.1787769645,0.0341737196,0.0166954305,0.5639643669,-0.1810501516,-0.0599438809,0.2342610359,-0.2342584878,0.0363017432,0.0825459138,0.1439548284,-0.038168285,-0.181454137,0.0570438057,0.7605223656,-0.2868873179,-0.0040725651,0.0645716265,-0.2386103421,0.3192780614,0.2190312594,-0.0627646372,0.0548263229,-0.3312347531,-0.0223843344,0.2203795016,-0.0492089242,0.3052375615,-0.1040297449,-0.1216966212,-0.0869738385,-0.1141555309,-0.1505255848,-0.4700725079,-0.0176507607,0.0689648986,0.0118586682,0.2829796374,-0.0816037804,0.0552300029,-0.0188461933,-0.0217032209,-0.0096874377,-0.0076297028,0.0964447707,0.0328528248,-0.2809767127,-0.37099576,0.1122867689,-0.0472183861,0.2059127092,-0.3807469606,0.4885462224,0.2925759852,0.0159944594,-0.0933833048,0.2137661129,-0.1594698429,0.0417959094,0.1802476794,0.0719100535,-0.2072035372,0.4196845889,0.1911338866,0.1224554256,0.1848059744,-0.3710247874,0.2803329825,0.1075646654,-0.1018229872,-0.1790726483,-0.2775411606,0.3033385277,0.1036186367,0.4381709397,-0.0083155483,0.1538201272,0.147460103,-0.1902009696,-0.244554624,0.1409129351,0.0918480381,-0.1703172177,0.2928293943,-0.0660921335,-0.5334093571,0.2663099468,0.2597095668,-0.0231016744,-0.0067645116,0.0656954944,-0.2278887928,-0.0272636358,-0.0800404102,0.5746644735,0.1669398397,0.2245680243,0.3080439568,0.0128028113,-0.3045589328,-0.0158982351,0.2137684673,0.1295600533,0.5631981492,0.0116662178,0.2122720331,-0.2407271266,-0.0450087413,-0.20996584,0.4044648409,0.0407788157,-0.0848146155,0.365542978,-0.3508611321,-0.4246234596,-0.6468885541,-0.185021773,-0.1552049071,-0.3520293832,-0.0254884977,0.3039807975,-0.1339466721,0.2258616388,-0.012356679,0.2932177782,-0.3522542119,0.0089806654,0.006000774,-0.4163363278,-0.0330904648,0.0651345029,0.5234524608,0.179954946,0.2201524079,0.4802725613,-0.2884744108,0.0228921548,-0.1987533867,0.2183101028,-0.0874336287,-0.0297724493,0.0596472658,-0.1122584566,0.2067912817,-0.2833401263,-0.101821214,-0.0160414036,-0.1360421628,-0.1377626657,-0.0377467163,0.3285039663,-0.0629631132,-0.165023908,-0.3227213323,-0.3119983077,0.350612402,0.036876291,0.1655424982,0.3113000691,0.0797174647,0.0510160699,0.1713953912,-0.0293928422,0.1030354202,-0.06838613,-0.1436235011,-0.277015835,0.2035324425,-0.2974310517,-0.2434927523,0.3983848989,0.128205657,-0.2371731848,0.402256906,-0.0576604716,-0.0482834689,-0.1453525871,-0.3861068785,0.0729031414,0.0853465497,0.0584493913,-0.0381551385,0.2228053659,0.1892830282,0.1013861746,0.1968120039,0.2205891758,0.1887685508,0.1956256628,0.8005090952,0.0264447257,-0.1882139146,-0.1373371035,0.1523606926,-0.3494243026,0.0859080628,-0.0366805531,-0.1492536515,-0.0511701368,0.0791835934,0.0792008489,0.0323397517,-0.3823565543,0.0108708041,-0.0044438266,-0.0840590298,-0.2789991498,0.3753066063,-0.6413503885,-0.0014688548,-0.036607597,-0.3795254827,-0.2813989818,0.025526993,-0.3960128725,-0.0763851553,0.1283549815,-0.0253086165,-0.5550844073,0.2587534189,-0.4043120444,0.1065532118,-0.149710685,0.1082975119,0.13370426,-0.2157505006,-0.0077795987,-0.2512643039,0.7265209556,-0.0621115156,0.0774380341,-0.0218980163,0.2443380654,-0.5533462763,0.1186911911,-0.1643367708,0.4846512675,0.2180559933,0.1855850518,-0.3536521196,-0.2611513138,0.1544137746,-0.1339903027,-0.237119332,0.0346159115,0.0901250541,-0.0753720775,-0.2876888216,-0.0491993576,0.1133300811,-0.0835008845,-0.2009040415,-0.0356262214,-0.2293891013,-0.1023474038,0.1347273141,-0.0209326856,-0.2114559561,0.0026985337,0.19488132,-0.0913314521,0.2166853845,0.1778644472,0.5501309037,0.1080018729,-0.4033303261,-0.0694692582,0.1753069162,0.3590191305,0.0023801096,0.4114625752,0.5049766302,0.4489447474,-0.2014081627,-0.1918383986,0.1467467695,-0.1210090145,0.1619018763,-0.4613928199,-0.3900202215,0.443344593,0.2551623285,-0.0187098645,0.3786338866,-0.292050302,-0.3776566684,0.346784234,0.0968668684,0.7605614066,-0.1275992543,0.2465569377,0.2234187275,-0.4836920798,0.8631589413,-0.4748440087,0.0444693938,-0.099927336,0.1351713389,-0.0582546815,-0.1513382047,-0.0857994258,0.5734060407,-0.0948485956,0.113084875,0.2172960639,-0.3962779641,0.0110708,0.4008001983,0.0025829009,-0.2414869964,-0.123406291,-0.0043781921,-0.381698519,-0.1698631197,0.0784130618,-0.1844027638,-0.1299361438,0.0124698617,-0.137490809,0.1000869125,0.1305066347,0.0557211936,0.1529760063,-0.4156391621,0.1985136718,-0.1558767408,0.1504121125,-0.3637373447,-0.0930848196,0.0381137915,-0.1674919426,0.0735989809,-0.094191581,-0.2420564592,0.4414747953,0.1090821773,-0.2119496465,-0.1252068579,0.0377472416,-0.2322396189,-0.0288273878,-0.0495088212,0.0902471766,0.2352702916,-0.2116117179,0.479379952,-0.3060407043,-0.1758318692,0.0093611032,0.0517830588,-0.0910814032,-0.207832247,-0.036681205,-0.051427722,0.0955743566,0.024005726,0.2492675036,0.0429390781,0.4369225502,0.3884042203,0.2362613082,-0.173646003,-0.1126041785,0.3920320868,-0.4630275071,0.0654894859,-0.3011799753,-0.2409796268,0.1630825847,-0.0893410891,0.0437426493,0.3343807459,-0.1670923382,-0.3785154223,-0.2100020796,0.5031381845,-0.1560942829,0.1168874577,-0.0720458329,0.2750499845,0.0285883546,0.4572605193,-0.1817756742,0.2492102087,0.0724975467,0.3490790427,0.0043703835,-0.0120640034,-0.1322572976,0.1297162771,-0.0616885312,-0.0028908164,0.1225645542,-0.1128114015,-0.1961969137,0.1171422973,0.0778800845,-0.1447005123,0.0897052363,-0.2446609437,0.0302836839,0.0119504835,-0.1140206009,-0.323420763,-0.067685023,0.2282803953,0.0907379836,0.1796731502,-0.2158172876,0.1726088226,0.1036064625,-0.008180758,-0.1344136149,0.3630339503,0.2937322557,-0.0827696621,-0.536431253,-0.063246049,0.2846280336,0.0123414043,-0.0230602231,-0.1194349378,-0.0778796747,-0.2318897396,0.2651407123,0.4332116544,0.0061927368,0.1934037358,0.2808112502,0.1222852692,-0.0499348789,0.0790431872,0.4891517162,-0.3450668752,0.0849794969,0.2614116073,-0.1177534834,0.1229925901,-0.3080476224,0.0012804935,-0.031697236,-0.2097047716,0.3735480011,0.2310183942,-0.2958547771,-0.0831735134,0.1448839158,-0.1254052371,0.3239160776,0.118822135,0.2797049582,0.4925565124,0.3353676796,-0.094581306,0.5092293024,-0.1900788695,-0.332527101,-0.0650516748,0.1259595752,0.2740227878,0.2377756834,0.6080209613,0.2367740422,-0.2535492778,0.096094884,0.4803633392,-0.0080231186,-0.2975946963,-0.6598876715,0.1000537723,-0.128250286,-0.0272627342,-0.0683542863,-0.2648769021,0.2347921878,0.1211791039,0.1713861227,-0.3783199489,0.2370422781,0.1215775087,0.0208646506,-0.0896722451,0.1988062412,0.1996976584,-0.1532538831,0.1174783483,0.0530359931,0.1160582528,-0.0431916937,-0.4731788635,0.0513174385,0.1138667613,-0.0585206784,-0.2365315855,0.1224347576,0.292062223,0.0109610036,0.3345223665,0.0191289019,-0.1169025525,0.0096034911,0.4403293729,-0.0293264035,-0.4878627658,0.4588002563,0.2077514529,-0.2303960472,0.2585619688,0.2308828086,-0.3879238367,-0.1584391445,0.3280764818,-0.2374748439,-0.0505244918,0.1283551604,0.0263886955,-0.0424837321,0.5274134874,0.1472370028,0.3815220892,0.0467581078,-0.0388125516,-0.3185816109,-0.0278855916,-0.2809956968,-0.0226377603,-0.3062281013,0.1478826255,0.3063928187,0.3770582676,-0.2242881656,0.0479024723,0.2318756878,0.3143723607,-0.0799173862,-0.2844037116,0.1349049807,0.2836181223,0.1767903417,-0.2951724231,-0.0581756905,0.5080987215,-0.1928969473,0.0708648264,-0.0207990594,0.2869108021,-0.0296810064,0.0483182855,0.0642770603,0.45362854,-0.0697833076,-0.4069775641,-0.2457739115,-0.132628426,-0.5627467036,0.2836138308,-0.0907640755,0.6159169078,-0.5590016842,0.2587600946,-0.0405212045,-0.5787869692,-0.076295726,0.0816037357,-0.408157289,0.2503029704,-0.2056222558,0.1880093813,0.0762926042,0.5517143011,0.0693137273,0.0199302845,-0.2070097178,0.060447976,0.2192196101,-0.1770373136,-0.0585656799,0.0832618326,-0.0925600305,0.1186421961,0.1360613406,-0.5521398187,-0.1466623396,0.2632931471,-0.3044005334,-0.0315192901,0.1822451651,0.2029423714,-0.1305771619,-0.0826253891,0.3133210242,0.1231720373,-0.3210164011,0.0717638507,-0.3371941447]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/326","title":"Large dataset in Squad2-format","comments":"I'm pretty sure you can get some inspiration from the squad_v2 script. It looks like the dataset is quite big so it will take some time for the users to generate it, but it should be reasonable.\r\n\r\nAlso you are saying that you are still making the dataset grow in size right ?\r\nIt's probably good practice to let the users do their training\/evaluations with the exact same version of the dataset.\r\nWe allow for each dataset to specify a version (ex: 1.0.0) and increment this number every time there are new samples in the dataset for example. Does it look like a good solution for you ? Or would you rather have one final version with the full dataset ?","body":"At the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?","comment_length":121,"text":"Large dataset in Squad2-format\nAt the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?\nI'm pretty sure you can get some inspiration from the squad_v2 script. It looks like the dataset is quite big so it will take some time for the users to generate it, but it should be reasonable.\r\n\r\nAlso you are saying that you are still making the dataset grow in size right ?\r\nIt's probably good practice to let the users do their training\/evaluations with the exact same version of the dataset.\r\nWe allow for each dataset to specify a version (ex: 1.0.0) and increment this number every time there are new samples in the dataset for example. Does it look like a good solution for you ? Or would you rather have one final version with the full dataset ?","embeddings":[-0.0997780412,-0.2414989769,-0.1118135303,0.1810322106,0.1970849931,-0.0152267907,0.0232527517,0.4885350168,-0.0912805051,0.1816319078,-0.2126611471,0.0559425987,-0.2452746332,0.3290416896,0.2647447884,-0.0670856759,0.0118722217,-0.0230084192,-0.1232943535,0.0525797717,0.0917730629,-0.1653291434,0.0919112861,-0.1205425188,-0.2654374838,-0.1660659611,-0.218966648,0.0960968584,-0.3439052105,-0.4318127632,-0.1703834683,0.2185877264,0.3581572771,0.6168690324,-0.0001192068,-0.2639674544,-0.1069821045,-0.3270840645,-0.3645772338,0.0406492725,-0.0091303037,-0.3253715634,-0.1189130396,-0.1130117327,-0.1868093908,-0.2391312569,0.1240132153,-0.2879826128,0.8988136649,-0.2257035673,0.0568251647,-0.0995194614,-0.1425751299,0.0200350657,0.1949758828,0.264967978,-0.0998327062,0.0275652818,0.7631838918,0.2729378045,-0.1379783005,0.1158566996,0.1028813273,-0.2239881754,0.0953689665,-0.2059375197,-0.0160013381,-0.3111697435,-0.0992633924,0.2254624814,0.4971261919,-0.231215775,-0.2989939153,-0.3494286239,0.0848995298,0.0883637816,-0.0953715444,0.3142319918,-0.0064442726,0.3016768992,-0.3257999122,-0.4009292126,-0.4019841254,-0.0720786899,0.2573390007,-0.1161980927,-0.0808769986,0.0244955681,-0.1471570432,-0.219811812,-0.0584718958,-0.1525382102,0.0715532452,0.2345580757,-0.0675154701,-0.3732780516,-0.1337952465,-0.1824221909,0.6460013986,0.1989931762,0.1953395307,-0.0929704607,0.1765871197,-0.086455822,0.3882081509,-0.0132769868,0.0069345436,-0.3535681069,0.1602530628,0.0566458814,0.015451747,-0.0166388899,-0.3482196331,0.1194508523,-0.0780948922,-0.1198165491,0.201299727,0.1279423088,0.1323255002,0.1488922238,-0.1224228665,0.0335170478,0.1445465982,0.2388262749,-0.1281583011,0.1534231901,-0.0863457993,0.0834686011,0.1609511524,-0.2093755305,0.0299157258,-0.0308845695,-0.105106391,0.330355376,0.2067345232,-0.0710580349,0.0653580055,0.0247822478,0.3602226377,-0.4267211854,0.1748328358,0.1002815887,0.0175224412,-0.1147703007,0.1647482067,0.0815888867,-0.2149352431,-0.1712434739,-0.2071367651,0.1294359118,-0.1964801997,-0.2423818558,-0.0544869825,0.0102973459,-0.3162892461,0.2142867744,0.0043270453,0.3083494604,0.0286607444,-0.026775999,-0.0905512422,0.2809029222,-0.2474737912,-0.2410406768,0.1092095897,0.6122671962,-0.2852084339,-0.0807601064,0.6636146307,0.0836110115,-0.5348961353,0.3374616504,-0.0009944156,0.1033220664,0.10663376,-0.2227986902,0.3268621564,-0.4240097106,-0.3810765743,0.4306105673,-0.2452858537,-0.1893838793,0.3636598587,-0.0155090056,0.728866756,-0.2947201133,-0.0227383524,0.3314570487,-0.0771267861,0.2354422808,-0.3025572896,-0.3982694745,-0.2546396852,0.2839325368,-0.2140710056,-0.1160830259,-0.1799302697,0.2502491474,0.2469008416,-0.1099305227,0.3165059686,-0.1018164009,0.3639549315,0.0839579701,-0.0795993358,-0.0867857933,-0.7564765811,-0.1431139708,-0.0840265155,-0.3280597031,0.186029613,-0.4696921706,-0.3212245703,-0.09506163,-0.071246326,-0.0954309925,-0.0386249088,-0.1620655358,0.2539693713,-0.079684414,-0.3652379811,0.1896637976,-0.0390120111,0.111178264,-0.1952002496,0.1073803604,0.113776505,-0.0149338366,0.2077325881,0.2578850091,0.0083844624,-0.3317626715,0.3545594215,0.1319757849,-0.1336446255,0.0812666565,0.1226986796,0.1317264885,0.3408952057,0.2437700927,0.3483716249,-0.0439622961,-0.0989810452,-0.1442310661,-0.5008667111,0.3706684113,-0.2921047509,0.1103616059,-0.1532380879,-0.1345479339,0.1723105758,0.069111295,-0.2613067031,0.0681070685,-0.1646813601,0.1101450771,0.344286263,0.0879935175,-0.0871090293,0.133363232,0.3919232488,-0.2169865966,0.0179702621,0.0516892448,0.1255054474,-0.0312472712,-0.0564965904,-0.2630293667,0.1382662058,0.1746782064,-0.128372401,0.0647285059,0.4244443774,-0.0671781078,-0.046236176,0.0543141514,0.2298889905,0.2028607279,0.3700594008,-0.1126025617,0.1116563529,0.0417751372,0.1205795705,-0.056896247,-0.1817296743,0.0054915077,0.1262895018,0.2495287359,-0.2198900133,0.235609889,-0.2114339471,0.2134434432,0.2909429967,0.1771951169,0.1042567864,0.0839726403,0.2970293164,0.8125568032,0.0710540861,-0.2512123883,0.1529860348,-0.2007253617,0.3018532395,-0.0371446386,0.3549658954,-0.4794067144,0.2100296468,-0.1089795157,0.0352245346,-0.2428153306,-0.1994035393,0.1299562752,-0.0244705826,0.1560817361,-0.1350118369,0.2721752524,-0.2808153331,-0.0307091903,0.1175482273,-0.0479569733,-0.1901573241,-0.2650650144,0.019728668,-0.0658896416,0.0830077156,-0.3270969987,-0.279815346,-0.2791698277,0.5616511106,0.1398451924,0.0851490423,-0.3313442171,0.17753461,-0.1528159827,-0.4699180126,-0.0373300873,-0.1394499689,-0.6506544948,0.1868302375,-0.2666991353,-0.2523920834,-0.197853446,0.094690986,-0.1084744781,-0.0140739512,-0.2855157256,-0.3582021594,0.1891213208,0.0941210687,-0.109618254,0.2015245706,0.3242135346,-0.2208324224,0.1413422376,0.0029685134,-0.1006147265,0.2027544826,0.2374801189,0.3559163511,-0.1488821059,0.3959993422,0.1915974766,0.8294717669,0.6231415272,0.1226647869,-0.192465961,0.1587837636,0.2736447155,0.0276941471,0.0053977608,0.4714065194,0.0324955806,-0.032116998,0.2436349988,0.4521153867,0.1945264637,-0.0915699899,-0.1980015486,-0.1404840946,-0.2652004063,0.3011285067,0.3607007563,0.1096629798,-0.093091242,-0.1706572473,-0.1621377468,-0.2697302699,-0.1159680337,0.1859961599,0.3794797063,0.2098849863,-0.574067235,0.3515959978,-0.2495236397,0.1586683691,0.1407624185,-0.1898964047,0.1042380333,-0.080692783,0.1455284804,-0.0503932089,0.6973086596,0.2146672308,-0.1506148726,-0.2447744161,-0.0417027585,-0.3592642248,0.1269333065,-0.1086162552,0.445112884,0.4036061168,0.3771094978,-0.4491505623,-0.2118763626,0.12485376,0.128811568,-0.2049937099,-0.1730587929,-0.1876726896,-0.1642362028,-0.3772872388,0.2050733417,0.1352116913,-0.0898914486,0.1681875437,-0.0176595245,0.0716029033,0.0904713646,0.0740086883,0.2989824116,0.0289299153,0.2948360741,0.1590932161,0.25293383,0.157222718,0.050047487,0.2393588871,-0.0101681957,0.0080422303,0.2292132378,-0.1190328002,0.5664660335,0.5696610808,0.1238998994,-0.1912186295,-0.4843904674,0.2945775986,-0.3479851484,-0.1847197413,0.1818924397,-0.1415814161,-0.2552084625,-0.2871849835,0.4790969491,-0.0409049802,-0.0847551972,-0.0073112431,0.2140118927,0.0477845334,0.3040321767,0.0754184425,1.2021838427,-0.3742401004,0.4124627709,0.0296174642,0.0230721794,0.5967461467,-0.6018073559,0.0579553209,-0.4674832821,0.3419056237,0.0099589257,-0.1942117214,0.0003660472,0.425829798,-0.3690877259,0.1852419823,-0.0992699116,0.4017491937,-0.2606285512,0.6654484868,-0.1651355922,-0.2708646059,-0.5053533316,0.026853662,-0.1865068972,-0.1574734747,-0.1243917421,-0.1319155842,-0.1164253801,0.0982790068,0.0090546301,-0.2376277745,-0.1858522147,0.0519593284,0.2512977123,-0.4156517982,0.3881489336,0.2541199625,0.038778387,0.217483744,-0.0712950304,0.1556489319,-0.2833273709,-0.1055983827,0.0734278113,0.0579539053,0.1931797564,-0.047286924,-0.2664129436,0.1490267664,0.279138118,0.0264125653,-0.5277001262,0.1213886961,0.3254843056,-0.6069250703,-0.1746605784,0.0657310411,-0.1984012723,0.0847045481,-0.0115274042,0.1129429266,-0.072688289,0.3060582876,0.5013999939,-0.1398217529,0.0687152743,0.1370402724,-0.1541064978,-0.2838394642,0.1902315319,0.0829016492,-0.045996774,-0.1814293265,0.7017863393,-0.0522638857,-0.1654691845,0.0023640755,-0.0472078398,-0.2197894603,-0.2392238826,-0.1758710295,0.3445545733,-0.0046082903,0.3115481138,-0.1323165894,-0.3046638966,0.2430802733,0.1918094754,-0.089282833,0.1325090975,0.0541319884,-0.0426912419,0.3417143226,-0.207131356,0.1975441724,0.0837184191,0.0973502398,-0.0291046947,-0.2228957415,0.0407824777,-0.1490175128,-0.1275978684,0.0000936241,-0.4162393212,-0.108900778,-0.1852497011,0.2129579782,-0.004815462,0.0500317588,-0.1606012285,-0.2276467681,-0.0708597824,-0.0704790652,0.2910270691,0.11542961,0.3595733941,-0.2725123763,-0.1087475196,-0.2252641767,-0.0278413277,0.1439148635,0.2232496589,0.3191729188,0.1773954034,0.1701991409,-0.0306842756,-0.1751803607,-0.0639691502,0.4981127977,0.5956969857,0.1428248733,0.259172529,-0.0916399509,0.0654430166,0.1437786818,0.1729608029,0.4379863739,-0.2449678779,0.1943517327,0.0463157557,0.067883797,0.0662233084,0.1516551226,0.4133435488,0.0555792116,-0.0465088896,0.4429084659,-0.0431636311,-0.0476899259,-0.1148670465,0.1415475905,0.5450333953,0.1170207039,-0.123279497,-0.062059138,-0.0437409729,-0.2136630118,0.4176063836,0.2818903625,0.148614198,-0.07480748,-0.1218436733,0.1668071896,-0.0911474898,-0.0005370748,0.2861866355,0.05708386,-0.2213194072,0.3472097516,0.3019327819,-0.0726384446,-0.0794752985,0.7899215221,0.2486392409,-0.302010417,-0.1370299906,0.1681548357,0.0844282508,-0.1639673114,-0.1675735861,-0.1863093376,-0.2095090598,0.0558713563,-0.2976074517,-0.5191299915,0.212048009,0.2914510071,0.3409846723,-0.1082467809,0.1394076049,0.1057222337,0.3179252148,-0.2273094356,0.287184,0.198670879,0.2035995573,0.1773162931,0.196261242,0.1064812467,-0.1156354025,-0.2542619109,0.3833452761,-0.3681040406,0.0694546476,0.10002698,0.0070876069,0.1284895092,-0.3347006738,0.2503018975,-0.0142431557,-0.006766526,-0.1343007833,0.1866585016,-0.4374960959,-0.2581464052,-0.2774604857,-0.1609300524,-0.2395208627,-0.1330559701,-0.0723278597,-0.4454123974,0.0130424555,0.1621605754,-0.1418965012,0.2449498028,-0.0729923919,0.0302721485,-0.0948017538,0.2382570058,0.0311881155,-0.1155207828,0.009239669,-0.117547594,-0.2322068363,0.0562544689,-0.189939633,0.1283693314,-0.1862628013,0.5108602047,0.1144768968,0.5743860602,-0.1525535583,-0.0378384963,-0.414138943,-0.1009266153,-0.4872943759,0.0679348856,0.0629954562,0.57944417,0.2389270067,-0.0967685878,0.2277923673,0.0356472358,-0.182469964,-0.045581501,0.136337921,-0.3447841704,-0.5008977056,0.0660584867,-0.0611398853,0.0389656574,0.0871229544,-0.2023596168,-0.2704786658,-0.0451458469,-0.1055658162,0.2444244921,0.2704154551,0.1172907054,-0.3241795301,-0.3725755513,0.0234501623,0.0022231492,-0.1681125164,-0.1866528988,-0.184044838,-0.2977952361,-0.0467762351,0.2462061048,-0.0516110584,0.500359714,-0.4273846447,0.1211485043,-0.1787669659,-0.0257166885,0.6111912131,-0.1560891271,-0.225562185,0.0675891712,0.3299677372,0.2188187838,-0.1582418084,-0.2939794064,0.3241057396,0.1196348667,-0.0128143802,-0.1734105647,0.1478224993,-0.3205628991,-0.2897049785,-0.14446114,0.2264078856,-0.0070546842,-0.0719314888,-0.1444687247,-0.2662914097]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/326","title":"Large dataset in Squad2-format","comments":"It would also be good if there is any possibility for versioning, I think this way is much better than the dynamic way.\nIf you mean that part to put the tiles into one is the generation it would take up to 15-20 minutes on home computer hardware.\nAre there any compression or optimization algorithms while generating the dataset ?\nOtherwise the hardware limit is around 32 GB ram at the moment.\nIf everything works well we will add some more gigabytes of data in future what would make it pretty memory costly.","body":"At the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?","comment_length":93,"text":"Large dataset in Squad2-format\nAt the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?\nIt would also be good if there is any possibility for versioning, I think this way is much better than the dynamic way.\nIf you mean that part to put the tiles into one is the generation it would take up to 15-20 minutes on home computer hardware.\nAre there any compression or optimization algorithms while generating the dataset ?\nOtherwise the hardware limit is around 32 GB ram at the moment.\nIf everything works well we will add some more gigabytes of data in future what would make it pretty memory costly.","embeddings":[-0.0783762857,-0.3369032741,-0.1021032855,0.3794317544,0.2913861275,-0.0967132673,-0.1217115596,0.5122983456,-0.0272251945,0.1946527064,-0.2856161296,0.0614134073,-0.2502669692,0.2593443692,0.3205595911,-0.0886087418,0.1549791992,0.0823775753,-0.2040606886,0.0569107495,0.0637137368,-0.1032855734,0.0612261407,-0.1653318405,-0.3009232581,-0.1412977427,-0.288174212,0.113891311,-0.3934970498,-0.3375560343,-0.2448669225,0.2171664685,0.1611695588,0.647347033,-0.0001227837,-0.1961065531,-0.1437792182,-0.3495814502,-0.2894806564,0.1330737621,-0.0990354195,-0.3333496749,-0.1614259183,-0.1127134338,-0.1404915005,-0.2584019601,0.1062348038,-0.3380098045,0.7726531625,-0.185261786,0.0068999734,-0.1393992007,-0.0483219214,0.0074566915,0.2478711307,0.2530166507,-0.1029361188,0.0350588039,0.624014616,0.3569599688,-0.0175891798,0.0792302564,0.0387504995,-0.1803491265,0.0834922567,-0.1305760592,-0.0905045271,-0.3459359109,-0.0884473249,0.2567327321,0.4648344815,-0.2146343887,-0.2226941884,-0.1702958196,-0.0250918809,0.1221155077,0.0624894463,0.412058264,-0.1602798402,0.3520800769,-0.3435668647,-0.4413897693,-0.3995464742,-0.0935547352,0.2771208286,-0.0322672129,-0.1175544858,-0.0084751481,-0.1265733689,-0.1373977065,-0.1024049148,-0.1397244185,0.0183658656,0.1883071661,-0.0581437014,-0.3634128869,-0.3048736751,-0.2218268216,0.7251416445,0.2197364718,0.3265017271,-0.0457135327,0.3206260204,-0.0938945189,0.2684270144,-0.0021524255,-0.2191832513,-0.4172424376,0.1518899649,-0.0420654975,0.005759927,-0.0518738441,-0.2956549227,0.1691857874,-0.1634948254,-0.2331241816,0.2861676514,0.1393417269,0.0892824307,-0.0214983411,-0.2318848521,0.0085954964,0.0988225043,0.2212495953,-0.1391449124,0.1223293319,-0.1243348643,-0.034625981,0.2165811956,-0.2578242123,0.0529310964,0.0112324674,-0.071440123,0.347055316,0.2074620873,-0.0121732913,0.0128071811,0.0960285291,0.3481047153,-0.363953799,0.1384862661,-0.0333423615,0.104071565,0.0320234001,0.1444004476,0.0015518777,-0.2644882202,-0.0145273749,-0.297685653,0.0540181771,-0.2035525143,-0.3303642571,-0.0882694274,-0.0250275768,-0.1353702098,0.3116150796,-0.1619094014,0.3599739969,0.0874314308,0.0490743369,-0.113785997,0.4489847124,-0.2182917148,-0.2930988073,0.1602414846,0.5045282245,-0.2007538974,0.0229594186,0.6979801655,0.0560174808,-0.2464072555,0.4481517673,-0.0043962421,0.1359654069,0.0660853833,-0.1463229507,0.2523387671,-0.4034131169,-0.393864423,0.5243748426,-0.1237470433,-0.2651093602,0.5103117228,0.0308409072,0.7380927801,-0.2041797936,0.006787654,0.4657239318,-0.0756600201,0.2678143084,-0.2925254703,-0.3666589856,-0.1064671427,0.1976788044,-0.3410522044,-0.1301359832,-0.1742994934,0.2694671154,0.085177891,-0.1051846445,0.3670702875,-0.1038185284,0.3626278341,0.0368106104,-0.0351086482,-0.0798206329,-0.706956923,-0.1754608452,-0.147510618,-0.1815338433,0.1756395847,-0.4423541129,-0.2127514631,-0.0871234462,-0.0436072201,-0.0111228097,-0.0349128731,-0.1196978018,0.2897192836,-0.2063937038,-0.4174753428,0.3462837338,-0.0044269282,0.0942240283,-0.0401619785,0.1176773384,0.0820627958,-0.0380719863,0.2208338231,0.1245052367,-0.0227843635,-0.3221158385,0.2823005915,0.1625126451,-0.1242315695,0.051511649,0.1698478609,0.0649467111,0.3528203368,0.184691295,0.4536722302,0.0004359923,-0.0661331266,-0.1742687076,-0.5200099945,0.248151958,-0.3177943826,0.1537080109,-0.0739712864,-0.1496352255,0.2833902538,0.02389233,-0.1768892854,0.0129621392,-0.0836063623,0.1472509205,0.3329172432,0.2726940811,-0.1475754976,0.1436816901,0.3839886189,-0.1751174331,0.0854555815,0.1256877333,0.1638944149,-0.28808406,-0.0547322556,-0.2870210111,0.0964189395,0.2132014334,-0.1603959799,0.0219216496,0.3833084106,-0.0723183453,-0.0162212215,0.035926085,0.2771426141,0.238175869,0.3160497248,-0.0604221039,0.0092027159,-0.0110462792,0.2079131007,-0.0585529953,-0.0915654078,-0.0322172232,0.0701841116,0.2753458023,-0.076545231,0.2308813334,-0.2097187191,0.2561206818,0.2956341505,0.052866485,0.1372467726,0.0791646242,0.4016389549,0.7862680554,0.0490032993,-0.1759125739,0.03916144,-0.1600472331,0.3815576136,-0.0766045973,0.3648779094,-0.234054327,0.2852953672,-0.0613864325,-0.018728612,-0.2080342174,-0.0753421038,0.2011942565,0.014133635,0.166196242,-0.1607528925,0.2517768145,-0.2142984569,-0.003629582,0.1522033215,0.100092046,-0.3120241463,-0.1916678995,-0.1176366061,-0.150331527,0.0887132958,-0.2238148153,-0.2860088348,-0.3231348395,0.5092948079,0.0681103617,0.1934662908,-0.4624144137,0.1838062853,-0.0821284205,-0.4521048963,0.0322987325,-0.1644722223,-0.6774635315,0.2262754291,-0.2423985749,-0.2435352802,-0.3196680546,0.1436239034,-0.0644794106,0.1405666769,-0.1824054569,-0.186115995,0.0617913976,0.0786859319,-0.0438218005,0.1198416352,0.4211664796,-0.1456255913,0.0989599824,-0.0075257407,-0.0118148373,0.1747241765,0.2913619876,0.213571161,0.0852979869,0.45124951,0.1385156065,0.7558050156,0.6317539215,0.2200107872,0.0183603205,0.1004882902,0.2332389951,-0.0699919164,0.0299024284,0.4696426392,0.0082808286,0.0165152736,0.2120725214,0.3998869658,0.168670252,-0.1937046647,-0.2370015681,-0.1864973009,-0.1431632489,0.324611634,0.3430022895,0.1465075761,-0.1356237382,-0.3484300077,-0.1745059639,-0.2955263555,-0.1150360182,0.1036045551,0.3208520114,0.1508605629,-0.5605376959,0.2728163004,-0.387901336,0.0726884082,0.1514025033,-0.2291636318,0.0294449776,0.0330013633,0.2058227509,-0.1313835979,0.6935654283,0.2362894416,-0.2052285075,-0.2848107219,-0.0359155945,-0.2962574959,0.2041636407,-0.1618206799,0.5082671642,0.4002027214,0.4520650208,-0.5513502359,-0.2969151139,0.0719205067,0.1676055193,-0.1722218394,-0.1284497529,-0.1660272032,-0.2056511492,-0.4376517832,0.0472209193,0.2290110439,-0.0335298665,0.3272571862,-0.237556085,0.0065540145,0.1567453444,0.1071840599,0.20760414,0.0108358758,0.0887902379,0.1709646136,0.1252224594,0.1800397933,0.0761865303,0.236060828,-0.008259167,0.1217346713,0.174582094,-0.0739689544,0.5832386613,0.5046699643,0.1077761129,-0.2305302471,-0.4071765542,0.3388228416,-0.4252403975,-0.205082804,0.1855164468,-0.1453351229,-0.3102443814,-0.1364741772,0.506379962,0.0255233049,-0.0999761298,0.0701604858,0.3052079678,-0.0552936681,0.3334284127,0.2119613588,1.3463050127,-0.4943513274,0.3390041888,0.0765283406,0.0472708046,0.5080577731,-0.6566591263,0.101321049,-0.3028968275,0.403685689,-0.0521732084,-0.1514641941,0.009356793,0.4282864928,-0.4516573548,0.0899517164,-0.0964669138,0.5072119236,-0.2389473468,0.6394389868,-0.1891430914,-0.4351724684,-0.5924040079,-0.0121851042,-0.2395348996,-0.2794212103,-0.0270986371,-0.1196104065,-0.1436289847,0.1217974648,-0.0202174485,-0.1796540022,-0.1806991696,0.15668194,0.2039212584,-0.3671700954,0.2174374312,0.1902770996,0.0093956795,0.1367942095,-0.0319665037,0.0622283295,-0.2395299822,-0.278228879,-0.0029730394,0.0376254581,0.0989116579,-0.1177292541,-0.2227041274,0.0241677333,0.2037927508,-0.0684641823,-0.3939576745,0.0621111654,0.2243216187,-0.5619941354,-0.1515211761,0.0598465055,-0.2130888253,0.0352248698,-0.0463196859,0.1275071204,-0.1143166944,0.3470391929,0.4972392023,-0.0826011077,0.0394032598,0.1234395877,0.0842721835,-0.2444636971,0.0904162601,0.0175518692,-0.1470633596,-0.1831600815,0.6808387637,-0.1280508488,-0.1843752861,0.1817666143,-0.109754853,-0.2629247606,-0.2936284244,-0.106205225,0.3944666386,0.0346633494,0.2410103679,-0.1528388709,-0.3036343753,0.3529018164,0.1658291817,-0.064214088,0.0199692827,-0.0322629772,-0.0536767244,0.4111748934,-0.1823551208,0.1870317012,-0.0284987576,0.10009747,0.0070200963,-0.2458523214,0.0349533558,-0.0541424975,-0.1004567221,-0.0587206855,-0.3416643143,-0.0649622753,-0.1126522496,0.2282746136,0.0350276828,-0.0051635704,-0.1717702001,-0.2853633761,-0.2156925052,-0.1201679334,0.2445326895,0.1376601309,0.2847688794,-0.4812834263,-0.0700335503,-0.2733211517,-0.0208819713,0.1764972657,0.2403638363,0.4069962502,0.0657456964,0.1164228544,0.025741186,-0.2583557665,-0.1290076524,0.5130090714,0.6452715993,-0.0499842875,0.2964611948,-0.1205539256,-0.044970423,0.1943706274,0.0887431055,0.3183452189,-0.2586114705,0.2076938152,-0.0824153796,0.0033087481,-0.0045900098,0.0779855847,0.4187324345,-0.0612084195,-0.0862048417,0.3684048057,0.0509825163,-0.1309064627,-0.1961303353,0.0988644958,0.5369048119,0.1139391437,-0.2775096297,-0.1387204826,0.0444371626,-0.1400185227,0.3737178743,0.3407344222,0.2538031638,0.1215060428,-0.0502987988,0.2065762132,-0.0334895849,0.1297723353,0.3383570015,0.1459894776,-0.1616298109,0.5025840998,0.1897349656,-0.0794500411,-0.1083898991,0.7325359583,0.2057285309,-0.223921299,-0.1951461136,0.2171822935,0.0751383603,-0.0824083984,-0.1666920036,-0.2654822767,-0.1495411992,0.1284290105,-0.1550442576,-0.4121450782,0.3357381523,0.3109161556,0.3543123305,-0.0016943153,0.304844141,0.1883201152,0.2788808048,-0.2487300932,0.3172352314,0.2447416782,0.0605279952,0.0113727441,0.279351294,0.1860786825,0.0290334206,-0.3341409266,0.466157645,-0.2991470695,-0.0593523979,0.0884759277,0.0756327435,0.1885124147,-0.2831536233,0.093200013,-0.0624140017,0.015281559,-0.2855525613,0.2223765254,-0.2917849422,0.0057018455,-0.3255158663,-0.2108808905,-0.2124382406,-0.0355385616,-0.0335170627,-0.4937590063,-0.1424138695,0.0535334013,-0.2226224393,0.1521715075,-0.0021527139,0.0078372154,-0.1211436391,0.3348330259,-0.0853761882,-0.0122719221,0.015988728,-0.2567740381,-0.1191204414,0.0258222949,-0.221897006,0.0955955982,-0.1849513799,0.6031038165,0.0857097954,0.6102767587,-0.227463901,0.0018812886,-0.3210496902,-0.0815692469,-0.4122192264,0.1538254172,0.1515185535,0.5324704051,0.2032715529,-0.2592150271,0.4004763067,-0.0103034331,-0.1934349835,-0.1263658106,0.2541647553,-0.2773366868,-0.4143603146,0.0231942479,0.0662839562,0.0857775882,0.0313286334,-0.237930119,-0.0508216657,-0.0020791173,-0.0265584216,0.1445007771,0.453494966,-0.0059606028,-0.3527190089,-0.2872933447,0.1072689816,0.2053967267,-0.2066513747,-0.1218008399,-0.0356567651,-0.3302613199,-0.0969768688,0.1723267734,0.0188210197,0.3120371997,-0.3507348895,-0.0494740158,-0.2220898569,-0.0489442758,0.4450863004,-0.0950232819,-0.262028724,-0.0134486845,0.2686888576,0.1577760279,-0.0673215091,-0.3039141893,0.2149633616,0.0775245875,-0.0427081957,-0.2025122643,0.2626698017,-0.1757562459,-0.2539327741,-0.1284459531,0.1256445795,-0.083100602,-0.1376430839,-0.2562361956,-0.1987886578]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/326","title":"Large dataset in Squad2-format","comments":"15-20 minutes is fine !\r\nAlso there's no RAM limitations as we save to disk every 1000 elements while generating the dataset by default.\r\nAfter generation, the dataset is ready to use with (again) no RAM limitations as we do memory-mapping.","body":"At the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?","comment_length":41,"text":"Large dataset in Squad2-format\nAt the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?\n15-20 minutes is fine !\r\nAlso there's no RAM limitations as we save to disk every 1000 elements while generating the dataset by default.\r\nAfter generation, the dataset is ready to use with (again) no RAM limitations as we do memory-mapping.","embeddings":[-0.1409041733,-0.2805253863,-0.1005571783,0.2925080955,0.2343988717,-0.0667674467,-0.0066274032,0.4827185273,-0.0046411366,0.1012585014,-0.2878576517,0.0170630533,-0.240396753,0.2706007659,0.2778697908,-0.0782357678,0.0859452188,-0.0013715298,-0.1129796654,0.0685329288,0.1218710691,-0.1785973459,0.0677223429,-0.1662177593,-0.3052946329,-0.1456225216,-0.1592085809,0.1375635117,-0.3753189445,-0.373906672,-0.2109333277,0.2612871528,0.2146432996,0.5602203608,-0.0001202043,-0.1613312215,-0.0782385916,-0.3438850343,-0.3753677309,0.114575617,-0.1013673097,-0.3235392272,-0.1401280612,-0.1104314923,-0.2038571388,-0.1932175905,0.1400700957,-0.332315892,0.7989575267,-0.1259247661,0.0317238234,-0.0380418748,-0.0403262377,-0.0242629591,0.242518723,0.2816529274,-0.0963999555,0.0187547877,0.6647381783,0.3354052007,-0.1760163307,0.0907357857,-0.0098588532,-0.1400218606,0.1098083109,-0.1842206717,-0.1575744301,-0.2615042031,-0.0657859743,0.2858432531,0.50301373,-0.2781634629,-0.1896306127,-0.2847802937,-0.0658238083,0.1196045727,0.0054941443,0.3605892956,-0.0875722915,0.3904540241,-0.2851333916,-0.4337237477,-0.3542928696,-0.0511342846,0.2782336473,-0.0421493873,-0.1095748246,0.0358316749,-0.1736291051,-0.1611446738,-0.1055147126,-0.1820171177,-0.0435582623,0.195127368,-0.0500388965,-0.3236088157,-0.122834608,-0.2667889595,0.7373473048,0.2351019531,0.2762565017,-0.0252917502,0.2214840949,-0.055995997,0.3499588966,0.000797126,-0.1983939558,-0.3956947923,0.1951480955,0.0411391705,-0.0147488173,-0.1023097858,-0.2278895229,0.1696128249,-0.1113717034,-0.2389044464,0.2255721837,0.1624681652,0.0749958381,0.0988065898,-0.119023703,-0.0564194247,0.1590619981,0.2800046504,-0.1008720621,0.1065988094,-0.0931565464,0.0426950194,0.2466717809,-0.2234209776,0.0033720608,-0.0549612306,-0.1067692637,0.2609632611,0.2497688085,0.0787077397,0.0633990914,-0.0406716168,0.3254795074,-0.2624429464,0.1954028904,-0.0300896242,0.0556440949,0.0270401016,0.1214144602,0.0723473206,-0.2507173419,-0.037362583,-0.2716755271,0.0664346591,-0.1702026725,-0.3618006408,-0.0369636826,-0.000042417,-0.2681352198,0.2985506952,-0.127226606,0.3165004849,0.0211924613,0.0033346754,-0.0748790801,0.4382675886,-0.2252081931,-0.2618328929,0.0783004463,0.5588118434,-0.1850253046,-0.016268529,0.6108691096,-0.0059924601,-0.3543802798,0.3948062956,0.0075898021,0.1504805088,0.0346252993,-0.1278951168,0.2690083683,-0.4436065853,-0.4289497137,0.5240533948,-0.1946698129,-0.2246601731,0.4402690232,0.0410940126,0.6278269291,-0.242933929,0.0849857032,0.4986977279,-0.022191653,0.301345408,-0.2948740125,-0.3466869593,-0.1763312966,0.3227970302,-0.2370019406,-0.19503887,-0.2147206664,0.2757806182,0.1182711422,-0.0507150628,0.3458012938,-0.0659842789,0.3465012312,-0.0189724527,-0.109103851,-0.0796734467,-0.7441965938,-0.1172305197,-0.0487518944,-0.1769839525,0.1242545471,-0.3950254023,-0.2437802255,-0.0272528436,-0.0753839612,-0.0389587134,-0.0110309022,-0.1466056108,0.2431043833,-0.0624290034,-0.42396608,0.2198781222,0.0312689021,0.108875744,-0.0517491549,0.1104810834,0.1083669439,0.0403883792,0.2250187397,0.1406447887,-0.0452486873,-0.2843993604,0.3054430783,0.1263492703,-0.1156675667,0.0338573977,0.0939443111,0.124429673,0.3205665648,0.2131882906,0.3468077779,-0.0148217175,-0.0581191108,-0.2420777231,-0.5757752657,0.2814452648,-0.3202792704,0.1856947839,-0.1016290709,-0.1921323687,0.3548845351,0.02245868,-0.2167508304,0.0203052033,-0.0939081684,0.1647088826,0.4010157585,0.2314790338,-0.1396001726,0.1183304042,0.4107497931,-0.1698398441,0.085956201,0.1295030415,0.1112831309,-0.2252754122,-0.0007226478,-0.3508841693,0.1351929605,0.1990710944,-0.1331205815,-0.0045663142,0.4250622094,-0.1296106577,-0.0679164976,-0.0029469032,0.2279081345,0.2689410448,0.4185391068,-0.1249468029,0.0409382097,-0.0680257082,0.1883984506,-0.0257454589,-0.0917408541,-0.0187229235,0.1817604154,0.2700441778,0.0131419664,0.244939059,-0.1544848382,0.2557690442,0.3178787231,0.1129627526,0.0277429018,0.0555845946,0.3220249116,0.7086125612,0.1092418134,-0.3408541083,0.0270894449,-0.2112098336,0.4393980801,-0.0549041107,0.4286058545,-0.3180980086,0.2616630197,-0.0354951173,-0.0042358991,-0.1711000651,-0.0890160277,0.2154260129,0.0497117825,0.2052056789,-0.1179560125,0.3070597053,-0.2029962838,-0.0046422826,0.1793079972,0.1094417125,-0.2314153463,-0.1409710199,-0.0624786653,-0.139859125,0.0873501077,-0.1704577655,-0.2106005847,-0.3970167339,0.5515090823,0.0806931257,0.0940168872,-0.4090313613,0.1804582626,-0.1239489764,-0.415743351,0.0509497188,-0.1454322636,-0.7269209623,0.204637289,-0.1891167462,-0.2277164012,-0.3275865018,0.0779181123,-0.0304900706,0.0789903328,-0.3037673831,-0.2016584873,0.0868927017,0.143485263,-0.0471298397,0.0897522792,0.3968957663,-0.1114342362,0.0900504217,0.0583491474,-0.0112755001,0.1375245303,0.2120562494,0.2483747751,-0.0091642449,0.4458053708,0.1821142584,0.7829383612,0.4851695299,0.2196800262,0.0542259477,0.1727016866,0.275483638,-0.1015152708,0.019838972,0.442499578,0.0342320092,-0.0314423963,0.1802765578,0.3919911981,0.1098082587,-0.2056760788,-0.2954733372,-0.1570940167,-0.2163414657,0.3759028316,0.4015352428,0.139361918,-0.0524315387,-0.2696832418,-0.1410923302,-0.2912504375,-0.1208146662,0.1347859502,0.4174838662,0.1133308411,-0.5920678973,0.3075262606,-0.3974340856,0.0312752314,0.032543771,-0.2528342009,-0.0121985655,-0.00514179,0.1667618603,-0.1427384168,0.6802515984,0.2327903509,-0.044901628,-0.1984276921,-0.0244643744,-0.3251370192,0.2406827211,-0.1463400722,0.4969035089,0.4652558267,0.4515276551,-0.5926408172,-0.2258896828,0.0128850061,0.1392452121,-0.1968651116,-0.212809667,-0.1484229416,-0.2182441056,-0.4754295349,0.1036966294,0.2383098006,-0.043960534,0.3554881513,-0.0784144625,-0.01936814,0.1624195278,0.0449253544,0.2307016104,-0.0484963767,0.0834056959,0.2023898065,0.1278285682,0.1726068258,0.1032778472,0.2818635702,0.0337035954,0.0903203115,0.1233193427,-0.0085219648,0.4882116318,0.5183218122,0.143508926,-0.1499529332,-0.451541692,0.3554843068,-0.406318754,-0.1295851171,0.1299921572,-0.1523332149,-0.3432395458,-0.1775955707,0.465143919,-0.0613240227,-0.1151962727,-0.0126695652,0.248480469,-0.0358067229,0.3706608415,0.1062932387,1.2893745899,-0.4414869845,0.3910543621,0.075217545,0.0515571907,0.5166640282,-0.7395325899,0.068565622,-0.3477133811,0.3740773499,0.0072500934,-0.1791781187,-0.0398703739,0.4383758903,-0.3424449563,0.1081087738,-0.1384115815,0.481470257,-0.2609600425,0.7284736037,-0.1351394355,-0.3896680474,-0.5786449909,0.0151581299,-0.2396695763,-0.2095908821,-0.0710889325,-0.1450633109,-0.0771521255,0.0842653662,-0.0611644275,-0.2490497977,-0.1270482689,0.1089506522,0.2076436132,-0.368516326,0.2751114964,0.2031453997,-0.0403284319,0.2738548219,-0.049689956,0.1429559886,-0.2844427228,-0.2363253385,0.0290148836,0.0374846086,0.149535045,-0.0959479883,-0.2634082139,0.0929517671,0.1497385204,-0.0544797704,-0.3979588747,0.1247566119,0.2053522319,-0.5689501762,-0.2819284499,-0.0178323351,-0.2618292868,0.0751783624,-0.025300594,0.1577112079,-0.0735873431,0.2400544882,0.4915740192,-0.090997301,-0.0074916207,0.182062462,-0.0016059802,-0.2429550439,0.1702003926,0.118800506,-0.1442970783,-0.1969185472,0.6800792813,-0.0105392048,-0.1847034693,0.1404966116,-0.0037365044,-0.2887126803,-0.3659242988,-0.2099346071,0.3386257887,-0.0608957149,0.3054252863,-0.1639448404,-0.3691946268,0.3123013973,0.1857190132,-0.0890095532,0.0912581086,-0.0316654965,-0.0656616688,0.3900138438,-0.2072121501,0.2037009001,-0.0131286439,0.0936944708,-0.0155118965,-0.2478198856,0.0741371214,-0.1320341527,-0.073975727,-0.0639434755,-0.3755156696,-0.1119010299,-0.1430509388,0.2250719666,0.0496232994,-0.0759969652,-0.1169423312,-0.3477752805,-0.1839897186,-0.1219804138,0.2752690315,0.1719344407,0.2944331169,-0.4831789732,-0.1036857143,-0.2874085605,-0.0232172068,0.2090702355,0.2285811156,0.4083471596,0.1109557822,0.0994686782,0.0517555885,-0.1989712864,-0.0265359748,0.5146842599,0.6444916725,0.0973289832,0.350682199,-0.1245428249,-0.0084523847,0.1809659302,0.121012114,0.2879489362,-0.2791025937,0.2501028776,-0.0425305441,0.0509493463,-0.0465703644,0.0771989152,0.3802145123,0.0022436453,0.0142667163,0.3488169312,-0.0120910676,-0.0625255629,-0.1603214294,0.1511818618,0.5711217523,0.2238799334,-0.2705689073,-0.1246121973,-0.0449252389,-0.1669439226,0.3977625072,0.3459735513,0.2794704735,0.0859215707,0.0093365973,0.1830564886,-0.0191013236,0.0822952837,0.2479453981,0.0310269911,-0.1983717531,0.4128300846,0.2662302852,-0.0667299852,-0.1031140834,0.6514553428,0.2968407869,-0.1370080411,-0.1608640552,0.1697643101,0.0429072045,-0.143020764,-0.1048239917,-0.2766443193,-0.1550352871,0.1510589123,-0.2287661284,-0.4610011876,0.2058522254,0.2624771595,0.3755548596,-0.0212105196,0.3653208315,0.180487439,0.2209096551,-0.2141548246,0.3091766536,0.2817086875,0.1266262531,0.1267700642,0.201847598,0.1727323234,-0.0063071065,-0.3402145207,0.4396518767,-0.3761489093,-0.0535261035,-0.0502605513,0.1068723649,0.2075186819,-0.2797419727,0.1731987745,-0.0248784982,-0.0364816301,-0.2087846547,0.2289432138,-0.37090078,-0.0412518159,-0.3318804204,-0.2246046811,-0.2519364357,-0.0864920095,-0.1284934729,-0.4553546011,-0.0998796523,0.133851096,-0.1839465201,0.252609998,-0.0441518687,0.0109281186,-0.1301109642,0.3143854141,-0.0782676488,-0.0487818606,-0.0527889393,-0.1989189237,-0.2017159462,0.0232770946,-0.2842140496,0.1540017873,-0.1881955415,0.5003314018,0.1217845678,0.58472085,-0.2400676012,0.0708098784,-0.3883979321,-0.0536916815,-0.3941109478,0.0806145817,0.0448235497,0.5533946157,0.1747605205,-0.2300333977,0.2786175907,0.0650682226,-0.1497405767,-0.1604817212,0.2179172784,-0.319742471,-0.4120720625,0.1225446314,0.0465596803,0.0845955163,0.0053298827,-0.2420692444,-0.1329422146,-0.0373048373,-0.0050690514,0.1964519471,0.3821803629,0.0558502674,-0.3367772698,-0.3499859571,0.0738332719,0.1201986521,-0.2507041991,-0.1465725601,-0.1684758812,-0.3080821335,-0.174803853,0.2735372782,-0.0702432021,0.3493786752,-0.3827394247,-0.0261934306,-0.2011385709,-0.0583673529,0.5014270544,-0.1039119884,-0.2714805901,0.0784859881,0.194337815,0.1541591734,-0.0468975045,-0.356761843,0.2978197932,0.085859552,-0.0381602012,-0.2461090088,0.1744018793,-0.211289838,-0.2963839173,-0.1303201318,0.1746084988,-0.0511534996,-0.1387960315,-0.2719121277,-0.1896782517]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/326","title":"Large dataset in Squad2-format","comments":"Wow, that sounds pretty cool.\nActually I have the problem of running out of memory while tokenization on our local machine.\nThat wouldn't happen again, would it ?","body":"At the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?","comment_length":28,"text":"Large dataset in Squad2-format\nAt the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?\nWow, that sounds pretty cool.\nActually I have the problem of running out of memory while tokenization on our local machine.\nThat wouldn't happen again, would it ?","embeddings":[-0.0048634037,-0.2359244674,-0.0683037639,0.2993221283,0.2564298809,-0.1159560755,0.0191383474,0.4660230875,-0.0981997773,0.1532815844,-0.2599274218,0.0292080212,-0.2432349324,0.2509172261,0.3011697829,-0.0480412319,0.1664877385,-0.0175047852,-0.0354582816,0.0578787178,0.1524417847,-0.1213407144,0.0502193794,-0.1121684983,-0.3433024585,-0.1410203576,-0.2227213234,0.1154344901,-0.382927388,-0.3995330334,-0.3136951029,0.2566596568,0.2417446673,0.5677559376,-0.000122708,-0.2018594295,-0.0883843675,-0.3403649628,-0.311219573,0.0545596182,-0.0006793588,-0.3438971639,-0.1172072738,-0.1090767756,-0.1672704965,-0.2621129453,0.1045962647,-0.287291795,0.8023911715,-0.185851872,0.0164926872,-0.016265979,-0.1020802408,0.0746310577,0.1943787187,0.3218402267,-0.0402478278,-0.0143877249,0.6634316444,0.2563692927,-0.1557704061,0.0979119316,-0.0329955295,-0.1636565626,0.0506472439,-0.1566651314,-0.1782923639,-0.1981963068,-0.0039009722,0.1783846915,0.3981475532,-0.3300261796,-0.1373391002,-0.2405589074,-0.1100009009,0.0567669086,0.0279380679,0.3217486441,-0.0983806923,0.3943783343,-0.3018188775,-0.4261778295,-0.3479032516,-0.0081264824,0.1863970608,0.0747401938,-0.1123981699,0.0162326712,-0.2107531875,-0.2084162682,-0.1750331819,-0.1212716326,-0.056275703,0.1857766509,-0.0294533614,-0.3806006014,-0.1769762039,-0.3344237208,0.7397287488,0.2407940477,0.2041114718,0.0006103968,0.2247766703,-0.0235044863,0.4215138257,0.0339765437,-0.1606846154,-0.2854788601,0.1769555509,-0.0767192245,0.007158454,-0.1267490238,-0.290168494,0.2594229877,-0.1199794859,-0.1985271275,0.2078863233,0.118146643,0.1250086576,0.1459026337,-0.1532063782,-0.0004797382,0.1926344037,0.2919611633,-0.0999253094,0.125199154,-0.1555048525,0.0132864499,0.1941238046,-0.2200011015,0.0151871825,-0.048892241,-0.1426734179,0.2775087953,0.1601999849,0.1904596239,0.0550614037,-0.0313021056,0.3454264402,-0.3316127956,0.1718173176,-0.0120868161,0.0918696374,-0.087102659,0.0966745988,0.1071960628,-0.1860133111,-0.0634552538,-0.280477494,0.0286109857,-0.1590733826,-0.2859141231,-0.0899710357,-0.0131995054,-0.2573435009,0.2753234506,-0.155698657,0.3370774388,0.0433541499,0.0276191849,-0.0760421306,0.4177356064,-0.1996265054,-0.2468789965,0.086254552,0.5264920592,-0.21275267,-0.0589169338,0.7256123424,0.0039113704,-0.2323844284,0.4522058666,0.0341198407,0.0996495932,0.0330233425,0.0171210337,0.1975421309,-0.5084146261,-0.3917007446,0.4872522354,-0.2710965574,-0.2888275087,0.4501453638,-0.0021801847,0.7244879007,-0.1779860854,0.1248246729,0.4527071416,-0.0112068933,0.2866778672,-0.3418312967,-0.3517705798,-0.1432753801,0.2145599872,-0.2476702631,-0.233724311,-0.1678469479,0.3443205953,0.1442345232,-0.0370383672,0.3658793271,-0.1258073151,0.4081765115,0.0815366507,-0.1075155735,-0.045349028,-0.6363104582,-0.1759152859,-0.0668965355,-0.2045734376,0.0875467882,-0.4391719401,-0.2339888364,-0.0397208706,-0.0037445631,-0.0227721911,-0.0105006043,-0.072168909,0.1975910366,0.0416513979,-0.4777804613,0.1027794331,0.0681987926,0.0919448584,0.0677362606,0.1153919026,0.0635745078,-0.027806107,0.1835619509,0.1694440395,0.013266569,-0.276183188,0.3075852394,0.136549443,-0.137611717,0.0771178082,0.0643713921,0.1885312498,0.3677638173,0.2471440583,0.3335402906,-0.0359484702,-0.0795969144,-0.1956493109,-0.4011143148,0.2823254168,-0.2059862465,0.1607313752,-0.1087658256,-0.1630889028,0.3603003025,0.0271318257,-0.1765018851,-0.007229628,-0.1080074459,0.1386487484,0.3167693913,0.2005201578,-0.1099159569,0.1504927576,0.419385761,-0.1662428975,0.1410766393,0.0860968456,0.0657812282,-0.2105067372,0.0731403157,-0.4206044078,0.1161367744,0.1916219145,-0.1124330088,0.0174667053,0.4644149244,-0.1069171578,-0.054423485,0.0615024157,0.1119858995,0.2068028748,0.3783842921,-0.1515436918,0.0632615536,-0.0613132752,0.1851516366,-0.018007718,-0.1255821288,0.0122132264,0.0384273268,0.2140253782,-0.0323152766,0.1993446648,-0.133793667,0.2352688462,0.3411096931,0.1273874789,-0.0041495147,0.0475620069,0.3337056339,0.7932181954,0.0033719291,-0.2434069216,0.0602368973,-0.2733080983,0.3194569349,-0.0889341235,0.4855739772,-0.3599860072,0.3138049543,0.0023291709,-0.0271195248,-0.1954461634,-0.1236829534,0.251401335,0.059504047,0.1228087917,-0.1811424792,0.2243694067,-0.3469936252,-0.0619622394,0.1353172213,0.0441457443,-0.2819411755,-0.1478100568,-0.0207597669,-0.1053292751,0.0789885521,-0.1557479352,-0.2638137043,-0.3474766016,0.56192559,0.0495658964,0.0918339193,-0.4269669056,0.123789601,-0.0336596966,-0.4764581919,-0.0442670435,-0.2064770162,-0.690692544,0.129042387,-0.2044298649,-0.1791326106,-0.2571054697,0.1340753138,-0.0105617484,0.0266334023,-0.2865175307,-0.2164182067,0.173415795,0.0656809136,-0.0573247857,0.1034618542,0.3878521025,-0.1101513952,0.0678931549,0.0896940306,0.0188467521,0.1214547008,0.1467282623,0.2475127131,-0.0135703562,0.4285876751,0.2638728321,0.7901008129,0.3915772438,0.1970973313,0.0834625512,0.1033207402,0.2767242789,-0.0046272958,-0.0110786567,0.4460401535,0.0492831655,-0.0639370307,0.2406007499,0.3872227371,0.1359610856,-0.1608263403,-0.3288183808,-0.1550691128,-0.2619346678,0.3843181431,0.360247016,0.099875696,-0.0669145137,-0.2637202442,-0.1609507352,-0.3486646414,-0.0912175328,0.1258564442,0.3103367984,0.1596029997,-0.6825463772,0.3048129678,-0.3812506795,0.0732673556,0.0437277816,-0.1244026124,-0.0309955813,-0.0406831317,0.1786164045,-0.1328398734,0.6704143286,0.3626085818,-0.1161901727,-0.2579666078,-0.1084468141,-0.2668247819,0.1988123655,-0.1225398257,0.460937202,0.4595271349,0.471809715,-0.5613796115,-0.2062008232,-0.0660793483,0.0868468359,-0.2086280137,-0.1341801584,-0.1570603698,-0.177868858,-0.5508025289,0.1023475602,0.2720913887,-0.0412047729,0.3813020289,-0.1734056324,0.0085241878,0.0298007634,0.0893564299,0.2183732241,-0.0154075958,0.1148341894,0.0835977867,0.1694617718,0.2123957127,0.1255591661,0.1472934484,0.1005322933,0.0518699214,0.1355925947,0.0317407027,0.6307205558,0.5391409397,0.1274539232,-0.0374043733,-0.4805533588,0.3121168613,-0.4224782288,-0.0316593535,0.0885877833,-0.1388895959,-0.3014279902,-0.08810848,0.3730335236,-0.0140165053,-0.1200728714,0.0496170856,0.2515343428,-0.105543904,0.4523071647,0.1032071635,1.392983675,-0.4514721632,0.4316121042,0.0266745873,0.0658509582,0.6378914714,-0.819429636,0.075372301,-0.3553821743,0.4845213294,0.0279831644,-0.1089505404,-0.0716067404,0.3980713189,-0.3112871945,0.0931032673,-0.1793703735,0.5494177938,-0.3038335741,0.6470788717,-0.1158551201,-0.388877362,-0.5553728938,0.0002270192,-0.2226498872,-0.2617736459,-0.1231742054,-0.0763366595,-0.0987590849,0.076595597,-0.0189995579,-0.2811443508,-0.1306397468,0.1832190305,0.1906512231,-0.2936618924,0.1789595336,0.1591617167,0.0245656092,0.3643526435,-0.0403414294,0.0743629932,-0.2797788084,-0.2025123388,0.0273776017,0.0013685566,0.1315407902,-0.09157902,-0.297621876,0.1341985613,0.2777326703,-0.0907610655,-0.5014808178,0.1273467243,0.3642624617,-0.5198207498,-0.2096793354,-0.0355665945,-0.2438249141,0.0729561448,-0.0320861936,0.1741944849,-0.1531539708,0.2272573858,0.5170201063,-0.0452102944,0.0120190913,0.199554503,-0.0853138492,-0.3673037291,0.2237195373,0.061376635,-0.1142330393,-0.1685552448,0.6826539636,0.0097083887,-0.1210052818,0.0746244267,0.0038238165,-0.3500802517,-0.3896212876,-0.1243440658,0.3901152611,-0.041337911,0.2755552828,-0.1811108887,-0.428543359,0.3383806646,0.1108337492,-0.0619493052,0.1698971689,0.1088584214,-0.1757758409,0.3394358158,-0.1865459532,0.1728892773,-0.1168090925,0.10828612,-0.0540545024,-0.2186049819,-0.0072936323,-0.1137067899,-0.0955777913,0.0397076756,-0.2682885528,-0.090334177,-0.1170661673,0.2406668812,0.111438565,-0.0461360067,-0.129383564,-0.2641299069,-0.0955235586,-0.1117089391,0.2225815654,0.2459102422,0.2900201976,-0.425491184,-0.0774688646,-0.2499217838,0.0075125154,0.169882983,0.3039051294,0.4021051526,0.1170350909,0.0978744626,0.0417186692,-0.1809555292,-0.0057735234,0.5137708187,0.7084438801,0.0902363956,0.2484451532,-0.0522909276,-0.0526193716,0.1871719956,-0.0088200308,0.2816484869,-0.2950824201,0.1594905406,0.1631025672,0.038095735,0.0317641422,0.0598829016,0.3637104332,-0.021534944,-0.0839360356,0.3634069562,0.0386961401,-0.0915406048,-0.1826538295,0.1232215986,0.5751820207,0.2144142687,-0.2730650008,-0.2374277413,-0.1275796592,-0.1843101233,0.3792657554,0.2954376042,0.2662119269,0.0717538595,-0.0446468033,0.215488404,-0.0599906594,0.1828856021,0.2250173688,0.0842254236,-0.2262138128,0.3378059268,0.3512512147,-0.0287638567,-0.1293130815,0.6900850534,0.3549058735,-0.1810753495,-0.0834742635,0.2515942454,0.0005791547,-0.1217798516,-0.1444719434,-0.2440908551,-0.1194600463,0.137125954,-0.1653459072,-0.4723704755,0.1429694295,0.3088828325,0.315862447,-0.0227740556,0.3322768807,0.2209158093,0.2178193778,-0.2352552116,0.3177048266,0.2284146547,0.2080644071,0.133549735,0.1596847177,0.1499284804,0.0468914472,-0.2949413061,0.4081040323,-0.438805759,0.0154482657,-0.0470500253,0.1584734768,0.1952102482,-0.2692917883,0.0826459527,-0.0452418029,-0.016630251,-0.2600587904,0.2939871252,-0.4036771655,-0.0206773374,-0.2553531826,-0.1937350333,-0.2414623946,-0.0963660553,-0.1352775544,-0.53027004,-0.0924401581,0.139797166,-0.1652861983,0.2769560814,-0.0640742555,0.0000356266,-0.0711736903,0.2686168551,-0.0679077655,-0.1023266762,-0.0763239264,-0.1858243495,-0.2031057328,0.1488704234,-0.2374664098,0.1368871778,-0.2526893914,0.4799545109,0.143812865,0.5463255048,-0.3531483412,0.1395496577,-0.4323921502,-0.1334947646,-0.4240529835,0.1600681692,0.0628734753,0.5529722571,0.2093481123,-0.1600458622,0.2817632854,0.0416953303,-0.1844717562,-0.1637284607,0.1966307163,-0.3929294348,-0.4286778271,0.1412266642,0.072345823,0.1066837385,0.0655791163,-0.2056892514,-0.1574924886,-0.0432484746,-0.0586747266,0.1555761099,0.3227590024,-0.0323966034,-0.3471820354,-0.3678293228,0.0114205144,0.1143780723,-0.2356737405,-0.196918875,-0.1125311852,-0.2713981569,-0.1514755785,0.2202937752,-0.1036953777,0.3974577487,-0.3559351861,0.0480310768,-0.2433433235,-0.0645577386,0.458535701,-0.1786570698,-0.3056346774,0.0793514401,0.244796142,0.2134696096,-0.0733820423,-0.4194638133,0.2626645565,0.0804453641,0.0447421037,-0.2134111375,0.1362763494,-0.2202365547,-0.2952007055,-0.1171937361,0.1270846874,-0.0512986481,-0.0708289295,-0.1670592874,-0.1908711046]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/326","title":"Large dataset in Squad2-format","comments":"You can do the tokenization step using `my_tokenized_dataset = my_dataset.map(my_tokenize_function)` that writes the tokenized texts on disk as well. And then `my_tokenized_dataset` will be a memory-mapped dataset too, so you should be fine :)","body":"At the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?","comment_length":34,"text":"Large dataset in Squad2-format\nAt the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?\nYou can do the tokenization step using `my_tokenized_dataset = my_dataset.map(my_tokenize_function)` that writes the tokenized texts on disk as well. And then `my_tokenized_dataset` will be a memory-mapped dataset too, so you should be fine :)","embeddings":[-0.0358469076,-0.247777611,-0.0586996078,0.246388644,0.2725090384,-0.0800485983,0.0325766578,0.4199478924,-0.1291513592,0.0406961478,-0.2698194087,0.1034108475,-0.2411333025,0.2175490558,0.2678762674,-0.0001746843,0.1660441458,-0.0381541513,0.0205726288,0.1060394868,0.1364686489,-0.0983715579,0.0692259222,-0.098045446,-0.3837281466,-0.2225388438,-0.2114028782,0.0674200729,-0.3816430867,-0.4354844391,-0.2983789444,0.2929075062,0.2630604506,0.5424135923,-0.0001224383,-0.2122880071,-0.0822095945,-0.355851531,-0.3333749473,0.0622420982,-0.0192783251,-0.3454458416,-0.1117130443,-0.1266127825,-0.1577755809,-0.2789115608,0.1271944791,-0.2555705011,0.8061004877,-0.1531662047,0.0214495026,-0.0215730332,-0.0481791385,0.0517999865,0.1918322295,0.3268030286,-0.0728117302,-0.0447113775,0.6106310487,0.2984131575,-0.184012413,0.1200443208,-0.0851198211,-0.1498504877,0.1096886098,-0.1629780978,-0.1704001427,-0.1921743155,-0.0237779431,0.2037169486,0.4222000241,-0.3385160565,-0.1862229705,-0.3305151165,-0.1230701283,0.0508738048,0.0350252986,0.3019004762,0.0060198815,0.3412303627,-0.3143237829,-0.4462244213,-0.3264684677,-0.0131069003,0.1752528548,0.1036070511,-0.1256473064,0.0252939034,-0.1996139437,-0.2140208185,-0.1438999474,-0.1332382113,-0.0655241832,0.188609004,-0.0463632196,-0.3719006479,-0.2080021352,-0.3558419347,0.7443717718,0.2502095401,0.1825263202,-0.0044432594,0.1786465645,-0.0180104822,0.3476016521,0.0791966021,-0.1344190389,-0.2228825241,0.2056197524,-0.0692040771,0.0441734232,-0.1626304239,-0.2966942191,0.3006076813,-0.1087574214,-0.2092300504,0.1497644633,0.1247977018,0.086250253,0.1220537871,-0.1460025162,0.0125121465,0.1612524539,0.3237168193,-0.0606556125,0.1013348028,-0.1591730565,0.0791071579,0.251031369,-0.1393309981,0.0203039441,-0.0825997218,-0.0755887032,0.2826264501,0.169211179,0.1916387081,0.1067391932,-0.0948662907,0.3511994183,-0.3164987862,0.1982142478,-0.0433665663,0.1172584742,-0.0536002629,0.1517839879,0.17542243,-0.2161428034,-0.0682597533,-0.2824198604,0.0142466454,-0.2111289352,-0.3164356351,-0.0480192788,-0.0132391006,-0.2957737744,0.2394763231,-0.1990526915,0.3587201536,0.0694444552,0.0011227143,-0.0613626428,0.3853585422,-0.2045324147,-0.2523868382,0.1157232821,0.4234727621,-0.165144816,-0.0755069181,0.6700637341,0.0387734957,-0.256937176,0.4542044699,0.0799492598,0.1556759775,0.0222006142,0.1333848983,0.1889381558,-0.5267193317,-0.4063301086,0.4948351085,-0.1921532303,-0.2608957291,0.3926098943,-0.027362138,0.7847687006,-0.224837631,0.0983915627,0.4408167899,0.0235162061,0.3045356274,-0.3460551202,-0.3247426152,-0.1307469457,0.2371600866,-0.2409185767,-0.2785597742,-0.1992860734,0.3613536954,0.0979598016,0.0059875655,0.3558785617,-0.0728290826,0.445025593,0.0458202623,-0.1008841172,-0.034379337,-0.6144986749,-0.1862875223,-0.0533258282,-0.1896263361,0.0219220128,-0.4489982426,-0.2040059865,-0.0548460037,-0.0650103986,-0.0996520445,-0.0107854232,-0.0867551491,0.1808195561,0.0646054968,-0.4297647774,0.0958206132,0.0679547489,0.121974729,0.1080547348,0.0707083717,0.1028700098,-0.0089806505,0.1983807087,0.1354849339,0.033813674,-0.2850415707,0.2865187526,0.1933010966,-0.122553274,0.0360604264,0.0323351398,0.1992418468,0.3765419424,0.2387423366,0.3395583332,-0.0528761707,-0.1034863591,-0.2263322324,-0.4660384059,0.3321805894,-0.172244817,0.1516910046,-0.1176439151,-0.1708733588,0.3835427463,0.0604184307,-0.1853945255,0.0116602741,-0.1172106266,0.1202117503,0.3583424985,0.1511693895,-0.0837305859,0.1534204334,0.4169473052,-0.188087061,0.1579304487,0.0922692046,0.0193203297,-0.2057335824,0.0339047126,-0.3745421469,0.1113106683,0.2045274973,-0.1165310144,-0.032033205,0.4354265034,-0.0989613235,-0.0211146828,0.0098216245,0.0550688617,0.2727505863,0.3929631412,-0.1295709163,0.0289757382,-0.0325896516,0.202135995,0.0086919954,-0.1281420588,-0.0460867733,0.0887818262,0.1770050675,0.0038029887,0.2265064269,-0.1486002207,0.1947254986,0.3439479768,0.1537312716,0.0585660562,0.1289347857,0.3415546715,0.7686249614,0.0309741292,-0.3049447238,0.0493489094,-0.2318971306,0.3245337307,-0.0766607448,0.43103984,-0.3617490232,0.2851543725,0.0097461278,-0.0011517979,-0.2504860163,-0.1496311426,0.2477441877,0.0177355893,0.1506178826,-0.1847651601,0.2407236993,-0.3253250718,-0.0264812615,0.1529536843,0.0557004921,-0.2686520517,-0.150921002,-0.0527387373,-0.1377538294,0.0886375308,-0.1250995845,-0.2744232714,-0.3307993412,0.516752243,0.0543538928,0.0817118734,-0.3696947992,0.0508889034,-0.0841617957,-0.4887417555,-0.0598633923,-0.189150244,-0.6689900756,0.2267435938,-0.2004956454,-0.1961240321,-0.2640039027,0.121244736,-0.0289270934,0.0722713619,-0.2839437127,-0.2136158347,0.1547720581,0.0793500543,-0.0785542503,0.0479094014,0.3909583986,-0.0872953907,0.0794652477,0.0557640977,-0.0035893854,0.1672790498,0.1231797263,0.2721035182,-0.0246197153,0.4078175128,0.2914268672,0.8061866164,0.3848915398,0.1959751844,0.0897487029,0.1119625866,0.2658505142,-0.0288793165,-0.007531255,0.4521793425,0.0277555175,-0.0088196034,0.2692959011,0.3838596046,0.1242348552,-0.0907758847,-0.3080970645,-0.1906188428,-0.2369171381,0.3694243729,0.3165629208,0.1075833812,-0.0840523392,-0.1937538385,-0.1551182717,-0.3198577762,-0.0943368524,0.1596541554,0.3163716495,0.1879037321,-0.6765270233,0.3549472392,-0.4088259935,0.0791276321,0.0022836793,-0.108303912,0.0055813733,-0.0226307176,0.1526863426,-0.1477396935,0.6616031528,0.3229282498,-0.0914456993,-0.2203113735,-0.0861198679,-0.2898351252,0.1937619746,-0.1531159729,0.4884580374,0.4400158226,0.5459082723,-0.6200933456,-0.1717019081,-0.1250707805,0.0530040339,-0.207223475,-0.1415389329,-0.1294427067,-0.188338086,-0.5278856158,0.1244241148,0.2873000801,-0.0425458699,0.3593448997,-0.1278073192,-0.0597834922,0.007527614,0.0739024878,0.2289947271,0.0508404449,0.0989776924,0.1173745766,0.170946613,0.2239158452,0.1616594642,0.1182694137,0.053000249,0.0410059132,0.120805636,0.0097249942,0.5497874618,0.4826801717,0.1355557889,-0.0334792808,-0.4917354882,0.2676380873,-0.4027078152,0.0017770105,0.0890903845,-0.0926766172,-0.3960807621,-0.0910904706,0.3552254736,-0.0016621603,-0.1386159062,0.0111448318,0.2135969996,-0.1599932462,0.444526732,0.1186923757,1.3274313211,-0.415669173,0.4368340969,0.0171288941,0.1409609914,0.6164472103,-0.8450108767,0.0388019681,-0.3598651886,0.4247410893,0.0297946259,-0.1085491702,-0.0936315879,0.4053987563,-0.3267577291,0.1037577093,-0.1090011597,0.548972249,-0.2766166627,0.676764071,-0.1359956414,-0.3900085986,-0.5071853995,0.0100568561,-0.24796924,-0.2652220428,-0.0769328997,-0.1078804359,-0.0773208961,0.0363461599,-0.0582711585,-0.2953857481,-0.1762766838,0.1564580351,0.1803081483,-0.2939793766,0.1962884218,0.1551854759,0.0704066306,0.3557056785,-0.0940317884,0.1165550798,-0.3090676367,-0.1790238917,0.0290010404,-0.0475667194,0.1323616952,-0.0566526242,-0.208013922,0.0645951405,0.2362639159,-0.0711735711,-0.4781263471,0.070945248,0.3397392929,-0.5559619665,-0.2173306942,-0.0462065339,-0.2046856433,0.0542465746,-0.0296260566,0.1538581848,-0.0953046456,0.232022509,0.5043819547,-0.057784386,-0.0062329792,0.2101411521,-0.1361910552,-0.3036091924,0.2424916625,0.0141589092,-0.1571898907,-0.1543036848,0.6779859066,-0.0044226651,-0.0961270481,0.0694539249,-0.0159717575,-0.3258942664,-0.4036073983,-0.1101429164,0.3476105928,-0.056240838,0.2685332,-0.2250954658,-0.4499088824,0.3364256918,0.0950102583,-0.0445492938,0.1758455932,0.0866953805,-0.1857729107,0.3449880183,-0.1888073981,0.1834845841,-0.0558894798,0.1168352142,-0.0679826438,-0.2087293118,0.0415927134,-0.1210605279,-0.0929049477,0.0625022799,-0.2411754578,-0.0948671252,-0.1557496935,0.2416904569,0.1395371556,-0.0594771057,-0.1366859972,-0.2649320066,-0.1226402372,-0.1376508474,0.2260236442,0.2327340543,0.2882927656,-0.3781515658,-0.021363942,-0.1897070706,-0.0122640682,0.1488154233,0.2742528319,0.33082515,0.1299727708,0.1034356207,0.0452063866,-0.1667609513,0.0202369429,0.5075788498,0.7554006577,0.0878295377,0.2576687932,-0.0609876886,0.0023058681,0.1873064488,0.0194995012,0.3032789528,-0.2602253258,0.1139059588,0.1739158928,0.057910651,0.0190147553,0.0631897673,0.3593915701,-0.0394090414,-0.0887115672,0.3182035685,0.0547095165,-0.0583629049,-0.1832580268,0.1159028411,0.7018389106,0.1988418251,-0.2889252901,-0.2357717454,-0.10478165,-0.164562434,0.3734858334,0.2908750772,0.2932788432,0.1042086557,-0.0919696316,0.2733827531,-0.0017071998,0.0982302055,0.2374656796,0.0866244957,-0.2261619866,0.3672831357,0.3661682308,-0.0586356483,-0.1150513366,0.7066913247,0.2983399928,-0.1959277987,-0.065762043,0.266495645,-0.0261348113,-0.1698407531,-0.2138808519,-0.2286643535,-0.1545480192,0.1448762715,-0.2393973619,-0.4700552225,0.0861263126,0.2842110991,0.3571471572,-0.0751369074,0.3392672837,0.1775911599,0.2466921061,-0.2510386705,0.370790273,0.2150051743,0.1839134097,0.1710395813,0.1525605023,0.2048102915,0.0100610359,-0.2900560498,0.3880849481,-0.4903666079,0.0233788285,-0.0258816443,0.1521198899,0.1927025467,-0.2843948305,0.1054351553,-0.0499683507,-0.0337048732,-0.2527788281,0.2708311975,-0.4384321272,0.0505168252,-0.295776397,-0.2173876911,-0.2327020615,-0.0732440725,-0.1300655007,-0.4757884741,-0.0652542859,0.176982522,-0.1273905784,0.3105265498,-0.0461530983,0.0060550617,-0.0735064447,0.2283282578,-0.0213605836,-0.1428967267,-0.1040666774,-0.1637127697,-0.2287216038,0.126767844,-0.1789252758,0.1197195724,-0.2743480206,0.4088410139,0.1407328844,0.5077558756,-0.3582148552,0.2111836076,-0.4199120104,-0.119510524,-0.4327723682,0.1643100381,0.0219518654,0.5623592138,0.2280697674,-0.1457137614,0.2083114535,0.1127192155,-0.1706877202,-0.1604555249,0.177902326,-0.3798595369,-0.4262799621,0.1012804732,0.1023630425,0.1308807135,0.0253151059,-0.1624245644,-0.0846507102,0.0042993878,-0.0882550627,0.1528504193,0.3082260787,0.0007684805,-0.3237611055,-0.3393217921,0.0228501689,0.0924466178,-0.2554957271,-0.2394739687,-0.2016498893,-0.2231935412,-0.131410569,0.1888444722,-0.1389206201,0.3724318743,-0.3437929451,0.0827581584,-0.2518282533,-0.1133990437,0.4629928172,-0.2120740116,-0.213694945,0.107956171,0.1733536124,0.2590770125,-0.0126357526,-0.4606678188,0.2509465516,0.1250086874,0.0348838232,-0.2539888322,0.1203561947,-0.1843298972,-0.3070386946,-0.1378579438,0.19339481,-0.0168408137,-0.105762735,-0.2072908878,-0.1747219563]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/326","title":"Large dataset in Squad2-format","comments":"Does it have an affect to the trainings speed ?","body":"At the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?","comment_length":10,"text":"Large dataset in Squad2-format\nAt the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?\nDoes it have an affect to the trainings speed ?","embeddings":[-0.0916126966,-0.4042828381,-0.1202021316,0.3813428581,0.2523303628,-0.0287667736,0.0082553821,0.5306742191,-0.064587146,0.0953805149,-0.3321513832,-0.0024423997,-0.2484575808,0.3060847521,0.3058586717,-0.0864770338,0.1506180018,-0.0174991135,-0.0542508103,0.0358302221,0.1572890133,-0.1436901391,0.069373779,-0.1060646176,-0.3642028868,-0.0933875665,-0.179266274,0.1341924369,-0.3202067316,-0.4064352512,-0.2470843494,0.2088611424,0.2356254309,0.5584247708,-0.0001231781,-0.1314166188,-0.0989727378,-0.336345464,-0.3076656759,0.1573538184,-0.0952426344,-0.3345980942,-0.1271150857,-0.0993398279,-0.2453672141,-0.1974134892,0.0718541741,-0.2014814913,0.7306617498,-0.0731905326,0.022987213,-0.087091364,-0.1596626192,-0.0083322264,0.2412008196,0.2269033939,-0.0632916689,0.0760342628,0.6111437678,0.4336218238,-0.1252911985,0.1135400832,0.0811840072,-0.0756719783,0.0250094943,-0.1589274108,-0.0232892223,-0.1409396827,-0.0525071509,0.2532564998,0.4559463263,-0.1516250372,-0.179346025,-0.3007984459,-0.0143626137,0.0367577784,0.0596697293,0.3672980964,0.0212330818,0.4015032649,-0.311835438,-0.4134724736,-0.3809514344,-0.0437755994,0.1721665561,-0.0353086293,-0.1233486384,0.0090859346,-0.1354377419,-0.0976504311,-0.0973675698,-0.1690478325,0.0261418726,0.0770013779,-0.0536871254,-0.4047775865,-0.1907617897,-0.2821290493,0.6157487035,0.2288684398,0.2707796097,-0.0243750438,0.1818035543,-0.0828597471,0.3759941757,-0.0103529496,-0.2171521485,-0.3132341802,0.1601893157,0.0580190867,-0.0444505289,-0.1122536957,-0.2349018753,0.1642772704,-0.1989032328,-0.2507750988,0.1437778473,0.1386446357,0.0332526863,0.1288330406,-0.2039417475,-0.0646190643,0.1904866993,0.3212021291,-0.1486820579,0.1543858349,-0.164518103,-0.0120715834,0.2111258507,-0.184465766,0.0076945294,-0.0729106814,-0.1521609575,0.2869651914,0.1556565762,0.03936572,0.130776301,-0.0663250908,0.4189457893,-0.2681811452,0.1091777831,-0.0201172773,0.0880568177,-0.0120541714,0.0227620229,0.0560149848,-0.274425,-0.0393881351,-0.2884218097,0.0639432147,-0.3208286166,-0.3354717791,0.027911596,-0.0150307361,-0.2816033661,0.3206745982,-0.1286547929,0.3574277163,0.0078329733,-0.0657507256,-0.0299785584,0.3860720396,-0.2450544834,-0.2481837422,0.0441330932,0.3686077893,-0.1767125577,-0.109467186,0.7416324615,0.0436633341,-0.3615175486,0.5149348378,-0.035298191,0.1686182469,0.0188290682,-0.1246155873,0.3292810023,-0.5186533928,-0.4533594251,0.5223739147,-0.1990065277,-0.2254761606,0.4550673366,0.0241298862,0.6667519212,-0.2429820448,0.100282222,0.4482401311,-0.0119752027,0.272213757,-0.2971560359,-0.319277823,-0.1374312639,0.2890134156,-0.2861047089,-0.2058042288,-0.204691723,0.3215744197,0.1572571993,-0.0090343487,0.2498401403,-0.0431218892,0.2932082415,0.0188304894,-0.0815509409,-0.0314200185,-0.7061689496,-0.212572366,-0.0462023206,-0.2721521854,0.1652285904,-0.429628253,-0.217059657,-0.0990324765,-0.097302869,-0.0565076955,-0.0664960518,-0.1425842196,0.2659662068,-0.0439968146,-0.4217074215,0.2180887312,-0.0452391282,0.0789796486,0.02997238,0.1312630624,0.1380649656,0.0156936198,0.174017489,0.1793350726,-0.0716552064,-0.3312722147,0.2747027576,0.1326962262,-0.1385497302,0.0185910854,0.0932146087,0.1765732467,0.4135818779,0.2316063344,0.2986782789,0.0407959335,-0.1252027154,-0.1905712932,-0.5221540928,0.3203805089,-0.209299624,0.1401908845,-0.0893606991,-0.2676163912,0.3060512841,0.0408627056,-0.2894815803,0.0328151397,-0.1001744643,0.163201049,0.3546434939,0.2183327079,-0.2144828737,0.0665503815,0.441785574,-0.2434233725,0.0259222388,0.1150624529,0.1095405966,-0.1915593892,0.0171695184,-0.4007899761,0.2272356898,0.1783146858,-0.1232612208,0.0191831123,0.4372399151,-0.1665040404,-0.0621678829,0.0646262765,0.1894619763,0.2750098109,0.4520285726,-0.1864707172,0.0195201952,-0.0055148136,0.1991055161,-0.0426723137,-0.1303481311,0.0642774254,0.1005320773,0.2667022049,-0.1048860103,0.226990208,-0.110503234,0.2854580283,0.3328084648,0.0973438844,0.0390585102,0.0445366651,0.3426456451,0.7608938217,0.0987153202,-0.3017770648,0.0983486995,-0.1843998879,0.3573459387,-0.0738219395,0.3636404276,-0.4229421914,0.22680749,-0.0813430324,-0.0611035191,-0.1735879183,-0.1101067439,0.2012299299,0.0727454796,0.0752517134,-0.1547593325,0.2625203133,-0.2889209986,-0.0308708716,0.2100631297,0.0481527038,-0.2193451971,-0.2001577914,-0.0281255729,-0.0269068666,0.0957517996,-0.2764951289,-0.214973703,-0.3410955667,0.5687466264,0.1201086566,0.0643916577,-0.4032885134,0.1418689191,-0.0697640032,-0.5342755318,-0.0181987248,-0.2230718434,-0.7341262698,0.1730323434,-0.2358429581,-0.1993463635,-0.3527305424,0.0833306313,0.0220471118,0.0487269834,-0.2667472959,-0.2415149808,0.1391977221,0.0706828162,-0.1427094787,0.1220772639,0.3512186408,-0.2006009072,0.0935015827,0.0995284468,0.0691325665,0.1533355117,0.2650023997,0.3367179334,-0.0890145078,0.3839758039,0.2097111195,0.7681170106,0.4958949089,0.1190320477,0.0133673912,0.0779620185,0.2716044784,-0.081872277,0.1250738949,0.4699003994,0.0490523726,-0.0399784334,0.2928282917,0.3644386232,0.1131416112,-0.142047748,-0.2861054242,-0.0531813465,-0.1990703493,0.3596060872,0.4102487266,0.0896230265,-0.0933372229,-0.2360648066,-0.0962993428,-0.2707057595,-0.0872498006,0.1683529913,0.3103118837,0.065793395,-0.6208825111,0.1898634881,-0.3027372658,0.0444141999,0.0762274265,-0.2392734587,0.037635047,-0.0227056388,0.2431533635,-0.1084110439,0.6912202835,0.27093032,-0.1677735299,-0.2711745501,-0.0683654398,-0.3059767187,0.1472732872,-0.2079284936,0.4564658403,0.4670689106,0.5726323724,-0.5202747583,-0.2585200369,0.0667098537,0.0764833987,-0.1692353636,-0.2194424868,-0.1986652762,-0.1810703427,-0.4672773778,0.1477922797,0.275218457,-0.0227187406,0.3350333273,-0.1775729358,0.027091451,0.1001702473,0.0855989084,0.2307860553,0.0277902503,0.2118676305,0.1043541133,0.1933348775,0.2854239047,0.1081575751,0.2289895713,0.0310069099,0.0718102753,0.1330859959,0.0222871639,0.5797349811,0.6130539775,0.1358132213,-0.1072482243,-0.5417235494,0.3559287488,-0.3476579785,-0.0061235055,0.201814115,-0.1168734729,-0.2939154506,-0.1963528246,0.5307025909,-0.0031462074,-0.0590298921,-0.0232014786,0.193176806,-0.0485927016,0.3130855262,0.0879588202,1.3217196465,-0.3237430751,0.4942190349,0.171477437,0.0820490047,0.5228345394,-0.8318601251,0.114087157,-0.2734008431,0.3540699184,0.0155422669,-0.1822998673,0.0063438043,0.4263727665,-0.3061113954,0.1395881027,-0.1096654162,0.5178824663,-0.3440020084,0.6639131904,-0.0220051184,-0.3023633361,-0.6163764,-0.0264277421,-0.1770665795,-0.2394095361,-0.0622027591,-0.1403306723,-0.1094851568,0.1113021299,0.0625196695,-0.3017003536,-0.0968281925,0.1679456979,0.2561889291,-0.456579566,0.2545068562,0.2652526498,-0.041043032,0.1756804138,-0.063569963,0.1716645062,-0.3235403597,-0.1783404648,0.0310756061,0.0184748415,0.1472890377,-0.1148857102,-0.2475634664,0.1597598493,0.1994359195,-0.0742175356,-0.3675849736,0.1316123754,0.1394975185,-0.5993118286,-0.3203550875,-0.0089434283,-0.29776299,0.1042052582,-0.0319812484,0.1308025271,-0.1825190783,0.329039216,0.5678566694,-0.1727849394,-0.0012927389,0.2557198703,0.0024991101,-0.2375738323,0.2286910564,-0.0148140667,-0.1164351553,-0.139200002,0.7083777785,-0.0056811897,-0.1215051785,0.134644866,0.0451077931,-0.29563573,-0.3921365738,-0.1181707308,0.3174936175,-0.004009455,0.2157169431,-0.0342098065,-0.381000042,0.3066597879,0.1041047275,-0.0544749461,0.0865837261,0.0849232897,-0.0865914226,0.3233177662,-0.1702817827,0.2031590044,-0.0354600921,0.1048431844,-0.0185020976,-0.2273291945,0.0978923291,-0.0760607943,-0.094339788,0.0184802283,-0.3141469061,-0.0956102461,-0.0938350335,0.2336195409,0.0532137565,0.065269269,-0.0898632482,-0.2784270644,-0.1202516854,-0.0323939137,0.2646833956,0.0964104608,0.387796849,-0.3711254895,-0.1426847726,-0.354429394,0.0388233848,0.1742878854,0.2018381655,0.4254310131,0.1002885103,-0.0073273024,-0.0076980069,-0.21579656,-0.0734341517,0.4901344478,0.6201223731,0.0464963503,0.2732168734,-0.1004570797,0.026077982,0.2045013607,0.1082574949,0.3544520438,-0.3042118251,0.1283393502,0.0637707636,-0.0043686703,-0.0269529596,0.0958542228,0.4251118898,0.0041053444,-0.0587896444,0.3523084521,0.0593168437,-0.0306868218,-0.1675902605,0.1525611579,0.4552884102,0.1828337908,-0.2365012616,-0.0397225767,0.0098593375,-0.0971809998,0.4306254983,0.2666046917,0.2211074978,0.0914626345,-0.0446465388,0.1639152616,-0.0424362235,0.0643641278,0.2530822754,0.1033748612,-0.2768442631,0.3642916977,0.2594676018,-0.0954416469,-0.0482327268,0.7156041861,0.3176779747,-0.2587218583,-0.1472114921,0.1638071239,-0.0175652858,-0.1387815922,-0.1087421849,-0.1981767416,-0.1276632547,0.2129512727,-0.283661902,-0.3594246209,0.275038749,0.2914137244,0.3570228815,0.0170237124,0.3186583519,0.1382165849,0.2792609632,-0.1799263209,0.3111912608,0.1682050079,0.1750209183,0.1870756,0.199069798,0.073661156,-0.0792623758,-0.2346976399,0.3576026559,-0.4047879279,-0.0112472465,0.0569776334,0.0856991485,0.31053707,-0.3215403259,0.1451633275,-0.060956385,-0.0386352167,-0.0931214839,0.1838677526,-0.4105437994,-0.1801652163,-0.2551710308,-0.2193188816,-0.2728280127,-0.1589117646,-0.1903887093,-0.4832399189,-0.072700873,0.0936466381,-0.1538515538,0.2072837055,-0.0095468927,0.0010129351,-0.1576219946,0.4083144069,-0.0808308497,-0.0657128766,-0.0543870814,-0.1590171605,-0.1780878603,0.1162344441,-0.3210654557,0.18604289,-0.1884109974,0.486431241,0.1087997854,0.5651049018,-0.1554285139,0.0131747723,-0.3438344002,-0.1537238061,-0.4773207009,0.0615523979,-0.0061547444,0.5800632834,0.1662338525,-0.1655191034,0.3096906543,0.0935579166,-0.2040202022,-0.1020785794,0.2551623583,-0.4508683681,-0.3944252133,0.1119323149,-0.0190125052,0.0819005594,0.0387638882,-0.1675011367,-0.1152724847,-0.0428180881,-0.0406464897,0.2243502736,0.3280459642,0.0473625511,-0.3841529787,-0.3087951839,0.1102562845,0.1016925871,-0.1988608241,-0.1906075329,-0.1499968022,-0.2955136299,-0.1205309704,0.240582481,-0.0204649884,0.4274500608,-0.3767482638,-0.1153090671,-0.1961193979,-0.0726418421,0.5283040404,-0.12924698,-0.2634772658,0.1182462499,0.2273479998,0.2202874273,-0.0399565287,-0.4235613644,0.2382873148,0.1132951379,-0.0317568816,-0.2883372307,0.170847252,-0.1752184629,-0.3100946248,-0.1222964525,0.1375980377,-0.0336696431,-0.1499437988,-0.2536101937,-0.2185370326]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/326","title":"Large dataset in Squad2-format","comments":"In your training loop, loading the tokenized texts is going to be fast and pretty much negligible compared to a forward pass. You shouldn't expect any slow down.","body":"At the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?","comment_length":28,"text":"Large dataset in Squad2-format\nAt the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?\nIn your training loop, loading the tokenized texts is going to be fast and pretty much negligible compared to a forward pass. You shouldn't expect any slow down.","embeddings":[-0.0998464674,-0.3400501907,-0.072254546,0.2740523219,0.233014524,-0.134118855,0.0534660295,0.4866881669,-0.0811557695,0.0389788896,-0.23714149,0.0705089793,-0.2566494942,0.309556216,0.2704547048,-0.0745398998,0.1412350386,-0.0320492834,0.0148968324,0.0299363341,0.1451997757,-0.1339221299,0.0269506257,-0.0343342684,-0.3312619925,-0.147840932,-0.1654999107,0.1486865729,-0.3085708022,-0.4827928841,-0.3069819808,0.2874170542,0.2753958702,0.4879121184,-0.0001232336,-0.1371826828,-0.0272386987,-0.3751021624,-0.3914620876,0.0622953884,0.0496604741,-0.2882883251,-0.0649245158,-0.0946050212,-0.1764810383,-0.2693074346,0.1124373823,-0.1765535325,0.9333483577,-0.0912950411,-0.001321081,-0.0185958948,-0.1706783324,0.0342927203,0.2240020037,0.1699875593,-0.0316277221,0.0094061894,0.653195858,0.3252064884,-0.2057896852,0.1627473533,0.00764257,-0.1622393876,0.0985202119,-0.15895392,-0.0159675162,-0.1528659612,-0.0900060683,0.2489254326,0.3597235978,-0.2345626652,-0.2427043319,-0.3881295919,-0.031607341,0.008495491,0.0847673863,0.2578563988,-0.0303193182,0.4028695822,-0.3450419307,-0.4935597479,-0.3452898264,-0.0025279061,0.1328754425,0.002325847,-0.0820589289,-0.0195719227,-0.1442527026,-0.1430077106,-0.0312887356,-0.1512053013,0.05735939,0.1760599464,0.0076766806,-0.3633486331,-0.1832547635,-0.3150159121,0.6128088236,0.2020144314,0.1659007668,0.0921715796,0.1930484474,-0.0760334879,0.3041756451,0.0039584069,-0.1741166562,-0.2357874066,0.1530188918,0.0032970435,-0.0408503562,-0.0910827443,-0.4117302001,0.2907006741,-0.2795992792,-0.2131229639,0.0398327783,0.1076254994,0.0285532866,0.1241719276,-0.1663440168,-0.0343140848,0.197809428,0.3367663622,-0.1896688044,0.2010002434,-0.0523402914,0.0550012775,0.193146199,-0.2182546258,0.0245846696,-0.0677965134,-0.1678190678,0.3149712086,0.1374841332,0.0973450169,0.1068066582,-0.0808860883,0.3495577574,-0.292362839,0.0944587961,0.0606519543,0.0520383343,-0.103023462,0.0815087631,0.0854252502,-0.2245240659,-0.0646581203,-0.2271315455,0.0724633411,-0.3223105967,-0.3137343824,0.0708979517,-0.0243879799,-0.2372407317,0.266965121,-0.1464351416,0.4468391836,-0.066742599,-0.0000228255,-0.0586486124,0.3950932324,-0.2568891943,-0.1637609303,0.1426975578,0.4170689881,-0.2085472345,-0.1686359644,0.767531395,0.1190056875,-0.2828905284,0.4321811199,-0.0407247879,0.1763848513,0.0575946942,0.0287949704,0.2927293181,-0.6150716543,-0.3437715471,0.5754372478,-0.2338035852,-0.2194793671,0.4251846075,-0.018134838,0.7501599193,-0.2715392709,0.1670282632,0.5265708566,0.0275566597,0.2613630593,-0.3428591192,-0.3132781684,-0.1130179688,0.352265507,-0.2507222891,-0.2015239447,-0.2499058098,0.3206180632,0.1277277023,-0.0399459191,0.2356550097,-0.0552047081,0.1949426532,0.107216157,-0.1261142045,0.068676129,-0.6354449987,-0.1800173968,-0.0353042781,-0.2504541576,0.0739325956,-0.4319304526,-0.2057418972,-0.0524757467,-0.0826043487,0.0094190594,-0.0734386444,-0.1099154279,0.1536721587,0.089274399,-0.4430952072,0.1576543003,-0.0317029841,0.1350343376,0.103367269,0.0933032632,0.1940133721,0.0167998485,0.2226579189,0.1754046977,0.019663956,-0.3061259091,0.3164180517,0.2310002595,-0.1803950965,0.107820563,0.0286296103,0.1224039197,0.3870017231,0.2914449871,0.2757450342,-0.020401625,-0.0890708119,-0.2014707327,-0.4354258478,0.4695908129,-0.1399256587,0.1029195413,-0.0469945371,-0.1922258437,0.2470768541,0.0279658064,-0.3278740346,0.0785472691,-0.041131489,0.1010354385,0.3779505789,0.1883982271,-0.2185820937,0.1035120264,0.5042096376,-0.2499515861,0.0631477311,0.1456290781,0.0739471838,-0.1624071002,-0.0654814914,-0.3691547215,0.1035001874,0.1509120315,-0.1132974997,0.0365013964,0.3595151901,-0.1414895505,-0.0264396388,0.0400507823,0.1576570719,0.2307530046,0.336263299,-0.1535274982,-0.012593152,-0.0430570208,0.1768823862,-0.020825278,-0.1473274678,0.0703120157,0.1081736684,0.2291824222,-0.1080216989,0.0923992395,-0.0407775417,0.2209887952,0.3710334003,0.1220967844,0.0494286269,0.1232839748,0.357422024,0.7777605653,0.0740785897,-0.2954411507,0.0345302224,-0.2509972751,0.2499700785,-0.109903276,0.4066915214,-0.3923880458,0.2088262588,-0.1002375409,-0.0704726055,-0.088101685,-0.1080953255,0.2357533723,-0.0128257424,0.040228229,-0.2133981436,0.2169346362,-0.3434363902,-0.0746857226,0.1513520479,-0.0094621992,-0.2644820213,-0.1470504105,-0.0022058873,-0.0863671303,0.0888572112,-0.2968888581,-0.2773435414,-0.2985850275,0.4928011894,0.0754599497,0.0640896633,-0.3779042959,0.0815671235,-0.0153703783,-0.4000914991,-0.06912449,-0.1580911875,-0.6986131072,0.2222498506,-0.165712297,-0.2289089561,-0.2681019604,0.0822325274,-0.0588756762,0.0815771818,-0.3671654463,-0.2016134858,0.2188031673,0.0321707651,-0.1601269394,0.0376335457,0.3266258538,-0.1108786836,0.09103778,0.0565655977,0.0788849667,0.1923375726,0.0437459983,0.2103662938,-0.045556847,0.5006765723,0.2444775254,0.8879320621,0.4185721874,0.1632732898,-0.0356384031,0.0654547885,0.164102599,-0.001624089,0.0714655146,0.4287313819,-0.0277694613,-0.0217478089,0.2838865221,0.3437896967,0.0208334737,-0.2105353177,-0.2243260592,-0.1663825512,-0.2680464387,0.3488025665,0.301602304,0.0969368815,-0.0536190346,-0.1421072632,-0.1444643885,-0.2485810816,-0.0418124646,0.0893358588,0.3514489532,0.1249718666,-0.6669871807,0.271928966,-0.3750924468,0.0028047864,0.1075820774,-0.101604335,0.0590999611,-0.062687315,0.1974937618,-0.1959409267,0.6504691839,0.3883799314,-0.0828795284,-0.2534446418,-0.1487042755,-0.308142066,0.148733899,-0.1808010638,0.410020858,0.4926868379,0.4620301127,-0.5409819484,-0.2708087265,-0.0290107727,-0.0087571982,-0.1661582887,-0.0959237888,-0.1718364507,-0.1978621483,-0.545214355,0.181826219,0.3202822208,-0.0085399654,0.2829557955,-0.1344594657,0.0567006581,0.0850086436,0.1073691994,0.187283501,0.0188476555,0.17818214,0.0895374715,0.2845397592,0.2857456803,0.0903420448,0.1812286228,0.1071604267,0.0152681405,0.1032182947,0.0495482348,0.5974902511,0.6548079252,0.1536557674,-0.0044996343,-0.4806945622,0.2829881012,-0.3764107525,0.0277704149,0.1394282728,-0.0948402211,-0.3397719562,-0.1617484838,0.426053822,0.0052198838,-0.1099753007,-0.0014933853,0.107256569,-0.1306718737,0.4067906737,0.0306712855,1.3512606621,-0.3563821912,0.4912479222,0.1727318168,0.2168494165,0.511880517,-0.8861692548,0.0963311344,-0.3162590265,0.387607336,-0.0304378681,-0.1310204417,-0.1044384092,0.3876610398,-0.2980631292,0.105726026,-0.0487184152,0.5162269473,-0.254567802,0.6119434834,-0.0532312542,-0.3716334105,-0.6192556024,-0.0415631644,-0.1461667269,-0.2977314293,-0.0556899458,-0.1509171575,-0.0802600309,0.0464850813,0.0569289178,-0.2324222922,-0.1602412611,0.1106458008,0.3232052028,-0.3037297726,0.2290357053,0.2459354252,0.053067375,0.2225888669,-0.0827510133,0.200500831,-0.4225772917,-0.1420373619,0.048864916,-0.0342628956,0.2505094111,-0.0836913809,-0.2108563483,0.2170357406,0.1719896495,-0.0862250775,-0.4488454461,0.0593666695,0.3781777918,-0.5011570454,-0.2208089232,0.0211799983,-0.1953941584,0.0459365882,-0.051065769,0.1341562867,-0.2042076588,0.2137427628,0.5658396482,-0.1569910049,0.0337467454,0.2941561937,-0.1210546866,-0.3052670956,0.3507449627,0.0744544119,-0.1224070042,-0.1303799897,0.6699294448,0.0875537694,-0.1203663424,0.1151726842,0.0244883485,-0.3150231838,-0.3878088593,0.0355331227,0.294903487,-0.0681043491,0.2688258588,-0.1532762796,-0.3670754433,0.3225948215,0.0986827388,-0.081185244,0.1734327376,0.1403615922,-0.1298323125,0.2685960233,-0.1680327356,0.2072601616,-0.062691085,0.1555480212,-0.0826551542,-0.2120367587,0.075601168,-0.0420167074,-0.1184657961,0.0713897571,-0.283853054,-0.0739535242,-0.114384003,0.2430031598,0.0862281099,-0.0285969526,-0.1452964991,-0.2620357275,-0.1635843664,-0.0432470143,0.2405247241,0.1436322629,0.3013543189,-0.3893832862,-0.1232066453,-0.1667733639,0.0185694154,0.1054696664,0.2372743785,0.3904150724,0.0940390751,0.0173790269,-0.0278386194,-0.2103467137,-0.0388990231,0.530207932,0.7436003685,0.1578486711,0.306425035,-0.1554288417,0.0537384786,0.1845121831,0.0057741404,0.3656205237,-0.3396030664,0.0393394865,0.1936206371,0.000806375,-0.0336848162,0.0404309928,0.3611772358,-0.0661933869,-0.0720051825,0.343506664,0.0511191115,-0.0448913574,-0.1086415201,0.1211734414,0.4851032197,0.1997783929,-0.2774341702,-0.1347250193,-0.0625217855,-0.1407152861,0.2962168157,0.3051091433,0.1989794523,0.1717758328,-0.0957488716,0.1877719164,-0.0528719649,0.1498090476,0.0664777458,0.0434193425,-0.2457429469,0.3780344427,0.2527337074,-0.0672855377,-0.1209337339,0.7141617537,0.3640082479,-0.3019389808,-0.0925271958,0.2697065771,-0.0135872914,-0.2406313419,-0.2118638009,-0.1876576245,-0.210034892,0.2540382445,-0.2479699254,-0.3314100802,0.1338115036,0.3403826952,0.4442258477,-0.0892456248,0.2357784808,0.1061919704,0.3274318874,-0.2305300683,0.3929686248,0.1338681877,0.2263185233,0.2677948475,0.0403166302,0.0432622209,-0.0105757359,-0.3140460253,0.3602118492,-0.4900680184,0.0539307334,0.0804791525,0.1562832892,0.3114764392,-0.2725020051,0.1654483378,-0.0721389651,-0.04260654,-0.0997516066,0.1755214036,-0.4402987659,-0.0918197781,-0.2306340188,-0.1336801052,-0.1629974693,-0.2142193317,-0.1493446082,-0.4683659971,-0.0881209224,0.1197658181,-0.1382948458,0.2415235341,-0.0225761104,0.0092840316,-0.077987738,0.3606652617,0.0151161496,0.0071027856,-0.1119028106,-0.1744412631,-0.1661728173,0.0979146063,-0.1681888998,0.1738532782,-0.1981576085,0.4040406644,0.1810657233,0.5885439515,-0.3318024874,0.1254516095,-0.3861024976,-0.1623791158,-0.4126793444,0.170403406,0.0535815395,0.6294961572,0.156255424,-0.1591611654,0.2456200421,0.0447323509,-0.1992175281,-0.0732060224,0.1863850653,-0.4293965697,-0.4038084149,0.0291541554,-0.0140282707,0.0959746167,0.0468867011,-0.1298421025,-0.1176438704,-0.0317665674,-0.0767430365,0.0994961634,0.3009855449,-0.0041389358,-0.335178256,-0.2618252933,0.0422287025,0.06298545,-0.1492683887,-0.176075384,-0.2143103182,-0.1981463283,-0.1694892645,0.2960017323,-0.073968932,0.4579182863,-0.3707360625,0.0830326304,-0.2062580734,-0.0976960883,0.5351342559,-0.2897498012,-0.2838290036,0.1446030736,0.1777909696,0.3432672322,0.0217710156,-0.4834470153,0.2382177562,0.0589201152,0.0281439833,-0.3584786654,0.1546441168,-0.151853472,-0.2857326567,-0.1427768469,0.1686160266,-0.035278555,-0.1014008895,-0.2121017873,-0.2255224735]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/326","title":"Large dataset in Squad2-format","comments":"Closing this one. Feel free to re-open if you have other questions","body":"At the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?","comment_length":12,"text":"Large dataset in Squad2-format\nAt the moment we are building an large question answering dataset and think about sharing it with the huggingface community.\r\nCaused the computing power we splitted it into multiple tiles, but they are all in the same format.\r\nRight now the most important facts about are this:\r\n- Contexts: 1.047.671\r\n- questions: 1.677.732\r\n- Answers: 6.742.406\r\n- unanswerable: 377.398\r\n\r\nIt is already cleaned\r\n\r\n<pre><code>\r\ntrain_data = [\r\n {\r\n 'context': \"this is the context\",\r\n 'qas': [\r\n {\r\n 'id': \"00002\",\r\n 'is_impossible': False,\r\n 'question': \"whats is this\",\r\n 'answers': [\r\n {\r\n 'text': \"answer\",\r\n 'answer_start': 0\r\n }\r\n ]\r\n },\r\n {\r\n 'id': \"00003\",\r\n 'is_impossible': False,\r\n 'question': \"question2\",\r\n 'answers': [\r\n {\r\n 'text': \"answer2\",\r\n 'answer_start': 1\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n]\r\n<\/code><\/pre>\r\n\r\nCause it is growing every day we are thinking about an structure like this:\r\nWe host an Json file, containing all the download links and the script can load it dynamically.\r\nAt the moment it is around ~20GB\r\n\r\nAny advice how to handle this, or an ready to use template ?\nClosing this one. Feel free to re-open if you have other questions","embeddings":[-0.0545848347,-0.2852962911,-0.1122599542,0.3391782939,0.2328381836,-0.0566980094,-0.0175540186,0.5261056423,-0.0142832221,0.1231510192,-0.3440013528,0.0504089557,-0.229718104,0.3335837424,0.3050529659,-0.0663990304,0.0991353616,-0.0230698269,-0.0906955674,0.0547824614,0.1276933402,-0.1004391164,0.1072412282,-0.1047568619,-0.2884635925,-0.093202807,-0.222323209,0.1555062532,-0.3487177491,-0.3709892929,-0.2218335867,0.2514761984,0.2274508625,0.6068015695,-0.0001188301,-0.1579607874,-0.0874531195,-0.371132791,-0.3409987092,0.0094217882,-0.1466664821,-0.3082870543,-0.0981236845,-0.0886728242,-0.2364851534,-0.2456978261,0.0708934069,-0.2652221918,0.7986325622,-0.1562239379,0.040568836,-0.0741084889,-0.0615096726,-0.0237576291,0.2651692927,0.2826077938,-0.0469143316,0.0399456173,0.6693655252,0.3571805656,-0.180837974,0.134596616,-0.0290354975,-0.1296238452,0.052055154,-0.1796750426,-0.0924363583,-0.1878467351,-0.0436573438,0.254371047,0.4576273561,-0.2599056363,-0.1651836932,-0.2734099329,-0.0756251886,0.090472661,-0.0101308571,0.3502885401,-0.0599994361,0.4242154658,-0.3057413399,-0.4321588576,-0.3922760189,-0.0251425449,0.2084842324,0.0153483879,-0.1310133189,0.0095553454,-0.1547452956,-0.1431264877,-0.1099632978,-0.1536833942,-0.0281856768,0.0873341039,-0.0452755652,-0.3583117723,-0.1056862921,-0.2419136167,0.6853433251,0.3096212745,0.2690627277,-0.0268813036,0.192909956,-0.0758467913,0.4402515292,-0.0013245249,-0.1264142543,-0.3822415173,0.1982431859,0.0387519598,0.0194518808,-0.0875381753,-0.2284790277,0.2042650133,-0.202286303,-0.1769857109,0.2542684972,0.1477951258,0.0615179166,0.0988259092,-0.1426522881,-0.0617678538,0.1746450961,0.296431154,-0.0769100487,0.1381340623,-0.133950606,0.0615477599,0.2669119835,-0.2018676996,0.0029150478,-0.0869786069,-0.0910444781,0.2636756301,0.1737741381,0.029667316,0.0747386143,-0.0200970955,0.3805565238,-0.3227533996,0.1361899674,0.056305632,0.060102433,-0.023110969,0.1018716618,0.0852776393,-0.167336911,-0.0754767433,-0.2787410617,0.0729582086,-0.2432889491,-0.2852827013,-0.1052969024,0.0039116777,-0.2607374787,0.2669786513,-0.1322562099,0.3141429126,-0.0433117822,-0.0602901913,-0.0708397776,0.4598161578,-0.1612255871,-0.28749457,0.0635453686,0.4792439342,-0.1848201603,-0.0332709588,0.6506721377,0.0098671978,-0.3962332308,0.3958572745,0.0280587152,0.1006739289,0.0429837033,-0.0953601152,0.2543296814,-0.5074138641,-0.4217803478,0.5393114686,-0.2443239838,-0.2605347037,0.4463006258,0.02251973,0.6679827571,-0.274846077,0.0703996494,0.4080550969,-0.014516796,0.2741247714,-0.2839857042,-0.3567275405,-0.1484094411,0.3062367439,-0.3425218463,-0.2006350905,-0.1711972058,0.2088799179,0.1523037404,-0.0232417192,0.2746721208,-0.0690224245,0.40429461,0.0372944996,-0.1022000462,-0.0887025669,-0.7477780581,-0.2250112593,-0.069215104,-0.2493715733,0.1117637232,-0.4512871206,-0.3244495392,-0.0638480633,-0.0496278852,-0.0896574333,0.0126809012,-0.1090564877,0.166079998,-0.0230733827,-0.4688578248,0.1698388606,0.1008234695,0.0587109067,-0.0044329353,0.1339686066,0.0953638256,0.0488674082,0.2446302772,0.2108036727,-0.005060039,-0.295340538,0.3269506395,0.1623059958,-0.151600644,0.0877644792,0.121279344,0.1465947926,0.3568014503,0.2260500491,0.3174600601,0.0255620945,-0.0981428325,-0.2124676406,-0.5974236131,0.3083744645,-0.2883080244,0.16850923,-0.140621841,-0.1570011973,0.3334652781,-0.0196557529,-0.228318736,0.0455330424,-0.1109139472,0.1101010293,0.3244772553,0.2205685526,-0.1349337399,0.1587058306,0.4085155427,-0.1457024962,0.1181895211,0.1060414314,0.0916224122,-0.1905660182,0.0182619728,-0.3546364307,0.1374901682,0.2153110206,-0.1474412531,0.0021607582,0.5082352161,-0.1415697336,-0.0676461086,0.0492413044,0.18884103,0.2652682662,0.4495715201,-0.1489767134,0.0804625973,-0.0213381536,0.1054086909,-0.0780347362,-0.1114864945,0.0161430892,0.0912111104,0.2461767793,-0.0398584381,0.1670926511,-0.1500355452,0.2235209793,0.3684130609,0.1299192607,0.0133003797,0.0381014161,0.2943356335,0.7636796236,0.0570075028,-0.2626447976,0.0036083341,-0.2698512375,0.38417539,-0.0453445762,0.4312041104,-0.36567536,0.2673391998,-0.0407720096,0.0235107075,-0.2142787129,-0.1119701192,0.2205979377,0.0079548741,0.1361624002,-0.1149552912,0.2640152872,-0.2739056647,-0.0674614981,0.1824938506,0.0918694437,-0.2202754468,-0.1399842799,0.0020498883,-0.0622197352,0.0810446069,-0.2393205762,-0.2569669485,-0.3818123937,0.5645319223,0.086575605,0.070909895,-0.4298084378,0.1229222342,-0.0865108147,-0.4690917432,-0.0160322245,-0.1980783194,-0.6748290062,0.143210873,-0.2502937019,-0.203292042,-0.3053654432,0.1458678395,-0.051773686,-0.0408364497,-0.2877068818,-0.2797086239,0.1760564297,0.1035666615,-0.0456568263,0.1112796962,0.3561869562,-0.1140365377,0.059164878,0.0341261886,0.0009803192,0.1286022365,0.26684919,0.2798267305,-0.0579746291,0.3776391149,0.1794314235,0.7163404822,0.531054914,0.1917656064,0.0540054515,0.1231840253,0.2911298871,-0.0455718488,0.0215350203,0.4439651072,0.0467009768,-0.0146872588,0.2414316088,0.4362855852,0.1281382293,-0.1801180393,-0.3035360873,-0.1919242889,-0.2355510741,0.3212239444,0.4272121489,0.0462151989,-0.1079428792,-0.2916294038,-0.093813777,-0.2869363427,-0.0791633055,0.1776308715,0.3459144831,0.1117928103,-0.5519428253,0.293671906,-0.3180081844,0.1219263151,0.0590970553,-0.2796956301,-0.0142198773,-0.0413150266,0.1754069179,-0.1301075667,0.7278891206,0.3003417552,-0.1499204338,-0.2634136975,-0.0044607162,-0.3165515959,0.2332229167,-0.1521376818,0.493489325,0.4323267341,0.4612773061,-0.6056853533,-0.2030160874,0.0458809659,0.093345359,-0.1894075274,-0.215196684,-0.1415353715,-0.2396707088,-0.4877483845,0.0602480434,0.250908792,-0.0617534034,0.3000924289,-0.124380596,0.0593019165,0.0947006419,0.0857962221,0.2245158851,0.0212442279,0.1737191081,0.0476025008,0.1834724098,0.2410461009,0.0723522529,0.2561670244,0.0488519594,0.0902033448,0.1421739459,-0.0441983901,0.56233567,0.5247654319,0.148141101,-0.1474722475,-0.4702962637,0.3038349152,-0.3573546112,-0.1650463045,0.1265571266,-0.1162284911,-0.2377074957,-0.1285062879,0.5184246898,-0.08365307,-0.1164585277,-0.0138003472,0.2649804056,0.0120417727,0.3868060112,0.0362979099,1.319506526,-0.3614299595,0.4213482141,0.1105842888,0.0414198749,0.5679705739,-0.7674894929,0.0683086812,-0.3215796649,0.3989891112,-0.0260411445,-0.1408991367,-0.1086198837,0.393196404,-0.3729823232,0.1213622913,-0.1435122341,0.4954285622,-0.2756084502,0.6325696707,-0.1103373617,-0.3016379774,-0.6230740547,0.0143794967,-0.2379069328,-0.1937564462,-0.0749492943,-0.1897688359,-0.0972293913,0.1173614189,-0.0211160742,-0.2446141094,-0.0903439969,0.1673309207,0.2581494153,-0.3689734042,0.1945202947,0.2363354117,-0.028643772,0.2357154489,-0.1074637994,0.0941565037,-0.2790004313,-0.1535018682,0.0162235945,0.0073205931,0.0916679874,-0.0805359781,-0.294167161,0.1308050901,0.2717175782,-0.0983364508,-0.4141730666,0.1455357671,0.2113202214,-0.5678485632,-0.2634512186,0.0058454731,-0.2490126491,0.0622973107,-0.0128140599,0.1382781118,-0.0861099511,0.2460179478,0.5355675817,-0.0939638242,0.0122149726,0.2041075528,-0.0697063506,-0.3262436092,0.1600855589,0.0908686742,-0.1199678779,-0.1862561256,0.6869305968,-0.0673871115,-0.1712520719,0.128244251,0.0116519677,-0.3052785397,-0.3731777966,-0.1423394084,0.3646332622,0.005840688,0.2755899727,-0.1635782123,-0.3519570827,0.3817750812,0.1399418712,-0.0735855028,0.0975961611,0.001163895,-0.0712351203,0.3171833456,-0.2154521644,0.1800551862,-0.0462435782,0.0269218441,0.0353401937,-0.2214651257,0.0625348091,-0.1180963591,-0.0771713927,-0.0593445525,-0.3488052487,-0.1104722992,-0.0980194733,0.2203945965,0.0336769558,0.0261243787,-0.150252834,-0.2695866227,-0.11599534,-0.078188926,0.2673688233,0.1409072876,0.3275106549,-0.4315177202,-0.1666244268,-0.3083327711,-0.0145831946,0.1815347224,0.2680336833,0.3731192946,0.08237385,0.0242914446,-0.0004601251,-0.1893420219,0.0109661967,0.5105850101,0.6762283444,0.1111970544,0.3195176423,-0.0521444976,-0.0073331855,0.2021808177,0.1085489392,0.2828254402,-0.2943923473,0.2507990003,-0.0147381742,0.0488657914,-0.0176654588,0.1149610952,0.4137041271,0.0498753414,-0.0360794254,0.3354976475,0.0046700179,-0.1078227684,-0.1293975413,0.1676985919,0.6167354584,0.1922197789,-0.2577420771,-0.1376618147,-0.089029111,-0.1640285105,0.3908017874,0.2801336646,0.2462409139,0.0316047147,-0.0221294407,0.1565683484,-0.0305095464,0.0273377858,0.3041538,0.1134856492,-0.2019307762,0.3264491856,0.3181265593,-0.0799670964,-0.0348047689,0.6973232627,0.3104599416,-0.1900941133,-0.1182084456,0.2160139084,0.0718547776,-0.1131831035,-0.1479928941,-0.258043766,-0.1297799647,0.1194079369,-0.2619001269,-0.4273075163,0.2420234084,0.3363099992,0.3412489593,-0.0850116685,0.2789873481,0.1681714356,0.2500574589,-0.1632449925,0.3264158368,0.3324593008,0.1959136575,0.1672491431,0.2142611891,0.1272670925,-0.0537584275,-0.3080976605,0.4086636305,-0.4157254696,-0.0443495512,0.0032410235,0.0797544643,0.1889351457,-0.3090522587,0.1499455124,-0.0006962738,-0.0417117029,-0.1305248737,0.210966751,-0.3677361012,-0.0813946277,-0.3251884282,-0.2253836542,-0.2650460601,-0.1199197918,-0.2188532948,-0.5024661422,-0.108942993,0.0995794982,-0.1453827471,0.2456161678,-0.04667807,0.0203641895,-0.0840874091,0.2370767593,-0.0457655266,-0.0457704924,-0.0423955247,-0.1394757032,-0.2250744849,0.1166563854,-0.2081105411,0.1229943484,-0.2126381695,0.4581338167,0.1515930891,0.6246837974,-0.2316151261,0.0611712188,-0.3553406894,-0.1483782083,-0.4707787335,0.119021751,0.0865825117,0.5708333254,0.2162055671,-0.1196238846,0.2779859006,0.1111526564,-0.145048216,-0.1551150084,0.2132126093,-0.3833232224,-0.4324444532,0.0267935898,-0.0024119725,0.0714242011,-0.0178702716,-0.2116694301,-0.1384190023,-0.0719527677,-0.0324094221,0.1772026718,0.3590931892,0.0158050712,-0.3466786146,-0.3906984627,0.0637737066,0.10832344,-0.2145219296,-0.1224574894,-0.1497722119,-0.2969071865,-0.1506321579,0.228689149,-0.0643118769,0.3435363173,-0.3928129673,0.002390492,-0.1474832892,-0.0695451126,0.5182189941,-0.0942520276,-0.3023611903,0.1273453683,0.2690095603,0.1394510716,-0.0395422056,-0.3600625992,0.2956174612,0.0815311447,0.0148546714,-0.1936296076,0.152564615,-0.2224375755,-0.3109474182,-0.1281321794,0.1289637983,0.0271986779,-0.1219877377,-0.2279237807,-0.1971742809]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/324","title":"Error when calculating glue score","comments":"The glue metric for cola is a metric for classification. It expects label ids as integers as inputs.","body":"I was trying glue score along with other metrics here. But glue gives me this error;\r\n\r\n```\r\nimport nlp\r\nglue_metric = nlp.load_metric('glue',name=\"cola\")\r\n\r\nglue_score = glue_metric.compute(predictions, references)\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-8-b9210a524504> in <module>()\r\n----> 1 glue_score = glue_metric.compute(predictions, references)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 191 \"\"\"\r\n 192 if predictions is not None:\r\n--> 193 self.add_batch(predictions=predictions, references=references)\r\n 194 self.finalize(timeout=timeout)\r\n 195 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in add_batch(self, predictions, references, **kwargs)\r\n 207 if self.writer is None:\r\n 208 self._init_writer()\r\n--> 209 self.writer.write_batch(batch)\r\n 210 \r\n 211 def add(self, prediction=None, reference=None, **kwargs):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 155 if self.pa_writer is None:\r\n 156 self._build_writer(pa_table=pa.Table.from_pydict(batch_examples))\r\n--> 157 pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n 158 if writer_batch_size is None:\r\n 159 writer_batch_size = self.writer_batch_size\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/types.pxi in __iter__()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\nTypeError: an integer is required (got type str)\r\n```\r\nI'm not sure whether I'm doing this wrong or whether it's an issue. I would like to know a workaround. Thank you.","comment_length":18,"text":"Error when calculating glue score\nI was trying glue score along with other metrics here. But glue gives me this error;\r\n\r\n```\r\nimport nlp\r\nglue_metric = nlp.load_metric('glue',name=\"cola\")\r\n\r\nglue_score = glue_metric.compute(predictions, references)\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-8-b9210a524504> in <module>()\r\n----> 1 glue_score = glue_metric.compute(predictions, references)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 191 \"\"\"\r\n 192 if predictions is not None:\r\n--> 193 self.add_batch(predictions=predictions, references=references)\r\n 194 self.finalize(timeout=timeout)\r\n 195 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in add_batch(self, predictions, references, **kwargs)\r\n 207 if self.writer is None:\r\n 208 self._init_writer()\r\n--> 209 self.writer.write_batch(batch)\r\n 210 \r\n 211 def add(self, prediction=None, reference=None, **kwargs):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 155 if self.pa_writer is None:\r\n 156 self._build_writer(pa_table=pa.Table.from_pydict(batch_examples))\r\n--> 157 pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n 158 if writer_batch_size is None:\r\n 159 writer_batch_size = self.writer_batch_size\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/types.pxi in __iter__()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\nTypeError: an integer is required (got type str)\r\n```\r\nI'm not sure whether I'm doing this wrong or whether it's an issue. I would like to know a workaround. Thank you.\nThe glue metric for cola is a metric for classification. It expects label ids as integers as inputs.","embeddings":[-0.1694530249,-0.2395812869,-0.0304733198,0.1611590832,0.236740306,-0.0491650477,0.1073005199,0.3623578846,0.4330275059,0.0516285449,-0.3168233335,0.2763474584,-0.0798071995,-0.0241100863,0.1026181132,-0.2831492722,-0.1162880808,0.1604485661,-0.2202495039,0.3102566004,-0.324510783,0.2713671029,-0.1858564019,0.3197874725,-0.1767067164,-0.0903876349,0.0204841644,-0.196170941,-0.3263719678,-0.362935394,0.2481628209,-0.1168349832,0.02158558,0.1766418815,-0.0001102634,-0.0120561915,0.458953023,0.0023093964,0.1588456184,-0.1421633959,-0.3889809847,-0.083480984,0.0444146022,-0.2361500263,0.1346656531,0.2275212109,0.0206140038,-0.1826084852,0.2193325311,0.2709797919,0.2774609029,0.3871142864,0.3126004338,-0.1989937872,0.1724911779,-0.2082325965,0.0265877806,0.8870652914,-0.327216506,-0.4651894271,-0.0078018066,0.2069443315,-0.1057071611,0.2180079669,0.2240158021,0.0521962196,0.3716999292,-0.3202626705,-0.1326412559,-0.0713140443,-0.2181349844,-0.2285419106,-0.1499314606,0.0416573919,0.0471167676,-0.3763104677,0.0972394794,-0.0726784766,0.0581606068,-0.2293186486,-0.0549694002,0.1402760744,-0.1144136488,0.2752963901,0.2400909662,0.3214469552,-0.1981527656,0.296900332,0.4147271514,0.2154424191,-0.1390433908,0.0435455628,-0.2401257157,0.1653348356,-0.2926098704,-0.0423279516,-0.0104532884,0.1914083958,0.0239265561,-0.2354671657,0.0487309434,-0.2080052644,0.4857490361,0.3427954614,0.0602983236,0.3389247358,0.3534550667,0.1564653218,0.127425313,0.2417826504,0.002334035,-0.1694935113,0.2388441265,-0.3881970942,0.3655791283,0.2413906008,0.1085240468,0.0770600438,-0.4262593985,-0.1771688163,-0.0480151996,-0.0049847378,0.036921978,0.0718623921,-0.0187913645,0.0006606558,0.1049904004,0.1906935126,-0.1910691857,0.057039503,-0.1342848837,0.4053150415,-0.3368168175,0.1215166077,0.1513898671,-0.033291731,0.2043572962,-0.1073043272,0.1692515314,-0.1872263104,0.0212524477,-0.2450529933,-0.0659325495,0.1954428554,0.0480438098,0.0028141201,0.3261824846,-0.5611881614,-0.1635352671,-0.2155396938,-0.2283369601,-0.3000667095,-0.1469439864,0.2351692468,-0.1796001494,0.0300378092,-0.4003524482,-0.18054685,0.237726599,0.0449476801,0.1126893461,-0.2676103711,-0.2484904975,-0.2921720147,0.2016495019,0.1388746202,-0.08172521,-0.1147434115,-0.1918641925,-0.1030081287,0.1174745634,0.3144727945,0.1723901331,0.43979159,-0.2319522053,0.1515319943,0.5892598629,-0.6347135305,-0.2369804829,0.1007082835,0.0968757421,-0.2819084525,-0.0695398524,0.0950889289,0.2874577045,0.0750943869,0.4693049788,0.1775257438,-0.1200407296,-0.1910822093,-0.3003584146,0.0345603265,0.0852224529,0.0522735454,0.2045292258,-0.0504421517,0.1096888632,-0.0649215803,-0.1438204199,0.1973367929,0.0194501802,0.0301419217,0.331820786,-0.2376646549,0.0082451282,-0.5078004003,-0.0287532061,0.0765153319,-0.5539197922,0.198527962,0.0962853581,-0.0163019728,-0.5231425762,0.1050595492,-0.0699512661,-0.0393085517,0.2326908261,-0.2064474374,0.3046903312,0.0231801495,-0.0384821072,0.171043694,0.0765673891,0.0847157612,-0.0891413689,0.2145031542,-0.2762120366,-0.4148796499,0.138267681,0.3498590291,0.2727716267,0.0353043973,-0.1092623249,0.4997556508,-0.1424180567,-0.2285176069,0.0557274185,-0.0633073077,-0.0071699293,0.0306306798,-0.0186237693,-0.1132494584,0.1906821728,-0.029351959,0.3210164309,0.1958460063,0.0103191854,0.2974856794,-0.1180674285,0.327396512,-0.1225315109,-0.0584307276,-0.1500223279,-0.1341111809,0.4181139767,-0.297301203,-0.0799402669,-0.1827307642,0.3219168186,-0.2306055278,0.3350412548,-0.0496505089,0.1854770035,-0.0289649013,-0.2190921903,0.107579872,-0.1537142396,0.1282522231,0.5975934267,-0.0049649971,-0.1247712001,0.0250173025,-0.1926951706,-0.0234322157,0.1754012406,0.0674831867,0.0308192037,0.1081281975,0.1688194871,-0.0281777438,-0.0380484313,0.1020587161,-0.2313185483,0.2311968207,-0.322055459,-0.0178321414,-0.0451471135,-0.4319617748,-0.1677066386,-0.1099136993,-0.2902896106,-0.6119655967,-0.057234671,0.0513690375,0.0873179585,0.1892627627,0.322621733,0.0651795939,0.1116792187,0.0968883038,0.052267164,-0.2201310247,-0.0294674635,0.0525177605,0.0137362741,0.2037277669,0.4438380897,-0.1036847085,-0.0145902503,-0.1050944924,-0.3087322116,0.0173460748,-0.284208715,0.6033970118,0.4132930934,-0.1238139421,-0.0578880683,-0.3781408966,0.4658437669,-0.4250400364,0.0171624403,0.2245608568,-0.2028575689,-0.0457102694,-0.1568728834,0.0534092486,-0.2870553434,-0.3602614999,0.3244150281,0.1362109035,0.0870978311,0.0247128438,-0.2230973095,0.378121376,-0.1902915835,-0.1985457391,-0.1898382753,-0.0287204385,0.3248326778,-0.2061116695,-0.3413917124,-0.088632822,-0.0667745844,0.3206015527,-0.2410281748,-0.2447149307,-0.5733760595,-0.1851422489,0.0286152307,-0.1296885163,0.1928846389,0.0440016761,-0.1671079099,-0.064908646,-0.0949185565,-0.2339582145,0.1628335267,0.2192826271,0.092607379,-0.1301689148,0.3421754539,0.0853842348,0.3098371029,0.1035997644,-0.0924218148,0.2702838182,0.0431021787,0.2309778929,0.0065481076,-0.4910922647,0.1871620119,-0.1212332994,0.3261258006,-0.116968669,-0.0228460543,-0.0533289127,-0.2160975337,-0.142928794,-0.1235205829,-0.221789971,0.0744251758,0.0152623169,0.1742431521,0.0593148656,-0.2170495242,-0.2897800803,-0.2181468904,0.1944755912,0.2156715095,0.0103608659,-0.0712358952,-0.0382272005,-0.2960558534,-0.3888400793,0.4718077183,0.1434947699,0.3303483725,-0.1176615953,0.2998418212,-0.1516719311,0.1147296652,0.2170622349,-0.3560436964,0.0797637701,0.289891392,0.1962627769,-0.5538406968,-0.1381310374,-0.4275742471,0.3352845907,0.5549218655,0.4441614449,-0.2514420748,0.0707286522,0.4990741014,-0.0577811487,-0.0494861081,-0.4405652881,-0.4276404083,-0.1975727975,-0.134593159,-0.0735190138,0.1434215605,0.2061714977,0.3873663843,0.0714507997,-0.0928121805,-0.2010710984,0.0717374757,-0.2025193423,0.0308996141,0.0319244713,0.0896566287,-0.2613339722,-0.0328474119,-0.5163642764,-0.1181045771,-0.0492085591,-0.0583241209,0.1162236407,0.006550475,0.3921333551,-0.1031682417,-0.1336190701,0.232488364,-0.2945844829,0.4098879993,0.04337411,0.0918040425,0.344673574,0.2249441594,-0.1188238487,0.1059011966,0.4642900229,0.2396440357,0.0409607179,0.3867133856,0.2303498238,-0.2547430396,0.2100529075,0.1255099028,0.7316866517,-0.0363749973,-0.2081698477,0.2821577787,-0.1958425492,0.6601784229,0.2018424273,0.2584860623,-0.2552503347,-0.240117535,0.0362041257,-0.1886274219,0.1207558587,-0.2527753115,-0.4338819087,0.4902597964,0.0216529164,0.1240034103,-0.1225317046,0.1950635314,-0.1712214351,-0.1687160432,-0.0379025266,0.1914608777,-0.333596915,0.1056694016,-0.0188455451,-0.0623411946,0.1282528341,-0.07512784,-0.2535679638,-0.1657012552,-0.3714176118,0.1900477111,0.2342687398,0.0006985788,0.0667436123,-0.256436944,0.2165791541,0.1683935821,-0.0986778289,-0.0945844427,-0.0107043562,-0.2469310313,0.1475564241,0.1228892058,-0.1027315333,-0.1388566792,-0.1170235425,0.0286960378,0.0947821066,-0.260346204,0.2329951674,0.111068137,0.0089348322,-0.2510281205,-0.2027098089,-0.1664216369,-0.0228838976,-0.0703513324,0.1712865978,0.004673244,-0.1367912591,0.2804692388,-0.065941751,-0.1624079347,0.0864233673,0.2396939844,-0.0793856904,-0.0198109224,0.1160599962,-0.1933647692,-0.2555323839,-0.075857617,-0.153219983,-0.2182993144,-0.2836368978,0.1234618053,-0.119727917,-0.0283316765,-0.1401845962,0.1114247739,0.4184065163,0.2592692673,0.0969694182,-0.4761908352,-0.1332361549,0.277235359,-0.0170346592,-0.1239774302,-0.1470875591,0.322176218,0.1539585143,0.3371094167,-0.340765208,-0.0861637741,-0.3368872106,0.1727570742,0.5577085614,-0.3882571459,0.03208258,0.0896285996,0.1925500333,0.2407439053,-0.1684650928,-0.2781485617,-0.0453970693,0.088414669,-0.0080468161,-0.1094610766,0.1246851608,0.0089911344,-0.1972405463,-0.0705862939,-0.2875487208,-0.1359250695,-0.1910493523,0.0623337924,-0.0399488844,0.1979519427,0.3852122426,0.3219557405,-0.0956423357,0.1455473304,-0.1070941314,0.1637578756,0.0806565061,0.0923649892,-0.1439367384,-0.1045835987,-0.1080604568,-0.0023589039,0.2476173043,-0.217047736,-0.2661941051,-0.1419357508,0.3768404126,-0.0228843149,-0.035823565,-0.0351780467,-0.1800120175,0.1641356945,-0.1655005217,0.1204651296,0.2528325021,0.2045561373,0.1759787649,0.1867296845,-0.0148430066,0.0641106591,0.0059583378,0.0965781733,0.3417929709,-0.1481131911,-0.1568119675,-0.3146477044,0.2342946976,-0.0372508503,0.3855211139,-0.1936479062,0.1915557832,-0.1926371604,0.0036960652,0.1574132144,-0.0652413517,0.0030420481,0.2391289175,0.1975857019,-0.1815176755,0.2595298886,0.1466700733,0.1041979119,-0.0200797468,0.516079843,-0.3852641582,-0.0782804415,0.0359458067,0.1538161933,0.0529209971,0.0623459443,0.0045074518,-0.0066839065,-0.3290788531,-0.155305177,0.0652153343,0.1220391914,0.3950314224,0.1611405611,0.0191938709,-0.048688747,-0.2124672532,0.3583585918,0.1800563484,-0.3143035173,0.1221407503,0.1270214617,0.2304644138,-0.2367886007,0.6592804193,0.5281543732,0.3133310378,-0.3758439124,0.1271784157,-0.0340991281,-0.1760197282,0.2069451958,0.1847516,0.3671351671,0.0971979424,0.2671824694,0.2085577548,-0.2633956969,-0.0485939421,-0.0723655745,-0.0130673712,-0.1685698628,0.4525218606,-0.2508077323,-0.2331060469,-0.1536246091,0.0573425516,-0.4240994155,0.1701419502,0.035439305,-0.022380935,0.0756966844,-0.2836797237,0.0981923938,-0.0239031259,0.5166293383,0.1721064448,0.5441957116,-0.5894020796,-0.2243749946,-0.4315952063,0.0422833636,-0.0980620757,-0.0537153855,0.0169731099,0.0522209108,-0.2406836897,0.3361567855,0.1721651852,0.498277992,-0.1158373654,-0.282445997,-0.302873075,-0.0565930903,0.0526053309,-0.4606128931,0.2394749522,-0.2231238037,0.092904076,-0.100292787,0.14038302,-0.0083689336,-0.2390481085,0.2770601213,0.5875069499,0.0879226848,0.2794111669,0.5532605052,-0.0172410365,-0.0034356373,0.0521484464,-0.3737610281,-0.1366231889,0.0499606021,0.0168516561,0.2484907508,-0.087572068,-0.0866054073,-0.3632865548,0.5056182146,-0.1624269336,0.0168920252,-0.0326465443,0.0752586946,0.1267989278,0.0208195522,-0.0418231077,0.0828972608,-0.0186353456,0.3271137178,-0.2005010694,-0.2827176452,0.3735974729,-0.4797991216,-0.3367664516,0.0954478085,0.2530971169,0.2006332725,-0.0931799114,-0.6303013563,-0.2510200441,0.3387457132,0.0636691377,-0.1801239103,0.3238251805,-0.1578463018,0.1805446595,0.1582315266,0.0388474241,0.1575522572,-0.308976233,0.0498125739,-0.2299158275]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/324","title":"Error when calculating glue score","comments":"I want to evaluate a sentence pair whether they are semantically equivalent, so I used MRPC and it gives the same error, does that mean we have to encode the sentences and parse as input?\r\n\r\nusing BertTokenizer;\r\n```\r\nencoded_reference=tokenizer.encode(reference, add_special_tokens=False)\r\nencoded_prediction=tokenizer.encode(prediction, add_special_tokens=False)\r\n```\r\n\r\n`glue_score = glue_metric.compute(encoded_prediction, encoded_reference)`\r\n```\r\n\r\nValueError Traceback (most recent call last)\r\n<ipython-input-9-4c3a3ce7b583> in <module>()\r\n----> 1 glue_score = glue_metric.compute(encoded_prediction, encoded_reference)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 198 predictions = self.data[\"predictions\"]\r\n 199 references = self.data[\"references\"]\r\n--> 200 output = self._compute(predictions=predictions, references=references, **metrics_kwargs)\r\n 201 return output\r\n 202 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metrics\/glue\/27b1bc63e520833054bd0d7a8d0bc7f6aab84cc9eed1b576e98c806f9466d302\/glue.py in _compute(self, predictions, references)\r\n 101 return pearson_and_spearman(predictions, references)\r\n 102 elif self.config_name in [\"mrpc\", \"qqp\"]:\r\n--> 103 return acc_and_f1(predictions, references)\r\n 104 elif self.config_name in [\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"]:\r\n 105 return {\"accuracy\": simple_accuracy(predictions, references)}\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metrics\/glue\/27b1bc63e520833054bd0d7a8d0bc7f6aab84cc9eed1b576e98c806f9466d302\/glue.py in acc_and_f1(preds, labels)\r\n 60 def acc_and_f1(preds, labels):\r\n 61 acc = simple_accuracy(preds, labels)\r\n---> 62 f1 = f1_score(y_true=labels, y_pred=preds)\r\n 63 return {\r\n 64 \"accuracy\": acc,\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/sklearn\/metrics\/_classification.py in f1_score(y_true, y_pred, labels, pos_label, average, sample_weight, zero_division)\r\n 1097 pos_label=pos_label, average=average,\r\n 1098 sample_weight=sample_weight,\r\n-> 1099 zero_division=zero_division)\r\n 1100 \r\n 1101 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/sklearn\/metrics\/_classification.py in fbeta_score(y_true, y_pred, beta, labels, pos_label, average, sample_weight, zero_division)\r\n 1224 warn_for=('f-score',),\r\n 1225 sample_weight=sample_weight,\r\n-> 1226 zero_division=zero_division)\r\n 1227 return f\r\n 1228 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/sklearn\/metrics\/_classification.py in precision_recall_fscore_support(y_true, y_pred, beta, labels, pos_label, average, warn_for, sample_weight, zero_division)\r\n 1482 raise ValueError(\"beta should be >=0 in the F-beta score\")\r\n 1483 labels = _check_set_wise_labels(y_true, y_pred, average, labels,\r\n-> 1484 pos_label)\r\n 1485 \r\n 1486 # Calculate tp_sum, pred_sum, true_sum ###\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/sklearn\/metrics\/_classification.py in _check_set_wise_labels(y_true, y_pred, average, labels, pos_label)\r\n 1314 raise ValueError(\"Target is %s but average='binary'. Please \"\r\n 1315 \"choose another average setting, one of %r.\"\r\n-> 1316 % (y_type, average_options))\r\n 1317 elif pos_label not in (None, 1):\r\n 1318 warnings.warn(\"Note that pos_label (set to %r) is ignored when \"\r\n\r\nValueError: Target is multiclass but average='binary'. Please choose another average setting, one of [None, 'micro', 'macro', 'weighted'].\r\n\r\n```","body":"I was trying glue score along with other metrics here. But glue gives me this error;\r\n\r\n```\r\nimport nlp\r\nglue_metric = nlp.load_metric('glue',name=\"cola\")\r\n\r\nglue_score = glue_metric.compute(predictions, references)\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-8-b9210a524504> in <module>()\r\n----> 1 glue_score = glue_metric.compute(predictions, references)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 191 \"\"\"\r\n 192 if predictions is not None:\r\n--> 193 self.add_batch(predictions=predictions, references=references)\r\n 194 self.finalize(timeout=timeout)\r\n 195 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in add_batch(self, predictions, references, **kwargs)\r\n 207 if self.writer is None:\r\n 208 self._init_writer()\r\n--> 209 self.writer.write_batch(batch)\r\n 210 \r\n 211 def add(self, prediction=None, reference=None, **kwargs):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 155 if self.pa_writer is None:\r\n 156 self._build_writer(pa_table=pa.Table.from_pydict(batch_examples))\r\n--> 157 pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n 158 if writer_batch_size is None:\r\n 159 writer_batch_size = self.writer_batch_size\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/types.pxi in __iter__()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\nTypeError: an integer is required (got type str)\r\n```\r\nI'm not sure whether I'm doing this wrong or whether it's an issue. I would like to know a workaround. Thank you.","comment_length":297,"text":"Error when calculating glue score\nI was trying glue score along with other metrics here. But glue gives me this error;\r\n\r\n```\r\nimport nlp\r\nglue_metric = nlp.load_metric('glue',name=\"cola\")\r\n\r\nglue_score = glue_metric.compute(predictions, references)\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-8-b9210a524504> in <module>()\r\n----> 1 glue_score = glue_metric.compute(predictions, references)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 191 \"\"\"\r\n 192 if predictions is not None:\r\n--> 193 self.add_batch(predictions=predictions, references=references)\r\n 194 self.finalize(timeout=timeout)\r\n 195 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in add_batch(self, predictions, references, **kwargs)\r\n 207 if self.writer is None:\r\n 208 self._init_writer()\r\n--> 209 self.writer.write_batch(batch)\r\n 210 \r\n 211 def add(self, prediction=None, reference=None, **kwargs):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 155 if self.pa_writer is None:\r\n 156 self._build_writer(pa_table=pa.Table.from_pydict(batch_examples))\r\n--> 157 pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n 158 if writer_batch_size is None:\r\n 159 writer_batch_size = self.writer_batch_size\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/types.pxi in __iter__()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\nTypeError: an integer is required (got type str)\r\n```\r\nI'm not sure whether I'm doing this wrong or whether it's an issue. I would like to know a workaround. Thank you.\nI want to evaluate a sentence pair whether they are semantically equivalent, so I used MRPC and it gives the same error, does that mean we have to encode the sentences and parse as input?\r\n\r\nusing BertTokenizer;\r\n```\r\nencoded_reference=tokenizer.encode(reference, add_special_tokens=False)\r\nencoded_prediction=tokenizer.encode(prediction, add_special_tokens=False)\r\n```\r\n\r\n`glue_score = glue_metric.compute(encoded_prediction, encoded_reference)`\r\n```\r\n\r\nValueError Traceback (most recent call last)\r\n<ipython-input-9-4c3a3ce7b583> in <module>()\r\n----> 1 glue_score = glue_metric.compute(encoded_prediction, encoded_reference)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 198 predictions = self.data[\"predictions\"]\r\n 199 references = self.data[\"references\"]\r\n--> 200 output = self._compute(predictions=predictions, references=references, **metrics_kwargs)\r\n 201 return output\r\n 202 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metrics\/glue\/27b1bc63e520833054bd0d7a8d0bc7f6aab84cc9eed1b576e98c806f9466d302\/glue.py in _compute(self, predictions, references)\r\n 101 return pearson_and_spearman(predictions, references)\r\n 102 elif self.config_name in [\"mrpc\", \"qqp\"]:\r\n--> 103 return acc_and_f1(predictions, references)\r\n 104 elif self.config_name in [\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"]:\r\n 105 return {\"accuracy\": simple_accuracy(predictions, references)}\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metrics\/glue\/27b1bc63e520833054bd0d7a8d0bc7f6aab84cc9eed1b576e98c806f9466d302\/glue.py in acc_and_f1(preds, labels)\r\n 60 def acc_and_f1(preds, labels):\r\n 61 acc = simple_accuracy(preds, labels)\r\n---> 62 f1 = f1_score(y_true=labels, y_pred=preds)\r\n 63 return {\r\n 64 \"accuracy\": acc,\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/sklearn\/metrics\/_classification.py in f1_score(y_true, y_pred, labels, pos_label, average, sample_weight, zero_division)\r\n 1097 pos_label=pos_label, average=average,\r\n 1098 sample_weight=sample_weight,\r\n-> 1099 zero_division=zero_division)\r\n 1100 \r\n 1101 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/sklearn\/metrics\/_classification.py in fbeta_score(y_true, y_pred, beta, labels, pos_label, average, sample_weight, zero_division)\r\n 1224 warn_for=('f-score',),\r\n 1225 sample_weight=sample_weight,\r\n-> 1226 zero_division=zero_division)\r\n 1227 return f\r\n 1228 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/sklearn\/metrics\/_classification.py in precision_recall_fscore_support(y_true, y_pred, beta, labels, pos_label, average, warn_for, sample_weight, zero_division)\r\n 1482 raise ValueError(\"beta should be >=0 in the F-beta score\")\r\n 1483 labels = _check_set_wise_labels(y_true, y_pred, average, labels,\r\n-> 1484 pos_label)\r\n 1485 \r\n 1486 # Calculate tp_sum, pred_sum, true_sum ###\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/sklearn\/metrics\/_classification.py in _check_set_wise_labels(y_true, y_pred, average, labels, pos_label)\r\n 1314 raise ValueError(\"Target is %s but average='binary'. Please \"\r\n 1315 \"choose another average setting, one of %r.\"\r\n-> 1316 % (y_type, average_options))\r\n 1317 elif pos_label not in (None, 1):\r\n 1318 warnings.warn(\"Note that pos_label (set to %r) is ignored when \"\r\n\r\nValueError: Target is multiclass but average='binary'. Please choose another average setting, one of [None, 'micro', 'macro', 'weighted'].\r\n\r\n```","embeddings":[-0.1694530249,-0.2395812869,-0.0304733198,0.1611590832,0.236740306,-0.0491650477,0.1073005199,0.3623578846,0.4330275059,0.0516285449,-0.3168233335,0.2763474584,-0.0798071995,-0.0241100863,0.1026181132,-0.2831492722,-0.1162880808,0.1604485661,-0.2202495039,0.3102566004,-0.324510783,0.2713671029,-0.1858564019,0.3197874725,-0.1767067164,-0.0903876349,0.0204841644,-0.196170941,-0.3263719678,-0.362935394,0.2481628209,-0.1168349832,0.02158558,0.1766418815,-0.0001102634,-0.0120561915,0.458953023,0.0023093964,0.1588456184,-0.1421633959,-0.3889809847,-0.083480984,0.0444146022,-0.2361500263,0.1346656531,0.2275212109,0.0206140038,-0.1826084852,0.2193325311,0.2709797919,0.2774609029,0.3871142864,0.3126004338,-0.1989937872,0.1724911779,-0.2082325965,0.0265877806,0.8870652914,-0.327216506,-0.4651894271,-0.0078018066,0.2069443315,-0.1057071611,0.2180079669,0.2240158021,0.0521962196,0.3716999292,-0.3202626705,-0.1326412559,-0.0713140443,-0.2181349844,-0.2285419106,-0.1499314606,0.0416573919,0.0471167676,-0.3763104677,0.0972394794,-0.0726784766,0.0581606068,-0.2293186486,-0.0549694002,0.1402760744,-0.1144136488,0.2752963901,0.2400909662,0.3214469552,-0.1981527656,0.296900332,0.4147271514,0.2154424191,-0.1390433908,0.0435455628,-0.2401257157,0.1653348356,-0.2926098704,-0.0423279516,-0.0104532884,0.1914083958,0.0239265561,-0.2354671657,0.0487309434,-0.2080052644,0.4857490361,0.3427954614,0.0602983236,0.3389247358,0.3534550667,0.1564653218,0.127425313,0.2417826504,0.002334035,-0.1694935113,0.2388441265,-0.3881970942,0.3655791283,0.2413906008,0.1085240468,0.0770600438,-0.4262593985,-0.1771688163,-0.0480151996,-0.0049847378,0.036921978,0.0718623921,-0.0187913645,0.0006606558,0.1049904004,0.1906935126,-0.1910691857,0.057039503,-0.1342848837,0.4053150415,-0.3368168175,0.1215166077,0.1513898671,-0.033291731,0.2043572962,-0.1073043272,0.1692515314,-0.1872263104,0.0212524477,-0.2450529933,-0.0659325495,0.1954428554,0.0480438098,0.0028141201,0.3261824846,-0.5611881614,-0.1635352671,-0.2155396938,-0.2283369601,-0.3000667095,-0.1469439864,0.2351692468,-0.1796001494,0.0300378092,-0.4003524482,-0.18054685,0.237726599,0.0449476801,0.1126893461,-0.2676103711,-0.2484904975,-0.2921720147,0.2016495019,0.1388746202,-0.08172521,-0.1147434115,-0.1918641925,-0.1030081287,0.1174745634,0.3144727945,0.1723901331,0.43979159,-0.2319522053,0.1515319943,0.5892598629,-0.6347135305,-0.2369804829,0.1007082835,0.0968757421,-0.2819084525,-0.0695398524,0.0950889289,0.2874577045,0.0750943869,0.4693049788,0.1775257438,-0.1200407296,-0.1910822093,-0.3003584146,0.0345603265,0.0852224529,0.0522735454,0.2045292258,-0.0504421517,0.1096888632,-0.0649215803,-0.1438204199,0.1973367929,0.0194501802,0.0301419217,0.331820786,-0.2376646549,0.0082451282,-0.5078004003,-0.0287532061,0.0765153319,-0.5539197922,0.198527962,0.0962853581,-0.0163019728,-0.5231425762,0.1050595492,-0.0699512661,-0.0393085517,0.2326908261,-0.2064474374,0.3046903312,0.0231801495,-0.0384821072,0.171043694,0.0765673891,0.0847157612,-0.0891413689,0.2145031542,-0.2762120366,-0.4148796499,0.138267681,0.3498590291,0.2727716267,0.0353043973,-0.1092623249,0.4997556508,-0.1424180567,-0.2285176069,0.0557274185,-0.0633073077,-0.0071699293,0.0306306798,-0.0186237693,-0.1132494584,0.1906821728,-0.029351959,0.3210164309,0.1958460063,0.0103191854,0.2974856794,-0.1180674285,0.327396512,-0.1225315109,-0.0584307276,-0.1500223279,-0.1341111809,0.4181139767,-0.297301203,-0.0799402669,-0.1827307642,0.3219168186,-0.2306055278,0.3350412548,-0.0496505089,0.1854770035,-0.0289649013,-0.2190921903,0.107579872,-0.1537142396,0.1282522231,0.5975934267,-0.0049649971,-0.1247712001,0.0250173025,-0.1926951706,-0.0234322157,0.1754012406,0.0674831867,0.0308192037,0.1081281975,0.1688194871,-0.0281777438,-0.0380484313,0.1020587161,-0.2313185483,0.2311968207,-0.322055459,-0.0178321414,-0.0451471135,-0.4319617748,-0.1677066386,-0.1099136993,-0.2902896106,-0.6119655967,-0.057234671,0.0513690375,0.0873179585,0.1892627627,0.322621733,0.0651795939,0.1116792187,0.0968883038,0.052267164,-0.2201310247,-0.0294674635,0.0525177605,0.0137362741,0.2037277669,0.4438380897,-0.1036847085,-0.0145902503,-0.1050944924,-0.3087322116,0.0173460748,-0.284208715,0.6033970118,0.4132930934,-0.1238139421,-0.0578880683,-0.3781408966,0.4658437669,-0.4250400364,0.0171624403,0.2245608568,-0.2028575689,-0.0457102694,-0.1568728834,0.0534092486,-0.2870553434,-0.3602614999,0.3244150281,0.1362109035,0.0870978311,0.0247128438,-0.2230973095,0.378121376,-0.1902915835,-0.1985457391,-0.1898382753,-0.0287204385,0.3248326778,-0.2061116695,-0.3413917124,-0.088632822,-0.0667745844,0.3206015527,-0.2410281748,-0.2447149307,-0.5733760595,-0.1851422489,0.0286152307,-0.1296885163,0.1928846389,0.0440016761,-0.1671079099,-0.064908646,-0.0949185565,-0.2339582145,0.1628335267,0.2192826271,0.092607379,-0.1301689148,0.3421754539,0.0853842348,0.3098371029,0.1035997644,-0.0924218148,0.2702838182,0.0431021787,0.2309778929,0.0065481076,-0.4910922647,0.1871620119,-0.1212332994,0.3261258006,-0.116968669,-0.0228460543,-0.0533289127,-0.2160975337,-0.142928794,-0.1235205829,-0.221789971,0.0744251758,0.0152623169,0.1742431521,0.0593148656,-0.2170495242,-0.2897800803,-0.2181468904,0.1944755912,0.2156715095,0.0103608659,-0.0712358952,-0.0382272005,-0.2960558534,-0.3888400793,0.4718077183,0.1434947699,0.3303483725,-0.1176615953,0.2998418212,-0.1516719311,0.1147296652,0.2170622349,-0.3560436964,0.0797637701,0.289891392,0.1962627769,-0.5538406968,-0.1381310374,-0.4275742471,0.3352845907,0.5549218655,0.4441614449,-0.2514420748,0.0707286522,0.4990741014,-0.0577811487,-0.0494861081,-0.4405652881,-0.4276404083,-0.1975727975,-0.134593159,-0.0735190138,0.1434215605,0.2061714977,0.3873663843,0.0714507997,-0.0928121805,-0.2010710984,0.0717374757,-0.2025193423,0.0308996141,0.0319244713,0.0896566287,-0.2613339722,-0.0328474119,-0.5163642764,-0.1181045771,-0.0492085591,-0.0583241209,0.1162236407,0.006550475,0.3921333551,-0.1031682417,-0.1336190701,0.232488364,-0.2945844829,0.4098879993,0.04337411,0.0918040425,0.344673574,0.2249441594,-0.1188238487,0.1059011966,0.4642900229,0.2396440357,0.0409607179,0.3867133856,0.2303498238,-0.2547430396,0.2100529075,0.1255099028,0.7316866517,-0.0363749973,-0.2081698477,0.2821577787,-0.1958425492,0.6601784229,0.2018424273,0.2584860623,-0.2552503347,-0.240117535,0.0362041257,-0.1886274219,0.1207558587,-0.2527753115,-0.4338819087,0.4902597964,0.0216529164,0.1240034103,-0.1225317046,0.1950635314,-0.1712214351,-0.1687160432,-0.0379025266,0.1914608777,-0.333596915,0.1056694016,-0.0188455451,-0.0623411946,0.1282528341,-0.07512784,-0.2535679638,-0.1657012552,-0.3714176118,0.1900477111,0.2342687398,0.0006985788,0.0667436123,-0.256436944,0.2165791541,0.1683935821,-0.0986778289,-0.0945844427,-0.0107043562,-0.2469310313,0.1475564241,0.1228892058,-0.1027315333,-0.1388566792,-0.1170235425,0.0286960378,0.0947821066,-0.260346204,0.2329951674,0.111068137,0.0089348322,-0.2510281205,-0.2027098089,-0.1664216369,-0.0228838976,-0.0703513324,0.1712865978,0.004673244,-0.1367912591,0.2804692388,-0.065941751,-0.1624079347,0.0864233673,0.2396939844,-0.0793856904,-0.0198109224,0.1160599962,-0.1933647692,-0.2555323839,-0.075857617,-0.153219983,-0.2182993144,-0.2836368978,0.1234618053,-0.119727917,-0.0283316765,-0.1401845962,0.1114247739,0.4184065163,0.2592692673,0.0969694182,-0.4761908352,-0.1332361549,0.277235359,-0.0170346592,-0.1239774302,-0.1470875591,0.322176218,0.1539585143,0.3371094167,-0.340765208,-0.0861637741,-0.3368872106,0.1727570742,0.5577085614,-0.3882571459,0.03208258,0.0896285996,0.1925500333,0.2407439053,-0.1684650928,-0.2781485617,-0.0453970693,0.088414669,-0.0080468161,-0.1094610766,0.1246851608,0.0089911344,-0.1972405463,-0.0705862939,-0.2875487208,-0.1359250695,-0.1910493523,0.0623337924,-0.0399488844,0.1979519427,0.3852122426,0.3219557405,-0.0956423357,0.1455473304,-0.1070941314,0.1637578756,0.0806565061,0.0923649892,-0.1439367384,-0.1045835987,-0.1080604568,-0.0023589039,0.2476173043,-0.217047736,-0.2661941051,-0.1419357508,0.3768404126,-0.0228843149,-0.035823565,-0.0351780467,-0.1800120175,0.1641356945,-0.1655005217,0.1204651296,0.2528325021,0.2045561373,0.1759787649,0.1867296845,-0.0148430066,0.0641106591,0.0059583378,0.0965781733,0.3417929709,-0.1481131911,-0.1568119675,-0.3146477044,0.2342946976,-0.0372508503,0.3855211139,-0.1936479062,0.1915557832,-0.1926371604,0.0036960652,0.1574132144,-0.0652413517,0.0030420481,0.2391289175,0.1975857019,-0.1815176755,0.2595298886,0.1466700733,0.1041979119,-0.0200797468,0.516079843,-0.3852641582,-0.0782804415,0.0359458067,0.1538161933,0.0529209971,0.0623459443,0.0045074518,-0.0066839065,-0.3290788531,-0.155305177,0.0652153343,0.1220391914,0.3950314224,0.1611405611,0.0191938709,-0.048688747,-0.2124672532,0.3583585918,0.1800563484,-0.3143035173,0.1221407503,0.1270214617,0.2304644138,-0.2367886007,0.6592804193,0.5281543732,0.3133310378,-0.3758439124,0.1271784157,-0.0340991281,-0.1760197282,0.2069451958,0.1847516,0.3671351671,0.0971979424,0.2671824694,0.2085577548,-0.2633956969,-0.0485939421,-0.0723655745,-0.0130673712,-0.1685698628,0.4525218606,-0.2508077323,-0.2331060469,-0.1536246091,0.0573425516,-0.4240994155,0.1701419502,0.035439305,-0.022380935,0.0756966844,-0.2836797237,0.0981923938,-0.0239031259,0.5166293383,0.1721064448,0.5441957116,-0.5894020796,-0.2243749946,-0.4315952063,0.0422833636,-0.0980620757,-0.0537153855,0.0169731099,0.0522209108,-0.2406836897,0.3361567855,0.1721651852,0.498277992,-0.1158373654,-0.282445997,-0.302873075,-0.0565930903,0.0526053309,-0.4606128931,0.2394749522,-0.2231238037,0.092904076,-0.100292787,0.14038302,-0.0083689336,-0.2390481085,0.2770601213,0.5875069499,0.0879226848,0.2794111669,0.5532605052,-0.0172410365,-0.0034356373,0.0521484464,-0.3737610281,-0.1366231889,0.0499606021,0.0168516561,0.2484907508,-0.087572068,-0.0866054073,-0.3632865548,0.5056182146,-0.1624269336,0.0168920252,-0.0326465443,0.0752586946,0.1267989278,0.0208195522,-0.0418231077,0.0828972608,-0.0186353456,0.3271137178,-0.2005010694,-0.2827176452,0.3735974729,-0.4797991216,-0.3367664516,0.0954478085,0.2530971169,0.2006332725,-0.0931799114,-0.6303013563,-0.2510200441,0.3387457132,0.0636691377,-0.1801239103,0.3238251805,-0.1578463018,0.1805446595,0.1582315266,0.0388474241,0.1575522572,-0.308976233,0.0498125739,-0.2299158275]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/324","title":"Error when calculating glue score","comments":"MRPC is also a binary classification task, so its metric is a binary classification metric.\r\n\r\nTo evaluate if pairs of sentences are semantically equivalent, maybe you could take a look at models that compute if one sentence entails the other or not (typically the kinds of model that could work well on the MRPC task).","body":"I was trying glue score along with other metrics here. But glue gives me this error;\r\n\r\n```\r\nimport nlp\r\nglue_metric = nlp.load_metric('glue',name=\"cola\")\r\n\r\nglue_score = glue_metric.compute(predictions, references)\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-8-b9210a524504> in <module>()\r\n----> 1 glue_score = glue_metric.compute(predictions, references)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 191 \"\"\"\r\n 192 if predictions is not None:\r\n--> 193 self.add_batch(predictions=predictions, references=references)\r\n 194 self.finalize(timeout=timeout)\r\n 195 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in add_batch(self, predictions, references, **kwargs)\r\n 207 if self.writer is None:\r\n 208 self._init_writer()\r\n--> 209 self.writer.write_batch(batch)\r\n 210 \r\n 211 def add(self, prediction=None, reference=None, **kwargs):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 155 if self.pa_writer is None:\r\n 156 self._build_writer(pa_table=pa.Table.from_pydict(batch_examples))\r\n--> 157 pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n 158 if writer_batch_size is None:\r\n 159 writer_batch_size = self.writer_batch_size\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/types.pxi in __iter__()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\nTypeError: an integer is required (got type str)\r\n```\r\nI'm not sure whether I'm doing this wrong or whether it's an issue. I would like to know a workaround. Thank you.","comment_length":55,"text":"Error when calculating glue score\nI was trying glue score along with other metrics here. But glue gives me this error;\r\n\r\n```\r\nimport nlp\r\nglue_metric = nlp.load_metric('glue',name=\"cola\")\r\n\r\nglue_score = glue_metric.compute(predictions, references)\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-8-b9210a524504> in <module>()\r\n----> 1 glue_score = glue_metric.compute(predictions, references)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 191 \"\"\"\r\n 192 if predictions is not None:\r\n--> 193 self.add_batch(predictions=predictions, references=references)\r\n 194 self.finalize(timeout=timeout)\r\n 195 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in add_batch(self, predictions, references, **kwargs)\r\n 207 if self.writer is None:\r\n 208 self._init_writer()\r\n--> 209 self.writer.write_batch(batch)\r\n 210 \r\n 211 def add(self, prediction=None, reference=None, **kwargs):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 155 if self.pa_writer is None:\r\n 156 self._build_writer(pa_table=pa.Table.from_pydict(batch_examples))\r\n--> 157 pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n 158 if writer_batch_size is None:\r\n 159 writer_batch_size = self.writer_batch_size\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/types.pxi in __iter__()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\nTypeError: an integer is required (got type str)\r\n```\r\nI'm not sure whether I'm doing this wrong or whether it's an issue. I would like to know a workaround. Thank you.\nMRPC is also a binary classification task, so its metric is a binary classification metric.\r\n\r\nTo evaluate if pairs of sentences are semantically equivalent, maybe you could take a look at models that compute if one sentence entails the other or not (typically the kinds of model that could work well on the MRPC task).","embeddings":[-0.1694530249,-0.2395812869,-0.0304733198,0.1611590832,0.236740306,-0.0491650477,0.1073005199,0.3623578846,0.4330275059,0.0516285449,-0.3168233335,0.2763474584,-0.0798071995,-0.0241100863,0.1026181132,-0.2831492722,-0.1162880808,0.1604485661,-0.2202495039,0.3102566004,-0.324510783,0.2713671029,-0.1858564019,0.3197874725,-0.1767067164,-0.0903876349,0.0204841644,-0.196170941,-0.3263719678,-0.362935394,0.2481628209,-0.1168349832,0.02158558,0.1766418815,-0.0001102634,-0.0120561915,0.458953023,0.0023093964,0.1588456184,-0.1421633959,-0.3889809847,-0.083480984,0.0444146022,-0.2361500263,0.1346656531,0.2275212109,0.0206140038,-0.1826084852,0.2193325311,0.2709797919,0.2774609029,0.3871142864,0.3126004338,-0.1989937872,0.1724911779,-0.2082325965,0.0265877806,0.8870652914,-0.327216506,-0.4651894271,-0.0078018066,0.2069443315,-0.1057071611,0.2180079669,0.2240158021,0.0521962196,0.3716999292,-0.3202626705,-0.1326412559,-0.0713140443,-0.2181349844,-0.2285419106,-0.1499314606,0.0416573919,0.0471167676,-0.3763104677,0.0972394794,-0.0726784766,0.0581606068,-0.2293186486,-0.0549694002,0.1402760744,-0.1144136488,0.2752963901,0.2400909662,0.3214469552,-0.1981527656,0.296900332,0.4147271514,0.2154424191,-0.1390433908,0.0435455628,-0.2401257157,0.1653348356,-0.2926098704,-0.0423279516,-0.0104532884,0.1914083958,0.0239265561,-0.2354671657,0.0487309434,-0.2080052644,0.4857490361,0.3427954614,0.0602983236,0.3389247358,0.3534550667,0.1564653218,0.127425313,0.2417826504,0.002334035,-0.1694935113,0.2388441265,-0.3881970942,0.3655791283,0.2413906008,0.1085240468,0.0770600438,-0.4262593985,-0.1771688163,-0.0480151996,-0.0049847378,0.036921978,0.0718623921,-0.0187913645,0.0006606558,0.1049904004,0.1906935126,-0.1910691857,0.057039503,-0.1342848837,0.4053150415,-0.3368168175,0.1215166077,0.1513898671,-0.033291731,0.2043572962,-0.1073043272,0.1692515314,-0.1872263104,0.0212524477,-0.2450529933,-0.0659325495,0.1954428554,0.0480438098,0.0028141201,0.3261824846,-0.5611881614,-0.1635352671,-0.2155396938,-0.2283369601,-0.3000667095,-0.1469439864,0.2351692468,-0.1796001494,0.0300378092,-0.4003524482,-0.18054685,0.237726599,0.0449476801,0.1126893461,-0.2676103711,-0.2484904975,-0.2921720147,0.2016495019,0.1388746202,-0.08172521,-0.1147434115,-0.1918641925,-0.1030081287,0.1174745634,0.3144727945,0.1723901331,0.43979159,-0.2319522053,0.1515319943,0.5892598629,-0.6347135305,-0.2369804829,0.1007082835,0.0968757421,-0.2819084525,-0.0695398524,0.0950889289,0.2874577045,0.0750943869,0.4693049788,0.1775257438,-0.1200407296,-0.1910822093,-0.3003584146,0.0345603265,0.0852224529,0.0522735454,0.2045292258,-0.0504421517,0.1096888632,-0.0649215803,-0.1438204199,0.1973367929,0.0194501802,0.0301419217,0.331820786,-0.2376646549,0.0082451282,-0.5078004003,-0.0287532061,0.0765153319,-0.5539197922,0.198527962,0.0962853581,-0.0163019728,-0.5231425762,0.1050595492,-0.0699512661,-0.0393085517,0.2326908261,-0.2064474374,0.3046903312,0.0231801495,-0.0384821072,0.171043694,0.0765673891,0.0847157612,-0.0891413689,0.2145031542,-0.2762120366,-0.4148796499,0.138267681,0.3498590291,0.2727716267,0.0353043973,-0.1092623249,0.4997556508,-0.1424180567,-0.2285176069,0.0557274185,-0.0633073077,-0.0071699293,0.0306306798,-0.0186237693,-0.1132494584,0.1906821728,-0.029351959,0.3210164309,0.1958460063,0.0103191854,0.2974856794,-0.1180674285,0.327396512,-0.1225315109,-0.0584307276,-0.1500223279,-0.1341111809,0.4181139767,-0.297301203,-0.0799402669,-0.1827307642,0.3219168186,-0.2306055278,0.3350412548,-0.0496505089,0.1854770035,-0.0289649013,-0.2190921903,0.107579872,-0.1537142396,0.1282522231,0.5975934267,-0.0049649971,-0.1247712001,0.0250173025,-0.1926951706,-0.0234322157,0.1754012406,0.0674831867,0.0308192037,0.1081281975,0.1688194871,-0.0281777438,-0.0380484313,0.1020587161,-0.2313185483,0.2311968207,-0.322055459,-0.0178321414,-0.0451471135,-0.4319617748,-0.1677066386,-0.1099136993,-0.2902896106,-0.6119655967,-0.057234671,0.0513690375,0.0873179585,0.1892627627,0.322621733,0.0651795939,0.1116792187,0.0968883038,0.052267164,-0.2201310247,-0.0294674635,0.0525177605,0.0137362741,0.2037277669,0.4438380897,-0.1036847085,-0.0145902503,-0.1050944924,-0.3087322116,0.0173460748,-0.284208715,0.6033970118,0.4132930934,-0.1238139421,-0.0578880683,-0.3781408966,0.4658437669,-0.4250400364,0.0171624403,0.2245608568,-0.2028575689,-0.0457102694,-0.1568728834,0.0534092486,-0.2870553434,-0.3602614999,0.3244150281,0.1362109035,0.0870978311,0.0247128438,-0.2230973095,0.378121376,-0.1902915835,-0.1985457391,-0.1898382753,-0.0287204385,0.3248326778,-0.2061116695,-0.3413917124,-0.088632822,-0.0667745844,0.3206015527,-0.2410281748,-0.2447149307,-0.5733760595,-0.1851422489,0.0286152307,-0.1296885163,0.1928846389,0.0440016761,-0.1671079099,-0.064908646,-0.0949185565,-0.2339582145,0.1628335267,0.2192826271,0.092607379,-0.1301689148,0.3421754539,0.0853842348,0.3098371029,0.1035997644,-0.0924218148,0.2702838182,0.0431021787,0.2309778929,0.0065481076,-0.4910922647,0.1871620119,-0.1212332994,0.3261258006,-0.116968669,-0.0228460543,-0.0533289127,-0.2160975337,-0.142928794,-0.1235205829,-0.221789971,0.0744251758,0.0152623169,0.1742431521,0.0593148656,-0.2170495242,-0.2897800803,-0.2181468904,0.1944755912,0.2156715095,0.0103608659,-0.0712358952,-0.0382272005,-0.2960558534,-0.3888400793,0.4718077183,0.1434947699,0.3303483725,-0.1176615953,0.2998418212,-0.1516719311,0.1147296652,0.2170622349,-0.3560436964,0.0797637701,0.289891392,0.1962627769,-0.5538406968,-0.1381310374,-0.4275742471,0.3352845907,0.5549218655,0.4441614449,-0.2514420748,0.0707286522,0.4990741014,-0.0577811487,-0.0494861081,-0.4405652881,-0.4276404083,-0.1975727975,-0.134593159,-0.0735190138,0.1434215605,0.2061714977,0.3873663843,0.0714507997,-0.0928121805,-0.2010710984,0.0717374757,-0.2025193423,0.0308996141,0.0319244713,0.0896566287,-0.2613339722,-0.0328474119,-0.5163642764,-0.1181045771,-0.0492085591,-0.0583241209,0.1162236407,0.006550475,0.3921333551,-0.1031682417,-0.1336190701,0.232488364,-0.2945844829,0.4098879993,0.04337411,0.0918040425,0.344673574,0.2249441594,-0.1188238487,0.1059011966,0.4642900229,0.2396440357,0.0409607179,0.3867133856,0.2303498238,-0.2547430396,0.2100529075,0.1255099028,0.7316866517,-0.0363749973,-0.2081698477,0.2821577787,-0.1958425492,0.6601784229,0.2018424273,0.2584860623,-0.2552503347,-0.240117535,0.0362041257,-0.1886274219,0.1207558587,-0.2527753115,-0.4338819087,0.4902597964,0.0216529164,0.1240034103,-0.1225317046,0.1950635314,-0.1712214351,-0.1687160432,-0.0379025266,0.1914608777,-0.333596915,0.1056694016,-0.0188455451,-0.0623411946,0.1282528341,-0.07512784,-0.2535679638,-0.1657012552,-0.3714176118,0.1900477111,0.2342687398,0.0006985788,0.0667436123,-0.256436944,0.2165791541,0.1683935821,-0.0986778289,-0.0945844427,-0.0107043562,-0.2469310313,0.1475564241,0.1228892058,-0.1027315333,-0.1388566792,-0.1170235425,0.0286960378,0.0947821066,-0.260346204,0.2329951674,0.111068137,0.0089348322,-0.2510281205,-0.2027098089,-0.1664216369,-0.0228838976,-0.0703513324,0.1712865978,0.004673244,-0.1367912591,0.2804692388,-0.065941751,-0.1624079347,0.0864233673,0.2396939844,-0.0793856904,-0.0198109224,0.1160599962,-0.1933647692,-0.2555323839,-0.075857617,-0.153219983,-0.2182993144,-0.2836368978,0.1234618053,-0.119727917,-0.0283316765,-0.1401845962,0.1114247739,0.4184065163,0.2592692673,0.0969694182,-0.4761908352,-0.1332361549,0.277235359,-0.0170346592,-0.1239774302,-0.1470875591,0.322176218,0.1539585143,0.3371094167,-0.340765208,-0.0861637741,-0.3368872106,0.1727570742,0.5577085614,-0.3882571459,0.03208258,0.0896285996,0.1925500333,0.2407439053,-0.1684650928,-0.2781485617,-0.0453970693,0.088414669,-0.0080468161,-0.1094610766,0.1246851608,0.0089911344,-0.1972405463,-0.0705862939,-0.2875487208,-0.1359250695,-0.1910493523,0.0623337924,-0.0399488844,0.1979519427,0.3852122426,0.3219557405,-0.0956423357,0.1455473304,-0.1070941314,0.1637578756,0.0806565061,0.0923649892,-0.1439367384,-0.1045835987,-0.1080604568,-0.0023589039,0.2476173043,-0.217047736,-0.2661941051,-0.1419357508,0.3768404126,-0.0228843149,-0.035823565,-0.0351780467,-0.1800120175,0.1641356945,-0.1655005217,0.1204651296,0.2528325021,0.2045561373,0.1759787649,0.1867296845,-0.0148430066,0.0641106591,0.0059583378,0.0965781733,0.3417929709,-0.1481131911,-0.1568119675,-0.3146477044,0.2342946976,-0.0372508503,0.3855211139,-0.1936479062,0.1915557832,-0.1926371604,0.0036960652,0.1574132144,-0.0652413517,0.0030420481,0.2391289175,0.1975857019,-0.1815176755,0.2595298886,0.1466700733,0.1041979119,-0.0200797468,0.516079843,-0.3852641582,-0.0782804415,0.0359458067,0.1538161933,0.0529209971,0.0623459443,0.0045074518,-0.0066839065,-0.3290788531,-0.155305177,0.0652153343,0.1220391914,0.3950314224,0.1611405611,0.0191938709,-0.048688747,-0.2124672532,0.3583585918,0.1800563484,-0.3143035173,0.1221407503,0.1270214617,0.2304644138,-0.2367886007,0.6592804193,0.5281543732,0.3133310378,-0.3758439124,0.1271784157,-0.0340991281,-0.1760197282,0.2069451958,0.1847516,0.3671351671,0.0971979424,0.2671824694,0.2085577548,-0.2633956969,-0.0485939421,-0.0723655745,-0.0130673712,-0.1685698628,0.4525218606,-0.2508077323,-0.2331060469,-0.1536246091,0.0573425516,-0.4240994155,0.1701419502,0.035439305,-0.022380935,0.0756966844,-0.2836797237,0.0981923938,-0.0239031259,0.5166293383,0.1721064448,0.5441957116,-0.5894020796,-0.2243749946,-0.4315952063,0.0422833636,-0.0980620757,-0.0537153855,0.0169731099,0.0522209108,-0.2406836897,0.3361567855,0.1721651852,0.498277992,-0.1158373654,-0.282445997,-0.302873075,-0.0565930903,0.0526053309,-0.4606128931,0.2394749522,-0.2231238037,0.092904076,-0.100292787,0.14038302,-0.0083689336,-0.2390481085,0.2770601213,0.5875069499,0.0879226848,0.2794111669,0.5532605052,-0.0172410365,-0.0034356373,0.0521484464,-0.3737610281,-0.1366231889,0.0499606021,0.0168516561,0.2484907508,-0.087572068,-0.0866054073,-0.3632865548,0.5056182146,-0.1624269336,0.0168920252,-0.0326465443,0.0752586946,0.1267989278,0.0208195522,-0.0418231077,0.0828972608,-0.0186353456,0.3271137178,-0.2005010694,-0.2827176452,0.3735974729,-0.4797991216,-0.3367664516,0.0954478085,0.2530971169,0.2006332725,-0.0931799114,-0.6303013563,-0.2510200441,0.3387457132,0.0636691377,-0.1801239103,0.3238251805,-0.1578463018,0.1805446595,0.1582315266,0.0388474241,0.1575522572,-0.308976233,0.0498125739,-0.2299158275]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/324","title":"Error when calculating glue score","comments":"Closing this one. Feel free to re-open if you have other questions :)","body":"I was trying glue score along with other metrics here. But glue gives me this error;\r\n\r\n```\r\nimport nlp\r\nglue_metric = nlp.load_metric('glue',name=\"cola\")\r\n\r\nglue_score = glue_metric.compute(predictions, references)\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-8-b9210a524504> in <module>()\r\n----> 1 glue_score = glue_metric.compute(predictions, references)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 191 \"\"\"\r\n 192 if predictions is not None:\r\n--> 193 self.add_batch(predictions=predictions, references=references)\r\n 194 self.finalize(timeout=timeout)\r\n 195 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in add_batch(self, predictions, references, **kwargs)\r\n 207 if self.writer is None:\r\n 208 self._init_writer()\r\n--> 209 self.writer.write_batch(batch)\r\n 210 \r\n 211 def add(self, prediction=None, reference=None, **kwargs):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 155 if self.pa_writer is None:\r\n 156 self._build_writer(pa_table=pa.Table.from_pydict(batch_examples))\r\n--> 157 pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n 158 if writer_batch_size is None:\r\n 159 writer_batch_size = self.writer_batch_size\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/types.pxi in __iter__()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\nTypeError: an integer is required (got type str)\r\n```\r\nI'm not sure whether I'm doing this wrong or whether it's an issue. I would like to know a workaround. Thank you.","comment_length":13,"text":"Error when calculating glue score\nI was trying glue score along with other metrics here. But glue gives me this error;\r\n\r\n```\r\nimport nlp\r\nglue_metric = nlp.load_metric('glue',name=\"cola\")\r\n\r\nglue_score = glue_metric.compute(predictions, references)\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-8-b9210a524504> in <module>()\r\n----> 1 glue_score = glue_metric.compute(predictions, references)\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in compute(self, predictions, references, timeout, **metrics_kwargs)\r\n 191 \"\"\"\r\n 192 if predictions is not None:\r\n--> 193 self.add_batch(predictions=predictions, references=references)\r\n 194 self.finalize(timeout=timeout)\r\n 195 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/metric.py in add_batch(self, predictions, references, **kwargs)\r\n 207 if self.writer is None:\r\n 208 self._init_writer()\r\n--> 209 self.writer.write_batch(batch)\r\n 210 \r\n 211 def add(self, prediction=None, reference=None, **kwargs):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size)\r\n 155 if self.pa_writer is None:\r\n 156 self._build_writer(pa_table=pa.Table.from_pydict(batch_examples))\r\n--> 157 pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n 158 if writer_batch_size is None:\r\n 159 writer_batch_size = self.writer_batch_size\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/types.pxi in __iter__()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.asarray()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\nTypeError: an integer is required (got type str)\r\n```\r\nI'm not sure whether I'm doing this wrong or whether it's an issue. I would like to know a workaround. Thank you.\nClosing this one. Feel free to re-open if you have other questions :)","embeddings":[-0.1694530249,-0.2395812869,-0.0304733198,0.1611590832,0.236740306,-0.0491650477,0.1073005199,0.3623578846,0.4330275059,0.0516285449,-0.3168233335,0.2763474584,-0.0798071995,-0.0241100863,0.1026181132,-0.2831492722,-0.1162880808,0.1604485661,-0.2202495039,0.3102566004,-0.324510783,0.2713671029,-0.1858564019,0.3197874725,-0.1767067164,-0.0903876349,0.0204841644,-0.196170941,-0.3263719678,-0.362935394,0.2481628209,-0.1168349832,0.02158558,0.1766418815,-0.0001102634,-0.0120561915,0.458953023,0.0023093964,0.1588456184,-0.1421633959,-0.3889809847,-0.083480984,0.0444146022,-0.2361500263,0.1346656531,0.2275212109,0.0206140038,-0.1826084852,0.2193325311,0.2709797919,0.2774609029,0.3871142864,0.3126004338,-0.1989937872,0.1724911779,-0.2082325965,0.0265877806,0.8870652914,-0.327216506,-0.4651894271,-0.0078018066,0.2069443315,-0.1057071611,0.2180079669,0.2240158021,0.0521962196,0.3716999292,-0.3202626705,-0.1326412559,-0.0713140443,-0.2181349844,-0.2285419106,-0.1499314606,0.0416573919,0.0471167676,-0.3763104677,0.0972394794,-0.0726784766,0.0581606068,-0.2293186486,-0.0549694002,0.1402760744,-0.1144136488,0.2752963901,0.2400909662,0.3214469552,-0.1981527656,0.296900332,0.4147271514,0.2154424191,-0.1390433908,0.0435455628,-0.2401257157,0.1653348356,-0.2926098704,-0.0423279516,-0.0104532884,0.1914083958,0.0239265561,-0.2354671657,0.0487309434,-0.2080052644,0.4857490361,0.3427954614,0.0602983236,0.3389247358,0.3534550667,0.1564653218,0.127425313,0.2417826504,0.002334035,-0.1694935113,0.2388441265,-0.3881970942,0.3655791283,0.2413906008,0.1085240468,0.0770600438,-0.4262593985,-0.1771688163,-0.0480151996,-0.0049847378,0.036921978,0.0718623921,-0.0187913645,0.0006606558,0.1049904004,0.1906935126,-0.1910691857,0.057039503,-0.1342848837,0.4053150415,-0.3368168175,0.1215166077,0.1513898671,-0.033291731,0.2043572962,-0.1073043272,0.1692515314,-0.1872263104,0.0212524477,-0.2450529933,-0.0659325495,0.1954428554,0.0480438098,0.0028141201,0.3261824846,-0.5611881614,-0.1635352671,-0.2155396938,-0.2283369601,-0.3000667095,-0.1469439864,0.2351692468,-0.1796001494,0.0300378092,-0.4003524482,-0.18054685,0.237726599,0.0449476801,0.1126893461,-0.2676103711,-0.2484904975,-0.2921720147,0.2016495019,0.1388746202,-0.08172521,-0.1147434115,-0.1918641925,-0.1030081287,0.1174745634,0.3144727945,0.1723901331,0.43979159,-0.2319522053,0.1515319943,0.5892598629,-0.6347135305,-0.2369804829,0.1007082835,0.0968757421,-0.2819084525,-0.0695398524,0.0950889289,0.2874577045,0.0750943869,0.4693049788,0.1775257438,-0.1200407296,-0.1910822093,-0.3003584146,0.0345603265,0.0852224529,0.0522735454,0.2045292258,-0.0504421517,0.1096888632,-0.0649215803,-0.1438204199,0.1973367929,0.0194501802,0.0301419217,0.331820786,-0.2376646549,0.0082451282,-0.5078004003,-0.0287532061,0.0765153319,-0.5539197922,0.198527962,0.0962853581,-0.0163019728,-0.5231425762,0.1050595492,-0.0699512661,-0.0393085517,0.2326908261,-0.2064474374,0.3046903312,0.0231801495,-0.0384821072,0.171043694,0.0765673891,0.0847157612,-0.0891413689,0.2145031542,-0.2762120366,-0.4148796499,0.138267681,0.3498590291,0.2727716267,0.0353043973,-0.1092623249,0.4997556508,-0.1424180567,-0.2285176069,0.0557274185,-0.0633073077,-0.0071699293,0.0306306798,-0.0186237693,-0.1132494584,0.1906821728,-0.029351959,0.3210164309,0.1958460063,0.0103191854,0.2974856794,-0.1180674285,0.327396512,-0.1225315109,-0.0584307276,-0.1500223279,-0.1341111809,0.4181139767,-0.297301203,-0.0799402669,-0.1827307642,0.3219168186,-0.2306055278,0.3350412548,-0.0496505089,0.1854770035,-0.0289649013,-0.2190921903,0.107579872,-0.1537142396,0.1282522231,0.5975934267,-0.0049649971,-0.1247712001,0.0250173025,-0.1926951706,-0.0234322157,0.1754012406,0.0674831867,0.0308192037,0.1081281975,0.1688194871,-0.0281777438,-0.0380484313,0.1020587161,-0.2313185483,0.2311968207,-0.322055459,-0.0178321414,-0.0451471135,-0.4319617748,-0.1677066386,-0.1099136993,-0.2902896106,-0.6119655967,-0.057234671,0.0513690375,0.0873179585,0.1892627627,0.322621733,0.0651795939,0.1116792187,0.0968883038,0.052267164,-0.2201310247,-0.0294674635,0.0525177605,0.0137362741,0.2037277669,0.4438380897,-0.1036847085,-0.0145902503,-0.1050944924,-0.3087322116,0.0173460748,-0.284208715,0.6033970118,0.4132930934,-0.1238139421,-0.0578880683,-0.3781408966,0.4658437669,-0.4250400364,0.0171624403,0.2245608568,-0.2028575689,-0.0457102694,-0.1568728834,0.0534092486,-0.2870553434,-0.3602614999,0.3244150281,0.1362109035,0.0870978311,0.0247128438,-0.2230973095,0.378121376,-0.1902915835,-0.1985457391,-0.1898382753,-0.0287204385,0.3248326778,-0.2061116695,-0.3413917124,-0.088632822,-0.0667745844,0.3206015527,-0.2410281748,-0.2447149307,-0.5733760595,-0.1851422489,0.0286152307,-0.1296885163,0.1928846389,0.0440016761,-0.1671079099,-0.064908646,-0.0949185565,-0.2339582145,0.1628335267,0.2192826271,0.092607379,-0.1301689148,0.3421754539,0.0853842348,0.3098371029,0.1035997644,-0.0924218148,0.2702838182,0.0431021787,0.2309778929,0.0065481076,-0.4910922647,0.1871620119,-0.1212332994,0.3261258006,-0.116968669,-0.0228460543,-0.0533289127,-0.2160975337,-0.142928794,-0.1235205829,-0.221789971,0.0744251758,0.0152623169,0.1742431521,0.0593148656,-0.2170495242,-0.2897800803,-0.2181468904,0.1944755912,0.2156715095,0.0103608659,-0.0712358952,-0.0382272005,-0.2960558534,-0.3888400793,0.4718077183,0.1434947699,0.3303483725,-0.1176615953,0.2998418212,-0.1516719311,0.1147296652,0.2170622349,-0.3560436964,0.0797637701,0.289891392,0.1962627769,-0.5538406968,-0.1381310374,-0.4275742471,0.3352845907,0.5549218655,0.4441614449,-0.2514420748,0.0707286522,0.4990741014,-0.0577811487,-0.0494861081,-0.4405652881,-0.4276404083,-0.1975727975,-0.134593159,-0.0735190138,0.1434215605,0.2061714977,0.3873663843,0.0714507997,-0.0928121805,-0.2010710984,0.0717374757,-0.2025193423,0.0308996141,0.0319244713,0.0896566287,-0.2613339722,-0.0328474119,-0.5163642764,-0.1181045771,-0.0492085591,-0.0583241209,0.1162236407,0.006550475,0.3921333551,-0.1031682417,-0.1336190701,0.232488364,-0.2945844829,0.4098879993,0.04337411,0.0918040425,0.344673574,0.2249441594,-0.1188238487,0.1059011966,0.4642900229,0.2396440357,0.0409607179,0.3867133856,0.2303498238,-0.2547430396,0.2100529075,0.1255099028,0.7316866517,-0.0363749973,-0.2081698477,0.2821577787,-0.1958425492,0.6601784229,0.2018424273,0.2584860623,-0.2552503347,-0.240117535,0.0362041257,-0.1886274219,0.1207558587,-0.2527753115,-0.4338819087,0.4902597964,0.0216529164,0.1240034103,-0.1225317046,0.1950635314,-0.1712214351,-0.1687160432,-0.0379025266,0.1914608777,-0.333596915,0.1056694016,-0.0188455451,-0.0623411946,0.1282528341,-0.07512784,-0.2535679638,-0.1657012552,-0.3714176118,0.1900477111,0.2342687398,0.0006985788,0.0667436123,-0.256436944,0.2165791541,0.1683935821,-0.0986778289,-0.0945844427,-0.0107043562,-0.2469310313,0.1475564241,0.1228892058,-0.1027315333,-0.1388566792,-0.1170235425,0.0286960378,0.0947821066,-0.260346204,0.2329951674,0.111068137,0.0089348322,-0.2510281205,-0.2027098089,-0.1664216369,-0.0228838976,-0.0703513324,0.1712865978,0.004673244,-0.1367912591,0.2804692388,-0.065941751,-0.1624079347,0.0864233673,0.2396939844,-0.0793856904,-0.0198109224,0.1160599962,-0.1933647692,-0.2555323839,-0.075857617,-0.153219983,-0.2182993144,-0.2836368978,0.1234618053,-0.119727917,-0.0283316765,-0.1401845962,0.1114247739,0.4184065163,0.2592692673,0.0969694182,-0.4761908352,-0.1332361549,0.277235359,-0.0170346592,-0.1239774302,-0.1470875591,0.322176218,0.1539585143,0.3371094167,-0.340765208,-0.0861637741,-0.3368872106,0.1727570742,0.5577085614,-0.3882571459,0.03208258,0.0896285996,0.1925500333,0.2407439053,-0.1684650928,-0.2781485617,-0.0453970693,0.088414669,-0.0080468161,-0.1094610766,0.1246851608,0.0089911344,-0.1972405463,-0.0705862939,-0.2875487208,-0.1359250695,-0.1910493523,0.0623337924,-0.0399488844,0.1979519427,0.3852122426,0.3219557405,-0.0956423357,0.1455473304,-0.1070941314,0.1637578756,0.0806565061,0.0923649892,-0.1439367384,-0.1045835987,-0.1080604568,-0.0023589039,0.2476173043,-0.217047736,-0.2661941051,-0.1419357508,0.3768404126,-0.0228843149,-0.035823565,-0.0351780467,-0.1800120175,0.1641356945,-0.1655005217,0.1204651296,0.2528325021,0.2045561373,0.1759787649,0.1867296845,-0.0148430066,0.0641106591,0.0059583378,0.0965781733,0.3417929709,-0.1481131911,-0.1568119675,-0.3146477044,0.2342946976,-0.0372508503,0.3855211139,-0.1936479062,0.1915557832,-0.1926371604,0.0036960652,0.1574132144,-0.0652413517,0.0030420481,0.2391289175,0.1975857019,-0.1815176755,0.2595298886,0.1466700733,0.1041979119,-0.0200797468,0.516079843,-0.3852641582,-0.0782804415,0.0359458067,0.1538161933,0.0529209971,0.0623459443,0.0045074518,-0.0066839065,-0.3290788531,-0.155305177,0.0652153343,0.1220391914,0.3950314224,0.1611405611,0.0191938709,-0.048688747,-0.2124672532,0.3583585918,0.1800563484,-0.3143035173,0.1221407503,0.1270214617,0.2304644138,-0.2367886007,0.6592804193,0.5281543732,0.3133310378,-0.3758439124,0.1271784157,-0.0340991281,-0.1760197282,0.2069451958,0.1847516,0.3671351671,0.0971979424,0.2671824694,0.2085577548,-0.2633956969,-0.0485939421,-0.0723655745,-0.0130673712,-0.1685698628,0.4525218606,-0.2508077323,-0.2331060469,-0.1536246091,0.0573425516,-0.4240994155,0.1701419502,0.035439305,-0.022380935,0.0756966844,-0.2836797237,0.0981923938,-0.0239031259,0.5166293383,0.1721064448,0.5441957116,-0.5894020796,-0.2243749946,-0.4315952063,0.0422833636,-0.0980620757,-0.0537153855,0.0169731099,0.0522209108,-0.2406836897,0.3361567855,0.1721651852,0.498277992,-0.1158373654,-0.282445997,-0.302873075,-0.0565930903,0.0526053309,-0.4606128931,0.2394749522,-0.2231238037,0.092904076,-0.100292787,0.14038302,-0.0083689336,-0.2390481085,0.2770601213,0.5875069499,0.0879226848,0.2794111669,0.5532605052,-0.0172410365,-0.0034356373,0.0521484464,-0.3737610281,-0.1366231889,0.0499606021,0.0168516561,0.2484907508,-0.087572068,-0.0866054073,-0.3632865548,0.5056182146,-0.1624269336,0.0168920252,-0.0326465443,0.0752586946,0.1267989278,0.0208195522,-0.0418231077,0.0828972608,-0.0186353456,0.3271137178,-0.2005010694,-0.2827176452,0.3735974729,-0.4797991216,-0.3367664516,0.0954478085,0.2530971169,0.2006332725,-0.0931799114,-0.6303013563,-0.2510200441,0.3387457132,0.0636691377,-0.1801239103,0.3238251805,-0.1578463018,0.1805446595,0.1582315266,0.0388474241,0.1575522572,-0.308976233,0.0498125739,-0.2299158275]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"It looks like it comes from `mwparserfromhell`.\r\n\r\nWould it be possible to get the bad `section` that causes this issue ? The `section` string is from `datasets\/wikipedia.py:L548` ? You could just add a `try` statement and print the section if the line `section_text.append(section.strip_code().strip())` crashes.\r\n\r\nIt will help us know if we have to fix it on our side or if it is a `mwparserfromhell` issue.","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":65,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nIt looks like it comes from `mwparserfromhell`.\r\n\r\nWould it be possible to get the bad `section` that causes this issue ? The `section` string is from `datasets\/wikipedia.py:L548` ? You could just add a `try` statement and print the section if the line `section_text.append(section.strip_code().strip())` crashes.\r\n\r\nIt will help us know if we have to fix it on our side or if it is a `mwparserfromhell` issue.","embeddings":[-0.2354938686,-0.1382549554,0.0552539006,0.5212737918,0.0680223629,-0.0856205523,-0.1348956376,0.3071224689,0.2067664564,0.2888492048,0.347761035,-0.0715623647,0.0598848015,-0.4790312052,0.0541907251,0.0624246337,0.2404568046,0.1717450917,0.0140154986,-0.1562183797,-0.0564819202,0.1827820987,-0.318266809,0.4079273045,-0.2893714607,0.1333573014,0.1821404994,0.0453482494,-0.0873938873,-0.5024523139,0.1602192223,-0.3359048069,0.1328683645,0.2546606958,-0.0001277348,-0.1548775285,0.530970335,-0.0702340454,-0.2122372091,-0.0369732305,-0.0304517858,-0.1057011634,-0.0895878971,-0.2332394421,0.2424845099,0.2433309108,0.1983077228,-0.071317412,0.3108440042,0.3353014588,0.0761991441,0.3869360983,0.1124016941,0.0967790335,0.6936830878,0.1377621144,0.125202626,0.0944369435,0.0384380855,-0.2644349337,0.0345922709,0.1256115586,-0.336327672,-0.250264585,-0.1778667122,-0.0576803312,0.1472638547,-0.2539338768,0.30596596,0.2866597474,0.3650677502,-0.0381625369,0.2111723274,-0.0308051053,-0.1637750566,0.2019702494,0.3874559402,0.4430146813,-0.5954430103,0.1332688928,0.479385227,-0.1277801991,-0.0204470921,0.2640166879,-0.2832796872,0.9716570377,0.122401759,0.2389769107,0.1774604172,-0.0569028258,-0.2832579315,-0.1096520424,-0.0692288727,0.1504844725,-0.1025137082,-0.0143687679,0.0064465175,-0.1386348158,0.0563424863,-0.5317035317,-0.4928170741,0.1415746957,-0.0694425851,0.163571313,0.5938709378,-0.2485040873,0.2245710641,0.0420949981,0.0299751759,-0.0062607736,0.0381632559,0.1198109835,-0.0576130264,0.0200764239,-0.350661397,0.1356573254,0.2376644164,-0.3176876903,-0.2114284188,0.1597348154,-0.1107552648,-0.1038236395,-0.0693252012,0.5388111472,-0.0688824281,0.3479135036,-0.0419887006,-0.0867054388,-0.1887230724,-0.4938837886,0.0754925832,0.1780680716,-0.3806151748,0.1249615997,-0.193040207,0.1050055996,0.044943396,0.2112226635,-0.0776245296,-0.5137030482,-0.2019978613,-0.16895625,0.0394029059,0.1415690631,0.2810630202,0.329131633,0.2446107566,-0.3599990904,-0.0871145725,0.240745604,-0.3768997192,-0.2545084357,-0.4904033542,-0.0008211152,-0.0353646055,0.2822312117,-0.59474051,0.0508756116,-0.0120046344,-0.0575802922,0.1264092624,-0.1706045866,-0.1604581326,-0.057974752,0.1905232072,0.450928092,-0.0708901063,-0.1168515608,-0.1774329096,-0.2140928507,0.2010282427,0.2269358635,-0.1839905083,0.5195731521,-0.4584503472,0.4583002627,0.3113233745,-0.2849512994,-0.0571353063,-0.0979459509,-0.1427361071,-0.0916979015,0.0400379486,-0.1657854319,0.1380462795,-0.187332809,-0.27854532,0.0720921829,0.0015590468,0.2039587945,-0.1975086182,-0.3238655031,0.257368505,0.1239002645,0.1697516441,-0.0927250609,0.1561171263,0.3895823061,0.5115286708,-0.0813133717,-0.0571516119,0.0208800733,0.2228118181,0.1539544761,-0.0132474313,-0.11784105,-0.075242497,0.0509776585,0.1157508269,0.2266798019,-0.253945142,-0.2227997333,-0.2215527445,0.2793840766,0.0249313395,0.1636644602,0.0264547113,0.1939318627,-0.2000493407,0.2564404309,-0.075311251,-0.2795470059,-0.3256709576,0.1935569346,-0.0402257368,0.2280983031,-0.0192772765,-0.1150707304,-0.2168326825,-0.1158802882,0.51837641,0.090927422,-0.1253851503,-0.0055526793,0.2218338996,0.1994571984,-0.3905209899,-0.1928244233,-0.0735775605,-0.4623733461,0.0073243184,0.2881063819,0.0147594763,-0.0881737769,0.3043561578,0.1623749733,0.363026619,0.1495341361,-0.1418321729,0.1046718732,0.111119777,-0.0588465407,0.31650123,-0.1134981662,0.3225903511,0.4323560297,-0.0882901549,-0.2456407249,0.1771972775,-0.0821040571,0.5798689723,-0.1613234133,0.3835196197,0.2482497096,-0.2514739633,-0.0573471934,0.2625975907,-0.3426445425,0.0388191789,-0.0855645016,-0.1710690111,0.260841608,0.1124908105,0.0950722992,0.4415653944,0.1859731227,0.0493018255,0.0334592499,-0.2050507665,-0.0729158074,-0.3128654063,-0.0842659175,-0.5056683421,0.2636985779,-0.4026723504,-0.0203204527,-0.3042700887,-0.0564648621,-0.4614537656,-0.0538029037,-0.3293029368,-0.1057729051,-0.1545158923,0.239153564,-0.1852569878,0.0493119583,-0.0950176641,0.2174655497,-0.0937689021,0.4357780516,-0.1978890449,-0.4379402101,-0.4635023177,-0.1898544729,0.5398993492,0.2548497617,0.0520797633,-0.1376512945,-0.260935545,-0.2251266539,-0.1003714874,0.4269749224,0.1238296553,0.012356597,-0.0263437498,0.2941504419,0.140590176,-0.3974824548,-0.021061305,-0.2681171894,-0.0495684333,0.0712934509,0.0634675398,-0.0828441158,0.1557223797,-0.0571126379,-0.384139657,-0.0979475603,-0.2573730052,-0.1597690284,0.1611708701,-0.4041493535,-0.1138582602,0.1818556041,-0.4152784944,-0.2532142997,-0.1731048375,0.0803441405,0.083342433,0.0995669663,-0.1332220733,0.2256107032,0.0259850845,0.0232237428,0.0667399168,-0.484084785,0.3821097314,0.1489899009,-0.069530502,-0.2207824439,-0.0596512295,0.2172506005,-0.1444769949,0.1558894068,0.064577952,-0.1554575861,-0.3619950116,-0.2255566865,0.4314027429,-0.1706152409,0.3485953212,0.0884206668,0.7554531097,-0.1121196747,0.1790819019,0.2185539603,-0.0767456293,-0.0468219779,0.2099902332,-0.1366433948,0.1960787326,0.0041359323,-0.3386886716,0.5686203837,-0.1833814085,-0.4030360281,-0.1962027103,-0.1050607637,-0.5596653819,-0.5058274269,0.1222739592,0.435106039,0.1982286423,-0.0593299605,0.3139870167,-0.0833059028,-0.3276592195,0.1397939324,0.3305218518,-0.1592597365,-0.1249691844,-0.2837719619,-0.0923950821,-0.0676725283,0.0812114775,0.1704558134,0.4799149334,-0.0280493312,0.0361194164,0.0095500136,-0.0412681177,0.5193560719,0.2742488384,0.3625155389,-0.0101458365,0.1713814735,-0.0502313636,-0.3192234337,-0.1449991763,0.0795565173,0.4323585629,-0.040552564,-0.4624991417,0.080695942,-0.0178655162,0.1046397462,-0.1688562185,0.0097631384,-0.1207341552,-0.1927438229,-0.2643757463,0.3259371519,0.1250087172,0.0898501873,0.1837792695,0.1340372413,0.2293275446,0.1369440705,0.0553549379,0.0896382034,0.1292001605,0.0179472249,-0.235762164,-0.1744028926,0.2055360675,0.1944972277,0.2299951166,0.6723674536,-0.1057301834,0.0051195575,0.1353537738,0.1814277172,0.4497550428,0.0121885194,0.3555920124,0.2967022955,0.2326452732,-0.2024063021,0.1796643287,0.2087224275,-0.0304252375,-0.3745726943,0.0820444226,0.5366145968,-0.1198287085,-0.0700639337,0.1448965967,-0.0848960504,-0.3131203353,0.3141000867,0.3210225999,1.3559755087,0.1880858541,0.0667419508,0.1177503318,0.0406065173,0.8324738145,-0.51708287,-0.1036776826,-0.3959549963,0.0850283131,-0.1350558996,-0.0578899793,0.1359573007,0.041562058,-0.2892319262,0.3057876229,0.1148737818,0.8049696684,0.1543678045,0.2144236267,0.1321058422,-0.0270720441,-0.2793483734,-0.0562812388,-0.3469496071,0.2603121698,-0.3400978446,0.1158581674,0.1644753069,-0.3571952879,-0.7858266234,0.1694662571,-0.1605653018,0.1277381629,0.0488966256,0.005637899,0.2098803967,-0.1723473221,0.0802903846,0.3055694103,-0.2605264187,0.4705589414,-0.1250104159,-0.2333446741,0.0665498674,0.0901727751,-0.1304869056,-0.1100963801,-0.1946901828,0.3893638253,0.0325885452,-0.2813827395,-0.2252381891,-0.0576398969,0.3674284816,0.0094800694,-0.4063715935,0.0330692418,0.2243004441,-0.1330427229,0.0234274231,0.0034673293,-0.1397193968,-0.0447039604,0.1933462918,-0.1299545318,-0.0793371648,0.1953894198,-0.0774394199,-0.4564896822,0.3584900498,0.407104075,-0.1613590121,-0.1280613095,-0.0884089172,-0.3836039305,-0.554007709,-0.1997277141,-0.1511522681,-0.0128996391,-0.0689955577,0.123709023,0.369612366,-0.0693092123,-0.0782530904,-0.3825037479,0.0206221975,0.0397141874,-0.1970056295,-0.0393801741,-0.0365677401,0.1231562346,-0.2233594209,0.1175331846,-0.0984479487,-0.0244838484,-0.4243420362,0.185956195,0.4281347692,0.2743801773,-0.3331552744,-0.159072578,-0.0666217953,0.2899631858,0.0478476174,0.0016908725,-0.1511006355,0.2825160325,0.2611758709,-0.1019863933,0.2051037997,0.0092910314,0.3371180594,-0.0620773099,-0.0645479783,-0.0247716904,0.006665282,0.2926816344,0.3140418828,0.1397961378,-0.5953754187,0.2701896727,0.0647551715,0.5508186221,0.0569939353,0.3407200277,0.0993160903,0.0030709519,-0.1658892035,0.2401728332,0.1845929325,0.2494466305,0.0745811611,0.1662678868,-0.1344636083,0.119496204,-0.1731606424,0.1967640221,-0.2846857905,-0.1794301122,0.5265513659,0.0529444143,-0.1286949068,0.0018369763,0.4313596487,-0.0853968337,-0.291143924,0.2447131127,-0.1568017155,-0.2636803389,0.3740332425,0.1354050189,0.3472859263,0.0861542225,0.1139845252,0.0974198058,-0.0708163381,0.2093579769,-0.0196563322,-0.2225802839,0.3959622085,0.5164225101,0.0391366035,0.2192318738,0.1454229504,0.0987588316,-0.0583233014,-0.1502765864,0.2071991861,0.6587830186,0.0090755755,0.1181544587,-0.016238559,0.000362251,0.119025141,-0.1617891192,-0.4723658562,0.2215996534,-0.2770661712,-0.0208431967,-0.384003222,-0.1041665524,-0.3006263077,0.2633967698,0.0993540809,-0.0973135307,-0.2757288814,0.2871501744,-0.4027179778,0.1154580191,-0.0865751579,0.0279191975,0.011255499,-0.3068801165,0.105705671,-0.2201872766,0.023019895,-0.0036655006,0.1486098915,0.4637709558,0.5263075233,-0.2449313402,-0.0690445527,0.1363199055,0.1998604387,-0.0509442054,0.0471037515,0.0090568913,0.4417717159,0.1217150241,0.0332920142,-0.1301126629,-0.2672181427,0.0909221321,-0.103650555,-0.0854981318,-0.273617655,-0.1242916137,-0.0035987361,-0.2767068148,-0.0119254468,-0.2484145164,0.2276866138,0.1429166049,0.2154936343,0.1450828165,0.0181532521,-0.0332678147,0.0562599003,0.2175432295,0.3581020534,0.3341641426,-0.2006722391,-0.5086638331,-0.7261003852,0.3948657215,-0.3475843668,-0.3004252911,0.1989096999,-0.097822994,0.1319255382,0.4655018747,0.1446798295,0.2105081528,-0.2215521932,0.1711558551,-0.2206071317,-0.286298573,0.0909639373,0.004000973,-0.1730719358,-0.4309080243,0.3013596237,-0.1249778867,-0.1084995866,-0.1239066198,-0.3051841259,-0.0622992627,-0.3106020391,0.5700954199,0.0224325955,0.3696025014,-0.0837172419,-0.0149174277,-0.1199050099,-0.1047847122,0.005029934,-0.0278878864,-0.1431304365,0.3542475104,0.10514348,0.2677586377,-0.5126780272,0.3220697939,0.1184470579,0.149011001,-0.2159142643,-0.0274390206,-0.2320137471,0.0537014566,0.1611271054,0.0457919687,0.0687782094,0.2690194845,-0.3508996665,-0.1207110882,0.2304455638,-0.4841432273,-0.2777426243,-0.3321529627,0.4267872572,-0.0894673914,-0.2710385323,-0.2365677804,-0.1027486846,0.1876013726,0.1310238093,-0.0490227081,-0.2320203632,-0.1751692742,-0.0789413452,-0.0205453858,-0.0076944679,0.1971932054,-0.5537313223,0.2444155514,-0.4019878209]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"Hi, \r\n\r\nThank you for you answer.\r\nI have try to print the bad section using `try` and `except`, but it is a bit weird as the error seems to appear 3 times for instance, but the two first error does not print anything (as if the function did not go in the `except` part).\r\nFor the third one, I got that (I haven't display the entire text) :\r\n\r\n> error : ==== Parque nacional Cajas ====\r\n> {{AP|Parque nacional Cajas}}\r\n> [[Archivo:Ecuador cajas national park.jpg|thumb|left|300px|Laguna del Cajas]]\r\n> El parque nacional Cajas est\u00e1 situado en los [[Cordillera de los Andes|Andes]], al sur del [[Ecuador]], en la provincia de [[Provincia de Azuay|Azuay]], a 33\r\n> [[km]] al noroccidente de la ciudad de [[Cuenca (Ecuador)|Cuenca]]. Los accesos m\u00e1s comunes al parque inician todos en Cuenca: Desde all\u00ed, la v\u00eda Cuenca-Mol\r\n> leturo atraviesa en Control de [[Surocucho]] en poco m\u00e1s de 30 minutos de viaje; m\u00e1s adelante, esta misma carretera pasa a orillas de la laguna La Toreadora donde est\u00e1n el Centro Administrativo y de Informaci\u00f3n del parque. Siguiendo de largo hacia [[Molleturo]], por esta v\u00eda se conoce el sector norte del Cajas y se serpentea entre varias lagunas mayores y menores.\r\n> Para acceder al parque desde la costa, la v\u00eda Molleturo-Cuenca es tambi\u00e9n la mejor opci\u00f3n.\r\n\r\nHow can I display the link instead of the text ? I suppose it will help you more ","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":234,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nHi, \r\n\r\nThank you for you answer.\r\nI have try to print the bad section using `try` and `except`, but it is a bit weird as the error seems to appear 3 times for instance, but the two first error does not print anything (as if the function did not go in the `except` part).\r\nFor the third one, I got that (I haven't display the entire text) :\r\n\r\n> error : ==== Parque nacional Cajas ====\r\n> {{AP|Parque nacional Cajas}}\r\n> [[Archivo:Ecuador cajas national park.jpg|thumb|left|300px|Laguna del Cajas]]\r\n> El parque nacional Cajas est\u00e1 situado en los [[Cordillera de los Andes|Andes]], al sur del [[Ecuador]], en la provincia de [[Provincia de Azuay|Azuay]], a 33\r\n> [[km]] al noroccidente de la ciudad de [[Cuenca (Ecuador)|Cuenca]]. Los accesos m\u00e1s comunes al parque inician todos en Cuenca: Desde all\u00ed, la v\u00eda Cuenca-Mol\r\n> leturo atraviesa en Control de [[Surocucho]] en poco m\u00e1s de 30 minutos de viaje; m\u00e1s adelante, esta misma carretera pasa a orillas de la laguna La Toreadora donde est\u00e1n el Centro Administrativo y de Informaci\u00f3n del parque. Siguiendo de largo hacia [[Molleturo]], por esta v\u00eda se conoce el sector norte del Cajas y se serpentea entre varias lagunas mayores y menores.\r\n> Para acceder al parque desde la costa, la v\u00eda Molleturo-Cuenca es tambi\u00e9n la mejor opci\u00f3n.\r\n\r\nHow can I display the link instead of the text ? I suppose it will help you more ","embeddings":[-0.2193864137,-0.2222698182,0.0420608148,0.3609670401,0.0011143076,-0.0577766001,-0.2417550832,0.209553659,0.3074944019,0.1932076514,0.3478331268,0.0101021947,0.0042322748,-0.4349834919,-0.0932994261,-0.023367146,0.2952888906,0.1979200393,0.04570226,-0.2040556818,-0.007366735,0.1882302612,-0.4374738634,0.3260341585,-0.1881965697,0.1131407544,0.1581795365,-0.1393144429,0.014603612,-0.3755434752,-0.0074203629,-0.314427346,0.0927146226,0.3675207794,-0.0001274272,-0.2371362448,0.4585935175,-0.0967858285,-0.1138205379,-0.1958162785,-0.0912482142,-0.1459701657,-0.1145082563,-0.295796603,0.2706686258,0.0221738387,0.2623672187,-0.2641826868,0.3727796972,0.311512351,0.0584836528,0.1152181178,0.2354289144,0.0652595684,0.4728451967,0.1147156358,0.1719979048,-0.0168300997,0.0028644218,-0.0566985495,0.1747065783,0.226606369,-0.3517799973,-0.2806597054,-0.1432513744,0.0191190019,0.3303057551,-0.3932858109,0.4076831639,0.1749070436,0.3341667354,-0.0630978495,0.3528113365,-0.0091338204,-0.2008414268,0.1968026757,0.4043156207,0.5426412225,-0.4197180569,0.0691999942,0.2718425393,-0.0963856652,-0.1063987538,0.3659130633,-0.1545752883,0.8306202292,0.0233020242,0.2787495255,0.0719276592,-0.0208499245,-0.0637707785,-0.2760691643,-0.0240615029,0.2852091193,-0.0458304361,0.0604048781,-0.1112740263,-0.1204128712,-0.0344832502,-0.5352146626,-0.3188851178,0.016314907,0.0520349368,0.2536082268,0.5370830297,-0.1810673624,0.1744203269,0.0504513346,-0.1436626613,-0.0362118743,0.0503579937,-0.0149717974,0.0126637584,-0.0031581181,-0.4311356246,0.1612440199,0.207602188,-0.1855865717,-0.043176949,0.3412980437,-0.2080373168,-0.0389119461,-0.1002752483,0.4228107929,-0.0640100911,0.3921245337,0.0432952233,-0.0033581243,-0.1112078279,-0.4731747806,0.0116122067,0.3763631284,-0.3916964531,0.1339963228,-0.2620747685,0.2182763219,0.1239406243,0.2374808937,-0.0575655848,-0.6647424698,-0.1636941582,-0.3196177781,0.0889815837,0.0721609294,0.2993113399,0.4473589957,0.2203225344,-0.3982048333,-0.0111890361,0.2441385239,-0.3707463741,-0.1091459915,-0.5119053125,0.0327002518,-0.0072276462,0.2863237858,-0.6130176783,0.071035035,-0.0608309433,-0.2034663558,0.1553456932,-0.136609152,-0.0946944132,-0.040131256,0.1314945221,0.284376502,-0.1842534989,0.1015154868,-0.1600361168,-0.1020812616,0.443862319,0.1019083261,-0.0799845234,0.6557809114,-0.4682842493,0.6272986531,0.2062439322,-0.3382780552,0.1129751652,0.0943775177,-0.0297438093,-0.1602768451,-0.0175427813,-0.2283287346,0.1636880189,-0.0794315636,-0.3810733557,-0.0160949845,0.0205538664,0.2330003977,-0.195982486,-0.228737548,0.146527797,0.0198918972,0.1913723201,-0.0826181471,0.1403740942,0.4307594001,0.4653141797,-0.1570250094,0.0332013443,0.0829630569,0.1245785058,0.2342986017,-0.1293135881,0.0222032443,-0.0531830825,-0.0452171713,0.0789411962,0.1511837393,-0.2835966945,-0.0415176526,-0.2271316946,0.2197405547,-0.0927394703,0.0808862746,0.0659536347,0.1650982946,-0.1487154365,0.3934485912,0.0759446323,-0.2305076122,-0.2285712212,0.1141640618,-0.0720312521,0.2543570101,-0.1999925524,-0.1869710833,-0.3368486166,-0.1040874645,0.5295463204,0.0206557009,-0.1265149117,0.1563125253,0.2563987076,0.1210784316,-0.4433177114,-0.1928886771,-0.0400879569,-0.5070744157,0.1419572085,0.2667557597,0.2609446943,-0.0980919302,0.1688242406,0.0551044792,0.3346451223,0.0481992736,-0.1643210053,0.0471285656,0.0951245204,0.0196772814,0.385991931,-0.2309042066,0.2975935638,0.5082976818,-0.1537844539,-0.3612570763,0.1945492774,-0.1508444548,0.5056219101,-0.0185458716,0.1731353551,0.1171935648,-0.1968074441,-0.0598715618,0.2999535799,-0.3947967589,0.0685457736,-0.0993594378,-0.1491765082,0.4354537427,0.2558443248,0.0767732114,0.4650770128,0.2169238925,-0.0594202429,0.1005321741,-0.2642533779,-0.0237028766,-0.2029424757,-0.2103270888,-0.3739588261,0.2363000214,-0.2853994966,-0.0107833426,-0.2933445275,-0.0377447046,-0.3747020662,0.134397462,-0.3128294051,-0.1369610578,-0.07216952,0.2654456794,-0.2574843764,0.1088799164,0.0163196735,0.0890122503,-0.0874495506,0.382876724,-0.3068619668,-0.2860129774,-0.4424818754,-0.1794234216,0.4841916263,0.2453587353,0.0898851156,-0.1131181568,-0.034018755,-0.2677024603,-0.1557061523,0.4435267746,-0.1022695825,-0.0443903804,-0.0383539796,0.2406124026,-0.008651997,-0.3758123517,-0.0006716651,-0.3829981685,-0.04584774,0.0885857791,0.0353933424,-0.1559861153,0.0406866111,-0.0269830637,-0.2393774092,-0.068910405,-0.3220782876,-0.1188825145,0.1221832037,-0.3866707683,0.082772769,0.2016867101,-0.3991760015,-0.2127141207,-0.1958945841,0.0294108149,0.078367725,0.0208120719,-0.1779778898,0.378598541,0.0338088907,0.171837464,-0.1126878709,-0.5451945662,0.4099715948,0.0749999136,-0.177185744,-0.3284282088,0.0607920513,0.2713207603,-0.1899035424,0.1288836598,0.0628008693,-0.1509239823,-0.3709451258,-0.2282283455,0.3510294259,-0.1403122842,0.3559291065,0.049637679,0.6082237363,-0.0265615489,0.2332248986,0.3531579077,-0.0431798659,0.04477337,0.0899039805,-0.0445801504,0.3161122799,0.1238505021,-0.3487167358,0.5429698825,-0.2769103348,-0.2745007873,-0.2182301134,-0.0114584323,-0.6698663831,-0.4383551478,0.2053493261,0.5381639004,0.1744176745,-0.2152736336,0.2217828631,-0.0932663903,-0.3328768611,-0.1135145947,0.1445526928,-0.057660237,-0.0941787213,-0.2348814905,-0.0249772929,-0.1007095724,0.2660746872,0.1573954225,0.3585386276,-0.0458334051,0.0154836709,0.0825780332,0.0937844738,0.4519880116,0.2045943886,0.2119671255,-0.1028497964,0.0741628632,-0.0630848408,-0.3478733599,-0.3351353705,-0.0863322318,0.5785005689,-0.1238891929,-0.4815710485,-0.0188395344,-0.0802030414,0.065509133,-0.0575169101,-0.0926694497,-0.0509813987,-0.139069885,-0.414844811,0.3135458529,0.2891496122,0.1360023916,0.2378216535,0.1319931746,0.2526874244,0.1143099517,-0.0005188019,-0.0190731697,0.10958983,0.0324377939,-0.3732857406,-0.1598348171,0.1572122872,0.0943617523,0.1216591597,0.5638344884,-0.1961217374,-0.0114660896,-0.0403434895,0.284702152,0.437505424,0.0713800341,0.1214105263,0.3562132716,0.3327354789,-0.0398866273,0.237655133,0.2201416641,0.0089263851,-0.3136926293,0.1202797368,0.3556164205,0.1389291435,-0.0740801319,-0.0028917589,0.1456306577,-0.2708331943,0.3193234205,0.3195789754,1.372379899,0.2524777353,-0.0122908894,0.0701549649,0.1158581078,0.857478261,-0.5321191549,-0.0247942042,-0.3011968732,0.1800721586,-0.1156750694,0.0221411996,0.1793841124,0.009347355,-0.4308919013,0.32995978,0.0915310532,0.671652019,0.1398278028,0.0962220803,0.3284432888,-0.0460927971,-0.096460484,-0.0222466812,-0.2750360966,0.0543853678,-0.3748600185,0.0384025052,0.2037559301,-0.498711437,-0.6637941599,0.1588662714,-0.1476394832,0.0761889517,-0.0480739772,0.020771455,0.3031446636,-0.0541057736,0.1048752517,0.3243848383,-0.208473295,0.2775375247,-0.0919650197,-0.2851914763,-0.1617548764,-0.0309143104,-0.1557297409,-0.1114566401,-0.2150617838,0.4981532693,0.0000336375,-0.2002714127,-0.2003249377,0.0206613597,0.3636728525,0.0219874922,-0.3902974129,0.0601805039,0.0696899667,-0.0999049172,-0.001891361,-0.0827932879,-0.127290383,0.0096067796,0.2655427456,-0.1900436878,-0.0008517299,0.2447743863,-0.0815134272,-0.4689823389,0.3524132669,0.3540010154,-0.223430559,-0.1734398752,-0.2212518454,-0.473503828,-0.5677690506,-0.166605249,-0.1467722207,0.0511136614,-0.0456681401,0.2104507238,0.3831043541,-0.0226683337,0.0747486725,-0.3272097409,0.1152125299,0.0194205493,-0.3564961851,0.0015362462,-0.0434270725,0.2575899661,-0.0683798268,0.0001513091,-0.1080108806,0.0714942813,-0.5082094073,0.2036318034,0.2900798917,0.2227190435,-0.3865591884,-0.1188447252,0.0048736264,0.3294782639,-0.0960453004,-0.0196611136,-0.0784214661,0.280202806,0.4783464968,-0.1615958661,0.0482017063,0.1021149084,0.2719171643,-0.1299027652,-0.1013020277,-0.1592018902,-0.0340860076,0.3143925667,0.1749917418,0.2555951178,-0.5346778631,0.1084835231,0.109566249,0.6087119579,0.0399743021,0.304631114,0.1993795484,-0.0364806317,-0.1820123941,0.1866481453,0.3626476824,0.3018642366,0.0368546844,0.0486570969,-0.1126028672,0.1141702011,-0.2240505069,0.1915571988,-0.3164019883,-0.2697434127,0.5364493728,0.1027927101,-0.1424027681,0.2094241679,0.5686332583,0.1705683023,-0.2566481829,0.2080309987,-0.0559712239,-0.3840126693,0.42071715,0.1680961847,0.3814961314,-0.0084036896,0.0741008297,-0.0289221425,-0.2506685853,0.2643938959,-0.010890224,0.0071772235,0.2346202582,0.458139658,-0.1335859746,0.0907787085,0.1331705749,0.1849113703,-0.086978972,-0.2192963511,0.1429413408,0.5942344666,0.0661736801,-0.0409248844,-0.163282752,-0.3828209639,0.2021690309,-0.1307419986,-0.4244832695,0.2649624348,-0.2229965925,0.1415845156,-0.4221544862,-0.082507208,-0.287514776,0.1561113596,0.18956393,-0.1085221395,-0.2405119985,0.3406592607,-0.3534331322,0.0090759126,-0.2195271552,-0.0345242471,-0.1156718135,-0.3618557453,0.2184060365,-0.0604190752,-0.0227624588,-0.1466313899,0.3151837289,0.5144750476,0.5036407113,-0.332106024,0.0096955821,-0.0125398282,0.2140850425,0.0401925705,0.1396662742,0.1544336528,0.3083744347,0.0984195694,0.0442989282,-0.1031991243,-0.1443878859,0.2481876314,0.0352429077,0.0871720463,-0.3964233696,-0.0427595302,0.2113657445,-0.2974696755,-0.0624406748,-0.447871685,0.1580037922,0.1400851458,0.0591823049,0.1478371769,0.0333921909,-0.0252462011,0.0196449552,0.1699490845,0.3986816108,0.3376191556,-0.285045594,-0.4959666431,-0.5544905663,0.4598610997,-0.3090630174,-0.2179690003,-0.0053541041,0.0287983697,0.1246418133,0.4619228542,0.1005712524,0.1358379871,-0.3769991696,0.3481936157,-0.2102953643,-0.0202900451,0.257219702,-0.0340187177,-0.0518651977,-0.1813149452,0.2395747751,-0.2129166722,-0.0521688163,-0.1075860113,-0.1147371829,-0.2043813616,-0.1875986457,0.565671742,0.1807953864,0.3579048812,-0.1033656672,-0.0331893936,-0.0759954005,0.103763245,0.0134172142,0.0097625693,0.1276433021,0.1763846874,0.0184162576,0.2718162835,-0.4293498397,0.2451841831,0.0826668665,0.1352255493,-0.3105969131,-0.0488495454,-0.2055886686,0.1115624756,-0.1069975644,-0.0531726032,-0.0025394876,0.2643370032,-0.3947309554,-0.1286714673,0.2791518569,-0.3982184529,-0.3795571327,-0.3196076751,0.2929228842,-0.1239411384,-0.0497830287,-0.1190621927,0.0295184441,0.1805836111,0.1351723671,-0.1762548983,-0.23945418,-0.2590578496,-0.2007961273,0.0301584583,-0.0052949404,0.1539319754,-0.3781682849,0.1778814495,-0.395020932]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"The error appears several times as Apache Beam retries to process examples up to 4 times irc.\r\n\r\nI just tried to run this text into `mwparserfromhell` but it worked without the issue.\r\n\r\nI used this code (from the `wikipedia.py` script):\r\n```python\r\nimport mwparserfromhell as parser\r\nimport re\r\nimport six\r\n\r\nraw_content = r\"\"\"==== Parque nacional Cajas ====\r\n{{AP|Parque nacional Cajas}}\r\n[[Archivo:Ecuador cajas national park.jpg|thumb|left|300px|Laguna del Cajas]]\r\nEl parque nacional Cajas est\u00e1 situado en los [[Cordillera de los Andes|Andes]], al sur del [[Ecuador]], en la provincia de [[Provincia de Azuay|Azuay]], a 33\r\n[[km]] al noroccidente de la ciudad de [[Cuenca (Ecuador)|Cuenca]]. Los accesos m\u00e1s comunes al parque inician todos en Cuenca: Desde all\u00ed, la v\u00eda Cuenca-Mol\r\nleturo atraviesa en Control de [[Surocucho]] en poco m\u00e1s de 30 minutos de viaje; m\u00e1s adelante, esta misma carretera pasa a orillas de la laguna La Toreadora donde est\u00e1n el Centro Administrativo y de Informaci\u00f3n del parque. Siguiendo de largo hacia [[Molleturo]], por esta v\u00eda se conoce el sector norte del Cajas y se serpentea entre varias lagunas mayores y menores.\r\n\"\"\"\r\n\r\nwikicode = parser.parse(raw_content)\r\n\r\n# Filters for references, tables, and file\/image links.\r\nre_rm_wikilink = re.compile(\"^(?:File|Image|Media):\", flags=re.IGNORECASE | re.UNICODE)\r\n\r\ndef rm_wikilink(obj):\r\n return bool(re_rm_wikilink.match(six.text_type(obj.title)))\r\n\r\ndef rm_tag(obj):\r\n return six.text_type(obj.tag) in {\"ref\", \"table\"}\r\n\r\ndef rm_template(obj):\r\n return obj.name.lower() in {\"reflist\", \"notelist\", \"notelist-ua\", \"notelist-lr\", \"notelist-ur\", \"notelist-lg\"}\r\n\r\ndef try_remove_obj(obj, section):\r\n try:\r\n section.remove(obj)\r\n except ValueError:\r\n # For unknown reasons, objects are sometimes not found.\r\n pass\r\n\r\nsection_text = []\r\nfor section in wikicode.get_sections(flat=True, include_lead=True, include_headings=True):\r\n for obj in section.ifilter_wikilinks(matches=rm_wikilink, recursive=True):\r\n try_remove_obj(obj, section)\r\n for obj in section.ifilter_templates(matches=rm_template, recursive=True):\r\n try_remove_obj(obj, section)\r\n for obj in section.ifilter_tags(matches=rm_tag, recursive=True):\r\n try_remove_obj(obj, section)\r\n\r\n section_text.append(section.strip_code().strip())\r\n```","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":264,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nThe error appears several times as Apache Beam retries to process examples up to 4 times irc.\r\n\r\nI just tried to run this text into `mwparserfromhell` but it worked without the issue.\r\n\r\nI used this code (from the `wikipedia.py` script):\r\n```python\r\nimport mwparserfromhell as parser\r\nimport re\r\nimport six\r\n\r\nraw_content = r\"\"\"==== Parque nacional Cajas ====\r\n{{AP|Parque nacional Cajas}}\r\n[[Archivo:Ecuador cajas national park.jpg|thumb|left|300px|Laguna del Cajas]]\r\nEl parque nacional Cajas est\u00e1 situado en los [[Cordillera de los Andes|Andes]], al sur del [[Ecuador]], en la provincia de [[Provincia de Azuay|Azuay]], a 33\r\n[[km]] al noroccidente de la ciudad de [[Cuenca (Ecuador)|Cuenca]]. Los accesos m\u00e1s comunes al parque inician todos en Cuenca: Desde all\u00ed, la v\u00eda Cuenca-Mol\r\nleturo atraviesa en Control de [[Surocucho]] en poco m\u00e1s de 30 minutos de viaje; m\u00e1s adelante, esta misma carretera pasa a orillas de la laguna La Toreadora donde est\u00e1n el Centro Administrativo y de Informaci\u00f3n del parque. Siguiendo de largo hacia [[Molleturo]], por esta v\u00eda se conoce el sector norte del Cajas y se serpentea entre varias lagunas mayores y menores.\r\n\"\"\"\r\n\r\nwikicode = parser.parse(raw_content)\r\n\r\n# Filters for references, tables, and file\/image links.\r\nre_rm_wikilink = re.compile(\"^(?:File|Image|Media):\", flags=re.IGNORECASE | re.UNICODE)\r\n\r\ndef rm_wikilink(obj):\r\n return bool(re_rm_wikilink.match(six.text_type(obj.title)))\r\n\r\ndef rm_tag(obj):\r\n return six.text_type(obj.tag) in {\"ref\", \"table\"}\r\n\r\ndef rm_template(obj):\r\n return obj.name.lower() in {\"reflist\", \"notelist\", \"notelist-ua\", \"notelist-lr\", \"notelist-ur\", \"notelist-lg\"}\r\n\r\ndef try_remove_obj(obj, section):\r\n try:\r\n section.remove(obj)\r\n except ValueError:\r\n # For unknown reasons, objects are sometimes not found.\r\n pass\r\n\r\nsection_text = []\r\nfor section in wikicode.get_sections(flat=True, include_lead=True, include_headings=True):\r\n for obj in section.ifilter_wikilinks(matches=rm_wikilink, recursive=True):\r\n try_remove_obj(obj, section)\r\n for obj in section.ifilter_templates(matches=rm_template, recursive=True):\r\n try_remove_obj(obj, section)\r\n for obj in section.ifilter_tags(matches=rm_tag, recursive=True):\r\n try_remove_obj(obj, section)\r\n\r\n section_text.append(section.strip_code().strip())\r\n```","embeddings":[-0.2439157814,-0.1970420778,0.0670820624,0.3134252429,0.0271570385,-0.0312242601,-0.1195121184,0.2109857798,0.3312293291,0.1418250799,0.3229496777,0.0170424618,0.0512088351,-0.4928643703,-0.1744854003,-0.0037909967,0.2605278194,0.126694575,-0.0522866212,-0.2089977413,0.0076541505,0.142170921,-0.4332667291,0.2717127204,-0.2520084083,0.1304321736,0.0929765031,-0.0105750877,0.0373966098,-0.469048351,-0.0403118506,-0.2240466177,0.1670672446,0.2844329476,-0.0001267064,-0.2108095735,0.4160577655,-0.0950901434,-0.18317312,-0.1716534048,0.0071890042,-0.1880869716,-0.044271756,-0.3014479578,0.2884559929,0.0185891744,0.2699138522,-0.286364913,0.3683305085,0.2292597592,0.0456214696,0.1972637475,0.2272142768,0.0291248728,0.450504601,0.0643212572,0.1650123149,-0.052069135,-0.0432651341,-0.0965783298,0.0542214997,0.2810028493,-0.3560197651,-0.1097043753,-0.1074362621,-0.03395373,0.3892698288,-0.2821047604,0.4064489007,0.0963136032,0.3552184999,0.0164924208,0.2594856918,0.0158013031,-0.0724432319,0.2757148743,0.3752604723,0.5088477135,-0.3879244924,0.0651457906,0.3225114644,-0.0166436248,-0.0790882036,0.2455794513,-0.0721082017,0.7677217126,0.1227155477,0.3747815788,-0.0259417295,-0.0633261204,0.0517114289,-0.3146010637,-0.0190164428,0.353582263,-0.1086524203,0.159908995,-0.156347692,-0.0853176191,-0.0002928512,-0.4934480786,-0.4367776811,0.0614360645,0.1088035107,0.2573606372,0.5199109316,-0.1411513686,0.1863741428,-0.0718507543,-0.0114212791,-0.0882609114,-0.0732454658,0.0659644008,-0.0933441818,-0.0662893206,-0.332628876,0.106274195,0.1409582198,-0.0423314087,-0.0592529289,0.2894517779,-0.3620603085,-0.0768241808,-0.2041373998,0.4364101887,-0.0336193815,0.5642702579,0.2208687216,0.0534734465,-0.1216491312,-0.6564360857,0.0444866419,0.1973567754,-0.3663577437,0.1661610454,-0.2031689584,0.2760706842,0.1157172173,0.1909150332,-0.070945099,-0.6753615141,-0.1433453411,-0.1477840245,-0.1164991707,0.0726512447,0.3854917884,0.4602942765,0.2932551503,-0.3723100126,0.005006732,0.3135888278,-0.3685143888,-0.0950808302,-0.3885682821,0.0424730852,0.0737980381,0.1866634488,-0.49588269,-0.0535191745,-0.0021891752,-0.1071103737,0.0766583607,-0.1638298184,-0.0927644372,-0.0717015788,-0.0266354829,0.3540619016,-0.0045451773,0.0871735662,-0.1042548344,0.0562980101,0.2985343039,0.1747146845,-0.1192095578,0.6493728757,-0.4247390032,0.6064457893,0.1375234872,-0.2236287892,0.0176330581,0.1443516314,0.0657686144,-0.2023114115,0.0165342167,-0.1092079058,0.1594649553,-0.1335683316,-0.4115052819,0.1158019677,0.0724975392,0.3051943481,-0.2148230076,-0.2910426259,0.1255895048,0.0655348375,0.2211541384,-0.1284651458,0.1120066941,0.4881643355,0.4339236617,-0.1908745766,0.101739496,0.0762710497,0.0587165207,0.1764653176,0.0147018321,0.0942199752,-0.0240320209,-0.0028472797,-0.0596542954,0.3432411849,-0.2033087164,-0.0566712767,-0.206263721,0.2809075713,-0.1454792321,-0.0117504541,0.0624536723,0.1869612336,-0.23657915,0.4068617225,0.0701657161,-0.1439375728,-0.1734018922,0.1468380094,0.0166568328,0.2651125193,-0.2953373194,-0.1468981802,-0.3140820563,-0.1859704703,0.4367952943,0.0147473076,-0.18538858,0.1461973488,0.2881453633,0.1721403003,-0.2892115712,-0.2493927181,0.0595733747,-0.4366012216,0.246480599,0.3074889183,0.3146463335,-0.1487218589,0.2846551239,0.0954824239,0.389208138,0.1007718593,-0.1450612098,0.1643443555,0.1466246843,-0.0050317668,0.3946208656,-0.0911520943,0.2289427221,0.5108888745,-0.273743093,-0.3508856893,0.2493451983,-0.1987016648,0.4050527811,-0.0175306182,0.2443616241,0.1115922183,-0.3060934544,-0.1200667471,0.2761844695,-0.4623517096,0.0935930386,-0.0911205113,-0.1070426106,0.3911009729,0.1414718926,0.0316717476,0.5108950138,0.1668330282,-0.0531923659,0.0966076776,-0.1132935286,-0.01941186,-0.2840810716,-0.2290095836,-0.2573992014,0.1794416904,-0.2868689001,0.0622661002,-0.3053764701,0.0137665346,-0.3138224781,0.1955426037,-0.4067344368,-0.216516301,-0.065108791,0.2800655663,-0.2110290676,0.2325728387,0.1181807816,0.0451997556,0.0876482502,0.3888649344,-0.2219836414,-0.3762750328,-0.5698435307,-0.2166003734,0.496360898,0.1641568094,0.102134265,0.0199719369,-0.0413026735,-0.3580471277,-0.1427371949,0.4228231907,-0.0582674593,-0.1273169518,-0.0762526169,0.2410059571,-0.1527664065,-0.3478518426,0.0598887801,-0.439645797,-0.0904707387,0.118610926,-0.0064835497,-0.1578289121,0.0286916588,-0.0873538032,-0.1545055509,-0.0583377182,-0.3719154,-0.2516958416,0.0327769816,-0.2780528963,0.0770787075,0.1496904194,-0.3611339629,-0.2150790244,-0.270290345,0.087501727,0.1113564372,0.0349699743,-0.2326234132,0.4163252413,-0.1056798473,0.1810459048,0.0059799314,-0.49488011,0.224997893,0.1587656289,-0.0572449118,-0.3003888428,0.1743222624,0.2831252217,-0.0949867666,0.110473983,0.0769067183,-0.0483278446,-0.2259436697,-0.2423432618,0.3762096763,-0.097295396,0.4344439209,0.1302692443,0.6834610105,0.0098076779,0.2261824459,0.3966525793,-0.0428141393,-0.0869307369,0.0329573266,0.0051952377,0.2468228787,0.0007923529,-0.2669427395,0.5544753671,-0.2428961545,-0.2323537618,-0.2343099415,0.0112680588,-0.5420507193,-0.4312593937,0.1656498462,0.3878288567,0.3233222663,-0.1296134889,0.2052385509,-0.0331263952,-0.2678649426,-0.067101039,0.162764594,-0.1086312532,-0.0825973749,-0.1447860003,0.0598139428,-0.0589149073,0.3201695383,0.2619362473,0.3989565969,0.0827726945,0.0591874607,0.1992621273,0.0684879646,0.4733504355,0.136343658,0.2132397741,-0.1031907871,0.09984079,-0.2521496117,-0.3287770152,-0.2264018953,-0.2229650617,0.4933878779,-0.2217371017,-0.4186546504,0.0808050185,-0.0878418759,0.0857847556,-0.0771816224,-0.0917203501,-0.0175335929,-0.2081461996,-0.5254966617,0.225408107,0.1914092451,0.1042892411,0.2822269499,0.2272935212,0.1271851659,0.0314455591,-0.039807979,-0.0457646325,0.1292539388,-0.0511221811,-0.2227943838,-0.2396635711,0.1692955047,-0.0000278021,0.1498710662,0.5779361725,-0.2408649474,-0.1187972873,0.0199819133,0.2297452092,0.3995102942,0.0831126794,0.1178379208,0.245277822,0.2715998888,-0.0809119269,0.2353726476,0.2509082258,-0.0124395182,-0.2929123044,0.0465999283,0.4117345512,0.0055364058,-0.035712909,-0.0581246093,0.0718830302,-0.2986934185,0.4252857566,0.2390448898,1.3299759626,0.1397980452,0.1313474625,0.140861243,0.0778507069,0.7298778892,-0.6040632725,-0.0365348086,-0.28748101,0.4025506973,-0.1426401734,0.0622505955,0.1719701439,0.0775756836,-0.4281995296,0.3375902176,0.140128687,0.5520429015,0.1729460657,0.1936335862,0.1028396115,-0.1207297072,-0.0260301176,-0.0211091246,-0.3125188649,0.0798731744,-0.2915139198,0.0305853505,0.2290643603,-0.4973370433,-0.6621001363,0.1700566411,-0.0738014504,0.3350213766,-0.1431546658,0.0403593741,0.3150550127,-0.1090003774,0.0684927925,0.2554074526,-0.1021942943,0.3669730723,-0.1564946771,-0.3252677023,-0.194176808,0.0501585864,-0.0715721548,-0.1006421372,-0.23043181,0.2871089876,0.0131227449,-0.1832987219,-0.2851789892,0.0635359138,0.4719486237,0.1042991504,-0.2770496011,0.0652645379,-0.0046193725,-0.0838661194,-0.0041122073,-0.1452438831,-0.1851820648,-0.0104157897,0.2468069196,-0.224134773,0.01138571,0.3839807212,-0.0465630554,-0.4790755808,0.4670300782,0.3768467009,-0.2421934456,-0.2099692672,-0.1298051924,-0.5138652921,-0.6772114635,-0.0528040677,-0.1277865171,0.0812091976,-0.0430266447,0.2475567162,0.2362623513,-0.0568304434,0.0207592733,-0.3045709133,0.2019405067,-0.0589142181,-0.4229645431,-0.0700269714,0.0111651057,0.2940684855,-0.0126251522,-0.0341525264,-0.0879460201,0.1143524125,-0.3084953129,0.2707249224,0.2566701174,0.0909439698,-0.3615843058,0.0093988618,0.0135606369,0.2861456275,-0.0697459131,-0.0071091615,-0.1328020692,0.2703225315,0.3827153146,-0.2338981181,0.1011144444,-0.0475615487,0.2174996883,-0.1566308439,-0.0396186188,-0.0734601766,-0.1314592659,0.2485080957,0.2089192569,0.2581070065,-0.4424045086,0.1211530715,0.0721941665,0.5967171192,0.1402492672,0.1848926544,0.2156502753,-0.0069222795,-0.2561196685,0.2114630491,0.3707139194,0.130171299,0.1372703314,-0.1082836837,-0.1917605847,0.0671046749,-0.1397372931,0.1556293517,-0.3254968524,-0.2894349396,0.5366226435,0.0886183679,-0.157933265,0.2036015242,0.4370489419,0.1117445379,-0.2860326767,0.1864867955,0.0879067183,-0.4010625184,0.262422353,0.1603426784,0.3800903857,-0.050483346,0.0829644576,-0.0977439806,-0.1310412139,0.1996890455,-0.1301872134,-0.0960257947,0.3383076191,0.3912017941,-0.1181891114,0.214364782,0.1854749322,0.1639163345,-0.0239626858,-0.2019262165,0.1884983629,0.50477916,0.1020760089,-0.0256599002,-0.0984646901,-0.2070857733,0.1636532843,-0.0989288613,-0.4667608738,0.3099548519,-0.1447390169,0.0302568916,-0.2093651593,-0.0231614411,-0.3640719056,0.2160027027,0.1275570542,-0.0629101768,-0.3295396268,0.3941777945,-0.3697271943,-0.1015253887,-0.2433308214,0.0054133786,-0.1028017625,-0.3456289172,0.1546686143,0.0314430669,-0.0609713607,-0.073105745,0.1959453076,0.5290171504,0.4877013266,-0.527261734,0.0753338635,-0.067380026,0.2905634344,0.0232849047,0.0573291965,0.134273082,0.36784935,0.0767161325,0.0019117357,-0.0978144109,-0.3419983387,0.2982867062,-0.0415504724,0.0392421111,-0.341746062,-0.0561950244,0.2586339712,-0.2751139998,0.0397304706,-0.5000792146,0.160036087,0.1681087613,0.0802950114,0.0784617886,-0.0440976508,-0.0353153832,-0.1571895629,0.2149786949,0.486931324,0.2715578973,-0.2093726546,-0.5163052678,-0.5368038416,0.4130154252,-0.3302844465,-0.2191963643,-0.0158568546,0.0947176665,0.1577572823,0.4018924534,0.1277690083,0.0380507149,-0.3444069028,0.3325327933,-0.2369214892,-0.1468168795,0.3220621049,-0.0791135132,-0.0747827441,-0.3180965483,0.1554449499,-0.1519023776,-0.0578025877,-0.1321987361,-0.1099865809,-0.0585876144,-0.2227220237,0.5239775777,0.1362691373,0.4043087065,-0.0057475958,0.0159791131,-0.1776844859,-0.0300557371,-0.0219237693,-0.0359030478,0.0388205051,0.1936432421,-0.0677346066,0.1720441282,-0.4300882518,0.2831495702,0.0526977926,0.1748317033,-0.398062408,-0.0310369059,-0.2528933585,0.2896255851,-0.2356533259,-0.0801538154,0.0222835541,0.2795817554,-0.3736264706,-0.1903099716,0.2450011969,-0.4076629877,-0.5217408538,-0.2219766676,0.2311699241,-0.3900434077,0.0033997113,-0.1855126768,0.0024275794,0.2290036231,0.1364833862,-0.2730088234,-0.2439188808,-0.0786357298,-0.2442163527,-0.0058558127,-0.0487845093,0.1136256978,-0.3359308839,0.1151103154,-0.4321381152]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"Not sure why we're having this issue. Maybe could you get also the file that's causing that ?","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":18,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nNot sure why we're having this issue. Maybe could you get also the file that's causing that ?","embeddings":[-0.2438737899,-0.1548451036,0.0903029814,0.4575005174,0.1244493946,-0.0543386936,-0.1172973961,0.2066474408,0.336443454,0.3008716106,0.2952870727,-0.0024946679,0.052222982,-0.3813283145,-0.0767171532,-0.0030087871,0.3680462241,0.1951178759,0.0440426283,-0.2419860959,-0.0590003878,0.2005376667,-0.3453764021,0.3041818738,-0.3107658625,0.0585562922,0.1418303251,-0.0247239042,-0.0764364898,-0.3120390773,0.0276921,-0.2750107944,0.1651804149,0.3228823543,-0.0001273461,-0.1857159138,0.5115447044,-0.1040547341,-0.1468497366,-0.2338778526,-0.0839670599,-0.1510858685,-0.0317203365,-0.3054434657,0.313524127,0.0639472082,0.2530777156,-0.2387647778,0.260045588,0.2302283645,0.0824697837,0.2199033499,0.2952759862,-0.0069826036,0.5136982203,0.1219803318,0.204753235,-0.0710389838,-0.0341430269,-0.0303169619,0.1266936362,0.0969860628,-0.4162338376,-0.2447789013,-0.1444558054,-0.0935338661,0.307528168,-0.4182212055,0.4121519327,0.1747305989,0.5269611478,-0.0202695616,0.350112766,0.1336372197,-0.0502374619,0.2415279597,0.4162761867,0.5217395425,-0.3993159235,0.0340274237,0.3900630772,-0.1469513923,-0.042469427,0.2651845813,-0.1742010415,0.8081724048,0.0326359831,0.3084394336,0.0507775359,-0.0183148943,-0.0851952806,-0.3319167495,-0.071087487,0.2830992341,-0.1364086866,0.032097768,-0.1755914986,-0.0749992207,0.0263354573,-0.5093848705,-0.4673197269,0.048002068,0.0184483379,0.2484384626,0.4564214647,-0.1146916524,0.1845860332,-0.0138370367,-0.0089980075,-0.144598335,-0.0538723581,0.0191114284,-0.2315288186,-0.0125745088,-0.4853001237,0.0780624151,0.095707573,-0.1704505533,-0.0402032472,0.2572525144,-0.2312712967,-0.0955126882,-0.1507345587,0.4952979088,-0.0619028546,0.391949147,-0.0485467128,0.0242672097,-0.137969628,-0.609387219,0.0128734466,0.1973122656,-0.3547971845,-0.0110471919,-0.1646290123,0.2207181454,0.133810088,0.3228284419,-0.0421385616,-0.4779711664,-0.0261187013,-0.2950364649,-0.1444501579,0.0541155748,0.4805473387,0.4583336413,0.199652046,-0.3672831357,-0.0564435907,0.4087544084,-0.4907042384,-0.1547195315,-0.5020549893,0.0364796184,0.0233863983,0.1601269841,-0.5785993934,0.0256016534,-0.0694199428,-0.1577776223,0.0533760712,-0.223376587,-0.1592518687,-0.1376728117,0.1006360278,0.3055128753,-0.1845821142,0.011036451,-0.0979872718,-0.1748080999,0.372736752,0.1509964615,-0.1719998866,0.659750104,-0.4492008686,0.5568723679,0.3406916559,-0.2888245583,-0.0907193422,0.0462819487,-0.064474389,-0.206685707,0.0266750138,-0.1909818053,0.2845521867,-0.0467438735,-0.3265430033,0.095997259,0.0915887803,0.2053874135,-0.2147358805,-0.3104133606,0.2907601893,0.0747240409,0.1716153026,-0.0663381964,0.171904996,0.465339154,0.4322321713,-0.2057147175,0.1602537483,0.115256682,0.1596601456,0.2181689441,-0.1131922901,0.1417740732,0.0020725175,-0.0545437187,0.0297569558,0.2059301436,-0.1255842447,-0.0830164254,-0.1817033142,0.2146631181,-0.0889083669,0.0787186027,0.0539172553,0.1325584203,-0.2125540823,0.3692745268,0.0476852544,-0.1967649013,-0.2540505528,0.110928975,0.1038986072,0.2972656786,-0.1605814397,-0.0571832471,-0.3508009911,-0.1678777635,0.5068054199,-0.013373916,-0.1443748772,0.0362109318,0.2480144054,0.0601928681,-0.272565335,-0.2072727382,-0.0175658297,-0.2871756554,0.2105044723,0.1699792594,0.2152316719,-0.1655080914,0.3415623009,0.088574186,0.3863001168,0.0712016597,-0.1960386485,0.1259265393,0.2389337122,0.0077397125,0.4671167731,-0.0711373538,0.2976313531,0.4176380336,-0.092295073,-0.2519229949,0.1286691278,-0.1277332604,0.5343972445,-0.069598116,0.1919314414,0.205151096,-0.3437318802,-0.0107756574,0.2965949774,-0.3911828697,0.1748312116,-0.0276965648,-0.1048231795,0.273758918,0.2530734837,0.0135616856,0.5135731697,0.0879016817,0.0418716595,0.1194233671,-0.261066705,-0.0293091163,-0.2766619623,-0.142339021,-0.2583797574,0.2075479627,-0.2546713054,-0.1088880748,-0.2646128833,-0.0210305918,-0.3601520658,0.1243790835,-0.3825925887,-0.0819281489,-0.161075145,0.254398793,-0.1200082675,0.1283909976,-0.0101331845,0.1112431213,-0.0355769284,0.2304631025,-0.2162801474,-0.2744900584,-0.5513046384,-0.152410239,0.5267803669,0.1114875078,0.0004289396,-0.0261537973,0.0205525216,-0.2660392225,-0.1641802639,0.4210189879,-0.0297951121,-0.0914885849,-0.085352622,0.1795302331,-0.1181223467,-0.3170990348,0.0051146424,-0.3451727033,-0.1308822036,0.0749335438,0.0684557408,-0.1905052811,0.06678655,-0.0430342034,-0.3454267383,-0.0301899146,-0.2688142955,-0.1228383631,0.1198506206,-0.5135689974,-0.0029473407,0.1931803226,-0.3985514343,-0.1948124766,-0.2629196644,0.0681876764,0.1352652162,-0.0106285997,-0.24457784,0.4565593898,0.0086395303,0.190732792,-0.0298137143,-0.6047429442,0.340696305,0.1188653782,-0.0786786899,-0.2802816331,0.0917687714,0.3030137122,-0.149182871,0.1661013663,0.034210708,-0.0800376087,-0.2589765191,-0.3078549802,0.3895877302,-0.0943142399,0.4309289753,-0.0003129902,0.5969522595,-0.0308421291,0.2419363856,0.3254552186,-0.0590361729,-0.025696395,-0.028413089,-0.082196109,0.3089119792,0.1155977473,-0.2447577864,0.5990236402,-0.2307508588,-0.2992627323,-0.2009941041,-0.1291095316,-0.4419801235,-0.3891987503,0.1840630472,0.3259333968,0.2409102321,-0.0811513364,0.1840967685,-0.0948016271,-0.3454446495,-0.0010839218,0.2366651595,-0.129824549,-0.0529925935,-0.3019225299,0.0408882909,-0.2444856614,0.2875695527,0.1513236016,0.4145076573,-0.0659282953,0.0903609246,0.1396747679,0.1131952852,0.6657505631,0.187389791,0.2014040798,-0.0544616655,0.1106777787,-0.2019364536,-0.349933058,-0.2815024555,0.0279457532,0.481844902,0.011475008,-0.4433548748,0.0159885455,-0.1085627526,0.0753663629,-0.0337470844,-0.0558141693,-0.1241060272,-0.170123443,-0.4060508013,0.2221248448,0.0994759351,0.2381229252,0.2016291022,0.1031922847,0.0659616664,0.079952307,0.0138164517,-0.1495929807,0.1812981367,0.1463815123,-0.2607524395,-0.1961850077,0.2767329812,0.0634215102,0.0852847546,0.5467895269,-0.2350268811,0.03947144,-0.1000388563,0.2835973203,0.319547236,0.0626400709,0.2859419882,0.1207229197,0.2047338635,-0.092656672,0.3206840754,0.2085364163,0.0531719923,-0.3946620226,0.0233648382,0.2817932367,0.0174179077,-0.0482948795,0.1377719641,0.0552910343,-0.2594876587,0.3033360541,0.1993102282,1.3601082563,0.2606369257,0.0605073944,0.1108948216,0.004398392,0.7963571548,-0.629807651,-0.0189270526,-0.3128011227,0.1327948868,-0.1389375329,-0.0097087817,0.1000773236,0.1444195956,-0.3683395386,0.2792601585,0.1070212275,0.6044886112,0.1850079596,0.2106743604,0.1967149228,-0.0565817989,-0.0108237779,-0.0222474486,-0.2962784469,0.1750328988,-0.3528541327,0.0459063239,0.0608250611,-0.3904224038,-0.5212930441,0.1844305247,-0.1486295164,0.2001838684,-0.0063782446,-0.0557130091,0.3860985041,-0.0757062286,0.1267178655,0.3899019063,-0.1791909933,0.3868409693,-0.112023145,-0.3065803051,-0.0284324102,0.0907422006,-0.2048860192,-0.1419510692,-0.3151187301,0.3814617395,-0.0807908103,-0.2983235121,-0.301841706,0.002633241,0.4286645651,0.0016240231,-0.4058823287,0.0637470409,0.1333778352,-0.0259324368,0.0339153633,-0.1300430298,-0.1188359708,-0.0287064184,0.3110334277,-0.2470197827,-0.0372652411,0.3636563122,-0.0766465738,-0.4688628018,0.5109813213,0.3010251522,-0.2208765,-0.1478509456,-0.0444754213,-0.4849438965,-0.6720795035,-0.1798169464,-0.1361635327,0.0911537707,-0.159246698,0.2470299006,0.3108221591,-0.0055389232,0.1253396869,-0.3382136822,0.0786659196,-0.0317672901,-0.4044038951,-0.0921047255,0.1533713192,0.165863812,-0.0442025624,-0.0804517046,-0.0990913585,0.0889642835,-0.3104228675,0.1767553836,0.4049191475,0.1613198221,-0.2886877656,-0.0289093722,-0.0078201788,0.2113430202,-0.0162184332,-0.0310199652,-0.1206210405,0.2824006379,0.3431390524,-0.1013106927,0.134145543,-0.0402459949,0.1312258244,-0.063074097,-0.0840538219,-0.1243697032,-0.1189361364,0.2311473638,0.1561111659,0.3033761084,-0.6175177097,0.1567089111,0.2122364044,0.5182578564,0.0101781692,0.2413633317,0.2124583423,-0.0096967574,-0.2173803747,0.1985197216,0.3253695071,0.191080153,0.0740262419,-0.0056872442,-0.029286528,0.1216231808,-0.2189952135,0.1084223166,-0.3840617239,-0.2678412795,0.6546400785,0.1048715487,-0.1199909225,0.1165197119,0.5963562727,0.1986391991,-0.3658186495,0.1173019335,0.0538518988,-0.4070931971,0.338241756,0.1766294241,0.3714925051,0.008865484,0.121261768,0.0626199171,-0.1921663731,0.2916052341,0.0621880293,-0.037034113,0.2124628425,0.5630069971,-0.0889249146,0.0685871243,0.1866466254,0.0757058337,-0.0532362312,-0.2476690412,0.1553548276,0.5270798206,0.0840765461,-0.1004327461,-0.1171279848,-0.0364124402,0.1418990344,-0.1766620725,-0.4514766335,0.2105668336,-0.3536434472,0.081978038,-0.3765353858,-0.0953383967,-0.2704105079,0.2513715625,0.1225418821,-0.0408375449,-0.4008167982,0.325764358,-0.4385765493,-0.0100147342,0.0306566488,0.1095701233,-0.1127486825,-0.3423266113,0.1936748177,-0.1967988312,0.0354592614,-0.1107283086,0.223788619,0.4590551853,0.5882257223,-0.3402910829,-0.0576722547,0.1371684372,0.2367716283,0.0181316342,0.1504067332,0.2121137083,0.4006197155,0.0605678633,0.0274565071,-0.1269885004,-0.2661486268,0.3369753063,0.0030717407,0.0527200885,-0.335983932,-0.1048269719,0.1239413619,-0.2843911052,-0.0802126601,-0.5245041251,0.2972145975,0.1381940395,0.1650841236,0.0839600116,-0.113404274,-0.026124401,-0.1292490512,0.1829048395,0.3374114037,0.23513107,-0.2557026446,-0.5928301215,-0.5923689008,0.5155175328,-0.2085255384,-0.1719621718,-0.0731403008,0.0601413287,0.0623319633,0.3624555171,0.1018301174,0.1438013464,-0.3306406736,0.2046477199,-0.2297417819,-0.0936587155,0.1909531057,-0.0157703143,-0.0849764273,-0.3382996619,0.2290946692,-0.1622266024,-0.0698235035,-0.1829075068,-0.2145685703,-0.136461705,-0.2413012236,0.5710756779,0.1211626679,0.4785780609,-0.0656350404,0.1872144043,-0.0684072673,0.0168244224,-0.0815816596,0.0378776118,-0.0192383267,0.1439688802,0.0423603952,0.3129379451,-0.3990135491,0.3354873657,0.1386944652,0.2268176526,-0.362531811,-0.0345895812,-0.3117889166,0.0808433667,-0.0334708095,-0.0563613661,0.0263212621,0.1504728794,-0.3717042506,-0.0480518006,0.1737814844,-0.4795393944,-0.3824426532,-0.2534619868,0.2423761785,-0.1136700734,-0.2396759093,-0.1374624521,0.0430728756,0.256423533,0.1280902177,-0.1610542834,-0.1142303646,-0.2987760901,-0.1495218575,-0.0021701669,0.0535574518,0.1717064977,-0.4144586027,0.1750483215,-0.4345165193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"thanks for your answer.\r\nHow can I know which file is causing the issue ? \r\nI am trying to load the spanish wikipedia data. ","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":24,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nthanks for your answer.\r\nHow can I know which file is causing the issue ? \r\nI am trying to load the spanish wikipedia data. ","embeddings":[-0.3805398047,-0.3404875696,0.0569592193,0.6597746015,-0.023943508,-0.1370638311,-0.2214765549,0.355119735,0.3205151558,0.2352275848,0.393688798,-0.0477157906,0.0279491432,-0.4523819983,0.1012191325,0.1489469558,0.4252944291,0.1956285387,0.1554450542,-0.2860964835,-0.1436806172,0.3324868679,-0.2521990538,0.476505816,-0.3799897134,0.1468522549,0.1703818142,0.1342841834,-0.0388246886,-0.3340257406,0.1467204094,-0.2287000865,0.2494850755,0.4108291864,-0.0001294859,-0.1358284801,0.4792954922,-0.2167567611,-0.0711504743,-0.2471745759,-0.1766093224,-0.1024994627,0.0518087409,-0.2762674391,0.3678812981,-0.0481221825,0.40333125,-0.2109247744,0.2742952406,0.3901074529,0.0669516176,0.2184573114,0.168709293,-0.0131089808,0.555941999,-0.0039246865,0.2948894203,-0.0138163939,0.0548495725,-0.044605121,0.0094287368,0.0913252681,-0.411995858,-0.1316954643,-0.0811175779,-0.1147457138,0.3158606291,-0.3811001182,0.3990839422,0.2864823639,0.5265470147,0.0244398471,0.453728646,0.1320969164,-0.2025650889,0.3474864364,0.5818397999,0.4023046792,-0.3517132998,-0.0165661462,0.3878603876,-0.0085276598,-0.0155066941,0.239936471,-0.3587429523,0.7549949884,-0.0231264494,0.3411198556,0.1030806378,-0.020427309,-0.182003051,-0.2388801873,0.0318420939,0.2876210213,-0.1689518988,0.0090212068,-0.0637496784,-0.137619853,-0.0729253441,-0.4824035764,-0.395334065,0.093685396,0.0208178386,0.2795303166,0.4600297213,-0.1279717088,0.2268948555,-0.1281471997,0.036125835,-0.0654530227,-0.1618681103,0.0330298841,-0.3590043187,-0.1693919301,-0.3182511628,-0.0556118451,0.1524051279,-0.2356127948,-0.1032150686,0.2423217148,-0.2417160124,-0.2401424646,-0.0160587989,0.5928182602,0.0460029356,0.361269474,-0.0883764476,-0.1704967618,-0.1643166393,-0.5626623631,0.0214125067,0.1993014365,-0.3200522363,-0.0310706757,-0.1405968517,0.1645071805,0.1247262135,0.2809581459,-0.1108282879,-0.5334955454,-0.0293393154,-0.3077511787,-0.0599958487,0.1015643328,0.3478440344,0.3976214528,0.2785716951,-0.4330412447,-0.122121498,0.2768184841,-0.5401117206,-0.1904270947,-0.3655868173,0.0073637529,-0.0813632384,0.2184054852,-0.5214793086,0.0352887176,-0.1589038968,-0.3161871135,0.0865637138,-0.2566933036,-0.0714406669,-0.1356313676,0.0950770602,0.3783811629,-0.1570718437,-0.086536698,-0.1100410819,-0.2945402563,0.2481389493,0.0776856914,-0.1522156,0.5089014769,-0.3929197192,0.460716784,0.4020156264,-0.3885554075,-0.0774063244,0.0032989245,-0.1051693261,-0.1443250626,0.1704610437,-0.0770815238,0.1641698182,-0.1816911995,-0.2103979886,0.0381533839,0.158132419,0.1202652752,-0.2180195153,-0.2795444429,0.0681381971,0.1130153611,0.1243659928,-0.223666653,0.1109133661,0.4432348609,0.2955294549,-0.0204175524,-0.0141053721,0.1304380149,0.1346194148,0.1634792536,-0.0504479669,0.0271425527,-0.0039531025,-0.0722234473,-0.0548003241,-0.0843851492,-0.2033379227,-0.0244715977,-0.1983307004,0.1904845685,-0.1336470246,0.1339884251,0.0251239911,0.2611249983,-0.2519678473,0.2475951165,-0.1870209277,-0.1587837785,-0.2405169755,0.1432477534,0.0596063845,0.1816093028,0.0290165041,0.0079204086,-0.198838532,-0.258022964,0.5626895428,-0.1092529371,-0.1604559869,0.0930312574,0.2045786679,0.239459902,-0.1246015728,-0.0929321721,-0.1977773458,-0.2928369343,0.129445225,0.1361512244,0.1311367452,-0.0346000828,0.357167542,0.0213129688,0.2833710611,0.0054765814,-0.1536316872,0.0290363058,0.1711270958,-0.035795588,0.5272099376,-0.0998658389,0.3510566056,0.4659900367,0.0529832728,-0.0824356079,0.0491601788,-0.1945196986,0.747695744,-0.160817802,0.1886676848,0.2848615944,-0.3441036642,-0.0493515581,0.2152215242,-0.2402270883,0.1980990767,-0.0112592597,-0.1218492389,0.1306510121,0.1797141433,-0.0068306462,0.5171731114,0.0912542343,-0.0789402574,0.1184268445,-0.3818058968,-0.045304656,-0.3620023131,-0.1243045852,-0.3254980743,0.2499777228,-0.2731672525,0.079643786,-0.2528131604,0.0217751395,-0.5295092463,0.0954803526,-0.382840097,-0.0397872925,-0.0676159784,0.2131757289,-0.0127077838,0.0428512357,-0.063901484,0.1982924491,-0.120386824,0.1586452127,-0.1872363985,-0.2542199492,-0.567986846,-0.1637690514,0.6536368132,0.1832839996,-0.0434576534,-0.2405002713,0.0162723139,-0.1531118006,-0.1163961217,0.4006673098,0.0613882281,0.0006478506,-0.1108780801,0.2774047256,-0.0537718982,-0.3318936527,0.0361514315,-0.3673259914,-0.1134881973,0.0765604898,0.1452403814,-0.1888056844,0.2144370824,-0.1124294773,-0.2642555237,-0.1026667506,-0.3206094801,-0.1488945931,0.202453047,-0.6096335649,-0.2209496498,0.3382605612,-0.2790539861,-0.0657678768,-0.2418012172,-0.0273332875,0.0993971154,0.1111358032,-0.0780663788,0.4412772954,0.1103961468,0.0849888846,0.0741496533,-0.5340140462,0.2013530284,0.0820121095,-0.084700264,-0.2322696149,-0.0517664589,0.345007062,-0.0914912894,0.1431186944,-0.0692384392,-0.0494285412,-0.2028441131,-0.3041860759,0.3847751617,-0.056521073,0.3137366772,-0.0374561213,0.6825994253,0.0240975507,0.0873805359,0.2980571985,-0.0160124432,-0.1377187073,0.0320341997,-0.0759115592,0.2200619578,0.1820499748,-0.2735163271,0.7422610521,-0.1768074036,-0.2366903126,-0.1763128936,-0.1182141155,-0.6048867702,-0.2558742762,0.1936539412,0.376486361,0.2533043623,-0.2265684754,0.1147230342,-0.1012261137,-0.2793629169,-0.0246510692,0.5404154062,-0.1634945273,-0.0173954833,-0.2303690612,0.0607682802,0.0173090734,0.3330138326,0.027889898,0.3645259142,-0.3009794056,0.0584818348,0.1851803362,-0.0664419234,0.7375569344,0.2812922299,0.1952385902,-0.1391331106,0.107912086,-0.1095714495,-0.3150516748,-0.2284460217,0.0664265007,0.3795223534,0.1186044291,-0.3667223155,-0.0681228861,-0.0757430568,0.1249100566,-0.0661777928,-0.0422979631,-0.108165741,-0.1178233996,-0.2608387768,0.1928352416,-0.0408295579,0.1302670985,0.2843620479,0.1572833508,0.0470321476,0.0588468947,0.1461844295,0.0110596186,0.1652793139,0.0087653911,-0.3772045076,0.0184284579,0.3099291623,0.1846915781,0.1031495333,0.6288958788,-0.1746204644,0.0785402954,0.0281988885,0.1029504687,0.4168391228,-0.0111830672,0.2691119909,0.0275780968,0.1334395707,-0.0615870655,0.2459012717,0.3112945557,0.1098641828,-0.535733819,0.0679985657,0.5120007992,0.0847099647,-0.1974281669,0.1787637472,-0.1756362766,-0.2289909422,0.1538375318,0.2000306547,1.449960947,0.1524119377,0.0745109469,0.0575494282,-0.0130344704,0.6243776083,-0.7486425042,-0.0916425958,-0.2428949326,0.0912720338,-0.2430552542,-0.1300571114,0.2165142298,0.1415092647,-0.3455143273,0.2901128829,0.1440196484,0.5426395535,0.1806188673,0.189116165,0.1602579653,0.1230422333,-0.2093893439,-0.0700892136,-0.3668535352,0.3385096788,-0.2858900726,0.0531713068,-0.0187168811,-0.3675303757,-0.491145432,-0.0380056798,-0.1358438432,0.1349930614,-0.0147033622,-0.1296633631,0.2623561025,0.0590343289,0.0528733023,0.1975555867,-0.1332390606,0.4760961831,-0.0304879863,-0.1726081371,-0.0841306224,0.0062038209,-0.136168614,-0.1384036094,-0.217701152,0.4074857831,-0.0213674884,-0.2046982646,-0.1616490036,0.0572449788,0.4449587166,0.0596503951,-0.5058220625,-0.0652864203,0.0823032707,0.0425924063,0.0263170376,-0.051326707,-0.0141033726,-0.1445216984,0.1523774117,-0.2435978502,-0.0430079624,0.2790696323,-0.1172523871,-0.4595456421,0.4491488039,0.0857147127,-0.1793924868,-0.0937565193,0.0898449048,-0.4517296851,-0.5832412243,-0.1965381205,-0.0496163219,-0.0063538854,-0.0469761826,0.2073540092,0.368640244,-0.0207541473,0.0667915493,-0.4061916471,-0.0126442946,-0.034560591,-0.2390427142,0.0164285079,0.0966110155,-0.0399434417,0.0398916677,-0.0132918339,-0.0764273107,0.1367548108,-0.379655242,0.1044781655,0.4097523391,0.3177155554,-0.2552176416,-0.1267582327,-0.0576285459,0.0908844993,-0.0615499429,-0.0139187286,-0.091700241,0.3072162569,0.2673517764,-0.1160517558,0.1397969127,-0.1129705682,0.3074871898,-0.0424607322,-0.1413345486,-0.0809102133,-0.0709223598,0.1589272916,0.1567128897,0.0765150115,-0.6586506367,0.2011527717,0.1272816956,0.4618560672,0.1002872959,0.2400373369,0.0315800495,-0.0718209594,-0.2901139557,0.2863998711,0.2572478354,0.1348080337,-0.0249973703,0.0859440863,-0.0142002115,0.1062781885,-0.3425728381,0.1890192628,-0.4209905565,-0.2592411339,0.465523392,0.0506162234,-0.034381412,0.0872588977,0.5667284727,0.0473153405,-0.3760892153,0.0875772983,0.0093162274,-0.327293396,0.3672846854,0.1612736136,0.326990217,0.0370093733,0.2489853352,0.1275031716,-0.0427399389,0.4304687381,0.0737182125,-0.0544305108,0.3077162206,0.5992571712,-0.1523355842,0.14097175,0.2159289718,0.0921862274,-0.0274928659,-0.2794288099,0.1667838395,0.6952610016,0.0112622529,-0.0266947635,-0.0134635679,0.1441679001,0.0314754434,-0.0918721855,-0.5693455935,0.1723654717,-0.3039981723,-0.0418752059,-0.4359058142,-0.1733748168,-0.2718034387,0.2712936401,0.0846751332,0.0075344658,-0.4959266186,0.4745282829,-0.3862531185,0.165972814,-0.001613775,-0.0352983885,-0.0768922716,-0.2510151863,0.2767047584,-0.1009471193,0.0900154933,0.1785899848,0.1078794897,0.4518149495,0.4608573914,-0.1981662959,-0.0717676729,0.14640975,0.1992852092,-0.080668278,0.1715554297,0.0444156043,0.4488538802,0.0613123588,0.0411935523,-0.1205742508,-0.2306892574,0.1679256856,0.0329201259,-0.0584573261,-0.3236364722,-0.2525622845,0.1214088351,-0.3276172876,-0.0439043306,-0.4012697339,0.2954159677,0.0826759487,0.3247374296,0.0995691866,0.0124490662,-0.0507526994,0.0161239952,0.0842737183,0.2671217024,0.2433388084,-0.1979233027,-0.6244456768,-0.7142555118,0.5119056702,-0.1751797795,-0.1437796056,0.0119918073,-0.1430342048,0.1733769476,0.2121593505,0.2462121546,0.0237591602,-0.2725060582,0.3092392385,-0.0916107744,-0.1433475465,0.1217948273,0.0482311286,-0.1714284867,-0.3135451674,0.231139645,-0.1329478174,-0.1006049514,-0.1388890147,-0.2052157968,-0.1508550197,-0.3039788306,0.5101718903,0.0416046679,0.6107798815,-0.1415695697,0.2125366628,0.0572180152,0.0902412757,-0.0502583385,0.0778098032,-0.1217024624,0.2651923895,-0.0064842179,0.4433827698,-0.3522951603,0.3404264748,0.0431978814,0.1644783616,-0.1872654259,-0.2675868571,-0.2149734497,0.0145270526,0.2357357591,0.175092712,0.0116047664,0.0811224133,-0.3216751814,0.0332718752,0.078437008,-0.5061957836,-0.3236168623,-0.0828239471,0.3031010032,-0.20152466,-0.0986846834,-0.0766227543,-0.0515723974,0.3096536994,0.1090635434,-0.1457051933,-0.1472847313,-0.2215688378,-0.1554242522,0.0013425947,0.0195943993,0.2298020571,-0.4885845184,0.2433447391,-0.4309993684]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"Because of the way Apache Beam works we indeed don't have access to the file name at this point in the code.\r\nWe'll have to use some tricks I think :p \r\n\r\nYou can append `filepath` to `title` in `wikipedia.py:L512` for example. [[EDIT: it's L494 my bad]]\r\nThen just do `try:...except:` on the call of `_parse_and_clean_wikicode` L500 I guess.\r\n\r\nThanks for diving into this ! I tried it myself but I run out of memory on my laptop\r\nAs soon as we have the name of the file it should be easier to find what's wrong.","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":95,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nBecause of the way Apache Beam works we indeed don't have access to the file name at this point in the code.\r\nWe'll have to use some tricks I think :p \r\n\r\nYou can append `filepath` to `title` in `wikipedia.py:L512` for example. [[EDIT: it's L494 my bad]]\r\nThen just do `try:...except:` on the call of `_parse_and_clean_wikicode` L500 I guess.\r\n\r\nThanks for diving into this ! I tried it myself but I run out of memory on my laptop\r\nAs soon as we have the name of the file it should be easier to find what's wrong.","embeddings":[-0.1209630445,-0.070509471,0.0564221703,0.3891802132,0.1034142226,0.0170317814,-0.1214456111,0.3558146954,0.2514157593,0.264095217,0.2141057104,-0.0685191453,0.0406434871,-0.4498924315,-0.0869592652,-0.0778382346,0.3508561254,0.123478502,0.0555284396,-0.1732604504,-0.1695192754,0.1376484931,-0.281078428,0.4107494354,-0.1864092797,0.1112359762,0.0803811401,-0.0238146037,-0.0770189911,-0.3404312134,-0.1463585049,-0.168971613,0.2026515007,0.3334973752,-0.0001258567,-0.1855412722,0.5518958569,-0.1446356773,-0.2137812823,-0.1335806847,0.0046767704,-0.2667896748,-0.059188243,-0.3690649867,0.2201614976,-0.0541450605,0.3302631676,-0.2550405562,0.3609259725,0.2255510688,0.0779779851,0.1390962452,0.2775888741,0.0408106744,0.654358089,0.0034429049,0.1290146261,-0.224219203,-0.0136564439,-0.1525859684,-0.0030372231,0.2453565449,-0.2965688109,-0.199443087,-0.0675252303,-0.1138773635,0.2302039415,-0.2996582389,0.3847329617,0.1337920725,0.5107129216,-0.1019761339,0.1809576601,0.0640801936,-0.1153809279,0.2564850152,0.3498593569,0.4780451655,-0.4989578426,0.060281802,0.4844412208,-0.1789774746,-0.1038604304,0.298196286,-0.0074597811,0.7731332779,0.1381369978,0.3352176845,0.0184042919,-0.0671589151,-0.1388245672,-0.2876846492,-0.0027429822,0.3892367482,-0.0984645411,0.1176916435,-0.1336878538,-0.0798086748,0.0004215051,-0.4752697349,-0.3524791896,0.092343621,0.0655017346,0.257134527,0.3952152133,-0.088089034,0.2560520172,-0.1058404669,0.0649590343,-0.0035789935,0.031890519,0.0092387004,-0.1560273468,-0.1251688749,-0.5585864186,0.1572338194,0.1669795066,-0.0611287653,0.0867830813,0.2184611112,-0.198992908,-0.0944182649,-0.1439749002,0.5513176918,-0.0361210816,0.5017178655,0.1081734598,-0.0195120219,-0.2092121243,-0.6590039134,0.072628431,0.2387360185,-0.3837993145,0.0940520912,-0.146284461,0.2706366777,0.0817148611,0.1626767963,-0.1101477593,-0.4127649665,-0.0177707318,-0.1296966374,-0.1106903851,0.0623148791,0.4266298115,0.5009470582,0.2024616897,-0.4213831127,0.0324726477,0.3185638785,-0.4508625567,-0.1133395955,-0.3865263164,0.0219051,-0.0549518429,0.2963490486,-0.511759758,-0.0257374663,-0.0611916408,-0.1805220246,0.0579591729,-0.0974291936,-0.0502929054,-0.1684542298,0.1400205493,0.3438383639,-0.0868770704,0.0030445298,-0.1516425312,-0.1271295696,0.2069042474,0.0125187179,-0.0878881142,0.5803267956,-0.4632230401,0.5278597474,0.4331421256,-0.4172807932,-0.044425793,-0.0509065613,-0.0690591112,-0.1800686568,0.1532539129,-0.24177441,0.2028295994,-0.1651450992,-0.4169232547,0.1614945978,0.0947089121,0.1545333415,-0.2393751293,-0.2976189554,0.1369993687,0.0699921399,0.1159614846,-0.0554039627,0.2703327239,0.5132507086,0.5763164759,-0.2047070414,0.2086592317,0.1209062189,0.0107676694,0.177630797,-0.0026646431,0.1111140475,0.1003481075,-0.0080743851,-0.1155226752,0.19045192,-0.2601393759,-0.0636972785,-0.2423490286,0.1893315464,-0.2007885426,-0.0428602658,0.077681005,0.2335370481,-0.2802915871,0.3116144836,0.0665516779,-0.2075055391,-0.2028139532,0.1360837817,-0.0770246759,0.1176707223,-0.1301356256,-0.0456713773,-0.328032732,-0.1180562451,0.5273607373,0.0116142714,-0.1616954803,0.1149473786,0.2832993865,0.0874383971,-0.1325665861,-0.1691450626,-0.0989626125,-0.3816641569,0.1865923554,0.2674351633,0.2412260473,-0.1106106043,0.1445416361,0.1156604663,0.3446531594,0.1922703385,-0.0792750344,0.1298828423,0.1943041384,0.0106886914,0.2934870124,-0.0030797762,0.3046086431,0.499006778,-0.0706461146,-0.3424893916,0.32921803,-0.2922590375,0.4965381026,-0.113088049,0.1492168903,0.1261763275,-0.3448246121,0.0666225404,0.3503163457,-0.4112823904,-0.0033078764,0.0153660644,-0.1353278905,0.1966359615,0.3104834259,0.0533887148,0.3721376657,0.2042023391,0.0186498351,0.0301543158,-0.1795443147,-0.0855605751,-0.1612430811,-0.1782986075,-0.2557564378,0.1765659302,-0.2424640656,-0.0852350891,-0.1452090293,-0.0867879242,-0.4174173176,0.2448441684,-0.4018446505,-0.2548186779,-0.078374058,0.2720758915,-0.2720501721,0.1508205831,0.074082464,0.0892747641,0.0460765511,0.1715887934,-0.3309583068,-0.3067087829,-0.4702958167,-0.1624400914,0.5332528353,0.1399260014,0.0548889786,0.0383490734,-0.0015598963,-0.3136587739,-0.1001237929,0.3807926476,0.0008735982,-0.0985232592,-0.1401782483,0.2973882556,0.0664051101,-0.3065908253,0.0340856016,-0.3090626597,-0.0541217215,0.157516554,0.2250111103,-0.0126016084,0.1113775671,-0.0902564302,-0.3435205519,-0.1660826653,-0.1793526858,-0.0840207487,0.1074519679,-0.3885612786,-0.0375832282,0.167233333,-0.2739905119,-0.126218304,-0.1190674007,0.1150188521,0.1819080263,-0.0487881042,-0.2680825591,0.3679413497,-0.0389905758,0.069161512,0.02138846,-0.3876900673,0.2129806727,0.2526650131,-0.0327753536,-0.2018021047,0.1148760095,0.3068694174,-0.097650148,0.1494807303,-0.0055844239,0.0456261002,-0.2713431716,-0.3473313451,0.5005397201,0.0510833226,0.2984991968,0.139131248,0.7935022116,-0.0914077759,0.4170954823,0.3806064725,-0.050907854,-0.0616207384,-0.0068951314,-0.0638709962,0.2387561351,0.0170538314,-0.2286790311,0.5545761585,-0.1608418077,-0.1884116679,-0.304302305,-0.030096883,-0.5362465978,-0.4509307146,0.1638864875,0.3014622629,0.1265941411,-0.0399286821,0.2464965731,-0.1392953843,-0.354782939,0.1180578321,0.2957460284,-0.1534203142,-0.0618201718,-0.0451543964,0.080497697,-0.214286983,0.3920486867,0.1126289964,0.4006482065,-0.0302445814,-0.0606715269,0.1365611702,0.0421154536,0.6749553084,0.1139661744,0.2668806314,-0.0483190976,0.1517101973,-0.2012027949,-0.337048769,-0.2402469367,-0.1292212903,0.4147008955,-0.053943526,-0.4135914147,0.0053389715,-0.1468929499,0.0956119075,-0.1625174433,-0.0295244027,-0.1250679344,-0.186701864,-0.4802512527,0.1160195693,0.1534693092,0.2812899947,0.223260209,0.2467782348,0.1022344604,-0.0221610367,-0.0066826916,-0.0260741394,0.1148944348,0.0415336341,-0.390858978,-0.1406547427,0.3413005471,-0.1051336974,0.0629710406,0.5062643886,-0.2861008346,0.0114388028,-0.1281368285,0.2277383953,0.3643339574,0.0093669342,0.1643534601,0.0891771838,0.1918284595,-0.1667552739,0.3696986139,0.1991132498,-0.0413460173,-0.3877440989,-0.0128740259,0.4758426547,-0.007365664,-0.1285909861,0.1480601132,0.0824897438,-0.3054510951,0.4694943726,0.1594971567,1.4028254747,0.0743103847,0.1060868427,0.0582725108,-0.0234280452,0.7709746957,-0.8501492143,0.0102290399,-0.3112947345,0.1336631179,-0.1457691044,0.0992764309,0.2456333339,0.1884240806,-0.1797171235,0.4108475745,0.2099640369,0.5706493855,0.1618009061,0.2777124643,0.1584195346,-0.1224527359,-0.1224455535,-0.0375858955,-0.339430809,0.2648872137,-0.2823267579,-0.0010853326,0.0000824482,-0.4254595041,-0.4824260771,0.170297429,-0.174490571,0.1645847112,-0.3101142943,-0.1270029247,0.1680432856,0.0147916796,0.0111144166,0.3238190114,-0.1544265002,0.4299156368,-0.2235511392,-0.2313754857,-0.094600521,0.1680569947,-0.0222958624,-0.1653202623,-0.196807757,0.4138615727,-0.0622830614,-0.3473204076,-0.1739187241,0.0197294857,0.5572702885,0.0266521238,-0.326171577,-0.0611718521,-0.0020259989,-0.0870087221,0.0478479862,-0.1802935302,-0.3034142852,-0.1223326474,0.1891368926,-0.1419883221,0.0763695613,0.3687142432,-0.1562960148,-0.5203355551,0.5429047346,0.3616911471,-0.1847402453,-0.1230106652,-0.0202622302,-0.3258296251,-0.6497552395,-0.2177561969,-0.0770271569,0.0285630263,-0.164137423,0.1911832243,0.2515867651,-0.0383396819,0.0510081202,-0.3938291967,0.0517926514,-0.0127603505,-0.2740117908,-0.0835078061,0.1159851477,0.1871450543,-0.0092620943,-0.0447382145,-0.1161634922,-0.0225915089,-0.3062828183,0.1924285591,0.4826489091,0.2326569855,-0.2700323462,-0.0137210749,0.0105523011,0.1974306852,-0.0218309257,-0.0344476923,-0.090861015,0.2667791843,0.339341104,-0.0351116471,0.1336829662,-0.133158043,0.1975799352,-0.1563782394,-0.1528541893,-0.1540423036,-0.0454743579,0.2977791727,0.3360817134,0.2830898762,-0.5487601161,0.2178022265,0.1383381784,0.6532803774,0.0054280232,0.1927553713,0.155905813,0.0525586009,-0.2453491837,0.0929764733,0.2692934573,0.160554558,0.1048110574,0.062946558,-0.1325593293,0.0959392041,-0.1208665445,0.1463815719,-0.3814380467,-0.2813280523,0.632344842,0.0745457634,-0.1512728482,0.1893349439,0.6089372635,0.0491309986,-0.2840152383,0.1259601116,-0.1253487319,-0.2812301517,0.2148862779,0.1608722359,0.259193629,0.0149762696,0.1351060271,0.097361818,-0.0875413269,0.2604701221,-0.0326113589,-0.0643942133,0.2623575628,0.6413194537,-0.2414705455,0.1756813824,0.1518553793,0.1070120335,-0.1117627174,-0.2240461707,0.2313757986,0.5097441673,0.1159885824,0.066604197,-0.1439887434,0.0434135795,0.2388588637,-0.1132748425,-0.4433374703,0.2972133756,-0.1541454643,0.0485983193,-0.2984187305,-0.0929191932,-0.3210144043,0.2261300683,0.1480429322,-0.1404351145,-0.2817440927,0.3320850134,-0.3453849256,0.0739464089,-0.0042846031,0.0188350733,0.0351630747,-0.3815689683,0.0740954503,-0.0593812764,0.0998515412,0.0616395921,0.0515931062,0.4012039304,0.4144957066,-0.3987133801,-0.0537901819,0.0305591132,0.1961592734,0.0257299934,0.0914187282,0.1472989023,0.3724621236,0.1365384012,0.0411321819,-0.1063813865,-0.3492832482,0.260429889,-0.0233520158,-0.0546865761,-0.289498955,-0.0983108208,0.0189807415,-0.3697973192,-0.1028269455,-0.4696650505,0.2777038217,0.0003977382,0.1927059591,0.0835435539,-0.036285419,-0.0346115232,-0.0832945332,0.0816679969,0.384372443,0.1786179245,-0.2805787027,-0.5300997496,-0.6154990792,0.3956102729,-0.2236165106,-0.2988999486,-0.0640619919,-0.0280656237,0.1025355384,0.2608192563,0.055098027,0.0836734101,-0.2400378287,0.3680185974,-0.2253845334,-0.2801788747,0.1484095752,0.1170547009,-0.113599807,-0.3118135631,0.2139999866,-0.0381483883,-0.08435826,-0.0995994732,-0.2943398654,0.0373275988,-0.4005628526,0.4856735766,0.0347065851,0.3801331818,-0.0710382015,0.1127722785,-0.1277603954,0.0742859319,-0.1226383373,-0.0107810088,-0.1104756743,0.201518923,0.0309020914,0.21080257,-0.4701856375,0.3865426779,-0.082533434,0.1990238279,-0.3401548564,-0.0396733694,-0.1996798962,0.2241757512,-0.0465959832,-0.077884376,-0.051911056,0.2063358128,-0.4808903039,-0.1081218943,0.2249006629,-0.5813328028,-0.4714143276,-0.0415451713,0.3231417239,-0.3128690422,-0.0618979447,-0.2354566157,0.0353623666,0.1874027401,0.243143186,-0.1491308212,-0.1797588468,-0.2889468372,-0.1373033673,-0.0080988398,-0.1090167463,0.2330000252,-0.341154784,0.10055352,-0.3634814322]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"Thanks for your help.\r\n\r\nI tried to print the \"title\" of the document inside the` except (mwparserfromhell.parser.ParserError) as e`,the title displayed was : \"Campeonato Mundial de futsal de la AMF 2015\". (Wikipedia ES) Is it what you were looking for ?","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":41,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nThanks for your help.\r\n\r\nI tried to print the \"title\" of the document inside the` except (mwparserfromhell.parser.ParserError) as e`,the title displayed was : \"Campeonato Mundial de futsal de la AMF 2015\". (Wikipedia ES) Is it what you were looking for ?","embeddings":[-0.0982404053,-0.1460119784,0.0579146668,0.4570070207,0.0399687998,0.0833492652,-0.2478527278,0.1913093477,0.3329879642,0.2941777706,0.2543114126,0.1080052555,0.0952981412,-0.3402068019,-0.1696409434,-0.0679538921,0.3555515409,0.1214177683,0.1624230444,-0.2468806505,-0.0501843765,0.1787006557,-0.3427317142,0.382601887,-0.1634172499,0.1669341028,0.1231548637,-0.1326595694,-0.1403289288,-0.2993070185,-0.0200498365,-0.2091211826,0.1560906768,0.3652326763,-0.0001279058,-0.3193122149,0.4868043661,-0.0847794339,-0.1486693174,-0.2189909369,0.0152916275,-0.1522385776,-0.0410055257,-0.4044004083,0.3248384893,0.1153984368,0.3341578841,-0.2809922099,0.1205001771,0.2197394818,0.0640458316,0.1066202298,0.3839101195,0.0759176463,0.480653286,0.0734132007,0.2054694891,-0.060079664,-0.016876379,-0.0532576554,0.149125278,0.2014700323,-0.3659819067,-0.1884918511,-0.163210243,0.0662546456,0.2595542669,-0.4452034533,0.4108956754,0.1394709051,0.61737746,-0.1191039309,0.2614586651,0.1654366851,-0.0580870099,0.2580067515,0.3013699353,0.5481126308,-0.3672949374,-0.0253248662,0.3877830803,-0.1134894118,-0.1369290501,0.3051530719,-0.2432048023,0.9164056182,0.1252373457,0.3172872663,0.035912808,0.0480789393,-0.0662555993,-0.251314193,-0.0493606776,0.334076643,0.0214808527,-0.0258580539,-0.1511856019,-0.0784204379,0.001178946,-0.625526607,-0.4130973518,0.0299955718,0.1209483594,0.3402904868,0.454151839,-0.2407009155,0.2941499054,-0.0054475013,-0.0771103427,-0.1497201025,0.0046592373,-0.0258476324,-0.1439839154,0.0026517075,-0.4691490531,-0.0029376033,0.2461206168,-0.2048180699,0.0865612179,0.3560785353,-0.2949361503,-0.1116065904,-0.2578809857,0.4341236651,-0.0717999339,0.3773729801,0.0606103987,-0.0367316529,-0.1395096928,-0.5028152466,-0.0180455018,0.2353696376,-0.3238372505,0.0698220283,-0.2390310317,0.226461634,0.1722396165,0.1966182142,-0.0110010495,-0.5448348522,-0.0428097472,-0.1836660951,-0.071903713,0.0472855903,0.406760335,0.5007110238,0.2680571079,-0.40572837,-0.0525238104,0.3967243731,-0.3704168499,-0.0222618766,-0.5510870814,0.0479742251,0.0428075977,0.1816760898,-0.5636150837,0.0236339998,-0.0637113526,-0.1677027494,0.2126279175,-0.0518712103,-0.1455517709,-0.0310135093,0.1124990955,0.2195572853,-0.1851063818,0.0615859032,-0.1012386531,-0.1575168818,0.2370682061,-0.0327689201,-0.0864664093,0.636880517,-0.3505786955,0.5782259107,0.2652637959,-0.4091205001,-0.0882848203,0.0071630459,-0.0275748428,-0.222207427,0.0340306722,-0.1393164545,0.1705590934,0.0085355034,-0.4780333042,-0.0440130718,0.0305705108,0.18126522,-0.1381178051,-0.2846617699,0.2403111309,0.0836685821,0.2543239892,-0.0825863183,0.1528435946,0.4838692844,0.6022037268,-0.1535146981,0.145203799,0.0552840605,0.0634704307,0.2141290158,-0.0845406055,0.0701060891,-0.0034425391,-0.0194442309,0.0826343819,0.1979782879,-0.146543026,-0.1380741596,-0.1379958242,0.2703558207,-0.1237415373,0.0237798672,0.0232579205,0.2133961022,-0.198654905,0.3907471597,0.0831923932,-0.2709238529,-0.2775025666,0.0981848761,-0.0184138268,0.1946448684,-0.1272756755,-0.0370278805,-0.3745705187,-0.1918487698,0.4063358903,0.1215744838,-0.1520330608,0.023461353,0.2480846196,0.12266922,-0.2635534406,-0.2410593182,-0.1296415031,-0.4255687892,0.1146242395,0.2754119933,0.2598394454,-0.1263753623,0.1595700383,0.0801101103,0.2849260271,0.0874281824,-0.1943712533,0.063453652,0.1280266345,-0.0008983696,0.382299006,-0.1133902892,0.2536658645,0.4015442729,-0.1322078109,-0.3764730692,0.1788337529,-0.2094368637,0.6232566833,-0.0297614131,0.1618248075,0.1505903751,-0.2678955793,0.0095711956,0.3111509681,-0.5253850818,0.1216418967,-0.0011219424,-0.1208436713,0.3589930236,0.3209595084,0.0280894954,0.4389223456,0.2746983767,-0.0538246557,0.0729867071,-0.2088221163,-0.0472809002,-0.2550527751,-0.0936595425,-0.249587521,0.1308402568,-0.2461634278,-0.0539874807,-0.1664287746,0.0453462154,-0.3202031255,0.044147931,-0.4074264467,-0.1982375681,-0.1249895766,0.4425379336,-0.1859439313,0.1449581236,0.0070478665,0.040186733,-0.0461187288,0.3184624314,-0.2622790933,-0.3390574753,-0.4685198069,-0.161923483,0.4164783061,0.1619075239,0.1300805956,-0.0816562623,0.0647757575,-0.3107574582,-0.1337570995,0.4158418477,-0.1041708291,-0.0374839753,-0.1843907237,0.1631098241,-0.2027106285,-0.3040036857,0.0082991933,-0.335069567,-0.0869709253,0.0554570556,0.2022849619,-0.1650488228,0.1563441306,-0.0912574306,-0.3194941878,-0.0305149443,-0.2652244866,-0.1757873148,0.1309884936,-0.5440905094,-0.1037883535,0.2221551985,-0.1957859546,-0.1582685262,-0.2416283339,0.0511857495,0.1386105865,-0.0271180645,-0.1973048896,0.4221188426,0.0898568109,0.1270066351,-0.1342784762,-0.531072557,0.3187642992,0.2872523665,-0.2071967572,-0.2658961117,0.0953802317,0.3380016387,-0.1261989325,0.1300807148,-0.0205747448,-0.0200337227,-0.3060702384,-0.2612015307,0.3659541011,-0.0696886629,0.3568731248,0.0052714818,0.5611320734,0.0077976589,0.3498138785,0.299805969,-0.0559647977,-0.1249294654,0.0478359722,-0.1055808663,0.2657906711,0.12482737,-0.2373918593,0.7256162763,-0.2425276935,-0.2778725922,-0.2114525288,-0.1085004732,-0.4927818477,-0.3511238098,0.1340619177,0.4256851971,0.2151128948,-0.1582023799,0.2645699978,-0.0461710878,-0.3970320225,0.0409245156,0.2700924575,-0.1457710564,-0.0752224177,-0.2231326699,-0.0068608266,-0.2216742188,0.2792330384,0.1658106595,0.2630618811,-0.0474534817,0.037865676,0.0920640752,0.1062745452,0.7170780897,0.0799824521,0.2855763435,0.0076335068,0.1088951901,-0.0855669081,-0.3343628645,-0.3213562667,0.0347119831,0.4114460945,-0.0811108723,-0.3828402758,0.011714993,-0.1130540669,-0.0276525952,-0.0771019906,-0.1594764143,-0.0631828383,-0.0750658736,-0.4859534204,0.2936125398,0.1052246988,0.1847798079,0.2181551456,0.1167586371,0.0645170733,0.1183754057,0.0404809341,-0.08190272,0.0652398765,0.1532297283,-0.3297492266,-0.1077964455,0.3382184803,0.1190087721,0.0855210572,0.466806829,-0.2860803008,0.0467538387,-0.2273383737,0.3570978343,0.3676089048,0.1378226876,0.1718908399,0.2850579917,0.1846428514,-0.2005975842,0.2463531196,0.2104185969,-0.0101194987,-0.4572693408,0.04766009,0.4190883934,0.0432882681,-0.1291730106,0.0807551146,0.1301420927,-0.3920389414,0.227646023,0.30420205,1.3743013144,0.2290473133,0.0000551489,0.0182178933,-0.0212042667,0.7775986791,-0.5432036519,-0.0416601263,-0.4133483171,0.2235513926,-0.2628512383,0.0718859434,0.1743706614,0.1549933255,-0.4109581411,0.3315269053,0.1195450276,0.6789625287,0.2007481307,0.2311806977,0.3003860712,-0.0696269125,-0.0471351072,-0.0562486798,-0.2997576594,0.1641265005,-0.3360463679,0.0040933033,0.0726586953,-0.4021903276,-0.5105246305,0.1497732699,-0.1560824662,0.1355955005,-0.2400936037,-0.0924371108,0.3394129574,0.0303811152,0.1530252248,0.3946716189,-0.2360749245,0.3084458411,-0.0790804327,-0.2827481031,-0.0845997408,0.0759869814,-0.1662607342,-0.0762496293,-0.2573314309,0.4024147093,0.0404142216,-0.2437894046,-0.1297530979,-0.0173817184,0.5050361753,-0.0179757234,-0.4633437097,-0.0243397225,0.0135538988,-0.0968402177,0.0099354535,-0.2525736094,-0.1889143586,-0.0800801665,0.3576352596,-0.2762170136,0.023976488,0.3802042007,-0.0508619435,-0.3608962893,0.4053850472,0.4141704738,-0.2110740691,-0.1729269922,-0.138873592,-0.4548933804,-0.6205152869,-0.2452708483,0.0036880078,0.0751879141,-0.09934517,0.2278403491,0.2664340436,-0.0028219644,0.0673345402,-0.3757095933,0.1301342398,0.0004426128,-0.29973647,-0.0541534983,0.150513351,0.144070521,-0.0772834644,0.0108918129,-0.1024203151,0.0501262024,-0.3457911909,0.1844022125,0.4043624401,0.2660724223,-0.3334859014,-0.1559803337,-0.0082400776,0.3081268668,-0.1110079214,-0.0437902063,-0.1429688632,0.2774117291,0.4025516212,-0.1154686213,0.1143009514,-0.056904193,0.1281068921,-0.1438817084,-0.1454659551,-0.1152153611,-0.0507896245,0.3200855255,0.2864226699,0.3363921642,-0.6627502441,0.1176775992,0.221680969,0.5979235768,-0.0016472841,0.2584121227,0.1601392031,0.0629687607,-0.0835658833,0.2112224549,0.3295860887,0.1443807483,0.0623542294,0.0162491128,-0.0463710465,0.1883053482,-0.1092370376,0.1076652333,-0.3437703252,-0.3186070621,0.5859051943,0.0780176371,-0.1367445588,0.0791248009,0.5128093362,0.0503338836,-0.2816810906,0.135364145,-0.0192648862,-0.3054715991,0.1443625689,0.1600497961,0.3883184493,-0.008088625,0.0976823345,0.0427978858,-0.2145092636,0.3497294188,0.0331430286,0.1103882715,0.2539881468,0.4191575944,-0.0092265448,0.0306610651,0.1890913397,0.2195819765,-0.1321515739,-0.1168117002,0.1506026089,0.4852175713,0.1417030394,-0.036451567,-0.1465826482,-0.2729983032,0.2765254378,-0.1070106775,-0.4260050654,0.2682324946,-0.20227772,0.0506315678,-0.2739011049,-0.1167620271,-0.1663043201,0.2934019864,0.1380989999,-0.2000113279,-0.343500942,0.2851376235,-0.3189735115,-0.0203512665,-0.0766399726,-0.0063008592,-0.0449219048,-0.4550501406,0.1171482205,-0.1794952601,0.0901087746,-0.1199383661,0.2348064035,0.3847108781,0.4226667881,-0.4357827902,0.0263300557,-0.0464054793,0.2207933068,0.1211030558,0.0787169039,0.2410830259,0.243532896,0.0952396467,0.0350228846,-0.0725517571,-0.3141316175,0.3812608421,-0.0295785088,0.0584462956,-0.308160305,0.0079684891,0.1974339783,-0.228430599,-0.1496478617,-0.5025536418,0.2157158554,0.1553734839,0.1363462359,0.0866229311,0.0585675016,-0.0431728177,-0.0426924229,0.1241076887,0.3479136229,0.2889395952,-0.2943775356,-0.4787543118,-0.5232974291,0.4851525724,-0.145958662,-0.1678867489,0.0266095232,-0.0969084874,0.1434943676,0.3562704027,0.0353256427,0.0365078039,-0.3478903472,0.3106482625,-0.1491153389,-0.1809680611,0.153692022,0.0774843916,-0.1319775581,-0.2365524024,0.1763026416,-0.217424944,-0.0963839963,-0.2786431909,-0.129952088,-0.127636984,-0.1647598743,0.5894804001,0.020112792,0.487385571,-0.1468153894,0.1515864134,0.1199431717,0.1168107465,-0.0840778798,0.1633863151,-0.0240935534,0.1927311867,-0.0205262545,0.1695741415,-0.3605150878,0.2414454669,-0.0204201601,0.2153668851,-0.4155307412,-0.0928949639,-0.1769775897,0.0274348129,-0.0958775952,-0.0296799801,0.0120360367,0.2343749404,-0.3720820248,0.000432994,0.278965503,-0.5013762116,-0.3355753124,-0.1907162666,0.2618013322,-0.1488953084,-0.1762331575,-0.2184724212,0.0482211672,0.2631866038,0.1811581403,-0.2186214328,-0.1442239285,-0.2311945558,-0.1701048315,-0.0474909991,0.0033463228,0.1215734482,-0.4107504189,0.062971577,-0.4289715588]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"Thanks a lot @Shiro-LK !\r\n\r\nI was able to reproduce the issue. It comes from [this table on wikipedia](https:\/\/es.wikipedia.org\/wiki\/Campeonato_Mundial_de_futsal_de_la_AMF_2015#Clasificados) that can't be parsed.\r\n\r\nThe file in which the problem occurs comes from the wikipedia dumps, and it can be downloaded [here](https:\/\/dumps.wikimedia.org\/eswiki\/20200501\/eswiki-20200501-pages-articles-multistream6.xml-p6424816p7924815.bz2)\r\n\r\nParsing the file this way raises the parsing issue:\r\n\r\n```python\r\nimport mwparserfromhell as parser\r\nfrom tqdm.auto import tqdm\r\nimport bz2\r\nimport six\r\nimport logging\r\nimport codecs\r\nimport xml.etree.cElementTree as etree\r\n\r\nfilepath = \"path\/to\/eswiki-20200501-pages-articles-multistream6.xml-p6424816p7924815.bz2\"\r\n\r\ndef _extract_content(filepath):\r\n \"\"\"Extracts article content from a single WikiMedia XML file.\"\"\"\r\n logging.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, \"rb\") as f:\r\n f = bz2.BZ2File(filename=f)\r\n if six.PY3:\r\n # Workaround due to:\r\n # https:\/\/github.com\/tensorflow\/tensorflow\/issues\/33563\r\n utf_f = codecs.getreader(\"utf-8\")(f)\r\n else:\r\n utf_f = f\r\n # To clear root, to free-up more memory than just `elem.clear()`.\r\n context = etree.iterparse(utf_f, events=(\"end\",))\r\n context = iter(context)\r\n unused_event, root = next(context)\r\n for unused_event, elem in tqdm(context, total=949087):\r\n if not elem.tag.endswith(\"page\"):\r\n continue\r\n namespace = elem.tag[:-4]\r\n title = elem.find(\".\/{0}title\".format(namespace)).text\r\n ns = elem.find(\".\/{0}ns\".format(namespace)).text\r\n id_ = elem.find(\".\/{0}id\".format(namespace)).text\r\n # Filter pages that are not in the \"main\" namespace.\r\n if ns != \"0\":\r\n root.clear()\r\n continue\r\n raw_content = elem.find(\".\/{0}revision\/{0}text\".format(namespace)).text\r\n root.clear()\r\n\r\n if \"Campeonato Mundial de futsal de la AMF 2015\" in title:\r\n yield (id_, title, raw_content)\r\n\r\nfor id_, title, raw_content in _extract_content(filepath):\r\n wikicode = parser.parse(raw_content)\r\n```\r\n\r\nThe copied the raw content that can't be parsed [here](https:\/\/pastebin.com\/raw\/ZbmevLyH).\r\n\r\nThe minimal code to reproduce is:\r\n```python\r\nimport mwparserfromhell as parser\r\nimport requests\r\n\r\nraw_content = requests.get(\"https:\/\/pastebin.com\/raw\/ZbmevLyH\").content.decode(\"utf-8\")\r\nwikicode = parser.parse(raw_content)\r\n\r\n```\r\n\r\nI will create an issue on mwparserfromhell's repo to see if we can fix that\r\n","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":248,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nThanks a lot @Shiro-LK !\r\n\r\nI was able to reproduce the issue. It comes from [this table on wikipedia](https:\/\/es.wikipedia.org\/wiki\/Campeonato_Mundial_de_futsal_de_la_AMF_2015#Clasificados) that can't be parsed.\r\n\r\nThe file in which the problem occurs comes from the wikipedia dumps, and it can be downloaded [here](https:\/\/dumps.wikimedia.org\/eswiki\/20200501\/eswiki-20200501-pages-articles-multistream6.xml-p6424816p7924815.bz2)\r\n\r\nParsing the file this way raises the parsing issue:\r\n\r\n```python\r\nimport mwparserfromhell as parser\r\nfrom tqdm.auto import tqdm\r\nimport bz2\r\nimport six\r\nimport logging\r\nimport codecs\r\nimport xml.etree.cElementTree as etree\r\n\r\nfilepath = \"path\/to\/eswiki-20200501-pages-articles-multistream6.xml-p6424816p7924815.bz2\"\r\n\r\ndef _extract_content(filepath):\r\n \"\"\"Extracts article content from a single WikiMedia XML file.\"\"\"\r\n logging.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, \"rb\") as f:\r\n f = bz2.BZ2File(filename=f)\r\n if six.PY3:\r\n # Workaround due to:\r\n # https:\/\/github.com\/tensorflow\/tensorflow\/issues\/33563\r\n utf_f = codecs.getreader(\"utf-8\")(f)\r\n else:\r\n utf_f = f\r\n # To clear root, to free-up more memory than just `elem.clear()`.\r\n context = etree.iterparse(utf_f, events=(\"end\",))\r\n context = iter(context)\r\n unused_event, root = next(context)\r\n for unused_event, elem in tqdm(context, total=949087):\r\n if not elem.tag.endswith(\"page\"):\r\n continue\r\n namespace = elem.tag[:-4]\r\n title = elem.find(\".\/{0}title\".format(namespace)).text\r\n ns = elem.find(\".\/{0}ns\".format(namespace)).text\r\n id_ = elem.find(\".\/{0}id\".format(namespace)).text\r\n # Filter pages that are not in the \"main\" namespace.\r\n if ns != \"0\":\r\n root.clear()\r\n continue\r\n raw_content = elem.find(\".\/{0}revision\/{0}text\".format(namespace)).text\r\n root.clear()\r\n\r\n if \"Campeonato Mundial de futsal de la AMF 2015\" in title:\r\n yield (id_, title, raw_content)\r\n\r\nfor id_, title, raw_content in _extract_content(filepath):\r\n wikicode = parser.parse(raw_content)\r\n```\r\n\r\nThe copied the raw content that can't be parsed [here](https:\/\/pastebin.com\/raw\/ZbmevLyH).\r\n\r\nThe minimal code to reproduce is:\r\n```python\r\nimport mwparserfromhell as parser\r\nimport requests\r\n\r\nraw_content = requests.get(\"https:\/\/pastebin.com\/raw\/ZbmevLyH\").content.decode(\"utf-8\")\r\nwikicode = parser.parse(raw_content)\r\n\r\n```\r\n\r\nI will create an issue on mwparserfromhell's repo to see if we can fix that\r\n","embeddings":[-0.1087391153,-0.2687698901,0.0588545799,0.3744995594,0.0680316836,0.0585111044,-0.0978485569,0.407232821,0.2762867808,0.2615219653,0.1648638099,0.1633142382,0.0697004646,-0.4300828576,-0.1707364321,-0.042684406,0.2212866992,0.118236579,0.059008494,-0.1014077961,-0.0859467611,0.1587132514,-0.3476288617,0.2824134529,-0.32690382,0.1148107052,0.1520309299,-0.065615125,-0.0889430493,-0.2003451288,-0.0028276851,-0.2012956291,0.2379267067,0.4786588848,-0.0001259802,-0.0412694849,0.3049947321,-0.093657434,0.0099669043,-0.2359312773,0.0336602069,-0.1631217003,-0.1159503236,-0.3689011633,0.2481287122,0.2071848214,0.1943381876,-0.1866433322,0.4095414579,0.2438102067,0.0544666685,0.163917467,0.2602112889,0.0912849233,0.514739275,-0.053879641,-0.0065108077,0.0079004159,0.0930049717,-0.1025947332,0.1022947431,0.2438964993,-0.420661658,-0.0717122629,-0.1377499551,0.1038781703,0.4466082454,-0.5326789618,0.4791566432,0.3168759346,0.3135971725,-0.022695519,0.1983242035,0.103534475,-0.1376882643,0.1046121716,0.3115144372,0.5089524984,-0.48955971,0.0953266397,0.4412794709,-0.0603463277,-0.2098179162,0.2732831836,-0.3379980326,0.8441174626,0.0538513251,0.1933108121,-0.0251846313,-0.0369251259,0.0193275176,-0.220910877,0.1413909644,0.378723979,-0.0517863333,-0.0091330903,-0.0343199112,-0.3973082006,-0.021723425,-0.6584088206,-0.3500780165,0.1816475093,0.0500032641,0.2907331586,0.3454244137,-0.337862879,0.1525789052,0.0352604389,0.0321704298,0.0292413831,-0.1271683574,0.0367852971,-0.2031404376,-0.0674105212,-0.5065873861,0.0622966178,0.2397127151,-0.0703944489,-0.0532746203,0.3891048431,-0.2007398307,0.0207773875,-0.1578020155,0.5316457748,-0.0691650882,0.4129965007,0.1543025076,0.0851949006,-0.1468566507,-0.5985732079,0.052663479,0.1228784323,-0.039723698,0.0134536037,-0.1145052537,0.1165548936,0.2161987126,0.2504267395,-0.0326060876,-0.5398552418,-0.0503432825,-0.3519892693,0.0567449853,0.1298061758,0.4478539824,0.34950158,0.215083316,-0.4113384783,-0.0219976194,0.4195792675,-0.4311141372,-0.1687888354,-0.4756811261,0.0521946289,0.0355714373,0.1392172575,-0.5501323938,0.1158676669,0.0992074832,-0.1055626422,-0.0358388014,-0.1966685802,-0.2580610216,-0.0632867962,0.1914139837,0.1181722507,-0.1422764808,0.126726836,-0.0599625632,0.0571596734,0.4125909805,0.1793214381,-0.163065955,0.7019707561,-0.2670831084,0.7228682041,0.1745021939,-0.2736831605,-0.0719350725,0.0337947607,0.0428299047,0.007124987,-0.1001690924,-0.2217167169,0.3694820106,-0.0816345289,-0.1842643321,0.1109004542,-0.003875616,0.2598793209,-0.2199014872,-0.2137375325,0.338465929,0.0155823752,0.2996574342,-0.2194487154,0.0218850672,0.4970418215,0.5048382282,-0.2736796141,0.1297584027,0.0826887786,0.1501490623,0.1439250559,-0.1093944609,-0.1013655439,-0.0906945467,0.0045490987,0.0658480749,0.3201473057,-0.2615011036,-0.158896327,-0.1252180934,0.2352294326,-0.2003859729,0.0534398891,0.0882508159,-0.0568753891,-0.292809099,0.3478443325,-0.0486512706,-0.4844489992,-0.3517448008,0.1733437777,0.1646933109,0.2952848077,-0.2190816253,-0.079315573,-0.3375533223,-0.0990324244,0.4453221858,-0.0610528551,-0.0529888757,-0.056188643,0.283511728,0.1120934263,-0.3946905434,-0.142821297,0.0544433855,-0.2703911364,0.0110277655,0.2399187833,0.2501266003,-0.1488434821,0.2306790799,0.2286777347,0.3325100541,0.0825077891,-0.1761720181,0.1144729629,0.2624149621,0.0093547711,0.3972715735,-0.1282763928,0.1555531919,0.4382648468,-0.5117701888,-0.3361923993,0.0879169554,-0.0545391068,0.5132347941,0.05086549,0.2247790992,0.0900133327,-0.1870054007,-0.0401293896,0.2298818529,-0.3210944533,0.1083917394,-0.1220101118,0.005767398,0.3593573868,0.1124442369,0.0358499289,0.5103645921,0.2113904655,0.0721914172,0.166925624,-0.1320455372,0.098707594,-0.1885120571,-0.035234198,-0.0812533349,0.2075201869,-0.3698166311,0.0277478304,-0.3147502542,0.0786038786,-0.1918708235,-0.0323620066,-0.3220308721,-0.2089153081,-0.0629463568,0.4096322358,-0.2718832791,0.0419884175,-0.0064639235,-0.1180246696,0.0215868615,0.2971606851,-0.2370031625,-0.3015561998,-0.3740295768,-0.1583939344,0.5779916644,0.0535107106,-0.0512510836,-0.0516674444,-0.2161799371,-0.35149014,-0.397675842,0.4286088943,-0.1134408191,0.0060054329,-0.1002593935,0.3180465102,-0.2111659199,-0.4990894496,0.1148440614,-0.2589308321,-0.0832984224,0.1671217531,0.0792736709,-0.2307243794,0.0519303642,-0.1746178865,-0.1400774121,-0.0965811983,-0.2825163305,-0.1297852546,0.1823790222,-0.3487784863,-0.0708561689,0.1796253771,-0.334259212,-0.1362028867,-0.2476677299,0.01838989,0.2872105837,-0.1329412907,-0.3237062097,0.4956243038,0.0269243158,0.3418346941,0.0566490144,-0.403057158,0.2267920822,0.2314048707,0.0250053797,-0.1748469323,0.0461068973,0.2127269655,-0.0847881883,0.1835100949,0.050592225,0.097263217,-0.3360319138,-0.1924429536,0.2831331491,-0.0903017372,0.5138000846,0.259572655,0.6778399348,-0.1312732249,0.2277941257,0.197369799,-0.1405770481,-0.1899258941,0.0949459448,-0.0761387199,0.1143446267,-0.1003623158,-0.1714862138,0.5857168436,-0.2173735648,-0.2346692085,-0.2378280908,-0.0568009578,-0.5688069463,-0.5239387155,0.1828358024,0.3403631449,0.3761722445,0.0175006408,0.2138800025,-0.0920705348,-0.3685364425,-0.0155466069,0.2618908584,-0.1244408339,-0.1231168956,-0.2195123732,-0.0545557737,-0.2247724384,0.2450663894,0.2223412693,0.5213936567,0.1216327474,0.0373340845,0.1914764494,0.0804083049,0.4079046249,0.1229577065,0.2616471648,-0.0548344851,-0.1593905687,-0.3506841063,-0.2592153549,-0.3739874661,-0.009043863,0.5089577436,-0.0336519592,-0.3929763734,-0.0688543916,-0.0405448191,-0.0219740234,-0.1007982045,0.0373904146,-0.0903743058,-0.3582050502,-0.3550568521,0.2206852883,0.3207581937,0.2599360347,0.2318659276,0.1107542813,0.0202271622,0.0940772519,0.1679939628,0.1295650005,0.0218961928,0.1068542674,-0.1335513741,-0.3471139073,0.0463485681,0.0771905109,0.1125213951,0.5233774781,-0.465544343,-0.0342264622,-0.1587858051,0.2762382925,0.3710380197,-0.009256782,0.2874032855,0.2022298872,0.2021597773,-0.0559518561,0.2551429868,0.3362346292,0.0627210811,-0.2458248436,0.1051725894,0.3895787001,-0.0080274297,-0.0241412483,-0.0015796397,0.0767262354,-0.2691985369,0.1893244535,0.3515012264,1.2506076097,0.2659700215,0.0169888642,0.2763577998,0.0442126878,0.6086792946,-0.6462409496,0.1094848886,-0.2827792764,0.0614246763,-0.20590505,0.1329853535,0.0934012458,0.1419589818,-0.2886447608,0.2039875388,0.1032896489,0.5259107947,0.2789724171,0.1803505421,0.2175611854,-0.0830236301,-0.1529653072,-0.0414283574,-0.2636484504,0.0002911442,-0.2863665223,-0.0215723123,0.0985354334,-0.3779264092,-0.4726229906,0.1767259538,-0.3571780622,0.2739375532,-0.0383762382,0.0106597589,0.3592602909,0.0099478411,-0.0358081535,0.3343281448,-0.1344681978,0.2318214476,-0.1697712392,-0.3552456498,0.0129977558,0.0338371769,-0.0132486178,-0.2199783921,-0.2343140393,0.4540521801,0.016178539,-0.2345126569,-0.1109379083,0.0964357406,0.2640200555,0.113412559,-0.4126053452,0.0684174001,0.0020151734,-0.0428782776,0.0237391051,-0.22108154,-0.3089178801,0.0455833226,0.1914596111,-0.298969835,0.1511256993,0.3534888625,-0.0079745678,-0.3089992702,0.5539166927,0.421487391,-0.2652538121,-0.1200853139,-0.146939531,-0.5995234251,-0.5139879584,-0.1296265572,-0.1517818272,0.0753184184,-0.0027998914,0.2987012267,0.3360542655,-0.1601730585,-0.0145235173,-0.2696587145,0.1962418854,-0.0144208176,-0.3134302199,-0.1097420231,0.0829666257,0.381965369,-0.0990817472,0.0859425217,-0.1187355742,0.2275440991,-0.216312021,0.2235797942,0.3537528813,0.0880687162,-0.3263801038,-0.0284316,-0.0219769888,0.3297540545,-0.0077916938,-0.0366446488,-0.2328091562,0.2400611937,0.3293090165,-0.07574974,0.1362517178,-0.0270531066,0.0028649538,-0.1123141423,-0.0395219103,-0.0460845418,-0.0571955852,0.3660621345,0.2320525646,0.3750469387,-0.5651993752,0.1087600216,0.1337923855,0.5745230913,0.0278675202,0.2287802696,0.2306108475,0.006489221,-0.1736459583,0.141430214,0.09748739,-0.0293035135,0.3401764929,0.0067823078,0.0067580664,0.1846049279,-0.2482762784,0.2205840051,-0.2990249991,-0.274192214,0.5228120089,0.0742757916,-0.2124698758,0.0874912813,0.4620794952,0.1062630638,-0.256289959,0.2262158394,0.0315433927,-0.2609515786,0.1855696291,0.1876291037,0.2801580727,-0.1200942397,0.2026831508,0.0164931901,-0.2598483264,0.3014808595,-0.1215875819,-0.0723850876,0.2637168169,0.3368169367,-0.041625794,0.0521306247,0.0975426733,0.0931923687,0.0775518641,-0.1573628038,0.0244251173,0.4214036763,-0.0256877188,-0.0009934366,-0.1551650763,-0.2060313821,0.2579576969,-0.0723037273,-0.44498083,0.3604534268,-0.2784141004,-0.0036356158,-0.2545226812,-0.1036021486,-0.2335770428,0.1375909299,0.163842082,-0.0178429373,-0.367467016,0.3523712754,-0.3744676113,0.1455938816,-0.0529463664,0.0031990996,-0.0197853893,-0.3990138471,0.1668805629,-0.0243334249,-0.0397890024,-0.1453586817,0.1307671666,0.4951551259,0.5052855611,-0.4910767972,0.0306373667,-0.0313672945,0.2223320901,0.0341138616,0.111014761,0.2423649877,0.4140434265,0.1523620933,0.0242620092,-0.1170482114,-0.3675275147,0.3683361411,0.1741411388,0.1025608778,-0.2432252467,-0.0427914448,0.1485030651,-0.1346629858,-0.1821497977,-0.3763532639,0.1467437446,0.2502170801,0.034485098,0.1311168224,0.0236348882,-0.0290091541,-0.0190008301,0.1255922467,0.3132602572,0.0382290371,-0.1958370656,-0.3837857544,-0.6240746975,0.4608494937,-0.2040307224,-0.3661982119,0.1476474404,0.2049539983,0.1240615845,0.3853283525,0.0820921361,-0.2400576621,-0.3210077286,0.2978592813,-0.1986877918,-0.0557459854,0.037449345,-0.0787730142,-0.068556942,-0.3496166766,0.158684358,-0.1043168455,-0.1114211157,-0.2411878109,-0.0280074626,-0.1553024948,-0.1515020579,0.3971240222,0.1063482612,0.3947918713,-0.1707577407,0.0838163719,-0.0741473511,0.1212983951,-0.1070879251,0.3608297408,0.0544192493,0.2073380947,0.0287489016,0.0233600102,-0.3573513031,0.1884986907,-0.0116838291,0.0379815362,-0.4710639417,0.0426907577,-0.2221050411,0.1887381375,0.017725246,0.0637415051,-0.0354145542,0.2457731962,-0.4298841059,-0.2522987723,0.1784219593,-0.5421526432,-0.3952219486,-0.3052185178,0.2859874964,-0.2357291877,0.0401068702,-0.2903586924,-0.1132241413,0.2577067316,0.0202623196,-0.2819970548,-0.1080349907,-0.1423834264,-0.2690139711,-0.1281462908,-0.0844760165,0.0411333442,-0.4689804912,-0.0842289925,-0.4250320792]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/321","title":"ERROR:root:mwparserfromhell","comments":"This going to be fixed in the next `mwparserfromhell` release :)","body":"Hi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n","comment_length":11,"text":"ERROR:root:mwparserfromhell\nHi,\r\n\r\nI am trying to download some wikipedia data but I got this error for spanish \"es\" (but there are maybe some others languages which have the same error I haven't tried all of them ).\r\n\r\n`ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token stack.`\r\n\r\nThe code I have use was : \r\n`dataset = load_dataset('wikipedia', '20200501.es', beam_runner='DirectRunner')`\r\n\r\n\nThis going to be fixed in the next `mwparserfromhell` release :)","embeddings":[-0.3029485643,-0.0829322264,0.0720134526,0.3769366145,0.0775637552,-0.0515599251,-0.152377516,0.2200460881,0.3627667725,0.2901905179,0.2781309187,0.0237164851,0.0404486656,-0.3342026472,-0.0660937726,0.0058835726,0.3668923676,0.1200451404,-0.0136656733,-0.2360523045,-0.0702550039,0.192976892,-0.3456954956,0.2923926711,-0.2800429165,0.0965652838,0.1483434141,-0.0293208938,-0.0942242369,-0.3477033377,0.0423753858,-0.2659012973,0.185654819,0.26504758,-0.0001254651,-0.1969166845,0.5230875611,-0.0988425314,-0.1292342693,-0.2017611265,-0.0502723046,-0.1946739107,-0.0505097881,-0.2473738194,0.2870600224,0.1359033734,0.2397641093,-0.2013242245,0.2441984415,0.2441843301,0.0946675986,0.2202967256,0.3456865549,0.0035016567,0.5350577235,0.0791022107,0.1970073134,-0.0495881848,-0.0026370171,-0.0815572292,0.1120341718,0.1665259153,-0.3865649402,-0.1595091671,-0.172017619,-0.0720978752,0.4146973491,-0.4126437008,0.3448693156,0.1789664924,0.5039269924,-0.0299378633,0.2935406566,0.1349442899,-0.0758551955,0.2268213332,0.4144261777,0.5591591001,-0.4472072124,0.0379805006,0.3564355075,-0.1130773351,-0.0021898381,0.2961555719,-0.1589553952,0.9120818377,0.0600433648,0.2825836241,0.0779816285,-0.0288758446,-0.0686025918,-0.3184456527,-0.0528797917,0.3071828187,-0.0788349435,-0.0126742795,-0.1766927242,-0.0274828319,0.0339195989,-0.5603187084,-0.5222287774,0.1071819291,0.053637743,0.2307014018,0.4832917154,-0.1516804099,0.161864996,-0.1065703854,0.030137023,-0.1179611459,-0.0221589897,0.0787290707,-0.1583978832,-0.055242002,-0.5625892878,0.0822617784,0.1529345959,-0.1296649873,0.0042249998,0.2084014863,-0.1665793061,-0.1831602901,-0.1925233305,0.4708549082,-0.0841871351,0.4451791048,-0.0163800828,0.0105431778,-0.1667971462,-0.5921105146,-0.0089016696,0.2119275033,-0.3405156434,0.0458269827,-0.1364572495,0.278416723,0.1301812977,0.280164808,-0.0438591689,-0.4739551246,-0.0590638183,-0.1921865195,-0.1043622941,0.0592605211,0.4727985561,0.4333671927,0.2275043279,-0.3252361119,-0.0617587864,0.4051963389,-0.4818452597,-0.1406025738,-0.5169247985,0.0508448817,0.0287560336,0.1664904803,-0.5577250123,0.0674454719,-0.0691415742,-0.1170719638,0.0528036356,-0.1763977706,-0.2052556872,-0.0891865715,0.0911266208,0.2792774141,-0.1259887666,0.0256753936,-0.0711159557,-0.1546889246,0.3039938211,0.084487319,-0.197310254,0.6330301762,-0.409743309,0.5306949615,0.3416782916,-0.259596169,-0.1143590286,-0.030259937,-0.0238158777,-0.2006143183,-0.0618273579,-0.200640291,0.2062762976,-0.0734071806,-0.3731717169,0.0886715055,0.080293797,0.2536610663,-0.2337891608,-0.411786139,0.2756696045,0.1127389744,0.1872544289,-0.0878776237,0.1487205029,0.4963518083,0.4902167618,-0.2167666703,0.1371979117,0.0370633975,0.1184269935,0.1509742141,-0.1012562215,0.0846989229,-0.0323145613,-0.0271223597,-0.0147716627,0.2664056718,-0.091021426,-0.0476138704,-0.1193670928,0.2096170038,-0.1423258781,0.0754076391,0.071723707,0.108855091,-0.1842883676,0.3457753956,0.0109437108,-0.2474152744,-0.2133909762,0.1129825562,0.1227825433,0.3172683418,-0.186101079,-0.0801841766,-0.3448589444,-0.1339995712,0.4830684662,-0.0002800195,-0.1444180906,0.0057741669,0.2872534692,0.0984590203,-0.3183758259,-0.1996192783,0.0300827976,-0.3453758359,0.1960908771,0.1838380396,0.2334690988,-0.1480435282,0.2587204278,0.0737140253,0.354667902,0.074269861,-0.1892767698,0.1319057345,0.2060496211,0.0357525833,0.44507128,-0.0852372274,0.2538257241,0.3840779364,-0.1878136992,-0.3168341815,0.2483221591,-0.1261252016,0.5331593752,-0.0553136878,0.2217689306,0.1856150329,-0.3393711746,-0.0646248758,0.3051406443,-0.4347032905,0.1501065791,-0.0349988565,-0.1036992148,0.319970727,0.2666831315,0.0264667552,0.5130885243,0.1019557714,0.0497149751,0.0872180015,-0.2187924683,-0.0500479601,-0.2669815421,-0.0673200563,-0.2563014925,0.1530962735,-0.2405896336,-0.1888009012,-0.2642006874,0.0397965908,-0.3129536808,0.0517788418,-0.3985456228,-0.11744003,-0.1706908047,0.3249405324,-0.1661039144,0.1449262202,0.0384359807,0.0847453997,-0.0761025175,0.2998120785,-0.19831644,-0.2898937464,-0.5092601776,-0.1374942809,0.549305439,0.1379248202,0.0075337947,-0.023877643,0.0403389744,-0.2714996934,-0.1966896653,0.4413952827,0.0145406509,-0.0995788053,-0.0505698062,0.1685789973,-0.0711347759,-0.3214116693,0.0223346408,-0.3345566988,-0.0945170596,0.0441667922,0.0414752215,-0.1770815402,0.0776892528,-0.0345472582,-0.3931075633,-0.0415507518,-0.306863606,-0.0890084803,0.1178176552,-0.5169578791,-0.0154268304,0.1738921702,-0.3952563405,-0.2057216167,-0.2827661037,0.0817331523,0.1339233965,-0.0144498562,-0.2312668115,0.4589441121,0.004200228,0.1678541601,-0.0953280851,-0.5733838081,0.2598474622,0.1584618986,-0.0741341487,-0.2329422534,0.1027510539,0.3399954736,-0.1276193261,0.1565315425,0.0434668958,-0.111754261,-0.2882436514,-0.2245563418,0.3738436401,-0.1177337542,0.3607291579,0.0242940355,0.6156001091,-0.0516428575,0.2615699768,0.3386620581,-0.0169664901,-0.0792291909,0.0198262818,-0.0468293093,0.3169243038,0.0696308538,-0.1983256936,0.6149777174,-0.2100994885,-0.3360970616,-0.2487807572,-0.1196329817,-0.4018968344,-0.3945290446,0.1581314653,0.3618605435,0.1838920265,-0.005302717,0.225314036,-0.0879833251,-0.348754853,0.0274726469,0.1851917654,-0.144183591,-0.0545389354,-0.2071504146,0.0098804263,-0.2948994935,0.2723393142,0.211619094,0.3745718598,-0.0335928425,0.0371324345,0.0801110789,0.0612480827,0.6396936774,0.0718813166,0.2684102356,-0.0240244512,0.1279690713,-0.190717876,-0.3709567189,-0.2314417064,-0.0953482687,0.4701828957,-0.0154063636,-0.3764448464,0.0237512868,-0.1062694117,0.0342046693,-0.0445448607,-0.0578089543,-0.1343722045,-0.1778761595,-0.4469788671,0.2323416471,0.0942371413,0.2168038934,0.2075772136,0.1202893108,0.1229405254,0.1166515648,-0.0237684548,-0.0850124881,0.1562514454,0.1733185798,-0.2898258865,-0.2218644619,0.2827845216,0.0343760476,0.1275027841,0.486094147,-0.2168280184,0.0209328011,-0.1462689936,0.3107803762,0.3316562176,0.1206492707,0.2066342533,0.1792324036,0.212858811,-0.1714295298,0.3644875884,0.1795254499,0.0479368903,-0.359797895,0.0508908667,0.3622294664,-0.0038440262,-0.0362552926,0.1544123888,0.0365234353,-0.2982138991,0.2883263528,0.29725191,1.2974915504,0.2560284734,0.0464656912,0.0794558898,-0.0758616179,0.8155955672,-0.6101807952,-0.0375648066,-0.2863833606,0.1917098165,-0.1548013538,0.0245117545,0.1110631824,0.159916684,-0.4004136026,0.2781204581,0.1337091923,0.6207278967,0.1958462298,0.2526569068,0.25427562,-0.1125703827,-0.0280439854,-0.0023172579,-0.3177216947,0.1961292475,-0.3502788842,0.00829612,0.1352054626,-0.3922473788,-0.5435349345,0.225776121,-0.1676459461,0.1795389503,-0.0912870765,-0.0112566901,0.3942167461,-0.1154548079,0.1764265448,0.3242470622,-0.2129909098,0.3695794344,-0.1648328006,-0.2413542718,-0.0297008883,0.1237418503,-0.1888778061,-0.1956759989,-0.2914250791,0.3017022908,-0.04686746,-0.2820705771,-0.2723125815,0.0242030416,0.4856597483,0.0461111553,-0.3975372016,0.0983780101,0.0978609696,-0.0441413559,0.0495704487,-0.1900753975,-0.1824429333,0.0288909618,0.3470702171,-0.2654399276,-0.0167593956,0.4026792943,-0.127075389,-0.4525042772,0.4531741738,0.3859378994,-0.1967300773,-0.180390209,-0.1222583726,-0.4449328184,-0.6737775207,-0.1479077339,-0.0996818691,0.0669129342,-0.1455649137,0.3657974005,0.2811386585,0.0052279197,0.125691697,-0.3032134175,0.1100802571,-0.0425736606,-0.3786770105,-0.0976213738,0.1424919516,0.2297766507,-0.017026281,-0.0283391494,-0.1148765981,0.0882541016,-0.2866210043,0.1495611668,0.3302124143,0.1623757035,-0.3309537172,0.0084612863,-0.0212530345,0.2480982393,-0.0758762956,-0.0523367003,-0.1137619689,0.2667938769,0.3297121525,-0.1845202446,0.1848600656,0.0214339588,0.1054702774,-0.054009676,-0.0799176246,-0.1602092981,-0.1457357109,0.2436101884,0.1808425635,0.2776176929,-0.5614239573,0.1134632379,0.2133646458,0.5263594389,0.0074057141,0.2616748512,0.2208904475,0.0182323419,-0.113282837,0.1661942005,0.3201208413,0.1643684655,0.1021758243,-0.0240193047,-0.1215306446,0.167190671,-0.2264683545,0.0798175633,-0.3431534171,-0.2567177713,0.5872564912,0.1164017022,-0.1667623818,0.0911753103,0.5686023831,0.1976274997,-0.3388956785,0.0743657798,0.1040031165,-0.3677705824,0.2901095152,0.1933323592,0.3959472477,-0.0799898878,0.1101180837,0.0594559424,-0.2199380249,0.3121532798,0.0542437099,-0.0465967916,0.2164448202,0.4973421395,-0.1014722064,0.0901036784,0.2103209347,0.0874460265,-0.0109352898,-0.202953279,0.2279531211,0.5357630253,0.1257718652,-0.0938544795,-0.1209633797,-0.0986412466,0.1271591187,-0.2292078137,-0.4707059264,0.2094444782,-0.2895659506,0.0469503552,-0.3775494099,-0.0613706373,-0.2470923662,0.2547228932,0.0786868483,-0.1165599078,-0.3547092974,0.3251737654,-0.4041842222,0.001259798,-0.1189509407,0.0802082568,-0.102227591,-0.357457608,0.1956189722,-0.1053231359,-0.0072876783,-0.1527670473,0.1588937193,0.475271374,0.5704238415,-0.3947329521,-0.0232710522,0.071799852,0.2546796799,-0.002400215,0.1421101242,0.2419520766,0.3946742415,0.0906296,0.0486305878,-0.1126253456,-0.277239114,0.2939385772,-0.0378828011,0.0715245605,-0.2892965376,-0.0706139877,0.1751295626,-0.2745937705,-0.0563146137,-0.4367267489,0.2913294435,0.1479400992,0.1088997796,0.095549196,-0.0909631476,-0.0187326167,-0.082537137,0.2150894701,0.2863754332,0.2178199589,-0.2333999574,-0.5668319464,-0.5961482525,0.4565844834,-0.1998843849,-0.1429218352,-0.0620981231,0.0742202252,0.0878898725,0.4057863653,0.0625690594,0.1612211168,-0.3067049682,0.1981173009,-0.191682145,-0.0914506763,0.1853131354,0.0456295423,-0.0771583542,-0.3388835192,0.2339658886,-0.1507810056,-0.0617948622,-0.216673851,-0.2325668782,-0.1130835041,-0.2805165648,0.5375950336,0.117700547,0.446600616,-0.1165416315,0.1682609618,-0.0988690928,0.0563070849,-0.1174238995,-0.034394376,-0.1014875099,0.164502576,0.0381110758,0.3399910927,-0.4156428576,0.3405227363,0.1148373857,0.2810961604,-0.3574502766,-0.0700453296,-0.2362692803,0.0656154901,-0.0179613698,-0.1202620417,0.0245079584,0.1756145358,-0.3756096661,-0.0250428878,0.205877021,-0.4832904041,-0.4173662066,-0.3155452311,0.2172105312,-0.2009653896,-0.2474095374,-0.1549579799,0.0460326411,0.2070869058,0.1569350362,-0.1632024348,-0.0624873936,-0.2600205243,-0.1470813006,-0.0675459728,-0.0197230689,0.1734541208,-0.3940017223,0.0942301154,-0.4644444287]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/320","title":"Blog Authorship Corpus, Non Matching Splits Sizes Error, nlp viewer","comments":"I wonder if this means downloading failed? That corpus has a really slow server.","body":"Selecting `blog_authorship_corpus` in the nlp viewer throws the following error: \r\n\r\n```\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 172, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 132, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\n```\r\n@srush @lhoestq ","comment_length":14,"text":"Blog Authorship Corpus, Non Matching Splits Sizes Error, nlp viewer\nSelecting `blog_authorship_corpus` in the nlp viewer throws the following error: \r\n\r\n```\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 172, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 132, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\n```\r\n@srush @lhoestq \nI wonder if this means downloading failed? That corpus has a really slow server.","embeddings":[-0.2341365665,0.1746635437,0.0528769828,0.2242121845,-0.054630965,-0.0274426695,0.0375437662,0.5022198558,-0.219807744,0.1949803382,-0.0882540569,-0.0733066052,-0.0650148839,-0.0218025353,-0.1484555602,-0.2776799202,-0.2665170133,0.3368445933,0.0040054526,0.0505702458,-0.1330253184,0.3972463012,-0.2592677772,-0.0288467146,-0.0560093895,-0.0945773348,0.1196756065,0.0992417708,-0.1994388103,-0.2846440971,-0.0320585892,-0.2276539057,0.0017965324,0.2405459434,-0.00011832,-0.0521318763,0.3046364784,-0.2610529363,0.1739033759,-0.2233773023,0.0804656073,-0.0924178734,-0.1535943151,-0.238028571,-0.0968984962,0.1424372941,0.236124903,0.0534084402,0.2056047469,0.4859830439,0.1864061654,0.2100055367,-0.0193582587,-0.0822991505,0.0310317855,0.0691109672,-0.2774814069,-0.1139206514,-0.0661811009,-0.3515653908,-0.1020692363,0.4920199513,-0.3369567096,0.0684974566,0.0590126961,-0.1667571515,0.4507972002,-0.3006088436,0.396671921,0.1131951436,0.3707208931,-0.0808910578,0.1284293383,-0.4112168849,-0.0608802065,0.0957361311,0.1510840952,0.519241631,-0.1550329924,-0.1399211586,-0.3092520535,0.1647305042,-0.1608047783,-0.0776638612,0.0328211412,0.4027679265,0.0867302045,0.0405257307,0.0317681022,-0.0571879745,0.2219107747,-0.0434500128,-0.0588857457,0.2149184346,-0.1624520719,-0.253229022,0.0767144412,0.0638756901,0.0667607486,0.2960712016,0.430201292,-0.2747676969,0.3439981341,-0.0774326175,0.236544922,0.1986503452,-0.0894369408,0.713735342,-0.1469518393,0.1481848508,0.0643167347,0.0327584781,0.057278838,-0.4348082542,-0.2825953662,0.1944416165,-0.0128615508,-0.2423688769,-0.4029657543,0.0894374996,-0.2352208942,0.0019084418,0.0437303334,0.3538170755,-0.1384624392,0.3725380599,-0.1321423054,0.1701880246,-0.4958769381,-0.0299406145,-0.2553322315,0.1997325867,-0.2340430766,-0.1152052581,0.1064902097,-0.2341362089,0.4890707731,-0.3374592066,0.2625008523,-0.1843208671,0.0667942911,-0.1719837487,-0.0532936752,0.1553213745,0.0046408223,-0.0583574921,0.051119294,-0.0762766376,-0.2232853025,0.0823742449,0.1880490631,-0.4331123829,0.0729910731,0.2012612522,-0.1427208185,-0.0085852854,0.3555385768,0.2001307905,0.5317850709,-0.2940734625,-0.0616269372,-0.3155483007,-0.1892229617,-0.1323689967,-0.047828842,0.3900066912,0.2923545241,-0.188269496,0.0479633883,0.2060771883,0.6760650873,0.2515402138,0.2933115363,-0.0934091806,-0.0703201145,0.592335403,0.2992485166,-0.36816594,-0.3501451015,0.3049783111,0.0814271122,0.1100988835,0.4589308202,-0.0226863567,0.3694957197,-0.2830116749,-0.1694007516,0.3530231416,-0.3465046585,0.244116649,-0.3757184446,-0.1867751628,0.1422659755,0.0473582484,0.5175161362,-0.2810153067,0.0719522163,0.4198400974,0.2445137799,-0.0958585516,0.124492228,0.0326473378,-0.3409708738,-0.045334734,0.1311379075,-0.1735829711,0.0578606799,0.0684507713,-0.5445121527,0.0909091607,0.319963485,-0.0716683418,-0.4699730873,-0.2672962546,-0.0123809166,-0.1456942111,0.0786250159,0.0600812696,0.2171099633,0.016241258,0.0395699814,0.015812533,-0.0227924157,0.1286682636,-0.4914089739,0.195477441,-0.4119060934,-0.2882962525,0.4236942232,0.4241124094,0.1636872739,0.1043479294,-0.0915229246,0.2704001069,-0.3040343225,0.5339012146,-0.2337966412,-0.119123213,0.2068128586,-0.1963584721,0.2390304059,0.2573575974,0.0397712812,-0.1976047009,0.1838828325,0.223078385,-0.2207979858,0.0989043936,-0.0605803803,0.0326407216,-0.0623252355,-0.2171825171,0.100692749,-0.4573519528,0.6223407984,-0.0146442382,-0.365305692,-0.0328069702,-0.2899564207,0.1150955707,0.2411938161,0.207904011,-0.0253726579,0.0363536589,-0.1495159417,-0.0534571931,0.0021606518,0.1482886076,0.3844685555,0.1326701045,0.169029057,0.0748086274,-0.0515363589,-0.4746567011,0.2813267708,-0.0864149034,0.0323343277,0.6150076985,0.0690596327,0.0600598976,-0.3489787877,0.2118938118,0.1659850776,0.0698252767,-0.2140687853,-0.1929686815,-0.2077674121,-0.6258753538,-0.1261740029,-0.0894451067,-0.1052093059,-0.2106352299,0.2958927155,0.3621723652,-0.2283607125,0.1697584242,-0.166931808,0.0953921825,-0.1257456541,0.2700557709,0.0592220277,-0.2429376841,-0.1207748204,0.1339834481,0.2380245328,0.1384688169,0.2744051516,-0.2108682245,-0.0906060562,0.0105304141,-0.3887266517,-0.0287208948,-0.0631743371,-0.1951724142,0.0804462954,-0.2456770241,0.4262956977,-0.0986747071,0.076004982,-0.0535424016,-0.3460762203,0.1199562326,0.0806695595,0.0065333289,-0.2380538136,-0.21551992,-0.0629038364,-0.6166680455,-0.0915685371,0.442366451,0.3254469931,0.1410059184,-0.2032932341,-0.2182094902,-0.0612077154,0.3252719045,-0.11649777,0.3335394561,0.3101631701,-0.1765716076,-0.2433776557,-0.0459042564,-0.0428632088,0.2199945301,-0.059804827,-0.3879623413,0.0073031373,-0.0993811339,-0.5655648112,-0.0869102031,-0.1579058319,0.4219523072,-0.0200160611,0.0186505076,-0.0466031618,-0.0752872005,-0.0593093596,0.2329156846,0.4591051936,-0.2379428893,0.2790832818,0.2447115779,0.4417745173,0.6506685019,0.2129516155,0.0574815944,0.225023672,-0.1328054965,0.116304785,-0.1153213307,0.8192629218,0.1328855455,0.010454936,0.5705657005,0.0549003221,-0.3037194312,-0.1108631119,-0.1135463566,-0.1897535175,-0.1329934895,0.2126721591,-0.3850331306,0.3022395074,0.0784463882,0.2738236785,-0.1990062296,-0.4706105888,-0.2083248198,0.0380734429,0.047337193,0.1468967497,0.1081883386,0.1974835098,-0.7274335623,0.3355390728,0.199482739,-0.1215412021,-0.2427461147,-0.1555061936,0.0317805409,-0.0126241026,0.0950671956,-0.3479501903,-0.1723843962,-0.1101593971,0.13786681,-0.3510380089,-0.1615861058,0.143239975,0.2933990061,0.2112995535,0.0223892946,-0.2287220657,0.1370001137,0.4354553819,-0.1594472229,-0.0951063707,-0.1316633224,-0.2110888362,-0.316167891,-0.1280887723,0.03689982,0.1188953146,0.3483349383,0.0419903696,-0.0243431944,0.1267143041,-0.2376783341,0.2830742896,0.1491045058,0.1082170084,-0.0503536053,0.096150361,0.1413806528,0.2073110789,-0.0036892528,0.4415619075,0.2652776539,-0.0842813477,-0.0397404619,-0.0708220378,-0.0191242751,0.4573628902,0.2647045851,0.0330888145,-0.0254692435,0.2242768407,-0.2058826834,-0.046867203,0.1015853509,-0.1914954484,-0.2139109373,-0.264880389,0.2391638756,0.0708578825,0.0038733089,0.0017736965,-0.3494585156,-0.2807813287,0.0960862339,0.1922041476,0.9271713495,0.4109401405,-0.2112888694,0.0063017947,-0.238844499,0.3870353401,-0.1661747396,0.4693213701,-0.4435597062,-0.0792939141,-0.0438741781,-0.1080553159,0.0606209561,-0.0772376284,-0.3219486177,0.2458067238,0.1625927538,0.0112755485,-0.1145927832,0.0712137446,-0.367110014,-0.0781507194,0.1206116453,0.0460896678,-0.4622894228,0.4662287235,0.0374447219,-0.0012152063,-0.0522680432,-0.1625233889,-0.5065690875,0.3188253343,-0.2463776916,-0.1297203302,0.1193743721,-0.3374151886,0.288929522,0.3410134315,0.2015173882,-0.2078058422,-0.0850292072,0.2851286232,0.0443106554,-0.2324414253,0.1442331374,0.236148715,0.1982560009,-0.25729087,0.0606638789,-0.0226811599,0.0351091251,-0.0661078915,-0.192957893,-0.0678708032,0.2031794935,-0.1245058998,-0.1689302325,-0.0318932198,-0.1866504848,-0.0103467293,0.0658945963,-0.3126449585,-0.2356596142,0.1632260233,-0.2036496699,-0.1821251214,-0.0514600426,0.2283282876,0.3686068654,0.0155298021,0.4415656924,0.0983654112,-0.2256162614,-0.1847764701,-0.3693659306,-0.4083600044,-0.3803112209,0.1882166266,-0.1573256999,-0.345345825,0.3613977134,0.2624169886,0.1682598144,0.2915204167,-0.1140110046,-0.2506527901,0.0613001361,0.0830985829,-0.3708007634,0.007301881,-0.1438212246,0.5562916398,0.0027539129,0.4266015887,-0.2346434593,-0.1049391925,-0.423753351,0.3038520515,0.0987907872,-0.3049312234,-0.213527903,0.1262171119,-0.0047565871,0.3495808542,-0.2498695999,-0.0964310095,-0.0877258852,0.1599424928,0.1759069413,-0.0947139487,-0.0970179811,0.1231445968,-0.2138710469,-0.0877314657,-0.0194452908,0.0372393988,-0.0256323498,0.0434237532,-0.2768428922,-0.1410388201,0.0307203308,0.2040995359,-0.3766620457,0.2821726799,-0.1075306833,0.2419945598,0.1007174999,-0.0940886065,-0.1642549187,0.0379455499,-0.0284836758,-0.0710049421,0.2844097614,-0.0343592167,0.1698396802,0.2508736253,0.1483680904,0.2195601016,0.102233015,-0.1501446217,0.0848930925,0.1776736826,-0.41090855,0.1344022453,-0.1264771223,0.1109707877,0.0501564816,0.2803552449,0.1536609232,-0.0477125309,-0.123025991,0.2995511591,0.2638212442,-0.2314505279,0.2281989455,0.2588667274,0.1039078087,-0.0019342967,0.3497002721,-0.1886776686,0.2835480869,-0.0265321061,0.0613954142,0.4275463223,0.0060779024,0.2399892807,0.0765475035,-0.1597391665,-0.3672069907,0.3541344106,0.213839516,0.3802344203,0.1098118722,0.4013881981,-0.3140049875,-0.0161995664,-0.3187513649,0.1151976436,-0.051793512,0.0464309268,-0.6466081142,-0.1458352953,-0.229615733,-0.09039893,-0.0282124635,0.3658715487,0.2216522098,-0.037403442,-0.2877688706,-0.3670092523,-0.0355923995,-0.0140140252,0.3070979416,-0.3762354851,0.2843185365,0.1654314995,-0.0390431695,0.1079916283,0.7267916799,0.5513197184,0.2456115484,-0.2686264217,-0.3699727058,0.088211447,-0.0545165613,0.093668744,0.3815398812,0.3370877206,0.0162146967,0.243761003,0.0145370271,-0.1759470105,0.2181914002,0.1912271082,-0.1670649499,0.0372190215,0.4260977209,-0.2845408022,-0.049009189,-0.171482563,-0.0105553549,-0.5934745669,0.0114102317,0.0997597203,-0.1636206657,0.2351559401,-0.1910562515,0.0395667516,0.0959613994,0.3803490996,0.2955439091,-0.4618961811,-0.1453795284,-0.2333026081,-0.1304208338,0.3442293704,-0.3623976111,-0.1869298667,-0.0343966782,0.393037051,-0.2402021736,0.0096203377,0.0207280256,0.434984535,0.0459797829,0.2090399712,-0.1640455574,0.1065761,-0.1200246215,-0.3561321199,0.0761813149,-0.1705269963,0.0934684798,-0.2003173977,0.0382783487,0.0361620411,0.0556777827,0.2787977457,0.3713519275,-0.2958137691,0.294085741,0.2899463773,-0.0592014343,0.0587415881,-0.1465968043,-0.3278466463,-0.4173358679,0.1447609961,0.1439013183,0.5233875513,-0.2440175265,0.2875046432,-0.0943316817,0.2279621661,-0.0677676424,0.0213423818,-0.138580054,-0.0584821701,-0.1580844671,0.0957547054,0.2369379848,0.3930021226,0.0405315794,-0.0469656289,-0.2543900013,-0.6759303808,0.3311416507,-0.2184527665,-0.18531847,0.1943044215,0.0694573745,0.2764319479,0.0643568188,-0.6895744801,0.0946858823,0.0502297357,-0.0715436488,0.2251195312,-0.0590328798,-0.0983622894,-0.3830584288,0.0175965466,-0.4123789966,0.2721537054,-0.2841884494,0.2350588739,-0.0290080775]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/320","title":"Blog Authorship Corpus, Non Matching Splits Sizes Error, nlp viewer","comments":"This dataset seems to have a decoding problem that results in inconsistencies in the number of generated examples.\r\nSee #215.\r\nThat's why we end up with a `NonMatchingSplitsSizesError `.","body":"Selecting `blog_authorship_corpus` in the nlp viewer throws the following error: \r\n\r\n```\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 172, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 132, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\n```\r\n@srush @lhoestq ","comment_length":29,"text":"Blog Authorship Corpus, Non Matching Splits Sizes Error, nlp viewer\nSelecting `blog_authorship_corpus` in the nlp viewer throws the following error: \r\n\r\n```\r\nNonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}]\r\nTraceback:\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 172, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/streamlit\/lib\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp-viewer\/run.py\", line 132, in get\r\n builder_instance.download_and_prepare()\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 432, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 488, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\nFile \"\/home\/sasha\/.local\/share\/virtualenvs\/lib-ogGKnCK_\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\n```\r\n@srush @lhoestq \nThis dataset seems to have a decoding problem that results in inconsistencies in the number of generated examples.\r\nSee #215.\r\nThat's why we end up with a `NonMatchingSplitsSizesError `.","embeddings":[-0.2341365665,0.1746635437,0.0528769828,0.2242121845,-0.054630965,-0.0274426695,0.0375437662,0.5022198558,-0.219807744,0.1949803382,-0.0882540569,-0.0733066052,-0.0650148839,-0.0218025353,-0.1484555602,-0.2776799202,-0.2665170133,0.3368445933,0.0040054526,0.0505702458,-0.1330253184,0.3972463012,-0.2592677772,-0.0288467146,-0.0560093895,-0.0945773348,0.1196756065,0.0992417708,-0.1994388103,-0.2846440971,-0.0320585892,-0.2276539057,0.0017965324,0.2405459434,-0.00011832,-0.0521318763,0.3046364784,-0.2610529363,0.1739033759,-0.2233773023,0.0804656073,-0.0924178734,-0.1535943151,-0.238028571,-0.0968984962,0.1424372941,0.236124903,0.0534084402,0.2056047469,0.4859830439,0.1864061654,0.2100055367,-0.0193582587,-0.0822991505,0.0310317855,0.0691109672,-0.2774814069,-0.1139206514,-0.0661811009,-0.3515653908,-0.1020692363,0.4920199513,-0.3369567096,0.0684974566,0.0590126961,-0.1667571515,0.4507972002,-0.3006088436,0.396671921,0.1131951436,0.3707208931,-0.0808910578,0.1284293383,-0.4112168849,-0.0608802065,0.0957361311,0.1510840952,0.519241631,-0.1550329924,-0.1399211586,-0.3092520535,0.1647305042,-0.1608047783,-0.0776638612,0.0328211412,0.4027679265,0.0867302045,0.0405257307,0.0317681022,-0.0571879745,0.2219107747,-0.0434500128,-0.0588857457,0.2149184346,-0.1624520719,-0.253229022,0.0767144412,0.0638756901,0.0667607486,0.2960712016,0.430201292,-0.2747676969,0.3439981341,-0.0774326175,0.236544922,0.1986503452,-0.0894369408,0.713735342,-0.1469518393,0.1481848508,0.0643167347,0.0327584781,0.057278838,-0.4348082542,-0.2825953662,0.1944416165,-0.0128615508,-0.2423688769,-0.4029657543,0.0894374996,-0.2352208942,0.0019084418,0.0437303334,0.3538170755,-0.1384624392,0.3725380599,-0.1321423054,0.1701880246,-0.4958769381,-0.0299406145,-0.2553322315,0.1997325867,-0.2340430766,-0.1152052581,0.1064902097,-0.2341362089,0.4890707731,-0.3374592066,0.2625008523,-0.1843208671,0.0667942911,-0.1719837487,-0.0532936752,0.1553213745,0.0046408223,-0.0583574921,0.051119294,-0.0762766376,-0.2232853025,0.0823742449,0.1880490631,-0.4331123829,0.0729910731,0.2012612522,-0.1427208185,-0.0085852854,0.3555385768,0.2001307905,0.5317850709,-0.2940734625,-0.0616269372,-0.3155483007,-0.1892229617,-0.1323689967,-0.047828842,0.3900066912,0.2923545241,-0.188269496,0.0479633883,0.2060771883,0.6760650873,0.2515402138,0.2933115363,-0.0934091806,-0.0703201145,0.592335403,0.2992485166,-0.36816594,-0.3501451015,0.3049783111,0.0814271122,0.1100988835,0.4589308202,-0.0226863567,0.3694957197,-0.2830116749,-0.1694007516,0.3530231416,-0.3465046585,0.244116649,-0.3757184446,-0.1867751628,0.1422659755,0.0473582484,0.5175161362,-0.2810153067,0.0719522163,0.4198400974,0.2445137799,-0.0958585516,0.124492228,0.0326473378,-0.3409708738,-0.045334734,0.1311379075,-0.1735829711,0.0578606799,0.0684507713,-0.5445121527,0.0909091607,0.319963485,-0.0716683418,-0.4699730873,-0.2672962546,-0.0123809166,-0.1456942111,0.0786250159,0.0600812696,0.2171099633,0.016241258,0.0395699814,0.015812533,-0.0227924157,0.1286682636,-0.4914089739,0.195477441,-0.4119060934,-0.2882962525,0.4236942232,0.4241124094,0.1636872739,0.1043479294,-0.0915229246,0.2704001069,-0.3040343225,0.5339012146,-0.2337966412,-0.119123213,0.2068128586,-0.1963584721,0.2390304059,0.2573575974,0.0397712812,-0.1976047009,0.1838828325,0.223078385,-0.2207979858,0.0989043936,-0.0605803803,0.0326407216,-0.0623252355,-0.2171825171,0.100692749,-0.4573519528,0.6223407984,-0.0146442382,-0.365305692,-0.0328069702,-0.2899564207,0.1150955707,0.2411938161,0.207904011,-0.0253726579,0.0363536589,-0.1495159417,-0.0534571931,0.0021606518,0.1482886076,0.3844685555,0.1326701045,0.169029057,0.0748086274,-0.0515363589,-0.4746567011,0.2813267708,-0.0864149034,0.0323343277,0.6150076985,0.0690596327,0.0600598976,-0.3489787877,0.2118938118,0.1659850776,0.0698252767,-0.2140687853,-0.1929686815,-0.2077674121,-0.6258753538,-0.1261740029,-0.0894451067,-0.1052093059,-0.2106352299,0.2958927155,0.3621723652,-0.2283607125,0.1697584242,-0.166931808,0.0953921825,-0.1257456541,0.2700557709,0.0592220277,-0.2429376841,-0.1207748204,0.1339834481,0.2380245328,0.1384688169,0.2744051516,-0.2108682245,-0.0906060562,0.0105304141,-0.3887266517,-0.0287208948,-0.0631743371,-0.1951724142,0.0804462954,-0.2456770241,0.4262956977,-0.0986747071,0.076004982,-0.0535424016,-0.3460762203,0.1199562326,0.0806695595,0.0065333289,-0.2380538136,-0.21551992,-0.0629038364,-0.6166680455,-0.0915685371,0.442366451,0.3254469931,0.1410059184,-0.2032932341,-0.2182094902,-0.0612077154,0.3252719045,-0.11649777,0.3335394561,0.3101631701,-0.1765716076,-0.2433776557,-0.0459042564,-0.0428632088,0.2199945301,-0.059804827,-0.3879623413,0.0073031373,-0.0993811339,-0.5655648112,-0.0869102031,-0.1579058319,0.4219523072,-0.0200160611,0.0186505076,-0.0466031618,-0.0752872005,-0.0593093596,0.2329156846,0.4591051936,-0.2379428893,0.2790832818,0.2447115779,0.4417745173,0.6506685019,0.2129516155,0.0574815944,0.225023672,-0.1328054965,0.116304785,-0.1153213307,0.8192629218,0.1328855455,0.010454936,0.5705657005,0.0549003221,-0.3037194312,-0.1108631119,-0.1135463566,-0.1897535175,-0.1329934895,0.2126721591,-0.3850331306,0.3022395074,0.0784463882,0.2738236785,-0.1990062296,-0.4706105888,-0.2083248198,0.0380734429,0.047337193,0.1468967497,0.1081883386,0.1974835098,-0.7274335623,0.3355390728,0.199482739,-0.1215412021,-0.2427461147,-0.1555061936,0.0317805409,-0.0126241026,0.0950671956,-0.3479501903,-0.1723843962,-0.1101593971,0.13786681,-0.3510380089,-0.1615861058,0.143239975,0.2933990061,0.2112995535,0.0223892946,-0.2287220657,0.1370001137,0.4354553819,-0.1594472229,-0.0951063707,-0.1316633224,-0.2110888362,-0.316167891,-0.1280887723,0.03689982,0.1188953146,0.3483349383,0.0419903696,-0.0243431944,0.1267143041,-0.2376783341,0.2830742896,0.1491045058,0.1082170084,-0.0503536053,0.096150361,0.1413806528,0.2073110789,-0.0036892528,0.4415619075,0.2652776539,-0.0842813477,-0.0397404619,-0.0708220378,-0.0191242751,0.4573628902,0.2647045851,0.0330888145,-0.0254692435,0.2242768407,-0.2058826834,-0.046867203,0.1015853509,-0.1914954484,-0.2139109373,-0.264880389,0.2391638756,0.0708578825,0.0038733089,0.0017736965,-0.3494585156,-0.2807813287,0.0960862339,0.1922041476,0.9271713495,0.4109401405,-0.2112888694,0.0063017947,-0.238844499,0.3870353401,-0.1661747396,0.4693213701,-0.4435597062,-0.0792939141,-0.0438741781,-0.1080553159,0.0606209561,-0.0772376284,-0.3219486177,0.2458067238,0.1625927538,0.0112755485,-0.1145927832,0.0712137446,-0.367110014,-0.0781507194,0.1206116453,0.0460896678,-0.4622894228,0.4662287235,0.0374447219,-0.0012152063,-0.0522680432,-0.1625233889,-0.5065690875,0.3188253343,-0.2463776916,-0.1297203302,0.1193743721,-0.3374151886,0.288929522,0.3410134315,0.2015173882,-0.2078058422,-0.0850292072,0.2851286232,0.0443106554,-0.2324414253,0.1442331374,0.236148715,0.1982560009,-0.25729087,0.0606638789,-0.0226811599,0.0351091251,-0.0661078915,-0.192957893,-0.0678708032,0.2031794935,-0.1245058998,-0.1689302325,-0.0318932198,-0.1866504848,-0.0103467293,0.0658945963,-0.3126449585,-0.2356596142,0.1632260233,-0.2036496699,-0.1821251214,-0.0514600426,0.2283282876,0.3686068654,0.0155298021,0.4415656924,0.0983654112,-0.2256162614,-0.1847764701,-0.3693659306,-0.4083600044,-0.3803112209,0.1882166266,-0.1573256999,-0.345345825,0.3613977134,0.2624169886,0.1682598144,0.2915204167,-0.1140110046,-0.2506527901,0.0613001361,0.0830985829,-0.3708007634,0.007301881,-0.1438212246,0.5562916398,0.0027539129,0.4266015887,-0.2346434593,-0.1049391925,-0.423753351,0.3038520515,0.0987907872,-0.3049312234,-0.213527903,0.1262171119,-0.0047565871,0.3495808542,-0.2498695999,-0.0964310095,-0.0877258852,0.1599424928,0.1759069413,-0.0947139487,-0.0970179811,0.1231445968,-0.2138710469,-0.0877314657,-0.0194452908,0.0372393988,-0.0256323498,0.0434237532,-0.2768428922,-0.1410388201,0.0307203308,0.2040995359,-0.3766620457,0.2821726799,-0.1075306833,0.2419945598,0.1007174999,-0.0940886065,-0.1642549187,0.0379455499,-0.0284836758,-0.0710049421,0.2844097614,-0.0343592167,0.1698396802,0.2508736253,0.1483680904,0.2195601016,0.102233015,-0.1501446217,0.0848930925,0.1776736826,-0.41090855,0.1344022453,-0.1264771223,0.1109707877,0.0501564816,0.2803552449,0.1536609232,-0.0477125309,-0.123025991,0.2995511591,0.2638212442,-0.2314505279,0.2281989455,0.2588667274,0.1039078087,-0.0019342967,0.3497002721,-0.1886776686,0.2835480869,-0.0265321061,0.0613954142,0.4275463223,0.0060779024,0.2399892807,0.0765475035,-0.1597391665,-0.3672069907,0.3541344106,0.213839516,0.3802344203,0.1098118722,0.4013881981,-0.3140049875,-0.0161995664,-0.3187513649,0.1151976436,-0.051793512,0.0464309268,-0.6466081142,-0.1458352953,-0.229615733,-0.09039893,-0.0282124635,0.3658715487,0.2216522098,-0.037403442,-0.2877688706,-0.3670092523,-0.0355923995,-0.0140140252,0.3070979416,-0.3762354851,0.2843185365,0.1654314995,-0.0390431695,0.1079916283,0.7267916799,0.5513197184,0.2456115484,-0.2686264217,-0.3699727058,0.088211447,-0.0545165613,0.093668744,0.3815398812,0.3370877206,0.0162146967,0.243761003,0.0145370271,-0.1759470105,0.2181914002,0.1912271082,-0.1670649499,0.0372190215,0.4260977209,-0.2845408022,-0.049009189,-0.171482563,-0.0105553549,-0.5934745669,0.0114102317,0.0997597203,-0.1636206657,0.2351559401,-0.1910562515,0.0395667516,0.0959613994,0.3803490996,0.2955439091,-0.4618961811,-0.1453795284,-0.2333026081,-0.1304208338,0.3442293704,-0.3623976111,-0.1869298667,-0.0343966782,0.393037051,-0.2402021736,0.0096203377,0.0207280256,0.434984535,0.0459797829,0.2090399712,-0.1640455574,0.1065761,-0.1200246215,-0.3561321199,0.0761813149,-0.1705269963,0.0934684798,-0.2003173977,0.0382783487,0.0361620411,0.0556777827,0.2787977457,0.3713519275,-0.2958137691,0.294085741,0.2899463773,-0.0592014343,0.0587415881,-0.1465968043,-0.3278466463,-0.4173358679,0.1447609961,0.1439013183,0.5233875513,-0.2440175265,0.2875046432,-0.0943316817,0.2279621661,-0.0677676424,0.0213423818,-0.138580054,-0.0584821701,-0.1580844671,0.0957547054,0.2369379848,0.3930021226,0.0405315794,-0.0469656289,-0.2543900013,-0.6759303808,0.3311416507,-0.2184527665,-0.18531847,0.1943044215,0.0694573745,0.2764319479,0.0643568188,-0.6895744801,0.0946858823,0.0502297357,-0.0715436488,0.2251195312,-0.0590328798,-0.0983622894,-0.3830584288,0.0175965466,-0.4123789966,0.2721537054,-0.2841884494,0.2350588739,-0.0290080775]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/319","title":"Nested sequences with dicts","comments":"Oh yes, this is a backward compatibility feature with tensorflow_dataset in which a `Sequence` or `dict` is converted in a `dict` of `lists`, unfortunately it is not very intuitive, see here: https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/features.py#L409\r\n\r\nTo avoid this behavior, you can just define the list in the feature with a simple list or a tuple (which is also simpler to write).\r\nIn your case, the features could be as follow:\r\n``` python\r\n...\r\nfeatures=nlp.Features({\r\n \"title\": nlp.Value(\"string\"),\r\n \"vertexSet\": [[{\r\n \"name\": nlp.Value(\"string\"),\r\n \"sent_id\": nlp.Value(\"int32\"),\r\n \"pos\": nlp.features.Sequence(nlp.Value(\"int32\")),\r\n \"type\": nlp.Value(\"string\"),\r\n }]],\r\n ...\r\n }),\r\n...\r\n```","body":"Am pretty much finished [adding a dataset](https:\/\/github.com\/ghomasHudson\/nlp\/blob\/DocRED\/datasets\/docred\/docred.py) for [DocRED](https:\/\/github.com\/thunlp\/DocRED), but am getting an error when trying to add a nested `nlp.features.sequence(nlp.features.sequence({key:value,...}))`. \r\n\r\nThe original data is in this format:\r\n```python\r\n{\r\n 'title': \"Title of wiki page\",\r\n 'vertexSet': [\r\n [\r\n { 'name': \"mention_name\", \r\n 'sent_id': \"mention in which sentence\", \r\n 'pos': [\"postion of mention in a sentence\"], \r\n 'type': \"NER_type\"},\r\n {another mention}\r\n ], \r\n [another entity]\r\n ]\r\n ...\r\n}\r\n```\r\nSo to represent this I've attempted to write:\r\n```\r\n...\r\nfeatures=nlp.Features({\r\n \"title\": nlp.Value(\"string\"),\r\n \"vertexSet\": nlp.features.Sequence(nlp.features.Sequence({\r\n \"name\": nlp.Value(\"string\"),\r\n \"sent_id\": nlp.Value(\"int32\"),\r\n \"pos\": nlp.features.Sequence(nlp.Value(\"int32\")),\r\n \"type\": nlp.Value(\"string\"),\r\n })),\r\n ...\r\n }),\r\n...\r\n```\r\nThis is giving me the error:\r\n```\r\npyarrow.lib.ArrowTypeError: Could not convert [{'pos': [[0,2], [2,4], [3,5]], \"type\": [\"ORG\", \"ORG\", \"ORG\"], \"name\": [\"Lark Force\", \"Lark Force\", \"Lark Force\", \"sent_id\": [0, 3, 4]}..... with type list: was not a dict, tuple, or recognized null value for conversion to struct type\r\n```\r\nDo we expect the pyarrow stuff to break when doing this deeper nesting? I've checked that it still works when you do `nlp.features.Sequence(nlp.features.Sequence(nlp.Value(\"string\"))` or `nlp.features.Sequence({key:value,...})` just not nested sequences with a dict.\r\n\r\nIf it's not possible, I can always convert it to a shallower structure. I'd rather not change the DocRED authors' structure if I don't have to though.","comment_length":88,"text":"Nested sequences with dicts\nAm pretty much finished [adding a dataset](https:\/\/github.com\/ghomasHudson\/nlp\/blob\/DocRED\/datasets\/docred\/docred.py) for [DocRED](https:\/\/github.com\/thunlp\/DocRED), but am getting an error when trying to add a nested `nlp.features.sequence(nlp.features.sequence({key:value,...}))`. \r\n\r\nThe original data is in this format:\r\n```python\r\n{\r\n 'title': \"Title of wiki page\",\r\n 'vertexSet': [\r\n [\r\n { 'name': \"mention_name\", \r\n 'sent_id': \"mention in which sentence\", \r\n 'pos': [\"postion of mention in a sentence\"], \r\n 'type': \"NER_type\"},\r\n {another mention}\r\n ], \r\n [another entity]\r\n ]\r\n ...\r\n}\r\n```\r\nSo to represent this I've attempted to write:\r\n```\r\n...\r\nfeatures=nlp.Features({\r\n \"title\": nlp.Value(\"string\"),\r\n \"vertexSet\": nlp.features.Sequence(nlp.features.Sequence({\r\n \"name\": nlp.Value(\"string\"),\r\n \"sent_id\": nlp.Value(\"int32\"),\r\n \"pos\": nlp.features.Sequence(nlp.Value(\"int32\")),\r\n \"type\": nlp.Value(\"string\"),\r\n })),\r\n ...\r\n }),\r\n...\r\n```\r\nThis is giving me the error:\r\n```\r\npyarrow.lib.ArrowTypeError: Could not convert [{'pos': [[0,2], [2,4], [3,5]], \"type\": [\"ORG\", \"ORG\", \"ORG\"], \"name\": [\"Lark Force\", \"Lark Force\", \"Lark Force\", \"sent_id\": [0, 3, 4]}..... with type list: was not a dict, tuple, or recognized null value for conversion to struct type\r\n```\r\nDo we expect the pyarrow stuff to break when doing this deeper nesting? I've checked that it still works when you do `nlp.features.Sequence(nlp.features.Sequence(nlp.Value(\"string\"))` or `nlp.features.Sequence({key:value,...})` just not nested sequences with a dict.\r\n\r\nIf it's not possible, I can always convert it to a shallower structure. I'd rather not change the DocRED authors' structure if I don't have to though.\nOh yes, this is a backward compatibility feature with tensorflow_dataset in which a `Sequence` or `dict` is converted in a `dict` of `lists`, unfortunately it is not very intuitive, see here: https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/features.py#L409\r\n\r\nTo avoid this behavior, you can just define the list in the feature with a simple list or a tuple (which is also simpler to write).\r\nIn your case, the features could be as follow:\r\n``` python\r\n...\r\nfeatures=nlp.Features({\r\n \"title\": nlp.Value(\"string\"),\r\n \"vertexSet\": [[{\r\n \"name\": nlp.Value(\"string\"),\r\n \"sent_id\": nlp.Value(\"int32\"),\r\n \"pos\": nlp.features.Sequence(nlp.Value(\"int32\")),\r\n \"type\": nlp.Value(\"string\"),\r\n }]],\r\n ...\r\n }),\r\n...\r\n```","embeddings":[0.1117630079,0.0667421669,-0.0714067966,-0.0210161116,-0.1091683283,-0.0410184972,0.198786512,0.1697288007,-0.0871766582,-0.4004120827,0.2879687548,0.3965295851,-0.0898342356,0.203804642,-0.0607240349,-0.2475784123,0.3430915177,0.1289698482,0.3750735223,0.2590895891,-0.2045025676,0.3637642562,-0.2815240622,0.1494687051,-0.2481655627,-0.3701804876,-0.3637534678,0.1604201645,-0.0595387518,-0.7306870818,0.327549696,0.2191215903,0.1450784504,0.1276401281,-0.0001186873,-0.0521878116,0.5560513139,-0.0750035197,-0.1546786278,-0.0694720596,0.007484843,-0.307399869,0.296666503,-0.2497453988,0.0215846244,-0.4088245034,-0.0574822165,-0.0613234304,0.3271240592,0.2150971889,0.1544761658,0.0688155666,0.3399304748,0.0271114036,0.4079800844,0.3201031387,-0.1407759339,0.1663142145,0.4056244791,-0.3216360211,0.0209346395,-0.097893022,-0.1353191733,-0.2777380347,0.3606657982,0.2268982381,0.0491360761,-0.1841319948,-0.5392251015,-0.0223106332,0.1228881702,-0.2456724495,-0.1384012252,-0.4200358987,0.0036535414,-0.4405346215,-0.0079782829,-0.1407092363,-0.2733757794,-0.0963975191,0.2459769398,-0.2561305463,-0.2779055834,0.3951284587,0.1447310299,0.2310712934,0.1184960008,0.0992585048,-0.1329668909,-0.4175079465,0.1348029524,-0.2447813153,0.0278784782,0.1005075127,0.0098711848,-0.1878099591,-0.0269921366,-0.1887381971,0.1825390458,0.0152063826,0.1940030903,-0.0737635717,-0.0238401555,0.0430311337,0.3456587493,0.1764035821,0.010860147,-0.1715756059,-0.066734232,0.311568588,0.0790737867,0.0870562047,-0.1188078448,0.0132933408,-0.2246382236,0.214470163,0.3883276284,0.0419625193,-0.1750754863,0.0624773428,-0.3113225698,0.0604519732,-0.0119610773,0.0920138583,0.0700005367,0.2033096254,0.1066470966,0.3950109184,0.0666037574,0.0462857634,-0.0807010829,0.007730695,-0.104136087,0.0657802448,-0.0892218798,0.1868112236,0.1713126451,-0.1115265265,0.1340780258,-0.0093421172,-0.2380873412,-0.0547784753,0.2477760762,-0.1615775973,-0.2585242987,0.1531671137,0.1702297181,-0.4171901941,-0.0409002751,0.1030413508,-0.1227635741,-0.2476044893,-0.3633392453,0.0623023622,-0.3166223764,-0.0854944289,-0.1410175264,0.3202441037,0.5429484844,-0.1632453352,-0.1814660281,-0.1992941946,-0.2664177418,-0.3964797556,-0.0749630183,0.1141208187,-0.2213226855,0.1271057725,-0.3402887285,0.1647937149,0.1423597783,0.1691092998,-0.0593792237,0.328269124,-0.2519994378,0.7448478937,0.2469037622,-0.1542544216,-0.0559407398,0.1489606947,-0.0351052023,0.1376333386,-0.0552322716,0.0722646639,0.1872952878,-0.3772203028,0.2196923494,-0.0028167854,0.0423983037,-0.3566228151,-0.1894056201,-0.0168206748,0.4003439248,-0.2266631722,-0.3518343568,-0.1862152517,-0.110184744,0.146551773,0.1216539294,-0.1506718546,0.2942144573,0.328137964,0.0982828662,0.0306445342,-0.0221919771,-0.1801386476,-0.6082834601,-0.0778388083,-0.4564249814,0.2655048966,-0.5156535506,-0.1381058097,-0.2986238003,0.2917177081,-0.2939352989,-0.0874444768,0.1657339483,0.0457946025,-0.0814892352,-0.1121656895,-0.0471532084,-0.2919530869,0.1465135962,-0.0049193334,-0.2575115561,0.4828610718,-0.0993677378,-0.3981579542,0.347478956,0.6011694074,0.2996060848,0.0108097121,0.2203307748,0.0839651003,0.0614008829,-0.270619154,-0.2162835747,-0.1576783806,0.1663332582,-0.5653815269,-0.0550803356,0.2663568854,-0.0157449953,-0.3664388955,-0.1052337065,0.3159486651,-0.0346250944,0.407774955,-0.1733401865,0.2705794871,-0.0348707028,0.1374605447,-0.1708569527,-0.333034128,-0.1048401818,0.420381546,-0.3121208847,-0.1443029046,-0.1692670435,0.479344815,0.3534285724,0.1225575507,0.0461264104,0.1554606557,-0.0794245601,-0.2559583485,0.1802677661,-0.2929849625,0.2055215836,0.2246131599,0.1453634948,0.0807232931,-0.2061256766,-0.0572400056,0.1440231949,-0.0237121787,0.3502274454,0.2081848532,0.3714060485,-0.1997290254,-0.0740707442,-0.410828352,-0.0180833675,0.1100207567,-0.2863145173,0.0280116592,-0.1382289827,-0.4561456442,-0.0179674197,-0.4906527698,-0.1596070677,-0.5329732299,-0.0305234063,0.1939590275,-0.1003117636,0.2125135064,0.1825372428,-0.2348060459,0.2344148457,-0.3206106722,-0.2240641713,-0.4881903827,-0.244237259,0.1311269253,0.151558429,0.0488087609,0.0672167316,0.2802769244,-0.0738631263,-0.1561627984,-0.5505623221,0.03042458,-0.4866124988,0.0957419798,0.4155690074,0.2472601235,0.1292892694,-0.4227377176,0.3008897007,0.5386326909,-0.3785912991,-0.0500573143,-0.0442277305,-0.0153104411,-0.0904574469,-0.0681429282,-0.074659206,-0.2739903033,0.3004772067,0.3360037506,0.1556708068,0.1452372372,-0.0839370787,0.0632244572,0.1560751498,-0.2299247682,0.073064968,0.4016881585,0.3970456123,-0.270750016,-0.3289731443,0.0066126189,-0.4844303131,0.0295713749,-0.1177048162,-0.1482520252,0.1077740118,0.1482848078,0.4257537127,0.3425695896,-0.3354375362,0.3598234951,0.4960214496,-0.0522433035,-0.0866467729,-0.0341401473,0.2497750968,0.1586211473,0.2001061589,0.2461517751,0.0882550776,0.1381200552,0.3682236671,0.450081557,-0.0455315635,0.41033113,-0.1635897905,-0.0244688429,-0.2582240701,0.0070215468,-0.3686869144,-0.1982266307,0.0572778471,0.0289594047,-0.1232874915,-0.3941147625,0.0881601498,0.0466540568,0.2459708005,-0.1358118206,0.1118961275,-0.5101607442,0.2923166156,-0.1387941986,-0.3272828758,-0.0665582344,-0.4218162894,-0.0783467144,-0.179451257,0.0777285248,-0.0237211529,-0.3296523392,-0.0878597051,-0.4834107161,0.210983336,0.3980477154,0.1980572045,-0.0530762039,-0.102268979,-0.1097588763,-0.223867029,0.1809992939,-0.1099721715,-0.1893151253,0.4855351746,0.1183651313,-0.421487838,-0.0338473022,-0.089671351,0.1774788201,0.0912995264,0.4397976696,-0.6211690903,0.0020221369,0.0713741854,0.0922099054,-0.067706719,-0.0741995052,-0.0735139027,-0.2639946043,-0.1719910204,0.0082074087,0.3025734425,0.4619008303,0.0121734142,-0.2804292142,-0.0899339616,-0.1619850844,0.0362528078,-0.1058387384,0.2589027882,-0.1245844662,-0.0514228679,0.0134861628,0.2822630107,0.1386035383,0.2392872274,0.3220471144,0.1425617933,-0.1194690391,-0.0797488391,0.3786869049,0.1363582909,0.0171595495,0.2974936366,0.1562426388,-0.1546482146,-0.1992692798,-0.0542862676,0.3101430833,-0.1684685498,-0.3162243962,-0.0562878288,0.2843524516,0.0125213098,-0.0801618174,-0.0388851017,0.0294120815,-0.4231781662,0.4896873236,0.2276365161,0.9587829113,0.1433469802,0.3334242105,0.7011730075,0.1009090543,0.5256799459,0.1235473305,0.0278157499,-0.3482498527,0.331215471,-0.0938304365,-0.1970284432,0.0337508209,0.1511791497,-0.3989120424,0.207625851,-0.0439409018,-0.0097749382,0.0969633833,0.0773602203,0.4406331778,-0.039881628,-0.1946097463,0.0343465842,-0.0421703458,-0.0641203746,-0.029214168,-0.5158143044,-0.4526414871,-0.1939998418,-0.1262547225,0.0731220618,0.0002202183,0.1277303249,0.1020550802,-0.3356178403,0.1802397668,0.5281002522,-0.0187885799,-0.3472119272,-0.2047382146,-0.2765871584,0.0460747667,-0.3822996318,0.0796269625,-0.2330664098,0.1702176929,-0.1389924735,0.0583875291,0.0667981058,-0.102518484,-0.1171067953,-0.1245573685,0.1298495531,-0.2100258172,-0.2777292728,-0.242689997,0.2711513638,-0.264326483,-0.005320841,0.0688451976,0.0013985023,-0.2149065435,0.0882434919,0.006676903,0.2009322494,0.1078080833,0.1676503867,0.3902579844,0.2498512566,0.6531678438,0.3535989523,-0.269903034,-0.0370758586,-0.2008210719,0.135663569,-0.2612209916,0.273147881,0.0095654866,-0.1079514325,-0.1851249039,0.5661982894,-0.0359059982,0.3272174001,-0.0294842795,-0.2730292678,-0.4289377034,0.3417748511,-0.3376172185,0.1372971386,-0.2105008364,0.3417845666,-0.0200437475,-0.0866845027,-0.2319801748,-0.0338974334,-0.2695265114,0.2831506729,0.0727759823,-0.2418663204,0.1432742476,0.1637407839,0.1134920493,0.1377165616,0.0091405399,-0.1189611331,-0.0946801454,0.1483018547,0.2462124825,-0.0146383718,0.1611179262,0.1037167013,0.0072389306,-0.0934785977,-0.1661732644,-0.2070735246,-0.0978035405,0.2616793215,0.2590046227,0.060883116,-0.1613887399,0.1721501201,-0.0749573857,0.2543973029,-0.1327088177,0.2022132576,-0.13597022,-0.0658146665,0.0559127592,0.2971179187,-0.0877368674,-0.0780258477,0.418602258,-0.08177834,-0.4433643222,-0.271982342,0.3999477625,0.2681146562,-0.1009983644,-0.0830782428,-0.1121801957,0.0747596994,-0.3760126531,-0.0077785593,0.2860894501,0.0323821567,0.0204305686,0.2159101814,0.3058033884,-0.0404888801,-0.2148513347,-0.0401835069,0.0060540643,-0.0112841111,0.0295729246,0.0908346325,0.1468585283,0.0643878654,0.1996053457,0.0061445376,0.3186852336,0.3405147791,0.2255375832,0.0747069493,0.7099593878,-0.0206038561,0.4179551303,0.0624392107,-0.0959222093,0.2301591337,0.8155221939,-0.0801276267,0.2281221598,0.2109827548,0.4051736891,-0.1539326459,0.2298147231,0.4492286146,-0.069603391,-0.2035358995,-0.1884580255,0.0116918851,-0.2171999365,-0.4033898711,-0.2008129507,-0.2008017153,0.1609137207,-0.085551843,0.3172802031,-0.3578862846,0.6291168928,0.235927999,0.1579395682,-0.0687289014,0.398370713,0.2859929204,-0.1227333024,0.1200389117,0.3106322587,0.5210928321,0.2319259793,-0.4977974296,-0.0201339852,-0.1575997174,-0.3119957745,0.0452808589,0.155756712,0.1033885777,0.1714435667,0.4270237684,0.0027688146,0.0035369045,-0.00669146,0.3002275229,-0.1084526852,0.2530638576,0.5017457604,-0.2810668349,-0.3682081401,0.163339287,0.028058093,-0.2052233666,-0.1081204265,0.3654825687,0.2288480848,0.2794364691,0.0771409497,0.0592629127,-0.0933699161,0.595961988,0.1037611663,0.3081622422,-0.1546451747,-0.2343511879,-0.3536604345,0.0036337508,-0.0565190315,-0.3216262758,0.3073219657,-0.0111772381,0.0589239858,0.1984140128,-0.1791730821,0.0193470046,-0.2994587123,0.2204196155,-0.3973325193,0.0502125509,0.3911755085,0.0467459001,-0.08370433,0.023023665,0.0510608181,0.0286290664,-0.0086941039,-0.1330474764,-0.1229645461,-0.0937700272,0.2239985168,-0.0495796539,0.2552890182,0.2741926908,0.0716451183,-0.1535004526,-0.1562321335,-0.2056280673,-0.0690802857,0.6255393624,0.089099586,0.4591796696,-0.2144263387,0.2980739772,-0.0728881434,-0.0675850287,0.0347643346,-0.0315892696,-0.3935305178,0.3084968925,-0.2421967834,0.2191549987,-0.0817307234,0.0946832672,0.0168624241,0.2577867508,-0.2422178686,-0.4130553007,0.3898530304,-0.1153757796,-0.2464152277,0.4650785923,-0.1455704123,0.424063921,0.0369960032,-0.1442358643,-0.2059316635,0.4398931563,-0.1090039387,-0.3411955833,-0.1297612488,-0.0677590519,-0.1385493129,-0.181083262,0.1994468868,-0.1636886597,-0.0651477948,-0.2459084094,-0.4080276489]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/317","title":"Adding a dataset with multiple subtasks","comments":"For one dataset you can have different configurations that each have their own `nlp.Features`.\r\nWe imagine having one configuration per subtask for example.\r\nThey are loaded with `nlp.load_dataset(\"my_dataset\", \"my_config\")`.\r\n\r\nFor example the `glue` dataset has many configurations. It is a bit different from your case though because each configuration is a dataset by itself (sst2, mnli).\r\nAnother example is `wikipedia` that has one configuration per language.","body":"I intent to add the datasets of the MT Quality Estimation shared tasks to `nlp`. However, they have different subtasks -- such as word-level, sentence-level and document-level quality estimation, each of which having different language pairs, and some of the data reused in different subtasks.\r\n\r\nFor example, in [QE 2019,](http:\/\/www.statmt.org\/wmt19\/qe-task.html) we had the same English-Russian and English-German data for word-level and sentence-level QE. \r\n\r\nI suppose these datasets could have both their word and sentence-level labels inside `nlp.Features`; but what about other subtasks? Should they be considered a different dataset altogether?\r\n\r\nI read the discussion on #217 but the case of QE seems a lot simpler.","comment_length":66,"text":"Adding a dataset with multiple subtasks\nI intent to add the datasets of the MT Quality Estimation shared tasks to `nlp`. However, they have different subtasks -- such as word-level, sentence-level and document-level quality estimation, each of which having different language pairs, and some of the data reused in different subtasks.\r\n\r\nFor example, in [QE 2019,](http:\/\/www.statmt.org\/wmt19\/qe-task.html) we had the same English-Russian and English-German data for word-level and sentence-level QE. \r\n\r\nI suppose these datasets could have both their word and sentence-level labels inside `nlp.Features`; but what about other subtasks? Should they be considered a different dataset altogether?\r\n\r\nI read the discussion on #217 but the case of QE seems a lot simpler.\nFor one dataset you can have different configurations that each have their own `nlp.Features`.\r\nWe imagine having one configuration per subtask for example.\r\nThey are loaded with `nlp.load_dataset(\"my_dataset\", \"my_config\")`.\r\n\r\nFor example the `glue` dataset has many configurations. It is a bit different from your case though because each configuration is a dataset by itself (sst2, mnli).\r\nAnother example is `wikipedia` that has one configuration per language.","embeddings":[-0.0709952563,-0.1814528406,-0.0437028371,0.3849149942,-0.2169752419,0.0946593061,0.3582087159,-0.1335341632,0.1271398067,-0.1552469432,-0.2816208005,0.1489305347,-0.1571817994,0.6809743047,0.3387671411,-0.2282722294,0.1595131904,-0.3534742594,-0.0652335063,-0.0188693497,-0.1657261997,0.0523259714,-0.0485033765,-0.0104857823,-0.3616020381,-0.2339914739,-0.2940639555,-0.1222254187,0.3124145567,-0.30023247,0.0982159302,0.2476523668,-0.0327625684,0.0659818202,-0.0001160392,0.1200204343,0.0303151142,-0.1052031666,-0.3548584282,-0.1189137399,-0.2348762155,-0.4712661803,-0.046874024,-0.166179806,0.1921762079,-0.3059154749,0.3483878672,-0.2273127884,0.1708153784,0.1654318273,0.0961554796,0.300424844,-0.3444350064,0.1898005158,0.1419133991,-0.1465310901,-0.0864705145,0.148295477,0.1591269374,0.1215689927,0.1044884771,0.0899096653,0.1305589229,0.1565486789,0.3428242803,0.0436940528,0.2532322705,-0.3091905415,-0.1941089779,0.3868044019,0.5393717289,0.0430028588,-0.4733121693,-0.4100723267,0.0619714856,-0.1059595793,-0.0183863845,0.3538913131,0.0619389564,-0.2496780157,0.2122739404,-0.0840408579,-0.092327401,0.2020340413,0.1703463644,0.4990210831,0.0561957099,0.2428796887,-0.2365749329,-0.1611414701,-0.1720635742,-0.1442683935,-0.184507817,-0.01231483,-0.2962135077,-0.0726034865,-0.0922355503,-0.4936167896,0.276032865,0.048081588,0.3868015409,-0.0181516968,-0.0343011282,0.3849567175,0.577732861,-0.0719930306,0.3023167849,-0.0365592279,-0.0505065285,-0.5055100918,-0.1048416346,-0.1112041101,-0.0039245044,0.0578346066,-0.317707032,-0.1836052537,0.1430572271,0.0773606747,-0.229382202,-0.1281397343,-0.1291473359,-0.3175265789,-0.203420043,-0.0726115182,0.2471601963,0.2654174566,-0.3013916016,0.5123954415,0.0432337299,0.0602396354,-0.0664111152,-0.0779026523,-0.3201749623,0.2367756814,0.0463304892,0.0716184974,0.4474036098,-0.0044772103,-0.0284687411,-0.3521718383,0.1882112473,-0.0790787786,-0.069163084,-0.0300837122,0.0427798964,0.0656022727,-0.2253544927,0.2875692844,-0.4145509303,0.0200992022,-0.3364298642,-0.1666955352,0.0898739174,0.0010788244,-0.1832784414,-0.1988341212,-0.5550371408,0.6362175941,0.0819481984,-0.0798847005,-0.0208134074,-0.1413581073,-0.2756295502,-0.1657590568,-0.3760385513,-0.100439243,-0.2967472374,-0.4861107469,-0.0149357766,-0.16964975,0.0170274973,0.3792363107,-0.2067186981,0.4938260019,-0.0901047587,0.3840901852,0.5942388177,-0.2019268274,-0.0835365579,0.3388235569,0.1239388585,0.0426066257,0.2889140248,0.431168437,0.2352268398,0.0221798625,0.0045318026,0.345169872,-0.1242875457,-0.1570481658,0.0157906059,-0.1439424306,0.4175118804,-0.0080850162,-0.0104979211,-0.2487357259,0.1275279671,-0.0433607511,0.1509204656,0.0679420382,0.1376361698,-0.1213252917,-0.1397398263,0.372022599,-0.0812838599,-0.2937514782,-0.3058458865,0.1738557965,-0.1913663596,0.1177506074,0.5350754261,-0.1869017184,0.097227104,-0.257556349,-0.1604676843,-0.4902059436,0.0914234519,0.4530514479,-0.5109923482,-0.39187482,-0.4492608309,0.1753063947,0.1212878823,-0.1223334149,-0.1563709527,0.2713687122,0.1602035165,0.0259517711,0.124001056,0.395866096,-0.0544895902,-0.0062389779,0.0937291831,0.0853292942,0.2098017186,0.2352549732,0.269169122,-0.1274307072,0.1423422545,-0.0064535015,0.0906715095,-0.1277323216,-0.0298698451,-0.2179368585,0.2128388435,0.2513683438,0.1943630427,0.3742973804,0.0365513936,-0.091596745,0.2419490665,0.0824622884,-0.3555667996,-0.2242487222,0.1361577064,0.0217424035,-0.0685300305,0.1027367339,-0.2026437372,0.2370167673,0.3304443061,0.1169455498,0.1193799973,-0.1372857541,-0.1324388683,-0.1327700317,0.1277343184,0.3263095617,0.4737481475,0.1827244014,0.3626636267,-0.0206293743,-0.0155712636,-0.1086177751,-0.0022715032,0.184429124,0.0015732168,0.213168174,-0.0769162774,-0.1609203964,-0.0590055436,0.1598151475,0.309805423,-0.1948590875,-0.1620259881,0.0063783028,-0.4035834074,-0.4256437719,-0.294324398,0.0231658742,-0.3156166673,-0.2660319507,0.0897278786,-0.0803450644,-0.1905919909,0.3742902577,0.0285845958,0.4266754389,-0.4384796619,0.1727481633,0.2919113636,-0.3651601076,-0.0391863845,0.0752993077,0.4533345997,0.517778337,0.5196759701,0.0819657966,-0.2692104578,-0.1495772302,-0.1766426414,0.1827419251,-0.3511623144,0.6441907883,-0.1898634732,-0.0635559261,0.0654105917,-0.3773079813,0.1839828193,0.3056868017,-0.1688381582,-0.0997271761,-0.0991654173,-0.1515535861,-0.0166971162,-0.5070537329,-0.4871918857,-0.2386375517,0.2469993681,-0.1593508273,0.3552494049,0.0109289121,-0.3600106537,-0.0445909277,-0.0510283485,0.022551354,-0.2522810698,-0.0990662724,-0.0281147789,-0.4457640946,-0.0753629655,0.1458103955,-0.1100309193,0.5756061077,0.0492919944,-0.2368786484,-0.0609803572,0.1762208939,-0.1357149333,0.2101591229,0.0292389244,0.1821368486,-0.1060563549,0.0636286065,-0.0608032905,-0.376483798,0.4122048318,0.4863766134,0.1986308694,-0.0532781854,0.0621095076,-0.2124482989,0.4311757386,0.6173303723,0.2669888139,0.1374833733,0.0152749326,-0.0156508349,0.1343345195,-0.173108995,0.537343204,-0.1520589888,-0.074188821,0.5096476078,0.0771290436,-0.1001929343,-0.0349095538,-0.2184156775,-0.2870620489,-0.2746579051,0.2158991545,-0.3821329772,0.3279493749,-0.1112555712,-0.5277523398,-0.3999684751,0.1223082319,-0.2745324969,0.427723527,0.0146912141,0.1539448351,-0.7118865848,-0.2398282737,-0.2504858375,0.0774448961,0.1625681818,0.408361733,-0.0458819047,0.1651030332,-0.0488121472,0.0752664432,-0.0932260379,-0.4428070486,-0.1155128255,0.1816355884,0.3534649014,-0.2903153896,-0.1967025846,-0.1320837289,-0.0732421651,0.3702421784,0.2047517747,-0.3799410164,-0.0781611353,0.0396044813,0.0743631423,-0.2015172988,-0.2565541565,-0.1636669338,-0.0791399404,-0.1459154338,-0.0708374605,-0.155073598,-0.0993754491,0.1038083807,-0.0523312911,-0.0232256707,-0.1716654897,0.1622259617,0.149948284,-0.0173113123,0.1483186483,0.1659760177,0.035667859,-0.0070451749,0.2020060569,0.6304768324,-0.1841425747,-0.0286972541,0.0473016649,-0.2431912422,0.5375525951,0.0800168291,0.1366087943,0.3033048213,-0.0836982355,-0.3534986377,-0.0829221532,0.1730746627,0.4156101942,0.2072981,-0.0930943936,-0.4728394747,0.3493261039,0.1529119164,-0.1637110263,0.0936644375,-0.2141213864,-0.2796530426,0.1748834848,0.1563460827,0.5925903916,0.170648858,0.160150677,0.019666858,0.1427400708,0.4723138511,-0.0432676375,0.050777439,0.070451811,0.126202181,0.0176399909,-0.10342253,0.1013849229,0.3143951595,-0.1051504016,0.2027142495,0.4885597825,0.073881574,-0.2054989934,0.1652051657,0.022402728,-0.1119206473,0.0245740321,0.0051422003,-0.3381905556,-0.0644522905,-0.1580274105,-0.0954706743,0.0202368945,0.1888591945,-0.2974637151,-0.2856971622,-0.2164370865,-0.0907142609,-0.0883827955,0.0741450787,0.1049054042,-0.1123193428,0.2329506129,-0.2949129641,0.0757784843,-0.061065197,-0.1601297706,-0.0580370873,-0.2413576096,-0.1626470387,0.359668076,-0.1328413486,-0.3288187087,-0.0213939846,0.2124659717,-0.2141181231,-0.2237711251,0.084416315,0.0128096249,0.0242490452,0.0298856888,0.6271650791,-0.1276538968,-0.080985412,0.0115360441,0.081005238,-0.0663654655,0.454793036,-0.2411862761,0.1281153113,-0.0983927548,-0.2685524225,0.395503521,-0.1430410743,0.0615327284,-0.1615624726,0.0560225099,-0.1229566708,0.3657936156,-0.2234287262,0.1669932902,-0.0406630002,0.1672368944,-0.2348572463,-0.0983647928,0.4191531241,0.174447313,0.1659699827,-0.0214052014,-0.3090891838,-0.2814281881,0.4315332174,-0.2380815446,0.2182317674,-0.3361181915,0.4560978711,0.1006002054,0.7375183105,-0.2038106769,-0.05855041,-0.0582380891,0.3075500727,0.6464055777,0.1227310076,0.0144951995,0.1748090088,-0.1364329308,-0.1318371594,-0.2538304031,-0.04798555,-0.2725458145,0.1584842652,0.4529352784,-0.0597900078,0.1185261682,-0.0837063789,0.1968557984,-0.1743214726,0.1540426761,-0.0864887685,-0.1704003513,0.5633848906,0.1775725186,0.1963306814,-0.0934634507,0.4911976755,0.1067792624,0.1471948028,0.1488702297,-0.0069482122,0.0328473859,0.0564046763,-0.1065667719,0.0379991196,0.061036732,0.3045606613,-0.1991023719,0.0441172794,-0.0095202457,-0.073287569,-0.0040003909,0.4577015638,0.1057154313,0.1770026982,0.0548630916,0.0329394713,-0.2006079108,0.13137272,0.2279487103,-0.0367365554,0.1702223867,0.2453300506,0.3747324347,0.0627739653,0.1034654677,-0.1887230277,-0.3689180911,-0.3820825517,0.3947607279,0.0812270939,0.1540253758,0.0747903436,0.3632289469,0.2061947584,0.1635936201,0.2591200769,0.1666923314,0.3571113944,0.4458702803,-0.0454629324,0.4074284732,0.0597958975,-0.0866564959,0.3271298707,0.0100587364,0.1574143767,0.0893088803,0.1816973537,-0.1257469952,-0.3956254721,-0.0027835267,0.1656653583,-0.0827641338,-0.224467665,-0.4635028243,-0.2103762478,-0.1292487234,-0.1795023978,-0.3159166276,-0.2240985632,0.1001911387,0.0159406215,0.0836430863,-0.1032661945,-0.3410227299,0.0406259038,-0.2416045368,-0.0881714821,0.4383792579,-0.0384349078,-0.1242976189,0.3325560689,0.5182185173,-0.008541666,-0.0434249416,-0.094701618,0.0789728314,0.1035142541,0.1742795259,0.0017959324,0.0661077574,0.1495991647,-0.4645207226,0.4282025993,-0.0746208727,0.044543758,0.3763595521,0.0591672957,-0.3788801432,-0.5645083785,0.3745102584,-0.0740086809,-0.1072400361,0.052270256,0.1893500984,-0.4506507516,-0.3078399003,0.3906122446,-0.0274621528,0.1043380126,-0.2164419293,0.0426098034,-0.1539153159,0.5759131908,0.1403675079,0.1214707643,-0.1890482008,0.0632346347,-0.7028479576,0.0908487141,-0.2161919624,-0.0121223535,-0.0317214876,0.0810013041,0.1939312667,0.4688183963,0.1265925169,0.2147818804,-0.1042783782,-0.0566413589,-0.5983037353,-0.1369603723,0.1484322697,0.2320921272,0.073820889,0.0875400975,-0.1203623116,-0.0382536389,-0.1451860815,0.119889915,-0.2839845717,0.3616273701,0.3176860511,0.4177701473,0.025027182,0.6134974957,-0.0264578909,-0.0929829106,-0.1020673141,-0.2681879103,-0.310608387,-0.114380002,-0.0434537344,0.3557246029,-0.5675235391,0.4912192225,0.1228709519,-0.2394486219,-0.2712240517,-0.1315515935,0.002027953,0.2404697984,0.0851934925,-0.1821265817,0.207594499,0.5532332063,0.0460416265,-0.1255834252,-0.0077400887,0.0181448758,0.1712376624,0.0711743832,-0.1821929067,-0.1840198785,0.2387349755,0.0736822411,0.3201595247,-0.4600193202,-0.0938368663,0.4359745383,0.1077983752,0.1193397865,0.1448532045,0.1202413216,-0.4784877002,-0.2822836637,-0.2116905451,-0.0737450495,-0.2070970237,-0.0961488932,-0.3565064073]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"Update: I think I've found a solution.\r\n\r\n```python\r\noutput_types = {\"input_ids\": tf.int64, \"token_type_ids\": tf.int64, \"attention_mask\": tf.int64}\r\ndef train_dataset_gen():\r\n for i in range(len(train_dataset)):\r\n yield train_dataset[i]\r\ntf_dataset = tf.data.Dataset.from_generator(train_dataset_gen, output_types=output_types)\r\n```\r\n\r\nloads WikiText-2 in 20 ms, and WikiText-103 in 20 ms. It appears to be lazily loading via indexing train_dataset.","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":48,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nUpdate: I think I've found a solution.\r\n\r\n```python\r\noutput_types = {\"input_ids\": tf.int64, \"token_type_ids\": tf.int64, \"attention_mask\": tf.int64}\r\ndef train_dataset_gen():\r\n for i in range(len(train_dataset)):\r\n yield train_dataset[i]\r\ntf_dataset = tf.data.Dataset.from_generator(train_dataset_gen, output_types=output_types)\r\n```\r\n\r\nloads WikiText-2 in 20 ms, and WikiText-103 in 20 ms. It appears to be lazily loading via indexing train_dataset.","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"Yes this is the current best solution. We should probably show it in the tutorial notebook.\r\n\r\nNote that this solution unfortunately doesn't allow to train on TPUs (yet). See #193 ","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":30,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nYes this is the current best solution. We should probably show it in the tutorial notebook.\r\n\r\nNote that this solution unfortunately doesn't allow to train on TPUs (yet). See #193 ","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"This approach still seems quite slow. When using TFRecords with a similar training loop, I get ~3.0-3.5 it\/s on multi-node, multi-GPU training. I notice a pretty severe performance regression when scaling, with observed performance numbers. Since the allreduce step takes less than 100ms\/it and I've achieved 80% scaling efficiency up to 64 GPUs, it must be the data pipeline.\r\n\r\n| Nodes | GPUs | Iterations\/Second |\r\n| --- | --- | --- |\r\n| 1 | 2 | 2.01 |\r\n| 1 | 8 | 0.81 |\r\n| 2 | 16 | 0.37 |\r\n\r\nHere are performance metrics over 10k steps. The iteration speed appears to follow some sort of caching pattern. I would love to use `nlp` in my project, but a slowdown from 3.0 it\/s to 0.3 it\/s is too great to stomach.\r\n\r\n<img width=\"1361\" alt=\"Screen Shot 2020-07-02 at 8 29 22 AM\" src=\"https:\/\/user-images.githubusercontent.com\/4564897\/86378156-2f8d3900-bc3e-11ea-918b-c395c3df5377.png\">\r\n","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":146,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nThis approach still seems quite slow. When using TFRecords with a similar training loop, I get ~3.0-3.5 it\/s on multi-node, multi-GPU training. I notice a pretty severe performance regression when scaling, with observed performance numbers. Since the allreduce step takes less than 100ms\/it and I've achieved 80% scaling efficiency up to 64 GPUs, it must be the data pipeline.\r\n\r\n| Nodes | GPUs | Iterations\/Second |\r\n| --- | --- | --- |\r\n| 1 | 2 | 2.01 |\r\n| 1 | 8 | 0.81 |\r\n| 2 | 16 | 0.37 |\r\n\r\nHere are performance metrics over 10k steps. The iteration speed appears to follow some sort of caching pattern. I would love to use `nlp` in my project, but a slowdown from 3.0 it\/s to 0.3 it\/s is too great to stomach.\r\n\r\n<img width=\"1361\" alt=\"Screen Shot 2020-07-02 at 8 29 22 AM\" src=\"https:\/\/user-images.githubusercontent.com\/4564897\/86378156-2f8d3900-bc3e-11ea-918b-c395c3df5377.png\">\r\n","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"An interesting alternative to investigate here would be to use the tf.io library which has some support for Arrow to TF conversion: https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset\r\n\r\nThere are quite a few types supported, including lists so if the unsupported columns are dropped then we could maybe have a zero-copy mapping from Arrow to TensorFlow, including tokenized inputs and 1D tensors like the ones we mostly use in NLP: https:\/\/github.com\/tensorflow\/io\/blob\/322b3170c43ecac5c6af9e39dbd18fd747913e5a\/tensorflow_io\/arrow\/python\/ops\/arrow_dataset_ops.py#L44-L72\r\n\r\nHere is an introduction on Arrow to TF using tf.io: https:\/\/medium.com\/tensorflow\/tensorflow-with-apache-arrow-datasets-cdbcfe80a59f","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":77,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nAn interesting alternative to investigate here would be to use the tf.io library which has some support for Arrow to TF conversion: https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset\r\n\r\nThere are quite a few types supported, including lists so if the unsupported columns are dropped then we could maybe have a zero-copy mapping from Arrow to TensorFlow, including tokenized inputs and 1D tensors like the ones we mostly use in NLP: https:\/\/github.com\/tensorflow\/io\/blob\/322b3170c43ecac5c6af9e39dbd18fd747913e5a\/tensorflow_io\/arrow\/python\/ops\/arrow_dataset_ops.py#L44-L72\r\n\r\nHere is an introduction on Arrow to TF using tf.io: https:\/\/medium.com\/tensorflow\/tensorflow-with-apache-arrow-datasets-cdbcfe80a59f","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"Interesting. There's no support for strings, but it does enable int and floats so that would work for tokenized inputs. \r\n\r\nArrowStreamDataset requires loading from a \"record batch iterator\", which can be instantiated from in-memory arrays as described here: https:\/\/arrow.apache.org\/docs\/python\/ipc.html. \r\n\r\nBut the nlp.Dataset stores its data as a `pyarrow.lib.Table`, and the underlying features are `pyarrow.lib.ChunkedArray`. I can't find any documentation about lazily creating a record batch iterator from a ChunkedArray or a Table. Have you had any success?\r\n\r\nI can't find [any uses](https:\/\/grep.app\/search?q=ArrowDataset&filter[lang][0]=Python) of tfio.arrow.ArrowDataset on GitHub.","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":86,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nInteresting. There's no support for strings, but it does enable int and floats so that would work for tokenized inputs. \r\n\r\nArrowStreamDataset requires loading from a \"record batch iterator\", which can be instantiated from in-memory arrays as described here: https:\/\/arrow.apache.org\/docs\/python\/ipc.html. \r\n\r\nBut the nlp.Dataset stores its data as a `pyarrow.lib.Table`, and the underlying features are `pyarrow.lib.ChunkedArray`. I can't find any documentation about lazily creating a record batch iterator from a ChunkedArray or a Table. Have you had any success?\r\n\r\nI can't find [any uses](https:\/\/grep.app\/search?q=ArrowDataset&filter[lang][0]=Python) of tfio.arrow.ArrowDataset on GitHub.","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"You can use `to_batches` maybe?\r\nhttps:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.Table.html#pyarrow.Table.to_batches","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":6,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nYou can use `to_batches` maybe?\r\nhttps:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.Table.html#pyarrow.Table.to_batches","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"Also note that since #322 it is now possible to do\r\n```python\r\nids = [1, 10, 42, 100]\r\nbatch = dataset[ids]\r\n```\r\nFrom my experience it is quite fast but it can take lots of memory for large batches (haven't played that much with it).\r\nLet me know if you think there could be a better way to implement it. (current code is [here](https:\/\/github.com\/huggingface\/nlp\/blob\/78628649962671b4aaa31a6b24e7275533416845\/src\/nlp\/arrow_dataset.py#L463))","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":64,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nAlso note that since #322 it is now possible to do\r\n```python\r\nids = [1, 10, 42, 100]\r\nbatch = dataset[ids]\r\n```\r\nFrom my experience it is quite fast but it can take lots of memory for large batches (haven't played that much with it).\r\nLet me know if you think there could be a better way to implement it. (current code is [here](https:\/\/github.com\/huggingface\/nlp\/blob\/78628649962671b4aaa31a6b24e7275533416845\/src\/nlp\/arrow_dataset.py#L463))","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"Thanks @lhoestq! That format is much better to work with.\r\n\r\nI put together a benchmarking script. This doesn't measure the CPU-to-GPU efficiency, nor how it scales with multi-GPU multi-node training where many processes are making the same demands on the same dataset. But it does show some interesting results:\r\n\r\n```python\r\nimport nlp\r\nimport numpy as np\r\nimport tensorflow as tf\r\nimport time\r\n\r\ndset = nlp.load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", split=\"train\")\r\ndset = dset.filter(lambda ex: len(ex[\"text\"]) > 0)\r\nbsz = 1024\r\nn_batches = 100\r\n\r\ndef single_item_gen():\r\n for i in range(len(dset)):\r\n yield dset[i]\r\n\r\ndef sequential_batch_gen():\r\n for i in range(0, len(dset), bsz):\r\n yield dset[i:i+bsz]\r\n\r\ndef random_batch_gen():\r\n for i in range(len(dset)):\r\n indices = list(np.random.randint(len(dset), size=(bsz,)))\r\n yield dset[indices]\r\n\r\noutput_types = {\"text\": tf.string}\r\nsingle_item = tf.data.Dataset.from_generator(single_item_gen, output_types=output_types).batch(bsz)\r\ninterleaved = tf.data.Dataset.range(10).interleave(\r\n lambda idx: tf.data.Dataset.from_generator(single_item_gen, output_types=output_types),\r\n cycle_length=10,\r\n)\r\nsequential_batch = tf.data.Dataset.from_generator(sequential_batch_gen, output_types=output_types)\r\nrandom_batch = tf.data.Dataset.from_generator(random_batch_gen, output_types=output_types)\r\n\r\ndef iterate(tf_dset):\r\n start = time.perf_counter()\r\n for i, batch in enumerate(tf_dset.take(n_batches)):\r\n pass\r\n elapsed = time.perf_counter() - start\r\n print(f\"{tf_dset} took {elapsed:.3f} secs\")\r\n\r\niterate(single_item)\r\niterate(interleaved)\r\niterate(sequential_batch)\r\niterate(random_batch)\r\n```\r\n\r\nResults:\r\n```\r\n<BatchDataset shapes: {text: <unknown>}, types: {text: tf.string}> took 23.005 secs\r\n<InterleaveDataset shapes: {text: <unknown>}, types: {text: tf.string}> took 0.135 secs\r\n<FlatMapDataset shapes: {text: <unknown>}, types: {text: tf.string}> took 0.074 secs\r\n<FlatMapDataset shapes: {text: <unknown>}, types: {text: tf.string}> took 0.550 secs\r\n```\r\n\r\n- Batching a generator which fetches a single item is terrible.\r\n- Interleaving performs well on a single process, but doesn't scale well to multi-GPU training. I believe the bottleneck here is in Arrow dataset locking or something similar. The numbers from the table above are with interleaving.\r\n- The sequential access dominates the random access (7x faster). Is there any way to bring random access times closer to sequential access? Maybe re-indexing the dataset after shuffling each pass over the data.","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":285,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nThanks @lhoestq! That format is much better to work with.\r\n\r\nI put together a benchmarking script. This doesn't measure the CPU-to-GPU efficiency, nor how it scales with multi-GPU multi-node training where many processes are making the same demands on the same dataset. But it does show some interesting results:\r\n\r\n```python\r\nimport nlp\r\nimport numpy as np\r\nimport tensorflow as tf\r\nimport time\r\n\r\ndset = nlp.load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", split=\"train\")\r\ndset = dset.filter(lambda ex: len(ex[\"text\"]) > 0)\r\nbsz = 1024\r\nn_batches = 100\r\n\r\ndef single_item_gen():\r\n for i in range(len(dset)):\r\n yield dset[i]\r\n\r\ndef sequential_batch_gen():\r\n for i in range(0, len(dset), bsz):\r\n yield dset[i:i+bsz]\r\n\r\ndef random_batch_gen():\r\n for i in range(len(dset)):\r\n indices = list(np.random.randint(len(dset), size=(bsz,)))\r\n yield dset[indices]\r\n\r\noutput_types = {\"text\": tf.string}\r\nsingle_item = tf.data.Dataset.from_generator(single_item_gen, output_types=output_types).batch(bsz)\r\ninterleaved = tf.data.Dataset.range(10).interleave(\r\n lambda idx: tf.data.Dataset.from_generator(single_item_gen, output_types=output_types),\r\n cycle_length=10,\r\n)\r\nsequential_batch = tf.data.Dataset.from_generator(sequential_batch_gen, output_types=output_types)\r\nrandom_batch = tf.data.Dataset.from_generator(random_batch_gen, output_types=output_types)\r\n\r\ndef iterate(tf_dset):\r\n start = time.perf_counter()\r\n for i, batch in enumerate(tf_dset.take(n_batches)):\r\n pass\r\n elapsed = time.perf_counter() - start\r\n print(f\"{tf_dset} took {elapsed:.3f} secs\")\r\n\r\niterate(single_item)\r\niterate(interleaved)\r\niterate(sequential_batch)\r\niterate(random_batch)\r\n```\r\n\r\nResults:\r\n```\r\n<BatchDataset shapes: {text: <unknown>}, types: {text: tf.string}> took 23.005 secs\r\n<InterleaveDataset shapes: {text: <unknown>}, types: {text: tf.string}> took 0.135 secs\r\n<FlatMapDataset shapes: {text: <unknown>}, types: {text: tf.string}> took 0.074 secs\r\n<FlatMapDataset shapes: {text: <unknown>}, types: {text: tf.string}> took 0.550 secs\r\n```\r\n\r\n- Batching a generator which fetches a single item is terrible.\r\n- Interleaving performs well on a single process, but doesn't scale well to multi-GPU training. I believe the bottleneck here is in Arrow dataset locking or something similar. The numbers from the table above are with interleaving.\r\n- The sequential access dominates the random access (7x faster). Is there any way to bring random access times closer to sequential access? Maybe re-indexing the dataset after shuffling each pass over the data.","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"Hey @jarednielsen \r\n\r\nThanks for this very interesting analysis!! IMHO to read text data one should use `tf.data.TextLineDataset`. It would be interesting to compare what you have done with simply load with a `TextLineDataset` and see if there is a difference.\r\n\r\nA good example can be found here https:\/\/www.tensorflow.org\/tutorials\/load_data\/text","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":48,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nHey @jarednielsen \r\n\r\nThanks for this very interesting analysis!! IMHO to read text data one should use `tf.data.TextLineDataset`. It would be interesting to compare what you have done with simply load with a `TextLineDataset` and see if there is a difference.\r\n\r\nA good example can be found here https:\/\/www.tensorflow.org\/tutorials\/load_data\/text","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"Thanks! I'm not actually loading in raw text data, that was just the synthetic data I created for this benchmark. A more realistic use case would be a dataset of tokenized examples, which would be a dict of lists of integers. TensorFlow's TextLineDataset greedily loads the dataset into the graph itself, which can lead to out-of-memory errors - one of the main reason I'm so drawn to the `nlp` library is its zero-copy no-RAM approach to dataset loading and mapping. \r\n\r\nIt's quite helpful for running a preprocessing pipeline - a sample ELECTRA pipeline I've built is here: https:\/\/github.com\/jarednielsen\/deep-learning-models\/blob\/nlp\/models\/nlp\/common\/preprocess.py.","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":98,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nThanks! I'm not actually loading in raw text data, that was just the synthetic data I created for this benchmark. A more realistic use case would be a dataset of tokenized examples, which would be a dict of lists of integers. TensorFlow's TextLineDataset greedily loads the dataset into the graph itself, which can lead to out-of-memory errors - one of the main reason I'm so drawn to the `nlp` library is its zero-copy no-RAM approach to dataset loading and mapping. \r\n\r\nIt's quite helpful for running a preprocessing pipeline - a sample ELECTRA pipeline I've built is here: https:\/\/github.com\/jarednielsen\/deep-learning-models\/blob\/nlp\/models\/nlp\/common\/preprocess.py.","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/315","title":"[Question] Best way to batch a large dataset?","comments":"Sorry, I think I badly expressed myself, my bad. What I suggested is to compare with the usual loading textual data in pure TF with `TextLineDataset` with `nlp`. I know it is not recommended with very large datasets to use it, but I was curious to see how it behaves compared to a processing with `nlp` on smaller datasets.\r\n\r\nBTW your script looks very interesting, thanks for sharing!!","body":"I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?","comment_length":68,"text":"[Question] Best way to batch a large dataset?\nI'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb), I see the following recommended for TensorFlow:\r\n\r\n```python\r\ntrain_tf_dataset = train_tf_dataset.filter(remove_none_values, load_from_cache_file=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x].to_tensor(default_value=0, shape=[None, tokenizer.max_len]) for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"].to_tensor(default_value=0, shape=[None, 1])}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"].to_tensor(default_value=0, shape=[None, 1])\r\n### Question about this last line ###\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nThis code works for something like WikiText-2. However, scaling up to WikiText-103, the last line takes 5-10 minutes to run. I assume it is because tf.data.Dataset.from_tensor_slices() is pulling everything into memory, not lazily loading. This approach won't scale up to datasets 25x larger such as Wikipedia.\r\n\r\nSo I tried manual batching using `dataset.select()`:\r\n\r\n```python\r\nidxs = np.random.randint(len(dataset), size=bsz)\r\nbatch = dataset.select(idxs).map(lambda example: {\"input_ids\": tokenizer(example[\"text\"])})\r\ntf_batch = tf.constant(batch[\"ids\"], dtype=tf.int64)\r\n```\r\n\r\nThis appears to create a new Apache Arrow dataset with every batch I grab, and then tries to cache it. The runtime of `dataset.select([0, 1])` appears to be much worse than `dataset[:2]`. So using `select()` doesn't seem to be performant enough for a training loop.\r\n\r\nIs there a performant scalable way to lazily load batches of nlp Datasets?\nSorry, I think I badly expressed myself, my bad. What I suggested is to compare with the usual loading textual data in pure TF with `TextLineDataset` with `nlp`. I know it is not recommended with very large datasets to use it, but I was curious to see how it behaves compared to a processing with `nlp` on smaller datasets.\r\n\r\nBTW your script looks very interesting, thanks for sharing!!","embeddings":[-0.3098398447,-0.3549019396,0.040070232,-0.1314413697,0.043479424,0.1706920415,0.5454062223,0.3945962787,0.1853833199,0.0758509487,-0.0109290816,0.0905708447,-0.0458247811,0.1636851728,0.1464808881,-0.3626331091,0.0164577533,-0.2653392255,-0.0260757729,-0.1447885334,-0.2456075102,-0.2963108122,-0.3499272764,-0.2272665501,-0.3148984015,-0.2920510471,0.0221025404,-0.1786595732,0.2146256864,-0.0490131006,0.0490412712,0.0216406938,0.300095439,0.4448573887,-0.0001156601,-0.1326184422,0.0600120351,-0.0445338674,-0.0088806693,0.3237411678,-0.0612367652,-0.388153106,-0.0979186371,-0.4061881304,0.1105616391,0.1923644096,0.1939423829,-0.0472506657,0.1358859092,-0.0325134471,0.063217558,0.3727622628,-0.0247660466,0.1722879857,0.02888369,-0.0639728978,0.0118579278,0.0609920211,0.2795847654,-0.0697013512,-0.0007684377,0.3284053802,-0.1674046367,-0.0159423053,0.2925575972,0.2313953638,0.1201131791,-0.6053037047,0.031976074,0.3724275529,0.2989349961,0.003854932,-0.2751540542,-0.6016538739,-0.1216737404,-0.2600981593,-0.2543848455,0.2779847682,-0.2282436937,-0.1784859449,-0.2848558724,-0.0607504658,-0.111662358,0.1517273933,-0.17424573,0.1897965819,0.1561438292,0.2513833046,0.1800088286,-0.259157747,0.0723085105,-0.2944973111,0.391967684,0.2885423005,-0.5199069381,-0.161050126,0.1525548995,-0.2357104123,0.0620710067,-0.2695543766,-0.0892135128,0.0071072425,0.0950781703,0.046422746,0.1840151697,0.2864991128,-0.1267043054,0.0481852777,0.1040561199,-0.2625125349,-0.2421223223,0.3590080142,-0.108374469,-0.2902850807,0.0475210175,-0.1952093393,-0.358223021,0.1385153085,-0.0088153258,0.0312358961,0.0680118501,-0.0676135719,-0.031125294,0.1281121075,-0.2842437327,0.0787763745,-0.1488966644,-0.1395684034,-0.3805860281,0.1666765809,-0.1485890448,0.0174979772,-0.0054796408,-0.0163144097,0.1355892569,-0.139675349,0.1104079336,-0.1817804277,0.0023557388,0.2069333345,-0.1126759872,-0.2320376039,0.1354833096,0.2346876562,0.0579938069,0.2410531789,0.0312836021,0.1558558941,-0.2094338685,0.3726697862,-0.3099021912,-0.3395798504,0.1943150908,0.0668445751,0.0166570023,-0.3149917126,-0.5945531726,0.3981288075,0.321008116,0.0930782035,-0.0770611241,-0.2863120437,-0.1548867822,-0.2435482889,0.1875332594,0.2121235579,-0.4263499677,0.1576705426,0.4017448723,0.3291971684,0.3743069172,0.6147413254,-0.2548478842,0.4410940707,0.175022468,0.1125394627,0.1075415313,-0.25900051,-0.2935233712,0.4176078737,-0.2239006311,0.0911498666,0.0280616544,0.2754259109,0.290940702,-0.235840559,0.4677284658,0.5002925396,-0.1060514823,0.3253593147,-0.2121745497,-0.2624133229,0.1294199079,0.1681138873,0.1687644422,-0.1044366434,-0.2262644023,0.2172021866,0.2736003399,-0.1601257324,-0.1335326135,0.0144466078,0.0447470471,0.0450545736,-0.2553613484,0.0899209753,-0.5541221499,0.2491139024,0.104594931,-0.0720571429,0.2975907326,-0.3017668426,0.2677598596,-0.0850059688,0.0262452774,0.2439461052,-0.0638855994,-0.1151607782,0.2073930055,-0.0009455864,-0.2045055181,-0.0694354549,-0.45423159,0.0335031152,-0.3397281468,-0.1507048756,0.3422574997,0.1017849892,-0.1304057837,0.1610793471,-0.0711504072,-0.1975040138,0.0639175996,0.1212921143,-0.0829551965,-0.0109530622,-0.2208668292,0.3150741458,0.2363253683,0.3301177025,0.0516277552,-0.1459756196,0.0075899232,-0.3100179136,-0.2521992922,0.2785663307,-0.0317873061,0.3897755146,0.1379317939,-0.3446787596,-0.1478319615,0.1022378057,0.0575494617,0.1419411749,0.1398730576,0.1266010553,0.0635658801,0.1517291218,-0.4028158188,0.2708552778,0.5878329873,-0.1192695946,-0.069387868,0.2395924926,0.059559077,-0.1695690155,-0.0338157713,-0.217757225,0.1716604531,0.1619177461,0.1790524721,-0.0999094173,0.124293372,-0.1298576891,0.3223983645,0.2120993882,0.004855027,0.022191342,0.116418615,0.0722042248,0.0016253602,-0.0955381989,0.1305938661,0.3099609613,-0.1252200752,0.0044579892,0.1954187155,-0.1072649211,-0.1237375364,-0.0269091949,0.0157357194,-0.3178576529,-0.1369980425,0.248922646,-0.1166380346,0.0364544466,0.1061013117,0.3592701256,0.1271025687,-0.364212513,0.1010778546,-0.1170879453,-0.2541911304,-0.0160684511,0.1550412625,-0.0097691929,0.1884468198,0.0886323899,-0.1106292754,0.015857242,-0.1376692057,-0.0592114329,-0.4576979876,0.0734886229,-0.1006338224,0.335616827,-0.2249776274,-0.437174648,-0.0026174795,-0.0687950104,0.2533338964,-0.2674293518,0.0078697056,-0.0599730387,0.1682515144,-0.2091825753,-0.3979697526,-0.2773379683,0.1616996378,-0.0804175064,0.1120989919,0.3758215606,0.0861320272,0.0614354759,0.1984479427,-0.1117398441,0.0645064414,-0.3359968662,0.2933339775,-0.1133206114,-0.2322174907,-0.2972748578,-0.0569187999,-0.0541822575,0.3564843833,-0.5518575907,0.1781203598,-0.3643219471,0.1989420503,-0.005830613,0.5625964403,-0.0943654105,-0.4863793552,-0.0153563311,0.1136103272,0.1111946926,0.0132406522,0.2653706968,0.2140454203,0.2785107493,0.4317783415,0.133768037,1.1682617664,0.1721532643,-0.2398597002,-0.2142921984,-0.0006967018,-0.0267474819,-0.3446087837,-0.2672060728,0.3804666996,-0.0651110634,-0.1218410656,0.2086466253,-0.1688445508,-0.2933580875,-0.0375190042,0.1244465709,-0.1118379459,-0.2858682573,0.4226363897,0.0917632133,0.2491028309,0.1349134147,0.0886393115,-0.4993352592,-0.3593015075,-0.0525136217,-0.2120135576,0.4119946659,-0.1316411048,-0.2977095544,-0.1367171705,-0.9755147696,0.2176832259,0.0309694838,0.2673722506,0.0691984817,-0.3508167565,0.0928070396,0.2425126731,0.8538385034,-0.0693197176,-0.1168282256,-0.0046312055,-0.3046589792,-0.4152691066,0.036352776,-0.0984104499,0.168717429,0.3588329852,0.0521052442,-0.3808854818,-0.1210648865,0.2053185999,0.2232019454,-0.0370706171,-0.062758103,-0.0237995256,-0.1611630917,0.0037862598,0.1920000911,0.1095616147,0.1918833107,-0.1529378593,0.0767399371,-0.1166860983,0.3052934706,0.1655394286,0.1245308071,-0.06460049,0.684384644,0.2821177244,-0.004588732,0.0273228213,0.2122226059,0.0313410088,-0.2250824869,-0.0758854523,0.0017626418,0.1497220993,0.3183744252,0.4158222377,0.0518370084,-0.1018585637,-0.0244697351,0.4759161174,-0.0445717685,0.2652131915,0.225712657,0.1480261534,-0.4993152618,-0.5580793619,0.6823279858,0.1741636842,0.164427191,0.3666120172,-0.1283598095,-0.2107718885,0.3020576835,0.0786754712,0.6478743553,-0.1737246513,0.4172235131,0.314653635,0.8153584003,0.6308771968,-0.0986908004,0.2893836796,-0.1265120655,-0.1174631044,0.0805878639,-0.0940856189,-0.2830126584,0.2102293819,0.119722046,0.2607629001,-0.0577676408,0.2022132576,-0.2976982892,0.2678610682,0.1215580106,-0.488828063,-0.1490101218,0.0078631928,-0.0250545572,0.1897145808,-0.0738680214,0.1231594533,0.0115910126,0.0521978065,-0.0586847477,-0.1983020008,-0.2293966264,0.2710294127,-0.103302516,-0.3641133308,-0.0581126064,0.2246191353,0.0554123856,0.1029742807,0.0519132838,-0.1379094422,0.074050203,0.0670989305,0.2410013378,-0.272134155,0.404617846,0.0719942003,0.0943093076,0.0911531672,-0.0158273242,-0.266985029,-0.3697374165,0.1058784425,0.1580176353,-0.0957546085,-0.5968505144,0.2684338987,-0.2813648283,0.0503226556,0.0037910382,-0.0731907263,-0.0849763006,0.7211557031,-0.2573437095,-0.1843225509,0.0496472456,0.1847871393,0.124970451,-0.2062485069,0.3921863139,0.3405942917,-0.1757613122,0.0322537906,0.2664454877,0.179825455,-0.0468793623,0.1488729566,0.0724646896,0.3175610602,0.0450942479,0.4393854439,-0.1473887861,-0.1433921307,-0.1848055571,0.0507971905,-0.3096582592,0.2169217169,0.1730131805,0.2440240979,0.0664235279,0.2008535713,0.0833128765,0.085666053,-0.1693710238,0.0468313769,0.0415404253,0.232238397,-0.0782681331,0.0980027318,0.1146478504,0.1879298985,-0.0071328366,0.1172509938,-0.0932628587,-0.2555660605,0.012203333,0.149570778,0.1200918928,-0.2321113795,0.1148166135,-0.2494661212,-0.2574733198,-0.3281428814,0.5115011334,0.1916114092,-0.01690658,-0.2077613175,0.0609314628,0.136243552,0.0935205743,0.2389405519,0.0112467678,0.1360512823,0.2009719014,0.1953348219,0.0930073187,-0.0966623947,-0.2422027886,-0.0926665366,0.2189396471,0.0491773002,-0.0516622663,-0.2504856586,-0.3024640977,0.0773895681,0.1929149926,0.3313835561,-0.1175881401,-0.2975839078,0.232012108,0.0940037966,0.0618037544,-0.0243890118,0.1837354898,-0.0688407719,0.1389096975,0.3248267472,-0.0041171955,0.2021241486,-0.3089779615,-0.0432885997,0.3552232683,-0.1468783766,0.0196461324,0.2772547603,0.0090765134,-0.070874624,0.0298670363,0.2787351012,0.1938203573,0.1448668092,0.3068629801,0.300632,-0.0135255437,-0.0194718204,-0.1182344705,-0.4309397638,-0.2160276771,0.2103411555,-0.232204318,0.4129007459,-0.0953074396,-0.0314271227,-0.1800400466,-0.2994455695,-0.0706428513,0.2695058286,-0.1646300703,-0.1954697222,-0.2913822234,0.0539110601,0.346334517,0.3158673942,-0.1297771931,-0.3905594349,-0.0089005595,0.1537624151,0.0923350006,-0.0572748296,0.0585334226,-0.1711357832,0.4078229666,0.0008700194,0.0353841223,-0.2385188639,-0.104023546,-0.3525911272,0.1843398809,0.2311054766,0.248517856,-0.1820403934,0.2531327307,-0.0024807588,0.1195366383,-0.099759534,0.1101076007,-0.0316892602,-0.0866467729,0.350288868,-0.0026457885,-0.0533601567,-0.0128743276,0.1048911735,-0.1012807414,-0.2076451033,-0.109804906,0.2251523882,-0.1399892271,-0.1239384934,-0.0363590792,-0.3962470293,-0.1025206223,0.4084668756,-0.3164224029,0.0528209917,0.0610715859,0.0466571115,-0.5405774713,0.7912288904,0.2515087724,0.3168759346,-0.4403222501,0.1956537068,-0.2404488176,-0.2135183364,-0.258079052,-0.0204233248,0.0934651047,0.3934063315,0.0163257737,0.2544080317,-0.3146833479,-0.2966714501,-0.3043270707,0.2985673547,-0.1539195776,-0.0309771746,-0.5753907561,0.2206028104,0.0312658623,-0.3881157339,0.3129305542,-0.1637300402,-0.1229502186,-0.2661139965,0.1841339022,0.357529819,0.2094400972,0.2181301266,-0.0055505666,0.1824417412,0.1804746836,0.0365026146,-0.1549041122,0.193511501,-0.2196915001,0.1634927243,-0.006641068,0.3492395282,-0.0434845723,-0.2172168195,0.0128389066,0.0358813778,0.2151203305,-0.1570638418,0.198191151,-0.2509370148,0.0548378043,0.0537038632,0.4235469997,0.6069497466,-0.2101473957,0.1147366166,-0.3305140138,-0.2574831545,0.4739016891,-0.7701929808,-0.3619873524,-0.3649832606,0.0039873808,0.0836719796,0.2129297256,-0.1846251786,0.0348593108,0.2442564368,-0.1608198136,-0.3659533858,0.2732785046,-0.1406791955,-0.0961095169,-0.0890373513,0.1838718057,-0.1871087849,-0.3062251508,-0.3018523157,-0.3897609413]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/312","title":"[Feature request] Add `shard()` method to dataset","comments":"Hi Jared,\r\nInteresting, thanks for raising this question. You can also do that after loading with `dataset.select()` or `dataset.filter()` which let you keep only a specific subset of rows in a dataset.\r\nWhat is your use-case for sharding?","body":"Currently, to shard a dataset into 10 pieces on different ranks, you can run\r\n\r\n```python\r\nrank = 3 # for example\r\nsize = 10\r\ndataset = nlp.load_dataset('wikitext', 'wikitext-2-raw-v1', split=f\"train[{rank*10}%:{(rank+1)*10}%]\")\r\n```\r\n\r\nHowever, this breaks down if you have a number of ranks that doesn't divide cleanly into 100, such as 64 ranks. Is there interest in adding a method shard() that looks like this?\r\n\r\n```python\r\nrank = 3\r\nsize = 64\r\ndataset = nlp.load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", split=\"train\").shard(rank=rank, size=size)\r\n```\r\n\r\nTensorFlow has a similar API: https:\/\/www.tensorflow.org\/api_docs\/python\/tf\/data\/Dataset#shard. I'd be happy to contribute this code.","comment_length":38,"text":"[Feature request] Add `shard()` method to dataset\nCurrently, to shard a dataset into 10 pieces on different ranks, you can run\r\n\r\n```python\r\nrank = 3 # for example\r\nsize = 10\r\ndataset = nlp.load_dataset('wikitext', 'wikitext-2-raw-v1', split=f\"train[{rank*10}%:{(rank+1)*10}%]\")\r\n```\r\n\r\nHowever, this breaks down if you have a number of ranks that doesn't divide cleanly into 100, such as 64 ranks. Is there interest in adding a method shard() that looks like this?\r\n\r\n```python\r\nrank = 3\r\nsize = 64\r\ndataset = nlp.load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", split=\"train\").shard(rank=rank, size=size)\r\n```\r\n\r\nTensorFlow has a similar API: https:\/\/www.tensorflow.org\/api_docs\/python\/tf\/data\/Dataset#shard. I'd be happy to contribute this code.\nHi Jared,\r\nInteresting, thanks for raising this question. You can also do that after loading with `dataset.select()` or `dataset.filter()` which let you keep only a specific subset of rows in a dataset.\r\nWhat is your use-case for sharding?","embeddings":[-0.0902190059,0.0335450843,-0.1385268271,-0.0104933362,-0.1253202856,-0.0391625874,0.4583161175,0.4131177068,-0.1107414141,0.1866177917,0.1924446821,0.2701655626,-0.2045647353,0.3323414326,0.5994433761,-0.4415318668,-0.1018194929,0.0852711573,0.149879232,0.1868807077,-0.0623615421,-0.0934392959,-0.0521033406,-0.3756529093,-0.0738371238,0.0836312249,-0.106206283,-0.2192081213,0.2495964617,-0.2033970505,0.1981488764,0.0128354235,0.176435113,0.1089037955,-0.0001193276,-0.1619028896,0.1732525229,0.1418514401,-0.0407615453,-0.1341593862,-0.4363878965,-0.3930357397,-0.0248603262,-0.2685240805,-0.0763259456,0.4600835145,0.0376816317,0.0504966155,0.4158009291,0.0452725515,0.0574582145,0.0144584598,0.0706741959,0.2210827023,0.4204316735,0.1398421079,-0.2779010236,0.0068351403,-0.0258939732,0.2300103903,0.1723851413,0.1819727123,-0.0076822857,-0.1983823329,0.151426211,0.1420772225,-0.2785290182,-0.1684783101,-0.0783310905,0.4818460345,-0.2934041917,-0.0841815248,-0.3170888126,-0.4590073228,0.2600898445,-0.7193084359,-0.3265034556,0.2590245306,-0.2154867947,-0.1617243439,-0.1539446115,-0.2942079008,-0.2469418496,0.0869933292,0.1184398085,0.2054739743,0.1094351187,-0.0066468925,0.2596993744,-0.0784598142,-0.1302294582,0.0487918742,0.2060503513,0.1232482344,-0.6726326346,-0.349616617,0.151286304,0.3256623149,0.2887284458,0.1294467002,0.0897560567,0.3228828013,0.1871463805,0.0715471655,0.0216732733,-0.1212604046,0.2701044381,-0.0877677575,-0.0579494201,-0.2449847311,0.2663830519,0.0972311795,0.23987858,-0.2317632437,-0.0781483501,-0.1192791685,-0.1509444565,0.0851148069,0.0345103033,-0.2624329627,0.3014739454,-0.2090816349,0.2798621356,0.0331917107,0.2475590408,-0.2252887189,-0.3614956439,0.0553383306,-0.1140266731,-0.2651718855,-0.0561420918,0.0309462454,-0.1502551585,0.1724989712,0.0432409197,-0.280846566,-0.2295240164,-0.0929453671,0.2970185876,0.1192361861,0.0993343368,0.1004474238,0.119814001,0.181933105,-0.3314205408,-0.4015157521,-0.0938133523,-0.0558922775,-0.3344355822,0.0863510668,-0.484593153,-0.359936595,-0.1270346045,0.0499380194,0.0325772315,-0.1788466722,-0.6782975793,0.5066899061,0.1495945156,-0.1023567095,0.0311424267,-0.1672639102,-0.700308919,-0.2102397978,0.3099248111,-0.0475842394,-0.1396569312,-0.1577159166,-0.7277149558,0.1584021002,0.1372088045,0.0691284612,-0.0961400345,-0.0049770996,0.116054289,0.7563393116,0.9136978388,0.1492827684,-0.363624841,0.2881655395,-0.0713724792,-0.2263800502,0.101901032,0.2832311094,0.312705338,-0.0604075417,0.1139606759,0.3264332712,-0.2208423764,-0.1373609304,0.0063825226,-0.1372893304,0.2847053111,0.4371267557,0.1175920293,-0.0439295135,0.425958097,-0.2272982746,0.2227542251,-0.1066037789,-0.0137880147,-0.3090628088,0.1172641441,0.072384499,-0.0667463318,-0.5428250432,-0.0849225223,0.0770921186,0.0037511138,-0.0932952315,-0.0223909728,-0.3046743572,0.0948518887,-0.0015547101,0.116472654,0.1285893917,-0.0430854522,-0.258610785,0.0833002478,-0.1595627666,-0.3628672659,-0.4273320735,-0.4661524594,-0.0019465975,-0.3521492779,0.4206297994,0.3815576434,-0.2036995143,-0.0961357579,0.4029873013,0.0117044104,-0.2372408956,0.2422447354,0.2740434706,0.0853389353,-0.1135813519,0.0520370677,0.3304307759,0.1716755778,0.1707373261,-0.2814111412,-0.0287463311,-0.2243514061,-0.0752126351,-0.0451592542,0.4841749668,-0.4256809652,0.4008709788,-0.106112577,0.006309777,-0.0138321863,-0.0932262912,-0.1130578965,-0.1825072616,0.0705758408,0.200553298,-0.1365466565,-0.2851176262,0.0616345555,0.1975689977,0.5304974318,-0.2233927399,0.2951188385,0.3810320199,0.3891986609,0.108373493,0.1387320608,0.4842126966,0.4051701725,0.1774630696,0.0463757627,0.2085753977,0.0684135258,-0.2003530115,0.0488263071,0.2136185914,-0.2074343562,0.096232906,0.0300726052,0.0040939962,-0.1740270704,-0.1103289127,-0.105906114,0.0203829631,-0.0459676273,-0.1507743597,-0.2782159746,0.1470051408,-0.1693458706,0.198210597,-0.253113389,0.0215052031,0.0074579101,-0.1364809722,-0.1525467187,0.0969379246,0.151262939,0.4196157455,-0.1436355561,0.0009898827,0.0976945609,-0.0651281178,0.2122161686,-0.0127151748,0.1586858481,0.0745873824,0.5542141199,-0.0527701527,-0.1078748181,-0.1906172335,-0.4497658014,-0.0322911516,-0.2761199772,0.2822998464,0.1809542924,0.0685620159,-0.3565734923,-0.1503163427,-0.0005949574,0.0145139135,0.0318046696,0.1161490828,0.1840159297,0.3239157498,-0.0685231239,0.0758896768,-0.1115966812,-0.1751188487,0.3468532562,0.1412599087,0.1919829696,-0.2804489732,-0.0493816845,0.1952583939,-0.0353983268,-0.1059934273,0.1141613871,-0.3425091803,0.2413946092,-0.3241923749,-0.1880903989,-0.1552043557,-0.0892999992,-0.0342253931,0.6443938613,-0.1353792548,-0.0285966881,-0.0249955337,0.1881885082,-0.0791523978,0.014095949,0.4491424859,-0.1632229388,0.1547424346,0.214763552,0.1615513563,-0.026202539,0.0258960873,0.0613388903,0.3198076785,0.7087903619,0.3468720019,0.8575080633,0.0832482651,-0.0617544651,-0.1029329002,-0.1273305267,-0.1556652337,-0.0813511163,-0.1452895105,0.2121749222,0.0523634851,0.1028443426,0.2388823032,0.0831348151,-0.1338017136,0.0735519901,-0.5296788812,0.0641091019,-0.3044583499,0.2565576434,-0.1438810676,0.4951171875,-0.254214853,-0.1213730574,-0.0095497407,-0.3806798458,0.0099296141,0.1526161879,0.247575596,-0.2041944563,-0.3471156061,-0.3430949152,-0.2364535481,0.2977704108,0.4020830393,-0.0270211603,-0.047871206,-0.035225302,-0.0881270543,0.193875879,0.3996646702,-0.3142029643,0.0104300082,-0.0993912444,-0.2552524507,-0.1477756053,0.3164612353,0.0915722325,-0.0733203068,-0.1286980063,-0.0911492333,-0.4403863251,0.0224146955,0.4155964255,0.0775736794,-0.0431421921,-0.1472555399,-0.1289371997,0.1835328788,0.0123265367,-0.0253041983,-0.0034029305,-0.0076587223,0.0344733708,-0.0503452346,-0.1979504228,-0.1843430847,0.0883341581,-0.0167948212,-0.0880382955,0.4143818021,0.0103638228,-0.0808538944,-0.2035551518,-0.1448958218,-0.0131287454,0.0345769264,-0.0898909867,-0.2419079542,0.1191693917,0.5527589321,-0.0734019354,-0.1386561394,0.2675365806,-0.0382712595,0.4198622406,-0.1796670258,-0.0393933207,0.3280020356,-0.0487394296,-0.2213156968,-0.4990406334,0.8817309141,-0.1676863879,-0.0312045347,0.2908248603,0.5615332723,-0.3180469573,0.2762522399,0.3688995838,0.8417679667,-0.0185979363,0.1977707744,-0.0053540068,0.2951343656,0.5957835317,-0.0239933729,0.169372499,-0.0181187335,-0.2097397298,-0.1149546653,-0.0317051075,0.1744549721,0.2964930832,0.0774027035,0.1000413075,-0.0626665205,0.1996498704,-0.0005216507,0.2713555694,-0.0904924423,-0.2058008313,-0.1727835089,0.0736935809,-0.0616924241,-0.5089703202,-0.130548045,0.06081618,0.388145864,0.1675246507,-0.2187086493,-0.1979362071,0.1505219489,0.3563309312,0.0353714041,-0.6842543483,0.239434734,0.1809998453,0.1021335199,0.0643519834,0.0524835624,0.168722868,-0.3369868398,0.0178163834,0.0921940953,-0.1664424837,0.1632102728,-0.1282479614,-0.2016726583,-0.0558165684,0.5025715232,-0.7185156941,-0.5703017712,0.307919085,0.0859804824,-0.0334446579,0.0350983962,0.2057153136,-0.0551436767,-0.0677528381,0.104827255,0.0879895464,-0.2392212301,0.1627586633,0.0502838343,0.1598158032,-0.0491296947,-0.0106623732,-0.0490779206,0.2898964882,0.2006638497,-0.1472039521,-0.0374788195,-0.1213695705,0.0482644066,-0.0772889256,0.2324201465,-0.1316272169,-0.2182991207,0.407448113,-0.0788135529,-0.2187265754,-0.0651693046,0.0112348842,-0.1179828495,-0.3917020857,0.0955931544,0.0507767536,0.1208870411,0.1601776928,-0.3732312918,0.3524297178,-0.2587747276,0.3208484352,-0.1767922789,0.0786029547,0.2764823139,0.2889790237,0.356307596,0.1760884076,-0.1193452477,0.0895463526,-0.0482947305,0.2396871001,-0.1932824999,-0.1602515727,-0.3207264841,0.0599376447,0.0313454792,0.12892434,-0.0888889134,-0.033944916,-0.2797559202,-0.0945420414,0.1314073503,-0.0925764292,0.0310408473,0.3647340834,0.5323236585,-0.0947637036,-0.0361987352,0.0760579184,-0.0945942551,0.2145778686,0.0402089059,0.3685859442,0.0791758224,0.2802962959,-0.1241713837,0.1485524774,0.186453253,-0.2994237244,0.3278986812,0.0338353403,0.0155247897,-0.0748774707,0.408010453,0.1803857386,-0.1150495335,0.0149803497,0.085078612,0.1334756017,-0.1475752592,-0.0743364245,0.3132946491,0.2744504213,0.0356729813,0.2053097039,-0.2430108935,0.1697616577,-0.1594981551,-0.2489584684,0.1026420072,-0.6579891443,0.1840436608,0.2643785179,0.0433319546,-0.180830285,0.0193112809,0.1276965737,0.0166517794,0.1881911755,-0.16838108,0.8092685342,-0.1840478927,-0.1278982908,0.1848487854,0.0316966735,-0.2059449852,0.5624418855,-0.0316811167,0.1091760099,-0.0289333481,0.3271669447,0.2532167435,-0.0290780347,0.0739532262,0.2886963785,0.211538747,-0.0732331201,-0.1257724613,0.0407457538,-0.203605175,0.0540896356,0.2131660879,-0.1853108257,0.2970460355,0.3495043516,0.2327223122,0.182858333,0.2566036284,-0.011837638,0.410710603,-0.2600878179,-0.4787492752,0.0283813104,0.0989527032,0.1630559862,-0.1380243748,-0.1195316166,0.0746000558,-0.0368577652,0.1559404731,-0.2296406478,0.1435088068,-0.1377499253,0.1587155163,0.0420501195,0.3200648129,0.1499036252,0.022486398,-0.0490840562,0.1850377321,0.1309322566,0.0197506379,-0.9951924086,0.4971162975,0.0797643214,-0.3842939734,-0.2463439405,0.0587401614,0.0508809537,-0.1601072997,0.3139059544,-0.2779950798,-0.1711705476,-0.0219555311,0.0174336527,-0.054586485,0.3563877642,-0.2912062705,0.1353495419,-0.1470326483,-0.2613186836,-0.2869455814,-0.3777897656,0.1908326447,-0.0458546095,0.2734177411,0.1231935024,-0.1553794891,0.3742073178,-0.3006043434,-0.7832527161,-0.3912226558,-0.0866847262,-0.0562231652,0.1022988185,-0.433424741,-0.0630541444,-0.0159938913,-0.1879207641,0.0509356149,-0.1320949793,-0.1596623659,0.0530988239,0.1379394084,0.1119832993,0.0330968127,0.0941122547,0.0477158763,0.2522908449,0.0549468659,0.1803002805,-0.0122475913,-0.0034881623,-0.2898731232,0.525408268,-0.1495328844,0.2207558453,0.0101914592,-0.2680514455,-0.1662404388,0.282289654,-0.0988341123,-0.1422419548,-0.0521846972,-0.0936433375,0.1360420436,0.086337328,0.1389862746,0.4783402979,-0.2211990803,0.1171759292,0.2113041729,0.010494641,0.1683929265,-0.679317534,-0.2071979493,-0.0406591035,0.3552740216,0.0423375778,-0.1263894439,-0.4131152332,-0.4194979668,0.3656519055,-0.0062356475,0.160014078,0.4764111936,0.2468638718,-0.2728731334,0.0561023392,-0.0114117945,-0.1101367027,-0.1423386186,0.0051894519,-0.6049827933]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/312","title":"[Feature request] Add `shard()` method to dataset","comments":"Thanks for the pointer to those functions! It's still a little more verbose since you have to manually calculate which ids each rank would keep, but definitely works.\r\n\r\nMy use case is multi-node, multi-GPU training and avoiding global batches of duplicate elements. I'm using horovod. You can shuffle indices, or set random seeds, but explicitly sharding the dataset up front is the safest and clearest way I've found to do so.","body":"Currently, to shard a dataset into 10 pieces on different ranks, you can run\r\n\r\n```python\r\nrank = 3 # for example\r\nsize = 10\r\ndataset = nlp.load_dataset('wikitext', 'wikitext-2-raw-v1', split=f\"train[{rank*10}%:{(rank+1)*10}%]\")\r\n```\r\n\r\nHowever, this breaks down if you have a number of ranks that doesn't divide cleanly into 100, such as 64 ranks. Is there interest in adding a method shard() that looks like this?\r\n\r\n```python\r\nrank = 3\r\nsize = 64\r\ndataset = nlp.load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", split=\"train\").shard(rank=rank, size=size)\r\n```\r\n\r\nTensorFlow has a similar API: https:\/\/www.tensorflow.org\/api_docs\/python\/tf\/data\/Dataset#shard. I'd be happy to contribute this code.","comment_length":71,"text":"[Feature request] Add `shard()` method to dataset\nCurrently, to shard a dataset into 10 pieces on different ranks, you can run\r\n\r\n```python\r\nrank = 3 # for example\r\nsize = 10\r\ndataset = nlp.load_dataset('wikitext', 'wikitext-2-raw-v1', split=f\"train[{rank*10}%:{(rank+1)*10}%]\")\r\n```\r\n\r\nHowever, this breaks down if you have a number of ranks that doesn't divide cleanly into 100, such as 64 ranks. Is there interest in adding a method shard() that looks like this?\r\n\r\n```python\r\nrank = 3\r\nsize = 64\r\ndataset = nlp.load_dataset(\"wikitext\", \"wikitext-2-raw-v1\", split=\"train\").shard(rank=rank, size=size)\r\n```\r\n\r\nTensorFlow has a similar API: https:\/\/www.tensorflow.org\/api_docs\/python\/tf\/data\/Dataset#shard. I'd be happy to contribute this code.\nThanks for the pointer to those functions! It's still a little more verbose since you have to manually calculate which ids each rank would keep, but definitely works.\r\n\r\nMy use case is multi-node, multi-GPU training and avoiding global batches of duplicate elements. I'm using horovod. You can shuffle indices, or set random seeds, but explicitly sharding the dataset up front is the safest and clearest way I've found to do so.","embeddings":[-0.0903999358,0.01625273,-0.1279894263,-0.0729840547,-0.0588933155,-0.0729780346,0.4991557598,0.4099660814,-0.0554774366,0.2099151611,0.2360797673,0.3211589158,-0.3143125176,0.4003581405,0.6131051183,-0.4336142242,-0.032296069,0.00938618,0.215573594,0.0877346918,-0.0656637102,-0.1401859075,-0.0400372222,-0.3743532002,-0.0052027269,0.1019332334,-0.1808533818,-0.2872148752,0.2049991786,-0.2163394541,0.166055724,0.0582099371,0.1823759824,0.1601049304,-0.0001183805,-0.191446349,0.1588507891,0.1180886403,0.0012730663,-0.139281705,-0.4328423738,-0.4010118246,-0.0340188183,-0.2599122524,-0.0119534042,0.4012717605,0.0698875487,-0.0125155281,0.392546773,0.0025384095,0.0516128317,-0.060711991,0.1544547975,0.1872902066,0.4036003351,0.2214667499,-0.2822946012,-0.0089455582,-0.0786861405,0.1254140884,0.1136673093,0.1217194945,-0.0275734589,-0.1082073972,0.1557449698,0.1765587926,-0.3047389686,-0.1531616449,-0.1950687766,0.4781175852,-0.3025984764,-0.1509925127,-0.4132490456,-0.3879085779,0.2934780717,-0.7473744154,-0.3074724674,0.2714187205,-0.2418925315,-0.27542454,-0.2380059063,-0.27680704,-0.2071870416,0.077737093,0.2221379429,0.263864696,0.1681626588,-0.0175340548,0.3311504424,-0.0402620845,-0.0945930332,0.0755236372,0.2097242475,0.1188467443,-0.614027977,-0.3843820095,0.1218651533,0.3380563557,0.3469469249,0.1159338206,0.1381366998,0.2720663846,0.1744915545,0.030086508,0.0391739421,-0.0918158963,0.220510453,-0.0723302066,-0.0907713547,-0.2289366275,0.2165081501,0.1132825091,0.2126709074,-0.2785083652,-0.0142555535,0.0081662051,-0.2290856242,0.1169550493,0.1137102172,-0.2728081346,0.2444725037,-0.2083791494,0.2825330198,-0.0320820659,0.2095986605,-0.2082325071,-0.3277778327,0.0349674299,-0.1531598717,-0.3048244119,-0.0477774031,0.0721688047,-0.1881659329,0.2256807834,-0.0001613488,-0.2330842167,-0.27420187,-0.0648637041,0.4346828461,0.1786167026,0.084065333,0.2126348764,0.1357833147,0.088941291,-0.257251054,-0.3714539409,-0.0325550213,-0.0415745378,-0.3541635871,0.0491739549,-0.587732017,-0.3894829452,-0.1214290634,0.0379108265,0.0685505718,-0.1741915941,-0.6619852781,0.4921070933,0.0940611511,-0.1506498903,0.0367395729,-0.1810275912,-0.6687296629,-0.2130110413,0.3053415716,0.0020801318,-0.0641986951,-0.1369177401,-0.7781395912,0.1745422035,0.1327816695,-0.0098212939,-0.0979701653,0.0036863945,0.06320519,0.6894866824,0.7901318073,0.0799240917,-0.259321481,0.249647662,-0.0498580597,-0.2648639977,0.0904735252,0.302419275,0.2453381717,-0.0833153278,0.0976925492,0.353602618,-0.2485919893,-0.0741748363,0.0000437247,-0.1481006593,0.2624673843,0.4019246697,0.1613715738,-0.099804841,0.4458760321,-0.2314695865,0.2351807207,-0.0646500513,-0.0449084193,-0.1955905259,0.1874001026,0.0401863866,-0.1198532954,-0.480161339,-0.0211492348,0.0615554862,0.0188921336,-0.0062778052,0.0000303681,-0.289360553,0.1609876752,-0.0148133347,0.015560302,0.0897698998,-0.024887925,-0.2294120044,0.0718062147,-0.1665593982,-0.3250158727,-0.4209618866,-0.4153961539,-0.0026576987,-0.3153293431,0.4252645075,0.3351682127,-0.1314558983,-0.1251170039,0.350163728,0.0106083928,-0.1990663111,0.2325002402,0.3144398332,0.1239589378,-0.1851849705,0.1384314001,0.318557173,0.1622317731,0.1150949895,-0.3427917063,0.021769261,-0.2812332511,-0.0201554876,-0.0249481257,0.4834856987,-0.4536584616,0.3446874917,-0.0948198065,0.0413978621,-0.0674914867,-0.0322815329,-0.082105495,-0.1548915505,0.0170999821,0.2470054626,-0.1021986157,-0.288100481,0.1100903898,0.1803701967,0.5856354237,-0.1634201407,0.2819884717,0.2988931835,0.4121267498,0.1175744161,0.1000778899,0.490033567,0.3577459455,0.1889505535,0.0540700667,0.1835535169,0.007465404,-0.1562810391,0.0177055821,0.1334724575,-0.1327423155,0.0875066817,0.0698189884,-0.0496655107,-0.1410765946,-0.1763355136,-0.0049803359,0.0110957921,-0.074994266,-0.2284101099,-0.2371111661,0.2976135015,-0.2259353101,0.1298104078,-0.3122288287,0.0116350716,0.0476827845,-0.1525698751,-0.1405093521,0.1476093829,0.2023098767,0.3973855972,-0.0970733762,-0.047663901,0.2124168575,-0.0644404441,0.1698105633,-0.0129557028,0.2171652317,0.0514901616,0.5093231797,0.0358145088,-0.0743879899,-0.2529281378,-0.4815770686,-0.0534883998,-0.2781827748,0.2317032069,0.156866774,-0.032728333,-0.408772558,-0.1373485923,0.0282857083,-0.1450952739,-0.071896106,0.1404522508,0.1567685306,0.2897759378,-0.0762949064,0.0976015851,-0.1189652979,-0.1235760674,0.3168518543,0.1304379255,0.2200096697,-0.2814023197,-0.0102225123,0.106131494,-0.0294495523,-0.1092444509,0.1849551201,-0.310992986,0.2382128984,-0.3582490087,-0.1694325507,-0.2485969216,-0.1250095218,-0.0642140955,0.7491729259,-0.0403823555,-0.066857405,0.0108843669,0.1439590752,-0.1054384038,0.0316976011,0.5483807325,-0.0572200306,0.1360258907,0.2081536502,0.2660071254,0.0347035527,0.0306900982,0.0958150104,0.2953639627,0.6451733708,0.3203065693,0.8731663227,0.0987205207,-0.0327910855,-0.0986943766,-0.1933798939,-0.1365426034,-0.0296432823,-0.129579857,0.1816159487,0.0352501906,0.0240287017,0.2045759708,0.0447306149,-0.1147902161,0.020648187,-0.3997074962,0.0664206445,-0.247212097,0.1935067177,-0.0922368467,0.4955633879,-0.3516761363,-0.1373445988,-0.1173922196,-0.4128207862,0.0296420511,0.0561973751,0.2447420955,-0.1990012079,-0.4037176669,-0.3870333731,-0.2138939053,0.2600160837,0.4698672593,-0.0854348466,0.0004026716,-0.0650440454,-0.1698829681,0.174891606,0.3240430951,-0.3477110565,-0.0149209602,-0.1251921505,-0.1557665467,-0.0763009042,0.2319218665,0.0541759357,-0.0669330359,-0.2182870358,-0.1152840927,-0.3981708586,0.0540198423,0.4199634492,0.0178792104,-0.0279563777,-0.0852494389,-0.1949022561,0.2019471973,0.0781320632,-0.0159119498,0.0266292784,0.0168030262,-0.0477672704,-0.0187462419,-0.1511361897,-0.144604668,0.1339359134,-0.1123268828,-0.0347809345,0.4088841081,0.0423426665,-0.0951987579,-0.2162930965,-0.1434260905,-0.0225667749,-0.0611088984,-0.1042453423,-0.3321190774,0.0490562767,0.5573011637,-0.0762399733,-0.116803892,0.2294700891,-0.1556234956,0.4777894914,-0.1714918762,-0.0785924122,0.2965258658,-0.1075154617,-0.208022505,-0.5555896759,0.8288854957,-0.1245793849,0.0166231375,0.393732965,0.642480135,-0.2583677471,0.2959198654,0.4128728211,0.7760377526,-0.0605917312,0.1950274706,-0.018349817,0.2824895978,0.5539624691,-0.0106745483,0.2589059472,-0.0291999057,-0.2499285191,-0.1381643564,0.0118396683,0.13045156,0.3750299811,0.0805963278,0.1500472128,-0.0049862168,0.2164103091,-0.1083190441,0.2598162591,0.0379077382,-0.2282008231,-0.0523660704,0.0943668708,-0.0569094159,-0.4457859695,-0.1336326599,0.0326853506,0.3783545792,0.1265669018,-0.2668841779,-0.1820002049,0.207795918,0.4457689524,0.0431359373,-0.5865632892,0.1645914614,0.2436376065,0.1872726083,0.0645156875,0.0499899462,0.1614743173,-0.2400116473,0.0393421799,0.0433748923,-0.1680683047,0.1748079956,-0.0558401644,-0.2017071843,-0.1359783113,0.5046851635,-0.7938537598,-0.6001187563,0.2508292496,0.087469928,0.0404776074,0.1548402309,0.2515393794,-0.0826455802,-0.0686346665,0.1006217897,-0.0030941176,-0.2071591467,0.1797482222,0.0274524484,0.1658818871,0.0211009327,0.0017025947,-0.0602237768,0.2708492875,0.2140356004,-0.1862253994,-0.0247167889,-0.1242655888,0.0516229197,0.0239041075,0.2162699401,-0.1021242738,-0.3356439471,0.3976138234,0.0025816767,-0.2233180553,-0.0806260481,0.1447622329,-0.0104240533,-0.3164904118,0.0591628402,0.0351316482,0.1229835078,0.1719331592,-0.3931394815,0.3110384643,-0.2184207588,0.288292706,-0.1775633991,0.0338153988,0.2850924432,0.195794642,0.2985618412,0.2079675496,-0.1609911919,0.0786816776,-0.0359992869,0.2371992022,-0.1223623678,-0.1275121272,-0.2695200443,0.0688310713,0.027952617,0.2211998403,-0.037553668,-0.0935258493,-0.3762474358,-0.1821535081,0.0634019449,0.052871421,0.0136505775,0.3772658408,0.5071145892,-0.0989447236,-0.0019492386,0.0522293597,-0.0482016727,0.2042929977,0.0609989949,0.3407793343,0.1450764835,0.2970131934,-0.0868910253,0.1984772831,0.2186187506,-0.3008509278,0.3319802284,0.0414673537,-0.0337259099,-0.041127272,0.4718823731,0.0758411884,-0.0829894096,0.0459229797,0.071422711,0.1331994534,-0.2280475348,-0.1132473946,0.2901129127,0.2939963937,0.000862708,0.2336083651,-0.1820556819,0.1447457969,-0.1569230407,-0.2128676176,0.0376844518,-0.6488708258,0.2275586277,0.2615426481,0.0582479909,-0.2244460285,0.0148985377,0.0247305278,0.0251929089,0.2591375709,-0.205508396,0.8164564967,-0.2379744649,-0.1572355926,0.2347330451,0.0409793183,-0.1476863474,0.5167104006,0.0377834141,0.1450148672,-0.016497409,0.3107829094,0.1570971608,-0.0756548494,-0.0013405869,0.2672916055,0.2398793548,-0.0155857988,-0.1386007667,0.0024057401,-0.3002158701,0.0806353018,0.2542969286,-0.1714050174,0.3803829253,0.3777695894,0.3144789636,0.1301078796,0.1567281038,0.1348094046,0.4964342415,-0.2856993675,-0.5428259373,0.0903881937,-0.0121220602,0.2014399618,-0.1113077253,-0.1230747327,0.1279772073,-0.074117139,0.1503968984,-0.1897893697,0.1648390144,-0.0647135973,0.0841424167,0.0424119756,0.3490308225,0.0915729627,0.0008262708,-0.0268216431,0.2064237744,0.1759447604,0.098636277,-0.8560948968,0.4871613383,0.1096793413,-0.3625624776,-0.2355593145,0.0789653137,-0.0330509245,-0.2587110102,0.287732929,-0.1654184461,-0.1679375768,-0.0751760677,0.0251645315,-0.0437027812,0.3141371906,-0.246983856,-0.016326867,-0.1131262332,-0.3326271176,-0.3796083033,-0.3455867171,0.1628321558,-0.0077939238,0.2399308085,0.1390847117,-0.096791707,0.2934943438,-0.2871805727,-0.7887305617,-0.3314759731,-0.0421139374,-0.1008521318,0.0531088896,-0.4707805514,-0.1179287285,0.0159951542,-0.1709063351,-0.0109251626,-0.085184373,-0.1493895054,0.0672284886,0.1867273152,0.0734201893,0.0204159301,0.0058119018,0.0098144086,0.263355881,0.0031536403,0.1583027393,0.0423823856,0.0770889074,-0.3178185821,0.5716711879,-0.1982651055,0.2122271806,0.0214331541,-0.2425095439,-0.2888360023,0.2472967505,-0.1003843173,-0.1212279201,-0.1928638667,-0.0719003007,0.1775726229,0.0586772971,0.1352394521,0.4553301334,-0.2573918402,0.1884123236,0.2233216912,-0.0276321284,0.1007479206,-0.7333005667,-0.1903629303,-0.1361529082,0.3577404618,-0.0181629527,-0.040285375,-0.3952936232,-0.3722831011,0.3207330406,0.0382906497,0.1953834295,0.4563092589,0.1442922652,-0.2179151177,0.0417596065,-0.0161644202,-0.1048095226,-0.1096249297,0.0603449382,-0.5192135572]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/302","title":"Question - Sign Language Datasets","comments":"Even more complicating - \r\n\r\nAs I see it, datasets can have \"addons\".\r\nFor example, the WebNLG dataset is a dataset for data-to-text. However, a work of mine and other works enriched this dataset with text plans \/ underlying text structures. In that case, I see a need to load the dataset \"WebNLG\" with \"plans\" addon.\r\n\r\nSame for sign language - if there is a dataset of videos, one addon can be to run OpenPose, another to run ARKit4 pose estimation, and another to run PoseNet, or even just a video embedding addon. (which are expensive to run individually for everyone who wants to use these data)\r\n\r\nThis is something I dabbled with my own implementation to a [research datasets library](https:\/\/github.com\/AmitMY\/meta-scholar\/) and I love to get the discussion going on these topics.","body":"An emerging field in NLP is SLP - sign language processing.\r\n\r\nI was wondering about adding datasets here, specifically because it's shaping up to be large and easily usable.\r\nThe metrics for sign language to text translation are the same.\r\n\r\nSo, what do you think about (me, or others) adding datasets here?\r\n\r\n\r\nAn example dataset would be [RWTH-PHOENIX-Weather 2014 T](https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/)\r\nFor every item in the dataset, the data object includes:\r\n1. video_path - path to mp4 file\r\n2. pose_path - a path to `.pose` file with human pose landmarks\r\n3. openpose_path - a path to a `.json` file with human pose landmarks\r\n4. gloss - string\r\n5. text - string\r\n6. video_metadata - height, width, frames, framerate\r\n\r\n\r\n------\r\n\r\nTo make it a tad more complicated - what if sign language libraries add requirements to `nlp`? for example, sign language is commonly annotated using `ilex`, `eaf`, or `srt` files, which are all loadable as text, but there is no reason for the dataset to parse that file by itself, if libraries exist to do so.","comment_length":131,"text":"Question - Sign Language Datasets\nAn emerging field in NLP is SLP - sign language processing.\r\n\r\nI was wondering about adding datasets here, specifically because it's shaping up to be large and easily usable.\r\nThe metrics for sign language to text translation are the same.\r\n\r\nSo, what do you think about (me, or others) adding datasets here?\r\n\r\n\r\nAn example dataset would be [RWTH-PHOENIX-Weather 2014 T](https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/)\r\nFor every item in the dataset, the data object includes:\r\n1. video_path - path to mp4 file\r\n2. pose_path - a path to `.pose` file with human pose landmarks\r\n3. openpose_path - a path to a `.json` file with human pose landmarks\r\n4. gloss - string\r\n5. text - string\r\n6. video_metadata - height, width, frames, framerate\r\n\r\n\r\n------\r\n\r\nTo make it a tad more complicated - what if sign language libraries add requirements to `nlp`? for example, sign language is commonly annotated using `ilex`, `eaf`, or `srt` files, which are all loadable as text, but there is no reason for the dataset to parse that file by itself, if libraries exist to do so.\nEven more complicating - \r\n\r\nAs I see it, datasets can have \"addons\".\r\nFor example, the WebNLG dataset is a dataset for data-to-text. However, a work of mine and other works enriched this dataset with text plans \/ underlying text structures. In that case, I see a need to load the dataset \"WebNLG\" with \"plans\" addon.\r\n\r\nSame for sign language - if there is a dataset of videos, one addon can be to run OpenPose, another to run ARKit4 pose estimation, and another to run PoseNet, or even just a video embedding addon. (which are expensive to run individually for everyone who wants to use these data)\r\n\r\nThis is something I dabbled with my own implementation to a [research datasets library](https:\/\/github.com\/AmitMY\/meta-scholar\/) and I love to get the discussion going on these topics.","embeddings":[-0.32776618,0.4525655508,-0.0732839257,0.0123400809,-0.0788401142,-0.0115335379,0.2201107293,0.2610009015,0.2013961822,-0.3687560558,0.0878804475,0.2947927713,-0.3939641416,0.2834338248,0.3449884951,-0.176551044,-0.0004947542,-0.0626648739,0.0087343119,-0.0980993658,-0.2763954997,-0.1909826398,0.0531514324,-0.007089871,-0.3837481141,-0.332061857,-0.1520041823,0.0552289262,-0.0798219368,-0.3670464754,-0.4654010832,0.1338605583,0.3811520338,-0.0573748946,-0.0001092122,-0.2131741047,0.0867840946,0.1366245598,-0.2981795073,-0.1120385602,-0.1225187331,-0.6363066435,-0.1847902238,-0.2197489291,0.3014643192,-0.3671759665,0.1177984476,-0.4387135506,0.0204097852,0.2650385797,0.2657268345,0.0734952241,-0.1763009578,0.2845447361,0.1073937044,0.2880243063,0.0539344735,-0.3114773035,0.3666415513,-0.0040134597,-0.1532109678,0.2822242379,0.0871672481,-0.5369473696,0.2831125557,-0.0832231119,-0.1681210697,-0.0665763915,-0.0609676987,0.0545954406,0.3695205152,-0.3887362778,-0.5512011051,-0.692758441,0.3190686107,-0.0417574681,-0.1850506514,0.3131968975,0.2486976981,-0.0441939384,-0.2032430619,-0.6055930257,0.0735249966,0.2032313198,0.0504943393,0.3771199882,0.1068511754,0.0853631496,-0.0598349199,-0.05807098,0.1020340174,-0.4142607152,0.0059124534,0.1247747168,-0.2750384808,-0.1684951633,0.0074741784,0.1095340252,0.2164651155,0.1288278699,-0.0004105956,0.0027560038,-0.336345166,0.0445142612,0.3745672107,-0.0391658172,-0.0283417664,0.144290179,0.1519366056,-0.2677285969,0.4116028547,0.0693721548,-0.0097140595,0.0064761494,-0.2346767187,-0.2014696002,0.119522281,-0.0326781608,-0.1603718996,0.0468571745,-0.1952341199,-0.2574142814,0.1204497814,-0.0789810494,0.0028640912,0.2978838086,0.0192716494,0.3454528153,0.1967356801,-0.168796137,0.0183068346,0.1815776974,-0.3700411618,0.2450752258,0.3831808269,0.0594607778,0.0507775694,0.0166308731,-0.1397555321,0.0706387311,0.3091606796,-0.0534301437,0.0902294517,-0.2173833251,0.0065714717,0.0854129121,-0.3741538823,-0.0530162305,-0.3044703603,0.11985939,0.0816684663,-0.0236122105,0.0384660922,0.2177266181,-0.3923450708,-0.3972732723,-0.3264209032,0.6124715805,0.0717181116,-0.0422144085,0.1649869084,-0.1042282134,-0.3048969507,-0.0357939191,-0.2105996013,0.4268853962,-0.3384500742,-0.1672324985,-0.1232564673,-0.0646909699,-0.2797655165,0.2127159387,-0.2385702431,0.1813210249,0.0024056698,0.2150002271,0.5056704879,-0.2569220066,-0.0756710097,0.3999677002,-0.0459146053,-0.0478044972,0.0169659127,0.3778139353,0.13293311,-0.1663077772,-0.2145010382,0.3691989481,-0.2047019601,-0.0959278271,-0.162826106,-0.1164556295,0.420647949,0.3798087239,0.1131544113,-0.0518590808,0.0357221775,0.3774410188,0.0435267687,-0.196443826,0.394294858,-0.1603751779,-0.017672481,-0.0753514692,-0.1399202645,-0.3019448817,-0.289950788,0.0062835142,0.0455887057,0.0574385636,0.1755376011,-0.0628907084,0.2070700079,0.0029975781,0.1741631925,0.1203707755,0.1408800334,0.1355981231,-0.2524044514,-0.1082754731,-0.4150164425,0.5105982423,-0.2547084391,-0.0392756015,-0.061357893,0.252538234,0.0246329252,0.1412556916,0.1342978776,0.6351205707,-0.2699644566,0.0135254422,0.2918762267,-0.1600831151,-0.2763584256,0.1621051282,0.3680669963,0.2745388746,0.3609109521,-0.1977920085,0.3337173164,-0.2925091088,0.2045737356,-0.2085653543,-0.0754081234,0.190060094,0.2169563025,0.1152892187,0.1386888027,-0.0741415396,0.0885092244,-0.0570485815,-0.2784792185,-0.0423538387,0.2771929204,-0.0012598449,0.4623275399,0.2014108598,-0.3186942637,0.1495314538,0.313357681,-0.1303913891,-0.013362282,0.2636477947,-0.0268867984,-0.1714869738,-0.0405314527,0.2153082192,0.414595753,0.3478784263,0.0927986503,0.046907533,-0.1673026681,-0.1603662968,0.0263926629,0.1448447108,-0.0611667261,0.2120802552,0.0907922462,0.1768361777,-0.1688581407,-0.2871466279,0.0197021011,-0.0383438766,-0.0456265509,0.1216837764,0.0811139867,-0.4516416788,-0.2804162204,0.0658235773,-0.2214261591,-0.3173054159,0.1547326893,-0.3717639148,-0.0085091023,0.0050645825,-0.3108474016,0.6126178503,-0.6380677223,-0.0599995442,-0.0429215617,-0.3297202587,-0.1528772861,0.1691616178,0.3665487766,0.152233243,0.5014584064,0.0929036438,0.2722419798,-0.2442156672,-0.5257211328,-0.1177096665,-0.3639345467,-0.1025497019,-0.118492797,-0.2681989372,0.3248793185,-0.1772309095,-0.1412790269,0.061333891,-0.0901023895,-0.3070274889,-0.0102554616,-0.3610468507,-0.0657807142,-0.2139178962,-0.2801144719,-0.3550033271,0.2222713381,0.1472243965,-0.1143486351,-0.082553491,-0.1171493009,0.2302026898,-0.0559170246,0.0973561779,-0.1091556102,0.221532777,0.0966774449,0.0737006962,-0.2337379009,0.0355563127,-0.059986189,0.085655205,0.2619890571,-0.5661298037,0.1449562758,-0.069379352,0.1110566705,-0.0226870645,-0.0608457588,0.0330215357,0.28169173,-0.0182203129,-0.1531530619,-0.4342613816,0.032331273,0.0822540969,0.2794435024,-0.0637432709,0.0371206515,-0.038826637,0.2141556442,0.2831184268,-0.3406554759,0.2703122795,0.3039261699,0.3658166528,0.1094877496,-0.1096167639,0.4175211191,0.0055683153,0.112026751,0.200948894,0.1855513752,-0.2546825409,-0.1513792723,0.1681476682,-0.0529038534,0.0332621634,0.3392641246,-0.3259800375,0.2681798041,0.0694595575,-0.2936660945,-0.3240588307,-0.1189767197,-0.052773159,0.2530348003,0.2221096307,0.1111335531,-0.7394844294,0.1310262531,-0.3390880227,0.1766641736,0.2675981522,0.3600482345,-0.2742620111,0.0379048064,-0.0245678145,-0.0034533308,-0.0596033186,-0.3821159303,-0.4413818717,-0.0014769159,0.137836799,-0.19347848,0.2766523957,-0.0731511191,-0.0202834979,0.2345368564,0.280819267,-0.1643844545,-0.0972538516,0.1360974461,0.3822792768,0.0138400001,-0.3639440835,0.2148842365,-0.1549583375,-0.1572473794,-0.2554562092,-0.1148084402,-0.1886718273,0.1932781339,0.0202524997,-0.1647088379,-0.1592662036,0.3284504712,-0.0759799704,-0.3297518492,0.1766274869,0.1995892078,-0.1882819384,0.3775122166,0.0268080439,0.4057422578,-0.1532982439,-0.2642554045,0.0313087702,-0.1481190622,0.5700646043,0.3201323748,0.0756625757,0.2256222516,0.0236608498,-0.2887450159,-0.3492979407,0.2807303965,0.2456750721,0.3173533082,-0.349888742,-0.5515734553,0.4434360862,0.1949841082,-0.0818312466,-0.0699673146,-0.1479028463,-0.2677091062,0.2063097805,0.0511063524,0.776835084,-0.1354139894,0.3382517397,0.0252800677,-0.455047369,0.5315403342,-0.1426776648,-0.1611883491,-0.0259686038,0.1625661254,0.0131331086,-0.0763376281,0.0255653188,0.2910048664,0.1330581605,-0.0140585545,0.210493058,-0.1506529152,0.0964930058,0.7823001146,-0.2208659649,-0.3575436771,-0.3073402643,0.1696504653,-0.2132996172,-0.105435349,-0.0683997646,-0.2008948326,-0.1548354179,0.3081341684,-0.3293710649,-0.0127807362,0.3329559267,-0.1626687348,-0.1592619121,-0.0814422593,0.6952056885,0.0678873658,0.3631236255,-0.0760263279,-0.112370111,-0.0397282057,-0.3975079358,0.0418185964,0.1506001651,0.0597330555,0.2984387279,-0.063408494,-0.0220114924,0.2372189909,0.1376472414,0.0222477075,-0.4388074279,0.0340041146,0.30602476,-0.1512724757,-0.0749820843,0.588683486,-0.0940099433,0.0257536042,0.1371878982,0.2333139181,0.1604954898,0.1172072291,0.1413751841,-0.1444667429,-0.164048776,0.028744597,-0.1075515822,-0.190089941,0.185216099,0.0035882802,-0.1800153255,-0.1428713202,0.0276998281,-0.0418267548,-0.3136711121,-0.2985225022,0.2279629558,-0.0665528849,-0.2161617875,-0.0897091329,0.1252862811,-0.2783396244,0.1514790952,-0.2320459336,-0.3384659886,0.2309880108,-0.2137141675,0.1870666146,-0.0068737143,0.292260766,-0.1624262035,0.1355932206,-0.3200897276,-0.1659261882,0.0637978986,0.1973965615,0.3663798571,-0.0828705877,-0.2532849312,-0.10485439,0.0668522194,-0.2257048935,-0.1411182731,-0.1541334391,-0.0019798619,0.1152066365,0.4015695453,-0.2354111969,0.0057657105,-0.1589365304,0.0295653604,-0.1020472199,-0.1464466453,-0.0335812941,-0.2479206026,0.3239008784,-0.4041709602,-0.0752839744,-0.0530449301,0.3244349658,0.0446136035,0.2183506489,0.1103093028,0.401884824,0.2877154946,0.1919245422,-0.1951209158,0.2794585526,0.1016832516,0.2987418473,0.2239165902,-0.1098160818,0.4686130881,0.2652967274,0.1452230364,0.2806883156,-0.1026005968,-0.1588004977,0.3392032981,0.1666077971,-0.2367142588,0.1088848785,0.0988428444,0.190088883,0.1052080542,0.3454721868,0.3633002639,0.1701591909,-0.1871370226,-0.1343658864,0.2054473758,-0.1705106497,-0.1407376528,-0.0169620756,0.027636297,0.0264549591,0.1786226183,0.5877977014,0.0195205361,0.2158789635,-0.1132897809,0.570566535,0.1660265028,0.292609334,0.2518772781,-0.1501765996,0.1990005523,0.3131928742,0.1905715168,0.0940514877,0.099716723,0.3216205835,0.0940650776,-0.1549425274,0.2753291726,0.2987986207,-0.1212520078,0.0610472746,-0.179418236,-0.1137369201,-0.2061056644,0.0304278471,-0.2563616335,-0.2359250635,0.2408980727,0.0669321567,-0.0476648882,-0.4425084889,0.3935589194,0.0000653322,0.0177864432,-0.0888029635,0.1822496951,-0.1903011948,-0.0559269786,0.3564329743,0.2742901146,0.1248257235,0.184041664,-0.0299153123,0.3358231187,-0.1316554397,0.1448099315,-0.1163842231,0.4367336929,0.1713324934,-0.0272590164,0.4199909866,0.0901308954,-0.0635482371,0.2165465802,-0.0082664583,-0.3266595304,-0.19000642,0.5533900857,0.0743042007,-0.1784687191,-0.1753480285,-0.013800228,-0.0982370004,-0.3667751849,0.5476843715,-0.2209514827,0.0580913089,-0.1955658644,0.1026873738,-0.0537687466,0.2913796604,0.3408299983,0.251635164,-0.0849949345,0.1212783456,-0.4803477526,-0.0578564741,-0.0219765957,-0.1819757521,-0.273789376,-0.083331503,0.0723078549,0.325225383,-0.0984482169,0.2862285376,-0.0201152172,-0.2694889307,-0.207025215,-0.0562772527,0.2990217805,0.208694309,-0.066426307,0.0786418542,-0.1919184178,-0.0120330146,0.0305999033,-0.3177833855,-0.0572423711,-0.0207219198,0.1644587815,0.1714894027,0.0603464544,0.5298666358,-0.025094619,0.010118451,-0.2049946487,-0.3195369542,-0.2065575868,0.1681701541,-0.0357229412,0.0005479701,-0.3461435735,0.3349312842,0.0087435478,-0.0104638841,-0.0248257536,-0.1052329242,-0.2663779557,0.1646501124,-0.2904594243,0.1872282922,0.1142694131,0.3710082769,0.1334945261,0.0081687029,0.0290064663,-0.0068436475,0.1138112471,-0.0040212651,-0.0990746394,0.3302085698,0.0384699292,0.3264529407,-0.0020176813,-0.6941640973,-0.2996281087,0.4133155346,-0.2155339271,0.1270000637,0.1994938403,-0.0874264389,-0.3565057814,-0.2023378909,-0.0215880796,-0.2799351811,-0.1478655338,0.2382547259,-0.214569062]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/302","title":"Question - Sign Language Datasets","comments":"This is a really cool idea !\r\nThe example for data objects you gave for the RWTH-PHOENIX-Weather 2014 T dataset can totally fit inside the library.\r\n\r\nFor your point about formats like `ilex`, `eaf`, or `srt`, it is possible to use any library in your dataset script.\r\nHowever most user probably won't need these libraries, as most datasets don't need them, and therefore it's unlikely that we will have them in the minimum requirements to use `nlp` (we want to keep it as light-weight as possible). If a user wants to load your dataset and doesn't have the libraries you need, an error is raised asking the user to install them.\r\n\r\nMore generally, we plan to have something like a `requirements.txt` per dataset. This could also be a place for addons as you said. What do you think ?","body":"An emerging field in NLP is SLP - sign language processing.\r\n\r\nI was wondering about adding datasets here, specifically because it's shaping up to be large and easily usable.\r\nThe metrics for sign language to text translation are the same.\r\n\r\nSo, what do you think about (me, or others) adding datasets here?\r\n\r\n\r\nAn example dataset would be [RWTH-PHOENIX-Weather 2014 T](https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/)\r\nFor every item in the dataset, the data object includes:\r\n1. video_path - path to mp4 file\r\n2. pose_path - a path to `.pose` file with human pose landmarks\r\n3. openpose_path - a path to a `.json` file with human pose landmarks\r\n4. gloss - string\r\n5. text - string\r\n6. video_metadata - height, width, frames, framerate\r\n\r\n\r\n------\r\n\r\nTo make it a tad more complicated - what if sign language libraries add requirements to `nlp`? for example, sign language is commonly annotated using `ilex`, `eaf`, or `srt` files, which are all loadable as text, but there is no reason for the dataset to parse that file by itself, if libraries exist to do so.","comment_length":139,"text":"Question - Sign Language Datasets\nAn emerging field in NLP is SLP - sign language processing.\r\n\r\nI was wondering about adding datasets here, specifically because it's shaping up to be large and easily usable.\r\nThe metrics for sign language to text translation are the same.\r\n\r\nSo, what do you think about (me, or others) adding datasets here?\r\n\r\n\r\nAn example dataset would be [RWTH-PHOENIX-Weather 2014 T](https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/)\r\nFor every item in the dataset, the data object includes:\r\n1. video_path - path to mp4 file\r\n2. pose_path - a path to `.pose` file with human pose landmarks\r\n3. openpose_path - a path to a `.json` file with human pose landmarks\r\n4. gloss - string\r\n5. text - string\r\n6. video_metadata - height, width, frames, framerate\r\n\r\n\r\n------\r\n\r\nTo make it a tad more complicated - what if sign language libraries add requirements to `nlp`? for example, sign language is commonly annotated using `ilex`, `eaf`, or `srt` files, which are all loadable as text, but there is no reason for the dataset to parse that file by itself, if libraries exist to do so.\nThis is a really cool idea !\r\nThe example for data objects you gave for the RWTH-PHOENIX-Weather 2014 T dataset can totally fit inside the library.\r\n\r\nFor your point about formats like `ilex`, `eaf`, or `srt`, it is possible to use any library in your dataset script.\r\nHowever most user probably won't need these libraries, as most datasets don't need them, and therefore it's unlikely that we will have them in the minimum requirements to use `nlp` (we want to keep it as light-weight as possible). If a user wants to load your dataset and doesn't have the libraries you need, an error is raised asking the user to install them.\r\n\r\nMore generally, we plan to have something like a `requirements.txt` per dataset. This could also be a place for addons as you said. What do you think ?","embeddings":[-0.2392410785,0.4322065115,-0.0678170472,0.0264536738,-0.0018659234,-0.0265170839,0.2193028182,0.3350127339,0.2512377799,-0.298207283,0.0776481628,0.2235988677,-0.3672275245,0.3933984637,0.3731903434,-0.1788406372,-0.141786173,0.1445282102,-0.0459923074,-0.0068757799,-0.1793410927,-0.1383463293,0.0843173191,0.0032767651,-0.3872087002,-0.2946252227,-0.2022378892,0.0103004035,-0.2074933946,-0.3936846256,-0.3563819528,0.2389658093,0.4327610135,-0.0413167961,-0.000111818,-0.1965150535,0.0004837274,-0.0264123771,-0.3759377301,-0.2181890011,-0.1599296778,-0.6538229585,-0.0996578187,-0.2734122276,0.2359748483,-0.4207744598,0.1932359934,-0.4131361246,0.0240745321,0.3100079894,0.2551298738,0.2083654106,-0.1462490559,0.281671226,0.2171062529,0.422010988,-0.04849305,-0.3106179237,0.3796490133,0.0535862073,-0.1197089329,0.1756723523,0.1521621644,-0.5902650952,0.3719701469,-0.048252359,-0.2144599557,-0.1157986671,-0.0788700357,0.1758687645,0.4622625113,-0.4509210587,-0.6121043563,-0.5556027293,0.2440145165,-0.1052371785,-0.2322888821,0.4563677013,0.1665138155,0.0677278712,-0.2239359617,-0.5427981019,-0.0680542886,0.2172320634,0.1704595387,0.4320033789,0.1302766502,0.0282241441,-0.1478251219,-0.0363157056,0.0833854526,-0.4088713527,-0.054743398,0.0755022243,-0.1521911025,-0.2674987614,-0.0443733595,0.3259717524,0.2731189132,0.1776399463,-0.0119318385,-0.0192898959,-0.3869999647,0.0939607695,0.346458137,-0.0282107238,0.1112058461,-0.0207225345,0.1145520657,-0.3524111509,0.4173051715,-0.0421943292,-0.1174831688,0.0822774917,-0.2719350457,-0.1131578833,0.1355942935,-0.0246956851,-0.0708161145,0.0333586372,-0.0378730819,-0.1779818982,0.1380895227,-0.0731866136,0.0021023108,0.3331817687,0.0288685858,0.2900317609,0.1525739282,-0.2059981227,0.0905216783,0.3386293352,-0.3930284381,0.2288369685,0.3849957287,-0.073299937,0.1623158008,-0.0317895599,-0.0328403637,0.0333853848,0.3529451191,-0.1149327159,0.1789229214,-0.1788972318,-0.0192502458,0.0197334774,-0.3452151418,-0.1483695954,-0.3417448103,0.0457391031,-0.0134694092,-0.0248545855,-0.0725097805,0.1527008265,-0.3737695813,-0.3657545745,-0.3084922433,0.4737797379,0.0162905268,-0.1361462176,0.1009619609,-0.2238278836,-0.32142663,-0.0759492293,-0.2590917647,0.5499836802,-0.4043999612,-0.1308394521,-0.0857312456,-0.0600163378,-0.2899185121,0.1903125346,-0.2675342262,0.2019135356,-0.0615801401,0.1260316521,0.5230683684,-0.3338760138,0.0530650206,0.4913771451,0.0523525104,-0.0058542043,0.1632335037,0.3408497274,0.245995,-0.2945617139,-0.3590253294,0.3687414229,-0.2249614596,-0.0728120357,-0.0955336466,-0.1786126345,0.4174608886,0.2728397846,0.0470745377,-0.0589042045,0.0063619055,0.3632199764,0.0339313261,-0.212763682,0.368443042,-0.1088632941,0.0076051224,0.0101275891,-0.0747040585,-0.2783149183,-0.313542366,-0.013272156,-0.0191969667,0.0899144188,0.2209035158,-0.1653003544,0.1142111048,-0.0527443625,0.1920663267,0.0783628896,0.1011962742,0.1069428101,-0.2299120277,-0.1301419735,-0.4460364878,0.34607023,-0.0295050628,-0.0311456732,-0.0771115571,0.1165247187,0.0345047265,0.1958824247,0.1797948182,0.6314455271,-0.2389091849,-0.0438045859,0.2426503897,0.0283519365,-0.1118899435,0.1752606481,0.2202176452,0.3296031356,0.3699393868,-0.0704124495,0.3901994526,-0.2260192335,0.2912960649,-0.1420739144,-0.0708417594,0.2859691978,0.1660409272,0.0717799515,0.2095985264,-0.0442502946,0.0785996094,-0.0545328893,-0.3388661146,-0.1812788844,0.1592064053,-0.0405674204,0.3809363842,0.217303887,-0.3826072514,0.0980519205,0.4414026737,-0.031172609,0.0333945565,0.2406980842,-0.0047639427,-0.1045906693,-0.0378986001,0.1969153285,0.448066026,0.3596537113,-0.076795429,0.0328083336,-0.0605686493,-0.2623927593,0.1412644833,0.1670076251,-0.1443777084,0.1062702686,-0.0300214794,0.1540267467,-0.1510039121,-0.4089504778,-0.0720310882,-0.1273516119,-0.0816775337,0.2121165544,-0.1215317473,-0.5161290169,-0.347641021,0.1148849949,-0.0987305939,-0.2064289302,0.2614905238,-0.3910255432,0.0113179926,0.0768273026,-0.2987986505,0.643124342,-0.7239416242,0.0112155946,0.0167229194,-0.3049184382,-0.2272376865,0.140206486,0.4569424689,0.1615840793,0.5446454883,0.0485961288,0.2454333901,-0.2471980453,-0.4581659734,-0.0586194731,-0.3301423192,-0.0572842881,-0.0713345706,-0.1475493461,0.2797916532,-0.1151470393,-0.1959896833,0.0035615342,-0.0417222716,-0.2534704506,-0.0614898726,-0.3246528208,-0.193864882,-0.2606068254,-0.2597326934,-0.3310178816,0.1313392073,0.1773203611,0.0304214023,0.0264847223,-0.0815234333,0.1796399802,-0.1072296873,0.0963068232,0.0082849348,0.1083381623,0.1894970089,-0.0520259924,-0.331607163,0.0873930901,-0.022807112,0.1444895715,0.3162441552,-0.4225608408,0.0671007484,-0.0635087118,-0.0721583441,-0.0089917015,-0.0036776734,0.1186179668,0.2058078647,0.0179417096,-0.1530881226,-0.3764448762,0.0828270912,-0.0834402069,0.1790902466,0.0388725773,0.1381020546,-0.0103971222,0.1412599087,0.1929103285,-0.3087898493,0.3499720395,0.2729178667,0.4398564994,0.1196243763,-0.0162531864,0.3699266016,0.1009175405,0.2219851315,0.335272193,0.2073653042,-0.0974008888,-0.2005241364,0.1487076133,-0.1287809461,0.0392954387,0.4502097368,-0.3502020836,0.3220341802,0.0612318814,-0.1601587534,-0.2231767327,-0.2075295299,-0.0421403311,0.3462342918,0.3087904453,0.1874963641,-0.6915634871,0.0281013027,-0.1986745745,0.1414340734,0.1980623901,0.3719217777,-0.2181531042,0.1971213967,-0.0247206092,0.0649770051,-0.1725747436,-0.3487086296,-0.5127555728,-0.0992467776,0.0486986265,-0.2950547934,0.2454340011,-0.1107888743,-0.0176842064,0.099362269,0.249022454,-0.2287604511,-0.2461145669,0.0234906543,0.3472217619,-0.052196119,-0.4224450588,0.2001164705,-0.1370128393,-0.2033844143,-0.3195958734,-0.0662866607,-0.2527782321,0.1040271521,0.0779755712,-0.1703685671,-0.1411121339,0.4347405732,-0.1184365377,-0.2026239634,0.1415362656,0.0448184609,-0.2620019317,0.3238806427,0.0831574649,0.3049007058,-0.1074472442,-0.0429565832,0.045096159,-0.2195607871,0.6228874922,0.334086746,0.081206426,0.2502073348,-0.0182954036,-0.3800028861,-0.3088877499,0.1887248904,0.1485966146,0.0082774926,-0.37031883,-0.6633868217,0.3367486894,0.165156588,0.0058126603,-0.0749252588,-0.0494813025,-0.3187718689,0.3240888715,0.1464211196,0.867760241,-0.1301515847,0.3352992833,0.0063564223,-0.4858841896,0.596965611,-0.2343140244,-0.1293558478,-0.1530999988,0.1366398185,-0.120457463,-0.0773331001,0.1317735612,0.3120001554,0.0009116311,0.0709005818,0.1817296743,-0.116382435,0.0805340037,0.8477870822,-0.3845674098,-0.4652253389,-0.3680853248,0.1518678516,-0.2248250246,-0.1328030527,-0.1276872903,-0.2470861971,-0.1100302115,0.1595669389,-0.3610002995,-0.0753130466,0.3345501423,-0.2068341672,-0.087397106,-0.1178590804,0.7803842425,0.1436693668,0.3111174107,0.0839653313,-0.1246314198,-0.0810206234,-0.4270599484,-0.0110175544,0.1402982026,0.0720452443,0.3278112411,-0.1053301096,0.0581178516,0.2616516054,0.1077198237,0.0724793747,-0.361097306,-0.0051488164,0.3375664055,-0.154358536,-0.1817443669,0.3534454107,-0.118458584,0.0354432017,0.1111353859,0.2420222312,0.2134124488,0.1371783018,0.1566456258,-0.0693950504,-0.2307569981,-0.0652631,-0.07055749,-0.1846166998,0.2672117054,0.0754229575,-0.1197473258,-0.2136300653,0.1633310169,-0.0294777993,-0.3949569166,-0.3176743984,0.2599502504,-0.0235395152,-0.2561844289,-0.0813505948,0.0380744971,-0.1845514625,0.1519241631,-0.3738636672,-0.214790076,0.2328856885,-0.1929453909,0.1649006456,0.0228643306,0.2064365894,-0.1811629832,0.1503593177,-0.3148274124,-0.2083697468,0.1537433118,0.2667753696,0.3345374465,0.0792360231,-0.2752797604,-0.1176693216,0.0220045894,-0.0906073302,-0.1588032246,-0.1168166846,-0.1460376084,0.1377998143,0.3670161366,-0.2457677871,-0.0201742947,-0.1032267511,-0.1363053769,-0.1141639426,-0.1847239733,0.0239026826,-0.3247128427,0.3724550009,-0.3234480023,-0.016898822,-0.1468380839,0.2027985901,0.0441219062,0.3206821084,0.3426313698,0.3007251024,0.1799403131,0.0947604701,-0.2369461209,0.3371214867,0.1722479463,0.3267541826,0.1307581514,-0.0150973387,0.4982776344,0.280049324,0.0831123292,0.2850404978,-0.00626081,-0.0925764441,0.4295139015,0.1191856414,-0.1168936566,0.1504354626,0.1159252077,0.1750214845,0.0257161278,0.4962119758,0.3500537574,0.1437888592,-0.0675532892,-0.135002315,0.0826280117,-0.0661146939,-0.0666804761,-0.1447851807,0.0724013746,0.0437454917,0.0006443571,0.4399148822,0.0695736706,0.0663825572,-0.079854548,0.5231013894,0.2074706405,0.2747345269,0.1934806108,-0.2048034966,0.1035544798,0.2038632184,0.3004034758,0.1193206757,0.1591387242,0.354221344,0.1260891706,-0.1194892004,0.1709068269,0.2110641897,-0.0806493089,0.1110622808,-0.2004024684,-0.0725307167,-0.2158038169,0.0621195585,-0.2387410402,-0.1798122525,0.0963684544,0.1368214488,0.0067452267,-0.3859777451,0.4528263509,-0.0402101278,0.0091665788,-0.1356382519,0.305226773,-0.1014413163,0.0659833699,0.3549158275,0.2100659758,0.0990711153,0.2855636179,0.1362924427,0.2355001122,-0.2102711499,0.1438736469,-0.0006832101,0.4406963289,0.2373733371,0.2129765302,0.4536242485,0.0913677514,0.0018618176,0.1205106974,0.0514007322,-0.4279730022,-0.2207470834,0.5370098352,0.2514302433,-0.16851978,-0.1581876129,0.0029706836,-0.0860072151,-0.4035359323,0.5033171177,0.0157627575,0.0497165658,-0.2786099315,0.0830867589,0.0463866889,0.3376850188,0.3966375887,0.2780683637,-0.1849700511,0.0493984707,-0.3838036954,-0.0559991263,-0.1525279433,-0.2819629312,-0.2818736434,-0.1282318532,0.0995177478,0.3470939696,-0.1571502537,0.3710315228,0.0179469418,-0.3060075045,-0.2353046089,-0.1054626405,0.2477034926,0.2829425335,-0.1134491339,0.0560064167,-0.2223588228,-0.0812143832,0.0233647097,-0.4031130075,-0.0080643976,-0.068729341,0.1518035382,0.1368705183,0.0243091714,0.5706079006,0.0106302872,0.06802468,-0.2428244948,-0.1464605331,-0.1731858701,0.1491960436,0.1218693852,0.0671291053,-0.3864066601,0.1353756487,-0.0885000601,0.1795297712,-0.1564372331,-0.0657016262,-0.1479658037,0.2039208412,-0.4316804111,0.0830748081,0.1596536189,0.3315754533,0.0236460008,0.0015567733,0.0224986635,0.0170324966,0.1190907732,0.0040986449,-0.0942800343,0.3629531264,0.0592557192,0.2928446233,-0.0104444809,-0.6340758204,-0.1954901665,0.3198603094,-0.0335238576,0.1436676383,0.1693872064,-0.0985883921,-0.2704084516,-0.1893446147,0.0106363688,-0.21411933,-0.1467208862,0.1295877844,-0.1905368567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/302","title":"Question - Sign Language Datasets","comments":"Thanks, Quentin, I think a `requirements.txt` per dataset will be a good thing.\r\nI will work on adding this dataset next week, and once we sort all of the kinks, I'll add more.","body":"An emerging field in NLP is SLP - sign language processing.\r\n\r\nI was wondering about adding datasets here, specifically because it's shaping up to be large and easily usable.\r\nThe metrics for sign language to text translation are the same.\r\n\r\nSo, what do you think about (me, or others) adding datasets here?\r\n\r\n\r\nAn example dataset would be [RWTH-PHOENIX-Weather 2014 T](https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/)\r\nFor every item in the dataset, the data object includes:\r\n1. video_path - path to mp4 file\r\n2. pose_path - a path to `.pose` file with human pose landmarks\r\n3. openpose_path - a path to a `.json` file with human pose landmarks\r\n4. gloss - string\r\n5. text - string\r\n6. video_metadata - height, width, frames, framerate\r\n\r\n\r\n------\r\n\r\nTo make it a tad more complicated - what if sign language libraries add requirements to `nlp`? for example, sign language is commonly annotated using `ilex`, `eaf`, or `srt` files, which are all loadable as text, but there is no reason for the dataset to parse that file by itself, if libraries exist to do so.","comment_length":33,"text":"Question - Sign Language Datasets\nAn emerging field in NLP is SLP - sign language processing.\r\n\r\nI was wondering about adding datasets here, specifically because it's shaping up to be large and easily usable.\r\nThe metrics for sign language to text translation are the same.\r\n\r\nSo, what do you think about (me, or others) adding datasets here?\r\n\r\n\r\nAn example dataset would be [RWTH-PHOENIX-Weather 2014 T](https:\/\/www-i6.informatik.rwth-aachen.de\/~koller\/RWTH-PHOENIX-2014-T\/)\r\nFor every item in the dataset, the data object includes:\r\n1. video_path - path to mp4 file\r\n2. pose_path - a path to `.pose` file with human pose landmarks\r\n3. openpose_path - a path to a `.json` file with human pose landmarks\r\n4. gloss - string\r\n5. text - string\r\n6. video_metadata - height, width, frames, framerate\r\n\r\n\r\n------\r\n\r\nTo make it a tad more complicated - what if sign language libraries add requirements to `nlp`? for example, sign language is commonly annotated using `ilex`, `eaf`, or `srt` files, which are all loadable as text, but there is no reason for the dataset to parse that file by itself, if libraries exist to do so.\nThanks, Quentin, I think a `requirements.txt` per dataset will be a good thing.\r\nI will work on adding this dataset next week, and once we sort all of the kinks, I'll add more.","embeddings":[-0.1974006295,0.4022528231,-0.079109177,-0.0050231428,-0.0767897144,0.0436102264,0.11855492,0.2503882945,0.2235915363,-0.3281885386,0.1770670414,0.2271167487,-0.4437556565,0.2644597292,0.3408164084,-0.1764732599,-0.084280245,0.0804555714,-0.023515759,-0.0963775143,-0.2208419442,-0.091707252,0.0955999345,-0.0577017106,-0.4422942698,-0.332775116,-0.178117156,0.0741666704,-0.204051435,-0.3707835078,-0.4634328485,0.1567583084,0.4244990051,-0.1583760083,-0.000114478,-0.2469081879,0.0123989526,-0.0152046494,-0.2423187345,-0.1023321301,-0.1577181667,-0.7473182678,-0.1494281739,-0.1643093228,0.2529296577,-0.443055898,0.1376864612,-0.2573286295,0.0975727811,0.3461552858,0.2374007404,0.1211299673,-0.1563312709,0.2964705527,0.1227010116,0.3389026225,-0.0006764327,-0.3003685772,0.3348721862,-0.0234885812,-0.2104610056,0.2818766236,0.1933553666,-0.5736477375,0.2710379958,-0.0685172454,-0.2079850733,-0.0676307306,-0.066908814,0.0754446387,0.3453674912,-0.3600135446,-0.4973927438,-0.7469828725,0.2285517901,-0.092820555,-0.1126513183,0.3469284475,0.1550317854,-0.010782606,-0.0998056531,-0.67773664,0.0045277481,0.109010458,0.1395772099,0.3664364815,0.1597377658,-0.0140163815,0.0213861279,-0.0601192638,0.0780716687,-0.3396180868,-0.0897082761,0.1468113363,-0.2279662341,-0.2546849251,0.0456875637,0.1795903295,0.1582259536,0.1055883467,-0.0377676859,-0.0872028396,-0.349296093,0.0652103499,0.3826233447,-0.0524613671,0.0003892317,0.0625457689,0.028248854,-0.3815040588,0.3718284667,0.0065557933,-0.037904989,0.0085483519,-0.1868052334,-0.1111243367,0.1067432761,-0.0339871123,0.0236690547,0.1212499067,-0.1666444838,-0.2111101747,0.0759290904,-0.0982967243,0.0746060163,0.3548745215,0.0401231274,0.2771657705,0.2093126178,-0.2290028632,0.0685553253,0.172265321,-0.3457767665,0.2788127363,0.4381219447,0.0669046342,0.1423583031,0.0603194907,-0.198967129,0.061864052,0.3503090143,-0.0628166124,0.1613736153,-0.1447441131,0.0385888293,0.0365707166,-0.3883935511,-0.0259430632,-0.2925308347,0.119131498,0.1366383731,-0.0804761723,-0.0388111621,0.1629549414,-0.4021618664,-0.3853216469,-0.3291303515,0.6202965379,0.1007386521,-0.0582490191,0.1596972644,-0.1684682071,-0.2560831606,0.014038221,-0.2791418135,0.428856492,-0.2821624279,-0.1059000939,-0.1297005266,-0.0637688488,-0.2600839436,0.2443723083,-0.2492683977,0.2110896707,-0.0272822473,0.2751368582,0.507065475,-0.2396633178,0.0208983328,0.4749806821,-0.0797762647,0.0007162541,0.1055531129,0.3611428142,0.1435505599,-0.1722275317,-0.3293012679,0.323480159,-0.2784798145,-0.0933738053,-0.2011948675,-0.1771761924,0.4429148436,0.2742109001,0.1469441056,-0.0275893416,-0.0663536265,0.406054318,0.1270267963,-0.1566640139,0.340169251,-0.0624594204,0.0352788046,0.0136017017,-0.0927390233,-0.3162546158,-0.1547408402,-0.0304822903,0.0125783207,0.1696051657,0.2088414431,-0.1213485077,0.2039934695,-0.1071820706,0.2525424361,0.0955638066,0.111690782,0.0382408984,-0.1850869954,-0.1132753566,-0.3947124779,0.4281033874,-0.2745808959,-0.0476967953,-0.0114997318,0.0725866556,0.0295501482,0.1697958857,0.1357233077,0.5882997513,-0.2614141405,0.0273805503,0.2882312536,-0.1669662595,-0.17252101,0.1844573468,0.3149498701,0.3089549541,0.3211069405,-0.1385314614,0.1999233961,-0.1884872913,0.1615176499,-0.184594214,-0.0175686348,0.195334211,0.1894375086,0.109466061,0.1455932856,-0.0465479717,0.1012184545,-0.0567432679,-0.3234140575,-0.0316369124,0.2800018489,-0.1444944888,0.3401680589,0.2117850333,-0.3706463575,0.1214513555,0.3513514996,-0.0525903329,0.0116539979,0.2522886395,0.0804406628,-0.1467967629,0.0063111018,0.2170946747,0.4361246824,0.3528653085,0.0151232611,0.0879413038,-0.1288971901,-0.2453832626,0.0566704609,0.1276622266,-0.1084780023,0.1986526251,0.0719914362,0.1995277703,-0.2864269614,-0.2640377283,-0.0211481228,-0.0154110361,0.0046084379,0.073966369,0.049952101,-0.5297708511,-0.365387857,0.1866244078,-0.2011543959,-0.2340789884,0.1968279183,-0.3085400164,0.0091536632,0.0580987334,-0.3039529324,0.6582548022,-0.6731199026,-0.0403692015,-0.0351419821,-0.3434549868,-0.1652197093,0.1436959654,0.3100822866,0.1160766482,0.4283594787,0.065305002,0.2960450053,-0.2109526098,-0.5485614538,-0.1787150651,-0.3139559925,-0.1921653599,-0.1019282117,-0.1495178491,0.3705406785,-0.1683762223,-0.2032722682,0.0717448294,-0.1242419183,-0.3720764816,-0.0146370251,-0.3222450614,-0.135581851,-0.172160387,-0.2186243534,-0.3289655149,0.2249587029,0.1005765647,0.0016788457,-0.0216464605,-0.1645269245,0.2052087784,-0.1831042916,0.0600906536,-0.0031858417,0.1764277965,0.1062486097,0.0799410865,-0.309022516,0.0749585256,-0.1006588638,0.1314843148,0.2565287352,-0.525586009,0.1268876642,-0.1069988981,0.0641059577,-0.0696773678,-0.1026674882,0.1247734725,0.2839758098,0.0247925073,-0.1574664414,-0.4571686387,0.0672450736,0.1272676438,0.343780607,-0.139229387,0.0594512597,-0.0022461715,0.2647852302,0.2434586287,-0.2642883658,0.3422489464,0.2653169334,0.4194591939,0.1689602584,-0.176170364,0.3868718743,0.063046068,0.1371614635,0.2577235699,0.1927354038,-0.2039498091,-0.1808024943,0.1392495483,-0.0761858448,0.0204944592,0.4147284329,-0.3725945354,0.3546568155,0.1537199169,-0.195804581,-0.2766859531,-0.1765551567,-0.147179082,0.3364880979,0.3284614086,0.1707978398,-0.6410788298,0.0368303768,-0.3105720282,0.0462168418,0.2745634019,0.4319670796,-0.3112004697,0.097614482,-0.0557938591,0.036180906,-0.1055780798,-0.4134283066,-0.4755859971,-0.1296423674,0.1433908194,-0.1986414492,0.249585703,-0.0398073457,-0.020132076,0.3165594935,0.2909579277,-0.1088272259,-0.0914976075,0.0264580902,0.3308044076,-0.0296886265,-0.4027970433,0.1051179394,-0.1942285597,-0.0985469371,-0.1701009274,-0.0668628365,-0.1866215914,0.2033026069,0.1028683409,-0.1352116466,-0.2271438539,0.3835304081,-0.0793187842,-0.3464402854,0.1934305429,0.1189389303,-0.2414215207,0.3323133588,-0.0264916327,0.3513551056,-0.164599061,-0.2012975961,0.0291503556,-0.2196477354,0.6314375401,0.3542383611,0.0856016651,0.3629098237,0.0632553771,-0.3654575944,-0.3602575362,0.3150689006,0.2369967997,0.1942257136,-0.3291448653,-0.507660687,0.4011161327,0.1376362443,-0.0364062004,-0.0677825511,-0.1430400014,-0.2811064422,0.2408285886,-0.0081887534,0.7889874578,-0.1493361294,0.2496530563,-0.0765776485,-0.5114427209,0.6271572113,-0.1674836129,-0.0804274231,-0.0143442219,0.183827728,-0.0191932395,-0.0812127739,0.1340786666,0.3162007034,0.2134095281,-0.0462760478,0.2883575559,-0.0703793913,0.1002650261,0.8568924069,-0.3034603,-0.3963684738,-0.3659341037,0.1264202744,-0.2464094758,-0.0266110953,-0.0649897531,-0.2947364151,-0.1323821843,0.1267938912,-0.4832511544,-0.0693684742,0.3888428509,-0.2350245565,-0.1511719227,-0.076807946,0.7176423669,0.077208966,0.3300268054,0.0314688459,-0.1952888072,0.0241292454,-0.4325172901,-0.0863858089,0.1471736282,0.1219456568,0.140382871,-0.1454220116,0.023473328,0.2262597829,0.0684597045,-0.0145148085,-0.5050381422,0.0522782169,0.1950868368,-0.1231159568,-0.0689273849,0.5138173699,-0.0681847483,0.0207653455,0.1027641073,0.1838267893,0.1689915955,0.0913116261,0.1520513892,-0.0374575593,-0.2160595059,0.0429682545,-0.0753262043,-0.2015752196,0.2343836129,0.0330457613,-0.055522576,-0.1840870529,0.0751046911,-0.1497548521,-0.3243551552,-0.310466975,0.3003430963,-0.0759452656,-0.2433856577,-0.056804616,0.1733276695,-0.2264055163,0.2474827766,-0.2429959476,-0.3314934075,0.2082413584,-0.3118906021,0.1174037606,-0.0232219156,0.3572558165,-0.1480077505,0.2578612268,-0.2804976404,-0.1546102017,0.0882635266,0.3105075657,0.3099578321,-0.084576413,-0.268591553,-0.1762034893,0.0161962304,-0.148924619,-0.105613336,-0.0866901204,-0.0364282317,0.1263318658,0.4104903042,-0.1653853208,-0.0419939496,-0.1324810982,-0.0544064268,-0.0490908884,-0.2426356226,-0.0707253218,-0.2757899761,0.4180639386,-0.3472617865,-0.0658145696,-0.1480024904,0.1774568111,-0.0457438268,0.2827943265,0.2108148634,0.3214334548,0.315312922,0.2342091501,-0.1881673187,0.2993411422,0.0602922626,0.1830218434,0.1445746869,-0.0468505137,0.4479170442,0.2959773839,0.1411787421,0.2675271034,-0.0132851992,-0.1491369903,0.3432057798,0.1370578408,-0.2571586967,0.1818166971,0.0800358728,0.2518636882,0.06291008,0.4056019783,0.3765972257,0.1017413065,-0.0252517201,-0.0959125683,0.1464974433,-0.152830109,-0.0624994971,-0.0497926623,0.0969401971,-0.0138497055,0.1129713058,0.5702890158,0.0413039811,0.1716400832,-0.0898811072,0.502348721,0.1305228323,0.2156416029,0.2177160978,-0.2426010668,0.059822306,0.2534345388,0.2359832823,0.0247118622,0.0812095702,0.4312993288,0.0741241127,-0.1881437451,0.1945919693,0.2206702381,-0.2081833482,0.1194122657,-0.0904785767,-0.1270567924,-0.1531333476,0.0348230079,-0.1282280535,-0.2318030298,0.1473718733,0.0851196349,-0.0189014263,-0.5047826171,0.3316527903,0.0130494824,0.080071196,-0.0840639323,0.2831828594,-0.1745671183,-0.0947864056,0.4462602735,0.2171249986,0.0983691588,0.2315948755,0.060385447,0.3669403195,-0.2215574533,0.1218077913,-0.0624685809,0.5107766986,0.2789461315,0.1489985287,0.4225064516,0.0446111746,-0.0198277757,0.1922050118,0.0385442711,-0.3531961143,-0.1963355988,0.6884595156,0.1349356025,-0.1358707249,-0.2045065314,-0.0006397834,-0.0450138301,-0.3699133098,0.4495587945,-0.0729041919,0.0396340303,-0.248564899,0.0680037066,-0.0703242943,0.3127764761,0.3956477046,0.3038144112,-0.1383784413,0.066347383,-0.4806429148,-0.1293336004,-0.1926625967,-0.1880766153,-0.2800236344,-0.0714380816,0.0998372883,0.3237400055,-0.1876550317,0.4049898386,0.0694449916,-0.3457067609,-0.228866145,-0.052359771,0.3079678118,0.1860607564,-0.1223305836,0.114272289,-0.1667231023,0.0130047286,0.018354116,-0.3734303415,-0.0654889271,-0.0763705894,0.3184005916,0.0978388488,0.1062641665,0.5531413555,-0.0079616932,0.0114619276,-0.2154726535,-0.2639873922,-0.24832578,0.1283883899,0.101777792,0.0674709454,-0.371889174,0.3322956264,-0.027088834,0.1103986055,-0.0758958608,-0.0841120854,-0.2196467221,0.1999896318,-0.3208364248,0.1973865926,0.1238451675,0.3181504011,0.1143124476,0.0478019901,-0.0739382654,-0.0178604908,0.167812407,-0.0040872791,-0.0524947569,0.3718813062,0.1876739711,0.395350337,-0.0232653916,-0.7201074362,-0.3790508807,0.3896291554,-0.1420317739,0.1743503213,0.1914739013,-0.0830183104,-0.3334847689,-0.1527766138,-0.0322004221,-0.242893368,-0.1434964389,0.1909088939,-0.182506457]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/301","title":"Setting cache_dir gives error on wikipedia download","comments":"Whoops didn't mean to close this one.\r\nI did some changes, could you try to run it from the master branch ?","body":"First of all thank you for a super handy library! I'd like to download large files to a specific drive so I set `cache_dir=my_path`. This works fine with e.g. imdb and squad. But on wikipedia I get an error:\r\n```\r\nnlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=my_path)\r\n```\r\n```\r\nOSError Traceback (most recent call last)\r\n<ipython-input-2-23551344d7bc> in <module>\r\n 1 import nlp\r\n----> 2 nlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=path)\r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 385 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 386 reader = ArrowReader(self._cache_dir, self.info)\r\n--> 387 reader.download_from_hf_gcs(self._cache_dir, self._relative_data_dir(with_version=True))\r\n 388 downloaded_info = DatasetInfo.from_directory(self._cache_dir)\r\n 389 self.info.update(downloaded_info)\r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py in download_from_hf_gcs(self, cache_dir, relative_data_dir)\r\n 231 remote_dataset_info = os.path.join(remote_cache_dir, \"dataset_info.json\")\r\n 232 downloaded_dataset_info = cached_path(remote_dataset_info)\r\n--> 233 os.rename(downloaded_dataset_info, os.path.join(cache_dir, \"dataset_info.json\"))\r\n 234 if self._info is not None:\r\n 235 self._info.update(self._info.from_directory(cache_dir))\r\n\r\nOSError: [Errno 18] Invalid cross-device link: '\/home\/local\/NTU\/nn\/.cache\/huggingface\/datasets\/025fa4fd4f04aaafc9e939260fbc8f0bb190ce14c61310c8ae1ddd1dcb31f88c.9637f367b6711a79ca478be55fe6989b8aea4941b7ef7adc67b89ff403020947' -> '\/data\/nn\/nlp\/wikipedia\/20200501.de\/1.0.0.incomplete\/dataset_info.json'\r\n```","comment_length":22,"text":"Setting cache_dir gives error on wikipedia download\nFirst of all thank you for a super handy library! I'd like to download large files to a specific drive so I set `cache_dir=my_path`. This works fine with e.g. imdb and squad. But on wikipedia I get an error:\r\n```\r\nnlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=my_path)\r\n```\r\n```\r\nOSError Traceback (most recent call last)\r\n<ipython-input-2-23551344d7bc> in <module>\r\n 1 import nlp\r\n----> 2 nlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=path)\r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 385 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 386 reader = ArrowReader(self._cache_dir, self.info)\r\n--> 387 reader.download_from_hf_gcs(self._cache_dir, self._relative_data_dir(with_version=True))\r\n 388 downloaded_info = DatasetInfo.from_directory(self._cache_dir)\r\n 389 self.info.update(downloaded_info)\r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py in download_from_hf_gcs(self, cache_dir, relative_data_dir)\r\n 231 remote_dataset_info = os.path.join(remote_cache_dir, \"dataset_info.json\")\r\n 232 downloaded_dataset_info = cached_path(remote_dataset_info)\r\n--> 233 os.rename(downloaded_dataset_info, os.path.join(cache_dir, \"dataset_info.json\"))\r\n 234 if self._info is not None:\r\n 235 self._info.update(self._info.from_directory(cache_dir))\r\n\r\nOSError: [Errno 18] Invalid cross-device link: '\/home\/local\/NTU\/nn\/.cache\/huggingface\/datasets\/025fa4fd4f04aaafc9e939260fbc8f0bb190ce14c61310c8ae1ddd1dcb31f88c.9637f367b6711a79ca478be55fe6989b8aea4941b7ef7adc67b89ff403020947' -> '\/data\/nn\/nlp\/wikipedia\/20200501.de\/1.0.0.incomplete\/dataset_info.json'\r\n```\nWhoops didn't mean to close this one.\r\nI did some changes, could you try to run it from the master branch ?","embeddings":[-0.016413182,0.2276869714,-0.008926251,0.1092670187,0.122345984,0.199938044,0.2557855248,0.3317468464,0.4485464692,-0.1023546979,-0.1903628558,0.1178650558,0.0666657314,-0.5188595653,0.1406508833,-0.1102182716,0.0300002266,-0.0613132231,-0.0226511266,0.0516940393,-0.2747196257,0.2622623146,-0.0441821441,0.0540891103,-0.2433815002,-0.1050464287,-0.072632812,0.107462205,0.1136655733,-0.3596624136,0.4128053188,-0.0299529247,0.2771955132,0.4403319061,-0.000117424,-0.0152614946,0.3624263406,-0.2030952275,-0.5579511523,-0.2163559198,-0.2246102244,-0.2418591231,0.2105771154,-0.4761382043,0.3626435697,0.3921292424,0.3601877689,-0.2884105444,-0.1154452339,0.3803896308,0.1888131052,-0.2133142799,-0.0241495632,0.1325340867,0.4576618373,0.0064487583,-0.0460410677,0.4196564853,-0.5581924915,-0.182408005,-0.1978879273,0.3190628886,-0.3348386884,0.361063391,0.4996759593,0.1689672768,0.0551661104,-0.1576720774,0.1740521789,-0.0217134003,0.708702445,-0.3186961412,-0.1667527556,-0.1298629493,-0.1916632652,-0.3419360816,0.3887316585,0.3142914474,-0.2168809623,-0.1313343197,-0.0785200372,-0.1864174902,-0.2055503577,0.5084335208,0.0353473499,0.242180258,-0.0118902205,0.1077330783,0.3105280995,0.1559653878,-0.0152277891,-0.1883089542,0.0289224386,0.2680670321,-0.0917278454,0.0463293567,0.0164938383,0.1754599512,0.0469526201,-0.1358773112,-0.0364212319,-0.1584873199,-0.0180105586,0.2070783824,0.2671858966,0.4584434032,-0.2714021802,0.0725077316,0.345495373,0.2596298158,-0.1769945621,-0.0114095518,0.2153182775,-0.2061931342,0.0493750125,-0.0884282887,0.1682378054,-0.1910486072,-0.0727941915,-0.0134651763,-0.1205733865,-0.2002833188,0.0497518852,0.3913376331,-0.1108429283,0.0725893304,0.0393974856,0.3830533624,-0.2109767348,0.3658888936,-0.0863929614,0.1849439144,-0.0562552474,0.3640348613,0.3974510431,0.1643991321,0.2825456858,-0.2996708453,0.0105315642,-0.1080213264,0.0263676383,-0.355806917,0.0472795293,0.2118555009,0.1468363553,0.5156888962,0.2154467702,-0.1385783553,-0.3362908661,0.1985244751,-0.3842075467,-0.4099826813,0.0202297848,0.1066173166,-0.2228432149,-0.1802170724,-0.067409046,-0.0913380012,0.4154752791,-0.2722593844,0.2022363096,0.0881238207,-0.3097882867,-0.3220295906,0.0165564232,0.3973737955,0.0871979296,0.0317643508,-0.1866356879,0.471336633,0.2485871315,0.4971693456,-0.1658027619,0.3217129111,-0.2038130611,0.1477424651,0.5421157479,-0.408711195,-0.5469186902,0.2316524982,0.0064263502,-0.1150686443,0.1368109435,0.2783952951,0.1917880177,-0.0924685001,0.0588237122,0.5033121705,0.1440496892,0.2762487531,-0.4132539034,-0.1885228008,0.2845275402,-0.0427296385,-0.0707916096,0.0257842727,-0.0464463308,0.5719918013,0.133411184,-0.0220252387,0.0729474127,0.3717224896,0.0085802888,0.0894883573,0.0432725474,0.1852875799,-0.2929489315,0.0871187523,0.07798253,-0.0258016698,-0.1508039981,-0.0566889718,-0.206571281,-0.2405162305,-0.1476676166,-0.3601519167,0.1087893322,0.0835640505,0.4531356692,0.1844787002,0.0979959741,0.0464844815,0.1498785317,-0.0275073517,-0.1222449318,0.2475824654,-0.3175865114,-0.0543072,-0.2789750993,0.0616955161,0.0693085566,0.1027576625,-0.3141680062,0.2036349922,0.1215253547,-0.1842951477,0.0691569299,0.0348740034,0.0998132303,-0.1528588831,0.2042676359,0.4087335765,0.1507891715,0.0722956136,-0.4435012639,0.0004099495,-0.1792548448,0.1177941039,-0.0772794634,0.0483716987,0.3536260426,0.0317918509,0.3313835561,-0.0596574061,0.4391058087,0.2169996053,0.1532561779,-0.1297254562,0.0815455243,-0.129945308,0.1299858987,0.1706516743,0.2419709712,0.3518333435,-0.12397594,-0.3393210471,-0.0108792502,0.2468971163,0.4208874404,0.1096465737,0.2762743831,-0.1904606074,-0.0932282135,-0.1360231489,0.2923359871,-0.0277508367,0.0862574279,-0.0453641713,0.1577400416,-0.0386776738,-0.0647119582,-0.1363081187,0.0061430531,0.4186868668,-0.2352326363,0.2236887068,-0.0264421515,-0.5045306087,-0.1359294057,-0.1209725216,-0.2691159248,-0.3751654625,0.0155547122,0.5011117458,0.1319015622,0.0480404533,-0.2572091222,-0.0303094629,0.143494606,-0.4754090011,-0.0204601791,-0.3241183162,-0.1442911923,-0.0856804848,0.3514268994,-0.1224092692,0.0593446568,0.0404588059,-0.0966990665,-0.450017035,0.0011224836,-0.0597250126,0.0585074201,-0.1251749098,-0.0632037595,0.4381805658,-0.1269313544,-0.342823863,0.0628386065,-0.0527419336,0.1600717306,0.1398719102,0.102086544,0.2967457473,-0.1913740784,-0.0091445595,-0.1177046672,-0.3937448859,0.1796115637,0.1178223938,0.2193236202,0.5895373225,-0.0777764395,0.1191114113,0.1746857762,0.0949008465,-0.2395320535,-0.2428200841,0.5289345384,-0.1670523137,-0.1949594468,-0.0525024384,-0.0015181918,0.0463569909,0.1917174906,-0.7021474242,-0.1443182081,0.0001712659,0.3861824572,0.0606087372,0.0247652903,0.323287487,-0.2217648327,0.0259407721,0.0210828073,-0.2541020811,0.0653096586,0.2154113203,0.3576698303,-0.0258156843,0.2474928647,0.0925820619,0.7005705833,-0.0984533206,0.136724636,0.5910069942,0.2630674839,0.3123771846,-0.1081122383,-0.1415640563,0.0339756422,-0.2543646693,-0.412769258,0.2169578522,0.1130403504,-0.5909470916,-0.1157220006,0.0774082616,-0.0475133024,-0.268640101,0.1714813262,-0.1132490486,0.074770242,-0.0087889293,0.0085813748,-0.3124426007,-0.4371773601,0.0273659304,0.3386439979,0.0391051993,0.2308510542,0.0825516731,-0.1985098273,-0.4888904095,0.2085332423,0.1187444329,0.1058765799,-0.1161029488,-0.0153072672,0.0350529738,0.0316421539,0.2744382024,0.0489111133,0.1794004142,-0.0356769376,0.2467146814,-0.4296489656,-0.0889283419,0.0598028153,0.5444708467,0.3421790004,0.1669621319,-0.3141854107,0.0472568125,0.02781092,0.1755387485,-0.0761562288,-0.336455375,-0.2055232674,-0.2697413266,-0.3697783947,-0.121017769,-0.0007539184,0.0290363394,0.2287291139,0.3266450763,0.1051136404,0.1288546771,-0.2328955084,-0.143301174,0.2951718867,0.2628570199,-0.1047429964,0.0527541488,0.0472778603,-0.3289657831,0.259980917,-0.1244601086,-0.0068397038,0.0860870853,0.1059930101,-0.2923348248,0.0587152988,0.092934832,-0.1338738799,-0.0450499095,-0.1996911317,0.0666895136,0.1933498234,0.1434355676,-0.145165652,-0.1140734777,-0.3475617766,0.702922523,0.0235502366,0.1170054451,0.4338641167,-0.1011485681,-0.3010015488,0.1858596504,0.1176165119,0.7429116964,-0.0762345344,0.4683906436,0.1714242548,-0.0857286453,0.7115085125,-0.3196153045,0.3685220778,-0.2188479155,0.1506065428,0.0260210037,-0.1230166405,-0.3293698728,-0.0714216158,-0.4267324209,0.2769703865,-0.0165867973,0.0066245086,-0.0299112312,0.3736410141,-0.0640191138,-0.4104448855,-0.3334049582,0.0971938819,-0.3908355534,0.4360163808,-0.0453331247,0.0207806174,-0.1653063744,-0.0980777219,-0.3404096663,0.0587033071,-0.2888369262,0.6166497469,-0.3921460807,-0.3520502746,-0.289283067,0.3093366623,0.0379140303,-0.0305529032,-0.2429151237,-0.007158983,-0.2245877087,-0.0948365927,0.1090735421,-0.0716660768,0.0674552247,-0.1958344877,0.052010458,-0.1056766137,-0.1442223787,-0.4680242836,0.1711761355,0.0920493528,-0.0723904595,0.1589509696,-0.3950124681,-0.0522698909,-0.2481306195,-0.009481594,0.0887654722,0.0874714106,-0.0033819522,-0.0348755457,-0.1981602162,-0.3373417258,0.0506110825,0.4414917827,-0.0977906436,-0.0076964167,0.5240352154,-0.0147900209,-0.405535996,-0.0018626133,-0.1336486191,-0.7260406613,-0.2480150014,0.017976189,-0.2769371271,0.1718014628,-0.4101255238,-0.2733956277,0.2185526937,0.1418948025,-0.0173094645,-0.440484643,-0.3008424342,0.0739918873,-0.345628798,-0.0219845865,0.1455744952,-0.3908933401,0.3372860849,-0.0159876458,-0.1853238791,0.0363634117,-0.1945848763,0.0507463478,0.3406376541,-0.1769328266,0.1941384077,-0.0636530817,-0.000184961,-0.1336811334,0.0347101726,-0.168287009,-0.1124802679,0.1542429924,0.1287359595,-0.2547098994,-0.0174708925,-0.4300861955,0.1838490665,-0.1178447157,-0.2559455633,0.1882424206,-0.0693532974,0.1610838473,0.0240280256,0.0328600705,-0.2047186196,0.1565233469,-0.2831510603,0.2583550811,0.1093620881,0.3405154943,0.1600188911,0.0259883758,-0.6841739416,-0.0502973236,-0.2113686055,0.0981021598,0.0084515484,-0.3884676993,0.0061391103,0.2412358969,-0.0807253644,0.1798380464,-0.0610382408,-0.2179076374,0.3269448876,0.0505858921,0.0200486016,0.0322229341,0.3912015259,-0.1006629169,-0.1535358727,0.1829553545,0.0332769565,-0.085233897,0.0559886806,0.1523456126,0.2491252869,-0.0587820448,0.1242922768,0.0178924929,-0.0504797064,0.1344889998,0.0726617947,0.2149319053,0.2057393938,0.1494083107,0.0041573369,0.1710852981,0.2343474329,-0.3078148961,-0.1340903789,-0.1303470582,-0.1543121189,-0.1020756736,0.0668803304,0.0982456654,0.000094463,0.2079461366,-0.2135246992,0.1449691057,-0.0599694327,0.1155374274,-0.0978124291,-0.0659403205,0.0603346229,0.0263001993,0.1375108063,0.1075834855,-0.1826429963,-0.167969346,0.2871863544,0.3152040839,-0.1891338974,-0.2231884748,0.4620463848,0.1735917777,0.0819258243,-0.1670199186,0.1947743297,-0.0585565791,-0.1475930363,-0.1181575581,0.1586479843,0.515963316,0.4165537655,-0.4197716117,-0.0678833574,0.0299602803,0.0922663659,-0.0381526537,-0.0678887144,0.2026346624,-0.0942471176,0.2924463153,0.0378864706,-0.0711590424,0.2384535223,0.1849591136,0.0534991622,0.1574970335,-0.0276708361,0.0916943476,0.0433514044,0.0144207617,0.1833705753,-0.361844182,0.2814033031,0.1314228624,0.0616641343,0.028415842,0.1367622465,0.0209254082,0.1875360608,0.5410298109,0.2292957455,0.268312186,-0.5026729107,0.0878733024,-0.5525556207,0.2962695062,-0.3617638946,-0.1734833866,-0.1476112306,0.3638663888,-0.339184016,0.1839359999,0.0175153743,-0.101767607,0.0994249433,0.16353257,-0.2389966249,-0.1956478059,-0.1113411188,0.0518909879,-0.0803839266,-0.4228535891,0.0348168202,0.0322074816,0.0673615336,-0.0769157857,-0.0576009341,0.0864319727,0.5537905693,0.1799575388,-0.0772539452,0.4416600168,0.0061583705,-0.2031337619,-0.3520434201,-0.036679294,-0.1375380307,-0.2156297565,-0.0738635361,0.2568704486,-0.3955132961,-0.0910812616,-0.4942489266,0.7418962121,-0.1484315097,-0.0339943878,0.2366783619,-0.261036396,0.0082721096,0.2684926987,0.1910718828,0.0464204773,-0.1763649136,0.3552641869,-0.4896112382,-0.2099740058,0.3372948766,-0.486291796,-0.3294306397,0.0979153812,0.262027353,0.1855799407,0.1044572368,-0.3378351033,0.0451950394,0.2677620947,0.044468306,-0.4406727552,-0.1964246482,-0.0974905267,0.1088973358,-0.0088242134,0.1406001449,0.0823216587,-0.304253608,-0.1703522056,-0.2422835976]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/301","title":"Setting cache_dir gives error on wikipedia download","comments":"Now it works, thanks!","body":"First of all thank you for a super handy library! I'd like to download large files to a specific drive so I set `cache_dir=my_path`. This works fine with e.g. imdb and squad. But on wikipedia I get an error:\r\n```\r\nnlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=my_path)\r\n```\r\n```\r\nOSError Traceback (most recent call last)\r\n<ipython-input-2-23551344d7bc> in <module>\r\n 1 import nlp\r\n----> 2 nlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=path)\r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 385 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 386 reader = ArrowReader(self._cache_dir, self.info)\r\n--> 387 reader.download_from_hf_gcs(self._cache_dir, self._relative_data_dir(with_version=True))\r\n 388 downloaded_info = DatasetInfo.from_directory(self._cache_dir)\r\n 389 self.info.update(downloaded_info)\r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py in download_from_hf_gcs(self, cache_dir, relative_data_dir)\r\n 231 remote_dataset_info = os.path.join(remote_cache_dir, \"dataset_info.json\")\r\n 232 downloaded_dataset_info = cached_path(remote_dataset_info)\r\n--> 233 os.rename(downloaded_dataset_info, os.path.join(cache_dir, \"dataset_info.json\"))\r\n 234 if self._info is not None:\r\n 235 self._info.update(self._info.from_directory(cache_dir))\r\n\r\nOSError: [Errno 18] Invalid cross-device link: '\/home\/local\/NTU\/nn\/.cache\/huggingface\/datasets\/025fa4fd4f04aaafc9e939260fbc8f0bb190ce14c61310c8ae1ddd1dcb31f88c.9637f367b6711a79ca478be55fe6989b8aea4941b7ef7adc67b89ff403020947' -> '\/data\/nn\/nlp\/wikipedia\/20200501.de\/1.0.0.incomplete\/dataset_info.json'\r\n```","comment_length":4,"text":"Setting cache_dir gives error on wikipedia download\nFirst of all thank you for a super handy library! I'd like to download large files to a specific drive so I set `cache_dir=my_path`. This works fine with e.g. imdb and squad. But on wikipedia I get an error:\r\n```\r\nnlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=my_path)\r\n```\r\n```\r\nOSError Traceback (most recent call last)\r\n<ipython-input-2-23551344d7bc> in <module>\r\n 1 import nlp\r\n----> 2 nlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=path)\r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 385 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 386 reader = ArrowReader(self._cache_dir, self.info)\r\n--> 387 reader.download_from_hf_gcs(self._cache_dir, self._relative_data_dir(with_version=True))\r\n 388 downloaded_info = DatasetInfo.from_directory(self._cache_dir)\r\n 389 self.info.update(downloaded_info)\r\n\r\n~\/anaconda3\/envs\/fastai2\/lib\/python3.7\/site-packages\/nlp\/arrow_reader.py in download_from_hf_gcs(self, cache_dir, relative_data_dir)\r\n 231 remote_dataset_info = os.path.join(remote_cache_dir, \"dataset_info.json\")\r\n 232 downloaded_dataset_info = cached_path(remote_dataset_info)\r\n--> 233 os.rename(downloaded_dataset_info, os.path.join(cache_dir, \"dataset_info.json\"))\r\n 234 if self._info is not None:\r\n 235 self._info.update(self._info.from_directory(cache_dir))\r\n\r\nOSError: [Errno 18] Invalid cross-device link: '\/home\/local\/NTU\/nn\/.cache\/huggingface\/datasets\/025fa4fd4f04aaafc9e939260fbc8f0bb190ce14c61310c8ae1ddd1dcb31f88c.9637f367b6711a79ca478be55fe6989b8aea4941b7ef7adc67b89ff403020947' -> '\/data\/nn\/nlp\/wikipedia\/20200501.de\/1.0.0.incomplete\/dataset_info.json'\r\n```\nNow it works, thanks!","embeddings":[-0.016413182,0.2276869714,-0.008926251,0.1092670187,0.122345984,0.199938044,0.2557855248,0.3317468464,0.4485464692,-0.1023546979,-0.1903628558,0.1178650558,0.0666657314,-0.5188595653,0.1406508833,-0.1102182716,0.0300002266,-0.0613132231,-0.0226511266,0.0516940393,-0.2747196257,0.2622623146,-0.0441821441,0.0540891103,-0.2433815002,-0.1050464287,-0.072632812,0.107462205,0.1136655733,-0.3596624136,0.4128053188,-0.0299529247,0.2771955132,0.4403319061,-0.000117424,-0.0152614946,0.3624263406,-0.2030952275,-0.5579511523,-0.2163559198,-0.2246102244,-0.2418591231,0.2105771154,-0.4761382043,0.3626435697,0.3921292424,0.3601877689,-0.2884105444,-0.1154452339,0.3803896308,0.1888131052,-0.2133142799,-0.0241495632,0.1325340867,0.4576618373,0.0064487583,-0.0460410677,0.4196564853,-0.5581924915,-0.182408005,-0.1978879273,0.3190628886,-0.3348386884,0.361063391,0.4996759593,0.1689672768,0.0551661104,-0.1576720774,0.1740521789,-0.0217134003,0.708702445,-0.3186961412,-0.1667527556,-0.1298629493,-0.1916632652,-0.3419360816,0.3887316585,0.3142914474,-0.2168809623,-0.1313343197,-0.0785200372,-0.1864174902,-0.2055503577,0.5084335208,0.0353473499,0.242180258,-0.0118902205,0.1077330783,0.3105280995,0.1559653878,-0.0152277891,-0.1883089542,0.0289224386,0.2680670321,-0.0917278454,0.0463293567,0.0164938383,0.1754599512,0.0469526201,-0.1358773112,-0.0364212319,-0.1584873199,-0.0180105586,0.2070783824,0.2671858966,0.4584434032,-0.2714021802,0.0725077316,0.345495373,0.2596298158,-0.1769945621,-0.0114095518,0.2153182775,-0.2061931342,0.0493750125,-0.0884282887,0.1682378054,-0.1910486072,-0.0727941915,-0.0134651763,-0.1205733865,-0.2002833188,0.0497518852,0.3913376331,-0.1108429283,0.0725893304,0.0393974856,0.3830533624,-0.2109767348,0.3658888936,-0.0863929614,0.1849439144,-0.0562552474,0.3640348613,0.3974510431,0.1643991321,0.2825456858,-0.2996708453,0.0105315642,-0.1080213264,0.0263676383,-0.355806917,0.0472795293,0.2118555009,0.1468363553,0.5156888962,0.2154467702,-0.1385783553,-0.3362908661,0.1985244751,-0.3842075467,-0.4099826813,0.0202297848,0.1066173166,-0.2228432149,-0.1802170724,-0.067409046,-0.0913380012,0.4154752791,-0.2722593844,0.2022363096,0.0881238207,-0.3097882867,-0.3220295906,0.0165564232,0.3973737955,0.0871979296,0.0317643508,-0.1866356879,0.471336633,0.2485871315,0.4971693456,-0.1658027619,0.3217129111,-0.2038130611,0.1477424651,0.5421157479,-0.408711195,-0.5469186902,0.2316524982,0.0064263502,-0.1150686443,0.1368109435,0.2783952951,0.1917880177,-0.0924685001,0.0588237122,0.5033121705,0.1440496892,0.2762487531,-0.4132539034,-0.1885228008,0.2845275402,-0.0427296385,-0.0707916096,0.0257842727,-0.0464463308,0.5719918013,0.133411184,-0.0220252387,0.0729474127,0.3717224896,0.0085802888,0.0894883573,0.0432725474,0.1852875799,-0.2929489315,0.0871187523,0.07798253,-0.0258016698,-0.1508039981,-0.0566889718,-0.206571281,-0.2405162305,-0.1476676166,-0.3601519167,0.1087893322,0.0835640505,0.4531356692,0.1844787002,0.0979959741,0.0464844815,0.1498785317,-0.0275073517,-0.1222449318,0.2475824654,-0.3175865114,-0.0543072,-0.2789750993,0.0616955161,0.0693085566,0.1027576625,-0.3141680062,0.2036349922,0.1215253547,-0.1842951477,0.0691569299,0.0348740034,0.0998132303,-0.1528588831,0.2042676359,0.4087335765,0.1507891715,0.0722956136,-0.4435012639,0.0004099495,-0.1792548448,0.1177941039,-0.0772794634,0.0483716987,0.3536260426,0.0317918509,0.3313835561,-0.0596574061,0.4391058087,0.2169996053,0.1532561779,-0.1297254562,0.0815455243,-0.129945308,0.1299858987,0.1706516743,0.2419709712,0.3518333435,-0.12397594,-0.3393210471,-0.0108792502,0.2468971163,0.4208874404,0.1096465737,0.2762743831,-0.1904606074,-0.0932282135,-0.1360231489,0.2923359871,-0.0277508367,0.0862574279,-0.0453641713,0.1577400416,-0.0386776738,-0.0647119582,-0.1363081187,0.0061430531,0.4186868668,-0.2352326363,0.2236887068,-0.0264421515,-0.5045306087,-0.1359294057,-0.1209725216,-0.2691159248,-0.3751654625,0.0155547122,0.5011117458,0.1319015622,0.0480404533,-0.2572091222,-0.0303094629,0.143494606,-0.4754090011,-0.0204601791,-0.3241183162,-0.1442911923,-0.0856804848,0.3514268994,-0.1224092692,0.0593446568,0.0404588059,-0.0966990665,-0.450017035,0.0011224836,-0.0597250126,0.0585074201,-0.1251749098,-0.0632037595,0.4381805658,-0.1269313544,-0.342823863,0.0628386065,-0.0527419336,0.1600717306,0.1398719102,0.102086544,0.2967457473,-0.1913740784,-0.0091445595,-0.1177046672,-0.3937448859,0.1796115637,0.1178223938,0.2193236202,0.5895373225,-0.0777764395,0.1191114113,0.1746857762,0.0949008465,-0.2395320535,-0.2428200841,0.5289345384,-0.1670523137,-0.1949594468,-0.0525024384,-0.0015181918,0.0463569909,0.1917174906,-0.7021474242,-0.1443182081,0.0001712659,0.3861824572,0.0606087372,0.0247652903,0.323287487,-0.2217648327,0.0259407721,0.0210828073,-0.2541020811,0.0653096586,0.2154113203,0.3576698303,-0.0258156843,0.2474928647,0.0925820619,0.7005705833,-0.0984533206,0.136724636,0.5910069942,0.2630674839,0.3123771846,-0.1081122383,-0.1415640563,0.0339756422,-0.2543646693,-0.412769258,0.2169578522,0.1130403504,-0.5909470916,-0.1157220006,0.0774082616,-0.0475133024,-0.268640101,0.1714813262,-0.1132490486,0.074770242,-0.0087889293,0.0085813748,-0.3124426007,-0.4371773601,0.0273659304,0.3386439979,0.0391051993,0.2308510542,0.0825516731,-0.1985098273,-0.4888904095,0.2085332423,0.1187444329,0.1058765799,-0.1161029488,-0.0153072672,0.0350529738,0.0316421539,0.2744382024,0.0489111133,0.1794004142,-0.0356769376,0.2467146814,-0.4296489656,-0.0889283419,0.0598028153,0.5444708467,0.3421790004,0.1669621319,-0.3141854107,0.0472568125,0.02781092,0.1755387485,-0.0761562288,-0.336455375,-0.2055232674,-0.2697413266,-0.3697783947,-0.121017769,-0.0007539184,0.0290363394,0.2287291139,0.3266450763,0.1051136404,0.1288546771,-0.2328955084,-0.143301174,0.2951718867,0.2628570199,-0.1047429964,0.0527541488,0.0472778603,-0.3289657831,0.259980917,-0.1244601086,-0.0068397038,0.0860870853,0.1059930101,-0.2923348248,0.0587152988,0.092934832,-0.1338738799,-0.0450499095,-0.1996911317,0.0666895136,0.1933498234,0.1434355676,-0.145165652,-0.1140734777,-0.3475617766,0.702922523,0.0235502366,0.1170054451,0.4338641167,-0.1011485681,-0.3010015488,0.1858596504,0.1176165119,0.7429116964,-0.0762345344,0.4683906436,0.1714242548,-0.0857286453,0.7115085125,-0.3196153045,0.3685220778,-0.2188479155,0.1506065428,0.0260210037,-0.1230166405,-0.3293698728,-0.0714216158,-0.4267324209,0.2769703865,-0.0165867973,0.0066245086,-0.0299112312,0.3736410141,-0.0640191138,-0.4104448855,-0.3334049582,0.0971938819,-0.3908355534,0.4360163808,-0.0453331247,0.0207806174,-0.1653063744,-0.0980777219,-0.3404096663,0.0587033071,-0.2888369262,0.6166497469,-0.3921460807,-0.3520502746,-0.289283067,0.3093366623,0.0379140303,-0.0305529032,-0.2429151237,-0.007158983,-0.2245877087,-0.0948365927,0.1090735421,-0.0716660768,0.0674552247,-0.1958344877,0.052010458,-0.1056766137,-0.1442223787,-0.4680242836,0.1711761355,0.0920493528,-0.0723904595,0.1589509696,-0.3950124681,-0.0522698909,-0.2481306195,-0.009481594,0.0887654722,0.0874714106,-0.0033819522,-0.0348755457,-0.1981602162,-0.3373417258,0.0506110825,0.4414917827,-0.0977906436,-0.0076964167,0.5240352154,-0.0147900209,-0.405535996,-0.0018626133,-0.1336486191,-0.7260406613,-0.2480150014,0.017976189,-0.2769371271,0.1718014628,-0.4101255238,-0.2733956277,0.2185526937,0.1418948025,-0.0173094645,-0.440484643,-0.3008424342,0.0739918873,-0.345628798,-0.0219845865,0.1455744952,-0.3908933401,0.3372860849,-0.0159876458,-0.1853238791,0.0363634117,-0.1945848763,0.0507463478,0.3406376541,-0.1769328266,0.1941384077,-0.0636530817,-0.000184961,-0.1336811334,0.0347101726,-0.168287009,-0.1124802679,0.1542429924,0.1287359595,-0.2547098994,-0.0174708925,-0.4300861955,0.1838490665,-0.1178447157,-0.2559455633,0.1882424206,-0.0693532974,0.1610838473,0.0240280256,0.0328600705,-0.2047186196,0.1565233469,-0.2831510603,0.2583550811,0.1093620881,0.3405154943,0.1600188911,0.0259883758,-0.6841739416,-0.0502973236,-0.2113686055,0.0981021598,0.0084515484,-0.3884676993,0.0061391103,0.2412358969,-0.0807253644,0.1798380464,-0.0610382408,-0.2179076374,0.3269448876,0.0505858921,0.0200486016,0.0322229341,0.3912015259,-0.1006629169,-0.1535358727,0.1829553545,0.0332769565,-0.085233897,0.0559886806,0.1523456126,0.2491252869,-0.0587820448,0.1242922768,0.0178924929,-0.0504797064,0.1344889998,0.0726617947,0.2149319053,0.2057393938,0.1494083107,0.0041573369,0.1710852981,0.2343474329,-0.3078148961,-0.1340903789,-0.1303470582,-0.1543121189,-0.1020756736,0.0668803304,0.0982456654,0.000094463,0.2079461366,-0.2135246992,0.1449691057,-0.0599694327,0.1155374274,-0.0978124291,-0.0659403205,0.0603346229,0.0263001993,0.1375108063,0.1075834855,-0.1826429963,-0.167969346,0.2871863544,0.3152040839,-0.1891338974,-0.2231884748,0.4620463848,0.1735917777,0.0819258243,-0.1670199186,0.1947743297,-0.0585565791,-0.1475930363,-0.1181575581,0.1586479843,0.515963316,0.4165537655,-0.4197716117,-0.0678833574,0.0299602803,0.0922663659,-0.0381526537,-0.0678887144,0.2026346624,-0.0942471176,0.2924463153,0.0378864706,-0.0711590424,0.2384535223,0.1849591136,0.0534991622,0.1574970335,-0.0276708361,0.0916943476,0.0433514044,0.0144207617,0.1833705753,-0.361844182,0.2814033031,0.1314228624,0.0616641343,0.028415842,0.1367622465,0.0209254082,0.1875360608,0.5410298109,0.2292957455,0.268312186,-0.5026729107,0.0878733024,-0.5525556207,0.2962695062,-0.3617638946,-0.1734833866,-0.1476112306,0.3638663888,-0.339184016,0.1839359999,0.0175153743,-0.101767607,0.0994249433,0.16353257,-0.2389966249,-0.1956478059,-0.1113411188,0.0518909879,-0.0803839266,-0.4228535891,0.0348168202,0.0322074816,0.0673615336,-0.0769157857,-0.0576009341,0.0864319727,0.5537905693,0.1799575388,-0.0772539452,0.4416600168,0.0061583705,-0.2031337619,-0.3520434201,-0.036679294,-0.1375380307,-0.2156297565,-0.0738635361,0.2568704486,-0.3955132961,-0.0910812616,-0.4942489266,0.7418962121,-0.1484315097,-0.0339943878,0.2366783619,-0.261036396,0.0082721096,0.2684926987,0.1910718828,0.0464204773,-0.1763649136,0.3552641869,-0.4896112382,-0.2099740058,0.3372948766,-0.486291796,-0.3294306397,0.0979153812,0.262027353,0.1855799407,0.1044572368,-0.3378351033,0.0451950394,0.2677620947,0.044468306,-0.4406727552,-0.1964246482,-0.0974905267,0.1088973358,-0.0088242134,0.1406001449,0.0823216587,-0.304253608,-0.1703522056,-0.2422835976]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/297","title":"Error in Demo for Specific Datasets","comments":"Thanks for reporting these errors :)\r\n\r\nI can actually see two issues here.\r\n\r\nFirst, datasets like `natural_questions` require apache_beam to be processed. Right now the import is not at the right place so we have this error message. However, even the imports are fixed, the nlp viewer doesn't actually have the resources to process NQ right now so we'll have to wait until we have a version that we've already processed on our google storage (that's what we've done for wikipedia for example).\r\n\r\nSecond, datasets like `newsroom` require manual downloads as we're not allowed to redistribute the data ourselves (if I'm not wrong). An error message should be displayed saying that we're not allowed to show the dataset.\r\n\r\nI can fix the first issue with the imports but for the second one I think we'll have to see with @srush to show a message for datasets that require manual downloads (it can be checked whether a dataset requires manual downloads if `dataset_builder_instance.manual_download_instructions is not None`).\r\n\r\n","body":"Selecting `natural_questions` or `newsroom` dataset in the online demo results in an error similar to the following.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/60150701\/85347842-ac861900-b4ae-11ea-98c4-a53a00934783.png)\r\n","comment_length":165,"text":"Error in Demo for Specific Datasets\nSelecting `natural_questions` or `newsroom` dataset in the online demo results in an error similar to the following.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/60150701\/85347842-ac861900-b4ae-11ea-98c4-a53a00934783.png)\r\n\nThanks for reporting these errors :)\r\n\r\nI can actually see two issues here.\r\n\r\nFirst, datasets like `natural_questions` require apache_beam to be processed. Right now the import is not at the right place so we have this error message. However, even the imports are fixed, the nlp viewer doesn't actually have the resources to process NQ right now so we'll have to wait until we have a version that we've already processed on our google storage (that's what we've done for wikipedia for example).\r\n\r\nSecond, datasets like `newsroom` require manual downloads as we're not allowed to redistribute the data ourselves (if I'm not wrong). An error message should be displayed saying that we're not allowed to show the dataset.\r\n\r\nI can fix the first issue with the imports but for the second one I think we'll have to see with @srush to show a message for datasets that require manual downloads (it can be checked whether a dataset requires manual downloads if `dataset_builder_instance.manual_download_instructions is not None`).\r\n\r\n","embeddings":[-0.1981071234,0.4033780098,-0.0345742553,0.0916168988,-0.0609801449,-0.040471375,0.2643227577,0.3957089484,-0.0206499416,0.1149799004,0.0365383849,0.1268178821,-0.1433795094,0.1707723886,0.192117855,-0.4499635398,-0.0038807646,0.1249553859,-0.0355561934,-0.069319725,-0.1204564869,0.1653505713,-0.4725416005,0.2649035454,-0.0176118817,-0.2618624568,-0.0384727344,0.0392732657,-0.3994920552,-0.4657357633,0.0809272528,0.1968013793,0.1599261612,-0.0005180874,-0.0001131029,0.0382400304,0.4839167297,-0.0483166315,-0.5484042764,-0.1696925312,-0.1286290139,-0.3020377755,0.1021560729,-0.0430534966,0.0262297224,-0.6082082391,0.1922516376,-0.0785224736,0.3578095734,0.5196115375,0.2201360911,0.5076093078,0.0161029883,-0.3282738924,0.4393441975,0.294282943,-0.1518373042,0.1352204829,0.0477910787,-0.0510030426,-0.0364651866,0.3086336255,0.0257014297,0.1527827829,-0.0602618046,-0.1421571821,-0.3206353188,-0.4986156225,-0.1197519824,0.4459383488,0.6633608341,0.0743716136,-0.1835986972,-0.1939502209,0.1495735347,0.0478801057,0.0631624535,0.4371371269,-0.2138037831,-0.0197302848,-0.3787341416,-0.2751147747,-0.2470450699,0.440015614,0.0848660022,0.0987464115,-0.0264036208,0.2691401839,-0.0508009307,-0.1801290363,0.1278465986,-0.0587071739,0.054145664,0.0016125643,-0.4455092251,-0.1112408042,-0.0311796758,0.040599037,0.0372829102,0.1945744157,0.1294780672,-0.011735267,0.0395428501,0.2793652713,0.2831104398,0.3266352713,0.3146273494,0.3872416914,0.4034557343,0.0530774109,0.1779122651,0.2051409781,-0.1392932832,-0.3384784162,-0.5065163374,0.0323855057,0.3175575137,-0.2519685328,-0.5763885975,0.2063839883,-0.2966667712,-0.0094444947,-0.0547546633,0.3192156851,0.0911048874,0.0232136268,0.1775573045,0.4780529439,-0.2147263139,-0.711591959,0.0356920063,0.3353013098,-0.2933809161,0.1421969831,0.2954668403,0.3753692508,0.1908024848,0.0171957389,0.1530253291,-0.0349088386,0.3172849715,-0.0082977293,-0.0913595706,0.2614539564,0.1109581515,0.1435957998,0.0197853297,-0.0449606888,-0.1087214723,0.1304396093,-0.0174404085,-0.4220673144,-0.0218519904,0.1434950233,-0.1796658039,0.0001322241,-0.1549212188,0.3701185286,-0.1066482663,-0.2454149276,-0.0933342949,-0.1988843232,0.0739992037,-0.0620109551,0.0609777942,0.4929059148,-0.3572953045,-0.014104777,-0.2872876227,-0.073183164,0.1695153862,0.3137457669,-0.3498011529,0.2884297073,-0.3124473095,0.0360883139,0.6690323353,-0.2652320564,-0.1361675262,0.3889521658,0.1521363407,-0.209536463,-0.1556265503,-0.0827601254,0.1384109259,-0.1652575582,-0.3314625621,0.5041031837,-0.0445608459,0.0315623507,-0.3019278347,-0.2324235886,0.4035744369,0.4479389489,0.1155449823,0.039159961,0.0323922746,0.0754637197,0.019848315,0.0247900635,0.061422646,0.1102890596,0.0182904135,-0.2559670806,-0.046059832,0.1362552643,-0.3674871325,0.0775811672,-0.5094981194,0.3944992721,0.1208044663,-0.0041109659,-0.4444130957,-0.0959618986,-0.2176892757,-0.2749752998,0.1779189706,0.0647481307,0.1631305218,0.2521955669,0.0554000326,0.0935483426,-0.1057463735,0.0657625347,-0.2442820966,0.2586836219,-0.1812973171,-0.0616905056,0.2920763493,-0.1391947269,0.3943233192,-0.0689119548,-0.1508922875,0.3955641091,-0.1765989959,0.2377078533,0.2862996459,-0.1709468663,0.2178450078,-0.5033503771,0.158449173,0.1945571303,-0.0532759465,0.2152076811,-0.1912836134,0.3722969592,0.0492303893,0.0986861661,0.0538784787,0.4147879481,0.1812629551,-0.0780160427,-0.0885619968,-0.1992299408,0.2647400796,0.4113313556,0.0834801346,-0.244946301,-0.4166576564,0.0249638725,0.1811139286,-0.0482652709,0.2014044374,0.0623043552,0.067928195,-0.1010516658,0.4000891447,0.4113735855,0.3827387989,0.0318042599,0.047577437,0.1044109687,-0.2516432405,-0.3253677487,0.1144971028,0.0414313301,0.1646135747,0.114533633,0.1352063566,0.0400645733,-0.364893198,0.0471885949,0.361200422,0.2409502119,-0.2400978506,-0.0440332964,-0.4294613004,-0.5175081491,-0.1684653759,0.1302052736,-0.256873101,-0.540941298,0.379493624,0.0211482923,-0.4098774493,0.2334162444,0.1367081851,0.4418497682,-0.2019059062,0.4963968992,-0.0761230364,-0.0833589584,-0.3440181017,0.0976666585,0.3494063318,0.3973334432,0.4992439747,-0.1747453958,-0.0280649159,-0.4055843949,-0.3387173712,0.2234098464,0.1596518457,0.5951380134,-0.094484508,0.3512040079,-0.0552666597,-0.1646458358,0.1619519144,-0.1427626908,-0.0991604999,0.1441089809,-0.048195865,-0.0626619682,-0.2645437717,-0.6977814436,-0.4090589285,-0.4599585533,-0.2028884143,-0.1979246885,0.2438067645,0.233697772,-0.1097362265,-0.0175722465,0.0205496885,-0.0083965752,-0.1718058437,-0.0689448118,0.2719592154,-0.6593863368,-0.4282494783,0.0517175868,0.1068603173,0.1274673343,0.2793439329,-0.4582374096,0.0910539702,-0.1234815419,-0.1770926565,0.1160819754,-0.285520792,0.2548550665,-0.1953578293,-0.0440754741,-0.1461909562,-0.0373638235,0.0814282671,0.1666160375,0.2665866911,-0.0415649414,0.1822211295,0.0857862011,0.6566123962,0.2117146999,0.2254574299,0.5008478761,0.1887433827,0.3019886017,0.0633157566,0.1152487695,0.2509574294,-0.1392772943,0.0063596871,0.2092837095,-0.1221354231,-0.2425933778,-0.2960003018,0.2020198256,-0.3664731383,-0.2441666424,0.0956480205,-0.376596719,0.273622185,0.1494428813,-0.0745368227,0.2203223407,-0.0372570492,-0.1145704314,0.0814961344,0.243890211,0.1516800374,-0.5252455473,-0.0177630242,-0.2078651935,0.1415897906,-0.2651933134,0.6571936607,-0.3658405542,0.0038033004,0.103701666,0.1937462389,0.1175400391,-0.574888885,-0.1773446947,0.0385760218,0.0863676071,-0.4662769735,-0.2195714563,-0.3589482009,0.2049835026,0.0694776401,0.009078742,-0.3030896187,-0.1169041097,0.2541904151,0.0720313713,-0.0167576913,-0.1788928807,0.176734224,0.0978535339,-0.6125366092,-0.0765250549,-0.0403105579,-0.0741792023,-0.5367669463,0.0563098229,-0.0627236441,0.3256813884,0.0540901795,0.1125026122,0.1614133716,0.0182712842,-0.1359257102,0.118976295,0.1313613355,0.1471993327,0.3836542964,0.2623980641,-0.1001564711,0.1386063099,0.0536006168,-0.0071892329,0.6157602668,-0.0630247071,0.0111972904,-0.2287762463,0.098449491,0.0687738582,0.2590809166,0.1843771338,-0.4804312885,-0.136100769,-0.1868962198,0.3830971718,-0.1327648908,-0.2401434332,0.0129684145,0.005827575,-0.0356359445,0.2383591235,0.2463853359,1.0048420429,-0.0366891362,-0.0580973141,0.1096551865,-0.1480350196,0.5812598467,-0.1890586466,0.0243846495,-0.2451957315,-0.0613478124,0.0328001454,-0.0935575068,0.2910012007,0.1977647543,-0.4215864241,0.3042161763,0.1378444433,0.0149517739,-0.0448677614,0.1483917087,-0.1559313387,0.0169657543,-0.3573397994,0.1450378597,-0.2827439606,0.3498194516,-0.1727623641,-0.2722786963,-0.1499145925,-0.2056366056,-0.4191603363,0.2725084424,0.0186253414,0.1489472836,-0.0192661565,-0.2983196676,0.3597486615,0.4683237672,0.227259174,0.0411649607,-0.387283653,0.1479939371,-0.2452413291,-0.171693787,0.0424132757,0.0621139221,0.3522135913,-0.1797749102,-0.1785710603,0.1292779297,-0.1743715256,0.0613239743,-0.5083681345,-0.1256967187,0.3128819466,-0.0220857058,-0.0847223923,-0.1698158681,-0.1195576116,-0.1661525369,0.1224929243,0.0256287716,-0.0317761488,-0.0418570042,0.2421578318,-0.0308180563,0.0293046944,0.09251789,0.109559007,0.0804674327,0.3666873872,0.1818652451,-0.0637805611,-0.2809884846,0.181927681,0.13144584,-0.5027119517,0.1247771978,-0.2719728649,-0.0659671724,0.130744487,0.1952537,-0.0860158801,-0.0579278693,0.0905041769,-0.5099039078,-0.058102034,0.1989432573,-0.3042751551,0.1795779318,-0.1578614712,-0.1306471378,0.219164446,0.2721655369,-0.2921165228,-0.0975070819,-0.453220427,0.2679082155,-0.030494675,0.1292979568,-0.2409978211,0.0211091768,-0.1047429815,-0.1094215289,-0.3323505223,-0.1012883335,-0.0497397929,0.1616226584,0.3272652626,-0.1539123207,0.1339713782,-0.1522318572,0.1495538205,-0.013419087,0.0836758912,-0.1746895164,-0.1118207797,0.2551409006,0.0492250919,0.1118847057,-0.0186915826,0.1929567158,-0.0048646191,0.1877949387,0.1860693693,0.1059036851,0.2087068558,-0.1594392061,-0.4427710176,0.0332291871,0.0594918393,0.211483568,0.1299265325,-0.2598473132,0.3293116987,-0.1000092328,0.2081318349,0.2163577825,-0.3254671395,0.058816772,0.245834589,0.1340456307,-0.2870177031,-0.0178881548,0.0739801824,0.1056993231,-0.1536821425,-0.0619853176,0.4100386798,-0.0365211852,0.1999367476,0.0675371811,0.0827364624,-0.0851633996,0.3317163885,0.3098627627,-0.2236455232,0.0956706926,-0.0871709734,-0.0253126416,0.0420190804,0.0752343014,-0.1770765483,0.5282194018,0.2105175108,0.1798924506,0.4375527203,-0.3083563149,0.1126559526,0.4501841068,0.2951728106,0.1711338907,-0.0317279622,0.1150983796,-0.1164854467,-0.2579783499,-0.3172601759,0.1533726901,-0.0162426997,0.1317508221,-0.3802255392,-0.1947349906,-0.4711546302,0.0005563481,-0.0924248844,-0.2355914861,0.0471298322,0.0281307232,-0.3189848661,-0.6951646209,0.4120100737,0.1451242268,-0.0403121188,-0.201300323,0.2788971364,-0.0809678137,-0.1254065633,0.2645362616,0.4675653279,0.3362561762,0.4843823016,-0.1837614477,-0.0637231767,0.0368750729,0.0664663464,0.1376530528,0.0835364088,0.2619531155,0.0336950049,0.4762539566,0.0290664434,-0.0878242403,0.1557623446,0.2547259629,-0.03855123,-0.3124664426,0.1196742728,-0.2325899303,-0.0463379994,-0.0239864402,-0.2767800987,-0.3721568584,0.0665058941,0.1047416627,0.0164745674,0.1063959673,-0.4064308405,0.0641500577,-0.2264083475,0.6622602344,0.2785651982,0.264800787,-0.2889125645,-0.4203690588,-0.6882609129,0.1602888554,-0.0588793643,0.0852314904,0.2896304131,0.0950779542,0.1803101301,0.3617956936,-0.1563683897,0.2062145621,-0.1581948549,0.2250670791,-0.5480691195,-0.1541029513,-0.025679145,0.0663804114,-0.0899209529,-0.1367210895,0.1477311254,-0.0781787261,-0.0748638362,0.1385719627,-0.0447308943,0.1264448762,-0.033868216,0.4230632186,0.2661769688,0.5350294709,-0.0470936298,-0.1777535081,-0.1697889566,-0.2589526474,-0.1855103225,0.03117867,-0.0058573135,0.4107149243,0.0844550058,0.2602502108,-0.0681210607,-0.0036673832,-0.0222858638,0.2091827244,-0.1971850544,-0.1468544006,0.0094436118,0.1340765357,0.1453395486,-0.0731143355,-0.0483140722,0.0692163333,-0.3202484846,-0.2129823416,0.3976902366,-0.2018700093,-0.0935952663,0.0902116969,0.0843887776,0.0418250524,-0.1034202874,-0.4077992141,-0.064644523,0.1281601936,-0.1711639613,0.2017368823,0.1048664823,0.1690252125,-0.216597572,0.033037886,0.2552473247,0.0999631807,-0.2482897341,0.1486691535,-0.1618630737]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/297","title":"Error in Demo for Specific Datasets","comments":"I added apache-beam to the viewer. We can think about how to add newsroom. ","body":"Selecting `natural_questions` or `newsroom` dataset in the online demo results in an error similar to the following.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/60150701\/85347842-ac861900-b4ae-11ea-98c4-a53a00934783.png)\r\n","comment_length":14,"text":"Error in Demo for Specific Datasets\nSelecting `natural_questions` or `newsroom` dataset in the online demo results in an error similar to the following.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/60150701\/85347842-ac861900-b4ae-11ea-98c4-a53a00934783.png)\r\n\nI added apache-beam to the viewer. We can think about how to add newsroom. ","embeddings":[-0.3623429239,0.3009083867,-0.1162245274,-0.0324354358,0.0338272192,0.0242333543,0.2281297296,0.5185450912,-0.2168404758,0.2707802057,0.0630017743,0.036783956,-0.1188950613,0.0918900743,0.0908572748,-0.5335237384,0.0483457111,0.159782216,-0.1572801918,0.0543620661,-0.1289060116,0.0827646703,-0.4261883199,-0.0367698073,0.2244355381,-0.0536911935,0.0499440432,-0.0661230162,-0.2885765731,-0.3081310391,-0.0638132542,0.0017089768,-0.0695624053,0.1742803305,-0.0001052855,0.1647574604,0.4729024172,-0.0856304616,-0.319884032,-0.182856068,-0.0905942023,-0.0869245231,-0.0627100542,-0.050597623,-0.2193138897,-0.5875648856,0.168865025,-0.1970734,0.243577987,0.4494462907,0.2764870822,0.5245394707,-0.041165363,-0.3444510996,0.4281018078,0.2209647447,-0.324834764,-0.0052569159,0.1535055041,0.1334927976,-0.0392677486,0.410176456,0.2203124166,0.1388502717,-0.1566029489,-0.2136279047,-0.2606455684,-0.3286661804,0.0987082496,0.3112487793,0.5066688657,0.2114200443,0.1001504436,0.0953961536,0.0426682346,0.2376235425,-0.0002131298,0.3652146459,-0.3464745283,0.0018490097,-0.2157098502,-0.2062072158,-0.3920075893,0.4749272764,0.1907761991,0.1885902435,-0.0241194703,0.3781839609,-0.0090242792,-0.145168975,0.2207816392,0.0598816574,0.0647076517,-0.0859767944,-0.3202159703,-0.0029358061,-0.0569291674,0.0523197949,0.1211097464,0.1935017705,0.0160511415,0.0754169151,0.2046560496,0.1370635182,0.1956804693,0.2097136676,0.3343025446,0.3300239742,0.273989886,0.2336458564,0.0414491519,0.2486662269,-0.0735867694,-0.4597388804,-0.4802018702,-0.018059032,0.3435602486,-0.1571822315,-0.4540173113,0.1431483775,-0.1942049712,0.0090669161,-0.0985812172,0.4441694915,0.2508785725,0.1666858792,0.1835840642,0.4457639456,-0.1455017626,-0.6824918389,-0.0345748849,0.3889251053,-0.256134212,0.1625041366,0.2663961649,0.2427647114,0.1842086315,-0.070424132,0.1929592937,0.0156153198,0.2662349641,0.1207852364,0.0541140735,0.2540975213,0.1881237924,0.0845803767,0.0427188836,0.110017024,0.0897782072,0.2652072906,0.0761895627,-0.3332854211,-0.1166523695,0.1980212033,-0.0604243763,0.1225272641,0.0246667024,0.2610880733,-0.0985624939,-0.0708850399,-0.0954760686,0.0009642757,0.1687113941,-0.0407631844,0.0654056817,0.1548583806,-0.3124711514,0.2060682923,-0.1431952417,-0.0834169015,-0.0170815885,0.0291447397,-0.3622713983,0.1193842068,-0.2883768976,-0.0621791519,0.6603797674,-0.4686746299,-0.2051570266,0.2564142048,0.114558056,-0.2622223794,-0.0501458831,-0.0445790179,0.0764862746,-0.0085375207,-0.4026421607,0.2980052233,-0.0710549504,0.0574286059,-0.2191120237,-0.0664185658,0.3306861222,0.4803160131,0.2118498683,-0.0788347274,0.0377376527,0.0624519885,0.1333312988,-0.0869454592,0.1017175168,0.1980014443,0.1445952356,-0.3197898865,0.1529982239,-0.0701811686,-0.2009867877,-0.0084703732,-0.2972554564,0.3554427624,0.1504304558,0.0292556603,-0.525724411,-0.0371090546,-0.3052665293,-0.3648298979,0.2602067292,0.1595058739,-0.1385194063,0.1160908714,-0.0250084363,0.0525438003,-0.0541093722,-0.011948308,-0.2600813508,0.2742295861,-0.3030941486,-0.01165058,0.0852732956,-0.3488719761,0.3879089653,-0.2243597209,-0.305000484,0.5158025622,-0.0814987868,0.108517617,0.2646965384,-0.2140788138,0.0415734947,-0.6912126541,-0.000953956,0.1348267794,-0.0670914277,0.2601461709,-0.1416931599,0.0994022042,0.068166934,-0.0096712997,-0.0344449617,0.2783018649,0.2556334734,-0.138050884,0.0617240854,-0.2102879882,0.2494916022,0.2797809243,0.1084106266,-0.3202016354,-0.1885743737,-0.0773749352,0.0678195804,-0.0815589428,0.1741874218,0.0923500508,-0.1428790092,-0.1356107891,0.3731648028,0.2647998929,0.3713067472,0.0772122592,-0.1247442812,0.0602696426,-0.0800545663,-0.2166391462,0.1959565878,0.0388894491,0.3644139469,0.1782515645,0.1294572651,-0.1243834421,-0.4658937752,0.0410244763,0.247364834,0.1951269209,-0.1468995512,-0.0830669999,-0.2777916193,-0.3351710737,-0.0049498035,0.265312463,0.0115139009,-0.4119620919,0.4411451221,0.0451128744,-0.365075171,0.346390605,0.1065148711,0.2986083329,0.129397735,0.4744853973,-0.164458409,0.0863720775,-0.2023964226,0.1801735908,0.1969937384,0.1798449457,0.3981493115,-0.1571900845,0.0293326899,-0.543558836,-0.3691233993,0.257466197,0.1747429669,0.6795602441,0.0108832894,0.412422806,-0.0294348616,-0.1636982709,0.2195400596,-0.1174150556,-0.1785393208,0.1683798879,-0.0887914225,-0.2071157843,-0.2177865654,-0.5987266302,-0.1472931057,-0.4713504612,-0.2494814247,-0.0316274501,0.0408827662,0.1347648203,-0.0535102822,-0.1954322457,0.0472554453,0.0263493042,-0.3268038034,-0.0720232502,0.2251747102,-0.6916334629,-0.4978787601,-0.0295852534,0.0896753147,0.2314048558,0.2394470274,-0.3884697855,-0.028992746,0.0730315372,-0.2170680463,-0.0736569166,-0.2897116244,0.2310525775,-0.2257317007,-0.1571763605,-0.3630778193,0.0681398213,0.126051113,-0.0217405502,0.2511957288,-0.0557682477,0.0883072466,0.2543094456,0.7101435065,0.1096762493,0.1082784086,0.5768988729,-0.0149545344,0.2890913785,-0.063432768,0.0362518914,0.2678189874,0.0242620818,0.0779329091,0.2236642838,-0.1455107033,-0.2455453426,-0.3396669924,0.1709852517,-0.4035922885,-0.1133110821,-0.006332614,-0.1859242618,0.2207345515,0.1539418697,0.0100953486,0.1253578216,-0.0983336195,0.0966916904,0.0934348106,-0.0212582275,-0.004981834,-0.3707380295,0.1460265964,-0.2536875606,0.3355291486,-0.2228789181,0.557528317,-0.2508241534,-0.1189875379,0.0808894187,0.0306016281,0.3592997789,-0.6540563107,-0.0445851013,-0.0415691324,0.0945316255,-0.4348349571,-0.181246683,-0.2166664898,0.0986945182,-0.0665634274,0.1067984402,-0.1103434935,-0.0362197272,0.3595634401,-0.0638353229,-0.0339543112,-0.2546600401,0.1403944492,0.082800962,-0.5122399926,-0.2728426158,-0.003166795,0.0195119362,-0.2317467928,-0.1375415474,0.0185078196,0.1421114206,-0.1097459793,0.1520959884,0.1334134638,-0.0753636807,-0.076086998,0.0723843127,0.1765175611,0.1515869498,0.2707836926,0.309251368,-0.2188845426,0.2426971048,-0.0551181734,-0.0749349222,0.4301261604,-0.1235091761,-0.1117789671,-0.153553009,0.1609404236,0.1388329268,0.423569411,0.2100700289,-0.4108397365,-0.089979589,0.003335407,0.3074577749,-0.1785590053,-0.2501660585,0.1935059726,-0.1663736999,-0.029223118,0.2041820884,0.1902505159,0.8807116747,-0.0149098495,-0.1561986804,0.0951144174,-0.192697987,0.2310751081,-0.3896405399,0.0339262113,-0.2682411671,-0.3249900341,0.0050605042,-0.1100000516,0.3294169009,0.0979960412,-0.2539628744,0.1824405491,0.1471115053,0.2180142701,-0.1446362883,0.0335878544,-0.0929208174,0.1644198149,-0.1864789128,0.2132768035,-0.2278141975,0.214994967,-0.2022428811,-0.1177483723,-0.1734209508,-0.2141223103,-0.3320559263,0.2936701775,-0.0337805487,0.272256881,-0.0458591729,-0.2479580492,0.1384177506,0.4140980244,0.0551914498,-0.0248644426,-0.3544671834,0.4893522561,-0.0692789406,-0.1991543025,-0.0263211261,0.0955645517,0.2840194106,-0.3197496831,-0.198846817,0.1112514362,-0.1758234799,-0.0773689076,-0.3750770986,-0.1158508882,0.5806762576,-0.1152786613,-0.1510683894,-0.2359441221,-0.1044986397,-0.2858364582,0.1887188107,-0.2110757381,-0.1610310972,-0.2885418832,0.3168165982,-0.0322963484,0.1040810421,0.0581443571,0.1063775867,-0.0431313738,0.2312036455,0.1638109535,-0.0754529908,-0.3730278313,0.3407803178,0.3266140819,-0.556599319,0.0061547216,-0.1750511825,-0.1513467133,0.3143404722,0.2544751465,-0.0039683138,-0.1125159785,-0.0055154702,-0.3408062756,-0.0000768057,0.0424053222,-0.292670995,0.2844547033,-0.1021705717,-0.1304452568,0.1832727194,0.1594712734,-0.3756566346,-0.211932525,-0.3385868967,0.2786122561,-0.1814442128,0.1245421544,-0.2101648748,0.1457734853,0.1080007255,-0.0682267398,-0.362959981,-0.2117476165,-0.0268329587,0.1186683849,0.2476547211,-0.2157118469,-0.002138207,-0.0269962549,-0.0378832892,0.0523147397,0.0929763466,-0.2464913428,-0.0698743835,0.0346027128,0.0573871396,-0.0890415385,0.0716868266,0.0649105832,-0.0332227461,0.331902802,0.1317518353,0.0483181849,0.1674661338,-0.2225241661,-0.4281189442,0.0090397084,0.1810574383,0.2750798166,0.0664852187,-0.2213718444,0.2245073169,-0.0512755997,0.3039289117,-0.0774134472,-0.4922519028,0.1696482897,0.1524526775,0.2675626576,-0.3546560705,-0.0748686343,0.1480820328,-0.0103072105,0.0134367989,-0.1244990751,0.25838992,0.0321143083,0.274078995,0.0293150675,0.0791089833,-0.0558249354,0.3594169915,0.2140268087,-0.2696737051,0.0619308501,-0.1574441344,-0.1210918054,0.2031831294,0.3114805222,-0.1015325934,0.5991835594,0.308200866,0.1559717208,0.409091711,-0.0732291117,0.2296034396,0.3326820135,0.1529065669,0.2550998032,0.1194175631,0.0192269459,0.065058127,-0.2710794806,-0.336486876,0.1416061074,0.0506398976,0.1363090873,-0.3025844991,-0.1810855567,-0.3955387473,-0.0085002165,0.0344318599,-0.2121089548,0.1137886271,0.0778332278,-0.1955097318,-0.2754269242,0.4487147033,0.1953806281,-0.0530361719,-0.1281651706,0.3577865362,-0.0144904228,-0.0715004206,0.3132150173,0.3449337482,0.2004213333,0.3004429042,-0.1662529856,-0.1240843311,0.1289983392,-0.0480430909,0.0998752266,-0.0761870518,0.243853271,0.0680228397,0.4719704688,0.1620904654,-0.2017084956,0.1421660483,0.331371367,-0.0275845975,-0.1691557616,-0.2577398717,-0.3578885198,0.0864160433,-0.1673452407,-0.2924648821,-0.3883420825,-0.0244349409,-0.1067918912,-0.1783203483,0.0984557569,-0.2929445803,0.1001924947,-0.223522082,0.4852666855,0.0011606137,0.2253392041,-0.1295037717,-0.6446211338,-0.4436828494,0.3517274857,0.0365034565,-0.1001022533,0.1760151684,-0.0150731718,0.2836758196,0.2285196334,0.3561014831,0.1722804457,-0.0371525548,0.356086731,-0.578962326,-0.0420900732,0.170110479,0.1700550467,0.0137204155,-0.1882076412,0.066900745,-0.0401038341,0.0884415507,0.1754565388,-0.2579244375,0.1676590294,-0.3121231794,0.5520763993,0.0037828712,0.4563382864,-0.2645252943,-0.1578644514,-0.0423481725,-0.2787791491,-0.1499988884,-0.0033783205,0.1585431248,0.3546237946,0.0814399943,0.0896243677,0.0608374663,0.0370592736,0.0411487296,0.1684162617,-0.2945582569,0.0320342183,-0.0837303326,0.1510302722,0.1648328304,-0.217610985,-0.1133730114,-0.0237656906,-0.2496639043,-0.2587513924,0.5243250728,-0.105070278,-0.303271234,0.1458043754,-0.0888774842,-0.0263658166,-0.0098181888,-0.4287532866,-0.0963772833,0.1240788549,0.0447013713,0.2255605906,0.0760767385,0.0453674383,-0.1278154999,0.0928347483,0.2171385884,0.2121539265,-0.2124197632,0.1185542941,-0.1737919599]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/297","title":"Error in Demo for Specific Datasets","comments":"We don't plan to host the source files of newsroom ourselves for now.\r\nYou can still get the dataset if you follow the download instructions given by `dataset = load_dataset('newsroom')` though.\r\nThe viewer also shows the instructions now.\r\n\r\nClosing this one. If you have other questions, feel free to re-open :)","body":"Selecting `natural_questions` or `newsroom` dataset in the online demo results in an error similar to the following.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/60150701\/85347842-ac861900-b4ae-11ea-98c4-a53a00934783.png)\r\n","comment_length":51,"text":"Error in Demo for Specific Datasets\nSelecting `natural_questions` or `newsroom` dataset in the online demo results in an error similar to the following.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/60150701\/85347842-ac861900-b4ae-11ea-98c4-a53a00934783.png)\r\n\nWe don't plan to host the source files of newsroom ourselves for now.\r\nYou can still get the dataset if you follow the download instructions given by `dataset = load_dataset('newsroom')` though.\r\nThe viewer also shows the instructions now.\r\n\r\nClosing this one. If you have other questions, feel free to re-open :)","embeddings":[-0.3872511983,0.3188091815,-0.1640232354,-0.0211658124,0.013869918,-0.0003127987,0.2441099882,0.4248589873,-0.1058463231,0.250238657,0.076460585,0.0662668496,-0.1547729522,0.198998034,0.1576560885,-0.3058364987,0.034139093,0.106728889,0.0080381259,0.0768803954,0.0208985768,0.1381000876,-0.3802264631,0.0807855427,0.0555668212,0.070566453,-0.0612187684,0.1005788371,-0.3853769004,-0.2824477255,0.1910808682,0.1865963787,0.1066664681,0.2580335736,-0.0000968782,0.1013725176,0.3720093668,-0.0740894452,-0.3853502274,-0.245289579,-0.1144347638,-0.1170114875,-0.1747405678,-0.0778490603,-0.3093830645,-0.3932517171,0.1433220655,-0.1347689331,0.3808894753,0.5134056211,0.3451370895,0.3197131455,0.0641769618,-0.3982389867,0.2445176691,0.3347768784,-0.2866628468,0.2551392317,0.1351358891,-0.0103753284,0.1773409843,0.269554466,0.131067425,0.1185721233,-0.1659903228,-0.0658473969,-0.1224851683,-0.3944908082,0.1921982467,0.3081759214,0.4608187079,0.2094991356,-0.040614184,0.0182142761,0.0734412521,0.0097016767,0.0295343827,0.3768169284,-0.2834703922,0.1123722047,-0.2613152564,-0.3204788268,-0.3009408116,0.3629044294,0.0215595607,0.2622927129,-0.2182945013,0.2114819884,-0.120443739,-0.1118349358,0.1694317311,0.05978873,-0.0276937988,-0.0231773928,-0.086862646,-0.0712509826,0.0011098736,-0.0449180752,0.2504673302,0.1148645133,-0.0322841145,-0.0790894106,0.1651591212,0.0767433271,0.2547540367,0.1350129694,0.3106298149,0.4236186445,0.2886647284,0.3024428189,-0.063519381,0.0966585651,-0.1747273952,-0.3139651716,-0.2929464281,0.045825161,0.4366629422,-0.3682164252,-0.371797353,0.1540751308,-0.0505330339,0.127332136,-0.1139572337,0.2986803651,0.1218006909,0.0631627142,0.0573716387,0.4153439999,-0.1190730631,-0.4704567492,-0.1539423019,0.1874102801,-0.191104874,0.0657421201,0.3246703446,0.1526779085,0.2951782942,0.0576418191,0.2079437971,-0.088049382,0.364474535,0.1267647743,0.1516768783,0.2503456771,0.1138969734,0.0518857576,0.0352837481,-0.0781417191,0.038864594,0.2460532188,-0.0280619748,-0.4773592949,-0.1569186747,0.3296296299,-0.0704427809,0.0011815069,-0.0395828821,0.2608208954,-0.0548480302,-0.1282824725,-0.1434522122,-0.0872637182,0.1408842802,0.0421402678,0.213459298,0.2267336845,-0.4083755612,0.2936526239,-0.1156534776,-0.0255732946,0.17785725,0.0088781919,-0.4150947332,0.2261175364,-0.302259326,-0.0209094603,0.4404123724,-0.3577588797,-0.2100269496,0.0945873335,-0.0135330679,-0.262396723,-0.230164513,-0.1978841275,-0.0164954923,0.0489151813,-0.1584012955,0.3208877146,0.0073946668,0.1409168094,-0.2430867702,-0.1350844502,0.2573469579,0.3529310524,0.014426575,-0.0176272206,-0.0714033917,0.0495391861,0.1344290674,-0.1043756902,-0.0211508982,0.1938442439,0.1335433871,-0.230402112,-0.0459466614,-0.043356806,-0.243736729,0.0554044768,-0.1315820813,0.3995983303,0.1226348802,0.0420893133,-0.4179421663,-0.037828166,-0.2738242149,-0.5199094415,0.3302708268,0.2020736635,0.0733386427,0.0772274882,-0.1083025038,-0.054993432,-0.0550580621,-0.0631411448,-0.2199083716,0.1707648486,-0.2846331596,-0.0056708436,0.096703425,-0.3065670431,0.283233881,-0.1726743877,-0.2059425861,0.4193875194,-0.0198369939,0.0575467013,0.4182517827,-0.1971727014,0.0525954291,-0.4988575876,-0.0080730114,0.1652863473,-0.1155634671,0.249667421,-0.1285469234,0.2141994238,0.1445358843,-0.0236836411,-0.0095277978,0.2160440236,0.3346383572,-0.0974224582,0.0177045129,-0.1007091701,0.1811771393,0.2839544415,0.0924012437,-0.3739606738,-0.3749496341,0.1001344174,0.0793851018,-0.031919498,0.0861245245,0.0888968855,-0.074501656,-0.2598748207,0.2304984331,0.3278220296,0.4170775115,0.0774272606,0.2484454662,0.1300415397,-0.1720896661,-0.2319050729,0.2817389071,-0.1592697948,0.1291060299,0.1429406404,0.1644272059,-0.0998535901,-0.4131441116,0.0557120442,0.1864568591,0.1960126013,-0.188638255,-0.0089930519,-0.2010096908,-0.1755679697,0.1071400717,0.0476272851,-0.0659389049,-0.3055223525,0.4321427941,0.0323332883,-0.3655504584,0.1023598313,-0.0179279149,0.361975044,-0.0456877425,0.4299207032,-0.0444134697,-0.0463662557,-0.3281881809,0.2481746972,0.1598792672,0.184389025,0.3590472937,-0.2736911774,-0.2064645886,-0.4784511328,-0.2711727321,0.2871100903,0.2367945313,0.7645660639,0.0121669192,0.4614914954,-0.112027809,-0.1231091172,0.2972155213,-0.1684675217,-0.0643439963,0.0680412129,-0.0657359734,-0.2628162205,-0.2710855603,-0.6054618359,-0.2764243782,-0.2533033788,-0.0740088224,-0.0343925059,0.0808572546,0.3402635157,0.0184953418,-0.0887746066,-0.0075627905,0.0359902084,-0.4659329355,-0.43918401,0.1614693701,-0.7218429446,-0.4966351688,0.0258448478,0.2140137106,0.2463016808,0.053470064,-0.3788907528,-0.0630356818,0.0697180703,-0.0572842136,0.0916902423,-0.1630955338,0.2149606198,-0.2943845689,-0.2422394454,-0.2595788836,0.0747232065,0.1373461038,0.0962980464,0.1859037131,-0.2009457797,-0.0745711029,0.0676011592,0.6230891347,0.085409604,0.0992675722,0.3722313046,-0.1327311397,0.3359413147,-0.1386457533,-0.0050208149,0.1669450253,-0.0674061403,-0.149818331,0.2395372689,0.0259458069,-0.3132219613,-0.3852425218,0.2405520976,-0.3015001714,-0.1989703625,-0.0538906008,-0.0208727792,0.2608452737,0.2406805903,-0.0016866003,0.1040038168,-0.074603036,0.0243886262,0.057032831,0.0445708111,0.0031950118,-0.6048046947,0.0267211851,-0.3515688479,0.1591261178,-0.2577735186,0.4119017422,-0.2407067865,-0.0085330531,0.0138827693,-0.0864780024,0.1769098788,-0.5102113485,-0.0276702773,0.0149352141,0.1109923273,-0.1924203485,-0.2355775386,-0.274546653,0.2641191781,0.07852301,0.184619382,-0.142986998,-0.1287338585,0.2427589744,-0.0134870009,0.0207486525,-0.0805817246,0.0630601719,-0.0967873707,-0.3988783658,-0.0911620483,0.0446013175,-0.0463696681,-0.3687047362,-0.1156627461,0.0793376714,0.3159848452,-0.085884355,0.1336700767,0.202038765,0.0479650833,-0.0114965755,0.2605047524,0.0677075014,0.1446578056,0.3633370697,0.2215911597,-0.081919916,0.1926191747,-0.1205545366,-0.0424277186,0.5388208628,-0.1751190871,-0.2817951739,-0.1047949791,0.141799733,0.0087282341,0.3472476304,0.2992350459,-0.3133321404,0.020970583,-0.2378170937,0.1973052621,-0.1814101487,-0.1668504328,0.1333498657,-0.1642518044,-0.0373821557,0.0062748301,0.1289378256,0.8680382371,0.1448338032,0.0411031619,0.1716700047,-0.1048961878,0.2332078218,-0.1474513263,-0.0045340457,-0.2011489421,-0.2723900378,0.0424751677,-0.0872676671,0.1309123188,0.1768802553,-0.2160094529,0.1602906734,0.0324721076,0.0871359184,-0.1521574855,0.0969389156,-0.0012679123,0.2295019031,-0.1676352918,0.3235172927,-0.0564787649,0.1517691165,-0.2461147308,-0.0918712988,-0.1434954256,-0.1628860384,-0.3564282954,0.2898541689,-0.0516658165,0.154307723,0.0112391692,-0.1803199202,-0.0429256484,0.3814529479,0.1313355267,0.1240388155,-0.3673484623,0.2598138452,-0.0172449257,-0.2312887758,0.0838156939,0.1056737602,0.1600036621,-0.3114282489,-0.2772390246,0.0913434699,-0.1630017906,-0.0109796459,-0.3528248966,-0.1768669486,0.2859664857,-0.1401664913,-0.2297171652,-0.1381976455,-0.1162360162,-0.2525806129,0.2790488601,-0.0838594437,-0.223707363,-0.3463155329,0.3273395002,-0.169348672,-0.0576971956,0.0563041084,0.0639566034,0.0882272571,0.370128423,0.2061975747,-0.062180236,-0.4052585363,0.1842390746,0.2291396707,-0.393409133,0.0596216545,-0.1618469656,-0.0213870481,0.2566961348,0.2476495653,-0.0533656552,-0.2109645009,0.0578687415,-0.3806421757,-0.0182860903,-0.0595317632,-0.1610081643,0.2897926867,-0.0500007309,-0.2112185657,0.1830375046,0.0295494013,-0.4641394317,-0.058506839,-0.339315623,0.1648101211,-0.3402887881,0.0772090629,-0.0520791188,-0.0520761311,0.1725098938,-0.1219627634,-0.2785654664,-0.3615244627,-0.0546621159,0.0916810483,0.2551635206,-0.2535848916,0.0210730825,-0.1381287128,-0.134653762,0.140510276,0.1434345543,-0.23038131,0.0182863772,0.0926127955,0.0201504417,0.0239399355,0.0874806121,0.1017661765,0.0954296589,0.3023668826,0.0443399139,0.1146038547,0.0687186792,-0.2319999635,-0.3179428875,0.0059678582,0.0862289369,0.479133755,0.1176342964,-0.2625258863,0.2080416381,0.0123037184,0.1937862486,0.2482506633,-0.4686806202,0.2854803503,0.1489261389,0.3802307248,-0.2535903454,-0.1025018618,-0.0712967217,-0.0021172403,-0.0405235961,-0.1618707925,0.2719226182,0.0380862653,0.1705889553,0.1537543088,0.0261664949,-0.0461504012,0.2621161342,0.3201759756,-0.379897058,0.1222417057,-0.1708783507,-0.0001218276,0.0392642207,0.2162192762,0.0725905895,0.4372932911,0.3349696994,0.205059737,0.3065778613,-0.1398878545,0.1267788261,0.2541255653,0.1811512411,0.2138209492,-0.0361263417,0.0263570752,-0.2686807811,-0.1644441634,-0.2801771462,0.1193192005,0.0402753688,0.0170687288,-0.2899395525,-0.29945755,-0.2605128586,0.0391094349,-0.1277347207,-0.1199796647,0.0078242812,-0.0688198432,-0.1869893223,-0.3241331577,0.3364488482,0.1578710377,-0.032998275,-0.1121716276,0.3772613108,0.043862313,-0.1366412491,0.2384351343,0.5030642152,0.3806135952,0.3672260344,-0.2398211807,-0.0130413556,0.0730472729,-0.0788550004,0.0051911213,0.0679344907,0.2401355654,-0.0756513253,0.4011612833,0.2593036294,-0.1929460019,0.1742027253,0.2846862972,0.0487600565,-0.0909425989,-0.0725556314,-0.1908063293,0.1058025956,-0.0252529867,-0.3271608949,-0.280761987,0.0386263728,0.1342496127,-0.0300225522,0.1884898692,-0.2734946311,0.1709487587,-0.1840057224,0.5951343775,0.087907508,0.2517850697,-0.0925460011,-0.5342869759,-0.5470656157,0.2486828566,-0.0967421308,-0.0746892244,0.1531457901,-0.0305123311,0.2696278989,0.2860680223,0.0512862131,0.0799893141,-0.1488135308,0.2649143636,-0.5765098929,-0.0189915597,0.0992281139,0.2678034306,0.0091333091,-0.2319775522,0.1317761689,-0.1062836945,0.0971587971,0.3270369768,-0.0200540461,0.068199493,-0.0729768649,0.4364084899,0.0643488467,0.3692887425,-0.2329207212,-0.2746248245,-0.2033756673,-0.1324523687,-0.1983074695,0.104489781,0.1142907813,0.1928530186,0.0795367509,0.0659575015,0.0516958423,-0.0922978818,0.0949281007,0.0406889804,-0.3376064599,-0.0771070272,0.0768262371,0.0217755083,0.1538318694,-0.0804214254,-0.1722629219,-0.1288499683,-0.2211908698,-0.1158022061,0.5106816292,-0.0435347259,-0.2114216834,-0.0710551366,0.0075596706,0.0878862515,-0.0936626345,-0.3844065964,0.0481376126,0.1446213871,-0.0520434044,0.0914302543,0.1257634163,0.1857575923,0.0164519567,-0.0249970127,0.5657560825,0.1693138182,-0.2248267829,0.0039623305,-0.1232460439]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/296","title":"snli -1 labels","comments":"@jxmorris12 , we use `-1` to label examples for which `gold label` is missing (`gold label = -` in the original dataset). ","body":"I'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels?\r\n```\r\nimport nlp\r\nfrom collections import Counter\r\ndata = nlp.load_dataset('snli')['train']\r\nprint(Counter(data['label']))\r\nCounter({0: 183416, 2: 183187, 1: 182764, -1: 785})\r\n```\r\n","comment_length":22,"text":"snli -1 labels\nI'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels?\r\n```\r\nimport nlp\r\nfrom collections import Counter\r\ndata = nlp.load_dataset('snli')['train']\r\nprint(Counter(data['label']))\r\nCounter({0: 183416, 2: 183187, 1: 182764, -1: 785})\r\n```\r\n\n@jxmorris12 , we use `-1` to label examples for which `gold label` is missing (`gold label = -` in the original dataset). ","embeddings":[0.4284365177,-0.3220365644,-0.072350055,0.2394536138,0.2163777351,0.1463885605,0.5796833038,0.165535748,0.170363158,0.1855384856,-0.1904290468,0.3528851271,-0.1502354145,0.0844011679,0.2737314105,0.007783385,0.420153439,0.2741086185,0.3514617383,-0.4586472213,-0.1709999442,0.1382910013,-0.1522455215,0.3128228784,-0.4414059818,0.3461144269,0.0009128025,-0.3870435059,-0.0180653129,-0.7808856368,0.170510307,-0.0521242283,0.2494774014,-0.2046043128,-0.0001184899,-0.2829562724,0.111255303,0.1001419276,-0.6039978862,-0.1291385889,-0.1478041112,-0.2091726214,0.2700715959,-0.0226308107,-0.0561759919,-0.2235001922,0.2570438385,0.1213434637,-0.0377139598,0.2486657947,0.1201034412,0.0074542216,-0.3962410092,0.2241046727,0.2631272674,-0.2490135729,0.0693386346,0.0233187918,0.0102578299,-0.1695575118,0.4104804993,0.480410248,-0.1212427467,-0.1633181721,0.1586790979,0.0758184195,0.2860766947,-0.3415988684,-0.0064109308,0.2240253091,0.091682069,-0.2839529514,-0.2545427978,-0.3441044986,0.2020335495,-0.2572908998,-0.2316446304,0.1768646836,0.3629240394,0.0125367874,-0.1831649989,0.0134287514,-0.1046056002,0.1106877998,0.2206532806,0.6899819374,0.1222342774,0.2554949224,0.3940927088,0.2337832749,-0.2395154387,-0.0751245618,0.03525896,0.4097285867,-0.5185235739,-0.2892338932,0.0263588745,0.0373357572,-0.0361391269,0.1992178112,-0.0142051755,-0.1323988587,-0.1135496721,0.0498196892,0.0985607058,0.034002848,0.3828135133,0.2338069528,0.0019421491,-0.2110381871,0.2714414299,0.037879698,-0.2870954573,-0.0507952832,0.4360054433,-0.0839964375,-0.2353134751,-0.1152166128,-0.1464784592,-0.0289740562,-0.4838451147,-0.0168020632,0.1664530337,-0.2647247016,0.1997204274,0.2434001863,-0.2316465974,0.0599176213,-0.1101179048,-0.4850924015,-0.1473666131,0.0291762762,-0.1275295466,-0.0946524814,0.0863943622,-0.1976353973,0.1781769097,0.0272483211,-0.3649908602,-0.0937846676,0.3724904656,0.1010765806,-0.0501782596,0.3409380019,-0.306878835,0.6312388182,-0.0613477938,-0.142670095,-0.0895807445,-0.1582084745,-0.2209960967,-0.3180475831,0.0200446639,0.1640287489,0.0441541784,-0.2355019301,0.0694767833,0.1258792728,-0.0116781835,-0.1452896446,0.1268828958,-0.260232091,-0.0830596462,-0.025969509,0.0630622059,-0.0034207613,-0.1305021197,-0.1461331248,-0.0012306088,-0.28161937,0.3323378265,0.4054003358,-0.0714260936,-0.1050572172,0.0750620514,0.0698342621,0.4805379212,-0.3734812438,-0.3728365302,0.1543425918,-0.2553902268,-0.481998831,-0.030120898,0.6453999877,-0.1843071431,-0.0288854409,0.0026406678,-0.1771391928,0.0531971604,-0.1375496984,-0.3707210422,0.0868156329,0.5049743056,0.2264974117,-0.0512811989,-0.0258728005,-0.3803990185,0.1161504462,0.3642605543,-0.0719467402,0.1627792567,0.0299467184,0.2692738473,0.1729638129,0.1708913893,0.0442218706,-0.1966613084,0.1785045415,-0.0578384325,0.4421191514,0.4727391303,0.0412294604,-0.0882817879,-0.101877287,-0.1216624677,-0.1106762439,0.0536741465,0.1870279014,0.2011844963,-0.063810125,-0.1342466325,0.0688203573,-0.1588870585,0.1239439622,-0.3422494531,-0.0206103064,0.0580910258,0.0811992362,-0.1915920824,0.242266342,0.3230985105,0.1615242213,0.1224323511,0.0802264884,-0.1514077038,-0.2654776573,0.2320171148,0.3996815383,0.1741261631,-0.2977236211,0.0530991293,0.4022228718,0.0653087124,-0.2728527188,-0.071168907,0.1225158274,-0.012123418,0.2192324251,-0.2900401056,0.1457639039,0.1062537506,0.2003894746,-0.3209221661,-0.244148016,0.1283989102,-0.3559906185,0.0873631909,0.3305926621,-0.3314161301,0.0643265173,0.0423999131,-0.1731135249,-0.045882225,0.0996168181,-0.1229182556,-0.1310167611,0.340793997,0.0693279281,0.1192153841,0.215751633,-0.0695673004,-0.0096834265,-0.5985792875,-0.2397589386,0.1104309857,0.1005831435,-0.2163856775,-0.2415601313,0.3826270401,0.0610465221,-0.2787593603,-0.543517828,-0.2947511673,0.217355445,-0.2428180128,0.2016547918,-0.1393660456,-0.4047529101,-0.4046653211,-0.2777750194,-0.1012286469,-0.1519326121,0.2774133384,-0.2894717455,0.1986037493,0.0449058004,0.0757452101,0.2232674062,-0.1936234534,0.4343918562,-0.2699826062,-0.2043341994,-0.3394888639,-0.056992393,-0.3640294075,0.0277949907,0.2382814735,-0.0908759981,-0.3378667235,-0.009052095,-0.2354523242,-0.0937164277,-0.1507401168,-0.0995300412,0.1783875823,-0.1449024081,-0.2105629444,0.3857544363,0.0033595806,0.1552433968,0.0179583654,0.1333090365,-0.2987652123,0.1111499891,-0.389303118,-0.1397668421,-0.0535636097,0.2671801448,0.1522258371,0.0830078349,0.1453879178,0.4180040658,-0.1087657511,0.0565101579,-0.125598073,0.3179318607,-0.1365964115,-0.312974304,0.0649632439,0.2428857237,-0.0245366246,-0.0346248858,0.1262348145,-0.0309290327,-0.3222071528,-0.4038827419,0.0878098309,0.1511195451,-0.1705161482,-0.0218577497,0.040397469,-0.0944384187,0.2482842207,0.1087397709,-0.0642673969,0.1175234914,0.2205933332,-0.1593246013,0.4910188317,-0.3128922284,0.1855703741,-0.1263467669,-0.0193296131,0.5070952773,-0.0816486925,0.0644335076,0.0194319915,-0.1754759103,-0.0944150984,0.0038916778,0.3216068447,0.2923545539,-0.0655866861,0.3540462554,-0.0613373891,-0.0122186849,0.0131008821,0.2314555943,0.308468312,0.135984078,-0.0961407274,-0.0906891301,-0.0644003749,-0.0772187039,-0.2330174893,-0.1829549223,0.0067274445,-0.1595621258,-0.1179517508,-0.0307790767,0.2115188688,-0.3535703123,-0.2933385074,-0.1993021965,-0.1975818127,0.0419871956,0.2009849995,0.1904274374,0.2532204688,0.3469063044,0.4513167739,0.5280640721,-0.4117277861,0.1360704005,0.1708026677,0.4598353207,-0.3179312944,-0.1705050468,-0.3953025639,-0.2422589362,0.5076133609,0.3429151475,0.0351547785,-0.2855625153,0.4906057119,0.1188024208,-0.1419799477,-0.1873476952,-0.3700501323,-0.1589188576,0.1474743634,-0.1222152784,-0.060292881,-0.0016447213,-0.0889269337,0.0327814631,0.0389969908,-0.0667444468,0.179100275,0.1249870956,-0.0924705416,0.3167528808,-0.0236193016,0.2637498975,0.1502431929,-0.2214195281,0.3227817118,-0.0809332356,-0.5003842115,-0.1104294583,-0.0361135788,0.5688650608,-0.0528076775,0.1631466001,-0.1516097337,-0.1663205326,0.1289234161,-0.1146494597,0.1187446937,0.2864684761,0.1708457172,-0.3608555794,-0.5239185691,0.7567542195,0.1284826696,-0.236388579,0.0332072265,-0.2266273499,-0.2464483231,0.2846199274,0.3165576458,0.827855587,-0.0867288634,0.1937095374,0.1711327881,-0.5426453352,0.7377099395,-0.145880118,-0.0943196788,-0.2976086736,0.1872493327,0.0683358088,0.1426670551,-0.3443259299,0.6471413374,0.2203018516,0.3624387383,-0.0533214286,-0.157554552,0.2381697595,0.0301150512,0.5198007822,-0.1437744945,-0.2588644624,0.0772798881,-0.0972329974,0.0621345863,-0.2950023115,0.0429999046,0.160084933,-0.0188936833,-0.0083389422,-0.3128891289,0.3336398005,0.0106456075,0.2420302778,0.0306867044,-0.2397782952,0.3733062446,0.6426190138,0.1260436624,0.0152517604,-0.0982021317,0.1515772343,0.2396794409,-0.0440464765,0.1718333215,0.4269835055,-0.0039459374,-0.2366537005,-0.0985567942,0.0861790329,0.3245222569,-0.4428792894,-0.0249277093,-0.2609945536,0.2397548407,0.1858021617,0.088026382,0.1297471821,-0.2200129777,0.1239404678,0.2014139742,-0.2534951568,-0.1374191046,-0.062463399,-0.3123234212,-0.1925273538,0.2476595193,0.1644963622,0.1697164476,0.1050714329,-0.0963925496,-0.0613990054,-0.107551299,-0.0449472368,0.1630550176,-0.4371390045,0.1433756948,0.1855713725,-0.0526938327,-0.1781574637,0.297272265,0.0493203215,0.0285776109,0.1651865095,-0.0441845991,-0.4084078968,-0.2584608495,-0.2334872782,0.2323039323,0.0240254067,0.1225391328,0.0778636336,-0.1157705113,-0.2326527834,-0.3047423661,-0.328630358,0.403794229,-0.0644716024,0.1157140285,-0.3726020455,0.2319677174,0.1643986255,-0.1496162415,0.0667824894,-0.10291484,-0.1187080666,0.1555552185,0.130763337,-0.0619291253,-0.0685671642,0.2700779438,-0.1134969518,0.0676686168,-0.0447699577,0.1581243277,0.2739683986,0.8259962201,-0.1983125061,-0.0651813671,0.0930642188,-0.1075300351,0.0344628543,-0.0490144193,0.1830633581,-0.0047629252,0.2388305813,-0.0363173932,-0.3358845413,0.0341251679,-0.366240859,-0.0998134539,-0.0278984997,-0.354850322,0.5007499456,0.3900423646,0.177192986,0.1306580603,-0.1592106819,-0.1007733718,0.3226483464,0.1610521227,-0.0154710002,-0.144355908,0.3236383498,0.2663088441,-0.1328849494,0.1708846837,-0.0090133138,-0.2163248807,-0.0779211149,-0.1670884192,0.0538468622,-0.4135082364,-0.0067833406,0.3144771755,-0.0523298196,-0.072926335,0.2095649987,-0.114401944,-0.0529608056,0.4292707741,-0.0870924741,0.359690398,0.149671793,0.0044433693,-0.2610094547,0.3264990747,-0.2433244735,0.3039928079,-0.5721048117,0.1433635652,-0.0432602577,0.1583163589,-0.1123581678,-0.0117625352,0.2700254023,0.1538223028,-0.1934385747,-0.2539489269,-0.6602365375,0.3764891624,-0.1580034047,0.3186723888,-0.1989757121,0.0066090687,0.1375488788,0.0733638331,0.0162582584,-0.4448853135,-0.344065398,0.3040152788,0.4286315739,-0.1847045273,-0.0650628656,-0.0834733024,-0.0622958578,0.0360379964,-0.2250067294,0.1916491538,0.069441475,0.2670170665,-0.0361203402,-0.3936543763,-0.0636710152,-0.1485050321,0.0541361459,0.2923434973,0.0850882903,0.0918397754,-0.0210300796,-0.0421336442,0.5177637339,-0.1064379364,0.0594720691,-0.4241501689,0.3440137804,-0.1012062654,-0.4397434294,-0.3080053031,0.0708568171,-0.195152998,0.161849007,0.0613503829,0.219890371,0.17382586,-0.3444452882,0.0481266826,-0.1016844288,0.3673909605,0.0180784017,-0.0119424378,-0.0636007264,-0.0380199328,-0.3053799868,-0.0980094448,-0.1097090542,0.0181857124,0.1245696247,-0.0273207668,-0.413241148,-0.0354944132,0.0295253024,0.3789497614,-0.012074504,0.0577982739,0.0264078975,0.05632535,-0.2031363696,0.138843596,0.0511715412,0.1974755079,0.0491330773,-0.2189598382,0.0126609923,-0.028321255,0.1865625381,-0.1376838535,0.3892340362,0.1794745177,0.0042666392,0.3903366029,0.2284800112,0.189984858,-0.2603660524,0.0886811689,-0.2987479568,0.0385253169,-0.1958940327,0.2179410309,-0.1111422777,0.2162517458,0.039323356,-0.0394804664,0.2872755527,-0.0464996584,-0.419990927,0.2190901339,-0.0755965337,0.1324478239,-0.007037316,0.0159025248,0.4599244297,-0.1328625977,-0.1711117327,-0.4088754952,0.2925471961,-0.225391373,-0.0967535675,0.0106842173,0.2912589908,0.276157409,-0.0652413219,-0.7549912333,-0.1509407163,0.4303013384,0.0497179702,-0.5622780919,0.1153017357,0.2581783533,0.10885901,0.0058466238,-0.0553258471,0.2274071127,0.3328619897,0.4013539255,-0.4327887893]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/296","title":"snli -1 labels","comments":"Thanks @mariamabarham! so the original dataset is missing some labels? That is weird. Is standard practice just to discard those examples training\/eval?","body":"I'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels?\r\n```\r\nimport nlp\r\nfrom collections import Counter\r\ndata = nlp.load_dataset('snli')['train']\r\nprint(Counter(data['label']))\r\nCounter({0: 183416, 2: 183187, 1: 182764, -1: 785})\r\n```\r\n","comment_length":22,"text":"snli -1 labels\nI'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels?\r\n```\r\nimport nlp\r\nfrom collections import Counter\r\ndata = nlp.load_dataset('snli')['train']\r\nprint(Counter(data['label']))\r\nCounter({0: 183416, 2: 183187, 1: 182764, -1: 785})\r\n```\r\n\nThanks @mariamabarham! so the original dataset is missing some labels? That is weird. Is standard practice just to discard those examples training\/eval?","embeddings":[0.3789802194,-0.2363891453,0.0022033609,0.1978055984,0.1676168442,0.2278822511,0.6327139139,0.1720219404,0.048420202,0.1103978008,-0.1079240814,0.3746734858,-0.2219463885,0.1496949047,0.1942365319,0.1201866493,0.3345723152,0.3273703456,0.3262062371,-0.4975576699,-0.164640218,0.1366201341,-0.1091011316,0.2648751736,-0.4185903668,0.2373347431,-0.1538481414,-0.3429230154,0.104502432,-0.6462287903,0.2199270278,-0.1023126468,0.4217903912,-0.2809385657,-0.0001216671,-0.2675632536,0.0543847121,-0.0170518328,-0.5345324874,0.0139202308,-0.0729482174,-0.1938710213,0.290718168,-0.1057281047,-0.0797977522,-0.1985406131,0.2433385104,0.0082859797,0.074079141,0.2916976213,0.1208982617,-0.00731995,-0.4337947667,0.3430729508,0.3437729478,-0.2276860923,0.0630370229,0.0863161534,0.0201416314,-0.0772902891,0.4642212093,0.438005656,-0.1158606634,-0.2857222557,0.1743275672,0.1308352202,0.4274489284,-0.387041539,0.0074046087,0.2546112537,0.1439848244,-0.1291528344,-0.3260562122,-0.4347472489,0.2493036985,-0.311229974,-0.2360923588,0.2196952701,0.3017023504,0.1511388123,-0.2829215527,-0.0516654812,-0.1889192164,0.1414294839,0.1295459419,0.6374177337,0.1101341918,0.1782872379,0.1365814656,0.2269991934,-0.1342483014,-0.110567905,-0.0120628253,0.3769011497,-0.388661772,-0.2018814236,0.0016416555,0.0251683593,-0.0831509829,0.2363758236,0.0536901578,-0.0864013061,0.0708268061,0.0604363792,0.1855139136,0.1493222266,0.4604480565,0.3609256446,-0.0067780134,-0.2038100362,0.2374268025,-0.0178337768,-0.4318780899,0.0751332417,0.2864333391,-0.1190738007,-0.1524971426,-0.2075908333,-0.1685065478,-0.0080020847,-0.5624997616,0.0549807958,0.0852830261,-0.1243278682,0.0734069422,0.2966636419,-0.230012387,0.1054317728,-0.1578510106,-0.6040849686,-0.1277037412,0.1315526217,-0.1773129404,-0.2071239501,0.1995102465,-0.2129683048,0.2379116118,-0.0117013454,-0.3459094167,-0.0630118847,0.4673307836,0.0577621646,0.1729013324,0.3453713953,-0.1311792433,0.5312720537,-0.0620600693,-0.1381403953,-0.0748060942,0.0297695827,-0.2154980749,-0.3328003883,0.1816056669,0.1727150381,-0.1253916025,-0.153014496,0.0454456285,0.1467695981,0.0849798396,-0.2330768108,0.0582960956,-0.3490657806,0.042917598,-0.0943082795,0.1568502784,0.0098793786,-0.1807923168,-0.1234242022,0.0030136926,-0.1748470664,0.3881748617,0.4391433299,-0.0888142362,0.0585124977,-0.0513508096,0.0846712962,0.5011434555,-0.2736699283,-0.2654494941,0.1133242175,-0.1075648591,-0.2903260887,-0.1185356006,0.513369143,0.0103101078,-0.0220843237,-0.0262202695,-0.0576498248,-0.0469392613,-0.1908290982,-0.4078170657,0.0439910553,0.5144603252,0.1584966481,-0.103487052,0.1325870156,-0.46264413,0.1176945195,0.383008033,-0.1768920124,0.18461667,0.1184726954,0.2052664906,0.1556284726,0.0737579241,0.0711098909,-0.2450219393,0.2019115984,-0.0962226018,0.2569848299,0.5588251352,-0.1285279691,-0.0360272974,-0.18266581,-0.1530183256,-0.2860913575,0.0586607195,0.15145953,0.1641748101,-0.0385966934,-0.1827423871,0.1158315614,-0.2991262674,0.0803115442,-0.468260169,-0.0258936062,0.1107965484,0.0876264349,-0.1554864794,0.2683875859,0.216019839,0.0232611056,0.1758345813,0.097792007,-0.0976430029,-0.2183565795,0.1102543473,0.2468774468,0.0685358867,-0.2757854462,0.0964060649,0.4207382798,0.0108609125,-0.2066677362,-0.2059896439,0.2684469521,0.0079928413,0.2270089835,-0.1075849608,0.0882544816,0.0733909681,0.0765407085,-0.3425277174,-0.1568909883,0.1437413543,-0.3022581637,0.1347488016,0.2816454172,-0.3767746687,0.0181833934,0.0338232629,-0.1499401182,-0.0465881191,0.1338830441,-0.1493709832,-0.1522812396,0.3809719086,0.0114321765,0.163294971,0.2967981398,0.1102058738,0.0073573957,-0.4906772971,-0.1788840443,0.2380926907,0.1433742046,-0.2497667372,-0.1835832447,0.3590720594,0.1449327767,-0.3770843446,-0.4780540168,-0.2569273412,0.212987572,-0.2778189778,0.2787250578,-0.1515138298,-0.4487258792,-0.2847844958,-0.1748629361,-0.1372944117,-0.2256820053,0.2676307857,-0.2786779106,0.0791948736,0.0603089221,0.1153282523,0.2810527682,-0.3082813323,0.4000158012,-0.1465576887,-0.2746135592,-0.5887207389,-0.0012312586,-0.4473215044,0.0778465047,0.2033810318,-0.1895958632,-0.4071713984,0.0100831548,-0.2798158526,-0.0498667918,-0.0165852681,-0.0519570969,0.1574081331,-0.0276500024,-0.2775493264,0.3261582553,-0.0807925463,0.1119641885,-0.0124512371,0.043721687,-0.1787853986,0.1360493004,-0.4149359465,-0.2756350935,-0.0465781055,0.1876485646,0.2455367595,0.1591165066,0.1108182818,0.4958164692,-0.0989606306,0.0194263142,-0.1318602264,0.2323003262,-0.1400011927,-0.2537880838,0.1410228014,0.1477815956,-0.1040073186,-0.0014571632,0.0027141643,-0.0066846069,-0.1398134679,-0.3399149179,0.0674358383,0.2494532168,-0.1703867614,-0.0908595175,0.0514993705,0.0219412409,0.1329831183,0.2189226747,-0.0586002208,0.0718773082,0.3512069881,-0.1890821755,0.3821302056,-0.2982612848,0.1768516749,-0.1004799157,0.0370757096,0.4577076137,-0.0518539809,-0.0058301282,-0.047437299,-0.0291824881,0.0231919773,0.1906560063,0.3272813857,0.1197554916,-0.2978253365,0.3343204558,0.0545965806,0.040574301,-0.0613857098,0.2818070054,0.1798592359,-0.0218219385,-0.0816583708,-0.1425870657,0.0308791399,0.1000034958,-0.1886535734,-0.208334282,0.037570104,-0.1164194793,-0.0889541656,0.0256187189,0.1456847489,-0.5623360872,-0.264572233,-0.2910386026,-0.0964044333,0.0268385671,0.2349053174,0.2702010572,0.0919688568,0.2787650228,0.307972312,0.5773761272,-0.454192102,0.1010529101,0.2169912159,0.437235415,-0.3506586552,-0.2670773864,-0.43105492,-0.2427933812,0.5924478173,0.183787927,0.0631688237,-0.178306371,0.4404780567,0.211356774,-0.1671170592,-0.0091177234,-0.3676311672,-0.2801332176,0.1514922529,-0.225518465,0.0998723134,-0.2316171676,-0.1847835779,0.0377952717,-0.0299500562,-0.0386960097,0.2709361613,0.1719335914,0.0662582889,0.2084994763,0.0524799712,0.1953190118,0.2376717031,-0.1929769367,0.4753508866,-0.1004653275,-0.585526228,-0.0582429282,-0.191585958,0.632401228,0.0503111742,0.1660875678,-0.2600183785,-0.2560628057,0.0257666856,-0.0416339114,0.1995993406,0.3217670918,0.059987314,-0.2406296879,-0.4686696231,0.6235411167,0.1689916402,-0.2375259995,0.0783445239,-0.4431008697,-0.3350730836,0.2247922868,0.2869834602,0.8856855631,0.1262027025,0.3661869168,0.3354715705,-0.3373045921,0.8705614209,-0.1879982799,0.0127250003,-0.3058094382,0.118599765,0.0700375438,0.0823480189,-0.227914691,0.656393528,0.2238608897,0.3990519345,-0.0981678367,-0.1551802158,0.2914149165,0.1050477475,0.5124004483,-0.066476047,-0.4091573656,0.0350258909,0.0152163124,0.0287250429,-0.3384090066,0.0999068618,0.0955638215,-0.0726382285,0.1350623071,-0.0836107731,0.3146389127,0.085230343,0.2030789256,0.0409643017,-0.2102829665,0.3841661215,0.644598186,0.1013747901,0.0158997867,-0.0758418441,0.1029417366,0.1986329854,-0.0099775242,0.156191349,0.3324311078,-0.037679486,-0.123859562,-0.0617924444,-0.0871214494,0.3406481445,-0.4597951472,-0.0132420864,-0.2539182901,0.0986298174,0.1966833919,0.1201200187,0.1116019636,-0.206735462,0.0915943906,0.2102094293,-0.2533970177,-0.1335427463,-0.0884481221,-0.2359195352,-0.2024580538,0.2868642509,0.2457191795,0.123208262,0.2502731681,-0.1167795062,0.0769653097,-0.0672692508,0.0656147972,0.1356365979,-0.4576823413,0.1038213,0.0951975584,-0.0379903093,-0.0826910362,0.3962375224,0.1526851058,0.0466478206,0.1268286556,-0.0097043784,-0.3443364501,-0.1337976605,-0.1481460184,0.2108782679,0.0822313875,0.2009558827,0.0558289587,-0.2324184477,-0.2361392379,-0.2393503189,-0.3015692532,0.4574971795,0.0366838686,0.0851949751,-0.2945260406,0.1102593914,0.0999091044,-0.1245957389,0.1262038946,-0.1113535687,-0.0942032039,0.1619904488,0.1065685749,-0.0592296906,0.00959587,0.30823946,-0.2213831842,0.1064027026,-0.1534722,0.081743069,0.2111967951,0.7996432781,-0.3170280159,-0.0095319012,0.0008499373,-0.0802497864,-0.0886885077,-0.0254132599,0.1548059434,0.0184291601,0.2977449298,-0.0485095084,-0.3181404471,-0.0409381427,-0.4460301697,-0.091598019,-0.0327100195,-0.3658599555,0.5111541748,0.2165148258,0.0875851512,0.3626528978,-0.2400742173,-0.129108116,0.3561379015,0.1767558157,-0.1472703964,-0.1037261859,0.2747886479,0.1189866811,-0.2216212451,0.2710149884,-0.0114516821,-0.114124313,-0.0148248738,-0.1753042042,0.0686898455,-0.3968212605,0.0040427181,0.3071716428,-0.1414183676,0.0307053607,0.0707298294,-0.0678538904,-0.0127555225,0.3585958481,-0.0802129284,0.3504011929,0.10297627,0.0636831447,-0.2606744766,0.0631324798,-0.2640179992,0.34533301,-0.519277513,0.289784044,-0.0208266862,0.1285162717,-0.1185199693,-0.1662263721,0.2212132215,0.1108655259,-0.2267230749,-0.2458800972,-0.6276063919,0.2741487026,-0.1437020749,0.2110807747,-0.1500760168,0.0283959396,0.15617612,0.1212303713,0.0291352086,-0.4544233382,-0.3073241413,0.1885603815,0.4231506288,-0.1460372061,-0.0479931645,-0.0789800137,-0.0787819847,0.0149004105,-0.1647477746,0.2416423708,0.0391856469,0.3086338937,-0.0157489404,-0.3157174587,-0.0413154587,-0.1969047785,0.000433933,0.3784509003,0.0629483908,0.1716902554,-0.0153633691,-0.0419245325,0.6420857906,-0.1040763855,-0.0137656797,-0.5475209355,0.4322214425,-0.0443516634,-0.4740867317,-0.262745738,0.0081886966,-0.2035595328,0.166473195,0.1110773757,0.2830700278,0.122277528,-0.181273967,0.0362667069,-0.209881708,0.379041791,0.1606827229,0.0271956194,-0.1502490044,-0.0526475124,-0.376784265,-0.0111139389,-0.2767833769,-0.0702105314,0.2621850371,-0.0410405211,-0.3554293215,-0.0356917344,-0.1404940337,0.2957130373,-0.0004415192,-0.0161137022,-0.1004497781,-0.0170594845,-0.2887943685,0.274664104,0.0790032297,0.0116287116,0.0686541274,-0.2450374961,-0.0474188738,0.0484855287,0.202288717,-0.0413750708,0.3697508276,0.1213603243,0.0822205693,0.3480968177,0.2035218179,0.0409392454,-0.1323343515,0.0810141861,-0.2230849564,0.1789840311,-0.1970832795,0.2060554773,-0.208886072,0.2160958797,0.009711381,-0.1077621207,0.460260421,0.1112254709,-0.5706871152,0.3810386062,-0.1481636763,0.2599347532,0.1089124978,0.1522687376,0.2566180229,-0.32181108,-0.3548780382,-0.2834203839,0.3448897898,-0.4080424905,-0.1270120293,-0.1688765883,0.3636052012,0.1903061718,-0.070861429,-0.7023304701,-0.1694470197,0.5747876167,-0.0646351799,-0.5643618703,0.0875896588,0.2418245375,-0.0498726219,0.0293722637,-0.0877610296,0.1584820747,0.1231741607,0.3682658076,-0.5045647621]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/296","title":"snli -1 labels","comments":"Yes the original dataset is missing some labels maybe @sleepinyourhat , @gangeli can correct me if I'm wrong \r\nFor my personal opinion at least if you want your model to learn to predict no answer (-1) you can leave it their but otherwise you can discard them. ","body":"I'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels?\r\n```\r\nimport nlp\r\nfrom collections import Counter\r\ndata = nlp.load_dataset('snli')['train']\r\nprint(Counter(data['label']))\r\nCounter({0: 183416, 2: 183187, 1: 182764, -1: 785})\r\n```\r\n","comment_length":47,"text":"snli -1 labels\nI'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels?\r\n```\r\nimport nlp\r\nfrom collections import Counter\r\ndata = nlp.load_dataset('snli')['train']\r\nprint(Counter(data['label']))\r\nCounter({0: 183416, 2: 183187, 1: 182764, -1: 785})\r\n```\r\n\nYes the original dataset is missing some labels maybe @sleepinyourhat , @gangeli can correct me if I'm wrong \r\nFor my personal opinion at least if you want your model to learn to predict no answer (-1) you can leave it their but otherwise you can discard them. ","embeddings":[0.4120676816,-0.2979552448,-0.0225758217,0.2181213647,0.1907149702,0.2088953108,0.5165299773,0.1553622186,0.1881432235,0.1818924844,-0.1560331881,0.3227591813,-0.2433932722,0.1784412861,0.2389490157,0.0606847554,0.3916952014,0.3287827373,0.3345844448,-0.502890408,-0.2930355966,0.0634000525,-0.173450619,0.355927974,-0.4537765384,0.2747114301,-0.0779209659,-0.3143007457,-0.0579689853,-0.6953605413,0.2357222289,-0.0917173848,0.3343616426,-0.1888742298,-0.0001194382,-0.2986067832,0.0320900045,-0.0003681006,-0.4734799862,-0.0767838284,-0.1925250441,-0.2501674294,0.3062699437,-0.1513938159,-0.1258653998,-0.2164394259,0.1940398067,0.0685293749,0.0365864411,0.313518852,0.1114537045,-0.0359704792,-0.4204098284,0.319221288,0.352468431,-0.2056467533,0.0593444929,0.0503150187,-0.0299880952,-0.0958659798,0.4762229919,0.4895938337,-0.1497641504,-0.2577138841,0.1788147986,0.0883318856,0.3778859377,-0.470713079,0.0388765372,0.1814833581,0.2218284756,-0.2397468835,-0.2664619088,-0.4241093099,0.293340981,-0.3391400874,-0.1773945391,0.1882055998,0.3220196962,0.0577190779,-0.182820335,-0.1295999587,-0.0896184146,0.1390957981,0.1569350362,0.7602541447,0.1406560838,0.2380843312,0.2896889448,0.249429673,-0.126504302,-0.132482931,0.0078537939,0.4254180193,-0.4600966573,-0.2374998331,-0.0690263286,0.0206214897,-0.0608253293,0.1892069131,-0.0739322305,-0.2107034475,-0.062701948,-0.0006546288,0.1791669577,0.0348445661,0.409309268,0.3669911623,-0.0277556386,-0.2307464331,0.2878878117,-0.0333755054,-0.3174919188,0.0533652566,0.3976795971,-0.0820273533,-0.180334568,-0.1435499489,-0.0571226403,0.011618251,-0.4993458092,0.0274719447,0.1146375984,-0.1753023267,0.1298204809,0.3081150651,-0.1446696669,0.0642628372,-0.1655151248,-0.5688715577,-0.141046375,0.0696520954,-0.1543442011,-0.0887165517,0.1715065837,-0.1474168301,0.225234583,0.0379579179,-0.4195157588,-0.1014051661,0.3598073125,-0.0088054687,0.0723962262,0.3655067682,-0.2805555165,0.5722590089,-0.0470122583,-0.1811730266,-0.0567510463,-0.0131022595,-0.2438603789,-0.2372651845,0.0339744575,0.1756392419,-0.07299418,-0.1955127567,0.0229026526,0.083822526,0.001399285,-0.2158325911,0.1908331811,-0.3064798713,-0.0159682594,-0.051163163,0.1275995523,0.0489357933,-0.1882529259,-0.1532481611,-0.0311978944,-0.2510912716,0.3765103519,0.3837752044,-0.0085928272,-0.0096009709,0.0293977465,0.0564486571,0.4859068096,-0.2925832868,-0.3611377478,0.1270952076,-0.1321854442,-0.3597744703,-0.0624113791,0.6516435742,-0.1079123467,0.0389409512,-0.011343657,-0.23585926,0.0670659989,-0.1416018009,-0.3873313069,0.0147746848,0.4654177129,0.2721132636,0.026565386,0.031527333,-0.515111804,0.0976451784,0.3338267505,-0.0722983554,0.2375006229,0.1364919096,0.2828272879,0.0465550199,0.0920957848,0.0238438006,-0.3373574018,0.1459613889,-0.1382603049,0.3534272313,0.464759022,-0.0086071314,-0.0576968268,-0.1500570625,-0.146410346,-0.1785556227,0.0540154874,0.1326824427,0.0954848677,-0.0072237425,-0.1951130629,0.0866376609,-0.2494881898,0.0913205445,-0.3926183283,0.015047201,0.1165630147,0.1016098484,-0.0797537193,0.2640586197,0.2795211375,0.1297652125,0.1358781904,0.0003686027,-0.1316739917,-0.2979814112,0.1232521161,0.3700306118,0.1434331089,-0.3118866086,0.0499982797,0.3983569145,0.0920298174,-0.2388046384,-0.04840957,0.1889577806,-0.1347482055,0.2445468903,-0.3007158041,0.0992037877,0.0971002951,0.0403351821,-0.3237749338,-0.1682404876,0.1457582861,-0.332526654,0.1793392897,0.3762755096,-0.323881954,0.0190690998,0.1378279328,-0.1528866887,-0.046198789,0.1275491416,-0.177638635,-0.1051504314,0.2849455178,0.0955817774,0.2118729055,0.246341899,-0.0961582586,0.0231638048,-0.4363025725,-0.1902455389,0.1616933495,0.1675454825,-0.1787641048,-0.146193251,0.3087539971,0.1443947554,-0.2746993899,-0.4500887394,-0.3060742915,0.2510749996,-0.2683726847,0.1634267718,-0.1412920058,-0.5088816881,-0.2976730168,-0.2456631064,-0.1944870353,-0.2243620306,0.2605407238,-0.3658292294,0.1340445131,0.0514947884,0.0158295538,0.2225194573,-0.2364580631,0.3548371494,-0.3228273988,-0.2669400275,-0.3862211406,-0.0414766558,-0.3748985529,0.0453385301,0.2227772325,-0.1961697042,-0.3911016583,0.030558927,-0.1571665406,-0.0725814551,-0.1245048866,-0.1254419982,0.1159212962,-0.0132749574,-0.2571898401,0.3905107081,0.0143683106,0.131772548,0.0094248699,0.0172660965,-0.2333174199,0.1317771971,-0.3792231381,-0.2530622184,-0.1266162694,0.24661614,0.1063783616,0.0922698975,0.134298563,0.4017403722,-0.1403211355,0.0462529063,-0.1153584942,0.2229656875,-0.1725610048,-0.3250908554,0.1119552478,0.3089745045,-0.1082744002,0.0492471755,0.0864282623,0.0389346145,-0.3106299341,-0.412858367,0.1307025254,0.1852618158,-0.1554034948,-0.0966402516,-0.021869285,-0.0284482464,0.2581855953,0.1251472682,-0.0609103926,0.0154040353,0.2919617891,-0.1403698921,0.504345417,-0.318820864,0.186768949,-0.2205400169,-0.0358247533,0.5008656383,-0.0741827637,0.05216204,-0.0028762829,-0.0713223442,-0.0926257223,-0.0057193367,0.3170007467,0.2775494158,-0.1601706743,0.3516402841,-0.0462263711,0.0496385619,-0.0731873214,0.1602074653,0.1834719032,0.1161020175,-0.0383036211,-0.1100278944,-0.0571384169,-0.0247605387,-0.2262963653,-0.1800100356,-0.0809920058,-0.1655825526,-0.0547471903,0.0727091953,0.1859612614,-0.4748190045,-0.3318730295,-0.2425298244,-0.1980276406,0.0416520685,0.1889446825,0.1611314267,0.2330145389,0.3268891275,0.445779264,0.4638791382,-0.3988007903,0.0417249277,0.1874551773,0.4624137282,-0.3448842168,-0.1333738267,-0.3770235777,-0.2464068979,0.533461988,0.2832220197,0.008938414,-0.3169467449,0.4711284339,0.2187291235,-0.218674019,-0.1684420705,-0.4044206738,-0.1621641219,0.1867662966,-0.154246524,-0.0060061566,-0.1126523539,-0.1067885533,0.026667716,0.0724604726,-0.0175954551,0.1845531613,0.1021759212,-0.02632525,0.2213954479,-0.0649815053,0.2034005076,0.2449342906,-0.1937403083,0.4518113434,-0.0410845503,-0.5115538239,-0.0335898846,-0.0907112062,0.589289546,-0.0325844064,0.1608941257,-0.155228883,-0.1838810891,0.059899129,-0.1085603833,0.050505247,0.3003541529,0.0890518054,-0.3902947307,-0.436368674,0.7228566408,0.0815630555,-0.1719101071,0.0059370426,-0.2486792654,-0.2880950272,0.2744540572,0.3458246291,0.9324253798,-0.0712630674,0.2072614282,0.2928717136,-0.3646984696,0.8622067571,-0.0975194648,-0.0897947848,-0.226690799,0.113545239,0.0219359752,0.1276339442,-0.3245335221,0.6223574281,0.2274999917,0.4053783715,-0.0418421216,-0.204715997,0.2640422285,0.0660800636,0.6422035694,-0.0533036776,-0.4470774531,0.0390949771,-0.1335031837,0.128235653,-0.2708443403,0.0703797042,0.1040463373,0.0090866331,0.0478217229,-0.1962493807,0.3035810292,-0.0276349895,0.2179737836,0.0770641118,-0.173365131,0.4282211065,0.6543002129,0.1119264886,-0.0282819066,-0.1202464923,0.0882358104,0.1786637306,0.033064343,0.2490722388,0.2887359262,-0.0233737454,-0.2041571885,-0.0659400299,-0.0300763175,0.310383141,-0.4120935798,-0.0461028516,-0.1786436141,0.164609313,0.1362209767,0.0737087205,0.1547456086,-0.2045824975,0.1059292108,0.1657689959,-0.1347656846,-0.122678861,0.0002052126,-0.3513312042,-0.2387323976,0.2964492142,0.2478313893,0.1747175008,0.3137504458,-0.1312606484,-0.0554102249,-0.1325775534,0.0515651479,0.1521888226,-0.4057530165,0.033632338,0.2777855396,0.0419270694,-0.113506034,0.3011670411,0.1136480868,-0.0191672575,0.1740302444,-0.0611903146,-0.3874638677,-0.1178925782,-0.2103268504,0.1481331587,0.0996732488,0.1258939505,0.1087310538,-0.1069824621,-0.2287208587,-0.2903219759,-0.3800808787,0.3530940115,0.0464963242,0.1237651259,-0.3487107158,0.1256732941,0.1340165287,-0.2506502867,0.1234280318,-0.0979514271,-0.0834476873,0.1621094793,0.1471530944,-0.0378139131,-0.097318992,0.2508469522,-0.1084837541,0.0788034126,-0.1278063804,0.1197321117,0.2841045856,0.9013696909,-0.2116038352,-0.129084155,-0.0734342933,0.002056841,-0.0550003722,-0.1138261855,0.1674351692,0.0308064725,0.3321894705,-0.0369180664,-0.4046168029,-0.0674338043,-0.4131424725,-0.0754169896,0.0291554593,-0.3671227694,0.5702582598,0.372582972,0.0709696263,0.2155797631,-0.2156759799,-0.1370783299,0.3358867764,0.1587321311,-0.1546134353,-0.0737885311,0.3120911419,0.2377635688,-0.2022255659,0.1658437401,-0.1147421524,-0.2452222705,0.0237955954,-0.1320172846,0.098356612,-0.4627384543,-0.0020070225,0.2875806093,0.0211351365,-0.0386360735,0.1688721776,-0.0603207238,-0.089213945,0.3625588715,-0.0979538411,0.2860642672,0.1389857382,0.0911918283,-0.2694695294,0.2116659582,-0.2799436152,0.3901804984,-0.6274667382,0.0955670774,-0.0342851877,0.1606092155,-0.0645186827,-0.0760620087,0.3054051697,0.1958965212,-0.209768787,-0.191042006,-0.7020381093,0.3145955503,-0.1027280167,0.1935329288,-0.2056529373,-0.0455451347,0.1961165816,0.1518577188,0.0099518001,-0.4093883634,-0.3378508985,0.3046694398,0.4196786582,-0.1412096918,-0.035309732,-0.0490983129,-0.0423775166,0.0393350273,-0.128957361,0.1613066792,0.0981490016,0.2882367074,-0.0141032897,-0.2982613444,-0.0761911124,-0.1293286234,0.0636347979,0.3996810019,0.1183513477,0.1223066673,-0.0125679774,-0.0191139095,0.5593019128,-0.1612658203,0.0404150411,-0.4187419116,0.3836141527,-0.0556760058,-0.4298972189,-0.3160558045,-0.0121609904,-0.1664955169,0.1803803593,0.0493980944,0.2256157547,0.196999684,-0.2624002695,0.0364301652,-0.2123407423,0.3262475431,0.1140016913,-0.0221190322,-0.1866138875,0.0389246233,-0.2679362595,-0.1143456995,-0.1252754778,-0.0990084186,0.2258652002,-0.0921942964,-0.3923729956,-0.0536669716,-0.061584197,0.3732468784,-0.0354195647,0.0634638146,-0.0916888192,0.0390190147,-0.2757048011,0.2917639315,0.0259665251,0.1461022347,0.0774336681,-0.1551792771,-0.0456009395,0.0881757513,0.248096481,-0.0357053317,0.4333093166,0.0886676908,0.0669216514,0.426424861,0.1809559166,0.1041544527,-0.1060265228,0.0758252069,-0.2426319718,0.0849782452,-0.0950708538,0.2359129488,-0.1151680127,0.2538585365,-0.0044187712,-0.1171191484,0.2143302709,0.0435275212,-0.5104151368,0.2887328267,0.0654882118,0.167499885,0.0151929138,-0.0162008181,0.3670670986,-0.1828430146,-0.2723827064,-0.2747081518,0.308242321,-0.2324135005,0.0141178332,0.0368095227,0.3590264916,0.2286821008,-0.0832487717,-0.7226815224,-0.1563959271,0.5330834389,0.0519067161,-0.5233260989,0.073201552,0.2085354924,0.1135643497,0.0224633627,-0.0436311737,0.1975239515,0.2256666422,0.3895891607,-0.4311517179]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/296","title":"snli -1 labels","comments":"thanks @mariamabarham :)","body":"I'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels?\r\n```\r\nimport nlp\r\nfrom collections import Counter\r\ndata = nlp.load_dataset('snli')['train']\r\nprint(Counter(data['label']))\r\nCounter({0: 183416, 2: 183187, 1: 182764, -1: 785})\r\n```\r\n","comment_length":3,"text":"snli -1 labels\nI'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels?\r\n```\r\nimport nlp\r\nfrom collections import Counter\r\ndata = nlp.load_dataset('snli')['train']\r\nprint(Counter(data['label']))\r\nCounter({0: 183416, 2: 183187, 1: 182764, -1: 785})\r\n```\r\n\nthanks @mariamabarham :)","embeddings":[0.3987343311,-0.4084069133,-0.0900833234,0.2612741888,0.275809139,0.1109563783,0.6256619096,0.1450338662,0.1475811154,0.1977067292,-0.1651434451,0.294881314,-0.1345397085,0.1564467102,0.232513994,-0.0078926003,0.4151615798,0.2522506118,0.3198880553,-0.4785743654,-0.1987515241,0.1606215388,-0.1503230035,0.2942630947,-0.4702290595,0.3279315829,-0.030081328,-0.3784004748,-0.0688133538,-0.7466760874,0.1631505638,-0.0540600233,0.3223309517,-0.0843609869,-0.0001192235,-0.2504051328,0.0853463039,0.0806941465,-0.517065227,-0.0799218416,-0.1311120987,-0.2710482478,0.2943291962,-0.0917066336,-0.0488167964,-0.2053107619,0.24899064,0.0526542924,-0.0359104127,0.2849620581,0.1341539025,-0.0238160267,-0.461992383,0.2522328496,0.3160700798,-0.2337110788,0.0591383353,0.0338711105,0.0021752978,-0.2291318327,0.4288457334,0.4324343801,-0.1107585654,-0.1963166147,0.1667013168,0.1325818151,0.3821352422,-0.4211915135,-0.0158296674,0.213446185,0.1480621696,-0.2670800686,-0.2047597021,-0.3256912529,0.1893228143,-0.2206353396,-0.2233840674,0.1592287421,0.3690971732,-0.0451078638,-0.1786305755,0.0110537959,-0.167760253,0.1558137536,0.1760710627,0.7095534801,0.1223420799,0.2733263969,0.3797571659,0.2261905968,-0.1539127231,-0.0752856508,0.0476237237,0.412199527,-0.5126996636,-0.2778637409,-0.0342820622,-0.0023436169,-0.0652115941,0.2106858492,-0.0773180723,-0.184254691,-0.0757396147,0.0189383142,0.1597939283,0.055862017,0.391593188,0.2523008883,-0.014880891,-0.1784004718,0.2020897567,-0.0042516682,-0.3058371842,-0.0689531416,0.3880011141,-0.104612194,-0.1994222105,-0.0924650505,-0.0873844847,-0.0756511018,-0.4989043474,-0.0181070678,0.1963099241,-0.2116565108,0.1760822386,0.3059530854,-0.1789648682,0.079594247,-0.1298741847,-0.4986465275,-0.1302460879,0.0456669219,-0.1546678543,-0.1399698555,0.1102641225,-0.2272677571,0.1511800885,0.0285894349,-0.2997407913,-0.1429174244,0.3865942359,0.076977931,-0.0372783951,0.3069441617,-0.2824966311,0.6067794561,-0.0130851557,-0.1807052046,-0.076055035,-0.1414889097,-0.2130059004,-0.267372787,0.0118019227,0.1651759148,0.1175245866,-0.2009219527,0.0822198465,0.1663183868,0.0889379233,-0.1846438348,0.1355158538,-0.2221369743,-0.0863017738,-0.0530552603,0.0605667531,-0.0509615205,-0.1147635728,-0.0789151639,-0.046673663,-0.2590727508,0.3599107265,0.4097955525,-0.0492697544,-0.0667038634,0.0573865175,0.1105437875,0.4777913392,-0.3274585009,-0.3396186829,0.1782696396,-0.2292699218,-0.5447964668,-0.0298417751,0.6681706309,-0.1352801323,0.0035758419,0.0284783449,-0.1840630025,0.0669306815,-0.1063869521,-0.3477205336,0.0998641923,0.4748425186,0.216326952,-0.0765627995,0.0254456252,-0.3653620183,0.1196247637,0.3829890192,-0.0683675334,0.1414222866,0.0211291574,0.251851052,0.1717103422,0.1643458903,0.0135429585,-0.2385726869,0.1409929693,-0.1278825402,0.4191736579,0.5016694069,0.0232640784,-0.0869285762,-0.0610104874,-0.0809347481,-0.0815209001,0.0747126788,0.1616963297,0.1181602776,-0.0841823071,-0.190299511,0.0409037396,-0.1646622717,0.0779731497,-0.3627045751,0.0166864879,0.0645643622,0.0794695988,-0.1878774613,0.2097447962,0.3173929751,0.1602073461,0.142130509,0.0913412496,-0.0495763235,-0.2585815787,0.1787399501,0.3717321455,0.1238322183,-0.2509697676,0.0813107565,0.4089393318,0.0978341848,-0.2808463275,-0.0843668357,0.1484227926,-0.0538442805,0.2225773484,-0.2784974873,0.1305761635,0.0888644978,0.2094953209,-0.359069407,-0.1931104958,0.1779943854,-0.2938699722,0.0606439374,0.3463154435,-0.3052708805,0.0297923107,0.0213646609,-0.1342235804,-0.0788112879,0.0506730452,-0.0978986248,-0.1150251403,0.274023205,0.0157313012,0.1210815012,0.1923609078,-0.0856459364,-0.0141434772,-0.5560248494,-0.2502607405,0.1212233603,0.0504451841,-0.2257017046,-0.221174866,0.4221006036,0.050660897,-0.3182944655,-0.5752282143,-0.3277903795,0.2732390761,-0.2825601697,0.2688110173,-0.0612338707,-0.5022619963,-0.368255049,-0.2860470116,-0.1424925178,-0.1440639496,0.2393574268,-0.2579886317,0.2351369262,0.0100434637,0.062140502,0.2715739608,-0.1865747273,0.4824827611,-0.2564046681,-0.1823573261,-0.3763328791,-0.0515378267,-0.344930619,0.037900079,0.2488895804,-0.0926909521,-0.3658179641,0.0044285618,-0.2446585894,-0.1468926072,-0.1330826432,-0.0755954683,0.2088576853,-0.0371194184,-0.2422650456,0.3791138232,0.0233050752,0.2084567845,0.0079700574,0.1034533754,-0.2512502074,0.1058707088,-0.3321133256,-0.1565256268,0.01568643,0.2573852837,0.1842478663,0.1211880147,0.1785460562,0.4095543921,-0.1433460861,0.0808044896,-0.1058648676,0.248714,-0.1355242729,-0.2810254693,0.0694207251,0.3107701242,-0.0695885047,-0.014880307,0.1368800253,-0.0254076924,-0.3475432098,-0.3588249087,0.0685172528,0.1010663733,-0.1529740691,-0.0687912628,0.0627229735,-0.048512008,0.2182376683,0.1140146405,-0.0614577048,0.0822617859,0.2247988582,-0.0744252354,0.4702658951,-0.2975918651,0.160977751,-0.1489894092,-0.0262471661,0.5425739288,-0.0795618892,0.1187845692,-0.0614113733,-0.217340827,-0.0573926009,-0.0592207909,0.3427761197,0.2520189881,-0.1670706868,0.3735430539,-0.0628584325,0.0206350666,-0.0332832448,0.2293984592,0.32515046,0.1826937646,-0.0883889198,-0.1601848751,-0.0812882259,-0.10038618,-0.2393651754,-0.1825386137,-0.0013210137,-0.1783058494,-0.1034359634,-0.1043486893,0.1928486079,-0.4112859368,-0.3529286683,-0.2379438281,-0.1748145372,0.0771940649,0.1884561181,0.1573356092,0.2622513771,0.3932051957,0.4015141129,0.6309692264,-0.3769318759,0.0712738484,0.1455408037,0.4204767644,-0.3498201668,-0.1471611857,-0.3831292093,-0.218957305,0.4833292067,0.3602923453,0.0396392718,-0.2319043875,0.4273100495,0.0968189165,-0.1621137708,-0.2164921463,-0.3960735202,-0.1606957912,0.176470086,-0.1259320229,-0.0267554596,-0.0382425599,-0.0456048734,-0.0003733425,0.0226760171,-0.101280652,0.175225094,0.1553212553,-0.0810705125,0.3016564846,-0.0711258426,0.2457137555,0.1832224727,-0.2675262392,0.3570610285,-0.12131311,-0.5296924114,-0.1089468375,-0.0081399772,0.5931622982,-0.0229157191,0.1707222164,-0.120859921,-0.113607578,0.1498273462,-0.1015961394,0.1249929443,0.3587319851,0.1539218575,-0.3779093325,-0.5426743627,0.8230270147,0.1075943485,-0.2050906569,0.0423561558,-0.2815334797,-0.2521683574,0.2689492404,0.2687506378,0.9241307378,-0.0964315906,0.2383432388,0.2189648002,-0.4042148292,0.7685850263,-0.2023452669,-0.0420876443,-0.362216115,0.1734749973,0.0180191565,0.1429168135,-0.3856374323,0.5889232755,0.175512135,0.32432881,-0.1104557216,-0.2174020708,0.1989078522,0.08066912,0.5287909508,-0.1138789728,-0.379406631,0.0776260719,-0.0850876942,0.0489892289,-0.3110431135,0.083852388,0.141527459,-0.0513939448,0.0012405546,-0.3118574917,0.3193747103,-0.0064877779,0.2586152554,0.0622075461,-0.2187107801,0.4063127041,0.6349969506,0.1174195483,-0.0363764279,-0.1223883405,0.1841396987,0.1883280277,-0.057308007,0.1606715322,0.3482441306,-0.0448835716,-0.2401543558,-0.1121036336,0.1162502989,0.3234719038,-0.4272460043,0.0414191,-0.2674103379,0.1323269606,0.1775025278,0.0574747361,0.0988177061,-0.1712965965,0.1190593317,0.1958473623,-0.2576484084,-0.126029551,0.0601523779,-0.334966749,-0.1584065259,0.2911749482,0.1824248135,0.1345459074,0.1491423547,0.0220174138,-0.0771296099,-0.12822707,0.003854295,0.1037132964,-0.3919375837,0.1338397413,0.203226164,0.0135820759,-0.1560951918,0.409904182,0.0417747349,0.030453762,0.1526101977,-0.0525368862,-0.3999591172,-0.2487495691,-0.2570881546,0.259172529,0.0862174928,0.0685965195,0.1256168336,-0.0879581347,-0.2240544707,-0.3036743701,-0.3200731575,0.3996984661,-0.0002789477,0.0646984428,-0.3695366383,0.2584276795,0.186556235,-0.162207976,0.0612782836,-0.1254805773,-0.0692482069,0.1581467092,0.1465917826,-0.0765823573,-0.1029291525,0.2035356611,-0.179144606,0.0699855089,-0.1134569645,0.1777899265,0.2410763204,0.7865943909,-0.1661373675,-0.1198010445,0.0866685063,-0.0662573203,-0.0065286951,-0.1041296571,0.2007730752,-0.001229464,0.2673496306,-0.0635587424,-0.3520231545,0.0714267716,-0.3061510324,-0.1040539816,-0.0725136921,-0.3303344846,0.5096989274,0.3600275517,0.1648560762,0.2420961261,-0.1873249561,-0.0886109322,0.3326611221,0.1591019183,-0.0295576658,-0.1310192049,0.2920713425,0.2780195475,-0.1549559981,0.2133645713,-0.0677046925,-0.1991872191,-0.0910229385,-0.1694119573,0.0351271816,-0.4364935458,0.0013390903,0.2698520124,-0.0441436395,-0.0914554074,0.1500889361,-0.0710248128,-0.0413864031,0.4403309524,-0.1253806353,0.2588404119,0.1305698007,-0.0180710386,-0.2866958082,0.3295022547,-0.2448032796,0.3250710666,-0.6085312963,0.0866659507,-0.0104327463,0.2135892808,-0.0869555995,0.0046762135,0.2024992406,0.1303240657,-0.1703120321,-0.2713127434,-0.656375885,0.3279542327,-0.1520035118,0.2764295936,-0.1880948544,0.053242892,0.2002066821,0.1201247945,0.0446417816,-0.5113631487,-0.3235946894,0.3121114075,0.4268139899,-0.1948346645,-0.0145882042,-0.1162049398,-0.0110258935,0.0654759929,-0.2298766077,0.2273275107,0.1456444561,0.2437868863,-0.0310394373,-0.4153619409,-0.0812190175,-0.1423537433,-0.0015552434,0.3088946342,0.1628900021,0.1164509878,-0.0101484377,-0.028676644,0.5247452855,-0.1285450608,0.061597839,-0.4535705447,0.3755066693,-0.120842047,-0.3805136383,-0.330206275,0.0197297987,-0.234413296,0.1581130624,0.0375814848,0.160307616,0.2151566446,-0.2985954285,0.0452095941,-0.1409590244,0.4087360501,0.0233332869,0.0470010638,-0.0738471523,-0.0355514698,-0.2601795793,-0.0750003606,-0.1792423129,-0.0283230934,0.0793777257,-0.0519932285,-0.3998752534,-0.0316269808,0.0566334501,0.3448258936,-0.0480732396,0.1143400595,-0.0010922406,0.088761434,-0.244311586,0.2094343454,0.01895223,0.1898720115,0.0912581384,-0.2311501503,0.0058532786,-0.0177665967,0.2022081465,-0.1223527342,0.3983100355,0.1166011542,0.0411041193,0.4090609252,0.1833141893,0.2579068244,-0.2777833045,0.0851021335,-0.3178979754,0.0657710209,-0.090366587,0.2273607999,-0.1708717048,0.2454805821,0.0489836782,-0.0336684883,0.3235521317,0.0527044833,-0.3495848179,0.1940379888,-0.0234043542,0.1165349334,-0.0466226563,-0.0332418606,0.4848484695,-0.1705670208,-0.166915074,-0.4205538034,0.3065153062,-0.2826459408,-0.1666634828,0.0043030423,0.2944333255,0.1917605549,-0.0244050398,-0.7229029536,-0.177951321,0.4721991718,0.0581004098,-0.5893354416,0.1077289358,0.2178434134,0.0444203652,-0.0123551451,-0.0764181092,0.2389422655,0.3563856781,0.3796823919,-0.4596875608]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/294","title":"Cannot load arxiv dataset on MacOS?","comments":"I couldn't replicate this issue on my macbook :\/\r\nCould you try to play with different encodings in `with open(path, encoding=...) as f` in scientific_papers.py:L108 ?","body":"I am having trouble loading the `\"arxiv\"` config from the `\"scientific_papers\"` dataset on MacOS. When I try loading the dataset with:\r\n\r\n```python\r\narxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n```\r\n\r\nI get the following stack trace:\r\n\r\n```bash\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-2-8e00c55d5a59> in <module>\r\n----> 1 arxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 662 \r\n 663 generator = self._generate_examples(**split_generator.gen_kwargs)\r\n--> 664 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 665 example = self.info.features.encode_example(record)\r\n 666 writer.write(example)\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1106 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1107 \r\n-> 1108 for obj in iterable:\r\n 1109 yield obj\r\n 1110 # Update and possibly print the progressbar.\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/datasets\/scientific_papers\/107a416c0e1958cb846f5934b5aae292f7884a5b27e86af3f3ef1a093e058bbc\/scientific_papers.py in _generate_examples(self, path)\r\n 114 # \"section_names\": list[str], list of section names.\r\n 115 # \"sections\": list[list[str]], list of sections (list of paragraphs)\r\n--> 116 d = json.loads(line)\r\n 117 summary = \"\\n\".join(d[\"abstract_text\"])\r\n 118 # In original paper, <S> and <\/S> are not used in vocab during training\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)\r\n 346 parse_int is None and parse_float is None and\r\n 347 parse_constant is None and object_pairs_hook is None and not kw):\r\n--> 348 return _default_decoder.decode(s)\r\n 349 if cls is None:\r\n 350 cls = JSONDecoder\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in decode(self, s, _w)\r\n 335 \r\n 336 \"\"\"\r\n--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())\r\n 338 end = _w(s, end).end()\r\n 339 if end != len(s):\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in raw_decode(self, s, idx)\r\n 351 \"\"\"\r\n 352 try:\r\n--> 353 obj, end = self.scan_once(s, idx)\r\n 354 except StopIteration as err:\r\n 355 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 46983 (char 46982)\r\n\r\n163502 examples [02:10, 2710.68 examples\/s] \r\n```\r\n\r\nI am not sure how to trace back to the specific JSON file that has the \"Unterminated string\". Also, I do not get this error on colab so I suspect it may be MacOS specific. Copy pasting the relevant lines from `transformers-cli env` below:\r\n\r\n- Platform: Darwin-19.5.0-x86_64-i386-64bit\r\n- Python version: 3.7.5\r\n- PyTorch version (GPU?): 1.5.0 (False)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n\r\nAny ideas?","comment_length":26,"text":"Cannot load arxiv dataset on MacOS?\nI am having trouble loading the `\"arxiv\"` config from the `\"scientific_papers\"` dataset on MacOS. When I try loading the dataset with:\r\n\r\n```python\r\narxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n```\r\n\r\nI get the following stack trace:\r\n\r\n```bash\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-2-8e00c55d5a59> in <module>\r\n----> 1 arxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 662 \r\n 663 generator = self._generate_examples(**split_generator.gen_kwargs)\r\n--> 664 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 665 example = self.info.features.encode_example(record)\r\n 666 writer.write(example)\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1106 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1107 \r\n-> 1108 for obj in iterable:\r\n 1109 yield obj\r\n 1110 # Update and possibly print the progressbar.\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/datasets\/scientific_papers\/107a416c0e1958cb846f5934b5aae292f7884a5b27e86af3f3ef1a093e058bbc\/scientific_papers.py in _generate_examples(self, path)\r\n 114 # \"section_names\": list[str], list of section names.\r\n 115 # \"sections\": list[list[str]], list of sections (list of paragraphs)\r\n--> 116 d = json.loads(line)\r\n 117 summary = \"\\n\".join(d[\"abstract_text\"])\r\n 118 # In original paper, <S> and <\/S> are not used in vocab during training\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)\r\n 346 parse_int is None and parse_float is None and\r\n 347 parse_constant is None and object_pairs_hook is None and not kw):\r\n--> 348 return _default_decoder.decode(s)\r\n 349 if cls is None:\r\n 350 cls = JSONDecoder\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in decode(self, s, _w)\r\n 335 \r\n 336 \"\"\"\r\n--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())\r\n 338 end = _w(s, end).end()\r\n 339 if end != len(s):\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in raw_decode(self, s, idx)\r\n 351 \"\"\"\r\n 352 try:\r\n--> 353 obj, end = self.scan_once(s, idx)\r\n 354 except StopIteration as err:\r\n 355 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 46983 (char 46982)\r\n\r\n163502 examples [02:10, 2710.68 examples\/s] \r\n```\r\n\r\nI am not sure how to trace back to the specific JSON file that has the \"Unterminated string\". Also, I do not get this error on colab so I suspect it may be MacOS specific. Copy pasting the relevant lines from `transformers-cli env` below:\r\n\r\n- Platform: Darwin-19.5.0-x86_64-i386-64bit\r\n- Python version: 3.7.5\r\n- PyTorch version (GPU?): 1.5.0 (False)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n\r\nAny ideas?\nI couldn't replicate this issue on my macbook :\/\r\nCould you try to play with different encodings in `with open(path, encoding=...) as f` in scientific_papers.py:L108 ?","embeddings":[-0.0581401661,0.0877689496,-0.0419098586,0.1886685789,0.2178975642,0.0454704054,-0.0241877399,0.3616167009,0.0815683678,-0.1818822473,-0.0505227782,0.691057086,-0.1919905692,-0.2139135003,-0.1082298383,0.0094932541,-0.1276049614,0.0114534106,0.1146815494,-0.1110842973,-0.1230693087,0.1993306875,-0.1573242992,-0.0993064716,0.4090956151,0.0958301648,0.1364857703,0.3957486153,-0.0027542866,-0.3556491733,0.1349909604,0.0240957309,0.2171578258,0.5375875831,-0.0001107034,0.2718124986,0.3134252429,-0.0939889029,-0.2043009251,-0.5840528011,-0.4115153253,-0.306637615,0.1335057318,-0.2136874497,-0.0360715277,-0.2607661784,0.1242143437,-0.1539173871,0.3288322389,0.3756518662,0.207921654,0.0798494294,0.3830094635,0.1164422184,-0.0670725107,-0.02313504,0.0021167931,0.4649830163,0.0422845446,0.1263236105,0.1612235159,0.1448102146,-0.1489691883,0.1127980724,0.2092457116,-0.0296654217,-0.1075037271,-0.1436872929,0.0375226587,0.1304574609,0.2721568644,-0.0974486545,-0.2403058112,0.0564082675,0.2218078375,-0.1019054577,0.295181632,0.430082649,-0.1124939024,0.0761579052,-0.3062997162,-0.3497585952,-0.4034608901,0.4736581147,-0.0218985379,-0.1115582362,-0.200718835,0.2033873051,0.1595942974,-0.2415931225,0.0368958823,0.0939481407,0.1707556397,0.11694552,-0.4864809811,0.2469767183,0.0796729624,-0.3753705025,0.0753835738,-0.150146991,-0.1037795395,-0.0815444365,-0.2829775214,0.2728938162,0.1120163128,0.018602727,-0.1558181047,-0.0018167398,0.0582292303,-0.0125903161,-0.1686109453,0.0329580121,-0.3595866859,-0.1540408581,-0.3627384901,-0.3801209033,0.2619623542,-0.3207472861,0.0621837713,-0.0250361618,-0.0250417013,0.0168774351,0.0238443669,0.228412956,-0.0542670265,0.1421990842,0.1716236323,0.4296216667,-0.3466945887,0.1758017391,-0.0509758741,0.0754750222,-0.1097994223,-0.2968219519,0.2044046372,0.0846939757,0.1635968983,-0.0225838032,0.0854115933,-0.201628834,0.1252433211,0.1339762807,-0.1017934233,0.4808894694,0.3299891949,0.0564984679,0.2672579885,0.0386227481,-0.1250129193,0.2215334028,-0.4679540396,-0.1227610484,-0.2679323256,0.162654683,-0.1997473538,-0.1317337751,-0.1605494618,-0.1432931125,0.1379330009,0.2394692749,-0.3634818196,-0.2293667346,0.0001346738,-0.3333007991,0.2488360554,0.0392104574,-0.3133456707,0.1613029242,-0.1056164876,0.1475966722,0.2834636867,0.3652399778,-0.2056614608,0.1426015943,-0.0255504828,-0.1135711595,0.3589995801,-0.3276819885,-0.1195762455,0.0364844464,-0.1393856555,0.076270476,-0.0743433014,0.0231164042,-0.1838792413,0.2579210103,0.2040625364,0.2711192966,0.0826367363,0.0304179639,-0.1805005819,-0.2009442002,0.45898664,0.3523835838,-0.0857428014,0.0294668283,-0.0705721304,-0.072366327,0.1135509834,-0.099117592,-0.3065351248,0.2054903805,0.0127701126,0.0597909503,0.0696247518,0.1258382499,0.1664613187,0.0967324823,0.2396769971,0.2965689003,-0.2159079313,0.3827721477,-0.3872411549,0.0926724225,-0.0310393795,0.0711300671,0.065654926,0.0439680256,-0.0934773609,0.1950520128,-0.1820219457,0.2183669358,-0.3807091713,-0.0371009409,-0.8843935132,0.0560739338,-0.1170641184,0.1107353494,0.2100989223,0.1402559727,-0.1191787124,-0.2400333434,-0.1165203899,0.2139781266,0.2940783203,-0.0699314922,0.2279684991,0.042080678,0.0583746955,-0.3201346993,0.2476847917,0.0171775892,0.0454673953,0.0980466679,-0.2400624603,0.3954666555,0.0848648995,0.1853234768,0.1478035152,0.1319267154,0.4758432508,-0.0060153292,-0.0275846701,-0.0089764493,0.4516296685,0.1805775911,-0.340288192,-0.1420488507,-0.1855588555,-0.0007562374,0.0774995834,0.1592724323,-0.0840466544,0.3020502329,-0.1074541733,0.0607566088,-0.1180131286,0.0291220751,0.3401017785,0.1709996909,-0.3174942732,0.0375214443,-0.2074111104,-0.0640699863,0.1576645225,0.0042496887,0.0024907775,-0.1562239826,0.130223617,0.1190572008,-0.2286053598,-0.2915285528,-0.0149250245,0.4036516249,-0.3374558687,0.0624639392,-0.3534377515,-0.2601831853,-0.0542715788,-0.1631045043,-0.1600035727,-0.2046508193,-0.1103503406,0.1549523771,0.1971672177,0.1569382101,-0.3404233754,-0.0687447339,0.2280444354,-0.212758258,0.2178190351,-0.3201696575,-0.090573445,0.0915043354,0.2567333579,0.115356423,0.2168773413,-0.0711518824,-0.1991698444,0.1951156259,-0.3111149967,-0.0690635741,-0.2529280484,0.1622450054,-0.0127729326,0.2829394639,0.0663689077,-0.2645323873,0.2123935223,0.0352514498,-0.1587046087,0.302259326,0.0569703095,-0.4054757655,-0.1261679381,-0.6859310865,-0.2402991205,-0.2965233922,0.0207346324,0.2632628381,0.1707098782,0.4518840313,0.3053714931,0.021538198,0.0888195485,0.0629290193,-0.1581660807,-0.0866923556,0.2277038097,-0.1510714889,-0.4011754692,0.1772225201,0.2070526481,0.5639321804,0.1303535402,-0.4385875762,0.1577425152,-0.1061138362,-0.0755192265,-0.2182867229,0.1015386134,0.0663023368,-0.210977301,-0.0257533733,0.0079541886,-0.1914936155,-0.0301366076,0.1466688365,0.2630581856,0.2407557815,0.4278973341,-0.2681685686,0.5905352831,0.0142622264,-0.1609193683,0.069476828,-0.1589577198,0.4026772082,-0.0031177653,-0.5069510341,0.2158525139,0.284262687,0.0171935838,0.1744769961,0.0340316966,-0.2551556826,-0.1661703885,0.2379712909,-0.3800166845,-0.2908549011,0.196081847,-0.2611018419,0.07894703,0.0448552743,-0.0590058006,-0.0890889242,-0.1388099492,-0.1887180358,0.4358404875,0.064318262,0.0315002762,-0.4632994235,0.0505689085,-0.342972517,0.2074735016,-0.1587536782,0.6160392761,-0.1219579875,-0.0999062806,0.1105678827,-0.014026559,0.3819845617,0.0812925398,0.1661246568,0.3375768065,-0.1239420772,-0.4966859519,-0.0414071269,0.0024599754,0.4472510517,0.4753023982,0.1587123126,0.0748985484,-0.1119315773,-0.1855520904,0.1059326828,-0.0940444842,-0.2817312777,-0.3055519164,-0.3186794221,-0.2693054676,-0.4744183123,-0.1720836312,0.2304685116,-0.0999512672,0.1676111668,-0.1212368608,-0.216672942,0.149696812,0.3861213326,0.0398848876,-0.0878831297,0.1504201591,0.215640977,-0.1163051724,0.1365493387,0.3614488244,0.1097372323,-0.7085859179,-0.0461754575,-0.0668326467,0.1575988531,-0.2619423568,-0.1674872935,-0.1746036857,0.076938197,-0.1360736489,0.002871929,0.1660889685,0.2037215233,0.0454277992,-0.204467237,-0.2929310203,0.2632623315,-0.1085660085,0.0342324302,0.1920119822,-0.2135142386,-0.1601367593,0.3781900108,-0.021024093,0.5590004325,-0.3097172678,0.1412570328,0.2761947513,0.0139164403,0.6699483395,-0.1115628928,0.2104437798,-0.3174932301,-0.0141911358,-0.0565616973,-0.2007324994,0.026437737,0.1123935506,-0.1216480434,0.1802656204,-0.0242364183,-0.2814278305,0.1291989833,0.2311947495,0.5022940636,-0.244632259,-0.7192257047,0.2023690343,-0.217751801,0.0522673614,0.0517729037,-0.1352651715,-0.2248557061,0.0489743501,-0.3124033511,0.3894600272,0.0901355222,0.3495639861,0.3082270324,-0.0468276963,0.0555090979,0.2309151143,0.1750369817,0.227895841,-0.1704107672,0.2341102362,-0.0087976633,-0.2591122687,0.248261109,0.1001886129,0.1886753291,-0.1627509743,-0.2001486868,0.1331465244,0.0853323266,-0.5239751339,0.1386449933,0.1783330888,-0.0720260888,0.0992175639,-0.2829616666,0.1097199842,0.1071184203,0.2075476944,0.1173229814,0.3670862317,0.0736812353,0.3407841921,0.0079562264,-0.4690263867,0.1429866999,0.2529992759,0.1792363971,0.0090548825,0.3015254736,0.142531231,0.0471658967,-0.1045872048,-0.1746542007,-0.461915791,0.010912667,-0.1224688143,0.3127566576,0.2229425609,0.1521368474,-0.1818508506,0.0410407372,-0.2236406952,0.0997594744,-0.859870553,-0.1354100406,-0.0548281781,-0.4878775179,-0.0006446862,0.2166990638,-0.0469971113,-0.0847147852,-0.1913502067,-0.2497873902,0.0197665431,-0.2032182515,-0.1099875793,0.2819565833,0.0619983673,0.1283500195,-0.2295306325,0.1299589574,-0.4033862948,-0.1084756553,-0.198766306,-0.0028026998,0.1285460293,0.052987881,-0.2573785782,-0.2084369361,-0.2459956408,-0.3299534023,0.1025005355,-0.040987689,-0.0201559719,-0.154088214,0.0581756756,-0.5069238544,0.0165901165,0.1186364442,0.3586595356,-0.1164739281,0.2467471063,0.2674191296,0.2644314468,-0.1753063053,-0.0909070447,-0.5406355262,-0.066910632,-0.2757559121,0.1637987047,0.3702971339,-0.1089690551,0.1324842572,0.0816420168,0.0535019226,0.5262340903,-0.1782574058,-0.0461499244,0.1299265176,0.1488399208,-0.2839462459,-0.1335304677,0.3298020065,0.309981674,-0.2143112719,0.2018111199,-0.0222959127,0.0226881001,0.1296839714,0.0303221513,0.5294115543,-0.2493306249,0.0884126201,-0.151979059,-0.2476819903,-0.0916667953,-0.2119928598,0.2499311864,0.2411658913,0.372854948,-0.0014118456,0.1080931425,0.1908807456,-0.1238163337,-0.013418478,-0.3544937372,-0.464689672,0.1635971665,0.0233234875,0.3950412869,0.0079648141,0.0928862691,0.0519949831,-0.2455949187,-0.065728046,0.237026006,0.1193934754,-0.000631774,-0.2060878128,0.0416782387,-0.0317267925,-0.1816642731,0.037583489,-0.1187287867,0.0865239799,0.3490435183,-0.3035861552,-0.7911368608,0.2285946459,0.0662597418,-0.1032067463,0.0720451027,0.3918698132,0.023687683,0.0435211957,0.3245864809,0.1088312864,0.5102598071,0.4114189446,-0.0492335819,-0.0457732938,0.0242724456,0.1189988926,-0.0850605667,0.3223448098,0.033200141,0.3221098781,0.2746065557,0.1198492125,0.0574153103,0.2889924943,0.1625372469,0.0712472424,0.0226034895,0.0666723996,0.1023148671,0.2394506186,-0.2568710446,0.0724869147,-0.3361437619,-0.1103490889,0.2590260506,0.2074869275,0.1140956059,0.0336433798,0.0582498871,-0.1319148839,0.7058607936,0.510782063,0.371773243,-0.1432923228,-0.0604136288,-0.5617278218,0.1689248532,-0.2268105,-0.209326908,0.1078344584,0.1282417923,0.4357310832,0.1924453825,-0.027367983,-0.1725472957,-0.400547415,-0.0237439033,-0.342823118,-0.2178272456,0.0078974469,-0.1119740531,-0.1383003145,-0.0180382617,0.0899634212,-0.1823913902,-0.0019597169,0.0270195864,0.1578267366,-0.0332669131,0.0914950147,0.4042009115,0.015184517,0.4290499091,-0.0270545203,-0.2645837665,-0.4509984851,-0.1991005838,-0.3306635618,0.3517012894,0.2352200449,0.1762435734,-0.1694796979,0.1872226745,-0.168528229,0.4838164151,-0.0565907434,-0.5422750115,0.1255182773,0.073942557,0.0957553387,0.1444119662,-0.0106463991,0.053534545,0.0044792788,0.5422230959,-0.2458273768,-0.1378215253,0.2703121006,-0.2595653236,-0.3543881774,0.2645863295,-0.1342298537,-0.3655931354,0.3093405664,-0.5371618867,0.1654866189,0.3419696987,-0.1092510149,-0.1710449755,-0.0504334494,0.1009506658,-0.2060299069,-0.0858500078,0.5699597597,-0.119342804,-0.2334057987,-0.0469614193,0.0090484619]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/294","title":"Cannot load arxiv dataset on MacOS?","comments":"I was able to track down the file causing the problem by adding the following to `scientific_papers.py` (starting at line 116):\r\n\r\n```python\r\n from json import JSONDecodeError\r\n try:\r\n d = json.loads(line)\r\n summary = \"\\n\".join(d[\"abstract_text\"])\r\n except JSONDecodeError:\r\n print(path, line)\r\n```\r\n\r\n\r\n\r\nFor me it was at: `\/Users\/johngiorgi\/.cache\/huggingface\/datasets\/f87fd498c5003cbe253a2af422caa1e58f87a4fd74cb3e67350c635c8903b259\/arxiv-dataset\/train.txt` with `\"article_id\": \"1407.3051\"`.\r\n\r\nNot really 100% sure at the moment, but it looks like this specific substring from `\"article_text\"` may be causing the problem?\r\n\r\n```\r\n\"after the missing - mass scale adjustment , the validity of the corrections was tested in the @xmath85 productions at 1.69 gev\/@xmath1 . in fig . [\", \"fig : calibrations ] ( a ) , we show the missing - mass spectrum in the @xmath86 region in the @xmath87 reaction at 1.69 gev\/@xmath1 . a fitting result with a lorentzian function for the @xmath86 ( dashed line ) and the three - body phas\r\n```\r\n\r\nperhaps because it appears to be truncated. I (think) I can recreate the problem by doing the following:\r\n\r\n```python\r\nimport json\r\n\r\n# A minimal example of the json file that causes the error\r\ninvalid_json = '{\"article_id\": \"1407.3051\", \"article_text\": [\"the missing - mass resolution was obtained to be 2.8 @xmath3 0.1 mev\/@xmath4 ( fwhm ) , which corresponds to the missing - mass resolution of 3.2 @xmath3 0.2 mev\/@xmath4 ( fwhm ) at the @xmath6 cusp region in the @xmath0 reaction .\", \"this resolution is at least by a factor of 2 better than the previous measurement with the same reaction ( 3.2@xmath595.5 mev\/@xmath4 in @xmath84 ) @xcite .\", \"after the missing - mass scale adjustment , the validity of the corrections was tested in the @xmath85 productions at 1.69 gev\/@xmath1 . in fig . [\", \"fig : calibrations ] ( a ) , we show the missing - mass spectrum in the @xmath86 region in the @xmath87 reaction at 1.69 gev\/@xmath1 . a fitting result with a lorentzian function for the @xmath86 ( dashed line ) and the three - body phas' \r\n# The line of code from `scientific_papers.py` which appears to cause the error\r\njson.loads(invalid_json)\r\n```\r\n\r\nThis is as far as I get before I am stumped.","body":"I am having trouble loading the `\"arxiv\"` config from the `\"scientific_papers\"` dataset on MacOS. When I try loading the dataset with:\r\n\r\n```python\r\narxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n```\r\n\r\nI get the following stack trace:\r\n\r\n```bash\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-2-8e00c55d5a59> in <module>\r\n----> 1 arxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 662 \r\n 663 generator = self._generate_examples(**split_generator.gen_kwargs)\r\n--> 664 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 665 example = self.info.features.encode_example(record)\r\n 666 writer.write(example)\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1106 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1107 \r\n-> 1108 for obj in iterable:\r\n 1109 yield obj\r\n 1110 # Update and possibly print the progressbar.\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/datasets\/scientific_papers\/107a416c0e1958cb846f5934b5aae292f7884a5b27e86af3f3ef1a093e058bbc\/scientific_papers.py in _generate_examples(self, path)\r\n 114 # \"section_names\": list[str], list of section names.\r\n 115 # \"sections\": list[list[str]], list of sections (list of paragraphs)\r\n--> 116 d = json.loads(line)\r\n 117 summary = \"\\n\".join(d[\"abstract_text\"])\r\n 118 # In original paper, <S> and <\/S> are not used in vocab during training\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)\r\n 346 parse_int is None and parse_float is None and\r\n 347 parse_constant is None and object_pairs_hook is None and not kw):\r\n--> 348 return _default_decoder.decode(s)\r\n 349 if cls is None:\r\n 350 cls = JSONDecoder\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in decode(self, s, _w)\r\n 335 \r\n 336 \"\"\"\r\n--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())\r\n 338 end = _w(s, end).end()\r\n 339 if end != len(s):\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in raw_decode(self, s, idx)\r\n 351 \"\"\"\r\n 352 try:\r\n--> 353 obj, end = self.scan_once(s, idx)\r\n 354 except StopIteration as err:\r\n 355 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 46983 (char 46982)\r\n\r\n163502 examples [02:10, 2710.68 examples\/s] \r\n```\r\n\r\nI am not sure how to trace back to the specific JSON file that has the \"Unterminated string\". Also, I do not get this error on colab so I suspect it may be MacOS specific. Copy pasting the relevant lines from `transformers-cli env` below:\r\n\r\n- Platform: Darwin-19.5.0-x86_64-i386-64bit\r\n- Python version: 3.7.5\r\n- PyTorch version (GPU?): 1.5.0 (False)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n\r\nAny ideas?","comment_length":352,"text":"Cannot load arxiv dataset on MacOS?\nI am having trouble loading the `\"arxiv\"` config from the `\"scientific_papers\"` dataset on MacOS. When I try loading the dataset with:\r\n\r\n```python\r\narxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n```\r\n\r\nI get the following stack trace:\r\n\r\n```bash\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-2-8e00c55d5a59> in <module>\r\n----> 1 arxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 662 \r\n 663 generator = self._generate_examples(**split_generator.gen_kwargs)\r\n--> 664 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 665 example = self.info.features.encode_example(record)\r\n 666 writer.write(example)\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1106 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1107 \r\n-> 1108 for obj in iterable:\r\n 1109 yield obj\r\n 1110 # Update and possibly print the progressbar.\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/datasets\/scientific_papers\/107a416c0e1958cb846f5934b5aae292f7884a5b27e86af3f3ef1a093e058bbc\/scientific_papers.py in _generate_examples(self, path)\r\n 114 # \"section_names\": list[str], list of section names.\r\n 115 # \"sections\": list[list[str]], list of sections (list of paragraphs)\r\n--> 116 d = json.loads(line)\r\n 117 summary = \"\\n\".join(d[\"abstract_text\"])\r\n 118 # In original paper, <S> and <\/S> are not used in vocab during training\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)\r\n 346 parse_int is None and parse_float is None and\r\n 347 parse_constant is None and object_pairs_hook is None and not kw):\r\n--> 348 return _default_decoder.decode(s)\r\n 349 if cls is None:\r\n 350 cls = JSONDecoder\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in decode(self, s, _w)\r\n 335 \r\n 336 \"\"\"\r\n--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())\r\n 338 end = _w(s, end).end()\r\n 339 if end != len(s):\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in raw_decode(self, s, idx)\r\n 351 \"\"\"\r\n 352 try:\r\n--> 353 obj, end = self.scan_once(s, idx)\r\n 354 except StopIteration as err:\r\n 355 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 46983 (char 46982)\r\n\r\n163502 examples [02:10, 2710.68 examples\/s] \r\n```\r\n\r\nI am not sure how to trace back to the specific JSON file that has the \"Unterminated string\". Also, I do not get this error on colab so I suspect it may be MacOS specific. Copy pasting the relevant lines from `transformers-cli env` below:\r\n\r\n- Platform: Darwin-19.5.0-x86_64-i386-64bit\r\n- Python version: 3.7.5\r\n- PyTorch version (GPU?): 1.5.0 (False)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n\r\nAny ideas?\nI was able to track down the file causing the problem by adding the following to `scientific_papers.py` (starting at line 116):\r\n\r\n```python\r\n from json import JSONDecodeError\r\n try:\r\n d = json.loads(line)\r\n summary = \"\\n\".join(d[\"abstract_text\"])\r\n except JSONDecodeError:\r\n print(path, line)\r\n```\r\n\r\n\r\n\r\nFor me it was at: `\/Users\/johngiorgi\/.cache\/huggingface\/datasets\/f87fd498c5003cbe253a2af422caa1e58f87a4fd74cb3e67350c635c8903b259\/arxiv-dataset\/train.txt` with `\"article_id\": \"1407.3051\"`.\r\n\r\nNot really 100% sure at the moment, but it looks like this specific substring from `\"article_text\"` may be causing the problem?\r\n\r\n```\r\n\"after the missing - mass scale adjustment , the validity of the corrections was tested in the @xmath85 productions at 1.69 gev\/@xmath1 . in fig . [\", \"fig : calibrations ] ( a ) , we show the missing - mass spectrum in the @xmath86 region in the @xmath87 reaction at 1.69 gev\/@xmath1 . a fitting result with a lorentzian function for the @xmath86 ( dashed line ) and the three - body phas\r\n```\r\n\r\nperhaps because it appears to be truncated. I (think) I can recreate the problem by doing the following:\r\n\r\n```python\r\nimport json\r\n\r\n# A minimal example of the json file that causes the error\r\ninvalid_json = '{\"article_id\": \"1407.3051\", \"article_text\": [\"the missing - mass resolution was obtained to be 2.8 @xmath3 0.1 mev\/@xmath4 ( fwhm ) , which corresponds to the missing - mass resolution of 3.2 @xmath3 0.2 mev\/@xmath4 ( fwhm ) at the @xmath6 cusp region in the @xmath0 reaction .\", \"this resolution is at least by a factor of 2 better than the previous measurement with the same reaction ( 3.2@xmath595.5 mev\/@xmath4 in @xmath84 ) @xcite .\", \"after the missing - mass scale adjustment , the validity of the corrections was tested in the @xmath85 productions at 1.69 gev\/@xmath1 . in fig . [\", \"fig : calibrations ] ( a ) , we show the missing - mass spectrum in the @xmath86 region in the @xmath87 reaction at 1.69 gev\/@xmath1 . a fitting result with a lorentzian function for the @xmath86 ( dashed line ) and the three - body phas' \r\n# The line of code from `scientific_papers.py` which appears to cause the error\r\njson.loads(invalid_json)\r\n```\r\n\r\nThis is as far as I get before I am stumped.","embeddings":[-0.0581401661,0.0877689496,-0.0419098586,0.1886685789,0.2178975642,0.0454704054,-0.0241877399,0.3616167009,0.0815683678,-0.1818822473,-0.0505227782,0.691057086,-0.1919905692,-0.2139135003,-0.1082298383,0.0094932541,-0.1276049614,0.0114534106,0.1146815494,-0.1110842973,-0.1230693087,0.1993306875,-0.1573242992,-0.0993064716,0.4090956151,0.0958301648,0.1364857703,0.3957486153,-0.0027542866,-0.3556491733,0.1349909604,0.0240957309,0.2171578258,0.5375875831,-0.0001107034,0.2718124986,0.3134252429,-0.0939889029,-0.2043009251,-0.5840528011,-0.4115153253,-0.306637615,0.1335057318,-0.2136874497,-0.0360715277,-0.2607661784,0.1242143437,-0.1539173871,0.3288322389,0.3756518662,0.207921654,0.0798494294,0.3830094635,0.1164422184,-0.0670725107,-0.02313504,0.0021167931,0.4649830163,0.0422845446,0.1263236105,0.1612235159,0.1448102146,-0.1489691883,0.1127980724,0.2092457116,-0.0296654217,-0.1075037271,-0.1436872929,0.0375226587,0.1304574609,0.2721568644,-0.0974486545,-0.2403058112,0.0564082675,0.2218078375,-0.1019054577,0.295181632,0.430082649,-0.1124939024,0.0761579052,-0.3062997162,-0.3497585952,-0.4034608901,0.4736581147,-0.0218985379,-0.1115582362,-0.200718835,0.2033873051,0.1595942974,-0.2415931225,0.0368958823,0.0939481407,0.1707556397,0.11694552,-0.4864809811,0.2469767183,0.0796729624,-0.3753705025,0.0753835738,-0.150146991,-0.1037795395,-0.0815444365,-0.2829775214,0.2728938162,0.1120163128,0.018602727,-0.1558181047,-0.0018167398,0.0582292303,-0.0125903161,-0.1686109453,0.0329580121,-0.3595866859,-0.1540408581,-0.3627384901,-0.3801209033,0.2619623542,-0.3207472861,0.0621837713,-0.0250361618,-0.0250417013,0.0168774351,0.0238443669,0.228412956,-0.0542670265,0.1421990842,0.1716236323,0.4296216667,-0.3466945887,0.1758017391,-0.0509758741,0.0754750222,-0.1097994223,-0.2968219519,0.2044046372,0.0846939757,0.1635968983,-0.0225838032,0.0854115933,-0.201628834,0.1252433211,0.1339762807,-0.1017934233,0.4808894694,0.3299891949,0.0564984679,0.2672579885,0.0386227481,-0.1250129193,0.2215334028,-0.4679540396,-0.1227610484,-0.2679323256,0.162654683,-0.1997473538,-0.1317337751,-0.1605494618,-0.1432931125,0.1379330009,0.2394692749,-0.3634818196,-0.2293667346,0.0001346738,-0.3333007991,0.2488360554,0.0392104574,-0.3133456707,0.1613029242,-0.1056164876,0.1475966722,0.2834636867,0.3652399778,-0.2056614608,0.1426015943,-0.0255504828,-0.1135711595,0.3589995801,-0.3276819885,-0.1195762455,0.0364844464,-0.1393856555,0.076270476,-0.0743433014,0.0231164042,-0.1838792413,0.2579210103,0.2040625364,0.2711192966,0.0826367363,0.0304179639,-0.1805005819,-0.2009442002,0.45898664,0.3523835838,-0.0857428014,0.0294668283,-0.0705721304,-0.072366327,0.1135509834,-0.099117592,-0.3065351248,0.2054903805,0.0127701126,0.0597909503,0.0696247518,0.1258382499,0.1664613187,0.0967324823,0.2396769971,0.2965689003,-0.2159079313,0.3827721477,-0.3872411549,0.0926724225,-0.0310393795,0.0711300671,0.065654926,0.0439680256,-0.0934773609,0.1950520128,-0.1820219457,0.2183669358,-0.3807091713,-0.0371009409,-0.8843935132,0.0560739338,-0.1170641184,0.1107353494,0.2100989223,0.1402559727,-0.1191787124,-0.2400333434,-0.1165203899,0.2139781266,0.2940783203,-0.0699314922,0.2279684991,0.042080678,0.0583746955,-0.3201346993,0.2476847917,0.0171775892,0.0454673953,0.0980466679,-0.2400624603,0.3954666555,0.0848648995,0.1853234768,0.1478035152,0.1319267154,0.4758432508,-0.0060153292,-0.0275846701,-0.0089764493,0.4516296685,0.1805775911,-0.340288192,-0.1420488507,-0.1855588555,-0.0007562374,0.0774995834,0.1592724323,-0.0840466544,0.3020502329,-0.1074541733,0.0607566088,-0.1180131286,0.0291220751,0.3401017785,0.1709996909,-0.3174942732,0.0375214443,-0.2074111104,-0.0640699863,0.1576645225,0.0042496887,0.0024907775,-0.1562239826,0.130223617,0.1190572008,-0.2286053598,-0.2915285528,-0.0149250245,0.4036516249,-0.3374558687,0.0624639392,-0.3534377515,-0.2601831853,-0.0542715788,-0.1631045043,-0.1600035727,-0.2046508193,-0.1103503406,0.1549523771,0.1971672177,0.1569382101,-0.3404233754,-0.0687447339,0.2280444354,-0.212758258,0.2178190351,-0.3201696575,-0.090573445,0.0915043354,0.2567333579,0.115356423,0.2168773413,-0.0711518824,-0.1991698444,0.1951156259,-0.3111149967,-0.0690635741,-0.2529280484,0.1622450054,-0.0127729326,0.2829394639,0.0663689077,-0.2645323873,0.2123935223,0.0352514498,-0.1587046087,0.302259326,0.0569703095,-0.4054757655,-0.1261679381,-0.6859310865,-0.2402991205,-0.2965233922,0.0207346324,0.2632628381,0.1707098782,0.4518840313,0.3053714931,0.021538198,0.0888195485,0.0629290193,-0.1581660807,-0.0866923556,0.2277038097,-0.1510714889,-0.4011754692,0.1772225201,0.2070526481,0.5639321804,0.1303535402,-0.4385875762,0.1577425152,-0.1061138362,-0.0755192265,-0.2182867229,0.1015386134,0.0663023368,-0.210977301,-0.0257533733,0.0079541886,-0.1914936155,-0.0301366076,0.1466688365,0.2630581856,0.2407557815,0.4278973341,-0.2681685686,0.5905352831,0.0142622264,-0.1609193683,0.069476828,-0.1589577198,0.4026772082,-0.0031177653,-0.5069510341,0.2158525139,0.284262687,0.0171935838,0.1744769961,0.0340316966,-0.2551556826,-0.1661703885,0.2379712909,-0.3800166845,-0.2908549011,0.196081847,-0.2611018419,0.07894703,0.0448552743,-0.0590058006,-0.0890889242,-0.1388099492,-0.1887180358,0.4358404875,0.064318262,0.0315002762,-0.4632994235,0.0505689085,-0.342972517,0.2074735016,-0.1587536782,0.6160392761,-0.1219579875,-0.0999062806,0.1105678827,-0.014026559,0.3819845617,0.0812925398,0.1661246568,0.3375768065,-0.1239420772,-0.4966859519,-0.0414071269,0.0024599754,0.4472510517,0.4753023982,0.1587123126,0.0748985484,-0.1119315773,-0.1855520904,0.1059326828,-0.0940444842,-0.2817312777,-0.3055519164,-0.3186794221,-0.2693054676,-0.4744183123,-0.1720836312,0.2304685116,-0.0999512672,0.1676111668,-0.1212368608,-0.216672942,0.149696812,0.3861213326,0.0398848876,-0.0878831297,0.1504201591,0.215640977,-0.1163051724,0.1365493387,0.3614488244,0.1097372323,-0.7085859179,-0.0461754575,-0.0668326467,0.1575988531,-0.2619423568,-0.1674872935,-0.1746036857,0.076938197,-0.1360736489,0.002871929,0.1660889685,0.2037215233,0.0454277992,-0.204467237,-0.2929310203,0.2632623315,-0.1085660085,0.0342324302,0.1920119822,-0.2135142386,-0.1601367593,0.3781900108,-0.021024093,0.5590004325,-0.3097172678,0.1412570328,0.2761947513,0.0139164403,0.6699483395,-0.1115628928,0.2104437798,-0.3174932301,-0.0141911358,-0.0565616973,-0.2007324994,0.026437737,0.1123935506,-0.1216480434,0.1802656204,-0.0242364183,-0.2814278305,0.1291989833,0.2311947495,0.5022940636,-0.244632259,-0.7192257047,0.2023690343,-0.217751801,0.0522673614,0.0517729037,-0.1352651715,-0.2248557061,0.0489743501,-0.3124033511,0.3894600272,0.0901355222,0.3495639861,0.3082270324,-0.0468276963,0.0555090979,0.2309151143,0.1750369817,0.227895841,-0.1704107672,0.2341102362,-0.0087976633,-0.2591122687,0.248261109,0.1001886129,0.1886753291,-0.1627509743,-0.2001486868,0.1331465244,0.0853323266,-0.5239751339,0.1386449933,0.1783330888,-0.0720260888,0.0992175639,-0.2829616666,0.1097199842,0.1071184203,0.2075476944,0.1173229814,0.3670862317,0.0736812353,0.3407841921,0.0079562264,-0.4690263867,0.1429866999,0.2529992759,0.1792363971,0.0090548825,0.3015254736,0.142531231,0.0471658967,-0.1045872048,-0.1746542007,-0.461915791,0.010912667,-0.1224688143,0.3127566576,0.2229425609,0.1521368474,-0.1818508506,0.0410407372,-0.2236406952,0.0997594744,-0.859870553,-0.1354100406,-0.0548281781,-0.4878775179,-0.0006446862,0.2166990638,-0.0469971113,-0.0847147852,-0.1913502067,-0.2497873902,0.0197665431,-0.2032182515,-0.1099875793,0.2819565833,0.0619983673,0.1283500195,-0.2295306325,0.1299589574,-0.4033862948,-0.1084756553,-0.198766306,-0.0028026998,0.1285460293,0.052987881,-0.2573785782,-0.2084369361,-0.2459956408,-0.3299534023,0.1025005355,-0.040987689,-0.0201559719,-0.154088214,0.0581756756,-0.5069238544,0.0165901165,0.1186364442,0.3586595356,-0.1164739281,0.2467471063,0.2674191296,0.2644314468,-0.1753063053,-0.0909070447,-0.5406355262,-0.066910632,-0.2757559121,0.1637987047,0.3702971339,-0.1089690551,0.1324842572,0.0816420168,0.0535019226,0.5262340903,-0.1782574058,-0.0461499244,0.1299265176,0.1488399208,-0.2839462459,-0.1335304677,0.3298020065,0.309981674,-0.2143112719,0.2018111199,-0.0222959127,0.0226881001,0.1296839714,0.0303221513,0.5294115543,-0.2493306249,0.0884126201,-0.151979059,-0.2476819903,-0.0916667953,-0.2119928598,0.2499311864,0.2411658913,0.372854948,-0.0014118456,0.1080931425,0.1908807456,-0.1238163337,-0.013418478,-0.3544937372,-0.464689672,0.1635971665,0.0233234875,0.3950412869,0.0079648141,0.0928862691,0.0519949831,-0.2455949187,-0.065728046,0.237026006,0.1193934754,-0.000631774,-0.2060878128,0.0416782387,-0.0317267925,-0.1816642731,0.037583489,-0.1187287867,0.0865239799,0.3490435183,-0.3035861552,-0.7911368608,0.2285946459,0.0662597418,-0.1032067463,0.0720451027,0.3918698132,0.023687683,0.0435211957,0.3245864809,0.1088312864,0.5102598071,0.4114189446,-0.0492335819,-0.0457732938,0.0242724456,0.1189988926,-0.0850605667,0.3223448098,0.033200141,0.3221098781,0.2746065557,0.1198492125,0.0574153103,0.2889924943,0.1625372469,0.0712472424,0.0226034895,0.0666723996,0.1023148671,0.2394506186,-0.2568710446,0.0724869147,-0.3361437619,-0.1103490889,0.2590260506,0.2074869275,0.1140956059,0.0336433798,0.0582498871,-0.1319148839,0.7058607936,0.510782063,0.371773243,-0.1432923228,-0.0604136288,-0.5617278218,0.1689248532,-0.2268105,-0.209326908,0.1078344584,0.1282417923,0.4357310832,0.1924453825,-0.027367983,-0.1725472957,-0.400547415,-0.0237439033,-0.342823118,-0.2178272456,0.0078974469,-0.1119740531,-0.1383003145,-0.0180382617,0.0899634212,-0.1823913902,-0.0019597169,0.0270195864,0.1578267366,-0.0332669131,0.0914950147,0.4042009115,0.015184517,0.4290499091,-0.0270545203,-0.2645837665,-0.4509984851,-0.1991005838,-0.3306635618,0.3517012894,0.2352200449,0.1762435734,-0.1694796979,0.1872226745,-0.168528229,0.4838164151,-0.0565907434,-0.5422750115,0.1255182773,0.073942557,0.0957553387,0.1444119662,-0.0106463991,0.053534545,0.0044792788,0.5422230959,-0.2458273768,-0.1378215253,0.2703121006,-0.2595653236,-0.3543881774,0.2645863295,-0.1342298537,-0.3655931354,0.3093405664,-0.5371618867,0.1654866189,0.3419696987,-0.1092510149,-0.1710449755,-0.0504334494,0.1009506658,-0.2060299069,-0.0858500078,0.5699597597,-0.119342804,-0.2334057987,-0.0469614193,0.0090484619]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/294","title":"Cannot load arxiv dataset on MacOS?","comments":"I just checked inside `train.txt` and this line isn't truncated for me (line 163577).\r\nCould you try to clear your cache and re-download the dataset ?","body":"I am having trouble loading the `\"arxiv\"` config from the `\"scientific_papers\"` dataset on MacOS. When I try loading the dataset with:\r\n\r\n```python\r\narxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n```\r\n\r\nI get the following stack trace:\r\n\r\n```bash\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-2-8e00c55d5a59> in <module>\r\n----> 1 arxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 662 \r\n 663 generator = self._generate_examples(**split_generator.gen_kwargs)\r\n--> 664 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 665 example = self.info.features.encode_example(record)\r\n 666 writer.write(example)\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1106 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1107 \r\n-> 1108 for obj in iterable:\r\n 1109 yield obj\r\n 1110 # Update and possibly print the progressbar.\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/datasets\/scientific_papers\/107a416c0e1958cb846f5934b5aae292f7884a5b27e86af3f3ef1a093e058bbc\/scientific_papers.py in _generate_examples(self, path)\r\n 114 # \"section_names\": list[str], list of section names.\r\n 115 # \"sections\": list[list[str]], list of sections (list of paragraphs)\r\n--> 116 d = json.loads(line)\r\n 117 summary = \"\\n\".join(d[\"abstract_text\"])\r\n 118 # In original paper, <S> and <\/S> are not used in vocab during training\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)\r\n 346 parse_int is None and parse_float is None and\r\n 347 parse_constant is None and object_pairs_hook is None and not kw):\r\n--> 348 return _default_decoder.decode(s)\r\n 349 if cls is None:\r\n 350 cls = JSONDecoder\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in decode(self, s, _w)\r\n 335 \r\n 336 \"\"\"\r\n--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())\r\n 338 end = _w(s, end).end()\r\n 339 if end != len(s):\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in raw_decode(self, s, idx)\r\n 351 \"\"\"\r\n 352 try:\r\n--> 353 obj, end = self.scan_once(s, idx)\r\n 354 except StopIteration as err:\r\n 355 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 46983 (char 46982)\r\n\r\n163502 examples [02:10, 2710.68 examples\/s] \r\n```\r\n\r\nI am not sure how to trace back to the specific JSON file that has the \"Unterminated string\". Also, I do not get this error on colab so I suspect it may be MacOS specific. Copy pasting the relevant lines from `transformers-cli env` below:\r\n\r\n- Platform: Darwin-19.5.0-x86_64-i386-64bit\r\n- Python version: 3.7.5\r\n- PyTorch version (GPU?): 1.5.0 (False)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n\r\nAny ideas?","comment_length":26,"text":"Cannot load arxiv dataset on MacOS?\nI am having trouble loading the `\"arxiv\"` config from the `\"scientific_papers\"` dataset on MacOS. When I try loading the dataset with:\r\n\r\n```python\r\narxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n```\r\n\r\nI get the following stack trace:\r\n\r\n```bash\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-2-8e00c55d5a59> in <module>\r\n----> 1 arxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 662 \r\n 663 generator = self._generate_examples(**split_generator.gen_kwargs)\r\n--> 664 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 665 example = self.info.features.encode_example(record)\r\n 666 writer.write(example)\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1106 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1107 \r\n-> 1108 for obj in iterable:\r\n 1109 yield obj\r\n 1110 # Update and possibly print the progressbar.\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/datasets\/scientific_papers\/107a416c0e1958cb846f5934b5aae292f7884a5b27e86af3f3ef1a093e058bbc\/scientific_papers.py in _generate_examples(self, path)\r\n 114 # \"section_names\": list[str], list of section names.\r\n 115 # \"sections\": list[list[str]], list of sections (list of paragraphs)\r\n--> 116 d = json.loads(line)\r\n 117 summary = \"\\n\".join(d[\"abstract_text\"])\r\n 118 # In original paper, <S> and <\/S> are not used in vocab during training\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)\r\n 346 parse_int is None and parse_float is None and\r\n 347 parse_constant is None and object_pairs_hook is None and not kw):\r\n--> 348 return _default_decoder.decode(s)\r\n 349 if cls is None:\r\n 350 cls = JSONDecoder\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in decode(self, s, _w)\r\n 335 \r\n 336 \"\"\"\r\n--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())\r\n 338 end = _w(s, end).end()\r\n 339 if end != len(s):\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in raw_decode(self, s, idx)\r\n 351 \"\"\"\r\n 352 try:\r\n--> 353 obj, end = self.scan_once(s, idx)\r\n 354 except StopIteration as err:\r\n 355 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 46983 (char 46982)\r\n\r\n163502 examples [02:10, 2710.68 examples\/s] \r\n```\r\n\r\nI am not sure how to trace back to the specific JSON file that has the \"Unterminated string\". Also, I do not get this error on colab so I suspect it may be MacOS specific. Copy pasting the relevant lines from `transformers-cli env` below:\r\n\r\n- Platform: Darwin-19.5.0-x86_64-i386-64bit\r\n- Python version: 3.7.5\r\n- PyTorch version (GPU?): 1.5.0 (False)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n\r\nAny ideas?\nI just checked inside `train.txt` and this line isn't truncated for me (line 163577).\r\nCould you try to clear your cache and re-download the dataset ?","embeddings":[-0.0581401661,0.0877689496,-0.0419098586,0.1886685789,0.2178975642,0.0454704054,-0.0241877399,0.3616167009,0.0815683678,-0.1818822473,-0.0505227782,0.691057086,-0.1919905692,-0.2139135003,-0.1082298383,0.0094932541,-0.1276049614,0.0114534106,0.1146815494,-0.1110842973,-0.1230693087,0.1993306875,-0.1573242992,-0.0993064716,0.4090956151,0.0958301648,0.1364857703,0.3957486153,-0.0027542866,-0.3556491733,0.1349909604,0.0240957309,0.2171578258,0.5375875831,-0.0001107034,0.2718124986,0.3134252429,-0.0939889029,-0.2043009251,-0.5840528011,-0.4115153253,-0.306637615,0.1335057318,-0.2136874497,-0.0360715277,-0.2607661784,0.1242143437,-0.1539173871,0.3288322389,0.3756518662,0.207921654,0.0798494294,0.3830094635,0.1164422184,-0.0670725107,-0.02313504,0.0021167931,0.4649830163,0.0422845446,0.1263236105,0.1612235159,0.1448102146,-0.1489691883,0.1127980724,0.2092457116,-0.0296654217,-0.1075037271,-0.1436872929,0.0375226587,0.1304574609,0.2721568644,-0.0974486545,-0.2403058112,0.0564082675,0.2218078375,-0.1019054577,0.295181632,0.430082649,-0.1124939024,0.0761579052,-0.3062997162,-0.3497585952,-0.4034608901,0.4736581147,-0.0218985379,-0.1115582362,-0.200718835,0.2033873051,0.1595942974,-0.2415931225,0.0368958823,0.0939481407,0.1707556397,0.11694552,-0.4864809811,0.2469767183,0.0796729624,-0.3753705025,0.0753835738,-0.150146991,-0.1037795395,-0.0815444365,-0.2829775214,0.2728938162,0.1120163128,0.018602727,-0.1558181047,-0.0018167398,0.0582292303,-0.0125903161,-0.1686109453,0.0329580121,-0.3595866859,-0.1540408581,-0.3627384901,-0.3801209033,0.2619623542,-0.3207472861,0.0621837713,-0.0250361618,-0.0250417013,0.0168774351,0.0238443669,0.228412956,-0.0542670265,0.1421990842,0.1716236323,0.4296216667,-0.3466945887,0.1758017391,-0.0509758741,0.0754750222,-0.1097994223,-0.2968219519,0.2044046372,0.0846939757,0.1635968983,-0.0225838032,0.0854115933,-0.201628834,0.1252433211,0.1339762807,-0.1017934233,0.4808894694,0.3299891949,0.0564984679,0.2672579885,0.0386227481,-0.1250129193,0.2215334028,-0.4679540396,-0.1227610484,-0.2679323256,0.162654683,-0.1997473538,-0.1317337751,-0.1605494618,-0.1432931125,0.1379330009,0.2394692749,-0.3634818196,-0.2293667346,0.0001346738,-0.3333007991,0.2488360554,0.0392104574,-0.3133456707,0.1613029242,-0.1056164876,0.1475966722,0.2834636867,0.3652399778,-0.2056614608,0.1426015943,-0.0255504828,-0.1135711595,0.3589995801,-0.3276819885,-0.1195762455,0.0364844464,-0.1393856555,0.076270476,-0.0743433014,0.0231164042,-0.1838792413,0.2579210103,0.2040625364,0.2711192966,0.0826367363,0.0304179639,-0.1805005819,-0.2009442002,0.45898664,0.3523835838,-0.0857428014,0.0294668283,-0.0705721304,-0.072366327,0.1135509834,-0.099117592,-0.3065351248,0.2054903805,0.0127701126,0.0597909503,0.0696247518,0.1258382499,0.1664613187,0.0967324823,0.2396769971,0.2965689003,-0.2159079313,0.3827721477,-0.3872411549,0.0926724225,-0.0310393795,0.0711300671,0.065654926,0.0439680256,-0.0934773609,0.1950520128,-0.1820219457,0.2183669358,-0.3807091713,-0.0371009409,-0.8843935132,0.0560739338,-0.1170641184,0.1107353494,0.2100989223,0.1402559727,-0.1191787124,-0.2400333434,-0.1165203899,0.2139781266,0.2940783203,-0.0699314922,0.2279684991,0.042080678,0.0583746955,-0.3201346993,0.2476847917,0.0171775892,0.0454673953,0.0980466679,-0.2400624603,0.3954666555,0.0848648995,0.1853234768,0.1478035152,0.1319267154,0.4758432508,-0.0060153292,-0.0275846701,-0.0089764493,0.4516296685,0.1805775911,-0.340288192,-0.1420488507,-0.1855588555,-0.0007562374,0.0774995834,0.1592724323,-0.0840466544,0.3020502329,-0.1074541733,0.0607566088,-0.1180131286,0.0291220751,0.3401017785,0.1709996909,-0.3174942732,0.0375214443,-0.2074111104,-0.0640699863,0.1576645225,0.0042496887,0.0024907775,-0.1562239826,0.130223617,0.1190572008,-0.2286053598,-0.2915285528,-0.0149250245,0.4036516249,-0.3374558687,0.0624639392,-0.3534377515,-0.2601831853,-0.0542715788,-0.1631045043,-0.1600035727,-0.2046508193,-0.1103503406,0.1549523771,0.1971672177,0.1569382101,-0.3404233754,-0.0687447339,0.2280444354,-0.212758258,0.2178190351,-0.3201696575,-0.090573445,0.0915043354,0.2567333579,0.115356423,0.2168773413,-0.0711518824,-0.1991698444,0.1951156259,-0.3111149967,-0.0690635741,-0.2529280484,0.1622450054,-0.0127729326,0.2829394639,0.0663689077,-0.2645323873,0.2123935223,0.0352514498,-0.1587046087,0.302259326,0.0569703095,-0.4054757655,-0.1261679381,-0.6859310865,-0.2402991205,-0.2965233922,0.0207346324,0.2632628381,0.1707098782,0.4518840313,0.3053714931,0.021538198,0.0888195485,0.0629290193,-0.1581660807,-0.0866923556,0.2277038097,-0.1510714889,-0.4011754692,0.1772225201,0.2070526481,0.5639321804,0.1303535402,-0.4385875762,0.1577425152,-0.1061138362,-0.0755192265,-0.2182867229,0.1015386134,0.0663023368,-0.210977301,-0.0257533733,0.0079541886,-0.1914936155,-0.0301366076,0.1466688365,0.2630581856,0.2407557815,0.4278973341,-0.2681685686,0.5905352831,0.0142622264,-0.1609193683,0.069476828,-0.1589577198,0.4026772082,-0.0031177653,-0.5069510341,0.2158525139,0.284262687,0.0171935838,0.1744769961,0.0340316966,-0.2551556826,-0.1661703885,0.2379712909,-0.3800166845,-0.2908549011,0.196081847,-0.2611018419,0.07894703,0.0448552743,-0.0590058006,-0.0890889242,-0.1388099492,-0.1887180358,0.4358404875,0.064318262,0.0315002762,-0.4632994235,0.0505689085,-0.342972517,0.2074735016,-0.1587536782,0.6160392761,-0.1219579875,-0.0999062806,0.1105678827,-0.014026559,0.3819845617,0.0812925398,0.1661246568,0.3375768065,-0.1239420772,-0.4966859519,-0.0414071269,0.0024599754,0.4472510517,0.4753023982,0.1587123126,0.0748985484,-0.1119315773,-0.1855520904,0.1059326828,-0.0940444842,-0.2817312777,-0.3055519164,-0.3186794221,-0.2693054676,-0.4744183123,-0.1720836312,0.2304685116,-0.0999512672,0.1676111668,-0.1212368608,-0.216672942,0.149696812,0.3861213326,0.0398848876,-0.0878831297,0.1504201591,0.215640977,-0.1163051724,0.1365493387,0.3614488244,0.1097372323,-0.7085859179,-0.0461754575,-0.0668326467,0.1575988531,-0.2619423568,-0.1674872935,-0.1746036857,0.076938197,-0.1360736489,0.002871929,0.1660889685,0.2037215233,0.0454277992,-0.204467237,-0.2929310203,0.2632623315,-0.1085660085,0.0342324302,0.1920119822,-0.2135142386,-0.1601367593,0.3781900108,-0.021024093,0.5590004325,-0.3097172678,0.1412570328,0.2761947513,0.0139164403,0.6699483395,-0.1115628928,0.2104437798,-0.3174932301,-0.0141911358,-0.0565616973,-0.2007324994,0.026437737,0.1123935506,-0.1216480434,0.1802656204,-0.0242364183,-0.2814278305,0.1291989833,0.2311947495,0.5022940636,-0.244632259,-0.7192257047,0.2023690343,-0.217751801,0.0522673614,0.0517729037,-0.1352651715,-0.2248557061,0.0489743501,-0.3124033511,0.3894600272,0.0901355222,0.3495639861,0.3082270324,-0.0468276963,0.0555090979,0.2309151143,0.1750369817,0.227895841,-0.1704107672,0.2341102362,-0.0087976633,-0.2591122687,0.248261109,0.1001886129,0.1886753291,-0.1627509743,-0.2001486868,0.1331465244,0.0853323266,-0.5239751339,0.1386449933,0.1783330888,-0.0720260888,0.0992175639,-0.2829616666,0.1097199842,0.1071184203,0.2075476944,0.1173229814,0.3670862317,0.0736812353,0.3407841921,0.0079562264,-0.4690263867,0.1429866999,0.2529992759,0.1792363971,0.0090548825,0.3015254736,0.142531231,0.0471658967,-0.1045872048,-0.1746542007,-0.461915791,0.010912667,-0.1224688143,0.3127566576,0.2229425609,0.1521368474,-0.1818508506,0.0410407372,-0.2236406952,0.0997594744,-0.859870553,-0.1354100406,-0.0548281781,-0.4878775179,-0.0006446862,0.2166990638,-0.0469971113,-0.0847147852,-0.1913502067,-0.2497873902,0.0197665431,-0.2032182515,-0.1099875793,0.2819565833,0.0619983673,0.1283500195,-0.2295306325,0.1299589574,-0.4033862948,-0.1084756553,-0.198766306,-0.0028026998,0.1285460293,0.052987881,-0.2573785782,-0.2084369361,-0.2459956408,-0.3299534023,0.1025005355,-0.040987689,-0.0201559719,-0.154088214,0.0581756756,-0.5069238544,0.0165901165,0.1186364442,0.3586595356,-0.1164739281,0.2467471063,0.2674191296,0.2644314468,-0.1753063053,-0.0909070447,-0.5406355262,-0.066910632,-0.2757559121,0.1637987047,0.3702971339,-0.1089690551,0.1324842572,0.0816420168,0.0535019226,0.5262340903,-0.1782574058,-0.0461499244,0.1299265176,0.1488399208,-0.2839462459,-0.1335304677,0.3298020065,0.309981674,-0.2143112719,0.2018111199,-0.0222959127,0.0226881001,0.1296839714,0.0303221513,0.5294115543,-0.2493306249,0.0884126201,-0.151979059,-0.2476819903,-0.0916667953,-0.2119928598,0.2499311864,0.2411658913,0.372854948,-0.0014118456,0.1080931425,0.1908807456,-0.1238163337,-0.013418478,-0.3544937372,-0.464689672,0.1635971665,0.0233234875,0.3950412869,0.0079648141,0.0928862691,0.0519949831,-0.2455949187,-0.065728046,0.237026006,0.1193934754,-0.000631774,-0.2060878128,0.0416782387,-0.0317267925,-0.1816642731,0.037583489,-0.1187287867,0.0865239799,0.3490435183,-0.3035861552,-0.7911368608,0.2285946459,0.0662597418,-0.1032067463,0.0720451027,0.3918698132,0.023687683,0.0435211957,0.3245864809,0.1088312864,0.5102598071,0.4114189446,-0.0492335819,-0.0457732938,0.0242724456,0.1189988926,-0.0850605667,0.3223448098,0.033200141,0.3221098781,0.2746065557,0.1198492125,0.0574153103,0.2889924943,0.1625372469,0.0712472424,0.0226034895,0.0666723996,0.1023148671,0.2394506186,-0.2568710446,0.0724869147,-0.3361437619,-0.1103490889,0.2590260506,0.2074869275,0.1140956059,0.0336433798,0.0582498871,-0.1319148839,0.7058607936,0.510782063,0.371773243,-0.1432923228,-0.0604136288,-0.5617278218,0.1689248532,-0.2268105,-0.209326908,0.1078344584,0.1282417923,0.4357310832,0.1924453825,-0.027367983,-0.1725472957,-0.400547415,-0.0237439033,-0.342823118,-0.2178272456,0.0078974469,-0.1119740531,-0.1383003145,-0.0180382617,0.0899634212,-0.1823913902,-0.0019597169,0.0270195864,0.1578267366,-0.0332669131,0.0914950147,0.4042009115,0.015184517,0.4290499091,-0.0270545203,-0.2645837665,-0.4509984851,-0.1991005838,-0.3306635618,0.3517012894,0.2352200449,0.1762435734,-0.1694796979,0.1872226745,-0.168528229,0.4838164151,-0.0565907434,-0.5422750115,0.1255182773,0.073942557,0.0957553387,0.1444119662,-0.0106463991,0.053534545,0.0044792788,0.5422230959,-0.2458273768,-0.1378215253,0.2703121006,-0.2595653236,-0.3543881774,0.2645863295,-0.1342298537,-0.3655931354,0.3093405664,-0.5371618867,0.1654866189,0.3419696987,-0.1092510149,-0.1710449755,-0.0504334494,0.1009506658,-0.2060299069,-0.0858500078,0.5699597597,-0.119342804,-0.2334057987,-0.0469614193,0.0090484619]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/294","title":"Cannot load arxiv dataset on MacOS?","comments":"Ah the turn-it-off-turn-it-on again solution! That did it, thanks a lot :) ","body":"I am having trouble loading the `\"arxiv\"` config from the `\"scientific_papers\"` dataset on MacOS. When I try loading the dataset with:\r\n\r\n```python\r\narxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n```\r\n\r\nI get the following stack trace:\r\n\r\n```bash\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-2-8e00c55d5a59> in <module>\r\n----> 1 arxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 662 \r\n 663 generator = self._generate_examples(**split_generator.gen_kwargs)\r\n--> 664 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 665 example = self.info.features.encode_example(record)\r\n 666 writer.write(example)\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1106 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1107 \r\n-> 1108 for obj in iterable:\r\n 1109 yield obj\r\n 1110 # Update and possibly print the progressbar.\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/datasets\/scientific_papers\/107a416c0e1958cb846f5934b5aae292f7884a5b27e86af3f3ef1a093e058bbc\/scientific_papers.py in _generate_examples(self, path)\r\n 114 # \"section_names\": list[str], list of section names.\r\n 115 # \"sections\": list[list[str]], list of sections (list of paragraphs)\r\n--> 116 d = json.loads(line)\r\n 117 summary = \"\\n\".join(d[\"abstract_text\"])\r\n 118 # In original paper, <S> and <\/S> are not used in vocab during training\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)\r\n 346 parse_int is None and parse_float is None and\r\n 347 parse_constant is None and object_pairs_hook is None and not kw):\r\n--> 348 return _default_decoder.decode(s)\r\n 349 if cls is None:\r\n 350 cls = JSONDecoder\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in decode(self, s, _w)\r\n 335 \r\n 336 \"\"\"\r\n--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())\r\n 338 end = _w(s, end).end()\r\n 339 if end != len(s):\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in raw_decode(self, s, idx)\r\n 351 \"\"\"\r\n 352 try:\r\n--> 353 obj, end = self.scan_once(s, idx)\r\n 354 except StopIteration as err:\r\n 355 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 46983 (char 46982)\r\n\r\n163502 examples [02:10, 2710.68 examples\/s] \r\n```\r\n\r\nI am not sure how to trace back to the specific JSON file that has the \"Unterminated string\". Also, I do not get this error on colab so I suspect it may be MacOS specific. Copy pasting the relevant lines from `transformers-cli env` below:\r\n\r\n- Platform: Darwin-19.5.0-x86_64-i386-64bit\r\n- Python version: 3.7.5\r\n- PyTorch version (GPU?): 1.5.0 (False)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n\r\nAny ideas?","comment_length":12,"text":"Cannot load arxiv dataset on MacOS?\nI am having trouble loading the `\"arxiv\"` config from the `\"scientific_papers\"` dataset on MacOS. When I try loading the dataset with:\r\n\r\n```python\r\narxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n```\r\n\r\nI get the following stack trace:\r\n\r\n```bash\r\nJSONDecodeError Traceback (most recent call last)\r\n<ipython-input-2-8e00c55d5a59> in <module>\r\n----> 1 arxiv = nlp.load_dataset(\"scientific_papers\", \"arxiv\")\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 522 download_mode=download_mode,\r\n 523 ignore_verifications=ignore_verifications,\r\n--> 524 save_infos=save_infos,\r\n 525 )\r\n 526 \r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 430 verify_infos = not save_infos and not ignore_verifications\r\n 431 self._download_and_prepare(\r\n--> 432 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 433 )\r\n 434 # Sync info\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 481 try:\r\n 482 # Prepare split will record examples associated to the split\r\n--> 483 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 484 except OSError:\r\n 485 raise OSError(\"Cannot find data file. \" + (self.manual_download_instructions or \"\"))\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 662 \r\n 663 generator = self._generate_examples(**split_generator.gen_kwargs)\r\n--> 664 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 665 example = self.info.features.encode_example(record)\r\n 666 writer.write(example)\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/tqdm\/std.py in __iter__(self)\r\n 1106 fp_write=getattr(self.fp, 'write', sys.stderr.write))\r\n 1107 \r\n-> 1108 for obj in iterable:\r\n 1109 yield obj\r\n 1110 # Update and possibly print the progressbar.\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/site-packages\/nlp\/datasets\/scientific_papers\/107a416c0e1958cb846f5934b5aae292f7884a5b27e86af3f3ef1a093e058bbc\/scientific_papers.py in _generate_examples(self, path)\r\n 114 # \"section_names\": list[str], list of section names.\r\n 115 # \"sections\": list[list[str]], list of sections (list of paragraphs)\r\n--> 116 d = json.loads(line)\r\n 117 summary = \"\\n\".join(d[\"abstract_text\"])\r\n 118 # In original paper, <S> and <\/S> are not used in vocab during training\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)\r\n 346 parse_int is None and parse_float is None and\r\n 347 parse_constant is None and object_pairs_hook is None and not kw):\r\n--> 348 return _default_decoder.decode(s)\r\n 349 if cls is None:\r\n 350 cls = JSONDecoder\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in decode(self, s, _w)\r\n 335 \r\n 336 \"\"\"\r\n--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())\r\n 338 end = _w(s, end).end()\r\n 339 if end != len(s):\r\n\r\n~\/miniconda3\/envs\/t2t\/lib\/python3.7\/json\/decoder.py in raw_decode(self, s, idx)\r\n 351 \"\"\"\r\n 352 try:\r\n--> 353 obj, end = self.scan_once(s, idx)\r\n 354 except StopIteration as err:\r\n 355 raise JSONDecodeError(\"Expecting value\", s, err.value) from None\r\n\r\nJSONDecodeError: Unterminated string starting at: line 1 column 46983 (char 46982)\r\n\r\n163502 examples [02:10, 2710.68 examples\/s] \r\n```\r\n\r\nI am not sure how to trace back to the specific JSON file that has the \"Unterminated string\". Also, I do not get this error on colab so I suspect it may be MacOS specific. Copy pasting the relevant lines from `transformers-cli env` below:\r\n\r\n- Platform: Darwin-19.5.0-x86_64-i386-64bit\r\n- Python version: 3.7.5\r\n- PyTorch version (GPU?): 1.5.0 (False)\r\n- Tensorflow version (GPU?): 2.2.0 (False)\r\n\r\nAny ideas?\nAh the turn-it-off-turn-it-on again solution! That did it, thanks a lot :) ","embeddings":[-0.0581401661,0.0877689496,-0.0419098586,0.1886685789,0.2178975642,0.0454704054,-0.0241877399,0.3616167009,0.0815683678,-0.1818822473,-0.0505227782,0.691057086,-0.1919905692,-0.2139135003,-0.1082298383,0.0094932541,-0.1276049614,0.0114534106,0.1146815494,-0.1110842973,-0.1230693087,0.1993306875,-0.1573242992,-0.0993064716,0.4090956151,0.0958301648,0.1364857703,0.3957486153,-0.0027542866,-0.3556491733,0.1349909604,0.0240957309,0.2171578258,0.5375875831,-0.0001107034,0.2718124986,0.3134252429,-0.0939889029,-0.2043009251,-0.5840528011,-0.4115153253,-0.306637615,0.1335057318,-0.2136874497,-0.0360715277,-0.2607661784,0.1242143437,-0.1539173871,0.3288322389,0.3756518662,0.207921654,0.0798494294,0.3830094635,0.1164422184,-0.0670725107,-0.02313504,0.0021167931,0.4649830163,0.0422845446,0.1263236105,0.1612235159,0.1448102146,-0.1489691883,0.1127980724,0.2092457116,-0.0296654217,-0.1075037271,-0.1436872929,0.0375226587,0.1304574609,0.2721568644,-0.0974486545,-0.2403058112,0.0564082675,0.2218078375,-0.1019054577,0.295181632,0.430082649,-0.1124939024,0.0761579052,-0.3062997162,-0.3497585952,-0.4034608901,0.4736581147,-0.0218985379,-0.1115582362,-0.200718835,0.2033873051,0.1595942974,-0.2415931225,0.0368958823,0.0939481407,0.1707556397,0.11694552,-0.4864809811,0.2469767183,0.0796729624,-0.3753705025,0.0753835738,-0.150146991,-0.1037795395,-0.0815444365,-0.2829775214,0.2728938162,0.1120163128,0.018602727,-0.1558181047,-0.0018167398,0.0582292303,-0.0125903161,-0.1686109453,0.0329580121,-0.3595866859,-0.1540408581,-0.3627384901,-0.3801209033,0.2619623542,-0.3207472861,0.0621837713,-0.0250361618,-0.0250417013,0.0168774351,0.0238443669,0.228412956,-0.0542670265,0.1421990842,0.1716236323,0.4296216667,-0.3466945887,0.1758017391,-0.0509758741,0.0754750222,-0.1097994223,-0.2968219519,0.2044046372,0.0846939757,0.1635968983,-0.0225838032,0.0854115933,-0.201628834,0.1252433211,0.1339762807,-0.1017934233,0.4808894694,0.3299891949,0.0564984679,0.2672579885,0.0386227481,-0.1250129193,0.2215334028,-0.4679540396,-0.1227610484,-0.2679323256,0.162654683,-0.1997473538,-0.1317337751,-0.1605494618,-0.1432931125,0.1379330009,0.2394692749,-0.3634818196,-0.2293667346,0.0001346738,-0.3333007991,0.2488360554,0.0392104574,-0.3133456707,0.1613029242,-0.1056164876,0.1475966722,0.2834636867,0.3652399778,-0.2056614608,0.1426015943,-0.0255504828,-0.1135711595,0.3589995801,-0.3276819885,-0.1195762455,0.0364844464,-0.1393856555,0.076270476,-0.0743433014,0.0231164042,-0.1838792413,0.2579210103,0.2040625364,0.2711192966,0.0826367363,0.0304179639,-0.1805005819,-0.2009442002,0.45898664,0.3523835838,-0.0857428014,0.0294668283,-0.0705721304,-0.072366327,0.1135509834,-0.099117592,-0.3065351248,0.2054903805,0.0127701126,0.0597909503,0.0696247518,0.1258382499,0.1664613187,0.0967324823,0.2396769971,0.2965689003,-0.2159079313,0.3827721477,-0.3872411549,0.0926724225,-0.0310393795,0.0711300671,0.065654926,0.0439680256,-0.0934773609,0.1950520128,-0.1820219457,0.2183669358,-0.3807091713,-0.0371009409,-0.8843935132,0.0560739338,-0.1170641184,0.1107353494,0.2100989223,0.1402559727,-0.1191787124,-0.2400333434,-0.1165203899,0.2139781266,0.2940783203,-0.0699314922,0.2279684991,0.042080678,0.0583746955,-0.3201346993,0.2476847917,0.0171775892,0.0454673953,0.0980466679,-0.2400624603,0.3954666555,0.0848648995,0.1853234768,0.1478035152,0.1319267154,0.4758432508,-0.0060153292,-0.0275846701,-0.0089764493,0.4516296685,0.1805775911,-0.340288192,-0.1420488507,-0.1855588555,-0.0007562374,0.0774995834,0.1592724323,-0.0840466544,0.3020502329,-0.1074541733,0.0607566088,-0.1180131286,0.0291220751,0.3401017785,0.1709996909,-0.3174942732,0.0375214443,-0.2074111104,-0.0640699863,0.1576645225,0.0042496887,0.0024907775,-0.1562239826,0.130223617,0.1190572008,-0.2286053598,-0.2915285528,-0.0149250245,0.4036516249,-0.3374558687,0.0624639392,-0.3534377515,-0.2601831853,-0.0542715788,-0.1631045043,-0.1600035727,-0.2046508193,-0.1103503406,0.1549523771,0.1971672177,0.1569382101,-0.3404233754,-0.0687447339,0.2280444354,-0.212758258,0.2178190351,-0.3201696575,-0.090573445,0.0915043354,0.2567333579,0.115356423,0.2168773413,-0.0711518824,-0.1991698444,0.1951156259,-0.3111149967,-0.0690635741,-0.2529280484,0.1622450054,-0.0127729326,0.2829394639,0.0663689077,-0.2645323873,0.2123935223,0.0352514498,-0.1587046087,0.302259326,0.0569703095,-0.4054757655,-0.1261679381,-0.6859310865,-0.2402991205,-0.2965233922,0.0207346324,0.2632628381,0.1707098782,0.4518840313,0.3053714931,0.021538198,0.0888195485,0.0629290193,-0.1581660807,-0.0866923556,0.2277038097,-0.1510714889,-0.4011754692,0.1772225201,0.2070526481,0.5639321804,0.1303535402,-0.4385875762,0.1577425152,-0.1061138362,-0.0755192265,-0.2182867229,0.1015386134,0.0663023368,-0.210977301,-0.0257533733,0.0079541886,-0.1914936155,-0.0301366076,0.1466688365,0.2630581856,0.2407557815,0.4278973341,-0.2681685686,0.5905352831,0.0142622264,-0.1609193683,0.069476828,-0.1589577198,0.4026772082,-0.0031177653,-0.5069510341,0.2158525139,0.284262687,0.0171935838,0.1744769961,0.0340316966,-0.2551556826,-0.1661703885,0.2379712909,-0.3800166845,-0.2908549011,0.196081847,-0.2611018419,0.07894703,0.0448552743,-0.0590058006,-0.0890889242,-0.1388099492,-0.1887180358,0.4358404875,0.064318262,0.0315002762,-0.4632994235,0.0505689085,-0.342972517,0.2074735016,-0.1587536782,0.6160392761,-0.1219579875,-0.0999062806,0.1105678827,-0.014026559,0.3819845617,0.0812925398,0.1661246568,0.3375768065,-0.1239420772,-0.4966859519,-0.0414071269,0.0024599754,0.4472510517,0.4753023982,0.1587123126,0.0748985484,-0.1119315773,-0.1855520904,0.1059326828,-0.0940444842,-0.2817312777,-0.3055519164,-0.3186794221,-0.2693054676,-0.4744183123,-0.1720836312,0.2304685116,-0.0999512672,0.1676111668,-0.1212368608,-0.216672942,0.149696812,0.3861213326,0.0398848876,-0.0878831297,0.1504201591,0.215640977,-0.1163051724,0.1365493387,0.3614488244,0.1097372323,-0.7085859179,-0.0461754575,-0.0668326467,0.1575988531,-0.2619423568,-0.1674872935,-0.1746036857,0.076938197,-0.1360736489,0.002871929,0.1660889685,0.2037215233,0.0454277992,-0.204467237,-0.2929310203,0.2632623315,-0.1085660085,0.0342324302,0.1920119822,-0.2135142386,-0.1601367593,0.3781900108,-0.021024093,0.5590004325,-0.3097172678,0.1412570328,0.2761947513,0.0139164403,0.6699483395,-0.1115628928,0.2104437798,-0.3174932301,-0.0141911358,-0.0565616973,-0.2007324994,0.026437737,0.1123935506,-0.1216480434,0.1802656204,-0.0242364183,-0.2814278305,0.1291989833,0.2311947495,0.5022940636,-0.244632259,-0.7192257047,0.2023690343,-0.217751801,0.0522673614,0.0517729037,-0.1352651715,-0.2248557061,0.0489743501,-0.3124033511,0.3894600272,0.0901355222,0.3495639861,0.3082270324,-0.0468276963,0.0555090979,0.2309151143,0.1750369817,0.227895841,-0.1704107672,0.2341102362,-0.0087976633,-0.2591122687,0.248261109,0.1001886129,0.1886753291,-0.1627509743,-0.2001486868,0.1331465244,0.0853323266,-0.5239751339,0.1386449933,0.1783330888,-0.0720260888,0.0992175639,-0.2829616666,0.1097199842,0.1071184203,0.2075476944,0.1173229814,0.3670862317,0.0736812353,0.3407841921,0.0079562264,-0.4690263867,0.1429866999,0.2529992759,0.1792363971,0.0090548825,0.3015254736,0.142531231,0.0471658967,-0.1045872048,-0.1746542007,-0.461915791,0.010912667,-0.1224688143,0.3127566576,0.2229425609,0.1521368474,-0.1818508506,0.0410407372,-0.2236406952,0.0997594744,-0.859870553,-0.1354100406,-0.0548281781,-0.4878775179,-0.0006446862,0.2166990638,-0.0469971113,-0.0847147852,-0.1913502067,-0.2497873902,0.0197665431,-0.2032182515,-0.1099875793,0.2819565833,0.0619983673,0.1283500195,-0.2295306325,0.1299589574,-0.4033862948,-0.1084756553,-0.198766306,-0.0028026998,0.1285460293,0.052987881,-0.2573785782,-0.2084369361,-0.2459956408,-0.3299534023,0.1025005355,-0.040987689,-0.0201559719,-0.154088214,0.0581756756,-0.5069238544,0.0165901165,0.1186364442,0.3586595356,-0.1164739281,0.2467471063,0.2674191296,0.2644314468,-0.1753063053,-0.0909070447,-0.5406355262,-0.066910632,-0.2757559121,0.1637987047,0.3702971339,-0.1089690551,0.1324842572,0.0816420168,0.0535019226,0.5262340903,-0.1782574058,-0.0461499244,0.1299265176,0.1488399208,-0.2839462459,-0.1335304677,0.3298020065,0.309981674,-0.2143112719,0.2018111199,-0.0222959127,0.0226881001,0.1296839714,0.0303221513,0.5294115543,-0.2493306249,0.0884126201,-0.151979059,-0.2476819903,-0.0916667953,-0.2119928598,0.2499311864,0.2411658913,0.372854948,-0.0014118456,0.1080931425,0.1908807456,-0.1238163337,-0.013418478,-0.3544937372,-0.464689672,0.1635971665,0.0233234875,0.3950412869,0.0079648141,0.0928862691,0.0519949831,-0.2455949187,-0.065728046,0.237026006,0.1193934754,-0.000631774,-0.2060878128,0.0416782387,-0.0317267925,-0.1816642731,0.037583489,-0.1187287867,0.0865239799,0.3490435183,-0.3035861552,-0.7911368608,0.2285946459,0.0662597418,-0.1032067463,0.0720451027,0.3918698132,0.023687683,0.0435211957,0.3245864809,0.1088312864,0.5102598071,0.4114189446,-0.0492335819,-0.0457732938,0.0242724456,0.1189988926,-0.0850605667,0.3223448098,0.033200141,0.3221098781,0.2746065557,0.1198492125,0.0574153103,0.2889924943,0.1625372469,0.0712472424,0.0226034895,0.0666723996,0.1023148671,0.2394506186,-0.2568710446,0.0724869147,-0.3361437619,-0.1103490889,0.2590260506,0.2074869275,0.1140956059,0.0336433798,0.0582498871,-0.1319148839,0.7058607936,0.510782063,0.371773243,-0.1432923228,-0.0604136288,-0.5617278218,0.1689248532,-0.2268105,-0.209326908,0.1078344584,0.1282417923,0.4357310832,0.1924453825,-0.027367983,-0.1725472957,-0.400547415,-0.0237439033,-0.342823118,-0.2178272456,0.0078974469,-0.1119740531,-0.1383003145,-0.0180382617,0.0899634212,-0.1823913902,-0.0019597169,0.0270195864,0.1578267366,-0.0332669131,0.0914950147,0.4042009115,0.015184517,0.4290499091,-0.0270545203,-0.2645837665,-0.4509984851,-0.1991005838,-0.3306635618,0.3517012894,0.2352200449,0.1762435734,-0.1694796979,0.1872226745,-0.168528229,0.4838164151,-0.0565907434,-0.5422750115,0.1255182773,0.073942557,0.0957553387,0.1444119662,-0.0106463991,0.053534545,0.0044792788,0.5422230959,-0.2458273768,-0.1378215253,0.2703121006,-0.2595653236,-0.3543881774,0.2645863295,-0.1342298537,-0.3655931354,0.3093405664,-0.5371618867,0.1654866189,0.3419696987,-0.1092510149,-0.1710449755,-0.0504334494,0.1009506658,-0.2060299069,-0.0858500078,0.5699597597,-0.119342804,-0.2334057987,-0.0469614193,0.0090484619]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/290","title":"ConnectionError - Eli5 dataset download","comments":"It should ne fixed now, thanks for reporting this one :)\r\nIt was an issue on our google storage.\r\n\r\nLet me now if you're still facing this issue.","body":"Hi, I have a problem with downloading Eli5 dataset. When typing `nlp.load_dataset('eli5')`, I get ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/eli5\/LFQA_reddit\/1.0.0\/explain_like_im_five-train_eli5.arrow\r\n\r\nI would appreciate if you could help me with this issue.","comment_length":28,"text":"ConnectionError - Eli5 dataset download\nHi, I have a problem with downloading Eli5 dataset. When typing `nlp.load_dataset('eli5')`, I get ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/eli5\/LFQA_reddit\/1.0.0\/explain_like_im_five-train_eli5.arrow\r\n\r\nI would appreciate if you could help me with this issue.\nIt should ne fixed now, thanks for reporting this one :)\r\nIt was an issue on our google storage.\r\n\r\nLet me now if you're still facing this issue.","embeddings":[-0.2497588396,0.0893835127,-0.0778242871,0.3270986974,0.2081756443,0.0021226979,0.2745363414,0.1381624788,0.049143821,0.0405926295,-0.0316886939,0.0492954031,0.1268917173,0.332809031,0.0971176103,-0.1882539243,-0.0128969112,0.0856067389,-0.0095131677,0.107113421,-0.0457852446,0.1713877767,-0.1018920094,0.1726023257,-0.1952628344,0.1230360121,-0.0542774051,-0.0766947418,-0.3689592183,-0.2835099101,0.343839556,0.1248752102,0.1802075505,-0.0275570005,-0.0001161579,-0.0773379281,0.2845264673,-0.0068009025,-0.3015561402,-0.4885488451,-0.1385288537,-0.199436605,0.3694058955,-0.0722754598,0.2015680373,0.1652351767,0.4329591393,0.1027759016,0.4838895798,0.1664904654,0.1847645193,-0.033776179,0.3706132174,0.0979095697,0.069197841,-0.3523107469,0.0500903428,0.3374673724,-0.1407166868,-0.1857194006,0.1142202765,0.0921674445,0.1664485037,0.1101124659,0.1821706891,0.1309393942,-0.0719027966,-0.3520626426,0.0242044348,0.1167074814,0.997747004,-0.0397392809,-0.2393235117,0.0057089836,0.1492985785,-0.4184676111,0.5427045822,0.2712637782,-0.147865504,0.0503337421,-0.3415909708,-0.3982751071,-0.3517916799,0.3603954017,-0.0355244651,0.0635365844,-0.0816120878,0.1403473765,0.0928828716,-0.0811193436,-0.108157225,0.1471655369,-0.2366290092,0.2940263748,-0.3973180056,-0.0846187025,-0.2899481654,0.1905506849,0.2292871624,0.3845526278,0.0899111181,0.0555599593,-0.1263232529,0.1376539022,0.3958095312,0.2708767951,0.1903690696,0.0780157372,0.187390089,0.174881652,-0.0046946746,-0.0981151238,-0.1432070434,-0.0560345501,-0.1115585491,-0.2207060307,0.080759801,-0.3565367758,-0.3290453553,0.099036172,-0.4623711407,-0.1824753582,-0.0268413592,0.1823046505,-0.2728771865,0.0126221757,0.0528683886,0.3319451511,-0.1053778827,-0.2310124487,-0.1688058674,0.0135987885,-0.2198605984,-0.0283150654,0.1741315722,-0.1911731213,0.4688150883,-0.211623624,-0.1507113278,-0.1241675541,0.1094485894,-0.1262895465,-0.5147399306,0.0752008483,0.2205120325,0.0992490426,-0.0520787165,-0.273128897,0.0105490126,-0.1274969578,-0.2243053019,-0.1969204694,-0.4014467001,0.1562357694,-0.1802914143,-0.2092911154,-0.3340261281,0.1914106905,-0.0650037825,-0.2015036494,-0.1894827485,0.0536635891,-0.1326780617,-0.148902297,0.2816813886,0.2782120109,0.0476502664,-0.4040094614,-0.179675281,-0.1528379172,0.1077236012,0.4209942222,-0.0597602911,-0.1324637383,-0.1739879698,0.3515346646,0.6295803189,-0.1592783481,-0.9718024731,0.2665201426,-0.2599483132,-0.2456943542,-0.0510173216,0.1613211036,0.1380274147,0.1287878454,0.1731065363,0.4931621552,-0.1016826704,-0.0801599994,-0.1647747904,-0.3046265841,0.1237145513,0.0984527096,0.0070138671,0.08282125,0.2047087699,0.0117723318,0.121151872,0.2219158411,0.1049161106,0.1853275746,0.2719538212,-0.0845817104,-0.1859809011,0.0755384043,-0.5517844558,0.1320389956,-0.2430137545,0.2367604226,-0.3266122043,0.1815017909,-0.1378211081,-0.0250577684,-0.0871811658,-0.0196262524,0.0588310659,-0.2522036135,0.0864872783,0.36637941,-0.0120975785,0.4874075949,0.0745212212,0.2470305115,-0.3514170945,0.5134061575,-0.2010786831,-0.0856424868,0.1534211189,-0.1430327147,0.0028649687,0.0159374196,-0.2361585498,0.0854357108,-0.5183472037,0.1810060591,0.2617428303,0.4170763493,0.1843682975,-0.4226613939,0.2094871253,-0.2681599557,0.0940827876,0.0769026577,0.0457851514,0.0614014305,0.2973328531,0.0859338418,-0.0368909761,0.3502425849,0.3250182867,-0.032225877,0.0745291933,-0.1063506529,0.4227878451,0.2946760654,0.2827172875,-0.2102748156,-0.2436743975,0.2049201876,0.364874959,0.1089647487,0.0341823362,0.3822146356,-0.2421706766,-0.4425079525,-0.1141517386,-0.0854029432,0.3881253898,0.0548147075,-0.0007224341,0.1530988514,0.0520122349,-0.1004080027,0.1607725173,0.0738225654,0.0998866633,0.0479518995,0.1733155251,0.0522657521,-0.1489476562,-0.1117216125,0.0068897391,0.3233179748,-0.1116506308,0.1294988543,-0.0682972223,-0.2203239202,0.1653384119,-0.2381619662,-0.4337794483,-0.3147348762,0.0085767368,0.2693352401,0.1083348691,0.0735155046,-0.0276876055,0.0659514591,0.200631246,0.0095318379,-0.0412142724,-0.2337365597,-0.3130791783,0.071206592,0.3317569494,0.1017727405,0.1627256423,-0.175133571,-0.067290619,-0.4746840894,-0.2086940855,0.0796019286,-0.124481529,0.4008778334,0.0662376881,0.5214606524,-0.1920004785,-0.1050177738,0.1928012669,-0.1713085175,-0.1980884671,-0.028674027,-0.0131229199,0.1640209407,0.1673440933,-0.2523571551,-0.4018141925,-0.2180769891,0.4885856211,0.147012338,0.0510380492,-0.0898293257,0.1064495519,0.0771180466,0.1504940987,-0.1009458154,-0.2762487829,-0.4101827145,0.1920643151,-0.0836889148,-0.318220675,0.1295692176,0.1608963013,0.2064031959,0.0575537719,-0.5428879857,-0.1649180055,-0.0307975076,0.2012941539,0.0202498566,-0.1885973066,0.2706339955,-0.1079489961,0.0085336426,0.2917350829,-0.082199052,0.287941128,0.2989262044,0.4235113859,0.099251464,0.6737154722,-0.0067914613,0.4394054115,0.1497519314,-0.1561596543,0.5129275322,0.217713356,0.0470972024,-0.0618590079,-0.0526073612,0.048192367,-0.0784617364,0.0367627889,0.0775884688,-0.0073583196,-0.2609281242,-0.3035992384,-0.2135281563,-0.2406092137,-0.1818315685,-0.0094042225,-0.3316059411,-0.0316459462,0.2277188003,-0.3477561772,-0.0815766156,-0.3281505704,-0.1538904905,0.1137425303,0.1255898327,0.098878175,-0.401640445,-0.164854154,-0.7408914566,0.4798649549,-0.0366738625,0.4076276124,-0.3573444188,0.1984847635,0.1532742083,-0.0475021712,0.3783436418,-0.1877074093,-0.1075773686,-0.1808519363,0.4171361625,-0.0990632027,-0.1739494056,0.032793235,0.3198185563,0.2112649381,0.0048578014,-0.4678073525,-0.0418552049,0.2367250174,-0.0191198606,-0.1713786125,-0.1697035283,-0.1911770105,-0.1812907904,-0.225057438,-0.1508511156,-0.0715977475,0.2053385079,0.0614306666,0.07760977,0.0926986337,-0.0552904345,-0.1057058945,-0.2637782395,0.1229469478,-0.0824101865,0.2335932404,0.3358117938,0.0829344243,-0.066221498,0.7518094778,0.0722035617,-0.3747424185,0.2129178494,0.0726347864,0.0504959598,0.0436930433,0.1096118838,0.2167548686,0.2070121765,0.3138194084,0.1138103753,0.0818902701,0.3213292956,0.1495512575,-0.2284025699,-0.3196690977,0.5979219675,-0.0772457495,0.026629338,0.3614577651,0.025448218,-0.0527034216,0.1322512329,-0.1671893448,0.9752023816,-0.0637270957,0.0888512284,0.290612489,-0.036392428,0.5495631099,-0.3213944137,0.1422257423,-0.4006831944,-0.3116593659,-0.1080612317,-0.1506585628,-0.1514257342,0.0388589278,-0.070682846,0.3890458345,-0.1569599211,-0.0307698194,0.1514774114,0.3611959517,-0.3257433474,0.001074414,-0.4061712921,0.1243985891,-0.1115576997,0.4765816629,-0.1790934503,-0.1770949513,-0.152822867,-0.3629816175,-0.4525060654,0.2904271483,-0.1186133698,0.281265229,0.1837593168,-0.2143530101,0.0821661055,0.3698530197,0.2788470387,-0.113462761,-0.3012597561,0.213904798,-0.3641892076,-0.2792799473,0.1909885854,0.1420849264,0.1644794643,-0.1711326689,-0.3458022773,0.33033517,0.0300877374,-0.3150250912,0.038394995,-0.063007161,-0.1650114954,-0.0111640254,-0.0491429977,-0.1779934168,-0.2788163722,0.0189603455,0.085422039,0.2785882652,0.1299882233,-0.0747781098,0.318372041,-0.1061498448,-0.1094591096,0.6303485632,-0.0367152579,-0.2569094002,0.6961232424,0.0715997741,-0.0523191206,-0.1027355418,-0.1806363463,-0.0620620847,-0.4142879546,-0.1203576177,-0.1896400303,0.2682061493,-0.240369767,0.0138671622,0.0832831189,0.0245375931,0.2988289893,-0.5705106258,-0.2339953929,0.3107908666,-0.6751425266,0.0912516788,-0.0619551875,-0.0405350775,0.2073073685,0.0465177633,-0.2335014045,0.0168364067,-0.5630574226,-0.1096439883,0.4691391587,-0.3179607093,0.3390700519,0.0330767594,0.1282810867,0.0709152967,-0.4158704579,-0.1232693195,0.1404625028,0.1342479289,0.0744658187,-0.1471761763,-0.0891750753,-0.0494111404,-0.1203631982,0.2037829608,-0.0831429958,-0.0802062824,0.0088241166,-0.04099584,0.0388887264,0.0532427914,-0.1792834997,0.1862981915,0.057067126,0.4040949643,-0.0663293004,0.1107206866,-0.0117272921,0.0073303138,-0.5277674198,-0.0233648382,-0.1421176642,-0.1028678343,0.2993016541,-0.3585296273,-0.031915199,-0.093376413,-0.0398168266,0.3959412277,-0.3606976867,0.0450688638,0.4536335766,0.2047598511,-0.115044266,-0.0489416979,0.1618078351,-0.0657101125,-0.1915210336,-0.1048973799,0.3560284972,-0.142533958,0.0074949795,0.1164888367,0.5072419047,-0.22874704,0.0539118983,-0.1903980523,-0.1098055467,0.0040934454,0.1579043716,0.2635743618,0.1804443449,0.0299622435,-0.2974122763,-0.3058331907,-0.3694431782,0.358181268,-0.0327682421,-0.3869180083,-0.4250785708,0.379189074,-0.1424476057,0.0945188776,0.0651360229,0.2678670883,0.0183982626,0.0278205033,-0.4035420418,0.2810949981,0.0372345522,-0.0772875994,0.0590378679,-0.2060252875,-0.2482370734,0.1561208218,-0.0507645123,-0.2287588716,-0.0081866384,0.2536241114,-0.1223288551,-0.3326306343,0.0595651828,0.4721235931,0.0557702854,-0.2593837082,0.2974856496,-0.054521773,-0.0840746164,0.0643685162,0.2882589996,0.4291482866,0.2077864259,-0.1113613024,0.1713638753,0.1456240863,-0.0680618063,0.0589853413,0.0404421873,0.4193636775,-0.3132342398,0.1662859768,0.1087354943,-0.1322739124,0.3147642314,-0.0274648331,0.7163369656,-0.0503231175,-0.0370516032,-0.1435165256,-0.0121168531,-0.2358630449,-0.1394416094,-0.627419889,0.1356130391,-0.0447005965,0.0773403049,0.1117776558,-0.2120007873,0.034464594,-0.2287580669,0.5761770606,0.5365760326,0.4533986747,-0.2004956156,-0.2497022599,-0.4111526012,0.1546654552,-0.0705276206,0.1328837425,0.0882303864,0.3086768985,-0.1362995505,0.1898849905,0.1029489264,0.4398308098,-0.1915183216,0.0976926163,-0.4273041785,-0.1192632541,0.2379492819,0.314021349,-0.1223980188,-0.0331743062,0.1472319812,-0.2594808042,0.0268345848,-0.2557217777,0.1261706054,-0.1776338369,0.2399816662,0.4202907979,0.2550695837,0.4653469622,0.0930489525,-0.2224843949,-0.2390590906,-0.3754879832,-0.1918814331,-0.0437452532,0.1129088774,0.3058715165,-0.1066821888,0.3847686648,-0.2915793955,0.1641982496,-0.1191589534,0.154640466,-0.1672753394,-0.1117936075,0.1817937195,0.1824832112,0.0058611701,0.058616966,0.0868314579,-0.06921646,-0.342007339,-0.0011577326,0.317796886,-0.065762721,0.0364116244,0.109814167,0.2601310611,0.2480129451,-0.0584874675,-0.5286949873,0.2906092703,0.4041068852,-0.166374445,-0.3493910432,0.0796217844,0.3144536018,0.1672551334,-0.0884463564,0.1378218532,-0.0906176418,-0.167847231,0.0019331727,-0.1552881896]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/290","title":"ConnectionError - Eli5 dataset download","comments":"It works now, thanks for prompt help!","body":"Hi, I have a problem with downloading Eli5 dataset. When typing `nlp.load_dataset('eli5')`, I get ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/eli5\/LFQA_reddit\/1.0.0\/explain_like_im_five-train_eli5.arrow\r\n\r\nI would appreciate if you could help me with this issue.","comment_length":7,"text":"ConnectionError - Eli5 dataset download\nHi, I have a problem with downloading Eli5 dataset. When typing `nlp.load_dataset('eli5')`, I get ConnectionError: Couldn't reach https:\/\/storage.googleapis.com\/huggingface-nlp\/cache\/datasets\/eli5\/LFQA_reddit\/1.0.0\/explain_like_im_five-train_eli5.arrow\r\n\r\nI would appreciate if you could help me with this issue.\nIt works now, thanks for prompt help!","embeddings":[-0.3160368502,0.0045427657,-0.1026679501,0.3319367766,0.1839226633,-0.0415495113,0.1790215969,0.131200701,0.007808601,0.0381670408,-0.0683693811,0.0912042186,0.1721084714,0.3547389209,0.1306024194,-0.2975560725,-0.033266671,0.0736123919,-0.1012925804,0.0856248513,0.0020400353,0.1946745813,-0.0948224813,0.2516215146,-0.1924663484,0.044993028,-0.0502812043,-0.0528582409,-0.3837417364,-0.2750585377,0.3265641332,0.1005097479,0.2225845307,-0.0137834772,-0.0001139173,-0.0807033107,0.3218512535,0.0097366972,-0.2407580614,-0.4909570813,-0.1903416365,-0.193009153,0.3726085722,-0.0932032913,0.1401702762,0.1824145466,0.4301486313,0.1288565248,0.5509679914,0.1866014153,0.2074299455,0.0629197583,0.3544031382,0.0760362819,-0.0195346456,-0.4353753924,0.0775880665,0.2196241915,-0.1523749232,-0.145176962,0.146446988,0.0353464782,0.1881281883,0.0862256587,0.2287599593,0.1018260941,-0.0655143559,-0.3710187078,-0.0280808471,0.1380922645,0.9783646464,-0.0200114008,-0.2317021489,0.0363727957,0.1289185584,-0.4365830123,0.482468009,0.2028784156,-0.2004730701,0.0544587076,-0.2527811527,-0.4132383764,-0.3581619263,0.3985892534,-0.0097074946,0.0679710135,-0.0744853988,0.1385287642,0.1021021456,-0.0746349916,-0.1109414473,0.1464940012,-0.1988526434,0.2861682177,-0.3778584599,-0.0950319543,-0.3001132607,0.2769494057,0.2221837193,0.4199337363,0.0532788858,0.0444657542,-0.1783898175,0.1163920835,0.429517597,0.207937628,0.2395010144,0.0266846661,0.1733655781,0.1585086882,-0.048569832,-0.1086407676,-0.1763888001,-0.0489146858,-0.0833867118,-0.1472421885,0.0760516003,-0.3683397174,-0.2615665495,0.0734527782,-0.4209763706,-0.20089297,0.0098391371,0.2081648111,-0.3167134821,0.0498013869,0.0350839719,0.3735655546,-0.1430878788,-0.3338142931,-0.1828426123,0.0196580533,-0.2310293615,-0.0865519792,0.1113857254,-0.1830300838,0.5268388987,-0.2229688466,-0.1035659015,-0.1028980389,0.1236847937,-0.0759308264,-0.5204966664,0.0408669598,0.2015656829,0.1232855096,-0.0654342175,-0.1801025271,0.0317311399,-0.1207792982,-0.1685443074,-0.1894882917,-0.3336519897,0.1629578173,-0.169285655,-0.2539440095,-0.3197716177,0.1558083594,-0.032701131,-0.1694931984,-0.1344596148,-0.017633358,-0.0746494085,-0.1405123323,0.247550711,0.2616276443,0.1152003035,-0.3918740451,-0.125500083,-0.1986340135,0.0540743619,0.4012035728,-0.1061619073,-0.1177819371,-0.093926385,0.3386781514,0.7145435214,-0.2443551272,-0.9375604987,0.2798621655,-0.2377164215,-0.2310992479,-0.0483525358,0.2109748274,0.0885634199,0.1124960557,0.1400562972,0.5366325974,-0.0998463035,-0.1371956468,-0.1546759903,-0.2763475776,0.1022544876,0.0657645687,0.0594135337,0.0935057625,0.1791470647,-0.014633432,0.159552142,0.2364636064,0.0806134939,0.1565670073,0.2013326436,-0.061794918,-0.1816622615,-0.0031095024,-0.4122925997,0.0719535351,-0.2384841591,0.2344830334,-0.2922262549,0.1454318315,-0.156547159,-0.0449511558,-0.096853748,-0.0015951368,0.0591616854,-0.2565131783,0.1380239278,0.3532640338,-0.0661934614,0.5161086917,0.0693629161,0.2881492674,-0.3974658549,0.3992271721,-0.2192782909,-0.0273861308,0.1187607273,-0.1292223781,0.001576082,-0.013898911,-0.1872392297,0.1293759793,-0.5173515677,0.0370742157,0.2605731189,0.3767189384,0.177911967,-0.4607554078,0.2164403945,-0.2104597539,0.0874192193,0.133039251,0.1221644655,0.073869288,0.2634145021,0.1085143834,0.0194467641,0.335746944,0.3537298739,-0.0318488032,-0.0065332535,-0.0058061415,0.4179523289,0.2531592846,0.2236450762,-0.2858195305,-0.2907445729,0.1915371567,0.3070197105,0.0814474076,-0.0186484363,0.3171163201,-0.2093933523,-0.336625427,-0.1152605489,-0.0247751754,0.3463065028,0.0942294374,0.0424830131,0.1912262142,-0.0128308749,-0.0515230186,0.137325868,0.1144298315,0.1184134632,-0.0116039887,0.0665901229,0.0841599181,-0.1127033457,-0.135372147,-0.0210222788,0.3140085638,-0.1468463838,0.1426988393,-0.0797545612,-0.3158383071,0.1680417955,-0.2185231298,-0.453882575,-0.2795265317,-0.0296052694,0.216522485,0.1327160448,0.1090508327,-0.112633355,0.0502641909,0.1885900944,0.0328844897,-0.0455818325,-0.255454123,-0.2951177061,0.0928670764,0.311414808,0.1454956681,0.1704753935,-0.1821504384,-0.0232925043,-0.360301882,-0.1881218255,0.0703011155,-0.0796602145,0.4291812778,0.0839627013,0.4878391922,-0.2272328734,-0.1091045439,0.2288642973,-0.1903076023,-0.217649743,-0.0921856612,-0.0290001668,0.1314664185,0.2117783576,-0.2726451755,-0.3695249856,-0.235184148,0.4805596769,0.2086895853,0.0271114521,0.023069825,0.0757762119,0.0548871197,0.1292813271,-0.0135728335,-0.2675594389,-0.4135936201,0.1615177393,-0.1268728971,-0.3471901417,0.0968066901,0.1263737381,0.2421748191,0.0806886703,-0.4920715988,-0.1922685653,0.0032060531,0.2498799264,0.0430412814,-0.0819776133,0.226433143,-0.1707025766,-0.0097296052,0.2855619192,-0.0271705128,0.3289905787,0.2387275398,0.511643827,0.0752004012,0.6946192384,-0.0126228686,0.4729540646,0.1548519582,-0.1521751285,0.4943776429,0.197595045,0.0842869878,-0.0742511004,-0.073938109,-0.0294875409,-0.067583546,0.1234454513,0.0484690405,-0.0038519951,-0.1856038123,-0.3634330332,-0.2853129208,-0.2817204595,-0.1685829908,-0.0180381313,-0.3750763237,-0.0743290856,0.2504639626,-0.2698898017,-0.1221158877,-0.2198711187,-0.079963088,0.1809663028,0.1034230664,0.0833252072,-0.4670000374,-0.1130940467,-0.7431433797,0.4614853859,-0.0700187981,0.3784041703,-0.337802738,0.1550221145,0.1286080182,-0.0673319772,0.3952168822,-0.229441449,-0.1443521827,-0.169029817,0.3753963709,-0.1166919842,-0.0979988426,0.0127120018,0.3622352779,0.2275296152,-0.1162611246,-0.5046649575,-0.0436666645,0.3047867119,-0.0446848571,-0.177913636,-0.2161074281,-0.2119750082,-0.16515477,-0.2116857618,-0.1308485568,-0.0346586779,0.2334235311,0.1338651478,0.1149108484,0.0884743705,-0.0686227605,-0.0470713638,-0.2337338477,0.063989006,-0.0658720359,0.2143191695,0.2929542065,0.0752285346,-0.1964729428,0.7046284676,0.0077454871,-0.4625891149,0.0806436092,0.0918087289,0.1537889391,-0.0240046922,0.1056209728,0.2098708004,0.2925037742,0.2464945167,0.2034423053,0.120404318,0.3335337937,0.1850096434,-0.2429071218,-0.397901237,0.5674382448,-0.1087838784,0.0492952317,0.3720018864,0.022827426,0.0302229691,0.1204979792,-0.2101639956,0.9542780519,-0.0402382649,0.0609088466,0.293402642,-0.044665806,0.6406154037,-0.2861580253,0.1353301853,-0.3864431679,-0.2551797926,-0.083816804,-0.1253445596,-0.1368719786,0.0222669132,-0.0037661358,0.4349326789,-0.154635787,-0.1221467331,0.1525245756,0.3490351737,-0.2496964186,0.0229969509,-0.4476905465,0.1271594614,-0.0980421007,0.4811492562,-0.2184044719,-0.2082701176,-0.1576489508,-0.3766268194,-0.3848199844,0.2594406307,-0.1663315892,0.2301400602,0.2001096457,-0.2506795526,0.0472492836,0.3786344826,0.325529784,-0.1675427854,-0.258977592,0.2281455696,-0.4195118248,-0.3168659508,0.2344556898,0.1285440922,0.2107559144,-0.2144292146,-0.3443413973,0.3838177919,0.0787208006,-0.317181915,0.1518627256,-0.0634934083,-0.1334641278,-0.0345478877,-0.0938001499,-0.0743825734,-0.2164732218,0.0441893563,0.0918599889,0.3079007566,0.0479347035,-0.0961261168,0.3376711607,-0.0877837613,-0.1124651507,0.581284523,-0.024288537,-0.3195082545,0.6194197536,0.0824854821,-0.0649360567,-0.0999992713,-0.1936226785,-0.0064651645,-0.4632235467,-0.1601490974,-0.1494054943,0.2323148847,-0.2309333533,0.006432184,0.0157065988,-0.0167086553,0.2950148284,-0.5576795936,-0.299726069,0.290625155,-0.636588037,0.0268891081,-0.0509755611,-0.0057226634,0.1910416335,0.075412713,-0.2469491363,-0.0170038026,-0.4883578718,-0.0570491478,0.5541730523,-0.3202654123,0.3292177916,0.0623539016,0.1383268833,0.0295031667,-0.4250778556,-0.1361382753,0.1031910554,0.1384622157,0.0335072093,-0.1780405045,-0.0361627303,-0.0238904003,-0.1282079667,0.2113081068,-0.1136227548,-0.1290283352,0.0082335584,0.0235489253,0.052362565,0.0499235466,-0.227284953,0.2069149017,0.090122804,0.4216292799,-0.0296231993,0.1294012219,-0.0317956619,-0.0160980467,-0.5163780451,-0.0609489791,-0.1062006205,-0.1474477351,0.3244429529,-0.3436755836,-0.0169408824,-0.0710314959,0.0155345919,0.4355965555,-0.2565228641,0.1441595852,0.4565818608,0.2100705504,-0.1330444068,-0.0118310805,0.2298415601,-0.0797452703,-0.1902022064,-0.0872210786,0.2715658545,-0.1359399855,0.0930446684,0.0812214315,0.5006201267,-0.2839563787,0.0720356777,-0.1499709487,-0.0820337012,-0.038264025,0.1716970652,0.3664487004,0.1821278185,0.1205260605,-0.2891526222,-0.293291837,-0.3452041149,0.3345153928,-0.0677632093,-0.3844204843,-0.4393613935,0.3376511037,-0.0764751583,0.0737514198,0.0395449698,0.242352739,0.0678258762,0.0479226783,-0.4397285581,0.2790709436,0.0414649211,-0.1126435921,0.0344572365,-0.2079639882,-0.1932045072,0.0931373462,-0.047252994,-0.2270544767,0.04408006,0.2209251821,-0.0517194048,-0.3738093972,-0.0461221375,0.4147742391,0.0443672575,-0.2373033166,0.3063954413,-0.0310536269,-0.0511533357,0.0768313184,0.3156601787,0.4749594033,0.1733636111,-0.1726780832,0.1965103149,0.0941312611,-0.0678417683,0.0243662391,0.0213001724,0.4319781363,-0.1796840727,0.18923226,0.1107271388,-0.1351695061,0.3186559975,-0.0422804579,0.6442863345,-0.119432658,-0.0642274469,-0.1275049895,-0.0159435701,-0.3089057207,-0.1232550144,-0.6747678518,0.1923536062,-0.073125951,0.0552760698,0.0775244087,-0.2649259269,0.0385074355,-0.2782240212,0.5649307966,0.5436359644,0.58193326,-0.206044063,-0.1744198501,-0.4281288087,0.1201868877,-0.0674615353,0.0793178976,0.0577690527,0.2816166878,-0.0723190233,0.247568056,0.0973985046,0.4298544824,-0.1519849747,0.1054624766,-0.3610313237,-0.1206574962,0.1938289553,0.3366586566,-0.0669528097,0.0138713187,0.0910863057,-0.2385233045,0.0148785897,-0.1940485984,0.0515712164,-0.1389554441,0.1693493873,0.4025813639,0.2613092363,0.5460969806,0.0831302106,-0.1994835883,-0.2263431549,-0.3681236207,-0.1976021975,0.0070038815,0.1340086013,0.2715317905,-0.1087941229,0.3736822605,-0.289509505,0.1479497552,-0.1139303073,0.1242488176,-0.1591225564,-0.1138858646,0.1919621527,0.1387163997,0.0083056157,0.062266469,0.0084918002,-0.0922491327,-0.3244762123,0.0769593716,0.3551872075,-0.0705968812,0.0438473634,0.0342362449,0.2609830201,0.267786622,-0.0585033521,-0.503410995,0.3216814697,0.3932133317,-0.1854239702,-0.3536376059,0.1007943451,0.3017788529,0.1856780946,-0.0778148174,0.1197636053,-0.1719077379,-0.1776840836,0.0284031238,-0.1550272405]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/288","title":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'","comments":"It looks like the bug comes from `dill`. Which version of `dill` are you using ?","body":"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'","comment_length":16,"text":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'\nIt looks like the bug comes from `dill`. Which version of `dill` are you using ?","embeddings":[0.0769836977,-0.4020940661,-0.1799209863,0.0659818947,0.1492687613,-0.003346167,0.55732584,0.3920768797,-0.0114857536,0.2672935724,-0.0524433628,0.3347723782,-0.0505946018,-0.3120033741,0.0111635383,-0.3291765749,0.0835466534,0.2736496031,-0.37093997,-0.2723834813,-0.2335172594,0.2800564766,-0.0776633248,0.2484975308,-0.1481041014,-0.0349433534,0.1214392036,0.1059611663,0.0675767437,-0.2886580229,0.1965774745,-0.1745566279,0.2275901288,0.4768029451,-0.0001043869,0.1361589134,0.3375531137,-0.0346087962,-0.1342187971,-0.3505931497,0.4713913202,0.0593214221,0.3363775015,-0.3500793874,-0.0507910028,-0.2969646454,-0.0969579443,-0.30893448,0.4441095889,0.4418491125,0.3010698855,0.1026526392,0.2446943521,-0.169915244,-0.0208351463,-0.0297836401,-0.0677934811,0.3753479719,0.1383522898,0.1474164873,0.3823755085,0.342466712,0.160931766,0.0776078403,0.3393225968,0.0322579108,0.6600615978,-0.4541978836,-0.0775012821,0.2736726701,0.1812872142,-0.1184436977,-0.3209805787,0.1269780993,0.0060105366,-0.2475019395,0.0616756529,-0.06774728,-0.2496780753,-0.032574065,0.4458368123,-0.0442663617,-0.1771987081,0.1953953058,-0.0185109787,0.1636362374,-0.044930961,-0.0506545231,0.0271611586,-0.3340252638,0.3872074783,0.0928654596,0.4484997094,0.0960105956,-0.0759613588,-0.1219853684,0.0625471696,-0.6255782843,-0.2034097016,-0.2914340496,0.0739517361,-0.0122585492,0.1180225462,0.2710880339,-0.1634567976,0.0667565018,0.1067271829,0.160689339,0.335481286,0.0793941543,0.328011781,0.1698743701,-0.3021524251,-0.125138551,-0.1567854881,0.0640961453,0.3410718143,0.0939639807,-0.5376574993,-0.0486492701,-0.1924375296,0.0474512838,-0.0055814907,0.2568105459,-0.0379515067,0.1916810572,0.4150514603,0.11126706,-0.2454229593,0.0046106484,-0.2229356468,0.3555662632,-0.0703071654,-0.2020584792,0.0587319173,0.1554544717,0.3014602363,-0.0993096381,0.1310544163,-0.0825211927,0.1695728451,0.0146686807,-0.0396083631,0.1542768478,-0.2942762375,-0.0451418795,0.2622388005,-0.1855244786,-0.2611550987,0.091217421,-0.2335643321,-0.0095623313,-0.4541479945,0.3482017517,0.250795722,0.110424757,0.2338398695,0.0304904748,0.1052182466,0.0470620319,-0.101509586,-0.2446223497,-0.1023621485,-0.3306634724,0.2848880291,0.1544385999,-0.0588522553,-0.1574843526,-0.1423849314,-0.0397362337,0.1201601923,0.2202340066,-0.1351770014,-0.0135140996,0.1174672022,-0.1719239056,0.6146841645,-0.5745144486,0.1056552455,-0.0611012615,0.1371087134,-0.0482239611,-0.0108275972,-0.0786206573,0.1418243051,-0.0617279485,0.3198719323,0.3495156169,0.2515051663,-0.1104580089,-0.3646187782,0.012272289,0.1300430149,0.0910370499,0.0965716168,0.0635002851,-0.0523701161,-0.0480599813,0.0647452176,-0.1598062068,0.004075455,0.123443149,0.0875847489,-0.0867793262,0.1008282751,-0.1719257832,-0.3230362236,-0.0569832884,-0.3147102892,0.0343750939,-0.0152414292,-0.1913605928,-0.1983193457,0.0127634313,0.119129926,-0.2337505966,0.3220001459,0.0348381661,-0.0130126383,0.1630757153,-0.2321627587,-0.0329379737,-0.1512498111,0.181532681,-0.1353039891,0.1641567945,-0.4232264459,-0.2587807477,-0.1687233597,0.35354954,0.0873158574,-0.3229031265,-0.0398910791,0.3210483193,-0.0660059303,-0.0963132754,-0.3560896814,-0.0151594486,0.3564194143,-0.1473701298,0.044008974,0.3544332087,0.1194430888,-0.106795162,0.066043511,0.2442780733,0.2975997329,-0.0210516416,0.2813411057,0.1925182492,0.0381851532,0.1245420799,-0.0583589785,-0.1804328114,0.1559882909,0.3096714914,0.1182772219,-0.1786590964,-0.3969525397,-0.0277689844,0.2105372548,0.0029683402,0.2126078606,-0.0307502821,-0.0751878694,0.1257900149,0.1558475643,-0.0766748637,0.1838900596,0.0091597186,0.1322796494,0.0727801621,-0.2034448534,-0.1412041932,0.1152334958,-0.0268808324,0.3497841656,-0.0709351376,-0.0873165056,0.0785132051,-0.0844333768,-0.3782753944,0.05666437,0.1657810658,-0.0829095691,0.0378239527,-0.26275298,-0.0227957889,-0.0947818086,-0.2296084017,0.1324962825,-0.1437409669,-0.0149990348,0.1761470586,-0.1750652939,0.2052592635,0.1883973628,-0.0567656383,0.0960681811,0.3126211762,-0.2592220008,-0.0779284015,-0.1663522869,0.1553854793,0.0330693349,-0.0856425464,0.1994274557,-0.2755847275,-0.120114252,0.1499013901,-0.4520840049,0.0524351485,0.0140112396,0.1148459166,0.2227616012,0.2215283513,0.0248776749,-0.3447566926,0.2151990533,-0.3349262178,-0.0258942172,-0.0942451507,0.053977523,-0.0956519395,-0.4885933995,-0.6345106363,-0.3095667958,-0.538667798,0.0189340804,0.1400108635,0.3738848865,0.3208152652,0.2370781004,0.4575053453,0.3019067645,0.445266813,0.0590105243,0.0270972084,0.3070289791,-0.4646860957,-0.3787640929,-0.1151326671,-0.1516299397,0.3044160306,-0.1614550352,-0.0524057709,-0.3177403808,-0.2742889822,0.2044684142,-0.2654448748,-0.0834642574,0.0780671909,0.1683275998,-0.2234417647,-0.1472693682,0.0398055241,0.1982986629,-0.0899709836,0.0503881052,0.1322678328,0.4105213583,-0.1355243474,0.641283989,0.1535908431,-0.5416244864,0.1315539628,0.0941807106,0.1784284562,0.1610788852,-0.0770013109,0.0641207471,0.1061399058,0.3015512526,-0.0606331378,-0.1528750956,-0.2624876797,-0.2485653609,0.0551917031,-0.2093648165,-0.0992520601,-0.1903233081,-0.222233355,0.0266475081,-0.1704294384,0.1294929087,-0.0302802324,0.2669641674,0.1404934525,0.3743702769,-0.029914448,0.0309316143,-0.1446655095,-0.1721434444,-0.168138504,0.2689387798,0.0391585045,0.350931108,0.0693768784,-0.0899103284,0.0954246521,0.1021384597,0.3387723565,-0.141647324,-0.2548166513,0.3434726298,-0.0369338654,-0.426663667,-0.1019397676,-0.298001945,-0.0479726493,0.4439178407,0.0099498928,-0.2568550408,-0.271645844,0.2455054373,-0.1302042454,-0.0676388666,0.0721810982,-0.1339272708,-0.4609440267,-0.337305218,-0.1392961591,0.195825085,0.1834748089,-0.1435066611,-0.1498092711,0.001615681,0.0100157037,0.0503716096,0.1900882125,0.2087095976,-0.1278277785,-0.1401393712,0.0382278301,0.2863094807,0.3246237338,0.0855566636,-0.3335414827,-0.2565034926,0.1705903262,-0.0674866214,0.145268783,0.0241156202,-0.1518263668,-0.0574669726,-0.4039544761,0.0352798887,0.3928391635,0.2366948873,0.4728352726,0.233066231,0.0414866954,-0.31564188,0.278352052,0.1127391607,0.0168916248,0.3914880157,-0.1173036098,-0.1339904368,0.1071419939,0.2551783919,0.7796511054,0.1979432702,0.0798775107,0.3982567787,-0.1491081417,0.2713763118,0.0214910451,0.1287784278,-0.2085438073,-0.1011961401,0.0327130482,-0.0950420499,0.2960813046,-0.1658862233,-0.476200968,-0.0088990666,-0.3046584427,-0.0855448246,-0.0103940498,0.25138551,0.0512061715,0.0007532024,0.083183229,0.1936745793,0.1429111958,0.1862857044,-0.2173997909,-0.1922255903,-0.1607504934,-0.3412706852,-0.0518728457,0.153183043,-0.2182635367,0.4361849129,0.1955285221,-0.3141541779,0.0928765908,0.3122994304,0.4542844892,-0.1912746727,-0.1373024136,-0.0474196039,-0.1302075237,-0.0439016484,-0.1522315741,0.1801161766,0.5092203021,-0.2898780406,-0.1360089183,0.2658412457,0.0044673975,0.2290174514,-0.4029322565,0.0144238602,-0.2015995383,-0.0062909569,0.207975924,-0.0527221859,-0.5108942986,-0.1215923205,0.2427589297,-0.0168557093,-0.3328695297,-0.0293962657,0.1110495627,-0.263274312,0.0632196739,0.0704504997,0.3525013626,0.1392199397,0.3493540287,-0.0360304974,-0.2955444157,-0.2647211254,-0.2726528943,0.0311377272,0.0112488735,0.3244880438,-0.2756336033,-0.3246463537,-0.0597305261,0.211653322,-0.0009634426,0.2978005111,-0.2246960104,-0.1632337719,-0.1972972453,0.1918155849,-0.1714804322,0.3040003479,-0.3490905762,0.4225679636,0.1137231663,-0.0430724025,-0.4419589937,0.1668350548,-0.0368808508,0.020018857,-0.1581536829,-0.2147509009,0.1676616073,0.2444770932,0.2284802049,0.0694871545,-0.1888405383,-0.2605834603,-0.170472011,0.0698552579,-0.2116407305,-0.1817622185,0.2404185832,0.0999821424,-0.1844204962,0.0863327235,-0.0292884465,-0.1320893615,-0.0019985072,0.5582348108,0.0387344882,0.5644739866,0.3869163394,0.0089126034,-0.1002072021,0.1331441849,-0.1345453113,0.021350868,0.0597498007,-0.2254040837,-0.3498137891,-0.1373037398,-0.1616208702,-0.1402896494,0.2207787186,-0.16355443,0.0553813726,-0.2637145817,0.1210697517,0.303383112,-0.3440740407,0.1544483453,0.0866090953,0.3013632298,-0.2628340721,0.2679738402,-0.0742736459,0.0446530357,0.2171688825,0.2485523224,-0.1163876876,-0.0176999439,0.0716812983,0.0811027512,-0.0912493244,-0.2889398634,0.3252561092,0.2326954752,-0.1652373821,0.0873996466,0.2099814266,-0.0217311438,0.1657243967,0.1753938794,-0.1132669076,0.3435326517,-0.0894798413,-0.1929242909,0.3750655651,0.1159625128,0.2090840489,0.2096315026,0.1593326479,0.2247309387,0.0840807483,0.165045768,-0.1709304899,-0.0828605443,-0.0514771938,0.134366408,-0.3270949125,-0.1075968072,0.0350296423,-0.1609695405,-0.1508623064,-0.4307509065,0.0921481475,-0.0709812194,0.1977212429,0.0450752638,-0.1878480315,-0.3252618015,0.1265001595,-0.0690778419,0.3112432957,-0.0885865241,0.1946066916,0.3189254403,-0.1001254544,-0.3750415742,0.3328012526,0.3252197206,0.5558351874,-0.2463056743,0.0623565912,-0.0271787271,-0.2585277259,0.1358801425,0.0069714678,0.1525976956,0.0096471645,0.2664268017,0.1872951984,-0.2842236459,0.093714267,0.0491212532,-0.0283321068,-0.2225133032,0.605899632,-0.3091025054,0.0241864193,-0.1419742405,-0.0250129309,-0.5140055418,-0.189438194,0.139050588,0.0070464471,0.0270313472,-0.2933402359,0.1228031516,-0.1128694788,0.3584936261,0.0038748723,0.3682335615,0.029320417,-0.0045026881,-0.3142127395,0.303709954,-0.1117279828,-0.0305645373,0.1000814959,0.0731207728,-0.2099718451,0.0324984714,0.2157515734,-0.1540497392,0.0103491219,0.1331142485,-0.4652892649,-0.0519434549,-0.3202302158,-0.1183811426,0.3270791173,-0.2909780145,0.0002811941,-0.2657406926,0.1885832846,-0.3409994543,0.0875840336,0.0808703229,0.4352749586,0.1772854477,0.1798701733,0.305210948,-0.043241553,-0.3241828382,0.1009184644,-0.1834964305,-0.2110900134,0.2475279272,0.2351478636,-0.0651715025,-0.1614931524,-0.4883197546,-0.2035161555,0.53690207,-0.0041251061,-0.2707729936,-0.2305567861,0.2251336724,0.0629380122,-0.0009813353,0.0747715309,0.2314361781,-0.0140736336,0.1076158136,-0.4403277338,-0.4213162661,0.6965722442,-0.0531169921,-0.2233040333,-0.3283874393,0.1428037137,0.0477949269,0.2180052996,-0.7143525481,-0.1181045771,0.3435353637,-0.1029728353,-0.366305083,0.2125729769,0.1637896448,-0.1559287012,-0.1214725077,0.311553061,0.0770348534,-0.0543998964,-0.028303735,-0.3904367983]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/288","title":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'","comments":"Thank you. It is version 0.2.6, which version is better?","body":"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'","comment_length":10,"text":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'\nThank you. It is version 0.2.6, which version is better?","embeddings":[0.0769836977,-0.4020940661,-0.1799209863,0.0659818947,0.1492687613,-0.003346167,0.55732584,0.3920768797,-0.0114857536,0.2672935724,-0.0524433628,0.3347723782,-0.0505946018,-0.3120033741,0.0111635383,-0.3291765749,0.0835466534,0.2736496031,-0.37093997,-0.2723834813,-0.2335172594,0.2800564766,-0.0776633248,0.2484975308,-0.1481041014,-0.0349433534,0.1214392036,0.1059611663,0.0675767437,-0.2886580229,0.1965774745,-0.1745566279,0.2275901288,0.4768029451,-0.0001043869,0.1361589134,0.3375531137,-0.0346087962,-0.1342187971,-0.3505931497,0.4713913202,0.0593214221,0.3363775015,-0.3500793874,-0.0507910028,-0.2969646454,-0.0969579443,-0.30893448,0.4441095889,0.4418491125,0.3010698855,0.1026526392,0.2446943521,-0.169915244,-0.0208351463,-0.0297836401,-0.0677934811,0.3753479719,0.1383522898,0.1474164873,0.3823755085,0.342466712,0.160931766,0.0776078403,0.3393225968,0.0322579108,0.6600615978,-0.4541978836,-0.0775012821,0.2736726701,0.1812872142,-0.1184436977,-0.3209805787,0.1269780993,0.0060105366,-0.2475019395,0.0616756529,-0.06774728,-0.2496780753,-0.032574065,0.4458368123,-0.0442663617,-0.1771987081,0.1953953058,-0.0185109787,0.1636362374,-0.044930961,-0.0506545231,0.0271611586,-0.3340252638,0.3872074783,0.0928654596,0.4484997094,0.0960105956,-0.0759613588,-0.1219853684,0.0625471696,-0.6255782843,-0.2034097016,-0.2914340496,0.0739517361,-0.0122585492,0.1180225462,0.2710880339,-0.1634567976,0.0667565018,0.1067271829,0.160689339,0.335481286,0.0793941543,0.328011781,0.1698743701,-0.3021524251,-0.125138551,-0.1567854881,0.0640961453,0.3410718143,0.0939639807,-0.5376574993,-0.0486492701,-0.1924375296,0.0474512838,-0.0055814907,0.2568105459,-0.0379515067,0.1916810572,0.4150514603,0.11126706,-0.2454229593,0.0046106484,-0.2229356468,0.3555662632,-0.0703071654,-0.2020584792,0.0587319173,0.1554544717,0.3014602363,-0.0993096381,0.1310544163,-0.0825211927,0.1695728451,0.0146686807,-0.0396083631,0.1542768478,-0.2942762375,-0.0451418795,0.2622388005,-0.1855244786,-0.2611550987,0.091217421,-0.2335643321,-0.0095623313,-0.4541479945,0.3482017517,0.250795722,0.110424757,0.2338398695,0.0304904748,0.1052182466,0.0470620319,-0.101509586,-0.2446223497,-0.1023621485,-0.3306634724,0.2848880291,0.1544385999,-0.0588522553,-0.1574843526,-0.1423849314,-0.0397362337,0.1201601923,0.2202340066,-0.1351770014,-0.0135140996,0.1174672022,-0.1719239056,0.6146841645,-0.5745144486,0.1056552455,-0.0611012615,0.1371087134,-0.0482239611,-0.0108275972,-0.0786206573,0.1418243051,-0.0617279485,0.3198719323,0.3495156169,0.2515051663,-0.1104580089,-0.3646187782,0.012272289,0.1300430149,0.0910370499,0.0965716168,0.0635002851,-0.0523701161,-0.0480599813,0.0647452176,-0.1598062068,0.004075455,0.123443149,0.0875847489,-0.0867793262,0.1008282751,-0.1719257832,-0.3230362236,-0.0569832884,-0.3147102892,0.0343750939,-0.0152414292,-0.1913605928,-0.1983193457,0.0127634313,0.119129926,-0.2337505966,0.3220001459,0.0348381661,-0.0130126383,0.1630757153,-0.2321627587,-0.0329379737,-0.1512498111,0.181532681,-0.1353039891,0.1641567945,-0.4232264459,-0.2587807477,-0.1687233597,0.35354954,0.0873158574,-0.3229031265,-0.0398910791,0.3210483193,-0.0660059303,-0.0963132754,-0.3560896814,-0.0151594486,0.3564194143,-0.1473701298,0.044008974,0.3544332087,0.1194430888,-0.106795162,0.066043511,0.2442780733,0.2975997329,-0.0210516416,0.2813411057,0.1925182492,0.0381851532,0.1245420799,-0.0583589785,-0.1804328114,0.1559882909,0.3096714914,0.1182772219,-0.1786590964,-0.3969525397,-0.0277689844,0.2105372548,0.0029683402,0.2126078606,-0.0307502821,-0.0751878694,0.1257900149,0.1558475643,-0.0766748637,0.1838900596,0.0091597186,0.1322796494,0.0727801621,-0.2034448534,-0.1412041932,0.1152334958,-0.0268808324,0.3497841656,-0.0709351376,-0.0873165056,0.0785132051,-0.0844333768,-0.3782753944,0.05666437,0.1657810658,-0.0829095691,0.0378239527,-0.26275298,-0.0227957889,-0.0947818086,-0.2296084017,0.1324962825,-0.1437409669,-0.0149990348,0.1761470586,-0.1750652939,0.2052592635,0.1883973628,-0.0567656383,0.0960681811,0.3126211762,-0.2592220008,-0.0779284015,-0.1663522869,0.1553854793,0.0330693349,-0.0856425464,0.1994274557,-0.2755847275,-0.120114252,0.1499013901,-0.4520840049,0.0524351485,0.0140112396,0.1148459166,0.2227616012,0.2215283513,0.0248776749,-0.3447566926,0.2151990533,-0.3349262178,-0.0258942172,-0.0942451507,0.053977523,-0.0956519395,-0.4885933995,-0.6345106363,-0.3095667958,-0.538667798,0.0189340804,0.1400108635,0.3738848865,0.3208152652,0.2370781004,0.4575053453,0.3019067645,0.445266813,0.0590105243,0.0270972084,0.3070289791,-0.4646860957,-0.3787640929,-0.1151326671,-0.1516299397,0.3044160306,-0.1614550352,-0.0524057709,-0.3177403808,-0.2742889822,0.2044684142,-0.2654448748,-0.0834642574,0.0780671909,0.1683275998,-0.2234417647,-0.1472693682,0.0398055241,0.1982986629,-0.0899709836,0.0503881052,0.1322678328,0.4105213583,-0.1355243474,0.641283989,0.1535908431,-0.5416244864,0.1315539628,0.0941807106,0.1784284562,0.1610788852,-0.0770013109,0.0641207471,0.1061399058,0.3015512526,-0.0606331378,-0.1528750956,-0.2624876797,-0.2485653609,0.0551917031,-0.2093648165,-0.0992520601,-0.1903233081,-0.222233355,0.0266475081,-0.1704294384,0.1294929087,-0.0302802324,0.2669641674,0.1404934525,0.3743702769,-0.029914448,0.0309316143,-0.1446655095,-0.1721434444,-0.168138504,0.2689387798,0.0391585045,0.350931108,0.0693768784,-0.0899103284,0.0954246521,0.1021384597,0.3387723565,-0.141647324,-0.2548166513,0.3434726298,-0.0369338654,-0.426663667,-0.1019397676,-0.298001945,-0.0479726493,0.4439178407,0.0099498928,-0.2568550408,-0.271645844,0.2455054373,-0.1302042454,-0.0676388666,0.0721810982,-0.1339272708,-0.4609440267,-0.337305218,-0.1392961591,0.195825085,0.1834748089,-0.1435066611,-0.1498092711,0.001615681,0.0100157037,0.0503716096,0.1900882125,0.2087095976,-0.1278277785,-0.1401393712,0.0382278301,0.2863094807,0.3246237338,0.0855566636,-0.3335414827,-0.2565034926,0.1705903262,-0.0674866214,0.145268783,0.0241156202,-0.1518263668,-0.0574669726,-0.4039544761,0.0352798887,0.3928391635,0.2366948873,0.4728352726,0.233066231,0.0414866954,-0.31564188,0.278352052,0.1127391607,0.0168916248,0.3914880157,-0.1173036098,-0.1339904368,0.1071419939,0.2551783919,0.7796511054,0.1979432702,0.0798775107,0.3982567787,-0.1491081417,0.2713763118,0.0214910451,0.1287784278,-0.2085438073,-0.1011961401,0.0327130482,-0.0950420499,0.2960813046,-0.1658862233,-0.476200968,-0.0088990666,-0.3046584427,-0.0855448246,-0.0103940498,0.25138551,0.0512061715,0.0007532024,0.083183229,0.1936745793,0.1429111958,0.1862857044,-0.2173997909,-0.1922255903,-0.1607504934,-0.3412706852,-0.0518728457,0.153183043,-0.2182635367,0.4361849129,0.1955285221,-0.3141541779,0.0928765908,0.3122994304,0.4542844892,-0.1912746727,-0.1373024136,-0.0474196039,-0.1302075237,-0.0439016484,-0.1522315741,0.1801161766,0.5092203021,-0.2898780406,-0.1360089183,0.2658412457,0.0044673975,0.2290174514,-0.4029322565,0.0144238602,-0.2015995383,-0.0062909569,0.207975924,-0.0527221859,-0.5108942986,-0.1215923205,0.2427589297,-0.0168557093,-0.3328695297,-0.0293962657,0.1110495627,-0.263274312,0.0632196739,0.0704504997,0.3525013626,0.1392199397,0.3493540287,-0.0360304974,-0.2955444157,-0.2647211254,-0.2726528943,0.0311377272,0.0112488735,0.3244880438,-0.2756336033,-0.3246463537,-0.0597305261,0.211653322,-0.0009634426,0.2978005111,-0.2246960104,-0.1632337719,-0.1972972453,0.1918155849,-0.1714804322,0.3040003479,-0.3490905762,0.4225679636,0.1137231663,-0.0430724025,-0.4419589937,0.1668350548,-0.0368808508,0.020018857,-0.1581536829,-0.2147509009,0.1676616073,0.2444770932,0.2284802049,0.0694871545,-0.1888405383,-0.2605834603,-0.170472011,0.0698552579,-0.2116407305,-0.1817622185,0.2404185832,0.0999821424,-0.1844204962,0.0863327235,-0.0292884465,-0.1320893615,-0.0019985072,0.5582348108,0.0387344882,0.5644739866,0.3869163394,0.0089126034,-0.1002072021,0.1331441849,-0.1345453113,0.021350868,0.0597498007,-0.2254040837,-0.3498137891,-0.1373037398,-0.1616208702,-0.1402896494,0.2207787186,-0.16355443,0.0553813726,-0.2637145817,0.1210697517,0.303383112,-0.3440740407,0.1544483453,0.0866090953,0.3013632298,-0.2628340721,0.2679738402,-0.0742736459,0.0446530357,0.2171688825,0.2485523224,-0.1163876876,-0.0176999439,0.0716812983,0.0811027512,-0.0912493244,-0.2889398634,0.3252561092,0.2326954752,-0.1652373821,0.0873996466,0.2099814266,-0.0217311438,0.1657243967,0.1753938794,-0.1132669076,0.3435326517,-0.0894798413,-0.1929242909,0.3750655651,0.1159625128,0.2090840489,0.2096315026,0.1593326479,0.2247309387,0.0840807483,0.165045768,-0.1709304899,-0.0828605443,-0.0514771938,0.134366408,-0.3270949125,-0.1075968072,0.0350296423,-0.1609695405,-0.1508623064,-0.4307509065,0.0921481475,-0.0709812194,0.1977212429,0.0450752638,-0.1878480315,-0.3252618015,0.1265001595,-0.0690778419,0.3112432957,-0.0885865241,0.1946066916,0.3189254403,-0.1001254544,-0.3750415742,0.3328012526,0.3252197206,0.5558351874,-0.2463056743,0.0623565912,-0.0271787271,-0.2585277259,0.1358801425,0.0069714678,0.1525976956,0.0096471645,0.2664268017,0.1872951984,-0.2842236459,0.093714267,0.0491212532,-0.0283321068,-0.2225133032,0.605899632,-0.3091025054,0.0241864193,-0.1419742405,-0.0250129309,-0.5140055418,-0.189438194,0.139050588,0.0070464471,0.0270313472,-0.2933402359,0.1228031516,-0.1128694788,0.3584936261,0.0038748723,0.3682335615,0.029320417,-0.0045026881,-0.3142127395,0.303709954,-0.1117279828,-0.0305645373,0.1000814959,0.0731207728,-0.2099718451,0.0324984714,0.2157515734,-0.1540497392,0.0103491219,0.1331142485,-0.4652892649,-0.0519434549,-0.3202302158,-0.1183811426,0.3270791173,-0.2909780145,0.0002811941,-0.2657406926,0.1885832846,-0.3409994543,0.0875840336,0.0808703229,0.4352749586,0.1772854477,0.1798701733,0.305210948,-0.043241553,-0.3241828382,0.1009184644,-0.1834964305,-0.2110900134,0.2475279272,0.2351478636,-0.0651715025,-0.1614931524,-0.4883197546,-0.2035161555,0.53690207,-0.0041251061,-0.2707729936,-0.2305567861,0.2251336724,0.0629380122,-0.0009813353,0.0747715309,0.2314361781,-0.0140736336,0.1076158136,-0.4403277338,-0.4213162661,0.6965722442,-0.0531169921,-0.2233040333,-0.3283874393,0.1428037137,0.0477949269,0.2180052996,-0.7143525481,-0.1181045771,0.3435353637,-0.1029728353,-0.366305083,0.2125729769,0.1637896448,-0.1559287012,-0.1214725077,0.311553061,0.0770348534,-0.0543998964,-0.028303735,-0.3904367983]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/288","title":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'","comments":"0.2.6 is three years old now, maybe try a more recent one, e.g. the current 0.3.2 if you can?","body":"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'","comment_length":19,"text":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'\n0.2.6 is three years old now, maybe try a more recent one, e.g. the current 0.3.2 if you can?","embeddings":[0.0769836977,-0.4020940661,-0.1799209863,0.0659818947,0.1492687613,-0.003346167,0.55732584,0.3920768797,-0.0114857536,0.2672935724,-0.0524433628,0.3347723782,-0.0505946018,-0.3120033741,0.0111635383,-0.3291765749,0.0835466534,0.2736496031,-0.37093997,-0.2723834813,-0.2335172594,0.2800564766,-0.0776633248,0.2484975308,-0.1481041014,-0.0349433534,0.1214392036,0.1059611663,0.0675767437,-0.2886580229,0.1965774745,-0.1745566279,0.2275901288,0.4768029451,-0.0001043869,0.1361589134,0.3375531137,-0.0346087962,-0.1342187971,-0.3505931497,0.4713913202,0.0593214221,0.3363775015,-0.3500793874,-0.0507910028,-0.2969646454,-0.0969579443,-0.30893448,0.4441095889,0.4418491125,0.3010698855,0.1026526392,0.2446943521,-0.169915244,-0.0208351463,-0.0297836401,-0.0677934811,0.3753479719,0.1383522898,0.1474164873,0.3823755085,0.342466712,0.160931766,0.0776078403,0.3393225968,0.0322579108,0.6600615978,-0.4541978836,-0.0775012821,0.2736726701,0.1812872142,-0.1184436977,-0.3209805787,0.1269780993,0.0060105366,-0.2475019395,0.0616756529,-0.06774728,-0.2496780753,-0.032574065,0.4458368123,-0.0442663617,-0.1771987081,0.1953953058,-0.0185109787,0.1636362374,-0.044930961,-0.0506545231,0.0271611586,-0.3340252638,0.3872074783,0.0928654596,0.4484997094,0.0960105956,-0.0759613588,-0.1219853684,0.0625471696,-0.6255782843,-0.2034097016,-0.2914340496,0.0739517361,-0.0122585492,0.1180225462,0.2710880339,-0.1634567976,0.0667565018,0.1067271829,0.160689339,0.335481286,0.0793941543,0.328011781,0.1698743701,-0.3021524251,-0.125138551,-0.1567854881,0.0640961453,0.3410718143,0.0939639807,-0.5376574993,-0.0486492701,-0.1924375296,0.0474512838,-0.0055814907,0.2568105459,-0.0379515067,0.1916810572,0.4150514603,0.11126706,-0.2454229593,0.0046106484,-0.2229356468,0.3555662632,-0.0703071654,-0.2020584792,0.0587319173,0.1554544717,0.3014602363,-0.0993096381,0.1310544163,-0.0825211927,0.1695728451,0.0146686807,-0.0396083631,0.1542768478,-0.2942762375,-0.0451418795,0.2622388005,-0.1855244786,-0.2611550987,0.091217421,-0.2335643321,-0.0095623313,-0.4541479945,0.3482017517,0.250795722,0.110424757,0.2338398695,0.0304904748,0.1052182466,0.0470620319,-0.101509586,-0.2446223497,-0.1023621485,-0.3306634724,0.2848880291,0.1544385999,-0.0588522553,-0.1574843526,-0.1423849314,-0.0397362337,0.1201601923,0.2202340066,-0.1351770014,-0.0135140996,0.1174672022,-0.1719239056,0.6146841645,-0.5745144486,0.1056552455,-0.0611012615,0.1371087134,-0.0482239611,-0.0108275972,-0.0786206573,0.1418243051,-0.0617279485,0.3198719323,0.3495156169,0.2515051663,-0.1104580089,-0.3646187782,0.012272289,0.1300430149,0.0910370499,0.0965716168,0.0635002851,-0.0523701161,-0.0480599813,0.0647452176,-0.1598062068,0.004075455,0.123443149,0.0875847489,-0.0867793262,0.1008282751,-0.1719257832,-0.3230362236,-0.0569832884,-0.3147102892,0.0343750939,-0.0152414292,-0.1913605928,-0.1983193457,0.0127634313,0.119129926,-0.2337505966,0.3220001459,0.0348381661,-0.0130126383,0.1630757153,-0.2321627587,-0.0329379737,-0.1512498111,0.181532681,-0.1353039891,0.1641567945,-0.4232264459,-0.2587807477,-0.1687233597,0.35354954,0.0873158574,-0.3229031265,-0.0398910791,0.3210483193,-0.0660059303,-0.0963132754,-0.3560896814,-0.0151594486,0.3564194143,-0.1473701298,0.044008974,0.3544332087,0.1194430888,-0.106795162,0.066043511,0.2442780733,0.2975997329,-0.0210516416,0.2813411057,0.1925182492,0.0381851532,0.1245420799,-0.0583589785,-0.1804328114,0.1559882909,0.3096714914,0.1182772219,-0.1786590964,-0.3969525397,-0.0277689844,0.2105372548,0.0029683402,0.2126078606,-0.0307502821,-0.0751878694,0.1257900149,0.1558475643,-0.0766748637,0.1838900596,0.0091597186,0.1322796494,0.0727801621,-0.2034448534,-0.1412041932,0.1152334958,-0.0268808324,0.3497841656,-0.0709351376,-0.0873165056,0.0785132051,-0.0844333768,-0.3782753944,0.05666437,0.1657810658,-0.0829095691,0.0378239527,-0.26275298,-0.0227957889,-0.0947818086,-0.2296084017,0.1324962825,-0.1437409669,-0.0149990348,0.1761470586,-0.1750652939,0.2052592635,0.1883973628,-0.0567656383,0.0960681811,0.3126211762,-0.2592220008,-0.0779284015,-0.1663522869,0.1553854793,0.0330693349,-0.0856425464,0.1994274557,-0.2755847275,-0.120114252,0.1499013901,-0.4520840049,0.0524351485,0.0140112396,0.1148459166,0.2227616012,0.2215283513,0.0248776749,-0.3447566926,0.2151990533,-0.3349262178,-0.0258942172,-0.0942451507,0.053977523,-0.0956519395,-0.4885933995,-0.6345106363,-0.3095667958,-0.538667798,0.0189340804,0.1400108635,0.3738848865,0.3208152652,0.2370781004,0.4575053453,0.3019067645,0.445266813,0.0590105243,0.0270972084,0.3070289791,-0.4646860957,-0.3787640929,-0.1151326671,-0.1516299397,0.3044160306,-0.1614550352,-0.0524057709,-0.3177403808,-0.2742889822,0.2044684142,-0.2654448748,-0.0834642574,0.0780671909,0.1683275998,-0.2234417647,-0.1472693682,0.0398055241,0.1982986629,-0.0899709836,0.0503881052,0.1322678328,0.4105213583,-0.1355243474,0.641283989,0.1535908431,-0.5416244864,0.1315539628,0.0941807106,0.1784284562,0.1610788852,-0.0770013109,0.0641207471,0.1061399058,0.3015512526,-0.0606331378,-0.1528750956,-0.2624876797,-0.2485653609,0.0551917031,-0.2093648165,-0.0992520601,-0.1903233081,-0.222233355,0.0266475081,-0.1704294384,0.1294929087,-0.0302802324,0.2669641674,0.1404934525,0.3743702769,-0.029914448,0.0309316143,-0.1446655095,-0.1721434444,-0.168138504,0.2689387798,0.0391585045,0.350931108,0.0693768784,-0.0899103284,0.0954246521,0.1021384597,0.3387723565,-0.141647324,-0.2548166513,0.3434726298,-0.0369338654,-0.426663667,-0.1019397676,-0.298001945,-0.0479726493,0.4439178407,0.0099498928,-0.2568550408,-0.271645844,0.2455054373,-0.1302042454,-0.0676388666,0.0721810982,-0.1339272708,-0.4609440267,-0.337305218,-0.1392961591,0.195825085,0.1834748089,-0.1435066611,-0.1498092711,0.001615681,0.0100157037,0.0503716096,0.1900882125,0.2087095976,-0.1278277785,-0.1401393712,0.0382278301,0.2863094807,0.3246237338,0.0855566636,-0.3335414827,-0.2565034926,0.1705903262,-0.0674866214,0.145268783,0.0241156202,-0.1518263668,-0.0574669726,-0.4039544761,0.0352798887,0.3928391635,0.2366948873,0.4728352726,0.233066231,0.0414866954,-0.31564188,0.278352052,0.1127391607,0.0168916248,0.3914880157,-0.1173036098,-0.1339904368,0.1071419939,0.2551783919,0.7796511054,0.1979432702,0.0798775107,0.3982567787,-0.1491081417,0.2713763118,0.0214910451,0.1287784278,-0.2085438073,-0.1011961401,0.0327130482,-0.0950420499,0.2960813046,-0.1658862233,-0.476200968,-0.0088990666,-0.3046584427,-0.0855448246,-0.0103940498,0.25138551,0.0512061715,0.0007532024,0.083183229,0.1936745793,0.1429111958,0.1862857044,-0.2173997909,-0.1922255903,-0.1607504934,-0.3412706852,-0.0518728457,0.153183043,-0.2182635367,0.4361849129,0.1955285221,-0.3141541779,0.0928765908,0.3122994304,0.4542844892,-0.1912746727,-0.1373024136,-0.0474196039,-0.1302075237,-0.0439016484,-0.1522315741,0.1801161766,0.5092203021,-0.2898780406,-0.1360089183,0.2658412457,0.0044673975,0.2290174514,-0.4029322565,0.0144238602,-0.2015995383,-0.0062909569,0.207975924,-0.0527221859,-0.5108942986,-0.1215923205,0.2427589297,-0.0168557093,-0.3328695297,-0.0293962657,0.1110495627,-0.263274312,0.0632196739,0.0704504997,0.3525013626,0.1392199397,0.3493540287,-0.0360304974,-0.2955444157,-0.2647211254,-0.2726528943,0.0311377272,0.0112488735,0.3244880438,-0.2756336033,-0.3246463537,-0.0597305261,0.211653322,-0.0009634426,0.2978005111,-0.2246960104,-0.1632337719,-0.1972972453,0.1918155849,-0.1714804322,0.3040003479,-0.3490905762,0.4225679636,0.1137231663,-0.0430724025,-0.4419589937,0.1668350548,-0.0368808508,0.020018857,-0.1581536829,-0.2147509009,0.1676616073,0.2444770932,0.2284802049,0.0694871545,-0.1888405383,-0.2605834603,-0.170472011,0.0698552579,-0.2116407305,-0.1817622185,0.2404185832,0.0999821424,-0.1844204962,0.0863327235,-0.0292884465,-0.1320893615,-0.0019985072,0.5582348108,0.0387344882,0.5644739866,0.3869163394,0.0089126034,-0.1002072021,0.1331441849,-0.1345453113,0.021350868,0.0597498007,-0.2254040837,-0.3498137891,-0.1373037398,-0.1616208702,-0.1402896494,0.2207787186,-0.16355443,0.0553813726,-0.2637145817,0.1210697517,0.303383112,-0.3440740407,0.1544483453,0.0866090953,0.3013632298,-0.2628340721,0.2679738402,-0.0742736459,0.0446530357,0.2171688825,0.2485523224,-0.1163876876,-0.0176999439,0.0716812983,0.0811027512,-0.0912493244,-0.2889398634,0.3252561092,0.2326954752,-0.1652373821,0.0873996466,0.2099814266,-0.0217311438,0.1657243967,0.1753938794,-0.1132669076,0.3435326517,-0.0894798413,-0.1929242909,0.3750655651,0.1159625128,0.2090840489,0.2096315026,0.1593326479,0.2247309387,0.0840807483,0.165045768,-0.1709304899,-0.0828605443,-0.0514771938,0.134366408,-0.3270949125,-0.1075968072,0.0350296423,-0.1609695405,-0.1508623064,-0.4307509065,0.0921481475,-0.0709812194,0.1977212429,0.0450752638,-0.1878480315,-0.3252618015,0.1265001595,-0.0690778419,0.3112432957,-0.0885865241,0.1946066916,0.3189254403,-0.1001254544,-0.3750415742,0.3328012526,0.3252197206,0.5558351874,-0.2463056743,0.0623565912,-0.0271787271,-0.2585277259,0.1358801425,0.0069714678,0.1525976956,0.0096471645,0.2664268017,0.1872951984,-0.2842236459,0.093714267,0.0491212532,-0.0283321068,-0.2225133032,0.605899632,-0.3091025054,0.0241864193,-0.1419742405,-0.0250129309,-0.5140055418,-0.189438194,0.139050588,0.0070464471,0.0270313472,-0.2933402359,0.1228031516,-0.1128694788,0.3584936261,0.0038748723,0.3682335615,0.029320417,-0.0045026881,-0.3142127395,0.303709954,-0.1117279828,-0.0305645373,0.1000814959,0.0731207728,-0.2099718451,0.0324984714,0.2157515734,-0.1540497392,0.0103491219,0.1331142485,-0.4652892649,-0.0519434549,-0.3202302158,-0.1183811426,0.3270791173,-0.2909780145,0.0002811941,-0.2657406926,0.1885832846,-0.3409994543,0.0875840336,0.0808703229,0.4352749586,0.1772854477,0.1798701733,0.305210948,-0.043241553,-0.3241828382,0.1009184644,-0.1834964305,-0.2110900134,0.2475279272,0.2351478636,-0.0651715025,-0.1614931524,-0.4883197546,-0.2035161555,0.53690207,-0.0041251061,-0.2707729936,-0.2305567861,0.2251336724,0.0629380122,-0.0009813353,0.0747715309,0.2314361781,-0.0140736336,0.1076158136,-0.4403277338,-0.4213162661,0.6965722442,-0.0531169921,-0.2233040333,-0.3283874393,0.1428037137,0.0477949269,0.2180052996,-0.7143525481,-0.1181045771,0.3435353637,-0.1029728353,-0.366305083,0.2125729769,0.1637896448,-0.1559287012,-0.1214725077,0.311553061,0.0770348534,-0.0543998964,-0.028303735,-0.3904367983]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/288","title":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'","comments":"Thanks guys! I upgraded dill and it works.","body":"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'","comment_length":8,"text":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'\nThanks guys! I upgraded dill and it works.","embeddings":[0.0769836977,-0.4020940661,-0.1799209863,0.0659818947,0.1492687613,-0.003346167,0.55732584,0.3920768797,-0.0114857536,0.2672935724,-0.0524433628,0.3347723782,-0.0505946018,-0.3120033741,0.0111635383,-0.3291765749,0.0835466534,0.2736496031,-0.37093997,-0.2723834813,-0.2335172594,0.2800564766,-0.0776633248,0.2484975308,-0.1481041014,-0.0349433534,0.1214392036,0.1059611663,0.0675767437,-0.2886580229,0.1965774745,-0.1745566279,0.2275901288,0.4768029451,-0.0001043869,0.1361589134,0.3375531137,-0.0346087962,-0.1342187971,-0.3505931497,0.4713913202,0.0593214221,0.3363775015,-0.3500793874,-0.0507910028,-0.2969646454,-0.0969579443,-0.30893448,0.4441095889,0.4418491125,0.3010698855,0.1026526392,0.2446943521,-0.169915244,-0.0208351463,-0.0297836401,-0.0677934811,0.3753479719,0.1383522898,0.1474164873,0.3823755085,0.342466712,0.160931766,0.0776078403,0.3393225968,0.0322579108,0.6600615978,-0.4541978836,-0.0775012821,0.2736726701,0.1812872142,-0.1184436977,-0.3209805787,0.1269780993,0.0060105366,-0.2475019395,0.0616756529,-0.06774728,-0.2496780753,-0.032574065,0.4458368123,-0.0442663617,-0.1771987081,0.1953953058,-0.0185109787,0.1636362374,-0.044930961,-0.0506545231,0.0271611586,-0.3340252638,0.3872074783,0.0928654596,0.4484997094,0.0960105956,-0.0759613588,-0.1219853684,0.0625471696,-0.6255782843,-0.2034097016,-0.2914340496,0.0739517361,-0.0122585492,0.1180225462,0.2710880339,-0.1634567976,0.0667565018,0.1067271829,0.160689339,0.335481286,0.0793941543,0.328011781,0.1698743701,-0.3021524251,-0.125138551,-0.1567854881,0.0640961453,0.3410718143,0.0939639807,-0.5376574993,-0.0486492701,-0.1924375296,0.0474512838,-0.0055814907,0.2568105459,-0.0379515067,0.1916810572,0.4150514603,0.11126706,-0.2454229593,0.0046106484,-0.2229356468,0.3555662632,-0.0703071654,-0.2020584792,0.0587319173,0.1554544717,0.3014602363,-0.0993096381,0.1310544163,-0.0825211927,0.1695728451,0.0146686807,-0.0396083631,0.1542768478,-0.2942762375,-0.0451418795,0.2622388005,-0.1855244786,-0.2611550987,0.091217421,-0.2335643321,-0.0095623313,-0.4541479945,0.3482017517,0.250795722,0.110424757,0.2338398695,0.0304904748,0.1052182466,0.0470620319,-0.101509586,-0.2446223497,-0.1023621485,-0.3306634724,0.2848880291,0.1544385999,-0.0588522553,-0.1574843526,-0.1423849314,-0.0397362337,0.1201601923,0.2202340066,-0.1351770014,-0.0135140996,0.1174672022,-0.1719239056,0.6146841645,-0.5745144486,0.1056552455,-0.0611012615,0.1371087134,-0.0482239611,-0.0108275972,-0.0786206573,0.1418243051,-0.0617279485,0.3198719323,0.3495156169,0.2515051663,-0.1104580089,-0.3646187782,0.012272289,0.1300430149,0.0910370499,0.0965716168,0.0635002851,-0.0523701161,-0.0480599813,0.0647452176,-0.1598062068,0.004075455,0.123443149,0.0875847489,-0.0867793262,0.1008282751,-0.1719257832,-0.3230362236,-0.0569832884,-0.3147102892,0.0343750939,-0.0152414292,-0.1913605928,-0.1983193457,0.0127634313,0.119129926,-0.2337505966,0.3220001459,0.0348381661,-0.0130126383,0.1630757153,-0.2321627587,-0.0329379737,-0.1512498111,0.181532681,-0.1353039891,0.1641567945,-0.4232264459,-0.2587807477,-0.1687233597,0.35354954,0.0873158574,-0.3229031265,-0.0398910791,0.3210483193,-0.0660059303,-0.0963132754,-0.3560896814,-0.0151594486,0.3564194143,-0.1473701298,0.044008974,0.3544332087,0.1194430888,-0.106795162,0.066043511,0.2442780733,0.2975997329,-0.0210516416,0.2813411057,0.1925182492,0.0381851532,0.1245420799,-0.0583589785,-0.1804328114,0.1559882909,0.3096714914,0.1182772219,-0.1786590964,-0.3969525397,-0.0277689844,0.2105372548,0.0029683402,0.2126078606,-0.0307502821,-0.0751878694,0.1257900149,0.1558475643,-0.0766748637,0.1838900596,0.0091597186,0.1322796494,0.0727801621,-0.2034448534,-0.1412041932,0.1152334958,-0.0268808324,0.3497841656,-0.0709351376,-0.0873165056,0.0785132051,-0.0844333768,-0.3782753944,0.05666437,0.1657810658,-0.0829095691,0.0378239527,-0.26275298,-0.0227957889,-0.0947818086,-0.2296084017,0.1324962825,-0.1437409669,-0.0149990348,0.1761470586,-0.1750652939,0.2052592635,0.1883973628,-0.0567656383,0.0960681811,0.3126211762,-0.2592220008,-0.0779284015,-0.1663522869,0.1553854793,0.0330693349,-0.0856425464,0.1994274557,-0.2755847275,-0.120114252,0.1499013901,-0.4520840049,0.0524351485,0.0140112396,0.1148459166,0.2227616012,0.2215283513,0.0248776749,-0.3447566926,0.2151990533,-0.3349262178,-0.0258942172,-0.0942451507,0.053977523,-0.0956519395,-0.4885933995,-0.6345106363,-0.3095667958,-0.538667798,0.0189340804,0.1400108635,0.3738848865,0.3208152652,0.2370781004,0.4575053453,0.3019067645,0.445266813,0.0590105243,0.0270972084,0.3070289791,-0.4646860957,-0.3787640929,-0.1151326671,-0.1516299397,0.3044160306,-0.1614550352,-0.0524057709,-0.3177403808,-0.2742889822,0.2044684142,-0.2654448748,-0.0834642574,0.0780671909,0.1683275998,-0.2234417647,-0.1472693682,0.0398055241,0.1982986629,-0.0899709836,0.0503881052,0.1322678328,0.4105213583,-0.1355243474,0.641283989,0.1535908431,-0.5416244864,0.1315539628,0.0941807106,0.1784284562,0.1610788852,-0.0770013109,0.0641207471,0.1061399058,0.3015512526,-0.0606331378,-0.1528750956,-0.2624876797,-0.2485653609,0.0551917031,-0.2093648165,-0.0992520601,-0.1903233081,-0.222233355,0.0266475081,-0.1704294384,0.1294929087,-0.0302802324,0.2669641674,0.1404934525,0.3743702769,-0.029914448,0.0309316143,-0.1446655095,-0.1721434444,-0.168138504,0.2689387798,0.0391585045,0.350931108,0.0693768784,-0.0899103284,0.0954246521,0.1021384597,0.3387723565,-0.141647324,-0.2548166513,0.3434726298,-0.0369338654,-0.426663667,-0.1019397676,-0.298001945,-0.0479726493,0.4439178407,0.0099498928,-0.2568550408,-0.271645844,0.2455054373,-0.1302042454,-0.0676388666,0.0721810982,-0.1339272708,-0.4609440267,-0.337305218,-0.1392961591,0.195825085,0.1834748089,-0.1435066611,-0.1498092711,0.001615681,0.0100157037,0.0503716096,0.1900882125,0.2087095976,-0.1278277785,-0.1401393712,0.0382278301,0.2863094807,0.3246237338,0.0855566636,-0.3335414827,-0.2565034926,0.1705903262,-0.0674866214,0.145268783,0.0241156202,-0.1518263668,-0.0574669726,-0.4039544761,0.0352798887,0.3928391635,0.2366948873,0.4728352726,0.233066231,0.0414866954,-0.31564188,0.278352052,0.1127391607,0.0168916248,0.3914880157,-0.1173036098,-0.1339904368,0.1071419939,0.2551783919,0.7796511054,0.1979432702,0.0798775107,0.3982567787,-0.1491081417,0.2713763118,0.0214910451,0.1287784278,-0.2085438073,-0.1011961401,0.0327130482,-0.0950420499,0.2960813046,-0.1658862233,-0.476200968,-0.0088990666,-0.3046584427,-0.0855448246,-0.0103940498,0.25138551,0.0512061715,0.0007532024,0.083183229,0.1936745793,0.1429111958,0.1862857044,-0.2173997909,-0.1922255903,-0.1607504934,-0.3412706852,-0.0518728457,0.153183043,-0.2182635367,0.4361849129,0.1955285221,-0.3141541779,0.0928765908,0.3122994304,0.4542844892,-0.1912746727,-0.1373024136,-0.0474196039,-0.1302075237,-0.0439016484,-0.1522315741,0.1801161766,0.5092203021,-0.2898780406,-0.1360089183,0.2658412457,0.0044673975,0.2290174514,-0.4029322565,0.0144238602,-0.2015995383,-0.0062909569,0.207975924,-0.0527221859,-0.5108942986,-0.1215923205,0.2427589297,-0.0168557093,-0.3328695297,-0.0293962657,0.1110495627,-0.263274312,0.0632196739,0.0704504997,0.3525013626,0.1392199397,0.3493540287,-0.0360304974,-0.2955444157,-0.2647211254,-0.2726528943,0.0311377272,0.0112488735,0.3244880438,-0.2756336033,-0.3246463537,-0.0597305261,0.211653322,-0.0009634426,0.2978005111,-0.2246960104,-0.1632337719,-0.1972972453,0.1918155849,-0.1714804322,0.3040003479,-0.3490905762,0.4225679636,0.1137231663,-0.0430724025,-0.4419589937,0.1668350548,-0.0368808508,0.020018857,-0.1581536829,-0.2147509009,0.1676616073,0.2444770932,0.2284802049,0.0694871545,-0.1888405383,-0.2605834603,-0.170472011,0.0698552579,-0.2116407305,-0.1817622185,0.2404185832,0.0999821424,-0.1844204962,0.0863327235,-0.0292884465,-0.1320893615,-0.0019985072,0.5582348108,0.0387344882,0.5644739866,0.3869163394,0.0089126034,-0.1002072021,0.1331441849,-0.1345453113,0.021350868,0.0597498007,-0.2254040837,-0.3498137891,-0.1373037398,-0.1616208702,-0.1402896494,0.2207787186,-0.16355443,0.0553813726,-0.2637145817,0.1210697517,0.303383112,-0.3440740407,0.1544483453,0.0866090953,0.3013632298,-0.2628340721,0.2679738402,-0.0742736459,0.0446530357,0.2171688825,0.2485523224,-0.1163876876,-0.0176999439,0.0716812983,0.0811027512,-0.0912493244,-0.2889398634,0.3252561092,0.2326954752,-0.1652373821,0.0873996466,0.2099814266,-0.0217311438,0.1657243967,0.1753938794,-0.1132669076,0.3435326517,-0.0894798413,-0.1929242909,0.3750655651,0.1159625128,0.2090840489,0.2096315026,0.1593326479,0.2247309387,0.0840807483,0.165045768,-0.1709304899,-0.0828605443,-0.0514771938,0.134366408,-0.3270949125,-0.1075968072,0.0350296423,-0.1609695405,-0.1508623064,-0.4307509065,0.0921481475,-0.0709812194,0.1977212429,0.0450752638,-0.1878480315,-0.3252618015,0.1265001595,-0.0690778419,0.3112432957,-0.0885865241,0.1946066916,0.3189254403,-0.1001254544,-0.3750415742,0.3328012526,0.3252197206,0.5558351874,-0.2463056743,0.0623565912,-0.0271787271,-0.2585277259,0.1358801425,0.0069714678,0.1525976956,0.0096471645,0.2664268017,0.1872951984,-0.2842236459,0.093714267,0.0491212532,-0.0283321068,-0.2225133032,0.605899632,-0.3091025054,0.0241864193,-0.1419742405,-0.0250129309,-0.5140055418,-0.189438194,0.139050588,0.0070464471,0.0270313472,-0.2933402359,0.1228031516,-0.1128694788,0.3584936261,0.0038748723,0.3682335615,0.029320417,-0.0045026881,-0.3142127395,0.303709954,-0.1117279828,-0.0305645373,0.1000814959,0.0731207728,-0.2099718451,0.0324984714,0.2157515734,-0.1540497392,0.0103491219,0.1331142485,-0.4652892649,-0.0519434549,-0.3202302158,-0.1183811426,0.3270791173,-0.2909780145,0.0002811941,-0.2657406926,0.1885832846,-0.3409994543,0.0875840336,0.0808703229,0.4352749586,0.1772854477,0.1798701733,0.305210948,-0.043241553,-0.3241828382,0.1009184644,-0.1834964305,-0.2110900134,0.2475279272,0.2351478636,-0.0651715025,-0.1614931524,-0.4883197546,-0.2035161555,0.53690207,-0.0041251061,-0.2707729936,-0.2305567861,0.2251336724,0.0629380122,-0.0009813353,0.0747715309,0.2314361781,-0.0140736336,0.1076158136,-0.4403277338,-0.4213162661,0.6965722442,-0.0531169921,-0.2233040333,-0.3283874393,0.1428037137,0.0477949269,0.2180052996,-0.7143525481,-0.1181045771,0.3435353637,-0.1029728353,-0.366305083,0.2125729769,0.1637896448,-0.1559287012,-0.1214725077,0.311553061,0.0770348534,-0.0543998964,-0.028303735,-0.3904367983]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/288","title":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'","comments":"Awesome","body":"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'","comment_length":1,"text":"Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:470: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:471: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:472: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:473: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/tensorflow\/python\/framework\/dtypes.py:476: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) \/ '(1,)type'.\r\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/importlib\/_bootstrap.py:219: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6\r\n return f(*args, **kwds)\r\n\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/h5py\/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`.\r\n from ._conv import register_converters as _register_converters\r\nTraceback (most recent call last):\r\n File \"\/Users\/parasol_tree\/Resource\/019 - Github\/AcademicEnglishToolkit \/test.py\", line 7, in <module>\r\n import nlp\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/__init__.py\", line 27, in <module>\r\n from .arrow_dataset import Dataset\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/arrow_dataset.py\", line 31, in <module>\r\n from nlp.utils.py_utils import dumps\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/__init__.py\", line 20, in <module>\r\n from .download_manager import DownloadManager, GenerateMode\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py\", line 25, in <module>\r\n from .py_utils import flatten_nested, map_nested, size_str\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 244, in <module>\r\n class Pickler(dill.Pickler):\r\n File \"\/Users\/parasol_tree\/anaconda3\/lib\/python3.6\/site-packages\/nlp\/utils\/py_utils.py\", line 247, in Pickler\r\n dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy())\r\nAttributeError: module 'dill' has no attribute '_dill'\nAwesome","embeddings":[0.0769836977,-0.4020940661,-0.1799209863,0.0659818947,0.1492687613,-0.003346167,0.55732584,0.3920768797,-0.0114857536,0.2672935724,-0.0524433628,0.3347723782,-0.0505946018,-0.3120033741,0.0111635383,-0.3291765749,0.0835466534,0.2736496031,-0.37093997,-0.2723834813,-0.2335172594,0.2800564766,-0.0776633248,0.2484975308,-0.1481041014,-0.0349433534,0.1214392036,0.1059611663,0.0675767437,-0.2886580229,0.1965774745,-0.1745566279,0.2275901288,0.4768029451,-0.0001043869,0.1361589134,0.3375531137,-0.0346087962,-0.1342187971,-0.3505931497,0.4713913202,0.0593214221,0.3363775015,-0.3500793874,-0.0507910028,-0.2969646454,-0.0969579443,-0.30893448,0.4441095889,0.4418491125,0.3010698855,0.1026526392,0.2446943521,-0.169915244,-0.0208351463,-0.0297836401,-0.0677934811,0.3753479719,0.1383522898,0.1474164873,0.3823755085,0.342466712,0.160931766,0.0776078403,0.3393225968,0.0322579108,0.6600615978,-0.4541978836,-0.0775012821,0.2736726701,0.1812872142,-0.1184436977,-0.3209805787,0.1269780993,0.0060105366,-0.2475019395,0.0616756529,-0.06774728,-0.2496780753,-0.032574065,0.4458368123,-0.0442663617,-0.1771987081,0.1953953058,-0.0185109787,0.1636362374,-0.044930961,-0.0506545231,0.0271611586,-0.3340252638,0.3872074783,0.0928654596,0.4484997094,0.0960105956,-0.0759613588,-0.1219853684,0.0625471696,-0.6255782843,-0.2034097016,-0.2914340496,0.0739517361,-0.0122585492,0.1180225462,0.2710880339,-0.1634567976,0.0667565018,0.1067271829,0.160689339,0.335481286,0.0793941543,0.328011781,0.1698743701,-0.3021524251,-0.125138551,-0.1567854881,0.0640961453,0.3410718143,0.0939639807,-0.5376574993,-0.0486492701,-0.1924375296,0.0474512838,-0.0055814907,0.2568105459,-0.0379515067,0.1916810572,0.4150514603,0.11126706,-0.2454229593,0.0046106484,-0.2229356468,0.3555662632,-0.0703071654,-0.2020584792,0.0587319173,0.1554544717,0.3014602363,-0.0993096381,0.1310544163,-0.0825211927,0.1695728451,0.0146686807,-0.0396083631,0.1542768478,-0.2942762375,-0.0451418795,0.2622388005,-0.1855244786,-0.2611550987,0.091217421,-0.2335643321,-0.0095623313,-0.4541479945,0.3482017517,0.250795722,0.110424757,0.2338398695,0.0304904748,0.1052182466,0.0470620319,-0.101509586,-0.2446223497,-0.1023621485,-0.3306634724,0.2848880291,0.1544385999,-0.0588522553,-0.1574843526,-0.1423849314,-0.0397362337,0.1201601923,0.2202340066,-0.1351770014,-0.0135140996,0.1174672022,-0.1719239056,0.6146841645,-0.5745144486,0.1056552455,-0.0611012615,0.1371087134,-0.0482239611,-0.0108275972,-0.0786206573,0.1418243051,-0.0617279485,0.3198719323,0.3495156169,0.2515051663,-0.1104580089,-0.3646187782,0.012272289,0.1300430149,0.0910370499,0.0965716168,0.0635002851,-0.0523701161,-0.0480599813,0.0647452176,-0.1598062068,0.004075455,0.123443149,0.0875847489,-0.0867793262,0.1008282751,-0.1719257832,-0.3230362236,-0.0569832884,-0.3147102892,0.0343750939,-0.0152414292,-0.1913605928,-0.1983193457,0.0127634313,0.119129926,-0.2337505966,0.3220001459,0.0348381661,-0.0130126383,0.1630757153,-0.2321627587,-0.0329379737,-0.1512498111,0.181532681,-0.1353039891,0.1641567945,-0.4232264459,-0.2587807477,-0.1687233597,0.35354954,0.0873158574,-0.3229031265,-0.0398910791,0.3210483193,-0.0660059303,-0.0963132754,-0.3560896814,-0.0151594486,0.3564194143,-0.1473701298,0.044008974,0.3544332087,0.1194430888,-0.106795162,0.066043511,0.2442780733,0.2975997329,-0.0210516416,0.2813411057,0.1925182492,0.0381851532,0.1245420799,-0.0583589785,-0.1804328114,0.1559882909,0.3096714914,0.1182772219,-0.1786590964,-0.3969525397,-0.0277689844,0.2105372548,0.0029683402,0.2126078606,-0.0307502821,-0.0751878694,0.1257900149,0.1558475643,-0.0766748637,0.1838900596,0.0091597186,0.1322796494,0.0727801621,-0.2034448534,-0.1412041932,0.1152334958,-0.0268808324,0.3497841656,-0.0709351376,-0.0873165056,0.0785132051,-0.0844333768,-0.3782753944,0.05666437,0.1657810658,-0.0829095691,0.0378239527,-0.26275298,-0.0227957889,-0.0947818086,-0.2296084017,0.1324962825,-0.1437409669,-0.0149990348,0.1761470586,-0.1750652939,0.2052592635,0.1883973628,-0.0567656383,0.0960681811,0.3126211762,-0.2592220008,-0.0779284015,-0.1663522869,0.1553854793,0.0330693349,-0.0856425464,0.1994274557,-0.2755847275,-0.120114252,0.1499013901,-0.4520840049,0.0524351485,0.0140112396,0.1148459166,0.2227616012,0.2215283513,0.0248776749,-0.3447566926,0.2151990533,-0.3349262178,-0.0258942172,-0.0942451507,0.053977523,-0.0956519395,-0.4885933995,-0.6345106363,-0.3095667958,-0.538667798,0.0189340804,0.1400108635,0.3738848865,0.3208152652,0.2370781004,0.4575053453,0.3019067645,0.445266813,0.0590105243,0.0270972084,0.3070289791,-0.4646860957,-0.3787640929,-0.1151326671,-0.1516299397,0.3044160306,-0.1614550352,-0.0524057709,-0.3177403808,-0.2742889822,0.2044684142,-0.2654448748,-0.0834642574,0.0780671909,0.1683275998,-0.2234417647,-0.1472693682,0.0398055241,0.1982986629,-0.0899709836,0.0503881052,0.1322678328,0.4105213583,-0.1355243474,0.641283989,0.1535908431,-0.5416244864,0.1315539628,0.0941807106,0.1784284562,0.1610788852,-0.0770013109,0.0641207471,0.1061399058,0.3015512526,-0.0606331378,-0.1528750956,-0.2624876797,-0.2485653609,0.0551917031,-0.2093648165,-0.0992520601,-0.1903233081,-0.222233355,0.0266475081,-0.1704294384,0.1294929087,-0.0302802324,0.2669641674,0.1404934525,0.3743702769,-0.029914448,0.0309316143,-0.1446655095,-0.1721434444,-0.168138504,0.2689387798,0.0391585045,0.350931108,0.0693768784,-0.0899103284,0.0954246521,0.1021384597,0.3387723565,-0.141647324,-0.2548166513,0.3434726298,-0.0369338654,-0.426663667,-0.1019397676,-0.298001945,-0.0479726493,0.4439178407,0.0099498928,-0.2568550408,-0.271645844,0.2455054373,-0.1302042454,-0.0676388666,0.0721810982,-0.1339272708,-0.4609440267,-0.337305218,-0.1392961591,0.195825085,0.1834748089,-0.1435066611,-0.1498092711,0.001615681,0.0100157037,0.0503716096,0.1900882125,0.2087095976,-0.1278277785,-0.1401393712,0.0382278301,0.2863094807,0.3246237338,0.0855566636,-0.3335414827,-0.2565034926,0.1705903262,-0.0674866214,0.145268783,0.0241156202,-0.1518263668,-0.0574669726,-0.4039544761,0.0352798887,0.3928391635,0.2366948873,0.4728352726,0.233066231,0.0414866954,-0.31564188,0.278352052,0.1127391607,0.0168916248,0.3914880157,-0.1173036098,-0.1339904368,0.1071419939,0.2551783919,0.7796511054,0.1979432702,0.0798775107,0.3982567787,-0.1491081417,0.2713763118,0.0214910451,0.1287784278,-0.2085438073,-0.1011961401,0.0327130482,-0.0950420499,0.2960813046,-0.1658862233,-0.476200968,-0.0088990666,-0.3046584427,-0.0855448246,-0.0103940498,0.25138551,0.0512061715,0.0007532024,0.083183229,0.1936745793,0.1429111958,0.1862857044,-0.2173997909,-0.1922255903,-0.1607504934,-0.3412706852,-0.0518728457,0.153183043,-0.2182635367,0.4361849129,0.1955285221,-0.3141541779,0.0928765908,0.3122994304,0.4542844892,-0.1912746727,-0.1373024136,-0.0474196039,-0.1302075237,-0.0439016484,-0.1522315741,0.1801161766,0.5092203021,-0.2898780406,-0.1360089183,0.2658412457,0.0044673975,0.2290174514,-0.4029322565,0.0144238602,-0.2015995383,-0.0062909569,0.207975924,-0.0527221859,-0.5108942986,-0.1215923205,0.2427589297,-0.0168557093,-0.3328695297,-0.0293962657,0.1110495627,-0.263274312,0.0632196739,0.0704504997,0.3525013626,0.1392199397,0.3493540287,-0.0360304974,-0.2955444157,-0.2647211254,-0.2726528943,0.0311377272,0.0112488735,0.3244880438,-0.2756336033,-0.3246463537,-0.0597305261,0.211653322,-0.0009634426,0.2978005111,-0.2246960104,-0.1632337719,-0.1972972453,0.1918155849,-0.1714804322,0.3040003479,-0.3490905762,0.4225679636,0.1137231663,-0.0430724025,-0.4419589937,0.1668350548,-0.0368808508,0.020018857,-0.1581536829,-0.2147509009,0.1676616073,0.2444770932,0.2284802049,0.0694871545,-0.1888405383,-0.2605834603,-0.170472011,0.0698552579,-0.2116407305,-0.1817622185,0.2404185832,0.0999821424,-0.1844204962,0.0863327235,-0.0292884465,-0.1320893615,-0.0019985072,0.5582348108,0.0387344882,0.5644739866,0.3869163394,0.0089126034,-0.1002072021,0.1331441849,-0.1345453113,0.021350868,0.0597498007,-0.2254040837,-0.3498137891,-0.1373037398,-0.1616208702,-0.1402896494,0.2207787186,-0.16355443,0.0553813726,-0.2637145817,0.1210697517,0.303383112,-0.3440740407,0.1544483453,0.0866090953,0.3013632298,-0.2628340721,0.2679738402,-0.0742736459,0.0446530357,0.2171688825,0.2485523224,-0.1163876876,-0.0176999439,0.0716812983,0.0811027512,-0.0912493244,-0.2889398634,0.3252561092,0.2326954752,-0.1652373821,0.0873996466,0.2099814266,-0.0217311438,0.1657243967,0.1753938794,-0.1132669076,0.3435326517,-0.0894798413,-0.1929242909,0.3750655651,0.1159625128,0.2090840489,0.2096315026,0.1593326479,0.2247309387,0.0840807483,0.165045768,-0.1709304899,-0.0828605443,-0.0514771938,0.134366408,-0.3270949125,-0.1075968072,0.0350296423,-0.1609695405,-0.1508623064,-0.4307509065,0.0921481475,-0.0709812194,0.1977212429,0.0450752638,-0.1878480315,-0.3252618015,0.1265001595,-0.0690778419,0.3112432957,-0.0885865241,0.1946066916,0.3189254403,-0.1001254544,-0.3750415742,0.3328012526,0.3252197206,0.5558351874,-0.2463056743,0.0623565912,-0.0271787271,-0.2585277259,0.1358801425,0.0069714678,0.1525976956,0.0096471645,0.2664268017,0.1872951984,-0.2842236459,0.093714267,0.0491212532,-0.0283321068,-0.2225133032,0.605899632,-0.3091025054,0.0241864193,-0.1419742405,-0.0250129309,-0.5140055418,-0.189438194,0.139050588,0.0070464471,0.0270313472,-0.2933402359,0.1228031516,-0.1128694788,0.3584936261,0.0038748723,0.3682335615,0.029320417,-0.0045026881,-0.3142127395,0.303709954,-0.1117279828,-0.0305645373,0.1000814959,0.0731207728,-0.2099718451,0.0324984714,0.2157515734,-0.1540497392,0.0103491219,0.1331142485,-0.4652892649,-0.0519434549,-0.3202302158,-0.1183811426,0.3270791173,-0.2909780145,0.0002811941,-0.2657406926,0.1885832846,-0.3409994543,0.0875840336,0.0808703229,0.4352749586,0.1772854477,0.1798701733,0.305210948,-0.043241553,-0.3241828382,0.1009184644,-0.1834964305,-0.2110900134,0.2475279272,0.2351478636,-0.0651715025,-0.1614931524,-0.4883197546,-0.2035161555,0.53690207,-0.0041251061,-0.2707729936,-0.2305567861,0.2251336724,0.0629380122,-0.0009813353,0.0747715309,0.2314361781,-0.0140736336,0.1076158136,-0.4403277338,-0.4213162661,0.6965722442,-0.0531169921,-0.2233040333,-0.3283874393,0.1428037137,0.0477949269,0.2180052996,-0.7143525481,-0.1181045771,0.3435353637,-0.1029728353,-0.366305083,0.2125729769,0.1637896448,-0.1559287012,-0.1214725077,0.311553061,0.0770348534,-0.0543998964,-0.028303735,-0.3904367983]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/281","title":"Private\/sensitive data","comments":"Hi @MFreidank, you should already be able to load a dataset from local sources, indeed. (ping @lhoestq and @jplu)\r\n\r\nWe're also thinking about the ability to host private datasets on a hosted bucket with permission management, but that's further down the road.","body":"Hi all,\r\nThanks for this fantastic library, it makes it very easy to do prototyping for NLP projects interchangeably between TF\/Pytorch. \r\n\r\nUnfortunately, there is data that cannot easily be shared publicly as it may contain sensitive information. \r\nIs there support\/a plan to support such data with NLP, e.g. by reading it from local sources?\r\n\r\nUse case flow could look like this: use NLP to prototype an approach on similar, public data and apply the resulting prototype on sensitive\/private data without the need to rethink data processing pipelines. \r\n\r\nMany thanks for your responses ahead of time and kind regards,\r\nMFreidank","comment_length":42,"text":"Private\/sensitive data\nHi all,\r\nThanks for this fantastic library, it makes it very easy to do prototyping for NLP projects interchangeably between TF\/Pytorch. \r\n\r\nUnfortunately, there is data that cannot easily be shared publicly as it may contain sensitive information. \r\nIs there support\/a plan to support such data with NLP, e.g. by reading it from local sources?\r\n\r\nUse case flow could look like this: use NLP to prototype an approach on similar, public data and apply the resulting prototype on sensitive\/private data without the need to rethink data processing pipelines. \r\n\r\nMany thanks for your responses ahead of time and kind regards,\r\nMFreidank\nHi @MFreidank, you should already be able to load a dataset from local sources, indeed. (ping @lhoestq and @jplu)\r\n\r\nWe're also thinking about the ability to host private datasets on a hosted bucket with permission management, but that's further down the road.","embeddings":[-0.0707458109,0.4125887752,-0.1550576985,0.2336954176,-0.3745613098,-0.3104677498,0.1758047789,0.2741076052,-0.187402606,0.0843257084,-0.099034436,0.235512346,-0.1740504205,0.4147525132,0.156123817,-0.141062513,-0.1745984107,0.3279705346,0.0204366203,0.123477526,0.2168058455,-0.0035431683,-0.1398066282,0.1810401976,0.0957490802,-0.0827764943,0.0768581554,-0.0709032044,-0.0051912307,-0.2377834469,0.1994736344,0.2674770951,0.3569665551,0.1103215218,-0.0001181756,-0.0618794262,-0.0077543999,-0.0915450752,0.1014710963,-0.3869082034,0.0909538865,-0.3763212264,0.1161264405,-0.0942868665,-0.6016390324,-0.2237698138,0.223329857,-0.2056765258,0.2856585681,-0.1414274871,0.0898133889,0.4285238683,-0.0426420681,0.4334448576,0.2679482102,0.0425534435,-0.3776562512,-0.2314343005,0.4403627515,0.0660785809,-0.1837641597,0.0795590803,-0.1058342829,0.0370284282,0.2091089934,-0.0535619333,-0.2636175454,-0.4229537845,-0.0417968594,0.4476955831,0.6776380539,-0.0553220026,-0.3027353585,-0.1976625919,0.0497470535,-0.1227308437,0.0437942371,0.3050142825,-0.3012392819,0.2311728448,-0.2347911745,-0.5339365602,-0.5878560543,0.4005849063,0.3250871301,0.172593683,-0.1147480085,0.0097190253,-0.1834225655,-0.1467662305,0.3118985593,0.0840847716,0.3460735381,0.3613792658,0.2314935625,-0.3157440424,-0.123499766,-0.1307466477,0.1263952702,0.0600157343,0.0627787486,0.3807478547,-0.4343865812,0.1869550794,0.5641598105,-0.3314897716,-0.27328071,0.2127691805,0.6141747832,-0.2264549583,0.1624816805,0.1731202006,-0.0866314843,0.1338378638,-0.3697251081,-0.1365144104,0.0459270217,-0.3774811029,-0.0641482994,0.1029619426,-0.2001150548,0.0772923082,0.0345184803,0.1005116478,0.0606263764,0.0826020911,0.2194363475,-0.2596558928,0.1002445668,-0.4008681476,0.0736742765,-0.3296131194,-0.230513081,0.1907190084,0.358513236,0.052518107,0.1227400005,-0.1362953931,0.1661897004,0.0949334279,0.182130456,-0.0007126121,0.2064194828,0.2837126851,-0.1496316344,-0.2135072201,-0.2083340436,-0.0681238249,-0.2654196024,0.4042986333,-0.2182285637,-0.3589202166,-0.0729403868,-0.0733458102,-0.3603952825,0.0497680828,0.1830289215,0.1923413873,-0.0132968742,-0.357829839,0.1233032793,0.0540306941,-0.3091997802,-0.3107890487,0.2044462562,0.0079360837,-0.21543625,0.1870301515,0.0801904276,-0.0899683759,0.0907674581,0.3384080529,-0.3551779389,-0.1171650141,0.1238018721,0.1438644379,0.6366678476,-0.1484181285,-0.4632449448,0.773643136,0.0050810995,-0.349994719,-0.0010337892,0.5067375898,0.0797293782,-0.0977019444,-0.1144037992,0.7414237261,0.0494311638,0.0938225463,-0.1705517024,-0.452065438,0.4799664319,0.3717123866,-0.007751755,0.044731807,0.2441882193,0.3608098328,0.2845618129,-0.0610938668,0.0669619665,-0.0351169221,0.1061966494,0.2022458762,-0.0028933052,-0.2733085752,0.0834963098,-0.1555231512,-0.5438956022,0.149952814,0.033630468,-0.1679928154,0.228863582,-0.0996507481,0.1972913593,-0.0412272178,0.0136140008,-0.0050666025,-0.0585574917,-0.0237949733,-0.3576771021,-0.0481377579,-0.1494691968,-0.153427124,-0.0364810079,0.2351106852,0.0654012412,-0.0547394156,0.3041162193,0.5591760278,-0.0135732358,-0.1076304913,0.1822916567,0.1524810195,-0.185741961,0.2442169636,0.0726282597,0.5805903673,0.3432220519,-0.1757588387,0.2598481178,0.3666487038,-0.0341157392,0.1733643711,0.0984331816,0.7042987347,-0.2638027668,-0.1616846174,-0.0321972966,-0.149155587,-0.004895024,-0.173579365,-0.3520435691,-0.0304909572,0.010926743,-0.2666922212,0.1307037622,-0.0318236388,-0.5742003918,0.2258845866,-0.1088710055,-0.0877471045,0.305164665,-0.0261934493,-0.0274327938,-0.0649805889,0.1479395777,0.2978618741,0.3659072816,0.1113848761,-0.1728381664,-0.0731302276,-0.1175538823,-0.1416867375,-0.0020144924,-0.0593261346,0.1523485929,0.0491004735,0.1965176165,0.354804337,-0.0670658574,0.1788393408,-0.073028028,0.0943154618,-0.0095762732,0.1214590222,-0.2423714846,-0.38159886,0.072255075,-0.0107747633,-0.1310466826,0.2822643518,0.155437842,0.1999242604,-0.0822517946,-0.2093529701,-0.2692062855,0.4863938689,-0.2974672318,-0.157526046,-0.0836522803,-0.1240205914,-0.096003361,0.0475061834,0.3210487962,-0.0360234231,0.3541307151,0.4143390656,0.0107448259,-0.1943289042,-0.3171733916,0.0647922307,0.0316571966,0.2427788079,-0.0861499459,-0.0453930534,0.087456964,-0.2234255224,-0.0734251589,-0.419618398,-0.0061414493,-0.2804930806,-0.2080231458,-0.0215628836,-0.2728872299,-0.1179999635,-0.5042937398,-0.1746373326,0.3917945921,0.0175452866,0.4474543035,0.0974578485,-0.2773035169,0.0967274904,0.0109266415,-0.0492402613,-0.0719493032,-0.4233448207,0.2278181612,-0.5833615065,-0.4579364359,-0.116012685,-0.0033489377,-0.0401588045,0.206042245,-0.625767529,-0.037111897,0.1528884619,0.2928291261,-0.0337190218,-0.2740913033,0.2435399741,0.0247672219,0.070463635,-0.1655634791,-0.1309505552,-0.2497606277,0.0240281913,0.2235518694,0.3631480932,-0.3084203005,0.4264250398,0.3714079857,-0.0315979458,-0.1078972965,0.3445717096,0.2310940176,0.158288762,0.2725990117,0.0424835645,0.1564814299,-0.0483453274,-0.1651373059,0.3218414187,0.1150265485,-0.471721679,-0.3430369496,-0.4360589683,-0.3518178165,-0.1752528995,0.1419397742,0.1484615207,0.6546312571,0.0345410593,-0.276591599,-0.2643121183,-0.5170109868,-0.0706208572,-0.0597595908,0.47088328,0.1222507805,-0.0337745771,-0.227718994,-0.4081262946,-0.089446418,0.1126338318,0.3884765208,-0.1986562908,0.3296014667,0.2983247042,-0.0142935663,-0.0284239668,-0.1028787047,0.1847768724,0.0703571066,-0.124977231,-0.094941169,0.3198770881,-0.1120712161,0.3463827968,0.1639566422,0.3733924925,-0.1324216872,-0.4414402843,0.1851907521,-0.0816765204,-0.2308580577,-0.194366917,0.1598500311,-0.0552421026,-0.1051101461,-0.0835655481,-0.1508893222,0.0903198272,-0.1171172485,0.1277925372,0.088006258,0.315584749,0.2185907364,0.049270831,-0.1255630404,0.2209627777,-0.0733037964,0.236970678,0.3940101564,-0.2094965577,0.2412099987,0.3167013526,0.1162359789,0.0532840975,0.0491740294,0.4401218593,0.4852981567,0.1008363366,-0.0215202477,-0.0505710803,-0.1575636864,-0.2853379548,-0.0430659428,0.0431453548,-0.3380031884,0.1288782656,-0.150150001,0.4707925916,-0.1566841602,-0.1687040627,0.0655271485,0.0078111985,-0.1811357737,0.4684836268,0.316080302,1.0504853725,-0.3008869588,0.4145292938,-0.0155621832,-0.0584205352,0.5179578662,-0.4314320087,0.021201225,-0.1692847013,0.1825361848,-0.1650936007,-0.017549118,0.1086113974,0.0474041514,0.1744170487,0.1485565454,0.3124348521,-0.1689552814,0.0563683063,0.3495236039,0.355099231,-0.2769204676,-0.4173377752,-0.0087036304,-0.2215892673,0.235939011,-0.0035172498,-0.5324001908,-0.0602276437,0.2643843889,-0.2098199129,-0.0754235685,-0.1397769898,0.0012054958,0.1134634987,-0.3987623453,0.5135905147,0.4438513219,0.2498894781,0.1626673043,-0.5347725153,-0.2794633508,-0.2712433636,-0.338013798,0.1475773752,0.2529192865,0.2925344706,-0.2589453757,-0.1438816935,0.0959938392,0.036560744,-0.6549979448,-0.1946158707,0.0389002971,0.4339012206,-0.1514926553,-0.1014143601,0.2259866893,-0.3171100616,0.1088032946,0.0551686175,0.4022719562,-0.1132711694,-0.030477643,0.2982135415,-0.0944538563,0.0758535042,0.2888768017,0.2115994692,-0.0030284629,0.2133069634,-0.0831056312,0.0497213155,-0.0599176921,0.3089555204,-0.1253231019,-0.1244070083,-0.2226338387,0.269151181,-0.1552933306,-0.1963363588,0.3266524673,-0.028323926,-0.2674144506,-0.2095963657,-0.5977570415,-0.20257999,0.338437438,-0.1217569336,-0.061131347,0.0242835451,0.4509021342,0.0306481775,0.1603468359,-0.1998890191,0.2316725999,-0.2023933828,0.2082710713,0.2860322893,-0.0122923534,-0.1575119495,0.1134107113,-0.1109398231,0.0907798782,-0.1975744665,-0.0771207809,-0.1202204078,0.2204321623,0.3247850537,-0.414647907,-0.2224480361,-0.7372852564,-0.2496774346,0.0521291271,0.0286329146,0.1434051096,0.009585348,0.2350489348,-0.0954044089,0.2745676637,-0.0399392024,0.3295143843,0.1933569461,0.4057556689,-0.1316053718,0.3267180622,-0.0094615705,-0.1354019791,0.0131138805,-0.0508942753,0.2899572849,-0.0574265458,0.6205945611,-0.1105115339,-0.0442554466,0.0270666387,0.1421189755,0.0180960093,-0.0618105754,0.3354998529,0.1648721695,-0.024212461,-0.2641049325,-0.2132457793,-0.1139493361,0.1806272417,-0.2042880058,-0.0469639599,0.6455292106,0.2042319179,0.1332355142,-0.2595514357,-0.0175489448,-0.2535315752,-0.3921943009,0.0616936758,-0.2349587083,-0.0407698378,-0.1246158183,0.3356882334,0.0016509632,0.0995056927,0.2140596807,0.3517780602,0.2666942477,0.0436918885,0.1604704559,-0.1709220707,-0.2023498118,-0.0662173331,0.3786062598,0.2248146385,-0.3483100832,0.3978134096,0.0678842142,0.114485994,0.2501385212,0.1266714931,0.0771578699,-0.2194930464,-0.1399848759,-0.0571312271,0.1511866599,-0.1383663416,-0.2714174092,-0.3089694977,0.0563507788,0.2587431073,0.1964287162,-0.1202956289,-0.0749596879,0.3919233978,0.100972265,-0.2236845195,0.468072176,-0.2789300978,0.1513322592,0.1175628901,0.1943866462,-0.0025959935,0.0940575302,-0.4987980425,0.0357724763,-0.5205508471,-0.0558123626,-0.0178785641,0.3984953165,0.3403483629,-0.1176034585,0.3315959573,-0.1105009094,0.0863524899,0.1713413745,0.0579947047,0.1456905305,-0.2098474354,0.4859402776,0.0796210021,-0.4297902882,0.0770101845,-0.1739656776,0.1241070181,0.1104916558,0.4481084645,-0.1365601271,0.0984852687,-0.1572194397,0.0321087651,-0.1452292651,0.3510954678,0.2182414383,0.5691325068,-0.0877609253,-0.206383571,-0.3088741601,0.4102728963,0.1998161525,-0.0833827704,0.061262954,0.1454944909,0.3595326543,0.602091074,-0.2580795586,-0.2385394275,0.0807672888,-0.0594958067,0.0656703264,0.1543276161,-0.2401903719,0.3685511351,-0.1283996254,-0.1714860797,0.2581997514,-0.3168207407,-0.1080095395,-0.2567026615,0.1452495158,-0.0999767333,0.2646899521,0.4240325093,-0.0464151353,0.0738045946,0.0059066433,-0.1054795459,-0.1852464825,-0.0544457957,-0.0390874334,-0.1180694997,0.183038339,-0.0789943114,-0.3894793689,0.0768444836,-0.1321230382,-0.0878608599,-0.3873867393,0.2484509051,0.2884972394,-0.1554188281,-0.3809860051,0.100615494,-0.0149357701,0.3789590597,-0.1933852285,-0.2485029548,-0.1345480829,0.1518797278,0.2221588343,-0.7115601301,0.052677352,0.1163988933,0.1823837459,0.111415185,-0.1442591548,-0.4165101647,0.0034092073,0.1030431315,0.0760392398,-0.0012032549,-0.0935162529,-0.0271772612,-0.1414024681,-0.1616559476,0.3910855651,-0.1132720783,0.0102489796,-0.3302137852,-0.3756519258]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/281","title":"Private\/sensitive data","comments":"Hi @MFreidank, it is possible to load a dataset from your local storage, but only CSV\/TSV and JSON are supported. To load a dataset in JSON format:\r\n\r\n```\r\nnlp.load_dataset(path=\"json\", data_files={nlp.Split.TRAIN: [\"path\/to\/train.json\"], nlp.Split.TEST: [\"path\/to\/test.json\"]})\r\n```\r\n\r\nFor CSV\/TSV datasets, you have to replace `json` by `csv`.","body":"Hi all,\r\nThanks for this fantastic library, it makes it very easy to do prototyping for NLP projects interchangeably between TF\/Pytorch. \r\n\r\nUnfortunately, there is data that cannot easily be shared publicly as it may contain sensitive information. \r\nIs there support\/a plan to support such data with NLP, e.g. by reading it from local sources?\r\n\r\nUse case flow could look like this: use NLP to prototype an approach on similar, public data and apply the resulting prototype on sensitive\/private data without the need to rethink data processing pipelines. \r\n\r\nMany thanks for your responses ahead of time and kind regards,\r\nMFreidank","comment_length":44,"text":"Private\/sensitive data\nHi all,\r\nThanks for this fantastic library, it makes it very easy to do prototyping for NLP projects interchangeably between TF\/Pytorch. \r\n\r\nUnfortunately, there is data that cannot easily be shared publicly as it may contain sensitive information. \r\nIs there support\/a plan to support such data with NLP, e.g. by reading it from local sources?\r\n\r\nUse case flow could look like this: use NLP to prototype an approach on similar, public data and apply the resulting prototype on sensitive\/private data without the need to rethink data processing pipelines. \r\n\r\nMany thanks for your responses ahead of time and kind regards,\r\nMFreidank\nHi @MFreidank, it is possible to load a dataset from your local storage, but only CSV\/TSV and JSON are supported. To load a dataset in JSON format:\r\n\r\n```\r\nnlp.load_dataset(path=\"json\", data_files={nlp.Split.TRAIN: [\"path\/to\/train.json\"], nlp.Split.TEST: [\"path\/to\/test.json\"]})\r\n```\r\n\r\nFor CSV\/TSV datasets, you have to replace `json` by `csv`.","embeddings":[-0.0166605357,0.3993093073,-0.1244864762,0.2548536658,-0.2739622295,-0.2752925754,0.2181889862,0.308070004,-0.0690384805,0.0078153489,-0.1233371347,0.2415469289,-0.1098864228,0.431302011,0.2261041403,-0.1605279893,-0.2306254357,0.3290660679,0.0751647204,0.1278129816,0.1798416227,0.0108177625,-0.027710788,0.1978319138,0.1053762361,-0.0224126279,0.0391383097,-0.0110528544,0.0493090414,-0.2807991803,0.2194712162,0.1461137533,0.3711489141,0.1042129025,-0.0001194865,0.0547816083,0.0042786831,-0.1700389832,0.014008496,-0.5010938048,0.118179366,-0.4590620697,0.2127284855,-0.1323252469,-0.5507670641,-0.3828507364,0.195341289,-0.2550372779,0.4459568262,-0.0134742884,0.0972172618,0.3948141634,-0.0238350388,0.4301288128,0.1516950428,0.1558122784,-0.2978108227,-0.1546912193,0.3256403804,0.0576554276,-0.1445109546,-0.0260695852,-0.0631829277,0.108187668,0.1694032699,0.0011481006,-0.2196285576,-0.3277963698,-0.0628166795,0.399366051,0.5945553184,-0.0275267567,-0.2495580465,-0.2670380175,0.0319775194,-0.1998359114,0.0711245313,0.4118881226,-0.289634496,0.2282141596,-0.3382027447,-0.4331628382,-0.5370650291,0.4045125246,0.3479412496,0.0709438473,-0.0933347046,0.0860698596,-0.1627711803,-0.1994019747,0.197381556,-0.0023525564,0.3907437623,0.3041099906,0.1776646227,-0.3675431609,-0.1347600371,-0.3807192147,0.0989573076,-0.0090187527,0.0206542239,0.3384737968,-0.3744046092,0.212100625,0.5879611969,-0.2345270962,-0.2241452485,0.2144237012,0.4542486966,-0.2493967116,0.0432523601,0.0823550448,-0.052257441,0.1146748811,-0.4078924656,-0.1083300561,0.0567062013,-0.3220865428,-0.0730881244,0.0340624973,-0.2571914792,0.0476955101,-0.0437502377,0.1794607937,0.0698910952,0.104365401,0.28446576,-0.1818287075,0.1176628172,-0.3972922266,0.0611345358,-0.254008472,-0.1454838961,0.2200012803,0.4175849855,0.094150953,0.1839074492,-0.2577002645,0.1028045192,-0.0206567999,0.1476295441,-0.1018630341,0.2499348968,0.246468544,-0.0304197464,-0.2639297545,-0.0956272781,-0.1191156656,-0.3232686818,0.3597046435,-0.2853327096,-0.3630567491,-0.1132007241,-0.0605303459,-0.4056692123,0.099638015,0.1104567423,0.1905024499,0.0251190625,-0.3158795536,0.0254389681,0.0203312244,-0.3600462973,-0.2895978689,0.2112156898,-0.0422562957,-0.2529892325,0.0836327374,0.0734239519,-0.1013646945,0.0622593053,0.3870765865,-0.3909443915,-0.0891318396,0.0643113926,0.242074728,0.680860579,-0.0877781883,-0.3745702207,0.8095407486,-0.0852320418,-0.3504332304,-0.0121514369,0.4251954556,0.1100581288,0.0039028972,-0.0741009638,0.8450555801,0.048078645,0.1301187575,-0.2199296802,-0.5326390266,0.5446045399,0.3769712746,-0.0556492992,0.0064388844,0.2360025048,0.276570797,0.3048285246,-0.0953290686,0.1123543009,0.0592962503,0.0538814664,0.2307034582,-0.0348530076,-0.2128134817,-0.0293981824,-0.1252918839,-0.4451179504,0.0170486793,0.0235721376,-0.1477998644,0.1749249697,-0.09533225,0.1014212146,0.0133184325,0.0436021797,0.0738188699,0.0279040486,0.0492114946,-0.2668811083,-0.0932446718,-0.1285628378,-0.1814226657,-0.0982199833,0.3334493041,0.0696581826,-0.0688517541,0.3223309219,0.5017097592,0.0012493958,-0.1183680519,0.118309997,0.1488822252,-0.1287769079,0.3108648956,-0.025181774,0.5072873235,0.2461736947,-0.1875599325,0.2004159838,0.347764045,0.0099599473,0.1449291259,0.0077634887,0.7716162801,-0.2840053141,-0.1025070325,-0.0444631949,-0.1314356923,0.0763065144,-0.1933103651,-0.285468787,-0.0716620088,0.013611136,-0.2660618126,0.1443147063,-0.0389845259,-0.5130639672,0.1489171982,0.0806592926,-0.0316482037,0.272053957,0.0470779724,-0.0526903495,-0.1681950837,0.1684117466,0.2464767545,0.4391745925,0.0823374689,-0.1612999737,-0.0511493944,-0.1021538153,-0.1314149648,-0.0250146147,-0.1019916683,0.2686356604,0.1020199209,0.2694602907,0.2681826651,-0.0358641073,0.2545010448,-0.0103044305,0.0745376125,-0.0335338153,0.1670872718,-0.3016680479,-0.3807734549,-0.0498780422,-0.0134986546,-0.0696675926,0.245334506,0.0742738098,0.1491188705,-0.1015831977,-0.2358610183,-0.2040970922,0.3944403529,-0.3479895592,-0.2545349896,-0.0299717207,-0.1484679133,-0.1017738208,0.0162803475,0.3572624922,0.1227777824,0.3335718513,0.3912318945,-0.0778014809,-0.0446859039,-0.2103272825,0.050436724,0.080468215,0.2831274569,-0.1011571959,0.0038761778,0.0902633145,-0.2536836565,0.0319102742,-0.3446621597,-0.0655551404,-0.1926712692,-0.1188364625,-0.0175068323,-0.3231666088,-0.1083950102,-0.6310309768,-0.2055350989,0.5200822353,0.0638064742,0.4213783741,0.0322194248,-0.2250576317,0.1284990907,-0.0232753698,-0.0056158407,-0.0860952064,-0.3206884861,0.1609563529,-0.5506543517,-0.4848898351,-0.0580535531,-0.0164761432,-0.0689202771,0.141336903,-0.6018842459,-0.0163094085,0.1117276475,0.2389303148,-0.0290665906,-0.373802036,0.3266041279,-0.0315826051,0.1052386239,-0.1330154836,-0.1841297895,-0.1981923133,0.0980571136,0.1647399515,0.360386163,-0.1718626767,0.4904503524,0.4867066741,0.0318252631,-0.2435638607,0.3236861229,0.144201532,0.1402071118,0.2634196281,-0.0255519953,0.088593781,-0.1194417551,-0.1527338326,0.3284191489,0.0189495254,-0.3911525011,-0.3489370346,-0.3677772284,-0.3942302465,-0.2331292629,0.1649867743,0.1421860605,0.6201733351,0.046695523,-0.2608411312,-0.2799127698,-0.5743591785,-0.1883243322,0.0333730169,0.4541786015,0.1794544905,-0.2198924124,-0.3218363822,-0.3483028114,-0.0576323494,0.1259959042,0.5013283491,-0.176070109,0.2361574769,0.2682358027,-0.0649791211,-0.0827489719,-0.0625152811,0.1840816587,0.1204325259,-0.2159570009,-0.1733563095,0.2778737843,-0.1154482365,0.2914932966,0.1579726636,0.451898396,-0.1268479973,-0.4344648123,0.0740376264,-0.0018674722,-0.2597160935,-0.1623038501,0.0858571604,-0.0005458181,-0.1680851728,-0.1617956609,-0.2033610046,0.1895053834,-0.1780154109,0.2198022306,0.0369274393,0.3316866755,0.1751579642,0.021388188,-0.0622160733,0.1842966676,-0.0651343763,0.1851908416,0.4601458311,-0.1989147514,0.2502401471,0.3350355923,0.0330981687,0.0061629075,0.0343217514,0.4424754083,0.4467557967,0.1427552402,0.0831139088,-0.06380108,-0.2052033544,-0.3313947022,0.0394384339,0.0263970289,-0.3888713121,0.2493765354,-0.1147730649,0.4262862504,-0.1439487487,-0.1177986935,0.0192917716,-0.1739882529,-0.2069434077,0.4962059259,0.3118307889,0.962321341,-0.2769501209,0.4369778633,-0.0007589597,-0.0796760023,0.5795754194,-0.3859019876,0.0353772081,-0.2179757804,0.2022815198,-0.1116758063,-0.0795930922,0.1102604121,0.0471171811,0.1516530812,0.102854237,0.352925241,-0.1564784944,0.0187946912,0.4241891503,0.3827804327,-0.2476407886,-0.458424598,-0.018389795,-0.2846644521,0.1845097989,0.0521561019,-0.5859048367,-0.0993527025,0.1903272569,-0.2218104601,-0.0261950772,-0.0932061151,0.1091063619,0.1207724139,-0.375426054,0.4350094199,0.4593996406,0.1446845829,0.150602296,-0.4995850325,-0.3325147033,-0.3276655376,-0.3055796623,0.1467470527,0.2023526579,0.2135422081,-0.1806581914,-0.1123799905,0.0807516724,0.0361318886,-0.624740541,-0.1462794244,0.0490614511,0.3756270111,-0.1372780204,-0.2192110717,0.1635213494,-0.3295556903,0.1200830936,0.0558869466,0.3473139405,-0.0922017097,-0.0771196708,0.3084156513,-0.1292666048,0.0950123072,0.3675634861,0.1777257323,0.0451069735,0.3061799705,-0.0806228295,0.0671476871,-0.072373569,0.4255792499,-0.1506490409,-0.1059943363,-0.2445386946,0.2478553057,-0.1326415837,-0.1967224777,0.3327677548,-0.0178121552,-0.2137249708,-0.1364901513,-0.6354273558,-0.2036847919,0.4110691547,-0.1202099621,-0.0279876404,0.1253854334,0.4923393428,-0.0341548994,0.1964405328,-0.1920211762,0.2586788833,-0.2582669258,0.2504993081,0.2631959319,-0.0244439133,-0.1339118779,0.0571432747,-0.0981990024,0.1221238524,-0.2253525853,-0.0848291591,-0.0873340815,0.2244150788,0.3185943067,-0.3540015817,-0.3149128258,-0.7644426823,-0.2544848323,0.0528769568,-0.0703585371,0.0833143815,-0.0135876331,0.2956675887,-0.1838445216,0.299015969,-0.1141425595,0.2512926459,0.06109621,0.4579073787,-0.2031513304,0.306196332,0.0342514366,-0.0534237847,-0.1696287096,-0.0755228698,0.2967054248,0.017716242,0.5531839132,-0.1011196002,-0.054283116,-0.0767906681,0.1179896966,0.1121239886,-0.162612319,0.4437756836,0.2562170923,-0.0280372929,-0.1709076315,-0.1344102323,0.0505611077,0.188720569,-0.1945667863,-0.0385908559,0.6310536861,0.2019317448,0.1544494629,-0.2350540757,0.0282641761,-0.2131879926,-0.3182070851,0.1499154121,-0.2367614359,-0.0561610088,-0.1369699091,0.3565436602,0.1027592272,0.0230532438,0.2235872597,0.3232924938,0.1492786705,0.0956351385,0.1313735247,-0.1607819945,-0.1441289783,0.0050666998,0.3178525269,0.2633243501,-0.341727823,0.4735431969,0.0169945545,0.0939850658,0.1979446113,0.1597337127,0.0963841602,-0.172723949,-0.2211094052,0.0090337163,0.0167273972,-0.0787443444,-0.2220301181,-0.3007415831,0.0090921428,0.2761209309,0.1736833006,-0.1054708511,0.0242292639,0.415009439,-0.0088845985,-0.2298019081,0.4949308336,-0.2747698128,0.1947166324,0.1332121193,0.1098014265,-0.0170675926,0.1371527761,-0.4706299603,-0.0387309529,-0.4303734899,-0.0753472596,-0.0319211371,0.4246704876,0.3504907489,-0.1474107206,0.4288035333,-0.0736394152,0.0725055933,0.1342594773,0.0807934925,0.1512714624,-0.214133814,0.5109365582,0.057175871,-0.3916727006,0.1018428653,-0.2378976345,0.0655674785,0.0608693771,0.3961156607,-0.0111908857,0.1875630915,-0.1559856981,0.0220439602,-0.1954658329,0.3519825041,0.1441507936,0.4340550005,-0.0215129498,-0.1880168021,-0.338827163,0.3261443377,0.1111396179,-0.1446434855,-0.0292378291,0.237992093,0.3607325256,0.5662312508,-0.4303020537,-0.089550823,0.0791641474,-0.0186343584,0.0788817108,0.0707130581,-0.1803404093,0.3250213563,-0.146503076,-0.2175437361,0.2301376462,-0.2784760594,-0.0956142843,-0.267098099,0.1264287084,-0.1632897556,0.3355629146,0.4106147587,-0.0469925739,0.1489314139,-0.012029958,-0.052776929,-0.159295544,-0.0594433844,-0.0192186367,-0.0094691878,0.1541529149,0.0798312798,-0.4466328323,0.2209000736,-0.168168962,-0.1616940498,-0.4171390533,0.1572962701,0.3395829499,-0.169963181,-0.4020066261,0.1447924376,0.0951193422,0.3582449555,-0.2050878704,-0.2601104677,-0.1358627826,0.2258004248,0.2183452994,-0.689233005,0.0244710613,0.1786981374,0.1260713786,0.1074208692,-0.1344649494,-0.3104922473,-0.0233730134,0.1823252141,0.0039113699,0.042208828,-0.0490769334,0.0117351804,-0.1120160669,-0.1942620277,0.3277128041,-0.0371014327,-0.0271971002,-0.3719162345,-0.366928041]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/281","title":"Private\/sensitive data","comments":"Hi @julien-c @jplu,\r\nThanks for sharing this solution with me, it helps, this is what I was looking for. \r\nIf not already there and only missed by me, this could be a great addition in the docs.\r\n\r\nClosing my issue as resolved, thanks again.","body":"Hi all,\r\nThanks for this fantastic library, it makes it very easy to do prototyping for NLP projects interchangeably between TF\/Pytorch. \r\n\r\nUnfortunately, there is data that cannot easily be shared publicly as it may contain sensitive information. \r\nIs there support\/a plan to support such data with NLP, e.g. by reading it from local sources?\r\n\r\nUse case flow could look like this: use NLP to prototype an approach on similar, public data and apply the resulting prototype on sensitive\/private data without the need to rethink data processing pipelines. \r\n\r\nMany thanks for your responses ahead of time and kind regards,\r\nMFreidank","comment_length":44,"text":"Private\/sensitive data\nHi all,\r\nThanks for this fantastic library, it makes it very easy to do prototyping for NLP projects interchangeably between TF\/Pytorch. \r\n\r\nUnfortunately, there is data that cannot easily be shared publicly as it may contain sensitive information. \r\nIs there support\/a plan to support such data with NLP, e.g. by reading it from local sources?\r\n\r\nUse case flow could look like this: use NLP to prototype an approach on similar, public data and apply the resulting prototype on sensitive\/private data without the need to rethink data processing pipelines. \r\n\r\nMany thanks for your responses ahead of time and kind regards,\r\nMFreidank\nHi @julien-c @jplu,\r\nThanks for sharing this solution with me, it helps, this is what I was looking for. \r\nIf not already there and only missed by me, this could be a great addition in the docs.\r\n\r\nClosing my issue as resolved, thanks again.","embeddings":[0.0258489307,0.3822511435,-0.1474190801,0.1858093888,-0.3208057284,-0.2908139527,0.1759027839,0.31238994,-0.2786038816,0.079381302,-0.0625791848,0.2926683426,-0.0952913463,0.4064874947,0.1631554365,-0.2160253972,-0.2121718824,0.3283576965,0.0399660729,0.1607280374,0.3044917285,0.0305082686,-0.0668024793,0.2438914329,0.1339483112,-0.1141139269,0.1144153848,-0.1650368124,0.0475930087,-0.2738578618,0.1451488137,0.1825234443,0.2605031729,0.0681904629,-0.000118519,-0.0630127862,0.0080701588,-0.1219932362,0.1805238575,-0.3296790421,0.169830218,-0.4010601044,0.1010971516,-0.075986743,-0.6062310338,-0.1920445263,0.2262507528,-0.1614921689,0.233020559,-0.1536771208,0.1125909239,0.3779383898,0.1014638543,0.4482401311,0.2235453129,0.0362447686,-0.3435452282,-0.3351929486,0.4353991151,0.0831449628,-0.1078776717,0.1273633689,-0.0848964453,-0.0166317523,0.0692506805,-0.0495769009,-0.1109369695,-0.4897403717,0.0032607289,0.4364346266,0.6228167415,-0.0677119493,-0.2318775505,-0.1826434731,0.1105767116,-0.1296249628,0.0149735576,0.2753042877,-0.3189122379,0.1809320152,-0.1930097789,-0.4642404914,-0.5439631939,0.3049429357,0.3671753407,0.0410555787,-0.1337723732,0.0076692468,-0.1846585721,-0.2136173695,0.1772072017,0.0620255135,0.3948307037,0.2782180011,0.3060156405,-0.4116638005,-0.109462887,-0.2372367978,0.1159115061,-0.0435348153,0.0135567142,0.3441677988,-0.4206798375,0.1662000865,0.4476668537,-0.3055676222,-0.2571874857,0.2101721615,0.601036489,-0.2004008889,0.1096265018,0.1507365406,0.016363034,0.1081607938,-0.3722083867,-0.1131602004,0.0403443314,-0.3706246614,-0.1459184736,0.140654102,-0.2036221176,0.1335197985,0.0404557995,0.133901149,0.0462411642,0.0996247455,0.2197240144,-0.2955405116,0.1329736114,-0.4402605295,0.1208012402,-0.3411676884,-0.2346657515,0.2266023904,0.3104593456,0.1657722145,0.1307254583,-0.0847361684,0.102638863,0.051932428,0.1602874994,-0.0611958317,0.2653575838,0.2281562835,-0.1429543197,-0.205706805,-0.2207786441,-0.0444301516,-0.2550649941,0.4310934544,-0.2649871707,-0.3226992488,-0.0442522764,-0.0621587262,-0.3281317949,0.0434703268,0.3314890563,0.2257246971,0.0723519996,-0.3221593201,0.1531202495,0.053211268,-0.3242949247,-0.3337930441,0.1474688053,-0.112137951,-0.1244513318,0.1684965938,0.145300135,-0.0814065337,0.1459770203,0.3059071302,-0.3773403764,-0.2175031155,0.10776788,0.1623148769,0.7060899138,-0.0930079445,-0.4636054933,0.787204206,-0.0351001769,-0.3584575653,0.058801122,0.5080543756,0.0448329598,-0.1144134924,-0.1989552826,0.6941428185,0.0089125969,0.1344359517,-0.2006943971,-0.4680550098,0.5286426544,0.2506704628,-0.1092492491,0.0714813545,0.1706616879,0.2525339723,0.3444802463,-0.0604808405,0.0776817501,0.0565261282,0.1300673932,0.2442748398,0.0165426824,-0.2678135633,0.1944401711,-0.2378713042,-0.5962439775,0.1379382759,0.0044525987,-0.2270783484,0.2211285383,-0.1197347939,0.1918445379,-0.1130478084,0.0704182684,0.0728209391,-0.0637513101,0.0359921232,-0.1966178864,-0.130409345,-0.121753715,-0.1547304839,0.0039050996,0.2122249156,0.0312975273,-0.0897065103,0.3491774201,0.592879951,0.0300890785,-0.0807663575,0.1476536095,0.0750993863,-0.1631666124,0.2906976342,0.0195028614,0.6055518389,0.3023273647,-0.2451256514,0.1431187838,0.3871825635,-0.0319288336,0.2201086283,0.2057910115,0.6911458969,-0.1732664704,-0.2011304796,-0.0359998271,-0.0938760191,-0.0558814369,-0.2427799851,-0.3530630767,0.0092328265,-0.000906169,-0.2990610003,0.1023767292,-0.0631882399,-0.5435878634,0.2988537252,0.0494468585,-0.0285582934,0.3031966984,0.0058652665,0.0068820301,-0.0320099294,0.13250193,0.2278660834,0.29939273,0.1043857485,-0.1883507669,-0.0772017837,-0.2015258372,-0.113483794,0.0535620935,-0.0895324424,0.2411708236,0.0412974916,0.2565224171,0.4144940972,0.0141732339,0.2222730219,-0.0423172601,0.0757144839,0.0019608631,0.1507309973,-0.2659513354,-0.4235904217,-0.040043436,0.0292924866,-0.1178217083,0.2622919977,0.1448417008,0.1711197942,-0.0769885629,-0.173371166,-0.2095347345,0.3818931878,-0.3229215145,-0.1560789943,-0.060211733,-0.1195016727,-0.0817985162,0.0512832291,0.2745808661,-0.0639762059,0.3549925387,0.4507524073,-0.0075816363,-0.1911910176,-0.4339693189,0.01850456,0.0506454892,0.2169718891,-0.0752496123,-0.0754859149,0.1266775429,-0.2356222123,-0.0945631713,-0.4045507908,-0.0439575203,-0.2115865946,-0.1797845066,-0.0180365704,-0.306456387,-0.0857300609,-0.4518187046,-0.1782670766,0.4111146629,-0.1131807864,0.4838880301,0.1294627488,-0.3423334062,0.0717174858,-0.0214707889,0.1171506643,-0.0848596394,-0.2901393175,0.1469908953,-0.5867846608,-0.4462214112,-0.1477696896,-0.0219206382,0.0593528524,0.1654686183,-0.6037142277,-0.0400836915,0.1073319316,0.2574415505,-0.0711598396,-0.3257621825,0.3278180063,0.0800223053,0.0438431203,-0.203488633,-0.1508628726,-0.2604581416,-0.0460256562,0.2202904224,0.3321332335,-0.296866715,0.5189398527,0.3058698773,0.0639551133,-0.1734406054,0.2994426787,0.2157346308,0.0945931599,0.3629154563,0.0330703966,0.1078675166,-0.0555293895,-0.0243010055,0.2675091028,0.0170696657,-0.4916903079,-0.2860345542,-0.451344341,-0.3015532196,-0.219483301,0.1638204008,0.1265518963,0.7068167329,0.0367454849,-0.2715640664,-0.2608793378,-0.5439243913,-0.0529007651,0.0158363394,0.4337777793,0.1420951784,-0.0288519058,-0.3666352332,-0.3698000312,-0.0949658304,0.1492105722,0.423633337,-0.281070739,0.3498114049,0.2100923061,0.0038342371,-0.1465034038,-0.1172953472,0.2059975713,0.0863503739,-0.1597217619,-0.0630377457,0.3592987359,-0.1413587332,0.3985812962,0.1767798811,0.3208022118,-0.154697597,-0.4560801685,0.1793266833,-0.1133945882,-0.1558441222,-0.1579609662,0.1927947104,0.0148718953,-0.0956526399,-0.0750457793,-0.1636987478,0.058543779,-0.2120420039,0.1596367955,0.1009083018,0.3026943505,0.2189251781,0.0021816676,-0.1707447767,0.2089253962,-0.0617238358,0.1306860149,0.4106842279,-0.216710642,0.0466802157,0.3789792955,0.1262377948,0.0297859404,0.0142840799,0.486010462,0.4820488989,0.1425084919,0.0966434479,-0.1211734638,-0.1833414137,-0.2815524042,-0.0723592117,0.0842825249,-0.3303744197,0.2198017091,-0.0357437395,0.339728713,-0.1757730395,-0.1982981563,0.0899436101,-0.0755008534,-0.1988654435,0.4128501117,0.272546351,1.0386303663,-0.137367025,0.3834441602,-0.0196912549,-0.1977071315,0.5154652596,-0.2702786028,0.0434942245,-0.1980787218,0.2364216894,-0.1128875986,-0.0440913849,0.1217052117,-0.0623143651,0.0357630327,0.1358496696,0.3239833713,-0.2967234552,0.0686847866,0.3357771933,0.3997140527,-0.2651272118,-0.3937948942,-0.0248376522,-0.2576268613,0.2632714808,0.0359754749,-0.651030004,-0.0908366814,0.2664673626,-0.2422221154,-0.0749068186,-0.2591437399,-0.0026981526,0.0872524902,-0.3631939888,0.5744145513,0.4063818753,0.2704287469,0.2025235146,-0.49369964,-0.320350498,-0.243557632,-0.3420480788,0.1934927553,0.1921103299,0.3104508221,-0.2376258075,-0.1474412531,0.0690200254,0.0732318237,-0.5395254493,-0.2327624559,-0.0456406511,0.4490931928,-0.1730309874,-0.0285154507,0.1894908547,-0.388124913,0.0708843321,0.0714194551,0.3598955572,-0.1813544631,0.0008804844,0.2582433224,-0.0391001254,0.1079902202,0.2064123899,0.1971379668,0.047597114,0.2663504183,-0.1916865706,0.0337111317,-0.0851709992,0.307285279,-0.1627016068,-0.1438990533,-0.242314145,0.2242262512,-0.2233062238,-0.1870152652,0.3021792173,-0.1381717622,-0.1680936962,-0.1998982877,-0.5592111945,-0.2277266532,0.2929150164,-0.161445111,-0.0962273851,-0.0182938129,0.6015933752,0.0106297191,0.2008111328,-0.2129144371,0.2726168334,-0.2326035947,0.25893417,0.2685910761,-0.1004848108,-0.2009237111,0.1522371918,-0.0964115709,0.1559232026,-0.2094217688,-0.0753234625,-0.1471346468,0.1975392401,0.2788585126,-0.4151136279,-0.1919861287,-0.6648947597,-0.25729388,0.120734334,-0.0260887723,0.0700993761,0.0771594569,0.2283808291,-0.2003264874,0.2988661528,0.0192689896,0.2921557724,0.1198109388,0.4512789547,-0.1237811744,0.2936354876,-0.0524201468,-0.1377349198,-0.0534585454,-0.0698507652,0.2850040495,-0.0946616158,0.6047859192,-0.048358202,-0.1324421614,0.0857051387,0.0298936684,-0.0660922676,-0.1131201461,0.3623418808,0.1669359058,-0.0142751057,-0.182414934,-0.178582415,-0.0631525964,0.161721468,-0.1925421804,-0.0501026735,0.7015508413,0.2607524991,0.3034802377,-0.2565442324,-0.0276143793,-0.3427503407,-0.323777169,-0.0100466246,-0.2235169262,-0.0360938087,-0.1198541299,0.3481209874,0.0244554803,-0.0460344702,0.2468602806,0.3549416661,0.2359410375,0.006491161,0.225161925,0.0285884701,-0.2206171304,-0.0536102057,0.3036887944,0.2582814097,-0.3693703115,0.4333002865,0.0842205733,0.1218345091,0.3150321543,0.1767142117,0.0025231394,-0.2023657858,-0.1296456903,0.0232997425,0.0620394535,-0.2328781188,-0.2434094399,-0.1838622838,0.0827336386,0.2335701734,0.1629696041,-0.1065342724,-0.1186969578,0.4715518057,0.1982187182,-0.1951876134,0.4647734165,-0.304754585,0.1782439202,0.1193440855,0.2213312536,0.0246078204,0.0459150635,-0.5211941004,-0.0455410965,-0.4762412012,-0.0726612732,0.0163057577,0.3716422021,0.3945476413,-0.1800387204,0.3248778284,-0.1216431707,0.0831220448,0.0959899575,0.1107639745,0.0947462022,-0.1965453774,0.4842384458,0.0574740097,-0.413012445,0.0628565326,-0.1627618372,0.1354077458,0.1741879135,0.4179721773,-0.1049791723,0.0809338316,-0.1722981483,0.0453132652,-0.1430998445,0.2850532234,0.0829049274,0.5021238327,-0.0567174517,-0.2211393565,-0.2373366803,0.4116211236,0.3190974891,-0.0813912079,0.0870059729,0.1805799305,0.3197391927,0.5312252641,-0.2244132757,-0.3075534403,0.1514248699,-0.0981154516,0.0387726799,0.1747317314,-0.2453689277,0.2914696038,-0.0822882205,-0.1117826551,0.2605931461,-0.2142180353,-0.0640668795,-0.3192116618,0.1888645738,-0.18936041,0.3761972785,0.3925131261,0.0111439619,0.0873611867,-0.0546622314,-0.1511514932,-0.1009999886,-0.0073149479,-0.0337223448,-0.0754650757,0.192099154,-0.0729418918,-0.3759573996,0.0269919652,-0.1672229469,-0.1299835742,-0.3631918132,0.2177123874,0.3509697914,-0.1358592808,-0.3087919056,0.1354793012,0.0118541755,0.3943217695,-0.2195737362,-0.1936658621,-0.1383873522,0.0823545456,0.193824172,-0.5943899751,0.0494542159,0.1075957119,0.2410897315,0.0790658519,-0.0995782465,-0.396078676,-0.0185219087,0.1131387576,0.1150475368,-0.0218859483,-0.0809203312,-0.0165451188,-0.0930114537,-0.1034654304,0.4703731537,-0.0470546447,0.0389227048,-0.2427382469,-0.3359514773]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/279","title":"Dataset Preprocessing Cache with .map() function not working as expected","comments":"When you're processing a dataset with `.map`, it checks whether it has already done this computation using a hash based on the function and the input (using some fancy serialization with `dill`). If you found that it doesn't work as expected in some cases, let us know !\r\n\r\nGiven that, you can still force to re-process using `.map(my_func, load_from_cache_file=False)` if you want to.\r\n\r\nI am curious about the problem you have with splits. It makes me think about #160 that was an issue of version 0.1.0. What version of `nlp` are you running ? Could you give me more details ?","body":"I've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!","comment_length":101,"text":"Dataset Preprocessing Cache with .map() function not working as expected\nI've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!\nWhen you're processing a dataset with `.map`, it checks whether it has already done this computation using a hash based on the function and the input (using some fancy serialization with `dill`). If you found that it doesn't work as expected in some cases, let us know !\r\n\r\nGiven that, you can still force to re-process using `.map(my_func, load_from_cache_file=False)` if you want to.\r\n\r\nI am curious about the problem you have with splits. It makes me think about #160 that was an issue of version 0.1.0. What version of `nlp` are you running ? Could you give me more details ?","embeddings":[0.0064097252,0.1274485886,-0.0176681671,0.0728261843,-0.0148931192,-0.0171939824,-0.0071921707,0.2866290808,0.2888216376,-0.0975390375,0.2722377777,0.4488555789,-0.0316165946,-0.0931173116,0.0500858091,0.1241735518,0.1228096411,0.300366044,-0.0334364139,0.1303222626,-0.3629049957,0.1479874551,-0.3015810549,0.1129778251,-0.3348096013,-0.2727018297,0.0172574595,0.095787093,-0.0188568141,-0.3256303072,0.2679966092,0.3375692368,-0.1044381335,0.3873212934,-0.000121221,-0.0419374481,0.4864032567,-0.0222951807,-0.3611887693,-0.1947378963,-0.2702814937,-0.2420765311,0.1916081309,-0.1074838191,-0.0039147022,0.1268712431,0.0983078554,-0.8171732426,0.362642169,-0.1187922433,0.1225599945,0.0427530892,-0.3778812885,0.3451079428,0.1409942657,0.2448941022,-0.1506149024,0.0156823527,0.249256894,-0.378675431,-0.1975915432,0.2599431574,-0.1872857362,0.0981651619,0.1893691868,-0.0960441977,-0.1506905556,-0.5925680399,0.2482265532,0.0938565284,0.3833966553,-0.222224772,-0.2484366894,-0.3732969761,-0.2066570669,-0.2435506284,0.4230378568,0.039680317,0.000003425,0.121342808,-0.8589971662,-0.241796568,0.1990735531,0.0515707023,0.3050819933,0.1497192979,-0.1075390428,0.1488328874,0.2520419359,-0.0432583205,0.1955893636,-0.3810196817,-0.1168789789,0.54404217,-0.2710390985,-0.0157404207,-0.1196785495,0.327714771,0.0930754244,0.080923304,0.1217122674,0.3118827641,-0.1256469637,0.1597678363,0.4004644156,0.413469404,0.2639719248,0.1381666809,0.1329961568,-0.2675252259,-0.1418130547,0.0852220133,0.2552706003,-0.2441579551,0.0248955078,0.1836040914,-0.0821266174,-0.2886599004,0.1652615219,-0.204752937,-0.3881452084,-0.3014133275,0.0466032811,-0.0346511304,0.0642717779,0.3691862822,-0.0679801404,0.0237702783,-0.3667443693,-0.2312962413,-0.0323533528,-0.1859189123,-0.450309962,0.2917486429,0.4802574813,-0.022991458,0.1479587853,0.0876553878,0.0076955394,-0.2525999546,0.3581242263,-0.3446438015,0.5570878983,0.0797214583,-0.1528120488,0.2126494944,0.0044821785,0.2268762887,-0.1894952804,0.2410020977,-0.358746022,-0.1022452414,0.4592746496,-0.0000665783,-0.2909381986,0.2537983656,-0.0992955118,0.3521670997,0.4363529384,-0.4755979478,-0.0124220718,0.0694461614,-0.4158191383,-0.2272676975,0.2569476962,0.4410957098,-0.1938149333,-0.1198772788,-0.214117378,0.0179559756,0.3563316166,0.1271065772,-0.1938841641,0.2086488605,-0.1683626175,0.290021956,0.3815331161,0.1351912469,-0.4795579016,0.4713776708,0.2284037471,0.227691859,-0.3854047954,0.0375707448,0.0791010335,-0.1269278824,-0.1764232665,0.3006061614,0.0627405345,0.0413747616,-0.275734514,-0.3874903023,0.3989247084,-0.1624772996,-0.0575778857,0.4469153583,-0.131718412,-0.2132024318,0.0778076947,-0.0879961699,0.266974926,0.0754674599,-0.1761226952,-0.128610909,0.0832729638,-0.1762315035,-0.4962344468,0.2724436522,-0.3240979314,-0.2273713648,0.1025934666,0.0042819437,0.1540934443,-0.1819059998,-0.1151033863,-0.1337401122,0.0199798215,-0.0929003432,0.2611244619,-0.1285678893,-0.1667251736,0.4882050157,0.2561483681,-0.1032717302,-0.2471446991,0.0003949931,0.0874283835,-0.1719672233,-0.2900414765,0.1959816813,0.2051539272,-0.0753096864,-0.191265747,0.4112710953,0.2673026323,0.2019968927,0.003380531,0.0145018063,0.1268532425,-0.0009968209,0.0589546487,-0.0359402336,-0.0375745222,-0.1615540087,-0.0593969822,0.3376890421,-0.1741296798,0.3409543037,0.0023423238,-0.0108288974,0.1117193103,-0.1357339919,-0.028474791,-0.3059627414,0.0659516603,-0.0474209897,0.4817927182,0.0373971984,0.0783253908,-0.034675546,0.2912854254,0.1321613342,0.0123990774,-0.0403266773,0.0076511316,-0.276273191,0.1063793004,0.2353534997,0.3839116096,0.0215854365,0.3465099335,-0.0069529675,-0.0878050029,-0.0466011353,-0.0562839583,0.0393119156,0.3893088996,-0.0059825908,0.0933693498,0.0178924724,-0.158398509,0.1463744044,0.1444567293,-0.1171731576,-0.0746455193,0.2454182953,-0.4893401861,-0.2945530713,0.0313844234,-0.0402236879,-0.4356859624,-0.2097008228,0.0017259953,0.1465749294,0.0630527362,0.280010432,-0.0101847155,0.3138221204,-0.3168032765,-0.1929496378,-0.1992981136,-0.3507534862,-0.245791629,-0.0110542495,0.1349146664,-0.0263965428,0.2472252548,0.0327347815,-0.1674441844,-0.0104458155,-0.0082206028,0.2668375075,0.1025384367,-0.0167490169,-0.2774066925,-0.160661906,-0.2169780582,0.0876844972,-0.0550031848,-0.4442885518,-0.1832215935,-0.0269125402,0.0297259949,0.0075627044,-0.3920730054,-0.0672996789,-0.0958922505,-0.0754189864,0.0948507935,-0.2142595649,0.203998968,0.0861299485,-0.0828519836,-0.0784939378,-0.0981289595,-0.2145887315,-0.3659687042,-0.2436297536,0.1614861339,-0.1592233032,-0.1678786278,0.0809421018,0.0509378798,0.1613148153,0.7081987858,-0.2456416637,0.0816381723,-0.0514448993,-0.1534805149,0.0792530626,-0.0001609416,0.4673969746,0.1011039019,0.1088512242,-0.2240369171,-0.3684271276,0.2269667387,0.1711654514,0.2487744391,0.4404617846,0.0246732384,0.1091503501,0.8266453147,0.4055009782,-0.1263312101,0.2748199999,0.260612607,0.2630581856,-0.251149565,-0.0069565796,-0.1072506234,-0.417815119,-0.0213203281,0.2703511119,0.014496115,-0.7684549093,-0.2149749249,0.3027918339,-0.3229091465,-0.2363556027,0.0526559688,-0.5057756305,0.5118843913,0.1472564936,0.0512242727,-0.2993236184,-0.2363729328,-0.1283421516,-0.1672179252,0.7289863825,0.0014056746,-0.4842478037,-0.0221381877,-0.3337204456,0.2105125189,0.134055689,0.5019213557,-0.3132720888,-0.0918582827,0.0779155791,0.1511161327,0.4856851697,-0.4133760035,-0.0130138341,0.0894729719,-0.1434801072,-0.0367490835,0.1446467638,0.2511833906,0.6791267991,0.3776953518,0.1472017318,-0.0270599518,-0.2170239836,-0.3850435019,0.0746969134,-0.0957727805,-0.355293721,-0.0739812106,0.2762762904,-0.2644429803,-0.2000414282,-0.0776650161,-0.2506612539,-0.1604654789,-0.0336911716,-0.0417762324,0.0736012533,0.177468136,0.2350426465,0.3998059332,-0.0228430126,-0.1672782004,0.5847776532,0.2105851918,-0.3136142492,0.434009105,-0.2198306024,0.2310994565,0.2186964005,-0.0291400049,0.2093359232,0.4655958414,-0.0588965714,-0.1210478917,0.0213386212,-0.2066892385,-0.4833388627,-0.0316990912,0.1694447547,-0.1293512285,-0.3581573665,-0.3836258352,0.3327217996,0.3341460228,-0.3819208145,0.3549835086,-0.2618377507,-0.3019738793,0.3253311217,0.0239499453,1.0048357248,-0.1948838234,-0.124831751,-0.4032354653,0.1095948666,0.4291089475,-0.3195338547,0.3689870834,-0.244198814,-0.0008735668,-0.0588079207,-0.1247972921,0.2702732086,0.74614048,-0.1136487126,0.371930927,0.1079635769,0.0601338297,-0.0092218053,0.1000101939,0.2591951787,-0.384386152,-0.0235667266,0.0723723099,-0.1698031276,0.066347681,0.0669573322,-0.0424921103,0.0070673097,0.1354987323,-0.1099957526,0.1884273291,-0.1476852596,0.0728608295,-0.1071482599,-0.0396831185,0.3360202014,0.149270013,0.188094303,-0.0078011062,-0.2221501172,-0.0517125018,0.3180361986,0.2039188892,-0.1083686575,0.1831480712,-0.0406702235,0.0020897132,-0.1245263442,-0.2530969679,-0.2252915204,-0.3147900105,-0.0537824742,0.0620624274,0.1692891717,-0.1764383614,-0.1892218739,0.2640277743,0.1497406811,0.1316511333,0.0689578876,0.1283252239,0.1121489182,0.3821691573,-0.0173573922,-0.187462002,0.0064482288,0.2096821666,-0.007049968,-0.1948873401,0.1713548601,-0.0716789067,-0.2365669757,-0.123602502,-0.2610141635,-0.2897469401,-0.2461136729,-0.0908444002,-0.0871591344,-0.086447306,0.0302132126,-0.2308715135,-0.05317178,0.1910160333,-0.197961852,-0.2889900208,-0.1136422455,-0.0676406771,-0.2768877149,0.2176239789,0.132608369,-0.1775802374,0.1786129177,0.5122457147,-0.1615325958,-0.0735867545,0.1160416305,0.30809322,0.1839488596,0.0873081759,-0.1668291837,-0.2434577644,-0.2275370955,0.0758873597,-0.3767928183,-0.1049954593,0.0864345506,0.2187223136,0.0256573688,-0.4668272436,-0.0737527534,-0.1002885848,0.078784503,-0.4276940227,-0.0428023785,-0.083055079,-0.0892085582,0.442238301,0.308439225,0.0207795277,-0.2366683632,0.1384489834,-0.065609619,0.1551499218,0.0468529575,0.1583068371,0.2034460157,0.070876278,-0.2576165795,-0.1523811221,0.063849166,-0.0055509303,0.2288091034,-0.201515466,0.2854418159,0.0784549266,0.1884370148,0.2013293058,0.048989892,-0.1246558204,0.2670288086,0.0231149551,-0.0306590628,-0.2331513762,0.8174630404,0.2363247871,-0.0470781289,0.2947334945,0.3611822426,0.1418642402,0.2163415253,0.0106498487,0.3291228116,-0.491774559,0.2105628252,0.2103260309,0.1886697114,0.1961784065,0.2223790586,0.1365744472,0.0211851168,0.1992658526,0.4374975264,0.2988744974,0.3040632606,0.4697696865,-0.0494114943,-0.5832751393,0.0528543256,0.4703790545,-0.0583236963,0.3675434291,0.0255619604,0.2290538847,-0.1327771097,-0.4096173942,0.0087854881,0.1852755398,-0.0325410515,-0.1582809985,0.0817119479,-0.0903602391,0.039803803,-0.0113361655,-0.1455932558,0.1309079528,0.82354182,0.0522282533,-0.0294989329,-0.08354339,-0.0587167069,-0.0552506782,0.3492958844,-0.0578359812,0.1647347212,-0.4313991368,0.084247537,0.0666564554,0.3254653811,0.0088739572,0.3660203516,-0.3733350635,-0.0598591007,0.0430035293,0.0146141918,-0.2489277422,0.5730592012,0.2041498274,-0.2849006951,0.1977986544,-0.057655625,0.141520083,-0.088421084,0.1105032042,0.4440021515,-0.0085370326,0.3120677769,0.2348242402,-0.1558400393,0.2259768099,0.0543211624,-0.2092333883,0.1087222025,0.3301640451,-0.204140529,0.1764392853,-0.0717589483,0.0158242155,0.0650586784,0.4250699282,0.288059622,0.0499758907,-0.2027334273,-0.2592552602,-0.49176085,-0.0618209802,-0.1175158545,0.1678069979,0.046928972,-0.0659302548,-0.0730159804,0.3085063994,-0.1196215227,-0.3544443846,0.0894507468,0.0075958823,-0.3419650197,0.0648221299,-0.2347061783,0.1187276021,-0.1004283726,-0.3049357831,0.1247803345,0.1861409396,-0.1550418437,0.0788923278,-0.2164865583,-0.0080199055,0.4857319891,0.2058416158,0.5476559401,0.1450551301,0.0631096885,0.0693128482,-0.3424026966,0.034314312,-0.0426799878,-0.0744053423,0.0526133291,0.2084495127,-0.5877813697,0.2476355135,-0.2361845672,-0.1315245032,-0.1014916748,0.1991197914,-0.2181738615,-0.2774003446,-0.047554303,0.142956987,0.2431960255,0.5879483819,-0.0743660405,0.1885570884,-0.4346538782,0.1965996176,0.1331104636,-0.4808310568,-0.3672071397,0.171833694,0.1114938855,-0.0621825121,-0.4134124815,-0.3246700466,-0.3102245927,0.2567914426,-0.3081018627,0.0306069087,0.0750459731,-0.3588088453,-0.0928641334,-0.1293269992,0.3822349608,0.254710108,-0.2054184377,-0.0491093062,-0.2069127709]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/279","title":"Dataset Preprocessing Cache with .map() function not working as expected","comments":"Thanks, that's helpful! I was running 0.1.0, but since upgraded to 0.2.1. I can't reproduce the issue anymore as I've cleared the cache & everything now seems to be running fine since the upgrade. I've added some checks to my code, so if I do encounter it again I will reopen this issue.","body":"I've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!","comment_length":53,"text":"Dataset Preprocessing Cache with .map() function not working as expected\nI've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!\nThanks, that's helpful! I was running 0.1.0, but since upgraded to 0.2.1. I can't reproduce the issue anymore as I've cleared the cache & everything now seems to be running fine since the upgrade. I've added some checks to my code, so if I do encounter it again I will reopen this issue.","embeddings":[-0.0595274158,0.0904483497,-0.0686693415,0.1003901288,0.1207222939,0.0799236521,0.0432367027,0.2624785304,0.285457164,0.0177971534,0.2515242994,0.3141241372,0.1379653215,-0.0756428763,-0.123119697,0.1792807281,0.1890160441,0.1980445683,-0.1283153892,0.2175145,-0.3310685158,0.0507617407,-0.2609412372,0.0022761859,-0.2656695247,-0.1827593297,0.1562230736,0.0252245814,-0.0041876552,-0.2667355835,0.2764956653,0.3308138251,-0.1584998816,0.5676491261,-0.0001187327,0.1042143404,0.4598644674,-0.0732590184,-0.2360089272,0.0055008717,-0.3911941946,-0.0423367731,0.1526716501,-0.039945066,-0.0345012173,0.1100187749,-0.0018390432,-0.9861100316,0.3587905765,-0.0971710533,0.1350430846,-0.0097585646,-0.3999705017,0.2026426941,0.06718833,0.2150055766,-0.1796841919,0.0030361803,0.4216757417,-0.1683782637,-0.1409471482,0.183951214,-0.2047435343,0.2275711745,0.2218047231,-0.1214689985,-0.0648612306,-0.6551277041,0.454888165,0.0899076834,0.4829883575,-0.2404881865,-0.3372349441,-0.1862241626,-0.2037635595,-0.1001643687,0.4849649668,-0.0276624057,0.0601207502,0.2609480023,-0.8676720262,-0.2269795835,0.1066075787,-0.1002234668,0.1879886836,-0.0380950607,-0.1921433806,0.1235531867,0.1384325773,-0.0326317213,0.3090585768,-0.4443030953,-0.1275442988,0.4259271026,-0.2078385502,0.109044455,-0.1491014957,0.2548020184,0.1343227327,0.0548303127,0.1186302081,0.4010969698,-0.055815693,0.1257498413,0.270812422,0.3804577291,0.1999516934,0.0539859459,0.2071965635,-0.2867138386,-0.1800409406,0.049800355,0.2067424804,-0.2862972617,0.1496955901,0.0931982175,0.0283403564,-0.3782491982,0.1323273927,-0.1684764475,-0.3022691011,-0.2636650503,-0.0303665139,0.07098867,0.0451824702,0.3284302056,-0.1247797236,-0.0048172628,-0.2795673311,-0.1862279475,-0.0313262716,-0.2699474096,-0.303024143,0.3095788956,0.3854755759,-0.1776116937,0.1134093404,0.2070455551,0.11043524,-0.3046349287,0.409712255,-0.3524376452,0.6011015177,0.1649867445,-0.177268818,0.2531388998,0.0884696841,0.3341836631,-0.0959151089,0.3465947211,-0.4138706326,-0.0795571208,0.5190355778,0.0128131751,-0.2568217814,0.2969447076,-0.1490399241,0.206664592,0.385432303,-0.3386059403,-0.0406747572,0.07820566,-0.3557200134,-0.2343351543,0.318288058,0.4376455545,-0.3451401889,-0.0244084932,-0.045419123,0.030383978,0.3146098852,-0.0508987159,-0.1541089267,0.2512764037,-0.2133024037,0.0842325315,0.1131338477,0.0664180666,-0.4564329386,0.4595998228,0.2408317775,0.3241764903,-0.3276294768,-0.0310352035,0.0604769103,-0.0584328882,-0.2508625388,0.0946025327,0.0931590348,0.0463019311,-0.1457220465,-0.3507972062,0.2563435137,-0.209526062,0.0044859131,0.5632539392,-0.1481964439,-0.4645827711,0.0373644941,-0.1506569535,0.2761831582,0.1701615602,-0.1055879295,-0.0880764499,0.0879942179,-0.1398374438,-0.3425427377,0.3794329762,-0.1424757391,-0.3455411196,0.0506214835,-0.0093509275,0.1030253619,-0.158650741,-0.2077572942,-0.0500634611,0.0235334542,-0.0339243934,0.165765211,-0.2036354542,-0.1153057814,0.5470615029,0.1891899556,-0.0793976784,-0.1510235965,-0.0350825824,0.2097179592,-0.1502480358,-0.3731407225,-0.0245712269,0.1470717639,-0.1356713027,-0.1894614398,0.4257082343,0.3730862439,0.2214649618,-0.0048203338,0.1351879835,0.0575566739,0.0619351603,0.0366932116,-0.0725549981,0.0498205759,-0.159109652,-0.0514163189,0.3344888389,-0.0911414027,0.3732474148,-0.0100102108,-0.0544526912,0.149504289,-0.0102278274,-0.0336267054,-0.277566582,-0.1447202712,-0.0774002969,0.5285404325,0.0067476733,0.0939818323,-0.0636564046,0.1063500941,0.1443342417,-0.009743209,-0.1614425778,-0.0838730782,-0.1745381504,0.083735764,0.1592413783,0.3290292621,-0.0062201428,0.3409837782,-0.0573556237,0.0048366413,-0.0275136847,-0.093927525,0.0857732818,0.4193117023,-0.0040599136,0.0293809436,-0.0115951002,-0.0849877372,0.1703855097,0.1218730286,-0.2071109861,-0.1827341914,0.209192574,-0.4168634117,-0.0502886176,0.130165562,-0.0189147964,-0.2983778715,-0.1598962545,-0.0456071831,0.1653749794,0.0913636088,0.2640117407,-0.0585686825,0.2686870396,-0.1954088211,-0.34221071,-0.1986907721,-0.2056871057,-0.0732575804,0.007954157,0.1256839037,-0.1541635692,0.2510619462,-0.0887085497,-0.1257594973,-0.1457536668,0.0405122563,0.2885510623,0.0854327381,0.0356912464,-0.3706345856,-0.0089761941,-0.2044826597,0.2617307007,-0.0208569523,-0.4014315009,-0.1830391884,-0.012171261,0.0049402998,-0.0473406166,-0.300817132,-0.0567367896,0.082481496,-0.0315570012,0.0004642938,-0.2242860943,0.0952961445,-0.0120264934,0.0699421763,-0.0765579119,-0.0082293339,-0.1531716287,-0.3678612411,-0.4966880381,0.2230886668,-0.186309278,-0.1567218006,0.2615591884,0.2169646621,0.1964536756,0.7846165895,-0.2175372392,0.019605102,-0.1301593632,-0.0638763309,0.0554211289,-0.0219341982,0.4251804054,0.1836566776,0.0809685066,-0.2568446398,-0.3909493387,0.1640685201,0.0993056372,0.0962958336,0.4736777842,0.0479007438,-0.0114189768,0.9212653637,0.2891286612,-0.0608736873,0.2333852798,0.1780751795,0.3517637849,-0.3186762631,0.060920056,-0.1786854863,-0.487529546,-0.1493476629,0.1119745597,0.0378801338,-0.54543221,-0.2041718364,0.3990443349,-0.2837381363,-0.1479299366,-0.0620149225,-0.4005838037,0.5391529799,0.0512313358,0.1854630411,-0.2467219085,-0.1070202067,0.0533123724,-0.1604481339,0.7593111396,-0.0761718825,-0.4067287147,0.027426349,-0.3406668603,0.3155036271,0.109916307,0.3678984642,-0.2778316736,-0.1029741019,0.0640021563,0.1480658352,0.4910818338,-0.4214988947,0.0462923199,0.0597036816,-0.3347888887,0.0465977117,0.0716415867,0.2156669647,0.6267978549,0.1488234997,0.0873892605,0.0531236082,-0.1913238913,-0.5050851703,0.1265778691,-0.1179731563,-0.3258690536,-0.1146571711,0.2298448384,-0.2708108723,-0.1699538678,-0.0046588541,-0.266330719,-0.1930045038,-0.1560005993,-0.1828419715,0.0982100442,0.1806223989,0.2443591952,0.4060881734,-0.0617608465,-0.0840655714,0.6687200069,0.179963693,0.0247224756,0.4776473641,-0.1866742671,0.0995090827,0.2201077044,-0.1249553561,0.1327976286,0.4130031466,-0.2108190507,-0.2021465003,0.0196395051,-0.2465939373,-0.5624935031,0.0012003819,0.1807606965,-0.0328983925,-0.3743573129,-0.4673776031,0.222466737,0.3381887376,-0.4176740944,0.4210338593,-0.1740472019,-0.2860767543,0.2021721005,0.0786116943,0.8869552612,-0.3360162079,-0.2119333297,-0.4440293014,0.1028308123,0.0187481996,-0.2972941101,0.322561413,-0.1631958783,-0.2073845267,-0.1289112419,-0.1997932494,0.378428638,0.6281092167,-0.1103118807,0.3274941146,-0.0922457725,0.2576876879,-0.0604381748,-0.0698908865,0.2925606072,-0.3084753156,0.0903789327,0.1228207201,-0.043243397,-0.0639645755,0.0824610591,0.0709595606,-0.0437480621,0.1999803483,0.0007836522,0.1129752249,-0.3217524588,0.0409507155,-0.2048680782,0.0339203365,0.3039815426,0.213608101,0.2289962023,0.0165269636,-0.2403589338,0.0455541238,0.3730081618,0.2545779049,-0.1677660793,0.1442346573,-0.1464838833,0.0663699359,-0.1275036335,-0.1230001152,-0.2934537828,-0.289478302,-0.0214957539,0.0044730864,0.1485633254,-0.2627685666,-0.1725668162,0.1919095218,0.2542985976,0.1264149398,0.0789371356,0.1525603384,0.1094983146,0.4262723625,-0.0785970092,-0.2673990726,-0.0245011728,0.2424902171,-0.0646107197,-0.2331254929,0.120426923,-0.1018592715,-0.2137970626,-0.1323664337,-0.1807457358,-0.1752254963,-0.1102776751,-0.0490811355,-0.1408293247,-0.087481916,0.0110823447,-0.2049196213,-0.0157990549,0.0756649673,-0.1835899204,-0.2050801963,-0.18387568,-0.1153943911,-0.1625483185,0.3037607074,-0.0263226312,-0.3170805275,0.1527165622,0.3556807041,-0.1814111322,-0.1099747196,0.2332231998,0.1967246085,0.0409177616,0.2854223549,-0.0525153577,-0.2287794501,-0.1743582636,0.045916345,-0.3606043756,-0.0996347889,0.1999198049,0.2232604474,-0.1286651045,-0.4671007991,0.0114449561,-0.1160565987,-0.0024374819,-0.4780494869,0.1506440341,0.1062405035,-0.0600690544,0.2874806523,0.3222017884,0.0086491844,-0.26315552,0.0568574145,-0.0380975716,0.1541025788,0.0113711869,0.0845213234,0.1962245256,0.0202192049,-0.0057640774,-0.0955257416,0.1170663163,-0.0202945247,0.2922273576,-0.2207029611,0.2975327969,0.2276818454,0.2156784385,0.2388552874,-0.0729439929,-0.0508699045,0.1377165765,0.0935816094,0.0310327727,-0.2592692375,0.7433461547,0.2427433878,-0.0444670431,0.3320808709,0.3685330749,0.1457959116,0.2659347057,0.0639651269,0.3084392548,-0.451889962,0.223938033,0.2170553952,0.1885970384,0.2354581505,0.3086994886,0.2615504861,-0.1081755906,0.3223491013,0.4323149621,0.3706761003,0.3667143285,0.4741810262,-0.1612984836,-0.6234046817,0.1654551923,0.377599597,-0.2864930928,0.2705142498,0.0158286318,0.0046207169,-0.0707075149,-0.4744415581,-0.1246583015,0.020509785,-0.058468204,-0.1547642648,0.2466968149,-0.1890918165,0.1996325999,-0.0538561977,-0.0013489571,0.097172156,0.8700405955,-0.0679967403,0.0033369202,-0.0345676653,-0.0846283138,-0.1373525411,0.3553052545,-0.0189707,0.0931442156,-0.4131549299,0.0328657366,0.150625214,0.2757976353,-0.0293727703,0.3324168324,-0.3319463432,-0.0078491094,0.0610645637,-0.0090504987,-0.2035997212,0.5143904686,0.0165024009,-0.2763305306,0.1810613573,-0.0044946671,0.1436165124,-0.1180866137,-0.0925453305,0.5848078132,-0.1371866912,0.1569663286,0.167405948,-0.1190475449,0.2598944604,-0.011829189,-0.082967259,0.0437074415,0.4165822268,-0.2856684625,0.1008843854,0.0432971604,0.0310411714,0.0695139617,0.3478196561,0.3020911515,-0.0002089784,-0.2623051703,-0.2043788284,-0.4773701429,-0.0119126346,-0.04181692,0.2782687545,0.0137273353,-0.1041591614,0.0244097505,0.2227221578,0.1266430616,-0.6021293998,0.1470221877,0.0990480408,-0.3059791028,0.0324513651,-0.1631774902,0.0956097245,-0.0724131167,-0.3027044237,0.2240176797,0.2514533401,-0.0895237923,0.0851065591,-0.1930757016,0.0079133147,0.2854198515,0.4247600734,0.431709677,0.0910286605,0.0702404901,0.129233852,-0.2891432345,0.0006375942,-0.0060301698,-0.0338432305,0.0000886461,0.1067610309,-0.547657609,0.0628275946,-0.2475527227,-0.1833358556,0.0174734667,0.2171714902,-0.3235859871,-0.2488898486,-0.0122499634,0.1204671189,0.2119780332,0.4282517135,-0.0275552664,0.2745811939,-0.3904275,0.2230901867,0.1394861639,-0.414837271,-0.4747494459,0.0775707811,0.0630936697,-0.1698837876,-0.4320859313,-0.2826708257,-0.1736196876,0.2840861082,-0.3131862879,-0.0337035209,0.0827469379,-0.485971719,-0.0049079801,-0.1403232217,0.4262045324,0.1982597709,-0.0894490629,-0.1782263219,-0.1588245481]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/279","title":"Dataset Preprocessing Cache with .map() function not working as expected","comments":"Just checking in, the cache sometimes still does not work when I make changes in my processing function in version `1.2.1`. The changes made to my data processing function only propagate to the dataset when I use `load_from_cache_file=False` or clear the cache. Is this a system-specific issue?","body":"I've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!","comment_length":47,"text":"Dataset Preprocessing Cache with .map() function not working as expected\nI've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!\nJust checking in, the cache sometimes still does not work when I make changes in my processing function in version `1.2.1`. The changes made to my data processing function only propagate to the dataset when I use `load_from_cache_file=False` or clear the cache. Is this a system-specific issue?","embeddings":[-0.0422306918,0.0891577378,-0.0817682147,0.0919544026,0.0423214361,0.0622354262,0.0255286153,0.2098047435,0.3237959445,0.0284722876,0.3000636697,0.2643457651,0.1776193827,-0.0800424367,-0.1128168851,0.1895933896,0.1895669103,0.2083784789,-0.0732567832,0.1837684959,-0.3083145916,0.0548591986,-0.2294518799,0.0391611382,-0.2585252821,-0.2065357119,0.1432308853,0.0060473899,0.0013979918,-0.2522467971,0.2973943949,0.33280918,-0.128296718,0.5881773829,-0.0001169303,0.0842275843,0.4561638832,-0.0505358204,-0.2349918932,-0.0099979267,-0.3796210885,-0.0702175722,0.1602348387,-0.0243576802,-0.0101384278,0.1162179038,0.0201478228,-0.9627584219,0.3745887876,-0.0856605768,0.148074165,0.0059180004,-0.369587034,0.1530929208,0.0492118634,0.2058280706,-0.1800124645,0.0365164429,0.4169890285,-0.1854203939,-0.139328897,0.1435709447,-0.1768898666,0.2246657759,0.224708274,-0.0970955789,-0.0745331347,-0.6531959772,0.4681518674,0.0929749683,0.502574563,-0.2382342964,-0.3564358354,-0.1886161566,-0.1998536289,-0.1284057647,0.4736365676,-0.0521153845,0.0480240099,0.2472449988,-0.8149891496,-0.2441741079,0.139962092,-0.0658177733,0.1589182317,-0.0606695451,-0.2202188224,0.1006012708,0.1753266901,-0.0335370228,0.2819596827,-0.4145128727,-0.0930637792,0.4279217422,-0.2198819518,0.0837464333,-0.12302766,0.2560547292,0.0892047137,-0.006210912,0.1486397833,0.4178132117,-0.0385146625,0.1505216211,0.2692398429,0.3919083476,0.1836466938,0.0295404531,0.2262314856,-0.3068936467,-0.181799382,0.0224110987,0.2195631564,-0.3019469082,0.1666239947,0.0519048311,-0.0434889719,-0.3644348383,0.147913903,-0.1613026857,-0.2701402009,-0.268709451,-0.0240079425,0.0419540741,0.0624130517,0.3686458468,-0.1567596197,-0.0583760887,-0.2892080545,-0.1535946876,-0.0568067878,-0.2732104361,-0.3307188153,0.3337659836,0.372641027,-0.1995537877,0.0967079625,0.2127576619,0.1194793731,-0.2556492984,0.402323097,-0.3718317747,0.6122225523,0.1673039049,-0.1996267587,0.313372463,0.0886345878,0.3282322884,-0.1233412698,0.319057107,-0.4271486998,-0.0732033551,0.5425621271,0.0426246487,-0.2365996838,0.272485286,-0.1576716304,0.1831423044,0.3819362223,-0.3411971927,-0.0193041209,0.0559044182,-0.3614040315,-0.2492461205,0.3011446893,0.4044575989,-0.313049078,-0.0062669669,-0.0966528431,0.0089608803,0.2724154294,0.0020196759,-0.1514182985,0.2569433451,-0.1890785396,0.0740880445,0.179725647,0.0462378711,-0.4824351668,0.4619077444,0.2666216493,0.3159234524,-0.3200908899,-0.0344842374,0.0754926205,-0.0292528663,-0.2274345905,0.0733203217,0.1126248091,0.0184201524,-0.1365385056,-0.3354551494,0.2387078404,-0.172780931,0.0098830936,0.5577960014,-0.1166372299,-0.453373909,-0.0055344477,-0.1171777472,0.2688828409,0.1867070496,-0.1241810918,-0.0835297853,0.112226367,-0.117894195,-0.2957796454,0.3868889809,-0.1607989222,-0.3888332546,0.0870992765,-0.0017753701,0.1489263773,-0.1508038789,-0.170879975,-0.0603685081,0.0341724642,-0.0398584269,0.2085989416,-0.2214135379,-0.1505843848,0.5996130109,0.2288369387,-0.111903742,-0.1441234946,-0.0906638652,0.1879694164,-0.1929447204,-0.3765650988,-0.0420685671,0.1240160316,-0.0717114881,-0.18283692,0.4112227261,0.364466995,0.2317550927,-0.0571934879,0.1484310776,0.0776228756,0.1053875834,0.0395125709,-0.0438413434,0.0481653288,-0.1259491593,-0.0457108915,0.2844944298,-0.0756516084,0.3332164288,0.025585454,-0.0784998611,0.1512382478,-0.0068959491,-0.032718908,-0.2883858979,-0.1371816546,-0.0856276006,0.5550134778,0.0193321556,0.0998920947,-0.0798730627,0.1644963175,0.1588204652,-0.0057599596,-0.1098353043,-0.0765024945,-0.222566992,0.0785853863,0.1421719342,0.3094837964,-0.003141151,0.3505507708,-0.0958959609,-0.0081953593,-0.0534472391,-0.1273302138,0.0642364025,0.3782822192,-0.0128259202,0.044713188,0.0554708354,-0.058836624,0.1394155622,0.0954171792,-0.2367599607,-0.1391814053,0.2114250213,-0.4382856786,-0.1031098217,0.1216110215,-0.0038585875,-0.2828514278,-0.1596031338,-0.0491864756,0.1624259353,0.077183269,0.2626119256,-0.0405067727,0.3092963398,-0.241481334,-0.3057519197,-0.2155146748,-0.1971511841,-0.1173083335,0.0255007688,0.0658701509,-0.1796990633,0.2166001499,-0.0736564919,-0.1284490526,-0.1553097069,0.0631600767,0.2894631922,0.1655323207,0.0497437865,-0.38831231,-0.0344401598,-0.2452546954,0.2555259466,0.0039207712,-0.434815973,-0.1521812528,-0.0571384616,0.0033029579,-0.0713680834,-0.3041388094,0.0055960943,0.0284589883,-0.0661439821,0.0078730276,-0.2158997953,0.0964919031,-0.0588539615,0.0630993769,-0.0804712921,0.0026187764,-0.1112724915,-0.3895037472,-0.5156039596,0.233699441,-0.2174013555,-0.1643688083,0.2138955593,0.2171540558,0.2134817839,0.8205273747,-0.1731255203,0.0194019359,-0.131186828,-0.052393306,0.0817247108,0.0411988609,0.4442794025,0.1278310716,0.046868775,-0.2690100372,-0.403957814,0.1687218249,0.097603336,0.1209678501,0.462066561,0.0261512864,-0.0262697395,0.9281291962,0.2807488441,-0.1106205508,0.2399621606,0.2195242196,0.3571163416,-0.2936383486,0.0774127245,-0.1929932088,-0.4637962878,-0.1529190838,0.1066411063,0.0307705253,-0.5778459311,-0.2362836897,0.3495679796,-0.2912435532,-0.1249611303,-0.0557608306,-0.3199297488,0.5669843554,0.0419913232,0.1650812477,-0.262395978,-0.091460906,0.065309301,-0.1480160058,0.7365278006,-0.0787315145,-0.3571311235,-0.0101976776,-0.3159860075,0.2897166908,0.0940380841,0.3959237337,-0.2671494782,-0.1186773032,0.0130470581,0.1042243093,0.4674056768,-0.4032789469,-0.0089583872,0.0771920234,-0.3594956994,0.0886775404,0.076360859,0.2463447154,0.6016659141,0.1338369548,0.0870831683,0.0809590742,-0.1844836026,-0.5479585528,0.1127626151,-0.1154177859,-0.3176874518,-0.093901068,0.23005867,-0.2388248295,-0.1784872264,-0.0259604901,-0.2603915036,-0.1423581392,-0.1217116714,-0.2099802941,0.0934630483,0.1732192487,0.2643479109,0.3928101659,-0.1045462191,-0.1035260931,0.6716063619,0.1652072221,0.0172619447,0.4291106164,-0.198196277,0.0951779634,0.2537599206,-0.0903434902,0.1081709713,0.4060548842,-0.2222407162,-0.1741388887,0.0773169547,-0.2306131274,-0.5374788046,-0.0068941307,0.1347159296,-0.0479667857,-0.3338080347,-0.4914812744,0.2252886146,0.3845279515,-0.4131149054,0.4236252904,-0.2135852724,-0.259031862,0.216243431,0.0664065406,0.8740595579,-0.3066954017,-0.1755008996,-0.4766534865,0.1166640446,0.0304037221,-0.3100541532,0.3236524761,-0.1557642221,-0.2019037753,-0.1310507506,-0.1979898065,0.377512306,0.6080793142,-0.0753356144,0.3647038043,-0.0378579497,0.2650955319,-0.0334518701,-0.036090713,0.2536982,-0.3418264389,0.0720515996,0.1362070888,0.0110294204,-0.0121827666,0.0611505397,0.0795126483,-0.0313272439,0.1683038026,-0.0111202635,0.0724824667,-0.3772874177,0.0573920906,-0.1906112283,0.0522163361,0.2531183064,0.2159038484,0.262776494,0.0325004086,-0.2449139357,0.0404223986,0.4043719769,0.2383799851,-0.1943042278,0.130157724,-0.0948349088,0.0416071787,-0.1057258993,-0.1037327349,-0.2594006658,-0.2851214409,-0.0310628507,0.016957026,0.1296373755,-0.1928440034,-0.1775474399,0.1650242656,0.2473141551,0.1264149547,0.0947827324,0.1902931482,0.1477006525,0.4358248413,-0.1344919652,-0.2479329854,-0.0683749169,0.2577914,-0.0767048001,-0.2712801397,0.0920242518,-0.1371069998,-0.2443401366,-0.1362231374,-0.2375420928,-0.2199594676,-0.0874551684,-0.0592236742,-0.1308646947,-0.1317467541,0.017549701,-0.2109548897,0.0080014709,0.083369717,-0.1773945689,-0.2219512314,-0.1304798722,-0.1015324444,-0.1600721776,0.2814302444,0.0138343498,-0.3935483694,0.1601390392,0.3114461601,-0.1936087757,-0.1269919276,0.2641711831,0.1755345166,0.0822361484,0.2998536229,-0.0627641305,-0.229340896,-0.1649951041,0.0461198352,-0.3588995636,-0.1267245412,0.1888725907,0.2124417871,-0.1174060479,-0.4723454714,0.028111415,-0.116847232,0.0473523699,-0.474732548,0.1565852612,0.1065794453,-0.0406275094,0.2895567417,0.2889657617,0.0402337313,-0.2808891833,0.0816791579,-0.0083992286,0.1487882882,0.0188261084,0.0819855258,0.1746985167,0.0304050297,-0.0159742534,-0.1004500911,0.0711596683,-0.0503383167,0.2550847828,-0.220710814,0.3024341166,0.2557480633,0.1917849481,0.2408600003,-0.06531398,-0.0652860701,0.191171065,0.1037215143,0.0645918995,-0.2489881516,0.6774760485,0.2197593153,-0.0522208102,0.3222691417,0.3933651149,0.1262727827,0.2538275421,0.0428762212,0.3401603699,-0.485670954,0.2219836414,0.1781566888,0.2305345386,0.2444677204,0.3148039579,0.3094749749,-0.1206241548,0.3351335526,0.4036638439,0.3411260545,0.4024530351,0.4882265031,-0.1643557101,-0.6150530577,0.2145674229,0.3988510072,-0.276207298,0.2673067451,-0.0315820016,-0.0624847747,-0.0675519109,-0.4185516834,-0.093900159,-0.0122019909,-0.0555926412,-0.1606380194,0.2629877925,-0.17795524,0.1947116852,-0.0509007908,0.0031111506,0.1350650936,0.8500322104,-0.0747897476,-0.0081708999,0.0172974747,-0.0792996064,-0.1756221652,0.3416588902,0.0094196796,0.0998893976,-0.4530889094,0.0114731705,0.1211261302,0.2790077925,-0.0302848592,0.356625855,-0.3317677677,-0.01689329,0.0610288195,-0.0380293913,-0.2270005792,0.4959072471,0.0214729253,-0.2843986154,0.149572432,0.0257511251,0.1173516512,-0.1088980064,-0.1391935199,0.6097431183,-0.1634706706,0.1619003713,0.167699486,-0.1008380279,0.2936282754,0.0041687787,-0.0798387975,0.0745744556,0.4469641447,-0.2630356848,0.0946160331,-0.0006471122,0.0438775569,0.0921604261,0.3493202925,0.3144757748,0.0399240367,-0.2234862447,-0.1632773876,-0.4827316999,-0.0178312082,-0.0551409312,0.2705824375,-0.0138036627,-0.0808224529,0.0153241428,0.2147803605,0.1249227375,-0.5909251571,0.121857211,0.0748963133,-0.2518156171,0.0605006367,-0.1980689466,0.1091288552,-0.0460439101,-0.3268637359,0.2052775919,0.166438207,-0.0613013059,0.0705628321,-0.2173718214,0.0365064964,0.2530453801,0.4496693015,0.4407114685,0.0578314811,0.0564716049,0.1126218289,-0.2935213745,0.0059686396,0.0041786069,-0.0368721113,-0.0319409929,0.076609008,-0.5562335253,0.0530161858,-0.2510072887,-0.1855866611,0.0044257441,0.2043251246,-0.2433253676,-0.2719516754,0.0289719496,0.0505871326,0.2746568024,0.4501057863,-0.0314241089,0.2331926376,-0.4069156647,0.251046747,0.0878073275,-0.4496797323,-0.3951035142,0.0139955916,0.0611329451,-0.1779440641,-0.4009954333,-0.2629910409,-0.1546336114,0.2852095664,-0.3229401112,-0.0784366354,0.042717129,-0.4692257345,-0.0031509362,-0.1504492909,0.4665024579,0.2042724788,-0.109808296,-0.1756791323,-0.1627427638]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/279","title":"Dataset Preprocessing Cache with .map() function not working as expected","comments":"Hi @sarahwie \r\nThe data are reloaded from the cache if the hash of the function you provide is the same as a computation you've done before. The hash is computed by recursively looking at the python objects of the function you provide.\r\n\r\nIf you think there's an issue, can you share the function you used or a google colab please ?","body":"I've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!","comment_length":61,"text":"Dataset Preprocessing Cache with .map() function not working as expected\nI've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!\nHi @sarahwie \r\nThe data are reloaded from the cache if the hash of the function you provide is the same as a computation you've done before. The hash is computed by recursively looking at the python objects of the function you provide.\r\n\r\nIf you think there's an issue, can you share the function you used or a google colab please ?","embeddings":[0.0126092965,0.0226763878,-0.0835920572,0.110905908,0.1039312929,0.0935757384,0.0071254992,0.1588908881,0.326246798,-0.0069931992,0.2460050434,0.2585703433,0.141965732,-0.043747548,-0.0952617526,0.1991635561,0.1659029126,0.1899385899,-0.0175594129,0.1777570695,-0.3379134238,0.0617026128,-0.2909172773,-0.0424195267,-0.2941628397,-0.1983861774,0.1261185855,0.0343156345,-0.0428413227,-0.2630408406,0.301135987,0.3228375018,-0.1433222592,0.6143494844,-0.0001192959,0.0433152243,0.3999326825,-0.097369954,-0.195039615,-0.0425278246,-0.3642002046,-0.0409483537,0.1866872609,-0.0762814879,-0.0729197934,0.1407180578,-0.0348699652,-1.0032604933,0.4038298726,-0.0395837873,0.1329070926,0.0171495415,-0.4026246667,0.2707398832,0.0218685232,0.1524098963,-0.1357863396,0.0711580813,0.3388117254,-0.2379151881,-0.0876900703,0.169557333,-0.2306657284,0.2240425646,0.2353822887,-0.0649860501,-0.1396852732,-0.7217977047,0.3800291717,0.0573786609,0.5168449283,-0.2660503685,-0.3717952967,-0.1826729327,-0.2236893624,-0.1478828043,0.447894454,0.0156683493,0.04886204,0.2076193243,-0.8422005773,-0.1816661507,0.1444846988,-0.0510664396,0.1792800575,0.0006038909,-0.1845416427,0.1082615852,0.1442225873,-0.0036395558,0.3374118805,-0.4216638803,-0.0585633628,0.4882795215,-0.2444188744,0.0681823492,-0.1582017541,0.3357010484,0.0762660727,0.0582534857,0.1465757042,0.3805178702,-0.075184375,0.1383454502,0.2568175495,0.3748097122,0.1340918839,0.0845250264,0.2072888166,-0.3187167943,-0.2340548486,0.0206100643,0.2593968809,-0.2871665955,0.2802867591,0.0512089431,0.0161273442,-0.3326468468,0.1532606184,-0.1525700688,-0.3575713038,-0.3165353835,-0.0289574247,0.1038471833,0.0137027502,0.2877177,-0.1415009499,-0.0398748107,-0.2923080921,-0.0593898036,-0.0368018746,-0.2417571843,-0.3055945337,0.2998381853,0.3559484184,-0.2185664624,0.1573401242,0.2191765904,0.1640818119,-0.3253117204,0.4204201102,-0.3591628969,0.6152808666,0.1765166521,-0.1864677817,0.2618684471,0.1314363331,0.2853624225,-0.1162240878,0.3068481088,-0.4481901824,-0.09910243,0.5191603899,0.0269613843,-0.2732546926,0.2550114989,-0.1367578655,0.1432133168,0.408195734,-0.2797901034,-0.0005557017,0.0615212284,-0.3477475345,-0.3059697151,0.3221846223,0.439150542,-0.2739111185,-0.0165224429,-0.054977715,0.0793558285,0.2468385249,0.0144589916,-0.1570525765,0.2650979459,-0.2132819146,0.08846917,0.1639108509,0.0004317645,-0.4677479863,0.3946530223,0.3026471138,0.3626728058,-0.3520169258,0.0434836969,0.0295975357,-0.0650653318,-0.1620073617,0.1532619894,0.0934985206,0.0354307853,-0.1248074844,-0.2823025286,0.2753602862,-0.1980088502,0.0008318454,0.5466743112,-0.1341930926,-0.498740375,0.0262129251,-0.1358769238,0.1735254824,0.1752534658,-0.0878476799,-0.0739189908,0.1387533098,-0.1402402669,-0.3800908625,0.3710728586,-0.1506705731,-0.3001328111,-0.0492646471,-0.0100126052,0.0820081532,-0.1136208326,-0.2048955113,-0.0743711889,0.0389905982,-0.1065050513,0.2533738911,-0.190398246,-0.1162975878,0.6129919887,0.2305235416,-0.0750195235,-0.1576407254,-0.0453655459,0.1819269657,-0.1382559836,-0.3760130405,-0.0317540951,0.218743369,-0.1027434766,-0.1763006151,0.433142364,0.3840714097,0.1836776286,-0.0544663295,0.1820138544,0.0737151802,0.0586111285,0.1019367203,-0.015754519,0.1110462323,-0.1827043295,0.0047957664,0.3525701463,-0.1156669855,0.3307572603,0.0242873654,-0.0883694738,0.1098253205,-0.0161415078,-0.0593330525,-0.2415944636,-0.130489096,-0.0739989877,0.5344321132,0.0149761755,0.0760662779,-0.1101099551,0.1147863343,0.1351981908,0.0436054505,-0.1515237391,-0.1055284441,-0.1825639904,0.0495039262,0.1613128036,0.3329404294,-0.0445548706,0.3212518692,-0.039583277,0.0073794462,-0.0218328256,-0.0697137192,0.0958448201,0.3745647371,-0.048412662,0.0538952686,0.0223970562,-0.0785183832,0.1887511909,0.0866197422,-0.209569037,-0.1836449653,0.2747165561,-0.3728384376,-0.0721570477,0.1207606345,-0.1489290297,-0.3147017658,-0.198555097,-0.0873114616,0.1540362239,0.1480253786,0.2385779321,-0.0211449787,0.1770698726,-0.1963211894,-0.3962063491,-0.210881412,-0.2509934306,-0.0241288282,0.0130318105,0.058223594,-0.1628418714,0.2691910267,-0.0793194547,-0.1402595043,-0.1621198505,0.031683486,0.2521141768,0.062940225,0.0975080878,-0.3867300749,-0.0323973931,-0.2633266747,0.2480117679,-0.0150135942,-0.3994628489,-0.1009921953,-0.051359538,0.0463383794,-0.0289350785,-0.2453615069,-0.0325764865,-0.0533973686,0.0107636917,-0.0171334893,-0.1349080056,0.1015388444,0.0019319599,0.0572788939,-0.141506657,0.0444805138,-0.1505845934,-0.3941085935,-0.5403397679,0.2928726971,-0.174210161,-0.115863964,0.3231078088,0.1441299468,0.1809595376,0.8046814203,-0.2107740045,-0.0886807591,-0.1395693421,-0.0463575646,0.0705094859,0.0854880512,0.3992874622,0.1413275301,0.0925284773,-0.2513990998,-0.3856617808,0.0727317855,0.0985859707,0.1149828061,0.4612811804,0.0300731901,-0.0849621147,0.9317190647,0.2833195627,-0.0950024128,0.2669970989,0.1702574939,0.3451196849,-0.2881950736,0.0708171204,-0.2367360145,-0.4759666622,-0.1828915626,0.1063376591,0.0879182443,-0.5318136811,-0.2258446068,0.3867982626,-0.39133811,-0.1599842757,-0.0657072961,-0.3975029588,0.5221547484,0.0705342665,0.1518422067,-0.2838066816,-0.1268579215,0.0202970337,-0.164227888,0.7392302155,-0.0950199589,-0.3028748631,-0.0011110203,-0.338042289,0.3037940562,0.1055577397,0.3277979195,-0.1865992546,-0.1020381451,0.0340127423,0.130504027,0.4988676608,-0.4583169818,-0.0022161966,0.0495560355,-0.3301033676,0.0877544582,0.0777620971,0.2153149694,0.6324646473,0.0839978009,0.0645758584,0.0656525269,-0.1327894777,-0.4931657016,0.1411961615,-0.1203804389,-0.3293852806,-0.1417140216,0.1360767037,-0.2405113876,-0.1840698868,0.0228153206,-0.3011866212,-0.1076294854,-0.0868260786,-0.2012244165,0.12941508,0.2172813267,0.2465334982,0.4548769593,-0.0538460948,-0.0747747645,0.6173526049,0.1336002201,-0.0564041622,0.5255818367,-0.2111590207,0.0876726359,0.1218397021,-0.0213752575,0.1304306537,0.3328437805,-0.198222205,-0.178743571,0.0583595745,-0.2208533585,-0.5552138686,-0.0314421356,0.1663862318,0.014575283,-0.4451220036,-0.5543464422,0.2807045579,0.3444280624,-0.3840258718,0.4244666398,-0.1937873811,-0.2579278052,0.2693218291,0.0766967535,0.9050422907,-0.3999024928,-0.1579403728,-0.4110604227,0.1713767201,0.0624810532,-0.3120304048,0.3149038553,-0.1649387926,-0.2067506462,-0.1401660144,-0.217041418,0.2949486673,0.5736173987,-0.1126932055,0.3896788359,-0.0785669163,0.238388896,-0.071928747,-0.0379633754,0.2574771941,-0.2842621207,0.0468977615,0.1334727257,-0.025584992,-0.0182046779,0.0485945418,0.1120329499,-0.048060935,0.1283046752,-0.0036543061,0.0686466545,-0.3537276387,0.066544883,-0.2221658975,-0.0480659492,0.2516109645,0.207630828,0.2489759475,0.0558759645,-0.2269952893,0.0364830159,0.3870263994,0.2033219337,-0.1639775932,0.1635318547,-0.1215399727,0.0856486931,-0.1350528449,-0.0916703716,-0.2422161847,-0.3425152898,-0.0437733345,0.0718054548,0.1247105077,-0.2651447952,-0.1970210522,0.1820755601,0.2007448524,0.1349459141,0.0791841,0.1869639754,0.1148314476,0.478656739,-0.1105520576,-0.2748798728,-0.053982757,0.2129807025,0.0490762331,-0.3119797707,0.1249650568,-0.0675830096,-0.2460440397,-0.1226750538,-0.2822341025,-0.2911476791,-0.0076243146,-0.0106205326,-0.0671787933,-0.0076076025,-0.0173694734,-0.2160056084,-0.0351177193,0.1080899164,-0.1765678078,-0.2673583329,-0.1527624428,-0.1297766566,-0.1080530956,0.3174190819,0.0085240705,-0.3597826362,0.1688007563,0.3486437798,-0.1827609688,-0.0618790761,0.2007534206,0.223243013,0.09663257,0.2734761238,0.0089167822,-0.234675169,-0.1388262808,0.04787625,-0.3410569131,-0.1204113886,0.23244147,0.2194065303,-0.1128771305,-0.5000815988,0.0173114091,-0.1419840306,0.064032957,-0.4966711104,0.1426207721,0.1659921408,-0.0278108623,0.2998484373,0.3571559191,-0.0319407806,-0.2766461372,0.0705921501,-0.0489757881,0.1647659242,0.0491188914,0.0631782413,0.1680635661,0.047551237,0.0094872145,-0.1291360408,0.0904792249,-0.0207948908,0.2781990469,-0.228110984,0.2825012803,0.1824586838,0.2479124367,0.2677654028,-0.0781529173,-0.0337517709,0.1696862876,0.0921782479,0.0074982867,-0.2400321215,0.6981934905,0.2228987068,-0.0108822333,0.3533830345,0.3265448511,0.1854860634,0.2035680711,0.0393245369,0.3794471622,-0.4553790092,0.2298572361,0.1159884334,0.2000046074,0.2070942819,0.3283661306,0.2765349448,-0.1325063109,0.2599083185,0.4038740993,0.3374174237,0.3235001266,0.3901452124,-0.1910293847,-0.5856961608,0.2203454375,0.3824098408,-0.2621112168,0.3146702051,0.0561044924,-0.0249895342,-0.0380959548,-0.3837385178,-0.0846043229,0.0354876593,-0.0287532136,-0.1529734284,0.3428881764,-0.1619920433,0.240214169,-0.0478732213,0.0208126437,0.0174209233,0.8757129312,-0.0359549373,-0.0011877005,0.0544965267,-0.0359385833,-0.1531930417,0.3803105056,-0.0210077334,0.0297988765,-0.3976700306,0.0471714363,0.0692416951,0.2290407419,0.0288785249,0.3294484913,-0.376318872,0.0959211141,0.0928716809,-0.0346876755,-0.1654845923,0.4780681729,0.0181273464,-0.2641535103,0.1079873219,0.0246186107,0.1039289832,-0.1104386821,-0.0960275605,0.609519124,-0.1541203111,0.1431068629,0.1908957511,-0.1261518896,0.2218962014,0.0165834166,-0.1446873695,0.0633091927,0.4511490166,-0.2422026694,0.0936993808,0.0137431361,0.0275313202,0.0701824054,0.383351028,0.3801325858,0.1401961148,-0.2852891684,-0.1369883865,-0.5016660094,-0.050531216,-0.1271043718,0.200231567,-0.0037908398,-0.0853466243,0.0287951492,0.2967400551,0.142073378,-0.5314871669,0.0824197456,0.0986172259,-0.2431383878,-0.0272087697,-0.2360907495,0.0832537264,-0.0522726104,-0.3047014475,0.1619851142,0.1390913725,-0.0621040501,0.0414521508,-0.186995402,0.0434081852,0.252084583,0.4884822667,0.4370420277,0.1001728401,0.0623514876,0.1365892738,-0.2371641248,0.0068935952,0.0162835605,-0.0231283214,0.0041449778,0.0810713172,-0.5207995772,0.0373872742,-0.2671990395,-0.164969787,-0.0113176098,0.1435789317,-0.219568342,-0.2876952589,0.0659516156,0.0706464425,0.1852206737,0.4934231341,-0.0560478717,0.3082307279,-0.3612108529,0.2046155483,0.1274631619,-0.449752748,-0.4256587029,0.084297739,0.04652128,-0.2532380223,-0.3669771254,-0.3475477695,-0.080381386,0.2992354631,-0.319239974,-0.077626586,0.0949622989,-0.4346528351,-0.0272769779,-0.1388857216,0.4334099293,0.1447268873,-0.0748715699,-0.2155508846,-0.1994420737]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/279","title":"Dataset Preprocessing Cache with .map() function not working as expected","comments":"I can't reproduce it, so I'll close for now.","body":"I've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!","comment_length":9,"text":"Dataset Preprocessing Cache with .map() function not working as expected\nI've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. \r\n\r\nIs there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I want to be able to be certain the data is being re-processed rather than loaded from a cached file. \r\n\r\nCould you also help me understand a bit more about how the caching functionality is used for pre-processing? E.g. how is it determined when to load from a cache vs. reprocess. \r\nI was particularly having an issue where the correct dataset splits were loaded, but as soon as I applied the `.map()` function to each split independently, they somehow all exited this process having been converted to the test set.\r\nThanks!\nI can't reproduce it, so I'll close for now.","embeddings":[-0.0642543286,0.073205933,-0.0901178494,0.0947499424,0.0731438473,0.1082207263,0.0287376568,0.1852729768,0.3192409277,0.0131260911,0.2575820088,0.2563031018,0.1704948097,-0.07015495,-0.1194637045,0.2134769559,0.1760123819,0.2094057053,-0.076728493,0.2321373969,-0.298135519,0.0194161739,-0.2643180192,0.0082778055,-0.2709945738,-0.2383398116,0.1570444107,0.000467016,-0.027681781,-0.2326743156,0.2483049929,0.3339700401,-0.1761068255,0.6005613208,-0.0001187849,0.0829218253,0.4273579121,-0.0707126334,-0.1809850335,0.0168098304,-0.4140092731,-0.0191223416,0.1584269106,-0.0735043734,-0.034011323,0.1436064392,-0.0175851788,-1.0318850279,0.4004627466,-0.0961132795,0.1347199082,0.0296500344,-0.4564859569,0.2203024924,0.0246544424,0.2457195371,-0.1750435829,0.0212770943,0.3948353827,-0.1924095601,-0.148536548,0.1539518237,-0.2031824142,0.2306886166,0.206250146,-0.1109267771,-0.1310993731,-0.6701208353,0.4431790709,0.0803262815,0.51756984,-0.2539178729,-0.3116612434,-0.1801199764,-0.2065656334,-0.101361908,0.4279097021,-0.0077014733,0.054032445,0.2377340049,-0.8794162869,-0.2159480155,0.1123230159,-0.0952932611,0.2001775652,-0.0396350697,-0.193849802,0.1033156216,0.1683716178,-0.0202536285,0.3116076887,-0.447042197,-0.0928675458,0.4298344254,-0.2048047334,0.0968075916,-0.1849260628,0.2759923339,0.1281939,0.0121191088,0.1444097608,0.3975607753,-0.0556762777,0.1462874413,0.2767176628,0.3704788685,0.1794525981,0.0637636334,0.1959695518,-0.299277395,-0.199398309,0.0359522589,0.2376385182,-0.2886595726,0.1884309798,0.0639365464,-0.0184910893,-0.3533971906,0.1309803575,-0.1945502907,-0.3018546999,-0.2950330675,-0.0458489768,0.0696208924,0.0497265048,0.3037765622,-0.1779419184,-0.0294902455,-0.2920389175,-0.1746696979,-0.0375172198,-0.2673027515,-0.3128673434,0.3117409945,0.3511958122,-0.1950274259,0.1216545105,0.2388013154,0.1108484939,-0.2918173373,0.4341749549,-0.3656937778,0.5997356176,0.1562200785,-0.1584409177,0.2680432796,0.0978954211,0.3243422806,-0.1060099155,0.3188342452,-0.422909081,-0.0811137855,0.5318481922,0.0204297137,-0.2572138608,0.3012301326,-0.1675409377,0.1828514487,0.3730722368,-0.3211669624,-0.0034920808,0.0994122848,-0.3560469151,-0.2417685688,0.3288591802,0.4395881593,-0.3270152509,-0.025372047,-0.0781236589,0.010453878,0.2769451737,-0.0297118202,-0.1696924716,0.2949768007,-0.2082954645,0.1021738872,0.1455881745,0.0679650456,-0.4431673586,0.4476607144,0.2549217045,0.328125298,-0.3392567635,-0.0026792327,0.0534857363,-0.044234585,-0.2237678468,0.1233003363,0.0811230987,0.0294859372,-0.0851368681,-0.3305629492,0.244791016,-0.1889278591,-0.0346214622,0.5566186905,-0.1254718453,-0.4521719515,0.0208362415,-0.1296036541,0.2310631126,0.1705949903,-0.1569027752,-0.079436481,0.0810648277,-0.1226192266,-0.3685045242,0.3548410833,-0.1367144287,-0.3758755624,0.0265500732,-0.0388001129,0.1184129789,-0.1249430999,-0.1746519357,-0.0392298885,0.024747055,-0.0673912168,0.1775679439,-0.2105638385,-0.1057249457,0.5639373064,0.2329111993,-0.0971977338,-0.1584948748,-0.0946007073,0.1893877238,-0.1152765527,-0.3787905574,-0.0518404134,0.1428803205,-0.1256902516,-0.1760009378,0.4163169563,0.3595170081,0.2526300848,-0.0253741369,0.1532533169,0.0699197501,0.0447338074,0.0408341438,-0.0831177682,0.0632235855,-0.1795672029,-0.0331982411,0.308226943,-0.1143115461,0.3736973107,-0.0164994765,-0.0940247625,0.1179452315,-0.0247455351,-0.0203075465,-0.2651259005,-0.1252876371,-0.0751151294,0.545052886,0.0371935703,0.0911043212,-0.0972280204,0.0921362415,0.1375543028,0.0176053271,-0.1450755,-0.070068568,-0.1760185063,0.0711057559,0.1213739961,0.329858005,-0.0206218529,0.3431257904,-0.078681998,0.0176548734,-0.0427761525,-0.115988791,0.1039007753,0.4203600883,-0.0217926614,0.0207029339,0.0197794661,-0.0474516675,0.1479420066,0.112266168,-0.2461065948,-0.1497390121,0.2741379738,-0.3988142312,-0.1021442115,0.1469732821,-0.0510156043,-0.2693574131,-0.1503246874,-0.061356768,0.1275265068,0.0782962367,0.264336437,-0.0586082488,0.2740935683,-0.2225950956,-0.3690599501,-0.1946647167,-0.2103028744,-0.0535515472,0.0209052395,0.1011796668,-0.140369609,0.247580871,-0.0648234338,-0.0986543298,-0.1609881073,0.0932724103,0.2903539836,0.1387459487,0.0618954226,-0.3944920599,0.0078159766,-0.2133785039,0.2387480736,-0.0488522127,-0.3820850253,-0.1328073889,-0.0613593645,0.0190122575,-0.0369105004,-0.2799393535,-0.0137577532,0.0295983627,-0.0395518765,-0.0439448766,-0.2388339043,0.0958059654,-0.0420886688,0.0629545376,-0.088976793,0.0081206616,-0.1484750211,-0.3483912945,-0.5123869181,0.240133062,-0.1766752005,-0.1293018609,0.2652855814,0.1915601045,0.1918118,0.8058345318,-0.2255976498,0.0340134948,-0.121242851,-0.0652323067,0.0417090021,-0.0051776618,0.3981150985,0.144962281,0.0733316541,-0.2452372015,-0.3716436625,0.1349642426,0.0355058387,0.0914454162,0.4914748073,0.0340115875,-0.0512700342,0.9258566499,0.2630104423,-0.0842563286,0.2342631668,0.1890634447,0.357784152,-0.275234133,0.0844378993,-0.2125851065,-0.4665112793,-0.1470557749,0.1381558329,0.0331352018,-0.5519959331,-0.204642579,0.3729569316,-0.307672292,-0.1453412324,-0.0315500051,-0.3851319551,0.5506080389,0.040236067,0.1412619203,-0.2391425371,-0.095091626,0.0581548065,-0.1852921844,0.7536293864,-0.0660331771,-0.365557313,0.0234326944,-0.3114485443,0.2954856157,0.0934804678,0.3469876647,-0.2957945764,-0.0868679211,0.0407428145,0.1494769305,0.4968132377,-0.4228538275,0.0559574738,0.0411475524,-0.3146812022,0.0783553869,0.0924348012,0.2197737694,0.6072331071,0.1314183027,0.0920003802,0.0707011968,-0.1566975415,-0.5227555633,0.134720847,-0.126308471,-0.3459532857,-0.0930745155,0.2194395363,-0.2516708672,-0.1631875932,0.0110519566,-0.2937455773,-0.1332925558,-0.1292919219,-0.206194818,0.1295624673,0.1947051436,0.2615760863,0.4058592021,-0.0605436638,-0.1063292474,0.6564358473,0.1463507861,-0.0241792351,0.4839569032,-0.1860382557,0.1004431769,0.2192501128,-0.0785700306,0.1341471523,0.35989362,-0.1822933704,-0.1763911545,0.0641244575,-0.209010303,-0.5714933276,0.0642261133,0.1504212171,-0.0223851223,-0.4162478745,-0.4810862839,0.2473085076,0.3539803326,-0.4030331075,0.4126446843,-0.1873064041,-0.2631314099,0.2369556874,0.094971247,0.877112031,-0.3526954949,-0.1969468892,-0.4836047888,0.12538746,0.0219178703,-0.3451305628,0.2675075531,-0.1241892204,-0.2336265594,-0.1425690949,-0.2087043822,0.337813437,0.6316816807,-0.0859175548,0.333188504,-0.0812785998,0.255066663,-0.0488901436,-0.0474422835,0.3061315119,-0.3283050954,0.1332689226,0.1293193549,-0.0468079373,-0.0497225113,0.0762384087,0.0979403853,-0.0190630313,0.1915667951,0.0018939496,0.0956882983,-0.3437044919,0.0601391904,-0.2351562232,0.0239960477,0.3369866312,0.2366082221,0.2131494582,0.0495412014,-0.2127911448,0.0496345721,0.3964692652,0.2176328152,-0.2085827589,0.1372404546,-0.1663018763,0.0700102672,-0.1147300005,-0.0894428492,-0.2696544528,-0.2693307996,-0.0086840317,0.0490797572,0.1861182153,-0.2610078156,-0.206616655,0.1684306413,0.2721994221,0.1507770717,0.077571325,0.1678542197,0.1366927922,0.4171350002,-0.1100237966,-0.2671218514,-0.0453911982,0.1907973289,-0.0516295843,-0.2755194604,0.0929729268,-0.0883883089,-0.2396983802,-0.1389001608,-0.2135817856,-0.2098720819,-0.0647448823,-0.0668612272,-0.1062989607,-0.0822474509,-0.0245773196,-0.226402998,-0.0050022504,0.0772378743,-0.1714434773,-0.2369203418,-0.1518269926,-0.1311544776,-0.1730875224,0.3020019531,0.0420446657,-0.3933740556,0.1709345579,0.3817010224,-0.173580423,-0.1134989858,0.2236302942,0.2117866427,0.0953334793,0.3007218838,-0.0397548564,-0.223056823,-0.1641917527,0.0783545598,-0.3369297981,-0.1186468676,0.1936853379,0.2179947197,-0.1329143345,-0.4666802287,0.0451200642,-0.1093195602,0.0477811992,-0.4709630609,0.1482819468,0.0732669607,-0.0338634811,0.2663696408,0.3245076239,-0.0201443825,-0.3431280553,0.0725691989,-0.0019133072,0.1532038301,0.0040964927,0.0718396157,0.1887148023,0.0208511893,-0.0227855947,-0.1202193573,0.1660329401,-0.0270408969,0.2413929701,-0.1891215146,0.339345783,0.2359952182,0.2070957571,0.2037919611,-0.0754078329,-0.0410554782,0.1524400562,0.0905714259,0.0442158133,-0.2899561822,0.6700962782,0.2248651534,-0.041307155,0.3282375336,0.3558692932,0.1936858743,0.2397996336,0.0431971736,0.327388078,-0.4276858568,0.2290772349,0.1906648725,0.1824388057,0.2185592502,0.3220193684,0.300450772,-0.1150748357,0.2740242183,0.438610822,0.3694756329,0.3763111532,0.4455956519,-0.1795429438,-0.6296697259,0.2190631032,0.3999482393,-0.2699860632,0.2956845164,0.0297539197,-0.0240612719,-0.0415993929,-0.4620105922,-0.0826060772,0.0175855085,-0.0358851627,-0.1323750466,0.300150454,-0.1527070552,0.2279150933,-0.0342872292,0.0172737893,0.0888161808,0.8618578315,-0.0571181625,-0.0080409534,0.0515064225,-0.0248153433,-0.1414364278,0.3187177479,-0.0085592894,0.0572038032,-0.4151867926,0.0506014451,0.1312770993,0.2625763416,-0.0364704356,0.2989040613,-0.3125161827,0.0267034937,0.0785817206,-0.0244839471,-0.2027123868,0.4691613019,0.010644516,-0.3057741821,0.1503091902,0.016486967,0.1117262989,-0.1291390508,-0.0992134288,0.6119531393,-0.1006679311,0.1239662394,0.1853126585,-0.1195292547,0.2403075546,-0.0042259139,-0.1008480266,0.0967844725,0.4360745847,-0.288295418,0.097972773,0.0187748689,0.0235742237,0.0693057701,0.3824178874,0.3238574862,0.0610464923,-0.2780866921,-0.1840503663,-0.5073093772,-0.029870078,-0.0611304268,0.2359251827,0.0188522916,-0.1246347055,0.0137615465,0.2617409825,0.1270293146,-0.5703418851,0.1114103124,0.0963407308,-0.2865358293,0.0121407434,-0.1846765578,0.0932122171,-0.0762682781,-0.2957464755,0.2319012582,0.2071189731,-0.0710082501,0.0811378881,-0.2116896957,0.0264447499,0.2714107931,0.420237869,0.4084970057,0.074518092,0.0490913615,0.1649799049,-0.2458753735,0.0178881716,0.0170974359,-0.0477920175,0.0341312326,0.0675949752,-0.5608987212,0.0780344829,-0.2362792194,-0.2116004229,0.0098703587,0.1770815849,-0.2771661878,-0.2832274735,0.0266608149,0.1098801717,0.2440285236,0.4374373257,-0.0243376326,0.2472857833,-0.3518696725,0.2348784953,0.1074367166,-0.4272186458,-0.4389258921,0.0981944203,0.0675507784,-0.1809666157,-0.3986749053,-0.2722114623,-0.1331200153,0.2900718749,-0.3456783295,-0.0441665016,0.0813193172,-0.479052484,-0.0221751705,-0.1491462141,0.4906468093,0.1798167676,-0.1102665216,-0.1707056761,-0.179850027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/278","title":"MemoryError when loading German Wikipedia","comments":"Hi !\r\n\r\nAs you noticed, \"big\" datasets like Wikipedia require apache beam to be processed.\r\nHowever users usually don't have an apache beam runtime available (spark, dataflow, etc.) so our goal for this library is to also make available processed versions of these datasets, so that users can just download and use them right away.\r\n\r\nThis is the case for english and french wikipedia right now: we've processed them ourselves and now they are available from our google storage. However we've not processed the german one (yet).","body":"Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!","comment_length":87,"text":"MemoryError when loading German Wikipedia\nHi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!\nHi !\r\n\r\nAs you noticed, \"big\" datasets like Wikipedia require apache beam to be processed.\r\nHowever users usually don't have an apache beam runtime available (spark, dataflow, etc.) so our goal for this library is to also make available processed versions of these datasets, so that users can just download and use them right away.\r\n\r\nThis is the case for english and french wikipedia right now: we've processed them ourselves and now they are available from our google storage. However we've not processed the german one (yet).","embeddings":[-0.0479339696,0.1261895299,0.0182390139,0.6689180136,0.2445327044,0.1557507068,0.004935035,0.2083568722,0.3758642077,-0.0514769666,0.2673629224,-0.3312878907,-0.1734624952,-0.3341802955,0.2296689302,-0.4861274064,0.0930675492,0.0584275424,-0.0968580395,0.021348808,-0.2210392207,0.2842619419,-0.2860866785,-0.1296746433,-0.2414990813,0.0682763606,-0.0606940202,0.0540070944,-0.1613469124,-0.4543804526,0.1093896404,-0.2572765052,0.1181677878,0.1811023504,-0.0001217219,0.0400367081,0.4013991356,-0.1707563549,-0.4441632926,0.0392563045,-0.0454732403,-0.3407937586,0.0367344134,-0.2606716752,0.0609828457,-0.1065008789,0.3253657222,-0.082839556,0.3636316657,0.1792971939,0.1830790192,0.0583989583,0.2596977651,-0.0829991326,0.5757554173,0.0603940561,0.0517875366,0.1204346195,-0.3371914327,-0.3669273555,-0.266692996,0.2971111834,0.0168598853,0.1016100198,0.3428906202,-0.1183491275,0.014594771,-0.2151484787,0.2146662027,-0.0155843385,0.6984869242,-0.222539112,0.0544035435,-0.2987264991,-0.0819756612,0.1372263432,0.3699685335,0.3768196404,-0.53745538,-0.1752103865,-0.1816793531,-0.3489992321,-0.1634562314,0.3382711709,0.0580682494,0.2427105755,-0.046096243,0.165600881,0.3660611212,-0.0474638939,-0.2768139839,-0.1119219139,-0.0121806534,0.4462251961,-0.0702341348,0.0682196021,-0.0799727216,-0.0619321316,0.2351161987,-0.2792296708,-0.0940648913,-0.0614943281,0.2538709641,0.0773239955,0.2765713036,0.4570385516,-0.1349075586,-0.1987938285,0.1601509452,0.1806961447,0.0185495038,-0.1625983566,-0.0348524787,-0.163364619,-0.0773852915,-0.0258760657,-0.103963092,-0.0469388105,0.0200344175,0.1309570819,-0.096364513,-0.133965537,-0.0321695395,0.4189065993,-0.1397236735,0.3382306099,0.2956480384,0.0687169135,-0.384778291,-0.2950851619,-0.1493525356,0.1582213342,-0.5230259895,0.1237957925,0.2662939131,0.2771739364,0.3613065481,-0.1229094267,-0.2683769166,0.0498335995,0.2657951117,-0.1831974089,-0.0323753022,0.2156431675,0.14843674,0.2886892855,0.1391697973,-0.2281753868,0.000194584,0.0619199313,-0.1770699173,-0.1940729618,0.0692587718,0.0176814776,-0.1571019292,0.154286474,-0.4031710923,0.2367206365,0.2888612151,-0.2128570974,-0.0976049975,-0.0422633328,-0.23210226,-0.2615077198,0.2098641694,0.5494705439,-0.0914283767,-0.2689732015,-0.0759014413,-0.0987920538,0.4090284109,0.2375489473,-0.1052161306,0.2960037291,-0.2315713167,0.0987135246,0.3416324556,-0.1739220619,-0.331951797,0.1309767216,0.1102416217,0.1638558507,0.0562583506,0.0016745509,0.0118619101,0.1984750479,0.1997092515,0.3413085043,-0.0352610424,0.2907262743,-0.267072022,-0.2709881663,0.2600758076,0.0086501343,0.3516753912,0.0555153303,0.0945201367,0.8308191299,0.2724248767,-0.0689333975,0.3318047822,0.5638161898,-0.227718845,0.0573095493,-0.0112689119,-0.0633697733,-0.2486257851,0.1927720159,-0.3067270517,0.3094541728,-0.0214819368,0.1321331114,-0.1640296876,-0.0969142392,-0.0566746257,-0.3699617386,0.0888292938,0.1652462482,0.1101260856,0.2140970677,0.1379154623,0.2627291083,-0.0111610973,0.0226750709,-0.6933624744,0.2087740302,-0.1417605281,-0.1521971226,-0.0109318141,0.1096379384,0.1333905607,0.1438275576,-0.1323129535,0.1214753836,0.0990371704,-0.0469947085,0.1362223029,-0.0250016525,0.2658998072,-0.1880667061,0.4338091016,0.0586686023,0.1608509421,-0.1543041915,0.0469034724,-0.2138385028,0.072046563,0.2993028164,0.0846952274,0.0347199701,0.1672719419,0.1503384262,0.1518827975,-0.1656813025,0.4092136621,0.0806189775,0.1114226282,-0.0011588708,0.0266635604,-0.4050886929,0.4183508158,0.1737964302,0.1427270025,0.2619020641,-0.4456169605,-0.2904942334,0.2247490287,-0.057462208,0.2343137711,0.0811882243,0.1338569522,0.0289964098,0.1593956351,0.037909966,0.0936210901,0.058900822,0.5227373838,0.0115124211,0.0125517687,-0.1341618747,-0.2134275883,0.0808712468,0.0312990136,0.4169813991,-0.1902172863,-0.0390307494,-0.0026984455,-0.5402334332,-0.3053193986,0.2616876364,-0.4882035851,-0.2883309424,-0.2413621992,0.280143708,-0.0662873089,-0.0134526268,0.068596065,0.1989407092,0.1088028774,-0.1494421959,-0.1298904121,-0.2826367915,-0.3521508873,-0.0132163968,0.1780460626,0.0457917042,0.2574231327,0.0474953204,-0.129376784,-0.2252615541,-0.1390877813,0.193174839,0.0193976611,0.1323526949,0.0808379054,0.6441944838,-0.0916180611,-0.1589263827,0.3123699129,-0.1345023215,-0.1123504639,0.2058447748,0.0021174341,0.1894079149,0.0238474421,-0.0355216302,-0.1239717528,-0.4503405094,0.1686739624,0.1654811502,-0.0202529617,0.347874105,0.1282710731,0.0782882795,0.1775042415,0.1381355822,-0.190916121,-0.0510670096,0.2469130903,-0.3309720457,-0.3059126437,-0.0715213791,-0.1601582468,0.2242604643,0.4027511775,-0.6984960437,0.1061492711,-0.1276174188,-0.2190024704,0.0490920655,0.0372534879,0.3683996797,-0.1916541755,-0.0193818528,-0.0021286565,-0.0695347115,-0.0335374884,-0.2112904936,0.6732001901,0.1878428757,0.3525752127,0.2811793983,0.9196683168,0.2102591395,0.4203176796,0.4592570961,0.1845643818,0.2118268609,-0.1392084062,-0.2165400535,0.0542722419,-0.2958499491,-0.1469565779,0.1930077374,0.0539988689,-0.5334852934,-0.336306721,-0.2229558825,-0.1735776961,-0.3317177892,0.3225135505,-0.0072969436,0.3478627801,-0.0577862076,-0.0329394154,-0.2737751603,-0.4534512758,0.0244151521,0.0525489785,0.036143139,0.1982194632,-0.0706402957,0.0925828964,-0.7637271285,0.0521496274,0.0224501938,0.2525706291,-0.2339808792,-0.037597999,-0.0533665158,-0.0019641947,0.6382448673,-0.122154586,-0.1083132252,-0.0036729013,-0.0728005245,-0.6327511072,-0.0034906983,-0.0141837699,0.0197430234,0.2299287319,-0.0092491293,-0.4394513369,-0.017309675,0.1475987136,0.2754671276,-0.3025964797,-0.0935151726,-0.3661640882,-0.1195315123,-0.5057441592,-0.1286639273,0.0648095906,0.2224610299,0.3846001923,0.3160492778,0.0997214988,0.0857022554,-0.0120870713,-0.0698755458,0.2582469881,0.1297812015,-0.1368017495,0.2899037004,0.0841262639,-0.2411785573,0.3528220952,0.0503834598,-0.2506101429,0.0791513175,0.1490257233,0.1158474162,0.2734086812,0.0989278853,-0.1237156764,0.1667004228,-0.0585143901,-0.4219202399,0.2355522513,0.2815838456,-0.1854799986,-0.3053144217,-0.5116246343,0.615860939,0.0559713878,0.0304511804,0.2086319327,0.0223995503,-0.412781179,0.4413424134,0.2683010101,0.9731247425,-0.204042688,0.424998045,0.2317784727,0.1344753057,0.5237073898,-0.5167378187,0.3831552863,-0.3056226373,0.1268451512,-0.0133372135,-0.1085634455,0.0364827737,0.114719294,0.0654074848,0.3253664672,0.2079365999,0.0961439908,-0.0773934275,0.4588253498,-0.0800674111,-0.3420723975,-0.1955794841,0.1220066622,-0.3448764682,0.3898961246,-0.3645066917,0.0217100196,0.0776721239,-0.1105872989,-0.4916018248,0.0795153826,-0.3250877857,0.4055930078,-0.3284801543,-0.1371369958,-0.1421707869,0.3159619272,0.0592573695,0.1933968067,-0.249840185,0.2242742479,-0.3976946771,-0.2356773764,0.0208289362,0.2337531894,0.4087526202,-0.2069119662,-0.2438365221,0.0980038419,-0.1904449761,-0.3769418299,-0.1676462144,-0.1064466536,0.0883559063,-0.014652296,-0.3172785044,-0.00905261,-0.1413644105,0.0560814254,0.0824531466,0.0137014808,-0.1196166202,-0.0904117227,0.0306326207,-0.1625874341,0.0537343882,0.4452787936,0.1993274242,-0.1737648696,0.6748771667,0.318600148,-0.2701166868,-0.0588436611,-0.145492211,0.026524296,-0.3916162848,0.0803605095,-0.1223210543,-0.0326891243,-0.2329388112,-0.0158797093,0.2639751136,-0.128901422,0.2503610849,-0.3973307312,-0.4390514493,0.3276472688,-0.1340383142,0.1806013137,0.2681422234,-0.3036365807,0.1483330727,-0.0007178176,-0.1766967624,-0.0336865559,-0.2832540572,0.1993031353,0.3011004329,-0.1552578509,0.0712757185,0.134473592,0.0281189606,0.1406504214,-0.167130813,-0.1209419593,-0.0215271469,0.1702215821,-0.0211845562,-0.1575117856,0.0079886634,-0.4215225875,0.1337191612,-0.3100504279,0.0949737802,0.0940818116,-0.0231312681,0.3049455881,0.1652727872,0.0589501262,-0.0955347419,0.4590955079,0.004906415,0.4881094098,0.1949438304,0.1460272819,0.21745345,0.0095139295,-0.529992342,0.0595963299,-0.1048070416,-0.0354540385,0.129529193,-0.1712080538,0.015818838,0.1867771298,-0.040361397,0.102782838,-0.1173781157,-0.0446845815,0.2029732615,0.1301267743,-0.129020974,0.1219610199,0.343000412,-0.038141951,-0.2553494275,0.1517674178,0.1166727766,-0.0218609311,-0.0463967025,0.138712272,0.3468563259,-0.0800423026,0.0823748708,-0.1870741248,0.0247108527,-0.0429122858,-0.1344307065,0.2051127702,0.1837250888,0.4019264877,0.0027958902,0.1415004283,-0.1637841016,0.075403668,-0.1910852641,-0.3289398253,0.0203157384,0.5687055588,-0.0469907708,0.2843696475,-0.1446273178,0.1780040264,0.3612075746,-0.1369031966,-0.367102474,0.2313219756,-0.1447919011,0.0617254116,-0.058048822,-0.2364676446,-0.2097277045,0.2906796634,0.0528513603,-0.4837180972,0.0807831958,0.1928961873,-0.1719204783,-0.2963343859,0.4201445282,0.1663333476,-0.0192703065,-0.3831961155,0.0990737677,0.1448647529,-0.0447314866,-0.0872007981,0.3097783327,0.4060985744,0.3071936667,-0.2704363167,0.0093551604,-0.0059454297,-0.0063552903,-0.2495992184,0.2559792399,0.4065111876,0.3059679568,0.1803972274,0.0541909225,-0.1025387272,-0.3101823628,0.3593687117,-0.0628880933,-0.1212561727,-0.0192109924,-0.2770347297,0.0351019092,-0.2277391851,0.0494226553,-0.5648443103,0.1031810939,0.2087474763,-0.132521525,0.1167968512,-0.0363964811,0.0014875523,-0.1657541692,0.5369609594,0.2511483133,0.1122913584,-0.4637848437,-0.3375482261,-0.5109270215,0.0780812204,-0.3646364808,0.0649989098,-0.0553256199,0.3731818795,-0.1717679948,0.1566798538,-0.2668590546,0.1098729745,-0.1133813411,0.2389042079,-0.3001950383,-0.2587818205,0.0470050648,0.0787234753,-0.0804847926,-0.2608900666,0.0874856189,-0.0274887383,-0.0548563041,-0.2284462303,-0.2025665939,0.0881140232,0.2646359205,0.2633498013,0.1222851872,0.5243082047,0.0749866292,-0.3782134652,-0.2976669967,-0.1673058867,0.0435902327,0.0065886723,-0.080036357,0.5198352337,-0.2667166889,0.1582237184,-0.4382742643,0.4756548703,-0.1714072227,0.1068449765,-0.3627738953,-0.1564667523,0.0345594808,0.226611346,0.2080149949,0.1816634089,-0.091755636,0.077336058,-0.2539073229,-0.1309581548,0.3351980746,-0.5924607515,-0.3291895986,0.0524134412,-0.058366973,-0.0152939847,-0.1152605638,-0.5271500349,0.0343527161,0.1688707769,0.0881754756,-0.1708844006,-0.0418434329,-0.0630329624,-0.0228900574,-0.0843612254,0.0717218816,-0.1209484562,-0.3793058991,0.1373226494,-0.2624127865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/278","title":"MemoryError when loading German Wikipedia","comments":"Hi @lhoestq \r\n\r\nThank you for your quick reply. I thought this might be the case, that the processing was done for some languages and not for others. Is there any set timeline for when other languages (German, Italian) will be processed?\r\n\r\nGiven enough memory, is it possible to process the data ourselves by specifying the `beam_runner`?","body":"Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!","comment_length":56,"text":"MemoryError when loading German Wikipedia\nHi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!\nHi @lhoestq \r\n\r\nThank you for your quick reply. I thought this might be the case, that the processing was done for some languages and not for others. Is there any set timeline for when other languages (German, Italian) will be processed?\r\n\r\nGiven enough memory, is it possible to process the data ourselves by specifying the `beam_runner`?","embeddings":[-0.0479339696,0.1261895299,0.0182390139,0.6689180136,0.2445327044,0.1557507068,0.004935035,0.2083568722,0.3758642077,-0.0514769666,0.2673629224,-0.3312878907,-0.1734624952,-0.3341802955,0.2296689302,-0.4861274064,0.0930675492,0.0584275424,-0.0968580395,0.021348808,-0.2210392207,0.2842619419,-0.2860866785,-0.1296746433,-0.2414990813,0.0682763606,-0.0606940202,0.0540070944,-0.1613469124,-0.4543804526,0.1093896404,-0.2572765052,0.1181677878,0.1811023504,-0.0001217219,0.0400367081,0.4013991356,-0.1707563549,-0.4441632926,0.0392563045,-0.0454732403,-0.3407937586,0.0367344134,-0.2606716752,0.0609828457,-0.1065008789,0.3253657222,-0.082839556,0.3636316657,0.1792971939,0.1830790192,0.0583989583,0.2596977651,-0.0829991326,0.5757554173,0.0603940561,0.0517875366,0.1204346195,-0.3371914327,-0.3669273555,-0.266692996,0.2971111834,0.0168598853,0.1016100198,0.3428906202,-0.1183491275,0.014594771,-0.2151484787,0.2146662027,-0.0155843385,0.6984869242,-0.222539112,0.0544035435,-0.2987264991,-0.0819756612,0.1372263432,0.3699685335,0.3768196404,-0.53745538,-0.1752103865,-0.1816793531,-0.3489992321,-0.1634562314,0.3382711709,0.0580682494,0.2427105755,-0.046096243,0.165600881,0.3660611212,-0.0474638939,-0.2768139839,-0.1119219139,-0.0121806534,0.4462251961,-0.0702341348,0.0682196021,-0.0799727216,-0.0619321316,0.2351161987,-0.2792296708,-0.0940648913,-0.0614943281,0.2538709641,0.0773239955,0.2765713036,0.4570385516,-0.1349075586,-0.1987938285,0.1601509452,0.1806961447,0.0185495038,-0.1625983566,-0.0348524787,-0.163364619,-0.0773852915,-0.0258760657,-0.103963092,-0.0469388105,0.0200344175,0.1309570819,-0.096364513,-0.133965537,-0.0321695395,0.4189065993,-0.1397236735,0.3382306099,0.2956480384,0.0687169135,-0.384778291,-0.2950851619,-0.1493525356,0.1582213342,-0.5230259895,0.1237957925,0.2662939131,0.2771739364,0.3613065481,-0.1229094267,-0.2683769166,0.0498335995,0.2657951117,-0.1831974089,-0.0323753022,0.2156431675,0.14843674,0.2886892855,0.1391697973,-0.2281753868,0.000194584,0.0619199313,-0.1770699173,-0.1940729618,0.0692587718,0.0176814776,-0.1571019292,0.154286474,-0.4031710923,0.2367206365,0.2888612151,-0.2128570974,-0.0976049975,-0.0422633328,-0.23210226,-0.2615077198,0.2098641694,0.5494705439,-0.0914283767,-0.2689732015,-0.0759014413,-0.0987920538,0.4090284109,0.2375489473,-0.1052161306,0.2960037291,-0.2315713167,0.0987135246,0.3416324556,-0.1739220619,-0.331951797,0.1309767216,0.1102416217,0.1638558507,0.0562583506,0.0016745509,0.0118619101,0.1984750479,0.1997092515,0.3413085043,-0.0352610424,0.2907262743,-0.267072022,-0.2709881663,0.2600758076,0.0086501343,0.3516753912,0.0555153303,0.0945201367,0.8308191299,0.2724248767,-0.0689333975,0.3318047822,0.5638161898,-0.227718845,0.0573095493,-0.0112689119,-0.0633697733,-0.2486257851,0.1927720159,-0.3067270517,0.3094541728,-0.0214819368,0.1321331114,-0.1640296876,-0.0969142392,-0.0566746257,-0.3699617386,0.0888292938,0.1652462482,0.1101260856,0.2140970677,0.1379154623,0.2627291083,-0.0111610973,0.0226750709,-0.6933624744,0.2087740302,-0.1417605281,-0.1521971226,-0.0109318141,0.1096379384,0.1333905607,0.1438275576,-0.1323129535,0.1214753836,0.0990371704,-0.0469947085,0.1362223029,-0.0250016525,0.2658998072,-0.1880667061,0.4338091016,0.0586686023,0.1608509421,-0.1543041915,0.0469034724,-0.2138385028,0.072046563,0.2993028164,0.0846952274,0.0347199701,0.1672719419,0.1503384262,0.1518827975,-0.1656813025,0.4092136621,0.0806189775,0.1114226282,-0.0011588708,0.0266635604,-0.4050886929,0.4183508158,0.1737964302,0.1427270025,0.2619020641,-0.4456169605,-0.2904942334,0.2247490287,-0.057462208,0.2343137711,0.0811882243,0.1338569522,0.0289964098,0.1593956351,0.037909966,0.0936210901,0.058900822,0.5227373838,0.0115124211,0.0125517687,-0.1341618747,-0.2134275883,0.0808712468,0.0312990136,0.4169813991,-0.1902172863,-0.0390307494,-0.0026984455,-0.5402334332,-0.3053193986,0.2616876364,-0.4882035851,-0.2883309424,-0.2413621992,0.280143708,-0.0662873089,-0.0134526268,0.068596065,0.1989407092,0.1088028774,-0.1494421959,-0.1298904121,-0.2826367915,-0.3521508873,-0.0132163968,0.1780460626,0.0457917042,0.2574231327,0.0474953204,-0.129376784,-0.2252615541,-0.1390877813,0.193174839,0.0193976611,0.1323526949,0.0808379054,0.6441944838,-0.0916180611,-0.1589263827,0.3123699129,-0.1345023215,-0.1123504639,0.2058447748,0.0021174341,0.1894079149,0.0238474421,-0.0355216302,-0.1239717528,-0.4503405094,0.1686739624,0.1654811502,-0.0202529617,0.347874105,0.1282710731,0.0782882795,0.1775042415,0.1381355822,-0.190916121,-0.0510670096,0.2469130903,-0.3309720457,-0.3059126437,-0.0715213791,-0.1601582468,0.2242604643,0.4027511775,-0.6984960437,0.1061492711,-0.1276174188,-0.2190024704,0.0490920655,0.0372534879,0.3683996797,-0.1916541755,-0.0193818528,-0.0021286565,-0.0695347115,-0.0335374884,-0.2112904936,0.6732001901,0.1878428757,0.3525752127,0.2811793983,0.9196683168,0.2102591395,0.4203176796,0.4592570961,0.1845643818,0.2118268609,-0.1392084062,-0.2165400535,0.0542722419,-0.2958499491,-0.1469565779,0.1930077374,0.0539988689,-0.5334852934,-0.336306721,-0.2229558825,-0.1735776961,-0.3317177892,0.3225135505,-0.0072969436,0.3478627801,-0.0577862076,-0.0329394154,-0.2737751603,-0.4534512758,0.0244151521,0.0525489785,0.036143139,0.1982194632,-0.0706402957,0.0925828964,-0.7637271285,0.0521496274,0.0224501938,0.2525706291,-0.2339808792,-0.037597999,-0.0533665158,-0.0019641947,0.6382448673,-0.122154586,-0.1083132252,-0.0036729013,-0.0728005245,-0.6327511072,-0.0034906983,-0.0141837699,0.0197430234,0.2299287319,-0.0092491293,-0.4394513369,-0.017309675,0.1475987136,0.2754671276,-0.3025964797,-0.0935151726,-0.3661640882,-0.1195315123,-0.5057441592,-0.1286639273,0.0648095906,0.2224610299,0.3846001923,0.3160492778,0.0997214988,0.0857022554,-0.0120870713,-0.0698755458,0.2582469881,0.1297812015,-0.1368017495,0.2899037004,0.0841262639,-0.2411785573,0.3528220952,0.0503834598,-0.2506101429,0.0791513175,0.1490257233,0.1158474162,0.2734086812,0.0989278853,-0.1237156764,0.1667004228,-0.0585143901,-0.4219202399,0.2355522513,0.2815838456,-0.1854799986,-0.3053144217,-0.5116246343,0.615860939,0.0559713878,0.0304511804,0.2086319327,0.0223995503,-0.412781179,0.4413424134,0.2683010101,0.9731247425,-0.204042688,0.424998045,0.2317784727,0.1344753057,0.5237073898,-0.5167378187,0.3831552863,-0.3056226373,0.1268451512,-0.0133372135,-0.1085634455,0.0364827737,0.114719294,0.0654074848,0.3253664672,0.2079365999,0.0961439908,-0.0773934275,0.4588253498,-0.0800674111,-0.3420723975,-0.1955794841,0.1220066622,-0.3448764682,0.3898961246,-0.3645066917,0.0217100196,0.0776721239,-0.1105872989,-0.4916018248,0.0795153826,-0.3250877857,0.4055930078,-0.3284801543,-0.1371369958,-0.1421707869,0.3159619272,0.0592573695,0.1933968067,-0.249840185,0.2242742479,-0.3976946771,-0.2356773764,0.0208289362,0.2337531894,0.4087526202,-0.2069119662,-0.2438365221,0.0980038419,-0.1904449761,-0.3769418299,-0.1676462144,-0.1064466536,0.0883559063,-0.014652296,-0.3172785044,-0.00905261,-0.1413644105,0.0560814254,0.0824531466,0.0137014808,-0.1196166202,-0.0904117227,0.0306326207,-0.1625874341,0.0537343882,0.4452787936,0.1993274242,-0.1737648696,0.6748771667,0.318600148,-0.2701166868,-0.0588436611,-0.145492211,0.026524296,-0.3916162848,0.0803605095,-0.1223210543,-0.0326891243,-0.2329388112,-0.0158797093,0.2639751136,-0.128901422,0.2503610849,-0.3973307312,-0.4390514493,0.3276472688,-0.1340383142,0.1806013137,0.2681422234,-0.3036365807,0.1483330727,-0.0007178176,-0.1766967624,-0.0336865559,-0.2832540572,0.1993031353,0.3011004329,-0.1552578509,0.0712757185,0.134473592,0.0281189606,0.1406504214,-0.167130813,-0.1209419593,-0.0215271469,0.1702215821,-0.0211845562,-0.1575117856,0.0079886634,-0.4215225875,0.1337191612,-0.3100504279,0.0949737802,0.0940818116,-0.0231312681,0.3049455881,0.1652727872,0.0589501262,-0.0955347419,0.4590955079,0.004906415,0.4881094098,0.1949438304,0.1460272819,0.21745345,0.0095139295,-0.529992342,0.0595963299,-0.1048070416,-0.0354540385,0.129529193,-0.1712080538,0.015818838,0.1867771298,-0.040361397,0.102782838,-0.1173781157,-0.0446845815,0.2029732615,0.1301267743,-0.129020974,0.1219610199,0.343000412,-0.038141951,-0.2553494275,0.1517674178,0.1166727766,-0.0218609311,-0.0463967025,0.138712272,0.3468563259,-0.0800423026,0.0823748708,-0.1870741248,0.0247108527,-0.0429122858,-0.1344307065,0.2051127702,0.1837250888,0.4019264877,0.0027958902,0.1415004283,-0.1637841016,0.075403668,-0.1910852641,-0.3289398253,0.0203157384,0.5687055588,-0.0469907708,0.2843696475,-0.1446273178,0.1780040264,0.3612075746,-0.1369031966,-0.367102474,0.2313219756,-0.1447919011,0.0617254116,-0.058048822,-0.2364676446,-0.2097277045,0.2906796634,0.0528513603,-0.4837180972,0.0807831958,0.1928961873,-0.1719204783,-0.2963343859,0.4201445282,0.1663333476,-0.0192703065,-0.3831961155,0.0990737677,0.1448647529,-0.0447314866,-0.0872007981,0.3097783327,0.4060985744,0.3071936667,-0.2704363167,0.0093551604,-0.0059454297,-0.0063552903,-0.2495992184,0.2559792399,0.4065111876,0.3059679568,0.1803972274,0.0541909225,-0.1025387272,-0.3101823628,0.3593687117,-0.0628880933,-0.1212561727,-0.0192109924,-0.2770347297,0.0351019092,-0.2277391851,0.0494226553,-0.5648443103,0.1031810939,0.2087474763,-0.132521525,0.1167968512,-0.0363964811,0.0014875523,-0.1657541692,0.5369609594,0.2511483133,0.1122913584,-0.4637848437,-0.3375482261,-0.5109270215,0.0780812204,-0.3646364808,0.0649989098,-0.0553256199,0.3731818795,-0.1717679948,0.1566798538,-0.2668590546,0.1098729745,-0.1133813411,0.2389042079,-0.3001950383,-0.2587818205,0.0470050648,0.0787234753,-0.0804847926,-0.2608900666,0.0874856189,-0.0274887383,-0.0548563041,-0.2284462303,-0.2025665939,0.0881140232,0.2646359205,0.2633498013,0.1222851872,0.5243082047,0.0749866292,-0.3782134652,-0.2976669967,-0.1673058867,0.0435902327,0.0065886723,-0.080036357,0.5198352337,-0.2667166889,0.1582237184,-0.4382742643,0.4756548703,-0.1714072227,0.1068449765,-0.3627738953,-0.1564667523,0.0345594808,0.226611346,0.2080149949,0.1816634089,-0.091755636,0.077336058,-0.2539073229,-0.1309581548,0.3351980746,-0.5924607515,-0.3291895986,0.0524134412,-0.058366973,-0.0152939847,-0.1152605638,-0.5271500349,0.0343527161,0.1688707769,0.0881754756,-0.1708844006,-0.0418434329,-0.0630329624,-0.0228900574,-0.0843612254,0.0717218816,-0.1209484562,-0.3793058991,0.1373226494,-0.2624127865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/278","title":"MemoryError when loading German Wikipedia","comments":"Adding them is definitely in our short term objectives. I'll be working on this early next week :)\r\n\r\nAlthough if you have an apache beam runtime feel free to specify the beam runner. You can find more info [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/docs\/beam_dataset.md) on how to make it work on Dataflow but you can adapt it for Spark or any other beam runtime (by changing the `runner`).\r\n\r\nHowever if you don't have a beam runtime and even if you have enough memory, I discourage you to use the `DirectRunner` on the german or italian wikipedia. According to Apache Beam documentation it was made for testing purposes and therefore it is memory-inefficient.","body":"Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!","comment_length":107,"text":"MemoryError when loading German Wikipedia\nHi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!\nAdding them is definitely in our short term objectives. I'll be working on this early next week :)\r\n\r\nAlthough if you have an apache beam runtime feel free to specify the beam runner. You can find more info [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/docs\/beam_dataset.md) on how to make it work on Dataflow but you can adapt it for Spark or any other beam runtime (by changing the `runner`).\r\n\r\nHowever if you don't have a beam runtime and even if you have enough memory, I discourage you to use the `DirectRunner` on the german or italian wikipedia. According to Apache Beam documentation it was made for testing purposes and therefore it is memory-inefficient.","embeddings":[-0.0479339696,0.1261895299,0.0182390139,0.6689180136,0.2445327044,0.1557507068,0.004935035,0.2083568722,0.3758642077,-0.0514769666,0.2673629224,-0.3312878907,-0.1734624952,-0.3341802955,0.2296689302,-0.4861274064,0.0930675492,0.0584275424,-0.0968580395,0.021348808,-0.2210392207,0.2842619419,-0.2860866785,-0.1296746433,-0.2414990813,0.0682763606,-0.0606940202,0.0540070944,-0.1613469124,-0.4543804526,0.1093896404,-0.2572765052,0.1181677878,0.1811023504,-0.0001217219,0.0400367081,0.4013991356,-0.1707563549,-0.4441632926,0.0392563045,-0.0454732403,-0.3407937586,0.0367344134,-0.2606716752,0.0609828457,-0.1065008789,0.3253657222,-0.082839556,0.3636316657,0.1792971939,0.1830790192,0.0583989583,0.2596977651,-0.0829991326,0.5757554173,0.0603940561,0.0517875366,0.1204346195,-0.3371914327,-0.3669273555,-0.266692996,0.2971111834,0.0168598853,0.1016100198,0.3428906202,-0.1183491275,0.014594771,-0.2151484787,0.2146662027,-0.0155843385,0.6984869242,-0.222539112,0.0544035435,-0.2987264991,-0.0819756612,0.1372263432,0.3699685335,0.3768196404,-0.53745538,-0.1752103865,-0.1816793531,-0.3489992321,-0.1634562314,0.3382711709,0.0580682494,0.2427105755,-0.046096243,0.165600881,0.3660611212,-0.0474638939,-0.2768139839,-0.1119219139,-0.0121806534,0.4462251961,-0.0702341348,0.0682196021,-0.0799727216,-0.0619321316,0.2351161987,-0.2792296708,-0.0940648913,-0.0614943281,0.2538709641,0.0773239955,0.2765713036,0.4570385516,-0.1349075586,-0.1987938285,0.1601509452,0.1806961447,0.0185495038,-0.1625983566,-0.0348524787,-0.163364619,-0.0773852915,-0.0258760657,-0.103963092,-0.0469388105,0.0200344175,0.1309570819,-0.096364513,-0.133965537,-0.0321695395,0.4189065993,-0.1397236735,0.3382306099,0.2956480384,0.0687169135,-0.384778291,-0.2950851619,-0.1493525356,0.1582213342,-0.5230259895,0.1237957925,0.2662939131,0.2771739364,0.3613065481,-0.1229094267,-0.2683769166,0.0498335995,0.2657951117,-0.1831974089,-0.0323753022,0.2156431675,0.14843674,0.2886892855,0.1391697973,-0.2281753868,0.000194584,0.0619199313,-0.1770699173,-0.1940729618,0.0692587718,0.0176814776,-0.1571019292,0.154286474,-0.4031710923,0.2367206365,0.2888612151,-0.2128570974,-0.0976049975,-0.0422633328,-0.23210226,-0.2615077198,0.2098641694,0.5494705439,-0.0914283767,-0.2689732015,-0.0759014413,-0.0987920538,0.4090284109,0.2375489473,-0.1052161306,0.2960037291,-0.2315713167,0.0987135246,0.3416324556,-0.1739220619,-0.331951797,0.1309767216,0.1102416217,0.1638558507,0.0562583506,0.0016745509,0.0118619101,0.1984750479,0.1997092515,0.3413085043,-0.0352610424,0.2907262743,-0.267072022,-0.2709881663,0.2600758076,0.0086501343,0.3516753912,0.0555153303,0.0945201367,0.8308191299,0.2724248767,-0.0689333975,0.3318047822,0.5638161898,-0.227718845,0.0573095493,-0.0112689119,-0.0633697733,-0.2486257851,0.1927720159,-0.3067270517,0.3094541728,-0.0214819368,0.1321331114,-0.1640296876,-0.0969142392,-0.0566746257,-0.3699617386,0.0888292938,0.1652462482,0.1101260856,0.2140970677,0.1379154623,0.2627291083,-0.0111610973,0.0226750709,-0.6933624744,0.2087740302,-0.1417605281,-0.1521971226,-0.0109318141,0.1096379384,0.1333905607,0.1438275576,-0.1323129535,0.1214753836,0.0990371704,-0.0469947085,0.1362223029,-0.0250016525,0.2658998072,-0.1880667061,0.4338091016,0.0586686023,0.1608509421,-0.1543041915,0.0469034724,-0.2138385028,0.072046563,0.2993028164,0.0846952274,0.0347199701,0.1672719419,0.1503384262,0.1518827975,-0.1656813025,0.4092136621,0.0806189775,0.1114226282,-0.0011588708,0.0266635604,-0.4050886929,0.4183508158,0.1737964302,0.1427270025,0.2619020641,-0.4456169605,-0.2904942334,0.2247490287,-0.057462208,0.2343137711,0.0811882243,0.1338569522,0.0289964098,0.1593956351,0.037909966,0.0936210901,0.058900822,0.5227373838,0.0115124211,0.0125517687,-0.1341618747,-0.2134275883,0.0808712468,0.0312990136,0.4169813991,-0.1902172863,-0.0390307494,-0.0026984455,-0.5402334332,-0.3053193986,0.2616876364,-0.4882035851,-0.2883309424,-0.2413621992,0.280143708,-0.0662873089,-0.0134526268,0.068596065,0.1989407092,0.1088028774,-0.1494421959,-0.1298904121,-0.2826367915,-0.3521508873,-0.0132163968,0.1780460626,0.0457917042,0.2574231327,0.0474953204,-0.129376784,-0.2252615541,-0.1390877813,0.193174839,0.0193976611,0.1323526949,0.0808379054,0.6441944838,-0.0916180611,-0.1589263827,0.3123699129,-0.1345023215,-0.1123504639,0.2058447748,0.0021174341,0.1894079149,0.0238474421,-0.0355216302,-0.1239717528,-0.4503405094,0.1686739624,0.1654811502,-0.0202529617,0.347874105,0.1282710731,0.0782882795,0.1775042415,0.1381355822,-0.190916121,-0.0510670096,0.2469130903,-0.3309720457,-0.3059126437,-0.0715213791,-0.1601582468,0.2242604643,0.4027511775,-0.6984960437,0.1061492711,-0.1276174188,-0.2190024704,0.0490920655,0.0372534879,0.3683996797,-0.1916541755,-0.0193818528,-0.0021286565,-0.0695347115,-0.0335374884,-0.2112904936,0.6732001901,0.1878428757,0.3525752127,0.2811793983,0.9196683168,0.2102591395,0.4203176796,0.4592570961,0.1845643818,0.2118268609,-0.1392084062,-0.2165400535,0.0542722419,-0.2958499491,-0.1469565779,0.1930077374,0.0539988689,-0.5334852934,-0.336306721,-0.2229558825,-0.1735776961,-0.3317177892,0.3225135505,-0.0072969436,0.3478627801,-0.0577862076,-0.0329394154,-0.2737751603,-0.4534512758,0.0244151521,0.0525489785,0.036143139,0.1982194632,-0.0706402957,0.0925828964,-0.7637271285,0.0521496274,0.0224501938,0.2525706291,-0.2339808792,-0.037597999,-0.0533665158,-0.0019641947,0.6382448673,-0.122154586,-0.1083132252,-0.0036729013,-0.0728005245,-0.6327511072,-0.0034906983,-0.0141837699,0.0197430234,0.2299287319,-0.0092491293,-0.4394513369,-0.017309675,0.1475987136,0.2754671276,-0.3025964797,-0.0935151726,-0.3661640882,-0.1195315123,-0.5057441592,-0.1286639273,0.0648095906,0.2224610299,0.3846001923,0.3160492778,0.0997214988,0.0857022554,-0.0120870713,-0.0698755458,0.2582469881,0.1297812015,-0.1368017495,0.2899037004,0.0841262639,-0.2411785573,0.3528220952,0.0503834598,-0.2506101429,0.0791513175,0.1490257233,0.1158474162,0.2734086812,0.0989278853,-0.1237156764,0.1667004228,-0.0585143901,-0.4219202399,0.2355522513,0.2815838456,-0.1854799986,-0.3053144217,-0.5116246343,0.615860939,0.0559713878,0.0304511804,0.2086319327,0.0223995503,-0.412781179,0.4413424134,0.2683010101,0.9731247425,-0.204042688,0.424998045,0.2317784727,0.1344753057,0.5237073898,-0.5167378187,0.3831552863,-0.3056226373,0.1268451512,-0.0133372135,-0.1085634455,0.0364827737,0.114719294,0.0654074848,0.3253664672,0.2079365999,0.0961439908,-0.0773934275,0.4588253498,-0.0800674111,-0.3420723975,-0.1955794841,0.1220066622,-0.3448764682,0.3898961246,-0.3645066917,0.0217100196,0.0776721239,-0.1105872989,-0.4916018248,0.0795153826,-0.3250877857,0.4055930078,-0.3284801543,-0.1371369958,-0.1421707869,0.3159619272,0.0592573695,0.1933968067,-0.249840185,0.2242742479,-0.3976946771,-0.2356773764,0.0208289362,0.2337531894,0.4087526202,-0.2069119662,-0.2438365221,0.0980038419,-0.1904449761,-0.3769418299,-0.1676462144,-0.1064466536,0.0883559063,-0.014652296,-0.3172785044,-0.00905261,-0.1413644105,0.0560814254,0.0824531466,0.0137014808,-0.1196166202,-0.0904117227,0.0306326207,-0.1625874341,0.0537343882,0.4452787936,0.1993274242,-0.1737648696,0.6748771667,0.318600148,-0.2701166868,-0.0588436611,-0.145492211,0.026524296,-0.3916162848,0.0803605095,-0.1223210543,-0.0326891243,-0.2329388112,-0.0158797093,0.2639751136,-0.128901422,0.2503610849,-0.3973307312,-0.4390514493,0.3276472688,-0.1340383142,0.1806013137,0.2681422234,-0.3036365807,0.1483330727,-0.0007178176,-0.1766967624,-0.0336865559,-0.2832540572,0.1993031353,0.3011004329,-0.1552578509,0.0712757185,0.134473592,0.0281189606,0.1406504214,-0.167130813,-0.1209419593,-0.0215271469,0.1702215821,-0.0211845562,-0.1575117856,0.0079886634,-0.4215225875,0.1337191612,-0.3100504279,0.0949737802,0.0940818116,-0.0231312681,0.3049455881,0.1652727872,0.0589501262,-0.0955347419,0.4590955079,0.004906415,0.4881094098,0.1949438304,0.1460272819,0.21745345,0.0095139295,-0.529992342,0.0595963299,-0.1048070416,-0.0354540385,0.129529193,-0.1712080538,0.015818838,0.1867771298,-0.040361397,0.102782838,-0.1173781157,-0.0446845815,0.2029732615,0.1301267743,-0.129020974,0.1219610199,0.343000412,-0.038141951,-0.2553494275,0.1517674178,0.1166727766,-0.0218609311,-0.0463967025,0.138712272,0.3468563259,-0.0800423026,0.0823748708,-0.1870741248,0.0247108527,-0.0429122858,-0.1344307065,0.2051127702,0.1837250888,0.4019264877,0.0027958902,0.1415004283,-0.1637841016,0.075403668,-0.1910852641,-0.3289398253,0.0203157384,0.5687055588,-0.0469907708,0.2843696475,-0.1446273178,0.1780040264,0.3612075746,-0.1369031966,-0.367102474,0.2313219756,-0.1447919011,0.0617254116,-0.058048822,-0.2364676446,-0.2097277045,0.2906796634,0.0528513603,-0.4837180972,0.0807831958,0.1928961873,-0.1719204783,-0.2963343859,0.4201445282,0.1663333476,-0.0192703065,-0.3831961155,0.0990737677,0.1448647529,-0.0447314866,-0.0872007981,0.3097783327,0.4060985744,0.3071936667,-0.2704363167,0.0093551604,-0.0059454297,-0.0063552903,-0.2495992184,0.2559792399,0.4065111876,0.3059679568,0.1803972274,0.0541909225,-0.1025387272,-0.3101823628,0.3593687117,-0.0628880933,-0.1212561727,-0.0192109924,-0.2770347297,0.0351019092,-0.2277391851,0.0494226553,-0.5648443103,0.1031810939,0.2087474763,-0.132521525,0.1167968512,-0.0363964811,0.0014875523,-0.1657541692,0.5369609594,0.2511483133,0.1122913584,-0.4637848437,-0.3375482261,-0.5109270215,0.0780812204,-0.3646364808,0.0649989098,-0.0553256199,0.3731818795,-0.1717679948,0.1566798538,-0.2668590546,0.1098729745,-0.1133813411,0.2389042079,-0.3001950383,-0.2587818205,0.0470050648,0.0787234753,-0.0804847926,-0.2608900666,0.0874856189,-0.0274887383,-0.0548563041,-0.2284462303,-0.2025665939,0.0881140232,0.2646359205,0.2633498013,0.1222851872,0.5243082047,0.0749866292,-0.3782134652,-0.2976669967,-0.1673058867,0.0435902327,0.0065886723,-0.080036357,0.5198352337,-0.2667166889,0.1582237184,-0.4382742643,0.4756548703,-0.1714072227,0.1068449765,-0.3627738953,-0.1564667523,0.0345594808,0.226611346,0.2080149949,0.1816634089,-0.091755636,0.077336058,-0.2539073229,-0.1309581548,0.3351980746,-0.5924607515,-0.3291895986,0.0524134412,-0.058366973,-0.0152939847,-0.1152605638,-0.5271500349,0.0343527161,0.1688707769,0.0881754756,-0.1708844006,-0.0418434329,-0.0630329624,-0.0228900574,-0.0843612254,0.0717218816,-0.1209484562,-0.3793058991,0.1373226494,-0.2624127865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/278","title":"MemoryError when loading German Wikipedia","comments":"German is [almost] done @gregburman","body":"Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!","comment_length":5,"text":"MemoryError when loading German Wikipedia\nHi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!\nGerman is [almost] done @gregburman","embeddings":[-0.0479339696,0.1261895299,0.0182390139,0.6689180136,0.2445327044,0.1557507068,0.004935035,0.2083568722,0.3758642077,-0.0514769666,0.2673629224,-0.3312878907,-0.1734624952,-0.3341802955,0.2296689302,-0.4861274064,0.0930675492,0.0584275424,-0.0968580395,0.021348808,-0.2210392207,0.2842619419,-0.2860866785,-0.1296746433,-0.2414990813,0.0682763606,-0.0606940202,0.0540070944,-0.1613469124,-0.4543804526,0.1093896404,-0.2572765052,0.1181677878,0.1811023504,-0.0001217219,0.0400367081,0.4013991356,-0.1707563549,-0.4441632926,0.0392563045,-0.0454732403,-0.3407937586,0.0367344134,-0.2606716752,0.0609828457,-0.1065008789,0.3253657222,-0.082839556,0.3636316657,0.1792971939,0.1830790192,0.0583989583,0.2596977651,-0.0829991326,0.5757554173,0.0603940561,0.0517875366,0.1204346195,-0.3371914327,-0.3669273555,-0.266692996,0.2971111834,0.0168598853,0.1016100198,0.3428906202,-0.1183491275,0.014594771,-0.2151484787,0.2146662027,-0.0155843385,0.6984869242,-0.222539112,0.0544035435,-0.2987264991,-0.0819756612,0.1372263432,0.3699685335,0.3768196404,-0.53745538,-0.1752103865,-0.1816793531,-0.3489992321,-0.1634562314,0.3382711709,0.0580682494,0.2427105755,-0.046096243,0.165600881,0.3660611212,-0.0474638939,-0.2768139839,-0.1119219139,-0.0121806534,0.4462251961,-0.0702341348,0.0682196021,-0.0799727216,-0.0619321316,0.2351161987,-0.2792296708,-0.0940648913,-0.0614943281,0.2538709641,0.0773239955,0.2765713036,0.4570385516,-0.1349075586,-0.1987938285,0.1601509452,0.1806961447,0.0185495038,-0.1625983566,-0.0348524787,-0.163364619,-0.0773852915,-0.0258760657,-0.103963092,-0.0469388105,0.0200344175,0.1309570819,-0.096364513,-0.133965537,-0.0321695395,0.4189065993,-0.1397236735,0.3382306099,0.2956480384,0.0687169135,-0.384778291,-0.2950851619,-0.1493525356,0.1582213342,-0.5230259895,0.1237957925,0.2662939131,0.2771739364,0.3613065481,-0.1229094267,-0.2683769166,0.0498335995,0.2657951117,-0.1831974089,-0.0323753022,0.2156431675,0.14843674,0.2886892855,0.1391697973,-0.2281753868,0.000194584,0.0619199313,-0.1770699173,-0.1940729618,0.0692587718,0.0176814776,-0.1571019292,0.154286474,-0.4031710923,0.2367206365,0.2888612151,-0.2128570974,-0.0976049975,-0.0422633328,-0.23210226,-0.2615077198,0.2098641694,0.5494705439,-0.0914283767,-0.2689732015,-0.0759014413,-0.0987920538,0.4090284109,0.2375489473,-0.1052161306,0.2960037291,-0.2315713167,0.0987135246,0.3416324556,-0.1739220619,-0.331951797,0.1309767216,0.1102416217,0.1638558507,0.0562583506,0.0016745509,0.0118619101,0.1984750479,0.1997092515,0.3413085043,-0.0352610424,0.2907262743,-0.267072022,-0.2709881663,0.2600758076,0.0086501343,0.3516753912,0.0555153303,0.0945201367,0.8308191299,0.2724248767,-0.0689333975,0.3318047822,0.5638161898,-0.227718845,0.0573095493,-0.0112689119,-0.0633697733,-0.2486257851,0.1927720159,-0.3067270517,0.3094541728,-0.0214819368,0.1321331114,-0.1640296876,-0.0969142392,-0.0566746257,-0.3699617386,0.0888292938,0.1652462482,0.1101260856,0.2140970677,0.1379154623,0.2627291083,-0.0111610973,0.0226750709,-0.6933624744,0.2087740302,-0.1417605281,-0.1521971226,-0.0109318141,0.1096379384,0.1333905607,0.1438275576,-0.1323129535,0.1214753836,0.0990371704,-0.0469947085,0.1362223029,-0.0250016525,0.2658998072,-0.1880667061,0.4338091016,0.0586686023,0.1608509421,-0.1543041915,0.0469034724,-0.2138385028,0.072046563,0.2993028164,0.0846952274,0.0347199701,0.1672719419,0.1503384262,0.1518827975,-0.1656813025,0.4092136621,0.0806189775,0.1114226282,-0.0011588708,0.0266635604,-0.4050886929,0.4183508158,0.1737964302,0.1427270025,0.2619020641,-0.4456169605,-0.2904942334,0.2247490287,-0.057462208,0.2343137711,0.0811882243,0.1338569522,0.0289964098,0.1593956351,0.037909966,0.0936210901,0.058900822,0.5227373838,0.0115124211,0.0125517687,-0.1341618747,-0.2134275883,0.0808712468,0.0312990136,0.4169813991,-0.1902172863,-0.0390307494,-0.0026984455,-0.5402334332,-0.3053193986,0.2616876364,-0.4882035851,-0.2883309424,-0.2413621992,0.280143708,-0.0662873089,-0.0134526268,0.068596065,0.1989407092,0.1088028774,-0.1494421959,-0.1298904121,-0.2826367915,-0.3521508873,-0.0132163968,0.1780460626,0.0457917042,0.2574231327,0.0474953204,-0.129376784,-0.2252615541,-0.1390877813,0.193174839,0.0193976611,0.1323526949,0.0808379054,0.6441944838,-0.0916180611,-0.1589263827,0.3123699129,-0.1345023215,-0.1123504639,0.2058447748,0.0021174341,0.1894079149,0.0238474421,-0.0355216302,-0.1239717528,-0.4503405094,0.1686739624,0.1654811502,-0.0202529617,0.347874105,0.1282710731,0.0782882795,0.1775042415,0.1381355822,-0.190916121,-0.0510670096,0.2469130903,-0.3309720457,-0.3059126437,-0.0715213791,-0.1601582468,0.2242604643,0.4027511775,-0.6984960437,0.1061492711,-0.1276174188,-0.2190024704,0.0490920655,0.0372534879,0.3683996797,-0.1916541755,-0.0193818528,-0.0021286565,-0.0695347115,-0.0335374884,-0.2112904936,0.6732001901,0.1878428757,0.3525752127,0.2811793983,0.9196683168,0.2102591395,0.4203176796,0.4592570961,0.1845643818,0.2118268609,-0.1392084062,-0.2165400535,0.0542722419,-0.2958499491,-0.1469565779,0.1930077374,0.0539988689,-0.5334852934,-0.336306721,-0.2229558825,-0.1735776961,-0.3317177892,0.3225135505,-0.0072969436,0.3478627801,-0.0577862076,-0.0329394154,-0.2737751603,-0.4534512758,0.0244151521,0.0525489785,0.036143139,0.1982194632,-0.0706402957,0.0925828964,-0.7637271285,0.0521496274,0.0224501938,0.2525706291,-0.2339808792,-0.037597999,-0.0533665158,-0.0019641947,0.6382448673,-0.122154586,-0.1083132252,-0.0036729013,-0.0728005245,-0.6327511072,-0.0034906983,-0.0141837699,0.0197430234,0.2299287319,-0.0092491293,-0.4394513369,-0.017309675,0.1475987136,0.2754671276,-0.3025964797,-0.0935151726,-0.3661640882,-0.1195315123,-0.5057441592,-0.1286639273,0.0648095906,0.2224610299,0.3846001923,0.3160492778,0.0997214988,0.0857022554,-0.0120870713,-0.0698755458,0.2582469881,0.1297812015,-0.1368017495,0.2899037004,0.0841262639,-0.2411785573,0.3528220952,0.0503834598,-0.2506101429,0.0791513175,0.1490257233,0.1158474162,0.2734086812,0.0989278853,-0.1237156764,0.1667004228,-0.0585143901,-0.4219202399,0.2355522513,0.2815838456,-0.1854799986,-0.3053144217,-0.5116246343,0.615860939,0.0559713878,0.0304511804,0.2086319327,0.0223995503,-0.412781179,0.4413424134,0.2683010101,0.9731247425,-0.204042688,0.424998045,0.2317784727,0.1344753057,0.5237073898,-0.5167378187,0.3831552863,-0.3056226373,0.1268451512,-0.0133372135,-0.1085634455,0.0364827737,0.114719294,0.0654074848,0.3253664672,0.2079365999,0.0961439908,-0.0773934275,0.4588253498,-0.0800674111,-0.3420723975,-0.1955794841,0.1220066622,-0.3448764682,0.3898961246,-0.3645066917,0.0217100196,0.0776721239,-0.1105872989,-0.4916018248,0.0795153826,-0.3250877857,0.4055930078,-0.3284801543,-0.1371369958,-0.1421707869,0.3159619272,0.0592573695,0.1933968067,-0.249840185,0.2242742479,-0.3976946771,-0.2356773764,0.0208289362,0.2337531894,0.4087526202,-0.2069119662,-0.2438365221,0.0980038419,-0.1904449761,-0.3769418299,-0.1676462144,-0.1064466536,0.0883559063,-0.014652296,-0.3172785044,-0.00905261,-0.1413644105,0.0560814254,0.0824531466,0.0137014808,-0.1196166202,-0.0904117227,0.0306326207,-0.1625874341,0.0537343882,0.4452787936,0.1993274242,-0.1737648696,0.6748771667,0.318600148,-0.2701166868,-0.0588436611,-0.145492211,0.026524296,-0.3916162848,0.0803605095,-0.1223210543,-0.0326891243,-0.2329388112,-0.0158797093,0.2639751136,-0.128901422,0.2503610849,-0.3973307312,-0.4390514493,0.3276472688,-0.1340383142,0.1806013137,0.2681422234,-0.3036365807,0.1483330727,-0.0007178176,-0.1766967624,-0.0336865559,-0.2832540572,0.1993031353,0.3011004329,-0.1552578509,0.0712757185,0.134473592,0.0281189606,0.1406504214,-0.167130813,-0.1209419593,-0.0215271469,0.1702215821,-0.0211845562,-0.1575117856,0.0079886634,-0.4215225875,0.1337191612,-0.3100504279,0.0949737802,0.0940818116,-0.0231312681,0.3049455881,0.1652727872,0.0589501262,-0.0955347419,0.4590955079,0.004906415,0.4881094098,0.1949438304,0.1460272819,0.21745345,0.0095139295,-0.529992342,0.0595963299,-0.1048070416,-0.0354540385,0.129529193,-0.1712080538,0.015818838,0.1867771298,-0.040361397,0.102782838,-0.1173781157,-0.0446845815,0.2029732615,0.1301267743,-0.129020974,0.1219610199,0.343000412,-0.038141951,-0.2553494275,0.1517674178,0.1166727766,-0.0218609311,-0.0463967025,0.138712272,0.3468563259,-0.0800423026,0.0823748708,-0.1870741248,0.0247108527,-0.0429122858,-0.1344307065,0.2051127702,0.1837250888,0.4019264877,0.0027958902,0.1415004283,-0.1637841016,0.075403668,-0.1910852641,-0.3289398253,0.0203157384,0.5687055588,-0.0469907708,0.2843696475,-0.1446273178,0.1780040264,0.3612075746,-0.1369031966,-0.367102474,0.2313219756,-0.1447919011,0.0617254116,-0.058048822,-0.2364676446,-0.2097277045,0.2906796634,0.0528513603,-0.4837180972,0.0807831958,0.1928961873,-0.1719204783,-0.2963343859,0.4201445282,0.1663333476,-0.0192703065,-0.3831961155,0.0990737677,0.1448647529,-0.0447314866,-0.0872007981,0.3097783327,0.4060985744,0.3071936667,-0.2704363167,0.0093551604,-0.0059454297,-0.0063552903,-0.2495992184,0.2559792399,0.4065111876,0.3059679568,0.1803972274,0.0541909225,-0.1025387272,-0.3101823628,0.3593687117,-0.0628880933,-0.1212561727,-0.0192109924,-0.2770347297,0.0351019092,-0.2277391851,0.0494226553,-0.5648443103,0.1031810939,0.2087474763,-0.132521525,0.1167968512,-0.0363964811,0.0014875523,-0.1657541692,0.5369609594,0.2511483133,0.1122913584,-0.4637848437,-0.3375482261,-0.5109270215,0.0780812204,-0.3646364808,0.0649989098,-0.0553256199,0.3731818795,-0.1717679948,0.1566798538,-0.2668590546,0.1098729745,-0.1133813411,0.2389042079,-0.3001950383,-0.2587818205,0.0470050648,0.0787234753,-0.0804847926,-0.2608900666,0.0874856189,-0.0274887383,-0.0548563041,-0.2284462303,-0.2025665939,0.0881140232,0.2646359205,0.2633498013,0.1222851872,0.5243082047,0.0749866292,-0.3782134652,-0.2976669967,-0.1673058867,0.0435902327,0.0065886723,-0.080036357,0.5198352337,-0.2667166889,0.1582237184,-0.4382742643,0.4756548703,-0.1714072227,0.1068449765,-0.3627738953,-0.1564667523,0.0345594808,0.226611346,0.2080149949,0.1816634089,-0.091755636,0.077336058,-0.2539073229,-0.1309581548,0.3351980746,-0.5924607515,-0.3291895986,0.0524134412,-0.058366973,-0.0152939847,-0.1152605638,-0.5271500349,0.0343527161,0.1688707769,0.0881754756,-0.1708844006,-0.0418434329,-0.0630329624,-0.0228900574,-0.0843612254,0.0717218816,-0.1209484562,-0.3793058991,0.1373226494,-0.2624127865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/278","title":"MemoryError when loading German Wikipedia","comments":"I added the German and the Italian Wikipedia to our google cloud storage:\r\nFirst update the `nlp` package to 0.3.0:\r\n```bash\r\npip install nlp --upgrade\r\n```\r\nand then\r\n```python\r\nfrom nlp import load_dataset\r\nwiki_de = load_dataset(\"wikipedia\", \"20200501.de\")\r\nwiki_it = load_dataset(\"wikipedia\", \"20200501.it\")\r\n```\r\nThe datasets are downloaded and directly ready to use (no processing).","body":"Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!","comment_length":53,"text":"MemoryError when loading German Wikipedia\nHi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!\nI added the German and the Italian Wikipedia to our google cloud storage:\r\nFirst update the `nlp` package to 0.3.0:\r\n```bash\r\npip install nlp --upgrade\r\n```\r\nand then\r\n```python\r\nfrom nlp import load_dataset\r\nwiki_de = load_dataset(\"wikipedia\", \"20200501.de\")\r\nwiki_it = load_dataset(\"wikipedia\", \"20200501.it\")\r\n```\r\nThe datasets are downloaded and directly ready to use (no processing).","embeddings":[-0.0479339696,0.1261895299,0.0182390139,0.6689180136,0.2445327044,0.1557507068,0.004935035,0.2083568722,0.3758642077,-0.0514769666,0.2673629224,-0.3312878907,-0.1734624952,-0.3341802955,0.2296689302,-0.4861274064,0.0930675492,0.0584275424,-0.0968580395,0.021348808,-0.2210392207,0.2842619419,-0.2860866785,-0.1296746433,-0.2414990813,0.0682763606,-0.0606940202,0.0540070944,-0.1613469124,-0.4543804526,0.1093896404,-0.2572765052,0.1181677878,0.1811023504,-0.0001217219,0.0400367081,0.4013991356,-0.1707563549,-0.4441632926,0.0392563045,-0.0454732403,-0.3407937586,0.0367344134,-0.2606716752,0.0609828457,-0.1065008789,0.3253657222,-0.082839556,0.3636316657,0.1792971939,0.1830790192,0.0583989583,0.2596977651,-0.0829991326,0.5757554173,0.0603940561,0.0517875366,0.1204346195,-0.3371914327,-0.3669273555,-0.266692996,0.2971111834,0.0168598853,0.1016100198,0.3428906202,-0.1183491275,0.014594771,-0.2151484787,0.2146662027,-0.0155843385,0.6984869242,-0.222539112,0.0544035435,-0.2987264991,-0.0819756612,0.1372263432,0.3699685335,0.3768196404,-0.53745538,-0.1752103865,-0.1816793531,-0.3489992321,-0.1634562314,0.3382711709,0.0580682494,0.2427105755,-0.046096243,0.165600881,0.3660611212,-0.0474638939,-0.2768139839,-0.1119219139,-0.0121806534,0.4462251961,-0.0702341348,0.0682196021,-0.0799727216,-0.0619321316,0.2351161987,-0.2792296708,-0.0940648913,-0.0614943281,0.2538709641,0.0773239955,0.2765713036,0.4570385516,-0.1349075586,-0.1987938285,0.1601509452,0.1806961447,0.0185495038,-0.1625983566,-0.0348524787,-0.163364619,-0.0773852915,-0.0258760657,-0.103963092,-0.0469388105,0.0200344175,0.1309570819,-0.096364513,-0.133965537,-0.0321695395,0.4189065993,-0.1397236735,0.3382306099,0.2956480384,0.0687169135,-0.384778291,-0.2950851619,-0.1493525356,0.1582213342,-0.5230259895,0.1237957925,0.2662939131,0.2771739364,0.3613065481,-0.1229094267,-0.2683769166,0.0498335995,0.2657951117,-0.1831974089,-0.0323753022,0.2156431675,0.14843674,0.2886892855,0.1391697973,-0.2281753868,0.000194584,0.0619199313,-0.1770699173,-0.1940729618,0.0692587718,0.0176814776,-0.1571019292,0.154286474,-0.4031710923,0.2367206365,0.2888612151,-0.2128570974,-0.0976049975,-0.0422633328,-0.23210226,-0.2615077198,0.2098641694,0.5494705439,-0.0914283767,-0.2689732015,-0.0759014413,-0.0987920538,0.4090284109,0.2375489473,-0.1052161306,0.2960037291,-0.2315713167,0.0987135246,0.3416324556,-0.1739220619,-0.331951797,0.1309767216,0.1102416217,0.1638558507,0.0562583506,0.0016745509,0.0118619101,0.1984750479,0.1997092515,0.3413085043,-0.0352610424,0.2907262743,-0.267072022,-0.2709881663,0.2600758076,0.0086501343,0.3516753912,0.0555153303,0.0945201367,0.8308191299,0.2724248767,-0.0689333975,0.3318047822,0.5638161898,-0.227718845,0.0573095493,-0.0112689119,-0.0633697733,-0.2486257851,0.1927720159,-0.3067270517,0.3094541728,-0.0214819368,0.1321331114,-0.1640296876,-0.0969142392,-0.0566746257,-0.3699617386,0.0888292938,0.1652462482,0.1101260856,0.2140970677,0.1379154623,0.2627291083,-0.0111610973,0.0226750709,-0.6933624744,0.2087740302,-0.1417605281,-0.1521971226,-0.0109318141,0.1096379384,0.1333905607,0.1438275576,-0.1323129535,0.1214753836,0.0990371704,-0.0469947085,0.1362223029,-0.0250016525,0.2658998072,-0.1880667061,0.4338091016,0.0586686023,0.1608509421,-0.1543041915,0.0469034724,-0.2138385028,0.072046563,0.2993028164,0.0846952274,0.0347199701,0.1672719419,0.1503384262,0.1518827975,-0.1656813025,0.4092136621,0.0806189775,0.1114226282,-0.0011588708,0.0266635604,-0.4050886929,0.4183508158,0.1737964302,0.1427270025,0.2619020641,-0.4456169605,-0.2904942334,0.2247490287,-0.057462208,0.2343137711,0.0811882243,0.1338569522,0.0289964098,0.1593956351,0.037909966,0.0936210901,0.058900822,0.5227373838,0.0115124211,0.0125517687,-0.1341618747,-0.2134275883,0.0808712468,0.0312990136,0.4169813991,-0.1902172863,-0.0390307494,-0.0026984455,-0.5402334332,-0.3053193986,0.2616876364,-0.4882035851,-0.2883309424,-0.2413621992,0.280143708,-0.0662873089,-0.0134526268,0.068596065,0.1989407092,0.1088028774,-0.1494421959,-0.1298904121,-0.2826367915,-0.3521508873,-0.0132163968,0.1780460626,0.0457917042,0.2574231327,0.0474953204,-0.129376784,-0.2252615541,-0.1390877813,0.193174839,0.0193976611,0.1323526949,0.0808379054,0.6441944838,-0.0916180611,-0.1589263827,0.3123699129,-0.1345023215,-0.1123504639,0.2058447748,0.0021174341,0.1894079149,0.0238474421,-0.0355216302,-0.1239717528,-0.4503405094,0.1686739624,0.1654811502,-0.0202529617,0.347874105,0.1282710731,0.0782882795,0.1775042415,0.1381355822,-0.190916121,-0.0510670096,0.2469130903,-0.3309720457,-0.3059126437,-0.0715213791,-0.1601582468,0.2242604643,0.4027511775,-0.6984960437,0.1061492711,-0.1276174188,-0.2190024704,0.0490920655,0.0372534879,0.3683996797,-0.1916541755,-0.0193818528,-0.0021286565,-0.0695347115,-0.0335374884,-0.2112904936,0.6732001901,0.1878428757,0.3525752127,0.2811793983,0.9196683168,0.2102591395,0.4203176796,0.4592570961,0.1845643818,0.2118268609,-0.1392084062,-0.2165400535,0.0542722419,-0.2958499491,-0.1469565779,0.1930077374,0.0539988689,-0.5334852934,-0.336306721,-0.2229558825,-0.1735776961,-0.3317177892,0.3225135505,-0.0072969436,0.3478627801,-0.0577862076,-0.0329394154,-0.2737751603,-0.4534512758,0.0244151521,0.0525489785,0.036143139,0.1982194632,-0.0706402957,0.0925828964,-0.7637271285,0.0521496274,0.0224501938,0.2525706291,-0.2339808792,-0.037597999,-0.0533665158,-0.0019641947,0.6382448673,-0.122154586,-0.1083132252,-0.0036729013,-0.0728005245,-0.6327511072,-0.0034906983,-0.0141837699,0.0197430234,0.2299287319,-0.0092491293,-0.4394513369,-0.017309675,0.1475987136,0.2754671276,-0.3025964797,-0.0935151726,-0.3661640882,-0.1195315123,-0.5057441592,-0.1286639273,0.0648095906,0.2224610299,0.3846001923,0.3160492778,0.0997214988,0.0857022554,-0.0120870713,-0.0698755458,0.2582469881,0.1297812015,-0.1368017495,0.2899037004,0.0841262639,-0.2411785573,0.3528220952,0.0503834598,-0.2506101429,0.0791513175,0.1490257233,0.1158474162,0.2734086812,0.0989278853,-0.1237156764,0.1667004228,-0.0585143901,-0.4219202399,0.2355522513,0.2815838456,-0.1854799986,-0.3053144217,-0.5116246343,0.615860939,0.0559713878,0.0304511804,0.2086319327,0.0223995503,-0.412781179,0.4413424134,0.2683010101,0.9731247425,-0.204042688,0.424998045,0.2317784727,0.1344753057,0.5237073898,-0.5167378187,0.3831552863,-0.3056226373,0.1268451512,-0.0133372135,-0.1085634455,0.0364827737,0.114719294,0.0654074848,0.3253664672,0.2079365999,0.0961439908,-0.0773934275,0.4588253498,-0.0800674111,-0.3420723975,-0.1955794841,0.1220066622,-0.3448764682,0.3898961246,-0.3645066917,0.0217100196,0.0776721239,-0.1105872989,-0.4916018248,0.0795153826,-0.3250877857,0.4055930078,-0.3284801543,-0.1371369958,-0.1421707869,0.3159619272,0.0592573695,0.1933968067,-0.249840185,0.2242742479,-0.3976946771,-0.2356773764,0.0208289362,0.2337531894,0.4087526202,-0.2069119662,-0.2438365221,0.0980038419,-0.1904449761,-0.3769418299,-0.1676462144,-0.1064466536,0.0883559063,-0.014652296,-0.3172785044,-0.00905261,-0.1413644105,0.0560814254,0.0824531466,0.0137014808,-0.1196166202,-0.0904117227,0.0306326207,-0.1625874341,0.0537343882,0.4452787936,0.1993274242,-0.1737648696,0.6748771667,0.318600148,-0.2701166868,-0.0588436611,-0.145492211,0.026524296,-0.3916162848,0.0803605095,-0.1223210543,-0.0326891243,-0.2329388112,-0.0158797093,0.2639751136,-0.128901422,0.2503610849,-0.3973307312,-0.4390514493,0.3276472688,-0.1340383142,0.1806013137,0.2681422234,-0.3036365807,0.1483330727,-0.0007178176,-0.1766967624,-0.0336865559,-0.2832540572,0.1993031353,0.3011004329,-0.1552578509,0.0712757185,0.134473592,0.0281189606,0.1406504214,-0.167130813,-0.1209419593,-0.0215271469,0.1702215821,-0.0211845562,-0.1575117856,0.0079886634,-0.4215225875,0.1337191612,-0.3100504279,0.0949737802,0.0940818116,-0.0231312681,0.3049455881,0.1652727872,0.0589501262,-0.0955347419,0.4590955079,0.004906415,0.4881094098,0.1949438304,0.1460272819,0.21745345,0.0095139295,-0.529992342,0.0595963299,-0.1048070416,-0.0354540385,0.129529193,-0.1712080538,0.015818838,0.1867771298,-0.040361397,0.102782838,-0.1173781157,-0.0446845815,0.2029732615,0.1301267743,-0.129020974,0.1219610199,0.343000412,-0.038141951,-0.2553494275,0.1517674178,0.1166727766,-0.0218609311,-0.0463967025,0.138712272,0.3468563259,-0.0800423026,0.0823748708,-0.1870741248,0.0247108527,-0.0429122858,-0.1344307065,0.2051127702,0.1837250888,0.4019264877,0.0027958902,0.1415004283,-0.1637841016,0.075403668,-0.1910852641,-0.3289398253,0.0203157384,0.5687055588,-0.0469907708,0.2843696475,-0.1446273178,0.1780040264,0.3612075746,-0.1369031966,-0.367102474,0.2313219756,-0.1447919011,0.0617254116,-0.058048822,-0.2364676446,-0.2097277045,0.2906796634,0.0528513603,-0.4837180972,0.0807831958,0.1928961873,-0.1719204783,-0.2963343859,0.4201445282,0.1663333476,-0.0192703065,-0.3831961155,0.0990737677,0.1448647529,-0.0447314866,-0.0872007981,0.3097783327,0.4060985744,0.3071936667,-0.2704363167,0.0093551604,-0.0059454297,-0.0063552903,-0.2495992184,0.2559792399,0.4065111876,0.3059679568,0.1803972274,0.0541909225,-0.1025387272,-0.3101823628,0.3593687117,-0.0628880933,-0.1212561727,-0.0192109924,-0.2770347297,0.0351019092,-0.2277391851,0.0494226553,-0.5648443103,0.1031810939,0.2087474763,-0.132521525,0.1167968512,-0.0363964811,0.0014875523,-0.1657541692,0.5369609594,0.2511483133,0.1122913584,-0.4637848437,-0.3375482261,-0.5109270215,0.0780812204,-0.3646364808,0.0649989098,-0.0553256199,0.3731818795,-0.1717679948,0.1566798538,-0.2668590546,0.1098729745,-0.1133813411,0.2389042079,-0.3001950383,-0.2587818205,0.0470050648,0.0787234753,-0.0804847926,-0.2608900666,0.0874856189,-0.0274887383,-0.0548563041,-0.2284462303,-0.2025665939,0.0881140232,0.2646359205,0.2633498013,0.1222851872,0.5243082047,0.0749866292,-0.3782134652,-0.2976669967,-0.1673058867,0.0435902327,0.0065886723,-0.080036357,0.5198352337,-0.2667166889,0.1582237184,-0.4382742643,0.4756548703,-0.1714072227,0.1068449765,-0.3627738953,-0.1564667523,0.0345594808,0.226611346,0.2080149949,0.1816634089,-0.091755636,0.077336058,-0.2539073229,-0.1309581548,0.3351980746,-0.5924607515,-0.3291895986,0.0524134412,-0.058366973,-0.0152939847,-0.1152605638,-0.5271500349,0.0343527161,0.1688707769,0.0881754756,-0.1708844006,-0.0418434329,-0.0630329624,-0.0228900574,-0.0843612254,0.0717218816,-0.1209484562,-0.3793058991,0.1373226494,-0.2624127865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/278","title":"MemoryError when loading German Wikipedia","comments":"Hi @lhoestq \r\n\r\nWow, thanks so much, that's **really** incredible! I was considering looking at creating my own Beam Dataset, as per the doc you linked, but instead opted to process the data myself using `wikiextractor`. However, now that this is available, I'll definitely switch across and use it.\r\n\r\nThanks so much for the incredible work, this really helps out our team considerably!\r\n\r\nHave a great (and well-deserved ;) weekend ahead!\r\n\r\nP.S. I'm not sure if I should close the issue here - if so I'm happy to do so.","body":"Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!","comment_length":89,"text":"MemoryError when loading German Wikipedia\nHi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!\nHi @lhoestq \r\n\r\nWow, thanks so much, that's **really** incredible! I was considering looking at creating my own Beam Dataset, as per the doc you linked, but instead opted to process the data myself using `wikiextractor`. However, now that this is available, I'll definitely switch across and use it.\r\n\r\nThanks so much for the incredible work, this really helps out our team considerably!\r\n\r\nHave a great (and well-deserved ;) weekend ahead!\r\n\r\nP.S. I'm not sure if I should close the issue here - if so I'm happy to do so.","embeddings":[-0.0479339696,0.1261895299,0.0182390139,0.6689180136,0.2445327044,0.1557507068,0.004935035,0.2083568722,0.3758642077,-0.0514769666,0.2673629224,-0.3312878907,-0.1734624952,-0.3341802955,0.2296689302,-0.4861274064,0.0930675492,0.0584275424,-0.0968580395,0.021348808,-0.2210392207,0.2842619419,-0.2860866785,-0.1296746433,-0.2414990813,0.0682763606,-0.0606940202,0.0540070944,-0.1613469124,-0.4543804526,0.1093896404,-0.2572765052,0.1181677878,0.1811023504,-0.0001217219,0.0400367081,0.4013991356,-0.1707563549,-0.4441632926,0.0392563045,-0.0454732403,-0.3407937586,0.0367344134,-0.2606716752,0.0609828457,-0.1065008789,0.3253657222,-0.082839556,0.3636316657,0.1792971939,0.1830790192,0.0583989583,0.2596977651,-0.0829991326,0.5757554173,0.0603940561,0.0517875366,0.1204346195,-0.3371914327,-0.3669273555,-0.266692996,0.2971111834,0.0168598853,0.1016100198,0.3428906202,-0.1183491275,0.014594771,-0.2151484787,0.2146662027,-0.0155843385,0.6984869242,-0.222539112,0.0544035435,-0.2987264991,-0.0819756612,0.1372263432,0.3699685335,0.3768196404,-0.53745538,-0.1752103865,-0.1816793531,-0.3489992321,-0.1634562314,0.3382711709,0.0580682494,0.2427105755,-0.046096243,0.165600881,0.3660611212,-0.0474638939,-0.2768139839,-0.1119219139,-0.0121806534,0.4462251961,-0.0702341348,0.0682196021,-0.0799727216,-0.0619321316,0.2351161987,-0.2792296708,-0.0940648913,-0.0614943281,0.2538709641,0.0773239955,0.2765713036,0.4570385516,-0.1349075586,-0.1987938285,0.1601509452,0.1806961447,0.0185495038,-0.1625983566,-0.0348524787,-0.163364619,-0.0773852915,-0.0258760657,-0.103963092,-0.0469388105,0.0200344175,0.1309570819,-0.096364513,-0.133965537,-0.0321695395,0.4189065993,-0.1397236735,0.3382306099,0.2956480384,0.0687169135,-0.384778291,-0.2950851619,-0.1493525356,0.1582213342,-0.5230259895,0.1237957925,0.2662939131,0.2771739364,0.3613065481,-0.1229094267,-0.2683769166,0.0498335995,0.2657951117,-0.1831974089,-0.0323753022,0.2156431675,0.14843674,0.2886892855,0.1391697973,-0.2281753868,0.000194584,0.0619199313,-0.1770699173,-0.1940729618,0.0692587718,0.0176814776,-0.1571019292,0.154286474,-0.4031710923,0.2367206365,0.2888612151,-0.2128570974,-0.0976049975,-0.0422633328,-0.23210226,-0.2615077198,0.2098641694,0.5494705439,-0.0914283767,-0.2689732015,-0.0759014413,-0.0987920538,0.4090284109,0.2375489473,-0.1052161306,0.2960037291,-0.2315713167,0.0987135246,0.3416324556,-0.1739220619,-0.331951797,0.1309767216,0.1102416217,0.1638558507,0.0562583506,0.0016745509,0.0118619101,0.1984750479,0.1997092515,0.3413085043,-0.0352610424,0.2907262743,-0.267072022,-0.2709881663,0.2600758076,0.0086501343,0.3516753912,0.0555153303,0.0945201367,0.8308191299,0.2724248767,-0.0689333975,0.3318047822,0.5638161898,-0.227718845,0.0573095493,-0.0112689119,-0.0633697733,-0.2486257851,0.1927720159,-0.3067270517,0.3094541728,-0.0214819368,0.1321331114,-0.1640296876,-0.0969142392,-0.0566746257,-0.3699617386,0.0888292938,0.1652462482,0.1101260856,0.2140970677,0.1379154623,0.2627291083,-0.0111610973,0.0226750709,-0.6933624744,0.2087740302,-0.1417605281,-0.1521971226,-0.0109318141,0.1096379384,0.1333905607,0.1438275576,-0.1323129535,0.1214753836,0.0990371704,-0.0469947085,0.1362223029,-0.0250016525,0.2658998072,-0.1880667061,0.4338091016,0.0586686023,0.1608509421,-0.1543041915,0.0469034724,-0.2138385028,0.072046563,0.2993028164,0.0846952274,0.0347199701,0.1672719419,0.1503384262,0.1518827975,-0.1656813025,0.4092136621,0.0806189775,0.1114226282,-0.0011588708,0.0266635604,-0.4050886929,0.4183508158,0.1737964302,0.1427270025,0.2619020641,-0.4456169605,-0.2904942334,0.2247490287,-0.057462208,0.2343137711,0.0811882243,0.1338569522,0.0289964098,0.1593956351,0.037909966,0.0936210901,0.058900822,0.5227373838,0.0115124211,0.0125517687,-0.1341618747,-0.2134275883,0.0808712468,0.0312990136,0.4169813991,-0.1902172863,-0.0390307494,-0.0026984455,-0.5402334332,-0.3053193986,0.2616876364,-0.4882035851,-0.2883309424,-0.2413621992,0.280143708,-0.0662873089,-0.0134526268,0.068596065,0.1989407092,0.1088028774,-0.1494421959,-0.1298904121,-0.2826367915,-0.3521508873,-0.0132163968,0.1780460626,0.0457917042,0.2574231327,0.0474953204,-0.129376784,-0.2252615541,-0.1390877813,0.193174839,0.0193976611,0.1323526949,0.0808379054,0.6441944838,-0.0916180611,-0.1589263827,0.3123699129,-0.1345023215,-0.1123504639,0.2058447748,0.0021174341,0.1894079149,0.0238474421,-0.0355216302,-0.1239717528,-0.4503405094,0.1686739624,0.1654811502,-0.0202529617,0.347874105,0.1282710731,0.0782882795,0.1775042415,0.1381355822,-0.190916121,-0.0510670096,0.2469130903,-0.3309720457,-0.3059126437,-0.0715213791,-0.1601582468,0.2242604643,0.4027511775,-0.6984960437,0.1061492711,-0.1276174188,-0.2190024704,0.0490920655,0.0372534879,0.3683996797,-0.1916541755,-0.0193818528,-0.0021286565,-0.0695347115,-0.0335374884,-0.2112904936,0.6732001901,0.1878428757,0.3525752127,0.2811793983,0.9196683168,0.2102591395,0.4203176796,0.4592570961,0.1845643818,0.2118268609,-0.1392084062,-0.2165400535,0.0542722419,-0.2958499491,-0.1469565779,0.1930077374,0.0539988689,-0.5334852934,-0.336306721,-0.2229558825,-0.1735776961,-0.3317177892,0.3225135505,-0.0072969436,0.3478627801,-0.0577862076,-0.0329394154,-0.2737751603,-0.4534512758,0.0244151521,0.0525489785,0.036143139,0.1982194632,-0.0706402957,0.0925828964,-0.7637271285,0.0521496274,0.0224501938,0.2525706291,-0.2339808792,-0.037597999,-0.0533665158,-0.0019641947,0.6382448673,-0.122154586,-0.1083132252,-0.0036729013,-0.0728005245,-0.6327511072,-0.0034906983,-0.0141837699,0.0197430234,0.2299287319,-0.0092491293,-0.4394513369,-0.017309675,0.1475987136,0.2754671276,-0.3025964797,-0.0935151726,-0.3661640882,-0.1195315123,-0.5057441592,-0.1286639273,0.0648095906,0.2224610299,0.3846001923,0.3160492778,0.0997214988,0.0857022554,-0.0120870713,-0.0698755458,0.2582469881,0.1297812015,-0.1368017495,0.2899037004,0.0841262639,-0.2411785573,0.3528220952,0.0503834598,-0.2506101429,0.0791513175,0.1490257233,0.1158474162,0.2734086812,0.0989278853,-0.1237156764,0.1667004228,-0.0585143901,-0.4219202399,0.2355522513,0.2815838456,-0.1854799986,-0.3053144217,-0.5116246343,0.615860939,0.0559713878,0.0304511804,0.2086319327,0.0223995503,-0.412781179,0.4413424134,0.2683010101,0.9731247425,-0.204042688,0.424998045,0.2317784727,0.1344753057,0.5237073898,-0.5167378187,0.3831552863,-0.3056226373,0.1268451512,-0.0133372135,-0.1085634455,0.0364827737,0.114719294,0.0654074848,0.3253664672,0.2079365999,0.0961439908,-0.0773934275,0.4588253498,-0.0800674111,-0.3420723975,-0.1955794841,0.1220066622,-0.3448764682,0.3898961246,-0.3645066917,0.0217100196,0.0776721239,-0.1105872989,-0.4916018248,0.0795153826,-0.3250877857,0.4055930078,-0.3284801543,-0.1371369958,-0.1421707869,0.3159619272,0.0592573695,0.1933968067,-0.249840185,0.2242742479,-0.3976946771,-0.2356773764,0.0208289362,0.2337531894,0.4087526202,-0.2069119662,-0.2438365221,0.0980038419,-0.1904449761,-0.3769418299,-0.1676462144,-0.1064466536,0.0883559063,-0.014652296,-0.3172785044,-0.00905261,-0.1413644105,0.0560814254,0.0824531466,0.0137014808,-0.1196166202,-0.0904117227,0.0306326207,-0.1625874341,0.0537343882,0.4452787936,0.1993274242,-0.1737648696,0.6748771667,0.318600148,-0.2701166868,-0.0588436611,-0.145492211,0.026524296,-0.3916162848,0.0803605095,-0.1223210543,-0.0326891243,-0.2329388112,-0.0158797093,0.2639751136,-0.128901422,0.2503610849,-0.3973307312,-0.4390514493,0.3276472688,-0.1340383142,0.1806013137,0.2681422234,-0.3036365807,0.1483330727,-0.0007178176,-0.1766967624,-0.0336865559,-0.2832540572,0.1993031353,0.3011004329,-0.1552578509,0.0712757185,0.134473592,0.0281189606,0.1406504214,-0.167130813,-0.1209419593,-0.0215271469,0.1702215821,-0.0211845562,-0.1575117856,0.0079886634,-0.4215225875,0.1337191612,-0.3100504279,0.0949737802,0.0940818116,-0.0231312681,0.3049455881,0.1652727872,0.0589501262,-0.0955347419,0.4590955079,0.004906415,0.4881094098,0.1949438304,0.1460272819,0.21745345,0.0095139295,-0.529992342,0.0595963299,-0.1048070416,-0.0354540385,0.129529193,-0.1712080538,0.015818838,0.1867771298,-0.040361397,0.102782838,-0.1173781157,-0.0446845815,0.2029732615,0.1301267743,-0.129020974,0.1219610199,0.343000412,-0.038141951,-0.2553494275,0.1517674178,0.1166727766,-0.0218609311,-0.0463967025,0.138712272,0.3468563259,-0.0800423026,0.0823748708,-0.1870741248,0.0247108527,-0.0429122858,-0.1344307065,0.2051127702,0.1837250888,0.4019264877,0.0027958902,0.1415004283,-0.1637841016,0.075403668,-0.1910852641,-0.3289398253,0.0203157384,0.5687055588,-0.0469907708,0.2843696475,-0.1446273178,0.1780040264,0.3612075746,-0.1369031966,-0.367102474,0.2313219756,-0.1447919011,0.0617254116,-0.058048822,-0.2364676446,-0.2097277045,0.2906796634,0.0528513603,-0.4837180972,0.0807831958,0.1928961873,-0.1719204783,-0.2963343859,0.4201445282,0.1663333476,-0.0192703065,-0.3831961155,0.0990737677,0.1448647529,-0.0447314866,-0.0872007981,0.3097783327,0.4060985744,0.3071936667,-0.2704363167,0.0093551604,-0.0059454297,-0.0063552903,-0.2495992184,0.2559792399,0.4065111876,0.3059679568,0.1803972274,0.0541909225,-0.1025387272,-0.3101823628,0.3593687117,-0.0628880933,-0.1212561727,-0.0192109924,-0.2770347297,0.0351019092,-0.2277391851,0.0494226553,-0.5648443103,0.1031810939,0.2087474763,-0.132521525,0.1167968512,-0.0363964811,0.0014875523,-0.1657541692,0.5369609594,0.2511483133,0.1122913584,-0.4637848437,-0.3375482261,-0.5109270215,0.0780812204,-0.3646364808,0.0649989098,-0.0553256199,0.3731818795,-0.1717679948,0.1566798538,-0.2668590546,0.1098729745,-0.1133813411,0.2389042079,-0.3001950383,-0.2587818205,0.0470050648,0.0787234753,-0.0804847926,-0.2608900666,0.0874856189,-0.0274887383,-0.0548563041,-0.2284462303,-0.2025665939,0.0881140232,0.2646359205,0.2633498013,0.1222851872,0.5243082047,0.0749866292,-0.3782134652,-0.2976669967,-0.1673058867,0.0435902327,0.0065886723,-0.080036357,0.5198352337,-0.2667166889,0.1582237184,-0.4382742643,0.4756548703,-0.1714072227,0.1068449765,-0.3627738953,-0.1564667523,0.0345594808,0.226611346,0.2080149949,0.1816634089,-0.091755636,0.077336058,-0.2539073229,-0.1309581548,0.3351980746,-0.5924607515,-0.3291895986,0.0524134412,-0.058366973,-0.0152939847,-0.1152605638,-0.5271500349,0.0343527161,0.1688707769,0.0881754756,-0.1708844006,-0.0418434329,-0.0630329624,-0.0228900574,-0.0843612254,0.0717218816,-0.1209484562,-0.3793058991,0.1373226494,-0.2624127865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/278","title":"MemoryError when loading German Wikipedia","comments":"Thanks for your message, glad I could help :)\r\nClosing this one.","body":"Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!","comment_length":12,"text":"MemoryError when loading German Wikipedia\nHi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :)\r\n\r\nI'm trying to download the German Wikipedia dataset as follows:\r\n\r\n```\r\nwiki = nlp.load_dataset(\"wikipedia\", \"20200501.de\", split=\"train\")\r\n```\r\n\r\nHowever, when I do so, I get the following error:\r\n\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.de (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/ubuntu\/.cache\/huggingface\/datasets\/wikipedia\/20200501.de\/1.0.0...\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 433, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/ubuntu\/anaconda3\/envs\/albert\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 824, in _download_and_prepare\r\n \"\\n\\t`{}`\".format(usage_example)\r\nnlp.builder.MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https:\/\/beam.apache.org\/documentation\/runners\/capability-matrix\/\r\nIf you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). \r\nExample of usage: \r\n\t`load_dataset('wikipedia', '20200501.de', beam_runner='DirectRunner')`\r\n```\r\n\r\nSo, following on from the example usage at the bottom, I tried specifying `beam_runner='DirectRunner`, however when I do this after about 20 min after the data has all downloaded, I get a `MemoryError` as warned.\r\n\r\nThis isn't an issue for the English or French Wikipedia datasets (I've tried both), as neither seem to require that `beam_runner` be specified. Can you please clarify why this is an issue for the German dataset?\r\n\r\nMy nlp version is 0.2.1.\r\n\r\nThank you!\nThanks for your message, glad I could help :)\r\nClosing this one.","embeddings":[-0.0479339696,0.1261895299,0.0182390139,0.6689180136,0.2445327044,0.1557507068,0.004935035,0.2083568722,0.3758642077,-0.0514769666,0.2673629224,-0.3312878907,-0.1734624952,-0.3341802955,0.2296689302,-0.4861274064,0.0930675492,0.0584275424,-0.0968580395,0.021348808,-0.2210392207,0.2842619419,-0.2860866785,-0.1296746433,-0.2414990813,0.0682763606,-0.0606940202,0.0540070944,-0.1613469124,-0.4543804526,0.1093896404,-0.2572765052,0.1181677878,0.1811023504,-0.0001217219,0.0400367081,0.4013991356,-0.1707563549,-0.4441632926,0.0392563045,-0.0454732403,-0.3407937586,0.0367344134,-0.2606716752,0.0609828457,-0.1065008789,0.3253657222,-0.082839556,0.3636316657,0.1792971939,0.1830790192,0.0583989583,0.2596977651,-0.0829991326,0.5757554173,0.0603940561,0.0517875366,0.1204346195,-0.3371914327,-0.3669273555,-0.266692996,0.2971111834,0.0168598853,0.1016100198,0.3428906202,-0.1183491275,0.014594771,-0.2151484787,0.2146662027,-0.0155843385,0.6984869242,-0.222539112,0.0544035435,-0.2987264991,-0.0819756612,0.1372263432,0.3699685335,0.3768196404,-0.53745538,-0.1752103865,-0.1816793531,-0.3489992321,-0.1634562314,0.3382711709,0.0580682494,0.2427105755,-0.046096243,0.165600881,0.3660611212,-0.0474638939,-0.2768139839,-0.1119219139,-0.0121806534,0.4462251961,-0.0702341348,0.0682196021,-0.0799727216,-0.0619321316,0.2351161987,-0.2792296708,-0.0940648913,-0.0614943281,0.2538709641,0.0773239955,0.2765713036,0.4570385516,-0.1349075586,-0.1987938285,0.1601509452,0.1806961447,0.0185495038,-0.1625983566,-0.0348524787,-0.163364619,-0.0773852915,-0.0258760657,-0.103963092,-0.0469388105,0.0200344175,0.1309570819,-0.096364513,-0.133965537,-0.0321695395,0.4189065993,-0.1397236735,0.3382306099,0.2956480384,0.0687169135,-0.384778291,-0.2950851619,-0.1493525356,0.1582213342,-0.5230259895,0.1237957925,0.2662939131,0.2771739364,0.3613065481,-0.1229094267,-0.2683769166,0.0498335995,0.2657951117,-0.1831974089,-0.0323753022,0.2156431675,0.14843674,0.2886892855,0.1391697973,-0.2281753868,0.000194584,0.0619199313,-0.1770699173,-0.1940729618,0.0692587718,0.0176814776,-0.1571019292,0.154286474,-0.4031710923,0.2367206365,0.2888612151,-0.2128570974,-0.0976049975,-0.0422633328,-0.23210226,-0.2615077198,0.2098641694,0.5494705439,-0.0914283767,-0.2689732015,-0.0759014413,-0.0987920538,0.4090284109,0.2375489473,-0.1052161306,0.2960037291,-0.2315713167,0.0987135246,0.3416324556,-0.1739220619,-0.331951797,0.1309767216,0.1102416217,0.1638558507,0.0562583506,0.0016745509,0.0118619101,0.1984750479,0.1997092515,0.3413085043,-0.0352610424,0.2907262743,-0.267072022,-0.2709881663,0.2600758076,0.0086501343,0.3516753912,0.0555153303,0.0945201367,0.8308191299,0.2724248767,-0.0689333975,0.3318047822,0.5638161898,-0.227718845,0.0573095493,-0.0112689119,-0.0633697733,-0.2486257851,0.1927720159,-0.3067270517,0.3094541728,-0.0214819368,0.1321331114,-0.1640296876,-0.0969142392,-0.0566746257,-0.3699617386,0.0888292938,0.1652462482,0.1101260856,0.2140970677,0.1379154623,0.2627291083,-0.0111610973,0.0226750709,-0.6933624744,0.2087740302,-0.1417605281,-0.1521971226,-0.0109318141,0.1096379384,0.1333905607,0.1438275576,-0.1323129535,0.1214753836,0.0990371704,-0.0469947085,0.1362223029,-0.0250016525,0.2658998072,-0.1880667061,0.4338091016,0.0586686023,0.1608509421,-0.1543041915,0.0469034724,-0.2138385028,0.072046563,0.2993028164,0.0846952274,0.0347199701,0.1672719419,0.1503384262,0.1518827975,-0.1656813025,0.4092136621,0.0806189775,0.1114226282,-0.0011588708,0.0266635604,-0.4050886929,0.4183508158,0.1737964302,0.1427270025,0.2619020641,-0.4456169605,-0.2904942334,0.2247490287,-0.057462208,0.2343137711,0.0811882243,0.1338569522,0.0289964098,0.1593956351,0.037909966,0.0936210901,0.058900822,0.5227373838,0.0115124211,0.0125517687,-0.1341618747,-0.2134275883,0.0808712468,0.0312990136,0.4169813991,-0.1902172863,-0.0390307494,-0.0026984455,-0.5402334332,-0.3053193986,0.2616876364,-0.4882035851,-0.2883309424,-0.2413621992,0.280143708,-0.0662873089,-0.0134526268,0.068596065,0.1989407092,0.1088028774,-0.1494421959,-0.1298904121,-0.2826367915,-0.3521508873,-0.0132163968,0.1780460626,0.0457917042,0.2574231327,0.0474953204,-0.129376784,-0.2252615541,-0.1390877813,0.193174839,0.0193976611,0.1323526949,0.0808379054,0.6441944838,-0.0916180611,-0.1589263827,0.3123699129,-0.1345023215,-0.1123504639,0.2058447748,0.0021174341,0.1894079149,0.0238474421,-0.0355216302,-0.1239717528,-0.4503405094,0.1686739624,0.1654811502,-0.0202529617,0.347874105,0.1282710731,0.0782882795,0.1775042415,0.1381355822,-0.190916121,-0.0510670096,0.2469130903,-0.3309720457,-0.3059126437,-0.0715213791,-0.1601582468,0.2242604643,0.4027511775,-0.6984960437,0.1061492711,-0.1276174188,-0.2190024704,0.0490920655,0.0372534879,0.3683996797,-0.1916541755,-0.0193818528,-0.0021286565,-0.0695347115,-0.0335374884,-0.2112904936,0.6732001901,0.1878428757,0.3525752127,0.2811793983,0.9196683168,0.2102591395,0.4203176796,0.4592570961,0.1845643818,0.2118268609,-0.1392084062,-0.2165400535,0.0542722419,-0.2958499491,-0.1469565779,0.1930077374,0.0539988689,-0.5334852934,-0.336306721,-0.2229558825,-0.1735776961,-0.3317177892,0.3225135505,-0.0072969436,0.3478627801,-0.0577862076,-0.0329394154,-0.2737751603,-0.4534512758,0.0244151521,0.0525489785,0.036143139,0.1982194632,-0.0706402957,0.0925828964,-0.7637271285,0.0521496274,0.0224501938,0.2525706291,-0.2339808792,-0.037597999,-0.0533665158,-0.0019641947,0.6382448673,-0.122154586,-0.1083132252,-0.0036729013,-0.0728005245,-0.6327511072,-0.0034906983,-0.0141837699,0.0197430234,0.2299287319,-0.0092491293,-0.4394513369,-0.017309675,0.1475987136,0.2754671276,-0.3025964797,-0.0935151726,-0.3661640882,-0.1195315123,-0.5057441592,-0.1286639273,0.0648095906,0.2224610299,0.3846001923,0.3160492778,0.0997214988,0.0857022554,-0.0120870713,-0.0698755458,0.2582469881,0.1297812015,-0.1368017495,0.2899037004,0.0841262639,-0.2411785573,0.3528220952,0.0503834598,-0.2506101429,0.0791513175,0.1490257233,0.1158474162,0.2734086812,0.0989278853,-0.1237156764,0.1667004228,-0.0585143901,-0.4219202399,0.2355522513,0.2815838456,-0.1854799986,-0.3053144217,-0.5116246343,0.615860939,0.0559713878,0.0304511804,0.2086319327,0.0223995503,-0.412781179,0.4413424134,0.2683010101,0.9731247425,-0.204042688,0.424998045,0.2317784727,0.1344753057,0.5237073898,-0.5167378187,0.3831552863,-0.3056226373,0.1268451512,-0.0133372135,-0.1085634455,0.0364827737,0.114719294,0.0654074848,0.3253664672,0.2079365999,0.0961439908,-0.0773934275,0.4588253498,-0.0800674111,-0.3420723975,-0.1955794841,0.1220066622,-0.3448764682,0.3898961246,-0.3645066917,0.0217100196,0.0776721239,-0.1105872989,-0.4916018248,0.0795153826,-0.3250877857,0.4055930078,-0.3284801543,-0.1371369958,-0.1421707869,0.3159619272,0.0592573695,0.1933968067,-0.249840185,0.2242742479,-0.3976946771,-0.2356773764,0.0208289362,0.2337531894,0.4087526202,-0.2069119662,-0.2438365221,0.0980038419,-0.1904449761,-0.3769418299,-0.1676462144,-0.1064466536,0.0883559063,-0.014652296,-0.3172785044,-0.00905261,-0.1413644105,0.0560814254,0.0824531466,0.0137014808,-0.1196166202,-0.0904117227,0.0306326207,-0.1625874341,0.0537343882,0.4452787936,0.1993274242,-0.1737648696,0.6748771667,0.318600148,-0.2701166868,-0.0588436611,-0.145492211,0.026524296,-0.3916162848,0.0803605095,-0.1223210543,-0.0326891243,-0.2329388112,-0.0158797093,0.2639751136,-0.128901422,0.2503610849,-0.3973307312,-0.4390514493,0.3276472688,-0.1340383142,0.1806013137,0.2681422234,-0.3036365807,0.1483330727,-0.0007178176,-0.1766967624,-0.0336865559,-0.2832540572,0.1993031353,0.3011004329,-0.1552578509,0.0712757185,0.134473592,0.0281189606,0.1406504214,-0.167130813,-0.1209419593,-0.0215271469,0.1702215821,-0.0211845562,-0.1575117856,0.0079886634,-0.4215225875,0.1337191612,-0.3100504279,0.0949737802,0.0940818116,-0.0231312681,0.3049455881,0.1652727872,0.0589501262,-0.0955347419,0.4590955079,0.004906415,0.4881094098,0.1949438304,0.1460272819,0.21745345,0.0095139295,-0.529992342,0.0595963299,-0.1048070416,-0.0354540385,0.129529193,-0.1712080538,0.015818838,0.1867771298,-0.040361397,0.102782838,-0.1173781157,-0.0446845815,0.2029732615,0.1301267743,-0.129020974,0.1219610199,0.343000412,-0.038141951,-0.2553494275,0.1517674178,0.1166727766,-0.0218609311,-0.0463967025,0.138712272,0.3468563259,-0.0800423026,0.0823748708,-0.1870741248,0.0247108527,-0.0429122858,-0.1344307065,0.2051127702,0.1837250888,0.4019264877,0.0027958902,0.1415004283,-0.1637841016,0.075403668,-0.1910852641,-0.3289398253,0.0203157384,0.5687055588,-0.0469907708,0.2843696475,-0.1446273178,0.1780040264,0.3612075746,-0.1369031966,-0.367102474,0.2313219756,-0.1447919011,0.0617254116,-0.058048822,-0.2364676446,-0.2097277045,0.2906796634,0.0528513603,-0.4837180972,0.0807831958,0.1928961873,-0.1719204783,-0.2963343859,0.4201445282,0.1663333476,-0.0192703065,-0.3831961155,0.0990737677,0.1448647529,-0.0447314866,-0.0872007981,0.3097783327,0.4060985744,0.3071936667,-0.2704363167,0.0093551604,-0.0059454297,-0.0063552903,-0.2495992184,0.2559792399,0.4065111876,0.3059679568,0.1803972274,0.0541909225,-0.1025387272,-0.3101823628,0.3593687117,-0.0628880933,-0.1212561727,-0.0192109924,-0.2770347297,0.0351019092,-0.2277391851,0.0494226553,-0.5648443103,0.1031810939,0.2087474763,-0.132521525,0.1167968512,-0.0363964811,0.0014875523,-0.1657541692,0.5369609594,0.2511483133,0.1122913584,-0.4637848437,-0.3375482261,-0.5109270215,0.0780812204,-0.3646364808,0.0649989098,-0.0553256199,0.3731818795,-0.1717679948,0.1566798538,-0.2668590546,0.1098729745,-0.1133813411,0.2389042079,-0.3001950383,-0.2587818205,0.0470050648,0.0787234753,-0.0804847926,-0.2608900666,0.0874856189,-0.0274887383,-0.0548563041,-0.2284462303,-0.2025665939,0.0881140232,0.2646359205,0.2633498013,0.1222851872,0.5243082047,0.0749866292,-0.3782134652,-0.2976669967,-0.1673058867,0.0435902327,0.0065886723,-0.080036357,0.5198352337,-0.2667166889,0.1582237184,-0.4382742643,0.4756548703,-0.1714072227,0.1068449765,-0.3627738953,-0.1564667523,0.0345594808,0.226611346,0.2080149949,0.1816634089,-0.091755636,0.077336058,-0.2539073229,-0.1309581548,0.3351980746,-0.5924607515,-0.3291895986,0.0524134412,-0.058366973,-0.0152939847,-0.1152605638,-0.5271500349,0.0343527161,0.1688707769,0.0881754756,-0.1708844006,-0.0418434329,-0.0630329624,-0.0228900574,-0.0843612254,0.0717218816,-0.1209484562,-0.3793058991,0.1373226494,-0.2624127865]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/277","title":"Empty samples in glue\/qqp","comments":"We are only wrapping the original dataset.\r\n\r\nMaybe try to ask on the GLUE mailing list or reach out to the original authors?","body":"```\r\nqqp = nlp.load_dataset('glue', 'qqp')\r\nprint(qqp['train'][310121])\r\nprint(qqp['train'][362225])\r\n```\r\n```\r\n{'question1': 'How can I create an Android app?', 'question2': '', 'label': 0, 'idx': 310137}\r\n{'question1': 'How can I develop android app?', 'question2': '', 'label': 0, 'idx': 362246}\r\n```\r\nNotice that question 2 is empty string. \r\nBTW, I have checked and these two are the only naughty ones in all splits of qqp.","comment_length":23,"text":"Empty samples in glue\/qqp\n```\r\nqqp = nlp.load_dataset('glue', 'qqp')\r\nprint(qqp['train'][310121])\r\nprint(qqp['train'][362225])\r\n```\r\n```\r\n{'question1': 'How can I create an Android app?', 'question2': '', 'label': 0, 'idx': 310137}\r\n{'question1': 'How can I develop android app?', 'question2': '', 'label': 0, 'idx': 362246}\r\n```\r\nNotice that question 2 is empty string. \r\nBTW, I have checked and these two are the only naughty ones in all splits of qqp.\nWe are only wrapping the original dataset.\r\n\r\nMaybe try to ask on the GLUE mailing list or reach out to the original authors?","embeddings":[0.3944284618,-0.2275950313,-0.1333640814,0.1721872091,-0.0090440977,-0.1809310764,0.3399785757,0.4980908334,0.3016017675,0.2876112461,-0.1628892273,0.4013324678,-0.0513740256,0.0415244102,0.1211948544,-0.2321262211,-0.1426435709,0.4839450121,-0.2110554874,-0.0486818142,-0.2900344431,0.3497726321,-0.1607245803,0.0145700229,-0.2572481036,-0.0267112777,-0.4356914759,0.1250949204,-0.0838717371,-0.1141366884,-0.0953655615,0.088230148,-0.3356195092,0.0275134351,-0.0001076749,-0.1921067536,0.2097817063,-0.0667193532,0.0979350433,0.0208776332,-0.1886062175,-0.1601732671,-0.0197192859,-0.3324033618,-0.1081115827,0.0603447631,0.0312088747,0.0091472585,0.3878845572,0.2867927551,0.231938839,0.1878136843,-0.0533459149,0.0011456627,0.5737708211,-0.4523113966,0.1663807482,0.1940990835,0.1904550791,-0.0723294169,0.2012558132,0.2850463092,-0.0291922819,-0.1413354278,-0.3117975295,0.073603563,-0.0549966656,-0.5886821151,0.11005532,0.5429876447,-0.1368231475,0.0054109218,-0.1368394196,-0.2885940075,0.0308143701,-0.1221046373,-0.0853625312,0.289188534,-0.1083173379,0.1899968237,-0.091862075,-0.1836483926,0.0907038078,0.111167796,-0.1785192937,0.3162373602,-0.0265359003,0.0967656299,-0.0690471306,0.0069285724,-0.3025737703,-0.0438517183,-0.2633654177,-0.05087642,-0.0054775872,-0.1263584048,0.0265676584,-0.1544861495,0.014671864,0.0555241778,0.0235987026,0.0789896846,0.3578359783,0.1973656118,0.0603472888,0.1082852259,0.4745409489,0.2059003264,-0.2136047333,-0.1027712151,-0.0928279161,-0.1192266196,-0.0344808735,0.1058450416,-0.1463371068,-0.0181441456,0.1527648568,0.01987423,-0.1597275138,-0.0852263942,-0.2839925587,0.093300961,-0.2367840409,0.4214972854,0.0020284581,0.088876389,0.0664033443,0.1870643348,-0.1252355725,-0.4068473577,-0.3048885465,0.14692536,-0.0183423571,-0.2193646878,0.0506461672,0.0939232856,0.3305933774,0.0040266337,-0.3367695212,-0.0640058443,0.1815020293,-0.1282046139,0.2904049456,0.3732571304,0.1224220395,0.0452854969,0.0463478342,-0.1798280925,-0.0764913335,0.305179745,-0.2295915335,-0.0494144224,-0.2482259721,0.2842925489,-0.2753049433,0.2139467448,0.301047653,0.2676579356,-0.0028213784,0.0312818959,0.245218724,-0.0584058985,0.0224559084,-0.3709172904,0.1769923568,0.1493795067,-0.4684355259,-0.2842772603,-0.040240176,-0.0014727769,0.0031715131,0.2163157314,-0.0184906013,0.267180711,-0.3092720211,0.5628643632,0.3224934042,-0.3252393901,-0.2990133166,0.2551974952,-0.2539637685,-0.0020691596,0.0715375543,-0.0273460504,0.4085083306,0.2399876565,0.2770467699,0.2180753797,-0.007983976,-0.1678893268,-0.384680301,-0.1650359035,0.3145085275,-0.0528995022,-0.0659816638,0.1639251411,-0.112235412,-0.0719398186,0.2754436433,0.1192446798,0.0017785862,0.1193085462,0.3002591133,-0.2346257865,-0.0166703388,-0.0893278196,-0.2737095058,0.084728919,-0.5236303806,0.163312301,0.091062285,-0.1931527853,-0.1754205674,-0.0931807309,0.0721271187,-0.2961351573,0.2624043822,-0.0942217261,0.1309671551,-0.0039560669,-0.0591877177,0.2372475117,0.0441201329,0.1758411825,-0.2782508135,0.1913246214,0.110832192,0.0012173938,0.1393664628,0.603890717,0.1739385277,0.0278028324,0.3377902508,0.3833782673,-0.1247403473,-0.0124142682,-0.0620077699,-0.4666868746,-0.0670435056,-0.2006079704,-0.0397650041,0.1223443225,0.0825217515,-0.0669386536,-0.0259498823,0.582020998,0.2762618661,0.1307536066,0.026604116,0.1524035186,-0.3962775767,-0.4450353086,-0.1695312113,-0.307769835,0.1928272992,0.1617302597,0.3633176684,0.2564749718,-0.1709272712,0.3790645599,0.5739662051,-0.0692600757,0.3076536357,-0.2247022241,-0.2381239533,-0.1790424734,0.0701891482,0.3703540266,0.1997359395,0.3989004195,0.0155678988,-0.0836394578,0.0834258273,-0.3214094639,0.2073069364,0.1239553243,-0.229490906,0.1494203061,0.3624715507,-0.1701753438,-0.201259017,0.5741873384,0.0243121926,-0.0494152382,-0.1920899004,-0.1492207497,-0.3281409442,-0.2790329754,-0.0113154044,-0.2292650193,-0.2454906255,-0.4493693411,0.2525092959,-0.3933390677,-0.1659594923,0.0172137953,0.131049931,0.3514339328,-0.1337381005,0.0004678103,-0.012865006,-0.037069425,-0.3257400393,0.2296125144,-0.4175542295,0.3865434229,0.3919174373,-0.0575511046,-0.0735447109,0.1745669097,-0.280805856,0.107807599,-0.3091467023,0.4351453781,0.1558913887,0.0820875242,-0.0727378726,-0.0775733292,0.109332405,-0.0259627961,-0.1517684013,-0.0908388868,0.2362346351,0.232569918,-0.3710660338,-0.4433030486,-0.2700110972,-0.3494876027,0.4619119465,-0.2531074584,0.3448719084,0.0024914034,-0.1901856065,0.2387921065,-0.3217241764,-0.2853172421,-0.166195333,-0.1441185623,0.2234225124,-0.2578462064,-0.2959876359,0.0591706596,-0.2494992018,0.1469355822,-0.188769415,-0.3285140097,-0.3279092908,0.0730063245,0.0583032221,0.3050025702,-0.0555207282,0.1158126891,-0.2326511741,-0.1398031414,-0.1862350553,0.0044557997,0.4342602491,0.0017085961,0.0043306304,-0.0346563794,0.2183077931,-0.0416686498,0.2817901969,0.120159857,0.4252009988,0.1642449051,-0.0652260631,0.3786019981,0.0359279513,-0.0413284376,0.17578502,0.2188123316,0.0603291951,0.3531430066,0.0979494154,-0.0609142333,0.0641166791,0.0818066746,-0.449634105,-0.1471253186,0.1239634007,-0.0709403679,-0.0928324834,-0.0041723629,-0.0717100874,0.0253139641,-0.0989011675,-0.2842521966,-0.1268379688,0.2005868107,-0.0387797616,-0.4878808856,-0.1508969963,-0.4676205516,0.3085104823,0.0355732702,0.2575332522,-0.0858508497,-0.1335566193,0.2047252506,0.0086970953,0.6842465401,-0.2168975621,-0.2390174717,0.2540599108,0.1206727326,-0.2901621759,-0.0238329042,-0.5618231893,-0.1005836874,0.3230280578,0.0700251386,-0.4161573052,-0.0658373162,0.0631185323,-0.0540161245,-0.2018034905,0.0775722414,-0.2286893576,-0.1632472277,0.0213699229,-0.1781850606,0.1592742205,0.3052087128,-0.2303267568,-0.0947498456,-0.0207480434,-0.1084983796,0.2755572796,-0.0635360852,0.4202828407,-0.1727812588,-0.150795266,-0.1001292765,-0.0251149293,-0.5728302002,0.2940683067,-0.2971131802,0.3505969644,0.1686253101,-0.3859302104,0.6130178571,0.3698877394,0.0828891248,0.3258310556,-0.2908847332,0.2322841287,0.0301062278,-0.0556588247,0.3734821081,-0.2357524931,-0.0026337458,-0.0946691111,0.201309979,-0.1164964437,-0.1116221398,-0.098670803,0.1508992165,-0.140408501,0.4762075841,-0.0653576404,0.8699271083,-0.1022026911,0.0288003106,0.3639542758,-0.2930316627,0.7299194932,0.1254662722,0.0636992902,-0.3811820149,-0.0855522826,0.0106716556,0.0865421668,-0.1003331617,0.276278913,-0.3786797523,0.3030673862,-0.2192403525,-0.1692805588,0.2131244838,0.0334584191,0.102320224,0.1487868428,-0.2839200497,0.1911133826,-0.029039707,0.0549911596,0.046912726,-0.2098493874,-0.0838778615,-0.1398311704,-0.0975531787,0.2478506714,-0.2676294148,-0.1409444809,0.4564511776,-0.0569871403,-0.6023692489,0.0812257156,-0.0145741478,0.3054861426,-0.0113750817,0.1296513826,0.1420897096,-0.0746205747,0.3581736386,0.2338319868,0.2600750029,-0.0419935919,-0.1917475909,0.2148963958,-0.1435481906,-0.2315842658,-0.3384072185,0.2367616743,0.0128476536,-0.2314345241,-0.4329336882,-0.0506715141,0.2014591247,-0.2351711988,0.1825748533,-0.0134257749,-0.2096028775,-0.1215615869,0.1633191258,-0.3453218341,0.0054446189,0.1530146152,0.1750057936,0.2415855527,0.4682160318,0.0946297497,0.0640941858,-0.3818280697,-0.1119730845,0.3699917495,-0.2195798904,0.2393682599,-0.0464700833,-0.160519287,-0.0169032738,0.1385279894,-0.0591162406,0.0342295654,-0.1641802341,-0.3515660465,-0.0441923477,0.4166790545,0.0003629217,0.1878093779,-0.0020641848,-0.022774158,0.1879604012,0.4029609561,-0.4251336455,0.1461168975,-0.3965063393,0.1516744345,-0.0415749028,-0.3134778738,0.1091136113,0.0059090517,0.1896004379,-0.0712062344,-0.2308023274,-0.3434695899,-0.1895463765,0.0522821657,-0.0676267296,0.2147141993,-0.09183871,0.0750328079,0.1201309785,0.0360366628,0.2240687162,-0.19221659,0.1864978522,0.2153778374,0.3040243685,0.2316845506,-0.2960215509,0.1838497221,0.0454704054,-0.0309836641,0.0210340396,-0.1541690528,-0.1896822006,-0.0810721889,0.1892766207,-0.0356468186,-0.0048923944,-0.0027142104,0.0540352352,-0.1590565443,0.0598057695,-0.2890315652,0.1538439691,0.0911514312,-0.2610992491,-0.0006663236,-0.0365212448,0.3149941564,-0.1158927307,0.1256052554,0.1385217607,0.140899688,0.1090192124,0.2486696541,0.1067240089,-0.2185636759,0.1103905514,0.1677855402,0.1849579662,-0.1511172205,-0.1075606719,0.3153118789,0.0948790312,-0.0257509016,0.2668917477,0.0092154322,-0.0205845628,-0.1160523146,0.1357221007,0.4513293803,0.2103834897,0.0945565775,0.3860871792,0.0188437328,-0.0540039055,0.4714006186,0.1002651155,0.2675414979,-0.0798869431,0.658287406,-0.1432418525,-0.2179017067,0.0811085328,0.3425239325,-0.1608319581,-0.2232286781,-0.0157118365,0.0582658052,-0.5974663496,-0.1774604023,0.1642058194,-0.2975628674,0.2071583271,0.0825633332,-0.165267989,-0.0391679928,-0.4906109571,0.0012091475,0.1394328624,-0.2860247791,0.278367281,-0.1906278133,0.1866098344,0.3655211926,0.4616247118,0.1054278761,0.1244624853,-0.323874265,-0.0424992032,-0.0298828036,-0.2519010305,-0.1712537408,0.2433502525,0.2742091119,0.2474101037,0.1131661907,0.2691096663,-0.3204631507,0.1359612942,0.2478088439,0.2534888089,-0.1750777662,0.27544716,-0.1274197102,-0.16556032,-0.220307529,-0.3327860832,-0.7666961551,0.0359205827,-0.0018588174,-0.1566048563,0.1184495986,0.0482258014,0.1376516074,-0.2596834898,0.1411233395,0.1483082771,0.0900063515,-0.2347634584,-0.2263382375,-0.4332302213,0.045763813,-0.1479471475,-0.1933630556,0.190292567,0.2700456977,-0.1034115702,-0.0357105546,-0.5100876689,0.4358816445,-0.3994014263,-0.259019047,-0.5443627834,0.1356120408,-0.1030209437,0.0729569346,0.1266053915,-0.1627966762,0.209971562,0.0596195608,0.1007718444,0.2535577416,-0.1036771685,0.3059473038,0.0692840815,0.0093490342,0.2776673138,0.1050219387,-0.1615355015,-0.0111971051,-0.1702062786,-0.1551191062,-0.0892969146,-0.0054647722,0.0261082835,0.4342817068,0.2479162812,0.1311109513,0.1950038821,0.1177413985,0.2375914603,0.2182043344,-0.0309675187,0.1526060253,-0.0342965536,0.0360255614,0.1560852081,0.2993510067,0.002967444,-0.0918378681,-0.0538026877,-0.1845365316,0.3916791975,-0.511018157,-0.0479526371,-0.0726110116,0.4356440306,0.2411486506,-0.1386790574,-0.6016255021,-0.4045550227,0.3187651932,-0.168705076,-0.0915843695,0.2373186648,0.1765239686,-0.0155380815,-0.0576656349,0.1002713218,0.2680636644,-0.1565011591,-0.0210756734,-0.2616681755]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/277","title":"Empty samples in glue\/qqp","comments":"Tanks for the suggestion, I'll try to ask GLUE benchmark.\r\nI'll first close the issue, post the following up here afterwards, and reopen the issue if needed. ","body":"```\r\nqqp = nlp.load_dataset('glue', 'qqp')\r\nprint(qqp['train'][310121])\r\nprint(qqp['train'][362225])\r\n```\r\n```\r\n{'question1': 'How can I create an Android app?', 'question2': '', 'label': 0, 'idx': 310137}\r\n{'question1': 'How can I develop android app?', 'question2': '', 'label': 0, 'idx': 362246}\r\n```\r\nNotice that question 2 is empty string. \r\nBTW, I have checked and these two are the only naughty ones in all splits of qqp.","comment_length":27,"text":"Empty samples in glue\/qqp\n```\r\nqqp = nlp.load_dataset('glue', 'qqp')\r\nprint(qqp['train'][310121])\r\nprint(qqp['train'][362225])\r\n```\r\n```\r\n{'question1': 'How can I create an Android app?', 'question2': '', 'label': 0, 'idx': 310137}\r\n{'question1': 'How can I develop android app?', 'question2': '', 'label': 0, 'idx': 362246}\r\n```\r\nNotice that question 2 is empty string. \r\nBTW, I have checked and these two are the only naughty ones in all splits of qqp.\nTanks for the suggestion, I'll try to ask GLUE benchmark.\r\nI'll first close the issue, post the following up here afterwards, and reopen the issue if needed. ","embeddings":[0.271572113,-0.2135789394,-0.0960008353,0.2048930079,0.0334747694,-0.225374341,0.2624841034,0.5297278166,0.2512718141,0.3464791477,-0.1416842639,0.3634182811,0.0181651134,0.01799234,0.094574675,-0.2144176066,-0.1223716512,0.4743580222,-0.2266591638,-0.0161949974,-0.2788559198,0.2712185085,-0.2013012618,-0.0525031015,-0.2173058242,-0.0292863809,-0.2624510527,0.0808574483,-0.0969525501,-0.2160422653,-0.1603916138,-0.0079348674,-0.4718260467,0.1135303974,-0.0001081613,-0.2366446108,0.3038535416,0.0301308166,-0.0165314879,0.0587730184,-0.1760877073,-0.1438621432,-0.0316167511,-0.2702952623,-0.1310332268,0.1710726321,0.0563666709,-0.0414781757,0.29091236,0.2894429266,0.2498814315,0.3229873478,-0.1677517444,0.0585022941,0.6316813231,-0.4834994078,0.0495391414,0.2030827552,0.1330105066,-0.1508524567,-0.002295092,0.31549263,-0.0769410059,0.0050608828,-0.2792363167,0.0500551052,0.0341887288,-0.3719379306,0.1219384149,0.6385517716,-0.2257316858,-0.0109339533,-0.1305674314,-0.2927179039,0.0347197801,-0.1861673892,-0.0095988782,0.2188460827,-0.1121698394,0.0842830464,-0.0766551718,-0.0182503108,0.1095781475,0.0213664509,-0.1605271697,0.2946777046,0.0150829675,0.0994072482,0.0831712559,-0.0842941403,-0.367105037,-0.0496092066,-0.2779007554,-0.0806365386,-0.2737316787,-0.0291984845,0.1033329666,-0.2223591655,0.091556333,0.0383832641,0.0451059788,0.1216750145,0.3454152942,0.1836071908,0.0129121933,0.2131349742,0.5253053308,0.1997385919,-0.2258029431,0.0194709748,-0.2615400553,0.0119850943,0.0702880323,0.0163297243,0.0182542503,0.0525818244,0.0214352142,-0.0071948986,-0.181463778,-0.0523411185,-0.260275811,0.0578023233,-0.2140987217,0.3981827796,-0.0354133621,0.0819521248,-0.0214646235,0.1872767359,-0.2436059266,-0.4802745879,-0.2614175081,0.1371790916,-0.041058179,-0.1456342489,0.0946939737,0.1308218092,0.3209361136,0.0506912582,-0.2557221949,-0.0488579534,0.3602681458,-0.2660951912,0.253059268,0.2802160382,0.2013659775,0.1034067273,-0.0073912353,-0.1239401475,-0.0878083333,0.2330834121,-0.0307327639,0.0183601044,-0.0971510857,0.30669716,-0.112555556,0.1843148768,0.1896537095,0.2996610105,0.0256747957,-0.0152722085,0.1179271191,0.0051100929,-0.0761498138,-0.336587429,0.072602734,0.1973284185,-0.3691127896,-0.2344619036,-0.0350852534,0.1252648532,0.1360246241,0.2471339107,-0.0183297172,0.171940878,-0.2010870874,0.4859287441,0.1703246385,-0.4958887398,-0.3176754117,0.3468904495,-0.2425451726,0.0385901816,0.1114358157,-0.009177452,0.4340400398,0.226268217,0.3273109198,0.3090688884,0.0549235158,-0.0748044029,-0.5140237808,-0.1311775148,0.3648431897,-0.0378127173,-0.1155951396,-0.0513603836,-0.2063602954,-0.0577512048,0.1394093931,0.1814709455,-0.0263048895,0.1331997365,0.271083653,-0.387337625,-0.0003699263,-0.1440971941,-0.2968893051,0.1813849658,-0.5925250053,0.1939084679,0.2127396911,-0.2041662931,-0.1179871559,-0.0427297652,0.0378256962,-0.2277549356,0.2820472717,-0.1904683113,0.1155307665,0.0669649988,0.0325520299,0.1890295744,0.0194214918,0.1492248625,-0.1549046785,0.1268672049,0.0374020301,0.0111428406,0.1795771569,0.4414916039,0.2346164733,0.1180062816,0.1957749426,0.3532408476,-0.0556187928,0.0250540692,0.0326963961,-0.4226806164,-0.0368220955,-0.0464887954,-0.1070588827,0.2304226011,0.0195707902,-0.1014252454,-0.0760976225,0.5049564838,0.3220842481,0.2473686486,0.0445962623,0.0596207716,-0.3528153002,-0.4345826209,-0.0700776428,-0.3270904422,0.3353847265,0.0409794599,0.4298760295,0.2962331176,-0.0394995995,0.3853868246,0.566626966,-0.0476820208,0.3834279776,-0.2004968077,-0.1923364103,-0.2509758174,0.0331685469,0.3896905184,0.2012426406,0.3364197314,0.0272395927,-0.1449757218,-0.001338392,-0.2931687236,0.225712508,-0.0030841869,-0.1504805684,0.2093813419,0.4381842315,-0.1580533981,-0.2193771601,0.5158028603,0.0432374552,-0.0212625563,-0.1248267815,-0.1486525387,-0.3147306144,-0.2375023961,0.0446571112,-0.1595415175,-0.3207473457,-0.4617287219,0.2919749618,-0.3784709573,-0.1189679429,0.0736282468,0.133728981,0.3025200069,-0.0549640916,-0.0463313125,0.0555662885,0.0127427727,-0.3448801339,0.1689077765,-0.3408043087,0.3934916854,0.3623147607,-0.07537321,-0.1350661665,0.1690968871,-0.3209129572,0.0623841211,-0.1388616711,0.5314763784,0.1236593723,0.0662679821,-0.0634511486,-0.0176467318,0.1697911769,-0.1296929419,-0.1724248081,-0.0788312107,0.2004493475,0.2160006762,-0.2482372224,-0.372839272,-0.1128423885,-0.4325796068,0.30550313,-0.2845984399,0.3849646449,0.0027652327,-0.2537881136,0.2153553963,-0.2340178341,-0.1688700616,-0.1356313974,-0.0948432758,0.170809105,-0.2498506606,-0.2468600571,-0.1370799094,-0.2538340688,0.1085635573,-0.2012091577,-0.3262814879,-0.350831598,-0.0811013579,0.1288247854,0.2127402723,-0.125134334,0.0760346279,-0.1489944756,-0.1821599454,-0.1802304834,-0.0539148785,0.4500750601,0.0466834828,0.0400915146,0.065026179,0.2793383598,0.0598585308,0.3947966695,0.081266284,0.3368882537,0.1455902606,-0.0832585916,0.2390438616,-0.0501812287,-0.1090916097,0.133584559,0.2538377047,0.0705687702,0.3102230728,0.0782697126,-0.1035052687,0.1188375056,0.1034650356,-0.3536418378,-0.1636374146,0.1618609577,-0.1495027989,-0.0493764877,0.0581560433,-0.0129013555,0.0112935798,-0.0950225368,-0.2126782238,-0.2355821431,0.1287619174,-0.129789114,-0.4396393597,-0.0927489921,-0.5289298892,0.2870887816,0.1076285169,0.2634969652,-0.1335857362,-0.0815236717,0.1261907071,-0.0176507402,0.67053473,-0.2847555578,-0.0952090621,0.208931759,0.0658688173,-0.293711096,0.0069536273,-0.5354694128,-0.0933928713,0.4221332073,0.0908454135,-0.4992766082,0.0195101053,0.0867794752,-0.0175729748,-0.1837095171,0.001816474,-0.3305701017,-0.0826737508,0.1090015993,-0.1573685259,0.1065743938,0.338825047,-0.1372347474,-0.115798749,-0.0957573503,-0.1001980305,0.2797958851,-0.0216259919,0.4361642897,-0.2024807185,0.0819801167,-0.2250949442,-0.1014466733,-0.4868714809,0.2278878838,-0.2688478529,0.3672320247,0.0876362845,-0.3222587407,0.4701736867,0.3631215394,0.104958348,0.3797159195,-0.4211321175,0.2606576085,0.0894977972,-0.0786487162,0.3746202588,-0.0858164802,-0.0036107185,-0.0107164728,0.2240367979,-0.128774941,-0.1424565315,-0.0925841033,0.1049007922,-0.1471141279,0.4613465369,-0.0970066786,0.8855526447,-0.2199768573,0.0159163736,0.3063757122,-0.346252352,0.6145241261,0.1481975168,0.1310078651,-0.4615854323,-0.1653729081,0.0520295613,0.0537128784,-0.1075655073,0.2394584715,-0.3754077852,0.2790742517,-0.2547707856,-0.1699656397,0.1922533363,0.1166923717,0.1106619537,0.0704090595,-0.3522309959,0.2006762922,-0.1097978279,0.0111758923,0.1191120744,-0.1464597285,-0.0004636755,-0.1248831525,-0.1091600731,0.1569407731,-0.27499035,-0.051670678,0.4869017899,-0.0137333581,-0.5811046958,0.012960705,-0.1753713042,0.3412328959,0.0577264279,0.1832609177,0.1572469026,0.0159747321,0.4486148655,0.1998182535,0.1861381978,-0.083266817,-0.2632910013,0.16956833,-0.1590086967,-0.2312678397,-0.3078417778,0.2483602017,0.0651456043,-0.197381109,-0.4733287394,-0.1210266724,0.1677023768,-0.2819992006,0.1979310811,-0.1748457104,-0.2156783193,-0.1259829551,0.2328980714,-0.3480013907,0.0187583268,0.2295622975,0.1346795708,0.3314605653,0.40629673,0.0644754544,0.033801958,-0.3555426598,0.0092971353,0.286111027,-0.1667911112,0.2771489918,-0.1347801834,-0.2053501159,0.020119356,0.0766439661,0.0268658791,0.0483691134,-0.1119209528,-0.1995908022,0.0433881208,0.3028692901,0.0323548913,0.1065402776,-0.0033674198,-0.1801121384,0.1364152879,0.5124989152,-0.4190405905,0.0656815097,-0.4481351674,0.1751643866,-0.0405018926,-0.338576436,0.0638697669,0.0559462234,0.238538295,-0.0952181369,-0.2287250459,-0.3541775942,-0.2276967913,0.0429377295,-0.1539235264,0.2242744863,-0.1185686365,0.0866218433,0.0212837998,-0.0759802461,0.1834658831,-0.162136063,0.1149160713,0.1317553073,0.1907269955,0.1856644303,-0.2394363582,0.2190048099,0.0049079414,-0.0542157441,-0.0466692597,-0.1727114171,-0.0454999991,-0.0933001861,0.1147966832,-0.1161117032,-0.0677731708,-0.0060928594,0.035699293,-0.2034540623,-0.0031631228,-0.1465425789,0.2589978874,0.0424591079,-0.2457573116,-0.1235968918,-0.1482927054,0.3284456134,-0.1436349303,0.0351734757,0.123941876,0.0319518149,0.2316007167,0.2193279266,0.0596662909,-0.1983868778,0.1996989399,0.1561886221,0.2040750831,-0.2150550187,-0.1669159234,0.1465061009,0.1589642316,-0.0556071401,0.3031100929,-0.0819411576,0.0339737386,-0.0441479124,0.2292451113,0.4877113402,0.2321505547,0.0732760206,0.3107292354,0.1487121135,-0.0385643877,0.6244726181,-0.0582798161,0.2489005476,0.0008571615,0.6918641925,-0.2370527536,-0.3201916814,0.0244695209,0.2090952992,-0.1568623632,-0.3546380997,0.0819513947,0.0982556045,-0.568009913,-0.1150459051,0.203902036,-0.1209931076,0.2469462454,0.1300608069,-0.0405334421,-0.1789054126,-0.4321320951,0.0488939695,0.2448887527,-0.3571634293,0.2981532812,-0.1213966906,0.2621774375,0.2796632648,0.392200619,0.1777597368,0.2161088586,-0.3681357205,-0.0762846917,-0.0678273812,-0.1473537683,-0.2059291005,0.2012602091,0.2145915478,0.1683612913,0.2282188833,0.2457999885,-0.3538049161,0.0785038322,0.2948915064,0.1354816258,-0.1899468899,0.2803794742,-0.1798201203,-0.0304178074,-0.2985871136,-0.1843944788,-0.7166969776,0.0251029693,-0.0364420004,-0.210380733,0.1387125999,-0.0234352686,0.1443626136,-0.2412384897,0.2342242151,0.0859630406,-0.0024150764,-0.3239735961,-0.0998622775,-0.4623683393,0.1171024069,-0.182349667,-0.2512120008,0.1596691161,0.3514060378,-0.2212231606,-0.0091357138,-0.3307335973,0.3748611808,-0.3800550997,-0.2198624611,-0.6355608106,0.1669418067,-0.1612004787,-0.1145603508,0.1135703251,-0.3320986032,0.1983606666,0.0898916572,0.1090103388,0.2725344002,-0.1427312195,0.3721240461,0.1770854145,-0.0130674485,0.2649458349,0.248229295,-0.143791616,-0.1124340594,-0.0677280128,-0.1696067303,-0.0574696288,-0.0727654323,0.0774099901,0.4261710644,0.2542338073,0.0795843229,0.1739833951,0.3036410213,0.3054361641,0.117518492,0.0004170346,0.0838074535,0.0642225891,0.0732087493,0.0369904377,0.3366991878,0.0280322041,-0.0276938435,-0.0612352788,-0.3449518085,0.36647138,-0.503739953,-0.0737484917,-0.0516071245,0.449161768,0.3405846655,-0.0983002335,-0.5108272433,-0.4513756037,0.2980215847,-0.1225269064,-0.1465785056,0.3283297122,0.0530243888,0.029550653,-0.0211667027,-0.0443437882,0.2474568188,-0.2251920253,0.018570967,-0.1794684976]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/275","title":"NonMatchingChecksumError when loading pubmed dataset","comments":"For some reason the files are not available for unauthenticated users right now (like the download service of this package). Instead of downloading the right files, it downloads the html of the error.\r\nAccording to the error it should be back again in 24h.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/42851186\/84751599-096c6580-afbd-11ea-97f3-ee4aef791711.png)\r\n","body":"I get this error when i run `nlp.load_dataset('scientific_papers', 'pubmed', split = 'train[:50%]')`.\r\nThe error is:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-2-7742dea167d0> in <module>()\r\n----> 1 df = nlp.load_dataset('scientific_papers', 'pubmed', split = 'train[:50%]')\r\n 2 df = pd.DataFrame(df)\r\n 3 gc.collect()\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 431 verify_infos = not save_infos and not ignore_verifications\r\n 432 self._download_and_prepare(\r\n--> 433 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 434 )\r\n 435 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 468 # Checksums verification\r\n 469 if verify_infos:\r\n--> 470 verify_checksums(self.info.download_checksums, dl_manager.get_recorded_sizes_checksums())\r\n 471 for split_generator in split_generators:\r\n 472 if str(split_generator.split_info.name).lower() == \"all\":\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums)\r\n 34 bad_urls = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]]\r\n 35 if len(bad_urls) > 0:\r\n---> 36 raise NonMatchingChecksumError(str(bad_urls))\r\n 37 logger.info(\"All the checksums matched successfully.\")\r\n 38 \r\n\r\nNonMatchingChecksumError: ['https:\/\/drive.google.com\/uc?id=1b3rmCSIoh6VhD4HKWjI4HOW-cSwcwbeC&export=download', 'https:\/\/drive.google.com\/uc?id=1lvsqvsFi3W-pE1SqNZI0s8NR9rC1tsja&export=download']\r\n```\r\nI'm currently working on google colab.\r\n\r\nThat is quite strange because yesterday it was fine.\r\n","comment_length":45,"text":"NonMatchingChecksumError when loading pubmed dataset\nI get this error when i run `nlp.load_dataset('scientific_papers', 'pubmed', split = 'train[:50%]')`.\r\nThe error is:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nNonMatchingChecksumError Traceback (most recent call last)\r\n<ipython-input-2-7742dea167d0> in <module>()\r\n----> 1 df = nlp.load_dataset('scientific_papers', 'pubmed', split = 'train[:50%]')\r\n 2 df = pd.DataFrame(df)\r\n 3 gc.collect()\r\n\r\n3 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 431 verify_infos = not save_infos and not ignore_verifications\r\n 432 self._download_and_prepare(\r\n--> 433 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 434 )\r\n 435 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 468 # Checksums verification\r\n 469 if verify_infos:\r\n--> 470 verify_checksums(self.info.download_checksums, dl_manager.get_recorded_sizes_checksums())\r\n 471 for split_generator in split_generators:\r\n 472 if str(split_generator.split_info.name).lower() == \"all\":\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/utils\/info_utils.py in verify_checksums(expected_checksums, recorded_checksums)\r\n 34 bad_urls = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]]\r\n 35 if len(bad_urls) > 0:\r\n---> 36 raise NonMatchingChecksumError(str(bad_urls))\r\n 37 logger.info(\"All the checksums matched successfully.\")\r\n 38 \r\n\r\nNonMatchingChecksumError: ['https:\/\/drive.google.com\/uc?id=1b3rmCSIoh6VhD4HKWjI4HOW-cSwcwbeC&export=download', 'https:\/\/drive.google.com\/uc?id=1lvsqvsFi3W-pE1SqNZI0s8NR9rC1tsja&export=download']\r\n```\r\nI'm currently working on google colab.\r\n\r\nThat is quite strange because yesterday it was fine.\r\n\nFor some reason the files are not available for unauthenticated users right now (like the download service of this package). Instead of downloading the right files, it downloads the html of the error.\r\nAccording to the error it should be back again in 24h.\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/42851186\/84751599-096c6580-afbd-11ea-97f3-ee4aef791711.png)\r\n","embeddings":[0.058758609,0.1301917434,0.0369352363,0.0248560738,0.2648719549,-0.0135630183,0.216802299,0.5417689085,0.2348525822,0.1036150381,-0.139252767,0.3025202751,-0.1345059574,-0.469260931,-0.1694715023,0.0336121656,0.0182243753,0.278247118,0.083115153,0.0176543035,-0.2718373835,0.3171768188,-0.2022533268,-0.0674488246,0.1895201951,-0.1108570844,0.1367979944,0.217150107,0.0484668277,-0.3944905996,0.2232331336,0.1793066412,0.1068236604,0.2187495083,-0.0001170882,0.1765800714,0.3026967645,0.0146549763,-0.2404377311,-0.3675601482,-0.306748867,-0.5333061218,0.2033980042,-0.3149736226,-0.0939296708,0.0818529204,0.0985179842,-0.0667017177,0.1352661699,0.2718558311,0.1855047345,0.4645368159,-0.0121475356,-0.0389801152,0.2210695595,-0.0710691288,-0.1167197675,0.4443893135,0.2470195442,-0.205421567,-0.4585720003,0.2826867402,-0.2289150953,0.4773514867,-0.0487231873,0.0054706582,0.4340510666,-0.2538340092,0.1287406385,0.3172093332,0.124449648,-0.1102622896,0.0408434309,-0.2443514317,-0.0612620674,-0.1005057618,0.3322421908,0.1642604917,-0.3393443227,-0.164248243,-0.2672503293,0.1548261195,-0.0184418168,0.2772804499,0.4511999488,0.0823994428,0.0259934198,0.0779543221,0.3885330856,-0.1400493532,0.4081314206,-0.1891864985,0.0559792668,0.2718269527,-0.3856696188,-0.1050526798,-0.0299018398,0.0830047354,0.2192598432,0.5085961819,0.1739616841,-0.0206740312,0.1601028144,0.1740664244,0.2345796525,-0.0325501524,-0.20497334,0.3127065301,0.2428222746,0.2654270232,-0.0817287639,0.1325980276,-0.1243766472,-0.2750798166,0.4795861244,-0.1543896943,0.1567379683,-0.3283813,-0.1853384376,0.2568361461,0.1256311685,-0.037092898,0.150780201,0.379488349,-0.0067011532,0.2513456941,-0.0467179157,-0.0900826231,-0.2285945415,-0.0017539149,-0.1619568467,0.078827545,-0.2954720855,0.0364404917,0.291310519,-0.1616549641,0.3400786221,-0.0261939988,0.1431849599,-0.1925871819,0.2771724463,-0.2258296162,-0.1703099459,0.400228858,0.0789528787,0.0937558636,0.259924233,-0.0361443907,-0.3035447896,0.171074912,-0.1880393922,-0.3761236668,-0.354393512,0.1679352224,-0.3283609748,-0.1124940366,0.1948510855,-0.1655438989,0.4185551405,-0.5822321773,0.0198790077,-0.3036311269,-0.2127438039,-0.1403591037,-0.0047661471,0.1324805021,0.3329315484,-0.0629779398,-0.0698211119,0.0543446913,0.5059974194,0.3451900184,-0.2003525347,-0.0730774403,-0.1836931109,-0.0042758714,0.6614866853,-0.1024645865,-0.5011329651,0.3050438166,0.0577829257,0.4503071904,-0.033370968,0.1673368067,-0.1188712195,-0.101896219,0.3278857768,0.0811803341,0.0006618578,0.2791234553,-0.3819270134,-0.1303378493,0.5268676877,0.2277288735,0.1866333634,-0.0063327644,0.0629571304,0.2206354439,0.1839743108,-0.0855364129,-0.0306455232,0.2708131969,0.0185403861,0.0692929104,0.0581090748,-0.2218934298,-0.3858188689,0.2098036557,-0.380053401,0.2502179742,0.0859001353,0.0226947721,-0.3482504189,-0.3304007053,-0.210401997,0.0228056498,0.0646701306,0.3360591531,0.1912897527,0.0297807064,-0.1321470737,0.2002444714,-0.3121551275,0.0384585224,-0.5431321859,0.2622187436,-0.2126975209,-0.1852943301,0.164986223,0.1927392185,0.0837604105,0.0281460565,-0.2496987879,0.3297187686,0.0849613547,-0.0043527321,0.0416009612,0.0834872201,-0.0649434477,-0.09139283,-0.003358939,0.450768888,0.0452840663,-0.0097539034,-0.2259782404,0.3569218516,-0.0916158184,0.1400891393,0.0180972219,0.111937426,0.292778492,-0.1206601635,-0.2199435085,-0.1293094158,0.5966858864,0.0419388562,-0.1533253789,0.1737088412,-0.4326075912,-0.295591563,0.3696337044,-0.0256391652,0.0271185786,0.1284158677,0.0668641329,-0.1737574339,-0.013439131,0.3423480392,0.502830863,0.1839086413,-0.1045734361,0.0342627466,-0.3598060608,-0.0776783898,-0.0758731961,0.1211585626,0.4027727842,0.5767911673,0.3505509794,0.0097552771,-0.5195763111,-0.2499975264,0.0892635733,0.5005633235,-0.3145840466,-0.0588865057,-0.2020569295,-0.1500694752,-0.2669269741,-0.2294199616,-0.2996838987,-0.2502132654,-0.0903199464,0.4568479955,0.013585669,0.3313672841,-0.7557685375,0.0948301107,-0.0509436615,-0.1814254373,-0.1037207767,-0.1316320896,-0.0341805406,-0.0372446664,0.4121696651,0.0776028782,0.4435406327,-0.0992219895,-0.1393858939,-0.2177254558,-0.2293803543,-0.0338992439,-0.0765170306,-0.0230811145,0.2885401845,0.0395289585,-0.1497325003,-0.3457106948,0.3683997691,-0.2874168754,-0.1831320226,0.2736922503,-0.0688725263,0.0477568209,0.0877032802,-0.2555535436,-0.1006214321,-0.4179308712,-0.0921681896,0.2130959332,0.1583324075,0.377029717,-0.1883330047,0.1201434508,0.0769679248,0.3528823853,-0.2467336208,-0.3859330118,0.3995440602,-0.0446745269,-0.2828446627,-0.0373046882,-0.2464389801,0.0727727264,0.0861012265,-0.4358641207,-0.3477205336,-0.2699185014,-0.2323732972,-0.0331604108,-0.1730334908,0.3908375204,0.0942935273,-0.1394832581,-0.0061373133,-0.1883255392,0.1998282671,0.0757934898,0.3481656909,-0.1597508639,0.2954999804,-0.136423707,0.572486639,0.4148236513,0.0519084893,0.0926354006,0.0320627168,0.3804651499,-0.1218233407,-0.3274921477,-0.0851674974,0.018538259,0.1922067851,0.1199336872,-0.0905446559,-0.2683863938,-0.0759817436,-0.1040291116,-0.1687823683,-0.275039047,0.1403013915,-0.3935478926,0.3114085495,0.2206878513,-0.0049183383,-0.174995333,-0.4817243814,-0.1533858627,0.3081018627,-0.1542447507,-0.0217623487,-0.2281146646,0.2059494108,-0.4059179127,0.2446061522,0.332151711,0.3792577684,-0.0965353251,-0.1378999799,0.1744908988,-0.1074006334,0.2112129331,-0.1568175107,0.0229365714,0.1802372783,-0.193263799,-0.3291547894,-0.1285332441,-0.0708794594,0.2246686071,0.1628459394,0.3866912127,-0.1893853992,-0.2253923714,0.5273262262,-0.0171767455,-0.1391472816,-0.5009344816,-0.4964928329,-0.3288419247,-0.4223902524,-0.0244218502,-0.0691626817,0.3058442175,-0.0602593049,-0.0070520178,0.0879858434,0.007744805,0.2474460155,0.0470835082,0.1417860687,0.2906084359,0.2893086076,0.0748605281,0.2209791988,-0.0449585095,0.4247762859,-0.2667268217,-0.2374444455,0.0583176166,0.0230593272,-0.1651921272,0.0698721111,0.1089125127,-0.0540041551,0.354783088,0.0694028884,0.1609755903,0.0149473846,0.3397999704,-0.1320673674,-0.4824078679,-0.3375608921,0.2068861723,-0.0034058893,0.0006137264,0.141939491,-0.1403355598,-0.0144427167,0.1813034117,0.0700120628,0.9123491645,-0.0502093993,0.2700971961,0.155629456,-0.1263743937,0.484785825,-0.1906324029,0.3136151731,-0.4442590773,-0.3440862,0.0332065076,-0.2150554359,0.0016659602,-0.0846742317,-0.0313050523,0.4170595109,0.3165467978,-0.0888693258,-0.0297749508,0.2389581949,-0.0411700569,-0.1754746437,-0.316509217,0.1362701207,-0.4005624056,0.3194374144,-0.1510509104,-0.002036839,-0.002230247,-0.1660822928,-0.2145355195,0.2455095649,0.1838780046,0.4514034986,0.5331104398,-0.1979886293,-0.145332858,0.0426795073,0.073144719,-0.1107757986,-0.2280195504,0.0841793343,0.4728123248,-0.2130319625,-0.027643105,0.1677720249,0.5750768781,-0.1611853987,-0.0488367081,-0.0175486393,-0.1586233228,-0.5636658669,0.0165213812,-0.0144087197,0.2287370414,-0.2249049395,-0.2408615798,-0.006057377,-0.1177294701,-0.2581005692,0.1344264448,0.1428399384,-0.2220241129,0.1995378286,0.095813565,-0.3081702888,0.0535598658,0.4131399989,0.1912595034,-0.2264447212,0.4888488352,-0.0402879864,-0.323217392,-0.1999334395,0.0013781581,-0.0778845623,-0.3243076503,0.2354054898,-0.0050533335,0.2434721142,0.1977971196,0.4563693702,0.1538855731,0.1639074385,0.0640108809,-0.4290203452,-0.3424910009,-0.0983150378,-0.058682166,0.3464601934,0.1933617443,0.025077153,0.0282806996,-0.0141406339,-0.2789706886,-0.0362044089,-0.6158044934,0.0808647722,0.1627991199,-0.1786714643,0.0985638499,0.0848488584,0.1080764681,0.0622780211,-0.118789196,-0.2138739824,-0.2060462832,0.16817981,0.1312129945,-0.119877331,-0.0978473276,-0.1324732602,-0.1538299769,-0.1558457911,-0.2439978719,0.1678182483,-0.0650607273,-0.0109392712,-0.1312608272,-0.0744362026,0.2280102372,0.1402037591,-0.4103318155,-0.1182083189,0.0809397474,0.009015345,0.0280953031,0.0839531645,-0.0772586018,0.0766524747,0.314153254,-0.2686917484,0.2941831648,-0.3255680799,-0.136290729,0.0649695322,0.337026298,0.2133489996,-0.4081490338,-0.0788637623,0.2357676178,0.1682230532,-0.5343015194,0.0118744299,0.3354955614,0.1092594191,0.0180379935,0.1105092242,-0.146067366,-0.3263048232,0.3797014058,0.1548517048,0.2326482981,-0.6414141655,-0.0072425562,0.2353511602,0.1024449542,-0.1425143182,0.2817765176,0.0523321256,0.1128085628,-0.0291441642,-0.0650060922,0.1464861929,-0.1957941353,0.1531431824,-0.23318322,-0.2960094512,0.2189733088,0.4795645177,0.2377023101,0.0864738598,0.0658563375,0.3420736194,-0.0713060647,0.2049432993,-0.3094731569,0.2302245945,-0.1054363474,-0.0936600938,0.0239870343,-0.0296238996,-0.1709246039,0.2104807943,-0.0133724296,0.2012875974,0.0409114808,0.133392334,-0.0445417985,-0.5335198641,0.1446587592,0.2360322922,-0.0156143066,-0.2741888165,0.2356105894,0.1141975299,-0.0736607239,0.0423975512,0.3567353189,0.3995248675,0.3630995452,-0.007847202,-0.1293052286,0.1155804843,0.1168572828,0.137637645,0.6320483088,0.3914758265,0.2891390324,0.3663424253,0.015223301,-0.199554503,0.2551238835,-0.0997853503,-0.3354570568,0.0263992604,0.7012009621,0.1500085443,-0.005062894,-0.2052838802,0.1908212304,-0.3925946951,-0.2459540069,-0.0374747925,0.0676875785,0.1558868587,-0.1156782359,0.0563672967,-0.1069598645,0.4185028672,0.144592911,-0.0550763905,-0.3177075684,-0.3510942161,-0.7823400497,0.4147538841,-0.2222436517,-0.00453933,0.1111758947,0.1239203736,-0.1260117888,0.1213584468,0.0527518094,0.0829532519,-0.4774822891,-0.1319554746,-0.026100779,-0.0201770123,-0.1110374779,-0.4083555639,-0.0465532728,-0.183736071,0.0283312201,0.1031557694,0.0224977136,-0.0233072154,-0.0576357879,-0.1868527234,0.4520632625,-0.086367406,0.1859265119,0.4873575866,0.0534144938,-0.0161941312,-0.0614363737,-0.320417285,-0.2496080101,0.2517541647,0.1122905686,0.3334973454,-0.0513677225,0.0783071071,-0.3139564395,0.1882484406,-0.0214784723,-0.1734167337,-0.0735203028,-0.0597056188,-0.0415067486,0.2270751297,-0.0535887852,0.0115463585,0.2179293036,0.3108489811,-0.3252769709,-0.2350171059,0.5604361892,-0.4640592337,-0.3845112026,0.0114489021,0.1940463036,-0.2139998823,-0.2046723217,-0.7420552373,0.041729331,0.35193187,0.1089389324,-0.0904256031,0.1255107224,-0.5100414753,0.0548629686,0.0101072807,-0.0520602502,0.1499003321,-0.2129961401,-0.1807079315,-0.2111466229]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/274","title":"PG-19","comments":"Sounds good! Do you want to give it a try?","body":"Hi, and thanks for all your open-sourced work, as always!\r\n\r\nI was wondering if you would be open to adding PG-19 to your collection of datasets. https:\/\/github.com\/deepmind\/pg19 It is often used for benchmarking long-range language modeling.","comment_length":10,"text":"PG-19\nHi, and thanks for all your open-sourced work, as always!\r\n\r\nI was wondering if you would be open to adding PG-19 to your collection of datasets. https:\/\/github.com\/deepmind\/pg19 It is often used for benchmarking long-range language modeling.\nSounds good! Do you want to give it a try?","embeddings":[-0.5126307011,-0.1761040092,-0.2935127318,-0.1989256144,-0.1018454134,-0.2595919967,0.2872336209,0.1724146307,0.1119277254,0.1124251708,0.0146612767,0.1633377671,-0.2597104609,0.6453084946,0.2245080322,-0.217996195,0.1738731861,0.1629879922,0.0161929987,0.0263118614,-0.210307315,-0.0009418928,-0.0609546117,-0.0293721594,-0.153445974,-0.2185863554,-0.1228498891,-0.1642298102,-0.144705832,-0.3071727753,-0.1048137695,0.2273892611,-0.0353660993,0.4580079019,-0.0000963435,-0.1812041104,0.1176715195,0.099162288,0.0605724566,-0.0838028714,-0.0466728322,-0.108197622,-0.4152242541,-0.1079576612,-0.193397209,-0.219022572,-0.0584228672,-0.3317202628,0.0887641013,-0.0907837301,0.342676878,0.1770369411,-0.0626607165,-0.1001416966,0.052455008,-0.0563108809,-0.3139868677,-0.0125307236,0.5930511951,0.0646359548,0.0444495752,0.345264554,-0.1163422912,-0.2921469808,0.1733080894,-0.1843825877,0.3484627903,-0.1069671661,-0.0095315771,0.4466868043,0.2213231772,-0.0587126128,-0.2991367579,-0.2157846987,-0.1840537339,-0.4881046712,-0.3688715696,0.001142538,0.0140885906,0.1794083863,-0.0644765049,-0.3897499442,-0.2155816555,-0.1199553236,-0.1089384481,0.6541976333,0.1075777486,-0.1839014292,0.0972457901,-0.2248331159,-0.2553008497,-0.1224016696,0.0114882672,-0.2614639401,-0.08700037,-0.2851990759,0.1456878334,-0.0524603315,0.1427945197,0.0502997302,0.3418531716,0.4433546066,-0.0788899437,0.0418084003,0.2402025163,0.0176838897,0.2205898166,-0.0861656293,0.2555288076,-0.0986990184,0.0518829934,0.2364498973,0.0755826086,0.0245434009,-0.1087003499,-0.0738762543,-0.2641077042,-0.2555109859,-0.0259135235,0.1004649624,-0.0126436129,-0.0887333751,0.1940126717,0.1086170822,0.0343950652,0.2014126182,-0.3181008697,0.0377138555,-0.0838305876,-0.4906457365,-0.1411121041,0.1114607006,-0.2194106877,0.0752042308,0.1401462853,0.0673766434,-0.1006386727,0.0828527734,0.1435897946,0.1647769362,0.2244122475,-0.2252596915,0.2712406218,-0.1926237047,-0.1139699295,-0.1987221688,-0.1827926189,-0.1446853876,-0.3433270454,0.0658375993,0.1648783088,-0.0804647133,0.0121131511,0.3092423975,-0.1935508996,-0.3863457441,-0.2556341887,0.7408767343,-0.2745271325,-0.1691603959,0.176751256,-0.0976697654,-0.3768149614,-0.0455304198,0.007399709,0.0055518709,-0.4687010348,-0.0805176646,0.0215688348,0.152158469,0.0924061686,0.3089047074,-0.1068031043,-0.21748285,0.1140835509,0.4357493818,-0.0505722761,0.042368304,-0.0184639413,0.0417244099,-0.1103862002,-0.2090153545,0.0710432902,0.2034190446,0.1934332401,-0.1692401171,-0.083932735,0.3048947752,0.1548392624,-0.0083636949,-0.3046762049,-0.1060405448,0.3896465003,0.3044659495,0.089753449,-0.671235323,0.0024322495,-0.1404026449,0.1175394654,-0.4017933607,-0.2844379544,-0.0872629359,0.5141513348,-0.1352359653,-0.0672467053,-0.2490990609,0.0266256481,0.1227529943,-0.0028476606,0.3094342351,0.2397734374,0.0428009033,0.071607627,0.3270642161,0.1446396261,-0.06140076,0.2668775618,0.10049624,-0.1467854083,0.2928030193,-0.1353300363,-0.1282717437,-0.1348582804,-0.3140166402,0.0401216671,0.1334693581,-0.0323231518,0.0014805486,0.1534515321,0.2107992023,-0.0387961082,0.0191623233,0.1277361065,-0.0227311477,-0.1847362816,0.2153800726,0.3817623258,0.5547032952,0.2757416964,-0.4261909425,0.2264729887,-0.0162936281,-0.1267527044,-0.0767043456,0.1454199553,0.7669002414,0.0143261813,0.0903871432,0.0665061548,-0.2357234806,-0.0028041815,-0.0108345635,-0.1683485806,0.0196558498,0.6008583903,-0.1911173612,-0.0359870754,0.152912572,-0.2329977751,0.2330082953,0.3642849624,-0.2657931447,0.3738533854,0.0633426458,0.0263957065,-0.4406180382,0.1355842948,0.0050484254,0.052089747,0.3670420349,0.1095177978,-0.1052287593,-0.0053256014,-0.0983759761,0.3326379359,-0.0026700601,0.041670166,0.2171014696,0.1983632594,-0.0415298827,-0.2745167613,-0.0962188616,-0.1052224338,-0.1048155725,0.1781594306,0.233644709,-0.2286733985,-0.2076814473,0.1222539395,-0.1893917918,-0.4590973556,-0.1195045933,0.2277909964,-0.3869766593,-0.2276978791,0.2703569233,-0.1341267675,0.5442224145,-0.2728540003,-0.2167185843,-0.0392089784,-0.3243042827,-0.0506057404,0.2372019887,0.233404398,0.0149502698,0.5011799335,0.2006667405,0.1406459212,-0.045720391,-0.4003968835,0.1295113564,-0.2163079083,0.2071751952,-0.1766599864,-0.0231508762,0.0639420003,-0.1490370184,0.0117096333,-0.6909408569,-0.0157959722,-0.2422846109,-0.1060820669,-0.171243012,-0.1001053452,-0.1436968893,-0.2947438657,-0.3808382154,0.2272181064,-0.0178993884,-0.0346371569,0.2135966718,0.0814144611,0.0990634337,0.089181751,0.0076440908,-0.2161564678,-0.0702136233,0.1994208544,-0.0187892262,-0.0659325942,-0.4216670692,-0.0831394419,0.2103223801,-0.0422084406,-0.3344091773,0.2081085294,0.1506344527,0.2896404266,0.1385584921,-0.259375453,0.1453278661,0.0466367491,-0.2272958606,-0.0189494565,-0.2089347392,0.0695226043,-0.0163707044,0.039071966,-0.0298519246,-0.0330172852,0.1450807303,0.3929818273,-0.3364672065,-0.2332237661,0.2548675537,0.2610044181,-0.0385804735,-0.1555027366,-0.1884396076,-0.1710208356,-0.0741623938,-0.0617806949,0.4373865724,-0.0210950878,-0.6137695909,-0.2004270703,-0.3370293081,0.1119359508,0.0392897688,0.23287718,0.1460085362,0.4156832397,-0.0419372991,-0.166515097,-0.4151753187,-0.1904855967,0.1136209443,-0.3453754783,0.0358295478,-0.1236987263,-0.4079249501,0.2444113046,-0.4268715382,0.444680959,0.245643273,-0.0174617879,-0.2984775603,-0.0526701882,-0.053881444,-0.173797816,-0.0344009735,-0.2434274107,-0.0560562201,0.1139093265,-0.1975735128,0.086801149,0.1867579222,-0.2600535154,-0.1794107854,0.4649822712,0.4084643126,-0.149366051,0.1590474546,0.0975670367,0.1283168197,0.0442187227,0.0235477705,-0.0090263858,-0.0540635884,0.1500522643,0.1582226008,-0.0514673926,0.0617797449,0.1030218899,0.0550241247,0.1351497769,0.209940508,0.319093585,0.1381499469,-0.12176653,-0.040127866,0.1686651111,0.0311687998,-0.0219102837,0.2002087981,0.073613897,0.047376886,0.1333673,0.0210622065,-0.1910039335,0.4274230003,0.2836173177,0.1522073448,0.3660613894,-0.048194278,0.1374277622,-0.3553670943,-0.1003708616,0.1829285473,0.2708808184,-0.382145822,-0.3422937989,0.3883626461,0.0385016389,-0.14062576,0.1913367808,-0.1892472059,-0.0230779126,0.499361515,0.2391396314,0.9347373247,0.1184306294,0.1233900338,0.1041216403,-0.0204855688,0.2259597033,-0.0808296055,0.1786568314,-0.0964968875,-0.1529129893,0.0423285328,-0.0732681826,0.0202860907,0.0819402337,-0.1380239576,0.0534920618,0.4442096055,-0.329800427,-0.0029543824,0.3332248628,0.2657414377,-0.1317483187,-0.2611036897,0.3187591434,0.0526831113,-0.3015160561,-0.1200555116,-0.2463829517,-0.034341421,0.0820214376,-0.3689948916,-0.0022971248,-0.0130741848,0.2231006622,0.0219187587,-0.3969063163,0.3445467651,-0.1339026839,-0.1106051207,0.0591890067,0.0151262833,0.1847786754,-0.0599648803,0.1118347049,0.1876412034,-0.2143540233,0.2718743086,-0.162827298,-0.2433512956,0.070222944,0.2459698617,-0.4103738666,-0.4771713614,-0.0456456393,0.1357752979,0.2090177387,0.2092150897,0.2671512067,-0.0952533633,-0.237161845,0.3045389056,0.1210457236,-0.0640007257,0.0435147434,0.2257530689,-0.1564164758,-0.2439830303,0.0940434188,-0.0643871203,0.1042193919,0.2737454474,0.3155331016,-0.2336291224,-0.3805211782,0.230743736,-0.3191204369,-0.2042043507,0.0808248594,-0.0428389721,-0.2605423629,-0.0661297739,0.2362848818,0.3767623603,-0.120698072,0.0124042844,0.0583825819,0.0447493568,0.0923056453,-0.1121252477,0.0663274899,-0.2130412906,-0.134809792,-0.1188309863,0.0970108882,-0.5205464363,-0.0833178312,-0.0887905732,-0.1034913659,-0.1455655098,-0.3263027668,-0.1641404778,0.0180084538,0.228150174,0.0470732786,-0.2191725522,-0.3848870695,-0.1553585827,0.0584349111,-0.0594017729,-0.0253289677,0.0542642921,0.1371599138,0.0203721039,-0.0476218536,-0.0940494165,0.0741817653,0.1530056,-0.0372929312,0.060491696,-0.1426716298,-0.0078340145,0.1413531601,0.2352652401,0.3498039842,0.0806141496,0.226106301,0.1478580534,-0.1234634221,0.0674342364,0.2899062932,-0.067453213,-0.0481359586,0.1280870885,0.0480846316,0.0014394484,0.3560382724,0.3975288272,-0.0614660755,-0.2220668644,0.0904286131,0.0499134399,0.3683255017,-0.2881943285,-0.2960068882,0.1939570904,0.0004356363,0.1652399749,0.1455531716,0.5494647622,0.1001306698,0.194343403,-0.0777363703,-0.0669891462,-0.1649994552,0.1259850711,0.0064761881,-0.0016705315,0.0259862766,0.2260163277,0.3083998859,0.2309625149,0.124909848,0.1765643805,0.2626186609,0.0447718799,0.275998652,-0.0460531078,-0.011447737,-0.0891803131,0.2113222778,0.0397780091,0.3396508694,0.178538546,0.4945549667,-0.1306825578,-0.0444906428,-0.2327438593,0.23655352,-0.0336741656,-0.0706334189,-0.0907688886,-0.2241376787,-0.574275434,0.1097048745,-0.0664891452,0.0822447985,-0.1916867793,0.2179244459,0.03177898,0.1321626008,0.18901591,-0.1899942011,0.3055337965,-0.219530344,0.1539508998,0.3059490025,0.1777801663,0.0324944556,0.3135772347,0.1298933923,-0.1529071927,-0.4904436469,0.176704213,-0.3528056145,-0.022632733,-0.1571272165,0.2596318126,0.2080160975,0.0567524061,0.4195333123,0.247559607,-0.2488086224,0.3735337555,0.2217034847,0.1827244312,-0.2168640494,0.2580948472,0.1460710317,-0.112155281,-0.2853206396,0.1214867011,-0.191778034,0.206745863,0.403439641,-0.1551926732,-0.0333431177,-0.0947370753,0.1795106679,-0.2210630029,0.4828570187,0.1849085689,0.4236961305,-0.1667782515,-0.1919882447,-0.3975341916,0.1940866411,-0.102627188,0.2651212811,-0.0787498578,-0.0042769471,0.4946236312,0.3058127463,0.0905643851,0.1416250169,-0.4265452027,-0.0149007533,-0.4560219646,0.4376765788,0.119118534,0.0173900761,0.0671509951,-0.0206203721,-0.0334486738,-0.1232622266,0.2697928548,-0.0960945487,-0.1866591424,0.0986146778,0.4367916584,0.3186466098,-0.0249449071,0.3662141263,-0.2249716669,-0.127970174,-0.002691221,-0.2295082361,-0.0255526192,0.1135320067,0.0500262752,0.2533611357,-0.1177459657,-0.2373097539,0.2241134942,0.1128375232,-0.1376680881,-0.2046186179,-0.1499467939,-0.2161746025,0.0753207803,-0.2886715531,-0.0501785353,0.3365187645,-0.0402044356,-0.1250142008,-0.1105701625,0.0232149269,0.2157826573,0.0715387538,-0.2091135085,-0.2039482594,-0.0332957953,0.3826673925,0.1604720801,-0.3555296659,-0.0079844762,0.2520947158,0.2153432369,-0.1210279092,0.1987080127,-0.0667188168,-0.0488391183,-0.0950163454,0.347135067,-0.073942028,-0.1770797223,-0.1504715234,-0.1380832195]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/274","title":"PG-19","comments":"Ok, I'll see if I can figure it out tomorrow!","body":"Hi, and thanks for all your open-sourced work, as always!\r\n\r\nI was wondering if you would be open to adding PG-19 to your collection of datasets. https:\/\/github.com\/deepmind\/pg19 It is often used for benchmarking long-range language modeling.","comment_length":10,"text":"PG-19\nHi, and thanks for all your open-sourced work, as always!\r\n\r\nI was wondering if you would be open to adding PG-19 to your collection of datasets. https:\/\/github.com\/deepmind\/pg19 It is often used for benchmarking long-range language modeling.\nOk, I'll see if I can figure it out tomorrow!","embeddings":[-0.5338121057,-0.2250035703,-0.2964842618,-0.1924934238,-0.0885891989,-0.2439277023,0.2673106492,0.2458212972,0.114312306,0.1618493348,0.0500397645,0.1839947104,-0.2065932006,0.5722091794,0.2329687923,-0.2390552312,0.1682980061,0.1916936338,0.0059723398,0.0262924246,-0.2420404404,0.0371678248,-0.0449330248,-0.073142454,-0.1640958637,-0.223195523,-0.104751803,-0.1311286539,-0.1234088019,-0.3491167426,-0.0768397525,0.2193989009,-0.0437050574,0.3950951993,-0.0000942984,-0.1999860555,0.146998927,0.0508122407,0.0260569993,-0.0772619396,-0.0679110214,-0.1257577986,-0.388799876,-0.1084327102,-0.1892160922,-0.1993840784,-0.014044689,-0.3031511903,0.070976533,-0.0391894467,0.366147995,0.1828207821,-0.1132933944,-0.1373940855,0.0561337434,-0.0755661353,-0.2972946465,-0.0507907942,0.5458198786,0.0066660121,-0.0254676435,0.3580365777,-0.0753665268,-0.2364685386,0.1781750917,-0.177515015,0.337569654,-0.1030480042,0.0021551475,0.4568752646,0.2070023119,-0.0420040227,-0.2420396358,-0.2048085034,-0.2021146566,-0.4620453417,-0.307464093,-0.0175120998,0.0353488624,0.1261096001,-0.0202540923,-0.291488409,-0.2124174237,-0.0698690787,-0.1311433315,0.601446867,0.0879452229,-0.1954554766,0.1141644716,-0.2364279628,-0.2956936359,-0.0956518203,0.0297242701,-0.1985495687,-0.0958722085,-0.2742513716,0.1948746294,-0.0630261898,0.1291188747,0.0348583721,0.3037083447,0.4033204615,-0.046559155,0.0925611183,0.1891300678,0.0303826183,0.2458780259,-0.1243174672,0.2509370148,-0.0638558045,0.0437468477,0.2180932015,0.0722477362,0.0380863138,-0.0878272206,-0.0372029506,-0.2547000945,-0.2485124916,-0.0808542669,0.1326923966,-0.0108131301,-0.0636730939,0.1609649062,0.1452969462,0.0599686503,0.233375296,-0.2870554328,0.0669591501,-0.1074026302,-0.4664393961,-0.1864501238,0.1384679377,-0.228884846,0.0110824304,0.1542925537,0.0921745747,-0.0255175307,0.090399757,0.099547036,0.1011850014,0.2940362394,-0.2359867394,0.3242656291,-0.1261761039,-0.0851115808,-0.185944885,-0.1401110888,-0.1528126299,-0.3729571998,0.0200681798,0.1989293098,-0.0765977204,0.0112003321,0.3295197785,-0.1618836224,-0.3658371866,-0.167678073,0.7459632754,-0.2522744536,-0.1561533213,0.1796578914,-0.0809734762,-0.3654853404,-0.0379184969,0.0167060569,0.0147150084,-0.4930289984,-0.1057978794,-0.001766863,0.1152893826,0.0701887533,0.2564147115,-0.0975063145,-0.2305862606,0.1029023528,0.4147396386,-0.0196047071,0.0249125399,-0.0288533662,0.0085139973,-0.1499758512,-0.1797928363,0.0890259966,0.2099573463,0.1788900793,-0.1786276102,-0.0849948451,0.3338852525,0.1036428213,-0.0003031258,-0.2797570527,-0.0978854522,0.3382207751,0.3104135096,0.1332267374,-0.6796185374,0.0010989287,-0.0852837041,0.0707224458,-0.3568122089,-0.2621306479,-0.0512926728,0.4270576239,-0.1183550134,-0.0597649515,-0.3036614358,0.0104615064,0.1255683452,-0.0064949607,0.2985114753,0.2536064684,0.0511322543,-0.0072163851,0.2778095603,0.1502713263,-0.0249939244,0.3115442395,0.1351449341,-0.1968158036,0.2723963559,-0.1183470413,-0.1026574001,-0.1047382802,-0.2976559997,0.0365082808,0.1054628864,-0.0387628786,-0.0573316626,0.2086678892,0.1847008467,-0.0186079443,0.0467942171,0.0949789286,0.0176645927,-0.1392452866,0.238650322,0.3752193749,0.5300636888,0.1959181428,-0.384483695,0.2212546915,0.009543499,-0.1241195798,-0.0480489098,0.116986312,0.752907753,0.1031509787,0.128127709,0.0797261968,-0.1877406538,0.0195771102,-0.0201137606,-0.1108019128,-0.0691101477,0.6327946186,-0.1857826561,-0.0004419827,0.1722396016,-0.2589039505,0.2579300106,0.4333907366,-0.2748917937,0.3943804801,0.0913248509,0.0145099927,-0.4266555309,0.1156502888,0.0076002297,-0.0006620407,0.3515398502,0.1298430711,-0.1225245222,-0.0634085983,-0.1447985321,0.3360650837,-0.0007123169,0.0569918416,0.17654185,0.2043727636,-0.0650676638,-0.3457590342,-0.0809137151,-0.1253281832,-0.0796888694,0.1865699589,0.255092442,-0.2642982006,-0.1799986511,0.1135271341,-0.1343258619,-0.4622514844,-0.1391333044,0.248648718,-0.3670700788,-0.2468735725,0.2646222115,-0.1381864101,0.5538969636,-0.2300098389,-0.1741389781,-0.048845686,-0.3722328842,-0.1011865214,0.2669962049,0.1744315922,0.0500913002,0.5129259229,0.120019339,0.1920259595,0.0220498238,-0.4392673075,0.110516794,-0.2284948379,0.2069483995,-0.1423638314,0.0113196932,0.0273552686,-0.1697660834,0.0343714319,-0.6640248299,-0.0404647663,-0.1837237924,-0.1404888332,-0.1423297822,-0.0716561899,-0.1670654267,-0.2310409099,-0.3831319511,0.2314613461,-0.0422587208,-0.0011927345,0.176378116,0.0477511175,0.1010505483,0.0398944467,0.039130982,-0.2410282493,-0.1167377904,0.2024116218,-0.0190567411,-0.0684883818,-0.4100342393,-0.0780256242,0.1992030889,-0.0565484688,-0.3131441176,0.1788439602,0.0802104473,0.2955591977,0.1335738152,-0.2585727572,0.1679107696,0.0657388568,-0.2541445494,-0.0744686499,-0.2153200954,0.0977984667,-0.0095135998,0.0383424684,-0.0902253464,-0.0231683739,0.1500937939,0.3610503972,-0.3247041702,-0.2081914842,0.2374608368,0.299572736,-0.0196775347,-0.1594928205,-0.1590788364,-0.1072138101,-0.0373966135,-0.0127198426,0.4947528839,0.0193359461,-0.6196320057,-0.1451299042,-0.2927596867,0.0674621463,0.026704818,0.2769380212,0.0816296861,0.389929384,-0.0392400287,-0.14346385,-0.4027861953,-0.1497973502,0.0862224326,-0.2975481153,-0.0040260102,-0.1056158394,-0.3922156394,0.2469236553,-0.4006909132,0.4426881969,0.2306683213,0.0109231537,-0.3408552706,-0.0570554137,-0.0801034197,-0.1442507505,-0.0052487375,-0.2516210973,-0.0010296402,0.1285383701,-0.216754213,0.0514426269,0.1760033667,-0.3028281927,-0.1659906507,0.4703876376,0.371863842,-0.130891934,0.1135945544,0.1338473409,0.1361092925,0.0361871719,0.0358648896,-0.0367148295,-0.1033096537,0.1158915907,0.1993099153,-0.0404810645,0.0637905076,0.0731047243,0.0363196954,0.1285290271,0.1600630581,0.3005918562,0.1555932909,-0.1027341634,-0.0154281603,0.1482216865,0.0602447502,-0.0183744822,0.1687557995,0.0271394588,0.0290747583,0.0911584049,0.0186826047,-0.2072654217,0.4296904206,0.2786149979,0.1548066884,0.350825429,-0.0583146662,0.1094834879,-0.3746370375,-0.0348961838,0.1976129711,0.2631520927,-0.3504345715,-0.317904681,0.4234219193,0.0498130508,-0.1697402745,0.1802553535,-0.1843903065,-0.0612634271,0.5094426274,0.2380772233,0.9417066574,0.0489503704,0.1216089204,0.0766973644,-0.0452697389,0.2165809274,-0.1111877859,0.1896009892,-0.0909776911,-0.1776074022,0.022636218,-0.056976784,0.0244592447,0.0748268366,-0.1539719105,0.0643789545,0.39220348,-0.293181181,0.0429655164,0.2920527756,0.2349118441,-0.1357379109,-0.2929735482,0.3304252028,0.0199170858,-0.2843275368,-0.1234237626,-0.1942331046,-0.024469737,0.0453879982,-0.3454398811,-0.0085163666,-0.0191935971,0.2232915163,0.0033211191,-0.3943594396,0.2245959044,-0.0967894197,-0.0795821995,0.0444108546,0.0121331569,0.2483435124,-0.0254079048,0.101903066,0.2050745338,-0.2166849226,0.2815557122,-0.1560412347,-0.2040401548,0.0519858152,0.2033017129,-0.3419334292,-0.4735870361,-0.0579155721,0.0973810777,0.2426350713,0.1909070313,0.2109098136,-0.0913755149,-0.2834507227,0.3226492405,0.1014433652,-0.0366420373,0.0254378673,0.2162097096,-0.165224418,-0.2183376551,0.0766888112,-0.0518770926,0.0838366151,0.2643003166,0.3100331724,-0.2390948087,-0.41031304,0.2053474635,-0.2721509337,-0.1819975376,0.126980558,-0.0672460571,-0.3076694012,-0.0702891573,0.2348424047,0.3584672213,-0.137729615,0.0107491594,0.0193136856,0.0221988223,0.0901667476,-0.101194337,0.0742193758,-0.2073342949,-0.1410028189,-0.1305303276,0.1294336319,-0.5457097292,-0.1376450062,-0.1348996758,-0.0441568345,-0.1623402238,-0.3284669816,-0.1346250176,0.029836623,0.2456754744,-0.0099642305,-0.2493411005,-0.4110135436,-0.2198423445,0.0503775701,-0.1087766588,-0.0128378086,0.0693586022,0.1396546364,0.0168735702,-0.1118446812,-0.1059896573,0.0392160006,0.1432077885,0.024566574,0.0403044708,-0.1036084443,0.059623763,0.08658766,0.2322267145,0.3226573169,0.0296093877,0.1701818705,0.1287757307,-0.1305343807,0.1204759628,0.298708111,-0.0925813764,-0.0350255258,0.0972074419,0.0943672135,-0.015285084,0.3551599085,0.343968451,-0.0618101358,-0.2039737999,0.0502544381,-0.0117060915,0.4028504193,-0.2746878266,-0.2617481351,0.2163854539,0.0278817359,0.1767932326,0.1275058836,0.4748389721,0.0820169523,0.2059003264,-0.027314242,-0.0529150441,-0.2058266401,0.1538242996,0.0426043943,0.0611602515,0.0427200124,0.2251801044,0.2639462352,0.28350088,0.1716182232,0.1852702647,0.3008708954,0.0638703555,0.2838211954,-0.0346225202,0.0014745957,-0.0785695016,0.2367218286,0.0040536006,0.357765913,0.1595300585,0.530660212,-0.1574803293,-0.0421089642,-0.2785945535,0.2211979777,-0.070511058,-0.13117598,-0.1057495102,-0.2256947756,-0.5487551689,0.0809351653,-0.030422518,0.0932186171,-0.1500343233,0.2290568352,0.0606501028,0.0424292572,0.1513429284,-0.1853091717,0.3263218403,-0.2256911397,0.1939611882,0.340052098,0.1265377402,0.0734115243,0.2791900933,0.1300249249,-0.172373414,-0.5292206407,0.2044355571,-0.3356584609,-0.0516980998,-0.1785787791,0.2300801873,0.2312515527,0.1207343116,0.4114298224,0.2705027461,-0.2854255736,0.337467581,0.2243564427,0.1519232839,-0.1590354294,0.2496271133,0.1258194894,-0.1061265767,-0.2925453782,0.142898187,-0.2088370323,0.1750535965,0.3760270178,-0.1050586775,-0.0181051865,-0.0967056826,0.1867471784,-0.1728384942,0.4929760993,0.1414275467,0.3855079114,-0.1699278951,-0.2186744511,-0.3894589245,0.1718560457,-0.1435579807,0.2549887002,-0.059821438,-0.0064551854,0.4720339179,0.2445926517,0.1023145691,0.1554280818,-0.4060432911,-0.0545075424,-0.4657457769,0.3790454268,0.1168249473,-0.024141578,0.0473590456,-0.0368034244,-0.029837871,-0.1166372299,0.2997751832,-0.1454803497,-0.2145262212,0.1421095133,0.3876218498,0.3363460898,-0.0053485902,0.3803088069,-0.254442513,-0.1544955522,-0.0229225382,-0.2346801907,-0.0626207888,0.0969196334,0.0420703478,0.3033840358,-0.1303761601,-0.2123456001,0.236772567,0.1345872432,-0.1637651324,-0.2017033547,-0.1257972568,-0.2080723792,0.0574901141,-0.2300475091,-0.0453169532,0.3624753654,-0.021898644,-0.1024932116,-0.1239110604,-0.0460221469,0.2022431642,0.0599170551,-0.2378761768,-0.2145025581,-0.0052859182,0.3751157522,0.1530994475,-0.2960548997,0.0091885775,0.2491264045,0.2405334413,-0.1600066274,0.1940912455,-0.0423701853,-0.0388324112,-0.086166054,0.3680944145,-0.0142264012,-0.2025555372,-0.1593997777,-0.1574810147]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/274","title":"PG-19","comments":"Got around to this today, and so far so good, I'm able to download and load pg19 locally. However, I think there may be an issue with the dummy data, and testing in general.\r\n\r\nThe problem lies in the fact that each book from pg19 actually resides as its own text file in a google cloud folder that denotes the split, where the book id is the name of the text file. https:\/\/console.cloud.google.com\/storage\/browser\/deepmind-gutenberg\/train\/ I don't believe there's anywhere else (even in the supplied metadata), where the mapping of id -> split can be found.\r\n\r\nTherefore I end up making a network call `tf.io.gfile.listdir` to get all the files within each of the split directories. https:\/\/github.com\/lucidrains\/nlp\/commit\/adbacbd85decc80db2347d0882e7dab4faa6fd03#diff-cece8f166a85dd927caf574ba303d39bR78\r\n\r\nDoes this network call need to be eventually stubbed out for testing?","body":"Hi, and thanks for all your open-sourced work, as always!\r\n\r\nI was wondering if you would be open to adding PG-19 to your collection of datasets. https:\/\/github.com\/deepmind\/pg19 It is often used for benchmarking long-range language modeling.","comment_length":127,"text":"PG-19\nHi, and thanks for all your open-sourced work, as always!\r\n\r\nI was wondering if you would be open to adding PG-19 to your collection of datasets. https:\/\/github.com\/deepmind\/pg19 It is often used for benchmarking long-range language modeling.\nGot around to this today, and so far so good, I'm able to download and load pg19 locally. However, I think there may be an issue with the dummy data, and testing in general.\r\n\r\nThe problem lies in the fact that each book from pg19 actually resides as its own text file in a google cloud folder that denotes the split, where the book id is the name of the text file. https:\/\/console.cloud.google.com\/storage\/browser\/deepmind-gutenberg\/train\/ I don't believe there's anywhere else (even in the supplied metadata), where the mapping of id -> split can be found.\r\n\r\nTherefore I end up making a network call `tf.io.gfile.listdir` to get all the files within each of the split directories. https:\/\/github.com\/lucidrains\/nlp\/commit\/adbacbd85decc80db2347d0882e7dab4faa6fd03#diff-cece8f166a85dd927caf574ba303d39bR78\r\n\r\nDoes this network call need to be eventually stubbed out for testing?","embeddings":[-0.2726430595,0.1023975164,-0.0761308819,0.0389732197,-0.3389706612,-0.1271181107,0.6120792031,0.1419056952,0.2338510901,0.0364079662,0.0945236534,-0.0275518056,-0.0765411258,0.8012996912,0.396630764,-0.0443375036,0.211777553,0.0587302968,0.0876412019,-0.0332939029,-0.313323468,0.3008314073,-0.0816275403,0.008407576,-0.5192525983,-0.4210985005,-0.2140348107,0.1362223029,-0.0564975664,-0.0908730552,0.0232849214,0.3214838207,-0.0683648884,0.7272160649,-0.0001202966,0.0550454371,0.0457218662,-0.1240919605,-0.2383431792,-0.2450154871,-0.0297817923,0.2153979242,-0.1817445755,-0.1727653742,-0.0562552996,0.0125058303,-0.0156625323,-0.0883732364,0.0534064882,0.1536814123,0.0577766858,0.1069048643,-0.1264804304,0.0249941312,0.3883236945,0.216906935,-0.3077532053,0.2421517074,0.4530335963,-0.1050634906,0.0934706926,0.3744819462,0.1002711728,-0.1324554533,0.2871436477,0.201012969,0.2314558476,-0.2365605384,-0.1251798123,0.3853826523,0.1542103142,-0.1849046797,-0.4427044094,-0.4211259186,-0.3595270813,-0.3846863806,0.1833167374,0.0560664907,-0.1714600921,0.1323085725,-0.0573672689,-0.5777279735,-0.0575946607,0.0093365954,-0.2454632521,0.6067339182,0.1900549978,-0.0070566242,0.3556369245,0.0511804558,-0.342284739,-0.2946127355,-0.0425567254,-0.3199821115,0.21934928,-0.2322885692,-0.0531322807,-0.1331907958,0.0561144575,0.174808085,0.2768690586,0.1526039988,-0.1641702354,0.1510071307,0.3916577101,0.2029719353,0.2193077058,0.3185265362,0.4196359813,0.1732527614,0.071432896,0.1075261906,-0.0424622297,-0.0206191055,-0.3254282773,-0.1253993809,0.0022389609,-0.2448345125,-0.0526662879,0.0094290199,-0.1056966186,-0.0337773263,0.1963540018,0.158449769,0.2853690386,-0.0132637536,-0.045060616,0.2140438706,-0.3329660594,-0.4193009138,-0.069764711,-0.0168797132,-0.160025999,0.3653959334,0.3247297406,0.0730307922,0.0274488293,-0.2186322808,0.0043942216,0.046492625,0.1876645386,-0.2817598879,0.3472299576,0.2262933552,-0.0165260583,-0.0629668012,-0.1108629107,-0.356213063,-0.3561598957,0.3167044818,0.0032073788,-0.1998966932,0.3542391062,0.0610938333,-0.2482744753,-0.0860883966,-0.4490329921,0.5363029838,-0.0586841106,-0.4931678176,0.1699602455,-0.1751819253,-0.5470999479,0.0817420334,0.0991437659,0.3235040009,-0.579390347,-0.2148498744,-0.1032194719,0.1039886922,0.0887611061,0.373023659,-0.1905645281,0.0898620114,-0.1698357463,0.9594690204,-0.0993583947,0.1975433379,-0.1675347984,0.2848426998,-0.3256671131,-0.0090381065,0.0218295734,0.0426345319,0.1131571978,-0.2048590928,0.0868954137,0.2947108746,0.0620340444,-0.0600580573,-0.2760382295,-0.1112532243,0.3911851048,0.0289650783,-0.0009515805,-0.4765183628,-0.1512331069,-0.2698189616,0.1740778685,-0.2358109951,-0.1551565081,0.0308812764,0.3874010146,0.0375196598,-0.0096713472,0.3352309167,-0.4020671248,0.2687511742,-0.012287871,0.1768016666,-0.1955323666,-0.2127089053,0.0497965924,0.1878638715,0.0376365334,-0.138183251,-0.0923651606,0.2016936839,0.0606909133,0.1877214462,-0.0458053388,-0.4293634295,-0.3604776561,-0.0348171927,-0.342639178,0.171873033,-0.0285636801,-0.0352945775,0.0439403094,0.0776204616,-0.0934830308,-0.2785622478,0.0286997221,0.2729766071,-0.001540517,0.3509218693,0.3272588253,0.6057779789,0.2587488294,-0.3310555816,0.0790191889,0.0507937931,-0.1362638921,-0.2429079413,-0.282789886,0.535728693,-0.2918588519,0.2532324493,0.0408431664,-0.2861541808,-0.1022042558,-0.0917572528,-0.2435637861,0.0972796902,0.3314112723,-0.4470750093,0.0934003145,0.0728303269,-0.2139514685,0.289252758,0.3497867584,-0.3756168485,0.2537577748,0.0915027782,0.0730308741,-0.3983814418,0.1558253467,-0.0336742252,0.3324222267,0.1943069398,0.3082286716,-0.1299704313,-0.0000342827,-0.3080719113,0.3471041024,0.0973096117,-0.2300127894,0.2968508601,-0.1393949986,-0.2148397416,-0.2560610175,0.076181002,0.1825117767,0.0446547084,-0.2624447644,0.2580497265,0.0524353907,-0.2476713806,-0.1175240278,-0.2869992256,-0.3177367449,-0.0545947403,0.2922195792,-0.174159795,-0.3121162653,0.1548203528,0.1175682917,0.521911025,-0.3469548821,-0.4283790588,-0.1055355668,-0.2466436923,-0.3328722417,0.0004457505,0.6615573168,0.1518442929,0.2738092542,-0.0665475428,-0.2734891176,0.0185178947,-0.3513025343,0.2038461715,-0.2112296969,0.3707543612,-0.1882704049,0.1462554038,0.1249106675,-0.1929099709,-0.0940069929,-0.729865253,-0.1402025223,-0.1654230952,0.0874549374,-0.0837273076,-0.3110301197,-0.5655396581,-0.2192698568,-0.0760293975,0.4412038326,0.0793059096,0.0764595345,0.3570248485,0.190431729,0.0222899653,0.3863989413,-0.2552841306,-0.1555346996,-0.0833690688,0.3025322556,0.0770509318,-0.0985095948,0.0111520272,-0.1882321835,0.0394399539,0.0151489293,-0.705918014,0.3068466187,0.2903669178,0.2597416937,0.0925674886,-0.3836504221,0.249778226,-0.1992362291,-0.0331020691,0.0072217989,-0.1347773224,0.351996094,0.1103857011,0.0250624456,0.0806694478,-0.1839890182,0.1030256078,0.6840956211,-0.2482760251,0.0036132797,0.0213546697,0.1816469878,-0.0131375417,-0.1698090732,-0.1296664476,-0.244847551,-0.2465710044,-0.4585032165,0.5736465454,-0.0593144037,-0.5932353139,-0.1298754066,-0.0980080068,0.0081369383,-0.3383710086,0.4411448836,0.2071642727,0.3476293087,0.0230940636,-0.1251212806,-0.1863764524,-0.1950679421,0.0506098792,-0.2279559672,0.2261791527,0.0466553122,-0.1756548434,-0.134314701,-0.355281055,0.3489300013,0.2551191151,0.1749023199,-0.3630930185,-0.0755884945,0.0227093659,-0.2219688892,0.3317024708,-0.1779330373,-0.0093995836,0.2630961537,0.0396290161,0.2976771891,-0.0351604111,-0.3403724134,-0.0567593351,0.5999386907,0.5400308967,-0.2375167459,0.1011598781,0.0914801061,0.1242351159,-0.110856235,0.2190769911,-0.1244838387,-0.1575267166,-0.0539802536,0.1433664262,0.163253516,-0.1406032443,-0.2619319856,0.1616323292,0.3779697716,0.2200572938,0.2634662092,0.043119248,0.1848716438,-0.0737323537,0.083595641,0.4948540926,-0.055793792,0.3722937405,0.5668678284,0.0574347973,0.4507494569,0.2335363925,-0.1680429578,0.3974840045,0.8063243032,0.1405192018,0.0611205176,0.079563871,0.0767573938,-0.3686742187,0.167948544,0.3405255079,0.0315885879,-0.2728308737,-0.3024204373,0.4311411679,0.1669760942,-0.1041932181,0.3814059496,-0.2048694789,-0.2611415982,0.5249561071,-0.166748032,1.1697301865,0.1368249208,0.2993491292,0.0809977278,0.4191167355,0.1900576651,-0.4884206355,0.1485393941,0.0181904696,0.0397296548,-0.1441792995,-0.1387221813,0.0473325774,0.1548210531,-0.3110102117,0.1173854321,0.3908536732,-0.0234330744,-0.0294194072,0.3881385326,0.1213229001,-0.0899503529,0.2296048254,0.0345119126,0.1406188011,-0.0271195285,0.0435832292,-0.2933674455,-0.4227819443,-0.106984742,-0.2674905062,-0.0756566226,-0.3035304844,0.0649928898,-0.1064226031,-0.3392098248,0.1419107765,-0.0207054019,-0.2713129818,0.0388257466,-0.1473768055,0.1514954716,-0.2022618353,0.0859386101,0.148146227,-0.0872521326,0.257950604,0.1409018636,-0.2104291767,-0.1471411586,-0.077167511,-0.2018414885,-0.4417015612,-0.3046221137,0.0685091317,-0.1561025828,0.1771123409,0.2317343801,-0.2645566463,-0.1544072926,0.0970616713,-0.0195703376,-0.2739710808,0.01865693,0.0106249489,-0.2378074527,-0.2421089709,0.2205875367,-0.194207266,0.3820822537,0.412489742,0.3496235609,0.0817624703,-0.2193911672,0.2541910112,-0.1768300086,-0.2578894198,0.1494480222,-0.0552133434,-0.1352878809,-0.3229970932,0.4571163058,0.4444956183,-0.0257337913,-0.0892640501,-0.0830921382,0.2062606961,0.2840724289,0.2498887181,0.4437870979,-0.0897608772,-0.0079296948,-0.0950463936,0.2125563025,-0.2410940081,0.0542754456,-0.2605373561,0.0188402422,-0.0955619067,-0.1735625863,0.1620260477,-0.1036812663,-0.0303379372,0.271404922,0.0478602871,-0.0993204713,-0.002333055,0.2161145061,0.0434173271,-0.0459819809,-0.0324763954,0.0988567472,0.161376074,0.0162634812,0.0282701049,-0.0949476883,0.1034392715,-0.2527863383,0.0637743622,0.0171489902,-0.2364721298,0.180967927,0.0299480725,0.429618001,-0.0823633745,0.2004124969,-0.0734905601,-0.0253967047,0.0024158603,0.2037974745,0.0448093489,-0.0230372995,0.1332615465,-0.1504480094,0.1010665447,-0.0339514352,0.4799183309,0.3069323897,-0.3384394348,0.0659379363,-0.0241238344,0.1052003577,-0.1385966837,-0.2768109739,0.1725607365,-0.1998487264,0.0095903091,0.4495720565,0.7169886827,0.0763067827,-0.0125052026,-0.028907815,-0.1171094626,0.1782934964,0.093899101,0.293269664,-0.3195465505,0.1339810938,0.0540504083,0.2150758803,0.2383484095,-0.0519319326,0.0629424974,-0.0008559275,-0.0867704675,0.2609267831,-0.2424817234,-0.4443643391,-0.1959195435,0.1845454425,0.1359019578,0.3729007542,0.0096737249,0.7265553474,-0.525434494,-0.1399843991,-0.168155551,0.1144493669,-0.129572764,0.0528739803,-0.4104481936,-0.2534973621,-0.7404458523,0.2486055195,-0.0349163599,0.0342244133,-0.2049781829,0.2294397652,-0.1614019573,0.1850544363,0.1864523888,-0.4132696688,0.2457098812,-0.3718052804,0.3357541561,0.1753768772,0.0800139457,0.178754285,0.2018623203,0.2501917481,-0.2673251927,-0.2864138186,0.0612864904,-0.1897000074,-0.0508238636,-0.0377173983,0.3149444759,0.1342568696,-0.0198590439,0.4170704782,0.0201172028,-0.0488515012,0.2668664157,0.3733466864,0.2590516508,-0.1369504929,0.4088306129,-0.2147059143,-0.2509362698,-0.2189854383,-0.2040759623,-0.2603701055,0.142344296,0.3417740464,-0.2362694591,0.073910825,-0.0458525233,0.0226539429,-0.1943698972,0.6136355996,0.1349563003,0.301556468,-0.4537058473,-0.1868509203,-0.4564939141,0.0626424104,-0.2675774992,0.1378531456,0.1533735394,0.0983270854,0.3756082058,0.3372198045,-0.089620769,0.0860821903,-0.363081485,0.1857966632,-0.3647125959,0.2050547898,-0.2918778658,0.0423910618,0.0448625274,-0.1548424214,0.1335463226,-0.1325845122,-0.1612814665,0.0911176279,-0.2634660602,0.1195243746,0.7089613676,0.2730458975,-0.1385563612,0.5048616529,0.0449625514,0.071045205,0.0308075771,-0.2031416148,-0.1938755512,0.1018581539,-0.3042173088,0.2639802992,-0.2278987616,-0.1310273856,-0.0010580582,-0.2271565199,-0.0590113811,-0.0204833597,-0.1881715655,-0.2876839638,-0.3768277168,-0.1003793702,0.3555051982,0.5817753673,-0.0689639747,-0.2167637348,-0.1972569823,0.0218351986,0.5056435466,0.0783831775,-0.0010331844,0.0162262507,0.0620123632,0.6226617098,0.1065224856,-0.4134062231,-0.1098800302,0.2143742591,0.1830640286,-0.1542503685,0.0011518802,0.157707274,-0.0097032478,-0.0284749176,0.5589556694,0.0301841628,-0.1919053346,0.0709358156,-0.2202951163]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/274","title":"PG-19","comments":"Ohh nevermind, I think I can use `download_custom` here with `listdir` as the custom function. Ok, I'll keep trying to make the dummy data work!","body":"Hi, and thanks for all your open-sourced work, as always!\r\n\r\nI was wondering if you would be open to adding PG-19 to your collection of datasets. https:\/\/github.com\/deepmind\/pg19 It is often used for benchmarking long-range language modeling.","comment_length":25,"text":"PG-19\nHi, and thanks for all your open-sourced work, as always!\r\n\r\nI was wondering if you would be open to adding PG-19 to your collection of datasets. https:\/\/github.com\/deepmind\/pg19 It is often used for benchmarking long-range language modeling.\nOhh nevermind, I think I can use `download_custom` here with `listdir` as the custom function. Ok, I'll keep trying to make the dummy data work!","embeddings":[-0.4496191442,-0.0325611532,-0.2415347844,0.0113263791,0.0196675546,-0.2363707572,0.3253330886,0.2554726601,0.3786719143,0.1375896037,-0.1407309175,0.1234682724,-0.2844292819,0.6086748838,0.2480923682,-0.172456935,0.1320279837,0.1412584335,0.0229877122,-0.0359676145,-0.3278507292,0.1621127874,0.0156882498,-0.1066643074,0.0994011313,-0.051836092,-0.1455862224,0.1118728817,-0.2036280185,-0.2397161573,0.0195881221,0.248938486,0.0873469114,0.3737565577,-0.0000990424,-0.1411964595,0.1471233666,-0.1750644296,-0.1681011617,-0.177079469,0.0087815961,-0.1589248925,-0.3505608141,-0.071904704,-0.2097429186,-0.1847786009,-0.0747273117,-0.2173962146,-0.0259989146,0.17593427,0.3331770301,0.1625520885,-0.0765999183,-0.1970894337,0.3186580837,0.205319494,-0.3026254773,0.1041432694,0.5350350142,0.0575420745,0.226691708,0.1472870708,-0.1392449886,-0.0871092007,0.1038857847,0.0163980406,0.3358004391,-0.1456792206,-0.0325880758,0.4370433092,0.2743554115,-0.2226375192,-0.1406696737,-0.1740690917,-0.2891650498,-0.5619824529,-0.1779934615,0.0258650444,-0.1150368303,0.1747485399,0.1253533512,-0.2363759279,-0.1189032719,0.1020026132,-0.1371812522,0.5368193388,0.1239046082,-0.1856379509,0.2636006773,-0.2258602828,-0.2271918505,-0.3083097041,-0.0334562063,-0.2204840928,0.024958536,-0.3071707189,0.0811232328,-0.0263818279,0.2497797906,0.0611995757,0.403575778,0.3741003871,-0.0422348492,-0.0319192,0.0785553828,0.1675983369,-0.026160175,0.1110615581,0.1165724248,0.1275400668,0.0928875804,0.2510779202,-0.019871138,0.2266147286,-0.3101396561,0.1070979536,-0.2500466704,-0.1984442174,-0.173829183,0.1036912948,0.0853611529,-0.0045561823,0.1078005731,0.3054639995,-0.0083418963,0.0905207396,-0.2202184498,0.1306272298,-0.1280262172,-0.4333171844,-0.1433338076,0.1360117197,-0.1882008761,-0.0605785958,0.2425047904,0.2005563974,0.0558305383,0.065563038,0.2242944539,0.2239504457,0.4162147939,-0.3934520185,0.2677246332,-0.0238544345,-0.0135099487,-0.2324438393,-0.0271218661,-0.2948922217,-0.4005902708,0.1251236349,0.1765996367,-0.0966551304,-0.0518541113,0.2899715006,-0.2850761116,-0.2268649936,-0.1316840351,0.6907501817,-0.3521029651,-0.2408408076,-0.000249237,-0.0981628969,-0.3826604486,-0.1678152829,-0.0682638437,0.2230497748,-0.7102313042,-0.0014367804,-0.0881650895,0.0876414925,0.1221907213,0.1479886919,-0.1221750304,-0.0569751337,-0.1786092818,0.4285411835,-0.0073873335,0.0634549111,-0.1386584938,0.0479716659,-0.1762387156,-0.0238591805,0.0830367431,0.244820863,0.266395539,-0.1704757959,-0.2056723237,0.3559737504,0.1017345339,-0.0471819565,-0.2808577418,-0.2230838239,0.3721271455,0.2732970715,0.0245760772,-0.37380144,-0.0200691894,-0.1720552891,0.2108671367,-0.3505474329,-0.1919389963,0.1363509148,0.369763732,-0.0901834667,0.0271727443,-0.1142846793,-0.3181369603,0.224214986,-0.1626094431,0.2529934943,0.139137581,-0.0102434354,-0.1896639317,0.2022197247,0.1474494487,-0.092829816,0.2675631046,0.3086658716,0.0451175719,0.1465414762,0.0040356345,-0.1700763553,-0.082858555,-0.214568153,-0.0482161976,0.168878451,0.0446545221,-0.0821390003,0.0733653083,0.0036325641,0.0582691059,-0.0483385585,0.096712701,-0.0101715019,0.0568860881,0.1058743894,0.3047047257,0.4284171164,0.1185785756,-0.184403941,0.233685866,0.3589049578,0.0456297994,-0.1627823263,-0.1102421731,0.7311078906,0.0216261838,0.0830856636,-0.0175756346,-0.0984648168,0.1634133458,-0.0129959052,-0.0342781581,-0.068840988,0.6526178718,-0.0427350178,0.219409138,0.0551450774,-0.2367326766,0.3291032612,0.2710447907,-0.3213002086,0.390402019,0.1643798947,0.0217107311,-0.3258359432,0.2371312976,0.086586386,0.3153596818,0.3786853254,0.2012796998,-0.0551148392,-0.0395657755,-0.1276569217,0.513776958,-0.1240727007,-0.1921029389,0.1431730092,0.166623041,-0.2695827186,-0.2591991425,-0.0274482388,-0.0819627047,0.0517943278,0.0231285542,0.0882518291,-0.1653073132,-0.1684926599,0.1481920928,-0.2837992609,-0.3677750826,-0.1133735925,0.206483528,-0.1017498001,-0.3563189507,0.1745751947,0.05030163,0.4389922023,-0.3019811511,-0.3839160502,-0.0820413157,-0.1343021989,-0.1664964259,0.2428562641,0.2744905353,0.0489178412,0.3300122321,-0.0091046281,0.0731624141,-0.0430292748,-0.347976625,0.1718115658,-0.0593698807,0.1950065941,0.0625412911,0.1053026542,0.1494202912,-0.0789449438,0.027824726,-0.5253397226,-0.0980188921,-0.1136682183,-0.1326740384,0.0018229119,-0.2441450059,-0.1484719068,-0.4197816849,-0.4171620011,0.1512208283,0.2760438025,0.1007145494,0.119177334,0.2399676591,0.085902065,0.2548600137,-0.1275292486,-0.3787027001,-0.2274725735,0.3552961349,-0.1580530107,-0.1372421384,-0.3094339967,-0.3070884347,0.1115585193,0.0119988611,-0.4052853882,0.1284037977,0.1792709082,0.2972586453,0.0215697009,-0.2398805022,0.3363254964,0.0931689814,-0.1702373922,-0.1667775661,-0.1932845116,0.1437280476,-0.1239264756,0.0541550629,0.0257418193,-0.1417310685,0.1266516149,0.6686883569,-0.3144363463,-0.1296614856,0.3166135848,0.2104554623,0.116296798,-0.1307317317,-0.1326116174,-0.2857694328,-0.0050991527,-0.2197553962,0.499222666,-0.033051841,-0.6249346733,-0.1822271645,-0.011847117,0.1265564412,0.0361822732,0.3441401124,0.2690191865,0.1792858094,0.0525719002,-0.0257917233,-0.4849683642,-0.1925655156,0.162387386,-0.3692144752,-0.1043040082,-0.1446269155,-0.3234770894,0.25035882,-0.3776231408,0.4971387982,0.2345060259,-0.1397440434,-0.2826575935,-0.1146584079,-0.1458333433,-0.1570810825,0.1001632959,-0.3264569938,-0.0554340594,0.2193365395,-0.0801129341,0.00940214,0.1120512336,-0.3385166824,-0.2300797999,0.4436789453,0.2835435867,-0.3600516021,0.116615437,0.021960292,0.2847560942,-0.0176445115,0.0540449806,-0.0834848434,-0.2652904987,-0.0221802816,0.1455366611,0.0209415257,-0.0475829355,0.009708656,0.034879785,0.1163578704,0.1714587957,0.3515789211,0.0141245211,0.0554833226,0.0487805828,0.1067455634,-0.023395149,-0.0305201337,0.153676331,0.374840647,-0.1545541584,-0.0265544113,0.0296901613,-0.2578403056,0.1693851501,0.4453938305,0.1684514731,0.1875218004,-0.2678750157,0.0785392374,-0.3695303202,-0.0039596721,0.0749214068,0.1194176525,-0.0936725661,-0.388566345,0.2877745926,0.0360755436,-0.1967306137,0.3596974909,-0.3817777932,-0.1403761953,0.5071076155,0.1715399921,1.069742322,0.1382474452,0.2277506739,0.2690109015,-0.1123169363,0.1769308299,-0.1013678014,0.0141016273,-0.0903584361,-0.1090837494,-0.0040301252,-0.1163992807,0.0125453398,0.2308494598,-0.3814270496,0.2143874615,0.1874138117,-0.0433947854,0.1430401504,0.2398749441,0.1523887217,-0.1314472705,-0.0872986838,0.303837359,0.0888430476,-0.2356543392,-0.0898464844,-0.2445459664,-0.1678673923,0.0147444746,-0.304255873,0.0222703982,-0.2616690993,-0.024637809,-0.2143444866,-0.3872902095,-0.0104752881,0.0437581725,-0.137022987,0.126951322,-0.1129901558,0.08242663,-0.0416604206,0.2649954557,0.1730926037,-0.2701754272,0.0655785501,-0.0969691277,-0.1505325437,0.054545261,0.0900318399,-0.2973219156,-0.4159227908,-0.0674654171,0.2315244824,-0.0399446413,0.0912665948,0.0526363924,-0.053316582,-0.2608888447,0.2784076631,0.1591484994,-0.1142067611,-0.0012002388,0.0115477266,-0.1131895408,-0.3527474701,0.0943329856,-0.0367180631,0.1103849933,0.2188691944,0.2783349454,-0.161568895,-0.3992879987,0.2303353101,-0.4040111601,-0.2551642656,0.234649986,-0.1682871431,-0.4499288499,-0.2994424701,0.1827792972,0.3982261419,0.0267371275,0.0362498797,-0.1406245679,-0.1324993372,-0.0263724532,0.1978906095,0.2039725929,-0.2525269985,-0.2408903986,-0.1247691959,0.0301049221,-0.5307038426,0.0459258854,-0.2104512751,-0.1087200344,-0.1318976432,-0.1910294741,-0.0001299965,0.0529372357,0.1701429933,0.1302685142,-0.3670071959,-0.3362665176,-0.2466502488,0.0947991908,-0.1898582429,0.020323528,0.1746068001,0.0729746595,0.0254575834,0.0235371925,-0.1450505853,0.0047437125,0.0142726172,0.0440730564,0.1607226431,-0.0179693047,-0.1078303009,0.1182438433,0.162570864,0.4822970331,0.0284038968,0.2059419304,0.0055254856,-0.1750848591,0.134917438,0.1915425509,0.0970890447,-0.0527100377,0.3109689951,0.1518612355,-0.0850276947,0.237843439,0.4349475503,0.0667007416,-0.2614818811,0.1674642414,0.2404797524,0.3586176038,-0.3505341113,-0.366356492,0.2699556649,-0.0571050048,0.2011396587,0.2634993494,0.5090160966,0.0559938774,0.3693817556,-0.043278072,-0.0208404772,0.091416873,-0.003949943,0.1869081855,-0.165469557,0.0350198373,0.1175502017,0.1931996942,0.1532424688,0.1545321941,0.0882990062,0.2996974587,-0.0034258533,0.1333684474,-0.0346180759,-0.2629000247,0.2150958329,0.3194667399,0.0775773898,0.457318306,0.2147779167,0.4042337239,-0.3567391038,0.0112961447,-0.2824988961,0.1134254485,-0.0433544181,0.0305098742,-0.0957247093,-0.2550141513,-0.6515776515,0.1020775363,0.05943878,-0.1026330516,-0.0217787474,0.1016777009,0.1845874488,0.0675997213,0.0181130134,-0.1648770571,0.1479998529,-0.3413234949,0.1909485459,0.3641978502,0.0704605877,0.165532887,0.2569619119,0.1940617859,-0.0338752754,-0.4499942362,0.0075200452,-0.2952581048,-0.0273489673,-0.4036979377,0.2954619527,0.1795177609,0.087289609,0.5714366436,0.2134309709,-0.3286542892,0.1666833907,0.2962334156,-0.02350308,-0.2522214055,0.465334326,0.1142127588,-0.1969916672,-0.303317219,-0.1175309941,-0.2748638988,0.0919513777,0.3361843228,-0.0773764774,0.1398292035,-0.1219176054,0.1788094938,-0.1884625256,0.4774327278,0.0518542789,0.3941797316,-0.3209950924,-0.1259215027,-0.4690095484,0.2998864353,-0.2483670712,0.024125129,-0.0713929906,0.0925808996,0.5485860109,0.4447729886,-0.0576307401,0.231860593,-0.2167255133,0.2589296997,-0.5738062859,0.2676143646,-0.0118651809,0.0888423994,0.1441946328,-0.3371603787,-0.0335576534,0.0031609554,0.2609309256,-0.1328394562,-0.3144959509,0.0535107516,0.3894987404,0.4873784482,-0.0653240681,0.3704745173,-0.2717778385,-0.2141725421,0.1481594443,0.0041300822,-0.0875169933,0.0066097104,0.0812319443,0.2170924246,-0.2944077551,-0.230971545,0.0386710837,0.0070218141,0.0198765285,0.004748852,-0.2540014982,-0.0574084409,-0.1479108632,-0.1070290208,-0.0397804826,0.4793085456,-0.0568372533,-0.0091183102,-0.1438842267,0.04604223,0.423973918,0.0731905252,-0.2884407043,-0.057094343,-0.0271566007,0.2436552942,0.0364439189,-0.287578851,0.1055402458,0.1928028166,0.236960426,-0.0671973601,0.1631648988,-0.0542596169,0.0880757868,-0.0786968097,0.4497817159,0.1118277833,-0.1634517759,-0.0498632863,-0.2840099037]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/270","title":"c4 dataset is not viewable in nlpviewer demo","comments":"C4 is too large to be shown in the viewer","body":"I get the following error when I try to view the c4 dataset in [nlpviewer](https:\/\/huggingface.co\/nlp\/viewer\/)\r\n\r\n```python\r\nModuleNotFoundError: No module named 'langdetect'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 54, in <module>\r\n configs = get_confs(option.id)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 48, in get_confs\r\n builder_cls = nlp.load.import_main_class(module_path, dataset=True)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 57, in import_main_class\r\n module = importlib.import_module(module_path)\r\nFile \"\/usr\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\nFile \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\nFile \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\nFile \"<frozen importlib._bootstrap>\", line 967, in _find_and_load_unlocked\r\nFile \"<frozen importlib._bootstrap>\", line 677, in _load_unlocked\r\nFile \"<frozen importlib._bootstrap_external>\", line 728, in exec_module\r\nFile \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/88bb1b1435edad3fb772325710c4a43327cbf4a23b9030094556e6f01e14ec19\/c4.py\", line 29, in <module>\r\n from .c4_utils import (\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/88bb1b1435edad3fb772325710c4a43327cbf4a23b9030094556e6f01e14ec19\/c4_utils.py\", line 29, in <module>\r\n import langdetect\r\n```","comment_length":10,"text":"c4 dataset is not viewable in nlpviewer demo\nI get the following error when I try to view the c4 dataset in [nlpviewer](https:\/\/huggingface.co\/nlp\/viewer\/)\r\n\r\n```python\r\nModuleNotFoundError: No module named 'langdetect'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 54, in <module>\r\n configs = get_confs(option.id)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 48, in get_confs\r\n builder_cls = nlp.load.import_main_class(module_path, dataset=True)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 57, in import_main_class\r\n module = importlib.import_module(module_path)\r\nFile \"\/usr\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\nFile \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\nFile \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\nFile \"<frozen importlib._bootstrap>\", line 967, in _find_and_load_unlocked\r\nFile \"<frozen importlib._bootstrap>\", line 677, in _load_unlocked\r\nFile \"<frozen importlib._bootstrap_external>\", line 728, in exec_module\r\nFile \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/88bb1b1435edad3fb772325710c4a43327cbf4a23b9030094556e6f01e14ec19\/c4.py\", line 29, in <module>\r\n from .c4_utils import (\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/88bb1b1435edad3fb772325710c4a43327cbf4a23b9030094556e6f01e14ec19\/c4_utils.py\", line 29, in <module>\r\n import langdetect\r\n```\nC4 is too large to be shown in the viewer","embeddings":[-0.2870647609,0.1350371242,-0.0991052836,0.3264829218,0.2287869304,0.0880193412,0.2990790009,0.4050920606,-0.0533804744,0.0054588234,0.018971486,0.2833926678,-0.3927465379,0.247559458,0.1984093189,-0.1616899818,-0.0224667061,0.3656271696,-0.0160909574,0.0760527104,-0.0129785687,0.0204963647,-0.3416695595,-0.0439890102,-0.0663000345,-0.1799120903,0.1171200648,0.1152004078,-0.1176986471,-0.8207744956,0.2420960367,-0.0077897329,0.3344399333,0.2266331464,-0.0001138984,0.1619270891,0.290607214,-0.2695138752,-0.2590911388,-0.0794834867,0.0853165984,-0.1979011744,0.2246707678,-0.2014019191,-0.3439427912,-0.2466224283,0.0826018155,-0.4486275315,0.0570239089,0.3786529005,0.2180176675,0.0090096863,0.0853899047,0.4327436686,-0.4526798427,-0.1593962163,-0.1778487563,0.2329738736,0.056256216,0.2875347435,-0.154855147,0.228363499,0.1218683794,0.1884315759,0.1365233958,-0.1501538754,0.057672482,-0.2386935055,0.0576883182,0.0910521373,0.3618771732,-0.3305667937,-0.0773282051,-0.0625548288,0.1799002588,-0.3523288667,0.0812347233,0.2702074349,0.0723463446,0.0691400394,-0.214759618,-0.0136543047,-0.2788703442,0.3241462111,0.0734650642,0.3181271553,-0.1727860868,0.2107585222,0.1218228713,0.0313147083,0.2299191505,0.0770477206,-0.018878758,0.2382939309,-0.2737355828,-0.1076846048,0.0811361745,0.5283503532,-0.474501431,-0.0499806255,-0.1871330738,0.0612604842,0.0799970925,-0.0546450689,-0.2083643079,0.0701337904,0.2313616127,-0.0048931055,0.3078097701,0.076201275,0.067602694,-0.0265761688,-0.2556738257,-0.2339130044,-0.5874486566,0.1961396337,0.1800377071,-0.2396205515,-0.3553000093,-0.0261300914,-0.3746516705,-0.1725396663,0.0147467945,0.5665973425,0.0212136135,0.1899191737,0.170911774,0.0499776378,-0.2455647886,-0.5403926373,-0.0685602427,0.1485345364,-0.4316160679,-0.1358396411,0.2088176012,-0.017401902,0.3493711948,-0.1879731715,-0.060478393,0.1171407104,0.1290458888,-0.1122285724,-0.0465062,-0.1039340347,0.0105684604,-0.0262273327,0.327206105,-0.2364487946,-0.0831541717,-0.1740554422,0.0550550036,0.0268701613,-0.2307028919,0.1571808904,-0.2268617153,-0.0044164355,-0.3582608998,-0.0116348602,0.0416895784,-0.1791594177,-0.0085718939,-0.3142136335,0.1430772692,-0.2192335278,-0.0162893422,0.342220217,-0.3193509579,-0.1713897884,-0.2082974464,-0.0523265041,0.2417826504,-0.0073816334,-0.1995040029,0.0999416709,-0.1920406967,-0.0121488636,0.6231903434,-0.3457804918,-0.4665177464,0.186451748,-0.0825869441,-0.0050302031,0.0772351623,-0.0692925528,-0.1019246951,0.0291265212,-0.1724943519,0.2861809731,0.1145680621,0.1394236684,-0.1769938618,-0.0381891876,0.8263379335,-0.0603609495,0.0172987543,-0.1059594378,0.3000529706,0.2110196799,0.4284287095,-0.2560963333,0.1173827276,0.0859128013,0.1346181482,-0.0881804526,-0.0641892105,0.2379666418,-0.5763319731,0.0518020429,-0.1164698303,0.2236281335,-0.083804369,0.0918257385,0.0109704267,-0.341456145,-0.4050937891,-0.364410013,0.2298740894,0.4313929975,0.1295526773,0.2657046318,0.1496113986,0.4117803276,-0.3463265002,-0.1504212171,-0.1508454233,0.3716615736,-0.1539832205,-0.1502525508,0.2048738301,0.3593421578,0.3174504042,-0.3790105283,-0.0842761695,-0.0407627933,0.1110021845,0.2320052981,-0.0022006326,-0.0114519671,0.2539611459,-0.5675045252,0.334597528,0.4556268752,0.1816173941,0.0998782367,0.1549402773,0.1761346161,0.1574990749,0.0831375718,0.2813569903,0.0613164268,0.4688104093,0.0454806834,0.2703546584,-0.1168802753,0.3929905593,-0.0687038228,0.5313211679,-0.0947694927,-0.182457,0.352322787,0.4382533133,0.219332993,-0.0656407401,-0.0846904963,-0.2076654285,-0.2240628302,0.6114386916,0.245934695,0.2101754546,0.1751014441,-0.1641331464,0.0811554715,0.0931098536,-0.0003330333,0.0386621393,-0.1065187752,0.0869482532,0.0829689801,0.1369113773,-0.0585939661,-0.2131668329,-0.4350110888,0.1604523212,0.4031144381,-0.165035978,0.1324010044,-0.4361777306,-0.4954040945,0.1311332732,-0.1790635139,0.0054828799,-0.3354031742,-0.3390111625,0.1187374294,0.0397867002,0.1147261038,-0.6923130751,-0.1468776613,0.0068628816,0.2009674907,-0.1160109863,-0.1438724101,-0.348908782,0.138984561,0.3802813888,0.2230485231,0.416467011,-0.1873108596,0.0529181026,-0.0521972179,-0.457906723,0.1693881601,-0.221523121,0.0065255589,-0.1714200974,0.1143924147,-0.1437484771,-0.2654653192,0.0510844551,-0.2595849037,-0.1993218511,0.0423389338,-0.2049298137,-0.1166988537,-0.1274091303,-0.5350187421,-0.6607378125,-0.4705260992,0.0444788076,0.2436685562,0.0528715514,0.196878776,-0.2456681728,0.3473730683,0.0343967006,0.0491862409,-0.0574500561,0.1691003144,-0.0360736847,-0.3192757666,-0.3829473555,-0.0415667966,-0.1333092749,0.5097364187,-0.178141281,-0.2892674506,0.3398669064,-0.1057196632,0.2687402368,-0.028729707,-0.2405575812,0.3990261257,0.0757973045,0.1185734123,-0.1885519028,0.1063823551,-0.1426251233,-0.0129227843,0.2990038395,-0.1431217045,0.2302153111,0.1308002174,0.3550739288,0.3223090172,-0.3977171481,0.5291985869,0.0197732579,0.417801559,-0.0247180779,-0.215519011,0.4562401474,0.1188106686,0.20398058,0.3340131044,-0.3398112059,-0.3325009346,-0.0217493661,0.0522003323,-0.428809166,-0.0658312291,-0.0586220771,-0.0236377213,0.1317363679,-0.1042404696,-0.0192027837,-0.0431746058,-0.5361050367,-0.0928066671,0.3632381856,0.0956871808,0.2891862988,-0.3549180329,-0.3607620895,-0.4302686155,0.2223438174,0.2162128538,0.5014207363,-0.1036703885,-0.0847114623,0.096204035,-0.1246119067,-0.0728956982,-0.1046550497,0.1449789554,0.4536512792,0.0713677555,-0.2213303298,-0.0027203003,-0.2214238346,0.3880017698,0.3344833255,-0.0200232957,-0.0933639407,-0.2613000274,-0.0482367873,0.1953405142,-0.0297672208,0.074329108,-0.1043660417,0.157108292,-0.2835039794,-0.1474068612,-0.0723413154,0.1833611131,-0.0197628327,0.1329741776,0.1695031375,-0.1664313823,0.0997185484,0.5710487962,0.2408226877,0.1240664795,-0.005974608,0.0336864181,0.1323225498,0.1731116772,0.3377902508,0.0717414767,0.0202849824,0.2542455196,0.0389269702,0.3234991133,0.0355049074,-0.0938878655,0.2057034075,-0.258734107,0.0879523307,-0.170290634,0.4216626585,0.1850753576,-0.0326217823,-0.2741675079,-0.2102848738,0.2583295405,-0.3840016127,0.0301290173,0.3236259818,0.2880037725,-0.3082944751,0.0652989447,0.0101186102,0.5352099538,0.1937631816,0.0627235547,0.324326545,0.1368076056,0.2695686817,-0.281973213,0.1914853454,-0.2641299367,0.2558116615,0.0259507038,-0.0901728421,0.2693756819,-0.0653592423,-0.3506756127,0.1445467919,0.0965498611,-0.2095048577,-0.0816913247,0.2557323575,0.1812137812,0.0458588712,-0.4563609064,0.1596089453,-0.3980472386,0.1779974997,-0.1269298196,-0.0311075784,-0.1528747529,-0.0097895516,-0.1149867699,0.2417108864,-0.0898342952,0.1279369146,-0.1632106006,-0.2416040897,0.4109326899,0.2450141311,0.1619673669,0.132191211,-0.2230581939,0.0757318512,-0.253793925,-0.1448033601,0.1731558889,0.1899707913,0.0731192082,-0.0926003382,-0.1370249689,0.1975016147,0.0767103434,-0.3791917264,0.1430263519,0.0413949564,-0.1246516332,-0.0060671228,-0.2319290191,0.1259390861,-0.2088929415,-0.0563498177,0.1024319381,0.0703455508,-0.2670338154,-0.236674428,0.3569884896,-0.1304947734,0.0763656273,0.1659492552,0.2389461994,0.0041117799,0.3647618592,0.3525829613,-0.403768003,-0.2828049958,-0.1536041051,-0.0238464978,-0.1358999461,-0.1468826979,0.148407355,-0.0671765506,-0.0144890482,0.2038818598,-0.0935997963,0.0242095795,-0.1564353257,-0.4283746481,-0.4843542278,0.0953257978,-0.5270940065,0.0300761275,0.1867851913,0.4279649556,0.0232755914,0.1716532409,-0.2890460491,-0.1047454104,-0.4422086179,0.1190160662,0.2209243178,-0.2009991705,0.0708543807,0.0906298831,0.1046380326,0.0203907639,-0.2586152852,-0.1900250167,0.0240565389,0.090353556,0.3225194812,-0.3218845725,0.1099673808,0.0172846951,-0.0589997992,-0.054112047,0.1439170241,-0.0616090223,0.0052360762,0.4631858468,0.0014976517,-0.2747288346,0.1219943687,0.0338947475,-0.060014464,0.0284332186,0.1997825354,0.2326895297,0.1485240757,0.1944516748,-0.1662381738,-0.0349726975,-0.2350473851,0.2438787371,0.1601391137,-0.3656629026,0.2828674018,-0.1989877075,-0.0242623966,-0.014225577,-0.152508378,-0.0583976731,0.5339083672,0.1066264883,-0.4307010472,0.1287889034,0.3404809833,0.0456478372,0.2525120974,-0.036950767,0.1129873767,-0.1264264733,0.2687448263,-0.0623844899,0.3088028133,-0.1222504824,0.2135567814,0.208974719,-0.2066344917,-0.0283270571,0.2681111097,0.108915396,-0.0498913117,0.3627180457,-0.1166649908,0.1354022175,-0.1222662851,0.0208917577,0.1319900155,-0.240756914,-0.0642354488,0.0499209464,-0.0628646538,0.1348443478,0.0553603359,0.2418654114,0.0081448564,-0.158197403,-0.0761432499,0.068813242,-0.321374476,0.2022487968,0.0478126109,-0.0174537189,-0.1494280398,0.0503273681,-0.0715504512,-0.2208399773,-0.0018962671,0.2852265835,-0.3207240403,-0.114421308,0.003813379,0.1872372925,-0.0400857478,0.0647593737,0.3710288405,-0.0556325763,0.2005630434,-0.3217252493,0.1599368602,0.1289785802,0.0930161625,-0.0034734346,-0.1035784855,0.1779958755,-0.1408418268,-0.0568981767,0.337477982,0.333646208,0.2515533566,0.148223713,0.0779190063,-0.1132166907,-0.0000257711,0.2125873715,-0.1301632226,0.1344485432,0.1911236793,-0.1465982646,-0.0624839924,-0.1491891146,-0.035388317,-0.5590645075,0.246627599,0.2361249328,-0.1137324646,0.0813888833,-0.3415862918,0.0571262017,-0.0229264162,0.3480712175,0.0832609385,0.3589696884,-0.0287969578,-0.4093111157,-0.4237415195,0.3132614195,-0.2937932909,-0.2125501484,-0.1463072151,0.0417004339,0.2168198079,0.192244187,0.3889135122,0.6867372394,0.263964206,-0.1327627301,-0.3481219709,-0.1302443147,0.1323644519,0.0344807431,-0.2482826412,-0.354300648,0.0622293577,0.0757573247,0.0914267153,-0.1689586788,-0.0130717242,-0.1445459723,0.0065388572,0.3430753648,0.0747320279,0.8262761235,-0.100974597,0.0035116253,-0.3622245491,-0.1715513468,-0.094602935,-0.0301982593,-0.0007193316,0.4857544601,-0.1177248135,-0.0789732486,-0.2859633267,0.5481299758,0.1516160071,0.2824723125,-0.0252542663,0.358494401,-0.4694617987,0.1657602787,0.0002465965,-0.0335053653,0.0734402388,-0.1404440701,-0.2260234207,-0.0908410698,0.4036456645,-0.7680633664,-0.2236157954,0.1093816683,0.115406014,0.0040023886,-0.1272070855,-0.3960362971,0.0908848867,0.1806368828,-0.0713818371,0.1000078395,0.3248119652,-0.192165643,-0.0043443339,-0.0770248994,0.3598084748,0.0226368643,0.0075843902,-0.0804032236,-0.254881233]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/267","title":"How can I load\/find WMT en-romanian?","comments":"I will take a look :-) ","body":"I believe it is from `wmt16`\r\n\r\nWhen I run\r\n\r\n```python\r\nwmt = nlp.load_dataset('wmt16')\r\n```\r\nI get:\r\n```python\r\nAssertionError: The dataset wmt16 with config cs-en requires manual data. \r\n Please follow the manual download instructions: Some of the wmt configs here, require a manual download.\r\n Please look into wmt.py to see the exact path (and file name) that has to\r\n be downloaded.\r\n . \r\n Manual data can be loaded with `nlp.load(wmt16, data_dir='<path\/to\/manual\/data>')\r\n```\r\nThere is no wmt.py,as the error message suggests, and wmt16.py doesn't have manual download instructions.\r\n\r\nAny idea how to do this?\r\n\r\nThanks in advance!\r\n\r\n\r\n","comment_length":6,"text":"How can I load\/find WMT en-romanian?\nI believe it is from `wmt16`\r\n\r\nWhen I run\r\n\r\n```python\r\nwmt = nlp.load_dataset('wmt16')\r\n```\r\nI get:\r\n```python\r\nAssertionError: The dataset wmt16 with config cs-en requires manual data. \r\n Please follow the manual download instructions: Some of the wmt configs here, require a manual download.\r\n Please look into wmt.py to see the exact path (and file name) that has to\r\n be downloaded.\r\n . \r\n Manual data can be loaded with `nlp.load(wmt16, data_dir='<path\/to\/manual\/data>')\r\n```\r\nThere is no wmt.py,as the error message suggests, and wmt16.py doesn't have manual download instructions.\r\n\r\nAny idea how to do this?\r\n\r\nThanks in advance!\r\n\r\n\r\n\nI will take a look :-) ","embeddings":[-0.2032090873,-0.2609338462,0.0847411156,0.4479438066,0.1586989015,0.087288484,-0.0284962058,0.1094017923,0.1474179924,0.0993779078,0.119574219,0.0673467517,-0.3464436233,0.0702806562,0.4917528331,0.2811710835,0.0055067469,0.1168495193,-0.2584639192,-0.2314175665,-0.1898065209,-0.0004141324,-0.2789565921,0.3030253947,-0.0382775553,0.146947816,-0.0801305473,0.2744104266,-0.1502615213,-0.1514229029,0.3784234524,-0.0559977442,0.4748617709,0.323204726,-0.0001256382,-0.0879309326,0.1764269024,-0.3160282969,-0.1157878935,-0.5956392884,-0.7075554729,-0.4119486809,0.0216787215,-0.1963145435,0.0960646123,-0.0046631284,0.1328487545,0.0466791727,0.1787003875,0.4759156704,0.0590568185,0.2938591242,-0.2106312215,-0.0816476792,-0.0877719,0.0263922252,0.1549755335,0.140037626,0.1670156717,-0.1636904329,0.39506495,0.1765840352,0.0234547127,-0.0069893938,0.302061975,-0.1114228293,0.2137804627,-0.1281391829,0.0937663391,0.186934948,0.5313196182,0.2266560942,0.0463274494,0.2488426268,-0.0014376147,-0.1782383472,0.1975896358,0.5925871134,-0.4668091238,0.180858627,-0.302390635,-0.4619061947,-0.2188048959,0.4127159417,-0.1715803295,0.8694525361,0.030385131,0.0613516942,0.2514004409,-0.1347563565,0.0728246346,0.1910019368,0.2712785006,0.5263392925,0.2084175497,0.0495299473,0.0019297422,0.0845109075,-0.0817976668,-0.0611642562,0.1113890558,0.0957554579,-0.2016190886,0.1494632065,0.2542185783,0.2351026386,-0.1131266803,-0.2297756374,-0.0463355407,0.0409856588,-0.0246483348,0.0712571144,-0.4625157118,-0.4067932367,-0.1101554409,-0.079536885,0.0637305602,-0.1397788525,-0.02959463,-0.3082052171,-0.0959859192,-0.3214743435,0.0682087541,0.3794014454,-0.1877456307,0.1921484321,0.1643137932,0.019907305,-0.3785243034,-0.0758857355,-0.0024163141,0.2469579428,-0.1332971901,-0.1710459441,0.1933169216,0.1206538826,0.1897274405,-0.0481291264,-0.040068198,0.018518962,0.0645022616,0.1634059101,0.0298730731,0.1265003234,-0.1129997298,0.1455018371,0.0825650692,-0.4395136535,-0.2857097685,0.2563743591,-0.3056466281,-0.5183034539,0.0274471249,0.0649998933,-0.0983979329,-0.166884169,0.1392841041,0.2687206566,-0.2615542114,0.1162484288,-0.1320860684,0.0391998291,-0.0242384877,-0.0261780135,0.3038374186,0.2482482642,-0.3718227744,-0.1807021201,-0.212978974,0.0155458525,0.3557551503,0.4238137901,-0.2677719593,0.0711493641,-0.277731657,-0.2089508772,0.4677195549,-0.2200302184,0.0476302654,-0.0012112861,-0.066855669,-0.0510162897,0.0309800766,0.0972692147,-0.4598389864,-0.125750646,0.1554449499,0.5852584243,-0.0392331667,-0.2184479237,-0.0849417821,-0.2738132477,0.7465341091,0.0852456838,0.4042404294,-0.2914254963,0.2697165012,0.3450931609,0.2901440561,-0.1004640982,-0.3245882094,-0.0096301017,0.1832448393,0.1309990287,-0.0653810203,-0.0161950458,0.1026474014,0.0904115215,-0.1369473934,0.2843018174,-0.0880293846,0.1580686569,-0.2195499986,0.0791667253,-0.1404673308,-0.0824625641,-0.0589747652,0.2487696409,-0.1295334101,-0.0478750952,0.118860729,0.0918182656,-0.0499598794,0.0407426655,-0.1342005581,-0.0098155327,-0.181139797,0.2203399688,-0.2095156759,0.2100671828,0.0372780301,-0.1170799732,-0.1383994073,0.0344720148,0.0343551971,0.2549994588,0.1356590688,0.0831474289,0.1353216469,-0.6820371747,0.1211575121,0.3320544362,0.1453852952,-0.0344374962,0.0944000781,0.0496791452,-0.2669276297,0.1390612274,0.3048036098,0.1054372862,0.366114974,0.1059423909,-0.1252699643,-0.1226584688,0.2658379674,0.350168258,-0.22903952,0.1456327289,-0.3829849064,-0.0691741556,0.4373107553,0.0856833607,0.2803063095,0.106266126,-0.2537598312,-0.0502616838,0.0649071485,0.0101542957,0.1912236214,0.0028322414,0.1219542101,0.0433564447,-0.0796474293,0.0388576649,0.3390303552,-0.0293928422,0.0340381116,-0.1732019633,-0.2281238288,-0.0200032759,-0.1404071897,0.2680369914,-0.0820908323,0.1529053301,-0.210799709,0.0918323249,-0.4845943451,-0.1214153916,-0.720187366,-0.7958168387,-0.3554048836,-0.0210778546,-0.0982050225,-0.3160211146,0.1440277249,0.1651905477,0.1124088839,0.1148556024,-0.2427779585,0.2218849063,0.1640188098,-0.3285734951,-0.3183276355,-0.1509352624,0.4058619142,0.3754656911,0.3025496006,-0.5222748518,0.2195051163,0.2599745095,-0.5646830201,-0.0135503327,0.2248778343,0.2938760519,0.0385970958,0.2492535859,-0.1057242453,-0.2908221185,0.1712700874,-0.1055032536,0.135911867,0.2751491964,-0.1171409637,0.1235249862,0.0437096842,-0.584317863,-0.827403307,-0.110429652,-0.1263685077,0.3493086696,0.1217573881,0.3987831473,-0.0009900174,0.202378422,0.172352314,0.3017950654,-0.1117335185,0.0252560284,0.4651223123,-0.1116987169,-0.2405575514,0.3765815198,-0.0502404869,0.1986069232,0.0075646653,-0.0630991161,0.3415356576,0.0452283807,-0.0104959738,0.0267905034,0.2149285227,0.3703528047,-0.0319755264,0.1513753533,-0.0674680769,-0.160446763,-0.1588190645,0.06318409,0.6361302137,-0.1132637858,0.0249636173,0.2146148682,0.7155183554,-0.3504467607,-0.0466186628,-0.0467684157,0.1204661876,0.3089801371,0.2776693106,-0.6533325315,0.0064092758,0.0417467318,0.0211619288,0.4683234394,0.0919104144,-0.2348324507,-0.4354780912,-0.0545058697,-0.3411700726,-0.068179287,0.1992225051,0.125066489,0.3940690756,0.3035158813,-0.0696883723,-0.3903332651,-0.2156504691,-0.0497697331,0.63053298,-0.2102300078,0.2741852105,0.1058743149,-0.5548173785,-0.1673139781,0.0871955901,0.0743479505,0.5199397802,0.0447409414,-0.2091834843,-0.0232724417,-0.1645722836,-0.0201311689,-0.0334580094,0.1135781854,-0.0320171006,-0.021835275,-0.3885374963,-0.0643817261,-0.2841005325,-0.1655434668,0.4871377647,0.0347852372,-0.0462299064,-0.0984231755,0.0087897554,0.4295887053,-0.4000371993,-0.0154372938,-0.2484155148,-0.3337549269,-0.1355270147,-0.026453374,-0.2068770826,0.0757999495,0.1207684353,0.2506487072,0.037703719,0.2758741379,0.1679621041,0.0422733314,0.0707733482,0.2588818371,-0.3893169463,-0.0023834188,0.3079742491,-0.2496248633,0.2805514932,0.049708873,-0.2128594965,-0.3946429193,0.0611289889,0.5383645892,0.1366768777,-0.0559212603,-0.0898159444,0.3559780121,-0.0449533053,-0.4754359126,0.3646655381,0.1486343592,0.1966448277,-0.3012838662,-0.4342274964,0.2005020827,-0.0572193898,-0.0246877577,-0.0884506106,-0.3878523707,-0.4024187922,-0.0983402058,0.1435993761,1.1405118704,0.0115537373,0.1424195468,0.085057795,-0.0599198788,0.1389082521,-0.0251544286,-0.1814221889,0.027518142,0.0723845586,-0.1423634291,-0.0674828291,0.0798184499,0.3640608788,-0.3223861754,0.4465421438,0.2066072524,0.1566264927,0.3413663208,0.2470058799,0.1186033934,-0.2108003497,-0.2809899449,-0.0655043125,-0.1914004385,0.2620660663,-0.2859320343,-0.0771497563,0.2549057007,-0.2463358939,-0.2266603261,-0.0365771614,-0.44798401,0.4240052402,-0.0907414258,-0.1566342264,0.2508070469,-0.008536445,0.20689331,0.4868449569,-0.2094397843,0.1476731151,-0.041045513,-0.3048643768,0.1159800738,0.2707941532,0.1655560881,-0.0684229136,-0.244354248,0.0148500968,-0.047958605,0.0014085558,0.1433942169,0.0184628051,0.1139994338,0.3630373776,-0.4942835867,0.2492495477,0.033946272,-0.003221052,0.0266965386,0.1335860193,-0.3562712073,0.2182810009,0.157754153,-0.1472541243,0.0038580976,0.1055263057,-0.000769733,0.0244378783,0.2575092912,0.1771205515,-0.1759036779,-0.1795732081,0.1544300765,-0.4579264224,-0.2024080753,-0.2777024209,0.3311907649,0.0727928504,-0.0513931774,0.2486350536,0.3979563415,-0.2974713445,-0.1497603953,-0.6814543605,-0.3206688464,0.2324930131,0.2381200641,-0.0383701734,0.1146195531,0.0728929564,0.1902510375,0.3014506698,-0.1593848318,0.0519206077,-0.2154660523,0.0668959543,0.2884461582,0.015409017,-0.1840073019,0.1464990377,-0.0906351209,-0.3029601276,-0.007445815,-0.0280529615,0.020559771,0.2018748224,0.0800618082,-0.0769100338,0.2922613919,-0.2874342501,0.2070834488,-0.0540392846,-0.2033355683,0.0260139201,-0.028652247,0.4373197854,0.0781343356,-0.0596907772,-0.3463966548,0.3778165281,0.1480235159,0.4820056558,0.121567972,0.1550164223,0.1109765992,-0.1504181921,0.0744265839,0.0033688864,-0.018836556,0.1148206964,0.1950344443,-0.1725692898,0.5341673493,0.062966302,-0.2056033462,0.2087480426,-0.0423731767,0.2821823359,0.4809158146,0.0285460111,-0.6245885491,0.1084474102,-0.1914929748,-0.0311708059,-0.1689195633,0.2262417972,0.074824743,-0.1845858544,0.2299911678,0.0367297381,0.0251142513,-0.1321828812,-0.026936762,0.1548585594,-0.1086892486,0.0423979945,-0.1922752261,0.0764321089,0.1230324656,0.0625291839,-0.2523228526,0.0761692077,0.0379106514,0.0927578285,-0.0046008215,-0.4079039097,-0.2479270846,0.488722086,0.1277362555,0.0824421793,-0.1357576698,0.6117275953,-0.0793123096,0.0828473195,-0.2280544192,0.3201188743,0.0085704597,-0.1079234108,-0.0688482448,-0.246837005,-0.2222373933,0.0081493556,0.2671044171,-0.0972433537,-0.4033998251,0.2094486207,0.0150436927,-0.2704956234,-0.0935829133,0.1093493253,-0.5271977782,-0.2371833324,0.3296177685,0.0876162276,0.0557524636,0.2086587101,0.0454592109,0.2435111254,0.3303770125,-0.3565886617,0.2710071802,-0.0956967324,-0.0946971476,-0.1964198053,0.0524836779,0.0446269698,0.1757940054,0.2156232744,-0.0008953036,-0.0572181195,0.0688706115,0.0837493688,-0.3943404853,-0.2059688419,0.2409687787,0.2719535232,0.1971836984,-0.3996480405,0.2140151858,-0.49484393,0.2268268019,0.3449775577,0.1031446457,0.1349351108,-0.0634646788,-0.0493180677,0.071547702,0.5398412943,0.196633637,0.3716632724,-0.0457697287,-0.536670208,-1.0705215931,0.0391415842,-0.0838204175,-0.0789416656,0.0751691386,-0.185544461,0.3193065822,0.1499446332,0.1614266485,0.6218615174,0.1247115284,-0.1874615997,-0.0437187999,-0.27184093,0.1919305772,0.2601183653,-0.226719588,-0.2280879766,-0.1288041174,-0.271841228,-0.1472595483,-0.1110556349,0.1084749326,-0.0248638708,0.5920586586,0.2828031182,-0.0653474852,0.5304176211,-0.0407716781,0.3095166385,0.1364389956,-0.085079439,0.0983227566,0.3748431206,-0.0030461906,0.0022479475,-0.3218943477,0.3769010007,-0.2531032264,0.044324711,-0.0314539447,-0.3877043426,-0.1718785912,-0.0478404649,-0.3519325256,0.0394589752,0.2682402134,0.3427316248,-0.099458456,-0.0833334699,0.001738596,-0.0707609653,0.1979726255,-0.1345929801,-0.4991217554,-0.3667905033,-0.0142705347,-0.4264901876,0.2073093355,-0.0814641416,0.0856033638,0.2151794136,0.2329836935,-0.0420237482,0.0630842671,-0.0680122375,-0.1133282706,-0.179614976,0.2484604269,0.1494685858,-0.0406302586,-0.3853634,-0.1905454546]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/263","title":"[Feature request] Support for external modality for language datasets","comments":"Thanks a lot, @aleSuglia for the very detailed and introductive feature request.\r\nIt seems like we could build something pretty useful here indeed.\r\n\r\nOne of the questions here is that Arrow doesn't have built-in support for generic \"tensors\" in records but there might be ways to do that in a clean way. We'll probably try to tackle this during the summer.","body":"# Background\r\n\r\nIn recent years many researchers have advocated that learning meanings from text-based only datasets is just like asking a human to \"learn to speak by listening to the radio\" [[E. Bender and A. Koller,2020](https:\/\/openreview.net\/forum?id=GKTvAcb12b), [Y. Bisk et. al, 2020](https:\/\/arxiv.org\/abs\/2004.10151)]. Therefore, the importance of multi-modal datasets for the NLP community is of paramount importance for next-generation models. For this reason, I raised a [concern](https:\/\/github.com\/huggingface\/nlp\/pull\/236#issuecomment-639832029) related to the best way to integrate external features in NLP datasets (e.g., visual features associated with an image, audio features associated with a recording, etc.). This would be of great importance for a more systematic way of representing data for ML models that are learning from multi-modal data. \r\n\r\n# Language + Vision\r\n\r\n## Use case\r\nTypically, people working on Language+Vision tasks, have a reference dataset (either in JSON or JSONL format) and for each example, they have an identifier that specifies the reference image. For a practical example, you can refer to the [GQA](https:\/\/cs.stanford.edu\/people\/dorarad\/gqa\/download.html#seconddown) dataset.\r\n\r\nCurrently, images are represented by either pooling-based features (average pooling of ResNet or VGGNet features, see [DeVries et.al, 2017](https:\/\/arxiv.org\/abs\/1611.08481), [Shekhar et.al, 2019](https:\/\/www.aclweb.org\/anthology\/N19-1265.pdf)) where you have a single vector for every image. Another option is to use a set of feature maps for every image extracted from a specific layer of a CNN (see [Xu et.al, 2015](https:\/\/arxiv.org\/abs\/1502.03044)). A more recent option, especially with large-scale multi-modal transformers [Li et. al, 2019](https:\/\/arxiv.org\/abs\/1908.03557), is to use FastRCNN features. \r\n\r\nFor all these types of features, people use one of the following formats:\r\n1. [HD5F](https:\/\/pypi.org\/project\/h5py\/)\r\n2. [NumPy](https:\/\/numpy.org\/doc\/stable\/reference\/generated\/numpy.savez.html)\r\n3. [LMDB](https:\/\/lmdb.readthedocs.io\/en\/release\/)\r\n\r\n## Implementation considerations\r\n\r\nI was thinking about possible ways of implementing this feature. As mentioned above, depending on the model, different visual features can be used. This step usually relies on another model (say ResNet-101) that is used to generate the visual features for each image used in the dataset. Typically, this step is done in a separate script that completes the feature generation procedure. The usual processing steps for these datasets are the following:\r\n\r\n1. Download dataset\r\n2. Download images associated with the dataset\r\n3. Write a script that generates the visual features for every image and store them in a specific file\r\n4. Create a DataLoader that maps the visual features to the corresponding language example\r\n\r\nIn my personal projects, I've decided to ignore HD5F because it doesn't have out-of-the-box support for multi-processing (see this PyTorch [issue](https:\/\/github.com\/pytorch\/pytorch\/issues\/11929)). I've been successfully using a NumPy compressed file for each image so that I can store any sort of information in it.\r\n\r\nFor ease of use of all these Language+Vision datasets, it would be really handy to have a way to associate the visual features with the text and store them in an efficient way. That's why I immediately thought about the HuggingFace NLP backend based on Apache Arrow. The assumption here is that the external modality will be mapped to a N-dimensional tensor so easily represented by a NumPy array. \r\n\r\nLooking forward to hearing your thoughts about it!","comment_length":61,"text":"[Feature request] Support for external modality for language datasets\n# Background\r\n\r\nIn recent years many researchers have advocated that learning meanings from text-based only datasets is just like asking a human to \"learn to speak by listening to the radio\" [[E. Bender and A. Koller,2020](https:\/\/openreview.net\/forum?id=GKTvAcb12b), [Y. Bisk et. al, 2020](https:\/\/arxiv.org\/abs\/2004.10151)]. Therefore, the importance of multi-modal datasets for the NLP community is of paramount importance for next-generation models. For this reason, I raised a [concern](https:\/\/github.com\/huggingface\/nlp\/pull\/236#issuecomment-639832029) related to the best way to integrate external features in NLP datasets (e.g., visual features associated with an image, audio features associated with a recording, etc.). This would be of great importance for a more systematic way of representing data for ML models that are learning from multi-modal data. \r\n\r\n# Language + Vision\r\n\r\n## Use case\r\nTypically, people working on Language+Vision tasks, have a reference dataset (either in JSON or JSONL format) and for each example, they have an identifier that specifies the reference image. For a practical example, you can refer to the [GQA](https:\/\/cs.stanford.edu\/people\/dorarad\/gqa\/download.html#seconddown) dataset.\r\n\r\nCurrently, images are represented by either pooling-based features (average pooling of ResNet or VGGNet features, see [DeVries et.al, 2017](https:\/\/arxiv.org\/abs\/1611.08481), [Shekhar et.al, 2019](https:\/\/www.aclweb.org\/anthology\/N19-1265.pdf)) where you have a single vector for every image. Another option is to use a set of feature maps for every image extracted from a specific layer of a CNN (see [Xu et.al, 2015](https:\/\/arxiv.org\/abs\/1502.03044)). A more recent option, especially with large-scale multi-modal transformers [Li et. al, 2019](https:\/\/arxiv.org\/abs\/1908.03557), is to use FastRCNN features. \r\n\r\nFor all these types of features, people use one of the following formats:\r\n1. [HD5F](https:\/\/pypi.org\/project\/h5py\/)\r\n2. [NumPy](https:\/\/numpy.org\/doc\/stable\/reference\/generated\/numpy.savez.html)\r\n3. [LMDB](https:\/\/lmdb.readthedocs.io\/en\/release\/)\r\n\r\n## Implementation considerations\r\n\r\nI was thinking about possible ways of implementing this feature. As mentioned above, depending on the model, different visual features can be used. This step usually relies on another model (say ResNet-101) that is used to generate the visual features for each image used in the dataset. Typically, this step is done in a separate script that completes the feature generation procedure. The usual processing steps for these datasets are the following:\r\n\r\n1. Download dataset\r\n2. Download images associated with the dataset\r\n3. Write a script that generates the visual features for every image and store them in a specific file\r\n4. Create a DataLoader that maps the visual features to the corresponding language example\r\n\r\nIn my personal projects, I've decided to ignore HD5F because it doesn't have out-of-the-box support for multi-processing (see this PyTorch [issue](https:\/\/github.com\/pytorch\/pytorch\/issues\/11929)). I've been successfully using a NumPy compressed file for each image so that I can store any sort of information in it.\r\n\r\nFor ease of use of all these Language+Vision datasets, it would be really handy to have a way to associate the visual features with the text and store them in an efficient way. That's why I immediately thought about the HuggingFace NLP backend based on Apache Arrow. The assumption here is that the external modality will be mapped to a N-dimensional tensor so easily represented by a NumPy array. \r\n\r\nLooking forward to hearing your thoughts about it!\nThanks a lot, @aleSuglia for the very detailed and introductive feature request.\r\nIt seems like we could build something pretty useful here indeed.\r\n\r\nOne of the questions here is that Arrow doesn't have built-in support for generic \"tensors\" in records but there might be ways to do that in a clean way. We'll probably try to tackle this during the summer.","embeddings":[0.0000579982,-0.1425000429,-0.1048433781,-0.1137593612,-0.0208638422,-0.2757304013,0.3570465446,0.1619891822,-0.1701952517,-0.115410395,-0.1062394306,0.3741975129,-0.2327695489,0.0247858353,0.2276581675,-0.1631375402,0.1156109571,0.0206510182,0.1751564294,0.0377422199,-0.0802958757,-0.1862993091,-0.0996891856,0.0158086233,-0.1543438584,-0.0065530962,-0.2325278074,-0.0663946643,0.0174453668,-0.1567316949,-0.4168612659,0.1775110662,-0.0064961114,-0.0587129258,-0.000105389,0.0410651825,0.0508660674,-0.1544797868,-0.0636066273,-0.2922307551,-0.1718059927,-0.5856842995,-0.0051841163,-0.2138669938,0.1689768434,-0.2611020803,0.303366214,-0.0154541899,0.219957605,0.042448286,0.2338732183,0.1619654,0.0570976958,0.3144838214,0.1790788174,0.7218409181,-0.1756500155,-0.0262282193,0.3759553134,-0.1980069876,-0.1714316756,0.5319548249,0.0707272217,-0.1726881266,0.1047964245,-0.1554219872,-0.0742541477,-0.1862260252,-0.182682246,0.26961869,0.5540597439,-0.0946797803,-0.3956744969,-0.2771704197,0.101535894,0.2120435387,-0.1491226256,0.4695759714,-0.084704265,0.2618876398,-0.1411739141,-0.4172467589,-0.2269273251,0.4005026221,-0.0107411835,-0.1144736856,0.1139273718,0.1064259782,0.1271232069,-0.1394746155,0.2405346036,-0.0852413252,0.1974173635,0.1384702623,0.1199183315,-0.147744298,0.092940703,-0.0756618828,0.1010376588,-0.1379873306,0.0141931195,0.1652175933,-0.6861980557,0.3720004857,0.1384753585,-0.3171008229,-0.3466371894,0.1551730186,0.298522681,-0.0579590537,-0.0268653706,-0.0378671661,0.0169135183,0.0440881662,-0.3268453181,0.0000947688,0.1069344431,-0.0153924404,-0.0973811075,-0.292539537,-0.2431646138,0.0581131242,-0.0261005349,-0.0754403844,0.2316321433,0.295160681,-0.0103263939,0.2380757183,0.0051021767,-0.6885832548,0.0688787028,0.0822422132,-0.4507437944,0.112957038,0.0196862388,0.5769885778,0.1056848541,0.0224712547,0.0795072168,-0.0589267202,0.1154198125,-0.3274696171,0.0944871306,0.0146099059,-0.3158082962,-0.282207936,-0.0010033672,0.2286927402,-0.193028748,-0.0952018723,0.0137523692,-0.2000210136,-0.1908557564,0.121775277,-0.0717911124,0.0397867337,0.3123777509,0.9068838954,-0.3320865631,-0.4840673804,-0.0372311957,-0.0365264602,-0.324636817,-0.1530091465,-0.1083315238,0.2535992861,-0.4206463397,-0.1457893252,0.0214555152,-0.0782916471,-0.0730273351,0.2533953488,-0.3021765351,-0.0619365349,0.0733146667,-0.1439300179,0.422486335,-0.2967081666,-0.2142054886,0.1942324638,0.0423240773,0.2526317537,-0.0438900702,0.3999971449,-0.0188864972,-0.1381499916,0.1212410256,0.7113153338,-0.150968194,0.0519453287,-0.1938297302,-0.3600857258,0.2658928335,0.4229554236,0.0130498447,-0.3205803931,0.6397942305,-0.000615358,0.1015567109,-0.2774530053,0.1214610413,-0.193411395,0.2937177122,-0.0562877953,-0.1427305341,-0.2755998075,-0.2272177041,0.0019212904,-0.2704661489,0.2816795111,-0.1663463861,-0.0009555613,0.208978802,-0.1393232346,0.0956419855,0.0041683582,0.2037206292,0.1011403576,-0.4169175923,0.2500483096,-0.0902082399,0.0251920149,-0.0975428373,0.038838096,0.1815153807,0.2594423592,0.0138022294,0.1161562577,0.0984601974,0.5246058702,-0.2238256782,0.0367368646,0.0321959108,-0.0006297429,0.1992013603,0.480219692,-0.0132267922,0.3972020745,0.480912894,-0.6762456298,0.3488892317,-0.1610369831,0.2448391616,-0.1503302455,-0.042594105,0.2094171345,0.248840332,0.0853460506,-0.0793127269,0.0688090697,0.0853889957,0.0471753962,-0.3372054994,-0.1611867249,-0.0550510958,0.194349274,0.0049326029,0.03765646,-0.3816787601,0.1459192485,0.534516871,0.2316180319,0.179944545,0.1668641865,-0.2088233531,-0.3059071302,0.2450826764,-0.2178815007,0.0730601624,0.2838123739,-0.2241439521,0.0624792688,-0.212926656,0.0168988183,0.1520752311,0.06433817,0.4003802836,0.0581446253,0.379850626,0.0560440756,-0.0656488761,-0.4201531112,-0.0221136883,-0.2725446224,0.1746994704,-0.0094886972,-0.189365685,-0.6152865887,-0.1843091547,-0.0398294479,-0.445258677,-0.0595044978,0.0666307211,-0.2852560878,-0.1706150472,0.0751235336,0.0053999494,0.378551811,-0.3618397117,-0.0995696262,0.1389790475,-0.260455668,-0.1546472907,0.0843451992,0.4543462694,0.0610960312,0.2402535081,0.2454445511,-0.0476959907,-0.133087188,-0.4327585399,0.2702728808,0.0578765273,0.1886128038,-0.002010792,-0.5663678646,0.0909249336,-0.0657481998,0.2020861208,0.0889275745,-0.0042456607,-0.0664625615,-0.0416938066,0.1243292913,-0.2534283996,0.0937701836,-0.4923431575,-0.5119807124,0.3841014206,-0.1436410397,0.1656590104,0.5681470633,0.0995381847,-0.0953869373,-0.0172695182,0.2712105215,0.0300444029,0.3557389677,-0.0596465357,-0.000171215,-0.1874789894,-0.2881140709,-0.2174292356,0.1192458197,0.3493945003,-0.3474397063,-0.3242609501,0.0953799263,0.156168282,0.0391207188,0.0027884229,0.0905975923,0.0016481931,-0.0056026275,-0.0388501734,0.0010586765,-0.0685858503,0.3693791628,0.1696063131,0.3322416246,0.010456631,0.4449982047,0.3984428048,-0.0735823065,-0.1130097508,0.4104923606,0.341801554,0.2031654716,-0.1458218843,-0.2318649739,0.461813271,-0.099810183,-0.0304083619,0.2845420837,-0.0223187096,-0.5936282277,-0.230983153,-0.0808477029,-0.4445461929,-0.193429783,0.3490699232,-0.1369313598,0.4724569321,0.1747802347,-0.5142968297,-0.2779099643,-0.0862949044,0.0622660592,0.0467646345,0.230810672,0.0712513477,-0.3800601959,-0.0887860432,0.094128795,0.0836932957,0.2238935381,0.2547115982,-0.1767908186,-0.127438426,0.0343180485,0.1126118004,0.1010146737,-0.3075811863,-0.4294191003,0.1311664283,0.0697894841,-0.2317713052,-0.0498642474,-0.1488519907,-0.0059812935,0.1850325316,0.2465440184,-0.0023722476,-0.1417336166,-0.006456967,0.3172285259,0.0406909883,-0.2438861132,0.1223148108,0.1655285358,-0.2372305393,-0.2288937867,-0.13118276,0.0266881436,0.243184939,0.1105710045,0.0440215431,-0.0077721127,0.1166662052,0.1760064214,-0.1860144734,0.0162820574,-0.1535492837,-0.0094976211,0.2016980499,-0.1232051104,0.2899032235,0.1724204421,-0.2464565188,-0.0342411175,-0.2041366845,0.5062916875,0.3987755179,0.555439353,0.1340870708,-0.2353624701,-0.1000528038,-0.4058998823,0.0784540176,0.1703917235,-0.0879821107,0.0928243622,-0.2557076216,0.177929759,-0.1484358013,-0.2434948832,-0.0016531135,0.1146700084,-0.3192114532,0.4626779556,0.2678561509,0.9083020091,0.1739370823,0.3773852289,0.0632027984,-0.139299795,0.4011537135,-0.5117281079,0.1901806742,-0.2407249808,0.0361801311,0.0237822328,0.0500327013,0.0928927436,0.1420457512,-0.0994795635,-0.0832984149,0.6208557487,0.1133584604,0.1424345523,0.2351865172,-0.0402040407,-0.5092383027,-0.0592830367,0.1355792284,-0.1756825447,-0.2223032415,-0.1440446973,-0.3420350254,0.081941247,0.2642249763,-0.2147074491,-0.0011316941,-0.316054076,0.1708391458,0.0972714871,0.0274491161,0.4782885015,0.0652947873,0.103047885,-0.1676760763,0.0214994159,-0.112146154,0.1028367281,0.1289995611,-0.0645298436,-0.2993836999,0.5649161935,-0.0850202814,-0.2367733568,0.0830471367,0.0265793484,-0.0678922683,-0.3287886977,-0.1408794671,0.3525542319,0.0692692399,-0.204262495,0.3383693695,0.0336397365,0.0056782272,0.1585759819,0.1495663524,-0.0072706854,0.254992038,0.0907109082,0.0957420543,-0.0411472321,-0.1602442861,-0.1721546501,-0.0682337284,0.0020590571,-0.0395036861,0.0266543794,-0.2097666264,0.4513889253,-0.165593937,-0.0678607747,-0.2193889022,0.1647765487,-0.1918230802,-0.1911372095,0.0416479334,0.1102282107,0.0128573095,-0.056775935,-0.3587937653,-0.0227739848,0.4008126855,-0.4294159412,-0.0583019592,-0.0565981828,0.150400877,-0.0135832634,0.4569157362,-0.3693554699,-0.0621005334,0.2285134196,0.2409587353,-0.072363928,-0.3310264945,-0.1456676275,0.0843308046,0.1410640925,0.1694407761,-0.2751306593,-0.16777955,-0.075637877,0.1569445878,0.2780585885,0.0947420076,-0.0844768286,-0.3365341425,-0.1436418742,-0.1429385245,-0.3242317736,-0.1453953534,-0.2079687715,0.422285229,-0.6543209553,0.2695195079,0.2933262885,0.0965970084,-0.1245065331,0.3523571193,0.2550243139,0.0950312018,0.0871967599,0.1092824489,-0.2875727713,0.2659047842,-0.1545007229,0.001340931,0.3109216392,0.4030639231,0.1597541869,-0.0666116849,0.1476008296,-0.1723828763,-0.240617767,0.0150780082,0.3615274727,0.2620097101,-0.203440845,0.0838944167,0.566072166,-0.1636136323,0.1882482171,0.1468002498,0.8290458918,0.2113883197,0.1585373282,-0.3222102523,0.031565316,-0.1468386799,-0.0865758136,-0.0459187441,0.181945309,0.0793437287,0.0689206421,0.2830908597,0.4801355898,0.3831429482,-0.3467300534,0.1410560906,-0.0798541233,0.1351424158,0.5679847002,-0.0949716717,0.2906667292,0.1286545843,0.1338963211,0.1971307099,-0.3649289906,0.1564904451,-0.0954354033,-0.0901187509,-0.0247576255,0.2911919057,0.1542281657,-0.0017625219,-0.3724502921,-0.039304208,-0.4641978741,0.0071829511,-0.0693771392,0.0064607002,0.037394993,0.228744477,0.1294110417,-0.1567295045,0.1997879148,0.2505199313,0.4216548502,-0.2381192893,0.0469047427,0.0342185535,0.0821622536,0.1136531308,0.1388760656,-0.025581643,0.0380305126,-0.3868378103,-0.0602942519,-0.3789958954,-0.0223439839,-0.1458857358,0.0431689508,0.1438975781,0.3281206787,0.3592202663,0.078514576,-0.141339913,0.045587033,0.0547987781,-0.248927176,-0.3126660883,-0.0954528376,0.1773027927,0.030504439,-0.0231337044,0.0370859206,-0.2534484565,-0.4080215991,0.4324475229,-0.1037863418,-0.0051232195,-0.0782962739,0.0743996873,-0.1761177778,-0.0367306247,0.4093006849,0.1028505489,0.1296515912,-0.195560798,-0.2923826277,-0.0247532614,0.1040319726,-0.4462167621,-0.0645439923,0.2829323709,0.3547560871,0.3475917876,-0.2640645206,0.1405351162,0.2678466439,-0.2227523178,-0.1304278076,0.177875489,-0.1076065302,0.2173901796,-0.1289331913,0.0439179651,-0.081497103,-0.1311272681,0.08348497,-0.430336237,-0.1338418573,-0.0598375127,0.2861565948,0.3971822858,0.1749394983,0.0044677169,0.0369785987,-0.3076259792,-0.1168514043,-0.0747248083,-0.0788575038,-0.0553862154,-0.0985004976,0.7773543596,-0.286259979,0.1690689176,-0.02445562,0.2367896438,-0.0450646318,-0.2427914441,-0.0508058034,0.0319436751,-0.1385505646,-0.0690859258,0.1078336611,0.4174990952,0.0375690758,-0.1575308144,-0.0531437695,-0.1485236287,-0.0462270901,-0.2435237616,-0.2073337138,0.2868824601,0.0515742823,-0.1594630033,0.3107645214,-0.437997669,-0.2390692085,0.337269336,0.0692639798,0.0916248187,0.3000227511,0.2653143406,-0.370274961,-0.1716799885,-0.2494577914,-0.2490811497,-0.1187176555,-0.125312224,-0.1293733567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/263","title":"[Feature request] Support for external modality for language datasets","comments":"I was looking into Facebook MMF and apparently they decided to use LMDB to store additional features associated with every example: https:\/\/github.com\/facebookresearch\/mmf\/blob\/master\/mmf\/datasets\/databases\/features_database.py\r\n\r\n","body":"# Background\r\n\r\nIn recent years many researchers have advocated that learning meanings from text-based only datasets is just like asking a human to \"learn to speak by listening to the radio\" [[E. Bender and A. Koller,2020](https:\/\/openreview.net\/forum?id=GKTvAcb12b), [Y. Bisk et. al, 2020](https:\/\/arxiv.org\/abs\/2004.10151)]. Therefore, the importance of multi-modal datasets for the NLP community is of paramount importance for next-generation models. For this reason, I raised a [concern](https:\/\/github.com\/huggingface\/nlp\/pull\/236#issuecomment-639832029) related to the best way to integrate external features in NLP datasets (e.g., visual features associated with an image, audio features associated with a recording, etc.). This would be of great importance for a more systematic way of representing data for ML models that are learning from multi-modal data. \r\n\r\n# Language + Vision\r\n\r\n## Use case\r\nTypically, people working on Language+Vision tasks, have a reference dataset (either in JSON or JSONL format) and for each example, they have an identifier that specifies the reference image. For a practical example, you can refer to the [GQA](https:\/\/cs.stanford.edu\/people\/dorarad\/gqa\/download.html#seconddown) dataset.\r\n\r\nCurrently, images are represented by either pooling-based features (average pooling of ResNet or VGGNet features, see [DeVries et.al, 2017](https:\/\/arxiv.org\/abs\/1611.08481), [Shekhar et.al, 2019](https:\/\/www.aclweb.org\/anthology\/N19-1265.pdf)) where you have a single vector for every image. Another option is to use a set of feature maps for every image extracted from a specific layer of a CNN (see [Xu et.al, 2015](https:\/\/arxiv.org\/abs\/1502.03044)). A more recent option, especially with large-scale multi-modal transformers [Li et. al, 2019](https:\/\/arxiv.org\/abs\/1908.03557), is to use FastRCNN features. \r\n\r\nFor all these types of features, people use one of the following formats:\r\n1. [HD5F](https:\/\/pypi.org\/project\/h5py\/)\r\n2. [NumPy](https:\/\/numpy.org\/doc\/stable\/reference\/generated\/numpy.savez.html)\r\n3. [LMDB](https:\/\/lmdb.readthedocs.io\/en\/release\/)\r\n\r\n## Implementation considerations\r\n\r\nI was thinking about possible ways of implementing this feature. As mentioned above, depending on the model, different visual features can be used. This step usually relies on another model (say ResNet-101) that is used to generate the visual features for each image used in the dataset. Typically, this step is done in a separate script that completes the feature generation procedure. The usual processing steps for these datasets are the following:\r\n\r\n1. Download dataset\r\n2. Download images associated with the dataset\r\n3. Write a script that generates the visual features for every image and store them in a specific file\r\n4. Create a DataLoader that maps the visual features to the corresponding language example\r\n\r\nIn my personal projects, I've decided to ignore HD5F because it doesn't have out-of-the-box support for multi-processing (see this PyTorch [issue](https:\/\/github.com\/pytorch\/pytorch\/issues\/11929)). I've been successfully using a NumPy compressed file for each image so that I can store any sort of information in it.\r\n\r\nFor ease of use of all these Language+Vision datasets, it would be really handy to have a way to associate the visual features with the text and store them in an efficient way. That's why I immediately thought about the HuggingFace NLP backend based on Apache Arrow. The assumption here is that the external modality will be mapped to a N-dimensional tensor so easily represented by a NumPy array. \r\n\r\nLooking forward to hearing your thoughts about it!","comment_length":22,"text":"[Feature request] Support for external modality for language datasets\n# Background\r\n\r\nIn recent years many researchers have advocated that learning meanings from text-based only datasets is just like asking a human to \"learn to speak by listening to the radio\" [[E. Bender and A. Koller,2020](https:\/\/openreview.net\/forum?id=GKTvAcb12b), [Y. Bisk et. al, 2020](https:\/\/arxiv.org\/abs\/2004.10151)]. Therefore, the importance of multi-modal datasets for the NLP community is of paramount importance for next-generation models. For this reason, I raised a [concern](https:\/\/github.com\/huggingface\/nlp\/pull\/236#issuecomment-639832029) related to the best way to integrate external features in NLP datasets (e.g., visual features associated with an image, audio features associated with a recording, etc.). This would be of great importance for a more systematic way of representing data for ML models that are learning from multi-modal data. \r\n\r\n# Language + Vision\r\n\r\n## Use case\r\nTypically, people working on Language+Vision tasks, have a reference dataset (either in JSON or JSONL format) and for each example, they have an identifier that specifies the reference image. For a practical example, you can refer to the [GQA](https:\/\/cs.stanford.edu\/people\/dorarad\/gqa\/download.html#seconddown) dataset.\r\n\r\nCurrently, images are represented by either pooling-based features (average pooling of ResNet or VGGNet features, see [DeVries et.al, 2017](https:\/\/arxiv.org\/abs\/1611.08481), [Shekhar et.al, 2019](https:\/\/www.aclweb.org\/anthology\/N19-1265.pdf)) where you have a single vector for every image. Another option is to use a set of feature maps for every image extracted from a specific layer of a CNN (see [Xu et.al, 2015](https:\/\/arxiv.org\/abs\/1502.03044)). A more recent option, especially with large-scale multi-modal transformers [Li et. al, 2019](https:\/\/arxiv.org\/abs\/1908.03557), is to use FastRCNN features. \r\n\r\nFor all these types of features, people use one of the following formats:\r\n1. [HD5F](https:\/\/pypi.org\/project\/h5py\/)\r\n2. [NumPy](https:\/\/numpy.org\/doc\/stable\/reference\/generated\/numpy.savez.html)\r\n3. [LMDB](https:\/\/lmdb.readthedocs.io\/en\/release\/)\r\n\r\n## Implementation considerations\r\n\r\nI was thinking about possible ways of implementing this feature. As mentioned above, depending on the model, different visual features can be used. This step usually relies on another model (say ResNet-101) that is used to generate the visual features for each image used in the dataset. Typically, this step is done in a separate script that completes the feature generation procedure. The usual processing steps for these datasets are the following:\r\n\r\n1. Download dataset\r\n2. Download images associated with the dataset\r\n3. Write a script that generates the visual features for every image and store them in a specific file\r\n4. Create a DataLoader that maps the visual features to the corresponding language example\r\n\r\nIn my personal projects, I've decided to ignore HD5F because it doesn't have out-of-the-box support for multi-processing (see this PyTorch [issue](https:\/\/github.com\/pytorch\/pytorch\/issues\/11929)). I've been successfully using a NumPy compressed file for each image so that I can store any sort of information in it.\r\n\r\nFor ease of use of all these Language+Vision datasets, it would be really handy to have a way to associate the visual features with the text and store them in an efficient way. That's why I immediately thought about the HuggingFace NLP backend based on Apache Arrow. The assumption here is that the external modality will be mapped to a N-dimensional tensor so easily represented by a NumPy array. \r\n\r\nLooking forward to hearing your thoughts about it!\nI was looking into Facebook MMF and apparently they decided to use LMDB to store additional features associated with every example: https:\/\/github.com\/facebookresearch\/mmf\/blob\/master\/mmf\/datasets\/databases\/features_database.py\r\n\r\n","embeddings":[0.0000579982,-0.1425000429,-0.1048433781,-0.1137593612,-0.0208638422,-0.2757304013,0.3570465446,0.1619891822,-0.1701952517,-0.115410395,-0.1062394306,0.3741975129,-0.2327695489,0.0247858353,0.2276581675,-0.1631375402,0.1156109571,0.0206510182,0.1751564294,0.0377422199,-0.0802958757,-0.1862993091,-0.0996891856,0.0158086233,-0.1543438584,-0.0065530962,-0.2325278074,-0.0663946643,0.0174453668,-0.1567316949,-0.4168612659,0.1775110662,-0.0064961114,-0.0587129258,-0.000105389,0.0410651825,0.0508660674,-0.1544797868,-0.0636066273,-0.2922307551,-0.1718059927,-0.5856842995,-0.0051841163,-0.2138669938,0.1689768434,-0.2611020803,0.303366214,-0.0154541899,0.219957605,0.042448286,0.2338732183,0.1619654,0.0570976958,0.3144838214,0.1790788174,0.7218409181,-0.1756500155,-0.0262282193,0.3759553134,-0.1980069876,-0.1714316756,0.5319548249,0.0707272217,-0.1726881266,0.1047964245,-0.1554219872,-0.0742541477,-0.1862260252,-0.182682246,0.26961869,0.5540597439,-0.0946797803,-0.3956744969,-0.2771704197,0.101535894,0.2120435387,-0.1491226256,0.4695759714,-0.084704265,0.2618876398,-0.1411739141,-0.4172467589,-0.2269273251,0.4005026221,-0.0107411835,-0.1144736856,0.1139273718,0.1064259782,0.1271232069,-0.1394746155,0.2405346036,-0.0852413252,0.1974173635,0.1384702623,0.1199183315,-0.147744298,0.092940703,-0.0756618828,0.1010376588,-0.1379873306,0.0141931195,0.1652175933,-0.6861980557,0.3720004857,0.1384753585,-0.3171008229,-0.3466371894,0.1551730186,0.298522681,-0.0579590537,-0.0268653706,-0.0378671661,0.0169135183,0.0440881662,-0.3268453181,0.0000947688,0.1069344431,-0.0153924404,-0.0973811075,-0.292539537,-0.2431646138,0.0581131242,-0.0261005349,-0.0754403844,0.2316321433,0.295160681,-0.0103263939,0.2380757183,0.0051021767,-0.6885832548,0.0688787028,0.0822422132,-0.4507437944,0.112957038,0.0196862388,0.5769885778,0.1056848541,0.0224712547,0.0795072168,-0.0589267202,0.1154198125,-0.3274696171,0.0944871306,0.0146099059,-0.3158082962,-0.282207936,-0.0010033672,0.2286927402,-0.193028748,-0.0952018723,0.0137523692,-0.2000210136,-0.1908557564,0.121775277,-0.0717911124,0.0397867337,0.3123777509,0.9068838954,-0.3320865631,-0.4840673804,-0.0372311957,-0.0365264602,-0.324636817,-0.1530091465,-0.1083315238,0.2535992861,-0.4206463397,-0.1457893252,0.0214555152,-0.0782916471,-0.0730273351,0.2533953488,-0.3021765351,-0.0619365349,0.0733146667,-0.1439300179,0.422486335,-0.2967081666,-0.2142054886,0.1942324638,0.0423240773,0.2526317537,-0.0438900702,0.3999971449,-0.0188864972,-0.1381499916,0.1212410256,0.7113153338,-0.150968194,0.0519453287,-0.1938297302,-0.3600857258,0.2658928335,0.4229554236,0.0130498447,-0.3205803931,0.6397942305,-0.000615358,0.1015567109,-0.2774530053,0.1214610413,-0.193411395,0.2937177122,-0.0562877953,-0.1427305341,-0.2755998075,-0.2272177041,0.0019212904,-0.2704661489,0.2816795111,-0.1663463861,-0.0009555613,0.208978802,-0.1393232346,0.0956419855,0.0041683582,0.2037206292,0.1011403576,-0.4169175923,0.2500483096,-0.0902082399,0.0251920149,-0.0975428373,0.038838096,0.1815153807,0.2594423592,0.0138022294,0.1161562577,0.0984601974,0.5246058702,-0.2238256782,0.0367368646,0.0321959108,-0.0006297429,0.1992013603,0.480219692,-0.0132267922,0.3972020745,0.480912894,-0.6762456298,0.3488892317,-0.1610369831,0.2448391616,-0.1503302455,-0.042594105,0.2094171345,0.248840332,0.0853460506,-0.0793127269,0.0688090697,0.0853889957,0.0471753962,-0.3372054994,-0.1611867249,-0.0550510958,0.194349274,0.0049326029,0.03765646,-0.3816787601,0.1459192485,0.534516871,0.2316180319,0.179944545,0.1668641865,-0.2088233531,-0.3059071302,0.2450826764,-0.2178815007,0.0730601624,0.2838123739,-0.2241439521,0.0624792688,-0.212926656,0.0168988183,0.1520752311,0.06433817,0.4003802836,0.0581446253,0.379850626,0.0560440756,-0.0656488761,-0.4201531112,-0.0221136883,-0.2725446224,0.1746994704,-0.0094886972,-0.189365685,-0.6152865887,-0.1843091547,-0.0398294479,-0.445258677,-0.0595044978,0.0666307211,-0.2852560878,-0.1706150472,0.0751235336,0.0053999494,0.378551811,-0.3618397117,-0.0995696262,0.1389790475,-0.260455668,-0.1546472907,0.0843451992,0.4543462694,0.0610960312,0.2402535081,0.2454445511,-0.0476959907,-0.133087188,-0.4327585399,0.2702728808,0.0578765273,0.1886128038,-0.002010792,-0.5663678646,0.0909249336,-0.0657481998,0.2020861208,0.0889275745,-0.0042456607,-0.0664625615,-0.0416938066,0.1243292913,-0.2534283996,0.0937701836,-0.4923431575,-0.5119807124,0.3841014206,-0.1436410397,0.1656590104,0.5681470633,0.0995381847,-0.0953869373,-0.0172695182,0.2712105215,0.0300444029,0.3557389677,-0.0596465357,-0.000171215,-0.1874789894,-0.2881140709,-0.2174292356,0.1192458197,0.3493945003,-0.3474397063,-0.3242609501,0.0953799263,0.156168282,0.0391207188,0.0027884229,0.0905975923,0.0016481931,-0.0056026275,-0.0388501734,0.0010586765,-0.0685858503,0.3693791628,0.1696063131,0.3322416246,0.010456631,0.4449982047,0.3984428048,-0.0735823065,-0.1130097508,0.4104923606,0.341801554,0.2031654716,-0.1458218843,-0.2318649739,0.461813271,-0.099810183,-0.0304083619,0.2845420837,-0.0223187096,-0.5936282277,-0.230983153,-0.0808477029,-0.4445461929,-0.193429783,0.3490699232,-0.1369313598,0.4724569321,0.1747802347,-0.5142968297,-0.2779099643,-0.0862949044,0.0622660592,0.0467646345,0.230810672,0.0712513477,-0.3800601959,-0.0887860432,0.094128795,0.0836932957,0.2238935381,0.2547115982,-0.1767908186,-0.127438426,0.0343180485,0.1126118004,0.1010146737,-0.3075811863,-0.4294191003,0.1311664283,0.0697894841,-0.2317713052,-0.0498642474,-0.1488519907,-0.0059812935,0.1850325316,0.2465440184,-0.0023722476,-0.1417336166,-0.006456967,0.3172285259,0.0406909883,-0.2438861132,0.1223148108,0.1655285358,-0.2372305393,-0.2288937867,-0.13118276,0.0266881436,0.243184939,0.1105710045,0.0440215431,-0.0077721127,0.1166662052,0.1760064214,-0.1860144734,0.0162820574,-0.1535492837,-0.0094976211,0.2016980499,-0.1232051104,0.2899032235,0.1724204421,-0.2464565188,-0.0342411175,-0.2041366845,0.5062916875,0.3987755179,0.555439353,0.1340870708,-0.2353624701,-0.1000528038,-0.4058998823,0.0784540176,0.1703917235,-0.0879821107,0.0928243622,-0.2557076216,0.177929759,-0.1484358013,-0.2434948832,-0.0016531135,0.1146700084,-0.3192114532,0.4626779556,0.2678561509,0.9083020091,0.1739370823,0.3773852289,0.0632027984,-0.139299795,0.4011537135,-0.5117281079,0.1901806742,-0.2407249808,0.0361801311,0.0237822328,0.0500327013,0.0928927436,0.1420457512,-0.0994795635,-0.0832984149,0.6208557487,0.1133584604,0.1424345523,0.2351865172,-0.0402040407,-0.5092383027,-0.0592830367,0.1355792284,-0.1756825447,-0.2223032415,-0.1440446973,-0.3420350254,0.081941247,0.2642249763,-0.2147074491,-0.0011316941,-0.316054076,0.1708391458,0.0972714871,0.0274491161,0.4782885015,0.0652947873,0.103047885,-0.1676760763,0.0214994159,-0.112146154,0.1028367281,0.1289995611,-0.0645298436,-0.2993836999,0.5649161935,-0.0850202814,-0.2367733568,0.0830471367,0.0265793484,-0.0678922683,-0.3287886977,-0.1408794671,0.3525542319,0.0692692399,-0.204262495,0.3383693695,0.0336397365,0.0056782272,0.1585759819,0.1495663524,-0.0072706854,0.254992038,0.0907109082,0.0957420543,-0.0411472321,-0.1602442861,-0.1721546501,-0.0682337284,0.0020590571,-0.0395036861,0.0266543794,-0.2097666264,0.4513889253,-0.165593937,-0.0678607747,-0.2193889022,0.1647765487,-0.1918230802,-0.1911372095,0.0416479334,0.1102282107,0.0128573095,-0.056775935,-0.3587937653,-0.0227739848,0.4008126855,-0.4294159412,-0.0583019592,-0.0565981828,0.150400877,-0.0135832634,0.4569157362,-0.3693554699,-0.0621005334,0.2285134196,0.2409587353,-0.072363928,-0.3310264945,-0.1456676275,0.0843308046,0.1410640925,0.1694407761,-0.2751306593,-0.16777955,-0.075637877,0.1569445878,0.2780585885,0.0947420076,-0.0844768286,-0.3365341425,-0.1436418742,-0.1429385245,-0.3242317736,-0.1453953534,-0.2079687715,0.422285229,-0.6543209553,0.2695195079,0.2933262885,0.0965970084,-0.1245065331,0.3523571193,0.2550243139,0.0950312018,0.0871967599,0.1092824489,-0.2875727713,0.2659047842,-0.1545007229,0.001340931,0.3109216392,0.4030639231,0.1597541869,-0.0666116849,0.1476008296,-0.1723828763,-0.240617767,0.0150780082,0.3615274727,0.2620097101,-0.203440845,0.0838944167,0.566072166,-0.1636136323,0.1882482171,0.1468002498,0.8290458918,0.2113883197,0.1585373282,-0.3222102523,0.031565316,-0.1468386799,-0.0865758136,-0.0459187441,0.181945309,0.0793437287,0.0689206421,0.2830908597,0.4801355898,0.3831429482,-0.3467300534,0.1410560906,-0.0798541233,0.1351424158,0.5679847002,-0.0949716717,0.2906667292,0.1286545843,0.1338963211,0.1971307099,-0.3649289906,0.1564904451,-0.0954354033,-0.0901187509,-0.0247576255,0.2911919057,0.1542281657,-0.0017625219,-0.3724502921,-0.039304208,-0.4641978741,0.0071829511,-0.0693771392,0.0064607002,0.037394993,0.228744477,0.1294110417,-0.1567295045,0.1997879148,0.2505199313,0.4216548502,-0.2381192893,0.0469047427,0.0342185535,0.0821622536,0.1136531308,0.1388760656,-0.025581643,0.0380305126,-0.3868378103,-0.0602942519,-0.3789958954,-0.0223439839,-0.1458857358,0.0431689508,0.1438975781,0.3281206787,0.3592202663,0.078514576,-0.141339913,0.045587033,0.0547987781,-0.248927176,-0.3126660883,-0.0954528376,0.1773027927,0.030504439,-0.0231337044,0.0370859206,-0.2534484565,-0.4080215991,0.4324475229,-0.1037863418,-0.0051232195,-0.0782962739,0.0743996873,-0.1761177778,-0.0367306247,0.4093006849,0.1028505489,0.1296515912,-0.195560798,-0.2923826277,-0.0247532614,0.1040319726,-0.4462167621,-0.0645439923,0.2829323709,0.3547560871,0.3475917876,-0.2640645206,0.1405351162,0.2678466439,-0.2227523178,-0.1304278076,0.177875489,-0.1076065302,0.2173901796,-0.1289331913,0.0439179651,-0.081497103,-0.1311272681,0.08348497,-0.430336237,-0.1338418573,-0.0598375127,0.2861565948,0.3971822858,0.1749394983,0.0044677169,0.0369785987,-0.3076259792,-0.1168514043,-0.0747248083,-0.0788575038,-0.0553862154,-0.0985004976,0.7773543596,-0.286259979,0.1690689176,-0.02445562,0.2367896438,-0.0450646318,-0.2427914441,-0.0508058034,0.0319436751,-0.1385505646,-0.0690859258,0.1078336611,0.4174990952,0.0375690758,-0.1575308144,-0.0531437695,-0.1485236287,-0.0462270901,-0.2435237616,-0.2073337138,0.2868824601,0.0515742823,-0.1594630033,0.3107645214,-0.437997669,-0.2390692085,0.337269336,0.0692639798,0.0916248187,0.3000227511,0.2653143406,-0.370274961,-0.1716799885,-0.2494577914,-0.2490811497,-0.1187176555,-0.125312224,-0.1293733567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/263","title":"[Feature request] Support for external modality for language datasets","comments":"I saw the Mozilla common_voice dataset in model hub, which has mp3 audio recordings as part it. It's use predominantly maybe in ASR and TTS, but dataset is a Language + Voice Dataset similar to @aleSuglia's point about Language + Vision. \r\n\r\nhttps:\/\/huggingface.co\/datasets\/common_voice","body":"# Background\r\n\r\nIn recent years many researchers have advocated that learning meanings from text-based only datasets is just like asking a human to \"learn to speak by listening to the radio\" [[E. Bender and A. Koller,2020](https:\/\/openreview.net\/forum?id=GKTvAcb12b), [Y. Bisk et. al, 2020](https:\/\/arxiv.org\/abs\/2004.10151)]. Therefore, the importance of multi-modal datasets for the NLP community is of paramount importance for next-generation models. For this reason, I raised a [concern](https:\/\/github.com\/huggingface\/nlp\/pull\/236#issuecomment-639832029) related to the best way to integrate external features in NLP datasets (e.g., visual features associated with an image, audio features associated with a recording, etc.). This would be of great importance for a more systematic way of representing data for ML models that are learning from multi-modal data. \r\n\r\n# Language + Vision\r\n\r\n## Use case\r\nTypically, people working on Language+Vision tasks, have a reference dataset (either in JSON or JSONL format) and for each example, they have an identifier that specifies the reference image. For a practical example, you can refer to the [GQA](https:\/\/cs.stanford.edu\/people\/dorarad\/gqa\/download.html#seconddown) dataset.\r\n\r\nCurrently, images are represented by either pooling-based features (average pooling of ResNet or VGGNet features, see [DeVries et.al, 2017](https:\/\/arxiv.org\/abs\/1611.08481), [Shekhar et.al, 2019](https:\/\/www.aclweb.org\/anthology\/N19-1265.pdf)) where you have a single vector for every image. Another option is to use a set of feature maps for every image extracted from a specific layer of a CNN (see [Xu et.al, 2015](https:\/\/arxiv.org\/abs\/1502.03044)). A more recent option, especially with large-scale multi-modal transformers [Li et. al, 2019](https:\/\/arxiv.org\/abs\/1908.03557), is to use FastRCNN features. \r\n\r\nFor all these types of features, people use one of the following formats:\r\n1. [HD5F](https:\/\/pypi.org\/project\/h5py\/)\r\n2. [NumPy](https:\/\/numpy.org\/doc\/stable\/reference\/generated\/numpy.savez.html)\r\n3. [LMDB](https:\/\/lmdb.readthedocs.io\/en\/release\/)\r\n\r\n## Implementation considerations\r\n\r\nI was thinking about possible ways of implementing this feature. As mentioned above, depending on the model, different visual features can be used. This step usually relies on another model (say ResNet-101) that is used to generate the visual features for each image used in the dataset. Typically, this step is done in a separate script that completes the feature generation procedure. The usual processing steps for these datasets are the following:\r\n\r\n1. Download dataset\r\n2. Download images associated with the dataset\r\n3. Write a script that generates the visual features for every image and store them in a specific file\r\n4. Create a DataLoader that maps the visual features to the corresponding language example\r\n\r\nIn my personal projects, I've decided to ignore HD5F because it doesn't have out-of-the-box support for multi-processing (see this PyTorch [issue](https:\/\/github.com\/pytorch\/pytorch\/issues\/11929)). I've been successfully using a NumPy compressed file for each image so that I can store any sort of information in it.\r\n\r\nFor ease of use of all these Language+Vision datasets, it would be really handy to have a way to associate the visual features with the text and store them in an efficient way. That's why I immediately thought about the HuggingFace NLP backend based on Apache Arrow. The assumption here is that the external modality will be mapped to a N-dimensional tensor so easily represented by a NumPy array. \r\n\r\nLooking forward to hearing your thoughts about it!","comment_length":42,"text":"[Feature request] Support for external modality for language datasets\n# Background\r\n\r\nIn recent years many researchers have advocated that learning meanings from text-based only datasets is just like asking a human to \"learn to speak by listening to the radio\" [[E. Bender and A. Koller,2020](https:\/\/openreview.net\/forum?id=GKTvAcb12b), [Y. Bisk et. al, 2020](https:\/\/arxiv.org\/abs\/2004.10151)]. Therefore, the importance of multi-modal datasets for the NLP community is of paramount importance for next-generation models. For this reason, I raised a [concern](https:\/\/github.com\/huggingface\/nlp\/pull\/236#issuecomment-639832029) related to the best way to integrate external features in NLP datasets (e.g., visual features associated with an image, audio features associated with a recording, etc.). This would be of great importance for a more systematic way of representing data for ML models that are learning from multi-modal data. \r\n\r\n# Language + Vision\r\n\r\n## Use case\r\nTypically, people working on Language+Vision tasks, have a reference dataset (either in JSON or JSONL format) and for each example, they have an identifier that specifies the reference image. For a practical example, you can refer to the [GQA](https:\/\/cs.stanford.edu\/people\/dorarad\/gqa\/download.html#seconddown) dataset.\r\n\r\nCurrently, images are represented by either pooling-based features (average pooling of ResNet or VGGNet features, see [DeVries et.al, 2017](https:\/\/arxiv.org\/abs\/1611.08481), [Shekhar et.al, 2019](https:\/\/www.aclweb.org\/anthology\/N19-1265.pdf)) where you have a single vector for every image. Another option is to use a set of feature maps for every image extracted from a specific layer of a CNN (see [Xu et.al, 2015](https:\/\/arxiv.org\/abs\/1502.03044)). A more recent option, especially with large-scale multi-modal transformers [Li et. al, 2019](https:\/\/arxiv.org\/abs\/1908.03557), is to use FastRCNN features. \r\n\r\nFor all these types of features, people use one of the following formats:\r\n1. [HD5F](https:\/\/pypi.org\/project\/h5py\/)\r\n2. [NumPy](https:\/\/numpy.org\/doc\/stable\/reference\/generated\/numpy.savez.html)\r\n3. [LMDB](https:\/\/lmdb.readthedocs.io\/en\/release\/)\r\n\r\n## Implementation considerations\r\n\r\nI was thinking about possible ways of implementing this feature. As mentioned above, depending on the model, different visual features can be used. This step usually relies on another model (say ResNet-101) that is used to generate the visual features for each image used in the dataset. Typically, this step is done in a separate script that completes the feature generation procedure. The usual processing steps for these datasets are the following:\r\n\r\n1. Download dataset\r\n2. Download images associated with the dataset\r\n3. Write a script that generates the visual features for every image and store them in a specific file\r\n4. Create a DataLoader that maps the visual features to the corresponding language example\r\n\r\nIn my personal projects, I've decided to ignore HD5F because it doesn't have out-of-the-box support for multi-processing (see this PyTorch [issue](https:\/\/github.com\/pytorch\/pytorch\/issues\/11929)). I've been successfully using a NumPy compressed file for each image so that I can store any sort of information in it.\r\n\r\nFor ease of use of all these Language+Vision datasets, it would be really handy to have a way to associate the visual features with the text and store them in an efficient way. That's why I immediately thought about the HuggingFace NLP backend based on Apache Arrow. The assumption here is that the external modality will be mapped to a N-dimensional tensor so easily represented by a NumPy array. \r\n\r\nLooking forward to hearing your thoughts about it!\nI saw the Mozilla common_voice dataset in model hub, which has mp3 audio recordings as part it. It's use predominantly maybe in ASR and TTS, but dataset is a Language + Voice Dataset similar to @aleSuglia's point about Language + Vision. \r\n\r\nhttps:\/\/huggingface.co\/datasets\/common_voice","embeddings":[0.0000579982,-0.1425000429,-0.1048433781,-0.1137593612,-0.0208638422,-0.2757304013,0.3570465446,0.1619891822,-0.1701952517,-0.115410395,-0.1062394306,0.3741975129,-0.2327695489,0.0247858353,0.2276581675,-0.1631375402,0.1156109571,0.0206510182,0.1751564294,0.0377422199,-0.0802958757,-0.1862993091,-0.0996891856,0.0158086233,-0.1543438584,-0.0065530962,-0.2325278074,-0.0663946643,0.0174453668,-0.1567316949,-0.4168612659,0.1775110662,-0.0064961114,-0.0587129258,-0.000105389,0.0410651825,0.0508660674,-0.1544797868,-0.0636066273,-0.2922307551,-0.1718059927,-0.5856842995,-0.0051841163,-0.2138669938,0.1689768434,-0.2611020803,0.303366214,-0.0154541899,0.219957605,0.042448286,0.2338732183,0.1619654,0.0570976958,0.3144838214,0.1790788174,0.7218409181,-0.1756500155,-0.0262282193,0.3759553134,-0.1980069876,-0.1714316756,0.5319548249,0.0707272217,-0.1726881266,0.1047964245,-0.1554219872,-0.0742541477,-0.1862260252,-0.182682246,0.26961869,0.5540597439,-0.0946797803,-0.3956744969,-0.2771704197,0.101535894,0.2120435387,-0.1491226256,0.4695759714,-0.084704265,0.2618876398,-0.1411739141,-0.4172467589,-0.2269273251,0.4005026221,-0.0107411835,-0.1144736856,0.1139273718,0.1064259782,0.1271232069,-0.1394746155,0.2405346036,-0.0852413252,0.1974173635,0.1384702623,0.1199183315,-0.147744298,0.092940703,-0.0756618828,0.1010376588,-0.1379873306,0.0141931195,0.1652175933,-0.6861980557,0.3720004857,0.1384753585,-0.3171008229,-0.3466371894,0.1551730186,0.298522681,-0.0579590537,-0.0268653706,-0.0378671661,0.0169135183,0.0440881662,-0.3268453181,0.0000947688,0.1069344431,-0.0153924404,-0.0973811075,-0.292539537,-0.2431646138,0.0581131242,-0.0261005349,-0.0754403844,0.2316321433,0.295160681,-0.0103263939,0.2380757183,0.0051021767,-0.6885832548,0.0688787028,0.0822422132,-0.4507437944,0.112957038,0.0196862388,0.5769885778,0.1056848541,0.0224712547,0.0795072168,-0.0589267202,0.1154198125,-0.3274696171,0.0944871306,0.0146099059,-0.3158082962,-0.282207936,-0.0010033672,0.2286927402,-0.193028748,-0.0952018723,0.0137523692,-0.2000210136,-0.1908557564,0.121775277,-0.0717911124,0.0397867337,0.3123777509,0.9068838954,-0.3320865631,-0.4840673804,-0.0372311957,-0.0365264602,-0.324636817,-0.1530091465,-0.1083315238,0.2535992861,-0.4206463397,-0.1457893252,0.0214555152,-0.0782916471,-0.0730273351,0.2533953488,-0.3021765351,-0.0619365349,0.0733146667,-0.1439300179,0.422486335,-0.2967081666,-0.2142054886,0.1942324638,0.0423240773,0.2526317537,-0.0438900702,0.3999971449,-0.0188864972,-0.1381499916,0.1212410256,0.7113153338,-0.150968194,0.0519453287,-0.1938297302,-0.3600857258,0.2658928335,0.4229554236,0.0130498447,-0.3205803931,0.6397942305,-0.000615358,0.1015567109,-0.2774530053,0.1214610413,-0.193411395,0.2937177122,-0.0562877953,-0.1427305341,-0.2755998075,-0.2272177041,0.0019212904,-0.2704661489,0.2816795111,-0.1663463861,-0.0009555613,0.208978802,-0.1393232346,0.0956419855,0.0041683582,0.2037206292,0.1011403576,-0.4169175923,0.2500483096,-0.0902082399,0.0251920149,-0.0975428373,0.038838096,0.1815153807,0.2594423592,0.0138022294,0.1161562577,0.0984601974,0.5246058702,-0.2238256782,0.0367368646,0.0321959108,-0.0006297429,0.1992013603,0.480219692,-0.0132267922,0.3972020745,0.480912894,-0.6762456298,0.3488892317,-0.1610369831,0.2448391616,-0.1503302455,-0.042594105,0.2094171345,0.248840332,0.0853460506,-0.0793127269,0.0688090697,0.0853889957,0.0471753962,-0.3372054994,-0.1611867249,-0.0550510958,0.194349274,0.0049326029,0.03765646,-0.3816787601,0.1459192485,0.534516871,0.2316180319,0.179944545,0.1668641865,-0.2088233531,-0.3059071302,0.2450826764,-0.2178815007,0.0730601624,0.2838123739,-0.2241439521,0.0624792688,-0.212926656,0.0168988183,0.1520752311,0.06433817,0.4003802836,0.0581446253,0.379850626,0.0560440756,-0.0656488761,-0.4201531112,-0.0221136883,-0.2725446224,0.1746994704,-0.0094886972,-0.189365685,-0.6152865887,-0.1843091547,-0.0398294479,-0.445258677,-0.0595044978,0.0666307211,-0.2852560878,-0.1706150472,0.0751235336,0.0053999494,0.378551811,-0.3618397117,-0.0995696262,0.1389790475,-0.260455668,-0.1546472907,0.0843451992,0.4543462694,0.0610960312,0.2402535081,0.2454445511,-0.0476959907,-0.133087188,-0.4327585399,0.2702728808,0.0578765273,0.1886128038,-0.002010792,-0.5663678646,0.0909249336,-0.0657481998,0.2020861208,0.0889275745,-0.0042456607,-0.0664625615,-0.0416938066,0.1243292913,-0.2534283996,0.0937701836,-0.4923431575,-0.5119807124,0.3841014206,-0.1436410397,0.1656590104,0.5681470633,0.0995381847,-0.0953869373,-0.0172695182,0.2712105215,0.0300444029,0.3557389677,-0.0596465357,-0.000171215,-0.1874789894,-0.2881140709,-0.2174292356,0.1192458197,0.3493945003,-0.3474397063,-0.3242609501,0.0953799263,0.156168282,0.0391207188,0.0027884229,0.0905975923,0.0016481931,-0.0056026275,-0.0388501734,0.0010586765,-0.0685858503,0.3693791628,0.1696063131,0.3322416246,0.010456631,0.4449982047,0.3984428048,-0.0735823065,-0.1130097508,0.4104923606,0.341801554,0.2031654716,-0.1458218843,-0.2318649739,0.461813271,-0.099810183,-0.0304083619,0.2845420837,-0.0223187096,-0.5936282277,-0.230983153,-0.0808477029,-0.4445461929,-0.193429783,0.3490699232,-0.1369313598,0.4724569321,0.1747802347,-0.5142968297,-0.2779099643,-0.0862949044,0.0622660592,0.0467646345,0.230810672,0.0712513477,-0.3800601959,-0.0887860432,0.094128795,0.0836932957,0.2238935381,0.2547115982,-0.1767908186,-0.127438426,0.0343180485,0.1126118004,0.1010146737,-0.3075811863,-0.4294191003,0.1311664283,0.0697894841,-0.2317713052,-0.0498642474,-0.1488519907,-0.0059812935,0.1850325316,0.2465440184,-0.0023722476,-0.1417336166,-0.006456967,0.3172285259,0.0406909883,-0.2438861132,0.1223148108,0.1655285358,-0.2372305393,-0.2288937867,-0.13118276,0.0266881436,0.243184939,0.1105710045,0.0440215431,-0.0077721127,0.1166662052,0.1760064214,-0.1860144734,0.0162820574,-0.1535492837,-0.0094976211,0.2016980499,-0.1232051104,0.2899032235,0.1724204421,-0.2464565188,-0.0342411175,-0.2041366845,0.5062916875,0.3987755179,0.555439353,0.1340870708,-0.2353624701,-0.1000528038,-0.4058998823,0.0784540176,0.1703917235,-0.0879821107,0.0928243622,-0.2557076216,0.177929759,-0.1484358013,-0.2434948832,-0.0016531135,0.1146700084,-0.3192114532,0.4626779556,0.2678561509,0.9083020091,0.1739370823,0.3773852289,0.0632027984,-0.139299795,0.4011537135,-0.5117281079,0.1901806742,-0.2407249808,0.0361801311,0.0237822328,0.0500327013,0.0928927436,0.1420457512,-0.0994795635,-0.0832984149,0.6208557487,0.1133584604,0.1424345523,0.2351865172,-0.0402040407,-0.5092383027,-0.0592830367,0.1355792284,-0.1756825447,-0.2223032415,-0.1440446973,-0.3420350254,0.081941247,0.2642249763,-0.2147074491,-0.0011316941,-0.316054076,0.1708391458,0.0972714871,0.0274491161,0.4782885015,0.0652947873,0.103047885,-0.1676760763,0.0214994159,-0.112146154,0.1028367281,0.1289995611,-0.0645298436,-0.2993836999,0.5649161935,-0.0850202814,-0.2367733568,0.0830471367,0.0265793484,-0.0678922683,-0.3287886977,-0.1408794671,0.3525542319,0.0692692399,-0.204262495,0.3383693695,0.0336397365,0.0056782272,0.1585759819,0.1495663524,-0.0072706854,0.254992038,0.0907109082,0.0957420543,-0.0411472321,-0.1602442861,-0.1721546501,-0.0682337284,0.0020590571,-0.0395036861,0.0266543794,-0.2097666264,0.4513889253,-0.165593937,-0.0678607747,-0.2193889022,0.1647765487,-0.1918230802,-0.1911372095,0.0416479334,0.1102282107,0.0128573095,-0.056775935,-0.3587937653,-0.0227739848,0.4008126855,-0.4294159412,-0.0583019592,-0.0565981828,0.150400877,-0.0135832634,0.4569157362,-0.3693554699,-0.0621005334,0.2285134196,0.2409587353,-0.072363928,-0.3310264945,-0.1456676275,0.0843308046,0.1410640925,0.1694407761,-0.2751306593,-0.16777955,-0.075637877,0.1569445878,0.2780585885,0.0947420076,-0.0844768286,-0.3365341425,-0.1436418742,-0.1429385245,-0.3242317736,-0.1453953534,-0.2079687715,0.422285229,-0.6543209553,0.2695195079,0.2933262885,0.0965970084,-0.1245065331,0.3523571193,0.2550243139,0.0950312018,0.0871967599,0.1092824489,-0.2875727713,0.2659047842,-0.1545007229,0.001340931,0.3109216392,0.4030639231,0.1597541869,-0.0666116849,0.1476008296,-0.1723828763,-0.240617767,0.0150780082,0.3615274727,0.2620097101,-0.203440845,0.0838944167,0.566072166,-0.1636136323,0.1882482171,0.1468002498,0.8290458918,0.2113883197,0.1585373282,-0.3222102523,0.031565316,-0.1468386799,-0.0865758136,-0.0459187441,0.181945309,0.0793437287,0.0689206421,0.2830908597,0.4801355898,0.3831429482,-0.3467300534,0.1410560906,-0.0798541233,0.1351424158,0.5679847002,-0.0949716717,0.2906667292,0.1286545843,0.1338963211,0.1971307099,-0.3649289906,0.1564904451,-0.0954354033,-0.0901187509,-0.0247576255,0.2911919057,0.1542281657,-0.0017625219,-0.3724502921,-0.039304208,-0.4641978741,0.0071829511,-0.0693771392,0.0064607002,0.037394993,0.228744477,0.1294110417,-0.1567295045,0.1997879148,0.2505199313,0.4216548502,-0.2381192893,0.0469047427,0.0342185535,0.0821622536,0.1136531308,0.1388760656,-0.025581643,0.0380305126,-0.3868378103,-0.0602942519,-0.3789958954,-0.0223439839,-0.1458857358,0.0431689508,0.1438975781,0.3281206787,0.3592202663,0.078514576,-0.141339913,0.045587033,0.0547987781,-0.248927176,-0.3126660883,-0.0954528376,0.1773027927,0.030504439,-0.0231337044,0.0370859206,-0.2534484565,-0.4080215991,0.4324475229,-0.1037863418,-0.0051232195,-0.0782962739,0.0743996873,-0.1761177778,-0.0367306247,0.4093006849,0.1028505489,0.1296515912,-0.195560798,-0.2923826277,-0.0247532614,0.1040319726,-0.4462167621,-0.0645439923,0.2829323709,0.3547560871,0.3475917876,-0.2640645206,0.1405351162,0.2678466439,-0.2227523178,-0.1304278076,0.177875489,-0.1076065302,0.2173901796,-0.1289331913,0.0439179651,-0.081497103,-0.1311272681,0.08348497,-0.430336237,-0.1338418573,-0.0598375127,0.2861565948,0.3971822858,0.1749394983,0.0044677169,0.0369785987,-0.3076259792,-0.1168514043,-0.0747248083,-0.0788575038,-0.0553862154,-0.0985004976,0.7773543596,-0.286259979,0.1690689176,-0.02445562,0.2367896438,-0.0450646318,-0.2427914441,-0.0508058034,0.0319436751,-0.1385505646,-0.0690859258,0.1078336611,0.4174990952,0.0375690758,-0.1575308144,-0.0531437695,-0.1485236287,-0.0462270901,-0.2435237616,-0.2073337138,0.2868824601,0.0515742823,-0.1594630033,0.3107645214,-0.437997669,-0.2390692085,0.337269336,0.0692639798,0.0916248187,0.3000227511,0.2653143406,-0.370274961,-0.1716799885,-0.2494577914,-0.2490811497,-0.1187176555,-0.125312224,-0.1293733567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/261","title":"Downloading dataset error with pyarrow.lib.RecordBatch","comments":"When you install `nlp` for the first time on a Colab runtime, it updates the `pyarrow` library that was already on colab. This update shows this message on colab:\r\n```\r\nWARNING: The following packages were previously imported in this runtime:\r\n [pyarrow]\r\nYou must restart the runtime in order to use newly installed versions.\r\n```\r\nYou just have to restart the runtime and it should be fine.\r\nIf you don't restart, then it breaks like in your message.","body":"I am trying to download `sentiment140` and I have the following error\r\n\r\n```\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n 419 self._download_and_prepare(\r\n--> 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n 422 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 472 try:\r\n 473 # Prepare split will record examples associated to the split\r\n--> 474 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 475 except OSError:\r\n 476 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 652 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 653 example = self.info.features.encode_example(record)\r\n--> 654 writer.write(example)\r\n 655 num_examples, num_bytes = writer.finalize()\r\n 656 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write(self, example, writer_batch_size)\r\n 143 self._build_writer(pa_table=pa.Table.from_pydict(example))\r\n 144 if writer_batch_size is not None and len(self.current_rows) >= writer_batch_size:\r\n--> 145 self.write_on_file()\r\n 146 \r\n 147 def write_batch(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 127 else:\r\n 128 # All good\r\n--> 129 self._write_array_on_file(pa_array)\r\n 130 self.current_rows = []\r\n 131 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 96 def _write_array_on_file(self, pa_array):\r\n 97 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 98 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 99 self._num_bytes += pa_array.nbytes\r\n 100 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\r\n```\r\n\r\nI installed the last version and ran the following command:\r\n\r\n```python\r\nimport nlp\r\nsentiment140 = nlp.load_dataset('sentiment140', cache_dir='\/content')\r\n```","comment_length":77,"text":"Downloading dataset error with pyarrow.lib.RecordBatch\nI am trying to download `sentiment140` and I have the following error\r\n\r\n```\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n 419 self._download_and_prepare(\r\n--> 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n 422 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 472 try:\r\n 473 # Prepare split will record examples associated to the split\r\n--> 474 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 475 except OSError:\r\n 476 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 652 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 653 example = self.info.features.encode_example(record)\r\n--> 654 writer.write(example)\r\n 655 num_examples, num_bytes = writer.finalize()\r\n 656 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write(self, example, writer_batch_size)\r\n 143 self._build_writer(pa_table=pa.Table.from_pydict(example))\r\n 144 if writer_batch_size is not None and len(self.current_rows) >= writer_batch_size:\r\n--> 145 self.write_on_file()\r\n 146 \r\n 147 def write_batch(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 127 else:\r\n 128 # All good\r\n--> 129 self._write_array_on_file(pa_array)\r\n 130 self.current_rows = []\r\n 131 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 96 def _write_array_on_file(self, pa_array):\r\n 97 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 98 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 99 self._num_bytes += pa_array.nbytes\r\n 100 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\r\n```\r\n\r\nI installed the last version and ran the following command:\r\n\r\n```python\r\nimport nlp\r\nsentiment140 = nlp.load_dataset('sentiment140', cache_dir='\/content')\r\n```\nWhen you install `nlp` for the first time on a Colab runtime, it updates the `pyarrow` library that was already on colab. This update shows this message on colab:\r\n```\r\nWARNING: The following packages were previously imported in this runtime:\r\n [pyarrow]\r\nYou must restart the runtime in order to use newly installed versions.\r\n```\r\nYou just have to restart the runtime and it should be fine.\r\nIf you don't restart, then it breaks like in your message.","embeddings":[-0.4892075658,0.3927338719,-0.0271265376,0.1367761344,0.1485611051,0.1173222139,0.2704391778,0.5356116295,0.0611658767,-0.1925094575,-0.1285087019,-0.0726970136,-0.0604599528,0.0435563102,0.1996953934,-0.3966560066,-0.0350689851,0.3276588619,0.1510497481,0.1637248844,-0.172482878,0.1492370367,-0.3593584895,0.1444754452,-0.3186692894,-0.3128395081,-0.1183318347,0.1084738076,-0.3839229643,-0.4500899911,0.3358423114,0.0578327812,0.3932600319,0.2650594115,-0.000117448,0.1116172746,0.1905764639,-0.1977074444,-0.4337591827,-0.4006903768,-0.0441719815,-0.230258882,0.0015386335,-0.2227816284,0.1908577085,-0.4440507293,0.1007414684,0.2061419487,0.2961962521,0.3937441409,0.1861062348,0.3695085943,0.3249904215,-0.1696384549,0.4855354428,0.1941499412,-0.0995773673,0.4899009764,0.0804633275,-0.1694681495,-0.0587596558,0.1225319132,-0.2555664182,0.3563007116,-0.0102964174,0.1473158002,0.2881261706,-0.1210029796,0.1599364132,0.1933370084,0.6528946161,-0.3635427058,-0.3038655221,0.0054645441,0.2109954208,-0.352116853,0.037499506,0.443918258,-0.5277152061,0.1056669205,-0.3487102091,-0.096262902,-0.0807079524,0.3184595406,-0.0237489827,-0.034094654,-0.0054722689,0.1258220822,0.0778660998,0.0793051571,0.4359017909,0.0207992177,-0.1023334712,0.1800541282,-0.2871754766,0.109818086,-0.1221947819,0.2700802386,0.2193382233,-0.0537812561,0.1338996291,-0.1652329117,0.2458812594,0.1196235791,0.1868953705,0.2888101339,-0.1429705471,-0.2260627449,0.0829186216,0.0914286673,0.184903115,-0.0814399645,-0.0812835768,-0.3262293637,0.2995933294,-0.0103330687,0.2676487863,-0.1157412603,-0.4714718759,-0.0896422118,-0.2914097011,-0.1754256636,0.0908308774,0.3734006584,-0.1583016962,-0.0375185944,0.0479477458,0.2270958871,-0.1774018407,-0.1854207963,-0.1569811702,0.0567496531,-0.2840721011,-0.1020767167,0.3571476638,0.1962182224,0.4431747198,0.0573271587,-0.0920140594,0.0906986073,0.1681345105,-0.2038195282,0.0099528041,0.4029393792,0.1610419899,-0.0081352834,0.05793266,-0.0021745807,-0.0843068138,0.226747632,0.0306001659,-0.3220888972,-0.199762851,0.1254594326,-0.2575152516,-0.0970946923,-0.293526113,-0.1229139864,0.0695314184,-0.2298915684,-0.165683642,-0.2550082207,-0.0134389959,-0.3429088891,0.2889151871,0.1961348653,-0.5239393115,-0.0606309958,-0.2506811321,-0.1861683428,0.3700190187,0.1299467832,-0.3432075083,0.563421011,-0.2346496284,0.3071200252,0.6222015023,-0.287848711,-0.7235483527,0.2745607197,-0.1867569089,-0.090546079,0.001762665,0.2890398502,0.286121875,-0.0463762544,0.0073553002,0.2426165342,-0.0895427242,0.0145119717,-0.209961623,-0.1354868859,0.2433937788,0.2387472242,-0.0977547318,0.1114543006,0.0455193296,0.2303031832,0.4403785169,0.0453526527,0.2413324118,0.0603364222,0.1927467883,-0.0562493317,-0.0980714411,-0.0956363752,-0.4647476673,0.057584241,-0.4637873173,-0.0198911894,-0.5254127383,-0.2377028763,-0.5607109666,0.070500955,-0.3272235394,-0.033638265,0.0952113345,0.0652716681,0.1929218471,-0.010928872,-0.0410111099,0.2827192545,0.1616543233,0.1452789456,0.1667373478,0.3377925754,-0.2518784404,-0.4513130486,0.2220562398,0.2295018733,0.2298679948,0.0118251415,-0.003103157,0.3668798506,-0.1924512088,-0.1402219832,-0.0555098988,-0.4057013094,0.1735838801,-0.2317463011,0.1499910355,-0.0521119349,0.14243415,0.0576541834,0.1917726099,0.0500125848,-0.1557822376,-0.0402871035,0.0167806037,0.2500864863,0.209070608,-0.0379022621,0.3518637717,0.1167308167,0.1862058342,-0.0336596109,0.3855115771,-0.1243394911,-0.0647965968,-0.059909761,0.4924831092,-0.1291579753,0.2608294189,0.0292297117,-0.1564101875,-0.1048164964,0.0698584393,-0.0398321673,0.477727145,0.2122194916,0.048465874,0.0114498651,-0.1163005158,-0.1007153019,0.1045714393,0.1227829307,0.1494815201,0.1262163073,0.122737743,-0.1358026266,-0.2230208069,-0.2798749208,0.2818372548,0.5004068017,-0.2372696996,-0.1715032756,-0.2075822502,-0.168447122,-0.2163604647,-0.3563213348,-0.0726947039,-0.320979923,0.0179656278,0.1279006004,0.084592469,0.1943110675,-0.0190217104,-0.1591059119,0.0522343367,-0.2081071585,-0.0398574248,-0.2335075438,-0.3417904973,0.0147319352,0.130126819,0.1167403236,0.363530755,-0.3642138243,0.0080448585,-0.3366523385,-0.0863478556,0.1599424332,-0.1055041999,0.1739455462,0.349922955,0.5634797812,-0.321973145,-0.1788387746,0.1370937079,-0.0134375338,-0.2322247028,0.1821423322,-0.0834413394,0.0898994505,-0.0191618372,-0.4091536999,-0.5689952374,-0.3764414191,0.1992646009,-0.1053888053,0.1356163621,0.1178278029,0.145246312,0.2287512273,-0.0664414391,0.1976379901,-0.0514190905,-0.2429492176,0.3138307333,-0.1990071535,-0.342282176,0.158926025,-0.0846386477,0.1505794525,0.3035141826,-0.5758484006,-0.0574271679,-0.0322524346,0.589515388,0.119772993,0.0387480259,0.3558401763,0.0227574892,-0.0657013655,0.009441861,-0.0545164831,0.0367188342,-0.2966586053,0.4122065902,-0.082840547,-0.0544182509,0.0060928394,0.8633522987,-0.0095525086,-0.0644375384,0.3446916342,-0.1457499564,0.0865022689,0.0207300466,-0.2893950045,-0.0176182128,-0.1351748407,0.2453832626,0.166914165,-0.1340153813,-0.1906418949,-0.2510347962,-0.1637034267,-0.3817250431,-0.319003582,-0.1672717482,0.3672538698,0.4683549106,0.134638235,0.1167290285,0.0041070026,-0.2449621707,0.1111370325,0.1338043511,-0.0693313107,0.2152056545,-0.0087106489,-0.201621756,-0.3252120316,0.3502572477,0.1555796266,0.3106236458,-0.1071926579,0.0570253357,0.0328801163,-0.0282163154,0.77460289,0.1083680838,0.1324264854,-0.1934742332,0.2264248133,-0.228109926,-0.0688426495,-0.1239720359,0.502751708,0.159950763,0.0605911948,-0.4501763284,-0.0069930851,0.5193058252,0.3487828374,-0.0490767173,-0.3542655408,-0.2506860197,-0.4769188762,-0.2018501759,0.010334746,-0.0106577482,0.2906792462,0.0557930209,0.1190179363,0.1248076931,0.132576704,-0.0083418461,-0.0255833138,0.2934412062,-0.0517708547,0.1046133414,0.1440155953,0.2256674916,0.2276583761,0.3823162615,0.0290335435,-0.3141415417,-0.0089190798,-0.1611179262,0.0394921973,0.1635888964,-0.1109216809,-0.1015258059,0.2779714763,-0.0507505275,0.0397350788,0.1668094695,0.4143212438,-0.3867508769,-0.2249628603,-0.1781380326,0.4710519612,0.1122733206,0.1021010354,-0.2086734027,0.0267152023,0.0215778966,0.1718598306,-0.1007818058,0.8868646026,0.0107514448,0.2602873445,0.580855608,-0.3665026426,0.2276771814,-0.1868802905,0.1204863712,-0.2928270102,-0.0064375238,-0.1172911674,-0.1176122576,-0.0263489392,0.1964289099,-0.2615427077,0.2200115323,-0.1568207741,-0.0468588732,0.1609899104,0.0589762442,-0.1602376252,-0.1513265371,-0.3159479201,0.0455757752,-0.348588109,0.2227684408,-0.1817676276,-0.2723868489,-0.0910807699,-0.0686987787,-0.2695513666,0.0251842998,-0.338121891,0.036458496,0.1915876716,-0.2730433047,-0.0793624148,0.1309661865,0.1497946233,0.0633744597,-0.2067907602,0.1740848273,0.0971405059,-0.18934986,0.1945969611,0.0982569307,0.2922706008,-0.1235280558,-0.3940219283,0.294742018,-0.1324085593,0.1024965942,-0.06471809,-0.0542061254,0.2522399426,-0.3372144401,-0.1061169878,-0.2111095786,-0.0321292095,-0.2772430182,0.0872601867,0.0519403964,-0.2426812351,0.2357741743,-0.1999478787,-0.2372819036,-0.1814817935,0.3651655018,-0.0421556942,-0.0043106405,0.55330199,0.3192165792,-0.3143989444,-0.2441094369,-0.2828895748,-0.062918134,-0.9106292129,0.1967292726,0.2373434305,0.2654592395,-0.0982799977,0.5354986787,0.129310146,-0.103411369,0.117029354,-0.430021286,-0.3058278561,0.1651007831,0.3079100549,0.1183687672,0.3951390088,0.2511398196,0.2111161947,-0.070361428,-0.2607041597,-0.0741753131,-0.1678236723,0.1773357093,-0.2126110494,-0.0943821296,0.0820900351,0.104158558,0.09972132,0.1210818589,-0.0711982325,-0.1638398319,-0.079205513,0.1596506387,-0.0593797117,-0.2433150858,0.170558989,-0.1161612794,-0.1138617471,0.1257172525,0.1328082532,0.2116765827,0.0166086592,0.1139963716,0.0567951463,-0.0683958381,-0.2562175095,0.1302777976,-0.0985366702,0.1805930287,0.0665906742,0.2927458882,-0.0802249312,-0.0597005524,-0.0688115433,-0.2073355466,0.0362084545,-0.2590167224,0.2924411893,-0.526251018,-0.0051496443,-0.0315847583,0.4530531764,0.0591708459,-0.1012712196,-0.1370524466,0.2682368755,0.1523497999,-0.4767332673,0.0096806642,0.0931736156,0.1418686807,-0.0108119687,0.2186550498,-0.1074090451,-0.1157520115,-0.0313331783,0.1795442849,0.3452129662,-0.0052904999,-0.1898931563,0.5471910834,-0.0144446697,0.1763182878,0.2327429801,0.317261219,0.0170399155,0.3329916894,-0.5497567654,0.1368770897,-0.1789324284,0.0463458411,-0.0818761811,-0.3255369365,0.1990224421,0.3870962858,0.0940829292,-0.0993761569,-0.1549074054,0.6332028508,0.0930413082,0.2031233311,-0.0761756003,0.2235228419,-0.1746201068,0.1439854354,-0.1785447448,-0.1531858742,-0.2055397034,0.0028452263,-0.102657415,-0.2115647197,0.1880029291,0.2344942987,-0.0653129369,-0.5137712955,0.2927722633,0.1684475392,-0.2132462114,-0.2507573664,0.2595613897,0.2752043903,-0.1626134962,0.2250253409,0.4563612938,0.4324546158,0.2736489177,-0.316660285,-0.1096356064,-0.0387203246,-0.0693330467,-0.0625552982,0.1097142175,0.3515844047,0.4193814695,0.3758336306,0.1125433892,-0.0824197009,0.2361307144,0.2010086179,-0.1000992209,0.0734025016,0.4880063236,-0.2408223152,-0.3282158375,-0.1292497814,-0.0110953636,-0.3646892309,0.0840953514,0.1375420243,0.1415962428,0.0554552935,-0.1948576421,0.0436183661,-0.2801865339,0.4929964244,0.4946095645,0.4332982302,-0.4648807943,0.0311111864,-0.6018992662,0.2212471515,-0.3157522082,-0.3498498797,0.0853395686,0.1089675799,-0.1048840731,0.3607356846,0.2757783532,0.0039308947,-0.0179495793,0.1734892428,-0.2404170334,-0.1857463866,0.0377932712,0.1727247387,-0.0635666773,-0.3967821896,0.273336947,-0.1753042191,-0.0100033721,-0.3056263626,-0.3999144137,0.2436999381,0.1893974096,0.11431095,-0.0527985431,0.4787840545,0.1011428386,-0.0348378457,-0.0124196876,0.1526200026,-0.0452956706,0.000704905,0.1206668019,0.3290970922,-0.1067848057,0.3048799634,-0.3448428512,0.3387084305,0.0487286784,0.0968183875,-0.1237951815,-0.0099459942,-0.1182368398,-0.0295627601,0.147132203,-0.1387360692,-0.1716975421,0.0917624235,-0.5351992249,-0.4105947912,0.5157888532,-0.4964693785,-0.19944942,0.0637968853,0.2571902871,-0.0001189275,-0.224772647,-0.570217073,0.2695102394,0.3066809475,0.0865049586,-0.1899316758,0.1250606328,-0.1341867596,0.1082798392,0.1559746265,0.3093648851,-0.0818075761,-0.3986260593,-0.1257156581,-0.2106925845]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/261","title":"Downloading dataset error with pyarrow.lib.RecordBatch","comments":"Yeah, that worked! Thanks :) ","body":"I am trying to download `sentiment140` and I have the following error\r\n\r\n```\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n 419 self._download_and_prepare(\r\n--> 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n 422 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 472 try:\r\n 473 # Prepare split will record examples associated to the split\r\n--> 474 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 475 except OSError:\r\n 476 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 652 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 653 example = self.info.features.encode_example(record)\r\n--> 654 writer.write(example)\r\n 655 num_examples, num_bytes = writer.finalize()\r\n 656 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write(self, example, writer_batch_size)\r\n 143 self._build_writer(pa_table=pa.Table.from_pydict(example))\r\n 144 if writer_batch_size is not None and len(self.current_rows) >= writer_batch_size:\r\n--> 145 self.write_on_file()\r\n 146 \r\n 147 def write_batch(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 127 else:\r\n 128 # All good\r\n--> 129 self._write_array_on_file(pa_array)\r\n 130 self.current_rows = []\r\n 131 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 96 def _write_array_on_file(self, pa_array):\r\n 97 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 98 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 99 self._num_bytes += pa_array.nbytes\r\n 100 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\r\n```\r\n\r\nI installed the last version and ran the following command:\r\n\r\n```python\r\nimport nlp\r\nsentiment140 = nlp.load_dataset('sentiment140', cache_dir='\/content')\r\n```","comment_length":5,"text":"Downloading dataset error with pyarrow.lib.RecordBatch\nI am trying to download `sentiment140` and I have the following error\r\n\r\n```\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n 419 self._download_and_prepare(\r\n--> 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n 422 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 472 try:\r\n 473 # Prepare split will record examples associated to the split\r\n--> 474 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 475 except OSError:\r\n 476 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 652 for key, record in utils.tqdm(generator, unit=\" examples\", total=split_info.num_examples, leave=False):\r\n 653 example = self.info.features.encode_example(record)\r\n--> 654 writer.write(example)\r\n 655 num_examples, num_bytes = writer.finalize()\r\n 656 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write(self, example, writer_batch_size)\r\n 143 self._build_writer(pa_table=pa.Table.from_pydict(example))\r\n 144 if writer_batch_size is not None and len(self.current_rows) >= writer_batch_size:\r\n--> 145 self.write_on_file()\r\n 146 \r\n 147 def write_batch(\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 127 else:\r\n 128 # All good\r\n--> 129 self._write_array_on_file(pa_array)\r\n 130 self.current_rows = []\r\n 131 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 96 def _write_array_on_file(self, pa_array):\r\n 97 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 98 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 99 self._num_bytes += pa_array.nbytes\r\n 100 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\r\n```\r\n\r\nI installed the last version and ran the following command:\r\n\r\n```python\r\nimport nlp\r\nsentiment140 = nlp.load_dataset('sentiment140', cache_dir='\/content')\r\n```\nYeah, that worked! Thanks :) ","embeddings":[-0.4892075658,0.3927338719,-0.0271265376,0.1367761344,0.1485611051,0.1173222139,0.2704391778,0.5356116295,0.0611658767,-0.1925094575,-0.1285087019,-0.0726970136,-0.0604599528,0.0435563102,0.1996953934,-0.3966560066,-0.0350689851,0.3276588619,0.1510497481,0.1637248844,-0.172482878,0.1492370367,-0.3593584895,0.1444754452,-0.3186692894,-0.3128395081,-0.1183318347,0.1084738076,-0.3839229643,-0.4500899911,0.3358423114,0.0578327812,0.3932600319,0.2650594115,-0.000117448,0.1116172746,0.1905764639,-0.1977074444,-0.4337591827,-0.4006903768,-0.0441719815,-0.230258882,0.0015386335,-0.2227816284,0.1908577085,-0.4440507293,0.1007414684,0.2061419487,0.2961962521,0.3937441409,0.1861062348,0.3695085943,0.3249904215,-0.1696384549,0.4855354428,0.1941499412,-0.0995773673,0.4899009764,0.0804633275,-0.1694681495,-0.0587596558,0.1225319132,-0.2555664182,0.3563007116,-0.0102964174,0.1473158002,0.2881261706,-0.1210029796,0.1599364132,0.1933370084,0.6528946161,-0.3635427058,-0.3038655221,0.0054645441,0.2109954208,-0.352116853,0.037499506,0.443918258,-0.5277152061,0.1056669205,-0.3487102091,-0.096262902,-0.0807079524,0.3184595406,-0.0237489827,-0.034094654,-0.0054722689,0.1258220822,0.0778660998,0.0793051571,0.4359017909,0.0207992177,-0.1023334712,0.1800541282,-0.2871754766,0.109818086,-0.1221947819,0.2700802386,0.2193382233,-0.0537812561,0.1338996291,-0.1652329117,0.2458812594,0.1196235791,0.1868953705,0.2888101339,-0.1429705471,-0.2260627449,0.0829186216,0.0914286673,0.184903115,-0.0814399645,-0.0812835768,-0.3262293637,0.2995933294,-0.0103330687,0.2676487863,-0.1157412603,-0.4714718759,-0.0896422118,-0.2914097011,-0.1754256636,0.0908308774,0.3734006584,-0.1583016962,-0.0375185944,0.0479477458,0.2270958871,-0.1774018407,-0.1854207963,-0.1569811702,0.0567496531,-0.2840721011,-0.1020767167,0.3571476638,0.1962182224,0.4431747198,0.0573271587,-0.0920140594,0.0906986073,0.1681345105,-0.2038195282,0.0099528041,0.4029393792,0.1610419899,-0.0081352834,0.05793266,-0.0021745807,-0.0843068138,0.226747632,0.0306001659,-0.3220888972,-0.199762851,0.1254594326,-0.2575152516,-0.0970946923,-0.293526113,-0.1229139864,0.0695314184,-0.2298915684,-0.165683642,-0.2550082207,-0.0134389959,-0.3429088891,0.2889151871,0.1961348653,-0.5239393115,-0.0606309958,-0.2506811321,-0.1861683428,0.3700190187,0.1299467832,-0.3432075083,0.563421011,-0.2346496284,0.3071200252,0.6222015023,-0.287848711,-0.7235483527,0.2745607197,-0.1867569089,-0.090546079,0.001762665,0.2890398502,0.286121875,-0.0463762544,0.0073553002,0.2426165342,-0.0895427242,0.0145119717,-0.209961623,-0.1354868859,0.2433937788,0.2387472242,-0.0977547318,0.1114543006,0.0455193296,0.2303031832,0.4403785169,0.0453526527,0.2413324118,0.0603364222,0.1927467883,-0.0562493317,-0.0980714411,-0.0956363752,-0.4647476673,0.057584241,-0.4637873173,-0.0198911894,-0.5254127383,-0.2377028763,-0.5607109666,0.070500955,-0.3272235394,-0.033638265,0.0952113345,0.0652716681,0.1929218471,-0.010928872,-0.0410111099,0.2827192545,0.1616543233,0.1452789456,0.1667373478,0.3377925754,-0.2518784404,-0.4513130486,0.2220562398,0.2295018733,0.2298679948,0.0118251415,-0.003103157,0.3668798506,-0.1924512088,-0.1402219832,-0.0555098988,-0.4057013094,0.1735838801,-0.2317463011,0.1499910355,-0.0521119349,0.14243415,0.0576541834,0.1917726099,0.0500125848,-0.1557822376,-0.0402871035,0.0167806037,0.2500864863,0.209070608,-0.0379022621,0.3518637717,0.1167308167,0.1862058342,-0.0336596109,0.3855115771,-0.1243394911,-0.0647965968,-0.059909761,0.4924831092,-0.1291579753,0.2608294189,0.0292297117,-0.1564101875,-0.1048164964,0.0698584393,-0.0398321673,0.477727145,0.2122194916,0.048465874,0.0114498651,-0.1163005158,-0.1007153019,0.1045714393,0.1227829307,0.1494815201,0.1262163073,0.122737743,-0.1358026266,-0.2230208069,-0.2798749208,0.2818372548,0.5004068017,-0.2372696996,-0.1715032756,-0.2075822502,-0.168447122,-0.2163604647,-0.3563213348,-0.0726947039,-0.320979923,0.0179656278,0.1279006004,0.084592469,0.1943110675,-0.0190217104,-0.1591059119,0.0522343367,-0.2081071585,-0.0398574248,-0.2335075438,-0.3417904973,0.0147319352,0.130126819,0.1167403236,0.363530755,-0.3642138243,0.0080448585,-0.3366523385,-0.0863478556,0.1599424332,-0.1055041999,0.1739455462,0.349922955,0.5634797812,-0.321973145,-0.1788387746,0.1370937079,-0.0134375338,-0.2322247028,0.1821423322,-0.0834413394,0.0898994505,-0.0191618372,-0.4091536999,-0.5689952374,-0.3764414191,0.1992646009,-0.1053888053,0.1356163621,0.1178278029,0.145246312,0.2287512273,-0.0664414391,0.1976379901,-0.0514190905,-0.2429492176,0.3138307333,-0.1990071535,-0.342282176,0.158926025,-0.0846386477,0.1505794525,0.3035141826,-0.5758484006,-0.0574271679,-0.0322524346,0.589515388,0.119772993,0.0387480259,0.3558401763,0.0227574892,-0.0657013655,0.009441861,-0.0545164831,0.0367188342,-0.2966586053,0.4122065902,-0.082840547,-0.0544182509,0.0060928394,0.8633522987,-0.0095525086,-0.0644375384,0.3446916342,-0.1457499564,0.0865022689,0.0207300466,-0.2893950045,-0.0176182128,-0.1351748407,0.2453832626,0.166914165,-0.1340153813,-0.1906418949,-0.2510347962,-0.1637034267,-0.3817250431,-0.319003582,-0.1672717482,0.3672538698,0.4683549106,0.134638235,0.1167290285,0.0041070026,-0.2449621707,0.1111370325,0.1338043511,-0.0693313107,0.2152056545,-0.0087106489,-0.201621756,-0.3252120316,0.3502572477,0.1555796266,0.3106236458,-0.1071926579,0.0570253357,0.0328801163,-0.0282163154,0.77460289,0.1083680838,0.1324264854,-0.1934742332,0.2264248133,-0.228109926,-0.0688426495,-0.1239720359,0.502751708,0.159950763,0.0605911948,-0.4501763284,-0.0069930851,0.5193058252,0.3487828374,-0.0490767173,-0.3542655408,-0.2506860197,-0.4769188762,-0.2018501759,0.010334746,-0.0106577482,0.2906792462,0.0557930209,0.1190179363,0.1248076931,0.132576704,-0.0083418461,-0.0255833138,0.2934412062,-0.0517708547,0.1046133414,0.1440155953,0.2256674916,0.2276583761,0.3823162615,0.0290335435,-0.3141415417,-0.0089190798,-0.1611179262,0.0394921973,0.1635888964,-0.1109216809,-0.1015258059,0.2779714763,-0.0507505275,0.0397350788,0.1668094695,0.4143212438,-0.3867508769,-0.2249628603,-0.1781380326,0.4710519612,0.1122733206,0.1021010354,-0.2086734027,0.0267152023,0.0215778966,0.1718598306,-0.1007818058,0.8868646026,0.0107514448,0.2602873445,0.580855608,-0.3665026426,0.2276771814,-0.1868802905,0.1204863712,-0.2928270102,-0.0064375238,-0.1172911674,-0.1176122576,-0.0263489392,0.1964289099,-0.2615427077,0.2200115323,-0.1568207741,-0.0468588732,0.1609899104,0.0589762442,-0.1602376252,-0.1513265371,-0.3159479201,0.0455757752,-0.348588109,0.2227684408,-0.1817676276,-0.2723868489,-0.0910807699,-0.0686987787,-0.2695513666,0.0251842998,-0.338121891,0.036458496,0.1915876716,-0.2730433047,-0.0793624148,0.1309661865,0.1497946233,0.0633744597,-0.2067907602,0.1740848273,0.0971405059,-0.18934986,0.1945969611,0.0982569307,0.2922706008,-0.1235280558,-0.3940219283,0.294742018,-0.1324085593,0.1024965942,-0.06471809,-0.0542061254,0.2522399426,-0.3372144401,-0.1061169878,-0.2111095786,-0.0321292095,-0.2772430182,0.0872601867,0.0519403964,-0.2426812351,0.2357741743,-0.1999478787,-0.2372819036,-0.1814817935,0.3651655018,-0.0421556942,-0.0043106405,0.55330199,0.3192165792,-0.3143989444,-0.2441094369,-0.2828895748,-0.062918134,-0.9106292129,0.1967292726,0.2373434305,0.2654592395,-0.0982799977,0.5354986787,0.129310146,-0.103411369,0.117029354,-0.430021286,-0.3058278561,0.1651007831,0.3079100549,0.1183687672,0.3951390088,0.2511398196,0.2111161947,-0.070361428,-0.2607041597,-0.0741753131,-0.1678236723,0.1773357093,-0.2126110494,-0.0943821296,0.0820900351,0.104158558,0.09972132,0.1210818589,-0.0711982325,-0.1638398319,-0.079205513,0.1596506387,-0.0593797117,-0.2433150858,0.170558989,-0.1161612794,-0.1138617471,0.1257172525,0.1328082532,0.2116765827,0.0166086592,0.1139963716,0.0567951463,-0.0683958381,-0.2562175095,0.1302777976,-0.0985366702,0.1805930287,0.0665906742,0.2927458882,-0.0802249312,-0.0597005524,-0.0688115433,-0.2073355466,0.0362084545,-0.2590167224,0.2924411893,-0.526251018,-0.0051496443,-0.0315847583,0.4530531764,0.0591708459,-0.1012712196,-0.1370524466,0.2682368755,0.1523497999,-0.4767332673,0.0096806642,0.0931736156,0.1418686807,-0.0108119687,0.2186550498,-0.1074090451,-0.1157520115,-0.0313331783,0.1795442849,0.3452129662,-0.0052904999,-0.1898931563,0.5471910834,-0.0144446697,0.1763182878,0.2327429801,0.317261219,0.0170399155,0.3329916894,-0.5497567654,0.1368770897,-0.1789324284,0.0463458411,-0.0818761811,-0.3255369365,0.1990224421,0.3870962858,0.0940829292,-0.0993761569,-0.1549074054,0.6332028508,0.0930413082,0.2031233311,-0.0761756003,0.2235228419,-0.1746201068,0.1439854354,-0.1785447448,-0.1531858742,-0.2055397034,0.0028452263,-0.102657415,-0.2115647197,0.1880029291,0.2344942987,-0.0653129369,-0.5137712955,0.2927722633,0.1684475392,-0.2132462114,-0.2507573664,0.2595613897,0.2752043903,-0.1626134962,0.2250253409,0.4563612938,0.4324546158,0.2736489177,-0.316660285,-0.1096356064,-0.0387203246,-0.0693330467,-0.0625552982,0.1097142175,0.3515844047,0.4193814695,0.3758336306,0.1125433892,-0.0824197009,0.2361307144,0.2010086179,-0.1000992209,0.0734025016,0.4880063236,-0.2408223152,-0.3282158375,-0.1292497814,-0.0110953636,-0.3646892309,0.0840953514,0.1375420243,0.1415962428,0.0554552935,-0.1948576421,0.0436183661,-0.2801865339,0.4929964244,0.4946095645,0.4332982302,-0.4648807943,0.0311111864,-0.6018992662,0.2212471515,-0.3157522082,-0.3498498797,0.0853395686,0.1089675799,-0.1048840731,0.3607356846,0.2757783532,0.0039308947,-0.0179495793,0.1734892428,-0.2404170334,-0.1857463866,0.0377932712,0.1727247387,-0.0635666773,-0.3967821896,0.273336947,-0.1753042191,-0.0100033721,-0.3056263626,-0.3999144137,0.2436999381,0.1893974096,0.11431095,-0.0527985431,0.4787840545,0.1011428386,-0.0348378457,-0.0124196876,0.1526200026,-0.0452956706,0.000704905,0.1206668019,0.3290970922,-0.1067848057,0.3048799634,-0.3448428512,0.3387084305,0.0487286784,0.0968183875,-0.1237951815,-0.0099459942,-0.1182368398,-0.0295627601,0.147132203,-0.1387360692,-0.1716975421,0.0917624235,-0.5351992249,-0.4105947912,0.5157888532,-0.4964693785,-0.19944942,0.0637968853,0.2571902871,-0.0001189275,-0.224772647,-0.570217073,0.2695102394,0.3066809475,0.0865049586,-0.1899316758,0.1250606328,-0.1341867596,0.1082798392,0.1559746265,0.3093648851,-0.0818075761,-0.3986260593,-0.1257156581,-0.2106925845]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/259","title":"documentation missing how to split a dataset","comments":"this seems to work for my specific problem:\r\n\r\n`self.train_ds, self.test_ds, self.val_ds = map(_prepare_ds, ('train', 'test[:25%]+test[50%:75%]', 'test[75%:]'))`","body":"I am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n","comment_length":16,"text":"documentation missing how to split a dataset\nI am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n\nthis seems to work for my specific problem:\r\n\r\n`self.train_ds, self.test_ds, self.val_ds = map(_prepare_ds, ('train', 'test[:25%]+test[50%:75%]', 'test[75%:]'))`","embeddings":[-0.0186025947,0.0678199828,-0.0675335974,0.1767382026,-0.0039453311,0.1862957478,0.4544264674,0.3805294931,-0.0849740207,0.0096183019,0.157981649,0.2981640697,-0.1939596683,0.1171469241,0.1969036162,-0.3786970079,-0.1162521765,0.2196792066,0.2202848494,-0.0207722802,0.2683160305,-0.0878090486,-0.3013998568,0.1289382279,-0.3250898421,-0.4462779462,-0.162488386,-0.0560760982,-0.2404287905,-0.2245016843,0.2357390076,-0.2476490587,0.0229505431,0.2671960592,-0.0001169801,0.0858824998,0.2045546472,-0.1231242344,-0.0624568313,-0.2040258199,-0.3563699126,-0.0965228453,0.3687441051,-0.15157561,-0.0168993734,-0.2895177603,0.1156505421,-0.1246723235,0.5587252975,0.4705801606,0.0977748558,0.0605919696,0.1509509683,0.0747918859,-0.0116904005,0.3118744195,-0.1074403524,0.07410492,-0.1767975539,-0.1108882874,-0.1365486383,0.0705980286,0.1753382981,0.2046459615,0.0438751914,0.0106822299,-0.3863059282,-0.4449477196,-0.172561869,0.5125129819,0.5099626184,-0.3127263188,-0.3157458603,-0.343855232,0.0373566411,-0.2178646326,-0.1018407866,0.5787149072,-0.1215525493,0.1002738997,-0.1438938379,-0.1982794553,-0.3005862236,0.1699094027,0.0357606858,0.3928737938,-0.092836611,0.0900413319,0.2547083199,0.2649130523,0.2147781998,0.044595059,-0.0077961106,0.3059446216,-0.2638330758,-0.289809823,-0.1192446128,-0.4262012243,0.120447956,0.2507862151,0.13361305,-0.062240608,-0.197851792,0.1702111959,0.4436992109,0.234958902,0.1760654747,0.4938243032,0.0034347884,-0.0556002185,0.0040604332,-0.1536647081,-0.0026444367,-0.2552213371,-0.0048025316,0.1184806526,0.1445331126,0.0970536545,-0.2874823511,-0.3272711933,-0.3107089698,-0.1424574703,0.1399584413,0.1261718869,0.0506234169,-0.0749747679,-0.3134502769,0.2443236858,-0.3853572309,-0.1505621821,-0.1096616611,0.1972878575,-0.4032860398,0.1087486371,0.1017720103,0.0077641513,0.3967711926,-0.0983010009,0.0068024057,-0.3800592721,0.3235695064,-0.0060878438,0.2312981784,0.4229411781,0.1666384041,0.0244422127,0.0921589658,0.1239261255,-0.3356702328,-0.0331380926,-0.3759348691,-0.4590528309,-0.0085356832,0.0288570505,-0.011236514,0.2126413584,-0.2117210627,0.3877968788,0.0655360967,-0.0392750166,0.2040649354,0.0309319589,-0.1864057332,-0.3186094761,0.2620127499,-0.0355582573,-0.3765455484,-0.0840904489,-0.3531986773,-0.0105140554,-0.0375451483,0.1668779552,-0.3595998585,0.5219768882,-0.0765571743,0.4839746058,0.7543491125,-0.2417528182,-0.0982304886,0.2866485715,0.1120669767,-0.2819014192,-0.1141312718,-0.0469170697,0.2763130367,0.1434243768,-0.3279213309,0.4014891088,-0.3133448958,-0.1627570987,0.2922531962,-0.2228444666,0.237905249,-0.0628532991,-0.0785867423,-0.2706501186,0.2040354311,0.2698469758,0.4687471688,0.0151900779,0.1762865335,0.1804482639,0.0468781441,0.1226739883,-0.1034839973,-0.3599649668,-0.367416352,-0.0363817327,0.1209029034,-0.1167773381,0.0136490008,-0.2652081251,-0.1870301068,-0.1521000862,0.021210961,-0.2905018628,0.0500765219,-0.3772895634,0.1867760271,-0.102789849,-0.4393511117,0.0719605759,-0.1744220257,0.2400341481,-0.4450750351,0.1022045538,0.0919367373,-0.0668003112,0.0983601138,0.275967896,-0.0319311321,-0.113819167,0.0215031654,0.5539045334,0.5597566366,-0.0086867781,-0.2220508009,0.0216953121,0.4029690325,-0.4507803023,0.142492488,-0.0588942878,-0.0363886543,-0.0020125839,-0.4457806647,0.3606335819,-0.333240509,0.3146030605,0.1735803038,0.145176217,0.0354459733,-0.1915537566,-0.2770095468,-0.0292928629,-0.269133836,-0.3689307272,0.2141682059,-0.0282619838,-0.1507072747,0.0540156476,0.4456152022,-0.2017521709,0.1850414872,-0.1372690499,0.0644971356,-0.0147520527,0.164878726,0.4185732901,0.2900047898,-0.006122556,0.4006564617,-0.0790501311,-0.2734704912,-0.075665012,0.1154567376,0.2213155627,0.4420474768,-0.0329391249,-0.397331208,-0.086853683,-0.0056658448,0.0353841968,0.1722909659,0.0615746379,-0.2826004624,0.1121321395,-0.3698826134,-0.1115290821,-0.4540897906,-0.2572494745,-0.0854291394,-0.2304375172,-0.1461099237,-0.1568814069,-0.0309084263,-0.0247909334,-0.1584431678,-0.0758845955,-0.1544736028,0.1494784653,0.3911426365,-0.2639917731,-0.1655928046,0.0399479344,0.1162464842,0.595733285,0.2761847377,0.0691654757,-0.1688031852,0.1182705089,-0.2901183367,0.2202607691,-0.2356204689,0.008544609,0.2044394165,0.1658634245,0.0478423536,-0.2953163683,-0.2143628746,-0.1857382655,-0.1527206004,-0.0875526667,0.2334582359,0.0979503989,-0.1764697433,-0.7052025199,-0.2810895443,-0.1937929541,0.2698379159,-0.2516147494,0.2213246673,0.2111864388,0.1571695209,0.3712313473,0.107760489,-0.092757307,0.180675298,0.2621370554,0.0243270695,-0.4707919061,-0.2082182616,0.2600358427,-0.0280088428,0.2169561982,0.0294757113,-0.1711167097,0.1549678594,-0.1848180592,-0.0637829453,-0.0715663061,-0.021511456,0.2696675658,-0.1875339448,0.2374568582,-0.1606770456,0.1127752438,0.1508792937,-0.0787286758,0.1865571886,0.0295359045,0.3088374734,0.2784070075,0.7139962912,0.2629844844,-0.1093328744,-0.2752417028,-0.1456858069,0.0979027525,0.1112567931,-0.0715985149,0.0010460548,-0.0892213061,-0.1402039379,0.0315338485,0.2297079116,-0.249400571,0.0515752807,0.0418275259,-0.38899225,-0.4456570148,0.3325714469,-0.4490879178,0.2467711866,-0.0656051636,0.0451190136,0.0855962485,-0.0442842506,-0.3033319414,0.1236479729,0.2424200326,0.0250772517,-0.3818568289,-0.077334851,-0.2065925449,-0.1082406566,0.1195157096,0.1951927543,-0.1296493858,-0.015336764,0.2284470648,0.2635782361,0.2736069858,-0.1642381847,-0.0948624685,-0.1113729626,0.1148502752,-0.3509507179,0.1359980702,-0.2249267548,0.3831521869,0.0622994117,-0.0545457006,-0.376447618,-0.4656874239,0.4516023397,-0.0590134151,-0.1405365616,0.1336731911,-0.1243652329,-0.2335187942,0.0356477648,0.0920319185,0.3854584694,-0.1581973135,-0.0639085323,0.0587646104,-0.1279884726,-0.0762227401,0.1828724891,-0.0214057826,0.2896789014,0.0741285458,-0.1320736557,0.3965522647,0.0984618813,-0.0376278274,0.303712517,-0.1442089081,-0.1423947215,0.1643046737,-0.0801387802,0.6287440658,0.2862637043,0.0545748882,0.0391619653,-0.083234705,-0.2181189209,-0.2484401464,0.3092346191,0.2322084755,-0.1357232332,-0.2232725173,-0.599224925,0.3932131827,0.1075816751,-0.040671315,0.5002294779,0.2275572121,-0.3482246399,0.4041642845,0.2237427533,0.6722362638,0.0973736048,0.1001954004,0.1406404227,-0.2443891466,0.545237124,-0.4555911124,0.1662275195,-0.1840257943,-0.1389528662,-0.1402612329,-0.2584874928,0.0119016692,0.3969438374,0.0439563058,0.2609879375,-0.102680333,0.0034612573,-0.1664132327,0.3793567419,-0.1090973988,-0.0712629855,0.0334453098,0.0477947183,-0.1448665708,-0.0284888521,0.0470869131,-0.1183993965,-0.2446735948,0.0645509139,-0.384960115,-0.0247672629,-0.1915405244,0.1807472259,-0.0289593767,-0.1621683538,0.2941586077,0.2815640569,0.0504896902,-0.0685516745,-0.0369415246,0.2560127676,-0.2146633118,-0.0173473842,-0.1256807148,-0.0443102382,0.426299423,0.1620701998,-0.2744208872,0.0317089148,-0.1838224083,-0.1125470549,0.0872248113,0.3421586752,0.1746801287,-0.354388535,-0.1089740321,0.1438872069,0.0148540428,0.0391929522,0.0596705154,0.0258319564,-0.250734359,0.1831040531,-0.0185124688,-0.1873093992,0.0804374143,-0.1562560499,0.0968452692,0.1265891939,0.3459616005,0.0443077758,0.0166203883,-0.2344433367,0.3440574706,0.5174093246,-0.3070377707,-0.0465173684,-0.4088949561,0.1494287252,0.0895441398,-0.1740433127,0.0380897224,0.3333357573,-0.3144913018,-0.4444495142,0.0278803874,0.3417975307,-0.1780245602,0.3478838503,-0.2438233346,0.1185720861,0.0706982911,0.4230944514,-0.1612760425,0.1914980561,0.1569704711,0.3527011275,0.0430305265,0.1054606289,0.2481154948,-0.1356279999,-0.0801785961,0.076657325,0.096910499,-0.1624695659,-0.1440149248,0.1574914902,-0.0706927627,0.0307638105,-0.2150485367,-0.2650292814,-0.2186556756,-0.0473310016,0.0042865789,-0.0976666287,0.1837761402,0.4265225232,0.3082600236,0.160205096,-0.2202927768,0.0726649836,-0.0770183653,0.349851042,0.0786374733,0.2602566183,0.1435099393,0.0998276398,-0.2608238161,-0.0279920157,0.2979618609,-0.4033307731,-0.039421536,0.2067458928,0.2281352282,-0.6105564237,0.3794700503,0.4685684443,0.112731725,0.1652246714,0.0898432136,0.0635879561,0.146805793,0.2376798391,0.260668993,0.1859203428,-0.009955748,0.2944073379,0.1209144518,0.6356110573,-0.3971501589,-0.0745274872,0.1207860634,0.0151045658,0.3635282218,0.4176061153,-0.2986032963,-0.1151211858,0.2397633791,-0.3498406112,0.3001830876,0.163989827,0.0871047303,0.5904926062,0.0476487763,0.0470109023,0.399153918,-0.086830318,-0.5318815708,0.0537841618,0.1008657888,-0.0893194675,-0.0804368854,0.5899493098,-0.0614105202,-0.2744282782,-0.0472729094,0.1467450261,0.0745282918,-0.0203052405,-0.4402687252,-0.0016304037,-0.2677973509,-0.1046602353,0.1856311113,-0.3094089925,0.3539857268,0.062424615,0.1002440602,-0.0299452394,0.2893149853,0.2652161419,0.0779205337,-0.0167497322,0.0098033547,-0.1636151522,-0.0456190743,-0.0923274755,0.1761132777,0.1490956098,-0.086227715,-0.3744151592,0.1067139804,0.2882679403,-0.1811639071,0.1844592541,0.2109506875,-0.09573818,0.0581299216,0.144209519,0.099802874,-0.0359472819,-0.0017200718,0.1333235651,0.0671746135,-0.4371813238,0.4507311285,-0.1003253534,-0.1818158925,-0.115173012,-0.003097554,-0.2519983649,-0.1963945627,0.5572986007,-0.3428584635,0.0952700749,-0.1659739614,0.0025320882,-0.1193833351,0.6049438119,0.0485501476,0.2872435451,-0.2306391448,-0.2353113145,-0.3484077752,-0.1724646986,-0.2616736889,-0.1431108117,0.1164469197,0.3252502084,-0.2003195584,0.1976142079,-0.0523983426,0.0223706942,0.1338509619,0.2788351774,-0.0556447804,-0.2280526012,-0.3200216889,0.1941308826,-0.083384268,-0.2465542108,-0.0296685807,-0.0917253569,-0.2744737864,0.1414147764,-0.1091601849,0.5010811687,0.4788812995,-0.0708258599,0.1626960039,0.4706989825,0.2141718566,0.0618860871,-0.3631548285,0.0631008893,-0.3223737478,0.4241426885,-0.0720060468,0.2387050837,-0.3206014037,0.0645304024,-0.1823389828,0.1154283434,0.0456511341,-0.1840975881,-0.2306181788,-0.051425796,-0.3432809711,0.1858374327,0.8144020438,0.4527972639,-0.3223564923,-0.0065814331,0.1515847147,-0.12273141,-0.074863933,-0.425724566,-0.1299207956,0.1879647523,0.1765194833,0.4325864315,-0.0811382234,-0.7890359163,-0.2377249897,0.2145493478,-0.2427999526,0.0575002655,0.0796240792,0.3678286076,-0.1339787543,-0.0347645544,0.4012277424,0.135888353,-0.3040014505,0.070472233,-0.322542727]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/259","title":"documentation missing how to split a dataset","comments":"Currently you can indeed split a dataset using `ds_test = nlp.load_dataset('imdb, split='test[:5000]')` (works also with percentages).\r\n\r\nHowever right now we don't have a way to shuffle a dataset but we are thinking about it in the discussion in #166. Feel free to share your thoughts about it.\r\n\r\nOne trick that you can do until we have a better solution is to shuffle and split the indices of your dataset:\r\n```python\r\nimport nlp\r\nfrom sklearn.model_selection import train_test_split\r\n\r\nimdb = nlp.load_dataset('imbd', split='test')\r\ntest_indices, val_indices = train_test_split(range(len(imdb)))\r\n```\r\n\r\nand then to iterate each split:\r\n```python\r\nfor i in test_indices:\r\n example = imdb[i]\r\n ...\r\n```\r\n","body":"I am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n","comment_length":101,"text":"documentation missing how to split a dataset\nI am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n\nCurrently you can indeed split a dataset using `ds_test = nlp.load_dataset('imdb, split='test[:5000]')` (works also with percentages).\r\n\r\nHowever right now we don't have a way to shuffle a dataset but we are thinking about it in the discussion in #166. Feel free to share your thoughts about it.\r\n\r\nOne trick that you can do until we have a better solution is to shuffle and split the indices of your dataset:\r\n```python\r\nimport nlp\r\nfrom sklearn.model_selection import train_test_split\r\n\r\nimdb = nlp.load_dataset('imbd', split='test')\r\ntest_indices, val_indices = train_test_split(range(len(imdb)))\r\n```\r\n\r\nand then to iterate each split:\r\n```python\r\nfor i in test_indices:\r\n example = imdb[i]\r\n ...\r\n```\r\n","embeddings":[0.0600995272,0.059202835,-0.0615618974,0.1139700487,-0.0133908736,0.223149091,0.4341289997,0.3423463106,-0.0872614309,0.0904238746,0.112329334,0.3258261383,-0.2517769635,0.1168583408,0.1798712313,-0.3691290915,-0.066240415,0.1591636539,0.2443268448,0.0308177676,0.2214924395,-0.1523120105,-0.372136116,0.0719026774,-0.2917358577,-0.4570458829,-0.2463944703,-0.054137107,-0.2753708959,-0.236896053,0.1701683253,-0.0853734389,-0.0064511513,0.2993582189,-0.0001171174,-0.0006970303,0.221613571,-0.1150745973,-0.1520451605,-0.2749583423,-0.2659641504,0.0024026691,0.3048646152,-0.2019954473,0.0187097117,-0.2679358721,0.1041534394,-0.1138850152,0.5253756642,0.402502358,0.0973768905,-0.0187447518,0.0593397357,0.1293645203,0.0749443918,0.27727139,-0.0650620162,0.1716542989,-0.1823061854,-0.0956431925,-0.135114342,0.0488672815,0.1610796005,0.2262029648,0.1025451124,0.0239890199,-0.3529039025,-0.4566062391,-0.1725031585,0.5094365478,0.4698416889,-0.3314265609,-0.3592270613,-0.3861284554,0.1277313083,-0.2642101645,-0.1312361509,0.5653489828,-0.0973729268,-0.0550903566,-0.1545509845,-0.1297812462,-0.2560816407,0.1392175555,0.0853776261,0.5397906899,-0.032362543,0.0769093856,0.3139477372,0.204372853,0.4018697143,-0.0247702878,-0.0117319776,0.3584611714,-0.2561255097,-0.3121018112,-0.1303187609,-0.4234106243,0.1190472469,0.2578414977,0.1716760248,-0.0856156796,-0.178942889,0.1059741601,0.4475736618,0.2421585768,0.1864495426,0.4735388756,0.0036861252,0.0158502683,0.0550022162,-0.1519695222,-0.0304263625,-0.2850629389,-0.0540900081,0.1259509772,0.0423416011,0.1031108946,-0.3341119885,-0.3942473233,-0.3149623871,-0.1922617704,0.1769528389,0.1530546099,0.0330443643,-0.055171892,-0.3120959699,0.2518874407,-0.3018952608,-0.0045857793,-0.1249805763,0.1991766095,-0.4028161764,0.0909401029,0.137765035,0.0272540599,0.393316716,-0.0744716227,0.0851056948,-0.3233603835,0.2844261229,0.0135471094,0.2084345967,0.391625613,0.1960054189,0.0341712907,0.0379531942,0.1721700579,-0.3572608531,0.0106583713,-0.3762223423,-0.386336714,0.0590663776,0.0359242409,-0.0429656468,0.1375054419,-0.3059713244,0.3850699961,0.0028136522,-0.0842985436,0.1836826652,-0.0354707316,-0.1883960217,-0.3066426516,0.2635630667,-0.0423973612,-0.3014155328,-0.0850330666,-0.3118213713,0.0348725542,0.0092230486,0.2253821194,-0.3613464832,0.4354794025,-0.0787982419,0.4266362786,0.7096788883,-0.1668386012,-0.0830487385,0.1356394887,0.073025167,-0.2475161254,-0.0586425662,-0.0254105758,0.2792952359,0.0822786763,-0.2581385374,0.4794852138,-0.3317678571,-0.0684077665,0.2964524925,-0.2518995702,0.3039956987,0.0026511,-0.1129344031,-0.2537751794,0.1864834726,0.3416492045,0.5174593925,0.0002922888,0.1240524352,0.122132577,0.0347708501,0.2333393097,-0.0684840605,-0.3601818383,-0.2926032841,-0.0724027082,0.1696252078,-0.0901157707,0.0293577574,-0.2904085219,-0.2447777987,-0.2300977111,0.0267193448,-0.2659700215,0.0562235788,-0.3088188469,0.2587308586,-0.1421919763,-0.4538706243,-0.0228313711,-0.2623423636,0.1578739583,-0.4556034505,0.2260335088,0.0970976949,-0.1513530761,0.1235115901,0.3515447378,-0.0676233694,-0.1001838073,-0.0139017086,0.6143959761,0.5653508902,-0.0955953449,-0.1702238172,0.0294235516,0.4321298003,-0.3713900745,0.1643283516,-0.073419854,0.007383246,0.0067607639,-0.4206548035,0.4240078926,-0.5293163657,0.3825000226,0.2183520347,0.0554136634,0.0161225572,-0.0953370705,-0.3166333735,0.0221355837,-0.2015053928,-0.3675967157,0.1182452515,-0.0644110814,-0.206991449,0.0260133315,0.4118818343,-0.1996985674,0.1081074029,-0.1360199749,0.0132391714,-0.040350873,0.1119055226,0.4428805411,0.2646597326,0.0628821775,0.3699391186,-0.0981138274,-0.2832746208,-0.0941550434,0.1123309731,0.2549903393,0.4364367127,0.0110180536,-0.2843090296,-0.0864050016,-0.0240161959,-0.0805049613,0.1877817214,0.083736971,-0.2561196387,0.0638338253,-0.3536810577,-0.1584827751,-0.4681741893,-0.3202151954,-0.0568619035,-0.2556049824,-0.1227160916,-0.0339344889,-0.0542002209,0.0088231368,-0.1515870839,-0.0395451784,-0.124467805,0.1857990175,0.344530344,-0.3150709867,-0.1321641803,0.0329206064,0.1363002807,0.6758494377,0.3586135507,0.0377044044,-0.1696942449,0.1066739187,-0.2268465608,0.1913998276,-0.1884536445,-0.0037773808,0.1433348805,0.1277379245,-0.0425320528,-0.4658038318,-0.1937800497,-0.212925896,-0.1523442566,0.036781352,0.2246374488,0.1164393201,-0.1551230252,-0.6601370573,-0.343165189,-0.1810243577,0.1933230311,-0.207370609,0.2268755883,0.2438796163,0.1859435141,0.3274717629,0.0146223428,-0.117303066,0.1914955527,0.2556433678,0.0435611084,-0.4491837025,-0.1671505123,0.1876729876,-0.0665262043,0.0979472101,0.0923764408,-0.1088687405,0.1633361131,-0.2055927217,-0.054957144,-0.0582875721,0.0279396754,0.2760000229,-0.1686510444,0.2373516411,-0.1299887002,0.1291237622,0.1275572032,-0.0030874407,0.2730184495,0.0085010016,0.3296947479,0.2465957254,0.7530816197,0.2361808568,-0.0867044628,-0.2982821167,-0.1320105642,0.0575513244,0.0226036217,-0.0801631659,0.121966064,-0.0013240745,-0.169676125,0.0868730471,0.1664909869,-0.3392602205,0.0656096712,0.0736569092,-0.3866912723,-0.4350097477,0.3419037461,-0.4380700588,0.1782549471,-0.0613769069,-0.0173338745,0.0224900972,-0.0949844792,-0.3401085138,0.0985811278,0.2204699814,0.0309354067,-0.4048057497,-0.2065958381,-0.2126451135,-0.1574431956,0.1924208701,0.2665263712,-0.0264681727,-0.0211815275,0.2608845234,0.2814454734,0.2959349751,-0.1504775286,-0.0525011867,-0.0398290232,0.1245618239,-0.3295584321,0.0693333298,-0.3494670987,0.3621776104,0.1006673053,0.0253793504,-0.3897855282,-0.4066637158,0.4339585602,-0.0061200066,-0.1357493401,0.129872188,-0.1702317595,-0.1985980868,-0.0111465538,0.1198601276,0.3398828208,-0.0983911529,-0.1400348246,0.1323728859,-0.071853213,-0.0828451663,0.1359932125,0.0486008078,0.3254575729,0.2151256353,-0.1158573255,0.4347855747,0.0229687355,0.0072505753,0.2126640677,-0.1215586215,-0.1124666855,0.1488222778,-0.0820965767,0.5192819238,0.3348642886,0.1167123318,0.0019601283,-0.088852942,-0.1721151322,-0.2203487754,0.2200643271,0.1969327778,-0.1629000157,-0.2322635502,-0.7845417261,0.3825081587,0.1493138671,-0.0787292346,0.4942558408,0.2227549255,-0.3529698849,0.5404192805,0.2600698471,0.7107943296,0.0270935781,0.1087848395,0.2451622337,-0.2199293077,0.5987612605,-0.3728725314,0.1955491006,-0.2358036041,-0.2317264676,-0.0505060516,-0.2486106157,-0.0422449037,0.3854596913,-0.0192986745,0.3094728589,-0.0920213088,-0.1220292598,-0.2474119663,0.4114217162,-0.2356865108,-0.0378548242,0.0677265748,0.0294263754,-0.1674997956,-0.0603904799,0.1065952554,-0.0735361576,-0.2748829722,0.1142620891,-0.3794391751,-0.0345000885,-0.1439367533,0.1884531528,-0.1470765173,-0.1539306343,0.2308850586,0.1599219888,0.0916601643,-0.0648995489,-0.019740209,0.2676504552,-0.0960786268,-0.0289989822,-0.0911914334,0.014268321,0.4709145427,0.2043625861,-0.2958531082,-0.0350387804,-0.0829460397,-0.1843602806,0.0005770111,0.2810373604,0.2053861022,-0.286005646,-0.139171645,0.2051773369,0.032736171,0.0255577769,0.0658966377,-0.0198378377,-0.2188262343,0.2100426853,-0.0539858043,-0.1761043668,0.0770147741,-0.1460437775,0.0802083835,0.1033793315,0.3675252795,0.0514261536,0.0021328963,-0.2066353261,0.3037539423,0.4383648634,-0.2862949073,0.0494734012,-0.3851535916,0.1446261406,0.0824591145,-0.1435530484,0.090108797,0.4006581604,-0.3112068772,-0.4613431394,0.092972599,0.432479322,-0.0383319855,0.3395060301,-0.2784985602,0.1513526291,0.019041894,0.3190456629,-0.1888807714,0.1861719042,0.1802683771,0.4271774292,-0.0482407585,0.0949818864,0.1235107481,-0.1272827238,-0.0672807097,0.0396105014,0.1075892895,-0.1702583432,-0.1514644772,0.1383125484,-0.0558800623,0.0355234221,-0.2046453655,-0.3283611834,-0.2501081824,-0.1294766068,-0.0613244362,-0.06784264,0.2068958879,0.4533402622,0.4532244205,0.1227358431,-0.0972053111,0.1651911587,-0.1471459419,0.2250366658,0.0359441377,0.3269498348,0.1799125373,0.0998291671,-0.3032548726,-0.0818268806,0.2034571022,-0.3801213205,-0.0686107874,0.1789451391,0.2668625414,-0.4711160064,0.4318476021,0.5926184654,0.1806321591,0.0577298626,0.1817914546,0.0698221847,0.1219519675,0.2370398939,0.1880248636,0.2561095357,0.0035008553,0.3167130053,0.0351857804,0.4766220152,-0.4306510687,-0.0701407641,0.1311644614,-0.0663125068,0.366808176,0.4985340536,-0.2365569323,-0.2182097882,0.2221250683,-0.3801796734,0.2452473491,0.1780261248,0.1286341101,0.6150350571,0.0819283947,-0.0128322961,0.341775924,-0.0588045418,-0.4968168736,0.1731159985,0.0913525298,-0.0201136358,0.003830692,0.5978030562,-0.0783997923,-0.1626096815,-0.0143508771,0.1705261916,0.1151888892,-0.0637390167,-0.4369430244,-0.0614455342,-0.3473995328,-0.0609429143,0.1126243472,-0.2696665525,0.3236602545,0.0872591138,0.0400099829,-0.1954786927,0.3290708959,0.2906000614,0.1356917024,-0.0851239637,-0.0789255276,-0.1341346204,-0.0441853888,0.0419453382,0.2085326463,0.1113721728,-0.0033137575,-0.4120709598,0.1005015597,0.2396113724,-0.2206005752,0.1848676354,0.2230178714,0.0269466285,0.063488327,0.1378449053,0.1230117679,-0.0588529408,-0.0059314957,0.1754356623,0.0122867366,-0.4112175107,0.4289708436,-0.0460764728,-0.1350552887,-0.044596985,0.0144310882,-0.3357693255,-0.1644351035,0.3247780502,-0.3392886221,0.0555931069,-0.1636827588,0.0027706136,-0.0965679511,0.5712146759,0.1023399755,0.2575375736,-0.1659798771,-0.242295593,-0.3547655344,-0.2354462147,-0.1837561131,-0.1635542661,0.1517813206,0.4201254845,-0.2421985567,0.2116318345,-0.1346132606,0.0761364102,0.045459643,0.3558129668,-0.1069984213,-0.2201104611,-0.3341938257,0.2212844044,-0.1073456928,-0.2415921986,-0.0487651825,-0.0650852323,-0.2900025249,0.1640608013,-0.0977802575,0.4897425771,0.4773822427,-0.1549904346,0.1331690997,0.4795996249,0.2026745528,0.0995793939,-0.4411298633,-0.0005304368,-0.4071692526,0.3740480244,-0.0343128331,0.2950650454,-0.3066810966,-0.0082552023,-0.1573832184,0.1780373603,-0.0225819238,-0.2455612719,-0.3620999157,-0.0412084274,-0.3270583451,0.2329274267,0.6373425126,0.4534019232,-0.3058331013,0.0039184652,0.1617230624,-0.0940980017,-0.0126626836,-0.4618834257,-0.1761028916,0.1094803512,0.137165308,0.4254462123,-0.1104557589,-0.6365734935,-0.1691668928,0.1760030687,-0.1436435133,0.1249101087,0.1179802418,0.3617659211,-0.0888904929,-0.0511528403,0.3349886239,0.1007939726,-0.2868116498,-0.0284935813,-0.3389170468]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/259","title":"documentation missing how to split a dataset","comments":"I added a small guide [here](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) that explains how to split a dataset. It is very similar to the tensorflow datasets guide, as we kept the same logic.","body":"I am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n","comment_length":28,"text":"documentation missing how to split a dataset\nI am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n\nI added a small guide [here](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) that explains how to split a dataset. It is very similar to the tensorflow datasets guide, as we kept the same logic.","embeddings":[-0.0108554037,0.1039159372,-0.0716410503,0.1723709404,-0.009352929,0.1751483977,0.4139045775,0.3510569036,-0.1455844194,0.0215435736,0.1812569052,0.2846695781,-0.1473984122,0.110882692,0.1814140528,-0.3872855008,-0.1115607172,0.229946211,0.2616378665,-0.042115178,0.2874653637,-0.0557160713,-0.2707217634,0.1046781465,-0.347628653,-0.4612329006,-0.1970775723,-0.0371643528,-0.2252805233,-0.2295696884,0.2283925712,-0.2733032405,0.038595479,0.2595723867,-0.0001157783,0.0629990995,0.2459522337,-0.1078856215,-0.1020556837,-0.2007867396,-0.3612455428,-0.0809873492,0.4162449241,-0.1300804913,0.008889907,-0.2497373223,0.1623073071,-0.1735195816,0.50085181,0.4594459832,0.0998941362,0.0351319499,0.1878993064,0.0838824585,-0.0086505422,0.3338405788,-0.1121975332,0.0764554814,-0.1113737598,-0.0766184032,-0.1672828346,0.0434849784,0.1838376224,0.2061329037,0.0807435662,-0.0273496639,-0.3851658106,-0.4645871222,-0.1427486241,0.5450023413,0.605633378,-0.3358956873,-0.2980398834,-0.3045441806,0.0450891331,-0.1760812998,-0.1278075129,0.6296437979,-0.1171363294,0.0876608789,-0.1201952696,-0.2391177118,-0.3045322597,0.1573000997,0.022686664,0.389533788,-0.1093668193,0.0562407859,0.1918927729,0.2463200688,0.2808936238,0.0417102873,-0.0121208252,0.2608874738,-0.179040283,-0.2431488782,-0.090922907,-0.3770047128,0.0993314087,0.2342866361,0.1321774572,-0.0647843778,-0.1741934717,0.1682287008,0.4464684725,0.1732770354,0.2441631854,0.4896550775,0.0379738994,-0.039864704,-0.00231804,-0.1706026644,-0.0124308001,-0.1979073584,-0.0247887224,0.0710361302,0.1537875086,0.0901100263,-0.2858226001,-0.3360194564,-0.2909919322,-0.16204907,0.124273859,0.1403428316,0.0750277936,-0.0726471096,-0.3455613256,0.2231651992,-0.3875484765,-0.1767753065,-0.1235925108,0.2355868071,-0.3899735212,0.1046811789,0.1119999737,0.0107036149,0.4004684985,-0.1086489856,0.0191355199,-0.3715496659,0.306962043,0.0410933457,0.2116015255,0.4237737358,0.2138502598,0.020229198,0.0617886782,0.1171503589,-0.3336219788,-0.0221009646,-0.3795628548,-0.4628950357,-0.0616194271,0.039680291,-0.0250289794,0.1905612051,-0.1604062021,0.3714154661,0.0512458533,-0.048003722,0.2013643533,0.0819937959,-0.1747274697,-0.3574976325,0.2417288423,0.0053395969,-0.4272280931,-0.0731006041,-0.3806426823,-0.0314782411,-0.0557448119,0.1454569399,-0.3786714673,0.5079536438,-0.0999322757,0.4572858214,0.7189478874,-0.278337419,-0.0706058815,0.3035817444,0.1537110507,-0.2755768299,-0.1114162579,-0.0518568456,0.2611777484,0.1561566591,-0.4373097122,0.3556575477,-0.307607621,-0.1306802183,0.3071198165,-0.2115510553,0.2188747823,-0.0768824071,-0.0840034261,-0.2523697317,0.2786921859,0.3431999087,0.4787201583,0.0270526204,0.1831009388,0.2342686355,0.0657580048,0.1468948424,-0.0490043424,-0.3427709639,-0.3662566841,-0.0524929129,0.1418801695,-0.118029058,-0.0259524621,-0.2883369029,-0.1490792334,-0.135079816,0.0492266454,-0.2921785712,0.0640786663,-0.3532393277,0.1958754659,-0.1191392839,-0.4262670577,0.0873605236,-0.2295692265,0.2660537958,-0.4764243364,0.1153511256,0.0713808537,-0.0815802366,0.1057398915,0.2592471838,-0.0510609597,-0.1367089897,0.0266248398,0.5864788294,0.5731501579,0.0404473953,-0.2118848413,0.0319817029,0.4351481497,-0.4945230782,0.1646827757,-0.0357021466,-0.0406303331,0.0197371207,-0.4918183088,0.3089127839,-0.3423790634,0.2918802798,0.2441430837,0.1246892214,0.0534768105,-0.1709048152,-0.2978188694,-0.0349749476,-0.2654487491,-0.3422133923,0.2162171155,-0.0248491727,-0.1540777087,0.1124038547,0.4354673624,-0.1889245361,0.2056400776,-0.1116753146,0.0765462443,-0.0406405702,0.2076422274,0.427867651,0.249644652,0.0174676441,0.4108245075,-0.0889343172,-0.2685613334,-0.0661095902,0.1127064452,0.2360969037,0.4292649329,-0.0678519383,-0.3976389468,-0.0774865597,-0.0187645387,0.0069756387,0.1440650672,0.0676456764,-0.2933172286,0.1057011485,-0.3916169405,-0.1474703848,-0.427064389,-0.2854633629,-0.0807077289,-0.2603794336,-0.1695373654,-0.1485543698,0.0122485897,-0.0461445116,-0.1674004644,-0.0769603178,-0.1710436791,0.1795431823,0.3728763759,-0.2571858466,-0.1507111043,0.0603423268,0.1510509253,0.6260809898,0.2506404221,0.0201234054,-0.1515856236,0.0628092065,-0.328525871,0.2419970185,-0.205597207,-0.0152375707,0.2106800973,0.186044544,0.0140250605,-0.3126119971,-0.2665431201,-0.1900269091,-0.1723769307,-0.0817518085,0.1955081671,0.0547106974,-0.1628088206,-0.7155647278,-0.2830574214,-0.2016948611,0.292462647,-0.2403344214,0.2397518754,0.1748123467,0.1396600455,0.355679065,0.1154142022,-0.0146244513,0.2010682076,0.2305342406,0.0061793267,-0.4905734658,-0.2406401783,0.3171946406,-0.0099633206,0.227502808,0.0062271859,-0.2001265883,0.180793941,-0.2070200741,-0.0661455244,-0.0552999005,-0.0082790134,0.2614949644,-0.1543386132,0.2438741773,-0.1787375808,0.1214395463,0.0943631604,-0.1324971169,0.1973887682,0.0055295047,0.2760886848,0.2678914964,0.6091409922,0.2550810874,-0.0876832902,-0.2653273642,-0.140034005,0.1400356889,0.1263163239,-0.0411760435,0.0360700637,-0.1034703553,-0.1227900684,0.0487057716,0.2632274926,-0.2793763876,0.0587548539,0.007395762,-0.3954569101,-0.4270704091,0.2944313586,-0.445482403,0.2239471525,-0.05564766,0.0323885344,0.1070841625,-0.0738990307,-0.3157726526,0.174303472,0.2323838919,0.0493641943,-0.3276259005,-0.079636924,-0.1881275922,-0.1433453113,0.1301845014,0.1806937307,-0.1262061149,-0.0094609,0.2101885229,0.2227452993,0.2387933284,-0.1617364883,-0.0771251395,-0.113678962,0.1290969998,-0.3410099149,0.1779456288,-0.2492110878,0.4085025191,0.0621353686,-0.1024861485,-0.3521377444,-0.4546485245,0.4207289517,-0.059448611,-0.1069566384,0.1584547013,-0.143981114,-0.252548039,0.0172165018,0.0231680833,0.3640353978,-0.1785525531,-0.0514591075,0.0391216688,-0.0944004282,-0.0846269131,0.1701571047,-0.0255515762,0.2940468788,0.012692512,-0.1491545588,0.3754704595,0.1125066876,0.0303527117,0.2963512242,-0.1479453892,-0.190825358,0.2248423547,-0.0847334713,0.5996389389,0.2911293805,0.0861459374,0.0129284309,-0.0889667049,-0.2382880896,-0.1926456988,0.2559098899,0.1834310144,-0.1058775559,-0.1930080652,-0.6062171459,0.3586286306,0.1057371572,-0.0328869931,0.5307198167,0.2949858904,-0.3071547449,0.3670976162,0.1933377832,0.6510209441,0.1368878484,0.0759351403,0.1357086897,-0.3133864403,0.5474331975,-0.4472092092,0.1698411256,-0.1781846285,-0.1392391026,-0.1614579558,-0.2407468408,0.0084845154,0.390365839,0.0030031144,0.2652426064,-0.1245295256,-0.0353010483,-0.147969991,0.427919209,-0.1697594225,-0.0500480868,0.0545200892,0.0488322414,-0.1591653228,-0.0319739804,0.037780419,-0.1054269895,-0.226754114,0.100526005,-0.4233619869,-0.011748218,-0.1256310493,0.2007479221,-0.0605028458,-0.1697274894,0.3008081317,0.3042423129,0.0780729428,-0.0898846835,-0.0814772919,0.29293859,-0.2199408561,-0.0707832053,-0.1192722768,-0.0041615884,0.428314656,0.1577158719,-0.2871338129,0.0186188649,-0.1946074516,-0.1165346056,0.0995112881,0.342271924,0.1866003126,-0.281345278,-0.0550212376,0.1529145837,-0.0132908691,0.030043358,0.0666293278,0.0447445773,-0.2066187561,0.1536607742,0.0268991068,-0.1933045387,0.0379676558,-0.1451349258,0.0976542309,0.1223956868,0.3040711284,0.1059392914,-0.0430194996,-0.2244167775,0.3682890832,0.4998830855,-0.2677210271,-0.0637045279,-0.4257542193,0.1601600796,0.0921099931,-0.1946090758,0.0206690729,0.3484405279,-0.3426215351,-0.4485455453,0.0572179593,0.3609348238,-0.1981173903,0.3023910522,-0.2448360473,0.0572956502,0.10901618,0.3757447004,-0.1645894647,0.2210297883,0.1791006029,0.3216922283,0.0707693398,0.1065806597,0.2227225602,-0.1663650423,-0.0857793465,0.0274712797,0.1451640576,-0.1592143178,-0.133818388,0.1577709317,-0.1165202856,0.0336209014,-0.2108561695,-0.2579405308,-0.2165601701,-0.0290826708,0.0167508572,-0.0807706341,0.1555975527,0.4094020426,0.2860698104,0.1206945181,-0.2208314836,0.0381516963,-0.0696476474,0.384262383,0.0417539589,0.285574764,0.1301591843,0.0843061358,-0.2262097597,0.0168823786,0.2752984464,-0.3770772219,0.0048546456,0.2419471443,0.2393680364,-0.6506142616,0.3816908002,0.408328712,0.1248109713,0.1871681958,0.0762067884,0.0595989786,0.1614349335,0.167462647,0.2175455242,0.2154757529,0.00147141,0.306612283,0.1469423175,0.5802815557,-0.3549509943,-0.0654189363,0.2021237463,0.0163431522,0.3350754678,0.4409968257,-0.2998600602,-0.1287742406,0.2470846772,-0.3163169622,0.2875592709,0.2036587298,0.0750042498,0.6348125339,0.0664722994,0.0371623337,0.4465915561,-0.0589866377,-0.5137970448,0.0190000366,0.0977782905,-0.1053593457,-0.0819723532,0.5677708387,-0.0392151996,-0.2483867258,-0.047497768,0.1234580725,0.0699603781,-0.0060696225,-0.4208908379,-0.0004839349,-0.2475746423,-0.1170146838,0.1821567565,-0.2999718189,0.3403620422,0.0699844956,0.0929244682,-0.0393268242,0.3636172712,0.2126205117,0.0740130991,-0.003426013,0.0328676142,-0.1711405218,-0.0652417764,-0.1163490713,0.1826274544,0.176479876,-0.076879546,-0.3350023329,0.1181715578,0.3057120442,-0.2009143233,0.1847958118,0.2009731829,-0.1196334735,0.0735204145,0.1132428572,0.1103821024,-0.0363603868,0.0101403454,0.1846975386,0.0753990188,-0.4126007557,0.3575341702,-0.1487631649,-0.192284286,-0.1361565143,0.0094975401,-0.2746709287,-0.1855753362,0.5705321431,-0.3507171571,0.0634169206,-0.1747401059,0.0115937572,-0.0990992859,0.5687583089,0.0702375099,0.3151823878,-0.1661264598,-0.2857992351,-0.3418375552,-0.1546999961,-0.2733696401,-0.1908336133,0.1186345965,0.3484735191,-0.2203210145,0.171377033,-0.0750693455,0.0449249893,0.0984306633,0.288482964,-0.0572170429,-0.2440003008,-0.2553466856,0.1981169432,-0.1302741021,-0.2169383317,0.0023590813,-0.1172226146,-0.2558829784,0.1278804988,-0.126102075,0.5148493052,0.474675566,-0.1135859415,0.1391736418,0.4700137079,0.1789058745,0.0078657791,-0.3639888763,0.0673504621,-0.2879277766,0.4713402987,-0.0828115717,0.194986403,-0.3280073702,0.0448571555,-0.1578531265,0.1421331018,0.0347074904,-0.132641539,-0.2842836976,-0.0388768613,-0.3516290784,0.2111669332,0.7949632406,0.4312686622,-0.3226009011,-0.0499274209,0.1503359526,-0.1283041388,-0.0980373025,-0.398547709,-0.0930350572,0.1671239436,0.1579529196,0.406965673,-0.0927443057,-0.7600876689,-0.224818036,0.1973667592,-0.2541642189,-0.0030321646,0.0345053226,0.3735860586,-0.1024874002,-0.0669093877,0.4098187983,0.1500536054,-0.3145959377,0.0825325772,-0.3006183505]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/259","title":"documentation missing how to split a dataset","comments":"Thanks a lot, the new explanation is very helpful!\r\n\r\nAbout using train_test_split from sklearn: I stumbled across the [same error message as this user ](https:\/\/github.com\/huggingface\/nlp\/issues\/147 )and thought it can't be used at the moment in this context. Will check it out again.\r\n\r\nOne of the problems is how to shuffle very large datasets, which don't fit into the memory. Well, one strategy could be shuffling data in sections. But in a case where the data is sorted by the labels you have to swap larger sections first. \r\n","body":"I am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n","comment_length":87,"text":"documentation missing how to split a dataset\nI am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n\nThanks a lot, the new explanation is very helpful!\r\n\r\nAbout using train_test_split from sklearn: I stumbled across the [same error message as this user ](https:\/\/github.com\/huggingface\/nlp\/issues\/147 )and thought it can't be used at the moment in this context. Will check it out again.\r\n\r\nOne of the problems is how to shuffle very large datasets, which don't fit into the memory. Well, one strategy could be shuffling data in sections. But in a case where the data is sorted by the labels you have to swap larger sections first. \r\n","embeddings":[0.0933188573,-0.0180116799,-0.0440782048,0.2046089619,0.018560512,0.2293842435,0.3818902373,0.3100673556,-0.052347783,0.0560588539,0.173167184,0.2897090316,-0.2534625828,0.14297086,0.1826965511,-0.3846262991,-0.040949177,0.1238843575,0.2132634968,0.0340434983,0.2280826271,-0.1566811204,-0.3243975341,0.0833063573,-0.3787745535,-0.4607850909,-0.2420907617,0.0113340588,-0.2463061064,-0.2269048244,0.2115312815,-0.2230210155,0.0325529687,0.3177694082,-0.0001165538,-0.015661221,0.2516828477,-0.1430504322,-0.1428460777,-0.1684978604,-0.3076057732,-0.0614499822,0.292765677,-0.1962689012,0.0164931845,-0.1657263041,0.1286057234,-0.0931775346,0.4861189127,0.4358699918,0.1081555262,-0.0751257911,0.1664645821,0.0756355226,0.1132426932,0.2819303274,-0.0977016091,0.1179105043,-0.17264162,-0.076216802,-0.2115267068,0.053548038,0.1823757142,0.1566736102,0.125684917,-0.0841683,-0.4509294033,-0.4551001191,-0.1296016276,0.4934631288,0.5426617265,-0.3339476287,-0.3045299351,-0.3151073456,0.0716482848,-0.2239682972,-0.1004446298,0.657910645,-0.0995268598,0.0326450169,-0.2173854858,-0.2442063391,-0.2979713082,0.1379865557,0.0440278985,0.4149711132,-0.0704656467,0.0622478761,0.3162668347,0.2178592235,0.2886422276,0.0521348789,-0.0544531345,0.2982503474,-0.2664626837,-0.3006373942,-0.1505006105,-0.3768862784,0.1879127175,0.2897794843,0.1368898749,-0.0967512727,-0.1370047331,0.0968484282,0.4555282891,0.2691028714,0.1787383109,0.5023200512,-0.0326183476,0.0236787945,-0.00661773,-0.1938532442,-0.1231800541,-0.2649686337,-0.0859762207,0.0528143607,-0.010835099,0.0401656032,-0.2710323036,-0.3356516957,-0.3671219349,-0.1159258932,0.1128808931,0.2140519619,0.0547791831,-0.0869436041,-0.3301884532,0.2288992703,-0.344464004,-0.0770404339,-0.1473353952,0.2890575528,-0.4430508316,0.1214073002,0.150044769,0.0211206302,0.3825264275,-0.1478670239,0.0332543068,-0.3363251984,0.1904426366,-0.0261317939,0.1668536216,0.3849906027,0.1337274015,0.0464870483,0.0343871415,0.2436407059,-0.3323981762,0.0222001988,-0.4431276619,-0.467299968,0.1060873047,0.0153613314,-0.0748935044,0.1824659258,-0.2412878424,0.2949691713,0.0049221776,-0.0577361956,0.1639326215,0.0135808978,-0.1646177173,-0.3053277135,0.2806352675,-0.0270866584,-0.336346209,-0.0258346461,-0.3241038024,0.0049534161,0.0451364666,0.2634805441,-0.3417490125,0.46156919,-0.117293179,0.3866520226,0.6255981326,-0.2665971518,-0.0927596316,0.2199219316,0.0737655759,-0.2708795667,-0.0650176406,-0.0674834177,0.3004568219,0.0994572267,-0.3013006449,0.4073327184,-0.272626102,-0.1013282463,0.1903397292,-0.270242542,0.2623762488,-0.0125599466,-0.1003315225,-0.2614190578,0.2502726316,0.4529341459,0.4291930497,0.0457557254,0.1852958649,0.2299852222,0.0436483957,0.1743118614,-0.1062085852,-0.2931696773,-0.3414640427,-0.0624880902,0.1257124543,-0.0733819604,0.0498063453,-0.2440433502,-0.2545617521,-0.1727221906,0.076438725,-0.3039550185,0.0500836223,-0.4049403071,0.3367552161,-0.092987597,-0.2854805291,0.0210399162,-0.3029925525,0.1937655807,-0.3741206229,0.1780481339,0.0287899673,-0.0663610101,0.1689425707,0.3435349464,-0.0526507124,-0.0593009628,0.0205595829,0.5714374185,0.5855073333,-0.1178481206,-0.1375153661,0.0156595893,0.4043174386,-0.4950317144,0.1544603109,-0.1134900749,-0.0153809879,0.0020598613,-0.5048499703,0.2872421145,-0.4481737614,0.3501088917,0.2593263388,0.1163748801,-0.0042462517,-0.1661787778,-0.205637753,0.018674748,-0.1871286333,-0.3557155132,0.1887671947,-0.0312693641,-0.1677328646,0.0217781235,0.5045206547,-0.2093385607,0.0627252385,-0.1132500172,-0.0044349278,-0.0686507598,0.1848542243,0.3737941384,0.1804282516,0.0757313445,0.3668141365,-0.1050656661,-0.2559290528,-0.0592306815,0.0512705743,0.2312475294,0.4707044065,-0.0124679701,-0.3036999106,-0.1205679551,0.0491354801,-0.0395382941,0.2080675513,0.1352888793,-0.2522885203,0.0577648953,-0.3520688117,-0.1349604428,-0.5467683077,-0.2239441872,-0.1322238892,-0.2780596614,-0.1199290305,-0.098386474,-0.0212559048,-0.0185306985,-0.1862570792,0.0012156031,-0.1477627754,0.1817387342,0.3582047522,-0.3515107036,-0.1967376322,0.0476123393,0.1667869687,0.6737962961,0.2871239781,0.0911054984,-0.1230676249,0.0287224967,-0.2473865002,0.1865918189,-0.1735794693,-0.0139636779,0.1336762905,0.1673673093,-0.0490347035,-0.3744676113,-0.1865420789,-0.2479310781,-0.187598899,0.0365710855,0.2715950906,0.1522685885,-0.1694388986,-0.7023664117,-0.2842538357,-0.2441640049,0.2758953869,-0.1915767938,0.1825142801,0.1525360197,0.1924121678,0.3374585509,-0.0148653081,-0.1065972894,0.2025753409,0.2754484117,0.004536326,-0.4193987846,-0.2563293576,0.2754090726,-0.020039903,0.1350158006,0.093058005,-0.1924740374,0.2536008358,-0.2302522957,-0.1527380794,-0.0305495001,0.0003086976,0.2978792787,-0.1919722706,0.268724978,-0.1498731524,0.1061584726,0.1293998361,-0.1358627826,0.2764899731,-0.0195109397,0.2527690232,0.2697938681,0.7959001064,0.2456320226,0.024959879,-0.3783942759,-0.0816765726,0.1254327744,0.0581409074,-0.0164634101,0.1688447297,-0.0930107757,-0.1885501593,0.0894003063,0.2449335307,-0.2968895733,0.0558524318,0.0394880064,-0.3108976781,-0.4277210832,0.3284148276,-0.3879098892,0.1899999976,-0.1148421019,0.0036380943,0.0747987479,-0.1614018083,-0.300783217,0.1252616197,0.2318617254,0.0932432115,-0.4762066901,-0.1768801659,-0.2314636707,-0.186152488,0.1672003269,0.2446663231,-0.0827640668,-0.0784750581,0.218990311,0.2506454885,0.2507941127,-0.0950242728,-0.0387736484,-0.0366838463,0.1550470591,-0.2871447504,0.1177817062,-0.3360717893,0.2984229028,0.0974624157,-0.0903820768,-0.438329488,-0.4115644097,0.4284511209,-0.0124993501,-0.1128587499,0.1444638669,-0.140851751,-0.1821876019,-0.0462232716,0.0968006328,0.3686917424,-0.0926734433,-0.1924466938,0.1678459197,-0.0647783577,-0.0318966992,0.1361120492,-0.0034860484,0.3259032667,0.0583311841,-0.1247309148,0.3844856322,0.1558671594,0.0751603246,0.3125393093,-0.1388070583,-0.1505511254,0.1816422641,-0.0805362538,0.5899174809,0.3286976218,0.1370225102,0.0263521951,-0.0942243785,-0.2090475112,-0.2700946331,0.3084840775,0.1049625054,-0.1736936718,-0.3126144111,-0.6472595334,0.332508266,0.1801181585,-0.0424645059,0.431342274,0.2609110773,-0.3862370551,0.4432201087,0.2118071467,0.7920416594,0.0127372295,0.1475440711,0.1756327003,-0.2705809176,0.6010873318,-0.4157494307,0.2180844992,-0.2315271497,-0.1738532335,-0.0496462435,-0.2688383162,-0.0339606293,0.4327793121,-0.0066562989,0.2928547859,-0.0640101805,0.0115997959,-0.2614720464,0.4319429696,-0.0934858024,-0.0670329407,0.0783171654,0.043817874,-0.135327965,-0.0102024814,0.0497062914,-0.0717561692,-0.2648791373,0.1049554646,-0.3984519541,0.0411282852,-0.1009622738,0.176490128,-0.122671403,-0.1409531385,0.3335616589,0.1761310548,0.0255631767,-0.0076105329,-0.0946107954,0.27425614,-0.1884091198,-0.1489420384,-0.0407044999,0.0572011694,0.4263288975,0.1537602395,-0.2795599997,0.0411295816,-0.196339801,-0.1461555958,-0.0403789431,0.3116209805,0.1542691737,-0.3274499774,-0.1701220423,0.1862521023,0.0458233617,-0.0053726905,0.0634926856,-0.0247492827,-0.177155897,0.1570520401,-0.0269990917,-0.1244152039,0.04412302,-0.1326964051,0.0476041548,0.079356119,0.3646090031,0.0717743859,-0.0371207781,-0.2247209549,0.2982898355,0.4597319067,-0.3562633693,-0.0078187836,-0.375995636,0.1422559172,0.1080947518,-0.1250568032,0.116729252,0.3218547702,-0.2142826766,-0.4522641003,0.0040523582,0.4776442051,-0.1817855388,0.2796476185,-0.2963930368,0.170242846,-0.0041099233,0.3981889784,-0.1884837449,0.2016786635,0.0994691402,0.355776608,0.0422501601,0.0918667912,0.1398098022,-0.125966683,-0.1074641794,0.0369800217,0.140354082,-0.1437069923,-0.0898947865,0.1539746672,-0.0461107716,0.0616742969,-0.2260435373,-0.325638622,-0.1745230854,-0.0723528266,0.0053287689,-0.0444331951,0.1831970513,0.4326807857,0.3581690788,0.1834308654,-0.1896140128,0.1988265216,-0.0962906256,0.2902747095,0.0351094604,0.2474222034,0.2076576501,0.1259331554,-0.265509516,-0.0089306869,0.2336477339,-0.3311530054,0.0000617048,0.1874328852,0.2726083398,-0.4687635005,0.3732601106,0.4829464555,0.1707503647,0.0543365292,0.2456143051,0.0658320114,0.1364284605,0.1967912316,0.260305196,0.1590285599,-0.0102036987,0.3061126769,0.0887353569,0.512585938,-0.4226796031,-0.0317303538,0.1596478075,0.0053547332,0.3506353199,0.4653632045,-0.2353554368,-0.1776273251,0.2543795407,-0.3389159739,0.3259781301,0.2159130275,0.165623188,0.6236848831,0.0447951183,0.0662975311,0.3561085463,-0.1359663606,-0.5598917007,0.1638800204,0.086286895,-0.0507289097,-0.2131101489,0.6076283455,-0.0591780171,-0.2103677392,-0.0459646359,0.1484448612,0.0535346679,0.0184057131,-0.3480505049,-0.1128064543,-0.2202858627,-0.035500817,0.1682533026,-0.3167338073,0.3955667019,0.0233528856,0.0506865084,-0.0778124705,0.3185546994,0.3223921657,0.1612419784,-0.0516931191,-0.0384987742,-0.1833804101,-0.1202984601,0.0549172685,0.1042335108,0.1393318623,-0.0080134263,-0.3288946152,0.0258013941,0.2802728713,-0.1753753573,0.2165521681,0.3208757937,0.0261304043,-0.046777904,0.1568453759,0.0937315226,-0.0506097898,-0.047944136,0.1992452443,0.0109777777,-0.3903980851,0.3062102199,-0.0560246967,-0.1819999814,-0.0871342868,0.0212073568,-0.2622387111,-0.121505931,0.3463400006,-0.2991252244,0.0790508017,-0.1613875479,0.0048944689,-0.0572145358,0.6285125613,0.1339647323,0.220652163,-0.2274456918,-0.2880062163,-0.3233340681,-0.1600948423,-0.3285861015,-0.1975018531,0.1390625089,0.3364286423,-0.2651438415,0.171046555,-0.1804319471,0.0804285184,0.1338137239,0.2941442728,-0.1228971779,-0.2238142639,-0.258177489,0.1429474503,-0.1118860841,-0.278623879,0.0020922816,0.0149588482,-0.2938500345,0.191971451,-0.0257433821,0.4331857264,0.4787186384,-0.210207507,0.1381002814,0.5317716002,0.2446607947,0.0373651572,-0.3966127932,0.0103107383,-0.3066258729,0.3514079452,-0.058364179,0.2431861609,-0.3306998014,0.0308478847,-0.1717275381,0.1854444295,0.0223965086,-0.1288520694,-0.3340089619,-0.0516488962,-0.349940002,0.224088788,0.7802200913,0.4724250436,-0.3136300147,-0.0546364151,0.1696928889,-0.099482432,-0.0810545534,-0.4101920724,-0.1743736118,0.1650438458,0.1339748204,0.3939222395,-0.140280515,-0.7209003568,-0.1638847142,0.1605288982,-0.1027985066,0.0948313326,0.0975646153,0.2638238966,-0.1234307885,0.000655745,0.4191991389,0.07783968,-0.3093946576,0.0566997826,-0.3002833724]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/259","title":"documentation missing how to split a dataset","comments":"We added a way to shuffle datasets (shuffle the indices and then reorder to make a new dataset).\r\nYou can do `shuffled_dset = dataset.shuffle(seed=my_seed)`. It shuffles the whole dataset.\r\nThere is also `dataset.train_test_split()` which if very handy (with the same signature as sklearn).\r\n\r\nClosing this issue as we added the docs for splits and tools to split datasets. Thanks again for your feedback !","body":"I am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n","comment_length":64,"text":"documentation missing how to split a dataset\nI am trying to understand how to split a dataset ( as arrow_dataset). \r\nI know I can do something like this to access a split which is already in the original dataset : \r\n\r\n`ds_test = nlp.load_dataset('imdb, split='test') `\r\n\r\nBut how can I split ds_test into a test and a validation set (without reading the data into memory and keeping the arrow_dataset as container)?\r\nI guess it has something to do with the module split :-) but there is no real documentation in the code but only a reference to a longer description: \r\n\r\n> See the [guide on splits](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/docs\/splits.md) for more information.\r\n\r\nBut the guide seems to be missing.\r\n\r\nTo clarify: I know that this has been modelled after the dataset of tensorflow and that some of the documentation there can be used [like this one](https:\/\/www.tensorflow.org\/datasets\/splits). But to come back to the example above: I cannot simply split the testset doing this: \r\n`ds_test = nlp.load_dataset('imdb, split='test'[:5000]) `\r\n`ds_val = nlp.load_dataset('imdb, split='test'[5000:])`\r\n\r\nbecause the imdb test data is sorted by class (probably not a good idea anyway)\r\n\nWe added a way to shuffle datasets (shuffle the indices and then reorder to make a new dataset).\r\nYou can do `shuffled_dset = dataset.shuffle(seed=my_seed)`. It shuffles the whole dataset.\r\nThere is also `dataset.train_test_split()` which if very handy (with the same signature as sklearn).\r\n\r\nClosing this issue as we added the docs for splits and tools to split datasets. Thanks again for your feedback !","embeddings":[0.0403223671,0.0286834165,-0.0598337613,0.1646729261,-0.0391383842,0.2028904557,0.4111513495,0.3333023489,-0.0838269815,0.0418192372,0.1581265479,0.3008451462,-0.1978239268,0.1513804793,0.1966464221,-0.3646208346,-0.0580054745,0.1984985769,0.2289592177,-0.0036904679,0.256201148,-0.1071294993,-0.304423362,0.120927833,-0.3231808543,-0.4618531764,-0.2295882851,-0.0311393272,-0.2734042108,-0.1920207292,0.2099711448,-0.1823351085,0.0320916846,0.3054240644,-0.000116105,-0.010881043,0.2465662807,-0.1087708473,-0.1147388145,-0.2236365676,-0.3785127699,-0.030364342,0.3681846559,-0.1375394017,0.0056159082,-0.2517320812,0.1225903183,-0.1782120317,0.5087322593,0.4407333136,0.1115595773,0.0035519912,0.1767529845,0.1035160571,0.042402748,0.3595411777,-0.1065701693,0.0957077816,-0.1545086503,-0.1049503535,-0.1268551499,0.0406020321,0.1666917205,0.1925523877,0.1087984517,-0.035980694,-0.3644660413,-0.4900597334,-0.1254992932,0.5456266999,0.6005104184,-0.3356675804,-0.3338814378,-0.2798808217,0.0761559084,-0.228790462,-0.1258219928,0.6406729817,-0.1184810773,0.0291664936,-0.1475769877,-0.2527887821,-0.2896288335,0.139794454,0.0476697795,0.432497412,-0.0843734443,0.0583464317,0.2639944255,0.2229848951,0.3153964579,-0.0034207574,-0.081353724,0.292568326,-0.2424946725,-0.292462647,-0.1715589464,-0.3974406421,0.132484138,0.2986210585,0.1361417919,-0.0838172808,-0.1949217319,0.1345125586,0.4678732455,0.2058346272,0.2396071851,0.4790492654,-0.0016800514,-0.0263343584,0.0180896092,-0.1862372309,-0.0208235234,-0.2293924093,-0.029344935,0.0750173256,0.1216978952,0.0660451204,-0.2564261556,-0.3403445184,-0.3157760501,-0.1848141253,0.1066501439,0.132505253,0.0828250721,-0.080006741,-0.3198460042,0.2792649269,-0.3513965607,-0.1381861866,-0.1297213584,0.2163873166,-0.4161950052,0.1454454958,0.1128338203,-0.0038959899,0.3922364712,-0.083182998,0.0201491434,-0.3755839169,0.2846872807,0.0171303116,0.1891864538,0.3779692054,0.1809567511,0.0440621153,0.0598460361,0.1318244934,-0.3288018405,-0.0048243543,-0.3971320987,-0.4248842299,-0.0321050212,0.0282722227,-0.0414998494,0.1902054697,-0.2672340572,0.351020664,-0.0028942707,-0.0546868294,0.2016918063,0.0277158506,-0.1759249419,-0.3253219426,0.2509885728,0.0148803042,-0.3857682347,-0.0391694866,-0.3769051731,-0.0401922204,-0.0184912179,0.1650433242,-0.3774425983,0.4833278358,-0.1127269715,0.4063957334,0.6908680201,-0.1989747882,-0.0340800621,0.2310912609,0.1239069849,-0.2687203884,-0.1132232249,-0.0545059368,0.3031643033,0.116834797,-0.4216461778,0.3843097985,-0.3077020049,-0.1268476248,0.2832005322,-0.254827857,0.2440154254,-0.0654645637,-0.0825421885,-0.254324168,0.2538763881,0.3207102418,0.4920059144,0.0280294158,0.1801446527,0.1909128726,0.061658036,0.2179094106,-0.0664332882,-0.3025019169,-0.3309570253,-0.0793043599,0.1560174525,-0.071212098,0.0058269701,-0.2646156847,-0.2066113055,-0.172613129,0.0351849459,-0.3387405574,0.0594701134,-0.3565586805,0.2210893333,-0.1259914637,-0.4023621976,0.0217912626,-0.2359521091,0.242100656,-0.4350490868,0.1747662723,0.0647162572,-0.0816771388,0.1286371648,0.2814127803,-0.0714057758,-0.1138372868,-0.0043459455,0.5958948135,0.5878075361,-0.0264427308,-0.1346909702,0.0432598852,0.4055946469,-0.5084712505,0.2129133195,-0.0599845313,0.0160070825,0.0081074769,-0.4583743513,0.3185171783,-0.4548642933,0.3451863825,0.2045084238,0.1280926019,0.0154547524,-0.1400159001,-0.2886721492,0.02641619,-0.2369019687,-0.3842772245,0.1804403812,-0.048252888,-0.16476129,0.062357232,0.4256391227,-0.2152217478,0.1237346753,-0.1150745898,0.0095078796,-0.0433240235,0.1387760043,0.4403604865,0.2106843144,0.0526768081,0.3590402603,-0.0993303731,-0.248431623,-0.0596648753,0.1357475668,0.2258344591,0.4377182126,-0.0557044819,-0.4013492167,-0.0979817435,0.0026900461,-0.0214847047,0.1622288078,0.0848039016,-0.2820232809,0.0794488713,-0.3813348711,-0.1499004364,-0.5209005475,-0.2775146663,-0.1069980189,-0.2646991014,-0.1675828397,-0.0983399972,0.0030451696,-0.026531009,-0.1790220439,-0.0568118766,-0.1716521382,0.2067324072,0.3574621975,-0.3144453168,-0.1490798295,0.0507454239,0.2032431066,0.6589824557,0.2827355862,0.053942617,-0.1018270403,0.0052011949,-0.2792691588,0.2223029882,-0.1885470599,-0.0103698988,0.1580085605,0.1867484599,0.0028782091,-0.3813948333,-0.2566469312,-0.2440122217,-0.1778089404,-0.0182332359,0.2075580806,0.0996807292,-0.1586859822,-0.729760766,-0.2971604168,-0.1867208928,0.2442190349,-0.2553653121,0.2198871225,0.201742053,0.1517691016,0.3269978762,0.0214531217,-0.065773651,0.2023561746,0.2355849892,-0.000492701,-0.4409202039,-0.1999350041,0.3105662465,-0.0308506079,0.1985801458,0.0315750167,-0.1794963181,0.2157322764,-0.1878322512,-0.0676265657,-0.0577653721,-0.0059105507,0.2933513224,-0.1467363834,0.2593975365,-0.1576316953,0.0894758999,0.1173988357,-0.1268907636,0.2416574806,0.0256769899,0.3007251918,0.2090368569,0.7162435055,0.2169198394,-0.0496118627,-0.3078827262,-0.1198214665,0.1227457449,0.0832005069,-0.0693561584,0.0601780042,-0.0565507971,-0.1230469421,0.0571317002,0.2224975079,-0.3131632805,0.0492444821,0.0198015776,-0.3851720095,-0.4363874495,0.2914071381,-0.4392451644,0.1846540272,-0.0833713189,0.0155400643,0.0990051329,-0.0882876292,-0.3387968242,0.178387776,0.2523909807,0.0675498545,-0.3932358027,-0.1368214786,-0.1724320501,-0.1602956206,0.1363639385,0.2237350196,-0.0813033208,-0.0105788652,0.2096960992,0.2643876374,0.2860427499,-0.1441953927,-0.0868417919,-0.0686072335,0.1527087986,-0.319350034,0.1082289368,-0.2728942931,0.3731343448,0.0478409566,-0.0624198429,-0.3523035645,-0.4623895586,0.4459838569,-0.0225010663,-0.1089765579,0.131750226,-0.1218779758,-0.2083639652,0.0165919233,0.071805872,0.3153065145,-0.1571342796,-0.1344532073,0.1100082025,-0.0813710988,-0.0563573651,0.1561253816,-0.0222922526,0.3218558729,0.0912841111,-0.1056616977,0.3978521824,0.0925958306,0.0204186961,0.2542870939,-0.1303016543,-0.1756850332,0.1556497067,-0.1278392524,0.5831713676,0.283754617,0.1274223179,0.0390483737,-0.1054624692,-0.2130824178,-0.2523541152,0.2407673895,0.2048507631,-0.1252409518,-0.2370994091,-0.6877892613,0.3460371196,0.1299509108,-0.0463190563,0.5394695997,0.2873894572,-0.3330204189,0.4225696325,0.201598078,0.7477269769,0.0762851015,0.0732946917,0.1754496545,-0.329533577,0.568672657,-0.3917847872,0.1484681368,-0.2093200982,-0.1846489608,-0.1217485517,-0.2601589859,-0.0024031848,0.3796321154,-0.0206537507,0.3090942204,-0.1220182106,-0.0388306975,-0.1852693409,0.4369939864,-0.157868281,-0.0386330336,0.1171749234,0.0497081205,-0.1701854914,-0.031542141,0.0686156675,-0.0970599204,-0.2383664101,0.0897887051,-0.4137589335,-0.0039857212,-0.0936348215,0.1743156165,-0.1165762767,-0.1285838932,0.3310354054,0.2014929205,0.0769494101,-0.0470001735,-0.0862478316,0.2644329965,-0.1550586373,-0.0802347437,-0.1183241382,0.007898028,0.4244668186,0.1777239144,-0.3372201324,-0.0016750883,-0.1474841237,-0.1404349655,0.0317053832,0.3100647926,0.2073042244,-0.3079301417,-0.0597708188,0.2092687637,0.0176890623,0.0300950184,0.0701070279,0.0323982723,-0.1881749481,0.2046191245,-0.0057176072,-0.1675052196,0.0380118489,-0.1753720194,0.0619026385,0.1214928105,0.3558523953,0.0676505193,-0.0374215804,-0.235719949,0.3150270283,0.4389844835,-0.3237165511,-0.027406374,-0.4423149526,0.1874001622,0.1253643483,-0.2000238746,0.0556186028,0.377355963,-0.2882798314,-0.4561295211,0.0860733911,0.3866676092,-0.1715884954,0.3158802688,-0.2676703036,0.098412998,0.0552106425,0.3880164623,-0.1836767644,0.2403481007,0.2049387097,0.3693642616,0.0657507032,0.0962123573,0.1903946102,-0.1657320261,-0.0842860416,0.0015350453,0.1375169754,-0.1580260098,-0.1233465075,0.1513304114,-0.0974080637,0.0174096636,-0.1899469793,-0.2951283157,-0.231850177,-0.0545293242,-0.0332034305,-0.0677813217,0.1707528532,0.4092100859,0.3738699555,0.1342544407,-0.2121079862,0.12073569,-0.0866125003,0.3235029578,0.0302814003,0.2924144566,0.1557211429,0.0796316415,-0.2641418874,0.0043117581,0.2512800395,-0.3833703697,0.0201570615,0.2284638435,0.2473916709,-0.5429638624,0.4029043019,0.4616820514,0.1510469317,0.125165686,0.1613737345,0.0737377629,0.1638292819,0.1937484145,0.2226104289,0.2313716561,-0.0253922231,0.2896288931,0.0856892616,0.503330946,-0.399338752,-0.066138953,0.1873229742,-0.0252371207,0.3962180614,0.4388950765,-0.2788352072,-0.1600986868,0.2630686462,-0.3144689202,0.2578749061,0.1524720192,0.097441189,0.5958568454,0.0711128414,0.014083446,0.4374821186,-0.048614867,-0.5118055344,0.1202183217,0.0819862336,-0.105982095,-0.0782754347,0.6175237894,-0.0525770672,-0.2081803679,-0.0492200442,0.1359009892,0.0903205797,0.0175968166,-0.4294456244,-0.052839689,-0.2667381763,-0.0893903151,0.1585074663,-0.2950607836,0.3204170167,0.0735074803,0.0717599988,-0.077145122,0.3561198711,0.2811409831,0.1060580984,-0.0346498489,-0.0461558811,-0.1354419589,-0.094269596,-0.0097735496,0.2205352336,0.1571876556,-0.0435949638,-0.357588172,0.1080478281,0.2932514548,-0.1990911961,0.2499285936,0.2395002991,-0.0611463673,0.0016486547,0.1325111687,0.1182436198,-0.0238555577,-0.0155284088,0.1763847768,0.0615801029,-0.4181664586,0.3436127901,-0.0968577042,-0.1469762325,-0.1198127791,0.0243319273,-0.2717857063,-0.1448500752,0.4898412228,-0.2937377989,0.0498909838,-0.1740180701,0.0115811881,-0.0737591311,0.575776577,0.1035181209,0.2875744998,-0.1891214401,-0.2778648436,-0.3347233236,-0.1771017164,-0.2444074005,-0.1643835455,0.115304403,0.3453474641,-0.1990485489,0.1723533571,-0.0962464809,0.0553650707,0.1245097592,0.3176043332,-0.0720609426,-0.2481213361,-0.2612348795,0.1950792968,-0.1144715399,-0.2309558243,-0.0003361014,-0.0747895539,-0.2715563178,0.1687706113,-0.0619911253,0.5020113587,0.4536445737,-0.1745002419,0.1170869246,0.5048570037,0.2103504986,0.0413013063,-0.375962615,0.0523339622,-0.3360321224,0.4153881967,-0.0540983193,0.2203135192,-0.3131210804,0.0194664765,-0.1805739701,0.1590904146,-0.0025377299,-0.1574306488,-0.3457916975,-0.0679126382,-0.3383202851,0.1855581105,0.75955832,0.4215837419,-0.3276263177,-0.0323504731,0.1678061634,-0.0936116502,-0.0749757439,-0.3846354783,-0.1207671389,0.1181346476,0.1582869142,0.3857811093,-0.0868676975,-0.7019298077,-0.1841271967,0.1819613278,-0.1854793131,0.0895234421,0.0927058086,0.357991457,-0.0905299708,-0.0455036275,0.3920331597,0.1428621262,-0.2787569165,0.0489722155,-0.2914073765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/258","title":"Why is dataset after tokenization far more larger than the orginal one ?","comments":"Hi ! This is because `.map` added the new column `input_ids` to the dataset, and so all the other columns were kept. Therefore the dataset size increased a lot.\r\n If you want to only keep the `input_ids` column, you can stash the other ones by specifying `remove_columns=[\"title\", \"text\"]` in the arguments of `.map`","body":"I tokenize wiki dataset by `map` and cache the results.\r\n```\r\ndef tokenize_tfm(example):\r\n example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text']))\r\n return example\r\nwiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train']\r\nwiki.map(tokenize_tfm, cache_file_name=cache_dir\/\"wikipedia\/20200501.en\/1.0.0\/tokenized_wiki.arrow\")\r\n```\r\nand when I see their size\r\n```\r\nls -l --block-size=M\r\n17460M wikipedia-train.arrow\r\n47511M tokenized_wiki.arrow\r\n```\r\nThe tokenized one is over 2x size of original one.\r\nIs there something I did wrong ?","comment_length":53,"text":"Why is dataset after tokenization far more larger than the orginal one ?\nI tokenize wiki dataset by `map` and cache the results.\r\n```\r\ndef tokenize_tfm(example):\r\n example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text']))\r\n return example\r\nwiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train']\r\nwiki.map(tokenize_tfm, cache_file_name=cache_dir\/\"wikipedia\/20200501.en\/1.0.0\/tokenized_wiki.arrow\")\r\n```\r\nand when I see their size\r\n```\r\nls -l --block-size=M\r\n17460M wikipedia-train.arrow\r\n47511M tokenized_wiki.arrow\r\n```\r\nThe tokenized one is over 2x size of original one.\r\nIs there something I did wrong ?\nHi ! This is because `.map` added the new column `input_ids` to the dataset, and so all the other columns were kept. Therefore the dataset size increased a lot.\r\n If you want to only keep the `input_ids` column, you can stash the other ones by specifying `remove_columns=[\"title\", \"text\"]` in the arguments of `.map`","embeddings":[-0.05911934,-0.0847912058,0.051377032,0.0893474296,0.0501229502,0.0617089607,0.3212458491,0.4181883931,-0.0478091687,-0.0994747803,-0.1050057113,0.2905612588,0.0789179951,-0.1424425691,0.2202340662,-0.0885483623,0.3397546709,0.0273732729,0.1782818288,-0.2089830488,-0.0077200597,0.0045569316,-0.1589310467,0.0408784822,-0.4096499681,-0.0649461299,-0.193261385,-0.1158594117,-0.3763851821,-0.4542703629,0.0177729093,0.1124850512,0.4515711367,0.2357399762,-0.0001270274,-0.3008657098,0.015436911,-0.0012035987,0.0112030348,0.3585988581,0.1909619868,-0.5831577182,-0.1245546415,-0.2549140751,0.2653027773,-0.5587732196,-0.2960033119,0.2278251201,-0.010379578,-0.1184156165,0.0964032114,-0.1408597231,-0.0603276119,0.4277178347,0.0949645415,0.0258308407,-0.0870460644,-0.2361410111,-0.1481280178,-0.2496973127,-0.0193840619,0.2022605985,-0.0480351225,-0.2493420541,0.3461992145,-0.0334121771,-0.1799578965,-0.0225804951,0.3381353915,0.2231501788,0.3686365187,-0.0917311534,0.076419957,-0.2663859427,-0.1512706429,-0.1332911998,-0.1526265293,0.1295672655,0.3435021639,-0.0665261224,-0.3992856145,-0.0862823725,0.2093705386,0.2339636832,-0.2337776572,0.511779964,0.0349198878,0.2241724879,-0.2702212036,-0.1453222781,-0.1040053368,-0.1828158945,-0.1459912211,0.4562894702,-0.3604279459,-0.3879573643,0.0803292021,-0.0779577717,0.43373245,-0.2686167061,-0.3836771846,0.0286895335,-0.006413355,-0.2013777941,0.1444687098,0.5048397183,-0.2656995952,0.4378589392,-0.0957537144,-0.1229636818,-0.4596572816,-0.1897007823,0.0471173003,0.170946613,0.2345881313,-0.2991546988,-0.2995208204,-0.0388312005,0.2762614191,0.2337453514,-0.4753175676,0.0576496571,-0.1413362622,0.3265320361,-0.127571702,0.3444333375,-0.386506021,0.1868578345,-0.1330244541,0.0153495744,-0.1350686401,-0.1800366789,0.0989733934,0.1725703478,-0.2482747436,0.2382375896,0.3281407058,0.1204317734,-0.1469759494,-0.0860265866,0.4267893732,-0.1727076918,0.3936494887,0.144278273,0.1197035611,0.3475687504,-0.044130329,-0.1256456822,-0.197924763,0.1588634402,-0.5327580571,-0.1950257421,-0.0560805872,0.0217503421,0.005869606,-0.0263538267,-0.3628127277,0.3748996556,0.4107330441,-0.0662525371,0.175813511,-0.3005610108,-0.3777944446,-0.2144777477,-0.0133280465,-0.078125149,-0.1085254103,-0.0072020669,0.263325721,0.322047174,0.3666019142,0.6014493108,-0.225136742,0.506583333,-0.2451561689,0.3823338449,0.4053663313,-0.302113682,-0.737957418,0.3207946122,0.1591732204,0.2684880495,-0.1612892449,-0.0188451856,0.5017424226,0.0920706019,0.0527848154,0.2601010799,-0.0502847172,0.1213388517,-0.3101269603,-0.1889876276,0.3550931513,-0.5069997311,0.1982344687,0.0577119961,0.0366838276,0.4232871532,0.4101299047,0.0688466355,0.2923918664,0.2346351296,-0.0228509344,0.0889867097,0.2690239251,0.0733008012,0.0940393806,-0.0858660266,-0.080476515,-0.1107840985,0.1996038407,-0.4010845423,-0.1931902617,-0.0662963241,-0.0975160226,-0.3066636622,0.0048550181,-0.1004267931,0.1177812442,-0.044129245,0.211798504,0.1014265195,-0.0825838149,0.0912325084,0.0964552015,0.1563900709,0.1161217466,0.1823627055,-0.2617761493,0.2670944631,0.1819374114,-0.0376796611,-0.2108133137,0.0918716937,0.1051934883,0.0365116484,0.1522092521,0.3906073868,0.1717032492,0.0953818485,0.4608205855,-0.0380945206,0.0117277429,-0.1451399326,0.1543003917,0.3306969404,0.3767090738,0.1166266575,0.0138118789,-0.2462913096,0.1822610199,0.019084258,0.1345324069,-0.2554041445,-0.2080486119,0.2156358063,-0.0192125496,0.0240642168,0.1125004366,0.3427215219,0.4778441787,-0.1071349904,-0.1353248805,0.1246572733,-0.6855455637,-0.5055297613,0.157845512,0.0531001315,0.2458252013,0.0988888517,0.2933290005,0.0277646184,0.195890516,0.1115763932,0.1851910204,-0.0681385547,-0.0098555377,0.1779201031,0.4578203857,0.1271985471,-0.2502453029,0.251022011,0.0222988445,0.4215033948,-0.0949088484,0.1341169924,-0.218585372,-0.0506426878,0.0137004675,-0.0659863204,-0.1964899451,-0.431121707,-0.0915186331,-0.0314840563,0.3101437688,-0.1640016735,-0.1009414271,0.0141453203,0.1298680753,0.1178558767,0.130728513,-0.0481383689,-0.3677668869,-0.0437394977,-0.0823242143,-0.0946425796,0.2671999931,0.1329374313,-0.0695441887,-0.4135707319,-0.7815631628,0.0633474886,-0.1783969253,0.0664602593,-0.0457357951,0.0075658523,-0.4851557016,0.16965051,-0.0277932808,0.0267434381,-0.1989867538,-0.0523474142,-0.3482215106,0.1279136091,0.1593476236,-0.0251743756,-0.0163428914,0.0750915855,-0.0367845483,0.0466038659,0.0685766712,0.0380102061,-0.2096304893,-0.101762116,-0.0553878732,-0.1567961127,-0.5051301122,-0.2839381099,0.2793945968,-0.1403146088,-0.2329826802,-0.0317146927,-0.1785343289,0.1400174946,0.087827079,-0.5514478683,0.1583929509,-0.2124280185,0.1916601956,-0.2602857351,0.0172649939,0.1755646616,-0.1336281598,0.0733250827,0.0045240242,0.0095663751,-0.3130403459,0.1630244404,0.4565995336,-0.0777926967,0.4733356535,0.2586420774,0.8525633812,0.0157968346,0.0328100882,0.213544637,-0.2784926891,0.1436034739,-0.2989118993,-0.1292733997,0.1617760062,-0.1741361618,0.1807521731,0.4305220544,0.1738902479,-0.1181302443,0.6388402581,-0.2887889445,0.1047127768,-0.0959652141,0.2121330649,-0.2236872017,0.2641284168,0.2335649431,-0.1642081141,-0.6363502741,-0.5240095854,0.0157368146,0.1268444359,-0.0615629703,0.1331123114,-0.7828913927,-0.0653619319,-0.6774397492,0.2748371959,0.098265484,0.2737056315,0.3563112617,0.040792942,0.1143241078,0.0431324616,0.5410208106,-0.1418653429,-0.2233980298,-0.0854783654,-0.0231055152,-0.2765070498,-0.2000237703,-0.00210763,0.2203007787,0.4453181028,0.6754165292,0.0222675689,-0.1887315959,-0.0347929299,0.5779742599,-0.1647357047,-0.0580121391,-0.3751465082,-0.0498849489,-0.3190487623,0.0173757933,-0.2342173606,0.2088268548,0.0142801292,-0.2663803399,-0.2042874992,-0.2549265921,-0.0957637727,0.113132827,0.3051750958,0.3400226831,0.0601958111,-0.2264736742,0.2353183478,0.0076809037,-0.045271121,0.1082853451,0.069379203,-0.14581348,0.0627233088,0.3206310272,0.0908300653,0.0188077521,0.5403699279,-0.3301042914,0.145332098,-0.0838665888,0.4072865844,0.3613957465,0.1180799529,-0.451510787,-0.3181846142,0.1857718378,0.1650397032,-0.0691632703,0.3194537759,0.0845697969,-0.1370802671,0.4602780044,0.0660700276,0.9790534973,-0.1060549915,0.2866687775,0.0000893633,-0.003614259,0.3213832378,-0.5782904029,0.0396052636,-0.2436259687,0.1978450269,0.0367444679,0.0739025474,-0.02606773,0.1841714978,0.0307519566,0.0558932126,0.0134664262,0.4594786763,-0.3787733018,0.4584659338,0.1293183863,-0.0789743215,0.0771431252,0.0161333662,-0.1007995158,-0.2620897889,-0.176332742,0.2034144551,-0.1437988877,-0.2942550778,-0.0245791767,-0.2759604454,-0.0136327101,0.3781241775,-0.0393584333,-0.2465340346,0.0242826771,0.0788047835,0.123532854,0.471044898,0.0196333416,0.1571984887,0.0173716079,0.2313854247,0.20039092,-0.1912413687,-0.0178618915,0.0411435999,-0.2371105403,-0.0614400133,-0.0381814949,-0.3024937212,-0.3547362983,0.1559073776,0.5879484415,-0.3762474954,-0.1071729362,0.0762383267,-0.0869837552,-0.1380647868,0.0369467959,-0.1718323678,0.0611570664,0.2732652724,0.2821738422,-0.4257617593,-0.0756398365,0.3695720434,0.3958668709,0.0705835149,0.7585434318,-0.2851180434,-0.1089349464,-0.0336679928,0.2628870904,0.1575577259,0.1521796137,-0.2588036656,-0.0959916413,0.1300073266,0.046134688,0.0234391782,-0.1975012273,-0.0054345229,-0.2912546992,-0.2011929154,-0.560105145,0.0312362332,-0.3721835911,0.1879945546,0.2716309428,0.1745072454,-0.3227248192,0.5348668694,-0.1382447481,0.07519871,0.1726842523,0.1798041165,-0.196737498,0.0283319727,-0.0926893502,-0.032315895,-0.0569525585,0.1557477117,0.0888282061,-0.1387358457,-0.354619503,0.1914497614,0.2820284963,-0.1457911283,-0.0635601059,0.0200265627,-0.1527117044,-0.377066642,0.2815602422,-0.0416297503,0.1119715497,0.1002023667,0.2932283282,-0.2668458521,-0.0689943805,-0.028105706,0.1956409961,-0.0062503819,0.092851527,0.1738091856,0.2828683555,-0.0799672008,-0.4542538524,0.0702262968,0.2185202837,0.0287319552,0.4231923819,0.1460666209,0.1063066348,0.0702401325,-0.2505934536,0.161865145,0.0635208338,-0.4121688306,0.3658710718,0.0093763052,-0.060158249,-0.1242049113,0.3208639324,-0.1223801002,-0.1352737099,0.0048252353,0.4468995929,0.2352954149,-0.4104179442,-0.0677906126,0.409317106,-0.1909317374,-0.044676505,0.2629776001,-0.0758305788,0.078523092,0.2926717699,0.0625883713,0.0427169763,0.3937374055,0.3723279238,0.2766469121,0.3065327108,0.0783282965,0.025759276,-0.2941976786,-0.1397879422,0.3273918331,-0.1583966464,0.0905851573,0.1175371408,0.074456118,0.556946218,-0.3510777056,0.0387104824,0.4113403559,-0.3115716279,-0.222187072,-0.0707549974,-0.0338111632,0.081818074,0.109292917,-0.114134714,-0.2129836679,-0.0952965021,-0.1276652515,-0.2052709609,-0.2059244663,0.1625533253,0.1398721635,0.1289059371,-0.2399079353,0.0178099666,-0.3468590081,0.1177413166,-0.1573513299,0.113294214,0.279445976,0.2193481922,-0.0472948775,-0.0211972073,0.1454105079,0.0477933735,-0.0116612362,0.0766381249,0.173053205,-0.4105388224,-0.0847634152,-0.0995359272,-0.0577941798,-0.3521797955,0.3204970658,-0.0887681395,0.0776847601,0.1781221926,-0.0660474151,-0.2564237416,-0.104678221,0.1899688095,-0.370785594,0.1165261939,0.3587572277,0.034652546,0.3313942552,-0.3841510713,0.026846014,-0.323053211,0.3820792735,0.560506165,-0.3645080924,-0.2622018158,-0.2813616395,-0.2082095742,0.2300589085,-0.4183411002,-0.0099505913,-0.1728505492,0.2826250494,-0.1216621697,-0.3347283006,-0.0016157256,-0.1150317565,-0.0774231628,0.121335797,-0.0167385731,-0.0171659328,-0.3584667742,0.1542277932,-0.2032237053,-0.0423572101,0.2969509065,0.1775441021,-0.1146416441,-0.1453517228,0.0886183754,-0.0143000241,-0.0489379875,0.3547165096,0.2388485074,0.335685581,-0.064812839,0.2596982419,0.0184217077,0.0376023911,-0.1106703281,0.2143891901,-0.253337115,0.0371871814,0.1052204296,-0.0239919201,0.2082184404,0.015258993,0.0650137067,-0.2231423557,-0.1852751225,0.160099566,-0.1126485988,0.0329601392,-0.1247100532,0.5538580418,0.0568277687,0.2275034189,0.0008597392,-0.2433762699,0.2034361213,-0.5507060289,-0.2111866325,-0.2141315937,0.003544658,0.254959017,-0.2977746725,-0.3130095899,-0.0792184919,0.516102314,-0.0107197063,-0.0368984789,0.2389673889,0.0862852335,0.18690148,-0.1439185739,0.3211912513,-0.2282097936,0.0088535165,0.0608959496,-0.0403908417]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/258","title":"Why is dataset after tokenization far more larger than the orginal one ?","comments":"Hi ! Thanks for your reply.\r\n\r\nBut since size of `input_ids` < size of `text`, I am wondering why\r\nsize of `input_ids` + `text` > 2x the size of `text` \ud83e\udd14","body":"I tokenize wiki dataset by `map` and cache the results.\r\n```\r\ndef tokenize_tfm(example):\r\n example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text']))\r\n return example\r\nwiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train']\r\nwiki.map(tokenize_tfm, cache_file_name=cache_dir\/\"wikipedia\/20200501.en\/1.0.0\/tokenized_wiki.arrow\")\r\n```\r\nand when I see their size\r\n```\r\nls -l --block-size=M\r\n17460M wikipedia-train.arrow\r\n47511M tokenized_wiki.arrow\r\n```\r\nThe tokenized one is over 2x size of original one.\r\nIs there something I did wrong ?","comment_length":31,"text":"Why is dataset after tokenization far more larger than the orginal one ?\nI tokenize wiki dataset by `map` and cache the results.\r\n```\r\ndef tokenize_tfm(example):\r\n example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text']))\r\n return example\r\nwiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train']\r\nwiki.map(tokenize_tfm, cache_file_name=cache_dir\/\"wikipedia\/20200501.en\/1.0.0\/tokenized_wiki.arrow\")\r\n```\r\nand when I see their size\r\n```\r\nls -l --block-size=M\r\n17460M wikipedia-train.arrow\r\n47511M tokenized_wiki.arrow\r\n```\r\nThe tokenized one is over 2x size of original one.\r\nIs there something I did wrong ?\nHi ! Thanks for your reply.\r\n\r\nBut since size of `input_ids` < size of `text`, I am wondering why\r\nsize of `input_ids` + `text` > 2x the size of `text` \ud83e\udd14","embeddings":[-0.0532691069,-0.2086362988,0.0178327393,0.1911770701,0.0143836392,0.0087456573,0.2398011088,0.3740540445,-0.1149224043,-0.0138116116,-0.080173701,0.180669263,0.1408331692,-0.1129715592,0.1978251338,-0.1090740561,0.2893790305,-0.0118231466,0.1726615727,-0.1984837353,0.0641963705,0.0123498654,-0.204471454,0.0265595634,-0.4099167883,-0.0119645176,-0.143138513,-0.1234519556,-0.4040339291,-0.5144379139,-0.0526950508,0.0811511278,0.381595999,0.272680223,-0.000126695,-0.3327266574,0.0753403977,0.0116364425,-0.0123662101,0.267736882,0.1676332504,-0.5629211068,-0.1614192128,-0.253213048,0.2919107974,-0.5190222859,-0.2093334943,0.3381920159,0.0419585146,-0.1553609222,0.0847018063,-0.1341997832,0.033729028,0.4106487036,0.1461370587,0.094788827,-0.0445502438,-0.193176493,-0.125205189,-0.1861883402,-0.0148738222,0.1704450995,0.0846230239,-0.2187519372,0.3521589637,-0.0789667591,-0.1837906837,-0.0440309532,0.2250877321,0.2020062655,0.3780929446,-0.1213997677,0.0179003906,-0.2720344663,-0.1647694409,-0.0660214499,-0.1757876873,0.1807849854,0.2901870906,-0.0319655389,-0.3919959068,-0.0505461805,0.1163604036,0.2394372076,-0.1916068792,0.530661881,0.0373358503,0.1746242046,-0.2876830399,-0.171141848,-0.0772107616,-0.231599763,-0.1640636772,0.4628706574,-0.334744066,-0.3571234047,0.0454478711,-0.0553786941,0.5117006302,-0.3004132807,-0.3079538941,-0.0179550126,-0.0300103668,-0.1349734366,0.1017569602,0.5036150217,-0.320517689,0.4058420956,-0.1463568062,-0.0689106286,-0.5243211389,-0.1801228523,0.0581733361,0.0973942503,0.1247244701,-0.2626882493,-0.36893332,-0.0654681846,0.2246282846,0.2283162028,-0.5162183642,0.0501702949,-0.0462418012,0.353132993,-0.1877923906,0.398598671,-0.3270302415,0.2114166915,-0.136152789,0.0425623767,-0.1161684692,-0.1355384141,0.0962168351,0.1538248807,-0.2341097593,0.1981964558,0.3465766609,0.0936483592,0.0146889258,-0.0968535244,0.4784098566,-0.2083839923,0.3629956245,0.1219941303,0.1408423781,0.3056558073,-0.0901968256,-0.0957771912,-0.1201085597,0.1409538686,-0.5012376904,-0.169020012,0.0116774058,0.0399583317,0.0463613011,-0.0260105561,-0.3790266812,0.3396331072,0.445038259,-0.0485933945,0.2475073636,-0.2733230889,-0.3672395051,-0.1775032431,0.0153883873,-0.0809691027,-0.0650012121,-0.012406134,0.3479374945,0.2410304099,0.349231869,0.6582411528,-0.2303016782,0.5145463943,-0.2676454782,0.5004504919,0.4632590711,-0.4106018841,-0.7795951962,0.3592059314,0.2177218795,0.1909922659,-0.0905768275,-0.062267378,0.4403330982,0.1709681302,0.1503650248,0.2609789371,-0.0067491755,0.0891227871,-0.4096227586,-0.1871406883,0.3767555952,-0.5307529569,0.2135346979,-0.0000495143,0.0488071628,0.461697042,0.4492997229,0.0690544769,0.2840161324,0.2604551315,-0.0561473072,0.0318829864,0.2822740078,0.0226592347,0.2465954721,-0.0556628406,-0.2750283182,-0.0677846894,0.2460779548,-0.3916614354,-0.2138777971,0.0052284729,-0.1389308274,-0.2794297338,0.0234179758,-0.0932357311,0.1023214534,-0.0383189172,0.2316971719,0.1100384593,-0.1432964951,0.0935535952,0.1334498376,0.1645781398,0.0402019545,0.1588942707,-0.298945725,0.2747724354,0.150938496,0.0381687693,-0.2896446288,0.0561730787,0.1120401472,0.0079375952,0.1598926485,0.3029532433,0.2006729692,0.0209676512,0.3990192711,0.0519715957,0.0553002618,-0.1516284645,0.1953581423,0.2741399109,0.4093807042,0.0639325827,0.0830415934,-0.2988683581,0.1838340014,0.0164156202,0.1465729624,-0.210712105,-0.1486240029,0.2459668368,0.0925299302,0.0446782671,0.0847314447,0.3287924826,0.4958305061,-0.0970020592,-0.0951333717,0.092579633,-0.6602830291,-0.5403527021,0.148996532,-0.0497088321,0.2239417136,0.099999547,0.3472332358,0.0199308563,0.2298643291,0.0970973521,0.2053426355,-0.085895434,0.0185722504,0.2559397817,0.5013879538,0.1528055668,-0.2660228312,0.202513501,-0.0429480076,0.3088431358,-0.1124926284,0.0850555524,-0.1335549653,-0.1435682327,-0.0023192614,-0.039491009,-0.1966706067,-0.4011041224,-0.0901038498,-0.07314533,0.3624985516,-0.2025968432,-0.0469216406,0.0116731729,0.1543474793,0.1703260988,0.0842947587,0.0560884178,-0.3676507175,-0.0593170598,-0.0991245285,-0.0619967133,0.2375106215,0.1665263027,-0.0645672902,-0.3458328843,-0.8534325957,0.0825678632,-0.084121339,0.0708762929,-0.0375623852,-0.0424161553,-0.4549081028,0.1427483112,-0.0337599181,0.084429495,-0.2855614424,-0.1527382433,-0.3064069748,0.1108344123,0.0838663727,0.0327131152,-0.0781294554,-0.0054610954,0.0288356692,0.0639322698,0.0412345901,0.0331529342,-0.2343938798,-0.0055217049,-0.0295666941,-0.0494091026,-0.4937898219,-0.2349341959,0.2614579499,-0.186485216,-0.2871465981,-0.1379059851,-0.2127659768,0.1674518883,0.1235416904,-0.575129509,0.1784176081,-0.2078160346,0.191449523,-0.2898373604,0.0419713072,0.071193859,-0.1305159628,0.0412760638,-0.065564841,0.0627502576,-0.3446030915,0.1373836249,0.5308549404,-0.1019116789,0.441552937,0.3388991654,0.9411715865,0.0710766986,0.059741471,0.2623276114,-0.2129848003,0.1000787169,-0.2972133756,-0.1126083434,0.1811749935,-0.0857195929,0.1975196898,0.4480571151,0.1431238204,-0.1449812353,0.6234247088,-0.4005103111,0.0346171372,-0.048764471,0.1275820434,-0.2185366899,0.2929039896,0.206403628,-0.199714452,-0.5666322708,-0.4909052551,-0.0252021998,0.1508867443,-0.0897450596,0.1794929504,-0.6895117164,-0.1021163762,-0.6798046827,0.2289959192,0.1907180846,0.2268924117,0.2893172503,0.1228372604,0.0986716896,-0.0000464669,0.5573988557,-0.0954390615,-0.1696855128,-0.1001953781,-0.0893320665,-0.2187113166,-0.1574794203,0.0460026823,0.1913294494,0.4865813255,0.613401711,-0.0369437858,-0.2032911479,-0.0758208036,0.5636317134,-0.1356333345,-0.0907220244,-0.3391308486,-0.0056420513,-0.3226272762,-0.0101695936,-0.1441559494,0.2290710062,-0.0105876178,-0.3273718059,-0.2876749337,-0.2229082584,-0.1403618306,0.080937922,0.3359941542,0.2816732824,0.042550426,-0.2974095643,0.2304454595,0.0629048795,-0.0645938739,0.1954265684,0.0152798602,-0.1770285964,0.0828779042,0.2911258638,0.0938986316,-0.0147900553,0.5646324158,-0.2761473656,0.1382764429,-0.0533560961,0.475040257,0.3200357854,0.139131844,-0.4849229455,-0.2804906666,0.2338916361,0.1095558181,-0.051028192,0.2023665458,0.050314419,-0.1862108409,0.3887755573,0.0281450283,1.0094174147,-0.0559253842,0.3413280249,-0.0150667951,0.0341818444,0.2308701426,-0.502381146,0.0393605642,-0.2651951611,0.247264266,0.0367768034,-0.0200331211,-0.1346169114,0.218086496,0.0114376964,0.0484562106,0.0225089639,0.326221168,-0.3503052592,0.5328888297,0.078415148,-0.0115081938,0.0440585278,0.0034808861,-0.0942111313,-0.194536984,-0.2212127298,0.1340264827,-0.1591313779,-0.2852482796,-0.0846323371,-0.1929343343,-0.1119007468,0.4539324343,-0.0369846895,-0.2596124411,0.0445569679,0.0951860324,0.080141373,0.5051612854,-0.0039875014,0.1667694449,-0.0645396858,0.1965014338,0.2081844956,-0.1267438084,0.0821605548,-0.0444018319,-0.2410298288,-0.0791256502,-0.0179513376,-0.2659711838,-0.3112939,0.2222388685,0.4669784307,-0.3620292246,-0.1344530135,0.1378592253,-0.1779134721,-0.1051538587,0.0436931662,-0.1480315626,0.019449234,0.1870409995,0.3336391151,-0.4734423459,-0.073312819,0.3393525481,0.3918019831,0.0628241152,0.7738110423,-0.3206628263,-0.1390824467,-0.0217643455,0.2482616305,0.0813539028,0.1282113343,-0.1915438473,-0.0926737636,0.1157943234,-0.009256307,0.0023911558,-0.1650209278,-0.0261182152,-0.297467947,-0.2309001684,-0.5808985829,0.0390865579,-0.4215116203,0.238992095,0.2948336005,0.226332739,-0.2764213383,0.4198241234,-0.1429228932,0.0762160346,0.0843107924,0.1467740834,-0.3328534365,-0.0210713074,-0.1339057982,0.1198610812,-0.0671060979,0.1427742541,0.0320815817,-0.147787109,-0.3229166567,0.1828864962,0.3324050307,-0.1604850143,-0.0970514193,0.0103180297,-0.1327730566,-0.322101593,0.2611097991,-0.0599336438,0.1192526668,0.1499315351,0.3006532788,-0.1675349027,-0.0655774251,0.0538571551,0.2275741696,0.0984328911,0.1034617499,0.1702799201,0.2892483473,-0.0920974687,-0.5138878822,0.0483801961,0.3064833283,0.0442463979,0.3689006269,0.1666705608,0.0799836218,0.0112886038,-0.201993376,0.1030655876,0.0412109382,-0.3377054632,0.3317939043,0.0025519945,-0.0645574927,-0.0517049991,0.2551340759,-0.2453050613,-0.137802422,-0.0538866632,0.4683297575,0.2841570377,-0.4058928192,-0.0885316804,0.311581701,-0.2014083713,-0.0777834579,0.1521663666,-0.0061099478,0.0656590983,0.2132089436,0.0985811129,0.1229559034,0.3483352065,0.3747964203,0.3071369529,0.2125925124,0.0573883764,0.1475170851,-0.1972602457,-0.2420945764,0.3180951774,-0.1158268228,0.1136649027,0.0794959962,0.0469808057,0.5698667169,-0.3263924718,0.0916749462,0.3821566105,-0.2795787156,-0.2462017089,-0.0392531939,-0.0029507265,0.0290731527,0.061105974,-0.048690442,-0.1476872563,-0.2208212167,-0.0343722515,-0.1883826852,-0.1918771416,0.292958647,0.2847601771,0.1494754255,-0.2994998991,0.0408210903,-0.2890634835,0.1501983851,-0.1944538951,0.2535018921,0.3292703331,0.2624985278,-0.1754819751,-0.0608235449,0.1230738536,0.0706250221,-0.0877103582,0.0532670021,0.2275362164,-0.4091357589,-0.0322198085,-0.1101516187,-0.066395767,-0.3087646365,0.3365980983,-0.2456378937,0.0271532703,0.1550098956,-0.1166002452,-0.2735729218,-0.1233112887,0.2119901627,-0.5336232781,0.1218724623,0.4087361097,0.0128751574,0.3519569337,-0.3962770104,0.0306353997,-0.356400758,0.421952039,0.5847312212,-0.1407602727,-0.2208828181,-0.218722105,-0.1122579053,0.264161706,-0.3131760061,0.0552457199,-0.2011154592,0.3303229809,-0.2282004505,-0.3678624928,0.0640006661,-0.1404257417,-0.0727992356,0.1040831432,-0.0434604324,-0.0415736102,-0.3713268042,0.1005873978,-0.1629937738,-0.0538744517,0.2102728188,0.0239211041,-0.1234043241,-0.1336365193,0.1404177994,0.0241975542,0.0155283604,0.335054189,0.2555703521,0.3614542186,-0.1383251697,0.2697926462,0.1320325136,-0.056089431,-0.0695149601,0.2591328919,-0.2424577177,0.0142122954,0.1162445694,-0.0769676641,0.2797967792,-0.0198589489,0.0625004768,-0.282007426,-0.0978202596,0.1297300905,-0.1783958077,-0.0050885524,-0.127691716,0.522887826,0.0418595001,0.1902500093,0.0206489693,-0.252109766,0.2256608605,-0.4897980988,-0.1273146421,-0.2483121306,-0.0441544056,0.2793737054,-0.2569743991,-0.391503036,-0.0738655552,0.5597876906,0.1051387042,-0.097871922,0.2520550191,0.145486787,0.1653405875,-0.1739533544,0.3162582517,-0.2240762115,0.0436558202,0.1174812019,-0.033569742]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/258","title":"Why is dataset after tokenization far more larger than the orginal one ?","comments":"Hard to tell... This is probably related to the way apache arrow compresses lists of integers, that may be different from the compression of strings.","body":"I tokenize wiki dataset by `map` and cache the results.\r\n```\r\ndef tokenize_tfm(example):\r\n example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text']))\r\n return example\r\nwiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train']\r\nwiki.map(tokenize_tfm, cache_file_name=cache_dir\/\"wikipedia\/20200501.en\/1.0.0\/tokenized_wiki.arrow\")\r\n```\r\nand when I see their size\r\n```\r\nls -l --block-size=M\r\n17460M wikipedia-train.arrow\r\n47511M tokenized_wiki.arrow\r\n```\r\nThe tokenized one is over 2x size of original one.\r\nIs there something I did wrong ?","comment_length":25,"text":"Why is dataset after tokenization far more larger than the orginal one ?\nI tokenize wiki dataset by `map` and cache the results.\r\n```\r\ndef tokenize_tfm(example):\r\n example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text']))\r\n return example\r\nwiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train']\r\nwiki.map(tokenize_tfm, cache_file_name=cache_dir\/\"wikipedia\/20200501.en\/1.0.0\/tokenized_wiki.arrow\")\r\n```\r\nand when I see their size\r\n```\r\nls -l --block-size=M\r\n17460M wikipedia-train.arrow\r\n47511M tokenized_wiki.arrow\r\n```\r\nThe tokenized one is over 2x size of original one.\r\nIs there something I did wrong ?\nHard to tell... This is probably related to the way apache arrow compresses lists of integers, that may be different from the compression of strings.","embeddings":[-0.1123888046,-0.1193458214,0.0112903155,0.2027449757,-0.004323286,-0.0841799676,0.2501812875,0.4384357631,-0.1686826497,-0.0794514418,-0.0184094273,0.1791732907,0.1267877519,-0.4135742784,0.272151947,-0.0815012231,0.2910658121,-0.0246310961,0.0719703063,-0.1968753636,0.0347574987,0.0416620113,-0.1358028799,0.0729799196,-0.4001576602,-0.0449916124,-0.129962936,-0.0719472393,-0.2361072004,-0.5674791932,-0.0745736212,0.0121322563,0.4288644493,0.211513564,-0.000128962,-0.2799145281,0.1451965868,0.0475846492,-0.1298143268,0.2928497791,0.2264073938,-0.6245266199,-0.0985882804,-0.1812898219,0.2950395346,-0.6420503855,-0.2269112021,0.0821715295,0.1848838329,-0.1353145838,0.08482638,-0.0938159451,-0.107243076,0.3942354321,0.1767108738,0.1288815588,-0.0818373933,-0.2256771624,-0.1874342412,-0.1747066677,-0.050890673,0.1627078801,0.063438125,-0.1471032649,0.3064796925,-0.0714227855,-0.2241654843,0.0634685904,0.2754486203,0.2621813118,0.4169594944,-0.0156482328,-0.005996719,-0.2301370353,-0.201808393,-0.0389131866,-0.0931153372,0.2346327901,0.2906144559,-0.017525468,-0.2741985619,0.0263411086,0.0947591513,0.1975840926,-0.2298889756,0.4450588822,0.0401627533,0.217928648,-0.2753769457,-0.1545382589,-0.0586714633,-0.2145055979,-0.1798379123,0.4322636724,-0.1928647757,-0.2621163428,-0.0515298918,-0.2579568923,0.4600994587,-0.0762406737,-0.1475306153,0.1732755601,-0.041558221,-0.2559193969,0.042412404,0.4979791939,-0.1888276786,0.4086038768,-0.0788646042,-0.0163268335,-0.4559734464,-0.1477343738,-0.0141379097,0.0840102956,0.191803962,-0.18388246,-0.4125388265,-0.073911719,0.2471038848,0.2568508685,-0.5777155161,0.0105630895,-0.1531440914,0.2712470591,-0.1933001578,0.4041868448,-0.2657767832,0.225282222,-0.1282550693,0.0076675951,-0.0605316609,-0.3740162253,0.0974092335,0.2564538717,-0.2381279469,0.1824522614,0.3605071306,0.0697530508,-0.0042297896,-0.0986470431,0.5601621866,-0.1723465621,0.371450007,0.1005745307,0.0955010131,0.3552550972,0.0051985444,0.0059400466,-0.117996335,0.1684743762,-0.459382534,-0.121962145,0.0247707125,-0.0249905102,0.042194169,0.0598158501,-0.2653065026,0.2889685929,0.3526547849,0.0013087175,0.2010141313,-0.2585879564,-0.2672606409,-0.2102826238,-0.0157373771,-0.1517865509,-0.1470724642,0.0005063387,0.267262727,0.4102359712,0.3200981617,0.6782882214,-0.3026679754,0.3560176194,-0.246985361,0.4818019569,0.4356308877,-0.2668084502,-0.7089950442,0.3641351759,0.3064741194,0.1355556846,-0.1135172695,-0.1115586013,0.4555978477,0.1529850364,0.0637237206,0.3180859685,0.0895827338,0.0531876162,-0.3666683733,-0.1303328127,0.2926559746,-0.564604044,0.1038689464,-0.031335406,0.1121084616,0.3461750448,0.4183446169,0.0585711859,0.3294259012,0.3610934615,-0.1227441281,-0.0275731608,0.3681645989,0.0696429312,0.2804941833,-0.1416314542,-0.3820162714,-0.0580911115,0.0012402319,-0.3346402943,-0.2133373469,0.07413885,-0.026293179,-0.2358429432,-0.0214437116,-0.0417720377,0.0621552877,0.0222867243,0.2409428656,0.0759207681,-0.0031610993,0.1322761178,0.1277240962,0.1355972439,0.144617945,0.2024888545,-0.2423618585,0.1948191673,0.1212050095,-0.0206076019,-0.2502890825,0.0443870798,0.0982219204,-0.0688159764,0.2537713647,0.2087551206,0.2937293649,-0.070459269,0.5168308616,0.2028337121,0.0226970892,-0.1259239614,0.1228789985,0.4375647306,0.4034962654,0.1817870885,0.0409672707,-0.1742817909,0.1905086488,0.119663775,0.0280658994,-0.200728327,-0.2002355456,0.332288444,0.1952180117,0.0820949525,0.1325874627,0.3306999505,0.3930947781,-0.2206744701,-0.0446204059,0.1269198507,-0.7217383981,-0.5281842947,0.1752351075,-0.1243394092,0.1278612763,0.0770362243,0.3478259742,-0.0948631763,0.1967029721,0.1062638238,0.2983808517,-0.0800254717,0.132299304,0.1677952558,0.414031893,0.0981151536,-0.2518465817,0.1062500104,-0.017850941,0.2040759474,-0.0552605875,0.1663462073,-0.236582011,-0.0052166781,-0.0908513144,-0.1407175362,-0.4034102857,-0.4084809721,-0.0656422228,-0.1708032042,0.2646420598,-0.1635754704,-0.0970429331,0.0325585343,0.2891853154,0.1397779733,0.0746874511,0.0765789226,-0.4737802744,-0.0896960646,-0.0797646344,-0.132977441,0.2211560905,0.2187421173,-0.0599924959,-0.4326705933,-0.7699640989,0.1413439214,-0.0497472845,0.0092643667,-0.0213063862,-0.0506681502,-0.4660758078,0.1676148921,0.0186135676,0.0561215393,-0.2046419978,-0.0709274188,-0.2541188002,0.164123401,0.1320967227,0.0119778151,0.0709322616,0.0585376099,0.0074060177,0.1563634127,0.0609027781,0.156238243,-0.1296957582,-0.0550642535,-0.0065985862,-0.1961809695,-0.4393118918,-0.1822176129,0.3604336381,-0.2064551115,-0.3408961296,-0.1987121254,-0.2136014104,0.1834375709,0.2203746587,-0.5550186634,0.3347565532,-0.2044409513,0.165296942,-0.3110952377,-0.0124408538,0.0915132314,-0.0509849563,-0.0098378854,-0.0335424319,0.1119629815,-0.3022398055,0.0193320904,0.6221038103,-0.0128397001,0.4115440547,0.260751456,1.0486251116,-0.0229009874,0.1311082691,0.2479877174,-0.2006756812,-0.0039068745,-0.2587243319,0.0423114002,0.0611592494,-0.1547425538,0.1754122376,0.4836206138,0.2739487886,-0.1684340686,0.6691889167,-0.27701056,0.04234742,-0.1178961545,0.2433492541,-0.3538164794,0.2527452707,0.2253294438,-0.1377001554,-0.6037968993,-0.4350684881,0.0791640058,-0.0078942776,-0.0352805331,0.1069494188,-0.6138989925,0.0649231672,-0.7611913085,0.3399455845,0.0583171211,0.244396925,0.3503576517,-0.0373924375,0.0886533856,0.0104699926,0.4196232557,-0.0363927037,-0.2047560662,-0.1232365966,-0.0005622844,-0.4640890658,-0.1336940229,-0.0128834303,0.1218886301,0.3096955121,0.5771069527,-0.082464397,-0.2465149611,-0.1993007511,0.5381547213,-0.1532351375,0.0488490611,-0.3741458952,-0.0671851709,-0.3274954557,-0.0526400432,-0.163344726,0.2704223692,-0.0068172775,-0.320530355,-0.3182036877,-0.3208780885,-0.1378476918,-0.0051957909,0.4126265347,0.1711201072,-0.0614807345,-0.2433637828,0.3658642769,0.0413623936,-0.0728578717,0.326884985,-0.0405823439,-0.0946427509,0.0143989511,0.2570184171,0.1378474683,-0.0308172107,0.4664027691,-0.1954167038,0.1128077134,-0.014232507,0.3922351003,0.2835197747,0.1453584582,-0.4813140631,-0.2578438818,0.2884397209,0.1952428222,-0.0361283869,0.3587246835,-0.0895751491,-0.1282112002,0.4765395224,0.2143427432,1.0645099878,-0.1115282625,0.3996556997,-0.0170521438,-0.0154630216,0.1706447154,-0.5571756959,0.0537095331,-0.2833386958,0.2938237488,0.0126751503,0.0317040868,-0.0374304764,0.3189369738,0.0832906291,-0.0088179894,0.0144569501,0.2932939529,-0.342215836,0.4235132039,-0.0014370279,-0.075497523,0.0006077261,-0.0099989707,-0.0705031902,-0.1627050638,-0.1790656149,0.0928609967,-0.2279163748,-0.3124333918,-0.0365944654,-0.1596956998,-0.0972949192,0.3951896131,-0.0150334043,-0.3361245692,0.0076362672,0.0188493803,0.0126416022,0.3787163496,0.0008672501,0.156933859,-0.133917883,0.2694789469,0.1650175303,-0.2293627709,0.0245294664,0.0050407993,-0.1588830799,-0.1181002557,-0.0545048825,-0.3009825051,-0.1761486083,0.2057027966,0.504776001,-0.3452023864,0.0327578895,0.0874031186,-0.250889957,-0.1133787483,0.0193349626,-0.1655512601,0.063443996,0.090292573,0.4422696531,-0.3927358091,-0.0395509191,0.3326480389,0.331720382,0.0268771928,0.7461211681,-0.2636687756,-0.1310627162,-0.0517041124,0.2981928885,0.1601020992,-0.0195900537,-0.1013847142,-0.1895347834,0.0377715789,-0.0206094645,0.0548599958,-0.164143607,0.02651955,-0.3682690263,-0.1794842035,-0.560859859,0.0533695295,-0.4516380429,0.2238231003,0.1408394873,0.0199120231,-0.2691285014,0.3149071038,-0.0987985581,0.073181197,0.2684805989,0.168935582,-0.5337049365,-0.0469281077,-0.0677845851,0.197386995,-0.0738605633,0.2152202129,0.0942161977,-0.1097401977,-0.2969451249,0.1910550147,0.2100258023,-0.2378118187,-0.0930809677,0.0462536365,-0.0668524727,-0.4275300503,0.1565885842,-0.181319043,0.1088280752,-0.0223954972,0.4402322471,-0.1651485115,-0.13763915,0.0445432253,0.1723861992,0.1813743711,0.0975111276,0.189646557,0.2859797478,-0.0879364014,-0.506822288,0.0364461653,0.2951573431,0.0342006721,0.4252802432,0.1027521938,-0.0073546404,0.0630362481,-0.1609052122,0.1571957916,0.0794426873,-0.4598769546,0.2521470785,-0.0044134413,-0.0756356567,-0.0885662436,0.1311997622,-0.2073742449,-0.2714795172,-0.0602323934,0.4227359295,0.2131667733,-0.3525822163,-0.1477334499,0.3129835427,-0.2261547148,-0.0269443411,0.2111188918,0.1031536832,0.0894438177,0.3192686141,-0.0747434571,0.1782976389,0.3176268935,0.3104436994,0.3052541316,0.4638380706,0.0481074527,0.0405283011,-0.1779925227,-0.2076733559,0.3831957281,-0.024064526,0.1375554651,0.1992515624,0.1787275076,0.573466897,-0.3003483117,-0.01418206,0.3012088835,-0.2191375643,-0.2593933642,-0.1138371229,-0.0150000239,0.0567117967,0.0322715566,-0.0517925471,-0.1727198213,-0.1193908677,0.0002396529,-0.0766655952,-0.2563026547,0.2634271383,0.2316263765,0.127398476,-0.2074840963,0.0105991662,-0.2688136101,0.1709202528,-0.124040179,0.2004351914,0.3524907231,0.1364900023,-0.1879258603,-0.02572551,0.0813760161,0.1156183109,-0.0842894092,0.0669776127,0.2507988214,-0.4366205633,-0.0528750084,-0.1223691925,-0.0135233672,-0.3929215968,0.2501864433,-0.2340856791,0.0914523378,0.1331598312,-0.1589409262,-0.278450191,-0.1642617732,0.1304123849,-0.3757233024,0.1307713836,0.4261535406,-0.006274784,0.2964459062,-0.3734609187,-0.0024554373,-0.4488922358,0.3881143928,0.5730144382,-0.2211462259,-0.1471798569,-0.263792634,-0.029432388,0.2714197636,-0.3679638207,0.027342869,-0.1536201388,0.3586548865,-0.1316228509,-0.2930263877,0.1690470576,-0.0983446985,-0.020916475,0.1574370712,-0.100277029,0.004404563,-0.1687359512,0.2540200055,-0.2325387448,-0.1018429697,0.2121306956,0.1338328421,-0.1443418711,-0.124887161,-0.032014478,0.1201638281,0.0669568628,0.4428333044,0.3170431852,0.2394460589,-0.0581513308,0.1545061916,0.035499312,0.009113254,0.0697098821,0.3531410098,-0.2202854455,-0.0345106572,0.1050051972,0.0072602378,0.1861708313,0.0611907765,0.0673458725,-0.2035116404,-0.2058465928,0.049439989,-0.2224158496,-0.0305529051,-0.2114424109,0.5169645548,0.0888474584,0.1619945467,-0.0088011967,-0.3011767566,0.1538934708,-0.5309007764,-0.184634611,-0.1154258847,-0.0200892687,0.2848861217,-0.193269521,-0.3655531704,-0.1028587371,0.5226110816,0.0769377202,-0.0704751164,0.0471263267,0.1259960681,0.1774275005,-0.2241218239,0.4204753637,-0.2645287812,0.0459215492,0.0279825982,-0.0734163076]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/258","title":"Why is dataset after tokenization far more larger than the orginal one ?","comments":"Thanks for your point. \ud83d\ude00, It might be answer.\r\nSince this is hard to know, I'll close this issue.\r\nBut if somebody knows more details, please comment below ~ \ud83d\ude01","body":"I tokenize wiki dataset by `map` and cache the results.\r\n```\r\ndef tokenize_tfm(example):\r\n example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text']))\r\n return example\r\nwiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train']\r\nwiki.map(tokenize_tfm, cache_file_name=cache_dir\/\"wikipedia\/20200501.en\/1.0.0\/tokenized_wiki.arrow\")\r\n```\r\nand when I see their size\r\n```\r\nls -l --block-size=M\r\n17460M wikipedia-train.arrow\r\n47511M tokenized_wiki.arrow\r\n```\r\nThe tokenized one is over 2x size of original one.\r\nIs there something I did wrong ?","comment_length":30,"text":"Why is dataset after tokenization far more larger than the orginal one ?\nI tokenize wiki dataset by `map` and cache the results.\r\n```\r\ndef tokenize_tfm(example):\r\n example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text']))\r\n return example\r\nwiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train']\r\nwiki.map(tokenize_tfm, cache_file_name=cache_dir\/\"wikipedia\/20200501.en\/1.0.0\/tokenized_wiki.arrow\")\r\n```\r\nand when I see their size\r\n```\r\nls -l --block-size=M\r\n17460M wikipedia-train.arrow\r\n47511M tokenized_wiki.arrow\r\n```\r\nThe tokenized one is over 2x size of original one.\r\nIs there something I did wrong ?\nThanks for your point. \ud83d\ude00, It might be answer.\r\nSince this is hard to know, I'll close this issue.\r\nBut if somebody knows more details, please comment below ~ \ud83d\ude01","embeddings":[-0.072691001,-0.1665187478,0.0337734222,0.1605291814,-0.0135168238,-0.0379374921,0.287569344,0.3828713596,-0.0926128626,-0.0533792302,-0.1073254943,0.1967065334,0.082892172,-0.1633370072,0.2538659275,-0.0917038247,0.3183568418,0.0056694299,0.1641955674,-0.2462658137,0.0687176883,-0.0034690171,-0.1630095243,0.080906488,-0.4438725412,-0.0409448482,-0.1405430287,-0.0889516473,-0.3230459392,-0.4877051711,0.0544542223,0.0904541686,0.4469436705,0.2833976448,-0.0001280673,-0.3116829693,0.0528840944,0.0146437287,0.0397626869,0.2755564451,0.1965554357,-0.5918030143,-0.1092252284,-0.2343671024,0.2229949236,-0.5260525942,-0.2225880474,0.2866103649,0.0634340569,-0.1216695085,0.0849941671,-0.1538697332,-0.0244361199,0.4318528175,0.108915627,0.0561084896,-0.0649044737,-0.152423203,-0.1710953563,-0.2175376266,0.0182565302,0.1856680959,0.0337630883,-0.2018301487,0.334808439,-0.0902845562,-0.1766476333,-0.0227145962,0.216934517,0.2312103808,0.3756917119,-0.0655214861,0.036623843,-0.2437412292,-0.1951198131,-0.0647377819,-0.1502988487,0.2063044161,0.3383838832,-0.0508698896,-0.4059690535,-0.0109195858,0.1191241443,0.2104614675,-0.2516656518,0.5392639041,0.0459858887,0.1912549883,-0.3162843287,-0.1431555748,-0.0920566395,-0.207180351,-0.1333358437,0.4413743615,-0.3257778287,-0.3650396466,0.042679172,-0.0733757243,0.4625688791,-0.2051334232,-0.3422219455,0.0748761371,-0.0427960567,-0.2019629627,0.0850400701,0.4957807064,-0.2846673727,0.4225941002,-0.0632860959,-0.0480735786,-0.5810618997,-0.1435708553,0.0103813158,0.1031589434,0.157706663,-0.2445176542,-0.4083057642,-0.1312685907,0.2424326986,0.2182481885,-0.5596010685,0.0609375685,-0.091123946,0.3551521599,-0.2358727604,0.3478893936,-0.3164413869,0.1695366502,-0.1548074186,0.051054243,-0.106949009,-0.1914841682,0.0759078115,0.1784043312,-0.2429095507,0.1946538836,0.3395061791,0.0500411876,-0.0452153273,-0.1140034795,0.4913453758,-0.1974925995,0.3377678096,0.1062311903,0.1137147322,0.287622124,-0.0561812036,-0.0767778754,-0.152647987,0.1096385419,-0.4803295434,-0.1649407297,0.0217577796,0.0157407206,-0.012823469,-0.0166290887,-0.384519428,0.3187755942,0.3729155064,-0.0135454014,0.1754490882,-0.2576174438,-0.3746815622,-0.1883584112,0.0029015499,-0.1387889534,-0.0537368506,-0.0170464572,0.3274080753,0.2638149858,0.3902115226,0.6496686339,-0.2520271242,0.5214618444,-0.2812904418,0.4189815819,0.4458949566,-0.409520328,-0.7466168404,0.3305455744,0.2325271517,0.2159551829,-0.1696109623,-0.1010147631,0.45513767,0.1278775632,0.0742719322,0.3066436052,0.0433652177,0.1288255155,-0.3877939582,-0.1889545619,0.3930892348,-0.4821769297,0.2345898896,0.0410103276,0.0740305483,0.4521016181,0.4246041775,0.1116348803,0.2516742945,0.25370121,-0.0022564803,0.0424169339,0.3129778802,0.0177450981,0.2182937413,-0.0605856031,-0.3031754196,-0.0454438068,0.210212186,-0.3845507503,-0.1622486115,-0.0162679944,-0.131367743,-0.339032352,0.0077267531,-0.1016768441,0.1528059691,0.0177973043,0.2814898789,0.0604653098,-0.0813478976,0.083539322,0.1748135686,0.2153217494,0.0870512277,0.2072841972,-0.2524546981,0.2669246197,0.1636469066,-0.0030985218,-0.2674341798,0.0386190005,0.0881199613,0.0319925882,0.2041619271,0.2901285589,0.2043853849,-0.0191812757,0.4399230778,0.0906587914,0.0196062829,-0.1277063638,0.1646957844,0.3998301923,0.4183766246,0.0556076989,0.0607926995,-0.2680594027,0.1888647527,0.022559993,0.1557742506,-0.2136959285,-0.1384789795,0.2503273487,0.0691841841,0.0404487103,0.0831980556,0.3396417499,0.4841019213,-0.1333108693,-0.0797042474,0.1216859668,-0.6560309529,-0.541195333,0.198623091,-0.0623448752,0.1635153294,0.0901921391,0.3118586838,0.0150013231,0.2528746128,0.1363239735,0.2016536891,-0.1087834686,0.0637223423,0.2056415677,0.4870423079,0.1158730462,-0.2442601919,0.1812293977,-0.0427489951,0.3363183141,-0.0424735211,0.0907228142,-0.1933244616,-0.0987319425,-0.0518575758,-0.0370747261,-0.2787724733,-0.3823391497,-0.1147487164,-0.0611493289,0.3703647256,-0.1481208205,-0.0532165468,0.0151113374,0.1142131984,0.1771213561,0.1322360784,0.0367702246,-0.4259366691,-0.0662372857,-0.0845860168,-0.0618039668,0.2719150782,0.1413689703,-0.0514149927,-0.4133565128,-0.8184748292,0.0847441778,-0.0484303497,0.0695959702,-0.0285598375,-0.0496390015,-0.4696209431,0.1774952412,-0.0112362402,0.0417142175,-0.2167793363,-0.130121991,-0.3662998676,0.2011483908,0.1551592946,0.0365481749,-0.1027734131,0.0529157259,0.0019789981,0.0863968357,0.0523911342,0.027867699,-0.2423267215,-0.0767747015,-0.043700289,-0.1285436749,-0.4691095948,-0.2726513445,0.2572519481,-0.1844524145,-0.2743245959,-0.1149621755,-0.1867665797,0.1754225343,0.1072964817,-0.5669481754,0.2361117452,-0.1820842177,0.1882661879,-0.2580097318,-0.0654695109,0.1083400995,-0.1316745132,0.0668529421,-0.0261351559,0.0776201636,-0.3353712261,0.0728584602,0.514326334,-0.0541051179,0.4642397761,0.279201746,0.978761971,0.0283438731,0.0337975807,0.2215492427,-0.2428744733,0.0253304858,-0.2796728015,-0.0883811116,0.2139550894,-0.1028437763,0.1899794936,0.4913728535,0.1769777536,-0.1503149569,0.599945724,-0.4078009725,0.0194058735,-0.0921145752,0.1283288449,-0.244644478,0.2750113606,0.2108824551,-0.2089214027,-0.5871834755,-0.5047989488,0.0068574664,0.1920865625,-0.0535593033,0.1837977469,-0.7017826438,-0.0472627804,-0.6418154836,0.2449791133,0.1216787696,0.2035819441,0.3331296146,0.1106136143,0.0969289839,0.0028038556,0.5349250436,-0.0899631605,-0.123418726,-0.1303569525,-0.0831986964,-0.2667235732,-0.2034880519,0.0502668098,0.2079893202,0.422115624,0.6396846175,-0.0181745403,-0.2408898324,-0.1324923933,0.5823591948,-0.1514609307,-0.0366330296,-0.3615648746,0.0097441692,-0.314171046,0.0231520869,-0.2293952703,0.2083012015,-0.0395015888,-0.3119896054,-0.2788812816,-0.2269030958,-0.1332676113,0.0934266299,0.3569119871,0.2507148981,0.050375849,-0.2593683898,0.2653886378,0.1222787499,-0.0939074084,0.1829505265,0.0447234288,-0.1494618356,0.1109513789,0.2913945615,0.1125302538,-0.0218571257,0.5225077868,-0.2703311145,0.135404557,-0.0401770771,0.4362836182,0.2950796187,0.1874892116,-0.4897307754,-0.2730542421,0.215035513,0.1590602398,-0.0525603741,0.2352221906,-0.0011437321,-0.1771446615,0.4197638035,0.0752856135,1.0324589014,-0.081681259,0.3404111266,-0.0063023157,0.0415061414,0.2976212204,-0.5319032073,0.0493122637,-0.282063514,0.258089602,0.0262024663,0.0240562335,-0.1306201369,0.2523039877,0.0543722361,0.0716085583,0.0417851135,0.3470294178,-0.3743632734,0.5021025538,0.1454901546,-0.0348214805,-0.0076349787,-0.00885705,-0.1493139267,-0.1716518104,-0.2195957005,0.1203320846,-0.150847882,-0.2544943988,-0.0115729012,-0.2179704159,-0.0600833893,0.425565809,-0.0163853765,-0.283823818,0.0268752091,0.065737471,0.0774037689,0.4879695773,-0.0089885937,0.1861478239,-0.0499046296,0.2513028681,0.2068410814,-0.160418272,0.058146432,-0.0626068488,-0.2437385321,-0.1374602765,-0.0259486251,-0.2779219449,-0.2813147306,0.1817311198,0.4592375159,-0.3054692447,-0.0920101926,0.1078287512,-0.1723325104,-0.1259385049,0.0388393067,-0.1479885429,0.0421098918,0.2073057145,0.3386696279,-0.4318103492,-0.0612153001,0.3661260307,0.3979151845,0.0629090369,0.7458975315,-0.2787281573,-0.1260559112,-0.0545617417,0.2708778977,0.0743895918,0.1198962778,-0.1902819723,-0.0999418497,0.1631213129,0.0208523776,0.0699715838,-0.1912690103,-0.0321693756,-0.2584814727,-0.2195187211,-0.5605755448,0.0462479331,-0.4587098658,0.2077327669,0.2398873866,0.2475411743,-0.2975123227,0.3902420402,-0.1294227093,0.1252297759,0.0979913473,0.1446118057,-0.391613096,-0.0055658114,-0.1123101413,0.0953805,-0.0835842416,0.1655615866,0.0677868202,-0.1269427538,-0.3280826509,0.1974897087,0.2874694169,-0.1786658019,-0.0433699936,0.0477794483,-0.1499832273,-0.3028782606,0.2485170662,-0.0607960597,0.1575934589,0.0763471574,0.3379111588,-0.2188207805,-0.1000690386,-0.0007012108,0.2162156254,0.0646288469,0.0848572403,0.1430528015,0.2670669854,-0.0898596719,-0.48067379,0.008267154,0.2503557205,0.0806295946,0.3912357688,0.1591602862,0.0410780683,0.0603678823,-0.2062233984,0.1170331538,0.0504666008,-0.3848752379,0.3746520281,-0.0060268939,-0.0551555119,-0.044036895,0.2743842006,-0.2653957903,-0.1859815419,-0.0595585145,0.4661214054,0.2951004207,-0.4041041136,-0.0919860825,0.3661443889,-0.2241643667,-0.0641895086,0.0971227139,-0.0098463167,0.0934667438,0.2680348456,0.0548743606,0.113316752,0.3695070148,0.3923889697,0.2883475721,0.2603395879,0.0777942762,0.1032693535,-0.2553917766,-0.2653571665,0.3103382885,-0.1531152129,0.08997982,0.0528870635,0.1000420749,0.5368420482,-0.354167223,0.0480193943,0.4134216905,-0.2640724182,-0.2332316339,-0.0627764463,-0.023408059,0.068245925,0.0958754122,-0.0718544871,-0.1540004611,-0.2346418947,-0.0370184705,-0.1833231896,-0.1952455342,0.1980551332,0.2319127172,0.198203221,-0.2359384745,0.0169511046,-0.3175613284,0.127129972,-0.2200936228,0.2020678818,0.3063046634,0.2241675109,-0.1465413272,-0.0788502097,0.1130340397,0.1072754413,-0.0500626713,0.0720541626,0.2317835391,-0.4880773425,-0.0415647924,-0.1164265126,-0.0517769866,-0.3074439466,0.3058531582,-0.197778374,0.0481348969,0.1242211685,-0.1056796685,-0.2300592512,-0.1648836732,0.2180307806,-0.4399427474,0.1330220103,0.3766916692,0.0617166534,0.3358033895,-0.4036953747,0.0198818762,-0.3701347113,0.39908728,0.5262007713,-0.1209670007,-0.2154575437,-0.2646383643,-0.1108805016,0.2902738452,-0.3585973084,0.0505906492,-0.1812591851,0.2633677125,-0.191850245,-0.3611743152,0.0252476707,-0.0790434554,-0.0080266064,0.1392814517,-0.0423979983,-0.0342022702,-0.412083745,0.1087527722,-0.1911482215,-0.0748468265,0.2083022892,0.0654212758,-0.1304049492,-0.1132040471,0.0713812262,0.019046247,0.0372061543,0.3966077268,0.2608486116,0.3634545207,-0.0893924087,0.2499944568,0.1179165468,-0.0126813119,-0.0652046353,0.2503354549,-0.2663959265,0.0027404695,0.1508235484,-0.0277987048,0.2109278888,0.0145359114,0.1021498516,-0.2304242998,-0.1275306195,0.0639622957,-0.1374926418,-0.0361369587,-0.1238750964,0.5886523724,0.0487280004,0.1767513156,-0.025978528,-0.216731593,0.1924982965,-0.531894207,-0.1194510683,-0.2614589632,-0.0022514537,0.2452489138,-0.269659251,-0.3578719795,-0.076712057,0.5602203012,0.0736324936,-0.0247230753,0.1855686158,0.1773032248,0.1805512607,-0.1497805417,0.3128385544,-0.276702106,0.0337501727,0.0683869496,-0.0230376646]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/257","title":"Tokenizer pickling issue fix not landed in `nlp` yet?","comments":"Yes, the new release of tokenizers solves this and should be out soon.\r\nIn the meantime, you can install it with `pip install tokenizers==0.8.0-dev2`","body":"Unless I recreate an arrow_dataset from my loaded nlp dataset myself (which I think does not use the cache by default), I get the following error when applying the map function:\r\n\r\n```\r\ndataset = nlp.load_dataset('cos_e')\r\ntokenizer = GPT2TokenizerFast.from_pretrained('gpt2', cache_dir=cache_dir)\r\n\r\nfor split in dataset.keys():\r\n dataset[split].map(lambda x: some_function(x, tokenizer))\r\n```\r\n```\r\n06\/09\/2020 10:09:19 - INFO - nlp.builder - Constructing Dataset for split train[:10], from \/home\/sarahw\/.cache\/huggingface\/datasets\/cos_e\/default\/0.0.1\r\nTraceback (most recent call last):\r\n File \"generation\/input_to_label_and_rationale.py\", line 390, in <module>\r\n main()\r\n File \"generation\/input_to_label_and_rationale.py\", line 263, in main\r\n dataset[split] = dataset[split].map(lambda x: input_to_explanation_plus_label(x, tokenizer, max_length, datasource=data_args.task_name, wt5=(model_class=='t5'), expl_only=model_args.rationale_only), batched=False)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/arrow_dataset.py\", line 522, in map\r\n cache_file_name = self._get_cache_file_path(function, cache_kwargs)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/arrow_dataset.py\", line 381, in _get_cache_file_path\r\n function_bytes = dumps(function)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 257, in dumps\r\n dump(obj, file)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 250, in dump\r\n Pickler(file).dump(obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 445, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 485, in dump\r\n self.save(obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1410, in save_function\r\n pickler.save_reduce(_create_function, (obj.__code__,\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1147, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 884, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 601, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 715, in save_reduce\r\n save(state)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 912, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 969, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 995, in _batch_setitems\r\n save(v)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 601, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 715, in save_reduce\r\n save(state)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 912, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 969, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 995, in _batch_setitems\r\n save(v)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 576, in save\r\n rv = reduce(self.proto)\r\nTypeError: cannot pickle 'Tokenizer' object\r\n```\r\nFix seems to be in the tokenizers [`0.8.0.dev1 pre-release`](https:\/\/github.com\/huggingface\/tokenizers\/issues\/87), which I can't install with any package managers. ","comment_length":24,"text":"Tokenizer pickling issue fix not landed in `nlp` yet?\nUnless I recreate an arrow_dataset from my loaded nlp dataset myself (which I think does not use the cache by default), I get the following error when applying the map function:\r\n\r\n```\r\ndataset = nlp.load_dataset('cos_e')\r\ntokenizer = GPT2TokenizerFast.from_pretrained('gpt2', cache_dir=cache_dir)\r\n\r\nfor split in dataset.keys():\r\n dataset[split].map(lambda x: some_function(x, tokenizer))\r\n```\r\n```\r\n06\/09\/2020 10:09:19 - INFO - nlp.builder - Constructing Dataset for split train[:10], from \/home\/sarahw\/.cache\/huggingface\/datasets\/cos_e\/default\/0.0.1\r\nTraceback (most recent call last):\r\n File \"generation\/input_to_label_and_rationale.py\", line 390, in <module>\r\n main()\r\n File \"generation\/input_to_label_and_rationale.py\", line 263, in main\r\n dataset[split] = dataset[split].map(lambda x: input_to_explanation_plus_label(x, tokenizer, max_length, datasource=data_args.task_name, wt5=(model_class=='t5'), expl_only=model_args.rationale_only), batched=False)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/arrow_dataset.py\", line 522, in map\r\n cache_file_name = self._get_cache_file_path(function, cache_kwargs)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/arrow_dataset.py\", line 381, in _get_cache_file_path\r\n function_bytes = dumps(function)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 257, in dumps\r\n dump(obj, file)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 250, in dump\r\n Pickler(file).dump(obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 445, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 485, in dump\r\n self.save(obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1410, in save_function\r\n pickler.save_reduce(_create_function, (obj.__code__,\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1147, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 884, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 601, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 715, in save_reduce\r\n save(state)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 912, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 969, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 995, in _batch_setitems\r\n save(v)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 601, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 715, in save_reduce\r\n save(state)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 912, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 969, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 995, in _batch_setitems\r\n save(v)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 576, in save\r\n rv = reduce(self.proto)\r\nTypeError: cannot pickle 'Tokenizer' object\r\n```\r\nFix seems to be in the tokenizers [`0.8.0.dev1 pre-release`](https:\/\/github.com\/huggingface\/tokenizers\/issues\/87), which I can't install with any package managers. \nYes, the new release of tokenizers solves this and should be out soon.\r\nIn the meantime, you can install it with `pip install tokenizers==0.8.0-dev2`","embeddings":[0.0099123437,0.1761874706,0.0797633827,0.0332191736,-0.0187639873,-0.0983517841,-0.0439764857,0.3858543932,0.2328621596,-0.178395614,0.1533451229,0.7815694809,-0.1890581548,-0.2719969749,0.0566341542,-0.155905962,0.0068416926,0.1128314584,0.3384636045,0.0179074295,-0.2215670645,0.3663436174,-0.2035544813,0.2977550626,-0.4966283739,-0.2166752219,0.2067759931,-0.0472208932,-0.1231692284,-0.5728648305,0.1870387495,-0.0740047246,-0.1319389343,0.0616183169,-0.0001174416,-0.1495584548,0.3925046623,0.0270643979,-0.1482364833,-0.2368089408,0.1705570519,-0.0939855054,0.1326126754,-0.0232695937,-0.0295413285,-0.0769911557,0.1083666608,-0.1098065153,0.3249566257,-0.0344230719,0.1912767291,0.3061876297,0.1382566988,0.010819071,0.0969163477,-0.1552183479,-0.1684094965,-0.0005967876,0.090119198,-0.2653211653,-0.2927828133,0.4738675952,0.0086574946,-0.0582620651,0.06942375,0.0213256795,-0.1785932332,-0.0757203773,0.0937063918,-0.0271070879,0.0955460519,-0.185951829,-0.0445755795,-0.4417182207,-0.1290053278,-0.4550065696,0.1827386171,-0.1479194164,-0.2384156585,-0.044095207,-0.0090401918,-0.3884577453,0.1388706863,0.305347234,0.1289279908,0.6000414491,-0.0572790317,0.2780373394,0.2819808722,-0.0880403817,-0.2382824272,0.1567702591,-0.1878780276,0.4356526732,-0.3443645239,-0.1178101301,0.2205701023,0.0535372719,-0.0640307739,0.3975797296,-0.0016291763,0.2429233938,-0.1206074208,0.1644044071,0.0477179624,0.4591069818,0.3047444224,0.5499494672,0.0105261905,-0.3980909884,-0.0375846587,-0.0503841192,-0.0753701106,-0.2622773945,0.0318233259,0.0509526394,-0.117181316,-0.104768239,-0.1330924928,0.0608898513,-0.2686992288,0.0647368804,0.2208596468,0.1958482265,0.1352951229,0.1504516751,-0.1320771724,0.1724413484,-0.1877634674,-0.1661207825,-0.248142451,0.0797850415,-0.3150077462,0.181962356,0.097306259,0.2771179378,0.480322212,-0.1688139588,-0.0587520488,0.0628477111,0.1248926893,-0.0092322603,0.2322514802,0.0725742429,-0.1473613679,0.3290615082,0.3126963675,-0.141333133,-0.2816562951,-0.051630348,-0.2734317183,-0.3626489639,-0.1152614132,0.1254754364,-0.0622648075,-0.1384949982,-0.271987617,0.2420289963,0.4594833851,-0.1889908016,0.1101836562,-0.1990117133,-0.3264870942,-0.2962644696,0.0307122543,0.2880884409,0.0484273285,-0.3790308535,-0.1973203868,0.152703166,0.4366891086,0.5128081441,-0.3213967681,0.1518722922,-0.2618008852,0.814293623,0.5736558437,-0.2508249879,-0.4575694203,0.0358645134,-0.1898695827,0.0828077123,-0.2472429127,-0.0829209462,-0.0119433152,-0.1429459453,0.179187119,0.2489403337,-0.008139831,0.0076341168,-0.3845753372,-0.042985931,0.3610703051,-0.040211156,0.1359008253,-0.1101026833,-0.2231208682,0.0138509758,0.0019649428,-0.0150994463,0.3447351456,-0.0740484223,0.1137956753,-0.1738644242,0.0757068768,0.0411160365,-0.3737537563,0.1523695588,-0.2930880189,0.1511403322,-0.2010424584,0.021357473,-0.1287766695,0.0444968641,-0.3398889005,-0.233027786,0.1284675449,0.0046516308,0.1984073669,0.0344288088,-0.0145979878,-0.0610454157,0.0627615601,0.1549107283,-0.1303329468,-0.0565437041,-0.0467760414,-0.2752353847,-0.2991557121,0.1600325555,0.2399732471,-0.0254261903,-0.2603600323,0.294133842,0.0070313257,0.0486650318,-0.1207752451,-0.0605716966,0.0970292762,-0.1387098879,-0.1103155315,0.124253653,-0.0198693071,-0.2519939244,0.3183259368,0.3542825282,0.1884092987,0.0600378774,-0.2868296802,0.2985999286,0.0610981956,-0.0173833854,0.1314131021,-0.2714962363,0.066673331,-0.1250588149,0.2938562632,-0.1201208532,-0.0217081122,0.3552860916,0.3217470646,0.3191149235,0.0601604953,0.1106171981,-0.2048500776,-0.2746075094,0.0948237032,0.1242748946,0.1567907631,0.1859988868,0.0423683301,-0.0587046333,-0.1060367897,-0.0883650258,0.1340711117,0.0121885845,0.0686790571,0.159634769,0.0601588152,0.1171429902,-0.3626596928,0.0093214056,0.1092613563,0.3128890097,-0.1123162508,-0.0685740262,-0.3171990812,-0.0965782776,0.1861843467,-0.3486100435,-0.2434670776,-0.4538672864,0.1318599135,0.1396827549,-0.1204633787,0.4497210681,-0.1803721189,0.2464598715,0.0291105229,-0.0901661292,-0.3170962036,-0.5001772046,-0.3929176927,-0.0020681028,0.1086931378,-0.0656516925,0.1988438815,0.1700224876,0.0388037674,-0.0799686238,-0.3938218653,0.224536106,-0.1287586391,0.048669856,0.2407984734,-0.2067009807,-0.4435534477,-0.3115964532,0.1795250326,-0.3477775156,-0.3655576408,-0.1269761026,0.0443097316,-0.1161485836,-0.1567027569,-0.0487882085,-0.1834051311,-0.2831899822,0.2769730687,0.0246919971,0.1811643243,0.4258752167,-0.3476061225,0.0534039997,-0.3079631329,-0.1246998087,-0.5135576129,0.0901808962,0.3106078804,-0.320093751,-0.1387337446,-0.1729266346,-0.1581386179,0.3402467668,0.0410679355,-0.1939451247,0.1122732684,-0.0858655423,0.1610163599,-0.1077982262,-0.0623691827,0.4786886871,0.0970861316,-0.0032040772,-0.1829605103,-0.1646911353,0.4114802182,-0.0680685937,0.3304801583,0.0873666406,0.1052733064,0.2931070924,0.8183516264,0.2056207508,-0.1043714285,0.4881397486,0.233660996,0.0024523013,-0.2040912062,-0.2489936501,0.1088703945,-0.0217352081,0.3350535631,0.0500925519,0.0086000394,-0.4172491133,0.120437704,0.1518307179,-0.1542117,-0.2865460217,0.1852338761,-0.3836121857,0.1347557008,0.1777047366,0.0442596115,-0.5306248665,-0.1345783621,0.1172003746,-0.2568663657,0.3932621479,0.0826739147,-0.5393274426,0.0516309179,-0.7337316871,0.2864343524,0.2413739264,0.395411402,-0.1988530755,-0.173621133,0.0418537781,0.0748715401,0.3452197015,-0.1550505608,-0.0552804247,0.1675938666,0.0694346502,-0.135293141,0.1023075432,0.2772547603,0.5168840885,0.4541476071,0.3920671344,-0.3881210089,-0.2562205195,-0.1675943136,-0.1744462848,-0.0186568461,0.0365911275,0.0293230452,0.1137141362,-0.1721294224,0.1507237852,0.2761496007,0.338830322,0.3948661685,0.1762166023,-0.1165319607,-0.1274839342,-0.1879188567,0.1366696954,0.3615833223,0.0326662026,0.0474168211,0.2189311534,0.1186931729,-0.2655065954,0.1644708067,0.2509417534,-0.1118044183,0.16112037,0.2835705876,0.1217287332,0.2762947083,0.0352720916,0.4669424593,-0.4396469593,0.2055474669,0.1118125096,0.4023244381,0.0656892955,-0.0535079427,-0.2024742812,-0.0851320848,0.3267726898,0.125105083,-0.217855528,0.5619816184,-0.0305400882,-0.3434184194,0.4547913074,0.1010717377,0.8719286919,0.2725573778,0.2011611462,0.2368342876,0.0563699044,0.6379817724,0.0370739438,0.1459992528,-0.3191590011,0.3665098548,0.0357395113,-0.0379595421,-0.1278351098,0.0960807428,-0.0499115027,0.2048881799,0.239455536,0.2010641098,0.2482946813,0.1110169291,-0.0012670546,-0.3541906178,0.1831119359,0.0841663703,-0.3797288537,0.3635558486,0.0432655923,-0.1476145536,-0.0283209905,-0.3495350778,-0.2687566876,0.1171247214,-0.4619830847,0.390971452,0.4313520193,-0.207949996,0.1729608029,-0.1405204535,0.1785715222,-0.0680005252,-0.1257000118,0.3717615008,0.28187415,-0.0578295998,0.0317209437,-0.1306055188,0.2797636986,0.030441802,0.001976619,-0.1016213447,-0.1483108401,-0.4747525752,-0.4060358107,0.0260358769,0.1796491146,0.1115586534,-0.2101286948,0.133970961,-0.0188679732,-0.2450491339,0.0894215703,0.1031866297,-0.1743165106,-0.0313140936,0.107783556,-0.1305726916,0.0892200693,0.2922829688,-0.2033206522,-0.1827051044,0.5570828915,-0.1918235123,-0.1815356463,-0.1592898071,-0.174697414,0.2314829826,-0.4645784199,0.1009742618,-0.0894128829,-0.121676147,0.0410831235,0.3308238685,0.1877301335,0.1735885888,0.0708826333,-0.2569064498,-0.2817651927,0.1632193476,-0.5082813501,0.0636172071,0.0004067882,0.3584066331,-0.2887223661,0.3904939592,-0.2625270486,0.0943123773,-0.0914544761,0.0905122533,-0.0564934649,-0.3398321271,-0.0600878224,0.0083680376,-0.0068506976,0.4153741896,-0.2501581311,-0.164799273,-0.2339626253,0.1227046326,0.3353522122,-0.3256459832,-0.0476962738,0.2404371351,0.0712321177,-0.1416354477,-0.0511782989,-0.0344138667,-0.0987120941,0.1077831537,0.0902803764,0.0484989993,0.2448071092,0.1720362157,0.0000043339,0.0116087748,0.103373304,0.3003525138,-0.0566956587,0.0499827862,0.0096726269,0.0549555942,-0.0547022261,-0.1061150581,0.1037990525,-0.4290778041,-0.2767672837,-0.1523224413,0.051454097,0.0842219666,-0.0090162754,-0.4086213112,0.5563294888,0.177757144,-0.1644624621,-0.0730643198,0.4187648296,0.2304314673,0.0582475476,0.0615208931,0.2833521366,-0.096971713,0.0829341337,0.0213811267,0.499332875,-0.4074431658,0.1961526275,-0.0075245402,-0.0538639985,0.0389791019,0.5023404956,0.1022214144,0.2201762497,0.0315361694,-0.0683617741,0.3054454923,0.099632971,0.2393583059,-0.1012043878,-0.3959549367,-0.3370906711,0.3917880654,-0.0488352813,0.2234889269,0.0223838016,0.5310403705,-0.0221083537,-0.4205748439,0.1457373053,0.534075141,-0.3441424966,-0.3428340256,-0.4735886753,0.0072986092,-0.3503251076,0.0235188976,-0.0265383869,-0.0413034633,0.007149477,-0.0790733173,-0.1581784189,-0.3293400109,-0.2359717935,0.0574537069,0.2912587821,-0.2471500784,0.4018816948,-0.4688897431,-0.217081219,-0.2654462457,0.1990816444,0.2948338091,0.244736433,-0.4160876572,-0.2237972617,0.0119639235,0.1271606237,-0.1225386411,0.3143150806,0.2412485033,-0.3868137896,0.1421112418,-0.0401093401,-0.1103739366,0.0310608242,0.1389701068,0.222957328,0.3279425204,0.1659200042,-0.1513467729,-0.0700433701,-0.051429756,0.0774747059,-0.223693043,0.1334914416,0.022824904,0.0422992408,0.1793929487,-0.324831754,0.0863499418,0.1088384539,0.4186912775,0.3605180681,-0.2678305805,-0.4015859067,-0.4270046949,-0.3972657323,0.1965407878,-0.0541454852,-0.0473720096,-0.0405917242,0.0586175174,0.0444418862,-0.0199996606,-0.1949566901,0.4008176923,0.1366796643,0.1745637655,-0.3776068687,0.4529934227,0.0952504426,-0.0781205297,0.2346618772,-0.3071935475,0.0148459869,0.100093253,0.0566659942,-0.151372388,-0.1251380444,-0.0407404304,0.2070125639,0.2774702013,0.332254082,0.2022197694,0.0640883967,0.1838990897,-0.3896707892,-0.3934141099,-0.2413275391,-0.1081641614,-0.1865242422,0.4314035475,-0.0296274666,0.2332597822,-0.4199026227,0.0125968214,-0.0292735714,0.0756558776,-0.1747145206,0.3251663446,-0.2844869792,0.2631872594,-0.0517764725,0.3642563522,-0.0152900144,0.4393196702,-0.4197939336,-0.4348351657,0.5731081367,-0.7418988943,-0.2233948559,0.2329641283,0.0076883072,0.5054789186,-0.3683009446,-0.193344295,-0.2531439662,0.2136694938,-0.0410078503,-0.2464897037,0.050612662,-0.0451937281,-0.0969157368,-0.0427862369,0.6299054623,0.1238593236,-0.1319498271,0.1468424052,-0.171857819]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/257","title":"Tokenizer pickling issue fix not landed in `nlp` yet?","comments":"If others run into this issue, a quick fix is to use python 3.6 instead of 3.7+. Serialization differences between the 3rd party `dataclasses` package for 3.6 and the built in `dataclasses` in 3.7+ cause the issue.\r\n\r\nProbably a dumb fix, but it works for me.","body":"Unless I recreate an arrow_dataset from my loaded nlp dataset myself (which I think does not use the cache by default), I get the following error when applying the map function:\r\n\r\n```\r\ndataset = nlp.load_dataset('cos_e')\r\ntokenizer = GPT2TokenizerFast.from_pretrained('gpt2', cache_dir=cache_dir)\r\n\r\nfor split in dataset.keys():\r\n dataset[split].map(lambda x: some_function(x, tokenizer))\r\n```\r\n```\r\n06\/09\/2020 10:09:19 - INFO - nlp.builder - Constructing Dataset for split train[:10], from \/home\/sarahw\/.cache\/huggingface\/datasets\/cos_e\/default\/0.0.1\r\nTraceback (most recent call last):\r\n File \"generation\/input_to_label_and_rationale.py\", line 390, in <module>\r\n main()\r\n File \"generation\/input_to_label_and_rationale.py\", line 263, in main\r\n dataset[split] = dataset[split].map(lambda x: input_to_explanation_plus_label(x, tokenizer, max_length, datasource=data_args.task_name, wt5=(model_class=='t5'), expl_only=model_args.rationale_only), batched=False)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/arrow_dataset.py\", line 522, in map\r\n cache_file_name = self._get_cache_file_path(function, cache_kwargs)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/arrow_dataset.py\", line 381, in _get_cache_file_path\r\n function_bytes = dumps(function)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 257, in dumps\r\n dump(obj, file)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 250, in dump\r\n Pickler(file).dump(obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 445, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 485, in dump\r\n self.save(obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1410, in save_function\r\n pickler.save_reduce(_create_function, (obj.__code__,\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1147, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 884, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 601, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 715, in save_reduce\r\n save(state)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 912, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 969, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 995, in _batch_setitems\r\n save(v)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 601, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 715, in save_reduce\r\n save(state)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 912, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 969, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 995, in _batch_setitems\r\n save(v)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 576, in save\r\n rv = reduce(self.proto)\r\nTypeError: cannot pickle 'Tokenizer' object\r\n```\r\nFix seems to be in the tokenizers [`0.8.0.dev1 pre-release`](https:\/\/github.com\/huggingface\/tokenizers\/issues\/87), which I can't install with any package managers. ","comment_length":46,"text":"Tokenizer pickling issue fix not landed in `nlp` yet?\nUnless I recreate an arrow_dataset from my loaded nlp dataset myself (which I think does not use the cache by default), I get the following error when applying the map function:\r\n\r\n```\r\ndataset = nlp.load_dataset('cos_e')\r\ntokenizer = GPT2TokenizerFast.from_pretrained('gpt2', cache_dir=cache_dir)\r\n\r\nfor split in dataset.keys():\r\n dataset[split].map(lambda x: some_function(x, tokenizer))\r\n```\r\n```\r\n06\/09\/2020 10:09:19 - INFO - nlp.builder - Constructing Dataset for split train[:10], from \/home\/sarahw\/.cache\/huggingface\/datasets\/cos_e\/default\/0.0.1\r\nTraceback (most recent call last):\r\n File \"generation\/input_to_label_and_rationale.py\", line 390, in <module>\r\n main()\r\n File \"generation\/input_to_label_and_rationale.py\", line 263, in main\r\n dataset[split] = dataset[split].map(lambda x: input_to_explanation_plus_label(x, tokenizer, max_length, datasource=data_args.task_name, wt5=(model_class=='t5'), expl_only=model_args.rationale_only), batched=False)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/arrow_dataset.py\", line 522, in map\r\n cache_file_name = self._get_cache_file_path(function, cache_kwargs)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/arrow_dataset.py\", line 381, in _get_cache_file_path\r\n function_bytes = dumps(function)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 257, in dumps\r\n dump(obj, file)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/nlp\/utils\/py_utils.py\", line 250, in dump\r\n Pickler(file).dump(obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 445, in dump\r\n StockPickler.dump(self, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 485, in dump\r\n self.save(obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1410, in save_function\r\n pickler.save_reduce(_create_function, (obj.__code__,\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 899, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 1147, in save_cell\r\n pickler.save_reduce(_create_cell, (f,), obj=obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 690, in save_reduce\r\n save(args)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 884, in save_tuple\r\n save(element)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 601, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 715, in save_reduce\r\n save(state)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 912, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 969, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 995, in _batch_setitems\r\n save(v)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 601, in save\r\n self.save_reduce(obj=obj, *rv)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 715, in save_reduce\r\n save(state)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 558, in save\r\n f(self, obj) # Call unbound method with explicit self\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/site-packages\/dill\/_dill.py\", line 912, in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 969, in save_dict\r\n self._batch_setitems(obj.items())\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 995, in _batch_setitems\r\n save(v)\r\n File \"\/home\/sarahw\/miniconda3\/envs\/project_huggingface\/lib\/python3.8\/pickle.py\", line 576, in save\r\n rv = reduce(self.proto)\r\nTypeError: cannot pickle 'Tokenizer' object\r\n```\r\nFix seems to be in the tokenizers [`0.8.0.dev1 pre-release`](https:\/\/github.com\/huggingface\/tokenizers\/issues\/87), which I can't install with any package managers. \nIf others run into this issue, a quick fix is to use python 3.6 instead of 3.7+. Serialization differences between the 3rd party `dataclasses` package for 3.6 and the built in `dataclasses` in 3.7+ cause the issue.\r\n\r\nProbably a dumb fix, but it works for me.","embeddings":[0.0099123437,0.1761874706,0.0797633827,0.0332191736,-0.0187639873,-0.0983517841,-0.0439764857,0.3858543932,0.2328621596,-0.178395614,0.1533451229,0.7815694809,-0.1890581548,-0.2719969749,0.0566341542,-0.155905962,0.0068416926,0.1128314584,0.3384636045,0.0179074295,-0.2215670645,0.3663436174,-0.2035544813,0.2977550626,-0.4966283739,-0.2166752219,0.2067759931,-0.0472208932,-0.1231692284,-0.5728648305,0.1870387495,-0.0740047246,-0.1319389343,0.0616183169,-0.0001174416,-0.1495584548,0.3925046623,0.0270643979,-0.1482364833,-0.2368089408,0.1705570519,-0.0939855054,0.1326126754,-0.0232695937,-0.0295413285,-0.0769911557,0.1083666608,-0.1098065153,0.3249566257,-0.0344230719,0.1912767291,0.3061876297,0.1382566988,0.010819071,0.0969163477,-0.1552183479,-0.1684094965,-0.0005967876,0.090119198,-0.2653211653,-0.2927828133,0.4738675952,0.0086574946,-0.0582620651,0.06942375,0.0213256795,-0.1785932332,-0.0757203773,0.0937063918,-0.0271070879,0.0955460519,-0.185951829,-0.0445755795,-0.4417182207,-0.1290053278,-0.4550065696,0.1827386171,-0.1479194164,-0.2384156585,-0.044095207,-0.0090401918,-0.3884577453,0.1388706863,0.305347234,0.1289279908,0.6000414491,-0.0572790317,0.2780373394,0.2819808722,-0.0880403817,-0.2382824272,0.1567702591,-0.1878780276,0.4356526732,-0.3443645239,-0.1178101301,0.2205701023,0.0535372719,-0.0640307739,0.3975797296,-0.0016291763,0.2429233938,-0.1206074208,0.1644044071,0.0477179624,0.4591069818,0.3047444224,0.5499494672,0.0105261905,-0.3980909884,-0.0375846587,-0.0503841192,-0.0753701106,-0.2622773945,0.0318233259,0.0509526394,-0.117181316,-0.104768239,-0.1330924928,0.0608898513,-0.2686992288,0.0647368804,0.2208596468,0.1958482265,0.1352951229,0.1504516751,-0.1320771724,0.1724413484,-0.1877634674,-0.1661207825,-0.248142451,0.0797850415,-0.3150077462,0.181962356,0.097306259,0.2771179378,0.480322212,-0.1688139588,-0.0587520488,0.0628477111,0.1248926893,-0.0092322603,0.2322514802,0.0725742429,-0.1473613679,0.3290615082,0.3126963675,-0.141333133,-0.2816562951,-0.051630348,-0.2734317183,-0.3626489639,-0.1152614132,0.1254754364,-0.0622648075,-0.1384949982,-0.271987617,0.2420289963,0.4594833851,-0.1889908016,0.1101836562,-0.1990117133,-0.3264870942,-0.2962644696,0.0307122543,0.2880884409,0.0484273285,-0.3790308535,-0.1973203868,0.152703166,0.4366891086,0.5128081441,-0.3213967681,0.1518722922,-0.2618008852,0.814293623,0.5736558437,-0.2508249879,-0.4575694203,0.0358645134,-0.1898695827,0.0828077123,-0.2472429127,-0.0829209462,-0.0119433152,-0.1429459453,0.179187119,0.2489403337,-0.008139831,0.0076341168,-0.3845753372,-0.042985931,0.3610703051,-0.040211156,0.1359008253,-0.1101026833,-0.2231208682,0.0138509758,0.0019649428,-0.0150994463,0.3447351456,-0.0740484223,0.1137956753,-0.1738644242,0.0757068768,0.0411160365,-0.3737537563,0.1523695588,-0.2930880189,0.1511403322,-0.2010424584,0.021357473,-0.1287766695,0.0444968641,-0.3398889005,-0.233027786,0.1284675449,0.0046516308,0.1984073669,0.0344288088,-0.0145979878,-0.0610454157,0.0627615601,0.1549107283,-0.1303329468,-0.0565437041,-0.0467760414,-0.2752353847,-0.2991557121,0.1600325555,0.2399732471,-0.0254261903,-0.2603600323,0.294133842,0.0070313257,0.0486650318,-0.1207752451,-0.0605716966,0.0970292762,-0.1387098879,-0.1103155315,0.124253653,-0.0198693071,-0.2519939244,0.3183259368,0.3542825282,0.1884092987,0.0600378774,-0.2868296802,0.2985999286,0.0610981956,-0.0173833854,0.1314131021,-0.2714962363,0.066673331,-0.1250588149,0.2938562632,-0.1201208532,-0.0217081122,0.3552860916,0.3217470646,0.3191149235,0.0601604953,0.1106171981,-0.2048500776,-0.2746075094,0.0948237032,0.1242748946,0.1567907631,0.1859988868,0.0423683301,-0.0587046333,-0.1060367897,-0.0883650258,0.1340711117,0.0121885845,0.0686790571,0.159634769,0.0601588152,0.1171429902,-0.3626596928,0.0093214056,0.1092613563,0.3128890097,-0.1123162508,-0.0685740262,-0.3171990812,-0.0965782776,0.1861843467,-0.3486100435,-0.2434670776,-0.4538672864,0.1318599135,0.1396827549,-0.1204633787,0.4497210681,-0.1803721189,0.2464598715,0.0291105229,-0.0901661292,-0.3170962036,-0.5001772046,-0.3929176927,-0.0020681028,0.1086931378,-0.0656516925,0.1988438815,0.1700224876,0.0388037674,-0.0799686238,-0.3938218653,0.224536106,-0.1287586391,0.048669856,0.2407984734,-0.2067009807,-0.4435534477,-0.3115964532,0.1795250326,-0.3477775156,-0.3655576408,-0.1269761026,0.0443097316,-0.1161485836,-0.1567027569,-0.0487882085,-0.1834051311,-0.2831899822,0.2769730687,0.0246919971,0.1811643243,0.4258752167,-0.3476061225,0.0534039997,-0.3079631329,-0.1246998087,-0.5135576129,0.0901808962,0.3106078804,-0.320093751,-0.1387337446,-0.1729266346,-0.1581386179,0.3402467668,0.0410679355,-0.1939451247,0.1122732684,-0.0858655423,0.1610163599,-0.1077982262,-0.0623691827,0.4786886871,0.0970861316,-0.0032040772,-0.1829605103,-0.1646911353,0.4114802182,-0.0680685937,0.3304801583,0.0873666406,0.1052733064,0.2931070924,0.8183516264,0.2056207508,-0.1043714285,0.4881397486,0.233660996,0.0024523013,-0.2040912062,-0.2489936501,0.1088703945,-0.0217352081,0.3350535631,0.0500925519,0.0086000394,-0.4172491133,0.120437704,0.1518307179,-0.1542117,-0.2865460217,0.1852338761,-0.3836121857,0.1347557008,0.1777047366,0.0442596115,-0.5306248665,-0.1345783621,0.1172003746,-0.2568663657,0.3932621479,0.0826739147,-0.5393274426,0.0516309179,-0.7337316871,0.2864343524,0.2413739264,0.395411402,-0.1988530755,-0.173621133,0.0418537781,0.0748715401,0.3452197015,-0.1550505608,-0.0552804247,0.1675938666,0.0694346502,-0.135293141,0.1023075432,0.2772547603,0.5168840885,0.4541476071,0.3920671344,-0.3881210089,-0.2562205195,-0.1675943136,-0.1744462848,-0.0186568461,0.0365911275,0.0293230452,0.1137141362,-0.1721294224,0.1507237852,0.2761496007,0.338830322,0.3948661685,0.1762166023,-0.1165319607,-0.1274839342,-0.1879188567,0.1366696954,0.3615833223,0.0326662026,0.0474168211,0.2189311534,0.1186931729,-0.2655065954,0.1644708067,0.2509417534,-0.1118044183,0.16112037,0.2835705876,0.1217287332,0.2762947083,0.0352720916,0.4669424593,-0.4396469593,0.2055474669,0.1118125096,0.4023244381,0.0656892955,-0.0535079427,-0.2024742812,-0.0851320848,0.3267726898,0.125105083,-0.217855528,0.5619816184,-0.0305400882,-0.3434184194,0.4547913074,0.1010717377,0.8719286919,0.2725573778,0.2011611462,0.2368342876,0.0563699044,0.6379817724,0.0370739438,0.1459992528,-0.3191590011,0.3665098548,0.0357395113,-0.0379595421,-0.1278351098,0.0960807428,-0.0499115027,0.2048881799,0.239455536,0.2010641098,0.2482946813,0.1110169291,-0.0012670546,-0.3541906178,0.1831119359,0.0841663703,-0.3797288537,0.3635558486,0.0432655923,-0.1476145536,-0.0283209905,-0.3495350778,-0.2687566876,0.1171247214,-0.4619830847,0.390971452,0.4313520193,-0.207949996,0.1729608029,-0.1405204535,0.1785715222,-0.0680005252,-0.1257000118,0.3717615008,0.28187415,-0.0578295998,0.0317209437,-0.1306055188,0.2797636986,0.030441802,0.001976619,-0.1016213447,-0.1483108401,-0.4747525752,-0.4060358107,0.0260358769,0.1796491146,0.1115586534,-0.2101286948,0.133970961,-0.0188679732,-0.2450491339,0.0894215703,0.1031866297,-0.1743165106,-0.0313140936,0.107783556,-0.1305726916,0.0892200693,0.2922829688,-0.2033206522,-0.1827051044,0.5570828915,-0.1918235123,-0.1815356463,-0.1592898071,-0.174697414,0.2314829826,-0.4645784199,0.1009742618,-0.0894128829,-0.121676147,0.0410831235,0.3308238685,0.1877301335,0.1735885888,0.0708826333,-0.2569064498,-0.2817651927,0.1632193476,-0.5082813501,0.0636172071,0.0004067882,0.3584066331,-0.2887223661,0.3904939592,-0.2625270486,0.0943123773,-0.0914544761,0.0905122533,-0.0564934649,-0.3398321271,-0.0600878224,0.0083680376,-0.0068506976,0.4153741896,-0.2501581311,-0.164799273,-0.2339626253,0.1227046326,0.3353522122,-0.3256459832,-0.0476962738,0.2404371351,0.0712321177,-0.1416354477,-0.0511782989,-0.0344138667,-0.0987120941,0.1077831537,0.0902803764,0.0484989993,0.2448071092,0.1720362157,0.0000043339,0.0116087748,0.103373304,0.3003525138,-0.0566956587,0.0499827862,0.0096726269,0.0549555942,-0.0547022261,-0.1061150581,0.1037990525,-0.4290778041,-0.2767672837,-0.1523224413,0.051454097,0.0842219666,-0.0090162754,-0.4086213112,0.5563294888,0.177757144,-0.1644624621,-0.0730643198,0.4187648296,0.2304314673,0.0582475476,0.0615208931,0.2833521366,-0.096971713,0.0829341337,0.0213811267,0.499332875,-0.4074431658,0.1961526275,-0.0075245402,-0.0538639985,0.0389791019,0.5023404956,0.1022214144,0.2201762497,0.0315361694,-0.0683617741,0.3054454923,0.099632971,0.2393583059,-0.1012043878,-0.3959549367,-0.3370906711,0.3917880654,-0.0488352813,0.2234889269,0.0223838016,0.5310403705,-0.0221083537,-0.4205748439,0.1457373053,0.534075141,-0.3441424966,-0.3428340256,-0.4735886753,0.0072986092,-0.3503251076,0.0235188976,-0.0265383869,-0.0413034633,0.007149477,-0.0790733173,-0.1581784189,-0.3293400109,-0.2359717935,0.0574537069,0.2912587821,-0.2471500784,0.4018816948,-0.4688897431,-0.217081219,-0.2654462457,0.1990816444,0.2948338091,0.244736433,-0.4160876572,-0.2237972617,0.0119639235,0.1271606237,-0.1225386411,0.3143150806,0.2412485033,-0.3868137896,0.1421112418,-0.0401093401,-0.1103739366,0.0310608242,0.1389701068,0.222957328,0.3279425204,0.1659200042,-0.1513467729,-0.0700433701,-0.051429756,0.0774747059,-0.223693043,0.1334914416,0.022824904,0.0422992408,0.1793929487,-0.324831754,0.0863499418,0.1088384539,0.4186912775,0.3605180681,-0.2678305805,-0.4015859067,-0.4270046949,-0.3972657323,0.1965407878,-0.0541454852,-0.0473720096,-0.0405917242,0.0586175174,0.0444418862,-0.0199996606,-0.1949566901,0.4008176923,0.1366796643,0.1745637655,-0.3776068687,0.4529934227,0.0952504426,-0.0781205297,0.2346618772,-0.3071935475,0.0148459869,0.100093253,0.0566659942,-0.151372388,-0.1251380444,-0.0407404304,0.2070125639,0.2774702013,0.332254082,0.2022197694,0.0640883967,0.1838990897,-0.3896707892,-0.3934141099,-0.2413275391,-0.1081641614,-0.1865242422,0.4314035475,-0.0296274666,0.2332597822,-0.4199026227,0.0125968214,-0.0292735714,0.0756558776,-0.1747145206,0.3251663446,-0.2844869792,0.2631872594,-0.0517764725,0.3642563522,-0.0152900144,0.4393196702,-0.4197939336,-0.4348351657,0.5731081367,-0.7418988943,-0.2233948559,0.2329641283,0.0076883072,0.5054789186,-0.3683009446,-0.193344295,-0.2531439662,0.2136694938,-0.0410078503,-0.2464897037,0.050612662,-0.0451937281,-0.0969157368,-0.0427862369,0.6299054623,0.1238593236,-0.1319498271,0.1468424052,-0.171857819]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/256","title":"[Feature request] Add a feature to dataset","comments":"Do you have an example of what you would like to do? (you can just add a field in the output of the unction you give to map and this will add this field in the output table)","body":"Is there a straightforward way to add a field to the arrow_dataset, prior to performing map?","comment_length":38,"text":"[Feature request] Add a feature to dataset\nIs there a straightforward way to add a field to the arrow_dataset, prior to performing map?\nDo you have an example of what you would like to do? (you can just add a field in the output of the unction you give to map and this will add this field in the output table)","embeddings":[-0.3230097592,-0.125670135,-0.1297954321,-0.3137857914,0.0523874797,0.1927578896,0.0315824375,0.0829818398,0.1852330863,0.0769858584,0.6435584426,0.7557856441,-0.0890218318,0.4095419049,0.2446237057,0.032614138,0.0282653682,0.2254813313,0.0143262055,0.0299326628,-0.3113274574,-0.3316054344,0.0440422781,-0.1416705102,-0.2610886395,-0.1055848822,0.0917241275,-0.2881222665,-0.2672578692,-0.4562530518,-0.0757000521,0.0525777973,-0.0076457122,-0.2505634129,-0.0001156606,-0.247966975,0.1716554314,0.0429585762,-0.1068635136,-0.099915199,-0.652341783,-0.46455881,0.2334650755,-0.1411959231,-0.0307949446,-0.0986400843,0.0351928398,-0.3240500391,-0.0623200461,-0.0979386419,0.2118946612,-0.1266070306,0.1847655624,-0.0503193736,0.2521041632,0.4834830463,-0.369220674,-0.2472529262,0.0324977897,0.0674415156,0.1742890328,0.3334390819,0.2909506559,-0.2479885072,0.4152329862,0.1051029712,0.0374160185,-0.1611279547,-0.0902455524,0.1184934154,0.7849738002,-0.2754909098,-0.2820773423,0.0049742302,0.120862186,-0.3142980933,-0.0033539662,-0.0835787952,0.1403281838,0.0212115422,0.2073623836,-0.548999846,-0.0674224645,0.1288884878,0.1251959801,0.3817350864,0.0620457865,0.115920037,-0.1173211113,-0.1588683873,0.0172910094,-0.0275660828,-0.2013939768,0.5192962289,-0.2226473689,-0.2803495526,0.2294760495,-0.3391990066,0.0392387472,0.1831898838,0.184876278,0.0174749363,-0.3253744543,0.174706772,0.1322717518,-0.0281038992,0.2036432475,0.1825891435,0.0067139035,-0.4320653975,-0.0406626016,-0.3651522994,0.2856080234,-0.0994315073,0.2918941677,0.0474699661,0.4299737513,-0.1470286846,0.2040187865,-0.1468968093,0.1710050553,-0.0374894105,0.0053038467,0.0882869512,0.2963477373,-0.1005245596,-0.0368408151,0.2756737471,0.3390971422,0.0404423624,-0.0502534881,-0.3233454525,0.0756810755,0.3453169763,0.0271619223,0.204601109,0.1169341281,-0.0236117169,-0.0765623599,0.0535558946,0.273286134,-0.0373285264,0.2673544884,0.2196780592,-0.1547069699,0.1350218654,-0.0075077419,0.4994907677,-0.2631914318,0.3902081847,-0.1304592341,-0.2942380011,-0.1127515808,0.1767225266,0.0951572433,-0.3213917315,-0.2972108722,0.3748586178,-0.0889949277,-0.1698729694,0.2799259126,0.3327130973,-0.248543337,-0.3037452102,-0.068134971,0.1408743709,-0.1992450058,0.0187904555,-0.421851635,-0.0745439306,-0.2238362879,-0.4167459309,-0.1549995542,0.3106938899,0.0001319322,0.3453457952,0.8380850554,0.0523078553,-0.1164599136,0.1494684517,-0.4384101331,-0.1004165038,-0.2182468027,0.5195293427,0.2225422859,0.1640651673,-0.0957169309,0.3228781521,-0.2440561503,-0.0722122341,0.0837466121,-0.0319254696,0.1864677519,0.079782024,-0.1155349016,0.0723225921,0.3547309637,-0.3399245143,-0.0757753849,0.2287126482,0.1548018754,0.0624214709,0.6634558439,-0.1903357953,-0.1494433284,-0.2541691959,-0.3794186711,0.0446260683,0.0183061734,-0.2327300757,-0.0779548734,-0.363858372,0.1358209848,0.3931482136,-0.0465489514,0.1942265481,0.0388862118,-0.291077286,-0.2459311485,-0.127122432,-0.2429455668,-0.286403507,-0.0538008511,0.0218144748,0.0602553673,-0.0607900172,0.1765721291,-0.0910034254,-0.3659125268,0.1023856401,0.0408974178,0.0884168819,-0.2122453004,0.2533627152,0.1685374975,0.2339468151,0.513903439,0.2232467234,0.0511075296,-0.250493288,0.0126367444,0.1277921498,-0.0999452472,-0.2233893126,-0.1853364855,0.2213394493,0.070104003,0.1439287215,-0.2015253305,0.17729792,0.2823123336,0.0923544243,-0.0955821276,-0.1467555314,-0.337774545,-0.0845686346,-0.0994954631,-0.1259720325,-0.0281348322,0.2752671242,0.3266276717,-0.0365251973,0.0784481689,0.2283449024,-0.1694997996,0.115801461,0.3157924116,-0.1905090809,0.3854655027,0.2556683719,-0.0222615898,-0.017681526,-0.0211289767,0.0807342902,0.2407308072,-0.137873143,0.2707097232,-0.016586408,0.3332600594,-0.0678861663,-0.0527783036,-0.0566621535,0.1189701855,0.0997710973,-0.0448881052,-0.2961109579,0.1191897765,0.0921307877,-0.0042592441,-0.0634180903,-0.1733612269,-0.3518213928,0.2331577986,-0.0514189973,-0.2088760436,0.1475667506,-0.334433347,0.2592228055,0.0295759346,-0.3412538767,-0.0522108823,-0.2054463923,0.2089969218,0.0529111624,-0.0920000598,-0.155643478,0.5642309189,0.5620846748,0.3921133876,-0.5298598409,-0.4843091071,-0.0141239166,0.0818410516,-0.1659708172,0.024325639,0.0106068142,-0.2986755967,-0.215908885,0.0515596904,0.0790272877,-0.2116862983,-0.143138364,-0.1234049052,-0.0902454406,-0.0976947024,-0.165838182,-0.0106627587,-0.3048554361,0.4907079935,0.0869383961,0.1383474022,-0.0187547524,0.1312554777,-0.0678128973,-0.1683114469,-0.3420356214,-0.1610829085,0.1426559687,0.3432585001,-0.4180550277,0.0558481812,0.2637420893,-0.1318227053,-0.0202905796,0.4314632416,-0.2649515867,0.0927449688,-0.0009440525,0.6025671363,-0.1807419211,-0.0231507495,0.5678798556,0.2796882689,-0.0451871902,-0.0530390739,-0.1038935333,0.1486031413,0.4860756397,0.204756096,-0.0486469157,0.3489842117,-0.2798686028,0.6352140307,-0.2009723634,-0.0147747695,0.2248992473,-0.210642755,0.1744220257,-0.1710287035,0.0627716854,-0.2686358392,0.0295552518,-0.2321026474,-0.1637757421,0.0190743748,-0.2214834392,0.0246323776,0.0078162476,0.0106095588,-0.0550731942,0.1867983788,-0.6058994532,0.5232605934,-0.1445961595,-0.1736809909,-0.5607355833,-0.0098552732,0.0159520227,-0.1441912949,0.4654318988,-0.4178943038,-0.4612035155,-0.2174022794,-0.3871830404,0.1870581955,0.0808905289,-0.073593989,-0.2195816338,-0.1895532161,-0.144218713,0.2287197411,0.3513083756,-0.6429318786,-0.164701879,-0.1423741877,0.2122738212,-0.0974067301,0.059299536,0.1740359664,0.361941576,-0.2174941301,0.5275080204,-0.2049197704,0.244438082,0.3806366324,-0.0948139206,-0.2038997561,-0.0996651873,0.0244134553,-0.0027050998,-0.0446241237,0.0897795707,-0.0682212412,-0.2114564478,0.3042940795,-0.4484103024,-0.274082303,0.0758049339,-0.1722645015,0.122608982,0.249020189,0.2360480875,0.2840407491,0.014226377,0.0196377821,-0.0259283446,0.3448985815,0.025977768,-0.183177352,-0.1373799741,-0.4283925295,0.7307028174,0.1598422378,0.0783917606,0.7191215158,-0.0668129846,-0.118019551,-0.376565069,-0.094275564,0.2884323895,-0.1874753088,-0.0856973082,-0.2908050418,0.5907081962,-0.0530480407,-0.0741091967,0.6676121354,0.088445425,-0.2206763178,0.5987038612,0.220946461,0.9525287151,-0.1935117692,-0.1417441815,-0.0580418445,-0.250408262,0.2143838108,-0.1726959944,-0.0833499581,-0.1577391028,-0.2897281945,-0.0891431943,-0.118809849,0.091390565,0.3618845642,-0.2552137077,-0.2974041402,-0.1148666367,0.476352483,-0.0730989277,0.1706347764,0.2400482595,-0.2808719575,0.1182684675,0.0872067437,-0.02255922,-0.1019552872,-0.2490098178,-0.2241192311,0.0724807903,-0.2220399976,-0.1071740016,-0.1413418204,-0.3563657701,-0.0249510296,-0.1281701922,-0.0110547887,0.129417479,-0.105497919,-0.046638675,0.0432864167,-0.0783129483,0.2617404759,0.3378766477,0.3220254481,-0.0389920771,-0.0075217918,-0.0083564604,-0.067902036,-0.186436668,0.3520045877,-0.248849377,-0.4843153954,-0.1339602619,-0.0075569786,-0.0220865346,-0.2310641855,-0.3893095553,0.195092544,-0.1880642921,-0.0804353133,0.1191900447,0.0564944036,-0.0598608255,0.2180835605,0.3378434777,-0.0572931543,0.0812928975,-0.0387553535,-0.112873517,0.0439999215,0.2099787742,-0.2339549661,-0.0743980259,-0.1285293698,0.0564184785,0.3340590298,-0.4564085007,-0.1627862453,0.2021562159,0.0325256884,-0.1476847082,0.2081984282,-0.1721680313,0.4044394791,-0.2823361456,-0.0309269764,-0.3442201614,0.193166256,-0.0770243555,0.2968859375,-0.2407788634,-0.1100059003,-0.1486532986,0.4155266583,-0.2784255743,-0.2237745672,0.3204050958,0.3578750789,0.0001110953,0.490978092,0.0658963993,-0.2763756514,0.1371061951,-0.0523009636,-0.0367809758,-0.1670792848,0.1241114959,0.1868125051,0.0657421872,-0.2053637356,-0.256450057,-0.2793375552,-0.1318929791,-0.1434379518,-0.1424623132,-0.2237465233,-0.2377745956,-0.1527929753,0.0341303647,0.1899696589,-0.1506687701,0.1963237971,0.1409431398,0.1933099031,-0.0990123823,0.1034793779,-0.0658580139,0.0952963009,0.1380637139,0.0459183306,0.0148241222,-0.0612082146,0.2245169729,0.0737920403,0.0046500815,-0.0272118002,0.2537141442,-0.1963344663,-0.1477659345,0.2409239262,0.0296953879,0.2407490164,-0.2175312936,0.1210120693,0.4524828792,-0.0625467524,0.1484796852,0.1480235606,0.2969687283,0.432372272,0.0150218913,0.0962564349,-0.3143247366,-0.3172514439,0.1561307311,0.4448055327,0.1358335167,0.0962138474,0.2541843951,0.3600869179,0.1806252748,0.2444464266,0.1413002014,0.2294823974,0.4651010334,-0.2577500045,0.2078315318,-0.3314006329,0.2717255056,-0.0367860906,0.1530897468,0.0134542212,0.502402842,-0.0964631736,0.1339132339,-0.2314017713,0.0420858525,0.0395562574,-0.166744262,0.1457146108,0.1409804225,-0.1101643592,0.0133682629,-0.1592698246,0.1053883806,-0.0872302577,0.3674431145,-0.0022116464,0.3381568491,0.2733388543,0.4702838957,0.2463739365,0.2139004469,-0.1148894355,-0.1212856621,-0.4095312655,-0.0748468414,0.0479692407,0.2244103253,-0.1902330965,0.2198684216,-0.2684758306,0.697140336,-0.0221111327,-0.2750869095,0.0647155866,0.0996285751,0.2045362443,0.0011921842,-0.0783091784,0.1592434049,-0.0256542265,0.1731173545,0.0404459387,0.2446762919,-0.1540360004,0.1421146095,0.2685221136,-0.5370784998,-0.0397075675,0.0850884542,0.2115752101,-0.3055979908,0.3663336337,-0.0444387868,-0.1261988729,-0.1228776127,0.0503607132,-0.1525100768,0.1384806633,0.2217730433,0.0412503406,-0.2644842565,-0.1519628018,-0.387691021,0.0165584423,-0.0700688511,-0.1563557535,0.383461982,0.0403531604,0.2811020017,0.0801402181,0.3604902923,-0.4743566215,0.090317212,0.1819583327,0.002680768,-0.2182780653,0.0355388001,0.1078452766,-0.1794999391,0.0957138389,0.0828998834,0.308110863,0.0043472508,-0.2374207526,-0.2536135018,-0.0798514336,0.0157363284,0.5547657609,-0.2445423752,0.2568741143,-0.2621216178,0.2161101103,0.0188885611,-0.2626237869,-0.2059480399,0.1162952259,-0.0600693673,0.2213888913,0.1159595698,-0.0937421322,-0.3625070155,-0.0897852629,-0.3467420936,-0.2869196832,-0.2397892624,0.2558828592,-0.1416024715,-0.0630056337,0.0136026833,-0.0835833773,0.0138172442,0.1478604525,0.0845877901,-0.0859042406,0.1457944661,0.1201748997,-0.0749814659,0.1234502196,0.1308690608,0.2448751479,0.1080155671,-0.6100404859,-0.2564037442,0.4132263362,-0.1011890471,0.350841701,0.3517682254,-0.0846994221,0.2125426531,-0.1900782138,0.5408409834,-0.3125436306,0.3141437471,-0.3981516659,-0.3060057759]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/256","title":"[Feature request] Add a feature to dataset","comments":"Given another source of data loaded in, I want to pre-add it to the dataset so that it aligns with the indices of the arrow dataset prior to performing map.\r\n\r\nE.g. \r\n```\r\nnew_info = list of length dataset['train']\r\n\r\ndataset['train'] = dataset['train'].map(lambda x: some_function(x, new_info[index of x]))\r\n\r\ndef some_function(x, new_info_x):\r\n # adds new_info[index of x] as a field to x\r\n x['new_info'] = new_info_x\r\n return x\r\n```\r\nI was thinking to instead create a new field in the arrow dataset so that instance x contains all the necessary information when map function is applied (since I don't have index information to pass to map function).","body":"Is there a straightforward way to add a field to the arrow_dataset, prior to performing map?","comment_length":103,"text":"[Feature request] Add a feature to dataset\nIs there a straightforward way to add a field to the arrow_dataset, prior to performing map?\nGiven another source of data loaded in, I want to pre-add it to the dataset so that it aligns with the indices of the arrow dataset prior to performing map.\r\n\r\nE.g. \r\n```\r\nnew_info = list of length dataset['train']\r\n\r\ndataset['train'] = dataset['train'].map(lambda x: some_function(x, new_info[index of x]))\r\n\r\ndef some_function(x, new_info_x):\r\n # adds new_info[index of x] as a field to x\r\n x['new_info'] = new_info_x\r\n return x\r\n```\r\nI was thinking to instead create a new field in the arrow dataset so that instance x contains all the necessary information when map function is applied (since I don't have index information to pass to map function).","embeddings":[-0.2954547703,0.0480495133,-0.068608582,-0.1894563437,0.0562324785,0.2561274171,0.0807101727,0.1100029051,0.1961615384,-0.0207735319,0.4314366281,0.5331404209,-0.1051034182,0.2124332637,0.133281514,0.0656450167,-0.0400395468,0.3422541916,-0.0246693976,0.0526039377,-0.2568831444,-0.3627748787,0.0121611562,-0.2262430042,-0.2012324333,-0.0531029813,0.0925348029,-0.2065407038,-0.0499845408,-0.4349793494,0.0697100982,0.0547874197,-0.0611637011,-0.0775067657,-0.0001218519,-0.0082797566,0.2286789864,-0.0181808807,-0.0220237598,-0.0886155665,-0.6228856444,-0.4504071474,0.1403827667,-0.1462102532,-0.1161957681,-0.2814851701,0.0643228516,-0.4899307191,-0.0304019302,0.1176671013,0.1175913066,-0.0339107066,0.3083866239,0.0327693485,0.3443826437,0.2178616673,-0.2083014548,-0.2057977468,-0.0556959473,-0.0051278574,0.2442174554,0.1893960685,0.1678595394,-0.3340607882,0.4431711733,0.1553868949,-0.0249546971,-0.2453878671,-0.3294202387,0.0895995572,0.6357524991,-0.4809344113,-0.4458319843,-0.2342638373,0.3120888174,-0.290114969,0.0607441217,-0.1333394945,0.210457027,-0.0164413694,0.2804879546,-0.389656961,0.0144422483,0.2269136161,0.0619366057,0.476678133,0.1661749482,0.096834287,-0.1452657282,-0.1335329562,0.1443166733,-0.330093056,-0.06665048,0.5027745962,-0.2417925447,-0.2139225453,-0.0244842526,-0.438134104,-0.0016758052,0.1643468589,0.1435663551,0.0206871275,-0.2361611426,0.1020519063,0.2505217791,0.1508660913,-0.0119642578,0.2477612644,-0.1225104108,-0.4533706009,-0.0715743825,-0.2039059997,0.3690459728,-0.1363199055,0.3272722363,-0.010967344,0.3330433071,0.1570524573,0.1418698132,0.0278924368,-0.0564527139,-0.0701018795,0.0854346454,-0.0442520306,0.1940442473,-0.0491920784,-0.1129735634,0.411008358,0.2865205407,0.217422843,-0.0495889969,-0.2126169056,0.151268959,0.2483240217,-0.0767468885,0.0651071891,0.2182892412,0.0567841195,0.1074002311,0.2571184635,0.362067759,-0.0483580343,0.1796389967,0.2562591732,-0.1308117658,0.1176817268,-0.0223747231,0.2681906223,-0.2887227237,0.4301692247,-0.1666408032,-0.3181441128,-0.0180657003,0.0798423514,0.0267928187,-0.373313576,-0.3221887946,0.5601476431,0.0531609878,-0.3258930147,0.3006869256,0.0765359029,-0.4147170186,-0.3155430555,0.0478864126,0.202069357,-0.0447325408,-0.0481433496,-0.3672377765,-0.0581971146,-0.289432168,-0.3591374457,-0.1361573339,0.1373808086,-0.0681172982,0.3618087471,0.7887734771,-0.1264313608,-0.3442746103,0.045232445,-0.3030264974,0.1571156383,-0.375082165,0.5562777519,-0.1387429684,0.1508688033,0.0626032874,0.5208093524,-0.2876549959,-0.0548995435,0.1433007419,0.0251336675,0.0928553715,0.1853526533,-0.1041533202,0.0862533078,0.2928877771,-0.361335516,-0.1255096048,0.229325816,0.078717418,0.2501796186,0.2867565453,-0.1407821476,-0.2265393436,-0.3820767105,-0.4558905959,0.1355385333,0.1613528728,-0.1518810391,-0.0022721998,-0.2403021157,0.0519773215,0.3842323422,-0.0291440953,0.1029371843,-0.0354005322,-0.1379572302,-0.1275209337,-0.1152086407,-0.1917721182,-0.0880454406,0.0234132558,0.1266871393,-0.081543833,-0.0156095549,0.2958852351,-0.1038988382,-0.4534268677,0.1745183468,-0.0810087323,0.1316818893,-0.2303139865,0.2882097661,0.281044513,0.307657212,0.3753387034,0.3596666753,0.1202720925,-0.2022271901,-0.0492429622,0.2572801113,-0.0293284841,-0.152689442,-0.1535563171,0.5156170726,0.1086487621,0.2238612622,-0.2337494642,0.041144494,0.1870003939,0.1000285894,-0.2504220605,-0.103762418,-0.3828133047,0.0285269767,0.0936289281,0.0024606087,-0.0357764363,0.2042023689,0.0699372068,-0.0240182914,0.1600016057,0.2019366175,-0.2276441306,0.1079131439,0.0774967149,-0.0624136589,0.3290600181,0.2227063626,0.0126179587,0.0426483452,0.1092888191,0.160209924,0.1915973574,-0.0738721341,0.3904670775,0.1353820264,0.3821550012,0.155317679,0.0920716226,0.0518261939,-0.0285529736,0.0855422318,-0.1734751314,-0.123380743,0.1931572258,0.1512268186,-0.1907939464,-0.0727324188,-0.1309424341,-0.2570100129,-0.0422084853,0.0283493903,-0.2032609433,0.0837598816,-0.366987437,0.1900515258,0.1666791588,-0.6432436705,0.0507394411,-0.0893904567,0.2384161651,-0.0036567324,-0.0164030138,-0.174557671,0.3104876876,0.7045741677,0.3526602685,-0.4002373219,-0.5986948013,-0.1394598931,0.0326243453,0.0610160977,-0.056439463,-0.1356823742,-0.2187397331,-0.2760376334,0.0529641323,0.0978603065,-0.1257185191,-0.2376390845,-0.0766170472,0.0012320286,-0.0057821316,-0.011828755,-0.2295980006,-0.2544108927,0.3818815649,0.2282903641,0.0037640331,0.0251771454,0.1335633695,-0.1741533428,0.0961103886,-0.3459210992,-0.0677424967,0.0082773566,0.4669155777,-0.291755408,0.2223026305,-0.0565286539,-0.4282010794,-0.1681381166,0.5416530371,-0.2858798802,0.0335521512,-0.0685808808,0.4442356825,-0.00547755,0.0415530168,0.4493017793,0.1307611614,0.1019128188,-0.0397400931,-0.1687309444,0.0195706673,0.3857576847,0.1371400952,0.026542509,0.2526220083,-0.4078219831,0.8244695067,-0.096462138,-0.0982213393,0.2135765702,-0.3015334606,0.2136496007,-0.1988614351,-0.0391930006,-0.3540034592,0.1187202632,-0.2392914444,-0.1137767881,-0.1378688961,-0.3384617567,0.0115319826,-0.1034522802,-0.0715048462,-0.1136921793,0.2776060402,-0.7096518278,0.5136854053,-0.1167214587,-0.1627096236,-0.5571870208,-0.0105287135,-0.1344741136,-0.257779032,0.2695598602,-0.358886838,-0.2897860408,-0.1367036551,-0.3390552998,0.1599769592,0.328982085,0.0109233232,0.0694298893,-0.284619391,-0.2018526942,0.1255827993,0.1007939577,-0.7038961649,-0.3776725531,-0.0603376366,0.1350761056,-0.312774241,0.0321955271,0.2692115009,0.3959923983,-0.2950237393,0.332785219,-0.0970068574,0.2671418786,0.2079720944,-0.0519145988,-0.1870245337,0.0454990231,-0.0406397358,0.1094188839,-0.0325950533,0.1395483166,-0.0781616271,-0.2029489875,0.4935392439,-0.2078263909,-0.5050938129,0.2532685399,0.0099108415,0.1636029333,0.3125104308,0.3241735995,0.3094304204,0.120487377,-0.1788273156,0.0644044578,0.4180524647,-0.1141905636,-0.2228506505,-0.3180222213,-0.1268287152,0.6585425138,0.0218003429,0.0964458659,0.4333217144,-0.1049373299,-0.1404343545,-0.4417458475,0.0513241999,0.4564833939,-0.0933029056,-0.1556733251,-0.5510123372,0.6145581007,-0.0646322891,-0.0433357917,0.5349655747,0.206373319,-0.159887448,0.5062615275,0.2650838494,0.7346950173,-0.0204222854,0.0301517136,0.1933490634,-0.0894335881,0.1631153375,-0.0851109102,-0.0200578272,-0.1575649381,-0.4308118224,-0.0038920555,-0.1519746482,-0.1166022569,0.3548955917,-0.0452916957,-0.2973412275,-0.1605757028,0.3114293814,-0.1506595612,0.1867831647,0.0780762061,-0.2434274554,0.0507799089,0.0568329617,0.1301898807,-0.0525233522,-0.1661621481,-0.1270719171,-0.2141226083,-0.1048301458,-0.1963620931,0.0009955138,-0.3412139714,0.1004746035,-0.238670826,-0.0842973515,0.1819693893,0.1106549129,0.0588778853,-0.057705801,0.0113993576,0.3690600395,0.3624340594,0.3142381608,0.0789666921,-0.0832299739,0.0219900496,0.1261251867,-0.1139482111,0.3394486904,-0.1307381094,-0.5961117148,-0.1944648325,0.0659425408,-0.125480786,-0.1974856108,-0.382786721,0.2993823886,-0.0693841204,-0.0267402586,0.0489823446,-0.0073937178,-0.0317382105,0.1884375513,0.098294735,-0.2052383572,0.0361777581,0.0387698114,0.0022943523,0.207085073,0.4687119722,-0.3493294716,0.0531408079,0.0249813125,0.098645471,0.406825304,-0.1634347588,-0.1159841567,-0.0078762602,-0.0629752502,-0.109324269,0.1393965185,-0.1683247387,0.2923592031,-0.1629251838,-0.159529835,-0.3864021301,0.2419726402,0.0289385114,0.4287542403,-0.0443828739,-0.1234662533,-0.3896718621,0.438934952,-0.1900089532,-0.3230793476,0.3609903753,0.1855752766,0.2330751419,0.4451929629,0.3107380867,-0.3401559293,0.0582812428,-0.1498091072,0.1230471581,-0.113533251,0.0166522954,0.1590275615,0.0453396104,-0.2391647249,-0.2431190014,-0.3971074224,-0.038121447,-0.3407492936,-0.166481182,-0.0747392774,-0.1129997894,-0.1436064243,0.1304463595,0.1407217681,-0.1949779391,0.1896747351,0.0770143867,0.0530369915,0.0062448503,0.2368282527,-0.1569787264,0.1254940778,0.1131369472,0.0650525317,0.2635497451,-0.0008799365,0.1680134237,0.0128247049,0.1215806752,0.3147473037,0.3299047351,0.0364718139,-0.1927658767,0.228824839,-0.0919209644,0.1120904163,-0.2940257788,0.0671226159,0.3260783851,-0.1971905529,0.1943188161,0.2065911442,0.2097284794,0.6998457313,-0.2804930806,0.0025720384,-0.3698576391,-0.1945171207,0.2173478901,0.4785536528,0.2256541103,-0.1284535676,0.2805359662,0.5443066955,0.3776435256,0.3105355501,0.1697674841,0.2846684158,0.2513502538,-0.3804255426,0.0895137265,-0.3886937797,0.2942371666,-0.0175570175,0.2701082826,0.0546994954,0.5864534378,-0.0185757261,0.1098877117,-0.3700762391,0.0214762613,0.2235758156,-0.1024249569,-0.0375721678,0.3706280887,-0.0738497525,-0.0263352208,-0.2514857948,0.1283645928,-0.1546100527,0.2608329058,0.0299448129,0.3402260542,0.1871611923,0.4017398655,0.2882911563,0.2756876647,-0.1240538135,-0.0971367657,-0.452960521,-0.1925428063,0.0492265113,0.2918071151,-0.1111714393,0.061806079,-0.3906153142,0.4911114872,0.0729961842,-0.2280675769,-0.0058034728,-0.1085138544,0.1091511473,-0.1092999429,-0.016031066,0.1079543233,-0.0405579284,0.102297999,0.0666756704,0.2367082536,-0.1701485515,0.1763896197,0.2817983925,-0.222491309,0.0382450074,0.1261460334,0.2049907744,-0.3052233458,0.3017689288,-0.0709675923,-0.0111328606,-0.0160955656,0.0150355492,0.0476685353,0.1226761937,0.1894771606,0.0889020339,-0.191138953,-0.0355496034,-0.5502502322,0.01177694,-0.029770216,-0.1308160573,0.2801946998,0.0537481532,0.2917378545,-0.0245149639,0.2399183661,-0.4459002316,0.02412875,0.0805633739,0.0340039618,-0.1697068214,-0.3878952861,-0.0658751801,0.0525582246,0.0664348602,-0.1116261557,0.2987186015,-0.0268488545,-0.1348082572,-0.3300515115,-0.1046007499,-0.0229625367,0.4856193662,-0.2991344333,0.2528913915,-0.3834982514,0.2297551781,-0.0480500646,-0.4150995016,-0.4699653089,0.3128101826,-0.2131816745,0.2614042461,-0.1266082823,-0.0608519167,-0.1867714673,-0.1711273938,-0.2456919551,-0.5278000832,-0.2825757563,0.1686342657,-0.3479966819,0.1548517793,-0.0018807049,0.1725323349,0.0849721879,0.3282291293,0.2667157352,-0.079088524,0.2411333472,0.0584791005,0.1523552239,-0.004625102,-0.1585965008,0.209803775,0.2215560973,-0.4839754403,-0.0709364936,0.4147847593,-0.0901110843,0.260963738,0.1288648099,0.0136307552,0.1482875943,-0.2505509853,0.450189352,-0.3410737813,0.2927999198,-0.252279371,-0.0946431309]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/256","title":"[Feature request] Add a feature to dataset","comments":"This is what I have so far: \r\n\r\n```\r\nimport pyarrow as pa\r\nfrom nlp.arrow_dataset import Dataset\r\n\r\naug_dataset = dataset['train'][:]\r\naug_dataset['new_info'] = new_info\r\n\r\n#reformat as arrow-table\r\nschema = dataset['train'].schema\r\n\r\n# this line doesn't work:\r\nschema.append(pa.field('new_info', pa.int32()))\r\n\r\ntable = pa.Table.from_pydict(\r\n aug_dataset,\r\n schema=schema\r\n)\r\ndataset['train'] = Dataset(table) \r\n```","body":"Is there a straightforward way to add a field to the arrow_dataset, prior to performing map?","comment_length":45,"text":"[Feature request] Add a feature to dataset\nIs there a straightforward way to add a field to the arrow_dataset, prior to performing map?\nThis is what I have so far: \r\n\r\n```\r\nimport pyarrow as pa\r\nfrom nlp.arrow_dataset import Dataset\r\n\r\naug_dataset = dataset['train'][:]\r\naug_dataset['new_info'] = new_info\r\n\r\n#reformat as arrow-table\r\nschema = dataset['train'].schema\r\n\r\n# this line doesn't work:\r\nschema.append(pa.field('new_info', pa.int32()))\r\n\r\ntable = pa.Table.from_pydict(\r\n aug_dataset,\r\n schema=schema\r\n)\r\ndataset['train'] = Dataset(table) \r\n```","embeddings":[-0.1585631967,0.1218319461,-0.0199142918,-0.2379771769,0.1159518287,0.2459182143,0.0558399968,0.2358242124,-0.0567644648,-0.0269440133,0.5227229595,0.7792691588,-0.1148941293,0.2621577084,0.4461783767,0.0180788506,0.0926241353,0.3316138089,0.0632166937,0.053030584,-0.2023163885,-0.5101904869,-0.0154500231,-0.1047873199,-0.2550474703,0.0640518516,-0.0727066621,-0.2698785365,-0.1969700605,-0.6185671687,0.0552477315,-0.0052971891,0.0488106385,-0.2825003564,-0.0001257511,-0.0990261436,0.1238243207,0.0937622264,-0.1492837667,0.0070876474,-0.1820129901,-0.5508115888,0.3477323055,-0.25052917,-0.02778388,-0.3641993701,0.0755024999,0.0092849629,-0.0952078477,0.0579496585,0.0997853279,0.0939668193,0.2876735926,0.1071146429,0.3623693585,0.2798253298,-0.3339948356,-0.0428954586,-0.0220629424,-0.2151848674,0.1690187752,0.1127569675,0.1265625507,-0.2649116814,0.3952746689,0.2403582633,0.0384079106,-0.072328493,-0.2271326333,-0.1069110855,0.6156909466,-0.6202868819,-0.3876196146,-0.2214737386,0.3082877398,-0.4906707704,-0.0614589639,0.0448310971,0.1050458103,0.0437751263,0.3411901891,-0.4858388007,-0.1245170832,0.3369037807,0.1579572558,0.5274518728,0.3024938405,0.1550373435,-0.1368014067,0.0279536378,0.1394296288,0.0006353033,-0.0988960937,0.5474140048,-0.1916419119,-0.3288538456,0.1603990942,-0.3275590837,0.0637634099,0.225965783,0.072326228,-0.0258287359,-0.1592750549,0.1315601766,0.2263168544,0.0333113261,-0.0248911455,0.2305516154,-0.0531374067,-0.4014611244,0.0675160065,-0.1941779256,0.1835125685,-0.0240348242,0.5270875692,-0.0994340554,0.5241091251,0.0916185528,0.2641537189,-0.12292023,-0.0796188489,0.0289337803,0.0015675381,0.1388510913,0.1910752356,-0.0476425886,-0.1843174994,0.4463225305,0.2602829337,0.0229133125,-0.0585689172,-0.1158423573,0.1210389733,0.2123061568,0.053764984,0.1982225925,0.12853311,-0.0099368943,-0.0215493757,0.1386262029,0.1777260602,-0.0943022966,0.1793120801,0.3338983655,-0.0447518341,0.0700507686,-0.0151229054,0.2096881419,-0.2649386227,0.4940430522,-0.0427006558,-0.3354871869,-0.182712093,0.0792179629,-0.0760878697,-0.3811966777,-0.3896959424,0.2810362875,0.1529284716,-0.2052124441,0.3210585415,0.1022731066,-0.2074391693,-0.3183969557,0.0141020566,0.0050593172,0.020591436,-0.0731970966,-0.2081436962,-0.106873326,-0.254784435,-0.314312458,-0.2401307672,0.3415870965,0.0265480988,0.316865474,0.9826906323,-0.0947195664,-0.3102529645,-0.0044013709,-0.2885476947,-0.1376720071,-0.2810827494,0.5194311142,-0.000458518,0.1911732852,-0.1188182086,0.4386410117,-0.2460013777,0.0624340326,0.1339839846,-0.0622337013,0.3955076039,0.2214668244,-0.0669111311,0.0565835573,0.2403141856,-0.1898543835,0.078507632,0.1749491245,0.1417320371,-0.0234400928,0.5813657641,-0.1480278075,-0.1587654203,-0.3559047282,-0.5103592873,0.0481979623,0.0614799298,-0.0498385392,-0.2668022215,-0.3365260661,0.1379484087,0.410826236,-0.21857135,0.0828226581,-0.0337449946,-0.2657914162,0.0313534811,-0.0390634015,-0.3062026203,-0.2801218331,0.0827199146,0.2396755666,0.0705125704,0.1131925136,0.1532941163,-0.2490605712,-0.3930204511,0.2865010798,0.0241674427,0.1350463033,-0.112665385,0.2159660608,0.2123830169,0.190021053,0.2170919031,0.2195721418,0.0773657933,-0.2173906267,-0.0081111444,0.024139272,-0.0064256126,-0.1075872108,-0.0767975301,0.2526226938,0.1183936298,0.1508035362,-0.0551127419,0.1701215357,0.297231108,0.1012447327,-0.1686341017,-0.0361541808,-0.3948580325,0.082512103,0.0132988421,-0.0205064379,-0.1242800876,0.1715178043,0.3190536797,-0.1497215629,0.1284432113,0.193891719,-0.2011642456,0.0162294768,0.2585250139,-0.1799705178,0.2840501368,0.2247115523,-0.0071285721,0.1735942513,-0.1491429657,0.1777966022,0.1507430375,-0.0082725342,0.3791612387,0.0827663839,0.491884172,0.062659882,0.0409868211,-0.0164795388,0.1325978935,0.2742210925,-0.1277945936,-0.2125937641,0.1566524804,-0.0656216145,-0.1127957776,-0.2664131224,-0.1476638168,-0.3067302108,0.0748658925,0.0802909136,-0.0718798563,0.0658462271,-0.4538493156,-0.0331082083,0.0424247161,-0.6114932895,0.0338730961,-0.2173403203,0.1986373067,-0.0420190394,0.0002152516,-0.0223866981,0.4436041713,0.5154750347,0.3509569466,-0.4484282136,-0.5637549758,-0.0260125343,-0.0625211746,-0.0568652935,0.0072709452,-0.033459954,-0.1215717122,-0.2443925738,0.1578592807,0.0520993657,-0.1704946607,-0.0017590916,-0.1390076131,-0.0746620595,-0.1779552847,-0.1160608679,-0.253651917,-0.2001410425,0.4121743441,0.3110242486,0.1660983264,0.1023302376,0.0647426173,-0.0393917114,-0.1045548469,-0.2717067003,0.0465511717,0.2692864239,0.3771940768,-0.2905590534,-0.0700872988,0.1223501712,-0.3589475155,-0.0133673809,0.3452511132,-0.2681254447,-0.1045634449,-0.1315259039,0.6544305086,-0.1809949279,0.061110694,0.4917134643,0.1800765544,0.1259661913,0.0645482615,-0.0575284921,0.0581770651,0.4751567245,0.0536932535,-0.1974352598,0.3545249999,-0.1638989151,0.6824057102,-0.2674229443,-0.0324546099,0.2890657187,-0.4521875679,0.1243514866,-0.1754645258,-0.0518148616,-0.3754043877,-0.0402451158,-0.1911339015,-0.1986826956,-0.0592075065,-0.2571844757,-0.0269646514,-0.0767010972,-0.034670461,-0.1432774514,0.2570279241,-0.6129857302,0.4767407477,-0.023436334,-0.1636833549,-0.593467474,-0.0892284885,-0.160461992,-0.1027160734,0.164583236,-0.2988649011,-0.4919733405,-0.4300163686,-0.3920715749,0.0194917303,0.1957015991,0.1655498147,0.0377299264,-0.2686166465,-0.0311103016,0.0775839388,0.2406376004,-0.35993433,-0.1689047366,0.0284106433,0.2840740383,-0.2831538618,-0.0534141324,0.1423125863,0.2584647536,-0.2029890418,0.3765675426,-0.248746559,0.1561361104,0.4884978235,0.0460100472,-0.1258965135,0.0359547511,-0.2119247019,-0.1457538009,-0.1295851022,-0.0102055538,-0.1028543487,-0.104898572,0.2832601368,-0.1269048154,-0.3473350406,0.0521850511,-0.0832878426,0.2079725564,0.1858105808,0.1545908302,0.3581623137,-0.1583478749,-0.037782751,-0.0189065468,0.4107255936,0.0369465016,-0.3677837551,-0.2743690908,-0.2612393498,0.8111658096,-0.0533727072,0.0421168655,0.6886078715,-0.2510790527,-0.088377513,-0.3154602051,0.0237926822,0.3955509365,-0.221372053,-0.2090979666,-0.4495918453,0.6590726972,-0.0608053617,0.0022970031,0.2921187282,0.1764971465,-0.2020089179,0.6029342413,0.0289512016,0.7561174035,-0.1606501937,0.1596265882,0.3366757333,-0.2007087618,0.280625701,-0.1172663942,-0.064888373,-0.3031124175,-0.1477965266,-0.0394733027,-0.222287029,0.0863227174,0.1838237792,-0.114170894,-0.1544334143,-0.1009677723,0.391569674,-0.029187249,0.335498035,0.0553730913,-0.2120381296,-0.1536519825,-0.015315121,-0.1098567843,-0.1193850115,-0.1678930968,-0.155338496,-0.287281245,-0.2775656879,-0.1146816537,-0.0427805595,-0.2601376474,0.0398515835,-0.0253157262,-0.189904362,0.0646939725,0.0595824718,-0.0665888041,-0.1198209748,-0.0020896788,0.2460796833,0.2279047966,0.2359357327,0.1082561463,0.0183836631,0.1020189151,-0.0309780296,-0.2448406219,0.2957397103,-0.1605062932,-0.5123876333,-0.0357952602,0.0153745282,0.1052625477,-0.2330364138,-0.362177223,0.1411417872,-0.1420343965,-0.1177442148,0.0439496003,0.0383621529,-0.0065550474,0.1905516535,0.2257850766,-0.1166319028,0.0457632691,0.0029494548,-0.0380199961,0.1824090481,0.4506618977,-0.2103733569,-0.0039632875,0.0047603911,0.0415549465,0.2939327061,-0.4868574739,-0.2583299577,0.2763181329,0.1503428668,-0.172943905,0.1077881604,-0.2781784832,0.2693227232,-0.2398824543,-0.0899362043,-0.3913894296,0.3100816011,0.0720862076,0.2564846575,0.0913535133,0.3130819499,-0.2448113859,0.2852059603,-0.1528395414,-0.2153529078,0.3948774934,0.3118824065,-0.0182794929,0.3555015624,0.0666697249,-0.3369539678,0.1057712287,-0.079329811,0.19605878,-0.1148355976,0.0333924741,0.1906636357,0.1341796666,-0.2545539141,-0.27063936,-0.3031240106,-0.0277348254,-0.1833405495,-0.1302250028,-0.2740218937,-0.2145328224,0.1525447518,0.2862391174,0.2720698118,-0.1430950463,0.1244455948,0.0282883383,0.1105685011,-0.0409758203,0.2331154495,-0.2132423371,0.2029624134,-0.006632166,0.0055660191,-0.0514231399,-0.1476125866,0.3038796186,-0.1342989504,0.0842218623,-0.1251690388,0.2900972664,-0.0221133586,-0.1756536514,0.239649564,0.1706908941,0.1219791993,-0.3873328567,0.173508808,0.3757994175,-0.2174308151,0.1506309509,0.2236278057,0.19726336,0.615765214,-0.2723188102,0.0221008845,-0.1972346455,-0.1329652816,0.070118092,0.6301004887,0.0838815346,-0.0011548232,0.1873222739,0.4594738781,0.0896190777,0.2704044878,0.1308705658,0.2527268529,0.332793355,-0.4072865844,0.1903537363,-0.2755206525,0.0396065079,-0.0424991138,0.3121157587,-0.1131844595,0.3751172423,0.2397871614,0.4041926265,-0.1717334241,0.1288266182,0.1929618418,-0.0212712791,0.0574843474,0.2609863877,0.0153998658,0.0152062913,-0.1981087774,-0.047091756,-0.3349761963,0.0238036532,0.0317516588,0.1960808337,0.1262893975,0.6106854081,0.3089271784,0.1500250697,-0.1102879047,-0.1066460833,-0.3961720765,-0.1840778142,-0.192886129,0.0807411969,0.0282910764,0.1530740261,-0.386597693,0.5614286065,-0.0692833886,-0.2305684686,0.0821030065,0.0099925073,0.2477477491,0.1952481568,-0.1186825186,0.0815068036,-0.0082981857,0.0947486162,0.1438757926,0.1882583499,-0.2091565579,0.46249035,0.2958878875,-0.6133111119,0.0203604307,0.1428034157,0.0756600797,-0.1968321651,0.3214893937,0.1661864519,-0.0556000657,-0.243800655,-0.0070667188,0.0138259046,0.159683913,0.2941429913,0.2848108411,-0.1666688472,-0.0829461962,-0.5021372437,-0.0131307608,-0.1394735873,-0.2369990498,0.4343595803,0.0598335192,0.1677977145,0.0557848923,0.2752939463,-0.2091964185,0.0686089769,0.1043534577,0.1213505641,-0.2491119802,-0.2200087011,0.1554754078,-0.1498524845,-0.0473012477,-0.0012173576,0.1140875518,-0.0710764825,-0.3338567615,-0.1559708416,0.0447101593,0.1117583737,0.4848420918,-0.3177315295,0.4325461388,-0.0734915733,0.207606867,-0.1644307077,-0.3024151623,-0.3349880874,0.1736537367,-0.1974713206,0.1441137642,0.1021316648,-0.0277318154,-0.2601104379,-0.1121458784,-0.3994649053,-0.4051250517,-0.1675203294,0.2998760343,-0.0523840636,0.1132880896,-0.0925959796,-0.1197040826,-0.0616280809,0.2605071366,0.0103664696,-0.1340817064,0.2029551715,-0.0630800575,0.0187174585,0.1050214544,0.0810770169,0.0859403238,-0.0236257315,-0.5783276558,-0.181066215,0.4227893949,-0.0685619339,0.2102206647,0.3711635768,0.0659886152,0.2114470005,-0.2180529833,0.6425334811,-0.4958380461,0.2859469652,-0.4440027177,-0.2989968359]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/256","title":"[Feature request] Add a feature to dataset","comments":"Maybe you can use `with_indices`?\r\n\r\n```python\r\nnew_info = list of length dataset['train']\r\n\r\ndef some_function(indice, x):\r\n # adds new_info[index of x] as a field to x\r\n x['new_info'] = new_info_x[indice]\r\n return x\r\n\r\ndataset['train'] = dataset['train'].map(some_function, with_indices=True)\r\n```","body":"Is there a straightforward way to add a field to the arrow_dataset, prior to performing map?","comment_length":35,"text":"[Feature request] Add a feature to dataset\nIs there a straightforward way to add a field to the arrow_dataset, prior to performing map?\nMaybe you can use `with_indices`?\r\n\r\n```python\r\nnew_info = list of length dataset['train']\r\n\r\ndef some_function(indice, x):\r\n # adds new_info[index of x] as a field to x\r\n x['new_info'] = new_info_x[indice]\r\n return x\r\n\r\ndataset['train'] = dataset['train'].map(some_function, with_indices=True)\r\n```","embeddings":[-0.1078236252,-0.0950366557,-0.1245010421,-0.1514483988,0.1871568114,0.3007179201,0.1838353723,0.2558493316,0.3210432529,0.1587405056,0.528386116,0.625244379,-0.2347141653,0.3278807402,0.1876332462,-0.05337587,0.0141954916,0.2249573022,0.0846671462,0.0613886155,-0.3617742062,-0.4533320963,-0.006954384,-0.1957277209,-0.1881195903,-0.0918976218,0.0972512886,-0.3355694413,-0.131321758,-0.5429760814,-0.0301965326,-0.1167532429,-0.0923570395,0.0524035841,-0.0001186418,-0.0422028378,0.2212127596,0.1339968294,-0.0558137521,-0.1594056785,-0.5856688023,-0.5728530884,0.1825901419,-0.3350245357,-0.0429508463,-0.2290669084,0.0593489781,-0.4260806143,-0.1803305894,0.0572605245,0.1637644172,-0.0679824352,0.233760044,-0.0360014699,0.4558517039,0.1536865234,-0.3162890673,-0.2017423809,0.0060712672,0.0041402406,0.2546449006,0.2826150358,0.1268410981,-0.3548519313,0.391894877,0.1401714236,0.155552879,-0.2356152833,-0.2917853594,0.0804627389,0.5871175528,-0.2631477118,-0.4045743942,-0.0798135847,0.2646789551,-0.4920731783,-0.0741465986,-0.1404269487,0.1963732094,-0.0694496855,0.2091938555,-0.4332793653,0.0142719373,0.2940422297,0.2138616443,0.545632422,0.1143340468,0.0920650661,-0.0423269831,-0.1644961834,0.074111186,-0.0618493184,-0.0081348037,0.4773364663,-0.2043196112,-0.2738350928,0.127216801,-0.4587990046,0.044262033,0.1155173257,-0.0419312678,-0.0565995537,-0.1757650077,0.1606503874,0.1501177251,0.1076475605,0.0052302694,0.2677212358,-0.1015377939,-0.35801211,-0.0443412885,-0.2704422772,0.2254671454,-0.0649589673,0.3445422053,0.047936812,0.275036633,0.0916398391,0.1036507785,0.0120589342,-0.1663980633,-0.1238511652,0.0839854181,0.1230355203,0.2344736904,0.0054374924,-0.1229794025,0.3372631669,0.3458981216,0.3112016022,-0.107442759,-0.2126322687,0.1547345072,0.1393012255,-0.0765942559,0.0856375098,0.1823187768,-0.0269806851,0.1582476944,0.1024156809,0.2272258848,-0.1828977764,0.2606956363,0.2110901177,-0.1837568283,0.2089575231,0.0518257543,0.1167875156,-0.3144630194,0.3690935671,-0.2824328244,-0.3489378691,-0.0826415867,0.1523860693,0.1272735447,-0.3446595073,-0.19757393,0.4396517873,0.037065085,-0.189479351,0.3492400646,0.2072193325,-0.4341925681,-0.31297189,-0.0276727639,0.1455973983,0.0120927002,-0.0448054299,-0.2212402523,0.0928503871,-0.2608798444,-0.2146037519,-0.1981479973,0.2057473361,-0.0680169538,0.3534968197,0.8596593738,-0.1093169302,-0.3311024308,0.1419070363,-0.3039422035,-0.0511692688,-0.1390286088,0.5475512743,0.1005674899,0.0178816319,0.0748453587,0.2675637603,-0.1341206729,-0.081064418,0.0801983178,-0.0713017732,0.3110402822,0.0757067204,-0.1374980658,0.0827305615,0.2140202373,-0.4347311258,0.0104316166,0.1827216893,0.0353515632,0.1471180767,0.4865828753,-0.0622773357,-0.139246285,-0.4023250043,-0.315068692,0.0850114971,0.2560750246,-0.1315557212,-0.0564922057,-0.3838046789,0.0411904715,0.3376157582,-0.0930948332,0.1920477748,0.0180984214,-0.2580021322,-0.2771020532,-0.1142548919,-0.3849006593,-0.2202653587,0.0025686605,0.0340221003,0.0429672673,-0.0383333191,0.2549908161,-0.069407545,-0.4769134223,0.1799545288,0.0135041503,0.0666203126,-0.203277275,0.1755382866,0.2284716219,0.0279199891,0.3076652288,0.320923239,0.1065977663,-0.219881326,-0.0806425959,0.2785798311,-0.0108550657,-0.1345078349,-0.1162842661,0.2652430832,-0.0090507288,0.2704108357,-0.1731457859,0.0539249368,0.1970151514,0.0934338197,-0.2376272976,-0.1993034482,-0.3792415261,0.0042476188,-0.1010407358,-0.1289854646,-0.0370637812,0.0835909918,0.3455303311,-0.1041750461,0.1496624351,0.1654145569,-0.1737333685,0.2314660996,0.0443180017,-0.2211854011,0.2762586772,0.253798306,-0.0167220104,0.0387470871,0.0860501677,0.1844401509,0.2718241513,-0.0072902129,0.2178813666,0.0403656326,0.3167016506,0.076929979,0.0469002984,-0.18509534,0.0664847791,0.1358944625,-0.1457649767,-0.181527555,0.1890415996,-0.0139392493,-0.0700866729,-0.1962982118,-0.1267551035,-0.2679468095,0.027739577,-0.0038654837,-0.1771871299,0.1943042427,-0.3962599337,0.0961064696,0.2864940166,-0.6227661371,0.0914276838,-0.0902328119,0.2919758558,0.0225172043,-0.023435317,-0.1501042843,0.3781115413,0.5751972198,0.3162941039,-0.3572919965,-0.4256483614,-0.1076565832,0.0035933314,0.0003379413,-0.0602452047,0.040925093,-0.3032838404,-0.2962276042,0.0606922954,0.1335870475,-0.1346251369,-0.2070714682,0.0026516218,0.0664110258,-0.053892225,-0.1274783313,-0.145945847,-0.3532822728,0.4266149402,0.1954131424,0.1405219287,0.0611188598,0.1954496801,-0.0433334298,0.0780377537,-0.2725583911,-0.119337596,0.0527477227,0.410685569,-0.1730504781,0.1069868356,0.0351223573,-0.4268287122,-0.2257262766,0.3809039593,-0.2895036936,-0.0058323168,-0.1255987734,0.5389932394,-0.1341796219,0.3405480385,0.4850676656,0.2227980793,0.0294600651,0.0210356098,-0.1961565912,0.0219083466,0.4201479554,0.1148672998,0.0238693971,0.3983840644,-0.261421144,0.6585185528,0.0316712223,-0.0323824994,0.2063956261,-0.3531016707,0.1559422314,-0.2816554308,-0.22521469,-0.3181138933,-0.003316835,-0.2853698134,-0.2781780362,-0.1155590937,-0.3026209176,0.134417668,-0.0703768283,0.0452069789,-0.0313163549,0.3276205957,-0.6335791349,0.4230048954,-0.1774734706,-0.0864528269,-0.5579728484,-0.0281279795,-0.05671167,-0.171790272,0.3356387317,-0.4589656293,-0.3039745092,-0.3026862144,-0.3731932938,0.0912011117,0.3420955241,0.0302977059,0.0303825401,-0.2465752214,-0.1940120906,0.219925344,0.3768099248,-0.6040028334,-0.3728715479,0.0722521991,0.189648971,-0.2059538066,0.1057901159,0.1349869072,0.3845051229,-0.2128749788,0.6205476522,-0.1530641764,0.2649077773,0.3790736198,0.0336124934,-0.1378072351,-0.0759076253,-0.0962366462,0.0834231749,-0.0278922152,0.1827480346,-0.0310942736,-0.1528668106,0.3558150232,-0.3106110692,-0.3990319371,0.0445553325,-0.0276521631,0.0291532837,0.2408657223,0.3204190731,0.2974684834,0.036419332,-0.0602959432,-0.1577948481,0.3897464573,-0.1209320053,-0.2056259215,-0.2824397683,-0.1346027106,0.7198033333,0.1317309439,0.1265767366,0.6759861112,-0.1625969708,-0.009566037,-0.4439994991,-0.1029637083,0.5182792544,-0.1229854375,-0.2245127559,-0.5432346463,0.6463309526,-0.0048043896,-0.0282947365,0.5440329313,0.2531407773,-0.1354517192,0.5383806825,0.1178227514,0.8973997831,-0.1386689246,0.0428333543,0.1214509085,-0.0717411563,0.2672677934,-0.1020571589,0.0257764328,-0.1664579064,-0.4732201099,-0.0056553637,-0.1722011864,-0.0494398102,0.219719246,-0.1759848595,-0.0735456795,-0.2605407834,0.4527991116,-0.1141919494,0.1135432869,0.0445237532,-0.1808194965,0.0062238597,0.0587207824,0.1578676105,-0.1414965838,-0.1113810837,-0.214601323,-0.1542651951,-0.1647501439,-0.1084087193,-0.0796297416,-0.3826559782,0.2215468287,-0.2301650494,-0.1839784831,0.1791689396,-0.1137899756,-0.0146226669,0.0445290618,0.0196476076,0.1710761338,0.4610592127,0.3057665825,0.082207419,-0.051059451,0.0550204925,0.1192964166,-0.242134124,0.380489707,-0.1333564073,-0.6124716401,-0.1194011047,0.1356033981,-0.0315112732,-0.3286578655,-0.416554749,0.3146192133,-0.1165780947,-0.0995833576,0.0975473151,0.0506897345,-0.1680639535,0.2601736784,0.2528355718,-0.101414375,0.0754028782,0.0831803456,-0.0391134098,0.2192426026,0.3950485885,-0.3613190353,-0.0834449232,-0.0595598668,0.1318780333,0.4349039197,-0.1364325434,-0.1147079468,0.0125862733,0.0273435991,-0.2294478416,0.0816965103,-0.2032317668,0.4445791245,-0.3331821263,-0.1299837232,-0.5196351409,0.3745311797,0.0392712764,0.2512502372,-0.2702799141,-0.0673567951,-0.3912752569,0.1943269521,-0.2437036633,-0.0980274305,0.2978036702,0.2728988528,0.1146916747,0.3928788304,0.2577614486,-0.2183409631,0.1542443335,-0.0865044817,0.062284179,-0.2051128596,-0.0292495899,0.1338011473,0.0298959948,-0.2434543967,-0.2404440939,-0.3856900334,-0.0583352,-0.2634608448,-0.2187697142,-0.0388516188,-0.1461034864,0.0160241406,0.0778333098,0.1017725021,-0.0719098747,0.1752558202,0.0417016745,0.1398138851,0.0311976504,0.1463407874,-0.1106079966,0.1895393431,0.0193010643,0.0242182929,0.0564775243,-0.049092494,0.316378206,-0.091258578,0.1130166575,0.2513692677,0.3021382093,0.0482381172,-0.2788463533,0.2814736664,0.0508214831,0.2131782621,-0.3302413225,-0.0028981126,0.3826183975,-0.0630896017,0.288356632,0.215255633,0.1268400103,0.477486372,-0.1902843565,-0.086613223,-0.3725791276,-0.1918226033,0.1518883407,0.5361402631,0.254422158,-0.1971865594,0.1784892082,0.4863780141,0.3134151697,0.4343241453,0.2641363442,0.2552035451,0.4459419847,-0.263587594,0.1286549568,-0.2262490243,0.2911535501,-0.0178592391,0.1214748546,-0.0169968344,0.5025038719,-0.0656116605,0.1153911725,-0.1599333733,0.1441002488,0.0818688795,-0.0478898324,-0.0395784527,0.4195753932,0.0110246781,-0.0604544841,-0.265460372,0.0266909953,-0.2186066806,0.1847662628,0.0014630919,0.3479412496,0.1977501661,0.4370705187,0.3695476353,0.4313194454,-0.1343776584,-0.1909075826,-0.2854827046,-0.0731592998,-0.1457272321,0.2491332591,-0.0551345982,0.0963190347,-0.3794454634,0.5683107376,-0.0456500314,-0.266143918,-0.0055396017,-0.0142862508,0.1435000151,0.0050989268,-0.1383124888,0.1190921888,-0.0962361768,0.2658429444,0.009951896,0.2420334965,-0.277189672,0.1542340815,0.2391846031,-0.4078318477,-0.1245934516,0.026088208,0.1542280316,-0.1914583743,0.4852948487,0.0076216091,-0.0137679977,-0.0214271713,0.0503801778,-0.0738635436,0.0286055803,0.2243572623,0.0971973315,-0.2534860969,0.0217254423,-0.3139698803,0.0183783602,-0.0639911741,-0.0559776872,0.3210718632,0.1084072888,0.2041660696,-0.0006587787,0.2546184957,-0.4506436586,0.1094229594,0.2318435162,0.0801695213,-0.2506469488,-0.2650771737,0.0835507587,0.0279513747,0.0254489947,-0.0869682357,0.1255206317,0.0404844955,-0.201833874,-0.2339788377,-0.1614364386,0.0295683332,0.5203288198,-0.2210360616,0.2648938298,-0.258179903,0.1299267709,-0.0547797866,-0.2884075642,-0.3066644073,0.2825362086,-0.0308529548,0.1520637572,-0.0400991738,-0.066042833,-0.2453926206,-0.0955058038,-0.3251925707,-0.5171813369,-0.2370894998,0.1982637495,-0.2143962532,0.0078141056,-0.1609614342,0.0979323387,-0.0672003403,0.241296351,0.2947176695,-0.084528245,0.2829105854,0.0013365875,-0.017346397,0.0220883768,0.0324637182,0.2264678478,0.1670636833,-0.4723065197,-0.2758606374,0.427467376,-0.0708139092,0.0458423495,0.1943082511,0.0271451361,0.262915045,-0.141474694,0.4539399147,-0.3408811986,0.3687354028,-0.3093938529,-0.2806949317]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/256","title":"[Feature request] Add a feature to dataset","comments":"Oh great. That should work. I missed that in the documentation- thanks :) ","body":"Is there a straightforward way to add a field to the arrow_dataset, prior to performing map?","comment_length":13,"text":"[Feature request] Add a feature to dataset\nIs there a straightforward way to add a field to the arrow_dataset, prior to performing map?\nOh great. That should work. I missed that in the documentation- thanks :) ","embeddings":[-0.1759447008,-0.0733822212,-0.1920109987,-0.3037525415,0.0417599045,0.1488418132,0.0346042514,-0.0170161799,0.0460622013,0.1123961508,0.595590353,0.6084624529,-0.2051133364,0.3813489974,0.2170944512,0.0076125674,-0.0108325379,0.2042423785,0.0296585858,0.0058115721,-0.2059879303,-0.2643194199,0.0011693994,-0.2415930033,-0.2208486497,-0.033057332,0.1042241082,-0.2014854848,-0.3593073189,-0.4976098537,0.0551552474,0.0492845327,-0.075785622,-0.18980214,-0.0001112245,-0.2438334972,0.3065032661,0.0918232948,-0.0483567268,-0.0302910786,-0.6384589672,-0.4435905814,0.2224812657,-0.115723066,0.0504926369,-0.1066736653,0.1248935163,-0.3191322088,-0.095235303,-0.0390145779,0.2474898696,-0.0070581776,0.1878922582,-0.0952916145,0.305785358,0.485832721,-0.4654582739,-0.2230228037,0.0986990556,0.0354077742,0.1931687593,0.3301906586,0.4036188126,-0.2535547912,0.4949661493,0.0120538129,-0.0308156535,-0.166161105,-0.109607406,0.1072298214,0.8959718347,-0.1997755766,-0.3248124123,0.0919918045,0.0751302391,-0.308005631,0.0472406484,-0.0646991432,0.0928382799,-0.0448061936,0.1810807437,-0.6437373757,-0.1254555285,0.0619288944,0.172100842,0.3427742422,0.056829121,0.0215918217,-0.0751348585,-0.1433959603,0.1476491392,0.0897653773,-0.2590329051,0.5471345782,-0.172094509,-0.2818696201,0.2227012366,-0.1920773089,0.1385212243,0.2543992102,0.1764314175,0.0471491404,-0.3386818469,0.0838582739,0.1370497942,-0.0502337292,0.2019278109,0.0070030997,0.1488024443,-0.5398797393,0.071756795,-0.2692034841,0.1839501411,-0.2081000805,0.2088811547,-0.0095568011,0.4241875708,-0.2103933841,0.3040506244,-0.281291604,0.1438469291,-0.0777006447,-0.0409664661,0.0635333285,0.3448015749,-0.1625727713,-0.1122273132,0.2676674128,0.2695740759,-0.0117477756,-0.12772578,-0.2054504752,0.1205706298,0.2515414655,0.0438670143,0.0440148488,0.0799010918,-0.1124496087,-0.0289897788,0.1556315273,0.2398239225,0.1160014644,0.2260892689,0.2621347308,-0.1218148693,0.0816179514,-0.1144728586,0.458735466,-0.2181420773,0.3534145951,-0.1350898743,-0.2884908915,-0.2094864547,0.1569390446,0.1053900346,-0.4012045562,-0.3060803413,0.463357091,-0.1158992201,-0.1404856443,0.2023733407,0.3104778528,-0.1793439537,-0.3234949708,-0.1048511043,0.130946368,-0.2225179374,0.0526220053,-0.3565254211,-0.0880325213,-0.25321877,-0.3370993733,-0.2551502287,0.1900243014,-0.0565776527,0.4077303708,0.7711873651,0.082964316,-0.1602505296,0.0489903502,-0.3361584544,-0.2292371392,-0.1590624601,0.5162587166,0.0269715805,0.0595859215,-0.1958554536,0.2351290733,-0.20327878,-0.0201140083,0.0207378827,-0.0287145227,0.2034660727,-0.0464026742,-0.1066017747,-0.0146729061,0.4297304153,-0.2690733671,-0.1580260992,0.1549940109,0.0672026798,0.1366843134,0.6394635439,-0.1998380572,-0.1289678514,-0.1383248121,-0.4076742232,-0.0083827702,0.0405422151,-0.0870276019,0.0374885462,-0.3183628321,0.0923454762,0.3159960806,0.02374571,0.0926294774,0.0419945009,-0.1961257905,-0.2124907076,-0.1360932589,-0.313835144,-0.2463852167,-0.0117659848,0.0627397522,-0.0367532484,-0.0541328043,0.157144621,-0.1004553884,-0.2770577967,-0.0140291778,0.0454916023,-0.0049707945,-0.2471606284,0.2794454396,0.2747302651,0.2250546366,0.489841789,0.2481849194,0.1791098267,-0.3299202323,-0.0432663709,0.0882138163,-0.1112016216,-0.1639324725,-0.3578274548,0.1822962612,0.061235711,0.1304555386,-0.0898539945,0.0932942033,0.3306238055,0.1403736323,-0.2256841511,-0.0852602348,-0.1946161091,-0.129000321,-0.0231623575,-0.0937029496,-0.0544682369,0.264039278,0.2134603113,0.0462740101,0.1013944447,0.2260578275,-0.2014461309,0.0848286822,0.2510961294,-0.1164901629,0.3356184959,0.2461070716,0.0292861406,-0.0311410669,-0.0721651092,0.0228961874,0.195192486,-0.109509483,0.1711039841,-0.1392857283,0.4284159243,-0.0978800282,-0.1961822659,-0.0614874214,0.0534355976,0.0554629155,-0.1044739112,-0.3794037402,0.0720689893,0.0878792331,0.0678488612,-0.1331778318,-0.1964471638,-0.4041184783,0.2437756509,-0.0522014201,-0.2055332214,0.1712998748,-0.4008053839,0.3773549795,0.0181100741,-0.1986295432,-0.0269071143,-0.1295323819,0.1733337492,0.0825187266,0.0229714792,-0.0794078484,0.5610334277,0.5177881122,0.4958937764,-0.4893189371,-0.7186246514,0.0031198333,0.1103672013,-0.1324142814,-0.0208378267,0.014996809,-0.1802867353,-0.1644963473,-0.0129596321,0.0432465598,-0.2839616537,-0.1808524728,-0.1100297421,-0.0505092107,-0.1389073133,-0.1718642414,0.01963735,-0.2971659601,0.5095891356,0.1247996166,0.0743419901,0.167059198,0.1048156694,-0.0382466987,-0.1286999881,-0.2866785228,-0.1390971541,0.0650755465,0.4304070771,-0.504113555,-0.1002053171,0.2470229566,-0.1899534315,0.0271194335,0.3165247142,-0.316300422,0.0568572022,-0.0579582267,0.4452473223,-0.1084705889,0.0498865135,0.536970973,0.2458848506,-0.0686956942,-0.0900681093,-0.1618446559,0.1550595164,0.453130275,0.2329597175,-0.0848234892,0.2153467834,-0.1686339378,0.6327844858,-0.3037899435,-0.0288593303,0.2320919037,-0.2129059434,0.263835609,-0.1253626198,-0.0252138805,-0.1899874508,0.0309835467,-0.1824730039,-0.227408722,0.0664588511,-0.2953310311,0.0308150351,-0.01155305,0.1801395267,0.0149291186,0.1776416749,-0.5426855683,0.4097107947,-0.0184093267,-0.2120165974,-0.5512978435,0.0144996736,0.0296334736,-0.0740415081,0.4458283782,-0.4113937616,-0.4922577143,-0.1469960958,-0.370811522,0.106431298,0.096537374,-0.0180388819,-0.2116333097,-0.1550483555,-0.1942383498,0.1412564814,0.3423202932,-0.5900896788,-0.2477371097,-0.0877875239,0.303013742,-0.100831598,0.0843197703,0.1961771101,0.3526089489,-0.2266382128,0.4880464673,-0.131632641,0.2656256855,0.4100489914,-0.1533746868,-0.2306889892,-0.0195169616,-0.0332613811,-0.1937988102,0.0472520478,0.0872453675,-0.0086333128,-0.2983537912,0.3232935369,-0.3490186334,-0.2473191768,-0.0823447704,-0.2197220027,0.0699409917,0.1735234857,0.2561342418,0.2938518822,0.0047267373,-0.0663787797,-0.0234916117,0.3810061216,0.0877233073,-0.1735997796,-0.090668343,-0.3539049029,0.7498205304,0.211521998,0.0773003921,0.5638552308,-0.1244944334,-0.1909977049,-0.3640868068,0.0259907097,0.354080528,-0.2412651479,-0.0900705382,-0.3541325927,0.6308782101,-0.1062906384,-0.0355465338,0.6644345522,0.3536255956,-0.1561345011,0.5067573786,0.1490960866,0.9217481613,-0.0757349283,-0.1308370978,-0.0544849336,-0.327765882,0.3209289014,-0.1963284165,-0.0672444776,-0.1308742613,-0.3415208459,-0.0624738783,-0.1732718796,0.1107811257,0.2217134237,-0.1246354058,-0.3157531917,-0.0851817951,0.4309114516,-0.0703110397,0.3053085804,0.2038797736,-0.285571754,0.1434053779,0.1510976255,0.0352438726,-0.0946107879,-0.1458805799,-0.247580111,0.0269539561,-0.2556675971,-0.2042917162,-0.0923560783,-0.3938592076,0.0515766144,-0.1812850386,-0.0352962464,0.171188131,-0.1311840862,0.0626868084,0.0781110302,-0.1333906651,0.3390014172,0.3014665544,0.2827788889,-0.0436083376,0.0258998238,-0.0619130991,-0.0892335027,-0.1196411476,0.3073912263,-0.2535009086,-0.5707023144,-0.1328333914,-0.0677035898,-0.0933374763,-0.1862901002,-0.28427127,0.3122891188,-0.1346562505,-0.148265034,0.1591602862,0.0757816359,0.0201517027,0.155408293,0.3921865225,-0.0035730014,0.0764590427,0.0440043472,-0.1606482118,-0.1069919616,0.2352171391,-0.2236848921,-0.0948253199,-0.1349978596,0.1364825219,0.3430036306,-0.3970170319,-0.1395143569,0.0758047998,0.0305031259,-0.1695845723,0.285462141,-0.1811236441,0.253292352,-0.31456092,-0.0563513711,-0.3803770244,0.1879081875,-0.1696093231,0.3415038288,-0.271553576,-0.1091736332,-0.1403889805,0.3906421661,-0.3274582922,-0.2291321605,0.3621339798,0.3122864664,0.035373684,0.4304292202,0.1009896025,-0.2770510018,0.1309297681,-0.0163156148,0.0228664782,-0.1912415773,0.078393057,0.181237027,0.0948316306,-0.2290995717,-0.2030075788,-0.3033901453,-0.1436714977,-0.0272831898,-0.1530521214,-0.1956584901,-0.2253038585,-0.0722256452,0.0058962861,0.139968425,-0.0104205692,0.2435410172,0.2506772578,0.2481056154,-0.0310398806,0.0967946276,-0.1269595623,0.1132507697,0.3153254092,0.1105983928,-0.0727742612,0.0151272556,0.1181962788,-0.036609754,0.0214328486,-0.0449972823,0.214914754,-0.0761751086,-0.2247852981,0.3905928135,-0.0077108489,0.2206055522,-0.2195782065,0.0311974194,0.3938325644,-0.0659769624,0.1587066054,0.1233813912,0.3212665021,0.5096278787,0.1133506224,0.1183795705,-0.235823676,-0.3713224828,0.1258071065,0.3653272986,-0.0704204291,0.0147193056,0.3976859152,0.4479571581,0.2736951709,0.3090638816,0.0899485052,0.2570383847,0.528449893,-0.194883287,0.2011304647,-0.2169507891,0.247873351,-0.0589068644,0.2943241298,0.0356394462,0.4939140379,-0.0498885922,0.1575262696,-0.2055228651,0.0013182198,0.0318097025,-0.0695879161,0.0172527432,0.1327215433,-0.1380553842,0.0634795502,-0.1290285736,0.0602989867,-0.1606334746,0.329682529,-0.0415491909,0.3666081429,0.2011394203,0.4699809849,0.2646303177,0.1838025004,-0.0698144287,-0.0814351588,-0.3020267487,-0.1410304755,0.0276899356,0.2468126714,-0.1618933082,0.1577748358,-0.30534181,0.540478766,-0.0710220635,-0.2629101276,0.0353296101,0.1783780605,0.1733336449,0.0327229239,-0.0884035379,0.1564578712,-0.0487074777,0.383995533,0.0917257071,0.2186650336,-0.1738818437,0.016370533,0.193623364,-0.4201314747,-0.0875040069,0.1044017673,0.260751605,-0.3903516829,0.4288172424,-0.1055531874,-0.1075457335,-0.1590131074,0.0976660252,-0.0651312843,0.2530398965,0.2197268754,0.0392568111,-0.1791386455,-0.1670699567,-0.4713461995,0.042954579,0.0819193721,-0.0260123257,0.3006361723,0.1247145161,0.2919878066,0.1810035855,0.3848168254,-0.387802124,0.1981367469,0.0506188609,-0.0343500264,-0.1304331869,0.1055061892,0.1135013178,-0.139242962,0.1789427251,0.1164491996,0.2094475478,0.0559676029,-0.2658105493,-0.2905706167,0.0580536462,0.1172504053,0.4766252935,-0.2587964833,0.304125011,-0.1974978447,0.1291742176,-0.1513456553,-0.2842588127,-0.2017819881,0.0902022421,-0.0492575802,0.0903682485,0.0672260821,-0.0913735256,-0.2973533571,-0.0343988016,-0.3701976836,-0.2638867199,-0.2673138082,0.3149237931,-0.1187064499,-0.0827099532,-0.0444960073,-0.1633861661,-0.0521353483,0.0944548696,0.0943080336,-0.1435679495,0.1975381076,-0.0049306341,-0.0231339689,0.1689380258,0.0288334396,0.3615996838,0.1056978405,-0.5903561115,-0.1513820291,0.3293605447,-0.0909588113,0.1984420419,0.3765233159,-0.0813870728,0.1366890967,-0.2093211561,0.6297402978,-0.2184847891,0.3485204875,-0.4421598017,-0.2188996375]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/254","title":"[Feature request] Be able to remove a specific sample of the dataset","comments":"Oh yes you can now do that with the `dataset.filter()` method that was added in #214 ","body":"As mentioned in #117, it's currently not possible to remove a sample of the dataset.\r\n\r\nBut it is a important use case : After applying some preprocessing, some samples might be empty for example. We should be able to remove these samples from the dataset, or at least mark them as `removed` so when iterating the dataset, we don't iterate these samples.\r\n\r\nI think it should be a feature. What do you think ?\r\n\r\n---\r\n\r\nAny work-around in the meantime ?","comment_length":16,"text":"[Feature request] Be able to remove a specific sample of the dataset\nAs mentioned in #117, it's currently not possible to remove a sample of the dataset.\r\n\r\nBut it is a important use case : After applying some preprocessing, some samples might be empty for example. We should be able to remove these samples from the dataset, or at least mark them as `removed` so when iterating the dataset, we don't iterate these samples.\r\n\r\nI think it should be a feature. What do you think ?\r\n\r\n---\r\n\r\nAny work-around in the meantime ?\nOh yes you can now do that with the `dataset.filter()` method that was added in #214 ","embeddings":[0.1524707824,-0.0686902627,-0.146965012,-0.2211095989,-0.3081418276,0.0451667421,0.221041292,0.3083456457,-0.0878339708,0.3214425743,0.1547526717,0.3221749067,-0.1845388561,0.2033200562,-0.0725193173,-0.072534658,-0.1344357282,0.2064556628,-0.0416877307,0.1055373028,-0.1516434848,-0.092831254,-0.1631123573,0.0364665538,0.0001396739,-0.1590100527,0.1928139776,-0.3438686132,-0.23199597,-0.2637731433,0.2420486361,0.5169215798,0.0783414468,-0.1050742641,-0.0001050023,-0.1977978498,0.2854300737,0.0252290405,-0.3200468719,0.042735137,-0.2964933813,-0.0042866352,0.0999666303,-0.0253435373,0.05831879,-0.0285041407,-0.1237535253,-0.1715595424,0.0183938108,0.2911897302,0.3215633631,0.2354656458,-0.3738358021,-0.0379895419,0.1803824455,0.6441884041,-0.2491365969,-0.1939595491,0.4327890277,-0.138751626,0.0543087535,0.4814007282,-0.1351951808,0.0658717752,0.2164822221,-0.2131249458,-0.1743445992,-0.593526423,0.1151498705,0.3580580652,0.7855474353,-0.1481525302,-0.1514899433,-0.0695586726,-0.0738944784,-0.2212246805,-0.2048683465,0.0793955475,-0.4128407836,0.2999526262,-0.3055996597,-0.2568860352,-0.1369430125,-0.0562187284,0.0188549031,0.0579599068,0.0253004525,0.0301971976,0.2365388125,0.1984406859,0.3114848435,0.1610807925,0.0400727764,0.2651872635,-0.1797466278,-0.3887878358,-0.0115247136,-0.2442950308,0.4376913905,0.3863368928,0.0665562749,0.0049874787,0.0385149457,0.0994900241,0.5406312346,-0.0275661442,0.3464235365,-0.029067656,0.1584667265,-0.0394011959,0.1427334696,-0.0418224297,0.4167382121,0.0253681522,0.058960665,0.1128173321,0.4688585401,-0.2380289137,-0.4363637865,0.0637960881,-0.2282157838,-0.1035118997,-0.1743416339,0.01791027,0.4964860082,-0.1212322488,-0.2533375621,-0.025421001,-0.2252066582,-0.4202360213,-0.098812297,0.1154697761,-0.1027311683,-0.280154258,0.1578596085,-0.0488454029,0.3315295577,-0.1067183092,0.1442272365,0.02032125,0.3713670671,-0.099508293,0.4219819307,0.3568205535,-0.2871767581,-0.0621601269,-0.105441913,0.1676531732,-0.2552564144,0.4463689625,-0.0001658838,-0.2190548331,-0.2839265168,0.1990804374,0.0979910195,-0.1039063036,-0.0032374689,0.3359579444,-0.282286793,0.1548547745,0.1022531539,-0.0610412173,-0.1628772169,-0.0102469632,-0.1115598232,0.2041341662,-0.4011398852,0.0656782538,-0.3449183702,-0.3245483339,0.1461312771,-0.091811806,-0.3562164307,0.1757980436,-0.1945919693,0.2514718175,0.2005859464,-0.0443770774,-0.2376707643,0.3638539612,-0.0864177123,0.0928181112,-0.24696064,0.1849029511,0.105954662,-0.1491414905,-0.3297962248,0.0477576442,-0.1588001549,-0.1968523711,-0.0313144699,-0.3938298523,0.3564226329,-0.0755006671,0.3161650002,0.2220146209,0.2586925626,-0.1761614382,0.3409664035,0.0820167735,0.1753762066,0.100949809,0.699893713,-0.1835140735,0.0040006731,-0.3151400983,-0.2812793553,-0.0516267717,-0.1388118118,-0.1276285946,0.1939436048,-0.5064941645,-0.0145200659,-0.0448360369,-0.0656428561,-0.2465183884,0.0782189518,-0.2371264845,0.0340741426,-0.3830653727,-0.5653280616,0.0145148467,-0.2014602572,0.0020714083,-0.1319381595,-0.1468449831,0.2393127978,0.1131776944,0.1452044845,0.19616732,0.065799199,-0.0660896227,-0.0890317261,0.2003376037,0.278440088,-0.2950909436,-0.0846454278,0.0359112807,0.0992265791,-0.1669364274,-0.0112247188,0.1364192516,-0.0330095515,0.1066735014,-0.2476149946,0.3743722737,-0.2114716917,0.2994423211,-0.1655457765,0.0740624666,0.2526150346,-0.234357208,-0.1235059276,-0.1797049493,-0.0299263373,0.0271663107,-0.0210894514,-0.1740151197,-0.2960761487,0.1143345088,-0.0525955185,0.1503941715,-0.067528367,0.2828569114,0.113530077,-0.1267099828,0.1603176296,0.171514228,0.3910123408,0.1552311182,-0.0733541772,-0.0245978702,0.4557509124,0.0382879712,0.107889384,0.1832901239,-0.134327352,-0.0199933816,-0.0428867117,-0.0844735503,-0.3788739145,-0.0567793995,-0.0283361599,0.0868192315,-0.0168339871,-0.055122938,-0.2023082078,-0.5482844114,0.0757297054,0.1905257553,-0.1402407736,-0.1396197975,0.2762286663,0.1673987806,-0.2793147862,0.239706248,-0.0718530342,0.6686889529,-0.312120378,0.3265020549,0.0277346727,-0.0108191529,0.0574962646,0.2458640635,-0.1657044739,0.3721702695,0.605910778,0.2086495608,-0.1610233188,-0.4974036515,-0.2689962089,0.1812807322,0.0188983157,0.2495717555,0.2859071195,0.1040226966,0.0129304286,0.0437634364,-0.3359830081,-0.2038429976,-0.3055730462,-0.3022105098,0.0553853922,0.3313924074,-0.2298619151,-0.0797358006,-0.2824411094,-0.2069191635,0.1034878343,-0.0693598613,0.2219944745,0.0409385003,-0.227632612,-0.1130291745,-0.1559122652,-0.2918018997,-0.2431513369,-0.4499205649,-0.2808900476,-0.2144709527,-0.2396109104,0.0452260077,0.2854363918,0.0309104268,0.1576340348,-0.1407988369,-0.3538701236,-0.0193966497,0.2619760931,-0.0710366964,-0.0738758221,0.1476152688,0.0599192418,-0.065450646,-0.0256453324,-0.1464360058,0.0759128928,0.1596871167,0.006143813,0.1668523401,0.1987110078,0.0561457425,0.6262665391,0.1465247124,0.0718777105,0.2946554124,0.0982772708,0.2543916702,-0.1454623938,-0.2613230348,0.1503057778,0.0277463943,-0.1762808263,0.3424513936,-0.1752699465,-0.3346106708,0.1341099143,-0.0956781507,0.0274367388,-0.2173796296,0.0990592241,0.0027676113,0.2048640549,0.1039608344,-0.0469423272,-0.1334152967,-0.1100610718,-0.051242657,-0.023184523,0.5291541815,-0.1040534154,-0.7140022516,-0.0883681551,-0.3917911351,0.1126012951,0.1015192792,0.136847958,-0.0155847864,0.0694454461,0.0883977041,0.3237752318,0.4941772521,-0.1265718639,-0.1096089184,0.0995548368,-0.1879328638,0.2218267769,0.0690922812,0.1223987862,0.5511656404,-0.0633068383,-0.0556180738,-0.0565514304,-0.3648303449,0.5033463836,0.4331119359,-0.2701098025,-0.0743386745,0.1339772046,-0.4539625645,-0.1623749137,0.1874197274,0.3584504724,-0.0177895967,0.1046855822,-0.2119674087,0.0147275897,-0.2031919956,0.2260521203,-0.1402352899,0.0924686044,0.1695163697,0.0972865894,-0.0337339453,0.0443914831,0.335993588,0.1440193802,-0.2337584198,-0.1483154297,-0.0502981506,-0.4831162691,0.5759854913,0.4025353491,0.066128552,0.480656445,-0.161128521,0.1067109331,-0.1430115849,0.0351004042,-0.0789735615,-0.3594435751,-0.299898237,-0.3773574829,0.1452076435,-0.2390553951,-0.135429427,0.5034415722,0.0663282201,-0.3376877308,0.2600629628,0.1279743761,0.8335614204,0.2101512402,-0.1658747941,-0.3197425008,0.366795063,0.4445437193,-0.0661434159,0.1481736302,-0.1517112255,-0.4650948346,-0.0986901894,0.091161117,0.2458019555,0.1708971858,0.3036312759,0.1425014734,-0.3331363201,0.2382974327,0.096856609,0.2040428519,0.204479143,0.0866391957,0.2796413898,0.1444545537,-0.1537187546,0.1364238709,-0.0146688335,0.0147415288,0.4568224847,-0.1832941175,-0.0878518075,-0.2767695785,-0.2319262028,0.1737441868,-0.1677431613,-0.2251743227,0.037039157,-0.0412893072,0.2246524692,-0.1624141335,-0.0072510624,0.261611253,0.4305461347,0.2613465488,-0.182185486,0.0892930329,0.0674331412,-0.2170285732,-0.1740196645,0.1998308599,-0.0882273167,-0.6302384138,0.0451186262,-0.0370802656,0.1394983977,-0.2621927857,-0.1778276563,0.0797662213,0.1998777539,-0.2662315667,0.1791371852,0.4197267592,-0.1594320089,0.3291675448,0.1419424266,-0.1108380482,-0.1311237812,-0.1965718567,0.2485965192,-0.2229696512,0.2435824126,-0.1665102094,-0.217217207,-0.3615176082,0.0978451371,0.0420763046,-0.2994478345,0.3311510682,-0.157122314,0.1212192401,-0.0768241361,0.3186172545,-0.2623768747,-0.0779756159,-0.1458842009,-0.117030561,-0.1096847355,-0.1249668971,-0.2324270159,0.2765315771,-0.3172044158,0.0709193125,-0.2992806137,0.0949273109,-0.352022022,0.0131937647,0.201930061,0.330915451,0.2108596116,0.4198658764,-0.1643500328,-0.1381740868,-0.0254092403,0.090893954,0.1013619304,-0.2763561606,-0.2395838797,0.1283572912,-0.1529341042,-0.2567759156,0.0781970099,0.1844085306,-0.2350838929,0.0114884349,0.2410909534,0.081247434,0.2036150992,0.3696844578,0.2126162946,-0.3143373728,-0.1465889961,0.2053958327,0.1115212515,0.0902663022,0.0373106301,0.1750398576,0.1016101912,-0.0006025296,-0.0135762431,0.0851758793,-0.0580535345,-0.0368019864,-0.1707370877,0.1592190117,0.1380587071,0.1906599849,0.1146684736,0.0674367845,-0.2013959736,0.0874572247,0.2294460982,0.3001911342,-0.3757338524,0.0477156974,0.2190164328,-0.0372883528,0.0480040573,0.1176505089,0.3379618824,0.2758145928,0.1944613606,-0.1427142918,0.1466784775,-0.0433589034,0.1966171861,0.5363403559,0.07790602,-0.0912932605,0.2090110183,0.2249441296,0.0897134915,0.2429490983,-0.2368642688,0.7157301307,0.4475235343,0.1186498925,0.0449161977,-0.3907006383,0.306162864,0.3439212441,-0.0816653967,0.032626573,0.3975999057,0.054934822,-0.0216747466,-0.322686702,-0.1790700704,0.2538158,0.0093509881,-0.0372956544,-0.2358718961,-0.1525779814,0.0526143238,0.2089958489,-0.0267099775,-0.1131421849,0.2701874673,-0.1351611614,-0.0969169736,-0.0530758426,0.2112284005,-0.3297098875,0.2764938176,-0.1226812303,-0.5608581305,-0.3446208239,-0.0673453808,0.3321212828,0.3554744124,-0.2908454239,-0.1042802632,0.1098853573,-0.1022161469,-0.3607233167,0.0053532165,-0.2886288464,0.1703980118,-0.0004720959,0.1374028474,0.1133887321,0.0933977365,-0.0688142478,0.0351214893,0.3826909363,0.1045230404,-0.7485453486,0.0754265413,0.313344121,-0.1835143864,0.0224799402,-0.0573371165,0.3456203938,-0.0509058461,0.4373703599,-0.0212365314,0.1652185023,0.0199169852,0.143466264,-0.4360460937,-0.0386121199,0.0742475092,-0.238605231,-0.3269577622,-0.1742448956,-0.5445418358,-0.0526623987,-0.088560909,0.0918075964,0.5256829262,-0.0007511599,0.0551996492,0.2635044456,-0.0482157357,-0.2776601613,0.1035159081,0.0400072187,-0.3905744851,0.0418052413,-0.1188071668,0.3993772268,0.1124275848,-0.3543589413,0.0164864305,-0.2146885842,0.0513994694,-0.2378453761,-0.2160301954,0.4501122832,0.2699845135,0.3934665322,0.0797203332,-0.0902164876,-0.0125648612,-0.140745759,-0.1495807916,0.008123857,-0.2722079158,0.0070367255,0.1192824915,0.2322125584,0.0687059239,0.3768313527,-0.216968745,0.3306803703,-0.2001114786,0.1520103663,-0.3559267521,-0.0200535338,-0.0377804488,0.0082599372,0.2659196258,0.1712702811,-0.1545171291,0.0960396156,-0.3604622185,-0.0338159017,0.2870180607,-0.4551140964,-0.2992479801,-0.1718946844,0.3613574803,0.1554852724,-0.5626271367,-0.3370779753,-0.0743286312,0.2759281695,-0.1715849787,0.224470973,0.3413218856,-0.1381219327,0.0154976547,0.1342683583,0.225550428,0.3296641707,0.0109608481,-0.043953862,-0.2093375623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/252","title":"NonMatchingSplitsSizesError error when reading the IMDB dataset","comments":"I just tried on my side and I didn't encounter your problem.\r\nApparently the script doesn't generate all the examples on your side.\r\n\r\nCan you provide the version of `nlp` you're using ?\r\nCan you try to clear your cache and re-run the code ?","body":"Hi!\r\n\r\nI am trying to load the `imdb` dataset with this line:\r\n\r\n`dataset = nlp.load_dataset('imdb', data_dir='\/A\/PATH', cache_dir='\/A\/PATH')`\r\n\r\nbut I am getting the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 517, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 363, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 421, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=5929447, num_examples=4537, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n```\r\n\r\nAm I overlooking something? Thanks!","comment_length":45,"text":"NonMatchingSplitsSizesError error when reading the IMDB dataset\nHi!\r\n\r\nI am trying to load the `imdb` dataset with this line:\r\n\r\n`dataset = nlp.load_dataset('imdb', data_dir='\/A\/PATH', cache_dir='\/A\/PATH')`\r\n\r\nbut I am getting the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 517, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 363, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 421, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=5929447, num_examples=4537, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n```\r\n\r\nAm I overlooking something? Thanks!\nI just tried on my side and I didn't encounter your problem.\r\nApparently the script doesn't generate all the examples on your side.\r\n\r\nCan you provide the version of `nlp` you're using ?\r\nCan you try to clear your cache and re-run the code ?","embeddings":[-0.5058609843,-0.1288104951,0.0182252526,0.3497672975,0.1497054547,0.2042558789,0.0464905091,0.4214467704,0.2376457453,0.0744059384,-0.1168221682,0.1823245883,-0.2613818645,-0.101923354,-0.15735282,0.0979988351,-0.1106186137,0.2100696564,0.0744425282,0.0653068647,-0.4028107226,0.2740860283,-0.3829877377,-0.0700217038,-0.3913931847,-0.1332287341,0.1459162831,0.1515464187,0.191357404,-0.3842667341,0.3481659591,0.0109436065,0.1533981115,0.1625825912,-0.0001161558,0.010485352,0.6431316733,-0.2872529328,-0.2771607041,-0.2786664367,-0.2890476584,-0.2957006395,0.2998387218,-0.3556797504,-0.0921420455,-0.0069964286,0.3332283497,-0.4666482806,0.1093039885,0.1632326394,0.17784594,0.1451750994,-0.1186726689,0.1979513317,0.0227693487,-0.1154360548,-0.1067392081,0.0767500103,-0.0245760996,-0.0722568482,-0.1163303331,0.2309108973,-0.4465887249,0.1689139605,0.3901188076,0.175498262,0.3332034945,-0.2765008509,0.075174287,0.2043976337,0.5286696553,-0.1759230644,-0.2950552106,-0.3949430287,0.0285661221,-0.2426873296,0.164115116,0.3492821455,-0.2348972708,-0.0012560944,-0.3458954692,0.0242989697,-0.2227833569,0.1689743102,0.3957193494,0.1043043733,0.1111679003,0.0797282085,0.2857824266,-0.0581012443,0.4545836747,-0.0850620642,0.1459582746,0.1444033682,-0.2710443139,-0.0759877861,-0.2171054035,0.2746360898,0.0082770418,0.3213576674,0.1967979074,0.0516806841,0.1436003894,0.3109807074,0.2449324131,0.0170768797,0.1431279927,0.3343563378,0.0498044267,0.1574208736,-0.1261125505,-0.0362699814,0.0853060409,-0.0314564854,0.0814510062,0.1571653485,0.2686862648,-0.2099481523,-0.5460675359,-0.0010156493,-0.2332894206,-0.1651888639,0.0959894434,0.2754747272,-0.1536782682,0.3319395185,0.042252142,0.1424370408,-0.3492205441,-0.1216831952,-0.2368981838,0.172750473,-0.2033310533,0.0771291628,0.31588009,0.1044000685,0.4351963103,-0.1986221075,-0.1153535098,-0.0420896672,0.4212338924,-0.1903631389,-0.0745942444,0.2226096541,0.1970767379,0.1623467505,0.0870599002,-0.0507297404,-0.3188578784,0.1813304275,0.0028089394,-0.3774318993,0.0150839789,0.1790364236,-0.1909037083,-0.0862755403,-0.034753602,0.1134413704,0.1302699894,-0.4634093642,-0.2054534554,-0.1073764414,-0.2702164054,-0.1232669726,-0.0165222269,0.5770250559,0.0338287167,-0.0637336597,-0.1441337615,0.008088435,0.5104398727,0.3860457838,-0.0508839116,-0.1670527458,-0.3651423454,0.3389675915,0.2271953523,-0.4194648564,-0.4263260365,0.4837440848,0.006436381,0.2578959167,0.3255821764,0.4054284692,0.1204781532,-0.0325896367,0.0521423072,0.4240181148,-0.1317705214,-0.0142845968,-0.2874594629,-0.2950439453,0.6005905271,0.1494239867,0.0687649548,-0.0436367318,0.0187510271,0.4665596485,0.3268418014,-0.001648157,-0.0108888811,0.1649563313,-0.2478443682,0.3226459622,-0.1736309826,-0.1761586815,-0.5091295242,0.074178949,-0.1129090339,0.090829961,-0.1763046086,-0.1781263351,-0.0882440358,-0.3431604803,-0.2358764261,0.0028419192,0.1064237207,0.2504612803,-0.0604404472,0.139773801,-0.1747403294,0.5486708283,0.0023630376,0.082942225,-0.8483415246,0.3239140511,-0.1733106077,-0.2676151097,0.3004835248,0.2964940369,0.2826043665,-0.1530664861,-0.1141646951,0.3894184828,0.0910348073,0.3174045682,-0.1669467986,-0.0319519527,0.2927332819,-0.363134414,0.1423194408,0.2724507153,0.1341876537,-0.1109847426,-0.0392149165,0.2899010479,-0.3300355971,0.2139120996,0.0790896118,0.0629346669,0.1603439301,-0.1450172812,-0.0745699927,-0.4325863719,0.4534665942,0.1690415144,-0.071861811,0.0622173026,-0.3637724817,-0.1239177287,0.4598456919,0.1139727235,0.0168150458,0.0469442867,-0.0513484478,-0.1667591631,0.1503908932,0.1913670003,0.2858821452,0.3570718765,-0.1138631478,0.0858264789,-0.094161734,-0.1824414432,0.1085890383,0.1541863382,0.2679120898,0.3302231431,-0.1167620197,-0.0563683361,-0.4007021487,-0.2456368655,0.0161414798,0.3377129138,-0.2668766081,-0.1578804553,-0.4097839892,-0.4801176488,-0.056841068,-0.0843887404,-0.1401246637,-0.4054656029,-0.1167454571,0.5066664815,0.0902764648,0.2009690404,-0.386033684,-0.056571804,-0.0384274311,-0.145033434,-0.1585079283,-0.3335529864,-0.0711717978,-0.0457949005,0.3911580443,0.2297028601,0.3629732132,-0.4176098406,-0.0948796198,0.0653366596,0.0540711619,-0.1577545255,-0.0800911337,0.0756564736,0.0688458905,0.0790497363,0.1566694677,-0.3845457137,0.2842551768,-0.2741718292,-0.0760483891,0.1950515956,0.1936430484,0.0419287495,0.0353900678,-0.3608492613,-0.3879698217,-0.4397134185,0.0191146694,0.2849380672,0.1020080447,0.165165633,-0.0089086527,0.0223700646,0.0546234101,0.1785691381,-0.1242079139,0.0252420697,0.3226629198,-0.0290788095,-0.1172380075,0.0563454255,0.0549977571,0.1520994306,-0.0223549381,-0.4382071197,-0.1349186003,-0.1488203108,0.1116709709,-0.090152517,-0.2011874467,0.3662034869,0.1344199479,-0.0220552888,-0.0281651579,-0.2943358719,0.0167569909,0.0694208816,0.2837187648,-0.0141324084,0.5791600943,-0.0643859282,0.3804620802,0.4987338781,0.0455647632,0.2685537636,-0.1942719817,0.2392872423,-0.0748686567,-0.3963817358,0.0430059806,-0.065911904,-0.0903106108,0.102690056,-0.2037360817,-0.2834114134,-0.2474655956,-0.3123920262,-0.492813766,-0.2052918822,-0.0419776477,-0.3690470755,0.0671715587,0.1142082438,0.1570950001,-0.1366999,-0.4786142409,-0.0029995129,0.1720211953,0.2018553913,0.0454262644,0.002372029,0.112492986,-0.4285701513,0.4104145467,0.0891925171,0.4809066355,-0.1275958568,0.0004133504,0.106237635,-0.0396484695,0.5215788484,-0.347682178,0.1147932783,0.0772566944,-0.1183854863,-0.6588267088,0.0873671547,-0.1385788321,0.3216620386,0.4780519307,0.3114853501,-0.2311037034,-0.1851455867,0.1953789145,0.1498493105,-0.120704554,-0.2355740368,-0.2058528662,-0.1715175956,-0.4106161892,-0.1713676155,-0.0474788249,0.2916354537,0.1333781481,-0.0695790201,0.050281167,-0.1940112114,0.3146005273,0.0605621226,0.2779478729,0.1074624136,0.2539286017,0.306009382,0.31583938,0.111191839,0.4503264725,-0.014344017,0.0461661257,-0.0741442665,-0.1089513302,0.1159288138,0.1714711785,0.0884084702,-0.1365008056,-0.0308311414,-0.0622353926,-0.1701074094,-0.1018525735,0.3103106916,-0.07304921,-0.1800836921,-0.5892112255,0.4753007591,0.224488318,0.0644740239,-0.0521057881,0.0523687489,-0.2450018078,0.4438097179,0.1461969614,0.7782346606,0.2315705419,0.1015336439,0.0233296268,0.0174873788,0.5530241132,-0.0926305801,0.4083078206,-0.4835374951,-0.1268625706,-0.0691849291,-0.1424192041,0.0282572843,-0.0034954529,-0.2602635324,0.4872853756,-0.1277933121,-0.3951775134,-0.2352561951,0.4670633078,-0.3740275502,-0.0686999783,-0.052218888,0.1108666435,-0.2358676493,0.2975409329,0.0900800973,-0.1718620956,-0.047993388,-0.0603899099,-0.3981961906,0.210619539,-0.21927315,0.292272687,-0.2318606526,-0.1582142562,-0.0413427316,0.1540179849,0.1175688431,-0.2576344609,-0.2467580736,-0.1726966202,-0.0243189968,-0.2438659072,0.0316468552,0.1261455119,0.3964951634,-0.1396337003,-0.1136614457,0.0196182448,0.0792004988,-0.230617255,0.2163131386,-0.1996915936,0.0247094389,-0.1647003442,-0.5514370203,-0.0291414894,-0.1771263182,-0.1453780085,0.0900887325,-0.0960800722,-0.184115231,0.072912626,0.0623252206,-0.2890541255,-0.0627714917,0.5095923543,0.1822737157,0.1006661803,0.4814749658,0.0086050183,-0.2807370424,-0.2713316977,0.1044773683,-0.1335462481,-0.4408507347,0.1395952106,-0.0130844619,0.2914969921,0.2384903431,0.0768125579,0.2962935865,0.114862971,-0.0730927512,-0.4691395462,-0.2807099223,0.1527353972,-0.2083532512,0.3305924535,-0.1106204242,0.2401310205,0.0388927348,0.0792196989,-0.2634624839,-0.0190832336,-0.3192701936,0.1041337475,0.2587983608,-0.2227339,0.0652067885,0.1140528172,0.1187650934,0.0891325027,-0.1006643176,-0.1367288977,-0.0441542454,0.1490375996,0.0698425025,-0.2456952184,-0.3087036014,-0.114268966,-0.1854035556,-0.1146502942,-0.3026444614,0.29234761,-0.094547905,0.5159421563,-0.0256025419,-0.1894533187,-0.0149078304,0.2085004598,-0.311004132,0.3327851295,-0.1651076972,0.1961317062,0.0702773184,0.0952481627,-0.0319404192,-0.0785310417,-0.0573980846,0.0178687237,0.2745209336,-0.2878330946,0.0330613814,0.2350523323,0.1743424237,0.4963508546,-0.0423651524,-0.2141529173,0.0062150392,0.1651076525,-0.444652468,0.1129385233,0.5683695078,0.0282196701,-0.017522689,0.2885959744,0.201691553,-0.0269860215,-0.0795376301,0.0491537824,0.5171214342,-0.3017064929,0.1368759125,0.1469701678,0.0429963619,-0.1227517799,0.1644227654,0.2118979245,0.3436450064,0.5141329169,-0.1919100434,0.302053839,-0.2119131684,0.190798685,-0.304692179,-0.3163419366,-0.1183501855,0.2924265862,0.0155443205,0.1833102107,-0.0010102052,0.2147727311,-0.2041135728,0.1257146746,-0.2297851443,0.1350675076,-0.167255491,-0.0325336084,-0.1384450495,-0.0918379501,-0.3104115725,-0.0194232203,-0.009829985,0.1213115156,0.1699363738,0.2625784874,0.0348459594,-0.3329252899,0.0670570433,0.0607070513,0.1977004111,-0.3414990008,0.1506216675,0.04275316,0.1671671569,0.1254007816,0.5256478786,0.5326807499,0.381829083,-0.497646302,0.1000099182,0.078734085,-0.0744145662,0.0051493542,0.4151686132,0.2417095602,0.1278492659,0.3802852035,0.0687550455,-0.0718311965,0.329505533,0.0849409103,-0.0618023016,-0.0501097925,0.4606259763,-0.1114091501,0.0840820298,-0.2616661489,0.0943494961,-0.3821594119,-0.2704980671,0.3131412268,-0.2778038085,0.2021993995,0.0626498833,0.0418654345,-0.301295191,0.4437068403,0.323964864,-0.014916027,-0.3931873739,-0.0451750532,-0.6925978661,0.1487028152,-0.1757470667,-0.117818363,0.0530028194,0.3886677325,-0.2979029417,0.0005521679,-0.0622542314,0.5312165022,-0.1350671798,0.1048273221,-0.1133649871,-0.2026952654,-0.1523328722,-0.1748943776,-0.1197955459,-0.2434452027,0.0405483022,-0.316257894,0.0463291369,-0.1581527442,-0.0724611357,0.0085521806,0.6363357902,0.0301067885,0.3228030503,0.6062831879,-0.1531859338,-0.04710811,-0.1705669016,-0.3430226445,-0.2737978101,0.0975004137,0.3371979296,0.4055688679,-0.3082157373,0.3000153601,-0.2714779973,0.2924050093,-0.1004395559,-0.1348361522,-0.2340820581,0.0384989902,-0.1771728992,0.0795624107,-0.1113493741,0.1957411915,0.1067846417,0.2646092176,-0.2957310081,-0.1954199374,0.6806669831,-0.2425277829,-0.1901583374,0.0020970264,0.136228621,0.1533734351,-0.0369779095,-0.3961508572,0.0239590034,0.3804267645,-0.0542757213,-0.0332662016,0.1067489386,-0.0300816353,-0.2003993839,-0.0696348995,-0.2387509346,0.304954499,-0.2876032591,0.1136988476,-0.1027203202]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/252","title":"NonMatchingSplitsSizesError error when reading the IMDB dataset","comments":"I updated it, that was it, thanks!","body":"Hi!\r\n\r\nI am trying to load the `imdb` dataset with this line:\r\n\r\n`dataset = nlp.load_dataset('imdb', data_dir='\/A\/PATH', cache_dir='\/A\/PATH')`\r\n\r\nbut I am getting the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 517, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 363, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 421, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=5929447, num_examples=4537, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n```\r\n\r\nAm I overlooking something? Thanks!","comment_length":7,"text":"NonMatchingSplitsSizesError error when reading the IMDB dataset\nHi!\r\n\r\nI am trying to load the `imdb` dataset with this line:\r\n\r\n`dataset = nlp.load_dataset('imdb', data_dir='\/A\/PATH', cache_dir='\/A\/PATH')`\r\n\r\nbut I am getting the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 517, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 363, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 421, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=5929447, num_examples=4537, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n```\r\n\r\nAm I overlooking something? Thanks!\nI updated it, that was it, thanks!","embeddings":[-0.5058609843,-0.1288104951,0.0182252526,0.3497672975,0.1497054547,0.2042558789,0.0464905091,0.4214467704,0.2376457453,0.0744059384,-0.1168221682,0.1823245883,-0.2613818645,-0.101923354,-0.15735282,0.0979988351,-0.1106186137,0.2100696564,0.0744425282,0.0653068647,-0.4028107226,0.2740860283,-0.3829877377,-0.0700217038,-0.3913931847,-0.1332287341,0.1459162831,0.1515464187,0.191357404,-0.3842667341,0.3481659591,0.0109436065,0.1533981115,0.1625825912,-0.0001161558,0.010485352,0.6431316733,-0.2872529328,-0.2771607041,-0.2786664367,-0.2890476584,-0.2957006395,0.2998387218,-0.3556797504,-0.0921420455,-0.0069964286,0.3332283497,-0.4666482806,0.1093039885,0.1632326394,0.17784594,0.1451750994,-0.1186726689,0.1979513317,0.0227693487,-0.1154360548,-0.1067392081,0.0767500103,-0.0245760996,-0.0722568482,-0.1163303331,0.2309108973,-0.4465887249,0.1689139605,0.3901188076,0.175498262,0.3332034945,-0.2765008509,0.075174287,0.2043976337,0.5286696553,-0.1759230644,-0.2950552106,-0.3949430287,0.0285661221,-0.2426873296,0.164115116,0.3492821455,-0.2348972708,-0.0012560944,-0.3458954692,0.0242989697,-0.2227833569,0.1689743102,0.3957193494,0.1043043733,0.1111679003,0.0797282085,0.2857824266,-0.0581012443,0.4545836747,-0.0850620642,0.1459582746,0.1444033682,-0.2710443139,-0.0759877861,-0.2171054035,0.2746360898,0.0082770418,0.3213576674,0.1967979074,0.0516806841,0.1436003894,0.3109807074,0.2449324131,0.0170768797,0.1431279927,0.3343563378,0.0498044267,0.1574208736,-0.1261125505,-0.0362699814,0.0853060409,-0.0314564854,0.0814510062,0.1571653485,0.2686862648,-0.2099481523,-0.5460675359,-0.0010156493,-0.2332894206,-0.1651888639,0.0959894434,0.2754747272,-0.1536782682,0.3319395185,0.042252142,0.1424370408,-0.3492205441,-0.1216831952,-0.2368981838,0.172750473,-0.2033310533,0.0771291628,0.31588009,0.1044000685,0.4351963103,-0.1986221075,-0.1153535098,-0.0420896672,0.4212338924,-0.1903631389,-0.0745942444,0.2226096541,0.1970767379,0.1623467505,0.0870599002,-0.0507297404,-0.3188578784,0.1813304275,0.0028089394,-0.3774318993,0.0150839789,0.1790364236,-0.1909037083,-0.0862755403,-0.034753602,0.1134413704,0.1302699894,-0.4634093642,-0.2054534554,-0.1073764414,-0.2702164054,-0.1232669726,-0.0165222269,0.5770250559,0.0338287167,-0.0637336597,-0.1441337615,0.008088435,0.5104398727,0.3860457838,-0.0508839116,-0.1670527458,-0.3651423454,0.3389675915,0.2271953523,-0.4194648564,-0.4263260365,0.4837440848,0.006436381,0.2578959167,0.3255821764,0.4054284692,0.1204781532,-0.0325896367,0.0521423072,0.4240181148,-0.1317705214,-0.0142845968,-0.2874594629,-0.2950439453,0.6005905271,0.1494239867,0.0687649548,-0.0436367318,0.0187510271,0.4665596485,0.3268418014,-0.001648157,-0.0108888811,0.1649563313,-0.2478443682,0.3226459622,-0.1736309826,-0.1761586815,-0.5091295242,0.074178949,-0.1129090339,0.090829961,-0.1763046086,-0.1781263351,-0.0882440358,-0.3431604803,-0.2358764261,0.0028419192,0.1064237207,0.2504612803,-0.0604404472,0.139773801,-0.1747403294,0.5486708283,0.0023630376,0.082942225,-0.8483415246,0.3239140511,-0.1733106077,-0.2676151097,0.3004835248,0.2964940369,0.2826043665,-0.1530664861,-0.1141646951,0.3894184828,0.0910348073,0.3174045682,-0.1669467986,-0.0319519527,0.2927332819,-0.363134414,0.1423194408,0.2724507153,0.1341876537,-0.1109847426,-0.0392149165,0.2899010479,-0.3300355971,0.2139120996,0.0790896118,0.0629346669,0.1603439301,-0.1450172812,-0.0745699927,-0.4325863719,0.4534665942,0.1690415144,-0.071861811,0.0622173026,-0.3637724817,-0.1239177287,0.4598456919,0.1139727235,0.0168150458,0.0469442867,-0.0513484478,-0.1667591631,0.1503908932,0.1913670003,0.2858821452,0.3570718765,-0.1138631478,0.0858264789,-0.094161734,-0.1824414432,0.1085890383,0.1541863382,0.2679120898,0.3302231431,-0.1167620197,-0.0563683361,-0.4007021487,-0.2456368655,0.0161414798,0.3377129138,-0.2668766081,-0.1578804553,-0.4097839892,-0.4801176488,-0.056841068,-0.0843887404,-0.1401246637,-0.4054656029,-0.1167454571,0.5066664815,0.0902764648,0.2009690404,-0.386033684,-0.056571804,-0.0384274311,-0.145033434,-0.1585079283,-0.3335529864,-0.0711717978,-0.0457949005,0.3911580443,0.2297028601,0.3629732132,-0.4176098406,-0.0948796198,0.0653366596,0.0540711619,-0.1577545255,-0.0800911337,0.0756564736,0.0688458905,0.0790497363,0.1566694677,-0.3845457137,0.2842551768,-0.2741718292,-0.0760483891,0.1950515956,0.1936430484,0.0419287495,0.0353900678,-0.3608492613,-0.3879698217,-0.4397134185,0.0191146694,0.2849380672,0.1020080447,0.165165633,-0.0089086527,0.0223700646,0.0546234101,0.1785691381,-0.1242079139,0.0252420697,0.3226629198,-0.0290788095,-0.1172380075,0.0563454255,0.0549977571,0.1520994306,-0.0223549381,-0.4382071197,-0.1349186003,-0.1488203108,0.1116709709,-0.090152517,-0.2011874467,0.3662034869,0.1344199479,-0.0220552888,-0.0281651579,-0.2943358719,0.0167569909,0.0694208816,0.2837187648,-0.0141324084,0.5791600943,-0.0643859282,0.3804620802,0.4987338781,0.0455647632,0.2685537636,-0.1942719817,0.2392872423,-0.0748686567,-0.3963817358,0.0430059806,-0.065911904,-0.0903106108,0.102690056,-0.2037360817,-0.2834114134,-0.2474655956,-0.3123920262,-0.492813766,-0.2052918822,-0.0419776477,-0.3690470755,0.0671715587,0.1142082438,0.1570950001,-0.1366999,-0.4786142409,-0.0029995129,0.1720211953,0.2018553913,0.0454262644,0.002372029,0.112492986,-0.4285701513,0.4104145467,0.0891925171,0.4809066355,-0.1275958568,0.0004133504,0.106237635,-0.0396484695,0.5215788484,-0.347682178,0.1147932783,0.0772566944,-0.1183854863,-0.6588267088,0.0873671547,-0.1385788321,0.3216620386,0.4780519307,0.3114853501,-0.2311037034,-0.1851455867,0.1953789145,0.1498493105,-0.120704554,-0.2355740368,-0.2058528662,-0.1715175956,-0.4106161892,-0.1713676155,-0.0474788249,0.2916354537,0.1333781481,-0.0695790201,0.050281167,-0.1940112114,0.3146005273,0.0605621226,0.2779478729,0.1074624136,0.2539286017,0.306009382,0.31583938,0.111191839,0.4503264725,-0.014344017,0.0461661257,-0.0741442665,-0.1089513302,0.1159288138,0.1714711785,0.0884084702,-0.1365008056,-0.0308311414,-0.0622353926,-0.1701074094,-0.1018525735,0.3103106916,-0.07304921,-0.1800836921,-0.5892112255,0.4753007591,0.224488318,0.0644740239,-0.0521057881,0.0523687489,-0.2450018078,0.4438097179,0.1461969614,0.7782346606,0.2315705419,0.1015336439,0.0233296268,0.0174873788,0.5530241132,-0.0926305801,0.4083078206,-0.4835374951,-0.1268625706,-0.0691849291,-0.1424192041,0.0282572843,-0.0034954529,-0.2602635324,0.4872853756,-0.1277933121,-0.3951775134,-0.2352561951,0.4670633078,-0.3740275502,-0.0686999783,-0.052218888,0.1108666435,-0.2358676493,0.2975409329,0.0900800973,-0.1718620956,-0.047993388,-0.0603899099,-0.3981961906,0.210619539,-0.21927315,0.292272687,-0.2318606526,-0.1582142562,-0.0413427316,0.1540179849,0.1175688431,-0.2576344609,-0.2467580736,-0.1726966202,-0.0243189968,-0.2438659072,0.0316468552,0.1261455119,0.3964951634,-0.1396337003,-0.1136614457,0.0196182448,0.0792004988,-0.230617255,0.2163131386,-0.1996915936,0.0247094389,-0.1647003442,-0.5514370203,-0.0291414894,-0.1771263182,-0.1453780085,0.0900887325,-0.0960800722,-0.184115231,0.072912626,0.0623252206,-0.2890541255,-0.0627714917,0.5095923543,0.1822737157,0.1006661803,0.4814749658,0.0086050183,-0.2807370424,-0.2713316977,0.1044773683,-0.1335462481,-0.4408507347,0.1395952106,-0.0130844619,0.2914969921,0.2384903431,0.0768125579,0.2962935865,0.114862971,-0.0730927512,-0.4691395462,-0.2807099223,0.1527353972,-0.2083532512,0.3305924535,-0.1106204242,0.2401310205,0.0388927348,0.0792196989,-0.2634624839,-0.0190832336,-0.3192701936,0.1041337475,0.2587983608,-0.2227339,0.0652067885,0.1140528172,0.1187650934,0.0891325027,-0.1006643176,-0.1367288977,-0.0441542454,0.1490375996,0.0698425025,-0.2456952184,-0.3087036014,-0.114268966,-0.1854035556,-0.1146502942,-0.3026444614,0.29234761,-0.094547905,0.5159421563,-0.0256025419,-0.1894533187,-0.0149078304,0.2085004598,-0.311004132,0.3327851295,-0.1651076972,0.1961317062,0.0702773184,0.0952481627,-0.0319404192,-0.0785310417,-0.0573980846,0.0178687237,0.2745209336,-0.2878330946,0.0330613814,0.2350523323,0.1743424237,0.4963508546,-0.0423651524,-0.2141529173,0.0062150392,0.1651076525,-0.444652468,0.1129385233,0.5683695078,0.0282196701,-0.017522689,0.2885959744,0.201691553,-0.0269860215,-0.0795376301,0.0491537824,0.5171214342,-0.3017064929,0.1368759125,0.1469701678,0.0429963619,-0.1227517799,0.1644227654,0.2118979245,0.3436450064,0.5141329169,-0.1919100434,0.302053839,-0.2119131684,0.190798685,-0.304692179,-0.3163419366,-0.1183501855,0.2924265862,0.0155443205,0.1833102107,-0.0010102052,0.2147727311,-0.2041135728,0.1257146746,-0.2297851443,0.1350675076,-0.167255491,-0.0325336084,-0.1384450495,-0.0918379501,-0.3104115725,-0.0194232203,-0.009829985,0.1213115156,0.1699363738,0.2625784874,0.0348459594,-0.3329252899,0.0670570433,0.0607070513,0.1977004111,-0.3414990008,0.1506216675,0.04275316,0.1671671569,0.1254007816,0.5256478786,0.5326807499,0.381829083,-0.497646302,0.1000099182,0.078734085,-0.0744145662,0.0051493542,0.4151686132,0.2417095602,0.1278492659,0.3802852035,0.0687550455,-0.0718311965,0.329505533,0.0849409103,-0.0618023016,-0.0501097925,0.4606259763,-0.1114091501,0.0840820298,-0.2616661489,0.0943494961,-0.3821594119,-0.2704980671,0.3131412268,-0.2778038085,0.2021993995,0.0626498833,0.0418654345,-0.301295191,0.4437068403,0.323964864,-0.014916027,-0.3931873739,-0.0451750532,-0.6925978661,0.1487028152,-0.1757470667,-0.117818363,0.0530028194,0.3886677325,-0.2979029417,0.0005521679,-0.0622542314,0.5312165022,-0.1350671798,0.1048273221,-0.1133649871,-0.2026952654,-0.1523328722,-0.1748943776,-0.1197955459,-0.2434452027,0.0405483022,-0.316257894,0.0463291369,-0.1581527442,-0.0724611357,0.0085521806,0.6363357902,0.0301067885,0.3228030503,0.6062831879,-0.1531859338,-0.04710811,-0.1705669016,-0.3430226445,-0.2737978101,0.0975004137,0.3371979296,0.4055688679,-0.3082157373,0.3000153601,-0.2714779973,0.2924050093,-0.1004395559,-0.1348361522,-0.2340820581,0.0384989902,-0.1771728992,0.0795624107,-0.1113493741,0.1957411915,0.1067846417,0.2646092176,-0.2957310081,-0.1954199374,0.6806669831,-0.2425277829,-0.1901583374,0.0020970264,0.136228621,0.1533734351,-0.0369779095,-0.3961508572,0.0239590034,0.3804267645,-0.0542757213,-0.0332662016,0.1067489386,-0.0300816353,-0.2003993839,-0.0696348995,-0.2387509346,0.304954499,-0.2876032591,0.1136988476,-0.1027203202]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/252","title":"NonMatchingSplitsSizesError error when reading the IMDB dataset","comments":"Hello, I am facing the same problem... how do you clear the huggingface cache?","body":"Hi!\r\n\r\nI am trying to load the `imdb` dataset with this line:\r\n\r\n`dataset = nlp.load_dataset('imdb', data_dir='\/A\/PATH', cache_dir='\/A\/PATH')`\r\n\r\nbut I am getting the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 517, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 363, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 421, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=5929447, num_examples=4537, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n```\r\n\r\nAm I overlooking something? Thanks!","comment_length":14,"text":"NonMatchingSplitsSizesError error when reading the IMDB dataset\nHi!\r\n\r\nI am trying to load the `imdb` dataset with this line:\r\n\r\n`dataset = nlp.load_dataset('imdb', data_dir='\/A\/PATH', cache_dir='\/A\/PATH')`\r\n\r\nbut I am getting the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 517, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 363, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 421, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=5929447, num_examples=4537, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n```\r\n\r\nAm I overlooking something? Thanks!\nHello, I am facing the same problem... how do you clear the huggingface cache?","embeddings":[-0.5058609843,-0.1288104951,0.0182252526,0.3497672975,0.1497054547,0.2042558789,0.0464905091,0.4214467704,0.2376457453,0.0744059384,-0.1168221682,0.1823245883,-0.2613818645,-0.101923354,-0.15735282,0.0979988351,-0.1106186137,0.2100696564,0.0744425282,0.0653068647,-0.4028107226,0.2740860283,-0.3829877377,-0.0700217038,-0.3913931847,-0.1332287341,0.1459162831,0.1515464187,0.191357404,-0.3842667341,0.3481659591,0.0109436065,0.1533981115,0.1625825912,-0.0001161558,0.010485352,0.6431316733,-0.2872529328,-0.2771607041,-0.2786664367,-0.2890476584,-0.2957006395,0.2998387218,-0.3556797504,-0.0921420455,-0.0069964286,0.3332283497,-0.4666482806,0.1093039885,0.1632326394,0.17784594,0.1451750994,-0.1186726689,0.1979513317,0.0227693487,-0.1154360548,-0.1067392081,0.0767500103,-0.0245760996,-0.0722568482,-0.1163303331,0.2309108973,-0.4465887249,0.1689139605,0.3901188076,0.175498262,0.3332034945,-0.2765008509,0.075174287,0.2043976337,0.5286696553,-0.1759230644,-0.2950552106,-0.3949430287,0.0285661221,-0.2426873296,0.164115116,0.3492821455,-0.2348972708,-0.0012560944,-0.3458954692,0.0242989697,-0.2227833569,0.1689743102,0.3957193494,0.1043043733,0.1111679003,0.0797282085,0.2857824266,-0.0581012443,0.4545836747,-0.0850620642,0.1459582746,0.1444033682,-0.2710443139,-0.0759877861,-0.2171054035,0.2746360898,0.0082770418,0.3213576674,0.1967979074,0.0516806841,0.1436003894,0.3109807074,0.2449324131,0.0170768797,0.1431279927,0.3343563378,0.0498044267,0.1574208736,-0.1261125505,-0.0362699814,0.0853060409,-0.0314564854,0.0814510062,0.1571653485,0.2686862648,-0.2099481523,-0.5460675359,-0.0010156493,-0.2332894206,-0.1651888639,0.0959894434,0.2754747272,-0.1536782682,0.3319395185,0.042252142,0.1424370408,-0.3492205441,-0.1216831952,-0.2368981838,0.172750473,-0.2033310533,0.0771291628,0.31588009,0.1044000685,0.4351963103,-0.1986221075,-0.1153535098,-0.0420896672,0.4212338924,-0.1903631389,-0.0745942444,0.2226096541,0.1970767379,0.1623467505,0.0870599002,-0.0507297404,-0.3188578784,0.1813304275,0.0028089394,-0.3774318993,0.0150839789,0.1790364236,-0.1909037083,-0.0862755403,-0.034753602,0.1134413704,0.1302699894,-0.4634093642,-0.2054534554,-0.1073764414,-0.2702164054,-0.1232669726,-0.0165222269,0.5770250559,0.0338287167,-0.0637336597,-0.1441337615,0.008088435,0.5104398727,0.3860457838,-0.0508839116,-0.1670527458,-0.3651423454,0.3389675915,0.2271953523,-0.4194648564,-0.4263260365,0.4837440848,0.006436381,0.2578959167,0.3255821764,0.4054284692,0.1204781532,-0.0325896367,0.0521423072,0.4240181148,-0.1317705214,-0.0142845968,-0.2874594629,-0.2950439453,0.6005905271,0.1494239867,0.0687649548,-0.0436367318,0.0187510271,0.4665596485,0.3268418014,-0.001648157,-0.0108888811,0.1649563313,-0.2478443682,0.3226459622,-0.1736309826,-0.1761586815,-0.5091295242,0.074178949,-0.1129090339,0.090829961,-0.1763046086,-0.1781263351,-0.0882440358,-0.3431604803,-0.2358764261,0.0028419192,0.1064237207,0.2504612803,-0.0604404472,0.139773801,-0.1747403294,0.5486708283,0.0023630376,0.082942225,-0.8483415246,0.3239140511,-0.1733106077,-0.2676151097,0.3004835248,0.2964940369,0.2826043665,-0.1530664861,-0.1141646951,0.3894184828,0.0910348073,0.3174045682,-0.1669467986,-0.0319519527,0.2927332819,-0.363134414,0.1423194408,0.2724507153,0.1341876537,-0.1109847426,-0.0392149165,0.2899010479,-0.3300355971,0.2139120996,0.0790896118,0.0629346669,0.1603439301,-0.1450172812,-0.0745699927,-0.4325863719,0.4534665942,0.1690415144,-0.071861811,0.0622173026,-0.3637724817,-0.1239177287,0.4598456919,0.1139727235,0.0168150458,0.0469442867,-0.0513484478,-0.1667591631,0.1503908932,0.1913670003,0.2858821452,0.3570718765,-0.1138631478,0.0858264789,-0.094161734,-0.1824414432,0.1085890383,0.1541863382,0.2679120898,0.3302231431,-0.1167620197,-0.0563683361,-0.4007021487,-0.2456368655,0.0161414798,0.3377129138,-0.2668766081,-0.1578804553,-0.4097839892,-0.4801176488,-0.056841068,-0.0843887404,-0.1401246637,-0.4054656029,-0.1167454571,0.5066664815,0.0902764648,0.2009690404,-0.386033684,-0.056571804,-0.0384274311,-0.145033434,-0.1585079283,-0.3335529864,-0.0711717978,-0.0457949005,0.3911580443,0.2297028601,0.3629732132,-0.4176098406,-0.0948796198,0.0653366596,0.0540711619,-0.1577545255,-0.0800911337,0.0756564736,0.0688458905,0.0790497363,0.1566694677,-0.3845457137,0.2842551768,-0.2741718292,-0.0760483891,0.1950515956,0.1936430484,0.0419287495,0.0353900678,-0.3608492613,-0.3879698217,-0.4397134185,0.0191146694,0.2849380672,0.1020080447,0.165165633,-0.0089086527,0.0223700646,0.0546234101,0.1785691381,-0.1242079139,0.0252420697,0.3226629198,-0.0290788095,-0.1172380075,0.0563454255,0.0549977571,0.1520994306,-0.0223549381,-0.4382071197,-0.1349186003,-0.1488203108,0.1116709709,-0.090152517,-0.2011874467,0.3662034869,0.1344199479,-0.0220552888,-0.0281651579,-0.2943358719,0.0167569909,0.0694208816,0.2837187648,-0.0141324084,0.5791600943,-0.0643859282,0.3804620802,0.4987338781,0.0455647632,0.2685537636,-0.1942719817,0.2392872423,-0.0748686567,-0.3963817358,0.0430059806,-0.065911904,-0.0903106108,0.102690056,-0.2037360817,-0.2834114134,-0.2474655956,-0.3123920262,-0.492813766,-0.2052918822,-0.0419776477,-0.3690470755,0.0671715587,0.1142082438,0.1570950001,-0.1366999,-0.4786142409,-0.0029995129,0.1720211953,0.2018553913,0.0454262644,0.002372029,0.112492986,-0.4285701513,0.4104145467,0.0891925171,0.4809066355,-0.1275958568,0.0004133504,0.106237635,-0.0396484695,0.5215788484,-0.347682178,0.1147932783,0.0772566944,-0.1183854863,-0.6588267088,0.0873671547,-0.1385788321,0.3216620386,0.4780519307,0.3114853501,-0.2311037034,-0.1851455867,0.1953789145,0.1498493105,-0.120704554,-0.2355740368,-0.2058528662,-0.1715175956,-0.4106161892,-0.1713676155,-0.0474788249,0.2916354537,0.1333781481,-0.0695790201,0.050281167,-0.1940112114,0.3146005273,0.0605621226,0.2779478729,0.1074624136,0.2539286017,0.306009382,0.31583938,0.111191839,0.4503264725,-0.014344017,0.0461661257,-0.0741442665,-0.1089513302,0.1159288138,0.1714711785,0.0884084702,-0.1365008056,-0.0308311414,-0.0622353926,-0.1701074094,-0.1018525735,0.3103106916,-0.07304921,-0.1800836921,-0.5892112255,0.4753007591,0.224488318,0.0644740239,-0.0521057881,0.0523687489,-0.2450018078,0.4438097179,0.1461969614,0.7782346606,0.2315705419,0.1015336439,0.0233296268,0.0174873788,0.5530241132,-0.0926305801,0.4083078206,-0.4835374951,-0.1268625706,-0.0691849291,-0.1424192041,0.0282572843,-0.0034954529,-0.2602635324,0.4872853756,-0.1277933121,-0.3951775134,-0.2352561951,0.4670633078,-0.3740275502,-0.0686999783,-0.052218888,0.1108666435,-0.2358676493,0.2975409329,0.0900800973,-0.1718620956,-0.047993388,-0.0603899099,-0.3981961906,0.210619539,-0.21927315,0.292272687,-0.2318606526,-0.1582142562,-0.0413427316,0.1540179849,0.1175688431,-0.2576344609,-0.2467580736,-0.1726966202,-0.0243189968,-0.2438659072,0.0316468552,0.1261455119,0.3964951634,-0.1396337003,-0.1136614457,0.0196182448,0.0792004988,-0.230617255,0.2163131386,-0.1996915936,0.0247094389,-0.1647003442,-0.5514370203,-0.0291414894,-0.1771263182,-0.1453780085,0.0900887325,-0.0960800722,-0.184115231,0.072912626,0.0623252206,-0.2890541255,-0.0627714917,0.5095923543,0.1822737157,0.1006661803,0.4814749658,0.0086050183,-0.2807370424,-0.2713316977,0.1044773683,-0.1335462481,-0.4408507347,0.1395952106,-0.0130844619,0.2914969921,0.2384903431,0.0768125579,0.2962935865,0.114862971,-0.0730927512,-0.4691395462,-0.2807099223,0.1527353972,-0.2083532512,0.3305924535,-0.1106204242,0.2401310205,0.0388927348,0.0792196989,-0.2634624839,-0.0190832336,-0.3192701936,0.1041337475,0.2587983608,-0.2227339,0.0652067885,0.1140528172,0.1187650934,0.0891325027,-0.1006643176,-0.1367288977,-0.0441542454,0.1490375996,0.0698425025,-0.2456952184,-0.3087036014,-0.114268966,-0.1854035556,-0.1146502942,-0.3026444614,0.29234761,-0.094547905,0.5159421563,-0.0256025419,-0.1894533187,-0.0149078304,0.2085004598,-0.311004132,0.3327851295,-0.1651076972,0.1961317062,0.0702773184,0.0952481627,-0.0319404192,-0.0785310417,-0.0573980846,0.0178687237,0.2745209336,-0.2878330946,0.0330613814,0.2350523323,0.1743424237,0.4963508546,-0.0423651524,-0.2141529173,0.0062150392,0.1651076525,-0.444652468,0.1129385233,0.5683695078,0.0282196701,-0.017522689,0.2885959744,0.201691553,-0.0269860215,-0.0795376301,0.0491537824,0.5171214342,-0.3017064929,0.1368759125,0.1469701678,0.0429963619,-0.1227517799,0.1644227654,0.2118979245,0.3436450064,0.5141329169,-0.1919100434,0.302053839,-0.2119131684,0.190798685,-0.304692179,-0.3163419366,-0.1183501855,0.2924265862,0.0155443205,0.1833102107,-0.0010102052,0.2147727311,-0.2041135728,0.1257146746,-0.2297851443,0.1350675076,-0.167255491,-0.0325336084,-0.1384450495,-0.0918379501,-0.3104115725,-0.0194232203,-0.009829985,0.1213115156,0.1699363738,0.2625784874,0.0348459594,-0.3329252899,0.0670570433,0.0607070513,0.1977004111,-0.3414990008,0.1506216675,0.04275316,0.1671671569,0.1254007816,0.5256478786,0.5326807499,0.381829083,-0.497646302,0.1000099182,0.078734085,-0.0744145662,0.0051493542,0.4151686132,0.2417095602,0.1278492659,0.3802852035,0.0687550455,-0.0718311965,0.329505533,0.0849409103,-0.0618023016,-0.0501097925,0.4606259763,-0.1114091501,0.0840820298,-0.2616661489,0.0943494961,-0.3821594119,-0.2704980671,0.3131412268,-0.2778038085,0.2021993995,0.0626498833,0.0418654345,-0.301295191,0.4437068403,0.323964864,-0.014916027,-0.3931873739,-0.0451750532,-0.6925978661,0.1487028152,-0.1757470667,-0.117818363,0.0530028194,0.3886677325,-0.2979029417,0.0005521679,-0.0622542314,0.5312165022,-0.1350671798,0.1048273221,-0.1133649871,-0.2026952654,-0.1523328722,-0.1748943776,-0.1197955459,-0.2434452027,0.0405483022,-0.316257894,0.0463291369,-0.1581527442,-0.0724611357,0.0085521806,0.6363357902,0.0301067885,0.3228030503,0.6062831879,-0.1531859338,-0.04710811,-0.1705669016,-0.3430226445,-0.2737978101,0.0975004137,0.3371979296,0.4055688679,-0.3082157373,0.3000153601,-0.2714779973,0.2924050093,-0.1004395559,-0.1348361522,-0.2340820581,0.0384989902,-0.1771728992,0.0795624107,-0.1113493741,0.1957411915,0.1067846417,0.2646092176,-0.2957310081,-0.1954199374,0.6806669831,-0.2425277829,-0.1901583374,0.0020970264,0.136228621,0.1533734351,-0.0369779095,-0.3961508572,0.0239590034,0.3804267645,-0.0542757213,-0.0332662016,0.1067489386,-0.0300816353,-0.2003993839,-0.0696348995,-0.2387509346,0.304954499,-0.2876032591,0.1136988476,-0.1027203202]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/252","title":"NonMatchingSplitsSizesError error when reading the IMDB dataset","comments":"Hi ! The cache is at ~\/.cache\/huggingface\r\nYou can just delete this folder if needed :)","body":"Hi!\r\n\r\nI am trying to load the `imdb` dataset with this line:\r\n\r\n`dataset = nlp.load_dataset('imdb', data_dir='\/A\/PATH', cache_dir='\/A\/PATH')`\r\n\r\nbut I am getting the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 517, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 363, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 421, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=5929447, num_examples=4537, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n```\r\n\r\nAm I overlooking something? Thanks!","comment_length":16,"text":"NonMatchingSplitsSizesError error when reading the IMDB dataset\nHi!\r\n\r\nI am trying to load the `imdb` dataset with this line:\r\n\r\n`dataset = nlp.load_dataset('imdb', data_dir='\/A\/PATH', cache_dir='\/A\/PATH')`\r\n\r\nbut I am getting the following error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 517, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 363, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 421, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n File \"\/mounts\/Users\/cisintern\/antmarakis\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/utils\/info_utils.py\", line 70, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=5929447, num_examples=4537, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]\r\n```\r\n\r\nAm I overlooking something? Thanks!\nHi ! The cache is at ~\/.cache\/huggingface\r\nYou can just delete this folder if needed :)","embeddings":[-0.5058609843,-0.1288104951,0.0182252526,0.3497672975,0.1497054547,0.2042558789,0.0464905091,0.4214467704,0.2376457453,0.0744059384,-0.1168221682,0.1823245883,-0.2613818645,-0.101923354,-0.15735282,0.0979988351,-0.1106186137,0.2100696564,0.0744425282,0.0653068647,-0.4028107226,0.2740860283,-0.3829877377,-0.0700217038,-0.3913931847,-0.1332287341,0.1459162831,0.1515464187,0.191357404,-0.3842667341,0.3481659591,0.0109436065,0.1533981115,0.1625825912,-0.0001161558,0.010485352,0.6431316733,-0.2872529328,-0.2771607041,-0.2786664367,-0.2890476584,-0.2957006395,0.2998387218,-0.3556797504,-0.0921420455,-0.0069964286,0.3332283497,-0.4666482806,0.1093039885,0.1632326394,0.17784594,0.1451750994,-0.1186726689,0.1979513317,0.0227693487,-0.1154360548,-0.1067392081,0.0767500103,-0.0245760996,-0.0722568482,-0.1163303331,0.2309108973,-0.4465887249,0.1689139605,0.3901188076,0.175498262,0.3332034945,-0.2765008509,0.075174287,0.2043976337,0.5286696553,-0.1759230644,-0.2950552106,-0.3949430287,0.0285661221,-0.2426873296,0.164115116,0.3492821455,-0.2348972708,-0.0012560944,-0.3458954692,0.0242989697,-0.2227833569,0.1689743102,0.3957193494,0.1043043733,0.1111679003,0.0797282085,0.2857824266,-0.0581012443,0.4545836747,-0.0850620642,0.1459582746,0.1444033682,-0.2710443139,-0.0759877861,-0.2171054035,0.2746360898,0.0082770418,0.3213576674,0.1967979074,0.0516806841,0.1436003894,0.3109807074,0.2449324131,0.0170768797,0.1431279927,0.3343563378,0.0498044267,0.1574208736,-0.1261125505,-0.0362699814,0.0853060409,-0.0314564854,0.0814510062,0.1571653485,0.2686862648,-0.2099481523,-0.5460675359,-0.0010156493,-0.2332894206,-0.1651888639,0.0959894434,0.2754747272,-0.1536782682,0.3319395185,0.042252142,0.1424370408,-0.3492205441,-0.1216831952,-0.2368981838,0.172750473,-0.2033310533,0.0771291628,0.31588009,0.1044000685,0.4351963103,-0.1986221075,-0.1153535098,-0.0420896672,0.4212338924,-0.1903631389,-0.0745942444,0.2226096541,0.1970767379,0.1623467505,0.0870599002,-0.0507297404,-0.3188578784,0.1813304275,0.0028089394,-0.3774318993,0.0150839789,0.1790364236,-0.1909037083,-0.0862755403,-0.034753602,0.1134413704,0.1302699894,-0.4634093642,-0.2054534554,-0.1073764414,-0.2702164054,-0.1232669726,-0.0165222269,0.5770250559,0.0338287167,-0.0637336597,-0.1441337615,0.008088435,0.5104398727,0.3860457838,-0.0508839116,-0.1670527458,-0.3651423454,0.3389675915,0.2271953523,-0.4194648564,-0.4263260365,0.4837440848,0.006436381,0.2578959167,0.3255821764,0.4054284692,0.1204781532,-0.0325896367,0.0521423072,0.4240181148,-0.1317705214,-0.0142845968,-0.2874594629,-0.2950439453,0.6005905271,0.1494239867,0.0687649548,-0.0436367318,0.0187510271,0.4665596485,0.3268418014,-0.001648157,-0.0108888811,0.1649563313,-0.2478443682,0.3226459622,-0.1736309826,-0.1761586815,-0.5091295242,0.074178949,-0.1129090339,0.090829961,-0.1763046086,-0.1781263351,-0.0882440358,-0.3431604803,-0.2358764261,0.0028419192,0.1064237207,0.2504612803,-0.0604404472,0.139773801,-0.1747403294,0.5486708283,0.0023630376,0.082942225,-0.8483415246,0.3239140511,-0.1733106077,-0.2676151097,0.3004835248,0.2964940369,0.2826043665,-0.1530664861,-0.1141646951,0.3894184828,0.0910348073,0.3174045682,-0.1669467986,-0.0319519527,0.2927332819,-0.363134414,0.1423194408,0.2724507153,0.1341876537,-0.1109847426,-0.0392149165,0.2899010479,-0.3300355971,0.2139120996,0.0790896118,0.0629346669,0.1603439301,-0.1450172812,-0.0745699927,-0.4325863719,0.4534665942,0.1690415144,-0.071861811,0.0622173026,-0.3637724817,-0.1239177287,0.4598456919,0.1139727235,0.0168150458,0.0469442867,-0.0513484478,-0.1667591631,0.1503908932,0.1913670003,0.2858821452,0.3570718765,-0.1138631478,0.0858264789,-0.094161734,-0.1824414432,0.1085890383,0.1541863382,0.2679120898,0.3302231431,-0.1167620197,-0.0563683361,-0.4007021487,-0.2456368655,0.0161414798,0.3377129138,-0.2668766081,-0.1578804553,-0.4097839892,-0.4801176488,-0.056841068,-0.0843887404,-0.1401246637,-0.4054656029,-0.1167454571,0.5066664815,0.0902764648,0.2009690404,-0.386033684,-0.056571804,-0.0384274311,-0.145033434,-0.1585079283,-0.3335529864,-0.0711717978,-0.0457949005,0.3911580443,0.2297028601,0.3629732132,-0.4176098406,-0.0948796198,0.0653366596,0.0540711619,-0.1577545255,-0.0800911337,0.0756564736,0.0688458905,0.0790497363,0.1566694677,-0.3845457137,0.2842551768,-0.2741718292,-0.0760483891,0.1950515956,0.1936430484,0.0419287495,0.0353900678,-0.3608492613,-0.3879698217,-0.4397134185,0.0191146694,0.2849380672,0.1020080447,0.165165633,-0.0089086527,0.0223700646,0.0546234101,0.1785691381,-0.1242079139,0.0252420697,0.3226629198,-0.0290788095,-0.1172380075,0.0563454255,0.0549977571,0.1520994306,-0.0223549381,-0.4382071197,-0.1349186003,-0.1488203108,0.1116709709,-0.090152517,-0.2011874467,0.3662034869,0.1344199479,-0.0220552888,-0.0281651579,-0.2943358719,0.0167569909,0.0694208816,0.2837187648,-0.0141324084,0.5791600943,-0.0643859282,0.3804620802,0.4987338781,0.0455647632,0.2685537636,-0.1942719817,0.2392872423,-0.0748686567,-0.3963817358,0.0430059806,-0.065911904,-0.0903106108,0.102690056,-0.2037360817,-0.2834114134,-0.2474655956,-0.3123920262,-0.492813766,-0.2052918822,-0.0419776477,-0.3690470755,0.0671715587,0.1142082438,0.1570950001,-0.1366999,-0.4786142409,-0.0029995129,0.1720211953,0.2018553913,0.0454262644,0.002372029,0.112492986,-0.4285701513,0.4104145467,0.0891925171,0.4809066355,-0.1275958568,0.0004133504,0.106237635,-0.0396484695,0.5215788484,-0.347682178,0.1147932783,0.0772566944,-0.1183854863,-0.6588267088,0.0873671547,-0.1385788321,0.3216620386,0.4780519307,0.3114853501,-0.2311037034,-0.1851455867,0.1953789145,0.1498493105,-0.120704554,-0.2355740368,-0.2058528662,-0.1715175956,-0.4106161892,-0.1713676155,-0.0474788249,0.2916354537,0.1333781481,-0.0695790201,0.050281167,-0.1940112114,0.3146005273,0.0605621226,0.2779478729,0.1074624136,0.2539286017,0.306009382,0.31583938,0.111191839,0.4503264725,-0.014344017,0.0461661257,-0.0741442665,-0.1089513302,0.1159288138,0.1714711785,0.0884084702,-0.1365008056,-0.0308311414,-0.0622353926,-0.1701074094,-0.1018525735,0.3103106916,-0.07304921,-0.1800836921,-0.5892112255,0.4753007591,0.224488318,0.0644740239,-0.0521057881,0.0523687489,-0.2450018078,0.4438097179,0.1461969614,0.7782346606,0.2315705419,0.1015336439,0.0233296268,0.0174873788,0.5530241132,-0.0926305801,0.4083078206,-0.4835374951,-0.1268625706,-0.0691849291,-0.1424192041,0.0282572843,-0.0034954529,-0.2602635324,0.4872853756,-0.1277933121,-0.3951775134,-0.2352561951,0.4670633078,-0.3740275502,-0.0686999783,-0.052218888,0.1108666435,-0.2358676493,0.2975409329,0.0900800973,-0.1718620956,-0.047993388,-0.0603899099,-0.3981961906,0.210619539,-0.21927315,0.292272687,-0.2318606526,-0.1582142562,-0.0413427316,0.1540179849,0.1175688431,-0.2576344609,-0.2467580736,-0.1726966202,-0.0243189968,-0.2438659072,0.0316468552,0.1261455119,0.3964951634,-0.1396337003,-0.1136614457,0.0196182448,0.0792004988,-0.230617255,0.2163131386,-0.1996915936,0.0247094389,-0.1647003442,-0.5514370203,-0.0291414894,-0.1771263182,-0.1453780085,0.0900887325,-0.0960800722,-0.184115231,0.072912626,0.0623252206,-0.2890541255,-0.0627714917,0.5095923543,0.1822737157,0.1006661803,0.4814749658,0.0086050183,-0.2807370424,-0.2713316977,0.1044773683,-0.1335462481,-0.4408507347,0.1395952106,-0.0130844619,0.2914969921,0.2384903431,0.0768125579,0.2962935865,0.114862971,-0.0730927512,-0.4691395462,-0.2807099223,0.1527353972,-0.2083532512,0.3305924535,-0.1106204242,0.2401310205,0.0388927348,0.0792196989,-0.2634624839,-0.0190832336,-0.3192701936,0.1041337475,0.2587983608,-0.2227339,0.0652067885,0.1140528172,0.1187650934,0.0891325027,-0.1006643176,-0.1367288977,-0.0441542454,0.1490375996,0.0698425025,-0.2456952184,-0.3087036014,-0.114268966,-0.1854035556,-0.1146502942,-0.3026444614,0.29234761,-0.094547905,0.5159421563,-0.0256025419,-0.1894533187,-0.0149078304,0.2085004598,-0.311004132,0.3327851295,-0.1651076972,0.1961317062,0.0702773184,0.0952481627,-0.0319404192,-0.0785310417,-0.0573980846,0.0178687237,0.2745209336,-0.2878330946,0.0330613814,0.2350523323,0.1743424237,0.4963508546,-0.0423651524,-0.2141529173,0.0062150392,0.1651076525,-0.444652468,0.1129385233,0.5683695078,0.0282196701,-0.017522689,0.2885959744,0.201691553,-0.0269860215,-0.0795376301,0.0491537824,0.5171214342,-0.3017064929,0.1368759125,0.1469701678,0.0429963619,-0.1227517799,0.1644227654,0.2118979245,0.3436450064,0.5141329169,-0.1919100434,0.302053839,-0.2119131684,0.190798685,-0.304692179,-0.3163419366,-0.1183501855,0.2924265862,0.0155443205,0.1833102107,-0.0010102052,0.2147727311,-0.2041135728,0.1257146746,-0.2297851443,0.1350675076,-0.167255491,-0.0325336084,-0.1384450495,-0.0918379501,-0.3104115725,-0.0194232203,-0.009829985,0.1213115156,0.1699363738,0.2625784874,0.0348459594,-0.3329252899,0.0670570433,0.0607070513,0.1977004111,-0.3414990008,0.1506216675,0.04275316,0.1671671569,0.1254007816,0.5256478786,0.5326807499,0.381829083,-0.497646302,0.1000099182,0.078734085,-0.0744145662,0.0051493542,0.4151686132,0.2417095602,0.1278492659,0.3802852035,0.0687550455,-0.0718311965,0.329505533,0.0849409103,-0.0618023016,-0.0501097925,0.4606259763,-0.1114091501,0.0840820298,-0.2616661489,0.0943494961,-0.3821594119,-0.2704980671,0.3131412268,-0.2778038085,0.2021993995,0.0626498833,0.0418654345,-0.301295191,0.4437068403,0.323964864,-0.014916027,-0.3931873739,-0.0451750532,-0.6925978661,0.1487028152,-0.1757470667,-0.117818363,0.0530028194,0.3886677325,-0.2979029417,0.0005521679,-0.0622542314,0.5312165022,-0.1350671798,0.1048273221,-0.1133649871,-0.2026952654,-0.1523328722,-0.1748943776,-0.1197955459,-0.2434452027,0.0405483022,-0.316257894,0.0463291369,-0.1581527442,-0.0724611357,0.0085521806,0.6363357902,0.0301067885,0.3228030503,0.6062831879,-0.1531859338,-0.04710811,-0.1705669016,-0.3430226445,-0.2737978101,0.0975004137,0.3371979296,0.4055688679,-0.3082157373,0.3000153601,-0.2714779973,0.2924050093,-0.1004395559,-0.1348361522,-0.2340820581,0.0384989902,-0.1771728992,0.0795624107,-0.1113493741,0.1957411915,0.1067846417,0.2646092176,-0.2957310081,-0.1954199374,0.6806669831,-0.2425277829,-0.1901583374,0.0020970264,0.136228621,0.1533734351,-0.0369779095,-0.3961508572,0.0239590034,0.3804267645,-0.0542757213,-0.0332662016,0.1067489386,-0.0300816353,-0.2003993839,-0.0696348995,-0.2387509346,0.304954499,-0.2876032591,0.1136988476,-0.1027203202]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/249","title":"[Dataset created] some critical small issues when I was creating a dataset","comments":"Thanks for noticing all these :) They should be easy to fix indeed","body":"Hi, I successfully created a dataset and has made a pr #248.\r\nBut I have encountered several problems when I was creating it, and those should be easy to fix.\r\n\r\n1. Not found dataset_info.json\r\nshould be fixed by #241 , eager to wait it be merged.\r\n\r\n2. Forced to install `apach_beam`\r\nIf we should install it, then it might be better to include it in the pakcage dependency or specified in `CONTRIBUTING.md`\r\n```\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 10, in <module>\r\n from nlp.commands.run_beam import RunBeamCommand\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/commands\/run_beam.py\", line 6, in <module>\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n```\r\n\r\n3. `cached_dir` is `None`\r\n```\r\nFile \"\/home\/yisiang\/nlp\/src\/nlp\/datasets\/bookscorpus\/aea0bd5142d26df645a8fce23d6110bb95ecb81772bb2a1f29012e329191962c\/bookscorpus.py\", line 88, in _split_generators\r\n downloaded_path_or_paths = dl_manager.download_custom(_GDRIVE_FILE_ID, download_file_from_google_drive)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 128, in download_custom\r\n downloaded_path_or_paths = map_nested(url_to_downloaded_path, url_or_urls)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 126, in url_to_downloaded_path\r\n return os.path.join(self._download_config.cache_dir, hash_url_to_filename(url))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/nlppr\/lib\/python3.7\/posixpath.py\", line 80, in join\r\n a = os.fspath(a)\r\n```\r\nThis is because this line\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/src\/nlp\/commands\/test.py#L30-L32\r\nAnd I add `--cache_dir=\"....\"` to `python nlp-cli test datasets\/<your-dataset-folder> --save_infos --all_configs` in the doc, finally I could pass this error.\r\nBut it seems to ignore my arg and use `\/home\/yisiang\/.cache\/huggingface\/datasets\/bookscorpus\/plain_text\/1.0.0` as cahe_dir\r\n\r\n4. There is no `pytest`\r\nSo maybe in the doc we should specify a step to install pytest\r\n\r\n5. Not enough capacity in my `\/tmp`\r\nWhen run test for dummy data, I don't know why it ask me for 5.6g to download something, \r\n```\r\ndef download_and_prepare\r\n...\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n> utils.size_str(self.info.dataset_size or 0),\r\n )\r\n )\r\nE OSError: Not enough disk space. Needed: 5.62 GiB (download: 1.10 GiB, generated: 4.52 GiB)\r\n```\r\nI add a `processed_temp_dir=\"some\/dir\"; raw_temp_dir=\"another\/dir\"` to 71, and the test passed\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/a67a6c422dece904b65d18af65f0e024e839dbe8\/tests\/test_dataset_common.py#L70-L72\r\n\r\nI suggest we can create tmp dir under the `\/home\/user\/tmp` but not `\/tmp`, because take our lab server for example, everyone use `\/tmp` thus it has not much capacity. Or at least we can improve error message, so the user know is what directory has no space and how many has it lefted. Or we could do both.\r\n\r\n6. name of datasets\r\nI was surprised by the dataset name `books_corpus`, and didn't know it is from `class BooksCorpus(nlp.GeneratorBasedBuilder)` . I change it to `Bookscorpus` afterwards. I think this point shold be also on the doc.\r\n\r\n7. More thorough doc to how to create `dataset.py`\r\nI believe there will be.\r\n\r\n**Feel free to close this issue** if you think these are solved.","comment_length":13,"text":"[Dataset created] some critical small issues when I was creating a dataset\nHi, I successfully created a dataset and has made a pr #248.\r\nBut I have encountered several problems when I was creating it, and those should be easy to fix.\r\n\r\n1. Not found dataset_info.json\r\nshould be fixed by #241 , eager to wait it be merged.\r\n\r\n2. Forced to install `apach_beam`\r\nIf we should install it, then it might be better to include it in the pakcage dependency or specified in `CONTRIBUTING.md`\r\n```\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 10, in <module>\r\n from nlp.commands.run_beam import RunBeamCommand\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/commands\/run_beam.py\", line 6, in <module>\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n```\r\n\r\n3. `cached_dir` is `None`\r\n```\r\nFile \"\/home\/yisiang\/nlp\/src\/nlp\/datasets\/bookscorpus\/aea0bd5142d26df645a8fce23d6110bb95ecb81772bb2a1f29012e329191962c\/bookscorpus.py\", line 88, in _split_generators\r\n downloaded_path_or_paths = dl_manager.download_custom(_GDRIVE_FILE_ID, download_file_from_google_drive)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 128, in download_custom\r\n downloaded_path_or_paths = map_nested(url_to_downloaded_path, url_or_urls)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 126, in url_to_downloaded_path\r\n return os.path.join(self._download_config.cache_dir, hash_url_to_filename(url))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/nlppr\/lib\/python3.7\/posixpath.py\", line 80, in join\r\n a = os.fspath(a)\r\n```\r\nThis is because this line\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/src\/nlp\/commands\/test.py#L30-L32\r\nAnd I add `--cache_dir=\"....\"` to `python nlp-cli test datasets\/<your-dataset-folder> --save_infos --all_configs` in the doc, finally I could pass this error.\r\nBut it seems to ignore my arg and use `\/home\/yisiang\/.cache\/huggingface\/datasets\/bookscorpus\/plain_text\/1.0.0` as cahe_dir\r\n\r\n4. There is no `pytest`\r\nSo maybe in the doc we should specify a step to install pytest\r\n\r\n5. Not enough capacity in my `\/tmp`\r\nWhen run test for dummy data, I don't know why it ask me for 5.6g to download something, \r\n```\r\ndef download_and_prepare\r\n...\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n> utils.size_str(self.info.dataset_size or 0),\r\n )\r\n )\r\nE OSError: Not enough disk space. Needed: 5.62 GiB (download: 1.10 GiB, generated: 4.52 GiB)\r\n```\r\nI add a `processed_temp_dir=\"some\/dir\"; raw_temp_dir=\"another\/dir\"` to 71, and the test passed\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/a67a6c422dece904b65d18af65f0e024e839dbe8\/tests\/test_dataset_common.py#L70-L72\r\n\r\nI suggest we can create tmp dir under the `\/home\/user\/tmp` but not `\/tmp`, because take our lab server for example, everyone use `\/tmp` thus it has not much capacity. Or at least we can improve error message, so the user know is what directory has no space and how many has it lefted. Or we could do both.\r\n\r\n6. name of datasets\r\nI was surprised by the dataset name `books_corpus`, and didn't know it is from `class BooksCorpus(nlp.GeneratorBasedBuilder)` . I change it to `Bookscorpus` afterwards. I think this point shold be also on the doc.\r\n\r\n7. More thorough doc to how to create `dataset.py`\r\nI believe there will be.\r\n\r\n**Feel free to close this issue** if you think these are solved.\nThanks for noticing all these :) They should be easy to fix indeed","embeddings":[-0.2165798396,-0.0164848957,-0.1206973046,0.3510122001,0.2939466834,-0.0238549132,0.086690411,0.4066426158,-0.1998541057,0.0216472019,0.2210519165,0.2012127191,-0.2307064533,-0.0083013931,0.1113757342,-0.1998809725,0.2577252984,0.309070766,0.21963422,-0.1952063143,-0.4448091984,0.1668979228,-0.0202732198,0.0794764906,-0.1854971647,-0.3484351039,0.0012606009,0.3041497171,-0.27669245,-0.3347336054,0.0429069027,-0.0216965955,0.061426077,0.1791488975,-0.0001148037,-0.1664888561,0.3816398084,-0.1286742687,-0.4447022676,-0.0524816103,-0.2777306139,-0.5376020074,0.1146237329,-0.3695807159,0.1170074046,-0.4766104817,-0.2099819928,-0.2251408994,0.5323174,0.3592292964,0.2335041165,0.2630361617,0.2306550294,-0.0752441138,0.3409655988,0.0986781269,-0.2740069926,-0.0909317881,0.0840169713,0.191456601,0.0731679052,0.1320601255,0.1426730156,-0.1103613377,0.4137816429,-0.1641572863,-0.0283232555,-0.5333042741,-0.0329061784,-0.0450610407,0.4773083031,-0.3531982601,-0.3011023998,-0.3424141109,0.1612422019,-0.1395581961,0.238892436,0.4321993887,-0.1544614136,-0.1136258841,-0.1933754086,-0.1405055672,-0.2878140509,0.1813217252,0.4124358594,0.0853471979,0.1599856317,0.0504578277,-0.0027547576,-0.1099818051,0.1672928035,-0.3342158794,-0.0749942809,0.3552259803,-0.1991508752,-0.1965242326,-0.0443683229,0.0313737169,0.0599546842,0.2317240387,-0.2460917979,0.0044569359,-0.0118472353,0.0019146631,0.1471191496,0.1913504303,0.1334288269,0.0652819648,0.1037480012,-0.039223291,0.2899758518,0.0508449227,-0.0023807478,0.0406709425,-0.1865070015,0.0519573689,-0.0040365495,-0.0670203939,-0.0239564143,0.1387755275,-0.3734666407,-0.2500493526,0.042333439,0.2187863439,0.1310413629,0.2003273368,-0.0677166432,0.2659163773,-0.2425889075,-0.2212205827,-0.0890087709,0.1785895973,-0.3464155793,-0.008724126,0.4041240513,0.3678691387,0.3285503387,-0.0558720306,-0.1568581462,0.0804305598,0.3698014021,-0.1725635231,-0.164951548,0.3664527833,0.1163076907,0.3178191185,-0.2282937914,-0.0065453667,-0.1832158566,0.0733555034,0.0013044693,-0.2434278727,-0.0742869303,0.2257186919,-0.3786150217,0.3559951782,0.0095038917,0.085091956,0.1143025979,-0.2616052032,-0.1059470549,0.000416379,-0.0401857942,-0.3765689731,0.2222083956,0.3557888269,-0.2729791701,-0.117356427,-0.2392379493,-0.4322434068,-0.172145471,0.0143524855,-0.1099702269,0.4447313845,-0.1886315346,0.0590488017,0.5762619376,-0.1736322492,-0.3862493932,-0.0011030184,0.0378160216,-0.161188066,0.1443196833,0.0966827124,0.0397116542,-0.2814246118,-0.3508945405,0.3981647193,-0.0608353876,0.0863966048,-0.3207007647,-0.0957875624,0.1719620228,0.2355937362,-0.009895077,0.0999968201,0.2488787919,0.2356128842,-0.0183623042,-0.1480812132,0.2485683262,0.3226508498,0.3735436797,-0.3224754632,0.0843802467,0.0609972142,-0.4909184277,0.0059003527,-0.4745084047,0.0989301354,-0.1233595684,0.004432248,-0.2629866004,-0.0900884494,-0.0647096708,-0.3086581826,0.1512106508,-0.078992255,0.1919051856,-0.1069257483,-0.3420167267,0.5311259031,-0.1893766224,0.068848528,-0.5045070648,0.218655467,-0.1654846072,-0.121896334,0.3162867129,0.591373682,0.1859377474,-0.1297962517,-0.1571085155,0.376141876,-0.0072881156,0.1002138332,0.1314674765,0.108823508,0.3348307014,-0.1476595104,0.3715452254,-0.2893131375,0.108980827,-0.2142601311,-0.3619832397,0.2575721145,-0.0945694,0.1903479993,0.2109564394,-0.0040955762,0.2478683144,-0.0905265659,0.0391285047,-0.1222396418,0.2752211988,0.2415908575,0.3407287896,0.2105956227,0.1973709315,0.2309572399,0.2959473133,-0.0584616847,-0.1581475288,0.2084072828,-0.0477813818,-0.1669331044,0.2196256369,0.4565879405,0.5612210035,0.2158060372,-0.3815115392,0.1613033712,0.0972903296,-0.1790976673,0.1301805824,0.2057107389,0.267224133,0.0388933457,0.2177975923,0.0833001956,0.0917641595,-0.012623868,0.0895320624,0.3389057219,-0.1725993156,-0.156357497,-0.031914182,-0.3260663152,-0.1701599509,0.3311553299,-0.2817266881,-0.3784973323,-0.0004050345,-0.1697893739,-0.0229077768,0.2918809056,0.2162147611,0.2618064284,-0.1883648634,0.0976374298,-0.3757290244,-0.3888498545,-0.1720531881,0.1114448011,0.1218660325,0.3295152783,0.5941365957,-0.0154186925,0.12340004,-0.1669679731,-0.1062984541,0.0704004169,-0.0434974879,0.1907853782,0.0079046916,0.2060802579,-0.0516079739,-0.2771467268,0.1803607047,-0.0872100592,-0.2609621286,-0.0134241097,-0.147666499,-0.0718898848,-0.0679896995,-0.5194193125,-0.4250129759,-0.2148083299,0.1327863038,0.3100067079,0.1828389317,-0.0841341466,0.141230613,-0.0181512292,-0.0019709226,-0.0615515746,0.0066660936,0.0194676388,0.2024261504,-0.2755369544,-0.4629991949,0.0172609836,-0.2323603481,0.1478987336,0.1514241397,-0.3561945558,-0.0858034939,-0.1002788767,0.2035038769,0.0486067049,0.2767562568,0.3269117773,0.1419498324,0.0358995572,-0.1744469106,-0.3353277445,0.1088822708,-0.1908497959,0.59681952,-0.0053010136,0.1931440681,-0.3918654323,0.8372858167,0.4679619968,-0.1322748065,0.2019901276,0.0645503551,0.2530503273,-0.0333257355,-0.2155572325,0.5519694686,0.1865862757,0.1339824796,-0.0086955437,0.1465135515,-0.2078359425,-0.2886020243,0.1719544679,-0.7883492708,-0.1010141745,-0.1152329817,-0.132088989,0.2729951739,0.0884637088,-0.0757809579,-0.0928049907,-0.1519215107,0.0211826358,0.0190127604,0.1256043166,-0.0408476442,-0.2778846622,-0.1165947616,-0.1661058217,0.1609886736,-0.1069056168,0.2166684717,-0.0879853219,0.0053945784,0.0756809562,-0.0121861566,0.6508090496,-0.4458170235,-0.2502537668,0.1428911984,0.4055549204,-0.3814205229,0.1012092978,-0.1455695629,0.158567667,0.1066678092,0.2751333416,-0.4258680046,-0.0385760106,0.1401399225,0.3697842062,-0.160544157,-0.1771896034,-0.0975282714,-0.2522131801,-0.2250371128,-0.3424991667,-0.6088339686,0.1231244728,0.0420012511,0.2845884562,0.0299565531,-0.0380930267,0.0616338961,0.1306768805,0.4030514956,-0.1472469568,0.2574329376,0.0296916068,0.1881632656,0.2162958086,0.3045300543,-0.0867219493,-0.0720852688,0.0461534262,-0.11099042,0.3767855167,0.2177415192,-0.0714936703,-0.2043244392,-0.0884054229,0.0822213739,0.0422078259,0.0134127671,0.1663259268,-0.2151949853,-0.2809138596,-0.3256173134,0.4102795422,0.1630631685,-0.0794472471,0.1998076737,-0.1293647587,-0.0149041219,0.5705702305,0.1807837635,1.1641917229,0.0162149724,-0.1489294767,0.3272439241,-0.3185850382,0.7361056805,-0.9293342829,0.0379289575,-0.4258750677,0.2147131562,0.1235820502,-0.228075847,0.2144611925,0.0830759332,-0.1261196136,0.1880285591,0.1353084594,0.1012107059,-0.114755854,0.6493266225,-0.0398233831,-0.4592416883,-0.5051500201,0.1028768122,-0.2891237438,0.0760233328,-0.1944546551,-0.1979286373,-0.3322232366,-0.2097279429,-0.1891156435,0.140247792,-0.1220817715,0.1048285365,0.082154274,-0.5340114832,0.1189398244,0.1871011704,0.4441755414,0.0693396553,-0.2172711492,0.3504209816,0.0356042162,0.0122375796,0.1434174031,0.156142965,0.194249928,-0.223055765,0.0185003784,0.1097248495,-0.3166891038,0.1906911135,-0.2579829991,-0.2314601541,0.2656127214,-0.1680596322,-0.0141959451,0.1864646375,-0.179586336,-0.1864559352,0.1789260209,0.052234605,0.0230730865,0.2207249999,-0.0080895051,-0.0254745204,-0.2591033578,0.1850200295,0.2551265955,-0.4933585525,0.2920069993,0.0143672777,-0.1413042992,-0.3368189633,0.3393170238,-0.0217135046,-0.549806416,-0.0471096225,0.1549621969,-0.1732402891,-0.1225043088,-0.0646256283,0.1288732588,0.1655765772,0.1828327626,-0.3124121428,-0.3814761341,0.2685787678,-0.1478324682,0.1492538601,-0.3125891089,-0.0533510149,0.1307689697,0.2059382647,-0.3075966239,-0.2742983401,-0.0850652829,0.1111879051,0.0690172836,-0.0001548357,-0.1310928911,-0.3503128886,0.10763859,-0.1060899794,-0.3347283602,-0.1759790033,0.0289696325,0.1521108449,0.0870237499,-0.1402989328,-0.1382886469,-0.0890989006,-0.0205531791,-0.1737530529,-0.0808127373,-0.2732879519,-0.0885648057,0.0174123812,0.1678901166,0.1153389364,-0.2374764383,-0.026261678,-0.0853664726,0.2476085275,0.0834323093,-0.106934458,0.1269097924,-0.0297013968,-0.2137421668,-0.1161093339,0.0099907881,0.1641687602,0.1888871193,-0.1789885163,0.0328501016,0.0628117621,0.3357524276,0.2211795747,-0.2855054438,-0.0039075133,0.258782804,0.1599668264,-0.3051172197,0.1355106235,0.5688182712,0.116217427,-0.192642644,0.1513213515,0.1291161329,-0.0586773381,-0.0556202792,0.1372754425,0.5175210834,-0.1722085476,0.3115231097,0.2180622518,0.0318760425,0.0112811523,0.035239812,0.1726840585,0.0470054373,0.6041743159,-0.0963322371,0.5378553867,0.276620388,0.1715888083,-0.0179595184,-0.2979881167,0.0516243875,0.0835087001,0.0514033437,-0.0647090077,0.0556912832,0.475015074,0.0523283966,-0.2881971598,-0.1949679106,0.0889088064,-0.0953513607,0.4037155509,-0.1788719743,-0.2117675096,-0.0986685753,-0.1745070368,0.0179083161,-0.1983197927,0.364806056,0.185503602,-0.1240945458,-0.1778636426,0.4243930876,0.0772921667,0.2227859348,-0.0287255794,0.2480990291,0.0141337169,-0.0970877856,-0.1475265324,0.2929859459,0.3658083975,0.4779532254,-0.1898424327,0.0621615611,0.2308991849,-0.0886326879,0.0779972747,0.4141319394,0.3960823119,0.1710058153,0.1178681254,0.1362997293,-0.1313177794,-0.0407680348,0.2802560627,0.0961469188,-0.0316606537,0.1877766252,-0.2365777344,-0.4242991805,-0.0851354077,-0.0414729193,-0.3212204576,-0.1003718749,0.048656065,0.2468833178,0.0312168244,-0.288523823,0.0620972738,-0.1806997508,0.5545122623,0.1600003093,0.093418695,-0.4730508029,-0.339553833,-0.6459328532,-0.1185649484,-0.2552911043,0.0187612884,-0.0552571602,0.0170525331,0.0001157257,0.296777606,0.1137385443,0.5873936415,-0.045605015,-0.0323732086,-0.5817576051,-0.0594591089,0.0666519701,0.0310340524,-0.1627264768,-0.0509418994,0.1158232912,-0.099445127,0.0204731133,-0.0383134261,-0.3848719895,0.3540809453,0.2101806998,0.3670881987,0.2355558872,0.3461514115,-0.0735239387,-0.0712920278,-0.0996794179,-0.2440582514,0.1177210361,0.27832973,0.0427765362,0.1906647831,-0.0408141315,0.1475720257,-0.4616036415,-0.0035805705,-0.1504666656,0.295078665,-0.1972342879,0.0702444836,-0.0566571318,0.2186337709,0.1541198194,-0.1112893,-0.1131286323,0.291503489,-0.4034203291,-0.1097390428,0.4648357332,-0.1177231297,-0.0457417481,0.2879976034,0.2958430648,-0.1753144413,-0.0571581312,-0.5936287045,-0.0158088785,0.3534444869,-0.2711624801,-0.0148731638,0.0799034014,-0.1544134021,-0.1725821346,0.0343465582,0.1275982708,-0.1552858055,-0.1353527606,0.1812076569,0.0627781302]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/249","title":"[Dataset created] some critical small issues when I was creating a dataset","comments":"Alright I think I fixed all the problems you mentioned. Thanks again, that will be useful for many people.\r\nThere is still more work needed for point 7. but we plan to have some nice docs soon.","body":"Hi, I successfully created a dataset and has made a pr #248.\r\nBut I have encountered several problems when I was creating it, and those should be easy to fix.\r\n\r\n1. Not found dataset_info.json\r\nshould be fixed by #241 , eager to wait it be merged.\r\n\r\n2. Forced to install `apach_beam`\r\nIf we should install it, then it might be better to include it in the pakcage dependency or specified in `CONTRIBUTING.md`\r\n```\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 10, in <module>\r\n from nlp.commands.run_beam import RunBeamCommand\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/commands\/run_beam.py\", line 6, in <module>\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n```\r\n\r\n3. `cached_dir` is `None`\r\n```\r\nFile \"\/home\/yisiang\/nlp\/src\/nlp\/datasets\/bookscorpus\/aea0bd5142d26df645a8fce23d6110bb95ecb81772bb2a1f29012e329191962c\/bookscorpus.py\", line 88, in _split_generators\r\n downloaded_path_or_paths = dl_manager.download_custom(_GDRIVE_FILE_ID, download_file_from_google_drive)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 128, in download_custom\r\n downloaded_path_or_paths = map_nested(url_to_downloaded_path, url_or_urls)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 126, in url_to_downloaded_path\r\n return os.path.join(self._download_config.cache_dir, hash_url_to_filename(url))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/nlppr\/lib\/python3.7\/posixpath.py\", line 80, in join\r\n a = os.fspath(a)\r\n```\r\nThis is because this line\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/src\/nlp\/commands\/test.py#L30-L32\r\nAnd I add `--cache_dir=\"....\"` to `python nlp-cli test datasets\/<your-dataset-folder> --save_infos --all_configs` in the doc, finally I could pass this error.\r\nBut it seems to ignore my arg and use `\/home\/yisiang\/.cache\/huggingface\/datasets\/bookscorpus\/plain_text\/1.0.0` as cahe_dir\r\n\r\n4. There is no `pytest`\r\nSo maybe in the doc we should specify a step to install pytest\r\n\r\n5. Not enough capacity in my `\/tmp`\r\nWhen run test for dummy data, I don't know why it ask me for 5.6g to download something, \r\n```\r\ndef download_and_prepare\r\n...\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n> utils.size_str(self.info.dataset_size or 0),\r\n )\r\n )\r\nE OSError: Not enough disk space. Needed: 5.62 GiB (download: 1.10 GiB, generated: 4.52 GiB)\r\n```\r\nI add a `processed_temp_dir=\"some\/dir\"; raw_temp_dir=\"another\/dir\"` to 71, and the test passed\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/a67a6c422dece904b65d18af65f0e024e839dbe8\/tests\/test_dataset_common.py#L70-L72\r\n\r\nI suggest we can create tmp dir under the `\/home\/user\/tmp` but not `\/tmp`, because take our lab server for example, everyone use `\/tmp` thus it has not much capacity. Or at least we can improve error message, so the user know is what directory has no space and how many has it lefted. Or we could do both.\r\n\r\n6. name of datasets\r\nI was surprised by the dataset name `books_corpus`, and didn't know it is from `class BooksCorpus(nlp.GeneratorBasedBuilder)` . I change it to `Bookscorpus` afterwards. I think this point shold be also on the doc.\r\n\r\n7. More thorough doc to how to create `dataset.py`\r\nI believe there will be.\r\n\r\n**Feel free to close this issue** if you think these are solved.","comment_length":37,"text":"[Dataset created] some critical small issues when I was creating a dataset\nHi, I successfully created a dataset and has made a pr #248.\r\nBut I have encountered several problems when I was creating it, and those should be easy to fix.\r\n\r\n1. Not found dataset_info.json\r\nshould be fixed by #241 , eager to wait it be merged.\r\n\r\n2. Forced to install `apach_beam`\r\nIf we should install it, then it might be better to include it in the pakcage dependency or specified in `CONTRIBUTING.md`\r\n```\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 10, in <module>\r\n from nlp.commands.run_beam import RunBeamCommand\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/commands\/run_beam.py\", line 6, in <module>\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n```\r\n\r\n3. `cached_dir` is `None`\r\n```\r\nFile \"\/home\/yisiang\/nlp\/src\/nlp\/datasets\/bookscorpus\/aea0bd5142d26df645a8fce23d6110bb95ecb81772bb2a1f29012e329191962c\/bookscorpus.py\", line 88, in _split_generators\r\n downloaded_path_or_paths = dl_manager.download_custom(_GDRIVE_FILE_ID, download_file_from_google_drive)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 128, in download_custom\r\n downloaded_path_or_paths = map_nested(url_to_downloaded_path, url_or_urls)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 126, in url_to_downloaded_path\r\n return os.path.join(self._download_config.cache_dir, hash_url_to_filename(url))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/nlppr\/lib\/python3.7\/posixpath.py\", line 80, in join\r\n a = os.fspath(a)\r\n```\r\nThis is because this line\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/src\/nlp\/commands\/test.py#L30-L32\r\nAnd I add `--cache_dir=\"....\"` to `python nlp-cli test datasets\/<your-dataset-folder> --save_infos --all_configs` in the doc, finally I could pass this error.\r\nBut it seems to ignore my arg and use `\/home\/yisiang\/.cache\/huggingface\/datasets\/bookscorpus\/plain_text\/1.0.0` as cahe_dir\r\n\r\n4. There is no `pytest`\r\nSo maybe in the doc we should specify a step to install pytest\r\n\r\n5. Not enough capacity in my `\/tmp`\r\nWhen run test for dummy data, I don't know why it ask me for 5.6g to download something, \r\n```\r\ndef download_and_prepare\r\n...\r\nif not utils.has_sufficient_disk_space(self.info.size_in_bytes or 0, directory=self._cache_dir_root):\r\n raise IOError(\r\n \"Not enough disk space. Needed: {} (download: {}, generated: {})\".format(\r\n utils.size_str(self.info.size_in_bytes or 0),\r\n utils.size_str(self.info.download_size or 0),\r\n> utils.size_str(self.info.dataset_size or 0),\r\n )\r\n )\r\nE OSError: Not enough disk space. Needed: 5.62 GiB (download: 1.10 GiB, generated: 4.52 GiB)\r\n```\r\nI add a `processed_temp_dir=\"some\/dir\"; raw_temp_dir=\"another\/dir\"` to 71, and the test passed\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/a67a6c422dece904b65d18af65f0e024e839dbe8\/tests\/test_dataset_common.py#L70-L72\r\n\r\nI suggest we can create tmp dir under the `\/home\/user\/tmp` but not `\/tmp`, because take our lab server for example, everyone use `\/tmp` thus it has not much capacity. Or at least we can improve error message, so the user know is what directory has no space and how many has it lefted. Or we could do both.\r\n\r\n6. name of datasets\r\nI was surprised by the dataset name `books_corpus`, and didn't know it is from `class BooksCorpus(nlp.GeneratorBasedBuilder)` . I change it to `Bookscorpus` afterwards. I think this point shold be also on the doc.\r\n\r\n7. More thorough doc to how to create `dataset.py`\r\nI believe there will be.\r\n\r\n**Feel free to close this issue** if you think these are solved.\nAlright I think I fixed all the problems you mentioned. Thanks again, that will be useful for many people.\r\nThere is still more work needed for point 7. but we plan to have some nice docs soon.","embeddings":[-0.2165798396,-0.0164848957,-0.1206973046,0.3510122001,0.2939466834,-0.0238549132,0.086690411,0.4066426158,-0.1998541057,0.0216472019,0.2210519165,0.2012127191,-0.2307064533,-0.0083013931,0.1113757342,-0.1998809725,0.2577252984,0.309070766,0.21963422,-0.1952063143,-0.4448091984,0.1668979228,-0.0202732198,0.0794764906,-0.1854971647,-0.3484351039,0.0012606009,0.3041497171,-0.27669245,-0.3347336054,0.0429069027,-0.0216965955,0.061426077,0.1791488975,-0.0001148037,-0.1664888561,0.3816398084,-0.1286742687,-0.4447022676,-0.0524816103,-0.2777306139,-0.5376020074,0.1146237329,-0.3695807159,0.1170074046,-0.4766104817,-0.2099819928,-0.2251408994,0.5323174,0.3592292964,0.2335041165,0.2630361617,0.2306550294,-0.0752441138,0.3409655988,0.0986781269,-0.2740069926,-0.0909317881,0.0840169713,0.191456601,0.0731679052,0.1320601255,0.1426730156,-0.1103613377,0.4137816429,-0.1641572863,-0.0283232555,-0.5333042741,-0.0329061784,-0.0450610407,0.4773083031,-0.3531982601,-0.3011023998,-0.3424141109,0.1612422019,-0.1395581961,0.238892436,0.4321993887,-0.1544614136,-0.1136258841,-0.1933754086,-0.1405055672,-0.2878140509,0.1813217252,0.4124358594,0.0853471979,0.1599856317,0.0504578277,-0.0027547576,-0.1099818051,0.1672928035,-0.3342158794,-0.0749942809,0.3552259803,-0.1991508752,-0.1965242326,-0.0443683229,0.0313737169,0.0599546842,0.2317240387,-0.2460917979,0.0044569359,-0.0118472353,0.0019146631,0.1471191496,0.1913504303,0.1334288269,0.0652819648,0.1037480012,-0.039223291,0.2899758518,0.0508449227,-0.0023807478,0.0406709425,-0.1865070015,0.0519573689,-0.0040365495,-0.0670203939,-0.0239564143,0.1387755275,-0.3734666407,-0.2500493526,0.042333439,0.2187863439,0.1310413629,0.2003273368,-0.0677166432,0.2659163773,-0.2425889075,-0.2212205827,-0.0890087709,0.1785895973,-0.3464155793,-0.008724126,0.4041240513,0.3678691387,0.3285503387,-0.0558720306,-0.1568581462,0.0804305598,0.3698014021,-0.1725635231,-0.164951548,0.3664527833,0.1163076907,0.3178191185,-0.2282937914,-0.0065453667,-0.1832158566,0.0733555034,0.0013044693,-0.2434278727,-0.0742869303,0.2257186919,-0.3786150217,0.3559951782,0.0095038917,0.085091956,0.1143025979,-0.2616052032,-0.1059470549,0.000416379,-0.0401857942,-0.3765689731,0.2222083956,0.3557888269,-0.2729791701,-0.117356427,-0.2392379493,-0.4322434068,-0.172145471,0.0143524855,-0.1099702269,0.4447313845,-0.1886315346,0.0590488017,0.5762619376,-0.1736322492,-0.3862493932,-0.0011030184,0.0378160216,-0.161188066,0.1443196833,0.0966827124,0.0397116542,-0.2814246118,-0.3508945405,0.3981647193,-0.0608353876,0.0863966048,-0.3207007647,-0.0957875624,0.1719620228,0.2355937362,-0.009895077,0.0999968201,0.2488787919,0.2356128842,-0.0183623042,-0.1480812132,0.2485683262,0.3226508498,0.3735436797,-0.3224754632,0.0843802467,0.0609972142,-0.4909184277,0.0059003527,-0.4745084047,0.0989301354,-0.1233595684,0.004432248,-0.2629866004,-0.0900884494,-0.0647096708,-0.3086581826,0.1512106508,-0.078992255,0.1919051856,-0.1069257483,-0.3420167267,0.5311259031,-0.1893766224,0.068848528,-0.5045070648,0.218655467,-0.1654846072,-0.121896334,0.3162867129,0.591373682,0.1859377474,-0.1297962517,-0.1571085155,0.376141876,-0.0072881156,0.1002138332,0.1314674765,0.108823508,0.3348307014,-0.1476595104,0.3715452254,-0.2893131375,0.108980827,-0.2142601311,-0.3619832397,0.2575721145,-0.0945694,0.1903479993,0.2109564394,-0.0040955762,0.2478683144,-0.0905265659,0.0391285047,-0.1222396418,0.2752211988,0.2415908575,0.3407287896,0.2105956227,0.1973709315,0.2309572399,0.2959473133,-0.0584616847,-0.1581475288,0.2084072828,-0.0477813818,-0.1669331044,0.2196256369,0.4565879405,0.5612210035,0.2158060372,-0.3815115392,0.1613033712,0.0972903296,-0.1790976673,0.1301805824,0.2057107389,0.267224133,0.0388933457,0.2177975923,0.0833001956,0.0917641595,-0.012623868,0.0895320624,0.3389057219,-0.1725993156,-0.156357497,-0.031914182,-0.3260663152,-0.1701599509,0.3311553299,-0.2817266881,-0.3784973323,-0.0004050345,-0.1697893739,-0.0229077768,0.2918809056,0.2162147611,0.2618064284,-0.1883648634,0.0976374298,-0.3757290244,-0.3888498545,-0.1720531881,0.1114448011,0.1218660325,0.3295152783,0.5941365957,-0.0154186925,0.12340004,-0.1669679731,-0.1062984541,0.0704004169,-0.0434974879,0.1907853782,0.0079046916,0.2060802579,-0.0516079739,-0.2771467268,0.1803607047,-0.0872100592,-0.2609621286,-0.0134241097,-0.147666499,-0.0718898848,-0.0679896995,-0.5194193125,-0.4250129759,-0.2148083299,0.1327863038,0.3100067079,0.1828389317,-0.0841341466,0.141230613,-0.0181512292,-0.0019709226,-0.0615515746,0.0066660936,0.0194676388,0.2024261504,-0.2755369544,-0.4629991949,0.0172609836,-0.2323603481,0.1478987336,0.1514241397,-0.3561945558,-0.0858034939,-0.1002788767,0.2035038769,0.0486067049,0.2767562568,0.3269117773,0.1419498324,0.0358995572,-0.1744469106,-0.3353277445,0.1088822708,-0.1908497959,0.59681952,-0.0053010136,0.1931440681,-0.3918654323,0.8372858167,0.4679619968,-0.1322748065,0.2019901276,0.0645503551,0.2530503273,-0.0333257355,-0.2155572325,0.5519694686,0.1865862757,0.1339824796,-0.0086955437,0.1465135515,-0.2078359425,-0.2886020243,0.1719544679,-0.7883492708,-0.1010141745,-0.1152329817,-0.132088989,0.2729951739,0.0884637088,-0.0757809579,-0.0928049907,-0.1519215107,0.0211826358,0.0190127604,0.1256043166,-0.0408476442,-0.2778846622,-0.1165947616,-0.1661058217,0.1609886736,-0.1069056168,0.2166684717,-0.0879853219,0.0053945784,0.0756809562,-0.0121861566,0.6508090496,-0.4458170235,-0.2502537668,0.1428911984,0.4055549204,-0.3814205229,0.1012092978,-0.1455695629,0.158567667,0.1066678092,0.2751333416,-0.4258680046,-0.0385760106,0.1401399225,0.3697842062,-0.160544157,-0.1771896034,-0.0975282714,-0.2522131801,-0.2250371128,-0.3424991667,-0.6088339686,0.1231244728,0.0420012511,0.2845884562,0.0299565531,-0.0380930267,0.0616338961,0.1306768805,0.4030514956,-0.1472469568,0.2574329376,0.0296916068,0.1881632656,0.2162958086,0.3045300543,-0.0867219493,-0.0720852688,0.0461534262,-0.11099042,0.3767855167,0.2177415192,-0.0714936703,-0.2043244392,-0.0884054229,0.0822213739,0.0422078259,0.0134127671,0.1663259268,-0.2151949853,-0.2809138596,-0.3256173134,0.4102795422,0.1630631685,-0.0794472471,0.1998076737,-0.1293647587,-0.0149041219,0.5705702305,0.1807837635,1.1641917229,0.0162149724,-0.1489294767,0.3272439241,-0.3185850382,0.7361056805,-0.9293342829,0.0379289575,-0.4258750677,0.2147131562,0.1235820502,-0.228075847,0.2144611925,0.0830759332,-0.1261196136,0.1880285591,0.1353084594,0.1012107059,-0.114755854,0.6493266225,-0.0398233831,-0.4592416883,-0.5051500201,0.1028768122,-0.2891237438,0.0760233328,-0.1944546551,-0.1979286373,-0.3322232366,-0.2097279429,-0.1891156435,0.140247792,-0.1220817715,0.1048285365,0.082154274,-0.5340114832,0.1189398244,0.1871011704,0.4441755414,0.0693396553,-0.2172711492,0.3504209816,0.0356042162,0.0122375796,0.1434174031,0.156142965,0.194249928,-0.223055765,0.0185003784,0.1097248495,-0.3166891038,0.1906911135,-0.2579829991,-0.2314601541,0.2656127214,-0.1680596322,-0.0141959451,0.1864646375,-0.179586336,-0.1864559352,0.1789260209,0.052234605,0.0230730865,0.2207249999,-0.0080895051,-0.0254745204,-0.2591033578,0.1850200295,0.2551265955,-0.4933585525,0.2920069993,0.0143672777,-0.1413042992,-0.3368189633,0.3393170238,-0.0217135046,-0.549806416,-0.0471096225,0.1549621969,-0.1732402891,-0.1225043088,-0.0646256283,0.1288732588,0.1655765772,0.1828327626,-0.3124121428,-0.3814761341,0.2685787678,-0.1478324682,0.1492538601,-0.3125891089,-0.0533510149,0.1307689697,0.2059382647,-0.3075966239,-0.2742983401,-0.0850652829,0.1111879051,0.0690172836,-0.0001548357,-0.1310928911,-0.3503128886,0.10763859,-0.1060899794,-0.3347283602,-0.1759790033,0.0289696325,0.1521108449,0.0870237499,-0.1402989328,-0.1382886469,-0.0890989006,-0.0205531791,-0.1737530529,-0.0808127373,-0.2732879519,-0.0885648057,0.0174123812,0.1678901166,0.1153389364,-0.2374764383,-0.026261678,-0.0853664726,0.2476085275,0.0834323093,-0.106934458,0.1269097924,-0.0297013968,-0.2137421668,-0.1161093339,0.0099907881,0.1641687602,0.1888871193,-0.1789885163,0.0328501016,0.0628117621,0.3357524276,0.2211795747,-0.2855054438,-0.0039075133,0.258782804,0.1599668264,-0.3051172197,0.1355106235,0.5688182712,0.116217427,-0.192642644,0.1513213515,0.1291161329,-0.0586773381,-0.0556202792,0.1372754425,0.5175210834,-0.1722085476,0.3115231097,0.2180622518,0.0318760425,0.0112811523,0.035239812,0.1726840585,0.0470054373,0.6041743159,-0.0963322371,0.5378553867,0.276620388,0.1715888083,-0.0179595184,-0.2979881167,0.0516243875,0.0835087001,0.0514033437,-0.0647090077,0.0556912832,0.475015074,0.0523283966,-0.2881971598,-0.1949679106,0.0889088064,-0.0953513607,0.4037155509,-0.1788719743,-0.2117675096,-0.0986685753,-0.1745070368,0.0179083161,-0.1983197927,0.364806056,0.185503602,-0.1240945458,-0.1778636426,0.4243930876,0.0772921667,0.2227859348,-0.0287255794,0.2480990291,0.0141337169,-0.0970877856,-0.1475265324,0.2929859459,0.3658083975,0.4779532254,-0.1898424327,0.0621615611,0.2308991849,-0.0886326879,0.0779972747,0.4141319394,0.3960823119,0.1710058153,0.1178681254,0.1362997293,-0.1313177794,-0.0407680348,0.2802560627,0.0961469188,-0.0316606537,0.1877766252,-0.2365777344,-0.4242991805,-0.0851354077,-0.0414729193,-0.3212204576,-0.1003718749,0.048656065,0.2468833178,0.0312168244,-0.288523823,0.0620972738,-0.1806997508,0.5545122623,0.1600003093,0.093418695,-0.4730508029,-0.339553833,-0.6459328532,-0.1185649484,-0.2552911043,0.0187612884,-0.0552571602,0.0170525331,0.0001157257,0.296777606,0.1137385443,0.5873936415,-0.045605015,-0.0323732086,-0.5817576051,-0.0594591089,0.0666519701,0.0310340524,-0.1627264768,-0.0509418994,0.1158232912,-0.099445127,0.0204731133,-0.0383134261,-0.3848719895,0.3540809453,0.2101806998,0.3670881987,0.2355558872,0.3461514115,-0.0735239387,-0.0712920278,-0.0996794179,-0.2440582514,0.1177210361,0.27832973,0.0427765362,0.1906647831,-0.0408141315,0.1475720257,-0.4616036415,-0.0035805705,-0.1504666656,0.295078665,-0.1972342879,0.0702444836,-0.0566571318,0.2186337709,0.1541198194,-0.1112893,-0.1131286323,0.291503489,-0.4034203291,-0.1097390428,0.4648357332,-0.1177231297,-0.0457417481,0.2879976034,0.2958430648,-0.1753144413,-0.0571581312,-0.5936287045,-0.0158088785,0.3534444869,-0.2711624801,-0.0148731638,0.0799034014,-0.1544134021,-0.1725821346,0.0343465582,0.1275982708,-0.1552858055,-0.1353527606,0.1812076569,0.0627781302]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/246","title":"What is the best way to cache a dataset? ","comments":"Everything is already cached by default in \ud83e\udd17nlp (in particular dataset\nloading and all the \u201cmap()\u201d operations) so I don\u2019t think you need to do any\nspecific caching in streamlit.\n\nTell us if you feel like it\u2019s not the case.\n\nOn Sat, 6 Jun 2020 at 13:02, Fabrizio Milo <notifications@github.com> wrote:\n\n> For example if I want to use streamlit with a nlp dataset:\n>\n> @st.cache\n> def load_data():\n> return nlp.load_dataset('squad')\n>\n> This code raises the error \"uncachable object\"\n>\n> Right now I just fixed with a constant for my specific case:\n>\n> @st.cache(hash_funcs={pyarrow.lib.Buffer: lambda b: 0})\n>\n> But I was curious to know what is the best way in general\n>\n> \u2014\n> You are receiving this because you are subscribed to this thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/nlp\/issues\/246>, or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ABYDIHKAKO7CWGX2QY55UXLRVIO3ZANCNFSM4NV333RQ>\n> .\n>\n","body":"For example if I want to use streamlit with a nlp dataset:\r\n\r\n```\r\n@st.cache\r\ndef load_data():\r\n return nlp.load_dataset('squad')\r\n```\r\nThis code raises the error \"uncachable object\"\r\n\r\nRight now I just fixed with a constant for my specific case:\r\n```\r\n @st.cache(hash_funcs={pyarrow.lib.Buffer: lambda b: 0})\r\n```\r\nBut I was curious to know what is the best way in general\r\n\r\n","comment_length":151,"text":"What is the best way to cache a dataset? \nFor example if I want to use streamlit with a nlp dataset:\r\n\r\n```\r\n@st.cache\r\ndef load_data():\r\n return nlp.load_dataset('squad')\r\n```\r\nThis code raises the error \"uncachable object\"\r\n\r\nRight now I just fixed with a constant for my specific case:\r\n```\r\n @st.cache(hash_funcs={pyarrow.lib.Buffer: lambda b: 0})\r\n```\r\nBut I was curious to know what is the best way in general\r\n\r\n\nEverything is already cached by default in \ud83e\udd17nlp (in particular dataset\nloading and all the \u201cmap()\u201d operations) so I don\u2019t think you need to do any\nspecific caching in streamlit.\n\nTell us if you feel like it\u2019s not the case.\n\nOn Sat, 6 Jun 2020 at 13:02, Fabrizio Milo <notifications@github.com> wrote:\n\n> For example if I want to use streamlit with a nlp dataset:\n>\n> @st.cache\n> def load_data():\n> return nlp.load_dataset('squad')\n>\n> This code raises the error \"uncachable object\"\n>\n> Right now I just fixed with a constant for my specific case:\n>\n> @st.cache(hash_funcs={pyarrow.lib.Buffer: lambda b: 0})\n>\n> But I was curious to know what is the best way in general\n>\n> \u2014\n> You are receiving this because you are subscribed to this thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/nlp\/issues\/246>, or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ABYDIHKAKO7CWGX2QY55UXLRVIO3ZANCNFSM4NV333RQ>\n> .\n>\n","embeddings":[-0.0791417584,-0.0054426533,0.0433646031,0.0309611745,0.3594623804,-0.1114412323,-0.0066469391,0.1171398461,0.0686439276,0.1308163553,0.1577533782,0.0367619656,-0.1826368272,0.1247068271,0.4915667176,-0.0963596031,0.0291531328,0.1217107698,-0.0822886825,0.0933153853,-0.1159205288,-0.2149104029,-0.0649334118,-0.1643327922,-0.29099527,-0.1210641488,-0.1207960397,0.0799476579,0.0311479159,-0.2001817971,0.3986590505,0.191817075,0.1369281113,-0.0135208312,-0.0001179134,-0.0179088991,0.3497399092,-0.2958795726,-0.3403178453,-0.2064831704,-0.3478736877,-0.3250800073,0.1356263608,-0.0289090984,-0.316403389,-0.1195975393,0.2626536191,-0.440908879,0.4450412095,0.1855327934,0.1102296188,-0.0131289838,0.1342876852,0.3962053359,0.0525613651,-0.1016618237,-0.1999524236,-0.1110631078,0.0836441219,0.0082153268,0.0463165529,0.2444910556,-0.1063841432,-0.0312433932,0.6360678673,-0.0506441221,-0.464953959,-0.2141303718,-0.0844174176,0.1957267821,0.6592319012,-0.2575154901,-0.5362505317,-0.3742972016,0.0216725916,-0.4770931304,0.1869377196,0.1748289317,-0.0589700304,0.2702479661,-0.4797565043,-0.3037990034,0.044083152,0.1916121244,0.2693197131,0.0252833385,-0.0207847878,0.1667786837,0.0161773544,-0.1872114241,0.1719463915,-0.0714344233,-0.1203716174,0.2191658467,-0.3066838086,0.1190677658,0.0155530497,0.2072263509,0.1255341023,0.1493085772,0.5070616007,0.2352126241,0.128145799,0.076477088,-0.0187271051,0.4619321227,0.0078621535,-0.0626628101,0.102245681,-0.1494984925,-0.2285330147,-0.1756053567,0.1721443832,0.2349360883,0.2839807272,0.1700600982,0.0865028948,-0.3802713752,0.1799581796,-0.108186543,-0.2296739519,-0.2214444578,0.1230745986,0.3695170879,-0.1252463013,0.295811832,-0.2560270131,0.0321612917,-0.1035373136,-0.3668451607,-0.1925476789,0.008249661,-0.0521777943,0.3953482211,0.1942145824,-0.487600714,0.1344993711,-0.0439002626,-0.0759318992,-0.0127570201,0.2849001884,-0.179495275,0.2617857456,0.1858826131,-0.0053232717,0.0123102395,-0.0514554717,-0.0501557589,-0.2150112987,0.2134146988,-0.1483264267,-0.7214868665,0.2737378776,0.0565578006,-0.4841581285,-0.1595443636,-0.3712064326,0.5935316086,0.2733146846,-0.0156454053,0.1179438531,0.0013978113,-0.1824920177,-0.3100670874,0.2377763093,0.2972569764,-0.0023034422,-0.1332862973,0.1639194489,0.0847359374,0.2712289989,0.2201752216,-0.3504532874,0.0753714368,0.0694509149,-0.1782054901,0.640879333,-0.1535963416,-0.5387370586,-0.0192201193,0.2734828889,0.2440693825,0.1072872877,0.315964669,0.0151742958,-0.0607365854,-0.0661827102,0.4188011289,-0.1644411832,-0.0135387508,-0.312063396,-0.428517282,0.5561183691,0.0591845736,0.0112291435,0.0635528639,0.0541882142,0.5534829497,-0.1423951238,-0.156803906,0.0572758839,0.4245156348,-0.0728099421,-0.1383220255,-0.0748131797,-0.1953760684,-0.7898908257,0.1433190703,-0.1094537377,-0.0945420861,-0.0670008063,-0.2357566804,0.4408468902,-0.1568446904,-0.0268691014,-0.1255964488,-0.0126039758,0.0297756214,0.5305255055,-0.1319667995,-0.4664229751,0.57159096,0.1120155901,0.1440637261,-0.3560891151,-0.1206869408,0.067279242,-0.3497923613,0.1039767414,0.2700039446,0.1601337045,-0.1713483781,-0.1626326293,0.1947911531,-0.2408110052,0.3363534212,-0.1240032315,0.3826124668,0.0867004916,-0.1623900086,0.0480391346,0.1188599616,-0.0355735011,0.0517503619,-0.0069937925,0.2490597665,0.072094366,0.0005589007,0.0950622708,-0.2466789037,0.0446445569,-0.0481995195,-0.0835933238,0.0030469694,0.4813488722,-0.1932514161,0.433524549,0.3845349848,-0.2530942261,0.359615922,0.6983598471,0.0457331762,0.2178187668,0.1418946087,0.0875151604,-0.5609429479,0.2564837635,-0.1957597286,0.1106264442,0.1782255024,0.3006596565,-0.0745081827,0.0362934545,-0.1051021442,0.1754598469,-0.1084127873,0.0257788412,-0.2941855788,0.3341449201,0.0356033593,-0.0822640806,0.1576630473,0.06330695,0.2589777708,-0.0923607871,-0.0568176024,-0.343698293,-0.3930536807,0.0127122421,-0.1565709412,-0.4962465465,-0.0980674699,-0.0486263894,0.1884430349,0.0583713017,0.0401819311,-0.3619216681,0.2649368346,-0.274990797,-0.3841170967,-0.3828631938,-0.1650548726,-0.3346540034,-0.0545994788,0.1334372163,-0.3610694408,0.2618533373,0.1698355824,0.0411344692,-0.1050665155,-0.2228546441,-0.0311315693,0.1255652457,0.2201830596,-0.1370445192,0.1204437912,-0.0660410747,-0.0860445127,0.2279858291,-0.3752197623,-0.1063523069,-0.0495339595,-0.0477896109,0.2464321107,-0.0740276575,-0.1765855104,-0.4447836876,-0.2995262444,0.2155094594,0.0241212938,0.019564528,0.0071496568,-0.2222803086,-0.0165380016,0.0308464728,0.0575084165,-0.2335250974,-0.5710589886,0.3007079959,0.1204299182,-0.330660969,-0.1724932194,-0.0910754725,-0.0497761331,0.3589242101,-0.4684201479,-0.1763007939,-0.163080886,0.0598162822,0.133515954,-0.3932551742,0.3917959034,-0.0955056325,0.0246955268,0.001585937,-0.3434217274,-0.0301468764,0.0686947107,0.467453301,0.3109143376,0.1002315879,0.3690013587,0.6891123056,0.1336451173,0.0258559529,0.4306862354,0.26535815,0.3063560426,-0.1425518394,-0.2282550633,0.3457773626,-0.1200147346,-0.0282587279,-0.0342092998,0.2829247117,-0.1994740367,-0.2714641988,-0.0196008701,-0.2883556783,-0.1858429909,0.2942028642,-0.3300023973,0.157417506,0.3139321506,-0.0177909639,-0.1596551836,-0.5740087032,-0.0312711746,0.2470716089,0.2790088654,0.4540108144,-0.5400013328,-0.1497484297,-0.619351685,0.5108376145,0.1088171452,0.2713660896,-0.2866216898,0.0302325394,0.0103108874,-0.3334999382,0.1613865495,-0.072793819,-0.2789882421,-0.1065143347,-0.1353911906,-0.1662480831,0.0431875698,0.2015491426,0.3970452249,0.0055099628,0.0590565093,-0.2887251079,-0.1980149895,-0.167836383,0.0947323442,-0.2160205245,0.049985826,-0.1566386521,-0.2946563661,-0.1753402352,0.0266250093,-0.3478452563,0.0447502583,0.057437364,0.0806914717,-0.1845290065,0.2739923894,0.0841095746,-0.1747661382,0.1975331008,0.0990923494,0.1353946924,0.2745622694,0.1315345019,-0.1907171607,0.4376656413,0.0030059768,-0.1152983531,-0.2782945633,0.2551507652,0.2479103953,0.1165401191,-0.1841938794,0.2385423928,0.3922950327,-0.185226813,-0.3070905507,0.2162282318,-0.1751150191,0.1118567511,-0.2902914584,-0.7748069167,0.3710681796,0.1772645861,-0.0068604117,0.4467784166,-0.338110745,-0.2220686674,0.2109927684,0.2604379952,1.084956646,-0.2421605736,0.3112491667,-0.0048062406,0.0196265019,0.6438978314,-0.2434020042,0.241110757,-0.1757462621,-0.2439703196,-0.0254981015,-0.1419060826,0.0019547802,-0.0349228121,-0.0801820531,0.5324956179,0.3090100586,0.0917510688,0.0406324938,0.2257103026,-0.1218478605,-0.2654405534,-0.6026583314,0.1311962157,-0.5431944132,0.166269511,-0.089052178,-0.0208312646,0.1979009956,-0.2085479796,-0.1069360226,-0.0099033052,-0.5662989616,-0.0229298249,-0.0276683159,-0.3394787908,-0.4411351979,0.1882594228,0.0372430794,-0.1779053658,-0.073292844,0.0035077462,-0.0558985136,0.1271161586,0.0973175615,0.1275643408,0.4872748852,-0.0371156037,0.0352727361,0.0227141567,-0.0715246797,-0.5310295224,-0.0768446326,-0.1420334131,0.0875499919,0.0553521849,-0.1287155747,0.2509489059,-0.0207514316,-0.0314291604,0.0565884635,0.3216857612,-0.0008592388,0.0704560354,0.0511848778,-0.1460152417,-0.0167302489,0.3425778449,0.0300936457,-0.2030135989,0.3337711692,-0.1036188826,-0.1535341591,-0.0164782833,0.0557689369,-0.027832754,0.0011393939,-0.0771223381,-0.2109164298,-0.0136798834,-0.1393114328,0.1672440469,0.1111034229,-0.2254132479,-0.1048463359,-0.2211010158,-0.6179187894,0.3391147256,0.1685263813,0.1029957458,0.3037675619,-0.0936101377,0.0379217379,0.3011038899,-0.2404963821,-0.2288473994,-0.1420495212,0.2012725174,0.2291985899,-0.1686756611,0.3836634755,-0.1169655025,-0.0906308219,-0.0290362127,-0.3670320809,-0.2013751566,-0.0605911314,0.1516954899,0.177612558,-0.3068909049,-0.4292011559,-0.1945921183,0.3183097839,-0.2683458924,0.0335932784,0.2106244117,0.0352399759,0.1416852474,-0.1328350455,0.0779266581,-0.1401215196,0.0989072248,0.1196066216,0.1859522015,-0.0801290944,0.215586558,-0.0086469054,0.1103690192,0.0066190618,-0.1206855923,0.2880953848,0.1817902923,0.1528450847,-0.345980674,0.0909100994,-0.0258717295,0.141715914,-0.0427657999,-0.1329607368,0.2724699676,0.2716389298,0.0420847014,-0.0252033453,0.0708196089,0.1947019845,0.1654705703,-0.1097919047,0.2165461332,0.1546693891,0.173900485,0.0819318742,-0.0681846887,0.6233013272,-0.4634339511,-0.0610285103,-0.3458017111,-0.0111765219,-0.1796018481,0.320538491,0.3976679146,0.0294054709,0.4939382672,0.1547967643,0.287813127,0.3103057146,0.1773991734,-0.1119694561,-0.2488233745,-0.0878349021,0.5038475394,0.1372719109,0.3859373927,0.0950917602,0.3510256112,0.1113976687,-0.4690274,0.3331849575,0.2070385963,-0.0470947996,-0.2459054887,0.0073095178,0.1810505986,0.07215821,0.1140941679,0.1187763661,-0.1582904756,0.396315515,0.1978647262,0.2835291326,-0.290372014,0.1538341343,0.4110330343,0.238155216,-0.0564700328,0.6531962752,0.0223594159,-0.1039643586,0.1026450694,0.2962405086,0.2218841016,0.0270809419,-0.2486175746,0.3493115604,0.0568063073,-0.019457791,-0.4021539092,0.3465339243,0.1706774384,-0.0117615452,0.1978525817,0.057263542,-0.0054115122,0.4223130047,-0.0454500094,0.1003605872,-0.1707139611,0.3791207671,0.2329505831,-0.1478307992,0.0068826373,0.2115757912,-0.1929095536,0.0895629004,0.3285349905,-0.0156546086,0.2073533386,-0.1555515975,0.0604512915,0.2184076905,0.4291932583,0.2490694672,0.3821173608,-0.161790058,0.1574469358,-0.5565519929,-0.0210790895,0.0088307541,0.1861978769,0.0010196978,0.2689861953,0.0828867331,0.2567797005,-0.0606351197,0.1698958725,0.265424788,-0.105388619,-0.295450598,-0.1335500926,-0.0017767492,0.2544241548,0.059433274,-0.2419894338,0.2442355156,0.0143125756,-0.080450587,-0.2680755556,-0.161462307,0.1166236773,0.2480138987,0.2433748096,0.3024022281,0.208225742,0.0650372058,0.0396098718,-0.2174969763,-0.1397913098,-0.2429805249,0.0514190681,-0.2882683277,0.0840623751,-0.4057172239,0.0879705623,-0.3254971504,0.1394613981,-0.0682579949,-0.178735435,-0.0908429474,-0.077418007,-0.0623688623,-0.0214028787,-0.0698384717,0.5901696086,-0.0621928498,-0.0410887934,-0.3418334126,0.0770092458,0.4359880388,-0.4491095245,0.0984471217,0.0508569628,0.0931319222,0.6444492936,-0.22570391,-0.6652424335,0.0111795366,0.313685298,-0.2641386986,-0.0376281887,0.1506170481,-0.0427328311,0.1526153535,-0.1511933953,0.1468613744,-0.0875310376,-0.1962923855,0.1462199092,-0.0582072027]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/246","title":"What is the best way to cache a dataset? ","comments":"Closing this one. Feel free to re-open if you have other questions !","body":"For example if I want to use streamlit with a nlp dataset:\r\n\r\n```\r\n@st.cache\r\ndef load_data():\r\n return nlp.load_dataset('squad')\r\n```\r\nThis code raises the error \"uncachable object\"\r\n\r\nRight now I just fixed with a constant for my specific case:\r\n```\r\n @st.cache(hash_funcs={pyarrow.lib.Buffer: lambda b: 0})\r\n```\r\nBut I was curious to know what is the best way in general\r\n\r\n","comment_length":13,"text":"What is the best way to cache a dataset? \nFor example if I want to use streamlit with a nlp dataset:\r\n\r\n```\r\n@st.cache\r\ndef load_data():\r\n return nlp.load_dataset('squad')\r\n```\r\nThis code raises the error \"uncachable object\"\r\n\r\nRight now I just fixed with a constant for my specific case:\r\n```\r\n @st.cache(hash_funcs={pyarrow.lib.Buffer: lambda b: 0})\r\n```\r\nBut I was curious to know what is the best way in general\r\n\r\n\nClosing this one. Feel free to re-open if you have other questions !","embeddings":[-0.310174495,0.117244564,0.1047027335,0.2134909928,0.2868501246,-0.1405108571,0.0429240316,0.1152483076,-0.0369445235,0.0604081266,0.1200718731,0.1375265867,-0.2138116211,-0.0080181872,0.3084204197,-0.018134417,0.0109476084,0.1576344073,0.2011000365,0.0810203403,-0.2442805022,-0.1781571358,0.0018782815,-0.1148710325,-0.2517766953,-0.2657901645,0.2069189996,0.2394376397,0.0620810091,-0.1863553971,0.2036979645,0.1977326274,0.1545430869,0.1077425703,-0.0001296005,-0.0382198654,0.2750402391,-0.2156710327,-0.2973525524,-0.0484055877,-0.4325699508,-0.1956165284,0.1916221678,-0.0318454541,-0.1392918378,-0.121284917,0.1828835458,-0.7369990945,0.5135641098,0.1572404653,0.0443469025,-0.095142968,0.1268144697,0.2956528664,0.1125491112,-0.3512252569,-0.1182235181,-0.2383983284,-0.0935820267,0.0994641632,0.2067803591,0.2125538737,-0.3924664855,-0.1465841532,0.533039391,-0.0918784514,-0.315040648,-0.0715905204,0.0259507634,0.0310260877,0.91229707,-0.2706373632,-0.4271051288,-0.137110889,-0.0532278642,-0.4022369683,-0.0384425744,0.2170812041,-0.1618963331,0.3162564337,-0.4905697107,-0.34083125,-0.1033780649,0.1618768126,0.1457334757,-0.117395401,0.140711844,0.3386280537,0.1486148238,-0.2293973565,0.2932570279,0.0803423449,-0.0506959073,0.0634045824,-0.1928029805,-0.0625830963,0.0894509777,0.2609757185,0.1757541299,0.0993023664,0.3401486874,0.2752220035,0.3582753539,0.1301605701,0.1045230478,0.4695494473,-0.0771730542,-0.154479444,0.2217377126,-0.3294048607,-0.5228472948,-0.1104923338,0.1290091425,0.034694165,0.5628107786,0.1277716607,0.1267994195,-0.5021237731,-0.0084227854,0.2872204483,-0.399151206,-0.0460366458,-0.016607983,0.5013529062,-0.1084878743,0.2512266338,-0.1456984878,-0.1857033223,-0.1997852474,-0.2378776968,-0.1128893271,0.0125058787,-0.0116071971,0.4658547342,-0.0499161147,-0.3333698809,-0.0869756192,0.0455433279,-0.1159123033,-0.1236797124,0.4393692017,-0.1935649514,0.1853521615,0.2352021933,0.0553954616,0.0040019392,-0.0008859211,-0.3320551813,-0.2152854353,0.4344194233,-0.1514505893,-0.7042740583,0.1000124812,0.0274495594,-0.5554084778,-0.1002325639,-0.2592122257,0.3653868735,0.1946675181,-0.0796028748,0.1185564473,0.0440400727,-0.1238440648,-0.457710743,0.2446618527,0.3344351649,-0.0925476253,0.1320668906,0.0947054401,0.1523089707,0.3187779784,0.1925479621,-0.2207202911,0.0839874744,0.1111123785,-0.3816383183,0.6137375832,-0.0003954783,-0.5266039371,0.0941613317,0.1371959299,0.0737063959,-0.0289712548,0.4884966612,-0.0771400556,-0.0685536638,0.1089811847,0.2473538667,-0.1882804781,-0.1011370644,-0.475451529,-0.3102246225,0.6019746065,-0.0031702407,0.1586968005,-0.0486376993,0.086992234,0.5658985376,-0.2408880889,0.0493430197,-0.0219207425,0.3800468147,0.2166836858,-0.154464066,-0.17922993,-0.3710638583,-0.6975345016,0.0807353407,0.0179804079,-0.329495877,-0.0368140601,-0.160415709,0.2422823757,-0.0141147627,0.0433922596,0.0134448614,-0.0755264238,-0.0497363657,0.253487289,-0.2175176442,-0.3996038139,0.6263814569,0.2758665085,0.1517071426,-0.1713427901,-0.1011082605,0.1083116382,-0.3181031644,-0.1709849089,0.1742287427,0.2846879363,-0.0689709261,-0.0572295114,0.2362248451,-0.0127880424,0.0462782755,-0.1734223664,0.4421717823,-0.2258623093,-0.1840889156,-0.1090306267,0.1093352959,-0.1133695394,0.1129053757,-0.0658169091,0.239967525,0.0708876252,-0.0944351703,0.0355157852,-0.1831641346,-0.1483948529,-0.1586382538,-0.0531400144,-0.0362578444,0.3901351094,-0.0532368571,0.529699266,0.4877519906,-0.1660871655,0.2217973471,0.5489859581,0.1079675406,0.2193629444,0.1909153312,0.0562552847,-0.3215408921,0.1883885413,-0.1267296225,0.2300459445,0.1510009021,0.2377835661,0.1112894118,-0.1218946949,0.0237774234,0.1326234788,-0.0777588263,0.0393483229,-0.21955055,0.2960981429,0.0796394572,0.0451987013,0.0684474111,0.0469507799,0.254098177,-0.0058276979,0.1809200346,-0.2800860703,-0.1400198638,-0.0920192152,-0.2715394497,-0.40459916,-0.0538381822,0.0202657022,0.2653134167,0.0488662869,-0.0100077838,-0.5270440578,0.2226333022,-0.3770727813,-0.4685558677,-0.1273347586,-0.2794842422,-0.2316373289,-0.1342262179,0.0784613639,-0.4553861916,0.3064935505,0.1253527254,-0.003778416,-0.145124957,-0.1569648534,-0.1088038236,0.2286414653,0.3577004373,0.0126308417,0.2913803756,-0.3043758273,-0.0602014735,0.2862736285,-0.243032679,0.176401943,-0.0393018238,0.0241531152,0.1168635562,0.0910050496,-0.2425187677,-0.312468946,-0.38237077,0.1572480798,-0.1275885701,0.0017931114,0.010163581,-0.1268842518,-0.201471597,0.109632045,0.034159556,-0.1605504602,-0.6491348743,0.2054668665,0.0851330236,-0.1319423765,0.0626792237,-0.0259979703,0.0632822663,0.4642732739,-0.4234844148,-0.1802437603,0.037961334,0.0767253935,-0.0323930159,-0.3729501665,0.4331277311,0.0217464603,-0.0014251729,-0.1163872331,-0.2766215205,-0.0829273239,-0.0013629635,0.4571456015,0.2629882097,0.1544141769,0.2969298661,0.6611608267,0.2419227809,-0.0414889827,0.2784421444,0.1751492321,0.2554753125,-0.0745511577,-0.10328307,-0.0036714114,-0.0292501375,-0.438621968,-0.0156175448,0.3657493293,-0.2226841152,-0.2967616618,-0.0033127868,-0.2073325962,-0.2178803682,0.4150034189,-0.3251504898,0.2524392307,0.1976549625,-0.0355887264,-0.1803926826,-0.4845893681,-0.1734713912,0.3754156232,0.3035390377,0.3617029488,-0.5187466145,-0.2374311686,-0.3021793962,0.5728958249,0.0806482807,0.3727716804,-0.298358798,0.0216279197,-0.1111063883,-0.2184045017,0.2099183649,-0.1201362014,-0.3244763911,0.1045560092,-0.0571895577,-0.2886546552,0.0284071136,0.359686017,0.4399397671,-0.1261214018,0.0278801601,-0.3633699715,-0.0519037098,-0.1324366927,0.1329138577,-0.1350893676,0.0032891699,-0.1827584207,-0.2866635621,-0.0708864555,0.1905262619,-0.2738898098,-0.02323048,0.1611170173,-0.0286580715,-0.2541847825,0.1728293151,0.2580966651,-0.0129954908,0.1915144622,0.1561555117,0.2491767555,0.1669579595,0.2456421405,-0.0482978821,0.4810461402,-0.1418419927,-0.0544862002,-0.4107666612,0.4087016582,0.4222156405,0.0618970394,-0.2981919944,0.285579741,0.2698124945,-0.0103870276,-0.187557146,0.2725617588,-0.1658095568,-0.0242396109,-0.3841964602,-0.7331666946,0.4099384844,0.1559613943,-0.1666434109,0.5213503242,-0.2528426051,-0.0256350469,0.0918423608,0.1835488975,1.3651162386,-0.3260684013,0.35742414,-0.2397343963,-0.1082518026,0.5579128861,-0.3771129549,0.2191927582,-0.225015521,-0.2729457617,-0.01872232,-0.2409680933,-0.1412345916,-0.0416683294,-0.1516653448,0.4110102654,0.0946115255,0.1637973189,0.035332188,0.0431028754,0.0374777913,-0.0256547276,-0.6322296858,0.0393591896,-0.6259350181,0.0831289515,-0.2801426649,0.0814271495,0.2894120812,-0.2026140094,-0.1911465675,-0.1325080246,-0.5691111684,0.0642676502,-0.1135199815,-0.5292815566,-0.4682179689,0.1008734033,-0.0318486951,-0.0522718877,0.0092730746,-0.0758343413,-0.0587735623,-0.0175250545,0.024539208,0.0609365441,0.3859964311,0.0564886592,-0.0522642061,0.2242880911,-0.0601825453,-0.4772463143,0.0367591083,-0.0548664816,0.3230757713,-0.237673074,-0.0178658366,0.0715435222,-0.1057053953,0.0242510866,-0.0026605411,0.3479240835,0.1458689868,0.2555732727,0.0065960055,-0.251803726,0.0312688313,0.3102264702,0.0174279939,-0.1973046809,0.1200457662,-0.1292476356,-0.1439303309,-0.0404206403,0.0340083577,-0.1748500764,0.0879093036,0.159667179,-0.2284775674,0.0882633999,-0.1671569347,0.0506385341,0.1570065022,-0.252604574,-0.0344539508,-0.1644055545,-0.6256610751,0.2977868617,0.2139895856,0.2464789748,0.287897706,-0.0884669349,-0.0970480591,0.1199115068,-0.141135022,-0.2699642181,-0.3120149672,0.0798214823,0.1011507213,-0.0184895061,0.153736338,-0.2224977911,-0.1341478378,0.0184699502,-0.3420242667,-0.0771667734,-0.1191891804,0.2491384298,0.1717094034,-0.2477356046,-0.1606161743,-0.3698169887,0.4545792043,-0.1507902741,0.077307418,0.3152421713,0.0646357462,0.2337135077,-0.0551811643,0.0308994297,-0.3545413911,0.0641668513,0.2449958026,0.4928550422,-0.1272874773,0.1401847452,0.1815065891,-0.0215873607,-0.1233818829,-0.0882649049,0.5139016509,0.0560787842,0.1252999902,-0.2628160715,-0.0296860132,0.0188260525,0.0838598683,0.1168365702,-0.1294040978,0.3700538278,0.3033795357,0.000295462,0.0329629704,-0.033254642,0.1014010534,0.1746535003,-0.1555770785,0.3159380555,0.0607937835,0.0342485979,0.0813974068,-0.0127529558,0.7199373841,-0.3346892595,-0.0661141276,-0.3505709171,-0.0353707448,-0.2782585323,0.2499202937,0.4111675918,0.1078695059,0.4384518862,0.1956993192,0.3414745927,0.4407964647,0.2023214847,-0.205741331,-0.0811904669,0.0555878393,0.5246928334,0.2661111355,0.2130498588,0.1071494371,0.3052016199,0.1395684183,-0.2978796065,0.3547431231,0.1463856101,-0.0673620105,-0.1766556799,0.2594760358,0.0293518454,0.1861758083,0.1324825734,0.0129546858,-0.2061888278,0.4163241684,0.1540952027,0.3697729707,-0.1274185032,0.2786766291,0.6697968841,0.1807959676,-0.245277822,0.46640113,0.1738059223,-0.145856306,0.0374553613,0.1380504817,0.2496147156,-0.004768054,-0.3298011422,0.3833307028,0.0867591128,0.0610054322,-0.4341252148,0.2399909198,-0.0220644623,-0.1516919583,0.0082952846,-0.0084463805,-0.020532405,0.5849095583,-0.125159502,0.1440440714,-0.2933039963,0.3090782762,0.3011045754,-0.2954248786,0.0265338086,0.1752710193,0.0515204929,0.0268825777,0.3094967604,0.1028986722,0.1786962599,0.0438373871,0.0155035201,0.2277454883,0.4274324179,0.1050627157,0.5406873226,-0.2764897048,-0.0182581078,-0.5582947731,0.2105617225,-0.1665101498,-0.0254234467,0.2232838124,0.0506908298,0.0633384585,0.292270422,-0.0487163849,0.1978930235,0.1144294441,0.094570376,-0.2942007482,-0.2003487945,0.1136048064,0.0294003189,0.0777235553,-0.2208538949,0.2132159323,0.0908750221,-0.1085877344,-0.3333635032,-0.1862949282,0.1896972209,0.2165344656,0.3387879133,0.2866787314,0.3522740602,0.1767647117,0.0173678976,-0.0480225943,-0.0804897696,-0.1139160022,-0.1610632688,-0.0096825734,0.0530630536,-0.3831383884,0.0776906312,-0.3542157114,0.0091756945,-0.1333403289,0.1661383361,-0.0822231546,-0.1855175495,-0.1012335494,0.154444471,-0.098761335,0.6651298404,-0.2000395209,-0.0518577583,-0.400832355,-0.0116237728,0.3967906833,-0.4113268554,-0.11824058,0.0353530869,0.1134374663,0.1772118807,-0.1325220615,-0.4624225795,-0.0589967035,0.4347944558,-0.284042716,-0.2219363898,0.0567813888,-0.040338736,0.0548540987,0.0611052476,0.3005228043,-0.0701374039,-0.033817973,0.0889005065,-0.0471071042]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"this also happened to me with `nlp.load_dataset('glue', 'mnli')`","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":8,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\nthis also happened to me with `nlp.load_dataset('glue', 'mnli')`","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"Yes, this is because the test sets for glue are hidden so the labels are\nnot publicly available. You can read the glue paper for more details.\n\nOn Sat, 6 Jun 2020 at 18:16, Jack Morris <notifications@github.com> wrote:\n\n> this also happened to me with nlp.load_datasets('glue', 'mnli')\n>\n> \u2014\n> You are receiving this because you are subscribed to this thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/nlp\/issues\/245#issuecomment-640083980>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ABYDIHMVQD2EDX2HTZUXG5DRVJTWRANCNFSM4NVG3AKQ>\n> .\n>\n","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":82,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\nYes, this is because the test sets for glue are hidden so the labels are\nnot publicly available. You can read the glue paper for more details.\n\nOn Sat, 6 Jun 2020 at 18:16, Jack Morris <notifications@github.com> wrote:\n\n> this also happened to me with nlp.load_datasets('glue', 'mnli')\n>\n> \u2014\n> You are receiving this because you are subscribed to this thread.\n> Reply to this email directly, view it on GitHub\n> <https:\/\/github.com\/huggingface\/nlp\/issues\/245#issuecomment-640083980>,\n> or unsubscribe\n> <https:\/\/github.com\/notifications\/unsubscribe-auth\/ABYDIHMVQD2EDX2HTZUXG5DRVJTWRANCNFSM4NVG3AKQ>\n> .\n>\n","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"Thanks @thomwolf!","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":2,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\nThanks @thomwolf!","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"@thomwolf shouldn't this be visible in the .info and\/or in the .features?","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":12,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\n@thomwolf shouldn't this be visible in the .info and\/or in the .features?","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"It should be in the datasets card (the README.md and on the hub) in my opinion. What do you think @yjernite?","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":21,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\nIt should be in the datasets card (the README.md and on the hub) in my opinion. What do you think @yjernite?","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"I checked both before I got to looking at issues, so that would be fine as well.\r\n\r\nSome additional thoughts on this: Is there a specific reason why the \"test\" split even has a \"label\" column if it isn't tagged. Shouldn't there just not be any. Seems more transparent","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":49,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\nI checked both before I got to looking at issues, so that would be fine as well.\r\n\r\nSome additional thoughts on this: Is there a specific reason why the \"test\" split even has a \"label\" column if it isn't tagged. Shouldn't there just not be any. Seems more transparent","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"I'm a little confused with the data size.\r\n`sst2` dataset is referenced to `Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank` and the link of the dataset in the paper is https:\/\/nlp.stanford.edu\/sentiment\/index.html which is often shown in GLUE\/SST2 reference.\r\nFrom the original data, the standard train\/dev\/test splits split is 6920\/872\/1821 for binary classification. \r\nWhy in GLUE\/SST2 the train\/dev\/test split is 67,349\/872\/1,821 ? \r\n\r\n","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":64,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\nI'm a little confused with the data size.\r\n`sst2` dataset is referenced to `Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank` and the link of the dataset in the paper is https:\/\/nlp.stanford.edu\/sentiment\/index.html which is often shown in GLUE\/SST2 reference.\r\nFrom the original data, the standard train\/dev\/test splits split is 6920\/872\/1821 for binary classification. \r\nWhy in GLUE\/SST2 the train\/dev\/test split is 67,349\/872\/1,821 ? \r\n\r\n","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"> I'm a little confused with the data size.\r\n> `sst2` dataset is referenced to `Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank` and the link of the dataset in the paper is https:\/\/nlp.stanford.edu\/sentiment\/index.html which is often shown in GLUE\/SST2 reference.\r\n> From the original data, the standard train\/dev\/test splits split is 6920\/872\/1821 for binary classification.\r\n> Why in GLUE\/SST2 the train\/dev\/test split is 67,349\/872\/1,821 ?\r\n\r\nHave you figured out this problem? AFAIK, the original sst-2 dataset is totally different from the GLUE\/sst-2. Do you think so?","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":89,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\n> I'm a little confused with the data size.\r\n> `sst2` dataset is referenced to `Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank` and the link of the dataset in the paper is https:\/\/nlp.stanford.edu\/sentiment\/index.html which is often shown in GLUE\/SST2 reference.\r\n> From the original data, the standard train\/dev\/test splits split is 6920\/872\/1821 for binary classification.\r\n> Why in GLUE\/SST2 the train\/dev\/test split is 67,349\/872\/1,821 ?\r\n\r\nHave you figured out this problem? AFAIK, the original sst-2 dataset is totally different from the GLUE\/sst-2. Do you think so?","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/245","title":"SST-2 test labels are all -1","comments":"@yc1999 Sorry, I didn't solve this conflict. In the end, I just use a local data file provided by the previous work I followed(for consistent comparison), not use `datasets` package.\r\n\r\nRelated information: https:\/\/github.com\/thunlp\/OpenAttack\/issues\/146#issuecomment-766323571","body":"I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```","comment_length":33,"text":"SST-2 test labels are all -1\nI'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1.\r\n```\r\n>>> import nlp\r\n>>> glue = nlp.load_dataset('glue', 'sst2')\r\n>>> glue\r\n{'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 872), 'test': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 1821)}\r\n>>> list(l['label'] for l in glue['test'])\r\n[-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1]\r\n```\n@yc1999 Sorry, I didn't solve this conflict. In the end, I just use a local data file provided by the previous work I followed(for consistent comparison), not use `datasets` package.\r\n\r\nRelated information: https:\/\/github.com\/thunlp\/OpenAttack\/issues\/146#issuecomment-766323571","embeddings":[0.3236779869,-0.3612535298,-0.0937504917,0.1778616607,-0.0011129238,0.1007492021,0.3361848891,0.449954927,0.575297296,0.0148675088,-0.1256535202,0.1346777976,-0.1473744661,0.2773214579,0.3333793879,-0.0530476905,0.1083805263,0.3197503388,-0.1676009595,-0.2568297088,-0.0326385088,0.382578373,-0.1446127892,0.1994652748,-0.2805328667,0.2485164255,0.0017220888,-0.0963404924,-0.0316199176,-0.5869413018,0.2044335455,0.2065011263,0.2146628797,0.041023083,-0.0001139119,0.0438769571,0.199644655,-0.1350758076,-0.2542124689,-0.1216584072,-0.1056764275,-0.202876538,0.0458618514,-0.1059137732,-0.3367511034,0.158604309,0.0468212739,0.0070835771,0.1640624404,0.1519432068,0.1746294796,-0.0413381346,0.0038401268,0.0002421326,0.098652482,-0.3579421639,0.0070241201,0.0418982767,-0.0449892581,-0.1923692524,0.2850547135,0.4425401986,0.1635231972,0.3538922369,0.0891574696,-0.1086178869,0.2901045084,-0.2414478511,0.0570447817,0.1071272865,0.1578318179,-0.157120496,-0.2407914549,-0.116439715,-0.0565617196,0.1006575152,-0.0059571643,0.0276229605,0.2176365852,0.2936665714,-0.0193677172,-0.0927764326,-0.1666014344,0.0875245631,0.0444382131,0.5951553583,0.0525575913,0.1323291361,0.3667913377,0.0886464491,-0.4231613576,0.289028585,-0.2783505917,0.1741924286,-0.5723791718,-0.1834564209,0.1960399151,-0.002869542,-0.2150663882,0.1016736701,0.025422398,0.0275838878,0.3623879552,-0.0517081767,-0.0184163898,0.3278754354,0.5321333408,0.3578608632,0.0279294159,-0.1615818143,0.0433757678,0.1671692431,-0.3254519105,-0.2536082268,0.055789616,0.2035899758,0.0716067106,-0.3534903228,-0.3213402331,-0.0689652935,-0.3975773156,0.1182760447,0.1045105457,0.0138104791,0.2748116255,0.0285410937,-0.1252366751,-0.005990752,-0.1177942976,-0.3868844509,-0.2563748062,0.1317728013,-0.1777149588,0.0184122026,0.4044367671,0.1601787508,0.1664631814,-0.2921445668,-0.3622271121,-0.0422817767,0.1334693283,-0.0434931256,0.1852776706,0.3485696316,-0.2840801477,0.3784767985,0.2428891361,-0.5122995973,-0.1410859227,-0.2726314664,-0.0292389225,-0.0726896748,0.0673725232,0.1427009851,-0.3055776358,-0.0654691905,-0.0006719767,0.3467418253,-0.0410009809,0.1258835644,0.0487167463,-0.2369191796,0.0019332132,0.2537924349,0.2284991294,0.2446567267,-0.4592027366,-0.097838752,0.3145676553,-0.0179248098,-0.1690036952,0.1724620908,0.2161767483,0.1493662596,-0.0602655187,0.2051674575,0.3542272151,-0.3027257919,-0.3342137635,0.1045898274,-0.2881499231,-0.4994359016,0.032340914,0.0609023944,-0.0006794172,0.1040444374,-0.0421709493,-0.3407530189,0.1207541898,0.0356192514,-0.5237576365,0.2544219494,0.5446698666,-0.016373327,-0.025428297,-0.1288862079,-0.3430946171,0.1978552639,0.1403770298,0.0544712953,0.0367121734,0.1544160545,0.1745421588,-0.1268416196,0.0858040377,-0.0954349041,-0.1772256047,0.1945025176,0.0130313449,0.393972367,0.5494517088,0.0648840144,-0.3999761641,-0.2042219192,0.0357141532,-0.2271699905,0.1765029281,0.4278755784,0.3945477307,-0.0470248908,0.0751001462,-0.0110270521,0.2815508842,0.0096159391,-0.2109095901,0.2346245795,0.0818820819,-0.2415735871,0.078686811,0.2926479578,0.2893486917,-0.0917757004,0.0322616473,0.22111696,-0.201456666,-0.0546025857,0.0886546671,0.2956277728,-0.0888027027,-0.4509783089,-0.2570939958,0.5002954602,-0.0395481475,-0.0208692644,-0.075032942,0.0755238459,-0.0529433861,0.1584235281,-0.1140918955,0.1547783911,-0.0377261303,-0.0591299087,-0.4671587944,-0.0529296584,0.224119246,-0.4385807812,-0.0076648807,0.1317540407,-0.0462878197,0.317730397,0.2281946093,0.1615154892,0.0979470983,-0.2021484226,-0.2077597827,-0.0914531872,0.2668095827,0.2907688022,-0.0192162395,0.1956549287,0.0562129952,-0.2310011387,-0.5293570757,-0.0398037657,0.0130055901,-0.0980209783,-0.2299815416,-0.1332415789,0.3371716142,-0.3350443542,-0.1945854574,0.2070712298,-0.0692047477,0.3786220849,-0.6263083816,-0.0622233488,-0.1187708005,-0.1242091954,-0.2760334909,-0.0520232283,0.0103210332,-0.2586920559,0.3558694422,0.0493571386,0.0848145783,0.0825465173,-0.0007399836,0.1618949324,-0.149728626,0.2043101043,-0.3001515269,-0.1242339164,-0.2505884767,-0.0001663062,-0.2251261473,0.3416076303,0.0314049311,-0.1615663469,-0.4707098305,0.2010546178,-0.3075324595,-0.0509093888,-0.311673373,0.1930027157,0.2237749994,0.1517865509,-0.2070730627,0.1381038427,0.104045935,-0.1074146777,-0.1088032797,0.1196152195,0.0314826034,0.0655656606,-0.4639659822,-0.4390846491,-0.0139324414,0.0409170836,0.40105474,0.0659215972,-0.1223512739,0.0496974736,-0.2470991015,-0.1264885962,-0.1755560488,0.1663680524,-0.1875957698,-0.2554449737,0.2479190975,0.1119256765,-0.3266984522,-0.2127224356,0.0801267102,0.0762675032,-0.7130568027,-0.2000487894,-0.1134926751,-0.1474514008,-0.0590478256,0.0865231156,-0.1103315428,0.1984110028,0.1098318845,-0.02925843,-0.1576776057,-0.3824420273,0.6548931599,0.3079114258,0.0966887325,-0.4028122723,0.0745698959,0.051984299,0.1751458198,0.1673040241,-0.1668654531,-0.0060525783,-0.1098987684,-0.0408829637,0.1531369984,-0.3298499584,0.4263611734,0.2392539829,0.0783155113,0.2785067558,0.0004221499,0.0481248833,-0.0292709637,0.4904929996,-0.1470145136,0.0451524295,-0.2398291528,-0.1143433303,-0.1532816887,0.0303001497,0.0106283017,-0.2077560872,-0.1280959398,0.1492244154,-0.1617190838,-0.0372586921,0.275290221,-0.732552588,-0.0751448125,-0.0944725201,-0.0284636915,0.0452844575,0.4404402971,-0.108198747,-0.118861258,0.1701602787,-0.0082021086,0.6488139033,0.0617252067,0.4940756559,0.3102582097,0.1985439509,-0.3425166011,-0.3801377118,-0.4393658638,-0.1725531518,0.5286617875,0.1560548395,-0.2542396188,-0.2872360349,0.2137312889,-0.1023274213,-0.1947612166,0.0963857174,-0.4203895628,-0.2383720875,0.3322480619,-0.0373442955,0.1644622684,0.0595598817,-0.0016732523,0.0782699734,0.2368905842,-0.4608204067,0.4027179182,0.3578073084,0.1285772473,0.3503626883,0.0243714526,0.3749052584,0.1176716834,-0.464939177,0.2364984155,-0.1561971456,-0.0461930558,0.1719627976,-0.2602113187,0.341686666,-0.0659903735,-0.1073216945,-0.0489874035,-0.3839330673,0.2862161696,-0.3703196347,0.3019980788,0.3484676778,0.2424851358,0.0201282818,-0.1555307508,0.122219786,-0.1180984378,-0.2628390491,-0.2113470733,-0.0161536876,-0.4556924999,0.0849199072,-0.2104105055,0.644818604,0.0453515425,0.0006920958,0.1022992656,-0.3434745669,0.5957657695,-0.1162126809,0.13951765,-0.2881869972,-0.0822033212,0.0693964511,0.06554351,-0.1502910256,0.4253515303,-0.1543275863,0.4221574962,0.1851944029,0.0072499495,0.0367451422,0.1472464651,0.5044356585,0.1358224899,-0.2058896422,0.180686757,-0.227803126,0.1381678581,-0.1022923514,-0.0200239122,0.1115603149,-0.2711080313,-0.2080396861,-0.1458145082,-0.3024890423,-0.0870206729,0.4240999818,0.1037028208,-0.1666999906,0.0368835442,0.5649836659,0.2304651141,-0.2195448279,-0.0636169612,0.1668859273,0.3394561112,0.0962954462,0.2072866261,0.3198139668,0.0012471343,-0.2308737338,-0.0166411884,0.0801634118,0.1738085151,-0.3481507599,-0.1017419025,-0.0468938909,-0.1230679974,0.2246161252,0.2875225842,0.1345015466,-0.277755022,0.1477760077,0.042110499,-0.4823461473,0.0880549699,0.1862150431,-0.4466958046,0.1284505874,0.4110448658,-0.3931851685,0.3199150562,0.2890365422,-0.0767018721,-0.0086688176,-0.2683332264,0.0829298645,0.0207927041,-0.500256598,0.1510371715,0.0682421476,-0.312712878,-0.0617997646,0.3585748971,0.3742308915,-0.0457244702,-0.0459201038,0.0495658033,-0.126606226,-0.0315359011,-0.0525507554,0.2949239612,-0.0991967544,0.2035393417,-0.0005553397,0.2477180064,-0.3138470948,-0.1842354536,-0.3627864718,0.1797088981,-0.007426348,-0.0080497228,-0.2206862271,-0.0802493095,0.2063737214,-0.2019584179,-0.0771369413,-0.2378775775,0.0490760021,0.1097868159,-0.0619535595,0.0475424528,-0.2424401045,0.0774611533,0.0400333218,0.0109387664,-0.0615744963,0.0670595393,0.0970124602,0.8924075961,0.1089359,-0.0345239453,0.1264148206,-0.0472053476,-0.0908409506,-0.1398901343,-0.0353140831,0.220516488,0.2597257495,0.0708756149,-0.2479421049,0.1092262566,-0.4701394439,0.0670598075,-0.1003386676,-0.3349014223,0.1707273126,-0.1102902517,0.2450977266,-0.0020975294,-0.0565025732,-0.07533703,0.1967256963,0.2744044065,-0.1440592408,0.2991374433,0.3024626672,0.1285304278,0.206375733,0.2766053379,-0.0861407965,-0.0918454975,0.0446168147,0.1708595157,0.253320843,-0.1556815356,0.0148341106,0.2354191244,-0.3778322637,-0.1394178122,0.2338889837,-0.5454513431,-0.2599850595,-0.0635323301,0.3590611815,0.4730839431,0.2442908883,0.2248538435,-0.1174967885,0.5504229665,-0.1005650312,0.1148993075,-0.1903004348,0.1568087935,-0.0269046891,0.4228307903,-0.2254350334,-0.289681673,-0.1428186595,0.074194558,0.0123419613,-0.4227356315,-0.3647421896,0.0775581896,-0.2778157592,0.0937697068,0.0272423774,-0.0689352378,0.4262211919,-0.0156309623,0.029314233,-0.362406075,-0.6591207385,0.082730636,0.2454971224,-0.03081958,0.3564572632,0.0135354018,-0.0234169289,0.0826943517,0.4850860238,-0.0217086691,-0.1539367139,-0.1377496272,-0.2947840691,-0.1990977079,-0.1798348874,-0.1997739375,0.1143070161,-0.0689531788,-0.2047249079,0.2256322354,0.1332937479,-0.0338684805,0.4815896153,-0.3065578938,-0.1352907419,-0.1985166967,0.4735826552,0.0533327349,-0.0843383968,-0.0419882238,0.0042119497,-0.2606084645,-0.1726108342,0.010212604,-0.0690060481,0.2692813277,-0.3107885122,0.0877802894,0.0872605816,0.4814496934,-0.2841915786,0.1334250122,-0.1504841745,0.15816769,-0.1848749518,-0.1463611424,-0.0363074131,-0.1639991403,-0.1206063777,-0.0589600652,0.1612282991,-0.0776259229,0.0503929555,0.3595695496,-0.1182397008,-0.0655155182,-0.3488412499,0.3465177119,0.0619244948,0.3626910746,0.2391082942,-0.1259867549,0.1356225312,0.1125827581,0.0826555192,0.1062327474,0.1661250591,-0.0963607877,0.2769807875,0.1993840933,0.0897277519,0.4974392653,0.0567635298,0.1737076342,-0.2943834662,-0.3370407522,-0.3608500361,-0.1045531407,-0.2991927266,0.4071636498,-0.0568978973,0.0057604364,0.1097955108,-0.1208022684,0.1738740802,0.3284982741,-0.2332250178,0.1424823254,-0.0696453378,-0.0213694833,-0.0113614332,-0.1544405073,0.1508358419,-0.0689016432,-0.1547403038,-0.5150067806,0.6079291701,-0.1748900712,-0.129706189,-0.0743925571,0.3857112527,0.3296773136,-0.1316089183,-0.5880601406,-0.2272664458,0.4960272014,0.1941723078,-0.4108332396,0.3343149722,0.1390274763,0.2144568712,0.0944270715,-0.0870755985,0.3029446304,0.0461426415,-0.0372201763,-0.2824232876]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/242","title":"UnicodeDecodeError when downloading GLUE-MNLI","comments":"It should be good now, thanks for noticing and fixing it ! I would say that it was because you are on windows but not 100% sure","body":"When I run\r\n```python\r\ndataset = nlp.load_dataset('glue', 'mnli')\r\n```\r\nI get an encoding error (could it be because I'm using Windows?) :\r\n```python\r\n# Lots of error log lines later...\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\tqdm\\std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\glue\\5256cc2368cf84497abef1f1a5f66648522d5854b225162148cb8fc78a5a91cc\\glue.py in _generate_examples(self, data_file, split, mrpc_files)\r\n 529 \r\n--> 530 for n, row in enumerate(reader):\r\n 531 if is_cola_non_test:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\csv.py in __next__(self)\r\n 110 self.fieldnames\r\n--> 111 row = next(self.reader)\r\n 112 self.line_num = self.reader.line_num\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\encodings\\cp1252.py in decode(self, input, final)\r\n 22 def decode(self, input, final=False):\r\n---> 23 return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\n 24 \r\n\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 6744: character maps to <undefined>\r\n```\r\nAnyway this can be solved by specifying to decode in UTF when reading the csv file. I am proposing a PR if that's okay.","comment_length":27,"text":"UnicodeDecodeError when downloading GLUE-MNLI\nWhen I run\r\n```python\r\ndataset = nlp.load_dataset('glue', 'mnli')\r\n```\r\nI get an encoding error (could it be because I'm using Windows?) :\r\n```python\r\n# Lots of error log lines later...\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\tqdm\\std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\glue\\5256cc2368cf84497abef1f1a5f66648522d5854b225162148cb8fc78a5a91cc\\glue.py in _generate_examples(self, data_file, split, mrpc_files)\r\n 529 \r\n--> 530 for n, row in enumerate(reader):\r\n 531 if is_cola_non_test:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\csv.py in __next__(self)\r\n 110 self.fieldnames\r\n--> 111 row = next(self.reader)\r\n 112 self.line_num = self.reader.line_num\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\encodings\\cp1252.py in decode(self, input, final)\r\n 22 def decode(self, input, final=False):\r\n---> 23 return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\n 24 \r\n\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 6744: character maps to <undefined>\r\n```\r\nAnyway this can be solved by specifying to decode in UTF when reading the csv file. I am proposing a PR if that's okay.\nIt should be good now, thanks for noticing and fixing it ! I would say that it was because you are on windows but not 100% sure","embeddings":[0.0960785672,-0.0313783884,-0.0775830969,0.2640608847,0.4257141054,-0.0790004805,-0.0854370892,0.1514505595,0.085815832,0.1757878512,0.1594255567,0.113467671,-0.0249078646,-0.1914832145,0.1588652879,-0.2088525146,0.1939257681,0.1727608442,-0.1549511552,-0.015184924,-0.286002934,0.3500201702,0.0244884845,0.1565985084,-0.1144619659,-0.1852870286,-0.2307462692,0.2669430673,-0.3331380486,-0.5159603357,-0.0930942968,-0.1411760449,0.4058378339,-0.3424748182,-0.0001093308,-0.0721149892,0.3982744813,-0.1664467156,0.0547835156,0.1166156083,-0.0154737113,-0.3343654871,-0.1760163158,-0.2486954927,0.3781097233,0.3489088416,0.0731906444,-0.3212687671,0.0242051724,0.2659868002,0.2378252745,0.2477606535,0.1781280041,0.2954207361,0.244182229,-0.3906762898,-0.028495172,0.349504441,0.2638509572,-0.2385381609,0.1345652938,0.2473401576,-0.2612062097,0.1423325688,-0.2731612921,0.3587567508,0.1623549014,-0.5745044947,-0.0732909441,0.1375042945,-0.0825233459,-0.2108460367,-0.2324727029,-0.0214668028,0.3149498701,-0.3452878892,0.4538139701,0.3430697322,-0.139133513,0.202895388,0.2395725399,-0.005509668,0.1321135163,0.2591429055,0.0190588515,0.1714113504,-0.1949968934,-0.0338637345,0.315877974,-0.2977007031,-0.2757945359,-0.1427662075,-0.1771564931,0.0720318705,-0.0356128886,-0.0812036917,-0.2842743397,0.1435817331,0.0846680403,-0.0134624029,0.0946583226,-0.0573431253,0.0899615809,0.110853143,0.0208287444,0.2417101115,-0.14708215,-0.0004220317,0.4232667387,0.1088606417,0.3318261206,0.0181730501,-0.1834580898,-0.2916094959,-0.164936915,0.27084288,0.2871673703,-0.0648045614,-0.3927219212,0.0809699893,-0.1012425944,0.0655919239,-0.2223991454,0.0761764199,-0.180843696,0.2043089569,-0.1888338178,0.3469958007,-0.2571710944,-0.3519371152,0.0149911698,0.3389755189,-0.1232428849,-0.0285522565,0.1955817938,-0.0283842776,-0.0404230729,-0.0602286719,-0.158913672,0.0255914442,-0.2434097826,-0.0801455379,0.0578858331,0.2219349742,0.1136653051,0.1883814633,0.0987239555,-0.6630222201,-0.2351446301,0.0586407557,-0.1415205002,0.0448225476,-0.1568872333,0.15356417,-0.2686422169,-0.198261261,-0.0049656536,-0.05214829,0.0420172103,0.1156464741,0.0802602321,-0.0318687819,0.0376031138,-0.3339197934,0.1539804339,0.4237532318,-0.3073547184,0.0881555602,0.0343531072,-0.2626703382,0.2700485289,0.3727375567,0.2950722575,-0.1970290095,-0.271065414,0.5628251433,0.2969708741,-0.4510250688,-0.5493014455,0.5310085416,-0.1751743257,-0.0164244361,0.0871717334,0.1576889157,0.2352524251,0.1092154086,0.1837364733,0.5050608516,-0.0801191106,0.0982305631,-0.3659400642,-0.4618396163,0.3761814237,0.0676294491,-0.2985517979,-0.0729923025,0.1556464732,-0.440420419,0.2256357372,0.0287156813,0.2387967557,0.1448225826,0.2214071304,-0.1668151617,-0.0042063789,-0.1305445582,-0.1685751081,0.0079936199,-0.1055179387,0.2592512369,-0.6571520567,-0.2187514603,-0.0467980877,-0.1435094476,-0.2525244653,-0.1095065027,0.1855923086,0.0744087398,0.3862223625,0.237325117,-0.1318559796,0.5412459373,0.0361153558,0.1005556583,-0.2505453825,0.3868020773,-0.137332052,-0.2259554416,0.0624194108,0.3521325588,0.1498402655,-0.1148921996,-0.0219613351,0.2945459187,-0.1449641883,0.0458772704,0.1562280357,-0.0929330066,0.351041615,-0.3313707709,0.187915206,0.2839381695,0.2087194622,0.0116677396,0.405600071,0.2408409268,0.0773420483,0.2359721959,0.1807653159,0.2210743576,0.2278275788,0.0244149826,-0.2062000036,-0.461948365,0.2684993148,0.0331732891,0.1258951575,-0.0679659173,-0.2384745926,-0.0589276738,0.6475167871,-0.1361549646,0.136540994,-0.0186666474,-0.3090726733,0.3296530843,-0.1149338558,0.2858798206,0.4102845192,-0.0284524001,-0.1026120111,0.2041167319,0.0772209838,-0.1080849543,0.0749537051,-0.0833185986,-0.1447232217,-0.0337691084,-0.2238414437,0.0857315809,-0.2310127765,0.0127826612,-0.0808480307,0.2251496166,-0.1622379422,-0.178805843,-0.5618280172,-0.470752269,-0.2029812634,-0.2117941827,-0.5013468862,-0.2158409506,-0.1323920339,-0.2413420379,-0.0234204177,-0.0338389911,-0.1416575909,-0.1357844919,-0.0820841268,-0.1961432248,0.1037638709,-0.1461428553,-0.4340043962,0.1121312454,0.3220778704,0.1211995333,0.1814731956,-0.3321772218,0.3115805387,-0.1961656213,-0.0553119592,0.103661865,-0.1345860809,0.2402223796,0.0963666961,-0.012691414,0.2656985223,-0.259136796,0.2570508718,-0.0776764452,-0.1377224177,0.1200319603,0.1895545572,-0.4034391344,-0.0965973213,-0.1397151649,-0.7030835748,-0.2230282128,0.0615337193,-0.1293295175,0.013375177,-0.1952487528,-0.2366751432,0.1121903434,-0.145375371,-0.0304585416,-0.1451872438,-0.0018216852,0.5186716914,-0.1627871692,-0.4522241354,0.1555846632,-0.1912770271,-0.0331651457,-0.1888067871,-0.3169983625,0.0377703048,-0.4247045517,-0.0206879899,0.1168887541,0.1911672652,0.1616137326,-0.1768027991,0.1168118045,0.1030664369,-0.1106764898,-0.2166063488,0.2738518715,0.2781443298,0.1500487328,0.1878759861,0.1356651634,0.0090656755,0.1934974343,0.0491352715,0.5423480272,-0.2135621905,0.3014684319,0.3107925951,-0.0927606001,0.006523489,-0.1977573335,0.2089509219,0.1855590492,0.0166288987,-0.2993296385,-0.1678502113,0.2693977654,-0.3353396654,-0.3381195366,0.0509142391,-0.0387566686,-0.0562567711,0.0822191164,-0.0550042205,0.1378330439,-0.2853698134,0.1453665197,0.4659432471,-0.2176744789,0.1647081226,-0.142332688,-0.6985322833,-0.3042908907,0.1565534323,0.3822276592,0.6439914703,-0.0699912757,-0.0151534351,0.1850119233,-0.1884350479,0.4023388326,-0.4889247715,-0.1586768478,0.2194533795,0.4156270921,0.1141736284,-0.1248085052,-0.4630333185,-0.089999713,0.3163911998,0.1048720255,-0.0107841985,0.0117738005,0.1132075414,0.1722649932,-0.0974453613,-0.1195115224,-0.4737176299,-0.4600507617,-0.2901911736,0.0244040675,0.0629184172,0.3556575477,0.2049372196,0.2504357398,-0.0193112064,-0.1313568205,0.027361678,0.1637227237,0.1467247456,-0.0384015441,0.0112718381,-0.1745097786,0.2139522433,-0.1999609321,0.4932182729,0.1238479242,-0.1025703922,0.0527386293,-0.4481794834,0.4238167107,0.1645700186,-0.0512149818,-0.0725796074,0.0644961149,0.1399890184,-0.2324614078,-0.071623005,0.4307657778,0.0813888237,-0.2889288068,0.055474434,0.3026761711,0.1856369823,0.2076051533,-0.0527171418,0.1698961556,-0.0922633782,0.0060459459,-0.0335629918,0.8596192598,0.1393060535,0.0482402183,0.0805299804,-0.4887138307,0.4115574062,-0.072821565,0.0782398432,-0.2332847863,-0.269282043,-0.0748383924,0.2021314502,0.5368649364,0.0178877246,-0.4103850424,0.2123206407,0.2594627738,0.1733084023,0.2273137271,0.4100671709,-0.4128699303,-0.3461407423,-0.2607558668,0.1467491239,-0.0509291813,0.1575378925,-0.0025416685,-0.1808653474,0.1510878801,-0.1643083543,-0.3796912134,0.1481798589,-0.5830651522,-0.1059094891,-0.1159503907,-0.0505560488,0.0276000295,-0.0325349383,0.4301761091,0.0788147375,0.1462220252,0.0287767183,0.5023577809,-0.1347955614,0.2634490132,0.0649158433,0.0293695088,-0.1692399979,-0.2536670864,0.0503207333,0.1035284549,0.0931726247,0.0457430482,0.0309722815,-0.1655649394,-0.0099989334,-0.1781852394,-0.0597790107,0.1339139193,0.1969260573,0.1373153329,0.053902775,-0.1622062176,0.2158534676,-0.050471887,-0.2069851458,-0.1149858385,0.0962894186,-0.0865576565,-0.338765502,-0.0000448124,-0.0895199999,-0.3144166172,-0.1773141176,-0.0498673283,-0.2948665321,-0.1095600426,-0.2490810901,0.0947751477,-0.0037871096,-0.1460085213,0.0099655688,0.2050777078,-0.0175672267,0.0567150675,-0.6207973361,-0.2790296078,0.2701216638,0.11701902,-0.0511030592,0.0250979364,0.0435589291,0.1252604276,-0.2193239033,-0.2847450376,0.104562141,-0.0480232202,0.0091423467,0.5306373835,-0.0797183439,0.145195961,0.1064324901,0.1281823516,0.3685395718,-0.0100181457,-0.1860472411,-0.0728402436,0.0723790228,0.0700487643,-0.262648195,0.2065203041,-0.1436946094,0.0090934308,-0.0019019386,0.1096202508,-0.0420618951,-0.0786515325,0.2049996704,-0.124425821,0.1131497622,-0.0192081872,0.2128065825,-0.3043265939,0.5440462828,0.1988140494,0.1885215044,-0.1173685789,0.1726166159,0.1218018532,-0.1342391819,-0.300775528,0.0201400556,0.1969583482,-0.3434199989,0.0676229149,0.00423367,0.1457539946,0.4197669923,-0.1385031193,-0.1265950501,0.57941401,0.2061559558,-0.1375902742,0.1366768479,0.2641611397,-0.0864175409,0.0020647803,0.2470193356,0.2873879075,0.1410555691,0.2585171461,-0.0754051208,0.610485673,-0.0110168839,-0.0121187987,0.1957356781,0.0342017636,0.139329657,-0.0139013855,0.0641347542,-0.0379219763,0.2962032557,-0.3129601777,0.0142601253,0.3813517988,0.2522052526,-0.1619686335,-0.1429505199,0.0050678048,0.0638159439,-0.0613718368,0.3057824671,0.211539045,0.4559714496,-0.143042624,-0.1331113875,-0.0036832655,0.2221008241,-0.0244739689,0.03299357,0.0995862707,0.1223705038,-0.5756689906,0.2016313523,0.0297463965,-0.1971906275,0.2701058686,0.228193894,-0.0745060667,-0.3943826258,-0.199199006,0.0198579822,0.1872206479,-0.3332831562,0.2501318455,0.1700826436,0.1409678459,-0.0207064655,-0.0071998779,0.538606286,0.41171363,-0.4177429378,0.067883119,-0.2541178167,0.001823434,0.0017785013,0.1928388923,0.3674044311,0.3436054289,0.1835330278,0.1475745738,-0.0970902592,-0.1252169758,0.0820958987,0.0658996701,0.1669122577,0.2621804178,-0.1237127334,0.2254276276,-0.446105659,-0.1413818449,-0.6944621801,-0.0078035197,0.2011968791,-0.1833673716,0.2014449984,-0.1061415151,0.082534343,0.0894667283,0.2843176126,0.4724918604,0.1624284834,-0.1431423277,-0.1814238578,-0.773149848,0.1070468575,-0.2013553083,-0.3078991175,-0.1461916119,0.068603754,-0.1706115156,0.3500737846,0.0615050644,0.7089384794,-0.0908861756,0.0347727463,-0.1442360878,-0.0400933176,0.0791048184,-0.0747568682,-0.0370700136,-0.3391436934,0.1330929995,-0.065197587,0.0161507037,-0.1079834774,-0.1541165113,-0.3509855866,0.0500363447,0.2432803363,0.3958930075,0.6262511611,-0.1579274386,-0.0846172944,-0.3797359765,-0.0811664686,-0.1952175945,0.2209117413,-0.0506156571,-0.0309301168,-0.0845702887,0.1616847962,-0.5174289346,0.2082441896,0.0410947464,0.2399537414,-0.0459319353,0.0910194293,0.2638615966,0.0351101346,0.0531269498,0.1780718118,-0.0814524963,0.1501331925,0.0537901632,-0.3257542551,0.1363993883,-0.1620736867,-0.2424684018,0.1962384284,0.2875564396,0.2323967367,0.0941333398,-0.3354713321,-0.0860067457,0.0747777894,-0.1668247283,-0.2506377697,0.3078598082,-0.0706819147,-0.2163034976,-0.0388534479,0.1073542386,0.070081763,-0.2426321805,0.1463143378,-0.3746826649]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/242","title":"UnicodeDecodeError when downloading GLUE-MNLI","comments":"On Windows Python supports Unicode almost everywhere, but one of the notable exceptions is open() where it uses the locale encoding schema. So platform independent python scripts would always set the encoding='utf-8' in calls to open explicitly. \r\nIn the meantime: since Python 3.7 Windows users can set the default encoding for everything including open() to Unicode by setting this environment variable: set PYTHONUTF8=1 (details can be found in [PEP 540](https:\/\/www.python.org\/dev\/peps\/pep-0540\/))\r\n\r\nFor me this fixed the problem described by the OP.","body":"When I run\r\n```python\r\ndataset = nlp.load_dataset('glue', 'mnli')\r\n```\r\nI get an encoding error (could it be because I'm using Windows?) :\r\n```python\r\n# Lots of error log lines later...\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\tqdm\\std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\glue\\5256cc2368cf84497abef1f1a5f66648522d5854b225162148cb8fc78a5a91cc\\glue.py in _generate_examples(self, data_file, split, mrpc_files)\r\n 529 \r\n--> 530 for n, row in enumerate(reader):\r\n 531 if is_cola_non_test:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\csv.py in __next__(self)\r\n 110 self.fieldnames\r\n--> 111 row = next(self.reader)\r\n 112 self.line_num = self.reader.line_num\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\encodings\\cp1252.py in decode(self, input, final)\r\n 22 def decode(self, input, final=False):\r\n---> 23 return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\n 24 \r\n\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 6744: character maps to <undefined>\r\n```\r\nAnyway this can be solved by specifying to decode in UTF when reading the csv file. I am proposing a PR if that's okay.","comment_length":80,"text":"UnicodeDecodeError when downloading GLUE-MNLI\nWhen I run\r\n```python\r\ndataset = nlp.load_dataset('glue', 'mnli')\r\n```\r\nI get an encoding error (could it be because I'm using Windows?) :\r\n```python\r\n# Lots of error log lines later...\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\tqdm\\std.py in __iter__(self)\r\n 1128 try:\r\n-> 1129 for obj in iterable:\r\n 1130 yield obj\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\glue\\5256cc2368cf84497abef1f1a5f66648522d5854b225162148cb8fc78a5a91cc\\glue.py in _generate_examples(self, data_file, split, mrpc_files)\r\n 529 \r\n--> 530 for n, row in enumerate(reader):\r\n 531 if is_cola_non_test:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\csv.py in __next__(self)\r\n 110 self.fieldnames\r\n--> 111 row = next(self.reader)\r\n 112 self.line_num = self.reader.line_num\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\encodings\\cp1252.py in decode(self, input, final)\r\n 22 def decode(self, input, final=False):\r\n---> 23 return codecs.charmap_decode(input,self.errors,decoding_table)[0]\r\n 24 \r\n\r\nUnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 6744: character maps to <undefined>\r\n```\r\nAnyway this can be solved by specifying to decode in UTF when reading the csv file. I am proposing a PR if that's okay.\nOn Windows Python supports Unicode almost everywhere, but one of the notable exceptions is open() where it uses the locale encoding schema. So platform independent python scripts would always set the encoding='utf-8' in calls to open explicitly. \r\nIn the meantime: since Python 3.7 Windows users can set the default encoding for everything including open() to Unicode by setting this environment variable: set PYTHONUTF8=1 (details can be found in [PEP 540](https:\/\/www.python.org\/dev\/peps\/pep-0540\/))\r\n\r\nFor me this fixed the problem described by the OP.","embeddings":[0.0659664795,0.0692808926,-0.0075027859,0.1465992033,0.4890412092,-0.1362413168,-0.0057733776,0.0795608684,0.057453759,0.3163624704,0.1549845338,0.126315102,-0.0671637133,-0.1741780341,0.0777117684,-0.2971668839,0.1926772147,0.2051754296,-0.0492804386,-0.0339454375,-0.4367743731,0.3971391618,0.0737092793,0.2630495727,-0.0219343193,-0.0867550448,-0.152930513,0.2397748083,-0.3016114831,-0.57560426,-0.1237106919,-0.1266360134,0.3628610075,-0.2685573399,-0.0001157126,-0.012200146,0.3200551569,-0.2603190839,0.1376090944,0.1711819768,0.028887324,-0.294804424,-0.151617676,-0.177483201,0.3508855402,0.2826948166,0.1608464867,-0.4594718218,-0.1264068335,0.285561502,0.1487494558,0.3086079359,0.1480985582,0.4098854363,0.1059744209,-0.3754203618,-0.1038057655,0.3554778099,0.2898302078,-0.2861046493,-0.1215083003,0.1885137856,-0.2729696333,0.0802527964,-0.3073840737,0.3867275119,0.1543731093,-0.5894976854,-0.1191273183,0.0787643194,-0.1083671525,-0.2737489939,-0.3448261023,-0.1108141169,0.1624771506,-0.3434237242,0.3828711808,0.3813179433,-0.1983408779,0.3759185374,0.3234007657,0.0271237083,0.1297441721,0.1788926721,-0.1149421483,0.3030929863,-0.0484179333,0.0052201985,0.32939291,-0.1760675162,-0.2178891301,-0.0927065015,-0.028195003,-0.0783487111,-0.0077807754,-0.0275792498,-0.1730580926,0.0897668004,0.0650106892,0.0243771579,0.1349372417,0.0160542447,-0.0377602242,0.1903323233,-0.0071625346,0.1537086368,-0.1960662454,0.1105273664,0.5007560253,0.0054312591,0.3920165896,0.0856892392,-0.0916080102,-0.3336857855,-0.1481531411,0.3762517273,0.3734878898,0.0488297008,-0.3270634115,0.1245114505,-0.0932204127,0.0808945149,-0.268321991,-0.018830901,-0.1305695474,0.4093034267,-0.1884571761,0.3444765508,-0.412813127,-0.3085870445,0.0623199753,0.2654925287,-0.0576127358,-0.0216783267,0.1953626871,-0.0063774041,-0.0489238426,-0.0647775307,-0.0668369457,0.0914170519,-0.1626498997,-0.1058646441,0.1504982859,0.2132132053,0.1111661941,0.0766256154,0.1605852693,-0.7501699328,-0.3676734567,0.1320264041,-0.1136827841,-0.0327389948,-0.1624801457,0.0419723019,-0.2905904353,-0.268329531,0.0432097912,0.0893631652,-0.012850591,0.058380805,0.1423714906,-0.0658238605,-0.1256457716,-0.2885705233,0.1912132055,0.5120857358,-0.2714211643,0.11977835,0.2272893786,-0.2887923419,0.1852322519,0.3578078449,0.2416885197,-0.3323736787,-0.1923466474,0.4748018682,0.2483752966,-0.5571578145,-0.5454277396,0.4257736206,-0.1235029846,-0.0389658064,0.1891016364,-0.0144956205,0.1878460795,0.0291927364,0.1525562257,0.5387167931,-0.1974547654,0.1925505251,-0.2940203547,-0.4947428703,0.3141426444,0.1426318139,-0.1736194789,-0.0822926685,0.129146114,-0.4585199654,0.1723280698,-0.1112949252,0.2787634432,0.0282671954,0.3319406807,-0.211189881,0.0380765013,-0.0915697888,-0.0415454134,-0.01109784,-0.0237123705,0.2121428251,-0.69395715,-0.2628234029,0.0737558901,-0.1202406287,-0.2130515873,-0.0597773083,0.0705621764,0.1311570406,0.3312766552,0.3530176878,-0.2560486794,0.444412291,0.0291101765,0.0596346259,-0.1860978156,0.2995320857,-0.0332169123,-0.185224995,-0.0328014642,0.4689338803,0.0278450828,-0.1168617755,-0.0130393375,0.2994942367,0.0071417773,0.0874278173,0.099679254,0.0350923501,0.4462957084,-0.2842265069,0.139889732,0.4377090931,0.1405509412,0.1681509465,0.2921564579,0.4354314506,0.0450360291,0.140774563,0.2176133096,0.1677333117,0.2079701722,0.0885910019,-0.4142263234,-0.4824961424,0.1321980655,0.0846019536,0.2097947448,-0.0038142868,-0.3643803895,-0.1360848695,0.7756450176,-0.1477789134,0.1650189459,0.0773028359,-0.1030396968,0.343202889,-0.116912812,0.2165263444,0.4581326544,-0.0990952849,-0.0600243285,0.2631981075,0.0864502862,-0.0863333419,0.0601589531,-0.0021347655,-0.2038664222,0.0336265825,-0.1810392439,0.0602921583,-0.238687098,-0.0007922559,-0.0726954043,0.1344421059,-0.1566104144,0.018894786,-0.5958063006,-0.3180299103,-0.2905757427,-0.1569772214,-0.3728326559,-0.177833274,-0.08260452,-0.2205192745,-0.1535080373,-0.0098585598,-0.3806848228,-0.0583350398,-0.1645463705,-0.3551083207,0.287994653,-0.1519503444,-0.3765107691,0.0458634123,0.2971877456,-0.0745043233,0.0527410358,-0.3912444711,0.2198841721,-0.1161720082,-0.1598160565,0.1045536026,-0.0126124732,0.1350214183,0.0728747621,-0.2613962889,0.1342346221,-0.3289485276,0.2606318891,-0.1018458083,-0.0507454909,0.0579627044,0.1840224266,-0.3369160593,-0.0888148546,-0.0477336943,-0.6349591017,-0.2621896863,0.1085829437,-0.2164635956,-0.0296417847,-0.3215074539,-0.2497011125,-0.0321872681,-0.0824408755,-0.0988871232,-0.0484782159,-0.0408875011,0.4804697037,-0.2116928697,-0.3819882572,0.1253904253,-0.2614033222,-0.0949382335,-0.0084321955,-0.2596348822,-0.1206082031,-0.4150761962,0.0678995475,0.0816438198,0.2113174051,0.1888361722,-0.066246599,0.1889101863,0.0285259429,-0.2133290172,-0.2013958544,0.3152658939,0.1287043095,0.1994595081,0.1697904617,0.2229832262,-0.1316216737,0.1720539331,0.1581799388,0.5512159467,-0.2044785023,0.3849100173,0.2994396389,0.0794238672,0.0352438837,-0.1622576118,0.1220686436,0.2517001033,0.0881588459,-0.3062875867,-0.16598396,0.2079837918,-0.3473239541,-0.574014008,0.1600897908,-0.1192490309,0.0359022766,-0.0519875512,0.0414482132,0.1185077503,-0.1647282541,0.1501125544,0.4650923014,-0.0852173194,0.2689602673,-0.0288297459,-0.5663641095,-0.3229424357,0.0118277464,0.3666900098,0.7510473132,-0.0367795229,-0.0580359474,0.1367016286,-0.2418793589,0.2690557837,-0.4555700719,-0.171573326,0.0708443969,0.3321160376,0.049781967,-0.0853177831,-0.3413788378,-0.0995751619,0.0929016694,0.0179550014,0.0016852646,-0.0048173368,0.0494719297,0.1562419534,-0.0499142669,0.0009700467,-0.4999693036,-0.5834292769,-0.2735549212,0.0356477983,0.1798700094,0.2000611871,0.2440768033,0.3113577664,-0.0907899737,-0.1310891062,0.1207939088,0.1393080652,0.0824258104,-0.1110416427,-0.0769585297,-0.2178651839,0.0986213833,-0.1395837516,0.5598430037,0.1188126355,-0.3939176798,0.0808040127,-0.3913773894,0.4750382006,0.2205745727,-0.0482246988,-0.1526371539,0.0295976978,0.0546858162,-0.3251277506,-0.1033275574,0.3757904172,0.107118465,-0.3114070892,-0.0423335545,0.2519550025,0.1587755829,0.1981963068,0.0016101812,0.2698329687,-0.0852806121,0.1210038289,0.106746912,0.7687660456,0.1135380641,0.0985363573,0.1486109793,-0.487600565,0.3208132386,-0.0086525474,0.0897578746,-0.173513785,-0.2528032362,-0.1189965531,0.233820647,0.5631158948,-0.0795739517,-0.3419634402,0.1924080104,0.3476173282,0.1065314785,0.3265296519,0.3944890797,-0.5686751008,-0.3911683261,-0.3675346971,0.0375816636,0.0253796596,0.1202061772,0.0490008742,-0.1987502426,0.1520520598,-0.2151767015,-0.3568888605,0.0265987013,-0.561429143,-0.0691109225,-0.1016776413,-0.1371914446,0.0828204453,0.1239934638,0.5200181603,0.0427654795,0.1907686293,-0.0325326398,0.4909977913,-0.086387217,0.284296751,-0.0053275232,-0.0208366159,-0.0911151245,-0.244768396,0.0304845441,0.1265695989,0.0151044345,0.0485376082,0.0213345252,-0.1358715594,0.0794990808,-0.167414695,-0.2160183787,0.0747633055,0.3042266071,0.0533663519,-0.0310255419,-0.0037366312,0.196513921,-0.1772778183,-0.2435431331,-0.05394217,-0.0563774928,-0.2146795243,-0.3788553178,-0.0693967119,-0.1819036156,-0.2666616738,-0.0569930859,0.040492218,-0.2393921912,-0.0085531585,-0.2823258638,0.0370041318,-0.080323711,-0.1316894293,-0.0484637916,0.1440740228,-0.0317695178,0.0722286627,-0.6231847405,-0.2428381741,0.3059763312,0.1047405154,-0.1449358463,0.1105077043,0.130558908,0.0126918601,-0.3582857847,-0.1965313405,0.1318673193,0.0609747693,-0.0028748889,0.4972391427,-0.0034724458,0.0957305729,0.1236132011,0.0042776726,0.3776924014,0.0118919462,-0.0582519919,-0.1805169582,0.1287736893,0.0875171572,-0.2672631443,0.3187815547,-0.1304059178,0.0047070701,-0.1004949436,0.1612779349,0.1779608577,-0.2146264464,0.281083256,-0.1274720579,0.2242746502,-0.0038685985,0.1245329455,-0.1865690202,0.6320711374,0.2263941318,0.2795944512,-0.1015354022,0.2469105572,0.2319065183,-0.0249609128,-0.2458360046,0.0919650123,0.1280023158,-0.2220268548,0.1055594534,-0.0579249971,0.1261825264,0.3847450316,-0.1186875403,-0.1219965145,0.579387486,0.0930615664,-0.0575102568,0.1211392879,0.2755745351,-0.0688014776,0.0264898147,0.3882041872,0.408978194,0.1765515506,0.2406212538,-0.1253906041,0.5952467918,0.0011201998,0.1183713898,0.3623171151,-0.047286272,0.1879994571,-0.0708358213,0.0142517043,0.068466194,0.2516631782,-0.3138509989,-0.0044326116,0.3850510716,0.2570827901,-0.2002011836,-0.2042310387,0.0115689244,-0.1148200482,-0.0202361345,0.3192639053,0.29761976,0.353848815,-0.2145280987,-0.0250058733,-0.0047704037,0.2487775832,0.0457973555,-0.0978997946,0.1763565838,0.1071151197,-0.5134841204,0.2841794193,-0.0460393615,-0.177413255,0.3543101847,0.2523783743,0.0853629485,-0.2463068813,-0.1491591185,-0.0464271903,0.3513022363,-0.2995989919,0.1949407458,0.297938168,0.1131566167,0.0273726787,-0.1121849492,0.5564958453,0.3554884791,-0.4729909897,0.1146875471,-0.3460694253,0.0918731913,0.0000266945,0.1166412085,0.3284035325,0.4090204835,0.1822415888,0.0790033564,0.0478704646,-0.1397525966,0.0707262903,0.0442092121,0.10398186,0.3787470758,0.0144949704,0.3178913295,-0.2724893987,-0.1072372273,-0.4920473099,-0.0218807962,0.1937068403,-0.0931999907,0.185405165,0.016116336,0.0180905964,0.2927776873,0.1242395416,0.3251207769,0.1901239753,0.0053433203,-0.1124474779,-0.7355491519,0.1676690876,-0.0577995628,-0.4195208549,-0.1105898842,-0.091639854,-0.0929638743,0.4235082865,-0.1873225868,0.7202270031,0.014636483,0.064138636,-0.0472101085,-0.141022712,0.1466666311,-0.0634539649,-0.0797409415,-0.3795458972,0.0356135555,-0.0526407883,-0.0431566387,-0.1462205499,-0.0894531906,-0.4586206675,-0.0084653106,0.17162925,0.5348653793,0.6343557239,-0.264826566,-0.0515964776,-0.343350023,-0.0217514187,-0.3386199474,0.2185120285,-0.1061352715,0.001000747,-0.2635776103,-0.0179872178,-0.453659445,0.2437256128,0.0708178431,0.2721259296,0.0445527397,0.1515187174,0.2206400335,-0.0536919571,0.0263820067,0.3207274079,-0.0657147542,0.1566605419,0.1568377167,-0.3324783146,0.1384954751,-0.2053885609,-0.2074968517,0.1461064517,0.2187728435,0.1676390767,0.1232252866,-0.3770213723,0.0337976664,-0.0189497825,-0.1387718469,-0.3186836243,0.2775359452,-0.0934364945,-0.3046444952,-0.1472955197,0.2229944021,0.0477050059,-0.2741551101,0.1133416593,-0.4106862545]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/240","title":"Deterministic dataset loading","comments":"Yes good point !","body":"When calling:\r\n```python \r\nimport nlp\r\ndataset = nlp.load_dataset(\"trivia_qa\", split=\"validation[:1%]\")\r\n```\r\n\r\nthe resulting dataset is not deterministic over different google colabs. \r\nAfter talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/datasets\/trivia_qa\/trivia_qa.py#L180\r\n\r\nwhich seems to return an ordering of files that depends on the filesystem:\r\nhttps:\/\/stackoverflow.com\/questions\/6773584\/how-is-pythons-glob-glob-ordered\r\n\r\nI think we should go through all the dataset scripts and make sure to have deterministic behavior.\r\n\r\nA simple solution for `glob.glob()` would be to just replace it with `sorted(glob.glob())` to have everything sorted by name. \r\n\r\nWhat do you think @lhoestq?","comment_length":4,"text":"Deterministic dataset loading\nWhen calling:\r\n```python \r\nimport nlp\r\ndataset = nlp.load_dataset(\"trivia_qa\", split=\"validation[:1%]\")\r\n```\r\n\r\nthe resulting dataset is not deterministic over different google colabs. \r\nAfter talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/datasets\/trivia_qa\/trivia_qa.py#L180\r\n\r\nwhich seems to return an ordering of files that depends on the filesystem:\r\nhttps:\/\/stackoverflow.com\/questions\/6773584\/how-is-pythons-glob-glob-ordered\r\n\r\nI think we should go through all the dataset scripts and make sure to have deterministic behavior.\r\n\r\nA simple solution for `glob.glob()` would be to just replace it with `sorted(glob.glob())` to have everything sorted by name. \r\n\r\nWhat do you think @lhoestq?\nYes good point !","embeddings":[0.0945873633,0.0957692191,-0.0216399673,0.1657868624,0.1450249702,-0.13906537,0.2475258708,0.0627876297,0.1644714177,0.0390271358,0.098517932,0.4085090458,-0.1912948489,0.1137885079,-0.0135012707,-0.0098155076,-0.0236251093,0.2353990376,-0.0582889505,-0.2304395437,-0.3629297316,0.0748674572,-0.4281496704,-0.3791905046,-0.2330177575,0.0300920866,-0.07254228,0.1379635334,-0.1247970313,-0.455832094,0.0853844732,0.4657485187,0.1013024822,0.2091461867,-0.0001121603,-0.2553572953,0.3605852127,-0.1061380059,-0.4217329919,-0.3430823088,-0.4110149145,0.0089807315,0.0619817041,-0.0685706437,-0.338845402,0.0982361361,0.4541248977,-0.2306975424,0.2911240757,0.1210905015,0.197321713,-0.0830761343,-0.0154164918,0.1488040388,0.029082777,0.09761177,-0.0620057173,0.3898255527,0.0588201582,-0.2425620556,-0.0758030415,0.1715105027,0.0982830301,0.0862753391,0.4285977781,0.1529737711,-0.2098231167,-0.3419148922,-0.0863937736,0.2626880109,0.4616270363,-0.1745405793,-0.2642199397,-0.3310501873,0.2287247777,-0.1425824314,0.1147182733,0.4061127603,-0.1986084878,-0.0713726953,-0.0125610223,0.2050287575,-0.1536396593,0.0699101537,0.000433458,0.6247884035,-0.000666994,-0.0124212736,0.0731943399,0.2142144144,0.1652068645,-0.1966495514,-0.021718936,0.1023840979,-0.3065754473,-0.2278914601,0.2182696164,0.4656026363,0.0998885781,0.1660636216,0.1884724051,0.0810813084,-0.1540548652,0.0108460076,0.3998070955,0.5025461912,-0.0059920195,0.0808774978,0.4238455296,0.0402510799,-0.1507389247,0.2641663849,0.0238320883,-0.2590076327,0.1137864962,0.1261232942,0.1767624468,-0.1790841967,-0.4575904608,-0.2504689693,-0.2466423362,-0.0301307626,-0.0530577712,0.2711630464,-0.2853363752,0.1210361272,-0.2495712787,0.1719069034,-0.1196708828,-0.0325187035,-0.1031102091,-0.1106035262,-0.4100127816,0.1662596911,0.1473224759,0.1010723412,0.1508559138,0.1590875685,-0.2014039904,-0.0267433804,0.1373196393,-0.3891229331,0.0867620036,-0.0422472693,0.1134859249,-0.1292403489,-0.2180735022,-0.1961421371,-0.399430871,0.0577632226,-0.1517410427,-0.0710951164,-0.0508377478,0.174965173,-0.2922861874,-0.1105073467,-0.3568353653,0.3473115265,0.1291597486,-0.1031193063,-0.1125476509,-0.191427812,-0.3306518197,-0.344742924,-0.2137427479,0.5478838682,-0.2774924934,0.1105904877,-0.0455848649,0.2671298087,0.2401722372,0.3206399977,0.0061372444,0.1946215779,-0.2305689752,0.0254184268,0.1557523012,-0.0916064009,-0.1422694772,0.1202926636,-0.2160172909,0.3048881888,0.0088647725,0.2449086159,0.0910394043,-0.2860219181,0.1862355173,0.5252295732,0.0735167712,0.1972567737,-0.2246299237,-0.1493048221,0.1308275759,0.0337726399,-0.0980923399,0.1973364651,-0.0029889874,-0.0651959032,0.2506788373,0.0030806647,-0.3431002796,0.0812759995,0.479627341,0.0174300577,-0.0255804062,0.0177303217,-0.6688655019,0.2080661952,-0.1182608232,0.2904148698,-0.1074185446,-0.2168058306,-0.0278808735,-0.198851645,-0.3309913874,-0.0446577482,0.176835373,0.4819156229,0.2809567451,0.2744262516,0.089605853,0.4671111405,0.0528310984,-0.2160604447,-0.4221501052,0.3886922896,-0.0930935666,-0.1158286259,0.2605186999,0.2604146898,-0.0383868627,-0.2277856767,-0.0521066822,0.2976858318,0.194045946,0.1116610914,0.2023958564,-0.0433240645,0.0711556599,0.0090003368,0.1713510156,0.3018912375,-0.0588629097,-0.0893184245,-0.0719896555,0.462595731,-0.0990460366,0.2405830473,0.1309930384,-0.1047270149,0.1151052341,0.0127048381,0.0002707007,-0.2699932456,0.2006544024,0.2628410459,0.3840576708,0.4334153533,-0.6345706582,0.2992953956,0.7336896062,0.1120752692,0.0958999395,-0.0273070317,0.1125184819,-0.2618478239,0.2250386924,0.33389768,0.6063230038,0.1990608722,-0.0519668013,-0.0534234755,-0.1466808617,-0.2642047107,0.0272859801,-0.0275602657,0.1815913618,0.1141093895,0.3633785546,0.1708364785,-0.2874180079,-0.1045644581,-0.0912683532,-0.0106382947,-0.0793241262,0.1837315857,-0.2107187361,-0.3533428311,-0.3913383186,-0.0773745701,-0.2960610688,-0.2441903353,0.1438769102,0.239109382,0.0294077713,-0.0959670544,0.1386108547,0.3186073005,-0.0360313468,-0.4599055946,-0.1472352296,-0.0816993862,-0.3718604147,0.0193874948,0.3784902096,0.2507220507,0.5035094023,-0.1553113163,-0.2040880322,0.1087939814,-0.3416737616,-0.0796356499,-0.2907286584,0.3566903472,-0.0035402272,-0.0790899843,-0.0501508117,-0.0444994234,0.0035493742,0.0120701371,-0.1075341702,-0.0031892622,0.0745603815,-0.0086185085,-0.087840125,-0.3912994266,-0.2677439451,-0.1686951816,0.1127131581,0.3277488947,0.1459663361,0.2813071609,-0.0798528716,-0.2552734613,0.1426594108,0.0575575568,-0.1259056032,0.0003590172,0.3458511531,-0.2229515612,-0.2180695385,-0.3586640358,-0.2016667426,-0.0553641655,0.1115349531,-0.2697105408,-0.3426713943,-0.3564448059,0.0110133355,-0.0505338386,-0.2243531942,0.4319261014,0.030906193,0.0060656318,0.1223119497,0.0000989746,0.2013240904,0.3131582737,0.3817140162,0.1690115929,0.0313939825,0.0394893065,0.4641197324,0.0182936862,-0.3707867265,0.2189803272,0.0944704264,0.3979997635,-0.461389184,-0.0348865688,0.2702769637,0.2443663776,-0.1281114817,0.5115141869,-0.0273497868,-0.0656518936,-0.0997678861,0.2902362347,-0.2149215639,-0.2148267329,0.0273038708,-0.1791304648,-0.1976199001,0.1458735913,0.1309804618,0.0519328229,-0.3418011367,-0.2885336876,0.1562055051,0.0869726837,0.073423095,0.2224906236,0.0325090401,-0.6342589855,0.1043260321,0.0429229811,0.0417224132,0.1036004871,-0.1462161988,0.265581131,-0.0799555331,0.2075460851,-0.0019088186,0.1656284481,-0.0541342087,0.080527097,0.0689246878,-0.2224892825,-0.3352823257,0.0201770607,0.1791301817,0.290592134,-0.4598352909,-0.1265408546,0.0220872071,0.1633172333,0.0034228102,-0.449123621,-0.5496536493,-0.0504080653,0.1872368008,-0.0946921632,-0.0124750072,0.1276665181,-0.4247058332,-0.1642598212,-0.0609407239,-0.2513067722,0.2290420979,0.3448551893,0.2602017224,0.0191607103,0.0064181667,-0.007637925,0.1463221908,0.1707733423,0.3694819212,-0.0344083011,-0.1781571507,-0.2002730966,-0.1303507537,-0.0254274625,0.2777811587,-0.1038765162,-0.2058018744,-0.6165345907,0.0736133158,-0.2284697294,0.2640001774,0.1892525852,-0.1393611729,-0.4666115642,-0.4202456772,0.5441010594,0.1095592752,-0.3037032485,0.3586831689,0.0082505094,-0.4960257411,0.5214697719,0.290099144,0.739466548,-0.1992961168,0.1557766199,0.3237666786,0.05924749,0.6052289605,0.1179985255,0.4471523762,-0.5529441237,-0.2676442564,-0.0010401139,-0.0802007467,-0.1360311508,0.2899307013,-0.4143084884,0.5073554516,0.0618496537,0.1580729187,-0.0169406589,0.5113559961,-0.1448287964,0.0120042376,-0.338113457,0.1204675138,-0.1252302527,0.3908526301,-0.1911241412,-0.1620530039,0.0260569323,0.0100087877,-0.0710669532,0.290017277,-0.3537419736,-0.0213610679,-0.0051271645,-0.2210932672,-0.0128949368,0.3737125993,0.1311346442,0.255632937,-0.1189930886,0.0517843403,0.1799429804,0.1572644413,0.1347696334,-0.2112134099,0.4062017202,0.1680815667,0.0403020941,-0.357583791,0.1716174632,-0.3608961105,-0.4728999436,0.1663427204,-0.2715691924,0.302480787,-0.2245599031,0.1779046357,-0.020055715,-0.0283257198,0.1547915488,-0.0201600045,0.1645280272,0.4391562641,-0.0300175287,-0.4161788523,-0.0465727635,0.1495119184,0.0611990727,0.2546749711,0.3071766794,-0.1138417646,-0.0590823665,-0.2970591187,-0.1673080027,-0.4655505717,-0.0466488674,0.2917495966,0.0803639442,0.0962436721,0.1552900821,0.3003555536,-0.0084196199,0.1062903851,0.0476567,-0.4432260394,0.1681710929,-0.009553046,-0.1219599769,0.1747353822,-0.0397958159,0.2957703173,-0.0127706844,0.0249647647,-0.3482517898,-0.2375494838,-0.1657047272,0.0262653437,-0.131107837,-0.0982202142,0.070138365,0.104485482,0.110805966,-0.019820258,-0.0722890124,-0.2425189167,0.0795583576,0.110764116,0.0911601186,0.1991166174,-0.1499141753,-0.3796131611,-0.1504721045,-0.3850801885,0.1110565588,0.0472604856,0.0364918299,0.1306547523,0.2070592344,0.2492648512,0.2227492481,0.2118603289,-0.3964516819,-0.0593568943,0.1401616633,0.2004080862,0.074166283,0.1656965911,-0.2110730857,-0.0015268105,0.0333920419,0.2867673934,-0.1224057227,-0.1974633634,-0.0800315067,-0.1305103153,0.1998352259,0.6523944139,0.0341908857,-0.5614849925,0.1746048331,0.1900301427,-0.2814773619,0.1707381606,-0.0831928477,0.3381953239,0.0905876309,0.3099751472,0.2529183328,0.1261951327,-0.1003006995,0.0155376559,0.2616707087,-0.0275982767,0.3439424038,0.0590963215,0.1233297586,0.0092969593,0.1304981261,-0.1411222965,-0.1154860184,0.3067248166,0.0478547215,0.336376071,-0.2321509421,-0.2007266879,0.137896359,-0.0766112432,0.0155163035,0.7675790191,-0.0326105431,0.1634783894,0.3650996089,0.3771992624,-0.3772962689,-0.1003236175,-0.4719051719,-0.1214208454,-0.0755869448,-0.0842413828,-0.0723197609,-0.0026963218,-0.1283655167,0.0910646245,-0.0795390904,-0.0644986629,-0.0042182305,0.1274305582,-0.1452159435,-0.4883601964,0.3973898292,-0.0315437429,0.3258887827,-0.098228626,0.2660138607,-0.1981152147,0.1410780102,0.0584546737,0.1020254344,0.1237435341,0.7621641755,-0.3282700777,-0.2412863374,-0.2228099555,0.0422113724,-0.144883737,0.1850458533,-0.016135579,-0.1157470867,0.5171301961,0.1201481149,-0.1034071296,0.3082436919,0.4065875113,0.0108568808,-0.0712031052,0.4482508898,0.0358630307,0.045473963,-0.0355448648,0.1857178509,-0.3967195153,-0.1338685304,-0.0866045207,0.0320603251,-0.1045809388,-0.3045164049,0.0789407492,-0.0833216533,0.4066400826,0.2031700909,0.2332466394,-0.2899755239,-0.2169504911,-0.6358399987,0.2143093646,0.0397580229,-0.0254734289,-0.34296754,0.4159320295,-0.3521982729,-0.0057321796,-0.420889318,0.3622388542,-0.0947967619,0.0636213869,-0.5731700659,-0.0496977977,-0.4883071184,-0.2265347093,-0.0542441718,-0.4501628578,-0.1004313231,-0.2850353122,0.0023564589,-0.0007474959,0.3243255615,0.0572231822,0.432625562,0.3048868179,0.1729754508,0.3562880754,-0.0722509548,-0.0333645158,-0.3082391918,-0.0235099886,-0.4578950405,0.1674207896,0.1241309568,0.2045638412,-0.0154692018,0.3024786711,-0.1320949644,0.4648543298,0.0696463883,0.1436746866,-0.2570763826,-0.1196300089,-0.1480411291,0.0785077885,-0.129868865,0.1506782323,0.0280945282,0.0047676703,-0.2119609863,-0.2867402434,0.4253305495,-0.1997943521,-0.2538209558,-0.1128293127,0.2482505441,0.1291116178,-0.3027158678,-0.4496001005,0.0697408691,0.2309681028,0.0575402118,0.1337697357,0.3503191471,0.0135866255,0.0912835374,-0.2109602094,0.0981379375,-0.0754856914,-0.2982876301,-0.1027177051,-0.2487876266]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/240","title":"Deterministic dataset loading","comments":"I think using `sorted(glob.glob())` would actually solve this problem. Can you think of other reasons why dataset loading might not be deterministic? @mariamabarham @yjernite @lhoestq @thomwolf . \r\n\r\nI can do a sweep through the dataset scripts and fix the glob.glob() if you guys are ok with it","body":"When calling:\r\n```python \r\nimport nlp\r\ndataset = nlp.load_dataset(\"trivia_qa\", split=\"validation[:1%]\")\r\n```\r\n\r\nthe resulting dataset is not deterministic over different google colabs. \r\nAfter talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/datasets\/trivia_qa\/trivia_qa.py#L180\r\n\r\nwhich seems to return an ordering of files that depends on the filesystem:\r\nhttps:\/\/stackoverflow.com\/questions\/6773584\/how-is-pythons-glob-glob-ordered\r\n\r\nI think we should go through all the dataset scripts and make sure to have deterministic behavior.\r\n\r\nA simple solution for `glob.glob()` would be to just replace it with `sorted(glob.glob())` to have everything sorted by name. \r\n\r\nWhat do you think @lhoestq?","comment_length":47,"text":"Deterministic dataset loading\nWhen calling:\r\n```python \r\nimport nlp\r\ndataset = nlp.load_dataset(\"trivia_qa\", split=\"validation[:1%]\")\r\n```\r\n\r\nthe resulting dataset is not deterministic over different google colabs. \r\nAfter talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/datasets\/trivia_qa\/trivia_qa.py#L180\r\n\r\nwhich seems to return an ordering of files that depends on the filesystem:\r\nhttps:\/\/stackoverflow.com\/questions\/6773584\/how-is-pythons-glob-glob-ordered\r\n\r\nI think we should go through all the dataset scripts and make sure to have deterministic behavior.\r\n\r\nA simple solution for `glob.glob()` would be to just replace it with `sorted(glob.glob())` to have everything sorted by name. \r\n\r\nWhat do you think @lhoestq?\nI think using `sorted(glob.glob())` would actually solve this problem. Can you think of other reasons why dataset loading might not be deterministic? @mariamabarham @yjernite @lhoestq @thomwolf . \r\n\r\nI can do a sweep through the dataset scripts and fix the glob.glob() if you guys are ok with it","embeddings":[0.1278988868,0.0680350214,0.018163044,0.1612559557,0.1494209617,-0.0938441157,0.2877918482,0.0737954974,0.2083065808,-0.0574632734,0.1042929068,0.3704168797,-0.1524901688,0.1114908755,-0.0127119003,-0.0483885668,-0.0260270219,0.1764348149,-0.0584851317,-0.1872009039,-0.3069525063,0.0784734413,-0.4576539397,-0.3690684438,-0.2645183802,0.0899192244,-0.0991521105,0.1341389865,-0.0830376893,-0.4587273598,0.1570782363,0.5054252148,0.1628576219,0.1823225468,-0.0001150454,-0.2656959295,0.4034866095,-0.1031261235,-0.4438330531,-0.3913377523,-0.3945771158,-0.0304798484,0.1379951537,-0.0177153181,-0.2774758041,0.0264262781,0.3882413805,-0.2673333287,0.2669376433,0.0418426171,0.1728890985,-0.1249023452,0.0030792588,0.1387374699,0.0104721747,0.0695920885,-0.0228298008,0.4524676204,0.1382265389,-0.2738265991,-0.0689454302,0.1072151735,0.031256482,0.1048302799,0.4822502732,0.1279492378,-0.2176283002,-0.400513202,-0.1074987575,0.2443935126,0.582634747,-0.1721649319,-0.2514915168,-0.352547586,0.2645475864,-0.1000305116,0.1244716272,0.380508095,-0.134509325,-0.0959320664,-0.0022379844,0.1671114117,-0.1371524483,0.0710875019,0.0856458023,0.5879830718,-0.013878718,0.0256638918,0.0793035403,0.2684374452,0.2062194347,-0.2505574524,-0.0696311295,0.1474831551,-0.316983819,-0.1194183305,0.1632873714,0.5273426175,0.0931168944,0.1871911436,0.226766631,0.0674799159,-0.1524080336,0.008962173,0.3967241645,0.5014610887,0.0129531249,0.0889035016,0.3393018246,0.0102449907,-0.1401090324,0.2734246552,-0.060456071,-0.318338275,0.1536530554,0.0968183875,0.1490698606,-0.2396645695,-0.4463087618,-0.2819098234,-0.2935085297,-0.0425756872,0.0255235918,0.2448837459,-0.3041346967,0.1373276114,-0.2099104524,0.1781618744,-0.1122649014,-0.066726014,-0.0442476049,-0.1189903989,-0.3672465682,0.1553732008,0.1779651642,0.0710023493,0.1543083936,0.2629646659,-0.2230555415,-0.0933227241,0.0318714604,-0.3338704109,0.0340087451,-0.0125172529,0.1098929197,-0.0834866986,-0.2122244835,-0.0992376581,-0.3635241091,0.109333314,-0.2231628746,-0.1021245122,-0.1164172366,0.1394500434,-0.3203218877,-0.141747877,-0.4392810762,0.3463121653,0.0661278293,-0.085641861,-0.1390953958,-0.2519544363,-0.3000761867,-0.3598109484,-0.1437502056,0.5643336773,-0.333956331,0.0883162245,-0.0775070265,0.2793798149,0.3261723518,0.3118193746,-0.0433119647,0.2612887621,-0.2711884975,-0.0348494351,0.0915929154,-0.076557532,-0.1479871869,0.1560693979,-0.2097709924,0.3643498123,-0.0059251441,0.288792491,0.0269558374,-0.2857646048,0.2122264951,0.5817522407,0.1125245392,0.1455496252,-0.2228961289,-0.1160703376,0.1606504917,0.1341605335,-0.1529251933,0.2557583153,-0.0797494054,-0.1022738442,0.2364648879,0.0587814227,-0.3761658072,0.0654713288,0.4298803508,0.0566270873,-0.0059010242,0.0919143707,-0.802533567,0.2653383017,-0.1356774271,0.2197048217,-0.1007074937,-0.1710778326,-0.0598150566,-0.1929310113,-0.3671398461,0.0432257727,0.1095663309,0.3891427517,0.2721548975,0.2649303973,0.1429004669,0.5052405596,0.0226428248,-0.1303787231,-0.4089696407,0.4076074958,-0.0580010265,-0.1394962817,0.3009549677,0.2322888821,0.0106304036,-0.2568787038,-0.1099532694,0.3478696346,0.1922554821,0.1490982622,0.1422685087,-0.0063865264,0.0790573433,0.0635892227,0.2284526229,0.2303832173,-0.0306497291,-0.142277196,-0.0237642191,0.4832135737,-0.1408339143,0.2432553321,0.106881395,-0.0918823034,0.0576850101,0.0561521091,-0.0098581268,-0.1848978996,0.2872168422,0.2241625935,0.3984701931,0.4891129434,-0.6162044406,0.2685339451,0.6741495728,0.1172868982,0.0545707718,0.0422465354,0.1281605065,-0.2923390865,0.2435434461,0.3109410107,0.5828956962,0.1728081405,-0.1215939075,-0.0487129018,-0.1323100924,-0.256632477,0.0082253413,-0.0497053936,0.231030643,0.068370007,0.3204985857,0.2718552351,-0.2242341489,-0.1144805327,-0.1183806583,0.0355489403,-0.1283758134,0.1732635051,-0.2431495041,-0.3185338974,-0.3550515175,-0.0998179838,-0.3305409849,-0.2406738102,0.0725947246,0.2184052914,0.0571436323,-0.1266395003,0.1862879097,0.2984232008,-0.0690757111,-0.4595802724,-0.1809671819,-0.0450526848,-0.3760229945,-0.0321737155,0.4337281585,0.2527377307,0.48213467,-0.134592101,-0.2056815326,0.1416743249,-0.2974398136,-0.1305258572,-0.3035556972,0.3335773945,-0.0823215246,-0.0986532941,-0.0242812317,-0.0864902288,0.0119557856,-0.0091673518,-0.0324624293,0.0095238388,0.0695247874,0.0021101534,-0.048866868,-0.341485709,-0.2831537426,-0.1440596282,0.0879117325,0.2736336887,0.1332778633,0.3045849502,-0.0792741999,-0.2478583753,0.1371490657,0.0449613631,-0.0806969404,-0.0878233761,0.3428190053,-0.1608350575,-0.2361774892,-0.3503187299,-0.1138889492,-0.070620127,0.1005068496,-0.3323993087,-0.2848387361,-0.3624153435,0.010045087,-0.07570225,-0.2865492105,0.3644975722,0.0775023624,0.0695832297,0.1888955235,-0.0041415044,0.1965331286,0.292099148,0.3772507012,0.1817639917,0.0183896068,-0.0339224599,0.4845993817,-0.0126197385,-0.41455549,0.2060723752,0.15878658,0.3422958553,-0.4762473106,-0.068302013,0.2320453376,0.1910715252,-0.1054775789,0.5191330314,-0.0405323282,-0.0447408482,-0.1158997715,0.3423368633,-0.2025170773,-0.2299305648,0.0364479236,-0.2357175201,-0.225430131,0.1037498787,0.0921147317,0.0534585677,-0.3630147576,-0.2038318515,0.1484143883,0.1867225617,0.0684405789,0.2117994428,0.0499445498,-0.5956832767,0.105544433,0.0134531707,0.0538207218,0.0250208303,-0.0765682459,0.2728616297,-0.0241453573,0.2414186001,-0.024873428,0.2168451995,-0.0057119746,0.0849963725,0.0862814486,-0.3125590384,-0.2967254221,0.0863681361,0.1897643209,0.2033002228,-0.4483338892,-0.1149778888,0.0259382613,0.1493924707,0.0077363346,-0.51786834,-0.5376967788,0.0175013896,0.1657565832,-0.1861450672,-0.0360143892,0.1354911625,-0.5030944347,-0.1736017317,-0.0176034886,-0.2194736153,0.217848137,0.2394248545,0.2116100639,0.0699539632,0.0652293935,0.084498778,0.1672157943,0.2078924179,0.3494348526,-0.0006086368,-0.2274155021,-0.1979618371,-0.1107973307,0.0176916774,0.2727665007,-0.0900615305,-0.2178982049,-0.6402572393,0.0575642511,-0.2421330959,0.2026738226,0.2156065404,-0.1161156073,-0.5116728544,-0.4302211106,0.5949904919,0.1257106215,-0.2779100537,0.4554837346,-0.0401198529,-0.4783618152,0.4830440581,0.2197094113,0.7067087293,-0.293058008,0.0472508706,0.2937939167,0.0922698826,0.535089016,0.1695132852,0.3881170452,-0.5430061221,-0.2523645163,0.0206541214,-0.0875586495,-0.0799329951,0.3004677594,-0.4050023556,0.5030626655,-0.0255993027,0.1538592726,0.0421669856,0.549394846,-0.0957556292,-0.0577286817,-0.3514424264,0.0697952881,-0.1942262501,0.4118383527,-0.2053389847,-0.1589804143,0.0530072972,0.1006594375,-0.1176102757,0.3257851005,-0.3543224633,-0.0740365162,-0.0702572241,-0.2141356915,0.106462501,0.4508429468,0.079525955,0.2360016555,-0.1452397257,0.0526542515,0.0845048726,0.1127118394,0.1070463136,-0.2086251378,0.3768033683,0.2033645809,0.0071983635,-0.3862644136,0.1244903803,-0.3595887423,-0.459084481,0.125517264,-0.1770808101,0.2942007184,-0.2412121892,0.1894846112,0.0286817048,0.0004607666,0.125567317,0.0205137022,0.1814477295,0.3807770312,-0.0202603601,-0.4196606874,-0.0334517732,0.1662198156,0.0007932456,0.2514366508,0.3638491631,-0.0406991765,-0.0357814245,-0.2425656915,-0.2162797004,-0.4363132715,-0.1138789132,0.2943444848,0.0920971036,0.1739566773,0.115508005,0.2473016679,0.0570255667,0.0525809042,0.0801675692,-0.4885747433,0.1931038648,-0.0287792534,-0.1397668421,0.1369879395,-0.0747047663,0.2365794778,0.065707922,0.0516688153,-0.2957158387,-0.1675573438,-0.1178113744,-0.0258158669,-0.1618572623,-0.1259131134,0.0413538031,0.082714662,0.075128071,-0.0821870938,-0.0353246666,-0.1888912022,0.104170382,0.1273991019,0.0788291395,0.1892597377,-0.1316522062,-0.4576672316,-0.1448698491,-0.353197664,0.1129235178,0.0459807366,0.0741409212,0.0992606208,0.2913087606,0.312730372,0.1848030835,0.2687434852,-0.4418791234,-0.0093452921,0.1626241654,0.126436308,0.0636360645,0.168293491,-0.2337189764,-0.0421537347,0.0580376908,0.2631934285,-0.0512111709,-0.1948946863,0.0028901519,-0.1389047354,0.1837907583,0.6173468232,0.0152098024,-0.5521917343,0.1506928205,0.1556540877,-0.244145155,0.2116540968,-0.0547293872,0.4376530349,-0.010482965,0.2724595368,0.2875722945,0.1769684404,-0.1432020664,-0.0171978623,0.3120467663,0.0553477295,0.3701547682,0.0785536915,0.1238358319,0.0375577658,0.1584036052,-0.1039266214,-0.1650943905,0.2137206793,-0.0163127948,0.3138754666,-0.1796164066,-0.0992458388,0.1361103952,-0.1502589434,-0.038549345,0.8459666967,-0.0161345787,0.1608428955,0.3227763474,0.3865442872,-0.385250181,-0.1279685199,-0.4479801655,-0.1292566806,-0.0415172279,-0.0660078228,-0.0434928685,0.0215130728,-0.1841611713,0.1429527253,-0.1311571002,-0.1142279506,-0.0274459869,0.1246717721,-0.1655519307,-0.5515330434,0.4079733789,-0.038978111,0.3758791089,-0.0909089968,0.226477623,-0.2193036228,0.0974198952,0.119567655,0.0858764276,0.1730549634,0.7899259329,-0.3634607494,-0.2163263112,-0.1652639359,0.0856740773,-0.0864154845,0.2728657722,-0.0151236663,-0.134123072,0.4716494083,0.0751957074,-0.0771791041,0.2771446407,0.3830429316,0.0813128576,-0.067162931,0.4809931219,0.0537728667,0.023943048,-0.0430588499,0.2462714911,-0.3554511666,-0.1571777761,-0.0762023106,0.0188974645,-0.1588055193,-0.2214760184,0.0629831627,-0.0946714133,0.4574965239,0.2913885415,0.2396603227,-0.3542179763,-0.2383306772,-0.6219888926,0.1719860286,0.0691952258,-0.0235020351,-0.2315736711,0.3624781668,-0.381433636,-0.0087280935,-0.4201943576,0.3539237678,-0.0965951234,0.0381138511,-0.574028492,0.0038930578,-0.5121834278,-0.2196490318,-0.0817790776,-0.4147522151,-0.0534516685,-0.301007688,-0.0659993291,-0.0411644019,0.3028202951,0.085869506,0.4274740517,0.3033766747,0.1091611683,0.3528417349,0.055237826,-0.0755785033,-0.3462198079,-0.0992271751,-0.3638622761,0.2536395192,0.1099807918,0.1332284063,0.001337167,0.3380185366,-0.1864715815,0.4435076118,0.0251882728,0.1968639791,-0.3721699417,-0.2078318745,-0.1490301937,0.0654222518,-0.1195823774,0.1091083363,0.0582161881,0.0485962778,-0.2353591174,-0.1628765017,0.4360439777,-0.1476989239,-0.2143805325,-0.0688067898,0.274297744,0.1284197122,-0.2526897788,-0.4334496558,0.0609670356,0.2695488334,-0.0059939385,0.1471029967,0.3941020966,-0.0061943256,0.1405069977,-0.1948396862,0.0496964604,-0.130154103,-0.2397856712,-0.0945941657,-0.2226564437]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/240","title":"Deterministic dataset loading","comments":"I'm pretty sure it would solve the problem too.\r\n\r\nThe only other dataset that is not deterministic right now is `blog_authorship_corpus` (see #215) but this is a problem related to string encodings.","body":"When calling:\r\n```python \r\nimport nlp\r\ndataset = nlp.load_dataset(\"trivia_qa\", split=\"validation[:1%]\")\r\n```\r\n\r\nthe resulting dataset is not deterministic over different google colabs. \r\nAfter talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/datasets\/trivia_qa\/trivia_qa.py#L180\r\n\r\nwhich seems to return an ordering of files that depends on the filesystem:\r\nhttps:\/\/stackoverflow.com\/questions\/6773584\/how-is-pythons-glob-glob-ordered\r\n\r\nI think we should go through all the dataset scripts and make sure to have deterministic behavior.\r\n\r\nA simple solution for `glob.glob()` would be to just replace it with `sorted(glob.glob())` to have everything sorted by name. \r\n\r\nWhat do you think @lhoestq?","comment_length":32,"text":"Deterministic dataset loading\nWhen calling:\r\n```python \r\nimport nlp\r\ndataset = nlp.load_dataset(\"trivia_qa\", split=\"validation[:1%]\")\r\n```\r\n\r\nthe resulting dataset is not deterministic over different google colabs. \r\nAfter talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/datasets\/trivia_qa\/trivia_qa.py#L180\r\n\r\nwhich seems to return an ordering of files that depends on the filesystem:\r\nhttps:\/\/stackoverflow.com\/questions\/6773584\/how-is-pythons-glob-glob-ordered\r\n\r\nI think we should go through all the dataset scripts and make sure to have deterministic behavior.\r\n\r\nA simple solution for `glob.glob()` would be to just replace it with `sorted(glob.glob())` to have everything sorted by name. \r\n\r\nWhat do you think @lhoestq?\nI'm pretty sure it would solve the problem too.\r\n\r\nThe only other dataset that is not deterministic right now is `blog_authorship_corpus` (see #215) but this is a problem related to string encodings.","embeddings":[0.1137944087,0.1788549572,0.0189524554,0.1695959121,0.1481847465,-0.0693371296,0.2104988247,0.1159460098,0.1663966626,0.0036022905,0.1294664294,0.3810071647,-0.1616073549,0.1054980531,-0.0421155952,-0.0011640498,-0.0110056577,0.1828359514,0.0234035216,-0.2169196308,-0.3032742143,0.0948169231,-0.4922880232,-0.4265719652,-0.2638105452,0.050531961,-0.0494492315,0.0728175268,-0.0821795687,-0.4266438484,0.0884041637,0.5214291215,0.1297443062,0.2086421102,-0.0001158346,-0.2780812979,0.3429811597,-0.1384103,-0.4034918249,-0.4584883749,-0.4235266447,0.0029786567,0.089284651,-0.0171199627,-0.2904814482,0.110562034,0.4717698693,-0.2651779354,0.2556129992,0.1604441106,0.1706977189,-0.1325588971,-0.0069109569,0.1613929272,0.0084528038,0.1136064753,-0.0788914859,0.365737319,0.0599491559,-0.237290293,-0.1026322246,0.1262462288,-0.0123312455,0.0853772238,0.4108412862,0.1294186264,-0.2025597692,-0.4190253019,-0.048245538,0.2851984501,0.5857836008,-0.1489483416,-0.2884296775,-0.3177915215,0.2706146836,-0.0686203614,0.164473787,0.4762815535,-0.1837822497,-0.0772719234,0.0095022209,0.2165177166,-0.1672361642,0.1256327331,0.0550950542,0.5633864403,0.020567894,-0.046430178,0.0403209664,0.1976746172,0.2665305734,-0.2073193192,-0.0149173141,0.1045660973,-0.2946223021,-0.1331092864,0.1327408403,0.4168833196,0.1423321217,0.1887691021,0.1884180754,0.1247835606,-0.1409153789,0.0277523864,0.3660642505,0.4443119764,0.0469317734,0.0882839635,0.4415364563,0.0005819829,-0.1401025355,0.2631989121,-0.0451807156,-0.2447943091,0.1352208853,0.0917385444,0.1828014702,-0.2448383868,-0.4639567435,-0.2088102549,-0.2868311405,-0.0472718664,-0.0736261457,0.2372121364,-0.277225554,0.1708800495,-0.2723610103,0.2396499515,-0.094397299,-0.0668712035,-0.0513294861,-0.14786461,-0.3596418202,0.1718271673,0.1715838909,0.0433012284,0.1575044394,0.2333868295,-0.2306009829,-0.1313460767,0.0631222874,-0.3846156895,0.0707031041,-0.0532546565,0.1267556697,-0.0679464936,-0.2007551938,-0.2258478999,-0.39488253,0.0732177719,-0.1875868291,-0.0415990837,-0.0811909139,0.1269268543,-0.3006807566,-0.1601521969,-0.3778267205,0.4154417515,0.1039694026,-0.0399899669,-0.1212582663,-0.2215484828,-0.3228255808,-0.3081771135,-0.2047908008,0.5929179192,-0.250954926,0.1347161233,-0.0144268936,0.3353873193,0.3447002172,0.342304498,0.0949027538,0.1837269217,-0.2054882348,0.0315852463,0.0978194252,-0.1302893907,-0.1156396642,0.1216481104,-0.1305008531,0.3174394965,-0.0036311306,0.2452240735,0.0708586574,-0.295288533,0.1327050626,0.5699854493,0.0917882249,0.1661237925,-0.2098805159,-0.1239983886,0.1535026878,0.0869185552,-0.073324956,0.2437751889,-0.0263798069,-0.0693856403,0.1995629221,0.0381739251,-0.354409337,0.1135728732,0.4417835474,0.0259771757,0.0174017753,0.032162834,-0.6761294007,0.2036518455,-0.0810295492,0.2807785273,-0.1386486143,-0.1894450784,-0.031260848,-0.2177087814,-0.351120919,0.0180445854,0.1251591444,0.4610756934,0.296536535,0.3313344717,0.1129760072,0.4452210963,0.027772449,-0.1577457041,-0.4701626301,0.3422989249,-0.0647957176,-0.1605475694,0.2936730087,0.2650282979,-0.0115144039,-0.2074688971,-0.0890064538,0.3012985289,0.2034747452,0.1466956586,0.196500361,-0.0047538569,0.0571860559,-0.0478245988,0.1256566197,0.3327848911,-0.0741519332,-0.0917756706,-0.0437165014,0.4265551269,-0.1051935852,0.245553717,0.0603084601,-0.0633689091,0.1409069449,0.0101236105,0.0338064693,-0.2565670907,0.3141155839,0.1884608716,0.4168983698,0.4318349659,-0.6472997069,0.2532051802,0.6687130928,0.1311240345,0.1337899566,0.0470857248,0.1058642715,-0.2604922056,0.1974286735,0.3241913319,0.4786657095,0.1711111516,-0.0800079778,-0.0753903538,-0.1365277171,-0.3183978498,0.010920926,-0.051234249,0.1863211393,0.1105131954,0.3885009587,0.2145434469,-0.3057562113,-0.0753770545,-0.0978050232,0.0150893796,-0.1057908759,0.1774474978,-0.3200418055,-0.2972366512,-0.3805834055,-0.077479586,-0.3543400466,-0.2322288901,0.1233587861,0.2657910883,0.0365579315,-0.0707651377,0.1533031464,0.2784759998,-0.0448838733,-0.447578162,-0.2064391673,-0.0843680203,-0.3556398451,-0.0400758237,0.4606668353,0.2419704497,0.4747899175,-0.1962220073,-0.2092666477,0.0662846267,-0.3068489134,-0.0887665451,-0.3029898405,0.366817683,-0.0931768641,-0.1227117851,-0.0300362613,-0.0684262738,-0.0041240379,0.0355628133,-0.0885950103,0.1138387918,0.028828647,-0.0352056511,-0.0114942016,-0.3158364892,-0.2380981296,-0.1639774442,0.1147304475,0.3120846748,0.1828108728,0.3057503402,-0.0969119221,-0.3069820404,0.1150780767,0.0978914499,-0.0968039334,-0.0391394645,0.4157397747,-0.1858071536,-0.2301391661,-0.3322431743,-0.1592838615,-0.1092950851,0.1176007167,-0.2903528214,-0.2713468671,-0.3909585476,0.0572606102,-0.0290543437,-0.3220250607,0.4101304412,0.0630945712,0.0524797514,0.1303123236,0.0009941681,0.2009332031,0.2924929261,0.3757012188,0.1786862612,-0.0003599727,0.017692795,0.4301039875,0.0923483148,-0.3738052845,0.2287277579,0.1366498768,0.316393286,-0.4368558526,-0.0289837066,0.2667537332,0.262501657,-0.0819980428,0.5508411527,-0.0164088216,-0.0700132251,-0.0953407213,0.2730460763,-0.2342889756,-0.2162806839,0.0255584605,-0.2762463689,-0.1469480991,0.1263189763,0.1495503336,0.0850206688,-0.4425002337,-0.2343565226,0.124324508,0.1346532255,0.0692162588,0.2602708042,0.0762340501,-0.6774932742,0.1283614039,0.0047224485,0.118128933,0.0149708577,-0.0896790922,0.2548965812,-0.0382866524,0.1986334622,-0.1041279435,0.1607329547,-0.068526566,0.1113027036,0.0622030795,-0.2332188487,-0.3385125399,0.0199408345,0.1285448223,0.2212200165,-0.4335190952,-0.0937537402,0.0029252297,0.1343998015,0.0282583013,-0.4729935229,-0.5182256699,-0.0430604555,0.2043706477,-0.1321995109,-0.042318847,0.1051006094,-0.4083647728,-0.173217997,-0.0416861698,-0.2362694591,0.3166966736,0.3083198071,0.2321359515,0.0248305723,0.0248818956,0.0679002628,0.0919112936,0.2281070501,0.3187704384,0.0075668371,-0.2902910113,-0.191925481,-0.196143657,0.02288495,0.2732425034,-0.0638495982,-0.1582741886,-0.6268544197,0.045140367,-0.2855335176,0.2033276707,0.2178163379,-0.0905018225,-0.4762634039,-0.3762312829,0.5591688156,0.1271013767,-0.2329279482,0.3657148182,-0.0431958325,-0.4890762269,0.5494600534,0.3100154102,0.7622916698,-0.2927529514,0.0300721936,0.2363690436,0.0642995238,0.5886327624,0.1454038471,0.4201725125,-0.5156995654,-0.2689718902,-0.0281157624,-0.0444267429,-0.1150463894,0.285641849,-0.4419335127,0.4760812521,0.0140625248,0.1571817845,0.0187443588,0.5202624798,-0.1611385942,-0.0576367341,-0.3130173385,0.0954439566,-0.2046982199,0.3586152792,-0.1418949962,-0.1224704906,0.076926887,0.0946548432,-0.1601141691,0.3148660064,-0.3392430842,-0.0466804467,0.0413837358,-0.2778505981,0.0719876736,0.418651998,0.1913320422,0.2318597436,-0.1648662686,0.0527336709,0.1818175763,0.0941306502,0.1307239681,-0.2407595664,0.4110294282,0.2103672624,0.0368865281,-0.3947605491,0.1993491799,-0.3416984081,-0.491068989,0.09181007,-0.2426339537,0.3264848292,-0.2010719478,0.1421067715,0.0114906672,-0.0262001157,0.1275752634,-0.0468306281,0.2489815652,0.3928253353,-0.012045484,-0.4347582459,-0.0274208393,0.1308655292,0.0268866476,0.28422454,0.3605415225,-0.033177346,-0.0660561174,-0.2653252184,-0.1721902043,-0.5158190727,-0.0947568566,0.2979691625,0.0357195921,0.1108250767,0.1081376448,0.28122136,-0.0383829549,0.0102070346,0.061017219,-0.4703993201,0.2152050883,-0.006740476,-0.0676553473,0.0968679488,-0.0951212868,0.2797602415,-0.0007647623,0.0979794264,-0.2943920791,-0.2169244885,-0.1794053465,0.0438162312,-0.1472646743,-0.0650531352,0.0497813188,0.0632481948,0.0525726601,-0.0284618586,-0.0163565036,-0.1910466105,0.0360109285,0.1280143708,0.0751520172,0.187837407,-0.1114993095,-0.4070796371,-0.1892487705,-0.4432769418,0.0395662822,0.0888449773,0.0292988103,0.1543973535,0.1824588627,0.2711310089,0.193178311,0.1698934287,-0.4526246488,-0.0365547016,0.1458551437,0.1609910727,0.0773922801,0.1830900609,-0.24892582,-0.012200064,0.077471666,0.2076856941,-0.0640993938,-0.216020599,-0.040184658,-0.0609093346,0.1401114464,0.6141610742,0.0422424003,-0.5872521996,0.1972767264,0.1555660218,-0.2832266986,0.1996081024,-0.0918037966,0.4179368317,0.0250828657,0.2798302472,0.2690876722,0.0811951309,-0.1195590869,-0.0003816024,0.3048425317,-0.0419991538,0.3419761658,0.0364516154,0.2121537477,0.038688086,0.1549907178,-0.1711126864,-0.1503050178,0.229683429,0.0648857579,0.2850756645,-0.1867246181,-0.1570701599,0.1521878988,-0.0978184342,0.0413672328,0.7915148735,-0.007006302,0.2106830031,0.388454318,0.4452921152,-0.3818205297,-0.0955146328,-0.5060690641,-0.1313941479,-0.0353573188,-0.0941625163,-0.1014409214,0.0068015708,-0.1248142347,0.0856874883,-0.0818613693,-0.0589715764,0.0174777098,0.1732760221,-0.1736097932,-0.6126018763,0.4208393693,-0.0890995711,0.3969834149,-0.1226720661,0.2771202624,-0.2097463459,0.0411003083,0.1142856106,0.1573393345,0.1676439494,0.7388731241,-0.3403811753,-0.1936333328,-0.2682727873,0.0640630797,-0.0571776293,0.2944566309,-0.0123304836,-0.0872566178,0.4612799585,0.0693922862,-0.0657475218,0.2757764161,0.4123499095,0.0846694335,-0.0382272489,0.4324911237,0.1008390933,0.0294527207,-0.0278444309,0.1407195926,-0.3330128193,-0.1036380306,-0.0233594142,0.0021102647,-0.1246415302,-0.2778750956,0.0576745681,-0.0492705479,0.4179221094,0.288256228,0.1979347467,-0.2882437706,-0.2099870592,-0.573243916,0.2451582551,-0.0229606256,-0.0178858172,-0.2719829082,0.3699620068,-0.3169700801,-0.0046876473,-0.4700049162,0.344781369,-0.1375870258,0.0838931873,-0.593218863,-0.018963147,-0.5143905878,-0.2227605581,-0.0878076553,-0.4162074327,-0.0330266021,-0.2758480012,-0.0626303405,0.0095838672,0.302149117,0.0574647374,0.4434833527,0.2620721161,0.173350811,0.3690351844,-0.0431091227,-0.0240493622,-0.314622432,-0.0322641321,-0.4741529524,0.232226342,0.1099815071,0.1636190265,-0.0451350473,0.3210480511,-0.1569150835,0.5310435295,-0.0042384197,0.1621026397,-0.3588054478,-0.1423979849,-0.1609204262,0.0754326582,-0.1320369691,0.118378751,0.0437265895,0.0036475719,-0.2547435164,-0.2821332812,0.3977829218,-0.2398967296,-0.2634274662,-0.1107814908,0.2522282898,0.1657600105,-0.2589740157,-0.484128207,0.060800489,0.2251031697,0.0316521451,0.1219402701,0.3386109769,0.0220227744,0.1309368461,-0.2010101676,0.121544607,-0.0663709044,-0.3270967305,-0.1218009517,-0.2355503887]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/240","title":"Deterministic dataset loading","comments":"I think we should do the same also for `os.list_dir`","body":"When calling:\r\n```python \r\nimport nlp\r\ndataset = nlp.load_dataset(\"trivia_qa\", split=\"validation[:1%]\")\r\n```\r\n\r\nthe resulting dataset is not deterministic over different google colabs. \r\nAfter talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/datasets\/trivia_qa\/trivia_qa.py#L180\r\n\r\nwhich seems to return an ordering of files that depends on the filesystem:\r\nhttps:\/\/stackoverflow.com\/questions\/6773584\/how-is-pythons-glob-glob-ordered\r\n\r\nI think we should go through all the dataset scripts and make sure to have deterministic behavior.\r\n\r\nA simple solution for `glob.glob()` would be to just replace it with `sorted(glob.glob())` to have everything sorted by name. \r\n\r\nWhat do you think @lhoestq?","comment_length":10,"text":"Deterministic dataset loading\nWhen calling:\r\n```python \r\nimport nlp\r\ndataset = nlp.load_dataset(\"trivia_qa\", split=\"validation[:1%]\")\r\n```\r\n\r\nthe resulting dataset is not deterministic over different google colabs. \r\nAfter talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line:\r\n\r\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/2e0a8639a79b1abc848cff5c669094d40bba0f63\/datasets\/trivia_qa\/trivia_qa.py#L180\r\n\r\nwhich seems to return an ordering of files that depends on the filesystem:\r\nhttps:\/\/stackoverflow.com\/questions\/6773584\/how-is-pythons-glob-glob-ordered\r\n\r\nI think we should go through all the dataset scripts and make sure to have deterministic behavior.\r\n\r\nA simple solution for `glob.glob()` would be to just replace it with `sorted(glob.glob())` to have everything sorted by name. \r\n\r\nWhat do you think @lhoestq?\nI think we should do the same also for `os.list_dir`","embeddings":[0.0676127747,0.1204752475,-0.0174630824,0.1911819428,0.1381106824,-0.1441269517,0.245742172,0.010565985,0.25890854,-0.0011456147,0.072897926,0.4404055178,-0.1960840076,0.0664574057,0.0028411532,0.0097477194,0.023330614,0.209608525,-0.1424403787,-0.2771811485,-0.4136631787,0.0872348323,-0.4051647484,-0.4254803061,-0.1886669099,0.0028008479,-0.1131311283,0.1586726755,-0.0589208044,-0.4831140041,0.1158240288,0.4578330517,0.1553897858,0.231769979,-0.0001173349,-0.2050134987,0.3543515503,-0.0750666857,-0.3871086836,-0.4091807604,-0.3844179809,-0.0675263181,0.1537450254,-0.102655232,-0.2511365414,0.0454044528,0.4765380919,-0.1988138705,0.248959288,0.0882206857,0.1508889049,-0.1054181904,-0.0507281609,0.2201167643,0.0300248154,0.1662745923,-0.0157146733,0.3776882887,0.1216342077,-0.2507028282,-0.0987973884,0.094229579,0.0582893118,0.1180232689,0.4639378488,0.2276519537,-0.1651004404,-0.4076728225,-0.1524830759,0.2377247214,0.5327293277,-0.2271741331,-0.2651467621,-0.349399358,0.2528755963,-0.1996306777,0.1802554131,0.4329480231,-0.1892776042,-0.079970479,-0.0266739037,0.1992829442,-0.1538956612,0.0945633948,0.0427361839,0.6773807406,0.0420284122,-0.0530807637,0.1024855822,0.2500103116,0.2519853115,-0.2838736475,-0.0040165121,0.1334089488,-0.2870635986,-0.1770412475,0.1300590336,0.5201961398,0.1036506146,0.1729948968,0.1729411334,0.048603408,-0.2120702863,0.0253469218,0.3502322137,0.4936121404,-0.0493690595,0.1127336696,0.4221061766,0.0044044303,-0.1063583195,0.3290758133,-0.0011895215,-0.2811889946,0.1026756838,0.1347956657,0.1905919015,-0.1594237238,-0.457235992,-0.2517903745,-0.2654713094,-0.0799845085,0.0293004066,0.2331281751,-0.2537607849,0.1013528109,-0.2307681143,0.1985532045,-0.1107670888,0.0587217361,-0.0445978567,-0.1748357266,-0.3572506607,0.1256128997,0.1453422308,0.1156866327,0.1669139266,0.2820672691,-0.175132513,-0.0214802511,0.1099339575,-0.3788590133,0.0571946241,-0.0153922793,0.1026787013,-0.1275517046,-0.2347695231,-0.147217527,-0.4064051807,0.1363834292,-0.1927141547,-0.0696721375,-0.0451256298,0.1205202192,-0.2758949101,-0.1323228478,-0.3439425528,0.3509619236,0.0590413846,-0.0257197339,-0.0831590667,-0.2365435362,-0.3375295997,-0.3299326003,-0.2106283605,0.5382794738,-0.2951927185,0.1387576014,-0.0104393521,0.2452546358,0.2950542867,0.3874118328,0.028668683,0.2363507599,-0.2728895247,-0.0241659656,0.1837275177,-0.057662122,-0.1331442446,0.1174695566,-0.1896320432,0.2941460907,0.0579771884,0.2801850736,0.0786548778,-0.2705125511,0.1272821277,0.5779412985,0.0893113017,0.1965293586,-0.1964475363,-0.1813393086,0.117638737,0.0775561854,-0.1125515401,0.2245273441,0.0148602091,-0.0906401128,0.3022914231,0.0059622177,-0.3869940042,0.0930059925,0.4922510684,0.0645523742,-0.0266019367,0.1016856506,-0.7249949574,0.1874703169,-0.1082000211,0.2083099782,-0.1483163536,-0.2126608044,0.0408655368,-0.1778492779,-0.3601911366,0.0902945399,0.1201628745,0.4663893878,0.281768024,0.2958360314,0.138745442,0.4146415293,0.0191085506,-0.1260115057,-0.3625060916,0.495364815,-0.0672621876,-0.1745220721,0.2523348629,0.2486473173,-0.0494445413,-0.3091628253,-0.0818125308,0.2968206704,0.2325055152,0.1041861549,0.2044577301,0.0407962203,0.101945281,0.0600683242,0.2240531147,0.3102146685,-0.0779942125,-0.0696729496,-0.0849433765,0.4491431713,-0.1353060454,0.2575736642,0.061757043,-0.098753348,0.0809610784,0.0548459664,0.0547176339,-0.2337993979,0.2563711703,0.2752789855,0.4400291145,0.4657376111,-0.6469978094,0.2546949983,0.695559144,0.0421611704,0.1260314137,-0.0058973008,0.1873812526,-0.2641454339,0.2570747137,0.259453088,0.5881262422,0.1732174456,-0.0667833388,-0.0388625041,-0.1474103481,-0.2594450414,0.0586338006,-0.0025682172,0.1635771543,0.115889512,0.313755393,0.1545712948,-0.2859836221,-0.2004216313,-0.1046644971,0.0264418107,-0.0716638491,0.1482740194,-0.2354159206,-0.3257488608,-0.3718807995,-0.0853883922,-0.3342889547,-0.2048602402,0.0820932835,0.2134458274,0.0317665301,-0.0653247088,0.1831926852,0.2862431705,-0.033051908,-0.4692470431,-0.1943894178,-0.0249420535,-0.3590497077,-0.0564156696,0.477769196,0.2189507484,0.4552116394,-0.2010077536,-0.2036304772,0.0709943101,-0.3020977676,-0.0647454709,-0.2588724494,0.3009795547,-0.0372950509,-0.1481919438,-0.0349212289,-0.0522904173,-0.0371295027,0.0711836517,-0.0976509973,0.0052774767,0.0783758983,0.0090905922,-0.0577669442,-0.2997931242,-0.3189409375,-0.1757821441,0.0894751698,0.325665921,0.204213351,0.2679585218,-0.0870939717,-0.2577173114,0.1225666925,0.0559767522,-0.0361046381,-0.040765278,0.3400525749,-0.1827248037,-0.2083619237,-0.3892098367,-0.1972963363,-0.1285570264,0.1412935704,-0.3359766901,-0.3896195889,-0.3791182637,0.1010690108,-0.0703453124,-0.1874082386,0.3972537816,0.0787701681,0.0757751912,0.1538622379,-0.0110424394,0.2131811827,0.3018565476,0.3497726619,0.2366679162,0.0525389276,-0.0307400096,0.4713709652,0.0078851245,-0.4118443727,0.2272130549,0.1276096702,0.4077757895,-0.4115252197,0.0077406033,0.2415744215,0.3345225155,-0.1865605265,0.5477392077,-0.0281014219,-0.0742010623,-0.0190230999,0.3033543229,-0.2361363769,-0.2174751461,0.0556038283,-0.208275646,-0.2180844098,0.0374848694,0.1198203415,0.0556048378,-0.3526769876,-0.2360737324,0.2343190312,0.0879098922,0.108742483,0.2383951396,-0.0157440323,-0.6332278252,0.1143422499,-0.0025411418,0.0441102907,0.0698031709,-0.1419409364,0.2698062062,-0.0789229423,0.227145642,0.0018772074,0.1955931634,-0.0767468885,0.0794521496,0.1391445696,-0.283023268,-0.4153654575,0.0321455561,0.2209936231,0.2741949558,-0.4228038788,-0.1758719981,-0.0368038155,0.1785915047,0.0101873167,-0.5045804977,-0.4808104038,-0.0433028638,0.1987638921,-0.077638112,-0.0660080835,0.0439601019,-0.4369384348,-0.1880790889,-0.0551873334,-0.2663452923,0.215230003,0.2813841105,0.2680218518,0.0392722972,0.0362558626,0.0346394442,0.1902563572,0.2386232018,0.3561398089,-0.1320929527,-0.1777936965,-0.1839969307,-0.1574589759,-0.0378115475,0.2645226419,-0.0825859606,-0.2615571022,-0.6986196637,0.1129670367,-0.2518010139,0.2295775712,0.2172800153,-0.0277450196,-0.4461124539,-0.4245752692,0.5257621408,0.0945244357,-0.3399841189,0.4048508108,-0.047508765,-0.5039141774,0.5866379738,0.3134778142,0.7254549265,-0.2451274395,0.1454611868,0.3072448373,0.047667481,0.5500716567,0.143512696,0.4115539193,-0.5155009031,-0.2189630121,-0.0105029475,-0.1026752293,-0.0998997912,0.3241791427,-0.4445556104,0.4937404096,0.0816062689,0.0881265774,-0.0632434636,0.5105524659,-0.1695258617,-0.066612564,-0.3526331186,0.0531848483,-0.1085045263,0.3916579485,-0.2202751338,-0.1368001401,0.037677668,0.094388403,-0.0234888867,0.2596101761,-0.3688839972,0.0177028365,-0.0507030152,-0.3148470521,-0.000103056,0.4979455173,0.1266865581,0.2235281467,-0.0876186863,-0.0242538657,0.1685979515,0.207799226,0.1161563173,-0.247863099,0.3561452627,0.205753684,0.0354056172,-0.4921530187,0.1684441566,-0.3544316292,-0.4827631116,0.0853233784,-0.2584625185,0.3286777735,-0.1824278235,0.1751676351,-0.0092853811,0.0313890651,0.1003856212,0.0222648252,0.1770431697,0.4828917682,-0.1011147797,-0.4468161464,-0.0593978092,0.0758706704,0.0074765994,0.2661483884,0.3299107552,-0.1101291776,-0.0497225747,-0.2717190385,-0.1167534068,-0.4754927456,-0.0446551256,0.2578479052,0.0263564289,0.0915487558,0.1098831221,0.2464317828,0.0679905415,0.1184967458,0.0011822777,-0.4794232845,0.144156307,0.0084475037,-0.0649425983,0.1987503022,-0.0841582641,0.2053542584,0.0067710895,-0.0558721311,-0.2840541601,-0.1967890859,-0.1257465333,-0.015571313,-0.1692900211,-0.1176172495,0.1296312213,0.1315053105,0.0591150261,-0.1293962747,-0.080186598,-0.1700241566,0.0993927866,0.1283450276,0.0595964715,0.1808951646,-0.1222551987,-0.4250262082,-0.1110680327,-0.423316896,0.0869497508,0.0424898118,0.0491835847,0.1156355888,0.2615351379,0.2508137524,0.2112931609,0.2282256782,-0.4489279389,-0.0825298652,0.1402309537,0.2582982481,0.0465523079,0.1748823225,-0.1922614425,-0.0179167036,0.039166335,0.2845885456,-0.0705205798,-0.1623019278,-0.0417990498,-0.0919253752,0.2162376493,0.6603098512,0.0350710712,-0.6028288007,0.2188334018,0.1520504504,-0.2466173321,0.135587126,-0.0580469854,0.4319244027,0.0526980981,0.3563454747,0.266561836,0.1517220289,-0.1865209788,-0.0123151522,0.338814646,0.0599604957,0.3108624518,0.1614424139,0.1431802064,0.0481298827,0.1484708339,-0.1303227693,-0.1375814229,0.3049200773,-0.002479729,0.388061434,-0.1824528128,-0.1844075769,0.223432824,-0.1457731426,0.0041503864,0.733371377,0.0090870317,0.1433571577,0.342522949,0.3846418858,-0.3772748709,-0.1416930258,-0.4557829797,-0.1006794125,-0.0796555802,-0.0969317555,-0.1214003861,0.0220321193,-0.0700838789,0.0857587308,-0.0801992491,-0.0969088674,-0.0445695333,0.1942164302,-0.1309361458,-0.5106114149,0.3783435524,-0.0199544616,0.2922474146,-0.0813006833,0.28884238,-0.170119077,0.0884769931,0.064039804,0.1217122972,0.1648711562,0.7841171026,-0.3358785212,-0.2352009416,-0.1565757245,0.0881577432,-0.1621692032,0.1632869542,-0.0911418423,-0.0682322606,0.4752748609,0.0659388676,-0.0644207671,0.3580305278,0.345177561,-0.0264731888,-0.0091912784,0.4578896761,0.1401355267,0.023501765,-0.0082748327,0.1838605702,-0.4038133323,-0.1521533877,-0.0448149852,0.0318588093,-0.1084732264,-0.202456966,0.0522795394,-0.0995190442,0.3462148011,0.2262452841,0.2445534766,-0.2708851397,-0.1970269531,-0.6054341793,0.1865983009,0.0361955427,-0.0919485912,-0.2852300107,0.3733628392,-0.3746699095,-0.0271369927,-0.3345971704,0.3240602911,-0.1209401116,0.0953240097,-0.5969626904,-0.0585291982,-0.4772935808,-0.2180739641,-0.0759399235,-0.4648697078,-0.0902112275,-0.2088631094,-0.0599204563,-0.0401092172,0.3879909217,0.0307460483,0.419326663,0.3125143051,0.1698165238,0.3232603669,-0.0562496856,-0.0583361238,-0.3097096682,-0.0762499198,-0.4747547507,0.191297859,0.0992658138,0.1747409999,-0.1311907023,0.2783082128,-0.1838410944,0.4383296072,0.0187936388,0.1414869577,-0.2830063701,-0.1430493593,-0.1794183999,0.0661475137,-0.0722757056,0.1557535231,0.088846691,0.027473107,-0.2477003187,-0.2432231307,0.3702276349,-0.2058640271,-0.202639848,-0.1554628313,0.248641625,0.1364219785,-0.2524937987,-0.4131936729,0.0492477939,0.2359316647,0.0027211243,0.1043982208,0.315731287,0.0506600514,0.1342255324,-0.23735632,0.0713419989,-0.0963396281,-0.3574833572,-0.0947180092,-0.2265079916]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/239","title":"[Creating new dataset] Not found dataset_info.json","comments":"I think you can just `rm` this directory and it should be good :)","body":"Hi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n","comment_length":14,"text":"[Creating new dataset] Not found dataset_info.json\nHi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n\nI think you can just `rm` this directory and it should be good :)","embeddings":[-0.1165063456,0.1242502332,-0.0860032439,0.057063479,0.0176752172,0.1466908902,0.1669220626,0.3094637394,-0.2007142454,0.0218163859,0.0602891333,0.5790073276,-0.0935086533,0.0841157287,0.2967385054,-0.0632679313,0.0798937827,0.3547498286,0.4398046434,-0.1709556282,-0.3041882515,0.3338649273,-0.0877133608,0.01387031,-0.338992089,-0.1399389803,-0.1987462044,0.1517356932,-0.3689015806,-0.2440757155,0.4867274761,-0.0733882189,0.1036610082,0.5475575328,-0.0001140775,0.0102124736,0.3725022972,-0.1520183682,-0.463706404,-0.5580253005,-0.0097940266,-0.4027791023,0.1149742603,-0.3410841227,-0.1230617836,-0.3242848516,0.2014034241,-0.1395288706,-0.1161304712,0.4508829117,0.1798371077,0.0618324205,0.184210822,-0.0867190957,0.0001254161,0.2762156427,-0.2401407212,0.2098938376,0.5791471601,-0.03125735,0.2688177526,-0.0465215221,-0.0659272224,-0.2932347655,0.2591234148,0.159864217,0.0156177217,-0.4965935946,0.0261633452,0.0796523169,0.7736666799,-0.5069001317,-0.3475548625,0.0792120323,0.194891125,0.0815229118,0.0261326935,0.3870602548,-0.0036244129,-0.1468177885,-0.4909806252,-0.3836780488,-0.1394308358,0.3541182876,0.2187725157,0.2453514934,-0.0337953642,0.0422469042,-0.0160155166,-0.0535681583,-0.1302879304,-0.1859699339,-0.1286573559,0.2532887161,-0.053666506,-0.5121890903,0.1687165648,-0.0076868557,-0.0895367935,0.061128173,-0.0326223448,-0.274641186,-0.2865626812,0.1832764745,0.2046544701,-0.1106357574,0.279658407,0.0797307193,0.0774832591,0.055838082,0.0785688013,-0.0678011626,-0.0335660912,0.0156166032,-0.1277638078,-0.0994326547,0.3245657682,-0.0741777495,0.0718602091,0.0012805207,-0.1981693804,-0.133311823,0.1172707081,0.2936829925,-0.0719582215,-0.2800116837,-0.0251497515,0.2071141303,-0.4114207625,-0.0116609484,-0.1645233035,0.4944566786,-0.2056022882,-0.0262290053,0.3875793517,0.1016707048,0.3338846862,-0.0130871162,-0.368738085,-0.1556921601,0.3918465078,-0.2121744305,-0.1790122241,0.228953436,0.1160459667,0.1139768213,-0.0378630012,-0.3668879271,-0.3832911253,0.1916641742,-0.2350608408,-0.3481474817,-0.283872366,0.1933897138,0.0613683388,-0.2342845798,-0.3146072924,0.2751885056,0.3097748458,-0.3435207307,0.0633967742,0.0758629963,-0.1343005598,-0.2305130363,-0.0999691784,0.3998726904,-0.4836797416,-0.0721997991,-0.3215029836,-0.0094186161,0.1910135746,-0.0352825522,-0.2096750736,0.5172001123,-0.1338477433,0.1112853959,0.5153878927,-0.1893327534,-0.187461406,0.2695929706,-0.1940307021,-0.2928598523,0.0725557357,0.2507296503,-0.0414150171,-0.1687862128,0.1327093393,0.0917857811,-0.126192376,0.0494734347,0.0835390836,-0.0060451939,0.0510402881,0.1193809733,-0.2383529991,0.1086443737,0.1355073601,0.2590746582,0.097801879,-0.0312276371,0.1779914796,0.4847249985,0.3885298371,-0.0059467414,-0.0228138771,-0.0708329007,-0.3362148404,-0.0818720236,-0.2369996756,0.4043165743,0.0337655842,-0.002795812,-0.4716331661,-0.192667827,0.0334816724,-0.1176143065,0.0850239992,0.0162108652,0.3978447318,0.1003789753,-0.2538137734,0.0775243267,-0.1711559445,0.0665230826,-0.8675051332,0.3471957743,-0.2013229132,-0.2094378471,0.2470538467,0.2310829163,-0.0176213309,-0.1472965032,0.1502861083,0.2265686393,-0.1452424228,0.3361055255,0.2227082998,0.128842622,0.2885342836,-0.2038090229,0.1463573426,0.0493759997,0.1962344646,-0.0217218064,-0.0831139833,0.0535558686,0.1434141099,-0.0550300702,0.1617177576,-0.0439132974,0.1402127743,-0.1786086708,0.0769875199,-0.2421171665,0.1755624563,0.075510703,0.131608218,0.048625093,-0.0772270411,0.2307497859,0.3067060411,0.0531004742,0.0431879945,0.157994464,0.0398811735,-0.0914959162,-0.1635780036,0.2864362597,0.4337894619,0.1305443794,-0.1725736558,-0.0445974953,-0.1054399312,-0.2338734865,0.1762125492,-0.144586727,0.1815698743,0.0553759448,0.2064576298,-0.091882892,-0.1981725991,-0.4503073096,0.2582753003,0.2915617824,0.0125326188,0.2330195606,-0.0483596958,-0.2889663279,-0.1023572162,-0.2199970037,-0.027930662,-0.1097455919,0.0449513085,0.2148979455,0.2119145691,-0.1206713319,-0.1959022731,0.1891421676,-0.0882790238,-0.3575147688,-0.0885614678,-0.2564783394,-0.1869902909,0.1057349443,0.0348248258,0.1400998682,0.2495245636,-0.1843912452,0.0105685499,-0.1660453826,-0.2570820153,0.0180810299,-0.1595323384,0.19502756,0.1740631312,0.0597916804,-0.2608325481,-0.4254710078,0.0775657073,-0.0116267065,-0.4032616317,0.0009183167,-0.1733697355,-0.1220971048,-0.0407483503,-0.711997509,-0.3251382411,-0.213160634,-0.0155927436,0.1631813794,0.1977260411,0.4130190015,-0.0757454857,0.1082077697,0.1723190397,0.2316488922,-0.0692338049,-0.3634283841,0.3943191171,-0.5005172491,-0.2842501998,0.1889234185,-0.4618218243,0.2375343293,-0.1741824299,-0.5138138533,0.0168081373,-0.2980986834,0.0059204972,0.1343669146,0.2339942753,0.4963957071,-0.0279377773,-0.053906247,-0.1033315063,-0.4580044746,0.0228570011,0.4239569306,0.6483789682,-0.0955790132,0.1766877472,-0.4366553128,0.1395106316,0.3806612194,-0.0229976848,0.4596581459,-0.0642139837,0.2842175663,-0.2890721262,-0.3776216805,0.2021721154,0.2846024036,-0.0968416631,0.1915385127,0.0889659151,-0.1900496632,-0.1813447624,-0.1297150552,-0.4387228489,-0.0670411065,-0.1606695354,-0.0154463137,0.2671151459,0.0893691704,-0.063767001,-0.1456261724,-0.2757143676,-0.2235910594,0.4727536738,0.1974101365,0.1208718792,-0.2766308486,-0.0612071976,-0.307307303,-0.0377948396,0.0238852985,0.2937771976,-0.202954337,0.255422622,0.0629198477,-0.0123398248,0.5490016341,-0.2908591032,-0.2055511624,0.2633007467,-0.0580418445,-0.2241145223,0.1805141866,-0.1040338427,0.6076475382,0.2433528602,0.5753422379,-0.3115656376,0.2990731299,0.1214507297,0.3701847196,-0.2317983359,-0.2861895859,-0.1955006421,-0.4630236924,-0.1389040947,-0.3091448843,-0.3363452852,0.4752550423,0.2303315699,0.170401454,0.3116016984,-0.1196064502,0.2841059268,0.1153542474,0.2576376796,0.2335043997,0.4246392846,-0.1052278578,-0.0671053752,0.3669828773,0.6004720926,0.2031099647,-0.0956731215,0.1284203976,-0.0364036411,0.395367533,-0.0132449679,0.266926378,-0.1098193899,0.0210946109,-0.2184128016,-0.115531534,-0.0234098081,-0.0193970408,-0.2241461277,-0.6888647079,-0.3939317465,0.2423800826,-0.0106856376,0.3002980351,0.2865825295,0.1424811333,-0.2200430632,0.471485883,-0.2134755552,0.8308844566,-0.0526556,-0.0279705022,0.0454090759,0.1510666609,0.7034148574,-0.5142815709,0.4126512706,-0.3891026974,-0.0269086305,0.0515783466,-0.1758003682,0.1750633717,0.4178828299,-0.4325260818,0.1930044442,-0.0822679177,-0.1185833812,-0.1702666879,0.4646313787,-0.1232812628,-0.3681624234,-0.3878662288,0.1243901253,-0.1559613645,0.3885234296,-0.0936821178,-0.0173259676,-0.5373473763,-0.0237672329,-0.2104974687,-0.0220599901,0.3186563551,0.249805361,0.5425424576,-0.3958864212,0.1680613905,0.1537734866,0.5115233064,0.3233711123,-0.2130830884,-0.0431414843,-0.0085645812,-0.4150432348,0.0861745626,0.0233168025,0.3259398341,0.0027130737,-0.2486928701,0.1554380059,-0.0910228491,-0.0530705713,-0.0971550718,-0.1465338618,-0.2417299896,-0.2759905159,-0.3625261486,0.2587006092,-0.1688533127,-0.1415452212,0.0788361356,0.2971162498,-0.1353914887,0.1185149401,0.0202571284,-0.1812777519,-0.2206898779,0.23206532,0.3509170711,-0.1711783409,0.4593079388,0.2199606895,-0.0807433948,-0.2345264852,0.1070886403,0.1253063977,-0.1298018247,0.2112330049,0.3616847694,0.1504724026,0.2839156985,0.3608105481,-0.166363582,0.1632482409,0.0730139166,-0.3862999678,-0.436652571,0.2469571084,0.0244277362,0.0901544467,0.1384783983,0.3877499104,0.2723523378,0.035061609,-0.2648818195,-0.0455673523,-0.3696907163,0.2637805939,0.2491617203,0.0605015419,-0.2160846889,-0.057571128,0.0510149114,-0.2032997012,-0.0290426463,-0.1544078141,-0.2216129601,0.1520265043,0.1260631382,0.0154480394,-0.1245804504,-0.2296669036,-0.2332468629,-0.0199401714,-0.2605073452,0.1780531257,0.1019101664,0.1546177417,0.0935981572,-0.3269782066,-0.1073557809,-0.0148436576,-0.0619534105,0.3862453401,0.0843533278,0.0452335738,0.1161664277,0.0425391942,-0.4028134346,0.0701125935,-0.1516933888,0.0837625638,0.1490865946,-0.157145083,0.2667430937,0.0217986628,0.1099834368,0.3466070592,-0.1359733343,0.1405175328,0.260553211,0.1115451902,-0.1620906889,-0.0440804772,0.3279460669,0.0434803776,-0.0122272819,0.3146601915,-0.1331052035,0.131069079,-0.1478780508,0.1849879473,0.3246780336,-0.0631473958,0.2577349842,0.1364097297,0.0012924885,-0.0977015942,0.2594444156,0.1881525815,-0.0853210315,0.1823779792,0.0450893342,0.3973941207,0.302352041,0.2861189544,0.2109260857,-0.2249996215,0.0504466966,-0.2021240294,0.2507336736,-0.0419258364,0.0826024935,0.2187250257,-0.1958678067,0.3018075824,-0.1780408919,0.1030380204,-0.3208485246,-0.0656001419,-0.1125702336,-0.2674712539,0.2546589375,-0.0874804184,0.0292556714,-0.1072182357,-0.1549135,-0.0531782545,-0.0556121841,-0.327442795,0.4158905149,-0.014811635,0.0209267847,-0.1005132124,0.3173038661,-0.1687464118,-0.1228318959,-0.1224116385,0.3447618186,0.307612896,0.3469654322,0.1381114572,0.2645431161,-0.0776175037,-0.1906177551,0.1139080673,0.4997584224,0.4584063292,-0.1055591255,0.0251274351,0.0590283573,-0.0244646985,0.1337961853,0.5156471133,0.0381495841,0.1663969457,0.4495267272,-0.1199095845,-0.1236041114,-0.2566175163,0.2996979356,-0.4078597426,-0.1276357025,0.5206471682,0.0403465368,-0.0090243537,-0.3289625347,0.0792628452,-0.0467726104,0.2938723564,0.3701539338,0.1201424971,-0.262578249,-0.2484817952,-0.3296501338,-0.0624068305,-0.111139901,-0.1620778441,-0.2252211571,0.3631657958,0.0261462871,0.1804602295,-0.362882942,0.2894383967,-0.1485565156,0.0748760626,-0.1408980638,-0.338190645,-0.1541633755,0.1900706887,-0.0819498003,-0.0871439874,-0.0184988342,-0.2756521106,-0.0113097196,-0.432760179,-0.2158632725,0.2175652385,0.3473265171,0.1009334028,0.0736395046,0.3765215874,-0.2977442741,-0.1872087121,-0.1939508617,-0.2278358787,-0.0951205119,0.2785026729,0.2827716768,0.37742576,-0.301546216,0.3614298999,-0.1413934827,0.0988541469,-0.030947797,0.1174751371,-0.0910616368,0.0121015664,-0.2916093171,-0.0935944244,0.0717773363,0.2724191248,0.0033911539,0.3202648759,-0.0997141823,-0.449174583,0.4673528671,-0.2092696279,-0.188631773,-0.0279480461,0.0273697264,-0.2985510826,-0.2551029325,-0.2464308143,0.0702252463,0.3553676307,-0.2813268304,0.0182162989,0.0918259546,-0.1916161627,0.0478141867,-0.0269898754,0.1780251116,0.103638418,0.0088643283,-0.1135134101,0.0419382639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/239","title":"[Creating new dataset] Not found dataset_info.json","comments":"@lhoestq - this seems to happen quite often (already the 2nd issue). Can we maybe delete this automatically?","body":"Hi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n","comment_length":18,"text":"[Creating new dataset] Not found dataset_info.json\nHi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n\n@lhoestq - this seems to happen quite often (already the 2nd issue). Can we maybe delete this automatically?","embeddings":[-0.1165063456,0.1242502332,-0.0860032439,0.057063479,0.0176752172,0.1466908902,0.1669220626,0.3094637394,-0.2007142454,0.0218163859,0.0602891333,0.5790073276,-0.0935086533,0.0841157287,0.2967385054,-0.0632679313,0.0798937827,0.3547498286,0.4398046434,-0.1709556282,-0.3041882515,0.3338649273,-0.0877133608,0.01387031,-0.338992089,-0.1399389803,-0.1987462044,0.1517356932,-0.3689015806,-0.2440757155,0.4867274761,-0.0733882189,0.1036610082,0.5475575328,-0.0001140775,0.0102124736,0.3725022972,-0.1520183682,-0.463706404,-0.5580253005,-0.0097940266,-0.4027791023,0.1149742603,-0.3410841227,-0.1230617836,-0.3242848516,0.2014034241,-0.1395288706,-0.1161304712,0.4508829117,0.1798371077,0.0618324205,0.184210822,-0.0867190957,0.0001254161,0.2762156427,-0.2401407212,0.2098938376,0.5791471601,-0.03125735,0.2688177526,-0.0465215221,-0.0659272224,-0.2932347655,0.2591234148,0.159864217,0.0156177217,-0.4965935946,0.0261633452,0.0796523169,0.7736666799,-0.5069001317,-0.3475548625,0.0792120323,0.194891125,0.0815229118,0.0261326935,0.3870602548,-0.0036244129,-0.1468177885,-0.4909806252,-0.3836780488,-0.1394308358,0.3541182876,0.2187725157,0.2453514934,-0.0337953642,0.0422469042,-0.0160155166,-0.0535681583,-0.1302879304,-0.1859699339,-0.1286573559,0.2532887161,-0.053666506,-0.5121890903,0.1687165648,-0.0076868557,-0.0895367935,0.061128173,-0.0326223448,-0.274641186,-0.2865626812,0.1832764745,0.2046544701,-0.1106357574,0.279658407,0.0797307193,0.0774832591,0.055838082,0.0785688013,-0.0678011626,-0.0335660912,0.0156166032,-0.1277638078,-0.0994326547,0.3245657682,-0.0741777495,0.0718602091,0.0012805207,-0.1981693804,-0.133311823,0.1172707081,0.2936829925,-0.0719582215,-0.2800116837,-0.0251497515,0.2071141303,-0.4114207625,-0.0116609484,-0.1645233035,0.4944566786,-0.2056022882,-0.0262290053,0.3875793517,0.1016707048,0.3338846862,-0.0130871162,-0.368738085,-0.1556921601,0.3918465078,-0.2121744305,-0.1790122241,0.228953436,0.1160459667,0.1139768213,-0.0378630012,-0.3668879271,-0.3832911253,0.1916641742,-0.2350608408,-0.3481474817,-0.283872366,0.1933897138,0.0613683388,-0.2342845798,-0.3146072924,0.2751885056,0.3097748458,-0.3435207307,0.0633967742,0.0758629963,-0.1343005598,-0.2305130363,-0.0999691784,0.3998726904,-0.4836797416,-0.0721997991,-0.3215029836,-0.0094186161,0.1910135746,-0.0352825522,-0.2096750736,0.5172001123,-0.1338477433,0.1112853959,0.5153878927,-0.1893327534,-0.187461406,0.2695929706,-0.1940307021,-0.2928598523,0.0725557357,0.2507296503,-0.0414150171,-0.1687862128,0.1327093393,0.0917857811,-0.126192376,0.0494734347,0.0835390836,-0.0060451939,0.0510402881,0.1193809733,-0.2383529991,0.1086443737,0.1355073601,0.2590746582,0.097801879,-0.0312276371,0.1779914796,0.4847249985,0.3885298371,-0.0059467414,-0.0228138771,-0.0708329007,-0.3362148404,-0.0818720236,-0.2369996756,0.4043165743,0.0337655842,-0.002795812,-0.4716331661,-0.192667827,0.0334816724,-0.1176143065,0.0850239992,0.0162108652,0.3978447318,0.1003789753,-0.2538137734,0.0775243267,-0.1711559445,0.0665230826,-0.8675051332,0.3471957743,-0.2013229132,-0.2094378471,0.2470538467,0.2310829163,-0.0176213309,-0.1472965032,0.1502861083,0.2265686393,-0.1452424228,0.3361055255,0.2227082998,0.128842622,0.2885342836,-0.2038090229,0.1463573426,0.0493759997,0.1962344646,-0.0217218064,-0.0831139833,0.0535558686,0.1434141099,-0.0550300702,0.1617177576,-0.0439132974,0.1402127743,-0.1786086708,0.0769875199,-0.2421171665,0.1755624563,0.075510703,0.131608218,0.048625093,-0.0772270411,0.2307497859,0.3067060411,0.0531004742,0.0431879945,0.157994464,0.0398811735,-0.0914959162,-0.1635780036,0.2864362597,0.4337894619,0.1305443794,-0.1725736558,-0.0445974953,-0.1054399312,-0.2338734865,0.1762125492,-0.144586727,0.1815698743,0.0553759448,0.2064576298,-0.091882892,-0.1981725991,-0.4503073096,0.2582753003,0.2915617824,0.0125326188,0.2330195606,-0.0483596958,-0.2889663279,-0.1023572162,-0.2199970037,-0.027930662,-0.1097455919,0.0449513085,0.2148979455,0.2119145691,-0.1206713319,-0.1959022731,0.1891421676,-0.0882790238,-0.3575147688,-0.0885614678,-0.2564783394,-0.1869902909,0.1057349443,0.0348248258,0.1400998682,0.2495245636,-0.1843912452,0.0105685499,-0.1660453826,-0.2570820153,0.0180810299,-0.1595323384,0.19502756,0.1740631312,0.0597916804,-0.2608325481,-0.4254710078,0.0775657073,-0.0116267065,-0.4032616317,0.0009183167,-0.1733697355,-0.1220971048,-0.0407483503,-0.711997509,-0.3251382411,-0.213160634,-0.0155927436,0.1631813794,0.1977260411,0.4130190015,-0.0757454857,0.1082077697,0.1723190397,0.2316488922,-0.0692338049,-0.3634283841,0.3943191171,-0.5005172491,-0.2842501998,0.1889234185,-0.4618218243,0.2375343293,-0.1741824299,-0.5138138533,0.0168081373,-0.2980986834,0.0059204972,0.1343669146,0.2339942753,0.4963957071,-0.0279377773,-0.053906247,-0.1033315063,-0.4580044746,0.0228570011,0.4239569306,0.6483789682,-0.0955790132,0.1766877472,-0.4366553128,0.1395106316,0.3806612194,-0.0229976848,0.4596581459,-0.0642139837,0.2842175663,-0.2890721262,-0.3776216805,0.2021721154,0.2846024036,-0.0968416631,0.1915385127,0.0889659151,-0.1900496632,-0.1813447624,-0.1297150552,-0.4387228489,-0.0670411065,-0.1606695354,-0.0154463137,0.2671151459,0.0893691704,-0.063767001,-0.1456261724,-0.2757143676,-0.2235910594,0.4727536738,0.1974101365,0.1208718792,-0.2766308486,-0.0612071976,-0.307307303,-0.0377948396,0.0238852985,0.2937771976,-0.202954337,0.255422622,0.0629198477,-0.0123398248,0.5490016341,-0.2908591032,-0.2055511624,0.2633007467,-0.0580418445,-0.2241145223,0.1805141866,-0.1040338427,0.6076475382,0.2433528602,0.5753422379,-0.3115656376,0.2990731299,0.1214507297,0.3701847196,-0.2317983359,-0.2861895859,-0.1955006421,-0.4630236924,-0.1389040947,-0.3091448843,-0.3363452852,0.4752550423,0.2303315699,0.170401454,0.3116016984,-0.1196064502,0.2841059268,0.1153542474,0.2576376796,0.2335043997,0.4246392846,-0.1052278578,-0.0671053752,0.3669828773,0.6004720926,0.2031099647,-0.0956731215,0.1284203976,-0.0364036411,0.395367533,-0.0132449679,0.266926378,-0.1098193899,0.0210946109,-0.2184128016,-0.115531534,-0.0234098081,-0.0193970408,-0.2241461277,-0.6888647079,-0.3939317465,0.2423800826,-0.0106856376,0.3002980351,0.2865825295,0.1424811333,-0.2200430632,0.471485883,-0.2134755552,0.8308844566,-0.0526556,-0.0279705022,0.0454090759,0.1510666609,0.7034148574,-0.5142815709,0.4126512706,-0.3891026974,-0.0269086305,0.0515783466,-0.1758003682,0.1750633717,0.4178828299,-0.4325260818,0.1930044442,-0.0822679177,-0.1185833812,-0.1702666879,0.4646313787,-0.1232812628,-0.3681624234,-0.3878662288,0.1243901253,-0.1559613645,0.3885234296,-0.0936821178,-0.0173259676,-0.5373473763,-0.0237672329,-0.2104974687,-0.0220599901,0.3186563551,0.249805361,0.5425424576,-0.3958864212,0.1680613905,0.1537734866,0.5115233064,0.3233711123,-0.2130830884,-0.0431414843,-0.0085645812,-0.4150432348,0.0861745626,0.0233168025,0.3259398341,0.0027130737,-0.2486928701,0.1554380059,-0.0910228491,-0.0530705713,-0.0971550718,-0.1465338618,-0.2417299896,-0.2759905159,-0.3625261486,0.2587006092,-0.1688533127,-0.1415452212,0.0788361356,0.2971162498,-0.1353914887,0.1185149401,0.0202571284,-0.1812777519,-0.2206898779,0.23206532,0.3509170711,-0.1711783409,0.4593079388,0.2199606895,-0.0807433948,-0.2345264852,0.1070886403,0.1253063977,-0.1298018247,0.2112330049,0.3616847694,0.1504724026,0.2839156985,0.3608105481,-0.166363582,0.1632482409,0.0730139166,-0.3862999678,-0.436652571,0.2469571084,0.0244277362,0.0901544467,0.1384783983,0.3877499104,0.2723523378,0.035061609,-0.2648818195,-0.0455673523,-0.3696907163,0.2637805939,0.2491617203,0.0605015419,-0.2160846889,-0.057571128,0.0510149114,-0.2032997012,-0.0290426463,-0.1544078141,-0.2216129601,0.1520265043,0.1260631382,0.0154480394,-0.1245804504,-0.2296669036,-0.2332468629,-0.0199401714,-0.2605073452,0.1780531257,0.1019101664,0.1546177417,0.0935981572,-0.3269782066,-0.1073557809,-0.0148436576,-0.0619534105,0.3862453401,0.0843533278,0.0452335738,0.1161664277,0.0425391942,-0.4028134346,0.0701125935,-0.1516933888,0.0837625638,0.1490865946,-0.157145083,0.2667430937,0.0217986628,0.1099834368,0.3466070592,-0.1359733343,0.1405175328,0.260553211,0.1115451902,-0.1620906889,-0.0440804772,0.3279460669,0.0434803776,-0.0122272819,0.3146601915,-0.1331052035,0.131069079,-0.1478780508,0.1849879473,0.3246780336,-0.0631473958,0.2577349842,0.1364097297,0.0012924885,-0.0977015942,0.2594444156,0.1881525815,-0.0853210315,0.1823779792,0.0450893342,0.3973941207,0.302352041,0.2861189544,0.2109260857,-0.2249996215,0.0504466966,-0.2021240294,0.2507336736,-0.0419258364,0.0826024935,0.2187250257,-0.1958678067,0.3018075824,-0.1780408919,0.1030380204,-0.3208485246,-0.0656001419,-0.1125702336,-0.2674712539,0.2546589375,-0.0874804184,0.0292556714,-0.1072182357,-0.1549135,-0.0531782545,-0.0556121841,-0.327442795,0.4158905149,-0.014811635,0.0209267847,-0.1005132124,0.3173038661,-0.1687464118,-0.1228318959,-0.1224116385,0.3447618186,0.307612896,0.3469654322,0.1381114572,0.2645431161,-0.0776175037,-0.1906177551,0.1139080673,0.4997584224,0.4584063292,-0.1055591255,0.0251274351,0.0590283573,-0.0244646985,0.1337961853,0.5156471133,0.0381495841,0.1663969457,0.4495267272,-0.1199095845,-0.1236041114,-0.2566175163,0.2996979356,-0.4078597426,-0.1276357025,0.5206471682,0.0403465368,-0.0090243537,-0.3289625347,0.0792628452,-0.0467726104,0.2938723564,0.3701539338,0.1201424971,-0.262578249,-0.2484817952,-0.3296501338,-0.0624068305,-0.111139901,-0.1620778441,-0.2252211571,0.3631657958,0.0261462871,0.1804602295,-0.362882942,0.2894383967,-0.1485565156,0.0748760626,-0.1408980638,-0.338190645,-0.1541633755,0.1900706887,-0.0819498003,-0.0871439874,-0.0184988342,-0.2756521106,-0.0113097196,-0.432760179,-0.2158632725,0.2175652385,0.3473265171,0.1009334028,0.0736395046,0.3765215874,-0.2977442741,-0.1872087121,-0.1939508617,-0.2278358787,-0.0951205119,0.2785026729,0.2827716768,0.37742576,-0.301546216,0.3614298999,-0.1413934827,0.0988541469,-0.030947797,0.1174751371,-0.0910616368,0.0121015664,-0.2916093171,-0.0935944244,0.0717773363,0.2724191248,0.0033911539,0.3202648759,-0.0997141823,-0.449174583,0.4673528671,-0.2092696279,-0.188631773,-0.0279480461,0.0273697264,-0.2985510826,-0.2551029325,-0.2464308143,0.0702252463,0.3553676307,-0.2813268304,0.0182162989,0.0918259546,-0.1916161627,0.0478141867,-0.0269898754,0.1780251116,0.103638418,0.0088643283,-0.1135134101,0.0419382639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/239","title":"[Creating new dataset] Not found dataset_info.json","comments":"Yes I have an idea of what's going on. I'm sure I can fix that","body":"Hi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n","comment_length":15,"text":"[Creating new dataset] Not found dataset_info.json\nHi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n\nYes I have an idea of what's going on. I'm sure I can fix that","embeddings":[-0.1165063456,0.1242502332,-0.0860032439,0.057063479,0.0176752172,0.1466908902,0.1669220626,0.3094637394,-0.2007142454,0.0218163859,0.0602891333,0.5790073276,-0.0935086533,0.0841157287,0.2967385054,-0.0632679313,0.0798937827,0.3547498286,0.4398046434,-0.1709556282,-0.3041882515,0.3338649273,-0.0877133608,0.01387031,-0.338992089,-0.1399389803,-0.1987462044,0.1517356932,-0.3689015806,-0.2440757155,0.4867274761,-0.0733882189,0.1036610082,0.5475575328,-0.0001140775,0.0102124736,0.3725022972,-0.1520183682,-0.463706404,-0.5580253005,-0.0097940266,-0.4027791023,0.1149742603,-0.3410841227,-0.1230617836,-0.3242848516,0.2014034241,-0.1395288706,-0.1161304712,0.4508829117,0.1798371077,0.0618324205,0.184210822,-0.0867190957,0.0001254161,0.2762156427,-0.2401407212,0.2098938376,0.5791471601,-0.03125735,0.2688177526,-0.0465215221,-0.0659272224,-0.2932347655,0.2591234148,0.159864217,0.0156177217,-0.4965935946,0.0261633452,0.0796523169,0.7736666799,-0.5069001317,-0.3475548625,0.0792120323,0.194891125,0.0815229118,0.0261326935,0.3870602548,-0.0036244129,-0.1468177885,-0.4909806252,-0.3836780488,-0.1394308358,0.3541182876,0.2187725157,0.2453514934,-0.0337953642,0.0422469042,-0.0160155166,-0.0535681583,-0.1302879304,-0.1859699339,-0.1286573559,0.2532887161,-0.053666506,-0.5121890903,0.1687165648,-0.0076868557,-0.0895367935,0.061128173,-0.0326223448,-0.274641186,-0.2865626812,0.1832764745,0.2046544701,-0.1106357574,0.279658407,0.0797307193,0.0774832591,0.055838082,0.0785688013,-0.0678011626,-0.0335660912,0.0156166032,-0.1277638078,-0.0994326547,0.3245657682,-0.0741777495,0.0718602091,0.0012805207,-0.1981693804,-0.133311823,0.1172707081,0.2936829925,-0.0719582215,-0.2800116837,-0.0251497515,0.2071141303,-0.4114207625,-0.0116609484,-0.1645233035,0.4944566786,-0.2056022882,-0.0262290053,0.3875793517,0.1016707048,0.3338846862,-0.0130871162,-0.368738085,-0.1556921601,0.3918465078,-0.2121744305,-0.1790122241,0.228953436,0.1160459667,0.1139768213,-0.0378630012,-0.3668879271,-0.3832911253,0.1916641742,-0.2350608408,-0.3481474817,-0.283872366,0.1933897138,0.0613683388,-0.2342845798,-0.3146072924,0.2751885056,0.3097748458,-0.3435207307,0.0633967742,0.0758629963,-0.1343005598,-0.2305130363,-0.0999691784,0.3998726904,-0.4836797416,-0.0721997991,-0.3215029836,-0.0094186161,0.1910135746,-0.0352825522,-0.2096750736,0.5172001123,-0.1338477433,0.1112853959,0.5153878927,-0.1893327534,-0.187461406,0.2695929706,-0.1940307021,-0.2928598523,0.0725557357,0.2507296503,-0.0414150171,-0.1687862128,0.1327093393,0.0917857811,-0.126192376,0.0494734347,0.0835390836,-0.0060451939,0.0510402881,0.1193809733,-0.2383529991,0.1086443737,0.1355073601,0.2590746582,0.097801879,-0.0312276371,0.1779914796,0.4847249985,0.3885298371,-0.0059467414,-0.0228138771,-0.0708329007,-0.3362148404,-0.0818720236,-0.2369996756,0.4043165743,0.0337655842,-0.002795812,-0.4716331661,-0.192667827,0.0334816724,-0.1176143065,0.0850239992,0.0162108652,0.3978447318,0.1003789753,-0.2538137734,0.0775243267,-0.1711559445,0.0665230826,-0.8675051332,0.3471957743,-0.2013229132,-0.2094378471,0.2470538467,0.2310829163,-0.0176213309,-0.1472965032,0.1502861083,0.2265686393,-0.1452424228,0.3361055255,0.2227082998,0.128842622,0.2885342836,-0.2038090229,0.1463573426,0.0493759997,0.1962344646,-0.0217218064,-0.0831139833,0.0535558686,0.1434141099,-0.0550300702,0.1617177576,-0.0439132974,0.1402127743,-0.1786086708,0.0769875199,-0.2421171665,0.1755624563,0.075510703,0.131608218,0.048625093,-0.0772270411,0.2307497859,0.3067060411,0.0531004742,0.0431879945,0.157994464,0.0398811735,-0.0914959162,-0.1635780036,0.2864362597,0.4337894619,0.1305443794,-0.1725736558,-0.0445974953,-0.1054399312,-0.2338734865,0.1762125492,-0.144586727,0.1815698743,0.0553759448,0.2064576298,-0.091882892,-0.1981725991,-0.4503073096,0.2582753003,0.2915617824,0.0125326188,0.2330195606,-0.0483596958,-0.2889663279,-0.1023572162,-0.2199970037,-0.027930662,-0.1097455919,0.0449513085,0.2148979455,0.2119145691,-0.1206713319,-0.1959022731,0.1891421676,-0.0882790238,-0.3575147688,-0.0885614678,-0.2564783394,-0.1869902909,0.1057349443,0.0348248258,0.1400998682,0.2495245636,-0.1843912452,0.0105685499,-0.1660453826,-0.2570820153,0.0180810299,-0.1595323384,0.19502756,0.1740631312,0.0597916804,-0.2608325481,-0.4254710078,0.0775657073,-0.0116267065,-0.4032616317,0.0009183167,-0.1733697355,-0.1220971048,-0.0407483503,-0.711997509,-0.3251382411,-0.213160634,-0.0155927436,0.1631813794,0.1977260411,0.4130190015,-0.0757454857,0.1082077697,0.1723190397,0.2316488922,-0.0692338049,-0.3634283841,0.3943191171,-0.5005172491,-0.2842501998,0.1889234185,-0.4618218243,0.2375343293,-0.1741824299,-0.5138138533,0.0168081373,-0.2980986834,0.0059204972,0.1343669146,0.2339942753,0.4963957071,-0.0279377773,-0.053906247,-0.1033315063,-0.4580044746,0.0228570011,0.4239569306,0.6483789682,-0.0955790132,0.1766877472,-0.4366553128,0.1395106316,0.3806612194,-0.0229976848,0.4596581459,-0.0642139837,0.2842175663,-0.2890721262,-0.3776216805,0.2021721154,0.2846024036,-0.0968416631,0.1915385127,0.0889659151,-0.1900496632,-0.1813447624,-0.1297150552,-0.4387228489,-0.0670411065,-0.1606695354,-0.0154463137,0.2671151459,0.0893691704,-0.063767001,-0.1456261724,-0.2757143676,-0.2235910594,0.4727536738,0.1974101365,0.1208718792,-0.2766308486,-0.0612071976,-0.307307303,-0.0377948396,0.0238852985,0.2937771976,-0.202954337,0.255422622,0.0629198477,-0.0123398248,0.5490016341,-0.2908591032,-0.2055511624,0.2633007467,-0.0580418445,-0.2241145223,0.1805141866,-0.1040338427,0.6076475382,0.2433528602,0.5753422379,-0.3115656376,0.2990731299,0.1214507297,0.3701847196,-0.2317983359,-0.2861895859,-0.1955006421,-0.4630236924,-0.1389040947,-0.3091448843,-0.3363452852,0.4752550423,0.2303315699,0.170401454,0.3116016984,-0.1196064502,0.2841059268,0.1153542474,0.2576376796,0.2335043997,0.4246392846,-0.1052278578,-0.0671053752,0.3669828773,0.6004720926,0.2031099647,-0.0956731215,0.1284203976,-0.0364036411,0.395367533,-0.0132449679,0.266926378,-0.1098193899,0.0210946109,-0.2184128016,-0.115531534,-0.0234098081,-0.0193970408,-0.2241461277,-0.6888647079,-0.3939317465,0.2423800826,-0.0106856376,0.3002980351,0.2865825295,0.1424811333,-0.2200430632,0.471485883,-0.2134755552,0.8308844566,-0.0526556,-0.0279705022,0.0454090759,0.1510666609,0.7034148574,-0.5142815709,0.4126512706,-0.3891026974,-0.0269086305,0.0515783466,-0.1758003682,0.1750633717,0.4178828299,-0.4325260818,0.1930044442,-0.0822679177,-0.1185833812,-0.1702666879,0.4646313787,-0.1232812628,-0.3681624234,-0.3878662288,0.1243901253,-0.1559613645,0.3885234296,-0.0936821178,-0.0173259676,-0.5373473763,-0.0237672329,-0.2104974687,-0.0220599901,0.3186563551,0.249805361,0.5425424576,-0.3958864212,0.1680613905,0.1537734866,0.5115233064,0.3233711123,-0.2130830884,-0.0431414843,-0.0085645812,-0.4150432348,0.0861745626,0.0233168025,0.3259398341,0.0027130737,-0.2486928701,0.1554380059,-0.0910228491,-0.0530705713,-0.0971550718,-0.1465338618,-0.2417299896,-0.2759905159,-0.3625261486,0.2587006092,-0.1688533127,-0.1415452212,0.0788361356,0.2971162498,-0.1353914887,0.1185149401,0.0202571284,-0.1812777519,-0.2206898779,0.23206532,0.3509170711,-0.1711783409,0.4593079388,0.2199606895,-0.0807433948,-0.2345264852,0.1070886403,0.1253063977,-0.1298018247,0.2112330049,0.3616847694,0.1504724026,0.2839156985,0.3608105481,-0.166363582,0.1632482409,0.0730139166,-0.3862999678,-0.436652571,0.2469571084,0.0244277362,0.0901544467,0.1384783983,0.3877499104,0.2723523378,0.035061609,-0.2648818195,-0.0455673523,-0.3696907163,0.2637805939,0.2491617203,0.0605015419,-0.2160846889,-0.057571128,0.0510149114,-0.2032997012,-0.0290426463,-0.1544078141,-0.2216129601,0.1520265043,0.1260631382,0.0154480394,-0.1245804504,-0.2296669036,-0.2332468629,-0.0199401714,-0.2605073452,0.1780531257,0.1019101664,0.1546177417,0.0935981572,-0.3269782066,-0.1073557809,-0.0148436576,-0.0619534105,0.3862453401,0.0843533278,0.0452335738,0.1161664277,0.0425391942,-0.4028134346,0.0701125935,-0.1516933888,0.0837625638,0.1490865946,-0.157145083,0.2667430937,0.0217986628,0.1099834368,0.3466070592,-0.1359733343,0.1405175328,0.260553211,0.1115451902,-0.1620906889,-0.0440804772,0.3279460669,0.0434803776,-0.0122272819,0.3146601915,-0.1331052035,0.131069079,-0.1478780508,0.1849879473,0.3246780336,-0.0631473958,0.2577349842,0.1364097297,0.0012924885,-0.0977015942,0.2594444156,0.1881525815,-0.0853210315,0.1823779792,0.0450893342,0.3973941207,0.302352041,0.2861189544,0.2109260857,-0.2249996215,0.0504466966,-0.2021240294,0.2507336736,-0.0419258364,0.0826024935,0.2187250257,-0.1958678067,0.3018075824,-0.1780408919,0.1030380204,-0.3208485246,-0.0656001419,-0.1125702336,-0.2674712539,0.2546589375,-0.0874804184,0.0292556714,-0.1072182357,-0.1549135,-0.0531782545,-0.0556121841,-0.327442795,0.4158905149,-0.014811635,0.0209267847,-0.1005132124,0.3173038661,-0.1687464118,-0.1228318959,-0.1224116385,0.3447618186,0.307612896,0.3469654322,0.1381114572,0.2645431161,-0.0776175037,-0.1906177551,0.1139080673,0.4997584224,0.4584063292,-0.1055591255,0.0251274351,0.0590283573,-0.0244646985,0.1337961853,0.5156471133,0.0381495841,0.1663969457,0.4495267272,-0.1199095845,-0.1236041114,-0.2566175163,0.2996979356,-0.4078597426,-0.1276357025,0.5206471682,0.0403465368,-0.0090243537,-0.3289625347,0.0792628452,-0.0467726104,0.2938723564,0.3701539338,0.1201424971,-0.262578249,-0.2484817952,-0.3296501338,-0.0624068305,-0.111139901,-0.1620778441,-0.2252211571,0.3631657958,0.0261462871,0.1804602295,-0.362882942,0.2894383967,-0.1485565156,0.0748760626,-0.1408980638,-0.338190645,-0.1541633755,0.1900706887,-0.0819498003,-0.0871439874,-0.0184988342,-0.2756521106,-0.0113097196,-0.432760179,-0.2158632725,0.2175652385,0.3473265171,0.1009334028,0.0736395046,0.3765215874,-0.2977442741,-0.1872087121,-0.1939508617,-0.2278358787,-0.0951205119,0.2785026729,0.2827716768,0.37742576,-0.301546216,0.3614298999,-0.1413934827,0.0988541469,-0.030947797,0.1174751371,-0.0910616368,0.0121015664,-0.2916093171,-0.0935944244,0.0717773363,0.2724191248,0.0033911539,0.3202648759,-0.0997141823,-0.449174583,0.4673528671,-0.2092696279,-0.188631773,-0.0279480461,0.0273697264,-0.2985510826,-0.2551029325,-0.2464308143,0.0702252463,0.3553676307,-0.2813268304,0.0182162989,0.0918259546,-0.1916161627,0.0478141867,-0.0269898754,0.1780251116,0.103638418,0.0088643283,-0.1135134101,0.0419382639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/239","title":"[Creating new dataset] Not found dataset_info.json","comments":"Hi, I rebase my local copy to `fix-empty-cache-dir`, and try to run again `python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`.\r\n\r\nI got this, \r\n```\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 10, in <module>\r\n from nlp.commands.run_beam import RunBeamCommand\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/commands\/run_beam.py\", line 6, in <module>\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n```\r\nAnd after I installed it. I got this\r\n```\r\nFile \"\/home\/yisiang\/nlp\/src\/nlp\/datasets\/bookcorpus\/aea0bd5142d26df645a8fce23d6110bb95ecb81772bb2a1f29012e329191962c\/bookcorpus.py\", line 88, in _split_generators\r\n downloaded_path_or_paths = dl_manager.download_custom(_GDRIVE_FILE_ID, download_file_from_google_drive)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 128, in download_custom\r\n downloaded_path_or_paths = map_nested(url_to_downloaded_path, url_or_urls)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 126, in url_to_downloaded_path\r\n return os.path.join(self._download_config.cache_dir, hash_url_to_filename(url))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/nlppr\/lib\/python3.7\/posixpath.py\", line 80, in join\r\n a = os.fspath(a)\r\n```\r\nThe problem is when I print `self._download_config.cache_dir` using pdb, it is `None`.\r\n\r\nDid I miss something ? Or can you provide a workaround first so I can keep testing my script ?","body":"Hi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n","comment_length":142,"text":"[Creating new dataset] Not found dataset_info.json\nHi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n\nHi, I rebase my local copy to `fix-empty-cache-dir`, and try to run again `python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`.\r\n\r\nI got this, \r\n```\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 10, in <module>\r\n from nlp.commands.run_beam import RunBeamCommand\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/commands\/run_beam.py\", line 6, in <module>\r\n import apache_beam as beam\r\nModuleNotFoundError: No module named 'apache_beam'\r\n```\r\nAnd after I installed it. I got this\r\n```\r\nFile \"\/home\/yisiang\/nlp\/src\/nlp\/datasets\/bookcorpus\/aea0bd5142d26df645a8fce23d6110bb95ecb81772bb2a1f29012e329191962c\/bookcorpus.py\", line 88, in _split_generators\r\n downloaded_path_or_paths = dl_manager.download_custom(_GDRIVE_FILE_ID, download_file_from_google_drive)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 128, in download_custom\r\n downloaded_path_or_paths = map_nested(url_to_downloaded_path, url_or_urls)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/yisiang\/nlp\/src\/nlp\/utils\/download_manager.py\", line 126, in url_to_downloaded_path\r\n return os.path.join(self._download_config.cache_dir, hash_url_to_filename(url))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/nlppr\/lib\/python3.7\/posixpath.py\", line 80, in join\r\n a = os.fspath(a)\r\n```\r\nThe problem is when I print `self._download_config.cache_dir` using pdb, it is `None`.\r\n\r\nDid I miss something ? Or can you provide a workaround first so I can keep testing my script ?","embeddings":[-0.1165063456,0.1242502332,-0.0860032439,0.057063479,0.0176752172,0.1466908902,0.1669220626,0.3094637394,-0.2007142454,0.0218163859,0.0602891333,0.5790073276,-0.0935086533,0.0841157287,0.2967385054,-0.0632679313,0.0798937827,0.3547498286,0.4398046434,-0.1709556282,-0.3041882515,0.3338649273,-0.0877133608,0.01387031,-0.338992089,-0.1399389803,-0.1987462044,0.1517356932,-0.3689015806,-0.2440757155,0.4867274761,-0.0733882189,0.1036610082,0.5475575328,-0.0001140775,0.0102124736,0.3725022972,-0.1520183682,-0.463706404,-0.5580253005,-0.0097940266,-0.4027791023,0.1149742603,-0.3410841227,-0.1230617836,-0.3242848516,0.2014034241,-0.1395288706,-0.1161304712,0.4508829117,0.1798371077,0.0618324205,0.184210822,-0.0867190957,0.0001254161,0.2762156427,-0.2401407212,0.2098938376,0.5791471601,-0.03125735,0.2688177526,-0.0465215221,-0.0659272224,-0.2932347655,0.2591234148,0.159864217,0.0156177217,-0.4965935946,0.0261633452,0.0796523169,0.7736666799,-0.5069001317,-0.3475548625,0.0792120323,0.194891125,0.0815229118,0.0261326935,0.3870602548,-0.0036244129,-0.1468177885,-0.4909806252,-0.3836780488,-0.1394308358,0.3541182876,0.2187725157,0.2453514934,-0.0337953642,0.0422469042,-0.0160155166,-0.0535681583,-0.1302879304,-0.1859699339,-0.1286573559,0.2532887161,-0.053666506,-0.5121890903,0.1687165648,-0.0076868557,-0.0895367935,0.061128173,-0.0326223448,-0.274641186,-0.2865626812,0.1832764745,0.2046544701,-0.1106357574,0.279658407,0.0797307193,0.0774832591,0.055838082,0.0785688013,-0.0678011626,-0.0335660912,0.0156166032,-0.1277638078,-0.0994326547,0.3245657682,-0.0741777495,0.0718602091,0.0012805207,-0.1981693804,-0.133311823,0.1172707081,0.2936829925,-0.0719582215,-0.2800116837,-0.0251497515,0.2071141303,-0.4114207625,-0.0116609484,-0.1645233035,0.4944566786,-0.2056022882,-0.0262290053,0.3875793517,0.1016707048,0.3338846862,-0.0130871162,-0.368738085,-0.1556921601,0.3918465078,-0.2121744305,-0.1790122241,0.228953436,0.1160459667,0.1139768213,-0.0378630012,-0.3668879271,-0.3832911253,0.1916641742,-0.2350608408,-0.3481474817,-0.283872366,0.1933897138,0.0613683388,-0.2342845798,-0.3146072924,0.2751885056,0.3097748458,-0.3435207307,0.0633967742,0.0758629963,-0.1343005598,-0.2305130363,-0.0999691784,0.3998726904,-0.4836797416,-0.0721997991,-0.3215029836,-0.0094186161,0.1910135746,-0.0352825522,-0.2096750736,0.5172001123,-0.1338477433,0.1112853959,0.5153878927,-0.1893327534,-0.187461406,0.2695929706,-0.1940307021,-0.2928598523,0.0725557357,0.2507296503,-0.0414150171,-0.1687862128,0.1327093393,0.0917857811,-0.126192376,0.0494734347,0.0835390836,-0.0060451939,0.0510402881,0.1193809733,-0.2383529991,0.1086443737,0.1355073601,0.2590746582,0.097801879,-0.0312276371,0.1779914796,0.4847249985,0.3885298371,-0.0059467414,-0.0228138771,-0.0708329007,-0.3362148404,-0.0818720236,-0.2369996756,0.4043165743,0.0337655842,-0.002795812,-0.4716331661,-0.192667827,0.0334816724,-0.1176143065,0.0850239992,0.0162108652,0.3978447318,0.1003789753,-0.2538137734,0.0775243267,-0.1711559445,0.0665230826,-0.8675051332,0.3471957743,-0.2013229132,-0.2094378471,0.2470538467,0.2310829163,-0.0176213309,-0.1472965032,0.1502861083,0.2265686393,-0.1452424228,0.3361055255,0.2227082998,0.128842622,0.2885342836,-0.2038090229,0.1463573426,0.0493759997,0.1962344646,-0.0217218064,-0.0831139833,0.0535558686,0.1434141099,-0.0550300702,0.1617177576,-0.0439132974,0.1402127743,-0.1786086708,0.0769875199,-0.2421171665,0.1755624563,0.075510703,0.131608218,0.048625093,-0.0772270411,0.2307497859,0.3067060411,0.0531004742,0.0431879945,0.157994464,0.0398811735,-0.0914959162,-0.1635780036,0.2864362597,0.4337894619,0.1305443794,-0.1725736558,-0.0445974953,-0.1054399312,-0.2338734865,0.1762125492,-0.144586727,0.1815698743,0.0553759448,0.2064576298,-0.091882892,-0.1981725991,-0.4503073096,0.2582753003,0.2915617824,0.0125326188,0.2330195606,-0.0483596958,-0.2889663279,-0.1023572162,-0.2199970037,-0.027930662,-0.1097455919,0.0449513085,0.2148979455,0.2119145691,-0.1206713319,-0.1959022731,0.1891421676,-0.0882790238,-0.3575147688,-0.0885614678,-0.2564783394,-0.1869902909,0.1057349443,0.0348248258,0.1400998682,0.2495245636,-0.1843912452,0.0105685499,-0.1660453826,-0.2570820153,0.0180810299,-0.1595323384,0.19502756,0.1740631312,0.0597916804,-0.2608325481,-0.4254710078,0.0775657073,-0.0116267065,-0.4032616317,0.0009183167,-0.1733697355,-0.1220971048,-0.0407483503,-0.711997509,-0.3251382411,-0.213160634,-0.0155927436,0.1631813794,0.1977260411,0.4130190015,-0.0757454857,0.1082077697,0.1723190397,0.2316488922,-0.0692338049,-0.3634283841,0.3943191171,-0.5005172491,-0.2842501998,0.1889234185,-0.4618218243,0.2375343293,-0.1741824299,-0.5138138533,0.0168081373,-0.2980986834,0.0059204972,0.1343669146,0.2339942753,0.4963957071,-0.0279377773,-0.053906247,-0.1033315063,-0.4580044746,0.0228570011,0.4239569306,0.6483789682,-0.0955790132,0.1766877472,-0.4366553128,0.1395106316,0.3806612194,-0.0229976848,0.4596581459,-0.0642139837,0.2842175663,-0.2890721262,-0.3776216805,0.2021721154,0.2846024036,-0.0968416631,0.1915385127,0.0889659151,-0.1900496632,-0.1813447624,-0.1297150552,-0.4387228489,-0.0670411065,-0.1606695354,-0.0154463137,0.2671151459,0.0893691704,-0.063767001,-0.1456261724,-0.2757143676,-0.2235910594,0.4727536738,0.1974101365,0.1208718792,-0.2766308486,-0.0612071976,-0.307307303,-0.0377948396,0.0238852985,0.2937771976,-0.202954337,0.255422622,0.0629198477,-0.0123398248,0.5490016341,-0.2908591032,-0.2055511624,0.2633007467,-0.0580418445,-0.2241145223,0.1805141866,-0.1040338427,0.6076475382,0.2433528602,0.5753422379,-0.3115656376,0.2990731299,0.1214507297,0.3701847196,-0.2317983359,-0.2861895859,-0.1955006421,-0.4630236924,-0.1389040947,-0.3091448843,-0.3363452852,0.4752550423,0.2303315699,0.170401454,0.3116016984,-0.1196064502,0.2841059268,0.1153542474,0.2576376796,0.2335043997,0.4246392846,-0.1052278578,-0.0671053752,0.3669828773,0.6004720926,0.2031099647,-0.0956731215,0.1284203976,-0.0364036411,0.395367533,-0.0132449679,0.266926378,-0.1098193899,0.0210946109,-0.2184128016,-0.115531534,-0.0234098081,-0.0193970408,-0.2241461277,-0.6888647079,-0.3939317465,0.2423800826,-0.0106856376,0.3002980351,0.2865825295,0.1424811333,-0.2200430632,0.471485883,-0.2134755552,0.8308844566,-0.0526556,-0.0279705022,0.0454090759,0.1510666609,0.7034148574,-0.5142815709,0.4126512706,-0.3891026974,-0.0269086305,0.0515783466,-0.1758003682,0.1750633717,0.4178828299,-0.4325260818,0.1930044442,-0.0822679177,-0.1185833812,-0.1702666879,0.4646313787,-0.1232812628,-0.3681624234,-0.3878662288,0.1243901253,-0.1559613645,0.3885234296,-0.0936821178,-0.0173259676,-0.5373473763,-0.0237672329,-0.2104974687,-0.0220599901,0.3186563551,0.249805361,0.5425424576,-0.3958864212,0.1680613905,0.1537734866,0.5115233064,0.3233711123,-0.2130830884,-0.0431414843,-0.0085645812,-0.4150432348,0.0861745626,0.0233168025,0.3259398341,0.0027130737,-0.2486928701,0.1554380059,-0.0910228491,-0.0530705713,-0.0971550718,-0.1465338618,-0.2417299896,-0.2759905159,-0.3625261486,0.2587006092,-0.1688533127,-0.1415452212,0.0788361356,0.2971162498,-0.1353914887,0.1185149401,0.0202571284,-0.1812777519,-0.2206898779,0.23206532,0.3509170711,-0.1711783409,0.4593079388,0.2199606895,-0.0807433948,-0.2345264852,0.1070886403,0.1253063977,-0.1298018247,0.2112330049,0.3616847694,0.1504724026,0.2839156985,0.3608105481,-0.166363582,0.1632482409,0.0730139166,-0.3862999678,-0.436652571,0.2469571084,0.0244277362,0.0901544467,0.1384783983,0.3877499104,0.2723523378,0.035061609,-0.2648818195,-0.0455673523,-0.3696907163,0.2637805939,0.2491617203,0.0605015419,-0.2160846889,-0.057571128,0.0510149114,-0.2032997012,-0.0290426463,-0.1544078141,-0.2216129601,0.1520265043,0.1260631382,0.0154480394,-0.1245804504,-0.2296669036,-0.2332468629,-0.0199401714,-0.2605073452,0.1780531257,0.1019101664,0.1546177417,0.0935981572,-0.3269782066,-0.1073557809,-0.0148436576,-0.0619534105,0.3862453401,0.0843533278,0.0452335738,0.1161664277,0.0425391942,-0.4028134346,0.0701125935,-0.1516933888,0.0837625638,0.1490865946,-0.157145083,0.2667430937,0.0217986628,0.1099834368,0.3466070592,-0.1359733343,0.1405175328,0.260553211,0.1115451902,-0.1620906889,-0.0440804772,0.3279460669,0.0434803776,-0.0122272819,0.3146601915,-0.1331052035,0.131069079,-0.1478780508,0.1849879473,0.3246780336,-0.0631473958,0.2577349842,0.1364097297,0.0012924885,-0.0977015942,0.2594444156,0.1881525815,-0.0853210315,0.1823779792,0.0450893342,0.3973941207,0.302352041,0.2861189544,0.2109260857,-0.2249996215,0.0504466966,-0.2021240294,0.2507336736,-0.0419258364,0.0826024935,0.2187250257,-0.1958678067,0.3018075824,-0.1780408919,0.1030380204,-0.3208485246,-0.0656001419,-0.1125702336,-0.2674712539,0.2546589375,-0.0874804184,0.0292556714,-0.1072182357,-0.1549135,-0.0531782545,-0.0556121841,-0.327442795,0.4158905149,-0.014811635,0.0209267847,-0.1005132124,0.3173038661,-0.1687464118,-0.1228318959,-0.1224116385,0.3447618186,0.307612896,0.3469654322,0.1381114572,0.2645431161,-0.0776175037,-0.1906177551,0.1139080673,0.4997584224,0.4584063292,-0.1055591255,0.0251274351,0.0590283573,-0.0244646985,0.1337961853,0.5156471133,0.0381495841,0.1663969457,0.4495267272,-0.1199095845,-0.1236041114,-0.2566175163,0.2996979356,-0.4078597426,-0.1276357025,0.5206471682,0.0403465368,-0.0090243537,-0.3289625347,0.0792628452,-0.0467726104,0.2938723564,0.3701539338,0.1201424971,-0.262578249,-0.2484817952,-0.3296501338,-0.0624068305,-0.111139901,-0.1620778441,-0.2252211571,0.3631657958,0.0261462871,0.1804602295,-0.362882942,0.2894383967,-0.1485565156,0.0748760626,-0.1408980638,-0.338190645,-0.1541633755,0.1900706887,-0.0819498003,-0.0871439874,-0.0184988342,-0.2756521106,-0.0113097196,-0.432760179,-0.2158632725,0.2175652385,0.3473265171,0.1009334028,0.0736395046,0.3765215874,-0.2977442741,-0.1872087121,-0.1939508617,-0.2278358787,-0.0951205119,0.2785026729,0.2827716768,0.37742576,-0.301546216,0.3614298999,-0.1413934827,0.0988541469,-0.030947797,0.1174751371,-0.0910616368,0.0121015664,-0.2916093171,-0.0935944244,0.0717773363,0.2724191248,0.0033911539,0.3202648759,-0.0997141823,-0.449174583,0.4673528671,-0.2092696279,-0.188631773,-0.0279480461,0.0273697264,-0.2985510826,-0.2551029325,-0.2464308143,0.0702252463,0.3553676307,-0.2813268304,0.0182162989,0.0918259546,-0.1916161627,0.0478141867,-0.0269898754,0.1780251116,0.103638418,0.0088643283,-0.1135134101,0.0419382639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/239","title":"[Creating new dataset] Not found dataset_info.json","comments":"I'll close this issue because I brings more reports in another issue #249 .","body":"Hi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n","comment_length":14,"text":"[Creating new dataset] Not found dataset_info.json\nHi, I am trying to create Toronto Book Corpus. #131 \r\n\r\nI ran\r\n`~\/nlp % python nlp-cli test datasets\/bookcorpus --save_infos --all_configs`\r\nbut this doesn't create `dataset_info.json` and try to use it\r\n```\r\nINFO:nlp.load:Checking datasets\/bookcorpus\/bookcorpus.py for additional imports.\r\nINFO:filelock:Lock 139795325778640 acquired on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.load:Found main folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\r\nINFO:nlp.load:Found specific version folder for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\r\nINFO:nlp.load:Found script file from datasets\/bookcorpus\/bookcorpus.py to \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.py\r\nINFO:nlp.load:Couldn't find dataset infos file at datasets\/bookcorpus\/dataset_infos.json\r\nINFO:nlp.load:Found metadata file for dataset datasets\/bookcorpus\/bookcorpus.py at \/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/datasets\/bookcorpus\/8e84759446cf68d0b0deb3417e60cc331f30a3bbe58843de18a0f48e87d1efd9\/bookcorpus.json\r\nINFO:filelock:Lock 139795325778640 released on datasets\/bookcorpus\/bookcorpus.py.lock\r\nINFO:nlp.builder:Overwrite dataset info from restored data version.\r\nINFO:nlp.info:Loading Dataset info from \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\r\nTraceback (most recent call last):\r\n File \"nlp-cli\", line 37, in <module>\r\n service.run()\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/commands\/test.py\", line 78, in run\r\n builders.append(builder_cls(name=config.name, data_dir=self._data_dir))\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\n File \"\/home\/yisiang\/miniconda3\/envs\/ml\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/dataset_info.json'\r\n```\r\nbtw, `ls \/home\/yisiang\/.cache\/huggingface\/datasets\/book_corpus\/plain_text\/1.0.0\/` show me nothing is in the directory.\r\n\r\nI have also pushed the script to my fork [bookcorpus.py](https:\/\/github.com\/richardyy1188\/nlp\/blob\/bookcorpusdev\/datasets\/bookcorpus\/bookcorpus.py).\r\n\nI'll close this issue because I brings more reports in another issue #249 .","embeddings":[-0.1165063456,0.1242502332,-0.0860032439,0.057063479,0.0176752172,0.1466908902,0.1669220626,0.3094637394,-0.2007142454,0.0218163859,0.0602891333,0.5790073276,-0.0935086533,0.0841157287,0.2967385054,-0.0632679313,0.0798937827,0.3547498286,0.4398046434,-0.1709556282,-0.3041882515,0.3338649273,-0.0877133608,0.01387031,-0.338992089,-0.1399389803,-0.1987462044,0.1517356932,-0.3689015806,-0.2440757155,0.4867274761,-0.0733882189,0.1036610082,0.5475575328,-0.0001140775,0.0102124736,0.3725022972,-0.1520183682,-0.463706404,-0.5580253005,-0.0097940266,-0.4027791023,0.1149742603,-0.3410841227,-0.1230617836,-0.3242848516,0.2014034241,-0.1395288706,-0.1161304712,0.4508829117,0.1798371077,0.0618324205,0.184210822,-0.0867190957,0.0001254161,0.2762156427,-0.2401407212,0.2098938376,0.5791471601,-0.03125735,0.2688177526,-0.0465215221,-0.0659272224,-0.2932347655,0.2591234148,0.159864217,0.0156177217,-0.4965935946,0.0261633452,0.0796523169,0.7736666799,-0.5069001317,-0.3475548625,0.0792120323,0.194891125,0.0815229118,0.0261326935,0.3870602548,-0.0036244129,-0.1468177885,-0.4909806252,-0.3836780488,-0.1394308358,0.3541182876,0.2187725157,0.2453514934,-0.0337953642,0.0422469042,-0.0160155166,-0.0535681583,-0.1302879304,-0.1859699339,-0.1286573559,0.2532887161,-0.053666506,-0.5121890903,0.1687165648,-0.0076868557,-0.0895367935,0.061128173,-0.0326223448,-0.274641186,-0.2865626812,0.1832764745,0.2046544701,-0.1106357574,0.279658407,0.0797307193,0.0774832591,0.055838082,0.0785688013,-0.0678011626,-0.0335660912,0.0156166032,-0.1277638078,-0.0994326547,0.3245657682,-0.0741777495,0.0718602091,0.0012805207,-0.1981693804,-0.133311823,0.1172707081,0.2936829925,-0.0719582215,-0.2800116837,-0.0251497515,0.2071141303,-0.4114207625,-0.0116609484,-0.1645233035,0.4944566786,-0.2056022882,-0.0262290053,0.3875793517,0.1016707048,0.3338846862,-0.0130871162,-0.368738085,-0.1556921601,0.3918465078,-0.2121744305,-0.1790122241,0.228953436,0.1160459667,0.1139768213,-0.0378630012,-0.3668879271,-0.3832911253,0.1916641742,-0.2350608408,-0.3481474817,-0.283872366,0.1933897138,0.0613683388,-0.2342845798,-0.3146072924,0.2751885056,0.3097748458,-0.3435207307,0.0633967742,0.0758629963,-0.1343005598,-0.2305130363,-0.0999691784,0.3998726904,-0.4836797416,-0.0721997991,-0.3215029836,-0.0094186161,0.1910135746,-0.0352825522,-0.2096750736,0.5172001123,-0.1338477433,0.1112853959,0.5153878927,-0.1893327534,-0.187461406,0.2695929706,-0.1940307021,-0.2928598523,0.0725557357,0.2507296503,-0.0414150171,-0.1687862128,0.1327093393,0.0917857811,-0.126192376,0.0494734347,0.0835390836,-0.0060451939,0.0510402881,0.1193809733,-0.2383529991,0.1086443737,0.1355073601,0.2590746582,0.097801879,-0.0312276371,0.1779914796,0.4847249985,0.3885298371,-0.0059467414,-0.0228138771,-0.0708329007,-0.3362148404,-0.0818720236,-0.2369996756,0.4043165743,0.0337655842,-0.002795812,-0.4716331661,-0.192667827,0.0334816724,-0.1176143065,0.0850239992,0.0162108652,0.3978447318,0.1003789753,-0.2538137734,0.0775243267,-0.1711559445,0.0665230826,-0.8675051332,0.3471957743,-0.2013229132,-0.2094378471,0.2470538467,0.2310829163,-0.0176213309,-0.1472965032,0.1502861083,0.2265686393,-0.1452424228,0.3361055255,0.2227082998,0.128842622,0.2885342836,-0.2038090229,0.1463573426,0.0493759997,0.1962344646,-0.0217218064,-0.0831139833,0.0535558686,0.1434141099,-0.0550300702,0.1617177576,-0.0439132974,0.1402127743,-0.1786086708,0.0769875199,-0.2421171665,0.1755624563,0.075510703,0.131608218,0.048625093,-0.0772270411,0.2307497859,0.3067060411,0.0531004742,0.0431879945,0.157994464,0.0398811735,-0.0914959162,-0.1635780036,0.2864362597,0.4337894619,0.1305443794,-0.1725736558,-0.0445974953,-0.1054399312,-0.2338734865,0.1762125492,-0.144586727,0.1815698743,0.0553759448,0.2064576298,-0.091882892,-0.1981725991,-0.4503073096,0.2582753003,0.2915617824,0.0125326188,0.2330195606,-0.0483596958,-0.2889663279,-0.1023572162,-0.2199970037,-0.027930662,-0.1097455919,0.0449513085,0.2148979455,0.2119145691,-0.1206713319,-0.1959022731,0.1891421676,-0.0882790238,-0.3575147688,-0.0885614678,-0.2564783394,-0.1869902909,0.1057349443,0.0348248258,0.1400998682,0.2495245636,-0.1843912452,0.0105685499,-0.1660453826,-0.2570820153,0.0180810299,-0.1595323384,0.19502756,0.1740631312,0.0597916804,-0.2608325481,-0.4254710078,0.0775657073,-0.0116267065,-0.4032616317,0.0009183167,-0.1733697355,-0.1220971048,-0.0407483503,-0.711997509,-0.3251382411,-0.213160634,-0.0155927436,0.1631813794,0.1977260411,0.4130190015,-0.0757454857,0.1082077697,0.1723190397,0.2316488922,-0.0692338049,-0.3634283841,0.3943191171,-0.5005172491,-0.2842501998,0.1889234185,-0.4618218243,0.2375343293,-0.1741824299,-0.5138138533,0.0168081373,-0.2980986834,0.0059204972,0.1343669146,0.2339942753,0.4963957071,-0.0279377773,-0.053906247,-0.1033315063,-0.4580044746,0.0228570011,0.4239569306,0.6483789682,-0.0955790132,0.1766877472,-0.4366553128,0.1395106316,0.3806612194,-0.0229976848,0.4596581459,-0.0642139837,0.2842175663,-0.2890721262,-0.3776216805,0.2021721154,0.2846024036,-0.0968416631,0.1915385127,0.0889659151,-0.1900496632,-0.1813447624,-0.1297150552,-0.4387228489,-0.0670411065,-0.1606695354,-0.0154463137,0.2671151459,0.0893691704,-0.063767001,-0.1456261724,-0.2757143676,-0.2235910594,0.4727536738,0.1974101365,0.1208718792,-0.2766308486,-0.0612071976,-0.307307303,-0.0377948396,0.0238852985,0.2937771976,-0.202954337,0.255422622,0.0629198477,-0.0123398248,0.5490016341,-0.2908591032,-0.2055511624,0.2633007467,-0.0580418445,-0.2241145223,0.1805141866,-0.1040338427,0.6076475382,0.2433528602,0.5753422379,-0.3115656376,0.2990731299,0.1214507297,0.3701847196,-0.2317983359,-0.2861895859,-0.1955006421,-0.4630236924,-0.1389040947,-0.3091448843,-0.3363452852,0.4752550423,0.2303315699,0.170401454,0.3116016984,-0.1196064502,0.2841059268,0.1153542474,0.2576376796,0.2335043997,0.4246392846,-0.1052278578,-0.0671053752,0.3669828773,0.6004720926,0.2031099647,-0.0956731215,0.1284203976,-0.0364036411,0.395367533,-0.0132449679,0.266926378,-0.1098193899,0.0210946109,-0.2184128016,-0.115531534,-0.0234098081,-0.0193970408,-0.2241461277,-0.6888647079,-0.3939317465,0.2423800826,-0.0106856376,0.3002980351,0.2865825295,0.1424811333,-0.2200430632,0.471485883,-0.2134755552,0.8308844566,-0.0526556,-0.0279705022,0.0454090759,0.1510666609,0.7034148574,-0.5142815709,0.4126512706,-0.3891026974,-0.0269086305,0.0515783466,-0.1758003682,0.1750633717,0.4178828299,-0.4325260818,0.1930044442,-0.0822679177,-0.1185833812,-0.1702666879,0.4646313787,-0.1232812628,-0.3681624234,-0.3878662288,0.1243901253,-0.1559613645,0.3885234296,-0.0936821178,-0.0173259676,-0.5373473763,-0.0237672329,-0.2104974687,-0.0220599901,0.3186563551,0.249805361,0.5425424576,-0.3958864212,0.1680613905,0.1537734866,0.5115233064,0.3233711123,-0.2130830884,-0.0431414843,-0.0085645812,-0.4150432348,0.0861745626,0.0233168025,0.3259398341,0.0027130737,-0.2486928701,0.1554380059,-0.0910228491,-0.0530705713,-0.0971550718,-0.1465338618,-0.2417299896,-0.2759905159,-0.3625261486,0.2587006092,-0.1688533127,-0.1415452212,0.0788361356,0.2971162498,-0.1353914887,0.1185149401,0.0202571284,-0.1812777519,-0.2206898779,0.23206532,0.3509170711,-0.1711783409,0.4593079388,0.2199606895,-0.0807433948,-0.2345264852,0.1070886403,0.1253063977,-0.1298018247,0.2112330049,0.3616847694,0.1504724026,0.2839156985,0.3608105481,-0.166363582,0.1632482409,0.0730139166,-0.3862999678,-0.436652571,0.2469571084,0.0244277362,0.0901544467,0.1384783983,0.3877499104,0.2723523378,0.035061609,-0.2648818195,-0.0455673523,-0.3696907163,0.2637805939,0.2491617203,0.0605015419,-0.2160846889,-0.057571128,0.0510149114,-0.2032997012,-0.0290426463,-0.1544078141,-0.2216129601,0.1520265043,0.1260631382,0.0154480394,-0.1245804504,-0.2296669036,-0.2332468629,-0.0199401714,-0.2605073452,0.1780531257,0.1019101664,0.1546177417,0.0935981572,-0.3269782066,-0.1073557809,-0.0148436576,-0.0619534105,0.3862453401,0.0843533278,0.0452335738,0.1161664277,0.0425391942,-0.4028134346,0.0701125935,-0.1516933888,0.0837625638,0.1490865946,-0.157145083,0.2667430937,0.0217986628,0.1099834368,0.3466070592,-0.1359733343,0.1405175328,0.260553211,0.1115451902,-0.1620906889,-0.0440804772,0.3279460669,0.0434803776,-0.0122272819,0.3146601915,-0.1331052035,0.131069079,-0.1478780508,0.1849879473,0.3246780336,-0.0631473958,0.2577349842,0.1364097297,0.0012924885,-0.0977015942,0.2594444156,0.1881525815,-0.0853210315,0.1823779792,0.0450893342,0.3973941207,0.302352041,0.2861189544,0.2109260857,-0.2249996215,0.0504466966,-0.2021240294,0.2507336736,-0.0419258364,0.0826024935,0.2187250257,-0.1958678067,0.3018075824,-0.1780408919,0.1030380204,-0.3208485246,-0.0656001419,-0.1125702336,-0.2674712539,0.2546589375,-0.0874804184,0.0292556714,-0.1072182357,-0.1549135,-0.0531782545,-0.0556121841,-0.327442795,0.4158905149,-0.014811635,0.0209267847,-0.1005132124,0.3173038661,-0.1687464118,-0.1228318959,-0.1224116385,0.3447618186,0.307612896,0.3469654322,0.1381114572,0.2645431161,-0.0776175037,-0.1906177551,0.1139080673,0.4997584224,0.4584063292,-0.1055591255,0.0251274351,0.0590283573,-0.0244646985,0.1337961853,0.5156471133,0.0381495841,0.1663969457,0.4495267272,-0.1199095845,-0.1236041114,-0.2566175163,0.2996979356,-0.4078597426,-0.1276357025,0.5206471682,0.0403465368,-0.0090243537,-0.3289625347,0.0792628452,-0.0467726104,0.2938723564,0.3701539338,0.1201424971,-0.262578249,-0.2484817952,-0.3296501338,-0.0624068305,-0.111139901,-0.1620778441,-0.2252211571,0.3631657958,0.0261462871,0.1804602295,-0.362882942,0.2894383967,-0.1485565156,0.0748760626,-0.1408980638,-0.338190645,-0.1541633755,0.1900706887,-0.0819498003,-0.0871439874,-0.0184988342,-0.2756521106,-0.0113097196,-0.432760179,-0.2158632725,0.2175652385,0.3473265171,0.1009334028,0.0736395046,0.3765215874,-0.2977442741,-0.1872087121,-0.1939508617,-0.2278358787,-0.0951205119,0.2785026729,0.2827716768,0.37742576,-0.301546216,0.3614298999,-0.1413934827,0.0988541469,-0.030947797,0.1174751371,-0.0910616368,0.0121015664,-0.2916093171,-0.0935944244,0.0717773363,0.2724191248,0.0033911539,0.3202648759,-0.0997141823,-0.449174583,0.4673528671,-0.2092696279,-0.188631773,-0.0279480461,0.0273697264,-0.2985510826,-0.2551029325,-0.2464308143,0.0702252463,0.3553676307,-0.2813268304,0.0182162989,0.0918259546,-0.1916161627,0.0478141867,-0.0269898754,0.1780251116,0.103638418,0.0088643283,-0.1135134101,0.0419382639]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/238","title":"[Metric] Bertscore : Warning : Empty candidate sentence; Setting recall to be 0.","comments":"This print statement comes from the official implementation of bert_score (see [here](https:\/\/github.com\/Tiiiger\/bert_score\/blob\/master\/bert_score\/utils.py#L343)). The warning shows up only if the attention mask outputs no candidate.\r\nRight now we want to only use official code for metrics to have fair evaluations, so I'm not sure we can do anything about it. Maybe you can try to create an issue on their [repo](https:\/\/github.com\/Tiiiger\/bert_score) ?","body":"When running BERT-Score, I'm meeting this warning :\r\n\r\n> Warning: Empty candidate sentence; Setting recall to be 0.\r\n\r\nCode :\r\n\r\n```\r\nimport nlp\r\nmetric = nlp.load_metric(\"bertscore\")\r\nscores = metric.compute([\"swag\", \"swags\"], [\"swags\", \"totally something different\"], lang=\"en\", device=0)\r\n```\r\n\r\n---\r\n\r\n**What am I doing wrong \/ How can I hide this warning ?**","comment_length":61,"text":"[Metric] Bertscore : Warning : Empty candidate sentence; Setting recall to be 0.\nWhen running BERT-Score, I'm meeting this warning :\r\n\r\n> Warning: Empty candidate sentence; Setting recall to be 0.\r\n\r\nCode :\r\n\r\n```\r\nimport nlp\r\nmetric = nlp.load_metric(\"bertscore\")\r\nscores = metric.compute([\"swag\", \"swags\"], [\"swags\", \"totally something different\"], lang=\"en\", device=0)\r\n```\r\n\r\n---\r\n\r\n**What am I doing wrong \/ How can I hide this warning ?**\nThis print statement comes from the official implementation of bert_score (see [here](https:\/\/github.com\/Tiiiger\/bert_score\/blob\/master\/bert_score\/utils.py#L343)). The warning shows up only if the attention mask outputs no candidate.\r\nRight now we want to only use official code for metrics to have fair evaluations, so I'm not sure we can do anything about it. Maybe you can try to create an issue on their [repo](https:\/\/github.com\/Tiiiger\/bert_score) ?","embeddings":[0.2172083408,-0.1352753937,0.1404270083,0.1618988514,0.1744060963,0.0671908632,-0.0303219259,0.264549464,0.4367605448,0.3824990094,0.4003357589,0.1243717819,-0.2570009232,-0.4931720793,-0.1998976916,-0.1357424855,0.0677198842,0.4522253573,0.5661281347,-0.251927197,-0.2824742794,-0.0953284204,-0.3659345508,0.2246672958,-0.269020021,0.0552604273,0.0221389793,-0.2626544237,-0.1808631271,-0.5885035992,0.1738504618,-0.1883931458,-0.2331287712,0.2377419025,-0.0001283982,-0.2515506148,0.4155663252,-0.0837017745,0.2523778975,-0.4310386777,-0.4188614786,-0.1371326596,0.2361352146,-0.2389804721,0.0187330488,0.2030011117,0.1517081559,-0.1645323634,0.0686172247,0.1861715019,0.0121348435,0.2936764956,-0.1058877259,-0.0611063577,0.0227679927,-0.1313389391,-0.0275754463,0.0413871147,-0.192634508,-0.2139684707,-0.0495755672,0.5230774283,-0.3952677846,0.1201559156,0.6345873475,0.0905861557,0.2315475643,-0.2104471177,0.0147829996,0.2737215161,-0.1209857985,-0.1175498515,-0.0278368797,-0.0735123008,0.3021912575,-0.3431499898,0.1852401495,0.0706935227,-0.0142913405,-0.2116330415,-0.3099201024,-0.2357536852,-0.2747386098,-0.0452069156,-0.0472432859,0.3264420331,0.0771593675,-0.064789623,0.285643369,-0.1446775496,-0.0932134464,-0.1892850995,-0.1289047003,0.1941079348,-0.0856098011,-0.2921285927,0.1814135462,-0.3983608484,0.2413861603,-0.4822866023,0.1300214827,-0.1615929604,0.0125321653,0.0205259118,0.2635143101,0.4210568666,0.2929509878,0.210457325,-0.2632245421,0.183308199,0.1879091561,-0.0303863995,0.1382540464,0.1980026811,0.427703172,0.2182935923,-0.0025749032,-0.3060588837,-0.378777951,0.3667081296,-0.6508225203,-0.1554356366,0.177965939,0.0961790308,-0.1908064187,0.1977982074,-0.120195359,0.1526350081,-0.2993704081,-0.2251215428,-0.2421463728,0.1063049808,-0.2950397432,-0.0109460047,0.1267605275,0.1687352061,0.2033834159,-0.0032161898,0.4691496491,0.1447551847,-0.0030009658,-0.4175774157,0.365311265,0.2870754004,0.0016980814,-0.0793112293,0.3670933247,-0.1423478574,-0.1115620732,0.4028783143,-0.1367293,-0.1410658509,0.1487500966,0.0409576446,-0.5017951131,-0.010038102,-0.1201108545,0.4336763024,-0.1826024354,0.0453266464,0.1114468575,-0.0820746645,-0.1678718925,-0.0720645711,0.3950432837,0.6641337872,-0.0311525594,-0.1508771628,0.0343001485,0.1536741853,0.2091035545,-0.0224782676,0.1260635406,0.2954151034,0.0210266784,0.2190592587,0.1616791487,-0.5407570004,-0.0825528353,-0.1191290393,-0.088856034,-0.0452513881,-0.039928481,-0.0125782918,-0.144595921,-0.077084519,0.1568501741,-0.4318296313,-0.0828227997,0.0029682973,-0.3594414294,-0.2030716389,0.2426879704,-0.2542194426,0.1952984184,0.1229350492,-0.3715901971,0.6814710498,-0.0656106472,-0.0377174579,0.0286733266,-0.2180281281,0.2738712728,-0.0218699072,-0.1736154705,0.0716456547,0.1327930689,-0.0182188973,-0.4321421683,0.129990235,0.2906796932,-0.2665609121,-0.2060015351,-0.1474108994,-0.2199855894,-0.1980659217,0.012501549,-0.2633131146,0.0091224909,0.1802123189,-0.2141342163,-0.2851103544,-0.1754494905,0.088712588,0.1457494944,-0.1454801261,0.1112036109,-0.1030413955,0.079609707,0.4481427372,0.4937630296,0.3303021491,0.0161165334,0.206701979,0.0104518421,-0.132597208,0.0350257233,0.0539596416,0.1724950522,-0.4314732254,0.1335953921,0.1833998114,-0.0259705875,0.0959281996,-0.134539783,0.0390607342,0.0892209709,0.2943095565,-0.2372471541,-0.1204894558,-0.3590610921,-0.0762530491,0.1567562073,-0.2641415,-0.0215879679,0.0554708019,-0.0585485138,-0.0392433479,-0.134301722,-0.2703173459,0.2294283807,0.227495417,0.166766867,0.0886536464,0.0664336383,-0.202432096,-0.0473284535,-0.4831431508,0.2371384501,0.170296818,-0.0470597669,0.0829712749,-0.272865653,-0.1448379457,0.0531097353,0.4124594033,-0.2057670057,-0.0480520688,-0.0068303221,0.0476070605,0.0982540548,0.056281358,-0.0962264687,0.2895222306,-0.333522439,0.0828367248,0.0470942408,-0.1901102662,-0.1628680676,0.0326192603,-0.3311604857,-0.077011466,0.2044581473,-0.1896677464,-0.0671472251,0.1545462906,-0.0388696752,0.2940200865,-0.0448274277,-0.029598739,-0.1341499537,-0.0887797847,-0.2249080986,-0.0249913167,-0.3071593046,0.2817917466,0.1091584787,-0.188946262,-0.131850481,0.1648333818,-0.1732589155,0.1726527959,-0.2086050063,-0.0028176755,0.0246709567,-0.0231984351,-0.5214915276,0.0940884873,0.1555965096,-0.1103254259,0.0398592912,0.0614583492,0.1027319655,0.3387366235,-0.2706692815,-0.3621058762,0.2669712007,0.0326420702,0.0121546034,-0.0966365784,0.1388606876,-0.0262364186,0.0883742943,0.3146154881,-0.2663766742,-0.2583737969,-0.176283434,-0.2877731919,0.2703225315,0.0103182225,-0.1003650278,0.3193663955,-0.1577481478,0.1230217591,-0.4959367812,-0.3708249331,-0.6145672202,0.1839373857,-0.4545835555,-0.1591233909,-0.0006665088,0.1189976782,-0.076492399,0.003376527,-0.0885705426,0.0732885152,0.2426484078,-0.4084762633,0.4219225943,-0.2583235502,-0.2140686661,-0.1020656526,0.4282715917,0.3861439824,0.2228370607,-0.029499514,0.1557049602,0.3310789168,-0.0301491953,-0.0243961662,0.1702740192,0.5506826043,-0.014459135,0.1274460256,-0.2651349306,0.1473625451,-0.0279888343,-0.1648572385,0.0146296034,-0.2897578478,0.0721386969,0.0970281363,0.2634501457,-0.0200438146,0.1203501672,0.0114500942,-0.3101598918,-0.1134309843,0.0694142729,0.5498431325,0.0100728869,0.1118492857,-0.3050127029,-0.380580157,0.1755414754,0.1460656226,0.2111584097,-0.1778202355,0.0592719167,0.1955932826,0.2941496074,0.4730266333,-0.1243667975,0.1617894918,0.3088487387,0.2017328143,0.245978266,0.1690641344,-0.369078964,-0.4282012582,0.1055850685,-0.1016469225,-0.4301854074,-0.1570993811,0.3982708454,-0.0855370238,-0.1856926382,-0.1426783204,-0.6614421606,0.1982615739,0.1989426315,0.4220896363,0.0953003392,0.4285657108,0.2011746168,0.264169544,0.096722737,-0.0111392774,0.3545060754,0.0593799762,0.219525516,0.09096504,-0.0481325239,0.408518374,-0.1914631873,0.2222338915,0.0590281971,-0.1055684239,-0.5071742535,0.0968345702,0.0993156061,0.5091096759,0.1225114241,0.2801995277,0.0754743889,-0.2258046567,0.2871660888,-0.2117867172,0.3690875173,0.2239632457,-0.1188632995,0.1157871559,-0.1464339942,0.2851332426,0.1957730353,0.0480547398,-0.3150913119,0.4376888275,0.0144547606,0.1409292221,0.4275515378,0.9575660825,0.0730612352,0.043625202,0.2549076974,-0.0326426663,0.7441516519,-0.2148562521,0.0897927657,-0.0683507249,0.2148378044,0.0229260046,-0.0022365318,0.0607354753,-0.0439576544,-0.2171159536,0.6109153628,-0.1733095497,-0.1176037863,0.0246897917,-0.0270934608,0.1040150598,-0.1634567976,-0.08123485,-0.0723635256,-0.0522958934,0.228264764,-0.0107938908,-0.0551181249,0.2377988398,-0.2094125301,-0.2974385023,-0.0019059932,0.026966719,-0.5267813206,0.1464626491,-0.2675234973,0.3607989252,0.0321017168,0.4635218978,0.4184094071,-0.1807469875,0.1051165834,-0.1527440548,-0.0067377514,-0.1456900686,0.2480383664,0.0990754887,-0.0205115955,-0.5675644279,0.3666430414,-0.240654707,0.1931577325,-0.1059942693,-0.1459557712,0.5476580262,-0.0366551988,0.1045986339,0.0569630042,0.3916785419,0.071164161,0.053894978,0.0744633377,-0.0193626489,-0.1514970511,0.0909149125,-0.50534904,0.1630581617,0.0983706415,0.2054315656,-0.4433358908,0.3906980455,-0.4139960408,0.0797684267,-0.0992844403,-0.0609588772,0.2084016502,-0.1374356896,0.0007641098,0.3673326671,-0.0868682638,0.0770058706,0.4477491379,0.1302054524,0.2307725698,-0.0381083526,0.068226032,-0.1208127514,0.2397720665,-0.2980870605,-0.167578131,0.037303973,0.3075712025,0.330141753,-0.0143578053,-0.1338278651,-0.1941985339,-0.5053734779,0.1580909789,-0.434317708,0.1574808955,0.2092007846,-0.1337432861,-0.1719532609,0.1055791155,-0.2320793867,-0.0134112621,-0.1455431134,0.2351508439,0.0384776406,0.1823839843,-0.2304267138,-0.0599023141,0.1658803225,-0.0532423109,-0.1679231972,0.1551945806,0.0945272744,0.5458922386,-0.0637573898,0.417698741,-0.2863662243,-0.281524986,0.2375330627,0.2273809016,-0.1116865054,-0.0916641578,0.0589640848,-0.0863117278,0.156860441,0.0799442977,0.1232949123,0.1209655851,-0.0172105506,0.1809314638,0.1952173263,-0.0611528009,-0.347381711,-0.2122638375,0.0225347094,-0.3977859616,0.2605210543,0.0595740341,-0.2505081892,0.053491082,-0.1754831821,0.6088402867,0.0290054884,0.2386457175,-0.1407231092,-0.0194698963,0.2878546715,0.1851087958,0.1593906134,-0.2071877718,0.0104963807,0.1525929719,-0.2747837603,0.0343087465,0.3500215113,-0.0032987602,0.2054754496,0.3160900176,0.1943966448,0.5569949746,-0.3445813954,0.1917452961,-0.3911083639,0.036738649,-0.4599284232,0.3406831622,0.172146827,0.0862152874,-0.0170710348,0.338798672,-0.1774634272,-0.1692736149,0.130664587,-0.2018558234,0.0165588576,0.1536884904,-0.306333065,0.0559612736,-0.0162458587,-0.0321231596,0.329938978,0.031006325,-0.0224874187,-0.0865530521,-0.1718521565,0.2852021456,-0.2707394063,0.5228366256,0.3205713928,-0.0340741761,-0.0056130141,0.092193678,-0.0078322133,0.311041683,0.7324509025,0.0361053832,0.1816644818,0.1303397417,-0.3069489598,-0.1171539277,-0.0029303795,0.0315728225,0.1452591866,0.4266273975,-0.2096410692,-0.0988801122,-0.1129491106,-0.0223703738,0.3930734396,-0.0700594485,0.1703604311,-0.1708825082,0.513681829,0.3483603895,0.1214942932,-0.1642851233,-0.0570260957,-0.4837378263,0.4891711473,-0.1586375386,0.1853635758,-0.1041548774,-0.2171944976,-0.0022509024,0.0689254254,0.1474960595,0.2231620252,-0.3148412406,-0.6887147427,0.0700336471,-0.8960809112,0.2602880895,-0.0560411513,0.1209869534,0.2514045537,0.0339552313,-0.463675797,0.2557689846,-0.2205367982,0.1959713697,-0.2555966377,-0.2598942816,-0.2125663757,0.0627960265,-0.0594000109,-0.1981397122,0.0503525846,0.2090266198,0.1086269692,-0.2327677459,-0.0953567922,0.2643117905,-0.1359297186,-0.0132146943,0.9460549355,0.0335832685,0.2863444686,0.339969635,0.0554170087,-0.0282284841,0.1582957357,0.3654692769,-0.1359855086,-0.1616921574,0.0597932562,0.1284500957,-0.3944309354,0.1675124764,-0.0859491006,0.444544822,0.0825097933,-0.2452307492,-0.4501217902,0.2783803344,-0.0388340168,0.1493823677,-0.2521259189,-0.1294004619,0.0998462811,0.1244976446,-0.3044618964,-0.0143179484,0.3623262644,-0.5114169717,-0.1245884076,-0.3022936583,0.3139726818,0.0941005275,-0.3380868435,-1.1964467764,-0.0582515523,0.2086955458,0.3192472458,-0.2933107913,-0.1705363393,-0.4087942243,0.02903115,-0.0939240381,0.2171404213,0.0920901448,0.0459153056,0.3510318398,-0.4274113774]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/237","title":"Can't download MultiNLI","comments":"You should use `load_dataset('glue', 'mnli')`","body":"When I try to download MultiNLI with \r\n```python\r\ndataset = load_dataset('multi_nli')\r\n```\r\n\r\nI get this long error:\r\n```python\r\n---------------------------------------------------------------------------\r\nOSError Traceback (most recent call last)\r\n<ipython-input-13-3b11f6be4cb9> in <module>\r\n 1 # Load a dataset and print the first examples in the training set\r\n 2 # nli_dataset = nlp.load_dataset('multi_nli')\r\n----> 3 dataset = load_dataset('multi_nli')\r\n 4 # nli_dataset = nlp.load_dataset('multi_nli', split='validation_matched[:10%]')\r\n 5 # print(nli_dataset['train'][0])\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 514 \r\n 515 # Download and prepare data\r\n--> 516 builder_instance.download_and_prepare(\r\n 517 download_config=download_config,\r\n 518 download_mode=download_mode,\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 417 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n--> 419 self._download_and_prepare(\r\n 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 455 split_dict = SplitDict(dataset_name=self.name)\r\n 456 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 457 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 458 # Checksums verification\r\n 459 if verify_infos:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\multi_nli\\60774175381b9f3f1e6ae1028229e3cdb270d50379f45b9f2c01008f50f09e6b\\multi_nli.py in _split_generators(self, dl_manager)\r\n 99 def _split_generators(self, dl_manager):\r\n 100 \r\n--> 101 downloaded_dir = dl_manager.download_and_extract(\r\n 102 \"http:\/\/storage.googleapis.com\/tfds-data\/downloads\/multi_nli\/multinli_1.0.zip\"\r\n 103 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in download_and_extract(self, url_or_urls)\r\n 214 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 215 \"\"\"\r\n--> 216 return self.extract(self.download(url_or_urls))\r\n 217 \r\n 218 def get_recorded_sizes_checksums(self):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in extract(self, path_or_paths)\r\n 194 path_or_paths.\r\n 195 \"\"\"\r\n--> 196 return map_nested(\r\n 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 168 return tuple(mapped)\r\n 169 # Singleton\r\n--> 170 return function(data_struct)\r\n 171 \r\n 172 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in <lambda>(path)\r\n 195 \"\"\"\r\n 196 return map_nested(\r\n--> 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n 199 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 231 if is_zipfile(output_path):\r\n 232 with ZipFile(output_path, \"r\") as zip_file:\r\n--> 233 zip_file.extractall(output_path_extracted)\r\n 234 zip_file.close()\r\n 235 elif tarfile.is_tarfile(output_path):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in extractall(self, path, members, pwd)\r\n 1644 \r\n 1645 for zipinfo in members:\r\n-> 1646 self._extract_member(zipinfo, path, pwd)\r\n 1647 \r\n 1648 @classmethod\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in _extract_member(self, member, targetpath, pwd)\r\n 1698 \r\n 1699 with self.open(member, pwd=pwd) as source, \\\r\n-> 1700 open(targetpath, \"wb\") as target:\r\n 1701 shutil.copyfileobj(source, target)\r\n 1702 \r\n\r\nOSError: [Errno 22] Invalid argument: 'C:\\\\Users\\\\Python\\\\.cache\\\\huggingface\\\\datasets\\\\3e12413b8ec69f22dfcfd54a79d1ba9e7aac2e18e334bbb6b81cca64fd16bffc\\\\multinli_1.0\\\\Icon\\r'\r\n```\r\n","comment_length":5,"text":"Can't download MultiNLI\nWhen I try to download MultiNLI with \r\n```python\r\ndataset = load_dataset('multi_nli')\r\n```\r\n\r\nI get this long error:\r\n```python\r\n---------------------------------------------------------------------------\r\nOSError Traceback (most recent call last)\r\n<ipython-input-13-3b11f6be4cb9> in <module>\r\n 1 # Load a dataset and print the first examples in the training set\r\n 2 # nli_dataset = nlp.load_dataset('multi_nli')\r\n----> 3 dataset = load_dataset('multi_nli')\r\n 4 # nli_dataset = nlp.load_dataset('multi_nli', split='validation_matched[:10%]')\r\n 5 # print(nli_dataset['train'][0])\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 514 \r\n 515 # Download and prepare data\r\n--> 516 builder_instance.download_and_prepare(\r\n 517 download_config=download_config,\r\n 518 download_mode=download_mode,\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 417 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n--> 419 self._download_and_prepare(\r\n 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 455 split_dict = SplitDict(dataset_name=self.name)\r\n 456 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 457 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 458 # Checksums verification\r\n 459 if verify_infos:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\multi_nli\\60774175381b9f3f1e6ae1028229e3cdb270d50379f45b9f2c01008f50f09e6b\\multi_nli.py in _split_generators(self, dl_manager)\r\n 99 def _split_generators(self, dl_manager):\r\n 100 \r\n--> 101 downloaded_dir = dl_manager.download_and_extract(\r\n 102 \"http:\/\/storage.googleapis.com\/tfds-data\/downloads\/multi_nli\/multinli_1.0.zip\"\r\n 103 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in download_and_extract(self, url_or_urls)\r\n 214 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 215 \"\"\"\r\n--> 216 return self.extract(self.download(url_or_urls))\r\n 217 \r\n 218 def get_recorded_sizes_checksums(self):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in extract(self, path_or_paths)\r\n 194 path_or_paths.\r\n 195 \"\"\"\r\n--> 196 return map_nested(\r\n 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 168 return tuple(mapped)\r\n 169 # Singleton\r\n--> 170 return function(data_struct)\r\n 171 \r\n 172 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in <lambda>(path)\r\n 195 \"\"\"\r\n 196 return map_nested(\r\n--> 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n 199 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 231 if is_zipfile(output_path):\r\n 232 with ZipFile(output_path, \"r\") as zip_file:\r\n--> 233 zip_file.extractall(output_path_extracted)\r\n 234 zip_file.close()\r\n 235 elif tarfile.is_tarfile(output_path):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in extractall(self, path, members, pwd)\r\n 1644 \r\n 1645 for zipinfo in members:\r\n-> 1646 self._extract_member(zipinfo, path, pwd)\r\n 1647 \r\n 1648 @classmethod\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in _extract_member(self, member, targetpath, pwd)\r\n 1698 \r\n 1699 with self.open(member, pwd=pwd) as source, \\\r\n-> 1700 open(targetpath, \"wb\") as target:\r\n 1701 shutil.copyfileobj(source, target)\r\n 1702 \r\n\r\nOSError: [Errno 22] Invalid argument: 'C:\\\\Users\\\\Python\\\\.cache\\\\huggingface\\\\datasets\\\\3e12413b8ec69f22dfcfd54a79d1ba9e7aac2e18e334bbb6b81cca64fd16bffc\\\\multinli_1.0\\\\Icon\\r'\r\n```\r\n\nYou should use `load_dataset('glue', 'mnli')`","embeddings":[-0.0840254948,-0.0221919809,-0.0726869553,0.0423389748,0.1981838346,0.0664926246,0.457182169,0.0801374689,0.1033792794,0.0883751884,-0.1641247272,0.2453239113,-0.1955309659,-0.0816973373,-0.0615135431,-0.2790237069,-0.0329709835,-0.0675984398,0.028705474,0.1370823085,-0.2184431851,0.1371101737,-0.2122425735,0.2103482187,-0.1834195554,-0.1558727026,-0.2034406811,0.2328009605,-0.338126868,-0.360994488,0.1864654124,-0.1097058058,0.2480324358,0.0156800989,-0.0001071728,0.1085644364,0.1779890656,-0.0657087117,-0.406357944,-0.1557623595,-0.1127581894,-0.3171509802,0.1604038626,-0.3909189105,0.3036190569,0.3498140872,0.0443298407,0.0456054173,0.1174699664,0.4848155379,0.2677161992,0.4489547312,0.3633489907,-0.0720728859,-0.0470621996,-0.2057535052,0.0127581256,0.3778342307,-0.1025850102,-0.176113531,0.0909408852,0.0736363828,-0.1486242712,0.3225732148,-0.1937520951,-0.0052935621,0.3084739447,-0.2554347217,-0.0774453729,0.2574424744,0.0329429805,0.0830193385,-0.3229765594,0.0241578445,0.2695622742,-0.3627426624,0.0999657214,0.4350741804,-0.3351387978,-0.0922091752,-0.246600464,-0.1657491475,-0.1589851528,0.3905088007,-0.3112727106,0.5100572109,0.1047358438,0.1057526469,0.43421489,-0.2032002658,-0.2787941992,-0.053084895,0.0459684543,0.1960389763,-0.3286784589,0.0189728905,-0.0301640481,-0.2324981689,0.158440128,0.1301026046,-0.0234435536,-0.2342739105,-0.1908606142,0.2034369111,0.0483432971,0.0642103851,-0.2885984182,0.1249965057,0.10281533,0.2291225791,0.2289705127,0.0392173566,-0.1188764647,-0.0918307155,-0.116723232,0.0861231312,0.2541342378,0.0922482461,-0.4595086873,-0.1994170845,-0.5215702653,-0.019995559,0.2925751209,0.1465471834,-0.1126548126,0.3401513398,0.1732837707,0.3975913823,-0.2035049647,-0.3343687654,-0.1372718513,0.1671380103,-0.1231139451,-0.0419097692,0.435234338,-0.0595215447,0.3110079169,-0.0159450583,-0.0854558572,-0.0951857194,0.1114431769,-0.0950859785,-0.3110278845,0.2561810315,0.170438081,0.0722104535,0.1697930694,-0.1313999444,-0.0520505607,0.0058724899,-0.3391163945,-0.4312471151,-0.3035261929,0.2793096006,0.0968348831,-0.105003126,0.1401997209,-0.4007652998,0.2124014944,-0.0330316164,-0.0638782978,-0.2944987416,-0.0988622084,-0.2295627892,0.259360671,0.2700573802,0.1803033054,-0.0016141343,-0.349278003,-0.1656953394,0.2533250153,0.5466470718,-0.2035302222,0.1508474648,-0.1804694384,0.0348155797,0.5982265472,-0.4164569676,-0.0977124944,0.2224085629,-0.3346424699,0.0443879105,0.0076296059,0.179850027,0.0642710999,0.1057124212,0.3237139881,0.5148923397,-0.1797897667,0.079438813,-0.1894708723,-0.2525600791,0.1889620721,0.37766397,0.0521411523,0.0710817426,-0.1293640584,0.4729548395,0.4586094022,0.0054114168,0.0687954426,0.0822698623,0.145469591,-0.090170987,-0.3268409967,-0.0332494974,-0.3870306611,0.127906695,0.0395684429,0.103273347,-0.026721485,0.0882387906,-0.1036591455,0.0494505912,-0.1920175254,-0.1252931952,0.1495797187,0.2034477443,0.0919264555,0.0242259167,0.0234373137,0.516964972,-0.0308067165,0.1307230592,-0.4853770137,0.2479034066,-0.3664743602,-0.0430885442,0.0266262665,0.0217892621,-0.0637478679,0.0476308279,-0.0967497304,0.3046732545,-0.0396848843,-0.4234547317,-0.104751274,-0.1042764485,0.1154831797,-0.1144754738,0.1306092739,0.3364742696,0.0666871145,-0.0499959737,-0.3063918948,0.3337834477,0.0307656657,0.0900291875,-0.0031021968,0.3673850894,0.3018737435,-0.020218024,-0.0372559763,-0.1680344045,0.3337793052,0.0372740142,0.0485329591,-0.2252466381,-0.3169656098,-0.1664787829,0.3011681437,0.0414901115,0.0305451453,0.0421516337,0.1887793392,0.0492153913,-0.1162228882,0.0048489976,0.4669987857,0.0854075626,0.109497644,-0.0226920247,-0.1893053204,-0.2992962599,0.064426057,-0.0190867856,0.2057503611,-0.0591562502,0.130966872,-0.1099563316,-0.0401301496,-0.4957756102,0.2854011953,0.4715655744,-0.2018342465,0.0635068342,-0.3473848403,-0.5306869745,-0.2337204665,-0.2072554231,-0.3806086183,-0.3395351171,-0.1652444899,0.4490947723,0.0872477144,0.0202280376,-0.0163672008,-0.2310703546,0.2713745236,-0.0395486802,0.203712821,-0.071269013,-0.2772781551,0.1511174887,0.3364775479,-0.0214621387,0.2527517974,-0.1002357751,-0.173353374,-0.268103987,0.0158549752,-0.2235613465,0.0938835666,-0.264356792,0.2511912286,0.3671227098,0.046039965,-0.2325368524,0.2155634612,-0.0189510845,-0.1420431137,0.3279959261,0.0061510629,0.0973288193,-0.1499354541,-0.3068360388,-0.3677200377,-0.4596532583,0.2124171704,-0.0558876693,0.2125092894,0.374065578,0.0455109701,0.1976522058,0.0339022465,-0.1330120265,0.0363395475,0.0478519276,0.1572207958,-0.1895360351,-0.3011138439,-0.078162238,0.2041410804,0.2132148594,0.0037161643,-0.3780254722,-0.0257090516,0.0308423396,0.2746254802,-0.1909338534,-0.0438152179,0.2122442573,-0.3453694284,-0.0411612727,0.2166718096,0.0430256687,0.08033517,0.2186608016,0.4515864849,-0.1415331513,0.5075715184,0.0484841429,0.3951174617,0.2567915022,-0.0194629226,0.3725375533,-0.0210259911,-0.0696513951,0.0216038134,-0.4024845064,-0.0067745629,-0.1218655184,0.1562472284,-0.0280966498,-0.108148396,-0.3482307792,-0.17430453,0.0958466232,0.1026307717,-0.2052616328,0.0999547988,-0.243960917,0.026900338,0.0105684996,-0.0934093297,-0.0357636884,-0.1712870151,-0.3353484571,0.3295469582,-0.2056181431,-0.0075147203,-0.0646636635,-0.1637697369,-0.2538139224,0.0381051078,0.1786724627,0.5281776786,0.0313175805,0.1133058146,0.1517692059,-0.0579215512,0.366265893,0.0431779437,0.1123287231,0.3103982806,0.1655441672,-0.5614237785,-0.3184332252,-0.3913879693,0.2710352838,0.7224541306,0.2256345153,-0.1329276562,-0.0833349377,0.3776691854,0.1377887726,-0.0911204144,-0.1700523645,-0.3710435629,-0.3217700422,-0.4575541019,0.054056026,-0.0656715706,0.3394945264,0.1082355306,0.2655141056,0.122991398,0.001990254,-0.1314436942,-0.0960291401,-0.1669194996,0.2519840598,0.1203190908,-0.2048624605,0.1100886837,-0.0121516995,0.5492684841,-0.0537355542,-0.219985351,0.2330361307,-0.0417358167,0.0177323762,0.0470406264,0.1685984582,0.117260173,0.0961255133,0.1955198497,0.3162567317,0.1779068708,0.2874184251,-0.266925633,-0.2842011154,-0.2203262299,0.4090629518,0.0650444776,0.1939111203,0.2437657565,-0.5065380931,0.0695204362,-0.038530454,0.1428883374,0.5448486209,-0.1601898819,0.2056912631,0.369762361,-0.1422553509,0.4998179078,-0.3342645466,0.1792706996,-0.19360286,0.2306205481,0.084226191,0.0270975754,-0.0335593112,0.1626606733,-0.2206200808,0.1645316631,0.129934594,-0.0553147718,0.1815406978,0.5320366621,0.0094062435,-0.3416405618,-0.2274582535,0.2065631896,-0.1639429927,0.3153637052,-0.0948611796,-0.0771327764,0.109076336,-0.1655254811,-0.3335944116,0.2814983726,-0.0834192485,0.5441380739,-0.0700275153,-0.0536753051,-0.0654046685,0.2200279981,0.0794126615,-0.1387498081,-0.0544516519,-0.0917550325,-0.0884905159,-0.274774909,-0.1351374239,0.0227739383,0.2547706664,-0.31023857,-0.3599821627,0.2084232271,0.043487519,-0.2768305242,0.1300090998,0.2966113985,-0.1079709679,0.1171940863,0.0799803808,-0.0604163483,0.1019484773,-0.091481328,0.1378287226,0.1844536662,-0.2723405361,0.0384070612,-0.0725497305,-0.126545012,0.0302758999,0.5448255539,0.0795113966,0.0146676674,0.4382507503,0.2637680769,0.0487562343,-0.3331158757,0.1242254376,-0.2090209126,-0.2165759206,-0.0329977721,0.0389123447,0.2234701365,0.0592571534,0.1032000259,0.1449246407,0.1771889478,0.1525190771,-0.6146883965,-0.4120766521,-0.0507671908,-0.1135169566,-0.1017977074,-0.0972705856,0.2015437484,0.2837581635,0.0364494883,-0.2931175232,0.029571142,-0.3940633833,0.0747049004,0.443446219,-0.2955845296,-0.0098030576,0.2064347267,0.2212134749,-0.1965438724,-0.1449277401,-0.2474171668,-0.026847383,0.0683801845,0.0169644486,-0.1980533749,0.1576873064,-0.3512070179,-0.2484963834,0.1924411058,-0.0722915605,0.038387455,0.0708764419,0.122181423,-0.1602078825,0.1678972542,-0.0834501088,0.0146519588,-0.2705078423,0.0771098062,0.1117303222,0.3172302544,0.2172395736,-0.0045611658,-0.6347709894,-0.1739444435,-0.2452757657,-0.0842615291,0.0953229442,-0.3559564352,-0.0140274074,-0.2292230129,0.061665494,0.2297003418,-0.1112764478,0.0435652323,0.3644263148,0.2506194413,-0.1571569294,-0.0137904389,0.2758883238,0.0277180467,0.0098809311,0.2398214787,-0.104826726,-0.0761379525,0.1575665772,0.0776044205,0.1016805097,0.0976478979,-0.0092486562,0.3073760867,-0.1990081966,-0.0538323224,0.036504183,-0.004383883,0.2784198225,0.313498646,-0.2719894648,0.2696474195,-0.0701691806,0.0511826873,0.1469730884,-0.2084991187,-0.3138246834,-0.0141030736,-0.2458355427,-0.0295137689,-0.1594554037,0.3483742476,-0.3062147498,-0.2602591813,-0.3952552676,0.5272894502,-0.1780480593,0.0531285182,-0.3437132537,-0.0936110541,-0.2732801437,0.1869675666,0.0505930707,0.0776092857,0.4932879806,0.3244490027,-0.1509422213,-0.5531567335,-0.1860025376,0.2678630054,0.0156636816,-0.3500516415,0.2893440127,0.4177259803,-0.2784986794,-0.0091645466,0.0957595855,0.5214709044,0.2982955277,0.1787415594,0.1002323553,-0.0472967327,-0.0325361378,-0.0254061408,-0.0573556684,0.4776994884,0.2105083466,0.3830700815,0.0977295265,-0.2219573557,0.1700776219,0.0921241343,-0.105059281,-0.1036489606,0.4112376571,-0.0219001193,0.0097767729,-0.2397839427,0.0670564026,-0.5321535468,0.0814334303,0.0267071966,0.1320957392,0.1471248418,0.0453968905,0.0507086217,0.0097324513,0.5457183719,0.3034781218,0.2115044296,-0.2945046127,-0.2823817432,-0.6838545203,0.0972865373,-0.5811139941,-0.1010993719,-0.0463687479,0.2308626026,-0.2003297508,0.2188737988,0.0341939181,0.0345413908,0.1818779707,0.2735221684,-0.3600364923,-0.0847819075,0.2178517282,-0.0189787764,0.0338429622,-0.3433586657,0.0698510334,-0.2905687988,0.0890476033,-0.2007319629,-0.0783626437,0.1026028916,0.2820727527,0.0038187257,0.1270779669,0.4972711504,0.1928167641,-0.3524492383,-0.2249494642,0.2027591467,-0.1610056609,0.0131931314,0.2036267668,0.4022216499,-0.1029094979,0.1434143186,-0.3461368084,0.5090784431,-0.1239434853,-0.1700542718,-0.0795723647,0.0544737875,0.056328211,0.1147856936,0.2220103294,0.0350599065,-0.0253666025,0.0394006297,-0.3940591216,-0.2115023285,0.3609002233,-0.0975072682,-0.2655795813,-0.1804896444,0.2269531041,-0.3209328353,0.2880321443,-0.5555567145,0.1251621544,0.296951443,-0.0823288411,-0.5224146843,0.1200467423,0.1242874041,-0.1698413789,-0.0072111064,0.0002059927,0.0807665884,-0.4523905218,0.2310978621,-0.1104964763]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/237","title":"Can't download MultiNLI","comments":"Thanks! I thought I had to use the same code displayed in the live viewer:\r\n```python\r\n!pip install nlp\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('multi_nli', 'plain_text')\r\n```\r\nYour suggestion works, even if then I got a different issue (#242). ","body":"When I try to download MultiNLI with \r\n```python\r\ndataset = load_dataset('multi_nli')\r\n```\r\n\r\nI get this long error:\r\n```python\r\n---------------------------------------------------------------------------\r\nOSError Traceback (most recent call last)\r\n<ipython-input-13-3b11f6be4cb9> in <module>\r\n 1 # Load a dataset and print the first examples in the training set\r\n 2 # nli_dataset = nlp.load_dataset('multi_nli')\r\n----> 3 dataset = load_dataset('multi_nli')\r\n 4 # nli_dataset = nlp.load_dataset('multi_nli', split='validation_matched[:10%]')\r\n 5 # print(nli_dataset['train'][0])\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 514 \r\n 515 # Download and prepare data\r\n--> 516 builder_instance.download_and_prepare(\r\n 517 download_config=download_config,\r\n 518 download_mode=download_mode,\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 417 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n--> 419 self._download_and_prepare(\r\n 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 455 split_dict = SplitDict(dataset_name=self.name)\r\n 456 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 457 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 458 # Checksums verification\r\n 459 if verify_infos:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\multi_nli\\60774175381b9f3f1e6ae1028229e3cdb270d50379f45b9f2c01008f50f09e6b\\multi_nli.py in _split_generators(self, dl_manager)\r\n 99 def _split_generators(self, dl_manager):\r\n 100 \r\n--> 101 downloaded_dir = dl_manager.download_and_extract(\r\n 102 \"http:\/\/storage.googleapis.com\/tfds-data\/downloads\/multi_nli\/multinli_1.0.zip\"\r\n 103 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in download_and_extract(self, url_or_urls)\r\n 214 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 215 \"\"\"\r\n--> 216 return self.extract(self.download(url_or_urls))\r\n 217 \r\n 218 def get_recorded_sizes_checksums(self):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in extract(self, path_or_paths)\r\n 194 path_or_paths.\r\n 195 \"\"\"\r\n--> 196 return map_nested(\r\n 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 168 return tuple(mapped)\r\n 169 # Singleton\r\n--> 170 return function(data_struct)\r\n 171 \r\n 172 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in <lambda>(path)\r\n 195 \"\"\"\r\n 196 return map_nested(\r\n--> 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n 199 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 231 if is_zipfile(output_path):\r\n 232 with ZipFile(output_path, \"r\") as zip_file:\r\n--> 233 zip_file.extractall(output_path_extracted)\r\n 234 zip_file.close()\r\n 235 elif tarfile.is_tarfile(output_path):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in extractall(self, path, members, pwd)\r\n 1644 \r\n 1645 for zipinfo in members:\r\n-> 1646 self._extract_member(zipinfo, path, pwd)\r\n 1647 \r\n 1648 @classmethod\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in _extract_member(self, member, targetpath, pwd)\r\n 1698 \r\n 1699 with self.open(member, pwd=pwd) as source, \\\r\n-> 1700 open(targetpath, \"wb\") as target:\r\n 1701 shutil.copyfileobj(source, target)\r\n 1702 \r\n\r\nOSError: [Errno 22] Invalid argument: 'C:\\\\Users\\\\Python\\\\.cache\\\\huggingface\\\\datasets\\\\3e12413b8ec69f22dfcfd54a79d1ba9e7aac2e18e334bbb6b81cca64fd16bffc\\\\multinli_1.0\\\\Icon\\r'\r\n```\r\n","comment_length":40,"text":"Can't download MultiNLI\nWhen I try to download MultiNLI with \r\n```python\r\ndataset = load_dataset('multi_nli')\r\n```\r\n\r\nI get this long error:\r\n```python\r\n---------------------------------------------------------------------------\r\nOSError Traceback (most recent call last)\r\n<ipython-input-13-3b11f6be4cb9> in <module>\r\n 1 # Load a dataset and print the first examples in the training set\r\n 2 # nli_dataset = nlp.load_dataset('multi_nli')\r\n----> 3 dataset = load_dataset('multi_nli')\r\n 4 # nli_dataset = nlp.load_dataset('multi_nli', split='validation_matched[:10%]')\r\n 5 # print(nli_dataset['train'][0])\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 514 \r\n 515 # Download and prepare data\r\n--> 516 builder_instance.download_and_prepare(\r\n 517 download_config=download_config,\r\n 518 download_mode=download_mode,\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 417 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n--> 419 self._download_and_prepare(\r\n 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 455 split_dict = SplitDict(dataset_name=self.name)\r\n 456 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 457 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 458 # Checksums verification\r\n 459 if verify_infos:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\multi_nli\\60774175381b9f3f1e6ae1028229e3cdb270d50379f45b9f2c01008f50f09e6b\\multi_nli.py in _split_generators(self, dl_manager)\r\n 99 def _split_generators(self, dl_manager):\r\n 100 \r\n--> 101 downloaded_dir = dl_manager.download_and_extract(\r\n 102 \"http:\/\/storage.googleapis.com\/tfds-data\/downloads\/multi_nli\/multinli_1.0.zip\"\r\n 103 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in download_and_extract(self, url_or_urls)\r\n 214 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 215 \"\"\"\r\n--> 216 return self.extract(self.download(url_or_urls))\r\n 217 \r\n 218 def get_recorded_sizes_checksums(self):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in extract(self, path_or_paths)\r\n 194 path_or_paths.\r\n 195 \"\"\"\r\n--> 196 return map_nested(\r\n 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 168 return tuple(mapped)\r\n 169 # Singleton\r\n--> 170 return function(data_struct)\r\n 171 \r\n 172 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in <lambda>(path)\r\n 195 \"\"\"\r\n 196 return map_nested(\r\n--> 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n 199 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 231 if is_zipfile(output_path):\r\n 232 with ZipFile(output_path, \"r\") as zip_file:\r\n--> 233 zip_file.extractall(output_path_extracted)\r\n 234 zip_file.close()\r\n 235 elif tarfile.is_tarfile(output_path):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in extractall(self, path, members, pwd)\r\n 1644 \r\n 1645 for zipinfo in members:\r\n-> 1646 self._extract_member(zipinfo, path, pwd)\r\n 1647 \r\n 1648 @classmethod\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in _extract_member(self, member, targetpath, pwd)\r\n 1698 \r\n 1699 with self.open(member, pwd=pwd) as source, \\\r\n-> 1700 open(targetpath, \"wb\") as target:\r\n 1701 shutil.copyfileobj(source, target)\r\n 1702 \r\n\r\nOSError: [Errno 22] Invalid argument: 'C:\\\\Users\\\\Python\\\\.cache\\\\huggingface\\\\datasets\\\\3e12413b8ec69f22dfcfd54a79d1ba9e7aac2e18e334bbb6b81cca64fd16bffc\\\\multinli_1.0\\\\Icon\\r'\r\n```\r\n\nThanks! I thought I had to use the same code displayed in the live viewer:\r\n```python\r\n!pip install nlp\r\nfrom nlp import load_dataset\r\ndataset = load_dataset('multi_nli', 'plain_text')\r\n```\r\nYour suggestion works, even if then I got a different issue (#242). ","embeddings":[-0.0840254948,-0.0221919809,-0.0726869553,0.0423389748,0.1981838346,0.0664926246,0.457182169,0.0801374689,0.1033792794,0.0883751884,-0.1641247272,0.2453239113,-0.1955309659,-0.0816973373,-0.0615135431,-0.2790237069,-0.0329709835,-0.0675984398,0.028705474,0.1370823085,-0.2184431851,0.1371101737,-0.2122425735,0.2103482187,-0.1834195554,-0.1558727026,-0.2034406811,0.2328009605,-0.338126868,-0.360994488,0.1864654124,-0.1097058058,0.2480324358,0.0156800989,-0.0001071728,0.1085644364,0.1779890656,-0.0657087117,-0.406357944,-0.1557623595,-0.1127581894,-0.3171509802,0.1604038626,-0.3909189105,0.3036190569,0.3498140872,0.0443298407,0.0456054173,0.1174699664,0.4848155379,0.2677161992,0.4489547312,0.3633489907,-0.0720728859,-0.0470621996,-0.2057535052,0.0127581256,0.3778342307,-0.1025850102,-0.176113531,0.0909408852,0.0736363828,-0.1486242712,0.3225732148,-0.1937520951,-0.0052935621,0.3084739447,-0.2554347217,-0.0774453729,0.2574424744,0.0329429805,0.0830193385,-0.3229765594,0.0241578445,0.2695622742,-0.3627426624,0.0999657214,0.4350741804,-0.3351387978,-0.0922091752,-0.246600464,-0.1657491475,-0.1589851528,0.3905088007,-0.3112727106,0.5100572109,0.1047358438,0.1057526469,0.43421489,-0.2032002658,-0.2787941992,-0.053084895,0.0459684543,0.1960389763,-0.3286784589,0.0189728905,-0.0301640481,-0.2324981689,0.158440128,0.1301026046,-0.0234435536,-0.2342739105,-0.1908606142,0.2034369111,0.0483432971,0.0642103851,-0.2885984182,0.1249965057,0.10281533,0.2291225791,0.2289705127,0.0392173566,-0.1188764647,-0.0918307155,-0.116723232,0.0861231312,0.2541342378,0.0922482461,-0.4595086873,-0.1994170845,-0.5215702653,-0.019995559,0.2925751209,0.1465471834,-0.1126548126,0.3401513398,0.1732837707,0.3975913823,-0.2035049647,-0.3343687654,-0.1372718513,0.1671380103,-0.1231139451,-0.0419097692,0.435234338,-0.0595215447,0.3110079169,-0.0159450583,-0.0854558572,-0.0951857194,0.1114431769,-0.0950859785,-0.3110278845,0.2561810315,0.170438081,0.0722104535,0.1697930694,-0.1313999444,-0.0520505607,0.0058724899,-0.3391163945,-0.4312471151,-0.3035261929,0.2793096006,0.0968348831,-0.105003126,0.1401997209,-0.4007652998,0.2124014944,-0.0330316164,-0.0638782978,-0.2944987416,-0.0988622084,-0.2295627892,0.259360671,0.2700573802,0.1803033054,-0.0016141343,-0.349278003,-0.1656953394,0.2533250153,0.5466470718,-0.2035302222,0.1508474648,-0.1804694384,0.0348155797,0.5982265472,-0.4164569676,-0.0977124944,0.2224085629,-0.3346424699,0.0443879105,0.0076296059,0.179850027,0.0642710999,0.1057124212,0.3237139881,0.5148923397,-0.1797897667,0.079438813,-0.1894708723,-0.2525600791,0.1889620721,0.37766397,0.0521411523,0.0710817426,-0.1293640584,0.4729548395,0.4586094022,0.0054114168,0.0687954426,0.0822698623,0.145469591,-0.090170987,-0.3268409967,-0.0332494974,-0.3870306611,0.127906695,0.0395684429,0.103273347,-0.026721485,0.0882387906,-0.1036591455,0.0494505912,-0.1920175254,-0.1252931952,0.1495797187,0.2034477443,0.0919264555,0.0242259167,0.0234373137,0.516964972,-0.0308067165,0.1307230592,-0.4853770137,0.2479034066,-0.3664743602,-0.0430885442,0.0266262665,0.0217892621,-0.0637478679,0.0476308279,-0.0967497304,0.3046732545,-0.0396848843,-0.4234547317,-0.104751274,-0.1042764485,0.1154831797,-0.1144754738,0.1306092739,0.3364742696,0.0666871145,-0.0499959737,-0.3063918948,0.3337834477,0.0307656657,0.0900291875,-0.0031021968,0.3673850894,0.3018737435,-0.020218024,-0.0372559763,-0.1680344045,0.3337793052,0.0372740142,0.0485329591,-0.2252466381,-0.3169656098,-0.1664787829,0.3011681437,0.0414901115,0.0305451453,0.0421516337,0.1887793392,0.0492153913,-0.1162228882,0.0048489976,0.4669987857,0.0854075626,0.109497644,-0.0226920247,-0.1893053204,-0.2992962599,0.064426057,-0.0190867856,0.2057503611,-0.0591562502,0.130966872,-0.1099563316,-0.0401301496,-0.4957756102,0.2854011953,0.4715655744,-0.2018342465,0.0635068342,-0.3473848403,-0.5306869745,-0.2337204665,-0.2072554231,-0.3806086183,-0.3395351171,-0.1652444899,0.4490947723,0.0872477144,0.0202280376,-0.0163672008,-0.2310703546,0.2713745236,-0.0395486802,0.203712821,-0.071269013,-0.2772781551,0.1511174887,0.3364775479,-0.0214621387,0.2527517974,-0.1002357751,-0.173353374,-0.268103987,0.0158549752,-0.2235613465,0.0938835666,-0.264356792,0.2511912286,0.3671227098,0.046039965,-0.2325368524,0.2155634612,-0.0189510845,-0.1420431137,0.3279959261,0.0061510629,0.0973288193,-0.1499354541,-0.3068360388,-0.3677200377,-0.4596532583,0.2124171704,-0.0558876693,0.2125092894,0.374065578,0.0455109701,0.1976522058,0.0339022465,-0.1330120265,0.0363395475,0.0478519276,0.1572207958,-0.1895360351,-0.3011138439,-0.078162238,0.2041410804,0.2132148594,0.0037161643,-0.3780254722,-0.0257090516,0.0308423396,0.2746254802,-0.1909338534,-0.0438152179,0.2122442573,-0.3453694284,-0.0411612727,0.2166718096,0.0430256687,0.08033517,0.2186608016,0.4515864849,-0.1415331513,0.5075715184,0.0484841429,0.3951174617,0.2567915022,-0.0194629226,0.3725375533,-0.0210259911,-0.0696513951,0.0216038134,-0.4024845064,-0.0067745629,-0.1218655184,0.1562472284,-0.0280966498,-0.108148396,-0.3482307792,-0.17430453,0.0958466232,0.1026307717,-0.2052616328,0.0999547988,-0.243960917,0.026900338,0.0105684996,-0.0934093297,-0.0357636884,-0.1712870151,-0.3353484571,0.3295469582,-0.2056181431,-0.0075147203,-0.0646636635,-0.1637697369,-0.2538139224,0.0381051078,0.1786724627,0.5281776786,0.0313175805,0.1133058146,0.1517692059,-0.0579215512,0.366265893,0.0431779437,0.1123287231,0.3103982806,0.1655441672,-0.5614237785,-0.3184332252,-0.3913879693,0.2710352838,0.7224541306,0.2256345153,-0.1329276562,-0.0833349377,0.3776691854,0.1377887726,-0.0911204144,-0.1700523645,-0.3710435629,-0.3217700422,-0.4575541019,0.054056026,-0.0656715706,0.3394945264,0.1082355306,0.2655141056,0.122991398,0.001990254,-0.1314436942,-0.0960291401,-0.1669194996,0.2519840598,0.1203190908,-0.2048624605,0.1100886837,-0.0121516995,0.5492684841,-0.0537355542,-0.219985351,0.2330361307,-0.0417358167,0.0177323762,0.0470406264,0.1685984582,0.117260173,0.0961255133,0.1955198497,0.3162567317,0.1779068708,0.2874184251,-0.266925633,-0.2842011154,-0.2203262299,0.4090629518,0.0650444776,0.1939111203,0.2437657565,-0.5065380931,0.0695204362,-0.038530454,0.1428883374,0.5448486209,-0.1601898819,0.2056912631,0.369762361,-0.1422553509,0.4998179078,-0.3342645466,0.1792706996,-0.19360286,0.2306205481,0.084226191,0.0270975754,-0.0335593112,0.1626606733,-0.2206200808,0.1645316631,0.129934594,-0.0553147718,0.1815406978,0.5320366621,0.0094062435,-0.3416405618,-0.2274582535,0.2065631896,-0.1639429927,0.3153637052,-0.0948611796,-0.0771327764,0.109076336,-0.1655254811,-0.3335944116,0.2814983726,-0.0834192485,0.5441380739,-0.0700275153,-0.0536753051,-0.0654046685,0.2200279981,0.0794126615,-0.1387498081,-0.0544516519,-0.0917550325,-0.0884905159,-0.274774909,-0.1351374239,0.0227739383,0.2547706664,-0.31023857,-0.3599821627,0.2084232271,0.043487519,-0.2768305242,0.1300090998,0.2966113985,-0.1079709679,0.1171940863,0.0799803808,-0.0604163483,0.1019484773,-0.091481328,0.1378287226,0.1844536662,-0.2723405361,0.0384070612,-0.0725497305,-0.126545012,0.0302758999,0.5448255539,0.0795113966,0.0146676674,0.4382507503,0.2637680769,0.0487562343,-0.3331158757,0.1242254376,-0.2090209126,-0.2165759206,-0.0329977721,0.0389123447,0.2234701365,0.0592571534,0.1032000259,0.1449246407,0.1771889478,0.1525190771,-0.6146883965,-0.4120766521,-0.0507671908,-0.1135169566,-0.1017977074,-0.0972705856,0.2015437484,0.2837581635,0.0364494883,-0.2931175232,0.029571142,-0.3940633833,0.0747049004,0.443446219,-0.2955845296,-0.0098030576,0.2064347267,0.2212134749,-0.1965438724,-0.1449277401,-0.2474171668,-0.026847383,0.0683801845,0.0169644486,-0.1980533749,0.1576873064,-0.3512070179,-0.2484963834,0.1924411058,-0.0722915605,0.038387455,0.0708764419,0.122181423,-0.1602078825,0.1678972542,-0.0834501088,0.0146519588,-0.2705078423,0.0771098062,0.1117303222,0.3172302544,0.2172395736,-0.0045611658,-0.6347709894,-0.1739444435,-0.2452757657,-0.0842615291,0.0953229442,-0.3559564352,-0.0140274074,-0.2292230129,0.061665494,0.2297003418,-0.1112764478,0.0435652323,0.3644263148,0.2506194413,-0.1571569294,-0.0137904389,0.2758883238,0.0277180467,0.0098809311,0.2398214787,-0.104826726,-0.0761379525,0.1575665772,0.0776044205,0.1016805097,0.0976478979,-0.0092486562,0.3073760867,-0.1990081966,-0.0538323224,0.036504183,-0.004383883,0.2784198225,0.313498646,-0.2719894648,0.2696474195,-0.0701691806,0.0511826873,0.1469730884,-0.2084991187,-0.3138246834,-0.0141030736,-0.2458355427,-0.0295137689,-0.1594554037,0.3483742476,-0.3062147498,-0.2602591813,-0.3952552676,0.5272894502,-0.1780480593,0.0531285182,-0.3437132537,-0.0936110541,-0.2732801437,0.1869675666,0.0505930707,0.0776092857,0.4932879806,0.3244490027,-0.1509422213,-0.5531567335,-0.1860025376,0.2678630054,0.0156636816,-0.3500516415,0.2893440127,0.4177259803,-0.2784986794,-0.0091645466,0.0957595855,0.5214709044,0.2982955277,0.1787415594,0.1002323553,-0.0472967327,-0.0325361378,-0.0254061408,-0.0573556684,0.4776994884,0.2105083466,0.3830700815,0.0977295265,-0.2219573557,0.1700776219,0.0921241343,-0.105059281,-0.1036489606,0.4112376571,-0.0219001193,0.0097767729,-0.2397839427,0.0670564026,-0.5321535468,0.0814334303,0.0267071966,0.1320957392,0.1471248418,0.0453968905,0.0507086217,0.0097324513,0.5457183719,0.3034781218,0.2115044296,-0.2945046127,-0.2823817432,-0.6838545203,0.0972865373,-0.5811139941,-0.1010993719,-0.0463687479,0.2308626026,-0.2003297508,0.2188737988,0.0341939181,0.0345413908,0.1818779707,0.2735221684,-0.3600364923,-0.0847819075,0.2178517282,-0.0189787764,0.0338429622,-0.3433586657,0.0698510334,-0.2905687988,0.0890476033,-0.2007319629,-0.0783626437,0.1026028916,0.2820727527,0.0038187257,0.1270779669,0.4972711504,0.1928167641,-0.3524492383,-0.2249494642,0.2027591467,-0.1610056609,0.0131931314,0.2036267668,0.4022216499,-0.1029094979,0.1434143186,-0.3461368084,0.5090784431,-0.1239434853,-0.1700542718,-0.0795723647,0.0544737875,0.056328211,0.1147856936,0.2220103294,0.0350599065,-0.0253666025,0.0394006297,-0.3940591216,-0.2115023285,0.3609002233,-0.0975072682,-0.2655795813,-0.1804896444,0.2269531041,-0.3209328353,0.2880321443,-0.5555567145,0.1251621544,0.296951443,-0.0823288411,-0.5224146843,0.1200467423,0.1242874041,-0.1698413789,-0.0072111064,0.0002059927,0.0807665884,-0.4523905218,0.2310978621,-0.1104964763]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/237","title":"Can't download MultiNLI","comments":"Glad it helps !\nThough I am not one of hf team, but maybe you should close this issue first.","body":"When I try to download MultiNLI with \r\n```python\r\ndataset = load_dataset('multi_nli')\r\n```\r\n\r\nI get this long error:\r\n```python\r\n---------------------------------------------------------------------------\r\nOSError Traceback (most recent call last)\r\n<ipython-input-13-3b11f6be4cb9> in <module>\r\n 1 # Load a dataset and print the first examples in the training set\r\n 2 # nli_dataset = nlp.load_dataset('multi_nli')\r\n----> 3 dataset = load_dataset('multi_nli')\r\n 4 # nli_dataset = nlp.load_dataset('multi_nli', split='validation_matched[:10%]')\r\n 5 # print(nli_dataset['train'][0])\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 514 \r\n 515 # Download and prepare data\r\n--> 516 builder_instance.download_and_prepare(\r\n 517 download_config=download_config,\r\n 518 download_mode=download_mode,\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 417 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n--> 419 self._download_and_prepare(\r\n 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 455 split_dict = SplitDict(dataset_name=self.name)\r\n 456 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 457 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 458 # Checksums verification\r\n 459 if verify_infos:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\multi_nli\\60774175381b9f3f1e6ae1028229e3cdb270d50379f45b9f2c01008f50f09e6b\\multi_nli.py in _split_generators(self, dl_manager)\r\n 99 def _split_generators(self, dl_manager):\r\n 100 \r\n--> 101 downloaded_dir = dl_manager.download_and_extract(\r\n 102 \"http:\/\/storage.googleapis.com\/tfds-data\/downloads\/multi_nli\/multinli_1.0.zip\"\r\n 103 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in download_and_extract(self, url_or_urls)\r\n 214 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 215 \"\"\"\r\n--> 216 return self.extract(self.download(url_or_urls))\r\n 217 \r\n 218 def get_recorded_sizes_checksums(self):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in extract(self, path_or_paths)\r\n 194 path_or_paths.\r\n 195 \"\"\"\r\n--> 196 return map_nested(\r\n 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 168 return tuple(mapped)\r\n 169 # Singleton\r\n--> 170 return function(data_struct)\r\n 171 \r\n 172 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in <lambda>(path)\r\n 195 \"\"\"\r\n 196 return map_nested(\r\n--> 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n 199 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 231 if is_zipfile(output_path):\r\n 232 with ZipFile(output_path, \"r\") as zip_file:\r\n--> 233 zip_file.extractall(output_path_extracted)\r\n 234 zip_file.close()\r\n 235 elif tarfile.is_tarfile(output_path):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in extractall(self, path, members, pwd)\r\n 1644 \r\n 1645 for zipinfo in members:\r\n-> 1646 self._extract_member(zipinfo, path, pwd)\r\n 1647 \r\n 1648 @classmethod\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in _extract_member(self, member, targetpath, pwd)\r\n 1698 \r\n 1699 with self.open(member, pwd=pwd) as source, \\\r\n-> 1700 open(targetpath, \"wb\") as target:\r\n 1701 shutil.copyfileobj(source, target)\r\n 1702 \r\n\r\nOSError: [Errno 22] Invalid argument: 'C:\\\\Users\\\\Python\\\\.cache\\\\huggingface\\\\datasets\\\\3e12413b8ec69f22dfcfd54a79d1ba9e7aac2e18e334bbb6b81cca64fd16bffc\\\\multinli_1.0\\\\Icon\\r'\r\n```\r\n","comment_length":20,"text":"Can't download MultiNLI\nWhen I try to download MultiNLI with \r\n```python\r\ndataset = load_dataset('multi_nli')\r\n```\r\n\r\nI get this long error:\r\n```python\r\n---------------------------------------------------------------------------\r\nOSError Traceback (most recent call last)\r\n<ipython-input-13-3b11f6be4cb9> in <module>\r\n 1 # Load a dataset and print the first examples in the training set\r\n 2 # nli_dataset = nlp.load_dataset('multi_nli')\r\n----> 3 dataset = load_dataset('multi_nli')\r\n 4 # nli_dataset = nlp.load_dataset('multi_nli', split='validation_matched[:10%]')\r\n 5 # print(nli_dataset['train'][0])\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 514 \r\n 515 # Download and prepare data\r\n--> 516 builder_instance.download_and_prepare(\r\n 517 download_config=download_config,\r\n 518 download_mode=download_mode,\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 417 with utils.temporary_assignment(self, \"_cache_dir\", tmp_data_dir):\r\n 418 verify_infos = not save_infos and not ignore_verifications\r\n--> 419 self._download_and_prepare(\r\n 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 421 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 455 split_dict = SplitDict(dataset_name=self.name)\r\n 456 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 457 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 458 # Checksums verification\r\n 459 if verify_infos:\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\datasets\\multi_nli\\60774175381b9f3f1e6ae1028229e3cdb270d50379f45b9f2c01008f50f09e6b\\multi_nli.py in _split_generators(self, dl_manager)\r\n 99 def _split_generators(self, dl_manager):\r\n 100 \r\n--> 101 downloaded_dir = dl_manager.download_and_extract(\r\n 102 \"http:\/\/storage.googleapis.com\/tfds-data\/downloads\/multi_nli\/multinli_1.0.zip\"\r\n 103 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in download_and_extract(self, url_or_urls)\r\n 214 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 215 \"\"\"\r\n--> 216 return self.extract(self.download(url_or_urls))\r\n 217 \r\n 218 def get_recorded_sizes_checksums(self):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in extract(self, path_or_paths)\r\n 194 path_or_paths.\r\n 195 \"\"\"\r\n--> 196 return map_nested(\r\n 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\py_utils.py in map_nested(function, data_struct, dict_only, map_tuple)\r\n 168 return tuple(mapped)\r\n 169 # Singleton\r\n--> 170 return function(data_struct)\r\n 171 \r\n 172 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\download_manager.py in <lambda>(path)\r\n 195 \"\"\"\r\n 196 return map_nested(\r\n--> 197 lambda path: cached_path(path, extract_compressed_file=True, force_extract=False), path_or_paths,\r\n 198 )\r\n 199 \r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\site-packages\\nlp\\utils\\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs)\r\n 231 if is_zipfile(output_path):\r\n 232 with ZipFile(output_path, \"r\") as zip_file:\r\n--> 233 zip_file.extractall(output_path_extracted)\r\n 234 zip_file.close()\r\n 235 elif tarfile.is_tarfile(output_path):\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in extractall(self, path, members, pwd)\r\n 1644 \r\n 1645 for zipinfo in members:\r\n-> 1646 self._extract_member(zipinfo, path, pwd)\r\n 1647 \r\n 1648 @classmethod\r\n\r\n~\\Miniconda3\\envs\\nlp\\lib\\zipfile.py in _extract_member(self, member, targetpath, pwd)\r\n 1698 \r\n 1699 with self.open(member, pwd=pwd) as source, \\\r\n-> 1700 open(targetpath, \"wb\") as target:\r\n 1701 shutil.copyfileobj(source, target)\r\n 1702 \r\n\r\nOSError: [Errno 22] Invalid argument: 'C:\\\\Users\\\\Python\\\\.cache\\\\huggingface\\\\datasets\\\\3e12413b8ec69f22dfcfd54a79d1ba9e7aac2e18e334bbb6b81cca64fd16bffc\\\\multinli_1.0\\\\Icon\\r'\r\n```\r\n\nGlad it helps !\nThough I am not one of hf team, but maybe you should close this issue first.","embeddings":[-0.0840254948,-0.0221919809,-0.0726869553,0.0423389748,0.1981838346,0.0664926246,0.457182169,0.0801374689,0.1033792794,0.0883751884,-0.1641247272,0.2453239113,-0.1955309659,-0.0816973373,-0.0615135431,-0.2790237069,-0.0329709835,-0.0675984398,0.028705474,0.1370823085,-0.2184431851,0.1371101737,-0.2122425735,0.2103482187,-0.1834195554,-0.1558727026,-0.2034406811,0.2328009605,-0.338126868,-0.360994488,0.1864654124,-0.1097058058,0.2480324358,0.0156800989,-0.0001071728,0.1085644364,0.1779890656,-0.0657087117,-0.406357944,-0.1557623595,-0.1127581894,-0.3171509802,0.1604038626,-0.3909189105,0.3036190569,0.3498140872,0.0443298407,0.0456054173,0.1174699664,0.4848155379,0.2677161992,0.4489547312,0.3633489907,-0.0720728859,-0.0470621996,-0.2057535052,0.0127581256,0.3778342307,-0.1025850102,-0.176113531,0.0909408852,0.0736363828,-0.1486242712,0.3225732148,-0.1937520951,-0.0052935621,0.3084739447,-0.2554347217,-0.0774453729,0.2574424744,0.0329429805,0.0830193385,-0.3229765594,0.0241578445,0.2695622742,-0.3627426624,0.0999657214,0.4350741804,-0.3351387978,-0.0922091752,-0.246600464,-0.1657491475,-0.1589851528,0.3905088007,-0.3112727106,0.5100572109,0.1047358438,0.1057526469,0.43421489,-0.2032002658,-0.2787941992,-0.053084895,0.0459684543,0.1960389763,-0.3286784589,0.0189728905,-0.0301640481,-0.2324981689,0.158440128,0.1301026046,-0.0234435536,-0.2342739105,-0.1908606142,0.2034369111,0.0483432971,0.0642103851,-0.2885984182,0.1249965057,0.10281533,0.2291225791,0.2289705127,0.0392173566,-0.1188764647,-0.0918307155,-0.116723232,0.0861231312,0.2541342378,0.0922482461,-0.4595086873,-0.1994170845,-0.5215702653,-0.019995559,0.2925751209,0.1465471834,-0.1126548126,0.3401513398,0.1732837707,0.3975913823,-0.2035049647,-0.3343687654,-0.1372718513,0.1671380103,-0.1231139451,-0.0419097692,0.435234338,-0.0595215447,0.3110079169,-0.0159450583,-0.0854558572,-0.0951857194,0.1114431769,-0.0950859785,-0.3110278845,0.2561810315,0.170438081,0.0722104535,0.1697930694,-0.1313999444,-0.0520505607,0.0058724899,-0.3391163945,-0.4312471151,-0.3035261929,0.2793096006,0.0968348831,-0.105003126,0.1401997209,-0.4007652998,0.2124014944,-0.0330316164,-0.0638782978,-0.2944987416,-0.0988622084,-0.2295627892,0.259360671,0.2700573802,0.1803033054,-0.0016141343,-0.349278003,-0.1656953394,0.2533250153,0.5466470718,-0.2035302222,0.1508474648,-0.1804694384,0.0348155797,0.5982265472,-0.4164569676,-0.0977124944,0.2224085629,-0.3346424699,0.0443879105,0.0076296059,0.179850027,0.0642710999,0.1057124212,0.3237139881,0.5148923397,-0.1797897667,0.079438813,-0.1894708723,-0.2525600791,0.1889620721,0.37766397,0.0521411523,0.0710817426,-0.1293640584,0.4729548395,0.4586094022,0.0054114168,0.0687954426,0.0822698623,0.145469591,-0.090170987,-0.3268409967,-0.0332494974,-0.3870306611,0.127906695,0.0395684429,0.103273347,-0.026721485,0.0882387906,-0.1036591455,0.0494505912,-0.1920175254,-0.1252931952,0.1495797187,0.2034477443,0.0919264555,0.0242259167,0.0234373137,0.516964972,-0.0308067165,0.1307230592,-0.4853770137,0.2479034066,-0.3664743602,-0.0430885442,0.0266262665,0.0217892621,-0.0637478679,0.0476308279,-0.0967497304,0.3046732545,-0.0396848843,-0.4234547317,-0.104751274,-0.1042764485,0.1154831797,-0.1144754738,0.1306092739,0.3364742696,0.0666871145,-0.0499959737,-0.3063918948,0.3337834477,0.0307656657,0.0900291875,-0.0031021968,0.3673850894,0.3018737435,-0.020218024,-0.0372559763,-0.1680344045,0.3337793052,0.0372740142,0.0485329591,-0.2252466381,-0.3169656098,-0.1664787829,0.3011681437,0.0414901115,0.0305451453,0.0421516337,0.1887793392,0.0492153913,-0.1162228882,0.0048489976,0.4669987857,0.0854075626,0.109497644,-0.0226920247,-0.1893053204,-0.2992962599,0.064426057,-0.0190867856,0.2057503611,-0.0591562502,0.130966872,-0.1099563316,-0.0401301496,-0.4957756102,0.2854011953,0.4715655744,-0.2018342465,0.0635068342,-0.3473848403,-0.5306869745,-0.2337204665,-0.2072554231,-0.3806086183,-0.3395351171,-0.1652444899,0.4490947723,0.0872477144,0.0202280376,-0.0163672008,-0.2310703546,0.2713745236,-0.0395486802,0.203712821,-0.071269013,-0.2772781551,0.1511174887,0.3364775479,-0.0214621387,0.2527517974,-0.1002357751,-0.173353374,-0.268103987,0.0158549752,-0.2235613465,0.0938835666,-0.264356792,0.2511912286,0.3671227098,0.046039965,-0.2325368524,0.2155634612,-0.0189510845,-0.1420431137,0.3279959261,0.0061510629,0.0973288193,-0.1499354541,-0.3068360388,-0.3677200377,-0.4596532583,0.2124171704,-0.0558876693,0.2125092894,0.374065578,0.0455109701,0.1976522058,0.0339022465,-0.1330120265,0.0363395475,0.0478519276,0.1572207958,-0.1895360351,-0.3011138439,-0.078162238,0.2041410804,0.2132148594,0.0037161643,-0.3780254722,-0.0257090516,0.0308423396,0.2746254802,-0.1909338534,-0.0438152179,0.2122442573,-0.3453694284,-0.0411612727,0.2166718096,0.0430256687,0.08033517,0.2186608016,0.4515864849,-0.1415331513,0.5075715184,0.0484841429,0.3951174617,0.2567915022,-0.0194629226,0.3725375533,-0.0210259911,-0.0696513951,0.0216038134,-0.4024845064,-0.0067745629,-0.1218655184,0.1562472284,-0.0280966498,-0.108148396,-0.3482307792,-0.17430453,0.0958466232,0.1026307717,-0.2052616328,0.0999547988,-0.243960917,0.026900338,0.0105684996,-0.0934093297,-0.0357636884,-0.1712870151,-0.3353484571,0.3295469582,-0.2056181431,-0.0075147203,-0.0646636635,-0.1637697369,-0.2538139224,0.0381051078,0.1786724627,0.5281776786,0.0313175805,0.1133058146,0.1517692059,-0.0579215512,0.366265893,0.0431779437,0.1123287231,0.3103982806,0.1655441672,-0.5614237785,-0.3184332252,-0.3913879693,0.2710352838,0.7224541306,0.2256345153,-0.1329276562,-0.0833349377,0.3776691854,0.1377887726,-0.0911204144,-0.1700523645,-0.3710435629,-0.3217700422,-0.4575541019,0.054056026,-0.0656715706,0.3394945264,0.1082355306,0.2655141056,0.122991398,0.001990254,-0.1314436942,-0.0960291401,-0.1669194996,0.2519840598,0.1203190908,-0.2048624605,0.1100886837,-0.0121516995,0.5492684841,-0.0537355542,-0.219985351,0.2330361307,-0.0417358167,0.0177323762,0.0470406264,0.1685984582,0.117260173,0.0961255133,0.1955198497,0.3162567317,0.1779068708,0.2874184251,-0.266925633,-0.2842011154,-0.2203262299,0.4090629518,0.0650444776,0.1939111203,0.2437657565,-0.5065380931,0.0695204362,-0.038530454,0.1428883374,0.5448486209,-0.1601898819,0.2056912631,0.369762361,-0.1422553509,0.4998179078,-0.3342645466,0.1792706996,-0.19360286,0.2306205481,0.084226191,0.0270975754,-0.0335593112,0.1626606733,-0.2206200808,0.1645316631,0.129934594,-0.0553147718,0.1815406978,0.5320366621,0.0094062435,-0.3416405618,-0.2274582535,0.2065631896,-0.1639429927,0.3153637052,-0.0948611796,-0.0771327764,0.109076336,-0.1655254811,-0.3335944116,0.2814983726,-0.0834192485,0.5441380739,-0.0700275153,-0.0536753051,-0.0654046685,0.2200279981,0.0794126615,-0.1387498081,-0.0544516519,-0.0917550325,-0.0884905159,-0.274774909,-0.1351374239,0.0227739383,0.2547706664,-0.31023857,-0.3599821627,0.2084232271,0.043487519,-0.2768305242,0.1300090998,0.2966113985,-0.1079709679,0.1171940863,0.0799803808,-0.0604163483,0.1019484773,-0.091481328,0.1378287226,0.1844536662,-0.2723405361,0.0384070612,-0.0725497305,-0.126545012,0.0302758999,0.5448255539,0.0795113966,0.0146676674,0.4382507503,0.2637680769,0.0487562343,-0.3331158757,0.1242254376,-0.2090209126,-0.2165759206,-0.0329977721,0.0389123447,0.2234701365,0.0592571534,0.1032000259,0.1449246407,0.1771889478,0.1525190771,-0.6146883965,-0.4120766521,-0.0507671908,-0.1135169566,-0.1017977074,-0.0972705856,0.2015437484,0.2837581635,0.0364494883,-0.2931175232,0.029571142,-0.3940633833,0.0747049004,0.443446219,-0.2955845296,-0.0098030576,0.2064347267,0.2212134749,-0.1965438724,-0.1449277401,-0.2474171668,-0.026847383,0.0683801845,0.0169644486,-0.1980533749,0.1576873064,-0.3512070179,-0.2484963834,0.1924411058,-0.0722915605,0.038387455,0.0708764419,0.122181423,-0.1602078825,0.1678972542,-0.0834501088,0.0146519588,-0.2705078423,0.0771098062,0.1117303222,0.3172302544,0.2172395736,-0.0045611658,-0.6347709894,-0.1739444435,-0.2452757657,-0.0842615291,0.0953229442,-0.3559564352,-0.0140274074,-0.2292230129,0.061665494,0.2297003418,-0.1112764478,0.0435652323,0.3644263148,0.2506194413,-0.1571569294,-0.0137904389,0.2758883238,0.0277180467,0.0098809311,0.2398214787,-0.104826726,-0.0761379525,0.1575665772,0.0776044205,0.1016805097,0.0976478979,-0.0092486562,0.3073760867,-0.1990081966,-0.0538323224,0.036504183,-0.004383883,0.2784198225,0.313498646,-0.2719894648,0.2696474195,-0.0701691806,0.0511826873,0.1469730884,-0.2084991187,-0.3138246834,-0.0141030736,-0.2458355427,-0.0295137689,-0.1594554037,0.3483742476,-0.3062147498,-0.2602591813,-0.3952552676,0.5272894502,-0.1780480593,0.0531285182,-0.3437132537,-0.0936110541,-0.2732801437,0.1869675666,0.0505930707,0.0776092857,0.4932879806,0.3244490027,-0.1509422213,-0.5531567335,-0.1860025376,0.2678630054,0.0156636816,-0.3500516415,0.2893440127,0.4177259803,-0.2784986794,-0.0091645466,0.0957595855,0.5214709044,0.2982955277,0.1787415594,0.1002323553,-0.0472967327,-0.0325361378,-0.0254061408,-0.0573556684,0.4776994884,0.2105083466,0.3830700815,0.0977295265,-0.2219573557,0.1700776219,0.0921241343,-0.105059281,-0.1036489606,0.4112376571,-0.0219001193,0.0097767729,-0.2397839427,0.0670564026,-0.5321535468,0.0814334303,0.0267071966,0.1320957392,0.1471248418,0.0453968905,0.0507086217,0.0097324513,0.5457183719,0.3034781218,0.2115044296,-0.2945046127,-0.2823817432,-0.6838545203,0.0972865373,-0.5811139941,-0.1010993719,-0.0463687479,0.2308626026,-0.2003297508,0.2188737988,0.0341939181,0.0345413908,0.1818779707,0.2735221684,-0.3600364923,-0.0847819075,0.2178517282,-0.0189787764,0.0338429622,-0.3433586657,0.0698510334,-0.2905687988,0.0890476033,-0.2007319629,-0.0783626437,0.1026028916,0.2820727527,0.0038187257,0.1270779669,0.4972711504,0.1928167641,-0.3524492383,-0.2249494642,0.2027591467,-0.1610056609,0.0131931314,0.2036267668,0.4022216499,-0.1029094979,0.1434143186,-0.3461368084,0.5090784431,-0.1239434853,-0.1700542718,-0.0795723647,0.0544737875,0.056328211,0.1147856936,0.2220103294,0.0350599065,-0.0253666025,0.0394006297,-0.3940591216,-0.2115023285,0.3609002233,-0.0975072682,-0.2655795813,-0.1804896444,0.2269531041,-0.3209328353,0.2880321443,-0.5555567145,0.1251621544,0.296951443,-0.0823288411,-0.5224146843,0.1200467423,0.1242874041,-0.1698413789,-0.0072111064,0.0002059927,0.0807665884,-0.4523905218,0.2310978621,-0.1104964763]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/234","title":"Huggingface NLP, Uploading custom dataset","comments":"What do you mean 'custom' ? You may want to elaborate on it when ask a question.\r\n\r\nAnyway, there are two things you may interested\r\n`nlp.Dataset.from_file` and `load_dataset(..., cache_dir=)`","body":"Hello,\r\n\r\nDoes anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp.\r\n\r\nThank you!","comment_length":29,"text":"Huggingface NLP, Uploading custom dataset\nHello,\r\n\r\nDoes anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp.\r\n\r\nThank you!\nWhat do you mean 'custom' ? You may want to elaborate on it when ask a question.\r\n\r\nAnyway, there are two things you may interested\r\n`nlp.Dataset.from_file` and `load_dataset(..., cache_dir=)`","embeddings":[0.2511941493,-0.1868027151,0.0526289828,0.0816768184,-0.0106313126,0.0912558138,0.3054780662,-0.0391972624,0.1293654442,-0.148810491,-0.2689487338,0.1939389706,-0.0605024397,0.220899567,0.5498043895,-0.006288135,-0.0692309439,0.2022132725,-0.0570144914,-0.1058279723,-0.1856539249,0.2207611203,-0.1370268613,0.1703862697,-0.1214475632,-0.0225457828,-0.0661958754,0.5867702365,-0.1426349282,-0.4699003994,0.3383302391,0.1335825622,0.2301869094,0.3451362848,-0.0001174077,-0.242519781,0.0866902247,-0.2588348687,-0.2761338353,-0.2854745984,0.1859979481,-0.5186660886,0.2983292341,-0.3381852508,-0.0134237055,0.0814301968,0.2943903804,-0.1260166019,0.5597587824,0.0612912029,0.101152733,0.202267617,-0.2472454458,0.079847008,-0.1332534701,0.1941750646,0.0277568046,0.1069401503,0.3506956398,-0.2971981168,-0.0304007865,-0.0905299783,0.1101559401,-0.0872577131,0.5247397423,0.1906174868,-0.2882422805,-0.1282220036,-0.1692812294,0.2119090408,0.406278044,-0.3414674997,-0.1741296947,-0.3976282179,0.1134680137,-0.1795766801,0.2533883452,0.1868111789,-0.2927107215,0.0189503897,-0.2960908115,-0.1277569532,-0.2604834139,0.1948349029,0.3770096898,0.0312416796,-0.1505837291,0.1121048406,0.2019895315,-0.2020061165,-0.2376164794,0.0816826448,0.1484282911,0.6347585917,-0.3498253524,-0.2218444198,-0.180939585,0.1708837748,0.2634724975,0.0313833244,0.3042993844,-0.0190500077,-0.1780896187,0.0734458193,0.3316245973,0.0621524788,0.4791782498,-0.1950465441,-0.0832278728,-0.0602650829,-0.0906262696,-0.0278990362,-0.3807085454,0.0610165149,-0.1343231052,-0.2206491977,0.0271761008,-0.183277458,0.1038198248,-0.3341500163,0.3256415129,-0.1706353873,0.1045821905,0.2052087039,-0.0413789079,-0.2097400427,-0.084017992,0.0338136777,-0.0752505213,-0.2528449595,-0.0328084417,-0.0397825316,-0.1726889312,0.0984549671,0.344299525,-0.2336006463,0.4295724332,-0.0807401687,0.0600132644,0.0233578179,-0.0694897771,0.1254695952,-0.2132776976,-0.2606166303,0.2250982225,-0.0894164816,0.1706477553,-0.3729172051,-0.1835775822,-0.0088387672,-0.2209742218,-0.045483999,-0.3162697554,0.0511501282,-0.213724032,-0.4107028246,-0.2110774219,0.4020387232,-0.1021610275,-0.2433484197,-0.0482000448,0.230752781,-0.1970522255,-0.0970444828,0.0752107352,0.2437968254,-0.4955697358,-0.3121131659,0.1605289876,-0.1799182892,-0.118223995,0.2004851997,-0.2654595077,0.0357513018,-0.1790160537,0.2451011688,0.6685561538,-0.3615259826,-0.2545460165,0.1291518509,-0.3375508487,0.0503921211,-0.0511076376,0.2312552333,0.2727699578,-0.0261819325,0.1195365563,0.7563425899,0.0304378755,-0.0457668565,-0.0146195935,0.012817245,0.1913543493,0.0880464911,-0.3754507601,0.1940221786,0.1528025419,0.1638249904,0.1648674905,-0.0200040583,-0.1410517842,-0.0115677733,-0.0874648392,0.373365283,-0.0119616948,0.0937986299,-0.7424104214,0.250795573,-0.3227907419,0.0265963618,0.3197612762,-0.0572920516,-0.5034213066,-0.1567646861,-0.1250696778,0.1394387335,0.003527012,0.1284808964,0.1161632463,-0.1500425935,-0.3917634785,0.4788067639,-0.2969135642,0.259866178,-0.4144138396,0.0279623326,0.2513425648,-0.0867765546,0.0578437597,0.3452283442,0.1667339057,-0.0923562199,0.085404709,0.3543488681,-0.1966007203,0.3160957992,0.385817796,-0.0798922554,-0.0119260559,0.0590403415,0.2474550307,-0.1868411601,0.1045483053,-0.2373169214,-0.3755524755,0.2033917308,-0.1266898066,0.1138185263,0.058688052,-0.2212678194,0.0994337723,-0.050987225,-0.2495695204,0.0228641126,0.1393253654,0.1846900582,0.5775030851,0.0293774754,-0.1229536533,-0.0937971547,0.5493193865,-0.0694452748,0.0978062898,0.2314471006,-0.0945181176,-0.239956513,0.1245625094,-0.1773858964,0.2780827284,0.2319319248,-0.0777869299,0.0474088043,-0.1571730524,0.0460164286,-0.0632321313,-0.0598318987,-0.0755402446,-0.2413527519,0.2460099757,-0.1520587355,-0.4141062796,0.1202446893,-0.0843704417,-0.0887901708,-0.1791388243,0.0327836908,0.0258512758,-0.3187442124,-0.2578939199,-0.260623008,-0.5657714605,0.1098884866,-0.3193959892,-0.1699989587,0.5434654355,-0.042633187,0.0894463807,0.679009974,-0.1581475735,-0.1685850024,0.1133372039,-0.0597933605,-0.2973839939,-0.03805843,0.1349502206,0.2562676072,0.4579537213,-0.0091979131,0.1660220772,0.1255880147,-0.2280965,0.0864352435,-0.1482443362,0.2335582227,0.1508990079,0.1980547756,-0.1274206787,-0.0572060756,0.0523496196,-0.1385670304,-0.179023996,-0.2404356152,0.0019247442,0.2437565923,-0.0372643843,-0.1320925504,-0.4951516092,-0.2063333541,0.5470715165,0.363799274,0.100857161,0.2450055033,-0.337979883,0.436318785,-0.2013397217,-0.0298940167,-0.2431279719,-0.5314003825,0.1110533327,-0.2212736011,-0.1870451123,0.1094565764,0.04110967,0.1178323105,-0.2680106759,-0.2026875168,-0.2860255241,0.2530061603,-0.0214734487,-0.0531165637,0.06862472,0.1988119781,-0.3225416541,0.1168540716,-0.0986807272,-0.3786769211,0.1931208372,0.3867774308,0.244505465,0.2283461541,0.1312942803,-0.0158447921,0.6685090065,0.0275276974,-0.1056762487,0.3860656321,-0.0057337312,0.221611321,0.0189451911,-0.4455029666,0.0968725905,0.2294549793,0.2309973687,0.401912868,0.2349952459,0.0601207688,-0.2935670018,0.0045253271,-0.1796638072,0.1343782097,-0.1014319435,0.2586897314,-0.0557223745,0.1442110687,-0.1377077699,-0.5438347459,-0.2991214097,-0.2503632903,0.1944397986,0.1576123387,0.2664287388,-0.7415153384,0.0820779279,-0.4755269587,0.2280856222,0.0425124541,0.1476490945,-0.0042540417,-0.0085567115,0.2061494291,0.1683983058,0.5750076771,0.2543748915,-0.1479306668,-0.0584111996,0.0946148708,-0.4649529755,0.3017906249,0.1423837841,0.3281335235,0.3223803043,0.5257136226,-0.2388924062,-0.2450682819,0.1613656878,0.0109566115,-0.357976228,-0.3428820968,-0.3559771478,-0.1083637699,-0.0142478859,-0.1077906564,-0.1859328449,0.1585281044,0.1760942638,0.0644018352,0.1274660528,0.0265164729,0.0725328103,0.203369379,-0.0336538814,0.463539958,-0.1895828843,0.3946459293,0.2590387166,-0.537486434,0.6323096156,-0.4285534322,-0.337710619,-0.1661485583,-0.0027464293,0.5326382518,0.2596348524,0.4100420177,0.1135770306,-0.1531108469,-0.1127478853,-0.2799547017,0.2060446441,0.2188224494,-0.0887853727,-0.1724361628,-0.4115307331,0.6642174125,-0.1253691912,-0.0142120738,0.1717309207,0.1919846982,-0.2211806774,0.0917063653,-0.1531617194,0.9541105032,-0.3873251081,0.3526079357,0.1197058037,-0.0099990629,0.7712205052,-0.1817929447,-0.1625916958,-0.3657719791,0.0291690379,-0.0812450498,-0.0940425768,0.2180778086,0.6166026592,-0.1423174739,0.3255699575,0.1767976731,-0.0927708969,0.1079549044,0.4875271618,0.1843470633,-0.5538482666,-0.7102081776,0.0603516512,-0.3340833485,0.2049708813,-0.1655218452,-0.0905331597,0.034811683,-0.2015576065,0.0662600622,0.0148900934,-0.2054176033,-0.2979595363,0.3345995843,-0.3240866661,0.041493874,0.5139353871,0.1572243422,-0.0310506187,-0.0088659301,-0.0063475007,-0.091828756,-0.0301012117,-0.2045817822,0.3019344509,0.2518041134,-0.1172953993,-0.0317707211,0.0222395193,0.1166367158,-0.2117556632,-0.2835069001,0.237990424,0.199623093,-0.3022352159,-0.5194169879,-0.0023244908,0.2110832334,0.0864513591,0.0239311159,0.2017680258,-0.0706328973,-0.113901183,0.4379002452,-0.0418484248,0.128881678,0.4388933778,-0.3553407788,-0.4714884162,0.4188661575,0.2282076627,0.1188083366,0.0019888941,0.1696198583,0.0791053399,-0.1307053119,-0.1676656604,0.4686748683,-0.0138896564,-0.0796686932,0.1723967493,0.1618813872,-0.1481516212,0.1280463189,-0.1939021945,-0.3550582826,0.1070313752,-0.0264735706,-0.034784507,0.4684087336,-0.0369186513,0.2316412628,0.2830173969,-0.1894064546,0.1614175439,-0.2517104745,0.1376013458,0.5653265715,-0.1713072658,0.0128414268,0.1365103126,0.0181449428,0.0578298084,-0.1279695928,-0.1272273362,-0.2523893416,0.1987717301,0.1550896466,-0.104486838,-0.1222681329,-0.183529526,-0.0782115012,-0.1162468567,0.0186142791,0.1467441916,0.1244984046,0.3703163862,0.0719775632,0.2226103842,-0.00388914,0.0290055797,0.1488191038,0.2056493908,0.1384309828,-0.0402175821,-0.1797061265,0.0334232226,-0.2460646778,0.2219079733,0.5597800016,0.0894270316,0.0814632326,-0.0800017193,0.0924023613,-0.0762057751,0.0349499285,0.1303813308,-0.179288134,0.1431700885,0.2328599542,0.0574849062,-0.140128389,-0.126433745,0.6219967604,-0.1061830893,-0.0647737458,0.2504395843,-0.1310276538,0.2922578156,-0.1190719455,-0.2139939517,0.4340814352,0.1513227671,-0.1202651486,-0.1343465,-0.2251219749,-0.2415361702,0.275110662,0.3698018491,0.1239179298,0.0680181235,-0.1743521839,0.1309022009,-0.2866813242,0.139405027,0.3228420913,-0.44480744,-0.1280961484,0.3435077667,0.2870287895,0.5161547661,-0.252410382,0.9262969494,-0.2093737274,-0.1474722922,-0.0828234106,0.1101433039,0.1206443086,-0.1401916891,-0.1865784377,-0.0202412307,-0.1250214875,-0.0759110898,-0.2039105296,-0.4526928067,-0.2174873352,-0.1241643205,0.1628555208,-0.0601293817,-0.2560814023,0.2250674516,-0.0895432532,-0.0897463188,0.2354154587,-0.0840310231,0.2372182757,0.4916706681,0.122001037,-0.1269647032,0.2095475346,0.0639062002,0.1398533136,-0.0175693613,0.0440180525,0.0074536782,0.0787410587,0.0132794697,0.0604069754,0.1171359643,0.0465086028,-0.0496437959,-0.0185241848,0.1779904664,-0.018203767,-0.4296704531,0.1741535515,-0.1935662329,-0.1336492896,-0.4657391608,0.2527318001,-0.5220646858,-0.1287761331,0.0620663017,0.2569209933,0.0989430323,-0.2657575905,0.0298513304,-0.2185126245,0.5074788332,0.2122344375,0.0320805013,0.0951520577,-0.2340484709,-0.5880166888,0.1993702054,0.1645781547,-0.2712222338,0.0189883318,0.0646125972,-0.0401025712,0.1994287372,-0.0902371481,-0.0519047715,-0.1530139595,-0.0805527791,-0.2661845684,-0.2601546347,-0.1091302559,0.2602443397,0.0556840524,0.0285981447,-0.0816720128,0.2017632425,-0.1871013194,-0.133767277,0.0589811131,-0.0117320055,-0.0124997469,0.1130949259,-0.2479581982,0.3801167905,0.1310814619,0.0162369274,0.1262481511,0.0024312104,-0.2638449073,0.0547385775,-0.0713827983,0.3061409295,-0.2006978542,0.0520070978,-0.1553658247,0.1286509186,-0.0889888555,-0.1510299742,-0.1771154106,-0.1124609411,-0.1070469245,-0.0372111537,0.1936522573,0.5496388078,-0.0695770383,0.196214661,-0.0895118788,0.1812868267,0.4863636494,-0.4867604375,-0.0912506059,0.1694693118,0.2191723436,0.1114573032,-0.3302909136,-0.4294406474,0.1502025574,0.3653630912,0.0011041999,0.0691451505,0.4149712026,-0.4803500772,-0.0295142364,-0.0905225351,-0.0166546311,0.2801739275,0.1656362116,0.0842817798,-0.0744338632]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/234","title":"Huggingface NLP, Uploading custom dataset","comments":"To load a dataset you need to have a script that defines the format of the examples, the splits and the way to generate examples. As your dataset has the same format of squad, you can just copy the squad script (see the [datasets](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets) forlder) and just replace the url to load the data to your local or remote path.\r\n\r\nThen what you can do is `load_dataset(<path\/to\/your\/script>)`","body":"Hello,\r\n\r\nDoes anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp.\r\n\r\nThank you!","comment_length":67,"text":"Huggingface NLP, Uploading custom dataset\nHello,\r\n\r\nDoes anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp.\r\n\r\nThank you!\nTo load a dataset you need to have a script that defines the format of the examples, the splits and the way to generate examples. As your dataset has the same format of squad, you can just copy the squad script (see the [datasets](https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets) forlder) and just replace the url to load the data to your local or remote path.\r\n\r\nThen what you can do is `load_dataset(<path\/to\/your\/script>)`","embeddings":[0.1632618755,-0.2137292325,0.0653117523,0.1629524976,-0.0624897406,0.0928252488,0.2512159646,-0.0430666991,0.1474726647,-0.1472516805,-0.2029948384,0.1871040314,-0.1082477719,0.3745346367,0.6062445045,-0.0716352537,-0.1152896136,0.1753334403,-0.0515353531,-0.0377068184,-0.1981060505,0.2560156584,-0.1257559955,0.2088316083,-0.157189548,0.0645788833,-0.1292643845,0.6166014671,-0.0432282425,-0.4653248489,0.4585714042,0.1644788831,0.3075277805,0.3923658431,-0.0001183142,-0.2341071218,0.1320631802,-0.2658723593,-0.2585261762,-0.435149461,0.115730226,-0.4135122597,0.357668817,-0.2580885291,-0.075783655,0.1012234688,0.256700784,-0.071713537,0.5230125785,0.0841127262,0.0794004872,0.1752882004,-0.1973444819,0.0069302241,-0.0799260363,0.2747301161,0.118542023,0.2547322214,0.4028348923,-0.3194004297,-0.0109980796,-0.1588699222,0.078856498,-0.0765652731,0.5086486936,0.1393877268,-0.3838796914,-0.1076978445,-0.1860266775,0.2605612576,0.441868633,-0.4110240042,-0.1638529003,-0.3494079411,0.1157299951,-0.0819408968,0.2004316747,0.2470340133,-0.2333192229,-0.0007751902,-0.3248098493,-0.0814079493,-0.2789474428,0.2652963996,0.4247459471,-0.0282355379,-0.1757750809,0.1322721392,0.1465883106,-0.1842674017,-0.229835391,0.0410306044,0.1278071702,0.5799487233,-0.2658235133,-0.226796478,-0.1616683006,0.1954281032,0.3089895546,0.0729882643,0.2669959366,0.0085437084,-0.2398358136,0.0831922814,0.3722091615,0.0566619597,0.4918217361,-0.1684591919,-0.1371180713,0.019395059,-0.1130543426,-0.0279668532,-0.4432641864,0.0435884334,-0.2443509847,-0.209770605,0.0535984859,-0.2177799493,0.0456063561,-0.3099789619,0.2870750725,-0.1471867561,0.1001880392,0.2248855978,-0.0065349527,-0.1765411496,-0.0419527143,0.1313504279,-0.0876829848,-0.2424858361,0.0154616591,-0.0259964149,-0.1502151787,0.1816100031,0.3011452556,-0.1626353562,0.4779955447,-0.1116511077,0.0032149018,-0.0790123716,-0.0785928816,0.2046216875,-0.2747231126,-0.228053987,0.2984566987,-0.1199111268,0.2767028511,-0.2447419614,-0.1950594187,-0.0161996968,-0.1980811805,-0.099808082,-0.2604421973,0.0298508797,-0.2414761037,-0.42902565,-0.3082430363,0.4173445404,-0.1880047917,-0.2443479747,-0.0652576238,0.2210941166,-0.2627894878,-0.1088764146,0.1435061842,0.3203823566,-0.5170246363,-0.3434064984,0.1513788998,-0.0854873806,-0.16925776,0.1477366835,-0.3176531196,0.1524892151,-0.2047575563,0.221737802,0.5584897995,-0.2829348147,-0.1232941374,0.2048125714,-0.3473904133,-0.0599625893,-0.0623305887,0.2136587203,0.2472444624,-0.0182476752,0.0306184273,0.699007988,0.0460638888,-0.0494595096,0.0443298556,-0.055604808,0.1161719635,0.1239847168,-0.4393498302,0.2005514354,0.1385040432,0.1559725255,0.1280867904,-0.0063274037,-0.1663253754,0.0146783758,-0.0424188823,0.3889630735,-0.0186274182,0.0057169311,-0.6861996651,0.2428305745,-0.3102069199,-0.1179864556,0.3537498415,-0.0747481659,-0.5600332022,-0.1199524626,-0.052174475,0.1423639953,-0.0310133155,0.0715241432,0.235191673,-0.1559379399,-0.419976294,0.3658932447,-0.2654148638,0.2219631225,-0.4422061145,0.0715068355,0.2992537022,-0.1190320328,0.1711467803,0.3448594809,0.1859295219,-0.1832340956,0.1324022412,0.4043011665,-0.0898139775,0.3038014174,0.3436336517,-0.1167560294,-0.0457207374,0.0180007908,0.2573494613,-0.0722771212,0.1622330695,-0.2356225401,-0.39416641,0.2469726801,-0.1662818342,0.1955576986,0.0855974704,-0.1404212266,0.0753035918,-0.1076562181,-0.2879719138,0.0308810845,0.1261328757,0.1642033309,0.5432072282,0.0151495188,-0.0742928758,-0.137536034,0.4589149952,-0.1621593237,0.0483373962,0.2757790089,-0.2151982337,-0.1649657488,0.0967513621,-0.1139991134,0.3100784719,0.2174781263,-0.1915289909,0.1209520027,-0.1546864957,0.0167939328,-0.0776061267,0.0096650738,-0.0724230111,-0.2934227288,0.0289833359,-0.1682446152,-0.3165303171,0.0073105465,-0.0837204754,-0.05273946,-0.279763788,0.0846176967,-0.0238074176,-0.3330242038,-0.3326044083,-0.3089468181,-0.409217298,0.1175934896,-0.3757782578,-0.0898815915,0.5219320655,-0.067493096,0.078946881,0.5729561448,-0.1967441142,-0.1401185095,0.149264589,-0.1234442294,-0.3371569812,-0.0581831336,0.2175436765,0.3872212768,0.4830297828,-0.0962698013,0.1698093116,0.1431282461,-0.1550988853,0.0761236474,-0.1170581281,0.1787731051,0.2281749099,0.2039536834,-0.1120222136,-0.0922498927,0.069925271,-0.1441305876,-0.0947047323,-0.2146958709,0.0065226643,0.1487960219,-0.0692987218,-0.2026622593,-0.392604053,-0.2046023011,0.5442581177,0.317915678,0.094566159,0.1486199498,-0.2004937977,0.5334941149,-0.2726586461,-0.0955983847,-0.1534222215,-0.5053127408,0.043326918,-0.1828352511,-0.2392140329,0.1731575429,-0.0023852864,0.0803936869,-0.2721155286,-0.2141652256,-0.3095809519,0.1658876389,-0.1243156865,0.0111166248,0.0636945069,0.1931689084,-0.3519292474,0.1423199177,-0.0800969973,-0.3716738522,0.1727839261,0.3058595657,0.1259484887,0.3053837717,0.2932281792,-0.0838696137,0.5675312281,0.1291265339,-0.1714803129,0.3443665206,-0.0402669162,0.1789775193,0.067992419,-0.3560134768,0.0380213857,0.1795797497,0.2200683653,0.4753509164,0.266054064,0.1811908782,-0.2931441963,-0.0726755038,-0.1638056189,0.0977697745,-0.0492739528,0.1175727248,-0.0360350311,0.0642567798,-0.1507172883,-0.4546360373,-0.2977863848,-0.2510930598,0.2559709251,0.1712976545,0.2918236554,-0.6364765763,0.0713652894,-0.3645475805,0.2603876889,0.0204705242,0.1405488551,0.0067775678,0.0354191884,0.2148374319,0.1790653914,0.4744753838,0.2818923593,-0.1275414526,-0.0829936415,0.0145442737,-0.5307664871,0.2800444365,0.0936902538,0.3389353752,0.3042061925,0.5084884763,-0.2509047389,-0.2283212394,0.1943625063,0.0491243005,-0.304125905,-0.3914254904,-0.2928214967,-0.1052877679,-0.1562438011,-0.1447368264,-0.1588700265,0.0910777301,0.1252750754,0.0737436637,0.102108404,-0.0654225126,0.1005931273,0.1242032796,0.0259874389,0.4303339422,-0.20999749,0.3520077169,0.2423614115,-0.5084127188,0.6294357777,-0.2764791846,-0.3367935419,-0.2020911723,-0.0018835395,0.5229845643,0.0989648327,0.3765937388,0.0771805719,-0.0711238608,-0.1296425909,-0.3097154796,0.2332356274,0.1351251453,-0.0911625624,-0.1205738038,-0.4487238526,0.6946140528,-0.1061770022,0.0159464516,0.0789308026,0.1630154401,-0.1994716227,0.1325691342,-0.1868541539,0.7473629713,-0.3751449585,0.3902586401,0.1302455962,0.0035096956,0.7170394659,-0.0950674489,-0.2441701591,-0.3508577347,0.1099546626,-0.1457488239,-0.1118384078,0.1543856263,0.6320897341,-0.2097486556,0.3886799216,0.1037075147,-0.0029980917,0.1480758339,0.5094792843,0.1545839161,-0.5113891363,-0.6902395487,0.0567219518,-0.2997639477,0.2571177483,-0.1179746464,-0.0964411125,-0.0174417607,-0.2025994062,-0.0157738756,0.0593105778,-0.2245119363,-0.3230780363,0.2293545008,-0.2782453895,0.1624460071,0.5821105838,0.1242396086,-0.02593733,-0.0918034315,-0.0531459562,-0.1775504351,-0.0915537328,-0.19871144,0.2340759039,0.2042722106,-0.0426684879,-0.0167222172,-0.0191920679,0.1943223625,-0.1067332551,-0.2794774473,0.2197616994,0.1290057153,-0.3557678759,-0.6285941005,-0.0216856748,0.1941078156,0.0940655395,-0.0110009871,0.2376994938,-0.0637600198,-0.1185451895,0.4086724222,-0.0435629264,0.104027614,0.3902845383,-0.3402022421,-0.3956837058,0.4708957374,0.2909394801,0.2221646011,0.0028174068,0.2171220481,0.2111028284,-0.2119378299,-0.1450534463,0.5018584132,0.0719345957,0.0172119755,0.1696601957,0.1455610394,-0.1444828361,0.1608453542,-0.2607404888,-0.3615255952,0.1232202202,0.0234348644,0.0229995027,0.4183296263,0.0386551432,0.2411439866,0.2213960737,-0.1426660269,0.1920413077,-0.1899485886,0.0788211077,0.5557528734,-0.0537683889,0.0103268968,0.1289436519,-0.0232265722,0.0555938371,-0.0289824083,-0.0895431042,-0.2642658055,0.2132717818,0.1899088472,-0.1070596501,-0.1478244066,-0.1872739941,-0.0341319107,-0.1136232913,-0.012284399,0.1121048778,0.132039234,0.3177058995,0.062996991,0.2132779211,-0.0523691624,0.0645428821,0.0492332205,0.2295682132,0.1381445676,-0.0491981916,-0.060791377,0.0187435951,-0.3008621335,0.2291802317,0.615952909,0.0421675816,0.0542980246,-0.0486629903,0.1030887142,-0.1313527524,0.0990528166,0.1644878536,-0.1199643612,0.1464222223,0.3028815687,0.0181765482,-0.1055426374,0.0293668751,0.5654889941,-0.0895986259,-0.0829142481,0.2293394804,-0.1696101874,0.3017829955,-0.1773396581,-0.2049773186,0.3812517226,0.2803016603,-0.0552132279,-0.1186575368,-0.1581610292,-0.2488832772,0.1453041434,0.3515394926,0.1439852417,-0.0857212171,-0.2017533332,0.2122204006,-0.3304879069,0.1439915001,0.3951110244,-0.4431227446,-0.195223242,0.32062006,0.333793968,0.4531187117,-0.3040761948,0.9252217412,-0.2291449755,-0.0972212031,-0.1421980411,0.1182337925,0.1555022895,-0.1779163033,-0.2183939517,0.0511371568,-0.0641890839,-0.0802159309,-0.1675035059,-0.4903082252,-0.2019679248,-0.1363516897,0.1241889596,-0.1170485467,-0.2363885343,0.2489046156,-0.1850993633,-0.0798591152,0.2128397226,-0.0338640176,0.2575384378,0.5040982366,0.0844363123,-0.0476292334,0.196544528,0.0119716888,0.1720649749,-0.0335554108,0.0778158531,0.0594721399,0.0623539053,-0.0620012432,0.1188877523,0.1364641786,0.0433426425,-0.0046252506,-0.0681404099,0.153035894,-0.0025302521,-0.4499546885,0.2142452002,-0.2077808678,-0.0822022259,-0.4868535995,0.2683153152,-0.5577459335,-0.1333116889,0.0442229658,0.2668989003,0.0938567072,-0.2231746614,0.0126999663,-0.2235442698,0.5138816237,0.1905690581,0.1077861041,0.0324132182,-0.2708286643,-0.5493530631,0.1231707633,0.0953589827,-0.3297888041,0.02394161,-0.0093842065,-0.0909564942,0.1794672906,-0.2185569257,-0.0668046251,-0.2022578269,-0.0262499265,-0.1942459494,-0.3171683848,-0.1957118511,0.3553727567,0.0570356697,0.0877921805,-0.155713439,0.1583606005,-0.2210959196,-0.1377209127,0.1367911845,-0.0043681879,-0.0802055821,-0.0215765666,-0.2114246488,0.4362390041,0.1445487142,0.1189908162,0.0088032251,-0.0007139264,-0.2326341122,0.0927947089,-0.0973541141,0.2749655545,-0.26326406,0.0623066239,-0.1764576286,0.0602200031,-0.09631522,-0.1251280159,-0.0828753859,-0.0668040514,-0.1780525446,-0.0061445786,0.2377381623,0.5563530326,-0.1058877558,0.2305358201,0.0334423594,0.1536972374,0.4228229821,-0.4647780955,-0.0668636411,0.2159597427,0.2218232304,-0.0098113781,-0.3232257962,-0.418040514,0.2003141791,0.3468027413,0.0422861315,0.1356089711,0.4116926789,-0.3617994487,-0.0551312007,-0.0770757273,0.0878526047,0.2676624656,0.1341014206,0.0872888118,-0.1015202478]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/234","title":"Huggingface NLP, Uploading custom dataset","comments":"Also if you want to upload your script, you should be able to use the `nlp-cli`.\r\n\r\nUnfortunately the upload feature was not shipped in the latest version 0.2.0. so right now you can either clone the repo to use it or wait for the next release. We will add some docs to explain how to upload datasets.\r\n","body":"Hello,\r\n\r\nDoes anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp.\r\n\r\nThank you!","comment_length":57,"text":"Huggingface NLP, Uploading custom dataset\nHello,\r\n\r\nDoes anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp.\r\n\r\nThank you!\nAlso if you want to upload your script, you should be able to use the `nlp-cli`.\r\n\r\nUnfortunately the upload feature was not shipped in the latest version 0.2.0. so right now you can either clone the repo to use it or wait for the next release. We will add some docs to explain how to upload datasets.\r\n","embeddings":[0.1764683574,-0.1404461861,0.0849194527,-0.0182812419,-0.0717138723,0.075803332,0.3437079787,-0.0349298343,0.1954903901,-0.0800016373,-0.2074443698,0.2885677218,-0.0553389043,0.3174791336,0.6837016344,-0.0082312217,-0.1109836325,0.2270630151,-0.1107003391,-0.072724998,-0.1884524673,0.2482985109,-0.1515674591,0.1427882761,-0.0904977247,0.0327357911,-0.1368426234,0.5879563689,-0.0347582921,-0.4627845883,0.3801742792,0.1850762814,0.2705678046,0.2915583551,-0.0001199196,-0.3151123822,0.2196165025,-0.2283890694,-0.1992084533,-0.3886288106,0.2290245891,-0.4995836616,0.3296054304,-0.2545453608,0.0229102876,0.0480307788,0.3462797999,-0.0469349064,0.4474572241,0.1120043546,0.0682483017,0.2357687205,-0.2086908072,0.0165882111,-0.1261121035,0.2023485154,0.0364916101,0.0982222632,0.4745078087,-0.3767592609,-0.0204599723,-0.2292882949,0.1150449514,-0.2255510688,0.5860815048,0.1650473028,-0.2993729413,-0.1718950421,-0.1495643258,0.2630612552,0.4205902517,-0.3851597011,-0.1577212512,-0.3561770022,0.1602452248,-0.0932904631,0.2420265675,0.2168908119,-0.198820129,-0.0408759117,-0.3009008765,-0.083590284,-0.3048931062,0.2369063497,0.4349042773,0.0735771507,-0.0748693347,0.1235278696,0.2585088313,-0.2515004575,-0.1675399542,0.1206649542,0.0902183279,0.6502204537,-0.2848996818,-0.3007370234,-0.1803151667,0.1586339325,0.3465579748,0.0871054977,0.2340097725,-0.0379519314,-0.231588617,0.0632524639,0.4135237634,0.0722877532,0.5410382748,-0.2067291737,-0.1094599664,-0.0746562853,-0.0071321977,-0.0387752801,-0.3436836898,0.0666880682,-0.2101968825,-0.2071940452,-0.003997284,-0.2105837017,0.0884229988,-0.3016293347,0.3965158761,-0.2021564543,0.0337174647,0.2337683141,-0.0262790117,-0.1990767866,-0.0066285911,0.1639598608,-0.0240708608,-0.2833609283,0.0277406126,0.0290357713,-0.1441715509,0.1292189658,0.308316201,-0.119163394,0.4292005002,-0.1462306678,0.115179427,0.034417659,-0.1853409857,0.2004367262,-0.2792135775,-0.211660713,0.2990442514,-0.1502891481,0.1245240346,-0.3165315688,-0.1780049354,0.0566845685,-0.1173621565,-0.0222805347,-0.3355042636,-0.000865708,-0.1576640755,-0.5109561086,-0.2110397071,0.4058326483,-0.2044931352,-0.2549584508,-0.0943488777,0.1454541534,-0.233881861,-0.0956437364,0.0803910345,0.2395888716,-0.5089073181,-0.3144393265,0.1347630173,-0.0557294339,-0.0909917727,0.1236264408,-0.2257529497,0.0113541689,-0.1373366714,0.2648732662,0.6306740046,-0.3635216951,-0.170813024,0.1660255045,-0.284729749,-0.0126986876,-0.0783314705,0.2403729409,0.2574870884,-0.0433332622,-0.0072615067,0.7859998941,0.0204003416,-0.030374784,0.0199573878,-0.0455239676,0.1873829961,0.1366540343,-0.3432533741,0.2071154863,0.1615181714,0.1116489321,0.1957997233,0.0161686949,-0.1604333222,-0.0386950076,-0.0726060942,0.3785209954,0.0165354125,0.0375590101,-0.6924958229,0.2066673636,-0.418810606,-0.014900797,0.3599675596,-0.0527982041,-0.5091999173,-0.1182487607,-0.0001259361,0.1753367186,-0.0350242443,0.0883272365,0.1414326727,-0.1319975555,-0.4686155617,0.4200882912,-0.2656469345,0.2159391493,-0.383990407,0.0986372977,0.2503145933,-0.1872021258,0.1851067692,0.3410764337,0.1777301729,-0.1337010562,0.1010109782,0.3626877666,-0.1958237588,0.2949708998,0.4155726135,-0.0886107013,-0.052945476,0.0855376795,0.2579508126,-0.1100844741,0.0429702364,-0.2711121738,-0.392005235,0.2096698433,-0.1177296042,0.1833424121,0.0382431038,-0.1359160841,0.0667028725,-0.1294801086,-0.3018996716,0.0593983829,0.0913830251,0.1398960352,0.5469378829,-0.0844841823,-0.1183323562,-0.0494960733,0.5501151085,-0.1140015051,0.0779598504,0.2512154877,-0.1055456176,-0.1468648314,0.1260046959,-0.190888539,0.2596572936,0.2283757627,-0.2271471769,0.0737366378,-0.2061674297,0.0459526591,-0.0846225694,-0.1173196062,-0.0527106449,-0.2695628703,0.1740244776,-0.1882400364,-0.4044696391,0.0869704857,-0.1470432281,-0.0693099573,-0.2084822357,0.0153290853,0.0022148776,-0.3238930404,-0.2259030342,-0.2658852339,-0.4910959601,0.0526196323,-0.3498063982,-0.1062775552,0.5295960307,-0.0520726815,0.1573909074,0.683640182,-0.149899736,-0.066070497,0.0718422458,-0.0269821845,-0.2955059707,-0.059819933,0.1493528783,0.2859882712,0.4051009715,-0.0493484624,0.195365712,0.1278515905,-0.2362931222,0.074982129,-0.1610620469,0.1566876322,0.1923018098,0.1800318509,-0.0995386392,-0.0737074092,0.0470809266,-0.1955215633,-0.1439724863,-0.1943651289,0.0137509201,0.0518596806,-0.1114363894,-0.1601402909,-0.4644373953,-0.2090989947,0.5759682655,0.4488873482,0.1323640347,0.1834955066,-0.2322513461,0.4705428779,-0.2687545419,-0.0973618999,-0.1350892186,-0.4322135448,0.1380234659,-0.2309996188,-0.2651397288,0.2041730434,0.08503142,0.0729781091,-0.369435966,-0.1806652248,-0.2982805669,0.2305029929,-0.0645077378,0.0144791082,0.0836250484,0.2201632112,-0.3017017841,0.1690779179,-0.1146730855,-0.4005320668,0.1927462965,0.4177711308,0.2112343013,0.3016082048,0.1985858232,-0.121679306,0.4930608571,0.0962475762,-0.166952908,0.3587188721,-0.0510834903,0.2181683183,0.0389962494,-0.4225400984,0.0670210868,0.2781130075,0.2694309652,0.4079959989,0.1909218878,0.0648137033,-0.3097715974,-0.0376838185,-0.0374503992,0.14918001,-0.0980652049,0.2544055879,-0.1242515594,0.0780041367,-0.1170751676,-0.5316746235,-0.3109454215,-0.19868505,0.2032075226,0.2260257006,0.2644037604,-0.7062585354,0.0857222378,-0.5139964223,0.2198425233,-0.0145958308,0.1183047071,-0.0337569788,0.0045152465,0.1651839763,0.1520133317,0.522543788,0.1951159537,-0.1938698143,-0.0398447886,0.1011936218,-0.516787529,0.2903286517,0.0574816801,0.2678439617,0.3412585258,0.4525682926,-0.1791606545,-0.1964899302,0.2660779655,-0.0128566753,-0.3577558696,-0.4303629994,-0.3104012311,-0.1045342758,-0.1124195307,-0.1667010337,-0.1923720539,0.139787823,0.1549269855,-0.0169027206,0.1546111703,-0.0644374788,0.0979746878,0.172781989,-0.0299724769,0.5414875746,-0.2067313343,0.3115109205,0.203204006,-0.4954355657,0.5356373191,-0.3347657323,-0.3403835297,-0.1891472191,-0.0664271489,0.6380007863,0.1867889166,0.4127139151,0.0716604739,-0.1701103598,-0.1324496418,-0.2060520947,0.1522977352,0.1808265448,-0.0818560645,-0.0620702095,-0.3783703148,0.6578394771,-0.1484383345,0.0152406227,0.1428601891,0.1314786226,-0.2073984146,0.1678875238,-0.1756636202,0.9181867838,-0.4013736248,0.30450809,0.0658267364,-0.0023988115,0.6905835271,-0.1263768375,-0.1575800627,-0.4469189644,0.0318160132,-0.1245350465,-0.111707814,0.1562142521,0.6060447693,-0.2904159427,0.3330254853,0.1022632346,-0.1566510051,0.0857458636,0.5474103689,0.1954794675,-0.55217731,-0.6150876284,0.0507820025,-0.271309495,0.2174417675,-0.1225028038,-0.144471243,0.0176996663,-0.1526005119,0.0209327843,0.1539129466,-0.1413407922,-0.3787118793,0.3384369314,-0.1711135358,0.1288845241,0.5490393043,0.179016158,0.0211097971,-0.0140784886,-0.0278722718,-0.1363568902,-0.0982230157,-0.2008842677,0.3473661542,0.2184488177,-0.1002631411,-0.0878752917,0.0284437817,0.1698334068,-0.2046904862,-0.2876246274,0.2139721364,0.1774089485,-0.3118366301,-0.5614944696,0.0610748008,0.173177734,0.1174962819,-0.002452007,0.2749664783,-0.0268653575,-0.0674516857,0.4489811361,0.0327360295,0.1277525425,0.3642868102,-0.3437955379,-0.4746578336,0.3770820796,0.3455297947,0.186771512,0.0283463188,0.1017418727,0.126509428,-0.1253415793,-0.1523979008,0.463652879,-0.0035683229,-0.0824796706,0.2397326529,0.0979017094,-0.1180921495,0.095111616,-0.1770331115,-0.4044647217,0.0799606219,0.0342609733,-0.0030445524,0.4400976002,0.0292635839,0.2341208905,0.2732691765,-0.1530267298,0.0812960044,-0.2051444799,0.1095779315,0.64223665,-0.1181754395,0.0158409942,0.1295917034,-0.03905119,0.0026292743,-0.079917863,-0.0590643361,-0.2287062854,0.2000668347,0.2207164168,-0.063768588,-0.1922050118,-0.2222703695,-0.1011904925,-0.1705963016,-0.0201164484,0.1290994138,0.0393288843,0.3361453414,0.1420578957,0.1637385786,0.056226518,0.0069031692,0.1409445107,0.2213820666,0.087616615,-0.0147384843,-0.1236979589,0.0388005152,-0.2033963948,0.233669281,0.5907527804,0.1211756766,0.1096919328,-0.0694430843,0.0568298697,-0.1275362819,0.0116588566,0.0961743444,-0.1826571524,0.1929765195,0.2257422358,0.0098454598,-0.1360057145,-0.031072652,0.6075597405,-0.0004742048,-0.1009344384,0.2780642509,-0.1659267992,0.2091235369,-0.0450682379,-0.1710895449,0.4275469184,0.1424685121,-0.1523714662,-0.1849141866,-0.2427366823,-0.2652885616,0.2519276738,0.331461072,0.1331572086,-0.0282075778,-0.2160606831,0.1581285298,-0.2844087183,0.0729062483,0.355630666,-0.3562244475,-0.1422437429,0.3530160189,0.384842217,0.4920444191,-0.1813126951,0.9246274829,-0.1608626395,-0.1758285016,-0.1373135746,0.147454679,0.1082065627,-0.1151964813,-0.2877688706,0.0301647633,-0.0729135647,-0.120222494,-0.1534544379,-0.4575326741,-0.1800722033,-0.1211273372,0.1582866162,-0.0190677028,-0.2579399347,0.2206989974,-0.0839099288,-0.0515033379,0.2033429295,-0.1597410142,0.3006352782,0.4441347718,0.1342652887,-0.0668930933,0.3013826311,0.1333032548,0.2255007625,-0.0017807863,0.0367036648,0.1465034932,0.0834804624,-0.0356278643,0.1313693672,0.0863389671,0.013292999,-0.0139526371,-0.0809811056,0.1686645001,0.0239331629,-0.3790731132,0.205854252,-0.1466253549,-0.1522453129,-0.5125371218,0.2595129907,-0.5331436992,-0.1909729093,0.0540700592,0.2861725092,0.0584585816,-0.3261230588,0.0149251884,-0.2337177396,0.465311408,0.2223500162,0.0171876121,0.0972261354,-0.2738226056,-0.4709037542,0.1958318204,0.140380308,-0.2856005728,0.0062162112,0.0637056157,-0.0830347091,0.1276968271,-0.0189351328,-0.1104155555,-0.1545847505,-0.1286301315,-0.2602014542,-0.229012996,-0.1174246967,0.2523983419,0.0827328488,0.1446991265,-0.115443863,0.1431581825,-0.2001928389,-0.1366018355,0.123798877,-0.0176783167,-0.0245251153,0.1067199036,-0.2734106481,0.4452265799,0.1341597289,0.0438431352,0.0975182578,-0.0314512067,-0.2656061649,0.0596044324,-0.1217647493,0.2434993684,-0.1983081698,0.0861098915,-0.1347760707,0.0999794081,-0.1467711329,-0.1571557671,-0.1059980318,-0.081195198,-0.1691467315,0.0005242907,0.1726085097,0.4137925208,-0.0705687776,0.1774505973,-0.0448196679,0.1875441223,0.4689190984,-0.4195564985,-0.1246413291,0.1551665664,0.151924789,-0.0655928105,-0.3130161762,-0.4497193098,0.1387549788,0.2284657955,0.0515746102,0.0915838405,0.4112349749,-0.5065764189,-0.040448118,-0.1021146402,-0.0395150967,0.3436336517,0.0974308252,-0.0060921274,-0.0771244913]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/234","title":"Huggingface NLP, Uploading custom dataset","comments":"Since the latest release 0.2.1 you can use \r\n```bash\r\nnlp-cli upload_dataset <path\/to\/dataset>\r\n```\r\nwhere `<path\/to\/dataset>` is a path to a folder containing your script (ex: `squad.py`).\r\nThis will upload the script under your namespace on our S3.\r\n\r\nOptionally the folder can also contain `dataset_infos.json` generated using\r\n```bash\r\nnlp-cli test <path\/to\/dataset> --all_configs --save_infos\r\n```\r\n\r\nThen you should be able to do\r\n```python\r\nnlp.load_dataset(\"my_namespace\/dataset_name\")\r\n```","body":"Hello,\r\n\r\nDoes anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp.\r\n\r\nThank you!","comment_length":63,"text":"Huggingface NLP, Uploading custom dataset\nHello,\r\n\r\nDoes anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp.\r\n\r\nThank you!\nSince the latest release 0.2.1 you can use \r\n```bash\r\nnlp-cli upload_dataset <path\/to\/dataset>\r\n```\r\nwhere `<path\/to\/dataset>` is a path to a folder containing your script (ex: `squad.py`).\r\nThis will upload the script under your namespace on our S3.\r\n\r\nOptionally the folder can also contain `dataset_infos.json` generated using\r\n```bash\r\nnlp-cli test <path\/to\/dataset> --all_configs --save_infos\r\n```\r\n\r\nThen you should be able to do\r\n```python\r\nnlp.load_dataset(\"my_namespace\/dataset_name\")\r\n```","embeddings":[0.280041188,-0.2268893868,0.120161891,0.1158971786,-0.0624757409,-0.0111734392,0.2993092537,-0.0362044014,0.0847581029,-0.0956161693,-0.2538748085,0.3167442977,-0.0799233466,0.3306141794,0.5891800523,0.1280133426,0.0025305294,0.2043104917,-0.0012188745,-0.0587979071,-0.2672669888,0.3734842837,-0.1716811508,0.1110452861,-0.0410867184,0.1917255223,-0.1079308912,0.5931818485,-0.0679470599,-0.6077477336,0.4877357781,0.193832919,0.3708676398,0.409065038,-0.0001233944,-0.3133193552,0.2694946826,-0.2316975147,-0.3273893297,-0.3374843597,0.1249143034,-0.4609065354,0.3682494164,-0.2326495647,-0.0676464364,0.1287667751,0.235003069,-0.0852788314,0.4762735963,0.0823692679,0.0159708504,0.2765752077,-0.3785421252,0.0509530269,0.0250980575,0.2564996481,0.0095828464,0.0897440761,0.4341789484,-0.3698439598,0.0935192108,-0.0385057665,0.0429343171,-0.1325575262,0.6434711218,0.1918417066,-0.3280702233,-0.1321462542,-0.2390377969,0.2397617698,0.3419497013,-0.4234245121,-0.325639993,-0.4911620021,0.1315747201,-0.1563498974,0.2453739643,0.0697207898,-0.1712663323,-0.0648642853,-0.3384968042,-0.1390088201,-0.3118106127,0.2429273427,0.2955783606,0.2280941606,-0.2206053585,0.1583246142,0.235191226,-0.2058346719,-0.0886887908,0.1007154062,0.1064787805,0.564686954,-0.3737162948,-0.2899183929,-0.1165457219,0.10073971,0.2578741312,0.1344522089,0.1770187467,-0.0359660685,-0.1233959645,0.0913675278,0.390314877,-0.0014647252,0.5357534885,-0.1598761529,-0.0851837173,0.0251783356,0.0550952889,-0.0850275531,-0.4656982124,-0.0134292534,-0.2789924145,-0.2053917497,0.1622759849,-0.2037165463,0.0875260904,-0.2311067432,0.3427425325,-0.1016345471,0.1443826258,0.2277094871,-0.0353277661,-0.1291587502,-0.0205230527,0.1601542681,-0.0732647851,-0.1987454593,0.0590833277,-0.0491438583,-0.1831965744,0.0938047692,0.3327322006,-0.0904485211,0.4279941916,-0.1487330645,0.1311012954,-0.0102179246,-0.1246866733,0.24250108,-0.1801869869,-0.1258146018,0.2880768776,-0.0189978424,0.115399614,-0.3117888868,-0.2295581251,-0.0522932187,-0.1529917717,-0.0616694763,-0.3780478537,-0.0268930066,-0.3280336857,-0.4907149971,-0.2676517963,0.2743685246,-0.1532911658,-0.2338869572,-0.0303353649,0.1520068198,-0.2674449682,-0.1086116731,0.1839228123,0.2969834507,-0.4266986549,-0.302161783,0.0420350097,-0.0123940324,-0.1056095734,0.1387207955,-0.2472092658,-0.0542636998,-0.15707919,0.2555834949,0.574765563,-0.4837093353,-0.1388399899,0.2292327583,-0.2719882429,0.0131997671,0.0639390051,0.2463511378,0.1747874469,0.01045743,-0.0625673085,0.7491369247,0.0347712263,-0.0652870312,0.079337962,-0.0201671943,0.1566513777,0.1346314698,-0.3383793235,0.1531339139,0.0922186971,0.198026076,0.1743172258,-0.0980526507,-0.0739667267,0.034192659,0.1123767048,0.417150557,0.0026158318,0.0373882465,-0.695799768,0.2870245874,-0.2856880724,-0.0026063058,0.2908089459,-0.0849649012,-0.4383996725,-0.1123043969,-0.1052105948,0.1792892963,-0.0818910822,0.1696848273,0.2380019575,-0.1964533329,-0.4215344489,0.3596050739,-0.1228931174,0.2609522939,-0.4864855409,0.0444273315,0.2445634604,-0.1797391027,0.1138580963,0.3458587527,0.2149309367,-0.1512641758,0.1970122159,0.3514067531,-0.1065535247,0.3253590763,0.3652241528,0.0149927866,-0.0511014163,0.1266390532,0.2454586029,-0.1220536903,0.072183229,-0.1380444467,-0.3422394693,0.3357030451,-0.1649213731,0.2453159541,-0.0025176776,-0.1462320238,0.0532507189,-0.0891687125,-0.42794469,-0.0027036541,0.0352217779,0.1198277399,0.3981923163,-0.0135125397,-0.0769134611,-0.0973861739,0.6103428602,-0.1032547727,0.0376774333,0.2313651145,-0.1037518829,-0.1961955726,0.1441417187,-0.0011320115,0.3079428077,0.1700057536,-0.1886392981,0.0464922376,-0.0874040574,0.0712431967,-0.0334302746,0.0001723964,0.0166194774,-0.2834673226,0.2029100955,-0.1489429325,-0.317085743,0.0581376888,-0.0794910938,-0.1055355296,-0.257717669,0.1330084354,0.002569594,-0.3521052599,-0.2281140834,-0.3058689237,-0.4979859293,0.082530655,-0.3920715153,-0.0814008042,0.5837751031,-0.0671747774,0.1321170479,0.5852800608,-0.1560715139,-0.1986920238,0.0728469267,-0.0184183307,-0.2793127298,-0.1237122118,0.0595833659,0.1815273315,0.3396261632,-0.0761952922,0.1956444979,0.080611214,-0.2836532295,0.0786687732,-0.1990657598,0.1872478426,0.3521501422,0.1134731323,-0.1329292059,0.0083256094,0.1063299105,-0.1537938118,-0.1182949841,-0.0744151697,0.0029164657,0.07503362,-0.120759055,-0.0856740549,-0.3339976072,-0.1405495256,0.5175922513,0.3240983784,0.1171012595,0.1943234652,-0.2162727863,0.4645294845,-0.3083764017,-0.0089304494,-0.1356204152,-0.5117289424,0.032068789,-0.1415932626,-0.1379126757,0.17673783,0.1191289499,0.1214162558,-0.3518004119,-0.1929907054,-0.4288131297,0.219819665,0.0215193424,0.0349836424,0.1249322593,0.2229184061,-0.2214020044,0.1917724162,-0.1415399611,-0.4419384301,0.0736217573,0.418623656,0.2334130555,0.2650006413,0.2457137257,-0.0208604634,0.5214276314,0.1515869647,-0.1895602345,0.3977134228,-0.0435568728,0.2010843158,0.1313114464,-0.3892534971,0.0604921915,0.2669017911,0.2402645648,0.3672428429,0.1753045619,0.1485011727,-0.269067347,0.0715689436,-0.1303426623,0.1785045266,-0.1290689558,0.2322729379,-0.043136891,0.0453356653,-0.1490938365,-0.4444147944,-0.2447847873,-0.1872436404,0.2459124476,0.2474557757,0.3023855984,-0.665309608,-0.0887597203,-0.4830067456,0.1967628151,-0.09480948,0.1488329023,0.071200259,0.0455645956,0.1687134951,0.0544941723,0.6477319598,0.0724977702,-0.1894096583,-0.0538664609,0.0132946577,-0.6558192968,0.2496963143,0.0632444248,0.2618872821,0.3039624095,0.5747789741,-0.1860575527,-0.2908875346,0.1541653425,-0.0302494597,-0.3178240359,-0.3237392902,-0.3765396774,-0.1485574692,-0.1410322785,-0.2114433646,-0.1780858636,0.1467426419,0.2263942659,0.0475949086,0.1216973588,0.0249024779,0.0578252226,0.2943685949,-0.0584113374,0.5272747874,-0.1376094371,0.2554588914,0.1654529423,-0.3756713867,0.6000048518,-0.2558785081,-0.4437695444,-0.1565438807,0.1167188212,0.6394878626,0.196826309,0.323435843,0.1490115821,-0.0935884714,-0.1095406786,-0.3384194374,0.1336222142,0.2434947789,-0.0431515388,0.0663953125,-0.3870314658,0.5882586837,-0.1043478698,-0.034468323,-0.1031986997,0.2037136704,-0.2224707007,0.1742636263,-0.1180261523,0.8925431967,-0.4042233229,0.3991562724,0.1949102283,-0.0083395829,0.7712113261,-0.0710915402,-0.1952488571,-0.4421927631,0.2296813279,-0.1013706326,-0.1051748395,0.1328963488,0.4739500284,-0.2339793295,0.3612963259,0.1341591775,-0.1984679103,0.0603134073,0.56814152,0.2075704187,-0.5780225992,-0.6750756502,0.019689694,-0.3262565434,0.2848817706,-0.109548524,-0.0888957307,-0.0481779315,-0.154211536,0.0446484499,0.113869898,-0.0431925878,-0.4616310596,0.3202899098,-0.3091632426,0.157186538,0.5623395443,0.2044418901,0.0461316183,-0.0031004711,-0.0286613181,-0.0935123414,-0.1051057354,-0.1975781322,0.1908003986,0.0855261311,-0.0997290313,-0.0741496533,-0.0659804121,0.2464985251,-0.1773228794,-0.1360149682,0.210790351,0.128249228,-0.3042229116,-0.5435206294,-0.0227832161,0.1489761472,0.0947442129,-0.0221788194,0.2583086789,-0.1068245023,-0.1202172488,0.3054865301,-0.0312733948,0.1551933289,0.418386966,-0.3858992457,-0.363702476,0.338247627,0.3833316267,0.2017428577,0.0157984961,0.0901361927,0.0929368436,-0.0988090113,-0.129210487,0.4974895716,-0.1088051423,0.0067566917,0.1091715544,0.1795967519,-0.1827742308,0.1114075258,-0.1940450072,-0.2364113331,0.0853515714,0.0332723856,0.0233593769,0.4362819195,-0.1379793584,0.1552195251,0.0948854312,-0.1214115918,0.1092169583,-0.2049249858,0.182827279,0.4841090441,-0.1927681714,0.0605598241,0.0724810883,-0.0612081923,-0.0592073984,-0.0654788762,-0.0538957641,-0.2560448945,0.2277764529,0.2528777421,-0.0992402062,-0.1135668904,-0.204385072,-0.1438851655,-0.2003828585,0.0288844816,0.1565558314,0.1449348778,0.37540555,-0.0055717379,0.0753425285,0.0791385099,0.0301104337,0.1033793166,0.132243067,0.0779463723,0.0988947228,-0.0701759979,0.0210393313,-0.2054564357,0.2262936085,0.5412856936,0.1187421307,0.0873250961,-0.1680331677,0.0177013688,-0.1014919207,0.1515491009,0.1994531453,-0.1798024774,0.0797244012,0.1812307388,-0.0083946167,-0.051256761,-0.1340385079,0.4174162447,-0.021750005,-0.0597212017,0.3559253514,-0.1728089899,0.299113065,-0.140550524,-0.1669453233,0.4358364642,0.2823688686,-0.1205790117,-0.1596356481,-0.1951531172,-0.2329356223,0.2651916444,0.3151313663,0.0482668541,0.0240414143,-0.2267839015,0.1274550259,-0.2008691877,0.1938616037,0.3003180921,-0.3114257753,-0.134833917,0.2487302274,0.322021842,0.4606935084,-0.2512055337,0.8900499344,-0.2082508802,-0.1327005774,-0.1670141071,0.1701205373,0.1314441413,-0.2743003964,-0.2992436886,0.0876147896,-0.0265152473,-0.0721764565,-0.1698849052,-0.5302137136,-0.2090249509,-0.0866336823,0.2120303363,-0.0965778157,-0.2525299191,0.2364838123,-0.1175021529,-0.1024066359,0.2297776639,-0.1354084164,0.2758542597,0.3882068098,0.0975420251,0.0508956574,0.2564308941,0.0338268466,0.2085303068,0.0011941746,0.0248338692,0.0424416214,-0.0014728913,-0.0936867073,0.2220139354,0.0412091836,-0.0067534912,0.0134942364,-0.1261180937,0.0829006433,-0.0329392925,-0.3117902577,0.2223142534,-0.2063256353,-0.0874984562,-0.4465425909,0.4152689576,-0.619402051,-0.1329186112,0.2620144188,0.3689613938,0.1095535606,-0.2574582696,-0.0046852194,-0.2274147719,0.4847807884,0.1143018827,0.0827308446,0.055874154,-0.1872499436,-0.4674095511,0.1330847591,0.069446981,-0.3638960719,0.0465311483,0.1530483663,-0.1559281796,0.0824585631,-0.1072245836,-0.0501367487,-0.1673272848,0.020895509,-0.2143275738,-0.1970236152,-0.1190056279,0.2408183366,0.1204425022,0.0752051249,-0.1501968354,0.0925722942,-0.2727144659,-0.1340413243,0.149219647,0.0637369379,0.0014253729,0.2092085034,-0.2594481409,0.4553104341,0.0930928662,-0.0523215532,-0.0375298746,-0.0708064064,-0.309371084,0.102636233,-0.238237679,0.2546860278,-0.2111202776,0.1825409085,-0.2470241487,0.0496094041,-0.0556599274,-0.215736106,-0.0091989161,-0.0050395401,-0.2100753039,0.0972385854,0.1297473907,0.409524709,-0.0272164904,0.2013867944,-0.1416337788,0.0713424906,0.4824336469,-0.5757673979,-0.0996544361,0.1803994477,0.2315546721,-0.1007456183,-0.419064492,-0.5331852436,0.2177242786,0.2580074966,0.070747681,-0.0044049877,0.3593482673,-0.4332230985,-0.0631085411,-0.1109695882,-0.0292665344,0.296210587,0.0780341253,0.0435730517,-0.0895020664]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/233","title":"Fail to download c4 english corpus","comments":"Hello ! Thanks for noticing this bug, let me fix that.\r\n\r\nAlso for information, as specified in the changelog of the latest release, C4 currently needs to have a runtime for apache beam to work on. Apache beam is used to process this very big dataset and it can work on dataflow, spark, flink, apex, etc. You can find more info on beam datasets [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/docs\/beam_dataset.md).\r\n\r\nOur goal in the future is to make available an already-processed version of C4 (as we do for wikipedia for example) so that users without apache beam runtimes can load it.","body":"i run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?","comment_length":96,"text":"Fail to download c4 english corpus\ni run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?\nHello ! Thanks for noticing this bug, let me fix that.\r\n\r\nAlso for information, as specified in the changelog of the latest release, C4 currently needs to have a runtime for apache beam to work on. Apache beam is used to process this very big dataset and it can work on dataflow, spark, flink, apex, etc. You can find more info on beam datasets [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/docs\/beam_dataset.md).\r\n\r\nOur goal in the future is to make available an already-processed version of C4 (as we do for wikipedia for example) so that users without apache beam runtimes can load it.","embeddings":[0.0115546789,0.0910863578,-0.0556439348,0.2902231514,0.148630634,0.1652969122,-0.0632363707,0.3174011707,-0.1166139469,0.0394295342,-0.1543145627,0.0087515097,0.0267580524,0.084500052,0.1310365945,-0.5472636223,-0.2235968858,0.1527146846,-0.131220445,-0.1412544996,-0.0423424728,0.4258191586,-0.1137883365,-0.1747365743,-0.0600037314,-0.1374861747,-0.1974728554,-0.0550695211,-0.206342712,-0.544832468,0.4454170763,0.0049821967,0.1893910617,0.3470121026,-0.0001094098,-0.0178526044,0.410225302,-0.1208095476,-0.2488911003,-0.39947474,0.1793028861,-0.3214279711,-0.0999809876,-0.3226758242,-0.2282853127,-0.1305258572,0.1668989211,-0.2270071805,0.0431619808,0.3165642321,0.2735976577,-0.0255144741,0.2802649736,-0.1330512613,-0.1831474304,-0.2402020991,-0.0908977464,0.2393375039,0.1667280197,0.2533928156,-0.0951283351,0.1529464126,0.0502758324,0.0134145711,-0.1094024777,-0.2209978402,0.0489708111,-0.2720611691,0.3356462419,0.0553029366,0.7832083702,-0.2817870378,-0.1618426442,0.1612834483,0.283929795,-0.212313965,0.2203885913,0.4282406271,-0.1392534822,0.0007200533,-0.3249097168,-0.0999147892,-0.181789875,0.2037830502,0.0209814962,0.0087318849,-0.0804654807,-0.0100580025,0.2173687667,0.1014782935,0.0793812573,-0.0789981186,-0.2777691185,0.2961800694,-0.2832439542,0.020596629,-0.1462968141,0.5204554796,0.02538017,-0.3621297181,-0.1353999525,-0.1054494008,0.0756659955,0.0276800711,-0.0533361509,0.3394507468,0.1087685227,0.0285199098,0.079799287,0.130185023,0.0743425488,-0.0499666631,-0.1396474391,-0.2227846831,-0.2987947464,0.2747345567,-0.1234494448,-0.3272612691,-0.1521591097,0.0012162372,-0.2087304294,-0.0346010141,-0.0995942354,0.4414266348,-0.1579021513,0.1526206434,0.0880152136,0.1817054152,-0.1947121471,-0.3660129607,-0.1490098834,0.2008820623,-0.468621403,-0.3205323219,0.2491369843,-0.3165852427,0.4326317608,-0.0886029601,0.107325092,-0.1022875011,-0.1421988904,-0.2225233167,-0.2582617104,0.188044548,0.067876935,0.2651393414,0.0120807774,-0.0328292362,-0.0076496922,-0.2450187951,-0.129508391,-0.1909275204,-0.072525613,0.1784774959,-0.0103534712,-0.1338939667,-0.5229309797,0.0499758162,0.1910482049,0.0380627438,-0.002433571,-0.2078721821,-0.2314997613,-0.2085983604,0.117168434,0.5448202491,-0.0188138504,0.028379038,-0.0435673185,-0.0119555918,0.4996949434,0.3720760047,-0.0997387618,0.172167778,-0.2095858306,0.1426471472,0.1664265692,-0.5432876945,-0.2214052975,0.2418850362,-0.0492497012,0.0903555453,0.0329224057,-0.4252802134,0.2009977847,-0.130916357,0.1117296368,0.2759277225,0.0992452502,0.1265601665,-0.2411850691,-0.1267588139,0.2473011166,0.1046614796,0.1690177172,-0.1020453349,0.2604554296,0.1870794445,0.2613472939,-0.1208254769,0.3596661985,0.2470407784,-0.2625224292,-0.3072057366,0.1050221324,0.19859384,-0.3305431604,0.206600979,-0.2245477438,0.2911973596,0.0532628298,-0.0218927599,-0.3256198168,-0.2064074129,-0.2410363108,-0.2011760324,0.1774201095,-0.0429085456,0.3280696571,0.1672783643,0.1197206452,0.3532952666,-0.3063316047,-0.1096612364,-0.1067400351,0.1026701927,-0.2434271425,-0.0721904933,0.3241747916,0.2424546927,0.2653382123,-0.156836763,-0.1398927718,0.0729344636,-0.1470297724,0.4155639708,-0.1165600345,-0.1203745231,0.3694632649,-0.3810000718,0.4082131386,0.317717284,0.0645144135,-0.0745139942,0.3047270179,-0.2446958125,0.0516471826,0.0077799303,0.0869080424,-0.1271286458,0.331769824,-0.091515258,0.1778372526,-0.0359349363,0.5329710841,-0.0943983495,0.2545621097,-0.1527448297,-0.1970202029,0.027613027,0.4264924526,0.0807586312,-0.1145147458,0.171007663,-0.1288840175,-0.175828889,0.2758097649,0.1189039946,0.1102279648,0.1651006788,0.1731343567,0.1311655343,0.1335800737,-0.2187917084,0.2529250383,-0.2158294767,0.1718282849,0.3353054821,0.1661093533,-0.0029510632,-0.4320878088,-0.0286971219,0.2143987715,0.1778079122,-0.0685272738,-0.0997375175,-0.2467764765,-0.4081358314,-0.3134776056,-0.1185436398,0.0371320471,0.0172460489,-0.1968258172,0.114120625,-0.0407053009,-0.0076273112,-0.2748734355,-0.3112739325,-0.1994776577,-0.2850935757,-0.1725654304,0.0059298952,-0.4090853035,0.1231618375,0.1412646472,0.0594344474,0.3566991687,-0.33560583,-0.0827376172,-0.223473832,-0.0705547929,0.2513357699,-0.0651990846,-0.1545362324,0.0221690554,0.2006978989,-0.2227170467,-0.1932851523,0.0155675327,-0.1274445355,-0.1874835044,0.0146911833,-0.2334830463,-0.155720219,-0.1595274955,-0.6133767366,-0.4746182561,-0.4864419997,0.0599954873,-0.1413523406,0.1965738386,0.3935822845,-0.2014529854,0.1657384634,0.13731879,0.1436540931,-0.2028478831,-0.0262709111,0.1178618744,-0.3232525885,-0.454369545,0.117317833,0.0238614138,0.4150661826,0.1647234112,-0.3046405315,0.3316366673,-0.032270696,0.1291528493,0.0913803279,0.1418595463,0.4941840768,-0.0997108519,-0.0022173061,0.0292065702,0.1109181046,-0.0067351903,0.2292229086,0.4109459519,-0.2725373507,0.1508922577,0.1694056392,0.3010313809,0.3988194764,-0.2326139212,0.6556056738,0.3299936354,0.1660488546,-0.0785605311,-0.223159954,0.4240542054,-0.0242026802,0.050451912,0.3943394423,-0.1681078672,-0.560706377,-0.1106758937,-0.4160575271,-0.4069871306,-0.0655990541,-0.0058622886,0.1179788038,0.4066350162,0.0412172377,0.179729417,0.2586691678,-0.2891934812,0.1588014811,0.222110793,0.0307035279,0.2962247729,-0.0844785944,-0.1370988339,-0.7133892775,0.3157611787,0.2892406583,0.1681810617,-0.1471838206,0.3811462522,0.0132091008,-0.1939187348,0.540183723,-0.2146276981,0.1172232628,0.0385664403,0.0394678414,-0.082289353,0.0103465356,0.0670095608,0.1176479086,0.4223914742,0.0589513667,-0.4840069711,-0.0881331116,-0.0218050499,0.4767236412,0.1708550751,-0.2185003012,-0.0499941818,-0.5328354239,-0.2335794419,0.007856288,0.0180500951,0.3688515723,-0.0234136004,-0.007777845,0.178348586,0.0718784854,0.1867159754,0.1425793469,0.2800192833,0.2922777236,0.200650543,0.3297092319,0.0043779006,0.1921951771,0.3183093369,-0.148680225,0.1402457356,-0.0530247316,0.1164921597,0.1367502511,0.2045944482,0.0301326998,-0.0627524033,-0.1415639669,-0.197998628,0.146508038,0.0835249424,0.0963886976,0.1257433295,-0.5703757405,-0.3882588744,0.2796278596,-0.3202518225,0.1595812887,0.1544818729,0.3057815433,-0.339107275,0.2559328675,0.0795460567,0.7368466854,-0.0402690209,0.0766436979,0.0286077913,0.0771994665,0.2957970202,-0.3393850923,0.179179728,-0.2431397438,0.1196998507,-0.0036815894,0.077733703,-0.0203926973,0.1365586519,-0.3911959827,0.5633780956,0.0777336881,-0.0448268466,0.0642819181,0.3738422096,-0.1419715285,-0.3766852617,-0.4003916681,0.1819061935,-0.3179160058,0.3305777907,-0.208878845,-0.0886835158,0.1062344611,-0.1021086574,-0.3304612637,0.4157050848,-0.2829901278,0.0253621247,-0.0954853967,-0.1803955734,0.4980296493,0.24747926,0.3305479288,0.374456197,-0.2370951921,0.120631963,0.1032213718,-0.5300790071,0.2447067946,0.021626763,0.0547361709,-0.3400761485,-0.2175462395,0.0930889472,-0.0411833115,-0.0525409877,0.2623298168,-0.1425350606,-0.3336849809,0.0105105313,-0.3549546599,0.2626432478,-0.0046562646,0.0392499901,0.1276262254,-0.0834263116,-0.1788953841,0.1112684011,0.1640378982,-0.2192924768,-0.0932556242,0.311394304,0.1225598529,-0.3694102466,0.470108211,0.2693347335,-0.2435082346,-0.1901333928,-0.6724639535,-0.2965880632,-0.3600971103,-0.0081366608,-0.0068643484,0.0362955108,-0.1432041675,0.3669493794,-0.0289171077,-0.0073453444,0.2249272019,-0.3883952796,-0.2859397829,0.0696738884,-0.1597448587,0.0285757296,0.2387289405,0.2828913033,0.1914266944,0.1129984111,-0.2709113061,-0.0451796874,-0.3741313815,0.0734555945,0.2913359702,-0.1015430987,0.3328825235,0.1120014787,0.1091698781,0.2557425201,-0.2520971298,-0.2090447694,-0.1103292257,0.0847762004,0.1579771787,0.0817384645,0.3228664398,-0.3832597435,-0.1494952142,-0.028482357,0.126094833,0.3018396795,-0.0339369401,0.1260331422,0.2406696826,0.0628555119,-0.0867634267,0.1424496472,0.1167202741,0.0945004001,-0.0211362783,0.1423729509,0.067919977,-0.0810720176,-0.1436301917,-0.0144849047,0.1109489575,0.0499675982,0.1543722302,-0.0368774943,0.1027513221,-0.0054851673,-0.0454933383,0.0074303565,-0.0234520882,-0.0358009003,0.3045820594,0.2066614926,-0.2701682448,0.1760812104,-0.0184231084,0.1721590906,0.0045921775,0.0139468545,0.3261063695,-0.2998712361,0.1954528838,0.0973289311,0.3590956032,-0.1300854832,-0.1550516337,-0.0282243006,0.0612748973,0.0415270887,0.2245525867,-0.1463569552,0.0086984821,0.1348892301,-0.0893119946,-0.0116788819,0.2734456062,0.2953779697,-0.0654611513,-0.2115878016,0.0072166775,0.3011555672,0.2955462039,0.2288535535,0.1197468266,0.1959207058,0.1732964814,-0.200048998,-0.3275516331,0.3751668632,-0.0789166242,0.0393313244,-0.1925189644,-0.2186932266,-0.0818450674,0.1360602379,-0.1872272342,-0.0659432039,0.1962611973,0.2184791416,-0.2186942101,-0.4077454209,-0.0435331203,0.0160435289,0.1576144248,-0.2169872075,0.4991546273,0.2705328763,-0.054660134,-0.1632510126,0.314086169,0.4485455155,0.2007255554,-0.1284498572,0.0977757648,0.1225194037,-0.0104406327,-0.0364384986,0.4297582507,0.3452330232,0.1135912612,0.282918483,0.1092250496,-0.2205705345,-0.0985294208,0.0933237672,-0.0567233674,0.1250903606,0.451225847,0.0044774292,0.0749910474,-0.3351687193,0.0485384502,-0.480506897,0.4305048883,0.4933682084,-0.1941915303,-0.0126123652,-0.1042474657,0.0864552259,0.0925589874,0.3288391829,0.2744688094,0.4959403872,-0.4177182019,-0.3438930511,-0.3343045115,0.1863348931,-0.2957495749,0.1797835678,-0.2519392371,0.0594345629,0.143458575,0.1007816344,-0.1272498667,0.366692394,0.1858077347,0.2237451226,-0.1296077967,0.1150657088,-0.1630086899,0.0071013104,-0.006841823,-0.4851977825,-0.1454279423,-0.0727040321,0.0664691925,-0.2589695156,0.0846332908,0.0441818535,-0.1518297344,0.135022223,0.0279058758,0.6574344039,-0.2393392473,-0.1950959712,-0.2961331308,-0.2016223073,-0.1850347966,-0.0286590047,0.0989930332,0.3526166975,0.0663160384,0.1508650333,-0.1567121297,0.4380097985,0.1698628068,0.3453728259,-0.1365328282,0.1078100652,-0.3517187536,-0.1205077767,0.0466760062,0.506477654,0.0218654592,-0.1784384698,-0.2926490903,-0.5317479372,0.5479229093,-0.8124235868,-0.2576377988,0.0784190819,0.0957219377,-0.0114729004,-0.2659358084,-0.5267435908,0.2552785277,0.0984909907,-0.0782416686,-0.327549547,0.1157159433,-0.0942824036,-0.0646765679,-0.0526354015,0.3443565071,0.1893855482,-0.2799617648,-0.1954742223,-0.3156386018]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/233","title":"Fail to download c4 english corpus","comments":"@lhoestq I am facing `IsADirectoryError` while downloading with this command.\r\nCan you pls look into it & help me.\r\nI'm using version 0.4.0 of `nlp`.\r\n\r\n```\r\ndataset = load_dataset(\"c4\", 'en', data_dir='.', beam_runner='DirectRunner')\r\n```\r\n\r\nHere's the complete stack trace.\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/devops\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0\/096df5a27756d51957c959a2499453e60a08154971fceb017bbb29f54b11bef7...\r\n\r\n---------------------------------------------------------------------------\r\nIsADirectoryError Traceback (most recent call last)\r\n<ipython-input-11-f622e6705e03> in <module>\r\n----> 1 dataset = load_dataset(\"c4\", 'en', data_dir='.', beam_runner='DirectRunner')\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 547 # Download and prepare data\r\n 548 builder_instance.download_and_prepare(\r\n--> 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n 551 \r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 461 if not downloaded_from_gcs:\r\n 462 self._download_and_prepare(\r\n--> 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n 465 # Sync info\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos)\r\n 964 pipeline = beam_utils.BeamPipeline(runner=beam_runner, options=beam_options,)\r\n 965 super(BeamBasedBuilder, self)._download_and_prepare(\r\n--> 966 dl_manager, verify_infos=False, pipeline=pipeline,\r\n 967 ) # TODO handle verify_infos in beam datasets\r\n 968 # Run pipeline\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 516 split_dict = SplitDict(dataset_name=self.name)\r\n 517 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 518 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/datasets\/c4\/096df5a27756d51957c959a2499453e60a08154971fceb017bbb29f54b11bef7\/c4.py in _split_generators(self, dl_manager, pipeline)\r\n 187 if self.config.realnewslike:\r\n 188 files_to_download[\"realnews_domains\"] = _REALNEWS_DOMAINS_URL\r\n--> 189 file_paths = dl_manager.download_and_extract(files_to_download)\r\n 190 \r\n 191 if self.config.webtextlike:\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py in download_and_extract(self, url_or_urls)\r\n 218 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 219 \"\"\"\r\n--> 220 return self.extract(self.download(url_or_urls))\r\n 221 \r\n 222 def get_recorded_sizes_checksums(self):\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py in download(self, url_or_urls)\r\n 156 lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n 157 )\r\n--> 158 self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n 159 return downloaded_path_or_paths\r\n 160 \r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py in _record_sizes_checksums(self, url_or_urls, downloaded_path_or_paths)\r\n 106 flattened_downloaded_path_or_paths = flatten_nested(downloaded_path_or_paths)\r\n 107 for url, path in zip(flattened_urls_or_urls, flattened_downloaded_path_or_paths):\r\n--> 108 self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n 109 \r\n 110 def download_custom(self, url_or_urls, custom_download):\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/utils\/info_utils.py in get_size_checksum_dict(path)\r\n 77 \"\"\"Compute the file size and the sha256 checksum of a file\"\"\"\r\n 78 m = sha256()\r\n---> 79 with open(path, \"rb\") as f:\r\n 80 for chunk in iter(lambda: f.read(1 << 20), b\"\"):\r\n 81 m.update(chunk)\r\n\r\nIsADirectoryError: [Errno 21] Is a directory: '\/'\r\n\r\n```\r\n\r\nCan anyone please try to see what I am doing wrong or is this a bug?","body":"i run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?","comment_length":341,"text":"Fail to download c4 english corpus\ni run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?\n@lhoestq I am facing `IsADirectoryError` while downloading with this command.\r\nCan you pls look into it & help me.\r\nI'm using version 0.4.0 of `nlp`.\r\n\r\n```\r\ndataset = load_dataset(\"c4\", 'en', data_dir='.', beam_runner='DirectRunner')\r\n```\r\n\r\nHere's the complete stack trace.\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, post-processed: Unknown sizetotal: Unknown size) to \/home\/devops\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0\/096df5a27756d51957c959a2499453e60a08154971fceb017bbb29f54b11bef7...\r\n\r\n---------------------------------------------------------------------------\r\nIsADirectoryError Traceback (most recent call last)\r\n<ipython-input-11-f622e6705e03> in <module>\r\n----> 1 dataset = load_dataset(\"c4\", 'en', data_dir='.', beam_runner='DirectRunner')\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 547 # Download and prepare data\r\n 548 builder_instance.download_and_prepare(\r\n--> 549 download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications,\r\n 550 )\r\n 551 \r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 461 if not downloaded_from_gcs:\r\n 462 self._download_and_prepare(\r\n--> 463 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 464 )\r\n 465 # Sync info\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos)\r\n 964 pipeline = beam_utils.BeamPipeline(runner=beam_runner, options=beam_options,)\r\n 965 super(BeamBasedBuilder, self)._download_and_prepare(\r\n--> 966 dl_manager, verify_infos=False, pipeline=pipeline,\r\n 967 ) # TODO handle verify_infos in beam datasets\r\n 968 # Run pipeline\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 516 split_dict = SplitDict(dataset_name=self.name)\r\n 517 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)\r\n--> 518 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n 519 # Checksums verification\r\n 520 if verify_infos:\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/datasets\/c4\/096df5a27756d51957c959a2499453e60a08154971fceb017bbb29f54b11bef7\/c4.py in _split_generators(self, dl_manager, pipeline)\r\n 187 if self.config.realnewslike:\r\n 188 files_to_download[\"realnews_domains\"] = _REALNEWS_DOMAINS_URL\r\n--> 189 file_paths = dl_manager.download_and_extract(files_to_download)\r\n 190 \r\n 191 if self.config.webtextlike:\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py in download_and_extract(self, url_or_urls)\r\n 218 extracted_path(s): `str`, extracted paths of given URL(s).\r\n 219 \"\"\"\r\n--> 220 return self.extract(self.download(url_or_urls))\r\n 221 \r\n 222 def get_recorded_sizes_checksums(self):\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py in download(self, url_or_urls)\r\n 156 lambda url: cached_path(url, download_config=self._download_config,), url_or_urls,\r\n 157 )\r\n--> 158 self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n 159 return downloaded_path_or_paths\r\n 160 \r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/utils\/download_manager.py in _record_sizes_checksums(self, url_or_urls, downloaded_path_or_paths)\r\n 106 flattened_downloaded_path_or_paths = flatten_nested(downloaded_path_or_paths)\r\n 107 for url, path in zip(flattened_urls_or_urls, flattened_downloaded_path_or_paths):\r\n--> 108 self._recorded_sizes_checksums[url] = get_size_checksum_dict(path)\r\n 109 \r\n 110 def download_custom(self, url_or_urls, custom_download):\r\n\r\n\/data\/anaconda\/envs\/hf\/lib\/python3.6\/site-packages\/nlp\/utils\/info_utils.py in get_size_checksum_dict(path)\r\n 77 \"\"\"Compute the file size and the sha256 checksum of a file\"\"\"\r\n 78 m = sha256()\r\n---> 79 with open(path, \"rb\") as f:\r\n 80 for chunk in iter(lambda: f.read(1 << 20), b\"\"):\r\n 81 m.update(chunk)\r\n\r\nIsADirectoryError: [Errno 21] Is a directory: '\/'\r\n\r\n```\r\n\r\nCan anyone please try to see what I am doing wrong or is this a bug?","embeddings":[0.0115546789,0.0910863578,-0.0556439348,0.2902231514,0.148630634,0.1652969122,-0.0632363707,0.3174011707,-0.1166139469,0.0394295342,-0.1543145627,0.0087515097,0.0267580524,0.084500052,0.1310365945,-0.5472636223,-0.2235968858,0.1527146846,-0.131220445,-0.1412544996,-0.0423424728,0.4258191586,-0.1137883365,-0.1747365743,-0.0600037314,-0.1374861747,-0.1974728554,-0.0550695211,-0.206342712,-0.544832468,0.4454170763,0.0049821967,0.1893910617,0.3470121026,-0.0001094098,-0.0178526044,0.410225302,-0.1208095476,-0.2488911003,-0.39947474,0.1793028861,-0.3214279711,-0.0999809876,-0.3226758242,-0.2282853127,-0.1305258572,0.1668989211,-0.2270071805,0.0431619808,0.3165642321,0.2735976577,-0.0255144741,0.2802649736,-0.1330512613,-0.1831474304,-0.2402020991,-0.0908977464,0.2393375039,0.1667280197,0.2533928156,-0.0951283351,0.1529464126,0.0502758324,0.0134145711,-0.1094024777,-0.2209978402,0.0489708111,-0.2720611691,0.3356462419,0.0553029366,0.7832083702,-0.2817870378,-0.1618426442,0.1612834483,0.283929795,-0.212313965,0.2203885913,0.4282406271,-0.1392534822,0.0007200533,-0.3249097168,-0.0999147892,-0.181789875,0.2037830502,0.0209814962,0.0087318849,-0.0804654807,-0.0100580025,0.2173687667,0.1014782935,0.0793812573,-0.0789981186,-0.2777691185,0.2961800694,-0.2832439542,0.020596629,-0.1462968141,0.5204554796,0.02538017,-0.3621297181,-0.1353999525,-0.1054494008,0.0756659955,0.0276800711,-0.0533361509,0.3394507468,0.1087685227,0.0285199098,0.079799287,0.130185023,0.0743425488,-0.0499666631,-0.1396474391,-0.2227846831,-0.2987947464,0.2747345567,-0.1234494448,-0.3272612691,-0.1521591097,0.0012162372,-0.2087304294,-0.0346010141,-0.0995942354,0.4414266348,-0.1579021513,0.1526206434,0.0880152136,0.1817054152,-0.1947121471,-0.3660129607,-0.1490098834,0.2008820623,-0.468621403,-0.3205323219,0.2491369843,-0.3165852427,0.4326317608,-0.0886029601,0.107325092,-0.1022875011,-0.1421988904,-0.2225233167,-0.2582617104,0.188044548,0.067876935,0.2651393414,0.0120807774,-0.0328292362,-0.0076496922,-0.2450187951,-0.129508391,-0.1909275204,-0.072525613,0.1784774959,-0.0103534712,-0.1338939667,-0.5229309797,0.0499758162,0.1910482049,0.0380627438,-0.002433571,-0.2078721821,-0.2314997613,-0.2085983604,0.117168434,0.5448202491,-0.0188138504,0.028379038,-0.0435673185,-0.0119555918,0.4996949434,0.3720760047,-0.0997387618,0.172167778,-0.2095858306,0.1426471472,0.1664265692,-0.5432876945,-0.2214052975,0.2418850362,-0.0492497012,0.0903555453,0.0329224057,-0.4252802134,0.2009977847,-0.130916357,0.1117296368,0.2759277225,0.0992452502,0.1265601665,-0.2411850691,-0.1267588139,0.2473011166,0.1046614796,0.1690177172,-0.1020453349,0.2604554296,0.1870794445,0.2613472939,-0.1208254769,0.3596661985,0.2470407784,-0.2625224292,-0.3072057366,0.1050221324,0.19859384,-0.3305431604,0.206600979,-0.2245477438,0.2911973596,0.0532628298,-0.0218927599,-0.3256198168,-0.2064074129,-0.2410363108,-0.2011760324,0.1774201095,-0.0429085456,0.3280696571,0.1672783643,0.1197206452,0.3532952666,-0.3063316047,-0.1096612364,-0.1067400351,0.1026701927,-0.2434271425,-0.0721904933,0.3241747916,0.2424546927,0.2653382123,-0.156836763,-0.1398927718,0.0729344636,-0.1470297724,0.4155639708,-0.1165600345,-0.1203745231,0.3694632649,-0.3810000718,0.4082131386,0.317717284,0.0645144135,-0.0745139942,0.3047270179,-0.2446958125,0.0516471826,0.0077799303,0.0869080424,-0.1271286458,0.331769824,-0.091515258,0.1778372526,-0.0359349363,0.5329710841,-0.0943983495,0.2545621097,-0.1527448297,-0.1970202029,0.027613027,0.4264924526,0.0807586312,-0.1145147458,0.171007663,-0.1288840175,-0.175828889,0.2758097649,0.1189039946,0.1102279648,0.1651006788,0.1731343567,0.1311655343,0.1335800737,-0.2187917084,0.2529250383,-0.2158294767,0.1718282849,0.3353054821,0.1661093533,-0.0029510632,-0.4320878088,-0.0286971219,0.2143987715,0.1778079122,-0.0685272738,-0.0997375175,-0.2467764765,-0.4081358314,-0.3134776056,-0.1185436398,0.0371320471,0.0172460489,-0.1968258172,0.114120625,-0.0407053009,-0.0076273112,-0.2748734355,-0.3112739325,-0.1994776577,-0.2850935757,-0.1725654304,0.0059298952,-0.4090853035,0.1231618375,0.1412646472,0.0594344474,0.3566991687,-0.33560583,-0.0827376172,-0.223473832,-0.0705547929,0.2513357699,-0.0651990846,-0.1545362324,0.0221690554,0.2006978989,-0.2227170467,-0.1932851523,0.0155675327,-0.1274445355,-0.1874835044,0.0146911833,-0.2334830463,-0.155720219,-0.1595274955,-0.6133767366,-0.4746182561,-0.4864419997,0.0599954873,-0.1413523406,0.1965738386,0.3935822845,-0.2014529854,0.1657384634,0.13731879,0.1436540931,-0.2028478831,-0.0262709111,0.1178618744,-0.3232525885,-0.454369545,0.117317833,0.0238614138,0.4150661826,0.1647234112,-0.3046405315,0.3316366673,-0.032270696,0.1291528493,0.0913803279,0.1418595463,0.4941840768,-0.0997108519,-0.0022173061,0.0292065702,0.1109181046,-0.0067351903,0.2292229086,0.4109459519,-0.2725373507,0.1508922577,0.1694056392,0.3010313809,0.3988194764,-0.2326139212,0.6556056738,0.3299936354,0.1660488546,-0.0785605311,-0.223159954,0.4240542054,-0.0242026802,0.050451912,0.3943394423,-0.1681078672,-0.560706377,-0.1106758937,-0.4160575271,-0.4069871306,-0.0655990541,-0.0058622886,0.1179788038,0.4066350162,0.0412172377,0.179729417,0.2586691678,-0.2891934812,0.1588014811,0.222110793,0.0307035279,0.2962247729,-0.0844785944,-0.1370988339,-0.7133892775,0.3157611787,0.2892406583,0.1681810617,-0.1471838206,0.3811462522,0.0132091008,-0.1939187348,0.540183723,-0.2146276981,0.1172232628,0.0385664403,0.0394678414,-0.082289353,0.0103465356,0.0670095608,0.1176479086,0.4223914742,0.0589513667,-0.4840069711,-0.0881331116,-0.0218050499,0.4767236412,0.1708550751,-0.2185003012,-0.0499941818,-0.5328354239,-0.2335794419,0.007856288,0.0180500951,0.3688515723,-0.0234136004,-0.007777845,0.178348586,0.0718784854,0.1867159754,0.1425793469,0.2800192833,0.2922777236,0.200650543,0.3297092319,0.0043779006,0.1921951771,0.3183093369,-0.148680225,0.1402457356,-0.0530247316,0.1164921597,0.1367502511,0.2045944482,0.0301326998,-0.0627524033,-0.1415639669,-0.197998628,0.146508038,0.0835249424,0.0963886976,0.1257433295,-0.5703757405,-0.3882588744,0.2796278596,-0.3202518225,0.1595812887,0.1544818729,0.3057815433,-0.339107275,0.2559328675,0.0795460567,0.7368466854,-0.0402690209,0.0766436979,0.0286077913,0.0771994665,0.2957970202,-0.3393850923,0.179179728,-0.2431397438,0.1196998507,-0.0036815894,0.077733703,-0.0203926973,0.1365586519,-0.3911959827,0.5633780956,0.0777336881,-0.0448268466,0.0642819181,0.3738422096,-0.1419715285,-0.3766852617,-0.4003916681,0.1819061935,-0.3179160058,0.3305777907,-0.208878845,-0.0886835158,0.1062344611,-0.1021086574,-0.3304612637,0.4157050848,-0.2829901278,0.0253621247,-0.0954853967,-0.1803955734,0.4980296493,0.24747926,0.3305479288,0.374456197,-0.2370951921,0.120631963,0.1032213718,-0.5300790071,0.2447067946,0.021626763,0.0547361709,-0.3400761485,-0.2175462395,0.0930889472,-0.0411833115,-0.0525409877,0.2623298168,-0.1425350606,-0.3336849809,0.0105105313,-0.3549546599,0.2626432478,-0.0046562646,0.0392499901,0.1276262254,-0.0834263116,-0.1788953841,0.1112684011,0.1640378982,-0.2192924768,-0.0932556242,0.311394304,0.1225598529,-0.3694102466,0.470108211,0.2693347335,-0.2435082346,-0.1901333928,-0.6724639535,-0.2965880632,-0.3600971103,-0.0081366608,-0.0068643484,0.0362955108,-0.1432041675,0.3669493794,-0.0289171077,-0.0073453444,0.2249272019,-0.3883952796,-0.2859397829,0.0696738884,-0.1597448587,0.0285757296,0.2387289405,0.2828913033,0.1914266944,0.1129984111,-0.2709113061,-0.0451796874,-0.3741313815,0.0734555945,0.2913359702,-0.1015430987,0.3328825235,0.1120014787,0.1091698781,0.2557425201,-0.2520971298,-0.2090447694,-0.1103292257,0.0847762004,0.1579771787,0.0817384645,0.3228664398,-0.3832597435,-0.1494952142,-0.028482357,0.126094833,0.3018396795,-0.0339369401,0.1260331422,0.2406696826,0.0628555119,-0.0867634267,0.1424496472,0.1167202741,0.0945004001,-0.0211362783,0.1423729509,0.067919977,-0.0810720176,-0.1436301917,-0.0144849047,0.1109489575,0.0499675982,0.1543722302,-0.0368774943,0.1027513221,-0.0054851673,-0.0454933383,0.0074303565,-0.0234520882,-0.0358009003,0.3045820594,0.2066614926,-0.2701682448,0.1760812104,-0.0184231084,0.1721590906,0.0045921775,0.0139468545,0.3261063695,-0.2998712361,0.1954528838,0.0973289311,0.3590956032,-0.1300854832,-0.1550516337,-0.0282243006,0.0612748973,0.0415270887,0.2245525867,-0.1463569552,0.0086984821,0.1348892301,-0.0893119946,-0.0116788819,0.2734456062,0.2953779697,-0.0654611513,-0.2115878016,0.0072166775,0.3011555672,0.2955462039,0.2288535535,0.1197468266,0.1959207058,0.1732964814,-0.200048998,-0.3275516331,0.3751668632,-0.0789166242,0.0393313244,-0.1925189644,-0.2186932266,-0.0818450674,0.1360602379,-0.1872272342,-0.0659432039,0.1962611973,0.2184791416,-0.2186942101,-0.4077454209,-0.0435331203,0.0160435289,0.1576144248,-0.2169872075,0.4991546273,0.2705328763,-0.054660134,-0.1632510126,0.314086169,0.4485455155,0.2007255554,-0.1284498572,0.0977757648,0.1225194037,-0.0104406327,-0.0364384986,0.4297582507,0.3452330232,0.1135912612,0.282918483,0.1092250496,-0.2205705345,-0.0985294208,0.0933237672,-0.0567233674,0.1250903606,0.451225847,0.0044774292,0.0749910474,-0.3351687193,0.0485384502,-0.480506897,0.4305048883,0.4933682084,-0.1941915303,-0.0126123652,-0.1042474657,0.0864552259,0.0925589874,0.3288391829,0.2744688094,0.4959403872,-0.4177182019,-0.3438930511,-0.3343045115,0.1863348931,-0.2957495749,0.1797835678,-0.2519392371,0.0594345629,0.143458575,0.1007816344,-0.1272498667,0.366692394,0.1858077347,0.2237451226,-0.1296077967,0.1150657088,-0.1630086899,0.0071013104,-0.006841823,-0.4851977825,-0.1454279423,-0.0727040321,0.0664691925,-0.2589695156,0.0846332908,0.0441818535,-0.1518297344,0.135022223,0.0279058758,0.6574344039,-0.2393392473,-0.1950959712,-0.2961331308,-0.2016223073,-0.1850347966,-0.0286590047,0.0989930332,0.3526166975,0.0663160384,0.1508650333,-0.1567121297,0.4380097985,0.1698628068,0.3453728259,-0.1365328282,0.1078100652,-0.3517187536,-0.1205077767,0.0466760062,0.506477654,0.0218654592,-0.1784384698,-0.2926490903,-0.5317479372,0.5479229093,-0.8124235868,-0.2576377988,0.0784190819,0.0957219377,-0.0114729004,-0.2659358084,-0.5267435908,0.2552785277,0.0984909907,-0.0782416686,-0.327549547,0.1157159433,-0.0942824036,-0.0646765679,-0.0526354015,0.3443565071,0.1893855482,-0.2799617648,-0.1954742223,-0.3156386018]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/233","title":"Fail to download c4 english corpus","comments":"I have the same problem as @prashant-kikani","body":"i run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?","comment_length":7,"text":"Fail to download c4 english corpus\ni run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?\nI have the same problem as @prashant-kikani","embeddings":[0.0115546789,0.0910863578,-0.0556439348,0.2902231514,0.148630634,0.1652969122,-0.0632363707,0.3174011707,-0.1166139469,0.0394295342,-0.1543145627,0.0087515097,0.0267580524,0.084500052,0.1310365945,-0.5472636223,-0.2235968858,0.1527146846,-0.131220445,-0.1412544996,-0.0423424728,0.4258191586,-0.1137883365,-0.1747365743,-0.0600037314,-0.1374861747,-0.1974728554,-0.0550695211,-0.206342712,-0.544832468,0.4454170763,0.0049821967,0.1893910617,0.3470121026,-0.0001094098,-0.0178526044,0.410225302,-0.1208095476,-0.2488911003,-0.39947474,0.1793028861,-0.3214279711,-0.0999809876,-0.3226758242,-0.2282853127,-0.1305258572,0.1668989211,-0.2270071805,0.0431619808,0.3165642321,0.2735976577,-0.0255144741,0.2802649736,-0.1330512613,-0.1831474304,-0.2402020991,-0.0908977464,0.2393375039,0.1667280197,0.2533928156,-0.0951283351,0.1529464126,0.0502758324,0.0134145711,-0.1094024777,-0.2209978402,0.0489708111,-0.2720611691,0.3356462419,0.0553029366,0.7832083702,-0.2817870378,-0.1618426442,0.1612834483,0.283929795,-0.212313965,0.2203885913,0.4282406271,-0.1392534822,0.0007200533,-0.3249097168,-0.0999147892,-0.181789875,0.2037830502,0.0209814962,0.0087318849,-0.0804654807,-0.0100580025,0.2173687667,0.1014782935,0.0793812573,-0.0789981186,-0.2777691185,0.2961800694,-0.2832439542,0.020596629,-0.1462968141,0.5204554796,0.02538017,-0.3621297181,-0.1353999525,-0.1054494008,0.0756659955,0.0276800711,-0.0533361509,0.3394507468,0.1087685227,0.0285199098,0.079799287,0.130185023,0.0743425488,-0.0499666631,-0.1396474391,-0.2227846831,-0.2987947464,0.2747345567,-0.1234494448,-0.3272612691,-0.1521591097,0.0012162372,-0.2087304294,-0.0346010141,-0.0995942354,0.4414266348,-0.1579021513,0.1526206434,0.0880152136,0.1817054152,-0.1947121471,-0.3660129607,-0.1490098834,0.2008820623,-0.468621403,-0.3205323219,0.2491369843,-0.3165852427,0.4326317608,-0.0886029601,0.107325092,-0.1022875011,-0.1421988904,-0.2225233167,-0.2582617104,0.188044548,0.067876935,0.2651393414,0.0120807774,-0.0328292362,-0.0076496922,-0.2450187951,-0.129508391,-0.1909275204,-0.072525613,0.1784774959,-0.0103534712,-0.1338939667,-0.5229309797,0.0499758162,0.1910482049,0.0380627438,-0.002433571,-0.2078721821,-0.2314997613,-0.2085983604,0.117168434,0.5448202491,-0.0188138504,0.028379038,-0.0435673185,-0.0119555918,0.4996949434,0.3720760047,-0.0997387618,0.172167778,-0.2095858306,0.1426471472,0.1664265692,-0.5432876945,-0.2214052975,0.2418850362,-0.0492497012,0.0903555453,0.0329224057,-0.4252802134,0.2009977847,-0.130916357,0.1117296368,0.2759277225,0.0992452502,0.1265601665,-0.2411850691,-0.1267588139,0.2473011166,0.1046614796,0.1690177172,-0.1020453349,0.2604554296,0.1870794445,0.2613472939,-0.1208254769,0.3596661985,0.2470407784,-0.2625224292,-0.3072057366,0.1050221324,0.19859384,-0.3305431604,0.206600979,-0.2245477438,0.2911973596,0.0532628298,-0.0218927599,-0.3256198168,-0.2064074129,-0.2410363108,-0.2011760324,0.1774201095,-0.0429085456,0.3280696571,0.1672783643,0.1197206452,0.3532952666,-0.3063316047,-0.1096612364,-0.1067400351,0.1026701927,-0.2434271425,-0.0721904933,0.3241747916,0.2424546927,0.2653382123,-0.156836763,-0.1398927718,0.0729344636,-0.1470297724,0.4155639708,-0.1165600345,-0.1203745231,0.3694632649,-0.3810000718,0.4082131386,0.317717284,0.0645144135,-0.0745139942,0.3047270179,-0.2446958125,0.0516471826,0.0077799303,0.0869080424,-0.1271286458,0.331769824,-0.091515258,0.1778372526,-0.0359349363,0.5329710841,-0.0943983495,0.2545621097,-0.1527448297,-0.1970202029,0.027613027,0.4264924526,0.0807586312,-0.1145147458,0.171007663,-0.1288840175,-0.175828889,0.2758097649,0.1189039946,0.1102279648,0.1651006788,0.1731343567,0.1311655343,0.1335800737,-0.2187917084,0.2529250383,-0.2158294767,0.1718282849,0.3353054821,0.1661093533,-0.0029510632,-0.4320878088,-0.0286971219,0.2143987715,0.1778079122,-0.0685272738,-0.0997375175,-0.2467764765,-0.4081358314,-0.3134776056,-0.1185436398,0.0371320471,0.0172460489,-0.1968258172,0.114120625,-0.0407053009,-0.0076273112,-0.2748734355,-0.3112739325,-0.1994776577,-0.2850935757,-0.1725654304,0.0059298952,-0.4090853035,0.1231618375,0.1412646472,0.0594344474,0.3566991687,-0.33560583,-0.0827376172,-0.223473832,-0.0705547929,0.2513357699,-0.0651990846,-0.1545362324,0.0221690554,0.2006978989,-0.2227170467,-0.1932851523,0.0155675327,-0.1274445355,-0.1874835044,0.0146911833,-0.2334830463,-0.155720219,-0.1595274955,-0.6133767366,-0.4746182561,-0.4864419997,0.0599954873,-0.1413523406,0.1965738386,0.3935822845,-0.2014529854,0.1657384634,0.13731879,0.1436540931,-0.2028478831,-0.0262709111,0.1178618744,-0.3232525885,-0.454369545,0.117317833,0.0238614138,0.4150661826,0.1647234112,-0.3046405315,0.3316366673,-0.032270696,0.1291528493,0.0913803279,0.1418595463,0.4941840768,-0.0997108519,-0.0022173061,0.0292065702,0.1109181046,-0.0067351903,0.2292229086,0.4109459519,-0.2725373507,0.1508922577,0.1694056392,0.3010313809,0.3988194764,-0.2326139212,0.6556056738,0.3299936354,0.1660488546,-0.0785605311,-0.223159954,0.4240542054,-0.0242026802,0.050451912,0.3943394423,-0.1681078672,-0.560706377,-0.1106758937,-0.4160575271,-0.4069871306,-0.0655990541,-0.0058622886,0.1179788038,0.4066350162,0.0412172377,0.179729417,0.2586691678,-0.2891934812,0.1588014811,0.222110793,0.0307035279,0.2962247729,-0.0844785944,-0.1370988339,-0.7133892775,0.3157611787,0.2892406583,0.1681810617,-0.1471838206,0.3811462522,0.0132091008,-0.1939187348,0.540183723,-0.2146276981,0.1172232628,0.0385664403,0.0394678414,-0.082289353,0.0103465356,0.0670095608,0.1176479086,0.4223914742,0.0589513667,-0.4840069711,-0.0881331116,-0.0218050499,0.4767236412,0.1708550751,-0.2185003012,-0.0499941818,-0.5328354239,-0.2335794419,0.007856288,0.0180500951,0.3688515723,-0.0234136004,-0.007777845,0.178348586,0.0718784854,0.1867159754,0.1425793469,0.2800192833,0.2922777236,0.200650543,0.3297092319,0.0043779006,0.1921951771,0.3183093369,-0.148680225,0.1402457356,-0.0530247316,0.1164921597,0.1367502511,0.2045944482,0.0301326998,-0.0627524033,-0.1415639669,-0.197998628,0.146508038,0.0835249424,0.0963886976,0.1257433295,-0.5703757405,-0.3882588744,0.2796278596,-0.3202518225,0.1595812887,0.1544818729,0.3057815433,-0.339107275,0.2559328675,0.0795460567,0.7368466854,-0.0402690209,0.0766436979,0.0286077913,0.0771994665,0.2957970202,-0.3393850923,0.179179728,-0.2431397438,0.1196998507,-0.0036815894,0.077733703,-0.0203926973,0.1365586519,-0.3911959827,0.5633780956,0.0777336881,-0.0448268466,0.0642819181,0.3738422096,-0.1419715285,-0.3766852617,-0.4003916681,0.1819061935,-0.3179160058,0.3305777907,-0.208878845,-0.0886835158,0.1062344611,-0.1021086574,-0.3304612637,0.4157050848,-0.2829901278,0.0253621247,-0.0954853967,-0.1803955734,0.4980296493,0.24747926,0.3305479288,0.374456197,-0.2370951921,0.120631963,0.1032213718,-0.5300790071,0.2447067946,0.021626763,0.0547361709,-0.3400761485,-0.2175462395,0.0930889472,-0.0411833115,-0.0525409877,0.2623298168,-0.1425350606,-0.3336849809,0.0105105313,-0.3549546599,0.2626432478,-0.0046562646,0.0392499901,0.1276262254,-0.0834263116,-0.1788953841,0.1112684011,0.1640378982,-0.2192924768,-0.0932556242,0.311394304,0.1225598529,-0.3694102466,0.470108211,0.2693347335,-0.2435082346,-0.1901333928,-0.6724639535,-0.2965880632,-0.3600971103,-0.0081366608,-0.0068643484,0.0362955108,-0.1432041675,0.3669493794,-0.0289171077,-0.0073453444,0.2249272019,-0.3883952796,-0.2859397829,0.0696738884,-0.1597448587,0.0285757296,0.2387289405,0.2828913033,0.1914266944,0.1129984111,-0.2709113061,-0.0451796874,-0.3741313815,0.0734555945,0.2913359702,-0.1015430987,0.3328825235,0.1120014787,0.1091698781,0.2557425201,-0.2520971298,-0.2090447694,-0.1103292257,0.0847762004,0.1579771787,0.0817384645,0.3228664398,-0.3832597435,-0.1494952142,-0.028482357,0.126094833,0.3018396795,-0.0339369401,0.1260331422,0.2406696826,0.0628555119,-0.0867634267,0.1424496472,0.1167202741,0.0945004001,-0.0211362783,0.1423729509,0.067919977,-0.0810720176,-0.1436301917,-0.0144849047,0.1109489575,0.0499675982,0.1543722302,-0.0368774943,0.1027513221,-0.0054851673,-0.0454933383,0.0074303565,-0.0234520882,-0.0358009003,0.3045820594,0.2066614926,-0.2701682448,0.1760812104,-0.0184231084,0.1721590906,0.0045921775,0.0139468545,0.3261063695,-0.2998712361,0.1954528838,0.0973289311,0.3590956032,-0.1300854832,-0.1550516337,-0.0282243006,0.0612748973,0.0415270887,0.2245525867,-0.1463569552,0.0086984821,0.1348892301,-0.0893119946,-0.0116788819,0.2734456062,0.2953779697,-0.0654611513,-0.2115878016,0.0072166775,0.3011555672,0.2955462039,0.2288535535,0.1197468266,0.1959207058,0.1732964814,-0.200048998,-0.3275516331,0.3751668632,-0.0789166242,0.0393313244,-0.1925189644,-0.2186932266,-0.0818450674,0.1360602379,-0.1872272342,-0.0659432039,0.1962611973,0.2184791416,-0.2186942101,-0.4077454209,-0.0435331203,0.0160435289,0.1576144248,-0.2169872075,0.4991546273,0.2705328763,-0.054660134,-0.1632510126,0.314086169,0.4485455155,0.2007255554,-0.1284498572,0.0977757648,0.1225194037,-0.0104406327,-0.0364384986,0.4297582507,0.3452330232,0.1135912612,0.282918483,0.1092250496,-0.2205705345,-0.0985294208,0.0933237672,-0.0567233674,0.1250903606,0.451225847,0.0044774292,0.0749910474,-0.3351687193,0.0485384502,-0.480506897,0.4305048883,0.4933682084,-0.1941915303,-0.0126123652,-0.1042474657,0.0864552259,0.0925589874,0.3288391829,0.2744688094,0.4959403872,-0.4177182019,-0.3438930511,-0.3343045115,0.1863348931,-0.2957495749,0.1797835678,-0.2519392371,0.0594345629,0.143458575,0.1007816344,-0.1272498667,0.366692394,0.1858077347,0.2237451226,-0.1296077967,0.1150657088,-0.1630086899,0.0071013104,-0.006841823,-0.4851977825,-0.1454279423,-0.0727040321,0.0664691925,-0.2589695156,0.0846332908,0.0441818535,-0.1518297344,0.135022223,0.0279058758,0.6574344039,-0.2393392473,-0.1950959712,-0.2961331308,-0.2016223073,-0.1850347966,-0.0286590047,0.0989930332,0.3526166975,0.0663160384,0.1508650333,-0.1567121297,0.4380097985,0.1698628068,0.3453728259,-0.1365328282,0.1078100652,-0.3517187536,-0.1205077767,0.0466760062,0.506477654,0.0218654592,-0.1784384698,-0.2926490903,-0.5317479372,0.5479229093,-0.8124235868,-0.2576377988,0.0784190819,0.0957219377,-0.0114729004,-0.2659358084,-0.5267435908,0.2552785277,0.0984909907,-0.0782416686,-0.327549547,0.1157159433,-0.0942824036,-0.0646765679,-0.0526354015,0.3443565071,0.1893855482,-0.2799617648,-0.1954742223,-0.3156386018]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/233","title":"Fail to download c4 english corpus","comments":"Looks like a bug in the dataset script, can you open an issue ?","body":"i run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?","comment_length":14,"text":"Fail to download c4 english corpus\ni run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?\nLooks like a bug in the dataset script, can you open an issue ?","embeddings":[0.0115546789,0.0910863578,-0.0556439348,0.2902231514,0.148630634,0.1652969122,-0.0632363707,0.3174011707,-0.1166139469,0.0394295342,-0.1543145627,0.0087515097,0.0267580524,0.084500052,0.1310365945,-0.5472636223,-0.2235968858,0.1527146846,-0.131220445,-0.1412544996,-0.0423424728,0.4258191586,-0.1137883365,-0.1747365743,-0.0600037314,-0.1374861747,-0.1974728554,-0.0550695211,-0.206342712,-0.544832468,0.4454170763,0.0049821967,0.1893910617,0.3470121026,-0.0001094098,-0.0178526044,0.410225302,-0.1208095476,-0.2488911003,-0.39947474,0.1793028861,-0.3214279711,-0.0999809876,-0.3226758242,-0.2282853127,-0.1305258572,0.1668989211,-0.2270071805,0.0431619808,0.3165642321,0.2735976577,-0.0255144741,0.2802649736,-0.1330512613,-0.1831474304,-0.2402020991,-0.0908977464,0.2393375039,0.1667280197,0.2533928156,-0.0951283351,0.1529464126,0.0502758324,0.0134145711,-0.1094024777,-0.2209978402,0.0489708111,-0.2720611691,0.3356462419,0.0553029366,0.7832083702,-0.2817870378,-0.1618426442,0.1612834483,0.283929795,-0.212313965,0.2203885913,0.4282406271,-0.1392534822,0.0007200533,-0.3249097168,-0.0999147892,-0.181789875,0.2037830502,0.0209814962,0.0087318849,-0.0804654807,-0.0100580025,0.2173687667,0.1014782935,0.0793812573,-0.0789981186,-0.2777691185,0.2961800694,-0.2832439542,0.020596629,-0.1462968141,0.5204554796,0.02538017,-0.3621297181,-0.1353999525,-0.1054494008,0.0756659955,0.0276800711,-0.0533361509,0.3394507468,0.1087685227,0.0285199098,0.079799287,0.130185023,0.0743425488,-0.0499666631,-0.1396474391,-0.2227846831,-0.2987947464,0.2747345567,-0.1234494448,-0.3272612691,-0.1521591097,0.0012162372,-0.2087304294,-0.0346010141,-0.0995942354,0.4414266348,-0.1579021513,0.1526206434,0.0880152136,0.1817054152,-0.1947121471,-0.3660129607,-0.1490098834,0.2008820623,-0.468621403,-0.3205323219,0.2491369843,-0.3165852427,0.4326317608,-0.0886029601,0.107325092,-0.1022875011,-0.1421988904,-0.2225233167,-0.2582617104,0.188044548,0.067876935,0.2651393414,0.0120807774,-0.0328292362,-0.0076496922,-0.2450187951,-0.129508391,-0.1909275204,-0.072525613,0.1784774959,-0.0103534712,-0.1338939667,-0.5229309797,0.0499758162,0.1910482049,0.0380627438,-0.002433571,-0.2078721821,-0.2314997613,-0.2085983604,0.117168434,0.5448202491,-0.0188138504,0.028379038,-0.0435673185,-0.0119555918,0.4996949434,0.3720760047,-0.0997387618,0.172167778,-0.2095858306,0.1426471472,0.1664265692,-0.5432876945,-0.2214052975,0.2418850362,-0.0492497012,0.0903555453,0.0329224057,-0.4252802134,0.2009977847,-0.130916357,0.1117296368,0.2759277225,0.0992452502,0.1265601665,-0.2411850691,-0.1267588139,0.2473011166,0.1046614796,0.1690177172,-0.1020453349,0.2604554296,0.1870794445,0.2613472939,-0.1208254769,0.3596661985,0.2470407784,-0.2625224292,-0.3072057366,0.1050221324,0.19859384,-0.3305431604,0.206600979,-0.2245477438,0.2911973596,0.0532628298,-0.0218927599,-0.3256198168,-0.2064074129,-0.2410363108,-0.2011760324,0.1774201095,-0.0429085456,0.3280696571,0.1672783643,0.1197206452,0.3532952666,-0.3063316047,-0.1096612364,-0.1067400351,0.1026701927,-0.2434271425,-0.0721904933,0.3241747916,0.2424546927,0.2653382123,-0.156836763,-0.1398927718,0.0729344636,-0.1470297724,0.4155639708,-0.1165600345,-0.1203745231,0.3694632649,-0.3810000718,0.4082131386,0.317717284,0.0645144135,-0.0745139942,0.3047270179,-0.2446958125,0.0516471826,0.0077799303,0.0869080424,-0.1271286458,0.331769824,-0.091515258,0.1778372526,-0.0359349363,0.5329710841,-0.0943983495,0.2545621097,-0.1527448297,-0.1970202029,0.027613027,0.4264924526,0.0807586312,-0.1145147458,0.171007663,-0.1288840175,-0.175828889,0.2758097649,0.1189039946,0.1102279648,0.1651006788,0.1731343567,0.1311655343,0.1335800737,-0.2187917084,0.2529250383,-0.2158294767,0.1718282849,0.3353054821,0.1661093533,-0.0029510632,-0.4320878088,-0.0286971219,0.2143987715,0.1778079122,-0.0685272738,-0.0997375175,-0.2467764765,-0.4081358314,-0.3134776056,-0.1185436398,0.0371320471,0.0172460489,-0.1968258172,0.114120625,-0.0407053009,-0.0076273112,-0.2748734355,-0.3112739325,-0.1994776577,-0.2850935757,-0.1725654304,0.0059298952,-0.4090853035,0.1231618375,0.1412646472,0.0594344474,0.3566991687,-0.33560583,-0.0827376172,-0.223473832,-0.0705547929,0.2513357699,-0.0651990846,-0.1545362324,0.0221690554,0.2006978989,-0.2227170467,-0.1932851523,0.0155675327,-0.1274445355,-0.1874835044,0.0146911833,-0.2334830463,-0.155720219,-0.1595274955,-0.6133767366,-0.4746182561,-0.4864419997,0.0599954873,-0.1413523406,0.1965738386,0.3935822845,-0.2014529854,0.1657384634,0.13731879,0.1436540931,-0.2028478831,-0.0262709111,0.1178618744,-0.3232525885,-0.454369545,0.117317833,0.0238614138,0.4150661826,0.1647234112,-0.3046405315,0.3316366673,-0.032270696,0.1291528493,0.0913803279,0.1418595463,0.4941840768,-0.0997108519,-0.0022173061,0.0292065702,0.1109181046,-0.0067351903,0.2292229086,0.4109459519,-0.2725373507,0.1508922577,0.1694056392,0.3010313809,0.3988194764,-0.2326139212,0.6556056738,0.3299936354,0.1660488546,-0.0785605311,-0.223159954,0.4240542054,-0.0242026802,0.050451912,0.3943394423,-0.1681078672,-0.560706377,-0.1106758937,-0.4160575271,-0.4069871306,-0.0655990541,-0.0058622886,0.1179788038,0.4066350162,0.0412172377,0.179729417,0.2586691678,-0.2891934812,0.1588014811,0.222110793,0.0307035279,0.2962247729,-0.0844785944,-0.1370988339,-0.7133892775,0.3157611787,0.2892406583,0.1681810617,-0.1471838206,0.3811462522,0.0132091008,-0.1939187348,0.540183723,-0.2146276981,0.1172232628,0.0385664403,0.0394678414,-0.082289353,0.0103465356,0.0670095608,0.1176479086,0.4223914742,0.0589513667,-0.4840069711,-0.0881331116,-0.0218050499,0.4767236412,0.1708550751,-0.2185003012,-0.0499941818,-0.5328354239,-0.2335794419,0.007856288,0.0180500951,0.3688515723,-0.0234136004,-0.007777845,0.178348586,0.0718784854,0.1867159754,0.1425793469,0.2800192833,0.2922777236,0.200650543,0.3297092319,0.0043779006,0.1921951771,0.3183093369,-0.148680225,0.1402457356,-0.0530247316,0.1164921597,0.1367502511,0.2045944482,0.0301326998,-0.0627524033,-0.1415639669,-0.197998628,0.146508038,0.0835249424,0.0963886976,0.1257433295,-0.5703757405,-0.3882588744,0.2796278596,-0.3202518225,0.1595812887,0.1544818729,0.3057815433,-0.339107275,0.2559328675,0.0795460567,0.7368466854,-0.0402690209,0.0766436979,0.0286077913,0.0771994665,0.2957970202,-0.3393850923,0.179179728,-0.2431397438,0.1196998507,-0.0036815894,0.077733703,-0.0203926973,0.1365586519,-0.3911959827,0.5633780956,0.0777336881,-0.0448268466,0.0642819181,0.3738422096,-0.1419715285,-0.3766852617,-0.4003916681,0.1819061935,-0.3179160058,0.3305777907,-0.208878845,-0.0886835158,0.1062344611,-0.1021086574,-0.3304612637,0.4157050848,-0.2829901278,0.0253621247,-0.0954853967,-0.1803955734,0.4980296493,0.24747926,0.3305479288,0.374456197,-0.2370951921,0.120631963,0.1032213718,-0.5300790071,0.2447067946,0.021626763,0.0547361709,-0.3400761485,-0.2175462395,0.0930889472,-0.0411833115,-0.0525409877,0.2623298168,-0.1425350606,-0.3336849809,0.0105105313,-0.3549546599,0.2626432478,-0.0046562646,0.0392499901,0.1276262254,-0.0834263116,-0.1788953841,0.1112684011,0.1640378982,-0.2192924768,-0.0932556242,0.311394304,0.1225598529,-0.3694102466,0.470108211,0.2693347335,-0.2435082346,-0.1901333928,-0.6724639535,-0.2965880632,-0.3600971103,-0.0081366608,-0.0068643484,0.0362955108,-0.1432041675,0.3669493794,-0.0289171077,-0.0073453444,0.2249272019,-0.3883952796,-0.2859397829,0.0696738884,-0.1597448587,0.0285757296,0.2387289405,0.2828913033,0.1914266944,0.1129984111,-0.2709113061,-0.0451796874,-0.3741313815,0.0734555945,0.2913359702,-0.1015430987,0.3328825235,0.1120014787,0.1091698781,0.2557425201,-0.2520971298,-0.2090447694,-0.1103292257,0.0847762004,0.1579771787,0.0817384645,0.3228664398,-0.3832597435,-0.1494952142,-0.028482357,0.126094833,0.3018396795,-0.0339369401,0.1260331422,0.2406696826,0.0628555119,-0.0867634267,0.1424496472,0.1167202741,0.0945004001,-0.0211362783,0.1423729509,0.067919977,-0.0810720176,-0.1436301917,-0.0144849047,0.1109489575,0.0499675982,0.1543722302,-0.0368774943,0.1027513221,-0.0054851673,-0.0454933383,0.0074303565,-0.0234520882,-0.0358009003,0.3045820594,0.2066614926,-0.2701682448,0.1760812104,-0.0184231084,0.1721590906,0.0045921775,0.0139468545,0.3261063695,-0.2998712361,0.1954528838,0.0973289311,0.3590956032,-0.1300854832,-0.1550516337,-0.0282243006,0.0612748973,0.0415270887,0.2245525867,-0.1463569552,0.0086984821,0.1348892301,-0.0893119946,-0.0116788819,0.2734456062,0.2953779697,-0.0654611513,-0.2115878016,0.0072166775,0.3011555672,0.2955462039,0.2288535535,0.1197468266,0.1959207058,0.1732964814,-0.200048998,-0.3275516331,0.3751668632,-0.0789166242,0.0393313244,-0.1925189644,-0.2186932266,-0.0818450674,0.1360602379,-0.1872272342,-0.0659432039,0.1962611973,0.2184791416,-0.2186942101,-0.4077454209,-0.0435331203,0.0160435289,0.1576144248,-0.2169872075,0.4991546273,0.2705328763,-0.054660134,-0.1632510126,0.314086169,0.4485455155,0.2007255554,-0.1284498572,0.0977757648,0.1225194037,-0.0104406327,-0.0364384986,0.4297582507,0.3452330232,0.1135912612,0.282918483,0.1092250496,-0.2205705345,-0.0985294208,0.0933237672,-0.0567233674,0.1250903606,0.451225847,0.0044774292,0.0749910474,-0.3351687193,0.0485384502,-0.480506897,0.4305048883,0.4933682084,-0.1941915303,-0.0126123652,-0.1042474657,0.0864552259,0.0925589874,0.3288391829,0.2744688094,0.4959403872,-0.4177182019,-0.3438930511,-0.3343045115,0.1863348931,-0.2957495749,0.1797835678,-0.2519392371,0.0594345629,0.143458575,0.1007816344,-0.1272498667,0.366692394,0.1858077347,0.2237451226,-0.1296077967,0.1150657088,-0.1630086899,0.0071013104,-0.006841823,-0.4851977825,-0.1454279423,-0.0727040321,0.0664691925,-0.2589695156,0.0846332908,0.0441818535,-0.1518297344,0.135022223,0.0279058758,0.6574344039,-0.2393392473,-0.1950959712,-0.2961331308,-0.2016223073,-0.1850347966,-0.0286590047,0.0989930332,0.3526166975,0.0663160384,0.1508650333,-0.1567121297,0.4380097985,0.1698628068,0.3453728259,-0.1365328282,0.1078100652,-0.3517187536,-0.1205077767,0.0466760062,0.506477654,0.0218654592,-0.1784384698,-0.2926490903,-0.5317479372,0.5479229093,-0.8124235868,-0.2576377988,0.0784190819,0.0957219377,-0.0114729004,-0.2659358084,-0.5267435908,0.2552785277,0.0984909907,-0.0782416686,-0.327549547,0.1157159433,-0.0942824036,-0.0646765679,-0.0526354015,0.3443565071,0.1893855482,-0.2799617648,-0.1954742223,-0.3156386018]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/233","title":"Fail to download c4 english corpus","comments":"I see the same issue as @prashant-kikani. I'm using `datasets` version 1.2.0 to download C4.","body":"i run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?","comment_length":15,"text":"Fail to download c4 english corpus\ni run following code to download c4 English corpus.\r\n\r\n```\r\ndataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner'\r\n, data_dir='\/mypath')\r\n```\r\n\r\nand i met failure as follows\r\n\r\n```\r\nDownloading and preparing dataset c4\/en (download: Unknown size, generated: Unknown size, total: Unknown size) to \/home\/adam\/.cache\/huggingface\/datasets\/c4\/en\/2.3.0...\r\nTraceback (most recent call last):\r\n File \"download_corpus.py\", line 38, in <module>\r\n , data_dir='\/home\/adam\/data\/corpus\/en\/c4')\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 520, in load_dataset\r\n save_infos=save_infos,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 420, in download_and_prepare\r\n dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 816, in _download_and_prepare\r\n dl_manager, verify_infos=False, pipeline=pipeline,\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 457, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/adam\/anaconda3\/envs\/adam\/lib\/python3.7\/site-packages\/nlp\/datasets\/c4\/f545de9f63300d8d02a6795e2eb34e140c47e62a803f572ac5599e170ee66ecc\/c4.py\", line 175, in _split_generators\r\n dl_manager.download_checksums(_CHECKSUMS_URL)\r\nAttributeError: 'DownloadManager' object has no attribute 'download_checksums\r\n\r\n```\r\ncan i get any advice?\nI see the same issue as @prashant-kikani. I'm using `datasets` version 1.2.0 to download C4.","embeddings":[0.0115546789,0.0910863578,-0.0556439348,0.2902231514,0.148630634,0.1652969122,-0.0632363707,0.3174011707,-0.1166139469,0.0394295342,-0.1543145627,0.0087515097,0.0267580524,0.084500052,0.1310365945,-0.5472636223,-0.2235968858,0.1527146846,-0.131220445,-0.1412544996,-0.0423424728,0.4258191586,-0.1137883365,-0.1747365743,-0.0600037314,-0.1374861747,-0.1974728554,-0.0550695211,-0.206342712,-0.544832468,0.4454170763,0.0049821967,0.1893910617,0.3470121026,-0.0001094098,-0.0178526044,0.410225302,-0.1208095476,-0.2488911003,-0.39947474,0.1793028861,-0.3214279711,-0.0999809876,-0.3226758242,-0.2282853127,-0.1305258572,0.1668989211,-0.2270071805,0.0431619808,0.3165642321,0.2735976577,-0.0255144741,0.2802649736,-0.1330512613,-0.1831474304,-0.2402020991,-0.0908977464,0.2393375039,0.1667280197,0.2533928156,-0.0951283351,0.1529464126,0.0502758324,0.0134145711,-0.1094024777,-0.2209978402,0.0489708111,-0.2720611691,0.3356462419,0.0553029366,0.7832083702,-0.2817870378,-0.1618426442,0.1612834483,0.283929795,-0.212313965,0.2203885913,0.4282406271,-0.1392534822,0.0007200533,-0.3249097168,-0.0999147892,-0.181789875,0.2037830502,0.0209814962,0.0087318849,-0.0804654807,-0.0100580025,0.2173687667,0.1014782935,0.0793812573,-0.0789981186,-0.2777691185,0.2961800694,-0.2832439542,0.020596629,-0.1462968141,0.5204554796,0.02538017,-0.3621297181,-0.1353999525,-0.1054494008,0.0756659955,0.0276800711,-0.0533361509,0.3394507468,0.1087685227,0.0285199098,0.079799287,0.130185023,0.0743425488,-0.0499666631,-0.1396474391,-0.2227846831,-0.2987947464,0.2747345567,-0.1234494448,-0.3272612691,-0.1521591097,0.0012162372,-0.2087304294,-0.0346010141,-0.0995942354,0.4414266348,-0.1579021513,0.1526206434,0.0880152136,0.1817054152,-0.1947121471,-0.3660129607,-0.1490098834,0.2008820623,-0.468621403,-0.3205323219,0.2491369843,-0.3165852427,0.4326317608,-0.0886029601,0.107325092,-0.1022875011,-0.1421988904,-0.2225233167,-0.2582617104,0.188044548,0.067876935,0.2651393414,0.0120807774,-0.0328292362,-0.0076496922,-0.2450187951,-0.129508391,-0.1909275204,-0.072525613,0.1784774959,-0.0103534712,-0.1338939667,-0.5229309797,0.0499758162,0.1910482049,0.0380627438,-0.002433571,-0.2078721821,-0.2314997613,-0.2085983604,0.117168434,0.5448202491,-0.0188138504,0.028379038,-0.0435673185,-0.0119555918,0.4996949434,0.3720760047,-0.0997387618,0.172167778,-0.2095858306,0.1426471472,0.1664265692,-0.5432876945,-0.2214052975,0.2418850362,-0.0492497012,0.0903555453,0.0329224057,-0.4252802134,0.2009977847,-0.130916357,0.1117296368,0.2759277225,0.0992452502,0.1265601665,-0.2411850691,-0.1267588139,0.2473011166,0.1046614796,0.1690177172,-0.1020453349,0.2604554296,0.1870794445,0.2613472939,-0.1208254769,0.3596661985,0.2470407784,-0.2625224292,-0.3072057366,0.1050221324,0.19859384,-0.3305431604,0.206600979,-0.2245477438,0.2911973596,0.0532628298,-0.0218927599,-0.3256198168,-0.2064074129,-0.2410363108,-0.2011760324,0.1774201095,-0.0429085456,0.3280696571,0.1672783643,0.1197206452,0.3532952666,-0.3063316047,-0.1096612364,-0.1067400351,0.1026701927,-0.2434271425,-0.0721904933,0.3241747916,0.2424546927,0.2653382123,-0.156836763,-0.1398927718,0.0729344636,-0.1470297724,0.4155639708,-0.1165600345,-0.1203745231,0.3694632649,-0.3810000718,0.4082131386,0.317717284,0.0645144135,-0.0745139942,0.3047270179,-0.2446958125,0.0516471826,0.0077799303,0.0869080424,-0.1271286458,0.331769824,-0.091515258,0.1778372526,-0.0359349363,0.5329710841,-0.0943983495,0.2545621097,-0.1527448297,-0.1970202029,0.027613027,0.4264924526,0.0807586312,-0.1145147458,0.171007663,-0.1288840175,-0.175828889,0.2758097649,0.1189039946,0.1102279648,0.1651006788,0.1731343567,0.1311655343,0.1335800737,-0.2187917084,0.2529250383,-0.2158294767,0.1718282849,0.3353054821,0.1661093533,-0.0029510632,-0.4320878088,-0.0286971219,0.2143987715,0.1778079122,-0.0685272738,-0.0997375175,-0.2467764765,-0.4081358314,-0.3134776056,-0.1185436398,0.0371320471,0.0172460489,-0.1968258172,0.114120625,-0.0407053009,-0.0076273112,-0.2748734355,-0.3112739325,-0.1994776577,-0.2850935757,-0.1725654304,0.0059298952,-0.4090853035,0.1231618375,0.1412646472,0.0594344474,0.3566991687,-0.33560583,-0.0827376172,-0.223473832,-0.0705547929,0.2513357699,-0.0651990846,-0.1545362324,0.0221690554,0.2006978989,-0.2227170467,-0.1932851523,0.0155675327,-0.1274445355,-0.1874835044,0.0146911833,-0.2334830463,-0.155720219,-0.1595274955,-0.6133767366,-0.4746182561,-0.4864419997,0.0599954873,-0.1413523406,0.1965738386,0.3935822845,-0.2014529854,0.1657384634,0.13731879,0.1436540931,-0.2028478831,-0.0262709111,0.1178618744,-0.3232525885,-0.454369545,0.117317833,0.0238614138,0.4150661826,0.1647234112,-0.3046405315,0.3316366673,-0.032270696,0.1291528493,0.0913803279,0.1418595463,0.4941840768,-0.0997108519,-0.0022173061,0.0292065702,0.1109181046,-0.0067351903,0.2292229086,0.4109459519,-0.2725373507,0.1508922577,0.1694056392,0.3010313809,0.3988194764,-0.2326139212,0.6556056738,0.3299936354,0.1660488546,-0.0785605311,-0.223159954,0.4240542054,-0.0242026802,0.050451912,0.3943394423,-0.1681078672,-0.560706377,-0.1106758937,-0.4160575271,-0.4069871306,-0.0655990541,-0.0058622886,0.1179788038,0.4066350162,0.0412172377,0.179729417,0.2586691678,-0.2891934812,0.1588014811,0.222110793,0.0307035279,0.2962247729,-0.0844785944,-0.1370988339,-0.7133892775,0.3157611787,0.2892406583,0.1681810617,-0.1471838206,0.3811462522,0.0132091008,-0.1939187348,0.540183723,-0.2146276981,0.1172232628,0.0385664403,0.0394678414,-0.082289353,0.0103465356,0.0670095608,0.1176479086,0.4223914742,0.0589513667,-0.4840069711,-0.0881331116,-0.0218050499,0.4767236412,0.1708550751,-0.2185003012,-0.0499941818,-0.5328354239,-0.2335794419,0.007856288,0.0180500951,0.3688515723,-0.0234136004,-0.007777845,0.178348586,0.0718784854,0.1867159754,0.1425793469,0.2800192833,0.2922777236,0.200650543,0.3297092319,0.0043779006,0.1921951771,0.3183093369,-0.148680225,0.1402457356,-0.0530247316,0.1164921597,0.1367502511,0.2045944482,0.0301326998,-0.0627524033,-0.1415639669,-0.197998628,0.146508038,0.0835249424,0.0963886976,0.1257433295,-0.5703757405,-0.3882588744,0.2796278596,-0.3202518225,0.1595812887,0.1544818729,0.3057815433,-0.339107275,0.2559328675,0.0795460567,0.7368466854,-0.0402690209,0.0766436979,0.0286077913,0.0771994665,0.2957970202,-0.3393850923,0.179179728,-0.2431397438,0.1196998507,-0.0036815894,0.077733703,-0.0203926973,0.1365586519,-0.3911959827,0.5633780956,0.0777336881,-0.0448268466,0.0642819181,0.3738422096,-0.1419715285,-0.3766852617,-0.4003916681,0.1819061935,-0.3179160058,0.3305777907,-0.208878845,-0.0886835158,0.1062344611,-0.1021086574,-0.3304612637,0.4157050848,-0.2829901278,0.0253621247,-0.0954853967,-0.1803955734,0.4980296493,0.24747926,0.3305479288,0.374456197,-0.2370951921,0.120631963,0.1032213718,-0.5300790071,0.2447067946,0.021626763,0.0547361709,-0.3400761485,-0.2175462395,0.0930889472,-0.0411833115,-0.0525409877,0.2623298168,-0.1425350606,-0.3336849809,0.0105105313,-0.3549546599,0.2626432478,-0.0046562646,0.0392499901,0.1276262254,-0.0834263116,-0.1788953841,0.1112684011,0.1640378982,-0.2192924768,-0.0932556242,0.311394304,0.1225598529,-0.3694102466,0.470108211,0.2693347335,-0.2435082346,-0.1901333928,-0.6724639535,-0.2965880632,-0.3600971103,-0.0081366608,-0.0068643484,0.0362955108,-0.1432041675,0.3669493794,-0.0289171077,-0.0073453444,0.2249272019,-0.3883952796,-0.2859397829,0.0696738884,-0.1597448587,0.0285757296,0.2387289405,0.2828913033,0.1914266944,0.1129984111,-0.2709113061,-0.0451796874,-0.3741313815,0.0734555945,0.2913359702,-0.1015430987,0.3328825235,0.1120014787,0.1091698781,0.2557425201,-0.2520971298,-0.2090447694,-0.1103292257,0.0847762004,0.1579771787,0.0817384645,0.3228664398,-0.3832597435,-0.1494952142,-0.028482357,0.126094833,0.3018396795,-0.0339369401,0.1260331422,0.2406696826,0.0628555119,-0.0867634267,0.1424496472,0.1167202741,0.0945004001,-0.0211362783,0.1423729509,0.067919977,-0.0810720176,-0.1436301917,-0.0144849047,0.1109489575,0.0499675982,0.1543722302,-0.0368774943,0.1027513221,-0.0054851673,-0.0454933383,0.0074303565,-0.0234520882,-0.0358009003,0.3045820594,0.2066614926,-0.2701682448,0.1760812104,-0.0184231084,0.1721590906,0.0045921775,0.0139468545,0.3261063695,-0.2998712361,0.1954528838,0.0973289311,0.3590956032,-0.1300854832,-0.1550516337,-0.0282243006,0.0612748973,0.0415270887,0.2245525867,-0.1463569552,0.0086984821,0.1348892301,-0.0893119946,-0.0116788819,0.2734456062,0.2953779697,-0.0654611513,-0.2115878016,0.0072166775,0.3011555672,0.2955462039,0.2288535535,0.1197468266,0.1959207058,0.1732964814,-0.200048998,-0.3275516331,0.3751668632,-0.0789166242,0.0393313244,-0.1925189644,-0.2186932266,-0.0818450674,0.1360602379,-0.1872272342,-0.0659432039,0.1962611973,0.2184791416,-0.2186942101,-0.4077454209,-0.0435331203,0.0160435289,0.1576144248,-0.2169872075,0.4991546273,0.2705328763,-0.054660134,-0.1632510126,0.314086169,0.4485455155,0.2007255554,-0.1284498572,0.0977757648,0.1225194037,-0.0104406327,-0.0364384986,0.4297582507,0.3452330232,0.1135912612,0.282918483,0.1092250496,-0.2205705345,-0.0985294208,0.0933237672,-0.0567233674,0.1250903606,0.451225847,0.0044774292,0.0749910474,-0.3351687193,0.0485384502,-0.480506897,0.4305048883,0.4933682084,-0.1941915303,-0.0126123652,-0.1042474657,0.0864552259,0.0925589874,0.3288391829,0.2744688094,0.4959403872,-0.4177182019,-0.3438930511,-0.3343045115,0.1863348931,-0.2957495749,0.1797835678,-0.2519392371,0.0594345629,0.143458575,0.1007816344,-0.1272498667,0.366692394,0.1858077347,0.2237451226,-0.1296077967,0.1150657088,-0.1630086899,0.0071013104,-0.006841823,-0.4851977825,-0.1454279423,-0.0727040321,0.0664691925,-0.2589695156,0.0846332908,0.0441818535,-0.1518297344,0.135022223,0.0279058758,0.6574344039,-0.2393392473,-0.1950959712,-0.2961331308,-0.2016223073,-0.1850347966,-0.0286590047,0.0989930332,0.3526166975,0.0663160384,0.1508650333,-0.1567121297,0.4380097985,0.1698628068,0.3453728259,-0.1365328282,0.1078100652,-0.3517187536,-0.1205077767,0.0466760062,0.506477654,0.0218654592,-0.1784384698,-0.2926490903,-0.5317479372,0.5479229093,-0.8124235868,-0.2576377988,0.0784190819,0.0957219377,-0.0114729004,-0.2659358084,-0.5267435908,0.2552785277,0.0984909907,-0.0782416686,-0.327549547,0.1157159433,-0.0942824036,-0.0646765679,-0.0526354015,0.3443565071,0.1893855482,-0.2799617648,-0.1954742223,-0.3156386018]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/228","title":"Not able to access the XNLI dataset","comments":"Added pull request to change the name of the file from dataset_infos.json to dataset_info.json","body":"When I try to access the XNLI dataset, I get the following error. The option of plain_text get selected automatically and then I get the following error.\r\n\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/sasha\/.cache\/huggingface\/datasets\/xnli\/plain_text\/1.0.0\/dataset_info.json'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 86, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 72, in get\r\n builder_instance = builder_cls(name=conf)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\n```\r\n\r\nIs it possible to see if the dataset_info.json is correctly placed?","comment_length":14,"text":"Not able to access the XNLI dataset\nWhen I try to access the XNLI dataset, I get the following error. The option of plain_text get selected automatically and then I get the following error.\r\n\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/sasha\/.cache\/huggingface\/datasets\/xnli\/plain_text\/1.0.0\/dataset_info.json'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 86, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 72, in get\r\n builder_instance = builder_cls(name=conf)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\n```\r\n\r\nIs it possible to see if the dataset_info.json is correctly placed?\nAdded pull request to change the name of the file from dataset_infos.json to dataset_info.json","embeddings":[-0.1408677548,-0.1892592013,-0.033536762,0.4693672359,0.4054459035,0.1120007858,0.0560360961,0.3544986248,-0.1094894633,0.2927758694,-0.1404097974,0.1621204913,0.0502276979,-0.0280459467,0.0505416356,-0.0353243835,-0.1691273153,0.1510201693,0.132073015,0.0053231702,-0.1707711071,0.0455822125,-0.2690639794,0.2587417364,-0.286706537,-0.0852832273,0.0914464816,0.3859780431,-0.3081700802,-0.53493011,0.0450817719,-0.1443354785,0.2407738417,0.1957660019,-0.000111058,0.0493330322,0.2629134953,-0.0573490635,-0.5286895037,-0.1454180032,-0.3568858802,-0.1231440604,-0.0146996509,-0.3132304847,-0.1287814528,-0.2209604383,0.4022442102,-0.3689225614,0.2718051374,0.3152759969,0.2039080113,0.2000430375,0.1218272746,-0.054714378,0.049337551,0.1351935565,-0.1314477473,-0.0139704924,0.2392853945,0.1932622641,0.1636672765,0.4578833878,-0.1711953729,-0.1116460562,0.255207777,0.2438426167,-0.0302988421,-0.1604979932,0.3209942281,0.2331341654,0.4437571764,-0.4332896769,-0.4334993362,-0.1114859581,-0.0038188207,-0.2649722695,0.2488645911,0.1001946703,-0.3177075386,0.1554604769,-0.1199381873,-0.3665495813,-0.3681380153,0.1454014629,0.2627976835,0.3802297413,-0.0596888699,0.0670860037,-0.0611994453,-0.1958376616,0.0528130345,-0.0894943327,-0.1641328484,0.3026693761,-0.2269841582,0.1276466697,0.1352884024,0.0571524575,0.1198465228,0.3628030717,0.2032430172,-0.1840142161,-0.3755405843,0.06257765,-0.2645841837,0.4228107333,0.2957006991,0.1408858001,0.3827148676,-0.0089444388,0.2218957394,-0.3649003506,-0.2945749164,-0.1601772606,-0.114383772,0.2509687543,0.3615804911,-0.1595380753,-0.2476482093,-0.0595754758,-0.1263961792,0.0193485729,0.2029277384,0.5645589828,0.0536458157,0.1894836426,0.0351332389,0.3217778802,0.053703472,-0.5332477689,-0.1743085086,0.2165416032,-0.3087960184,0.1456948817,0.3105262816,-0.3313245475,0.3001628816,-0.1157548949,-0.3032845855,0.0032358915,0.1294727772,-0.1749177426,0.1020331085,0.0779607892,0.0931859761,0.1030509546,0.1904263943,-0.0799686313,0.0655035079,-0.0332771763,-0.0471923687,-0.4852254391,0.0255070888,0.1306854486,0.0716528967,-0.009506152,-0.0878667012,-0.0543031245,0.1962302923,-0.2217216939,-0.1117357835,0.112898834,-0.0255647469,-0.2713457048,0.2126425505,0.5540782213,-0.3136119843,0.2092669159,-0.1985261887,-0.1348921061,0.163162753,0.3628430367,-0.1167293563,0.0775886104,-0.3848822117,0.3266322613,0.6447759867,-0.5202516913,-0.507689178,0.2816505432,-0.2223585993,-0.0715679377,0.5067173243,0.2407118678,-0.0645311177,-0.113183476,0.0470005237,0.117008321,0.0467213541,-0.0885009691,-0.1114888489,-0.2450889647,0.3402366638,0.1130200699,-0.0306807663,0.1320907325,-0.029109424,0.1613193005,0.3716828525,-0.2910831869,0.1511864364,0.2840608954,0.3005669415,0.0372093096,-0.1425783336,-0.1196034774,-0.2946609557,0.2399859577,-0.0920302644,-0.0201166086,-0.3088808358,0.0002530748,-0.1130894944,-0.0409190543,-0.3876262605,-0.2783128321,0.1904964447,0.4527053237,0.227343753,-0.0196524505,-0.2899426222,0.3173918426,-0.4826749563,0.0027164356,-0.6354672313,0.1280632615,-0.0931880027,-0.1490348577,-0.1639639139,0.1639258713,0.0235497002,-0.0028483428,-0.092747651,0.212470293,-0.2380763143,0.0813123882,0.2638707161,0.1023620442,0.0382039137,-0.5709121823,0.3375821114,0.4717598259,0.2284216285,0.0894116163,-0.2693724036,0.1183070689,-0.0825481787,0.0903004631,0.0437282436,0.0172032174,0.4666255414,0.0148105649,0.0699819252,-0.212793991,0.3003255725,0.216489777,0.4621777833,-0.086125508,-0.5364867449,0.0861315429,0.2796010077,0.0605193675,-0.0986444056,-0.0109362863,-0.3510042727,-0.0416343696,0.0550630279,0.2289790064,0.2813903093,0.2278090268,-0.0653548166,0.0831116587,0.0983732715,-0.2621799409,0.1807291061,-0.2579271495,0.1012303159,0.2698712349,0.1998011768,0.0320530571,-0.0356441364,-0.5033630133,0.0721699521,0.4454903901,-0.270344466,-0.1728234887,-0.1369042546,-0.6854811907,0.0606506467,0.0763859451,-0.11054077,-0.3432247937,-0.0222284645,0.18776007,0.2145616412,-0.0719953775,-0.0950528979,0.0755412653,0.0686472133,-0.109409146,-0.419331491,-0.3149599433,-0.2077754438,0.046856489,0.2707952857,-0.0228072517,0.2469428778,-0.2802794874,0.0639284253,-0.4144373536,-0.0254895911,0.0234772526,0.0292966571,-0.0004744022,-0.0340683497,0.4818875194,0.4775179625,-0.0623232983,0.3475892544,0.0415477231,-0.1742302775,0.2578029335,-0.2152395099,-0.2344868481,-0.0418086126,-0.5249136686,-0.3563694656,-0.3282742202,0.0827530324,-0.1398809403,0.0803990811,0.3180901706,-0.1464771181,0.2062613368,-0.038883388,0.1958673,-0.0754591152,-0.2256141007,0.3382492363,-0.161397025,-0.2159953713,0.2360935062,0.0649059862,0.2100883275,-0.0412420146,-0.2306012511,0.1721257716,-0.2255693376,0.2844632864,-0.1305034161,-0.1519210339,0.3499826789,-0.0105040036,-0.001459451,0.0043800203,0.0118807135,0.1178084835,0.0100269215,0.571620822,0.2039424479,0.4134452343,0.1445199102,0.3723564744,0.4486645758,0.0267595891,0.3283901513,-0.2621395588,0.0669077039,-0.0658819005,-0.3156565428,0.1194543019,0.1747564822,-0.1733243912,-0.1079721004,0.1386827379,0.3889435828,-0.3315194845,-0.1330213994,-0.2874074876,-0.0854833126,-0.1463778317,0.0202502068,0.0111190295,0.1291114837,0.2093752027,-0.3425545394,-0.3424660563,-0.1121322811,0.4633776844,0.0023366199,0.1523140669,-0.2744363844,-0.2841807306,-0.5618222952,0.4806259871,0.1180058867,0.2120662779,-0.2758866251,0.1250002533,0.266346544,0.0225754622,0.3283257484,-0.3004835844,0.2563642263,0.0807866678,-0.0989301056,-0.2358802706,0.1696390212,-0.1501615494,0.198500365,0.335871011,0.1939541698,-0.160841018,-0.1110729948,-0.016944252,0.1083815396,-0.1749693602,-0.2175936401,-0.0177838448,-0.0037040839,-0.1634147614,-0.3284824491,-0.1715682,0.2472744733,0.1329616308,0.3358918428,0.0468720645,-0.0742730275,0.1784730852,0.2204153389,0.339287132,-0.0961351916,-0.1533091366,0.1652644426,0.2498364896,0.0717588365,0.4986754954,0.2898446321,-0.3622654378,-0.0141012259,-0.2284084857,0.1272356212,0.2903348804,-0.2587492168,0.1040537134,0.0343668945,-0.0078272121,0.0616949275,-0.0959052071,0.2513180673,0.0315440856,-0.1259422749,-0.3986914158,0.3272973299,0.1011688933,0.0039729285,0.4540026486,-0.146319136,-0.2276169211,0.0237343851,0.0572314076,0.9369785786,0.1299279779,0.2292374223,0.3930836022,0.1683117896,0.5260003209,-0.3035136759,0.3688016534,-0.3871545494,-0.0803950131,-0.0029479368,-0.1137043014,-0.0116645973,-0.1774899364,-0.0803141966,0.2340769768,-0.1563339084,0.0061115115,-0.0985916182,0.2628973126,-0.3349470198,-0.0904692635,-0.4606150687,0.0847525224,-0.2080034167,0.334980011,-0.1050385013,-0.1498461068,-0.3853043914,0.1126162186,-0.4202090502,0.3758969307,-0.0859740525,0.1225233302,-0.0736165866,-0.3276994228,-0.0553329438,0.6137100458,0.5807358027,0.050126832,-0.3391729295,0.0077568712,-0.2113013864,-0.0230903607,-0.0938924998,0.1553285122,0.3382046223,-0.0408551991,-0.3091776371,0.5142868757,-0.1125862598,0.016242044,-0.2366230637,-0.2566816807,-0.240296945,-0.0432238802,-0.1230145022,0.0199265126,-0.1297458559,0.0251795202,0.0997499153,0.0126655791,-0.2329041064,-0.2464849204,-0.1339305639,-0.1444867551,0.1830044985,0.325381577,0.0474451073,0.1761954576,0.6140019894,0.2216762155,-0.0501693785,-0.2481475621,0.0654041916,0.1556771398,-0.2556731999,-0.2123807222,0.1041428968,0.1746927202,0.0965442285,0.2774943709,0.1864701658,-0.0062570614,-0.1428339779,-0.4487022161,-0.5422468781,0.1517494321,-0.1925470531,0.1390153468,-0.1516945809,0.2596283257,0.1187472492,-0.2652499974,-0.3370279372,-0.0957017839,-0.3234106898,-0.005745166,0.3934482038,0.0590532944,0.1968781054,0.0436726399,0.1371066868,-0.0367001854,-0.2849012911,-0.1598269492,0.0488781109,0.1000243649,0.1371713132,-0.4027868807,-0.1165812388,-0.2882617414,-0.0881268978,-0.1604430079,0.0502879359,0.2985337675,0.0316240191,0.2532197535,-0.3133458495,0.0580079556,-0.2136869878,0.1573789269,-0.1528376639,0.184106335,0.0598614365,0.1261301488,0.017676102,0.0098836925,-0.5507247448,-0.0625745282,-0.0907989293,0.1829756051,0.1152569503,-0.4182234406,0.228056103,-0.1419947892,0.171419993,0.1962636858,-0.0333172902,0.2832373977,0.1811421067,0.1862851232,-0.2386663258,0.2086086422,0.2157798111,0.0946389586,-0.0938897207,0.0430795886,-0.0494546257,-0.025663266,0.2692067325,-0.1024496406,0.3867939115,-0.4072207212,-0.052489385,0.1194665208,0.0713919997,-0.0525237285,0.1153041348,0.3662636876,0.1645925641,0.7827575207,-0.2470326126,0.1465714276,-0.184984982,0.3281104267,-0.2487826943,-0.3345987499,-0.3359787464,0.1134815142,-0.2942698896,-0.1261737496,-0.1852843314,0.2995412946,-0.1018136367,0.1427347064,-0.2357794791,0.3921541274,-0.1488359123,-0.0453949682,-0.5625841022,-0.2266335785,0.1058561429,-0.1797054708,0.0250908062,0.0159932282,-0.0091492422,0.1309443563,-0.2652380168,-0.4018542171,0.3093726933,0.4047755599,-0.0163565241,-0.2793292999,0.4106734395,0.2905892134,0.0659506842,0.2053315043,0.3192927837,0.4276020825,0.4022081494,0.0665072724,0.2000069916,-0.1069403887,-0.2528617382,-0.0138815902,0.0525235943,0.3014642596,0.0584575385,0.312531054,0.1497073919,-0.0062345145,0.1430208981,0.045691397,0.3462027311,0.0922235548,-0.0150208259,-0.1465478987,-0.2455284297,-0.1894638687,-0.1285670996,-0.558146894,0.1378033608,0.2128199339,0.215043202,0.2549575567,-0.0042081224,0.0490093678,-0.0665702969,0.1328429878,0.5250784755,-0.0340116769,-0.2703349888,-0.3209694922,-0.5026276112,0.1653779298,-0.0252851732,-0.420865953,-0.2485361844,0.1211538464,-0.1755242199,-0.2355565578,0.2410198301,0.1014041528,0.069594197,0.3392386734,-0.2268995792,-0.3327465951,0.2696409523,0.0531273633,-0.1709012687,-0.1728681773,-0.0946457535,-0.1486764699,0.0490003414,-0.0460385568,0.3904182315,0.0071133403,0.2057304978,0.1977957785,0.1998120844,0.4811964035,0.0680666938,-0.0404394753,-0.3731051385,-0.1117367521,-0.2012594044,-0.0415709764,0.3111200333,0.5081033111,-0.3583825231,-0.0083459811,-0.3815684319,0.1278301626,-0.1665936261,-0.0174836386,-0.2955942452,0.3703538477,-0.2835772932,0.2372785211,0.1641732454,-0.0455508903,-0.0290332139,0.1030434892,-0.3331671655,0.0036255466,0.5232225657,-0.2101983577,-0.2365404814,0.0900798738,0.1933183372,0.1135041565,0.0103466297,-0.3378685713,0.1756677479,0.1824224442,-0.0053519118,-0.2406205684,-0.0816917047,0.0652151182,0.1135519743,-0.0653273016,-0.1427070946,0.159398526,-0.0728523582,-0.0136100473,-0.0865384638]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/228","title":"Not able to access the XNLI dataset","comments":"Thanks for reporting this bug !\r\nAs it seems to be just a cache problem, I closed your PR.\r\nI think we might just need to clear and reload the `xnli` cache @srush ? ","body":"When I try to access the XNLI dataset, I get the following error. The option of plain_text get selected automatically and then I get the following error.\r\n\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/sasha\/.cache\/huggingface\/datasets\/xnli\/plain_text\/1.0.0\/dataset_info.json'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 86, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 72, in get\r\n builder_instance = builder_cls(name=conf)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\n```\r\n\r\nIs it possible to see if the dataset_info.json is correctly placed?","comment_length":34,"text":"Not able to access the XNLI dataset\nWhen I try to access the XNLI dataset, I get the following error. The option of plain_text get selected automatically and then I get the following error.\r\n\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/sasha\/.cache\/huggingface\/datasets\/xnli\/plain_text\/1.0.0\/dataset_info.json'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 86, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 72, in get\r\n builder_instance = builder_cls(name=conf)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\n```\r\n\r\nIs it possible to see if the dataset_info.json is correctly placed?\nThanks for reporting this bug !\r\nAs it seems to be just a cache problem, I closed your PR.\r\nI think we might just need to clear and reload the `xnli` cache @srush ? ","embeddings":[-0.1408677548,-0.1892592013,-0.033536762,0.4693672359,0.4054459035,0.1120007858,0.0560360961,0.3544986248,-0.1094894633,0.2927758694,-0.1404097974,0.1621204913,0.0502276979,-0.0280459467,0.0505416356,-0.0353243835,-0.1691273153,0.1510201693,0.132073015,0.0053231702,-0.1707711071,0.0455822125,-0.2690639794,0.2587417364,-0.286706537,-0.0852832273,0.0914464816,0.3859780431,-0.3081700802,-0.53493011,0.0450817719,-0.1443354785,0.2407738417,0.1957660019,-0.000111058,0.0493330322,0.2629134953,-0.0573490635,-0.5286895037,-0.1454180032,-0.3568858802,-0.1231440604,-0.0146996509,-0.3132304847,-0.1287814528,-0.2209604383,0.4022442102,-0.3689225614,0.2718051374,0.3152759969,0.2039080113,0.2000430375,0.1218272746,-0.054714378,0.049337551,0.1351935565,-0.1314477473,-0.0139704924,0.2392853945,0.1932622641,0.1636672765,0.4578833878,-0.1711953729,-0.1116460562,0.255207777,0.2438426167,-0.0302988421,-0.1604979932,0.3209942281,0.2331341654,0.4437571764,-0.4332896769,-0.4334993362,-0.1114859581,-0.0038188207,-0.2649722695,0.2488645911,0.1001946703,-0.3177075386,0.1554604769,-0.1199381873,-0.3665495813,-0.3681380153,0.1454014629,0.2627976835,0.3802297413,-0.0596888699,0.0670860037,-0.0611994453,-0.1958376616,0.0528130345,-0.0894943327,-0.1641328484,0.3026693761,-0.2269841582,0.1276466697,0.1352884024,0.0571524575,0.1198465228,0.3628030717,0.2032430172,-0.1840142161,-0.3755405843,0.06257765,-0.2645841837,0.4228107333,0.2957006991,0.1408858001,0.3827148676,-0.0089444388,0.2218957394,-0.3649003506,-0.2945749164,-0.1601772606,-0.114383772,0.2509687543,0.3615804911,-0.1595380753,-0.2476482093,-0.0595754758,-0.1263961792,0.0193485729,0.2029277384,0.5645589828,0.0536458157,0.1894836426,0.0351332389,0.3217778802,0.053703472,-0.5332477689,-0.1743085086,0.2165416032,-0.3087960184,0.1456948817,0.3105262816,-0.3313245475,0.3001628816,-0.1157548949,-0.3032845855,0.0032358915,0.1294727772,-0.1749177426,0.1020331085,0.0779607892,0.0931859761,0.1030509546,0.1904263943,-0.0799686313,0.0655035079,-0.0332771763,-0.0471923687,-0.4852254391,0.0255070888,0.1306854486,0.0716528967,-0.009506152,-0.0878667012,-0.0543031245,0.1962302923,-0.2217216939,-0.1117357835,0.112898834,-0.0255647469,-0.2713457048,0.2126425505,0.5540782213,-0.3136119843,0.2092669159,-0.1985261887,-0.1348921061,0.163162753,0.3628430367,-0.1167293563,0.0775886104,-0.3848822117,0.3266322613,0.6447759867,-0.5202516913,-0.507689178,0.2816505432,-0.2223585993,-0.0715679377,0.5067173243,0.2407118678,-0.0645311177,-0.113183476,0.0470005237,0.117008321,0.0467213541,-0.0885009691,-0.1114888489,-0.2450889647,0.3402366638,0.1130200699,-0.0306807663,0.1320907325,-0.029109424,0.1613193005,0.3716828525,-0.2910831869,0.1511864364,0.2840608954,0.3005669415,0.0372093096,-0.1425783336,-0.1196034774,-0.2946609557,0.2399859577,-0.0920302644,-0.0201166086,-0.3088808358,0.0002530748,-0.1130894944,-0.0409190543,-0.3876262605,-0.2783128321,0.1904964447,0.4527053237,0.227343753,-0.0196524505,-0.2899426222,0.3173918426,-0.4826749563,0.0027164356,-0.6354672313,0.1280632615,-0.0931880027,-0.1490348577,-0.1639639139,0.1639258713,0.0235497002,-0.0028483428,-0.092747651,0.212470293,-0.2380763143,0.0813123882,0.2638707161,0.1023620442,0.0382039137,-0.5709121823,0.3375821114,0.4717598259,0.2284216285,0.0894116163,-0.2693724036,0.1183070689,-0.0825481787,0.0903004631,0.0437282436,0.0172032174,0.4666255414,0.0148105649,0.0699819252,-0.212793991,0.3003255725,0.216489777,0.4621777833,-0.086125508,-0.5364867449,0.0861315429,0.2796010077,0.0605193675,-0.0986444056,-0.0109362863,-0.3510042727,-0.0416343696,0.0550630279,0.2289790064,0.2813903093,0.2278090268,-0.0653548166,0.0831116587,0.0983732715,-0.2621799409,0.1807291061,-0.2579271495,0.1012303159,0.2698712349,0.1998011768,0.0320530571,-0.0356441364,-0.5033630133,0.0721699521,0.4454903901,-0.270344466,-0.1728234887,-0.1369042546,-0.6854811907,0.0606506467,0.0763859451,-0.11054077,-0.3432247937,-0.0222284645,0.18776007,0.2145616412,-0.0719953775,-0.0950528979,0.0755412653,0.0686472133,-0.109409146,-0.419331491,-0.3149599433,-0.2077754438,0.046856489,0.2707952857,-0.0228072517,0.2469428778,-0.2802794874,0.0639284253,-0.4144373536,-0.0254895911,0.0234772526,0.0292966571,-0.0004744022,-0.0340683497,0.4818875194,0.4775179625,-0.0623232983,0.3475892544,0.0415477231,-0.1742302775,0.2578029335,-0.2152395099,-0.2344868481,-0.0418086126,-0.5249136686,-0.3563694656,-0.3282742202,0.0827530324,-0.1398809403,0.0803990811,0.3180901706,-0.1464771181,0.2062613368,-0.038883388,0.1958673,-0.0754591152,-0.2256141007,0.3382492363,-0.161397025,-0.2159953713,0.2360935062,0.0649059862,0.2100883275,-0.0412420146,-0.2306012511,0.1721257716,-0.2255693376,0.2844632864,-0.1305034161,-0.1519210339,0.3499826789,-0.0105040036,-0.001459451,0.0043800203,0.0118807135,0.1178084835,0.0100269215,0.571620822,0.2039424479,0.4134452343,0.1445199102,0.3723564744,0.4486645758,0.0267595891,0.3283901513,-0.2621395588,0.0669077039,-0.0658819005,-0.3156565428,0.1194543019,0.1747564822,-0.1733243912,-0.1079721004,0.1386827379,0.3889435828,-0.3315194845,-0.1330213994,-0.2874074876,-0.0854833126,-0.1463778317,0.0202502068,0.0111190295,0.1291114837,0.2093752027,-0.3425545394,-0.3424660563,-0.1121322811,0.4633776844,0.0023366199,0.1523140669,-0.2744363844,-0.2841807306,-0.5618222952,0.4806259871,0.1180058867,0.2120662779,-0.2758866251,0.1250002533,0.266346544,0.0225754622,0.3283257484,-0.3004835844,0.2563642263,0.0807866678,-0.0989301056,-0.2358802706,0.1696390212,-0.1501615494,0.198500365,0.335871011,0.1939541698,-0.160841018,-0.1110729948,-0.016944252,0.1083815396,-0.1749693602,-0.2175936401,-0.0177838448,-0.0037040839,-0.1634147614,-0.3284824491,-0.1715682,0.2472744733,0.1329616308,0.3358918428,0.0468720645,-0.0742730275,0.1784730852,0.2204153389,0.339287132,-0.0961351916,-0.1533091366,0.1652644426,0.2498364896,0.0717588365,0.4986754954,0.2898446321,-0.3622654378,-0.0141012259,-0.2284084857,0.1272356212,0.2903348804,-0.2587492168,0.1040537134,0.0343668945,-0.0078272121,0.0616949275,-0.0959052071,0.2513180673,0.0315440856,-0.1259422749,-0.3986914158,0.3272973299,0.1011688933,0.0039729285,0.4540026486,-0.146319136,-0.2276169211,0.0237343851,0.0572314076,0.9369785786,0.1299279779,0.2292374223,0.3930836022,0.1683117896,0.5260003209,-0.3035136759,0.3688016534,-0.3871545494,-0.0803950131,-0.0029479368,-0.1137043014,-0.0116645973,-0.1774899364,-0.0803141966,0.2340769768,-0.1563339084,0.0061115115,-0.0985916182,0.2628973126,-0.3349470198,-0.0904692635,-0.4606150687,0.0847525224,-0.2080034167,0.334980011,-0.1050385013,-0.1498461068,-0.3853043914,0.1126162186,-0.4202090502,0.3758969307,-0.0859740525,0.1225233302,-0.0736165866,-0.3276994228,-0.0553329438,0.6137100458,0.5807358027,0.050126832,-0.3391729295,0.0077568712,-0.2113013864,-0.0230903607,-0.0938924998,0.1553285122,0.3382046223,-0.0408551991,-0.3091776371,0.5142868757,-0.1125862598,0.016242044,-0.2366230637,-0.2566816807,-0.240296945,-0.0432238802,-0.1230145022,0.0199265126,-0.1297458559,0.0251795202,0.0997499153,0.0126655791,-0.2329041064,-0.2464849204,-0.1339305639,-0.1444867551,0.1830044985,0.325381577,0.0474451073,0.1761954576,0.6140019894,0.2216762155,-0.0501693785,-0.2481475621,0.0654041916,0.1556771398,-0.2556731999,-0.2123807222,0.1041428968,0.1746927202,0.0965442285,0.2774943709,0.1864701658,-0.0062570614,-0.1428339779,-0.4487022161,-0.5422468781,0.1517494321,-0.1925470531,0.1390153468,-0.1516945809,0.2596283257,0.1187472492,-0.2652499974,-0.3370279372,-0.0957017839,-0.3234106898,-0.005745166,0.3934482038,0.0590532944,0.1968781054,0.0436726399,0.1371066868,-0.0367001854,-0.2849012911,-0.1598269492,0.0488781109,0.1000243649,0.1371713132,-0.4027868807,-0.1165812388,-0.2882617414,-0.0881268978,-0.1604430079,0.0502879359,0.2985337675,0.0316240191,0.2532197535,-0.3133458495,0.0580079556,-0.2136869878,0.1573789269,-0.1528376639,0.184106335,0.0598614365,0.1261301488,0.017676102,0.0098836925,-0.5507247448,-0.0625745282,-0.0907989293,0.1829756051,0.1152569503,-0.4182234406,0.228056103,-0.1419947892,0.171419993,0.1962636858,-0.0333172902,0.2832373977,0.1811421067,0.1862851232,-0.2386663258,0.2086086422,0.2157798111,0.0946389586,-0.0938897207,0.0430795886,-0.0494546257,-0.025663266,0.2692067325,-0.1024496406,0.3867939115,-0.4072207212,-0.052489385,0.1194665208,0.0713919997,-0.0525237285,0.1153041348,0.3662636876,0.1645925641,0.7827575207,-0.2470326126,0.1465714276,-0.184984982,0.3281104267,-0.2487826943,-0.3345987499,-0.3359787464,0.1134815142,-0.2942698896,-0.1261737496,-0.1852843314,0.2995412946,-0.1018136367,0.1427347064,-0.2357794791,0.3921541274,-0.1488359123,-0.0453949682,-0.5625841022,-0.2266335785,0.1058561429,-0.1797054708,0.0250908062,0.0159932282,-0.0091492422,0.1309443563,-0.2652380168,-0.4018542171,0.3093726933,0.4047755599,-0.0163565241,-0.2793292999,0.4106734395,0.2905892134,0.0659506842,0.2053315043,0.3192927837,0.4276020825,0.4022081494,0.0665072724,0.2000069916,-0.1069403887,-0.2528617382,-0.0138815902,0.0525235943,0.3014642596,0.0584575385,0.312531054,0.1497073919,-0.0062345145,0.1430208981,0.045691397,0.3462027311,0.0922235548,-0.0150208259,-0.1465478987,-0.2455284297,-0.1894638687,-0.1285670996,-0.558146894,0.1378033608,0.2128199339,0.215043202,0.2549575567,-0.0042081224,0.0490093678,-0.0665702969,0.1328429878,0.5250784755,-0.0340116769,-0.2703349888,-0.3209694922,-0.5026276112,0.1653779298,-0.0252851732,-0.420865953,-0.2485361844,0.1211538464,-0.1755242199,-0.2355565578,0.2410198301,0.1014041528,0.069594197,0.3392386734,-0.2268995792,-0.3327465951,0.2696409523,0.0531273633,-0.1709012687,-0.1728681773,-0.0946457535,-0.1486764699,0.0490003414,-0.0460385568,0.3904182315,0.0071133403,0.2057304978,0.1977957785,0.1998120844,0.4811964035,0.0680666938,-0.0404394753,-0.3731051385,-0.1117367521,-0.2012594044,-0.0415709764,0.3111200333,0.5081033111,-0.3583825231,-0.0083459811,-0.3815684319,0.1278301626,-0.1665936261,-0.0174836386,-0.2955942452,0.3703538477,-0.2835772932,0.2372785211,0.1641732454,-0.0455508903,-0.0290332139,0.1030434892,-0.3331671655,0.0036255466,0.5232225657,-0.2101983577,-0.2365404814,0.0900798738,0.1933183372,0.1135041565,0.0103466297,-0.3378685713,0.1756677479,0.1824224442,-0.0053519118,-0.2406205684,-0.0816917047,0.0652151182,0.1135519743,-0.0653273016,-0.1427070946,0.159398526,-0.0728523582,-0.0136100473,-0.0865384638]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/228","title":"Not able to access the XNLI dataset","comments":"Update: The dataset_info.json error is gone, but we have a new one instead:\r\n```\r\nConnectionError: Couldn't reach https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip\r\n```\r\nI am not able to reproduce on my side unfortunately. Any idea @srush ?","body":"When I try to access the XNLI dataset, I get the following error. The option of plain_text get selected automatically and then I get the following error.\r\n\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/sasha\/.cache\/huggingface\/datasets\/xnli\/plain_text\/1.0.0\/dataset_info.json'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 86, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 72, in get\r\n builder_instance = builder_cls(name=conf)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\n```\r\n\r\nIs it possible to see if the dataset_info.json is correctly placed?","comment_length":33,"text":"Not able to access the XNLI dataset\nWhen I try to access the XNLI dataset, I get the following error. The option of plain_text get selected automatically and then I get the following error.\r\n\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/sasha\/.cache\/huggingface\/datasets\/xnli\/plain_text\/1.0.0\/dataset_info.json'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 86, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 72, in get\r\n builder_instance = builder_cls(name=conf)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\n```\r\n\r\nIs it possible to see if the dataset_info.json is correctly placed?\nUpdate: The dataset_info.json error is gone, but we have a new one instead:\r\n```\r\nConnectionError: Couldn't reach https:\/\/www.nyu.edu\/projects\/bowman\/xnli\/XNLI-1.0.zip\r\n```\r\nI am not able to reproduce on my side unfortunately. Any idea @srush ?","embeddings":[-0.1408677548,-0.1892592013,-0.033536762,0.4693672359,0.4054459035,0.1120007858,0.0560360961,0.3544986248,-0.1094894633,0.2927758694,-0.1404097974,0.1621204913,0.0502276979,-0.0280459467,0.0505416356,-0.0353243835,-0.1691273153,0.1510201693,0.132073015,0.0053231702,-0.1707711071,0.0455822125,-0.2690639794,0.2587417364,-0.286706537,-0.0852832273,0.0914464816,0.3859780431,-0.3081700802,-0.53493011,0.0450817719,-0.1443354785,0.2407738417,0.1957660019,-0.000111058,0.0493330322,0.2629134953,-0.0573490635,-0.5286895037,-0.1454180032,-0.3568858802,-0.1231440604,-0.0146996509,-0.3132304847,-0.1287814528,-0.2209604383,0.4022442102,-0.3689225614,0.2718051374,0.3152759969,0.2039080113,0.2000430375,0.1218272746,-0.054714378,0.049337551,0.1351935565,-0.1314477473,-0.0139704924,0.2392853945,0.1932622641,0.1636672765,0.4578833878,-0.1711953729,-0.1116460562,0.255207777,0.2438426167,-0.0302988421,-0.1604979932,0.3209942281,0.2331341654,0.4437571764,-0.4332896769,-0.4334993362,-0.1114859581,-0.0038188207,-0.2649722695,0.2488645911,0.1001946703,-0.3177075386,0.1554604769,-0.1199381873,-0.3665495813,-0.3681380153,0.1454014629,0.2627976835,0.3802297413,-0.0596888699,0.0670860037,-0.0611994453,-0.1958376616,0.0528130345,-0.0894943327,-0.1641328484,0.3026693761,-0.2269841582,0.1276466697,0.1352884024,0.0571524575,0.1198465228,0.3628030717,0.2032430172,-0.1840142161,-0.3755405843,0.06257765,-0.2645841837,0.4228107333,0.2957006991,0.1408858001,0.3827148676,-0.0089444388,0.2218957394,-0.3649003506,-0.2945749164,-0.1601772606,-0.114383772,0.2509687543,0.3615804911,-0.1595380753,-0.2476482093,-0.0595754758,-0.1263961792,0.0193485729,0.2029277384,0.5645589828,0.0536458157,0.1894836426,0.0351332389,0.3217778802,0.053703472,-0.5332477689,-0.1743085086,0.2165416032,-0.3087960184,0.1456948817,0.3105262816,-0.3313245475,0.3001628816,-0.1157548949,-0.3032845855,0.0032358915,0.1294727772,-0.1749177426,0.1020331085,0.0779607892,0.0931859761,0.1030509546,0.1904263943,-0.0799686313,0.0655035079,-0.0332771763,-0.0471923687,-0.4852254391,0.0255070888,0.1306854486,0.0716528967,-0.009506152,-0.0878667012,-0.0543031245,0.1962302923,-0.2217216939,-0.1117357835,0.112898834,-0.0255647469,-0.2713457048,0.2126425505,0.5540782213,-0.3136119843,0.2092669159,-0.1985261887,-0.1348921061,0.163162753,0.3628430367,-0.1167293563,0.0775886104,-0.3848822117,0.3266322613,0.6447759867,-0.5202516913,-0.507689178,0.2816505432,-0.2223585993,-0.0715679377,0.5067173243,0.2407118678,-0.0645311177,-0.113183476,0.0470005237,0.117008321,0.0467213541,-0.0885009691,-0.1114888489,-0.2450889647,0.3402366638,0.1130200699,-0.0306807663,0.1320907325,-0.029109424,0.1613193005,0.3716828525,-0.2910831869,0.1511864364,0.2840608954,0.3005669415,0.0372093096,-0.1425783336,-0.1196034774,-0.2946609557,0.2399859577,-0.0920302644,-0.0201166086,-0.3088808358,0.0002530748,-0.1130894944,-0.0409190543,-0.3876262605,-0.2783128321,0.1904964447,0.4527053237,0.227343753,-0.0196524505,-0.2899426222,0.3173918426,-0.4826749563,0.0027164356,-0.6354672313,0.1280632615,-0.0931880027,-0.1490348577,-0.1639639139,0.1639258713,0.0235497002,-0.0028483428,-0.092747651,0.212470293,-0.2380763143,0.0813123882,0.2638707161,0.1023620442,0.0382039137,-0.5709121823,0.3375821114,0.4717598259,0.2284216285,0.0894116163,-0.2693724036,0.1183070689,-0.0825481787,0.0903004631,0.0437282436,0.0172032174,0.4666255414,0.0148105649,0.0699819252,-0.212793991,0.3003255725,0.216489777,0.4621777833,-0.086125508,-0.5364867449,0.0861315429,0.2796010077,0.0605193675,-0.0986444056,-0.0109362863,-0.3510042727,-0.0416343696,0.0550630279,0.2289790064,0.2813903093,0.2278090268,-0.0653548166,0.0831116587,0.0983732715,-0.2621799409,0.1807291061,-0.2579271495,0.1012303159,0.2698712349,0.1998011768,0.0320530571,-0.0356441364,-0.5033630133,0.0721699521,0.4454903901,-0.270344466,-0.1728234887,-0.1369042546,-0.6854811907,0.0606506467,0.0763859451,-0.11054077,-0.3432247937,-0.0222284645,0.18776007,0.2145616412,-0.0719953775,-0.0950528979,0.0755412653,0.0686472133,-0.109409146,-0.419331491,-0.3149599433,-0.2077754438,0.046856489,0.2707952857,-0.0228072517,0.2469428778,-0.2802794874,0.0639284253,-0.4144373536,-0.0254895911,0.0234772526,0.0292966571,-0.0004744022,-0.0340683497,0.4818875194,0.4775179625,-0.0623232983,0.3475892544,0.0415477231,-0.1742302775,0.2578029335,-0.2152395099,-0.2344868481,-0.0418086126,-0.5249136686,-0.3563694656,-0.3282742202,0.0827530324,-0.1398809403,0.0803990811,0.3180901706,-0.1464771181,0.2062613368,-0.038883388,0.1958673,-0.0754591152,-0.2256141007,0.3382492363,-0.161397025,-0.2159953713,0.2360935062,0.0649059862,0.2100883275,-0.0412420146,-0.2306012511,0.1721257716,-0.2255693376,0.2844632864,-0.1305034161,-0.1519210339,0.3499826789,-0.0105040036,-0.001459451,0.0043800203,0.0118807135,0.1178084835,0.0100269215,0.571620822,0.2039424479,0.4134452343,0.1445199102,0.3723564744,0.4486645758,0.0267595891,0.3283901513,-0.2621395588,0.0669077039,-0.0658819005,-0.3156565428,0.1194543019,0.1747564822,-0.1733243912,-0.1079721004,0.1386827379,0.3889435828,-0.3315194845,-0.1330213994,-0.2874074876,-0.0854833126,-0.1463778317,0.0202502068,0.0111190295,0.1291114837,0.2093752027,-0.3425545394,-0.3424660563,-0.1121322811,0.4633776844,0.0023366199,0.1523140669,-0.2744363844,-0.2841807306,-0.5618222952,0.4806259871,0.1180058867,0.2120662779,-0.2758866251,0.1250002533,0.266346544,0.0225754622,0.3283257484,-0.3004835844,0.2563642263,0.0807866678,-0.0989301056,-0.2358802706,0.1696390212,-0.1501615494,0.198500365,0.335871011,0.1939541698,-0.160841018,-0.1110729948,-0.016944252,0.1083815396,-0.1749693602,-0.2175936401,-0.0177838448,-0.0037040839,-0.1634147614,-0.3284824491,-0.1715682,0.2472744733,0.1329616308,0.3358918428,0.0468720645,-0.0742730275,0.1784730852,0.2204153389,0.339287132,-0.0961351916,-0.1533091366,0.1652644426,0.2498364896,0.0717588365,0.4986754954,0.2898446321,-0.3622654378,-0.0141012259,-0.2284084857,0.1272356212,0.2903348804,-0.2587492168,0.1040537134,0.0343668945,-0.0078272121,0.0616949275,-0.0959052071,0.2513180673,0.0315440856,-0.1259422749,-0.3986914158,0.3272973299,0.1011688933,0.0039729285,0.4540026486,-0.146319136,-0.2276169211,0.0237343851,0.0572314076,0.9369785786,0.1299279779,0.2292374223,0.3930836022,0.1683117896,0.5260003209,-0.3035136759,0.3688016534,-0.3871545494,-0.0803950131,-0.0029479368,-0.1137043014,-0.0116645973,-0.1774899364,-0.0803141966,0.2340769768,-0.1563339084,0.0061115115,-0.0985916182,0.2628973126,-0.3349470198,-0.0904692635,-0.4606150687,0.0847525224,-0.2080034167,0.334980011,-0.1050385013,-0.1498461068,-0.3853043914,0.1126162186,-0.4202090502,0.3758969307,-0.0859740525,0.1225233302,-0.0736165866,-0.3276994228,-0.0553329438,0.6137100458,0.5807358027,0.050126832,-0.3391729295,0.0077568712,-0.2113013864,-0.0230903607,-0.0938924998,0.1553285122,0.3382046223,-0.0408551991,-0.3091776371,0.5142868757,-0.1125862598,0.016242044,-0.2366230637,-0.2566816807,-0.240296945,-0.0432238802,-0.1230145022,0.0199265126,-0.1297458559,0.0251795202,0.0997499153,0.0126655791,-0.2329041064,-0.2464849204,-0.1339305639,-0.1444867551,0.1830044985,0.325381577,0.0474451073,0.1761954576,0.6140019894,0.2216762155,-0.0501693785,-0.2481475621,0.0654041916,0.1556771398,-0.2556731999,-0.2123807222,0.1041428968,0.1746927202,0.0965442285,0.2774943709,0.1864701658,-0.0062570614,-0.1428339779,-0.4487022161,-0.5422468781,0.1517494321,-0.1925470531,0.1390153468,-0.1516945809,0.2596283257,0.1187472492,-0.2652499974,-0.3370279372,-0.0957017839,-0.3234106898,-0.005745166,0.3934482038,0.0590532944,0.1968781054,0.0436726399,0.1371066868,-0.0367001854,-0.2849012911,-0.1598269492,0.0488781109,0.1000243649,0.1371713132,-0.4027868807,-0.1165812388,-0.2882617414,-0.0881268978,-0.1604430079,0.0502879359,0.2985337675,0.0316240191,0.2532197535,-0.3133458495,0.0580079556,-0.2136869878,0.1573789269,-0.1528376639,0.184106335,0.0598614365,0.1261301488,0.017676102,0.0098836925,-0.5507247448,-0.0625745282,-0.0907989293,0.1829756051,0.1152569503,-0.4182234406,0.228056103,-0.1419947892,0.171419993,0.1962636858,-0.0333172902,0.2832373977,0.1811421067,0.1862851232,-0.2386663258,0.2086086422,0.2157798111,0.0946389586,-0.0938897207,0.0430795886,-0.0494546257,-0.025663266,0.2692067325,-0.1024496406,0.3867939115,-0.4072207212,-0.052489385,0.1194665208,0.0713919997,-0.0525237285,0.1153041348,0.3662636876,0.1645925641,0.7827575207,-0.2470326126,0.1465714276,-0.184984982,0.3281104267,-0.2487826943,-0.3345987499,-0.3359787464,0.1134815142,-0.2942698896,-0.1261737496,-0.1852843314,0.2995412946,-0.1018136367,0.1427347064,-0.2357794791,0.3921541274,-0.1488359123,-0.0453949682,-0.5625841022,-0.2266335785,0.1058561429,-0.1797054708,0.0250908062,0.0159932282,-0.0091492422,0.1309443563,-0.2652380168,-0.4018542171,0.3093726933,0.4047755599,-0.0163565241,-0.2793292999,0.4106734395,0.2905892134,0.0659506842,0.2053315043,0.3192927837,0.4276020825,0.4022081494,0.0665072724,0.2000069916,-0.1069403887,-0.2528617382,-0.0138815902,0.0525235943,0.3014642596,0.0584575385,0.312531054,0.1497073919,-0.0062345145,0.1430208981,0.045691397,0.3462027311,0.0922235548,-0.0150208259,-0.1465478987,-0.2455284297,-0.1894638687,-0.1285670996,-0.558146894,0.1378033608,0.2128199339,0.215043202,0.2549575567,-0.0042081224,0.0490093678,-0.0665702969,0.1328429878,0.5250784755,-0.0340116769,-0.2703349888,-0.3209694922,-0.5026276112,0.1653779298,-0.0252851732,-0.420865953,-0.2485361844,0.1211538464,-0.1755242199,-0.2355565578,0.2410198301,0.1014041528,0.069594197,0.3392386734,-0.2268995792,-0.3327465951,0.2696409523,0.0531273633,-0.1709012687,-0.1728681773,-0.0946457535,-0.1486764699,0.0490003414,-0.0460385568,0.3904182315,0.0071133403,0.2057304978,0.1977957785,0.1998120844,0.4811964035,0.0680666938,-0.0404394753,-0.3731051385,-0.1117367521,-0.2012594044,-0.0415709764,0.3111200333,0.5081033111,-0.3583825231,-0.0083459811,-0.3815684319,0.1278301626,-0.1665936261,-0.0174836386,-0.2955942452,0.3703538477,-0.2835772932,0.2372785211,0.1641732454,-0.0455508903,-0.0290332139,0.1030434892,-0.3331671655,0.0036255466,0.5232225657,-0.2101983577,-0.2365404814,0.0900798738,0.1933183372,0.1135041565,0.0103466297,-0.3378685713,0.1756677479,0.1824224442,-0.0053519118,-0.2406205684,-0.0816917047,0.0652151182,0.1135519743,-0.0653273016,-0.1427070946,0.159398526,-0.0728523582,-0.0136100473,-0.0865384638]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/228","title":"Not able to access the XNLI dataset","comments":"xnli is now properly shown in the viewer.\r\nClosing this one.","body":"When I try to access the XNLI dataset, I get the following error. The option of plain_text get selected automatically and then I get the following error.\r\n\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/sasha\/.cache\/huggingface\/datasets\/xnli\/plain_text\/1.0.0\/dataset_info.json'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 86, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 72, in get\r\n builder_instance = builder_cls(name=conf)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\n```\r\n\r\nIs it possible to see if the dataset_info.json is correctly placed?","comment_length":11,"text":"Not able to access the XNLI dataset\nWhen I try to access the XNLI dataset, I get the following error. The option of plain_text get selected automatically and then I get the following error.\r\n\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/home\/sasha\/.cache\/huggingface\/datasets\/xnli\/plain_text\/1.0.0\/dataset_info.json'\r\nTraceback:\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 86, in <module>\r\n dts, fail = get(str(option.id), str(conf_option.name) if conf_option else None)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 591, in wrapped_func\r\n return get_or_create_cached_value()\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/caching.py\", line 575, in get_or_create_cached_value\r\n return_value = func(*args, **kwargs)\r\nFile \"\/home\/sasha\/nlp_viewer\/run.py\", line 72, in get\r\n builder_instance = builder_cls(name=conf)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 610, in __init__\r\n super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/builder.py\", line 152, in __init__\r\n self.info = DatasetInfo.from_directory(self._cache_dir)\r\nFile \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/info.py\", line 157, in from_directory\r\n with open(os.path.join(dataset_info_dir, DATASET_INFO_FILENAME), \"r\") as f:\r\n```\r\n\r\nIs it possible to see if the dataset_info.json is correctly placed?\nxnli is now properly shown in the viewer.\r\nClosing this one.","embeddings":[-0.1408677548,-0.1892592013,-0.033536762,0.4693672359,0.4054459035,0.1120007858,0.0560360961,0.3544986248,-0.1094894633,0.2927758694,-0.1404097974,0.1621204913,0.0502276979,-0.0280459467,0.0505416356,-0.0353243835,-0.1691273153,0.1510201693,0.132073015,0.0053231702,-0.1707711071,0.0455822125,-0.2690639794,0.2587417364,-0.286706537,-0.0852832273,0.0914464816,0.3859780431,-0.3081700802,-0.53493011,0.0450817719,-0.1443354785,0.2407738417,0.1957660019,-0.000111058,0.0493330322,0.2629134953,-0.0573490635,-0.5286895037,-0.1454180032,-0.3568858802,-0.1231440604,-0.0146996509,-0.3132304847,-0.1287814528,-0.2209604383,0.4022442102,-0.3689225614,0.2718051374,0.3152759969,0.2039080113,0.2000430375,0.1218272746,-0.054714378,0.049337551,0.1351935565,-0.1314477473,-0.0139704924,0.2392853945,0.1932622641,0.1636672765,0.4578833878,-0.1711953729,-0.1116460562,0.255207777,0.2438426167,-0.0302988421,-0.1604979932,0.3209942281,0.2331341654,0.4437571764,-0.4332896769,-0.4334993362,-0.1114859581,-0.0038188207,-0.2649722695,0.2488645911,0.1001946703,-0.3177075386,0.1554604769,-0.1199381873,-0.3665495813,-0.3681380153,0.1454014629,0.2627976835,0.3802297413,-0.0596888699,0.0670860037,-0.0611994453,-0.1958376616,0.0528130345,-0.0894943327,-0.1641328484,0.3026693761,-0.2269841582,0.1276466697,0.1352884024,0.0571524575,0.1198465228,0.3628030717,0.2032430172,-0.1840142161,-0.3755405843,0.06257765,-0.2645841837,0.4228107333,0.2957006991,0.1408858001,0.3827148676,-0.0089444388,0.2218957394,-0.3649003506,-0.2945749164,-0.1601772606,-0.114383772,0.2509687543,0.3615804911,-0.1595380753,-0.2476482093,-0.0595754758,-0.1263961792,0.0193485729,0.2029277384,0.5645589828,0.0536458157,0.1894836426,0.0351332389,0.3217778802,0.053703472,-0.5332477689,-0.1743085086,0.2165416032,-0.3087960184,0.1456948817,0.3105262816,-0.3313245475,0.3001628816,-0.1157548949,-0.3032845855,0.0032358915,0.1294727772,-0.1749177426,0.1020331085,0.0779607892,0.0931859761,0.1030509546,0.1904263943,-0.0799686313,0.0655035079,-0.0332771763,-0.0471923687,-0.4852254391,0.0255070888,0.1306854486,0.0716528967,-0.009506152,-0.0878667012,-0.0543031245,0.1962302923,-0.2217216939,-0.1117357835,0.112898834,-0.0255647469,-0.2713457048,0.2126425505,0.5540782213,-0.3136119843,0.2092669159,-0.1985261887,-0.1348921061,0.163162753,0.3628430367,-0.1167293563,0.0775886104,-0.3848822117,0.3266322613,0.6447759867,-0.5202516913,-0.507689178,0.2816505432,-0.2223585993,-0.0715679377,0.5067173243,0.2407118678,-0.0645311177,-0.113183476,0.0470005237,0.117008321,0.0467213541,-0.0885009691,-0.1114888489,-0.2450889647,0.3402366638,0.1130200699,-0.0306807663,0.1320907325,-0.029109424,0.1613193005,0.3716828525,-0.2910831869,0.1511864364,0.2840608954,0.3005669415,0.0372093096,-0.1425783336,-0.1196034774,-0.2946609557,0.2399859577,-0.0920302644,-0.0201166086,-0.3088808358,0.0002530748,-0.1130894944,-0.0409190543,-0.3876262605,-0.2783128321,0.1904964447,0.4527053237,0.227343753,-0.0196524505,-0.2899426222,0.3173918426,-0.4826749563,0.0027164356,-0.6354672313,0.1280632615,-0.0931880027,-0.1490348577,-0.1639639139,0.1639258713,0.0235497002,-0.0028483428,-0.092747651,0.212470293,-0.2380763143,0.0813123882,0.2638707161,0.1023620442,0.0382039137,-0.5709121823,0.3375821114,0.4717598259,0.2284216285,0.0894116163,-0.2693724036,0.1183070689,-0.0825481787,0.0903004631,0.0437282436,0.0172032174,0.4666255414,0.0148105649,0.0699819252,-0.212793991,0.3003255725,0.216489777,0.4621777833,-0.086125508,-0.5364867449,0.0861315429,0.2796010077,0.0605193675,-0.0986444056,-0.0109362863,-0.3510042727,-0.0416343696,0.0550630279,0.2289790064,0.2813903093,0.2278090268,-0.0653548166,0.0831116587,0.0983732715,-0.2621799409,0.1807291061,-0.2579271495,0.1012303159,0.2698712349,0.1998011768,0.0320530571,-0.0356441364,-0.5033630133,0.0721699521,0.4454903901,-0.270344466,-0.1728234887,-0.1369042546,-0.6854811907,0.0606506467,0.0763859451,-0.11054077,-0.3432247937,-0.0222284645,0.18776007,0.2145616412,-0.0719953775,-0.0950528979,0.0755412653,0.0686472133,-0.109409146,-0.419331491,-0.3149599433,-0.2077754438,0.046856489,0.2707952857,-0.0228072517,0.2469428778,-0.2802794874,0.0639284253,-0.4144373536,-0.0254895911,0.0234772526,0.0292966571,-0.0004744022,-0.0340683497,0.4818875194,0.4775179625,-0.0623232983,0.3475892544,0.0415477231,-0.1742302775,0.2578029335,-0.2152395099,-0.2344868481,-0.0418086126,-0.5249136686,-0.3563694656,-0.3282742202,0.0827530324,-0.1398809403,0.0803990811,0.3180901706,-0.1464771181,0.2062613368,-0.038883388,0.1958673,-0.0754591152,-0.2256141007,0.3382492363,-0.161397025,-0.2159953713,0.2360935062,0.0649059862,0.2100883275,-0.0412420146,-0.2306012511,0.1721257716,-0.2255693376,0.2844632864,-0.1305034161,-0.1519210339,0.3499826789,-0.0105040036,-0.001459451,0.0043800203,0.0118807135,0.1178084835,0.0100269215,0.571620822,0.2039424479,0.4134452343,0.1445199102,0.3723564744,0.4486645758,0.0267595891,0.3283901513,-0.2621395588,0.0669077039,-0.0658819005,-0.3156565428,0.1194543019,0.1747564822,-0.1733243912,-0.1079721004,0.1386827379,0.3889435828,-0.3315194845,-0.1330213994,-0.2874074876,-0.0854833126,-0.1463778317,0.0202502068,0.0111190295,0.1291114837,0.2093752027,-0.3425545394,-0.3424660563,-0.1121322811,0.4633776844,0.0023366199,0.1523140669,-0.2744363844,-0.2841807306,-0.5618222952,0.4806259871,0.1180058867,0.2120662779,-0.2758866251,0.1250002533,0.266346544,0.0225754622,0.3283257484,-0.3004835844,0.2563642263,0.0807866678,-0.0989301056,-0.2358802706,0.1696390212,-0.1501615494,0.198500365,0.335871011,0.1939541698,-0.160841018,-0.1110729948,-0.016944252,0.1083815396,-0.1749693602,-0.2175936401,-0.0177838448,-0.0037040839,-0.1634147614,-0.3284824491,-0.1715682,0.2472744733,0.1329616308,0.3358918428,0.0468720645,-0.0742730275,0.1784730852,0.2204153389,0.339287132,-0.0961351916,-0.1533091366,0.1652644426,0.2498364896,0.0717588365,0.4986754954,0.2898446321,-0.3622654378,-0.0141012259,-0.2284084857,0.1272356212,0.2903348804,-0.2587492168,0.1040537134,0.0343668945,-0.0078272121,0.0616949275,-0.0959052071,0.2513180673,0.0315440856,-0.1259422749,-0.3986914158,0.3272973299,0.1011688933,0.0039729285,0.4540026486,-0.146319136,-0.2276169211,0.0237343851,0.0572314076,0.9369785786,0.1299279779,0.2292374223,0.3930836022,0.1683117896,0.5260003209,-0.3035136759,0.3688016534,-0.3871545494,-0.0803950131,-0.0029479368,-0.1137043014,-0.0116645973,-0.1774899364,-0.0803141966,0.2340769768,-0.1563339084,0.0061115115,-0.0985916182,0.2628973126,-0.3349470198,-0.0904692635,-0.4606150687,0.0847525224,-0.2080034167,0.334980011,-0.1050385013,-0.1498461068,-0.3853043914,0.1126162186,-0.4202090502,0.3758969307,-0.0859740525,0.1225233302,-0.0736165866,-0.3276994228,-0.0553329438,0.6137100458,0.5807358027,0.050126832,-0.3391729295,0.0077568712,-0.2113013864,-0.0230903607,-0.0938924998,0.1553285122,0.3382046223,-0.0408551991,-0.3091776371,0.5142868757,-0.1125862598,0.016242044,-0.2366230637,-0.2566816807,-0.240296945,-0.0432238802,-0.1230145022,0.0199265126,-0.1297458559,0.0251795202,0.0997499153,0.0126655791,-0.2329041064,-0.2464849204,-0.1339305639,-0.1444867551,0.1830044985,0.325381577,0.0474451073,0.1761954576,0.6140019894,0.2216762155,-0.0501693785,-0.2481475621,0.0654041916,0.1556771398,-0.2556731999,-0.2123807222,0.1041428968,0.1746927202,0.0965442285,0.2774943709,0.1864701658,-0.0062570614,-0.1428339779,-0.4487022161,-0.5422468781,0.1517494321,-0.1925470531,0.1390153468,-0.1516945809,0.2596283257,0.1187472492,-0.2652499974,-0.3370279372,-0.0957017839,-0.3234106898,-0.005745166,0.3934482038,0.0590532944,0.1968781054,0.0436726399,0.1371066868,-0.0367001854,-0.2849012911,-0.1598269492,0.0488781109,0.1000243649,0.1371713132,-0.4027868807,-0.1165812388,-0.2882617414,-0.0881268978,-0.1604430079,0.0502879359,0.2985337675,0.0316240191,0.2532197535,-0.3133458495,0.0580079556,-0.2136869878,0.1573789269,-0.1528376639,0.184106335,0.0598614365,0.1261301488,0.017676102,0.0098836925,-0.5507247448,-0.0625745282,-0.0907989293,0.1829756051,0.1152569503,-0.4182234406,0.228056103,-0.1419947892,0.171419993,0.1962636858,-0.0333172902,0.2832373977,0.1811421067,0.1862851232,-0.2386663258,0.2086086422,0.2157798111,0.0946389586,-0.0938897207,0.0430795886,-0.0494546257,-0.025663266,0.2692067325,-0.1024496406,0.3867939115,-0.4072207212,-0.052489385,0.1194665208,0.0713919997,-0.0525237285,0.1153041348,0.3662636876,0.1645925641,0.7827575207,-0.2470326126,0.1465714276,-0.184984982,0.3281104267,-0.2487826943,-0.3345987499,-0.3359787464,0.1134815142,-0.2942698896,-0.1261737496,-0.1852843314,0.2995412946,-0.1018136367,0.1427347064,-0.2357794791,0.3921541274,-0.1488359123,-0.0453949682,-0.5625841022,-0.2266335785,0.1058561429,-0.1797054708,0.0250908062,0.0159932282,-0.0091492422,0.1309443563,-0.2652380168,-0.4018542171,0.3093726933,0.4047755599,-0.0163565241,-0.2793292999,0.4106734395,0.2905892134,0.0659506842,0.2053315043,0.3192927837,0.4276020825,0.4022081494,0.0665072724,0.2000069916,-0.1069403887,-0.2528617382,-0.0138815902,0.0525235943,0.3014642596,0.0584575385,0.312531054,0.1497073919,-0.0062345145,0.1430208981,0.045691397,0.3462027311,0.0922235548,-0.0150208259,-0.1465478987,-0.2455284297,-0.1894638687,-0.1285670996,-0.558146894,0.1378033608,0.2128199339,0.215043202,0.2549575567,-0.0042081224,0.0490093678,-0.0665702969,0.1328429878,0.5250784755,-0.0340116769,-0.2703349888,-0.3209694922,-0.5026276112,0.1653779298,-0.0252851732,-0.420865953,-0.2485361844,0.1211538464,-0.1755242199,-0.2355565578,0.2410198301,0.1014041528,0.069594197,0.3392386734,-0.2268995792,-0.3327465951,0.2696409523,0.0531273633,-0.1709012687,-0.1728681773,-0.0946457535,-0.1486764699,0.0490003414,-0.0460385568,0.3904182315,0.0071133403,0.2057304978,0.1977957785,0.1998120844,0.4811964035,0.0680666938,-0.0404394753,-0.3731051385,-0.1117367521,-0.2012594044,-0.0415709764,0.3111200333,0.5081033111,-0.3583825231,-0.0083459811,-0.3815684319,0.1278301626,-0.1665936261,-0.0174836386,-0.2955942452,0.3703538477,-0.2835772932,0.2372785211,0.1641732454,-0.0455508903,-0.0290332139,0.1030434892,-0.3331671655,0.0036255466,0.5232225657,-0.2101983577,-0.2365404814,0.0900798738,0.1933183372,0.1135041565,0.0103466297,-0.3378685713,0.1756677479,0.1824224442,-0.0053519118,-0.2406205684,-0.0816917047,0.0652151182,0.1135519743,-0.0653273016,-0.1427070946,0.159398526,-0.0728523582,-0.0136100473,-0.0865384638]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/227","title":"Should we still have to force to install apache_beam to download wikipedia ?","comments":"Thanks for your message \ud83d\ude0a \r\nIndeed users shouldn't have to install those dependencies","body":"Hi, first thanks to @lhoestq 's revolutionary work, I successfully downloaded processed wikipedia according to the doc. \ud83d\ude0d\ud83d\ude0d\ud83d\ude0d\r\n\r\nBut at the first try, it tell me to install `apache_beam` and `mwparserfromhell`, which I thought wouldn't be used according to #204 , it was kind of confusing me at that time.\r\n\r\nMaybe we should not force users to install these ? Or we just add them to`nlp`'s dependency ?","comment_length":13,"text":"Should we still have to force to install apache_beam to download wikipedia ?\nHi, first thanks to @lhoestq 's revolutionary work, I successfully downloaded processed wikipedia according to the doc. \ud83d\ude0d\ud83d\ude0d\ud83d\ude0d\r\n\r\nBut at the first try, it tell me to install `apache_beam` and `mwparserfromhell`, which I thought wouldn't be used according to #204 , it was kind of confusing me at that time.\r\n\r\nMaybe we should not force users to install these ? Or we just add them to`nlp`'s dependency ?\nThanks for your message \ud83d\ude0a \r\nIndeed users shouldn't have to install those dependencies","embeddings":[0.2446198016,0.0130939912,-0.0891243443,-0.066920422,-0.1474457979,0.0496424697,0.0804633126,0.1371218711,0.4622112811,0.0668714494,0.3250715733,0.0682467148,-0.0905696005,-0.3348688483,0.0582655594,-0.2800340652,0.2024766505,0.4069409966,-0.2649184167,-0.0817563534,-0.2541305721,0.1852510124,-0.3694931567,0.0649296045,0.0125303548,0.0580261461,0.130345121,-0.0985765904,-0.1722903252,-0.1781694442,0.0866206661,0.1242235526,0.0225401968,-0.2248182148,-0.0001165492,-0.2372857928,0.5186698437,0.1110714749,-0.2589124739,0.1537023038,-0.239737004,-0.3942042887,0.2031645477,-0.2616474628,0.1686137021,0.0332645625,0.2613850832,0.0664162561,0.4339972734,-0.2898262143,0.1988468915,-0.0019789319,0.234162733,0.1456386596,0.5924622416,0.1768086255,-0.0879393741,-0.0517001376,-0.0034399324,0.041547291,-0.052085802,0.3426978588,0.0136886518,-0.1642922312,0.3849707544,-0.2119929194,0.177713573,-0.4257761836,-0.2198912352,-0.0529488735,0.5435855985,0.3380150497,0.154843986,-0.0576722249,-0.123079434,0.2515039146,0.3171243072,0.2080609649,-0.2463086545,0.018080974,0.0341095738,-0.6527433991,-0.0025674929,0.422757715,0.075358063,0.5232874751,0.122743234,0.1802853048,0.2645194232,0.082767725,-0.2792278528,-0.1966829896,0.0143069373,0.1617138684,0.4230162799,-0.1276502013,0.0116472449,0.5010073185,-0.0214173663,-0.0452334844,-0.7031511068,0.0555360466,-0.0603572167,0.0119941011,0.1162712276,-0.1062542051,0.1352649033,-0.2602572441,0.3755447567,0.0574286282,0.0144630894,0.278780967,0.3960641325,-0.0833910629,-0.4671495259,0.1172726452,-0.0873062983,-0.237896651,0.2881557643,-0.0628644973,-0.3532311022,-0.2149534076,-0.2277917415,-0.1005704999,0.1020648032,0.6225602627,0.2566706836,0.0548897162,-0.1560446024,-0.3090133369,0.163907662,0.062131919,-0.4238885641,0.4711208344,0.3289444447,0.1117802784,0.2113279998,0.1240838245,-0.1704949588,0.1371936351,0.0083208289,-0.0181670059,-0.185092628,0.176097095,-0.0318739638,0.4335996211,-0.1562292725,0.042909652,-0.1212239191,0.2367304265,-0.0830687657,-0.1185167134,0.2857126594,0.145342201,-0.0649793372,0.0668248385,-0.0815040171,-0.0845051333,-0.061819654,-0.2596541941,-0.1117011234,-0.0808025077,-0.0608834624,-0.2068494409,-0.2194053084,0.2934026718,0.1014209017,0.0098169828,-0.0945713967,0.0297964029,0.0487501658,0.257104218,-0.4123511314,0.3413085043,-0.0987464339,-0.0994859934,0.5789161921,-0.3160024881,-0.0694689527,0.2186360061,0.0945573002,0.2285039872,0.1132990867,0.0397056341,0.1196055114,-0.0948324129,-0.3678954244,0.3329326808,0.1418384761,-0.016634576,-0.4066388905,-0.2610394657,0.3380825818,0.0057817367,0.2940638363,0.1383563876,-0.0389516428,0.6986958981,0.446233958,-0.1549313813,-0.0425503068,0.0576422028,0.0211352725,0.0393828973,0.2348136753,0.0202226564,0.0625482947,-0.0691960156,-0.2666794062,0.2564023137,0.1397589594,-0.2002720386,-0.1532441229,-0.2887431085,0.078099452,0.0189801399,0.1134133786,-0.1951522976,-0.0376567617,0.1301323175,0.0058298395,0.1070665792,-0.0207431428,0.0369384289,-0.1841975451,0.0052531757,-0.5288869739,0.2307525575,0.0157016944,0.0768733323,0.1314488202,-0.0023457897,0.1089648753,-0.1929395348,0.0156098967,-0.0579310209,-0.1075209677,0.0042477008,0.2274409384,-0.4918248057,0.1675375402,0.1051742509,0.056562528,-0.1659292877,0.1011516005,-0.1956137121,0.2557918727,0.1877886951,-0.0357757248,0.1361494064,-0.0469180755,-0.1118580028,-0.1198473051,-0.1624366641,0.075458698,0.3319222033,-0.2390035242,-0.0334404595,0.1907775104,-0.1631901264,0.3353628814,0.0328630581,0.0293252133,0.1115239784,-0.0960061103,-0.1292257607,0.1719414145,-0.4052529335,0.1307689399,0.2882908583,-0.0114805968,-0.078628324,0.0757556185,-0.2480688095,0.2207279354,0.2572032213,0.3559134603,-0.3287228346,-0.2230074555,0.0831404328,-0.060021922,0.1009534225,-0.1240116358,-0.1793875694,-0.1860514581,-0.2958185375,-0.1255558431,-0.4677653909,-0.1901199818,0.1161759943,-0.4890214205,-0.4144528806,0.1601338685,0.1484371275,-0.0110522518,0.178822875,0.2163344771,0.4716217518,-0.3596199453,0.1400886029,-0.5110132694,-0.0373138674,-0.3928462863,0.1050539389,0.1853459477,0.0670074597,0.272188127,0.0709460154,0.1367011219,-0.2216336876,-0.4155196846,0.0186279733,-0.1277099103,-0.1634066701,-0.0607512631,0.1858847439,0.0178690311,-0.238449797,-0.0280932169,-0.2246584445,0.0529071577,-0.1814700663,-0.0386308916,0.0223998632,-0.2050188035,-0.3306724131,0.0039922325,-0.2147909254,0.092685543,0.3615141213,0.102489002,0.0860062391,-0.2389694601,-0.3309353292,0.0659217983,0.2130295634,-0.0798818916,0.3565091789,0.2534737885,-0.1688145846,-0.401096344,0.0157650225,-0.1169241667,-0.1218054444,0.2998740673,-0.4411546886,0.1277819574,-0.0395987146,-0.057521645,-0.0025908982,0.0858235732,0.3900952041,0.045499865,-0.0928252637,0.0953293368,-0.0585943423,-0.0198036414,0.0542747751,0.4455971122,0.121629931,0.0269674547,0.177009955,0.8922283053,-0.1040067896,0.3158734739,0.1370339841,0.5298061371,0.1785096079,-0.004610131,-0.0704985708,-0.0204652026,-0.0492254198,0.1967444718,0.0551157407,0.0332571454,-0.4209341109,-0.4648382366,-0.1266326904,-0.1584310383,-0.2156068981,-0.0170078985,-0.0173098203,0.1789842844,0.2620748878,0.0320588946,0.1746374369,-0.2196478099,0.2934723794,0.3544837832,0.2508067787,-0.0463892408,0.2875230312,-0.0454592556,-0.1654664278,0.05701226,0.0553821698,0.0869144797,0.0495314635,-0.0600033216,0.1176918298,-0.010628297,0.2261506468,-0.1871011555,-0.0668707639,0.047124479,0.332108885,-0.4427402914,-0.0658470988,-0.0763424858,-0.213191554,0.5812987089,-0.0607256405,-0.1488045752,-0.1186398566,0.2828840017,0.1159314811,-0.2748245299,-0.430763334,-0.2337029576,-0.0551129431,-0.2714946568,-0.0633526519,-0.0634099245,0.0364792757,-0.0750235245,0.0900689736,0.0579039715,-0.1780055165,-0.0336088948,-0.1737173796,-0.1703972518,0.2048915327,-0.4096993804,-0.176509425,0.235672906,-0.6182338595,-0.0288391467,0.2542042434,0.2856945992,-0.238024801,-0.1891553551,-0.0040508392,0.3939364851,-0.1153890043,-0.024144737,-0.0866590068,-0.161577329,0.0399780646,0.0448180698,0.098074697,-0.4639957249,-0.164118588,-0.0353063717,0.5154393315,0.0459496044,-0.1951599866,0.1463132799,0.3284173906,-0.2365347594,0.302192688,0.2705010176,1.2291427851,0.0155105339,0.1423094124,0.101213485,-0.043369133,0.8882835507,-0.6015267372,0.1801952422,0.0029766182,0.3330521882,-0.0866774619,-0.0878940299,0.3064578474,0.0066888719,-0.2399189621,0.3944647908,0.403222084,0.0860352516,0.0523828343,0.5380289555,0.0843236297,-0.2775313556,-0.0443431027,0.099770382,-0.0348905548,0.7091772556,-0.2842359245,-0.1486795843,-0.0492713191,-0.4032531083,-0.1010501087,0.1945832968,0.0919675678,0.4346583486,-0.5558318496,0.0163180269,0.2917838395,0.0542022064,-0.0252567586,0.3777986765,-0.2626724839,0.0872949883,-0.2574772537,-0.3481225669,0.1059801504,0.2926985025,0.2376554012,-0.3273339272,-0.0859305412,0.0467761382,-0.3025826514,-0.2863506973,-0.3629891276,0.0050918111,0.0860686153,0.2504910231,-0.1019057855,0.5315440893,-0.2387700975,0.2002315372,0.1866994798,-0.2683489621,0.1310982853,0.4128800929,-0.0071879444,0.0048525827,-0.0240440667,0.342780143,0.1026680246,-0.3707816303,0.1143002659,-0.0895528868,-0.3334814906,-0.2638309598,-0.1910144538,-0.6437434554,-0.2724263072,-0.4147028625,0.1461161077,0.1352545768,-0.2878035903,-0.0407136753,0.2257370204,-0.0071638012,0.189378351,-0.1936937422,-0.0854522735,0.1990075111,-0.2948545516,-0.2889157534,0.2397874743,-0.04907557,0.4777294397,0.3120134473,-0.2727058828,-0.0493969247,0.1292632222,0.1806970388,0.294585228,0.0004275755,-0.2289635837,0.1078893021,0.0627639815,0.0100498125,-0.0433502086,-0.1735381484,0.2289326638,0.140783295,0.3668328524,-0.0233540442,-0.0409975126,-0.1372997761,0.2401371002,0.2020289004,-0.1817909479,-0.2090060115,-0.1237656474,0.1871080101,0.1636610031,0.4585908949,-0.3604903817,0.3520172238,-0.2508359253,0.3677151501,0.1196492985,0.0093327351,0.5427495837,-0.0149291474,-0.5152511597,-0.0684627667,-0.0860340744,0.2307449579,0.1279824078,0.0226062462,0.0153334606,-0.026154289,0.0493677072,0.0912937075,-0.1244275719,-0.2255239636,0.3153662086,0.1334530264,-0.1067404896,0.2748770416,0.4646950364,0.1691553742,-0.3824142516,0.1444131285,0.3613497615,-0.2431235462,0.189331606,0.1667882502,0.0147403004,-0.1201449409,0.0828158483,-0.2815381289,0.1872271001,0.0599794202,-0.162145853,0.049096711,-0.0569435172,0.0716157109,-0.1114521697,0.1538869888,0.261572957,0.1096606478,-0.0109382384,0.232075423,0.1380886883,0.4898463488,0.3151062727,-0.0652199909,0.2032988667,-0.2082937658,0.3104665577,-0.3804796636,-0.276088953,0.1167013049,0.0131286187,0.3409247994,-0.1288496405,-0.1042587087,-0.0871039703,0.1235693246,0.01902318,-0.2292315513,-0.2156100869,0.3477909863,-0.1850703508,-0.2419151962,0.311971277,-0.2020308673,0.3866702318,0.0511943437,-0.2368149012,-0.4035667479,-0.0742183477,0.0921366811,0.2079989314,-0.1635267287,0.4188520908,-0.1443609595,-0.198236376,0.0839695558,0.0288205836,0.007113432,0.2555744946,0.3519836068,0.3192527592,0.2225464731,0.1221299767,-0.1700528711,-0.2491873056,0.0848912373,0.0179590471,-0.2049821764,0.0714939088,0.2450677603,-0.2589274049,0.0158887133,0.0952222645,-0.272526741,0.2246219814,-0.1802003086,0.0876030847,-0.0518097021,-0.5503715277,0.0513026528,0.0626696274,0.4043436646,0.2720525563,0.2342482507,-0.3204140365,-0.3471774757,-0.3423415422,0.0302632898,-0.4476004243,-0.0679915771,-0.0646424964,0.109714523,0.0065281279,0.4196603894,-0.295537889,0.0911133215,-0.0304729193,-0.1518169045,-0.1221902966,0.221159324,0.2861980498,-0.0551341847,-0.4457647502,-0.0230171289,0.2051029503,-0.0803489983,0.0591469929,0.0075972262,-0.1710661948,0.2732309997,0.4684543908,0.3485272229,0.1351471692,0.4676908255,0.0979134887,-0.1857223958,-0.0297829024,-0.0368849225,-0.1427031308,-0.063441202,-0.1957740486,-0.0561565273,-0.0928153992,0.2216898501,-0.2810493112,0.5262457132,-0.1247942075,0.4991675913,-0.3192785978,-0.3337046206,-0.0409453847,0.2665943205,-0.0474797562,0.0599222071,-0.2368090302,0.0375283957,-0.6649406552,-0.0232288279,0.0744399726,-0.307040751,-0.1072660387,0.0720021054,0.4139966369,0.0603642575,0.0124818385,-0.3473639488,-0.3968195021,0.1919914037,-0.0101176891,0.1247980446,0.0852710605,-0.0880123898,-0.4133383334,-0.1577422023,-0.4007852376,-0.0660069808,-0.2637928724,-0.2826306522,-0.2877759635]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/227","title":"Should we still have to force to install apache_beam to download wikipedia ?","comments":"Got it, feel free to close this issue when you think it\u2019s resolved.","body":"Hi, first thanks to @lhoestq 's revolutionary work, I successfully downloaded processed wikipedia according to the doc. \ud83d\ude0d\ud83d\ude0d\ud83d\ude0d\r\n\r\nBut at the first try, it tell me to install `apache_beam` and `mwparserfromhell`, which I thought wouldn't be used according to #204 , it was kind of confusing me at that time.\r\n\r\nMaybe we should not force users to install these ? Or we just add them to`nlp`'s dependency ?","comment_length":13,"text":"Should we still have to force to install apache_beam to download wikipedia ?\nHi, first thanks to @lhoestq 's revolutionary work, I successfully downloaded processed wikipedia according to the doc. \ud83d\ude0d\ud83d\ude0d\ud83d\ude0d\r\n\r\nBut at the first try, it tell me to install `apache_beam` and `mwparserfromhell`, which I thought wouldn't be used according to #204 , it was kind of confusing me at that time.\r\n\r\nMaybe we should not force users to install these ? Or we just add them to`nlp`'s dependency ?\nGot it, feel free to close this issue when you think it\u2019s resolved.","embeddings":[0.2748143971,0.0078004715,-0.0634392649,-0.0371804722,-0.1197935417,0.0317761004,0.0673677102,0.1615340412,0.4366588891,0.0271489527,0.3276711702,0.0695527494,-0.0811151192,-0.3630872965,0.1157696992,-0.2694873214,0.2364326566,0.3207990229,-0.2336010039,-0.088569209,-0.2709094584,0.229886353,-0.3910895884,0.0754870698,0.0434878394,0.08707726,0.1321543455,-0.1019102335,-0.1423687339,-0.1891032755,0.0769925341,0.1202062368,0.0297117084,-0.2323913425,-0.0001168005,-0.2491332889,0.5718727708,0.0994127169,-0.2416625768,0.1094475389,-0.1978023797,-0.3674424887,0.1811747402,-0.2417230159,0.1923616678,0.0450589508,0.2649121284,0.0944410786,0.4132536352,-0.2801751196,0.2157087624,0.0000203253,0.1834342778,0.1159586981,0.5696032643,0.1846647114,-0.0760828331,-0.0751974732,-0.004875408,-0.0019424642,-0.0998023823,0.3364252746,-0.0060818037,-0.1416127384,0.3725314438,-0.1917694062,0.1237814128,-0.3830720186,-0.1799664497,-0.0695522651,0.5519137383,0.3178074062,0.1560356319,-0.0483906083,-0.1274670511,0.2794784307,0.3661907911,0.2269649804,-0.2507709265,-0.0130480332,0.0766979456,-0.6399183869,0.0161945149,0.4132294059,0.0890448838,0.5311231017,0.1536768526,0.1875827014,0.2599598467,0.0840928331,-0.2706597447,-0.2099635452,0.0045354818,0.170250982,0.4037900269,-0.0637243092,0.0235906057,0.4871280193,0.0125224516,-0.038751822,-0.7044503689,0.0454189591,-0.0398939028,0.0081898933,0.1355768442,-0.0960431844,0.1462605149,-0.2394984663,0.3812178373,0.087448597,-0.0309701897,0.3052235246,0.3919187486,-0.0928936005,-0.4697985947,0.0789260715,-0.1153241917,-0.2775153816,0.2924650908,-0.0813761353,-0.3373028338,-0.2393512428,-0.1978448331,-0.0892564729,0.1319694817,0.6208693981,0.2763876319,0.0827938393,-0.1688168794,-0.3035110831,0.1669003814,0.0378399193,-0.4347311556,0.4764412344,0.3539639413,0.1235609204,0.1854469925,0.1074918881,-0.2011757344,0.0777369142,-0.0927033424,0.0023575407,-0.1988269687,0.1639906615,-0.0471120961,0.4525875151,-0.1603766233,0.0241677184,-0.135025844,0.2273339182,-0.0404379033,-0.1197024807,0.3018125296,0.1333943158,-0.0861556828,0.0567937121,-0.0638741851,-0.0516243689,-0.049093511,-0.2886353135,-0.1049350798,-0.0704436749,-0.0522596091,-0.2192856371,-0.1786438227,0.2962258458,0.1454265118,0.0051063956,-0.1259044409,0.0598391257,0.05857189,0.2522310615,-0.3887612522,0.2947293818,-0.1304460168,-0.0626030341,0.5455432534,-0.3060099483,-0.085085094,0.2093704939,0.0962843224,0.2324951291,0.0975483954,0.0160479657,0.0861520618,-0.1570855081,-0.3460661471,0.3663809597,0.1222263053,-0.0566235259,-0.4348639548,-0.2427128404,0.3333208263,-0.0264673289,0.3172929585,0.1549649686,-0.0659906641,0.677456677,0.4461002052,-0.0986688659,-0.0209038686,0.0564063601,-0.0149442991,0.0091479048,0.258041203,0.0077398974,0.0469126962,-0.0667845309,-0.2771674693,0.278865099,0.1382557601,-0.1874772608,-0.1566244513,-0.278814137,0.0987687707,-0.0057690223,0.1223855168,-0.1821679771,-0.0269590784,0.1565103829,-0.0097756814,0.0641517714,-0.0009433922,0.0329878069,-0.1815867424,0.0078847045,-0.4995204806,0.1881662458,0.0167792197,0.0603669509,0.1719312817,0.0261964705,0.0719121322,-0.1782757193,0.0018341256,-0.0293600801,-0.1148826033,-0.0197825022,0.2160892487,-0.4672567844,0.1289764196,0.0882611647,0.0281784758,-0.1957051903,0.0831944868,-0.1907754242,0.2582963407,0.2405159175,-0.0229772273,0.1651382297,-0.0737889186,-0.1067181602,-0.1021990553,-0.1157354489,0.1030794233,0.3315971196,-0.2680439651,-0.0570422485,0.1917720288,-0.1241113991,0.3429201245,0.0018557493,-0.0013174475,0.1709908545,-0.1097843722,-0.1084476486,0.1641055048,-0.3630203903,0.1244768351,0.2606999278,-0.0060551548,-0.0649372339,0.0413434803,-0.2312878221,0.2491803169,0.2768023014,0.3415308297,-0.3485076725,-0.1709125489,0.0767101049,-0.0479930192,0.1325669289,-0.1309074163,-0.1678110361,-0.201655671,-0.2644404769,-0.1778741628,-0.4829831123,-0.2163885981,0.1288011521,-0.5250201225,-0.4144893587,0.1584857404,0.1718190312,-0.052798707,0.172060594,0.1756826788,0.4450922906,-0.3825868964,0.1289535165,-0.5001274943,-0.0720856041,-0.4501284957,0.0973835588,0.2497394979,0.0619918481,0.270609349,0.0805037096,0.1142046079,-0.2180117816,-0.4173668623,0.0286251679,-0.1106673405,-0.1316116005,-0.0714051574,0.1463192105,-0.0492449217,-0.2415171415,0.0202688985,-0.2717556357,0.0266381502,-0.1411543041,-0.0287691299,0.021642521,-0.2259905189,-0.323395431,0.0165797379,-0.2359779775,0.0994062498,0.3211908937,0.0921243951,0.101377584,-0.2943655849,-0.3379602134,0.0803511888,0.1723923981,-0.0688349754,0.3617862463,0.2694005668,-0.1866010427,-0.4540376067,0.0118118562,-0.1060136259,-0.1320341676,0.2964220941,-0.4234122038,0.1523374766,0.0134571567,-0.0542510673,0.0125913052,0.0693890601,0.3725950718,0.0401908159,-0.0745327622,0.0977744386,-0.0527845658,0.0236982666,0.0507917851,0.4809034765,0.1484586447,0.0057266359,0.1909611523,0.8792365789,-0.1133417413,0.33551085,0.1563151926,0.516890347,0.1069930345,-0.024247108,-0.0805399045,0.0132935615,-0.0810872242,0.1936483234,0.1036282554,0.0671464801,-0.4288949072,-0.4611457288,-0.0701319277,-0.1638614088,-0.2607288063,0.0016191154,-0.0462076068,0.1789624393,0.296420157,0.040254306,0.1529657841,-0.2518259883,0.3095807433,0.3165799081,0.2746379375,-0.0570089705,0.2810706198,-0.0560701303,-0.1815083474,0.0575229675,0.0651270375,0.0955072865,0.0209432114,-0.0702489838,0.1012261361,-0.0346191749,0.2815839052,-0.1311762333,-0.0569475815,0.0534091406,0.3696132302,-0.4685488045,-0.0794667974,-0.0313796923,-0.2378507406,0.5764037967,-0.0916332528,-0.1282452643,-0.0789760053,0.2554793656,0.1224697009,-0.2621190846,-0.450838834,-0.178050831,-0.1059624553,-0.2195964307,-0.1030162796,-0.035997469,0.0416527577,-0.121032998,0.1487276107,0.0909844413,-0.2041527331,-0.0066563855,-0.1842706949,-0.139265433,0.1600509882,-0.3331230283,-0.1217199787,0.2656206489,-0.6402502656,-0.0395338647,0.27456972,0.2866266966,-0.2524844706,-0.1540360451,-0.0175737739,0.3891254067,-0.0875402763,0.0058354065,-0.0910438746,-0.1839067787,0.0003879282,0.0013028899,0.0903427079,-0.4618931413,-0.1812961698,-0.0287367646,0.5740661025,0.03422039,-0.2174896002,0.1583594233,0.309381485,-0.2457152009,0.3266430795,0.2053706795,1.2344312668,0.015669724,0.1246645674,0.1249638721,-0.079372324,0.9138650894,-0.5928034186,0.2070136815,-0.0262801591,0.3316482604,-0.0647452846,-0.0707642138,0.2740448713,0.0150998626,-0.2438926697,0.3777088523,0.4117679298,0.1072243601,0.0886734352,0.5057560205,0.0651552901,-0.2981933355,-0.0645387247,0.0990737975,-0.0932218954,0.7188455462,-0.2691364288,-0.1455814391,-0.0520078056,-0.3998077214,-0.1627568454,0.223431617,0.1278929561,0.4270605445,-0.4777855575,0.0211905837,0.2629927695,0.0513986312,-0.0539611578,0.3251715899,-0.2479368448,0.1480788589,-0.2121216804,-0.3141354024,0.0972623378,0.2825272977,0.2261869609,-0.3207364976,-0.0911298245,-0.0081072096,-0.2784562409,-0.2940298915,-0.403962791,-0.0129819242,0.1256658733,0.2691884041,-0.1051892638,0.503359139,-0.1900981963,0.1644203663,0.1918489784,-0.2513630986,0.1473794281,0.427470088,-0.0242891274,0.0643609464,-0.0078156823,0.3644704223,0.0765909106,-0.4198506474,0.1340643913,-0.007348395,-0.3390941918,-0.2455573678,-0.1971064955,-0.6696767211,-0.2827181816,-0.3450080454,0.0709241405,0.126584813,-0.2537822425,-0.0582630523,0.2259661704,-0.0650966465,0.1828828156,-0.2033688575,-0.0928099826,0.1999897808,-0.2837431133,-0.2849574983,0.1835133582,-0.0309998579,0.4848365784,0.3157718182,-0.2689302266,-0.084214434,0.1336892247,0.2092580497,0.2704539597,-0.0032700677,-0.2550079823,0.094474785,0.0604103804,0.0279655196,-0.0437363982,-0.1672390997,0.1982818395,0.1434854865,0.3524117768,-0.0550551713,-0.0341910459,-0.1440791786,0.2693960369,0.2028776109,-0.2432800084,-0.2050202489,-0.1360799521,0.1440088451,0.1447623968,0.4497365355,-0.3072914779,0.3590389192,-0.2563644648,0.4427172244,0.1592887789,0.0214702971,0.5551270843,-0.0187354777,-0.5270379186,-0.0959492624,-0.110794276,0.2158040255,0.1660255045,-0.0141792381,-0.0028298432,-0.0308025833,0.0766491443,0.0935910866,-0.1152695715,-0.2156375796,0.3363888562,0.1386354417,-0.1247162297,0.318706125,0.4836500287,0.2141149342,-0.3869612217,0.1218713745,0.3173529208,-0.2686143517,0.1576929837,0.1700649858,0.0700979307,-0.1348339617,0.1206531972,-0.306355387,0.1908868402,0.052469369,-0.2053644061,0.0196355972,-0.0257396717,0.1054220498,-0.0618960075,0.1307540834,0.2620040178,0.1225056425,0.0453500934,0.2071783543,0.1152841821,0.5003185272,0.3322413266,-0.0215930548,0.2097761333,-0.1585236639,0.2627293468,-0.39336887,-0.3044616282,0.1574279368,0.023196198,0.2831748426,-0.1705623567,-0.124156177,-0.0749957338,0.1786723286,0.0120998295,-0.2232909799,-0.1855235398,0.3793740869,-0.1646424383,-0.27691257,0.2858635783,-0.2221413106,0.4177688956,0.0682717934,-0.2273440957,-0.367102772,-0.0888399333,0.1416286677,0.2020217925,-0.1404814571,0.4336936772,-0.1995248199,-0.2037180066,0.0558524616,0.0599392727,-0.0056593306,0.2893559337,0.372063905,0.3246130645,0.2127655745,0.113627173,-0.1694768369,-0.2452015132,0.0910462439,0.0019636755,-0.1718837917,0.1048373282,0.1943781674,-0.2675011158,-0.0115736285,0.0756610483,-0.2657034397,0.1868653148,-0.1798228323,0.1044487432,-0.0376538485,-0.5285056233,0.0485298485,0.0721229315,0.3763345182,0.2497669756,0.2486885786,-0.3022731841,-0.3652233481,-0.308398366,0.0496143028,-0.4223908782,-0.0634882078,-0.0745351985,0.1227472946,-0.0083761737,0.3661311567,-0.2718897164,0.1329461634,-0.0302335471,-0.0775821358,-0.1315026879,0.1607951522,0.2535859048,-0.0770869851,-0.4476134479,-0.0155834248,0.1911858022,-0.1051224545,0.0563082844,0.017177999,-0.2014185637,0.3026939929,0.4592659473,0.3531355858,0.1757409275,0.4753050804,0.1130645052,-0.2138544023,-0.0827318579,-0.0734594688,-0.1349764615,-0.0603691526,-0.231573686,-0.0387806818,-0.1251840591,0.2414177656,-0.3016349971,0.5039690733,-0.1549558491,0.4535832107,-0.3581418693,-0.3322253525,-0.086155653,0.3075013757,-0.0366064012,0.0736202598,-0.2472952455,0.0559724122,-0.6575323939,-0.0695156828,0.1145542637,-0.3766417503,-0.1496248543,0.1368951052,0.3914020658,0.0296699759,0.0084746042,-0.3610907495,-0.4737573564,0.1666428447,-0.0022621241,0.1435267031,0.0228989758,-0.0607574917,-0.4421140254,-0.1326929927,-0.3891044855,-0.0339758918,-0.2826497853,-0.2425036281,-0.2527907193]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/227","title":"Should we still have to force to install apache_beam to download wikipedia ?","comments":"It should be good now :)","body":"Hi, first thanks to @lhoestq 's revolutionary work, I successfully downloaded processed wikipedia according to the doc. \ud83d\ude0d\ud83d\ude0d\ud83d\ude0d\r\n\r\nBut at the first try, it tell me to install `apache_beam` and `mwparserfromhell`, which I thought wouldn't be used according to #204 , it was kind of confusing me at that time.\r\n\r\nMaybe we should not force users to install these ? Or we just add them to`nlp`'s dependency ?","comment_length":6,"text":"Should we still have to force to install apache_beam to download wikipedia ?\nHi, first thanks to @lhoestq 's revolutionary work, I successfully downloaded processed wikipedia according to the doc. \ud83d\ude0d\ud83d\ude0d\ud83d\ude0d\r\n\r\nBut at the first try, it tell me to install `apache_beam` and `mwparserfromhell`, which I thought wouldn't be used according to #204 , it was kind of confusing me at that time.\r\n\r\nMaybe we should not force users to install these ? Or we just add them to`nlp`'s dependency ?\nIt should be good now :)","embeddings":[0.2472063452,0.0186057873,-0.0738593638,-0.064613305,-0.1378612965,0.0391498096,0.051202476,0.1830706,0.4381152689,0.0200804956,0.3018423319,0.0314680673,-0.1004439741,-0.3616685867,0.1208752915,-0.2720413208,0.2450285703,0.3336304128,-0.219149515,-0.0692702606,-0.2652941048,0.2096895725,-0.3791621029,0.0740171969,0.0415063351,0.088495791,0.1096919477,-0.1205175146,-0.1538891494,-0.1848382056,0.057132192,0.1392800808,0.0467153601,-0.2489881665,-0.0001192447,-0.2453527153,0.566367507,0.1133624911,-0.248910293,0.1653359532,-0.1963997632,-0.3990516663,0.1940122545,-0.2378218025,0.2101906538,0.0187364612,0.256421715,0.0929562822,0.4043541551,-0.2825421691,0.1884855628,0.0026967223,0.1707480848,0.1388926506,0.5972271562,0.2003780305,-0.088599138,-0.0850463957,0.0156326257,0.0136661008,-0.0710509792,0.3158338368,-0.0310829673,-0.1514919847,0.3600970507,-0.1817032546,0.1464955658,-0.3786808848,-0.224596262,-0.053970091,0.5465427637,0.328669101,0.1476625502,-0.0856468752,-0.1124154478,0.2535328269,0.3269347847,0.1881121248,-0.2405373752,0.0046198443,0.0789670274,-0.6705468297,0.0121832434,0.4482376873,0.136829868,0.564463973,0.160164997,0.1796644926,0.2905110121,0.1056556031,-0.2540817261,-0.1890245527,-0.003687728,0.1587104797,0.4213813245,-0.1052331254,0.0177143048,0.5076078773,-0.0135938739,-0.0470340028,-0.7063977718,0.0466062687,-0.0518937036,0.0018859348,0.141389221,-0.1225820556,0.1322128773,-0.2369433343,0.4035377204,0.0970229581,0.0175646506,0.3204857111,0.4075762331,-0.0728022978,-0.4788781703,0.0908933207,-0.1060164124,-0.2539862394,0.3088773489,-0.0819411054,-0.331479758,-0.2454436421,-0.2062579989,-0.1128599271,0.1062674075,0.6253712177,0.2567373812,0.0531763062,-0.1459178329,-0.3201905489,0.1787190884,0.0465443097,-0.4307447076,0.4746362269,0.3522799611,0.1237772331,0.1938075125,0.0891929492,-0.2102387398,0.0927686617,-0.1032044441,0.0160436798,-0.2041494995,0.1471203864,-0.0578598939,0.4377883375,-0.173197329,0.0116610043,-0.1552034169,0.2005246729,-0.0768003836,-0.1259854287,0.2713133395,0.1096292734,-0.0860721767,0.0714391395,-0.0940254405,-0.0360783041,-0.0534680821,-0.3158332109,-0.0921980888,-0.0564010777,-0.0401551351,-0.1826924533,-0.1817854345,0.2934085727,0.1292048246,0.0069372309,-0.1437687129,0.0681938455,0.0561935417,0.2744719982,-0.3825570047,0.3251407444,-0.1106820777,-0.1004387289,0.5770716667,-0.3033978641,-0.0693861768,0.1997365654,0.1280519068,0.2140218168,0.1060887948,0.0344589576,0.0916808248,-0.1366193742,-0.378264308,0.335580349,0.1257877499,-0.0731414258,-0.419351995,-0.2671109736,0.3513250351,-0.0149361156,0.3109129667,0.1444948167,-0.0633926019,0.7235403657,0.4343249202,-0.1053515077,-0.0415500849,0.0178813599,-0.0299832188,0.0086953556,0.2565057576,0.0042891144,0.028841598,-0.0712836087,-0.27643314,0.2418263406,0.1454981416,-0.2061527818,-0.1220118925,-0.2938793004,0.1162699088,-0.0171234645,0.0972493663,-0.193753913,-0.029813569,0.1520168334,-0.0313926153,0.0473691449,-0.0027682364,0.0339803956,-0.1978851855,-0.0112222815,-0.4956209362,0.2196675092,0.0027222743,0.0837604553,0.1273443997,0.0096214861,0.104096204,-0.1850111485,-0.0137436381,-0.0388323702,-0.1006382182,0.0016975615,0.2164263874,-0.4862236083,0.1259864718,0.0990565568,0.0137156127,-0.1895494312,0.0843128115,-0.2167939544,0.2599415183,0.2318005562,-0.0356461592,0.1439863294,-0.0891140029,-0.0930937007,-0.1120578945,-0.1287517548,0.1169466376,0.3432826996,-0.2719149888,-0.0575503819,0.2009168416,-0.1678021997,0.3501737714,-0.0089386525,0.0043990496,0.1593509763,-0.0730914101,-0.1381716132,0.1698401421,-0.3899224997,0.1115431339,0.2466763258,-0.0304211602,-0.0495139286,0.0363514051,-0.2332179099,0.2370179892,0.2992062867,0.3467028439,-0.3441755772,-0.2122264951,0.1014564633,-0.0477966219,0.0893947259,-0.157044664,-0.1331450343,-0.1893628389,-0.2627694011,-0.1612012833,-0.5017128587,-0.1973694861,0.1363632381,-0.4952442646,-0.4255935848,0.1450634301,0.1548395455,-0.0336986557,0.1825169772,0.1794682145,0.4639363289,-0.3777204752,0.1261137277,-0.542375803,-0.0518979132,-0.4316887558,0.0773489028,0.2343160361,0.0505846143,0.2453637868,0.0642766207,0.0903228968,-0.2346888483,-0.4218876362,0.01335245,-0.1053596735,-0.1423488557,-0.0576094538,0.1327813715,-0.0513914004,-0.259075284,0.0027453287,-0.2612827122,0.0341093354,-0.1309790164,-0.0555325747,0.0318196081,-0.2087814659,-0.3081711829,-0.0086352527,-0.2240538895,0.0959179029,0.3233079314,0.0781163648,0.1158988029,-0.2953200936,-0.3134902716,0.0656502768,0.180085808,-0.0570785366,0.3666347861,0.2821495235,-0.16459921,-0.4026023448,-0.0414841957,-0.1106568053,-0.1463215053,0.297696203,-0.4485727251,0.1502521634,0.0022721663,-0.0592360571,0.021311488,0.0803058594,0.3689055741,0.0549507998,-0.0670422241,0.1487596333,-0.0415369309,-0.0357206166,0.0630926117,0.4806652069,0.1380686909,0.0329425186,0.180999428,0.8747443557,-0.1260405779,0.3497968316,0.1562542468,0.517224133,0.1430882961,-0.0186426509,-0.0848499686,0.0223338697,-0.0669321567,0.1930594444,0.0685657635,0.0425925851,-0.4229611754,-0.4593239725,-0.0530653894,-0.1678703874,-0.2274362892,-0.0115208644,-0.0485618003,0.1910009384,0.2854011357,0.0536581613,0.185604915,-0.2787999809,0.3526863158,0.3088949919,0.2809146047,-0.0234971512,0.2731418908,-0.0577642433,-0.1702924073,0.0690130591,0.0719229504,0.0951800719,0.0528830662,-0.0469850227,0.0803720132,-0.0073588649,0.2389998883,-0.1361490488,-0.0717015564,0.0664934963,0.3729545772,-0.466278851,-0.0932268798,-0.043419417,-0.2526747286,0.5902059674,-0.066320166,-0.1502955705,-0.0815399364,0.3016233742,0.1536295116,-0.2843514979,-0.4219278395,-0.1836749017,-0.0999058187,-0.2233143896,-0.1010538489,-0.0690705925,0.0361164324,-0.1013473123,0.1283825338,0.066457741,-0.211840868,-0.0062986328,-0.1690768003,-0.1600365937,0.18660748,-0.3652723134,-0.141590625,0.2584387958,-0.6688750386,-0.0415033326,0.2870156765,0.2873589098,-0.2448557168,-0.1377452463,0.0013971173,0.3641349375,-0.1034053043,0.0292215701,-0.0864851028,-0.1481215209,0.0103531806,0.0314709432,0.1043556407,-0.4534614086,-0.2145400941,-0.0411948897,0.5942464471,0.0460875146,-0.2090319097,0.1541107744,0.3015604913,-0.2145303041,0.3127565086,0.2240789384,1.238273859,0.0131259942,0.1314777434,0.1184309423,-0.0489436314,0.9514651299,-0.6006151438,0.1969712824,0.0286416803,0.3458749354,-0.0718540624,-0.0497328974,0.2864045501,-0.023266701,-0.2216778547,0.4158082604,0.4392150939,0.1122086272,0.0698517114,0.5271015763,0.1145871952,-0.2936191857,-0.0465885848,0.0634236038,-0.098439239,0.7136672139,-0.2723886371,-0.1316305101,-0.032642182,-0.3688149452,-0.1523054987,0.1877704859,0.1210836843,0.4396972358,-0.4917708635,0.0046417611,0.2532221079,0.0757499039,-0.0496149845,0.3545223475,-0.2622582316,0.1043488756,-0.2429852933,-0.3269899189,0.1120721623,0.2780137658,0.2204815596,-0.3137126863,-0.0755053535,0.0342451967,-0.259191066,-0.3356886804,-0.38823089,-0.0397152528,0.1357142329,0.3033197522,-0.0845744982,0.523496151,-0.2105595767,0.1649869233,0.1659919769,-0.2544507086,0.1530148536,0.420901835,-0.0157096311,0.0548037104,-0.0214468539,0.3505286872,0.082891874,-0.4094794989,0.1152559519,-0.0281367302,-0.3119941652,-0.225326106,-0.1953686029,-0.6666992903,-0.2826287746,-0.3667685688,0.0819659755,0.1188484132,-0.2680344284,-0.0641478524,0.2502600551,-0.0235632043,0.1783081293,-0.1770825386,-0.1167900711,0.2355379164,-0.2506605685,-0.2948665619,0.2354711741,-0.0402340144,0.458129108,0.3172661364,-0.2369950116,-0.1034680381,0.1513505578,0.2064378113,0.2817827463,0.0148974024,-0.2662698925,0.1176661551,0.0405256674,0.0405793004,-0.0240223836,-0.1387639344,0.175786525,0.1593704969,0.3592247069,-0.0322534852,-0.0243956614,-0.1321640164,0.2703487873,0.182383731,-0.2759569287,-0.226425454,-0.1016083062,0.1856765151,0.1693298668,0.4331024885,-0.3252469003,0.3601582944,-0.2429873347,0.4308718145,0.159759447,0.057720568,0.5806457996,-0.0192566216,-0.5497182012,-0.0784763917,-0.1238122359,0.2070155442,0.1543318182,0.0270362142,0.0148701174,-0.0395463519,0.081565775,0.1134267002,-0.0902717188,-0.2350678593,0.338252306,0.1025031433,-0.1270905733,0.2942824066,0.4622237384,0.2414018214,-0.3845683038,0.1346678734,0.3119357824,-0.2605312765,0.1719597131,0.1401409656,0.0357871503,-0.1440440863,0.1166485548,-0.3025191426,0.2119766772,0.0560965352,-0.2057952583,0.0103471968,-0.0388872586,0.0992154777,-0.0717543587,0.1470499486,0.2564660013,0.1418765336,0.0157331936,0.2346607298,0.1114373133,0.4937179983,0.3288315237,-0.0094373021,0.2088294178,-0.1769491285,0.286896497,-0.3880236149,-0.2993973196,0.1378220767,0.047819633,0.2730929255,-0.1470289528,-0.0969229639,-0.075537622,0.1830147952,0.0225667376,-0.234015733,-0.1820381731,0.3611768186,-0.178880617,-0.2737088799,0.330653429,-0.2333948314,0.4402555823,0.0729690269,-0.2443786561,-0.3939413726,-0.0951759145,0.1100308225,0.2257938087,-0.1717655063,0.4276257157,-0.1913049668,-0.180676356,0.0594646856,0.0527510345,-0.0028008707,0.2873922586,0.3664677143,0.3048979342,0.2117294967,0.0978022814,-0.1483484507,-0.2264681906,0.0761365294,-0.0136966165,-0.1820750386,0.1285013556,0.2333121747,-0.2943252921,0.0168061294,0.1013150439,-0.2615515888,0.199432075,-0.1917427331,0.0962199643,-0.0497421809,-0.5233650208,0.0322953388,0.0861431882,0.3851620853,0.2781930864,0.2684805095,-0.2989846468,-0.3491475582,-0.3217816055,0.0264994651,-0.4009693265,-0.0564034395,-0.0716800839,0.0996445641,-0.017473124,0.4153852463,-0.267655164,0.1149416789,-0.033991944,-0.1237686649,-0.1024170816,0.1778185666,0.2382426709,-0.0490603298,-0.4593217671,-0.0193720255,0.1843339503,-0.1038650125,0.0248355698,0.0060639167,-0.2000221014,0.2997540236,0.4761345685,0.3499394059,0.1655303091,0.4272982776,0.1119098514,-0.2004213333,-0.1006661579,-0.0587176792,-0.110659793,-0.0696466416,-0.2329115719,-0.0587297641,-0.128278479,0.2367153615,-0.2892487347,0.5170767903,-0.1430669874,0.4641082883,-0.3638104796,-0.3358713686,-0.0533424616,0.2839073539,-0.0139990477,0.0458306894,-0.2388958782,0.072619237,-0.6618736386,-0.0484089367,0.0737429336,-0.3460507095,-0.1085151806,0.1046689004,0.4014132619,0.0822344571,0.0271102618,-0.36003986,-0.4979735017,0.185539037,-0.0020759362,0.1451871842,0.0447946712,-0.0501455739,-0.4365459085,-0.1455812156,-0.3939907849,-0.0384376533,-0.3021946251,-0.2507568896,-0.2719042897]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/225","title":"[ROUGE] Different scores with `files2rouge`","comments":"@Colanim unfortunately there are different implementations of the ROUGE metric floating around online which yield different results, and we had to chose one for the package :) We ended up including the one from the google-research repository, which does minimal post-processing before computing the P\/R\/F scores. If I recall correctly, files2rouge relies on the Perl, script, which among other things normalizes all numbers to a special token: in the case you presented, this should account for a good chunk of the difference.\r\n\r\nWe may end up adding in more versions of the metric, but probably not for a while (@lhoestq correct me if I'm wrong). However, feel free to take a stab at adding it in yourself and submitting a PR if you're interested!","body":"It seems that the ROUGE score of `nlp` is lower than the one of `files2rouge`.\r\n\r\nHere is a self-contained notebook to reproduce both scores : https:\/\/colab.research.google.com\/drive\/14EyAXValB6UzKY9x4rs_T3pyL7alpw_F?usp=sharing\r\n\r\n---\r\n\r\n`nlp` : (Only mid F-scores)\r\n\r\n>rouge1 0.33508031962733364\r\nrouge2 0.14574333776191592\r\nrougeL 0.2321187823256159\r\n\r\n`files2rouge` :\r\n\r\n>Running ROUGE...\r\n===========================\r\n1 ROUGE-1 Average_R: 0.48873 (95%-conf.int. 0.41192 - 0.56339)\r\n1 ROUGE-1 Average_P: 0.29010 (95%-conf.int. 0.23605 - 0.34445)\r\n1 ROUGE-1 Average_F: 0.34761 (95%-conf.int. 0.29479 - 0.39871)\r\n===========================\r\n1 ROUGE-2 Average_R: 0.20280 (95%-conf.int. 0.14969 - 0.26244)\r\n1 ROUGE-2 Average_P: 0.12772 (95%-conf.int. 0.08603 - 0.17752)\r\n1 ROUGE-2 Average_F: 0.14798 (95%-conf.int. 0.10517 - 0.19240)\r\n===========================\r\n1 ROUGE-L Average_R: 0.32960 (95%-conf.int. 0.26501 - 0.39676)\r\n1 ROUGE-L Average_P: 0.19880 (95%-conf.int. 0.15257 - 0.25136)\r\n1 ROUGE-L Average_F: 0.23619 (95%-conf.int. 0.19073 - 0.28663)\r\n\r\n---\r\n\r\nWhen using longer predictions\/gold, the difference is bigger. \r\n**How can I reproduce same score as `files2rouge` ?**\r\n\r\n@lhoestq \r\n","comment_length":124,"text":"[ROUGE] Different scores with `files2rouge`\nIt seems that the ROUGE score of `nlp` is lower than the one of `files2rouge`.\r\n\r\nHere is a self-contained notebook to reproduce both scores : https:\/\/colab.research.google.com\/drive\/14EyAXValB6UzKY9x4rs_T3pyL7alpw_F?usp=sharing\r\n\r\n---\r\n\r\n`nlp` : (Only mid F-scores)\r\n\r\n>rouge1 0.33508031962733364\r\nrouge2 0.14574333776191592\r\nrougeL 0.2321187823256159\r\n\r\n`files2rouge` :\r\n\r\n>Running ROUGE...\r\n===========================\r\n1 ROUGE-1 Average_R: 0.48873 (95%-conf.int. 0.41192 - 0.56339)\r\n1 ROUGE-1 Average_P: 0.29010 (95%-conf.int. 0.23605 - 0.34445)\r\n1 ROUGE-1 Average_F: 0.34761 (95%-conf.int. 0.29479 - 0.39871)\r\n===========================\r\n1 ROUGE-2 Average_R: 0.20280 (95%-conf.int. 0.14969 - 0.26244)\r\n1 ROUGE-2 Average_P: 0.12772 (95%-conf.int. 0.08603 - 0.17752)\r\n1 ROUGE-2 Average_F: 0.14798 (95%-conf.int. 0.10517 - 0.19240)\r\n===========================\r\n1 ROUGE-L Average_R: 0.32960 (95%-conf.int. 0.26501 - 0.39676)\r\n1 ROUGE-L Average_P: 0.19880 (95%-conf.int. 0.15257 - 0.25136)\r\n1 ROUGE-L Average_F: 0.23619 (95%-conf.int. 0.19073 - 0.28663)\r\n\r\n---\r\n\r\nWhen using longer predictions\/gold, the difference is bigger. \r\n**How can I reproduce same score as `files2rouge` ?**\r\n\r\n@lhoestq \r\n\n@Colanim unfortunately there are different implementations of the ROUGE metric floating around online which yield different results, and we had to chose one for the package :) We ended up including the one from the google-research repository, which does minimal post-processing before computing the P\/R\/F scores. If I recall correctly, files2rouge relies on the Perl, script, which among other things normalizes all numbers to a special token: in the case you presented, this should account for a good chunk of the difference.\r\n\r\nWe may end up adding in more versions of the metric, but probably not for a while (@lhoestq correct me if I'm wrong). However, feel free to take a stab at adding it in yourself and submitting a PR if you're interested!","embeddings":[-0.1457264721,-0.1697432995,-0.1972297877,0.3397462964,-0.1194976568,-0.4242601097,-0.420399636,0.1476738602,-0.270043999,0.1725059152,-0.0828669071,0.1803876609,0.120330058,0.0080935061,-0.1003900841,-0.0979746804,0.1656547487,0.0388290472,0.0924810246,-0.3720678091,0.066559352,0.5696008801,-0.0313320532,0.21043396,0.0540217571,0.2702380717,0.2851136923,0.3035303652,-0.1485799253,0.0284368172,0.1280016154,0.0306413826,-0.023698464,0.2221485972,-0.0001088679,-0.2347189039,0.1026096642,-0.0031664839,-0.2757230699,-0.3026088178,-0.0345568508,0.0643639937,-0.1229285002,-0.0703938454,0.0858464167,-0.044892218,0.0487728454,0.2095425725,0.4293982685,0.1475735754,0.2262544334,-0.3024373949,-0.1853493452,0.0240821615,0.5763342977,0.1626760215,-0.1144831404,0.4472635388,0.0056994664,-0.3526385427,-0.2401907593,-0.0085632736,0.3430406153,-0.1725598574,0.1325590014,0.1931548268,0.148585543,-0.045940455,-0.079535611,0.2447285503,-0.2702822387,0.0322115794,-0.2271714061,-0.2778382003,-0.0763156638,-0.1997605413,0.1168917641,-0.0814458728,0.1642860919,0.0122475997,0.0441323034,0.2954271436,0.0496390946,0.007164266,-0.0540564582,-0.1078455895,0.0508458652,0.0984653682,0.2952796817,-0.0505435802,-0.5163154006,0.0107727246,-0.2593673468,0.0749608874,-0.302280724,-0.0378559306,0.3144967258,0.46597296,-0.200871408,0.499627769,0.2691969573,-0.0934173465,-0.132207945,-0.0394466408,-0.1378237605,0.1081323028,-0.0289463978,0.0206935201,0.0881181508,0.2246808112,-0.133582443,-0.0120077161,0.3982224762,-0.5637685657,-0.4777997434,0.1025914326,-0.2610344291,-0.432397604,-0.6891934872,-0.0673610121,-0.2554311156,-0.1192424744,0.0989084914,0.2272011787,-0.174884215,0.2922985256,0.0467879772,0.1833910197,-0.289503634,0.0196620282,-0.1839990616,0.0635797158,-0.3143928647,0.1345241964,0.1705666035,0.1127464101,0.3180577457,0.2560533583,0.2735133171,0.0386443697,0.3298657835,-0.2839507163,-0.022634726,-0.2111162692,-0.211391747,0.197453782,0.0271293335,0.0918488353,-0.0087018358,-0.0748942494,-0.1808407903,0.1442367584,0.4666180313,0.2065799385,0.1496509761,-0.1105541289,0.0406925417,0.116938062,-0.2401628494,-0.5936908722,0.2040610909,-0.136552155,-0.1270114928,-0.0581774674,0.1098197326,-0.0345822014,0.0931209251,-0.1174313724,0.3731989563,0.0635824502,0.2603128254,0.4534834325,-0.0408312865,0.0118638165,-0.1628564298,-0.0708854869,0.4665318727,-0.4385315478,-0.2222961336,0.3354998529,-0.1650701761,-0.065185748,0.239046663,0.0392757431,-0.3886532784,-0.166007787,0.1774507761,0.1975248605,0.0777985081,-0.0483966693,-0.2970068455,0.134958744,0.3166256249,0.0654775649,-0.2318288237,-0.2993499935,-0.1114049256,-0.0643908754,0.5141965151,0.1328068972,-0.2751375139,0.1049917638,-0.2721597254,0.2936930954,-0.0516953692,0.118573375,0.4230363369,-0.0233754385,-0.7951256037,-0.1602763534,0.4370327294,-0.05248693,-0.2122911811,-0.4714090824,0.2405331284,-0.2394646406,0.2232830822,0.0550486632,0.1873215437,0.0303547215,0.0240709484,0.0966067314,-0.2364263833,-0.1586188674,-0.1715354323,0.2036774307,-0.2210983336,-0.0074328831,-0.0105401725,0.3596087694,0.3116413951,0.0875802562,0.0373043455,0.263418287,0.0501557477,0.0980502665,0.3672500849,0.1448055357,-0.0007641261,0.1889616251,-0.1226600558,0.6879813671,0.0164625514,0.03225106,-0.3400395513,0.4259150624,-0.2170387655,0.2006913573,0.4093107283,-0.0717375576,-0.0500977673,0.0577801615,-0.0067999209,-0.0171536021,0.7467250228,0.195269227,0.2502554655,0.0029566153,-0.3220258057,0.2596178651,0.399140805,-0.0153429508,-0.0208532233,-0.1199541613,0.1937142164,-0.0933883861,-0.0367203467,0.2507390678,0.3810696304,0.1951338202,0.1557074785,-0.1311991662,-0.338236481,-0.3438020647,0.0351328626,-0.182103917,0.0874467045,0.4963860214,0.1016074717,0.2162867785,-0.3020778,-0.5422854424,-0.2536484003,-0.2223142087,0.0613259934,0.1616452485,-0.1439131945,-0.3678351939,-0.1536603719,-0.0919036344,-0.2918887734,-0.2732761204,0.2300974727,0.0208752602,-0.0724995062,-0.2189383954,0.0232325289,0.1487477869,-0.3703945577,0.1783805788,-0.0173747577,-0.0280433036,-0.2795380056,0.2228176445,-0.0567981303,-0.0193044487,-0.0314762257,-0.279186964,-0.4508359432,0.1696880311,-0.3762677908,-0.0109601086,0.0646441653,-0.0032368647,0.0596726015,-0.2506172359,-0.0705748871,0.1464014053,0.0922076777,-0.1741306931,-0.3507432342,-0.0896511376,0.0367848724,0.0142136998,-0.191406846,-0.2775212228,0.0793746784,-0.1433201283,0.6197952032,0.32701388,0.005700672,0.3858786821,-0.2369560748,0.0000112303,0.1717422307,0.5837681293,-0.2605839074,-0.0444282405,0.1986091286,-0.2997124791,-0.442188561,-0.2730611861,0.188300699,0.0323222205,-0.0084293075,-0.1786166281,0.0310044196,-0.1309438795,-0.0489977524,-0.2130453289,-0.0558184981,0.3378269076,-0.1836595982,-0.2421921641,-0.0644743517,0.2167336792,0.0299405474,0.286546886,0.4599354267,-0.4840275347,0.1646117717,-0.1400938034,0.1275601089,0.0440830775,0.5117965341,0.1533818245,0.1332314163,0.000875253,-0.1406763494,0.0981805027,0.4896937013,0.056188345,-0.1449028701,0.3308489919,0.2087339163,0.2498039454,-0.1116498262,-0.0419149771,-0.076322034,-0.0079122381,0.1035077944,0.0772862658,-0.0560009852,0.0171361193,0.0741459429,-0.0036135165,-0.1537922323,0.2124625593,0.072868526,0.3913019896,-0.1112450063,-0.2594474852,-0.0057734381,-0.4689766765,-0.0055840118,0.0318537094,-0.0215430595,-0.1584112644,-0.1728700995,0.0590714328,-0.3485899866,0.2623428702,0.1606943607,0.2371293604,0.1690757126,-0.1341534704,-0.52725178,-0.2941477299,-0.3322209418,0.1811343133,0.5960763693,0.1151443645,-0.1382219791,-0.1728235483,0.1099917367,-0.0518502444,-0.1482473761,0.1006681323,-0.3312714398,0.2024192959,0.1513287723,-0.2806144059,0.0458374247,0.375363946,-0.2438099384,0.124802433,0.0661579967,-0.14999336,0.2123298347,0.1349525005,0.3393088579,-0.1040217653,-0.2439708114,-0.293197304,0.3013551831,-0.3514430821,0.3701419234,0.1212450117,0.0003360837,0.3002912104,-0.0935872495,0.1068416983,0.310128808,-0.1816409081,0.1664065272,-0.6221519709,0.1127027199,-0.4461789429,-0.2102091163,0.3684732318,-0.0753581375,0.2850719988,-0.1410285085,0.208453089,0.2306411862,-0.4770971537,-0.0794984475,-0.1620940566,-0.1941264272,0.2774369419,-0.0204179231,0.7596278191,0.4354731739,0.5634003282,0.2302164435,-0.1473510563,0.3893977404,-0.5955858827,0.1912795603,-0.3415154219,-0.3702346385,0.0957428738,0.1044336855,-0.0651132092,0.2145491242,0.2076508403,0.2953763902,0.3931223452,-0.0991116613,-0.1082807481,0.176552847,0.1234301403,0.1875409931,-0.2870904505,0.1991364509,0.3464435339,0.0045762225,-0.0596935861,0.0959675908,-0.7420678139,0.0225155354,0.0424916819,0.1004469842,-0.0955998227,0.3547414839,0.1195802242,-0.2243282497,-0.0487181768,0.3672380745,0.1598750502,-0.0159130674,0.0071960702,0.2100245059,-0.3194080293,0.3263314664,0.2423100621,-0.0388033167,0.0314080268,-0.2202143669,-0.0787131488,-0.0707605928,0.1557241678,-0.168481499,-0.5484755039,0.1821554154,-0.3347036242,0.0072074863,-0.0299168285,0.100690186,-0.0996055901,-0.0534304827,0.1672593504,-0.0568627976,-0.0626536757,-0.0348853096,0.0930720717,-0.33285743,-0.0504637584,0.4744251966,-0.1367388517,-0.0460099652,0.0648741722,-0.3305659294,0.0352666937,-0.243176356,0.0389244296,-0.0807575434,-0.1971541643,-0.1043005362,-0.0075798221,-0.2291497886,0.2527276278,0.2584365606,0.2495662868,0.2130372673,0.0328528471,-0.3630851805,-0.0983250961,0.0810795501,0.1679992825,-0.0347334705,-0.0196181145,-0.1595884413,0.2588406503,0.230631575,-0.3546353579,-0.1501723379,-0.0899762064,0.1268173754,-0.2210238129,-0.318739295,0.0224867295,0.1468538046,0.1430855095,-0.0081499545,-0.057829652,-0.2477172017,-0.205165863,0.083225742,-0.0483468585,-0.0974408388,-0.1320016086,0.2183995098,0.1043626517,-0.14063555,-0.2313222587,-0.1339266151,0.3573982418,0.1217451915,-0.0855392069,0.182516098,-0.1426294595,0.1086662039,0.0188238062,-0.0253852066,0.0735768974,-0.0265123807,0.1094571352,-0.2018416524,-0.2564111948,-0.0742407218,-0.2803894281,0.1471730322,0.1041463539,-0.0529725924,0.0087823989,-0.1832898557,0.2844219208,0.3649749458,0.1178048849,-0.0996367931,-0.0139173409,0.2252826989,-0.2222018987,0.094253704,0.4496181607,0.0439867266,0.0656358302,-0.0282832906,0.1585787684,-0.207698375,-0.0182847306,-0.0592486821,-0.2744233012,-0.0181700382,0.1728071719,-0.1911957115,0.2515231073,0.0736314803,0.155381009,-0.3791483045,0.1871821731,0.3507705629,-0.0025788394,0.1941527128,-0.2856431901,0.1495129317,-0.0236760825,0.090320833,-0.2703602612,0.3451122046,0.0853992477,-0.0218057595,0.0836358815,0.1691215634,-0.1083337441,-0.3376137614,-0.0048827888,-0.0880927593,-0.0704750121,-0.5443993211,-0.4664245844,-0.2264700234,-0.3945831656,0.0710501149,-0.168741256,0.3817866743,0.0295190457,0.2477214336,-0.0967243165,-0.1826776713,-0.4617654681,0.0102461642,0.4208806753,-0.0467574298,0.1774321795,0.2167157233,0.4143132865,0.3571799695,0.1616108865,0.2387460768,0.188827306,-0.2336149365,-0.2279311121,0.0324662104,0.0893204585,-0.1961580068,0.1742165089,0.2599774301,0.043390654,0.3002180159,0.1322067231,-0.2410817444,0.0774203986,0.2649872303,-0.0035752088,-0.2141294479,0.2947366536,-0.0391615815,0.0186822563,-0.3314743042,0.3797481954,-0.5077442527,-0.0865890756,-0.5439399481,0.1514041871,0.1786459982,-0.124354206,0.1039321572,0.2688948512,0.417313993,-0.1272355914,0.3387746513,-0.2841375768,-0.4613207579,-0.2653644681,-0.0458504297,0.039508719,0.270529747,-0.4309380352,0.2887354195,-0.2974143326,0.05145007,0.2745599747,0.4283306003,0.2469193488,0.0664969906,-0.3382286429,0.1559213102,0.1984027177,-0.0882449746,-0.1195693016,0.1306326389,-0.1739066392,0.0198163018,0.1097153276,0.0855560899,-0.3465361595,-0.2886542678,0.4207560122,0.3385222256,0.2750663757,0.2680857182,-0.3725818992,-0.3853949606,0.0310038142,-0.1237521991,-0.0627099201,0.0622824617,-0.0658418089,0.3876439035,0.0250640213,0.0747741759,0.0738607943,0.5041282773,0.0559554696,0.0017199109,-0.2270302922,-0.0262061767,0.0936457366,0.0634231493,-0.6129953861,0.4074434042,0.0005171182,0.1871438771,-0.2337078154,-0.2538995147,0.3750461042,-0.3584223688,0.2720425427,-0.0997014269,0.3427538872,0.2217618674,0.2968046069,0.0467911996,-0.105895713,0.071754165,0.2957480848,0.1622443944,0.2303458303,-0.2334202379,-0.0838037357,-0.0016768635,0.2927948236,0.0649266243,-0.2919711173,0.4134346247,-0.1841075718]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/225","title":"[ROUGE] Different scores with `files2rouge`","comments":"Thank you for your kind answer.\r\n\r\nAs a side question : Isn't it better to have a package that normalize more ?\r\n\r\nI understand to idea of having a package that does minimal post-processing for transparency.\r\n\r\nBut it means that people using different architecture (with different tokenizers for example) will have difference in ROUGE scores even if their predictions are actually similar. \r\nThe goal of `nlp` is to have _one package to rule them all_, right ?\r\n\r\nI will look into it but I'm not sure I have the required skill for this ^^ ","body":"It seems that the ROUGE score of `nlp` is lower than the one of `files2rouge`.\r\n\r\nHere is a self-contained notebook to reproduce both scores : https:\/\/colab.research.google.com\/drive\/14EyAXValB6UzKY9x4rs_T3pyL7alpw_F?usp=sharing\r\n\r\n---\r\n\r\n`nlp` : (Only mid F-scores)\r\n\r\n>rouge1 0.33508031962733364\r\nrouge2 0.14574333776191592\r\nrougeL 0.2321187823256159\r\n\r\n`files2rouge` :\r\n\r\n>Running ROUGE...\r\n===========================\r\n1 ROUGE-1 Average_R: 0.48873 (95%-conf.int. 0.41192 - 0.56339)\r\n1 ROUGE-1 Average_P: 0.29010 (95%-conf.int. 0.23605 - 0.34445)\r\n1 ROUGE-1 Average_F: 0.34761 (95%-conf.int. 0.29479 - 0.39871)\r\n===========================\r\n1 ROUGE-2 Average_R: 0.20280 (95%-conf.int. 0.14969 - 0.26244)\r\n1 ROUGE-2 Average_P: 0.12772 (95%-conf.int. 0.08603 - 0.17752)\r\n1 ROUGE-2 Average_F: 0.14798 (95%-conf.int. 0.10517 - 0.19240)\r\n===========================\r\n1 ROUGE-L Average_R: 0.32960 (95%-conf.int. 0.26501 - 0.39676)\r\n1 ROUGE-L Average_P: 0.19880 (95%-conf.int. 0.15257 - 0.25136)\r\n1 ROUGE-L Average_F: 0.23619 (95%-conf.int. 0.19073 - 0.28663)\r\n\r\n---\r\n\r\nWhen using longer predictions\/gold, the difference is bigger. \r\n**How can I reproduce same score as `files2rouge` ?**\r\n\r\n@lhoestq \r\n","comment_length":94,"text":"[ROUGE] Different scores with `files2rouge`\nIt seems that the ROUGE score of `nlp` is lower than the one of `files2rouge`.\r\n\r\nHere is a self-contained notebook to reproduce both scores : https:\/\/colab.research.google.com\/drive\/14EyAXValB6UzKY9x4rs_T3pyL7alpw_F?usp=sharing\r\n\r\n---\r\n\r\n`nlp` : (Only mid F-scores)\r\n\r\n>rouge1 0.33508031962733364\r\nrouge2 0.14574333776191592\r\nrougeL 0.2321187823256159\r\n\r\n`files2rouge` :\r\n\r\n>Running ROUGE...\r\n===========================\r\n1 ROUGE-1 Average_R: 0.48873 (95%-conf.int. 0.41192 - 0.56339)\r\n1 ROUGE-1 Average_P: 0.29010 (95%-conf.int. 0.23605 - 0.34445)\r\n1 ROUGE-1 Average_F: 0.34761 (95%-conf.int. 0.29479 - 0.39871)\r\n===========================\r\n1 ROUGE-2 Average_R: 0.20280 (95%-conf.int. 0.14969 - 0.26244)\r\n1 ROUGE-2 Average_P: 0.12772 (95%-conf.int. 0.08603 - 0.17752)\r\n1 ROUGE-2 Average_F: 0.14798 (95%-conf.int. 0.10517 - 0.19240)\r\n===========================\r\n1 ROUGE-L Average_R: 0.32960 (95%-conf.int. 0.26501 - 0.39676)\r\n1 ROUGE-L Average_P: 0.19880 (95%-conf.int. 0.15257 - 0.25136)\r\n1 ROUGE-L Average_F: 0.23619 (95%-conf.int. 0.19073 - 0.28663)\r\n\r\n---\r\n\r\nWhen using longer predictions\/gold, the difference is bigger. \r\n**How can I reproduce same score as `files2rouge` ?**\r\n\r\n@lhoestq \r\n\nThank you for your kind answer.\r\n\r\nAs a side question : Isn't it better to have a package that normalize more ?\r\n\r\nI understand to idea of having a package that does minimal post-processing for transparency.\r\n\r\nBut it means that people using different architecture (with different tokenizers for example) will have difference in ROUGE scores even if their predictions are actually similar. \r\nThe goal of `nlp` is to have _one package to rule them all_, right ?\r\n\r\nI will look into it but I'm not sure I have the required skill for this ^^ ","embeddings":[-0.1457264721,-0.1697432995,-0.1972297877,0.3397462964,-0.1194976568,-0.4242601097,-0.420399636,0.1476738602,-0.270043999,0.1725059152,-0.0828669071,0.1803876609,0.120330058,0.0080935061,-0.1003900841,-0.0979746804,0.1656547487,0.0388290472,0.0924810246,-0.3720678091,0.066559352,0.5696008801,-0.0313320532,0.21043396,0.0540217571,0.2702380717,0.2851136923,0.3035303652,-0.1485799253,0.0284368172,0.1280016154,0.0306413826,-0.023698464,0.2221485972,-0.0001088679,-0.2347189039,0.1026096642,-0.0031664839,-0.2757230699,-0.3026088178,-0.0345568508,0.0643639937,-0.1229285002,-0.0703938454,0.0858464167,-0.044892218,0.0487728454,0.2095425725,0.4293982685,0.1475735754,0.2262544334,-0.3024373949,-0.1853493452,0.0240821615,0.5763342977,0.1626760215,-0.1144831404,0.4472635388,0.0056994664,-0.3526385427,-0.2401907593,-0.0085632736,0.3430406153,-0.1725598574,0.1325590014,0.1931548268,0.148585543,-0.045940455,-0.079535611,0.2447285503,-0.2702822387,0.0322115794,-0.2271714061,-0.2778382003,-0.0763156638,-0.1997605413,0.1168917641,-0.0814458728,0.1642860919,0.0122475997,0.0441323034,0.2954271436,0.0496390946,0.007164266,-0.0540564582,-0.1078455895,0.0508458652,0.0984653682,0.2952796817,-0.0505435802,-0.5163154006,0.0107727246,-0.2593673468,0.0749608874,-0.302280724,-0.0378559306,0.3144967258,0.46597296,-0.200871408,0.499627769,0.2691969573,-0.0934173465,-0.132207945,-0.0394466408,-0.1378237605,0.1081323028,-0.0289463978,0.0206935201,0.0881181508,0.2246808112,-0.133582443,-0.0120077161,0.3982224762,-0.5637685657,-0.4777997434,0.1025914326,-0.2610344291,-0.432397604,-0.6891934872,-0.0673610121,-0.2554311156,-0.1192424744,0.0989084914,0.2272011787,-0.174884215,0.2922985256,0.0467879772,0.1833910197,-0.289503634,0.0196620282,-0.1839990616,0.0635797158,-0.3143928647,0.1345241964,0.1705666035,0.1127464101,0.3180577457,0.2560533583,0.2735133171,0.0386443697,0.3298657835,-0.2839507163,-0.022634726,-0.2111162692,-0.211391747,0.197453782,0.0271293335,0.0918488353,-0.0087018358,-0.0748942494,-0.1808407903,0.1442367584,0.4666180313,0.2065799385,0.1496509761,-0.1105541289,0.0406925417,0.116938062,-0.2401628494,-0.5936908722,0.2040610909,-0.136552155,-0.1270114928,-0.0581774674,0.1098197326,-0.0345822014,0.0931209251,-0.1174313724,0.3731989563,0.0635824502,0.2603128254,0.4534834325,-0.0408312865,0.0118638165,-0.1628564298,-0.0708854869,0.4665318727,-0.4385315478,-0.2222961336,0.3354998529,-0.1650701761,-0.065185748,0.239046663,0.0392757431,-0.3886532784,-0.166007787,0.1774507761,0.1975248605,0.0777985081,-0.0483966693,-0.2970068455,0.134958744,0.3166256249,0.0654775649,-0.2318288237,-0.2993499935,-0.1114049256,-0.0643908754,0.5141965151,0.1328068972,-0.2751375139,0.1049917638,-0.2721597254,0.2936930954,-0.0516953692,0.118573375,0.4230363369,-0.0233754385,-0.7951256037,-0.1602763534,0.4370327294,-0.05248693,-0.2122911811,-0.4714090824,0.2405331284,-0.2394646406,0.2232830822,0.0550486632,0.1873215437,0.0303547215,0.0240709484,0.0966067314,-0.2364263833,-0.1586188674,-0.1715354323,0.2036774307,-0.2210983336,-0.0074328831,-0.0105401725,0.3596087694,0.3116413951,0.0875802562,0.0373043455,0.263418287,0.0501557477,0.0980502665,0.3672500849,0.1448055357,-0.0007641261,0.1889616251,-0.1226600558,0.6879813671,0.0164625514,0.03225106,-0.3400395513,0.4259150624,-0.2170387655,0.2006913573,0.4093107283,-0.0717375576,-0.0500977673,0.0577801615,-0.0067999209,-0.0171536021,0.7467250228,0.195269227,0.2502554655,0.0029566153,-0.3220258057,0.2596178651,0.399140805,-0.0153429508,-0.0208532233,-0.1199541613,0.1937142164,-0.0933883861,-0.0367203467,0.2507390678,0.3810696304,0.1951338202,0.1557074785,-0.1311991662,-0.338236481,-0.3438020647,0.0351328626,-0.182103917,0.0874467045,0.4963860214,0.1016074717,0.2162867785,-0.3020778,-0.5422854424,-0.2536484003,-0.2223142087,0.0613259934,0.1616452485,-0.1439131945,-0.3678351939,-0.1536603719,-0.0919036344,-0.2918887734,-0.2732761204,0.2300974727,0.0208752602,-0.0724995062,-0.2189383954,0.0232325289,0.1487477869,-0.3703945577,0.1783805788,-0.0173747577,-0.0280433036,-0.2795380056,0.2228176445,-0.0567981303,-0.0193044487,-0.0314762257,-0.279186964,-0.4508359432,0.1696880311,-0.3762677908,-0.0109601086,0.0646441653,-0.0032368647,0.0596726015,-0.2506172359,-0.0705748871,0.1464014053,0.0922076777,-0.1741306931,-0.3507432342,-0.0896511376,0.0367848724,0.0142136998,-0.191406846,-0.2775212228,0.0793746784,-0.1433201283,0.6197952032,0.32701388,0.005700672,0.3858786821,-0.2369560748,0.0000112303,0.1717422307,0.5837681293,-0.2605839074,-0.0444282405,0.1986091286,-0.2997124791,-0.442188561,-0.2730611861,0.188300699,0.0323222205,-0.0084293075,-0.1786166281,0.0310044196,-0.1309438795,-0.0489977524,-0.2130453289,-0.0558184981,0.3378269076,-0.1836595982,-0.2421921641,-0.0644743517,0.2167336792,0.0299405474,0.286546886,0.4599354267,-0.4840275347,0.1646117717,-0.1400938034,0.1275601089,0.0440830775,0.5117965341,0.1533818245,0.1332314163,0.000875253,-0.1406763494,0.0981805027,0.4896937013,0.056188345,-0.1449028701,0.3308489919,0.2087339163,0.2498039454,-0.1116498262,-0.0419149771,-0.076322034,-0.0079122381,0.1035077944,0.0772862658,-0.0560009852,0.0171361193,0.0741459429,-0.0036135165,-0.1537922323,0.2124625593,0.072868526,0.3913019896,-0.1112450063,-0.2594474852,-0.0057734381,-0.4689766765,-0.0055840118,0.0318537094,-0.0215430595,-0.1584112644,-0.1728700995,0.0590714328,-0.3485899866,0.2623428702,0.1606943607,0.2371293604,0.1690757126,-0.1341534704,-0.52725178,-0.2941477299,-0.3322209418,0.1811343133,0.5960763693,0.1151443645,-0.1382219791,-0.1728235483,0.1099917367,-0.0518502444,-0.1482473761,0.1006681323,-0.3312714398,0.2024192959,0.1513287723,-0.2806144059,0.0458374247,0.375363946,-0.2438099384,0.124802433,0.0661579967,-0.14999336,0.2123298347,0.1349525005,0.3393088579,-0.1040217653,-0.2439708114,-0.293197304,0.3013551831,-0.3514430821,0.3701419234,0.1212450117,0.0003360837,0.3002912104,-0.0935872495,0.1068416983,0.310128808,-0.1816409081,0.1664065272,-0.6221519709,0.1127027199,-0.4461789429,-0.2102091163,0.3684732318,-0.0753581375,0.2850719988,-0.1410285085,0.208453089,0.2306411862,-0.4770971537,-0.0794984475,-0.1620940566,-0.1941264272,0.2774369419,-0.0204179231,0.7596278191,0.4354731739,0.5634003282,0.2302164435,-0.1473510563,0.3893977404,-0.5955858827,0.1912795603,-0.3415154219,-0.3702346385,0.0957428738,0.1044336855,-0.0651132092,0.2145491242,0.2076508403,0.2953763902,0.3931223452,-0.0991116613,-0.1082807481,0.176552847,0.1234301403,0.1875409931,-0.2870904505,0.1991364509,0.3464435339,0.0045762225,-0.0596935861,0.0959675908,-0.7420678139,0.0225155354,0.0424916819,0.1004469842,-0.0955998227,0.3547414839,0.1195802242,-0.2243282497,-0.0487181768,0.3672380745,0.1598750502,-0.0159130674,0.0071960702,0.2100245059,-0.3194080293,0.3263314664,0.2423100621,-0.0388033167,0.0314080268,-0.2202143669,-0.0787131488,-0.0707605928,0.1557241678,-0.168481499,-0.5484755039,0.1821554154,-0.3347036242,0.0072074863,-0.0299168285,0.100690186,-0.0996055901,-0.0534304827,0.1672593504,-0.0568627976,-0.0626536757,-0.0348853096,0.0930720717,-0.33285743,-0.0504637584,0.4744251966,-0.1367388517,-0.0460099652,0.0648741722,-0.3305659294,0.0352666937,-0.243176356,0.0389244296,-0.0807575434,-0.1971541643,-0.1043005362,-0.0075798221,-0.2291497886,0.2527276278,0.2584365606,0.2495662868,0.2130372673,0.0328528471,-0.3630851805,-0.0983250961,0.0810795501,0.1679992825,-0.0347334705,-0.0196181145,-0.1595884413,0.2588406503,0.230631575,-0.3546353579,-0.1501723379,-0.0899762064,0.1268173754,-0.2210238129,-0.318739295,0.0224867295,0.1468538046,0.1430855095,-0.0081499545,-0.057829652,-0.2477172017,-0.205165863,0.083225742,-0.0483468585,-0.0974408388,-0.1320016086,0.2183995098,0.1043626517,-0.14063555,-0.2313222587,-0.1339266151,0.3573982418,0.1217451915,-0.0855392069,0.182516098,-0.1426294595,0.1086662039,0.0188238062,-0.0253852066,0.0735768974,-0.0265123807,0.1094571352,-0.2018416524,-0.2564111948,-0.0742407218,-0.2803894281,0.1471730322,0.1041463539,-0.0529725924,0.0087823989,-0.1832898557,0.2844219208,0.3649749458,0.1178048849,-0.0996367931,-0.0139173409,0.2252826989,-0.2222018987,0.094253704,0.4496181607,0.0439867266,0.0656358302,-0.0282832906,0.1585787684,-0.207698375,-0.0182847306,-0.0592486821,-0.2744233012,-0.0181700382,0.1728071719,-0.1911957115,0.2515231073,0.0736314803,0.155381009,-0.3791483045,0.1871821731,0.3507705629,-0.0025788394,0.1941527128,-0.2856431901,0.1495129317,-0.0236760825,0.090320833,-0.2703602612,0.3451122046,0.0853992477,-0.0218057595,0.0836358815,0.1691215634,-0.1083337441,-0.3376137614,-0.0048827888,-0.0880927593,-0.0704750121,-0.5443993211,-0.4664245844,-0.2264700234,-0.3945831656,0.0710501149,-0.168741256,0.3817866743,0.0295190457,0.2477214336,-0.0967243165,-0.1826776713,-0.4617654681,0.0102461642,0.4208806753,-0.0467574298,0.1774321795,0.2167157233,0.4143132865,0.3571799695,0.1616108865,0.2387460768,0.188827306,-0.2336149365,-0.2279311121,0.0324662104,0.0893204585,-0.1961580068,0.1742165089,0.2599774301,0.043390654,0.3002180159,0.1322067231,-0.2410817444,0.0774203986,0.2649872303,-0.0035752088,-0.2141294479,0.2947366536,-0.0391615815,0.0186822563,-0.3314743042,0.3797481954,-0.5077442527,-0.0865890756,-0.5439399481,0.1514041871,0.1786459982,-0.124354206,0.1039321572,0.2688948512,0.417313993,-0.1272355914,0.3387746513,-0.2841375768,-0.4613207579,-0.2653644681,-0.0458504297,0.039508719,0.270529747,-0.4309380352,0.2887354195,-0.2974143326,0.05145007,0.2745599747,0.4283306003,0.2469193488,0.0664969906,-0.3382286429,0.1559213102,0.1984027177,-0.0882449746,-0.1195693016,0.1306326389,-0.1739066392,0.0198163018,0.1097153276,0.0855560899,-0.3465361595,-0.2886542678,0.4207560122,0.3385222256,0.2750663757,0.2680857182,-0.3725818992,-0.3853949606,0.0310038142,-0.1237521991,-0.0627099201,0.0622824617,-0.0658418089,0.3876439035,0.0250640213,0.0747741759,0.0738607943,0.5041282773,0.0559554696,0.0017199109,-0.2270302922,-0.0262061767,0.0936457366,0.0634231493,-0.6129953861,0.4074434042,0.0005171182,0.1871438771,-0.2337078154,-0.2538995147,0.3750461042,-0.3584223688,0.2720425427,-0.0997014269,0.3427538872,0.2217618674,0.2968046069,0.0467911996,-0.105895713,0.071754165,0.2957480848,0.1622443944,0.2303458303,-0.2334202379,-0.0838037357,-0.0016768635,0.2927948236,0.0649266243,-0.2919711173,0.4134346247,-0.1841075718]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/225","title":"[ROUGE] Different scores with `files2rouge`","comments":"You're right, there's a pretty interesting trade-off here between robustness and sensitivity :) The flip side of your argument is that we also still want the metric to be sensitive to model mistakes. How we think about number normalization for example has evolved a fair bit since the Perl script was written: at the time, ROUGE was used mostly to evaluate short-medium text summarization systems, where there were only a few numbers in the input and it was assumed that the most popular methods in use at the time would get those right. However, as your example showcases, that assumption does not hold any more, and we do want to be able to penalize a model that generates a wrong numerical value.\r\n\r\nAlso, we think that abstracting away tokenization differences is the role of the model\/tokenizer: if you use the \ud83e\udd17Tokenizers library for example, it will handle that for you ;)\r\n\r\nFinally, there is a lot of active research on developing model-powered metrics that are both more sensitive and more robust than ROUGE. Check out for example BERTscore, which is implemented in this library!","body":"It seems that the ROUGE score of `nlp` is lower than the one of `files2rouge`.\r\n\r\nHere is a self-contained notebook to reproduce both scores : https:\/\/colab.research.google.com\/drive\/14EyAXValB6UzKY9x4rs_T3pyL7alpw_F?usp=sharing\r\n\r\n---\r\n\r\n`nlp` : (Only mid F-scores)\r\n\r\n>rouge1 0.33508031962733364\r\nrouge2 0.14574333776191592\r\nrougeL 0.2321187823256159\r\n\r\n`files2rouge` :\r\n\r\n>Running ROUGE...\r\n===========================\r\n1 ROUGE-1 Average_R: 0.48873 (95%-conf.int. 0.41192 - 0.56339)\r\n1 ROUGE-1 Average_P: 0.29010 (95%-conf.int. 0.23605 - 0.34445)\r\n1 ROUGE-1 Average_F: 0.34761 (95%-conf.int. 0.29479 - 0.39871)\r\n===========================\r\n1 ROUGE-2 Average_R: 0.20280 (95%-conf.int. 0.14969 - 0.26244)\r\n1 ROUGE-2 Average_P: 0.12772 (95%-conf.int. 0.08603 - 0.17752)\r\n1 ROUGE-2 Average_F: 0.14798 (95%-conf.int. 0.10517 - 0.19240)\r\n===========================\r\n1 ROUGE-L Average_R: 0.32960 (95%-conf.int. 0.26501 - 0.39676)\r\n1 ROUGE-L Average_P: 0.19880 (95%-conf.int. 0.15257 - 0.25136)\r\n1 ROUGE-L Average_F: 0.23619 (95%-conf.int. 0.19073 - 0.28663)\r\n\r\n---\r\n\r\nWhen using longer predictions\/gold, the difference is bigger. \r\n**How can I reproduce same score as `files2rouge` ?**\r\n\r\n@lhoestq \r\n","comment_length":184,"text":"[ROUGE] Different scores with `files2rouge`\nIt seems that the ROUGE score of `nlp` is lower than the one of `files2rouge`.\r\n\r\nHere is a self-contained notebook to reproduce both scores : https:\/\/colab.research.google.com\/drive\/14EyAXValB6UzKY9x4rs_T3pyL7alpw_F?usp=sharing\r\n\r\n---\r\n\r\n`nlp` : (Only mid F-scores)\r\n\r\n>rouge1 0.33508031962733364\r\nrouge2 0.14574333776191592\r\nrougeL 0.2321187823256159\r\n\r\n`files2rouge` :\r\n\r\n>Running ROUGE...\r\n===========================\r\n1 ROUGE-1 Average_R: 0.48873 (95%-conf.int. 0.41192 - 0.56339)\r\n1 ROUGE-1 Average_P: 0.29010 (95%-conf.int. 0.23605 - 0.34445)\r\n1 ROUGE-1 Average_F: 0.34761 (95%-conf.int. 0.29479 - 0.39871)\r\n===========================\r\n1 ROUGE-2 Average_R: 0.20280 (95%-conf.int. 0.14969 - 0.26244)\r\n1 ROUGE-2 Average_P: 0.12772 (95%-conf.int. 0.08603 - 0.17752)\r\n1 ROUGE-2 Average_F: 0.14798 (95%-conf.int. 0.10517 - 0.19240)\r\n===========================\r\n1 ROUGE-L Average_R: 0.32960 (95%-conf.int. 0.26501 - 0.39676)\r\n1 ROUGE-L Average_P: 0.19880 (95%-conf.int. 0.15257 - 0.25136)\r\n1 ROUGE-L Average_F: 0.23619 (95%-conf.int. 0.19073 - 0.28663)\r\n\r\n---\r\n\r\nWhen using longer predictions\/gold, the difference is bigger. \r\n**How can I reproduce same score as `files2rouge` ?**\r\n\r\n@lhoestq \r\n\nYou're right, there's a pretty interesting trade-off here between robustness and sensitivity :) The flip side of your argument is that we also still want the metric to be sensitive to model mistakes. How we think about number normalization for example has evolved a fair bit since the Perl script was written: at the time, ROUGE was used mostly to evaluate short-medium text summarization systems, where there were only a few numbers in the input and it was assumed that the most popular methods in use at the time would get those right. However, as your example showcases, that assumption does not hold any more, and we do want to be able to penalize a model that generates a wrong numerical value.\r\n\r\nAlso, we think that abstracting away tokenization differences is the role of the model\/tokenizer: if you use the \ud83e\udd17Tokenizers library for example, it will handle that for you ;)\r\n\r\nFinally, there is a lot of active research on developing model-powered metrics that are both more sensitive and more robust than ROUGE. Check out for example BERTscore, which is implemented in this library!","embeddings":[-0.1457264721,-0.1697432995,-0.1972297877,0.3397462964,-0.1194976568,-0.4242601097,-0.420399636,0.1476738602,-0.270043999,0.1725059152,-0.0828669071,0.1803876609,0.120330058,0.0080935061,-0.1003900841,-0.0979746804,0.1656547487,0.0388290472,0.0924810246,-0.3720678091,0.066559352,0.5696008801,-0.0313320532,0.21043396,0.0540217571,0.2702380717,0.2851136923,0.3035303652,-0.1485799253,0.0284368172,0.1280016154,0.0306413826,-0.023698464,0.2221485972,-0.0001088679,-0.2347189039,0.1026096642,-0.0031664839,-0.2757230699,-0.3026088178,-0.0345568508,0.0643639937,-0.1229285002,-0.0703938454,0.0858464167,-0.044892218,0.0487728454,0.2095425725,0.4293982685,0.1475735754,0.2262544334,-0.3024373949,-0.1853493452,0.0240821615,0.5763342977,0.1626760215,-0.1144831404,0.4472635388,0.0056994664,-0.3526385427,-0.2401907593,-0.0085632736,0.3430406153,-0.1725598574,0.1325590014,0.1931548268,0.148585543,-0.045940455,-0.079535611,0.2447285503,-0.2702822387,0.0322115794,-0.2271714061,-0.2778382003,-0.0763156638,-0.1997605413,0.1168917641,-0.0814458728,0.1642860919,0.0122475997,0.0441323034,0.2954271436,0.0496390946,0.007164266,-0.0540564582,-0.1078455895,0.0508458652,0.0984653682,0.2952796817,-0.0505435802,-0.5163154006,0.0107727246,-0.2593673468,0.0749608874,-0.302280724,-0.0378559306,0.3144967258,0.46597296,-0.200871408,0.499627769,0.2691969573,-0.0934173465,-0.132207945,-0.0394466408,-0.1378237605,0.1081323028,-0.0289463978,0.0206935201,0.0881181508,0.2246808112,-0.133582443,-0.0120077161,0.3982224762,-0.5637685657,-0.4777997434,0.1025914326,-0.2610344291,-0.432397604,-0.6891934872,-0.0673610121,-0.2554311156,-0.1192424744,0.0989084914,0.2272011787,-0.174884215,0.2922985256,0.0467879772,0.1833910197,-0.289503634,0.0196620282,-0.1839990616,0.0635797158,-0.3143928647,0.1345241964,0.1705666035,0.1127464101,0.3180577457,0.2560533583,0.2735133171,0.0386443697,0.3298657835,-0.2839507163,-0.022634726,-0.2111162692,-0.211391747,0.197453782,0.0271293335,0.0918488353,-0.0087018358,-0.0748942494,-0.1808407903,0.1442367584,0.4666180313,0.2065799385,0.1496509761,-0.1105541289,0.0406925417,0.116938062,-0.2401628494,-0.5936908722,0.2040610909,-0.136552155,-0.1270114928,-0.0581774674,0.1098197326,-0.0345822014,0.0931209251,-0.1174313724,0.3731989563,0.0635824502,0.2603128254,0.4534834325,-0.0408312865,0.0118638165,-0.1628564298,-0.0708854869,0.4665318727,-0.4385315478,-0.2222961336,0.3354998529,-0.1650701761,-0.065185748,0.239046663,0.0392757431,-0.3886532784,-0.166007787,0.1774507761,0.1975248605,0.0777985081,-0.0483966693,-0.2970068455,0.134958744,0.3166256249,0.0654775649,-0.2318288237,-0.2993499935,-0.1114049256,-0.0643908754,0.5141965151,0.1328068972,-0.2751375139,0.1049917638,-0.2721597254,0.2936930954,-0.0516953692,0.118573375,0.4230363369,-0.0233754385,-0.7951256037,-0.1602763534,0.4370327294,-0.05248693,-0.2122911811,-0.4714090824,0.2405331284,-0.2394646406,0.2232830822,0.0550486632,0.1873215437,0.0303547215,0.0240709484,0.0966067314,-0.2364263833,-0.1586188674,-0.1715354323,0.2036774307,-0.2210983336,-0.0074328831,-0.0105401725,0.3596087694,0.3116413951,0.0875802562,0.0373043455,0.263418287,0.0501557477,0.0980502665,0.3672500849,0.1448055357,-0.0007641261,0.1889616251,-0.1226600558,0.6879813671,0.0164625514,0.03225106,-0.3400395513,0.4259150624,-0.2170387655,0.2006913573,0.4093107283,-0.0717375576,-0.0500977673,0.0577801615,-0.0067999209,-0.0171536021,0.7467250228,0.195269227,0.2502554655,0.0029566153,-0.3220258057,0.2596178651,0.399140805,-0.0153429508,-0.0208532233,-0.1199541613,0.1937142164,-0.0933883861,-0.0367203467,0.2507390678,0.3810696304,0.1951338202,0.1557074785,-0.1311991662,-0.338236481,-0.3438020647,0.0351328626,-0.182103917,0.0874467045,0.4963860214,0.1016074717,0.2162867785,-0.3020778,-0.5422854424,-0.2536484003,-0.2223142087,0.0613259934,0.1616452485,-0.1439131945,-0.3678351939,-0.1536603719,-0.0919036344,-0.2918887734,-0.2732761204,0.2300974727,0.0208752602,-0.0724995062,-0.2189383954,0.0232325289,0.1487477869,-0.3703945577,0.1783805788,-0.0173747577,-0.0280433036,-0.2795380056,0.2228176445,-0.0567981303,-0.0193044487,-0.0314762257,-0.279186964,-0.4508359432,0.1696880311,-0.3762677908,-0.0109601086,0.0646441653,-0.0032368647,0.0596726015,-0.2506172359,-0.0705748871,0.1464014053,0.0922076777,-0.1741306931,-0.3507432342,-0.0896511376,0.0367848724,0.0142136998,-0.191406846,-0.2775212228,0.0793746784,-0.1433201283,0.6197952032,0.32701388,0.005700672,0.3858786821,-0.2369560748,0.0000112303,0.1717422307,0.5837681293,-0.2605839074,-0.0444282405,0.1986091286,-0.2997124791,-0.442188561,-0.2730611861,0.188300699,0.0323222205,-0.0084293075,-0.1786166281,0.0310044196,-0.1309438795,-0.0489977524,-0.2130453289,-0.0558184981,0.3378269076,-0.1836595982,-0.2421921641,-0.0644743517,0.2167336792,0.0299405474,0.286546886,0.4599354267,-0.4840275347,0.1646117717,-0.1400938034,0.1275601089,0.0440830775,0.5117965341,0.1533818245,0.1332314163,0.000875253,-0.1406763494,0.0981805027,0.4896937013,0.056188345,-0.1449028701,0.3308489919,0.2087339163,0.2498039454,-0.1116498262,-0.0419149771,-0.076322034,-0.0079122381,0.1035077944,0.0772862658,-0.0560009852,0.0171361193,0.0741459429,-0.0036135165,-0.1537922323,0.2124625593,0.072868526,0.3913019896,-0.1112450063,-0.2594474852,-0.0057734381,-0.4689766765,-0.0055840118,0.0318537094,-0.0215430595,-0.1584112644,-0.1728700995,0.0590714328,-0.3485899866,0.2623428702,0.1606943607,0.2371293604,0.1690757126,-0.1341534704,-0.52725178,-0.2941477299,-0.3322209418,0.1811343133,0.5960763693,0.1151443645,-0.1382219791,-0.1728235483,0.1099917367,-0.0518502444,-0.1482473761,0.1006681323,-0.3312714398,0.2024192959,0.1513287723,-0.2806144059,0.0458374247,0.375363946,-0.2438099384,0.124802433,0.0661579967,-0.14999336,0.2123298347,0.1349525005,0.3393088579,-0.1040217653,-0.2439708114,-0.293197304,0.3013551831,-0.3514430821,0.3701419234,0.1212450117,0.0003360837,0.3002912104,-0.0935872495,0.1068416983,0.310128808,-0.1816409081,0.1664065272,-0.6221519709,0.1127027199,-0.4461789429,-0.2102091163,0.3684732318,-0.0753581375,0.2850719988,-0.1410285085,0.208453089,0.2306411862,-0.4770971537,-0.0794984475,-0.1620940566,-0.1941264272,0.2774369419,-0.0204179231,0.7596278191,0.4354731739,0.5634003282,0.2302164435,-0.1473510563,0.3893977404,-0.5955858827,0.1912795603,-0.3415154219,-0.3702346385,0.0957428738,0.1044336855,-0.0651132092,0.2145491242,0.2076508403,0.2953763902,0.3931223452,-0.0991116613,-0.1082807481,0.176552847,0.1234301403,0.1875409931,-0.2870904505,0.1991364509,0.3464435339,0.0045762225,-0.0596935861,0.0959675908,-0.7420678139,0.0225155354,0.0424916819,0.1004469842,-0.0955998227,0.3547414839,0.1195802242,-0.2243282497,-0.0487181768,0.3672380745,0.1598750502,-0.0159130674,0.0071960702,0.2100245059,-0.3194080293,0.3263314664,0.2423100621,-0.0388033167,0.0314080268,-0.2202143669,-0.0787131488,-0.0707605928,0.1557241678,-0.168481499,-0.5484755039,0.1821554154,-0.3347036242,0.0072074863,-0.0299168285,0.100690186,-0.0996055901,-0.0534304827,0.1672593504,-0.0568627976,-0.0626536757,-0.0348853096,0.0930720717,-0.33285743,-0.0504637584,0.4744251966,-0.1367388517,-0.0460099652,0.0648741722,-0.3305659294,0.0352666937,-0.243176356,0.0389244296,-0.0807575434,-0.1971541643,-0.1043005362,-0.0075798221,-0.2291497886,0.2527276278,0.2584365606,0.2495662868,0.2130372673,0.0328528471,-0.3630851805,-0.0983250961,0.0810795501,0.1679992825,-0.0347334705,-0.0196181145,-0.1595884413,0.2588406503,0.230631575,-0.3546353579,-0.1501723379,-0.0899762064,0.1268173754,-0.2210238129,-0.318739295,0.0224867295,0.1468538046,0.1430855095,-0.0081499545,-0.057829652,-0.2477172017,-0.205165863,0.083225742,-0.0483468585,-0.0974408388,-0.1320016086,0.2183995098,0.1043626517,-0.14063555,-0.2313222587,-0.1339266151,0.3573982418,0.1217451915,-0.0855392069,0.182516098,-0.1426294595,0.1086662039,0.0188238062,-0.0253852066,0.0735768974,-0.0265123807,0.1094571352,-0.2018416524,-0.2564111948,-0.0742407218,-0.2803894281,0.1471730322,0.1041463539,-0.0529725924,0.0087823989,-0.1832898557,0.2844219208,0.3649749458,0.1178048849,-0.0996367931,-0.0139173409,0.2252826989,-0.2222018987,0.094253704,0.4496181607,0.0439867266,0.0656358302,-0.0282832906,0.1585787684,-0.207698375,-0.0182847306,-0.0592486821,-0.2744233012,-0.0181700382,0.1728071719,-0.1911957115,0.2515231073,0.0736314803,0.155381009,-0.3791483045,0.1871821731,0.3507705629,-0.0025788394,0.1941527128,-0.2856431901,0.1495129317,-0.0236760825,0.090320833,-0.2703602612,0.3451122046,0.0853992477,-0.0218057595,0.0836358815,0.1691215634,-0.1083337441,-0.3376137614,-0.0048827888,-0.0880927593,-0.0704750121,-0.5443993211,-0.4664245844,-0.2264700234,-0.3945831656,0.0710501149,-0.168741256,0.3817866743,0.0295190457,0.2477214336,-0.0967243165,-0.1826776713,-0.4617654681,0.0102461642,0.4208806753,-0.0467574298,0.1774321795,0.2167157233,0.4143132865,0.3571799695,0.1616108865,0.2387460768,0.188827306,-0.2336149365,-0.2279311121,0.0324662104,0.0893204585,-0.1961580068,0.1742165089,0.2599774301,0.043390654,0.3002180159,0.1322067231,-0.2410817444,0.0774203986,0.2649872303,-0.0035752088,-0.2141294479,0.2947366536,-0.0391615815,0.0186822563,-0.3314743042,0.3797481954,-0.5077442527,-0.0865890756,-0.5439399481,0.1514041871,0.1786459982,-0.124354206,0.1039321572,0.2688948512,0.417313993,-0.1272355914,0.3387746513,-0.2841375768,-0.4613207579,-0.2653644681,-0.0458504297,0.039508719,0.270529747,-0.4309380352,0.2887354195,-0.2974143326,0.05145007,0.2745599747,0.4283306003,0.2469193488,0.0664969906,-0.3382286429,0.1559213102,0.1984027177,-0.0882449746,-0.1195693016,0.1306326389,-0.1739066392,0.0198163018,0.1097153276,0.0855560899,-0.3465361595,-0.2886542678,0.4207560122,0.3385222256,0.2750663757,0.2680857182,-0.3725818992,-0.3853949606,0.0310038142,-0.1237521991,-0.0627099201,0.0622824617,-0.0658418089,0.3876439035,0.0250640213,0.0747741759,0.0738607943,0.5041282773,0.0559554696,0.0017199109,-0.2270302922,-0.0262061767,0.0936457366,0.0634231493,-0.6129953861,0.4074434042,0.0005171182,0.1871438771,-0.2337078154,-0.2538995147,0.3750461042,-0.3584223688,0.2720425427,-0.0997014269,0.3427538872,0.2217618674,0.2968046069,0.0467911996,-0.105895713,0.071754165,0.2957480848,0.1622443944,0.2303458303,-0.2334202379,-0.0838037357,-0.0016768635,0.2927948236,0.0649266243,-0.2919711173,0.4134346247,-0.1841075718]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/224","title":"[Feature Request\/Help] BLEURT model -> PyTorch","comments":"Is there any update on this? \r\n\r\nThanks!","body":"Hi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Twitter).\r\n\r\nI had a go of just like manually using the checkpoint that they publish which includes the weights. It seems like the architecture is exactly aligned with the out-of-the-box BertModel in transformers just with a single linear layer on top of the CLS embedding. I loaded all the weights to the PyTorch model but I am not able to get the same numbers as the BLEURT package's python api. Here is my colab notebook where I tried https:\/\/colab.research.google.com\/drive\/1Bfced531EvQP_CpFvxwxNl25Pj6ptylY?usp=sharing . If you have any pointers on what might be going wrong that would be much appreciated!\r\n\r\nThank you muchly!","comment_length":7,"text":"[Feature Request\/Help] BLEURT model -> PyTorch\nHi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Twitter).\r\n\r\nI had a go of just like manually using the checkpoint that they publish which includes the weights. It seems like the architecture is exactly aligned with the out-of-the-box BertModel in transformers just with a single linear layer on top of the CLS embedding. I loaded all the weights to the PyTorch model but I am not able to get the same numbers as the BLEURT package's python api. Here is my colab notebook where I tried https:\/\/colab.research.google.com\/drive\/1Bfced531EvQP_CpFvxwxNl25Pj6ptylY?usp=sharing . If you have any pointers on what might be going wrong that would be much appreciated!\r\n\r\nThank you muchly!\nIs there any update on this? \r\n\r\nThanks!","embeddings":[0.0769645944,-0.2721210122,0.0911164731,0.266956687,0.2280693799,-0.3040116727,0.3290455341,0.1226881593,0.1514686048,0.1511580348,-0.2440004647,0.2983187139,-0.4640855491,0.2562192082,0.0681594387,-0.0891910866,-0.0481648669,0.0602022,0.1281574517,-0.1624639332,-0.1408660114,0.2177770287,0.0296547264,-0.0940112919,-0.2574591637,0.3580862582,-0.0639657825,0.0777869448,-0.2171576172,-0.0603650399,0.2810967863,-0.0772188902,0.2021557242,0.7916839719,-0.0001253193,-0.3410066068,0.1413801312,-0.0118958699,-0.318888098,-0.2242820561,0.4198358357,-0.2784964442,0.1322808564,-0.1702308357,-0.0185153093,-0.0713392943,-0.2348389626,-0.1212117448,0.1161928996,0.2510134578,0.0197905432,0.128056407,-0.020217875,0.0736400634,0.1776612699,-0.1582725048,-0.0421606749,0.5778595209,0.3364526629,-0.0287013408,-0.1524122506,0.1890420169,0.0651289672,0.1304624677,0.7059246898,0.0288367383,-0.1223891899,-0.0963791683,0.052343592,0.2919279635,-0.0581676178,-0.1375780255,-0.258885175,0.1402303576,0.1161526367,-0.2550997436,-0.2629127204,0.1321908683,0.0986306369,-0.2374990731,-0.3676900566,-0.4275257289,-0.2053250968,0.0825206041,-0.268489331,0.3987672925,-0.0548062585,-0.1267796904,0.1209228262,0.071789436,0.3665702045,-0.0829115212,0.3362879455,0.2581480742,0.0665699095,-0.2493262738,0.2149145156,-0.0185621586,-0.0121646235,-0.2788236439,-0.2064609081,0.1907791197,-0.2383274734,-0.0010505973,-0.0137834484,0.3618043363,-0.0057348115,0.3418517411,0.304445833,0.3778805435,0.0558601916,-0.0572867282,-0.1138474271,-0.1294468939,0.1950473338,0.1685471833,-0.2413292527,-0.3131300509,-0.0632238239,-0.0361835808,-0.2427953929,-0.1951813996,0.3276986778,0.1461552233,-0.4845784903,-0.0377782658,0.5439999104,0.0379455537,-0.1878189296,0.0748964623,-0.1068077087,-0.1142934859,-0.3253027201,0.0774332434,0.2264290452,-0.1199535951,0.2847441435,-0.0198139511,0.5112751126,0.0804554969,0.1947977394,0.4007608593,0.1299037039,-0.193135947,-0.2331850082,-0.3177248836,0.1568032354,0.0306287464,-0.1348334551,0.0323535055,-0.156435132,-0.2841642499,-0.1703152955,0.0239797905,-0.3012142181,-0.0575961284,0.0461821891,0.3378869593,0.095436208,-0.2571758032,-0.0328019969,-0.2965148091,-0.1462129802,0.0748948753,0.399148196,-0.2302847952,-0.3461496532,-0.4428849518,0.1181931198,0.4895263612,0.2493035197,0.1565327793,-0.2389797866,-0.1123697609,0.1919502616,-0.3175570369,0.3271844685,0.1847557724,-0.2943373322,-0.2386659384,0.0995219722,-0.3109753728,-0.211658448,0.1105794087,-0.1736912131,-0.0435990952,0.3402058482,0.2117206305,-0.1473146081,-0.0641849935,-0.3295137882,-0.2367307842,0.2762959003,0.1204185337,0.3501949906,0.0270363893,0.1754651815,0.517464757,0.1218558401,-0.3649024069,-0.0323551744,-0.2058850676,0.4999886155,-0.3820032179,0.4073016346,-0.210660249,-0.0861462578,0.1872591823,-0.4036572278,0.3713673651,0.1549323201,0.0917610377,-0.1338422,-0.0609493442,0.2914878428,-0.0767591,-0.118681848,-0.1895484924,-0.2884859145,-0.1944032609,-0.2126965523,-0.2713961005,-0.1709916592,-0.0642361492,0.2283495218,0.3220317662,-0.0760638639,-0.1707732379,-0.0445991904,0.4014178216,0.1863695383,0.2054944187,-0.0261892695,-0.2806453705,0.1536019146,-0.2377888113,0.2374673188,0.1608981192,0.4179314673,-0.0443292446,0.3283953667,0.2252492756,-0.0966056585,-0.079722926,-0.190968588,0.2646250129,0.3012746871,0.2598773539,0.2824874222,0.0714970604,-0.254739821,0.1175601706,-0.0205440801,0.1484289467,0.3380433023,-0.1376534253,-0.0222462919,-0.1932622045,0.1560545713,0.2216446996,0.2904651463,-0.0162393246,-0.0231615789,0.2041871995,-0.4346252084,-0.3120645285,-0.0206515603,-0.0155996997,0.1145655662,0.1011385173,0.4645623863,0.1070848256,0.001588529,0.0285844002,-0.1383199394,0.1434418708,-0.1761867106,-0.0830774754,0.0821303874,0.1061413288,0.1075521111,0.0448154695,0.0344020315,0.1248625442,-0.1739302576,0.3498610258,0.0934181362,0.3603532016,-0.2327635884,-0.383299768,-0.2340296954,-0.2262019366,0.0013943901,0.2929693162,0.0406900086,-0.1669992805,0.1653419733,0.5148428082,-0.0630059019,-0.0532075502,-0.0763593614,-0.2328189015,-0.3136793077,-0.0055914833,-0.0567345694,0.0757192671,-0.0327562802,0.0523231551,-0.072154738,-0.1812010854,-0.8160647154,0.0253611878,0.068970792,0.3185171187,0.167893663,-0.1326136291,-0.2992697358,0.0651082695,0.0287988968,-0.398483187,-0.0306567401,0.2902272642,-0.213417381,0.0559903532,-0.0336719677,-0.2146829069,-0.0926501974,0.0014972251,0.0055362382,0.1884091645,0.0475645848,-0.1968725026,0.1078413129,0.3295027614,0.423233211,-0.1249096915,-0.0734033734,0.2304234952,0.4280174375,-0.2964344025,-0.3310207427,-0.1199903414,-0.0148909492,0.0618359223,-0.0396632627,-0.3148698509,-0.328761369,-0.1521040946,-0.0896924734,0.1944935769,0.2178356051,0.0613545924,0.1368204206,0.0870406181,-0.2135280371,-0.1670393348,-0.0497920513,0.2071353048,0.7003551126,-0.2024998665,-0.188724637,0.040008273,0.8280011415,0.1723065227,-0.4270895422,0.0641201735,0.2739430368,0.1011408716,-0.0738350898,-0.2518183291,0.1426744759,-0.2595513761,-0.2639357448,0.3241861761,0.1183645129,-0.2561791241,-0.2919297516,-0.2228096575,0.5202057362,-0.0984480828,0.0380027853,0.035373915,-0.0094121238,-0.0286628716,-0.1023548171,-0.2464041412,-0.1216186583,0.2071378231,0.0050566532,0.1019346938,-0.0198710915,0.463164866,-0.0543646812,-0.6182572842,0.436309576,0.0506023616,-0.5876846313,0.1629425287,-0.1271802187,-0.014589305,-0.0793007836,-0.3722271025,0.1293336153,-0.0109154684,-0.1242982,0.0334329642,-0.1958976239,0.0740289018,-0.1957906783,-0.3438901007,0.5401678681,0.3791955411,-0.3765573204,0.0668348968,0.2670398057,0.4072266221,-0.181815356,-0.0329587311,-0.3715336025,-0.2523064911,0.1500943303,-0.0691625699,-0.0102747511,0.2743607163,-0.0704310462,-0.0410309918,0.177914694,-0.1281831712,0.18814753,0.0573490784,0.0468375795,-0.0433239602,0.1347927749,0.1158179641,0.0059143789,0.1953403354,0.4235871732,-0.4545798898,-0.4546498954,0.178659901,0.1672542691,0.2195011824,0.1211151332,0.1002714112,-0.0702131391,0.4462929368,0.264413178,0.1970801502,0.254620254,0.1296408921,0.3872816563,0.0546697974,-0.1699094474,0.3792904317,-0.150835216,0.0167802442,0.2487807125,0.0080236513,-0.0005889082,0.0788227692,0.4726569057,0.9947401285,0.1680686623,0.1260013431,0.2837894559,-0.2398847789,0.3168632686,-0.0974229798,0.0428152122,-0.3445183039,-0.3001939952,-0.0061897333,-0.0673773065,-0.4302369356,-0.2562252581,-0.2130158693,0.162420854,-0.3646328449,-0.2657547295,0.2764922082,0.3529539406,0.060947869,0.1062527671,0.2761923671,0.0428881645,-0.1808521301,0.34008497,-0.0883081928,-0.2205734551,-0.2350035459,0.1006353647,-0.2168741524,0.3344072998,-0.0550317951,-0.0410257429,-0.1011451632,-0.3731209636,0.0711917728,0.0079898685,0.7282720208,0.1195480004,-0.3702606857,0.1090696603,-0.1054101214,0.128511101,0.1490149349,-0.1364171058,0.3434342146,-0.1699344367,-0.188618958,-0.07605277,0.1814926565,-0.3237537146,0.121294789,-0.1347643137,-0.0836343542,-0.0856381804,0.376080811,0.3359513879,0.1569563299,-0.2248822898,0.0701787546,0.2131723762,-0.2436327934,0.07614398,0.1400217116,-0.235157758,0.0105179446,-0.0226342194,0.0192107223,-0.0963735655,0.229185313,0.1805318892,-0.1202457771,0.232861951,-0.5785297751,0.554854393,-0.4296751618,0.1343450099,-0.12044058,-0.0095181419,0.0732625499,0.6476778984,0.4668928087,0.1500619054,-0.409280926,0.1528281868,-0.4188937247,0.0921545103,-0.2781175375,0.3855577111,-0.2110980451,-0.0390777513,-0.0295199007,0.2699607015,-0.1931181252,0.0205822457,0.1771633476,-0.0868559331,-0.0096985614,-0.4409418106,0.0442096405,0.0664688498,-0.2710798681,-0.1437865645,0.127825439,0.0040698457,-0.4155914485,0.1709961593,0.2193405628,0.0609645247,-0.3221980929,-0.4108484685,-0.0798236057,-0.1234761626,0.0228744429,0.0087468186,0.1059882343,0.3117083311,0.1795488894,-0.0030759696,-0.2125156969,0.0258429088,0.4270035028,0.1624821872,0.0603471026,-0.0718994662,0.4109271467,-0.0007013993,0.1376493573,0.054505907,-0.0193503387,0.0946052521,0.4180380404,0.0866205245,0.1997704953,-0.3755164444,0.2741155028,0.2588984966,0.1955222785,-0.1807197779,0.236701712,0.0726467669,-0.2636015117,-0.1805230975,-0.2213550508,0.084280327,-0.1819964051,0.1934349835,-0.1221713051,0.3824150562,0.3864810467,-0.0277277846,-0.2486477494,-0.2182094157,0.0774412975,0.0272261444,-0.199457407,0.288564682,0.5504803658,0.1077059284,0.0377418324,-0.399854809,-0.1308544278,0.017924428,-0.3060232997,-0.1341054887,0.389215827,0.2295185477,-0.1349049062,-0.1731546819,0.2194181234,-0.1119645983,-0.1071755737,0.4391743243,-0.3436707854,-0.170295164,-0.0277561843,0.3397659957,-0.0170952026,0.100055553,0.1326625049,-0.1883736402,0.0623624139,-0.0218484569,0.0266221017,0.1018815562,0.038616959,0.0821561664,-0.1073714569,0.1879786402,-0.6621757746,0.3763124645,0.296117723,0.0575760864,-0.1578894556,0.1165532023,-0.0545291305,-0.1867948174,0.3422041833,-0.0557101704,0.2486488372,-0.3532544374,-0.0420944653,0.3424499333,-0.0373449102,0.0799797773,0.133983627,0.4808753133,-0.1488703638,-0.0635126904,-0.1001357362,-0.13806732,0.4367698431,-0.4165613353,0.2054197341,-0.0508183092,0.8885181546,0.3101511002,-0.2712148428,-0.2761763632,-0.1144971848,-0.3296041489,0.1014064923,0.0763991401,-0.0723793507,-0.0215982366,-0.0538462698,-0.0123991333,0.1156896949,0.1746277511,0.2992973626,0.44558689,-0.3280488253,0.2537563741,-0.2758664489,0.1422647238,0.2241567522,0.0756553933,-0.1699276268,-0.0376273394,0.0024203258,-0.111474216,0.129204616,-0.2548455596,0.0801516846,-0.1834201813,-0.2611302137,-0.245191887,0.0314407907,0.0308087468,-0.0120533239,-0.057877373,0.0210760366,-0.1224084646,-0.1401793361,0.0036706359,-0.3364719152,0.3516384065,0.3228633404,0.109443374,0.0220393594,0.6532315016,0.1701945513,-0.1387836933,0.2232505977,0.1590741426,-0.0774094984,-0.2128058225,0.0261546131,0.1958025545,-0.0829149112,-0.2472068518,-0.5097185969,0.5806443691,0.1176257133,-0.4995141029,-0.1679649204,-0.2020861357,-0.2037510574,-0.1470912695,0.2841015756,0.3640090227,0.1141774207,-0.0655971318,-0.4236589074,0.061200466,0.4607792497,-0.4134958386,-0.1489250362,-0.0981427133,0.1875570714,0.2807607949,-0.185521394,-0.8837897182,-0.1519079059,0.1487349421,0.0166335832,-0.1198668182,0.2793745995,0.1818922162,0.2272649556,-0.2627471089,0.5660734773,-0.0031166929,-0.0982478335,-0.5282688141,-0.2625540495]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/224","title":"[Feature Request\/Help] BLEURT model -> PyTorch","comments":"Hitting this error when using bleurt with PyTorch ...\r\n\r\n```\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n... and I'm assuming because it was built for TF specifically. Is there a way to use this metric in PyTorch?","body":"Hi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Twitter).\r\n\r\nI had a go of just like manually using the checkpoint that they publish which includes the weights. It seems like the architecture is exactly aligned with the out-of-the-box BertModel in transformers just with a single linear layer on top of the CLS embedding. I loaded all the weights to the PyTorch model but I am not able to get the same numbers as the BLEURT package's python api. Here is my colab notebook where I tried https:\/\/colab.research.google.com\/drive\/1Bfced531EvQP_CpFvxwxNl25Pj6ptylY?usp=sharing . If you have any pointers on what might be going wrong that would be much appreciated!\r\n\r\nThank you muchly!","comment_length":38,"text":"[Feature Request\/Help] BLEURT model -> PyTorch\nHi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Twitter).\r\n\r\nI had a go of just like manually using the checkpoint that they publish which includes the weights. It seems like the architecture is exactly aligned with the out-of-the-box BertModel in transformers just with a single linear layer on top of the CLS embedding. I loaded all the weights to the PyTorch model but I am not able to get the same numbers as the BLEURT package's python api. Here is my colab notebook where I tried https:\/\/colab.research.google.com\/drive\/1Bfced531EvQP_CpFvxwxNl25Pj6ptylY?usp=sharing . If you have any pointers on what might be going wrong that would be much appreciated!\r\n\r\nThank you muchly!\nHitting this error when using bleurt with PyTorch ...\r\n\r\n```\r\nUnrecognizedFlagError: Unknown command line flag 'f'\r\n```\r\n... and I'm assuming because it was built for TF specifically. Is there a way to use this metric in PyTorch?","embeddings":[0.0004879119,-0.1944001168,0.0977797955,0.1410354078,0.2860070765,-0.2194753885,0.2981606722,0.1757618785,0.1684712321,0.1584296376,-0.2239150256,0.2047086954,-0.4715835452,0.2436532378,-0.0394526012,-0.0937426761,-0.1899132431,-0.0336314701,0.1052042097,-0.015585267,-0.0906981975,0.2365888655,0.045650728,-0.0287116207,-0.3174606562,0.3504799306,0.0552200526,0.0559271649,-0.1453370899,-0.0327646509,0.3463766277,-0.1924692392,0.2557664514,0.7926077843,-0.0001235562,-0.2130512893,0.0671901554,-0.0112896813,-0.2755617499,-0.3058053255,0.4654067755,-0.2523365915,0.0014712351,-0.2047467679,-0.0270921495,-0.0622269325,-0.2024288625,-0.0805595145,0.0086525586,0.3162073791,0.0406865142,0.2029185295,0.0005698704,0.0706276149,0.0959588811,-0.1069801301,-0.1193717793,0.5899261832,0.1905050278,-0.0390190519,-0.1993449926,0.1651122421,0.1557437032,0.2218760997,0.7811316252,0.03963238,-0.1051483601,0.0078610759,0.0517428033,0.3505223095,-0.0570162758,-0.2330993563,-0.2421548218,0.1831979007,0.068287909,-0.281570822,-0.2522399426,0.0521995761,0.0469888858,-0.2774746716,-0.382629782,-0.354801625,-0.2562071383,0.1344293654,-0.2618772089,0.3999883831,-0.0648725256,-0.0589506961,0.1828382462,0.0548060238,0.209708631,-0.0122973248,0.3575534225,0.3185590506,0.0450001433,-0.2273323238,0.1566396207,-0.1462717354,0.0655388013,-0.2430343479,-0.2348572016,0.1291662008,-0.3112175465,0.0090820882,-0.0051812772,0.4701384902,0.0389155224,0.3749932647,0.3503911793,0.2842978239,-0.0297545362,-0.0786910653,-0.0677850991,-0.2594396472,0.2883807421,0.1613946557,-0.2171094269,-0.311069876,-0.1479407549,-0.0581485331,-0.1836255044,-0.2136000395,0.3854407668,0.2447400242,-0.3576726019,-0.1165842265,0.5851029754,0.0581300519,-0.2100519985,0.0859041661,-0.0929606929,-0.0970908031,-0.2978732288,0.0861729831,0.2061848044,-0.1500409245,0.209319979,-0.1078692079,0.6836862564,0.135818094,0.1862277985,0.3334340155,0.0832948312,-0.1664729863,-0.2499709129,-0.40275383,0.2584887743,0.0936398804,-0.1719942242,0.0645830929,-0.1143134162,-0.3340193331,-0.2537507415,0.0032774769,-0.2937202156,-0.1393714994,0.0467373393,0.3141675293,0.0852746069,-0.2999877334,0.0218755882,-0.2990497649,-0.2519973516,0.0970712528,0.2996884882,-0.3436619043,-0.3732590377,-0.4258345366,0.1134296656,0.3820712864,0.2269518673,0.1458987743,-0.1502639353,-0.1124973819,0.1804845333,-0.3285117447,0.3918120265,0.0386061333,-0.181296289,-0.2300805897,0.1016746685,-0.286238879,-0.2184966356,0.0639600009,-0.1890777797,-0.1219702139,0.3215975463,0.2692631185,-0.1522388756,-0.1192740947,-0.2392320782,-0.2757670581,0.2382387817,0.1066083983,0.3859713674,-0.0185017623,0.1580834538,0.5193030834,0.018723486,-0.4093893766,-0.0457703918,-0.1325852722,0.381831944,-0.2926695645,0.3908918798,-0.3206767738,-0.1213139519,0.2370038182,-0.2204726338,0.3617454469,0.199922502,0.1151999384,-0.1020439342,-0.0449570455,0.3409549594,0.0391774774,-0.1111276299,-0.1455000043,-0.2357196659,-0.1997219622,-0.2001805305,-0.3327550888,-0.291932404,-0.0543323532,0.4235497117,0.2041716129,-0.038499888,-0.0883931443,-0.0854253992,0.4208586514,0.13108477,0.2169723213,-0.1212121621,-0.2075590044,0.1365715414,-0.2814387977,0.2433534116,0.2878023982,0.345333606,-0.0836210623,0.1704854667,0.2630348206,-0.1229204163,-0.055307474,-0.0595499054,0.2311901003,0.298914969,0.2667693794,0.2532306015,0.0738804117,-0.2816995978,0.0544116236,0.0128507838,0.0123666888,0.4185680747,-0.2691677511,-0.1440791041,-0.1935264468,0.2079549879,0.0653934926,0.3745555878,-0.026033273,0.04296536,0.3506178856,-0.2904786766,-0.2164357752,-0.1224425733,-0.0230483133,0.1722002178,0.101308383,0.4216381907,0.1246316284,-0.1053852886,0.0762932077,-0.0540655442,0.1142887175,-0.199891746,-0.1042663828,0.0411479957,0.0894846171,0.1685812771,-0.0190609042,0.063017346,0.129863143,-0.2475427091,0.3689163625,0.0334745683,0.367395103,-0.1088698804,-0.3008708358,-0.2557474971,-0.1884493828,0.0084981322,0.2539114356,0.0916624069,-0.142018646,0.0540209822,0.4649915695,-0.0766244456,-0.1419015378,-0.0292699002,-0.3652302027,-0.3453891277,0.0093076173,0.0419486761,0.042096708,0.0964580029,0.0896503329,-0.0313360505,-0.2331273854,-0.7229447365,0.0043879054,-0.0067760991,0.2805350721,0.211117819,-0.0831347629,-0.3391121626,0.0040894351,0.0523563065,-0.5004475713,-0.0058011105,0.3323715925,-0.3564035594,0.0889470428,-0.0786410049,-0.1056697816,-0.0934436843,-0.0775417089,-0.0709223747,0.1984406114,0.0537065193,-0.0583876409,-0.0204657782,0.4024657011,0.5086855292,-0.0480325595,-0.1685403734,0.2404119968,0.3423588574,-0.3283778131,-0.3037554026,-0.0282775518,0.0658796206,0.029017875,0.0055691032,-0.3014440835,-0.4165248275,-0.1035077125,-0.0752799958,0.051425457,0.2057031095,0.0248824731,0.0247537456,0.0788896754,-0.2982492149,-0.2002647966,-0.0193369724,0.1483561546,0.7090571523,-0.1582219452,-0.1775605232,0.1485042125,0.8652088046,0.0129403165,-0.4598026872,0.1080646664,0.1744071394,0.0926937014,0.0157062057,-0.219392702,0.2692294717,-0.3033925891,-0.1727742255,0.3044237196,0.1061276123,-0.1255859882,-0.2744297683,-0.1240753159,0.5246574879,-0.1307867914,0.0432769693,0.0808661059,0.0277936924,-0.059137661,-0.1332893521,-0.3359974623,-0.1166523769,0.2886205614,0.0849208012,-0.0168991797,0.0468008965,0.4622923136,-0.119692184,-0.5215771794,0.477922529,0.1486181766,-0.6269401908,0.0303427931,-0.0907730684,-0.0149778184,-0.0043394212,-0.3057304323,0.0829900652,0.0220561549,-0.0230037011,0.0475445576,-0.2639552355,0.111073792,-0.1353975981,-0.3203111589,0.4749357104,0.2835913002,-0.3579865396,0.0602517575,0.2661552131,0.3918273747,-0.1682068408,-0.0442588292,-0.3871447444,-0.2475208491,0.0511751063,-0.1605274975,0.0598302819,0.2689395547,0.0639897734,0.1575710028,0.1485835612,-0.0523847006,0.1533213854,0.0633866563,0.0676541477,-0.0354301929,0.0365983956,0.1509349346,-0.0538212694,0.1668811142,0.3117909431,-0.4149143696,-0.4968371689,0.1397722512,0.2661863565,0.2174071372,0.0349097624,0.0777493119,-0.1071273237,0.3252798021,0.3136738837,0.1727414578,0.3225979507,0.1536158323,0.3155540228,0.114199087,-0.1838609427,0.3391724825,-0.1192089245,0.1223259494,0.348988086,0.0299479142,-0.0275351945,0.0645965189,0.5371320248,0.9507375956,0.0719896853,0.1056296825,0.3453920782,-0.2159121037,0.3318788111,-0.1346594244,-0.0547708049,-0.2090287507,-0.2768645287,-0.0287369955,-0.0094007719,-0.416841656,-0.3024944961,-0.245131582,0.1704130322,-0.3722744286,-0.2893781364,0.273917526,0.2895247638,-0.012750363,0.026801791,0.2843710184,0.0721207485,-0.2462864965,0.3137528896,-0.0689199641,-0.1371177137,-0.1217496172,0.1112828851,-0.3442812264,0.1823105514,-0.143615678,-0.0431867242,-0.137941137,-0.2727847099,0.0484643094,0.0430409834,0.7600719929,0.0196650065,-0.3796815872,0.2225043476,-0.0693353042,0.1874959916,0.1232801527,-0.1162858903,0.3619279861,-0.2187030315,-0.0835879743,-0.1212912202,0.2360728979,-0.4464293718,0.1687423587,-0.1505320668,-0.1622354984,0.0115138926,0.245608747,0.2365314662,0.1182114482,-0.1011223495,0.0586779565,0.1474028826,-0.2392086238,0.0327095985,0.043618124,-0.3119702637,0.0290990248,-0.0414380133,-0.0783292279,-0.0092197349,0.2251460105,0.1811981052,-0.0850266963,0.2331380248,-0.5456883907,0.4800318182,-0.3977860808,0.1631653607,0.0829263255,-0.0336070955,0.0867536217,0.7250293493,0.4419584572,0.1383617818,-0.4551554322,0.103372246,-0.3853457272,0.0554811209,-0.187551111,0.3016719222,-0.2130830288,0.1631274819,-0.051664222,0.1955759972,-0.1942232698,-0.0192081295,0.1287049502,-0.1361641139,0.0578930788,-0.3698289096,0.1801576018,0.0709929392,-0.2820509076,-0.1298039854,0.0791756511,-0.0011408869,-0.4950146377,0.162168324,0.230059877,-0.0153498724,-0.3231398463,-0.4455380738,-0.0574519821,-0.1060748175,0.0210535228,0.1417102069,0.0447904617,0.3292350173,-0.0591244362,-0.0471910611,-0.1035596281,-0.0545505658,0.4670524299,0.2202817053,0.115329124,0.0120212464,0.3437568545,0.0014604587,0.1320913881,0.0297615174,0.0062268204,0.0604617186,0.2978816628,0.1449952275,0.19285281,-0.3699392974,0.2290145606,0.1482077986,0.1222192198,-0.161981225,0.1719021201,0.0346485041,-0.2363875061,-0.2088801563,-0.3177432418,0.0820757076,-0.1735854596,0.1348427385,-0.2804161012,0.4191925824,0.3138982952,-0.0982554629,-0.242277801,-0.2245112956,0.0681361333,0.1616402268,-0.1870094389,0.3133465052,0.4076449573,0.1146761402,0.2044928521,-0.4011959136,-0.142490536,0.0656723082,-0.4225869477,-0.0895751268,0.322717011,0.1163295507,-0.0532086939,-0.1357431412,0.1743603945,-0.0714455843,-0.1919140071,0.3954488933,-0.315526247,-0.2116173059,-0.0874658152,0.35621804,0.1253100932,0.0961894616,0.1022959501,-0.0866241306,0.2011722624,-0.0553172044,0.0332171209,0.1526265889,-0.1583859622,0.1244634539,-0.0410244204,0.265127629,-0.7464025617,0.2926133275,0.2878527045,0.1497822851,-0.2172665298,0.1176650524,-0.0134346476,-0.0537018552,0.3168705702,0.0329933949,0.1935813576,-0.37752828,-0.0368585214,0.2186944783,0.0019623372,0.1600331217,0.1604831964,0.4990955591,-0.1896755546,0.0575371534,-0.1026487723,-0.1495476067,0.4168822765,-0.3948406875,0.1670202762,-0.0611447804,0.8726229668,0.2835285962,-0.1502934545,-0.2694928646,-0.0804476142,-0.2182772756,0.0842755437,0.0871815681,-0.1201402321,0.0740615204,-0.0515723526,0.0068646353,0.1870849878,0.1844526082,0.2467506677,0.3651891649,-0.276065141,0.3165981472,-0.3954797983,0.1557580084,0.1742842644,0.0607457571,-0.0484895594,-0.1410810351,0.0234865174,-0.1150446162,0.2646304965,-0.2252410501,0.0511214882,-0.1827779859,-0.2393108755,-0.3419876993,-0.0500369817,0.0389033258,-0.0317364298,-0.1246154681,-0.0637544021,-0.0963194147,-0.1280702502,-0.0501852557,-0.2914326489,0.3976371586,0.370854944,0.150457114,-0.0725284293,0.768453896,0.2032062709,0.0067161205,0.3011975586,0.2014060616,-0.1147808284,-0.2022213042,-0.0375080667,0.1938225031,-0.2518007457,-0.2619609535,-0.4261301458,0.61869663,0.1093906611,-0.4776004851,-0.0888891369,-0.2265508771,-0.2260579616,-0.130824402,0.2619839311,0.4220674336,0.0790817887,0.0322288871,-0.3935738802,0.0753868893,0.5604330301,-0.4282587767,-0.1913508326,-0.0126961432,0.1809550822,0.2391916215,-0.3084867001,-0.7673503757,-0.1244050339,0.1188996807,0.1689778715,-0.1666519046,0.2352272421,0.1324500889,0.2679490149,-0.2229093462,0.5535248518,-0.0783578083,-0.1384928375,-0.412629813,-0.0999660343]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/224","title":"[Feature Request\/Help] BLEURT model -> PyTorch","comments":"We currently provide a wrapper on the TensorFlow implementation: https:\/\/huggingface.co\/metrics\/bleurt\r\n\r\nWe have long term plans to better handle model-based metrics, but they probably won't be implemented right away\r\n\r\n@adamwlev it would still be cool to add the BLEURT checkpoints to the transformers repo if you're interested, but that would best be discussed there :) \r\n\r\nclosing for now","body":"Hi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Twitter).\r\n\r\nI had a go of just like manually using the checkpoint that they publish which includes the weights. It seems like the architecture is exactly aligned with the out-of-the-box BertModel in transformers just with a single linear layer on top of the CLS embedding. I loaded all the weights to the PyTorch model but I am not able to get the same numbers as the BLEURT package's python api. Here is my colab notebook where I tried https:\/\/colab.research.google.com\/drive\/1Bfced531EvQP_CpFvxwxNl25Pj6ptylY?usp=sharing . If you have any pointers on what might be going wrong that would be much appreciated!\r\n\r\nThank you muchly!","comment_length":57,"text":"[Feature Request\/Help] BLEURT model -> PyTorch\nHi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Twitter).\r\n\r\nI had a go of just like manually using the checkpoint that they publish which includes the weights. It seems like the architecture is exactly aligned with the out-of-the-box BertModel in transformers just with a single linear layer on top of the CLS embedding. I loaded all the weights to the PyTorch model but I am not able to get the same numbers as the BLEURT package's python api. Here is my colab notebook where I tried https:\/\/colab.research.google.com\/drive\/1Bfced531EvQP_CpFvxwxNl25Pj6ptylY?usp=sharing . If you have any pointers on what might be going wrong that would be much appreciated!\r\n\r\nThank you muchly!\nWe currently provide a wrapper on the TensorFlow implementation: https:\/\/huggingface.co\/metrics\/bleurt\r\n\r\nWe have long term plans to better handle model-based metrics, but they probably won't be implemented right away\r\n\r\n@adamwlev it would still be cool to add the BLEURT checkpoints to the transformers repo if you're interested, but that would best be discussed there :) \r\n\r\nclosing for now","embeddings":[0.0396036766,-0.294549346,0.0695946515,0.1643861979,0.2727992237,-0.2811159492,0.38066715,0.1402266473,0.1454873532,0.1657412946,-0.2105591744,0.2934762537,-0.4552504122,0.2241601944,0.0695941672,-0.1098843887,-0.0509779267,0.0371574424,0.1063593552,-0.1357742697,-0.1087616831,0.1649904698,0.0593500435,-0.1525469422,-0.2916157842,0.3176142275,0.0192493163,0.0257068686,-0.2094441354,-0.0577423349,0.2948090732,-0.0604142509,0.2058348507,0.7945739031,-0.0001224944,-0.2677385807,0.1325636506,-0.0253161397,-0.3406816423,-0.2211294025,0.3839514554,-0.3060537279,0.0823291987,-0.1754459292,0.0427883156,-0.0340652652,-0.2432406396,-0.0426557995,0.1874929368,0.2166575789,0.0422483832,0.2304356843,0.0012260919,0.1125814095,0.1391409039,-0.1055864319,-0.1043705791,0.5303354859,0.2758409083,-0.0624103881,-0.1922634393,0.2381688207,0.0778375641,0.101423353,0.6716713309,0.0390604064,-0.0555954054,-0.0681458116,0.005765269,0.3251337707,-0.114808701,-0.2151495516,-0.3067008257,0.0872814134,0.0710418969,-0.2673069239,-0.2179705352,0.106988363,0.1001451537,-0.2690118849,-0.4190244377,-0.364830941,-0.1958520859,0.0797381252,-0.2726512849,0.3813111484,0.009816546,-0.1550256461,0.1310456097,0.1450663954,0.3880723417,-0.0743771195,0.3471343517,0.2746399343,-0.0165070072,-0.3021687567,0.2685927153,-0.1360906065,-0.0050387089,-0.2247710675,-0.1665416211,0.2024484724,-0.2630818188,0.0775952637,-0.0429920703,0.401499778,-0.0002396844,0.372813046,0.3213628232,0.4031051695,0.104582943,-0.0196256172,-0.0739620849,-0.0863811821,0.2585074306,0.1930990219,-0.2239641994,-0.25219661,-0.0677620322,-0.0203725789,-0.2209860682,-0.2402275652,0.3333803415,0.1934585422,-0.4989129305,-0.0231807753,0.5008059144,0.0060213157,-0.1655693054,0.1050276309,-0.1418714076,-0.1260456741,-0.3094723821,0.0759087801,0.2528241277,-0.1601814777,0.2812950313,-0.076871708,0.4981269836,0.0475040898,0.3035526574,0.3707239926,0.0962487012,-0.1441248059,-0.2020063102,-0.3485707641,0.1855834424,0.0507264175,-0.1315112561,-0.0309726689,-0.114903979,-0.2794413269,-0.075020656,0.0509738848,-0.2576736212,-0.057642512,0.0253990646,0.3534557819,0.0978759676,-0.2471972257,0.0006365551,-0.2485276461,-0.167667836,0.0672941059,0.3956460655,-0.1806036234,-0.3170186579,-0.5069212317,0.1681725085,0.4395365715,0.1879903972,0.1661761254,-0.2475459576,-0.0329421535,0.2556152046,-0.3426558673,0.2737840414,0.1172076166,-0.3056722879,-0.2301910073,0.0894582793,-0.3044480085,-0.2189075649,0.0856976211,-0.1461465508,-0.0644501522,0.2756146193,0.2686757445,-0.1765097529,-0.012937896,-0.2883855999,-0.2562455535,0.246227473,0.1788948625,0.3769879639,-0.0379012376,0.1523504704,0.5387817621,0.0669796616,-0.3838602304,-0.0300253555,-0.1569999456,0.5144113898,-0.4600306153,0.3816400766,-0.2148496062,-0.117012158,0.1749205589,-0.3829161227,0.3466677368,0.1580247432,0.0597301014,-0.0823241696,0.0048180511,0.3204125762,-0.0448577777,-0.0798355788,-0.1922538877,-0.2519005835,-0.1517390609,-0.2314921319,-0.2959868014,-0.1850219518,-0.0406598635,0.2206051499,0.3695568442,-0.0213718712,-0.1467802376,-0.0849675387,0.3680124283,0.1280675977,0.1687961519,-0.0365346745,-0.2077385783,0.1696754694,-0.1845837981,0.2581163645,0.2752383053,0.407127738,0.0261994246,0.2872467041,0.2546685636,-0.100231804,-0.1049645543,-0.2070124,0.2988640964,0.3075419068,0.2385030687,0.2911813855,0.0936933905,-0.2713990808,0.0976055786,-0.0604119487,0.1711318344,0.3944917023,-0.1385252178,-0.0819859877,-0.1640891731,0.1472574621,0.1415482163,0.360725373,-0.052627977,0.0177914649,0.2372604907,-0.414567709,-0.2849433124,-0.0858028904,0.0554222129,0.1426686198,0.0641591698,0.4149323702,0.0514041185,-0.0188012589,0.044747971,-0.1211606413,0.1547817886,-0.2346344292,-0.0462681651,0.089615114,0.0987383351,0.090155676,0.0251399856,0.0745601952,0.1143124029,-0.1435763985,0.3518405259,0.1326592863,0.4257067144,-0.232666254,-0.364344418,-0.2572859228,-0.2122784108,0.0386481173,0.3008292019,0.0732662752,-0.085523665,0.2842047215,0.5591579676,-0.0492170826,-0.1014965698,-0.0635784343,-0.2979995012,-0.2619434893,0.0130494479,-0.0470721014,0.0210433081,-0.0121315392,0.1273240149,-0.0729090124,-0.1648213267,-0.7918711305,-0.0071387189,0.061953228,0.2949561179,0.1277306527,-0.1378466636,-0.327054441,-0.0012699482,0.0172598958,-0.4089310467,-0.0230823662,0.3334011436,-0.2403262109,0.0200988147,-0.0331980512,-0.1950121224,-0.1217609644,-0.0371470749,0.0350580998,0.1774721146,0.0372842364,-0.1275971979,0.1132897139,0.3214526772,0.4272084534,-0.1533857286,-0.0384360254,0.1742305607,0.4179562628,-0.2875500321,-0.3364490867,-0.1131410822,-0.0429405347,0.0972191468,0.0200368837,-0.3338596225,-0.3707358837,-0.1350100636,-0.1547125876,0.1676237881,0.2113265246,0.0806474239,0.1357162893,0.0812390894,-0.164341554,-0.1194946989,-0.0512098521,0.1652881056,0.6772111654,-0.2268800139,-0.1361592114,0.0998500213,0.865185678,0.147551924,-0.4560471177,0.0362508558,0.317102015,0.0258227959,-0.0309558529,-0.291069597,0.1881337017,-0.2465849519,-0.2843211889,0.2827787101,0.1024310887,-0.2974280417,-0.3031903505,-0.2026455551,0.5068925023,-0.1223453432,0.07363078,0.1130760089,0.0693345368,-0.004519497,-0.0654088855,-0.2938379347,-0.1253665686,0.240793094,-0.0076102004,0.0948043913,-0.0374576449,0.4476364553,0.0032588257,-0.5484255552,0.3487662077,0.0529347323,-0.5662446618,0.1476329416,-0.0729421005,-0.0393105373,-0.046537742,-0.3226353824,0.0747007504,0.0361691751,-0.1483704448,-0.061204575,-0.2529002726,0.092325002,-0.2258281261,-0.4737833142,0.5314677954,0.3373503387,-0.3827210367,0.0932153761,0.2227109671,0.315443933,-0.2028792948,-0.0146928914,-0.2976086736,-0.2209087908,0.0764063075,-0.0746564046,-0.0264334269,0.2257058322,-0.0512012169,-0.0273023676,0.12710917,-0.101631768,0.1812264025,0.0748959109,0.0083874781,-0.0267140176,0.1550222635,0.1112856343,-0.078166239,0.1924811751,0.3351226151,-0.4193533361,-0.5140190125,0.1123961285,0.1502209902,0.1733609736,0.096206516,0.0997608453,-0.0103324344,0.4023586214,0.2830177844,0.1623783559,0.2686342299,0.1312973052,0.4590050876,0.0657718331,-0.1467003673,0.3369626403,-0.1843064576,0.0292250346,0.3064768016,-0.0933981538,-0.0260423794,0.1751044989,0.4582521319,0.954762578,0.1138849482,0.0766222402,0.2924112678,-0.1200177744,0.3372704983,-0.2470247746,0.0528967865,-0.3119491935,-0.3774301708,-0.0375666395,-0.0892937407,-0.4033792913,-0.2799663246,-0.2035803646,0.1163453534,-0.3447310627,-0.2265592813,0.304363668,0.3243243694,0.0753794461,0.0215697307,0.2870031893,0.061904747,-0.2093516737,0.3146812022,-0.0779998675,-0.2303091735,-0.2361436635,0.101312086,-0.1874045134,0.2655441761,-0.0571903251,-0.0067226514,0.0100549646,-0.3199194372,0.0706599206,0.0693367124,0.7521306276,0.1196119338,-0.3720659912,0.1482444704,-0.1703068763,0.1371913105,0.1343544871,-0.1684082747,0.3660239577,-0.1432255507,-0.1668988019,-0.0825686976,0.2166524529,-0.2843281329,0.1679222137,-0.173920393,-0.1123276129,-0.071442537,0.4080491662,0.3213432133,0.0798186585,-0.2015874386,0.0962599441,0.2228288054,-0.2782490849,0.0879303589,0.1365130097,-0.236873284,-0.0030592971,0.0210609734,-0.0454434417,-0.0760928914,0.2099469751,0.1896344423,-0.1224563941,0.2027594149,-0.5506263375,0.5572549105,-0.4292548001,0.1107014492,-0.1371678412,-0.022318203,0.0341504999,0.6142625809,0.4629493356,0.177575022,-0.4233263731,0.1225864813,-0.4270440638,0.1048761308,-0.1800378859,0.3510752916,-0.273953557,-0.006761584,0.0067098117,0.389806807,-0.2270165384,-0.0263610184,0.1260772496,-0.07171648,-0.0216518044,-0.4830073118,0.11179322,0.0455929562,-0.2318330258,-0.1076429635,0.0828796625,-0.0429535434,-0.4671590924,0.151980713,0.1879009604,0.0512357093,-0.2913528383,-0.3445456922,-0.0923873112,-0.181612283,-0.0012822374,-0.0359881558,0.0712981075,0.3768666089,0.16384691,0.0322551653,-0.2014025599,-0.0582619533,0.3841692507,0.1152417436,0.0622107126,-0.0822442845,0.3912833035,-0.0234722849,0.0268530082,0.0503187701,0.0277119298,0.087253958,0.4256474376,0.0977621898,0.1573863626,-0.4112392366,0.2807865143,0.1774874926,0.1182877943,-0.1377708018,0.1800993383,0.1043133736,-0.2249921411,-0.2332541794,-0.1972156167,0.1246391833,-0.1126857176,0.1643534899,-0.1148823947,0.3551516235,0.3632742465,-0.0097659891,-0.2480158061,-0.3071395159,0.0422256999,0.0548876561,-0.1848622561,0.2769478559,0.4547138512,0.068892926,0.1006450653,-0.4465751648,-0.0926669836,0.0869694054,-0.3843234777,-0.0882020965,0.3419310153,0.1956781,-0.1506090909,-0.205921039,0.1579788923,-0.1140966192,-0.1324256212,0.3602488339,-0.2412534952,-0.1418252736,-0.0540720001,0.3252484202,0.0262757726,0.0775646567,0.0578758717,-0.1066066772,0.0652185902,-0.0064457343,0.107477352,0.0673594326,-0.0129648875,0.1285708994,-0.1101794019,0.20703803,-0.6235055327,0.3515489995,0.3406662345,0.0404872708,-0.2386523336,0.1311085373,-0.065376237,-0.2679315507,0.3332059979,-0.0799207613,0.2008474767,-0.3434508443,0.08213377,0.2826623321,-0.0541475117,0.0699153766,0.2067144662,0.4017029107,-0.0886710137,-0.0517617986,-0.0598030984,-0.1769834906,0.3869960606,-0.442759186,0.1662397236,-0.0957949311,0.9571204185,0.3474615514,-0.2492387742,-0.2494022846,-0.1470510513,-0.3462857008,0.1076830328,0.1776347905,-0.0383405201,-0.0039488678,-0.0639107898,0.0028834105,0.0523778722,0.1550376415,0.2236462086,0.53174752,-0.3545685709,0.2993887067,-0.2283246219,0.1032901704,0.2751065791,0.1322727203,-0.1272502393,-0.0176834743,-0.0093516922,-0.0492710769,0.026653165,-0.3291062713,0.0641793385,-0.1466959417,-0.2652193308,-0.217539221,-0.0490310714,0.0365920998,0.0408451594,-0.0651323944,0.0326203592,-0.1210747808,-0.1073820889,-0.0522526205,-0.3160063922,0.3485418558,0.3667268753,0.0898686796,-0.0278411098,0.6427884102,0.1550075859,-0.1114492118,0.2014160454,0.1498375088,-0.1177357435,-0.2359752059,0.0229507461,0.3052551746,-0.1143825501,-0.3438960016,-0.4947920144,0.6355996132,0.1798554659,-0.5693950057,-0.1701572686,-0.2464660406,-0.1722551435,-0.1156327501,0.3047385216,0.3367748559,0.1070007458,-0.0171868652,-0.3678515553,0.0672688037,0.4848318994,-0.4563578069,-0.0733272582,-0.1196065918,0.166735813,0.2647439837,-0.1948134601,-0.9240174294,-0.1374066919,0.1441078782,-0.0140729547,-0.178728655,0.298996985,0.2460008711,0.2937105,-0.2554591596,0.5315845013,-0.0495374463,-0.0996771157,-0.4590679705,-0.2120485902]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/224","title":"[Feature Request\/Help] BLEURT model -> PyTorch","comments":"Hi there. We ran into the same problem this year (converting BLEURT to PyTorch) and thanks to @adamwlev found his colab notebook which didn't work but served as a good starting point. Finally, we **made it work** by doing just two simple conceptual fixes: \r\n\r\n1. Transposing 'kernel' layers instead of 'dense' ones when copying params from the original model;\r\n2. Taking pooler_output as a cls_state in forward function of the BleurtModel class.\r\n\r\nPlus few minor syntactical fixes for the outdated parts. The result is still not exactly the same, but is very close to the expected one (1.0483 vs 1.0474).\r\n\r\nFind the fixed version here (fixes are commented): https:\/\/colab.research.google.com\/drive\/1KsCUkFW45d5_ROSv2aHtXgeBa2Z98r03?usp=sharing \r\n","body":"Hi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Twitter).\r\n\r\nI had a go of just like manually using the checkpoint that they publish which includes the weights. It seems like the architecture is exactly aligned with the out-of-the-box BertModel in transformers just with a single linear layer on top of the CLS embedding. I loaded all the weights to the PyTorch model but I am not able to get the same numbers as the BLEURT package's python api. Here is my colab notebook where I tried https:\/\/colab.research.google.com\/drive\/1Bfced531EvQP_CpFvxwxNl25Pj6ptylY?usp=sharing . If you have any pointers on what might be going wrong that would be much appreciated!\r\n\r\nThank you muchly!","comment_length":109,"text":"[Feature Request\/Help] BLEURT model -> PyTorch\nHi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Twitter).\r\n\r\nI had a go of just like manually using the checkpoint that they publish which includes the weights. It seems like the architecture is exactly aligned with the out-of-the-box BertModel in transformers just with a single linear layer on top of the CLS embedding. I loaded all the weights to the PyTorch model but I am not able to get the same numbers as the BLEURT package's python api. Here is my colab notebook where I tried https:\/\/colab.research.google.com\/drive\/1Bfced531EvQP_CpFvxwxNl25Pj6ptylY?usp=sharing . If you have any pointers on what might be going wrong that would be much appreciated!\r\n\r\nThank you muchly!\nHi there. We ran into the same problem this year (converting BLEURT to PyTorch) and thanks to @adamwlev found his colab notebook which didn't work but served as a good starting point. Finally, we **made it work** by doing just two simple conceptual fixes: \r\n\r\n1. Transposing 'kernel' layers instead of 'dense' ones when copying params from the original model;\r\n2. Taking pooler_output as a cls_state in forward function of the BleurtModel class.\r\n\r\nPlus few minor syntactical fixes for the outdated parts. The result is still not exactly the same, but is very close to the expected one (1.0483 vs 1.0474).\r\n\r\nFind the fixed version here (fixes are commented): https:\/\/colab.research.google.com\/drive\/1KsCUkFW45d5_ROSv2aHtXgeBa2Z98r03?usp=sharing \r\n","embeddings":[0.0310527328,-0.2103960514,0.1103106588,0.2494825572,0.2850297391,-0.3368865252,0.298024863,0.1486453265,0.0918903202,0.1669930518,-0.3092391789,0.3939811289,-0.4846557379,0.2546459734,0.0468796566,-0.0895405188,-0.0679999739,0.0589459985,0.1028213054,-0.1792573035,-0.1125464216,0.1559166908,0.1110634729,-0.1056422293,-0.1948655099,0.3722063899,-0.066964075,0.1105119884,-0.2348370403,-0.0558178537,0.2399424314,-0.093134962,0.1971631199,0.7407881021,-0.0001230124,-0.2967883348,0.1056205407,-0.0138176922,-0.3956662118,-0.2290660292,0.3934163749,-0.3064046502,0.1475394517,-0.1882185936,-0.0232359432,-0.0468298905,-0.2829231024,-0.1112557352,0.1309197545,0.2489177138,0.0356830359,0.1206979081,-0.0400364399,0.1071800739,0.173642844,-0.192170009,-0.0252399687,0.5683526993,0.3351753652,-0.0194226112,-0.2249428928,0.1777407378,0.0114936503,0.1007300913,0.661795795,0.0406068414,-0.0507665649,-0.046323657,0.0426528454,0.3301609159,-0.1137454659,-0.1908951998,-0.2550768852,0.1223970577,0.0724894702,-0.2415703833,-0.2309271246,0.1607208848,0.1085655168,-0.2415809184,-0.3910423815,-0.3491120636,-0.1409149766,0.1379454583,-0.2982500196,0.309289068,-0.0680692419,-0.1271192729,0.1534783542,0.0853556842,0.3418859243,-0.0955543742,0.3543405831,0.1970385611,0.0863438398,-0.1823852807,0.222573489,0.0421038233,-0.0624436848,-0.3093143702,-0.1253534406,0.2069179267,-0.2814028859,0.0117968516,-0.027306417,0.3203355372,0.040418759,0.3597744405,0.2862763405,0.3707205355,0.0775256529,-0.0453638025,-0.0890704319,-0.1548777819,0.2409738749,0.1922279596,-0.2800593674,-0.3287021518,-0.0337699726,-0.0541970655,-0.2176968604,-0.1059070975,0.3486952484,0.1363616735,-0.4923024774,-0.0643925518,0.6397174001,0.0468702875,-0.1819374114,0.0798487812,-0.1414922029,-0.0950845033,-0.3073863387,0.0675833821,0.2125045359,-0.1938531697,0.2788228095,0.0320009142,0.5187744498,0.1252882928,0.1617130041,0.4335998297,0.1615476757,-0.1924098581,-0.2416658998,-0.2961929142,0.226289317,0.0264559891,-0.1124014631,0.0893841907,-0.1287229359,-0.3263468742,-0.1520962715,0.0239600185,-0.2871274352,-0.0191908292,-0.0109536378,0.2780379951,0.1098789126,-0.2487730086,-0.0321970731,-0.3951019347,-0.1244252101,0.0946157575,0.4095953405,-0.203890264,-0.4032551944,-0.4273768365,0.1044918299,0.4528109729,0.3346205354,0.0944755077,-0.2226804495,-0.2060235441,0.1822289079,-0.3569924235,0.2496623844,0.1920628697,-0.2746546566,-0.2299157977,0.1334045827,-0.2530925572,-0.2153822482,0.0259451009,-0.1508406401,-0.0636427253,0.3645896614,0.1104431152,-0.1822930276,-0.0231459234,-0.3769241273,-0.1707117409,0.2408783883,0.1452450305,0.3112163544,0.0442182533,0.1397858858,0.5452295542,0.109492965,-0.3767819405,0.0386493839,-0.2087074518,0.4394527376,-0.3888003826,0.362914592,-0.132294625,-0.043620877,0.2286232859,-0.3252677619,0.4498291314,0.1517201066,0.1373924017,-0.1164109409,-0.1146542281,0.2857363224,-0.0415584221,-0.0985841379,-0.1753855646,-0.2840968966,-0.2204536945,-0.1479412317,-0.254697293,-0.2391453236,-0.0517394692,0.1921084821,0.3260833323,-0.063496843,-0.1639235467,-0.0386807695,0.3862399757,0.1850811243,0.2176853418,-0.0757845417,-0.2285495698,0.1139377654,-0.2082346231,0.1406573802,0.196056515,0.385183841,0.0316243209,0.4031256735,0.2144813538,-0.1114635989,-0.0912845656,-0.1220181286,0.2812281251,0.3246671557,0.2669116259,0.2754970491,0.1210632324,-0.2493489385,0.1069417074,-0.0453744121,0.1015534475,0.3515598178,-0.1683177948,0.0646203905,-0.1682455838,0.1615700424,0.2956642807,0.2568995357,-0.0597676449,-0.0450268425,0.2192204893,-0.4693132937,-0.2722736299,0.0111395614,0.0704007968,0.0824523643,0.0731727779,0.4600592256,0.1447232962,-0.0168324653,0.0277309604,-0.086017102,0.1478748173,-0.1126614809,-0.0631481409,0.1250304729,0.0686542541,0.0781423524,0.1039386019,0.0552336797,0.1424942166,-0.1808573455,0.3385642171,-0.0062667448,0.4345985651,-0.2493534684,-0.3888390958,-0.2261955589,-0.2483255863,-0.0348907895,0.3237876296,-0.0133988503,-0.1362865865,0.1992494613,0.4045144618,-0.0856476873,-0.0558412038,-0.1134730801,-0.2637962103,-0.366084516,-0.001267113,-0.0641445741,0.0651266128,-0.0870532617,0.0358788744,-0.1546071768,-0.1930067986,-0.7795089483,0.0602129996,0.0392719209,0.300018847,0.205970645,-0.1349012256,-0.3109205067,0.0444471091,0.1018146574,-0.3742019832,-0.026989758,0.417681843,-0.2349728495,0.0767093226,-0.0332292654,-0.2210605443,-0.1288671196,-0.0225678999,-0.0551790744,0.1210982949,0.0652783662,-0.2043250501,0.1393239349,0.3219200671,0.4152190387,-0.098772198,-0.0671397075,0.2249228954,0.427801013,-0.2569454908,-0.3013522327,-0.1849201471,-0.0005780608,0.0172542986,0.000376916,-0.323782891,-0.3601538539,-0.1146064624,-0.1512562037,0.202199921,0.1691578627,0.061578963,0.2242957354,0.0646770298,-0.2167789489,-0.1627930999,-0.0702314898,0.2430413514,0.6987832785,-0.2315392196,-0.0999535173,0.08970429,0.7524694204,0.1305748224,-0.4231865108,0.0625411049,0.260733664,0.0608615875,-0.113940388,-0.2671453357,0.1556501985,-0.3145405054,-0.2764021754,0.3156627119,0.1146013513,-0.280421257,-0.2611299753,-0.2004965544,0.4979617894,-0.1756325811,0.0696367994,0.034685351,-0.0499780141,0.0228224136,-0.1170801222,-0.2740621865,-0.1145454645,0.2060011476,0.0220862851,0.0904170349,0.036680501,0.4527597725,0.0155743454,-0.7020940781,0.4291401207,0.0912268013,-0.6415277123,0.2007907033,-0.1509196907,-0.0830677152,-0.0532336161,-0.3633197248,0.1148818061,0.0365857817,-0.0634057522,-0.0249022041,-0.2808913589,0.0199358985,-0.2606969178,-0.3109962046,0.6108304262,0.3093993068,-0.3783528805,0.0513239987,0.2382778078,0.4477616549,-0.1608959585,0.035822764,-0.3568872809,-0.2157523632,0.0927283838,-0.1170455515,-0.039722614,0.2884736061,-0.1285486668,-0.0787201598,0.2440050542,-0.0877478793,0.1997198015,0.0497005768,0.1096034497,-0.0948295891,0.1706187129,0.0504971705,0.0647140145,0.2465958148,0.4144496918,-0.5077793002,-0.4713501334,0.2168165147,0.1612162441,0.1713764369,0.1093557253,0.1049197391,-0.1360114366,0.4354983866,0.2722870111,0.1856850535,0.2338026613,0.0391736701,0.3682483435,0.1187831759,-0.1948879063,0.3924103677,-0.1706740707,0.0144212283,0.2270653695,0.066492565,-0.0122716231,0.1219392121,0.4337999225,0.9371419549,0.142219618,0.0944437161,0.2991998494,-0.1356031448,0.3273809552,-0.0459537059,0.0288540274,-0.3419891298,-0.26656425,0.0436376408,-0.0495897681,-0.4452964664,-0.1622665077,-0.1985961497,0.1473296732,-0.3593905568,-0.2568362355,0.2827571332,0.28083691,0.1088479385,0.0309322998,0.2770261765,0.0805539638,-0.1751108468,0.2692439556,-0.0848746896,-0.1843997389,-0.2621049583,0.1320784837,-0.1802485138,0.3719127476,-0.1005960405,-0.0448558517,-0.0609015934,-0.3492490947,0.0737573579,0.0327626206,0.7054663897,0.0923163295,-0.3070918322,0.0802794918,-0.0621383898,0.1774423867,0.1961669177,-0.1495553851,0.3595603406,-0.1338005364,-0.1622400284,-0.1679967642,0.1469708532,-0.3749420643,0.2212458551,-0.1645991951,-0.0455489196,-0.1182498336,0.4033854902,0.334535718,0.1593397856,-0.255320698,0.0771498829,0.203279987,-0.2602018714,0.0458195284,0.116062142,-0.2364223003,0.0402359217,-0.0099044545,0.0215246771,-0.0013387915,0.2349434346,0.1742056906,-0.1145671234,0.2359017581,-0.6396334767,0.5488190055,-0.3684133887,0.175715059,-0.1360732019,-0.055177968,0.1222446933,0.5733029842,0.435783118,0.2145937532,-0.3986991048,0.1854686439,-0.3889478147,0.0877297595,-0.1864099354,0.3603316545,-0.1763778925,0.0029712766,-0.1169335991,0.2312009037,-0.2067342103,0.0399377421,0.1982919723,-0.0306832828,0.0095589301,-0.4418539107,0.0274868663,0.0358331688,-0.2390349358,-0.0639577508,0.1150896773,-0.0228963252,-0.4367915988,0.1588521898,0.2320314348,0.0091410177,-0.3205921948,-0.3893790245,-0.1304402351,-0.1896283329,0.06611339,0.0273705442,0.0649443045,0.2162599415,0.207943961,0.0203054696,-0.1601992548,-0.0488762669,0.4102194309,0.1664810926,0.0091140894,-0.0343408659,0.4261155725,-0.0280621611,0.1916098595,0.0241176616,-0.0257500336,0.1640705168,0.4158225358,0.1315031052,0.1530122012,-0.3915197849,0.3136760294,0.2520202696,0.2004140317,-0.2012351602,0.2113129348,0.0853198767,-0.2471276522,-0.1780573577,-0.2090687454,0.0501729622,-0.137120679,0.1760263294,-0.1112417802,0.3604824245,0.3827470243,-0.0612405799,-0.2361859381,-0.2675483227,0.050848864,0.0448315963,-0.2170985788,0.3127976954,0.5844194293,0.0440050624,0.0522158407,-0.3825649321,-0.1198428348,0.0084645124,-0.2900799215,-0.1088458598,0.358864218,0.2046005875,-0.1092017144,-0.0990248844,0.1512819976,-0.0717027932,-0.1213439405,0.3947104812,-0.4014695585,-0.1622598767,-0.0369143896,0.3568064272,-0.0266297571,0.0682641864,0.1088075861,-0.2161393017,0.0773333162,-0.0150254108,-0.0182749964,0.0922215655,0.0747444704,0.1013407186,-0.094475992,0.1506614685,-0.7600205541,0.3500680029,0.2666804492,0.0775581002,-0.1919765025,0.1510921866,-0.1153282002,-0.236151576,0.383100152,-0.0166963339,0.2577486634,-0.3252430856,-0.0414283201,0.4066012502,-0.0235648658,0.0434331074,0.1191044226,0.4557109177,-0.137135312,-0.0801688433,-0.0849497989,-0.1710611433,0.334772557,-0.3921130598,0.2219242454,0.057304617,0.8864088655,0.3305407763,-0.2043890357,-0.2726921439,-0.0908331424,-0.2980068922,0.0467532873,0.1034178957,-0.0821732357,0.0275858641,-0.0277837627,0.0060583879,0.1241200715,0.1421762258,0.267565757,0.2982497811,-0.3599913418,0.2530737519,-0.2842273712,0.1555223167,0.182070449,0.0662230849,-0.2323411405,-0.0077620847,0.071177341,-0.1051210463,0.0571733415,-0.2300207615,0.0792560726,-0.2101911604,-0.2341049165,-0.2609882355,0.0678260624,-0.0144277206,0.0045178039,-0.0863476247,0.0231519304,-0.114505887,-0.0945779234,0.0212414637,-0.287833333,0.3728129268,0.3003582656,0.1216574609,0.05715679,0.5850401521,0.1294901073,-0.1710172296,0.2642609179,0.1361714154,-0.1034356952,-0.1814621389,-0.0433797389,0.2637507319,-0.0508974046,-0.263594389,-0.5351043344,0.5506891608,0.1140035614,-0.5117967129,-0.228952527,-0.161784336,-0.185589835,-0.1099147871,0.2482443154,0.381172508,0.1388430893,-0.0286270641,-0.4515781701,0.0040644482,0.4650700688,-0.4754207134,-0.1233692467,-0.121230267,0.1840686947,0.2990711629,-0.1291980445,-0.8295275569,-0.1543298662,0.0958310738,-0.0380874947,-0.1388419718,0.2332542986,0.1172419861,0.254024595,-0.2943967283,0.5520856977,0.0463992134,-0.1617400795,-0.4791437387,-0.1919283122]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/223","title":"[Feature request] Add FLUE dataset ","comments":"Hi @lbourdois, yes please share it with us","body":"Hi,\r\n\r\nI think it would be interesting to add the FLUE dataset for francophones or anyone wishing to work on French.\r\n\r\nIn other requests, I read that you are already working on some datasets, and I was wondering if FLUE was planned.\r\n\r\nIf it is not the case, I can provide each of the cleaned FLUE datasets (in the form of a directly exploitable dataset rather than in the original xml formats which require additional processing, with the French part for cases where the dataset is based on a multilingual dataframe, etc.).","comment_length":8,"text":"[Feature request] Add FLUE dataset \nHi,\r\n\r\nI think it would be interesting to add the FLUE dataset for francophones or anyone wishing to work on French.\r\n\r\nIn other requests, I read that you are already working on some datasets, and I was wondering if FLUE was planned.\r\n\r\nIf it is not the case, I can provide each of the cleaned FLUE datasets (in the form of a directly exploitable dataset rather than in the original xml formats which require additional processing, with the French part for cases where the dataset is based on a multilingual dataframe, etc.).\nHi @lbourdois, yes please share it with us","embeddings":[-0.1571975648,0.1648371816,-0.2408254296,-0.0027440556,-0.0637981668,0.0776519179,0.2919570804,-0.1484845132,-0.2900182009,0.1033369824,-0.0504840128,-0.0560541861,-0.1837443411,0.3868722022,0.0721644163,-0.3618884087,0.0490605198,0.10371124,-0.026661614,-0.3453487158,-0.1034600362,0.1838845462,-0.0789502636,0.0476215407,-0.1176375076,0.1978092194,-0.1170755923,-0.1613847017,-0.2835241854,-0.0693714693,0.0700475052,0.2195388079,0.0306842979,0.1603794098,-0.000099218,-0.3003520072,0.2816765308,-0.0343055464,-0.0578518622,0.1505931318,-0.0804583132,-0.2974053919,-0.3745985925,0.1108591631,-0.2186505497,-0.1276953816,0.0423198417,-0.5864566565,-0.1697543561,0.3775184453,0.2590579689,-0.1254160851,-0.0951790065,-0.0867417157,0.3427216113,0.367356658,-0.3650538921,-0.1493513882,0.6243764162,0.099039197,0.145412311,0.318125397,0.3143677413,-0.4097227156,-0.1712392271,-0.1058821008,-0.0443409421,-0.2879481018,0.0939455554,0.3221336901,0.6892565489,-0.0986394137,-0.4605006874,-0.042046193,0.2227955908,-0.253100425,-0.0734540373,-0.1941858977,0.1624700129,0.1415291727,0.2902344465,-0.5130160451,-0.1363303065,0.0179304052,-0.0340539739,0.1751239002,-0.044549372,-0.1482945681,-0.1187046543,-0.1098076925,0.0394850709,0.2203493565,-0.2215445042,0.2641579509,-0.1088355854,-0.2480177134,0.0253410283,-0.0114109823,0.3961032331,0.2010546327,-0.5701758265,0.0982544273,-0.2604819834,-0.112342976,0.198819682,-0.0722158179,0.2700899243,-0.3748680353,0.2370253503,-0.4488073885,0.1510453373,-0.0929473117,-0.0470911004,0.0488158278,-0.3344818652,0.2024626732,0.0666543469,-0.3761241436,0.0804480165,0.05834122,0.334414959,-0.1874759197,-0.07924591,0.1485387087,0.1221116185,0.2276098132,-0.4002845287,0.1112362593,-0.1317908168,-0.73826617,-0.1412623674,0.0038522563,-0.2967792153,0.0140440948,0.1113244593,-0.2570217252,-0.0079946499,0.12115293,0.0649513379,0.3607256711,0.2045423388,-0.2728264332,-0.1144444495,-0.2638740242,-0.3034692705,0.0180145949,-0.0791552514,0.3360207379,-0.2861189246,-0.1150518879,0.0870771334,-0.0017120857,-0.6371824741,0.2577339113,0.2396671921,-0.3007862568,-0.0405846089,0.699113965,-0.1627551913,-0.3098374307,0.1391069889,0.0000067655,-0.1779897064,-0.1293698698,-0.0661747754,0.1814620197,-0.4428762197,-0.0397279896,-0.2063411176,-0.4188448191,0.3121559322,0.1745320261,0.008925559,0.024764169,-0.2959823906,0.0205167271,0.1559781432,-0.0523086265,-0.0205835029,0.0922434852,-0.1739788502,0.0598629564,0.0756726786,0.1191535294,0.2125421315,-0.1125014946,0.2805592716,0.5226479769,-0.2591293454,-0.0936277062,-0.0616344474,-0.2690910399,0.1590843052,0.1563926935,0.0545583181,-0.0049814326,0.3129892647,-0.2451378703,0.1290516555,-0.062876977,0.1540843695,0.1881223172,0.2997649312,0.3700167835,0.0538810566,-0.0344637856,-0.165534392,-0.167508021,-0.2071055174,0.2932342589,0.3375841677,-0.2574677765,-0.0232769344,-0.0593027063,-0.0719427317,-0.2670461833,0.3032129705,0.2909139693,0.0212603547,0.3478708863,-0.3014670908,0.249381125,-0.1296218187,-0.1386229843,-0.1467915475,0.2266427279,0.0780680999,-0.0564776547,-0.150392279,0.2959656417,-0.0590339229,-0.0162381362,-0.0949034393,-0.1985612959,0.2961705923,0.2360093445,0.5194109678,0.1584222466,0.3195595145,-0.1748271883,-0.1215795651,0.046123255,0.0643622428,-0.1295460463,0.0640334412,0.2139478475,0.1973388791,-0.2100236416,-0.0475601368,-0.0004629883,0.3758299947,0.1486423314,-0.1205265149,-0.1645936668,0.3208877146,0.0982593,0.0833641365,-0.1091936752,-0.4560949504,0.0885794982,0.4289090335,-0.0012225818,0.0653099194,0.0670327395,-0.1020800471,-0.1186194718,0.3892419934,-0.4187993705,0.1184500158,0.2828919291,0.2697145641,-0.1149519682,0.2347000986,-0.035324499,0.1103394702,-0.3406774104,-0.1193151027,-0.0313177854,0.1866580397,-0.0385047644,-0.4352472126,-0.1694073379,-0.0166693311,0.0207060855,0.2124689221,-0.2278802246,-0.0134743294,-0.2197035849,0.2390443683,-0.2614557445,-0.2464563251,-0.1756892204,0.3139949739,-0.2487937957,-0.0845646039,0.0932388455,-0.1836949885,0.4474074543,-0.1566990018,0.0529643409,-0.3238340616,0.0199384131,-0.3906003833,0.2894725204,-0.0539696738,-0.1399634928,0.5517076254,-0.0779292062,0.3230818808,-0.4269552529,-0.5252372622,0.2225548327,-0.0323533379,0.4807910919,0.1109147072,-0.0655439273,0.1144865155,0.0016941131,0.0195783284,0.0311619639,-0.1813083589,-0.0867046416,-0.0572508797,0.3115543723,0.0019472563,-0.2780660093,-0.3355764747,-0.2513000071,0.2469803393,-0.2464488745,0.0110956775,0.0527452417,-0.2329930514,0.035325218,-0.1890259832,0.0882666707,-0.3286494315,-0.0084848851,0.1976601779,-0.4359731376,-0.2755902112,-0.0291002579,-0.0340199769,0.0838494971,-0.1890728176,-0.2776739895,0.2189752311,-0.0712550431,0.2682801783,0.0072015868,0.0334614292,0.196870327,0.1790084094,-0.1113813818,0.0880662575,-0.0140177039,-0.2135346383,0.0904931724,0.0727392063,0.0034335987,-0.1783781797,-0.0615563244,0.1817291975,-0.0467534885,0.124807559,0.4058946967,0.0587274246,0.2206020057,-0.2972090244,-0.44336766,0.3303575814,0.0229253788,0.0396362208,0.3588134944,0.2010423243,-0.2976070046,-0.1148850545,-0.100442186,0.2762099802,-0.0594169237,0.1257928461,-0.1565860063,0.1233516559,-0.1861754209,-0.1999118179,-0.1878920794,-0.0732987225,0.2711290419,0.1371707618,0.1785750836,-0.1404874027,-0.2788748145,0.1126251072,-0.3190094829,-0.0053064208,0.0426545069,0.2341727912,-0.3508066535,0.3369279802,-0.0109127136,0.1551950127,0.2397933751,-0.6747989058,-0.4994166791,0.1208339557,0.3115163743,0.2805324495,0.2205561697,-0.1329346001,-0.0775198266,0.1117772311,0.3010512888,-0.1597467959,-0.0255358294,0.1677839458,0.408077985,-0.1089599207,0.044226788,0.1292524934,-0.1761252433,-0.0881109089,0.3731768727,-0.0782836527,-0.036701072,0.1723708659,0.0737843663,-0.1237877607,0.2204201221,0.2653273344,-0.0537239984,-0.1067194119,0.1593315303,0.0732437596,-0.1609260887,-0.3088524342,0.210198462,0.2073995918,-0.0141431009,-0.479434818,0.0015724297,-0.3235547841,0.5554083586,0.2058332264,0.2389437854,0.2919135392,-0.2893553972,-0.1583154351,-0.1262310296,0.1107021198,0.3214547932,0.0454321541,-0.3737743795,-0.6810734868,0.0155299092,-0.0418271273,-0.0095058884,0.4384651184,0.34328264,-0.2063468844,0.1076428965,0.1432348639,0.9337934256,0.3423294723,0.0772115737,0.0219662152,-0.1552367657,0.1793423593,-0.3377551436,-0.0535795912,-0.0564112104,-0.3242915273,-0.0290465187,0.2230990231,0.3761298358,0.1372530758,-0.2707894146,0.0860958844,0.048509229,-0.2096797228,0.1587883979,0.448210746,-0.0074842917,-0.3871766925,0.1141603664,0.1951401681,0.1186353415,0.2135228217,-0.0956509858,-0.201292187,0.1937164664,0.072959587,-0.4014725983,-0.069808729,-0.3436788023,-0.0026994601,-0.4082935452,0.0512213036,0.1867562979,-0.2069029361,0.3767954707,0.4653349817,0.0097535197,0.0069082235,0.4573948383,0.0614480004,-0.168005079,-0.1681310534,0.2352194786,-0.1991744488,-0.2442047596,0.3919271231,-0.0505659766,-0.0337991603,-0.4278904498,-0.1029850021,-0.1984360665,-0.0419667661,-0.1513150185,0.3929704726,0.147668466,-0.2334414124,0.1934995055,0.4368274212,-0.1665956229,-0.1414902806,0.220107317,-0.04937502,-0.3094932437,-0.2025575787,0.2531112731,-0.7219631076,0.2119178325,0.3206340373,-0.1686739773,-0.3067215979,0.1742969453,0.2786849439,-0.4217892885,-0.3220278323,0.0279967561,-0.2021339387,-0.127965495,0.1845484078,0.0220193397,-0.2489657551,0.1679767519,-0.0503182635,-0.3562769592,0.0839450657,0.1296074986,0.1682798415,0.1553553641,-0.3406390846,0.0398576111,0.0417162254,-0.431232065,-0.0819240585,0.0017832083,0.0167377014,0.2815385461,0.0755653307,-0.1115679964,0.023248015,0.2111202031,-0.2056799084,-0.0845680982,-0.2183114588,-0.0620929897,0.0772673935,0.0285213739,0.1768316776,0.036295332,0.0107228868,-0.1081673801,-0.0803807601,0.0869112983,0.0216150247,-0.0723984241,0.3320108652,0.0197350793,0.0494479164,-0.0059196185,0.222582981,0.373696506,0.0949334353,0.1600058079,0.2186093479,0.1246466264,-0.1404054612,0.1987055838,0.2625482976,0.0276842378,0.2277781665,-0.1577544957,0.0778512806,0.2399376631,0.2765473127,-0.194527179,-0.0309092738,-0.1477248073,-0.0111637702,0.4479542673,0.3083030283,-0.3638509512,0.03355534,0.1551599056,-0.0815438107,0.1414859593,0.1062851995,0.3585386276,0.3576259613,0.3275825679,-0.2061493695,-0.2459660321,-0.0007695614,0.3349812031,0.2200025469,-0.0514787398,-0.0864467248,0.0092554381,0.4214717746,-0.0335159488,0.3784300983,-0.0055067074,0.2985835075,0.3266586661,0.1549786925,0.0481924154,-0.1259978265,0.4186437428,0.3189263642,0.1915394217,0.3080093563,0.2089341879,-0.0863794684,0.0601199567,-0.2050364614,-0.1806337237,0.2008330226,-0.1073885337,-0.0745093524,0.3193294704,-0.2526195347,-0.1534714848,-0.2526654303,0.0253671985,0.1117110774,-0.0868937671,0.0569785349,0.2066713572,0.000779769,0.2043112218,-0.0228094533,-0.0144156059,-0.0450566486,0.2603560388,-0.0614446439,0.1431421489,0.0953205004,0.3383303285,-0.1180848256,0.1090658009,-0.0381835252,0.2647869587,-0.2605814636,-0.0359260738,-0.0284845419,0.3812919259,0.1126691625,0.2138061821,0.2704371512,0.1473944485,-0.1745328307,0.0665364563,0.181237191,-0.2827352583,-0.3252443373,0.0041792621,0.1452616304,0.036043983,-0.3990622461,-0.104424566,-0.1946916133,-0.1772363186,0.4560508132,-0.1336128712,-0.0856784135,-0.0601633452,0.1528564394,0.0975019261,-0.1219761372,0.3668352664,0.3591458499,-0.0293083806,-0.2369991243,-0.3837893307,0.0992041081,0.0236333273,0.1430050284,-0.045018591,-0.0634007454,0.4633854628,0.3034484684,-0.0424557365,-0.0566498116,0.1488531828,0.25858742,-0.3820648789,0.0865302533,0.3817696273,0.360555768,-0.0500278547,-0.086008057,0.0860854462,0.2932324708,0.1851744205,-0.2023937553,-0.1528824568,0.1513191164,0.0435893163,0.3191004694,0.20204781,0.3946969807,-0.2910136878,-0.1622463614,-0.1358941495,-0.1925205141,-0.0476761982,0.1521410197,0.1536853462,-0.0596682616,-0.0221794937,-0.3544336557,-0.2077064812,0.3371202052,0.0366055816,0.3083453774,-0.4206288457,0.26571697,0.0954462513,-0.3897592723,-0.1978318095,0.064532347,0.0158567969,-0.09112259,-0.0549944155,-0.0972010866,0.0673231632,-0.26813519,-0.3471356928,-0.2114605904,0.0657384545,0.2888139188,0.1209265366,-0.2209200114,0.1416282356,0.2036983371,0.1010239646,0.3345006704,0.0956626236,0.2283110172,-0.0308867656,-0.2548967898,0.3389899731,0.1151027754,0.0849374905,0.0334648937,-0.2604377866]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/223","title":"[Feature request] Add FLUE dataset ","comments":"@mariamabarham \r\nI put all the datasets on this drive: https:\/\/1drv.ms\/u\/s!Ao2Rcpiny7RFinDypq7w-LbXcsx9?e=iVsEDh\r\n\r\n\r\nSome information : \r\n\u2022 For FLUE, the quote used is\r\n\r\n> @misc{le2019flaubert,\r\n> title={FlauBERT: Unsupervised Language Model Pre-training for French},\r\n> author={Hang Le and Lo\u00efc Vial and Jibril Frej and Vincent Segonne and Maximin Coavoux and Benjamin Lecouteux and Alexandre Allauzen and Beno\u00eet Crabb\u00e9 and Laurent Besacier and Didier Schwab},\r\n> year={2019},\r\n> eprint={1912.05372},\r\n> archivePrefix={arXiv},\r\n> primaryClass={cs.CL}\r\n> }\r\n\r\n\u2022 The Github repo of FLUE is avaible here : https:\/\/github.com\/getalp\/Flaubert\/tree\/master\/flue\r\n\r\n\r\n\r\nInformation related to the different tasks of FLUE : \r\n\r\n**1. Classification**\r\nThree dataframes are available: \r\n- Book\r\n- DVD\r\n- Music\r\nFor each of these dataframes is available a set of training and test data, and a third one containing unlabelled data.\r\n\r\nCitation : \r\n>@dataset{prettenhofer_peter_2010_3251672,\r\n author = {Prettenhofer, Peter and\r\n Stein, Benno},\r\n title = {{Webis Cross-Lingual Sentiment Dataset 2010 (Webis- \r\n CLS-10)}},\r\n month = jul,\r\n year = 2010,\r\n publisher = {Zenodo},\r\n doi = {10.5281\/zenodo.3251672},\r\n url = {https:\/\/doi.org\/10.5281\/zenodo.3251672}\r\n}\r\n\r\n\r\n**2. Paraphrasing** \r\nFrench part of the PAWS-X dataset (https:\/\/github.com\/google-research-datasets\/paws).\r\nThree dataframes are available: \r\n- train\r\n- dev\r\n- test \r\n\r\nCitation : \r\n> @InProceedings{pawsx2019emnlp,\r\n> title = {{PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification}},\r\n> author = {Yang, Yinfei and Zhang, Yuan and Tar, Chris and Baldridge, Jason},\r\n> booktitle = {Proc. of EMNLP},\r\n> year = {2019}\r\n> }\r\n\r\n\r\n\r\n**3. Natural Language Inference**\r\nFrench part of the XNLI dataset (https:\/\/github.com\/facebookresearch\/XNLI).\r\nThree dataframes are available: \r\n- train\r\n- dev\r\n- test \r\n\r\nFor the dev and test datasets, extra columns compared to the train dataset were available so I left them in the dataframe (I didn't know if these columns could be useful for other tasks or not). \r\nIn the context of the FLUE benchmark, only the columns gold_label, sentence1 and sentence2 are useful.\r\n\r\n\r\nCitation : \r\n\r\n> @InProceedings{conneau2018xnli,\r\n> author = \"Conneau, Alexis\r\n> and Rinott, Ruty\r\n> and Lample, Guillaume\r\n> and Williams, Adina\r\n> and Bowman, Samuel R.\r\n> and Schwenk, Holger\r\n> and Stoyanov, Veselin\",\r\n> title = \"XNLI: Evaluating Cross-lingual Sentence Representations\",\r\n> booktitle = \"Proceedings of the 2018 Conference on Empirical Methods\r\n> in Natural Language Processing\",\r\n> year = \"2018\",\r\n> publisher = \"Association for Computational Linguistics\",\r\n> location = \"Brussels, Belgium\",\r\n\r\n\r\n**4. Parsing**\r\nThe dataset used by the FLUE authors for this task is not freely available.\r\nUsers of your library will therefore not be able to access it.\r\nNevertheless, I think maybe it is useful to add a link to the site where to request this dataframe: http:\/\/ftb.linguist.univ-paris-diderot.fr\/telecharger.php?langue=en \r\n(personally it was sent to me less than 48 hours after I requested it).\r\n\r\n\r\n**5. Word Sense Disambiguation Tasks**\r\n5.1 Verb Sense Disambiguation\r\n\r\nTwo dataframes are available: train and test\r\nFor both dataframes, 4 columns are available: document, sentence, lemma and word.\r\nI created the document column thinking that there were several documents in the dataset but afterwards it turns out that there were not: several sentences but only one document. It's up to you to keep it or not when importing these two dataframes.\r\n\r\nThe sentence column is used to determine to which sentence the word in the word column belongs. It is in the form of a dictionary {'id': 'd000.s001', 'idx': '1'}. I thought for a while to keep only the idx because the id doesn't matter any more information. Nevertheless for the test dataset, the dictionary has an extra value indicating the source of the sentence. I don't know if it's useful or not, that's why I left the dictionary just in case. The user is free to do what he wants with it.\r\n\r\nCitation : \r\n\r\n> Segonne, V., Candito, M., and Crabb \u0301e, B. (2019). Usingwiktionary as a resource for wsd: the case of frenchverbs. InProceedings of the 13th International Confer-ence on Computational Semantics-Long Papers, pages259\u2013270\r\n\r\n5.2 Noun Sense Disambiguation\r\nTwo dataframes are available: 2 train and 1 test\r\n\r\nI confess I didn't fully understand the procedure for this task.\r\n\r\nCitation : \r\n\r\n> @dataset{loic_vial_2019_3549806,\r\n> author = {Lo\u00efc Vial},\r\n> title = {{French Word Sense Disambiguation with Princeton \r\n> WordNet Identifiers}},\r\n> month = nov,\r\n> year = 2019,\r\n> publisher = {Zenodo},\r\n> version = {1.0},\r\n> doi = {10.5281\/zenodo.3549806},\r\n> url = {https:\/\/doi.org\/10.5281\/zenodo.3549806}\r\n> }\r\n\r\nFinally, additional information about FLUE is available in the FlauBERT publication : \r\nhttps:\/\/arxiv.org\/abs\/1912.05372 (p. 4).\r\n\r\n\r\nHoping to have provided you with everything you need to add this benchmark :) \r\n","body":"Hi,\r\n\r\nI think it would be interesting to add the FLUE dataset for francophones or anyone wishing to work on French.\r\n\r\nIn other requests, I read that you are already working on some datasets, and I was wondering if FLUE was planned.\r\n\r\nIf it is not the case, I can provide each of the cleaned FLUE datasets (in the form of a directly exploitable dataset rather than in the original xml formats which require additional processing, with the French part for cases where the dataset is based on a multilingual dataframe, etc.).","comment_length":723,"text":"[Feature request] Add FLUE dataset \nHi,\r\n\r\nI think it would be interesting to add the FLUE dataset for francophones or anyone wishing to work on French.\r\n\r\nIn other requests, I read that you are already working on some datasets, and I was wondering if FLUE was planned.\r\n\r\nIf it is not the case, I can provide each of the cleaned FLUE datasets (in the form of a directly exploitable dataset rather than in the original xml formats which require additional processing, with the French part for cases where the dataset is based on a multilingual dataframe, etc.).\n@mariamabarham \r\nI put all the datasets on this drive: https:\/\/1drv.ms\/u\/s!Ao2Rcpiny7RFinDypq7w-LbXcsx9?e=iVsEDh\r\n\r\n\r\nSome information : \r\n\u2022 For FLUE, the quote used is\r\n\r\n> @misc{le2019flaubert,\r\n> title={FlauBERT: Unsupervised Language Model Pre-training for French},\r\n> author={Hang Le and Lo\u00efc Vial and Jibril Frej and Vincent Segonne and Maximin Coavoux and Benjamin Lecouteux and Alexandre Allauzen and Beno\u00eet Crabb\u00e9 and Laurent Besacier and Didier Schwab},\r\n> year={2019},\r\n> eprint={1912.05372},\r\n> archivePrefix={arXiv},\r\n> primaryClass={cs.CL}\r\n> }\r\n\r\n\u2022 The Github repo of FLUE is avaible here : https:\/\/github.com\/getalp\/Flaubert\/tree\/master\/flue\r\n\r\n\r\n\r\nInformation related to the different tasks of FLUE : \r\n\r\n**1. Classification**\r\nThree dataframes are available: \r\n- Book\r\n- DVD\r\n- Music\r\nFor each of these dataframes is available a set of training and test data, and a third one containing unlabelled data.\r\n\r\nCitation : \r\n>@dataset{prettenhofer_peter_2010_3251672,\r\n author = {Prettenhofer, Peter and\r\n Stein, Benno},\r\n title = {{Webis Cross-Lingual Sentiment Dataset 2010 (Webis- \r\n CLS-10)}},\r\n month = jul,\r\n year = 2010,\r\n publisher = {Zenodo},\r\n doi = {10.5281\/zenodo.3251672},\r\n url = {https:\/\/doi.org\/10.5281\/zenodo.3251672}\r\n}\r\n\r\n\r\n**2. Paraphrasing** \r\nFrench part of the PAWS-X dataset (https:\/\/github.com\/google-research-datasets\/paws).\r\nThree dataframes are available: \r\n- train\r\n- dev\r\n- test \r\n\r\nCitation : \r\n> @InProceedings{pawsx2019emnlp,\r\n> title = {{PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification}},\r\n> author = {Yang, Yinfei and Zhang, Yuan and Tar, Chris and Baldridge, Jason},\r\n> booktitle = {Proc. of EMNLP},\r\n> year = {2019}\r\n> }\r\n\r\n\r\n\r\n**3. Natural Language Inference**\r\nFrench part of the XNLI dataset (https:\/\/github.com\/facebookresearch\/XNLI).\r\nThree dataframes are available: \r\n- train\r\n- dev\r\n- test \r\n\r\nFor the dev and test datasets, extra columns compared to the train dataset were available so I left them in the dataframe (I didn't know if these columns could be useful for other tasks or not). \r\nIn the context of the FLUE benchmark, only the columns gold_label, sentence1 and sentence2 are useful.\r\n\r\n\r\nCitation : \r\n\r\n> @InProceedings{conneau2018xnli,\r\n> author = \"Conneau, Alexis\r\n> and Rinott, Ruty\r\n> and Lample, Guillaume\r\n> and Williams, Adina\r\n> and Bowman, Samuel R.\r\n> and Schwenk, Holger\r\n> and Stoyanov, Veselin\",\r\n> title = \"XNLI: Evaluating Cross-lingual Sentence Representations\",\r\n> booktitle = \"Proceedings of the 2018 Conference on Empirical Methods\r\n> in Natural Language Processing\",\r\n> year = \"2018\",\r\n> publisher = \"Association for Computational Linguistics\",\r\n> location = \"Brussels, Belgium\",\r\n\r\n\r\n**4. Parsing**\r\nThe dataset used by the FLUE authors for this task is not freely available.\r\nUsers of your library will therefore not be able to access it.\r\nNevertheless, I think maybe it is useful to add a link to the site where to request this dataframe: http:\/\/ftb.linguist.univ-paris-diderot.fr\/telecharger.php?langue=en \r\n(personally it was sent to me less than 48 hours after I requested it).\r\n\r\n\r\n**5. Word Sense Disambiguation Tasks**\r\n5.1 Verb Sense Disambiguation\r\n\r\nTwo dataframes are available: train and test\r\nFor both dataframes, 4 columns are available: document, sentence, lemma and word.\r\nI created the document column thinking that there were several documents in the dataset but afterwards it turns out that there were not: several sentences but only one document. It's up to you to keep it or not when importing these two dataframes.\r\n\r\nThe sentence column is used to determine to which sentence the word in the word column belongs. It is in the form of a dictionary {'id': 'd000.s001', 'idx': '1'}. I thought for a while to keep only the idx because the id doesn't matter any more information. Nevertheless for the test dataset, the dictionary has an extra value indicating the source of the sentence. I don't know if it's useful or not, that's why I left the dictionary just in case. The user is free to do what he wants with it.\r\n\r\nCitation : \r\n\r\n> Segonne, V., Candito, M., and Crabb \u0301e, B. (2019). Usingwiktionary as a resource for wsd: the case of frenchverbs. InProceedings of the 13th International Confer-ence on Computational Semantics-Long Papers, pages259\u2013270\r\n\r\n5.2 Noun Sense Disambiguation\r\nTwo dataframes are available: 2 train and 1 test\r\n\r\nI confess I didn't fully understand the procedure for this task.\r\n\r\nCitation : \r\n\r\n> @dataset{loic_vial_2019_3549806,\r\n> author = {Lo\u00efc Vial},\r\n> title = {{French Word Sense Disambiguation with Princeton \r\n> WordNet Identifiers}},\r\n> month = nov,\r\n> year = 2019,\r\n> publisher = {Zenodo},\r\n> version = {1.0},\r\n> doi = {10.5281\/zenodo.3549806},\r\n> url = {https:\/\/doi.org\/10.5281\/zenodo.3549806}\r\n> }\r\n\r\nFinally, additional information about FLUE is available in the FlauBERT publication : \r\nhttps:\/\/arxiv.org\/abs\/1912.05372 (p. 4).\r\n\r\n\r\nHoping to have provided you with everything you need to add this benchmark :) \r\n","embeddings":[-0.0987504795,0.2994920313,-0.1906657368,0.0841301605,-0.0912493095,0.1056410372,0.4507476687,-0.1468382031,-0.2177065462,0.004500614,-0.207936123,-0.1749657989,-0.1512342393,0.2744062841,0.0792365074,-0.3693341315,-0.0019665421,0.0335363708,0.0868884772,-0.3447933495,-0.1076890379,0.1783014983,-0.0497566573,0.1935892403,-0.0001795929,0.0912411287,-0.2292832583,-0.1584287286,-0.1509997249,-0.0223210566,0.1317371428,0.2434061617,0.0731366873,0.3848908246,-0.0001029056,-0.2126103938,0.1268322617,-0.1261634231,0.0290071759,0.0322246701,0.0220035128,-0.2267683446,-0.3399682045,0.0377826095,-0.1280567944,-0.2536050975,-0.1875289679,-0.5613648295,-0.1343907118,0.5067580938,0.2427782714,-0.1785595417,0.1348417997,0.0144604268,0.1603562087,0.2705264688,-0.3232008219,-0.1657286733,0.6291852593,0.1499028355,0.0344640873,0.2998957336,0.2246343642,-0.371067971,-0.2069713622,-0.131523177,-0.0522462502,-0.3681867123,0.0279233847,0.3565875888,0.6327362061,-0.0815203041,-0.5868777633,-0.0946210846,0.2437147945,-0.0755497366,-0.1009749994,-0.1344939321,0.2280339748,0.19478935,0.3044316173,-0.4035966098,-0.0697431564,0.099014543,0.081435211,0.0094177499,-0.0546343103,-0.0176309925,0.1192259565,-0.1604201794,0.0108385095,0.1647719145,-0.0601687282,0.3607103229,-0.2149062902,-0.3206715286,0.0690350533,-0.0593278743,0.3171031773,0.0718134195,-0.6085301638,0.0339402743,-0.4754464328,-0.1365335733,0.2307552993,0.1410774142,0.1131886095,-0.2693189979,0.1695144325,-0.3491479158,0.1283834726,-0.0089363325,-0.0067876312,-0.0257690623,-0.2819310129,0.0620592758,-0.0937890485,-0.3450576067,0.0095341895,0.0284710582,0.1938218176,-0.2102639973,-0.1038752347,0.1957834363,0.0845955759,0.3147147894,-0.1896250248,0.1951976568,-0.1819043458,-0.613976717,-0.1501671076,0.0466882586,-0.2589325011,0.0336541049,0.1179802939,-0.3058903217,0.0450367853,0.1060032099,-0.0355217606,0.2437459081,0.1896084249,-0.3306345344,-0.14126423,-0.2343382835,-0.2384477556,0.0829156712,0.0062551172,0.2903526723,-0.2742957473,-0.132715106,0.0041045304,-0.0777190924,-0.5493536592,0.2553771734,0.1287322789,-0.2849131823,-0.0503059328,0.7471109033,-0.1838762015,-0.2807388008,0.0672428682,-0.0590743795,-0.0985703692,-0.0643855706,0.0351849049,0.3334688544,-0.5780578256,-0.1407892853,-0.2153735012,-0.3041471541,0.2633823752,0.2401827723,-0.0136863776,0.1278267652,-0.2747952342,-0.0291897468,0.0547558405,-0.0510421954,-0.0077471253,0.1915834546,-0.1714612395,0.2518334091,0.0054828706,0.0792416707,-0.0153486952,-0.0773086026,0.2999676168,0.3543708026,-0.2617487609,-0.0703961179,-0.0361999683,-0.4247270226,0.1664165109,0.1873419136,0.0743337423,-0.0027772691,0.1444615126,-0.2353902012,0.1300503165,-0.0637475029,0.2238672674,0.2670314312,0.1356241107,0.5249230266,0.0607625395,0.0489258282,-0.1597839594,-0.0156642832,0.1047909334,0.4020428658,0.3152895272,-0.1573581547,-0.0252302866,-0.1231054962,-0.2486605793,-0.3672913909,0.246550858,0.263835758,0.0199128538,0.3339195848,-0.3577463031,0.2866940498,-0.2280166745,-0.2566809654,-0.1881205589,0.2682740986,-0.0802951679,0.0237505175,-0.1958561987,0.3590033054,-0.1000219434,-0.0650889575,-0.1096429154,-0.1666441262,0.214011997,0.1449781209,0.5165333152,0.2998847663,0.2574204206,-0.2457240522,-0.1387692988,0.1868895292,0.0629615188,-0.2090667635,0.143025443,0.2379076034,0.1825202107,0.0511173606,0.0688579008,-0.0243051145,0.3960514069,0.2139659226,-0.0781380907,-0.1446684152,0.3003722131,-0.0670024827,0.2817319036,-0.0789468959,-0.3979156911,-0.0214979928,0.3684980571,0.0226097032,0.11486011,0.1395166814,-0.1527694315,-0.1007094011,0.2167847306,-0.4813978076,0.1276504844,0.2626873255,0.2217152566,-0.1654064357,0.2598191202,0.0346075445,0.0458805598,-0.2548020482,-0.1095755622,-0.0398895964,0.1586689502,0.0767411292,-0.3456186056,-0.4028255939,-0.002589761,0.0819898322,0.0495549031,-0.1989218146,0.0789047852,-0.054054264,0.2363436222,-0.3461315036,-0.2608318925,-0.0998265743,0.2485681772,-0.3377684653,-0.006428007,0.042944327,-0.1900702566,0.3239147365,-0.0356499217,-0.0031526275,-0.2456724793,0.0058891075,-0.4746785164,0.2933714092,-0.0538147613,-0.3159295321,0.3817481399,-0.0754599944,0.147675097,-0.3301287889,-0.5107155442,0.1697301269,-0.1299942583,0.5753853917,0.1254235804,0.0257649999,0.0134682329,-0.1258265674,0.0239615273,-0.0546529628,-0.1641530246,-0.1144845933,-0.0887319818,0.4223630428,0.0386973806,-0.4531418383,-0.4203661382,-0.1723112464,0.2529330254,-0.1639656723,-0.0030101773,0.229985714,-0.1378036588,0.1382640451,0.0099065416,0.1318792552,-0.264727354,0.0278456025,0.2004290968,-0.3106866777,-0.2774636447,0.0611869432,-0.0906982794,0.0852127746,-0.2441659868,-0.3625802994,0.2580148876,-0.0958719105,0.2738499343,-0.1915207952,0.0749033019,0.191463545,0.1696659178,-0.0674576834,0.110429585,0.0078716241,-0.1321851462,0.0105574392,0.086990878,-0.1092168987,-0.0853853971,-0.0116310194,0.294559449,-0.0115905013,0.0675024763,0.2806142867,0.0267516412,0.2315541655,-0.3706284463,-0.4429990649,0.3553462923,0.0400360227,0.048634205,0.4195979238,0.2135556042,-0.2142562121,-0.2098941952,0.0525777638,0.1803082675,-0.0713451877,0.0407626778,-0.1082012579,0.1665843278,-0.244934231,-0.2167077959,-0.1730590612,-0.1011517718,0.2547444105,0.3276857734,0.0936207995,-0.074071236,-0.2278883904,0.2381794006,-0.3003156483,-0.0443584062,0.0840816945,0.254611522,-0.2716081738,0.3353055716,0.0300292093,0.1542760581,0.2186018229,-0.6904952526,-0.4939383268,0.1358559877,0.3060667515,0.1556712985,0.1767362058,-0.1837191731,-0.2265481353,0.3084176779,0.2455106527,-0.1624627262,0.0297820698,0.1941706091,0.4270160794,-0.0171993431,-0.0469216779,0.0955202654,-0.1658641994,-0.0973811746,0.3946689367,-0.0293121394,0.017172765,0.0857527182,0.1370745301,-0.0335640423,0.2016941607,0.2797218263,-0.0593954735,-0.1606953591,0.152137652,0.0868340209,-0.0403815806,-0.4850472808,0.0786107481,0.1992167383,-0.0208807886,-0.525150001,0.0876165032,-0.2386714518,0.4695504308,0.1602394879,0.2618011534,0.3291909397,-0.277551949,-0.2181497812,0.0303880945,0.2831467986,0.4071322978,-0.0389448293,-0.4350191355,-0.7765517235,-0.0014824506,-0.0912955478,0.0166433547,0.4128485918,0.4393833578,-0.1327073127,0.2475264668,0.0767596737,0.9190419912,0.2312945724,0.041550491,0.0771926269,-0.0802166685,0.1929474175,-0.3251571953,-0.1349240839,0.0496716537,-0.2305433452,0.0059842821,0.2619853914,0.3471663296,0.2499647439,-0.2336064726,0.100499779,0.0962322205,-0.1959431171,0.2961099744,0.3082038462,0.1067743376,-0.4312557578,0.0338468254,0.230436042,0.0437933318,0.2038466036,-0.0725455731,-0.0566840954,0.10696394,0.1314010322,-0.4548589289,-0.0760947168,-0.5070952773,0.1264172941,-0.472950995,0.0894166157,0.20291394,-0.0858417004,0.4214228392,0.4717181921,0.0409157239,0.0229251124,0.4762382507,0.0628408566,-0.0877066329,-0.2355531454,0.2755264938,-0.0505395271,-0.2636356056,0.3230834603,-0.2016752213,-0.1362146288,-0.3825470209,-0.1163197458,-0.106591396,-0.0457282513,-0.1752589047,0.4103654027,0.1723272055,-0.195346266,0.1480227858,0.3497507572,-0.2355995625,-0.194161281,0.1073292047,-0.0974726528,-0.3136829436,-0.2266597599,0.2490622848,-0.6465171576,0.2945066988,0.4389957488,-0.1503630877,-0.2353438735,0.0864051804,0.4400274158,-0.470153451,-0.3849063814,0.0059765349,-0.1480595618,-0.0426977426,0.2869281173,0.0666101426,-0.2061115056,0.1507675499,-0.0945869237,-0.3692905307,-0.0289407764,0.1487245709,0.1253795624,0.2711099088,-0.2238437086,0.0357854515,0.0790957212,-0.3972319365,-0.1092496216,-0.1065866798,0.0030479888,0.2427649945,0.0556869991,0.0141382068,0.0340109281,0.1984766424,-0.3470435441,-0.0730968341,-0.1833737046,-0.1146325916,0.0745272562,0.1349312663,0.1228228211,-0.0779468492,-0.0350560844,-0.1030374169,-0.1456647366,0.1173659116,0.0765322894,-0.0797197148,0.2564330101,-0.0932966918,0.0115412697,-0.1612450033,0.0922035798,0.2833352983,0.0420557447,0.288731277,0.168070361,0.1850283444,-0.1306830645,0.0826932415,0.2942379713,-0.0047163013,0.3498417139,-0.1388337016,0.037949007,0.2871198356,0.3044242263,-0.0796230361,0.0239542536,-0.1038728431,-0.0387403965,0.3412595987,0.2805144787,-0.3252315223,0.163378492,0.1060327441,-0.148153916,0.1985507756,0.2169762105,0.329796046,0.2921764553,0.4217361808,-0.1599366814,-0.1356134862,0.0108738653,0.4031977654,0.324082762,-0.0374286547,-0.1339367926,0.0970411673,0.2642424107,0.0107276971,0.2740622759,-0.1399390101,0.3285216391,0.1947271228,0.21688281,-0.0120684197,-0.2332323492,0.445066154,0.3071076572,0.2817720771,0.2235705256,0.1095092744,-0.1401675493,-0.1730021089,-0.2234975845,-0.0463561602,0.0801670775,-0.2020115256,-0.0010115803,0.4020216167,-0.343310833,-0.1807059348,-0.2253137827,-0.0533671863,-0.0010703797,-0.2768157423,0.1128140986,0.1313049197,-0.3100359142,0.069966726,0.0297105927,0.0266484022,-0.0611058138,0.2780019343,0.0103994468,0.1770153642,0.0477205105,0.4197740555,-0.1397374421,0.0833046734,0.0168744605,0.3180228174,-0.2049370408,-0.018013224,0.0879957676,0.3278907537,0.0627762228,0.2459598929,0.3205837905,0.1203444973,-0.1926162392,-0.0150433984,0.1718025357,-0.1916517764,-0.4136287868,0.0126996385,-0.038408123,0.0860797763,-0.3797051311,-0.0310510825,-0.1285461485,-0.1348854154,0.3256211281,-0.2585844696,-0.1083818227,-0.0789923146,0.1384611726,0.1661755145,-0.2029033452,0.4855581522,0.2340565026,-0.1668652594,-0.2174425423,-0.438138783,0.0512649231,0.0801497474,0.1744616628,-0.1245624647,0.0625836924,0.383197099,0.1167944744,-0.0427718125,-0.0642128736,0.1631996781,0.3232921958,-0.3479463458,0.0629393682,0.285934031,0.4084410667,-0.0302078053,-0.018503271,-0.0732908919,0.4125907719,0.1505441368,-0.1334849894,-0.2462115139,0.130390659,0.1272662431,0.3557190597,0.2265091687,0.5275244713,-0.2215262502,-0.0379721448,-0.2685339153,-0.2232326567,-0.069254294,0.1755631864,0.028544765,0.0636214837,-0.1147708148,-0.3132456243,-0.24616009,0.2030116618,0.1759140193,0.2730841935,-0.5236433744,0.2106972337,0.0956702605,-0.3745821416,-0.182957679,0.145992741,0.0419237353,-0.0368626192,-0.09746667,-0.0534512438,0.043697387,-0.1767215431,-0.2392448485,-0.1110521331,0.1325585395,0.3138341606,0.1159163266,-0.3667922914,0.1731071323,0.2874090374,0.2281908095,0.2834448814,-0.0266075935,-0.0202200972,0.0379456617,-0.2808719873,0.3004857898,0.1155652925,0.0965197459,0.1939515322,-0.2216003686]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/223","title":"[Feature request] Add FLUE dataset ","comments":"https:\/\/github.com\/huggingface\/datasets\/pull\/943","body":"Hi,\r\n\r\nI think it would be interesting to add the FLUE dataset for francophones or anyone wishing to work on French.\r\n\r\nIn other requests, I read that you are already working on some datasets, and I was wondering if FLUE was planned.\r\n\r\nIf it is not the case, I can provide each of the cleaned FLUE datasets (in the form of a directly exploitable dataset rather than in the original xml formats which require additional processing, with the French part for cases where the dataset is based on a multilingual dataframe, etc.).","comment_length":1,"text":"[Feature request] Add FLUE dataset \nHi,\r\n\r\nI think it would be interesting to add the FLUE dataset for francophones or anyone wishing to work on French.\r\n\r\nIn other requests, I read that you are already working on some datasets, and I was wondering if FLUE was planned.\r\n\r\nIf it is not the case, I can provide each of the cleaned FLUE datasets (in the form of a directly exploitable dataset rather than in the original xml formats which require additional processing, with the French part for cases where the dataset is based on a multilingual dataframe, etc.).\nhttps:\/\/github.com\/huggingface\/datasets\/pull\/943","embeddings":[-0.1138885617,0.0201090612,-0.2017977983,0.0217149276,-0.0668713525,0.0821919218,0.3086242974,-0.1284856647,-0.2158084363,0.1436840296,-0.1844947934,-0.1214750931,-0.20977512,0.5184231997,0.1598236412,-0.4242121279,0.0963141695,0.0893124193,-0.047694061,-0.2990930676,-0.1068487093,0.3054482937,-0.0820136964,0.0682070032,-0.1604556739,0.1779842973,-0.139450267,-0.127796784,-0.2459370345,-0.0678748265,0.0994971693,0.2189712524,0.0391843468,0.2230267972,-0.0001015445,-0.2954333425,0.2343072295,-0.0634732395,-0.0264172051,0.0867693946,0.0696646199,-0.3254773617,-0.3584456444,0.1709452122,-0.2474468648,-0.1480086148,0.0245093312,-0.5102183223,-0.0450543165,0.3521412313,0.2359467596,0.0198289137,0.0081400629,-0.062642701,0.2780300677,0.4293572903,-0.3688441813,-0.0855347291,0.5897359848,0.0770706683,0.1600646228,0.3144943416,0.3135752976,-0.4009285271,-0.1344370395,-0.0327120312,-0.144745037,-0.2994815111,0.0897829905,0.3849660754,0.5242367387,-0.1251841336,-0.5157444477,-0.1727780551,0.1900622547,-0.2994328141,-0.0387694165,-0.1870084405,0.1547386795,0.1784988493,0.2350331694,-0.5080826283,-0.1128546596,0.0094866799,-0.013204244,0.1299175471,-0.107795909,-0.1607280672,0.0062949611,-0.109047927,-0.1912190914,0.2520527244,-0.2155878246,0.3287084401,-0.1594628394,-0.2782716751,0.0386597514,0.05867479,0.4727493525,0.1642737687,-0.6082844138,0.1006324515,-0.3255391121,-0.1096884608,0.2019578964,-0.0241068136,0.2688207626,-0.4416929185,0.314711839,-0.3632261753,0.2481133044,-0.0825778395,-0.046788279,0.0902497172,-0.3716180325,0.0927796587,0.0917398185,-0.3658213615,0.1365671754,-0.0040910607,0.3316003382,-0.2228597701,-0.0935935155,0.1955501437,0.1002018452,0.1811708212,-0.3678531051,0.1571492404,-0.1591082513,-0.6946117282,-0.1176549196,-0.0219255462,-0.3143663406,0.0707229674,0.0980724543,-0.3814487159,0.0500857309,0.1172858477,0.0995244086,0.3047376871,0.0924147815,-0.2358619273,-0.1666651964,-0.2237225175,-0.319883287,-0.0155412769,-0.0391905271,0.2665649951,-0.278601855,-0.170559451,0.0432315804,-0.0043939543,-0.6395406723,0.2074639797,0.1764385998,-0.2963139117,-0.2078758031,0.7851132154,-0.1884597987,-0.2765944302,0.139747709,0.0138394535,-0.2307183146,-0.0953442082,0.0395940728,0.2619876862,-0.3342693746,-0.1667086929,-0.0931585282,-0.4214458466,0.2501997352,0.270355016,0.024593886,-0.0205930211,-0.3487709165,0.0829154253,0.1158074662,-0.0393293574,-0.00177065,0.0567786992,-0.1647604853,0.1437230557,0.0571674518,0.088888444,0.1753341109,-0.1030140445,0.2356736362,0.5018596053,-0.2238969505,-0.0792217106,-0.0110896034,-0.2871869802,0.1027709916,0.1681192666,0.0008932425,0.0163263492,0.2335603237,-0.293659389,0.1548333019,-0.0898406282,0.1592120081,0.2514831424,0.3071253002,0.4548807442,0.0485869423,-0.079530865,-0.2037005574,-0.1477669328,-0.1813721508,0.3329716623,0.2487141043,-0.328715533,0.0106255868,-0.0968171582,-0.0984854773,-0.298507452,0.2378116995,0.2721960247,0.1287450045,0.4026982188,-0.3247589469,0.2588820755,-0.105080612,-0.0753099993,-0.2685526013,0.2486502528,0.1089590713,-0.0787349865,-0.0951423943,0.3274614811,-0.123279199,-0.104499802,-0.0512945987,-0.1483477056,0.2365709543,0.2250610292,0.4748602211,0.1681679785,0.3335720003,-0.1235011891,-0.1767534763,-0.0118875466,0.0554476976,-0.0917486101,0.0557496212,0.2535169125,0.1828804612,-0.1737778485,-0.0485761985,0.0037774206,0.3666087389,0.2000831962,-0.1340346187,-0.1392384022,0.3624212444,-0.014685221,0.1847584099,-0.0574189499,-0.5186154246,0.0417577587,0.5025678277,-0.079734765,0.0922047123,0.1683186293,-0.0969098657,-0.0899185613,0.3574519753,-0.4517807961,0.0792188048,0.2698593438,0.2755875289,-0.1178241521,0.2493128031,-0.006914895,0.060043551,-0.3654628098,-0.1314292252,-0.063933827,0.1789279133,-0.0100425882,-0.4999206066,-0.1357219666,-0.0660044923,-0.019306168,0.1797360778,-0.1826199889,-0.0155396862,-0.2161450684,0.1404782832,-0.2606761456,-0.3649349511,-0.1392542124,0.2645755112,-0.1837913245,0.0040845703,0.0960211977,-0.0807193443,0.4634726048,-0.1770851463,0.0419322625,-0.3494104445,0.0565438904,-0.3668341935,0.2498271912,-0.0197698101,-0.1921157986,0.5456587076,-0.1872946024,0.2446567863,-0.3836437464,-0.582433641,0.2473492026,-0.1733819544,0.4785887003,0.1164908037,-0.0571063198,0.0749577358,0.0540427566,0.0747212619,-0.03625172,-0.2532536983,-0.0984374657,-0.0272287037,0.3045600355,-0.0477262363,-0.1901037395,-0.3738543391,-0.2519306242,0.4367754459,-0.2620405853,-0.0068415077,0.1369989663,-0.3052266538,0.1199357137,-0.2720456421,0.108550474,-0.333094418,-0.0908817127,0.1823411882,-0.436514318,-0.3016332686,-0.0383085981,-0.0099141337,0.0396622792,-0.2274070084,-0.3121877909,0.0721885711,-0.1189221144,0.2920047045,0.0024692863,0.1431138664,0.1818999052,0.1361443102,-0.0710018799,0.0811748952,-0.0937239304,-0.1603352278,0.0872937441,0.0172097664,0.0162340496,-0.1010431796,-0.0889645666,0.2598108947,0.04108366,0.14817895,0.4497583508,-0.0096500199,0.31185022,-0.3455944061,-0.4183941483,0.354195863,0.0252964851,0.1211395711,0.3665655851,0.2435645312,-0.159531787,-0.1144566312,-0.1076586023,0.2420655191,-0.0800322369,0.0799738765,-0.0678856969,0.093979463,-0.2281181514,-0.2025313526,-0.1466539651,-0.0604755729,0.3128253222,0.229050532,0.2154658586,-0.0439726561,-0.3569872677,0.0527690314,-0.4021570385,0.0153603964,0.0385448225,0.1461385339,-0.3851723671,0.3062187433,-0.0355841033,0.1325179338,0.3316948414,-0.556666553,-0.5483426452,-0.0026717412,0.1944116056,0.2223478854,0.2488846332,-0.0820797905,-0.0723484904,0.1478296369,0.3898028135,-0.2635067403,-0.0788385123,0.187915951,0.4531806111,-0.0741610676,0.0303438287,-0.0015062279,-0.1804097593,-0.1315688938,0.3853216767,-0.0684295148,-0.0315233544,0.227323398,0.1094939932,-0.0993234813,0.170826301,0.2476092875,0.0119622815,-0.0694407225,0.1986816823,0.0554743633,-0.0549677946,-0.2771814167,0.1714301258,0.3107222319,-0.0788160264,-0.6045917869,-0.0115149347,-0.2754111886,0.5594878197,0.190200597,0.2662421167,0.3682156801,-0.1560466737,-0.1088450849,-0.2235715687,0.1432557404,0.3500610292,0.213393569,-0.4145666659,-0.6312149167,0.0128930043,-0.0370542295,0.0650539845,0.444190383,0.4911472499,-0.1928778291,0.0549465418,0.0050673038,0.9323785901,0.2757569551,0.1095260456,0.0619538911,-0.1451611072,0.3098255098,-0.2564950883,-0.0942549407,-0.0364287756,-0.3113161027,-0.0169270393,0.2150949836,0.342048496,0.0363298059,-0.2537838519,0.1365037262,0.0911909789,-0.1712873131,0.1252582371,0.4180552661,-0.0181979705,-0.427887857,-0.0698327646,0.1715503484,0.0865090191,0.3040876389,-0.1068753004,-0.1772003919,0.1820554584,0.0579735413,-0.425165683,-0.1144025326,-0.361117363,-0.0598768592,-0.2309279442,0.0428453982,0.1855424792,-0.1454734206,0.4714531004,0.4429657757,-0.0204768088,-0.0327533074,0.3209601641,0.0622753166,-0.2078038156,-0.1486346126,0.2585280836,-0.1635700613,-0.2343317717,0.3384930789,-0.0019233468,-0.0439422466,-0.3819249868,-0.1376593411,-0.2852243483,-0.0360927135,-0.1253512353,0.3186258972,0.2274105102,-0.2296957523,0.1557340175,0.4411768317,-0.1816216856,-0.267029494,0.2178283632,-0.0468684472,-0.3197109997,-0.2190785706,0.165653646,-0.6938459277,0.2086727172,0.3843746483,-0.143990308,-0.2560334504,0.1496073008,0.351588428,-0.4802548289,-0.3098517954,0.0943921283,-0.2160508037,-0.2217015475,0.2539371252,0.0395524837,-0.3114626408,0.1934990436,-0.0757634267,-0.4300019443,0.1246587411,0.1136735678,0.1515266597,0.3023982942,-0.2794867754,0.039018821,0.0724134892,-0.408305794,-0.0742858499,-0.0413006283,-0.0209893156,0.3970454335,0.0774220973,0.0191371255,0.0002548459,0.169748649,-0.1589069515,-0.1214215383,-0.1844426244,-0.1273283213,0.0767418444,-0.0552942231,0.1376731843,0.0698021576,0.0565697327,-0.0795584247,-0.135974586,0.1790607721,0.0368830822,-0.0162411463,0.3001196384,0.0097360257,0.0365102626,0.0046745366,0.1731867641,0.3654302359,0.0286323037,0.2229610533,0.1604402065,0.1030708849,-0.1204337776,0.1970288008,0.2758999169,0.0796337426,0.164736554,-0.1628880352,0.038301643,0.1753684878,0.2651027143,-0.0592033118,-0.0615138151,-0.1406819969,0.0347973593,0.5000426769,0.2612418234,-0.3664906919,0.0377216302,0.2203569263,-0.051340811,0.1799161583,0.1758148819,0.3225748539,0.4118155241,0.3453775644,-0.238725245,-0.2496572286,0.0587866493,0.3513243198,0.1305018365,-0.0074110548,-0.0644745752,0.0813556537,0.4229746759,-0.0377206467,0.2821335495,-0.1352779865,0.2429374009,0.1674921215,0.2121413946,0.0669382215,-0.1495037079,0.3899597526,0.2926806808,0.1853998303,0.3047437072,0.1496168673,0.1143572032,-0.0285368506,-0.2658038139,-0.1364142001,0.2838530242,-0.1411199272,-0.1634356678,0.3637388051,-0.2277292162,-0.1306022853,-0.1897552609,0.0764073953,0.0383771844,-0.0606686324,0.0729350969,0.2120756209,-0.090633139,0.0675886497,0.0108893756,0.0781975836,0.0327546932,0.322040379,-0.0251949616,0.1209761649,0.0985147953,0.4045496285,-0.0608734563,-0.0136999357,0.0431616791,0.3287947476,-0.3073832095,0.0310837608,-0.0105175311,0.374122858,0.0980771929,0.225185886,0.2535538077,0.1147700474,-0.1465254277,0.0441563092,0.0989866257,-0.2217126787,-0.3753718436,0.0266148075,-0.0008455048,0.1175660267,-0.4875076711,-0.0598709248,-0.2055523843,-0.0722841322,0.401430279,-0.0940814763,-0.1136536002,-0.05028449,0.1372278929,0.1794716567,-0.0829110369,0.4571363926,0.385191083,0.0299591087,-0.2125200033,-0.4206179082,0.0362563133,0.0628759637,0.1491710991,-0.0458460152,-0.124552384,0.4984131455,0.2462384999,-0.0369965918,-0.061601419,0.1734962463,0.1452328116,-0.3390876949,0.1326269954,0.2672691643,0.3302906454,-0.0464530475,0.0011738616,0.1024570093,0.2656129003,0.1133318469,-0.1993967295,0.0173171796,-0.0373975001,-0.0688948557,0.290317297,0.2832114995,0.3385130763,-0.2962146699,-0.1852938682,-0.1294609606,-0.2267227322,-0.1243290976,0.2084195316,0.0694427118,-0.0011051119,-0.1145558879,-0.4551560283,-0.2850477695,0.3791589439,0.1475800574,0.2791262865,-0.3683819771,0.2741360068,0.1220559329,-0.4718966782,-0.097893022,0.1509690434,0.0157136843,-0.0993356332,-0.0345332362,-0.0499410741,0.144858703,-0.3249726295,-0.2884715199,-0.1391701251,0.0579119213,0.35103935,-0.0124164941,-0.3861205876,0.2045717686,0.1623386592,0.1367448568,0.2764475644,0.1550269574,0.1053869277,-0.0379872397,-0.239685148,0.3537153006,0.1984390467,0.0421215184,0.1602570713,-0.2403471172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/222","title":"Colab Notebook breaks when downloading the squad dataset","comments":"The notebook forces version 0.1.0. If I use the latest, things work, I'll run the whole notebook and create a PR.\r\n\r\nBut in the meantime, this issue gets fixed by changing:\r\n`!pip install nlp==0.1.0`\r\nto\r\n`!pip install nlp`","body":"When I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n","comment_length":38,"text":"Colab Notebook breaks when downloading the squad dataset\nWhen I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n\nThe notebook forces version 0.1.0. If I use the latest, things work, I'll run the whole notebook and create a PR.\r\n\r\nBut in the meantime, this issue gets fixed by changing:\r\n`!pip install nlp==0.1.0`\r\nto\r\n`!pip install nlp`","embeddings":[-0.2037729025,0.1056738496,-0.0336354487,0.0640668422,-0.1007536948,-0.1691397429,0.1365311444,0.1091818288,-0.3450911641,0.1280623674,-0.164352268,0.5187450051,0.1809385419,0.0821388215,0.1669569165,0.0765133947,0.0644943118,0.4605704844,0.0468692891,0.156682536,-0.213108331,0.4624915421,-0.4079595804,0.1537030637,-0.2386357337,-0.1677411795,-0.206126675,0.2790915072,-0.5138830543,-0.2758500576,0.2979657948,0.1417234689,-0.0084389485,0.1811356694,-0.0001169652,-0.2367392331,0.1207963079,-0.0518519096,-0.3485308588,-0.3542282879,-0.0085788863,-0.4436558187,0.1595697105,-0.3419606388,0.1409291476,0.3921480477,0.3511864245,0.3706688583,0.3492597342,0.3874811232,0.1548643261,0.3137634397,0.2158098072,-0.0544759072,0.1636968553,-0.2806620598,-0.2976787686,0.3243994713,0.5822194219,-0.1685713977,0.003550905,0.1059031114,-0.0955071971,0.0559050888,0.1220199764,0.1638884097,-0.2535198331,-0.4291177094,0.2994447947,0.1212930828,0.0607747249,-0.2001691163,-0.0181519743,0.0014585996,0.2585046589,-0.3444949687,0.2189854831,0.3701309264,-0.3430844843,0.0796014294,0.0347072147,-0.0327624939,-0.217060253,0.2308544666,0.0566510111,0.3647040427,-0.0647751614,0.0102592483,0.2563159466,0.0780712813,0.2204387486,0.2324579954,-0.1993964911,-0.0125770569,-0.020614529,-0.1966523677,-0.1525165141,0.3759249151,0.235389784,-0.3535800874,0.1968852282,-0.0626712218,0.1999604702,0.1969324648,0.0232112259,-0.0120320274,0.3583292365,-0.0416987538,0.3990332186,0.4018417299,-0.0248469114,0.0016650113,0.0479009189,-0.1774966717,-0.2790907621,-0.1293845326,0.0646892264,-0.1897190213,-0.2175166458,-0.016830435,-0.1500985622,0.278216511,0.0331732593,0.194628939,-0.0681748316,-0.2387417108,0.1087198034,-0.0225516856,-0.3674689829,-0.1743022501,0.0099496134,-0.0019469556,-0.187640354,0.1096844599,0.2619293332,0.1836888492,0.4778348207,-0.1534589827,0.0996062905,-0.1178124025,0.052396033,-0.2653231621,-0.101341635,0.314409405,0.3155173361,-0.2200689912,-0.2717237175,-0.5757359862,0.0451195054,0.1975996643,-0.0320335366,-0.0128898425,-0.1250274926,0.1084180698,-0.2844673693,-0.070265539,-0.3243390918,0.0549854003,-0.2155044228,-0.2589317262,0.048960574,-0.1539312303,-0.2456499338,-0.0019286483,-0.0584790818,0.1869665235,-0.0323922038,-0.0988170281,0.1605466604,0.0084136315,-0.0025028242,0.2557320893,-0.1460038722,-0.0871363506,0.0606881231,0.1924744993,0.3825965226,-0.3142406344,-0.7157806754,0.1233162656,-0.0762078762,-0.2864676714,-0.09852314,0.1813633144,0.4203037322,-0.1855093539,0.0700470582,0.4979850352,-0.1739978939,0.1083303615,-0.2697708309,-0.1738209724,-0.128898561,0.0501525812,0.1142117381,-0.0171914566,-0.0391070768,0.7044432759,0.1668141633,-0.0989093333,-0.2014872432,0.0599216856,0.3261883259,-0.3010640442,-0.0882215127,-0.1463488638,-0.506205976,0.0100470018,-0.3460903466,0.2132534832,0.0968355238,-0.2111298591,-0.0591825619,-0.0682137907,-0.012183357,-0.1781014204,0.1000677645,0.0204873681,0.048870232,-0.0486058295,0.0901986808,0.3979105353,-0.1955251098,0.4248863161,-0.2051782757,0.1899199635,-0.3483103216,-0.3624228835,0.0692607537,0.3132476211,0.0034313435,-0.170596689,-0.1000657305,0.257260561,0.0810529739,-0.0746914893,0.1012284756,-0.3666822016,0.3960621059,0.0606683865,-0.0403566696,-0.0608279929,0.0121467514,-0.0353698842,-0.0288012698,0.2245773971,-0.1860223264,-0.0414232239,0.1730560809,0.2094190866,0.0392654352,-0.1198861673,-0.2044531107,0.1320448965,0.323166132,0.236813575,0.0411805324,0.0109152021,-0.3629996777,0.2105794847,0.4938595295,0.027466055,0.24809964,0.0198727529,0.0485267714,0.002644903,0.3425408006,0.0880601555,0.3751697838,0.1141346842,-0.110131368,-0.0038270995,-0.1853574663,-0.3078569472,0.1314207166,0.0830455497,0.1299071312,-0.169739753,0.303027153,0.0148459617,-0.3646787107,-0.1110709012,-0.0707152486,0.235033825,0.1090317667,0.3059636056,-0.1434033215,-0.6106244922,-0.2032148987,-0.1151826084,-0.186147213,-0.1863025725,-0.0547748841,0.3990697563,0.3937791884,0.1321959198,0.3591193855,0.2093052119,-0.1175979078,0.1829980314,0.1719687581,-0.0438949093,-0.429446429,0.0313699432,0.3862446845,0.297998786,0.1085413694,-0.2365486473,-0.2492028624,0.0154726664,-0.4298043847,0.2603361309,-0.3362569809,0.279479444,0.0997630954,0.247821793,-0.5951432586,-0.1031071842,0.1649894267,-0.1144916117,-0.122908406,-0.0929281935,-0.0368827581,-0.4055460393,-0.2671480477,-0.2051927149,-0.4359725416,-0.2693944275,0.3358285129,0.0157515723,0.2115788162,0.3739767075,0.0929482728,0.3100524545,0.0495364405,-0.2053632885,0.1208358854,0.0357938856,0.0885275751,-0.2505455911,-0.498541981,0.0997747332,0.0517346002,0.0376448929,-0.2449922562,-0.4161091149,-0.5225148201,0.1261278987,0.1882442236,-0.001207964,-0.0788631067,0.2069427967,-0.1973300278,0.1095579416,-0.0130770393,-0.2461119145,-0.1278255284,0.0830923319,0.3247517049,-0.1507366598,0.1058794037,0.1483967453,0.1934161335,0.3145639598,-0.0725045502,0.1143689454,0.1201350763,0.2739352882,-0.2105091661,-0.3405758739,0.3424800336,0.0101347398,-0.0096188923,0.3454568386,-0.0597283803,-0.5129151344,-0.1813617647,0.1022461578,-0.2586407065,-0.3566066027,0.1080049947,0.4365469515,-0.0429634638,0.2192899883,0.1769399941,-0.0064321249,-0.2671233714,-0.3938916624,0.2829703093,-0.2330520898,0.1004543379,0.1866302937,-0.4540288448,-0.6995074153,0.1291138381,0.0201124884,0.2973970175,-0.0087549817,0.2319483608,0.1322501153,-0.0280305818,0.2421707809,0.0865303725,-0.1293123215,0.0555368885,0.0240424387,-0.2520784736,0.1438996643,-0.3590139449,0.5358413458,0.8802663088,0.3770962059,-0.1643478721,0.0577812567,0.5292019248,-0.0327910446,0.1895888746,0.108040154,-0.3044334054,-0.3213104308,-0.1440336108,0.0569630191,-0.0482934862,-0.0072514927,0.0971350819,-0.1026031747,0.2945193052,-0.1304945946,-0.1803795695,0.4900846481,-0.0431484058,0.0689838976,0.034631677,-0.2005060911,-0.1485117823,0.2358810902,0.278200984,-0.2147084922,-0.2510846257,0.1910683215,-0.2993556261,0.1263995469,0.5574210286,-0.0127046667,-0.0876765847,-0.1413716525,-0.0103574181,0.3766844273,0.0082043977,0.3475843072,0.1379762888,-0.0829979107,0.1521376371,0.2743895352,-0.0102352155,0.0606488623,-0.1266966909,-0.1330262721,-0.2665355504,0.431796968,0.0313236266,0.8313063979,-0.1980553269,0.1825906783,0.216595903,0.0345823094,0.7005783916,0.0981837511,0.2973754704,-0.1605034471,-0.2612912953,-0.0269173253,0.0341729037,0.2771368027,0.0035691557,-0.6976842284,0.3543384671,0.5069897175,-0.0666596591,0.1408876181,0.2046708763,-0.2041286826,-0.3309403956,-0.6929699779,0.149915278,-0.2472083867,0.212081939,-0.231827423,-0.0166068338,-0.2207427472,-0.0558516942,-0.3980884254,0.0978431106,-0.4928942323,-0.1491035074,0.3097957075,-0.2716408372,-0.0152843101,0.3180866838,0.2980517149,0.2150382251,-0.041175995,0.055800803,-0.0094887856,-0.4288060963,0.0696155801,0.1207265034,0.0707869902,-0.2315786183,-0.0116988411,-0.1561685503,0.191993773,-0.147300899,0.3778708279,0.2653465569,0.1137990579,0.1531701088,-0.1188110039,-0.0428827517,-0.1843663305,-0.2605476081,0.1095840111,-0.1028380841,-0.1080933362,0.1768663526,0.1704010963,-0.1674775779,-0.0181277543,0.2444477081,-0.109146066,0.0531569086,0.2179025561,0.2029288262,-0.048076015,-0.2186718434,0.1641601622,-0.4710913599,-0.2075667381,0.019297827,0.0043055471,-0.1733898818,-0.0564207621,0.6419909,0.2973118722,0.0328934751,-0.1205995679,-0.3671325147,-0.2201470733,0.2041160017,0.2204357237,-0.1217902303,0.3346557319,0.3340523243,0.2858063281,0.096517086,-0.2278681546,0.2888351977,-0.3742848337,0.1768328398,0.0350636989,-0.2467701733,0.1611362547,-0.1358114034,-0.0565140508,0.5478605628,0.0633948222,-0.2198669761,-0.1331758052,0.068712689,0.1189699396,-0.1074636579,-0.3171965778,-0.0681325421,-0.0985211879,-0.0403521843,0.0776518285,-0.0682112277,0.1093354002,-0.0690529421,0.2667797506,0.2946943045,0.2482579798,0.0718109906,-0.0543679036,0.2110072821,0.1292755455,0.2501770258,0.2140119374,0.0341803096,0.1410023868,-0.0003102113,-0.0785295367,0.1575947702,0.2667445838,-0.4023347199,0.108538948,-0.0636997446,-0.0467179939,0.3891509473,0.0595789999,-0.1627654582,0.2553032637,0.073186487,-0.1452823728,0.0629553124,0.538184464,0.1647594571,0.0742015541,0.4867260754,0.1154769883,-0.0481447577,0.1442152262,0.1059578955,0.2352437824,0.2533346713,0.0116077438,-0.0835580006,0.0138906483,0.1132432669,0.088474147,0.0577534437,0.194741711,0.0087063201,-0.002556727,0.1844158769,-0.1320847869,0.099448055,0.2976964116,-0.0365743674,-0.2163549066,0.0441060364,0.2792088687,0.0621829852,0.1355578601,0.839427948,-0.5352561474,-0.256313622,-0.1129696742,0.1396200657,-0.3616019487,-0.0069119781,-0.3097963333,0.1597084999,-0.1319256574,-0.0187680088,-0.1546417773,-0.046836365,0.3026385009,0.2967253625,-0.1764535308,-0.1194128767,-0.2104070932,0.0963282585,-0.1580306143,-0.0432194546,0.5650874376,-0.4615538716,0.1558709592,-0.1157205775,0.5128487349,0.1959529817,0.2344564199,-0.1721441299,-0.1699351519,-0.0400432684,0.0216363799,-0.0968862772,-0.0191170443,0.4978402853,0.2487420589,0.1886569858,0.0300807264,-0.1156955212,0.1956358999,0.1585228741,-0.2628220916,0.0586590953,-0.1024087444,-0.1074168608,-0.1466408819,-0.2744992375,-0.043056123,-0.3473244607,0.5429142714,-0.2994098067,-0.1069813296,0.0449773185,-0.0946458504,0.0484454408,-0.0046039196,0.5350646377,0.4032155871,0.0254755635,-0.2152014226,-0.4277662337,-0.3889572322,0.2161229253,-0.2563636899,0.011096416,0.0596353374,0.1823772043,-0.1012397036,0.4139778912,-0.0789307728,0.0418472998,-0.1264992058,-0.0653211251,-0.4724362195,-0.1160057634,0.2852906883,-0.2760683894,-0.0696646199,-0.223663792,0.1521486938,-0.2058594078,-0.0717563927,-0.0071329228,-0.3104135692,-0.2900138497,-0.1044819802,0.07206247,0.1701776832,0.1010769829,0.0196082704,-0.321094811,-0.0465757288,-0.2240406722,-0.271959275,0.0755751878,0.2256835401,0.2975488901,-0.0523131713,0.3451108932,-0.1659618318,0.502035141,0.1008222923,0.03670194,-0.0122194579,0.0015544911,-0.1229908392,0.0852454007,-0.1476600021,0.4588993788,-0.2785682082,0.1753114909,-0.4899502397,-0.3166498542,0.3832266927,-0.4803468883,-0.0998111665,0.0017301765,0.2095846981,0.1394143403,-0.3002653122,-0.397646606,-0.02579505,-0.1576420963,-0.1750727296,-0.4597726166,0.1397780031,-0.2021505237,0.0512658544,-0.0826171711,0.4182423949,0.0798269659,-0.2558418512,0.0152477333,-0.1556756198]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/222","title":"Colab Notebook breaks when downloading the squad dataset","comments":"It still breaks very near the end\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83312264-aa96a600-a1df-11ea-987f-2f4a0474247e.png)\r\n","body":"When I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n","comment_length":8,"text":"Colab Notebook breaks when downloading the squad dataset\nWhen I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n\nIt still breaks very near the end\r\n\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83312264-aa96a600-a1df-11ea-987f-2f4a0474247e.png)\r\n","embeddings":[-0.2453231215,0.1265578568,-0.0555746034,0.1304289848,0.0279392004,-0.0962962061,0.3019449711,0.060063079,-0.3697974384,0.2233843207,-0.3876703978,0.3216298521,0.3497347534,0.1432359517,0.0889978632,0.1818201095,0.0537408926,0.3886002898,0.012332092,0.1648053974,-0.2017745823,0.4017664194,-0.3926337063,0.0158984289,-0.2998062074,-0.2220029235,-0.1767001003,0.2989171743,-0.5018666387,-0.165115118,0.2830982208,0.2360524982,-0.0754956231,0.2501781881,-0.0001191332,-0.2283937633,-0.0112816961,-0.1019016281,-0.3363925219,-0.1540449262,-0.0779848918,-0.1849163026,0.2255862802,-0.3756937683,0.1171497256,0.5302575827,0.2409542799,0.2563079596,0.2924793363,0.3243489861,0.1345028579,0.3440576196,0.1688451469,-0.0868403316,0.2246469408,-0.3074944019,-0.2844084203,0.448713541,0.6297420263,-0.1123408005,-0.0861666948,0.1253518164,-0.1093073785,0.2034886032,0.0487844497,0.2166916877,-0.3280586302,-0.3790917397,0.3763224185,0.0941118002,0.1185160205,-0.1446860284,0.0698929578,0.0947691873,0.1905350238,-0.3466815054,0.1868062317,0.3183270693,-0.2513887882,0.1379859298,0.0149199953,0.0461157374,-0.2383907437,0.1744912863,-0.117530629,0.2330750674,-0.138239786,-0.0350207984,0.2172722369,0.1106862351,0.2375907451,0.1811290532,-0.1785479784,-0.1562733948,-0.1141214296,-0.1901065707,-0.242567718,0.3013167977,0.2904114425,-0.2642318606,0.2496067584,0.0468524396,0.2806406915,0.1392719299,0.0114067756,-0.1382704973,0.2972939312,-0.1053575128,0.4562666416,0.4486875832,-0.218829155,-0.11720898,-0.0229074024,-0.1778700203,-0.2186070085,-0.1036233008,0.09196686,-0.2420670837,-0.3222458661,0.0422538668,-0.1715873778,0.2961679101,0.0821555331,0.2942343056,0.0112507679,-0.4121108651,0.0061260038,-0.0693361238,-0.2935739756,-0.2261812389,0.0069349902,-0.102093704,-0.1511829495,0.2321178615,0.2353571057,0.1157502234,0.4534631968,-0.1892378479,0.2598941922,-0.1736151278,0.0472240373,-0.409904182,-0.0757593364,0.3351511955,0.3496727943,-0.1784581095,-0.2145988941,-0.575247705,0.0800927579,0.0586226955,-0.021283228,-0.0544846468,-0.1074840799,0.1092832237,-0.2410878688,0.050709486,-0.3382655978,-0.0397421569,-0.260494113,-0.1580533832,0.026609201,-0.0161283016,-0.2404824048,0.0141883427,-0.0000181466,0.2760188282,-0.1189388707,0.0308604594,0.2333335578,0.0129814781,-0.0330094472,0.2950106859,-0.1414042711,-0.0187999364,0.0020186005,0.1664506346,0.2828995287,-0.2026683241,-0.7979896665,0.0832012743,-0.0566353984,-0.282338053,-0.0937410071,0.201145485,0.5667233467,-0.2144225091,0.1327502131,0.4811082184,-0.1894403845,0.0902004614,-0.137004599,-0.1125776246,-0.1508377641,0.0957249254,0.1230854839,0.0062259571,0.0822793767,0.6854791641,0.2129116654,-0.0908508524,-0.3390220404,0.0640895963,0.3202149272,-0.3624691665,-0.1078615263,-0.0986004844,-0.541628778,0.0213609077,-0.1869624406,0.0795968622,0.0530606806,-0.2351961136,-0.1345613599,-0.0401674323,-0.1213025749,-0.0622985177,0.0866819024,-0.0230373126,-0.0562794916,0.0134326909,0.1565249115,0.3330442309,-0.3413953483,0.4667951465,-0.0588091649,0.2859535813,-0.2668130994,-0.3211674392,0.0870695338,0.1329878718,0.0275296941,-0.241979152,-0.0967459008,0.2850730121,0.1996399909,-0.0130983358,0.1167884246,-0.2627469003,0.4218431413,0.1034982428,-0.0782400668,-0.1054529548,0.0841142163,-0.1800044477,-0.0533400998,0.2148131281,-0.2216628194,-0.0132613452,0.1533936113,0.1641588509,0.015306009,-0.1271155179,-0.1604656577,0.0889107659,0.2193324119,0.239746809,-0.0274503827,-0.0221298449,-0.3079191744,0.227667138,0.5567532778,-0.0424381085,0.2276684344,-0.0804107711,-0.0315105617,0.0279734246,0.4267624617,0.0068932972,0.364069134,0.0928153247,-0.031600751,-0.0190654695,-0.0149856675,-0.3365435302,0.1044545844,0.1223558113,0.2370118648,-0.1881807446,0.1898399442,-0.0552033149,-0.3603596687,-0.1638644785,-0.0811956078,0.2189472914,0.117426075,0.3775480986,-0.1030246317,-0.4632872939,-0.0875489414,-0.2462123036,-0.0253116563,-0.2575010061,-0.11910256,0.5656509995,0.4043952525,-0.0111908587,0.3115215898,0.2589112222,0.1273617148,0.2172053605,0.2074344754,-0.0615432002,-0.2876619697,0.0592360497,0.43509835,0.2604832649,0.1314091086,-0.3033001423,-0.1886035353,-0.070758298,-0.2939302921,0.2809461653,-0.1921732873,0.2130649388,0.0172271878,0.3088534772,-0.5568756461,-0.0753799826,0.1010542661,-0.036046423,-0.081667833,-0.024008574,0.0006763902,-0.4188296795,-0.1600800753,-0.1385560185,-0.3366317749,-0.2498149127,0.2148475945,-0.1401640624,0.1636652946,0.2667544782,0.109254241,0.3947454095,0.028874306,-0.2368699014,0.0512005985,-0.1127891466,0.1850448102,-0.2830388546,-0.4966156483,0.1444291174,0.2005981505,-0.0313918963,-0.2790839076,-0.4428981841,-0.362439394,0.0972916409,0.1206676513,0.0090648504,-0.169683814,0.0511740521,-0.3109427989,0.120229885,0.0200014636,-0.2105628103,-0.1684599668,0.0481169559,0.3485707343,-0.2287339121,0.1866552085,0.0794697478,0.187930733,0.2457389981,-0.1361026764,0.1350384951,0.0697257295,0.2220817208,-0.3352567852,-0.4632652104,0.3551285565,-0.18098782,-0.0516771078,0.3466368914,-0.085087955,-0.4633395374,-0.1754775792,0.0800909102,-0.2351770252,-0.4409416318,0.0885253102,0.4809088409,-0.1918515712,0.1235429868,0.2136677206,0.0844107196,-0.2375703156,-0.3362895548,0.3413932025,-0.3234403431,0.0106020058,0.2350167483,-0.3993977904,-0.699509263,0.1593644917,-0.0436109565,0.1805248559,-0.0709799379,0.3067024052,0.1086820513,0.0543065667,0.2739142776,0.2026852667,-0.0724871755,-0.0258288272,-0.083966352,-0.1446738392,0.1033396125,-0.4084560871,0.5936658978,0.8362895846,0.3447614908,-0.0858405456,0.1498965025,0.5353718996,-0.095133476,0.1576646864,0.1470438838,-0.2406691015,-0.3096102774,-0.1644270122,0.0611357614,-0.0076310718,-0.0195758585,0.0935922116,-0.2360296547,0.2996901274,-0.1664403826,-0.048628144,0.5512909293,-0.0872458518,0.1496978253,0.039768897,-0.1267764568,-0.0907599926,0.3838977814,0.2856781483,-0.1334917098,-0.2468074262,0.3276807964,-0.4627605081,0.1122647896,0.5596466064,-0.0332666375,-0.0585605986,-0.2044997364,-0.0123231681,0.3476112485,0.098767139,0.3777164221,0.2349455655,-0.1303193867,0.1893964261,0.2808528244,-0.1563521028,0.0756530017,-0.0417462662,-0.0589980148,-0.2524244785,0.4154141843,0.057306774,0.7484630942,-0.184226051,0.095510684,0.1445388496,0.0805775747,0.5118159652,-0.0250699632,0.2520190179,-0.0926053002,-0.4317754507,-0.0519282185,-0.0414654538,0.2385220379,-0.0329219662,-0.6439391971,0.1869892627,0.2817979157,0.1590924114,0.0790408179,0.1211267188,-0.1674723774,-0.2149860263,-0.6116220355,0.1563452184,-0.2347584367,0.184977904,-0.259016186,0.0425268672,-0.2347459793,0.0682056621,-0.4274705052,0.0468204021,-0.5017384887,-0.0231104698,0.0967881233,-0.2575035393,0.0338137262,0.2469957918,0.3325936794,0.2208962739,-0.0564555749,-0.0197764561,-0.0631271675,-0.3819863498,-0.0259927809,0.000365873,0.0099488907,-0.2179792225,-0.0704587251,-0.0512257889,0.327614814,-0.2398294061,0.4378131628,0.2369854301,0.1733787656,-0.0800396204,-0.0842868835,-0.1140626371,-0.2060979754,-0.2984067202,0.0887568668,-0.132003054,-0.0957367793,0.2461615205,0.181780979,-0.135325253,-0.0282633398,0.2736862898,-0.2333385944,0.1273755282,0.2449137866,0.2772300839,-0.1004786193,-0.2261549681,0.3546503484,-0.4100020826,-0.1566384435,0.0233212244,-0.1344187558,-0.0871138871,-0.0896110237,0.6583247781,0.3499514461,-0.0065295785,-0.1185708866,-0.4120721817,-0.2178361863,0.1046433821,0.2497419417,0.008250908,0.2459557354,0.2641913891,0.2142043412,0.0833953246,-0.2146106213,0.3500331342,-0.374658376,0.0755376294,-0.0074990871,-0.2500043511,0.2678196728,-0.175318405,-0.0513927154,0.5655355453,0.1172767133,-0.2094032019,-0.0756600127,0.0777587891,-0.0310178678,-0.1184916273,-0.2619901001,-0.0154083231,-0.074468568,-0.0656998008,0.1800580472,-0.0593145676,0.3062013388,-0.1910471171,0.366353184,0.2466706485,0.1254995167,0.0064900964,0.0190982297,0.2675904036,0.1629308164,0.1737812459,0.1013245657,-0.0099710971,0.2221306413,0.037849471,-0.0623672418,-0.0433543213,0.3860449493,-0.3838050067,0.1322408468,-0.0661099032,-0.1274482012,0.3931994736,0.0290647205,-0.1975455135,0.078133449,0.0806993395,-0.0399006382,0.0190204456,0.5886649489,0.1773424298,0.1154906303,0.5630693436,0.0066812406,-0.0213514231,0.1922916174,0.049026195,0.3004421592,0.5331115127,0.0164422654,-0.0981482565,0.0061964812,0.2058698535,0.1393382698,0.0792384148,0.1991034001,-0.0253571365,-0.0270041563,0.1573637426,-0.2493803501,0.1094206199,0.3698721528,-0.0629122555,-0.2240565419,0.0601731762,0.0997510999,0.0466873199,0.0939840898,0.7557439208,-0.5047186613,-0.2255714685,-0.1573780328,0.1021829247,-0.3851438165,0.0681544244,-0.3081605732,0.1300881207,-0.1052721813,0.0375627019,-0.1344478577,-0.0858308002,0.3644750118,0.2311086059,-0.1592925191,0.0620199442,-0.1140034199,0.0496043898,-0.1363233775,-0.0335229449,0.4314320683,-0.3111242056,0.2663336396,-0.057091821,0.520616591,0.1208990142,0.1597694606,-0.1548982412,-0.1944621652,0.0743442103,-0.0978165492,-0.0274728611,-0.0823092535,0.4119105935,0.0931080803,0.2672014236,0.045587115,-0.1056642681,0.2473553121,0.1823610514,-0.1945943236,0.0805978626,-0.3283516467,-0.0841693208,-0.1258638203,-0.3422870636,-0.1345846355,-0.3091433644,0.4372245371,-0.3044753969,-0.2261115909,0.0734470859,0.014386395,0.0463111736,0.0259448476,0.5657601953,0.347810477,0.0687318072,-0.3280747533,-0.5009368658,-0.2743065655,0.2324104458,-0.1336620003,0.0627062917,0.1136501655,0.2207620144,-0.0087640276,0.5008965135,0.0758293197,-0.0522707626,-0.216488421,-0.0691635087,-0.4581098557,-0.1733273417,0.2111210674,-0.2787963748,-0.1786858141,-0.196995914,0.3493988216,-0.194004029,-0.0229812004,-0.041431617,-0.2876906693,-0.3614017665,-0.235806182,0.0972526893,0.0884520933,0.0678717792,0.0050975382,-0.2949171066,0.1019273475,-0.2035692334,-0.1339605302,-0.0024364754,0.2962260842,0.2668511569,0.0414958745,0.2763688266,-0.1197731122,0.520411551,0.1509914547,-0.0242840033,-0.0954514369,-0.0916449428,-0.1316718161,0.0642815679,-0.1062249169,0.438719213,-0.3403731287,0.2207971066,-0.4136596024,-0.2704654038,0.3816092312,-0.3998782039,-0.1090688854,0.0644962564,0.2203929126,0.1161998659,-0.1676816195,-0.347753495,0.0486111976,-0.1597260684,-0.2790006995,-0.511741221,0.2768796086,-0.0829671696,0.1372447163,-0.0990942717,0.3966557384,0.1149107069,-0.2394762635,-0.1801224202,-0.2176103145]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/222","title":"Colab Notebook breaks when downloading the squad dataset","comments":"When you install `nlp` for the first time on a Colab runtime, it updates the `pyarrow` library that was already on colab. This update shows this message on colab:\r\n```\r\nWARNING: The following packages were previously imported in this runtime:\r\n [pyarrow]\r\nYou must restart the runtime in order to use newly installed versions.\r\n```\r\nYou just have to restart the runtime and it should be fine.\r\nIf you don't restart, then it breaks like in your first message ","body":"When I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n","comment_length":78,"text":"Colab Notebook breaks when downloading the squad dataset\nWhen I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n\nWhen you install `nlp` for the first time on a Colab runtime, it updates the `pyarrow` library that was already on colab. This update shows this message on colab:\r\n```\r\nWARNING: The following packages were previously imported in this runtime:\r\n [pyarrow]\r\nYou must restart the runtime in order to use newly installed versions.\r\n```\r\nYou just have to restart the runtime and it should be fine.\r\nIf you don't restart, then it breaks like in your first message ","embeddings":[-0.1622934341,0.2886501253,-0.0163853634,0.0890750438,-0.0951395929,-0.1352742016,0.2415599525,0.0423933864,-0.444473207,0.0557704158,-0.342538625,0.5006802082,0.1902989149,0.0517215766,0.2317513376,0.1294944435,0.0527208261,0.4562418163,-0.0051696734,0.151572302,-0.137189582,0.4322423935,-0.3422395289,0.1929836124,-0.2837288082,-0.206650421,-0.338755697,0.3366293311,-0.3838270307,-0.3679392934,0.3007687032,0.1165230796,-0.0106510911,0.2421114743,-0.0001158728,-0.2633949518,0.0717012957,0.0509735085,-0.3235515356,-0.1854397655,0.1579290479,-0.2767647505,0.3587703109,-0.3696824014,0.2319737971,0.2630354762,0.1968207508,0.4131796956,0.3784220219,0.2550897896,0.1543263346,0.362324059,0.1514990628,-0.0499028414,0.3280057609,-0.240154177,-0.2422332317,0.5541874766,0.5191043019,-0.1760158986,-0.0643188581,-0.02365526,-0.1500819623,0.1081864014,0.09065561,0.2510838211,-0.3042800725,-0.3595671356,0.1477923691,0.0752122998,0.0100565674,-0.2550982535,0.0877754316,-0.0887733027,0.254827708,-0.4677815735,0.2201659232,0.2952525914,-0.3333260715,0.0473990291,0.0693074241,-0.093594119,-0.1557637602,0.1831721514,0.1144590974,0.3580800295,-0.0566371605,0.0117057459,0.2579820156,0.1104667336,0.2554870546,0.2613551319,-0.1273595095,-0.1209307611,-0.0481666811,-0.1080249175,-0.1122742146,0.353577584,0.1700024158,-0.2848011851,0.1891995668,-0.0828458592,0.1804108322,0.1894937456,-0.0258666314,0.0118322466,0.1780892611,-0.2273446918,0.4025895298,0.2651009262,-0.1325009912,-0.0196064375,0.0696858689,-0.0828318894,-0.1812394857,-0.2002736777,0.1114613116,-0.1733044386,-0.2119898051,-0.0605254471,-0.2101204246,0.1031340957,0.030671658,0.226370424,-0.0486257151,-0.3248869181,-0.0088088252,0.0097236205,-0.3067674041,-0.1658763289,-0.0207025446,-0.0158130545,-0.3158730865,0.2927928269,0.256110996,0.0141710509,0.4369165301,-0.0667461082,0.1125650778,-0.0615690388,-0.0365417302,-0.250507921,-0.1376703531,0.1945447773,0.2850417495,-0.1370413601,-0.274626106,-0.5182225108,-0.0294751823,0.1341778487,-0.0913712829,-0.1003817618,-0.1418798119,0.1134282127,-0.2824954987,-0.0845008194,-0.2444850653,-0.0112388711,-0.2775984704,-0.3270464838,0.0648815259,-0.1133406609,-0.1195804477,-0.1733658165,-0.0513908975,0.2277743369,-0.1550365239,-0.060792502,0.1266469508,0.0868893862,0.0564630665,0.3691760302,-0.2238779217,0.0588181689,-0.0302966218,0.2738117576,0.4754826725,-0.1313362718,-0.5511142612,0.0222484414,-0.0927393138,-0.2200119048,-0.1776846945,0.185340777,0.4359275103,-0.1713544577,0.1205334887,0.4776637554,-0.1279907227,0.089275308,-0.2525898218,-0.2019875348,0.0210615434,0.0683600083,0.1030825973,-0.0479242206,0.0108466875,0.6572160721,0.1523803025,-0.1279436052,-0.2924892008,0.0120234145,0.2637312114,-0.1593085527,-0.0861503333,-0.1020768061,-0.49761796,-0.0189151708,-0.3826644719,0.2217239439,0.1123476997,-0.1261179149,-0.0196895897,-0.029196512,0.0130854743,-0.1026265398,0.1155181155,0.0912156776,0.0312762894,0.0355225652,0.0110383537,0.3079181612,-0.2108149976,0.4070410132,-0.1642879844,0.33948493,-0.3726189435,-0.4543426633,0.0388311446,0.3175660968,-0.0054071872,-0.185867995,-0.0847697854,0.3038392365,-0.0331054032,-0.0015203939,-0.0419727787,-0.216797784,0.4175575674,-0.0338797942,0.0402369164,-0.1533365697,0.0110290274,-0.1177541688,0.1212889031,0.1527857184,-0.2726380825,-0.0467420332,0.1193272471,0.080814898,0.0008182354,-0.1234039962,-0.1216048822,0.2013993263,0.2504892945,0.2321150005,0.0556145906,0.0039871279,-0.3448368311,0.1688613445,0.6388934851,0.0096107749,0.1933993697,0.0026634368,-0.0553557277,-0.0918763578,0.3654557168,-0.1382660419,0.4473465383,0.0853601694,-0.1106263846,0.0461161211,-0.2001615167,-0.359598875,0.0843449533,0.1504205763,0.3393526375,-0.1611248851,0.3281327486,0.1892798096,-0.4165047705,-0.1966777146,-0.1132400781,0.1603195816,0.0463760197,0.3105352521,-0.2258269191,-0.6363144517,-0.2138471752,-0.2246873528,-0.1736547202,-0.2121387273,-0.0654355213,0.4171052575,0.3795734942,0.1048296541,0.3908360302,0.291942358,-0.1407936513,0.1514684558,0.0768474117,-0.1165332571,-0.4451972246,0.0355963856,0.4403401613,0.3027220666,0.186685428,-0.1826223433,-0.1473050416,0.0416513383,-0.4739942253,0.2552631795,-0.2829425037,0.2505657673,0.0056363838,0.2146390676,-0.5915743709,-0.2966302335,0.1783751994,-0.1917441934,-0.0638269112,-0.0714156181,-0.0907543525,-0.3687000573,-0.2321415544,-0.1907423884,-0.3892237246,-0.2519963682,0.3063248098,0.0470906608,0.1246845797,0.3451837599,-0.0578474961,0.4203667939,-0.0066610253,-0.0449649617,0.0683777705,0.0322420262,0.0629432872,-0.2506498396,-0.3901831508,0.0863297582,0.0814006254,0.0714772567,-0.3299230039,-0.4478314817,-0.4516491294,0.1052510813,0.2513576448,0.0090991259,-0.095794864,0.2081669569,-0.1809079349,0.0521468371,0.1224564537,-0.3523931503,-0.2072187364,0.1554124057,0.3015427589,-0.1551513523,0.1400202364,-0.0365807898,0.2472063452,0.1203057542,-0.162696898,0.2434937656,0.0470384546,0.2754901648,-0.2692047656,-0.3264662027,0.3509713709,-0.138790518,-0.0719210431,0.352723062,-0.1739966124,-0.5435373783,-0.1762143373,0.0883504823,-0.3520294428,-0.3512001336,0.1408772767,0.4890919924,-0.128390193,0.1615461856,0.1004031599,0.1366298944,-0.3128579855,-0.2801979482,0.1468798071,-0.3128330708,0.0150158638,0.0840377435,-0.5422756076,-0.6152954102,0.0904130563,0.0246044658,0.1647312939,0.0106369676,0.1695338786,0.1697925031,-0.0775957331,0.188816458,0.2393735647,-0.2116941214,0.1013790444,0.1028610095,-0.2493129075,0.0044774613,-0.4466750026,0.439561367,0.852701962,0.2162894011,-0.1608830541,0.1470594257,0.4967111647,0.0820081159,0.1896104366,0.1308539808,-0.2746914625,-0.3481545746,-0.2027966529,0.0279221572,-0.2091312557,0.0356773622,-0.0479238555,-0.0573757142,0.3363129199,-0.058140792,-0.1176591888,0.5659456849,-0.0507175215,0.1383468062,-0.0365656577,-0.1675105095,-0.0553859025,0.2056242526,0.3144642413,-0.1491708755,-0.1171460524,0.2581256628,-0.1791644394,0.1546858549,0.5725128651,0.0337055847,-0.0157271121,-0.2549785674,-0.027279675,0.3534435034,0.0384608358,0.391944319,0.0241016448,-0.0969147086,0.2232751846,0.3603267372,0.0063475724,0.0057392865,-0.0559191592,-0.1315788329,-0.1634043753,0.5269417763,0.0964612141,0.9199811816,-0.1351889968,0.232801035,0.3955861628,0.00545754,0.8213492036,0.0364424661,0.2902154028,-0.2200108916,-0.108880125,0.0386196747,-0.0274588279,0.1455073059,-0.1564489156,-0.6769124866,0.2420564592,0.4510639906,0.0814022422,0.0592839159,0.2267391682,-0.0405262224,-0.2793488801,-0.6700119972,0.1651622653,-0.2391075194,0.257181406,-0.3153550923,0.0061789579,-0.1215001643,-0.0194438957,-0.4194890261,0.1596044153,-0.3940165341,-0.1097309366,0.2596352994,-0.2490946949,-0.0036673455,0.1905277818,0.3236683607,0.1867326051,-0.0359215885,-0.0271981619,-0.0723295212,-0.3444035649,0.1388999224,0.0780336335,0.1385015994,-0.2367018461,-0.0868702158,0.0286540482,0.2405577451,-0.178883478,0.2510049343,0.2675097585,0.096522674,0.0387755521,0.0217721816,0.0493119769,-0.1244028285,-0.2563912272,0.0897575542,-0.1048298106,-0.0463675708,0.2597410977,0.1653760672,-0.152979672,-0.0932950005,0.1601149589,-0.1281466484,0.0916573033,0.189836666,0.2613015175,-0.077208586,-0.175209716,0.0454744436,-0.5221547484,-0.3000228107,-0.0752293095,0.0260112975,-0.0544920638,0.0741932094,0.6213572025,0.3452058136,0.1389109343,-0.0982713029,-0.300825268,-0.138250351,0.2220058888,0.3104009926,-0.1053813323,0.433082819,0.4210212529,0.2333650291,0.1701840609,-0.2340855747,0.3082622588,-0.34559986,0.1343726814,-0.0472303405,-0.2782140374,0.1894898117,-0.0686681792,-0.0509331785,0.5490371585,0.1302431673,-0.1948457211,-0.0554988012,0.0648596585,0.0895537585,-0.1063517332,-0.3555725515,-0.080477491,0.0717421994,0.0074633495,0.1199612021,-0.1320906878,0.1513804793,-0.0682617798,0.3728254735,0.2453472018,0.1389769018,0.1199063584,0.0181681886,0.2314891219,0.1649368554,0.2641074955,0.1112998053,-0.0083761625,0.1089110747,0.0121668829,-0.0440864824,0.0655947402,0.1964219958,-0.3878044784,0.0779643506,-0.2351735234,-0.0373917148,0.4894140065,0.1326068789,-0.182484284,0.2481169999,0.0697234198,-0.1159343719,0.0222765394,0.4485885799,0.1829520762,-0.019731136,0.4950132072,0.1573648453,-0.003334658,0.126298219,0.048122704,0.2687554359,0.3688949049,0.0004856915,-0.0677557215,-0.0621735863,0.0352386273,0.1761565059,0.0602183491,0.1521510333,-0.0674016923,-0.1278880984,0.1520903111,-0.2890118361,0.0638224781,0.306622684,-0.0542006977,-0.2557837963,0.1104950905,0.3738567531,-0.0290900692,0.0844082534,0.8333598971,-0.3544097245,-0.3209325671,-0.102040343,0.2567326427,-0.3450084329,0.0990152135,-0.272423774,0.1039467007,-0.0836279988,0.0878348053,-0.1785786599,-0.1320108622,0.3258433938,0.2735810876,-0.253875494,-0.1599981785,-0.1673966348,0.0544953756,-0.1816901863,0.0749422312,0.5572719574,-0.3313541412,0.0993546844,-0.233041048,0.5280355215,0.2896029055,0.2048740685,-0.1909984201,-0.1523385197,-0.0407430679,0.045050703,-0.0147360116,-0.0688188672,0.4951960146,0.100885354,0.1896581352,0.0497773103,-0.1397587359,0.2849245965,0.1317390352,-0.2151464522,0.0373054892,-0.1004583389,0.0037543094,-0.1756564975,-0.3053840101,-0.0506777838,-0.3146498203,0.5148189068,-0.1455922127,-0.1173355728,-0.0424038619,-0.1393557638,0.052861426,0.0952523276,0.6466189027,0.5057282448,0.1590747833,-0.194280535,-0.4506128728,-0.4104737937,0.268981576,-0.3099144995,0.1136656329,0.0504562669,0.1924664676,-0.1657028794,0.4790001512,-0.0266634692,0.1169879436,-0.248109743,-0.1550624818,-0.5336962342,-0.0150511535,0.112137489,-0.329529047,-0.1642404646,-0.0433056839,0.218234241,-0.3421744406,-0.0778337568,-0.0618812405,-0.3067423403,-0.2668486238,-0.075452812,0.2547647059,0.1035224274,0.133173123,0.076661557,-0.3111735284,0.0293025412,-0.3836374283,-0.1451120824,0.0473434851,0.1732350886,0.1977921575,0.0003891314,0.368388623,-0.1935978234,0.5045952201,0.0966085494,0.0646068528,-0.0029809326,-0.0904900581,-0.0720775798,0.086120829,-0.0878100693,0.4204276502,-0.2391373068,0.2577540576,-0.5477001071,-0.2630956173,0.4129124284,-0.4527491331,-0.0418625548,-0.0043557538,0.1946283132,0.1191179156,-0.1933592856,-0.4065580964,-0.0907721967,-0.1022673324,-0.1282304972,-0.4465615451,0.1807406247,-0.1265060753,-0.0096189426,-0.1133438647,0.2060468495,0.044329714,-0.2754811347,-0.0735132322,-0.1813736558]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/222","title":"Colab Notebook breaks when downloading the squad dataset","comments":"Thanks for reporting the second one ! We'll update the notebook to fix this one :)","body":"When I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n","comment_length":16,"text":"Colab Notebook breaks when downloading the squad dataset\nWhen I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n\nThanks for reporting the second one ! We'll update the notebook to fix this one :)","embeddings":[-0.2081260532,0.0852969661,-0.0881832987,0.1644764543,-0.0231155716,-0.1033349931,0.3287060261,0.0826022476,-0.312695682,0.2114281356,-0.3994072378,0.3362334669,0.2900559604,0.1640942395,0.1548372507,0.1645147949,0.0983086675,0.3727589548,-0.066974692,0.1471362859,-0.228515476,0.451759398,-0.3685064018,0.0600332208,-0.2734017372,-0.2093191594,-0.2376538068,0.2732954919,-0.4236905873,-0.2180943191,0.2879371047,0.1839033067,-0.1117460802,0.2599040568,-0.0001143353,-0.2412296832,-0.0203225743,-0.0653363764,-0.3073404729,-0.1606719941,-0.0806202888,-0.2332572192,0.2200177163,-0.409855932,0.1319079399,0.4122153223,0.2227527201,0.28621158,0.293985337,0.2952934802,0.1804483533,0.4040353298,0.1271449178,-0.1776516438,0.325496614,-0.2722629905,-0.2431535721,0.410019666,0.6274053454,-0.0883135051,-0.0656272545,0.1342762262,-0.1280885339,0.1249476671,0.0890843719,0.2477122247,-0.2894340158,-0.4838736355,0.3436742723,0.1162116155,0.1510486752,-0.123151198,0.0589256324,0.0589955933,0.1713833958,-0.2932754159,0.2615860999,0.3350569904,-0.2227059156,0.1343462467,0.0062894667,0.089873597,-0.1818215251,0.1315548718,-0.027879525,0.1601843238,-0.1480301172,-0.0957625434,0.2256306261,0.011602832,0.2465777993,0.0990663767,-0.1621572673,-0.0948356315,-0.1089966148,-0.2423676401,-0.2443491966,0.249994114,0.2372867465,-0.241565004,0.3187463582,-0.0291611906,0.2450104654,0.1847766042,-0.0319995806,-0.1474552155,0.2595814466,-0.102215454,0.4240207374,0.416279465,-0.1882743984,-0.091848217,-0.0244186036,-0.1793902367,-0.2249310911,-0.0915122256,0.0407344513,-0.2405347228,-0.2889055908,0.0778213814,-0.136849463,0.2789804637,0.0691426918,0.2426396757,-0.001410056,-0.4101406932,0.0328567028,-0.0101671061,-0.2695299983,-0.2202639133,-0.0301732346,-0.0911093429,-0.1644539535,0.2610429227,0.2496694326,0.1362547725,0.5002061129,-0.0945061669,0.2376801968,-0.2036774755,0.0361762382,-0.4252381027,-0.0955571607,0.3575981557,0.3498216569,-0.2039547414,-0.2214946151,-0.5597006679,0.1084224805,0.0706817433,-0.035561271,-0.0227709413,-0.0372449458,0.1511408091,-0.2052285224,0.0243086144,-0.3127945662,0.0479006767,-0.2800324559,-0.1421732754,0.0136019904,-0.0340065658,-0.2153875828,-0.0319699533,-0.0166847929,0.2605358064,-0.1001942307,0.0080268718,0.2057422996,-0.046099551,-0.0403769165,0.3128234744,-0.1565088034,-0.0095242597,-0.010018927,0.2053337693,0.3080560565,-0.2581493556,-0.7119438648,0.0582446717,0.0052469377,-0.2625623643,-0.1313274056,0.2376966625,0.5601799488,-0.1987905353,0.114585489,0.4635337591,-0.1890334636,0.0986378938,-0.1500823498,-0.151231721,-0.1275015324,0.1236408129,0.2079047114,-0.0061938888,0.0766653419,0.5572876334,0.1662478298,-0.1265100688,-0.3530400693,0.0721036047,0.2784352601,-0.3214615881,-0.0980319455,-0.1085832566,-0.5646517873,0.0625723302,-0.2864468992,0.0835127607,0.1364913434,-0.2309456617,-0.0580955297,-0.0626566112,-0.0772447884,-0.0365413912,0.1337591112,-0.095026806,-0.0334592611,0.0200085845,0.1139889285,0.3453361392,-0.2902797163,0.5099155903,0.008631492,0.3248543739,-0.2755912244,-0.3000109792,0.1337193698,0.0963592008,0.0810678303,-0.1677869558,-0.145418182,0.2911042571,0.1966377646,0.0310584195,0.145625934,-0.2731467783,0.4598881602,0.0589739382,-0.0357784703,-0.1337578893,0.0844368339,-0.1913496405,-0.0462984331,0.2320576608,-0.2428800315,-0.0481875464,0.1195827127,0.1858395636,0.0412908457,-0.1335269511,-0.1565676481,0.0985781848,0.2388100326,0.2070394456,-0.0013626686,-0.0433761142,-0.3433444798,0.1956359893,0.4620435536,-0.0054820143,0.2048797756,-0.0629556626,-0.0692540556,0.0003487886,0.421723187,0.041216135,0.4217000902,0.0865193158,-0.0906889588,0.0007990873,-0.0349027254,-0.3570867777,0.1690445095,0.1428683847,0.2106125355,-0.1635544896,0.2298246324,-0.0365560018,-0.3995292485,-0.1075950339,-0.0882363096,0.1315501779,0.1349104345,0.3200077116,-0.1169909611,-0.5118410587,-0.0717158765,-0.1881686151,-0.0910916254,-0.2565027177,-0.0470915921,0.5061331391,0.3253305554,0.0137885762,0.3485425413,0.3416335583,-0.0021227498,0.3129969835,0.1184988618,-0.0434658229,-0.289424181,0.0792036131,0.4227304459,0.2414837331,0.212882787,-0.3029239476,-0.1514097899,-0.1465930343,-0.2868507206,0.2941179872,-0.1998257041,0.2345814556,-0.0236976035,0.1992693245,-0.536819458,-0.1262854487,0.1590674222,-0.0104005616,-0.0841393992,-0.0523559637,-0.0166616403,-0.4458810091,-0.2034320235,-0.1740429103,-0.3917513192,-0.2414952666,0.1878699064,-0.1172189564,0.1827541292,0.2395986766,0.0655275881,0.4017173946,-0.0359086171,-0.248591423,0.0184466317,-0.1290365458,0.2076357603,-0.323977232,-0.5115101337,0.1064165905,0.1879418343,0.0032006914,-0.2690907419,-0.4355993569,-0.3778634965,0.0499061234,0.1191340461,0.023841206,-0.2214809954,0.0059342468,-0.2621512711,0.0446545333,-0.0101997964,-0.2739887834,-0.1110514253,0.0766931325,0.3078039289,-0.1971263587,0.1102869809,0.0315279998,0.252944231,0.3045102954,-0.1189804822,0.1805392057,0.1001783684,0.2179147899,-0.2815532386,-0.4495449066,0.3379516304,-0.1619501561,-0.0686090365,0.3591628075,-0.0910074562,-0.5044242144,-0.1716601253,0.0726669729,-0.2678212523,-0.4227606654,0.0708650202,0.4443021119,-0.1476943642,0.1125862822,0.2193157077,0.0198513847,-0.2291978747,-0.3125333786,0.3360985219,-0.2867803574,0.0014979164,0.1557379663,-0.4171317816,-0.6907590032,0.1618874818,-0.050364498,0.1151643023,-0.1106739566,0.2906994522,0.1050530449,0.0392547175,0.3141644895,0.1167536527,-0.1585943848,0.0435778871,-0.0749516264,-0.1944473237,0.0344503336,-0.4641061723,0.5842937231,0.7916564941,0.3910048306,-0.0427203141,0.1442574859,0.5085101724,-0.0329583846,0.1519339979,0.0828854367,-0.1406869888,-0.3031001687,-0.1240240261,0.059478607,-0.0273999143,-0.0606502108,0.0337664858,-0.2005619705,0.308885932,-0.0716144741,-0.0267871395,0.5213577151,-0.0195242446,0.0862450376,0.1034147739,-0.1057771742,-0.1115430072,0.4208713174,0.3046407104,-0.2092673928,-0.267480731,0.2845479846,-0.4900812507,0.0963560343,0.5848205686,-0.0106910989,-0.0310133547,-0.1739789695,0.005066331,0.3778678775,0.0531096235,0.4014867544,0.1770639271,-0.0295399055,0.1776052862,0.2574175596,-0.0942690447,0.0188700091,0.0140363863,-0.1412167251,-0.2320794761,0.4488869011,0.0771218091,0.8232882023,-0.1511875689,0.0664346516,0.1777308136,0.0053941072,0.5558371544,-0.0158804394,0.2272443175,-0.1396790296,-0.4450139403,-0.0225555226,-0.1046130657,0.2890550196,-0.0266328976,-0.7058284879,0.2149589509,0.2657941282,0.1526277512,0.0204157457,0.0557693541,-0.1450773627,-0.2428599298,-0.5719945431,0.2210025489,-0.1751994938,0.23505193,-0.285486877,0.0306153242,-0.2551965117,0.0434356891,-0.4520656168,0.0669053048,-0.482294023,0.0130235497,0.1449777186,-0.2509015799,0.0438169762,0.3447950482,0.387530297,0.1742631346,-0.0434334092,-0.0186491627,-0.0355726555,-0.303409636,-0.0630998015,-0.0169416238,0.0051166583,-0.2645794749,-0.0680678785,-0.070599705,0.3034117818,-0.149382323,0.3579691052,0.1762153506,0.1282219142,-0.0667859837,-0.0537995175,-0.1160093173,-0.1748101413,-0.2913313806,0.1321407408,-0.1205011308,-0.0050696139,0.2978075147,0.1835730821,-0.1227674261,-0.0355562568,0.2913227677,-0.204639405,0.1092683151,0.2187434882,0.2572445273,-0.1284499764,-0.2801272869,0.3003302217,-0.4932375848,-0.1742924899,0.1160794571,-0.1740199625,-0.1275655031,-0.1307646781,0.6403630376,0.4149587154,-0.0102124028,-0.0636111945,-0.3631927371,-0.2070252299,0.2027102262,0.2500629127,0.0259467848,0.2208793908,0.1948773414,0.2190669179,0.0669327006,-0.2703333497,0.373213321,-0.4460939765,0.0552397333,-0.0300208442,-0.2469734251,0.2390450835,-0.1518960595,-0.0337372385,0.5021680593,0.065666385,-0.2611508071,-0.0645296723,0.0680660158,0.0085308542,-0.1281811744,-0.2609419227,0.0047281864,-0.0303533152,-0.0665664226,0.1719762087,-0.0588683598,0.2700216174,-0.1260629296,0.3310445249,0.2087761909,0.1554993689,0.0846957043,0.0133671602,0.2023233324,0.1288287789,0.1213656217,0.1580016613,-0.0707897022,0.2570583522,0.0248776302,-0.0761775598,-0.0078411596,0.3335094154,-0.3386337757,0.1594213396,0.0026592759,-0.166464299,0.3518217206,-0.012282447,-0.1223125011,0.1025402918,0.139198333,-0.1038166955,-0.0515254065,0.5680367351,0.2006270736,0.0301216189,0.5001779795,0.0856855959,-0.0400309414,0.2109019905,0.0440146402,0.3199207485,0.4771461785,0.0748509616,-0.1132102683,0.0257390216,0.2348363101,0.1190771163,0.0564405732,0.1624103487,-0.0450437292,-0.0502601378,0.1110038981,-0.1779201776,0.0893599913,0.4992901385,-0.1034069955,-0.2284978628,0.0429438427,0.1015979573,0.0271739289,0.1244320422,0.7389981151,-0.4879821837,-0.2675542235,-0.1957976669,0.1438278258,-0.4215510488,0.0867404118,-0.3101120293,0.0644109622,-0.0946816355,-0.0019602808,-0.1672723889,-0.0573425889,0.3959063888,0.2621213198,-0.1906858832,0.036120601,-0.1225870475,0.1078355089,-0.1399859637,-0.0120213795,0.4424277544,-0.3063601553,0.2628051043,0.0143405031,0.6072160006,0.1863359362,0.2077461481,-0.1018162817,-0.2245778441,0.0604256056,-0.0875727087,0.0120500028,-0.0550679751,0.4431182146,0.0877959803,0.2683306038,0.0810200721,-0.1402062625,0.2214479297,0.1270504296,-0.1860233694,0.0344328247,-0.2981722951,-0.0920550674,-0.1460090131,-0.2923965752,-0.1762821823,-0.3040123582,0.463080138,-0.2374807298,-0.1588252485,0.0928770006,0.0261720382,0.0717899948,0.0003697877,0.5908598304,0.3205854595,-0.0013031818,-0.2988015711,-0.5059591532,-0.332513839,0.1911905855,-0.1613356918,0.1393865347,0.0850098133,0.18678312,-0.0231140107,0.4586078227,0.1640256494,-0.0808349848,-0.2100596577,-0.0628097057,-0.5517154336,-0.1353833973,0.2651444376,-0.2606633306,-0.1606647521,-0.1617872864,0.2897636294,-0.1578647494,0.0083364034,-0.0447942801,-0.2731545866,-0.3319762051,-0.2486228198,0.1524205953,0.12959297,0.1229709089,0.0075850692,-0.3331421614,0.1209472939,-0.288921535,-0.1068277583,0.0114265969,0.2162542343,0.2503011227,0.028534852,0.2635632157,-0.0987103805,0.4951657653,0.1588351727,0.0285597146,-0.1662837416,-0.0873691738,-0.0726618469,0.0599754602,-0.0456722677,0.4081939459,-0.3233311474,0.2676672041,-0.4084074497,-0.2796310782,0.3924647272,-0.3721630871,-0.073457405,-0.0174338203,0.2117196023,0.1328281313,-0.1360816807,-0.3291109502,-0.0749480277,-0.1162362024,-0.2900268137,-0.4548600912,0.2603672445,-0.0769028962,0.0640920997,-0.1068090051,0.340888679,0.1294019073,-0.2829602063,-0.1672974527,-0.2179584354]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/222","title":"Colab Notebook breaks when downloading the squad dataset","comments":"This trick from @thomwolf seems to be the most reliable solution to fix this colab notebook issue:\r\n\r\n```python\r\n# install nlp\r\n!pip install -qq nlp==0.2.0\r\n\r\n# Make sure that we have a recent version of pyarrow in the session before we continue - otherwise reboot Colab to activate it\r\nimport pyarrow\r\nif int(pyarrow.__version__.split('.')[1]) < 16:\r\n import os\r\n os.kill(os.getpid(), 9)\r\n```","body":"When I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n","comment_length":60,"text":"Colab Notebook breaks when downloading the squad dataset\nWhen I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n\nThis trick from @thomwolf seems to be the most reliable solution to fix this colab notebook issue:\r\n\r\n```python\r\n# install nlp\r\n!pip install -qq nlp==0.2.0\r\n\r\n# Make sure that we have a recent version of pyarrow in the session before we continue - otherwise reboot Colab to activate it\r\nimport pyarrow\r\nif int(pyarrow.__version__.split('.')[1]) < 16:\r\n import os\r\n os.kill(os.getpid(), 9)\r\n```","embeddings":[-0.2108584791,0.3130596578,-0.0493807122,0.0397214517,-0.0872684419,-0.1855655313,0.263124615,0.1595547199,-0.4475891888,0.1087364554,-0.3607413769,0.5732998252,0.1958849877,0.0819763094,0.1732438505,0.1950283349,-0.0088848015,0.3963033259,-0.0574665889,0.1644324511,-0.2231678516,0.3445674479,-0.3933261037,0.1032137573,-0.2219548374,-0.1970246583,-0.1591427624,0.2976085246,-0.3919450045,-0.2167628407,0.276063025,0.1715926081,-0.0599931553,0.1792940795,-0.0001147339,-0.2016182393,-0.0302634668,-0.0780356601,-0.372962594,-0.2322010249,0.0352734253,-0.3360317647,0.3040677011,-0.4480285048,0.1321958601,0.4575712681,0.2886243761,0.3576161563,0.2969409525,0.2759976089,0.1642184556,0.3264267743,0.1345086545,0.0051154834,0.2324992865,-0.2531131506,-0.2693352103,0.4263532162,0.5949573517,-0.2634857297,-0.1326695681,0.1055862978,-0.1691093296,0.1500724107,0.1109077632,0.2283695638,-0.3441568017,-0.4259015918,0.2229602337,0.0606838465,0.0265284982,-0.2624763846,0.0548274443,0.0058779279,0.2763140202,-0.3067046404,0.1578664929,0.3533741832,-0.375043571,0.1199814379,0.0242759734,0.033814501,-0.1993505806,0.1858353913,0.058838632,0.3432296216,-0.0243609287,-0.0138749415,0.2171689123,0.0821247026,0.3446105123,0.2156177759,-0.0392461382,-0.0030000403,-0.13630265,-0.2019575089,-0.1412383467,0.2344243526,0.2335304469,-0.3234660923,0.2888835073,0.0063873846,0.291126281,0.1948486567,-0.0589854792,-0.1128916666,0.2124782652,-0.1368218511,0.3969634771,0.3778134286,-0.1054439396,0.0028238315,-0.0443405323,-0.1690574437,-0.1508910507,-0.1000082046,0.1224658787,-0.174937889,-0.28056705,0.0237823091,-0.2251797915,0.2988460064,0.042632129,0.2417269796,-0.0615691878,-0.376301378,0.0669527054,-0.0250224341,-0.3426691592,-0.2041161358,-0.0023027989,-0.0190374404,-0.1679764092,0.0972097889,0.300075382,0.1617040783,0.4099463522,-0.1188572943,0.1719561517,-0.1090329289,0.0323678665,-0.297097832,-0.0930035561,0.2763929367,0.3880445361,-0.2414577454,-0.2290214598,-0.5571026206,0.0456698202,0.1680841893,-0.0125241354,-0.008397527,-0.1567080021,0.1424696892,-0.2702636421,-0.0481065959,-0.3073248565,0.0109501826,-0.2560473382,-0.1812841147,0.0316573083,-0.0766721144,-0.1764997393,-0.0791532248,0.001023306,0.1980478913,-0.1631062776,-0.0346457921,0.1262241453,0.0313840248,0.0473505743,0.2351509184,-0.1826515347,0.0541924946,0.0288398024,0.2262758911,0.4111819565,-0.2042845339,-0.6995330453,-0.0071706502,-0.0884722844,-0.2417576164,-0.1328661144,0.2476087362,0.4465599358,-0.1435135156,0.210465923,0.5739889741,-0.1675453335,0.1404290944,-0.2227224708,-0.1723982543,-0.0670320243,0.1846537888,0.1432439983,-0.1096738055,0.0119318664,0.6371200085,0.2305385321,-0.1181589216,-0.2321148813,-0.0043965261,0.2663673759,-0.30782336,-0.1000126675,-0.0542001985,-0.5955289602,0.0301957056,-0.2725139856,0.1995994151,-0.0125481524,-0.1999818385,-0.0322625339,0.0027676311,-0.0214746576,-0.0415035374,0.1490339041,-0.1039347872,0.0377619043,0.0164105464,0.1371258944,0.3391304016,-0.2325789034,0.4468241036,-0.0039640521,0.2574146688,-0.2883975208,-0.3253237307,0.1558014005,0.2438882887,0.0305850264,-0.1248803809,-0.0695794597,0.216554746,0.1029064059,-0.0850576013,0.0670518056,-0.3381918669,0.3775502443,0.086209558,-0.0529328287,-0.079106465,0.0766893253,-0.0767324194,0.0185327604,0.2233506739,-0.312307477,-0.0703106746,0.10400635,0.2196217924,0.068207413,-0.1373142898,-0.127741307,0.2108820677,0.229276672,0.2444898784,0.0157216489,0.0384413302,-0.365306288,0.2714901567,0.4710392356,0.0090444358,0.2805314958,-0.0276142675,-0.067835696,-0.0176486485,0.3719348013,0.0231899228,0.3294185996,0.1166572496,-0.1165710315,0.0111387186,-0.1021077707,-0.2971960902,0.11165905,0.1278631538,0.317874074,-0.1010108069,0.3305140436,-0.0343724191,-0.3430919349,-0.2000886053,-0.0932968929,0.2421557605,0.1284599006,0.2591359317,-0.1354102492,-0.6098308563,-0.0550798923,-0.1672679186,-0.0903608426,-0.22800906,-0.0238990728,0.4561626017,0.2761535645,0.0735805854,0.2938821614,0.2140804976,0.0367488675,0.0862001106,0.1559814662,-0.1196362302,-0.3185096085,0.071880497,0.3952266872,0.2435241491,0.0868615359,-0.2415145785,-0.1873185188,-0.0984901264,-0.3265631795,0.2032726705,-0.235672757,0.3059905767,-0.0177470334,0.2388731092,-0.626424849,-0.2101235539,0.1028081775,-0.1280056238,-0.1129534319,-0.0388736688,-0.0058582094,-0.3610163629,-0.2418566942,-0.1475109756,-0.5018795133,-0.2425875813,0.2574924529,-0.0037881753,0.1998457462,0.2799191177,0.0712406114,0.3175303638,0.099508673,-0.2388119847,0.0621660985,0.0220669843,0.2359624356,-0.2466723472,-0.5061873794,0.0659509525,0.0568285398,0.0324610546,-0.2706146836,-0.4002767503,-0.4651815891,0.1409190595,0.2356439829,-0.0490813144,-0.1365013123,0.1378424913,-0.1899331659,0.0716856644,0.0150077883,-0.2303993702,-0.1759864241,0.1301798373,0.2832844257,-0.2543376982,0.0459595025,0.1570527852,0.2229388207,0.1117681637,-0.0831335261,0.1563901007,0.0955341384,0.1974157095,-0.1802189052,-0.4202899039,0.3344489038,-0.11916852,-0.0932377353,0.3436365128,-0.1598406583,-0.6042146087,-0.1298759282,0.0671724975,-0.3866874576,-0.4638876915,0.1615070254,0.5007529855,-0.1398280412,0.250346005,0.1870742589,0.008182602,-0.3484297097,-0.3713101745,0.2292093784,-0.2666554451,0.0425610393,0.2036812603,-0.4245846272,-0.6710538268,0.1856248081,0.077300556,0.1294668317,0.0337335318,0.1617641598,0.1435936838,0.0106841922,0.2023093402,0.2327254117,-0.0438245125,0.0860022455,0.0578109212,-0.2841723263,0.0270791166,-0.3934455514,0.5534501672,0.8214973807,0.217286095,-0.1223302707,0.1616934389,0.5125233531,-0.025665354,0.1680225581,0.1329211146,-0.2282789797,-0.3713434637,-0.1419384331,0.1089479476,-0.0584200174,0.0799485371,0.0968727544,-0.0694009066,0.2451857775,-0.0798706785,-0.0581579283,0.5733703375,-0.0430971831,0.1728099734,0.0742162392,-0.1856635064,-0.1351661831,0.3057743907,0.2549996674,-0.0483992957,-0.2348626852,0.2255316675,-0.3175958991,0.169106096,0.5100568533,-0.028746197,-0.0087077552,-0.2693238556,0.0096917534,0.3701850772,0.0820198506,0.2566702962,0.0761722922,-0.0887145251,0.2006742358,0.2374024391,-0.123861298,0.0246816278,-0.1515250057,-0.1954528242,-0.2053948343,0.4353126884,0.0928364769,0.7878347635,-0.2323566973,0.1613813341,0.2782193422,0.0291964225,0.6408740878,-0.0315153226,0.3367303312,-0.18474181,-0.2963071465,0.0536854714,-0.0211553574,0.1532121301,0.0141970133,-0.6053224802,0.2352608144,0.3834217489,0.0298274588,0.0888908505,0.1362036318,-0.1128017157,-0.3289945126,-0.6385908127,0.1790299714,-0.3138793409,0.1982148737,-0.3054284751,0.0033893166,-0.2572371662,0.0561438799,-0.3699455261,0.1510041654,-0.5142832398,-0.048114147,0.2063174248,-0.3006663918,-0.1024026126,0.3006394804,0.178323701,0.204671368,0.0603949688,0.0615566485,-0.0585905947,-0.3044575453,0.0837135389,0.0371781327,0.0762243941,-0.2088420987,-0.0828780085,-0.0731536373,0.278616339,-0.2377698272,0.3863200247,0.2796519399,0.182061851,0.0202376563,-0.0653482303,-0.0376815759,-0.1724324077,-0.2808220983,0.1083841249,-0.1088260785,-0.0652204603,0.2896270454,0.1406569481,-0.1736134589,-0.0035927158,0.2703547776,-0.1790101826,0.1063856855,0.2071498781,0.2732165754,-0.1199106202,-0.21892865,0.192475751,-0.5636957288,-0.1794547439,0.0294615533,-0.0787200034,-0.0263616797,-0.0818362534,0.6315822005,0.3025016487,0.025656512,-0.0784140974,-0.4058826864,-0.1844429076,0.2054409534,0.1786117256,-0.1519641578,0.3767311573,0.3331971169,0.1933424026,0.0487479605,-0.2389083654,0.2942790091,-0.4090549648,0.1626531482,-0.0574547239,-0.3167048693,0.1807577312,-0.1634580791,-0.0068307254,0.5193841457,0.1276141852,-0.2376262397,-0.0832366794,0.0630093142,0.0208764374,-0.106518276,-0.3156402111,-0.0792458877,-0.0379783511,-0.0308686253,0.0922796428,-0.1109104455,0.212151885,-0.0450857915,0.2560541034,0.2237636447,0.1853878349,0.0336073339,-0.0266250446,0.2572000325,0.1035361215,0.2087652385,0.1726456136,0.0385892913,0.2261850685,-0.041376844,-0.1663669348,0.0798888803,0.3123319745,-0.3684279621,0.0663196594,-0.1207156479,-0.083349742,0.3914603293,0.0945454612,-0.2214663327,0.242218405,0.1205795631,-0.2034115344,0.0594406873,0.5620884299,0.1985332817,0.1175913364,0.5118359327,0.0033109202,-0.0001262217,0.1494716108,0.1118346527,0.2903253734,0.4005498886,-0.0594464801,-0.1089259684,-0.0407776386,0.1046375409,0.1224716827,0.1048654467,0.1897198409,-0.0031355624,-0.089384988,0.0911786705,-0.2128139436,0.0739936829,0.3456501663,-0.0707039833,-0.2760520577,0.0611987263,0.229126811,0.0624627694,0.0791781396,0.7251804471,-0.3464798629,-0.2281428576,-0.0240978412,0.1677068025,-0.3900536299,0.1112294719,-0.257876277,0.1036548242,-0.0732646883,0.0666912422,-0.1540766656,-0.1333083361,0.3501730263,0.2939986587,-0.1731884331,-0.0247330517,-0.1064611152,0.160211727,-0.1082707122,-0.0230335183,0.38537696,-0.3409851789,0.1951876432,-0.184858039,0.5654684305,0.1751043797,0.1869105101,-0.258970052,-0.192229718,-0.0465456806,-0.0091898888,-0.129123345,-0.1091575846,0.4811042845,0.1795620918,0.2597269416,0.0710273832,-0.1390820444,0.2577843964,0.1982862502,-0.2809137106,0.1496539563,-0.2773090601,0.0480881557,-0.170365423,-0.251079917,-0.1099212542,-0.2670922279,0.4196458161,-0.263029784,-0.088808462,0.0830573067,-0.0683716238,0.0652912855,0.0369425416,0.554795444,0.3388822377,0.0179610383,-0.2560866475,-0.4043070078,-0.3963007331,0.2296556085,-0.2289824486,-0.0777447447,0.0866922736,0.2964431047,-0.0500798114,0.4934329987,0.0145937186,0.0442014411,-0.1833363473,-0.0094637275,-0.5380403996,-0.0627145618,0.2135582119,-0.3324807882,-0.0747069269,-0.2555923462,0.2263029367,-0.3445777297,0.011229855,-0.0224714223,-0.3701818287,-0.2412638515,-0.0820788443,0.1729689837,0.079687424,0.0680850968,0.0123003097,-0.2333474904,0.034245491,-0.2586530149,-0.2084958553,0.0408483036,0.2012158036,0.2290243357,0.0860215127,0.3633449078,-0.0953330994,0.5152751803,0.0968579426,0.0451628119,-0.0416998193,-0.0242469404,-0.1285299212,0.1383097321,-0.229961887,0.4407962859,-0.3113746643,0.2767087817,-0.4855020642,-0.3400692344,0.3633518815,-0.5322359204,-0.1876487434,0.0400228389,0.2281888872,0.089812994,-0.1765879393,-0.297727108,-0.0246308949,-0.1129118949,-0.2359892577,-0.4163489938,0.2368520498,-0.1849526018,0.0413483866,-0.0967797711,0.2848264873,0.0388232209,-0.2442216575,-0.2131587118,-0.2190769017]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/222","title":"Colab Notebook breaks when downloading the squad dataset","comments":"The second part got fixed here: 2cbc656d6fc4b18ce57eac070baec05b31180d39\r\n\r\nThanks! I'm then closing this issue.","body":"When I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n","comment_length":13,"text":"Colab Notebook breaks when downloading the squad dataset\nWhen I run the notebook in Colab\r\nhttps:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb\r\nbreaks when running this cell:\r\n![image](https:\/\/user-images.githubusercontent.com\/338917\/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)\r\n\nThe second part got fixed here: 2cbc656d6fc4b18ce57eac070baec05b31180d39\r\n\r\nThanks! I'm then closing this issue.","embeddings":[-0.208000347,0.1311183125,-0.0689535886,0.1150273979,0.0729427859,-0.0872001797,0.3466608822,0.1237105057,-0.3627470136,0.1831555516,-0.3810037673,0.3678531647,0.2798992395,0.1973166317,0.1367719322,0.1388294846,0.0366866216,0.3587137163,0.0139983948,0.2096809,-0.1685106009,0.4450495541,-0.3919082284,0.0554381497,-0.2907463014,-0.1593082398,-0.2253351361,0.2821878791,-0.4978193939,-0.2392632514,0.2430011928,0.1657081395,-0.0679557696,0.213585794,-0.0001134583,-0.2207385302,0.0489730723,-0.0986029878,-0.3085245788,-0.2100302577,-0.0706255212,-0.1923166215,0.160473749,-0.3866461217,0.082940951,0.4993278682,0.2443554848,0.2640855908,0.3487640917,0.3297785819,0.1820563823,0.3344514072,0.2064901739,-0.1256887615,0.2052943856,-0.2290187776,-0.2609091997,0.4513041079,0.6641871929,-0.0891800076,-0.0662815645,0.1382651627,-0.0968769267,0.1690075547,0.0774053484,0.2069268823,-0.3114579618,-0.461828351,0.3329782784,0.0875952393,0.1897955388,-0.1667899489,0.0809335634,0.0814283192,0.2584767342,-0.27224347,0.1998937428,0.3771475554,-0.279560864,0.1227044165,0.0146196559,0.0036369271,-0.2179237306,0.1236137673,-0.0434865318,0.194566384,-0.1518022865,-0.0731945857,0.1731062979,0.0585276224,0.2142842412,0.1883176714,-0.174576655,-0.1037367284,-0.1298794597,-0.2080361098,-0.21136415,0.275696516,0.2824004591,-0.2456433922,0.2479110211,0.008167428,0.2401023507,0.1566415727,0.0072493674,-0.0851952881,0.2620649934,-0.0874027163,0.4075090885,0.4669760168,-0.1642587185,-0.0865847617,-0.03604706,-0.1799150258,-0.2560456097,-0.0964493379,0.1251984984,-0.1873200536,-0.2873094082,0.0005645549,-0.1095102727,0.2790719271,0.0598031841,0.2556667328,0.0077852462,-0.4348564744,0.0816448107,-0.0224618055,-0.2617791295,-0.1772375703,-0.0189602822,-0.0533857644,-0.143808797,0.1923813671,0.269818306,0.1297402531,0.4524067044,-0.1380488276,0.2315982729,-0.1464259773,0.0304498896,-0.3388303816,-0.0771465153,0.2879213095,0.3271437883,-0.2200546265,-0.2076251209,-0.5599486232,0.086480476,0.1126907989,-0.1299939901,-0.0187519267,-0.1415372044,0.1398571581,-0.3022015691,-0.0000323259,-0.4103733003,-0.008756293,-0.2740421295,-0.1606187373,0.0085855294,-0.0273398012,-0.2390177697,-0.0143411644,-0.0075066988,0.2684510946,-0.105573073,-0.0218830891,0.198012501,-0.0372498408,-0.0052462169,0.3029452562,-0.1671164632,-0.0487155952,-0.0716831386,0.252669096,0.2422215641,-0.2639681399,-0.6928662658,0.0702236891,-0.1069377735,-0.2724492848,-0.1525179297,0.1611669064,0.5922826529,-0.1866541803,0.1216317639,0.484823823,-0.1746391654,0.1619457603,-0.183711201,-0.1573564857,-0.1539111733,0.0924505442,0.0971689448,-0.0062494138,0.0376271307,0.6066026092,0.2498885244,-0.1390720457,-0.2768853605,0.0957440957,0.3872683942,-0.3514635861,-0.087607339,-0.1002064422,-0.5366275907,0.0517998748,-0.2348395437,0.1459478587,0.0439407788,-0.2563869655,-0.1318149716,-0.0430254042,-0.1197612211,-0.0752391815,0.1357151121,-0.0876767039,-0.0778127462,-0.0255184807,0.1217923984,0.3803865016,-0.3143617213,0.4444728792,-0.0782162175,0.3720927835,-0.301574707,-0.2973900437,0.13161093,0.1524012089,0.0054536806,-0.2084566802,-0.1773361415,0.2656125724,0.1904724538,0.0155849578,0.1693938822,-0.3337858617,0.4201911986,0.0122425957,-0.0584516488,-0.1337451488,0.095523648,-0.1709362864,-0.0886042267,0.1974509507,-0.2832219303,-0.0441522114,0.1147057042,0.1970719397,0.0903662443,-0.1407027394,-0.1471382231,0.1275740266,0.2020593882,0.2496666312,0.0005398996,-0.0014806493,-0.3013829291,0.2559508383,0.4821514189,0.0206564814,0.1956944168,-0.0156142013,-0.0197305828,0.006954384,0.4005321264,0.0680369735,0.4052345157,0.0753306299,-0.0618728288,0.002083183,0.0772517473,-0.3200520277,0.1458359808,0.1307084709,0.2591795325,-0.1847366095,0.2247592807,-0.036414206,-0.3502981067,-0.1516845822,-0.0358301438,0.1898030341,0.0587607473,0.2596814632,-0.1189309955,-0.4819943905,-0.1068547964,-0.1055798158,-0.0942494944,-0.2273810357,-0.0744827688,0.4537461102,0.3682450056,0.0779672191,0.3577260971,0.249156639,0.0456173196,0.1524127722,0.1455575973,-0.0446643755,-0.2564947307,0.0963600427,0.4505330026,0.2848294377,0.1667255908,-0.2544670999,-0.1176499128,-0.1418276131,-0.3689324856,0.2621166706,-0.2621749341,0.2157849818,0.0277536456,0.3629201949,-0.591100812,-0.1050993726,0.138580367,0.0048896461,-0.1269033253,-0.0483486317,0.0112027926,-0.4445777833,-0.2094198614,-0.1744266599,-0.4479544461,-0.2629992366,0.2177442163,-0.1145668775,0.1625525504,0.2711474895,0.1005349308,0.3782713711,0.0499411821,-0.2660321295,0.0235750601,-0.1346350461,0.164289102,-0.2985039353,-0.5037561655,0.1402677149,0.1419462711,0.0486711562,-0.3073495626,-0.4271419346,-0.4124543965,0.1172805503,0.1399711967,0.0329166837,-0.178127721,0.0529593155,-0.2759559751,0.0634527281,0.034428373,-0.1994224638,-0.1339199245,0.045047503,0.3332173526,-0.1881596744,0.1346915662,0.0900110081,0.2233042419,0.2953181565,-0.0872029662,0.1578608453,0.0370968729,0.2386831492,-0.2625629604,-0.4269649982,0.3910203576,-0.1376401782,-0.0601215027,0.3173478246,-0.1053234562,-0.4679929316,-0.146422714,0.0882774442,-0.2908807695,-0.4439518154,0.1352860481,0.4762610793,-0.1156432927,0.1536976248,0.1461688876,0.0267154165,-0.2437725663,-0.3218799829,0.3201253414,-0.2219921798,0.0005549483,0.2043695152,-0.3991651535,-0.7189021111,0.1619279385,-0.0400942974,0.1332280934,-0.04490944,0.2782712877,0.1231217533,0.0053053121,0.3151809573,0.2239215672,-0.121586889,0.0075316541,-0.0481389016,-0.1539657116,0.0961112753,-0.3275452256,0.556502223,0.8027661443,0.3761501908,-0.1492159516,0.1100958735,0.5063599348,-0.0659003779,0.1549176127,0.0793353543,-0.188364163,-0.3089991212,-0.2350319028,0.0319119059,0.0259234644,-0.0226873718,0.0692826733,-0.2025506943,0.3092567623,-0.1297615319,-0.1166027486,0.5049629807,0.0013170758,0.1368880123,0.1081445143,-0.1205634326,-0.1119055152,0.3644821048,0.3353237212,-0.1760385931,-0.2973375022,0.2741277218,-0.4633214176,0.0769549683,0.5603045225,-0.0735384002,-0.0785080865,-0.1335537881,0.0160793103,0.3251824677,0.0416391231,0.3266753256,0.1699848026,-0.0300997347,0.1980641484,0.2994544208,-0.1191482916,0.0837922245,-0.0577829294,-0.0403005555,-0.227744475,0.4167511463,0.0723613948,0.8054533601,-0.2156077921,0.0635764226,0.1947269142,0.0262326542,0.5052211881,0.0624465719,0.1975925118,-0.1528550088,-0.4009876847,-0.0182066876,-0.0629722551,0.252304107,0.0192841124,-0.667804122,0.1419852674,0.3226876855,0.1429054439,0.0673317388,0.1223155856,-0.1514829695,-0.2606187463,-0.6025719047,0.2146777213,-0.2428113818,0.2203584909,-0.2177027315,-0.0167237632,-0.2474472672,0.0523803718,-0.4209160209,0.131641835,-0.4955945313,-0.0233840011,0.1463971734,-0.2523238957,0.0915005729,0.315415442,0.2644820213,0.1589600146,-0.0836627856,0.0874233842,-0.0327432416,-0.3563277125,0.0259803683,-0.0120931668,0.0076508825,-0.2713181973,-0.0953677967,-0.111125268,0.2889449894,-0.1868305504,0.3706156611,0.2538784444,0.1988428086,-0.0556719638,-0.1301703304,-0.0955785289,-0.1931742728,-0.2583744824,0.1312688291,-0.0932532996,-0.0741196126,0.2414035499,0.1808232516,-0.0922572017,-0.0074896305,0.2870039642,-0.1994900852,0.05950455,0.2440324724,0.2315005511,-0.0971633494,-0.262082994,0.2788748443,-0.4314832985,-0.1937679648,0.1052012742,-0.0476255529,-0.0834446624,-0.0986733884,0.5808262229,0.3383184671,-0.0311041176,-0.0781417862,-0.3889973164,-0.2447638959,0.1507959813,0.1820154041,0.0094955936,0.2315458506,0.3078391254,0.2378011942,0.0656329989,-0.2772880197,0.4072390795,-0.3967259526,0.0705100223,-0.0062409267,-0.2565154433,0.2605752051,-0.1794032454,-0.0186540801,0.576428473,0.0812464729,-0.2645466924,-0.0373022594,0.0477209501,0.0210692547,-0.0735251382,-0.2560116649,-0.0271131825,-0.1455335915,-0.0385226384,0.145388633,-0.0539488383,0.2433751076,-0.1822547466,0.3193826675,0.1291271299,0.1661770493,0.0199560951,0.0548924357,0.2613896728,0.1471158117,0.1572144777,0.0600314848,-0.0045967777,0.2129301578,0.0159738921,-0.0501893461,0.0703098401,0.3648307621,-0.3742870986,0.1038021818,-0.0066659385,-0.0856384784,0.3218301535,-0.0066105546,-0.1178249046,0.1086240038,0.1236559078,-0.1271725148,0.0631069317,0.5514030457,0.1502996683,0.0726605207,0.4775371253,0.0522156209,0.0076981708,0.1544807851,0.0695264563,0.3542718589,0.4342391491,0.0169095173,-0.1059965491,-0.0320712253,0.1588783711,0.136624068,0.0621736906,0.1691346914,0.0067429291,-0.0634568706,0.147762984,-0.2397443354,0.1062158048,0.4204717577,-0.0891481042,-0.219744131,0.0541990288,0.1109663621,0.0245908201,0.1269030869,0.7777938843,-0.4915835857,-0.2723706067,-0.1230673939,0.1745546162,-0.4187075198,0.1487131864,-0.2641626596,0.1464376003,-0.1235950962,-0.0125346174,-0.139546141,-0.0999498442,0.3498202264,0.2686992884,-0.1834128499,0.024788538,-0.1911020726,0.1069845259,-0.144074738,-0.0402034037,0.4275154471,-0.3301756084,0.2518854737,-0.0028089311,0.5289728642,0.1914322972,0.1563474238,-0.0848890096,-0.1978517622,0.0224243198,-0.1007206067,-0.0124071091,-0.1029590964,0.4325856864,0.0316646025,0.2388427258,0.0848970711,-0.1426458955,0.1663380116,0.1534211636,-0.1956925988,0.0927766412,-0.2757639289,-0.1296402365,-0.1347715855,-0.3542976081,-0.2259337306,-0.3081461191,0.4624453783,-0.2740243077,-0.1815547645,0.0653083473,-0.0300256107,0.0781447589,-0.0190741625,0.5220407248,0.3500952721,0.0120395739,-0.2863483131,-0.4926591218,-0.3841700852,0.2048263699,-0.0896075293,0.0426191725,0.1019567922,0.233040601,-0.0275083054,0.4352608025,0.0975869521,-0.0636153072,-0.1705184132,-0.0946774706,-0.5537119508,-0.0990910903,0.242298618,-0.2517946661,-0.1135677174,-0.1680899858,0.2829031646,-0.1578952372,-0.0091562811,-0.0379313231,-0.3113216162,-0.3501734734,-0.2668178082,0.1066779345,0.0922556818,0.0777808577,-0.0217788517,-0.2777342796,0.1033152267,-0.2840825021,-0.1371327192,0.0635190904,0.3224415779,0.2819181383,0.0687812418,0.2750305533,-0.1914798766,0.4501879215,0.1188666001,-0.0525332317,-0.1590485871,-0.0355969369,-0.1117573977,0.1297900379,-0.1382116377,0.4617345631,-0.3245925903,0.2496958375,-0.4447593093,-0.3770932555,0.4675891995,-0.3777472079,-0.1186699495,0.0732677504,0.2061454207,0.0726084933,-0.169558838,-0.395750463,0.0263819322,-0.1597999334,-0.2242603749,-0.4217966199,0.2456750423,-0.1071864367,0.1469579339,-0.1178226769,0.3748516738,0.0999529734,-0.1928541511,-0.1147420928,-0.1635625958]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"I like this feature! I think the first question we should decide on is how to convert all datasets into the same format. In T5, the authors decided to format every dataset into a text-to-text format. If the dataset had \"multiple\" inputs like MNLI, the inputs were concatenated. So in MNLI the input:\r\n\r\n> - **Hypothesis**: The St. Louis Cardinals have always won.\r\n> \r\n> - **Premise**: yeah well losing is i mean i\u2019m i\u2019m originally from Saint Louis and Saint Louis Cardinals when they were there were uh a mostly a losing team but \r\n\r\nwas flattened to a single input:\r\n\r\n> mnli hypothesis: The St. Louis Cardinals have always won. premise:\r\n> yeah well losing is i mean i\u2019m i\u2019m originally from Saint Louis and Saint Louis Cardinals\r\n> when they were there were uh a mostly a losing team but.\r\n\r\nThis flattening is actually a very simple operation in `nlp` already. You would just need to do the following:\r\n\r\n```python \r\ndef flatten_inputs(example):\r\n return {\"input\": \"mnli hypothesis: \" + example['hypothesis'] + \" premise: \" + example['premise']}\r\n\r\nt5_ready_mnli_ds = mnli_ds.map(flatten_inputs, remove_columns=[<all columns except output>])\r\n```\r\n\r\nSo I guess converting the datasets into the same format can be left to the user for now. \r\nThen the question is how we can merge the datasets. I would probably be in favor of a simple \r\n\r\n```python \r\ndataset.add()\r\n```\r\n\r\nfunction that checks if the dataset is of the same format and if yes merges the two datasets. Finally, how should the sampling be implemented? **Examples-proportional mixing** corresponds to just merging the datasets and shuffling. For the other two sampling approaches we would need some higher-level features, maybe even a `dataset.sample()` function for merged datasets. \r\n\r\nWhat are your thoughts on this @thomwolf @lhoestq @ghomasHudson @enzoampil ?","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":291,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nI like this feature! I think the first question we should decide on is how to convert all datasets into the same format. In T5, the authors decided to format every dataset into a text-to-text format. If the dataset had \"multiple\" inputs like MNLI, the inputs were concatenated. So in MNLI the input:\r\n\r\n> - **Hypothesis**: The St. Louis Cardinals have always won.\r\n> \r\n> - **Premise**: yeah well losing is i mean i\u2019m i\u2019m originally from Saint Louis and Saint Louis Cardinals when they were there were uh a mostly a losing team but \r\n\r\nwas flattened to a single input:\r\n\r\n> mnli hypothesis: The St. Louis Cardinals have always won. premise:\r\n> yeah well losing is i mean i\u2019m i\u2019m originally from Saint Louis and Saint Louis Cardinals\r\n> when they were there were uh a mostly a losing team but.\r\n\r\nThis flattening is actually a very simple operation in `nlp` already. You would just need to do the following:\r\n\r\n```python \r\ndef flatten_inputs(example):\r\n return {\"input\": \"mnli hypothesis: \" + example['hypothesis'] + \" premise: \" + example['premise']}\r\n\r\nt5_ready_mnli_ds = mnli_ds.map(flatten_inputs, remove_columns=[<all columns except output>])\r\n```\r\n\r\nSo I guess converting the datasets into the same format can be left to the user for now. \r\nThen the question is how we can merge the datasets. I would probably be in favor of a simple \r\n\r\n```python \r\ndataset.add()\r\n```\r\n\r\nfunction that checks if the dataset is of the same format and if yes merges the two datasets. Finally, how should the sampling be implemented? **Examples-proportional mixing** corresponds to just merging the datasets and shuffling. For the other two sampling approaches we would need some higher-level features, maybe even a `dataset.sample()` function for merged datasets. \r\n\r\nWhat are your thoughts on this @thomwolf @lhoestq @ghomasHudson @enzoampil ?","embeddings":[-0.0606288463,-0.4335705638,-0.0566293858,-0.0154857263,-0.1655908674,0.0452480465,0.1855864227,0.1840500981,0.2041598558,-0.1490784883,-0.2486168891,0.2844127417,-0.1800078154,0.471552074,0.3802833557,-0.4318719506,0.0726252794,-0.148357302,-0.327953428,0.346521765,0.0434317663,0.1234800965,-0.1807239652,0.0299022216,-0.3513098955,-0.1616335064,-0.385071367,0.0748649538,0.0141172633,0.0314689688,0.0698803589,0.3947065771,-0.0416435264,0.1824833006,-0.0001149848,-0.1062701494,0.1102551967,-0.1087463349,0.0937329605,-0.0340781435,0.1338173896,-0.3790647686,0.0580893494,-0.1654161513,0.2026768774,0.1457469612,0.1934349239,0.0513041392,0.3670559525,-0.2185335159,0.0840293467,0.2644307911,-0.1348502785,0.3218304515,-0.0651739612,0.0204702914,0.0448855683,0.2399405688,0.6195868254,-0.4785763025,-0.0277771205,0.2252210081,-0.0118415989,-0.0123158842,0.1250130832,-0.0638720319,-0.2911251783,-0.2389596552,-0.3664334118,0.244552955,-0.0434679091,-0.2190153152,-0.2268621624,-0.5487818718,-0.025222566,-0.0166294333,-0.3268792927,0.0636666864,-0.1695635319,-0.0531394035,-0.1652792394,-0.0777100474,-0.0644057468,0.1307319701,0.29616189,0.5442544818,0.0387099311,0.0535660908,0.2903420925,0.036766436,-0.1127419621,-0.2751785219,0.3076987863,0.0787266791,-0.3209531307,-0.3482193351,0.0695451275,-0.3125925362,0.321803242,0.0821104124,0.1377024502,0.0635517836,-0.130617559,0.2175747156,0.3832532763,0.0085670156,-0.075083077,-0.1686054468,0.0076402049,-0.1426707357,0.1661439538,0.1513280123,0.1184850931,-0.0039990922,-0.449752301,-0.0621160902,-0.0710137859,0.0521856472,-0.1257834733,-0.5080183744,-0.1141923368,-0.0967443734,-0.1139835641,-0.1760416627,-0.2671769857,0.1769250333,0.0036277203,0.2718095183,-0.1638914794,0.039084278,-0.0379803069,0.0168294962,-0.4351599813,-0.0823584646,0.2814767659,0.0876100659,0.0532539971,0.1616655886,-0.0415853783,0.0655003637,0.3322538435,0.0286847167,0.1381819993,-0.045800969,0.1786010563,-0.2670104504,-0.010550187,0.1570933014,-0.3184620142,-0.1224097684,-0.1524710655,-0.2089324147,0.0530343391,0.0109315291,-0.1749028862,-0.2891996503,-0.4556137323,0.7248809934,0.0154134696,-0.0831962973,-0.1941859424,0.1386067867,-0.1514987499,-0.0932872444,0.2236024737,0.1155076027,-0.0781733617,-0.2775520086,0.0191004183,-0.0802846476,0.0127374558,0.3045518994,-0.0646618158,0.2548953295,0.0614999644,0.2096288353,0.3009223938,-0.3996349275,-0.0024381455,0.3445948064,-0.055489637,0.3846751153,0.102682367,0.1290329546,0.0022143286,-0.1920385659,0.2654208541,0.9147472382,-0.3811212182,0.0671110302,-0.1982171685,-0.3537553251,0.5127952099,0.2841843069,0.0111314887,-0.3445620537,-0.1330880076,-0.0049945884,0.0784334987,0.0030070671,0.0373707712,-0.0265195891,-0.3109757006,0.0754680708,-0.1949543655,-0.1152403504,-0.4247468412,-0.1003589481,0.0457797125,0.3210237622,0.2605730593,-0.1533923596,0.2268137187,-0.3086336255,-0.1556716412,-0.2565404773,0.133099854,-0.0267518703,-0.0144738918,-0.2060856074,-0.0964523628,0.3236779273,0.0577925332,-0.075898692,-0.4959031343,0.2482441366,-0.0777617618,0.0200575497,0.0795487389,0.6802396774,-0.1879968047,-0.0758890882,0.1854728758,0.2653409243,-0.3162189126,0.2538578808,0.2200596333,0.3308732212,0.2140356302,-0.0697963163,0.1053741053,-0.063786611,-0.0416564196,-0.2234189659,-0.1915897131,0.4375747144,-0.1759518981,0.4636526704,-0.0321757197,0.0291524902,-0.2214388549,-0.0169028547,-0.2342641801,0.3847416639,0.3131430447,-0.2077127546,0.3308096528,-0.0469722785,-0.4412020147,0.2577673495,0.2447435558,0.0681833327,0.0957373753,-0.00955196,-0.0455345325,-0.3349420428,-0.0602991097,0.3420481086,0.5666439533,0.269957602,0.198374629,0.0683952942,-0.2664469481,-0.112411201,0.069297798,0.162815243,0.0128067406,0.2764882147,0.0359069817,0.0701989755,-0.0163582657,-0.0904838219,0.139838621,-0.1848670244,-0.159550786,0.0135105485,-0.2328624427,-0.6612980962,-0.2655380964,0.0595658682,-0.3062059283,-0.3195210397,0.1293625385,0.0172318295,-0.1474232078,0.1712545007,0.2321963161,0.6074471474,-0.3907059431,-0.1505924165,-0.0688083544,-0.3318816125,-0.1426523924,0.0778740942,0.5498899817,0.3654055893,0.4354554713,0.1750751734,0.0016457507,0.0403154679,-0.4197554588,0.0945982933,-0.3344560564,0.1600539535,0.0628429726,-0.1625745147,0.1456708461,-0.4231190979,-0.0899715796,0.0243899524,-0.0978647098,0.0175211001,-0.039675504,0.0278717205,-0.2234395891,-0.3024547994,-0.57163167,-0.4520238638,0.3431043327,-0.2798540294,0.144758299,0.0694180578,-0.225675419,0.1413625777,-0.1748025268,0.0703379437,-0.113563478,-0.1060817465,0.0421908833,-0.3436838984,-0.2414879054,-0.3852685094,-0.0269934312,0.1701690406,-0.0015300703,-0.0601092353,-0.2366360724,0.073477976,-0.2192817777,0.1113261655,0.2957314253,0.1376898587,-0.1666657329,0.1337774694,-0.0499742515,-0.1868763119,0.2078098208,0.3281923532,0.3863656819,0.1527036577,0.0017181401,0.2775653601,0.6705088615,0.3482151628,0.0438914299,0.0903099328,0.2588980198,0.0604608692,0.1286640763,-0.1806706488,0.2668640614,-0.2000270188,0.5133734345,0.1512197107,0.1806649417,-0.105412066,-0.2131503969,0.1212695912,-0.3510890901,-0.1996563524,0.3020295203,-0.3192821443,0.0215822048,-0.0544899441,-0.3992006183,-0.2533683479,-0.0495545864,0.0174266789,0.3520646393,-0.0550807454,0.1988597512,-0.4190728962,-0.0917042866,-0.4600532949,0.1969708651,0.0897412524,0.3336328566,-0.0685614347,-0.1117724925,-0.0851778239,0.1311123073,0.5710863471,-0.5576235056,-0.2217064798,0.1770237088,0.150617063,-0.0835627019,-0.122414887,-0.3805663884,-0.0267149396,0.4663621783,0.4254289567,-0.4082848132,-0.3596423566,0.2669548988,-0.1686519533,-0.0206888039,-0.1301175058,-0.2035482079,-0.2153304368,-0.1732441336,0.0265814029,0.2245212346,0.1689639688,-0.1544281542,0.1507239491,-0.0460218713,0.2372141778,0.2111849785,-0.0128336959,-0.0503892936,0.2638848424,0.0650327131,0.0724654943,-0.0306108594,-0.2702559233,0.1069723964,-0.1329270154,0.0665324107,0.2735386491,-0.0788705721,0.7340449095,0.1507013738,0.386365205,0.1744665802,-0.1042374074,0.0202678684,-0.2474548817,0.2143669277,0.2087725401,0.2412242442,-0.1877697408,-0.1901138276,0.5577899814,0.330306679,-0.0156930443,0.4144617021,-0.4715434313,-0.2955442369,0.3173158467,0.0374611244,0.7747870684,-0.055943504,0.3117325306,0.0889692158,-0.4208640456,0.570625186,-0.4346334934,0.1579667032,-0.0996921584,-0.1840039343,0.0302602109,-0.1004111022,-0.0264998302,0.2533987164,-0.3821482956,0.3142702878,0.4203944504,-0.2569933832,-0.0318915062,0.5367091298,-0.053168308,-0.392829597,-0.0563138239,0.0418116227,-0.3273864985,0.088632755,-0.1278613955,-0.3133984804,0.2392905504,0.1586398333,-0.3130239248,0.1046496555,-0.0630730316,0.142688632,0.2242194116,-0.0350155234,0.2476468235,-0.139314279,0.0429926328,-0.1307621151,-0.0667176992,0.033009313,-0.102061592,0.0768488348,0.098597832,-0.1636138558,0.5141546726,-0.0033412864,0.0088264113,-0.2976303399,0.2612644136,-0.3895632327,-0.2769506276,-0.0246763229,0.1837065667,0.0999713764,-0.2911480665,0.6143852472,0.0873111486,0.0357684344,0.0402521044,0.138581723,-0.2340051383,0.5710292459,-0.2432587296,-0.0142896026,0.0837167203,-0.0137015525,-0.1130398586,-0.4529515505,-0.0207377318,0.3857305348,0.0724535808,-0.0222053658,0.0676672757,0.130558461,-0.0608335845,-0.0172829982,-0.251711756,-0.2992796898,0.1024886593,0.2488079816,0.2466531098,0.1551619619,-0.0419434868,-0.2431480736,-0.1397524923,0.5461684465,-0.030903006,0.262034744,-0.0514493547,0.3808258176,0.3815828264,0.5050747395,-0.2507598102,0.0735071525,0.0664884076,0.1924358308,0.441975981,-0.2678069472,-0.0323338844,0.2462565154,-0.135087207,0.1086009368,-0.3244997859,-0.1510504633,-0.2054608166,0.10554111,0.1972669661,0.0291513205,0.0475096181,-0.0335397683,-0.1311680228,-0.1067837104,0.021673711,-0.3718713522,-0.1279652417,0.2309479266,-0.1180998981,0.0367706679,0.0795671418,0.1350078285,0.0044633471,-0.0925265029,0.3175365329,0.0924281552,0.0663458183,-0.1891515404,-0.1855026335,0.042386964,0.2321739197,0.1891829967,0.0023197513,0.0865117386,-0.170593977,-0.1667358875,0.2126780599,0.2336452901,0.0103250016,-0.0913672596,0.1436550915,0.0954991058,-0.0544308424,-0.0052159633,0.2259073853,0.0599536262,-0.0072142021,0.0664542839,0.1982610673,0.3934859931,-0.1817295253,0.045478452,0.0352600962,-0.3973194957,0.1519804746,-0.0400994718,0.0002017888,0.0052864379,0.0600671582,0.0377881974,0.1249130368,-0.3215548694,0.3289367259,0.2629734874,0.0259109959,0.0355755091,0.1799469292,-0.0284975097,-0.1554695964,0.2796450257,0.0755622536,0.4324940741,0.0586093217,0.8388553262,-0.2928149402,-0.5137993693,0.1134905592,0.4244175851,-0.0020872834,-0.1941960305,0.0541372485,0.1188908517,-0.3490165472,0.0717266575,-0.0671637133,-0.1238759607,0.5954538584,-0.0036205787,0.1662233323,-0.2218689471,-0.1740963012,0.2818503678,0.239363879,-0.1145233735,0.1149064898,0.0892239213,-0.1297858357,-0.1515509784,0.5569121838,0.2587621212,0.0525709763,-0.4229050279,0.0142324986,0.0955644473,0.1088169366,0.0758141205,0.0894236341,0.4531348646,-0.1393468678,0.2778194547,-0.0114751989,-0.0116966125,-0.154130742,0.0335245654,-0.4807212055,-0.343821913,0.3919151425,0.1465265453,-0.1690303534,0.1038100198,0.2620704174,-0.7270407677,-0.2905247808,0.4198157787,-0.0289025381,-0.0365145579,-0.2069344819,0.0661968738,0.1884429455,0.5728139281,0.2763862908,0.3275372088,-0.3885541558,-0.1176983789,-0.5355229378,-0.1727130264,0.1998890787,-0.0491161607,-0.186818704,0.1479757577,-0.0214330927,0.3246808052,-0.2616149485,0.1089248657,0.1682987064,0.0734911412,-0.5057321191,0.1196029335,0.2068788558,0.0492258854,0.1989087909,-0.0570582636,-0.0849899277,0.1317054927,-0.0763208196,-0.0857615545,0.0058662011,0.2451672852,0.1640406549,-0.0264625214,0.0823260024,0.4630802572,-0.0138724959,-0.1993646771,-0.0198038984,0.0445742272,-0.4071347713,-0.0702898204,-0.0931831598,0.1952257752,-0.3953464329,0.2937243283,0.0084499363,-0.1166602895,0.0764623657,0.0287596285,-0.1516787112,-0.0498309359,0.0878543854,0.1956941932,0.2983800471,0.7401274443,-0.0052439589,0.1236115098,-0.1726138741,-0.0690208748,0.0274100304,-0.0532236919,-0.2343157828,-0.1020807028,0.061302498,0.1682105809,0.217233628,-0.6970360279,-0.3775851727,0.0372025073,-0.0556473583,0.0134245502,0.7778042555,0.1314222962,-0.1681364775,-0.1608691961,-0.3078275025,0.0111371186,-0.2971332371,0.0481088646,-0.4513056576]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"I agree that we should leave the flattening of the dataset to the user for now. Especially because although the T5 framing seems obvious, there are slight variations on how the T5 authors do it in comparison to other approaches such as gpt-3 and decaNLP.\r\n\r\nIn terms of sampling, Examples-proportional mixing does seem the simplest to implement so would probably be a good starting point.\r\n\r\nTemperature-scaled mixing would probably most useful, offering flexibility as it can simulate the other 2 methods by setting the temperature parameter. There is a [relevant part of the T5 repo](https:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/03c94165a7d52e4f7230e5944a0541d8c5710788\/t5\/data\/utils.py#L889-L1118) which should help with implementation.\r\n\r\nAccording to the T5 authors, equal-mixing performs worst. Among the other two methods, tuning the K value (the artificial dataset size limit) has a large impact.\r\n","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":126,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nI agree that we should leave the flattening of the dataset to the user for now. Especially because although the T5 framing seems obvious, there are slight variations on how the T5 authors do it in comparison to other approaches such as gpt-3 and decaNLP.\r\n\r\nIn terms of sampling, Examples-proportional mixing does seem the simplest to implement so would probably be a good starting point.\r\n\r\nTemperature-scaled mixing would probably most useful, offering flexibility as it can simulate the other 2 methods by setting the temperature parameter. There is a [relevant part of the T5 repo](https:\/\/github.com\/google-research\/text-to-text-transfer-transformer\/blob\/03c94165a7d52e4f7230e5944a0541d8c5710788\/t5\/data\/utils.py#L889-L1118) which should help with implementation.\r\n\r\nAccording to the T5 authors, equal-mixing performs worst. Among the other two methods, tuning the K value (the artificial dataset size limit) has a large impact.\r\n","embeddings":[-0.0724446997,-0.4672179222,-0.0401601158,-0.0550321825,-0.3045603633,0.0747443065,0.1744338721,0.0672955289,0.2653243542,-0.1193471253,-0.2597554028,0.3054882586,-0.1694985777,0.3780532479,0.3509474993,-0.4499824047,-0.0215500556,-0.146344617,-0.2921186686,0.3242069781,0.0543157682,0.1164425984,-0.1543542147,0.0661999434,-0.275216639,-0.1766957045,-0.3439988196,0.123409681,0.0848320276,0.0797640905,0.1358439922,0.3954449594,-0.0185932722,0.1279855967,-0.0001114127,-0.0947610736,0.0992532969,-0.0971221849,0.1526368111,0.0399923772,0.1779924929,-0.3429228663,0.0634030551,-0.1400868893,0.1315973103,0.2471916676,0.197243154,0.0887655392,0.4500890672,-0.3533523977,0.1048917621,0.3131383359,-0.2438293546,0.3010331988,-0.0856859535,-0.1528030336,0.0115065798,0.2888505161,0.6218152642,-0.5280056596,-0.0151433442,0.1048829183,-0.0462879315,0.0499921255,0.131043449,-0.066638656,-0.2823973,-0.253523469,-0.3627516925,0.2775124609,-0.1223972142,-0.0628897771,-0.1925864369,-0.565694809,-0.0037437642,0.0562744327,-0.3290131688,0.0878174901,-0.2160571516,0.0150353275,-0.124979414,-0.126868397,-0.007691354,0.1114815921,0.3081286848,0.5256880522,0.0743688941,0.1346539259,0.3524874151,0.1015161201,-0.1408928931,-0.2651450634,0.2907785773,0.1354706734,-0.3288773298,-0.377097249,0.1516861916,-0.2695581913,0.2802706063,0.0967135876,0.1468145251,0.052291058,-0.0826511756,0.264383316,0.3891301453,0.0126319667,-0.0967651606,-0.1197434366,0.0239591077,-0.105910629,0.1915580928,0.1394885629,0.1446611136,0.0090050101,-0.4349189997,-0.1123368219,-0.1909623295,0.0965473577,-0.180518657,-0.5517173409,-0.0857002214,-0.1018367335,-0.0204156153,-0.1243636906,-0.2886069715,0.1530839503,-0.0747468993,0.1504200399,-0.1804399937,0.0207882784,-0.0701470748,0.0663670599,-0.4629813135,-0.0349683575,0.2592130601,0.105593577,0.0714485869,0.1174032763,-0.0205039736,0.1679399163,0.426109314,0.0308056287,0.0287274756,-0.0310468953,0.1476192325,-0.3091061711,-0.0313938111,0.2360494584,-0.3518314362,-0.0972793326,-0.1445493251,-0.1775629669,0.1374153048,0.0454991013,-0.2158234864,-0.2692705393,-0.4415575564,0.7260335088,-0.0132843517,-0.0239318404,-0.1717699319,0.2594797313,-0.2092255652,-0.1180295646,0.1894045472,0.0955987945,-0.0213698354,-0.3190785348,-0.1060383692,-0.0609196313,-0.0156160584,0.3213067055,-0.1517840624,0.1992661208,0.1068785638,0.1746224463,0.2618967593,-0.4294566512,-0.0568967387,0.3527350724,-0.0614716336,0.3841181993,0.0508754663,0.1725042611,-0.0096142683,-0.2168364078,0.2445953488,0.9837132096,-0.4041246474,0.0024758638,-0.1859401464,-0.4244100451,0.563744843,0.3482621908,0.0648248196,-0.2935227156,-0.1886730939,0.03436758,0.0467180461,0.0923098326,0.0749095082,-0.0802322179,-0.2644890249,0.0103593497,-0.2241342962,-0.1698400527,-0.4216531515,-0.0087363916,-0.0369628407,0.3787116706,0.3727781773,-0.1403525323,0.2375966161,-0.3064366579,-0.1114936844,-0.3117160499,0.1445584893,-0.0680671409,0.0125660766,-0.2394195795,-0.0652765334,0.2053758502,0.0974027514,-0.1050228104,-0.4230077267,0.2914787829,-0.0407337584,0.0215077512,-0.0130759804,0.6146175861,-0.216123417,-0.1165854037,0.2079827189,0.2915941179,-0.3138335943,0.21099931,0.3093211353,0.3545559943,0.1958430856,-0.0614049137,0.0992705673,-0.012211035,-0.0899136811,-0.1710817516,-0.0809546337,0.4797086418,-0.2012840956,0.357545495,-0.0219102651,0.0348964967,-0.2077240497,-0.0569221973,-0.1906933188,0.3881003559,0.3106745183,-0.2279953659,0.3347797096,-0.0767852217,-0.4003093839,0.2434422076,0.2139726877,0.0134796128,0.090265505,-0.0230680667,0.0312409755,-0.3101988137,-0.0605027825,0.3375441432,0.5514169335,0.2894964516,0.1773703694,0.0167293996,-0.2175824344,-0.0726922899,0.0426138304,0.1653773636,-0.0631872639,0.1419843733,-0.041099038,0.098900713,0.0266580731,-0.1013484895,0.1228195503,-0.1445389092,-0.0426808633,-0.0375057459,-0.1547565758,-0.5833342671,-0.1911754608,0.0525763854,-0.3515312374,-0.2960594594,0.2087562084,0.0568543263,-0.1775723994,0.1679805666,0.1997511536,0.6398572326,-0.3923530579,-0.1751709133,-0.0524817295,-0.2999004722,-0.1152332574,0.137364313,0.4358258843,0.3023828566,0.4983601272,0.2307027131,-0.0328797661,0.1201639548,-0.386284858,0.1113508493,-0.3567822576,0.2921699286,0.0997610539,-0.1011456624,0.0970676467,-0.3661126196,-0.0739550591,-0.0943037197,-0.0616693757,-0.040393576,0.0080986591,0.0266808625,-0.2271050662,-0.2285712063,-0.5969745517,-0.4918062389,0.287645936,-0.2523241043,0.0960664153,0.0432314016,-0.2876449227,0.2085760832,-0.1413347572,0.0381555855,-0.1862201393,-0.0783927143,-0.0208623353,-0.3604214489,-0.1792836934,-0.4228295088,-0.0125843398,0.1930734217,-0.0777141228,-0.0673339814,-0.2891925573,0.0329765044,-0.1065610796,0.0893678963,0.2331583202,0.1441303045,-0.2032502443,0.0735996887,-0.027181806,-0.141878292,0.2734171748,0.3981159031,0.3655802906,0.1455877423,-0.0253874883,0.2197417766,0.7495043278,0.2815073729,0.0870178863,0.1091095433,0.2936366796,-0.0391648002,0.0949851051,-0.1791398525,0.2571568191,-0.112368837,0.497356683,0.1744978875,0.1585704684,-0.2029459476,-0.2715166211,0.0198238902,-0.2972191274,-0.2140105665,0.2666811049,-0.2744171917,-0.0028797421,-0.1053560972,-0.4169648886,-0.2343729734,-0.0854017138,-0.0201849062,0.2863888144,0.0423392542,0.1747760624,-0.3722772002,-0.0249106344,-0.4501192272,0.1350299418,0.0535184555,0.3062896132,-0.1200365871,-0.1052275151,-0.0998138711,0.1517460644,0.5571809411,-0.5467593074,-0.151578933,0.084992297,0.0808283538,-0.1378734112,-0.0768334195,-0.3483757377,-0.0130607905,0.4500732422,0.3170130849,-0.4330319464,-0.3620241582,0.2260713875,-0.1992087215,-0.0786068961,-0.1529190987,-0.2000931799,-0.1855941564,-0.1270882487,0.0141043616,0.2638647556,0.2031842023,-0.094907783,0.1841097325,-0.0221267398,0.2680427432,0.2252076417,0.0337842591,-0.0832911804,0.2344718874,0.1479487121,0.0731528178,-0.0278821252,-0.380559653,0.0787953138,-0.082515724,0.1195181608,0.3015067875,0.0060753617,0.6325337887,0.0844114572,0.335981071,0.2805612087,-0.1061390936,0.0846456513,-0.2193406522,0.1391851902,0.1587159485,0.2518540621,-0.0918342173,-0.2338797748,0.5279419422,0.2838383913,-0.0503707714,0.4278495014,-0.4882173836,-0.2839681208,0.2870752215,0.0938036516,0.7551212907,-0.0216809716,0.30531165,0.115718618,-0.3540652096,0.6180198193,-0.3570949137,0.1053024679,-0.0797434151,-0.1983340234,0.0331422016,-0.0539894886,-0.1225515082,0.2684196532,-0.3492087424,0.3354416788,0.430932343,-0.264454782,-0.0450977832,0.533550024,-0.0242248792,-0.3266536593,-0.0201139599,0.0714738071,-0.3004837036,0.0691719204,-0.1442538649,-0.2498896569,0.2116397172,0.1776304692,-0.2500084937,0.0597387366,-0.035751909,0.1542449296,0.2190822363,0.020608753,0.2647165954,-0.2386105955,-0.0747130066,-0.1682815254,-0.1355244964,0.0720591694,-0.1909756958,0.0447673872,0.1156839356,-0.0988850594,0.552538693,-0.0750091746,-0.1055041254,-0.2602420747,0.2500441074,-0.4098782539,-0.2142281681,0.0200006571,0.1741642803,0.1515709758,-0.3563396633,0.5705786347,0.0601530187,0.0522032343,0.0529621504,0.1652723998,-0.2551434338,0.472382158,-0.2441768497,0.0069487276,0.0578271039,0.0097470619,-0.1020133868,-0.3523835242,-0.0696466714,0.3133563995,0.0930529386,-0.0426413454,0.0319418162,0.1327231228,-0.1125105396,0.019939404,-0.170629859,-0.3691498041,0.1408931613,0.2494584471,0.2711483836,0.1825062335,-0.1180847511,-0.2086640745,-0.1090038419,0.5103976727,0.0016730331,0.2612941265,-0.130157575,0.3655259907,0.4257893562,0.4940811098,-0.2902448177,0.0371735021,0.0374405012,0.0964160189,0.4827964008,-0.2636587322,0.0217799116,0.2817059457,-0.1314581484,0.0585723035,-0.3447951674,-0.1682955921,-0.2200633585,0.1133542731,0.1033428982,-0.0206540748,0.0722570419,-0.0430538952,-0.1420415193,-0.0688853711,0.0457496084,-0.4992572963,-0.0805315375,0.2202773839,-0.1371800005,-0.0116098402,0.0203500912,0.2394145876,-0.0081128338,-0.125860244,0.2913144827,0.133093223,0.1239340827,-0.192498982,-0.1676721871,0.0198789407,0.212551713,0.1732404083,-0.0569359958,0.1953579932,-0.1702476591,-0.1422345787,0.2402340025,0.1274124682,0.0438320115,-0.1791524142,0.1771994978,0.1416485459,-0.0975617692,0.0651970729,0.2262755483,0.084558025,0.1207726821,0.090198338,0.1219430566,0.435793519,-0.2544052899,-0.0179031305,0.1419228613,-0.3514217734,0.1340958327,-0.045421157,-0.0777515769,0.0030960084,0.1171457469,-0.0741879046,0.124276042,-0.3563074768,0.3523232937,0.2731309831,-0.0179805532,0.0604428425,0.1776670218,0.0052215424,-0.139151141,0.3545140624,0.1029014811,0.4071168602,0.0383831449,0.921475172,-0.2826140225,-0.4910770059,0.0742701665,0.4407659471,0.0139050037,-0.2088594288,0.063380979,0.0767068192,-0.3490919173,0.0910916105,0.028142143,-0.1617660671,0.6003195643,-0.0926639438,0.197003901,-0.1205104068,-0.1863871813,0.3041652441,0.2193452269,-0.0848201737,0.0386625305,0.0810324103,-0.0965912789,-0.2082840651,0.539488256,0.1378857046,-0.0306699853,-0.4813677073,0.0265417844,0.1303128153,0.1619441658,-0.0422619916,0.1696632802,0.4215089381,-0.1417666376,0.290740937,0.0038775066,-0.0947467759,-0.1848485321,0.0713041946,-0.4407307208,-0.3587898314,0.380677551,0.148929581,-0.1816970557,0.1225422323,0.2738818526,-0.6083033085,-0.2265184522,0.288765341,-0.1661256999,-0.0840118602,-0.2262104303,0.0873255357,0.050029438,0.6215308905,0.2386115938,0.3424455822,-0.3124419153,-0.1101361737,-0.5452610254,-0.1223433912,0.1547336429,-0.1268244684,-0.165378198,0.1752588004,-0.0254002772,0.3293079436,-0.2648798227,0.1895421594,0.1383681297,0.0801013112,-0.4509570301,0.1427939534,0.0801549703,0.068593964,0.2718913853,-0.109807536,-0.1711378098,0.0831257477,-0.0650783032,-0.0660624281,-0.0689276308,0.2686027884,0.1619205028,0.0313880071,0.0771801695,0.4355561435,0.0049794856,-0.206292823,0.0015175104,0.001496396,-0.3515442312,-0.1913697869,-0.0727459341,0.168337971,-0.3356745541,0.2971018255,-0.0298553109,-0.118434526,0.1227512658,0.0619151108,-0.1948868632,-0.0924559385,0.1043046191,0.151451081,0.2151798159,0.7912338376,-0.033916384,0.0512432382,-0.1006975472,-0.0103480415,0.040326871,-0.0962446481,-0.2264849246,-0.0754592568,0.0344322324,0.086966224,0.1020471081,-0.6406651139,-0.3781975508,0.0370477624,-0.0406235158,-0.0003257316,0.7422400117,0.0914560333,-0.1913268864,-0.0755398348,-0.323364377,0.0082022762,-0.2920859754,0.1692124754,-0.3407365084]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"I agree with going with temperature-scaled mixing for its flexibility!\r\n\r\nFor the function that combines the datasets, I also find `dataset.add()` okay while also considering that users may want it to be easy to combine a list of say 10 data sources in one go.\r\n\r\n`dataset.sample()` should also be good. By the looks of it, we're planning to have as main parameters: `temperature`, and `K`.\r\n\r\nOn converting the datasets to the same format, I agree that we can leave these to the users for now. But, I do imagine it'd be an awesome feature for the future to have this automatically handled, based on a chosen *approach* to formatting :smile: \r\n\r\nE.g. T5, GPT-3, decaNLP, original raw formatting, or a contributed way of formatting in text-to-text. ","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":125,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nI agree with going with temperature-scaled mixing for its flexibility!\r\n\r\nFor the function that combines the datasets, I also find `dataset.add()` okay while also considering that users may want it to be easy to combine a list of say 10 data sources in one go.\r\n\r\n`dataset.sample()` should also be good. By the looks of it, we're planning to have as main parameters: `temperature`, and `K`.\r\n\r\nOn converting the datasets to the same format, I agree that we can leave these to the users for now. But, I do imagine it'd be an awesome feature for the future to have this automatically handled, based on a chosen *approach* to formatting :smile: \r\n\r\nE.g. T5, GPT-3, decaNLP, original raw formatting, or a contributed way of formatting in text-to-text. ","embeddings":[-0.0916316584,-0.4211589098,-0.0451324433,-0.0800991431,-0.2404671758,0.059045136,0.1587893665,0.0534610786,0.3097864985,-0.1581772864,-0.233313784,0.3175208271,-0.1927586049,0.352458626,0.4201157093,-0.4721596241,0.0277890563,-0.2041643858,-0.3079895079,0.3535071611,0.0077999351,0.1143809631,-0.1774505675,0.0832355618,-0.2980297804,-0.165975526,-0.3557687104,0.1336849928,0.0949501544,0.0192527734,0.177731514,0.37621364,-0.0357892551,0.0648583695,-0.0001110554,-0.1133713573,0.2117237896,-0.0824448541,0.0959482938,-0.0414975397,0.1416421086,-0.3843331039,0.0004239143,-0.0986680612,0.1814703494,0.1664436758,0.1900739521,0.0496184193,0.3385022581,-0.2940393984,0.1220094785,0.3003106117,-0.1604250073,0.2594369352,-0.0790458918,-0.0797488689,0.0491222851,0.2236527056,0.6150226593,-0.5491260886,-0.0011972644,0.1282316595,-0.0015520197,-0.0240313634,0.1823762804,-0.0457618311,-0.3587004244,-0.243575871,-0.3750149906,0.2361418456,-0.1185867861,-0.1537864506,-0.2185504586,-0.5235368609,0.0277197435,0.0167973619,-0.281206429,0.0299735218,-0.1675621718,-0.0484484099,-0.1656024903,-0.1226045862,0.0131077683,0.1534031034,0.3461082876,0.5681383014,0.0615272112,0.1679215133,0.4003536701,0.0016655425,-0.1527252793,-0.3198023736,0.2920105159,0.0985112861,-0.3879045844,-0.3371360004,0.1489798725,-0.1886467189,0.2458421141,0.1551025957,0.106896162,0.0871020854,-0.1153277531,0.2748218179,0.3207070231,0.0658689141,-0.0684718788,-0.1193760559,0.0277469829,-0.0849121213,0.2163795233,0.124135837,0.2167029232,-0.027859414,-0.4467782676,-0.1466722786,-0.1759258509,0.1282652915,-0.1764015108,-0.5241801143,-0.0522716567,-0.1035368517,0.0894393101,-0.0489563607,-0.2832755148,0.1021507904,-0.0768533498,0.2082914561,-0.0627481714,0.074933961,-0.0794379637,0.1259627193,-0.4375811517,-0.0149753382,0.3017206192,0.1120697409,0.1171111315,0.1599151641,-0.0283310469,0.109897621,0.3795183003,-0.0167509057,0.0509770885,-0.0071642892,0.1398807615,-0.3260160089,-0.0488302857,0.1761283278,-0.3429517448,-0.1465502828,-0.1350822747,-0.2072174549,0.1426710635,0.0524654761,-0.1970576644,-0.272662878,-0.4697827399,0.7273029089,-0.0646088123,-0.0431523621,-0.1394853592,0.2776136398,-0.2422700375,-0.0958008319,0.196518451,0.0900476053,-0.0083275428,-0.3287339807,-0.0937953368,-0.110215798,-0.0325427055,0.2691782117,-0.1734347641,0.2014368176,0.0487213917,0.2012597322,0.2725915909,-0.4674472809,-0.0660237968,0.3148572743,-0.0215985775,0.3724561334,0.0471571274,0.2215044051,-0.0091188755,-0.1910979897,0.2708729804,0.9994993806,-0.3909662664,-0.0052985186,-0.1448641121,-0.3675379455,0.518184185,0.3499698043,0.0654997826,-0.3326678574,-0.1424450576,-0.0237783417,0.0809661597,0.0770776123,0.0808113813,-0.0233958308,-0.3496007323,0.02572079,-0.2160172909,-0.1317536831,-0.3982790411,-0.0131928548,-0.0113769602,0.3362922072,0.2629064918,-0.1139772683,0.2364494801,-0.2684761286,-0.1313512772,-0.3461121023,0.1502706707,-0.0612663291,0.0510793515,-0.2666864395,-0.0367573202,0.3218361139,0.1138489172,-0.1146605834,-0.5048948526,0.2798417211,-0.085126318,0.0320101827,-0.0159208123,0.6123462915,-0.1356692165,-0.0903052613,0.1830991656,0.3067480624,-0.3286168873,0.2714980543,0.3248350918,0.2580251396,0.1833746433,-0.0236838032,0.1149599329,-0.0348003581,-0.0313694552,-0.2011304349,-0.1055303887,0.4779736996,-0.1793974489,0.4034452736,-0.0279473476,0.0275263581,-0.1995432973,-0.0360027291,-0.2270117104,0.3675666749,0.301556766,-0.1931542307,0.3859142661,-0.0909542963,-0.3488397896,0.1781443655,0.2008402944,0.013799048,0.0930301994,-0.0160654299,-0.0603158176,-0.3105753362,-0.0644355863,0.4159680009,0.6467363834,0.2788435519,0.1465167403,0.0539188795,-0.2370414436,-0.0721770376,0.0270975158,0.1797160357,-0.0220576562,0.1321817189,0.0289726704,0.0536681227,-0.019797774,-0.1459692419,0.1391361654,-0.1829010546,-0.0267881677,0.035044726,-0.1900615245,-0.5519416332,-0.2311933339,0.0460517295,-0.3759756088,-0.304053694,0.1278389543,0.0250950344,-0.1918633431,0.1835840642,0.2177855372,0.5667126775,-0.4283856452,-0.25661695,-0.1000010669,-0.2585214078,-0.1005810127,0.1240393892,0.4180926085,0.3008857667,0.4967921376,0.2893135846,-0.0144694103,0.0208391342,-0.3807587922,0.0476373062,-0.2881382108,0.2660784721,0.1090801656,-0.2344630659,0.1544487625,-0.4112606645,-0.0272319969,-0.0677763969,-0.0513928868,-0.0491179861,-0.0151990345,0.0207087733,-0.1768659204,-0.2257882506,-0.6239804029,-0.4909668565,0.2893363833,-0.2813093662,0.0613841563,0.0165350251,-0.2601398528,0.1938994825,-0.1552231312,0.0375991315,-0.1745573282,-0.1040851027,0.0465736277,-0.3616591692,-0.1606725901,-0.4154684544,-0.1143826991,0.2324681729,-0.0850747824,-0.0425746366,-0.2387888134,0.0480952933,-0.1275751442,0.0941575468,0.222891286,0.1751593202,-0.22287108,0.0990753919,-0.0367767476,-0.2646290958,0.1948918253,0.3380730152,0.427256763,0.177809909,0.0165802669,0.1727931052,0.7245492339,0.2591904402,0.0557640009,0.150517121,0.3070631325,0.0335696489,0.146548003,-0.2337548584,0.2416428179,-0.1647280008,0.5056424737,0.098433651,0.1719013005,-0.2043492794,-0.2838809192,-0.012782204,-0.257922262,-0.1879765093,0.2683837712,-0.2222358435,0.0202785283,-0.055379007,-0.372433722,-0.2528823912,-0.0427951477,0.0179480109,0.3158903122,-0.0309924372,0.152218461,-0.345153749,-0.041200038,-0.4830800295,0.1510534883,0.0881519243,0.2757006586,-0.0816482231,-0.0731475726,-0.1406987011,0.1045371816,0.5943595767,-0.5221490264,-0.1656587869,0.0909055322,0.0804265812,-0.0901894197,-0.087753959,-0.2720246017,-0.0436302833,0.44831568,0.3695802093,-0.3692051768,-0.3136214614,0.2565201223,-0.1805986315,-0.0316669755,-0.1990140676,-0.1823862493,-0.1892318726,-0.142061606,0.0321542174,0.1843507141,0.2113642097,-0.1040511876,0.17735219,-0.0519658141,0.2529607713,0.2023988366,0.013077395,-0.0403005667,0.2269546092,0.1543975025,0.0318597741,-0.1210527271,-0.2936052382,0.0963308737,-0.0699611753,0.1100494042,0.2714964747,-0.0030550973,0.6472851038,0.0571767502,0.3105741739,0.2889769971,-0.125620544,0.0590077527,-0.2301298082,0.1417718083,0.2295093685,0.2583223879,-0.1393097639,-0.1966267228,0.593556881,0.316298753,-0.0732627213,0.3897017837,-0.4620025754,-0.2712930739,0.2523411512,0.0279185567,0.6917433739,-0.0259476528,0.26716277,0.0650329739,-0.3845382631,0.6288003325,-0.3909120858,0.1575741321,-0.0777261481,-0.2354030609,0.0326018445,-0.0945741609,-0.0697192401,0.2683088779,-0.3717157841,0.3496938944,0.4773365557,-0.204410255,-0.0511797406,0.5518017411,-0.0291925482,-0.4002793431,-0.1072373167,0.0890614018,-0.281213969,0.1060334593,-0.1633670032,-0.2763769329,0.2486427277,0.1135004014,-0.2740924358,0.1133617833,-0.035165146,0.1855692416,0.2648394108,0.0163140129,0.2512059808,-0.1788924783,-0.0666228384,-0.1549472064,-0.0886360034,0.1312909424,-0.1466120631,0.0502584949,0.0305555034,-0.204527691,0.5105494261,-0.0761516392,-0.1078909263,-0.2437790781,0.2447650433,-0.4367167056,-0.1542506665,0.1101066917,0.1682256907,0.2132159919,-0.3162401617,0.6001377106,0.0864137039,-0.0095839836,0.0558025725,0.1773011386,-0.2368263304,0.4851781726,-0.2578954399,-0.0046098372,0.0524401926,0.0108783757,-0.1622541994,-0.3727542758,-0.0335331783,0.298137635,0.0863665044,-0.0161652137,-0.0029815889,0.1021040455,-0.079756245,-0.0030077493,-0.1554479748,-0.3322068155,0.1165377051,0.1472442746,0.1947061867,0.1670138091,-0.0539530776,-0.2400145233,-0.1859924048,0.4692339003,0.0406197235,0.2234195024,-0.1361565143,0.3405199647,0.4290604293,0.5344376564,-0.2816098034,0.0721736476,0.0327903964,0.0864349157,0.4831210077,-0.2349261492,0.0746664926,0.2786789238,-0.1192489937,0.0491989627,-0.3402354121,-0.1727287769,-0.2547361553,0.0983532071,0.1122342423,-0.0018389161,0.0054131085,-0.0612734333,-0.0791018158,-0.0815896913,0.0165715721,-0.38943097,-0.0693220422,0.1556466967,-0.1154112369,0.0159976929,0.0592389554,0.2030221522,0.0097939344,-0.0524284244,0.2911643684,0.127098605,0.1111847535,-0.1909795254,-0.1763176173,0.0046915212,0.2138266414,0.1481524557,-0.0290785786,0.1604177952,-0.2343684882,-0.1400951594,0.234292388,0.1486858279,0.020620225,-0.1204305962,0.105655387,0.1406226754,-0.1248989627,-0.0285314042,0.1781450063,0.1108474508,0.1366314441,0.1105991751,0.1587068141,0.449914515,-0.210016042,0.0259732921,0.1152657196,-0.3685956895,0.1291753948,-0.1154743135,-0.040456254,0.0053082923,0.147482872,-0.0111293895,0.1395928562,-0.3216474056,0.3060735166,0.3096959591,-0.0237122644,0.0586463772,0.2391148359,0.0148077691,-0.1331096292,0.3046792448,0.0185764674,0.3192404211,0.0427732952,0.8754268885,-0.2722271085,-0.4550594985,0.1064825729,0.4365846217,0.0321108513,-0.2155284733,0.15521653,0.0911859348,-0.4059762061,0.1209785268,0.0581079945,-0.1637461632,0.6574727893,-0.116494514,0.1646675915,-0.1581150293,-0.1499401033,0.3523963392,0.1967354864,-0.0954383537,0.010715507,0.1199256778,-0.1106929258,-0.169852525,0.5139836073,0.2285721749,0.0347937122,-0.4124754965,0.0373516828,0.107136108,0.1429030895,-0.0732118785,0.1380367279,0.4449816048,-0.1812671572,0.2542766929,0.0367873386,-0.0788206309,-0.1354016066,0.0433842465,-0.3884741664,-0.3973376751,0.4565663338,0.1096775159,-0.1101013571,0.0452650897,0.320700109,-0.6264185905,-0.2374823391,0.3614814579,-0.1362901181,-0.0500934198,-0.2448353022,0.0871589556,0.0840754956,0.6122642756,0.2146092355,0.3534938991,-0.4017143548,-0.1403850764,-0.5742545128,-0.2160328031,0.1719411463,-0.0766166449,-0.2584018409,0.1467674375,-0.0685920343,0.2895057499,-0.214011386,0.117865026,0.1777322888,0.0270575713,-0.430539459,0.115783602,0.0587909929,0.004421941,0.319761157,-0.1105256677,-0.1742179841,0.0824259073,-0.0569639914,-0.0908900946,-0.0679281205,0.2378151268,0.1677979827,0.0964897498,0.0813624263,0.4660520256,-0.0493248887,-0.2016543746,-0.0087855943,0.0125133656,-0.3770980239,-0.1661677361,-0.0771840215,0.1801460832,-0.3348351419,0.2549709678,-0.0166341439,-0.0848235711,0.1192190945,0.0184182711,-0.189348802,-0.0494577847,0.0742050558,0.1950691491,0.2228058577,0.7319662571,-0.0105619999,0.0857545808,-0.0929860324,-0.0342138484,0.0708658174,-0.0521122403,-0.2077330351,-0.1361374259,0.0242378227,0.0972727686,0.1356813312,-0.6967553496,-0.3358291388,0.0143320551,-0.0228635501,0.0317773297,0.7183235288,0.0666534528,-0.1448654681,-0.1175315231,-0.298699826,0.0534241386,-0.2820818424,0.1892061383,-0.3477088809]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"This is an interesting discussion indeed and it would be nice to make multi-task easier.\r\n\r\nProbably the best would be to have a new type of dataset especially designed for that in order to easily combine and sample from the multiple datasets.\r\n\r\nThis way we could probably handle the combination of datasets with differing schemas as well (unlike T5).","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":59,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nThis is an interesting discussion indeed and it would be nice to make multi-task easier.\r\n\r\nProbably the best would be to have a new type of dataset especially designed for that in order to easily combine and sample from the multiple datasets.\r\n\r\nThis way we could probably handle the combination of datasets with differing schemas as well (unlike T5).","embeddings":[-0.0776049644,-0.4301363528,-0.0512811989,-0.0076487889,-0.2415291816,0.1148416847,0.1743364334,0.0923100263,0.2852181494,-0.1179621741,-0.2437363416,0.2735090852,-0.1946441382,0.4169626534,0.3981999755,-0.4560344517,0.0079969484,-0.2196667641,-0.2606224716,0.3799697757,0.0418277532,0.095559299,-0.1473453641,0.0627828836,-0.2902197242,-0.1371725649,-0.3435868025,0.0805074498,0.1044687256,0.0340153202,0.1990051121,0.3737783432,0.0089969868,0.1089183465,-0.0001125286,-0.0909401104,0.1605088562,-0.0712215528,0.1302847117,-0.0230519362,0.1236006096,-0.3596464097,0.0639435723,-0.1071627513,0.1815183014,0.186756596,0.2017910182,0.0839319006,0.3181020319,-0.3155004382,0.0999815166,0.290384084,-0.1883252263,0.2807796896,-0.0540621616,-0.0642039403,0.0306815002,0.228564024,0.6673142314,-0.5304598212,0.044611454,0.0683927238,-0.0271339901,0.0415637158,0.117888324,-0.0851886496,-0.3070947528,-0.2394675463,-0.3735044599,0.2802624702,-0.0719252825,-0.1064395383,-0.2332081348,-0.5333688259,0.0157342274,0.0769624412,-0.295244813,0.0596089661,-0.1627973914,-0.0191707928,-0.1291299015,-0.1218795106,-0.0014559383,0.1736190617,0.3086414337,0.5555925369,0.0679185465,0.166067943,0.3354541063,0.0697011054,-0.0984865054,-0.2893274128,0.283675164,0.0658083037,-0.4028246999,-0.3567550182,0.1301074028,-0.26543203,0.2804774344,0.1396420002,0.1625594497,0.0956441611,-0.1668536514,0.251871109,0.3397616148,-0.0224791672,-0.1228724495,-0.1360303015,0.0175045673,-0.1088773608,0.1665438712,0.136100173,0.1756422222,-0.0186628886,-0.4268658459,-0.1198429242,-0.1464762688,0.0982045382,-0.1879583001,-0.603138864,-0.0907536671,-0.1077904478,0.0310064703,-0.1576607078,-0.2877174616,0.0926540196,-0.1034632698,0.243978709,-0.1317925006,0.0085406294,-0.0485766232,0.065514192,-0.4239288867,-0.0484915897,0.2941403091,0.0993547812,0.0483110882,0.1649857163,-0.0342779905,0.1286848038,0.4205949903,0.0290484969,0.0141222924,-0.0367381349,0.1730397344,-0.3192974329,-0.0118671656,0.2226764262,-0.3717390895,-0.1183751225,-0.1698476076,-0.2173647434,0.0976075679,0.0334337987,-0.2183186114,-0.3184990585,-0.4913890362,0.712941885,-0.0447418466,-0.0464624129,-0.1646772921,0.266639322,-0.232261166,-0.0979923159,0.1632626653,0.0849712193,0.0131495027,-0.2949581742,-0.0931307301,-0.1038354039,-0.0648113415,0.3017258048,-0.1689353883,0.2122735828,0.0658584759,0.1369742304,0.249456495,-0.4536504745,-0.0488989614,0.3324487507,-0.0249690302,0.357055068,0.0768328607,0.2123102993,-0.0030692071,-0.2040670812,0.1949262917,1.0177958012,-0.3952714205,-0.0217342824,-0.1102285758,-0.3906148076,0.5425784588,0.3610700071,0.067350097,-0.3310129941,-0.1318062097,-0.0087063769,0.0908718035,0.0797558352,0.0483210608,-0.0628936738,-0.3434605598,0.0450412817,-0.2458027899,-0.1973199546,-0.4192028344,-0.0202788245,0.0169575177,0.2923036218,0.3340206146,-0.1000260711,0.2713528872,-0.3039948642,-0.1537616104,-0.3133296371,0.119758606,-0.0056195688,0.0379966423,-0.2483401,-0.0306746587,0.3191803098,0.0992679,-0.1351188719,-0.4337623417,0.3398182094,-0.047643017,0.0467606746,-0.0463985913,0.620665431,-0.1988634318,-0.0915856659,0.1701807678,0.2487031817,-0.3234956563,0.256033361,0.3241401017,0.3500341177,0.1919656843,-0.0883700922,0.0861881971,-0.0120608928,-0.0577380769,-0.1945427656,-0.0840635747,0.4658958018,-0.1854711324,0.4102387726,-0.0535622612,0.0147549119,-0.1941016763,-0.0550692603,-0.1838123798,0.4107014239,0.3079739511,-0.197261408,0.3365919888,-0.1108615696,-0.363048315,0.196744144,0.185142979,0.0567776822,0.0761518553,-0.0324886292,-0.0191662144,-0.2904635668,-0.1110854819,0.4045722485,0.632398963,0.2805095613,0.1170600653,-0.021244524,-0.2500744164,-0.1028238013,0.011356975,0.2060652524,-0.0168435108,0.150283888,0.0468926094,0.0736096278,0.043971926,-0.1459352523,0.193490237,-0.1794005781,-0.0495684855,0.0246106274,-0.1928532273,-0.5652402639,-0.2191136479,0.0437883101,-0.3279028237,-0.3013747036,0.1503418088,0.0533402786,-0.1769550741,0.1552890986,0.1746625155,0.5224668384,-0.4443944097,-0.2338437885,-0.0641313642,-0.2828474045,-0.0807362571,0.1160765886,0.4930141866,0.2858503759,0.4589028656,0.2643532157,-0.045532275,0.0037957064,-0.3939478099,0.0290108677,-0.3069182634,0.2791732252,0.0830265433,-0.2171403021,0.1505553871,-0.394996196,-0.0644386038,-0.0488226414,-0.0779298097,-0.0286925919,-0.0250476599,0.0897226259,-0.1780427694,-0.2269402891,-0.6084065437,-0.4707541168,0.2826776206,-0.2821086943,0.0727403238,0.0387645252,-0.277934283,0.1508194059,-0.1559341252,0.0280285962,-0.1545214653,-0.1129903123,-0.0186402239,-0.3668550253,-0.1596241295,-0.4123146832,-0.0737837255,0.2263587266,-0.0729172528,-0.0498354658,-0.2201999873,0.0606832951,-0.1032453701,0.0564823523,0.1742459983,0.1718525589,-0.2245576829,0.1106522009,-0.0412322246,-0.2030341327,0.2085092813,0.3692530394,0.3985555768,0.1724753976,-0.0453424975,0.1586776972,0.698872447,0.225701347,0.0746953636,0.1957734674,0.300358057,-0.0084648076,0.1412200034,-0.2332861125,0.2162008733,-0.1369470507,0.4953300953,0.1153532192,0.1564605981,-0.1784892082,-0.2667302787,0.0535668172,-0.2868153453,-0.1803445965,0.2325089574,-0.2959995866,0.0018161524,-0.0648912415,-0.4627397954,-0.2831377089,-0.0490763448,-0.0285924859,0.3023614287,-0.0183279533,0.1623066217,-0.3587013185,-0.0239684246,-0.3594931662,0.1668869257,0.0956119895,0.2567149103,-0.081652984,-0.0452852473,-0.0754171461,0.1324962676,0.5941112638,-0.5726919174,-0.1146803871,0.1122617573,0.0694399402,-0.1070293933,-0.1124659106,-0.2531117499,-0.0227783881,0.4108986259,0.3724685609,-0.3512171209,-0.3359476924,0.2770609558,-0.1780257523,-0.0666031092,-0.1773483753,-0.1805448383,-0.1992293745,-0.1117431819,-0.0150128482,0.2250314653,0.2360183746,-0.0939791277,0.1691284925,-0.119028531,0.2343086004,0.1777767241,-0.0200170428,-0.086044088,0.2196727842,0.1901293099,0.043740306,-0.1204652563,-0.2854855955,0.0902776346,-0.085317947,0.0410683453,0.323145777,-0.0169591587,0.654355824,0.0035472142,0.3540613055,0.2970191538,-0.1518308371,0.0351194777,-0.2252553552,0.1592186242,0.1932839155,0.2645814419,-0.1670916975,-0.2223951668,0.5615090132,0.3239040375,-0.0417188406,0.3951153159,-0.5345015526,-0.274638921,0.2725146115,0.0293014757,0.6709386706,-0.0856773406,0.2816002965,0.012769972,-0.4185321629,0.6093620062,-0.3563193977,0.1455237716,-0.0799400359,-0.231561482,0.0210898537,-0.0719360411,-0.0998954326,0.2743530571,-0.3496188521,0.301412344,0.4439024031,-0.1805108041,-0.0398728661,0.5531432629,-0.0153129883,-0.3570756912,-0.038609013,0.066703856,-0.2900906801,0.0710930005,-0.1547671705,-0.2515334487,0.2441351265,0.1927368939,-0.3443206549,0.0698324367,0.0189762414,0.1855354905,0.2244637012,-0.0021207256,0.296608299,-0.1636162251,-0.0310924593,-0.1591515243,-0.1119893417,0.1209476441,-0.1884543002,0.0049204477,0.0426330753,-0.1961916834,0.5082734227,-0.0502668843,-0.1739011407,-0.2432445735,0.2771665752,-0.462369293,-0.2343650013,0.1224643141,0.1940979958,0.2181558907,-0.3146968186,0.5614804029,0.127996698,-0.0110748243,0.044554349,0.1349021196,-0.208381325,0.4753073752,-0.2853443921,-0.0258624703,0.066884242,-0.004448039,-0.0967397168,-0.3717646897,-0.0446483381,0.3177084029,0.1656932533,-0.0284103751,0.0740085319,0.138213858,-0.0919289738,-0.0074774642,-0.1643103659,-0.2981975675,0.1717769802,0.1826200634,0.2225332707,0.182265535,-0.0715930089,-0.2359989434,-0.1222001463,0.4769056439,-0.0089862617,0.2460337579,-0.1398747414,0.3845985532,0.4304974675,0.5406701565,-0.2570672929,0.1253073364,0.043342445,0.081369929,0.5019494891,-0.2255442441,0.0380042903,0.2741829753,-0.1421247274,0.0417179503,-0.2888581157,-0.1685370058,-0.216394037,0.1169298664,0.1269612908,0.0241434257,0.0317429192,-0.0548917241,-0.1292057782,-0.0581377707,-0.0107884696,-0.378121227,-0.0466113202,0.1347695887,-0.0621498823,0.0481393076,0.0455538332,0.2488150895,-0.0054837,-0.0489119031,0.3026914895,0.1137172356,0.1313488483,-0.1924309433,-0.2258507311,-0.0227676723,0.2781153023,0.1600535661,-0.0804908797,0.1509431005,-0.2103306949,-0.1457174122,0.1973721385,0.1054654121,0.0688102394,-0.0883574635,0.146637693,0.1075832099,-0.0755475909,0.0134025756,0.1942141652,0.0746555403,0.0725068823,0.1083762646,0.1365597397,0.4735893607,-0.2260929197,0.0031990437,0.1232500225,-0.3594174683,0.1541866064,-0.087498948,-0.066926606,-0.0310226046,0.1314771026,-0.0590046942,0.1314285398,-0.3177983761,0.3143518269,0.2902783155,-0.0166152418,-0.0238690805,0.244492963,0.038166903,-0.1479455382,0.297565639,0.042414099,0.3625642359,0.031823352,0.8982050419,-0.2870637774,-0.4490221143,0.085068509,0.4442305267,0.0931392089,-0.2184083909,0.1227295324,0.1058463529,-0.3336721361,0.1266605705,0.027429739,-0.1679346412,0.6969407797,-0.088710241,0.139469862,-0.1556522995,-0.1600604355,0.3791301847,0.1530621052,-0.0872004032,0.0051243785,0.1180316359,-0.1149556711,-0.1713051945,0.5400868058,0.1966331601,-0.0169869866,-0.4411090612,-0.0029190739,0.0928225741,0.1858453006,-0.0031086458,0.1166482195,0.427552253,-0.1776586324,0.275704056,0.0165012497,-0.0424515344,-0.1272155493,0.0419438854,-0.3996353447,-0.4405324161,0.4334970713,0.1266337186,-0.101330556,0.1223909333,0.3422189653,-0.6254710555,-0.2166346461,0.3283931911,-0.1351541132,-0.0997852236,-0.2127865702,0.0829946622,0.0520626642,0.6268000007,0.2669469416,0.3559404612,-0.3678622842,-0.100773707,-0.5913131833,-0.1538396925,0.1531481892,-0.0713014677,-0.2285646647,0.1566448212,-0.0457349084,0.2795037925,-0.1662358791,0.1557860523,0.2200660855,0.0632857531,-0.4629289508,0.0581695586,0.1567252725,0.0008711959,0.2697931528,-0.095621407,-0.1836102605,0.1072154418,-0.0865937248,-0.1330373734,-0.0830835402,0.2785076797,0.1443428844,0.1257952452,0.0605219081,0.4831179082,-0.0133496625,-0.1953569353,0.0003230957,-0.0172191616,-0.3659597635,-0.1334766448,-0.1121910363,0.1616244018,-0.4069583118,0.3112419248,0.0202682968,-0.1174988747,0.0788925067,0.082060352,-0.1986315101,-0.1077199876,0.092512764,0.1740233302,0.229268074,0.706404388,-0.0096554002,0.0815721601,-0.0816268921,-0.0246424228,0.0138051696,0.0019771534,-0.2365907282,-0.0972196683,0.0359875485,0.0470708609,0.1454031765,-0.6879176497,-0.3390018642,0.0712452456,0.0081003979,0.0602682233,0.7683100104,0.1340706497,-0.1433209777,-0.1091944128,-0.3065088987,0.0288375933,-0.2758437693,0.1856127679,-0.3361102045]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"@thomwolf Are you suggesting making a wrapper class which can take existing datasets as arguments and do all the required sampling\/combining, to present the same interface as a normal dataset?\r\n\r\nThat doesn't seem too complicated to implement.\r\n","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":37,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\n@thomwolf Are you suggesting making a wrapper class which can take existing datasets as arguments and do all the required sampling\/combining, to present the same interface as a normal dataset?\r\n\r\nThat doesn't seem too complicated to implement.\r\n","embeddings":[-0.0898510367,-0.3969319463,-0.0316652842,-0.0373525135,-0.2264484614,0.0873821825,0.1969549954,0.066023387,0.3327300847,-0.13962093,-0.2261487246,0.3318240941,-0.2025055885,0.3645176291,0.4015562236,-0.501919508,-0.0090313414,-0.1732093245,-0.3404358923,0.3434834182,0.0159044527,0.0897598043,-0.1299690306,0.0539926477,-0.2659037709,-0.1480943859,-0.3480380177,0.1039842591,0.0902505815,0.0002097426,0.26544559,0.3834895492,-0.0041850111,0.1146860421,-0.000110999,-0.1040945202,0.1779927462,-0.061887417,0.1340163648,-0.0227332041,0.2059171796,-0.374109447,0.04129548,-0.1429956704,0.0951040164,0.186511755,0.1486477405,0.0613156334,0.4389664233,-0.3271816373,0.113716282,0.2715312541,-0.2127840817,0.2160600722,-0.103535369,-0.0817166641,0.0361859128,0.2948392332,0.6407872438,-0.5255509615,-0.0261882171,0.0954857841,-0.0273622703,-0.0269174017,0.1653091908,-0.0557940342,-0.3531101644,-0.2346815318,-0.3933763802,0.261813283,-0.1963312328,-0.1110707745,-0.2282507718,-0.5546969771,0.0432708822,0.0565065481,-0.3279976845,0.0498439819,-0.1630619168,0.0150655881,-0.1603159606,-0.1137707978,-0.0351350382,0.1509450972,0.3342216611,0.5391833186,0.0717265755,0.1810311377,0.3762328923,0.0664318129,-0.1559690833,-0.2558469474,0.2604014874,0.1465934813,-0.3001440167,-0.3815107048,0.1029030308,-0.2020753026,0.2648243606,0.1646739095,0.0934806019,0.126980558,-0.1174718142,0.2820120156,0.3259015679,-0.0005844994,-0.0584376119,-0.0893254504,0.0330901705,-0.0813911855,0.1632658243,0.1214111671,0.1739641428,0.0617650561,-0.4309351146,-0.1242405325,-0.2083215266,0.1426919699,-0.1268387288,-0.5166335106,-0.0674271211,-0.0839091614,0.0690300688,-0.0892471224,-0.3010070622,0.1144687757,-0.0229355283,0.2146689892,-0.0696992055,0.0141832316,-0.0549568981,0.1181560084,-0.4216566086,-0.0169754606,0.3122235835,0.0995567366,0.0971219018,0.1739335507,-0.0777562186,0.1229841337,0.4227028489,0.0299971607,0.0288171247,-0.0377979577,0.14484559,-0.3256048262,-0.006403306,0.1831966639,-0.3624858558,-0.1541394442,-0.0966254026,-0.165102303,0.1237905324,0.0594376586,-0.2251965404,-0.2789829969,-0.3781819642,0.7292924523,-0.0695030987,-0.0614325218,-0.1721529067,0.2976944447,-0.2514333725,-0.0713770092,0.1812423468,0.0710493848,-0.0328213945,-0.3899051249,-0.1003962681,-0.1081335098,-0.0720615834,0.2687501907,-0.2001522928,0.2165666968,0.0400445387,0.1675095111,0.2588683963,-0.4050790668,-0.1015974209,0.2870190144,-0.0375538357,0.3858802915,0.0768328905,0.2454753518,0.0891938135,-0.1763921827,0.252664268,1.0071948767,-0.3976961672,-0.0693167076,-0.1097234637,-0.3863373101,0.5040088296,0.3668292761,0.1083000153,-0.2368410081,-0.1491304785,-0.0592130013,0.0653444901,0.045004636,0.1062985361,-0.0819681883,-0.3427742124,0.0306715593,-0.216996491,-0.2022387087,-0.3672421277,-0.0022583967,-0.065089874,0.3331253231,0.3727762103,-0.1935801804,0.1813594997,-0.2651145458,-0.0785318315,-0.2893358469,0.1399663538,-0.072068423,0.0579696819,-0.2934385836,-0.0477033295,0.32401824,0.0663003772,-0.0903619602,-0.4725524187,0.3164875507,-0.0660732836,0.0410398617,-0.0641026646,0.6213018298,-0.1682782322,-0.1320482939,0.2150791287,0.2251460403,-0.3285654187,0.2118606567,0.3367442787,0.2405318171,0.1710718274,-0.0543709733,0.1039706469,0.0191598162,-0.0123938769,-0.2296965718,-0.1085937917,0.5694272518,-0.0680288002,0.4326314628,-0.0260058511,0.0427017175,-0.1831449568,-0.1110736579,-0.2428102493,0.3811077178,0.2990497053,-0.2120331079,0.352360338,-0.0513427258,-0.3553395569,0.1363967806,0.13285321,0.0007291199,0.0813225135,0.0181137118,-0.0639313385,-0.2471843958,-0.0975961983,0.4116670489,0.5995679498,0.2978718877,0.1413642615,0.0798594058,-0.2595912516,-0.075013943,0.0389111228,0.1910626441,-0.0794535726,0.154210344,0.0102188662,0.0571128428,-0.0155991139,-0.0966532603,0.1261167377,-0.2244980037,-0.0257465728,0.0404702388,-0.1504003704,-0.5861170292,-0.231887877,0.0051614125,-0.3123074174,-0.2645249665,0.1403571963,0.0393285602,-0.1784060895,0.1565387398,0.1852187216,0.6711630821,-0.4588549435,-0.2282123715,-0.0570804998,-0.2689540684,-0.1103215367,0.1257170886,0.4261064231,0.2776331007,0.5188305974,0.3089579046,-0.0348959155,0.0689550042,-0.3962318301,0.0578260273,-0.2660859823,0.3105842471,0.1134503111,-0.2200812101,0.1703180522,-0.3742897809,0.0022690678,-0.1026544869,-0.0340799056,-0.0740579218,-0.0439088307,0.0822699293,-0.2027610391,-0.1946837455,-0.6382994056,-0.4739052951,0.30521366,-0.2133403718,0.0697752684,0.0194587391,-0.2589373291,0.178520143,-0.1391531825,-0.013688419,-0.194584623,-0.1289618611,0.0167612638,-0.3375965953,-0.1641010046,-0.4051863253,-0.2008609921,0.2261935025,-0.090013206,-0.0637000054,-0.2524306774,0.0771816447,-0.1279675812,0.1112925783,0.2199829966,0.1511761546,-0.2682071924,0.0936457291,-0.0291541815,-0.2080545127,0.2571524084,0.2949794829,0.3784078658,0.1561033577,0.0382041782,0.1333256662,0.706382513,0.2446941286,0.0319510251,0.1457705498,0.3078972995,0.0172534864,0.1577182859,-0.2260904163,0.1919377297,-0.1703614742,0.4489403367,0.1061675176,0.1941896826,-0.2040517181,-0.2649407089,0.0500605702,-0.2274236381,-0.230343312,0.2668191791,-0.1887025386,0.0117208976,-0.0904016793,-0.3691735268,-0.246988222,-0.0226006731,0.016349107,0.2658740878,-0.0231956355,0.125591293,-0.4113302231,-0.0363422185,-0.4307916164,0.182260111,0.0788303688,0.2600061893,-0.0604928918,-0.0711534843,-0.0895100757,0.1006803215,0.5856122375,-0.5367433429,-0.1742345095,0.1198838875,0.0764022693,-0.1405958384,-0.0693980604,-0.2734323144,-0.0443898365,0.4505381286,0.3297436535,-0.3563486338,-0.2867946923,0.2995759845,-0.1722754985,-0.0965834185,-0.2066871077,-0.1536158174,-0.1994932294,-0.1387160867,0.0547180399,0.1832149327,0.159693554,-0.1042321622,0.1223096699,-0.1110259369,0.2156316936,0.1978040934,0.0091421474,-0.0450818203,0.2142856866,0.1789846271,0.0323690884,-0.1414233893,-0.341366291,0.1296182871,-0.0628699958,0.1292474121,0.2377251536,-0.0153231649,0.6900310516,0.042732574,0.3528766036,0.3185285926,-0.1063950956,0.0094634369,-0.3053716719,0.1309756339,0.1652957052,0.2444354147,-0.1179362014,-0.2283634394,0.5872105956,0.329706043,-0.0860994235,0.3757959306,-0.4737700224,-0.310490042,0.3206133842,0.0482919142,0.7352602482,-0.0201501846,0.2903295159,0.0600588769,-0.3805786073,0.6396003962,-0.3495909572,0.1266001761,-0.1399344504,-0.2060180604,0.0089666946,-0.0815130919,-0.0617382228,0.2745797932,-0.360897392,0.3065805435,0.3756437004,-0.2247984707,-0.0269378833,0.5579151511,-0.0284860507,-0.4020187259,-0.0726894215,0.101701729,-0.2750024199,0.0695490092,-0.166645363,-0.2324197143,0.2085057944,0.1353928745,-0.2649182081,0.1090129167,-0.0219965503,0.156533286,0.2781090438,0.0218036063,0.2740699351,-0.2461938858,-0.0806766599,-0.1060426086,-0.088235274,0.1340869367,-0.2227403671,0.0591329485,0.07656461,-0.166191861,0.5294392109,-0.0293302741,-0.1059564203,-0.2020442933,0.2376759946,-0.4289616644,-0.1612630635,0.1050783768,0.1949498206,0.122625038,-0.3510895669,0.5494362116,0.0929958224,-0.030742852,0.0594922341,0.1608476788,-0.2850424051,0.466938287,-0.2044122815,0.0145053724,0.0667580739,-0.0349689201,-0.0981705487,-0.363201052,-0.0358230956,0.2880769372,0.1427470446,-0.0450333729,0.0082563851,0.1737511754,-0.0649225786,-0.0009055374,-0.1085997075,-0.3678089678,0.1381730288,0.1550519317,0.2818160057,0.2077013254,-0.1084916666,-0.2485352606,-0.2284657359,0.4540192485,0.0856426582,0.2493200749,-0.1309677362,0.3653291166,0.4037597179,0.5555805564,-0.2882810831,0.0474740602,0.0300739519,0.0926196873,0.4883191288,-0.2407028973,0.0004642997,0.2718991935,-0.1266115308,0.0699016377,-0.3354583085,-0.1801497638,-0.2789697349,0.1032844931,0.1215907708,0.0043789451,0.0194344781,-0.0251791291,-0.0932710767,-0.0762114823,-0.032402657,-0.4200459719,-0.0596199296,0.1347333491,-0.0789921582,0.0457461812,-0.0042994553,0.2168247849,-0.0051984838,-0.0450629033,0.27724576,0.0874798149,0.1442192495,-0.1765516102,-0.1422996521,0.0061336108,0.3090966046,0.1041073278,-0.0281759538,0.1752347201,-0.2041564137,-0.0734726712,0.2641974092,0.1913737208,0.0282777436,-0.0657374561,0.1219795123,0.1541512758,-0.1338307112,0.0374211706,0.2465266883,0.0528054312,0.1256937683,0.1320372075,0.1526354998,0.4616318941,-0.2277929634,-0.0095821144,0.1340144277,-0.3001977503,0.1663933396,-0.1424223185,-0.0036700214,-0.0188433416,0.1448399872,-0.0304602124,0.1771894842,-0.3630974591,0.2931548655,0.3141860664,-0.0120241046,0.0497263893,0.2157157958,0.0113860648,-0.1301043034,0.3093082905,0.0953726768,0.3369999528,0.0031129885,0.8959380388,-0.3085143566,-0.4987756312,0.0820809901,0.4227997661,0.0348251201,-0.199343279,0.0823081583,0.1272723526,-0.3698183596,0.0759746209,0.0784344971,-0.1865928024,0.6474289894,-0.1167951375,0.1635962278,-0.1702311337,-0.204961881,0.342982024,0.1596621871,-0.1373147815,0.0277027432,0.1110748425,-0.0684817433,-0.1626202017,0.5162655711,0.1743649244,0.0161034539,-0.4487465918,0.0412261747,0.0500359312,0.1521725208,-0.0871050283,0.1192083284,0.4416899681,-0.1661205888,0.2409606427,0.0384051576,-0.1061391607,-0.1549689621,0.0149721866,-0.3957521617,-0.4279153347,0.4407235086,0.1213496774,-0.1510049403,0.0182202365,0.2517423332,-0.6123271585,-0.227174744,0.3545630872,-0.1733734757,-0.0767516047,-0.2256261408,0.0838294625,0.0379488952,0.611551106,0.2173800021,0.3595839143,-0.3842931092,-0.1715813875,-0.5617926121,-0.186886549,0.1403197199,-0.1046297774,-0.2472453713,0.1625194103,-0.079118222,0.296741128,-0.2263427973,0.1232836023,0.222853452,-0.0245985743,-0.4108704031,0.1212959215,0.1244953945,0.0211477578,0.3134693503,-0.0808979124,-0.2103290111,0.0732444897,-0.0707960054,-0.0356407836,-0.0873444751,0.2631873786,0.1474715918,0.1126745939,0.0629213378,0.5086533427,-0.0262721963,-0.1629562825,0.0201170016,-0.0493787527,-0.3885860443,-0.1232485399,-0.0805888474,0.2154360712,-0.3562215269,0.2585203648,0.0042474316,-0.1549849063,0.1442205459,0.0326070189,-0.256231457,-0.0582585894,-0.0123900333,0.2157614678,0.2192552835,0.7426388264,0.027954489,0.0870670378,-0.0391521938,-0.0098251598,0.087929219,-0.040989805,-0.2214282751,-0.1632405221,0.0109186545,0.0154421031,0.0981712043,-0.6646654606,-0.3680722415,0.0423703566,-0.007606118,0.0981842428,0.7144812942,0.0821470544,-0.1468365192,-0.145465672,-0.2757920027,0.0348375551,-0.245181784,0.1932799667,-0.3910035491]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"I guess we're looking at the end user writing something like:\r\n``` python\r\nds = nlp.load_dataset('multitask-t5',datasets=[\"squad\",\"cnn_dm\",...], k=1000, t=2.0)\r\n```\r\nUsing the t5 method of combining here (or this could be a function passed in as an arg) \r\n\r\nPassing kwargs to each 'sub-dataset' might become tricky.","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":45,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nI guess we're looking at the end user writing something like:\r\n``` python\r\nds = nlp.load_dataset('multitask-t5',datasets=[\"squad\",\"cnn_dm\",...], k=1000, t=2.0)\r\n```\r\nUsing the t5 method of combining here (or this could be a function passed in as an arg) \r\n\r\nPassing kwargs to each 'sub-dataset' might become tricky.","embeddings":[-0.0292878691,-0.394043982,-0.0610294119,-0.0256624818,-0.2608686984,0.082581006,0.1810528636,0.0469555631,0.2868432105,-0.1594537348,-0.2330442071,0.3090499938,-0.2074714005,0.3853416443,0.3700452447,-0.446831882,-0.0071452288,-0.2071918547,-0.2622185349,0.3925702274,0.0369971208,0.1119897887,-0.1414716542,0.109804526,-0.3010460138,-0.136756286,-0.3321913779,0.1030589566,0.1164888591,0.0181283373,0.2122901678,0.3617036343,-0.0475504771,0.143895179,-0.0001149721,-0.0445622839,0.1405429542,-0.0745464861,0.1279869676,-0.0385321192,0.1727730036,-0.3879154325,0.0674856231,-0.1346675009,0.2163201421,0.1592068374,0.1809329093,0.0692611039,0.4056928158,-0.3285125494,0.0875926465,0.2849261165,-0.1900693029,0.3437137902,-0.106348373,-0.104068324,0.0376524441,0.2589140236,0.5864793658,-0.5919806361,0.0531647094,0.0858707279,0.0129780285,-0.0354294479,0.1583545208,-0.0285796337,-0.3110812306,-0.224444598,-0.4186833799,0.1887254119,-0.1044372246,-0.1551899761,-0.221528545,-0.5562813878,-0.0170732997,-0.0037981721,-0.3029250801,-0.0021318083,-0.1731025726,-0.0394975655,-0.1476764232,-0.1062141284,-0.0510747917,0.154971689,0.334846288,0.5709902048,0.1295996904,0.1710970849,0.319866538,0.0538763031,-0.1331725866,-0.2922166288,0.3081015646,0.1642882973,-0.3647081256,-0.3597905934,0.1917760223,-0.2589807212,0.2324257493,0.0934328586,0.127033025,0.0813640952,-0.1447874457,0.2692347169,0.3316326737,0.0368669368,-0.1052711457,-0.1273326576,0.0287625547,-0.1599454433,0.2352163494,0.0967027619,0.1859356016,-0.0221786052,-0.3954431713,-0.1633712053,-0.1346349418,0.1488838494,-0.1919805855,-0.5355588794,-0.1626533121,-0.1399561167,0.0474985354,-0.1304683983,-0.314229548,0.076599434,-0.0300824065,0.1895588785,-0.1426861584,0.1208322644,-0.0526403412,0.0712479576,-0.4480927885,-0.0460767262,0.2776448429,0.109525986,0.08483053,0.1338436902,0.0148406941,0.1254808456,0.401938498,0.0392921008,-0.0199368186,-0.033141423,0.1681917012,-0.3317155242,-0.0062631932,0.1757421792,-0.3530600965,-0.1490301788,-0.15961124,-0.1868329048,0.1211162359,0.0244976729,-0.2208403945,-0.3201468289,-0.4782351851,0.6768115163,-0.027634982,-0.0647312999,-0.16472359,0.2757855952,-0.2691693902,-0.1202769652,0.1744775176,0.0458281785,0.0011190666,-0.3560557067,-0.0667324588,-0.0792039707,-0.0776051804,0.2584712505,-0.1712939292,0.2623884678,0.0908837914,0.1857641786,0.3408804536,-0.4731499255,-0.0226834994,0.2970294952,-0.0556328595,0.4108914137,0.070467867,0.2279607058,-0.0812057629,-0.2089106292,0.2719022036,1.0307596922,-0.3764421046,-0.0222222898,-0.0957022458,-0.3904013932,0.5535189509,0.3494330049,0.0771400183,-0.3079709113,-0.1021800116,-0.0096340487,0.0667360649,0.0841099992,0.074733451,-0.0878115147,-0.3393520415,0.1139122844,-0.2102668732,-0.1518988907,-0.3530140519,0.0339234881,0.0241677295,0.3140191734,0.2573538721,-0.1301783174,0.2086983174,-0.2874013186,-0.1486529857,-0.2948302031,0.1073915586,-0.0712572262,0.0002874161,-0.2379792929,-0.0317758545,0.307772398,0.1399238557,-0.1362237781,-0.477207154,0.293318361,-0.0632221475,0.0578540787,-0.0563283488,0.648289144,-0.1621540636,-0.0672787055,0.192209065,0.3254882395,-0.3306884766,0.2363168746,0.3137443066,0.3130964339,0.1656142473,0.0061513577,0.1179905161,0.0116396798,-0.0012399972,-0.2144867331,-0.0947560817,0.5063752532,-0.1947730184,0.4248305261,-0.0239511244,-0.0024468424,-0.2053102404,-0.0152175212,-0.1752177924,0.4153912663,0.2905997634,-0.2353711426,0.2970917821,-0.1605678797,-0.3739417493,0.1721218228,0.1792156249,0.0188672002,0.0344038904,-0.0445342585,0.0077328938,-0.2754392922,-0.1060047299,0.3186826408,0.657732904,0.254458189,0.1376096606,0.0657043904,-0.2021635175,-0.0595594049,0.0485227667,0.2135917246,-0.0488939099,0.1590911001,0.0362472609,0.0808383077,0.0621163175,-0.1677286327,0.1392087489,-0.2044975758,-0.116217196,0.0541696772,-0.184016943,-0.6338432431,-0.1893022954,-0.0091773653,-0.3316513896,-0.2957566082,0.1372021437,0.0558245629,-0.1146438941,0.1387735605,0.1970451176,0.5815647244,-0.4308137894,-0.2768557668,-0.0366385765,-0.3137181103,-0.101734519,0.1042483672,0.4355737567,0.2959771752,0.5036947131,0.2990525961,-0.0688932687,0.0986968577,-0.4021719694,0.065417096,-0.3533987999,0.2585304081,0.1358294189,-0.2217337042,0.1458302438,-0.3897809684,-0.0530337431,-0.0256696921,-0.0749264508,-0.0513381362,0.0540681854,0.0693137795,-0.1985552162,-0.190367803,-0.6149933338,-0.4419181347,0.3062751889,-0.2279975116,0.082725279,0.0621675141,-0.2821539342,0.2345841825,-0.0894587934,0.0253634453,-0.1702772528,-0.0523394346,0.0101791928,-0.3848066926,-0.1447838843,-0.3393516243,-0.0398955233,0.2213984579,-0.0660574064,-0.0294074975,-0.2750928998,0.0828027353,-0.087789759,0.0721915066,0.262242943,0.1656526178,-0.2060345113,0.1317709833,-0.0811146721,-0.2719359398,0.2103470117,0.3302366436,0.3737383187,0.1844933927,0.0363181233,0.1870743483,0.7639582157,0.248142615,0.0630793497,0.1484031677,0.2611881196,-0.0126449708,0.1614734381,-0.1882690638,0.1840164065,-0.1733156145,0.5277926922,0.0812266916,0.1695803106,-0.1492176503,-0.3321087658,0.0180272721,-0.3142365217,-0.1845273674,0.190161258,-0.2771730721,0.0171557404,-0.0910211504,-0.4362221062,-0.2857394516,-0.0373797268,-0.0170006379,0.3356775045,-0.0067285937,0.1707703769,-0.3354098499,-0.0434185937,-0.4217490256,0.1442063153,0.0583045147,0.2372829616,-0.0807373896,-0.0721391886,-0.1010343507,0.1118271425,0.5621086359,-0.5473772883,-0.1090124026,0.0950557664,0.0797650218,-0.12696293,-0.1165482402,-0.2918789983,0.0204344355,0.4276299179,0.3768800199,-0.3781233132,-0.2943044007,0.2408283055,-0.2539578676,-0.0721550807,-0.1933034807,-0.2405900061,-0.2112835795,-0.1243794337,-0.0059253396,0.2536102533,0.234411031,-0.1205018461,0.1836338043,-0.0812310353,0.2604237497,0.2436981648,0.0103810718,-0.1104074419,0.2441927493,0.1717461348,0.0389957577,-0.1210240722,-0.3591462374,0.099854745,-0.0216738395,0.0665015951,0.2553038895,-0.0018159044,0.6958088875,0.046072565,0.3278181255,0.2777076066,-0.1480576396,-0.0029457514,-0.2239089906,0.1858537197,0.2340468913,0.2490611374,-0.1415129602,-0.2325429469,0.5836411119,0.3070313632,-0.0868651122,0.4385273755,-0.4941880107,-0.3337468803,0.2170913666,0.0257658698,0.7134880424,-0.0225321557,0.3297557831,0.1180819049,-0.4132733047,0.5890747309,-0.336222738,0.1240845397,-0.0548373349,-0.1727508008,-0.0110374019,-0.0759113953,-0.1029586121,0.2057562172,-0.3488744497,0.3369855285,0.4035097659,-0.2302371562,-0.0473927855,0.5434772372,-0.0144713148,-0.34057796,-0.070958592,0.0513050705,-0.2821181417,0.0694786757,-0.2126549184,-0.2552015185,0.2105032951,0.159284234,-0.2744202316,0.0667962953,0.0081969816,0.2008349746,0.2247757316,0.0355926827,0.2640016675,-0.1730505675,-0.1037317812,-0.1767416894,-0.0648873448,0.1057883203,-0.2313127667,0.0363349319,0.0953336582,-0.1507094353,0.4938498437,-0.0744175091,-0.1055260152,-0.225899443,0.2618069053,-0.4456818998,-0.2548679709,0.1317418665,0.1610529125,0.18171525,-0.3218866587,0.5945078731,0.0866671577,0.0534644015,0.0180057138,0.170484364,-0.2397265136,0.4645395279,-0.3223521709,0.0002991938,0.0991841555,0.0001488334,-0.1518300325,-0.4189449251,-0.0535859466,0.2748262286,0.1244942546,0.0553411245,-0.0169946533,0.1142993644,-0.0379925631,-0.0555110238,-0.1118920743,-0.2893212736,0.1342903674,0.2440163344,0.1765129417,0.1978478134,-0.0604548119,-0.2944520116,-0.130066976,0.4992496669,-0.0168993901,0.2320831567,-0.1513770372,0.3708628714,0.4106283188,0.4903714657,-0.2431000471,0.0570452549,0.098533161,0.1000434607,0.4676124752,-0.2268052399,0.0473871827,0.2821198106,-0.1461158246,0.0312455148,-0.3519589901,-0.1358806789,-0.2496926934,0.1245684698,0.128815487,-0.028475469,0.0284170788,-0.1112866476,-0.1046101078,-0.0577550158,0.0444400497,-0.3633012474,-0.0975726917,0.1614964008,-0.0941285342,0.0867119581,0.05705202,0.2070290595,0.0694729835,-0.0731083527,0.3387204111,0.1200710535,0.0868645012,-0.1681587547,-0.1408303976,0.0164401643,0.2955585122,0.1803165227,-0.0689326674,0.1422432661,-0.2093845457,-0.1997292787,0.2077007294,0.2024322897,0.0352877043,-0.1302379817,0.1069171354,0.0964848995,-0.0600155517,-0.0129073607,0.2301515043,0.0877920762,0.0724479258,0.1168100163,0.1136879623,0.4800716639,-0.2822709978,0.032748986,0.0944719836,-0.359455049,0.1163828969,-0.0824384242,-0.0855578408,-0.0144915152,0.1247710511,0.0000897483,0.1193355918,-0.3342020214,0.3240496814,0.2672713399,0.0153990667,0.0150010176,0.1387518495,0.0422903299,-0.1731810868,0.2679603398,0.0350200571,0.3751724362,0.0604134649,0.8366437554,-0.2134697288,-0.3915117383,0.1579958797,0.4222691059,0.0876292065,-0.2188608199,0.1318653226,0.0981128737,-0.3754850924,0.069795914,0.0569917336,-0.1820816249,0.6210576892,-0.0859890431,0.168260932,-0.117920056,-0.1851587892,0.3580678105,0.2187560946,-0.0734489933,-0.0396005027,0.1282267421,-0.0619420893,-0.1988057196,0.4743301272,0.2133032084,-0.0024091061,-0.4816867113,0.0333791673,0.1014535874,0.1790028811,0.0025960908,0.1369649768,0.446062237,-0.1502597183,0.2034935951,-0.0070444676,-0.0340181813,-0.1344564557,0.0641842932,-0.4029557109,-0.423715502,0.4620600343,0.1102590933,-0.1495261788,0.1175512373,0.3762098253,-0.6295855045,-0.2694912255,0.3343758881,-0.0582320802,-0.0850206539,-0.1960967928,0.0654972419,0.0751247779,0.6126101017,0.2080774754,0.4082014859,-0.3459770381,-0.096712023,-0.5682125092,-0.2036944181,0.1717404425,-0.1529812217,-0.2241835743,0.118387796,-0.0693334118,0.2762838602,-0.189055413,0.1030983031,0.2452890724,0.0498199388,-0.4366213679,0.0849430114,0.1906822026,0.0404026471,0.2738911808,-0.0639429241,-0.2266221941,0.0473255366,-0.0683634505,-0.0969071463,-0.114198789,0.2681549788,0.1839283854,0.0959584117,0.0570734367,0.5114495754,0.0109695597,-0.1574940979,0.0035756819,0.0437193774,-0.3633825481,-0.135695979,-0.0934822485,0.1329881847,-0.3916032314,0.3226499557,0.0057077692,-0.0780505538,0.0918762162,0.0003349573,-0.1227842569,-0.1121983007,0.0437778868,0.2214048952,0.1861128509,0.7520397305,-0.0126848361,0.1377276927,-0.0949488133,-0.0087831272,0.0272683632,-0.0643478408,-0.2542924583,-0.0940607414,0.0674461201,0.0680915862,0.1432077736,-0.6984319687,-0.3541297913,0.050463587,0.0161522999,0.0159998555,0.7800256014,0.0947367921,-0.1327628791,-0.0894504786,-0.3622466028,0.0005289685,-0.2262697965,0.1659860313,-0.374137044]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"From thinking upon @thomwolf 's suggestion, I've started experimenting:\r\n```python\r\nclass MultitaskDataset(DatasetBuilder):\r\n def __init__(self, *args, **kwargs):\r\n super(MultitaskDataset, self).__init__(*args, **kwargs)\r\n self._datasets = kwargs.get(\"datasets\")\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(\r\n description=_DESCRIPTION,\r\n features=nlp.Features({\r\n \"source\": nlp.Value(\"string\"),\r\n \"target\": nlp.Sequence(nlp.Value(\"string\"))\r\n })\r\n )\r\n\r\n def _get_common_splits(self):\r\n '''Finds the common splits present in all self._datasets'''\r\n min_set = None\r\n for dataset in self._datasets:\r\n if min_set != None:\r\n min_set.intersection(set(dataset.keys()))\r\n else:\r\n min_set = set(dataset.keys())\r\n return min_set\r\n\r\n....\r\n\r\n# Maybe this?:\r\nsquad = nlp.load_dataset(\"squad\")\r\ncnn_dm = nlp.load_dataset(\"cnn_dailymail\",\"3.0.0\")\r\nmultitask_dataset = nlp.load_dataset(\r\n 'multitask_dataset',\r\n datasets=[squad,cnn_dailymail], \r\n k=1000, \r\n t=2.0\r\n)\r\n\r\n```\r\n\r\nDoes anyone know what methods of `MultitaskDataset` I would need to implement? Maybe `as_dataset` and `download_and_prepare`? Most of these should be just calling the methods of the sub-datasets. \r\n\r\nI'm assuming DatasetBuilder is better than the more specific `GeneratorBasedBuilder`, `BeamBasedBuilder`, etc....\r\n\r\nOne of the other problems is that the dataset size is unknown till you construct it (as you can pick the sub-datasets). Am hoping not to need to make changes to `nlp.load_dataset` just for this class.\r\n\r\nI'd appreciate it if anyone more familiar with nlp's internal workings could tell me if I'm on the right track!","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":177,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nFrom thinking upon @thomwolf 's suggestion, I've started experimenting:\r\n```python\r\nclass MultitaskDataset(DatasetBuilder):\r\n def __init__(self, *args, **kwargs):\r\n super(MultitaskDataset, self).__init__(*args, **kwargs)\r\n self._datasets = kwargs.get(\"datasets\")\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(\r\n description=_DESCRIPTION,\r\n features=nlp.Features({\r\n \"source\": nlp.Value(\"string\"),\r\n \"target\": nlp.Sequence(nlp.Value(\"string\"))\r\n })\r\n )\r\n\r\n def _get_common_splits(self):\r\n '''Finds the common splits present in all self._datasets'''\r\n min_set = None\r\n for dataset in self._datasets:\r\n if min_set != None:\r\n min_set.intersection(set(dataset.keys()))\r\n else:\r\n min_set = set(dataset.keys())\r\n return min_set\r\n\r\n....\r\n\r\n# Maybe this?:\r\nsquad = nlp.load_dataset(\"squad\")\r\ncnn_dm = nlp.load_dataset(\"cnn_dailymail\",\"3.0.0\")\r\nmultitask_dataset = nlp.load_dataset(\r\n 'multitask_dataset',\r\n datasets=[squad,cnn_dailymail], \r\n k=1000, \r\n t=2.0\r\n)\r\n\r\n```\r\n\r\nDoes anyone know what methods of `MultitaskDataset` I would need to implement? Maybe `as_dataset` and `download_and_prepare`? Most of these should be just calling the methods of the sub-datasets. \r\n\r\nI'm assuming DatasetBuilder is better than the more specific `GeneratorBasedBuilder`, `BeamBasedBuilder`, etc....\r\n\r\nOne of the other problems is that the dataset size is unknown till you construct it (as you can pick the sub-datasets). Am hoping not to need to make changes to `nlp.load_dataset` just for this class.\r\n\r\nI'd appreciate it if anyone more familiar with nlp's internal workings could tell me if I'm on the right track!","embeddings":[-0.0410289392,-0.3800130486,-0.0561361499,-0.0553337559,-0.226961717,0.0381454341,0.191163376,0.0285417549,0.3140310943,-0.1693242192,-0.2113295496,0.3226813376,-0.2117785513,0.3609280288,0.4108115137,-0.4081549644,0.0496115424,-0.2385492176,-0.3018552065,0.3816812336,0.0294328276,0.1105419919,-0.1815611124,0.1116595641,-0.2753484547,-0.1470191181,-0.3376018703,0.1057681218,0.1258859038,0.0001230787,0.2541147768,0.405015111,-0.0645343959,0.1640153676,-0.0001138703,-0.0289735682,0.2369827926,-0.0645472929,0.1149068475,-0.0750551447,0.2082991898,-0.4271639287,0.0591521002,-0.1675683409,0.1729619354,0.1567192823,0.1583682895,0.0400266796,0.3936338425,-0.2916638255,0.1032506153,0.2875022888,-0.1544683576,0.2761384845,-0.1251125187,-0.0476874821,0.036738351,0.2582830191,0.5602095127,-0.6138654351,0.0037755559,0.1158641875,0.0011100308,0.0048890761,0.1531963944,-0.0207958259,-0.3521731496,-0.2176881582,-0.4039084017,0.1862858385,-0.1451549232,-0.2017221302,-0.2375371158,-0.544827044,0.0046707969,0.0061735087,-0.2698778808,-0.0183589067,-0.2017983645,-0.0632458106,-0.1232680306,-0.0896949619,-0.0336662307,0.1552536786,0.3182489872,0.5842538476,0.0905896127,0.1673257947,0.3679713309,0.0413935184,-0.1508580148,-0.2804937363,0.327536732,0.1458220184,-0.3698496819,-0.3030605018,0.2041640729,-0.2367646396,0.2298245132,0.104572162,0.0904757529,0.1172205508,-0.1479838192,0.2619043887,0.3045757413,0.0461219847,-0.096149154,-0.0954477936,0.0346761346,-0.0859440193,0.1927355975,0.1121786162,0.1854983568,-0.0123846577,-0.3950420022,-0.1339060962,-0.1293433756,0.164744541,-0.2152846307,-0.4849033356,-0.1085741743,-0.1195227951,0.1177139729,-0.1045871302,-0.3091354072,0.0987556651,-0.0389532857,0.2612789869,-0.1374977529,0.1306159794,-0.071044147,0.0840983689,-0.4412369728,-0.0771046504,0.3059286177,0.1518349051,0.0807021558,0.1679016352,-0.0014322727,0.1236036271,0.3893814683,0.0155780558,0.0013976353,-0.0009416561,0.1396428198,-0.3161430657,0.0204230193,0.1589076817,-0.3474511504,-0.1902999729,-0.1193476021,-0.1685406864,0.140574187,0.032154683,-0.2005953342,-0.2385002077,-0.4981879592,0.7015710473,-0.0572152957,-0.0934591815,-0.1759139746,0.3037517369,-0.2464628518,-0.0800704435,0.2119654566,0.0576393493,0.0113504892,-0.3442129493,-0.1054180935,-0.0986813679,-0.0215478763,0.2500798404,-0.1620027721,0.2792710364,0.0527335331,0.2090146989,0.2798385322,-0.4821227789,-0.0626416206,0.2596439421,-0.0441140719,0.422999233,0.0580770448,0.1961356997,-0.0260135718,-0.1701483727,0.2901467979,0.9980714917,-0.3438392878,-0.0194317736,-0.1091452911,-0.3598763347,0.5201604962,0.325127393,0.1099428758,-0.2907424569,-0.1358843744,-0.0152131552,0.1217473596,0.0748822615,0.0830932632,-0.0717745125,-0.3480929434,0.0625114515,-0.2146702111,-0.1692674905,-0.3402894139,0.0233767964,0.0537420772,0.3371377587,0.2115899175,-0.1451852024,0.195847854,-0.2553427219,-0.1432838887,-0.3022666276,0.1462630332,-0.0699497685,0.0136631252,-0.2408796698,-0.0318978019,0.3690548241,0.1162230149,-0.1267782897,-0.4974001646,0.3005133867,-0.0742243752,0.0666137859,-0.0325902961,0.6585513353,-0.1506325752,-0.0851470754,0.1811940074,0.2812555134,-0.2942326665,0.201296702,0.303630203,0.1976070404,0.1517353952,0.0046875426,0.073297821,0.0522425622,0.0247967355,-0.2336047441,-0.0985799506,0.5100664496,-0.1613469124,0.452299118,-0.0280021038,0.0251580551,-0.1460136026,-0.0296826884,-0.2038654834,0.3801978528,0.2894101441,-0.1866005212,0.3278831244,-0.1623674035,-0.3643948436,0.1550501585,0.1312014014,0.0126263304,0.0546307974,-0.043492347,-0.0132482331,-0.279115051,-0.107753031,0.3511256278,0.680601418,0.2346359193,0.1918638498,0.0699950233,-0.2855549753,-0.0711565241,0.0013609271,0.2216838598,-0.0406082161,0.1673647016,0.0622549988,0.0424435064,0.0219647475,-0.1528270543,0.1556803733,-0.1852088571,-0.0936624333,0.1074038446,-0.2216831148,-0.5938890576,-0.1774952114,0.0022366841,-0.2878167629,-0.3085010648,0.103927806,0.0533200093,-0.1458798796,0.1620602161,0.2049497664,0.5456448197,-0.42458269,-0.2756488323,-0.0417231359,-0.3142536581,-0.1109102294,0.1089711413,0.4567675889,0.3351971805,0.4890900254,0.3100687563,-0.0816786662,0.0496938787,-0.4016466141,0.0475736223,-0.3254904449,0.2670509517,0.1173540875,-0.2622727752,0.1621409059,-0.3722044528,-0.0118275424,-0.087348938,-0.085972406,-0.0219025519,-0.0195316207,0.0591497868,-0.2100077868,-0.1561510116,-0.6618935466,-0.4537318647,0.3190036714,-0.2383414656,0.0822201148,0.0593159348,-0.2954932153,0.2223705649,-0.0649874806,0.0163610484,-0.1773646623,-0.0488452837,0.0293169599,-0.3534869552,-0.1780765206,-0.3723646402,-0.1382026076,0.2802268267,-0.0606195256,-0.0071239159,-0.2503638566,0.0998615101,-0.0790094286,0.1057046577,0.246562615,0.1785149276,-0.2391591221,0.1344874799,-0.0601592287,-0.2647292614,0.2468385696,0.2817914486,0.3843086064,0.1833924949,0.0599358752,0.1708058119,0.6879074574,0.2298003733,0.0656740442,0.1472759098,0.2549452186,0.019310737,0.1698751003,-0.2620752752,0.2082164139,-0.2138110101,0.4608762264,0.0450818129,0.1656016558,-0.2366372794,-0.2956605256,0.024182396,-0.3138543069,-0.2150159627,0.1974213868,-0.2264554799,0.0024594467,-0.0736261606,-0.3792159557,-0.2847125828,-0.0331578329,0.0045121517,0.3330295682,-0.096041657,0.1180453449,-0.3570632935,-0.0310056992,-0.4697405696,0.1414586008,0.0083270064,0.2584218979,-0.0509750359,-0.0787402913,-0.0897191986,0.07503286,0.6008829474,-0.4771214426,-0.1246808693,0.1526512355,0.111176379,-0.1372941285,-0.126300782,-0.2912558019,0.046896819,0.4684960246,0.3723252416,-0.3875468373,-0.2634352446,0.2291092277,-0.2160620987,-0.0449889861,-0.1908899099,-0.2358557433,-0.256003648,-0.0991715342,0.0442418344,0.217778191,0.232020542,-0.1234763637,0.2110002041,-0.1089715958,0.2612601519,0.2202203572,0.0042052749,-0.0271496847,0.201087147,0.1709838361,0.0384250358,-0.1398618966,-0.3714193106,0.1328180283,0.0046475013,0.0788257197,0.2440619171,-0.0076570194,0.633480072,0.0739131942,0.2964058518,0.2896263301,-0.1486874819,0.0177449565,-0.2544413507,0.1916362196,0.2428325862,0.2626424432,-0.1403993964,-0.2564390302,0.5900325179,0.3069310486,-0.0994892791,0.4368517697,-0.4996849895,-0.3505488932,0.2613225877,0.0221008472,0.6753665209,0.0059887902,0.3173231184,0.1614212096,-0.3751908541,0.5796290636,-0.3777527511,0.1610089988,-0.0703779981,-0.1492097825,0.0048084618,-0.0804935247,-0.037119057,0.2290440798,-0.3471155763,0.3315229714,0.4464797974,-0.2043319494,-0.0685900748,0.515676856,-0.0236175247,-0.3742783964,-0.1238871142,0.0779337734,-0.2607240379,0.0987707451,-0.1555745453,-0.2480808645,0.1932965666,0.1464109123,-0.2879010141,0.1049910337,0.0308531169,0.2383003831,0.256922096,0.0098964563,0.2169825882,-0.1943752617,-0.0544420071,-0.1598699093,-0.0600800104,0.1227303594,-0.2081538588,0.0833739862,0.0815269575,-0.2225060165,0.5000631809,-0.0530828014,-0.1086613685,-0.2400417626,0.2608032823,-0.4327819645,-0.183997646,0.1675369889,0.1632853001,0.2320406437,-0.3219026327,0.6011402011,0.1171333641,-0.0090434384,0.035241425,0.1837954521,-0.2748790085,0.5102385879,-0.3566158414,0.01190913,0.0858415067,0.0122207794,-0.1473492682,-0.3761584759,-0.032249473,0.2993479073,0.085238643,0.0319864675,-0.0109043056,0.1051611453,-0.0341627412,-0.0448406711,-0.0973384678,-0.3260810971,0.1228291616,0.1837864965,0.1799663007,0.1721730977,-0.0854676887,-0.27731511,-0.180569455,0.4803913534,0.0230193175,0.2021276355,-0.1525309682,0.3534677625,0.4196875393,0.5000764132,-0.2531453967,0.0555125102,0.0724088699,0.0715796873,0.4242964089,-0.2460532784,0.0602255426,0.2450597435,-0.1215982437,0.0435207933,-0.351827085,-0.1462923288,-0.2637699842,0.1085054204,0.109384872,0.0003898287,0.0513631143,-0.0967009664,-0.0972466841,-0.0661537126,-0.0129816364,-0.3419176638,-0.0716903508,0.1208119094,-0.0695233643,0.0431894846,0.073523052,0.1863006353,0.039544303,-0.0755516812,0.311919719,0.0912711844,0.1235490665,-0.1606186926,-0.1367616355,0.0202979762,0.2227637172,0.1559672803,-0.0428765193,0.1189593375,-0.2020344734,-0.1555562317,0.273425132,0.2207930386,0.0274563301,-0.133365646,0.1312036216,0.1121091917,-0.1198824346,-0.0256842058,0.2396405786,0.0806817859,0.1064941585,0.1220116466,0.1149659082,0.4606329501,-0.289588362,0.0494988672,0.1124724522,-0.3702643216,0.1184360385,-0.1373876631,-0.0105006918,0.026157964,0.1451226175,-0.0045402567,0.115399912,-0.2439991236,0.323844552,0.3000831902,-0.0101888375,0.0390351415,0.1394428611,0.0214830562,-0.1702267677,0.2212381661,-0.0121520199,0.3521032631,0.0319326445,0.8429723382,-0.3010584712,-0.3967199028,0.1265186667,0.4000535905,0.0607582554,-0.2388640791,0.132171616,0.1059504822,-0.3579234779,0.0606999174,0.0641688108,-0.1542833894,0.6550419927,-0.1400014311,0.1647063643,-0.1756535172,-0.2119795084,0.3319477737,0.2027831525,-0.0779683366,-0.0116058402,0.1274465919,-0.0609823652,-0.1957384497,0.4920668304,0.2548992038,0.0515676066,-0.4656876028,0.0072833374,0.0970633402,0.1373630166,-0.0935615972,0.1530828774,0.4537776411,-0.1514495909,0.2070917934,0.0042733736,-0.0584352799,-0.0991597697,0.082723774,-0.3991208673,-0.4085475504,0.5045669079,0.0608776174,-0.1160669252,0.0809353068,0.3370766044,-0.6575293541,-0.2604760528,0.3494887054,-0.0848937407,-0.0373116024,-0.2238081694,0.0740792379,0.0727372319,0.6086327434,0.1636591852,0.3771531284,-0.3970235288,-0.1071991771,-0.6464769244,-0.2132060677,0.0997600555,-0.1670626551,-0.2602818608,0.1377420872,-0.0389875732,0.2751605809,-0.2075423896,0.0516323633,0.2532147765,0.0269209407,-0.456305027,0.0790516883,0.1370771676,0.0201094635,0.2837494016,-0.0778978392,-0.2213195115,0.0658564419,-0.0567041151,-0.0504934266,-0.0861395821,0.2779473066,0.175609678,0.0359932892,0.0480653867,0.5123876929,-0.0224155672,-0.1742225289,-0.0260288995,0.0285733473,-0.372587055,-0.1118155867,-0.1226736084,0.2113185376,-0.3768762648,0.2700481117,0.0169577952,-0.1017483026,0.122689046,-0.0095324144,-0.1750100106,-0.021082608,0.0441170409,0.2265998423,0.2110447437,0.7624977827,-0.0001504509,0.1709384471,-0.0762044117,-0.0665120259,0.0755371004,-0.1174957082,-0.2645829022,-0.1545645595,0.0648577288,0.0475234613,0.1106641516,-0.7056497335,-0.3468599319,0.0161692388,-0.0191768166,0.0539594144,0.7361037135,0.0726608709,-0.1439816207,-0.1310945004,-0.3231859207,0.0205402821,-0.2399547696,0.1891361475,-0.3676554561]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"I think I would probably go for a `MultiDataset` wrapper around a list of `Dataset`.\r\n\r\nI'm not sure we need to give it `k` and `t` parameters at creation, it can maybe be something along the lines of:\r\n```python\r\nsquad = nlp.load_dataset(\"squad\")\r\ncnn_dm = nlp.load_dataset(\"cnn_dailymail\",\"3.0.0\")\r\n\r\nmultitask_dataset = nlp.MultiDataset(squad, cnn_dm)\r\n\r\nbatch = multitask_dataset.sample(10, temperature=2.0, k=1000)\r\n```\r\n\r\nThe first proof-of-concept for multi-task datasets could definitely require that the provided datasets have the same name\/type for columns (if needed you easily rename\/cast a column prior to instantiating the `MultiDataset`).\r\n\r\nIt's good to think about it for some time though and don't overfit too much on the T5 examples (in particular for the ways\/kwargs for sampling among datasets).","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":114,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nI think I would probably go for a `MultiDataset` wrapper around a list of `Dataset`.\r\n\r\nI'm not sure we need to give it `k` and `t` parameters at creation, it can maybe be something along the lines of:\r\n```python\r\nsquad = nlp.load_dataset(\"squad\")\r\ncnn_dm = nlp.load_dataset(\"cnn_dailymail\",\"3.0.0\")\r\n\r\nmultitask_dataset = nlp.MultiDataset(squad, cnn_dm)\r\n\r\nbatch = multitask_dataset.sample(10, temperature=2.0, k=1000)\r\n```\r\n\r\nThe first proof-of-concept for multi-task datasets could definitely require that the provided datasets have the same name\/type for columns (if needed you easily rename\/cast a column prior to instantiating the `MultiDataset`).\r\n\r\nIt's good to think about it for some time though and don't overfit too much on the T5 examples (in particular for the ways\/kwargs for sampling among datasets).","embeddings":[-0.0309596602,-0.4186734557,-0.0396580435,-0.0468707941,-0.2247027457,0.0937634408,0.1678373814,0.0375866704,0.2795366049,-0.1670990288,-0.2208789289,0.3342223167,-0.2148639411,0.3291554749,0.4196427464,-0.4405975938,0.0240546688,-0.2191532105,-0.2579804361,0.3798784018,0.0655685812,0.1019967496,-0.1636252403,0.0597926565,-0.2679932714,-0.1647084355,-0.3614471853,0.1224154383,0.1369815916,-0.0105154719,0.2192510962,0.3853777051,-0.0239703581,0.1168468222,-0.0001133946,-0.0766079575,0.1594891995,-0.1200244948,0.0910891593,0.0065836376,0.1546203941,-0.427708894,0.0729347542,-0.1584373564,0.1449337304,0.15658544,0.1450680494,-0.0069006812,0.4177408218,-0.2556488514,0.0976841375,0.3037799597,-0.1650875509,0.2707971334,-0.1420085579,-0.0669919997,0.0279299561,0.2208573371,0.614895761,-0.5899104476,0.0041678394,0.0785674825,-0.0143497717,0.0074051167,0.1554833502,-0.0298841391,-0.3357186317,-0.2630695403,-0.3932988048,0.2161172032,-0.130057469,-0.1769689173,-0.2505916059,-0.5384424329,0.0170675442,0.0354164727,-0.2903538942,0.0315727927,-0.2046495676,-0.0545909144,-0.1233536974,-0.1029213518,-0.0065230848,0.1564510912,0.3066310883,0.571914196,0.1277765781,0.171288386,0.352591455,0.0238961987,-0.1490521431,-0.2876000404,0.2940413952,0.1621154547,-0.4089529216,-0.3507789373,0.1309266686,-0.2493540794,0.2844475806,0.0881018192,0.118369028,0.0559124313,-0.1370865554,0.2888890803,0.3011126518,0.0355755202,-0.0766883716,-0.1267349124,0.0218941979,-0.1330881119,0.1889595985,0.1262667626,0.1817833036,-0.0086608129,-0.3749386966,-0.1357930005,-0.1604589373,0.173871994,-0.1611000299,-0.5166141391,-0.110078156,-0.0932472199,0.0721659958,-0.0924680829,-0.2943404615,0.0761008114,-0.0488542356,0.2247578651,-0.0882168263,0.0521588661,-0.0636374578,0.0874368623,-0.4271279275,-0.0311003216,0.3111233115,0.0953136384,0.0791382715,0.1683205217,-0.0090656439,0.107864365,0.4213511944,0.0207369979,-0.0052913204,-0.0257206541,0.2052460611,-0.2978515923,0.0131257866,0.1394484788,-0.3443307877,-0.1233699843,-0.1408030689,-0.1617729217,0.1051673219,0.0261682812,-0.2270695865,-0.2634857297,-0.4787130952,0.7096787691,-0.0163178314,-0.0407355018,-0.1800363213,0.2876241207,-0.2645999789,-0.1131608337,0.2292123139,0.1003069729,-0.0439029336,-0.2998158038,-0.0713219419,-0.0835302398,-0.0897591934,0.2788395584,-0.1998572201,0.2227457315,0.0680124685,0.1548843086,0.3027401865,-0.4629935324,-0.0365784392,0.269169122,-0.0357774608,0.3846664429,0.0656009838,0.2249258608,0.0193822924,-0.2001308948,0.2639879584,1.0421270132,-0.4086365104,0.0009421139,-0.0963041261,-0.3568603396,0.5225280523,0.3448227048,0.0528929122,-0.289357245,-0.1473952085,-0.0054200753,0.0723923892,0.0581510626,0.0951675251,-0.0369884484,-0.3416343629,0.0695472807,-0.2200360298,-0.1500298381,-0.4430118203,0.0194523167,0.0445789285,0.3434177339,0.2352165729,-0.1388451904,0.2276782244,-0.2906135023,-0.1399959624,-0.3124925792,0.1317205429,-0.0509352908,0.0310703609,-0.220759511,-0.0353601761,0.3621313572,0.1493374705,-0.1270153522,-0.480266124,0.3278311789,-0.0759468377,0.036641106,-0.0272858683,0.6227853298,-0.1506923288,-0.0730190575,0.1906048506,0.2668913901,-0.2859715223,0.2114320993,0.3612276018,0.2709768116,0.1712347567,-0.03409243,0.095298022,0.0238095596,-0.0051739481,-0.2202403247,-0.0857056677,0.5233914852,-0.1781087816,0.4187562466,-0.0450783186,0.0366004072,-0.1355301589,-0.018147558,-0.1988152266,0.4192889631,0.2726354301,-0.1745443195,0.3594210446,-0.1392042488,-0.3428289294,0.160084784,0.1558927,0.0385063365,0.0907846987,-0.003107762,-0.0485005416,-0.2614943981,-0.1089155227,0.3499152958,0.6467790008,0.2586946785,0.1375753731,0.0132202739,-0.2505892217,-0.0694816485,0.0318481065,0.1693363041,-0.0284860265,0.1438650191,0.061590679,0.0490800515,0.0293323211,-0.1715495288,0.1837420762,-0.1622173339,-0.0755687952,0.0723839849,-0.1664850414,-0.5669133067,-0.2283149213,0.0277331453,-0.3815065026,-0.2889538109,0.1047791615,0.0514129698,-0.1202398837,0.1866274923,0.2017738074,0.572204411,-0.3877091706,-0.2930240035,-0.0291458052,-0.3389903903,-0.1145647168,0.112439476,0.4557248652,0.2603594363,0.4842223227,0.3259260952,-0.0339717716,0.0463758931,-0.398424834,0.0176438,-0.3012909591,0.2676162124,0.073774904,-0.1858690083,0.1533944756,-0.406892091,-0.0056130914,-0.0210666899,-0.0708756596,-0.0528934598,0.0044161165,0.0954201221,-0.1834784299,-0.1902592927,-0.6412388086,-0.4301627874,0.26739344,-0.260163337,0.0865977556,0.0140416119,-0.2555756271,0.1796784401,-0.1078665927,-0.0026081605,-0.1634770185,-0.0389524139,0.0466123745,-0.3563415408,-0.1756981164,-0.3697786331,-0.1142726988,0.2460547686,-0.0694768429,-0.0248483159,-0.2876038253,0.0945280716,-0.1120574623,0.0530560985,0.2090707868,0.1939486712,-0.2207102031,0.1345973313,-0.0736984089,-0.2940365076,0.230275467,0.2727073729,0.3875804543,0.1529555172,0.0712210163,0.1755918711,0.7356215119,0.2542084754,0.0182265267,0.1526494622,0.2519257665,-0.0085175931,0.1782764047,-0.2468549311,0.1766627729,-0.1547312886,0.4875793159,0.0987572148,0.1508025378,-0.2450172007,-0.3062398434,0.0589261465,-0.2685538828,-0.2171592414,0.2424305975,-0.2287578285,0.0140708582,-0.0896598846,-0.4162377715,-0.2906380892,-0.0538840182,-0.0223103166,0.3469152749,-0.0776843429,0.1573176831,-0.3413406312,-0.0460453741,-0.4354320168,0.1478283256,0.0733894557,0.2878928185,-0.0697131529,-0.0244402513,-0.1107752398,0.1029282957,0.6152383089,-0.5185818672,-0.1225132644,0.1040606201,0.0839116201,-0.1375814974,-0.1016150191,-0.2601816058,0.0148680266,0.4297328889,0.3357182145,-0.3773824275,-0.2934282422,0.1744582057,-0.209812358,-0.0844757482,-0.1624628156,-0.2532831728,-0.2284482569,-0.1250564009,0.041775886,0.2297199517,0.2200999856,-0.0644015372,0.1653501093,-0.1073816791,0.2017074972,0.2185195386,0.0230326578,-0.0400125198,0.2556946874,0.1977338642,0.0395456105,-0.126898244,-0.3682380319,0.065940313,-0.0586392879,0.0672692284,0.2543282509,-0.0315840691,0.6494701505,0.0423284359,0.339296788,0.2956757545,-0.1710401773,0.0038366246,-0.2279187888,0.1694688797,0.2228393853,0.2263696939,-0.1590794176,-0.2618803084,0.5629377365,0.2996866107,-0.0713931546,0.451677829,-0.489803791,-0.3304511905,0.2624638379,0.0708127767,0.6956490874,-0.0626806095,0.2947659492,0.0793052763,-0.3640843928,0.6168707609,-0.3644067049,0.1506235749,-0.1062212214,-0.1969085783,0.0126754437,-0.0935866833,-0.073100701,0.2685155869,-0.3825459778,0.3077968657,0.4287261069,-0.2022626549,-0.0253601577,0.5634387732,0.0562331788,-0.3985061049,-0.158394143,0.0687965006,-0.308154881,0.0543649457,-0.1665422469,-0.2580900788,0.2257057577,0.1558324248,-0.3013501763,0.0667902306,-0.0352305546,0.217782408,0.2128637582,0.0152647002,0.2679021358,-0.1160597429,-0.0378508531,-0.1535062641,-0.0839715973,0.1355269402,-0.1933338046,0.037814483,0.0182684343,-0.1867404133,0.4988892376,-0.0624800324,-0.1185215339,-0.2142510414,0.2802072763,-0.4461526275,-0.1915701032,0.1475750506,0.1756841391,0.1982741207,-0.3049497604,0.5695934296,0.1094507724,0.0111564733,0.0341043584,0.2101195157,-0.2711587548,0.5002741218,-0.2811296582,0.0099636614,0.1048395857,0.025051849,-0.1126541123,-0.3857493699,-0.0307866223,0.3170596957,0.1385989636,0.008726703,0.0285298266,0.1123752072,-0.0650159344,-0.0270695928,-0.1449051052,-0.3513539732,0.1318763644,0.1800637096,0.1875754446,0.2329903096,-0.0289730728,-0.271537751,-0.1660395563,0.4481062591,0.0352604352,0.2099884599,-0.1294511408,0.3465990424,0.4091414511,0.548371017,-0.2595945895,0.0745584518,0.0553667061,0.1109805629,0.4592460096,-0.2511611879,0.0760797337,0.277148664,-0.124742046,0.0313541479,-0.3204190731,-0.153558135,-0.2497406304,0.1109318137,0.1044218987,0.0042010942,0.0040182625,-0.10565155,-0.113112919,-0.0842528418,-0.0002993329,-0.3816280663,-0.0595905259,0.1242031381,-0.0380664729,0.0650604889,0.0368797742,0.187790677,0.0234356858,-0.1025380045,0.3493964672,0.117238991,0.1081334949,-0.1778067946,-0.1632226706,0.0233437438,0.2428954393,0.1519824415,-0.0641244873,0.1286801547,-0.2415773422,-0.1206695288,0.236313045,0.1691904068,0.0174480267,-0.1027126759,0.1332758367,0.1149811521,-0.0952079147,-0.0351001285,0.2745020092,0.0782757252,0.1340378374,0.1436426342,0.098759681,0.4728011191,-0.2388040721,0.009598989,0.1430582255,-0.3520101905,0.1235264689,-0.1303716302,-0.0364451222,-0.0115565741,0.103236787,-0.0243816748,0.1411909014,-0.2700304985,0.2807745039,0.3052041829,0.0004441649,-0.0339137167,0.2237343639,0.0560695529,-0.1437354237,0.2464372069,0.0478809848,0.3264904022,0.0214848351,0.889824748,-0.259321481,-0.4453705251,0.109149985,0.427682668,0.0437286347,-0.2092781961,0.1428684592,0.1025017649,-0.3528324068,0.0732433274,0.0668142214,-0.2198761106,0.6501641273,-0.1044894755,0.2117462009,-0.1332384497,-0.1671050638,0.3940386772,0.1751196831,-0.065077439,0.0157078095,0.1436770707,-0.1233619153,-0.1789791733,0.4386436045,0.2128615081,0.0278087165,-0.4619535208,0.0628166422,0.0905681998,0.1469302624,-0.0395544432,0.1065973863,0.4349875152,-0.1517576724,0.2386019081,0.0137401717,-0.0534979217,-0.1947309673,0.0798115805,-0.4298866391,-0.3788129091,0.4788790643,0.0675856918,-0.1132399961,0.059434969,0.3338850737,-0.6175370812,-0.2591400146,0.3267904818,-0.0954336151,-0.0662726983,-0.1992470175,0.0728195831,0.0632103011,0.639778614,0.1772638112,0.3605995178,-0.3750834465,-0.1291012019,-0.6346368194,-0.1560516357,0.113924548,-0.1761644185,-0.2531372309,0.1707490236,-0.0470046327,0.3227478862,-0.209546715,0.0975830704,0.2715468407,0.0271078311,-0.4724882841,0.0400513299,0.0950052366,0.0479600318,0.2817948163,-0.0757132322,-0.191032365,0.0638209805,-0.0673609227,-0.0838575363,-0.1188783944,0.2494051307,0.17717053,0.06206128,0.0344523266,0.5532022715,0.0108181182,-0.161627695,0.0310947653,0.0285957325,-0.3770722449,-0.1337216496,-0.0821675062,0.2083973736,-0.3929455876,0.2529568672,-0.0076548737,-0.0252354406,0.119383648,0.0114970449,-0.2017443478,-0.0240462851,0.0840792283,0.2401800901,0.2028209418,0.7438048124,-0.0144339707,0.1731562763,-0.0818765759,-0.0374525413,0.0660283864,-0.1376398206,-0.247848019,-0.1356829554,0.0355518386,0.0620834529,0.1208925843,-0.6768579483,-0.3293687403,0.0451084785,-0.0008661266,0.0040006163,0.7409059405,0.0650401637,-0.1160142273,-0.1170460731,-0.3568641841,0.0334187448,-0.2515329421,0.1817403287,-0.3713240623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"The problem with changing `k` and `t` per sampling is that you'd have to somehow remember which examples you'd already returned while re-weighting the remaining examples based on the new `k` and `t`values. It seems possible but complicated (I can't really see a reason why you'd want to change the weighting of datasets after you constructed the multidataset).\r\n\r\nWouldn't it be convenient if it implemented the dataset interface? Then if someone has code using a single nlp dataset, they can replace it with a multitask combination of more datasets without having to change other code. We would at least need to be able to pass it into a `DataLoader`.\r\n\r\n","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":109,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nThe problem with changing `k` and `t` per sampling is that you'd have to somehow remember which examples you'd already returned while re-weighting the remaining examples based on the new `k` and `t`values. It seems possible but complicated (I can't really see a reason why you'd want to change the weighting of datasets after you constructed the multidataset).\r\n\r\nWouldn't it be convenient if it implemented the dataset interface? Then if someone has code using a single nlp dataset, they can replace it with a multitask combination of more datasets without having to change other code. We would at least need to be able to pass it into a `DataLoader`.\r\n\r\n","embeddings":[-0.0834357738,-0.3772905171,-0.0540124886,-0.0832781196,-0.2122467607,0.016944591,0.1334990859,0.0681641847,0.3124590814,-0.0252599195,-0.2377852947,0.3599816561,-0.1587046683,0.3497785032,0.3823591769,-0.4798976183,-0.0307444744,-0.1330458373,-0.3688006699,0.3823723197,0.0409010462,0.0993345529,-0.1226546094,0.0074622524,-0.2362718433,-0.1522186995,-0.3432965577,0.0929011106,0.1845614463,-0.0550318137,0.1963454932,0.4499551356,-0.0172803458,0.0523207374,-0.0001107237,-0.0873194411,0.1317604184,-0.1200200543,0.0498344786,0.0378307067,0.2165661305,-0.4117277861,-0.0697145388,-0.1880381107,0.1087680385,0.2151244581,0.183059305,0.0244122688,0.4155212045,-0.3210864365,0.1293699592,0.3268467784,-0.2219879478,0.2345312536,-0.0946439728,-0.1533048153,0.0941295028,0.3127886057,0.4916836321,-0.5477143526,-0.061179392,0.1415119916,0.0077215391,0.0596176833,0.1025726348,-0.0515166223,-0.3251841962,-0.2065102458,-0.3548207283,0.2940816879,-0.1811466068,-0.1491725594,-0.1987767816,-0.5856713653,0.0600595288,-0.0108269686,-0.2997102737,0.020461211,-0.149675414,-0.0127926776,-0.1295457035,-0.0983141139,0.0747852847,0.0990781561,0.3274924159,0.560352385,0.0938607454,0.183635056,0.3918946385,0.0773890615,-0.2122483999,-0.2460267693,0.2851282358,0.0963052288,-0.282671541,-0.3462053537,0.1630013734,-0.2639792562,0.2329585999,0.0578402653,0.167758882,0.1127669662,-0.137882337,0.2494637668,0.2821424901,0.0598511584,0.010651404,-0.1180342063,0.0149835404,-0.057337299,0.1585505754,0.0927682146,0.2161785364,-0.0449696481,-0.4050146937,-0.0826194361,-0.2754174471,0.1954955161,-0.2060203105,-0.5297424793,-0.0186265651,-0.114803344,0.0239045992,-0.1002877131,-0.3187436461,0.1601452827,-0.0565391108,0.1747000515,-0.1174471155,0.0728757977,-0.0856354162,0.0885189176,-0.4986366928,0.0251944605,0.2598644197,0.0800495073,0.162202552,0.1625870764,-0.0261631124,0.1506989747,0.3910187185,0.0596011877,0.0018361512,-0.0581313297,0.1590105444,-0.2949698567,-0.042152863,0.1823380589,-0.4246416986,-0.1548048556,-0.1469173878,-0.076041542,0.1608907133,0.0455429628,-0.244696945,-0.2981191278,-0.4465657175,0.7699325085,-0.0095635736,-0.1129112318,-0.1095577925,0.2079466134,-0.3065616786,-0.0800890848,0.2194667459,0.0320920646,-0.0190592427,-0.4033974111,-0.0422706418,-0.0371791124,-0.1039554253,0.2644662857,-0.2160887867,0.1813715249,0.0740556717,0.1739475131,0.2999924123,-0.4679256976,-0.0848970637,0.2344106138,-0.0360876322,0.3289677501,0.0849847868,0.1241306141,0.0519661009,-0.1715914905,0.1760984659,0.9709120393,-0.3548180759,0.0243716575,-0.1063789576,-0.3863515556,0.5253027081,0.2912895679,0.0707936659,-0.2695538402,-0.1268907189,0.004664849,0.0922796503,0.0185630433,0.0876701027,-0.0491235293,-0.330820024,0.0091974754,-0.2163795233,-0.1552149802,-0.3733976483,0.0538267419,-0.0618969314,0.3879180849,0.3986713886,-0.1182984784,0.2000267953,-0.3578913212,-0.0335234813,-0.2823462188,0.1489843577,-0.0139880786,0.0484355241,-0.220911622,-0.0495129675,0.2356532812,0.1350265741,-0.1555173248,-0.4074779451,0.3045088649,0.0201744139,-0.0015043471,-0.0542013943,0.5833072662,-0.1648832411,-0.1087331846,0.2087197006,0.3047441244,-0.3626170158,0.294107765,0.310453862,0.1920404434,0.158066541,0.0199261736,0.157711491,-0.0145369517,-0.0647680387,-0.1608451307,-0.0415621698,0.6416841745,-0.0483003706,0.3482078016,-0.0292414036,-0.0311008077,-0.1747988015,-0.1673587859,-0.2596576512,0.3309375048,0.2930160165,-0.2784144878,0.3624516129,-0.1274982095,-0.4090746045,0.1585979015,0.1983101666,0.0533816777,0.1220826581,0.0116764652,-0.0540280864,-0.3014111221,-0.1033004671,0.344779551,0.6520264149,0.2875113785,0.168321982,0.0216041952,-0.2816530168,-0.0925326347,0.0023980346,0.1340303868,-0.0298968572,0.182002008,0.1024383008,0.0314744525,-0.0856320933,-0.0643137246,0.1644387841,-0.2171367258,-0.03231683,0.0398653746,-0.178645879,-0.5128598213,-0.2909932137,0.0355958454,-0.3008249104,-0.3023713231,0.1804727167,0.0911055058,-0.2213292569,0.2116441429,0.2378734052,0.6204200983,-0.3974946439,-0.1298799217,-0.109520629,-0.3464152813,-0.1027458161,0.1299117357,0.3555285335,0.2745725214,0.4790076911,0.2562711835,-0.0584109835,0.1019947752,-0.4040745497,0.0262862705,-0.319781512,0.2502519786,0.1185360178,-0.1891649663,0.1525095254,-0.3981902003,0.0118647767,-0.1402017325,-0.0298847072,-0.0593073927,0.0118865576,0.1486317068,-0.202735275,-0.2027551234,-0.608305335,-0.4239632785,0.2456141859,-0.2116802186,0.0454282649,0.0518572964,-0.2142136395,0.1482136697,-0.1654004604,0.0154590076,-0.1877848804,-0.1026330888,-0.0060610902,-0.3906424344,-0.10505227,-0.5042768717,-0.1248725355,0.2693509758,-0.1049095318,-0.1020210385,-0.3205795288,0.05710008,-0.0957068577,0.0169347022,0.2777613997,0.1074365526,-0.2277356386,0.034652058,-0.069547154,-0.2141720951,0.2184725702,0.4109443128,0.387539953,0.1111993939,0.003796594,0.1425023228,0.7288064957,0.2042254061,-0.0062154955,0.2417244911,0.2230892926,0.0590062328,0.1169358939,-0.1488067657,0.2164951265,-0.1390655637,0.5181331635,0.1340827644,0.1449039429,-0.236360684,-0.2616962492,0.0015480308,-0.2062596083,-0.2486131638,0.2672935426,-0.2329824865,-0.0389779583,-0.0430797487,-0.3545137048,-0.3271301687,-0.0430239998,0.021406794,0.1693399996,-0.0096904757,0.1349156797,-0.4466553032,0.0073103006,-0.4875246286,0.1748704463,0.0355296433,0.2063129693,-0.0146246338,-0.0955250412,-0.1373222172,0.1053149179,0.4926844537,-0.5116747022,-0.0789994374,0.0813728124,0.0604245663,-0.1340132207,-0.1311506033,-0.3227549195,-0.0770327598,0.4415514171,0.3774554431,-0.4230467975,-0.302546382,0.2241525054,-0.1646210253,-0.0954800323,-0.1715964377,-0.211120531,-0.1963449866,-0.1546305269,0.0741317123,0.1433064342,0.2671214342,-0.1253154725,0.1903274804,-0.1081381962,0.184588179,0.1798727363,0.0158274733,-0.0441924222,0.353923291,0.1102185026,0.0706376582,-0.1312555671,-0.3319585621,0.1054884344,-0.1351999938,0.2050825357,0.226970464,0.0906750336,0.55300349,0.1143625826,0.3576849103,0.303411752,-0.0634749234,0.1254376918,-0.3138454854,0.2028866857,0.1758409739,0.2550442815,-0.106840767,-0.2142483294,0.5952863693,0.2874777913,-0.0887916759,0.3823327124,-0.5090103149,-0.3641580045,0.2897240222,0.0848763436,0.6568555832,-0.0126665691,0.3372671902,0.059179645,-0.3040446043,0.6199080944,-0.2759535015,0.1473553032,-0.1350039393,-0.2456523478,0.036676392,-0.0486415066,-0.0824146271,0.2677820921,-0.3512746394,0.3676718771,0.4438673556,-0.199815169,-0.1293602288,0.5204604864,-0.0869241655,-0.3249285519,-0.1141844913,0.0993852168,-0.2334738523,0.1165745333,-0.1201618016,-0.2505452633,0.2240793407,0.1357116401,-0.2270085365,0.0455454253,-0.0658814311,0.1369323581,0.2686853409,0.0302283708,0.1925261617,-0.2200118154,-0.0255269427,-0.1079611033,-0.1062542647,0.1295534819,-0.1636425555,0.0705537722,0.1392044425,-0.147597,0.519082129,0.0120293479,-0.1147722676,-0.1963238269,0.2693729401,-0.3825458884,-0.1779464036,0.0240984596,0.0924794972,0.1034710258,-0.3090551794,0.5627148747,0.1250997186,-0.0283493381,0.0616695583,0.1021076292,-0.2298944443,0.5123814344,-0.185657829,-0.0395511948,0.0541520454,0.0949078947,-0.1626537293,-0.304113239,-0.0404062979,0.241284892,0.1346213818,-0.0144532928,0.014386463,0.1648528725,-0.1703094691,0.0433053449,-0.075347051,-0.3425022066,0.1565356106,0.2247280926,0.2189420462,0.2962498069,-0.1110983267,-0.1854039878,-0.1225558072,0.4910309613,0.0918164775,0.2689321041,-0.1198960096,0.3225711286,0.3547474444,0.5900266171,-0.3123570681,0.0776665732,0.0638771132,0.1348430514,0.5099006295,-0.2374131083,0.0089693759,0.3305617571,-0.1103553325,0.0758995414,-0.329567939,-0.1768559963,-0.2944237292,0.0876587778,0.073034972,0.012625521,0.0238950588,-0.0342636518,-0.0450683087,-0.1670869738,0.0055859541,-0.4168511629,-0.1002965122,0.2000104785,-0.0610241666,0.064385049,0.1422737688,0.2284630984,0.0096813086,-0.1569224745,0.2438973635,0.1689839214,0.104251489,-0.1232447252,-0.1688495576,-0.0028996903,0.2162174284,0.0883054584,-0.1214594692,0.1905929595,-0.2052346021,-0.0906113759,0.3178473711,0.1873665154,0.0489359386,-0.1805133671,0.1527684629,0.166452691,-0.1125443429,0.0003929338,0.1596732289,0.1168988794,0.1586342156,0.1467077136,0.1457315534,0.3819443882,-0.1910196692,-0.0410075337,0.0876435041,-0.3214346766,0.1108077243,-0.0604326278,-0.0206132345,-0.0321245976,0.1805832833,-0.0632683486,0.205195114,-0.4029551148,0.2911435068,0.3381802738,-0.0088558197,0.0301897228,0.1225874871,0.0114028323,-0.133850351,0.3777519166,0.0574886911,0.3323088884,0.0518808179,0.926881969,-0.2629575133,-0.4481846392,0.0635863915,0.5313951373,0.029138539,-0.3214191794,0.0613894239,0.064780429,-0.3745323122,0.1602953672,0.011611741,-0.1540063918,0.6641734242,-0.1534091979,0.1933291703,-0.1681732088,-0.2260020822,0.2835280597,0.1881143749,-0.1497500539,0.0448607653,0.0780442283,-0.0641849935,-0.1243786141,0.5293443799,0.1735812575,-0.0133229159,-0.4142281711,-0.0032771649,0.1442958117,0.1632393748,-0.1211546287,0.1270753592,0.4776850045,-0.1713275015,0.291972816,0.0513655171,-0.1416739821,-0.1650521606,0.053725604,-0.4065135419,-0.411121428,0.4177231789,0.1373805106,-0.0414789766,0.0670821816,0.2855620384,-0.5498073101,-0.2214986831,0.2549922168,-0.173150748,-0.0709798485,-0.2230932117,0.09759368,0.0447586253,0.6479759216,0.2121252567,0.2520438433,-0.3549043238,-0.0999404117,-0.5510977507,-0.1398828924,0.1789738387,-0.1073023006,-0.2773841918,0.1640440971,-0.0369555913,0.2786957026,-0.1839829683,0.1812702566,0.211436525,0.0091554793,-0.4085699618,0.1516012251,0.0187746082,0.0416056551,0.3125755489,-0.1152503639,-0.1505112499,0.0213750657,-0.0266356226,-0.0452439263,-0.0816569924,0.24790667,0.1851310879,0.1656779647,0.0134810135,0.4840938449,-0.0781924352,-0.1265276223,-0.0207882188,-0.0746106058,-0.4193074107,-0.2067510486,-0.0996450856,0.2181218415,-0.2795772552,0.3174084425,-0.0091724303,-0.1115296111,0.152082935,0.0676351041,-0.1043665633,-0.0530077592,0.1152552292,0.0915914997,0.1921043247,0.7487504482,0.0661452711,0.0984084308,-0.0825172216,-0.1133482605,0.0889336765,-0.0693514422,-0.1878779083,-0.1619138867,0.0123314923,0.1088350937,0.003593985,-0.6421051621,-0.3090016544,0.0285696685,0.0407858826,0.0361567475,0.7093769312,0.0520841405,-0.1664980948,-0.123911716,-0.3487196863,0.0836584345,-0.3357553482,0.0660819933,-0.3758982718]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"A very janky (but working) implementation of `multitask_dataset.sample()` could be something like this:\r\n```python\r\nimport nlp\r\nimport torch\r\n\r\nclass MultiDataset():\r\n def __init__(self, *args, temperature=2.0, k=1000, maximum=None, scale=1):\r\n self.datasets = args\r\n self._dataloaders = {}\r\n for split in self._get_common_splits():\r\n split_datasets = [ds[split] for ds in self.datasets]\r\n mixing_rates = self._calc_mixing_rates(split_datasets,temperature, k, maximum, scale)\r\n weights = []\r\n for i in range(len(self.datasets)):\r\n weights += [mixing_rates[i]]*len(self.datasets[i][split])\r\n self._dataloaders[split] = torch.utils.data.DataLoader(torch.utils.data.ConcatDataset(split_datasets),\r\n sampler=torch.utils.data.sampler.WeightedRandomSampler(\r\n num_samples=len(weights),\r\n weights = weights,\r\n replacement=True),\r\n shuffle=False)\r\n\r\n def _get_common_splits(self):\r\n '''Finds the common splits present in all self.datasets'''\r\n min_set = None\r\n for dataset in self.datasets:\r\n if min_set != None:\r\n min_set.intersection(set(dataset.keys()))\r\n else:\r\n min_set = set(dataset.keys())\r\n return min_set\r\n\r\n\r\n def _calc_mixing_rates(self,datasets, temperature=2.0, k=1000, maximum=None, scale=1):\r\n '''Work out the weighting of each dataset based on t and k'''\r\n mixing_rates = []\r\n for dataset in datasets:\r\n rate = len(dataset)\r\n rate *= scale\r\n if maximum:\r\n rate = min(rate, maximum)\r\n if temperature != 1.0:\r\n rate = rate ** (1.0\/temperature)\r\n mixing_rates.append(rate)\r\n return mixing_rates\r\n\r\n def sample(self,n,split):\r\n batch = []\r\n for example in self._dataloaders[split]:\r\n batch.append(example)\r\n n -= 1\r\n if n == 0:\r\n return batch\r\n\r\n\r\ndef flatten(dataset,flatten_fn):\r\n for k in dataset.keys():\r\n if isinstance(dataset[k],nlp.Dataset):\r\n dataset[k] = dataset[k].map(flatten_fn,remove_columns=dataset[k].column_names)\r\n\r\n# Squad\r\ndef flatten_squad(example):\r\n return {\"source\": \"squad context: \" + example['context'] + \" question: \" + example['question'],\"target\":example[\"answers\"][\"text\"]}\r\nsquad = nlp.load_dataset(\"squad\")\r\nflatten(squad,flatten_squad)\r\n\r\n# CNN_DM\r\ndef flatten_cnn_dm(example):\r\n return {\"source\": \"cnn_dm: \" + example['article'],\"target\":[example[\"highlights\"]]}\r\ncnn_dm = nlp.load_dataset(\"cnn_dailymail\", \"3.0.0\")\r\nflatten(cnn_dm,flatten_cnn_dm)\r\n\r\nmultitask_dataset = MultiDataset(squad, cnn_dm)\r\nbatch = multitask_dataset.sample(100,\"train\")\r\n```\r\n\r\nThere's definitely a more sensible way than embedding `DataLoader`s inside. ","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":231,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nA very janky (but working) implementation of `multitask_dataset.sample()` could be something like this:\r\n```python\r\nimport nlp\r\nimport torch\r\n\r\nclass MultiDataset():\r\n def __init__(self, *args, temperature=2.0, k=1000, maximum=None, scale=1):\r\n self.datasets = args\r\n self._dataloaders = {}\r\n for split in self._get_common_splits():\r\n split_datasets = [ds[split] for ds in self.datasets]\r\n mixing_rates = self._calc_mixing_rates(split_datasets,temperature, k, maximum, scale)\r\n weights = []\r\n for i in range(len(self.datasets)):\r\n weights += [mixing_rates[i]]*len(self.datasets[i][split])\r\n self._dataloaders[split] = torch.utils.data.DataLoader(torch.utils.data.ConcatDataset(split_datasets),\r\n sampler=torch.utils.data.sampler.WeightedRandomSampler(\r\n num_samples=len(weights),\r\n weights = weights,\r\n replacement=True),\r\n shuffle=False)\r\n\r\n def _get_common_splits(self):\r\n '''Finds the common splits present in all self.datasets'''\r\n min_set = None\r\n for dataset in self.datasets:\r\n if min_set != None:\r\n min_set.intersection(set(dataset.keys()))\r\n else:\r\n min_set = set(dataset.keys())\r\n return min_set\r\n\r\n\r\n def _calc_mixing_rates(self,datasets, temperature=2.0, k=1000, maximum=None, scale=1):\r\n '''Work out the weighting of each dataset based on t and k'''\r\n mixing_rates = []\r\n for dataset in datasets:\r\n rate = len(dataset)\r\n rate *= scale\r\n if maximum:\r\n rate = min(rate, maximum)\r\n if temperature != 1.0:\r\n rate = rate ** (1.0\/temperature)\r\n mixing_rates.append(rate)\r\n return mixing_rates\r\n\r\n def sample(self,n,split):\r\n batch = []\r\n for example in self._dataloaders[split]:\r\n batch.append(example)\r\n n -= 1\r\n if n == 0:\r\n return batch\r\n\r\n\r\ndef flatten(dataset,flatten_fn):\r\n for k in dataset.keys():\r\n if isinstance(dataset[k],nlp.Dataset):\r\n dataset[k] = dataset[k].map(flatten_fn,remove_columns=dataset[k].column_names)\r\n\r\n# Squad\r\ndef flatten_squad(example):\r\n return {\"source\": \"squad context: \" + example['context'] + \" question: \" + example['question'],\"target\":example[\"answers\"][\"text\"]}\r\nsquad = nlp.load_dataset(\"squad\")\r\nflatten(squad,flatten_squad)\r\n\r\n# CNN_DM\r\ndef flatten_cnn_dm(example):\r\n return {\"source\": \"cnn_dm: \" + example['article'],\"target\":[example[\"highlights\"]]}\r\ncnn_dm = nlp.load_dataset(\"cnn_dailymail\", \"3.0.0\")\r\nflatten(cnn_dm,flatten_cnn_dm)\r\n\r\nmultitask_dataset = MultiDataset(squad, cnn_dm)\r\nbatch = multitask_dataset.sample(100,\"train\")\r\n```\r\n\r\nThere's definitely a more sensible way than embedding `DataLoader`s inside. ","embeddings":[-0.083420068,-0.388119489,-0.0588925034,-0.0579484887,-0.2545415759,0.0694044679,0.1949307323,0.0514100231,0.3405600786,-0.148264572,-0.2024349868,0.3717310727,-0.2394476831,0.3475413024,0.3872979283,-0.4599964023,0.0241373293,-0.2675699592,-0.2698787153,0.3728720248,0.0622705556,0.0881300569,-0.1811579019,0.0803664848,-0.2539775372,-0.16344814,-0.349388212,0.1046217084,0.1572879404,-0.013107325,0.212037161,0.3812626898,-0.0404960923,0.1317787915,-0.000113172,-0.0745764598,0.2105259895,-0.0762380138,0.0865641236,-0.0652449355,0.2296273559,-0.4317651391,0.0367130935,-0.1433316171,0.140597403,0.151721701,0.1642498076,-0.0074314806,0.3844811022,-0.2736278772,0.1117818952,0.2874096334,-0.1595599204,0.2645365596,-0.108694531,-0.0474401824,0.0490127094,0.2619538009,0.6061142683,-0.6340856552,-0.0141250845,0.1008240655,-0.0498053692,0.0319435559,0.1234847605,-0.0822734758,-0.3827714324,-0.235299632,-0.3922517896,0.2454876453,-0.1774665415,-0.194536373,-0.2274479568,-0.5580018759,0.0062529175,0.0419942886,-0.3048384488,0.0502216555,-0.1998161525,-0.0673157945,-0.1375676841,-0.0884879827,0.0118301101,0.173216626,0.3156492412,0.5342768431,0.109470658,0.1746337265,0.3422690332,0.0428313501,-0.138672933,-0.2869670689,0.3420218825,0.1264587343,-0.371249944,-0.3075858057,0.18244946,-0.26114887,0.2541568875,0.1146515235,0.1163638383,0.0927190036,-0.124762021,0.2851980329,0.3020448387,0.0150043061,-0.0813578963,-0.0808274001,0.0235543028,-0.1328221709,0.2297943234,0.1204114556,0.1645004302,-0.0305917617,-0.408781141,-0.0694567934,-0.1612365395,0.1976407468,-0.2383099049,-0.4846026897,-0.1420591176,-0.0831547529,0.0792373419,-0.0919273868,-0.2979574502,0.1001808122,-0.0730477124,0.2468690872,-0.1000253484,0.071838066,-0.0563448779,0.0983509347,-0.43944031,-0.069917284,0.3233097494,0.1422465295,0.1164548993,0.1751130968,-0.0169182457,0.1301406473,0.4420316517,-0.001940197,0.0209581833,-0.0140591869,0.1952323765,-0.3251471817,0.0048835138,0.2275522649,-0.3726580441,-0.1428907067,-0.1280905753,-0.1560909748,0.1126976386,0.0356523506,-0.2359916568,-0.2588003576,-0.4986609519,0.6829746366,-0.0290034935,-0.0418765731,-0.1579708308,0.2522134483,-0.248444587,-0.1251500994,0.2169201672,0.0657935813,-0.0024442901,-0.3119379878,-0.1297095269,-0.0817762762,-0.0371012688,0.2253071517,-0.1600787044,0.2730769217,0.0468321294,0.2083573341,0.2641534507,-0.5000427961,-0.0438659117,0.2669200301,-0.042176865,0.4183540344,0.0484382063,0.2128078043,0.0348049775,-0.1617123932,0.2985625565,1.0328638554,-0.3860523999,0.0045572058,-0.0949142426,-0.3651699126,0.5233176351,0.3392055035,0.0907386392,-0.319514513,-0.1159666032,-0.053001862,0.1005862877,0.0858918875,0.0734527931,-0.0722969472,-0.3650389612,0.0597842149,-0.2487353981,-0.1938261539,-0.3421436846,0.0303587671,0.1119065508,0.2797664702,0.2486675233,-0.1320004463,0.2229710817,-0.2549453378,-0.1298151761,-0.3390695751,0.1418402195,-0.059790995,0.0374792144,-0.2510701716,0.0014998317,0.3411405981,0.127417773,-0.150571242,-0.4737476707,0.2840258777,-0.0559381731,0.0393687151,-0.0220185854,0.6338954568,-0.1646690965,-0.0881074071,0.1917863786,0.2877437174,-0.2685473859,0.2172980011,0.3182614446,0.1692703664,0.1466194838,-0.0228321888,0.0941672623,0.0621032529,-0.0240913481,-0.2137680501,-0.0874792933,0.5368933678,-0.1505941004,0.4393788874,-0.0656994134,0.0612126663,-0.1688978076,-0.017659124,-0.1878746748,0.4131905138,0.3052524924,-0.1949422061,0.3491775095,-0.1831379384,-0.3764329255,0.1571003199,0.0892982632,0.0298302621,0.0982699543,-0.0435948223,-0.0516432226,-0.2501394153,-0.0670482814,0.3425808549,0.6638852358,0.2629099488,0.1862417758,0.0443923548,-0.2409894019,-0.0611293316,0.0086076315,0.174716562,-0.0858282447,0.1637564301,0.055454623,0.0332644433,0.0495144911,-0.1321865767,0.1540714651,-0.1827156693,-0.076834254,0.0934861749,-0.2060637176,-0.5439799428,-0.1966896802,0.0224109758,-0.2867782414,-0.2818977237,0.0911154896,0.0946393162,-0.1386158019,0.1551924646,0.213478446,0.5341400504,-0.3972025514,-0.2968822122,-0.0215749629,-0.343986541,-0.1300344914,0.1255968511,0.4387215674,0.3332568407,0.519888401,0.3033445776,-0.089452751,0.0479334258,-0.3372718096,0.0109446943,-0.2858090699,0.2852503359,0.1332926303,-0.2421933711,0.1681117564,-0.3485575914,-0.0252340175,-0.0736944824,-0.0894651711,-0.0388505012,-0.0026784644,0.0700024292,-0.1797032952,-0.1755022854,-0.6369875669,-0.4389748871,0.2991121709,-0.2640550435,0.0913328007,-0.0168008823,-0.2485507578,0.185419023,-0.1101642326,0.0057686241,-0.1777974665,-0.0930138007,0.0267334562,-0.3839551508,-0.1826389134,-0.3830832541,-0.1138103679,0.2173306048,-0.0198789537,-0.0235891305,-0.2534296811,0.1055404395,-0.0667821541,0.0601394922,0.2171058506,0.1514701694,-0.2382515669,0.1347248256,-0.0588868298,-0.2346391976,0.2361765355,0.2706868649,0.4174253345,0.1683673263,0.0401896052,0.2072385997,0.6981121302,0.248654142,0.0284307785,0.14254798,0.2371987253,-0.0051188325,0.1888375878,-0.2463108897,0.2186690867,-0.2162004709,0.483057946,0.028078543,0.1475925744,-0.2197672129,-0.2751479745,0.0053660101,-0.2993879318,-0.1995485872,0.2212370783,-0.2730290592,0.0116015831,-0.1181709841,-0.347112745,-0.2714881003,-0.0367682166,0.0135527961,0.3051335514,-0.0817406178,0.1570122093,-0.3513635397,-0.0490090698,-0.4496106207,0.1521967947,0.041846022,0.2807392776,-0.0617478192,-0.0609675013,-0.103979975,0.0956170186,0.5847178698,-0.5166597962,-0.0955330655,0.1287488192,0.1254887581,-0.1312012672,-0.1191443652,-0.2858881652,0.0736256242,0.4597921073,0.3409377635,-0.3444131613,-0.2527537644,0.2195505798,-0.1727743447,-0.0575932898,-0.1876332611,-0.2361873984,-0.2354426831,-0.0820688307,0.0553614907,0.228926301,0.2281401306,-0.1115577444,0.1948904246,-0.1273253709,0.2405482531,0.2175449133,-0.0025140126,-0.024933286,0.1810740829,0.1787148565,0.0449776538,-0.1487914622,-0.3022294939,0.0906487554,-0.0572759099,0.1108094975,0.2247030139,-0.0207430739,0.6292212605,0.0477002971,0.3185665011,0.2778654993,-0.1944342703,-0.0029815361,-0.2579374611,0.2018365562,0.2203503251,0.2621208131,-0.1164098382,-0.2639241815,0.5500631928,0.3211113513,-0.0659314916,0.4420499206,-0.5858775973,-0.3123213053,0.2433002144,0.0673038289,0.6210483313,-0.0111364303,0.2869629562,0.0889558047,-0.3830381334,0.5410959125,-0.3828496635,0.1639021188,-0.0822685584,-0.1536532044,0.0152164884,-0.0960355029,-0.0407184735,0.2526595592,-0.3295798004,0.3504186273,0.412070334,-0.2068225294,-0.0675337315,0.5238789916,-0.0139637683,-0.35267663,-0.042019818,0.0823407248,-0.2836205363,0.1062170491,-0.1444061548,-0.2439012676,0.2170099169,0.1585817337,-0.3457738161,0.0826077983,-0.0359137505,0.2272751182,0.2137261629,-0.0362466015,0.2051754147,-0.1772377938,-0.0749314725,-0.1512720585,-0.0366719365,0.1263705343,-0.1463217139,0.0401932746,0.0808702409,-0.2181075215,0.45674631,-0.0464311279,-0.1222217903,-0.1997517645,0.2868718505,-0.422288835,-0.1803593189,0.1595398486,0.1727862507,0.252602309,-0.3115997016,0.5563429594,0.1483858377,-0.0102235675,0.0390546881,0.1805767417,-0.2604192495,0.5079048872,-0.3668675125,0.0146508515,0.0897602215,0.0209121257,-0.1365335137,-0.3640401959,-0.0482435934,0.3149688244,0.1139809862,0.0043769251,0.0660130903,0.1207532212,0.0136683667,-0.0403767377,-0.1621457338,-0.3431298137,0.1165363565,0.1604316384,0.1777598709,0.187232554,-0.0499622263,-0.2939020097,-0.1633794457,0.4305419922,0.0440779403,0.2024307251,-0.1349615157,0.3518213332,0.3814522326,0.4948146939,-0.2580099404,0.0770569295,0.0950552598,0.0960406139,0.4429807067,-0.2144621909,0.052082289,0.2533465326,-0.1187752411,0.0298748594,-0.3219685555,-0.1579003781,-0.2378179878,0.0977488756,0.0700351,0.0175647791,0.0414376743,-0.1024015844,-0.0886309743,-0.0767530203,-0.0111732064,-0.3248333037,-0.0651867315,0.147129342,-0.1090827733,0.0478110127,0.0703791827,0.1974060088,-0.0022588368,-0.0873296782,0.2813520133,0.1235958338,0.1836776584,-0.1655916125,-0.1435883045,-0.0252527464,0.2127145529,0.14612028,-0.0713942274,0.1613968462,-0.2196006477,-0.1550362408,0.253549844,0.2257743031,0.0245937537,-0.1096903831,0.1470750868,0.1197553426,-0.1243220717,-0.0021175796,0.2085060775,0.0980359688,0.1171832755,0.1034610718,0.1028523967,0.4659965932,-0.3051425219,0.0313752815,0.1053305715,-0.3483788073,0.1158024669,-0.0827245712,-0.0022729961,0.0024447199,0.1037559584,-0.0023494894,0.1406650096,-0.2328876555,0.3303033412,0.3737260997,0.0639894903,0.0070393453,0.171013996,0.0021038989,-0.116316624,0.2337644994,0.0202161819,0.3528516293,0.0363882706,0.8677213192,-0.3164050579,-0.4453215301,0.1056334674,0.3968662918,0.0561193675,-0.2312902212,0.1447217911,0.1036296636,-0.3610671163,0.0887256786,0.0692591742,-0.1786099225,0.6680483222,-0.1202966869,0.1674935073,-0.1809268594,-0.2223655283,0.3528714478,0.1628759205,-0.1073769778,-0.0356573015,0.1072134003,-0.1029248759,-0.1998307705,0.4940884411,0.2465527654,0.0487529635,-0.5133512616,0.0119723752,0.0694675446,0.1574361771,-0.0794642419,0.1021791697,0.4472756684,-0.1771998703,0.2422115803,0.0154473549,-0.0577622615,-0.1007939875,0.0960222781,-0.4450981617,-0.3861228824,0.4605053365,0.071214892,-0.1043768749,0.0653751194,0.3186513484,-0.6325953007,-0.2643204033,0.3530026078,-0.0863581374,-0.056383606,-0.2091519535,0.0816375241,0.0421116687,0.601259768,0.1831813902,0.3498593569,-0.3703469634,-0.1162075102,-0.6050462127,-0.197711274,0.0519247763,-0.1589618176,-0.2622522116,0.1496835202,-0.0680054054,0.3262112439,-0.2323730141,0.0937545747,0.2632222474,0.106297791,-0.4595507085,0.0564988814,0.1048321277,0.0307256542,0.3046694398,-0.100577645,-0.2276275903,0.0465229414,-0.0557563044,-0.061588034,-0.1410767585,0.2911343575,0.1857735068,0.08580672,0.0254511312,0.5246701837,-0.0034386239,-0.145720616,-0.0257828124,0.0549957342,-0.3764570057,-0.1225063205,-0.0945232734,0.1602575779,-0.3817418516,0.2723015845,0.01171041,-0.0452997014,0.1185203716,0.0029898011,-0.1963092834,-0.0308386851,0.0635424405,0.2315344214,0.1789814681,0.7314289212,-0.025483055,0.1557292938,-0.0754501373,-0.0353113972,0.0534503646,-0.1333976686,-0.2366469502,-0.2042827904,0.0666476637,0.059699662,0.1490631402,-0.6794454455,-0.3367588222,0.0338811465,-0.0028483721,0.0607693829,0.7086068392,0.079132475,-0.1290808767,-0.1146334857,-0.3271324039,-0.0003094555,-0.258300066,0.2029402852,-0.3788251579]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"There is an interesting related investigation by @zphang here https:\/\/colab.research.google.com\/github\/zphang\/zphang.github.io\/blob\/master\/files\/notebooks\/Multi_task_Training_with_Transformers_NLP.ipynb","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":10,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nThere is an interesting related investigation by @zphang here https:\/\/colab.research.google.com\/github\/zphang\/zphang.github.io\/blob\/master\/files\/notebooks\/Multi_task_Training_with_Transformers_NLP.ipynb","embeddings":[-0.0616743974,-0.4030636549,-0.0587598681,-0.05660294,-0.2403090745,0.0958516002,0.2449350208,0.0309040677,0.2734928131,-0.1538774818,-0.2086921036,0.2724952698,-0.1859995276,0.3055661321,0.4009512365,-0.4570112824,0.0277175661,-0.1648355424,-0.2883341312,0.4265743494,0.0288353749,0.1239134148,-0.1820209026,0.0948675349,-0.3265630007,-0.1581454128,-0.354532361,0.0755940229,0.1647677273,0.0049198614,0.2487198412,0.3566808999,-0.0168814994,0.1181095615,-0.0001134201,-0.0663655996,0.2123865932,-0.0603531525,0.1185644716,-0.0508664884,0.1217864752,-0.3454694152,0.0591727905,-0.1294933259,0.1822126359,0.1696784496,0.2258604765,0.1299049705,0.3747032583,-0.3169631362,0.10402935,0.3135300279,-0.1954854429,0.2895288765,-0.0791201293,-0.0435593724,0.0561123602,0.2353991717,0.6586665511,-0.5678440332,-0.0458429717,0.0517681353,0.0316256508,-0.0171265081,0.1221075282,-0.0569053218,-0.3396398723,-0.2484251708,-0.4055341184,0.1941120476,-0.1615510583,-0.1441366076,-0.2304933071,-0.5225094557,-0.0265979227,0.0680885762,-0.288040787,0.0339838192,-0.1639680117,-0.051234033,-0.1768516004,-0.09871874,-0.0358006135,0.1450547427,0.3266830444,0.6169630885,0.0902064145,0.1579366028,0.3712649941,0.0574238896,-0.109149754,-0.2845401764,0.3061270714,0.1165406778,-0.3882130682,-0.379874438,0.1458889693,-0.24857077,0.2148147225,0.1164527833,0.076679118,0.106009908,-0.1570290029,0.2948799729,0.2936117351,0.0914277881,-0.0904736966,-0.0444632433,0.0490958579,-0.1249789223,0.2062382102,0.1151304916,0.1839582622,0.0279885437,-0.4103009403,-0.1430208534,-0.1959688812,0.1602181792,-0.227739349,-0.5790047646,-0.0626430064,-0.0823180676,0.1086215898,-0.085670419,-0.2873387337,0.1323284805,-0.0455097668,0.2132026255,-0.0817604363,0.1001761779,-0.0488350764,0.0953395292,-0.4305899441,-0.0648703054,0.2955386937,0.1892100126,0.1014705077,0.142480284,-0.0273699667,0.1137672514,0.4328672588,-0.0001846436,0.01068608,0.0455161668,0.1572307199,-0.3379606605,-0.0552661568,0.1824395955,-0.3637215197,-0.1357915848,-0.1641555578,-0.1715346575,0.1166049615,0.0326385796,-0.2202457935,-0.2527571619,-0.4776314199,0.6450725794,-0.036172837,-0.0821385831,-0.160834834,0.2738803327,-0.2505134344,-0.0665800646,0.1511397958,0.0295407921,0.1012985408,-0.3903351426,-0.0444990881,-0.0757790878,-0.0589384958,0.3200056553,-0.2386002839,0.236978367,0.0766619369,0.1686909497,0.2585593164,-0.4714216888,-0.0701742321,0.3113951981,-0.0198390689,0.3610929549,0.0476374067,0.262045145,0.0655240566,-0.2042252123,0.2399825603,1.0404888391,-0.3620851934,0.0202352945,-0.0795727372,-0.424505502,0.5248884559,0.3485018313,0.0983812883,-0.2705312073,-0.1320813894,0.1150599271,0.0812662989,0.0761163458,0.083665207,-0.0614857785,-0.4083064497,-0.0046366788,-0.1817522049,-0.1656530648,-0.4069360793,-0.0049123592,-0.054383859,0.3123656809,0.2551773787,-0.110356614,0.2680911124,-0.281104207,-0.141548574,-0.3324490488,0.1194871664,-0.0590091869,0.0009686836,-0.2970698476,-0.0254777949,0.336435914,0.1426414996,-0.1405826211,-0.4955830574,0.303301841,-0.124328129,0.0734577328,-0.065977931,0.6147111058,-0.1923001558,-0.0863029286,0.1741208881,0.2798771858,-0.3188383281,0.2750887275,0.3764780164,0.3233194053,0.2231364846,-0.0793165043,0.111340113,0.0613170378,-0.0031405976,-0.2456864715,-0.1041324809,0.4925737977,-0.1519477516,0.435819149,-0.0091121467,0.0052541937,-0.1739591509,-0.0722342804,-0.2461475432,0.4230655134,0.3080744445,-0.1753254831,0.3175743222,-0.0763885677,-0.3401378095,0.0892913714,0.1764571369,0.0118823554,0.072868675,-0.0241384562,-0.0560488887,-0.3007332683,-0.1146625355,0.3404181898,0.6377299428,0.2435893863,0.1389026046,0.0134506729,-0.2340434641,-0.0864138305,0.0018437379,0.2342443168,-0.0257096495,0.1530040354,0.0587146133,0.0652816743,0.0276165791,-0.2019504607,0.1847080737,-0.2208680063,-0.0306528006,0.1013807282,-0.1632596999,-0.5573260784,-0.2678852677,-0.0340885445,-0.3232137561,-0.2626253963,0.0968863592,0.0567234792,-0.1561303437,0.1783869714,0.2043695748,0.5870597363,-0.4230824113,-0.2379776835,-0.0855724365,-0.2905391455,-0.0971932113,0.1199990958,0.4176603258,0.2810089588,0.4487708211,0.283652246,-0.0683798119,0.0942869261,-0.3813302517,0.0358224362,-0.2395788431,0.2595030069,0.0811144337,-0.2681236863,0.185184747,-0.4037990272,-0.0517439209,-0.0948867798,-0.0458319969,-0.0314148776,-0.0303108748,0.0510454699,-0.1934244484,-0.1882743239,-0.6345367432,-0.4523611367,0.2851805091,-0.273970753,0.1010354981,0.0791220143,-0.308419019,0.1990379989,-0.1091157496,-0.0056125941,-0.1172629148,-0.0509009473,0.0328124054,-0.3275642395,-0.1472611576,-0.401396811,-0.130066514,0.1957086921,-0.0655363053,0.0053202845,-0.1632416844,0.0780128613,-0.1281579137,0.1014050767,0.1947382092,0.1653583944,-0.2509833574,0.1348156631,-0.0238068961,-0.2310278863,0.2159998864,0.3495199978,0.4436815381,0.1892483681,-0.0263896324,0.1504189223,0.7141780853,0.2073559761,0.0048705046,0.1252636611,0.3148053885,-0.0379681401,0.1393802464,-0.2366067618,0.1670671403,-0.2115206271,0.4796368778,0.0938214511,0.1716295034,-0.2493775934,-0.3024826646,-0.0095374854,-0.2623519599,-0.1922474355,0.2359436154,-0.2367416471,0.0271204915,-0.087258175,-0.4234155715,-0.2225532383,-0.0434591211,-0.0067454018,0.3170894086,-0.0763104632,0.1279891282,-0.2881049812,-0.0986669809,-0.4289290309,0.1018759981,0.0690591782,0.266923666,-0.0626756176,-0.036089465,-0.0874618739,0.1278098971,0.5904068947,-0.442161262,-0.1078934222,0.0869421214,0.0703480244,-0.0597422384,-0.109522,-0.2924712598,-0.0295463223,0.4868549109,0.3915483356,-0.3377918303,-0.2490967214,0.2290728986,-0.2497049123,-0.0975541472,-0.2080566883,-0.2104599029,-0.1802069992,-0.1537937671,0.0447012186,0.2041049898,0.1868896484,-0.1296900213,0.1564889103,-0.0599087141,0.2402950078,0.1931001395,0.0075821243,-0.0978921726,0.1972799301,0.1173356101,0.0306627862,-0.2138191164,-0.3299969435,0.0592623688,0.0126416581,0.1121440381,0.3195419312,-0.0097195813,0.6160163879,0.0086463699,0.3521351814,0.2845170796,-0.0804285035,-0.0098592518,-0.2348027229,0.149761945,0.2129637301,0.2599853575,-0.1352886111,-0.1688059419,0.5758923292,0.2549642622,-0.085834384,0.3742805123,-0.5928196907,-0.3088377416,0.3361850977,0.0145976748,0.6599773765,-0.0385494418,0.304582119,0.0127965771,-0.3891928196,0.5857066512,-0.3749813139,0.1738288701,-0.0839255378,-0.1905792356,-0.0017030285,-0.1003019139,-0.0320342965,0.2585989833,-0.3913491368,0.3050724864,0.479696095,-0.1712929457,-0.0446745642,0.5438808799,0.0356936827,-0.3763020933,-0.0648204386,0.0598305427,-0.2854119241,0.1334889382,-0.1878360659,-0.2497649193,0.2231229246,0.1520056576,-0.2862971425,0.1538768411,0.0434885472,0.274836421,0.2645527124,0.0646805838,0.252843082,-0.1997053474,-0.1152665243,-0.1769365519,-0.084464699,0.1474606097,-0.1709125489,0.0535378531,0.0692555755,-0.190402776,0.5116901398,-0.0562890507,-0.1437815428,-0.2202372104,0.258553952,-0.4410494566,-0.1548662186,0.1606904864,0.1814071387,0.249714911,-0.2973168194,0.583029151,0.0841049403,0.0041677626,0.0449061096,0.1660609841,-0.2608247399,0.5072105527,-0.2971101701,-0.0028438803,0.0934946761,-0.0195504297,-0.0889348313,-0.3316812813,-0.0394355506,0.2780260444,0.136329487,0.0218856204,0.0055065551,0.1021276787,-0.0896360576,-0.0382074714,-0.1375712603,-0.3460658193,0.120507732,0.1936542541,0.2100903839,0.2285992056,-0.11580614,-0.2782965302,-0.1955383718,0.4550693929,0.0231611878,0.2288008332,-0.1242035702,0.3605649471,0.4379479289,0.5447672606,-0.2526240647,0.0478407033,-0.0109418603,0.1353413314,0.4404592216,-0.2815008163,0.0584718846,0.2853812873,-0.1464003325,0.0317675993,-0.3121000528,-0.1531744003,-0.2232525647,0.1097755879,0.1450815648,0.0180251598,-0.0554246008,-0.0708514079,-0.0540485606,-0.0630379692,-0.0109124565,-0.3755191565,-0.041020859,0.1647199392,-0.0544097647,0.0309705008,0.0477135517,0.2505131066,0.0340463668,-0.1114378273,0.3197989762,0.1281994581,0.1142905727,-0.1540795416,-0.2206783295,0.0015891988,0.2678453624,0.0931707844,-0.037989445,0.1760036051,-0.2167023271,-0.1696641147,0.1953092366,0.1496402174,0.0358050726,-0.139881745,0.0652608201,0.1094888598,-0.082877256,-0.0195493493,0.2185115516,0.0935390666,0.0729180053,0.1068008617,0.1648506224,0.454135865,-0.1977529377,0.0428820141,0.1544176787,-0.3959305584,0.1385396272,-0.1460273415,-0.0127816033,0.0313375853,0.1762670726,-0.0361696519,0.1695079654,-0.355047673,0.3259340525,0.3012701273,-0.0374844223,0.0380030349,0.2003496885,0.053239882,-0.2079365551,0.2384728193,0.0069530201,0.3030003011,0.0602162816,0.8306283951,-0.2528022528,-0.3663693964,0.1211216003,0.4441637397,0.0292005893,-0.243478179,0.1878977418,0.1554083526,-0.3547571301,0.0878802612,0.0710917041,-0.1365205497,0.6947050691,-0.1051899642,0.1230030581,-0.1135356352,-0.164108634,0.3785146177,0.1365296841,-0.1037326381,0.0043341485,0.1350303888,-0.0628346577,-0.1925319582,0.5210363865,0.233766973,0.0283584632,-0.4171896875,0.0069070421,0.1022468731,0.1276975721,-0.0602542832,0.1511264741,0.4607639909,-0.1426149905,0.1898851991,0.0111163985,-0.0546723083,-0.1406684369,0.0327074416,-0.3872248828,-0.3732790947,0.4672935903,0.0826261863,-0.1693595648,0.0787807256,0.3624109626,-0.6424055099,-0.2104563862,0.3071279228,-0.1780788153,-0.0925889537,-0.1932106167,0.0701412931,0.0435179695,0.6402999163,0.2050799876,0.3947570622,-0.3868149519,-0.120702371,-0.5609939694,-0.19189623,0.1852669716,-0.0635863394,-0.2498123497,0.1566394418,-0.1279915422,0.2554942369,-0.2307877541,0.0611648299,0.2218455225,0.016800575,-0.3991873264,0.1471133083,0.1339996904,0.0328808725,0.2699570358,-0.0735298991,-0.1791971028,0.1183224022,-0.0883278623,-0.0971778259,-0.1192633733,0.2889887691,0.1944593042,0.0610203221,0.0470021553,0.5264805555,-0.0267401095,-0.1453046799,0.0421751589,-0.0228772294,-0.3446043432,-0.2052574307,-0.0681848153,0.1870582104,-0.3782590628,0.2815651298,0.0116288373,-0.153806597,0.0678053126,-0.0044928929,-0.1987647414,-0.0877109617,-0.0218649134,0.2235018015,0.224001646,0.680414021,0.0410998911,0.0664435402,-0.0857363492,-0.0106835598,0.0228544511,-0.068418324,-0.2506635785,-0.1725943536,0.0283702109,-0.0149245486,0.1028400734,-0.7300102115,-0.3499115705,0.0209478997,0.01642951,0.0493146926,0.72006464,0.0870043933,-0.0887091979,-0.1200222149,-0.2801032066,0.0200268254,-0.2681728899,0.1370043755,-0.3643938899]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"Good spot! Here are my thoughts:\r\n\r\n- Aside: Adding `MultitaskModel` to transformers might be a thing to raise - even though having task-specific heads has become unfashionable in recent times in favour of text-to-text type models.\r\n- Adding the task name as an extra field also seems useful for these kind of models which have task-specific heads\r\n- There is some validation of our approach that the user should be expected to `map` datasets into a common form.\r\n- The size-proportional sampling (also called \"Examples-proportional mixing\") used here doesn't perform too badly in the T5 paper (it's comparable to temperature-scaled mixing in many cases but less flexible. This is only reasonable with a `K` maximum size parameter to prevent very large datasets dominating). This might be good for a first prototype using:\r\n ```python\r\n def __iter__(self):\r\n \"\"\"\r\n For each batch, sample a task, and yield a batch from the respective\r\n task Dataloader.\r\n\r\n We use size-proportional sampling, but you could easily modify this\r\n to sample from some-other distribution.\r\n \"\"\"\r\n task_choice_list = []\r\n for i, task_name in enumerate(self.task_name_list):\r\n task_choice_list += [i] * self.num_batches_dict[task_name]\r\n task_choice_list = np.array(task_choice_list)\r\n np.random.shuffle(task_choice_list)\r\n\r\n dataloader_iter_dict = {\r\n task_name: iter(dataloader) \r\n for task_name, dataloader in self.dataloader_dict.items()\r\n }\r\n for task_choice in task_choice_list:\r\n task_name = self.task_name_list[task_choice]\r\n yield next(dataloader_iter_dict[task_name]) \r\n ```\r\n We'd just need to pull samples from the raw datasets and not from `DataLoader`s for each task. We can assume the user has done `dataset.shuffle()` if they want to.\r\n\r\n Other sampling methods can later be implemented by changing how the `task_choice_list` is generated. This should allow more flexibility and not tie us to specific methods for sampling among datasets.\r\n","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":264,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nGood spot! Here are my thoughts:\r\n\r\n- Aside: Adding `MultitaskModel` to transformers might be a thing to raise - even though having task-specific heads has become unfashionable in recent times in favour of text-to-text type models.\r\n- Adding the task name as an extra field also seems useful for these kind of models which have task-specific heads\r\n- There is some validation of our approach that the user should be expected to `map` datasets into a common form.\r\n- The size-proportional sampling (also called \"Examples-proportional mixing\") used here doesn't perform too badly in the T5 paper (it's comparable to temperature-scaled mixing in many cases but less flexible. This is only reasonable with a `K` maximum size parameter to prevent very large datasets dominating). This might be good for a first prototype using:\r\n ```python\r\n def __iter__(self):\r\n \"\"\"\r\n For each batch, sample a task, and yield a batch from the respective\r\n task Dataloader.\r\n\r\n We use size-proportional sampling, but you could easily modify this\r\n to sample from some-other distribution.\r\n \"\"\"\r\n task_choice_list = []\r\n for i, task_name in enumerate(self.task_name_list):\r\n task_choice_list += [i] * self.num_batches_dict[task_name]\r\n task_choice_list = np.array(task_choice_list)\r\n np.random.shuffle(task_choice_list)\r\n\r\n dataloader_iter_dict = {\r\n task_name: iter(dataloader) \r\n for task_name, dataloader in self.dataloader_dict.items()\r\n }\r\n for task_choice in task_choice_list:\r\n task_name = self.task_name_list[task_choice]\r\n yield next(dataloader_iter_dict[task_name]) \r\n ```\r\n We'd just need to pull samples from the raw datasets and not from `DataLoader`s for each task. We can assume the user has done `dataset.shuffle()` if they want to.\r\n\r\n Other sampling methods can later be implemented by changing how the `task_choice_list` is generated. This should allow more flexibility and not tie us to specific methods for sampling among datasets.\r\n","embeddings":[-0.0052698166,-0.4564427137,-0.0409008488,-0.033154469,-0.2421440929,0.0899885297,0.2262675017,0.0337931812,0.3455826938,-0.1832486987,-0.1746232957,0.2782377899,-0.2171302736,0.3147800267,0.3954238594,-0.4474587739,0.0529760905,-0.2687866986,-0.2423506677,0.376804173,0.0238288157,0.1376441717,-0.1650483012,0.1027793512,-0.3490164876,-0.1493272185,-0.2910457551,0.1054907218,0.1436435431,0.0494331829,0.244187519,0.3417589068,-0.0319728144,0.1957802624,-0.0001177549,-0.0263047367,0.2295914143,-0.0533907004,0.1380272061,-0.0514310077,0.1381418407,-0.3776621819,0.0677296743,-0.1242497787,0.1998292506,0.2635385692,0.1815100163,0.1306600869,0.3444681168,-0.3399113119,0.0587665401,0.2794277668,-0.2045923024,0.3041356802,-0.081949845,0.0475111082,0.0006843437,0.2406243682,0.6316853166,-0.6122648716,0.0095472531,0.1181457564,0.0206318013,-0.0755933002,0.2131651789,-0.0758276805,-0.308111757,-0.2063162178,-0.4016385972,0.1982733011,-0.1354376078,-0.126031056,-0.1750893742,-0.572663486,-0.0965319499,0.0781988278,-0.2844801545,0.0156463217,-0.1918295175,-0.048152294,-0.132640928,-0.1277317256,-0.0317760147,0.1317657679,0.3510734737,0.622261703,0.1479907334,0.155235827,0.3822132051,0.1009348929,-0.0878843218,-0.2404812276,0.3139014244,0.1148010492,-0.3939471841,-0.362313807,0.204201147,-0.2805657089,0.2293514609,0.0694336668,0.166472435,0.0844727755,-0.1230995134,0.2612062097,0.3129147887,0.0046515092,-0.1349536479,-0.097687386,0.0522312447,-0.0729626864,0.2354493439,0.1074520722,0.1639313996,-0.0166475121,-0.4331251085,-0.0816401616,-0.135004431,0.160520345,-0.1213202775,-0.5605207086,-0.0321442336,-0.1420903504,0.1316174716,-0.0990331993,-0.3058051169,0.1391584426,-0.0849037021,0.2681118548,-0.0944984704,0.1329028755,-0.0396047086,0.0485027283,-0.4495005608,-0.0649096295,0.3281109035,0.213588044,0.0528768487,0.1371210515,-0.0014478324,0.0744328275,0.3536281586,0.0548316017,-0.0100005325,0.0429559201,0.0779072419,-0.3095371127,-0.0271462183,0.2130467594,-0.3466988802,-0.1356944293,-0.1648989022,-0.2356415987,0.1343441159,-0.0148965195,-0.1898042858,-0.2510464787,-0.5263698697,0.6592676044,-0.0305614118,-0.0570743084,-0.1278265864,0.325268209,-0.2402964383,-0.0782807469,0.1297526062,0.0458408371,0.0346587524,-0.3742161393,-0.0962451994,-0.0856850073,-0.0403569527,0.3011822701,-0.1886329651,0.2752712369,0.0955034196,0.1311880052,0.2198126018,-0.4872383773,-0.030618431,0.293540895,-0.0262911692,0.378112644,0.0742825419,0.2482448965,-0.0497759134,-0.1843891442,0.2528439164,0.9806962609,-0.3445678055,0.0078132134,-0.1259717643,-0.3848104477,0.5938204527,0.3565863967,0.1247872785,-0.2823462188,-0.135115698,0.0828916058,0.0904686525,0.1159305573,0.0790559202,-0.0920598134,-0.3260626495,0.0239144992,-0.1623580605,-0.1164990962,-0.4359433651,0.0106736897,-0.0292150863,0.359777391,0.2142833769,-0.1092868671,0.271482259,-0.279142499,-0.1207623109,-0.302593559,0.07068979,-0.1252287775,-0.0336627178,-0.2842171788,-0.0142541584,0.2878228128,0.1741374135,-0.0913474709,-0.4625684023,0.3106147945,-0.1135091856,0.0794752613,-0.0783736259,0.671086669,-0.1330433488,-0.0517084524,0.1867444366,0.2545598149,-0.305372268,0.260253489,0.3174130917,0.3043528199,0.1962373555,-0.0506784692,0.0802746788,0.0710069239,-0.0063327975,-0.2542899549,-0.1688382477,0.4554283321,-0.1798306555,0.4642221034,-0.0735618845,-0.0063692285,-0.2080316544,-0.044366017,-0.2130458355,0.4461049139,0.2914744914,-0.1696991026,0.2145278305,-0.1094082892,-0.3521736264,0.1489396095,0.1390203834,0.0311614908,0.0351359546,0.0125844833,0.0066877385,-0.2945314348,-0.114424333,0.2900375724,0.5758755803,0.2188544869,0.1371133924,0.0179561619,-0.2088220268,-0.0903355405,0.0136938551,0.2752796113,0.016923096,0.1939088106,0.0156420991,0.0645550787,0.0424579158,-0.2073298991,0.1576102227,-0.2026110142,-0.0598201603,0.0319160037,-0.1862059683,-0.5582320094,-0.2554792762,-0.0109240115,-0.367121309,-0.2729380131,0.1540726423,0.0434343591,-0.1620724499,0.1376047432,0.2474056929,0.6375625134,-0.464548409,-0.2303404957,-0.1406392306,-0.3725031912,-0.0714472756,0.0738391206,0.4597982168,0.286105603,0.3831216097,0.3352016807,-0.0749679133,0.1093754768,-0.4088929594,0.0103904335,-0.281027019,0.2483323216,0.0805492625,-0.2144775093,0.174296543,-0.3981611431,-0.063097015,-0.0273258686,-0.0667391717,-0.0043625529,-0.0064716209,0.0804154351,-0.19725658,-0.1686619967,-0.6268009543,-0.4757328629,0.3247357011,-0.28774333,0.0625835881,0.0964683592,-0.2704408765,0.1228222027,-0.0775507092,-0.0060857539,-0.1258472651,-0.0420749076,0.0276470743,-0.3368201256,-0.1450543702,-0.3748605549,-0.0966837406,0.2254825532,-0.0754836425,-0.0141698429,-0.1489147395,0.1013894454,-0.1627754867,0.1472621113,0.1359443963,0.1742740273,-0.2007130235,0.1626321971,-0.0187071189,-0.3295623362,0.2292136699,0.3730612397,0.4572847784,0.1779815704,0.0543223433,0.164114818,0.7202286124,0.2120819688,0.1129700169,0.1207047924,0.3397210538,-0.0986955017,0.165661037,-0.2394076139,0.2550593317,-0.1768271774,0.4767411649,0.0497106537,0.201895982,-0.2010157257,-0.2819044888,-0.0300928913,-0.2980362177,-0.2182206661,0.2677378058,-0.2083994299,0.0401242487,-0.0738252997,-0.400900811,-0.2804740369,-0.0633699447,0.0563447215,0.4238347411,-0.0578905009,0.1724988371,-0.3345383406,-0.0698195025,-0.4045552313,0.0591119081,0.0287392661,0.2381699681,-0.0914681554,-0.0620659031,-0.0657405183,0.1658874452,0.6495799422,-0.4798402786,-0.1041220874,0.0836596861,0.0829392597,-0.043582961,-0.1442500353,-0.2400393784,-0.0459776148,0.4381367266,0.4039086699,-0.3940845132,-0.2811399996,0.1908792257,-0.2568361759,-0.1257805824,-0.2316512614,-0.2114178091,-0.1747581065,-0.1496135443,0.0511502884,0.2645751536,0.2115340531,-0.1040391102,0.1365631372,-0.0197205618,0.2384251058,0.2117637396,-0.0146641284,-0.0997286066,0.2485462576,0.1289398521,0.0363273509,-0.1606733948,-0.3427597582,0.0713743195,0.0416475013,0.1356416643,0.2838735282,-0.0084102526,0.6613308191,0.0713406429,0.3820053041,0.2950716019,-0.0361781269,-0.0069730943,-0.2792735696,0.1381271631,0.1890255213,0.2439095229,-0.1907054186,-0.1732691973,0.5538418889,0.2548626959,-0.1085383743,0.3889471889,-0.5794236064,-0.2847029567,0.2820958793,0.066470772,0.7568143606,-0.0227548741,0.2946831584,0.0510727689,-0.4602007866,0.5372838378,-0.4241019785,0.1517775655,-0.0489807017,-0.119680129,-0.0108526945,-0.1056490615,-0.101798825,0.3079398274,-0.3381010592,0.3063101172,0.4919751287,-0.2321570665,-0.0793584585,0.5174077153,0.0909696072,-0.3882365823,0.0089381607,0.008923498,-0.2548416257,0.1692931652,-0.1970803142,-0.2969521582,0.1913824975,0.1274457872,-0.2990947366,0.0804291293,0.083655335,0.2613982558,0.321977824,0.0013858244,0.3056202531,-0.2154478431,-0.0835574716,-0.1875606626,-0.1017180234,0.1674012989,-0.2228167206,0.0701348931,0.0455236956,-0.1812613904,0.5122030377,-0.0778392404,-0.1275695562,-0.2751877904,0.2506486475,-0.4533955753,-0.2337380648,0.1493917257,0.1460471749,0.2355832607,-0.3322305977,0.5851309299,0.0823445618,-0.0055854176,0.0122653535,0.1860225797,-0.2039264739,0.5055804849,-0.3174492121,0.0485519059,0.0821749195,-0.0439546891,-0.1198657155,-0.4059040546,-0.0971243232,0.3543559909,0.0950753689,0.0254548099,-0.0568877012,0.1359050423,-0.1269838512,-0.059183307,-0.138688013,-0.3926324248,0.1044555083,0.2105823457,0.2017632723,0.1646538228,-0.075263299,-0.2580513954,-0.2193695605,0.5150886774,-0.0267387256,0.2010471821,-0.1641914994,0.3756381869,0.424295038,0.5418843627,-0.208677426,0.0648792237,0.0617794804,0.0810440257,0.3777204454,-0.2797585428,0.0444473661,0.2270772457,-0.1876550913,0.0065835668,-0.2786373794,-0.1091991812,-0.2729395628,0.138550669,0.147890538,0.0352205932,-0.0034776439,-0.0839107186,-0.0454937704,-0.0185140707,-0.0612044856,-0.3870376348,-0.0260496642,0.1879039407,-0.04360855,0.0557614714,0.0179283097,0.2448153049,0.0428881794,-0.0685513765,0.3669426441,0.0915794373,0.1558605582,-0.1403006762,-0.2409394383,0.0204288084,0.2335350513,0.1394292712,-0.0091820667,0.1907389462,-0.2308543324,-0.1371116638,0.1696343869,0.1269953847,0.0751055032,-0.1060685143,0.0985283852,0.0758041441,-0.1129791662,-0.0234450195,0.2159123421,0.0966682434,0.0432640724,0.115269579,0.1418036819,0.4941137135,-0.2662071884,0.0839686096,0.108830139,-0.3563461006,0.1345180124,-0.1554356813,-0.0202244185,0.0292805787,0.1533048153,-0.0182627197,0.1168425158,-0.3583212495,0.381285429,0.2211309671,-0.0039729262,0.0310310982,0.2010092884,0.0503938608,-0.2398852855,0.2065346837,0.0274656117,0.2873491049,0.0663268194,0.8376120329,-0.2126868963,-0.3841235638,0.1474457979,0.4088921547,0.0464997962,-0.2087956071,0.1675236374,0.1362494975,-0.3738659918,0.0666099712,0.049195312,-0.1454670876,0.6858702898,-0.1041823328,0.1188837439,-0.0746169984,-0.2131343782,0.3843899369,0.1691306978,-0.0546512343,-0.0669413358,0.1378991902,-0.1112387255,-0.1763102263,0.4905510247,0.2080181688,0.0090578506,-0.45847103,0.0202280208,0.1295265853,0.1386816651,-0.0580507852,0.1767341942,0.4701445699,-0.0721231028,0.156820789,-0.0337035581,-0.0108422982,-0.1236676872,0.0422087908,-0.4178423584,-0.3098797798,0.5140650272,0.0914872959,-0.1700217426,0.0972797051,0.3429419398,-0.6166444421,-0.2568539381,0.2994376719,-0.1155206561,-0.0902803391,-0.2339397669,0.046573136,0.1041120663,0.6218804717,0.1480693668,0.3906709254,-0.3981969357,-0.1069854945,-0.5736933947,-0.2316160947,0.1399529427,-0.0517088212,-0.1787454635,0.0944190174,-0.064928554,0.2246150672,-0.2561454475,-0.0194277037,0.2132712007,0.0900172591,-0.4094817638,0.168774128,0.1896282434,0.0682809278,0.3064978421,-0.0322039425,-0.1583385915,0.2072924078,-0.1309208572,-0.1387364715,-0.0877376869,0.2436159253,0.1898093969,0.0468430705,0.0211430769,0.4844205081,-0.0040277406,-0.2120718658,0.0108474754,-0.002689179,-0.3256477118,-0.1787250638,-0.0771346539,0.208040759,-0.360068351,0.2944652438,-0.009957361,-0.136324048,0.0238765664,-0.0087926295,-0.219227314,-0.0255230386,0.0336700194,0.2317088395,0.2493135929,0.6675162315,0.0166511089,0.0932118446,-0.0899906084,-0.0504892319,0.0260747373,-0.0150224073,-0.2413495332,-0.144943893,0.0330268629,0.0146863097,0.1847986579,-0.7851419449,-0.3945767283,-0.0219792314,0.0314114466,0.0848202556,0.7013661861,0.1230232045,-0.1750703901,-0.0973248556,-0.3418374062,0.0201752391,-0.2287635356,0.1751659811,-0.3607013822]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"Another thought: Multitasking over benchmarks (represented as Meta-datasets in nlp) is probably a common use case. Would be nice to pass an entire benchmark to our `MultiDataset` wrapper rather than having to pass individual components.","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":35,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nAnother thought: Multitasking over benchmarks (represented as Meta-datasets in nlp) is probably a common use case. Would be nice to pass an entire benchmark to our `MultiDataset` wrapper rather than having to pass individual components.","embeddings":[-0.0656119063,-0.3723547161,-0.0681932122,-0.0492688045,-0.2852725089,0.1337666065,0.2164802402,0.1206526011,0.3021679819,-0.142117992,-0.1687693298,0.3203654289,-0.1637993902,0.3486942053,0.4185857177,-0.4121801853,0.0316882245,-0.2107659131,-0.2145602852,0.424690038,0.0625791103,0.0833854899,-0.1656299382,0.1017269045,-0.296767503,-0.1877516061,-0.3303469718,0.1228973418,0.1802742928,-0.0067459373,0.2167422026,0.3386645317,-0.005248019,0.0762604773,-0.0001127535,-0.079764612,0.2180484086,-0.0533743948,0.1031291857,-0.0139625371,0.193061769,-0.3639018834,0.0466753058,-0.1068037078,0.16875875,0.1935687959,0.1722348034,-0.0764045641,0.3412645757,-0.25450176,0.1023415551,0.3156677186,-0.2399112135,0.2976697683,-0.1113831922,-0.0868199468,0.0119753554,0.2438913584,0.6718504429,-0.627389431,-0.0600717738,0.0572424307,-0.0323867463,0.0497779027,0.0926750675,-0.1301555783,-0.3021051586,-0.2108954191,-0.3377223611,0.2510986626,-0.1597300768,-0.1370611787,-0.2288980633,-0.5552752614,-0.0130585609,0.0512918793,-0.3080892265,0.0318329036,-0.1725958437,-0.0062471372,-0.1011449695,-0.1240547299,0.0575542711,0.0890993401,0.3357074857,0.5584859252,0.1207945719,0.1740791351,0.3698949516,0.0731535032,-0.0847848877,-0.2591187954,0.274199307,0.1001640707,-0.415807724,-0.3103589714,0.1395327002,-0.2330120653,0.3023065329,0.0970019847,0.1351532787,0.1109842435,-0.1236224025,0.2736952007,0.3402852118,0.0342561081,-0.0776993334,-0.1692268997,0.0107862754,-0.1416548043,0.1802741289,0.1135542169,0.2074791193,0.0258042999,-0.4392684698,-0.1369123459,-0.2284797579,0.1634885073,-0.2595624626,-0.5688914061,-0.1121168733,-0.0956882536,0.0620818324,-0.1388763189,-0.2614015639,0.1391816139,-0.0921640471,0.269533962,-0.0688549802,0.0062881103,-0.0510915071,0.0879677236,-0.4555426836,-0.0101290494,0.3323619366,0.1006498411,0.0862040222,0.1457322687,-0.0197839607,0.0622323789,0.3870140612,0.0119770197,-0.0044591278,-0.0121356696,0.1867395937,-0.314055413,0.0292185564,0.1831528246,-0.3368544579,-0.0813517347,-0.1455326527,-0.148672998,0.1425740868,0.0321329869,-0.225368157,-0.2860387266,-0.5482085943,0.6673361063,-0.0112446342,-0.0629958063,-0.1631250829,0.3023531139,-0.3023776412,-0.0930897295,0.1709520966,0.0131647279,0.0217474457,-0.2774837613,-0.1617570072,-0.070432432,-0.1019014791,0.3153512776,-0.192191571,0.3084415197,0.0420111939,0.1713878959,0.2346699983,-0.484659791,-0.0329162888,0.331353277,-0.0713307858,0.4018953145,0.0502471477,0.226153627,0.055644203,-0.1782498807,0.2328609526,0.9699891806,-0.3880668879,0.0095262546,-0.0859299898,-0.3920594156,0.543384254,0.3569907248,0.059594553,-0.2415657938,-0.1471162885,-0.01227789,0.0871227756,0.1098291799,0.0995172784,-0.0751654282,-0.420724839,0.0233743042,-0.22361058,-0.1986405104,-0.431843549,0.0181785617,0.0211398825,0.2296025604,0.3293683529,-0.1370198131,0.2478265911,-0.279499054,-0.0465501472,-0.375577569,0.1250411868,-0.0515294671,0.0089130327,-0.202386722,-0.0481616594,0.3567742109,0.1752245128,-0.1295220703,-0.3912127018,0.307117492,-0.0529440902,0.0381249189,-0.0405544713,0.6141855717,-0.2152328938,-0.057150811,0.1891903877,0.2451416552,-0.2727411985,0.2884795964,0.3536682427,0.2753916681,0.1926756799,-0.0359169692,0.0541778505,0.0602812544,-0.0329499058,-0.2539368272,-0.0517228842,0.4938859344,-0.1278951317,0.4670577645,-0.069475919,0.0535185635,-0.1416844428,-0.0365008079,-0.1959548295,0.4411692917,0.290050596,-0.2122812867,0.3862041235,-0.1093377545,-0.3189106286,0.1744893789,0.1380334646,0.0512453653,0.0646825656,-0.0090722945,-0.0365703031,-0.2727175057,-0.0716025606,0.3707677424,0.6577773094,0.2386050671,0.1774469763,0.0148176551,-0.2293145061,-0.0808017552,-0.0312070567,0.1939817816,0.0265200138,0.0988889784,0.0134747196,0.0223168992,0.048615586,-0.1400788873,0.1509891003,-0.1698394567,-0.0097850068,0.0413742438,-0.2205082625,-0.556699276,-0.2170759737,0.0291742794,-0.3283439279,-0.2532233,0.1027175263,0.1222499907,-0.1313920766,0.1512336433,0.2321764529,0.6160982847,-0.4460153282,-0.2688653469,-0.083535172,-0.30082196,-0.1132342294,0.129001677,0.4587604702,0.3014477491,0.3852969408,0.3329302371,-0.0090692043,-0.0216647424,-0.3423926234,-0.0389637984,-0.2504143715,0.2589896619,0.0364834853,-0.2416997552,0.144774124,-0.3871199787,-0.0762299225,-0.0768024102,-0.0706041306,-0.087010622,0.0312856399,0.1460472494,-0.1852710694,-0.1524782032,-0.5852788091,-0.4743441641,0.297499001,-0.2947556674,0.0538740195,-0.1080989987,-0.3133614063,0.1933502853,-0.1261135191,-0.0194320362,-0.1375253648,-0.0907317176,-0.0554949977,-0.3396884203,-0.1574500352,-0.3975775838,-0.0830503479,0.2783398032,-0.0493620038,-0.0424786881,-0.1559733897,0.1311392635,-0.1548351496,0.0600678958,0.092638351,0.1741753072,-0.2363420427,0.1292829514,-0.0282709207,-0.2620977163,0.2368300408,0.2830856144,0.3746196032,0.1705283523,0.0223739874,0.2380328178,0.7015311122,0.2394606918,0.0548744611,0.1664235741,0.2817083895,-0.023641102,0.1192622632,-0.2199314088,0.1519428492,-0.1884621233,0.4781728983,0.1039729193,0.1576092988,-0.2565976381,-0.2707085907,0.0301404595,-0.253130585,-0.1704377681,0.2131997496,-0.3159549236,0.0334231965,-0.0661188886,-0.3981959224,-0.2217507362,-0.018919535,-0.0145324897,0.2880031466,-0.0429943614,0.1606458724,-0.3352029324,-0.031863261,-0.4085684121,0.1178528145,0.0969888642,0.3318119943,-0.0865966529,-0.0450321995,-0.0849434957,0.1118619218,0.5522227883,-0.4672246873,-0.0927446634,0.1084990948,0.0698909834,-0.0965210274,-0.08832445,-0.2206002176,0.0458063632,0.4289894998,0.3607068956,-0.2990393639,-0.2886453271,0.1550899148,-0.2195772976,-0.0486546643,-0.1847695708,-0.2151086032,-0.1798702329,-0.1340259612,0.065688394,0.2318376303,0.2185507417,-0.0807618946,0.1529323757,-0.1103344709,0.1578861326,0.1860540509,-0.0437117927,-0.0945378542,0.1524299085,0.2028805166,0.0133355893,-0.1537550241,-0.3070273995,0.0210480541,0.0137784854,0.1322285831,0.2465014011,-0.0179677159,0.6386027932,-0.0434415564,0.3078645766,0.3434130847,-0.150402233,-0.0601095185,-0.2892884612,0.1642033756,0.186846301,0.2574941516,-0.1232691109,-0.1871760935,0.5725802779,0.3358720243,-0.0695810691,0.4539735615,-0.5722930431,-0.3126081228,0.2435651124,0.0592391305,0.5774688721,-0.0839801729,0.2616603374,-0.0199688822,-0.4101735353,0.531545043,-0.3737444878,0.1534080505,-0.0518805608,-0.2438551337,0.0206009112,-0.0840625241,-0.0070884353,0.3127882183,-0.3229174018,0.3246551156,0.4892107546,-0.2282972485,-0.0012170743,0.5655279756,0.0258646496,-0.3701806068,0.0169428084,0.0685838386,-0.2902960777,0.0819194466,-0.1532068551,-0.210727334,0.2203501612,0.1797535419,-0.3520199656,0.0588380285,0.0268049389,0.2144398093,0.2053644508,0.0353680924,0.2878020108,-0.1741971076,-0.0609709769,-0.1578622162,-0.0868908614,0.1287158281,-0.1930597275,0.0364055037,0.0944975242,-0.2095956802,0.4564698339,-0.0603312105,-0.1451976746,-0.2287000865,0.2702369988,-0.5092933774,-0.1495708078,0.1981717646,0.1473065466,0.2863526046,-0.2999467552,0.5590233207,0.1000286713,-0.015114082,0.0415357947,0.1996981502,-0.2363314927,0.5029978752,-0.3028038442,-0.0152307032,0.0607312731,0.0567980893,-0.1458500326,-0.3981849253,-0.0703275874,0.2884268761,0.1405790001,-0.00456734,0.0554144718,0.0992426649,0.0506524295,0.0197228864,-0.2005835325,-0.3242928088,0.1335985214,0.1471768469,0.2038132399,0.2274574637,-0.0952346846,-0.302659452,-0.1976301521,0.4594477713,-0.0290687792,0.1782679111,-0.1285364777,0.3476558626,0.4074636698,0.5269008875,-0.2459168732,0.1176432669,0.0456090458,0.126955241,0.5051093102,-0.2154181302,-0.0042853183,0.3125104308,-0.1364323646,0.0628677234,-0.2663428485,-0.1617915034,-0.1922883093,0.1079843566,0.1229564995,0.025694279,0.0297054369,-0.1108930856,-0.0953914598,-0.0409348831,-0.0558067523,-0.3989929855,-0.0612472631,0.0984429121,-0.0674813837,0.0765927061,0.0247240793,0.2167935669,-0.0350389071,-0.1232553199,0.2448349595,0.117604129,0.1735719293,-0.152463451,-0.1608665287,-0.036695566,0.224973768,0.1443036944,-0.0504222587,0.1905044317,-0.2337229252,-0.0754129738,0.2463466823,0.1047860086,0.0912302881,-0.0884775147,0.1727251858,0.1037008911,-0.0844673589,-0.0206098557,0.2310098857,0.0595982149,0.1164538935,0.1318477839,0.1963889152,0.4679244757,-0.1857157201,-0.0031026916,0.1126827747,-0.384761095,0.1262214929,-0.1879340112,-0.0724590644,-0.0286672916,0.1343452483,-0.0179315675,0.1367463022,-0.2621413171,0.3380007446,0.3889698982,0.0716751888,-0.007468001,0.2216957211,0.0650139973,-0.1531659812,0.2569547594,0.0542646945,0.3002224565,0.0122461673,0.8360202909,-0.324025631,-0.4433020353,0.1131606698,0.3656114638,0.051836051,-0.2163879424,0.2377729118,0.1352377534,-0.3363131583,0.0876114368,0.0637903363,-0.1809618026,0.7053230405,-0.1334044337,0.1490319967,-0.0833337605,-0.2459761947,0.3666487634,0.1151493937,-0.0711918101,-0.000482557,0.0570859239,-0.1151063964,-0.1761982143,0.5325782299,0.1809427738,0.0182039645,-0.4425997138,0.0100711267,0.0149302715,0.1756376028,-0.0728958398,0.1774266064,0.4411541522,-0.1820049733,0.2456374466,0.0059402124,-0.0461780019,-0.1398256272,0.0468495376,-0.4108140767,-0.4485566318,0.4289399087,0.0766378418,-0.1290846169,0.0875914171,0.3628630936,-0.5635621548,-0.2513415515,0.3186553419,-0.2019164711,-0.1316321492,-0.2096778303,0.0823572353,0.0499644056,0.6422708631,0.2010989785,0.3915332258,-0.3583629429,-0.1792931259,-0.5960265994,-0.1786123961,0.0864177346,-0.085186474,-0.2282559872,0.1549064964,-0.110094659,0.3391738534,-0.1758441627,0.0758246183,0.2411529273,0.0000381273,-0.4148183763,0.0921233892,0.1253775358,-0.0173915774,0.2854141295,-0.1076857597,-0.144165799,0.1217658669,-0.0711131096,-0.0758255273,-0.1164273173,0.2526833713,0.1565549076,0.1055899039,0.0100836242,0.4512912333,0.0178520735,-0.1485956013,-0.0515219867,0.0374910906,-0.3527216017,-0.1814722419,-0.1022648066,0.1490992755,-0.4346382916,0.2957696021,0.0213664174,-0.0618078262,0.1016725376,0.0056778598,-0.2547988892,-0.0585627109,0.031436529,0.1994291693,0.2543523014,0.732393086,-0.0023266317,0.1040791124,-0.0569206923,0.0309470184,-0.0281184278,-0.0886234865,-0.2501010895,-0.1346303672,0.0529886298,0.0765742213,0.0841164812,-0.673864007,-0.4056535363,0.0536899567,0.0224158391,0.0578993671,0.7152570486,0.0275556836,-0.1257441342,-0.1139770523,-0.3888227046,0.0085798884,-0.2279847264,0.2192235887,-0.4018602371]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"Here's a fully working implementation based on the `__iter__` function of @zphang.\r\n\r\n- I've generated the task choice list in the constructor as it allows us to index into the MultiDataset just like a normal dataset. I'm changing `task_choice_list` into a list of `(dataset_idx, example_idx)` so each entry references a unique dataset example. The shuffling has to be done before this as we don't want to shuffle within each task (we assume this is done by the user if this is what they intend).\r\n- I'm slightly concerned this list could become very large if many large datasets were used. Can't see a way round it at the moment though.\r\n- I've used `task.info.builder_name` as the dataset name. Not sure if this is correct.\r\n- I'd love to add some of the other `Dataset` methods (map, slicing by column, etc...). Would be great to implement the whole interface so a single dataset can be simply replaced by this.\r\n- This does everything on the individual example-level. If some application required batches all from a single task in turn we can't really do that.\r\n\r\n```python\r\nimport nlp\r\nimport numpy as np\r\n\r\nclass MultiDataset:\r\n def __init__(self,tasks):\r\n self.tasks = tasks\r\n\r\n # Create random order of tasks\r\n # Using size-proportional sampling\r\n task_choice_list = []\r\n for i, task in enumerate(self.tasks):\r\n task_choice_list += [i] * len(task)\r\n task_choice_list = np.array(task_choice_list)\r\n np.random.shuffle(task_choice_list)\r\n\r\n # Add index into each dataset\r\n # - We don't want to shuffle within each task\r\n counters = {}\r\n self.task_choice_list = []\r\n for i in range(len(task_choice_list)):\r\n idx = counters.get(task_choice_list[i],0)\r\n self.task_choice_list.append((task_choice_list[i],idx))\r\n counters[task_choice_list[i]] = idx + 1\r\n\r\n\r\n def __len__(self):\r\n return np.sum([len(t) for t in self.tasks])\r\n\r\n def __repr__(self):\r\n task_str = \", \".join([str(t) for t in self.tasks])\r\n return f\"MultiDataset(tasks: {task_str})\"\r\n\r\n def __getitem__(self,key):\r\n if isinstance(key, int):\r\n task_idx, example_idx = self.task_choice_list[key]\r\n task = self.tasks[task_idx]\r\n example = task[example_idx]\r\n example[\"task_name\"] = task.info.builder_name\r\n return example\r\n elif isinstance(key, slice):\r\n raise NotImplementedError()\r\n\r\n def __iter__(self):\r\n for i in range(len(self)):\r\n yield self[i]\r\n\r\n\r\ndef load_multitask(*datasets):\r\n '''Create multitask datasets per split'''\r\n\r\n def _get_common_splits(datasets):\r\n '''Finds the common splits present in all self.datasets'''\r\n min_set = None\r\n for dataset in datasets:\r\n if min_set != None:\r\n min_set.intersection(set(dataset.keys()))\r\n else:\r\n min_set = set(dataset.keys())\r\n return min_set\r\n\r\n common_splits = _get_common_splits(datasets)\r\n out = {}\r\n for split in common_splits:\r\n out[split] = MultiDataset([d[split] for d in datasets])\r\n return out\r\n\r\n\r\n##########################################\r\n# Dataset Flattening\r\n\r\ndef flatten(dataset,flatten_fn):\r\n for k in dataset.keys():\r\n if isinstance(dataset[k],nlp.Dataset):\r\n dataset[k] = dataset[k].map(flatten_fn,remove_columns=dataset[k].column_names)\r\n\r\n# Squad\r\ndef flatten_squad(example):\r\n return {\"source\": \"squad context: \" + example['context'] + \" question: \" + example['question'],\r\n \"target\":example[\"answers\"][\"text\"]}\r\nsquad = nlp.load_dataset(\"squad\")\r\nflatten(squad,flatten_squad)\r\n\r\n# CNN_DM\r\ndef flatten_cnn_dm(example):\r\n return {\"source\": \"cnn_dm: \" + example['article'],\"target\":[example[\"highlights\"]]}\r\ncnn_dm = nlp.load_dataset(\"cnn_dailymail\", \"3.0.0\")\r\nflatten(cnn_dm,flatten_cnn_dm)\r\n\r\n#############################################\r\n\r\nmtds = load_multitask(squad,cnn_dm)\r\n\r\nfor example in mtds[\"train\"]:\r\n print(example[\"task_name\"],example[\"target\"])\r\n```\r\nLet me know if you have any thoughts. I've started using this in some of my projects and it seems to work. If people are happy with the general approach for a first version, I can make a pull request.","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":469,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nHere's a fully working implementation based on the `__iter__` function of @zphang.\r\n\r\n- I've generated the task choice list in the constructor as it allows us to index into the MultiDataset just like a normal dataset. I'm changing `task_choice_list` into a list of `(dataset_idx, example_idx)` so each entry references a unique dataset example. The shuffling has to be done before this as we don't want to shuffle within each task (we assume this is done by the user if this is what they intend).\r\n- I'm slightly concerned this list could become very large if many large datasets were used. Can't see a way round it at the moment though.\r\n- I've used `task.info.builder_name` as the dataset name. Not sure if this is correct.\r\n- I'd love to add some of the other `Dataset` methods (map, slicing by column, etc...). Would be great to implement the whole interface so a single dataset can be simply replaced by this.\r\n- This does everything on the individual example-level. If some application required batches all from a single task in turn we can't really do that.\r\n\r\n```python\r\nimport nlp\r\nimport numpy as np\r\n\r\nclass MultiDataset:\r\n def __init__(self,tasks):\r\n self.tasks = tasks\r\n\r\n # Create random order of tasks\r\n # Using size-proportional sampling\r\n task_choice_list = []\r\n for i, task in enumerate(self.tasks):\r\n task_choice_list += [i] * len(task)\r\n task_choice_list = np.array(task_choice_list)\r\n np.random.shuffle(task_choice_list)\r\n\r\n # Add index into each dataset\r\n # - We don't want to shuffle within each task\r\n counters = {}\r\n self.task_choice_list = []\r\n for i in range(len(task_choice_list)):\r\n idx = counters.get(task_choice_list[i],0)\r\n self.task_choice_list.append((task_choice_list[i],idx))\r\n counters[task_choice_list[i]] = idx + 1\r\n\r\n\r\n def __len__(self):\r\n return np.sum([len(t) for t in self.tasks])\r\n\r\n def __repr__(self):\r\n task_str = \", \".join([str(t) for t in self.tasks])\r\n return f\"MultiDataset(tasks: {task_str})\"\r\n\r\n def __getitem__(self,key):\r\n if isinstance(key, int):\r\n task_idx, example_idx = self.task_choice_list[key]\r\n task = self.tasks[task_idx]\r\n example = task[example_idx]\r\n example[\"task_name\"] = task.info.builder_name\r\n return example\r\n elif isinstance(key, slice):\r\n raise NotImplementedError()\r\n\r\n def __iter__(self):\r\n for i in range(len(self)):\r\n yield self[i]\r\n\r\n\r\ndef load_multitask(*datasets):\r\n '''Create multitask datasets per split'''\r\n\r\n def _get_common_splits(datasets):\r\n '''Finds the common splits present in all self.datasets'''\r\n min_set = None\r\n for dataset in datasets:\r\n if min_set != None:\r\n min_set.intersection(set(dataset.keys()))\r\n else:\r\n min_set = set(dataset.keys())\r\n return min_set\r\n\r\n common_splits = _get_common_splits(datasets)\r\n out = {}\r\n for split in common_splits:\r\n out[split] = MultiDataset([d[split] for d in datasets])\r\n return out\r\n\r\n\r\n##########################################\r\n# Dataset Flattening\r\n\r\ndef flatten(dataset,flatten_fn):\r\n for k in dataset.keys():\r\n if isinstance(dataset[k],nlp.Dataset):\r\n dataset[k] = dataset[k].map(flatten_fn,remove_columns=dataset[k].column_names)\r\n\r\n# Squad\r\ndef flatten_squad(example):\r\n return {\"source\": \"squad context: \" + example['context'] + \" question: \" + example['question'],\r\n \"target\":example[\"answers\"][\"text\"]}\r\nsquad = nlp.load_dataset(\"squad\")\r\nflatten(squad,flatten_squad)\r\n\r\n# CNN_DM\r\ndef flatten_cnn_dm(example):\r\n return {\"source\": \"cnn_dm: \" + example['article'],\"target\":[example[\"highlights\"]]}\r\ncnn_dm = nlp.load_dataset(\"cnn_dailymail\", \"3.0.0\")\r\nflatten(cnn_dm,flatten_cnn_dm)\r\n\r\n#############################################\r\n\r\nmtds = load_multitask(squad,cnn_dm)\r\n\r\nfor example in mtds[\"train\"]:\r\n print(example[\"task_name\"],example[\"target\"])\r\n```\r\nLet me know if you have any thoughts. I've started using this in some of my projects and it seems to work. If people are happy with the general approach for a first version, I can make a pull request.","embeddings":[-0.0021329687,-0.393357724,-0.0449090786,0.0113289133,-0.223907724,0.0217612665,0.1978975385,0.0275363214,0.4143027067,-0.1978273988,-0.2370704561,0.3982106447,-0.1945328712,0.3028288484,0.3515968025,-0.4019325972,-0.0221231747,-0.2419832498,-0.2949595749,0.3513539135,0.060441345,0.1977681965,-0.2124996334,0.0691922233,-0.21596843,-0.1789111793,-0.3551908135,0.1598140448,0.0868028328,0.0085352156,0.2525814474,0.4207158685,-0.0479114838,0.1720012128,-0.0001153873,-0.0406288989,0.2224498838,-0.1226238683,0.0533427075,-0.0625768453,0.1727963388,-0.4180442393,0.0764216483,-0.1779254228,0.1683048904,0.1507713795,0.1285147369,0.0180312525,0.3956317604,-0.2608163953,0.1016746759,0.31338346,-0.1547463536,0.3200140297,-0.134758845,-0.0212383512,0.0250547845,0.2562729418,0.6043381095,-0.5478271842,-0.0398601629,0.09073174,0.0203817412,-0.0825083554,0.1283764392,-0.0271547455,-0.3430265784,-0.2517932057,-0.3898305297,0.2208646685,-0.0587252378,-0.1800947934,-0.2241705209,-0.5440157056,-0.0574220493,0.001468258,-0.2000159472,-0.0979659259,-0.2069236785,-0.07884489,-0.080557473,-0.1003158018,-0.0552609116,0.1227865815,0.286226958,0.5722645521,0.0916849449,0.1695954949,0.411288619,-0.0037697784,-0.1038940176,-0.2763192356,0.289000392,0.1087882146,-0.4652658999,-0.3517702818,0.2077233344,-0.273562789,0.2420556396,0.1181547046,0.1267685294,0.1409205943,-0.1725961566,0.2350256741,0.310200125,0.1616436839,-0.0854422748,-0.1401860565,-0.0408445299,-0.1004976854,0.2605678141,0.0931036323,0.1397622079,-0.0241557043,-0.4383574128,-0.1129162163,-0.1902085245,0.1750691384,-0.2368577123,-0.4651109278,-0.1156226918,-0.1334738284,0.1643810421,-0.0893167332,-0.326359123,0.0938572586,-0.063605383,0.2285141498,-0.1643101871,0.0954068378,-0.0509240516,0.0793650821,-0.4200201631,-0.1034961343,0.304156512,0.2007611394,0.0887783244,0.1018332765,0.0413549505,0.1614445001,0.3386406898,-0.0185898189,0.0711685494,-0.0067506242,0.1700589955,-0.2418665141,0.0236337911,0.085721463,-0.3276285827,-0.1432785988,-0.1162451878,-0.171212554,0.0955512598,0.0015723535,-0.247001797,-0.2189024687,-0.5296889544,0.6929865479,0.0090131285,-0.0752257481,-0.1737617254,0.2170035094,-0.2249669731,-0.1134622991,0.1857753545,0.1306339055,-0.0194604285,-0.2366183102,-0.0868393332,-0.1156580597,0.0077987011,0.2511880398,-0.1722691208,0.3111093342,0.0009651037,0.251101464,0.2696984708,-0.4574744999,-0.0800231472,0.2996055186,-0.0706133023,0.4915627539,0.0650966316,0.2020908892,0.0199483726,-0.2011215389,0.2741647065,0.972507,-0.3398758769,-0.0454664975,-0.1353515983,-0.3082438409,0.5843709707,0.2936512828,0.0460594334,-0.2365175784,-0.1625545621,0.0251358207,0.1369397491,0.1146137789,0.1020966321,-0.0433963537,-0.3110345602,0.0680137202,-0.1612396091,-0.1333116293,-0.4370239973,0.0378150642,0.0461094975,0.3442163467,0.1861216575,-0.0424990766,0.2013813108,-0.2429167181,-0.2007026523,-0.323187232,0.128223896,-0.0340776592,-0.0067648645,-0.3257844448,-0.0405895934,0.3911013603,0.1122538075,-0.1496475339,-0.5206973553,0.2977626622,-0.0831870511,0.0701502338,0.0002406276,0.6380757689,-0.0924057141,-0.1353616714,0.1831828803,0.3113519549,-0.2477384657,0.1465300024,0.3683392704,0.1787835509,0.1724849343,-0.0537876487,0.0411787517,0.0147792213,0.0640910119,-0.2540854514,-0.0417446531,0.4979396462,-0.2070444524,0.4949687123,-0.0594720133,0.0961595923,-0.0777202174,0.0036059322,-0.2113495022,0.4170465171,0.2573398948,-0.1638271213,0.3197738826,-0.1749095321,-0.3892853558,0.2181707472,0.1477795243,0.0507120639,-0.0239398666,-0.0040273569,-0.0041733584,-0.2835718393,-0.0840637833,0.2592914701,0.6727406383,0.214672938,0.1757933795,-0.003553381,-0.2128891796,-0.0756322518,0.0170345791,0.2131554037,-0.0183848534,0.1383441389,0.0632266924,0.0550123565,0.0272596665,-0.1688721776,0.1636389792,-0.1502516419,-0.1182814389,0.1191030592,-0.2400859594,-0.644918561,-0.1902624816,-0.0053812331,-0.337434113,-0.33798334,0.0619773827,0.0960919634,-0.0927924737,0.1745878905,0.3390874267,0.5370786786,-0.4226085842,-0.2787002623,-0.0702021793,-0.313002944,-0.228705585,0.1059597805,0.4761104882,0.3147264421,0.4362726808,0.3334596455,-0.0783116147,-0.0111065796,-0.4081683755,0.0738274381,-0.315710932,0.2591182888,0.1077924967,-0.1619821787,0.1569965482,-0.3664077222,-0.0544608198,0.0256724525,-0.1082522348,0.0057358658,-0.0213206448,0.0617834181,-0.224941045,-0.2057399303,-0.604914546,-0.3986760378,0.3317517638,-0.2684569359,0.1595519632,0.111350961,-0.2901903391,0.2159575075,-0.0665777177,-0.0354988389,-0.1976852715,-0.0266128462,0.0486370549,-0.3335725069,-0.2293011695,-0.3334796429,-0.0837493837,0.2851195037,-0.0606135279,-0.0192406643,-0.2144850194,0.1371329874,-0.094092682,0.1131651923,0.1819939911,0.2623928189,-0.2219088674,0.151617229,-0.1329756379,-0.3378427029,0.2892360687,0.3065322638,0.466779381,0.2628137469,0.0595809482,0.1985298693,0.7363936305,0.246727109,0.0307715479,0.1628224552,0.2447069585,-0.0085787317,0.1286433935,-0.2955615222,0.2436808199,-0.205701232,0.4891047776,0.0468424447,0.1214248314,-0.278498292,-0.319814533,0.097293891,-0.2960558534,-0.240993917,0.2288609147,-0.2286780626,-0.0296684373,-0.0462915935,-0.3796523809,-0.3028076589,-0.0377591178,0.0249534212,0.3650663495,-0.0835358948,0.1451676786,-0.333181113,-0.0588308461,-0.4737010002,0.1164907366,0.0272342898,0.3008978963,-0.0765070468,-0.1025140136,-0.0961406827,0.0985039026,0.7110226154,-0.4442332387,-0.1452287585,0.1435721815,0.0976991877,-0.0933028087,-0.1217572019,-0.2865295112,0.1103212461,0.5207916498,0.3619172275,-0.4547293186,-0.2635481656,0.1425701976,-0.2180935591,-0.0463845506,-0.1899246871,-0.2505950928,-0.2670265138,-0.0971078277,0.0422475599,0.2250931114,0.279640615,-0.1162925884,0.2169062495,-0.0720493048,0.170096904,0.2198312432,0.0008591422,-0.0268347897,0.2566709816,0.144443661,0.0839143991,-0.0964154303,-0.3558118939,0.1244944111,-0.0286754426,0.1375513971,0.3451258242,-0.0695467219,0.5981542468,0.1567118615,0.3002188802,0.221235469,-0.1367071718,-0.0138940141,-0.188893944,0.2365727872,0.2181071341,0.2083162665,-0.143136844,-0.2470641732,0.4849055409,0.2541722953,-0.1302737296,0.4766074419,-0.4629568458,-0.3232700229,0.2342954576,0.0823782682,0.7584080696,-0.02576603,0.2482214868,0.2218556255,-0.3872935772,0.5620595217,-0.2961156964,0.152210474,-0.0602108389,-0.1560702324,-0.003472877,-0.100529477,-0.0665955916,0.1920007169,-0.3005217314,0.3170953393,0.4741307199,-0.2545052469,-0.0254249908,0.5520509481,0.0740955696,-0.4448995888,-0.0632599518,0.059811376,-0.2775410116,0.1279592514,-0.1231737211,-0.3219970763,0.1960480064,0.1436817944,-0.3340365589,0.0841854885,0.0025316614,0.2179044336,0.302400738,-0.0187245514,0.2435373217,-0.1553887278,-0.0691190436,-0.2024183869,-0.0824261308,0.1233870387,-0.1531684995,0.1167893335,0.0909700692,-0.2312885225,0.5451815724,-0.0926450938,-0.0711481422,-0.2456399798,0.2246968001,-0.4014180303,-0.2000752389,0.2005418986,0.1613871008,0.1955031455,-0.2907215059,0.590513289,0.1267305017,-0.0039932202,0.0149421496,0.2240964025,-0.2379041165,0.4939911664,-0.3347179592,0.0327793472,0.10144189,0.0415830947,-0.1077694371,-0.4338435829,-0.0012162753,0.3342471421,0.0758200586,0.0514318533,-0.0814498663,0.0753405467,-0.0482136868,-0.0935331658,-0.1852850616,-0.343398571,0.0834069327,0.2625076175,0.140331924,0.2258595675,-0.0623512864,-0.3012261093,-0.1828821301,0.4282072186,-0.0905606672,0.1718102247,-0.1609401405,0.3156589866,0.4120322168,0.4628573358,-0.2444220483,0.0169075225,0.0225408543,0.0932232365,0.3734447658,-0.2316675633,0.0770936757,0.2429405302,-0.1389829665,0.0426353589,-0.3441099823,-0.1245713383,-0.2696318924,0.1124397963,0.1352790892,-0.0400485955,0.0043101036,-0.1392835975,-0.1367931813,-0.0682709739,-0.0343441516,-0.4005203843,-0.1031487212,0.0878288001,-0.0472272225,0.0667581111,0.06176126,0.1681156158,0.0527072214,-0.0813245028,0.3830560446,0.0915634632,0.0713892356,-0.1855533272,-0.1679846644,0.0547182336,0.2784900069,0.1424310654,-0.0489643142,0.0697531328,-0.2131606489,-0.1310157776,0.2193908542,0.2282589376,0.0514495932,-0.1970723867,0.1387758255,0.0869611502,-0.0729918554,-0.0242099129,0.258713305,0.0611560903,0.0581025332,0.155780524,0.1339227855,0.4340779483,-0.1839957237,0.04718576,0.1686084718,-0.3724967837,0.0530011691,-0.1925285757,-0.1239283532,0.0434690677,0.1779720187,-0.0207130462,0.1128004715,-0.1853487045,0.2707884312,0.3122720122,0.0097170044,0.0476661585,0.1371757686,-0.0593017861,-0.1810820997,0.2001945078,0.0575599112,0.3443952501,0.0186985787,0.8508674502,-0.2761963904,-0.3828791082,0.1721753925,0.4096405804,0.0317296125,-0.1823183149,0.0848926306,0.0774157643,-0.3451005816,0.0185377076,0.0384724587,-0.170385167,0.6866490841,-0.1312552989,0.1184320673,-0.2223552316,-0.2760975063,0.3105563521,0.1714401543,-0.0565256886,0.0606840104,0.1309128404,-0.096059978,-0.1630795002,0.4281637669,0.2053738534,0.0710359439,-0.4505880475,-0.0553794578,0.1022794992,0.1184583977,-0.0857543796,0.1549673676,0.4437691271,-0.1220208704,0.2705670297,-0.0237433761,-0.0232195724,-0.1321669966,0.163266778,-0.3313334882,-0.3735802174,0.4598419964,0.0828377753,-0.1162544861,0.0967511684,0.2910226583,-0.6527740955,-0.2447612137,0.2560928762,-0.0618293695,-0.0006648638,-0.1802778989,0.0690605193,0.086608544,0.6433349252,0.1645341218,0.3060011268,-0.4014544189,-0.125849396,-0.6316563487,-0.217475459,0.1793742329,-0.1384020597,-0.1703632921,0.1805836409,-0.0070432015,0.247763142,-0.2609244585,-0.0069709048,0.3292714059,0.0429664217,-0.5192559361,0.0393428504,0.1607203484,0.1130446419,0.2652929723,-0.0206837263,-0.2050859034,0.1075808778,-0.0601737499,-0.0842008442,-0.1182250902,0.2389039695,0.2081925273,0.0055059632,0.0628017858,0.4919819832,0.0065527819,-0.170136407,-0.0049099931,0.093305409,-0.3385925591,-0.1612519324,-0.1082477942,0.2223531306,-0.3420108557,0.3595687449,-0.0179372188,-0.0841150284,0.131893456,0.0257322527,-0.2636849284,0.0285335854,-0.0080914283,0.2732332945,0.2000536621,0.7247751355,0.0397646539,0.1726018339,-0.1615488529,-0.0414128155,0.1044561714,-0.171495527,-0.2823372483,-0.1106853038,0.0700389668,0.031754896,0.1334082037,-0.7072275877,-0.3839831352,0.006639468,-0.0326706804,0.0244258717,0.6573565006,0.045899909,-0.1025301069,-0.1390305459,-0.3231084943,0.0940262601,-0.231338203,0.1839510053,-0.3546337485]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"Hey! Happy to jump into the discussion here. I'm still getting familiar with bits of this code, but the reasons I sampled over data loaders rather than datasets is 1) ensuring that each sampled batch corresponds to only 1 task (in case of different inputs formats\/downstream models) and 2) potentially having different batch sizes per task (e.g. some tasks have very long\/short inputs). How are you currently dealing with these in your PR?","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":73,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nHey! Happy to jump into the discussion here. I'm still getting familiar with bits of this code, but the reasons I sampled over data loaders rather than datasets is 1) ensuring that each sampled batch corresponds to only 1 task (in case of different inputs formats\/downstream models) and 2) potentially having different batch sizes per task (e.g. some tasks have very long\/short inputs). How are you currently dealing with these in your PR?","embeddings":[-0.0896057039,-0.3705182076,-0.0441243164,-0.0641978309,-0.2286925614,0.0289512146,0.2676903605,0.0233601592,0.3950957656,-0.109334521,-0.1705919504,0.3203356266,-0.2015978396,0.3754858077,0.3984468877,-0.4378172457,-0.0025773386,-0.2555523515,-0.3273433149,0.3867934048,0.0030515916,0.1184610575,-0.1452497393,0.0565555431,-0.2847829163,-0.1290688217,-0.3320092857,0.1075519323,0.0970576555,-0.0115130125,0.2389508486,0.3638801575,-0.002958701,0.1261327714,-0.0001156875,-0.0413929038,0.2245578617,-0.0885122195,0.0648911446,-0.0456992388,0.1790006757,-0.3761777878,0.0587886237,-0.1425863653,0.2026034445,0.1782149673,0.1775251627,0.0471881181,0.3459135592,-0.3254346848,0.0824123248,0.320581913,-0.2082775831,0.2640051246,-0.1115657315,-0.0502124429,0.0747940615,0.3185981512,0.6831018329,-0.5693539977,-0.0752263293,0.1094844714,0.0188646726,-0.0638962463,0.1539567858,-0.0844456404,-0.3090027273,-0.2464439571,-0.436266005,0.2125426531,-0.1186305359,-0.1288326532,-0.2360724211,-0.6045432091,-0.0291276295,0.0723598972,-0.2734466493,-0.0096624047,-0.2152974904,-0.0498412289,-0.173860088,-0.1056186333,-0.0311045777,0.1421042234,0.353438884,0.5484316349,0.0771757662,0.2102186382,0.4161781371,0.0839211196,-0.1161305383,-0.2795420885,0.3173269629,0.1217986569,-0.3628462851,-0.3442519009,0.1783768237,-0.2727918327,0.2653506994,0.1431957632,0.1328389049,0.0566122085,-0.127791062,0.2421325296,0.3584896028,0.0815949067,-0.090619579,-0.1105466858,0.0110702831,-0.1044861078,0.2219539732,0.1037233397,0.1453360468,-0.0153993219,-0.4716430902,-0.084676832,-0.2308618426,0.1142077595,-0.1881165057,-0.5452076793,-0.0772759244,-0.1443603784,0.1103415936,-0.0899574906,-0.2681081891,0.1173449531,-0.0560462289,0.2535911202,-0.1384112686,0.0487506762,-0.0495746024,0.0613282546,-0.4584903717,-0.0933215618,0.2870365381,0.1720797718,0.1290307641,0.1202379391,-0.0191581734,0.144102037,0.3373575807,0.0285957996,0.0130509157,0.0109926695,0.186542213,-0.3049393296,0.0114053348,0.2314634621,-0.3490486443,-0.108992286,-0.2015261352,-0.1910812408,0.1673606187,0.0139306616,-0.2104642391,-0.2614148259,-0.5597922802,0.6944103241,-0.0327052698,-0.109348096,-0.1405427754,0.2590900362,-0.1600858718,-0.0846285,0.1551488638,0.0482387878,-0.0486482829,-0.3655715883,-0.0786841288,-0.1202846766,0.0507150069,0.2679366171,-0.2283245176,0.2701061368,0.0187885351,0.1974085569,0.2476556748,-0.476846993,-0.0329776853,0.3645942211,-0.0443459302,0.4032140672,0.1089770868,0.204736203,0.0427176319,-0.2009263635,0.2404806167,1.0242568254,-0.3771042824,0.0088222316,-0.1226192638,-0.3946356475,0.5419133902,0.3336593509,0.0873916745,-0.2646752596,-0.1131930202,0.0209318604,0.1400299817,0.068072021,0.091499716,-0.0633876696,-0.3728696406,0.045084402,-0.2132455558,-0.0941620916,-0.3640448451,0.0200417917,-0.070893757,0.3300634921,0.2768280208,-0.1019619852,0.2901110947,-0.2599795461,-0.1355253011,-0.2989259362,0.080994986,-0.0919807404,0.0196252745,-0.3017867804,-0.0435298048,0.3114897311,0.0931477696,-0.1114646718,-0.5201098323,0.2931295931,-0.0359981544,0.0615643486,-0.0562952273,0.5926681161,-0.1611738652,-0.1230566204,0.1540281773,0.3034602106,-0.2690724432,0.2271575332,0.2872880101,0.1880010813,0.1648530364,-0.0165219959,0.1333711892,0.0051811868,-0.0125397518,-0.222474739,-0.0711253434,0.4943287671,-0.1799316406,0.4873961508,-0.0446648411,0.0157403201,-0.1863206327,-0.0475344732,-0.2326929718,0.4075044692,0.3192885518,-0.179496929,0.3725852668,-0.0862381607,-0.3513277173,0.1785660088,0.0962764546,0.0001979609,0.0212551672,-0.01291491,-0.0091346251,-0.3338392675,-0.1100993603,0.335542202,0.6644569635,0.24936831,0.1772748083,-0.0286508407,-0.1969970316,-0.0943430513,0.0360674225,0.2123548388,-0.024821775,0.1570017487,-0.010826814,0.0367388166,-0.0019103896,-0.1451434344,0.1448417157,-0.2040403932,-0.0847463384,0.0830414295,-0.2019584924,-0.6090888381,-0.2475720197,-0.0173969977,-0.329018116,-0.288523823,0.1099270657,0.0824487135,-0.1586320847,0.1415619254,0.2259223461,0.649754703,-0.445845902,-0.2383549064,-0.0980748683,-0.2736441195,-0.1178834289,0.0897930041,0.4389152527,0.2987910509,0.4678762257,0.3357658088,-0.0729456469,0.0949567854,-0.3517068028,0.0355306976,-0.2930790782,0.309882462,0.1390076578,-0.1824035794,0.1960035115,-0.3770197928,-0.0679646358,-0.0490537733,-0.0927712917,-0.0687157884,-0.0261847433,0.0787532404,-0.1979088634,-0.1430408508,-0.6475348473,-0.4276148379,0.2624100745,-0.2563127875,0.0538112968,0.056188643,-0.2789657414,0.1672408581,-0.1140871122,-0.0301994085,-0.166753307,-0.1156590655,-0.0223159734,-0.3414051533,-0.1668613851,-0.4368031025,-0.0881424397,0.2031256109,-0.0194643922,-0.0194301363,-0.1868307441,0.1578907371,-0.1286054254,0.1338546425,0.2255497724,0.1400244534,-0.2490694523,0.1449827105,-0.0692463741,-0.2159969956,0.2266514599,0.310231477,0.4672707021,0.2042819411,0.021374898,0.1477927715,0.7359365225,0.2251019478,0.061652571,0.1580505669,0.2971984148,-0.0115221487,0.1084874123,-0.2090418637,0.2638306022,-0.1977590919,0.480946511,0.0856477544,0.1445294917,-0.1877961904,-0.2808532119,0.0089591537,-0.2754822969,-0.2205241472,0.2475836128,-0.2749197483,0.0085210092,-0.0770046115,-0.3628948629,-0.2862836719,-0.0392616168,-0.0072002732,0.3148202598,0.0302791372,0.163256675,-0.3991277814,-0.069873929,-0.479278177,0.1246207729,0.050041724,0.2723787129,-0.0745666921,-0.1068074703,-0.0809649453,0.1391135752,0.6718360186,-0.4680744708,-0.1201026216,0.0467568599,0.0820782483,-0.0597720817,-0.1270364374,-0.2370415628,0.0200048219,0.4522344172,0.3554956317,-0.4107632041,-0.2772546709,0.2263065726,-0.1687395573,-0.0812666044,-0.2215708196,-0.213723883,-0.1787930429,-0.1205179468,0.0229235403,0.1987098604,0.2567109764,-0.1758775711,0.1802802235,-0.0964086577,0.2749904394,0.1885722727,-0.0048555033,-0.0612270012,0.1952364445,0.1555700004,0.0769702867,-0.1139927655,-0.2888612747,0.1177170426,-0.0944041014,0.0906453729,0.2796690166,-0.0278011113,0.6893079877,0.0867532566,0.3350836039,0.2698395252,-0.084354274,-0.0158317201,-0.2550778091,0.1784933954,0.1570003331,0.2542004883,-0.1599310189,-0.2616675794,0.5813786387,0.3021725714,-0.0748824626,0.4123930633,-0.5658521652,-0.3181343377,0.3210618496,0.0082347598,0.6767171621,-0.011694707,0.2966558337,0.0964209586,-0.2917297482,0.5918288231,-0.3068065047,0.1356731504,-0.0862947926,-0.1570104957,-0.0007169754,-0.1022059917,-0.0682597011,0.2856683433,-0.2693982422,0.340690434,0.4515395164,-0.2593253553,-0.0624633543,0.4970202744,0.0402605683,-0.3902253509,-0.0533882864,0.0380669348,-0.2636736035,0.1580257714,-0.133980751,-0.2760332525,0.2424033433,0.0927569345,-0.2599033117,0.1079488099,0.0026800353,0.2147811502,0.3164937794,0.024473248,0.2476583123,-0.1895921379,-0.1502360553,-0.1708444804,-0.0608615577,0.1315781921,-0.1499564648,0.0749995932,0.102251552,-0.1870381236,0.5334513783,-0.0603728071,-0.080388926,-0.2575178146,0.220783636,-0.453248769,-0.1567564607,0.1557288468,0.172933057,0.1708424985,-0.3983150423,0.5807895064,0.1363253146,0.0160669349,-0.0007994123,0.1664056033,-0.2203896195,0.5007690787,-0.3107822239,0.0090387566,0.070701763,0.0062876847,-0.1292518079,-0.417629689,-0.0480889194,0.2977130115,0.0980949476,0.0504196137,-0.0091125024,0.1802222133,-0.1155387461,-0.030411806,-0.1128156036,-0.2936977446,0.1376633048,0.1897752583,0.2009891272,0.193451643,-0.0801460296,-0.3293387294,-0.1923337281,0.4907077253,-0.0351760536,0.2011379451,-0.0892684981,0.353466928,0.392242521,0.4573069513,-0.2312459946,0.0341992751,0.0725927949,0.0904192179,0.4238544703,-0.2111247629,0.0549249016,0.2996701002,-0.1596868485,0.0328931548,-0.2968644798,-0.1228589937,-0.2477564961,0.1177071333,0.1244917363,-0.019673733,0.0004339943,-0.1036358401,-0.0594100319,-0.0566953495,0.0161581561,-0.3881987035,-0.0707405806,0.0800229385,-0.0566110425,0.0591752119,0.0188969821,0.3053609133,0.0377403349,-0.0479899272,0.3157261312,0.1257237643,0.0871788934,-0.1531969309,-0.1950376481,0.0172451995,0.3017755449,0.1727597564,-0.0392716601,0.1613339037,-0.1671968699,-0.1590321511,0.2260904908,0.223197639,0.0400984101,-0.1603205204,0.111574091,0.0839042217,-0.0976881012,0.0128135476,0.1506920606,0.113215737,0.0610277206,0.1405307651,0.1759592295,0.4591096938,-0.2538208067,-0.0028150561,0.1472608447,-0.2417181432,0.0918258876,-0.0745332986,-0.0424235277,0.022125287,0.1766904444,0.0169104934,0.1270769387,-0.3183573484,0.2580788732,0.3101720512,0.0057896669,0.0146507006,0.0942975879,-0.0674938336,-0.1866591722,0.3170962036,0.0096133286,0.3464258611,0.0452972651,0.8753758669,-0.3337633908,-0.4275131226,0.15399459,0.4489398003,0.0568626076,-0.2221125066,0.0373577774,0.1282807738,-0.3681758642,0.1188101918,0.0164688993,-0.1489977837,0.6261688471,-0.1150686741,0.1415536106,-0.1920899153,-0.251724124,0.3088244498,0.1510937661,-0.1087191105,0.0039810147,0.0862382725,-0.0319542624,-0.0897647142,0.4948276579,0.2442640662,0.0446022898,-0.4217641354,-0.0527556911,0.06980519,0.158354938,-0.0927105844,0.1416839361,0.4574712813,-0.1465972811,0.2598172128,-0.0120581388,-0.0287763197,-0.1094954982,0.1056919396,-0.4016031921,-0.3663015068,0.4545690119,0.0963345468,-0.1090126261,0.0508982763,0.2974782884,-0.6373806596,-0.2375301719,0.3041934371,-0.1298773885,-0.0715737268,-0.1806360781,0.060531877,0.0579665639,0.6411343217,0.2447846085,0.3587518632,-0.4191928506,-0.1283710748,-0.5469255447,-0.2503887415,0.1415649354,-0.0787108317,-0.2634576261,0.1379139721,-0.0405018702,0.2539454997,-0.2701577246,0.101074405,0.2356556803,0.0957556367,-0.4365875125,0.0883835182,0.1312229633,0.1066771373,0.2530837953,-0.0953542888,-0.1709672511,0.1099364534,-0.1209902242,-0.0866635144,-0.0844486132,0.3115092814,0.2262790501,0.0481479615,0.0510243699,0.5658935905,0.008748062,-0.1783837527,0.0196677614,0.0094375778,-0.3435695469,-0.16456227,-0.0811555013,0.1542697549,-0.3199591339,0.3417582512,-0.0197361168,-0.1182359904,0.1042819098,0.0029664028,-0.2295612097,-0.0509691425,0.0050038914,0.1825328022,0.25139606,0.728310883,0.0282544047,0.0951501355,-0.0725119784,0.0025233615,0.0310272761,-0.1043358892,-0.2157590836,-0.1757676452,0.0448780283,0.0133038536,0.1238889992,-0.7102771401,-0.3642641604,0.0196176767,0.0132253896,0.084021233,0.7129796147,0.0747342482,-0.1425295919,-0.1384018213,-0.3290707171,0.0511019081,-0.2399123758,0.1911872178,-0.3383631408]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"The short answer is - I'm not! Everything is currently on a per-example basis. It would be fairly simple to add a `batch_size` argument which would ensure that every `batch_size` examples come from the same task. That should suit most use-cases (unless you wanted to ensure batches all came from the same task and apply something like `SortishSampler` on each task first)\r\n\r\nYour notebook was really inspiring by the way - thanks!","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":72,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nThe short answer is - I'm not! Everything is currently on a per-example basis. It would be fairly simple to add a `batch_size` argument which would ensure that every `batch_size` examples come from the same task. That should suit most use-cases (unless you wanted to ensure batches all came from the same task and apply something like `SortishSampler` on each task first)\r\n\r\nYour notebook was really inspiring by the way - thanks!","embeddings":[-0.0956542194,-0.4303838611,-0.046293769,-0.0674511418,-0.2578944266,0.0099473,0.2022491246,0.0762412697,0.2903968692,-0.1109249666,-0.1859128326,0.2946228385,-0.2300930023,0.3789541125,0.4100244641,-0.4520003498,0.0244413298,-0.1772062331,-0.2945506573,0.3627444506,0.0154231098,0.1428891122,-0.1547029465,0.0810638219,-0.3128835857,-0.1692617238,-0.3484938145,0.1008630618,0.137015149,-0.0246207006,0.2090304345,0.3778528869,0.0340089686,0.101393275,-0.0001114291,-0.0972870216,0.2050793469,-0.0906254649,0.1029761508,-0.0590030178,0.1640173197,-0.3679223657,-0.0027350979,-0.126950115,0.159407258,0.2251133025,0.202381447,0.0856840611,0.3988807797,-0.3097245395,0.1181306392,0.3286939263,-0.18170169,0.2379733622,-0.0835704729,-0.0625105426,0.0609891564,0.234527722,0.607134819,-0.5887556672,-0.0640491694,0.1405183822,0.0385453366,0.0214208961,0.1444118172,-0.0387359113,-0.2962240577,-0.2327660322,-0.4362361431,0.2806535065,-0.1612733603,-0.1306264997,-0.2634522617,-0.5633007884,-0.0008527267,0.001461901,-0.3300979435,0.0447877906,-0.2143581957,-0.0334663987,-0.1524393559,-0.0602673888,-0.0406011157,0.1120113805,0.3631240129,0.5616534352,0.0817182064,0.1304565072,0.378122896,0.0584178232,-0.1990291923,-0.2562061548,0.2837378979,0.1137411967,-0.3466748297,-0.3419148326,0.175062716,-0.1930901557,0.3228087425,0.1010895297,0.0934974104,0.0911375284,-0.1003000438,0.2920592129,0.3476570249,0.1223694384,-0.0566057116,-0.124246493,0.0397594348,-0.0633675605,0.2047535181,0.1504277736,0.1986275166,-0.0070350585,-0.4239506125,-0.1135905012,-0.1826789975,0.1420101374,-0.2142997831,-0.5234180093,-0.029956447,-0.1304443479,0.0822643116,-0.0567394048,-0.2853320837,0.0973653644,-0.0880560875,0.1967229098,-0.12846829,0.074834168,-0.0825338438,0.0717627406,-0.4690391719,-0.0485996157,0.3015267253,0.1074483246,0.1313358247,0.1381658614,-0.0377435125,0.1328812093,0.3394283652,0.0303127151,-0.0003308657,-0.0361671411,0.1575369984,-0.3345584273,-0.0437694192,0.1560831964,-0.4061700106,-0.1570382714,-0.1264192313,-0.2152492553,0.133074224,0.0548871644,-0.234551549,-0.2543685734,-0.457261771,0.7054930925,-0.0462399013,-0.0334073231,-0.1386663914,0.2801197767,-0.2386475354,-0.0716435388,0.1591424644,0.0954545885,0.0009730769,-0.3780564368,-0.0324621834,-0.1053331867,-0.0680371821,0.3193644285,-0.1771673262,0.189245373,0.0511361398,0.2022205293,0.2415230572,-0.4966229796,-0.033442203,0.3052602112,-0.0365063101,0.3519688845,0.0778744444,0.1823500395,0.0981876105,-0.1924562603,0.2653273344,1.0035316944,-0.3828692436,0.0220557079,-0.1274592131,-0.4010963142,0.4855629504,0.3276019692,0.0234205406,-0.3136387467,-0.1300783306,-0.0256047193,0.0993120521,0.0698810965,0.0765289366,-0.0697749406,-0.3187339902,0.0139779644,-0.1563160568,-0.135272786,-0.4236288667,0.0341507196,-0.0694630593,0.3329489529,0.3250610232,-0.1376506835,0.2176266909,-0.2634049356,-0.1004827917,-0.3129209876,0.1418480575,-0.1044748574,0.0539262258,-0.2799595296,-0.0749296695,0.33102566,0.0866704583,-0.1224152371,-0.4675805569,0.2426160723,-0.1065036058,0.0175462756,0.0072998689,0.6171317101,-0.1495946944,-0.0700064003,0.2050337493,0.2820171416,-0.2821091712,0.2547475398,0.288651526,0.2265316844,0.1838992089,0.0129133407,0.1047694832,-0.0182390753,-0.0330988839,-0.1843897551,-0.1276169866,0.5337747931,-0.1134037822,0.4237475097,0.0102806175,-0.0016260303,-0.185748294,-0.087344192,-0.2569580674,0.3532716036,0.324221015,-0.2155417055,0.3575747609,-0.0927835107,-0.3873851001,0.1576682478,0.1813380569,0.0224516224,0.0883683413,0.0224850625,-0.0439936221,-0.2892819941,-0.0826865882,0.3609900475,0.6006555557,0.3052539527,0.1733634025,0.0270603634,-0.3042919636,-0.1120566651,0.0152275572,0.1831427813,-0.0301966835,0.1528551579,0.041586902,0.0559994616,-0.0489573814,-0.166114971,0.0925271213,-0.175807789,-0.0818899274,0.0395757742,-0.1671430916,-0.5780368447,-0.2460929602,0.0441225097,-0.3262543976,-0.2550864816,0.1588418782,0.0294343699,-0.1699263006,0.1729965061,0.2292761654,0.6286779642,-0.4388614297,-0.1850003749,-0.1076191068,-0.3222373426,-0.152625069,0.1209607422,0.4160206616,0.2905067801,0.4932436645,0.2830930352,-0.0260008276,0.1047434583,-0.3773645163,0.0544515029,-0.2754267156,0.2851442695,0.1103931069,-0.2209584415,0.1578573138,-0.3922051191,-0.0500610732,-0.1231088415,-0.0999128968,-0.0815797374,-0.0265806131,0.0420912169,-0.1918628216,-0.1675719917,-0.6334230304,-0.4681071341,0.3418900371,-0.2122168094,0.10471154,0.0577006973,-0.2606832683,0.1661696583,-0.1555620134,0.027958598,-0.161150977,-0.0806015879,0.0271599293,-0.3650360703,-0.1609771103,-0.4260776341,-0.131566301,0.2117258012,-0.0857354775,-0.0181610957,-0.25038746,0.0716508627,-0.1331433058,0.1169309691,0.2675018311,0.1917032152,-0.2375511378,0.0919945017,-0.0308797769,-0.2376433462,0.2405965626,0.3334428072,0.3925411403,0.1405200511,0.03054028,0.1958531737,0.735360384,0.2564295232,0.0238263607,0.1866267473,0.2621122897,0.0246364214,0.1601391584,-0.1698180139,0.2595056593,-0.175712198,0.4492536783,0.1285618544,0.1845314801,-0.2333082706,-0.2457871884,-0.0160353761,-0.2677399814,-0.2062052339,0.2436526865,-0.2004664838,-0.0054405187,-0.0985629335,-0.3882952034,-0.2598915398,-0.0329046957,0.0177905522,0.287910223,0.0028491584,0.1643075496,-0.3194676638,-0.0556372106,-0.484513849,0.1495630592,0.0722147822,0.2689265907,-0.0660001785,-0.1026054919,-0.1133637279,0.0860087648,0.6084620953,-0.4609884322,-0.1289923787,0.0296746865,0.0947560146,-0.1248146221,-0.1021870598,-0.26836133,-0.0634319931,0.4624727964,0.3793345094,-0.3683626652,-0.3086759746,0.2527801096,-0.2167630196,-0.0553524084,-0.2354158759,-0.2197987437,-0.1814945042,-0.1528939605,0.0930824652,0.2238733172,0.1790589094,-0.1041070372,0.1205002815,-0.0705216825,0.2069912851,0.2310265303,0.0324198566,-0.0248918254,0.2021239698,0.1299928874,0.0729190707,-0.128123492,-0.3435957432,0.0520562641,-0.0787158906,0.1674800068,0.2720283866,0.0389471874,0.6369618773,0.093196325,0.3483477533,0.2464493364,-0.1000306085,0.0282848142,-0.2644092739,0.1170210019,0.1925373822,0.2408113927,-0.1332415342,-0.1824225783,0.6144344211,0.3210253716,-0.0753424913,0.391569078,-0.4802118838,-0.2988835573,0.3047889769,0.0476767868,0.6780328751,-0.0086582555,0.2885786593,0.1095239148,-0.3549047709,0.5664435625,-0.3462434709,0.15785487,-0.1039695442,-0.1910541505,0.0071838135,-0.0879500732,-0.04873145,0.2688722014,-0.3467080593,0.3435438275,0.4566832483,-0.235658884,-0.080805108,0.5609843731,-0.0415434018,-0.4095562696,-0.0828427523,0.0758704543,-0.2834270895,0.1413340271,-0.1746006459,-0.2745569348,0.2578775883,0.1064659953,-0.2738929987,0.0917182043,-0.0500287227,0.1728328764,0.275623709,0.0317588001,0.1917444021,-0.2028178275,-0.0630259663,-0.1964412481,-0.0796662122,0.1545014232,-0.1452483535,0.1055343524,0.087846674,-0.2098416388,0.5108203888,-0.0530673862,-0.1016972363,-0.2674705386,0.2604368031,-0.4042938054,-0.1775135547,0.10693454,0.13696073,0.2079698145,-0.2924810946,0.5627292395,0.090274848,-0.0058632996,0.0553343892,0.1522676647,-0.2723241746,0.545381844,-0.2550186217,0.0112165594,0.0519024432,0.0118958931,-0.148883149,-0.3637679815,-0.0874822214,0.2741844058,0.097383216,-0.0001757346,-0.013708734,0.148003757,-0.1133348048,-0.0204177275,-0.12666291,-0.3535958827,0.1792514622,0.2160956711,0.2557373941,0.2160107493,-0.0746574476,-0.2411785424,-0.1948136091,0.4819283485,0.0276688524,0.2639501691,-0.1327854991,0.3500427902,0.4172843695,0.5425971746,-0.2901419401,0.0194893237,0.045447506,0.1125650108,0.459817946,-0.2419662625,0.0767639279,0.2858804166,-0.1301107854,0.0701566562,-0.3621213436,-0.1707072854,-0.2653025389,0.1015310064,0.1083341315,0.0471891426,0.0031760449,-0.0196874589,-0.0795590729,-0.0957369581,0.0184835307,-0.3760213852,-0.0982820317,0.1537205428,-0.0695399344,0.0423936546,0.1204356104,0.2050907463,0.0466983579,-0.0693286508,0.3236749172,0.1258391589,0.0748259351,-0.152202338,-0.160260886,0.0070751803,0.2286843657,0.1364774555,-0.0615746789,0.1617529094,-0.2716623843,-0.1360904425,0.2408113182,0.1459708959,0.0706044734,-0.1579339802,0.1163192317,0.1445293576,-0.1124162376,0.0010165648,0.2037465423,0.0565463714,0.1279760003,0.1517565101,0.1645636111,0.4538230002,-0.2142958492,0.0364831239,0.142202124,-0.3709935844,0.1153039038,-0.1337974668,-0.0251028277,0.0242950935,0.1567711681,-0.009452736,0.2167471498,-0.346701175,0.2866841555,0.3190624118,-0.0313687697,0.0551364683,0.1972748041,-0.0020192596,-0.1668850482,0.2794353068,0.0713000894,0.3188148737,0.0553458855,0.8951668739,-0.2494303882,-0.4379865229,0.0884797573,0.395436883,0.0246199518,-0.2290678918,0.1032284498,0.0845765099,-0.3302845657,0.0999027342,0.041744262,-0.1267838776,0.6448740959,-0.0823335499,0.1476678252,-0.1503199935,-0.2330583632,0.3292801678,0.1781780124,-0.1130226627,0.0135086067,0.099020429,-0.1085522324,-0.1989584118,0.5276771188,0.2439904809,0.0559279211,-0.405901134,0.0070243767,0.1158634946,0.1438955218,-0.0914530531,0.1073773429,0.4503692985,-0.1493267417,0.2122389823,0.0340158828,-0.0687288642,-0.0985286236,0.0331135355,-0.4560569823,-0.3647688329,0.4568906128,0.1024256274,-0.089893274,0.042917069,0.3086785078,-0.6284584403,-0.188540414,0.2987583578,-0.1300970167,-0.0746596456,-0.2261145413,0.0887196511,0.0711193606,0.5859486461,0.218856886,0.3504165113,-0.4008607268,-0.0959887058,-0.6091045141,-0.2007196397,0.1193348542,-0.0879906416,-0.2137723267,0.1967515349,-0.1063144803,0.2746653855,-0.1899933815,0.1117385104,0.1823184192,0.0193658825,-0.4103377163,0.1525804549,0.0987507999,0.0113522336,0.3067415953,-0.0957220122,-0.1701719761,0.0845771432,-0.0756979063,-0.086051859,-0.0514211729,0.2711248696,0.1778288037,0.0453940183,0.0555258766,0.4602867961,-0.0497507341,-0.1465178132,-0.0209757499,-0.0213895962,-0.3999608159,-0.1774977297,-0.0890607238,0.1700310558,-0.3348315656,0.2429104149,-0.0452840701,-0.1060848162,0.1206903085,0.0124431914,-0.1316211969,-0.0324041247,0.0383027792,0.1873748302,0.2284657657,0.736084044,0.0374933556,0.0868958309,-0.0959964544,-0.0549130142,0.054358948,-0.0933595151,-0.2339465767,-0.1753937602,0.0794785544,0.0916746259,0.0778740942,-0.6713755727,-0.3535111547,0.0105253402,0.0101816365,0.0783306584,0.7260477543,0.0725801438,-0.1551395208,-0.1533461213,-0.3561759293,0.0600897744,-0.2716904283,0.1818745434,-0.4164089859]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"@zphang is having different batch sizes per task actually helpful? Would be interesting to know as it's not something I've come across as a technique used by any MTL papers.","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":30,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\n@zphang is having different batch sizes per task actually helpful? Would be interesting to know as it's not something I've come across as a technique used by any MTL papers.","embeddings":[-0.0969287083,-0.4622378647,-0.0760454535,0.0172497611,-0.3047629297,0.052863922,0.2263214886,0.0648956746,0.27689749,-0.1179004088,-0.1725133061,0.2042178363,-0.2584697008,0.4444765747,0.4080323875,-0.410513103,0.0141269602,-0.2733159363,-0.2500169873,0.3967761993,0.0505517311,0.0848195553,-0.1709941477,0.0634241253,-0.2406710684,-0.13419801,-0.3394118249,0.0895778686,0.1340943128,0.0273927934,0.1499073207,0.3057298362,0.0270691775,0.1451935321,-0.0001170387,-0.1031790301,0.2246217132,-0.054975871,0.1232346594,-0.0387537517,0.1176415235,-0.4690669477,0.0013242804,-0.0907583088,0.2729418874,0.1990503818,0.2408550084,0.0511381067,0.3071526587,-0.3036270738,0.0846202746,0.2716978192,-0.2392474413,0.3131467402,-0.040932253,-0.0333626755,0.0654355884,0.2598795295,0.703714788,-0.5388659835,0.0035534874,0.115961127,0.0281036198,-0.0219708756,0.1327913702,-0.0722325444,-0.2183326483,-0.2454032302,-0.3490259349,0.209401235,-0.0541955382,-0.0867402628,-0.2143402845,-0.5842268467,-0.0185684059,0.0522705875,-0.3170751035,0.1362725347,-0.2085549831,-0.0774573386,-0.1931690872,-0.1036474332,-0.0195465386,0.1008689851,0.3700278699,0.5515710711,0.0521445163,0.1643018425,0.3478100002,0.0228370391,-0.1069183499,-0.2457596809,0.2928499579,0.1259238571,-0.3565704525,-0.3113478422,0.1862839758,-0.2371802181,0.2858082652,0.0631814003,0.1380448192,0.0395162962,-0.101452969,0.2029659152,0.3857395649,0.099338226,-0.0804509521,-0.1455410868,0.0266585778,-0.1694739014,0.1910659075,0.1575957984,0.1533171386,-0.0812147111,-0.4698135257,-0.1395972669,-0.2463291883,0.1241494864,-0.1801916212,-0.554835856,-0.0691336244,-0.2043522447,0.027278794,-0.1281008422,-0.2890936732,0.1347260624,-0.0876295269,0.2149019837,-0.1360725015,0.0736108869,-0.0429180898,0.0612499826,-0.4702689648,-0.1085100099,0.2724398673,0.1882496178,0.0766225606,0.1188501492,-0.045111198,0.1172554493,0.3278563321,0.0157220773,-0.0059090392,0.0185948722,0.2151930034,-0.3085086644,-0.0937296748,0.2627733052,-0.3823623955,-0.0494957082,-0.2228003889,-0.175011307,0.1586115658,0.0130534125,-0.1333700418,-0.246748209,-0.5358598828,0.6971767545,0.0293767918,-0.0612542629,-0.1539637893,0.2474061996,-0.2110520452,-0.1469141543,0.0880283937,0.0615229718,-0.0715792179,-0.3473764658,-0.0398440771,-0.0877717286,0.0162400026,0.337949276,-0.1956971884,0.2697057724,0.0762313828,0.1210111603,0.2956900597,-0.4761198759,-0.0421399958,0.3908207417,-0.0297530852,0.356341809,0.1299750358,0.2487469614,-0.0191987436,-0.171899423,0.2716453373,0.9539659023,-0.4012833834,0.0324301347,-0.1284988225,-0.4258191288,0.5435499549,0.2887675464,0.0950734913,-0.295889318,-0.0811581165,0.0535383485,0.1472453326,0.1120333448,0.1212647185,-0.0678711608,-0.3955809772,0.0236528795,-0.1915204376,-0.1070595905,-0.4190426767,-0.0004109416,-0.0966718718,0.2867059708,0.3450793624,-0.0710882843,0.3036378324,-0.313336581,-0.0802019611,-0.3391845226,0.0749736801,-0.1129366234,0.0082677249,-0.3027859032,-0.0604139045,0.3214496374,0.1327207536,-0.158838734,-0.4637388885,0.2198480368,-0.0665505379,0.0729709342,-0.006660711,0.6013067365,-0.2415633649,-0.095002912,0.1452832818,0.2679245472,-0.2993795872,0.2369750291,0.2770886123,0.2627178431,0.2146077454,0.0410457402,0.1863824427,-0.1073941588,-0.0651388243,-0.2171696424,-0.0737280622,0.4642374516,-0.1360435188,0.4431438446,-0.0413003042,-0.0566459373,-0.2385150343,-0.0426922664,-0.2022107989,0.3690057397,0.3462931812,-0.1895817369,0.2973552048,-0.1013746783,-0.3327428699,0.1553505659,0.1189503893,0.093209289,0.0555716082,-0.0128284153,-0.0384556241,-0.2931998968,-0.1450025737,0.3609074056,0.6997567415,0.2174215466,0.1800073385,0.0275223851,-0.154819876,-0.1363481879,0.0157283265,0.1799951792,-0.0432103239,0.1741939187,0.0200825967,0.0439808704,0.0587825701,-0.1308028847,0.1434108317,-0.2014935464,-0.0472820662,0.0075202878,-0.2275293171,-0.6512371302,-0.3100253046,0.0243433788,-0.3236199021,-0.2421461493,0.0769897327,-0.0048474786,-0.1467949003,0.0919544175,0.1915386468,0.6404314637,-0.4318291843,-0.1751060933,-0.024588719,-0.2508376539,-0.1563907564,0.0936932117,0.4699907601,0.2440198064,0.444329828,0.2816878855,-0.0668450519,0.1097663939,-0.3688216805,0.0224753581,-0.2358665615,0.2514665425,0.1361660361,-0.2515865862,0.1918144226,-0.3306295872,-0.0595009699,-0.1071182191,-0.0802377835,-0.1446035057,-0.0179440323,0.0609189905,-0.2195332646,-0.1628661901,-0.6563555598,-0.427423358,0.2951755822,-0.2354214191,0.0920567811,0.0430998802,-0.2766208053,0.1610893309,-0.1153984964,0.0055053937,-0.110439226,-0.1234009564,-0.0643074289,-0.3830778301,-0.1486849338,-0.3794020414,-0.0327631086,0.2191527635,-0.0132183908,-0.0656777993,-0.1386013776,0.0997226164,-0.194048658,0.075104557,0.2888384461,0.0794515088,-0.3194803298,0.1374232322,-0.0416421033,-0.2468930036,0.1678508073,0.3495776653,0.4843294919,0.1840914935,0.0225304123,0.161805898,0.713855207,0.2392644137,0.076674886,0.1239158586,0.2778073549,-0.0276446603,0.1562187374,-0.1513791978,0.3467990756,-0.1360127628,0.4500522017,0.1239484623,0.213975206,-0.1946695894,-0.2624000609,-0.0084690843,-0.2684517801,-0.1703481525,0.2559321821,-0.255561471,0.0493648462,-0.0672018081,-0.4713046849,-0.2685273886,-0.0407105312,-0.0450105332,0.3134008348,-0.0762306899,0.190092355,-0.3363176286,-0.1390691698,-0.4542549253,0.1405400932,0.0639506653,0.3106074929,-0.0474827997,-0.0647498295,-0.0800795481,0.1626018882,0.6325606108,-0.4828881323,-0.1499164402,0.01430682,0.1030576751,-0.0869303122,-0.1750021279,-0.2449891865,0.0045389682,0.479752779,0.3700741529,-0.3253916204,-0.2802447677,0.227339983,-0.2051388323,-0.0509779453,-0.2206475437,-0.2203364074,-0.1259941161,-0.0952440649,0.0731644854,0.1776407212,0.2416189313,-0.1587636918,0.1878801733,-0.0206585005,0.2550893128,0.1945962757,-0.011928034,-0.0912824199,0.2523726523,0.1145732403,-0.0004394567,-0.0798574239,-0.2460370809,0.0720707327,-0.1097127721,0.1900295168,0.3117486835,-0.0549162216,0.682168901,0.0729086995,0.3269190192,0.2302170545,-0.0995151177,0.0434821099,-0.1924853176,0.1378151476,0.1768670529,0.2216971964,-0.1892231256,-0.2239367515,0.5282304883,0.3682228625,-0.0214326121,0.3635038137,-0.6384046674,-0.2950291336,0.2439962775,0.0369796008,0.6201880574,0.0012555739,0.3311549127,-0.0282984022,-0.3404566646,0.514526844,-0.3430120647,0.174777016,-0.0671413317,-0.1301719248,0.0227895807,-0.105239898,-0.0344008952,0.3134420514,-0.2745056152,0.335968107,0.4565877914,-0.2524255216,-0.120368734,0.5019975901,-0.0001568076,-0.3571079075,-0.0268041827,0.0257466063,-0.2453847677,0.1365209669,-0.2202273309,-0.2273559868,0.2272843421,0.1529850364,-0.3146009743,0.0430014469,-0.0222249143,0.1911665946,0.2674505413,0.0095348451,0.2880605161,-0.2060193419,-0.1112157553,-0.1548606902,-0.0664758906,0.0912941694,-0.110132128,0.0908158645,0.0855413973,-0.1879211664,0.4651316702,-0.0855835527,-0.0498766862,-0.2392313629,0.2423922271,-0.3915686905,-0.2209770232,0.2041262239,0.1990994513,0.2343909591,-0.3496171832,0.5967295766,0.1162031367,0.0651920214,0.0013030343,0.1779675633,-0.1788129061,0.5910231471,-0.3143082261,-0.0250469148,0.0401580706,-0.0617597476,-0.0313074477,-0.4336971343,-0.0760747939,0.2946577072,0.1473196447,0.0401139222,0.0717900246,0.0634470135,-0.0248606782,0.0066333748,-0.0821934268,-0.2764511406,0.2156533748,0.2148265243,0.1768285185,0.2470439225,-0.0632066876,-0.2921839356,-0.1157329306,0.467847079,-0.0305256303,0.2143304199,-0.1443568468,0.4293505549,0.480225116,0.5306457281,-0.2060042769,0.0863434151,0.1195471808,0.0916841924,0.4120353758,-0.115300253,0.0162344351,0.3485609293,-0.2005305588,0.0638985559,-0.3479442894,-0.1155063659,-0.2380967438,0.1280660033,0.1398667544,0.0241812542,0.0403776206,-0.1075509265,-0.059195593,-0.0253163651,0.0172385033,-0.3293403983,-0.067954652,0.1045309454,-0.0255148076,0.015602232,0.0902877152,0.3568157852,0.0729780048,-0.0980912223,0.3494694829,0.1336923838,0.1295309514,-0.151843518,-0.215858981,-0.0420444049,0.2591364682,0.1729898751,-0.141426295,0.1201443002,-0.2330189049,-0.1235195398,0.2052649409,0.2321888804,0.076841265,-0.1220108345,0.1369244456,0.0648540556,-0.0443277732,-0.0046797241,0.1894712746,0.0924203768,0.0212385077,0.1604591906,0.1898422241,0.4918516278,-0.2660149932,0.0257153492,0.0839059725,-0.3478931487,0.1050602421,-0.1580966413,-0.0516423918,0.019299373,0.190631181,-0.0537729487,0.1364478469,-0.2919659019,0.3588840365,0.2559456825,0.0586839616,-0.0133415982,0.1396422237,-0.039308127,-0.2043799013,0.3329131007,0.0015008618,0.3294965327,0.0227278657,0.8239639401,-0.2245136052,-0.4310368001,0.1678002477,0.4035536051,0.0226308852,-0.2492283136,0.1238541976,0.1080022678,-0.310462445,0.1577772796,0.0090817306,-0.1147008091,0.6107500792,-0.084351033,0.1686760485,-0.1848054677,-0.2317777872,0.360750854,0.0615057163,-0.1191217154,0.0314358883,0.0141739948,-0.089806363,-0.1386640072,0.5110020638,0.1978655159,0.0815287158,-0.3625520766,-0.0862528533,0.1203531995,0.2306906134,-0.1393298954,0.1242806241,0.5147861242,-0.1894363612,0.2248300314,-0.0314158686,-0.0270020235,-0.0778569356,0.0545180999,-0.4602066576,-0.4399230182,0.434615314,0.0410927571,-0.1198567301,0.0570860766,0.3916845322,-0.6262607574,-0.2196515203,0.3360734582,-0.1508597136,-0.0878331438,-0.2478646785,0.0547994524,0.0617589913,0.6547300816,0.2686652839,0.3616770804,-0.3674939871,-0.1317285597,-0.5774887204,-0.1885368675,0.0968733728,-0.0261814799,-0.2875890136,0.2382505685,-0.1182769164,0.2193915993,-0.128116712,0.1089356169,0.1749700755,0.078052789,-0.4262299538,0.1132441238,0.0924323797,0.0816536918,0.2756973505,-0.0810511038,-0.1450604796,0.0360996984,-0.1280025393,-0.148900643,-0.0980127081,0.3259688318,0.2252085358,0.092276603,0.0351395309,0.5028654933,0.028904628,-0.1926552504,-0.075686045,0.004324445,-0.3545350134,-0.163552925,-0.0584579781,0.1632131487,-0.449660629,0.3605157733,0.0492659248,-0.1113958806,0.0974153504,0.021949254,-0.0988561362,-0.0909757614,0.088994883,0.1587654501,0.2819908857,0.7396941185,0.018032141,0.04638182,-0.0954440832,0.0105421003,-0.0195192806,-0.0713995546,-0.2192038,-0.2008430958,0.0753066391,-0.038502112,0.0839715004,-0.6680579185,-0.4091371894,0.0318759866,0.0484711751,0.0724315792,0.7247980237,0.0539994873,-0.1840043813,-0.1185459569,-0.3672242463,0.0312310643,-0.2750074863,0.1617084891,-0.3264088631]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"mt-dnn's [batcher.py](https:\/\/github.com\/namisan\/mt-dnn\/blob\/master\/mt_dnn\/batcher.py) might be worth looking at.","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":7,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nmt-dnn's [batcher.py](https:\/\/github.com\/namisan\/mt-dnn\/blob\/master\/mt_dnn\/batcher.py) might be worth looking at.","embeddings":[-0.2076974511,-0.3673651814,-0.0738451183,-0.0650516152,-0.2374670953,0.04209847,0.2390546799,0.0979312286,0.2776735425,-0.070308663,-0.2149332911,0.3139759004,-0.2342388928,0.3991645873,0.3708443642,-0.4208821356,0.0562856384,-0.1494828016,-0.3255767822,0.4172060192,0.0344394632,0.0738429949,-0.1735055447,0.0347597338,-0.273057878,-0.1714401692,-0.3317059875,0.0895633325,0.2361969054,-0.0341240726,0.176637575,0.3958806694,-0.0026076692,0.1572055221,-0.0001086791,-0.0932532996,0.1682997048,-0.0620057471,0.0556789003,0.0447966419,0.1281989366,-0.3507369459,-0.0174212325,-0.1198843122,0.179868713,0.2159257233,0.1989056915,0.0314275511,0.360750407,-0.2843060493,0.1436028183,0.3558090925,-0.2178632617,0.2149711996,-0.0817069486,-0.120676659,0.0366869532,0.3027007878,0.6274591088,-0.5957628489,-0.0774922147,0.1361004412,-0.0440720059,0.083311066,0.0782675147,0.007790321,-0.1728049219,-0.2673889399,-0.4189687371,0.2541895807,-0.257733047,-0.1114530563,-0.2537297308,-0.5592598319,0.026675567,0.0300823394,-0.3480542004,0.0436244607,-0.2214386165,0.0018138153,-0.1299298257,-0.0501917601,0.0233420338,0.0938171595,0.2949436903,0.5093000531,0.0668222681,0.1126406863,0.3598817289,0.007789115,-0.1523690373,-0.2605414391,0.2257206589,0.0793277398,-0.381785661,-0.3228712082,0.1004694253,-0.2040178627,0.2584535778,0.0837872624,0.0993691161,0.0995142907,-0.128832981,0.2952461839,0.2830471992,0.0718976185,-0.0180519149,-0.1091418192,0.0804361328,-0.0543981567,0.1033044085,0.1793033481,0.1507689357,-0.0375800803,-0.3414257169,-0.0947586745,-0.2648119628,0.177894935,-0.1968309432,-0.5154265761,0.0086115887,-0.0893700346,-0.0094079534,-0.0175673254,-0.316917181,0.1632090807,-0.1169997603,0.1981626898,-0.0755866393,0.0406932086,-0.0621806942,0.0954151973,-0.4422881901,-0.0485548973,0.2628264725,0.0761643872,0.101761654,0.1111687422,-0.0222743135,0.13775222,0.3554097712,-0.0141497813,0.0453974791,-0.0336502008,0.2470337152,-0.3677882552,-0.0554333441,0.1981649399,-0.349703759,-0.1415183991,-0.103502892,-0.1207949966,0.1578953862,0.0882568434,-0.1950961947,-0.2594169974,-0.4291526973,0.6704946756,-0.0202403851,0.0121222166,-0.1852743775,0.2351790667,-0.2331383079,-0.1307143718,0.2067926824,0.0876273289,-0.068030715,-0.3218090534,-0.0605857112,-0.1136764586,-0.0493749268,0.3543159664,-0.139549017,0.2191323191,0.0560855158,0.1390577853,0.2260771096,-0.5107221007,-0.0428034998,0.3095140457,-0.0799322352,0.3148576021,0.077374801,0.2572003007,0.0988720432,-0.2225933522,0.3166373074,0.9992148876,-0.4029945731,0.0097743459,-0.1036312506,-0.4046426713,0.4877799153,0.3720866442,0.1026367173,-0.3362969756,-0.1353655159,-0.1212642863,0.0902017131,0.048293218,0.1103824005,-0.0845756009,-0.3388554752,-0.0029543338,-0.2282170057,-0.2185113281,-0.406547904,0.0264881402,-0.1055408344,0.3591872752,0.3193269074,-0.1338172257,0.2256210744,-0.2167739421,-0.1224121079,-0.3277767897,0.1804111898,-0.0869775414,0.0485275127,-0.2040988803,-0.0475766994,0.3399707079,0.1164889857,-0.1267402172,-0.3875377774,0.2835785747,-0.1153841764,0.0252404343,-0.0099662142,0.6074126959,-0.2303167135,-0.0617480688,0.2233481556,0.276881218,-0.3456274271,0.2654336691,0.3262331188,0.245673418,0.1720515043,-0.0018634588,0.0727813169,-0.067828849,-0.0613761432,-0.2026585042,-0.0566209592,0.5433101058,-0.057517983,0.4349549413,-0.0089872926,0.0424709618,-0.1521313637,-0.0899596214,-0.1792499572,0.3127523363,0.3223812878,-0.2526365221,0.3392109573,-0.0924655572,-0.3161202669,0.1837677211,0.1576503664,0.0611970164,0.1579421014,-0.0547582917,-0.1325289309,-0.2526540756,-0.1668549329,0.4109750688,0.5909309387,0.3046739995,0.1505135894,0.0363416597,-0.2415634096,-0.0879190713,0.0281429514,0.1640559435,-0.0379716679,0.1731133461,0.0734803453,0.0342617743,0.0478442609,-0.1699603945,0.1464897394,-0.1899967939,-0.0453927517,-0.0124660907,-0.1804527938,-0.5523465872,-0.2364174277,0.0144675532,-0.3520695269,-0.2914297283,0.1108384281,0.0581327416,-0.1985076219,0.1912301779,0.1443332583,0.5569744706,-0.3735994101,-0.1911146343,-0.0660159811,-0.3293915391,-0.134062469,0.1680898368,0.4217709005,0.2673872113,0.5149088502,0.2375043035,-0.0192370936,0.0916614607,-0.3239834309,0.0399805121,-0.2322693765,0.3277965188,0.1038877741,-0.1988802403,0.1909471601,-0.364737004,0.0129173463,-0.1405614913,-0.0762549937,-0.1055079103,0.0005039946,0.0165012255,-0.1929817796,-0.1729454249,-0.6458232999,-0.4851145148,0.2361215204,-0.2253522873,0.0873206332,0.0257202219,-0.2627359927,0.1622232348,-0.1915550977,-0.0228846651,-0.1234057397,-0.0983896181,0.0585860126,-0.3469005227,-0.1214597747,-0.4572554529,-0.1527115256,0.1834931076,-0.0909545645,-0.0678631589,-0.2241010368,0.0439485982,-0.0943500251,0.1211579889,0.211538583,0.1142202243,-0.2095148414,0.0529724061,-0.0378627144,-0.1711044759,0.2330944985,0.2667790949,0.3815742135,0.1345721036,0.0394301228,0.1608531922,0.675714016,0.173744306,-0.0248309709,0.1436371654,0.2006014735,0.0367774032,0.1979477704,-0.2311942279,0.2051779181,-0.173184514,0.4192678332,0.1358567178,0.1569612175,-0.1948163509,-0.2845562994,-0.0363990925,-0.2266527116,-0.1927213073,0.2402012795,-0.2727284431,0.0428051911,-0.1063239351,-0.3746302128,-0.2224036604,-0.0364886932,0.0215111542,0.2199456543,-0.0445474237,0.0826237425,-0.3357491195,-0.0255101789,-0.5168146491,0.1359252334,0.1028146967,0.3218726218,-0.0857382119,-0.0376612507,-0.1810026169,0.0449048243,0.5970202684,-0.5117413402,-0.0249312464,0.0911671966,0.0878027976,-0.1049171165,-0.0771120042,-0.2955877781,-0.09302672,0.4906239212,0.3708742261,-0.3239332139,-0.2372152358,0.239108026,-0.2319471091,-0.1035817415,-0.236010015,-0.1875895858,-0.1052303687,-0.12117403,0.0834823772,0.1817358136,0.1744107157,-0.0706619397,0.1382756829,-0.0691539198,0.2445854843,0.248391673,0.0663988665,-0.059733633,0.1888719499,0.2042692453,0.0142486217,-0.1175707504,-0.245519191,0.0035477874,-0.1021492407,0.1612408757,0.190676555,-0.0335102007,0.6854363084,-0.0237152241,0.3019927144,0.3059020638,-0.0822341517,0.0935761556,-0.2668831348,0.1283858865,0.19970873,0.2492165864,-0.1573028266,-0.1957288235,0.5134266019,0.3330728412,-0.0471026562,0.3305666149,-0.5406979322,-0.2907235026,0.3500781655,0.054904826,0.5989431143,0.0245677046,0.3231500387,0.0390670449,-0.3430183232,0.5542370081,-0.288482815,0.1300651729,-0.137060523,-0.2144306302,0.033767391,-0.0874542221,-0.0754860193,0.3302220106,-0.3693778217,0.329154551,0.4194609821,-0.1907510608,-0.0713855326,0.459371835,0.0420448892,-0.4103204012,-0.1080596969,0.1208890527,-0.238758862,0.0547281057,-0.2108647376,-0.2360800356,0.284799695,0.2105609328,-0.2530733347,0.1137635484,-0.1241538301,0.1563803107,0.2115986645,0.0499498732,0.2021616697,-0.2355986834,-0.0148323569,-0.1481189579,-0.0435935855,0.1561784297,-0.0892076567,0.1024799123,0.1393711269,-0.2243228406,0.4865354598,-0.0455993526,-0.138815552,-0.2081723362,0.2560088336,-0.4081323743,-0.1286186874,0.1002026051,0.121079728,0.2473077923,-0.2627466917,0.5085331202,0.1016072854,-0.033106558,0.0807759166,0.1394976079,-0.2634440064,0.5748269558,-0.2687936425,-0.0205280557,0.0544304252,0.0052119959,-0.1369255334,-0.3466038108,-0.0103466874,0.3355174363,0.1131227389,-0.0834679157,0.0263601635,0.1113988906,-0.0870880559,0.0703561977,-0.1201850995,-0.3428975642,0.2006440908,0.1599623561,0.2700123489,0.1919501126,-0.0980355814,-0.2834940255,-0.2316366732,0.397923857,0.0819127038,0.2328772694,-0.0907807797,0.3532246053,0.3958878815,0.5815290809,-0.318672955,-0.0211023185,0.0121023683,0.0946284458,0.4527620375,-0.2130060196,0.1195084229,0.3233140111,-0.0888117328,0.0803294927,-0.3118104637,-0.2039718777,-0.2293483466,0.0804484859,0.0539439879,0.0401582234,0.0308476333,-0.0103075188,-0.1001370773,-0.1357422024,-0.0051226937,-0.3537656367,-0.0508321971,0.1223369539,-0.0472351871,0.0340710357,0.106073685,0.2709329426,0.0080067674,-0.1260918528,0.2585589588,0.1585237086,0.1130853072,-0.1679638475,-0.1127229556,-0.056333296,0.2509329319,0.1052547619,-0.0353520252,0.1187621802,-0.3298318982,-0.0660515651,0.3152298629,0.1390660852,0.0621940196,-0.1140057594,0.1575483978,0.1775305718,-0.1673997641,0.0388699844,0.2354668826,0.076154016,0.1886115521,0.1205102876,0.107495375,0.4044182003,-0.1969071478,0.0523355976,0.1143799871,-0.4031263292,0.1029102504,-0.1603586525,0.084485814,0.0029881573,0.181211099,-0.0768313333,0.2365717143,-0.2395101637,0.3432586491,0.3479152322,0.0615955107,-0.0146556245,0.216815874,0.0713753253,-0.1913587749,0.3411487639,0.0108195767,0.3169242144,0.0479869172,0.8774011135,-0.2869527936,-0.4437455237,0.0655837804,0.415258646,0.0055032321,-0.2634510398,0.0957703814,0.0858110487,-0.3027241826,0.1233699173,0.0960182697,-0.1298991293,0.5840233564,-0.0607729629,0.1986581385,-0.1770490706,-0.2421307415,0.4069541395,0.1583691984,-0.1779643148,-0.0277700014,0.1451856196,-0.1058516353,-0.1708972156,0.568197906,0.3040748239,0.037289381,-0.4150836468,-0.0168342683,0.0439316295,0.1692272127,-0.1732147932,0.1139170751,0.3786076009,-0.1494240016,0.3404810429,0.0662068874,-0.121527411,-0.1454257667,0.0230719391,-0.4448575974,-0.3943826258,0.4050982893,0.0997112915,-0.1243741065,0.0185583904,0.2980574369,-0.5980736613,-0.1813094914,0.3696584702,-0.2196208686,-0.0581636094,-0.2430311143,0.1129714772,0.0234993156,0.702421248,0.2770282924,0.3207337558,-0.3671554029,-0.1284462363,-0.6306045055,-0.155954361,0.0843094736,-0.0676036701,-0.3216981292,0.1745730937,-0.1044036373,0.2874253988,-0.1477486491,0.2144468576,0.1302810609,0.0553060696,-0.4292401373,0.1491810083,-0.0020253214,-0.0179557297,0.3076362014,-0.1282019168,-0.1555056125,0.0524094515,-0.0283510387,-0.0405449681,-0.0967501253,0.2878812253,0.1789815426,0.1415934861,0.0615622625,0.4665371478,-0.0693525076,-0.1407132,-0.0511091612,-0.0221997313,-0.3707939684,-0.2383485436,-0.044772137,0.2014687657,-0.3661293983,0.2450284958,-0.0252390262,-0.1301497668,0.1316699237,0.0306567512,-0.1550978273,-0.0658618659,0.1591327786,0.1733117849,0.1891834289,0.6732527614,0.0294390246,0.0752260685,-0.0743890181,-0.1135599315,0.0005213352,-0.1016241461,-0.2621211112,-0.2327028662,0.0168432798,0.0198247395,0.0900606886,-0.6043961048,-0.3560249805,0.0340396427,0.0380192325,0.0516417995,0.6752895117,0.0065341359,-0.1162001863,-0.14076446,-0.3496855795,0.0313798189,-0.3007484972,0.1459585428,-0.3730647266]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"> @zphang is having different batch sizes per task actually helpful? Would be interesting to know as it's not something I've come across as a technique used by any MTL papers.\r\n\r\nI think having different batch sizes per task is particularly helpful in some scenarios where each task has different amount of data. For example, the problem I'm currently facing is one task has tens of thousands of samples while one task has a couple hundreds. I think in this case different batch size could help. But if using the same batch size is a lot simpler to implement, I guess it makes sense to go with that.","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":108,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\n> @zphang is having different batch sizes per task actually helpful? Would be interesting to know as it's not something I've come across as a technique used by any MTL papers.\r\n\r\nI think having different batch sizes per task is particularly helpful in some scenarios where each task has different amount of data. For example, the problem I'm currently facing is one task has tens of thousands of samples while one task has a couple hundreds. I think in this case different batch size could help. But if using the same batch size is a lot simpler to implement, I guess it makes sense to go with that.","embeddings":[-0.1528656781,-0.485108465,-0.0882897079,0.043889109,-0.3023048937,0.0259212796,0.2686049938,0.0912450477,0.2722305059,-0.0741184279,-0.1674394459,0.1211875752,-0.2572728097,0.4499771893,0.3634528816,-0.4062088728,0.0306342803,-0.273011893,-0.2287893146,0.4066846669,0.0369204804,0.0514032245,-0.137498185,0.0486458354,-0.2384936064,-0.1524757743,-0.3196730018,0.0555802025,0.1975445747,0.0446504354,0.114674367,0.301550895,0.047314927,0.157008335,-0.000116407,-0.0821221247,0.229915306,-0.0201301072,0.079476513,-0.0190348867,0.1309003085,-0.5154117942,-0.0374999307,-0.1042581946,0.2734789848,0.2135601193,0.2659732997,0.0325354896,0.2544078827,-0.2947160304,0.0986241624,0.3221176267,-0.297036469,0.2996921837,-0.0103497924,-0.0147421807,0.0532822572,0.2629302144,0.6999950409,-0.4993149936,0.0012701716,0.1606194526,0.0273506884,0.0273662992,0.1041140184,-0.0575499535,-0.1475427598,-0.2259554565,-0.307974726,0.2468973398,-0.0819254518,-0.0660735071,-0.2369565815,-0.6513184905,-0.007323836,0.0499326214,-0.3508639634,0.1869454831,-0.1884657294,-0.096497342,-0.158461079,-0.0536108688,-0.0076526753,0.0664810017,0.3757089078,0.5052568316,0.0134295747,0.1736917645,0.3467091024,0.0312123857,-0.1065133959,-0.2447149903,0.2853565812,0.10458491,-0.3702590466,-0.2722153366,0.1617219448,-0.2643080652,0.2712616324,0.0236643199,0.1709341109,0.0216553751,-0.0688679814,0.1701315641,0.4097624719,0.1125170588,-0.0951912776,-0.1455382705,0.0410936289,-0.177620694,0.1831089258,0.2079978585,0.1485933512,-0.1357662082,-0.4656375647,-0.0821628198,-0.256293714,0.1493212283,-0.2005464286,-0.5421450138,-0.0468392186,-0.2318474203,-0.0090745185,-0.1137375161,-0.2909087241,0.1440445632,-0.1199963987,0.2271471173,-0.160934329,0.0743538141,-0.0578991622,0.0046857088,-0.4462988973,-0.1151438802,0.2692194879,0.1677381843,0.0834861994,0.0706990734,-0.0673724413,0.1179812849,0.2836702168,-0.0043958463,-0.0329799205,0.0396016166,0.2211727053,-0.2768903971,-0.1352516711,0.3050740957,-0.3997874558,-0.022525534,-0.2343412787,-0.1942861229,0.1870928556,0.0207450055,-0.128689602,-0.2231705487,-0.5973207355,0.6830005646,0.0675176308,-0.0465238169,-0.1566710174,0.1782171428,-0.1940242201,-0.1789400578,0.0115919681,0.0831236541,-0.0616525114,-0.346134752,0.0005161235,-0.1184058934,0.0154259559,0.3973837793,-0.1924615204,0.286556989,0.094451949,0.1200645417,0.2664086521,-0.4741751552,-0.049382288,0.4507997632,-0.01909432,0.3506062925,0.1672836989,0.2602223456,0.0020203993,-0.1747531444,0.2372568995,0.8971976042,-0.4060545564,0.0835805088,-0.139383927,-0.4228145182,0.5240346789,0.3101249635,0.1059497669,-0.3104820549,-0.0698605105,0.0462274626,0.1368227154,0.0981059596,0.1255739927,-0.0577630438,-0.4325707257,-0.0024134761,-0.1919256151,-0.0683367178,-0.4964828491,0.0032335885,-0.1554141492,0.2653370202,0.3736354709,-0.0585895851,0.3415445387,-0.3247838616,-0.0682675242,-0.3196668923,0.0744401738,-0.0956722498,0.0392858349,-0.3197910488,-0.0620633177,0.3237813413,0.1203211248,-0.1444507837,-0.4255874157,0.1923217475,-0.0377244428,0.0809303075,0.0166975502,0.5754878521,-0.291128695,-0.1207504794,0.109301351,0.2480295151,-0.2801344693,0.2186086327,0.2109756768,0.266854912,0.2549037039,0.1275349557,0.2021845281,-0.120138973,-0.0972060338,-0.2197655588,-0.0321399309,0.4492028356,-0.0958451554,0.4679341316,-0.0414217934,-0.1082644835,-0.228671968,-0.0420632996,-0.2095148116,0.3518670201,0.3388044238,-0.1527584195,0.3212234676,-0.0875376463,-0.3210728168,0.1529049575,0.0971042737,0.0753896609,0.0598750114,-0.0155222835,-0.0494279452,-0.2951928079,-0.1674006581,0.3495880067,0.7306095362,0.2229883075,0.1998189688,-0.0016515177,-0.1664979309,-0.1618093699,-0.0072607663,0.1925978661,-0.019020902,0.206328705,0.0260776617,0.0107030608,0.0475568175,-0.1425853074,0.1281280518,-0.1943386495,-0.0390218683,0.0012265577,-0.2181596458,-0.6630983949,-0.3449249268,0.0396685712,-0.2795931697,-0.2296168059,0.0963747874,0.0062006586,-0.202216655,0.0974417925,0.254486382,0.6557933092,-0.4205450416,-0.1220016405,-0.0118119549,-0.2465248108,-0.1822436005,0.0804419816,0.4787092507,0.1971721351,0.42263183,0.2554949522,-0.070756413,0.0809149221,-0.3518064022,0.0257015005,-0.2472163886,0.2424323112,0.116212517,-0.235052973,0.1969836354,-0.3383502662,-0.0317518748,-0.1152959615,-0.0735269263,-0.1828646362,-0.0227659773,0.0458274186,-0.2143455446,-0.1471020132,-0.6761795878,-0.4236723185,0.3288802505,-0.2589175701,0.0831294954,0.0686502457,-0.2425622791,0.1101025715,-0.1208234131,0.0201772153,-0.1074977741,-0.1358648539,-0.1051701903,-0.3834719956,-0.1640853286,-0.4417363703,-0.0516647995,0.2295840979,0.0273398571,-0.1045602262,-0.1372971982,0.0961832255,-0.229307726,0.0718224719,0.3079031706,0.0516071208,-0.3560309112,0.1333211809,-0.0120461481,-0.2183779329,0.1295349002,0.3778129518,0.4913676083,0.1492086649,0.0247739144,0.1245520413,0.7539132833,0.2017935365,0.0599366464,0.1063675955,0.2488217056,-0.0169444662,0.1358464509,-0.1161102206,0.3977303803,-0.1366483122,0.4263159931,0.1709875911,0.1997522265,-0.2030820698,-0.19826594,-0.0447219312,-0.2616245151,-0.1858818978,0.2822232842,-0.2242667526,0.0526005924,-0.0805112273,-0.4398551583,-0.2861528397,-0.0304201413,-0.0606736615,0.2741152644,-0.0919947997,0.1816529483,-0.3203602135,-0.1597827524,-0.4608951211,0.1459830254,0.0887589753,0.3017211556,-0.0124571836,-0.0877730697,-0.073589541,0.2025708258,0.637381196,-0.4487626553,-0.1601751745,-0.0056170681,0.0606977604,-0.1331735402,-0.2124770284,-0.2515119016,-0.0027016306,0.4625549912,0.391032517,-0.3513211608,-0.2920806706,0.1962276548,-0.1626907736,-0.0559128411,-0.2322782278,-0.241131559,-0.1025428548,-0.1081378311,0.1177350581,0.1584842205,0.2345429212,-0.1683327854,0.1688388437,-0.0341643542,0.2175414115,0.1756078005,0.0023393894,-0.0844697654,0.2803276777,0.1175704896,-0.0187955517,-0.0523531474,-0.1899883598,0.0890300199,-0.165379107,0.1914774776,0.3219176531,-0.0371852219,0.6493061781,0.0971960425,0.3355260491,0.1974731237,-0.057175763,0.1065633893,-0.2064405829,0.1674438119,0.1632014066,0.195735842,-0.2039257437,-0.2500295043,0.5089167953,0.4079144895,-0.0057400656,0.3332826793,-0.6443984509,-0.2598767877,0.2356385142,0.0071036383,0.5431053638,0.0058985446,0.3782309592,-0.043667566,-0.3120466471,0.4807910621,-0.3299480379,0.166913867,-0.0498564243,-0.1427991688,0.0768362433,-0.0965186507,-0.0587039255,0.3432528377,-0.2126275897,0.3445492387,0.4961313307,-0.2248218954,-0.1798937172,0.4790877998,-0.0124899875,-0.3797692955,0.0026407568,0.029900806,-0.2305551469,0.1625847965,-0.2158857435,-0.2061110884,0.2247896791,0.1379057169,-0.3079603612,-0.0263846051,-0.0560447723,0.1925183386,0.2610128522,-0.0171634387,0.2552068233,-0.2469896227,-0.1174722239,-0.1743632257,-0.0593936294,0.053464964,-0.0917478576,0.1116493046,0.0715921819,-0.1961523145,0.4640114903,-0.075719811,-0.0211942345,-0.2242822647,0.2308786362,-0.3635624945,-0.2226501852,0.2092639208,0.1487720162,0.2319565862,-0.3144294024,0.5743622184,0.0871337801,0.0779512376,0.0109792557,0.1568933129,-0.1279672384,0.632996738,-0.2941183448,-0.0701303035,-0.0063627516,-0.0899034888,0.0158568658,-0.4475607276,-0.061900381,0.2750469744,0.137993589,0.0288331788,0.096075736,0.046322491,-0.0732135624,0.0244219415,-0.0565762185,-0.2628385425,0.2195621133,0.2410232425,0.1873136014,0.2717296779,-0.0618715473,-0.2633697391,-0.1018303186,0.4638352692,-0.0135357194,0.2427498102,-0.0998178795,0.4236365557,0.4712509811,0.5466598868,-0.2070381194,0.0737061277,0.1323154867,0.0952762216,0.3916059434,-0.0591542572,0.0413052291,0.401401937,-0.1892255247,0.0892091319,-0.3487738669,-0.1275124848,-0.2297233045,0.1274730563,0.15703848,0.0375285,0.0587672293,-0.0881555378,-0.0133891739,-0.0449567251,0.0643164665,-0.2907074094,-0.0705224797,0.0866255313,0.0191516522,0.014970676,0.1282893419,0.4145230353,0.0743789002,-0.0898252651,0.3278343976,0.1382104754,0.1572460979,-0.1297175139,-0.2171125859,-0.0814460143,0.2716127038,0.1716536283,-0.1672323644,0.0935874283,-0.2185099721,-0.0803372115,0.1846629232,0.2425148636,0.0767478794,-0.1286868155,0.1766479164,0.0730798915,-0.0234026201,-0.0032806576,0.136576131,0.0994799435,0.0391546898,0.2060373724,0.2117107809,0.5047960281,-0.2334801704,0.0165252984,0.0676290691,-0.3135483861,0.0996926352,-0.1387843788,-0.0161949042,0.028217718,0.2262631506,-0.0867814794,0.154657647,-0.3059607148,0.3273347318,0.2416606992,0.0356623083,-0.0238811169,0.1247806624,-0.0372341983,-0.2066291422,0.3375938833,-0.0289996713,0.3151317835,0.0090875346,0.7983554602,-0.2380810976,-0.3960989118,0.1543672979,0.4071560204,0.0084567955,-0.2650129497,0.1071793661,0.1345320195,-0.2611270547,0.2086908817,-0.0478960462,-0.0980667025,0.5933932066,-0.0655530691,0.1735772789,-0.1612147093,-0.2606355846,0.3475784659,0.0083999876,-0.1248348579,0.0582109541,-0.0296035744,-0.0951378793,-0.1037959009,0.5061764121,0.1873366684,0.0959926024,-0.3453290462,-0.12983419,0.1432609856,0.2453796715,-0.1899359971,0.116466634,0.4996983409,-0.1996123046,0.2697574496,-0.025633093,-0.0519575924,-0.0409294814,0.044216767,-0.4889709949,-0.4163157642,0.4167271256,0.0332930796,-0.097515963,0.0518683977,0.3810840845,-0.6517302394,-0.1931717098,0.3062868714,-0.1815160513,-0.0895547196,-0.2417754382,0.059558332,0.0392020643,0.7016068697,0.2924351394,0.3273228407,-0.3832713962,-0.0800290331,-0.5715478659,-0.1767016053,0.018201191,0.0000599523,-0.2943792045,0.2641197145,-0.1251856983,0.2216232568,-0.0940614119,0.0881457329,0.1416030973,0.0706021041,-0.4327945709,0.0911336839,0.0509616993,0.0823571831,0.3042579293,-0.0788245425,-0.133690536,0.0516081043,-0.1352268457,-0.1746715158,-0.0830981657,0.3104027212,0.2144218087,0.1087013334,0.0148897087,0.460781157,0.05151565,-0.1857535392,-0.1247479841,-0.0223559104,-0.344832927,-0.1339929998,-0.0500481799,0.1708672047,-0.4427901208,0.3610153794,0.0457413644,-0.1213214472,0.1155604348,0.0341142341,-0.0503877588,-0.0671817735,0.1217597499,0.1257563978,0.2790922821,0.699252069,0.0268937405,0.0202005357,-0.0898649916,-0.0243834946,-0.0330640189,-0.0525817014,-0.2045632601,-0.2189637274,0.0958387703,-0.0608087108,0.1070446298,-0.6410902739,-0.386803627,0.0544837564,0.0603059866,0.0588452704,0.6827635169,0.0714619309,-0.2238551527,-0.1189923137,-0.3725690544,0.0410194024,-0.2961887121,0.1502610594,-0.3555030227]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"I think that instead of proportional to size sampling you should specify weights or probabilities for drawing a batch from each dataset. We should also ensure that the smaller datasets are repeated so that the encoder layer doesn't overtrain on the largest dataset.","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":43,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nI think that instead of proportional to size sampling you should specify weights or probabilities for drawing a batch from each dataset. We should also ensure that the smaller datasets are repeated so that the encoder layer doesn't overtrain on the largest dataset.","embeddings":[-0.1033664644,-0.4331935346,-0.0278643239,-0.0140366647,-0.2707727253,0.0822168067,0.1549767107,0.0747744218,0.3294961154,-0.0903296396,-0.1661012769,0.2605519891,-0.1972150654,0.4019600749,0.353055656,-0.4648923874,-0.0536190048,-0.2005532533,-0.3664029539,0.3340199292,0.0156459548,0.0867469683,-0.1794653535,0.0190733057,-0.2719652057,-0.1105169579,-0.3949349225,0.1083944663,0.1104341596,-0.0139238499,0.1876084059,0.3442344964,0.065418452,0.0540162437,-0.000112304,-0.1213718653,0.2175680101,-0.0919179097,0.1306596249,-0.0005821201,0.078445144,-0.4705014229,-0.0077313916,-0.1358966976,0.20232898,0.1704598814,0.2904416621,0.0082693361,0.3929942846,-0.3308926523,0.0972656831,0.2530705929,-0.1532385349,0.262188226,-0.1310250461,-0.0202312507,0.1017855555,0.2531660199,0.6352720857,-0.5347996354,-0.0368329883,0.1069618836,0.0256538838,0.0307450779,0.1634944081,-0.0833835155,-0.3191305101,-0.3349867165,-0.402161181,0.2369948626,-0.1488322914,-0.0839300081,-0.2452061772,-0.5854564309,0.0016672276,0.0164770391,-0.3153260052,0.1160025746,-0.2037293017,-0.0343722031,-0.2512560785,-0.1469610482,-0.0434428975,0.1449494064,0.3343852162,0.5547481179,0.0837486237,0.1610546708,0.3695610464,0.0574746057,-0.126322642,-0.3117973506,0.2927979827,0.1079671904,-0.3691598773,-0.3550261259,0.1392053813,-0.2029859126,0.2949888408,0.1029914394,0.0250546783,0.0836864039,-0.0762171671,0.256097585,0.3717130721,0.0963822901,-0.0830019861,-0.0629903078,0.0029550462,-0.0684841126,0.1434204727,0.1766897887,0.12384516,0.003417067,-0.4794363678,-0.1296742111,-0.2586896718,0.0778095946,-0.1598022878,-0.4763876498,-0.0689308047,-0.1173167527,0.029383596,-0.0860244706,-0.2983442843,0.1183153987,-0.0927727669,0.2307393551,-0.0812855288,0.0370105356,-0.0483385436,0.0713717043,-0.4358136952,-0.0589106418,0.2667329609,0.1206187755,0.1062667221,0.1664061844,0.0139734978,0.1516633779,0.3776472211,0.0165607277,0.0059502772,-0.0649299324,0.2132984698,-0.3031078279,0.0006734845,0.1707685441,-0.3618325293,-0.145702526,-0.1903548092,-0.1629563123,0.0947004706,0.0172788873,-0.1942497343,-0.2632856965,-0.4468869567,0.7332515121,0.0268473737,-0.0171402432,-0.1802853942,0.260647893,-0.2235585898,-0.0639678612,0.2086047679,0.1247502118,-0.0169098713,-0.3514381051,-0.0467984602,-0.014809913,0.0345200635,0.361556381,-0.2216624618,0.127733022,0.061726097,0.1461113542,0.2180094272,-0.4322618842,-0.0981513187,0.406033963,-0.0164812692,0.3736613095,0.1182852909,0.1948014349,0.125598982,-0.195696041,0.2478088588,1.0400977135,-0.4225597382,0.0221931431,-0.1649434268,-0.3881129622,0.5052055717,0.3628339767,0.0946190059,-0.3206781447,-0.1631164551,-0.0266165622,0.1187750623,0.0414356031,0.1045107245,-0.0092350896,-0.4052459896,-0.0224868655,-0.2214433849,-0.0781611949,-0.3308986723,0.0152934827,-0.0874131769,0.3274636865,0.3405655324,-0.1058155596,0.2896716595,-0.2984873056,-0.1305810362,-0.3386105597,0.1144706458,-0.0395180695,0.0734675005,-0.2365604043,-0.0671809837,0.3598694503,0.0890429169,-0.138526693,-0.5096868873,0.3090384901,-0.0574002825,0.0636445209,-0.0251071099,0.5965573788,-0.1937595457,-0.1289070398,0.1448816508,0.2248516828,-0.3346215785,0.1949292421,0.3376871645,0.1734287739,0.2078784108,-0.0826792493,0.1719578356,-0.0299117044,-0.0945972428,-0.1404886842,-0.0303960219,0.500379622,-0.1195050031,0.4435099363,-0.0188361872,0.015262017,-0.2615095079,-0.0925573856,-0.2165374458,0.3594393134,0.3366407156,-0.1929520071,0.3607346117,-0.0521140322,-0.3594655395,0.1487440318,0.2037214488,0.0842020884,0.0460316949,0.020731654,-0.101657711,-0.2789615393,-0.1018219441,0.4615798891,0.5798780918,0.2789751887,0.1910253316,0.0029395046,-0.2511986196,-0.0801592097,0.0353561603,0.2158660144,-0.0584862642,0.1405456811,0.0246482417,0.0093100183,0.0106791882,-0.118537955,0.146829918,-0.1877782047,-0.0149025992,0.0633634925,-0.1983887106,-0.616386652,-0.305262804,-0.0150716519,-0.3515806198,-0.2869991362,0.1372047514,-0.0037930715,-0.1388682276,0.1657537669,0.1269964725,0.5480087399,-0.4202445447,-0.1703028828,-0.0235898457,-0.237498939,-0.1251214296,0.1078320071,0.4728634655,0.2611252964,0.4312515855,0.281291008,-0.0362429097,0.1217584386,-0.3444488943,0.0408480763,-0.2936266959,0.2540329993,0.0118084112,-0.2061464936,0.1346393675,-0.4294753075,-0.0605197847,-0.0193789825,-0.0533498451,-0.0287429355,-0.0282208417,0.0325957462,-0.1716638505,-0.2008928657,-0.6588302851,-0.4562951624,0.2835221887,-0.2002891004,0.0949147865,0.0251602102,-0.233953163,0.0898785219,-0.1671512872,-0.0019508234,-0.127606526,-0.1209832355,0.0836951509,-0.3582076728,-0.167183429,-0.4726531208,-0.1242689043,0.2418021858,-0.0711395219,-0.0734351948,-0.1877881289,0.0764655098,-0.1948927492,0.0740958601,0.2653906345,0.105487816,-0.2660581768,0.105151251,0.0029675125,-0.1324982196,0.2176911235,0.3457122147,0.4796200097,0.150229916,0.113351211,0.1909072995,0.6792353392,0.3296077251,0.0079647405,0.2135124505,0.3103106022,0.0240745042,0.1975460798,-0.21085006,0.3423010707,-0.1159056425,0.4543204308,0.1069952399,0.1511467695,-0.2443632185,-0.2434226424,0.0243287031,-0.2064081281,-0.1780358404,0.2593956888,-0.226490885,0.0201609079,-0.0977021828,-0.3649139404,-0.2355718911,-0.0543937534,-0.0312826522,0.2283368558,0.0116144167,0.1722326875,-0.3531282842,-0.1016592681,-0.4612159133,0.2046677321,0.0577432252,0.301775068,-0.0199868083,-0.0846936554,-0.1292435527,0.1270914525,0.5680521131,-0.4728297591,-0.2209475338,-0.0253830496,0.0902525112,-0.0921195969,-0.1057273,-0.2197123617,-0.0254542716,0.4199261069,0.3062418997,-0.3639726043,-0.277926147,0.2682929933,-0.1646309644,-0.0722038895,-0.1496962458,-0.2126786858,-0.1383238286,-0.170881182,0.0775838494,0.1296515763,0.1933687478,-0.1396950483,0.1670351624,-0.1088937595,0.2603108883,0.1859177798,0.052346047,-0.0275810938,0.1999797821,0.1719032973,0.0535123907,-0.1065677777,-0.2357194126,0.1216960102,-0.0325007699,0.0720008239,0.2976102233,0.0390415192,0.6938976645,0.042602405,0.3670551181,0.2250311226,-0.0647598654,0.0874698311,-0.2181842625,0.1336080879,0.2299721688,0.2078846693,-0.1668775529,-0.2211142778,0.5431402922,0.2985459268,-0.0096113849,0.3315673172,-0.5390052795,-0.2511655688,0.296756655,0.0401641503,0.6300411224,-0.0071011498,0.2938930988,-0.0329882503,-0.3371446133,0.5710727572,-0.3307278454,0.1919950694,-0.1401996166,-0.2416983098,-0.0027913779,-0.0857037306,-0.0793662742,0.276381433,-0.3371970952,0.369179219,0.4814934134,-0.2480581403,-0.1045870855,0.5050842166,-0.0564581379,-0.3726610839,-0.1111670807,0.0611550733,-0.2758666873,0.0950925872,-0.1186870262,-0.3008026183,0.2708373964,0.1216236427,-0.2423113436,0.1643852741,-0.0504326709,0.184522897,0.2530386448,0.0087022102,0.2706410289,-0.1624806225,-0.0896631479,-0.1169528514,-0.0792377219,0.1151816025,-0.1497708708,0.0128444778,0.1199054271,-0.2024590522,0.4892717004,-0.0095729213,-0.1359947026,-0.2155087888,0.3113106489,-0.3876417577,-0.1785182357,0.1087499931,0.1811384857,0.2295551598,-0.3597275019,0.578001976,0.1318092197,0.0273009799,0.0307870694,0.1418282986,-0.2512458265,0.5130012631,-0.2759882808,-0.0380961038,0.042087853,0.0073543843,-0.0880451575,-0.3510382175,-0.0318847746,0.2833531201,0.129230693,0.021556627,0.0655190647,0.1746210158,-0.0739603415,-0.0654002279,-0.1322672367,-0.3194482625,0.1565032899,0.0759360939,0.1961118281,0.1613200605,-0.0762889981,-0.2127536237,-0.2180176079,0.5112911463,0.021171933,0.2578120828,-0.1816395223,0.3082201481,0.3887069225,0.569747448,-0.2556869686,0.1172614396,0.0552382991,0.0697464719,0.4610331953,-0.2023731768,0.0436346941,0.2892815173,-0.1484194845,0.062354885,-0.3408697248,-0.1428144574,-0.2346566617,0.0988053456,0.1511790603,-0.0193569809,0.0094579151,-0.0679880753,-0.0965587944,-0.1163807064,0.0130106686,-0.4335088134,-0.0681892186,0.1105900109,-0.0371805355,-0.0454632454,0.074594073,0.2376605868,0.0012456762,-0.0200707968,0.3473860025,0.1378141344,0.132697165,-0.1667907387,-0.2762277126,-0.0126143834,0.2735834122,0.1634393781,-0.0712225139,0.1035203263,-0.2437407225,-0.0873654708,0.2628618181,0.206361711,0.0771168545,-0.1726030409,0.1514470279,0.1127183735,-0.0789476782,-0.0010414501,0.2089208364,0.071742408,0.0864508301,0.1702328771,0.1579513848,0.4282055199,-0.2442437559,-0.0075551067,0.1038690284,-0.3436974883,0.1442801654,-0.0746839494,-0.0168989524,-0.0583436713,0.1164394319,-0.0426750295,0.2247778624,-0.2821710706,0.2502270043,0.3058214188,0.0099686896,-0.0421239659,0.2031437308,0.0018030572,-0.2245410085,0.3329151869,0.0852006078,0.3855229318,0.0489542522,0.8560091853,-0.2478873432,-0.4657870829,0.0372967795,0.485901475,0.041509416,-0.2708474398,0.1433792859,0.095980294,-0.4118725359,0.1614195406,0.074984625,-0.1844240874,0.6622334719,-0.0584883951,0.16207771,-0.190612033,-0.1701134145,0.3675387204,0.1431510746,-0.1363285333,0.0741908625,0.1387913078,-0.1096967608,-0.1428728551,0.5203853846,0.2531951368,0.0331777781,-0.4345834255,-0.0220725592,0.1602118909,0.1631410122,-0.1098903567,0.1235360503,0.4092101157,-0.1710425168,0.2681719363,0.0134572927,-0.0643919036,-0.1700004637,-0.0010189215,-0.4380501211,-0.3881178796,0.4196399152,0.1148872599,-0.0514159612,0.0190206748,0.2460403442,-0.6877144575,-0.2315695584,0.3309123218,-0.1353483349,-0.1028082147,-0.2420896739,0.0723007321,-0.0153910285,0.6274802089,0.2935220897,0.3365944624,-0.4087726474,-0.1573865414,-0.5566298366,-0.142603606,0.1098332554,-0.1562363207,-0.2828677297,0.2381754071,-0.0958595425,0.2356737852,-0.2189867198,0.1331322193,0.2130482942,0.0983852446,-0.4491211176,0.1221684366,0.1000118479,0.0442810506,0.2792528272,-0.0982091203,-0.1715544015,0.0537212268,-0.1032232717,-0.0081488406,-0.0168788582,0.2455243319,0.0991366133,0.122775659,0.0672534853,0.5235670805,-0.0335241891,-0.2191504687,0.0056220717,-0.0402341075,-0.412208885,-0.1306318492,-0.0777165666,0.2438376099,-0.346826911,0.2638210952,-0.0402813815,-0.1249945834,0.1292157024,0.031964574,-0.1225517839,-0.058099065,0.0578557737,0.2173567712,0.2185801566,0.7121251225,0.0382614806,0.1066697985,-0.0289160479,-0.0241883192,0.0697462931,-0.0819060802,-0.2243683487,-0.2049362659,0.0255067721,0.0647983775,0.0754126906,-0.6441550851,-0.3265175819,0.0116075091,0.0044608521,0.0475226603,0.7026138306,0.1149973869,-0.1402956694,-0.1339814961,-0.2580882907,-0.0005192024,-0.3298357427,0.2000503391,-0.3035807014]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"Are there any references for people doing different batch sizes per task in the literature? I've only seen constant batch sizes with differing numbers of batches for each task which seems sufficient to prevent the impact of large datasets (Read 3.5.3 of the [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) for example).\r\n\r\n","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":47,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nAre there any references for people doing different batch sizes per task in the literature? I've only seen constant batch sizes with differing numbers of batches for each task which seems sufficient to prevent the impact of large datasets (Read 3.5.3 of the [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) for example).\r\n\r\n","embeddings":[-0.0643153861,-0.4659213722,-0.0846638903,-0.0310634468,-0.2937629223,0.0142678423,0.2748717368,-0.001531178,0.327396214,-0.0881961659,-0.1483937055,0.1096212044,-0.1666349918,0.4398270845,0.3444593847,-0.3664786518,0.0387695245,-0.2969401777,-0.2708536386,0.3853909075,0.0380840339,0.0419314951,-0.133543089,0.0882256776,-0.2808085084,-0.2636702061,-0.3897076845,0.0344354585,0.1399751455,0.051349543,0.2340892702,0.3075521588,0.048616495,0.2791934013,-0.0001168718,-0.0837765485,0.2308102399,-0.0191363506,0.1168078035,-0.0018158883,0.1538009793,-0.4783226252,0.0054652067,-0.1108367369,0.2739711404,0.1568047404,0.2591815591,0.1064631566,0.2886648774,-0.3447953463,0.0977224261,0.2746801376,-0.2108391523,0.2263471484,-0.1184610054,-0.0141929314,0.0773544461,0.260610193,0.7055603266,-0.541046679,-0.0727652535,0.079444997,0.0422349125,0.0119696679,0.1090794876,-0.1714967191,-0.2628783286,-0.2975201309,-0.3295665979,0.2208566517,-0.0858785659,-0.0988260135,-0.2601613402,-0.6264220476,-0.0364246666,0.0315535776,-0.3546115458,0.1694985628,-0.2045316696,-0.1291310042,-0.2346027792,-0.100240767,-0.0264209136,0.0753989592,0.3474641144,0.5395688415,0.0419702642,0.1272194982,0.3926811814,-0.0382644981,-0.0889718533,-0.2737716436,0.2567122579,0.1343369037,-0.3992006779,-0.2938930392,0.1546251923,-0.2645098269,0.3152874708,0.0821964294,0.0817464069,0.0516123809,-0.1418322772,0.1976729929,0.4014687836,0.1253283024,-0.0997797996,-0.1122030392,0.0472247489,-0.1615311056,0.2122469842,0.182415992,0.1426582485,-0.1386050284,-0.5396947861,-0.1671905518,-0.2762538791,0.1414396912,-0.1807385981,-0.5744892955,-0.0204468071,-0.1508353353,0.020509243,-0.0943684354,-0.2721249759,0.0960933492,-0.1183325797,0.2690684199,-0.1600548774,0.0736635029,-0.0505211204,0.0471133962,-0.4707634449,-0.1046871394,0.2876490057,0.1782876998,0.0942670777,0.187098816,-0.0319584049,0.0859320387,0.3025982082,-0.0619176254,-0.0510382839,0.0393544436,0.2460583746,-0.2680261731,-0.0879347697,0.2486284375,-0.3791604936,-0.0590336211,-0.2317093462,-0.1593207866,0.1585091799,-0.0022804707,-0.1443956792,-0.2504307032,-0.5782610774,0.6736926436,0.0623671338,-0.0566855557,-0.1603138298,0.2204205543,-0.1436540931,-0.1869812459,0.0599551685,0.018209774,-0.0328351818,-0.3329956234,0.036641337,-0.0352408588,0.0035878927,0.4364959598,-0.1851605177,0.2895739377,0.081008926,0.1218289509,0.266451925,-0.4424839914,-0.011515771,0.4173333347,0.0503261499,0.3307783902,0.1234065443,0.2318913341,0.044218421,-0.2073778659,0.2358899266,0.9741356969,-0.3745271564,0.0966806635,-0.0580495447,-0.4658020437,0.441191256,0.2258762419,0.0897612348,-0.2925755084,-0.0739047229,-0.0540940836,0.0573772229,0.0896451846,0.0959465355,0.0477457196,-0.4209629595,0.0964864492,-0.1834495217,-0.0505201481,-0.4460891783,-0.0105639482,-0.0827766359,0.1956135482,0.3746560812,-0.0908392295,0.3362392485,-0.326307416,-0.0925636515,-0.3413313627,0.0661156401,-0.1217467189,0.0976743922,-0.3077352345,0.0007398505,0.3354276717,0.0945950076,-0.1044937447,-0.4976728559,0.1759320498,-0.0364608504,-0.0086301519,0.0581482835,0.5408452153,-0.1912011355,-0.0841429532,0.1043066084,0.2897190452,-0.2685348094,0.2316520214,0.311085552,0.260737747,0.2113591433,0.1183923185,0.2055728585,-0.1097736582,-0.0994369611,-0.224929899,-0.1034714803,0.4607518911,-0.1345931292,0.5010110736,-0.0296973847,-0.0716863722,-0.1886713952,-0.0173037425,-0.1613265574,0.45697245,0.3395256102,-0.1707567573,0.3268986344,-0.0983332917,-0.3560428023,0.0538365319,0.022897765,0.0538638495,0.0811352506,-0.0080035813,-0.0308276545,-0.2953766584,-0.1262144446,0.3653461039,0.6546179056,0.2216181606,0.2182417661,-0.0614946783,-0.2254962474,-0.1522153765,-0.0258746129,0.2021449655,0.0066400524,0.1604983211,0.026298061,-0.0057370472,0.0847122446,-0.105338335,0.1184067503,-0.2252216488,-0.0851924419,0.0054050749,-0.2221914977,-0.6311863661,-0.3055882752,0.0150794471,-0.3049708307,-0.2313881069,0.0390989184,0.0737466365,-0.1142402738,0.0818827003,0.2115861475,0.6373746991,-0.4346647263,-0.1099825054,-0.0314767323,-0.247411713,-0.1409847885,0.0675955936,0.4684339464,0.2097015977,0.4605123401,0.3219372332,-0.1092415825,0.1142005548,-0.3225851357,0.02633816,-0.2005291432,0.2423814386,0.0437043533,-0.1586246341,0.1765051335,-0.3366844356,-0.0731435791,-0.0610848852,-0.0528527126,-0.1735841036,-0.0702571347,0.0564881675,-0.1488361657,-0.1003974155,-0.672786653,-0.3964664638,0.3388643563,-0.256965071,0.0655088797,0.0254215319,-0.1942610592,0.0753900036,-0.1441620886,-0.0177366994,-0.0749676153,-0.1228859052,-0.0291451365,-0.4028765261,-0.1595147252,-0.3732082248,-0.0216745213,0.1983463764,0.0170577336,-0.0676832348,-0.0986860394,0.029117329,-0.1789455265,0.1095878631,0.3481385112,0.0832409114,-0.3999963403,0.15140073,-0.0556364655,-0.2490878701,0.1466619372,0.2859275043,0.4913568795,0.2199198753,0.0089417072,0.1551789939,0.7662396431,0.273021251,0.0236228127,0.0985928252,0.2874363959,0.0255047753,0.1275713444,-0.118877843,0.3374462724,-0.1756728739,0.436619252,0.0679442286,0.2040231824,-0.1605823189,-0.2338019311,-0.0395273939,-0.3257651329,-0.1714380085,0.2507255971,-0.2517659366,0.0781192854,-0.1435815543,-0.431673795,-0.2927151322,-0.0169615634,-0.0676081553,0.3017232418,-0.0818984583,0.200298503,-0.3350657821,-0.1642793268,-0.3893588781,0.1740123183,-0.0023737913,0.2998749316,-0.0377968214,-0.0142647214,-0.0789808035,0.1751032323,0.5862995982,-0.4357357919,-0.1429322213,-0.0142579982,0.1116925329,-0.1409674436,-0.1962885708,-0.2515062392,0.0530730821,0.5074831843,0.3795273602,-0.3611527383,-0.316876322,0.197359249,-0.1927375942,-0.0932243466,-0.2261340171,-0.2140607387,-0.1682549566,-0.0791083723,0.1411770731,0.1395166814,0.1591914147,-0.2341115177,0.1494751424,-0.0783837661,0.2822116613,0.1764064133,-0.0188554116,-0.0787947401,0.2480200529,0.1028497815,0.0177690871,-0.0868243277,-0.1627185643,0.0372939743,-0.1313015819,0.1557683945,0.2967388034,0.0138122533,0.6570737958,0.1072499976,0.4018984437,0.200597629,-0.1156847626,0.0569849163,-0.1988922209,0.1270309985,0.2361158431,0.2462625206,-0.2228367478,-0.2472975105,0.5878696442,0.3816872239,-0.0544543788,0.4050363302,-0.624481678,-0.2605518997,0.2318006754,0.0248332415,0.59909302,-0.0131169055,0.2996223271,-0.0794400051,-0.3576934934,0.4593538642,-0.3115878403,0.1631145924,-0.0789993927,-0.0896800831,0.070220083,-0.1389348805,0.018397484,0.2944930494,-0.2356731892,0.3794520199,0.4121049345,-0.2506567836,-0.1553011239,0.4758174717,-0.0347515121,-0.3831046522,0.0556125902,0.0247368813,-0.2695636153,0.1704663485,-0.2148157805,-0.200602293,0.2319068909,0.1601424962,-0.3523415327,0.0233052615,-0.016125828,0.2278084159,0.2794571221,-0.0491650514,0.2548292875,-0.1672861725,-0.1095386222,-0.1255609393,-0.0387177877,0.1005441323,-0.1121817082,0.0360157155,0.0384791233,-0.1712269634,0.4399472773,-0.0540465303,-0.08606489,-0.2332589775,0.2231765091,-0.2976768613,-0.2963247001,0.1337808222,0.1284625381,0.1511071622,-0.3166044056,0.5542103052,0.1613519937,0.0931548402,0.0071567832,0.1634629965,-0.1255705506,0.6486728191,-0.3296866417,-0.0107023967,0.0314196497,-0.0907798856,0.0063266824,-0.4751661718,-0.0592223257,0.2901133597,0.1181517467,0.0453317612,0.0926242396,0.093167901,0.026724996,-0.0108418409,-0.1010245383,-0.2596003711,0.1727828383,0.1676546931,0.1894173324,0.1972040534,-0.0561285429,-0.2705430686,-0.1072380468,0.4969045222,-0.0192694142,0.2187630683,-0.1185727194,0.3374295831,0.4945824742,0.5996088386,-0.1953843087,0.0918989852,0.1511173397,0.0973040387,0.413069427,-0.0851140544,0.0158799998,0.3530909419,-0.2040972859,0.0378240943,-0.2977033854,-0.1182649061,-0.1918208301,0.1132503524,0.1212540194,0.04695981,0.0286998693,-0.1480216384,0.0092987968,-0.051961612,0.0610613748,-0.2568272948,-0.0778424069,0.1117901951,0.0003035945,0.0850616246,0.1450283378,0.3739887476,0.0439603217,-0.0639299676,0.3479342759,0.1078022346,0.178165406,-0.1582488716,-0.2107111067,-0.0542494208,0.2973983884,0.1429697126,-0.1791617125,0.0863332301,-0.1817284375,-0.122573331,0.1435169429,0.2863256633,0.0899970755,-0.1103153899,0.2158229202,0.0614149049,0.0060818247,0.0023050739,0.1719259024,0.1008209884,-0.0049417107,0.1974211484,0.1723508239,0.4390957355,-0.2788700759,0.029636547,0.0911108106,-0.3425987959,0.097302191,-0.1335474104,-0.0017446643,-0.0215119086,0.2189326584,0.0036816036,0.1451461613,-0.3186523914,0.3492980599,0.2836128771,0.0418655947,-0.0312635824,0.1697319597,-0.1356530637,-0.2388624698,0.2818861008,-0.0074546495,0.2914683819,0.0062579131,0.837166965,-0.2760738432,-0.4076001048,0.111873962,0.4058805704,-0.008414126,-0.241951555,0.0589165315,0.1356609911,-0.2189960927,0.18768543,0.0059454632,-0.1557210088,0.6309084296,-0.0868592337,0.1520136446,-0.1394614428,-0.1847779751,0.3204895258,0.0687221512,-0.0457346365,0.1023813114,-0.0006043201,-0.0821139738,-0.1377709061,0.5545242429,0.2713231444,0.0506941825,-0.3948325515,-0.0580212213,0.1216768995,0.1845169365,-0.0989896506,0.1296195686,0.4572369456,-0.218944326,0.2153020054,-0.0184873622,-0.0337184258,-0.0684043169,0.0327728726,-0.4699131548,-0.3582220078,0.4419861138,0.039242398,-0.1533874571,0.0986938998,0.3518231809,-0.6634213924,-0.2220765203,0.3930753469,-0.1551325768,-0.1185301691,-0.1978646219,0.0570379086,0.1237693727,0.6506407261,0.3044009805,0.3155852258,-0.3372244537,-0.1507799625,-0.5069010258,-0.2103444934,-0.0153449383,-0.0489148423,-0.3153477609,0.2121831924,-0.1383205354,0.1627191454,-0.1201684698,0.0935433581,0.2275585085,0.1809206754,-0.4607583284,0.1245459318,0.0937919468,0.105977349,0.2719216347,-0.0829593167,-0.1547923088,0.083118163,-0.1303596646,-0.0644192547,-0.0050100712,0.3701479435,0.133189708,0.0890017301,0.067532815,0.5516714454,0.0813446864,-0.161962077,-0.0657714233,0.0348555073,-0.3513208032,-0.1360027343,-0.0804252326,0.1183284149,-0.4727511406,0.3081994355,0.0680920556,-0.1501395255,0.0518047512,0.0167991426,-0.0354929306,-0.109620519,0.1088163704,0.1547782123,0.3421808481,0.7173807621,0.0134887034,0.0511242189,-0.112076655,0.0163791273,-0.0414475389,-0.1004003435,-0.2099619806,-0.242192328,0.0935538709,-0.0435492583,0.1260360777,-0.5950588584,-0.3449791074,0.0035267707,0.0569779351,0.0914710388,0.6922090054,0.0802129731,-0.1998311877,-0.1205540895,-0.3673869073,0.0513148122,-0.2942933142,0.0956128091,-0.3140668273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/217","title":"Multi-task dataset mixing","comments":"Hi,\r\nregarding building T5 dataset , I think we can use datasets https:\/\/github.com\/huggingface\/datasets and then need something similar to tf.data.experimental.sample_from_datasets, do you know if similar functionality exist in pytorch? Which can sample multiple datasets with the given rates. thanks. ","body":"It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n","comment_length":39,"text":"Multi-task dataset mixing\nIt seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).\r\n\r\nThe [T5 paper](https:\/\/arxiv.org\/pdf\/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:\r\n- **Examples-proportional mixing** - sample from tasks proportionally to their dataset size\r\n- **Equal mixing** - sample uniformly from each task\r\n- **Temperature-scaled mixing** - The generalized approach used by multilingual BERT which uses a temperature T, where the mixing rate of each task is raised to the power 1\/T and renormalized. When T=1 this is equivalent to equal mixing, and becomes closer to equal mixing with increasing T.\r\n\r\nFollowing this discussion https:\/\/github.com\/huggingface\/transformers\/issues\/4340 in [transformers](https:\/\/github.com\/huggingface\/transformers), @enzoampil suggested that the `nlp` library might be a better place for this functionality.\r\n\r\nSome method for combining datasets could be implemented ,e.g.\r\n```\r\ndataset = nlp.load_multitask(['squad','imdb','cnn_dm'], temperature=2.0, ...)\r\n```\r\n\r\nWe would need a few additions:\r\n- Method of identifying the tasks - how can we support adding a string to each task as an identifier: e.g. 'summarisation: '?\r\n- Method of combining the metrics - a standard approach is to use the specific metric for each task and add them together for a combined score.\r\n\r\nIt would be great to support common use cases such as pretraining on the GLUE benchmark before fine-tuning on each GLUE task in turn. \r\n\r\nI'm willing to write bits\/most of this I just need some guidance on the interface and other library details so I can integrate it properly.\r\n\r\n\nHi,\r\nregarding building T5 dataset , I think we can use datasets https:\/\/github.com\/huggingface\/datasets and then need something similar to tf.data.experimental.sample_from_datasets, do you know if similar functionality exist in pytorch? Which can sample multiple datasets with the given rates. thanks. ","embeddings":[-0.03330677,-0.5136845112,-0.027779568,0.0666488931,-0.2503274679,0.0426299907,0.1753954887,0.0604559705,0.2067265809,-0.1286961436,-0.259631902,0.1723333001,-0.1614574343,0.5526627898,0.3689047396,-0.4282800555,-0.0336442292,-0.2255317271,-0.3311311305,0.3586273193,0.1625532806,0.1267403215,-0.1310738772,0.0116256261,-0.3748582304,-0.0523343459,-0.3290129006,-0.0141582461,0.071282655,0.1565443575,0.1405654848,0.3437874019,-0.0457316004,0.3042807877,-0.0001206887,-0.0008575789,0.0616064556,-0.1261853129,0.112125434,-0.0386468098,0.2874489129,-0.3399040699,0.0531314574,-0.1707761139,0.1323983818,0.1373210996,0.2363459915,0.0732047185,0.4259683788,-0.2262076288,0.0427934304,0.466755867,-0.2890118957,0.2969698906,0.0031433103,-0.0374882929,-0.0336275995,0.2170954794,0.7237206101,-0.4042848349,0.0634199083,0.0402297899,-0.0138911428,-0.0113678733,0.2176286876,-0.0077172173,-0.3298627734,-0.2981940508,-0.3776238263,0.2490173429,0.0038087189,-0.1812826693,-0.2273195833,-0.5786265135,-0.0522978082,-0.008305395,-0.3058774173,0.0288804471,-0.2804423869,-0.079859525,-0.1518013328,-0.1365444511,-0.117049545,0.1804970056,0.3751618564,0.4616797864,0.069835946,0.1793563813,0.2523779571,0.135468632,-0.0544097424,-0.3006078005,0.2920220494,0.1848786771,-0.3568205833,-0.3436084986,0.0752806738,-0.3887389004,0.3567280173,0.120508261,0.1658619344,0.0104901865,-0.1414692104,0.119521901,0.3928689659,-0.0963776261,-0.1670013666,-0.1134783626,-0.0251007359,-0.2513065338,0.0701842457,0.0931534395,0.1699233353,-0.1099843606,-0.3625379503,-0.1245318875,-0.1382128596,0.0985791311,-0.0329475701,-0.5224346519,-0.2159511,-0.1834100336,-0.0349710248,-0.08714699,-0.2840888798,0.0498408563,-0.0025753472,0.2152228951,-0.2706935704,0.0415827259,-0.0405669808,0.0272745285,-0.3746246994,-0.0478326529,0.2949212492,0.0531737357,0.1146205142,0.1628558785,0.1929011792,0.1215476468,0.4489952326,-0.0281228125,-0.0100444248,-0.0239373669,0.1456849128,-0.2112798095,-0.0191193614,0.3398948014,-0.3035695851,-0.0699490383,-0.15541251,-0.1905197352,0.0961160809,-0.050944563,-0.2012615353,-0.3853225112,-0.4683860242,0.6835908294,0.0639651343,-0.0959399715,-0.2083034068,0.2052851468,-0.210993886,-0.0864719748,0.1614248455,0.0682333708,-0.0132021857,-0.3724399805,0.041983664,-0.1005393788,0.0085652824,0.3029560745,-0.1540870816,0.2219990939,0.0847674534,0.1121402457,0.3399459422,-0.4791812003,-0.0096985158,0.3526688516,-0.0426944047,0.2965771258,0.1470527351,0.2737155557,-0.0726067573,-0.1507540196,0.1296712458,0.9934558272,-0.380512923,-0.0172143113,-0.079598859,-0.3645020723,0.4801670313,0.3585850298,0.1976085901,-0.3657473922,-0.1877189428,-0.0051624207,0.0852505788,0.0067849234,0.0833077282,0.024007665,-0.2344198078,0.1030645519,-0.1965169162,-0.2335738838,-0.4084452689,0.0414350405,-0.0509235002,0.3780598044,0.3517454565,-0.0534373298,0.2586524785,-0.2782162428,-0.124343127,-0.310174644,0.0345657542,-0.0232688747,0.0006008443,-0.2645057142,-0.0932127833,0.2605561614,0.0361750498,-0.0949807465,-0.4310057759,0.4275178611,0.0226294138,-0.0346387401,-0.0316006057,0.5317655802,-0.2460516989,-0.1196652502,0.1047087237,0.2162629813,-0.2052464336,0.1573851407,0.2279515266,0.3987587094,0.239885211,-0.0832391754,0.0573171824,-0.0517178029,-0.064358592,-0.2189694494,-0.0042868373,0.4068593681,-0.2748056054,0.403734386,-0.0074854079,-0.0276026484,-0.2262642831,0.0104640126,-0.1238725707,0.4250918627,0.3405516446,-0.2496808767,0.3180718422,-0.1349045336,-0.3477177322,0.2337108999,0.126834631,0.0192520116,0.0002144778,-0.040310014,0.0153487558,-0.2949433029,-0.1368179172,0.3106762171,0.6342361569,0.1991339922,0.1362771839,-0.0617868826,-0.1840106249,-0.0617514066,0.0771273747,0.1738761067,-0.0764711946,0.2113827318,-0.0133699253,0.1067886874,0.0416921228,-0.168312341,0.2519669831,-0.1721642613,-0.1644183099,0.0485866256,-0.1701442301,-0.6914842725,-0.1162000224,0.08361651,-0.2977903485,-0.2556681335,0.1518947333,0.1305794567,-0.096426554,0.1751936078,0.1604330987,0.4907012284,-0.3046949804,-0.3377010524,-0.0032321604,-0.2629941404,-0.0777215511,0.0428458825,0.4357628524,0.418310225,0.4902553558,0.1984901577,-0.0715317875,-0.0420199893,-0.3898181319,0.1562466472,-0.3618428111,0.3624500632,0.0631991699,-0.1201191843,0.0701552555,-0.3186580837,-0.0146517139,-0.088305898,-0.1483508497,-0.0626275763,-0.0902746916,0.0764873028,-0.1306282133,-0.1616630852,-0.6078257561,-0.392036289,0.1981446147,-0.188529253,0.0759614706,0.1544653773,-0.2840279043,0.2427247167,-0.1273284703,0.0816208273,-0.1322252601,-0.1482547522,-0.0169296134,-0.4053654969,-0.2201673388,-0.4019719362,-0.0195649359,0.1676714271,0.0524572209,-0.0996366665,-0.3910073936,0.0016666166,0.0276719648,0.0583601631,0.2569244504,0.1197583377,-0.2818752229,0.168163538,-0.1412181407,-0.1873745471,0.2724559903,0.3120087683,0.3604863882,0.1784595698,-0.00215867,0.1279563755,0.9331130385,0.29474774,0.0042397054,0.1840080023,0.1302707195,0.0191586409,0.0827959329,-0.1914925128,0.2007605284,-0.1197106093,0.4675966799,0.1090208292,0.0879264623,-0.0168099534,-0.2927386761,0.095951952,-0.3363727331,-0.2571494281,0.2042870075,-0.3424231708,0.1067072898,-0.1269151717,-0.408624351,-0.3784381747,-0.0760098845,-0.0991915688,0.2852086723,0.1052639857,0.1705239713,-0.3895929456,-0.0501770936,-0.4277202487,0.1417074502,0.1219835877,0.2932144105,-0.0806671828,-0.0009065086,-0.0619512573,0.1394703239,0.576267004,-0.4894063771,-0.0926437601,0.13218458,-0.0951746702,-0.1648749709,-0.0293594394,-0.3096029758,0.0531122983,0.3403039277,0.3419826925,-0.4822686017,-0.4431769252,0.2855308652,-0.07070636,-0.0659741312,-0.166931197,-0.1628559381,-0.2332927734,-0.0486721285,-0.0859648436,0.3517221808,0.2464684248,-0.1400410384,0.2022760957,-0.0320151411,0.250880003,0.2081165016,-0.0859523043,-0.0796984732,0.1945642382,0.2547477484,0.0349110067,-0.1054151431,-0.3044124544,0.1429991871,-0.1448717564,-0.1175753698,0.2817606628,0.0620383471,0.6452471018,0.0720050856,0.3485736847,0.3418132961,-0.1439765245,0.0882176012,-0.1035491079,0.2066372186,0.2272496819,0.2593936324,-0.1113066375,-0.2703577578,0.5457837582,0.3118360639,0.0187966153,0.4668097496,-0.5664706826,-0.2669516504,0.2087347656,0.0435506292,0.7242987156,-0.0649959669,0.417283684,0.1482743025,-0.3961771131,0.5780624747,-0.2157806754,0.0419860668,-0.1446857154,-0.3205384314,0.0241160747,-0.1303834021,-0.203368634,0.2206207663,-0.2274598926,0.2735256553,0.3709965944,-0.2172449827,-0.0346691012,0.4872440696,-0.0832042098,-0.3102728128,-0.0078324731,-0.0050102687,-0.2720608711,0.1534217149,-0.1563308239,-0.2725106478,0.1802783906,0.2255573869,-0.3630880415,-0.0368645079,-0.0410724245,0.1753991395,0.2798948586,-0.0646176338,0.2281193584,-0.0589962378,-0.0443154871,-0.1844001263,-0.1313935667,0.110503912,-0.2541779876,0.0388137288,0.1308597326,-0.0886169448,0.4745720029,-0.1037547886,-0.1408064812,-0.2136499733,0.2007808238,-0.4532203078,-0.2897716463,-0.0282187834,0.0460365713,0.1446153373,-0.3120006919,0.5833745599,0.0206494089,0.055483751,-0.0323480554,0.2065171599,-0.1736719757,0.4253309369,-0.3541749418,-0.0430950597,0.0554851443,0.0381482989,-0.0492223464,-0.3953077495,0.0757807568,0.4156872928,0.1633378863,0.0314316787,0.1543500572,0.106190443,-0.185562253,0.1107318252,-0.074817732,-0.1934457272,0.1271695793,0.3633469045,0.1682660729,0.1208489165,-0.0636978373,-0.2312946767,-0.1816683412,0.5214852095,-0.014889597,0.2288407087,-0.1573705822,0.4722126424,0.4273086786,0.4741526246,-0.1921314448,0.119566597,0.1168095171,0.0899335593,0.5097170472,-0.2461033463,0.0499169119,0.265968591,-0.2102999687,-0.002056499,-0.3129921257,-0.088035062,-0.263815999,0.154606387,0.1326220185,0.017633656,0.0813220367,-0.1023319513,-0.208364293,-0.1031860933,0.1371436268,-0.2286574543,-0.0798503235,0.2186507583,0.0048443023,-0.0153452642,-0.0683045089,0.3252397478,0.0670611709,-0.003958703,0.3423421085,0.1584032327,0.1673063785,-0.1784553975,-0.1662210524,-0.0274074823,0.3266676068,0.0960848406,-0.0647890344,0.0588796735,-0.1197938323,-0.1888971925,0.1873276383,0.2522055805,0.0300345775,-0.027138764,0.2146416903,0.0322379917,-0.0368730873,0.0016331911,0.1335245222,0.0052560461,0.0875311643,0.1235710233,0.0099100135,0.4867089987,-0.1714266539,-0.0405505411,0.0002594585,-0.2595713139,0.1623914689,0.033877939,-0.054904364,-0.0183053631,0.072104916,-0.0367440246,0.1222649887,-0.3303040564,0.3012450635,0.2769789994,-0.0871913284,-0.0150711173,0.2303853333,-0.0835654512,-0.1979320496,0.3110865057,0.0794972032,0.417865932,-0.0099031497,0.9180456996,-0.3006822467,-0.4647100568,0.0114995725,0.5212082267,0.1481111497,-0.2523589134,0.0689193159,0.0899170637,-0.2977739275,0.0977388769,-0.0098519865,-0.1815325767,0.3974483609,-0.0101556433,0.1972368211,-0.130217284,-0.1488280445,0.3492199481,0.1798864901,-0.0063858442,0.083237797,0.0397266112,-0.1050305068,-0.1717853099,0.5120565891,0.0826808214,-0.0702283531,-0.5144466758,0.0228966773,0.1550083458,0.1871426255,0.1216676682,0.2046426535,0.4158383906,-0.1962637901,0.3652599752,-0.0822149217,-0.0337921642,-0.1805472523,0.0605737083,-0.4283542633,-0.4648220837,0.4682810009,0.0524262711,-0.1495815217,0.1635667533,0.2457020879,-0.6194171906,-0.2717808187,0.325309068,-0.1706252396,-0.0946717411,-0.1400815994,0.03522842,0.0151072852,0.7085173726,0.2406424433,0.4471990764,-0.3157653809,-0.0677022934,-0.5137099624,-0.1357126534,0.1859470457,-0.0755925924,-0.1854141951,0.055792205,0.0473425835,0.3172065914,-0.0829816312,0.1642474681,0.1825134009,0.1907541007,-0.5117718577,-0.0442933813,-0.0237516854,0.2337485701,0.2409577221,-0.1344553828,-0.2112939656,0.1133662686,-0.1192578971,-0.107766889,-0.1222838983,0.3494457304,0.0576205552,0.1040841192,0.0757716522,0.5753722191,0.0626541004,-0.1947407126,0.0476447977,-0.0511623621,-0.2982565165,-0.0918610767,-0.0732170492,0.1137058362,-0.3598735332,0.3518770039,0.0130550889,-0.0268467367,0.1197482944,-0.0537377745,-0.0798229352,-0.1371589005,0.060858123,0.1998551488,0.2587180436,0.7949311137,-0.1530765593,0.1390356869,-0.1364797205,0.0544778891,0.1046755314,-0.1295965314,-0.2167314589,-0.1166471615,0.0568679869,0.0448982529,0.1433512419,-0.5357412696,-0.2840762436,0.1221643761,0.0052578161,0.0174822975,0.7596600652,0.2824128866,-0.1751403213,-0.0517670438,-0.1894172281,0.0087007312,-0.2264745086,0.0978518203,-0.2878144979]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/216","title":"\u2753 How to get ROUGE-2 with the ROUGE metric ?","comments":"ROUGE-1 and ROUGE-L shouldn't return the same thing. This is weird","body":"I'm trying to use ROUGE metric, but I don't know how to get the ROUGE-2 metric.\r\n\r\n---\r\n\r\nI compute scores with :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add([lp], [lg])\r\nscore = rouge.compute()\r\n```\r\n\r\nthen : _(print only the F-score for readability)_\r\n\r\n```python\r\nfor k, s in score.items():\r\n print(k, s.mid.fmeasure)\r\n```\r\n\r\nIt gives :\r\n\r\n>rouge1 0.7915168355671788\r\nrougeL 0.7915168355671788\r\n\r\n---\r\n\r\n**How can I get the ROUGE-2 score ?**\r\n\r\nAlso, it's seems weird that ROUGE-1 and ROUGE-L scores are the same. Did I made a mistake ?\r\n\r\n@lhoestq ","comment_length":11,"text":"\u2753 How to get ROUGE-2 with the ROUGE metric ?\nI'm trying to use ROUGE metric, but I don't know how to get the ROUGE-2 metric.\r\n\r\n---\r\n\r\nI compute scores with :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add([lp], [lg])\r\nscore = rouge.compute()\r\n```\r\n\r\nthen : _(print only the F-score for readability)_\r\n\r\n```python\r\nfor k, s in score.items():\r\n print(k, s.mid.fmeasure)\r\n```\r\n\r\nIt gives :\r\n\r\n>rouge1 0.7915168355671788\r\nrougeL 0.7915168355671788\r\n\r\n---\r\n\r\n**How can I get the ROUGE-2 score ?**\r\n\r\nAlso, it's seems weird that ROUGE-1 and ROUGE-L scores are the same. Did I made a mistake ?\r\n\r\n@lhoestq \nROUGE-1 and ROUGE-L shouldn't return the same thing. This is weird","embeddings":[0.1360371709,-0.5163823962,-0.1058645472,0.4003202915,-0.0837259516,-0.1736314148,-0.3730712831,0.0904385671,0.0278450586,0.2418949902,-0.2364832163,0.201129064,0.0185962841,-0.073075369,0.05657598,-0.2852033973,0.0055105584,-0.0121435439,0.1270394474,-0.2076948285,-0.120224528,0.373644501,-0.1056409329,0.2618264854,0.0177329928,0.2201266736,0.2088925242,0.2172424346,-0.4059882462,-0.1264639348,0.1593453139,0.0436280258,0.1407730132,0.2047226131,-0.0001142584,-0.2329573482,0.1438633353,-0.1130825356,0.1307969689,-0.5953278542,0.0929474607,-0.2797275186,-0.076393269,-0.1107836738,-0.1216271371,-0.2069983929,-0.305781275,-0.1517527848,0.3890333176,0.2127696872,0.1311273426,-0.215700984,-0.0751665384,-0.0779070333,0.2592055798,-0.1458877325,0.1194248423,0.5944293737,0.2792600393,-0.0104805101,0.1109953001,0.1471078098,0.229794696,0.1018921733,0.3406981826,0.174297899,0.2830949426,-0.055155661,-0.1733991802,0.2740136683,0.150631845,-0.1756505966,-0.2071666867,0.1558533907,0.0349161997,-0.543654561,-0.2690013349,-0.107907109,0.1852721274,-0.1428599209,-0.2340891361,-0.0085080871,-0.1880713552,0.1093389839,0.0039877431,-0.0246578455,-0.099159278,0.1797499657,0.1736720502,0.0640677884,-0.5618701577,0.2438358963,-0.27815184,0.3103541136,-0.3501034081,-0.1382351071,0.1582736224,0.3849481344,0.1049436182,0.399248302,0.1927373856,-0.2031003237,-0.2617807388,0.1286067516,-0.0803721026,0.4281361699,0.1842336506,0.32833758,-0.0634373277,0.0583849773,0.1072399244,-0.0415584333,0.2296354175,-0.4406796098,-0.0633222908,0.3889028132,0.0079713454,-0.6664549708,-0.4520128369,0.0396071896,-0.3592873216,-0.1365735978,0.0907038227,0.0826283619,-0.1974706203,0.2749105096,-0.0353034921,0.2303760052,-0.3391576409,-0.1422199756,-0.1204594746,0.1955794394,-0.2020327896,0.2092438638,0.1289081275,0.1993837506,0.2680475116,0.1984259486,0.2979134321,-0.0251570921,0.2985784113,-0.2852603495,-0.0613670088,-0.1524571925,-0.0992055312,0.076664634,0.1640484333,-0.2876804173,-0.1195439696,-0.0882923007,-0.3167530596,-0.0237586405,0.1488341838,0.1334079206,0.1275358945,0.0581447482,-0.0701646879,0.3871821761,-0.1478076428,-0.2713274658,0.2108719945,-0.0548295043,-0.2792071104,-0.0908878148,0.221778959,-0.3260278702,-0.1604587734,0.0647399947,0.2390467376,-0.0650658235,0.0242237989,0.3850338757,0.0523710102,0.2161267698,-0.0920548588,-0.5852908492,0.677685082,-0.7142557502,-0.158855021,0.0080450056,-0.1488695741,-0.1315893531,0.1226200163,0.0195070263,-0.3265327513,-0.1401007622,0.3248001039,-0.006924327,-0.1028939709,-0.3365921974,-0.1186803877,0.3921813071,0.4203895628,-0.098817952,-0.0867639929,-0.1711474806,-0.0931309611,-0.002703815,0.381378442,0.0825933069,-0.1076742485,0.0879083872,-0.2046328932,0.1370293647,0.2415886223,-0.3313497603,0.3466296792,0.0850111991,-0.8778423667,-0.2402244061,0.4288021624,-0.2223017812,-0.706831634,-0.183924824,0.3250961304,-0.2990161777,0.112100251,-0.2064466178,0.3820970654,0.1248767823,-0.0847370178,0.1515344977,0.0083483364,-0.0219821017,-0.3105199337,-0.1406908929,0.1239418089,0.0720035359,0.2307563424,0.4563622177,0.4261050224,0.2524348199,0.1592902243,0.1775385141,0.2088915855,0.0717332363,0.2579978406,0.2365656495,0.1299290359,0.0037377048,-0.0925157443,0.4717117548,0.3099349141,-0.1613876373,-0.1957311183,0.4262432456,-0.473316282,0.0939860344,0.0262501668,-0.0144007728,-0.2124695778,0.0698103011,-0.408259809,0.0098474091,0.3337866068,-0.0767876059,0.1822853535,-0.2051221281,0.1247776523,-0.0584801696,0.2170782536,0.0171241052,-0.0512602627,-0.0924127996,0.0467981957,-0.06954173,-0.0964706764,-0.2101807296,0.1588654518,0.1439983696,0.1898375601,0.2017472088,-0.4167002141,-0.1971834004,-0.0422565676,-0.0887625813,-0.01096028,0.0425124243,0.0738901496,-0.0693168193,0.1129042432,-0.3143056631,-0.2715390623,0.1085753292,-0.324747622,0.1273213029,-0.0732392296,-0.1440971494,-0.1805692464,-0.4490249455,-0.1194290146,-0.2918983698,-0.0237174835,0.0617339499,0.3302869201,-0.2375909835,0.2181945294,0.2446240634,-0.055244159,0.3127217591,0.1232608333,-0.2930848002,-0.2357357889,0.1551420838,-0.2570504546,0.0031138652,0.2244349271,-0.1865047216,-0.0021259286,0.1485210508,-0.2689663172,-0.1071644351,-0.0120314555,0.1915832758,0.0359528624,-0.2909518778,-0.2327234149,0.330465287,0.1824977696,0.0188585036,-0.2729045153,0.0855205357,-0.17728284,0.1845188737,0.0448267162,-0.2233637571,0.0927831456,-0.053020291,0.5192238092,0.2406984717,-0.0376402698,0.2506181896,-0.331682831,0.2432362735,0.2582475245,0.4326589108,-0.3346865177,-0.3030149937,0.3488309979,-0.3870767653,-0.3465188146,-0.132283181,0.0735061243,0.0334181152,-0.0712343976,0.0300410353,-0.3585003614,0.0627359077,-0.0915831402,-0.1772151291,0.1375743002,0.2289388031,-0.3084358573,-0.0582156107,-0.1119265556,0.0988582745,0.0076421029,0.089752458,0.4865014851,-0.2721476853,0.125447154,0.158712253,0.3103817105,0.4348253608,0.449968785,0.0349394977,0.169075951,0.0145353833,0.0727576911,-0.1380455941,0.2822895646,0.0532210805,-0.1514517665,0.1277768761,0.1827632934,0.3677946925,-0.2821440101,-0.0774629265,-0.1221547276,0.0897745416,-0.0839898735,-0.2758703828,-0.0543606393,-0.0512115806,0.0221824199,-0.1187617257,-0.103745766,0.2709281743,0.0343674161,0.1774481535,0.1985769272,-0.3399243355,-0.047721941,-0.2667633891,0.4255414307,0.1029199734,-0.0283726361,0.1192408502,0.0972564518,0.0267865211,-0.2130282372,0.4061413109,0.1378523111,0.2272479981,0.2927297056,-0.2416942716,-0.6836922169,-0.1355914474,-0.3325050175,0.2740961611,0.161105141,0.0643063039,-0.1818179339,0.0935281739,0.1751582175,-0.0231398959,-0.2003567815,0.0539879017,-0.2955207825,-0.022973815,0.3432687521,-0.203514263,-0.1359041333,0.4902084172,-0.0525476672,0.3217339814,0.1378912032,-0.243312493,0.4190601408,0.0064634099,0.2591437101,0.0326131955,-0.222124204,0.2137242705,-0.1269951016,-0.6886135936,0.3523054421,0.1802395284,-0.6067422032,-0.1504710168,0.0103780311,0.4904713333,0.0521420538,-0.0478900149,-0.0440528579,-0.4857300818,-0.0511750914,-0.5080962181,-0.0873180032,0.2749328613,0.0087928521,-0.1041446775,-0.3016369045,0.5127965808,0.3558304012,-0.319501549,0.3271287382,-0.2311628312,-0.1869178265,0.2276449651,0.098293677,0.8402042985,0.2155022919,-0.0027391692,0.0663598254,-0.345328033,0.5277203918,-0.7417638898,0.0318238959,-0.2215472162,-0.3050875962,-0.0354462564,0.0388358869,0.0533297174,0.2313740999,0.075881727,0.5425155163,0.0226756968,-0.1676210314,0.0255906638,0.2087966502,0.1377377063,-0.0567585565,-0.187543422,0.1364589185,-0.1479850113,-0.0328673795,-0.2879722714,0.1271541119,-0.5454553962,0.0604605973,-0.123026602,-0.159808591,-0.1440057307,0.1570232362,0.1355050355,-0.4175264537,-0.3413097262,0.5611543655,0.6825196147,0.124435924,-0.3415252566,0.2535166442,-0.2157330215,0.118878983,0.1012442932,-0.0549407154,0.0577697083,-0.0421962291,-0.0438630842,0.3054068089,0.3313338757,-0.0065215984,-0.1633487344,0.0119619314,-0.1587856859,0.0725871995,-0.0391067155,0.1823221594,0.3115078807,0.0647634268,0.0897965878,0.230540067,-0.5638209581,-0.0620018691,0.2300350368,-0.2607152462,0.135207966,0.3556466103,0.0188001581,-0.2615954578,-0.221123606,0.0425205939,0.0530391634,0.006930077,-0.0388279557,-0.0389435105,-0.0516916811,0.0426022522,0.2408222705,0.1702113301,-0.0265549496,0.4584594667,0.3475965559,-0.1032896936,0.2224103957,-0.4771330655,-0.1962073296,0.1745526493,0.0762402862,-0.0950755775,-0.0163693856,-0.1346662939,0.1467684805,0.5626370907,-0.2810869813,-0.2149291039,-0.0777937174,0.1860041022,-0.1478135735,-0.2056770772,-0.1962113082,0.1369066238,0.0635420009,0.0913083926,-0.177254647,-0.1769512296,-0.2398809791,0.1427944303,-0.1575234979,0.1655424982,-0.0665867776,0.087037228,0.1815344244,-0.0881256238,-0.078183502,0.3319510221,0.3922883272,0.3993208706,-0.1100582257,0.1136449948,-0.1639082432,-0.0554028228,0.1831081808,0.2382262349,0.0480203852,-0.0347425193,0.1367378682,-0.0292313285,-0.2743677497,-0.065928109,-0.1672274172,-0.3081829548,0.2224558294,0.0084562618,0.2681055665,-0.2286851555,0.3888053298,0.1158234477,-0.0850733221,0.1656814665,-0.1622015834,0.1318025738,-0.115722999,0.3632443845,0.2054444402,0.2189696431,-0.2337181866,0.2338652462,-0.0470110774,0.1256176829,0.03593456,-0.0395274349,-0.0057363529,-0.318908155,0.1475919783,-0.3983777463,0.1694095731,-0.0912348926,0.2875302732,-0.1240280047,0.1184249595,0.2121197432,0.1277419031,0.2352429479,0.0020537255,0.4505256712,0.0284678116,0.2130913138,-0.2889871597,0.2680079937,0.0412848666,-0.0466873795,0.0461019389,0.1021155342,0.0978180319,-0.1739476919,-0.1319649518,0.0961580873,0.1269912273,-0.3811546862,-0.343639195,-0.2659492195,-0.2507070303,-0.0339107402,-0.0206719395,0.1195200458,-0.1125296205,0.0543072596,-0.0392487198,-0.2851574719,-0.3589871824,0.1040612757,0.3963492513,-0.1169946045,0.1518862844,0.1834174544,0.3502518237,0.3349504769,0.3218497634,0.1066421568,0.0562255979,-0.5301019549,0.0718588606,-0.1844808906,-0.0982018039,0.0361762159,0.3242705166,0.1913615614,-0.1691660285,0.3245176673,0.1348386407,-0.1030669063,0.3589080274,-0.1090165898,0.0368221886,-0.4506005347,0.4781188667,0.0480122492,0.1968317926,-0.2053103596,0.0397672243,-0.4899892211,-0.1937710345,-0.5417338014,0.1262193322,0.0806682482,-0.3077897727,0.0717704445,0.0968761221,0.0702999607,0.0469024293,0.3951916397,-0.3121458292,-0.2617610097,-0.3850396872,-0.301379323,0.0073791537,0.0810415596,-0.2019940317,0.2370334119,-0.0590975471,0.5858445168,0.1110456139,0.5105576515,0.0628203824,0.1424970627,-0.2336207181,-0.0157498159,-0.0755769685,-0.1632595509,-0.1421504617,0.224569574,-0.1736003906,-0.2012470216,0.0401781388,-0.1555122584,-0.2476873696,0.2316203713,0.7587873936,0.2767795622,0.3365957737,0.427384764,-0.0142284548,0.07858105,-0.0036376233,0.0616065748,-0.0023558077,0.2013890892,-0.3358893991,0.2631808221,-0.2623026669,-0.2294840813,0.179830566,0.6258000135,-0.0025758522,-0.2092171311,-0.3737502992,-0.1586409211,0.011391419,-0.1415488124,-0.4665008783,0.4006671309,-0.0023264051,0.4249284267,0.0061101993,-0.1299139559,0.4667284489,-0.3223503828,0.084335193,0.1299363822,0.3993564546,0.256772995,-0.1368271559,-0.1708006561,-0.0156345014,0.376863867,0.1625566036,-0.0041798456,0.184895888,-0.1698684841,-0.0668415427,-0.1546534896,0.0645657927,0.0596984178,0.073769331,0.108484365,-0.187854737]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/216","title":"\u2753 How to get ROUGE-2 with the ROUGE metric ?","comments":"For the rouge2 metric you can do\r\n\r\n```python\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\nscore = rouge.compute(rouge_types=[\"rouge2\"])\r\n```\r\n\r\nNote that I just did a PR to have both `.add` and `.add_batch` for metrics, that's why now this is `rouge.add(lp, lg)` and not `rouge.add([lp], [lg])`","body":"I'm trying to use ROUGE metric, but I don't know how to get the ROUGE-2 metric.\r\n\r\n---\r\n\r\nI compute scores with :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add([lp], [lg])\r\nscore = rouge.compute()\r\n```\r\n\r\nthen : _(print only the F-score for readability)_\r\n\r\n```python\r\nfor k, s in score.items():\r\n print(k, s.mid.fmeasure)\r\n```\r\n\r\nIt gives :\r\n\r\n>rouge1 0.7915168355671788\r\nrougeL 0.7915168355671788\r\n\r\n---\r\n\r\n**How can I get the ROUGE-2 score ?**\r\n\r\nAlso, it's seems weird that ROUGE-1 and ROUGE-L scores are the same. Did I made a mistake ?\r\n\r\n@lhoestq ","comment_length":56,"text":"\u2753 How to get ROUGE-2 with the ROUGE metric ?\nI'm trying to use ROUGE metric, but I don't know how to get the ROUGE-2 metric.\r\n\r\n---\r\n\r\nI compute scores with :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add([lp], [lg])\r\nscore = rouge.compute()\r\n```\r\n\r\nthen : _(print only the F-score for readability)_\r\n\r\n```python\r\nfor k, s in score.items():\r\n print(k, s.mid.fmeasure)\r\n```\r\n\r\nIt gives :\r\n\r\n>rouge1 0.7915168355671788\r\nrougeL 0.7915168355671788\r\n\r\n---\r\n\r\n**How can I get the ROUGE-2 score ?**\r\n\r\nAlso, it's seems weird that ROUGE-1 and ROUGE-L scores are the same. Did I made a mistake ?\r\n\r\n@lhoestq \nFor the rouge2 metric you can do\r\n\r\n```python\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\nscore = rouge.compute(rouge_types=[\"rouge2\"])\r\n```\r\n\r\nNote that I just did a PR to have both `.add` and `.add_batch` for metrics, that's why now this is `rouge.add(lp, lg)` and not `rouge.add([lp], [lg])`","embeddings":[0.0825815573,-0.4401053488,-0.0926093534,0.3807037771,-0.0594926998,-0.1371209025,-0.3506770432,0.0906175971,0.0235020909,0.2352947742,-0.2559311986,0.1684113145,-0.0215512644,0.0022067709,0.0974013582,-0.3178662658,0.037422806,-0.0347548649,0.1625474393,-0.2055855244,-0.1583114564,0.3564315438,-0.0662797838,0.2676070035,0.0549102947,0.2283007354,0.1845670342,0.2020991892,-0.4124771357,-0.1735796481,0.1488842368,0.0051129377,0.2301985919,0.2997406125,-0.0001134505,-0.2268581986,0.1634779423,-0.0996424481,0.108243227,-0.5389806032,0.068422243,-0.2803028226,-0.1171009764,-0.0503603891,-0.077018328,-0.2652797699,-0.3086017072,-0.1595863998,0.3000632823,0.1558423191,0.136604473,-0.2055416405,-0.0622154027,-0.1061264127,0.2618375719,-0.1017468497,0.1145922467,0.5414428115,0.2648924589,0.0035370674,0.1269048452,0.1456472129,0.2574687004,0.0655306354,0.3002898097,0.2113054693,0.3186536133,-0.112101078,-0.1920291781,0.2805584967,0.1296021342,-0.1342128664,-0.3024221957,0.1002639085,0.0776538029,-0.5770205855,-0.2875574231,-0.1081864834,0.1591961682,-0.2118961662,-0.1913761348,-0.043241445,-0.1544982642,0.1063505411,0.0733333454,-0.0112977447,-0.0536589101,0.1988221407,0.184065178,-0.0033984715,-0.5394589901,0.2152126282,-0.3142087162,0.2942136228,-0.3065520525,-0.0951433256,0.1869513541,0.4066514373,0.1032526344,0.3489806354,0.1195347905,-0.258071661,-0.2762503028,0.0872939229,-0.0699587241,0.4045149088,0.1960490644,0.3381276429,-0.0284422114,0.0692702308,0.1954560876,-0.0628730208,0.2469357103,-0.4772729576,-0.0635710508,0.4770826995,-0.0077677444,-0.5834079385,-0.4588001668,-0.0287328344,-0.3379411995,-0.20796974,0.1575303078,0.1054935157,-0.2072849274,0.3233111203,-0.0341942124,0.2794035077,-0.3262607753,-0.1136728674,-0.1501775235,0.2062861472,-0.202261284,0.2235945165,0.1539401561,0.1776582748,0.2562233508,0.2119543105,0.2691634595,0.0598442294,0.2062682062,-0.2372273058,-0.1175263748,-0.1481491625,-0.034479402,0.0774111077,0.188406527,-0.2724808753,-0.1226747632,-0.0750541613,-0.3015033007,-0.0087420894,0.1592018306,0.1332612932,0.1258597076,0.0164227914,-0.2135912329,0.3548077345,-0.1503959894,-0.2694523036,0.2013503164,-0.0721282884,-0.2583304048,-0.1263648421,0.2293128222,-0.326433748,-0.1708662957,0.0443102866,0.2678412199,-0.0854357705,0.0589666888,0.3666197956,0.001542429,0.1980131865,-0.1521051228,-0.592318356,0.665184021,-0.7773889303,-0.0657990947,-0.0005517157,-0.1136176586,-0.1577291936,0.1390874684,0.0293319412,-0.296353519,-0.2013741434,0.3400664628,-0.0311039146,-0.1501097679,-0.3194093108,-0.0926349238,0.3367892504,0.4528051913,-0.1628918052,-0.0424935967,-0.1178130805,-0.0854873806,-0.0417402908,0.4088159502,0.069665581,-0.1271550059,0.0218936987,-0.1831500381,0.1643564403,0.2299869061,-0.2909383178,0.2540509999,0.1076658666,-0.9638232589,-0.2275743484,0.4533105195,-0.254757911,-0.6550403833,-0.2037677616,0.3090624213,-0.2672728002,0.0978610367,-0.1616641879,0.3661294878,0.0615179986,-0.1269472837,0.0966967791,0.0601762086,-0.004192079,-0.3270719051,-0.1326188743,0.1051758304,0.0788128972,0.2529537976,0.4529838562,0.4550014734,0.2661785185,0.1476079077,0.2340745926,0.2281712145,0.1169733554,0.3165963292,0.2823441923,0.1658865958,0.068170093,-0.0669259802,0.4591511488,0.3160399795,-0.1627162695,-0.195754379,0.4131931067,-0.4846227467,0.1242772117,0.0279041398,0.0121889245,-0.2059757113,0.1272017211,-0.4066193402,0.0046629691,0.2397213429,-0.1088741049,0.1383101642,-0.2654803693,0.0767369121,-0.0614245497,0.1953823417,0.0215802733,-0.0366844498,-0.1353598088,-0.0005478906,-0.0433564112,-0.0906143486,-0.2082996368,0.1854825616,0.1760426611,0.2232332528,0.1223328337,-0.3682170212,-0.1841252595,-0.0535663925,-0.0573278256,0.0160507746,0.0237289872,0.111025393,-0.0454005003,0.138350904,-0.2840763628,-0.2592444718,0.1057535112,-0.3508684337,0.100574173,-0.0994009376,-0.1100445762,-0.1776327342,-0.5282853842,-0.1343939006,-0.3477758169,-0.0261526201,0.0755158067,0.2976169884,-0.2034307271,0.2193649858,0.1803251803,-0.1081453189,0.3281554878,0.1060339138,-0.2106098682,-0.2679209113,0.1780754328,-0.2215532362,0.0079398891,0.2463553399,-0.1296310127,-0.0070060128,0.1543599516,-0.2549619675,-0.0985813141,-0.0539383031,0.2266653478,0.0624235868,-0.2883571386,-0.2040559649,0.3045315444,0.23586604,-0.0245086085,-0.2616884112,0.0340933166,-0.1985142529,0.2060337663,0.0521294214,-0.1808520854,0.0633005723,-0.0806743503,0.4839153886,0.2462226748,-0.0729641691,0.2791992724,-0.3118186295,0.2519313097,0.3141075969,0.3775237501,-0.3375565112,-0.2846821249,0.3150162697,-0.4162310362,-0.3131348789,-0.0967826098,0.0625382811,-0.0130871767,-0.0851631984,0.0659299865,-0.3470347226,0.0388141908,-0.0413189791,-0.1230345145,0.1854137927,0.2472903132,-0.2666775286,-0.0591413565,-0.1068544686,0.0815363601,-0.0134485196,0.1035355926,0.5398586988,-0.278263092,0.1178236455,0.1310110092,0.3444875777,0.3714714348,0.4343833029,0.054828763,0.1600300223,0.0291566886,0.0896737799,-0.1720827818,0.3043068349,0.0772539824,-0.1583983898,0.0700278357,0.1549829394,0.4240893722,-0.2869345546,-0.032307636,-0.1081596166,0.1067900062,-0.0826005861,-0.2662002146,-0.0396020487,-0.0825940073,0.0422179438,-0.126497522,-0.1020594612,0.3080261946,0.0362890996,0.2076334804,0.1973802149,-0.3730540574,-0.0860797465,-0.3050500751,0.4734287262,0.0915925577,-0.0030344566,0.1242779419,0.0989555866,-0.0170555133,-0.1943629086,0.4446807802,0.0360861979,0.1932586879,0.2945865393,-0.1920692325,-0.6826241612,-0.1762093902,-0.3117485046,0.277772218,0.1699585915,0.0862139687,-0.1744393408,0.0578868687,0.1932872832,0.0027846845,-0.1897595972,0.0271974653,-0.3189562559,0.0152819064,0.3253988028,-0.2173693627,-0.1152776107,0.5167660117,-0.0853353441,0.324963659,0.0988218114,-0.2580694258,0.4550522566,-0.0021247326,0.2430943251,0.1178696603,-0.1818492115,0.1593848169,-0.1389511824,-0.5745753646,0.3348114491,0.090355143,-0.5968511105,-0.1088005155,0.0081194993,0.482616663,0.0303758774,-0.0505130477,-0.0839702189,-0.4418610632,-0.1123769879,-0.5094912648,-0.138792783,0.2701759338,-0.0065773972,-0.1140909418,-0.3621301055,0.5230096579,0.3907353282,-0.331646204,0.3504809737,-0.2329187691,-0.2078519613,0.2235562652,0.0786143243,0.7750360966,0.2355679572,-0.031628184,0.1210622266,-0.2569797039,0.5335091352,-0.650067389,0.0733489916,-0.2151560336,-0.2684257627,-0.0399848036,0.0600940585,0.0482110716,0.1822722852,0.0801441073,0.5681715608,-0.0600794703,-0.1525988132,-0.0309202801,0.2536117435,0.0996656939,-0.1294346154,-0.1757190675,0.1674544066,-0.1329161972,0.0001763422,-0.2708412409,0.1150747463,-0.5599823594,0.0669050217,-0.1474509686,-0.1370246857,-0.1837145835,0.1779040247,0.1358850002,-0.4225407243,-0.3310341239,0.5675348639,0.7250471115,0.0667937025,-0.2974849343,0.263471514,-0.1750845164,0.1154583693,0.0364187434,-0.0925655141,0.0564440079,-0.0283969212,-0.0201182067,0.3248104453,0.3469849825,-0.002153184,-0.1701786667,0.0094260061,-0.1436962932,0.044150494,-0.0592580922,0.2411259413,0.3410102427,0.0754311457,0.0871415883,0.2268550545,-0.5656334758,-0.0526458584,0.2004537284,-0.1952412575,0.1426816136,0.2905634046,0.0375765301,-0.2935366035,-0.2215372622,0.0875626802,0.0789480805,-0.0058351476,-0.0464159884,0.0164737776,-0.0633491427,0.0182739068,0.2872958779,0.1504456252,0.0399607457,0.5065686703,0.3182367682,-0.1310816854,0.2522671819,-0.4538986981,-0.2241073996,0.1905702055,0.084865883,-0.0875309557,-0.0893733501,-0.0839216858,0.1378519088,0.5422516465,-0.2892592847,-0.2375673801,-0.0423994996,0.1442876607,-0.1634004116,-0.1392491609,-0.1074954271,0.1407331377,0.0643899739,0.0449677408,-0.186126098,-0.1794566959,-0.2494577616,0.131387502,-0.1495238543,0.1688950956,-0.0628471822,0.0750807077,0.1846788377,-0.0817184746,-0.0588736124,0.3747639656,0.3735157251,0.3379548192,-0.0343931168,0.1021811441,-0.0563245267,-0.0178300422,0.2132943571,0.2078013867,0.079009451,-0.0085378848,0.1438590735,-0.0168682262,-0.2875352204,-0.0898592249,-0.1477959603,-0.2268956006,0.1447622627,-0.0185693931,0.2065712214,-0.213333413,0.3798777163,0.1326868683,-0.0760844052,0.189731136,-0.1829057783,0.142856732,-0.1198012605,0.4061802328,0.1726002097,0.253685534,-0.2150344849,0.258207351,-0.033284422,0.1541899443,-0.0118915886,-0.0089817969,-0.0274722297,-0.3529624641,0.210236609,-0.4512999654,0.1374830753,-0.0760852173,0.345169425,-0.1417327672,0.0866359919,0.0946650058,0.1131738797,0.2061865181,0.05709498,0.3819496036,0.0586539693,0.1647116691,-0.3350286186,0.2503337264,0.0372933857,-0.0560350604,0.0605785251,0.1348668635,0.0403822847,-0.1946031898,-0.1217291877,0.1045911089,0.1230269596,-0.393794626,-0.3598280549,-0.2577284276,-0.197532028,-0.0568852685,-0.0044242335,0.0933867991,-0.1468286216,0.056323994,-0.059599638,-0.2923721373,-0.3988260627,0.1562704146,0.3441057205,-0.0743312165,0.1079124883,0.2061595321,0.3127463758,0.315790683,0.3679594696,0.132741347,0.0896670148,-0.5088692904,0.0551005639,-0.1820521355,-0.0749365911,0.036594905,0.2758986056,0.1929275095,-0.1755716056,0.3166462481,0.1541430354,-0.1053234041,0.3391842842,-0.1422563344,0.0593743064,-0.4119406343,0.5525143743,0.0364964232,0.2078888714,-0.2044390887,0.0591734238,-0.4736988544,-0.1873233318,-0.5115628242,0.0655957311,0.0615225546,-0.3025524914,0.0779673457,0.0898442864,0.0240850076,0.086866051,0.4211097956,-0.3376673758,-0.1917365193,-0.3841718733,-0.3622975647,-0.0484893918,0.1175469309,-0.2421547025,0.2872101367,-0.0677459612,0.5412329435,0.1583778411,0.4834292233,0.0539926738,0.1774774939,-0.2007937133,-0.0122562805,-0.0912612528,-0.1045594141,-0.1025995612,0.2286397368,-0.2144898325,-0.1760859787,0.0368311182,-0.1579037309,-0.2663093805,0.2467842996,0.691580832,0.3145761192,0.3558573127,0.4649412632,-0.0191299058,0.0392693952,-0.0585152134,-0.0018532984,-0.0558214113,0.1583451927,-0.3841409981,0.3125037849,-0.3025662899,-0.2276663333,0.1016861424,0.6001902223,0.0094135115,-0.2009996623,-0.3931234777,-0.0709826574,0.0184791498,-0.1498626024,-0.4123916626,0.3378522694,-0.0151268151,0.4427173436,-0.0096193589,-0.131726861,0.4783711135,-0.3655971587,0.067218259,0.0616267845,0.3842909932,0.2618566453,-0.0777329877,-0.1968768686,-0.0338130929,0.3542476594,0.1926382184,-0.0012845766,0.1645161361,-0.1842230558,-0.1039483994,-0.1597597301,-0.0198018067,0.0633615553,0.0416535474,0.1020370051,-0.1657766998]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/216","title":"\u2753 How to get ROUGE-2 with the ROUGE metric ?","comments":"Well I just tested with the official script and both rouge1 and rougeL return exactly the same thing for the input you gave, so this is actually fine ^^\r\n\r\nI hope it helped :)","body":"I'm trying to use ROUGE metric, but I don't know how to get the ROUGE-2 metric.\r\n\r\n---\r\n\r\nI compute scores with :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add([lp], [lg])\r\nscore = rouge.compute()\r\n```\r\n\r\nthen : _(print only the F-score for readability)_\r\n\r\n```python\r\nfor k, s in score.items():\r\n print(k, s.mid.fmeasure)\r\n```\r\n\r\nIt gives :\r\n\r\n>rouge1 0.7915168355671788\r\nrougeL 0.7915168355671788\r\n\r\n---\r\n\r\n**How can I get the ROUGE-2 score ?**\r\n\r\nAlso, it's seems weird that ROUGE-1 and ROUGE-L scores are the same. Did I made a mistake ?\r\n\r\n@lhoestq ","comment_length":34,"text":"\u2753 How to get ROUGE-2 with the ROUGE metric ?\nI'm trying to use ROUGE metric, but I don't know how to get the ROUGE-2 metric.\r\n\r\n---\r\n\r\nI compute scores with :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add([lp], [lg])\r\nscore = rouge.compute()\r\n```\r\n\r\nthen : _(print only the F-score for readability)_\r\n\r\n```python\r\nfor k, s in score.items():\r\n print(k, s.mid.fmeasure)\r\n```\r\n\r\nIt gives :\r\n\r\n>rouge1 0.7915168355671788\r\nrougeL 0.7915168355671788\r\n\r\n---\r\n\r\n**How can I get the ROUGE-2 score ?**\r\n\r\nAlso, it's seems weird that ROUGE-1 and ROUGE-L scores are the same. Did I made a mistake ?\r\n\r\n@lhoestq \nWell I just tested with the official script and both rouge1 and rougeL return exactly the same thing for the input you gave, so this is actually fine ^^\r\n\r\nI hope it helped :)","embeddings":[0.1239903495,-0.4979508221,-0.088984549,0.3463133872,-0.0841549039,-0.2206081003,-0.3729408681,0.0803578719,0.0360454321,0.2210490704,-0.2184779197,0.2334054857,-0.0111088082,-0.0014882199,0.0768619254,-0.2843629718,-0.01158862,-0.0326346718,0.0853345841,-0.2482199073,-0.1077575386,0.4054566026,-0.1544699073,0.2524574697,0.0654295012,0.3166961372,0.1700794101,0.2143239379,-0.4328656495,-0.1821267009,0.1652713269,0.0503630638,0.1496954411,0.2326778769,-0.0001142661,-0.2925375402,0.1849433631,-0.1414395273,0.1738614589,-0.6005346179,0.0583251007,-0.2691497505,-0.0811973214,-0.0645638555,-0.111022912,-0.1932000518,-0.2885150015,-0.1273674369,0.3743640482,0.2322625369,0.128094852,-0.2020400912,-0.0613144375,-0.0511284806,0.2469224036,-0.0883085206,0.0855068937,0.560139358,0.2660447061,-0.0374686196,0.0814225748,0.1509750187,0.231712684,0.0097445082,0.3903579712,0.1786261052,0.2437990755,-0.0938263983,-0.2231394649,0.3274884224,0.1833605468,-0.1542437077,-0.2336723953,0.1524391025,0.0272758584,-0.5276965499,-0.2502237856,-0.0963611454,0.2021198571,-0.1536117941,-0.2424559593,-0.0307536498,-0.2164042443,0.0850343779,0.0001797684,-0.0070787435,-0.0669669732,0.1687986851,0.2031937093,0.0118653523,-0.5337010026,0.2507768869,-0.3072173595,0.3494211733,-0.370957464,-0.1641173661,0.1327220947,0.4145557284,0.1002429798,0.4011620581,0.2246950269,-0.1404869854,-0.2672238052,0.1099193469,-0.0330820121,0.446931988,0.2299090624,0.3148649633,-0.0278021004,0.0251831245,0.148238197,-0.0226450935,0.2313371748,-0.4569521844,-0.0380098373,0.3978686333,-0.0274011027,-0.6806560159,-0.4113565981,0.0514224954,-0.321680963,-0.1402047426,0.1125793308,0.0980577096,-0.192164585,0.3302403986,-0.0078441193,0.2373747081,-0.3315191865,-0.1402775198,-0.0784643069,0.228780821,-0.2513328195,0.2152004093,0.1888292134,0.215188086,0.2448729575,0.2056077719,0.3032256365,0.0292105619,0.2829253376,-0.2480921745,-0.0271209087,-0.1494309306,-0.0716711283,0.0604845621,0.1127492487,-0.2844979167,-0.0919369161,-0.0552065,-0.2921645641,0.0097009111,0.1461049765,0.1334361136,0.0784384087,0.0127959736,-0.1120268926,0.3773163557,-0.2059684694,-0.26619941,0.2162994444,-0.1462043226,-0.282631129,-0.125677973,0.2128628939,-0.3043458462,-0.1561815441,0.0847705603,0.2528171241,-0.05326242,0.0146218482,0.406099081,0.0686085597,0.1522846371,-0.0909740776,-0.523879528,0.6502172947,-0.7507551908,-0.0961676016,0.0344627537,-0.080551222,-0.1089170799,0.0822352841,0.0219223499,-0.3146571219,-0.1447133422,0.281575948,0.0107191913,-0.0870100632,-0.3107838333,-0.1504411697,0.3970875442,0.4027325213,-0.0719477087,-0.0827456638,-0.1770658493,-0.1273582876,-0.0169230569,0.4076839685,0.0375565253,-0.1250233799,0.1138971671,-0.196759209,0.1464032382,0.272482723,-0.3125997484,0.3493408561,0.0773144439,-0.8894868493,-0.1724050641,0.4622159898,-0.2256922275,-0.6651101708,-0.215151459,0.3109326363,-0.3100345731,0.1021758616,-0.2257461399,0.3516809642,0.113977775,-0.1536433101,0.206938535,0.0619798601,0.0069304379,-0.3190974891,-0.1620740891,0.1061552316,0.0585050546,0.2559978664,0.4742427468,0.4499497712,0.2268804908,0.168827042,0.2143981606,0.2538842559,0.0349721834,0.258299619,0.2095764726,0.1959803849,-0.0184683632,-0.0774406195,0.5199013948,0.3037608266,-0.1541365236,-0.1943849921,0.4662745595,-0.4903933406,0.0574423149,0.0301229544,0.0035356553,-0.234408319,0.0489471927,-0.4269281328,0.0082103843,0.3378639221,-0.072214745,0.1508091837,-0.2143831402,0.0776554719,-0.0587222949,0.1669665128,-0.0231587589,-0.0700709522,-0.0991679057,0.048283577,-0.0524791405,-0.0871328712,-0.1464817375,0.1367138177,0.1424647272,0.1714953184,0.2192103863,-0.4117482305,-0.1840170622,-0.0079123853,-0.1532541811,0.0121837473,0.0154098012,0.0615092665,-0.0738395602,0.0518397205,-0.3049067259,-0.3022273779,0.0679023713,-0.3260828555,0.0888367966,-0.101811111,-0.1412836909,-0.1743475646,-0.4596986473,-0.1507118344,-0.3067997098,-0.0374746583,0.073218368,0.299755156,-0.2984479964,0.2118731886,0.2086891681,-0.0963942558,0.3506272435,0.0990061983,-0.2984908819,-0.2188930213,0.1485156566,-0.2297235727,0.0109957904,0.2069463879,-0.1500375867,-0.0032068694,0.1473244578,-0.2885562181,-0.0999546126,-0.0252586491,0.164885357,0.0557277091,-0.3303431571,-0.2132251412,0.3580795825,0.157057777,-0.0432526506,-0.2501104772,0.0918164849,-0.1292127669,0.1360247433,-0.005322271,-0.2049795836,0.0756529272,-0.0168042276,0.4890827537,0.2321534604,-0.0259796474,0.2851453722,-0.3540248871,0.212523669,0.2565449774,0.3914870322,-0.3267377317,-0.2725757062,0.3942364752,-0.435858041,-0.3721909821,-0.1029465869,0.1051651612,0.029439263,-0.0444234647,0.0571284555,-0.3295976818,0.0541082136,-0.0908435509,-0.1752023697,0.1330196112,0.2248006463,-0.2669880986,-0.0401029736,-0.1279033422,0.096050106,-0.002311877,0.1314290166,0.5086541176,-0.2481072396,0.0772690028,0.1404628903,0.2756562829,0.4029237926,0.4460240901,0.049148038,0.1697974652,0.0227131192,0.0836678892,-0.1309285164,0.282029748,0.0421498567,-0.1210120469,0.100321807,0.1847018003,0.3316663206,-0.341480732,-0.0804625005,-0.0918758437,0.0923556685,-0.0638107806,-0.2443143725,-0.0343742669,-0.0548488833,0.0570918918,-0.0976630896,-0.1349437088,0.3102974892,0.0227056108,0.2167982012,0.1691870838,-0.3748894334,-0.0289550219,-0.3219945133,0.4225338399,0.0653614104,-0.0376358218,0.132422626,0.0652796701,0.0062130392,-0.2297491431,0.3863012791,0.0636502355,0.2125621885,0.2652561963,-0.2225480527,-0.6708530784,-0.1808366477,-0.3159849346,0.2895922959,0.1978675127,0.0230554231,-0.1561221033,0.0831149593,0.1846843809,-0.0281131025,-0.1870182902,0.0456978194,-0.2581055164,0.0099988915,0.3646014929,-0.2292433381,-0.1427789479,0.4782606959,-0.0607299358,0.3421844542,0.1498375982,-0.2154518962,0.4536171257,0.047775425,0.2763142884,0.0265907254,-0.2080786079,0.162491262,-0.0944197401,-0.6795374155,0.33890751,0.1927536279,-0.5905697346,-0.1646516472,-0.0376363732,0.476432085,0.0545883849,-0.0116788177,-0.0929591134,-0.4811432064,-0.062146984,-0.5296539068,-0.1239028499,0.2654697299,-0.0089569148,-0.1116511375,-0.3016604781,0.4860129356,0.3428494632,-0.3132482469,0.3125559688,-0.1947443187,-0.2029212862,0.257242769,0.1027379632,0.8113803267,0.213755548,0.0021945015,0.0761789232,-0.3513457477,0.4707512558,-0.7156766057,0.0449995846,-0.2629191875,-0.2532728314,-0.0418257006,0.0467178002,0.0449314713,0.2142265886,0.0668193847,0.5717445612,0.0646705329,-0.1611295938,-0.0035065378,0.2737398148,0.1399475187,-0.0535033196,-0.1496002674,0.1455361247,-0.1289558262,-0.0271256603,-0.2474437654,0.0909345299,-0.5433633327,0.0771605521,-0.1685913205,-0.1210749149,-0.1406954676,0.1165148914,0.130307287,-0.3935345709,-0.2577760518,0.6006622314,0.7157092094,0.1242058799,-0.3713186979,0.2642662823,-0.2064557225,0.1517821103,0.1099224836,-0.0541549698,0.0458781272,-0.0570328422,-0.0645753816,0.2766417861,0.3590575457,-0.0103579583,-0.2205699831,-0.0072343466,-0.1653816849,0.0703055114,-0.039750278,0.207115978,0.3018410802,0.1060423329,0.0886322185,0.2199644297,-0.5469431281,-0.029552212,0.2431462407,-0.2311170995,0.1657048911,0.3313014507,0.0089000305,-0.2933346331,-0.254380703,0.0747520626,0.082336165,-0.0051105819,-0.0354620405,-0.0204854272,-0.0411767401,0.020233145,0.2510381341,0.1330106854,0.021784259,0.4346135259,0.3064217567,-0.1111544296,0.2350937426,-0.4188061953,-0.1954860389,0.1380894035,0.0805040598,-0.1147240475,-0.0317436457,-0.1579879522,0.1743555069,0.5180917382,-0.2837796211,-0.1876215637,-0.0289492235,0.1559036374,-0.1345339864,-0.208983019,-0.2026926577,0.0788301155,0.0464248583,0.0644051656,-0.1808715165,-0.1644833982,-0.2384837866,0.1403975189,-0.1358500421,0.1390448213,-0.0938856378,0.0531134233,0.1827073693,-0.0889545381,-0.0719437525,0.3491812944,0.3699831963,0.4574721456,-0.1020185649,0.0841999054,-0.1154616699,-0.0345874876,0.2173670232,0.2292386144,0.0809069723,-0.024483759,0.1727303267,-0.0361928083,-0.247410804,-0.0541033819,-0.2052417845,-0.2769989967,0.2045040131,-0.0031600017,0.2293151617,-0.2594710886,0.4298005998,0.1612987518,-0.0934952796,0.1403725892,-0.1873000562,0.1359609663,-0.1194847375,0.3579465449,0.2065811902,0.2541401386,-0.2403785139,0.2233322561,-0.0160740949,0.0837634578,0.0647708178,-0.0318327062,-0.0286231954,-0.3264411092,0.1821546257,-0.4298461378,0.1596973836,-0.1026268899,0.2641341388,-0.1583989114,0.0626837015,0.2038579285,0.1104020923,0.2291985005,0.0251170304,0.4261502624,0.0412687548,0.247986421,-0.3098804653,0.2689863741,0.0796473175,-0.0312836803,0.1067346856,0.14089939,0.1134745851,-0.2333462834,-0.1448436379,0.0697262734,0.1230993494,-0.3971675038,-0.3614507318,-0.2509464324,-0.2693315744,-0.0179570839,-0.0517877042,0.1213875934,-0.1085511073,0.1001818255,-0.0448411554,-0.2422510982,-0.3757948577,0.0737765357,0.4461507201,-0.1080251113,0.1056251898,0.1896151304,0.3804330528,0.29856354,0.3300207555,0.1283008456,0.0885032192,-0.5871763825,0.1041395813,-0.1879853308,-0.0834155977,0.0767197311,0.321211189,0.178765431,-0.1607680172,0.3263960779,0.126121521,-0.0635787249,0.3487559855,-0.1164981425,0.0591412745,-0.3998602331,0.4350470006,0.1144597158,0.2115858942,-0.2502632141,0.0183307156,-0.5301825404,-0.1801619828,-0.5192389488,0.156250298,0.0573136955,-0.3319290876,0.0728541911,0.1262485534,0.0742610693,0.0557651892,0.3771968186,-0.3097304106,-0.2862945497,-0.412556529,-0.3241791725,0.0237126462,0.0798478574,-0.2076801062,0.2294334024,-0.0484357439,0.5633387566,0.1081698388,0.4685987532,0.0556875728,0.1747446954,-0.2406173497,0.0553699769,-0.0938051865,-0.1329702735,-0.1430653036,0.2436824888,-0.1766179502,-0.189197585,0.030148603,-0.1243876442,-0.2287807018,0.2140584141,0.7476040721,0.267261833,0.3172945678,0.4162442386,-0.0028226308,0.0363422856,-0.0075615114,0.0660380274,0.0075611626,0.2057744265,-0.3025841713,0.257532984,-0.2172202319,-0.2541283667,0.1553340405,0.6273363829,-0.0402461551,-0.2009104341,-0.3534558713,-0.1567508429,0.0596410558,-0.1647377759,-0.5186479688,0.390350312,-0.0414501354,0.4093332887,-0.0025360873,-0.0858867913,0.488684386,-0.3368971348,0.1139326319,0.1197451577,0.372789681,0.241696775,-0.1107578948,-0.2371815294,-0.0515486449,0.2927348614,0.1626878381,0.0006719815,0.2067310363,-0.1564822495,-0.1019813418,-0.1715914756,0.0690301284,0.1122841984,0.0605773777,0.0623745993,-0.1644407809]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/215","title":"NonMatchingSplitsSizesError when loading blog_authorship_corpus","comments":"I just ran it on colab and got this\r\n```\r\n[{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812,\r\ndataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train',\r\nnum_bytes=611607465, num_examples=533285, dataset_name='blog_authorship_corpus')},\r\n{'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277,\r\ndataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation',\r\nnum_bytes=35652716, num_examples=30804, dataset_name='blog_authorship_corpus')}]\r\n```\r\nwhich is different from the `dataset_infos.json` and also different from yours.\r\n\r\nIt looks like the script for generating examples is not consistent","body":"Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.","comment_length":53,"text":"NonMatchingSplitsSizesError when loading blog_authorship_corpus\nGetting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.\nI just ran it on colab and got this\r\n```\r\n[{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812,\r\ndataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train',\r\nnum_bytes=611607465, num_examples=533285, dataset_name='blog_authorship_corpus')},\r\n{'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277,\r\ndataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation',\r\nnum_bytes=35652716, num_examples=30804, dataset_name='blog_authorship_corpus')}]\r\n```\r\nwhich is different from the `dataset_infos.json` and also different from yours.\r\n\r\nIt looks like the script for generating examples is not consistent","embeddings":[-0.1318606585,0.1403096616,0.059409298,0.421038717,-0.1027606875,0.0639110357,-0.0993900523,0.2403311878,-0.1699515879,0.1674250662,-0.0025761358,0.2693625987,0.1059267074,-0.0292154495,-0.0612888411,0.116175577,-0.0757438838,0.2383864522,0.066430375,0.0098237591,-0.0296085756,0.4090332091,-0.2907909751,-0.0837625861,-0.2635079324,-0.2044773251,-0.0076957922,0.2664904296,-0.0280386489,-0.2925414443,0.2720763385,0.0110679027,0.1331819445,0.1919158548,-0.0001248043,0.0228433479,0.550614059,-0.2505318224,-0.2113020569,-0.2491973341,-0.4140854776,-0.2816134691,0.0394140258,-0.3059527576,-0.0753069445,0.1337391585,0.1635834575,-0.0529763326,0.2451409698,0.3204348981,0.1107066944,0.0796384364,-0.2686454356,0.1345798671,0.1266217679,0.0818086118,-0.0369906351,-0.0744712278,-0.1828916818,-0.1793120056,-0.0033597462,0.4617771208,-0.2776548862,0.1398821175,0.1918159872,0.1794870496,0.3482623994,-0.1386475712,0.1591957808,0.3170470893,0.4259499609,0.1577037722,-0.2048099488,-0.4242800474,-0.0219616164,0.1000380963,0.4799427688,0.3248815835,-0.2373103946,0.0331986733,-0.4330620468,-0.076351136,-0.0368182212,0.0093937339,0.1595861018,0.3218357861,0.0860608891,0.0775717348,0.0999623761,0.0130117293,0.3436033726,-0.2797834575,-0.2161586881,0.0420845933,-0.1020336077,-0.1977801323,-0.182736963,0.0624412261,0.156547457,0.0754775777,0.3709217012,-0.0862541124,0.1798445284,0.1177612022,0.3715391457,-0.0136769041,-0.0289674569,0.5226837397,0.1068688184,-0.009326756,0.1176358163,0.0450279489,0.2935001552,-0.1499844342,-0.1944580972,0.1255706251,0.1616524309,-0.6070366502,-0.4450771213,0.0165156908,-0.5715517998,-0.2078129947,-0.055248756,0.2461667955,-0.3623109162,0.2653523386,-0.1877928376,-0.0375909582,-0.3258183002,-0.0176446829,-0.1811062694,0.1677558273,-0.1367729008,0.257212013,0.1405435801,-0.1857023835,0.4198700786,-0.1785368174,-0.0307090133,-0.154903546,0.2606299818,-0.1932947189,0.0708691105,0.2253680974,0.0033081858,0.1662946194,-0.0920059085,-0.2952809036,-0.3580850065,0.1588877887,-0.2516792715,-0.5216350555,0.1034586579,0.1205809936,-0.4729642868,-0.0364872813,-0.0369016379,0.2329141051,0.4576260149,-0.3340524435,-0.0269594621,-0.1752264351,-0.2176076472,-0.0436786823,-0.2373847216,0.4392350614,0.1435437053,-0.2371553481,0.1398679465,0.0678040758,0.6044771671,0.3637804985,-0.002590589,0.0724760145,-0.1579887271,0.3344545662,0.1393306106,-0.3792095482,-0.220191434,0.4759375453,0.0770473778,0.2783876956,0.2617090642,0.1201213151,0.1928377748,-0.1457988918,0.1319236755,0.3967820108,-0.0160179604,0.0646982715,-0.4757282138,-0.3340060711,0.4510889947,0.1048206836,0.1060980335,0.0166583806,-0.0662410483,0.4533866048,0.2774912417,-0.1179685816,-0.0385755822,0.178915143,-0.2975054085,-0.003019769,-0.0089946594,0.0459393449,-0.4310577512,-0.0618973635,-0.1615540832,0.2573902011,0.0215470344,-0.3396947682,-0.1260259151,-0.3853749633,-0.0126097491,-0.0381515138,-0.0249999687,0.3279168308,0.389162451,0.0201133378,0.0473103113,0.516518712,-0.0756102204,0.1980000734,-0.7434920073,0.2497137785,-0.0970003754,-0.2209163755,0.4876715243,0.294072032,0.1883365959,-0.2082684189,-0.128199935,0.4203102589,-0.0991321951,0.4102058411,-0.3028308451,0.0094528515,0.2866004705,-0.2159036547,-0.0950905457,0.3127256334,-0.0143619888,-0.0296804067,-0.0852994472,0.4560612738,-0.355229497,0.1949430257,0.0601800233,-0.0744410902,0.1126468629,-0.2118273079,-0.2256216407,-0.4647096395,0.5550792217,-0.0112827644,-0.3389958441,0.0956973657,-0.2214524448,0.0983054042,0.4710589647,0.09217491,-0.1332444549,-0.1031344831,0.1075798199,-0.1303173751,0.2442013621,0.2132988572,0.4979693294,0.1266188174,-0.0254707485,0.0588753745,-0.1459996402,-0.3809733391,0.1322558671,-0.0173103679,0.143349424,0.2880009711,0.0936621428,-0.0403743759,-0.2727138102,-0.0497351997,0.1679785103,0.2461205721,-0.2921513617,-0.0409143791,-0.3865088224,-0.4747574925,-0.2583890259,-0.1584340036,-0.216693297,-0.3360840678,0.1613166928,0.2477845401,-0.132611528,0.0692499056,-0.1683563441,0.0661015287,-0.3122084141,-0.0732158124,-0.0626509041,-0.2862415612,-0.2431289405,-0.0438643284,0.1981613338,0.3729360998,0.2036013603,-0.42539379,-0.2250232846,0.1412787586,-0.2914526463,0.1450170428,-0.0265262555,-0.1244560331,0.0801766738,-0.217384696,0.2646995485,-0.2417492121,0.0449195132,-0.1044062376,-0.3310623467,-0.1003105938,0.1255846471,0.0045288811,-0.2103860229,-0.4195670485,-0.2590777576,-0.3447067142,-0.1063203663,0.2822982073,0.3047814667,0.1186570451,-0.1598611176,-0.1130956933,-0.0577273406,0.2196367383,-0.340500772,-0.0982074142,0.2392107099,-0.0872906819,-0.1138919741,0.0112999985,-0.0569706783,-0.0334286913,0.152232483,-0.5655209422,-0.2071268409,-0.1663601846,-0.1016507372,-0.2521025538,-0.2789226174,0.5203965306,-0.0568067804,0.0354509577,0.065879561,-0.4510561228,0.0322562382,0.167688325,0.5530168414,-0.1850311309,0.2017791122,-0.0593964048,0.4893557727,0.5270174146,0.0005125082,-0.1543316394,0.152440533,0.1464681327,0.0762568712,-0.0376843177,0.3673476577,0.054857593,-0.2222499698,0.448146522,-0.0057776375,-0.2257878333,-0.1870693117,-0.1095896214,-0.2654105723,-0.2833300829,0.1376718581,-0.1600021571,0.1310964078,0.2893314064,0.3210534453,-0.071069546,-0.5795770884,-0.0772110522,0.3904193938,0.1502411813,0.1212038994,-0.3957865834,0.0677135289,-0.5080139637,0.3432270885,0.2445323765,0.2829881012,-0.1522456557,-0.0715009496,0.1007093713,-0.2628659308,0.2946953177,-0.4508108795,-0.1030839235,-0.0566695593,-0.0329139382,-0.3269299865,-0.1900273561,-0.1659848839,0.4601231813,0.4731752872,0.4658534825,-0.3707212508,-0.0875442103,0.203804031,-0.1290634125,-0.1338875294,-0.1733285636,-0.3280024827,-0.4676932096,-0.1635874063,-0.1463002563,-0.1255976707,0.3344830275,-0.0138926618,-0.0477862172,0.0837841704,0.0919467807,0.3580942154,0.3425988555,0.0784649849,0.1107559651,0.1065607741,0.2138594091,0.5311777592,0.1292665899,0.5807299018,0.1455204934,-0.066334717,-0.0193501394,-0.2441666722,0.1489513516,0.4908935428,0.1351774633,0.0538152047,0.0374839082,-0.0820770785,-0.2781617343,0.0752347112,-0.0076918621,-0.0544719733,-0.4696714282,-0.3332740963,0.2758779824,0.2949086726,-0.1409775317,-0.1551347971,-0.2400292903,-0.3468586504,0.283521086,0.243718639,1.0874798298,0.241748482,-0.0221058242,-0.0018392171,-0.1520771682,0.9227554798,-0.2762621343,0.4379425347,-0.3907802105,-0.2073120028,-0.0421940722,-0.1648712456,0.1221634522,0.02207339,-0.3829013109,0.3094181716,0.3011404276,0.0201676413,-0.1154497936,0.3871205747,-0.2153320312,-0.1462295055,-0.2855425477,-0.0286062751,-0.3070122302,0.4314880073,-0.0250857286,-0.0584685504,-0.2192477733,-0.2192170769,-0.3474551737,0.2333544046,0.0373244211,-0.0337887704,0.2643848062,-0.2253137678,-0.0542298295,0.0519678816,0.3515725732,-0.0851534829,-0.1073273793,-0.0710336044,-0.1847058535,-0.1427914202,0.1284922361,0.01437137,0.23452501,-0.0847068205,0.0624510422,0.0678861663,-0.0975303799,-0.3518661559,-0.1145561561,-0.0611329749,-0.2019591331,-0.1538673788,-0.04333416,0.1310224533,-0.0852941126,-0.0387621932,0.0029934372,-0.1441456527,-0.221053049,0.2700074911,-0.0691212714,-0.3818732798,-0.1572827101,0.3526189029,0.4428406358,0.0365885049,0.3667507768,-0.1528479755,-0.0776589438,-0.1503261179,-0.1503044814,-0.3644887507,-0.2034792006,0.2349234074,-0.2366288602,-0.125941202,0.1424077153,0.4553365111,0.1807442904,0.3331869841,-0.0611840524,-0.510886848,-0.0081897276,0.1752655059,0.032078959,0.058280278,-0.0580653921,0.5259190202,-0.2515851259,0.2103163153,-0.1909003407,-0.1488889903,-0.4438942969,0.3342981935,0.1320314705,-0.0437716916,-0.0095720487,0.0000539759,-0.1639130116,0.3139312863,-0.0652245805,-0.0059553701,-0.0614597797,0.206500113,0.1785381734,-0.1087701246,-0.1746367067,0.124801144,0.0767789185,0.0375171788,0.0346367061,0.2473377883,0.0440506637,0.2818430066,-0.3356451392,0.1102281362,-0.078303881,0.2276510894,-0.2985176742,0.3604111671,-0.0591354631,0.3064757288,-0.1001032144,0.0155948969,0.1505693793,-0.0360641666,0.0484821722,0.0973271355,0.0767751485,-0.1798094213,0.1473214477,0.1841632426,-0.0891641378,0.6363039613,-0.0754889026,-0.1910720915,0.1043129191,0.045415666,-0.0859746635,0.0015474129,0.2222012132,0.0250473954,-0.0561294816,0.5138854384,0.0897829309,-0.1323118657,-0.1374367028,0.1116622537,0.3375567198,-0.0757068396,0.1039449722,0.1966679394,-0.2053923309,0.0955599099,0.304613024,-0.2812727988,0.1519570649,-0.0000517707,0.0272426233,0.4766048789,0.1611953974,0.3706960082,-0.0243174434,-0.1133221835,0.0299595594,0.6413370967,0.2822777629,0.3716787398,0.2842691839,0.3873715699,-0.2766483426,0.0287864227,-0.3678883314,-0.1754918545,-0.0638569221,-0.140515551,-0.4606207609,-0.0484437309,-0.1724057496,0.2062790543,-0.140235275,0.126040563,0.4599737227,-0.0754481331,-0.2075772285,-0.3750886917,0.1200777292,0.0288925301,0.1669862568,-0.2368590385,0.3165039718,-0.0766635016,-0.086644806,0.3203835189,0.5079476833,0.4745163918,0.3062892556,-0.057672359,-0.2090128958,0.081525743,0.1103424132,0.2296565324,0.763684094,0.2975424528,-0.2571271062,0.3881793022,-0.0647664815,-0.0767417401,0.4682405889,0.0705722198,-0.0875261649,-0.0221661851,0.4095027447,-0.0731790066,0.1157869846,0.0215554666,0.1373033971,-0.2554471791,-0.2709901929,0.068212159,-0.2707930505,0.3050962389,0.0131677156,0.0222035144,-0.3044300377,0.6118450165,0.3897584081,-0.1275226772,-0.2048282772,-0.0723047927,-0.4306271374,0.3466023207,-0.4226308167,-0.2399234027,-0.299687326,0.5030860305,-0.2274711132,0.4045400321,-0.28772524,0.3795650005,-0.0651755854,-0.0316751413,-0.2298383266,-0.0672011673,-0.1227159426,-0.296638459,0.0200491715,-0.2617562115,0.2225907296,0.0769959018,-0.1308633834,0.0015143164,-0.0164357312,0.0160175711,0.4078022838,-0.0129080061,0.5090906024,0.3904209733,-0.1421260238,-0.1503125727,-0.0055250116,-0.319763422,-0.3111401796,0.4149208069,0.2139554918,0.3597548902,-0.1741255969,0.6665084362,-0.0674729422,0.3192361593,0.0568280891,0.0226313211,-0.2512759566,-0.0570168309,-0.2459766865,0.0358288772,0.0675442815,0.4534134865,0.1536414325,-0.0277580507,-0.242596522,-0.3491298556,0.4797303975,-0.3962062597,0.029296577,-0.0066790264,0.1837035716,0.2450031042,-0.2263406962,-0.7466571331,-0.0700578764,0.2567740679,-0.2124135941,0.0507245101,-0.0207965504,-0.0501661077,-0.2243605256,-0.1288058758,-0.233914569,0.2149940878,-0.4172472358,0.2863393128,-0.129387632]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/215","title":"NonMatchingSplitsSizesError when loading blog_authorship_corpus","comments":"The files provided by the authors are corrupted and the script seems to ignore the xml files that can't be decoded (it does `try:... except UnicodeDecodeError`). Maybe depending of the environment some files can be opened and some others don't but not sure why","body":"Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.","comment_length":44,"text":"NonMatchingSplitsSizesError when loading blog_authorship_corpus\nGetting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.\nThe files provided by the authors are corrupted and the script seems to ignore the xml files that can't be decoded (it does `try:... except UnicodeDecodeError`). Maybe depending of the environment some files can be opened and some others don't but not sure why","embeddings":[-0.2156543285,0.1234759688,0.0276875217,0.4813607037,-0.0115497429,0.0988334715,-0.1319700629,0.3685873151,-0.0516962074,0.185156092,0.0752992705,0.1368834227,0.085427843,-0.1344738603,-0.0322380997,0.0252993703,-0.0925179124,0.1929506361,0.038546674,-0.0925541669,-0.1750836521,0.4164965749,-0.3340626359,0.0076592555,-0.2115804255,-0.1027778238,0.0826694518,0.4248708189,0.0054142582,-0.3878874481,0.0170962587,-0.0154913152,0.0771716461,0.203576237,-0.0001237728,0.0525185615,0.6025198102,-0.2609122694,-0.151607424,-0.2473248988,-0.3983466923,-0.3468357623,-0.1532549858,-0.3098120689,0.0434061065,-0.137420699,0.2702697515,-0.2101037949,0.3015905917,0.3589797914,0.1067679897,0.1728032082,-0.0673225299,0.1383743435,0.2782538235,0.0605984442,-0.0570703223,-0.0743035004,-0.0305489581,-0.2105677128,-0.0998369455,0.3154497743,-0.2957826555,-0.0047144433,0.1741676927,0.0982157588,0.3729239106,-0.2041340321,0.1727690995,0.2954415381,0.5296692848,0.1300405264,-0.2119538039,-0.447938025,-0.1549126208,0.1176742166,0.4065707028,0.4266488552,-0.2051057816,-0.0010258397,-0.2633985877,-0.0346608423,-0.192991361,0.0662728176,0.2431640178,0.2416525632,0.0814639777,0.0660806373,-0.0201202147,0.0718880221,0.0966890082,-0.2602756023,-0.1418869048,0.0954284593,0.0298395157,-0.0755187124,-0.1188722774,-0.0754125789,0.206089437,0.2048174739,0.3663540781,0.0198506024,0.2098523378,0.1350396276,0.3377993703,-0.0728869513,0.0488545261,0.4268063307,0.2000229657,-0.0353823379,0.0891042277,0.0133006312,0.2252698541,-0.2736573219,-0.1577363461,0.1777952909,0.2460042089,-0.4968622029,-0.4554835856,0.147542581,-0.4238386452,-0.2733584344,-0.2503490448,0.2649307549,-0.3046384156,0.3659855723,-0.1307987422,0.0885317549,-0.3846735358,-0.1032655388,-0.1302550733,0.1645716131,-0.1479643881,0.2090974301,0.1741017699,-0.1450188458,0.3679032922,-0.1771585941,-0.0643847659,-0.2137999386,0.3670299053,-0.1796639264,0.0469619483,0.2502399683,0.1062244624,0.1718141735,-0.1179553717,-0.2918157279,-0.285081625,0.0958579406,-0.2795931101,-0.3930387497,0.1186463088,0.1444409639,-0.3303497136,-0.0422152318,0.0714515299,0.049522806,0.4147756994,-0.4803286195,-0.097071521,-0.2282780558,-0.1588879377,-0.1285473853,-0.2621929049,0.6329419613,0.1380793452,-0.1023058966,0.2826188207,-0.0291142669,0.5553824306,0.3593012691,0.0627660677,0.1281593889,-0.1939798892,0.4745135307,0.1176655293,-0.4100925326,-0.1508032233,0.469144702,-0.0413343571,0.3392701745,0.4302490652,0.1024661288,0.284476757,-0.179268077,-0.0424578525,0.3146088123,0.0632298961,-0.0284431968,-0.4521337748,-0.1466735005,0.4367762208,0.104891777,0.1847425103,0.0101594049,0.1415487379,0.438993305,0.3131590784,-0.1066095158,0.1054219157,0.2982517779,-0.4476807415,0.1695096791,0.1006326973,-0.1074304953,-0.3479078412,-0.0729608983,-0.3559926748,0.1438062489,0.0009276181,-0.2238157243,-0.173315078,-0.3992128968,-0.0771565512,-0.0034569832,-0.0222622212,0.3663996458,0.2873929143,0.1545172483,-0.0590559244,0.3667797148,-0.0336964652,0.1865332425,-0.8968728185,0.2386494875,-0.1518678814,-0.2217142284,0.4185646772,0.31582883,0.1912763417,-0.1933243871,-0.0772928596,0.3810427785,-0.105867587,0.4759338498,-0.2341870815,0.0746665969,0.1891220957,-0.3120641112,-0.0202617515,0.4056883752,0.0988725349,-0.1194396764,-0.0373294465,0.4227373004,-0.3420764506,0.1942277402,0.141124025,-0.0657071322,0.0637537688,-0.2677146792,-0.2454346418,-0.5215464234,0.6221488714,-0.0673057288,-0.2690837681,0.0467984378,-0.2562191486,0.0259330738,0.572961688,0.0020749425,-0.083009474,-0.0628990307,0.0671948269,-0.0630131811,-0.0181172378,0.1845991313,0.4571344852,0.0909475982,-0.0555765145,0.067112416,-0.131819576,-0.4584284425,0.1281425953,-0.0483283475,-0.0279729385,0.572494626,-0.0118877217,0.066377379,-0.3477852941,-0.0212487858,0.114563331,0.1385264695,-0.3325794935,-0.092933841,-0.3299708664,-0.4057205617,-0.3795689642,-0.0761694536,-0.1871168464,-0.3854968548,0.1835044175,0.2175429314,-0.2500676513,0.028112866,-0.1977861226,0.0052076131,-0.3419668972,0.0488546453,-0.1312797666,-0.1937546283,-0.2105798572,-0.0651410595,0.2704782188,0.3898126483,0.220295608,-0.5454667211,-0.2100887597,-0.0447695553,-0.2075350285,0.020225374,-0.0599084683,-0.1776082814,0.0929601938,-0.1559806168,0.5214742422,-0.1610414386,0.1477635652,-0.1287196428,-0.2708491683,0.1249035075,0.2216656953,-0.1444784105,-0.2206292152,-0.5865989327,-0.1790791005,-0.4568950534,-0.0989776105,0.2818943262,0.2948868573,0.0423539914,-0.1621510834,-0.1063438803,-0.2143936902,0.3561641574,-0.1631315798,0.0558564812,0.3234172761,-0.183947891,-0.2702252567,-0.0217075609,0.1376061141,-0.0514777265,0.1415862292,-0.4621354938,-0.059683416,-0.183438465,-0.1534477621,-0.2829677463,-0.1117320359,0.469930321,-0.0341297574,-0.004535214,-0.0566163063,-0.2683965266,0.0170177985,0.0547959618,0.3007240295,-0.1281157136,0.2338898927,-0.0995773971,0.4160902202,0.5284249783,0.1846239716,-0.067755647,0.0757426098,0.2727709115,0.1663347185,-0.0166349281,0.3360998333,-0.0237723496,-0.0910194665,0.4331526458,-0.00234625,-0.0968587697,-0.1455361247,-0.2459460199,-0.3490150273,-0.1934382021,0.1462911814,-0.1666951925,0.1487079561,0.2208891958,0.3066445291,-0.0387202874,-0.515557766,0.0245413277,0.4532245398,0.3203481734,0.1850076169,-0.479675442,0.0912569016,-0.4247424304,0.2979340255,0.1449542195,0.3121423125,-0.1828319728,-0.149903059,0.1062321514,-0.2515309453,0.2325569689,-0.4161820114,0.0198437963,0.0590470694,-0.006551607,-0.3043077588,-0.1795511693,-0.2614229023,0.4053450823,0.3865743279,0.4312340021,-0.3831051588,-0.1649511158,0.0973510891,0.0051356284,-0.108641237,-0.0825343877,-0.2994445562,-0.3675915599,-0.229607299,-0.2945693433,-0.0884970799,0.2692146003,-0.0923358873,-0.027085457,0.0346864425,0.1000918746,0.3609344065,0.2708037794,0.0865865499,-0.1059833616,0.0742232502,0.1566508561,0.5734146237,0.1181423217,0.5895779133,0.068994686,-0.1811110377,0.0247475449,-0.4304126501,0.1916280538,0.4658268094,0.0818428248,0.0443003178,0.1261618137,-0.0242002718,-0.4446395934,-0.0597196184,0.1669707,-0.081094496,-0.4040863216,-0.4204249382,0.2781200409,0.2054547518,-0.1348199695,-0.1648530066,-0.1309440881,-0.2597299516,0.3258020878,0.1342848986,1.0393663645,0.2653284967,0.0710747913,-0.0667507723,-0.1936643869,0.7017489672,-0.2334567755,0.4610559642,-0.4527295828,-0.1818931252,-0.0957932174,-0.1857745349,0.2801088989,0.1366838962,-0.217437312,0.338727802,0.2043427378,-0.0272056758,-0.1010571122,0.4760469794,-0.2647897601,-0.0346379131,-0.1028649658,-0.052990932,-0.1834339797,0.4568297565,-0.0012247202,-0.13717556,-0.1618823111,-0.2715009749,-0.3544199765,0.2869670987,0.1157818809,-0.0782925561,0.1835903227,-0.1999579519,0.0545952432,0.2060694098,0.2277206481,-0.0842021555,-0.0386506617,-0.09670569,-0.1803751439,-0.1994976848,0.187104091,0.1252504736,0.3472335935,-0.1142901406,0.0623054691,0.1867661476,-0.1503843367,-0.278490603,-0.3265249729,0.0203608051,-0.2225668579,-0.1865185201,-0.1451813579,0.015972428,-0.166983977,-0.0472408421,-0.0031717452,-0.0846774802,-0.09634009,0.1906725466,-0.1776465625,-0.4502427876,-0.125515148,0.2783940136,0.364097029,-0.0301699731,0.4310379922,-0.0935875922,-0.1244844347,-0.1957195401,-0.1521315277,-0.2345648408,-0.4015834332,0.1722549051,-0.2061403841,-0.1045775935,0.1757640392,0.243963331,0.0878320783,0.1960273087,-0.1278719604,-0.6653317213,0.1264739931,0.1522253305,0.2188434005,-0.0190530363,-0.0489017256,0.3229842484,-0.2310793847,0.1628666669,-0.1806020588,-0.0751138851,-0.3088244796,0.3002741039,0.166003719,0.0407697298,-0.109468244,0.0291907862,-0.135567084,0.2487290949,-0.1774606407,0.0392681845,-0.1411899179,0.2227547914,0.1571930647,-0.059653867,-0.1089705229,0.0197486058,0.0933022872,-0.0217984207,0.0339215286,0.2995970249,-0.0386265069,0.3489004076,-0.3989144266,0.1466967762,-0.0903753415,0.2649861872,-0.346200794,0.3897478878,-0.2014950812,0.2406035066,-0.0393453427,0.0229206737,-0.0987975746,-0.031777218,0.1211959198,0.0641354322,0.3152731955,-0.158739537,0.21573174,0.2088231593,-0.0373391919,0.73200351,-0.1256159097,-0.1162344962,0.0620181747,0.0547891445,-0.1106708422,-0.0179837774,0.1833472103,-0.0888067335,-0.0823713019,0.3635841012,0.1948918849,-0.2530850768,-0.188580513,0.1076555923,0.4120694995,-0.0813133344,0.0942232013,0.2386840135,-0.0116018308,0.0179333892,0.1976092905,-0.1533744931,0.0799498409,0.1149349362,0.0946036726,0.5233209133,0.1414508373,0.435102731,-0.2317974269,-0.1278586537,0.0184497759,0.6700661778,0.1927615851,0.4858525991,0.2512132227,0.3217865825,-0.2963056564,0.1505878717,-0.2669786811,-0.1741832495,-0.0377619229,-0.1280358136,-0.4697092772,-0.1269298047,-0.200431481,0.0633760393,-0.0709589645,0.2730638683,0.3169324994,-0.0559199713,-0.1986408234,-0.2947253883,0.1727563292,-0.1547849923,0.1828127354,-0.3227284253,0.2761786878,0.1026152968,-0.0328720473,0.3759278655,0.5440003872,0.6268032193,0.4050430059,-0.1098948196,-0.1285424531,-0.054338228,0.0478570051,0.1882706881,0.8659353852,0.2104937434,-0.1732787341,0.3700048625,-0.0392648429,-0.0889640674,0.3375838101,0.1680566519,-0.0406325124,0.0470468998,0.388229996,-0.1983951777,-0.0021036332,-0.0224142876,0.0341693275,-0.4142059386,-0.2665125728,0.1787832081,-0.1447107494,0.2802109718,-0.1523040384,0.0190835502,-0.1773107201,0.5529254675,0.477463752,-0.1700437367,-0.0966543406,-0.0223147515,-0.5047854781,0.507260263,-0.4191425741,-0.3187648356,-0.2432263196,0.3742886484,-0.1979193985,0.2628780901,-0.0930087119,0.4435425401,-0.1374349296,0.1004913896,-0.21290721,-0.0198611934,0.0142651265,-0.1569485217,-0.098482728,-0.2454882711,0.2524291873,0.0409253724,-0.072093904,-0.0109718544,0.1572081298,-0.0024542708,0.2743947804,-0.0261716936,0.4778504074,0.4452598393,-0.2909836769,-0.2139381766,-0.1103196219,-0.2303617001,-0.2815766633,0.4011003971,0.2258203328,0.3603354394,-0.2049239278,0.5031100512,-0.0925665647,0.2626859546,-0.1571084112,0.145056501,-0.2188481539,0.0314995088,-0.2332499921,-0.0163316876,0.1572694182,0.4576335549,0.1887994409,-0.1253947169,-0.2311981171,-0.3733319938,0.4127678573,-0.3910878599,-0.0798832476,-0.0172338504,0.1939743906,0.1286227554,-0.0415325314,-0.6443729401,-0.0595880263,0.2283192724,-0.1138498485,0.0298231579,-0.0664223209,-0.1520508975,-0.352040261,-0.0343551449,-0.2507992089,0.2707358301,-0.4192504287,0.4005514681,-0.103509143]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/215","title":"NonMatchingSplitsSizesError when loading blog_authorship_corpus","comments":"Feel free to do `ignore_verifications=True` for now... The verifications only include a check on the checksums of the downloaded files, and a check on the number of examples in each splits.","body":"Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.","comment_length":31,"text":"NonMatchingSplitsSizesError when loading blog_authorship_corpus\nGetting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.\nFeel free to do `ignore_verifications=True` for now... The verifications only include a check on the checksums of the downloaded files, and a check on the number of examples in each splits.","embeddings":[-0.1591764539,0.0613832586,0.0613702349,0.3985078633,-0.09111882,0.0681395456,-0.1541567892,0.3743870556,-0.0539890267,0.1291232109,0.0949106291,0.1696771979,0.050890822,-0.0702365413,-0.0299417637,-0.0313884914,-0.0462678969,0.2041032165,0.0438870601,0.0010297163,-0.0196369626,0.4404026568,-0.3017500937,-0.0229673106,-0.2469935268,-0.1683312207,-0.0421307385,0.3281975985,0.0039551742,-0.3751760125,0.2439582944,-0.0751470774,0.1352920681,0.1908108741,-0.000121871,-0.0134052187,0.5785955787,-0.2156344354,-0.2062556744,-0.25492239,-0.4047058821,-0.3622810245,-0.083601512,-0.2355190068,-0.021686472,-0.0550388619,0.1720563173,-0.0859804302,0.2725291848,0.2854221761,0.1191038489,0.1523060501,-0.1194791868,0.0312914997,0.1314683855,0.1114799976,-0.0382259116,-0.0960215852,-0.0876140743,-0.2862399817,-0.1037817448,0.4399676025,-0.1971516311,0.1551972777,0.2258738726,0.11542999,0.400872916,-0.1145614833,0.0375812799,0.3904350996,0.5094953179,0.1603325605,-0.2086892426,-0.506644249,-0.0504567102,0.193820864,0.4156517982,0.364884764,-0.2391003668,0.0084717777,-0.4111099839,-0.0739310309,-0.0905095637,0.101740256,0.2668362558,0.1985677481,0.1183715761,0.0067396951,0.1296718568,-0.0410987772,0.1146330684,-0.244287923,-0.1399923265,0.1128277257,-0.1838782728,-0.193074137,-0.1448516548,-0.0174066816,0.2120881826,0.2155574709,0.3518286943,-0.0540539734,0.2256540507,0.1119684651,0.428317368,0.0479068607,-0.0107361032,0.5196453333,0.0832171366,0.0033718608,0.1151564568,0.0596730672,0.2659813464,-0.219095245,-0.1451855898,0.1199168414,0.1594631672,-0.5667434335,-0.4398617744,0.108157292,-0.5615888834,-0.2434341908,-0.0924841166,0.1756622344,-0.3614130318,0.3320044577,-0.1538780034,0.0477984622,-0.3616379201,-0.0594461039,-0.2066724449,0.170678705,-0.1974993199,0.205324918,0.2014609128,-0.1554444134,0.4026421309,-0.1078296602,-0.0586714819,-0.1968979836,0.401135087,-0.1081409603,0.0200106185,0.1572443396,-0.0291808303,0.1699854136,-0.0008439791,-0.2169329971,-0.3886170685,0.0795449987,-0.223024413,-0.5368488431,0.1046898589,0.1530231833,-0.4432890415,-0.0612912364,0.0962721482,0.2109623849,0.3643842638,-0.3324509263,-0.0974242464,-0.2202017158,-0.1294719428,-0.0545685366,-0.2055283487,0.4950897992,0.1402006447,-0.2660527229,0.130453065,-0.0182818323,0.5335456133,0.349309206,0.0343344957,0.0964401215,-0.1925892532,0.4875856042,0.1357104778,-0.3542971611,-0.1630119681,0.5454872251,0.0286830738,0.2807053328,0.2512045503,0.1191703603,0.1942970008,-0.0937625468,0.0107714497,0.4080866873,-0.0216021389,0.0351920836,-0.4514915943,-0.3096883297,0.4148988128,0.2066952139,0.1179958284,-0.0422624536,-0.0019413965,0.4964993894,0.3046291471,-0.1359486133,0.053966511,0.2919709682,-0.3540271819,0.1288687885,0.038535621,-0.1150446013,-0.3697339296,-0.0230219252,-0.2869818509,0.2818144858,0.1069893241,-0.2584026158,-0.1996174753,-0.4155744016,-0.0252256133,-0.0299624354,0.010713621,0.281521976,0.397457391,0.0511937663,0.0685104057,0.4284547865,-0.062188752,0.1739074141,-0.796795547,0.2320177257,-0.1187641472,-0.2111772895,0.5127651095,0.3413342535,0.2236994207,-0.1693107188,-0.0957317874,0.4365608394,-0.1037369147,0.4019582868,-0.1740678698,0.010552736,0.207886681,-0.295458734,0.0182179827,0.3747373223,-0.0284037571,-0.0686894059,-0.1335939765,0.4167664349,-0.3382882476,0.1654530913,0.0687626153,0.0150097655,0.0781649724,-0.238795951,-0.1913959235,-0.4554651678,0.6415628791,-0.095137015,-0.3222642243,0.0422930382,-0.2010635138,0.0330332108,0.4396526814,0.0218652375,-0.0908261016,-0.0264099967,0.0283896141,-0.160866186,0.0957219303,0.2750542164,0.5138388276,0.1237687096,0.0233142655,0.0150964083,-0.1677722335,-0.3314580917,0.1229685172,-0.0629390255,0.027277194,0.3670684099,-0.0033677937,-0.0319100954,-0.3108036518,-0.0123697128,0.1595615149,0.2488149554,-0.3282091022,-0.1341907233,-0.3583884537,-0.404402107,-0.3453231752,-0.1350653917,-0.2826974392,-0.3694410026,0.2020636797,0.1999648362,-0.1682595015,0.1388691515,-0.1668097675,0.0433350764,-0.3156359196,-0.0528267287,-0.0185160059,-0.2104287446,-0.2734283507,-0.0262423493,0.2240412682,0.4267170131,0.2425216883,-0.392234385,-0.1577690244,0.0467405058,-0.2580870688,0.0625039861,-0.0561855622,-0.151365906,0.176835835,-0.1386253834,0.3756527603,-0.2218138576,0.1016403735,-0.1899563521,-0.3374041617,0.0273164678,0.1791119277,0.035421472,-0.1923547089,-0.5341876745,-0.2685709596,-0.3635152578,-0.0623655543,0.2855772376,0.2418258786,0.0317275077,-0.1204401776,-0.1327033788,-0.1704918742,0.2503778636,-0.301707834,-0.0588897243,0.2404206246,-0.0978669599,-0.1675153971,-0.0314710289,0.0028389059,-0.0506655537,0.1228749007,-0.5468320251,-0.1095873564,-0.205603078,-0.0794581175,-0.3269419968,-0.2575340569,0.5263426304,-0.0734755918,-0.0241890699,0.0135372495,-0.3449537754,0.0683690906,0.1019062549,0.5138315558,-0.1611067802,0.1840385497,-0.0143205607,0.5651077032,0.5774591565,0.0766566023,-0.1154994592,0.1767917275,0.0924454182,0.1740482748,0.0098281307,0.3770256639,0.0225284211,-0.1074554622,0.4157789946,0.0836051777,-0.1998007447,-0.1143890396,-0.1578506082,-0.3104238212,-0.1871028841,0.1615623385,-0.229975,0.2006326467,0.2018061429,0.2388329357,-0.1851038039,-0.5294703841,-0.0126544321,0.3591071665,0.2272570133,0.1768595129,-0.4701031744,0.1427349895,-0.4302374721,0.2958562374,0.2112143636,0.2849109173,-0.2108566612,-0.1554042697,0.0810376108,-0.1988070458,0.2955514193,-0.405642122,-0.0150202429,-0.046210967,-0.0318967067,-0.3422423601,-0.2017220557,-0.1381534636,0.3731075525,0.5086650252,0.4213859737,-0.427229017,-0.1245947629,0.1632818878,-0.0594048649,-0.1309099048,-0.1694367826,-0.2525319159,-0.3333903849,-0.1716302186,-0.1472924501,-0.1254293621,0.300116837,-0.0219560023,-0.0020834857,0.1447328031,0.125219807,0.335575968,0.2571821511,0.1685252935,-0.0330616906,0.1076688021,0.2788697779,0.5404441357,0.0835823566,0.5131454468,0.1059583724,-0.0589108244,-0.0250916965,-0.3047665954,0.0786135569,0.4301792383,0.1747909039,0.0852493346,0.1424742043,-0.0274347439,-0.4631120265,-0.0257440377,0.1107125357,-0.0765662864,-0.3716811538,-0.3909309804,0.2588397264,0.2693785131,-0.1429104656,-0.1460681856,-0.1955788732,-0.2553567886,0.2318805903,0.1594075561,0.995031774,0.2600901425,-0.0260207765,-0.0276092105,-0.2229971439,0.8679141402,-0.2555350959,0.4108975232,-0.4282152653,-0.2512688041,-0.0623484477,-0.2089149207,0.1523510218,0.1181777939,-0.2996963561,0.3315711617,0.2687560916,-0.0116665736,-0.1033218801,0.4452231228,-0.2453591824,-0.0730988085,-0.1773478091,0.0148012135,-0.2884218693,0.5008416772,-0.0177221671,-0.1215852797,-0.2184642553,-0.2183345705,-0.4154768288,0.2596201003,0.1806758642,0.0051092128,0.278000772,-0.1839193851,-0.000858806,0.0915722251,0.328358531,-0.1827333272,-0.1748190671,0.0069741206,-0.2024377584,-0.1396858692,0.1929191649,0.0744873956,0.3397163749,-0.1254932433,0.00139124,0.0564299785,-0.1222895607,-0.3300777376,-0.3036001027,-0.1133046001,-0.202158913,-0.1876929104,-0.0881682262,0.0782769471,-0.103473261,-0.0614964329,0.0404773057,-0.0402333401,-0.2118255794,0.2360780239,-0.0904948041,-0.3304221928,-0.1168062836,0.3977996111,0.3670253754,-0.0341135226,0.4429043829,-0.1041081473,-0.0931104198,-0.2093406618,-0.1291898638,-0.3109356463,-0.3431043625,0.1643280983,-0.2684628665,-0.0866700932,0.1947073936,0.2930184305,0.2384102643,0.34192729,-0.001604328,-0.5291135311,0.0160753056,0.1371033043,-0.0249782745,0.075279586,-0.1436313838,0.4218314886,-0.2061970979,0.240417853,-0.2192303538,-0.0825468749,-0.4039361477,0.3224556148,0.1132717207,-0.1017626077,-0.038898211,0.0158689003,-0.1192549169,0.2172656208,-0.1423014104,0.0013304161,-0.1154846996,0.2013712078,0.1473946571,-0.0819742829,-0.1027795896,0.1140071079,0.071429424,0.0290166158,-0.0539864264,0.2069938332,-0.0221136399,0.3232142031,-0.3484607339,0.0423226729,-0.0693849549,0.2365977317,-0.349024117,0.2936600447,-0.0947008505,0.219861865,-0.0839357078,-0.0001582067,-0.0336337648,0.0133300079,0.0837266967,0.0724359974,0.1884958297,-0.1730406582,0.1039829105,0.2213937044,0.0217144247,0.5436677337,-0.0795347095,-0.1097536385,0.0791536868,0.1007067263,-0.1290856153,-0.031610325,0.2547107637,-0.0304624699,-0.1291433424,0.3505035043,0.1402949244,-0.1605657935,-0.2005096525,0.1079647243,0.3961581886,-0.1329913288,0.1237978041,0.2921994627,-0.1617975086,-0.0157105867,0.2393840253,-0.2638280988,0.1348326057,0.0109876497,0.044185821,0.4613366723,0.0975363404,0.3920544982,-0.1445141584,-0.1371357292,-0.0538605228,0.7125883102,0.3032739758,0.407548368,0.230524078,0.4754119515,-0.3398908079,-0.0037842232,-0.3541959524,-0.184643358,-0.0373532176,-0.101901792,-0.5155585408,-0.1083647609,-0.2322479934,0.122185111,-0.0881861374,0.143259868,0.4093429744,-0.0715943202,-0.1891854852,-0.4787333608,0.053078115,0.0148088718,0.1754537374,-0.3287909329,0.1579623073,0.0368430875,-0.1626701951,0.3949878514,0.5192049742,0.5169358253,0.3021615446,-0.1419557333,-0.2123039365,0.0545767657,0.1516158581,0.2379419059,0.7681572437,0.2752504647,-0.2857536077,0.3937153518,-0.0513014123,-0.0828840286,0.4031175077,0.056044139,-0.0689928606,-0.0268758666,0.4426218271,-0.1370540559,0.1148833409,-0.0047031129,0.1161605567,-0.2987504005,-0.3102475703,0.1035747752,-0.1718194783,0.288946867,-0.0753353015,0.034251716,-0.3127886057,0.5959126949,0.4353995323,-0.2124710828,-0.2104101181,-0.1160924658,-0.4998012781,0.3454267383,-0.4612213969,-0.232844919,-0.1896778494,0.4772252142,-0.2099053562,0.2936579585,-0.1712387502,0.4648705125,-0.1285225749,0.0775583684,-0.2698149979,-0.0518827178,-0.081744276,-0.1910948604,0.0219846107,-0.2261401415,0.1991904229,0.1046501249,-0.0841167495,0.0321385264,0.0574894659,0.0269974619,0.4276269376,-0.0821514353,0.5023901463,0.4290032685,-0.1451623887,-0.2173262239,-0.0876893997,-0.2888011634,-0.2974659204,0.3786478043,0.1362177879,0.4217302501,-0.175104171,0.647175014,-0.1287107319,0.3200820982,0.0195913576,0.0118474076,-0.3271249235,-0.0376971103,-0.214524433,0.0260582119,0.1519396305,0.471006304,0.1763303131,-0.017447222,-0.1815591156,-0.3055366278,0.42345348,-0.319660157,0.034292575,-0.026695773,0.1810691953,0.2032095343,-0.1110123768,-0.6931037307,-0.0768073872,0.2931550145,-0.1101622656,0.0857436135,-0.0357915387,-0.0713719577,-0.2900959253,-0.0223081,-0.2207989544,0.2341665924,-0.3639999926,0.384462297,-0.0856123567]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/215","title":"NonMatchingSplitsSizesError when loading blog_authorship_corpus","comments":"I'm getting this same issue when loading the `imdb` corpus via `dataset = load_dataset(\"imdb\")`. When I try `ignore_verifications=True`, no examples are read into the `train` portion of the dataset. ","body":"Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.","comment_length":29,"text":"NonMatchingSplitsSizesError when loading blog_authorship_corpus\nGetting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.\nI'm getting this same issue when loading the `imdb` corpus via `dataset = load_dataset(\"imdb\")`. When I try `ignore_verifications=True`, no examples are read into the `train` portion of the dataset. ","embeddings":[-0.1060908213,0.171928525,0.086146906,0.371528089,-0.0706182346,0.1975203604,-0.0885161087,0.3589629531,-0.070593968,0.1323675662,0.0109635042,0.1926111281,0.0197329056,-0.1354331821,-0.0613517575,0.001089591,-0.0496850312,0.1512279063,0.0807428062,-0.0007668391,-0.0190830566,0.4366812706,-0.3507992327,-0.06492351,-0.2797681689,-0.1714482605,0.0050599738,0.3068122268,0.0677898526,-0.3383013606,0.246760428,-0.0821123496,0.1046871692,0.2587444484,-0.0001232282,0.034662582,0.6713196039,-0.2309208065,-0.1916342527,-0.3499614596,-0.3364582956,-0.2255395353,-0.007184539,-0.2159268856,0.0023939852,-0.0712563694,0.1934566498,-0.1592053622,0.17167373,0.3034625351,0.0969203934,0.168262884,-0.1791486144,0.0531807207,0.2252613604,0.1842544079,-0.0827900097,-0.0836396515,-0.096333757,-0.1391977817,-0.1313942224,0.4392049909,-0.2431436926,0.1521119773,0.2753120065,0.1611486673,0.2986012101,-0.2021783143,0.0186963435,0.3758307695,0.4876859188,0.1042603627,-0.3134584129,-0.4765708148,-0.0225400496,0.1831901968,0.3666220903,0.3496968746,-0.2403942049,0.0368560702,-0.4264369011,-0.029453624,-0.1325446963,0.0899536535,0.2136926055,0.1008616164,0.1553698927,-0.0032718065,0.0446611755,-0.0954588428,0.1590351611,-0.2373013943,-0.1132026687,0.0426563062,-0.1377905756,-0.1338654459,-0.0908062235,0.0153885586,0.1906351745,0.2048220485,0.3135951459,0.0532241315,0.2250049561,0.2191824764,0.471835047,-0.0458391346,-0.0841466635,0.559345603,0.2446117997,0.0300255381,0.0602370501,-0.0100380322,0.2779305279,-0.1237685531,-0.0874430612,0.0251359604,0.2521752417,-0.5562335849,-0.4696909189,0.1025406718,-0.4754393697,-0.1108555794,-0.0499455221,0.2599775493,-0.3506717682,0.3180253208,-0.1121230498,0.111800082,-0.3332261443,-0.0451678149,-0.1908111125,0.1568987668,-0.1093215421,0.1896168888,0.303100884,-0.099793829,0.4644140601,-0.1483693421,0.0527665131,-0.1652436405,0.4266853034,-0.1766819656,0.1208889857,0.226992026,0.0736209899,0.2308332473,-0.0224069133,-0.1434951574,-0.3749013841,0.1423525661,-0.1875197738,-0.5286055207,0.1108296365,0.1250460744,-0.3993974626,-0.0579533428,-0.0107940845,0.1903902292,0.3466759026,-0.385879159,-0.0868320987,-0.1637050062,-0.0795989782,0.004998249,-0.1590004116,0.4613084793,0.1852805912,-0.2022434771,0.1531981677,0.0539302379,0.5437243581,0.3525790572,0.0245580897,0.0222527664,-0.1407122016,0.4579935372,0.1890598536,-0.4663132131,-0.1278700083,0.5851631761,0.0991248861,0.3170831501,0.2227521688,0.1468230635,0.2337536663,-0.0608233847,0.0406913161,0.4009259343,-0.089827925,0.0759612471,-0.3674883544,-0.314070642,0.4080738723,0.1790603995,0.1312666237,0.0107996715,-0.0512868501,0.4866945446,0.2602523863,-0.0848612264,0.0902554318,0.2709386647,-0.4059148729,0.1325216144,0.0775727183,-0.0969680697,-0.3680007756,0.0032126242,-0.1876494735,0.2800232768,0.0573012233,-0.3265427351,-0.1517819762,-0.4228248,-0.0079327887,0.0248874761,-0.0178245157,0.2539552152,0.3247401714,0.0933306366,-0.019458659,0.3759137094,-0.0265187435,0.2407079041,-0.8501710892,0.1395420432,-0.1763742566,-0.2839396894,0.476364851,0.2768564522,0.1811404675,-0.1480883509,-0.0746747851,0.4480014443,-0.0722123757,0.4661310315,-0.3489920199,0.0225239396,0.3167486191,-0.2829039991,-0.0194459856,0.3597464263,0.0122471917,-0.0611301996,-0.1798427999,0.3726072907,-0.3008643389,0.2264282554,0.1057961434,0.0095432913,0.1256736368,-0.2296975404,-0.2462262511,-0.57834059,0.5604595542,-0.1039939523,-0.3391008377,-0.0010969744,-0.2969707251,-0.0197271742,0.4288889766,0.1164335161,-0.0657130927,-0.0364805199,-0.0052561536,-0.1504425108,0.1934147626,0.1804919243,0.4251812696,0.1282400191,0.0847087204,0.0127721839,-0.1702791452,-0.3577148914,0.1167976186,0.0517169423,0.0034514088,0.3891897202,0.0633616149,0.0323283598,-0.3454785347,-0.0279580597,0.2083341926,0.1039574668,-0.3814579844,-0.1296589077,-0.4263542593,-0.3968114257,-0.2394592166,-0.1771837473,-0.1215176284,-0.3765020967,0.2047392726,0.3355308175,-0.2092926353,0.1140846014,-0.2720216811,-0.0122208893,-0.3346613944,-0.0690802336,-0.070718579,-0.2245487273,-0.2937884331,-0.0372127742,0.1956599653,0.4540426433,0.2249233127,-0.3926710486,-0.1600924432,0.0544624589,-0.2624689639,0.0824066848,-0.0852065012,-0.0717430115,0.1191216335,-0.1622812003,0.3704144061,-0.2849778533,0.0295722894,-0.1612708718,-0.2947843075,0.0364830159,0.1533923298,0.0339845903,-0.2557273805,-0.5162029266,-0.218612805,-0.3656134009,-0.1080161482,0.2695895731,0.1867019683,0.0947451964,-0.1517426074,-0.0998639241,-0.0336693861,0.3037445247,-0.2530958951,-0.1118111387,0.335735172,-0.2215276062,-0.1377462745,0.054698009,-0.0554154105,-0.0578569882,0.1281909198,-0.5356531143,-0.1128276885,-0.2253438383,-0.0531788655,-0.2187703848,-0.3081253767,0.4669430256,-0.0796346962,-0.0247839969,-0.0377102904,-0.3831616938,0.0780759752,0.061214678,0.4086661041,-0.1168704852,0.2762599587,-0.0010489066,0.4330562949,0.5189310312,0.0885115564,0.003335777,0.1372217834,0.1260214895,0.1028815135,0.0281622969,0.3039752245,-0.0209763739,-0.1285870522,0.3586798608,0.0653785542,-0.2829352319,-0.1473605484,-0.1174909845,-0.3157087862,-0.266025126,0.1234096959,-0.3443300426,0.1294119656,0.193455413,0.3407942951,-0.0818226561,-0.5395624042,-0.0183910392,0.2862555385,0.2730646729,0.135222435,-0.4073795378,0.1556653529,-0.4685332477,0.2718783319,0.1803425103,0.3475734591,-0.1694798917,-0.1578041613,0.0554876179,-0.2014870346,0.3394175768,-0.4746232033,0.0132008754,-0.0041793068,-0.07114508,-0.4237633049,-0.189686954,-0.1853719205,0.4030248225,0.4932019114,0.4149579704,-0.3782071769,-0.0771943927,0.2414444089,-0.084624894,-0.0541401953,-0.1513752192,-0.2128935754,-0.3216530681,-0.2734083831,-0.1536314934,0.0058206916,0.2583266795,-0.1015676111,-0.1375629306,0.0953451172,0.1440445036,0.4129194915,0.2945846319,0.1835896522,0.0181447584,0.1495036036,0.3504323661,0.4505655468,0.179333657,0.4881430566,0.0603254884,-0.1553129256,0.0629877076,-0.3202917278,0.0437899083,0.4678914845,0.199555248,0.0219019782,0.1291142553,-0.143758893,-0.3940744996,-0.0053221625,0.0670293346,-0.1741766036,-0.3361240625,-0.3677902818,0.2681986988,0.3401359916,-0.1109249592,-0.1660006046,-0.1509868801,-0.3498974144,0.3399803936,0.1721284986,0.9456439614,0.288297236,-0.0156865753,-0.0824363157,-0.1753872633,0.8163893819,-0.2360059321,0.4304829538,-0.3811824918,-0.2464488894,-0.0887557194,-0.1746107787,0.1384228766,0.1136195809,-0.3597877026,0.3247306049,0.1303015947,0.0346145369,-0.1917145997,0.4030131102,-0.2849452496,-0.1913117915,-0.1555311978,0.0236926135,-0.168181777,0.3843036592,0.0553742014,-0.0797490925,-0.2306410968,-0.2179843485,-0.4340411723,0.2383931726,0.1523971409,0.0419916883,0.2618045211,-0.211929664,-0.0214440245,0.1229697466,0.3339589536,-0.1773292273,-0.1743271202,-0.0057399604,-0.1033720747,-0.1869614273,0.1410845369,0.0593638532,0.384488225,-0.0987360105,-0.011703006,0.0815649703,-0.1594619751,-0.2650088966,-0.174231708,-0.1326058954,-0.1864432245,-0.1896418035,-0.1694425792,0.013096814,-0.06749513,-0.0892770439,0.0130885961,-0.1344005466,-0.1561055332,0.2143868357,-0.0344938375,-0.3317168057,-0.1097909287,0.4217988849,0.335377574,-0.0034993209,0.4691964686,-0.0698864236,-0.115427576,-0.1543066204,-0.0799780339,-0.1490003765,-0.3325285316,0.2099364698,-0.2215031385,-0.0826893821,0.2382518798,0.2852213681,0.1336510032,0.2915156186,-0.1226885021,-0.471200943,-0.0095611932,0.2567787468,-0.0191939529,0.1148132682,-0.1499743909,0.3957259953,-0.1862943769,0.1535694599,-0.1901642978,-0.0731281787,-0.4806300998,0.3539187908,0.0994266346,-0.007976667,-0.0888339281,-0.0276491493,-0.1656598002,0.2819246352,-0.0191594064,0.0067090979,-0.1328140199,0.1978199631,0.0979128703,-0.101104863,-0.139944762,0.1161259785,0.0205470156,-0.0048481212,0.0378846936,0.2670261562,-0.0089940373,0.3893877566,-0.367716223,0.090631716,-0.0589551926,0.1878202409,-0.3241512179,0.2810364366,-0.1203465164,0.2107488662,-0.0449329689,0.0443634018,0.0336269215,-0.0285209864,0.032807827,0.0435926653,0.1490328163,-0.1627855152,0.1406096369,0.3381780684,-0.0239354558,0.5712109804,-0.0635782555,-0.1615829766,0.0300997868,0.0879779235,-0.0846195444,-0.0033463186,0.2025799751,-0.0604711287,-0.0683539882,0.3554304242,0.1837961376,-0.1007057503,-0.1846408099,0.1138505042,0.4215512276,-0.1778310984,0.192425251,0.1978384256,-0.2146585733,-0.0157837477,0.1678561866,-0.2099667042,0.1965464056,0.0457775556,0.0202096645,0.4743586183,0.1578319073,0.3953711987,-0.0726073608,-0.2118465304,-0.0680852458,0.6018937826,0.3082668781,0.4337646067,0.2270940691,0.3257051408,-0.2783255875,-0.0001095997,-0.3504144549,-0.1837607771,-0.0114174336,-0.1646630913,-0.4718987942,-0.0938857347,-0.2394476682,0.0969730169,-0.1547926366,0.15203996,0.3951610029,-0.0444191657,-0.2059871405,-0.5548540354,0.1378757209,-0.1304259449,0.1931203753,-0.2679717243,0.1846107244,0.0576746725,-0.1153370887,0.4310167134,0.5478070974,0.5472561717,0.3214588463,-0.1735173613,-0.1106422469,0.0334420986,0.0327796154,0.190327853,0.7335736156,0.3221981823,-0.2711010277,0.3745810986,-0.0590752028,-0.0548163503,0.4139440358,0.1274645329,0.0154510075,-0.0634217188,0.5102357268,-0.1708881706,0.1025093719,0.0068509313,0.0252489094,-0.2746662796,-0.3511106968,0.2094674259,-0.3275750875,0.2222513109,-0.0307050943,0.0189562999,-0.2546964288,0.5723082423,0.4384699464,-0.1409322619,-0.2188490927,-0.0106535247,-0.484533757,0.3560202718,-0.3736160696,-0.1640263796,-0.1328224838,0.5531371236,-0.2862461805,0.2392934114,-0.2779918909,0.3856370151,-0.1746870577,0.1658014059,-0.2571769059,-0.0135681927,-0.1085549667,-0.2218365669,0.0066351276,-0.2400313318,0.1432487518,0.0427373052,-0.1245212778,0.0315433331,-0.0243344773,0.0671446398,0.4972291291,-0.0039025699,0.478713274,0.465631932,-0.1569473445,-0.2309437543,-0.1626692861,-0.3780436218,-0.3439884782,0.3681852818,0.1728878319,0.4330153465,-0.1747612506,0.5583011508,-0.0461825728,0.274816364,-0.0432048887,-0.004775241,-0.3202866614,0.012488246,-0.2711303532,0.0250669681,0.1263644546,0.3703691065,0.173021391,-0.0207456183,-0.18759045,-0.4143213034,0.430175215,-0.3541780114,0.0532094091,-0.1393235475,0.1548424512,0.2172611058,-0.142222032,-0.7538333535,-0.0806404427,0.2726385295,-0.1609559059,0.0506149903,-0.0131042274,-0.0147820078,-0.2709468603,-0.1554519385,-0.2552728653,0.2476238459,-0.3734526336,0.3072869778,-0.0980558321]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/215","title":"NonMatchingSplitsSizesError when loading blog_authorship_corpus","comments":"> I'm getting this same issue when loading the `imdb` corpus via `dataset = load_dataset(\"imdb\")`. When I try `ignore_verifications=True`, no examples are read into the `train` portion of the dataset.\r\n\r\nWhen the checksums don't match, it may mean that the file you downloaded is corrupted. In this case you can try to load the dataset again `load_dataset(\"imdb\", download_mode=\"force_redownload\")`\r\n\r\nAlso I just checked on my side and it worked fine:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"imdb\")\r\nprint(len(dataset[\"train\"]))\r\n# 25000\r\n```\r\n\r\nLet me know if redownloading fixes your issue @EmilyAlsentzer .\r\nIf not, feel free to open a separate issue.","body":"Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.","comment_length":100,"text":"NonMatchingSplitsSizesError when loading blog_authorship_corpus\nGetting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.\n> I'm getting this same issue when loading the `imdb` corpus via `dataset = load_dataset(\"imdb\")`. When I try `ignore_verifications=True`, no examples are read into the `train` portion of the dataset.\r\n\r\nWhen the checksums don't match, it may mean that the file you downloaded is corrupted. In this case you can try to load the dataset again `load_dataset(\"imdb\", download_mode=\"force_redownload\")`\r\n\r\nAlso I just checked on my side and it worked fine:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"imdb\")\r\nprint(len(dataset[\"train\"]))\r\n# 25000\r\n```\r\n\r\nLet me know if redownloading fixes your issue @EmilyAlsentzer .\r\nIf not, feel free to open a separate issue.","embeddings":[-0.1760099977,0.1923228204,0.0658075586,0.4156303108,0.012083455,0.193692103,-0.060682267,0.3878968954,-0.0591741167,0.1324820668,-0.0745147988,0.1716659218,0.0093409549,-0.1365589201,-0.0743669644,0.0001403463,-0.0610307083,0.148367539,0.0798779055,0.035928838,-0.0695639327,0.4301069975,-0.3462486863,-0.0888401791,-0.2378142327,-0.1127559021,0.0294194818,0.3264782727,0.0462553427,-0.362829715,0.2542954683,-0.0668971688,0.1275963783,0.301310122,-0.000123778,0.0187365673,0.6639293432,-0.2431373149,-0.1930371374,-0.3823130727,-0.2983905673,-0.2795032263,-0.0311899316,-0.2351135463,0.0000467804,-0.0455775484,0.1376455873,-0.1730502546,0.2249151021,0.3440396488,0.0974616855,0.1834928989,-0.0925603881,0.1112376601,0.2157849818,0.1207833514,-0.0489259101,-0.0144525822,-0.0177792516,-0.1584300846,-0.0385266319,0.386231631,-0.332886517,0.117399022,0.2570728064,0.149095431,0.3444834054,-0.1964065731,0.0314813405,0.3853250742,0.4552387893,0.0204213522,-0.3434735537,-0.438593775,-0.0033312282,0.0954965949,0.3857685626,0.3656494915,-0.2501454055,0.0412016883,-0.3448823988,0.0202553086,-0.123211056,0.1056362167,0.1657611728,0.0833722576,0.1527453512,-0.0133266831,0.0407765508,-0.0361177921,0.1446164101,-0.2656333447,-0.1059982851,0.0934185162,-0.1632417142,-0.1253025234,-0.1892376244,-0.0111017358,0.1671974361,0.1999484599,0.2925247252,0.0101394793,0.191058889,0.1882573068,0.469216913,-0.0550310351,-0.1077874675,0.5959754586,0.2141971886,0.0886910558,-0.0053858538,-0.0311124939,0.2485347092,-0.2011571229,0.0270569436,0.1297849119,0.2478101552,-0.5492246151,-0.5105435848,0.1785078794,-0.4724813402,-0.1294340193,-0.0333772078,0.2145411223,-0.3933617175,0.3246038258,-0.0649747923,0.1226364151,-0.3191176355,-0.0042859917,-0.1595644504,0.1274048835,-0.1345994771,0.1621820182,0.289172411,-0.1728476882,0.4029968977,-0.1103069633,0.0636114776,-0.2159257084,0.3070893884,-0.2068894506,0.065681994,0.2388720661,0.0471848585,0.2478989661,0.0007347426,-0.2099758983,-0.3573240042,0.1365561336,-0.2047909498,-0.4814672768,0.0713409409,0.1207398176,-0.3954432607,-0.0478433408,-0.0742641538,0.0361789316,0.3651829064,-0.4267303646,-0.0959469229,-0.2285137922,-0.0722583681,-0.0359340496,-0.1052930355,0.5175378919,0.1542671472,-0.1689531654,0.1386117041,0.0594591536,0.5408571959,0.3914321661,0.0583208688,0.0255098529,-0.1914317757,0.376696825,0.2035428286,-0.5307164788,-0.2367614061,0.5350233316,0.1370008439,0.3692914844,0.2253634185,0.1480951458,0.2658122778,-0.0620126352,0.016567653,0.423915863,-0.0463287123,0.0777939856,-0.3728453517,-0.2930844128,0.3921938837,0.1419514269,0.1503074318,0.043526344,-0.021335233,0.4459542632,0.323789537,-0.0458554961,0.0935382023,0.2675275505,-0.2806486189,0.2581963539,0.0494296029,-0.220071733,-0.3241138756,0.0276459549,-0.1992612481,0.2994350791,-0.0963057578,-0.2697671354,-0.1718753874,-0.4570285678,-0.0420803465,-0.0074901036,-0.0066508995,0.2086012214,0.2889246345,0.1307027191,-0.0683569536,0.4250882566,-0.0938529223,0.2411918938,-0.8024850488,0.2409891933,-0.1955877095,-0.3173593283,0.4423464239,0.2632887065,0.1987364143,-0.149531275,-0.1410844624,0.4424902797,-0.0311328247,0.4205708802,-0.3902110755,0.0709907487,0.3251236081,-0.3283196688,-0.0136464769,0.3792682588,0.0683734417,-0.091307804,-0.0997138619,0.3737076819,-0.3763327003,0.219367981,0.130830586,0.0296715721,0.1335918605,-0.2117361724,-0.272849232,-0.5291529894,0.5458315015,-0.0643115565,-0.312831223,0.0047126291,-0.2317238897,-0.063375555,0.4167566001,0.0390755311,-0.071549952,-0.0163541958,-0.0056102965,-0.097375907,0.0756489262,0.2083393335,0.4420288503,0.0971901342,0.0137554286,0.0784996748,-0.2087429613,-0.3179133534,0.1633073092,0.0187568143,0.0033221028,0.4489204884,0.0139329536,0.0849304721,-0.2789880037,-0.0302163623,0.1136982739,0.1889391243,-0.4271098375,-0.0949904174,-0.4230710566,-0.3812933564,-0.3252099454,-0.1917271614,-0.152148515,-0.3741566539,0.102218762,0.335670948,-0.0933515877,0.1520268768,-0.2747816145,-0.0771090239,-0.2856377661,-0.1671758443,-0.0400727093,-0.2553856969,-0.2222131193,-0.0603001043,0.2851201594,0.4162527919,0.2191535085,-0.4314364791,-0.1876335591,-0.0039614425,-0.2128283232,0.0471390933,-0.0921220183,-0.0109143779,0.1246257499,-0.0966826528,0.3530616462,-0.3084542751,0.1366530657,-0.1865907758,-0.2865263224,0.0870189667,0.1654970944,0.0059019439,-0.1674925834,-0.4755873978,-0.2604391873,-0.3520001173,-0.1506460309,0.2697541118,0.1826926023,0.0912531465,-0.1219488308,-0.1085171327,-0.0393231586,0.3502364457,-0.2149682194,-0.1261631995,0.2993681133,-0.1570667475,-0.1714879125,0.0066655269,-0.0190337375,-0.0280244853,0.1299344897,-0.5250765681,-0.1266783178,-0.2553530335,-0.0194481555,-0.2235793322,-0.2169738114,0.40171808,-0.0199700054,0.0072240285,-0.0070943772,-0.3523037732,0.0274134036,0.0531912521,0.3816002309,-0.134692505,0.3094494343,-0.034442164,0.4264933169,0.5960376859,0.0360532515,0.0276391506,0.071342811,0.1436524987,0.0747463033,-0.1054499149,0.2657156885,-0.0729660168,-0.0789038166,0.2505123317,0.0089181019,-0.2238316387,-0.1783442199,-0.1296456605,-0.4514029324,-0.2732064128,0.1005958095,-0.3199054599,0.1766841412,0.1375228167,0.2814928293,-0.0551787391,-0.5646659136,0.029397862,0.4150556326,0.214581579,0.1150398627,-0.3385314345,0.1420570016,-0.3411735594,0.2643878758,0.2061057091,0.3740829229,-0.1111693829,-0.1137643754,0.0390937887,-0.1925824434,0.3788112998,-0.474663347,0.055516839,-0.0040849685,-0.0704109818,-0.4118959904,-0.2191711068,-0.1826020479,0.3691924214,0.4657855928,0.4691717923,-0.3520938456,-0.0619661622,0.2556953132,0.0213910528,-0.1011046991,-0.1745802462,-0.2285864204,-0.3355115056,-0.2618886232,-0.1732980162,-0.0438754782,0.3112166226,-0.0753442496,-0.0824190676,0.0422509387,0.1665192395,0.425327301,0.2312405556,0.2076847255,-0.0324012712,0.2240991592,0.2517690957,0.4094746411,0.2128598988,0.5298275948,-0.0027636692,-0.1928169578,-0.0765536949,-0.3011442125,0.0054931254,0.4374139309,0.2044096291,-0.0304501634,0.1263709515,-0.1373873949,-0.4296751022,-0.0529444106,0.1266048551,-0.1013170481,-0.3854480684,-0.4973120987,0.2998217046,0.3504993916,-0.1128722206,-0.1433965117,-0.0995505899,-0.3347936273,0.3483667076,0.1721326858,0.9992435575,0.2270867825,-0.025044417,-0.0683604553,-0.2009780854,0.7844552398,-0.2129834145,0.3598971963,-0.435025394,-0.2963214219,-0.1060200036,-0.1812584549,0.1910452247,0.1032356992,-0.3326264024,0.3883803785,0.1092572138,-0.0173381343,-0.2038629353,0.4057893157,-0.2884773016,-0.1145381778,-0.1564046741,0.0208074059,-0.1877886355,0.4317868948,0.0824533924,-0.0572607145,-0.2624443471,-0.1683529913,-0.5494999886,0.221664682,0.1427233815,0.0263723992,0.2118000239,-0.2248521298,-0.0135899018,0.1342342794,0.3278855085,-0.1151459739,-0.1068169251,-0.0016029181,-0.0226165187,-0.1557308137,0.2007645518,0.0577108301,0.4340377152,-0.1015535966,-0.0617629029,0.1091218516,-0.0791868716,-0.2611496449,-0.1338943392,-0.1395938843,-0.1802152693,-0.2108575702,-0.2031707466,0.0140566202,-0.029255975,-0.0973197818,0.0170064196,-0.0592136644,-0.1978055686,0.2540107369,-0.0918475762,-0.3513323367,-0.0954819247,0.447970897,0.2841574252,-0.0474261381,0.4655070901,-0.0447688475,-0.1163854524,-0.142869994,-0.0799196213,-0.2282397598,-0.3608661592,0.170556426,-0.2503566444,0.0032193675,0.2146572471,0.2766200602,0.231381163,0.2816073298,-0.0997282937,-0.5418647528,-0.0058851074,0.1627014726,0.0923534781,0.1533018649,-0.1904150844,0.385199517,-0.2032115459,0.0885573849,-0.1887628138,0.0016787729,-0.421643734,0.3190796971,0.0986846536,-0.0103346314,-0.1064795628,-0.0536204353,-0.1095468849,0.2328643203,-0.0086235963,-0.0021025848,-0.1735310256,0.1991340816,0.0781842843,-0.146214202,-0.127779603,0.02929781,0.0719479024,-0.0399061665,-0.0328832939,0.3182603121,0.016946191,0.4303997457,-0.2967918217,0.007057738,-0.0522233434,0.2537626028,-0.3740801811,0.3204038441,-0.0648331568,0.2339768261,-0.002991342,0.0760770664,0.0052502281,-0.0109396074,-0.0333061293,0.0442507304,0.260317862,-0.1823098511,0.1516094208,0.3269650638,0.0630297512,0.6621741056,-0.0977921039,-0.1435750425,0.1002192125,0.0893855393,-0.1212947071,0.0176852439,0.2556922734,-0.0632998124,-0.080075711,0.3751478195,0.1603872478,-0.1173736677,-0.2403319329,0.0951454192,0.5035053492,-0.1352917254,0.1872151494,0.2672548592,-0.0520813055,-0.0351174921,0.1195930541,-0.213570416,0.1130558103,0.0828209743,-0.0163976662,0.4081757367,0.0597969927,0.381071806,-0.1107972488,-0.1928135157,-0.0168840997,0.5765302777,0.1995475143,0.4437114,0.2377901673,0.3454336226,-0.3376035094,0.0259019658,-0.3117018938,-0.1353363395,-0.013287724,-0.173533693,-0.4493536949,-0.1327101588,-0.2263336629,0.1173153892,-0.1327744722,0.1628895998,0.3212738037,0.033501476,-0.2087710202,-0.5127853751,0.0697427467,-0.119269684,0.2277920991,-0.2851020098,0.1193503663,0.1218326837,-0.068002902,0.3805288672,0.5955414772,0.6343856454,0.3927333951,-0.1902579665,-0.0560278445,0.036397431,0.0595012158,0.2065488249,0.7163820267,0.2461320162,-0.2202010453,0.3402116299,-0.033732377,-0.0654515624,0.3719545007,0.0700091124,0.0255958978,-0.0862710327,0.523667872,-0.1916219741,0.1169350445,-0.0510605648,0.0679495633,-0.320607692,-0.3236781657,0.2613776624,-0.2641171515,0.2465271652,-0.0030771296,0.0153056616,-0.204777196,0.5182003379,0.5190639496,-0.0969166532,-0.243395403,-0.011231144,-0.5534388423,0.3506766856,-0.3848561645,-0.219726041,-0.1404456347,0.5029476285,-0.2913239002,0.2718916237,-0.1669805348,0.4217167199,-0.1999900341,0.2392649353,-0.2612939179,-0.0482010692,-0.0856673494,-0.2491249889,-0.0412074551,-0.2363372296,0.1247370988,0.000604218,-0.0942483395,0.030081287,-0.0263997037,0.0613665879,0.4949748218,0.0248681605,0.4822668433,0.5218866467,-0.1439878494,-0.2633271515,-0.175556913,-0.3367417157,-0.3082946241,0.4396428764,0.1312469244,0.3744220734,-0.1879597753,0.4798935056,-0.1375112534,0.3361933231,-0.025598906,-0.1167669743,-0.3755326867,-0.0512626022,-0.2222052217,-0.0311384983,0.0988998413,0.4211770892,0.1588225812,0.0884603411,-0.1798774302,-0.4090625644,0.4370062351,-0.3273996413,0.0430927314,-0.1127626225,0.1238812506,0.1499227434,-0.0391520225,-0.7592730522,-0.0359741598,0.3005142212,-0.1770361364,-0.0273163095,-0.0137074068,-0.0194511954,-0.2217527628,-0.110640578,-0.3095343113,0.2275260538,-0.3596231937,0.3135592043,-0.0814565197]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/215","title":"NonMatchingSplitsSizesError when loading blog_authorship_corpus","comments":"It doesn't seem to fix the problem. I'll open a separate issue. Thanks. ","body":"Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.","comment_length":13,"text":"NonMatchingSplitsSizesError when loading blog_authorship_corpus\nGetting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.\nIt doesn't seem to fix the problem. I'll open a separate issue. Thanks. ","embeddings":[-0.1390997767,0.0543702841,0.0728878602,0.3970887065,-0.0233884025,0.1375876367,-0.158470884,0.3796317875,-0.1168576628,0.1151458621,0.0551695116,0.1690439284,0.0559653789,-0.0888678581,-0.0592985824,-0.0234100502,-0.0562453046,0.2135821879,0.0507533997,0.0292638354,-0.0354876779,0.4359353483,-0.3135043681,0.0213658568,-0.2447722256,-0.1752839088,-0.0137599148,0.2874711454,0.0382478088,-0.3574716747,0.2444120049,-0.0978415757,0.143829599,0.1876740009,-0.0001205273,0.0085881595,0.5626437068,-0.2080705613,-0.1513448954,-0.2702901661,-0.3612966537,-0.3176291287,-0.0461670645,-0.2352032363,-0.052327659,-0.0771211609,0.1522168964,-0.0610055327,0.2657972276,0.3098547161,0.1432257146,0.1211863458,-0.1492287368,0.0293803997,0.1000850722,0.0449182726,-0.0589837357,-0.1206048802,-0.1262924373,-0.278546989,-0.0756688565,0.446459353,-0.2479012609,0.1574765742,0.2414804101,0.127018109,0.4128561318,-0.1042793691,0.1137193814,0.3601647615,0.518358171,0.1458729059,-0.1546526551,-0.4566423595,-0.0276504643,0.184095487,0.4334308505,0.3489078283,-0.2319483459,0.0217791963,-0.3921668231,-0.1255369782,-0.1050187573,0.0779489353,0.2505572736,0.2124220878,0.1060580611,0.0274914689,0.0803744271,-0.0669397339,0.149447456,-0.2277576625,-0.1564390659,0.102301456,-0.2189915478,-0.1815382391,-0.144269079,-0.0212550163,0.1892972291,0.1721398979,0.3529469371,-0.03989847,0.2495046854,0.1080902219,0.4239104092,0.0010633267,-0.0012275751,0.5342261195,0.1096567735,-0.0335097723,0.1056251377,0.0285924952,0.2809047997,-0.1972643137,-0.1095186174,0.0956624746,0.1829982698,-0.5564141273,-0.4327253401,0.1205030382,-0.533192575,-0.171918124,-0.085267432,0.2069253176,-0.3531675339,0.3531312644,-0.137059316,0.057097964,-0.342676878,-0.0631342009,-0.2085461915,0.1511295736,-0.1496572345,0.2122242451,0.1586439461,-0.1569141895,0.3920149207,-0.1266897023,-0.1014367789,-0.1875592619,0.3459009528,-0.1052030846,-0.0298469998,0.1569549739,-0.0278774686,0.2118312716,-0.0190590937,-0.2045071572,-0.3717663288,0.0986876935,-0.2253087908,-0.5575869083,0.0677632093,0.1637931168,-0.4025993943,-0.0599454753,0.0977071673,0.2169932872,0.395413518,-0.3337693512,-0.113816455,-0.1973689049,-0.1567428857,-0.0613630414,-0.205927819,0.4546979666,0.1726840585,-0.2752794027,0.1627058983,-0.020069629,0.5716105103,0.3604242206,0.0381911844,0.1010349616,-0.1591083407,0.4509415329,0.1506563127,-0.3559421599,-0.1681277603,0.5244192481,0.0074063437,0.2745908499,0.2148532122,0.1449384391,0.1533195823,-0.0631873533,0.052672904,0.4378495812,-0.0385914557,0.0431385525,-0.4584094584,-0.2987143695,0.4536646307,0.2083923072,0.1272431314,-0.0245661326,-0.0370775498,0.4920611978,0.2890051901,-0.1327906698,0.0700629503,0.303817153,-0.35489586,0.1305745095,0.0462898053,-0.1391822994,-0.3921821117,-0.0335409828,-0.2430020422,0.2586559951,0.093001157,-0.2401617765,-0.1995246708,-0.4056037962,-0.0370845124,-0.0161027014,0.0139495833,0.2463205457,0.3808025718,0.0757320672,0.0407497585,0.4788648784,-0.0682670921,0.1844462603,-0.8004388213,0.2258880734,-0.1292402148,-0.2202089727,0.5145196915,0.3384109139,0.2321903557,-0.1342605948,-0.1344905943,0.4476515353,-0.193993926,0.4418577552,-0.2434104681,-0.018171994,0.2304038107,-0.2843372524,-0.0075340504,0.3477758169,-0.0039349631,-0.0941466391,-0.0651096553,0.3836824298,-0.3326398432,0.1853171736,0.0248837993,0.0434236676,0.0871467963,-0.2607019544,-0.1655773968,-0.4632093012,0.6249533892,-0.1311710924,-0.3337543905,0.0032160541,-0.2205518037,0.0165784154,0.4196485579,0.0764315426,-0.080501765,-0.0247135963,0.0127588445,-0.1294916421,0.0991129875,0.2229521424,0.5090211034,0.1164514869,-0.0435801148,0.0517957546,-0.1455774903,-0.3330379426,0.1437144428,-0.0494489148,0.0768647492,0.3589205146,0.0264354814,-0.0191162247,-0.3077361584,0.0396250784,0.1840156764,0.2518124878,-0.3365499377,-0.164702177,-0.3918977082,-0.4000873268,-0.3134508729,-0.1489486247,-0.2630591691,-0.3611450195,0.1987921596,0.2202336639,-0.1869626492,0.1338957399,-0.1747693419,0.0072872872,-0.2875570655,-0.005991898,-0.0316521972,-0.2366093397,-0.2643128633,-0.0289235692,0.1699922532,0.4165933132,0.245063141,-0.3717840612,-0.1537464112,0.0514584519,-0.2748544812,0.0393779837,-0.0625249743,-0.1421100497,0.1471513808,-0.163874045,0.3715073466,-0.1808080971,0.1715852022,-0.2139939368,-0.2942385674,0.0796331987,0.1824671626,0.048040729,-0.1639449298,-0.5371729136,-0.2623079419,-0.3687801361,-0.1039111391,0.2075683177,0.2490413487,0.0225225184,-0.1235681549,-0.1442054808,-0.1604176909,0.2494483292,-0.306520462,-0.0378759205,0.2630345523,-0.0995462984,-0.1798778176,-0.0217144676,0.046750132,-0.0549395904,0.0894518271,-0.5378001332,-0.117268987,-0.1986848861,-0.0839537829,-0.3446650505,-0.2603244185,0.5175863504,-0.0787451491,-0.0193446875,0.0230198726,-0.3850727081,0.0701628849,0.106983453,0.5072275996,-0.1491715908,0.266967386,-0.0226724613,0.5214694738,0.5547996163,0.0757616907,-0.1211482659,0.1618015021,0.0790335014,0.1535653472,-0.0393598229,0.3397096694,0.0173482019,-0.0686977655,0.4249884784,0.0637746155,-0.1927740872,-0.1462300569,-0.1903501302,-0.2735652626,-0.204897359,0.1564238667,-0.2341938764,0.1908648163,0.2662275434,0.2564398646,-0.1655824631,-0.5315518975,-0.0217110105,0.2929988801,0.1487350017,0.134141475,-0.4952874184,0.1257002056,-0.4656862617,0.3087286949,0.1978250146,0.3108772933,-0.1625578552,-0.1267448813,0.1045910269,-0.1974734813,0.2868149281,-0.40148592,-0.0534099601,-0.0560073256,-0.0161543395,-0.3624183238,-0.1790499091,-0.0917230994,0.4312985837,0.4808051288,0.4195918143,-0.4001683593,-0.1538100988,0.165156737,-0.1002651826,-0.1505977809,-0.2094482183,-0.2542838454,-0.3227287829,-0.167222932,-0.1480347663,-0.13412112,0.336027652,0.021187434,-0.048574917,0.1433198154,0.1000445932,0.3105566204,0.2294199914,0.1309671253,-0.0410830379,0.1467442811,0.2938431501,0.5359700918,0.0746081173,0.5317272544,0.1344560832,-0.1185739264,-0.0383086577,-0.2798675001,0.1016604751,0.4251677692,0.1805332005,0.1224529296,0.1310683191,-0.0048803673,-0.4382737875,-0.0700162873,0.1123324558,-0.0952986032,-0.3789006472,-0.3476237059,0.2823292315,0.2714604437,-0.1257779002,-0.1421943307,-0.2170397639,-0.2733003795,0.2479399294,0.1378018409,0.9928952456,0.2631305754,-0.0411443487,-0.0589332767,-0.2632732093,0.8655105829,-0.2491923422,0.4478179812,-0.4563337564,-0.2216699421,-0.0403639413,-0.1827856749,0.1389153153,0.0990182683,-0.283095032,0.3401036859,0.2533677816,0.0470274873,-0.1000816599,0.3919045627,-0.2229803056,-0.0953552574,-0.2100000232,0.020734787,-0.3256898522,0.4798717797,-0.0096596424,-0.1108754724,-0.1890643835,-0.2411209494,-0.3951302469,0.2748768628,0.1353138536,0.0379702188,0.304574281,-0.1886456758,-0.006769204,0.0582836941,0.30177176,-0.1987732053,-0.1912958622,0.0031839448,-0.1830683649,-0.1743674874,0.1406021863,0.1005823612,0.3192259371,-0.1242022514,0.0040336382,0.0901837423,-0.1302050054,-0.3371326029,-0.2739905715,-0.105758056,-0.1550403535,-0.1972668022,-0.1066540703,0.0842155516,-0.0413121395,-0.0734587163,0.0496897548,-0.0597261935,-0.2288083881,0.2492052019,-0.0929760784,-0.3050024509,-0.119021371,0.4055173993,0.4117463529,-0.0106978621,0.4902607501,-0.0795095265,-0.0809931904,-0.2063788623,-0.1317732483,-0.2697767317,-0.3592654467,0.1829225719,-0.2539055049,-0.0864631087,0.2013558596,0.2903880179,0.1880082488,0.2875227332,-0.00629606,-0.5056512952,0.0275460891,0.1305948198,-0.0482943989,0.0469470397,-0.1346324533,0.4510463476,-0.2433176041,0.2773605287,-0.2288202941,-0.0931466073,-0.412691772,0.3134311438,0.1175762564,-0.0927804038,-0.0487315692,0.0408960283,-0.116085276,0.2674724162,-0.0964981765,-0.0174068436,-0.0880300775,0.1952566057,0.173701629,-0.0705675483,-0.1341833919,0.0866906792,0.0282043777,0.040942654,-0.0358932577,0.2600072622,0.0068302257,0.3127038777,-0.3612632155,0.0481772721,-0.0730768964,0.233234331,-0.3470272422,0.3239580393,-0.1405519098,0.205767259,-0.0845177695,0.010039499,-0.0422914848,-0.0301118046,0.0919065103,0.0308137629,0.2036252022,-0.2016656697,0.0980508998,0.2867252529,-0.0225082953,0.5799649358,-0.0995382443,-0.128432706,0.084329091,0.1118413582,-0.1339660883,-0.0079067741,0.2356189191,-0.0360831097,-0.1452716887,0.3405309319,0.1387044191,-0.1628090888,-0.1737555712,0.1375172436,0.3945126235,-0.1678994596,0.1165120006,0.2251057774,-0.1423591226,-0.0392568856,0.2846504748,-0.2266076654,0.1406384259,0.0650246069,0.0500606,0.4054410458,0.1308926195,0.4265829325,-0.1227615327,-0.1302256733,-0.0763682052,0.6713598371,0.2720986009,0.3871884644,0.2006267756,0.4424688518,-0.3392941952,0.0020730118,-0.367798239,-0.1468293071,-0.0696284249,-0.1054331884,-0.4837701321,-0.0878322423,-0.2214512676,0.1060684621,-0.113295801,0.1402329654,0.4132087827,-0.111245729,-0.2018039823,-0.5195214748,0.0576438941,0.0400644094,0.1730602533,-0.3087953925,0.1884785146,-0.0024031452,-0.1787455231,0.3877169192,0.5563623309,0.493704021,0.336940676,-0.1534026116,-0.1664088815,0.047603596,0.1152781099,0.2952062488,0.7931679487,0.2812167406,-0.2830565572,0.4014455974,-0.0491339564,-0.0983837768,0.4062225521,0.0714808479,0.0061938362,-0.0295428149,0.4517222047,-0.1605320722,0.1174195483,-0.0383291319,0.0738706216,-0.3029996753,-0.2966123521,0.1334399879,-0.1928186268,0.2630923092,-0.0841693208,0.039964851,-0.3060409725,0.5812904239,0.4127190113,-0.2483303845,-0.2053968459,-0.0931272805,-0.4583533406,0.3687878251,-0.4389217794,-0.2055021971,-0.1790963262,0.4961315095,-0.2396727055,0.2699517608,-0.177440688,0.4333143532,-0.1360677928,0.0766976997,-0.2405069619,-0.0387560986,-0.1079072356,-0.2166958451,0.0337469503,-0.2174552381,0.2001355886,0.1094353646,-0.0870690942,-0.0068748756,0.0549366996,0.0106594153,0.422078222,-0.0467183255,0.5182469487,0.3783884346,-0.1094546318,-0.1790599972,-0.0941758901,-0.3594289124,-0.3048039377,0.4066346884,0.1956443191,0.4171393514,-0.1751425415,0.5918559432,-0.1291678846,0.2714194357,-0.0419240184,0.0052604275,-0.2923775613,-0.0006042605,-0.211395666,0.049250681,0.143063724,0.4527157247,0.1658990681,-0.015084384,-0.2100362331,-0.3582350016,0.4410968125,-0.3218519986,-0.0067762351,0.0091481004,0.1983474195,0.2335055023,-0.1307343841,-0.7347556353,-0.087390855,0.29023543,-0.1107677966,0.0886595026,-0.0301736146,-0.0668671057,-0.2910437584,-0.0282704215,-0.2307345718,0.2408433259,-0.3652136922,0.3988716304,-0.0741349012]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/215","title":"NonMatchingSplitsSizesError when loading blog_authorship_corpus","comments":"I wasn't aware of the \"force_redownload\" option and manually removed the '\/home\/me\/.cache\/huggingface\/datasets\/' dir, this worked for me (dataset 'cnn_dailymail')","body":"Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.","comment_length":19,"text":"NonMatchingSplitsSizesError when loading blog_authorship_corpus\nGetting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.\nI wasn't aware of the \"force_redownload\" option and manually removed the '\/home\/me\/.cache\/huggingface\/datasets\/' dir, this worked for me (dataset 'cnn_dailymail')","embeddings":[-0.0595206991,0.1531680971,0.0704791918,0.3878914118,-0.0151472939,0.1419075578,-0.0923853889,0.3200816214,-0.0493569039,0.0081106173,-0.0866535455,0.1576867253,0.0365480743,0.0461476073,-0.0197161119,0.089791663,-0.045651257,0.2041377276,0.0454466417,0.0290790945,0.0009838509,0.5112382174,-0.2444841117,-0.0645389184,-0.3506295979,-0.1498450339,0.0544163249,0.3629713655,0.0106631266,-0.3910149038,0.3197171986,-0.0472841188,0.1141889095,0.2401960343,-0.0001225521,-0.0028425201,0.4699682295,-0.2191404551,-0.2050274163,-0.2703422606,-0.2168163657,-0.2991623282,-0.0463629216,-0.2182562351,-0.066741392,-0.0665390417,0.0799646303,-0.0519919395,0.3719491065,0.2974589765,0.1255471259,0.0959382281,-0.1272954643,0.1026805043,0.0848564133,0.0665248334,-0.0118846782,-0.0418558493,-0.172001794,-0.3219947815,-0.1281981915,0.5180576444,-0.1915793717,0.1262466908,0.1819391698,0.0800840184,0.2830646634,-0.1194166392,0.0679115728,0.3065373302,0.4064788222,0.1085797623,-0.2249051183,-0.5339212418,0.0071086911,0.0513608046,0.4527995586,0.3414332271,-0.1976626068,0.1005414873,-0.4158658087,-0.2263238728,-0.0545980595,0.0072788424,0.213900879,0.2067903578,0.0388180427,-0.0100916652,0.1374930739,-0.0378837809,0.224408865,-0.0922396928,-0.1446522772,0.0849092007,-0.2126174271,-0.1594928801,-0.1238902807,0.0378153287,0.1740897149,0.153521955,0.2372107655,-0.0782316625,0.1446377188,0.0419277325,0.3693934977,0.1749555022,-0.0018965586,0.4282423556,0.1405402869,0.0332893133,0.1102929413,0.003799333,0.2939818799,-0.1572295725,-0.0670643449,0.0451962948,0.2099593282,-0.5436299443,-0.4165382385,0.0798673034,-0.4917491674,-0.2023522109,-0.1229509637,0.1320477128,-0.3499180973,0.3523450494,-0.0788428336,0.0378077105,-0.3732774258,-0.0559959561,-0.2014294714,0.1090007573,-0.1936449111,0.2363349497,0.2714067996,-0.2821205556,0.4370520711,-0.0514768139,-0.0933633447,-0.2561869025,0.2646710575,-0.0307807121,-0.0232919883,0.1904060394,-0.1313756704,0.2399367094,0.0085804611,-0.2008440346,-0.2873153687,0.0163729154,-0.2768751383,-0.6388433576,0.0812386647,0.1171502993,-0.4046169817,-0.0596416369,0.0174986944,0.2203211784,0.3705640733,-0.260425508,-0.1059131399,-0.1687602401,-0.1013388857,-0.029852597,-0.1653283238,0.5042775869,0.1327273846,-0.3224145472,0.1607013941,-0.0313843079,0.4741957188,0.4351542592,-0.0065021119,0.1073416546,-0.1725720763,0.3553568721,0.1428028047,-0.3949939013,-0.25507918,0.4000764191,-0.0010343348,0.2915739715,0.1832810938,0.1310875714,0.092016153,-0.0645947754,0.0591232218,0.4729750752,0.0742925406,0.0992293283,-0.452811271,-0.4169977903,0.4330647886,0.1688203067,0.1097708866,0.0770275891,-0.0199286789,0.4880301654,0.3150661588,-0.1443881243,0.0633854643,0.3153182268,-0.3313460946,0.1586517394,0.0493910871,-0.0894606039,-0.5167136192,0.0077179689,-0.2730098367,0.3195577264,0.0504484996,-0.1981106102,-0.2350796461,-0.4627166688,-0.0789123029,-0.180663079,0.0007562484,0.2350212336,0.4090325832,0.0545505546,-0.0182933733,0.5882976055,-0.0964862183,0.2448498905,-0.884516418,0.2083256245,-0.1277618259,-0.1689957678,0.412289232,0.28587237,0.2703772783,-0.1416527331,-0.1319459081,0.4632707238,-0.2258070111,0.4788300693,-0.2279370576,-0.0022640536,0.2731258571,-0.2956173122,0.021120755,0.2180161327,-0.0465648845,-0.0552712381,-0.0661935881,0.3710585833,-0.3695765734,0.1385298371,0.0175738316,-0.0447714627,0.1000527516,-0.2872813642,-0.1437971741,-0.4895391762,0.578715384,-0.1502457857,-0.2525633574,0.0214846786,-0.1338328421,0.0163826402,0.3793990314,0.0649042577,-0.0964649171,-0.0319759324,-0.0160528589,-0.2002706826,0.0535912029,0.2453751862,0.5202929378,0.1364887804,0.0263755564,0.0923336148,-0.1641586721,-0.3306786716,0.1524433047,-0.0643642396,0.082352072,0.3692166805,0.175146237,-0.0021436773,-0.5159095526,0.1042157263,0.2119722068,0.2446636707,-0.3040995896,-0.1348413527,-0.3648944199,-0.3866154552,-0.2667773068,-0.1904563159,-0.3162629008,-0.3300524652,0.2057839334,0.285590589,-0.1177580953,0.1592735499,-0.0924357772,0.0607687272,-0.3277837634,0.0091106473,-0.109533146,-0.275867939,-0.2055223733,-0.0503627546,0.1038396284,0.3221497536,0.1894622594,-0.3560625911,-0.1991535574,-0.0704813153,-0.3379549086,0.0474636145,-0.0875671953,-0.0716575757,0.0805038214,-0.0683810934,0.3928327858,-0.1585329026,0.1405427605,-0.2609811425,-0.3093508184,-0.0416108705,0.1713787615,0.1956088841,-0.1676418483,-0.5419095755,-0.2437300086,-0.3634908497,-0.0151568409,0.2141262144,0.1851948798,0.1718039215,-0.1324822605,-0.2091613561,-0.1113305613,0.2509459555,-0.392528981,-0.1390981078,0.195878163,-0.0768251568,-0.1303210109,-0.0714152902,0.1017391905,-0.0858046189,0.0706128478,-0.6237045527,-0.2201807648,-0.2419821471,-0.0938954949,-0.3028301895,-0.1780101806,0.5007231236,-0.0454118252,-0.0041329497,0.0809499621,-0.4596955478,0.0179481264,0.1331700832,0.4873123765,-0.1790713072,0.2381062061,-0.019556094,0.5533105731,0.5295944214,0.0531762056,-0.0099898465,0.1688424796,0.2262649983,0.0400814116,-0.1329494268,0.3330087364,-0.0689118057,0.0372624025,0.3456496298,0.1008637771,-0.1492968798,-0.1453624815,-0.1662190109,-0.3622741401,-0.180170089,0.1373778731,-0.1392010599,0.1204248145,0.2959349453,0.1435425133,-0.1756931245,-0.4716587663,-0.0087087313,0.2427840233,0.1436542571,0.1978633255,-0.4579145312,0.1711394191,-0.5147839189,0.2657485008,0.1584010273,0.2646610737,-0.1318333298,-0.1470313668,0.0993958116,-0.172826618,0.3940214515,-0.4049489498,-0.0772495121,-0.0796756819,-0.0167160556,-0.3363564014,-0.1784145534,-0.0354516357,0.3083250523,0.4751313925,0.4908389151,-0.3914109766,-0.1579650193,0.1868767142,-0.0096696522,-0.1367051005,-0.137030378,-0.2299737185,-0.3890261352,-0.276325345,-0.1139221787,-0.1300932616,0.3114847839,0.0253745876,0.0168256462,0.1222649887,0.0988522768,0.2510802448,0.1897909194,0.1321551651,0.0253745522,0.0777809918,0.3831085861,0.5103031397,0.0369195938,0.6296110153,-0.0108793834,-0.037916407,-0.0879920274,-0.0814360976,0.0274067819,0.550566256,0.2001908273,0.0538075715,0.1374895573,0.0104315737,-0.4142747223,-0.0408380516,0.1152987704,-0.0883520171,-0.4278704822,-0.294110328,0.316978395,0.2353852838,-0.1339420229,-0.1313208491,-0.1356355697,-0.2719816566,0.2453358024,0.0552335419,1.0419280529,0.207284376,0.0418646447,-0.0558132082,-0.1492008269,0.9380419254,-0.3000783622,0.417637229,-0.3517107069,-0.1780042052,-0.0575034693,-0.1474669725,0.0360339694,0.0908773318,-0.1540603638,0.3749519289,0.2888363302,0.0504398979,-0.1297356784,0.4234961569,-0.2495972812,-0.1523145586,-0.3030506074,0.0170570854,-0.3022836447,0.540066421,-0.0381429829,-0.0711969733,-0.0894839317,-0.2042330056,-0.3720652461,0.2525860071,0.0241727773,0.1293673664,0.2318016589,-0.1961918473,-0.0776413754,0.1056078672,0.4005057812,-0.1648745835,-0.1290960014,-0.0062112506,-0.1460076869,-0.1272876561,0.1189286634,0.0915602371,0.2880973518,-0.0980723128,-0.0300803334,-0.0138033461,-0.1449225992,-0.3308665156,-0.248413682,-0.0519222505,-0.160809353,-0.2244493365,-0.0149964169,0.1725162268,-0.028158294,-0.1344454587,0.0208315402,-0.0476447977,-0.1695937961,0.1148085296,-0.0948010758,-0.3547864854,-0.186608851,0.4653012455,0.4119220674,-0.1300046593,0.4889164567,-0.1077917293,-0.0471522845,-0.161250487,-0.1859859526,-0.2622637451,-0.425476104,0.2058079243,-0.2105523348,-0.1763063222,0.2077461183,0.260540694,0.2374656051,0.3127465248,-0.0139690917,-0.4817534387,-0.0392549522,0.1360348463,-0.0111888731,0.0911906213,-0.0245403089,0.3797795177,-0.240981698,0.3181025386,-0.2151935101,0.0392882861,-0.3406300843,0.3029212654,0.1933193952,-0.1392346025,0.0678023696,0.0129206814,-0.141055882,0.264136076,-0.1022029072,-0.017112961,-0.1405878514,0.1969378442,0.15148592,-0.111853689,-0.1557070911,0.0798286349,0.0645197555,0.0160162095,0.0632984564,0.2837772965,0.067524381,0.3050287664,-0.2870919406,-0.0395348072,0.0005167002,0.1990904361,-0.2437739968,0.2209280133,-0.0393500179,0.2127682418,-0.1637399793,0.0593699776,-0.0996329486,0.0008992943,0.086391896,0.0896608904,0.0894290358,-0.2511567473,0.0347975194,0.2602270246,0.0537947454,0.5116800666,-0.083156012,-0.1543081403,0.124841623,0.0871141404,-0.1410171241,-0.0620493479,0.3102421761,-0.0773708001,-0.0963805467,0.3810330927,0.1730477661,-0.1297150552,-0.1407459229,0.1068659276,0.4532064795,-0.1916285902,0.1509832889,0.1268143356,-0.1055742428,-0.0867417976,0.3743505478,-0.2010780424,0.1221473813,0.1104713604,0.0450539775,0.372074008,0.0547282398,0.4531292021,-0.1651648432,-0.1589671075,-0.0396111086,0.6585915089,0.1879720837,0.430300504,0.1421813965,0.5052545667,-0.3357345462,-0.0152163878,-0.3209920526,0.0213597771,-0.0927468762,-0.1665071994,-0.3016424477,-0.117558755,-0.1048950776,0.1544779539,-0.1509697437,0.0550446287,0.4208003283,-0.0675971508,-0.2946312428,-0.5726760626,-0.0037673607,0.0555781424,0.2770498693,-0.227297321,0.2844375968,0.0758500025,-0.204398796,0.3192939758,0.6775817275,0.5208280683,0.2321730703,-0.0327896215,-0.0775188729,0.0703338012,0.1269390732,0.2107738554,0.7450389862,0.2959690392,-0.3133507073,0.2754249871,-0.0380341038,-0.1119686738,0.3472535014,0.0250038225,0.0732772127,-0.0365803838,0.3870863914,-0.1618282646,0.083363995,-0.065473251,0.0979300737,-0.3249251246,-0.210972935,0.1634385586,-0.1725271791,0.3017396927,-0.0809573829,0.0441568866,-0.247383967,0.5227413774,0.5025141835,-0.3186611533,-0.1505271345,-0.0263786688,-0.526768744,0.296295464,-0.3916660547,-0.0813597739,-0.2027486861,0.4482113719,-0.1969201267,0.2801702023,-0.1755737513,0.4149185419,-0.0742016882,-0.0189373605,-0.2368476987,0.076500617,-0.0313404314,-0.1934026778,0.0715759993,-0.2461636811,0.2173758298,0.1577958167,-0.1043893248,0.0316086151,0.1015205383,-0.0161333866,0.3900723755,-0.0753022432,0.5313581228,0.2922849357,-0.0397480577,-0.2839522064,-0.1452298015,-0.4025912583,-0.3109636903,0.3422142565,0.202791512,0.4863946736,-0.2239797562,0.4722755551,-0.1662262529,0.311029315,0.0275125019,-0.0619015433,-0.3194892406,-0.026341686,-0.1740523428,0.0449577309,0.20531407,0.5501668453,0.147492066,0.0528295301,-0.2644664049,-0.4141324162,0.3889919519,-0.3854787052,0.0039214003,0.0736305192,0.2361077517,0.2745179236,-0.2053200006,-0.7636074424,0.0434479825,0.2215884477,-0.1319132149,-0.0081460457,0.0673481226,-0.188516587,-0.2726277709,-0.0148607902,-0.1771641523,0.3695700467,-0.4083259404,0.4173524082,-0.0197966658]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/215","title":"NonMatchingSplitsSizesError when loading blog_authorship_corpus","comments":"Yes I think this might not be documented well enough. Let\u2019s add it to the doc @lhoestq @SBrandeis.\r\nAnd everything on how to control the cache behavior better (removing, overriding, changing the path, etc)","body":"Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.","comment_length":34,"text":"NonMatchingSplitsSizesError when loading blog_authorship_corpus\nGetting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. \r\n\r\n```\r\nraise NonMatchingSplitsSizesError(str(bad_splits))\r\nnlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', \r\nnum_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='train', num_bytes=616473500, num_examples=536323, \r\ndataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', \r\nnum_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), \r\n'recorded': SplitInfo(name='validation', num_bytes=30786661, num_examples=27766, \r\ndataset_name='blog_authorship_corpus')}]\r\n```\r\n\r\nUpon checking it seems like there is a disparity between the information in `datasets\/blog_authorship_corpus\/dataset_infos.json` and what was downloaded. Although I can get away with this by passing `ignore_verifications=True` in `load_dataset`, I'm thinking doing so might give problems later on.\nYes I think this might not be documented well enough. Let\u2019s add it to the doc @lhoestq @SBrandeis.\r\nAnd everything on how to control the cache behavior better (removing, overriding, changing the path, etc)","embeddings":[-0.0478634462,0.2781361639,0.0776575059,0.3915254474,-0.1035541743,0.0364038125,-0.082682997,0.2459606081,-0.083533369,0.0461862758,0.0891772434,0.174503088,0.0406080708,-0.1594099849,0.0179090556,0.0527998805,-0.0644671917,0.1248318329,0.1059798971,0.0428972952,-0.0105633764,0.3961234689,-0.1937782615,-0.040419735,-0.3053342998,-0.2336776257,0.0290222913,0.2334820032,0.0402284861,-0.4769920707,0.246647805,0.0566287637,0.1005317718,0.1670151651,-0.0001217629,-0.0632687733,0.5389484763,-0.1948429793,-0.2518877983,-0.0297897793,-0.3836036026,-0.2797572911,0.0038723538,-0.2842761874,0.0151785146,0.0235366579,0.1792152226,-0.1652392,0.229137063,0.3027590215,0.1213609725,0.109807238,-0.1881100684,0.1678940505,0.1732044816,-0.0276303552,-0.0647589192,-0.1565779895,-0.1117098108,-0.2819305658,-0.2403718829,0.4598938525,-0.2357244492,0.1500597298,0.2640921175,0.0883090869,0.3216771185,-0.0872647092,0.0130944196,0.3902913928,0.5862080455,-0.0468333252,-0.2849666774,-0.4985668957,-0.129673779,0.0286405049,0.5057656169,0.2795498967,-0.1570384949,0.0537012741,-0.4670750201,-0.1777775139,-0.0526111387,-0.00081012,0.3168709576,0.1980743259,0.1256275326,-0.0516351461,0.0307559595,-0.0101765851,0.2684497237,-0.1816899776,-0.1249931604,0.0854497477,-0.2321656793,-0.205163464,-0.0267253444,0.0911648199,0.2567503452,0.2174571157,0.4005258977,0.0058857715,0.2139704376,0.124216035,0.3358573914,0.2084189802,-0.0148705738,0.3932438493,0.1530789882,-0.0024870699,0.1373304725,0.0575754307,0.3515795469,-0.2048659772,-0.0110226339,0.2019051462,0.1597347856,-0.4877766073,-0.3507695198,0.1007281244,-0.3997793198,-0.260506779,-0.0586719476,0.2424853891,-0.3720569313,0.3226313591,-0.2690640688,-0.0335225835,-0.449954778,-0.1279056221,-0.2333000749,0.1798157245,-0.1468560398,0.338239193,0.239137724,-0.079446502,0.3943142891,-0.1676047742,-0.0994360372,-0.1319212317,0.3565537632,-0.1095732972,0.0893370956,0.2085787952,-0.0928035825,0.1552503854,-0.0484711193,-0.09950158,-0.485160321,0.0410177112,-0.3274896443,-0.671756506,0.0978627577,0.1147465035,-0.3720282912,-0.0490013696,0.1598810554,0.3045910299,0.5192116499,-0.3798107803,0.0015793662,-0.087711826,-0.1982912868,-0.1586367488,-0.1705338061,0.5231297612,0.0726187676,-0.3068563044,0.1242673993,0.0117179537,0.4826217294,0.3697668314,-0.0356649272,0.1533680558,-0.164187178,0.3558746576,0.1711984575,-0.4163642526,-0.2768096328,0.380389452,0.1063763201,0.2749008536,0.2496744096,0.1237706095,0.2442066967,-0.1929773241,0.0163563173,0.4298614264,-0.0630663037,0.093382135,-0.5093785524,-0.418787837,0.4363307953,0.1158372611,0.0694449022,0.0362953991,-0.0036744236,0.4742714763,0.1634601802,-0.085600771,0.098512955,0.2544678748,-0.3477673829,0.139681682,0.0839437395,-0.0734169707,-0.4933947623,-0.0299287718,-0.2795148492,0.1781546026,0.1481744349,-0.3201975226,-0.0982877761,-0.4774520695,-0.0937388912,-0.1365209669,0.0294803251,0.2034240365,0.5127832294,0.0822759047,0.019414166,0.5455961227,-0.006967701,0.1613585651,-0.8287149072,0.0638813823,-0.1368913502,-0.1838449389,0.3626117706,0.3595058322,0.2163828164,-0.0509442054,-0.0946413875,0.5340718627,-0.1414176077,0.4370479286,-0.2205443531,0.1521022022,0.2498196065,-0.2066895664,0.1167725325,0.2149262279,-0.1372483075,-0.002253104,-0.0948957726,0.3331183195,-0.374682188,0.086203523,0.0674552694,-0.1066150814,0.057475891,-0.255736351,-0.2727244198,-0.4599678516,0.5414722562,-0.0531041995,-0.2366636992,0.0136255063,-0.119912751,0.1013706997,0.4620293081,0.1043645069,-0.0153187793,-0.0629961863,0.0948043838,-0.2510462701,0.2087735236,0.2167150825,0.4686730206,0.1513731033,0.0737202018,0.0190006085,-0.1884738803,-0.336895138,0.1008825153,-0.0162118748,-0.0242401138,0.2759978473,0.0113301482,-0.0014322649,-0.3033410311,0.0221082624,0.191207841,0.1761494726,-0.3325128555,-0.0619676672,-0.3928717375,-0.3235004842,-0.2495739311,-0.1301067173,-0.2025113106,-0.3588606119,0.3024480045,0.1672642231,-0.2591952085,0.1308683753,-0.2516002059,0.1859038919,-0.276306957,-0.1052640676,-0.073060967,-0.2778658271,-0.2761032283,-0.0637727082,0.116792582,0.3153985739,0.3178684711,-0.415217787,-0.2130653858,0.0406742804,-0.3145892024,0.0671834499,0.059937343,-0.1061162055,0.1001225263,-0.2141468376,0.3740353584,-0.2210167497,-0.0558545925,-0.2211214155,-0.2521099746,-0.0674769878,0.2016217709,0.2455025017,-0.1995059699,-0.5588034987,-0.3258545101,-0.3663263321,-0.0570198968,0.1090074778,0.2995633483,0.0821922049,-0.2107693851,-0.2101608515,-0.1878654957,0.3125407994,-0.3795473874,-0.1632731259,0.3584213555,-0.1916775703,-0.0763096586,0.044109486,-0.0424670316,-0.071176447,0.1271946579,-0.5488827825,-0.1745836586,-0.1155960485,-0.0415999331,-0.2191823721,-0.2206530273,0.639005065,-0.0446474664,0.0478066728,0.0388732255,-0.3613528907,0.0701284856,0.0347477123,0.5672549605,-0.041620668,0.2360640019,0.0665677413,0.5585208535,0.4746896625,0.144145444,-0.0248094015,0.3020043075,0.1700406373,0.0852005705,0.0948510394,0.3581473529,-0.0735527873,-0.2010853291,0.3572814763,0.1408153623,-0.1478683501,-0.0891999751,-0.0461827144,-0.2848126292,-0.2670897841,0.1735617369,-0.3263123333,0.2059570551,0.2735862732,0.1817926168,-0.1691222042,-0.4797645211,0.051178582,0.3250429928,0.2547905147,0.2598886788,-0.4101262391,0.1853186935,-0.5481973886,0.1673724353,0.1233103424,0.2695994675,-0.1897912771,-0.1771589816,0.0442053378,-0.1511922032,0.2935730517,-0.5013008714,-0.0103408704,-0.1180660576,-0.0182660613,-0.1910658777,-0.1754840016,0.0170905571,0.3075019717,0.4353837967,0.3108359575,-0.4779395759,-0.0729023293,0.1180316806,-0.0882898644,-0.1425054371,-0.2267030478,-0.1783085465,-0.2467597276,-0.1737643033,-0.1463427991,-0.0684433281,0.2047618777,-0.0036694754,0.0576347299,0.078076452,0.1945188046,0.2609216273,0.2361555099,0.2463884056,-0.07365641,0.1055103093,0.3441776037,0.3810715675,0.0458005592,0.4809647799,-0.0473776497,-0.0333528407,0.0333514139,-0.158203423,-0.0558537953,0.4643750191,0.1498104185,0.0582266562,0.0461264551,-0.116933845,-0.3775770366,0.0007057927,-0.0550767481,-0.1400307268,-0.2999065816,-0.4009565413,0.2616018653,0.3482020497,-0.1439936459,-0.1402900368,-0.1762956083,-0.3403508067,0.2590385973,0.2319077104,1.0419523716,0.2871278822,0.0786777139,-0.1415207833,-0.1959871948,0.8950446248,-0.458468914,0.3884294033,-0.379758209,-0.1699530184,-0.0497177429,-0.1392178386,0.0671912879,0.0782817602,-0.3491445184,0.4424476624,0.2311291546,0.061702814,-0.1435950398,0.3393722475,-0.2498143762,-0.1875227839,-0.1950545758,-0.0053702421,-0.2349935323,0.3488819003,-0.0095857354,-0.0852255896,-0.0598622039,-0.1827042103,-0.3571867943,0.1146433204,0.0939710289,0.0124900779,0.2369415909,-0.2003677189,-0.1160285249,0.0936949998,0.3660143316,-0.1871988922,-0.1647220254,-0.0503454246,-0.1375662684,-0.0733361468,0.1151698679,0.0633945093,0.3432282209,-0.0829621851,0.0611771233,0.117047444,-0.1985533386,-0.4443511069,-0.2135725915,-0.1307022274,-0.1482189,-0.1180649176,-0.0232007578,0.0821652263,-0.013900388,-0.0531080179,0.0421260856,0.0036888947,-0.1595687568,0.3063155413,-0.2527931929,-0.2312993258,-0.167319268,0.3863882124,0.353952378,-0.0227161329,0.383534044,-0.1494132429,-0.1633327305,-0.1423273683,-0.2202544659,-0.2453572154,-0.3556602299,0.113422066,-0.3243699074,-0.2375132293,0.0965532437,0.1988579333,0.1759380847,0.3981848955,-0.0593548529,-0.4329328239,-0.08512716,0.1618926972,-0.085626334,0.1074577942,-0.0345918126,0.2953036427,-0.2341654003,0.3081848323,-0.2241592705,-0.2144243419,-0.3940682709,0.2727066875,0.0848630667,-0.0385907516,0.0143694989,0.0451145954,-0.2161915302,0.2673090696,-0.1653190404,0.0056592645,-0.0517045334,0.2062431574,0.1315208077,-0.0317963623,-0.1851962507,0.1290425658,0.0924127027,0.0454734191,0.0289719403,0.2819375098,0.040343456,0.4037653804,-0.2569806278,0.0845032111,-0.0194844753,0.2034149468,-0.2717857659,0.3247481287,-0.0864554867,0.3065057993,-0.2035342455,0.0291813016,-0.0207640901,0.076771006,0.1272356957,0.0113389222,0.1532257199,-0.1426361054,0.0279517211,0.2767182589,-0.0762600452,0.5579014421,-0.1184023172,-0.2716967165,0.1353232861,0.0732981786,-0.0243129972,-0.0656945556,0.090873614,0.02758931,-0.0758271143,0.4246926904,0.2287524492,-0.1360167861,-0.0125128543,0.238831982,0.4896000028,-0.2690233886,0.0784405321,0.1427551359,-0.1867322773,0.0578754134,0.3953347206,-0.2336616218,0.1423676461,0.0633207858,0.1130387336,0.4625710547,0.2025783509,0.3750730753,-0.1887370944,-0.1080243662,-0.0569613166,0.6232267022,0.2760687768,0.3305197954,0.1989185661,0.3161476851,-0.1434346586,0.0108347312,-0.3369034231,-0.1892056316,-0.0378127471,-0.1468657851,-0.3757616878,-0.0962624922,-0.0860467479,0.08358071,-0.1365763098,0.1039007008,0.557007432,-0.1709658951,-0.1825269163,-0.5047011375,0.2223153114,0.0032934726,0.2047145814,-0.2249863744,0.3006279469,-0.044192519,-0.20166713,0.2878983915,0.5239020586,0.425033927,0.2935960591,-0.0403794982,-0.179570958,0.0701213479,0.1715540439,0.1945973188,0.7336786985,0.2264087051,-0.3669781983,0.2798478603,-0.0775047019,-0.0481333956,0.4749349058,0.0102369394,0.0326872543,-0.0273014102,0.425028801,0.0105760498,0.118249692,0.1175375506,0.1492505819,-0.2290163636,-0.2858249843,0.1325466782,-0.1783285439,0.2870568335,-0.1423878074,0.0456335694,-0.2089939862,0.6096130013,0.4575624764,-0.2356959432,-0.2216827422,0.001924647,-0.5511800647,0.3374859095,-0.3613300323,-0.200192377,-0.1751606911,0.5507522821,-0.3412739336,0.3287825882,-0.30485636,0.3189489841,-0.03687457,0.0302971099,-0.2732519805,-0.0915131792,-0.0899836347,-0.1491312087,0.0885565355,-0.2684880197,0.2510387599,0.2370918542,-0.1010105014,-0.0206655022,-0.0270448625,0.179448247,0.5358687639,-0.0289871283,0.552588284,0.2186670154,-0.117667906,-0.176179558,-0.2153855562,-0.2438699305,-0.3098506331,0.1533384621,0.1306292862,0.309350729,-0.275657326,0.4983601272,-0.1463240534,0.4165279269,-0.0168304238,0.0083348965,-0.2304276079,-0.0479770489,-0.093550697,0.1230999529,0.1896443516,0.4126790166,0.178463921,0.1410259604,-0.1901106834,-0.3322605789,0.3096322417,-0.359690547,0.0801034868,-0.0354273431,0.2624872029,0.2798449695,-0.2487338036,-0.7318596244,-0.0688436106,0.2349818647,-0.1255519837,0.0618446283,0.045738034,-0.1132706255,-0.2084642947,-0.0927694291,-0.1714313924,0.2879588306,-0.3046326041,0.3468719721,-0.1572413296]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/211","title":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type","comments":"Here the full error trace:\r\n\r\n```\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-1-7aaf3f011358> in <module>\r\n 1 import nlp\r\n 2 ds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\n----> 3 ds.map(lambda x: x, load_from_cache_file=False)\r\n\r\n~\/python_bin\/nlp\/arrow_dataset.py in map(self, function, with_indices, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, arrow_schema, disable_nullable)\r\n 549\r\n 550 if update_data:\r\n--> 551 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n 552\r\n 553 # Create new Dataset from buffer or file\r\n\r\n~\/python_bin\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 182 def finalize(self, close_stream=True):\r\n 183 if self.pa_writer is not None:\r\n--> 184 self.write_on_file()\r\n 185 self.pa_writer.close()\r\n 186 if close_stream:\r\n\r\n~\/python_bin\/nlp\/arrow_writer.py in write_on_file(self)\r\n 104 \"\"\"\r\n 105 if self.current_rows:\r\n--> 106 pa_array = pa.array(self.current_rows, type=self._type)\r\n 107 first_example = pa.array(self.current_rows[0:1], type=self._type)[0]\r\n 108 # Sanity check\r\n\r\n~\/hugging_face\/venv_3.7\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/hugging_face\/venv_3.7\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\n~\/hugging_face\/venv_3.7\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Could not convert TagMe with type str: converting to null type\r\n```","body":"Running the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.","comment_length":155,"text":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type\nRunning the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.\nHere the full error trace:\r\n\r\n```\r\nArrowInvalid Traceback (most recent call last)\r\n<ipython-input-1-7aaf3f011358> in <module>\r\n 1 import nlp\r\n 2 ds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\n----> 3 ds.map(lambda x: x, load_from_cache_file=False)\r\n\r\n~\/python_bin\/nlp\/arrow_dataset.py in map(self, function, with_indices, batched, batch_size, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, arrow_schema, disable_nullable)\r\n 549\r\n 550 if update_data:\r\n--> 551 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file\r\n 552\r\n 553 # Create new Dataset from buffer or file\r\n\r\n~\/python_bin\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 182 def finalize(self, close_stream=True):\r\n 183 if self.pa_writer is not None:\r\n--> 184 self.write_on_file()\r\n 185 self.pa_writer.close()\r\n 186 if close_stream:\r\n\r\n~\/python_bin\/nlp\/arrow_writer.py in write_on_file(self)\r\n 104 \"\"\"\r\n 105 if self.current_rows:\r\n--> 106 pa_array = pa.array(self.current_rows, type=self._type)\r\n 107 first_example = pa.array(self.current_rows[0:1], type=self._type)[0]\r\n 108 # Sanity check\r\n\r\n~\/hugging_face\/venv_3.7\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib.array()\r\n\r\n~\/hugging_face\/venv_3.7\/lib\/python3.7\/site-packages\/pyarrow\/array.pxi in pyarrow.lib._sequence_to_array()\r\n\r\n~\/hugging_face\/venv_3.7\/lib\/python3.7\/site-packages\/pyarrow\/error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Could not convert TagMe with type str: converting to null type\r\n```","embeddings":[0.0478234887,-0.0149945924,0.0999606401,0.4827658832,0.2526236176,0.118270196,0.1242586449,0.462997973,0.20459342,-0.1143017784,0.1954923421,0.5628960729,-0.1122900546,-0.214510262,-0.068543978,-0.18655321,0.1165156439,0.2422813624,-0.0039511644,-0.0194521397,-0.1214477792,0.124361366,-0.2644390762,0.1609538198,-0.450982064,-0.1343633235,0.1006035954,0.0705768988,-0.1665291041,-0.4529118836,0.13311553,-0.386316359,-0.0144939888,0.2676903605,-0.0001202561,-0.1023915559,0.3526799679,0.0484397747,-0.261015296,-0.2479179204,-0.4819888473,0.0986581743,0.3085628152,-0.0510609858,0.2251927555,0.0243791398,0.0238793492,-0.5974248052,0.1822833419,0.5890497565,0.1811947525,0.3117505908,-0.0586166196,0.0824080631,0.2270174176,0.0864315182,-0.3474987447,0.1586010456,-0.0418649204,-0.1637495905,-0.0972599834,0.4248969555,-0.1737870723,0.1287671775,0.6042653322,0.1013759524,0.1031220332,-0.3009963036,0.1894645393,0.0022626503,0.2738990486,-0.1859186739,-0.0806752667,0.0309147146,-0.1441309005,-0.3694722056,0.2096119672,0.1765302718,-0.3138512671,0.0176620092,0.0113149639,0.0840630755,-0.0632319376,-0.0736751035,-0.0597238019,0.067420505,0.0130822873,0.1142640635,-0.0543162897,-0.1860047132,-0.189728424,-0.2200310826,-0.2048834711,0.2414929867,-0.3378819525,0.048407536,-0.0579255559,-0.2089133859,-0.1648839414,-0.1189332604,0.2784720659,-0.0002529972,0.1446985453,0.1296414286,0.3333248198,0.2861360013,0.3204212785,0.2487848103,0.0986191705,-0.3073056638,-0.1854353249,-0.0136373509,-0.0114122313,-0.0097023286,0.5141096115,0.0163877103,0.3002519608,-0.3601064086,-0.3549937606,-0.0110491002,-0.7358067632,0.1599646956,-0.169023782,0.1066551134,-0.0082527203,0.4835471511,0.2476842701,0.2814028859,-0.3391996026,-0.1926279962,-0.1142075062,-0.0042975568,-0.2985197008,0.0156220309,0.3052985668,-0.0363648497,0.0539230071,0.0657088831,-0.4077064991,-0.2089091837,0.3739091754,-0.4226520061,0.1788685918,-0.0087085897,-0.3287378848,0.415250808,0.2411908805,-0.4089022577,-0.2129134387,0.1715905517,-0.13049604,0.0368987098,-0.1928086728,0.1247835532,0.1592132598,-0.2186264098,-0.1702641696,-0.0653255433,0.3925433457,-0.1083274633,0.1037179083,-0.2929559052,-0.0403226614,-0.2410746217,-0.0862942711,0.4788666368,-0.2231424302,0.0666108057,-0.0555775538,0.1835543215,0.3641396165,0.3781097531,-0.0203731935,0.1771148741,0.1071878076,0.2820343673,0.4739594162,-0.3007650971,-0.1000544131,0.0992059857,-0.0288218465,-0.0070865471,-0.0418070704,0.1214025915,0.0770688877,-0.1202662885,0.1674464494,0.1947706938,-0.0623199306,0.2353259474,-0.420689404,-0.1015318036,0.5093363523,-0.1742017865,-0.0756611675,0.0260948651,-0.0768755227,0.0502024256,-0.0676269755,-0.0130599225,0.4281769693,0.260076344,-0.0904399976,-0.0825848132,0.0832560062,-0.1882934272,-0.3873153925,-0.0520174839,-0.1053638533,0.133559376,-0.1724085957,-0.0403863043,-0.3455444574,0.0745703951,-0.0271313544,-0.1989130229,0.125344798,0.1082480922,0.1589182913,0.058027789,-0.2278685123,0.1251087636,0.0048239664,0.1273974925,-0.3979255557,-0.1481667906,-0.0802366436,-0.3554536402,-0.0991665199,0.1799599826,0.2376293093,0.167554155,-0.2698638439,0.1873165816,0.0186801143,0.2171432972,-0.254982084,0.0299019367,0.0187678691,-0.4021503031,0.2202013135,0.3269737065,0.361649245,-0.2435902208,-0.3964939117,-0.0214629062,0.0146005685,0.5858560205,-0.1325559169,0.108533524,-0.0767762065,-0.0362331606,0.09410505,-0.0797639191,0.0960107297,-0.1133068502,0.197952792,0.2924374342,-0.3097238243,0.2726490796,0.4216948152,0.3002494872,0.1342764795,0.1598110348,-0.3345211446,-0.0856646225,0.264416337,-0.0123851011,0.4610905945,-0.043286074,0.2145610303,-0.084927313,-0.1507028341,-0.1485962123,0.357840091,-0.0966183916,0.017265385,-0.0092660841,0.4907109439,0.0106255123,-0.1787142158,0.2008915693,0.0055136667,0.2383064181,-0.3552833498,0.3290377855,-0.3391131163,-0.1678956449,-0.3810411394,-0.2544403672,-0.0185093954,-0.5028193593,-0.1405358016,0.0792334303,0.0061113108,0.1203892753,-0.2541172802,-0.1418566704,0.0180851668,-0.1107123569,-0.3347356617,-0.3540616333,-0.3705020845,-0.0632876456,0.4646281004,0.3231955767,0.1873808503,0.053034436,-0.0629505813,-0.0690104067,-0.3603878021,0.0305633377,-0.1989224106,0.0571852922,0.1203932911,0.0761394128,-0.136030376,-0.1272963732,0.4256206751,0.0031831567,-0.2406023741,0.3291910887,-0.0207711682,0.0490220152,-0.2927949727,-0.4807073772,0.0135379611,-0.4682863951,0.1370299757,-0.1121180058,0.2093194723,0.1776323467,0.1237757653,0.3106841445,-0.0557440855,0.1851253062,0.2070276141,0.3653668463,0.6162765026,-0.1331360489,-0.2379468679,0.2084654719,-0.0769689456,0.3138104975,0.2074887305,-0.1788210571,0.3327707052,-0.4218785763,0.3207609653,-0.163307339,-0.1305053681,0.2217357457,0.2210210115,-0.0480525903,-0.0604383834,-0.0075168535,0.2183129489,-0.217765525,0.1517935246,0.1262533069,0.2142089605,-0.0534500591,0.3817186058,0.1439975351,-0.2284086347,0.283031106,0.0168088023,-0.047595825,-0.3022206724,-0.2784267068,-0.0881357342,-0.2115705013,0.1277086437,-0.1773896068,0.0339444876,-0.4855810106,-0.0222650152,0.2889747918,-0.1829325408,-0.1343129575,0.3022049963,-0.2138018757,0.2901404798,0.1346093565,0.1258190572,-0.1979225725,-0.4124386907,0.032936126,-0.0246567875,0.1568616033,0.0247551557,-0.1019372344,-0.0528978221,-0.4739771783,0.2884022295,0.1751558334,0.4326758683,-0.0764364004,-0.1432521045,0.1917068511,-0.1135197803,0.298640728,-0.1461240351,-0.0275959224,0.3758094013,0.2723764479,-0.4777469039,0.2693087459,-0.2311157882,0.3493030071,0.2616674602,0.2834546268,-0.389685154,-0.3131132126,0.2579025626,0.1610282958,0.0166522432,0.0899731293,-0.3169793487,-0.0035179665,-0.1555343419,-0.0867393911,0.2757054567,0.3106018901,0.0826226175,0.0549024865,-0.3301869929,0.1789491624,0.1806426793,0.0263724253,0.4282669425,-0.3036712706,0.1376903057,-0.1568315029,0.2089803964,-0.0867718309,0.3240396678,-0.0634894818,-0.1984016001,0.2668813765,-0.0575366803,0.3644377589,0.329267025,0.0360931307,0.347053349,-0.122374855,-0.0959219262,-0.2033011466,0.0296046231,0.1675886661,-0.0528092533,0.0434266701,-0.2120433897,0.4822666645,0.2981485724,-0.0335702822,0.6272780299,0.1370319128,-0.2833762169,0.237400189,0.1865324974,1.1821588278,0.1512260884,0.1496685147,0.1337822229,-0.4301369488,0.333907336,0.0619379841,0.3254597485,-0.462836504,0.3249582052,0.0244982224,-0.1689322144,0.067314513,-0.0338957459,-0.3024041355,0.0830444619,-0.1960292161,0.3092464805,0.1123863235,0.0580706596,-0.302675277,-0.3165791035,-0.3749186993,0.0806003064,-0.2505501509,-0.0766257793,0.11036098,-0.2155663818,-0.4914936423,-0.0455054305,-0.0685388446,0.1147970706,-0.1706724316,0.0465223342,0.2367302179,-0.157687366,0.1811179966,0.0725658089,-0.1037082598,-0.0658696666,-0.1338339299,-0.0860887021,0.1967154294,0.1009659767,0.1206195876,-0.0757202208,0.0112010352,0.0762333795,0.024822183,0.2276172638,-0.3680085838,-0.0135413334,-0.0647316873,0.1664160639,0.1192658022,-0.0927513316,-0.0884645954,0.099075675,-0.0605181307,0.0031136074,0.0932031348,0.0256682355,0.1516789049,0.0805700347,-0.1569107622,-0.2600081265,0.2690163255,0.3413309157,0.0477588065,0.2406034619,0.716996789,0.0595088191,-0.1855409592,-0.1779445112,0.077611886,-0.052394066,-0.7310900092,0.3293785751,-0.3314181566,-0.2210787535,0.0340336673,0.4544684589,-0.2029377073,0.0803395808,-0.1594133079,-0.2330781072,-0.2189446241,0.1790490448,-0.0306003522,0.0346716568,-0.4524493515,-0.057527937,0.0966438055,0.1214938611,-0.2271929383,-0.0100416457,-0.1742923111,0.3713478744,0.0921324715,0.0329645686,-0.0507325456,-0.1805099696,-0.0114440229,0.5151129365,-0.0971550196,-0.1443873197,-0.2621200979,0.1471551657,0.0711333752,-0.1661830246,0.0705537945,0.0241728742,0.0784357637,-0.212869063,-0.151078999,-0.1028865576,-0.1862980872,0.0560390167,-0.087701112,0.2029781193,-0.1787182689,0.0992549509,-0.1471330374,0.0568210967,-0.2434588969,0.107935518,0.2107048184,0.0571228787,-0.2081965357,0.0487785973,-0.0946377218,-0.1588643044,0.3047349453,-0.377263695,-0.1661876589,0.0565505996,-0.1906723678,0.4166314304,-0.0920560658,0.0020528038,-0.0814980417,0.0977553278,0.0703344047,0.3166258335,-0.1786950082,-0.1631126702,-0.1134640351,0.1911716908,-0.0239853933,-0.1549674571,-0.030291561,0.0948891714,0.033046551,-0.4421164691,-0.1640719771,0.1360544264,-0.0254782438,-0.0716066957,0.3569366038,-0.1045392007,0.0215371978,0.0203028284,0.3664782345,0.3340464532,0.4364641011,0.2034334391,-0.0676420555,0.1195063591,-0.0249721054,0.3891993463,0.2442485839,0.1511014104,0.1753364205,0.5244451165,-0.2980790734,-0.3281244636,0.06573385,0.1083270982,-0.2727236748,-0.0876113474,-0.5703164339,-0.1152269617,-0.4419082701,-0.088098608,-0.1905383468,0.2544889152,0.6928649545,0.2337905616,0.0166978724,-0.5845615864,-0.1298844069,0.1329542398,0.3920841217,-0.2488602698,0.4035942554,0.108685039,-0.0760025606,-0.1305522323,0.4997346401,0.4683460295,0.5678134561,-0.2883790433,-0.0676521212,0.0580816679,-0.1825811714,0.2377776206,0.5766586065,0.3382422924,0.1401809156,0.0350437462,-0.0379658528,-0.1073758975,-0.0900807232,-0.0297216605,0.277446419,-0.1637601256,0.2949459553,-0.2384104729,-0.2226386517,-0.0462517515,-0.2109304667,-0.3650142848,-0.0462197773,0.2055170387,0.0495744683,0.0156688001,-0.3341812193,0.0574212298,0.2021238953,0.5171064734,0.2711421847,0.1386129707,-0.4753841758,-0.1169540212,-0.7315415144,0.1763935387,-0.3148992658,-0.1567145735,-0.0145184388,0.363278538,0.0452327728,0.3146993518,-0.0349213369,0.2289497554,0.2513009906,0.3053611815,-0.341689378,-0.1197941452,0.353043735,-0.3037295043,0.023086071,-0.5888962746,0.2686021328,0.0512421951,0.0651120543,-0.0086924313,-0.2844811082,0.2985390723,0.4892457128,0.2755067348,0.2637933493,0.2856043577,0.2475817353,-0.5557769537,-0.2176061422,-0.1734557748,-0.2077278942,-0.0203035604,0.0898685902,0.5805657506,-0.140447855,0.0624375902,-0.3035693467,0.3277708292,-0.2265174985,0.0343786478,-0.1755734086,0.067006357,-0.3490558267,0.1630246043,0.0721893013,0.1204009727,0.1978320926,0.2425436825,-0.5347838402,-0.3874289393,0.3043855429,-0.5194076896,-0.4466890097,0.2931914032,0.1429883093,0.4468529224,-0.0124158626,-0.7071861029,-0.2688814998,0.2132193446,-0.0964377522,0.0253313426,0.0230048187,0.0265947171,0.0740291253,-0.1916030794,0.3619840741,0.0083017824,0.0492807887,0.0483300611,-0.2284108847]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/211","title":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type","comments":"Actually thinking a bit more about it, it's probably a data sample that is not correct in `trivia_qa`. But I'm a bit surprised though that we managed to write it in .arrow format and now cannot write it anymore after an \"identity\" mapping.","body":"Running the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.","comment_length":43,"text":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type\nRunning the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.\nActually thinking a bit more about it, it's probably a data sample that is not correct in `trivia_qa`. But I'm a bit surprised though that we managed to write it in .arrow format and now cannot write it anymore after an \"identity\" mapping.","embeddings":[0.1208235994,0.0813008696,0.088923201,0.4082977474,0.2814412713,0.0546012893,0.1137645766,0.3365483284,0.18600972,-0.186949715,0.2484580576,0.5634832382,0.0009860162,-0.0783177614,-0.0235590022,-0.1181143671,0.0730633289,0.2508925498,0.0506293997,-0.1244531348,-0.0924610645,0.1616433859,-0.1819394678,0.2225300521,-0.5130993128,-0.158728376,0.1441280693,0.080267176,-0.1196371838,-0.4990111291,0.0155159179,-0.3473426104,0.0308181904,0.2083607018,-0.0001191395,-0.165944159,0.3734931946,-0.0522149019,-0.1960327476,-0.2107961476,-0.5531388521,0.1664822996,0.3032821119,-0.0672414526,0.2399285287,-0.1021348611,0.0658057779,-0.5268741846,0.0961819738,0.4953772724,0.1949684322,0.1500273943,0.1189303175,0.0855403394,0.2670636475,0.1191722527,-0.2874541581,0.0432622246,-0.13040483,-0.2241957039,-0.0693887472,0.3475057781,-0.0491954423,-0.0045447596,0.5662901402,0.0445675813,0.1975944936,-0.1864271164,0.2564823329,0.0585066713,0.3549563587,-0.0617081113,-0.1086405069,0.0988542885,-0.1233241186,-0.3633818626,0.2884753346,0.2225210816,-0.2659069002,-0.0122918878,0.1163758263,0.0631242916,0.0097148437,-0.1246361732,-0.1565378457,0.1494541615,0.0186119024,0.11266765,-0.050113596,-0.197682336,-0.1804719716,-0.0793853924,-0.3031898737,0.2140202522,-0.2830839157,0.0116387224,-0.0516032428,-0.1638733,-0.2018789798,-0.0879100785,0.4137466848,-0.0001443101,-0.005353089,0.067548573,0.3144080937,0.4288606346,0.4697208703,0.0982727483,0.1366083026,-0.3048892617,-0.1826519817,-0.0344758928,-0.0252424069,-0.0097393803,0.4327159822,0.0240835566,0.0781966895,-0.3047506213,-0.2590415776,0.0268414505,-0.7115644813,0.2219687253,-0.216433391,0.1374626607,0.0648862645,0.4073728621,0.2952431142,0.2960962057,-0.2042858005,-0.2012081146,-0.1123460829,-0.0987347513,-0.3487420678,-0.1523423791,0.192709595,-0.0855007321,0.0856177956,0.0216545276,-0.5190793276,-0.2234737873,0.2967338264,-0.2855610549,0.2961819768,0.0604484342,-0.2745224535,0.4591630697,0.2122778594,-0.5164351463,-0.0973500162,0.0927968323,-0.2046000361,0.0706736818,-0.1819467545,0.1397913992,0.0373096131,-0.2744913399,-0.2020525485,-0.0635189414,0.3884651959,-0.1685949862,0.0939295813,-0.2046537399,-0.0126877688,-0.2841027379,-0.1489118785,0.3834084868,-0.1865783185,0.1084365919,-0.0751806051,0.1669818163,0.3044075966,0.3455912769,0.0228832755,0.1997970492,0.1055131555,0.3500428796,0.3927366734,-0.2295797467,-0.0097290073,-0.0022533732,-0.0176022891,-0.0076589468,-0.0582821034,-0.0028022679,0.042020902,-0.1174990311,0.0826956704,0.1232978255,0.0120985182,0.2591465414,-0.3936404586,-0.0621934682,0.4800761938,-0.3127318025,-0.2060414106,-0.0303945281,0.0164995361,0.1014273167,-0.0264296494,-0.0937464014,0.3080484867,0.1747931391,-0.0706622899,-0.1480063796,0.1571844518,-0.0412917696,-0.3996143341,-0.0596142188,-0.1100006402,0.1341490746,-0.0511530638,0.0029529678,-0.2747970819,0.0522805527,-0.0077958298,-0.2527814507,0.1424805522,0.0631686226,0.1646498889,0.1263308078,-0.2156060785,0.0616247617,0.0216923766,0.1292787641,-0.3672278523,-0.2625304163,0.0105466442,-0.3314329088,-0.1912827492,0.2312710881,0.2290376127,0.1343213916,-0.2543695867,0.1925375313,-0.0094262837,0.392041862,-0.2142837346,0.0745102167,-0.013634935,-0.5046822429,0.1931379139,0.4102119505,0.2995217443,-0.2957393825,-0.4034367502,-0.0535171553,-0.0316734761,0.5904880166,-0.093098402,0.2558086216,-0.0202942807,-0.06831415,0.0701616481,-0.1256335527,0.1901365817,-0.1690690517,0.1784565151,0.2348818928,-0.2307665944,0.3158989549,0.3804571629,0.327327013,0.2145264894,0.0387076735,-0.3840380907,-0.0601568036,0.2709558904,0.0669727698,0.3688614666,-0.0591402985,0.2242632508,-0.1338997483,-0.1544368416,-0.1763559282,0.3210907876,-0.2017697096,0.0710040629,0.0661776364,0.5125981569,0.017989248,-0.1894221157,0.3008747995,-0.0166824255,0.2317208946,-0.417604059,0.2389861047,-0.311637044,-0.0587846339,-0.3408745229,-0.2117813826,-0.0825011507,-0.5657789707,-0.0303330626,-0.0334508978,-0.0700718388,0.0360913277,-0.1194025055,-0.147991851,-0.1088990644,-0.0712956339,-0.3734095395,-0.4300117791,-0.2597958744,-0.0475394838,0.3729808033,0.2394888401,0.1047431976,-0.0126247574,-0.0753576159,0.0003270398,-0.4714281857,-0.004124477,-0.2229709774,0.067943275,0.1828233004,0.0210705195,-0.0303823911,-0.0144363465,0.3603616059,-0.0667396411,-0.2764347196,0.3501122594,0.0663592964,0.0544083752,-0.2934792936,-0.311521709,0.082166262,-0.4667845368,0.3006452918,-0.0964783654,0.2416380048,0.2114753276,-0.0179771408,0.2288765311,-0.0653321818,0.1617138833,0.1883900911,0.4239374697,0.540738225,-0.2293752283,-0.3174421489,0.2958124578,-0.0376458392,0.3881962299,0.1468117535,-0.0484379753,0.4057910144,-0.3843774498,0.1319413185,-0.1474188864,-0.0709135756,0.1809769124,0.2352913171,-0.0599621199,-0.1455631405,0.0331562944,0.2185854912,-0.0307559539,0.24850595,0.0911097676,0.0045341966,-0.1135431379,0.382022202,0.3040758669,-0.1570863426,0.2629907727,0.0060901893,-0.0635355636,-0.2300884128,-0.2707279921,-0.1288834065,-0.1735387743,0.1558312476,-0.1517203301,0.0335833505,-0.3819098175,-0.0314342193,0.3377803862,-0.1851005107,-0.1455885768,0.2932660282,-0.2784805,0.2370390743,0.0854838416,0.1900607198,-0.1247148588,-0.4073356092,0.1250004321,-0.0006940081,0.117848888,0.0363282487,-0.198661834,-0.1296972036,-0.4102330506,0.3197306991,0.0399338789,0.3601866364,-0.1612703502,-0.1563735753,0.1840424687,-0.1521959007,0.4350283742,-0.3664804995,-0.0590631478,0.3051371872,0.3421055675,-0.4739362299,0.3419201374,-0.2217651904,0.2468216717,0.1647553444,0.1495350301,-0.3234816492,-0.23489438,0.1747420281,0.0201658644,0.0237352811,0.1704344153,-0.3350857794,0.0266625248,-0.0769751295,-0.1739079952,0.2077367753,0.242021203,0.0186902843,0.0758925378,-0.4179531932,0.1937550604,0.2174450606,0.0871915594,0.3997777998,-0.2400731295,0.011526579,-0.0202289727,0.1994380802,-0.0704538971,0.4529924989,-0.0416535437,-0.1785088927,0.2557938397,-0.0836934671,0.4723671973,0.383490324,0.0856897533,0.3246254623,-0.2723218203,-0.2283095717,-0.3209863007,0.0069494122,0.0411790572,0.0026778749,0.143037051,-0.1038238481,0.5234208107,0.2486487627,-0.0763490275,0.5133995414,0.1670313776,-0.3357836902,0.3231590986,0.2279377729,1.1573792696,0.0584904961,0.1037837863,-0.0582524352,-0.4672042429,0.4406113327,0.0648869798,0.2514755726,-0.471750617,0.3540160358,0.0048875846,-0.2005732805,0.0443479382,-0.0519426949,-0.3008950949,0.1111357659,-0.2160280049,0.2448741943,0.096672751,0.097720623,-0.3100964129,-0.3052620292,-0.2676104307,0.0932707861,-0.1876251251,-0.007991448,0.117065385,-0.2904045284,-0.5326505303,-0.0551836751,-0.0885736868,0.1917649806,-0.1987987906,-0.0728590861,0.2525029778,-0.2095432729,0.2399552464,0.2320885956,0.0881659538,-0.0634798631,-0.1739142388,-0.0472855568,0.2920392752,0.1568921506,0.1202145889,-0.0331292599,0.0141852424,0.1007955372,0.0456057154,0.2517969608,-0.4120059311,0.0407633521,-0.0872244388,0.2069533169,0.0925787985,-0.0561166108,0.0583340079,0.0569283701,-0.0576385632,-0.0249361936,0.1084731743,0.0131436875,0.1690440625,0.0357055739,-0.1153753921,-0.2564199269,0.3108493388,0.3111067116,-0.011716038,0.2638026774,0.7341854572,0.0279941261,-0.1527635008,-0.198307544,0.0286303516,-0.0241596904,-0.7314620614,0.3125097752,-0.3160667419,-0.367892772,0.0200802684,0.5516511202,-0.2353504449,0.1846442372,-0.1853222847,-0.2600318193,-0.1918115765,0.2942543924,-0.0680490956,0.0068127387,-0.4693598449,-0.1069454029,0.0295032784,0.1694559604,-0.2477701008,0.0353245065,-0.1177467331,0.362618953,-0.0466864407,-0.0051180809,-0.0046822699,-0.2574215531,0.0100516593,0.5666003227,-0.1207700446,-0.1419090033,-0.283336848,0.145878002,0.0844229534,-0.1702077985,0.165379256,-0.0178168248,0.0622471012,-0.2151459008,-0.0726678595,-0.208088994,-0.1844862849,0.074850522,-0.0211002324,0.2446193099,-0.234215349,0.0805011317,-0.0700895712,0.0681645647,-0.3697081506,0.0071487743,-0.0146928336,0.0024232934,-0.0353164859,0.1522941887,-0.1864635199,-0.1116990894,0.28232041,-0.295342654,-0.1291115284,-0.0050050672,-0.1942414343,0.337996006,-0.1524696499,0.0480980836,-0.0423083678,0.1149393395,0.1366561204,0.35715729,-0.2620997429,0.0211083367,-0.0831819922,0.2757547498,0.1245065555,-0.2282408178,-0.1093333289,0.0660745502,-0.0643682852,-0.3699744344,-0.1312263161,0.1212939918,-0.1415522248,-0.0027171001,0.3157488108,-0.1595619619,-0.0392885618,-0.00207247,0.3984380662,0.310343802,0.5081047416,0.2068912834,0.0034981461,0.095891051,0.0087020965,0.2419118136,0.3457305431,0.3615862429,0.2970237434,0.6127424836,-0.3836428225,-0.2681466937,-0.0260095764,0.1027920991,-0.2622461021,-0.1417000443,-0.5835602283,-0.1210505962,-0.4420409501,-0.0805511475,-0.2170771807,0.197097823,0.5802696347,0.2196568102,-0.0454751179,-0.6290245652,-0.2098427415,0.0581791587,0.501736939,-0.2239837348,0.4559963346,0.0160691701,-0.0568241738,-0.0182563532,0.5782521963,0.4393102229,0.5464799404,-0.3213458657,-0.013242715,0.0590823293,-0.1702314764,0.2828425765,0.6259995699,0.3719356954,0.0720670298,0.0497732237,-0.0286371019,-0.0999384671,-0.2609732151,0.0218301211,0.4561325312,-0.0094540026,0.2189076692,-0.2967534363,-0.1307037622,-0.07994169,-0.2456392795,-0.3012332022,-0.077403605,0.1358343363,0.0164384861,0.0205107126,-0.3425386846,0.0712622255,0.1603754759,0.5518670678,0.2615605891,0.1266719848,-0.4171098471,-0.1187971905,-0.5855775476,0.2008875161,-0.2385324538,-0.1507570297,-0.099528037,0.3261132836,0.1053712964,0.2445042878,-0.0276821256,0.203497991,0.3321014643,0.3179222941,-0.3015890121,-0.062345732,0.3174630702,-0.327367574,0.0084630437,-0.5750220418,0.2966242135,0.0678647012,0.0684059262,0.1313178092,-0.3014710248,0.201868102,0.443870604,0.3714677691,0.2160167396,0.2539949715,0.2564385533,-0.4523240328,-0.2691609859,-0.2093788981,-0.238634482,0.1015376672,0.105218038,0.5344880819,-0.0808610693,0.1265619695,-0.2559090853,0.1533973366,-0.1773639768,0.1092867404,-0.2637517452,0.1466666311,-0.3174301684,0.0809876323,0.1202270836,0.1227272823,0.1513158679,0.1441603154,-0.4850179553,-0.4537804723,0.3317622244,-0.3799088597,-0.3346242309,0.3907825053,0.2151694,0.4396257401,-0.1042736247,-0.7881219387,-0.3570627272,0.1472139955,-0.139195323,-0.0508999974,-0.1161145121,-0.0566447675,-0.0052359211,-0.2111218572,0.3439011574,-0.0087314341,0.0837295353,0.0902874023,-0.1682761908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/211","title":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type","comments":"I don't have this error :x","body":"Running the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.","comment_length":6,"text":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type\nRunning the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.\nI don't have this error :x","embeddings":[0.0502192043,0.0254255403,0.101905182,0.52077806,0.3000376523,0.142213881,0.0560313016,0.4376222789,0.1972087175,-0.1212866828,0.2799324989,0.5526871681,-0.0702325851,-0.0881084055,-0.0679403916,-0.1755232811,0.1827031374,0.3164204359,0.0647871569,-0.0959385633,-0.1484212726,0.1534421295,-0.2729373574,0.1333859265,-0.4993882477,-0.1379195601,0.0714440569,0.1020549089,-0.1596959084,-0.409468472,0.0250396393,-0.3906134069,0.0772640482,0.318141371,-0.0001207889,-0.1483429819,0.3652456105,-0.022930121,-0.1752515137,-0.2438963801,-0.4935040772,0.0842962489,0.3324103355,-0.0262332354,0.2056907415,-0.1007987037,0.0022397523,-0.5988404155,0.1290585101,0.5737779737,0.1870961189,0.2394925654,-0.0132610034,0.0673661157,0.2730291486,0.0232043173,-0.3245367408,0.0233154055,-0.0332042985,-0.1740421504,-0.1096464768,0.3790385425,-0.1420375407,0.08526057,0.5985029936,0.0786195844,0.0622327365,-0.3152057827,0.2437920868,0.0691324547,0.4283641577,-0.1314541996,-0.0641643777,0.0499718562,-0.1139251739,-0.3613600135,0.2396758944,0.1860308647,-0.2563553452,0.0046818554,0.086418286,0.0437704325,-0.0303908903,-0.1217240095,-0.1165368035,0.1466338933,0.0111818705,0.0763885528,-0.071627371,-0.1874661297,-0.1046776026,-0.1955985129,-0.2411618233,0.2476038337,-0.3903852403,0.0465269685,-0.0835521668,-0.1091255471,-0.1685089767,-0.0955530107,0.2932846546,0.0139787141,0.0581211932,0.116863586,0.3874284625,0.3571122289,0.3653279543,0.1759212315,0.1401580423,-0.3458825946,-0.1937390119,-0.0369714275,-0.0060547693,-0.0715083331,0.4294492006,0.008488236,0.2238699496,-0.4272092879,-0.2747100592,-0.0150915636,-0.703797698,0.1927813888,-0.2282403409,0.145863235,0.0138694271,0.5156758428,0.2087256312,0.3592909873,-0.2759380341,-0.3043216467,-0.1446026266,-0.0302838851,-0.291159004,-0.0876049325,0.2124798298,-0.0397277847,0.0823528543,0.0349894613,-0.4623999894,-0.2475677431,0.2841292918,-0.3519536555,0.0994287953,0.0518670753,-0.2342057377,0.4207911491,0.2441084236,-0.4425330758,-0.1611364037,0.1406366676,-0.1844079047,0.0579288714,-0.1555852443,0.1232223958,0.06607721,-0.2522572279,-0.2161098868,-0.011176859,0.3552473485,-0.1290381253,0.06964802,-0.1802175641,-0.0736756623,-0.2789751291,-0.1007881314,0.4582741559,-0.2476515174,0.1171355844,-0.066092968,0.1461410373,0.342115134,0.2998822927,0.0206082854,0.1780554801,0.1346569806,0.342662096,0.4938897789,-0.2467238605,-0.0151300095,0.1301781833,-0.0191835035,-0.0450150855,-0.0194464605,0.0622619167,0.0346812271,-0.1224136204,0.1227941364,0.1675245464,-0.1098506525,0.2807259262,-0.4075272381,-0.0678772628,0.5993596315,-0.1878142655,-0.0837872997,0.0558384694,-0.020694606,0.0668349192,-0.0475385115,-0.0413598306,0.359100312,0.2864187658,-0.1268875897,-0.1312757134,0.0911588296,-0.1287420243,-0.4161080122,-0.0286726784,-0.0975540206,0.1213782802,-0.0094413189,-0.0001418326,-0.3612881005,0.0536301918,-0.0236291401,-0.2205624729,0.1308395416,0.0399238728,0.1299194247,0.1158692613,-0.2358194292,0.1258275062,0.0035593081,0.1397292912,-0.366400212,-0.1438017637,-0.0321220122,-0.3450489342,-0.0665233284,0.2336288095,0.3051717877,0.1464484334,-0.287078023,0.209471792,0.0206421185,0.3665635288,-0.2323050797,0.0417645387,0.0148661891,-0.4306673408,0.1914727241,0.3704781234,0.384057492,-0.2884925902,-0.3641670048,-0.0709718838,0.0427990481,0.6422808766,-0.1538197398,0.1818000227,-0.0549572185,-0.0708619207,0.1658537388,-0.072657764,0.1694231033,-0.1253948808,0.1700497121,0.2819743156,-0.2642347515,0.2751502991,0.3841131926,0.3088341355,0.1412779093,0.1335758716,-0.3551524282,-0.1182715297,0.3403661847,0.0364021398,0.4312329888,-0.035205204,0.2232781202,-0.0600684807,-0.1371300071,-0.1613727659,0.314581871,-0.1120489687,0.122614108,0.0388058499,0.4821978211,-0.0061702398,-0.1873642057,0.3051170707,0.0131628467,0.2754454315,-0.3962234855,0.2671076059,-0.3624043167,-0.1510012895,-0.3544068336,-0.1458511353,-0.0750991479,-0.574677527,-0.0784549415,-0.025946537,0.0102108847,0.0673998147,-0.2375373244,-0.136620447,-0.0876333788,-0.0593252219,-0.3463140726,-0.3964079618,-0.3674389422,-0.0677424148,0.4319347441,0.347234875,0.134986788,0.0160198305,-0.0239401385,-0.0382289328,-0.3891150653,0.047989618,-0.2328911424,0.0878112689,0.1226633787,0.0066549582,0.0000543864,-0.0209690239,0.3573572338,-0.0322766677,-0.2750020027,0.3728255033,0.033071097,0.0721934587,-0.2505608201,-0.4465438128,0.0520018451,-0.5248692036,0.1135949343,-0.095776625,0.1903294623,0.0852050856,0.1215549484,0.2903985381,0.0198360849,0.2046639919,0.2317455262,0.4132091999,0.5925531983,-0.1675267965,-0.3013586998,0.3180028498,-0.0124288118,0.3443676829,0.1628109962,-0.1405374259,0.3648334742,-0.4071855545,0.1470770836,-0.1501058638,-0.0884457752,0.225179866,0.1613943279,-0.0398418158,-0.089559041,0.0113511719,0.1909005791,-0.1657792926,0.2259114832,0.1029563099,0.1410241425,-0.0905456245,0.2925598919,0.1837338805,-0.1588553488,0.3564889729,-0.0035605896,-0.001701441,-0.2622153163,-0.239713192,-0.0607687868,-0.227107957,0.110696286,-0.1198342294,0.0724323243,-0.4626992345,-0.1070561334,0.2359886169,-0.219233036,-0.1028048247,0.2456588596,-0.2401552945,0.2194226682,0.0660237074,0.1433586329,-0.1114361808,-0.4345372617,0.042465169,-0.0195416287,0.0899852663,-0.0060185194,-0.1278405041,-0.0399852246,-0.4660598338,0.3376529813,0.0783782601,0.4010949135,-0.099181667,-0.1375371665,0.1914187074,-0.1769854873,0.4275419712,-0.2778253853,-0.0339010581,0.3285033703,0.3112407327,-0.4547242522,0.3086226285,-0.2257430106,0.2777044773,0.1855898052,0.1748583615,-0.3294090629,-0.3005191088,0.1791579276,0.1494628936,0.008634273,0.1485638618,-0.3394794762,-0.0016732878,-0.1113546342,-0.182601437,0.2354936302,0.3389024734,0.0575834699,0.0278852563,-0.4274666607,0.1114259362,0.22046718,0.0597962849,0.4646727145,-0.3555501699,0.0496690981,-0.1155369431,0.1643323004,-0.1048094854,0.4026122391,-0.1112103835,-0.2135509551,0.3041942716,-0.112002179,0.5195626616,0.3660560548,0.0167430565,0.3400315046,-0.2218255103,-0.1767712831,-0.249910593,0.0216227081,0.1132666692,-0.0408861227,0.1263376176,-0.1553531587,0.5446890593,0.3278231621,-0.0423442535,0.554161191,0.2117569894,-0.2942788303,0.3200251758,0.2080254406,1.1941516399,0.0975149125,0.1462044567,-0.0010855937,-0.4741478562,0.4118733704,-0.0163588934,0.2773300409,-0.4832222462,0.2552043498,-0.0313823335,-0.1429671198,0.0693818107,-0.028094355,-0.3118169308,0.1904614419,-0.2436841428,0.2909682691,0.1357038766,0.0610682033,-0.2717036903,-0.327760756,-0.4476536512,0.0764902756,-0.216713801,-0.0755661651,0.0710353106,-0.2106837183,-0.46296978,-0.1009377465,-0.1096979305,0.1593484581,-0.1239352077,-0.0755385384,0.2641775608,-0.1868511289,0.2251789123,0.2428012341,0.0051510916,-0.085886538,-0.1763661504,-0.0535290539,0.2447095364,0.1061554998,0.1058662832,-0.0118583869,0.0003787213,0.070486255,0.044414293,0.2821406722,-0.4312086403,-0.0090960078,-0.066575259,0.2011432648,0.0695093647,-0.0824180916,-0.0607964098,0.1203722432,0.0218506455,-0.0646410435,0.0904510245,0.0157693475,0.1222276241,0.0371176302,-0.0760432631,-0.2621574402,0.2522254884,0.3351610601,0.0323859677,0.2377116978,0.690772891,0.1275943667,-0.1585000306,-0.1919580847,0.0481658317,0.0197698083,-0.6897882223,0.3346213102,-0.3643206954,-0.3478064835,0.0628900602,0.5673362613,-0.1933099031,0.1077160537,-0.1368739158,-0.2788204551,-0.1844433993,0.1963490099,-0.0024157679,0.028380245,-0.499538362,-0.1366133839,0.1134828776,0.1590401232,-0.2193434536,0.0163507573,-0.2001546472,0.3430066407,0.0301612895,0.062082544,-0.0560723729,-0.1886552423,-0.0196915753,0.5024873614,-0.1026863754,-0.1211077496,-0.274211973,0.1608317345,0.1031468734,-0.0880671963,0.1394816935,0.0161973257,0.0141589334,-0.2199788839,-0.1095609814,-0.153038308,-0.1849958748,0.0294810813,-0.0220890082,0.2143386006,-0.2510090768,0.0601784326,-0.1326748729,0.0837212875,-0.2884572148,0.0229375921,0.1630709618,0.0336601548,-0.0639830828,0.0811350569,-0.1232751161,-0.1855483949,0.2869323492,-0.3388084769,-0.0838714242,0.0970267653,-0.170179978,0.382463932,-0.0943604857,0.081271559,-0.0530066118,0.0889465734,0.0513073206,0.2999001145,-0.2158815116,-0.0830496326,-0.1195335388,0.2756985128,0.0859593377,-0.1738395244,-0.0516573973,0.0674252138,0.0386295617,-0.3842224181,-0.1138981953,0.0971989781,-0.1142447665,-0.018068606,0.2924421132,-0.1032822132,-0.0458707735,0.040573366,0.3652764857,0.3483972847,0.401394397,0.1893411428,-0.0841978788,0.0716493353,0.0892937481,0.4148809612,0.3441501558,0.2213013172,0.2034816891,0.5567286015,-0.3023475707,-0.2982839644,-0.010283093,0.1112818867,-0.2391135842,-0.0918324664,-0.5055191517,-0.146472156,-0.4831380248,-0.1042034626,-0.1811038703,0.1869531274,0.6240800619,0.1954524666,0.0099638943,-0.5831868649,-0.0958307609,0.0742833465,0.3741016984,-0.2317101359,0.4682835042,0.0014006615,-0.0817828029,-0.0363667607,0.4819176197,0.434933275,0.593409121,-0.3435724676,-0.1032965854,0.1064883322,-0.1808678061,0.2682626247,0.585621655,0.30769068,0.2094869614,0.0841297433,-0.0577526949,-0.0918557495,-0.1141866595,-0.0435387753,0.3730556369,-0.1405731291,0.32953462,-0.2427956164,-0.1560450196,-0.1160674691,-0.2603744864,-0.3454605639,-0.0425232388,0.1088435277,-0.0311699118,0.0222361051,-0.3139159977,0.0586733334,0.2108155042,0.5190419555,0.253362447,0.1632292271,-0.4591372907,-0.1342713088,-0.7708756924,0.2027030587,-0.2689130306,-0.1755142361,-0.006510817,0.3170387745,0.0571975373,0.3095687032,-0.0267742239,0.2754848897,0.2724505961,0.2466341555,-0.3378659189,-0.1781610399,0.3060103357,-0.3140348494,-0.0229260717,-0.5555660725,0.2837131619,0.0565456524,0.0509872548,0.0589004606,-0.3053709269,0.2164331079,0.3871609271,0.3018264771,0.1883882433,0.2817597687,0.2313517183,-0.4715583026,-0.2614998817,-0.1822957546,-0.2568730116,-0.0096387994,0.0744564161,0.4999499321,-0.1720468551,0.160736531,-0.2870025933,0.2351617813,-0.1392562538,0.1039722562,-0.2287497818,0.1674924642,-0.3107847869,0.0993272141,0.1797893643,0.0760253444,0.1399240345,0.1800004095,-0.4830567539,-0.4031953216,0.3154704273,-0.4602033198,-0.3971212506,0.3065812588,0.151301384,0.5217301846,-0.0932722315,-0.7497122288,-0.3282910287,0.17236045,-0.0963836014,0.0480340868,-0.0193191804,-0.0399973392,-0.0425113998,-0.1998751611,0.2870979011,0.003773011,0.0928554609,0.0789516643,-0.235842675]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/211","title":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type","comments":"Interesting, maybe I have a very old cache of trivia_qa...thanks for checking","body":"Running the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.","comment_length":12,"text":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type\nRunning the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.\nInteresting, maybe I have a very old cache of trivia_qa...thanks for checking","embeddings":[0.0815915689,0.0226194877,0.1106822193,0.5200621486,0.2965884209,0.1182672158,0.0500505902,0.4389167428,0.1778379828,-0.1218920797,0.2872642577,0.5463170409,-0.049948521,-0.1204143167,-0.0644242391,-0.1754419506,0.1857061982,0.3033518791,0.0621347986,-0.0928809717,-0.1326537877,0.1615956128,-0.2754946351,0.149649471,-0.4977772832,-0.1065860987,0.0930235013,0.0984235257,-0.1430949122,-0.4206030071,0.024716476,-0.3854033351,0.0776064023,0.3144654632,-0.0001217892,-0.1654590219,0.3672672808,-0.035978131,-0.1888790578,-0.2179785818,-0.4966371655,0.1068594158,0.3482596576,0.0010817944,0.193729952,-0.0836432353,0.0069021033,-0.6113762259,0.104728058,0.582434237,0.1807434261,0.2383273542,-0.0089827515,0.0630786121,0.2789545953,0.0083455564,-0.3300781846,0.0195962414,-0.0186755322,-0.1566835493,-0.1206181198,0.3693993092,-0.1186642572,0.0821172893,0.5851637721,0.0704582632,0.0549570806,-0.2719405293,0.2508995831,0.0892555267,0.4095961154,-0.1283688396,-0.0612893142,0.0649682581,-0.1212382242,-0.3411201239,0.2489887923,0.1813202798,-0.2544384301,0.0110957017,0.137850076,0.0274826344,-0.0391927846,-0.1385638416,-0.1429374069,0.145560652,-0.0057646264,0.0617744103,-0.0741196722,-0.212978676,-0.0961189494,-0.1813616455,-0.2689131498,0.208840549,-0.3981795609,0.0529597513,-0.0856595114,-0.1160219908,-0.15999192,-0.0768951327,0.3025360405,0.0482324623,0.0429237597,0.0906082913,0.3786050975,0.3758144081,0.3749571741,0.1674407572,0.1688652039,-0.3206017017,-0.2299305201,-0.0357588157,-0.0216837507,-0.0469667837,0.4549809694,0.0011437803,0.2141607851,-0.4449777007,-0.2671929896,-0.0185125358,-0.7070002556,0.1942735612,-0.2352560163,0.1389518827,-0.0062718028,0.5296701193,0.2068359852,0.3754504621,-0.2865449488,-0.2881952226,-0.1452662796,-0.064258419,-0.2774612606,-0.0796211883,0.217228204,-0.0473301075,0.0621753968,0.0234667715,-0.4521855414,-0.2608604133,0.268643558,-0.3488604724,0.1003791392,0.0435360111,-0.2492947578,0.4156775177,0.2440104932,-0.4322059155,-0.1534494311,0.1448592097,-0.1953514516,0.0560564213,-0.1771434546,0.112880975,0.060354583,-0.2614082992,-0.2007490844,-0.011529428,0.3680129051,-0.1245643198,0.0732571706,-0.1721343249,-0.0726219118,-0.2826193869,-0.094201915,0.4472855031,-0.2383367121,0.1243113205,-0.0813081264,0.1487986445,0.3184575737,0.3028367162,0.0300524626,0.1525550485,0.1140047312,0.3298241496,0.5004433393,-0.2713876367,-0.0000985635,0.0926893353,-0.0180684347,-0.0075404835,-0.0125710303,0.0427040197,0.0204680935,-0.1337763369,0.1197645515,0.1738696992,-0.0754740089,0.2809634209,-0.4106211364,-0.0910562798,0.5827088952,-0.2179569453,-0.0770637468,0.0638042241,-0.022952985,0.0514820293,-0.0409885868,-0.023754904,0.3452984393,0.2517730892,-0.10026481,-0.1265173852,0.1149875298,-0.1364220232,-0.4045669734,-0.0295396913,-0.1145317554,0.1222889274,-0.0191522427,-0.0034767424,-0.3601198494,0.0610290468,-0.015682308,-0.245449096,0.1317766309,0.0594287217,0.1323270649,0.14673087,-0.2053805143,0.118571721,0.0319966823,0.1344992369,-0.3542610407,-0.1633894295,-0.0195974242,-0.3402493,-0.0875983611,0.2392680347,0.2881985307,0.1565015316,-0.295331955,0.2194465846,0.0384875312,0.3734241426,-0.2288529426,0.0440783389,0.021521505,-0.4434015751,0.2236722857,0.3638965189,0.3605910838,-0.288729012,-0.3465659618,-0.0840296671,0.0431649797,0.6536787152,-0.1293211132,0.1688137054,-0.0613750555,-0.0499125719,0.1808445901,-0.0902950466,0.1521717459,-0.1109834909,0.1464702338,0.2884094715,-0.2677832544,0.3154963553,0.4046091139,0.321891129,0.1573265195,0.1324456185,-0.3807038069,-0.1275446266,0.3086015284,0.041857928,0.4340999126,-0.0479465574,0.2310986966,-0.0735182688,-0.117130883,-0.1772957593,0.327114284,-0.1045562401,0.1126890108,0.0264416095,0.5187185407,-0.0129640102,-0.1658824682,0.3386836052,-0.0074376082,0.261108458,-0.3990004361,0.2832282484,-0.3860992491,-0.1258006692,-0.3739030957,-0.1528099328,-0.0885291398,-0.5627202392,-0.0693654269,-0.0241515562,0.0356989615,0.0504097641,-0.2267009467,-0.126089707,-0.1022065058,-0.0788678154,-0.3318128586,-0.3986173868,-0.37799263,-0.0735101402,0.4052711129,0.3438740373,0.1189498231,0.0196196232,-0.0129906628,-0.0235723294,-0.4130571187,0.0339666642,-0.2162513733,0.0964461118,0.1182405055,0.0063693607,0.0144237587,0.0012010403,0.3721850514,-0.0288288631,-0.2605857253,0.3654604852,0.0252587106,0.0851163194,-0.2505621314,-0.4256905615,0.0651857778,-0.510055244,0.1177264825,-0.1015890464,0.1872560978,0.1062870771,0.097426258,0.2695205212,0.0029033222,0.2182295173,0.2368136644,0.4346277714,0.5978099108,-0.1821014732,-0.2917025685,0.3067065477,-0.0059163305,0.3452472985,0.1536885351,-0.1407611221,0.3647142947,-0.4407851994,0.1739172041,-0.1430163234,-0.1187573224,0.2169482857,0.1583372951,-0.0375121608,-0.0942682624,0.0079690581,0.223945573,-0.182740137,0.220655635,0.1198150143,0.101973027,-0.1075193807,0.322506398,0.173394233,-0.1662156731,0.3483896554,-0.0251760408,0.0059236018,-0.2720957994,-0.2398208678,-0.0663424358,-0.2015595287,0.1385482401,-0.1173342764,0.0822045133,-0.4055430293,-0.0983133987,0.2351923287,-0.2028066069,-0.1271868199,0.2337110788,-0.2643714547,0.2180990726,0.0722748861,0.1563410908,-0.095654577,-0.4457072914,0.0519778915,-0.0141636068,0.0910366476,-0.0067789545,-0.0967145413,-0.0496284477,-0.4389213622,0.3275473416,0.0823504478,0.4239679575,-0.0975008011,-0.1606254578,0.2095021307,-0.1689560562,0.4330276549,-0.2708877623,-0.0106503041,0.3177721798,0.3219534457,-0.466535002,0.3080788851,-0.2129590213,0.2506359816,0.136405468,0.1743002385,-0.329318434,-0.3079715371,0.1784613878,0.1554884166,0.0130584128,0.1635409296,-0.3496623635,0.0218610335,-0.0782529712,-0.2022615671,0.2338024825,0.3496795297,0.0326647684,0.0381550901,-0.4410004914,0.1166195348,0.2030802518,0.0614642166,0.4602671862,-0.3707900047,0.0639919192,-0.1221063808,0.1342696697,-0.090495646,0.3787542582,-0.1180417091,-0.2310180664,0.2808675766,-0.1048983186,0.4982024133,0.3646539152,0.0022115509,0.3764047325,-0.2361522168,-0.189067021,-0.240858227,-0.0069727912,0.0989675447,-0.0392503031,0.1528442055,-0.1285971701,0.5343067646,0.3411870897,-0.0405854918,0.5309374928,0.209701106,-0.3049148023,0.3079588115,0.1918494701,1.1999578476,0.0926553085,0.1384346336,-0.0119858142,-0.5015724897,0.4178739786,-0.0080765188,0.2700544,-0.4936321974,0.244771719,-0.0117702456,-0.1354196519,0.0716568753,-0.0743890032,-0.3277114332,0.1766772717,-0.279453367,0.3034633696,0.1439434439,0.0373212546,-0.263276428,-0.3124392629,-0.4258271158,0.068856582,-0.2363269031,-0.0702392012,0.0860921964,-0.2203639001,-0.4575602412,-0.085522972,-0.1208222359,0.1703797132,-0.1379403174,-0.0903218016,0.2950904071,-0.1950427592,0.2248474061,0.2432579696,0.0333887823,-0.0614170246,-0.1902332008,-0.0210943986,0.2599402666,0.1037047133,0.1144587845,-0.000628828,-0.019878488,0.0739537403,0.020641882,0.2701588571,-0.4217082262,0.0186779611,-0.0880438611,0.1897701472,0.0867978036,-0.0550469682,-0.0380252153,0.1102067679,0.0389367305,-0.059441749,0.0870826989,0.0186647382,0.1479983479,0.033724267,-0.07112059,-0.2749660611,0.2642706931,0.3426049352,0.021853324,0.2309766412,0.6928747296,0.1368288249,-0.1652125567,-0.1830038279,0.033958219,-0.0257173292,-0.7012068033,0.346618861,-0.3432222307,-0.3580519259,0.0719118565,0.5678146482,-0.2179186195,0.0971808434,-0.1441485286,-0.2822344005,-0.1525725871,0.1779513061,-0.0320096351,-0.0021605291,-0.5196484923,-0.1589908451,0.0923639908,0.1492231786,-0.2114035636,0.0267121568,-0.1763073802,0.337174207,-0.0116460864,0.0532527789,-0.043628566,-0.1860941648,-0.0240351651,0.5130593181,-0.0993369222,-0.1158937439,-0.2747286856,0.1643834263,0.0851165354,-0.1045047492,0.1252050102,0.0162236243,0.0279101133,-0.2207286209,-0.1000375077,-0.156405583,-0.1859480739,0.0239612423,-0.03800565,0.2207030803,-0.2425322533,0.0963827521,-0.1353250742,0.0993392393,-0.2791275978,0.0260550343,0.1261902153,0.0191637054,-0.0700678304,0.0905529633,-0.1467940956,-0.1692093313,0.2653481364,-0.3271382153,-0.1118053049,0.081686683,-0.2002907842,0.3781558871,-0.0872770175,0.0625412762,-0.0578377806,0.0813074559,0.0737100914,0.3195416927,-0.2385116518,-0.0882035047,-0.1367934197,0.2827850878,0.0829074085,-0.1939115524,-0.0188091882,0.0784251913,0.0300739408,-0.3823366463,-0.0957712531,0.1073092967,-0.0921952426,-0.0430437364,0.2786695659,-0.1276631206,-0.0638678223,0.0326767899,0.373875767,0.3277689815,0.4184783697,0.1701004952,-0.0550385937,0.0824392065,0.0843420401,0.4281069934,0.3229798079,0.1949904263,0.1955960989,0.5824716687,-0.2847314179,-0.2831526995,-0.0107355202,0.0846421346,-0.2362597436,-0.1089303568,-0.4576545954,-0.1419727206,-0.4835977852,-0.1142712682,-0.1709546298,0.208017379,0.6045295596,0.2013069987,0.00072598,-0.6044605374,-0.0993287712,0.0511374474,0.3945424259,-0.212174654,0.4761853814,0.013403574,-0.0664891675,-0.0022782027,0.4948806167,0.4270305336,0.5960644484,-0.3401225209,-0.0965599045,0.0878546461,-0.170108363,0.288449049,0.5786643624,0.3149787784,0.1825464964,0.062185023,-0.0612134896,-0.0789353922,-0.1123909503,-0.03046198,0.3679826558,-0.1453763098,0.3177988231,-0.2454757392,-0.1367286593,-0.1291125566,-0.2617242038,-0.3267460465,-0.0593825392,0.1056932062,-0.0260101743,0.0002421726,-0.3355146348,0.0561855286,0.2162049562,0.5366623998,0.2416713834,0.1607522517,-0.432154566,-0.1306654662,-0.7495183349,0.2271977216,-0.2383941561,-0.1595100909,-0.0181975178,0.3165664673,0.0324803703,0.2985256314,-0.0164757948,0.3023664951,0.2919558585,0.2434955686,-0.345610708,-0.1781913489,0.3271206617,-0.3383353651,-0.0191375948,-0.5665605068,0.280026108,0.0362168662,0.0465900414,0.0737081543,-0.3153579235,0.2313855886,0.4097183645,0.3172701895,0.1874285787,0.2844011784,0.250995189,-0.4867772758,-0.2742004991,-0.1803050637,-0.2552199066,0.007198208,0.0816719681,0.4585078955,-0.1655096114,0.154154256,-0.28055498,0.2603527904,-0.1252953112,0.1182127222,-0.2094835192,0.161958918,-0.2981992662,0.0844459161,0.2035082877,0.0536603741,0.1480877697,0.2041311413,-0.489297092,-0.4145559669,0.3152243793,-0.4233131409,-0.3863821924,0.2862881422,0.1814534217,0.4807398915,-0.1290650368,-0.7846543193,-0.3337669373,0.1746289283,-0.0790510476,0.0585825965,-0.0074442881,-0.0535023957,-0.0094785476,-0.2113596648,0.310916543,0.0028026884,0.1127527356,0.084250316,-0.2091728449]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/211","title":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type","comments":"I'm running it right now on colab to double check","body":"Running the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.","comment_length":10,"text":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type\nRunning the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.\nI'm running it right now on colab to double check","embeddings":[0.0480893105,0.0667722374,0.1085615605,0.5296492577,0.2605206072,0.093816191,0.0609813184,0.3810651004,0.177226156,-0.1010037139,0.2677445114,0.6236650348,-0.0864928663,-0.0589846075,-0.1010638773,-0.1438142806,0.168724373,0.3264837265,0.0556953624,-0.0820062682,-0.1381243914,0.1824117303,-0.3065697253,0.0927224532,-0.5650300384,-0.1423870027,0.0535200164,0.0792569146,-0.1993207186,-0.3889350593,0.0666146502,-0.3421472907,0.0817118734,0.3262377977,-0.0001202366,-0.1601293832,0.3735945225,-0.0340500586,-0.2338624448,-0.2713955045,-0.4900063872,0.1130168661,0.3932447433,-0.026797168,0.1793713868,-0.0015622908,0.0326485299,-0.5390457511,0.1173131093,0.5812163949,0.1938722283,0.1990087777,-0.0549899042,0.0627977625,0.3000804782,0.0449625626,-0.3512454033,0.0400435217,-0.0210088491,-0.142594561,-0.1089738086,0.3847116232,-0.1198601797,0.0746500194,0.5543821454,0.1036097631,0.0238709748,-0.3169820607,0.2753407061,0.0738631934,0.4028307498,-0.1297866106,-0.0385312214,0.0511522442,-0.089764744,-0.3718752563,0.2867363393,0.1800092608,-0.2800912857,0.0143283708,0.0983970612,0.0866293013,-0.0118878866,-0.152230531,-0.148255676,0.2192691565,-0.0319387168,0.0491194054,-0.0583701879,-0.195120424,-0.0002668842,-0.1858894825,-0.2600758374,0.2030771822,-0.3533463478,0.0413501747,-0.0913854912,-0.0752388313,-0.192751959,-0.1416011751,0.3142448366,0.0260728616,0.0289717484,0.1312429607,0.3350276053,0.3655561209,0.372323513,0.1738983542,0.1687339395,-0.2659416497,-0.2505093813,-0.0276426952,-0.0392094404,-0.0201667324,0.3714633584,0.0199055821,0.192861855,-0.4345341623,-0.2945814431,-0.0609400868,-0.6962262988,0.2042704672,-0.1884779781,0.1659448445,-0.0227954611,0.4697283506,0.1882797182,0.3157243133,-0.2761719823,-0.262378484,-0.1396801174,-0.0287791379,-0.307292819,-0.0651656464,0.2434132397,-0.007647098,0.0915396735,0.0344397239,-0.4492713809,-0.2541579604,0.2337519825,-0.3950003982,0.1185083762,0.074523896,-0.2121037692,0.3887064457,0.2031974047,-0.4997521639,-0.1644617468,0.1716775745,-0.1989584416,0.0698010996,-0.1692725271,0.1199155673,0.020344805,-0.2690716088,-0.2313523293,0.0126240551,0.3633626103,-0.1264156848,0.0617253073,-0.2001249045,-0.1128140464,-0.2762353122,-0.0958559066,0.4426500797,-0.2401035577,0.1322120577,-0.0716500804,0.1952202022,0.4005531967,0.2952875495,0.0100136334,0.1351045668,0.1305546016,0.306068182,0.4828451872,-0.2078616768,-0.0031566496,0.0717464462,0.0038801383,-0.0099303853,-0.0236000717,0.0257324763,0.0453525856,-0.1451905817,0.1553202718,0.1556392312,-0.0699168816,0.2923382223,-0.4410095811,-0.0903935879,0.5686299205,-0.2131813318,-0.0759496242,0.0314371362,-0.0534612834,0.0378373228,-0.0363659039,-0.052888833,0.3161356151,0.2116544098,-0.0507344306,-0.1878997386,0.0862451047,-0.0975818411,-0.4051321447,-0.0262740888,-0.040670976,0.1617876291,-0.0345675237,-0.0287345536,-0.3501205444,0.0763498992,-0.0101381419,-0.2526894808,0.1395963728,0.0767020881,0.109268263,0.1248503998,-0.2293128222,0.1701932251,-0.0024294327,0.127233997,-0.29678455,-0.1245753914,-0.0584202483,-0.3832653165,-0.0644103289,0.2203728259,0.2826158106,0.126317203,-0.2955746949,0.1835253686,0.0488431938,0.3259671926,-0.2352430671,0.0178128928,0.0481123254,-0.4010072351,0.2000039816,0.3365230262,0.3892845213,-0.2663569748,-0.3703241348,-0.0240835287,0.00501404,0.6032654047,-0.1367916167,0.1407701373,-0.0350538045,-0.0556259193,0.1719469577,-0.0503982194,0.186479643,-0.1030768752,0.1363853067,0.3306151927,-0.2669204175,0.3567071855,0.4222005308,0.3470950425,0.1286945939,0.14308016,-0.3367902935,-0.1406753808,0.3474842012,0.0877628401,0.434011966,-0.0400012359,0.226492092,-0.0679607689,-0.1430413276,-0.2050662637,0.3189838529,-0.1201401427,0.1608909667,0.0037798251,0.5202202201,0.0009893073,-0.1903615743,0.2797827423,-0.0011332039,0.2842948735,-0.3685743809,0.3066794872,-0.3471457064,-0.192308262,-0.3562874794,-0.1500561386,-0.057535179,-0.5595604777,-0.0742150322,0.0008389877,0.0346365981,0.0443643071,-0.1984989643,-0.0949902311,-0.1322886944,-0.059345223,-0.3445601463,-0.4244661629,-0.3725248873,-0.0515830331,0.4386447072,0.3597255945,0.092051819,-0.0127100088,-0.0165250767,0.0652994365,-0.4110758305,0.0870807618,-0.2321134806,0.1129245982,0.1068437472,-0.0325406939,-0.0848812982,-0.0466840081,0.3297120035,-0.0364088342,-0.2586838007,0.2941873372,0.0088039208,0.0350860208,-0.2882985175,-0.3991149366,0.0406562313,-0.5034156442,0.1097567454,-0.0857853219,0.2218025923,0.1448938698,0.1105321348,0.3031277657,0.032282643,0.203843385,0.2327216268,0.3827218711,0.58363837,-0.1695624292,-0.308698386,0.3128097057,-0.0551659577,0.3356462717,0.1478131711,-0.137978375,0.3120506406,-0.4194453657,0.143567577,-0.1068474799,-0.1236817464,0.2147175372,0.1613297313,-0.0268474873,-0.0860782787,-0.0471903831,0.1776688695,-0.1420468539,0.2447874546,0.0746022463,0.0621449202,-0.1120416,0.2942593396,0.1499575377,-0.1879729182,0.3477005959,-0.0173813105,0.0186997242,-0.2608976662,-0.2472410351,-0.0164440796,-0.2032659352,0.0303548761,-0.0628026724,0.044541318,-0.46392259,-0.1134169549,0.2611301839,-0.2211518586,-0.1581289023,0.2651157975,-0.1938144416,0.1796446443,0.0932727009,0.1774209738,-0.1004327238,-0.4540745318,-0.0035542804,0.0121107288,0.068484351,-0.025868522,-0.05415833,-0.0980461538,-0.4913547635,0.3066632152,0.136081785,0.3819529414,-0.0986309797,-0.1379529387,0.208044529,-0.1993058324,0.4048247635,-0.2446316034,-0.0729183108,0.3385815322,0.3145735264,-0.4368359447,0.2855974436,-0.2499644607,0.2924447656,0.1797084808,0.1747944355,-0.2818630636,-0.2786753476,0.1962566972,0.1147008538,0.0305087753,0.152346611,-0.361515075,-0.0436360277,-0.0712600425,-0.1813866496,0.2328737527,0.3371839821,0.0136124054,0.0204635635,-0.4142884016,0.1119640395,0.2299677283,0.1587169617,0.4292090237,-0.2991289794,0.0476271957,-0.1693030447,0.1510454714,-0.0659513697,0.4033815265,-0.1256480068,-0.1730933487,0.3050337434,-0.1217171103,0.510941267,0.3822390735,0.0000274247,0.3279665112,-0.2961199582,-0.21294716,-0.1804462969,0.0782238841,0.0531015135,-0.0195612144,0.0994509533,-0.1356768459,0.5265542865,0.3268057406,-0.0606945753,0.5157170892,0.1824962348,-0.3107460737,0.3746007085,0.2188255787,1.1999254227,0.0748959035,0.1282500923,0.0060153906,-0.4453760684,0.4525971115,0.0110220704,0.3171631396,-0.4779934883,0.2194014043,0.00529241,-0.1315881014,0.0382815078,-0.1166551486,-0.3830088079,0.1568545699,-0.2088070065,0.2900951505,0.1332897991,0.0235067904,-0.2741872668,-0.336522162,-0.4775016606,0.0819807053,-0.2264422625,-0.0608479455,0.056310311,-0.2063721418,-0.4893695712,-0.0664893761,-0.1176345125,0.1827410609,-0.1963394433,-0.1419383585,0.2891915441,-0.1974031627,0.2290159613,0.227722317,0.0343407653,-0.0478773378,-0.1723489165,-0.0331872627,0.2483917475,0.1066316068,0.0993465856,-0.052338168,0.0149604231,0.0804746374,0.0788736269,0.2154269964,-0.3935135007,0.0298877917,0.0131340325,0.1995102018,0.0608656257,-0.059228301,-0.0425418355,0.1326788962,0.0333851539,-0.076442942,0.0985243097,-0.0036124219,0.1172437221,0.0828989074,-0.056620162,-0.2794885337,0.2534480393,0.308512032,0.0400965884,0.2681950927,0.6823507547,0.1304600835,-0.1574467272,-0.1911015362,0.0274574775,-0.0678633898,-0.6642822027,0.3597509861,-0.3682350814,-0.3779307604,0.0932276621,0.6675080657,-0.1640948206,0.1156947538,-0.1398541778,-0.2782929242,-0.184052363,0.1627959311,-0.0142584676,0.0081278402,-0.4958683848,-0.0925041512,0.1149824634,0.1261862963,-0.2294740528,0.025523955,-0.2245189995,0.3512918353,0.0011760981,0.0452192761,-0.0163312666,-0.1936998516,-0.0215060748,0.5412877202,-0.0846685693,-0.1373948157,-0.2794215977,0.1535052657,0.1188994795,-0.0973076448,0.0836042836,0.0376693271,0.0124376426,-0.2500221729,-0.0979005396,-0.1839407235,-0.1937756687,0.002970129,-0.037439052,0.2585334182,-0.2307790667,0.0846618712,-0.1189985201,0.1255868673,-0.2650847137,0.0384048633,0.1395966709,0.0303799659,0.0208656359,0.0852470845,-0.1784784645,-0.1424946934,0.2341419011,-0.3402422965,-0.1014682651,0.0709299296,-0.1941245049,0.3916609287,-0.0377354324,0.0379233398,-0.0290670507,0.0795696825,0.0816088915,0.349121958,-0.1900334656,-0.0250976961,-0.1007671431,0.3291122615,0.0881166384,-0.1584125608,-0.0384297259,0.0802881196,0.0354099423,-0.2982473075,-0.1102371216,0.0608013123,-0.1223330274,0.0177384336,0.2913070917,-0.1455857903,-0.0411873385,0.0187398307,0.3484690189,0.3644358516,0.4174796045,0.1777967662,-0.0359509028,0.0896256119,0.119268246,0.3929525614,0.3521960378,0.1986268759,0.2151096761,0.5894825459,-0.3430722058,-0.2800378501,-0.0404651985,0.105426982,-0.2712842524,-0.1069758013,-0.4706119895,-0.1515833437,-0.499683708,-0.0618520491,-0.1788115203,0.237350896,0.6190884709,0.2120283395,-0.0100686094,-0.5746662617,-0.1184434891,0.0892529339,0.3904535472,-0.2039970011,0.5170497298,-0.0480642803,-0.0741690844,-0.0587764941,0.4920181036,0.4421664476,0.6104923487,-0.3406369984,-0.1448977739,0.0786442384,-0.1615059525,0.2579060793,0.5589845777,0.3654216826,0.2255904526,0.0771813691,-0.0486107841,-0.0967769399,-0.0639988109,-0.0452559516,0.3317114413,-0.1019262671,0.2622689903,-0.2204122692,-0.1160770133,-0.1238849089,-0.2548454404,-0.3252126276,-0.0405538119,0.0717992336,-0.0783020854,0.0480034761,-0.2875605524,0.061994683,0.217863813,0.5684274435,0.2520700097,0.1921340376,-0.4465860426,-0.1596875787,-0.7349078655,0.2110713273,-0.2527894676,-0.1745941341,-0.0584236085,0.3102003932,0.0359558314,0.3145215213,-0.0766029879,0.3228542805,0.3275579512,0.2106305659,-0.3631190956,-0.170350194,0.3001970053,-0.4052186906,-0.0330678187,-0.5582467914,0.2596569955,-0.0104172528,0.0529890433,0.0557493083,-0.3415736556,0.1959198117,0.3701582551,0.3203967214,0.2081120163,0.2887986004,0.2131272554,-0.4462191463,-0.2478254139,-0.1931089163,-0.2600665689,-0.0268158671,0.1225120276,0.4825780392,-0.1449440569,0.2367801964,-0.2664595842,0.2475852966,-0.083092697,0.1303134263,-0.1923425645,0.1240696684,-0.3360658884,0.0913191289,0.1680865586,0.0852101222,0.1221005395,0.1767856032,-0.4870480299,-0.4038442671,0.3484995067,-0.4635193646,-0.3969098032,0.2636337578,0.1394027621,0.5172697902,-0.1342155039,-0.7863292694,-0.3409047425,0.1320021152,-0.1264244169,0.0223593824,0.0075459713,-0.03083173,-0.0152580962,-0.2320481986,0.3216435313,0.002031472,0.0673131123,0.0246587247,-0.2143123895]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/211","title":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type","comments":"Actually, I know what the problem is...I'm quite sure it's a bug. Here we take some test inputs: https:\/\/github.com\/huggingface\/nlp\/blob\/0e0ef12c14d2175e0b0bd7d8aa814b09e2cd7e1f\/src\/nlp\/arrow_dataset.py#L472\r\n\r\nIt might be that in the test inputs, a `Sequence` type value is an emtpy list. So in my case I have `ds[0][\"entity_pages'][\"wiki_context\"] = []`. => this leads to an `arrow_schema` equal to `null` for `[\"entity_pages'][\"wiki_context\"]` => see line: https:\/\/github.com\/huggingface\/nlp\/blob\/0e0ef12c14d2175e0b0bd7d8aa814b09e2cd7e1f\/src\/nlp\/arrow_dataset.py#L501 instead of list of string which it should for other examples. \r\n\r\nGuess it's an edge case, but it can happen.","body":"Running the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.","comment_length":79,"text":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type\nRunning the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.\nActually, I know what the problem is...I'm quite sure it's a bug. Here we take some test inputs: https:\/\/github.com\/huggingface\/nlp\/blob\/0e0ef12c14d2175e0b0bd7d8aa814b09e2cd7e1f\/src\/nlp\/arrow_dataset.py#L472\r\n\r\nIt might be that in the test inputs, a `Sequence` type value is an emtpy list. So in my case I have `ds[0][\"entity_pages'][\"wiki_context\"] = []`. => this leads to an `arrow_schema` equal to `null` for `[\"entity_pages'][\"wiki_context\"]` => see line: https:\/\/github.com\/huggingface\/nlp\/blob\/0e0ef12c14d2175e0b0bd7d8aa814b09e2cd7e1f\/src\/nlp\/arrow_dataset.py#L501 instead of list of string which it should for other examples. \r\n\r\nGuess it's an edge case, but it can happen.","embeddings":[0.0935577601,-0.0521807335,0.0917302594,0.4922743142,0.2525187731,0.1492104828,0.1087132543,0.4076998234,0.1446877569,-0.0935794711,0.3454068303,0.584651649,-0.1262975335,-0.081480436,-0.039066568,-0.1896718293,0.1572439373,0.2686877549,0.0862914771,-0.0600314774,-0.1431038082,0.150540933,-0.2789250314,0.1321711391,-0.4200149477,-0.1031522155,0.0767619088,0.0571165048,-0.0884185731,-0.4528536499,0.0761469081,-0.3256983161,-0.0065747625,0.2591454387,-0.0001200363,-0.1207483485,0.4130158126,-0.0403311551,-0.2655786276,-0.2495539635,-0.4358327687,0.0318891145,0.3316243291,-0.0424906984,0.1805734038,-0.0637463704,0.0238026846,-0.5875159502,0.2061086893,0.5833625793,0.176483199,0.2174149603,-0.0684086382,0.1302618235,0.2942125499,0.0531598404,-0.338309139,-0.082702823,0.0226327311,-0.1217891127,-0.1331658512,0.3214308619,-0.0617509335,0.0585705712,0.6247366667,0.0932786316,0.1312825382,-0.3116702735,0.1474959999,0.1445359737,0.3962809741,-0.0886665434,-0.08991009,-0.0177704319,-0.1631635725,-0.3528529704,0.2785193622,0.1806472689,-0.2984367311,-0.0033442734,0.0748486668,0.1483760923,-0.0872329921,-0.0642626882,-0.1187096238,0.1531071216,0.0393165834,0.0668651089,-0.1040986702,-0.2040971518,-0.1083757952,-0.2334066629,-0.2524464428,0.31610322,-0.4382790625,0.0379723832,-0.0514421239,-0.137963295,-0.1476160586,0.0629595444,0.3841155767,0.0998169333,0.027437998,0.0716608018,0.3587138951,0.3659048676,0.3182042241,0.1350381672,0.0632362217,-0.2223580182,-0.0815998092,0.0175917875,0.0166998059,-0.039309673,0.4090114534,0.0091384687,0.1497365087,-0.3318137527,-0.2487682402,0.0587406121,-0.7114661932,0.230327636,-0.1983433962,0.0958948359,-0.0247162059,0.5226482153,0.2696353793,0.3240777254,-0.2275396883,-0.1015938893,-0.1469991058,-0.1375631243,-0.2435155362,-0.0453521013,0.2112690806,-0.0735243782,0.1657229066,0.0246097576,-0.4914286137,-0.2029517591,0.266869396,-0.3210095167,0.1762523651,0.0648161396,-0.3296069503,0.4532786608,0.1861167103,-0.3901233375,-0.181904912,0.103387557,-0.1929475814,0.0950197428,-0.1574056447,0.1287296116,0.0248793215,-0.1485266238,-0.3023641407,-0.0322797969,0.3422367275,-0.1120218337,0.1528177857,-0.2467699349,-0.1065720692,-0.2501494586,-0.0631762892,0.4090130329,-0.1929151863,0.0867756382,-0.0184203107,0.1405178308,0.2578379214,0.2855256498,0.0161442813,0.1215701923,0.063467972,0.3946039379,0.4315240085,-0.279081434,0.0075398842,0.1328537911,-0.0355771817,-0.0118561182,0.0022210458,0.1064522564,0.0330913626,-0.2236135006,0.1074813232,0.199086979,-0.0646120831,0.2510938942,-0.4027025104,-0.0663079694,0.4951446652,-0.2019708455,-0.1811583042,-0.0547601432,-0.0537452698,0.1514837444,-0.0141628962,-0.0288183056,0.3827757835,0.3288468122,-0.1439033747,-0.1607178599,0.100635983,-0.1073824242,-0.4678545594,0.011793917,-0.1890222728,0.2026111037,-0.0866255239,0.0280609094,-0.3410979807,0.1253488809,-0.1016520411,-0.2273107171,0.1494240463,0.0541739501,0.1610251516,0.1465349197,-0.1223564819,0.1121611074,0.0134227639,0.0821260437,-0.4639506638,-0.0835918933,0.0000041866,-0.2866661251,-0.0877610743,0.2652682364,0.2704835534,0.1123003811,-0.2972468436,0.2152297646,0.113789238,0.2788777649,-0.4021189809,0.0717955157,0.0632042587,-0.4297591448,0.1421477497,0.4353225231,0.3686198294,-0.3424832821,-0.388802886,-0.0415758975,0.0223490335,0.6803262234,-0.1308950037,0.1441080868,-0.0329954512,-0.0373431332,0.0707539394,-0.0549177192,0.1654370874,0.0179862641,0.123321794,0.3098256588,-0.2486727685,0.2911898494,0.3783714473,0.2824710608,0.1109832972,0.1221673265,-0.3852927983,-0.2181705385,0.3125986457,0.0250501968,0.3757635951,-0.0526905395,0.1864141971,-0.0541679859,-0.1827583611,-0.1757693738,0.3537929952,-0.0584243014,0.1938470602,0.0364122987,0.5011973381,-0.0158940665,-0.1313813627,0.2498072982,-0.0088671697,0.2077242434,-0.4109279811,0.3374972939,-0.4077059627,-0.1314309537,-0.3705583215,-0.1527723372,-0.0243969318,-0.6165552735,-0.0448949672,-0.0183785371,-0.0306285154,0.1224283129,-0.284360379,-0.1316617578,0.0132932328,-0.0776505396,-0.402428329,-0.4299962521,-0.4094293416,-0.0479497463,0.4301170409,0.3920782506,0.1036347523,0.0321872458,-0.1071349531,-0.0362480395,-0.4125105143,0.0343748108,-0.2577631474,0.0903121084,0.1510433257,0.0609035008,0.0196858011,-0.0868504643,0.3246415257,0.0006719413,-0.3420940638,0.3952510655,0.0531295091,0.1424460709,-0.305187732,-0.4616943896,0.0397003852,-0.4655478299,0.1820293069,0.0063488008,0.1815882623,0.2361270338,0.0544263609,0.2801814973,-0.0430229679,0.1691674441,0.1944939494,0.4796045125,0.5730380416,-0.1482831091,-0.3612468243,0.2676405013,-0.133779496,0.2956104875,0.1793285906,-0.1509426236,0.3191586733,-0.3194127977,0.1534694582,-0.1375831217,-0.1262972355,0.239814043,0.1834496409,-0.0422922596,-0.1074573323,0.0315641649,0.2597672045,-0.1865331382,0.1925479919,0.1195823997,0.2033810169,-0.046481736,0.3153975606,0.1880695075,-0.2586817443,0.3378894627,-0.0351511016,-0.0082450388,-0.2894046605,-0.1092589647,-0.0871891677,-0.1459803432,0.1523726434,-0.0830237567,0.0713328496,-0.4544129372,-0.0287066549,0.2991620004,-0.1059682518,-0.1450847685,0.2503580749,-0.2715928853,0.232623294,0.0730432868,0.153418988,-0.1189835593,-0.5184165239,0.0441826098,-0.0319945328,0.0928395763,-0.0209227409,-0.2019021362,0.017810218,-0.4196454883,0.2854292095,0.1683174223,0.4798285663,-0.0479730815,-0.191658482,0.1719032973,-0.1533489972,0.3857499957,-0.2743554413,-0.0134874526,0.3624569178,0.3600081503,-0.4810972214,0.2678273916,-0.2580388486,0.2621384859,0.1798731238,0.1962708831,-0.4151628613,-0.3015871644,0.1404119283,0.1914488375,-0.0212409999,0.163366586,-0.2792500257,-0.0234499127,-0.0662855357,-0.1777407378,0.2659941614,0.3581658304,0.0398139767,-0.0013971606,-0.4330096245,0.1133106127,0.1220485568,0.0209403243,0.4554503858,-0.2969655693,0.1239918917,-0.2000167817,0.1642338037,-0.0678914785,0.3436698914,-0.0663608462,-0.0957201421,0.3054823875,-0.1287686229,0.3905346394,0.3394410014,-0.026473999,0.3637794256,-0.1764840782,-0.1536545008,-0.2444369644,-0.0253160261,0.1409700215,-0.0874010324,0.1084104478,-0.1961810589,0.5411810279,0.3007130623,-0.0316545106,0.4439766407,0.1896410286,-0.3277938366,0.3507962823,0.219153747,1.1128391027,0.1189077348,0.1753781438,0.0842513368,-0.3473073244,0.4513010681,0.0224267803,0.2541964054,-0.485802561,0.257393986,-0.0355489291,-0.1875911504,0.0403053127,0.0243441593,-0.2863897681,0.1932086349,-0.1737762243,0.2628734112,0.1290338039,0.0199113674,-0.2860625982,-0.335441798,-0.483994633,0.075141117,-0.1586629152,-0.0643386915,0.1097120419,-0.2614796162,-0.493329823,-0.0919074416,-0.1271521449,0.1436118186,-0.2413685322,-0.0622903593,0.2582913041,-0.179766953,0.2281717211,0.1625608951,-0.048863396,-0.0574199371,-0.1887585968,-0.1197771952,0.2058803141,0.0907302201,0.1580939293,-0.019870298,0.1427912861,0.0580303371,0.0911154673,0.2642857134,-0.4077537656,-0.0016336767,0.0684538037,0.1640792489,0.0351541974,-0.0211390946,-0.1293338537,0.1025453806,-0.0163643714,-0.0341865309,0.0898297727,0.0469812527,0.0825754032,0.0582077391,-0.1064051166,-0.2266682088,0.2634593844,0.3531934917,0.0571988001,0.1427273452,0.7486610413,0.1637577266,-0.1326283664,-0.1922147572,0.0859968811,0.0185724515,-0.7875832319,0.2842776477,-0.2714299858,-0.3222102225,0.035847716,0.646566093,-0.2318476737,0.0895197093,-0.1663874984,-0.2611513138,-0.1747756749,0.1724402755,-0.0371308289,0.061160516,-0.5098103285,-0.1845772862,0.1025232747,0.1319337636,-0.2177729458,0.0001780363,-0.2947470844,0.3535670638,-0.0148663204,0.1016850844,0.0053182272,-0.1716876924,-0.0046555246,0.4534418285,-0.1637967974,-0.1482589841,-0.2384617478,0.1535465717,0.1145353392,-0.1168936566,0.0697984695,-0.0215583462,0.0305701997,-0.2121921331,-0.1073695794,-0.2103573233,-0.1751916111,-0.0135188345,0.0763070881,0.2623258829,-0.2717634737,0.1227253675,-0.1723663658,0.0783821195,-0.3097914755,0.020728182,0.1177961677,0.0214300118,-0.146422565,0.1059763059,-0.1635253429,-0.2105932683,0.3242958486,-0.3493185937,-0.2074198127,0.0435973369,-0.159149304,0.3959812224,-0.1054929197,0.0026340212,-0.0795061588,0.0834646225,0.0542800575,0.3221637309,-0.1934608072,-0.1152182892,-0.0412476845,0.2851088941,0.1726252139,-0.2153888196,-0.0268781204,0.1149025187,0.0282918271,-0.418459326,-0.0813358873,0.1888807118,-0.0489892364,0.0191485807,0.2683743238,-0.1958972365,0.0285013281,-0.0704047903,0.4393867552,0.4198611677,0.4042415619,0.1670336872,-0.0248903278,0.0855499059,0.0549137592,0.3567959666,0.3431524634,0.2155419886,0.2542289495,0.5395469069,-0.1761781722,-0.2687590122,-0.0565695614,0.1204204634,-0.2214541286,-0.1392701268,-0.4548982382,-0.1539667845,-0.4800411761,-0.2001110315,-0.1649404168,0.1381530464,0.5681523681,0.2202720344,0.0426487029,-0.5250192285,-0.1378671229,0.1064426452,0.3856297433,-0.1744782925,0.4604706168,0.0065395003,-0.0706381127,-0.0305367094,0.4228681326,0.4281765223,0.5327760577,-0.327239126,-0.0688587502,0.1260793656,-0.1641944349,0.2407129854,0.5911153555,0.2477370054,0.066037178,0.0658218116,-0.0559582785,-0.0827453509,-0.1022573784,-0.0220899191,0.3411232531,-0.1244357303,0.4214778244,-0.2777307034,-0.1460214406,-0.10044346,-0.1861699075,-0.2864459455,0.032042332,0.1036734879,0.0309975799,0.0779098645,-0.3275087774,0.0678440854,0.1837611943,0.5826793909,0.2241733074,0.1032864377,-0.3924221098,-0.0916994959,-0.7618367672,0.1753278822,-0.2416836768,-0.210876897,0.1027612835,0.3020718992,0.0831139386,0.3048590422,-0.0429100357,0.2784475982,0.252065599,0.2867037952,-0.382651329,-0.1970775872,0.3445274234,-0.2810958326,-0.0528833419,-0.5603260398,0.2307481319,0.0163493175,0.0587673336,0.0324563533,-0.2147503346,0.1856906414,0.4447227418,0.2075991482,0.1835728735,0.1758577377,0.2728566825,-0.4320274889,-0.3502247334,-0.2179296762,-0.1933140308,0.0342108235,0.0356037542,0.4834497273,-0.1684565395,0.0357520804,-0.2757136524,0.1209486425,-0.1355480999,0.1336915195,-0.2686521709,0.1832184345,-0.4139767289,0.1996792406,0.2292813063,0.0224060453,0.1284523755,0.1492483914,-0.4297791123,-0.4989631474,0.2569375336,-0.4704404473,-0.3264241815,0.2818513513,0.1456384063,0.3511776328,-0.0605602115,-0.7402795553,-0.2556822002,0.2354611009,-0.1298456043,0.0179632623,-0.0434294492,-0.0465827398,0.0261091031,-0.2346627116,0.2395868897,0.0006645587,0.1402980834,0.0331320725,-0.2577147782]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/211","title":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type","comments":"Good point, I think the schema should be infered at the writing stage where we have a `writer_batch_size` number of examples (typically 10k) so it's even less likely to run into this scenario.","body":"Running the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.","comment_length":33,"text":"[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type\nRunning the following code \r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, load_from_cache_file=False)\r\n```\r\n\r\ntriggers a `ArrowInvalid: Could not convert TagMe with type str: converting to null type` error.\r\n\r\nOn the other hand if we remove a certain column of `trivia_qa` which seems responsible for the bug, it works:\r\n\r\n```\r\nimport nlp\r\nds = nlp.load_dataset(\"trivia_qa\", \"rc\", split=\"validation[:1%]\") # this might take 2.3 min to download but it's cached afterwards...\r\nds.map(lambda x: x, remove_columns=[\"entity_pages\"], load_from_cache_file=False)\r\n```\r\n\r\n. Seems quite hard to debug what's going on here... @lhoestq @thomwolf - do you have a good first guess what the problem could be?\r\n\r\n**Note** BTW: I think this could be a good test to check that the datasets work correctly: Take a tiny portion of the dataset and check that it can be written correctly.\nGood point, I think the schema should be infered at the writing stage where we have a `writer_batch_size` number of examples (typically 10k) so it's even less likely to run into this scenario.","embeddings":[-0.0682159588,0.0609310456,0.1053825989,0.4737014472,0.2520805001,0.1043800265,0.0982681587,0.3812307119,0.2232889384,-0.0625068694,0.3270524144,0.4799827635,-0.0850764513,-0.0221855119,-0.0457396731,-0.13106893,0.1612938344,0.3236938417,0.0635415837,-0.1218426973,-0.143834874,0.0843204781,-0.2351943105,0.0762036294,-0.5430988073,-0.1695071906,0.0593361631,0.1031856909,-0.1546960771,-0.444822371,0.0146039668,-0.2505456209,0.1371708214,0.310092181,-0.0001177422,-0.1311674863,0.382363826,-0.0452552214,-0.2170929015,-0.1553127021,-0.4626958966,0.060809955,0.2633104622,-0.0684483796,0.1900988221,-0.0022964545,0.0641070083,-0.499919951,0.1078316048,0.5157188177,0.2082080543,0.2796896994,-0.0188001394,0.0332892723,0.3279869854,0.0210578609,-0.365647614,0.000564051,-0.0283158161,-0.1554972231,-0.1619089246,0.3987061381,-0.0881612748,0.0523482934,0.6084603667,0.0969526768,0.1183453053,-0.2912596464,0.1959602386,0.1253908724,0.4451818466,-0.1049862951,-0.1137465537,-0.0039369301,-0.1345479786,-0.3904245496,0.2390954942,0.1906224191,-0.2509997189,-0.0031803055,0.0595967285,0.1065190434,-0.0358187519,-0.1929807365,-0.066250287,0.2038974762,0.0615907647,0.042670615,-0.1046710238,-0.1581255347,-0.0125593115,-0.2313856333,-0.3124711215,0.250197202,-0.3363642097,0.0420314781,0.0092312638,-0.1078686789,-0.1433404088,-0.0746568143,0.3727562726,0.0236874502,0.1381733865,0.1114934534,0.3602565825,0.3803673387,0.3734174967,0.1389397085,0.1727739871,-0.2554156184,-0.2099593133,-0.0185158681,-0.0258867219,-0.0269207396,0.4833751321,-0.0181952342,0.126923278,-0.342744559,-0.286416173,0.0379694588,-0.6072553396,0.1798172444,-0.2164932191,0.180911392,0.014958024,0.4404576719,0.141150564,0.3197541833,-0.2866175771,-0.2129358798,-0.1690673083,-0.0671930015,-0.352997452,-0.0474066809,0.2316350043,0.0253452919,0.1709929854,0.0457521938,-0.5268825293,-0.1939303279,0.2752550542,-0.3364925086,0.1010017619,0.097999312,-0.18175596,0.3801154792,0.1523963213,-0.395026654,-0.2050777376,0.0757711753,-0.1422836185,-0.0146571267,-0.1426357627,0.1625868529,0.0194878001,-0.2284371108,-0.2854866982,-0.0441706069,0.3768354058,-0.0824771598,0.0331031308,-0.1942527443,-0.0768310428,-0.2546732128,-0.149711147,0.491435349,-0.1705068648,0.1637139916,0.0520674698,0.1588347703,0.3200914562,0.3464753926,-0.0512800626,0.2360598296,0.0744530186,0.3285940289,0.4783325791,-0.2767097354,0.054584939,0.1438684613,-0.0497444421,-0.0141193578,0.0557803698,0.0047464315,0.113134712,-0.1860814691,0.1133473292,0.1749228984,-0.1087215468,0.2992883325,-0.4459868968,-0.0496777222,0.3986061513,-0.2075344771,-0.081623584,0.0590871796,0.010010629,0.0553928502,0.0179253165,-0.0859188437,0.3454482555,0.298235774,-0.124192439,-0.1070694104,0.0870341286,-0.0994404778,-0.4447140992,0.0324312299,-0.1577639729,0.1230496839,0.050344836,-0.0975475609,-0.2944736779,0.0961500108,-0.038226407,-0.2505195737,0.1545303464,0.023196375,0.1731840819,0.0592500903,-0.2021054775,0.0661340356,-0.0157438554,0.107415773,-0.3512534201,-0.1788741052,-0.0442574099,-0.3509330451,-0.0398886614,0.2398837656,0.237018019,0.1002113223,-0.2585669458,0.2389192581,-0.001657697,0.3618729115,-0.2429510951,0.0875752792,-0.041957356,-0.2605344057,0.1956302673,0.2751905024,0.3589321077,-0.2975812852,-0.4221465588,-0.0065395865,0.0379454531,0.5811731815,-0.0447000004,0.1357868016,0.0336808674,-0.0795980692,0.1372735947,-0.1051984504,0.1780075729,-0.0735728443,0.0638778731,0.3371360302,-0.2674142718,0.3107471168,0.4285442829,0.2990176976,0.1116487235,0.0571209788,-0.2933969796,-0.136866942,0.3519334197,0.1324796528,0.4555586278,-0.0157048721,0.2098059505,-0.0727058724,-0.1649343669,-0.2009925693,0.3175549507,-0.1493178606,0.1967449933,0.0662128404,0.4594133794,0.0200231094,-0.2195431441,0.2116740495,0.036517866,0.2187836468,-0.3783154786,0.2838196456,-0.2859784365,-0.0772421584,-0.4313696325,-0.0424440168,-0.0464689098,-0.5650830269,-0.0383662693,0.007924024,-0.0930386931,0.036544878,-0.3283863366,-0.1192555726,-0.1217019036,-0.065365009,-0.2936858833,-0.4041466117,-0.3502987623,-0.0584838949,0.4213877022,0.3856321573,0.1997771859,0.0164284967,-0.0054529719,-0.0281507317,-0.4237630665,0.0470186062,-0.2218863964,0.043297112,0.1605933458,-0.0058562402,0.0517997704,-0.1029949561,0.3697011173,-0.0511306785,-0.2518925071,0.3521769047,0.0223147199,-0.0376341715,-0.2629387379,-0.4058380723,0.1166723669,-0.5667974949,0.10078796,-0.0469362214,0.2017458379,0.1716721356,0.0820748881,0.2255028635,-0.0652560666,0.1951522082,0.2803921103,0.3783949018,0.5403063297,-0.2259801775,-0.3230759203,0.2950859666,-0.1108957976,0.3608184755,0.2307130396,-0.1520201415,0.4052866697,-0.4103542268,0.1700476855,-0.0837432668,-0.117233336,0.3123679161,0.1904549301,-0.0683736354,-0.0968231037,0.0179946031,0.2136168331,-0.2014836669,0.1650692821,0.008489931,0.1998053938,-0.1033324972,0.3400717974,0.1244998872,-0.1656523943,0.3614376187,0.0115997614,0.0120794913,-0.2510290146,-0.1440345347,0.0199724529,-0.214868173,0.0616214536,-0.0479258671,0.056544099,-0.4713304043,-0.034646146,0.2368118614,-0.1476400644,-0.1639094651,0.3225658834,-0.3387586772,0.2803710103,0.0947590023,0.1365748495,-0.1194651425,-0.4064585865,0.0236012004,-0.0585123487,0.1087666303,-0.016414525,-0.2646563947,0.0026101954,-0.5347059369,0.249577567,0.0521733724,0.3960878253,-0.0433575436,-0.1855838001,0.1821517497,-0.2292856872,0.4075532556,-0.3126878738,0.0187744256,0.2859610319,0.298610121,-0.4982848763,0.1848805249,-0.2240577191,0.3302088082,0.2257486433,0.1855438501,-0.3720530272,-0.3285739124,0.2298182845,0.0447390266,-0.0006505501,0.1036787406,-0.3524675369,-0.0532914326,-0.0697411522,-0.14215675,0.2665815949,0.3564811647,-0.0709257424,0.0079501467,-0.4361245632,0.1583185643,0.1579498798,0.130937919,0.3880293369,-0.3398945928,0.1248553544,-0.1658344269,0.1916581988,-0.0455009788,0.4288339913,-0.2038114071,-0.1900624037,0.3096034825,-0.1745516062,0.4862303138,0.3790232539,0.0421507061,0.3114559054,-0.2637415826,-0.1782966256,-0.2773715258,-0.0435143895,0.1307752728,-0.088910684,0.0912303329,-0.2324367166,0.553139627,0.3016223907,-0.0555479042,0.4027088881,0.1888201535,-0.3564832509,0.465826869,0.2532046437,1.1186988354,0.1057584211,0.2223715484,0.0254039336,-0.4089384377,0.3530541658,-0.033066567,0.2725175619,-0.4955558479,0.227224648,0.0277887899,-0.122936599,0.079643786,0.0528324246,-0.2791074514,0.157748118,-0.1973510832,0.3361300528,0.0758870393,0.1265095174,-0.4026111662,-0.3452480137,-0.4859898388,0.0826870278,-0.0839158744,-0.0800849944,0.0268063359,-0.2240912467,-0.4069483578,-0.1369730979,-0.1386088729,0.1344418228,-0.1580027789,-0.1193529069,0.2568900883,-0.1544817686,0.1984219551,0.2281150669,0.0449158251,-0.0506123304,-0.1676095426,-0.0149977794,0.2701363862,0.1192907915,0.1429568231,-0.027162075,0.0714478195,0.0423078276,0.0858859122,0.2309127301,-0.4100562334,0.0110836076,-0.1211377531,0.2061462849,-0.0234229807,-0.1338873804,-0.0878252983,0.1156158447,-0.0341634192,-0.0831446499,0.1168229952,0.0296710934,0.1122730896,0.0999394283,-0.1505665779,-0.2468963861,0.2206156701,0.3188544512,0.0564895123,0.2368175387,0.6745464206,0.080355756,-0.1407398582,-0.2325723171,0.0480734892,0.0512046702,-0.7669643164,0.3388836086,-0.2679667473,-0.4024946392,0.1367842704,0.591776669,-0.1855553389,0.0137405787,-0.0944648683,-0.2665635943,-0.0999528244,0.1929178536,0.0318585038,0.0735643283,-0.4767542183,-0.1263425201,0.0770130157,0.2244807631,-0.2680850625,-0.0489128269,-0.2327403575,0.3796295822,0.0144707151,0.080275692,-0.029821774,-0.1508358121,-0.0067596347,0.4313045442,-0.143753767,-0.1511628479,-0.2718142569,0.1370279789,0.0731023699,-0.0248396751,0.1579179913,0.0298142042,-0.0124052837,-0.2556419373,-0.0019756027,-0.1860365868,-0.2229129374,-0.0284394529,0.0585207641,0.2825692296,-0.2086721212,0.1272779256,-0.1877675653,0.0798761547,-0.3394127488,0.0716509968,0.0795838833,0.0157308634,0.0319569856,0.0908099413,-0.1425256282,-0.0751568824,0.3140852153,-0.3454215825,-0.104538314,0.1135717705,-0.117106922,0.3987213671,-0.0842996165,0.0142189916,-0.0262145419,0.1328930259,0.0771411061,0.2576906085,-0.2710583508,-0.126392439,-0.0510775633,0.3045227826,0.1776770353,-0.2135095298,-0.0405853987,0.1284638345,0.0542636327,-0.391677171,-0.1485806704,0.0897697657,-0.099477008,0.0070582028,0.3367123902,-0.1119775474,0.0215390697,0.0827948898,0.4989666045,0.458220005,0.3991214335,0.1686481535,-0.1109307855,-0.0017288941,0.0215453748,0.3171045184,0.3153476715,0.30454427,0.1548912376,0.5104090571,-0.3058184385,-0.2672063708,-0.0920049846,0.0621447563,-0.2543641031,-0.1457484066,-0.5064582229,-0.1419048905,-0.4231629074,-0.1009150892,-0.1865214258,0.2660886049,0.5050708055,0.1783146858,0.0029865301,-0.5828503966,-0.0561324097,0.0311091803,0.3965829015,-0.2468403727,0.5015381575,-0.0348526202,-0.0985032618,-0.0982269496,0.5166693926,0.4602744281,0.6580978632,-0.2757391632,-0.0701605454,0.075531818,-0.1338509619,0.152148366,0.5606760979,0.327612251,0.1597324461,0.1191944256,-0.0344189592,-0.09463384,-0.1187490746,-0.0343739018,0.2678046823,-0.0924472809,0.3559325635,-0.1539549828,-0.2433710992,-0.0867159441,-0.1551567167,-0.2883932889,-0.0563558377,0.1319946498,-0.0714782849,0.0546446331,-0.4272519946,0.0791518986,0.2093691826,0.6322180629,0.2378586233,0.1609148383,-0.4584115744,-0.0798666626,-0.7588373423,0.1408186704,-0.3934166729,-0.1612670422,0.0268487819,0.3510175645,-0.0143461665,0.2293818444,-0.0559111051,0.3212631643,0.1782591939,0.2982578874,-0.3454144895,-0.1791665256,0.2205493897,-0.344370842,-0.0490588211,-0.6017602682,0.3270027637,0.0500857607,0.0817348212,-0.0072829435,-0.2222327143,0.3362937272,0.288228035,0.3257759809,0.1670262665,0.2884242535,0.1926187426,-0.4587065876,-0.3980751932,-0.1998036653,-0.3066636324,-0.0998417139,0.1319087446,0.4690833688,-0.1796379089,0.1180591881,-0.2406579852,0.2599414885,-0.1059807763,0.1154151484,-0.1505337059,0.1351409107,-0.3236263096,0.1183072329,0.2612648308,0.1605714858,0.1331457496,0.1277522743,-0.4737323523,-0.4435833097,0.2808890045,-0.3823681474,-0.3833841085,0.2526187003,0.2217255235,0.39790079,-0.0912841484,-0.7310703397,-0.3031186163,0.1360622048,-0.0894443616,0.0044163922,0.0632362366,0.0770340711,-0.077324599,-0.2166134268,0.283967942,-0.0552300029,0.0264831614,0.0541517325,-0.2534835935]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/207","title":"Remove test set from NLP viewer","comments":"~is the viewer also open source?~\r\n[is a streamlit app!](https:\/\/docs.streamlit.io\/en\/latest\/getting_started.html)","body":"While the new [NLP viewer](https:\/\/huggingface.co\/nlp\/viewer\/) is a great tool, I think it would be best to outright remove the option of looking at the test sets. At the very least, a warning should be displayed to users before showing the test set. Newcomers to the field might not be aware of best practices, and small things like this can help increase awareness.","comment_length":10,"text":"Remove test set from NLP viewer\nWhile the new [NLP viewer](https:\/\/huggingface.co\/nlp\/viewer\/) is a great tool, I think it would be best to outright remove the option of looking at the test sets. At the very least, a warning should be displayed to users before showing the test set. Newcomers to the field might not be aware of best practices, and small things like this can help increase awareness.\n~is the viewer also open source?~\r\n[is a streamlit app!](https:\/\/docs.streamlit.io\/en\/latest\/getting_started.html)","embeddings":[-0.2110095769,0.098418735,-0.1958587915,-0.3097161353,-0.1695863307,-0.0453360192,0.2990845144,0.5401100516,0.0683938935,0.0143513093,0.0079117659,0.1889037937,-0.3738262653,0.0976688564,-0.0159248803,0.107477285,-0.269503653,0.4862023592,-0.129868716,-0.0009176507,-0.1562902033,-0.1299041808,-0.1428512484,0.3042346537,0.2094916403,-0.4386712015,0.1776946038,-0.1090400741,-0.0834257901,-0.578415513,-0.2426900417,0.171083346,-0.1658805609,-0.0196691472,-0.0000969443,0.0464205034,0.1829655021,0.140358761,-0.3207639754,0.0101503311,-0.1010943726,0.0497509204,0.2011304498,-0.159974128,-0.0944643468,-0.2562600374,0.1748544276,-0.2383663654,0.1214829981,0.4285504222,0.3378430605,0.3450808525,-0.1527639329,0.4335133433,0.1126120463,-0.0689012259,-0.2790054679,-0.0005679228,0.1863801926,0.1410393864,-0.3900102675,0.4503338635,-0.1646986157,-0.0894512013,0.0573407747,-0.2781302929,0.27737239,-0.3179795742,-0.2671356499,0.0944977403,0.2566865683,-0.141046524,-0.097925216,-0.4215677381,0.223830685,0.058974456,0.1668804586,0.4064157605,-0.075925678,0.2061407417,-0.4804442227,-0.4976486862,-0.1986560971,0.0851147324,0.2335148454,0.0352168009,0.0789648071,-0.0701415464,0.3335147202,-0.0647307187,0.1492420882,-0.1109645888,-0.3052551746,0.0518581495,-0.0511221774,-0.4391216934,0.1175679713,0.0213620476,-0.1357964128,0.2176984251,0.1161268055,0.109359026,-0.297233969,0.1078652889,0.1005073115,0.1596109271,0.3993632793,0.1500516683,0.1761577874,0.2030792683,0.2554517686,0.341781646,0.1851298213,-0.2128921747,-0.4279878139,0.1122513115,0.3657061458,-0.197429046,-0.3824758828,-0.0665570423,-0.4684957862,-0.0593758486,0.0993221328,0.1530266851,-0.0837139934,0.0720987767,-0.105842866,0.040105179,-0.415337652,-0.4231235981,-0.0010418892,0.2029075027,0.0051959045,-0.069755666,0.3149054348,-0.2352563888,0.2299571633,-0.048864577,-0.0018346056,0.097706534,0.3883042634,0.0114113847,0.2137249112,-0.1559121609,-0.0620094277,-0.2153007537,-0.2304398417,-0.2286459506,-0.1624726504,0.1054876074,0.0244567133,-0.2421095818,-0.0052438742,0.3551528156,-0.3165808916,-0.1079439372,0.0922569558,0.2553347051,-0.3320835829,-0.0069169179,0.1522148252,-0.1095538661,-0.1606890112,0.085455738,-0.2128683478,0.2769011259,-0.0243465975,-0.1200715527,-0.1003134325,-0.1185137555,-0.1749677807,-0.001016928,-0.3921530247,0.1828976274,-0.2154341787,-0.0758169964,0.5635842681,-0.2181901932,-0.0486899726,0.271805048,-0.2363864779,-0.1280036569,-0.1246191561,-0.0106946677,-0.1066288874,-0.492947787,-0.5431936383,0.105083786,0.0961205512,0.104576014,-0.4499922097,-0.417830497,0.3657763004,-0.0290010646,-0.0144070983,-0.1770738065,-0.0546132699,0.3216380477,0.3730461597,0.0057820813,-0.1513338983,-0.2302831113,0.221094206,-0.2816556394,-0.3044033051,-0.0591857769,-0.0600956567,-0.0891192108,-0.1052412167,0.1752845645,0.5634385943,-0.1852656901,-0.1090030819,0.0219746307,-0.1274543703,-0.1025511101,0.2927740812,0.0331262723,-0.038682282,0.1511259228,0.0159965474,-0.2642655671,-0.1791734695,0.0444151759,0.0255680699,-0.2292102128,-0.077826798,0.1733240783,0.3244105279,0.2880102396,0.0948420838,0.036018271,0.0332012288,0.2566795051,-0.1088421866,0.2451310605,0.2665672898,0.2301718593,0.0919485614,-0.4812614322,-0.0384574793,0.2001722753,-0.1254941523,0.2509694099,-0.2915358841,0.1663051248,0.1320274174,-0.0855017006,0.1130964309,0.0495440252,-0.0104498854,-0.2773448527,-0.0453637317,-0.2274691463,0.1060108468,0.2445632815,-0.1675885171,0.2023359686,-0.3135494292,0.1160324812,0.1391296089,-0.101396136,0.0021288109,0.1557213962,0.2271798104,-0.2552861869,0.26399073,0.1102552265,0.1146600246,0.2882956266,0.2065795213,0.1832588613,-0.1833538413,-0.2870340645,0.3689641953,0.39651227,0.0592316836,-0.0097879209,0.2500662506,-0.1288114339,-0.2464787811,0.0507526025,-0.0891418532,-0.0065524788,-0.2636340857,0.1174008623,-0.1275897473,-0.4871332347,-0.1909794509,0.1496600658,-0.0275191739,-0.2306660265,0.5265179873,0.0206678789,-0.3009591401,0.2918970585,-0.2886877954,0.6019861698,-0.1763487756,0.2291610241,0.0138939321,0.0629080534,-0.4042978883,0.3233160973,0.3609203398,0.008538736,0.5991446376,0.1156458184,-0.178562656,-0.2214954197,-0.2898885012,0.126807183,-0.0481432304,0.251290381,0.0149588594,-0.2167275846,0.0063951397,-0.1769767255,-0.3444831073,-0.2851008177,0.036449708,-0.1652980149,0.1558251381,0.2573947012,-0.2144459039,-0.5000034571,0.0081707789,-0.2610873282,0.1014148667,0.0877023786,0.2494539022,0.1684929729,-0.2967253327,0.056189172,-0.0541833714,0.0594887324,-0.0615343377,0.0483361445,0.0033303897,-0.2261133343,-0.3288852274,0.0775439963,-0.2068464607,0.1827494949,-0.1765133291,-0.6586421728,-0.1304549724,0.1728957891,-0.1367318779,0.0027185653,-0.5150877237,-0.078463003,0.2045714706,-0.2587800026,0.1202746183,0.1531234831,0.1504582018,-0.0998257995,0.2374103218,-0.0778154582,-0.1016727537,0.2533317506,0.1720400006,0.2538751662,0.1276361346,-0.0266434737,0.0774534792,0.1975186169,-0.040216092,-0.1319825202,0.1782912314,0.1374601126,0.0185832083,0.3544260263,-0.068297036,-0.3503843546,-0.0077317292,0.0694417953,-0.1824364513,-0.2977960706,0.0983995721,-0.4328030646,0.1930602193,0.0424537659,0.0531011894,0.0602749735,-0.5738990903,0.1118021533,-0.1808000803,0.1759440452,-0.0317799374,-0.3110312819,0.0717261508,-0.2879727781,0.028478317,0.0831452534,0.2392481565,-0.1004210934,-0.1711031049,0.2644880712,0.0813341662,0.0677803755,0.1052818298,0.0840146393,0.088661477,-0.0202748347,-0.3177227974,0.1060756072,-0.2457875311,-0.0399265066,0.2129860073,-0.0557714328,0.089696236,-0.2298427671,0.3375851512,0.004467472,-0.2229251862,0.0545118265,-0.0181190353,0.0014356377,-0.0999844968,0.1559976041,-0.2575559914,0.0135391569,-0.3392727375,-0.0217985138,-0.0034825841,-0.1262433827,0.4096284211,0.2077260166,-0.0836326182,0.2286781669,0.0436239615,0.2366491854,0.1331710517,-0.2581902146,0.2189852744,0.2688424885,0.4009682834,-0.0443110131,-0.1220856458,0.2807529271,0.3513801396,0.0957427248,0.0444907658,-0.2614059746,0.1459391266,-0.3391713202,0.3500922024,0.230684638,-0.1008594856,0.0231964514,-0.2474853843,0.171959132,-0.2257834375,-0.3778115511,0.0584739149,-0.0020695243,-0.14696531,0.1584005058,0.4334210157,0.7610743046,0.042847164,0.0889063925,0.5125403404,-0.2271983922,0.4708398283,-0.1086404845,0.2401003242,-0.1120538563,0.0447663665,0.0116358614,-0.0748184398,0.1587709337,0.1700593531,-0.1004478484,0.0498301387,0.4288856983,-0.2627731562,-0.0369021818,0.4541571438,-0.0092013245,0.0022247105,-0.3208705485,0.29389745,-0.1630053967,-0.1013899744,0.0543571599,-0.0082193073,0.3025199473,-0.0973829702,0.0116280764,0.1518553495,0.1357382387,0.1850661486,-0.0791243017,-0.1344356388,0.4936760366,0.3467678428,-0.1920359731,0.1141060516,-0.0663802475,0.0873126984,-0.1615683287,0.2018802315,0.2572804987,0.3101580143,0.2436260432,-0.2648150325,-0.1839911193,0.3971939981,0.1868317723,0.1200180575,-0.2297255397,0.0047027287,0.3946354985,-0.3516457081,-0.0154684661,0.0012249791,-0.0624606758,-0.047970783,0.2883714139,-0.0798144639,0.0239682198,-0.0855806395,0.2860001028,-0.1827711463,0.033473298,-0.0022943111,0.108193554,0.0156014813,0.5321605802,-0.1701311916,-0.2337392569,-0.4426830113,0.1816545427,-0.2850289941,-0.2269865721,-0.0494986214,0.0297453292,-0.1841274649,0.0476216078,0.1875641793,-0.0440020189,0.0461196341,-0.1155106202,0.0383939408,0.0052530188,0.0764350668,-0.1725788414,0.1708360463,0.1688668579,0.0846217498,0.018004911,0.1629189402,-0.4558541179,-0.2847618461,-0.2826936245,0.1131512672,-0.3297881484,0.1782002002,-0.0293430723,0.041354388,0.1260090321,-0.3040156364,-0.2104502171,-0.3726899326,-0.1712924242,0.0601177253,-0.0676673502,-0.2156576961,0.0381649323,-0.0129186865,0.1462463737,0.0384983979,-0.1718996018,-0.4224681258,0.1131170988,0.5824346542,-0.1368386149,-0.4722678363,0.0173839126,-0.0419753343,-0.3342505991,0.1952354461,0.2279641926,0.2937599421,0.1424869895,-0.1590654701,-0.1120851636,0.1706705242,-0.0279626809,0.3185492456,-0.1346552223,0.0267185196,0.1503969431,0.028955603,0.0339234881,-0.0873181671,-0.2584092319,-0.2617842853,0.3405790627,0.3514449596,-0.3713923991,0.0104858503,0.2083596289,0.0502852648,0.102541402,0.1752658039,-0.0426225215,-0.0864698067,0.1896003783,0.0918257013,-0.0697798505,-0.3390619755,0.1848524064,0.211419791,-0.3501787186,-0.248100549,0.0619539768,0.0392756537,0.1739765555,0.0269574448,0.1117746606,0.5362628698,0.0115227317,0.3597778082,0.079812713,-0.0946048275,-0.0743754581,0.0721334219,-0.0667097941,0.1730427742,0.4139368832,0.3854359984,0.2050907165,-0.1358352453,-0.0794083253,0.1991738975,0.0247483682,0.1256323755,-0.340770036,0.1341264546,-0.1058502719,0.1591543704,0.1036969796,-0.1134651005,-0.0273798518,0.1538315862,-0.1785514653,-0.1379165798,0.130788222,-0.0424810909,0.2984336317,-0.1246965304,-0.0713923201,-0.0669164285,0.3263811469,0.1210394725,0.3942393064,-0.0889097899,-0.0816664398,-0.0269044712,-0.1484699249,-0.1723447442,0.0439459905,-0.2665364146,0.2787845433,0.0480805449,0.2332234383,0.2355327904,0.1780675799,-0.097899884,0.2490069568,0.3099666834,-0.190275982,-0.2878647745,0.2561464906,0.5557044148,-0.3001677692,-0.1948769838,-0.0097109219,-0.0488247238,0.2410914302,0.3113406599,-0.1288788766,0.1270270944,-0.3052480519,0.1573911756,0.3088734746,0.4196659923,0.0971693471,0.2058307379,-0.0566592887,-0.1407429874,-0.5313760638,0.1443710923,0.083362475,-0.1197418272,0.2383284718,-0.0816779882,-0.1910484284,0.28799209,-0.2751011252,0.1785166115,-0.122614421,-0.2157988399,-0.2587634623,0.0841688812,0.0399449617,0.3282043934,-0.1637761295,-0.1516511887,0.2942381501,-0.2534238994,0.1497104466,0.1492098272,0.1221647561,0.1001508161,0.195648998,0.0935730264,0.133031249,0.0995843709,-0.0427774601,0.0872753635,-0.0819318593,-0.0165007096,-0.2196096629,-0.1176387742,0.1573791355,0.2499886453,-0.140234217,0.2416749001,0.0697295889,0.2645647824,0.0177912991,0.0558176823,-0.183909595,0.0680636168,0.0594602674,0.2942656875,-0.1009034589,0.2267340571,-0.0619211905,-0.1894118488,-0.1121734232,-0.1978343129,0.337618798,-0.4304279089,0.0293114036,0.0347426459,0.3356059492,0.1345634162,-0.2351190448,-0.4517480135,-0.2374988794,0.3017439246,-0.108202152,0.2997711003,0.301912874,-0.0528761856,-0.0870517418,-0.055745434,0.2578886747,0.010076548,-0.0681643859,0.2284893543,-0.2190278918]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/207","title":"Remove test set from NLP viewer","comments":"Appears that [two thirds of those polled on Twitter](https:\/\/twitter.com\/srush_nlp\/status\/1265734497632477185) are in favor of _some_ mechanism for averting eyeballs from the test data.","body":"While the new [NLP viewer](https:\/\/huggingface.co\/nlp\/viewer\/) is a great tool, I think it would be best to outright remove the option of looking at the test sets. At the very least, a warning should be displayed to users before showing the test set. Newcomers to the field might not be aware of best practices, and small things like this can help increase awareness.","comment_length":22,"text":"Remove test set from NLP viewer\nWhile the new [NLP viewer](https:\/\/huggingface.co\/nlp\/viewer\/) is a great tool, I think it would be best to outright remove the option of looking at the test sets. At the very least, a warning should be displayed to users before showing the test set. Newcomers to the field might not be aware of best practices, and small things like this can help increase awareness.\nAppears that [two thirds of those polled on Twitter](https:\/\/twitter.com\/srush_nlp\/status\/1265734497632477185) are in favor of _some_ mechanism for averting eyeballs from the test data.","embeddings":[-0.1373631805,0.1704822779,-0.1900716126,-0.2922053933,-0.1791144311,0.0449159294,0.2525579333,0.5567144156,0.0552688204,0.0512448139,0.0143776154,0.1753867716,-0.3614476025,0.0851456895,0.0760480314,0.1644966453,-0.302610606,0.5070192218,-0.1310983002,0.1275262684,-0.223049894,-0.2079615146,0.0001339324,0.2439561933,0.1862295568,-0.3947822154,0.1879740953,-0.1656561196,-0.0192317199,-0.5875660777,-0.1071892381,0.2003776282,-0.170941025,-0.115052484,-0.0000970101,0.0429436676,0.1681482643,0.1277318895,-0.2890504897,0.1069646403,-0.120312646,-0.0588310845,0.2786584795,-0.1375912279,-0.0829643011,-0.2241288424,0.1421548277,-0.2243094891,0.1030538753,0.3524736464,0.3387937248,0.3524858356,-0.308524549,0.4340732992,0.1373018026,-0.0117046619,-0.2752700746,0.0451889001,0.03301147,0.0157014728,-0.3985555172,0.5405015945,-0.0640977994,-0.0079533244,-0.0382868648,-0.1773437709,0.39336285,-0.2694500983,-0.3061868548,0.0869025439,0.1984508485,-0.1947709918,-0.0623507462,-0.4192020297,0.2383095324,0.0467297658,0.092314519,0.267162919,-0.0586765446,0.2318628281,-0.5527657866,-0.4114420712,-0.1109345928,0.0838955641,0.141981557,-0.1045450866,0.0836365968,-0.0565772541,0.3543232083,-0.0869212896,0.1727447212,-0.0558110848,-0.2580604255,0.0144453058,-0.0050552534,-0.5017730594,0.0999640524,-0.0190571472,-0.0335977115,0.2049784958,0.0802929103,0.1039300784,-0.2046397924,0.0635428876,0.1091244742,0.1234012544,0.3510047197,0.0623871833,0.2597543895,0.2474015653,0.2432095259,0.3043641746,0.3376289904,-0.2518724203,-0.2752173841,0.1266618371,0.354028672,-0.2346625477,-0.3460925817,0.0034453757,-0.529618144,0.014235273,0.0144802872,0.1092815399,-0.0006176675,0.120587036,0.0353117809,0.031737823,-0.3619567454,-0.3578194976,-0.0544233955,0.1079839692,-0.0906823501,0.1193413883,0.4057036042,0.0434084795,0.160461843,-0.0342020579,0.0933236852,0.0405896194,0.3698310554,0.0246248916,0.3463379741,-0.0575681292,-0.1173797548,-0.2100870907,-0.2158712,-0.1420142204,-0.187733829,0.1528857052,0.093543373,-0.2084829807,0.0526691414,0.3505088985,-0.4395667315,-0.0142180864,0.0917818695,0.234715119,-0.3213298023,-0.0450621657,0.1401267648,-0.1745882481,-0.1345764399,0.0738597959,-0.1733458936,0.2905427814,-0.0970522314,-0.0834150985,-0.0399765298,-0.0957051069,-0.2075757682,-0.0241719633,-0.4004933536,0.175130114,-0.2039684802,-0.0671897158,0.3619686365,-0.2131475359,-0.1023666561,0.2003478259,-0.2316076308,0.0169600472,-0.1434776038,0.0715370998,-0.1622181386,-0.4502269328,-0.4792477489,0.0716374069,0.0601606518,0.1433715969,-0.4493593276,-0.4256094694,0.4917645156,0.0457842834,-0.0530646294,-0.135975644,-0.1581977159,0.2587842047,0.4374871552,0.0289078746,-0.039322935,-0.2383409292,0.208314389,-0.3265894055,-0.2815673053,0.007924635,-0.1326745749,-0.1672874242,-0.0834711641,0.1869811565,0.4207233787,-0.282923907,-0.0900661424,-0.0967865512,-0.1517234743,-0.0977611616,0.2900044918,-0.1413697451,-0.0983162299,0.1374544352,0.0965314135,-0.1579170376,-0.1721680909,0.0381547734,0.0207199156,-0.1765452772,-0.0530540943,0.1011116728,0.311833322,0.3026477993,0.0712188408,0.0607295074,0.0764993355,0.1724077314,0.0663416758,0.1049052998,0.2324662656,0.2782417536,0.0744726136,-0.4416604936,-0.0143511454,0.1355341226,-0.1512922794,0.2366539538,-0.3167858422,0.2563723028,0.1614443213,-0.0726582259,0.0031077946,-0.0295037497,0.0489318892,-0.3398030996,-0.1381757557,-0.3378735483,-0.0264726914,0.2268940508,-0.1862310469,0.23691912,-0.3378424644,0.1711694896,0.0964340344,-0.0159740672,-0.0260867719,0.1054835916,0.2049351484,-0.3523616195,0.2360577136,0.0762561411,0.1091933623,0.3495015502,0.2247239202,0.1718702167,-0.2100956887,-0.2726424932,0.3151494265,0.3302741945,0.2265132964,-0.0248425025,0.2463130504,-0.1666455418,-0.2424008995,0.0709528029,-0.1072883829,-0.0506054126,-0.373771131,0.0248348601,-0.1493524015,-0.5540050268,-0.169296369,0.1952503771,-0.0923483521,-0.2506904304,0.5380782485,-0.0282843281,-0.3674877882,0.2202184647,-0.2902729809,0.6584700346,-0.209738642,0.2230773717,-0.0658226833,-0.0208164733,-0.3691449761,0.3089107573,0.1798364967,0.0608340017,0.5227045417,0.1013828963,-0.1756847501,-0.1762022972,-0.2720102966,0.0865088925,-0.0329688936,0.3253640532,-0.0012539041,-0.1657108963,-0.0476097576,-0.1483738869,-0.4186193645,-0.2240441442,-0.0004589475,-0.1523243785,0.145780012,0.3848961592,-0.2410242856,-0.4645642042,0.0017174884,-0.219922483,0.0201602615,-0.0149705894,0.2074321657,0.2041841447,-0.2656361461,0.0050036875,-0.1401869059,-0.0535539649,-0.1193710938,0.0445018746,0.0123348702,-0.2702019811,-0.2807666957,0.0192462318,-0.2571010888,0.1463605911,-0.1737017184,-0.6786448359,-0.3268862367,0.1829380095,-0.0971905813,-0.045817595,-0.5455365777,-0.1140745282,0.175135538,-0.2621878386,0.1255510002,0.0865338743,0.25623402,0.0054035732,0.145359993,-0.1071774065,-0.1061482131,0.2679671347,0.1563165933,0.2747744024,0.1175215244,-0.0072315801,0.0892388672,0.2606086433,-0.0087560881,-0.0351131782,0.0462686606,0.1289527714,0.0156991538,0.3321427405,-0.1192228422,-0.3133612275,0.0518869646,0.0816458911,-0.2134507,-0.4037640989,0.149780646,-0.4017365873,0.1390127987,0.1218221858,0.0431131236,0.0612428375,-0.574898243,0.0587844588,-0.1971301585,0.2268818915,-0.0664825216,-0.2071045488,-0.0136967618,-0.2750267982,-0.0112305731,0.1462321579,0.38458848,0.0106431991,-0.2336485982,0.2442918718,0.1683441401,0.2181381136,0.0617300943,0.2018915713,0.130258292,-0.1236294284,-0.3014467657,0.187930584,-0.219289571,-0.0962334946,0.1576205641,-0.0271109454,0.0297204517,-0.2470651716,0.3698942363,0.1215469912,-0.2316070199,0.0246657003,-0.0132846907,0.010629992,-0.0371820591,0.2463866025,-0.1701394022,-0.0281224865,-0.3634008169,-0.1015849486,-0.0245298259,-0.2028280497,0.3764393926,0.2882789075,-0.0159148816,0.2726744115,0.0973345041,0.2122037262,0.0391569249,-0.1911202967,0.1898477077,0.2478585243,0.3424509466,-0.0448404215,-0.2092552036,0.3474107087,0.4122868478,0.1390337199,0.0013973122,-0.3875182569,0.0572364852,-0.2808183134,0.2900633514,0.074675709,-0.1746477634,0.0538317934,-0.1256579608,0.2436386049,-0.1218832061,-0.3639516234,0.0006386196,0.1159496456,-0.1826442331,0.2092995942,0.4568059742,0.7854480743,0.0453129485,0.1997020096,0.4722804725,-0.157801792,0.4363840222,-0.1543487906,0.2508734167,-0.1341299117,0.0953466743,0.0595317557,-0.1178391725,0.1119930819,0.2078066021,0.0020549328,0.0006366753,0.262265861,-0.358350724,-0.1193865836,0.321534425,0.046340175,-0.0212350655,-0.3037971556,0.2726382613,-0.1611576527,-0.2050421983,0.1078737453,0.0280863065,0.3095526993,0.045405779,0.1852248907,0.1358149499,0.1121900305,0.1330635697,-0.1372622401,-0.1884017587,0.3965903819,0.2354937643,-0.1709231883,0.124363482,-0.0892935097,-0.0125479931,-0.1636732072,0.2965009212,0.2138888538,0.2833036482,0.3150731623,-0.1993123144,-0.1702395827,0.2649152875,0.1970163733,0.0597237311,-0.0826280713,-0.0058592591,0.2142528743,-0.4721771181,0.0323734991,0.0392038673,0.1139521897,-0.1505123228,0.2872461081,0.0440129153,-0.038295418,0.0432287306,0.3496758044,-0.2304551601,0.1006510705,0.0579811111,0.0989366993,0.1047594324,0.4397662282,-0.1681389064,-0.2487066984,-0.4134775996,0.2597318888,-0.1213966832,-0.2759042978,0.1033056527,0.1241275892,-0.0599198453,-0.0365880281,0.2192915082,-0.1041616946,-0.0069426321,-0.098561883,0.0047284244,0.0028258294,0.1621439159,-0.1353156567,0.2648953497,0.2367120683,0.1758651584,-0.0797419325,0.1682097167,-0.4586578012,-0.2357046008,-0.362959832,0.1697924882,-0.3387686312,0.1633306593,0.0259765554,0.0013717579,0.1311241835,-0.258605659,-0.2390023321,-0.3924417198,-0.1155549064,0.0577706397,-0.0145451482,-0.1714858711,0.0207366757,0.0118695917,0.1092912853,0.056947846,-0.110956572,-0.3612701595,0.1871643364,0.6549928188,-0.2263977826,-0.3768440485,0.0480532497,0.0021644807,-0.3364017606,0.1804554313,0.192345351,0.256988138,0.0788681358,-0.07736139,0.0921177119,0.1420775652,-0.003899521,0.2430170178,-0.220794633,0.0561312884,0.0168516748,-0.0299333967,-0.0258929785,-0.0279734582,-0.3015801907,-0.2945787609,0.2904886305,0.3381888568,-0.3103620708,0.0277291834,0.3251906931,0.0484479293,0.0795861855,0.2135294527,0.0804045424,0.0296844188,0.2951496243,0.0309762601,-0.0000426526,-0.3247303963,0.0886060297,0.1646475494,-0.4191390574,-0.2750465274,0.1233973876,0.1220628396,0.17781201,-0.0193050969,0.1929826587,0.6014680862,-0.0429051891,0.3327935636,-0.0291351099,-0.0498771705,-0.0677984282,0.0063619693,-0.1729627252,0.0800591707,0.3721141219,0.3768928647,0.2283657938,-0.120360896,-0.0646763667,0.0485597439,0.0171714816,0.1723565459,-0.2735921741,0.119028613,-0.0903881565,0.1302610785,0.070800148,-0.2096399069,0.1096053421,0.1186121255,-0.1082659587,-0.1014227867,0.1704086214,-0.0365316905,0.2522586882,-0.0929841772,-0.1904485971,-0.0768479183,0.3444375694,0.1669953912,0.3699306846,-0.1490647197,-0.1844051182,-0.0260890443,-0.1887125522,-0.0996549502,-0.0091504054,-0.306666255,0.2665651441,0.0757513121,0.1416874975,0.2545981109,0.2173605561,-0.0904578939,0.0599958785,0.3468178809,-0.2202679962,-0.2666848898,0.2640348077,0.5363553166,-0.3322301507,-0.06163086,-0.1159566864,0.0598429292,0.1073192954,0.2752534151,-0.2412468642,0.0973111391,-0.2812126875,0.1597730517,0.277683109,0.4304797053,0.005765608,0.1721989512,-0.0926213041,-0.072695829,-0.5211905241,0.1836291701,0.2208982557,-0.1433753669,0.2590455711,-0.0267201532,-0.1825477928,0.2621278763,-0.3430914879,0.1963402778,-0.0314461999,-0.2231308818,-0.2514567971,0.0356398523,-0.0318681635,0.2519475818,-0.1788478792,-0.1040058881,0.2029329389,-0.1882822812,0.1587802023,0.0695596114,0.0849523544,0.142560035,0.2857635021,0.2812420726,0.169296205,0.1204695627,-0.0395597592,0.0959666893,0.0005993018,-0.0189062897,-0.184280023,-0.205519557,0.0990809128,0.1971745938,-0.0958144516,0.2707025707,0.1259957105,0.1675496995,-0.0273809135,0.2204625458,-0.1136440113,0.164774999,0.0328160301,0.2928231657,-0.128454268,0.2289755642,-0.0958422422,-0.1429787725,-0.0789321959,-0.1923187375,0.2561528087,-0.2987466156,-0.0473077893,0.0288886614,0.372743845,0.0096922247,-0.3118504286,-0.4535419643,-0.2849321067,0.245809108,-0.0828377306,0.2426332235,0.440495193,-0.1116044223,-0.1418990344,-0.0897662118,0.3100312948,0.073642917,0.1364199668,0.0710305721,-0.3468836248]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/206","title":"[Question] Combine 2 datasets which have the same columns","comments":"We are thinking about ways to combine datasets for T5 in #217, feel free to share your thoughts about this.","body":"Hi,\r\n\r\nI am using ``nlp`` to load personal datasets. I created summarization datasets in multi-languages based on wikinews. I have one dataset for english and one for german (french is getting to be ready as well). I want to keep these datasets independent because they need different pre-processing (add different task-specific prefixes for T5 : *summarize:* for english and *zusammenfassen:* for german)\r\n\r\nMy issue is that I want to train T5 on the combined english and german datasets to see if it improves results. So I would like to combine 2 datasets (which have the same columns) to make one and train T5 on it. I was wondering if there is a proper way to do it? I assume that it can be done by combining all examples of each dataset but maybe you have a better solution.\r\n\r\nHoping this is clear enough,\r\n\r\nThanks a lot \ud83d\ude0a\r\nBest","comment_length":20,"text":"[Question] Combine 2 datasets which have the same columns\nHi,\r\n\r\nI am using ``nlp`` to load personal datasets. I created summarization datasets in multi-languages based on wikinews. I have one dataset for english and one for german (french is getting to be ready as well). I want to keep these datasets independent because they need different pre-processing (add different task-specific prefixes for T5 : *summarize:* for english and *zusammenfassen:* for german)\r\n\r\nMy issue is that I want to train T5 on the combined english and german datasets to see if it improves results. So I would like to combine 2 datasets (which have the same columns) to make one and train T5 on it. I was wondering if there is a proper way to do it? I assume that it can be done by combining all examples of each dataset but maybe you have a better solution.\r\n\r\nHoping this is clear enough,\r\n\r\nThanks a lot \ud83d\ude0a\r\nBest\nWe are thinking about ways to combine datasets for T5 in #217, feel free to share your thoughts about this.","embeddings":[-0.2445141971,0.0638994128,0.0241108686,0.3628230393,0.1096224189,0.4387528002,0.269988209,0.3603741229,0.0449791066,-0.1164474189,-0.3814806938,0.1237424836,0.0036282265,0.2719849944,0.271317482,-0.5642143488,0.0580447242,0.2187532783,-0.5200107098,0.1850373447,0.0075754696,0.005126602,-0.1654464453,0.0126771564,-0.2030433267,0.2349442244,-0.3422628641,-0.227939114,0.0463839993,0.1566460878,0.3019719422,0.1670797765,0.2146203518,0.3149155974,-0.0001199274,-0.0794955119,-0.0164892916,-0.0905234143,0.1418972611,-0.1987285167,0.0968255997,-0.3619499803,-0.0633546039,-0.156581834,0.0822774172,-0.2395310551,-0.2331836522,0.0375286452,0.2680856884,-0.1660301238,0.0541115366,-0.1883206815,-0.251149863,0.3409512341,-0.0712746456,-0.1605684906,0.0614655875,-0.0694343448,0.0818909332,-0.3195155561,0.577925384,0.2133333832,-0.0351103805,-0.1198773012,0.0823331848,0.1424835175,0.1433921158,-0.1095088795,0.0553349555,0.1900027841,0.7741910219,-0.059649311,0.008375858,-0.2915332317,0.4038740695,-0.1315465271,-0.1932114661,0.143045038,0.0170429479,0.0254980847,-0.2548200786,-0.1563202739,-0.016406402,0.3023144901,-0.0431645103,0.331581831,0.1102488413,0.2770792842,-0.0637215301,-0.1514382958,0.1660322398,-0.3364259899,0.0009921413,0.1547577828,-0.3073634803,-0.4046192765,-0.3330450058,-0.1842289269,0.2726255059,-0.1534878612,-0.1543884575,-0.0069713704,-0.1090398356,0.138723433,0.3026831448,0.1582723111,0.1386101842,0.2125378847,-0.1884720474,-0.3162833154,-0.3603146374,0.1616100669,0.0701176077,-0.0728297979,-0.1233948544,-0.2811374664,-0.0584523194,-0.3414462209,0.1455402374,-0.1935318559,-0.2755381167,-0.2310825884,-0.1025671437,-0.1516104788,-0.0181961898,0.3452565074,-0.1489931345,0.3091166914,-0.1060740799,-0.0139371408,-0.0250301678,0.0738149956,-0.1707864255,0.1282698512,0.0945632681,0.1703294069,-0.1308951527,0.4186178148,-0.1831040531,0.1015597805,0.3208992779,-0.1995742023,-0.0710753202,-0.2422419488,0.3129127324,0.163817659,0.1630841047,0.1007980108,-0.3006671667,0.1673175842,-0.1736656725,-0.0824754238,0.0157322194,0.0594997555,-0.0839434862,-0.3012195528,-0.1238102093,0.6368827224,0.2376348078,-0.3169041872,-0.1415876597,0.0955535248,-0.464184165,-0.0478923246,0.20856677,0.0251198672,-0.3501130342,-0.13585338,0.0010477458,-0.0063651754,-0.0501091294,0.2889217138,-0.2664492428,0.2065665275,0.09581846,0.0672085956,0.6602528095,0.0170827769,-0.2144925296,0.2215339541,-0.1507163346,0.132588014,-0.104750745,0.1174124926,-0.376698494,-0.0470037572,0.1844806373,0.8385096192,-0.2917749286,-0.1561029702,-0.0637751222,-0.3239151537,0.3379637003,0.1553742886,0.1021537781,-0.2498816997,-0.142595768,0.1694810838,0.1662994921,-0.3560666442,0.1682165414,0.1871119142,-0.0291251838,0.117339924,-0.227634564,-0.2857426107,-0.265973419,-0.1344855428,0.1124017239,0.0476165377,0.4460558593,-0.4509105086,0.0663456917,-0.6603160501,-0.0207441468,0.0910769999,-0.0005953129,0.3114060462,0.1058811396,-0.2892328501,-0.175703615,0.3350967765,-0.052014336,-0.1666298807,-0.1528071761,0.4420305789,0.0385310762,0.0737517253,0.0093516335,0.5092473626,0.048488263,0.034645427,0.1494198889,-0.1253807247,-0.5496708155,0.1898004413,0.1835702062,0.0285738744,-0.1553871185,-0.0680693835,0.3543197811,-0.3344199359,-0.005318779,-0.160242483,-0.0741474032,0.4110820889,-0.248981297,0.3699252307,-0.0203907881,-0.117495209,0.0169245657,-0.1534184664,0.0219020154,-0.1019337028,0.2140550166,-0.0077626077,0.0135509977,0.3495665789,-0.3878057897,0.246159941,0.2839236557,0.1036929712,-0.0467015766,-0.0142829837,-0.1404851228,-0.0473917983,0.0349292904,0.1737324595,0.5730592608,0.1965861171,0.2018268853,0.2332470417,-0.154273361,0.1257927865,0.0148142884,0.1872755885,0.0660852492,0.3079122007,0.0905814096,0.2213162035,0.1897095889,0.3987439275,0.2033168674,-0.0915317163,-0.1926696301,-0.0057482198,-0.2850170135,-0.6398733854,-0.5628804564,-0.128213644,-0.2211208194,-0.0530706234,-0.0937654004,-0.005285013,0.0087295771,0.1844754219,-0.1145416275,0.4212614596,-0.4750213623,-0.2778398693,0.2787618935,0.0416234657,-0.1818008274,0.070905216,0.407938838,0.2121359855,0.2545477152,-0.0574187525,-0.2327980399,-0.0601055101,-0.5316917896,0.0501952358,-0.3654118776,-0.1488939524,-0.2518881261,-0.1015012488,-0.2149767727,-0.3766777217,0.051559221,0.3118179142,0.052741304,-0.1287004203,0.0572642796,0.2869411707,-0.1050098315,-0.294531405,-0.4214810133,-0.059875574,0.0423832871,-0.1619119942,-0.0208190884,-0.3642593622,-0.0539556891,-0.0365090817,-0.2672071159,-0.011379499,-0.201623857,0.0008073317,0.1787796766,-0.1347468495,-0.1283462644,-0.0846113712,-0.1897967458,0.299744308,-0.0308160167,-0.1024918258,-0.151703611,-0.1399434954,-0.0023926657,-0.1428914666,0.3774222732,0.178900227,0.0176718123,0.1485676914,0.0718961582,-0.2294689864,0.0211154018,0.098553285,0.4574983418,-0.1967522502,0.2529270947,-0.1457327604,0.607717514,0.5213459134,0.166951254,0.0660566986,0.2555118799,0.1424033791,0.0633199066,-0.2097760737,0.2008449882,0.1946293563,0.0576697513,0.2699237168,0.1037869453,-0.053351894,-0.333976239,0.0900695026,-0.304264009,-0.117387034,0.3777521253,-0.2725746632,0.1191146895,0.011026945,-0.4641563594,-0.3556422889,-0.1309647113,-0.2164750248,0.1183902472,-0.2585560381,0.1361858994,-0.562223196,-0.1036447287,-0.1406893581,0.3238658011,0.1514757127,0.3083507419,0.0497300662,0.0550473519,-0.246095553,-0.0438207537,0.0648694634,-0.4924115241,-0.1347339153,0.0649792552,0.0140513182,-0.2310789824,-0.1758926511,0.0173759013,-0.0913957804,-0.182997942,0.336460501,-0.3184767962,-0.1562388986,0.6944687366,0.4387214482,-0.0757842362,-0.3850885034,0.0464348271,-0.12323156,-0.1433954835,-0.2600070834,-0.0475106649,0.1632035375,-0.3689458072,0.1504460573,0.0164117515,0.0803885087,0.4504481852,0.1033263877,0.0721364096,0.3526634276,0.3876460195,-0.240286991,-0.0815341622,-0.2526184916,0.4975561798,-0.075580731,-0.3227154016,-0.1711561978,-0.1438244879,0.7155781388,-0.0685486197,0.151131615,0.2316921651,-0.399597764,-0.3615083694,0.0019530731,0.202846542,0.2826134264,0.3221149445,-0.397128284,-0.4497555494,0.6233006716,0.410638988,-0.0207189061,0.2541218996,-0.2573800683,-0.1741577387,0.0994557142,-0.1014834642,0.8141262531,0.0218137987,0.381593734,-0.0004789964,-0.1681197435,0.7002717257,-0.0966337547,0.019921869,0.0201725513,-0.1437635422,-0.0353861079,-0.0300958548,-0.1292168647,0.3702215254,-0.2077860981,0.3171710372,-0.0004518146,-0.00524273,0.0035988514,0.4091910124,0.0666945279,-0.258772701,-0.3931368887,-0.0303383805,-0.3027257323,-0.0538799465,-0.1444964111,0.0714733154,-0.1313787103,0.0704780221,-0.1482685357,0.2999644876,0.5066033006,0.044584468,-0.4117031693,-0.4415704012,0.2938781977,0.14138107,0.3556128144,0.0290898737,0.0299527086,0.014914426,-0.3565639257,-0.0157545581,-0.1148370355,-0.2938964069,0.2594220638,0.0201874431,-0.2533400059,0.0690134317,0.4127650261,-0.0823659301,-0.2544031143,-0.1785518527,0.3314886987,-0.0594611317,-0.5873777866,0.6571654677,0.1512079239,-0.0779192224,0.0100546209,-0.1137165502,-0.289228797,0.3845161796,-0.2728395462,-0.2554663122,0.0601836555,0.1803910136,0.3060005009,-0.3658264577,0.4701863825,0.1839188188,0.2744098902,0.0229732934,-0.0615515746,0.2622483969,-0.0737726316,-0.3581317663,0.1025350615,0.2045335323,0.1125623211,0.1527302414,0.163315922,0.1222746521,0.2067320198,-0.0677797198,-0.3782422543,0.2728558779,0.2002412975,0.2617895901,0.2570177019,0.3128030002,0.0644642487,0.6031781435,-0.1696951985,-0.0190978311,0.1362243593,0.2852416635,0.3465159237,0.084506996,-0.4070691764,0.014368739,-0.0442210995,0.0018225503,0.1141061559,-0.1442724764,-0.0768620968,0.1759717613,0.3449318111,0.0011230757,0.0602332763,-0.2807407081,-0.0118225906,-0.0225088745,0.0665124655,-0.0925774872,-0.0002626651,0.380640924,-0.0736532807,0.0340386964,-0.2856475115,0.2642692029,-0.132349968,-0.0356788896,0.259003669,0.1017374694,0.3916665018,-0.0444830991,-0.2572684884,-0.2678640783,0.2171400487,0.0650550872,-0.3455165029,-0.0349697173,0.0460911654,-0.061143484,0.1521234363,0.3250306547,-0.1072149351,-0.0250685439,0.2345823497,0.0007481967,-0.1970007569,0.0721049085,0.2305070013,-0.0640268698,-0.1415854543,0.2195847332,0.0906120315,0.3772614896,-0.3111916482,-0.0127991037,-0.0074813496,-0.0360482559,0.3243546486,0.5464568734,-0.0324210487,-0.1031251922,-0.0120026413,0.0809040144,0.0588643588,-0.0501781218,-0.1465632468,-0.0674618408,-0.0896897167,-0.0796321109,0.2312423885,-0.082683675,0.0639867336,0.2914494574,0.2242588103,0.1321956217,0.0598770194,0.2759810984,0.0057381787,-0.4656766653,0.027942745,0.4166921377,-0.0797219276,0.0029161859,-0.3966145515,-0.0038867351,-0.1403136253,0.2057416141,-0.124699533,-0.3338140249,0.5012153983,0.092915982,-0.1495987773,-0.2275858521,0.126692608,0.5161340833,-0.0350954272,-0.0484623164,0.0520229116,0.1654694527,-0.1280072331,0.0376996361,0.2517570555,0.1108668894,0.0293356422,0.0011414289,0.0243029371,0.140440315,0.1221234947,0.1034303308,0.0191776343,0.0841458067,-0.0162049141,0.2771762013,0.1007976755,-0.0612561554,0.0731330439,-0.1236757562,0.0440761894,-0.5888588428,0.2657965124,0.3155158758,-0.0237797014,-0.0525276028,0.1717575788,-0.3293572664,-0.1878270209,0.5251409411,0.2926934361,-0.1560959071,-0.1533704549,0.0353381671,-0.029739935,0.3186296821,0.3571055233,0.0833356082,-0.4211035371,-0.1852588654,-0.7274476886,0.1118840501,-0.1998965442,-0.1187604144,-0.2600693107,0.0404853746,0.2637618184,0.2332361788,0.0643844381,0.4323761165,0.1418526471,0.1557846367,-0.4341234565,-0.1574389189,0.0854790807,0.186892733,0.0597092994,-0.1115175337,0.0242311712,0.354624927,-0.1866978109,-0.063682951,0.0197881311,0.1336735934,-0.0402191021,0.2397711128,0.0368367061,0.6354696751,-0.1660688519,0.0947615281,0.1124468371,-0.0509695485,-0.5395126343,0.3559369743,-0.1595758051,0.0192934517,-0.6011698246,0.6696316004,0.1246214136,-0.2897533178,-0.0893223211,0.1219838113,-0.053984534,-0.1360832304,0.2872998118,0.0695642084,0.2895839214,0.623124063,0.0448061936,0.0836290792,-0.1801166832,-0.0981299654,0.2389071286,0.1017700955,-0.4384738803,-0.0993013531,-0.1474408954,0.0205277018,-0.0874537453,-0.5327376127,-0.0361309536,0.3234039247,-0.1769235283,0.3889918923,0.5019960403,0.0166903511,0.0657179207,-0.2933852673,-0.0699704736,-0.0717824101,-0.0772056356,-0.1236256137,-0.2613339424]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/206","title":"[Question] Combine 2 datasets which have the same columns","comments":"Ok great! I will look at it. Thanks","body":"Hi,\r\n\r\nI am using ``nlp`` to load personal datasets. I created summarization datasets in multi-languages based on wikinews. I have one dataset for english and one for german (french is getting to be ready as well). I want to keep these datasets independent because they need different pre-processing (add different task-specific prefixes for T5 : *summarize:* for english and *zusammenfassen:* for german)\r\n\r\nMy issue is that I want to train T5 on the combined english and german datasets to see if it improves results. So I would like to combine 2 datasets (which have the same columns) to make one and train T5 on it. I was wondering if there is a proper way to do it? I assume that it can be done by combining all examples of each dataset but maybe you have a better solution.\r\n\r\nHoping this is clear enough,\r\n\r\nThanks a lot \ud83d\ude0a\r\nBest","comment_length":8,"text":"[Question] Combine 2 datasets which have the same columns\nHi,\r\n\r\nI am using ``nlp`` to load personal datasets. I created summarization datasets in multi-languages based on wikinews. I have one dataset for english and one for german (french is getting to be ready as well). I want to keep these datasets independent because they need different pre-processing (add different task-specific prefixes for T5 : *summarize:* for english and *zusammenfassen:* for german)\r\n\r\nMy issue is that I want to train T5 on the combined english and german datasets to see if it improves results. So I would like to combine 2 datasets (which have the same columns) to make one and train T5 on it. I was wondering if there is a proper way to do it? I assume that it can be done by combining all examples of each dataset but maybe you have a better solution.\r\n\r\nHoping this is clear enough,\r\n\r\nThanks a lot \ud83d\ude0a\r\nBest\nOk great! I will look at it. Thanks","embeddings":[-0.2462292016,0.0392308757,0.0202458072,0.3757518232,0.0873391852,0.4316775501,0.2603235841,0.3487795293,0.0517877601,-0.1122787371,-0.3563239872,0.1305461526,0.0073037283,0.2410883904,0.2701638639,-0.5845234394,0.0603629462,0.2350384593,-0.5094716549,0.1664738953,0.0056987433,-0.0168072358,-0.1949364692,0.048766423,-0.1879769266,0.2189734429,-0.3615565002,-0.228071183,0.0267051812,0.1545997113,0.3152414858,0.1468645781,0.2136625797,0.2993278801,-0.0001199382,-0.1017906219,-0.0215143245,-0.0818821341,0.1459322274,-0.2322112024,0.0839812607,-0.350405395,-0.0816048384,-0.1919169724,0.0722385049,-0.2187928259,-0.2384815812,0.0116234636,0.2671406865,-0.1720696837,0.0526122078,-0.2048604935,-0.2586128712,0.3155318499,-0.1063720733,-0.1548189968,0.0952617973,-0.048919782,0.0771162882,-0.3341005147,0.5842083097,0.2313624173,-0.0565949939,-0.1173531935,0.0629756749,0.1496124715,0.1336199939,-0.103453055,0.0542995669,0.1953942329,0.7345303297,-0.0679181963,0.0198912323,-0.2852583826,0.4254706502,-0.1215930283,-0.2052625418,0.156746164,0.0465959683,0.03628885,-0.2760698497,-0.1741163731,-0.0263090581,0.3009400368,-0.0289761908,0.3022716939,0.0963023081,0.2853766978,-0.0461376719,-0.1517715454,0.133155942,-0.3463117182,0.0015270703,0.1918369532,-0.2989035249,-0.4105059803,-0.3594518602,-0.1636505276,0.2609563768,-0.1377024949,-0.163616389,-0.0152862994,-0.1014450863,0.1785591096,0.299649775,0.1808148175,0.1670840383,0.2227442414,-0.1945527196,-0.308277905,-0.3622082472,0.1526679248,0.0498707071,-0.0499696992,-0.1119328365,-0.2710192204,-0.0919912383,-0.3342358768,0.1159588769,-0.1968550831,-0.276239574,-0.2365582585,-0.0679640546,-0.1234659404,-0.0190139171,0.33430776,-0.1593435258,0.3117427528,-0.0908356011,-0.0191352088,-0.0175026748,0.0943369269,-0.1885748655,0.1310970038,0.0552119426,0.1970192045,-0.1399723291,0.4155315161,-0.212935403,0.0826032534,0.326190412,-0.2172883153,-0.0519856103,-0.2649978399,0.3288455606,0.1679938734,0.1753892452,0.0698832124,-0.316290319,0.1673628241,-0.1988350451,-0.0797864944,0.0298992768,0.0650916249,-0.0711773708,-0.2831112146,-0.1363645345,0.6260247827,0.2307889909,-0.3304328024,-0.1205610707,0.1276813895,-0.4425558448,-0.0430400893,0.2091929913,0.0130166085,-0.3272294402,-0.1453946978,-0.0028809919,-0.0244484227,-0.0239759814,0.29794994,-0.2876935303,0.2288613021,0.0785391927,0.0840497166,0.6719129086,0.0320553258,-0.1957992017,0.2361942828,-0.15388982,0.1131294146,-0.0966633558,0.1168924123,-0.3593547344,-0.0410213433,0.1884824783,0.825480938,-0.2889354229,-0.1601201445,-0.0331963822,-0.3129266798,0.3230085671,0.1292878091,0.0886749104,-0.2022413909,-0.1348843873,0.1589255482,0.1823943406,-0.3419157863,0.178133741,0.2000879943,-0.0225500613,0.1100395918,-0.2569380999,-0.3087802529,-0.2221467793,-0.1579959393,0.1091386005,0.0247285422,0.4608998895,-0.4694780111,0.0531064309,-0.662129581,0.0013567019,0.1153621599,-0.0072078286,0.2931179106,0.1130413488,-0.3135974407,-0.1794667542,0.3638904393,-0.0659905896,-0.1725172102,-0.1710187942,0.4435697198,0.0317832008,0.0689671412,-0.0013820541,0.5118144751,0.0785241127,0.0301082991,0.1761602014,-0.1073080897,-0.5519584417,0.192995742,0.2123714983,-0.0115076415,-0.1701125205,-0.0639483184,0.3474167287,-0.354462564,0.0052786153,-0.1633695513,-0.0596100129,0.4179919958,-0.2213298827,0.4004319012,0.0010494922,-0.1250783354,0.0133685172,-0.1977724582,0.0275096912,-0.1000122353,0.2149534971,0.0132984407,0.0111059556,0.3463858962,-0.4032680392,0.2088038325,0.2665798366,0.0848960802,-0.0277324133,-0.0076080831,-0.1623271406,-0.0418510586,0.0470944084,0.1752145737,0.58192873,0.1945769042,0.2082436085,0.2439998686,-0.1587823033,0.1356320232,0.0087247193,0.1720085591,0.0778866559,0.2854037583,0.072208181,0.2139060646,0.1967326254,0.4000567794,0.1956569254,-0.0758016706,-0.1914799809,0.0141251693,-0.275908947,-0.6194735169,-0.5686379671,-0.1571910679,-0.2171020657,-0.0315489657,-0.1166099533,0.0020814489,0.0301658399,0.1797571927,-0.0999584943,0.441269815,-0.4759100974,-0.2850240171,0.275912255,0.0351756588,-0.1905329525,0.0711721405,0.3803145885,0.2260341942,0.2809574008,-0.0608219318,-0.2223816514,-0.0438821353,-0.5078878999,0.0274960678,-0.3211906254,-0.1652503461,-0.262701273,-0.1068482101,-0.2304242253,-0.3661924005,0.0562987924,0.301112771,0.0584581904,-0.1286415011,0.0660206303,0.2875005007,-0.0934300646,-0.2876338363,-0.4138934612,-0.0524945967,0.040002998,-0.1794305444,-0.0121972533,-0.4105261862,-0.0420894288,-0.0427131914,-0.2836411893,-0.0078490321,-0.2235364318,-0.0100143487,0.1864641756,-0.1121737882,-0.1284169555,-0.0901574567,-0.2208474874,0.300922364,-0.0442261882,-0.0890913457,-0.1355901212,-0.117485866,-0.035051737,-0.1481572092,0.4098297954,0.1798736602,-0.00141783,0.1359209567,0.0969908237,-0.2357369661,0.025294764,0.0719139576,0.4825357497,-0.2080997825,0.2561088204,-0.1358672082,0.5650169253,0.5228729248,0.1520342082,0.0557780303,0.2294268906,0.1456075013,0.091883637,-0.2336022407,0.1896544695,0.190536961,0.0453484915,0.2391518205,0.1038420573,-0.0557280369,-0.3467248082,0.0787999555,-0.3039862514,-0.1040312573,0.3871185184,-0.2572231889,0.128237471,0.0165017787,-0.4693047106,-0.3598932028,-0.1010472178,-0.2101020217,0.1398227364,-0.2884911895,0.1008528695,-0.5932107568,-0.1439267695,-0.0864324421,0.3346690834,0.1443274468,0.3028312624,0.0510242954,0.0783386603,-0.2327822149,-0.0427818559,0.0665289611,-0.4741987586,-0.1664965302,0.0562421083,0.043719396,-0.2107628733,-0.1584173143,0.0123191783,-0.0911419839,-0.1428962052,0.3604444265,-0.3158390224,-0.139541924,0.6847633719,0.4351844192,-0.0918107033,-0.4088252187,0.035783086,-0.1007225886,-0.1397768557,-0.228176102,-0.0792923123,0.1476534009,-0.3400196433,0.1398640573,-0.0074750641,0.0866891369,0.4558404982,0.109255746,0.0882800296,0.3551646471,0.3608547747,-0.237872988,-0.0937988758,-0.2465076894,0.4763747454,-0.0764016584,-0.2564037144,-0.1916312277,-0.1240514368,0.737531364,-0.0677308217,0.160190478,0.2516319752,-0.3857941329,-0.3315126598,-0.0440351516,0.1951160282,0.2974019051,0.3225687146,-0.4007861316,-0.4635491669,0.6269599795,0.3972868025,-0.0449303612,0.2501241267,-0.2331086099,-0.1653302759,0.0883352384,-0.0979430899,0.8542217016,0.0312047377,0.363712877,-0.0092869373,-0.1688133925,0.7092227936,-0.0858090818,0.0165944304,0.0021025084,-0.1009423733,-0.0560789369,-0.0514729507,-0.0784147233,0.3968943954,-0.2107149512,0.3320551217,-0.0085168537,0.0076557263,0.0134916333,0.4180110395,0.0753216371,-0.2600958645,-0.4098071158,-0.0334620811,-0.2767522931,-0.0691143498,-0.1324686855,0.0800087899,-0.1358558983,0.0739071965,-0.1377085745,0.3188558519,0.5059434772,0.0337084197,-0.4112940133,-0.4486439228,0.2891825736,0.1239328384,0.3445642292,0.0534640849,0.0401197001,0.0114125982,-0.334543407,-0.0307658538,-0.1332802474,-0.3084891737,0.2550241351,0.0147854388,-0.2755226791,0.0968580171,0.4055237472,-0.0532911569,-0.2059646845,-0.1520273834,0.3557443917,-0.0626813024,-0.5951275229,0.6546921134,0.1803673804,-0.0861011371,0.0093912305,-0.093334876,-0.3212585747,0.4044575691,-0.2682996094,-0.2440486401,0.0715385973,0.1769488305,0.3241307437,-0.3611799479,0.4716184437,0.1211067215,0.2728620768,0.0161509234,-0.0726753473,0.2756584883,-0.0379333831,-0.3772074878,0.1209108531,0.1943822503,0.1022251174,0.0969426036,0.1780990511,0.1646345407,0.1877178997,-0.0781163871,-0.4021426141,0.2571066022,0.2257483155,0.2549962401,0.2728697658,0.2871808112,0.0602159277,0.6250033975,-0.1725282073,-0.0278254468,0.1086073443,0.307844907,0.3561253846,0.0864789262,-0.3947932422,0.0075020068,-0.034382347,0.0026299225,0.1090426221,-0.1483805031,-0.0803895295,0.172063753,0.3327727616,-0.0002610579,0.0587773062,-0.2957621515,0.0103107821,-0.0293428637,0.02358751,-0.0943712667,0.0260126982,0.3711901009,-0.0697615221,0.0140658878,-0.3134975135,0.2526602447,-0.163520664,-0.0283070318,0.2559629381,0.0918010026,0.3915121853,-0.0527362563,-0.2770103812,-0.2725086212,0.2284308076,0.0564163178,-0.3176523745,-0.0262446664,0.0562692359,-0.0693162531,0.1759533882,0.3448542655,-0.0970876887,-0.0136901457,0.2431574762,-0.0004836301,-0.2018051147,0.0962192118,0.2223702967,-0.0962512568,-0.1416448057,0.2256428748,0.0890670419,0.4134961069,-0.3290946186,-0.0371147059,0.0004426221,-0.0059390576,0.2899927795,0.5370304585,-0.0131552704,-0.1049382016,0.0144801335,0.1008051932,0.0624632277,-0.0900856778,-0.1401129961,-0.0410150997,-0.0583412983,-0.0675123483,0.2209893316,-0.0770956799,0.0809085146,0.29317379,0.2306375206,0.0849293843,0.0484522022,0.2842150331,-0.0037265811,-0.4649014473,0.0530022867,0.4135476947,-0.0849256814,0.0034234351,-0.4107018113,0.0094673894,-0.1397847682,0.1823666841,-0.087873131,-0.3362466693,0.5161201358,0.0770550221,-0.1327658296,-0.226989761,0.1147273704,0.5311043859,-0.0566587672,-0.0764045864,0.04348737,0.1400088072,-0.1174498871,0.0481811799,0.2482474297,0.1292911321,0.0474780463,0.0328787155,0.0167909265,0.1432498395,0.0971780419,0.0906220227,0.0252057351,0.1053065956,-0.0059530842,0.2543993294,0.1074028984,-0.0703100264,0.0975648239,-0.1130568087,0.0481755547,-0.5737156272,0.2854653299,0.3069463968,-0.0403773338,-0.1046352834,0.1437233537,-0.3196513951,-0.1744475365,0.5229780674,0.2442848235,-0.1599405855,-0.1390698254,0.0304917656,-0.0163879432,0.3029136062,0.3589748442,0.0979440138,-0.4440992773,-0.2206578404,-0.7411022782,0.1027316824,-0.1995992959,-0.1541931927,-0.2829319835,0.0606674477,0.2568445802,0.2266532779,0.0483119488,0.4031189978,0.1573295295,0.1545588374,-0.4145733118,-0.1442804933,0.0626408532,0.1896729767,0.0659932122,-0.119809173,0.0241468791,0.3777406216,-0.1969908774,-0.0173075907,0.0469020903,0.1074436679,-0.0352429301,0.2467388064,0.0262114909,0.6594129205,-0.1808003634,0.1152167991,0.1525158435,-0.0300367568,-0.5246483088,0.365290761,-0.1306828111,0.0119006205,-0.6042098403,0.6608581543,0.1222017929,-0.2918258607,-0.0785648376,0.0738823563,-0.072966285,-0.1099675223,0.2887264788,0.0722248331,0.2925929129,0.6109800935,0.0610351376,0.0951410457,-0.1635624915,-0.0642807037,0.2527574599,0.0824901462,-0.4435220361,-0.1128600612,-0.1108442321,0.0134846615,-0.1236031801,-0.5434712172,-0.0421106741,0.3444767594,-0.185902223,0.4116524458,0.5014144778,-0.0173007082,0.0613083914,-0.2912843227,-0.0782511607,-0.0834210739,-0.0757092834,-0.1157432124,-0.2653634846]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/202","title":"Mistaken `_KWARGS_DESCRIPTION` for XNLI metric","comments":"Indeed, good catch ! thanks\r\nFixing it right now","body":"Hi!\r\n\r\nThe [`_KWARGS_DESCRIPTION`](https:\/\/github.com\/huggingface\/nlp\/blob\/7d0fa58641f3f462fb2861dcdd6ce7f0da3f6a56\/metrics\/xnli\/xnli.py#L45) for the XNLI metric uses `Args` and `Returns` text from [BLEU](https:\/\/github.com\/huggingface\/nlp\/blob\/7d0fa58641f3f462fb2861dcdd6ce7f0da3f6a56\/metrics\/bleu\/bleu.py#L58) metric:\r\n\r\n```\r\n_KWARGS_DESCRIPTION = \"\"\"\r\nComputes XNLI score which is just simple accuracy.\r\nArgs:\r\n predictions: list of translations to score.\r\n Each translation should be tokenized into a list of tokens.\r\n references: list of lists of references for each translation.\r\n Each reference should be tokenized into a list of tokens.\r\n max_order: Maximum n-gram order to use when computing BLEU score.\r\n smooth: Whether or not to apply Lin et al. 2004 smoothing.\r\nReturns:\r\n 'bleu': bleu score,\r\n 'precisions': geometric mean of n-gram precisions,\r\n 'brevity_penalty': brevity penalty,\r\n 'length_ratio': ratio of lengths,\r\n 'translation_length': translation_length,\r\n 'reference_length': reference_length\r\n\"\"\"\r\n```\r\n\r\nBut it should be something like:\r\n\r\n```\r\n_KWARGS_DESCRIPTION = \"\"\"\r\nComputes XNLI score which is just simple accuracy.\r\nArgs:\r\n predictions: Predicted labels.\r\n references: Ground truth labels.\r\nReturns:\r\n 'accuracy': accuracy\r\n```","comment_length":9,"text":"Mistaken `_KWARGS_DESCRIPTION` for XNLI metric\nHi!\r\n\r\nThe [`_KWARGS_DESCRIPTION`](https:\/\/github.com\/huggingface\/nlp\/blob\/7d0fa58641f3f462fb2861dcdd6ce7f0da3f6a56\/metrics\/xnli\/xnli.py#L45) for the XNLI metric uses `Args` and `Returns` text from [BLEU](https:\/\/github.com\/huggingface\/nlp\/blob\/7d0fa58641f3f462fb2861dcdd6ce7f0da3f6a56\/metrics\/bleu\/bleu.py#L58) metric:\r\n\r\n```\r\n_KWARGS_DESCRIPTION = \"\"\"\r\nComputes XNLI score which is just simple accuracy.\r\nArgs:\r\n predictions: list of translations to score.\r\n Each translation should be tokenized into a list of tokens.\r\n references: list of lists of references for each translation.\r\n Each reference should be tokenized into a list of tokens.\r\n max_order: Maximum n-gram order to use when computing BLEU score.\r\n smooth: Whether or not to apply Lin et al. 2004 smoothing.\r\nReturns:\r\n 'bleu': bleu score,\r\n 'precisions': geometric mean of n-gram precisions,\r\n 'brevity_penalty': brevity penalty,\r\n 'length_ratio': ratio of lengths,\r\n 'translation_length': translation_length,\r\n 'reference_length': reference_length\r\n\"\"\"\r\n```\r\n\r\nBut it should be something like:\r\n\r\n```\r\n_KWARGS_DESCRIPTION = \"\"\"\r\nComputes XNLI score which is just simple accuracy.\r\nArgs:\r\n predictions: Predicted labels.\r\n references: Ground truth labels.\r\nReturns:\r\n 'accuracy': accuracy\r\n```\nIndeed, good catch ! thanks\r\nFixing it right now","embeddings":[0.0913031623,-0.1331153661,-0.1436036229,-0.0117897512,0.473944962,-0.2548961639,0.0263647288,0.0046182205,-0.1071946323,0.2016375363,0.0250016432,0.2616060674,0.214892149,-0.0068727196,0.0780791193,-0.0280460808,0.0708082765,0.0549953878,0.2333020717,-0.3114714921,-0.4147129059,0.2125460804,-0.0851167813,0.2524889112,-0.1056487635,0.0472631566,0.1741354018,-0.0862730667,-0.5392819047,-0.4869367778,0.0916410238,-0.0525091998,-0.1495673656,-0.0804839581,-0.0001014295,-0.2816405892,0.3511400521,-0.124215275,-0.1381129771,-0.0019110686,-0.0010335266,-0.3420173824,-0.0469074622,-0.4117923975,0.0294243675,-0.2060037404,-0.0475532524,0.0261301175,0.046650175,0.3628437221,0.3242207766,0.2128038108,0.0099016307,-0.2303774953,0.1448630095,-0.3139017224,-0.1337860376,0.2891353369,0.0906216279,0.0960275978,-0.2856559455,0.4170868695,0.0395526402,-0.1205523834,0.355122298,0.0535635091,0.5309060216,0.139819622,-0.0162997134,0.0780915841,-0.0756656975,-0.2322330028,-0.3144854605,0.0481019355,0.053854946,-0.4329447448,-0.2117504925,0.0192745589,-0.1706722975,-0.3016397655,-0.2810908854,0.0226574335,-0.1243968531,0.0370333046,-0.1167585179,0.6443428993,-0.1139572561,-0.0201718807,0.3123733103,-0.0668362603,-0.1728663743,0.1095769778,-0.0894211754,0.0193563085,-0.3574930727,0.0852261856,0.3958698511,0.1596740484,0.3733268678,0.1976262182,0.2263416052,-0.014629391,-0.0499329157,0.0266790278,-0.2194540054,0.7317664623,0.3911005259,-0.0445533767,0.083444491,0.112422131,0.2049718499,-0.0954943672,0.1056989878,-0.1478087902,0.0454377867,0.297942251,-0.1795506179,-0.1115980595,-0.3943030536,0.1355581433,0.0782146901,0.0458859242,0.0596374385,0.2413335294,-0.1589093208,0.0312424358,0.0958371907,0.239886269,-0.0376458131,-0.1020236239,-0.3703525364,0.0258100946,-0.2997933924,-0.1015054882,-0.055476822,-0.0693620071,0.5972087383,-0.0496467873,0.2525544763,-0.1644564271,-0.1425579935,0.1384020299,0.1599333435,-0.0525066219,-0.1489494145,-0.0135121681,0.1852320731,-0.2575340271,-0.1666482836,0.0011929292,-0.1843489707,-0.1398599297,0.0886490941,0.3416350186,-0.1163693294,-0.044228211,-0.0305498615,0.1646389961,-0.0968324393,0.091997005,0.1423041821,-0.1476713121,-0.1400141567,-0.2290059328,0.3956643343,-0.1612187773,0.0756656975,-0.1550501138,-0.1212303266,-0.1030737087,0.3014608324,0.1897856742,0.0923919976,0.2719809711,-0.1890566498,0.2941067815,0.4172000289,-0.4714084268,-0.1421042234,0.0921419263,-0.0707140267,-0.3216891587,0.0431051515,0.1712417006,0.1064405069,0.1555982381,-0.0347233303,-0.0196382571,0.1421892494,0.0220257919,-0.2528151572,-0.3819434345,0.2180088609,-0.1288373321,0.1291838586,0.0347200967,-0.0986380503,0.317584157,0.3372029066,-0.1442604065,-0.0596122369,0.0461500287,0.2753565609,-0.3600518107,0.0652869418,-0.0913713947,-0.018466508,0.059623275,-0.2850225866,0.5023182034,0.2224199176,-0.1454465985,-0.5411144495,0.1094877571,-0.2130830735,-0.3527453542,0.3614654541,-0.0578467138,0.1027428731,0.1347662807,0.0490506105,-0.3115338981,-0.5133593082,-0.1106426492,-0.2601343691,-0.0932939053,-0.1994627863,0.1886433363,0.0662672818,0.4169675112,0.0915017352,0.2924634218,-0.1015203819,0.2434794009,0.0530753881,0.0023513332,0.0708712563,0.4144098759,0.0296752192,-0.1921940893,0.101201199,0.4699638486,0.0964562595,-0.2393660098,-0.0600608177,0.4061223269,0.0860064179,0.1796572655,0.0710561052,-0.0081990799,0.1775366664,-0.0254899394,-0.2151744515,-0.3072147369,0.2342716157,-0.1045499146,0.2066566795,-0.1409114003,-0.0641389042,-0.087007761,0.5566837192,0.0496754088,0.0264087692,0.200042069,-0.0311267655,-0.2027164996,-0.0927500278,-0.1245568618,0.073691912,0.3062234521,0.1577223539,0.1531265825,-0.1649117023,-0.297547251,0.073410742,-0.0554010868,-0.0667957813,0.2283528447,-0.0130790137,-0.0862286538,-0.0698211342,-0.0733000562,-0.3054862916,0.2219373882,-0.2722060978,-0.2704034746,-0.2007989287,-0.25801754,-0.1964435577,-0.2613890171,-0.340551734,-0.5245014429,0.1668943465,-0.0902778804,-0.0871064216,0.2387023121,-0.0503766313,0.1446223706,0.1540334523,0.1810173541,0.0466038659,-0.5167757869,-0.0927902982,0.2334498167,-0.1593430191,0.149737969,0.3403945863,-0.1336889118,0.1381664723,-0.1418753415,-0.5138449073,-0.0082100024,-0.0126104197,0.2038814127,0.2323678434,0.0532638952,-0.0295092128,-0.0621770956,0.3798666894,-0.2169951499,-0.3239910305,0.1851449907,-0.2267712653,-0.2684181929,-0.1575942636,-0.2714623809,0.1473836601,-0.3481718302,0.184158802,-0.0482608043,-0.023082044,0.3228403926,0.0467668101,0.2795554698,0.0956809297,0.0788498744,-0.3114857972,0.0379384905,0.0993182734,-0.0798990875,-0.4554591477,-0.1120561287,0.042210903,0.3603470922,-0.2355370671,-0.2201904804,-0.4256613255,0.0864519998,0.1697891951,0.1727754027,0.0364519916,0.0551593415,-0.1652566791,-0.1744143516,-0.2287353128,-0.2283568531,0.3513429463,0.0433142781,0.3869698644,0.0045193983,0.1727219075,-0.0424155258,0.1412830949,0.440323621,0.0445500687,-0.0770474374,-0.1415098011,0.0284438003,-0.0575263128,-0.1927537322,0.1435080469,-0.0686766058,-0.206150502,0.3447676003,0.0584892593,0.2385403514,-0.1374520361,0.0210238658,0.018545704,-0.1231213883,-0.0254255347,0.0727336034,0.0446744636,-0.0113361273,0.1267823577,-0.1058705747,0.0305381436,0.1734837294,0.0818149522,-0.0937137827,0.1495662481,-0.2399790436,-0.0784451365,-0.4077475667,0.0804990157,0.197838068,0.0154053606,-0.0292738061,0.0826744214,0.0117125781,-0.0244430993,0.0592269674,-0.1593294889,0.1784007102,0.2000536323,0.2301491946,-0.2531661987,0.0642248392,-0.1200727746,-0.3929280341,0.5351771712,0.2702521384,-0.2166949511,0.1514447331,-0.1865818202,0.1602769941,-0.1781022102,-0.0418614037,-0.365007937,-0.3178344369,-0.0555087216,-0.0292986855,-0.0219658762,0.0418316908,-0.0450684465,0.2439177036,0.2413087636,-0.099062942,-0.0749873742,0.2095419914,0.2372506261,-0.2419347018,-0.1718311161,-0.1309131384,-0.0334616266,-0.3138867915,0.3177529871,0.0228584204,-0.1841026694,-0.051685553,-0.2914482653,0.0888901725,0.1174185798,-0.1467112303,0.1664441377,-0.0556510687,0.2312776446,-0.0983234569,0.1176709309,0.2892652154,0.0641449392,0.1284820437,-0.0746404454,0.7054752707,-0.0298366882,-0.1252156943,0.2515487075,0.1729599983,-0.2008546442,0.0660970733,0.4432734847,0.8700276017,0.4452214837,-0.194978863,0.5145097971,-0.0782524645,0.4081746638,-0.0467613302,0.1445289552,-0.4260663092,0.1069942787,0.0587814115,0.0954893157,-0.2111860067,-0.1544042081,-0.1624738723,0.0431595929,0.0213436168,0.02931072,0.0258022789,0.1563716382,0.0165743511,-0.0591619574,-0.1796628833,0.3069216311,-0.0922963023,0.0444304496,-0.0859647319,-0.0137678757,-0.357563138,-0.0294889305,-0.2921321392,0.1741712093,0.0745079592,0.1682662666,-0.1278496236,-0.1513185501,-0.2097330838,0.0683465376,0.6282229424,-0.0899130553,-0.278554678,0.1217430606,-0.105559729,0.1514124125,0.1019031778,0.1161519364,0.0843695551,-0.4008859396,-0.1757768542,0.0735089034,0.0230208579,-0.107561402,-0.2303290516,-0.2065986097,-0.4881097972,0.0413241945,0.0790346488,0.0703974217,0.0963750631,-0.1455853134,0.2291780561,0.0405316502,-0.150069803,0.1001781002,0.2356710434,-0.1699620485,-0.0144047737,0.0476888642,0.065575704,0.0185552929,0.0526985787,-0.1892166138,-0.0922254771,-0.2141876221,-0.2875443399,0.0032738149,-0.2989550233,0.0458161086,-0.0348411314,-0.0331100039,0.1233772486,0.703329742,0.39414379,0.1650057435,-0.0339545235,-0.266589582,-0.2566013336,0.2141294628,-0.2365321219,0.1248265281,-0.416267395,0.3438065648,0.0718638226,0.1509110183,-0.463509798,-0.2999201417,-0.3189359605,0.0742821619,0.1051843986,-0.1026962325,0.0479929931,0.2037711143,0.1363294274,-0.0006609804,-0.3898887038,-0.2764728963,-0.2066550702,0.0182425044,0.2317039818,0.0654115453,0.204174161,0.2290054411,-0.0559089966,-0.0340522379,0.0603797585,0.0123409405,0.0150528569,0.4621249735,-0.4461763203,0.4763486683,0.2246710062,-0.21796529,0.2197590172,0.2746258676,0.0964155421,-0.0551987626,-0.009555663,0.0501711853,0.0315329842,0.2480598688,-0.1137206033,0.0997338593,0.2904588282,0.0357176587,0.1950410604,0.0064350585,0.3093019426,0.0193520803,-0.0816640407,-0.1589885205,-0.1193384975,0.3786769807,-0.3763186336,0.0875012875,0.0676518381,0.1290613264,0.06302993,0.1560230851,0.2978933156,-0.1407539546,0.2699257135,-0.1306687593,0.4374745786,-0.4784556627,-0.0480431728,0.0673940256,0.108040154,0.1508616805,0.2001731247,0.0351094119,0.3142001033,0.2808910012,0.0137465559,0.4560908079,0.0020093676,0.3586678505,-0.109046489,0.2185404152,-0.1799707413,0.0978265256,-0.3583185673,-0.0269754138,-0.015520622,0.3004013002,0.0769286454,-0.0828003213,-0.2976915538,0.2627044618,-0.1866275817,-0.0932098255,-0.5035442114,-0.2040738761,-0.1423978359,-0.1158673689,0.1060523093,0.0357418768,-0.0464108624,-0.15523763,-0.1750718355,-0.3117406964,-0.3257957697,0.0814104527,0.3454497159,-0.2599895895,0.0702060163,0.5307807326,0.0909716263,0.0393273234,0.2912560701,0.462515831,0.3504654467,0.1511204988,-0.0632917136,-0.0383732244,-0.0949105173,-0.1292603463,0.2247769386,0.4108567834,-0.2550311387,0.4320936799,0.2484071702,-0.2820832133,0.3031463027,0.0878292918,0.0816031173,0.2764619291,0.2655619085,-0.2403648645,0.0200516842,-0.3033416271,-0.0590196885,-0.2542726994,0.1564845145,-0.0761900842,0.1321182698,0.1742834747,-0.2337680906,0.1855120212,0.0040892274,0.3029404283,0.119407557,0.2508754134,-0.211511746,-0.3532811701,-0.5663256645,0.1275719702,0.1013564095,-0.0808271766,-0.1319871098,0.0458896644,-0.1410633326,-0.0407781303,-0.2571213245,0.1057315841,0.1030746922,0.0140695767,-0.305932492,0.0497276485,-0.0150109772,-0.2027113587,-0.0203236584,0.000745867,-0.1910482198,-0.3189174235,0.2289662957,0.0006279812,0.0945826396,0.1555752754,0.6825587749,0.139412716,0.2825695574,0.5506614447,-0.0993252844,0.0239486657,-0.1264734417,0.075157702,-0.2360568792,-0.0585919581,0.0814063028,0.1764822006,0.1525346488,-0.0643260553,-0.267955482,0.5475525856,0.1570397913,-0.4307727516,-0.4397523403,0.235425666,-0.0332453847,-0.2299886942,0.0498141199,0.0461977087,0.0920632035,0.1632872075,-0.14231731,-0.4310871363,0.3653732836,0.0641834289,-0.0040292237,0.206302762,0.2444920093,0.1649839729,-0.0334927998,-0.7273082733,0.0448329933,0.2741759717,0.1157497764,-0.0318673179,0.1096809357,-0.1534430832,0.0586740784,-0.0568376444,-0.2978876829,0.2642129958,-0.2824088633,0.1399552375,-0.3156256676]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/198","title":"Index outside of table length","comments":"Sounds like something related to the nlp viewer @srush ","body":"The offset input box warns of numbers larger than a limit (like 2000) but then the errors start at a smaller value than that limit (like 1955).\r\n\r\n> ValueError: Index (2000) outside of table length (2000).\r\n> Traceback:\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n> exec(code, module.__dict__)\r\n> File \"\/home\/sasha\/nlp_viewer\/run.py\", line 116, in <module>\r\n> v = d[item][k]\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n> output_all_columns=self._output_all_columns,\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 290, in _getitem\r\n> raise ValueError(f\"Index ({key}) outside of table length ({self._data.num_rows}).\")","comment_length":9,"text":"Index outside of table length\nThe offset input box warns of numbers larger than a limit (like 2000) but then the errors start at a smaller value than that limit (like 1955).\r\n\r\n> ValueError: Index (2000) outside of table length (2000).\r\n> Traceback:\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n> exec(code, module.__dict__)\r\n> File \"\/home\/sasha\/nlp_viewer\/run.py\", line 116, in <module>\r\n> v = d[item][k]\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n> output_all_columns=self._output_all_columns,\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 290, in _getitem\r\n> raise ValueError(f\"Index ({key}) outside of table length ({self._data.num_rows}).\")\nSounds like something related to the nlp viewer @srush ","embeddings":[-0.133238405,-0.2262257189,-0.1187238246,0.3765133023,-0.0489251092,-0.0109354034,0.1363869011,0.4777677357,-0.0094934907,0.1617606431,0.3663626611,0.3341218531,-0.2157080024,0.2377277166,-0.0497593544,-0.3229538798,-0.1366588473,0.2217988819,-0.0474102944,0.0397772379,-0.3260987699,-0.0536401719,-0.5164063573,0.3713660538,-0.0388097316,-0.1306577176,-0.0224845707,0.0818180367,-0.209180057,-0.4875222743,-0.2217749357,-0.4269779325,0.0597802885,0.3852278292,-0.0001012715,-0.1532192379,0.3529761136,0.0443058163,-0.3176680803,-0.283875674,0.0040273229,-0.1019400656,0.1928707063,-0.332344979,0.1429475099,-0.2921248078,0.0910207778,0.0592385568,0.3193912506,0.6411463618,0.3207916915,0.2698442042,0.0917200223,0.0577383004,0.3815025091,-0.2520274818,-0.1376731396,0.0564995855,0.0104213264,0.0082856491,-0.221185267,0.1932849139,-0.0372691229,-0.0454516821,0.1960100681,-0.0337274894,0.0828029811,-0.3358167112,0.0396296941,0.1939150244,0.0700672492,-0.2029587477,0.0320325419,-0.3417423964,0.059512645,-0.3892810643,0.1312550306,0.2315450609,-0.1338343471,-0.0395006761,-0.0499038398,0.4061078131,-0.3019096255,-0.0286885146,-0.1252758354,0.4065403342,-0.08723744,0.1566182077,0.3018948138,-0.266861707,0.3795895278,0.0262668841,-0.0884927958,0.2416110486,-0.2634061873,-0.0378945656,0.2463030815,-0.0496404842,-0.0821807757,0.0403162539,0.5320737362,-0.0171080381,0.0405471846,0.1588834375,0.1809242964,0.3228932321,0.0861226246,0.1504656672,0.1071606576,0.2897972167,0.0700114518,0.021468062,0.090175122,-0.3275979161,-0.0366015695,0.1192254722,0.1230448708,-0.1119540185,-0.4574782252,0.4688632488,-0.1460352838,0.1714396924,0.3362070322,0.2654515207,0.0790499598,0.1006226018,0.0420840867,0.1913594306,-0.2000142634,-0.1114705205,-0.1485529989,0.2197987288,-0.309001863,-0.1194914505,0.1783698201,0.0371320695,0.2479361445,0.0752476752,0.1128936633,-0.0240448508,0.0899483114,-0.2651390433,0.2356685102,0.1921767443,0.0518566258,-0.0876408517,0.1058339775,0.0142681403,-0.0453288406,0.1420902312,-0.0806454048,-0.1744566113,0.0777236745,0.3809062541,-0.059262272,0.0575511828,0.0472071953,0.1408001035,0.1391278505,-0.1432040036,0.1842824072,-0.373375088,-0.0750506818,-0.1875309497,-0.0067924899,0.4820596576,-0.0348659977,0.3262298107,-0.158228308,0.0748695582,0.0513832308,0.0617387407,-0.1701130271,0.2285058051,-0.2171483934,0.3299371302,0.5164395571,-0.149487406,0.0478117578,0.159953624,-0.2912843227,-0.0345228687,-0.0193789676,-0.0521632358,-0.0088932086,0.0336036235,0.1489809304,0.1439203471,0.0546148568,0.0263723992,-0.255946964,0.0948806852,0.0802264139,0.0365591533,-0.0241267737,-0.2969387472,0.1284704804,-0.1834527701,0.0553035624,-0.0118966773,-0.0509837531,0.219136402,-0.2892916799,0.0137257082,0.0780794621,-0.3570205569,-0.0146279009,-0.0668874681,-0.2444724739,-0.005655251,-0.1188067272,-0.0721564516,-0.3345106244,0.2136516422,-0.1766919643,0.0413508788,0.3529234529,-0.1696059406,0.1570164561,0.0070035136,-0.332271874,0.0054832804,-0.2140140384,0.0022727407,-0.4161435962,-0.008052092,-0.5051020384,-0.1797168255,0.3080259562,0.1510012299,0.2772066295,-0.035744898,0.0581195615,0.3263455331,0.0853090063,-0.1529144198,-0.0172382202,0.0307609681,0.1294930726,-0.1194191799,0.3906146586,0.1942807585,0.1542287767,-0.0545545369,0.0255146716,0.4117993414,-0.4132155478,0.2022223622,0.1215811595,-0.1306020916,0.1380481869,-0.0681509823,0.0957562625,-0.1901345402,0.4481594265,0.1647072583,-0.2402761132,-0.2005037814,-0.3936164379,0.2649053037,0.2466845661,-0.0391294211,0.063982971,0.3070986867,-0.0298833121,-0.198688671,-0.1541000009,0.1941155344,0.2751754224,0.1935089678,-0.0301015247,0.1393740624,-0.3416764438,-0.3664766252,0.2637137771,-0.0280995443,-0.013718388,0.4935802221,0.154754281,0.0004630288,-0.2708466947,-0.2757884562,-0.0908428654,0.4761658907,-0.3278590143,0.19213745,-0.20387128,-0.6890749335,-0.3032545447,-0.167348057,-0.0827879906,-0.4023782015,0.1184008047,0.1464321017,-0.3485916257,0.0803497359,-0.2606227696,0.1563516408,0.1396342218,0.3389835656,-0.0010203804,-0.1446295083,-0.1371298581,0.1889013946,0.174193427,0.3428304195,0.2457439899,0.0428294837,0.0768582821,-0.0531437919,-0.478351891,-0.0208371915,-0.0618384853,0.0522885807,0.2572689354,-0.0092602363,-0.1127746254,-0.3093059361,0.1834376901,-0.1338688135,-0.0469287001,0.3341228664,0.0844841376,-0.1665892005,-0.1900297403,-0.2247220725,-0.04969896,-0.5674669743,0.2856423259,0.3556131721,0.4707017243,0.270096153,0.1489456743,0.3565600216,0.0615097061,0.1508973092,0.0182869863,0.2672909796,0.2674719989,-0.283941716,-0.4705733359,-0.0933910087,-0.1202681139,-0.1404035836,-0.1052996442,-0.568720758,-0.1082083359,-0.1263613701,-0.1857580692,-0.0495415442,0.0280363243,0.2887747884,0.0848414451,-0.2035464346,-0.107989043,-0.0401713289,-0.0067494996,-0.0244193878,0.0302870385,-0.0379774272,0.4838706553,-0.1285218,0.1772360355,0.0865371153,-0.0445255674,0.2395199388,0.0241747964,0.2828855515,-0.3588386476,-0.0795336217,0.0018074247,0.0585237183,-0.1227958947,0.2333753109,-0.259943068,-0.2363991886,-0.2503324449,-0.197701022,-0.043788027,-0.0726653561,0.3003554344,-0.044970233,0.0391382165,-0.0265344046,0.0684025511,-0.2269888073,-0.1739422828,-0.3278630674,-0.1942297816,-0.0430127196,-0.1237261295,-0.2608478367,-0.2829256356,-0.2433509529,0.2859064937,0.1081261337,0.4123806953,-0.0057857353,-0.3465947807,0.1564339548,-0.2174128145,0.0591527,-0.0546771139,-0.1354223341,0.1021322012,0.2638218999,-0.5495522022,-0.123638913,-0.2913531065,0.5158879757,0.1156245098,0.1373420358,-0.1338265389,0.1171393096,0.5322049856,-0.0574999154,-0.0716552064,0.0200447682,-0.1239345968,-0.071058549,-0.10950955,0.1334229559,0.0364482515,0.0902175829,-0.18102099,0.0562338792,0.1419372261,0.0233271457,-0.1322594136,0.3057263792,0.2065767795,-0.2066662759,0.1177527085,-0.1567614824,0.0767366663,0.1709097773,0.3915001452,0.0470525026,0.0124365864,0.0429019779,-0.1284882873,0.0969407633,0.2476730049,0.1561854929,0.2156841755,-0.0749110878,0.4538375735,-0.0940474644,-0.1433058977,0.2753116488,-0.099817194,-0.0581159256,-0.2762084305,0.2647866607,-0.2276251018,-0.0400295928,0.2967908978,-0.1317806095,-0.2011369318,0.5465393066,0.1207216233,0.9295414686,0.2854915261,0.0684950873,0.24575679,-0.2783338428,0.1311043501,0.0063942177,0.27089867,-0.5571449995,-0.2443258017,0.0656882972,-0.199789077,-0.0498792008,-0.1114046052,-0.3538817465,0.1584634632,-0.3280289471,-0.2146938741,-0.3066506386,0.1497822106,-0.2506892085,0.0679661483,-0.4111403823,0.2196308821,-0.103968896,-0.2046980858,-0.1517551541,-0.1247228384,-0.029502552,-0.3921319246,-0.3081608117,0.1502212882,-0.0910602883,0.0347632505,-0.155927375,-0.1586449146,0.045044519,0.0049594408,0.0243094955,-0.0440678671,0.0024172151,-0.0004190792,0.0725956038,-0.3184443116,0.0572284721,0.3024685681,0.3180184066,-0.1191358343,-0.2576569021,0.1112317145,0.1186495349,-0.125362128,-0.0086533464,0.2626308799,-0.0332970805,-0.1776463389,-0.0154015981,-0.162494868,-0.4355313182,-0.0684611499,0.2529904842,-0.1484930664,-0.1568633616,0.0819307193,0.086387977,-0.3561632931,-0.1171076968,0.2700821161,0.1515426934,0.4332962036,0.3825482428,0.3150550723,-0.2954520583,-0.3444068432,0.0133733153,0.0219467413,-0.2640549839,0.0809913054,0.0101425368,-0.0127663305,0.253603816,0.1418090165,0.1763442755,0.2594761252,-0.2593019307,-0.460948348,-0.1588309258,0.1030695662,-0.0131948646,0.1044432521,-0.0516610444,-0.1483772397,0.2633280456,0.0263240524,-0.4421235025,0.0912130326,-0.3083191216,0.3241728246,0.1206309646,-0.2807697356,-0.0504540168,0.1301853508,0.2010685205,0.1774639785,-0.3504231274,-0.3401643932,0.0663654134,0.0778914392,0.0032963248,-0.342464,-0.3397029936,-0.3401270211,-0.1230338663,-0.0993687436,-0.1617391706,0.1107571647,0.0008492165,0.1063362882,0.257722497,0.1093791872,0.2833819985,0.1797163785,-0.1522388309,0.3850378096,-0.0008012282,0.1756022871,-0.1323809922,-0.0775726438,-0.2705686688,0.2831527293,-0.188418448,0.0371463448,0.29748559,-0.1989853233,0.0098073641,0.0782844946,0.2811998725,0.6313968897,-0.14683415,-0.1926251054,0.1549924314,0.2711911798,-0.3233412206,-0.0300567746,-0.2071067691,0.0001758744,-0.0507436879,0.1768577844,0.0627953783,-0.3913722336,0.0651489124,0.2327067256,0.262334913,-0.2258984596,-0.1150067672,0.2281734347,-0.1044958159,-0.2563814819,0.3202328384,0.1353222579,0.4605558515,0.3844752014,-0.1124932095,-0.0666311979,0.2495865077,-0.2212644368,0.0096593257,0.0239654519,-0.1367896199,0.402364403,0.1489294171,0.0880054608,0.239349246,0.2020855993,0.0422357097,0.0588697866,-0.2548345029,0.1123182774,-0.1619550288,0.112508826,-0.3357420564,-0.1203203499,-0.3245554864,-0.1064159945,0.0942437425,0.4068545103,0.0880302191,0.1292232573,-0.2584675848,-0.0992121771,0.1526081711,0.1043204293,0.1699525565,-0.3836728334,0.3275732398,0.4612177908,0.2695771158,0.1593056023,0.287178874,0.5918681026,0.5304178596,-0.3573252857,0.0912871733,0.0554921925,-0.192640245,-0.0331636854,0.2009248435,0.2197234035,0.3352774978,0.5446245074,0.2335324585,-0.2273259461,0.2913128138,0.3294221163,0.0526714548,-0.0010781689,0.1382812113,0.1690847725,-0.1910012662,-0.1128903478,-0.0153368851,-0.5508238077,0.3721139431,-0.021395728,-0.0576191358,0.176618591,-0.1396689266,0.1383371055,0.012731418,0.6826310158,0.4650734663,-0.0403937362,-0.2285886109,-0.2114341855,-0.4472535253,0.2634841502,-0.3624057174,-0.0898094252,0.1910163015,0.1311064959,-0.1349637657,0.0297842789,0.3164514601,-0.0276681967,-0.217275098,0.3570563793,-0.2385870218,-0.0089104585,0.1198351979,-0.0891094208,-0.1422021389,-0.1422027051,0.1729949862,-0.4607591331,0.2839155495,-0.0325919241,0.0608201027,-0.1452781856,0.201428622,0.220150575,0.1447535604,0.2547079027,-0.1325617582,0.10788358,-0.3166041374,-0.2383264601,-0.1965483725,0.2629193366,0.3339879811,0.099878259,-0.0436787233,-0.4462735355,-0.0821764022,0.2436748296,-0.1146495342,0.1059982628,-0.0040416145,0.0390990674,-0.0473223925,0.0303885993,-0.0006308843,0.074399434,0.2238695323,0.0690316334,-0.090462558,-0.5038040876,0.4410120249,-0.2285667509,-0.1545843184,0.1264416575,0.2720943093,0.0496783741,-0.0332020074,-0.4504557252,0.0156943277,0.2433985621,-0.0640896633,-0.3634355366,0.0986986235,-0.3859491944,-0.2755608261,-0.0004915376,-0.1275265366,0.1595144868,-0.1565993577,0.0067115691,-0.187379688]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/198","title":"Index outside of table length","comments":"Fixed. ","body":"The offset input box warns of numbers larger than a limit (like 2000) but then the errors start at a smaller value than that limit (like 1955).\r\n\r\n> ValueError: Index (2000) outside of table length (2000).\r\n> Traceback:\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n> exec(code, module.__dict__)\r\n> File \"\/home\/sasha\/nlp_viewer\/run.py\", line 116, in <module>\r\n> v = d[item][k]\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n> output_all_columns=self._output_all_columns,\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 290, in _getitem\r\n> raise ValueError(f\"Index ({key}) outside of table length ({self._data.num_rows}).\")","comment_length":1,"text":"Index outside of table length\nThe offset input box warns of numbers larger than a limit (like 2000) but then the errors start at a smaller value than that limit (like 1955).\r\n\r\n> ValueError: Index (2000) outside of table length (2000).\r\n> Traceback:\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/streamlit\/ScriptRunner.py\", line 322, in _run_script\r\n> exec(code, module.__dict__)\r\n> File \"\/home\/sasha\/nlp_viewer\/run.py\", line 116, in <module>\r\n> v = d[item][k]\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n> output_all_columns=self._output_all_columns,\r\n> File \"\/home\/sasha\/.local\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 290, in _getitem\r\n> raise ValueError(f\"Index ({key}) outside of table length ({self._data.num_rows}).\")\nFixed. ","embeddings":[-0.2290041447,-0.1375160068,-0.1168010011,0.3834364116,0.0230834354,-0.1093542352,0.1977808774,0.5292378068,-0.0128640002,0.2315807194,0.3296780586,0.3583359718,-0.1969748437,0.313515991,-0.1167158186,-0.2864811122,-0.1225041002,0.2326520979,-0.0685144663,0.0619008094,-0.3003188074,-0.0693498403,-0.4741615057,0.2245804667,0.0024794617,-0.1122638658,-0.0148344859,0.0532752536,-0.2355968654,-0.4581713974,-0.2489648908,-0.3824645579,0.1219595745,0.4385651052,-0.0001028612,-0.0989758372,0.3066659868,0.0019687708,-0.3103621602,-0.2512944937,0.0300170295,-0.0707657933,0.1825018823,-0.3371544182,0.0350745209,-0.3469803035,0.0166640338,0.0100743631,0.3727734387,0.6639379263,0.3033929765,0.2469041198,0.1873139888,0.0709917694,0.3185144663,-0.216634199,-0.1188883483,0.0575183742,0.0042842152,0.0654430613,-0.2602181733,0.2600080073,-0.022401439,0.0030027335,0.1475158781,-0.0409352556,0.0859302357,-0.2707749009,0.0154653369,0.1453943998,0.1314297169,-0.2809050977,-0.0662182495,-0.3357970119,0.0924672708,-0.4406609833,0.0438793972,0.1418468654,-0.0759172216,-0.0036910356,0.0350233056,0.3347105384,-0.3286305666,-0.0172205549,-0.0028420531,0.4352541268,-0.0539207868,0.1099453494,0.2628906071,-0.2523483634,0.4487836957,0.0818690285,-0.1459727734,0.1315634251,-0.2745288014,-0.031340979,0.2478069365,-0.1008085385,-0.057051897,0.0478765368,0.5200781226,0.1131842509,0.081541121,0.0971835777,0.1547946036,0.3320045471,0.08339037,0.2624536455,0.172392115,0.3851076066,-0.0038626811,-0.0226290468,0.0889475718,-0.3653741181,-0.0237417314,0.1983584911,0.1481001973,-0.1093857735,-0.3547264636,0.3757118285,-0.077771388,0.1630586237,0.2924057543,0.3498601615,0.0938729122,0.1693046093,0.1115283519,0.2726615667,-0.1729181707,-0.1453622878,-0.1700188369,0.1816277057,-0.2132821977,-0.1243538409,0.197705552,0.0323795937,0.261891216,0.0057491725,0.134632498,-0.0739455894,0.052576609,-0.2558829784,0.2255927175,0.1424025595,0.1174596623,-0.1140259504,0.1352451891,-0.0008992295,0.0269759316,0.1784808487,-0.132068947,-0.1783427447,0.0277002566,0.3667051196,0.0153741157,0.140461117,0.0187939629,0.1155282557,0.1999523491,-0.1628684551,0.1260977685,-0.3559342325,-0.0966231078,-0.1602923274,0.0778005719,0.4474713802,-0.0545128621,0.3515463471,-0.1433376968,0.097016789,0.0767563656,0.0518009104,-0.1469629705,0.1965745538,-0.248229295,0.2470377535,0.4916948974,-0.1780704409,-0.0345287658,0.1323575675,-0.283176899,-0.1036930382,0.052549284,-0.084081918,0.0891646296,0.0755866915,0.074917376,0.116875492,0.0451752469,0.1279543191,-0.2999439538,0.0830923393,0.065926075,0.0803880095,0.0452086702,-0.2939603329,0.1397746205,-0.2402184904,0.1205889732,-0.0845133066,0.0209866688,0.2763915658,-0.1763716489,0.027185509,0.14651981,-0.3263149261,-0.0860691592,-0.0694415495,-0.2273828536,-0.0185903218,-0.2183769494,-0.122279577,-0.2982162833,0.224076435,-0.2041163892,0.0716302395,0.3541257083,-0.2096327394,0.1501780152,-0.0677838102,-0.2856623828,-0.0043737888,-0.1862680018,0.0447034612,-0.3344715834,0.0624085516,-0.5087909698,-0.2284972668,0.3810814023,0.169438839,0.3257103562,-0.1004474387,-0.0052077193,0.3581848741,0.0798982978,-0.1017358601,-0.05023681,-0.0627405345,0.1189251766,-0.1621805429,0.3511015475,0.2010252476,0.1372337639,-0.0129099032,-0.0664590895,0.4866172969,-0.4441443682,0.1708923131,0.1060192138,-0.1639505327,0.1599204093,-0.0568987802,0.1096209586,-0.2167057395,0.4039099216,0.1221840084,-0.2314979136,-0.2478119433,-0.3556720018,0.2311990559,0.2187159806,0.121481888,0.1083020419,0.305421412,-0.0416579023,-0.1725354046,-0.1011398733,0.1564950496,0.2163029164,0.2630329132,-0.0212557334,0.0998064503,-0.2129559517,-0.3551459908,0.2616561055,-0.0735778138,0.0232112464,0.5582045913,0.1833824664,-0.1107759774,-0.2681664228,-0.1848720759,-0.0488784909,0.4843204916,-0.3535310328,0.0935651213,-0.2139147222,-0.6548931003,-0.3214907646,-0.1441179365,-0.0125679085,-0.3868684173,0.1454862654,0.2466431558,-0.4037140012,0.2167981416,-0.2161088139,0.0572301634,0.2187269926,0.2697334886,-0.0216245335,-0.1124144122,-0.0126001742,0.1559532732,0.1791496873,0.2972017825,0.2275326997,0.09011437,0.0763463825,-0.1081007645,-0.4858226478,-0.0897506997,0.0127906678,0.0313948877,0.1904121637,0.0803586766,-0.1451539099,-0.3487899005,0.2091731578,-0.1497933716,-0.0483703017,0.3853221238,0.0669738501,-0.1695479453,-0.201755479,-0.250988692,-0.0321795866,-0.5896877646,0.2484629303,0.3697091043,0.4624907374,0.1798277795,0.2835752964,0.1695379466,0.0736848041,0.1270151436,0.0553880148,0.2804184258,0.3034002185,-0.2880815268,-0.4810489416,-0.1055509895,-0.1271498799,-0.1246361658,-0.0996452644,-0.512486279,-0.1237407252,-0.0389066935,-0.1347715855,-0.0365946591,0.0793852955,0.2825539112,0.1378822029,-0.2032253742,-0.139186725,0.0313776582,0.0675884038,-0.0390417688,-0.0057984465,-0.0794390738,0.5325879455,-0.0619671419,0.1575868577,0.1140091568,-0.0144490469,0.2855709493,-0.0392228849,0.2843078375,-0.3779863417,-0.104572989,0.0220772512,0.0565323718,-0.0010952262,0.1910884529,-0.2565448284,-0.179779321,-0.2928996384,-0.1544565707,-0.1581046581,-0.1141016558,0.2958340049,-0.0410834029,0.0682612434,0.0291358009,0.0708325878,-0.2715314031,-0.1704979837,-0.2771558166,-0.239842698,-0.058675617,-0.1285798848,-0.2209395021,-0.2278408259,-0.2994347215,0.3481402695,0.0736461058,0.4333791137,0.032817781,-0.3242917657,0.0903988034,-0.2628679872,0.1329485178,-0.0532547459,-0.0917884409,0.0341859423,0.2910724878,-0.51796031,-0.0878946334,-0.2348316908,0.4028249979,0.016049087,0.1764225811,-0.2226371467,0.135680601,0.4611270428,-0.0569542497,-0.0732196271,0.0200684983,-0.1209761128,-0.0366455801,-0.2352845967,0.1271226108,0.0365538336,0.1743913889,-0.1095524058,-0.063998878,0.0987940654,-0.0942894071,-0.2096584886,0.2917173505,0.3237657249,-0.1531560868,0.1714385599,-0.1597562432,0.0724966675,0.3018145263,0.3614908159,0.0411586501,-0.0713876635,-0.01916706,-0.2208825499,0.0976115465,0.244847551,0.0685563013,0.2353926003,-0.0477552973,0.4707000256,-0.0081787845,-0.1329434067,0.1954274625,-0.0962207913,-0.0390914641,-0.2485679537,0.1689151078,-0.2677552998,0.0117906323,0.335701853,-0.1436322182,-0.2347403467,0.5492370129,0.0994339436,0.8743262887,0.259927094,-0.029165864,0.2315056324,-0.343446672,0.105356589,0.0368032306,0.2641950548,-0.6125146747,-0.3115994334,0.0995117277,-0.1769768447,-0.0995741561,-0.1063847616,-0.3265552819,0.1588816345,-0.2933765054,-0.2477368712,-0.3891933262,0.1981050968,-0.1871576756,0.0178200863,-0.3944045007,0.1764938533,-0.121597372,-0.2284388542,-0.1167938486,-0.1742409319,-0.0446888804,-0.4916746616,-0.2481038868,0.1595376581,-0.1961614341,-0.0083461013,-0.2384604216,-0.2075696141,-0.0155066783,-0.0084916959,0.0247931723,-0.0071160407,-0.0065313294,0.0296700764,0.1331669688,-0.2641233504,0.027516298,0.3408159316,0.1873677969,-0.1063347235,-0.195336163,0.0476671718,0.1493738145,-0.1805903167,0.1030348092,0.157987386,0.0639900863,-0.194463864,0.0147784362,-0.1807263494,-0.4378423989,-0.1918722242,0.2462718189,-0.1515770406,-0.1618925184,0.0875447467,0.1722149402,-0.2888276875,-0.1701065153,0.3804877996,0.1121263281,0.4354037344,0.3700926602,0.4018879533,-0.3255645633,-0.3656489551,0.0411105752,0.0887074396,-0.3419449925,0.0724280477,0.0310750604,-0.0012461775,0.227512911,0.1593825221,0.1803201735,0.3041673899,-0.2412903607,-0.4176249504,-0.2141372263,0.1519402117,-0.0412862562,0.1247800589,-0.0491235629,-0.1017751247,0.1730614007,0.0977439508,-0.4318977892,0.0816903636,-0.2344302535,0.3294631839,0.100409098,-0.2331771851,0.0036072391,0.1007971317,0.207833156,0.1737625152,-0.2890529931,-0.3395385146,0.083009623,0.0714183822,-0.0064166728,-0.2782253623,-0.3691451252,-0.3791833222,-0.242950514,-0.1135256216,-0.1650075316,0.0709863082,-0.0556924418,0.0841740668,0.2978561521,0.1191783175,0.2970304787,0.1133811399,-0.1280782223,0.3637623787,-0.077424638,0.1529646963,-0.1271490008,-0.043980062,-0.2035461664,0.2232094109,-0.1709265262,0.072147727,0.3320157826,-0.2585488558,-0.0335812047,0.1472819448,0.3666805327,0.6015073061,-0.1324405819,-0.0908403322,0.2007203847,0.2705457211,-0.3949107826,-0.0124267563,-0.197870627,-0.0449020714,-0.0203032251,0.2234343588,0.0279149432,-0.3858051002,0.1304169148,0.2902474999,0.2262700647,-0.2662835419,-0.0991126373,0.2385098189,-0.0776038542,-0.24510023,0.3247416019,0.2027331293,0.4037175477,0.4040175378,-0.1664659083,-0.1148602813,0.2091376781,-0.2543464005,-0.0566315763,0.0119295958,-0.1535087228,0.3529007435,0.1383945495,0.0198431332,0.244380489,0.2023603022,0.0699271858,0.0355905406,-0.2071412504,0.0555282421,-0.1854233593,0.1472988129,-0.4107872248,-0.0485374108,-0.3276323378,-0.1491126567,0.1114194766,0.4208343327,0.055115629,0.1727081686,-0.2184915692,-0.1441886127,0.1158232614,0.1612262875,0.2009258717,-0.4281047881,0.3499911726,0.4401037693,0.3009001911,0.1067221686,0.230506584,0.5381067991,0.4629309773,-0.3149114549,0.1076356471,0.0598461442,-0.1954316348,-0.0616464615,0.1432994604,0.2941911817,0.3352108896,0.5372707248,0.2220494151,-0.2414339632,0.2765609622,0.3478054106,0.0577322543,0.0758505911,0.1176186427,0.1999929249,-0.2179428637,-0.1286614239,-0.1789830178,-0.4712466598,0.392233789,-0.0908950344,-0.0896462053,0.1327992976,-0.1560448557,0.1361703724,0.0510122627,0.5293292403,0.3817704618,-0.0264513772,-0.260173589,-0.2490913421,-0.4916492403,0.2357487977,-0.295647949,-0.0786792785,0.1913115233,0.1255434155,-0.0897942483,0.0296579432,0.3834896386,-0.0655999184,-0.0696757734,0.4154008329,-0.2687748671,-0.0086307125,0.1079377979,-0.0275277924,-0.0669308677,-0.1737410873,0.1773691475,-0.4076396525,0.2727748454,-0.0077115702,0.0676878542,-0.1574425846,0.1660272181,0.2716854215,0.0663292482,0.187898621,-0.1952342242,0.1704770625,-0.2702898383,-0.1951018274,-0.2456133068,0.2598594725,0.3992114067,0.213436231,-0.0390908159,-0.5075240135,-0.1455917209,0.2119693309,-0.0894265547,0.0953139663,-0.0090446211,0.1187259555,0.0314623937,-0.0039624679,-0.0799485818,-0.0185557138,0.1622373164,0.0549488626,-0.1746305972,-0.5686410666,0.4154297113,-0.2183796018,-0.1622720808,0.1523312032,0.2936242521,0.0808549821,-0.0438135378,-0.4488560259,0.1100131124,0.2185388356,-0.0287870057,-0.3105335534,0.083390221,-0.365018189,-0.1618248373,-0.0342968255,-0.1087273508,0.1912637651,-0.1336429566,-0.0875502229,-0.1761401147]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/197","title":"Scientific Papers only downloading Pubmed","comments":"Hi so there are indeed two configurations in the datasets as you can see [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/scientific_papers\/scientific_papers.py#L81-L82).\r\n\r\nYou can load either one with:\r\n```python\r\ndataset = nlp.load_dataset('scientific_papers', 'pubmed')\r\ndataset = nlp.load_dataset('scientific_papers', 'arxiv')\r\n```\r\n\r\nThis issues is actually related to a similar user-experience issue with GLUE. When several configurations are available and the first configuration is loaded by default (see issue #152 and #130), it seems to be unexpected for users.\r\n\r\nI think we should maybe raise a (very explicit) error when there are several configurations available and the user doesn't specify one.\r\n\r\nWhat do you think @lhoestq @patrickvonplaten @mariamabarham ?","body":"Hi!\r\n\r\nI have been playing around with this module, and I am a bit confused about the `scientific_papers` dataset. I thought that it would download two separate datasets, arxiv and pubmed. But when I run the following:\r\n\r\n```\r\ndataset = nlp.load_dataset('scientific_papers', data_dir='.', cache_dir='.')\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5.05k\/5.05k [00:00<00:00, 2.66MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.90k\/4.90k [00:00<00:00, 2.42MB\/s]\r\nDownloading and preparing dataset scientific_papers\/pubmed (download: 4.20 GiB, generated: 2.33 GiB, total: 6.53 GiB) to .\/scientific_papers\/pubmed\/1.1.1...\r\nDownloading: 3.62GB [00:40, 90.5MB\/s]\r\nDownloading: 880MB [00:08, 101MB\/s]\r\nDataset scientific_papers downloaded and prepared to .\/scientific_papers\/pubmed\/1.1.1. Subsequent calls will reuse this data.\r\n```\r\n\r\nonly a pubmed folder is created. There doesn't seem to be something for arxiv. Are these two datasets merged? Or have I misunderstood something?\r\n\r\nThanks!","comment_length":98,"text":"Scientific Papers only downloading Pubmed\nHi!\r\n\r\nI have been playing around with this module, and I am a bit confused about the `scientific_papers` dataset. I thought that it would download two separate datasets, arxiv and pubmed. But when I run the following:\r\n\r\n```\r\ndataset = nlp.load_dataset('scientific_papers', data_dir='.', cache_dir='.')\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5.05k\/5.05k [00:00<00:00, 2.66MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.90k\/4.90k [00:00<00:00, 2.42MB\/s]\r\nDownloading and preparing dataset scientific_papers\/pubmed (download: 4.20 GiB, generated: 2.33 GiB, total: 6.53 GiB) to .\/scientific_papers\/pubmed\/1.1.1...\r\nDownloading: 3.62GB [00:40, 90.5MB\/s]\r\nDownloading: 880MB [00:08, 101MB\/s]\r\nDataset scientific_papers downloaded and prepared to .\/scientific_papers\/pubmed\/1.1.1. Subsequent calls will reuse this data.\r\n```\r\n\r\nonly a pubmed folder is created. There doesn't seem to be something for arxiv. Are these two datasets merged? Or have I misunderstood something?\r\n\r\nThanks!\nHi so there are indeed two configurations in the datasets as you can see [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/scientific_papers\/scientific_papers.py#L81-L82).\r\n\r\nYou can load either one with:\r\n```python\r\ndataset = nlp.load_dataset('scientific_papers', 'pubmed')\r\ndataset = nlp.load_dataset('scientific_papers', 'arxiv')\r\n```\r\n\r\nThis issues is actually related to a similar user-experience issue with GLUE. When several configurations are available and the first configuration is loaded by default (see issue #152 and #130), it seems to be unexpected for users.\r\n\r\nI think we should maybe raise a (very explicit) error when there are several configurations available and the user doesn't specify one.\r\n\r\nWhat do you think @lhoestq @patrickvonplaten @mariamabarham ?","embeddings":[0.3410282135,-0.0507830754,0.0498376898,0.2057841271,-0.0124936337,-0.0915830582,0.1314337552,0.2478140146,0.0751410052,-0.2193237543,-0.1656386554,0.1005809531,0.0638275295,-0.2900847793,0.2401754558,0.0783967748,0.0976624116,0.1165956408,0.1482492685,-0.1969807744,-0.1103292704,0.4662143588,-0.1177706718,-0.1241875067,0.0264826193,0.057828825,-0.2549771369,0.2292036861,-0.1758149266,-0.2693143189,0.2296958268,0.4607445002,0.0106472215,0.2442484796,-0.0001260212,-0.0266372915,0.4085169435,0.0374921151,-0.0753441975,-0.128483057,0.1525420547,-0.3127950728,0.397844553,-0.106541872,0.042281799,-0.3930912018,0.1760856211,-0.046389278,0.1798724681,0.0012882873,0.0871801004,0.1242344752,0.0638084337,0.0672760829,0.4933620095,0.0101154372,0.0661203563,0.4198920727,0.2697269619,-0.1633948535,0.1365495175,0.2445969731,-0.1149521098,0.1563950777,0.1796972901,0.4194620252,0.0596443713,-0.4671780169,-0.2021933943,0.3455394804,0.0148535902,0.0130189043,-0.0460986458,-0.4677795768,-0.0122009078,0.0002345637,0.2077282518,0.4563730955,-0.1608153731,-0.0648918375,0.0811817348,-0.2500706613,-0.0604624264,0.561724484,0.1643213481,0.048341386,-0.1973384768,0.3637367487,0.4003931284,0.1685729623,-0.2624977529,-0.287566334,0.1975403875,0.1875087172,-0.3029337227,0.0080817882,0.0892410204,-0.0668539926,0.2401620001,0.3428836167,0.0869829282,-0.043154031,0.1938953996,0.1217616498,0.5296929479,-0.0609327704,-0.0209741835,0.1862228066,0.3687579334,-0.0165914521,0.0886773691,0.0983025879,0.0916959196,-0.0254227165,-0.2339766324,-0.5981558561,0.2879770994,-0.2756765187,0.2487315238,-0.0236083772,-0.0111906305,-0.0559957027,-0.2106598765,0.0046254485,0.0189874526,0.2226313055,0.1190301552,0.2808913887,-0.2055922002,0.2087723464,-0.0782689825,0.0059810891,-0.3656755686,0.0397046246,0.2711539567,-0.2781743705,0.1672439575,0.0916653425,-0.1132247075,-0.18400985,0.1791826189,-0.1315311641,0.0667215213,0.3698316514,0.2778123021,0.508071959,-0.0019687526,-0.2545326948,-0.2754873335,-0.0611439943,-0.2797983587,-0.1683173478,-0.1868097186,0.0298579298,-0.3902364969,0.1666101366,0.000118563,0.2882416546,0.1552381516,-0.1705115885,-0.1155071557,-0.0045683435,-0.0851240903,-0.3703423738,0.1219870299,0.3437537849,0.1906010509,-0.1056915969,-0.1270908415,-0.1005398929,0.0820995644,0.3111583889,-0.2761954367,0.1420872211,-0.2416537255,0.2310815901,0.5904170871,-0.2710135281,-0.4526688159,0.3115585446,-0.0249654334,0.074895449,0.152230531,0.2749254405,-0.138735801,-0.0225615874,0.1659161896,0.366828233,0.0908415616,-0.146275267,-0.3049763143,-0.1942659616,0.3023201823,-0.0907404348,0.1159532443,0.0059388643,-0.0089752292,0.0697069168,0.306605041,0.2294956148,0.1860751212,0.246301949,-0.0749332532,-0.0090108374,0.0708625168,-0.2000005096,-0.3342154324,0.0563949533,-0.355522126,0.2431680262,0.2622113228,-0.0423657894,-0.232797116,-0.4931894541,-0.1797827929,-0.2683065534,0.0047272388,0.2189692855,0.1224113032,0.1607274115,-0.2678839266,0.1743549258,0.0098401709,0.0880597085,-0.8079344034,0.3179457486,-0.0178457256,-0.0704507381,0.0631345361,0.1433012336,0.0012621655,-0.1691232473,0.3244474828,0.4365813136,-0.1374934316,0.2282069325,0.2255856842,0.0575182065,0.2537788153,-0.1108493656,0.1513350457,-0.1257910579,0.1119107604,-0.2311308086,-0.3042535782,0.1168276817,-0.077087,0.3747087121,0.1469537616,-0.0114957523,0.0103485119,-0.2301260382,-0.474625349,-0.2160553038,0.2014878392,0.2280126214,-0.1119729355,0.1459307522,-0.2998267114,0.0780779719,0.070139572,0.0172316842,-0.0204678196,0.1736777574,-0.2714233398,-0.244095698,-0.1822142005,0.5162521005,0.8718118668,0.1978533715,0.3229739964,0.1088076681,-0.1763394624,-0.196975708,0.0869795531,0.0337333344,-0.1280592978,0.2140399367,0.1713266373,-0.0086421194,-0.4231769443,0.3805908263,0.1604651511,0.0212894902,-0.3763753772,-0.1903130859,-0.4012383521,-0.3012850881,-0.6075859666,-0.1843883842,-0.3624563217,-0.3466341197,-0.0647551119,0.1399495602,-0.0504873395,0.0246390924,-0.2065403014,0.1555749625,-0.2396001816,0.068418026,0.141157195,0.1248278171,-0.3013817668,-0.0807427615,0.0365163162,0.3887548745,0.5186214447,0.0546899512,-0.2073940039,-0.0898477808,-0.4579675496,0.2533678114,-0.2039087415,0.4136421978,0.28525576,0.0983376056,0.3634143174,-0.0833395272,0.2817313671,-0.1454360038,0.0304209553,0.0082155261,0.0834249482,-0.0202473532,-0.2850773931,-0.5693511367,-0.306091696,-0.2529954314,-0.110071376,-0.055023253,0.1214481741,0.2450508326,-0.0811236277,0.1239974424,-0.4648569524,-0.0455252044,-0.3640047908,-0.0893353149,0.2416918129,-0.2270824611,-0.4088026285,-0.0433627553,-0.0247738864,-0.0982614979,0.1594512016,-0.3808137774,-0.4189061522,-0.2150374353,-0.5069791079,0.1753206104,0.4170131087,0.3016415834,-0.3365058005,-0.1032984853,-0.0849574581,-0.1637304872,0.1886581033,0.2175232023,0.361068815,0.2080634087,-0.0973239541,-0.0154065378,0.5112137794,0.4535945356,0.1888438463,0.2588811815,-0.0305912402,0.6296029091,0.0801544785,-0.3526650667,0.0025885261,-0.023501467,0.2995642126,0.3215641677,0.2464164197,-0.1745372266,-0.0990550295,0.1481453329,-0.3503153026,-0.3976812661,0.2206423879,-0.1181029454,0.0334850885,0.2116403431,-0.3735963106,0.0567202605,-0.0505013429,-0.3560997844,0.3902478516,-0.1315305829,0.12844491,-0.5994895101,0.1178519204,-0.4363968372,0.1391756386,-0.122389324,0.1642433107,0.1163776144,-0.0674035996,0.3922002017,0.0280274823,0.5079416037,-0.1339227855,-0.2179355323,0.1831876785,-0.0962787792,-0.8417409062,-0.1688054353,-0.0777037293,-0.1152691245,0.2761751115,0.3028251827,-0.5176626444,-0.2508129478,0.2035355717,0.1481360197,-0.172186479,-0.2493509054,-0.2638268173,-0.4485579133,-0.3102764189,-0.391410023,-0.1227435842,-0.0523416288,-0.1175897717,-0.128795594,0.157445848,-0.1702956557,0.1424729526,-0.1609002799,0.2610478997,-0.0209659953,0.0318266377,0.4652027786,0.1336735934,-0.5750659108,0.5631842613,-0.1878596395,0.2896809876,0.1397071928,-0.1725722849,0.3634237349,0.5340476036,-0.0688849315,0.0670909286,0.0395315662,-0.0713623613,-0.1149133965,0.2556439042,0.5111373663,-0.3491491377,-0.1388002634,-0.3376772404,0.5596758127,0.081665054,-0.4084248245,0.1992585361,0.2666523755,-0.2441545278,0.2821602821,-0.353872776,1.083216548,0.0448712185,0.3187460303,0.0813922212,-0.2079670578,0.8912687898,-0.0198471453,0.0250940006,-0.2386832833,-0.0400296748,0.0160512552,-0.185227409,0.0590646639,0.0831571892,-0.0264114141,0.2931695879,0.2207300961,0.0986571833,-0.1022955328,0.5956616402,0.0409725495,-0.0823026821,-0.3292581141,0.0409089811,-0.0130933505,0.3151535094,0.0684854537,-0.1459354907,-0.2229434699,-0.4063076079,-0.2671608627,0.0490466245,0.0589771084,0.1323080659,0.1028683484,-0.1178767234,-0.2950004339,0.0290810149,-0.0423934646,0.0197905581,-0.2558149695,-0.0875733867,0.0039718049,-0.307387948,-0.1308574677,-0.2202379853,0.2706197202,-0.128891781,-0.1828336418,-0.2363054305,0.040532995,-0.1591939926,-0.4621099532,0.3166395426,0.2644231021,-0.0829142183,-0.2963554263,0.3617571294,0.1117045134,-0.0023635693,0.0293600634,0.3525855541,0.0783955455,0.3619930446,0.1353421211,-0.1916669607,-0.068237111,0.3341464996,0.5289818048,-0.0918833911,0.2554853857,-0.1154194176,0.0519082695,-0.0509983264,-0.1693790853,-0.0838309675,-0.3399846554,0.1229299158,-0.0123681715,0.0692097992,0.1314321756,0.043052379,0.0909333602,-0.2170269042,-0.1089520827,-0.4528342783,0.0652750731,0.2933410704,-0.1723696291,0.4097402394,0.0405625179,-0.1246984825,0.2502573431,0.3126403987,-0.160583213,0.2575595677,0.0629959852,0.2446402311,0.6299233437,0.0180326346,0.0886121467,-0.1104044616,-0.0822519511,-0.0571030416,-0.1773712635,-0.0963757336,-0.2691383064,0.226451546,0.1445830464,0.0587295145,0.0979401767,-0.235622853,0.0319421701,0.0085718865,0.2074660957,-0.0665720105,-0.0394215733,-0.025018122,-0.1616052687,0.038482774,0.1117451712,0.3407118022,-0.3641946316,-0.0427351557,0.3421165943,0.0081332903,-0.2114962935,-0.097881496,-0.2852179408,-0.0305371918,0.1042067781,-0.0482181124,0.1110077649,0.0064332704,-0.0299760215,-0.182214573,0.3628311157,0.5864511132,-0.2662344277,0.0876504779,-0.0524091534,0.0244773403,-0.1797003895,-0.0384821817,0.2230917215,0.2693853676,-0.0885820761,0.2684659064,0.3248122931,0.0253777429,-0.1227575392,-0.0197178759,0.6393306255,-0.1933536083,0.0859199911,-0.2401420772,-0.0418787934,-0.1109327897,-0.0177884232,-0.033447478,0.0075639933,-0.0986706764,0.2037725449,0.2015048265,-0.0354351066,0.0525603145,0.2308350652,-0.1845732927,0.0165102035,0.7022256851,0.2064049393,0.3067231476,0.1423377246,0.7357049584,-0.1173909679,-0.0500354692,-0.1399211138,0.5564060211,0.2685464919,-0.23760432,0.0679926127,-0.113695167,-0.142375499,0.0482382588,-0.0440424904,-0.3282141089,0.2714611888,-0.0276606083,-0.19461371,-0.7354896069,0.1625507921,0.078610912,0.1073667407,-0.1520143598,0.186259672,-0.0700537339,-0.0500729792,0.2191655636,0.518214345,0.2908774912,0.3593863547,-0.0064822119,-0.1620459557,0.2486836016,0.1770858765,-0.2094998658,0.1545950919,0.3400497139,0.0780140311,0.3201011121,0.0420921259,-0.1034836248,-0.0672274828,0.0774066821,0.0293805487,-0.3207194507,0.5537858009,-0.2837340832,0.0801004693,-0.0178243984,0.0625470951,-0.6169649959,-0.1906497478,0.2139840126,0.0052100141,0.1693683714,-0.4606561363,0.0140757384,0.1770661026,0.5300937891,0.0472881086,0.0663498789,-0.1545783281,-0.2441277504,-0.7898305655,0.1378142983,-0.2290432304,0.1441519856,-0.0481810831,0.3934707642,0.0328746065,0.3112465441,-0.0222207326,0.1306433827,-0.5399712324,-0.1363769323,-0.1759830862,-0.0487984419,-0.2168409973,-0.058268033,0.0051192804,-0.1132188812,0.100979358,-0.299600333,-0.0770939291,0.0974140465,0.0709076896,0.0748329163,0.0544952527,0.4069505334,0.2345132083,0.4126707911,-0.0002770416,-0.2844558954,-0.5316898227,0.0455550067,-0.4169572592,0.099364154,0.0215483587,0.2074605078,-0.4415888488,0.1541761905,-0.1178327352,-0.0703340098,0.1311770678,0.0565228797,0.0364829786,-0.0084074121,0.1445237994,0.1382408738,0.2334769666,0.418356061,-0.2253128737,0.1311776191,-0.1955818385,-0.2739392519,0.2897921801,-0.3623568416,-0.1348870099,0.1206963509,0.2468204796,0.0702302977,0.0666694567,-0.6980174184,-0.2639483213,0.3107896149,-0.0424809866,-0.1835703999,0.2500554025,-0.1316495985,-0.4783776999,-0.0227041468,0.0441046171,0.1565813869,-0.381315887,0.0116341263,-0.2301677018]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/197","title":"Scientific Papers only downloading Pubmed","comments":"Yes, it looks like the right thing to do ","body":"Hi!\r\n\r\nI have been playing around with this module, and I am a bit confused about the `scientific_papers` dataset. I thought that it would download two separate datasets, arxiv and pubmed. But when I run the following:\r\n\r\n```\r\ndataset = nlp.load_dataset('scientific_papers', data_dir='.', cache_dir='.')\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5.05k\/5.05k [00:00<00:00, 2.66MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.90k\/4.90k [00:00<00:00, 2.42MB\/s]\r\nDownloading and preparing dataset scientific_papers\/pubmed (download: 4.20 GiB, generated: 2.33 GiB, total: 6.53 GiB) to .\/scientific_papers\/pubmed\/1.1.1...\r\nDownloading: 3.62GB [00:40, 90.5MB\/s]\r\nDownloading: 880MB [00:08, 101MB\/s]\r\nDataset scientific_papers downloaded and prepared to .\/scientific_papers\/pubmed\/1.1.1. Subsequent calls will reuse this data.\r\n```\r\n\r\nonly a pubmed folder is created. There doesn't seem to be something for arxiv. Are these two datasets merged? Or have I misunderstood something?\r\n\r\nThanks!","comment_length":9,"text":"Scientific Papers only downloading Pubmed\nHi!\r\n\r\nI have been playing around with this module, and I am a bit confused about the `scientific_papers` dataset. I thought that it would download two separate datasets, arxiv and pubmed. But when I run the following:\r\n\r\n```\r\ndataset = nlp.load_dataset('scientific_papers', data_dir='.', cache_dir='.')\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5.05k\/5.05k [00:00<00:00, 2.66MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.90k\/4.90k [00:00<00:00, 2.42MB\/s]\r\nDownloading and preparing dataset scientific_papers\/pubmed (download: 4.20 GiB, generated: 2.33 GiB, total: 6.53 GiB) to .\/scientific_papers\/pubmed\/1.1.1...\r\nDownloading: 3.62GB [00:40, 90.5MB\/s]\r\nDownloading: 880MB [00:08, 101MB\/s]\r\nDataset scientific_papers downloaded and prepared to .\/scientific_papers\/pubmed\/1.1.1. Subsequent calls will reuse this data.\r\n```\r\n\r\nonly a pubmed folder is created. There doesn't seem to be something for arxiv. Are these two datasets merged? Or have I misunderstood something?\r\n\r\nThanks!\nYes, it looks like the right thing to do ","embeddings":[0.3959166706,0.1504321545,0.018072091,0.203631103,-0.0103302868,-0.0243965331,0.1065785289,0.3857787549,-0.0738985315,-0.3226345181,-0.1923356205,0.2351146489,0.0538930297,-0.3625418842,0.3902675509,0.1142595634,0.2550103664,0.1102476642,0.2166461796,-0.378508687,-0.1251147836,0.2891970575,-0.186277926,-0.0467706919,0.1726790518,0.1048895493,-0.1225016266,0.1983986497,-0.3486183584,-0.2520948946,0.2520580888,0.4304970801,0.1826574802,0.3474259973,-0.0001238361,-0.0867364258,0.3407195807,-0.0535739921,-0.1006907299,-0.1782799214,0.0591302216,-0.2186222374,0.366430223,-0.1232714057,0.1099014804,-0.5806162953,0.1716650277,-0.0899202153,0.0429002047,0.2748803198,0.1079711989,-0.018969804,0.0519718863,0.0863325447,0.37932235,0.0908046737,0.0856125653,0.2919959724,0.2590969205,0.1044060588,0.140103668,0.2420820594,-0.1500318348,0.1244707033,0.3200564682,0.4479098618,0.0883137435,-0.4270432591,-0.1110313982,0.2034568191,0.3907706738,-0.018475445,0.0331055522,-0.190754503,-0.1291194856,-0.0778770372,0.0411274917,0.4252302349,-0.1799102575,-0.063308835,0.047130309,-0.3850339949,-0.1660058796,0.524559319,0.1824227571,0.1062224731,-0.0994695947,0.3471193314,0.2172853351,0.1406968981,0.0860783234,-0.3598834872,0.2600196302,0.3090045452,-0.2331109494,-0.0677013472,-0.0396831483,-0.1418696493,0.2105996907,0.3073239028,0.1007979512,-0.0497795977,0.10980542,0.0696887672,0.5097495914,-0.1955111623,-0.0201111883,0.2931278944,0.2621225715,-0.1744323969,-0.0710667446,0.0385320112,-0.0673168674,0.1324801594,-0.2351941466,-0.64065063,0.1537436992,-0.5166800022,0.3932346702,0.011439967,0.2099223137,-0.1469558477,-0.2659932673,0.0273831021,0.1487458348,0.2456783801,0.0556613542,0.261498034,-0.1557826251,0.093250528,-0.0705886707,0.0177747402,-0.322021544,-0.0756882355,0.2050869018,-0.0052635567,0.1426070035,-0.0776145309,0.0184746031,-0.165638417,0.3408052623,-0.0992343426,0.2343129367,0.4231782556,0.2968872786,0.5744467378,0.0160214528,-0.15369609,-0.232446149,0.1894479543,-0.3552799523,-0.068483375,-0.090331696,0.0795289129,-0.2514751554,-0.0666767955,0.1043176726,0.1797104329,0.2141397148,-0.3368316293,-0.0278734192,0.0331107825,0.009999915,-0.3405792415,0.0950058252,0.0607407391,0.0837369412,0.0277316421,-0.3203540742,-0.0721770898,0.243752107,0.2607187033,-0.2989886701,0.1743740439,-0.1365516484,0.061175596,0.8015086651,-0.1988377273,-0.4292069674,0.2969690561,0.041076459,-0.1222852096,0.1047636494,0.3437669873,-0.2408702821,-0.0418527685,0.0213796869,0.296516031,0.0200532749,-0.1974789649,-0.2159682661,-0.0516750962,0.4024026394,-0.1249417812,0.3167509437,0.1159106866,-0.0302203409,0.208743766,0.451228857,0.1697276384,0.1933346093,0.1471362561,-0.0764680654,-0.1473486871,0.2330598682,-0.2036332786,-0.0645066872,-0.0079369675,-0.4860935211,0.2228990793,0.3158966601,-0.0557684414,-0.2538231313,-0.5095009208,-0.0705939457,-0.3024183214,-0.0001293868,0.2175567001,0.0525574349,0.1448907554,-0.3396776319,-0.0165594202,-0.1129389182,0.0077126729,-0.6607295275,0.1853622943,0.0162894428,0.0253454074,0.0308526941,0.0565715842,-0.0517388284,-0.0855785757,0.2685313523,0.3163889349,-0.0050182655,0.3895972371,0.4091255665,-0.0368640274,0.1770050377,-0.2470318526,0.1500460505,-0.0738253817,0.2780334353,-0.2206709832,-0.3823261559,0.1423967779,-0.1340411901,0.3008796275,0.1966796517,-0.0714786723,0.043148309,-0.0788626373,-0.3979916871,-0.0908191428,0.1665375233,0.290437609,-0.3282107711,0.1651885957,-0.2075400501,0.1347189993,0.1167022139,-0.0285308454,-0.1049225256,0.1495211869,-0.3517490923,-0.3019632399,-0.149885118,0.4068022966,0.8257782459,0.2022768557,0.5080541372,0.1016127765,0.0147803007,-0.1808220446,0.1325136572,0.1306848973,-0.0841655359,0.1970900893,-0.0200862996,-0.0669029355,-0.4762544334,0.4494064748,0.1441943198,0.0980720297,-0.2904407978,-0.0668784678,-0.3540065289,-0.2080260515,-0.5321944952,-0.2256080508,-0.2052943259,-0.308811903,-0.1042344496,0.0736040026,0.0636403263,-0.0994919688,-0.480512917,0.0806536973,-0.185642913,0.0115411673,0.0946913436,0.0008133942,-0.2257717997,0.0133583648,0.0661741123,0.4453667402,0.5318757296,0.0027830836,-0.0285802186,0.0232038219,-0.3873758316,0.1708813161,-0.2471655458,0.1195372418,0.1543351561,0.1095364019,0.2337512821,0.1351428479,0.0483170189,-0.3091568947,-0.0081614265,-0.0743774995,0.0891738161,-0.0587997958,-0.3182567656,-0.7676833272,-0.3253721893,-0.1414451301,-0.2651926577,0.1922815442,0.1167852432,0.1485257298,-0.0807698369,0.1748157591,-0.3000010848,-0.0156530999,-0.399220705,-0.1136221588,0.3586472273,-0.2526381612,-0.4404665232,0.1120961308,0.0899282768,0.0243354049,0.1258884072,-0.2701312006,-0.0983069539,-0.0105367592,-0.5769565105,0.235731706,0.2996129096,0.295748055,-0.4142804742,-0.1250360459,-0.0718086883,0.107732892,0.0604472868,0.2299746722,0.4317383468,0.0645113364,-0.111113809,-0.1430903822,0.5301935077,0.3745293915,0.1966028512,0.1579799801,-0.0120618697,0.4833180606,0.0491959676,-0.2183644921,0.1061320603,0.0035306632,0.0830086768,0.3890366554,0.2430838794,-0.2531003952,-0.0959959701,0.0715548694,-0.4126876891,-0.2853083909,0.2189106345,-0.1559254974,0.101811111,0.3217453361,-0.396060288,-0.068595387,-0.0600171238,-0.4515447021,0.3622123301,-0.0620398186,0.1302859187,-0.6392335892,0.1667152345,-0.4611138105,0.1280200183,-0.1261443943,0.0866442546,0.079898268,-0.0911503062,0.3281401694,-0.0172137897,0.2971670032,-0.0334694833,-0.2468724102,0.0876735374,0.0397714116,-0.7149338722,-0.0161625799,-0.0053610317,0.0256599765,0.0075423745,0.1599552035,-0.3036321402,-0.1913907826,0.2191802114,0.1463568956,-0.2061894089,-0.1700702608,-0.308626622,-0.5989918113,-0.186627239,-0.3636333644,-0.1232446581,-0.1428897232,-0.2532975376,-0.1071675122,0.1701985747,-0.0558529533,0.244312495,-0.1589072049,0.2367947698,-0.0523713641,-0.068072848,0.4633762836,0.110139437,-0.400960207,0.483864367,-0.0957050174,0.2324910313,0.0597433522,-0.4475977719,0.4421556294,0.3946597874,-0.1028086171,0.0271666925,0.1706982255,-0.320160836,-0.0291039534,0.139389798,0.4731815755,-0.3392164111,-0.2940172851,-0.5139220953,0.5263596177,0.0111339986,-0.3910853863,0.1609169692,0.0255699866,-0.2953643501,0.2366956621,-0.2221666723,1.1836323738,0.0553733073,0.4560599923,0.0363642089,-0.1406869292,0.833440721,-0.103254281,-0.0099316025,-0.1352983415,-0.053173542,-0.1410338432,-0.1605213434,0.0734583139,0.0908277854,-0.022463249,0.2250943333,0.19460693,0.0157102309,0.021462962,0.5123224854,0.1337672919,0.0966287777,-0.2895163298,0.0654652864,0.0680751726,0.1267767698,0.0490647778,-0.2040661275,-0.4792118967,-0.4042505622,-0.0556896143,0.0749620199,0.4498747587,0.086523369,-0.1262577474,-0.2362395078,-0.3070617616,0.1641602814,0.0665439218,-0.1244446933,-0.3125560284,0.046713341,-0.1435875297,-0.1976302564,-0.142303139,-0.2042481154,0.2308390141,-0.246438846,-0.184653461,-0.0379326306,-0.0007920843,-0.2449838221,-0.4237146378,0.3350668252,0.4471787214,-0.0183127113,-0.2794903517,0.442555964,-0.1675494909,0.0427529477,0.0550897159,0.4240184426,0.0461632647,0.4422499835,0.1828742027,-0.300884068,-0.1109974012,0.3273270428,0.8248306513,-0.0431772582,0.2815077305,-0.0814582109,-0.0329002514,-0.1332996488,-0.1253096759,-0.0879713967,-0.2788262367,0.0866630897,0.0122037129,0.2875427008,0.1617839932,0.0485396013,-0.036867518,-0.2480215281,-0.2995178998,-0.3827372789,-0.0802961886,0.0919744447,-0.3209562898,0.3896977007,0.0925335884,-0.1451360285,0.3206012249,0.1467559934,-0.1760659665,0.2421977818,0.0963970497,0.2509289682,0.4452475309,0.0725999475,0.0134416949,-0.1692312956,-0.0901200473,-0.0655807108,-0.1003449187,-0.1265940815,-0.1966161281,0.2218548805,0.2269011438,0.0116912508,0.0439522229,-0.2060518712,-0.1419996917,0.1232061684,0.0185772907,-0.1400308311,0.0543903373,0.1015706211,-0.1706557125,0.0736386254,0.0594827086,0.3314638138,-0.3500518501,0.1066403836,0.2954829037,0.0752548799,-0.0708181933,-0.0947223678,-0.2812251747,0.0102554644,0.2322774082,-0.1546034217,0.3164124787,0.0310524553,0.0743910745,-0.2701033354,0.270428896,0.793800354,-0.2367685139,0.1070799679,-0.1344353855,0.0135939606,-0.2117683142,0.014234866,0.1003419906,0.2355040312,-0.2498716414,0.1640328914,0.2003903538,-0.049401965,-0.1720753014,0.0022434751,0.7910020947,-0.193287164,0.090110369,-0.2300726324,-0.023328511,-0.0163496099,-0.0649897009,0.0913848281,-0.082042411,0.0882481411,0.0751049817,0.3778336048,0.0460765958,-0.0734473243,0.2012998462,-0.1045450717,-0.0622228496,0.759754777,0.1589895338,0.3499136269,0.2758592069,0.5091834068,0.0867078304,0.0014467956,-0.2250107974,0.411913842,0.3206475377,-0.1164441034,-0.1538159102,-0.2544588745,-0.0639634132,-0.0305803176,-0.0740676746,-0.1810921282,0.0898816958,0.01694078,-0.1669659019,-0.6380234361,0.3238992393,-0.0037070941,-0.0981229767,-0.1002142802,0.0410306007,-0.0284827556,0.0125309359,0.4013751745,0.3165032268,0.1383078098,0.2828140557,-0.0200626906,-0.1203311235,0.2423337102,0.0721850768,-0.1890788078,0.2267535627,0.2634907663,0.1558957398,0.2577943504,0.0997397974,-0.0526990853,0.0830609053,0.2113165557,-0.0267625656,-0.3071100414,0.5059782267,-0.1683929116,0.0424230471,-0.053442765,0.0968110785,-0.6198765635,-0.2533765733,0.1934898049,0.0094019435,0.1316732466,-0.4019291103,0.0083171567,0.1127972677,0.6105602384,-0.0263482053,0.2146595567,-0.1252292842,-0.3470762074,-0.7149713635,0.2636069655,-0.1125466749,0.1908268183,-0.1232085899,0.2924100757,0.2558422387,0.3376351595,0.1065177545,0.0320811905,-0.6842156649,-0.0058855885,-0.011755948,-0.1412537545,-0.3148771226,0.1568480134,-0.1587218344,-0.0317599699,0.1845270842,-0.3736484647,-0.0404720008,-0.0020952162,-0.1213476583,0.1859835535,0.2038654536,0.5454018116,0.1600001007,0.4761904478,-0.1611057073,-0.2588069439,-0.397616744,0.1869221181,-0.4153163731,0.1435747594,-0.0464464165,0.0658020452,-0.4542770386,0.3863350451,-0.029422421,-0.0195014495,-0.0189799555,-0.1669964045,-0.0153064383,-0.0615109913,0.0207030885,0.0632893145,0.1793730259,0.3121171892,-0.1765624583,0.0040680016,-0.1868929863,-0.0939822868,0.2849811316,-0.2870170176,-0.1885413378,0.0636232421,0.1083057821,-0.085563615,0.060708642,-0.6846598983,-0.1163038984,0.3382411003,-0.1090930477,-0.1235025153,0.0978627577,-0.0661016479,-0.3595549464,-0.1882988811,0.1866716594,0.0227386728,-0.1881103665,-0.1432878226,-0.2459077239]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/197","title":"Scientific Papers only downloading Pubmed","comments":"Now if you don't specify which part you want, it raises an error:\r\n```\r\nValueError: Config name is missing.\r\nPlease pick one among the available configs: ['pubmed', 'arxiv']\r\nExample of usage:\r\n\t`load_dataset('scientific_papers', 'pubmed')`\r\n```","body":"Hi!\r\n\r\nI have been playing around with this module, and I am a bit confused about the `scientific_papers` dataset. I thought that it would download two separate datasets, arxiv and pubmed. But when I run the following:\r\n\r\n```\r\ndataset = nlp.load_dataset('scientific_papers', data_dir='.', cache_dir='.')\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5.05k\/5.05k [00:00<00:00, 2.66MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.90k\/4.90k [00:00<00:00, 2.42MB\/s]\r\nDownloading and preparing dataset scientific_papers\/pubmed (download: 4.20 GiB, generated: 2.33 GiB, total: 6.53 GiB) to .\/scientific_papers\/pubmed\/1.1.1...\r\nDownloading: 3.62GB [00:40, 90.5MB\/s]\r\nDownloading: 880MB [00:08, 101MB\/s]\r\nDataset scientific_papers downloaded and prepared to .\/scientific_papers\/pubmed\/1.1.1. Subsequent calls will reuse this data.\r\n```\r\n\r\nonly a pubmed folder is created. There doesn't seem to be something for arxiv. Are these two datasets merged? Or have I misunderstood something?\r\n\r\nThanks!","comment_length":34,"text":"Scientific Papers only downloading Pubmed\nHi!\r\n\r\nI have been playing around with this module, and I am a bit confused about the `scientific_papers` dataset. I thought that it would download two separate datasets, arxiv and pubmed. But when I run the following:\r\n\r\n```\r\ndataset = nlp.load_dataset('scientific_papers', data_dir='.', cache_dir='.')\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5.05k\/5.05k [00:00<00:00, 2.66MB\/s]\r\nDownloading: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.90k\/4.90k [00:00<00:00, 2.42MB\/s]\r\nDownloading and preparing dataset scientific_papers\/pubmed (download: 4.20 GiB, generated: 2.33 GiB, total: 6.53 GiB) to .\/scientific_papers\/pubmed\/1.1.1...\r\nDownloading: 3.62GB [00:40, 90.5MB\/s]\r\nDownloading: 880MB [00:08, 101MB\/s]\r\nDataset scientific_papers downloaded and prepared to .\/scientific_papers\/pubmed\/1.1.1. Subsequent calls will reuse this data.\r\n```\r\n\r\nonly a pubmed folder is created. There doesn't seem to be something for arxiv. Are these two datasets merged? Or have I misunderstood something?\r\n\r\nThanks!\nNow if you don't specify which part you want, it raises an error:\r\n```\r\nValueError: Config name is missing.\r\nPlease pick one among the available configs: ['pubmed', 'arxiv']\r\nExample of usage:\r\n\t`load_dataset('scientific_papers', 'pubmed')`\r\n```","embeddings":[0.3337525129,0.1518255174,0.0403823219,0.1220906079,-0.011581541,0.0701765418,0.1087679043,0.3869033754,-0.1169545278,-0.2211576104,-0.1246671453,0.3359685242,0.0420467369,-0.3219137192,0.3353407681,0.0597895607,0.1705282032,0.1198785305,0.2469612211,-0.3221702874,-0.1509200633,0.3626807332,-0.2007835805,-0.0750292316,0.1717316508,0.1704218984,-0.1363465637,0.2355844378,-0.2803621292,-0.2833781838,0.3049050868,0.4118289053,0.1248244494,0.3285555542,-0.0001224009,-0.0510751307,0.3996664286,-0.1145327687,-0.1240762919,-0.2439041436,0.0256197173,-0.2772782743,0.361112684,-0.1576248407,0.1032313108,-0.501804173,0.2266834974,-0.1019288227,0.0027055452,0.2615558207,0.121453315,-0.082328327,0.1363512725,0.0036474764,0.3264043629,0.0186310876,0.0906237736,0.3079102635,0.1788234711,0.0836476088,0.1962370723,0.2178081572,-0.1276474744,0.0741181374,0.3337646723,0.3773284554,0.0491853617,-0.4533298314,-0.0953774229,0.2512944937,0.4695162475,-0.0825927407,0.0370471664,-0.2068697661,-0.0401060581,-0.0887057558,0.1081100479,0.3900105357,-0.2233262211,-0.0482567102,0.0210090317,-0.3863975704,-0.1963814646,0.5864767432,0.1692327112,0.1530348957,-0.0870963484,0.3638678789,0.2177418768,0.1375072896,0.0339366943,-0.3225851357,0.255410403,0.2819865346,-0.2922056913,-0.0079085166,-0.0350400247,-0.2091565132,0.1889663935,0.3261244297,0.0750263706,-0.0670474544,0.200577721,0.1209650263,0.4643370211,-0.1520892382,0.0867839679,0.3539930284,0.215610683,-0.1866046041,0.0340204164,0.0182454176,-0.0772188455,0.0717442855,-0.2296784818,-0.6142804027,0.1409284472,-0.4055893123,0.4478225708,-0.0252392478,0.2138482481,-0.1062910482,-0.2344533354,0.0817294493,0.105173789,0.2092121243,0.0853358954,0.1764572263,-0.1176484972,0.0951757878,-0.0868492872,0.0165046081,-0.3610500395,-0.0581570975,0.2363769412,-0.0132135935,0.2610935569,-0.0931467563,0.0739073083,-0.0901588872,0.2440550923,-0.0596809015,0.1449033171,0.4964265227,0.3115926981,0.5195352435,-0.0001566742,-0.1992076337,-0.2772220671,0.1592060477,-0.3895844519,-0.1244929358,-0.1183593273,0.0739863664,-0.3111489117,-0.0614527352,0.1078479663,0.2107418031,0.1814860553,-0.358273834,-0.1082867756,0.0256287791,0.0402183868,-0.3753201663,0.1704992652,0.1312366128,0.0658913404,-0.038904395,-0.2859943211,-0.0642904639,0.226696834,0.0782630667,-0.3078137934,0.1558772773,-0.1210069731,0.0539570004,0.864571929,-0.2199848443,-0.407153964,0.2794377208,0.0544365272,-0.1435594708,0.091761142,0.3109305203,-0.2846711874,0.0514984578,0.0913101882,0.2885004282,-0.0125902724,-0.24612014,-0.1448047906,-0.1127997264,0.3086381257,-0.0759325698,0.2943561971,0.1680896133,-0.0156032005,0.1497993022,0.4100165963,0.1875562817,0.2137866467,0.226598829,-0.1147905588,-0.1092510372,0.2086015493,-0.1760005653,-0.1642107219,0.0664224625,-0.4373755455,0.2642482817,0.3202396631,0.0615606867,-0.2903928161,-0.520103991,-0.0640273318,-0.3106152415,-0.0129940351,0.2909795046,-0.0014562281,0.0786508992,-0.3556837142,0.0474463776,-0.1270782799,0.0147784054,-0.7680270672,0.156828627,-0.00258573,-0.0141226333,0.0891618133,-0.0303555317,-0.0414165594,-0.0181819033,0.2261610627,0.4155773222,-0.0318803489,0.3567320406,0.3559613824,-0.1002450064,0.116301313,-0.2353426963,0.1572002321,-0.1083389297,0.2621348798,-0.2087305188,-0.3446258903,0.1056321934,-0.1530318856,0.3154357672,0.1534233242,-0.0027962315,0.0623033755,-0.1272726655,-0.4065901339,-0.1481285244,0.170978874,0.2773890793,-0.3150270879,0.0903303102,-0.2642230093,0.0373467393,0.1256252229,0.0065444163,-0.1745367348,0.1829848588,-0.3114659488,-0.2189500183,-0.1368601173,0.4669195712,0.9032630324,0.2567404509,0.3886661232,0.0915179029,-0.0587531216,-0.1681360751,0.1890705228,0.1316781193,0.0125590973,0.1673580706,-0.0691405162,-0.0487458222,-0.4017603397,0.2793741822,0.1685538441,0.0622381717,-0.302798152,-0.0377122723,-0.3783569634,-0.1974488497,-0.5116115212,-0.2571055293,-0.2551214099,-0.3293745518,-0.167897597,0.0100614792,0.0549546406,0.0197101347,-0.5319440365,0.0452209786,-0.1329424083,-0.1507658213,0.1229515299,-0.0229472239,-0.249567166,0.0073658298,-0.065862909,0.3898340762,0.4878613055,-0.0702571422,-0.0429270044,0.0528673567,-0.402985543,0.2302953154,-0.2439556867,0.2364079803,0.2415213734,0.1290678531,0.2934623361,-0.0029641523,0.1573615968,-0.1802592874,0.0249048714,-0.0403277613,0.1010967419,-0.1088800505,-0.2754941285,-0.8172607422,-0.3088196218,-0.1809062213,-0.238820076,0.189415589,0.1692732275,0.3574635386,0.0208667945,0.1280640513,-0.3271786273,0.0576983094,-0.368775785,-0.15246813,0.4007768631,-0.2736347914,-0.3794854283,0.1096101403,0.1409732252,0.0388059467,0.1496944129,-0.2708301842,-0.0501307733,-0.0078999493,-0.5291983485,0.1509534419,0.336230427,0.2992244065,-0.3369193673,-0.1137447059,-0.1398274899,-0.065526031,0.1196309403,0.242737487,0.5703095198,0.136684224,-0.0611464344,-0.1879300177,0.4702349305,0.3327802122,0.2193660885,0.1394729614,-0.0807742178,0.5153936148,0.0709885135,-0.2541883886,0.1776622981,0.0340987071,0.1152197719,0.4011803269,0.1475609094,-0.3179173172,-0.227119416,0.0292134136,-0.3635054827,-0.2917662859,0.2263847142,-0.1641098261,0.1019237116,0.2606688142,-0.4112976491,-0.1092570648,-0.0217902586,-0.4203908443,0.4306922257,-0.1537828892,0.153131023,-0.6548845768,0.167355597,-0.4477037787,0.1161155775,-0.2242058814,0.1797571033,0.0131827258,-0.1094079316,0.2008357048,0.0720379725,0.3785326183,-0.0923903063,-0.2777590454,0.1635866761,0.111032553,-0.692882061,-0.054513,0.035483595,0.0671537295,0.0704635903,0.1560642123,-0.3221655488,-0.2182293236,0.280612886,0.1294167936,-0.2067828029,-0.2330890596,-0.3231199682,-0.5719004869,-0.2354187965,-0.4505118132,-0.1429520249,-0.0453155562,-0.1363145411,-0.0938038081,0.2150984108,-0.120315522,0.2540311813,-0.0608652607,0.2660250068,-0.0378558226,-0.0336808376,0.4772066772,0.0953952372,-0.2791570127,0.4296422601,-0.1428151727,0.2030206025,0.1330480874,-0.372007221,0.4784093499,0.3028566837,-0.1099802628,-0.0059058745,0.1960080713,-0.2528348863,-0.0089569036,0.1406580061,0.4020999968,-0.3468554914,-0.3203017414,-0.5828097463,0.5492745638,0.0031258843,-0.368829608,0.255199492,-0.0107274307,-0.3315703571,0.232607916,-0.2238222212,1.1526607275,0.1274229735,0.3265951872,-0.0009176816,-0.1020976976,0.8224450946,-0.0544585511,0.0806767046,-0.109561272,-0.0431543775,-0.1330513656,-0.2059556991,0.1671733856,0.1299716234,-0.0915976316,0.3078963161,0.1755507886,0.0021262406,0.0728295892,0.4751612246,0.1268336922,0.0042932448,-0.3501727283,0.0622022524,0.031312976,0.1564971358,0.0473201945,-0.2376621068,-0.4849183857,-0.4106020927,-0.0966625735,0.092669718,0.3221292794,0.0742439777,-0.0798615962,-0.1980776042,-0.2268861532,0.2078130096,-0.0088430773,-0.0584799461,-0.3350397944,0.1760788709,-0.1439324915,-0.2794754505,-0.2007133663,-0.1652783453,0.2625400722,-0.258205086,-0.2651183903,-0.0107109202,-0.0790306181,-0.2800617814,-0.321008563,0.2931113839,0.426056385,0.0106831659,-0.2837384343,0.4091001749,-0.0241195206,-0.0042427871,0.0562464558,0.403786093,0.0247805752,0.4783608317,0.1997638792,-0.3077810109,-0.1137280911,0.3047425151,0.7428827882,-0.0702211931,0.3021061718,-0.098068364,-0.0116638178,-0.1056960747,-0.103985481,-0.2096777558,-0.1769360453,0.0328191929,0.0312708691,0.2512495518,0.2184283882,0.0117088128,-0.0115737533,-0.1802456528,-0.1883938611,-0.4614041448,-0.1324356347,0.0887052044,-0.3724980056,0.4005108476,0.0721591413,-0.1330239326,0.3243182302,0.1654842496,-0.206547454,0.1549439132,-0.0359784178,0.2704541981,0.461045146,0.1112741306,0.0557077043,-0.2033537924,-0.0635239035,-0.1928080916,-0.1160019711,-0.1025902778,-0.2786972523,0.2238666862,0.1265098751,0.0489879958,0.0139116077,-0.232564345,-0.1972351223,0.1069355756,0.0127897076,-0.0910225883,0.0463527702,0.1213922799,-0.1388391554,0.01466492,0.0451745316,0.3172230721,-0.2853802145,0.117525816,0.3289766312,0.1133164391,-0.2091715783,-0.1367751658,-0.2675711215,-0.0148695065,0.1204617694,-0.1788603216,0.3095262051,0.0236592852,0.0665046945,-0.2103354782,0.2348409742,0.7377381325,-0.2688874602,0.1399053931,-0.0722812936,0.0265908856,-0.2600740492,-0.0356545895,0.1488880366,0.2221623361,-0.1608440131,0.1575269699,0.1493343711,-0.0743371323,-0.1829707325,0.0108462097,0.7443412542,-0.1675218791,0.1086150482,-0.1393347979,-0.0065567279,0.0411769971,-0.0155051639,0.1737532467,0.021766521,0.1040309519,0.0990992561,0.2958821654,0.0469290651,-0.0813338757,0.156587407,-0.1968526542,-0.0238937791,0.708239913,0.100194104,0.3292535245,0.2145409584,0.5007230639,0.0461714529,0.0884247422,-0.1524611861,0.4521008432,0.3075078726,-0.1250043213,-0.1658215821,-0.2766610086,-0.0663160011,-0.1019154266,0.0041472143,-0.2759516537,0.1621086746,0.0137073034,-0.2040797025,-0.6190449595,0.347325325,0.0289919302,-0.0881713033,-0.1512662768,0.0559381023,-0.0515767261,-0.1062292308,0.3673670292,0.3335018456,0.2425772101,0.293913722,0.0698903054,-0.1643188745,0.3510327339,0.0308395457,-0.1833092272,0.1991123706,0.3185027838,0.0908833966,0.3007444739,0.1006976664,-0.0364370868,0.0384108722,0.2743441164,0.0016553778,-0.2547133863,0.491663605,-0.1473433673,0.0651002005,-0.0114305736,0.116379872,-0.6328658462,-0.2241937071,0.2930049002,0.0958335623,0.1150795817,-0.3418160677,0.0142992102,0.0983314216,0.6334847212,-0.0781698525,0.127401799,-0.1227282137,-0.3408302665,-0.6841961145,0.2265511304,-0.1193956509,0.1381511986,0.0011204744,0.2716942132,0.2850787938,0.2446324378,0.08168865,0.0556474067,-0.6260475516,0.018999638,-0.0436976105,-0.2273301482,-0.2766310871,0.2152405232,-0.0874168724,-0.0394859053,0.1012854055,-0.351221323,-0.0289569311,-0.0786874741,-0.0838476643,0.2922539413,0.1682392657,0.5404808521,0.1115912944,0.5250933766,-0.1857176572,-0.2845462561,-0.3882062733,0.1177378371,-0.3753583133,0.1025675982,-0.0020777895,0.0630463809,-0.4657906592,0.3216525316,-0.0728624314,0.0139918439,-0.048585739,-0.2292243242,-0.0022874963,-0.0510740392,0.0779461116,0.1083182245,0.1899458617,0.29205814,-0.2019281238,0.0743772462,-0.1835720688,-0.1397275925,0.3861727715,-0.2082259357,-0.1940260828,0.098742038,0.0187201779,-0.1719149649,0.1618814617,-0.7090775371,-0.0625187159,0.370385319,-0.0811425224,-0.189641729,0.1016625986,-0.0853833854,-0.3409075439,-0.107022278,0.2071163207,0.0117872898,-0.2749647498,-0.1934283376,-0.1898615956]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/193","title":"[Tensorflow] Use something else than `from_tensor_slices()`","comments":"I guess we can use `tf.data.Dataset.from_generator` instead. I'll give it a try.","body":"In the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?","comment_length":12,"text":"[Tensorflow] Use something else than `from_tensor_slices()`\nIn the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?\nI guess we can use `tf.data.Dataset.from_generator` instead. I'll give it a try.","embeddings":[-0.1409347206,-0.0397877693,0.042681057,0.1479547024,0.2412027717,0.1365036815,0.2253275067,0.4326060116,-0.0407430939,-0.034751568,-0.1478944868,0.6063245535,-0.2089432776,0.0376926847,0.3290449977,-0.2219539136,-0.1060984656,0.229232952,-0.1127187237,-0.0273320451,0.0037291825,-0.0847089812,-0.1059894115,-0.0533125289,0.0232873075,-0.0261256713,0.0398791395,-0.1227423623,0.4147325158,0.0254283343,0.3529399037,0.0426781736,-0.0221231487,0.120950304,-0.0001088615,0.3769887686,0.016036164,-0.1654341072,0.0633081645,-0.0422181599,0.1617602259,-0.2579286993,0.1915842593,-0.2873087525,-0.0200824756,-0.0243218206,0.2428526282,-0.0389479399,0.2584976256,0.1032079309,0.1527888477,0.4725500345,-0.195410639,0.3402891159,0.0960488692,-0.1460416168,-0.1956927329,-0.2277406454,-0.1521445662,0.010088779,0.0377243683,0.1987466365,-0.2611368597,0.1088565737,0.4884866178,0.1284533143,-0.1605415493,-0.3202162087,-0.3000129163,0.4102810323,0.2144949585,-0.0926308557,-0.1621852517,-0.502638042,-0.1559762061,-0.2924831808,-0.2031363845,0.0992671177,-0.2353352308,0.1632748395,-0.1315229833,-0.1706542522,-0.2592573166,0.1231111512,0.2136853784,-0.0140927797,0.1752077788,0.1251571178,0.1896412075,0.0864109471,0.3205930293,0.0744159445,0.3580952883,0.3751791716,-0.3328224123,-0.0275167488,0.141003564,-0.4992485046,0.0309257675,-0.2498933226,0.3459661007,0.2064808011,-0.3022649288,0.2010513544,-0.0338648632,-0.0825774819,-0.3987597227,0.137548402,-0.1352126896,-0.5132371783,0.0437403806,0.0318692699,-0.2663530409,0.0051901969,0.0728644952,-0.4817666709,-0.1014212519,0.2612994313,-0.2883616984,-0.436068356,-0.191645667,0.0486845002,0.0202060174,0.2446092814,-0.1857520789,0.4476057291,0.1263399869,-0.1067712978,-0.2473656237,-0.0531635359,-0.227863431,0.1555784494,-0.0398786403,-0.0967223719,0.1651026905,-0.1151918322,0.1991506517,-0.153809458,-0.157710433,0.2040363848,0.2463674843,-0.0580866747,0.0871400014,0.1810988337,-0.0030199629,-0.0551205948,0.0459883213,0.1176653802,-0.2035280317,0.3085222244,-0.5068644881,-0.2608271539,-0.0624382868,0.1220096201,0.0991592929,-0.2327023,-0.3272660971,0.3365271389,0.0722722635,-0.1087449864,-0.1966184825,-0.1683814228,-0.4454084933,-0.3013353944,0.3179026544,0.0582353063,-0.4354985058,-0.1160627082,-0.0056251409,0.0094417995,0.2342298329,0.4028907716,-0.20084773,0.4870214164,-0.0689427555,0.0512404963,0.7628546357,-0.1090366319,-0.1744182855,0.2350111753,-0.142566666,0.201859206,-0.0546651818,0.4633917212,0.0382821709,-0.0035981666,0.3900398314,0.6519755721,-0.1680696011,0.1043068767,-0.0904858634,-0.3153561056,0.351675123,0.2890952229,0.0509330109,-0.0883956626,-0.2781863213,0.364259541,0.1518984139,-0.0347819105,-0.2093384713,0.0918368548,-0.1512374133,0.0095414827,-0.1966376007,0.0203612186,-0.6435192227,0.2719981074,0.2847804427,0.068855308,0.182345435,-0.1317048371,0.3301715553,0.0974379405,0.1050711647,0.1873271614,0.0280963201,0.0601170324,-0.0034380967,-0.0033786378,-0.3160276711,0.2175225765,-0.3933930695,-0.1119013503,-0.3515869975,0.2046046704,0.3983450234,0.0389854647,-0.0761199892,0.2748106122,-0.3397636414,-0.134377867,0.1280925721,0.2436286807,-0.0544043817,0.2813970745,-0.639939785,0.4543115795,0.0289470218,-0.03582789,-0.1048391387,0.1185706481,0.0123092225,-0.1750340015,-0.0130058052,0.2444089502,-0.1826384068,0.1557325572,0.2239937335,-0.218827799,0.173074618,0.055373352,0.1282815784,-0.0618989505,0.409250319,0.0608538985,0.1361492127,0.2556216717,-0.5245555639,0.2471510768,0.621614635,-0.0122969877,-0.0702073798,0.2216038257,-0.0114707034,-0.173538819,-0.0564182065,-0.0993017033,0.2085119784,0.1542688161,0.0315333083,0.0024953792,-0.1689344049,-0.2026905715,0.2441782951,0.1088306084,0.3747203946,-0.15201509,0.0334789343,0.1682522446,-0.1975075006,0.0622640513,0.2425721884,0.2097354233,-0.2045061886,0.0080176778,-0.2737818658,-0.3007135093,-0.0594698898,0.0085593201,0.1285738349,-0.1725124568,-0.0580981784,0.3286811709,-0.0086252093,-0.0245217551,-0.1247152984,0.2882916629,0.1935690939,-0.4367447197,-0.06699121,-0.1079648733,-0.1585000604,0.0669637546,0.3928582668,0.2262012511,0.2742345035,0.0618730262,-0.0851139426,0.1133975014,-0.0312774964,0.1951853484,-0.1249054372,0.1211434677,-0.035332337,0.1436289549,-0.1149662286,-0.1957823187,-0.1217628047,-0.4099520445,0.0668761209,0.1720976233,0.0277666431,0.1272213608,-0.2163198739,-0.149335131,-0.3243352771,-0.3930965066,0.0075057768,0.1274996549,0.3029610813,0.4151462317,-0.0617149882,0.2673887908,0.3298665285,0.0775909275,0.0952552333,-0.1441962421,0.5147873163,-0.2761670947,-0.2554987371,-0.1480540037,-0.1769650429,0.1673787683,0.3174764216,-0.6976945996,0.1841605306,-0.1121976823,0.0650158748,-0.1575435549,-0.0471613631,0.2654766738,-0.2212903202,0.0290913619,-0.0066147358,0.1698371023,0.1384602487,0.0045943856,-0.2097464651,0.3477366865,0.4148645401,0.3325720727,0.7137749791,-0.0945490599,-0.2266963869,0.2107809782,-0.2298327535,0.1912611127,-0.1388211101,-0.0923397467,0.0553899258,0.0199901313,0.0098026516,0.2472161651,-0.1048920676,-0.1949158907,-0.0836586505,0.1823012978,0.1333325207,-0.2287987471,0.2615645528,-0.2387947142,0.1463840157,0.1324264705,0.1794880331,-0.3844417632,-0.4394450784,-0.0189712215,-0.2031944096,0.4250231087,-0.0442808457,-0.3283933699,0.1166362613,-0.7471625805,0.0892269462,0.0864815265,0.2800333202,0.1133908927,-0.2771334052,-0.0091197267,0.0746543407,0.592330277,0.1039594114,-0.0685183182,0.1774148345,-0.0960858986,-0.5455280542,0.0380561911,-0.0524484366,0.0059044198,0.2378611267,0.1319678128,-0.3045978248,-0.5067415833,-0.0346506797,0.0340148248,-0.3627942204,-0.0718546286,-0.1863619238,-0.2242756784,0.0131869996,-0.3128806353,0.0829178393,0.2340627611,-0.1163791344,0.0287218653,-0.157947138,0.0085561909,0.2650246918,0.0112915998,0.19040595,0.5062750578,0.071772173,-0.0457877591,0.0511225313,-0.2046560943,0.2081555873,-0.0787354186,-0.1815455556,-0.0267605558,0.120752044,0.3986008167,0.0089777289,0.0246591195,0.1277266592,-0.1983283311,0.0316771232,-0.2341875434,0.3961806893,0.0740638003,-0.2854062319,-0.1044449061,-0.4949408174,0.5380278826,0.2831946015,0.1325812042,0.4806931615,-0.1958259791,-0.3546281457,0.4835780561,0.3242929578,0.4913548827,-0.4050786793,0.5139689445,0.7280606031,0.5370472074,0.9396314025,-0.3598246574,0.2802140117,-0.1159181222,-0.3620766401,-0.0224016439,-0.2598165274,-0.4066272974,0.0639798865,0.0816169977,0.2466543317,0.0998922065,0.2375540733,0.1950658113,0.3606855869,0.3778621256,-0.4248371124,-0.1729185879,0.0875358135,-0.0187047329,-0.2411187589,-0.088189058,-0.1908686906,0.1960254759,0.063590467,-0.0205906183,0.0938665718,-0.1576786786,0.2796302438,0.0747307539,-0.0472754464,0.0552718975,0.2689385414,0.0999510512,-0.1969587803,-0.1229813546,-0.0885096416,0.0261118487,0.0004234183,0.1158393919,-0.063973099,0.6269551516,-0.0046182969,0.1818998456,0.1399400085,-0.0571434237,-0.6079422832,-0.1859938055,-0.1058256403,0.1016420498,0.1406126469,-0.5978609324,0.2527682781,-0.1828462929,0.2016829997,0.0921825022,0.123047322,-0.2186382264,0.0764122605,0.0789034888,-0.1849667728,0.2314775735,0.0220164787,0.0954612121,0.0578553863,0.3473750949,0.207287699,0.0355929397,-0.0247276686,0.0855079591,0.2810788751,-0.0397179127,0.0362828895,0.054094471,0.2883422971,0.0110249845,0.3032592833,-0.2676548362,0.0082324967,-0.2597387433,-0.1954426914,-0.2852175832,0.3962238431,-0.0500473157,0.1336082071,0.0143755917,0.2324933708,-0.1748609692,0.2158662975,-0.2421335876,-0.0108233746,0.0139223309,0.0773160383,-0.1144058853,-0.0176408626,0.072422564,0.1025957614,0.0747925565,0.1593625247,-0.3149865568,-0.2109951824,-0.0314414427,0.1672118008,0.0113942297,-0.1947380453,0.0074231494,-0.3738916218,-0.3464235961,-0.1538661271,0.0740350932,-0.1240451485,-0.2739440799,0.0633604005,-0.0516135544,0.2877509296,-0.1349065006,0.0624866895,-0.1092984378,0.3250006139,-0.1217821762,0.256128639,0.1026182771,-0.0545148775,-0.32748124,-0.130163461,0.1515771151,-0.108228527,0.0077440068,-0.2896370888,-0.0746207908,-0.2115842402,0.0690291077,0.3015310466,-0.2442645282,0.117809169,0.2520481646,0.1225372553,-0.1094106585,-0.1571881175,0.2543361187,0.0205630865,0.0305483043,0.1831340641,0.0294940323,0.1970743835,-0.0836086199,-0.0662004724,0.0369007215,-0.5280764699,-0.234131828,0.3301568627,-0.1551273167,-0.0636102483,-0.1533145308,0.1343732923,0.151738435,0.2252937853,0.2771205902,0.2047518194,-0.0797060654,-0.0946735889,-0.0928359255,-0.3584408462,-0.4110491574,0.4001929462,-0.0519228131,0.5727984905,-0.2530632019,0.0145148076,0.1577865928,-0.2577901483,-0.1493125558,0.3014575243,-0.1298013777,-0.0493225902,-0.2117246687,0.0662313551,-0.1325351894,0.2566750348,-0.0862678885,-0.2290097624,-0.1688479632,0.2810490429,0.1293066442,0.2221566588,0.1151630655,0.1490802169,0.2406370789,-0.1195043027,0.184768036,-0.2638046145,-0.0701185539,-0.3630492091,0.0245367233,0.0707910657,0.2926615775,-0.3433614969,0.1792566627,0.0490802042,-0.0367895551,-0.0436841547,0.1956543028,0.2057573795,0.0616461709,0.2895565331,-0.0077095418,-0.0513410196,0.1617046893,0.1544669122,0.1573547125,-0.5646912456,0.523044765,0.2359953076,-0.0357844718,-0.0479995944,0.1822000891,-0.1502666622,-0.1165200397,0.5037794709,-0.2524596155,0.066591315,0.0163789354,0.049683135,-0.2483112663,0.6843262315,0.0231799204,0.4848305583,-0.1064878628,0.0550828278,-0.4837979376,0.0678305253,-0.1406673938,0.153826192,0.2474338561,0.3122571707,0.1356092244,0.1992427707,-0.4410076141,-0.2103655785,-0.0597522147,0.0664114878,-0.1679532975,-0.1248881072,-0.4908430874,0.1584864557,-0.0768742785,-0.4446339309,0.0908274427,-0.4045913219,-0.0454269797,-0.066478081,0.2613281012,0.3015222549,0.4070885181,0.3174887896,-0.076704599,-0.0605536848,0.0371177644,-0.0975896418,-0.1404159665,0.0245529544,-0.3423188627,-0.0853138715,-0.1146729887,0.1339507103,-0.2993355989,-0.0613244213,-0.222176075,-0.2459430397,0.0424892455,-0.1332900971,-0.1426001489,0.1336318105,-0.2435173392,0.2245735526,0.1533162296,0.4772619903,-0.1780256629,-0.0838552862,-0.2299449295,0.0606660359,0.427815944,-0.6093042493,-0.2080089301,0.0816480294,-0.1429839283,0.2385161221,0.0769891813,-0.2797821164,-0.2049111575,0.2964864671,-0.1130065173,-0.1912153363,0.2732228339,0.3441916108,-0.0951953381,-0.1007728875,0.3284245133,-0.0725228637,-0.4218618572,-0.2013471127,-0.4865978956]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/193","title":"[Tensorflow] Use something else than `from_tensor_slices()`","comments":"Is `tf.data.Dataset.from_generator` working on TPU ?","body":"In the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?","comment_length":6,"text":"[Tensorflow] Use something else than `from_tensor_slices()`\nIn the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?\nIs `tf.data.Dataset.from_generator` working on TPU ?","embeddings":[-0.0868790671,-0.0382875428,0.0482940599,0.1714205444,0.2084365487,0.0964492857,0.296674341,0.3527236581,-0.1370872706,-0.0254568588,-0.116677396,0.4980401099,-0.1231487766,0.1001269072,0.3218100965,-0.210337162,-0.0379137881,0.250507921,-0.0893423557,-0.0461085252,-0.0414157063,-0.0643877536,-0.1123574451,-0.0385818668,0.0986470878,0.0257850029,0.0341697223,-0.1887057275,0.3934222162,0.0618938133,0.4107213914,0.0539116599,-0.0244123153,0.1719019562,-0.0001141647,0.4121471941,0.0030622999,-0.1883231997,0.0722049996,-0.0762645975,0.1668334603,-0.1545860022,0.1597396582,-0.221860528,-0.0530676767,0.0033171161,0.2462582886,0.0162675139,0.2594282031,0.1238675565,0.1010102332,0.5529751182,-0.1841063648,0.2754843831,0.13413921,-0.1264522523,-0.2384615093,-0.2517392933,-0.179214552,0.0110656293,0.076561667,0.0984649509,-0.2368704826,0.1184524298,0.4186603129,0.143612951,-0.0918490887,-0.3854108751,-0.1741825342,0.442309916,0.2887401581,-0.1036753953,-0.1881713271,-0.4221935272,-0.2038784772,-0.2771670222,-0.1426480412,0.0874801874,-0.2484312057,0.2208236605,-0.1377811283,-0.1060861275,-0.2178054005,0.1366947442,0.1752296537,-0.0345943384,0.1402255148,0.1056216583,0.175206691,0.1257175505,0.435202688,0.0324258916,0.3094073832,0.4140760303,-0.2841370106,-0.0077350759,0.1584937423,-0.5640808344,-0.0311937947,-0.217994079,0.3778815567,0.1842603236,-0.3150196075,0.2317814827,0.0491166189,-0.1425053477,-0.459792763,0.1587095559,-0.1488756537,-0.5498354435,0.0145998532,0.0235746931,-0.2570431232,0.0284709688,0.1105088219,-0.5949354768,-0.1643221825,0.2301956713,-0.2662009597,-0.4269599617,-0.225786835,-0.0412965566,-0.0519620068,0.1824286282,-0.0765518993,0.5810716152,0.1270314604,-0.148475334,-0.2758766711,-0.0526565239,-0.1867753714,0.0445794128,-0.0430880077,-0.0917053223,0.1392078996,-0.1681171209,0.145279482,-0.13762106,-0.1453580856,0.1465240419,0.2970141768,-0.0937450454,0.1162587553,0.2741129696,0.0512158014,-0.0577674955,0.031837415,0.0469731688,-0.2077866346,0.3693138063,-0.565476954,-0.2719416916,-0.0648437142,0.0945728049,0.0651214644,-0.2233650535,-0.3701817393,0.2973718345,0.1497465819,-0.1060447693,-0.1411957294,-0.1870077401,-0.4023851454,-0.294147253,0.3296283185,0.0415339842,-0.4960269332,-0.1258747727,0.0383129306,-0.0825375915,0.2689813673,0.4077807069,-0.2107131481,0.5374667048,-0.055959478,-0.0820979699,0.7938559055,-0.0799249634,-0.151213482,0.2690729201,-0.0793267861,0.1863521636,-0.004490552,0.3989875019,-0.0301650278,-0.0122997109,0.3383696675,0.6511124372,-0.1462607533,0.0387718752,-0.0673296228,-0.3458172381,0.2746907771,0.2619879544,0.1380167156,-0.0273355022,-0.2206364274,0.4034931958,0.1547115296,-0.0094104875,-0.1973343939,0.0846883953,-0.134975642,0.0426118784,-0.1300901473,0.0851268917,-0.5689669847,0.2139237374,0.2312868088,0.0820477009,0.2617073655,-0.0800104514,0.3686425686,0.1287549436,0.0993668735,0.1503095776,-0.0418795384,0.084476009,-0.0524989441,-0.0548431911,-0.4006290734,0.1775621176,-0.3553922176,-0.0977806672,-0.3451013267,0.3101783097,0.4172928631,-0.0048265979,-0.14458929,0.2693160176,-0.3808309138,-0.1057209298,0.105634287,0.216221422,0.019633051,0.3070557714,-0.6328960061,0.4891891181,0.0726112127,0.0011744341,-0.1228025183,0.1319280714,0.0274505951,-0.2082509398,-0.0083901705,0.2885394692,-0.2225587815,0.183632791,0.279163748,-0.2901498377,0.1492657214,0.087007381,0.1476023793,-0.1053594053,0.4300692379,0.1002079993,0.1608707458,0.3264430463,-0.546045959,0.1878330708,0.6462352276,-0.0723396018,-0.1331173182,0.2045637667,-0.0186855458,-0.2055647373,-0.0489198379,-0.063954331,0.2049931735,0.0899941102,0.0477440618,-0.0221383329,-0.0970062017,-0.256175667,0.2325128019,0.1052808687,0.3462228179,-0.1553347111,0.0279902574,0.2439808249,-0.1996866614,0.1128136665,0.2321854681,0.1703078896,-0.2488480508,0.0376762785,-0.2732077837,-0.3158566356,-0.0273458771,0.1046907008,0.1692209393,-0.1554248929,-0.1092404425,0.33228755,-0.0109994197,-0.1009610891,-0.1861438602,0.2959452569,0.1819048971,-0.4953805506,-0.0559136234,-0.0935554355,-0.1649338752,0.050442867,0.3622851968,0.1673533022,0.2112733275,-0.0186036099,-0.1050545126,0.1721324772,-0.0629894584,0.264986068,-0.0343403555,0.1291906238,-0.0493041314,0.1113944501,-0.1427369416,-0.1938512325,-0.1205644011,-0.435100615,0.0361392349,0.1680579334,0.0218876395,0.1713550091,-0.2209700346,-0.0979999676,-0.2794063091,-0.3239627182,-0.0061211903,0.2173490673,0.310030818,0.4120718539,-0.0018247541,0.2503640652,0.3802010715,0.0432004519,0.0822039619,-0.2004224211,0.4757316709,-0.2513023317,-0.2665078938,-0.1305096149,-0.1365016103,0.1785944998,0.369667083,-0.6938173771,0.1530750096,-0.0659872666,0.1021413878,-0.0625484064,0.0428579375,0.1617040634,-0.2464196384,0.0704630837,-0.0047482569,0.1537251323,0.1293190122,0.0123376455,-0.2074681967,0.413053602,0.4354855716,0.2712579966,0.7577574253,-0.1456844658,-0.2220643759,0.2551366985,-0.2340375185,0.2485228777,-0.199411422,0.0123600028,0.0588156655,0.0302227568,-0.0651638359,0.2442591935,-0.0913090408,-0.1668714285,-0.1491355747,0.2412701398,0.1087286994,-0.201990217,0.2068947405,-0.1890225559,0.1866950989,0.1092835143,0.2425516099,-0.3250059485,-0.4405439794,-0.0421527959,-0.1739014685,0.4348756075,-0.0344770327,-0.2303287685,0.11757195,-0.7444400191,0.0648477077,0.0167127009,0.2932910621,0.1287203729,-0.275075525,0.0161542241,0.0450951792,0.5570883751,0.0660036355,-0.096550405,0.2058465034,-0.2044467032,-0.4368213713,0.0291187912,-0.0645176023,-0.0775731057,0.1944186091,0.1384037584,-0.2407487482,-0.4863093793,-0.0586439446,-0.0343650766,-0.3453305662,-0.0687207356,-0.2134678662,-0.2833826542,0.030599637,-0.3453009725,-0.0022563501,0.2434642762,-0.1636660844,0.0670986548,-0.212389946,0.032506302,0.2624002993,0.0582192987,0.1128211692,0.4111317396,0.0384248905,-0.0794377029,-0.0576195344,-0.0982335955,0.1927478611,-0.0831467509,-0.1597531438,0.0517374128,0.1242432669,0.3738074005,-0.0026491582,0.0222294442,0.1399831325,-0.1522200853,0.0408977047,-0.2010893077,0.3419886827,0.0843338147,-0.351673305,-0.1460798681,-0.5555419326,0.5839936733,0.3055446744,0.1506786644,0.4637757838,-0.160209462,-0.3623106778,0.5067741871,0.3795339465,0.5239444971,-0.4398086965,0.5383066535,0.708378315,0.4780375361,0.8994936347,-0.407032609,0.2180324495,-0.1038847566,-0.3962178528,-0.0828646347,-0.2638950944,-0.3495996296,0.1286068112,0.1255734861,0.2415893972,0.1357893199,0.3337372243,0.2600526214,0.3337422311,0.3462898433,-0.3899857104,-0.2162538767,0.030539779,0.0770510137,-0.1888493598,-0.1234658957,-0.1931129694,0.2406459153,0.104681775,-0.0133505659,0.087328814,-0.2249604911,0.177028954,0.1519630551,-0.1446584612,0.0122466562,0.2360368669,0.1481110156,-0.1797491908,-0.1123814955,-0.0580198132,0.119147107,0.0241466053,0.1277098954,-0.0169567056,0.643368125,-0.0295675341,0.2117405832,0.1669529676,-0.0739856362,-0.588290751,-0.1401112974,-0.1069555879,0.0501961373,0.1929810941,-0.5117247105,0.2795544565,-0.2338129133,0.1936499476,0.0514804274,0.1372128129,-0.2026855648,0.0342969075,0.0360072516,-0.2128894031,0.1673933417,0.0239654016,0.1260778606,0.0159090944,0.3505676091,0.2047586739,0.0227788649,-0.0158319175,0.1358344704,0.3137167096,-0.0418020114,0.0582912713,0.0457840599,0.2251188457,0.0931271762,0.3558231294,-0.2323796004,0.0257653501,-0.3204656243,-0.1817831695,-0.2848242819,0.3500858247,-0.0227469672,0.0949235186,0.0009726239,0.2182132602,-0.0792957693,0.1770759374,-0.1861278564,-0.0463270843,0.0207368229,0.0588482283,-0.1170254052,0.0291039329,0.0328401998,0.115275085,0.018047018,0.1150165945,-0.3142249882,-0.1630314738,-0.0253931209,0.1936796308,0.0484493375,-0.2448044568,-0.0065182196,-0.3239510953,-0.3482934237,-0.2035849243,0.1374405324,-0.0891219005,-0.2771610916,-0.0017751701,-0.053516984,0.3299628198,-0.2434860617,0.055296801,-0.0159667674,0.3395178318,-0.0865429863,0.2617418468,0.1171222553,0.0056551374,-0.2938148081,-0.0991334841,0.18082349,-0.1866872311,0.1126643941,-0.3687258959,-0.0659749955,-0.2603381872,0.0486157052,0.3228302598,-0.2224782705,0.1184835956,0.2391194105,0.0597210638,-0.097965382,-0.1194930598,0.2441406846,0.0648180097,-0.0570436269,0.2482053638,0.0793604553,0.1906872392,-0.1229285374,-0.0590665825,0.0282491166,-0.5793239474,-0.172464326,0.484616518,-0.0900782794,0.0701521859,-0.2057543248,0.114920333,0.1055602953,0.1910131276,0.314153254,0.2383429706,-0.0578580536,-0.1026456952,-0.0692789257,-0.3893674016,-0.4487060606,0.4878886342,-0.0703287721,0.6088397503,-0.2418860793,-0.0022300973,0.163305372,-0.1996318549,-0.2007375658,0.2659332752,-0.1085631102,-0.0599249154,-0.2106957734,0.0603568777,-0.1803510785,0.3071591854,-0.1267035007,-0.2360008806,-0.2016383111,0.2791036665,0.1529101133,0.1867636889,0.1131368354,0.0568871684,0.2205639333,-0.1133483052,0.2331406176,-0.3376013935,-0.1013604924,-0.3239659071,-0.0028212932,0.0661067814,0.2674559653,-0.3128728569,0.1329340041,0.0869012028,-0.0437580496,0.004614661,0.2531923056,0.2291501909,0.0899050981,0.2440035343,-0.0447396599,-0.0136211766,0.1288625747,0.1823112071,0.1594631076,-0.6616834998,0.5682293773,0.1533733904,-0.0583017096,0.0372616425,0.1992427856,-0.0404181555,-0.077959165,0.504414618,-0.253192395,0.0502887778,0.0633289963,0.0200848915,-0.1864648312,0.657089591,0.0896958858,0.5332286954,-0.0522699431,0.03054942,-0.4698709548,0.0855261609,-0.1305527687,0.170535773,0.2553669512,0.260029614,0.1802605838,0.2118143588,-0.4193258584,-0.1828724891,0.0022854393,0.0821651816,-0.0988298133,-0.1477116644,-0.564810276,0.162801668,-0.1794117242,-0.4805109501,0.059715718,-0.4114361703,-0.0984551609,-0.0588008873,0.2186493576,0.363727659,0.4545892477,0.328710407,-0.0648781881,-0.0715276152,0.0608705804,-0.0450522006,-0.1312408298,-0.0235806238,-0.3535795808,-0.0586174503,-0.2232599109,0.0162127484,-0.2799730301,-0.0134688783,-0.216001451,-0.2641216516,-0.0051513431,-0.0930071473,-0.1143442243,0.0431109183,-0.2843485773,0.1310715526,0.2777351737,0.4117286503,-0.1524763852,-0.1560319662,-0.2199468911,0.1719738543,0.3004270196,-0.6727653742,-0.2039794922,0.0173229501,-0.2004119605,0.2078109384,0.0743437037,-0.308563143,-0.1518829018,0.2945124805,-0.1235081404,-0.2163621038,0.2080201507,0.3343550861,-0.0742866918,-0.1116485372,0.3370922506,-0.0725785941,-0.459418565,-0.2472846061,-0.4375533164]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/193","title":"[Tensorflow] Use something else than `from_tensor_slices()`","comments":"`from_generator` is not working on TPU, I met the following error :\r\n\r\n```\r\nFile \"\/usr\/local\/lib\/python3.6\/contextlib.py\", line 88, in __exit__\r\n next(self.gen)\r\n File \"\/home\/usr\/.venv\/bart\/lib\/python3.6\/site-packages\/tensorflow_core\/python\/eager\/context.py\", line 1900, in execution_mode\r\n executor_new.wait()\r\n File \"\/home\/usr\/.venv\/bart\/lib\/python3.6\/site-packages\/tensorflow_core\/python\/eager\/executor.py\", line 67, in wait\r\n pywrap_tensorflow.TFE_ExecutorWaitForAllPendingNodes(self._handle)\r\ntensorflow.python.framework.errors_impl.NotFoundError: No registered 'PyFunc' OpKernel for 'CPU' devices compatible with node {{node PyFunc}}\r\n . Registered: <no registered kernels>\r\n\r\n [[PyFunc]]\r\n```\r\n\r\n---\r\n\r\n@lhoestq It seems you merged some changes that allow lazy-loading. **Can you give an example of how to use ?** Maybe the Colab notebook should be updated with this method as well.","body":"In the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?","comment_length":87,"text":"[Tensorflow] Use something else than `from_tensor_slices()`\nIn the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?\n`from_generator` is not working on TPU, I met the following error :\r\n\r\n```\r\nFile \"\/usr\/local\/lib\/python3.6\/contextlib.py\", line 88, in __exit__\r\n next(self.gen)\r\n File \"\/home\/usr\/.venv\/bart\/lib\/python3.6\/site-packages\/tensorflow_core\/python\/eager\/context.py\", line 1900, in execution_mode\r\n executor_new.wait()\r\n File \"\/home\/usr\/.venv\/bart\/lib\/python3.6\/site-packages\/tensorflow_core\/python\/eager\/executor.py\", line 67, in wait\r\n pywrap_tensorflow.TFE_ExecutorWaitForAllPendingNodes(self._handle)\r\ntensorflow.python.framework.errors_impl.NotFoundError: No registered 'PyFunc' OpKernel for 'CPU' devices compatible with node {{node PyFunc}}\r\n . Registered: <no registered kernels>\r\n\r\n [[PyFunc]]\r\n```\r\n\r\n---\r\n\r\n@lhoestq It seems you merged some changes that allow lazy-loading. **Can you give an example of how to use ?** Maybe the Colab notebook should be updated with this method as well.","embeddings":[-0.1073188931,-0.0873600394,0.0580076836,0.1967505664,0.2114300877,0.0975350142,0.2802228928,0.3764089644,-0.1501921862,0.0037057144,-0.0684898049,0.5348148346,-0.1144309938,0.0036037848,0.3254978359,-0.1784625053,-0.034475565,0.2562462687,-0.0664242283,-0.0699758157,-0.0435513556,-0.0245687813,-0.0964964926,-0.0267755184,0.0568178408,0.0180752091,0.0855254531,-0.1464977264,0.4081127346,0.0464173742,0.3898690343,0.0233454518,-0.0336213112,0.1560434401,-0.0001147429,0.4061399698,0.0354181714,-0.1917031854,0.0718830749,-0.1159350872,0.1680437624,-0.1635877639,0.165232569,-0.2180948704,-0.0562067553,0.0229955856,0.2129852325,-0.0601646751,0.2803551257,0.1379718184,0.0911628157,0.5434450507,-0.1706320345,0.2633911371,0.1432108581,-0.1526538581,-0.2350259572,-0.2543812394,-0.1945093572,0.0209915917,0.114635393,0.0894467756,-0.2371267676,0.1298708171,0.461627394,0.1611918807,-0.0694114193,-0.3664709926,-0.2092974782,0.4291151166,0.261646688,-0.0654994994,-0.1750208437,-0.4465816319,-0.2336889505,-0.3110021949,-0.1049036086,0.1266529858,-0.2806473672,0.2254827321,-0.0982220843,-0.0873731896,-0.2566456199,0.176518172,0.1964699477,-0.0448146909,0.1274375319,0.1107757837,0.2077470422,0.1545114368,0.4083363116,0.0213579722,0.3032938838,0.4022803903,-0.2894174159,0.0291159265,0.1779970229,-0.5676332712,-0.0688966215,-0.2302413583,0.3917006254,0.167443797,-0.3058205843,0.2517029345,-0.0107252654,-0.0943367183,-0.4143875241,0.1737424135,-0.1531392932,-0.5154541135,0.0534288622,0.0449390523,-0.2540406883,-0.0101100812,0.0223361384,-0.5388683677,-0.1700576693,0.224955067,-0.3130659759,-0.4252868593,-0.252390027,-0.0495489761,0.0026704632,0.2106461972,-0.124899298,0.5658953786,0.1485735923,-0.1373212188,-0.3045383096,-0.0486366339,-0.199589178,0.0869318992,-0.0675951391,-0.0861238316,0.1349673718,-0.1756485254,0.1778348088,-0.154832691,-0.1665776521,0.1368266493,0.2998939753,-0.1021675169,0.0943407863,0.2443123162,0.0462976992,-0.0560603179,0.0474385247,0.0725710988,-0.210729003,0.3622113168,-0.523563087,-0.3000704348,-0.0414670072,0.0968171433,0.0267720185,-0.1920387149,-0.323153019,0.2882851362,0.1580442935,-0.1397832632,-0.1726091057,-0.1775747091,-0.4339949489,-0.2991339266,0.3465132713,0.0750124305,-0.5027424693,-0.122789152,-0.0129162846,-0.0579261295,0.2622863948,0.4277922511,-0.2220567316,0.5380476713,-0.0876769274,-0.0274706054,0.8010421991,-0.0747700632,-0.1710435599,0.2472113371,-0.112245895,0.1672291756,0.0287137441,0.3875941038,-0.0019180553,-0.0308192465,0.3812481165,0.6379414201,-0.1516855806,0.0254837181,-0.1061467379,-0.3495667577,0.2707493603,0.2419636101,0.1044935584,-0.0300451554,-0.2365112305,0.3986965716,0.1312617213,-0.0243411288,-0.1928764731,0.146145761,-0.1458996832,0.0251875259,-0.1839003712,0.1019969657,-0.6190458536,0.2335283309,0.1444827318,0.0838889778,0.2368183136,-0.044368919,0.3555767238,0.1226112768,0.0514346585,0.1591100097,-0.0466658473,0.1264287531,-0.0583882667,-0.0436277501,-0.3507683873,0.1307992935,-0.3505170941,-0.0940877646,-0.3933821321,0.2961205542,0.3782663345,-0.0369105823,-0.1161267683,0.2575261593,-0.3368946016,-0.1007997617,0.0596050285,0.2039327323,0.0084921652,0.3252086341,-0.7214280367,0.4762921035,0.057509426,-0.0179248527,-0.0847948566,0.1305512488,0.0099450285,-0.2209627181,-0.0181686245,0.2750521302,-0.2316989899,0.198826164,0.288766861,-0.2876959443,0.1481619328,0.0720495954,0.1864931583,-0.111014463,0.4762393832,0.1712875068,0.178589493,0.3287807703,-0.5321706533,0.2031076849,0.6362365484,-0.0151546085,-0.1437105536,0.1793667525,0.0180220138,-0.2019819617,-0.0362487137,-0.0788511112,0.2486162186,0.1055759043,0.0446877517,-0.0182140991,-0.0976979807,-0.284923166,0.2182042152,0.1002806947,0.4032101631,-0.0997613072,0.0260698888,0.2311626375,-0.1733278781,0.0878035575,0.2498807311,0.1755167991,-0.2438870966,0.0178609453,-0.2885760963,-0.279242456,-0.0557992123,0.0874663666,0.1396376491,-0.1693182141,-0.0913473815,0.3872905672,-0.0092273736,-0.0840872526,-0.1445636302,0.2974518836,0.1473646462,-0.4759062827,-0.106806539,-0.0650728643,-0.2130011767,0.0315233655,0.3749706447,0.1918806434,0.2223303616,0.0081341583,-0.1391958594,0.1811559051,-0.0839734524,0.2416811734,-0.0319390222,0.1378182918,-0.0469623469,0.1122353598,-0.0727227181,-0.2634594738,-0.0618372411,-0.4290964901,0.0396714695,0.1868334264,0.0174817163,0.1602581888,-0.2362188548,-0.1386365145,-0.289046973,-0.3960028887,-0.0182985999,0.1862496287,0.3226663768,0.4042642713,0.0039606448,0.2540592253,0.3930638433,0.0672450215,0.0852791071,-0.1554905027,0.4933819771,-0.2766977847,-0.2668296993,-0.1434481293,-0.1447981149,0.2446989566,0.3684467673,-0.6950929165,0.2113606632,-0.0967331231,0.1460209489,-0.0984132066,-0.0158658475,0.2280783206,-0.2083772123,0.0687461495,-0.0179371033,0.0990448818,0.1303305179,0.0258808881,-0.1436575055,0.4220212698,0.4722150266,0.3167876899,0.7827600241,-0.123018235,-0.2463125587,0.2838458717,-0.2280215323,0.2599783242,-0.1918618977,-0.0162549224,0.0662870631,0.0486704744,-0.0657788813,0.2300899923,-0.1085874513,-0.1398120821,-0.1555846632,0.2244049311,0.0955583826,-0.2274152786,0.225395754,-0.2080326229,0.2161852717,0.0942552984,0.2296181768,-0.3376866877,-0.4288316071,-0.0257984381,-0.1772107333,0.4478472769,-0.0550626628,-0.2819491625,0.1320954114,-0.6937566996,0.0947369412,0.0476857796,0.3188589215,0.077662982,-0.2423045337,0.0042074779,0.0652077422,0.6028192043,0.0443473831,-0.0851078779,0.2141727507,-0.1975292116,-0.5001955032,0.0225244258,-0.0315114483,-0.0888418704,0.2648511231,0.120264031,-0.2928029895,-0.5042570233,-0.0794369131,-0.0002031551,-0.3595190048,-0.0935727358,-0.2238886654,-0.2616745532,-0.0281927232,-0.3826907277,0.021887606,0.301037699,-0.1224641651,0.0201726276,-0.1732188165,0.0016034595,0.2339190394,0.0116848955,0.1428958625,0.4014416635,0.0406818874,-0.0614523515,-0.0342997499,-0.0987658575,0.2059077173,-0.0980634019,-0.1612316072,0.0365691893,0.1534313112,0.3691768944,0.0317898728,0.0232632961,0.1279191673,-0.1482195556,0.0383006893,-0.1444599181,0.3786459565,0.0908218697,-0.3453332186,-0.0725552812,-0.5526955128,0.5314416289,0.2988009155,0.1519264579,0.4885193408,-0.1953699887,-0.3557861447,0.491624862,0.396874547,0.5634972453,-0.3789258897,0.4882778823,0.7028642297,0.5219005346,0.9219176769,-0.3840250969,0.283741802,-0.1113030091,-0.3852226138,-0.0714363158,-0.2753491104,-0.3265313804,0.1312577873,0.133095175,0.2455310971,0.1267234832,0.284815222,0.2412852794,0.3513440192,0.3585632741,-0.4072935581,-0.2122301161,0.0137320077,0.0516492464,-0.1445596963,-0.114865534,-0.2079911828,0.2138840854,0.0581814833,-0.0140091283,0.1565972716,-0.2624652684,0.2152719945,0.1364260018,-0.1354517192,0.0678654611,0.2597345412,0.1231050044,-0.183456406,-0.0779030323,-0.0785767287,0.0610422045,-0.0116662662,0.1457231045,-0.0493846722,0.6279592514,-0.0561382808,0.194537282,0.1661252826,-0.0887127146,-0.5715491772,-0.1340987384,-0.1442833841,0.0763438046,0.1837495118,-0.5355891585,0.2812329829,-0.2824200094,0.2070798427,0.0510312729,0.1385145038,-0.2297199667,0.0398102105,0.016621789,-0.1924344152,0.2076534778,0.0350049436,0.0930196941,0.0312024895,0.3733727336,0.2235580534,0.0207172837,-0.017776873,0.0854412168,0.2849075496,-0.0793826133,0.0800855383,0.0267137792,0.2016975284,0.0777299926,0.373321563,-0.1820382476,0.065179579,-0.2921445072,-0.1697729975,-0.2766076028,0.3394334912,-0.0816019624,0.1277525872,-0.0494597033,0.2162125558,-0.1064228341,0.2410110831,-0.1848933548,-0.0539821573,0.0022460038,0.0434737802,-0.1490770578,-0.017296331,0.0480049066,0.1189086512,0.0274159871,0.1244164407,-0.3220611513,-0.1428018659,-0.027802432,0.2020145953,0.0519523695,-0.2506320775,-0.0238010734,-0.3716357052,-0.3312898278,-0.2035338432,0.1122819185,-0.1109283566,-0.2844443023,0.0224388074,-0.07873144,0.3604117036,-0.2519087195,0.0619550869,-0.086282894,0.3406103551,-0.0935959071,0.2555476129,0.1501893401,-0.0224365517,-0.3736366928,-0.1217738986,0.1738017499,-0.1542133093,0.1214050874,-0.328571409,-0.0720373541,-0.2575981021,0.0246385839,0.3140626848,-0.1965522319,0.1291936338,0.2268683165,0.0630627051,-0.1053430513,-0.1104498357,0.2144174278,0.0261524785,-0.0303495564,0.2293281108,0.0744983926,0.152593568,-0.1068632603,-0.030716721,0.0404675715,-0.5724332929,-0.1810044348,0.4083304107,-0.0796088204,0.0684160367,-0.1724674255,0.1047245935,0.1336467415,0.1383842826,0.3073664308,0.1716020852,-0.1019174159,-0.0875922516,-0.0475402288,-0.3773836493,-0.4539462328,0.4834331274,-0.0419627875,0.5714561343,-0.265255779,0.0016521319,0.1176674515,-0.1616041362,-0.1760980636,0.284598738,-0.1015034094,-0.018113466,-0.255777061,0.0582797304,-0.161958009,0.2735616863,-0.092643328,-0.2304716706,-0.2114285976,0.3268530369,0.1237489432,0.1698043644,0.0880312249,0.0887144431,0.2086015195,-0.141257748,0.2405456454,-0.2891352475,-0.0972325057,-0.3500281274,0.0153476745,0.0887003541,0.3145700395,-0.30286327,0.1288616061,0.1081296876,-0.0666766241,-0.0150089571,0.2600989938,0.2302950472,0.1159553826,0.2454722226,-0.0452115387,-0.025874475,0.1459776759,0.1800823659,0.1694280058,-0.5887799263,0.6027830243,0.1803486794,-0.0968204439,-0.0083421227,0.1899624616,-0.0592466854,-0.0397980623,0.489726156,-0.2684250474,0.0399196409,0.0568905175,0.0132762892,-0.1862024814,0.6324128509,0.0884497762,0.4917037189,-0.1087175235,0.026488971,-0.4697318375,0.1212520152,-0.1135223731,0.1386704296,0.2599015832,0.2758076191,0.1513178945,0.1794599295,-0.43608284,-0.1845361739,0.0054653734,0.0976129696,-0.1273300499,-0.153816089,-0.4934645891,0.1202743798,-0.1569866985,-0.498511672,0.0545866117,-0.4139023125,-0.1042355075,-0.0938287899,0.2531805933,0.3283684254,0.4639804661,0.281491071,-0.0335109606,-0.0495559014,0.0429597795,-0.0872585699,-0.1108316183,-0.028888721,-0.3553535342,-0.0496384017,-0.1767718792,0.0745736584,-0.2796947062,-0.025637405,-0.3028122783,-0.24400796,0.0179985985,-0.08510869,-0.1251230985,0.079366006,-0.2918237448,0.1800650805,0.3247947395,0.3962630332,-0.1290539056,-0.1300882697,-0.2926558256,0.1488636881,0.3681569695,-0.6422933936,-0.17142196,0.0493033864,-0.1518708766,0.2456263602,0.064132385,-0.2793038487,-0.2004631162,0.3068369925,-0.125970602,-0.1965666264,0.1995116025,0.3359586596,-0.051682543,-0.1160439178,0.2768474221,-0.0313044302,-0.4617127478,-0.2316129208,-0.4398329556]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/193","title":"[Tensorflow] Use something else than `from_tensor_slices()`","comments":"Could you send me the code you used to run create the dataset using `.from_generator` ? What version of tensorflow are you using ?","body":"In the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?","comment_length":24,"text":"[Tensorflow] Use something else than `from_tensor_slices()`\nIn the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?\nCould you send me the code you used to run create the dataset using `.from_generator` ? What version of tensorflow are you using ?","embeddings":[-0.1686090976,-0.0939882025,0.0118442392,0.1699765623,0.2723704576,0.0823565796,0.2256458849,0.3897307217,-0.0503767245,0.0051490618,-0.0811933503,0.6080824733,-0.1429212838,-0.0144129554,0.2508257926,-0.2165247649,-0.1014938354,0.2596186996,-0.1496244222,-0.0935945809,-0.0144119794,-0.0348175615,-0.1413571984,-0.0275248867,-0.0601284504,-0.0839252993,0.0791330338,-0.0956678763,0.3711793423,0.0390430689,0.2916139066,0.0025852246,-0.0244213697,0.1651110053,-0.0001032014,0.3215746284,0.0343726091,-0.1477330327,0.0488398597,-0.0791320577,0.1653178334,-0.2739711106,0.1919615418,-0.3028230667,-0.0543792732,0.0081698783,0.1938415617,-0.0288159121,0.3141269088,0.1588182747,0.2169723213,0.4294994175,-0.1177147701,0.3017023206,0.1172235608,-0.1770817935,-0.1900475323,-0.2290671021,-0.217553854,-0.0144306719,0.0329059288,0.2663014233,-0.2041040063,0.1305107474,0.4554982185,0.1430775374,-0.1110719219,-0.3703571856,-0.2649399936,0.3789654672,0.2244816422,-0.0839506388,-0.1515463144,-0.4547227919,-0.1709520817,-0.3064984381,-0.1505764425,0.118133828,-0.2600311637,0.1226125285,-0.101498723,-0.0880129784,-0.2320558578,0.1448360384,0.1663917601,-0.0429132916,0.0822631568,0.1081561074,0.1561127454,0.0490138382,0.2461539507,0.0588911548,0.3580822349,0.3697514832,-0.3404166102,0.0536975823,0.1779217869,-0.516630888,-0.0021290567,-0.2505258024,0.3231516778,0.1721430868,-0.2449933589,0.1835705638,-0.0725678429,0.012310721,-0.4178177416,0.1140060052,-0.1178707406,-0.4833922684,-0.0239285361,0.0217021257,-0.2394130379,-0.0485653132,0.1193628386,-0.4675555527,-0.0338964462,0.2352781147,-0.3758570254,-0.3709529936,-0.1966249645,0.0565313622,0.0138816126,0.2716265619,-0.1850642562,0.4295899868,0.1524079293,-0.1067645699,-0.3398886323,0.0041233669,-0.2882518768,0.1229175851,-0.1284395307,-0.1123177409,0.1438223869,-0.0785714909,0.2024256736,-0.0808299333,-0.1718904823,0.1340387315,0.2805534005,-0.1010799035,0.0838325694,0.2150489241,0.0513633229,-0.0131490817,0.0446443819,0.0967687666,-0.16510351,0.3000501394,-0.5144050121,-0.2434287816,-0.0340855084,0.2013803869,0.116215907,-0.149327457,-0.2542578578,0.3058813214,0.1306425184,-0.0581671111,-0.1796268672,-0.1691520065,-0.4182859361,-0.327901423,0.3185227513,0.0584320128,-0.4493255317,-0.0759404153,-0.0228949822,-0.000078702,0.2840332985,0.4118054509,-0.2025939673,0.4759196639,-0.0660431534,0.0623255298,0.7272694111,-0.0797071606,-0.2046185136,0.2472713292,-0.1875445694,0.1665145159,-0.0564950109,0.3881363273,0.0145793241,0.0066112438,0.3825980425,0.6268616319,-0.130277738,0.1262755394,-0.1689798385,-0.2685698867,0.3044632077,0.2909620702,0.0550540276,-0.0994176045,-0.2690316439,0.3530552685,0.1409158111,-0.0482356064,-0.1705192924,0.1743093431,-0.1014072672,0.0338936485,-0.1964698136,0.023537185,-0.6025986075,0.256714195,0.2188520581,0.1189933419,0.1482024044,-0.0970392376,0.2708938122,0.1100614667,0.032282345,0.1563063562,0.1124503687,0.1176761985,-0.0173104219,-0.0104484577,-0.2507628798,0.2139796764,-0.3762319088,-0.162748903,-0.3574134111,0.2085116506,0.3017752469,-0.0353901908,0.004790958,0.2599491179,-0.2557673454,-0.1147442386,0.0572551675,0.2753024995,-0.0281672217,0.2390883267,-0.6261082292,0.39822644,-0.0164093953,-0.0681696907,-0.0559390485,0.1453260183,0.0453526229,-0.1716874391,0.0163724478,0.2203736156,-0.1982612312,0.0816421136,0.2122797817,-0.2504985332,0.1703201383,0.0214308985,0.17780976,-0.0902487412,0.399679929,0.0517187938,0.1282099783,0.2929057777,-0.5027149916,0.2471539229,0.6455689073,0.0017061487,-0.0606105551,0.2191503942,-0.0046074232,-0.1864926517,-0.1146235764,-0.0755283237,0.2204622179,0.2333493531,0.0045612883,-0.0491562411,-0.1739037037,-0.2364361435,0.2298947722,0.1034812331,0.3950202763,-0.0964093208,0.0465534218,0.1559046358,-0.2110231072,0.0506055467,0.2233275622,0.2867364883,-0.2051321715,-0.0385508724,-0.2432170957,-0.2629223764,-0.0705011562,0.0183933731,0.1004700437,-0.1776507348,-0.0637674779,0.3268815875,-0.0052786567,-0.0022931152,-0.1070263535,0.2660950124,0.2539460063,-0.3675926924,-0.0481266864,-0.0965466201,-0.2597522438,0.1232628822,0.345780462,0.265076369,0.306099534,0.0424025282,-0.0293588545,0.1571471989,-0.0293092672,0.1693187058,-0.1896898746,0.1289821416,-0.0720650628,0.2025320381,-0.1092027426,-0.2199773192,-0.0401361063,-0.4229851067,0.0455172844,0.1925011128,0.0160646681,0.0531260408,-0.2264475822,-0.2200699598,-0.2942709327,-0.423545301,-0.0113177849,0.1150469258,0.2867074907,0.4668727219,-0.0324269757,0.2744036913,0.3595530391,0.1347762644,0.0210020859,-0.131843105,0.5142495036,-0.2914405167,-0.2923904955,-0.1259394288,-0.1577883065,0.2138791382,0.2645537257,-0.6998734474,0.1633132249,-0.1801667213,0.0906324387,-0.1719289273,-0.0230344385,0.2796549499,-0.1828378737,-0.0377399474,-0.0642140582,0.1282198876,0.1593093127,-0.0026167354,-0.1744403839,0.2911267579,0.4363160729,0.293207407,0.6823472977,-0.0495223328,-0.2472582161,0.245116502,-0.2034212351,0.1798978299,-0.1588044316,-0.1297353804,0.0691828653,0.0468609706,-0.0193073694,0.2627430558,-0.0791296512,-0.1922529787,-0.0854637921,0.1933406293,0.0232283641,-0.245051384,0.2765370011,-0.1978453696,0.1782018542,0.1024813578,0.1724616289,-0.3827110231,-0.4328967929,-0.0281906463,-0.2183413506,0.367569536,-0.0549692176,-0.3525262773,0.123317048,-0.6885595322,0.0714337304,0.0951660275,0.2950821817,0.0440070406,-0.2644685507,0.0274349786,0.0636136159,0.5770879388,0.0853455067,-0.0374396965,0.194219768,-0.0606220439,-0.5440306067,0.0276835449,-0.0838857666,-0.0265664849,0.2526631355,0.1372006834,-0.316308707,-0.5172993541,-0.0549013689,0.0118847154,-0.3238763213,-0.1030324474,-0.2559072673,-0.2436222434,-0.0124457944,-0.2928737998,0.069262363,0.2637311518,-0.1118286848,-0.0076461607,-0.1052022055,-0.0017560215,0.2264771014,0.0210246015,0.2179865241,0.4974187315,0.0919593424,-0.0067844032,0.0447604544,-0.1746150106,0.2100191414,-0.1032134444,-0.1212223247,-0.0029220122,0.1182395518,0.3274333179,0.0437070169,0.0457658023,0.0859020427,-0.2112959176,0.1172109917,-0.2157901525,0.4107753038,0.1617105156,-0.2892139256,-0.1019583195,-0.4516589046,0.4895881712,0.2225752473,0.1089935005,0.4912770987,-0.1722845435,-0.3628089428,0.4391488135,0.3384194672,0.5404883623,-0.3218807876,0.4336530864,0.752619803,0.4858801663,0.8558766246,-0.3515875041,0.2858245969,-0.1510767937,-0.3399315476,0.006260511,-0.2254667133,-0.4014182687,0.0677493066,0.0680918768,0.1856302917,0.0864809006,0.2248955369,0.1757579297,0.3943638504,0.3600965142,-0.4310414791,-0.2145257145,0.1498019844,-0.0365459882,-0.1551292688,-0.0597876348,-0.2087164223,0.1818536073,0.0639966652,0.0220146123,0.1505619138,-0.1861501634,0.2835200429,0.0552062392,-0.0665060058,-0.0130742453,0.2632392049,0.122078456,-0.1773328036,-0.098484017,-0.0576420948,0.0200409833,0.003613255,0.0865023956,-0.05334289,0.6400727034,-0.0491784476,0.1369739026,0.1090061739,-0.0721559376,-0.4897185266,-0.195905298,-0.1155196354,0.0858875215,0.0831958801,-0.5479533076,0.2364267856,-0.2011193335,0.1866301596,0.162153706,0.1226982176,-0.1949729174,0.0924104527,0.0998888388,-0.2133836448,0.2217306495,0.0351023152,0.1007732004,0.0912480652,0.3558176756,0.2210889012,-0.0259533357,-0.0795312077,0.0841825008,0.3034344614,-0.0642075464,0.0354571082,0.0473066904,0.2883157432,0.0701755807,0.3478029966,-0.2279077768,0.0145872962,-0.1992014199,-0.1821100414,-0.3068373203,0.3347510993,-0.0675087348,0.1276504695,-0.0259972047,0.2155905962,-0.074432537,0.2593864799,-0.3228650689,-0.0429342426,-0.0224945005,0.0756108537,-0.1000733301,-0.1244211197,0.0979777724,0.0984220803,0.1472499669,0.1296602637,-0.35898754,-0.2738022804,-0.0220634714,0.1193842813,0.0218150057,-0.2305416316,-0.0245690327,-0.3880213797,-0.3245274723,-0.1746474057,0.1003886312,-0.1464450955,-0.2916195393,0.0598720647,-0.1121730655,0.2773267627,-0.093602933,0.0084640151,-0.171801582,0.284094274,-0.062563315,0.191199556,0.1374153793,-0.0796872005,-0.388251096,-0.1348582059,0.1281253248,-0.1004951447,-0.0008419878,-0.3645167053,-0.0804586262,-0.2034343183,0.0906414092,0.2400402427,-0.2507552207,0.0349207446,0.2132697701,0.2012830824,-0.12772277,-0.1039654613,0.1998331547,0.0231969189,0.0618392229,0.1210763231,-0.0210874323,0.16453816,-0.1070282981,0.0151150748,0.0518668853,-0.5421464443,-0.2646963894,0.3222047985,-0.156806469,-0.0340582319,-0.0866416246,0.1356594414,0.1247228011,0.2581652701,0.3180680871,0.2337075472,-0.0799107701,-0.0754240304,-0.1047288775,-0.3642441332,-0.4094366133,0.4181026518,-0.1346697211,0.502153635,-0.2514807284,0.0272152964,0.1200003102,-0.2684730589,-0.1701834947,0.3052918911,-0.2132370323,-0.0114888409,-0.2474830151,0.0629416108,-0.1118252575,0.227255553,-0.0682515129,-0.1929841638,-0.1808990836,0.2902747393,0.0493138842,0.1948871315,0.0714777187,0.1825223565,0.2388371676,-0.1276654452,0.2389979511,-0.2271498889,-0.0909642056,-0.4048538208,0.0249794424,0.1032235771,0.335216552,-0.3399096727,0.1290879697,0.0999296382,-0.1025441885,-0.067951493,0.2378443629,0.2075454295,0.1221886128,0.2946442068,0.0724020749,-0.1298855543,0.1495373249,0.1964368671,0.1768039465,-0.4668309391,0.4900734723,0.1949758977,-0.0202424843,-0.1069500074,0.1463764161,-0.1831090301,-0.0890939534,0.5219665766,-0.2488794625,0.0497273616,-0.0882094875,0.0793081522,-0.2243309617,0.6939463615,0.0330313221,0.4624567032,-0.1568162292,0.010074581,-0.4859363139,0.0970560908,-0.1924959719,0.1519235075,0.247262761,0.3328306973,0.1067361236,0.1808577776,-0.4369946718,-0.1848951429,-0.0539362729,0.1127543151,-0.2026930302,-0.0650596768,-0.5096171498,0.114051275,-0.075770162,-0.470977664,0.1125386208,-0.4334460199,0.0563108847,-0.0910381898,0.278365612,0.2807071507,0.413672626,0.3336991966,-0.0414898805,-0.047225792,-0.0197918043,-0.1056934148,-0.1266563833,0.023457408,-0.3446066976,-0.0757180154,-0.0502072982,0.2180005461,-0.2361315787,-0.0374736264,-0.239501521,-0.205968827,0.0932820067,-0.1142156646,-0.0988611206,0.1141330674,-0.1716416776,0.1832166463,0.1843813956,0.4326111674,-0.092746906,-0.0565605313,-0.272002995,-0.0375748053,0.4600409865,-0.5970891118,-0.2347231209,0.0974619538,-0.0600755103,0.2257249802,0.0428931005,-0.2987774312,-0.1591787189,0.2910027504,-0.1312486231,-0.2402857095,0.2773437798,0.3229190111,-0.0826276541,-0.0808382034,0.2904894352,-0.0606851056,-0.3934634626,-0.1542896479,-0.4548883736]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/193","title":"[Tensorflow] Use something else than `from_tensor_slices()`","comments":"I'm using TF2.2\r\n\r\nHere is my code :\r\n```\r\nimport nlp\r\nfrom transformers import BartTokenizer\r\n\r\ntokenizer = BartTokenizer.from_pretrained('bart-large')\r\n\r\ndef encode(sample):\r\n article_inputs = tokenizer.encode_plus(sample[\"article\"], max_length=tokenizer.model_max_length, pad_to_max_length=True)\r\n summary_inputs = tokenizer.encode_plus(sample[\"highlights\"], max_length=tokenizer.model_max_length, pad_to_max_length=True)\r\n\r\n article_inputs.update({\"lm_labels\": summary_inputs['input_ids']})\r\n return article_inputs\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail', '3.0.0', split='test')\r\ncnn_dm = cnn_dm.map(encode)\r\n\r\ndef gen():\r\n for sample in cnn_dm:\r\n s = {}\r\n s['input_ids'] = sample['input_ids']\r\n s['attention_mask'] = sample['attention_mask']\r\n s['lm_labels'] = sample['lm_labels']\r\n yield s\r\n\r\ndataset = tf.data.Dataset.from_generator(gen, output_types={k: tf.int32 for k in ['input_ids', 'attention_mask', 'lm_labels']}, output_shapes={k: tf.TensorShape([tokenizer.model_max_length]) for k in ['input_ids', 'attention_mask', 'lm_labels']}\r\n```","body":"In the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?","comment_length":82,"text":"[Tensorflow] Use something else than `from_tensor_slices()`\nIn the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?\nI'm using TF2.2\r\n\r\nHere is my code :\r\n```\r\nimport nlp\r\nfrom transformers import BartTokenizer\r\n\r\ntokenizer = BartTokenizer.from_pretrained('bart-large')\r\n\r\ndef encode(sample):\r\n article_inputs = tokenizer.encode_plus(sample[\"article\"], max_length=tokenizer.model_max_length, pad_to_max_length=True)\r\n summary_inputs = tokenizer.encode_plus(sample[\"highlights\"], max_length=tokenizer.model_max_length, pad_to_max_length=True)\r\n\r\n article_inputs.update({\"lm_labels\": summary_inputs['input_ids']})\r\n return article_inputs\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail', '3.0.0', split='test')\r\ncnn_dm = cnn_dm.map(encode)\r\n\r\ndef gen():\r\n for sample in cnn_dm:\r\n s = {}\r\n s['input_ids'] = sample['input_ids']\r\n s['attention_mask'] = sample['attention_mask']\r\n s['lm_labels'] = sample['lm_labels']\r\n yield s\r\n\r\ndataset = tf.data.Dataset.from_generator(gen, output_types={k: tf.int32 for k in ['input_ids', 'attention_mask', 'lm_labels']}, output_shapes={k: tf.TensorShape([tokenizer.model_max_length]) for k in ['input_ids', 'attention_mask', 'lm_labels']}\r\n```","embeddings":[-0.1217657998,-0.0259314273,0.0470915586,0.189820081,0.1980044991,0.1544274986,0.2483752966,0.45679757,-0.0371510051,-0.056614697,-0.1416853964,0.6093758941,-0.1823763102,-0.0313647278,0.2451841533,-0.1602426022,-0.0515337177,0.2101798803,-0.0837807283,-0.0511187688,0.0414349586,-0.1030471548,-0.1133751571,-0.0896192342,-0.0841150582,0.0378104895,0.0268044118,-0.1593336612,0.4063458145,-0.0158472583,0.3480536044,0.0354809612,-0.0619997717,0.1019106805,-0.0001139468,0.3830084801,0.0316342488,-0.1697195768,0.0051154783,-0.0486841574,0.1872231215,-0.2209469974,0.247618407,-0.2797369361,0.0229602605,-0.043305736,0.2282474935,0.0214918461,0.2741083801,0.0563438572,0.1135803461,0.4625842571,-0.2261624783,0.3665566444,0.1534703374,-0.1067058221,-0.138231352,-0.2763141394,-0.2109407485,0.0413133502,0.0462952703,0.214860484,-0.2967178822,0.1001085714,0.4883716702,0.1346100718,-0.1158743277,-0.2901355326,-0.2041688114,0.3537537158,0.1901264191,-0.0716039985,-0.1768607646,-0.5691713095,-0.2341983765,-0.2198168486,-0.2141052783,0.0487432033,-0.1875362694,0.1386623681,-0.216715768,-0.0914960578,-0.1921276599,0.1696150899,0.1730032712,0.0884684846,0.1593900025,0.1527348012,0.1636829078,0.0402699262,0.2768391371,-0.0111793596,0.2941534221,0.4132718444,-0.3397202492,-0.0049699522,0.1499769241,-0.5529497266,-0.0025993306,-0.2191033214,0.3147853017,0.2147977352,-0.2501312494,0.1788448095,-0.0064106318,-0.0640124604,-0.3965274692,0.1439047456,-0.1776492149,-0.5946298838,0.0709003583,0.0254464708,-0.2908785343,0.0922625288,0.0541443415,-0.5405897498,-0.0903404206,0.2274273634,-0.3082315326,-0.4565548897,-0.2495396137,0.0322222933,0.0282584541,0.2659028471,-0.1724849194,0.4949054122,0.0468011573,-0.1260062009,-0.2357968688,0.0714636967,-0.2054802477,0.085484013,-0.0532731339,-0.0908422247,0.1894188523,-0.0446499586,0.2361996174,-0.1806654483,-0.1580970287,0.1609508097,0.2584713101,-0.0925418213,0.0716399103,0.2176814377,-0.0144632915,0.0617462657,0.0473540463,0.1133971438,-0.2338122278,0.3567083776,-0.551022768,-0.2275333405,-0.0048400629,0.0849091113,0.1035248414,-0.2339808941,-0.3640972674,0.3549089432,0.1024191156,-0.1246886849,-0.1437174082,-0.1485555172,-0.4110409915,-0.263394773,0.2886543274,0.1181835011,-0.3723021746,-0.1996258497,0.0139911743,0.0392890982,0.2183583677,0.444042474,-0.2287716419,0.4910322726,-0.0685439557,0.1765822321,0.7466109991,-0.1683420092,-0.2163927257,0.2665832639,-0.1092089191,0.2027740777,0.0070782541,0.4875001609,0.0556126051,-0.0282452423,0.3278444111,0.607059896,-0.1100429371,0.0909122229,-0.0626382828,-0.3540176451,0.3833684623,0.2236017287,0.0494338833,-0.0890882313,-0.2451320291,0.4838311374,0.1579272896,0.0060733361,-0.1822026521,0.1302783191,-0.1859731525,0.0057758423,-0.2210057527,0.0439639203,-0.6262143254,0.2945075333,0.2898077071,0.0198014192,0.2248506397,-0.1320512146,0.3046332598,0.0869716927,0.0584740601,0.1558163464,-0.0363753103,0.1532447785,-0.0082864892,-0.0561432801,-0.3398274481,0.2135779709,-0.3556129336,-0.0859222114,-0.2873385251,0.2686126828,0.446282357,0.0063380175,-0.1396049559,0.2558359802,-0.2394817173,-0.09090738,0.0943144783,0.2214565724,-0.0374283791,0.3102308214,-0.6426632404,0.5175514817,-0.0072998405,-0.0519855805,-0.0380340815,0.1512567103,0.0408390574,-0.2298714817,-0.045041319,0.2038133591,-0.1811840981,0.2521127164,0.2275706381,-0.2973989844,0.0864404067,0.0949699059,0.1474915743,-0.1060408801,0.3818445802,0.0820869729,0.2589592338,0.2943733633,-0.4201218486,0.2011387348,0.5873517394,-0.033839304,-0.0882355794,0.2429160923,-0.0898037925,-0.2195173353,-0.0496823303,-0.1579284072,0.2387478501,0.1556952447,0.0475634746,-0.0183602758,-0.0851338357,-0.1895781159,0.2330112159,0.138178423,0.3442912102,-0.1048949435,0.079455547,0.2439493537,-0.2006927133,0.1206365302,0.2041326016,0.1614872366,-0.1821660399,0.0208744239,-0.2740591168,-0.3233437836,-0.0688158199,-0.0156345777,0.1198373586,-0.1766926646,-0.0839560255,0.3564248383,0.0389873683,-0.0269095302,-0.1203829944,0.3014902174,0.1829601973,-0.485211879,-0.1409324706,-0.0722896978,-0.1876286119,0.0123512261,0.3414314985,0.2476367354,0.2139724642,0.1156194732,-0.1064011455,0.0933765098,-0.1371449828,0.22248061,-0.1411897838,0.0744171813,-0.0913883373,0.085876815,-0.1317083389,-0.2374128699,-0.1384253502,-0.4001226127,0.0152844125,0.1781018674,0.0439833067,0.1407825798,-0.1730547547,-0.0541110225,-0.3113821447,-0.368055433,0.0251365807,0.164383471,0.2630543113,0.4266567826,-0.0977155864,0.2875241041,0.2841942906,0.0566007383,0.0843756571,-0.0591102839,0.5510883927,-0.2119562179,-0.2075658888,-0.1972450912,-0.1682775766,0.1238679662,0.2879946828,-0.7350276709,0.2582063377,-0.1476224959,0.0522895828,-0.136073038,-0.0432344712,0.3010390103,-0.2462277859,0.0579819977,-0.0192133784,0.0429808348,0.0740937069,0.0150421076,-0.2124611437,0.2835455835,0.4253311157,0.3563779891,0.8198500872,-0.0815245658,-0.195712477,0.2159968168,-0.2431849986,0.2042169571,-0.2124856114,-0.0797761902,0.0732174963,0.0149209648,-0.0192074329,0.3055550456,-0.1208537892,-0.2351140231,-0.0657066405,0.1624546349,0.0711238459,-0.1895557791,0.253777653,-0.2382827103,0.1228615418,0.1436945051,0.2271651179,-0.3981941938,-0.5145754218,-0.0529263839,-0.1915570945,0.3439481854,-0.0387931205,-0.4047645628,0.1823065281,-0.7144542933,0.1373093426,0.0563871711,0.3346378505,0.1076642424,-0.2996100187,0.0232895035,0.0796905011,0.61192137,0.1648482382,-0.0709822029,0.1736160517,-0.1510481983,-0.5380704403,0.0128713315,-0.0515648723,0.0371389054,0.2749974728,0.1597069651,-0.3015443683,-0.5077008605,-0.0497351587,0.0067438036,-0.3528348207,-0.0636999682,-0.2292827517,-0.1595719308,-0.0365676843,-0.3108164668,0.0701916218,0.2330524772,-0.054863032,0.0353446528,-0.2360271811,-0.0731334686,0.2959994078,-0.0075319586,0.1719074547,0.5554138422,0.06869822,-0.0578230806,0.0280716587,-0.2228826284,0.2243764102,-0.0465518869,-0.1779031605,0.0205410719,0.1392533481,0.421295464,0.0915347114,0.0426735692,0.1779648662,-0.1819625646,0.0264815949,-0.2325233966,0.4516882598,0.1056406498,-0.2878837585,-0.1127602905,-0.4171382487,0.505194664,0.2666365802,0.0977834836,0.4054192007,-0.1845551133,-0.3795979619,0.4747128189,0.3681874573,0.5930817723,-0.3568761051,0.5228301287,0.6306706071,0.5135165453,0.9229069352,-0.369907558,0.2578589916,-0.1076136976,-0.2891258597,-0.0341490172,-0.2701135874,-0.3677411377,0.0976251513,0.1489039212,0.255405277,0.0596081018,0.3281603754,0.2287256271,0.421323061,0.4004003108,-0.4190790951,-0.1393467933,0.0488374792,-0.000468445,-0.1998969018,-0.0864965767,-0.1306748539,0.220348537,-0.0378471911,-0.0631247386,0.092715919,-0.1913190484,0.2733098567,0.0886060148,-0.0228782427,-0.0306523927,0.2359596938,0.0516060889,-0.1625456512,-0.0739810318,-0.0782154948,0.0206819586,0.0128320558,0.1324210763,-0.0732157379,0.5777006745,-0.003746324,0.2327096313,0.2019574791,-0.0701873302,-0.6475856304,-0.2420333773,-0.0774635375,0.164600417,0.1208362579,-0.5740083456,0.2775155604,-0.2164397687,0.20163773,0.040244136,0.0579118989,-0.1890513897,0.0572008304,0.0881507322,-0.2242264301,0.2380438596,0.0798751637,0.0802137852,0.0043422868,0.4003131986,0.094149217,0.0264455918,0.0004958752,0.1406666934,0.3189900815,-0.0531729423,-0.0355195887,-0.0320995264,0.2284281403,-0.0022153177,0.3170286417,-0.2210985422,0.0451947004,-0.3271238208,-0.2269155383,-0.2797814906,0.4097398818,-0.0322789885,0.1703535467,-0.0296617541,0.2365385294,-0.2120735943,0.2599015534,-0.194529891,-0.0003529537,0.05574416,0.0645967051,-0.2063010931,-0.0371828042,0.0859111622,0.1474546939,0.0297869444,0.1458420753,-0.2863135934,-0.1601678133,-0.055195421,0.2104236186,0.068900004,-0.1837074012,-0.0768120885,-0.3510490656,-0.305419445,-0.2109848857,0.0713349879,-0.1109613776,-0.2526502311,-0.017708702,-0.0166983604,0.303311944,-0.1873164624,0.0589580014,-0.0388102867,0.3205218613,-0.0678950995,0.2631837726,0.172121346,-0.0168109834,-0.3609753549,-0.1110453159,0.2617632449,-0.1292838305,-0.0193210393,-0.3542596996,-0.0280042365,-0.2465105504,0.0194333363,0.3817173243,-0.2598596215,0.0119022336,0.2552919686,0.0760612115,-0.0693088919,-0.1370902508,0.2898459136,0.0156411957,-0.0170291252,0.2157491148,0.0190553479,0.2301880866,-0.0763970241,-0.059450686,0.050927069,-0.4581305981,-0.2483108491,0.3442254663,-0.1479749382,-0.0445314348,-0.1566002965,0.1595490724,0.158819899,0.2009692043,0.2688859701,0.2720530331,-0.1309790313,-0.0804200768,-0.1090736315,-0.298851043,-0.4159928262,0.4332542717,-0.0309494399,0.5543479323,-0.301459372,-0.0442896895,0.2187474072,-0.2889840603,-0.1426355839,0.2879602909,-0.12157318,-0.0982647315,-0.1845833808,0.0626406744,-0.2162009776,0.2093061358,-0.1255273968,-0.2421465218,-0.1884736568,0.2828618586,0.1213045418,0.1418761313,0.033044856,0.1392564029,0.2624235451,-0.168439731,0.2853245735,-0.3038662076,-0.023454329,-0.3080236316,0.0524333194,0.030013673,0.3233006001,-0.3477111757,0.135496065,0.0732380599,-0.0523074307,-0.0218496993,0.2548743188,0.2141042799,0.1118784174,0.2443323284,-0.0474197008,-0.0380628519,0.1554596573,0.1356609613,0.1043485403,-0.6335433722,0.5223560929,0.258955121,-0.0348538272,0.0482245795,0.2361011207,-0.1764230579,-0.1175672859,0.492916584,-0.2662466764,0.0792281777,0.0896202847,0.0232440103,-0.2360375077,0.6891623139,0.015659254,0.511536479,-0.129479751,0.0808566585,-0.4833623171,0.046040155,-0.1313797683,0.1344795078,0.2439010292,0.2871463299,0.1548627615,0.2174822241,-0.3771543503,-0.2930710912,-0.0790357515,0.0058989278,-0.1204926446,-0.0916261896,-0.4695657194,0.2053018957,-0.0645814463,-0.4875364304,0.0535471812,-0.3142371476,-0.1048689708,-0.0682485253,0.2054504305,0.2708277702,0.3716531992,0.3245228827,-0.0592667684,0.0022978974,0.062266361,-0.1000932455,-0.1329056621,-0.0284085162,-0.3497491479,-0.0834475085,-0.1625831872,0.168179974,-0.3341850638,0.013359244,-0.2161339968,-0.3234885037,0.0504177995,-0.2010847479,-0.1404235363,0.0903067663,-0.3096468747,0.3197849691,0.137475878,0.4710556865,-0.1315186471,-0.0389615968,-0.1975430399,0.1254427731,0.4762639701,-0.6969991922,-0.18452923,0.0805965886,-0.1336675882,0.2864736915,0.0819308236,-0.3468104601,-0.2159187794,0.3673852086,-0.0630359948,-0.218921423,0.2391411662,0.3046946824,-0.108745873,-0.1180869862,0.3073832095,-0.0826756135,-0.3657678366,-0.1998540908,-0.5045338869]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/193","title":"[Tensorflow] Use something else than `from_tensor_slices()`","comments":"Apparently we'll have to wait for the next tensorflow release to use `.from_generator` and TPU. See https:\/\/github.com\/tensorflow\/tensorflow\/issues\/34346#issuecomment-598262489","body":"In the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?","comment_length":17,"text":"[Tensorflow] Use something else than `from_tensor_slices()`\nIn the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?\nApparently we'll have to wait for the next tensorflow release to use `.from_generator` and TPU. See https:\/\/github.com\/tensorflow\/tensorflow\/issues\/34346#issuecomment-598262489","embeddings":[-0.1305493712,-0.0308594704,0.0363049358,0.1227419302,0.2281208932,0.0917753205,0.2730562389,0.4065895379,-0.1111178175,0.0027906678,-0.088807933,0.5513763428,-0.1521194577,0.0472110957,0.3196076751,-0.2627645731,-0.0665044412,0.2488584071,-0.1298071444,-0.014312909,-0.0405507199,-0.0532309711,-0.0955616459,-0.0164401829,0.0150601091,-0.033534877,0.0511503406,-0.1520288438,0.4169293344,0.0210770834,0.3553362489,0.0021882898,-0.0340940319,0.1122701243,-0.0001092273,0.3845244944,0.0321453772,-0.1496059895,0.0620988235,-0.0740662292,0.2081267238,-0.2273654491,0.1988500506,-0.2661621571,-0.0468088314,-0.0052516502,0.2368553579,-0.034285564,0.2574249506,0.1193688512,0.1560954005,0.5252822638,-0.1733796746,0.2786276639,0.1325304955,-0.1595598012,-0.2168976963,-0.2793309987,-0.1323648393,-0.0109547814,0.0657459348,0.1479676813,-0.2282933295,0.0962995291,0.4325979948,0.1546453238,-0.1112498567,-0.3790851235,-0.2610158622,0.4085841179,0.2015358955,-0.0729810297,-0.1770699024,-0.4851972461,-0.1757751554,-0.2934440076,-0.1783652008,0.0868638456,-0.2625950873,0.1819095612,-0.1011813283,-0.1204154864,-0.2189192027,0.1201040894,0.2120116055,-0.0355753787,0.1604187936,0.1168624461,0.2244132608,0.1082769483,0.3494542241,0.0806608647,0.35487023,0.3675652742,-0.284050256,-0.0114905862,0.1692720056,-0.5806717873,-0.0444342121,-0.2525962889,0.3571074605,0.1853387356,-0.2919266522,0.2165090591,-0.0133072734,-0.0735663027,-0.4224945009,0.1319459379,-0.1399093866,-0.5050939918,0.0470280461,0.0463061109,-0.2208214402,-0.0199277624,0.0605984516,-0.51125139,-0.147703439,0.2690734267,-0.2849743664,-0.4418404996,-0.2301685661,0.0283928644,-0.0143321892,0.1737463325,-0.1279168576,0.5364353657,0.1373176873,-0.1248362809,-0.2813409865,-0.0295778885,-0.2323073447,0.1130894944,-0.0614650324,-0.0960282385,0.1556884646,-0.1606730819,0.1797935963,-0.1226887256,-0.127771765,0.1816211492,0.2595539987,-0.0819504037,0.1001746058,0.2579149306,0.0148862759,-0.0645697564,0.0382283106,0.0749141052,-0.2161310017,0.3539175391,-0.5225843787,-0.2890223861,-0.0487967283,0.1349671781,0.0877802894,-0.2136898041,-0.3331672847,0.3261962831,0.120963119,-0.1192807257,-0.1852190197,-0.2075044811,-0.4238689244,-0.3011671007,0.3041065931,0.0406429134,-0.4613336623,-0.1152338833,0.0065076631,-0.0165925473,0.2493128031,0.4456069171,-0.2329480797,0.4801923335,-0.053436324,0.0012751004,0.7598665357,-0.0854216591,-0.16154176,0.2256312221,-0.1233380213,0.2028128058,-0.0049250848,0.4053750932,0.0003449573,-0.0291188397,0.3868638277,0.6318570971,-0.1584129184,0.0767637864,-0.1092906669,-0.3459326923,0.3001140058,0.2797210515,0.0952894017,-0.0565825552,-0.2650605738,0.3954328597,0.130624339,-0.0343222432,-0.1820747703,0.067004025,-0.1286901981,0.0004341092,-0.1707858741,0.0657679886,-0.6022973061,0.2533168495,0.1912777126,0.0960817561,0.1893976778,-0.1119040027,0.3626312912,0.1090806499,0.118262887,0.1590494215,0.0190945361,0.0754810348,-0.0283455104,-0.0117259379,-0.3443227112,0.1589715779,-0.3754411638,-0.126958698,-0.3489677012,0.2644969821,0.3732426167,0.017966941,-0.084560357,0.3013327718,-0.3664574921,-0.1209989563,0.1176432893,0.2152287513,-0.0305702519,0.2932139337,-0.667224288,0.442422688,0.0515987575,-0.0227559302,-0.1092137024,0.1241608113,0.0025712894,-0.1974492967,0.0108399848,0.2805424631,-0.2258610129,0.1471134871,0.2610126734,-0.2442211807,0.1527157277,0.0638118982,0.1342630833,-0.1127064526,0.3706358671,0.0992518589,0.1377707124,0.2951819897,-0.5292932987,0.2312241346,0.670640409,-0.0221984033,-0.1136625782,0.2168765515,0.0373343788,-0.2118321955,-0.08162269,-0.0811692923,0.2019304633,0.145782724,0.0416364148,0.0004896394,-0.164139986,-0.2583947182,0.2476225644,0.1311677694,0.3808506727,-0.1395992339,0.0300449897,0.2035273761,-0.1915541738,0.0832396448,0.2106474191,0.1754733324,-0.2056151778,0.0302167274,-0.2745707035,-0.3066859841,-0.0898331255,0.035019286,0.1440560818,-0.1836826354,-0.0499239415,0.3294695318,-0.0494266897,-0.0536168925,-0.1188829541,0.3149219155,0.1649483591,-0.4449187815,-0.0658236668,-0.1055019423,-0.1850717217,0.0902298316,0.3607676625,0.1774414331,0.242978856,0.0263939425,-0.087532863,0.1681303084,-0.0779821351,0.2226692438,-0.111642234,0.1100937501,-0.0496085063,0.0718158409,-0.1287714988,-0.2588922381,-0.1155318022,-0.4312819242,0.0538533144,0.1550911814,-0.0024251451,0.1485794932,-0.2426673174,-0.1758734882,-0.3055381775,-0.3943730593,0.0180975385,0.1643054336,0.3020911217,0.4260999858,-0.0299287923,0.2822209895,0.3855684698,0.0524073578,0.0802656263,-0.1383150667,0.4784578681,-0.2761949003,-0.2551722527,-0.1560890377,-0.1742354333,0.1930478215,0.3186536431,-0.7124727964,0.1455635726,-0.0832740888,0.086586833,-0.0899093747,-0.0410923846,0.2317247093,-0.2033178508,0.0205108784,0.0046083168,0.1650743186,0.1451765895,0.0248066206,-0.2041558921,0.3703911006,0.4237145483,0.3312722147,0.7575733662,-0.084291786,-0.2142929286,0.234271273,-0.1822491884,0.197028935,-0.2328152657,-0.0642755106,0.042939052,-0.0136925736,-0.0053534764,0.2236841023,-0.1000081599,-0.1953769177,-0.1173397154,0.2207117677,0.1183625385,-0.2383675426,0.2582300305,-0.2144397348,0.1831323504,0.1111207306,0.2126103491,-0.329985112,-0.399720192,-0.0193105377,-0.2327251583,0.4178363085,-0.0403597206,-0.3286675811,0.1039649695,-0.7131770253,0.065324083,0.1196059734,0.3085822463,0.0899462551,-0.2990151644,-0.006996972,0.0677557886,0.548350215,0.0983364508,-0.0874561518,0.2097491622,-0.1773277968,-0.4892942309,0.0340788551,-0.0558376238,-0.0789329633,0.2494371384,0.1401745528,-0.2619264126,-0.491312623,-0.024556512,-0.0164186489,-0.3395267725,-0.0488195978,-0.1868013293,-0.2434477359,0.0010785362,-0.3156530559,0.0576133989,0.2791346014,-0.1710855216,0.0206963476,-0.1508034021,-0.0014766246,0.2122334838,0.0035509055,0.1631867886,0.4632868171,0.0473942794,-0.0489132889,-0.0080882134,-0.1308065802,0.2240553945,-0.0922200233,-0.152269125,0.0187585671,0.1585353017,0.3719956577,0.048455283,0.0360952765,0.1285239309,-0.1741507351,0.0331269763,-0.1986674666,0.367053628,0.0867629796,-0.3417378664,-0.0824657828,-0.5109403729,0.5615842938,0.2959671915,0.1288378388,0.4830286801,-0.197045505,-0.3616954684,0.5212906599,0.3826307952,0.5099015236,-0.3703688383,0.5261979699,0.7229401469,0.4783598483,0.9239939451,-0.3590678275,0.2710904181,-0.1238886192,-0.3586636484,-0.0287977532,-0.2556560338,-0.3750624359,0.0788878873,0.1031144336,0.2167330235,0.1310868561,0.2775279582,0.2179110795,0.3121580482,0.3466188908,-0.4222291708,-0.1941190213,0.0675997511,0.0296701938,-0.2340955436,-0.0818518102,-0.1964045018,0.2307493389,0.0648563653,0.0203567557,0.1200598478,-0.1848243475,0.2819259167,0.1395899057,-0.0990309045,0.0404783972,0.2417275012,0.1200778261,-0.2049072087,-0.1091784984,-0.0896357596,0.0653189197,0.0101710204,0.1468190849,-0.044765152,0.6814283133,-0.0219719969,0.1964143366,0.1639902294,-0.0786113143,-0.5933412313,-0.1823864877,-0.1144627333,0.0853359178,0.1685634404,-0.5447576046,0.2875629961,-0.1981662363,0.1810290217,0.097979337,0.1195998862,-0.2046099454,0.0549654216,0.0552382059,-0.1870274544,0.2211266458,0.0292155519,0.1272464842,0.0346336924,0.3529243171,0.2273524255,0.004736661,-0.0259114504,0.104435958,0.2849631011,-0.0461859554,0.0644358769,0.0366448835,0.2289125174,0.0641891733,0.3598418534,-0.2269446105,0.031078333,-0.2919078171,-0.1731797457,-0.3018861711,0.3662432134,-0.0597964562,0.1284084916,-0.0037674049,0.2189663202,-0.123719275,0.2288983464,-0.2416868657,-0.0227429327,0.0058932723,0.0676049292,-0.1260156631,-0.0586395524,0.0696458593,0.1308999509,0.0643854886,0.1422179341,-0.3747168481,-0.2018874139,-0.0400533266,0.165595755,0.0326660089,-0.2386942506,-0.0044226376,-0.3472806811,-0.3325613439,-0.1648240834,0.1165761054,-0.1250740886,-0.2715747356,0.0603117906,-0.0594609305,0.2978952825,-0.1393365264,0.0345381387,-0.107003361,0.3132716715,-0.0834672078,0.2498146147,0.0991906226,-0.0500934124,-0.3136666119,-0.1167935356,0.1376471519,-0.1159993559,0.0682292208,-0.3133593798,-0.0745455176,-0.2488780767,0.0428732708,0.27158916,-0.2371542901,0.0904113948,0.2457919419,0.1149822325,-0.0976052955,-0.1507215351,0.2068716288,0.0823462307,-0.02431711,0.2002537698,0.0880941674,0.1603425294,-0.0639608875,-0.0402886569,0.0101867868,-0.5695465207,-0.1938688308,0.408152163,-0.1363224834,-0.0142499534,-0.1612711102,0.0954625309,0.1648012996,0.2110168338,0.2794189453,0.2288536429,-0.0718000457,-0.0962202922,-0.0613992736,-0.3514273167,-0.4288904965,0.4405208826,-0.0612532198,0.5847460032,-0.2274622172,0.0355161652,0.1482982188,-0.2197300792,-0.1693271101,0.2957035601,-0.1487514377,-0.064812988,-0.2391251624,0.0648172945,-0.150669232,0.2991458178,-0.1023098901,-0.2143223584,-0.1676593721,0.2694037259,0.0965816304,0.2263259143,0.1180019528,0.0747155845,0.248540327,-0.1172579527,0.1978902966,-0.263451457,-0.1078858003,-0.369613111,0.0116132013,0.0912148133,0.2826962173,-0.333753854,0.126609996,0.0701259226,-0.0458486378,-0.0208070613,0.244070828,0.2368531674,0.0986188799,0.3046800196,-0.0035825372,-0.0568122566,0.1515498906,0.1570539922,0.1753869802,-0.5788909793,0.5717278123,0.2261798531,-0.0555760004,-0.0059855236,0.2032668889,-0.0892744437,-0.1207637861,0.5214152336,-0.2620499134,0.0367177539,-0.0025086321,0.0467145182,-0.1930740178,0.6902199984,0.0574214011,0.4553852081,-0.0869448557,0.0456429198,-0.441949904,0.0915855095,-0.1071960181,0.2119730711,0.260627985,0.2983136773,0.1464917809,0.2051634341,-0.4557715058,-0.2137386352,-0.0162081532,0.0483998582,-0.1488635093,-0.1037948206,-0.5044937134,0.1245413795,-0.1303907633,-0.4423992932,0.0877696797,-0.421377629,-0.034125872,-0.0869131237,0.2229063958,0.3233665228,0.4466145933,0.3336482644,-0.0396316722,-0.1049203724,0.0495973863,-0.0795436427,-0.1129743755,-0.027450813,-0.3176705241,-0.0786445513,-0.1492996067,0.1286467314,-0.2904022932,-0.0474702828,-0.2210418582,-0.2627427578,0.0252130646,-0.1029951721,-0.1358454823,0.0992458984,-0.2605099678,0.1753808409,0.2303076088,0.44776842,-0.1406971514,-0.1299131662,-0.2153410167,0.0818004981,0.3826012611,-0.6043995619,-0.2027837336,0.0683153793,-0.165810138,0.2554988861,0.0637983307,-0.2777389586,-0.2199787796,0.2798736095,-0.128414765,-0.2145448923,0.2562736571,0.3233981729,-0.1007619947,-0.1221335828,0.2780945301,-0.06114088,-0.4330615997,-0.2135881484,-0.486518234]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/193","title":"[Tensorflow] Use something else than `from_tensor_slices()`","comments":"Fixed by https:\/\/github.com\/huggingface\/datasets\/pull\/339","body":"In the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?","comment_length":3,"text":"[Tensorflow] Use something else than `from_tensor_slices()`\nIn the example notebook, the TF Dataset is built using `from_tensor_slices()` :\r\n\r\n```python\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions']\r\ntrain_tf_dataset.set_format(type='tensorflow', columns=columns)\r\nfeatures = {x: train_tf_dataset[x] for x in columns[:3]} \r\nlabels = {\"output_1\": train_tf_dataset[\"start_positions\"]}\r\nlabels[\"output_2\"] = train_tf_dataset[\"end_positions\"]\r\ntfdataset = tf.data.Dataset.from_tensor_slices((features, labels)).batch(8)\r\n```\r\n\r\nBut according to [official tensorflow documentation](https:\/\/www.tensorflow.org\/guide\/data#consuming_numpy_arrays), this will load the entire dataset to memory.\r\n\r\n**This defeats one purpose of this library, which is lazy loading.**\r\n\r\nIs there any other way to load the `nlp` dataset into TF dataset lazily ?\r\n\r\n---\r\n\r\nFor example, is it possible to use [Arrow dataset](https:\/\/www.tensorflow.org\/io\/api_docs\/python\/tfio\/arrow\/ArrowDataset) ? If yes, is there any code example ?\nFixed by https:\/\/github.com\/huggingface\/datasets\/pull\/339","embeddings":[-0.116660744,-0.1116051227,0.0466688164,0.1800771952,0.2203834057,0.1262966841,0.2330494523,0.3849458992,0.0186054613,-0.0674565583,-0.1806490421,0.5607097149,-0.1538081765,0.0479073375,0.3145027161,-0.1978612542,-0.06258367,0.2236665189,-0.1312798113,-0.0263682306,0.0066866903,-0.0463060923,-0.1073472276,-0.088554576,0.032015346,-0.0064491914,0.0265430324,-0.1394851357,0.4118453562,0.0290045142,0.4061566293,0.0131213637,-0.0511258431,0.1393944025,-0.0001119228,0.4072353542,0.0204419754,-0.1702505201,0.0472429879,-0.0584904701,0.1860137731,-0.2878918946,0.2309188694,-0.2544854879,-0.0059076468,-0.0094684074,0.2633423507,-0.0498082489,0.2980515361,0.1057504117,0.1348195374,0.5306053758,-0.2260842025,0.3367777467,0.0912019238,-0.1178961769,-0.2019496113,-0.2042852491,-0.1622491181,0.0314872712,0.0234517865,0.1973404735,-0.2255287617,0.1043074727,0.4564765692,0.1337218285,-0.2146449536,-0.3219486177,-0.2837553918,0.4095915258,0.2194810361,-0.0786180124,-0.1582235396,-0.5371283889,-0.1845138073,-0.2828838527,-0.1773084104,0.101016961,-0.1900253892,0.1556830406,-0.1723771691,-0.1207362562,-0.2255235314,0.1581420749,0.2015101612,-0.0289285146,0.1120013967,0.1296851784,0.1851750463,0.0795824379,0.2493613809,0.0471981019,0.3328034878,0.3882896006,-0.3782628477,-0.0086007193,0.146340102,-0.4826616943,0.0699963197,-0.2117191553,0.3515687585,0.2103997767,-0.3128161728,0.1989486367,0.0021088396,-0.0621975027,-0.3926285207,0.103847973,-0.1195091978,-0.4943547547,0.1039095446,0.0307303984,-0.2326624691,0.0503006801,0.0367160253,-0.5507951975,-0.072317712,0.252201438,-0.3098458648,-0.4001550376,-0.1980906725,0.0633779317,0.0164578427,0.2393256724,-0.164060235,0.4426185787,0.0971719027,-0.0955336317,-0.2753515542,0.0334721133,-0.2297487259,0.0979316384,-0.0463443771,-0.08643163,0.1844464988,-0.1457206607,0.2371601164,-0.1338221431,-0.1494655907,0.1965449005,0.2156960666,-0.0679585934,0.0890060291,0.2360676378,-0.0195685923,-0.0055341441,0.0628061369,0.1503440887,-0.2110894322,0.2668524086,-0.544651866,-0.2289150357,-0.0310259201,0.0997221768,0.0872092247,-0.2346030474,-0.3701964319,0.3713743985,0.067808494,-0.1272760928,-0.1358905882,-0.1777167767,-0.4521149397,-0.3030212522,0.3161785603,0.0828047395,-0.389200449,-0.1551015973,0.0260026157,-0.0260311197,0.2034539282,0.4642057419,-0.2139894813,0.4173429012,-0.0887104794,0.0802951455,0.7254242897,-0.1289593577,-0.2037173063,0.2343407869,-0.1199623719,0.2463909537,-0.0326416902,0.4582307935,0.0319910534,-0.0141893923,0.3389745057,0.666140914,-0.1247314513,0.0769019425,-0.0627815053,-0.3071115315,0.3318338394,0.2739611566,0.031126,-0.1117577404,-0.2229948491,0.3481730521,0.1386574358,-0.0103390859,-0.2150048018,0.1298813969,-0.1560617983,0.0236809906,-0.1622191519,0.0227314979,-0.6967723966,0.2830310166,0.3161193132,0.0377615653,0.1913648099,-0.1302850991,0.3528954983,0.1014092192,0.0714564249,0.162467733,0.0007523604,0.0581138283,-0.0149371549,0.027390644,-0.3347482681,0.2596382797,-0.4186545908,-0.0959955752,-0.366494596,0.226458475,0.4288584888,0.0451712571,-0.1098832488,0.305321157,-0.3418676555,-0.1708247066,0.1277873516,0.2295331508,-0.0453289971,0.2874976993,-0.6202182174,0.4778478146,0.0555733629,-0.0583342612,-0.0923146904,0.1159153432,0.0270916708,-0.2289398313,-0.0687075928,0.2539217174,-0.1892859042,0.1774619669,0.2326463461,-0.2715439796,0.1563649476,0.0818345323,0.0891130492,-0.0938787088,0.4394506514,0.0465944819,0.1967562586,0.2774636745,-0.5544955134,0.1731863767,0.6136518121,-0.0394200683,-0.0405358113,0.2563566864,-0.0228590034,-0.2173868865,-0.0561440624,-0.1544160247,0.2477199733,0.1352442354,0.064236261,-0.0038284287,-0.1434254348,-0.2045878321,0.234118849,0.0951962844,0.3736137748,-0.1295615733,0.0556388572,0.1781276017,-0.2461501211,0.0821595192,0.2261089087,0.1464416534,-0.2266122997,0.0176574644,-0.2886194885,-0.2937682569,-0.0674222782,0.0091330996,0.0897554979,-0.1653913558,-0.0988003835,0.3355363309,0.000246175,-0.0411604568,-0.1467612982,0.3236524761,0.1794085801,-0.4558125734,-0.1005948633,-0.0808983445,-0.1463606954,0.0499090962,0.3682181537,0.1957109421,0.2710556686,0.0392346829,-0.1113201231,0.0904184952,-0.0815649927,0.1862458885,-0.1482644379,0.1044952571,-0.0474045314,0.1632592082,-0.1106185615,-0.1593270153,-0.1023305506,-0.397680521,0.043872159,0.1466225535,0.0310158581,0.144037351,-0.2408084571,-0.1137516499,-0.324762851,-0.3627872765,0.0713944733,0.1202331856,0.2700927556,0.4462988973,-0.0547508486,0.2800708413,0.2731832564,0.0885060281,0.0673438087,-0.1455911845,0.5088889599,-0.2432525605,-0.2709149122,-0.1633565724,-0.1437267512,0.1417847574,0.3167287111,-0.7442822456,0.1360020787,-0.1496150792,0.086134553,-0.1320772916,-0.011281101,0.2550677359,-0.2994706631,0.0539492406,-0.0358297899,0.0780399591,0.1231113672,0.014447364,-0.210572198,0.3511354327,0.4424880445,0.3256440759,0.7581339478,-0.0915192589,-0.2126371711,0.2601986229,-0.2143874019,0.2724345922,-0.2106448561,-0.0849344805,0.0830767974,-0.0028790878,0.0346267782,0.2712316811,-0.0871085227,-0.2041716576,-0.0731351823,0.1692118645,0.1089330763,-0.1951062083,0.2374643683,-0.2068601549,0.1207391173,0.140493378,0.1942009926,-0.3582061827,-0.4447306097,-0.0272860508,-0.138525337,0.3929773867,-0.0016727933,-0.3647957444,0.1239695325,-0.7393345833,0.1111603603,0.0480134934,0.334731698,0.0711235031,-0.2776852548,0.0133157838,0.095844686,0.6223118901,0.1535318345,-0.0785680041,0.1433730721,-0.159861058,-0.5712636113,0.0204694308,-0.0490073487,-0.0065575219,0.241466254,0.1628280133,-0.3164613843,-0.5265593529,-0.040383067,0.0014244247,-0.3339433968,-0.0475279726,-0.2083853185,-0.2354675978,0.0185164157,-0.3124831915,0.1018286869,0.1757291853,-0.1033651009,0.0255397148,-0.1918083429,-0.0034628899,0.255782783,0.0019134685,0.1712108552,0.5160368681,0.065749079,-0.0358960219,0.0738140047,-0.1954537183,0.2673788071,-0.1233438775,-0.2258467823,-0.000462605,0.1437010914,0.4169843793,0.0335113741,0.0347682387,0.127243638,-0.1701928526,0.0135068502,-0.241546765,0.4162100255,0.1291033924,-0.2363507003,-0.1444260478,-0.4629584253,0.5663648844,0.288269043,0.1206229255,0.4634557962,-0.1333227009,-0.3694827557,0.461814791,0.3042482436,0.4902560115,-0.3755793273,0.5610494018,0.7292613983,0.542527318,0.9250772595,-0.3353103101,0.2741054893,-0.0839698613,-0.3626478314,-0.039656885,-0.2858976722,-0.3563285768,0.0284461118,0.1306633204,0.2696462274,0.0959525257,0.3089224696,0.1840004176,0.3724567592,0.3484193683,-0.4121619463,-0.1944043785,0.0756401941,0.0262276158,-0.2067767382,-0.0937602893,-0.1738624424,0.2146150619,0.0471815951,-0.0201035105,0.0518844649,-0.187252,0.2744739652,0.0940778479,-0.0738221332,0.0601249933,0.281165868,0.0832813829,-0.2018076181,-0.1275529414,-0.1018683165,-0.0147816371,0.0114682028,0.1153878644,-0.0640865937,0.6001852155,0.0176500566,0.2220742404,0.1015948504,-0.0679918155,-0.5797640085,-0.1944041252,-0.063203685,0.0570659973,0.109222278,-0.612977922,0.2294251472,-0.1389377862,0.1996071339,0.0692502707,0.1149780378,-0.1654835045,0.0411083251,0.1082573161,-0.2313316315,0.2095635533,0.0478097908,0.096393086,0.0504392944,0.3731194735,0.1880303472,0.0065255202,-0.0098005114,0.1005287319,0.3207699656,-0.0419555716,0.0228236318,0.0497651733,0.2286252528,-0.0217973534,0.3043098152,-0.2443228662,-0.0437134951,-0.2865981758,-0.2280520201,-0.2498139888,0.4317429066,-0.0849840045,0.1506949067,0.0163525064,0.1794147789,-0.1403457373,0.1965683997,-0.2113343775,0.0189389735,0.0380058177,0.0661324784,-0.0964129269,-0.0021987699,0.151997298,0.1196000129,0.0447120331,0.137424618,-0.3268282413,-0.1889668703,-0.0247736499,0.1811720133,0.012248259,-0.1697204262,-0.0007860431,-0.3830602169,-0.3293018639,-0.2037566751,0.1107422113,-0.0863150954,-0.2464134842,0.0272791982,-0.0510970838,0.2972998023,-0.09729103,0.0780955032,-0.0918337777,0.2862249613,-0.0990943313,0.2213068902,0.0873336196,-0.0276212972,-0.3003615439,-0.1133396924,0.2102910876,-0.1604006588,-0.0226330943,-0.3142517209,-0.0525763631,-0.2129766345,0.0955837518,0.3400069773,-0.2610240579,0.1050357595,0.2386422306,0.0895265117,-0.0798662901,-0.1526191086,0.3108833134,0.052581273,0.0088364808,0.1850762814,0.0579732507,0.2306202948,-0.0898612961,-0.0946376622,0.072391741,-0.4811540842,-0.2274774313,0.2926879525,-0.153472349,-0.0414161943,-0.133198902,0.1475238651,0.1432465166,0.2163837105,0.2558257282,0.2151753604,-0.1034736186,-0.0595532544,-0.0741897374,-0.3749053776,-0.4224845171,0.4127454162,-0.0853438377,0.5822167397,-0.2833659947,0.0583644807,0.1496585459,-0.2695186734,-0.1613488793,0.3052074015,-0.1659948677,-0.0829334036,-0.1647805125,0.0893788114,-0.0893406048,0.2548293173,-0.1071101204,-0.2667258382,-0.1290975511,0.2559062839,0.1318760961,0.1914573908,0.0816254541,0.121241644,0.2454720289,-0.0977455154,0.2623630464,-0.2892899811,-0.055555474,-0.3603016138,0.0254160948,0.06993182,0.2479657978,-0.2923053205,0.1680145413,0.0710464343,-0.032644812,-0.0449409373,0.2520143688,0.2019283772,0.0915923789,0.2859998643,-0.0184584353,-0.0295200981,0.1479072422,0.1325066686,0.1316167563,-0.6284454465,0.515147388,0.1556161344,-0.0190981254,-0.0340366848,0.1886773705,-0.1353802681,-0.1245198399,0.5322321653,-0.2552560866,0.0601330101,0.027194608,0.037578579,-0.2146689594,0.678129375,0.0429415591,0.4902234674,-0.0953047872,0.077189602,-0.5144267678,0.0473848805,-0.1438789666,0.1510405988,0.2482191473,0.2851995826,0.1650612056,0.1831350774,-0.4026344419,-0.2172547877,0.0010725436,0.0091492534,-0.1438698024,-0.1011202112,-0.5258147717,0.1658557653,-0.0715879202,-0.4269750118,0.0692956597,-0.4145279229,-0.0790869743,-0.0772747025,0.3149172664,0.261957109,0.3842814565,0.3185015619,-0.0652048513,-0.036174316,0.0677151456,-0.1221808568,-0.1518414617,-0.0092985546,-0.3222432435,-0.0636370406,-0.1727324873,0.1258154362,-0.3455685973,-0.0646836013,-0.1767425835,-0.2360853255,0.070718348,-0.1448781043,-0.1378395408,0.1311587691,-0.2639543116,0.2212919444,0.1916440129,0.5122210979,-0.1707575619,-0.0997463539,-0.1716266721,0.0624559484,0.4390285909,-0.6664078832,-0.1748321652,0.0870561153,-0.1456872225,0.2531285286,0.0543072671,-0.3634650409,-0.2087084204,0.2969222665,-0.1163136363,-0.2407892197,0.2655600607,0.3191566169,-0.1132362187,-0.1163752154,0.3534495234,-0.0521650054,-0.4390779436,-0.1520812958,-0.4916895628]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/192","title":"[Question] Create Apache Arrow dataset from raw text file","comments":"We store every dataset in the Arrow format. This is convenient as it supports nested types and memory mapping. If you are curious feel free to check the [pyarrow documentation](https:\/\/arrow.apache.org\/docs\/python\/)\r\n\r\nYou can use this library to load your covid papers by creating a dataset script. You can find inspiration from the ones we've already written in `\/datasets`. Here is a link to the steps to [add a dataset](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/CONTRIBUTING.md#how-to-add-a-dataset)","body":"Hi guys, I have gathered and preprocessed about 2GB of COVID papers from CORD dataset @ Kggle. I have seen you have a text dataset as \"Crime and punishment\" in Apache arrow format. Do you have any script to do it from a raw txt file (preprocessed as for BERT like) or any guide?\r\nIs the worth of send it to you and add it to the NLP library?\r\nThanks, Manu\r\n","comment_length":68,"text":"[Question] Create Apache Arrow dataset from raw text file\nHi guys, I have gathered and preprocessed about 2GB of COVID papers from CORD dataset @ Kggle. I have seen you have a text dataset as \"Crime and punishment\" in Apache arrow format. Do you have any script to do it from a raw txt file (preprocessed as for BERT like) or any guide?\r\nIs the worth of send it to you and add it to the NLP library?\r\nThanks, Manu\r\n\nWe store every dataset in the Arrow format. This is convenient as it supports nested types and memory mapping. If you are curious feel free to check the [pyarrow documentation](https:\/\/arrow.apache.org\/docs\/python\/)\r\n\r\nYou can use this library to load your covid papers by creating a dataset script. You can find inspiration from the ones we've already written in `\/datasets`. Here is a link to the steps to [add a dataset](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/CONTRIBUTING.md#how-to-add-a-dataset)","embeddings":[-0.1738835275,-0.0641050488,-0.0457872935,0.0695887059,-0.3965011537,0.136375919,0.0676105917,0.4003833532,-0.0682438016,-0.3222419024,0.1313437074,0.1833892018,-0.082505092,0.0789601728,0.3914310634,-0.0407906063,-0.0588860214,0.2828747034,-0.0328622721,-0.0365570188,0.1609279215,-0.0212868117,-0.1044443101,0.0072559575,-0.1917187274,-0.3496663272,-0.1661534458,0.1569089741,-0.1471924633,-0.4131221771,0.1118626893,0.0859132558,0.5545091629,0.1522884071,-0.000120873,-0.4648412466,0.146173358,-0.0721157268,-0.1525297612,-0.2088950127,-0.2326005995,-0.0918239728,0.1721589863,-0.3390492201,0.0773122162,-0.4477847517,0.1712197363,-0.2712031901,0.6657809019,0.3493712246,0.0733593181,-0.1018955708,0.1893394291,0.2285748869,0.479952991,0.5429444909,-0.3052916229,0.1529542953,0.2607792616,-0.2076222152,-0.2825189829,0.1587996036,-0.0976548865,-0.2390976697,0.3305785358,0.1180302352,-0.2703990638,-0.3330769241,0.048938185,0.396437645,0.369625777,-0.4938062727,0.0485387929,-0.0501641035,0.1894060224,-0.4890819192,-0.148752749,0.6180791259,-0.5020700693,0.3597390354,0.3344428837,-0.3816461265,-0.433984071,0.1284186244,0.4575878978,-0.0715703219,-0.0471398421,-0.0758280829,-0.0502895489,0.1744943112,0.2112369835,-0.2455614209,-0.1487576514,0.3470050693,-0.1437056065,-0.1627587527,-0.487794131,-0.2078185678,0.2758220434,-0.0297537744,0.3700711131,-0.1301807463,-0.0390953608,-0.0916772112,0.0394635387,-0.0950186923,0.4510582089,0.2062097341,-0.0090025906,-0.2627836466,0.1622353196,-0.0809461623,-0.2055873126,0.0966669247,-0.1564095169,0.0332183614,0.0496243276,-0.1927904636,0.0111123761,0.0060190335,-0.4366478324,-0.0245696492,-0.2640629709,0.2335275561,0.1825598478,-0.0913557038,0.2279809117,0.3980584145,0.0025834572,-0.1910063475,0.1229753047,0.0735131875,-0.2823687494,0.2808929086,0.247234419,0.0487667546,-0.0020069233,-0.0952516273,0.0477823392,-0.3019622564,0.1904058754,-0.0115250554,0.1669414788,-0.1192823574,0.2679129839,0.0407756977,-0.1499408185,-0.1275525689,-0.1013562828,0.369210273,-0.3687529564,-0.1211427599,-0.1366639584,-0.0014926877,0.0550353937,-0.1415070891,-0.2777965963,0.2512921989,0.1334351301,0.1928410381,0.0556442216,0.2801023126,-0.0139733898,-0.2512498498,0.0342513993,0.4760324955,-0.2327194661,0.1828200668,-0.1303556114,0.4220273793,0.0418921448,0.1999807358,-0.2509055138,0.3373116553,0.0285330359,0.3784561157,0.7308462262,-0.0782934576,0.2264495641,0.4033631384,0.1815540344,-0.3522212803,0.1490732729,-0.0432250313,0.1029437035,-0.0991303548,-0.3000408709,0.4925096035,0.1188171133,0.0380928889,-0.0789848194,-0.2024594098,-0.1806162894,0.0138285318,-0.1424853951,-0.152122587,0.0905092135,-0.4588907659,0.4787380397,-0.2984763384,0.4021227658,0.290397644,0.0415256247,-0.0141991246,0.2463943064,-0.1117685288,-0.1824048758,-0.3121178448,-0.406529814,0.2906285822,-0.2542756498,-0.2743141353,-0.2539533079,-0.1194468737,0.0768854916,-0.0305947848,-0.031456843,-0.2436130941,-0.2193366289,0.0177638438,-0.207388252,0.2594847679,-0.2797718942,0.1349848956,-0.2302947938,0.2156806141,0.0819953457,-0.1958412677,0.2466090024,0.2380749583,-0.2799951732,0.1489574611,0.0656598657,0.1418323517,-0.0535547696,0.2025629133,0.4277290106,-0.2808156908,0.0847726315,-0.6659778357,0.3037104905,-0.0413484052,0.2257367522,-0.1121066958,-0.3082047403,0.4449282885,-0.062422622,0.1924302876,-0.1610332429,0.0182838421,-0.0741204917,-0.030854173,-0.0838211328,0.1798774302,0.0551718883,0.4346465766,0.156640172,-0.1513433754,0.0273626857,0.085149169,-0.127518326,-0.1711125374,0.2346558124,0.2375770062,-0.3029533029,-0.103598401,-0.1672834307,0.0483935922,0.0925913677,0.1848359704,-0.150638476,-0.040357966,0.080506593,-0.0968746543,0.3708855808,-0.030180322,0.2931441367,-0.1610877663,-0.0907879397,-0.0233558584,-0.1795140058,-0.2806978226,-0.0475527011,0.1384664476,-0.2195436507,0.0284088142,-0.0660056621,-0.3738856018,-0.3311776817,0.1296107471,-0.0414256863,0.2527340651,0.0128344148,-0.2428025901,0.1606639624,0.0032889307,-0.0153890615,0.5403869152,0.0907274038,0.0718730465,-0.1886594892,-0.5223944783,-0.1820635647,0.0564001463,0.2462329566,0.420417279,0.3654943705,-0.1879571527,0.120198153,-0.0372109525,-0.0155866034,0.0392338969,-0.1758018136,-0.2756327987,-0.1856742948,0.1290773302,-0.5696607828,-0.1051902771,-0.2161619216,0.0416453741,-0.1757654697,-0.2118123025,-0.0726570338,-0.0401740186,0.1962092221,-0.5246594548,-0.1546359658,-0.126838237,0.3574329615,0.3072159886,0.2926772535,-0.0244840514,0.2397527844,0.1166200936,-0.307590872,-0.0535371862,0.1910519451,-0.0654017478,0.3787032068,-0.3172714412,-0.5370230675,0.0119816223,-0.0533121489,-0.0727550015,0.3109488785,-0.3548155725,0.4196316004,0.0823071674,0.0295326319,-0.1032703221,-0.0128584262,0.2451626956,-0.0004229056,0.0613565557,-0.0583934337,0.0904885307,0.3568631411,-0.128423661,0.3276356161,0.3448913991,0.1456566602,-0.0466144495,0.5362749696,0.2517481744,0.0607058927,0.17636621,-0.1187140569,0.0775540322,0.1054610312,0.0318697952,-0.0742828995,0.1269871593,0.0374814793,0.1802240461,0.3182495534,-0.3227260113,-0.200630188,-0.2434081435,-0.3722504377,-0.1931559443,0.3950964212,-0.2958456576,0.2528685629,-0.1175625622,-0.2101007551,-0.1779503524,-0.4720108211,-0.1027451977,0.2386796325,0.2641419172,0.1729047596,-0.0420056544,0.3613616228,-0.6206264496,0.2994303405,0.1829653829,-0.1163880825,0.1148842499,0.0537875108,0.2133235633,0.0867825225,-0.1169250682,0.1683696806,-0.0590527914,-0.0143140052,0.3365096152,-0.3689166009,0.3253724277,-0.2826525867,0.0372405872,0.2201362997,-0.1127414927,-0.4396895766,-0.220167622,0.1740350872,0.1490804255,-0.1108866185,-0.4020351171,0.1299974024,0.0058027715,-0.3960943222,-0.1459907144,0.3968577683,-0.0160333719,-0.124976404,-0.1553913951,-0.0196636766,0.1007988602,0.2227826715,-0.2305995077,0.201177299,-0.2508752048,0.0217714943,-0.0818568468,0.3278177083,0.0673632324,-0.0015108865,0.1869726032,-0.2221156806,-0.0643138066,-0.3994572461,0.5800675154,0.1728699654,0.1120888665,0.156195417,0.1852512956,-0.30442065,-0.1199287251,-0.0110874213,0.1350952387,-0.1108947471,-0.4100241363,-0.3892837465,0.5749381185,0.0482090861,-0.034146335,0.2953098714,0.0565155409,-0.2902967632,0.6706213951,0.2002020478,1.2255791426,-0.1759890765,0.2871077061,0.0548758581,-0.1076686233,0.5923514366,-0.4202053547,-0.2130983174,-0.5059722066,0.3042323291,-0.1102536172,-0.145443514,0.1546525508,0.3318037093,-0.0661814138,0.10463579,0.2753223479,-0.0457415581,0.1352182329,0.3248891532,-0.2197948843,-0.4144165814,-0.1502167881,-0.0940712616,-0.1285450906,0.0469284542,0.1130556315,-0.4174150229,-0.3201584816,0.0748212859,-0.4243671298,0.1003520116,-0.0081342524,-0.2222197652,-0.2661272287,-0.6077933908,0.3085106313,0.3429758251,0.0942162573,-0.007780374,-0.285155654,0.091727607,-0.2546536624,-0.3908624649,0.2797312737,-0.1298853904,0.0750142932,-0.1504433155,-0.205621168,0.151090458,0.053391017,-0.2871479988,-0.0065818764,-0.2250326276,0.4990297258,-0.1214828566,-0.0603453442,0.3270651698,-0.4100150466,0.0054831076,0.0734867677,0.3840418756,0.1089474484,0.0527783185,0.2629637718,-0.2531400621,-0.045207046,0.1750053614,0.3665904701,-0.2257860899,0.5001036525,0.4196083546,-0.1537497491,-0.2953295708,0.2761527002,0.3888967335,-0.8429343104,-0.1055258289,-0.0083409594,0.0986773148,0.0162296593,0.1006726176,-0.0775003433,-0.0465541072,0.0103118178,-0.5420808196,-0.2696763873,0.1499106884,-0.0966627523,0.0754451081,0.3981229961,0.0965250134,0.041537229,0.1355668753,-0.2073509991,0.0801068395,0.5916288495,0.1752158105,0.1062238514,0.0970886201,0.0198823661,0.0255561955,-0.0247851666,0.1171043515,0.1111515611,-0.1314202845,-0.015419431,0.2028451115,-0.0608616173,-0.2965806425,-0.3098234236,-0.4513745904,-0.2555300593,-0.1911258399,-0.1472960562,-0.1557670236,-0.1328599602,0.1309082806,0.0779716596,0.0143913049,-0.4827416837,0.2256817967,-0.2151062936,0.2515775561,-0.0972577259,0.1495199949,-0.0002763463,-0.0502340123,0.0445649363,0.1290832162,0.2469741553,0.1895885468,0.5016093254,-0.119362846,-0.0470201895,-0.3728876412,0.2044701874,0.2573816478,0.0903843269,0.0638088062,0.2950477004,0.0918234661,-0.2226513326,0.2888726294,-0.2305031419,-0.2013943046,-0.1863033026,0.1267650425,-0.0210701637,-0.1975570172,-0.3109132051,-0.1603404433,0.1802659631,-0.0151638128,-0.3553688824,0.1900040954,0.2255921811,-0.0418221541,-0.1311610192,0.2277486175,0.2554960549,-0.0632269755,-0.0539583862,0.1503238529,0.7298562527,0.266744405,0.1422621608,-0.1678468734,0.0415041745,0.4384757876,0.2494751215,0.0143217212,0.2309010029,0.50416255,0.6058551669,0.1081209183,-0.0487144254,0.2923032045,0.1237493679,0.0365413502,-0.4599585533,-0.0659711435,-0.0484821796,-0.2736521959,-0.1924760789,-0.0760047138,-0.147340402,0.3707124889,0.1580518633,-0.0557747856,0.1649748683,0.0714254305,0.3248580098,-0.2492867559,0.0821079612,-0.1200413331,0.0153636364,0.0374741964,0.4358315468,0.3315368891,0.1721892655,-0.1392871141,0.1491888762,-0.0002270731,-0.1071798727,0.1763477772,0.196033448,0.4006010592,0.0384391807,0.0730996132,0.0452424586,0.0279879496,-0.1260420531,0.3155634701,-0.2134244442,0.0998877361,-0.1506026387,0.0747957677,-0.3740047216,-0.1697544008,0.0998467132,-0.1613108516,0.3633498847,0.606145978,-0.23281385,-0.0464297645,-0.1182134226,0.011206436,-0.1033501923,0.4182167947,0.343481034,0.1270029545,-0.29066661,-0.2813269794,-0.0470382348,0.1085739881,-0.0395549089,-0.2372414172,-0.0532982945,0.0345827974,0.2613586485,0.4125305712,0.048847571,0.134253636,-0.1178583652,0.4718045592,-0.2880142033,-0.0419725329,0.273188442,0.5164784789,-0.1628136486,-0.1024344265,0.1522894502,0.3586497605,-0.0644897148,-0.2367387563,-0.0892239213,0.3275832534,-0.1779955328,0.2237705141,-0.0907740444,0.152932629,-0.0516743921,-0.1233341917,0.0046204906,-0.0874487013,0.1057164744,0.4439938366,0.3223761916,-0.1910901815,-0.2652257979,-0.0015395243,0.0471139923,0.2749512494,-0.2875999808,-0.0508479327,-0.2726785243,-0.0233767126,-0.2617739141,0.1474382579,-0.0348756872,0.2876520455,-0.21054627,0.1158325076,-0.1269012988,-0.1302878261,0.0916018113,-0.2509090006,0.0399660803,0.2685217857,0.2169538736,0.5681504011,0.3674057722,-0.5660687089,0.2106537074,0.2342372686,0.0702542588,0.103604883,0.0629384741,0.0025862455,0.1040326729,-0.0169692133,0.3965064287,-0.058482673,-0.0382047892,-0.4614567757,-0.4983665347]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/192","title":"[Question] Create Apache Arrow dataset from raw text file","comments":"Hello @mrm8488 and @lhoestq \r\n\r\nIs there a way to convert a dataset to Apache arrow format (locally\/personal use) & use it before sending it to hugging face?\r\n\r\nThanks :)","body":"Hi guys, I have gathered and preprocessed about 2GB of COVID papers from CORD dataset @ Kggle. I have seen you have a text dataset as \"Crime and punishment\" in Apache arrow format. Do you have any script to do it from a raw txt file (preprocessed as for BERT like) or any guide?\r\nIs the worth of send it to you and add it to the NLP library?\r\nThanks, Manu\r\n","comment_length":29,"text":"[Question] Create Apache Arrow dataset from raw text file\nHi guys, I have gathered and preprocessed about 2GB of COVID papers from CORD dataset @ Kggle. I have seen you have a text dataset as \"Crime and punishment\" in Apache arrow format. Do you have any script to do it from a raw txt file (preprocessed as for BERT like) or any guide?\r\nIs the worth of send it to you and add it to the NLP library?\r\nThanks, Manu\r\n\nHello @mrm8488 and @lhoestq \r\n\r\nIs there a way to convert a dataset to Apache arrow format (locally\/personal use) & use it before sending it to hugging face?\r\n\r\nThanks :)","embeddings":[-0.104102768,-0.0871481746,-0.0996507183,0.1544027179,-0.3756233454,0.1595188379,0.1173257455,0.4278326631,-0.0435959734,-0.2937572002,0.0730511174,0.1151047051,-0.0833791643,0.1475037634,0.2515309751,-0.1538529694,-0.0880172402,0.2165444642,-0.2933829427,0.0157115478,0.2740622461,-0.0247493405,-0.0877847821,0.0122401519,-0.1714434177,-0.236255154,-0.1281672567,-0.0033141167,-0.1450779289,-0.337683022,-0.0253848862,-0.0293533411,0.5116750002,0.2051157355,-0.0001217041,-0.3465603292,0.1274175495,-0.0041155894,-0.2411180586,-0.2082494944,-0.1519872993,-0.0090410477,0.1944746971,-0.3356013,-0.0600214079,-0.3797875345,0.0302869622,-0.468857348,0.839243114,0.3928316534,0.075771749,-0.1602265984,0.188713938,0.2179340422,0.3413510919,0.5511066914,-0.3402373195,-0.0774220377,0.1499584019,0.0333747119,-0.1650861204,0.1985009313,0.0439888611,-0.1532467604,0.3540650904,-0.0316225402,-0.2806330621,-0.2210430801,0.1231576055,0.326362133,0.4954088926,-0.4750736356,-0.0450375564,-0.0119728958,0.2050462216,-0.4249300361,-0.2257204354,0.5759353638,-0.4589058161,0.4854776561,0.1496035159,-0.4008947313,-0.4561852813,-0.0745547041,0.3925387859,-0.1183444932,-0.1044038236,-0.063812755,-0.1421257108,0.1629603058,0.2270020247,-0.3375430405,-0.1277933568,0.2690395713,-0.0427741259,-0.2151730806,-0.5973445773,-0.1973363012,0.1853763014,0.0773296952,0.3857090473,-0.1129084006,-0.0476093516,-0.2783759832,0.0771988481,-0.0727790147,0.4339829087,0.1329318434,-0.0070713861,-0.2632562816,0.1707571894,-0.1187153831,-0.0946807563,0.0173992645,-0.1367946863,0.2015711069,-0.0105000706,-0.2780988812,-0.0333565138,-0.0199090876,-0.38649562,-0.0769996718,-0.2138036489,0.0261282306,0.201365307,-0.146756202,0.2247111648,0.3776323199,0.0520545654,-0.2802136838,0.176209569,-0.0180223454,-0.2320825905,0.281655401,0.0870089158,0.165380612,-0.1353075802,-0.0685042292,-0.0582958013,-0.1452396214,0.2524908185,-0.036170084,0.2175890356,-0.0666913912,0.2069645077,0.0162982009,-0.1201662347,-0.0072476715,-0.0758111998,0.3549625874,-0.181900546,0.0353412069,-0.0445053093,0.0157345664,0.0706239417,-0.0845680982,-0.2357393354,0.1604427993,-0.0032508643,0.3336627185,0.0702999011,0.2868445218,0.0605237335,-0.142325744,-0.061350055,0.4451697171,-0.3385572433,0.1481717825,-0.0977572948,0.2030458152,-0.1150634736,0.3096966743,-0.2214799076,0.2855267525,-0.0217324067,0.2480599135,0.7609063983,-0.0031955305,0.1953616738,0.4988286793,0.032612469,-0.3793379366,0.0923480466,-0.0841630176,-0.0858473703,-0.0297111217,-0.2868655026,0.5281491876,0.1118569747,-0.0169506893,-0.1194049045,-0.17205365,-0.1165966019,-0.05393878,-0.2323151678,-0.0548298918,0.0921822786,-0.4823833406,0.4895274341,-0.3356698155,0.4532503784,0.237851724,0.0251209922,-0.0393288396,0.1749618948,-0.0319602005,-0.0986610278,-0.356051147,-0.3879675865,0.2141292244,-0.3269676268,-0.2382126153,-0.269683212,-0.1252905577,-0.0686217174,0.0822976828,-0.0197964236,-0.2352799475,-0.1557948887,0.0143629862,-0.2127346098,0.3180286288,-0.1101872921,0.1491360366,-0.2236446142,0.1163673326,0.1057078913,-0.004393863,0.2375984639,0.1367183775,-0.2308913916,0.0524221845,-0.0545833111,0.0032960109,-0.0711125582,0.1790067852,0.5197834373,-0.2920793891,0.0735270306,-0.6826456189,0.3104538023,0.0175315402,0.2184136361,-0.1010279655,-0.465402782,0.4807801247,-0.0809450373,0.1903368831,-0.1441085488,0.0432977416,-0.137209937,-0.1126281694,-0.1677837819,0.1439184099,-0.0678353235,0.3569740951,0.2148139775,-0.0920701772,0.0997203216,0.031152619,-0.1916439235,-0.1972690076,0.232971862,0.2144037038,-0.2332379818,-0.0765687302,-0.0611813888,0.0409181267,0.117594786,0.1491021514,0.0216915365,0.0765997991,0.2202986926,-0.1371531934,0.3341186047,-0.1321943402,0.3828599453,-0.1595306695,-0.1652635932,0.0211987104,-0.0997905806,-0.1932457834,-0.1010821089,0.1401480436,-0.0816581771,-0.0382212214,-0.1377267987,-0.4040078521,-0.3725430667,0.2617996633,-0.0313793011,0.2539508939,-0.0742642134,-0.4246178567,0.1473329961,-0.1559281051,-0.0753659233,0.5126544833,0.0790228024,0.0604605265,-0.283616811,-0.3982836306,-0.1630405188,0.0774199739,0.3069842458,0.4057583511,0.3491528034,-0.082093887,0.1948646009,-0.1915400624,0.0942462608,0.0271473583,-0.0349812098,-0.3089195788,-0.1563499421,0.1664443016,-0.554985404,0.1195530668,-0.1840698421,0.116887562,-0.2389353514,-0.2792014182,-0.0752747208,-0.0213916749,0.2210385203,-0.4112195671,-0.1371773481,-0.0841433927,0.4223146439,0.233963266,0.2500327229,-0.041657839,0.1320338845,0.1708656251,-0.23383075,-0.072331056,0.1905654371,-0.1538463831,0.2510851324,-0.3682453632,-0.5960060358,-0.0636397153,-0.0158421248,0.0210041441,0.3126493692,-0.421554476,0.5755354166,-0.0353645161,0.0539309606,-0.1453753859,-0.0059959465,0.2033376247,-0.0336186029,0.0440981165,-0.0300906785,0.0854771137,0.3154038489,-0.1693241596,0.3524861038,0.4192026258,0.1349170804,0.0818894729,0.6222640276,0.2455940098,0.1014534533,0.2365687639,-0.0248210039,0.0064176605,0.0260796119,0.0519021042,-0.1404498816,0.1951198727,0.1238486767,0.0736161694,0.2893808484,-0.095873706,-0.2961913049,-0.2651782334,-0.3516783118,-0.20116359,0.297000736,-0.3599576652,0.2795423567,-0.1559702754,-0.2171825767,-0.0630719811,-0.3113392591,0.0599671379,0.1286876649,0.3190680146,0.1490600854,-0.2706946433,0.3482155502,-0.6118110418,0.3497489989,0.0516242273,-0.1002971157,0.0567213632,0.0305633433,0.1468573958,0.0864249021,-0.1788238436,0.1109851077,-0.0253161155,0.009351843,0.2899272442,-0.4226803184,0.3903354704,-0.2288075238,0.1104587242,0.0124553479,0.1247145757,-0.3796065152,-0.2781962752,0.1344010085,0.2421630919,-0.1416956186,-0.364812851,0.2568531334,0.0087629668,-0.1879394948,-0.1105597094,0.3730459213,0.0208727662,-0.0755393282,-0.1968933493,-0.0598220788,0.1574221551,0.1272713393,-0.3288098574,0.1062931642,-0.2729620934,0.0260457769,0.0575465821,0.3884018064,0.0088483561,0.0876742825,0.1769960821,-0.21610412,-0.0689923465,-0.3054362237,0.6498405933,0.2223334312,0.1261841655,0.137787655,0.2107263803,-0.2504128814,-0.0710306242,-0.1461554021,0.041152209,-0.0688971579,-0.4261356592,-0.3491384387,0.5717190504,0.0539535992,0.0027204514,0.1981412321,0.0837309584,-0.2717004716,0.5873886347,0.2792506516,1.1855224371,-0.2742815912,0.2730496824,-0.1355073005,-0.3160175681,0.4253914058,-0.2786796391,-0.2547178864,-0.5003957152,0.4487968385,-0.0769572631,-0.1531478614,0.4294314682,0.4180829227,-0.044714883,0.0234496985,0.2134468406,-0.1962731332,0.1116667837,0.4447535872,-0.3399270773,-0.3009553254,-0.0971889347,-0.0543925725,-0.1266822964,0.1494973898,0.1293238401,-0.440577656,-0.372543633,0.0856238082,-0.2790436447,0.0675965846,-0.1604561359,-0.2300854176,-0.309452951,-0.6526297927,0.3845908344,0.3031029701,0.1611376107,0.0180328768,-0.1456814855,0.0652702153,-0.3059372306,-0.2146332115,0.1400080472,-0.0970303193,0.0149862384,-0.17341429,-0.1473096162,0.1327155083,0.0722906291,-0.2237472832,0.0369485058,-0.2066394836,0.5542367697,-0.220249027,-0.0573714972,0.2082991153,-0.3871695399,0.0955581218,0.0747874081,0.3437507153,0.0100530889,-0.0198437814,0.2923179567,-0.1984134912,0.0622264519,0.1961559355,0.2038142383,-0.1967984289,0.4399947822,0.3845140338,-0.0971687883,-0.3226231039,0.5589427352,0.4894665182,-0.7103244662,-0.0582709685,-0.0873331651,0.1748102009,0.0075119841,0.0737844333,-0.0477354303,-0.0242757592,-0.0415898673,-0.4507725239,-0.3029092848,0.0213054903,-0.0532927364,0.0931829661,0.3139229715,0.0181529522,0.0753159896,0.0419158898,-0.205198288,0.1590944082,0.6357872486,0.177786842,0.0277661327,0.0422961973,-0.0567162335,-0.0001517224,-0.0048280102,0.1385512501,0.0704174638,-0.1743817627,-0.032969851,0.2171880007,-0.0901578069,-0.3297179937,-0.3392930031,-0.4449429214,-0.1549398303,-0.0355866738,-0.1353699118,-0.0862888023,-0.0677067041,0.1622811109,0.1766920388,-0.0913197175,-0.562315762,0.1294585317,-0.1023567766,0.2173252404,-0.1320950687,0.169967398,0.0985718817,-0.0579895675,0.0479214601,0.0091299936,0.3277361095,0.1631406695,0.3537423313,-0.1717619449,0.0479510836,-0.206268698,0.3029599488,0.1416818202,-0.0085540218,0.0744752958,0.295039475,0.0773180798,-0.2086265683,0.2393234521,-0.2136078626,-0.2218509465,-0.2649488747,0.0597081929,0.0045554223,0.026558198,-0.4137258232,-0.2439279407,0.0400485173,0.0096180262,-0.2730487287,0.3426257074,0.1418376267,-0.0367221087,-0.012312944,0.2288075835,0.1612667739,-0.0576144941,-0.1003260389,0.1408076882,0.7164049149,0.2588324845,0.2297164947,-0.1495765299,0.106093891,0.3407496214,0.173584342,0.0604513921,0.3084306717,0.5016956925,0.4905331135,-0.0653373078,0.1039663777,0.2332952768,0.02620022,0.0924511477,-0.297519505,-0.0976213291,-0.0797418803,-0.2705273628,-0.2670359015,-0.0659566075,-0.128050521,0.378308773,0.2242160439,0.0065191844,0.0948653966,0.0762869939,0.3090258241,-0.1967897713,-0.0840060934,-0.0479254611,0.0642542988,0.0148204258,0.4466154873,0.1485183537,0.1166361198,-0.1821920425,0.2195746452,-0.0323844068,-0.134749651,0.2533148527,0.2493117303,0.4374347031,0.0246595573,0.1623628139,0.0377784036,0.060028486,0.0108931288,0.1967499107,-0.1606864929,-0.1082974821,-0.0483869277,-0.0743744001,-0.4557867944,-0.2137721181,0.0723626837,-0.1384686381,0.2017277032,0.6342799664,-0.302331388,-0.0542246737,-0.2243724912,0.0124598276,-0.1151777208,0.4238154292,0.3466531038,0.1434257179,-0.2957483828,-0.2863418162,-0.0682409257,0.1166286021,-0.0230770838,-0.0425324216,-0.0962423906,-0.0311659183,0.2795181572,0.5127834082,0.1344300807,0.1204082519,-0.175738588,0.4893507659,-0.2379411906,-0.0199939087,0.4501178563,0.5878040195,-0.1501936615,-0.1882657558,0.3224348128,0.6150754094,-0.060428191,-0.1997967809,0.0078305146,0.3342184126,-0.1705723852,0.2096038759,-0.0134609006,0.0970799401,0.0249953456,-0.0999708399,0.1413346827,0.0328710712,0.125862807,0.5581205487,0.2509441972,-0.1717868,-0.2090211511,-0.0683216974,0.015021693,0.2349607348,-0.3354085386,-0.0859615803,-0.3126040101,-0.0972639173,-0.271261692,0.0418932363,0.0168843325,0.3037470579,-0.1729043871,0.0122117717,-0.2069065273,0.0238523241,0.1182836518,-0.275769949,0.0307678357,0.3645249605,0.1029346064,0.6225432158,0.1504802406,-0.4848260581,0.1481480747,0.3166593313,-0.0226834137,0.2637164891,0.090539813,0.0571362972,0.1308138818,0.010113827,0.5547730923,-0.0568159334,0.004993719,-0.5158873796,-0.4087069929]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/192","title":"[Question] Create Apache Arrow dataset from raw text file","comments":"> Is there a way to convert a dataset to Apache arrow format (locally\/personal use) & use it before sending it to hugging face?\r\n\r\nSure, to get a dataset in arrow format you can either:\r\n- [load from local files (txt, json, csv)](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=csv#from-local-files)\r\n- OR [load from python data (dict, pandas)](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=csv#from-in-memory-data)\r\n- OR [create your own dataset script](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=csv#using-a-custom-dataset-loading-script)\r\n","body":"Hi guys, I have gathered and preprocessed about 2GB of COVID papers from CORD dataset @ Kggle. I have seen you have a text dataset as \"Crime and punishment\" in Apache arrow format. Do you have any script to do it from a raw txt file (preprocessed as for BERT like) or any guide?\r\nIs the worth of send it to you and add it to the NLP library?\r\nThanks, Manu\r\n","comment_length":58,"text":"[Question] Create Apache Arrow dataset from raw text file\nHi guys, I have gathered and preprocessed about 2GB of COVID papers from CORD dataset @ Kggle. I have seen you have a text dataset as \"Crime and punishment\" in Apache arrow format. Do you have any script to do it from a raw txt file (preprocessed as for BERT like) or any guide?\r\nIs the worth of send it to you and add it to the NLP library?\r\nThanks, Manu\r\n\n> Is there a way to convert a dataset to Apache arrow format (locally\/personal use) & use it before sending it to hugging face?\r\n\r\nSure, to get a dataset in arrow format you can either:\r\n- [load from local files (txt, json, csv)](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=csv#from-local-files)\r\n- OR [load from python data (dict, pandas)](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=csv#from-in-memory-data)\r\n- OR [create your own dataset script](https:\/\/huggingface.co\/nlp\/loading_datasets.html?highlight=csv#using-a-custom-dataset-loading-script)\r\n","embeddings":[-0.116234459,-0.1841169298,-0.0716124773,0.1751199812,-0.3280258477,0.139235884,0.11795903,0.3817427754,0.0022516297,-0.2727223337,0.0569148846,0.0983997434,-0.077237919,0.2192152888,0.3109797537,-0.0580277778,-0.1029317454,0.1872917861,-0.2899804413,0.0341969468,0.2486115843,0.0235500224,-0.0627381131,0.014807988,-0.1860498339,-0.2168583274,-0.1237601191,0.1029290184,-0.1195597574,-0.4056254625,0.0453822389,-0.0250188615,0.5199136138,0.1765504926,-0.0001225578,-0.3012937605,0.1491823643,-0.0089221923,-0.2559587955,-0.3521432877,-0.0911589414,-0.0514178649,0.3022276759,-0.2763149142,-0.1366523504,-0.3666140735,0.0447538495,-0.4632765949,0.8944173455,0.3823668361,0.066881977,-0.0697314069,0.1201731786,0.1746979356,0.2935884893,0.64989537,-0.298520267,0.0173336752,0.2269861549,-0.0276041813,-0.1870692223,0.2103879452,0.0335024111,-0.1601584703,0.4182694554,0.0329427719,-0.3197295964,-0.2667315304,0.1045991704,0.2918638289,0.4435722828,-0.5081265569,-0.0686361194,-0.15882954,0.195936814,-0.4047840536,-0.2023734301,0.5912524462,-0.4560746253,0.4850006402,0.1111562252,-0.3301271498,-0.4546756744,-0.0233673491,0.4130191803,-0.1192310303,-0.1424247622,-0.0342211239,-0.0878192484,0.1139591858,0.1446623802,-0.3402825296,-0.0819694698,0.2702946663,-0.0304175131,-0.1481847614,-0.5688718557,-0.11655467,0.2962391675,0.1948115677,0.3581244051,-0.1492294371,-0.0830274597,-0.2412634045,0.1170790344,-0.0359844863,0.4621579945,0.1229492575,-0.0228784699,-0.1033878848,0.2161062211,-0.1180476174,-0.0714673772,-0.0071603106,-0.1958083659,0.1252120286,0.0682130828,-0.3583876491,-0.063696757,-0.0635823607,-0.3160603642,-0.0648569912,-0.1641439497,0.121300675,0.1500302255,-0.148541376,0.2026017308,0.420989871,0.002455418,-0.2346609831,0.1313377917,-0.0109175872,-0.2892900407,0.34125489,0.151455,0.0410262048,-0.0336779617,-0.065133065,0.0418761894,-0.1866392046,0.2114453018,0.0035033743,0.2554564178,-0.0279187653,0.2255206704,0.0512548424,-0.0815235972,-0.0155559788,-0.1415526271,0.3194224834,-0.1971614808,-0.0013487145,-0.0195522457,-0.0125658587,-0.0712687001,-0.11709214,-0.2966439128,0.1667173505,-0.0553302169,0.3410833478,0.0515456945,0.3041132092,-0.004430512,-0.1568996459,0.0606452562,0.5751587749,-0.2798394859,0.0774808377,-0.0636781305,0.1921245456,-0.1320408732,0.3210464716,-0.2403280884,0.2045165598,-0.1077213213,0.2150873542,0.6738168597,-0.1115691736,0.1638736874,0.5188358426,0.0564497374,-0.2370747477,0.1192140207,-0.1164142415,-0.074343361,-0.0028069427,-0.3140346706,0.5571510792,0.0826089084,0.0182557777,-0.1299570054,-0.215236634,-0.0970101357,0.0046830564,-0.2340568751,-0.014258204,0.0603631474,-0.5500680804,0.4852304161,-0.3260752857,0.4003874362,0.2320333868,0.0025299124,0.0229876433,0.1399732828,-0.0311654154,-0.2683026791,-0.2566067278,-0.3144997358,0.2239765972,-0.4510672092,-0.2488138527,-0.3250751197,-0.1321393251,-0.1072458029,-0.0211220328,-0.0625171959,-0.2079957873,-0.1098496914,0.0530205593,-0.2107385993,0.3642225564,-0.0545293428,0.2473109215,-0.3691832721,0.1250728667,0.1314229667,-0.0102914404,0.233803913,0.176344797,-0.1896404177,0.03401158,0.0070233983,0.0841445401,0.0100615537,0.2041340321,0.4599420428,-0.2706058621,0.1242063195,-0.7130148411,0.2794928551,0.0074256659,0.1846458912,-0.0920473263,-0.4825463891,0.5747972131,-0.1449293643,0.2311107963,-0.1459647417,0.0274425037,-0.1334023923,-0.0777245685,-0.2256112099,0.0628116801,-0.0321381874,0.3259410262,0.3570249677,-0.1237271875,0.0172535423,0.0338917896,-0.2287790626,-0.220150128,0.1503489316,0.2633678019,-0.2448005676,-0.0723638907,-0.032712128,-0.0086683696,0.1141559407,0.1704614758,-0.025074387,0.1204680055,0.2024980336,-0.1406179965,0.3045222759,-0.100370951,0.4107743204,-0.1775361449,-0.1377764791,-0.0128715876,-0.1209045574,-0.2307565957,-0.1614343226,0.1069969758,-0.177055195,-0.0085621923,-0.1658426225,-0.4955016673,-0.4169091582,0.231995523,-0.0679309741,0.2515639365,-0.1629195064,-0.3335672021,0.1284248531,-0.1341255456,-0.1017299369,0.473461926,-0.0070495149,0.0162949879,-0.2771503329,-0.3815431595,-0.1648819149,0.0659755245,0.3686508238,0.4183353484,0.3976896703,-0.120051071,0.1377187222,-0.1911898404,0.0149206603,0.0485848896,-0.0908257589,-0.213777259,-0.1240315586,0.1324923187,-0.5243587494,0.0738317594,-0.1173050627,0.0836823061,-0.2369590402,-0.2871921957,-0.0981979892,0.0131122749,0.1493159533,-0.3908486664,-0.1651644558,-0.1018541083,0.5026175976,0.2510915101,0.1904059052,0.0437579565,0.1060587615,0.1319574416,-0.2570896745,-0.1064995229,0.1840862334,-0.2015497535,0.2977467477,-0.3921264112,-0.5845618248,-0.1108840927,-0.0068856301,0.0341881961,0.2856819928,-0.4618063271,0.4126827717,-0.071103029,0.0711725876,-0.1178861037,-0.0103956386,0.204182893,-0.0252516419,0.0849703252,-0.0204213206,-0.0172816776,0.2986920476,-0.2002993971,0.2878693938,0.4254688323,0.172322616,0.0580608957,0.5829063058,0.3013731539,0.1089275777,0.2840287983,-0.0800581425,0.0536097921,-0.0263136476,0.0135049112,-0.2094151974,0.1249234155,0.1409981698,0.1072886139,0.307251215,-0.0369289592,-0.332125634,-0.2866527438,-0.4147658646,-0.2130818665,0.3038921356,-0.3870820701,0.2348600179,-0.1826101094,-0.1835940629,-0.0711640567,-0.3148885667,0.0509033613,0.2234038115,0.3517571092,0.2050115168,-0.266577214,0.428185761,-0.6509037018,0.3689917922,0.0584447123,-0.1497983336,0.0500975475,-0.0044020992,0.1792496443,0.0758814439,-0.1050073653,0.1694640815,-0.0187861677,-0.0229876526,0.1463411301,-0.518840909,0.3962396085,-0.1713440716,0.0277930032,0.012688634,0.1945351809,-0.4316796362,-0.3492634892,0.112582624,0.2232151181,-0.121543102,-0.3311592042,0.2177020609,-0.0602723137,-0.2420450747,-0.069044821,0.3726757467,0.0339001492,-0.0476140641,-0.1307657659,-0.0470196269,0.1214322075,0.1133872792,-0.2611373067,0.1378375441,-0.2840891182,0.0931993872,0.1468054205,0.4258970916,0.0708256438,0.2117327005,0.1341323406,-0.3408623934,-0.1226655394,-0.2233575284,0.6217304468,0.2724975348,0.152947247,0.0979153514,0.2817002833,-0.217374891,-0.176301375,-0.0922019035,0.0291438606,-0.0699799284,-0.4213859737,-0.3758063614,0.5728223324,0.0364772901,-0.021539174,0.1576511711,0.1421734095,-0.2886929214,0.5798205137,0.2426298857,1.1859325171,-0.2814678848,0.3137070239,-0.0359622277,-0.3606176972,0.5188571215,-0.2017288208,-0.2778556347,-0.4951659441,0.3837973475,-0.0449065156,-0.1855308115,0.3748285174,0.3764266074,-0.0301546808,0.1241888851,0.1666641086,-0.1216058135,0.0595355332,0.475453794,-0.312310189,-0.3469765186,-0.1508463323,-0.0724912956,-0.1424747705,0.2328318357,0.1278476417,-0.4553176165,-0.3689219058,0.0639018789,-0.2599177063,0.1060091332,-0.2166539431,-0.2464063913,-0.2637865543,-0.6389374733,0.4127868712,0.3639934957,0.2206063718,-0.0478414893,-0.2109034508,0.0887874886,-0.3631047904,-0.1849596798,0.1291662753,-0.1481270045,0.030227406,-0.1457622796,-0.1170193404,0.0354309864,0.1157595068,-0.1776612103,-0.0198194589,-0.1916961968,0.4527115822,-0.2664204836,-0.1080395728,0.1988711208,-0.2783823907,0.0872326642,0.0340551175,0.311362952,0.0498485714,-0.075252004,0.2784464955,-0.2262177914,0.0278175883,0.2285946459,0.1652499437,-0.2648728192,0.4430117607,0.4258753359,-0.0970700011,-0.2546084523,0.4642234147,0.5814712048,-0.8045473695,-0.0238830745,-0.0502199233,0.1676585078,0.0249479897,0.1120149866,-0.0131254839,-0.0990689993,0.0351896621,-0.4654941857,-0.3229873478,0.0847475454,-0.0576357618,0.1001603305,0.3351854384,0.1084932014,0.0570534766,0.0697688982,-0.1920591593,0.1754724979,0.651486814,0.1672798544,0.1289995015,0.0443566628,0.0586687736,-0.0103521468,-0.0294238534,0.1542534977,0.0334840864,-0.1404446065,-0.055313427,0.2243406326,-0.1115115508,-0.3583464324,-0.3070650697,-0.3529404998,-0.0504779965,-0.1449858546,-0.0487509891,-0.0624277368,-0.0485144109,0.1196770892,0.091166921,-0.0660633966,-0.518817544,0.1658761054,-0.1233258769,0.21302858,-0.0277884565,0.1678664088,-0.0145818684,-0.0924043879,0.0633602589,0.0195915122,0.3470749557,0.2004653662,0.3311260045,-0.1111036241,-0.0190293975,-0.2046125531,0.3492095768,0.1420442164,0.0439638495,0.0486274436,0.3234450221,0.0612180494,-0.1930480897,0.2455331087,-0.1727201641,-0.188615188,-0.2519422472,0.0988531932,0.0315764733,0.1077814847,-0.4496343434,-0.2361927032,0.1720366478,0.0681495219,-0.2589747906,0.2703657746,0.1438450217,-0.0399567336,-0.007923007,0.2794288099,0.2292510122,-0.0827807486,-0.1410939544,0.1737507582,0.6221066117,0.3592702746,0.290322274,-0.2367406338,0.0642716587,0.3366841674,0.1842474639,0.0393268131,0.2048326731,0.6484080553,0.5068120956,-0.0744930208,0.1196092293,0.2935681641,0.0490094014,0.0308969188,-0.2590864599,-0.0874367505,-0.1244532391,-0.1997783035,-0.2683987021,-0.1559896618,-0.0770117939,0.3239068687,0.1942888498,-0.0283348151,0.0281010009,0.091700457,0.3089791238,-0.1727251858,0.0220066253,0.0015081854,0.01093404,0.0408979021,0.4318575263,0.2473563701,0.0948856473,-0.1277872324,0.228821367,0.026228983,-0.0929477438,0.2152305096,0.1996608675,0.3968430459,0.0358628407,0.1872346848,0.0375478119,0.0675320402,-0.0557286479,0.181980446,-0.1553837657,-0.0786666498,-0.1503335088,-0.0515294224,-0.3914928436,-0.1986079961,0.068926163,-0.1780564934,0.2038507015,0.5880897641,-0.2470193058,-0.0499373525,-0.1666356474,0.0004345702,-0.0319142677,0.4568960071,0.3662316799,0.1416304708,-0.3024466038,-0.3249180019,-0.1399530172,0.0904624611,-0.0144109009,-0.0964346379,-0.0539915897,-0.0555764474,0.2230503559,0.4540471137,0.023300238,0.1714805961,-0.125713855,0.4537191093,-0.2217680812,-0.0101133976,0.3725586832,0.5218807459,-0.1012815908,-0.1328432709,0.2542348504,0.5032443404,-0.0931749195,-0.1573284119,0.0569067709,0.2313571572,-0.250623256,0.20016855,0.0353727117,0.1330993623,-0.0070608892,-0.1503180414,0.0996686295,-0.0279952735,0.0591359958,0.6108753085,0.2494723648,-0.1020064652,-0.2170572281,-0.1134122759,-0.0557875894,0.2859753072,-0.2583235502,-0.1287274212,-0.3091918528,-0.0620906278,-0.2419235259,0.025370799,0.0015789529,0.4026344717,-0.1552554518,0.0643492788,-0.1687609553,-0.0473827384,0.2202804834,-0.3101287782,0.0821347311,0.3282675743,0.1244425252,0.6768016815,0.1087597087,-0.5213273168,0.1987707466,0.2741399705,0.0041087475,0.2193053365,0.0936859027,0.0532278046,0.0733025819,-0.0350291878,0.4586233199,-0.0295333676,-0.0294908527,-0.4290830791,-0.4309493601]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/189","title":"[Question] BERT-style multiple choice formatting","comments":"Hi @sarahwie, can you details this a little more?\r\n\r\nI'm not sure I understand what you refer to and what you mean when you say \"Previously, this was done by passing a list of InputFeatures to the dataloader instead of a list of InputFeature\"","body":"Hello, I am wondering what the equivalent formatting of a dataset should be to allow for multiple-choice answering prediction, BERT-style. Previously, this was done by passing a list of `InputFeatures` to the dataloader instead of a list of `InputFeature`, where `InputFeatures` contained lists of length equal to the number of answer choices in the MCQ instead of single items. I'm a bit confused on what the output of my feature conversion function should be when using `dataset.map()` to ensure similar behavior.\r\n\r\nThanks!","comment_length":44,"text":"[Question] BERT-style multiple choice formatting\nHello, I am wondering what the equivalent formatting of a dataset should be to allow for multiple-choice answering prediction, BERT-style. Previously, this was done by passing a list of `InputFeatures` to the dataloader instead of a list of `InputFeature`, where `InputFeatures` contained lists of length equal to the number of answer choices in the MCQ instead of single items. I'm a bit confused on what the output of my feature conversion function should be when using `dataset.map()` to ensure similar behavior.\r\n\r\nThanks!\nHi @sarahwie, can you details this a little more?\r\n\r\nI'm not sure I understand what you refer to and what you mean when you say \"Previously, this was done by passing a list of InputFeatures to the dataloader instead of a list of InputFeature\"","embeddings":[0.1898331344,-0.4081771076,-0.0266713481,-0.0789325163,0.1755467206,-0.1144005209,0.2933261991,0.2346474379,-0.1131381989,-0.0067010592,-0.0750766918,0.5407450199,-0.2563856542,0.2198833674,0.0012709242,-0.4498939216,0.1627261937,0.1324364245,-0.1213597208,-0.1114255339,-0.2248265594,-0.1842457056,-0.2530254722,-0.1031176224,-0.1576482356,0.0686321408,-0.3666487634,-0.2160354257,-0.3754042387,-0.3336399198,-0.1936650425,0.2823522687,-0.2183113098,-0.036676608,-0.0001251408,-0.4015754759,-0.1773592681,-0.1747857928,-0.1102396101,-0.1179136857,-0.8842166066,-0.0070362641,0.384401083,0.018587172,0.0077913688,0.026045125,0.0269380696,-0.3077802956,0.4810141623,-0.0087722009,0.0720062032,-0.0372220948,0.0368573032,0.0641980544,0.130898267,0.5086032748,0.0186721031,-0.0431833044,0.6300452352,0.0035539258,-0.2057007253,0.3613417745,0.0969025716,0.1061301306,0.3440503478,-0.0262019616,-0.0527589843,-0.0893975794,-0.2308890373,0.3512215912,0.7839892507,-0.0999960825,-0.2693175673,-0.037072368,0.0213868059,0.0687108338,-0.1864119172,0.0146086719,-0.1401160657,-0.0611535646,-0.4859075844,-0.3253764212,-0.3561828732,0.0693899021,-0.2764470279,0.4265608191,0.0081101945,-0.0351795331,-0.0522108674,-0.2725637257,-0.021641355,-0.4392718971,0.3904174268,0.1420880258,0.1063589901,-0.3791531324,-0.1874579042,0.0773358643,0.3710345626,-0.294662565,-0.093178831,-0.0026960964,-0.5204651952,-0.0124727376,0.5711633563,0.0022485678,-0.0457250066,0.0721669048,-0.3457533121,-0.2913954854,-0.2567957044,0.0416471735,0.1252153665,0.2284183502,0.0716412142,0.1218287572,0.0568136945,-0.3215624094,0.0638229027,0.0779039189,-0.3623918593,-0.2042133212,0.0336744189,0.1545162797,-0.0849718377,0.5170226097,-0.1964963078,0.2077666819,0.0878812,-0.3568956256,0.1389586031,-0.1019800454,-0.5272906423,-0.2540571988,0.1949184388,0.287102133,-0.010821675,0.3692451119,0.3783965707,-0.2300265431,0.0706946179,0.0249621775,0.5642062426,-0.1838817745,-0.3527994156,0.3267552257,0.2906134427,0.3297919333,-0.0886824653,0.501160562,-0.2108788788,-0.0140871303,-0.1533492059,0.0175843053,0.2166631371,-0.211597234,0.0411195457,0.3512015343,0.0450342745,-0.335567832,0.1301427782,-0.5346509814,0.0498925187,-0.2956683636,0.0025754108,0.4049502313,-0.4627058506,0.0901223496,-0.1962527037,0.2359769344,0.2237932086,0.2625330985,-0.0753694549,-0.0596093945,-0.0112677189,0.3569813073,0.3155224323,-0.1638092399,-0.3183713555,0.2297862917,0.1608127654,0.1174034551,-0.0363770016,0.2344695628,0.1547364444,0.2690249979,0.051156465,-0.1387481242,-0.3022445738,0.3662523031,-0.0180012248,-0.042903509,0.1209712774,-0.0448402725,-0.0799431279,-0.1422651857,-0.0801847503,0.1147543415,0.2186723202,-0.2054872066,-0.0574702248,-0.4273211658,0.1527346522,0.0290086959,-0.1200327277,-0.1726429462,-0.6358430386,-0.0798971504,-0.0288463924,-0.1870038807,0.2269694358,-0.2735058367,-0.2580794692,0.0582880527,-0.3363609612,0.1047036499,-0.0136203682,0.0200140812,-0.2457858473,-0.2998512685,-0.2771873176,-0.0625879392,0.523280561,0.0186696853,0.0530535057,0.1163390502,0.0966972783,0.0029881662,0.1322368383,0.355776757,0.1922871321,0.2523916662,-0.0313612297,0.2370244563,0.4595884979,-0.0483255535,-0.0386576504,0.1157881245,0.4056983292,-0.2333194911,0.323117435,0.1775736511,0.0954629853,-0.0138599724,-0.8099329472,0.1893574744,-0.0785276443,0.1169799492,-0.1494040787,0.2087834477,-0.1929719895,0.2378272712,-0.0603553131,-0.4787614942,-0.1216315329,0.2656581402,0.1503820866,0.0750132725,-0.4771557748,0.2781422138,0.3794206381,-0.0178738348,0.2281641662,-0.0602244213,-0.017266579,-0.0987433717,0.419705689,-0.2194470763,0.2319580317,0.1562121958,-0.064319253,-0.1458725035,0.0085766567,-0.128225863,-0.0028874232,-0.0382087901,-0.1035607606,0.1028818786,0.1906561106,-0.105451636,0.0022769831,0.073290132,-0.0567633025,-0.0949905738,-0.1219476461,0.0056225667,-0.5228880048,-0.0163592119,-0.1550268829,-0.5842822194,-0.1949106604,-0.1818955988,0.2008174211,-0.196906507,-0.0877336711,0.0063602412,0.0872827098,0.3509713113,0.1391626,0.0905076191,0.0792290345,-0.0174420942,0.0698145255,0.0482323579,0.0040308577,0.0327776819,0.1947336942,0.1055906713,0.0115956618,-0.353723675,0.015375359,0.1942841113,-0.0494602025,0.0103383223,0.3358051181,0.0873691738,-0.4954452217,0.139918074,0.1405892521,0.3530600667,-0.0523260869,0.0308349282,0.3473846614,0.0177804008,0.0939540118,-0.3488003016,-0.1854479164,0.0460721403,0.0916791409,-0.0240837857,0.1084490493,-0.2396626472,0.1782774627,-0.1272112429,-0.3487858474,-0.2479266226,-0.2057134509,0.1175255775,0.0489020571,-0.1411753446,-0.0173885059,-0.217268154,0.1409626454,-0.138920784,0.1800364107,0.2704903781,-0.2778686583,-0.0454747006,0.1224455088,0.0982366428,0.4216068685,0.302885741,0.3918448091,0.0813359991,-0.0775967836,0.1031870842,0.1100871935,0.3211499751,0.0097876936,0.1452916414,0.04953501,-0.1308119446,0.7850805521,0.4763058126,0.1447903216,0.1125237569,0.0027535111,-0.1625672132,-0.1379141957,-0.1045337245,0.2698309422,-0.0767841339,0.1905375868,0.457778722,0.0880915374,-0.0634215325,0.0359540321,0.2923195958,-0.2363472134,-0.2863628268,0.4252736866,-0.1718090177,0.2505713701,0.0141778477,-0.1171555519,-0.2502305806,-0.1201371178,-0.0614109337,-0.1116428673,0.3435658514,-0.3986702263,-0.0975905582,-0.2561987638,-0.1570489854,0.2989191711,0.4773337245,0.1038342565,-0.0135086905,-0.0769287646,0.1650286317,0.34249264,0.5172498822,-0.0393498726,-0.1162255853,-0.0920813382,0.0562989227,0.1732502431,-0.036810603,-0.6540709138,0.3220745325,-0.2926156819,0.1334904581,-0.2254888117,-0.3519280255,0.1718120873,0.1453809142,-0.1047327891,-0.2261109203,0.13815777,0.5807728767,0.0008980384,0.2521329522,0.2287891954,-0.1592706591,0.1919574887,-0.0932563171,0.1850654334,-0.006664217,0.3489797711,0.2143312097,0.0644307062,0.1648572236,-0.3918598294,0.229424715,0.1500616372,0.1957376897,0.2030125111,-0.6772521138,-0.060860794,0.084965229,-0.1169665009,0.6384949684,0.3110846281,0.2727365792,-0.0807491392,0.0904628262,-0.0148605909,-0.1867106259,-0.0781757757,0.3036339879,0.0814067498,-0.2078103721,-0.4161690772,0.38397488,-0.1178029925,-0.2680466175,-0.136865586,0.5329193473,0.0063751433,0.0325256288,0.41089499,0.8881412148,-0.1826613694,0.3090893924,0.1333591193,0.0751583576,0.3711303473,-0.1337946802,0.2235966474,-0.2209894061,0.1054657847,-0.1843819469,-0.1960986257,-0.0003092639,0.5590391755,-0.6933900118,0.2738440931,-0.1549046338,0.159049198,0.1415614635,-0.0622472875,0.3114886284,-0.1365657449,-0.1344510317,-0.0127958255,-0.2086643279,-0.1317790002,-0.1540084928,-0.0484585725,0.0319659151,0.0056092078,-0.314815551,-0.0654661879,-0.5026328564,-0.2293139398,0.0704678297,-0.132711634,0.4928192198,0.4730234742,0.1101296172,0.126377821,-0.0846125707,-0.1857496649,0.1338843256,0.2864260077,-0.2314484417,0.1073825955,0.505435586,-0.0287612826,-0.3097273707,0.2361543775,-0.1268925667,-0.089211829,-0.485024482,0.1137208939,0.7107091546,-0.5381381512,-0.4216567576,-0.0315876417,0.1221126989,0.0282334704,-0.0889573172,0.2638349831,0.110114783,0.5779927373,0.2360899448,0.1575015485,0.2445419878,-0.1331834197,0.142686367,-0.101616405,0.3163881898,0.2070098817,-0.1430161446,-0.1376813352,0.0818388164,0.2291909307,-0.1033827737,0.1376063824,-0.2527720034,-0.0819374621,-0.0388871916,0.3396040201,0.3437861204,0.3037846982,-0.0093047926,0.0497466326,-0.1195895299,0.2682490051,-0.2267332524,0.1715279371,0.197195068,-0.1525156647,0.0667894706,0.2607313693,-0.1768621206,-0.0941506177,0.1146863475,0.3973581791,-0.0025263282,-0.2013353705,-0.2250721753,-0.0516553484,-0.0407256857,-0.2295282185,-0.2201394439,-0.1214470193,-0.1096406281,0.1389662921,0.0955216736,-0.0764570087,-0.1243966147,-0.1279193461,-0.0087509276,-0.124936454,-0.0020424989,-0.0699351206,-0.1673056781,0.3306976855,0.138597548,-0.3417626619,-0.194118157,0.1241503507,0.327349782,-0.0463120043,0.1761830449,0.1268651634,0.0688236505,-0.2199276984,-0.0735193193,0.4398283064,0.2794688344,0.0344990604,0.0281555727,0.0986639634,0.1635745168,0.0910554752,0.29285115,0.0589943305,0.0481073782,0.0916692913,0.0448624529,0.0873927474,-0.1086841822,-0.2180770338,0.5362927318,0.1088522449,0.1672738343,0.3539822698,0.0586780645,0.0175231025,0.3759634793,-0.0288945287,-0.3581575751,0.0637383237,0.0148657616,0.6164770126,-0.096758604,0.0600409992,0.1579789221,0.2112783194,0.0631434917,-0.0031044625,0.0737843066,0.2152060121,0.1195627451,-0.1352908164,0.4986390471,-0.0163091924,0.2160300761,0.2218584567,0.1069494188,0.4482232928,0.3513821363,0.1383229345,-0.2611649036,-0.5070870519,-0.0058212024,0.0829524696,-0.1492794752,-0.1876381934,-0.0230208486,-0.2203644514,-0.3409635723,-0.0308203306,-0.6939738393,-0.1361892968,0.4685730636,0.3233006299,0.5033949018,-0.1203055307,-0.049626343,0.1058638543,0.2920263708,-0.1235241666,0.1855368465,0.0337446481,0.027988527,0.1073462218,0.3419787586,-0.0009041918,0.216677025,-0.1818514913,-0.0319618806,-0.3361879587,-0.0869803503,0.2088313103,-0.2359446883,0.222337231,-0.0785492584,0.1523193121,0.0113579929,0.051598344,0.15602988,0.0699555427,0.0065101795,-0.2356919497,-0.5836459994,0.216498524,-0.2501597106,0.0859224498,-0.0313992053,-0.6951825619,-0.0269292966,0.056830328,-0.1017039865,-0.0800097659,0.1481975764,-0.012619474,0.2639656365,-0.0298058763,0.361697495,0.0144492108,-0.3209574521,-0.1784156114,-0.2791795731,-0.1117577925,0.2492114604,-0.1221849397,0.0724093765,-0.021377448,0.0543237142,0.3683216274,0.0556712151,-0.5797137022,-0.259784013,0.4736153781,-0.1560892165,0.1644121706,0.2888821959,0.2590017617,-0.0282119755,-0.0676181614,0.2782746851,0.3107702434,-0.1515821666,-0.2304982543,0.0107389167,-0.2494552881,-0.0599178225,-0.2276154906,0.3659666181,0.276381284,0.1070736274,-0.2032271028,0.2864720821,0.4057309628,0.1192763895,0.0943275169,0.1111679971,0.6024984717,-0.1494071931,0.2883358002,-0.0545483232,0.0235118475,0.098096177,-0.1638507247,-0.6133044362,-0.0224438012,-0.0239446294,-0.3201176226,0.1384833455,0.4174029827,-0.1201283708,0.1861409545,-0.2441980392,0.2617794275,0.1363038868,-0.2238335311,-0.349663049,-0.3081929386,-0.0459221005,-0.2481650859,-0.0147775058,-0.7189196944,-0.2410822809,0.2345633358,-0.1155320778,-0.1438892931,0.0421162471,-0.0393075012,-0.1662568748,-0.3629511893,0.6250204444,0.0369418524,0.1013304442,-0.3119326532,-0.4519559741]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/189","title":"[Question] BERT-style multiple choice formatting","comments":"I think I've resolved it. For others' reference: to convert from using the [`MultipleChoiceDataset` class](https:\/\/github.com\/huggingface\/transformers\/blob\/a34a9896ac2a4a33ff9cd805c76eed914c8d8965\/examples\/multiple-choice\/utils_multiple_choice.py#L82)\/[`run_multiple_choice.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/a34a9896ac2a4a33ff9cd805c76eed914c8d8965\/examples\/multiple-choice\/run_multiple_choice.py) script in Huggingface Transformers, I've done the following for hellaswag:\r\n\r\n1. converted the `convert_examples_to_features()` function to only take one input and return a dictionary rather than a list:\r\n```\r\ndef convert_examples_to_features(example, tokenizer, max_length):\r\n\r\n choices_inputs = defaultdict(list)\r\n for ending_idx, ending in enumerate(example['endings']['ending']):\r\n text_a = example['ctx']\r\n text_b = ending\r\n\r\n inputs = tokenizer.encode_plus(\r\n text_a,\r\n text_b,\r\n add_special_tokens=True,\r\n max_length=max_length,\r\n pad_to_max_length=True,\r\n return_overflowing_tokens=True,\r\n )\r\n if \"num_truncated_tokens\" in inputs and inputs[\"num_truncated_tokens\"] > 0:\r\n logger.info(\r\n \"Attention! you are cropping tokens (swag task is ok). \"\r\n \"If you are training ARC and RACE and you are poping question + options,\"\r\n \"you need to try to use a bigger max seq length!\"\r\n )\r\n\r\n for key in inputs:\r\n choices_inputs[key].append(inputs[key])\r\n \r\n choices_inputs['label'] = int(example['label'])\r\n\r\n return choices_inputs\r\n```\r\n2. apply this directly (instance-wise) to dataset, convert dataset to torch tensors. Dataset is then ready to be passed to `Trainer` instance.\r\n\r\n```\r\ndataset['train'] = dataset['train'].map(lambda x: convert_examples_to_features(x, tokenizer, max_length), batched=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'label']\r\ndataset['train'].set_format(type='torch', columns=columns)\r\n```","body":"Hello, I am wondering what the equivalent formatting of a dataset should be to allow for multiple-choice answering prediction, BERT-style. Previously, this was done by passing a list of `InputFeatures` to the dataloader instead of a list of `InputFeature`, where `InputFeatures` contained lists of length equal to the number of answer choices in the MCQ instead of single items. I'm a bit confused on what the output of my feature conversion function should be when using `dataset.map()` to ensure similar behavior.\r\n\r\nThanks!","comment_length":168,"text":"[Question] BERT-style multiple choice formatting\nHello, I am wondering what the equivalent formatting of a dataset should be to allow for multiple-choice answering prediction, BERT-style. Previously, this was done by passing a list of `InputFeatures` to the dataloader instead of a list of `InputFeature`, where `InputFeatures` contained lists of length equal to the number of answer choices in the MCQ instead of single items. I'm a bit confused on what the output of my feature conversion function should be when using `dataset.map()` to ensure similar behavior.\r\n\r\nThanks!\nI think I've resolved it. For others' reference: to convert from using the [`MultipleChoiceDataset` class](https:\/\/github.com\/huggingface\/transformers\/blob\/a34a9896ac2a4a33ff9cd805c76eed914c8d8965\/examples\/multiple-choice\/utils_multiple_choice.py#L82)\/[`run_multiple_choice.py`](https:\/\/github.com\/huggingface\/transformers\/blob\/a34a9896ac2a4a33ff9cd805c76eed914c8d8965\/examples\/multiple-choice\/run_multiple_choice.py) script in Huggingface Transformers, I've done the following for hellaswag:\r\n\r\n1. converted the `convert_examples_to_features()` function to only take one input and return a dictionary rather than a list:\r\n```\r\ndef convert_examples_to_features(example, tokenizer, max_length):\r\n\r\n choices_inputs = defaultdict(list)\r\n for ending_idx, ending in enumerate(example['endings']['ending']):\r\n text_a = example['ctx']\r\n text_b = ending\r\n\r\n inputs = tokenizer.encode_plus(\r\n text_a,\r\n text_b,\r\n add_special_tokens=True,\r\n max_length=max_length,\r\n pad_to_max_length=True,\r\n return_overflowing_tokens=True,\r\n )\r\n if \"num_truncated_tokens\" in inputs and inputs[\"num_truncated_tokens\"] > 0:\r\n logger.info(\r\n \"Attention! you are cropping tokens (swag task is ok). \"\r\n \"If you are training ARC and RACE and you are poping question + options,\"\r\n \"you need to try to use a bigger max seq length!\"\r\n )\r\n\r\n for key in inputs:\r\n choices_inputs[key].append(inputs[key])\r\n \r\n choices_inputs['label'] = int(example['label'])\r\n\r\n return choices_inputs\r\n```\r\n2. apply this directly (instance-wise) to dataset, convert dataset to torch tensors. Dataset is then ready to be passed to `Trainer` instance.\r\n\r\n```\r\ndataset['train'] = dataset['train'].map(lambda x: convert_examples_to_features(x, tokenizer, max_length), batched=False)\r\ncolumns = ['input_ids', 'token_type_ids', 'attention_mask', 'label']\r\ndataset['train'].set_format(type='torch', columns=columns)\r\n```","embeddings":[0.1992308199,-0.5757998228,0.0108198822,-0.0699366182,0.2859191298,-0.1536672115,0.2446017712,0.2416757196,-0.0098712537,0.0180189833,-0.1549074948,0.4558004439,-0.2713595927,0.0753545314,0.1255868077,-0.3799526095,0.1526270956,0.0689935386,-0.201622799,-0.0675194636,-0.1130087897,0.0281150993,-0.1665605009,-0.0206472613,-0.351328373,0.0464573093,-0.3601536155,-0.0656297877,-0.302508533,-0.3981375098,-0.1327254027,0.1980736852,-0.256768018,0.1135846227,-0.0001305897,-0.3111949265,-0.2120848447,-0.2144840509,-0.0953421518,-0.0315434635,-0.7208180428,-0.0117110601,0.3938957155,0.1077938452,-0.0250609145,-0.0244077258,0.0035534585,-0.1321900189,0.6255206466,-0.0203529987,0.0092290873,0.0400460251,0.0817500129,0.0990012288,0.1613848209,0.5654866695,-0.0732178316,0.0807715878,0.5695483088,-0.0895334482,-0.2318441719,0.5064966083,0.0316469595,0.0750548765,0.2941819429,0.1139432862,-0.0156049412,-0.0805158168,-0.2227819115,0.4199120402,0.5713579655,-0.3070062101,-0.4449506998,-0.2811709642,0.077552855,-0.0533092804,-0.185295701,0.0442555472,-0.246355921,0.0007777097,-0.3870469928,-0.143476069,-0.2480071783,0.0358369835,-0.2692357004,0.5118787289,-0.050744392,0.0218754392,-0.0064533297,-0.2313463092,-0.1519096494,-0.3797215819,0.3062953651,0.2080069631,-0.0442397892,-0.3968067765,-0.1380621046,-0.0766860098,0.450856626,-0.2433416694,-0.0085721314,0.0231091548,-0.4934567213,-0.0227418467,0.6486942768,0.0418396853,-0.0119943768,-0.0025542285,-0.2303538024,-0.0869294554,-0.1043043584,0.0629381984,0.1495052874,0.1489631534,0.1232819036,0.1073075309,0.0246130433,-0.2623034418,-0.0393709689,0.0123500591,-0.3285461068,-0.138445273,0.1397070438,0.2474577129,-0.1276465207,0.5036812425,-0.1274145097,0.3086260855,-0.0092506576,-0.0996710584,0.0914359689,-0.158286944,-0.403457433,-0.1063968614,0.2117799073,0.3301987946,0.0904372707,0.4027234018,0.5290731192,-0.2610939145,0.021614287,0.0145388981,0.4841366112,0.022282023,-0.3328678906,0.4092332423,0.3574821353,0.2704078257,-0.1167816669,0.3718445003,-0.1770601422,-0.0374801345,-0.0164656676,-0.0591561496,0.1264020801,-0.1688946337,-0.1565997005,0.3634795547,0.1241389737,-0.1976222843,0.0890343264,-0.5058836341,-0.0126779247,-0.1831261963,0.1596045047,0.5324590206,-0.2871998549,-0.1431741863,-0.0231243819,0.1454491168,0.0770599991,0.36676386,-0.0832993761,-0.0956673846,-0.0285424031,0.4020661712,0.2019063383,-0.2790749669,-0.2886228263,0.1832921058,0.3203335404,0.3041915596,-0.1232166886,0.1302313954,0.0209683999,0.2288772762,0.1084147096,-0.1621450186,-0.1762799174,0.4218750596,-0.0592051782,-0.1318949908,0.0792518482,-0.0365051329,-0.1087161303,-0.2211264819,-0.2119843811,0.1571849883,0.3730957508,-0.1357677281,-0.1573298723,-0.1915747821,0.1454184055,-0.0841141045,-0.032169506,-0.1841538846,-0.6730769873,0.0318732187,-0.0588189177,-0.0139771802,-0.0235597268,-0.2371215969,-0.3311184049,0.1513948292,-0.3594202101,-0.0612323061,-0.0667840391,-0.0110397208,-0.0972960815,-0.2155428827,-0.2320966423,0.052623447,0.4880287647,0.1118904948,-0.0119324317,0.1936452538,0.1937323362,-0.1239278316,0.0630680695,0.3638551533,0.2385548502,0.1898649931,0.0156246871,0.2106879056,0.5333706141,-0.0716076717,-0.1271521747,0.236597091,0.4539563954,-0.1962525696,0.1511336416,0.2068120688,0.1926598549,-0.0478395596,-0.7790426016,0.2532613277,0.013714348,0.2599429786,-0.0932526216,0.1928777993,-0.0752389878,0.3315351009,-0.0595252439,-0.3936669827,-0.2014767826,0.1246120259,0.1480499357,-0.0295605622,-0.4988960326,0.3210488856,0.3738487959,-0.0521548726,0.2573857903,-0.0576330759,-0.1082976982,-0.1585582048,0.3733247221,-0.3535532951,0.2838003337,0.0691675022,-0.0836794227,-0.1215839759,-0.0424232483,-0.0347593985,0.0271577127,0.0674328208,-0.0719189718,0.1330576986,0.1848294139,-0.220438689,-0.0803161114,-0.005192922,-0.0877737924,-0.0895496309,-0.2064072192,0.0967238694,-0.5124549866,0.103811726,-0.2801758051,-0.5547010303,-0.3657353222,-0.2290549576,0.2649710178,-0.0899189562,-0.16554901,0.0586052202,0.1167238131,0.0993118212,0.1955944002,-0.0060084872,-0.0894494876,0.0460132658,0.1018408909,-0.0513067134,-0.0476841703,-0.0478527918,0.0376473926,0.1288282424,-0.1131999642,-0.4533536136,-0.0084238099,0.311193049,-0.1108270064,0.1549980193,0.3773562312,0.1029158831,-0.4873541296,0.0487930737,0.2122695744,0.2985990942,-0.0036083993,0.1128151342,0.3003280461,-0.0229476374,-0.0123437988,-0.1915768087,-0.1320614815,-0.003898083,0.1001858413,0.0269555002,0.1524860412,-0.1670546681,0.2679468095,-0.0359983742,-0.3355773389,-0.2415347695,-0.1400486678,0.0869605243,0.1467295587,-0.0843720436,-0.0522504188,-0.283863157,0.0482754037,-0.0188674629,0.1145023406,0.1831578165,-0.3538354635,-0.087027885,-0.0628895089,0.1746334583,0.4848902524,0.2857912183,0.2647238374,0.1194117144,-0.0561731532,-0.0061280429,0.163911581,0.3485115767,0.0466543473,0.004912517,-0.0949426815,0.0010067495,1.0838108063,0.4843838513,0.2481923103,-0.0036791754,-0.1181708276,-0.1263030767,-0.1567159295,-0.0900336877,0.1181086153,-0.073675476,0.3027005196,0.4651427567,0.2442463189,0.0120604401,0.0534656979,0.3858361244,-0.162182346,-0.4523173273,0.3965883255,-0.0745390058,0.1458475888,0.0180842672,0.0641306937,-0.1832008809,-0.1223577634,-0.0762117356,0.0476063378,0.3101421595,-0.1923480183,-0.040282011,-0.2107439339,-0.2597437501,0.2814572752,0.5223881006,0.1269291043,-0.0026259418,-0.1497015208,0.2083980292,0.2371642888,0.63562572,0.1785300523,-0.2098157555,-0.0090589337,-0.1218135729,-0.0105070053,-0.1275863945,-0.6342933774,0.1838950813,-0.2725805938,0.2263436466,-0.3975050449,-0.4662691057,-0.0781304836,0.1175883412,-0.1353406012,-0.1076996177,-0.0565055944,0.4158498645,-0.1580485851,0.3252096474,0.3609256744,-0.1426489651,0.171275869,-0.0645057634,0.2282080054,-0.0542897806,0.3435719907,0.332724452,0.1330375373,0.0988577902,-0.3660643697,0.2214113623,0.0841441378,-0.0092256432,0.3217580914,-0.5721676946,-0.2904270589,0.1490000039,-0.0495819859,0.4967508018,0.4423178732,0.2168344259,-0.0509799421,0.2642028034,0.079705447,-0.1399865896,0.1115491241,0.2770271599,0.0872878507,-0.1706185192,-0.3983028233,0.4744043648,-0.064919427,-0.2802017927,0.0136902286,0.5781549811,-0.0878418535,-0.0790271387,0.2935150266,0.9808763862,-0.2313354164,0.3998239636,0.4019371271,0.0262567997,0.3194388151,-0.2821464837,0.2668994665,-0.2296040952,0.1147755012,-0.1268308312,-0.1896536052,0.0910701156,0.4854975939,-0.6954209208,0.40398857,-0.0632095411,0.0795763433,0.0671938211,-0.0208081435,0.1637315154,-0.2564381063,-0.2649288177,-0.0667251423,-0.2586742938,0.0004014663,-0.1023565382,-0.0611613877,-0.0066718888,-0.1209413037,-0.3932929933,-0.0147965327,-0.4553059936,-0.1754757613,0.29111588,-0.1525873542,0.3953567445,0.4345214069,0.0012727951,0.1455685645,-0.1255717576,-0.0287105348,-0.0840090662,0.4058328569,-0.2616159916,0.091705732,0.566505909,-0.0066859289,-0.3421804905,0.1574377418,-0.080924049,-0.1010132283,-0.4346273243,0.1806266159,0.6312642097,-0.5566821098,-0.2709474564,0.1057049185,0.1123408824,-0.0415938795,-0.1443416327,0.390412569,-0.0377634987,0.5050051808,0.1526936442,0.0753725246,0.1519210041,-0.0246065985,0.0446850285,-0.2201678157,0.3574769199,0.2970623672,-0.1157664433,-0.1050685123,-0.0570651777,0.2073895931,-0.1948967725,0.3421349227,-0.2141124457,-0.2222320586,-0.0952239186,0.300144583,0.1865332425,0.1520579308,0.0417337865,-0.0323386975,-0.1166564226,0.3286195397,-0.1433885694,0.2544212639,0.1908645779,-0.031950742,0.0283952132,0.1957785636,-0.1289579719,-0.0874994248,0.0082763173,0.3492435515,-0.1132294014,-0.2655161619,-0.0298206154,0.0035730768,-0.0389758386,-0.0955162793,-0.2530175745,-0.0689895079,-0.2502748072,0.1610959768,0.0479435213,-0.1259590536,-0.0265385807,0.0312609747,-0.0231834687,-0.2388320863,0.1918046474,-0.0868987516,-0.1408482939,0.3501248658,0.0075803748,-0.2158453017,-0.1304399818,0.0271077491,0.1823078841,-0.0359775722,0.3349059224,0.1529586613,-0.0354759619,-0.2368503362,0.0391066596,0.4817046225,0.1581270844,-0.0139211528,0.073547706,0.0815035924,0.0789054111,0.0178396069,0.2860041559,0.0081111239,0.0782606378,0.046649795,0.1451670378,-0.0062226206,-0.1072502211,-0.2913448513,0.4584798813,0.1048625782,0.200246349,0.4067823887,0.0172926914,-0.0032191298,0.395906508,-0.0531290323,-0.2769778967,0.0695279315,0.0112389615,0.6116735339,-0.1678520143,0.1247135177,0.075269118,0.2117466033,0.1798260361,0.0357596762,-0.1800639927,0.208125636,-0.1359257102,-0.1532132775,0.5411538482,-0.1362390071,0.0149111114,-0.0067873024,0.0219409596,0.4590047896,0.2312012017,0.2578005493,-0.3163488805,-0.3504881561,0.0007419722,0.1056132615,-0.1132547036,-0.1969221234,0.2010505497,-0.110953629,-0.3095963299,-0.063642852,-0.5783091187,-0.3299179375,0.4866971076,0.2248498201,0.5121278763,-0.1676061451,-0.1180017143,0.1449580789,0.3679752648,-0.1525890082,0.1540867239,0.3022060692,-0.0390836969,0.0186225381,0.2026045918,0.0895345882,0.1358354539,-0.2048610747,-0.014349171,-0.3012128174,-0.0274298247,0.1880570203,-0.1548428386,0.1841393858,0.0267135818,0.223192215,-0.0609474219,0.0680275708,-0.1177255586,0.0317081697,0.1584727913,-0.1828480065,-0.3853935599,-0.0131455474,-0.1711121947,0.0476968884,-0.0006276328,-0.7032936811,-0.0632745698,0.0726504028,-0.0693798289,-0.0340656303,0.0992378891,-0.0431205332,0.2042517215,0.1544468999,0.2449753433,0.1417860389,-0.2718691826,0.1209633648,-0.2757837772,-0.1867609769,0.2845396399,-0.2318628281,0.0397874266,0.0207818802,0.1018791348,0.3410893083,0.0682227612,-0.5816962123,-0.1138581857,0.4657229781,-0.2012690604,0.1362157613,0.1844778061,0.26688236,0.1062909067,-0.1284307837,0.2124747336,0.3538291752,-0.2119625062,-0.1707962453,-0.071302183,-0.3326971829,-0.0588465333,-0.2403443903,0.5293842554,0.2018614411,0.122413896,-0.2680345178,0.1406505853,0.2860545516,0.0567033477,-0.0289777443,0.0273178536,0.7622184753,-0.0932823569,0.1050689816,-0.111196138,0.215980038,0.1279983521,-0.3299755454,-0.4948566854,0.0894029588,-0.0252885856,-0.2356467694,0.1079358235,0.483458221,-0.1225612685,0.1367388219,-0.267082274,0.1285667419,0.319991529,-0.2631484568,-0.3871836662,-0.399974376,-0.071160391,-0.0355825424,-0.0329202972,-0.8910264969,-0.1721324325,0.1124913618,-0.0381589793,-0.2536055148,0.0852270052,-0.0223568585,-0.2716103792,-0.3777169287,0.6619983315,0.0568977185,0.0339616947,-0.2267159522,-0.5398957133]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/188","title":"When will the remaining math_dataset modules be added as dataset objects","comments":"On a similar note it would be nice to differentiate between train-easy, train-medium, and train-hard","body":"Currently only the algebra_linear_1d is supported. Is there a timeline for making the other modules supported. If no timeline is established, how can I help?","comment_length":15,"text":"When will the remaining math_dataset modules be added as dataset objects\nCurrently only the algebra_linear_1d is supported. Is there a timeline for making the other modules supported. If no timeline is established, how can I help?\nOn a similar note it would be nice to differentiate between train-easy, train-medium, and train-hard","embeddings":[-0.3644658029,-0.2648747265,-0.2903042138,0.194593057,0.1727906764,0.124680981,0.1975679547,-0.0213291701,0.1545403749,0.0297257509,0.1898474097,0.1180286184,-0.4256516993,0.0820173845,0.1325089633,-0.4118052721,0.4158248007,-0.1826136857,-0.3111876249,-0.300373286,-0.1286771446,0.1022927538,-0.3951662481,-0.1257083267,-0.433021158,-0.0268127173,-0.20587264,-0.1951666176,-0.5059969425,-0.3827085197,0.0449119955,0.2684865892,0.2143680155,0.2264016867,-0.0001157814,-0.2044546902,0.1789987981,0.0488550477,-0.3329448104,0.017687032,-0.4842326045,-0.3468343318,0.3682309985,0.1491223723,0.1066027358,0.0573814958,0.0222014152,-0.4944730699,0.0617088601,0.3510648608,0.1423078924,0.1565252692,0.1572712362,-0.311665386,0.2181363553,0.0581543222,-0.1421015114,0.1875703186,0.4480463564,0.2472400516,0.1792710423,-0.0406296775,0.1944298893,0.0675512701,0.2683117688,0.0195669215,0.1858889014,-0.3522984684,-0.4321002364,0.2006406337,0.855134666,-0.1497007906,-0.3018468022,-0.2069049627,-0.0560239218,-0.2677705288,0.0007724975,-0.0302981045,0.2882270813,-0.0171022229,0.0047743865,-0.3053902984,-0.2625767589,0.3727519512,0.1801438332,0.4874795079,0.1238787547,0.0285503063,0.1972545981,-0.0654844865,0.4366559684,-0.0767939687,0.2811885476,0.2233610898,-0.1882552058,-0.289985925,-0.2867549956,0.1558763534,0.0279112831,-0.1612362564,-0.1736642718,0.2732505202,-0.1394060254,0.2448565215,0.1795983911,-0.1990219802,-0.0525474325,-0.0622730069,0.4067674279,-0.0307018235,-0.0700924471,0.274120301,-0.2092526704,-0.0341614299,0.0240539089,0.2047156692,-0.12432082,-0.0112075144,0.068880558,-0.4274086058,-0.0658677891,-0.4061034322,-0.1053384095,-0.0453025922,-0.2480791211,0.17754215,-0.197230652,0.0915088877,-0.1641460508,-0.1135701761,-0.0339112207,0.1255479753,-0.225080356,0.0014857913,0.1356467754,0.1127066389,-0.0598024204,-0.0032278465,0.1942936033,0.0249577276,0.0616857298,0.0136263091,-0.1491024792,0.1039120331,-0.3707094789,0.0684238821,-0.2370975316,0.4070960283,-0.1071934253,-0.2151251435,0.0041973484,-0.2066099644,0.0195777696,0.1919053048,0.0561825819,-0.3518363237,-0.1317759752,0.4196633101,-0.3962431848,0.2499533594,-0.0864939764,-0.1797175407,-0.0574861579,-0.2712155879,-0.0139527516,0.0848845467,-0.2083443999,-0.0321557373,-0.1134350523,-0.1710197479,-0.2814630866,-0.0236250982,-0.1127583012,0.1307081878,-0.1476295888,-0.21517241,0.4675810039,-0.4482759535,-0.0352573022,-0.0549371503,0.2837302983,-0.2728815973,0.0651676878,0.2512279451,-0.0025544427,-0.0611595735,-0.0300420616,0.3217568994,-0.1207588241,-0.3328774273,-0.0519570075,-0.1068715528,0.2400470227,0.1276910454,0.5425010324,0.0545941927,0.2041015327,0.2646133304,0.0812536329,0.2596727014,0.0581661612,0.1924337894,0.2933646142,-0.0380582735,0.2833443284,-0.5765680671,-0.3563037515,0.0752296373,0.2790786922,0.1422836334,0.2527941465,0.0201495383,-0.1220889613,0.0909124166,-0.1952799857,0.3942016065,0.0985529199,0.0160426628,0.2292000502,-0.1707041413,-0.4435397685,0.245998621,-0.0655789152,0.4079241455,-0.1897240579,0.2357727438,0.0182033461,-0.2050147653,0.2195056975,0.323605597,0.1670200229,-0.4342312217,0.1908692569,0.2243928462,0.3504455984,0.0343213901,0.0825936124,0.4194789231,0.4189201593,0.0073955129,0.2176106572,-0.2969796956,-0.0859405249,-0.0641399845,0.1655092984,0.2884207666,0.1193301454,-0.148262769,0.2614082992,0.0543346517,0.0208188798,0.3195665479,-0.0099433148,-0.3579170704,0.1247719526,0.0036714228,-0.0515403152,0.0315881446,-0.3361847997,0.0075408211,0.6508182883,-0.1386559755,0.1893228143,-0.0347049348,-0.1344849169,0.00913384,0.0954187959,0.1745010912,0.1093635112,0.1829327941,0.3274006546,-0.1745285541,-0.3105501235,-0.0807159021,0.0237666536,0.2065743953,0.1597675532,0.0170352571,-0.1185473874,-0.0630851164,-0.1452102065,-0.0490110032,-0.0164270755,0.3006424308,0.0512159728,-0.19254601,-0.529258728,-0.0996462479,-0.1366803348,-0.0864412412,-0.2118092179,-0.22282511,0.1816733479,-0.118044436,0.2812513709,0.1455670297,-0.053244736,0.1679355502,-0.3308936357,0.4408709407,-0.1783830225,0.0381352007,-0.2528570592,0.1072119325,0.2094201297,-0.2491567433,0.3661607802,-0.1580459476,0.1763253808,-0.0128310155,-0.4075423479,0.2864955664,-0.1042822897,0.04869752,-0.0581110567,-0.1230604127,0.0803887099,0.221511513,0.1215499043,-0.1631312519,-0.14509359,-0.0787131563,-0.1152203158,-0.0779212713,0.0838137865,-0.471778959,-0.6514961123,-0.2323558927,0.0939280614,0.1174430847,0.0401217118,0.0956270471,0.0859239474,0.016276747,-0.1322245449,0.3119650781,0.1842645407,-0.1823063642,0.2801010609,-0.1915035248,-0.2253230065,0.2531846762,-0.1177079007,0.216479525,0.3987162709,-0.1462939233,-0.1960320175,-0.051658202,0.0660405383,0.1206308082,-0.2179233581,0.2783581614,0.0918467864,0.0392639451,-0.1322747469,-0.221598804,-0.0312622078,0.2855889797,0.0005618947,-0.0090448083,0.266291678,-0.1801258028,0.4601090848,0.0281339921,-0.2381316125,0.1236537322,0.2463115156,0.0387105644,-0.0069235386,-0.0046405229,0.0074639958,0.006111904,0.1364608854,0.0765460432,0.0994092375,0.1046420112,-0.3060671091,0.4501691759,-0.274268955,-0.0165848602,-0.015086744,-0.538131237,0.5747333169,0.003132493,-0.1223154962,-0.1701026857,0.0826035962,0.1531162858,0.3635968268,0.2249917388,-0.1379969269,-0.6526823044,-0.2058461457,0.0367563963,0.3435265422,0.0096083693,0.2784567177,-0.403142333,0.0548211858,0.188397333,0.2267924547,0.118588753,-0.5769719481,-0.4399369061,0.3433743119,-0.0160034224,-0.22277686,-0.0615202785,-0.3217219412,0.1993919462,0.0049044071,0.619846642,-0.1000848785,-0.4446305931,0.2296218425,0.1828424782,0.0868440717,-0.1342187077,0.3056707978,0.3167255223,0.0466503464,0.0802313164,0.0609858558,-0.1189645082,-0.1985368878,-0.5542483926,-0.1113599762,-0.3961032629,0.2218829095,-0.1178487986,-0.0367410257,0.1817189604,0.09351293,0.1631679535,0.0809159502,0.4353494644,-0.0103403851,-0.5547611713,-0.1673024148,0.3071158528,-0.1017958298,0.4244991839,0.2658318579,-0.0473752953,0.3789803684,0.0672727078,-0.1474138498,-0.1304980665,0.1813167334,0.4732993543,0.1243743971,-0.4087257087,-0.4832827747,0.0223345831,0.0931497961,-0.1049791053,0.1278330684,-0.2108167112,-0.2256492674,0.0660234392,-0.0017277776,0.7435305119,-0.2578614056,-0.1990886629,-0.2611181438,-0.2900470495,0.1441708207,-0.4983547926,0.2261814028,0.0256028622,0.1773861349,-0.1435110122,-0.2341738343,0.2464895397,0.3061531782,-0.221584633,0.1286787987,0.0511214659,0.2013872564,-0.2390244305,0.5826625228,-0.0189615004,-0.0177878812,-0.3193980753,0.104107745,0.0037563657,-0.0489699878,-0.1131270975,-0.125956282,-0.0013959511,0.259002924,-0.096113719,-0.3823566735,-0.3600711524,0.2111903429,0.1050608978,-0.3539367914,0.2686483562,0.0319915861,0.5623453259,-0.1565311998,-0.0621476993,0.210537225,-0.2716684043,-0.0791129172,-0.2942237854,0.0024068446,0.5046663284,-0.2038248628,0.1473563612,-0.0805545002,-0.1571337581,-0.2189765126,-0.3816920221,-0.2212599218,0.3946541548,-0.0861575603,0.1731692702,0.210847795,0.0290962923,-0.1175232679,0.1429704279,0.3587446511,0.12759386,0.4317105114,0.237275213,-0.0406313799,-0.177256912,-0.0860608593,0.2126780599,-0.1159942076,0.1123088151,0.1777402163,-0.2262970209,-0.242099911,-0.1225478947,0.2858470678,0.0342889205,0.2459283322,0.1649119556,0.0479518436,0.0910646617,0.1329673976,0.1334471405,-0.5578346252,-0.3707943857,-0.1880691499,-0.233832553,0.3873942494,-0.0249369498,0.3690789044,-0.2060165256,0.0004832617,0.1673028171,0.096392408,-0.2586403191,-0.0959255546,0.0644191876,0.0528929159,0.1362149417,-0.1169165,-0.0599542186,0.0561056249,0.0801236928,-0.1204710305,-0.4501246512,-0.0511104763,0.0752350688,0.1613643318,-0.0321672224,-0.0232717581,0.1941988319,-0.1356695294,-0.1607011706,-0.0470942594,-0.242409423,-0.3358577788,-0.0912326872,0.4094567001,0.1476756632,0.2842506468,0.5460193157,0.058822833,0.15668419,0.0447849855,0.5059970021,0.1213838011,0.4129478335,-0.014011235,-0.1775554568,0.1235497147,-0.0456450395,-0.3077005744,0.2858734727,-0.195826456,0.0040665441,0.1682368517,0.0893537328,0.3202165067,-0.0025661313,-0.0071927086,0.1407453418,0.133114323,-0.1933725029,-0.1160169914,0.5323107839,0.4479610026,-0.0158973038,-0.0252860971,0.6380771995,0.1919617951,0.3682045639,0.1919296831,-0.0265921447,0.1073520705,0.4574215114,-0.0394915044,0.2033310682,0.0905972198,0.3273359239,0.2957485914,0.1112100556,0.3611391485,0.2318948358,0.5117670298,0.059779264,-0.0058635501,0.5339126587,-0.2083244175,0.1488219351,0.1422409564,0.1171240732,-0.0424936228,0.0013873227,0.6828011274,-0.1836397797,-0.2119149864,-0.2027072161,-0.3007080555,-0.2694805861,-0.1013774425,0.1462928802,-0.2233209908,-0.2626111507,0.0047918037,-0.0800817534,-0.1858712882,0.2529278994,-0.0592221655,0.2315878719,-0.207840398,0.3712092936,0.0858629495,-0.0211573411,0.0566335358,0.2486108094,-0.4426696599,-0.1022275537,0.1831142157,-0.1289658844,-0.0914923996,-0.0448100045,-0.5634875298,-0.0721865669,-0.0020705496,-0.1619859785,0.1221481413,0.4456851184,0.3167034686,0.4698498845,0.075992696,0.0536140427,0.0005980488,0.0440060571,-0.3302556276,-0.2370290309,-0.5462009907,-0.0928848833,0.1749064922,-0.2253658026,-0.0308791008,0.1244031265,-0.4345695376,-0.570279479,0.3623164594,-0.1988381296,-0.1633509696,-0.1715811789,0.0371312797,-0.0079857549,0.2425962538,0.0896148235,0.4450386167,0.0270550326,-0.451444149,-0.2322577089,-0.0339223109,0.117955029,0.4451994002,-0.1362770349,0.1203680485,0.3110707402,-0.0680060983,0.4802006185,-0.044460956,0.0065024267,0.055573523,-0.4572417438,0.0441687778,0.0710183308,0.0619556867,0.0274661053,0.289616257,0.0129017103,-0.2471276373,-0.1294254959,-0.2906501293,-0.2042718679,0.2422358245,0.0910198987,0.6578359008,0.1784980148,0.495441258,0.183399573,-0.0436401293,-0.3782809079,-0.1888428628,0.0550800748,0.1864576787,-0.1142552271,0.1264969707,-0.2986611426,0.2966116071,-0.1278170347,0.1972867101,-0.0617858768,-0.2728451192,0.1165686026,0.029142132,-0.0234942194,-0.1066807732,0.4389766753,0.0289153755,0.1931147277,-0.2722263336,-0.2121006399,0.0893116817,-0.150519371,0.188008979,-0.2057086527,-0.3503818512,-0.1673581451,0.4357657731,0.1038014814,-0.8470209241,-0.3639751375,0.4440367222,-0.1095328331,0.239404276,0.4266268313,0.4699425995,-0.0650025755,-0.119473584,0.0844103098,-0.1690747738,-0.0248870496,-0.1886462122,-0.2311225235]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/188","title":"When will the remaining math_dataset modules be added as dataset objects","comments":"Hi @tylerroost, we don't have a timeline for this at the moment.\r\nIf you want to give it a look we would be happy to review a PR on it.\r\nAlso, the library is one week old so everything is quite barebones, in particular the doc.\r\nYou should expect some bumps on the road.\r\n\r\nTo get you started, you can check the datasets scripts in the `.\/datasets` folder on the repo and find the one on math_datasets that will need to be modified. Then you should check the original repository on the math_dataset to see where the other files to download are located and what is the expected format for the various parts of the dataset.\r\n\r\nTo get a general overview on how datasets scripts are written and used, you can read the nice tutorial on how to add a new dataset for TensorFlow Dataset [here](https:\/\/www.tensorflow.org\/datasets\/add_dataset), our API is not exactly identical but it can give you a high-level overview.","body":"Currently only the algebra_linear_1d is supported. Is there a timeline for making the other modules supported. If no timeline is established, how can I help?","comment_length":160,"text":"When will the remaining math_dataset modules be added as dataset objects\nCurrently only the algebra_linear_1d is supported. Is there a timeline for making the other modules supported. If no timeline is established, how can I help?\nHi @tylerroost, we don't have a timeline for this at the moment.\r\nIf you want to give it a look we would be happy to review a PR on it.\r\nAlso, the library is one week old so everything is quite barebones, in particular the doc.\r\nYou should expect some bumps on the road.\r\n\r\nTo get you started, you can check the datasets scripts in the `.\/datasets` folder on the repo and find the one on math_datasets that will need to be modified. Then you should check the original repository on the math_dataset to see where the other files to download are located and what is the expected format for the various parts of the dataset.\r\n\r\nTo get a general overview on how datasets scripts are written and used, you can read the nice tutorial on how to add a new dataset for TensorFlow Dataset [here](https:\/\/www.tensorflow.org\/datasets\/add_dataset), our API is not exactly identical but it can give you a high-level overview.","embeddings":[-0.2745408714,-0.0747141987,-0.2828445137,0.0816329569,0.1013158038,0.0851158649,0.1622705013,0.181571573,0.1549656242,0.028999202,0.0906958506,0.2542609274,-0.2430138886,0.091335535,0.255644381,-0.342015326,0.2013438195,-0.0719502717,-0.2481056452,-0.359226495,-0.108699061,0.0939241946,-0.2666563094,-0.211000964,-0.1336212009,-0.0600566156,-0.2486087978,-0.0442815199,-0.4253519475,-0.2308004946,0.0661661476,0.2980558574,0.2317584157,0.3571213186,-0.0001025364,-0.1791470647,0.1952542514,0.0659597665,-0.0967570692,-0.3587574065,-0.3629669845,-0.2668958306,0.1784650981,0.108467035,0.0992728546,0.0903836563,-0.0391332321,-0.4096406698,0.0279256608,0.3547129631,0.2386996895,0.3567646742,0.4147938788,-0.1960519999,0.2396775931,0.0669638962,-0.1821236759,0.0879681706,0.4211678803,0.1152278408,0.3046656549,0.1468877345,0.1961058527,0.1099760458,0.3800205886,0.0183225162,0.0137876738,-0.5149787068,-0.3262500465,0.2852599323,0.7866033316,-0.2344968766,-0.2374750227,-0.1045813859,-0.0337128527,-0.2809472382,0.0365477316,0.0314570703,0.2146729827,-0.163695693,0.0673857555,-0.40398857,-0.3353688121,0.4025502503,0.1397747397,0.4864596725,0.1236416698,0.1261004061,0.0988131762,-0.1491748095,0.3951944709,-0.0036637334,0.216307655,0.2663553953,-0.2195387036,-0.2165071964,-0.0990744382,-0.0851168185,0.0185330827,-0.1848434955,-0.1483968496,0.3597196937,-0.3094288707,0.2582101226,0.2134395242,-0.1501586139,-0.1043486893,-0.0047548832,0.4263304174,-0.0336952358,0.0713230073,0.1971784234,-0.1961677223,0.0076112011,-0.2757908106,0.0272748265,0.0295240488,-0.0128217051,0.1225248352,-0.3832071722,0.0956238732,-0.3619911373,-0.047780741,-0.0845338106,-0.3333867192,0.1714140326,-0.0095392363,0.0272273235,-0.1498740464,-0.0231564753,-0.0724998415,0.1515766829,-0.2152209878,0.0198928639,0.2334244549,-0.1018989086,0.0881501958,0.0125770569,0.2016483247,0.1553193033,0.1876082569,0.108616896,-0.1332701743,0.229112938,-0.1691801101,0.0380621664,-0.2693698406,0.1862420589,-0.0961535275,-0.0502220318,-0.149671793,-0.2168762386,-0.1686042398,0.2626042068,-0.1451126337,-0.3540813029,-0.2239722908,0.3776288629,-0.2807542384,0.2039030343,-0.1772480756,-0.1275356859,-0.0954099596,-0.3391961753,0.0054367357,0.0747958869,-0.4191736579,0.1555751264,-0.1323328465,-0.0583360679,-0.2366979718,-0.0541884974,-0.1949496567,0.0152082779,-0.095291689,-0.1207710579,0.4563644826,-0.5523675084,-0.0646584705,-0.169996798,0.2277211845,-0.2298590839,-0.1641166955,0.1930214912,0.0397006869,-0.1482824236,-0.2039136887,0.3211025894,-0.1425623,-0.2917489409,-0.2070939243,-0.1105629578,0.0519356839,0.1116952151,0.3645806611,0.0175068658,0.1615471691,0.2132958919,0.0144435093,0.1034300923,0.0315665305,0.2074112892,0.4144863188,0.2146508098,0.46539101,-0.4261690378,-0.4392787814,0.1380775422,0.1953701079,0.1881584078,0.1049885005,0.0862472281,-0.0043757618,0.1029463112,-0.1205544248,0.3178802729,0.14472574,-0.0405566916,0.2105709612,-0.1462246627,-0.5968189836,0.0926127657,-0.1041231528,0.4140124917,-0.4701019824,0.3047421575,0.0153421648,-0.1535006464,0.1693399101,0.2441736311,0.0057320679,-0.4595024884,0.0803301707,0.2964682579,0.1613706797,0.2357262075,-0.0338062085,0.4449404478,0.4259872139,-0.0609193929,0.311979115,-0.3168822825,-0.0265079327,-0.0569371842,0.0102055147,0.2147308439,-0.037214946,-0.0947376564,0.2684358656,0.0165636614,0.1671214402,0.2664136589,-0.0421637744,-0.2157949805,0.0300439075,-0.0084280111,-0.0213499125,-0.1336777806,-0.2969611585,0.228976801,0.6034273505,-0.1287028044,0.0356038362,0.0619414672,-0.2499331981,-0.0089356387,0.0916371346,0.2701995969,0.1916397214,0.1497255117,0.2187143415,-0.0189697258,-0.2912887931,-0.1258143336,0.0364939533,0.1996529698,0.1430567652,-0.0602197126,-0.0137000708,0.0058994349,-0.0994801819,-0.0192238316,-0.0458139889,0.4715058506,0.0794292912,-0.1826927215,-0.4099810123,-0.1873921901,-0.0715556815,-0.1772991121,-0.1551743746,-0.3065158427,0.0328201838,-0.0146003403,0.1989654899,0.0164232943,0.198269397,0.0992313027,-0.2671225071,0.3675664663,-0.2708781362,0.0022397877,-0.2175015211,0.1902993172,0.256509155,-0.086112164,0.4182340503,-0.0340089835,0.233294189,-0.1385352314,-0.640366137,0.1763880551,-0.1495903134,-0.0268919356,-0.0202498063,-0.08343108,0.179927662,0.1192989126,0.072791107,-0.291844964,-0.0514951162,-0.0440872423,-0.1970873624,-0.1112021282,-0.1318808645,-0.5581411123,-0.4593666792,-0.2616687715,0.1764215082,0.3262957931,0.0808147788,0.1624584645,0.0841361061,0.1643991172,-0.0631444678,0.0394746102,0.1127866805,-0.2584373057,0.4067581892,-0.307110697,-0.3843967021,0.4961076081,-0.1042147502,0.349216193,0.2127223015,-0.3054118156,-0.1986851692,-0.0253096335,-0.0159913544,0.1663932949,-0.0196063165,0.2342217118,-0.009280446,0.052656129,-0.2524116635,-0.1440183669,-0.0861576274,0.376968801,0.0294452496,0.143426463,0.3347580731,-0.0332237892,0.4305326343,0.0364487432,-0.2452243418,0.0504049435,0.16883564,0.1086771637,0.0113822129,-0.0502178073,0.0771260932,0.0313266069,0.2990892529,0.0489156954,0.1208809018,-0.1074631885,-0.3140335679,0.3371354938,-0.1784428954,-0.0018073107,-0.0018803782,-0.2516934574,0.3295735717,0.0887835696,-0.1960984468,-0.2880051136,-0.1487681568,0.1574665606,0.3545742035,0.2419824004,-0.1290258914,-0.5378568769,-0.1852818131,-0.1731888503,0.4174763858,-0.1019777581,0.1792779565,-0.192293182,0.1106683016,0.1177428886,0.0859808624,0.1878304034,-0.4766139388,-0.5271233916,0.2791646719,-0.1362728179,-0.2341880202,-0.0242565647,-0.3106378317,0.1136726514,-0.0290838815,0.448890686,0.0154803386,-0.3676791787,0.2430874258,0.1308574528,0.0859105811,-0.1405642778,0.2629944086,-0.0101359738,-0.1173252165,-0.2620448768,-0.1052145883,-0.1123388708,-0.1167273223,-0.5168627501,0.0712359026,-0.332726866,0.1847617179,-0.0470486656,-0.0565901175,0.2749298513,0.0368309617,0.3168192804,-0.1453709155,0.4666672051,-0.0115131373,-0.3804034591,-0.1696540117,0.1740658283,-0.0916218683,0.1956475824,0.1504790634,-0.0529821515,0.1828366071,-0.0608185194,-0.0310345944,-0.1255170554,0.1641343683,0.4193398654,0.2279085368,-0.2852481306,-0.6071881056,0.114722766,0.0161522292,-0.0485952869,0.4207161367,-0.1339213401,-0.1745081246,0.1109836623,0.0410565399,0.5863262415,-0.2631882727,-0.2927809358,0.0347498357,-0.086299859,0.4732591808,-0.6111395955,0.1742395014,-0.0342668928,0.1087060049,-0.0851123482,-0.2548707426,0.2109448612,0.0722769126,-0.2013601363,0.0995427296,0.1770904511,0.1380800307,-0.1942107081,0.659774363,-0.0631279275,-0.1403733045,-0.0989978984,0.2516933382,-0.0283081383,0.0521816462,-0.0602608062,-0.1913780868,-0.0318768881,0.1811335832,-0.2849520743,-0.1687600762,-0.2213168591,0.0083194003,0.0913176909,-0.0432951637,0.3468939364,0.0962307751,0.5915400982,-0.0320243873,-0.2376018763,0.2797388136,-0.2682278454,-0.1178873554,-0.1228888333,0.0172072388,0.3721633255,-0.296742022,-0.0120822983,-0.0773553401,-0.0002786314,-0.2425985038,-0.3651345074,-0.0714226365,0.3298620284,0.0970856845,0.1923030615,0.3033461869,-0.0927824527,0.0075603235,0.2478661388,0.451178968,0.0147248451,0.4076645374,0.0479846336,0.000529936,-0.1513430029,-0.1505491585,0.2708599865,-0.287745297,0.0725747198,0.4370304346,-0.1949128211,-0.2097455412,-0.1988448948,0.1838006675,0.0571298003,0.0781837702,0.155958876,0.0962726697,0.1928562969,0.0999173746,-0.0235089734,-0.4111428857,-0.4108963311,-0.2932205498,-0.2243692279,0.2566511035,0.0494616479,0.2794498205,-0.1963748783,-0.0722374022,0.1741121113,0.1493712813,-0.3791826963,0.057014551,0.1513358206,0.0361884348,0.2261245847,-0.0725012124,-0.0361216255,-0.0717405081,0.0938639268,-0.0279478356,-0.4679452479,-0.1847219616,0.0420264378,0.1002145261,-0.0426371731,-0.2287183404,0.1266716719,-0.3141507208,-0.2827925682,-0.0249093808,-0.0397677571,-0.3050769866,-0.1421205997,0.3311749697,0.1819973886,0.1702482998,0.558208704,0.0951398164,0.212450847,0.2048585266,0.4021882713,0.1895201057,0.260106802,-0.0821143761,-0.1481980532,0.0781635195,-0.1382057667,-0.1794576645,0.3982890248,-0.1692571193,0.1535912305,0.0323557965,0.1014401391,0.2642419636,0.0210733861,0.0369644649,0.1414287984,0.2131100297,-0.134429872,-0.0558996871,0.4027067721,0.4979761839,0.101879634,0.0768641904,0.5620511174,0.1583319306,0.2849806845,0.1785443872,0.2334572524,-0.1421746463,0.4086755812,-0.1030534506,-0.0537332669,0.073440142,0.0899200812,0.3814928234,0.0946704447,0.3289197683,0.1840584576,0.4397294521,0.1460433155,-0.080079481,0.5982282162,-0.3191629648,0.0660488233,0.1627114266,0.0607869476,0.0938301012,0.0774185732,0.4690658152,-0.0093901744,-0.1389848441,-0.0540349297,-0.1167002544,-0.1087819859,-0.0059311637,-0.0414071456,-0.1765544415,-0.2063525319,-0.0955451801,0.0424556173,-0.2446394563,0.1084115356,-0.006784068,0.0250752829,-0.1830290705,0.3499516249,-0.0880802646,-0.0695842057,0.2082016617,0.4416110814,-0.2254701257,-0.2055891007,-0.092307128,-0.0775796473,-0.0699910223,0.0203931723,-0.3597539663,0.3094950914,0.0517013483,-0.1929582655,0.207717672,0.4019238949,0.0711452067,0.3664084077,-0.1109784991,0.1595133245,-0.0535554215,-0.022820387,-0.2833408713,-0.0211471878,-0.4163031876,-0.0843380466,0.110245198,-0.1518711001,-0.0453495011,-0.0458963662,-0.4722163975,-0.5669901967,0.5651461482,-0.1895702332,-0.1535135508,-0.2535608113,0.1081456318,-0.0953315347,0.2511740625,0.1214741617,0.4550127983,0.0582242981,-0.3774459362,-0.2873682678,0.0056762523,0.1491586566,0.2815353274,-0.1130204573,0.1998531222,0.2401165664,-0.0334272832,0.430388093,-0.2973402441,0.0923043713,-0.0040982221,-0.3275870085,0.0098558348,-0.2028948367,0.0415496379,0.1104596481,0.2590786219,0.0708226189,-0.2688297033,-0.0257426314,-0.4121479988,-0.0233589951,0.3395145833,0.0814252198,0.4212688804,0.1432822943,0.2954122722,0.180005163,-0.0455534607,-0.4166835248,-0.163995266,-0.0197580811,0.2397600859,-0.1299556941,0.0806001574,-0.1950434893,0.0144461617,-0.1657220274,0.1421409696,-0.1651901156,-0.3549889922,0.0366339497,0.0665656179,-0.0896536931,0.0327761918,0.4078936577,0.1349681914,0.2335764021,-0.2401528656,-0.2137975246,0.1258515269,0.055552341,-0.0777761117,-0.24822478,-0.1938058883,-0.1850714833,0.2629101276,0.113597393,-0.8757445216,-0.2172639072,0.2978822589,-0.090507932,0.0677325502,0.2611724138,0.4608124197,-0.0116976481,-0.1992035508,0.0876035541,0.0045112437,-0.0585020594,-0.1302704662,-0.1536990255]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/188","title":"When will the remaining math_dataset modules be added as dataset objects","comments":"Thanks I'll give it a look","body":"Currently only the algebra_linear_1d is supported. Is there a timeline for making the other modules supported. If no timeline is established, how can I help?","comment_length":6,"text":"When will the remaining math_dataset modules be added as dataset objects\nCurrently only the algebra_linear_1d is supported. Is there a timeline for making the other modules supported. If no timeline is established, how can I help?\nThanks I'll give it a look","embeddings":[-0.3410984576,-0.2424143106,-0.3088272512,0.2483329475,0.0874816626,0.0654156953,0.0565822683,-0.0906167403,0.2255054265,0.0141560398,0.1430434436,0.1250177622,-0.3328547478,0.1273879707,0.1434667408,-0.2845974863,0.3416379094,-0.1352493912,-0.2972112298,-0.2912487388,-0.2218004912,0.1958979517,-0.4259852171,-0.1691635549,-0.3144358099,0.0115833944,-0.196818769,-0.1117288768,-0.5340137482,-0.3484967947,-0.0822358131,0.2514548898,0.2053242177,0.2099120617,-0.0001108197,-0.3030692041,0.2637627125,0.0593355596,-0.3291651011,0.0975298285,-0.5096616149,-0.2071602792,0.3038553894,0.1753717214,0.0707104355,0.0929641277,-0.036935024,-0.5524175167,-0.0208500344,0.2459982783,0.2004097998,0.1997110099,0.362375021,-0.3322079778,0.1793606132,0.1371433735,-0.1137889549,0.1441750526,0.5609744191,0.2628084421,0.3038804233,0.0424359888,0.1489128023,0.0769527853,0.320342809,-0.0641629994,0.1916632503,-0.3881634176,-0.3662427664,0.094278276,0.9578533769,-0.0818588212,-0.2407768965,-0.0565493181,-0.1401897222,-0.2052854449,-0.0500607267,-0.0059195636,0.3265064061,-0.0660993829,0.1209754646,-0.3964315951,-0.2055218667,0.3055559099,0.134762302,0.6005659699,0.1247118413,0.043154113,0.0206987038,-0.1657928973,0.445425421,-0.1143606305,0.1040334553,0.2369570285,-0.1579467207,-0.2291116267,-0.2588785887,0.3935271204,0.0197362695,-0.2413694859,-0.2384201884,0.2992182076,-0.106755197,0.1807534099,0.2318617553,-0.203790307,-0.0436790399,-0.0899783447,0.4480135441,-0.078194499,-0.0250243507,0.2067690492,-0.1025762931,0.0160316788,-0.1090252176,0.1421003044,-0.0172292162,-0.0523946695,0.1510453075,-0.510448575,0.0398562476,-0.2978683412,-0.1304353029,-0.1046541408,-0.1687518805,0.1454890072,-0.0590146929,0.0536210053,-0.1249293759,-0.2156466097,-0.059027154,0.0777151138,-0.179419741,0.0523472838,0.1100503355,0.0582211874,-0.0302559454,-0.0288503859,0.2803779244,0.1171739101,0.0044810171,0.0781187415,-0.2440403104,0.1813822687,-0.2835539579,0.0769680068,-0.2954629958,0.3487300575,-0.0747748986,-0.1449867636,-0.0066689914,-0.1313420534,-0.0892087221,0.2227631509,-0.0173184313,-0.3274194896,-0.1466283947,0.3050264716,-0.3041628301,0.139053002,-0.1721872687,-0.119150795,-0.0836475417,-0.2807909846,-0.1013681293,0.1481173933,-0.3222332895,0.1368153542,-0.118338801,-0.1531817019,-0.3244924843,-0.0247105714,-0.0616077632,0.0380007252,-0.2519438863,-0.2546457648,0.310703367,-0.4753303528,-0.0350084156,-0.0792048424,0.2728956938,-0.2962027192,0.0365070179,0.2591666281,-0.0000215988,-0.1238663793,-0.1284648925,0.2618477345,-0.0564185455,-0.2676454484,-0.1588328928,-0.0110023599,0.076877676,0.074649781,0.4328509569,0.1243026033,0.2916391492,0.174426049,0.0106367227,0.2197401524,0.1898728013,0.2274948657,0.3218225837,0.0316124372,0.4869033992,-0.5355285406,-0.3917950094,0.0626019835,0.1223112494,0.1874697208,0.1910650134,0.0640428588,-0.0871293843,0.1931852698,-0.2099571228,0.3823239207,0.11548803,0.0455068015,0.2619288862,-0.2016807646,-0.3916635811,0.273655504,0.1031255499,0.383690089,-0.3613009751,0.2994484007,0.0044430629,-0.219841823,0.2597680688,0.3082515597,0.2168622762,-0.4528001547,0.190098688,0.2162635922,0.3785786629,0.1802647561,0.1793427318,0.3139493465,0.4298250079,-0.1205711216,0.4063449502,-0.2590812743,0.0518513992,-0.1011704877,0.2054608017,0.1953546852,0.0244446043,-0.1234561056,0.2051337212,0.0449759737,0.1375698298,0.3254035115,-0.0452507921,-0.3018839359,0.0936436057,0.1374624819,-0.1135712191,0.0710318387,-0.2753412426,0.1720346808,0.5329902768,-0.0740585551,0.1456964463,-0.0103696445,-0.3170343637,-0.0453641266,0.1482510567,0.0699487701,0.1546241045,0.2403211743,0.2494681478,-0.1814412475,-0.215448156,-0.056086164,0.0853165165,0.1934390813,0.0936708674,0.0593804531,-0.0380198993,-0.0966575518,-0.1394646615,0.0231874753,-0.0778502077,0.3343726695,0.0652648211,-0.2233352214,-0.442228049,-0.2188114673,0.0035248045,-0.1684443504,-0.1715833843,-0.3064672649,0.1162197217,0.0081726313,0.3046708703,0.0555125587,0.0612065569,0.1906786561,-0.3657407761,0.5672510862,-0.3419604897,0.0265102088,-0.205760166,0.1314567327,0.2076629102,-0.1254234314,0.3783016801,-0.0928680003,0.3216629028,-0.1843074411,-0.4933131933,0.2242249995,-0.0474697053,0.0546786971,-0.0518210232,-0.1726034135,0.2117599696,0.2580411434,0.0938757509,-0.1709171236,-0.1873796731,-0.0318491198,-0.1937524229,-0.076852195,0.0162990093,-0.2904217839,-0.6114137769,-0.2112315446,0.1360743046,0.2191262096,-0.0011316303,0.0192137007,0.0191244036,-0.0225539692,-0.2062042654,0.1421811879,0.1328959763,-0.1087811813,0.3139064908,-0.2096964419,-0.2792170644,0.3892032504,-0.1301432848,0.2542756498,0.2978237867,-0.1264096797,-0.251226902,0.0092503978,0.0040727975,0.1239010021,-0.1677628905,0.2840749323,0.129389748,0.0200561509,-0.155801028,-0.2746975124,-0.133927241,0.2842914164,-0.0268430263,0.0347536057,0.3027783334,-0.1961195916,0.3839207292,0.0493247695,-0.1198296174,0.1704283506,0.3409096301,0.1013951376,-0.1540590376,-0.0111325923,-0.062838763,0.0203241818,0.1734556854,0.0661011934,0.104527384,0.0236149989,-0.3092856705,0.3639851809,-0.3199703693,0.065634653,-0.1783097237,-0.4126060903,0.550579071,0.0860374421,-0.2709344923,-0.1835051924,-0.0288764387,0.0795209855,0.4045588374,0.3053305447,-0.1500576586,-0.6531464458,-0.1638761163,-0.0147076072,0.4232454002,-0.0879985765,0.2627893984,-0.3503319621,0.0970956907,0.1996590197,0.1756581664,0.1436232775,-0.5349427462,-0.4441000819,0.3492245078,-0.0991461724,-0.2224990278,0.0709012672,-0.2465140671,0.2019489259,-0.0773451105,0.5559631586,-0.0094667021,-0.3373194933,0.1913606972,0.2179750204,0.0615745299,-0.2169039994,0.2572614253,0.2461851239,-0.0068175406,-0.1070657745,0.0033146515,-0.1534252167,-0.0455975905,-0.6313536167,-0.1284250766,-0.5246870518,0.1400585324,-0.1413397491,-0.0076891291,0.14887622,0.1173981503,0.1488257647,-0.0251950938,0.4521000683,-0.0341398418,-0.4375915825,-0.205195114,0.2368364781,-0.1475585997,0.380628854,0.1791170985,-0.0253765974,0.361920923,-0.0429387987,-0.0996600464,-0.1822796762,-0.0520990416,0.3757476509,0.1437899172,-0.3243154585,-0.348243773,-0.0724535882,-0.003976691,-0.1171001345,0.2653421164,-0.0955626443,-0.2833817303,0.0375611298,0.0189009961,0.7353761196,-0.2171992809,-0.2202216983,-0.2844323218,-0.3146977723,0.1622819006,-0.3342192173,0.1789772958,-0.0368357077,0.183716476,-0.1002653614,-0.2505615056,0.2425124049,0.2195816785,-0.2756412327,0.0944701284,0.0468837172,0.0996489972,-0.2109134644,0.6839082241,-0.0065100566,0.0193858426,-0.207780242,0.16194655,-0.0093174549,-0.076972805,-0.158674106,-0.1635667682,0.053417027,0.2087747157,-0.190799132,-0.3126473129,-0.2707546651,0.0897511467,0.0353780128,-0.2313713729,0.2965646088,-0.0601539463,0.5443842411,-0.1155637056,-0.0842638463,0.1971820891,-0.2948530912,-0.1611840427,-0.3511893153,0.0602810718,0.2719040215,-0.2919768095,0.1141327322,-0.0579791628,-0.0566374809,-0.3514795005,-0.4117631614,-0.0802860558,0.3386409581,0.0423865095,0.2086243182,0.23321639,-0.0041238382,-0.1115069762,0.1742697358,0.4247562885,0.1579500288,0.4252861738,0.1859747022,-0.0870785043,-0.239648357,-0.059383098,0.2048105597,-0.1155908629,0.0038083997,0.2635618448,-0.3112888932,-0.2583000362,-0.1559938639,0.1112843454,-0.0153024215,0.2660019994,0.2332464308,0.084147431,0.0839342177,0.1474430263,0.0300922636,-0.5740519762,-0.4696795046,-0.2773629129,-0.2464189082,0.3738848269,-0.091226697,0.2899071276,-0.1435542703,-0.0871247873,0.1662177593,0.1430779099,-0.2999598384,-0.0073179896,0.1262932122,0.0280282367,0.2705736458,-0.2099667192,-0.1735126972,0.0676492229,0.1026945263,-0.144298926,-0.3999058902,-0.1017081439,0.2256655246,0.1449931115,-0.0172720086,-0.1272612661,0.1768806279,-0.2540200055,-0.180998534,-0.0142677082,-0.2206828296,-0.3479394615,-0.103079319,0.2945659161,0.3010712862,0.1956543624,0.5805670023,0.2237917632,0.2370733619,0.0705169067,0.4476995766,0.1601147801,0.3951753676,-0.0303282291,-0.0986498818,0.1390185803,-0.1054670736,-0.2182908505,0.3679852784,-0.1222517416,-0.0011161031,0.1878789365,0.1095593795,0.2149226964,0.0785478055,-0.0428632796,0.0980708376,0.1533266008,-0.2046226561,-0.197643593,0.533506155,0.4662817121,0.0038501048,0.0499265492,0.6637842655,0.1648730636,0.5425274372,0.2058667839,0.1580878943,0.0824878737,0.3279881775,-0.175767988,0.1875946671,0.121950753,0.3320611715,0.4474227428,0.1208775267,0.3971507251,0.2437744439,0.407977581,0.0394998118,-0.0094609912,0.642421782,-0.2210310996,0.1619430631,0.0962211788,0.1458352953,-0.0529050715,0.0949152187,0.5458686352,-0.1661498398,-0.104549326,-0.1932505816,-0.2343084812,-0.2304224968,-0.106684804,0.1738944948,-0.2109497637,-0.1755759418,-0.1384015381,-0.0718121529,-0.1312204152,0.187155351,-0.1066575497,0.1856520027,-0.2004297525,0.4355891049,0.0635900348,-0.1044566184,0.0802938789,0.3614404798,-0.3967157602,-0.037207868,0.128305912,-0.0226532407,-0.1281652749,-0.0477017947,-0.5543751717,0.052265998,-0.0355486274,-0.1705276519,0.0981559232,0.4788233042,0.1870301068,0.4486315846,0.0524710566,0.0965404585,-0.0100079319,-0.0549347699,-0.2736997604,-0.2160715461,-0.4027916193,-0.1407763511,0.1328972876,-0.2386531383,0.0616623349,0.1453961879,-0.4433215559,-0.5374725461,0.3164073229,-0.2818186581,-0.217234388,-0.1429801136,0.0603806041,-0.0602681302,0.1926898807,0.1315947622,0.4821421504,0.0596539304,-0.5285567045,-0.2749715447,0.0544138588,0.1499914676,0.41639328,-0.1898876876,0.1994531155,0.3728779256,-0.1343260854,0.5040715933,-0.0133562367,-0.0010539353,0.0425361097,-0.3714030087,0.0100937132,0.1026411504,-0.0203944426,-0.002054916,0.25466308,0.1042846367,-0.18468678,-0.0420416929,-0.3246247768,-0.2245954275,0.1962274462,0.0358183496,0.5386351943,0.1743864119,0.4056968689,0.1526355445,0.0082420819,-0.4106789827,-0.2031989545,0.1166818216,0.1915952414,-0.1002486274,0.028743973,-0.2824878395,0.1835410297,-0.0935911834,0.1585569978,-0.1943179667,-0.1567137986,0.0704187453,0.1370183825,-0.1302559376,-0.0575862974,0.3092171252,-0.0427297205,0.2472032607,-0.2669069469,-0.1988547444,0.1612575054,-0.1451902539,0.1671141833,-0.2448079437,-0.2710175514,-0.1565537602,0.2675302625,0.01454979,-0.6925178766,-0.3257705867,0.4773776531,-0.069229655,0.280571878,0.4048551917,0.4846706688,-0.1010207534,-0.1721891612,-0.0058501428,0.0180749781,0.0674914643,-0.2226111591,-0.1899529845]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/187","title":"[Question] How to load wikipedia ? Beam runner ?","comments":"I have seen that somebody is hard working on easierly loadable wikipedia. #129 \r\nMaybe I should wait a few days for that version ?","body":"When `nlp.load_dataset('wikipedia')`, I got\r\n* `WARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be used.`\r\n* `AttributeError: 'NoneType' object has no attribute 'size'`\r\n\r\nCould somebody tell me what should I do ? \r\n\r\n# Env\r\nOn Colab,\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp\r\ncd nlp\r\npip install -q .\r\n```\r\n```\r\n%pip install -q apache_beam mwparserfromhell\r\n-> ERROR: pydrive 1.3.1 has requirement oauth2client>=4.0.0, but you'll have oauth2client 3.0.0 which is incompatible.\r\nERROR: google-api-python-client 1.7.12 has requirement httplib2<1dev,>=0.17.0, but you'll have httplib2 0.12.0 which is incompatible.\r\nERROR: chainer 6.5.0 has requirement typing-extensions<=3.6.6, but you'll have typing-extensions 3.7.4.2 which is incompatible.\r\n```\r\n```\r\npip install -q apache-beam[interactive]\r\nERROR: google-colab 1.0.0 has requirement ipython~=5.5.0, but you'll have ipython 5.10.0 which is incompatible.\r\n```\r\n\r\n# The whole message\r\n```\r\nWARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be used.\r\n\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wikipedia\/20200501.aa\/1.0.0...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nAttributeError Traceback (most recent call last)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n44 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker.invoke_process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/iobase.py in process(self, element, init_result)\r\n 1081 writer.write(e)\r\n-> 1082 return [window.TimestampedValue(writer.close(), timestamp.MAX_TIMESTAMP)]\r\n 1083 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/filebasedsink.py in close(self)\r\n 422 def close(self):\r\n--> 423 self.sink.close(self.temp_handle)\r\n 424 return self.temp_shard_path\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in close(self, writer)\r\n 537 if len(self._buffer[0]) > 0:\r\n--> 538 self._flush_buffer()\r\n 539 if self._record_batches_byte_size > 0:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in _flush_buffer(self)\r\n 569 for b in x.buffers():\r\n--> 570 size = size + b.size\r\n 571 self._record_batches_byte_size = self._record_batches_byte_size + size\r\n\r\nAttributeError: 'NoneType' object has no attribute 'size'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nAttributeError Traceback (most recent call last)\r\n\r\n<ipython-input-9-340aabccefff> in <module>()\r\n----> 1 dset = nlp.load_dataset('wikipedia')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 370 verify_infos = not save_infos and not ignore_verifications\r\n 371 self._download_and_prepare(\r\n--> 372 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 373 )\r\n 374 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos)\r\n 770 with beam.Pipeline(runner=beam_runner, options=beam_options,) as pipeline:\r\n 771 super(BeamBasedBuilder, self)._download_and_prepare(\r\n--> 772 dl_manager, pipeline=pipeline, verify_infos=False\r\n 773 ) # TODO{beam} verify infos\r\n 774 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in __exit__(self, exc_type, exc_val, exc_tb)\r\n 501 def __exit__(self, exc_type, exc_val, exc_tb):\r\n 502 if not exc_type:\r\n--> 503 self.run().wait_until_finish()\r\n 504 \r\n 505 def visit(self, visitor):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in run(self, test_runner_api)\r\n 481 return Pipeline.from_runner_api(\r\n 482 self.to_runner_api(use_fake_coders=True), self.runner,\r\n--> 483 self._options).run(False)\r\n 484 \r\n 485 if self._options.view_as(TypeOptions).runtime_type_check:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in run(self, test_runner_api)\r\n 494 finally:\r\n 495 shutil.rmtree(tmpdir)\r\n--> 496 return self.runner.run_pipeline(self, self._options)\r\n 497 \r\n 498 def __enter__(self):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/direct\/direct_runner.py in run_pipeline(self, pipeline, options)\r\n 128 runner = BundleBasedDirectRunner()\r\n 129 \r\n--> 130 return runner.run_pipeline(pipeline, options)\r\n 131 \r\n 132 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_pipeline(self, pipeline, options)\r\n 553 \r\n 554 self._latest_run_result = self.run_via_runner_api(\r\n--> 555 pipeline.to_runner_api(default_environment=self._default_environment))\r\n 556 return self._latest_run_result\r\n 557 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_via_runner_api(self, pipeline_proto)\r\n 563 # TODO(pabloem, BEAM-7514): Create a watermark manager (that has access to\r\n 564 # the teststream (if any), and all the stages).\r\n--> 565 return self.run_stages(stage_context, stages)\r\n 566 \r\n 567 @contextlib.contextmanager\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_stages(self, stage_context, stages)\r\n 704 stage,\r\n 705 pcoll_buffers,\r\n--> 706 stage_context.safe_coders)\r\n 707 metrics_by_stage[stage.name] = stage_results.process_bundle.metrics\r\n 708 monitoring_infos_by_stage[stage.name] = (\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in _run_stage(self, worker_handler_factory, pipeline_components, stage, pcoll_buffers, safe_coders)\r\n 1071 cache_token_generator=cache_token_generator)\r\n 1072 \r\n-> 1073 result, splits = bundle_manager.process_bundle(data_input, data_output)\r\n 1074 \r\n 1075 def input_for(transform_id, input_id):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in process_bundle(self, inputs, expected_outputs)\r\n 2332 \r\n 2333 with UnboundedThreadPoolExecutor() as executor:\r\n-> 2334 for result, split_result in executor.map(execute, part_inputs):\r\n 2335 \r\n 2336 split_result_list += split_result\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in result_iterator()\r\n 584 # Careful not to keep a reference to the popped future\r\n 585 if timeout is None:\r\n--> 586 yield fs.pop().result()\r\n 587 else:\r\n 588 yield fs.pop().result(end_time - time.monotonic())\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 430 raise CancelledError()\r\n 431 elif self._state == FINISHED:\r\n--> 432 return self.__get_result()\r\n 433 else:\r\n 434 raise TimeoutError()\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/utils\/thread_pool_executor.py in run(self)\r\n 42 # If the future wasn't cancelled, then attempt to execute it.\r\n 43 try:\r\n---> 44 self._future.set_result(self._fn(*self._fn_args, **self._fn_kwargs))\r\n 45 except BaseException as exc:\r\n 46 # Even though Python 2 futures library has #set_exection(),\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in execute(part_map)\r\n 2329 self._registered,\r\n 2330 cache_token_generator=self._cache_token_generator)\r\n-> 2331 return bundle_manager.process_bundle(part_map, expected_outputs)\r\n 2332 \r\n 2333 with UnboundedThreadPoolExecutor() as executor:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in process_bundle(self, inputs, expected_outputs)\r\n 2243 process_bundle_descriptor_id=self._bundle_descriptor.id,\r\n 2244 cache_tokens=[next(self._cache_token_generator)]))\r\n-> 2245 result_future = self._worker_handler.control_conn.push(process_bundle_req)\r\n 2246 \r\n 2247 split_results = [] # type: List[beam_fn_api_pb2.ProcessBundleSplitResponse]\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in push(self, request)\r\n 1557 self._uid_counter += 1\r\n 1558 request.instruction_id = 'control_%s' % self._uid_counter\r\n-> 1559 response = self.worker.do_instruction(request)\r\n 1560 return ControlFuture(request.instruction_id, response)\r\n 1561 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/sdk_worker.py in do_instruction(self, request)\r\n 413 # E.g. if register is set, this will call self.register(request.register))\r\n 414 return getattr(self, request_type)(\r\n--> 415 getattr(request, request_type), request.instruction_id)\r\n 416 else:\r\n 417 raise NotImplementedError\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/sdk_worker.py in process_bundle(self, request, instruction_id)\r\n 448 with self.maybe_profile(instruction_id):\r\n 449 delayed_applications, requests_finalization = (\r\n--> 450 bundle_processor.process_bundle(instruction_id))\r\n 451 monitoring_infos = bundle_processor.monitoring_infos()\r\n 452 monitoring_infos.extend(self.state_cache_metrics_fn())\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/bundle_processor.py in process_bundle(self, instruction_id)\r\n 837 for data in data_channel.input_elements(instruction_id,\r\n 838 expected_transforms):\r\n--> 839 input_op_by_transform_id[data.transform_id].process_encoded(data.data)\r\n 840 \r\n 841 # Finish all operations.\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/bundle_processor.py in process_encoded(self, encoded_windowed_values)\r\n 214 decoded_value = self.windowed_coder_impl.decode_from_stream(\r\n 215 input_stream, True)\r\n--> 216 self.output(decoded_value)\r\n 217 \r\n 218 def try_split(self, fraction_of_remainder, total_buffer_size):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.Operation.output()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.Operation.output()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.SingletonConsumerSet.receive()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.DoOperation.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.DoOperation.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner._reraise_augmented()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/future\/utils\/__init__.py in raise_with_traceback(exc, traceback)\r\n 417 if traceback == Ellipsis:\r\n 418 _, _, traceback = sys.exc_info()\r\n--> 419 raise exc.with_traceback(traceback)\r\n 420 \r\n 421 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker.invoke_process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/iobase.py in process(self, element, init_result)\r\n 1080 for e in bundle[1]: # values\r\n 1081 writer.write(e)\r\n-> 1082 return [window.TimestampedValue(writer.close(), timestamp.MAX_TIMESTAMP)]\r\n 1083 \r\n 1084 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/filebasedsink.py in close(self)\r\n 421 \r\n 422 def close(self):\r\n--> 423 self.sink.close(self.temp_handle)\r\n 424 return self.temp_shard_path\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in close(self, writer)\r\n 536 def close(self, writer):\r\n 537 if len(self._buffer[0]) > 0:\r\n--> 538 self._flush_buffer()\r\n 539 if self._record_batches_byte_size > 0:\r\n 540 self._write_batches(writer)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in _flush_buffer(self)\r\n 568 for x in arrays:\r\n 569 for b in x.buffers():\r\n--> 570 size = size + b.size\r\n 571 self._record_batches_byte_size = self._record_batches_byte_size + size\r\n\r\nAttributeError: 'NoneType' object has no attribute 'size' [while running 'train\/Save to parquet\/Write\/WriteImpl\/WriteBundles']\r\n```","comment_length":24,"text":"[Question] How to load wikipedia ? Beam runner ?\nWhen `nlp.load_dataset('wikipedia')`, I got\r\n* `WARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be used.`\r\n* `AttributeError: 'NoneType' object has no attribute 'size'`\r\n\r\nCould somebody tell me what should I do ? \r\n\r\n# Env\r\nOn Colab,\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp\r\ncd nlp\r\npip install -q .\r\n```\r\n```\r\n%pip install -q apache_beam mwparserfromhell\r\n-> ERROR: pydrive 1.3.1 has requirement oauth2client>=4.0.0, but you'll have oauth2client 3.0.0 which is incompatible.\r\nERROR: google-api-python-client 1.7.12 has requirement httplib2<1dev,>=0.17.0, but you'll have httplib2 0.12.0 which is incompatible.\r\nERROR: chainer 6.5.0 has requirement typing-extensions<=3.6.6, but you'll have typing-extensions 3.7.4.2 which is incompatible.\r\n```\r\n```\r\npip install -q apache-beam[interactive]\r\nERROR: google-colab 1.0.0 has requirement ipython~=5.5.0, but you'll have ipython 5.10.0 which is incompatible.\r\n```\r\n\r\n# The whole message\r\n```\r\nWARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be used.\r\n\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wikipedia\/20200501.aa\/1.0.0...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nAttributeError Traceback (most recent call last)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n44 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker.invoke_process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/iobase.py in process(self, element, init_result)\r\n 1081 writer.write(e)\r\n-> 1082 return [window.TimestampedValue(writer.close(), timestamp.MAX_TIMESTAMP)]\r\n 1083 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/filebasedsink.py in close(self)\r\n 422 def close(self):\r\n--> 423 self.sink.close(self.temp_handle)\r\n 424 return self.temp_shard_path\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in close(self, writer)\r\n 537 if len(self._buffer[0]) > 0:\r\n--> 538 self._flush_buffer()\r\n 539 if self._record_batches_byte_size > 0:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in _flush_buffer(self)\r\n 569 for b in x.buffers():\r\n--> 570 size = size + b.size\r\n 571 self._record_batches_byte_size = self._record_batches_byte_size + size\r\n\r\nAttributeError: 'NoneType' object has no attribute 'size'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nAttributeError Traceback (most recent call last)\r\n\r\n<ipython-input-9-340aabccefff> in <module>()\r\n----> 1 dset = nlp.load_dataset('wikipedia')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 370 verify_infos = not save_infos and not ignore_verifications\r\n 371 self._download_and_prepare(\r\n--> 372 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 373 )\r\n 374 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos)\r\n 770 with beam.Pipeline(runner=beam_runner, options=beam_options,) as pipeline:\r\n 771 super(BeamBasedBuilder, self)._download_and_prepare(\r\n--> 772 dl_manager, pipeline=pipeline, verify_infos=False\r\n 773 ) # TODO{beam} verify infos\r\n 774 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in __exit__(self, exc_type, exc_val, exc_tb)\r\n 501 def __exit__(self, exc_type, exc_val, exc_tb):\r\n 502 if not exc_type:\r\n--> 503 self.run().wait_until_finish()\r\n 504 \r\n 505 def visit(self, visitor):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in run(self, test_runner_api)\r\n 481 return Pipeline.from_runner_api(\r\n 482 self.to_runner_api(use_fake_coders=True), self.runner,\r\n--> 483 self._options).run(False)\r\n 484 \r\n 485 if self._options.view_as(TypeOptions).runtime_type_check:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in run(self, test_runner_api)\r\n 494 finally:\r\n 495 shutil.rmtree(tmpdir)\r\n--> 496 return self.runner.run_pipeline(self, self._options)\r\n 497 \r\n 498 def __enter__(self):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/direct\/direct_runner.py in run_pipeline(self, pipeline, options)\r\n 128 runner = BundleBasedDirectRunner()\r\n 129 \r\n--> 130 return runner.run_pipeline(pipeline, options)\r\n 131 \r\n 132 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_pipeline(self, pipeline, options)\r\n 553 \r\n 554 self._latest_run_result = self.run_via_runner_api(\r\n--> 555 pipeline.to_runner_api(default_environment=self._default_environment))\r\n 556 return self._latest_run_result\r\n 557 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_via_runner_api(self, pipeline_proto)\r\n 563 # TODO(pabloem, BEAM-7514): Create a watermark manager (that has access to\r\n 564 # the teststream (if any), and all the stages).\r\n--> 565 return self.run_stages(stage_context, stages)\r\n 566 \r\n 567 @contextlib.contextmanager\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_stages(self, stage_context, stages)\r\n 704 stage,\r\n 705 pcoll_buffers,\r\n--> 706 stage_context.safe_coders)\r\n 707 metrics_by_stage[stage.name] = stage_results.process_bundle.metrics\r\n 708 monitoring_infos_by_stage[stage.name] = (\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in _run_stage(self, worker_handler_factory, pipeline_components, stage, pcoll_buffers, safe_coders)\r\n 1071 cache_token_generator=cache_token_generator)\r\n 1072 \r\n-> 1073 result, splits = bundle_manager.process_bundle(data_input, data_output)\r\n 1074 \r\n 1075 def input_for(transform_id, input_id):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in process_bundle(self, inputs, expected_outputs)\r\n 2332 \r\n 2333 with UnboundedThreadPoolExecutor() as executor:\r\n-> 2334 for result, split_result in executor.map(execute, part_inputs):\r\n 2335 \r\n 2336 split_result_list += split_result\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in result_iterator()\r\n 584 # Careful not to keep a reference to the popped future\r\n 585 if timeout is None:\r\n--> 586 yield fs.pop().result()\r\n 587 else:\r\n 588 yield fs.pop().result(end_time - time.monotonic())\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 430 raise CancelledError()\r\n 431 elif self._state == FINISHED:\r\n--> 432 return self.__get_result()\r\n 433 else:\r\n 434 raise TimeoutError()\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/utils\/thread_pool_executor.py in run(self)\r\n 42 # If the future wasn't cancelled, then attempt to execute it.\r\n 43 try:\r\n---> 44 self._future.set_result(self._fn(*self._fn_args, **self._fn_kwargs))\r\n 45 except BaseException as exc:\r\n 46 # Even though Python 2 futures library has #set_exection(),\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in execute(part_map)\r\n 2329 self._registered,\r\n 2330 cache_token_generator=self._cache_token_generator)\r\n-> 2331 return bundle_manager.process_bundle(part_map, expected_outputs)\r\n 2332 \r\n 2333 with UnboundedThreadPoolExecutor() as executor:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in process_bundle(self, inputs, expected_outputs)\r\n 2243 process_bundle_descriptor_id=self._bundle_descriptor.id,\r\n 2244 cache_tokens=[next(self._cache_token_generator)]))\r\n-> 2245 result_future = self._worker_handler.control_conn.push(process_bundle_req)\r\n 2246 \r\n 2247 split_results = [] # type: List[beam_fn_api_pb2.ProcessBundleSplitResponse]\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in push(self, request)\r\n 1557 self._uid_counter += 1\r\n 1558 request.instruction_id = 'control_%s' % self._uid_counter\r\n-> 1559 response = self.worker.do_instruction(request)\r\n 1560 return ControlFuture(request.instruction_id, response)\r\n 1561 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/sdk_worker.py in do_instruction(self, request)\r\n 413 # E.g. if register is set, this will call self.register(request.register))\r\n 414 return getattr(self, request_type)(\r\n--> 415 getattr(request, request_type), request.instruction_id)\r\n 416 else:\r\n 417 raise NotImplementedError\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/sdk_worker.py in process_bundle(self, request, instruction_id)\r\n 448 with self.maybe_profile(instruction_id):\r\n 449 delayed_applications, requests_finalization = (\r\n--> 450 bundle_processor.process_bundle(instruction_id))\r\n 451 monitoring_infos = bundle_processor.monitoring_infos()\r\n 452 monitoring_infos.extend(self.state_cache_metrics_fn())\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/bundle_processor.py in process_bundle(self, instruction_id)\r\n 837 for data in data_channel.input_elements(instruction_id,\r\n 838 expected_transforms):\r\n--> 839 input_op_by_transform_id[data.transform_id].process_encoded(data.data)\r\n 840 \r\n 841 # Finish all operations.\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/bundle_processor.py in process_encoded(self, encoded_windowed_values)\r\n 214 decoded_value = self.windowed_coder_impl.decode_from_stream(\r\n 215 input_stream, True)\r\n--> 216 self.output(decoded_value)\r\n 217 \r\n 218 def try_split(self, fraction_of_remainder, total_buffer_size):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.Operation.output()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.Operation.output()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.SingletonConsumerSet.receive()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.DoOperation.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.DoOperation.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner._reraise_augmented()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/future\/utils\/__init__.py in raise_with_traceback(exc, traceback)\r\n 417 if traceback == Ellipsis:\r\n 418 _, _, traceback = sys.exc_info()\r\n--> 419 raise exc.with_traceback(traceback)\r\n 420 \r\n 421 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker.invoke_process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/iobase.py in process(self, element, init_result)\r\n 1080 for e in bundle[1]: # values\r\n 1081 writer.write(e)\r\n-> 1082 return [window.TimestampedValue(writer.close(), timestamp.MAX_TIMESTAMP)]\r\n 1083 \r\n 1084 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/filebasedsink.py in close(self)\r\n 421 \r\n 422 def close(self):\r\n--> 423 self.sink.close(self.temp_handle)\r\n 424 return self.temp_shard_path\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in close(self, writer)\r\n 536 def close(self, writer):\r\n 537 if len(self._buffer[0]) > 0:\r\n--> 538 self._flush_buffer()\r\n 539 if self._record_batches_byte_size > 0:\r\n 540 self._write_batches(writer)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in _flush_buffer(self)\r\n 568 for x in arrays:\r\n 569 for b in x.buffers():\r\n--> 570 size = size + b.size\r\n 571 self._record_batches_byte_size = self._record_batches_byte_size + size\r\n\r\nAttributeError: 'NoneType' object has no attribute 'size' [while running 'train\/Save to parquet\/Write\/WriteImpl\/WriteBundles']\r\n```\nI have seen that somebody is hard working on easierly loadable wikipedia. #129 \r\nMaybe I should wait a few days for that version ?","embeddings":[-0.1516936719,0.0031176787,-0.0172203388,0.395411998,-0.0661803856,0.055734206,0.0433485359,0.3345056772,0.3488544822,0.116828151,0.3199312985,0.1732222289,-0.1478529274,-0.08189895,-0.0162129309,-0.4596623182,-0.0519881584,0.2463434488,0.0365840979,-0.055988159,-0.2965040207,0.2679029405,-0.2995700538,0.1040329561,-0.0625268295,-0.1476691663,0.1509711891,0.0239506122,-0.2917689383,-0.3756918311,0.0411315262,-0.2272832096,0.2764588892,0.1493850648,-0.0001126928,-0.033079192,0.6667644978,0.013357413,-0.4484111369,-0.3978943825,-0.1365267038,-0.4073244929,0.4514855444,-0.2783575654,0.0108114174,-0.183243528,0.3193400502,0.1798266172,0.4127219319,0.2974144518,0.2151003033,0.0132923853,0.3161814809,0.0289726257,0.4331317544,0.0234559532,0.0376815312,-0.082880117,-0.1319960505,0.0341566205,-0.1272097528,0.1512051821,-0.2554622293,0.085636802,0.3495388627,-0.1963473111,-0.0711795911,-0.5439116359,0.1323036849,0.1406851411,0.6836251616,-0.0462400094,0.1660724282,0.0973634869,0.1164916232,0.0578920729,0.2387089282,0.3242072165,-0.4263738096,-0.3079098165,-0.2179973125,-0.249674052,-0.2927875817,0.4647779167,0.0429484174,0.552410841,-0.0626783296,0.2276490927,-0.0418308824,-0.0239027701,-0.0688646659,-0.1292949915,0.0287907701,0.3013543785,0.0604953989,0.0935744867,0.0215015952,0.2295125425,0.1968696713,-0.0434167497,-0.1734355539,-0.1454145759,0.1463511586,0.1465433091,0.0333783031,0.2567222118,0.1367914528,-0.1662375182,0.1898491532,0.2154508084,-0.0659422502,0.087950483,-0.0317949057,-0.1787978262,-0.4661916196,0.0235311091,0.2609489262,-0.1763258427,-0.0362600125,0.0417041592,-0.425917685,-0.2416983396,-0.0293636434,0.3228060305,-0.2112709135,0.3525010645,0.3740545809,0.076737836,-0.3170039356,-0.3271934688,0.0201534927,0.29330194,-0.286726743,0.0849589035,0.2305870503,0.3047331572,0.3511444032,-0.0177782457,-0.077314876,0.00820887,0.1638403088,-0.0817262754,-0.2170658857,0.238672778,0.3142968714,0.2701243162,0.0584865883,-0.3325569332,-0.0798217133,0.2548536956,-0.33965832,-0.0750596598,-0.033942204,0.1641893834,-0.1057851315,0.0137594612,-0.3482524455,0.2475658506,0.0308313519,-0.2846438885,-0.1734679341,-0.1604102403,-0.0858308598,-0.3557700515,0.2626804113,0.3350428641,0.1315860003,-0.1066654027,-0.0884822458,0.1678850949,0.2250353396,0.0563847348,-0.1249937639,0.4726496041,-0.0444213226,-0.0787719712,0.742199719,-0.1341574937,-0.1724846959,0.2340096086,0.1318541765,0.0190685838,0.0614238754,-0.0625562072,0.1204571575,0.0860344321,-0.0515297279,0.5668154955,-0.0113387695,0.0343194455,-0.4098312557,-0.2514778972,0.1766582578,0.0684391409,0.3860545456,0.1507287621,0.0109391762,0.5279040337,0.0819583535,-0.0396940969,0.0264235213,0.3199964464,-0.2970614135,-0.0903743431,-0.0217857026,0.0699690059,-0.1751378328,-0.0254766699,-0.3948397934,0.1508505195,0.1584998071,0.106175594,-0.3540579379,-0.1429710537,-0.0524569601,-0.1548898518,0.1619729996,0.1855083555,0.0948660374,0.1506925076,0.0541367829,0.06682349,-0.1275253743,-0.04091952,-0.6120955944,0.1855074912,-0.273617208,-0.0768042579,0.1137746349,0.2102171928,0.2491545975,0.011265412,-0.1234422401,0.0068761823,-0.201151222,-0.0626368225,-0.0272328909,-0.0705084205,0.2022548467,-0.3002306819,0.3145110905,0.2933092713,0.1526011229,-0.132370472,0.0877334848,-0.0768909827,0.1184144169,0.340471983,0.074039489,0.103276886,0.0042449655,-0.0627106279,-0.0017371755,0.2277478278,0.3068399727,0.4291990101,0.0985184088,-0.1077882946,0.0011931499,-0.2093723416,0.3699370921,0.0447329767,0.0440538302,0.267177254,-0.4355993271,-0.0154325292,0.1121865287,-0.1628754288,0.1669237763,0.2249119431,-0.1555083841,-0.0398738533,-0.0813928023,-0.1273939759,0.2003766,0.1784821004,0.3696184754,0.1193412915,0.0897206366,0.0779439732,-0.0783375725,-0.1367147118,-0.1466954499,0.3170475364,-0.304097563,0.0782394558,-0.0234099664,-0.4767170846,-0.2430294603,0.0441244617,-0.4523335695,-0.3893648982,-0.0438593477,0.2663142383,0.1046174094,0.0952164233,0.3447324038,0.1438266933,-0.0165728834,-0.0947479755,-0.1618872583,-0.2388005406,-0.4626995325,0.0603272431,0.3643649817,0.2716529369,0.2625266314,-0.0358117782,0.0011127862,-0.0167188626,-0.2616984546,0.035663031,-0.0869233012,0.1974505633,-0.1058520079,0.4282002747,-0.067420736,-0.1935695857,0.2694040537,-0.035429474,-0.0946539193,0.0787811503,-0.1313625425,0.0198957752,-0.0187205374,-0.3938397765,-0.1960566044,-0.5722575784,0.0362823792,0.5017025471,0.1574628949,0.1577188671,0.2231185734,0.0792135075,0.3722141087,0.300367713,-0.0748495087,0.0828082561,0.3569961786,-0.2425844073,-0.453726083,0.006872945,-0.2535283864,0.2097843885,0.1102367938,-0.5408006907,-0.0630773604,-0.036176987,-0.0940902308,-0.0728066266,0.0922088176,0.4026173949,-0.0251933578,0.0642637908,0.0335506499,-0.0044596074,-0.1506370008,-0.2979058027,0.5397560596,0.3464972079,0.3857060075,-0.1238712668,0.9763130546,0.1193715855,-0.0108398749,0.2124872506,0.1578999907,0.0447257012,-0.2002599537,-0.2127510607,0.1421232671,-0.015238462,-0.0513092838,0.3709128797,-0.0905363485,-0.2564307153,-0.2897773385,-0.0581647195,-0.2621936798,-0.1228254884,0.2391225696,0.0429717563,0.1968317032,-0.1175461039,0.0472731665,-0.0609720498,-0.3704538047,0.0841386244,0.2635550201,-0.0596749224,0.0893104449,0.2353031188,-0.1747617722,-0.5001497269,0.4032238424,0.0295508914,0.0946017653,-0.232518062,-0.0259186719,0.1726958752,-0.043921534,0.2294401973,-0.1727760434,-0.1800782382,0.3131050169,0.2789322138,-0.4931188226,-0.0071605882,-0.2260558158,0.2164935768,0.3188700676,-0.0650155619,-0.4468346536,0.0118404608,0.3586078286,0.2625594437,-0.1611926407,-0.0678652152,-0.4241851568,-0.1054242998,-0.3580246568,-0.1601400673,-0.1432640851,0.3781589866,0.0143492268,0.0823294148,-0.0511637256,-0.075337261,0.0300968941,0.0565532334,0.1316145509,0.1576212645,-0.3129265904,-0.1773394942,0.257078588,-0.3667420447,0.2074061185,0.3398807049,0.0224398859,0.009034724,0.1605149209,0.1010871381,0.1211710796,0.0386613943,-0.0937231556,-0.0935021862,-0.0203376561,-0.1044694483,0.0488840565,0.2029829025,-0.1898672879,-0.1906206757,-0.1861124933,0.9091110826,0.1154260114,-0.084263809,0.3991340101,-0.1043891758,-0.3004438877,0.4643959701,0.2923505902,1.0337502956,-0.052130986,0.0741080195,0.3550423086,0.1453929842,0.5975183249,-0.5634702444,0.2468140721,-0.467713505,0.196641922,-0.0420756415,-0.0888727307,0.181854248,-0.0254621152,-0.4229598045,0.3079963624,0.2147316486,0.0531658866,-0.0045523536,0.4972426891,-0.0075476244,-0.2034151554,-0.2026691586,0.1187987775,-0.4527353644,0.4196880758,-0.3137229979,-0.0773909688,-0.0372701213,-0.266451031,-0.3212264776,0.2381910235,-0.2842641771,0.2605059743,-0.078880772,-0.4040437043,0.1364526153,0.2001174986,-0.0738272667,0.2218795568,-0.3675208986,0.3076855242,-0.3412058651,-0.3970908821,-0.1670372933,0.3038877547,0.3126798272,-0.2041235566,-0.0911179781,0.2697536349,-0.1574627012,-0.1934327781,-0.1162736937,-0.0458574593,0.4036424458,-0.0822805688,-0.2840816081,0.0886297822,-0.1174004674,0.1052743047,0.1288161874,-0.1578905284,-0.0238424763,-0.1117869169,0.4168770313,-0.1224572957,0.1530792266,0.2160123587,0.3512758017,-0.0435097553,0.6242598295,0.2489507943,-0.1267132163,-0.1933340877,-0.0821626782,-0.3634008765,-0.2203258723,-0.1326922625,0.0610433519,0.0260353405,-0.1617572755,0.3119337559,0.0457110554,0.1481468529,0.0854196176,-0.4704941511,-0.0857337192,0.3661580682,-0.2159737498,-0.0267136525,0.1270240545,-0.1144802347,0.4633326232,0.135235399,-0.251655668,-0.1968414187,-0.2700380981,0.2019437701,0.3335370719,-0.0315216184,-0.0963801742,0.1897742748,0.0387237296,-0.187765196,-0.1967302561,-0.1924377084,0.0781488568,0.1621488482,0.2096488029,-0.0580844805,-0.1004076973,-0.2537478507,0.0404575579,-0.1762682647,-0.1704919189,-0.2846744359,-0.1624414772,0.0790376738,-0.0626454353,0.2831457257,-0.391821295,0.2159593552,-0.1537296176,0.4537641406,0.1176498309,0.0420255847,0.5127453804,-0.0554911382,-0.7429888248,0.0311053991,0.0138802985,0.1119629219,0.2513465881,-0.1018091217,0.0759131983,-0.0500892848,-0.1866137385,0.0174274351,-0.1264566183,0.0703162849,0.113356024,0.1564352363,-0.1616566628,0.3631634414,0.3946045935,0.1727810502,-0.1709176749,0.078667067,0.0342517905,-0.0917056426,-0.0037510979,0.2260856479,-0.0059776963,-0.0557172783,0.0232091639,-0.127233237,-0.0253192931,-0.013457559,0.011188468,-0.0202513337,0.1281438768,0.1153361425,-0.1465821862,0.184061721,0.0543780252,0.006913282,0.0157424472,-0.1864059865,0.235301435,0.152033329,0.3711350858,0.0470636301,0.0498729907,0.0448101126,0.1176264957,-0.1444133222,-0.4065522254,0.1806305051,-0.1293535084,0.126243785,-0.4547826052,-0.075286299,-0.289144367,0.1415076703,0.1070522293,-0.2766278386,0.0571720414,0.2894271612,-0.3486205637,-0.4083663225,0.4397546053,0.1445158273,0.0271509122,-0.1534154415,0.3643988073,-0.1512813568,0.0181094352,-0.082647182,0.3836381137,0.3436254263,0.549742341,-0.4759600461,-0.0549185015,0.0556129329,-0.1505192965,-0.079155989,0.1190736517,0.4115951359,0.2340695113,0.1822738945,0.1373762935,-0.2314870656,-0.0316734016,0.1822689027,-0.105718337,-0.0826201364,0.1635469645,0.0025953457,-0.176818192,-0.2499957234,-0.0298536588,-0.6789048314,0.0847015828,-0.0178617705,0.193749398,-0.0992387012,-0.3172448277,0.0282293726,-0.1929134578,0.4531596005,0.3371498585,0.3244173825,-0.3690618277,-0.3166697323,-0.3890509009,0.1598980129,-0.3162232935,-0.32014817,-0.099747695,0.233864367,-0.1551486999,0.3409184813,-0.239705205,-0.0007633103,0.0764811337,0.0292470884,-0.2659039497,-0.1608331203,0.0497716404,-0.1193939671,-0.306185782,-0.1297397912,-0.0049936366,-0.2757493556,0.062485382,-0.0262239687,-0.2599042654,0.2788180709,0.0533178598,0.2739654481,0.0767083168,0.6266765594,-0.1028350592,-0.0684035644,-0.0048561031,-0.0697140396,-0.0635503978,0.3697392046,0.0368193388,0.5085330606,-0.0811913759,0.1783290952,-0.3596467078,0.4930325449,-0.2634870708,0.1496642977,-0.1138263345,-0.1301416159,-0.2357619703,0.3155688941,-0.0398642123,0.0872734264,-0.1160328463,0.0741733834,-0.4722552001,-0.2119027078,0.3164293766,-0.7915250063,-0.3263424635,0.1186925545,0.0936174616,-0.0843324438,0.0234400481,-0.5125064254,-0.0636687875,0.2648860216,-0.0716170967,-0.1045155451,0.0364237353,-0.1194525883,-0.0582114905,-0.1178621575,-0.229190886,-0.1291629523,-0.2262610346,-0.1860428452,-0.3009596765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/187","title":"[Question] How to load wikipedia ? Beam runner ?","comments":"Yes we (well @lhoestq) are very actively working on this.","body":"When `nlp.load_dataset('wikipedia')`, I got\r\n* `WARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be used.`\r\n* `AttributeError: 'NoneType' object has no attribute 'size'`\r\n\r\nCould somebody tell me what should I do ? \r\n\r\n# Env\r\nOn Colab,\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp\r\ncd nlp\r\npip install -q .\r\n```\r\n```\r\n%pip install -q apache_beam mwparserfromhell\r\n-> ERROR: pydrive 1.3.1 has requirement oauth2client>=4.0.0, but you'll have oauth2client 3.0.0 which is incompatible.\r\nERROR: google-api-python-client 1.7.12 has requirement httplib2<1dev,>=0.17.0, but you'll have httplib2 0.12.0 which is incompatible.\r\nERROR: chainer 6.5.0 has requirement typing-extensions<=3.6.6, but you'll have typing-extensions 3.7.4.2 which is incompatible.\r\n```\r\n```\r\npip install -q apache-beam[interactive]\r\nERROR: google-colab 1.0.0 has requirement ipython~=5.5.0, but you'll have ipython 5.10.0 which is incompatible.\r\n```\r\n\r\n# The whole message\r\n```\r\nWARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be used.\r\n\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wikipedia\/20200501.aa\/1.0.0...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nAttributeError Traceback (most recent call last)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n44 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker.invoke_process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/iobase.py in process(self, element, init_result)\r\n 1081 writer.write(e)\r\n-> 1082 return [window.TimestampedValue(writer.close(), timestamp.MAX_TIMESTAMP)]\r\n 1083 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/filebasedsink.py in close(self)\r\n 422 def close(self):\r\n--> 423 self.sink.close(self.temp_handle)\r\n 424 return self.temp_shard_path\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in close(self, writer)\r\n 537 if len(self._buffer[0]) > 0:\r\n--> 538 self._flush_buffer()\r\n 539 if self._record_batches_byte_size > 0:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in _flush_buffer(self)\r\n 569 for b in x.buffers():\r\n--> 570 size = size + b.size\r\n 571 self._record_batches_byte_size = self._record_batches_byte_size + size\r\n\r\nAttributeError: 'NoneType' object has no attribute 'size'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nAttributeError Traceback (most recent call last)\r\n\r\n<ipython-input-9-340aabccefff> in <module>()\r\n----> 1 dset = nlp.load_dataset('wikipedia')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 370 verify_infos = not save_infos and not ignore_verifications\r\n 371 self._download_and_prepare(\r\n--> 372 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 373 )\r\n 374 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos)\r\n 770 with beam.Pipeline(runner=beam_runner, options=beam_options,) as pipeline:\r\n 771 super(BeamBasedBuilder, self)._download_and_prepare(\r\n--> 772 dl_manager, pipeline=pipeline, verify_infos=False\r\n 773 ) # TODO{beam} verify infos\r\n 774 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in __exit__(self, exc_type, exc_val, exc_tb)\r\n 501 def __exit__(self, exc_type, exc_val, exc_tb):\r\n 502 if not exc_type:\r\n--> 503 self.run().wait_until_finish()\r\n 504 \r\n 505 def visit(self, visitor):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in run(self, test_runner_api)\r\n 481 return Pipeline.from_runner_api(\r\n 482 self.to_runner_api(use_fake_coders=True), self.runner,\r\n--> 483 self._options).run(False)\r\n 484 \r\n 485 if self._options.view_as(TypeOptions).runtime_type_check:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in run(self, test_runner_api)\r\n 494 finally:\r\n 495 shutil.rmtree(tmpdir)\r\n--> 496 return self.runner.run_pipeline(self, self._options)\r\n 497 \r\n 498 def __enter__(self):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/direct\/direct_runner.py in run_pipeline(self, pipeline, options)\r\n 128 runner = BundleBasedDirectRunner()\r\n 129 \r\n--> 130 return runner.run_pipeline(pipeline, options)\r\n 131 \r\n 132 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_pipeline(self, pipeline, options)\r\n 553 \r\n 554 self._latest_run_result = self.run_via_runner_api(\r\n--> 555 pipeline.to_runner_api(default_environment=self._default_environment))\r\n 556 return self._latest_run_result\r\n 557 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_via_runner_api(self, pipeline_proto)\r\n 563 # TODO(pabloem, BEAM-7514): Create a watermark manager (that has access to\r\n 564 # the teststream (if any), and all the stages).\r\n--> 565 return self.run_stages(stage_context, stages)\r\n 566 \r\n 567 @contextlib.contextmanager\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_stages(self, stage_context, stages)\r\n 704 stage,\r\n 705 pcoll_buffers,\r\n--> 706 stage_context.safe_coders)\r\n 707 metrics_by_stage[stage.name] = stage_results.process_bundle.metrics\r\n 708 monitoring_infos_by_stage[stage.name] = (\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in _run_stage(self, worker_handler_factory, pipeline_components, stage, pcoll_buffers, safe_coders)\r\n 1071 cache_token_generator=cache_token_generator)\r\n 1072 \r\n-> 1073 result, splits = bundle_manager.process_bundle(data_input, data_output)\r\n 1074 \r\n 1075 def input_for(transform_id, input_id):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in process_bundle(self, inputs, expected_outputs)\r\n 2332 \r\n 2333 with UnboundedThreadPoolExecutor() as executor:\r\n-> 2334 for result, split_result in executor.map(execute, part_inputs):\r\n 2335 \r\n 2336 split_result_list += split_result\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in result_iterator()\r\n 584 # Careful not to keep a reference to the popped future\r\n 585 if timeout is None:\r\n--> 586 yield fs.pop().result()\r\n 587 else:\r\n 588 yield fs.pop().result(end_time - time.monotonic())\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 430 raise CancelledError()\r\n 431 elif self._state == FINISHED:\r\n--> 432 return self.__get_result()\r\n 433 else:\r\n 434 raise TimeoutError()\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/utils\/thread_pool_executor.py in run(self)\r\n 42 # If the future wasn't cancelled, then attempt to execute it.\r\n 43 try:\r\n---> 44 self._future.set_result(self._fn(*self._fn_args, **self._fn_kwargs))\r\n 45 except BaseException as exc:\r\n 46 # Even though Python 2 futures library has #set_exection(),\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in execute(part_map)\r\n 2329 self._registered,\r\n 2330 cache_token_generator=self._cache_token_generator)\r\n-> 2331 return bundle_manager.process_bundle(part_map, expected_outputs)\r\n 2332 \r\n 2333 with UnboundedThreadPoolExecutor() as executor:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in process_bundle(self, inputs, expected_outputs)\r\n 2243 process_bundle_descriptor_id=self._bundle_descriptor.id,\r\n 2244 cache_tokens=[next(self._cache_token_generator)]))\r\n-> 2245 result_future = self._worker_handler.control_conn.push(process_bundle_req)\r\n 2246 \r\n 2247 split_results = [] # type: List[beam_fn_api_pb2.ProcessBundleSplitResponse]\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in push(self, request)\r\n 1557 self._uid_counter += 1\r\n 1558 request.instruction_id = 'control_%s' % self._uid_counter\r\n-> 1559 response = self.worker.do_instruction(request)\r\n 1560 return ControlFuture(request.instruction_id, response)\r\n 1561 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/sdk_worker.py in do_instruction(self, request)\r\n 413 # E.g. if register is set, this will call self.register(request.register))\r\n 414 return getattr(self, request_type)(\r\n--> 415 getattr(request, request_type), request.instruction_id)\r\n 416 else:\r\n 417 raise NotImplementedError\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/sdk_worker.py in process_bundle(self, request, instruction_id)\r\n 448 with self.maybe_profile(instruction_id):\r\n 449 delayed_applications, requests_finalization = (\r\n--> 450 bundle_processor.process_bundle(instruction_id))\r\n 451 monitoring_infos = bundle_processor.monitoring_infos()\r\n 452 monitoring_infos.extend(self.state_cache_metrics_fn())\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/bundle_processor.py in process_bundle(self, instruction_id)\r\n 837 for data in data_channel.input_elements(instruction_id,\r\n 838 expected_transforms):\r\n--> 839 input_op_by_transform_id[data.transform_id].process_encoded(data.data)\r\n 840 \r\n 841 # Finish all operations.\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/bundle_processor.py in process_encoded(self, encoded_windowed_values)\r\n 214 decoded_value = self.windowed_coder_impl.decode_from_stream(\r\n 215 input_stream, True)\r\n--> 216 self.output(decoded_value)\r\n 217 \r\n 218 def try_split(self, fraction_of_remainder, total_buffer_size):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.Operation.output()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.Operation.output()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.SingletonConsumerSet.receive()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.DoOperation.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.DoOperation.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner._reraise_augmented()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/future\/utils\/__init__.py in raise_with_traceback(exc, traceback)\r\n 417 if traceback == Ellipsis:\r\n 418 _, _, traceback = sys.exc_info()\r\n--> 419 raise exc.with_traceback(traceback)\r\n 420 \r\n 421 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker.invoke_process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/iobase.py in process(self, element, init_result)\r\n 1080 for e in bundle[1]: # values\r\n 1081 writer.write(e)\r\n-> 1082 return [window.TimestampedValue(writer.close(), timestamp.MAX_TIMESTAMP)]\r\n 1083 \r\n 1084 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/filebasedsink.py in close(self)\r\n 421 \r\n 422 def close(self):\r\n--> 423 self.sink.close(self.temp_handle)\r\n 424 return self.temp_shard_path\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in close(self, writer)\r\n 536 def close(self, writer):\r\n 537 if len(self._buffer[0]) > 0:\r\n--> 538 self._flush_buffer()\r\n 539 if self._record_batches_byte_size > 0:\r\n 540 self._write_batches(writer)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in _flush_buffer(self)\r\n 568 for x in arrays:\r\n 569 for b in x.buffers():\r\n--> 570 size = size + b.size\r\n 571 self._record_batches_byte_size = self._record_batches_byte_size + size\r\n\r\nAttributeError: 'NoneType' object has no attribute 'size' [while running 'train\/Save to parquet\/Write\/WriteImpl\/WriteBundles']\r\n```","comment_length":10,"text":"[Question] How to load wikipedia ? Beam runner ?\nWhen `nlp.load_dataset('wikipedia')`, I got\r\n* `WARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be used.`\r\n* `AttributeError: 'NoneType' object has no attribute 'size'`\r\n\r\nCould somebody tell me what should I do ? \r\n\r\n# Env\r\nOn Colab,\r\n```\r\ngit clone https:\/\/github.com\/huggingface\/nlp\r\ncd nlp\r\npip install -q .\r\n```\r\n```\r\n%pip install -q apache_beam mwparserfromhell\r\n-> ERROR: pydrive 1.3.1 has requirement oauth2client>=4.0.0, but you'll have oauth2client 3.0.0 which is incompatible.\r\nERROR: google-api-python-client 1.7.12 has requirement httplib2<1dev,>=0.17.0, but you'll have httplib2 0.12.0 which is incompatible.\r\nERROR: chainer 6.5.0 has requirement typing-extensions<=3.6.6, but you'll have typing-extensions 3.7.4.2 which is incompatible.\r\n```\r\n```\r\npip install -q apache-beam[interactive]\r\nERROR: google-colab 1.0.0 has requirement ipython~=5.5.0, but you'll have ipython 5.10.0 which is incompatible.\r\n```\r\n\r\n# The whole message\r\n```\r\nWARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be used.\r\n\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wikipedia\/20200501.aa\/1.0.0...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nAttributeError Traceback (most recent call last)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n44 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker.invoke_process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/iobase.py in process(self, element, init_result)\r\n 1081 writer.write(e)\r\n-> 1082 return [window.TimestampedValue(writer.close(), timestamp.MAX_TIMESTAMP)]\r\n 1083 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/filebasedsink.py in close(self)\r\n 422 def close(self):\r\n--> 423 self.sink.close(self.temp_handle)\r\n 424 return self.temp_shard_path\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in close(self, writer)\r\n 537 if len(self._buffer[0]) > 0:\r\n--> 538 self._flush_buffer()\r\n 539 if self._record_batches_byte_size > 0:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in _flush_buffer(self)\r\n 569 for b in x.buffers():\r\n--> 570 size = size + b.size\r\n 571 self._record_batches_byte_size = self._record_batches_byte_size + size\r\n\r\nAttributeError: 'NoneType' object has no attribute 'size'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nAttributeError Traceback (most recent call last)\r\n\r\n<ipython-input-9-340aabccefff> in <module>()\r\n----> 1 dset = nlp.load_dataset('wikipedia')\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 370 verify_infos = not save_infos and not ignore_verifications\r\n 371 self._download_and_prepare(\r\n--> 372 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 373 )\r\n 374 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos)\r\n 770 with beam.Pipeline(runner=beam_runner, options=beam_options,) as pipeline:\r\n 771 super(BeamBasedBuilder, self)._download_and_prepare(\r\n--> 772 dl_manager, pipeline=pipeline, verify_infos=False\r\n 773 ) # TODO{beam} verify infos\r\n 774 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in __exit__(self, exc_type, exc_val, exc_tb)\r\n 501 def __exit__(self, exc_type, exc_val, exc_tb):\r\n 502 if not exc_type:\r\n--> 503 self.run().wait_until_finish()\r\n 504 \r\n 505 def visit(self, visitor):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in run(self, test_runner_api)\r\n 481 return Pipeline.from_runner_api(\r\n 482 self.to_runner_api(use_fake_coders=True), self.runner,\r\n--> 483 self._options).run(False)\r\n 484 \r\n 485 if self._options.view_as(TypeOptions).runtime_type_check:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/pipeline.py in run(self, test_runner_api)\r\n 494 finally:\r\n 495 shutil.rmtree(tmpdir)\r\n--> 496 return self.runner.run_pipeline(self, self._options)\r\n 497 \r\n 498 def __enter__(self):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/direct\/direct_runner.py in run_pipeline(self, pipeline, options)\r\n 128 runner = BundleBasedDirectRunner()\r\n 129 \r\n--> 130 return runner.run_pipeline(pipeline, options)\r\n 131 \r\n 132 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_pipeline(self, pipeline, options)\r\n 553 \r\n 554 self._latest_run_result = self.run_via_runner_api(\r\n--> 555 pipeline.to_runner_api(default_environment=self._default_environment))\r\n 556 return self._latest_run_result\r\n 557 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_via_runner_api(self, pipeline_proto)\r\n 563 # TODO(pabloem, BEAM-7514): Create a watermark manager (that has access to\r\n 564 # the teststream (if any), and all the stages).\r\n--> 565 return self.run_stages(stage_context, stages)\r\n 566 \r\n 567 @contextlib.contextmanager\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in run_stages(self, stage_context, stages)\r\n 704 stage,\r\n 705 pcoll_buffers,\r\n--> 706 stage_context.safe_coders)\r\n 707 metrics_by_stage[stage.name] = stage_results.process_bundle.metrics\r\n 708 monitoring_infos_by_stage[stage.name] = (\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in _run_stage(self, worker_handler_factory, pipeline_components, stage, pcoll_buffers, safe_coders)\r\n 1071 cache_token_generator=cache_token_generator)\r\n 1072 \r\n-> 1073 result, splits = bundle_manager.process_bundle(data_input, data_output)\r\n 1074 \r\n 1075 def input_for(transform_id, input_id):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in process_bundle(self, inputs, expected_outputs)\r\n 2332 \r\n 2333 with UnboundedThreadPoolExecutor() as executor:\r\n-> 2334 for result, split_result in executor.map(execute, part_inputs):\r\n 2335 \r\n 2336 split_result_list += split_result\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in result_iterator()\r\n 584 # Careful not to keep a reference to the popped future\r\n 585 if timeout is None:\r\n--> 586 yield fs.pop().result()\r\n 587 else:\r\n 588 yield fs.pop().result(end_time - time.monotonic())\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in result(self, timeout)\r\n 430 raise CancelledError()\r\n 431 elif self._state == FINISHED:\r\n--> 432 return self.__get_result()\r\n 433 else:\r\n 434 raise TimeoutError()\r\n\r\n\/usr\/lib\/python3.6\/concurrent\/futures\/_base.py in __get_result(self)\r\n 382 def __get_result(self):\r\n 383 if self._exception:\r\n--> 384 raise self._exception\r\n 385 else:\r\n 386 return self._result\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/utils\/thread_pool_executor.py in run(self)\r\n 42 # If the future wasn't cancelled, then attempt to execute it.\r\n 43 try:\r\n---> 44 self._future.set_result(self._fn(*self._fn_args, **self._fn_kwargs))\r\n 45 except BaseException as exc:\r\n 46 # Even though Python 2 futures library has #set_exection(),\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in execute(part_map)\r\n 2329 self._registered,\r\n 2330 cache_token_generator=self._cache_token_generator)\r\n-> 2331 return bundle_manager.process_bundle(part_map, expected_outputs)\r\n 2332 \r\n 2333 with UnboundedThreadPoolExecutor() as executor:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in process_bundle(self, inputs, expected_outputs)\r\n 2243 process_bundle_descriptor_id=self._bundle_descriptor.id,\r\n 2244 cache_tokens=[next(self._cache_token_generator)]))\r\n-> 2245 result_future = self._worker_handler.control_conn.push(process_bundle_req)\r\n 2246 \r\n 2247 split_results = [] # type: List[beam_fn_api_pb2.ProcessBundleSplitResponse]\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/portability\/fn_api_runner.py in push(self, request)\r\n 1557 self._uid_counter += 1\r\n 1558 request.instruction_id = 'control_%s' % self._uid_counter\r\n-> 1559 response = self.worker.do_instruction(request)\r\n 1560 return ControlFuture(request.instruction_id, response)\r\n 1561 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/sdk_worker.py in do_instruction(self, request)\r\n 413 # E.g. if register is set, this will call self.register(request.register))\r\n 414 return getattr(self, request_type)(\r\n--> 415 getattr(request, request_type), request.instruction_id)\r\n 416 else:\r\n 417 raise NotImplementedError\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/sdk_worker.py in process_bundle(self, request, instruction_id)\r\n 448 with self.maybe_profile(instruction_id):\r\n 449 delayed_applications, requests_finalization = (\r\n--> 450 bundle_processor.process_bundle(instruction_id))\r\n 451 monitoring_infos = bundle_processor.monitoring_infos()\r\n 452 monitoring_infos.extend(self.state_cache_metrics_fn())\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/bundle_processor.py in process_bundle(self, instruction_id)\r\n 837 for data in data_channel.input_elements(instruction_id,\r\n 838 expected_transforms):\r\n--> 839 input_op_by_transform_id[data.transform_id].process_encoded(data.data)\r\n 840 \r\n 841 # Finish all operations.\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/bundle_processor.py in process_encoded(self, encoded_windowed_values)\r\n 214 decoded_value = self.windowed_coder_impl.decode_from_stream(\r\n 215 input_stream, True)\r\n--> 216 self.output(decoded_value)\r\n 217 \r\n 218 def try_split(self, fraction_of_remainder, total_buffer_size):\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.Operation.output()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.Operation.output()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.SingletonConsumerSet.receive()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.DoOperation.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/worker\/operations.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.worker.operations.DoOperation.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner._reraise_augmented()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/future\/utils\/__init__.py in raise_with_traceback(exc, traceback)\r\n 417 if traceback == Ellipsis:\r\n 418 _, _, traceback = sys.exc_info()\r\n--> 419 raise exc.with_traceback(traceback)\r\n 420 \r\n 421 else:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.DoFnRunner.process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker.invoke_process()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/runners\/common.cpython-36m-x86_64-linux-gnu.so in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window()\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/iobase.py in process(self, element, init_result)\r\n 1080 for e in bundle[1]: # values\r\n 1081 writer.write(e)\r\n-> 1082 return [window.TimestampedValue(writer.close(), timestamp.MAX_TIMESTAMP)]\r\n 1083 \r\n 1084 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/filebasedsink.py in close(self)\r\n 421 \r\n 422 def close(self):\r\n--> 423 self.sink.close(self.temp_handle)\r\n 424 return self.temp_shard_path\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in close(self, writer)\r\n 536 def close(self, writer):\r\n 537 if len(self._buffer[0]) > 0:\r\n--> 538 self._flush_buffer()\r\n 539 if self._record_batches_byte_size > 0:\r\n 540 self._write_batches(writer)\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/apache_beam\/io\/parquetio.py in _flush_buffer(self)\r\n 568 for x in arrays:\r\n 569 for b in x.buffers():\r\n--> 570 size = size + b.size\r\n 571 self._record_batches_byte_size = self._record_batches_byte_size + size\r\n\r\nAttributeError: 'NoneType' object has no attribute 'size' [while running 'train\/Save to parquet\/Write\/WriteImpl\/WriteBundles']\r\n```\nYes we (well @lhoestq) are very actively working on this.","embeddings":[-0.1516936719,0.0031176787,-0.0172203388,0.395411998,-0.0661803856,0.055734206,0.0433485359,0.3345056772,0.3488544822,0.116828151,0.3199312985,0.1732222289,-0.1478529274,-0.08189895,-0.0162129309,-0.4596623182,-0.0519881584,0.2463434488,0.0365840979,-0.055988159,-0.2965040207,0.2679029405,-0.2995700538,0.1040329561,-0.0625268295,-0.1476691663,0.1509711891,0.0239506122,-0.2917689383,-0.3756918311,0.0411315262,-0.2272832096,0.2764588892,0.1493850648,-0.0001126928,-0.033079192,0.6667644978,0.013357413,-0.4484111369,-0.3978943825,-0.1365267038,-0.4073244929,0.4514855444,-0.2783575654,0.0108114174,-0.183243528,0.3193400502,0.1798266172,0.4127219319,0.2974144518,0.2151003033,0.0132923853,0.3161814809,0.0289726257,0.4331317544,0.0234559532,0.0376815312,-0.082880117,-0.1319960505,0.0341566205,-0.1272097528,0.1512051821,-0.2554622293,0.085636802,0.3495388627,-0.1963473111,-0.0711795911,-0.5439116359,0.1323036849,0.1406851411,0.6836251616,-0.0462400094,0.1660724282,0.0973634869,0.1164916232,0.0578920729,0.2387089282,0.3242072165,-0.4263738096,-0.3079098165,-0.2179973125,-0.249674052,-0.2927875817,0.4647779167,0.0429484174,0.552410841,-0.0626783296,0.2276490927,-0.0418308824,-0.0239027701,-0.0688646659,-0.1292949915,0.0287907701,0.3013543785,0.0604953989,0.0935744867,0.0215015952,0.2295125425,0.1968696713,-0.0434167497,-0.1734355539,-0.1454145759,0.1463511586,0.1465433091,0.0333783031,0.2567222118,0.1367914528,-0.1662375182,0.1898491532,0.2154508084,-0.0659422502,0.087950483,-0.0317949057,-0.1787978262,-0.4661916196,0.0235311091,0.2609489262,-0.1763258427,-0.0362600125,0.0417041592,-0.425917685,-0.2416983396,-0.0293636434,0.3228060305,-0.2112709135,0.3525010645,0.3740545809,0.076737836,-0.3170039356,-0.3271934688,0.0201534927,0.29330194,-0.286726743,0.0849589035,0.2305870503,0.3047331572,0.3511444032,-0.0177782457,-0.077314876,0.00820887,0.1638403088,-0.0817262754,-0.2170658857,0.238672778,0.3142968714,0.2701243162,0.0584865883,-0.3325569332,-0.0798217133,0.2548536956,-0.33965832,-0.0750596598,-0.033942204,0.1641893834,-0.1057851315,0.0137594612,-0.3482524455,0.2475658506,0.0308313519,-0.2846438885,-0.1734679341,-0.1604102403,-0.0858308598,-0.3557700515,0.2626804113,0.3350428641,0.1315860003,-0.1066654027,-0.0884822458,0.1678850949,0.2250353396,0.0563847348,-0.1249937639,0.4726496041,-0.0444213226,-0.0787719712,0.742199719,-0.1341574937,-0.1724846959,0.2340096086,0.1318541765,0.0190685838,0.0614238754,-0.0625562072,0.1204571575,0.0860344321,-0.0515297279,0.5668154955,-0.0113387695,0.0343194455,-0.4098312557,-0.2514778972,0.1766582578,0.0684391409,0.3860545456,0.1507287621,0.0109391762,0.5279040337,0.0819583535,-0.0396940969,0.0264235213,0.3199964464,-0.2970614135,-0.0903743431,-0.0217857026,0.0699690059,-0.1751378328,-0.0254766699,-0.3948397934,0.1508505195,0.1584998071,0.106175594,-0.3540579379,-0.1429710537,-0.0524569601,-0.1548898518,0.1619729996,0.1855083555,0.0948660374,0.1506925076,0.0541367829,0.06682349,-0.1275253743,-0.04091952,-0.6120955944,0.1855074912,-0.273617208,-0.0768042579,0.1137746349,0.2102171928,0.2491545975,0.011265412,-0.1234422401,0.0068761823,-0.201151222,-0.0626368225,-0.0272328909,-0.0705084205,0.2022548467,-0.3002306819,0.3145110905,0.2933092713,0.1526011229,-0.132370472,0.0877334848,-0.0768909827,0.1184144169,0.340471983,0.074039489,0.103276886,0.0042449655,-0.0627106279,-0.0017371755,0.2277478278,0.3068399727,0.4291990101,0.0985184088,-0.1077882946,0.0011931499,-0.2093723416,0.3699370921,0.0447329767,0.0440538302,0.267177254,-0.4355993271,-0.0154325292,0.1121865287,-0.1628754288,0.1669237763,0.2249119431,-0.1555083841,-0.0398738533,-0.0813928023,-0.1273939759,0.2003766,0.1784821004,0.3696184754,0.1193412915,0.0897206366,0.0779439732,-0.0783375725,-0.1367147118,-0.1466954499,0.3170475364,-0.304097563,0.0782394558,-0.0234099664,-0.4767170846,-0.2430294603,0.0441244617,-0.4523335695,-0.3893648982,-0.0438593477,0.2663142383,0.1046174094,0.0952164233,0.3447324038,0.1438266933,-0.0165728834,-0.0947479755,-0.1618872583,-0.2388005406,-0.4626995325,0.0603272431,0.3643649817,0.2716529369,0.2625266314,-0.0358117782,0.0011127862,-0.0167188626,-0.2616984546,0.035663031,-0.0869233012,0.1974505633,-0.1058520079,0.4282002747,-0.067420736,-0.1935695857,0.2694040537,-0.035429474,-0.0946539193,0.0787811503,-0.1313625425,0.0198957752,-0.0187205374,-0.3938397765,-0.1960566044,-0.5722575784,0.0362823792,0.5017025471,0.1574628949,0.1577188671,0.2231185734,0.0792135075,0.3722141087,0.300367713,-0.0748495087,0.0828082561,0.3569961786,-0.2425844073,-0.453726083,0.006872945,-0.2535283864,0.2097843885,0.1102367938,-0.5408006907,-0.0630773604,-0.036176987,-0.0940902308,-0.0728066266,0.0922088176,0.4026173949,-0.0251933578,0.0642637908,0.0335506499,-0.0044596074,-0.1506370008,-0.2979058027,0.5397560596,0.3464972079,0.3857060075,-0.1238712668,0.9763130546,0.1193715855,-0.0108398749,0.2124872506,0.1578999907,0.0447257012,-0.2002599537,-0.2127510607,0.1421232671,-0.015238462,-0.0513092838,0.3709128797,-0.0905363485,-0.2564307153,-0.2897773385,-0.0581647195,-0.2621936798,-0.1228254884,0.2391225696,0.0429717563,0.1968317032,-0.1175461039,0.0472731665,-0.0609720498,-0.3704538047,0.0841386244,0.2635550201,-0.0596749224,0.0893104449,0.2353031188,-0.1747617722,-0.5001497269,0.4032238424,0.0295508914,0.0946017653,-0.232518062,-0.0259186719,0.1726958752,-0.043921534,0.2294401973,-0.1727760434,-0.1800782382,0.3131050169,0.2789322138,-0.4931188226,-0.0071605882,-0.2260558158,0.2164935768,0.3188700676,-0.0650155619,-0.4468346536,0.0118404608,0.3586078286,0.2625594437,-0.1611926407,-0.0678652152,-0.4241851568,-0.1054242998,-0.3580246568,-0.1601400673,-0.1432640851,0.3781589866,0.0143492268,0.0823294148,-0.0511637256,-0.075337261,0.0300968941,0.0565532334,0.1316145509,0.1576212645,-0.3129265904,-0.1773394942,0.257078588,-0.3667420447,0.2074061185,0.3398807049,0.0224398859,0.009034724,0.1605149209,0.1010871381,0.1211710796,0.0386613943,-0.0937231556,-0.0935021862,-0.0203376561,-0.1044694483,0.0488840565,0.2029829025,-0.1898672879,-0.1906206757,-0.1861124933,0.9091110826,0.1154260114,-0.084263809,0.3991340101,-0.1043891758,-0.3004438877,0.4643959701,0.2923505902,1.0337502956,-0.052130986,0.0741080195,0.3550423086,0.1453929842,0.5975183249,-0.5634702444,0.2468140721,-0.467713505,0.196641922,-0.0420756415,-0.0888727307,0.181854248,-0.0254621152,-0.4229598045,0.3079963624,0.2147316486,0.0531658866,-0.0045523536,0.4972426891,-0.0075476244,-0.2034151554,-0.2026691586,0.1187987775,-0.4527353644,0.4196880758,-0.3137229979,-0.0773909688,-0.0372701213,-0.266451031,-0.3212264776,0.2381910235,-0.2842641771,0.2605059743,-0.078880772,-0.4040437043,0.1364526153,0.2001174986,-0.0738272667,0.2218795568,-0.3675208986,0.3076855242,-0.3412058651,-0.3970908821,-0.1670372933,0.3038877547,0.3126798272,-0.2041235566,-0.0911179781,0.2697536349,-0.1574627012,-0.1934327781,-0.1162736937,-0.0458574593,0.4036424458,-0.0822805688,-0.2840816081,0.0886297822,-0.1174004674,0.1052743047,0.1288161874,-0.1578905284,-0.0238424763,-0.1117869169,0.4168770313,-0.1224572957,0.1530792266,0.2160123587,0.3512758017,-0.0435097553,0.6242598295,0.2489507943,-0.1267132163,-0.1933340877,-0.0821626782,-0.3634008765,-0.2203258723,-0.1326922625,0.0610433519,0.0260353405,-0.1617572755,0.3119337559,0.0457110554,0.1481468529,0.0854196176,-0.4704941511,-0.0857337192,0.3661580682,-0.2159737498,-0.0267136525,0.1270240545,-0.1144802347,0.4633326232,0.135235399,-0.251655668,-0.1968414187,-0.2700380981,0.2019437701,0.3335370719,-0.0315216184,-0.0963801742,0.1897742748,0.0387237296,-0.187765196,-0.1967302561,-0.1924377084,0.0781488568,0.1621488482,0.2096488029,-0.0580844805,-0.1004076973,-0.2537478507,0.0404575579,-0.1762682647,-0.1704919189,-0.2846744359,-0.1624414772,0.0790376738,-0.0626454353,0.2831457257,-0.391821295,0.2159593552,-0.1537296176,0.4537641406,0.1176498309,0.0420255847,0.5127453804,-0.0554911382,-0.7429888248,0.0311053991,0.0138802985,0.1119629219,0.2513465881,-0.1018091217,0.0759131983,-0.0500892848,-0.1866137385,0.0174274351,-0.1264566183,0.0703162849,0.113356024,0.1564352363,-0.1616566628,0.3631634414,0.3946045935,0.1727810502,-0.1709176749,0.078667067,0.0342517905,-0.0917056426,-0.0037510979,0.2260856479,-0.0059776963,-0.0557172783,0.0232091639,-0.127233237,-0.0253192931,-0.013457559,0.011188468,-0.0202513337,0.1281438768,0.1153361425,-0.1465821862,0.184061721,0.0543780252,0.006913282,0.0157424472,-0.1864059865,0.235301435,0.152033329,0.3711350858,0.0470636301,0.0498729907,0.0448101126,0.1176264957,-0.1444133222,-0.4065522254,0.1806305051,-0.1293535084,0.126243785,-0.4547826052,-0.075286299,-0.289144367,0.1415076703,0.1070522293,-0.2766278386,0.0571720414,0.2894271612,-0.3486205637,-0.4083663225,0.4397546053,0.1445158273,0.0271509122,-0.1534154415,0.3643988073,-0.1512813568,0.0181094352,-0.082647182,0.3836381137,0.3436254263,0.549742341,-0.4759600461,-0.0549185015,0.0556129329,-0.1505192965,-0.079155989,0.1190736517,0.4115951359,0.2340695113,0.1822738945,0.1373762935,-0.2314870656,-0.0316734016,0.1822689027,-0.105718337,-0.0826201364,0.1635469645,0.0025953457,-0.176818192,-0.2499957234,-0.0298536588,-0.6789048314,0.0847015828,-0.0178617705,0.193749398,-0.0992387012,-0.3172448277,0.0282293726,-0.1929134578,0.4531596005,0.3371498585,0.3244173825,-0.3690618277,-0.3166697323,-0.3890509009,0.1598980129,-0.3162232935,-0.32014817,-0.099747695,0.233864367,-0.1551486999,0.3409184813,-0.239705205,-0.0007633103,0.0764811337,0.0292470884,-0.2659039497,-0.1608331203,0.0497716404,-0.1193939671,-0.306185782,-0.1297397912,-0.0049936366,-0.2757493556,0.062485382,-0.0262239687,-0.2599042654,0.2788180709,0.0533178598,0.2739654481,0.0767083168,0.6266765594,-0.1028350592,-0.0684035644,-0.0048561031,-0.0697140396,-0.0635503978,0.3697392046,0.0368193388,0.5085330606,-0.0811913759,0.1783290952,-0.3596467078,0.4930325449,-0.2634870708,0.1496642977,-0.1138263345,-0.1301416159,-0.2357619703,0.3155688941,-0.0398642123,0.0872734264,-0.1160328463,0.0741733834,-0.4722552001,-0.2119027078,0.3164293766,-0.7915250063,-0.3263424635,0.1186925545,0.0936174616,-0.0843324438,0.0234400481,-0.5125064254,-0.0636687875,0.2648860216,-0.0716170967,-0.1045155451,0.0364237353,-0.1194525883,-0.0582114905,-0.1178621575,-0.229190886,-0.1291629523,-0.2262610346,-0.1860428452,-0.3009596765]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/186","title":"Weird-ish: Not creating unique caches for different phases","comments":"Looks like a duplicate of #120.\r\nThis is already fixed on master. We'll do a new release on pypi soon","body":"Sample code:\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load_dataset('boolq')\r\n\r\ndef func1(x):\r\n return x\r\n\r\ndef func2(x):\r\n return None\r\n\r\ntrain_output = dataset[\"train\"].map(func1)\r\nvalid_output = dataset[\"validation\"].map(func1)\r\nprint()\r\nprint(len(train_output), len(valid_output))\r\n# Output: 9427 9427\r\n```\r\n\r\nThe map method in both cases seem to be pointing to the same cache, so the latter call based on the validation data will return the processed train data cache.\r\n\r\nWhat's weird is that the following doesn't seem to be an issue:\r\n\r\n```python\r\ntrain_output = dataset[\"train\"].map(func2)\r\nvalid_output = dataset[\"validation\"].map(func2)\r\nprint()\r\nprint(len(train_output), len(valid_output))\r\n# 9427 3270\r\n```","comment_length":20,"text":"Weird-ish: Not creating unique caches for different phases\nSample code:\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load_dataset('boolq')\r\n\r\ndef func1(x):\r\n return x\r\n\r\ndef func2(x):\r\n return None\r\n\r\ntrain_output = dataset[\"train\"].map(func1)\r\nvalid_output = dataset[\"validation\"].map(func1)\r\nprint()\r\nprint(len(train_output), len(valid_output))\r\n# Output: 9427 9427\r\n```\r\n\r\nThe map method in both cases seem to be pointing to the same cache, so the latter call based on the validation data will return the processed train data cache.\r\n\r\nWhat's weird is that the following doesn't seem to be an issue:\r\n\r\n```python\r\ntrain_output = dataset[\"train\"].map(func2)\r\nvalid_output = dataset[\"validation\"].map(func2)\r\nprint()\r\nprint(len(train_output), len(valid_output))\r\n# 9427 3270\r\n```\nLooks like a duplicate of #120.\r\nThis is already fixed on master. We'll do a new release on pypi soon","embeddings":[0.3151272535,-0.1353009492,-0.0059518418,0.2090219706,0.0983513445,-0.1870373935,0.2608934641,0.326993674,-0.1120346189,-0.1279544383,-0.1073975712,0.4277701974,0.0819041356,-0.2038931996,0.0939095169,0.0684093907,0.1656414419,0.1147565693,-0.1145209223,-0.0431628749,-0.1762373596,0.1024518609,-0.0897262171,-0.1124547198,-0.4616794586,0.1460487545,-0.2929142416,-0.1581558585,0.2172628641,-0.3920987844,0.3604982197,0.0784055144,-0.2556280494,0.264415741,-0.0001164348,-0.0394376367,0.1898856312,-0.0604702123,-0.09947294,0.1508262753,0.0201879684,-0.1497223973,0.1603352427,-0.3118847013,-0.0166392494,0.1147761196,0.0591147095,0.1327996701,0.5603545904,0.0672881752,0.1915287673,0.2768823206,-0.3045631349,0.1994510591,0.158294782,-0.092304185,-0.1966276914,-0.0765136629,0.0185283851,-0.2760599852,-0.034622632,0.328553915,-0.0990364999,0.0971185118,-0.0188813191,0.0933788642,0.1654084176,-0.1915283799,0.0857186466,0.063975133,-0.1246067584,-0.3299968541,-0.228296414,-0.3659480512,-0.2273795456,-0.3725856245,0.1338074058,0.1214887798,-0.1510031819,-0.1150846258,-0.2887341678,0.2583762109,0.1800031066,0.0654662102,0.1485068798,0.3490792513,0.1705223769,0.0039953599,0.2054348439,0.1646481156,-0.0229042117,-0.1676710844,0.0393320397,0.3825148046,-0.3286457658,-0.1350692809,0.2580424845,-0.0376147553,-0.020884132,-0.0351710245,0.2935594618,-0.0580629595,0.0300326552,0.0378095396,0.1459481269,0.4408579469,-0.0584590062,0.2622619271,0.036314521,-0.1119010076,-0.4782032073,0.0715369284,0.3043486476,0.0876684859,0.3045755923,0.0555901378,-0.0760893896,-0.0986751765,0.0427362807,0.0919156,-0.3604730964,-0.1665520668,0.0415167473,0.1238828078,0.0105131911,0.1854733229,-0.261251539,0.0412367694,-0.5030216575,0.1690862328,-0.3158500493,0.0115684131,-0.4746893942,0.3021249175,0.2482239157,-0.2088350803,0.2746181786,0.3141805232,-0.2436191291,-0.2717681229,0.2984854579,-0.2096520215,0.3557657897,0.0919852629,-0.1123642474,0.2093498558,-0.0522226915,0.1795104444,-0.2035749406,0.1083410084,-0.1023566127,-0.1805285066,0.3443871439,0.1667652875,-0.262537986,0.156000495,0.3252080679,0.2136511952,0.7869169116,-0.3852569461,0.1696177125,-0.0772030801,-0.2634728551,-0.3530146778,-0.03340929,0.2723197043,0.1240135133,-0.3135713935,0.1695544273,0.2323350757,0.0021240255,0.5360870957,-0.1672530472,0.1547774225,-0.320800513,0.0302221905,0.4241330922,-0.2938268185,-0.4632191658,0.1235978603,-0.010486397,0.1751444787,0.1186526567,0.1941884607,0.1164395586,-0.2828813791,0.2041236907,0.0405303463,-0.1049065217,0.0916736275,-0.3781327903,0.0146321012,0.3601006269,-0.2455439419,0.0002860624,0.0618882664,-0.1713062972,-0.0000895789,0.2635816038,0.0021527619,0.1317975819,0.0214457791,-0.0550763719,-0.1804737896,0.0705210641,-0.1827155501,0.0103290863,0.2068815231,-0.2916733921,-0.0309084114,0.2403352857,-0.1194010898,0.1090476215,-0.2765060663,-0.2697038054,-0.4919208288,0.1786421537,0.0512994602,0.4849601984,-0.0087551521,-0.0870167017,0.2997877002,0.3912484646,-0.1188084409,-0.3205804229,0.1116602495,-0.0021976321,-0.1649197936,-0.3592388928,0.5168806314,0.2101716697,-0.0087017491,-0.1274024993,-0.0456695259,0.0050723683,0.0821779445,-0.2594387829,0.0856710821,0.1075563282,0.1489421874,0.032679271,0.1141881794,-0.0331496038,-0.2737233043,0.126744315,0.4829832613,0.3349766433,0.1038113162,-0.0001734637,-0.0866548046,-0.0636653602,-0.1597080827,-0.2271358669,-0.3474419117,0.2525951862,-0.0317491926,0.425699085,0.2751815617,-0.0497094281,0.2771771848,0.5075845718,0.1860628575,-0.0227918532,-0.0917818621,-0.0151921874,-0.2629902065,-0.0940146744,0.2185552567,0.546656251,0.0993317217,0.1461596787,0.1891763508,-0.3211840689,-0.2050013095,-0.0299408957,-0.1228665784,0.06329678,-0.0426501445,0.3720710278,-0.0843531564,-0.2066482902,0.3466504216,0.1877124459,-0.1285450906,0.0377684496,0.2113189697,-0.2382613868,-0.4322886765,-0.3985508382,-0.2014000863,-0.1205924079,-0.3052800596,0.1275994778,0.3070930243,0.0944512188,0.2894098759,-0.2078359574,0.0865359753,-0.1703333706,-0.1507722288,0.0021728619,-0.3452709019,-0.313408494,0.0322801918,-0.0164930504,-0.2743608356,0.2854223251,0.0184094366,-0.1734153628,-0.1414485127,-0.3916652799,0.2079160661,-0.0086820498,-0.2005304545,-0.1986222267,-0.3684796989,0.0900128633,0.0122216055,0.3374741971,-0.3655262291,-0.2437645793,0.0550228618,-0.0861459896,0.1092540771,-0.2796729803,-0.1074035615,-0.144877106,-0.0487776957,0.2008318156,-0.282566458,0.2501550615,0.3470872641,-0.2549814284,-0.0977213532,-0.1596532464,-0.162631914,-0.4625173211,-0.0368571319,0.2164185047,-0.3161965013,-0.1890570819,-0.3839232922,-0.3472076356,0.2669111192,0.1701958925,-0.4356412888,-0.4333494902,-0.0611902401,0.0735043585,0.170878008,0.0493226014,0.3877258897,-0.1781677902,-0.0182335246,-0.1533555686,-0.3007079363,0.2078815848,0.3963835835,0.3567890227,-0.1756995618,0.2263661474,0.1310709566,0.5557454824,0.6314631701,0.0691826493,0.3732168674,-0.0136739174,0.2152680755,-0.261855185,-0.1999861747,0.1839577705,-0.1526453644,-0.2572266757,0.1832874268,0.1761990041,-0.5797126889,-0.0666341633,0.1652233154,-0.2786337137,-0.354192555,0.091165334,-0.3137045205,0.2286206186,0.3108310401,0.2339533269,-0.4184881449,-0.1160449311,-0.1248804778,-0.3066105247,0.1455273181,0.0942525119,-0.0962530375,-0.2474363595,-0.4561165869,0.2061754018,0.2422879487,0.1278309524,0.0226648599,-0.2308399379,0.0699954927,0.0194040854,0.4736868739,-0.033959318,-0.0039525745,0.1012182906,-0.071103707,-0.4034294784,-0.2792933583,0.2319036871,0.4614940286,0.4238904715,0.2808870971,-0.211793676,-0.1070905179,-0.0721646622,0.1328403503,-0.238768369,-0.3905230165,-0.2379556447,-0.0629179031,0.0660507381,0.0712945238,-0.1263540983,-0.1895468086,-0.1981025189,-0.0718090534,0.1953405887,-0.0391486175,0.0315442905,0.2464415729,0.2672862113,-0.1262177974,0.1792062521,0.1782539934,-0.1076482236,-0.4886569381,0.411167711,-0.1816360503,0.3603586853,0.1854187995,-0.0567481294,-0.0200529192,0.4066057205,0.1994699389,0.1454411298,-0.0775603727,-0.0773251727,-0.026072491,-0.2263959646,0.3206346631,0.002967767,-0.2382083535,-0.455353111,0.236322403,0.1579216123,-0.1496899426,0.2283801585,-0.6113729477,-0.1163541749,0.3909174204,-0.1971503794,0.784121573,0.0242022686,0.4790374041,0.2232557535,-0.0524470843,0.3801921308,-0.0466326699,0.3636256158,-0.3955560327,0.1865363419,0.121713154,-0.1743369848,-0.3330661654,0.0213027857,0.0663675815,0.4315514266,0.2073083371,0.4841272235,0.0202805381,-0.2113770247,0.2055265307,0.0201400407,-0.3214280009,0.1726046503,0.051495377,0.2472478598,-0.0631636232,-0.2258721292,-0.071922116,-0.1654229164,-0.0318808742,0.2956015468,-0.1394860148,0.4080139697,0.3562555313,-0.201168403,-0.6638280749,0.0960988477,0.1375410855,-0.3959575593,-0.0154681662,-0.1118344441,0.4741959572,0.4154971242,-0.0515213944,-0.2530067265,0.3664119542,-0.0708456412,-0.0483004972,-0.0550031811,-0.0085996035,-0.4087693691,-0.2352359891,0.3175220788,-0.3334677815,0.0507550202,0.1347248405,0.1830377728,-0.0747365355,-0.114834547,0.110037744,0.0739766434,-0.1293460131,0.3518595994,-0.3545048237,-0.1078604758,0.034888979,0.2857217789,0.2867710888,0.08036495,0.3626126647,-0.2709967494,0.0617275797,-0.0367226154,-0.3326629102,-0.3149150014,-0.3201569021,0.1079035252,-0.2535047233,-0.187484175,0.1025049165,0.0071218642,0.1206917092,0.3715602458,-0.0805359185,-0.1893055886,-0.5517706871,-0.0539892651,-0.1033546403,0.1609765291,-0.0176028274,0.6225499511,0.0702669919,0.3935421705,-0.2894585729,-0.020571854,-0.0156568978,0.2276464254,-0.13291803,-0.4048033655,-0.0541590713,0.1453793496,0.0436574444,0.3360795379,-0.4868725538,-0.2018187046,-0.1691497266,0.1019671932,0.146708712,0.1432404518,-0.2514538467,0.1305137277,0.3261987567,-0.3462394178,0.1161710992,0.169873938,0.0224850439,0.1915470064,0.1641242206,0.0671848804,0.1441858113,0.1170728356,-0.1414046586,-0.059945371,0.1628075093,0.1886161119,-0.1398519725,0.1501303017,-0.1367886513,-0.1625826061,-0.1133772209,-0.0793283582,-0.1007553264,-0.2714386284,-0.1292559355,0.0730395243,0.2117285579,0.2205676585,-0.0802233294,-0.1748627573,-0.103484638,0.1678178012,0.069235675,0.1069815755,0.3573373854,0.2053391486,0.0706474707,0.3598749042,0.4193611443,-0.0002602877,0.4682641029,0.0346510261,-0.0801458731,-0.47159338,0.2675064802,0.0105707077,0.195057705,0.0840756297,0.3485644758,-0.2840741873,-0.0718415752,0.0636945441,0.2873610556,0.1606606543,-0.0294532441,-0.0715025961,0.0820235386,0.0195648763,-0.2752049565,0.0292071085,-0.0937786251,0.2505142391,0.0436383449,0.371586591,-0.1546397656,-0.1823446751,-0.4207514822,0.3165487051,-0.2892093956,-0.2860289812,0.2699078619,-0.3006246984,0.2749159038,0.1448920667,-0.044626344,0.1926159561,0.7117823958,0.0600888282,-0.0273296405,-0.1531989127,-0.3156096041,0.2999631763,0.3962082267,-0.1062417924,0.347476244,0.0668392405,-0.3460538387,-0.1596561372,0.1851089001,0.3952046633,0.2746726871,-0.2849202156,0.1653120071,0.3119156361,0.0437199064,-0.1427929848,0.2844059169,0.4644452631,-0.2753710747,0.1001332924,0.0773630664,-0.1619855016,0.3374540806,0.0922406614,0.262637347,-0.2047153115,0.5761632323,-0.0550600849,-0.0793758184,0.0089691496,0.2030296326,-0.5066410303,-0.1221208423,0.0619587339,0.0576235577,0.1631636173,-0.2219672799,0.0901636928,0.1087147817,0.3901248872,0.1252449155,0.1304319352,-0.285621196,-0.0167448502,-0.5698698163,0.3329299688,-0.3890035748,0.2458084226,-0.1836736947,0.2748156488,-0.2668350339,0.0394254401,0.1469787806,-0.2107889503,0.1848515272,0.1018854827,-0.453212291,0.025243599,0.0722882822,-0.1955050826,0.2549091578,-0.2316809893,0.0217405539,-0.0190826729,0.1310511082,0.0037309956,-0.0033554703,0.2672175467,0.506138742,0.2889290452,0.4251261652,0.2001944035,0.0405017287,-0.1885145158,-0.3905121684,0.0326012857,-0.0315821469,-0.2456115335,-0.0693586171,0.3341854513,-0.3125975132,-0.1130522713,-0.2842991948,0.2264569551,0.0809383988,-0.3933438659,0.178098619,-0.0364459157,-0.1294267476,0.0605262034,0.0307938773,0.725405097,0.0161453299,0.4419613481,-0.1853630543,-0.2571270764,0.381931752,-0.6724100113,-0.2614426613,0.0037970759,0.3030821681,0.0934228078,-0.1773825586,-0.53091079,-0.124379918,0.1777398437,-0.1675694287,-0.1212128326,0.1805972308,-0.19370538,-0.0350330062,-0.0106028188,0.1642269343,0.0944306403,-0.2684945762,0.2727923989,-0.2511074543]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/186","title":"Weird-ish: Not creating unique caches for different phases","comments":"Good catch, it looks fixed.\r\n","body":"Sample code:\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load_dataset('boolq')\r\n\r\ndef func1(x):\r\n return x\r\n\r\ndef func2(x):\r\n return None\r\n\r\ntrain_output = dataset[\"train\"].map(func1)\r\nvalid_output = dataset[\"validation\"].map(func1)\r\nprint()\r\nprint(len(train_output), len(valid_output))\r\n# Output: 9427 9427\r\n```\r\n\r\nThe map method in both cases seem to be pointing to the same cache, so the latter call based on the validation data will return the processed train data cache.\r\n\r\nWhat's weird is that the following doesn't seem to be an issue:\r\n\r\n```python\r\ntrain_output = dataset[\"train\"].map(func2)\r\nvalid_output = dataset[\"validation\"].map(func2)\r\nprint()\r\nprint(len(train_output), len(valid_output))\r\n# 9427 3270\r\n```","comment_length":5,"text":"Weird-ish: Not creating unique caches for different phases\nSample code:\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load_dataset('boolq')\r\n\r\ndef func1(x):\r\n return x\r\n\r\ndef func2(x):\r\n return None\r\n\r\ntrain_output = dataset[\"train\"].map(func1)\r\nvalid_output = dataset[\"validation\"].map(func1)\r\nprint()\r\nprint(len(train_output), len(valid_output))\r\n# Output: 9427 9427\r\n```\r\n\r\nThe map method in both cases seem to be pointing to the same cache, so the latter call based on the validation data will return the processed train data cache.\r\n\r\nWhat's weird is that the following doesn't seem to be an issue:\r\n\r\n```python\r\ntrain_output = dataset[\"train\"].map(func2)\r\nvalid_output = dataset[\"validation\"].map(func2)\r\nprint()\r\nprint(len(train_output), len(valid_output))\r\n# 9427 3270\r\n```\nGood catch, it looks fixed.\r\n","embeddings":[0.2467508912,-0.0976690948,-0.0372382626,0.1779502034,0.1058292836,-0.2055710554,0.2004635185,0.2624466419,-0.0925657824,-0.1227251887,-0.0374020599,0.408423245,-0.01656349,-0.2244389206,0.0681109875,0.0637722015,0.146532774,0.088406831,-0.119842805,-0.0834258422,-0.1803459823,0.174134478,-0.1668830067,-0.1523104459,-0.4404868782,0.1104758829,-0.2117645144,-0.1633782238,0.2750295699,-0.3381148875,0.3067492247,-0.0165863428,-0.2675566375,0.2301988751,-0.0001132115,-0.0356316455,0.1879738867,-0.0438100658,-0.0709830672,0.1662341803,0.0194791947,-0.1135900542,0.098683089,-0.2909155488,-0.0390187688,0.1928951591,0.151916787,-0.0348634459,0.5754731297,0.1556378156,0.2230450809,0.2614919841,-0.2655758262,0.2070192248,0.1937733442,-0.182730332,-0.1658739597,-0.1077026948,0.0576598234,-0.3044588864,-0.1241888404,0.4223961532,0.001210739,0.0891210586,0.0057430165,0.1261882186,0.2180008292,-0.1708450019,0.0840092152,-0.0263395961,-0.0641142577,-0.2926781178,-0.2009978443,-0.3674985468,-0.2090887725,-0.3260148466,0.2036847919,0.1392593384,-0.0990399793,-0.1188398972,-0.4234812558,0.2522473931,0.1797697246,0.1378898025,0.1009887382,0.3631615043,0.0732858628,0.0020991773,0.0584406815,0.1610176861,-0.0230813362,-0.1895124912,0.1000808105,0.4251976311,-0.2456512153,-0.1197907403,0.3103427589,0.1877243817,-0.0233557858,-0.0856424719,0.2379693836,-0.0791316181,-0.04346193,-0.0056142686,0.1237654239,0.3933605254,0.0073733148,0.2085866481,0.0503908768,-0.0757130831,-0.5614403486,0.1115200445,0.3435876369,-0.0386376753,0.296112895,0.1114469096,-0.0347221941,-0.1030851677,0.0701428279,0.1545524448,-0.3370786607,-0.099213466,0.0627561882,0.1969385594,-0.0627957508,0.1660838425,-0.189260751,-0.0102780024,-0.3818277121,0.0711264461,-0.3526243567,0.0007570832,-0.4996190071,0.2152942717,0.2934217453,-0.1024259627,0.2616894841,0.291500926,-0.3210468292,-0.297286123,0.2064352781,-0.2499745637,0.3831695318,-0.0042478954,-0.0869320109,0.2238081843,-0.0271280091,0.102530621,-0.2054039538,0.1013635844,-0.1037044451,-0.1046204716,0.3924096823,0.1826354712,-0.238179639,0.1474092156,0.3045396805,0.2964282632,0.765645206,-0.3975048363,0.1024252847,-0.0464564748,-0.4125587642,-0.3448876739,-0.0271145385,0.2197591215,0.1326428503,-0.3335547745,0.0990268514,0.2270097435,0.0293889046,0.5131651759,-0.1968905628,0.2084186673,-0.299221307,0.0094212368,0.5258950591,-0.2730056345,-0.4465505481,0.1349041462,-0.0040677618,0.2669122815,-0.0714582056,0.1764044464,-0.0311226733,-0.2639950812,0.2557732463,0.1529802829,-0.1528008431,0.160166949,-0.4040044248,0.1117502302,0.3810142577,-0.2785815597,0.0394047834,0.1532391906,-0.201672703,-0.0820174739,0.3216599822,0.0462060347,0.1553692073,0.1570504457,-0.1299113631,-0.2260966897,0.0065170014,-0.2638790607,-0.0069037196,0.1804569066,-0.3207966685,-0.083922036,0.2762281299,-0.1029161513,0.0425747335,-0.2761160135,-0.0610666722,-0.4243394732,0.2141405791,0.0641163886,0.5120956898,0.0435786396,-0.0272832736,0.3187514544,0.3601571918,-0.1739725918,-0.3843175769,0.122497119,0.0576691478,-0.2238493413,-0.3521160185,0.4367851317,0.1808225662,0.0591350123,-0.1777099818,0.0183074363,0.0065117842,0.1140706837,-0.3114301264,0.087683171,0.001036011,0.2572281361,-0.0356563553,0.040005412,0.0135687524,-0.2630066276,0.1080291718,0.4474736452,0.2676286995,0.0585392118,0.0447243452,-0.0707884505,-0.0536503457,-0.1922993958,-0.2210625857,-0.3467621207,0.2634724975,-0.1055716053,0.3947494626,0.2271264344,-0.1119230762,0.194916442,0.559027195,0.3266520798,-0.0117744114,-0.1898439378,-0.0528109521,-0.2433575094,-0.0810533762,0.2880263329,0.6204983592,0.1226324812,0.1981034875,0.1977858096,-0.3112400472,-0.2028662115,-0.0079309093,-0.1191845685,0.1986859292,0.0138078304,0.3383920491,-0.1086766273,-0.1590774655,0.3171004355,0.1418779343,-0.1224306747,0.0199537408,0.214849785,-0.1977145523,-0.3744459152,-0.417193532,-0.1782879829,-0.1268566251,-0.346491605,0.0851536542,0.2462391555,0.0559820272,0.2930340469,-0.2393258065,0.0119159725,-0.1574973762,-0.0721321851,0.0479729921,-0.2836326063,-0.2761681974,0.0441968739,0.0036349487,-0.2369939685,0.2841704786,0.021642305,-0.1901270002,-0.0109405005,-0.3311147094,0.1500100642,-0.0472496971,-0.0848964676,-0.149491936,-0.3801751733,-0.058589641,0.0596425571,0.3628844619,-0.4276551306,-0.1650507599,-0.0512862951,0.0130778095,0.0390434153,-0.3039737046,-0.0379316881,-0.0963370502,-0.1083372459,0.1797584742,-0.2976326048,0.221140936,0.360248208,-0.2819384933,-0.0676205903,-0.0959591568,-0.0871382952,-0.4527276754,-0.0060439003,0.2761670053,-0.3547703326,-0.214240253,-0.3392578363,-0.3599377275,0.3391325474,0.2193216681,-0.4053862095,-0.3313527405,-0.0954528302,-0.0130595397,0.0861419886,-0.0173430908,0.3935339153,-0.2347562164,-0.0259816274,-0.2739040256,-0.3110162318,0.1573457122,0.3807273507,0.3259324133,-0.1876170486,0.1919328421,0.1538222581,0.4242323935,0.5318962932,0.0135333352,0.3279158175,0.0675077662,0.1692203283,-0.1628544182,-0.1627050489,0.1646415591,-0.1214658618,-0.2525874376,0.2296363264,0.1180064827,-0.5839285254,-0.1022793651,0.115559876,-0.3014006615,-0.3825238347,0.1617707908,-0.3456188142,0.2782308757,0.3069998622,0.1891656965,-0.439653039,-0.0954160243,-0.1756658256,-0.3724215627,0.2109208405,0.0190771986,-0.168021515,-0.1556232423,-0.4371362627,0.229652226,0.2041510195,0.1077801213,0.048106771,-0.202324599,0.0136870975,0.0346206054,0.4651925564,-0.1334349215,0.0148656117,0.0631800368,-0.0193251073,-0.468541801,-0.1927761883,0.195048973,0.49910748,0.5111116171,0.2588692605,-0.1501128972,-0.1093959883,-0.1582272649,0.085035108,-0.1899686903,-0.3639028072,-0.2344601452,0.0845674127,0.159902662,0.0871433094,-0.0315834358,-0.1360756159,-0.1529981047,-0.0506386012,0.1398162842,0.0473557599,0.045095142,0.1537533849,0.2993910313,-0.0774668232,0.145727247,0.130602926,0.0381519645,-0.5952783823,0.3479579985,-0.1929633021,0.3034771979,0.1253344268,0.0252651852,-0.0143450592,0.4593455791,0.185751915,0.072895512,0.0131355654,-0.0379739217,-0.0626626462,-0.1546679586,0.2494883388,-0.0068512019,-0.2068850845,-0.415191263,0.1897114813,0.1349692792,-0.142161116,0.2226055264,-0.5840522051,-0.2404207885,0.3790788352,-0.2401004136,0.7024500966,0.0593972467,0.3947834969,0.1329906434,-0.0497637764,0.3217284679,0.0231300835,0.4130138159,-0.3828523755,0.1278813183,0.1250552237,-0.1894874871,-0.2984346747,0.0846430138,0.0488993004,0.4160304368,0.2737907767,0.4373629689,-0.0521589853,-0.1984528303,0.234675169,-0.0126885073,-0.4254105091,0.192848444,-0.0507681966,0.2421834767,-0.1066523194,-0.2300883681,-0.1383195668,-0.1871251315,0.0034877444,0.3396110535,-0.2003385276,0.3499772549,0.3910221159,-0.0907780007,-0.5435519814,0.3164889514,0.1823416799,-0.2773410976,-0.0872761607,-0.0239500236,0.3577364981,0.4057737589,-0.0280132536,-0.2632077932,0.3276540935,-0.058815591,-0.0225343108,-0.0252097957,-0.0625133067,-0.355622977,-0.1964541823,0.348646313,-0.3298530877,-0.0210578144,0.0330763794,0.2212704271,-0.0494459756,-0.1172596961,0.1370675862,0.1078752875,-0.1022947133,0.4229285717,-0.3408428133,-0.148040399,0.0473600999,0.3655653298,0.211040169,0.0641544387,0.3706263304,-0.2541859746,0.0235741306,-0.0640709773,-0.3895638883,-0.300804317,-0.3117941618,0.1512131691,-0.2426047921,-0.1166965291,0.0930571333,-0.0172237586,0.0766480193,0.3540396094,0.0398646109,-0.1762595028,-0.5084567666,-0.0029238546,-0.0446056463,0.0909313783,-0.0295880325,0.5607545376,0.0656637922,0.419238627,-0.313316226,0.0042274809,-0.0214848332,0.2667610049,-0.1807450205,-0.3946766555,-0.0281164665,0.1161584407,0.0691400617,0.3533011079,-0.476937145,-0.2462261319,-0.1677579284,0.0799037665,0.1578819007,0.11382664,-0.2150483578,0.1133637875,0.3391011953,-0.4293695688,0.0863320231,0.1784188449,0.0181651339,0.1590196341,0.1513860971,0.1806744188,0.230793938,0.0647560209,-0.1952196509,-0.0633836538,0.087912485,0.1788567901,-0.0533461384,0.1457526386,-0.1561160535,-0.229161337,-0.1491373032,-0.0309164748,-0.0445466042,-0.2233716547,-0.2040914595,0.18588835,0.1871411055,0.1911724657,-0.1303969175,-0.1466026902,-0.1282514483,0.1956232637,0.1116566733,0.1475288868,0.4513443112,0.2425233424,0.0893666446,0.3170326054,0.3521600962,0.0112977428,0.3836307228,0.0893443972,-0.0909909159,-0.542543292,0.209189415,-0.12400233,0.2365145534,0.0831066072,0.2714445293,-0.1687241793,-0.0728907585,0.0629372448,0.3204174936,0.0970883071,-0.0253819767,-0.1085503846,0.0637081787,0.0151691027,-0.2949649692,0.1420210302,-0.1623100042,0.3247112632,0.0305670314,0.2136269063,-0.1116685718,-0.1899912357,-0.4100203514,0.2646826506,-0.2626350224,-0.2959661484,0.2458176166,-0.2448193282,0.1881359369,0.1276459396,-0.0038352206,0.2449948788,0.6132069826,0.0613951236,-0.0243902057,-0.1274356693,-0.2767182589,0.2625440657,0.401470989,-0.1648859233,0.3250267208,-0.0083360141,-0.2131198198,-0.1048563048,0.0828499049,0.2853808701,0.3168081343,-0.2429727018,0.1203551069,0.3555525243,0.0084134312,-0.2312155366,0.3709870577,0.462357372,-0.2219891101,0.1620825529,0.1146784499,-0.1862500757,0.3663871884,0.0280910358,0.3507835567,-0.1754405946,0.4934214652,-0.0726342052,-0.025783319,-0.0762635171,0.1438708603,-0.4287461936,-0.1315491349,-0.0524005666,0.0165930279,0.2332145721,-0.2393474728,0.0917768925,0.1106150299,0.3234194815,0.1314888,0.2112924457,-0.2459359318,-0.1212511435,-0.543967247,0.241321072,-0.3162086904,0.1999905556,-0.2909040153,0.3201945424,-0.2975472212,0.0054112142,0.093231082,-0.3659472466,0.1648614556,0.1534638256,-0.4105710685,0.0745476112,0.0572776683,-0.2012097389,0.3014259338,-0.2386474311,0.0002776418,0.0264534596,0.1430317611,-0.0069312458,-0.0335653611,0.1602081656,0.5790530443,0.2908617556,0.4455899,0.1744848788,-0.1029497087,-0.1674015224,-0.4891856909,-0.0037958382,-0.0015222834,-0.1992305666,-0.0761097819,0.4431738853,-0.3390386701,-0.0033540379,-0.1907356679,0.2339560986,0.0990443602,-0.2916224599,0.2621689737,-0.0810242072,-0.1022300422,0.1372796148,0.0039339964,0.7269467711,-0.012151082,0.525315702,-0.2023798674,-0.224383086,0.3839532733,-0.5697457194,-0.3239184916,0.0510922894,0.3110687435,0.1868885309,-0.236168772,-0.5729414225,-0.2062658221,0.2269307673,-0.1704829186,-0.0907884091,0.1515610814,-0.1434978843,-0.0752695724,-0.006830649,0.0987630114,0.0738323554,-0.2451484799,0.2848592401,-0.2615081668]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/183","title":"[Bug] labels of glue\/ax are all -1 ","comments":"This is the test set given by the Glue benchmark. The labels are not provided, and therefore set to -1.","body":"```\r\nax = nlp.load_dataset('glue', 'ax')\r\nfor i in range(30): print(ax['test'][i]['label'], end=', ')\r\n```\r\n```\r\n-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, \r\n```","comment_length":20,"text":"[Bug] labels of glue\/ax are all -1 \n```\r\nax = nlp.load_dataset('glue', 'ax')\r\nfor i in range(30): print(ax['test'][i]['label'], end=', ')\r\n```\r\n```\r\n-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, \r\n```\nThis is the test set given by the Glue benchmark. The labels are not provided, and therefore set to -1.","embeddings":[0.2502776384,-0.3640763164,-0.0616975874,0.1393109411,-0.0684490055,-0.0784347355,0.3446009159,0.2630479932,0.3701296449,0.1910719126,-0.2715180814,0.6841737628,-0.0655928925,0.1188958809,0.1820769161,-0.1121413559,-0.1016677618,0.3471295238,-0.1504511684,-0.1217599735,-0.261205107,0.3785851896,-0.2576521337,0.1414736658,-0.1796225756,-0.1017393172,-0.1501422524,0.0767334104,-0.0330746435,-0.4014689624,0.1059926972,0.1474643201,-0.1076188013,-0.1203342974,-0.000105719,-0.2869568169,0.2894555628,0.0716744885,0.0324562751,0.1292870194,-0.1467484832,-0.2972307503,-0.0397070386,-0.3864676356,-0.0571407005,-0.0009058942,0.0459285527,-0.1283687502,0.1665222347,0.1647905558,0.2408470958,0.1582940668,0.0880785137,0.1361727566,0.4565768242,-0.4768173099,0.0443155132,0.2460297197,0.0988072753,-0.3428685069,-0.0160303526,0.4554493427,0.1296474934,0.1253583431,0.0396814495,0.1442377269,0.0331890434,-0.3599227369,-0.0838703588,0.4264652431,0.1315085888,-0.3524526358,-0.3100077808,-0.201657027,0.2139121145,-0.3125111759,-0.0647390932,0.1420345008,-0.0162971616,0.0097063538,0.0376629084,0.1028022766,-0.1219680756,0.1391842663,0.13890405,0.5842686296,-0.0203114767,0.0699541345,0.3692638874,-0.0291570164,-0.185024932,0.0921852887,-0.0200800244,0.1759937108,-0.2807863951,-0.1972289979,0.1095440388,0.1007739007,0.0210758969,0.0445978157,0.2250251621,0.0380703025,0.2073385268,0.1642030329,0.0804497227,0.2856959701,0.3848163486,0.2837710083,0.0289968792,0.0624472275,-0.1282516271,0.2089024037,-0.0141546503,-0.1045895666,0.2733450234,-0.0953949988,-0.074853614,-0.12386702,-0.3194419742,-0.0257939361,-0.2851247787,0.0020686025,-0.1796579659,0.2444522083,0.0278627556,0.015347749,-0.034649346,0.0959750712,-0.1576816738,-0.1615332067,-0.2969714403,0.3730207682,-0.2329174429,-0.2509106994,0.1420157552,0.2079280317,0.2833411098,0.0284869093,-0.1306972206,0.0372800976,0.1497273892,-0.0204422679,0.2775793076,0.1166959628,-0.0686403364,0.1049478129,-0.0104661845,-0.3798038363,-0.2766103745,-0.0801543668,0.0921468064,0.0742983446,0.023174474,0.2386460155,-0.1758063436,0.0270310473,0.1805446297,0.0985941589,-0.0963577181,0.2063302994,0.1449921131,-0.1402951628,-0.1153011769,-0.189779982,0.2916013002,0.2018365413,-0.3010580838,-0.1139512658,0.0818045884,-0.0352464207,-0.1266508847,0.0524620414,0.2215994149,0.0054084039,-0.2682152689,0.2221393436,0.2920090854,-0.5699797869,-0.4601472616,0.0261179712,-0.0982162878,-0.1154138446,0.0840379745,0.0799318999,-0.027337499,0.2073450983,0.3057430983,0.2759204805,0.1464020759,-0.0140554663,-0.3354370892,0.1455385834,0.3818062544,0.1446694285,-0.1724830121,-0.1629223228,-0.2000304759,-0.1649155617,0.2638297081,0.1147524565,-0.0881151855,0.268130064,0.0120114852,-0.3535585999,0.1252509356,-0.1460119784,-0.234487325,0.2485731393,-0.2334445268,0.3903802335,0.2793847919,-0.135439083,-0.3910467327,-0.0095400782,-0.0215908568,-0.222611323,0.27640149,0.0221559014,0.3353919685,0.0612459555,-0.0968827158,0.3609966934,0.0980221555,0.0416084677,-0.369710058,0.0968736932,-0.0590874664,-0.0400334708,0.2472639233,0.510469377,0.2441359758,-0.0534041636,0.1302887201,0.3582192361,-0.0236603208,0.0128956521,0.0835901871,-0.0173818693,-0.0429648645,-0.2957713306,0.0962691158,0.1672365367,0.0804634988,-0.0402015708,-0.1558143646,0.3673764765,0.0938399211,0.051864069,0.1129817963,0.0980570689,-0.0967075527,-0.2965038121,-0.3145454824,-0.1394534409,0.334841311,-0.2254430652,0.3553495705,0.0405211672,-0.2502380311,0.2089716494,0.4467054009,0.0246783774,0.2037923187,-0.1188760623,-0.2853447795,-0.1006391421,0.0548200794,0.5218162537,0.1674064845,0.4208612144,0.0176508427,-0.0597045571,-0.3939405978,-0.0734595507,0.1677392572,-0.0849689916,0.079692632,-0.0892410651,0.2013346553,-0.148259595,-0.1896372736,0.0556289256,-0.2480327189,0.234317556,-0.2628102899,-0.1954490989,-0.0848426297,-0.4971702397,-0.2795573175,-0.2954831123,-0.1468037665,-0.5272936225,0.1358091831,-0.2413761318,-0.0206852444,0.2252945751,0.0543605722,0.1611211896,-0.1830408871,0.078437835,-0.2059392482,-0.1355561018,-0.2107263654,0.1599219292,-0.2740511,0.5203288198,0.3989130855,-0.0967089012,0.0753886253,0.3190047443,-0.2143687457,-0.0005630524,-0.1737831682,0.2465425283,0.4136917889,0.0018733669,-0.127711162,-0.0129922275,0.2429467887,-0.4388088584,-0.0056831175,-0.0672175065,0.2072559297,-0.0391405188,-0.1482223123,-0.4048272371,-0.3648358881,-0.2212098688,0.2395437211,0.2358979732,0.2596203387,0.1151958182,-0.3311708868,0.3270048499,-0.1680489033,-0.0336020775,-0.1516866237,-0.0547351688,0.2490203679,-0.1052173227,-0.2982955277,-0.2496383637,-0.1970977187,0.2230586261,-0.2926648557,-0.2116057575,-0.1770162284,-0.0741497353,-0.1401940137,0.0278450288,0.2228148878,0.1176871508,0.0326344036,-0.1517806053,-0.1452549249,-0.2485678792,0.3365797102,0.0520556122,-0.1408715099,0.0385372303,0.0352318808,0.0140010361,0.1799610108,0.1178442538,-0.1231226102,0.1978878379,-0.2788589299,0.2952502966,0.2305527329,-0.2540290952,0.0103728268,0.3167246282,0.2669706047,0.1392034143,0.1171701178,-0.3290161192,-0.0998946801,0.1183217466,-0.262835145,-0.0590805411,0.0393484235,-0.1033325121,-0.2153227776,0.0206797197,-0.0360731371,-0.1080714017,-0.1725045592,-0.0427563936,-0.3998681903,-0.0838824362,-0.0383352339,-0.6121569276,-0.1214394495,-0.5194763541,0.1543166041,0.0483947508,0.4734871089,-0.0107190656,-0.0948591977,-0.0313768201,0.0304214656,0.5715873241,-0.414696902,0.213191703,0.4603512287,0.1821395755,-0.694996655,0.0056413086,-0.5489996076,-0.1169118881,0.6284407973,0.2622639239,-0.2489965856,-0.1502825171,0.2737720311,0.016337879,-0.0919275507,-0.1665162444,-0.2935800254,-0.2230233103,-0.0367949791,0.1478812844,-0.0091051292,0.2636658847,0.1972577572,-0.0781444684,0.0583056845,-0.2125874162,0.144697547,0.2239960432,0.2999828458,0.0681437403,-0.0385125503,0.1171308681,0.2122257948,-0.6786824465,0.121092394,-0.5201427341,0.3546108603,-0.2175754607,-0.2540550232,0.4651712775,0.1619275063,0.0466176346,-0.161482811,-0.5788980722,0.2101602107,-0.0478425473,0.020870043,0.208485052,0.1639750153,-0.2340152264,-0.0756695196,0.5497072935,0.0684776679,-0.224692598,0.3328180313,0.1943953931,-0.3568425477,0.4119116962,-0.0888912603,0.9092609286,-0.0902239904,-0.0075918632,0.2545849383,-0.3636752069,0.6806232333,0.1408301443,0.1234453544,-0.5268469453,-0.3948065042,0.0687417686,-0.0832069218,-0.1278981268,0.275515765,-0.2226205617,0.4155819118,0.0730142668,-0.1933465898,0.0223330799,0.3862771094,0.0031905486,-0.1048492789,-0.3415438235,0.1548099369,-0.0329706445,0.167489633,-0.0614623912,-0.0196787473,-0.132370308,-0.2114521861,-0.0483419374,0.197701335,-0.2182679623,-0.1657030731,0.3732071221,-0.0305131711,-0.452439636,0.130931586,0.1828851253,0.1245384961,0.0064065936,0.0526251607,0.1765982509,0.0018510445,0.5183132887,0.2572222352,0.2960578501,-0.1017704979,-0.0579164699,-0.1008150354,0.085979104,-0.1053322852,-0.0639912188,0.0586687587,-0.0525974818,-0.1444304734,-0.3400135934,0.0604076535,-0.0134995282,-0.197664991,0.1982205361,0.0939964801,-0.4110828936,0.2700432837,0.0935873836,-0.3252713382,-0.0368701853,0.3700682521,-0.0598653108,0.4367307723,0.172697261,0.0293653719,-0.0168950465,-0.2826912105,-0.0653925687,0.321457684,-0.2296519876,0.0666497871,-0.1234247759,-0.1088970676,-0.0217208378,0.0352228433,0.1964286715,0.1793636829,0.0944262818,-0.4698393643,-0.1270841062,0.1616881937,0.0665285438,0.3171401024,0.0377872027,-0.0941464603,0.2179002315,0.3500137031,-0.4358422458,-0.0236710943,-0.2066912651,-0.0781285465,0.4145665169,-0.0236212723,0.0742359906,0.0542693958,0.3157197237,0.0799782127,-0.0934089124,-0.3097279668,-0.1868813634,0.0197497662,-0.0782093853,-0.0117298001,0.0257643554,-0.1200133041,-0.0767245367,-0.0659571439,-0.0914931372,-0.146587044,0.007814656,0.4473216236,0.0004193679,0.0969923735,0.0640292391,0.1166830286,-0.0210979749,-0.2281682342,0.0321902819,0.1621512026,-0.0517000929,-0.098664239,-0.1505555362,-0.0471648164,-0.3882824183,-0.0687662363,0.0067755408,-0.1605019122,-0.0958236232,0.0773943439,0.634023428,0.0107848085,-0.2965523899,-0.192164883,0.0279807709,0.3456476033,-0.3949334323,-0.1269024611,0.4542478919,0.2467412651,0.2646099925,0.352965951,0.1391118169,-0.0840526149,-0.1915542781,0.0040004486,0.2913329303,-0.4078508317,-0.1634413004,0.1737110317,-0.1426516026,-0.1698190719,0.3206777275,0.0578653179,0.1680351049,0.0838930234,0.3906585574,0.5761871934,0.1416966915,0.0304269325,-0.0577272959,0.1151572466,-0.0707179308,0.5745173097,-0.0276498739,0.2495678514,0.1567799747,0.5897644162,-0.2613561749,-0.2176583707,-0.0409209616,0.2125947177,-0.0342461579,-0.2194951028,0.1371561736,-0.0714574158,-0.3883415759,-0.0043713059,0.0145801594,-0.2313610911,0.3803722262,-0.0484597385,0.1276195347,-0.2602134347,-0.1190984398,0.0248678122,0.4030855894,-0.3979931176,0.0643229485,-0.2506301999,0.0955432057,-0.0175976735,0.3547403514,0.171680823,0.0177510045,-0.251244545,-0.08426328,-0.1949608028,-0.1199971363,-0.2027817667,0.2279925644,0.2409190536,0.018384289,0.4225254059,0.1816040576,-0.2255632579,0.0491378456,0.1672346592,0.0984476283,-0.0070404834,0.4931470752,-0.1018495783,-0.2557189167,-0.2666806579,-0.0513108484,-0.402690053,-0.066655606,-0.1175006181,-0.0308531094,0.1739558876,-0.3835694194,0.1416160017,-0.0497349761,0.4453088045,0.0125162564,0.2989162803,-0.2112408876,-0.2104013711,-0.6262410879,-0.0229917727,-0.0937419236,-0.5266143084,0.2152057588,0.0295979828,-0.0642610043,0.020876782,0.0427762866,0.5819959641,-0.1631900519,-0.1065302491,-0.4230671525,0.115421541,-0.1829814911,-0.0502040125,0.1412268579,-0.2908743322,0.0111123519,-0.2384875566,0.1909265816,0.0728773773,0.0255425349,0.0676968172,0.2701843679,0.1387313604,0.0436714292,0.2616766095,-0.1445899308,0.0857321545,-0.1028873995,-0.0168459956,-0.2982357442,0.0394424424,0.1075921729,0.3338515759,0.041206304,0.0219927747,-0.0728792474,0.0933411717,0.1682982892,0.2568712533,0.03382539,0.1943797916,0.1423932612,-0.0410159156,0.0953965634,0.3407488465,0.0053744121,0.0550034158,-0.0129750976,-0.3489193618,0.4604707062,-0.3287241459,-0.1328889728,0.0940823555,0.3477377892,0.2625038922,-0.2408187091,-0.5395864844,-0.1369075477,0.3368124068,0.0426281095,-0.295516938,0.3126365244,-0.0480222479,0.1210707799,0.054203894,-0.0159835536,0.2987117767,-0.2579360008,-0.1043453738,-0.2538382709]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/183","title":"[Bug] labels of glue\/ax are all -1 ","comments":"Ah, yeah. Why it didn\u2019t occur to me. \ud83d\ude02\nThank you for your comment.","body":"```\r\nax = nlp.load_dataset('glue', 'ax')\r\nfor i in range(30): print(ax['test'][i]['label'], end=', ')\r\n```\r\n```\r\n-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, \r\n```","comment_length":14,"text":"[Bug] labels of glue\/ax are all -1 \n```\r\nax = nlp.load_dataset('glue', 'ax')\r\nfor i in range(30): print(ax['test'][i]['label'], end=', ')\r\n```\r\n```\r\n-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, \r\n```\nAh, yeah. Why it didn\u2019t occur to me. \ud83d\ude02\nThank you for your comment.","embeddings":[0.3708410263,-0.3751214147,-0.0895321965,0.1415943354,0.1954722106,-0.0417574495,0.3801866472,0.2917180061,0.3838462532,0.248697266,-0.191145733,0.6987327337,-0.064515166,0.1376854777,0.1227343827,-0.1220916361,-0.0170389097,0.3643182516,-0.0583360605,-0.2359318584,-0.2340733707,0.3038620055,-0.1788117588,0.1264441758,-0.2056163698,-0.0668441355,-0.199635759,0.0801730677,-0.0264840871,-0.3986214101,0.0102963736,0.0890546888,-0.0483393818,-0.1038806289,-0.0001039119,-0.2509903312,0.3352326155,0.0857229978,-0.0059774425,0.0903814211,-0.1505310237,-0.23976928,0.0136992056,-0.4266754985,-0.0367207751,-0.0614703447,0.0302944351,-0.0575460307,0.1346543878,0.0909186304,0.2731457055,0.0395969041,0.2134052515,0.1035881415,0.4497775435,-0.4262342155,0.1171754524,0.2512437403,0.1091182455,-0.2832530737,0.1284733117,0.4960659742,0.1312758327,-0.0407252684,0.0584582649,0.2323475629,0.0881044939,-0.4056469202,0.0032400119,0.2882003188,0.193777591,-0.224108547,-0.2230432183,-0.1668432057,0.2673069835,-0.3504532278,-0.1359534264,0.1640959978,-0.0465853959,0.0129830418,0.0730399415,0.0455797426,-0.1620164812,0.2187044919,0.1125380844,0.5738576651,-0.0324113071,0.0980833024,0.2174468935,-0.0471462347,-0.1732245982,0.1084051505,-0.056351725,0.1593599468,-0.1362866312,-0.1496627182,0.02645915,0.1117702797,-0.052768819,0.0223300606,0.1776747704,-0.0111025441,0.1576340646,0.1935818493,0.2058342695,0.12974222,0.3342539668,0.3229803443,-0.0305325035,-0.0014377469,-0.1205738708,0.0340765156,-0.0161010474,-0.0424883664,0.3488124013,-0.1647933722,0.0669106245,-0.0719071478,-0.2343203723,-0.0811898187,-0.2756981254,0.0681279376,-0.1644869149,0.380107969,0.0262698792,0.0868188217,0.0005348424,0.0823924989,-0.0488138087,-0.1965441853,-0.3257721066,0.375317961,-0.2257722318,-0.3333780169,0.0480744503,0.2518254817,0.2870738804,0.0379651152,-0.090074122,-0.0602398142,-0.0935175642,-0.0429137871,0.2130081207,0.0965193212,-0.1346526146,0.1224582568,-0.0215869043,-0.4195561707,-0.122476168,0.0225018486,-0.0503911264,0.0972119197,-0.1800504476,0.2941066623,0.0303152706,0.0211137161,0.2731417418,0.01033739,0.0665086284,0.1135925353,0.098746255,-0.1425992996,-0.0689813495,-0.2650093138,0.1759663224,0.0966576636,-0.260437727,-0.0711926222,0.0746783316,-0.0465483293,0.0619910657,0.068730779,0.1801342666,0.1296254396,-0.2671622634,0.2113316208,0.4376261532,-0.4657274187,-0.4293342233,0.0501546375,-0.098887451,-0.1770169139,0.1095170155,0.0935898721,-0.1395625174,0.224183321,0.4004226625,0.1936155856,0.0872539505,0.0731734112,-0.3200443685,0.2184984982,0.2777940631,0.1364082992,-0.2285876721,-0.0174396038,-0.0640681684,-0.1296937913,0.4404654503,-0.0485949218,-0.0058109104,0.3151887953,0.0711440742,-0.157704398,0.0714678764,-0.1961246282,-0.2621771991,0.0956668332,-0.1889840215,0.3343580365,0.0589387603,-0.1750100553,-0.3691319227,0.029068483,-0.0427946933,-0.1550528407,0.3209889531,0.0909484178,0.2393031567,0.0118562756,-0.1032157838,0.2728453279,0.0956975818,0.0690372735,-0.4254882038,0.1247428209,-0.1411261261,-0.0450714268,0.1916948259,0.5925878882,0.3196223676,0.0524961017,0.1605210006,0.3518287838,-0.0523260981,-0.0341062583,-0.070737645,-0.1885704547,0.0071762097,-0.2161056399,0.0190119669,0.1219974533,0.234840408,-0.0500215329,-0.1317513883,0.250231564,0.0231384318,0.0006106868,0.1130768061,0.0737835169,-0.031883318,-0.2460392416,-0.2480244339,-0.2095655501,0.259208411,-0.1635896266,0.1624386013,0.026269611,-0.3581899405,0.2608489096,0.5059005618,0.0557413064,0.105483152,-0.1402865052,-0.1835580617,-0.0249195937,-0.0150764864,0.4873280227,0.1408146918,0.4343192577,-0.0463023707,-0.0751455873,-0.3166305721,-0.1409009695,0.1793272495,0.0424594954,0.1620036364,-0.0079394812,0.2958650291,-0.1073197722,-0.1839337498,0.0266968589,-0.2430542558,0.2493199855,-0.2464721948,-0.1109418049,-0.1289537698,-0.558372736,-0.2665817738,-0.3323172629,-0.1311868429,-0.5264694691,0.1038264111,-0.1916606277,-0.1123015732,0.128316015,0.1056971103,-0.0243232902,-0.0444190837,0.1569665372,-0.2142945379,-0.1653515995,-0.1173278242,0.1560385972,-0.2708653212,0.4570830464,0.4338180125,-0.0473659672,0.0558324568,0.1900101602,-0.2157861739,0.0246518962,-0.2156127542,0.1521595567,0.3204288483,0.0782325119,-0.182861641,-0.0671179444,0.25352332,-0.3202412128,-0.0237710886,-0.0198656581,0.2225714922,-0.0155037967,-0.2021040618,-0.4335986674,-0.394795984,-0.3217040896,0.1859988868,0.1891804188,0.2835555971,0.2589893341,-0.2054529488,0.2091143727,-0.1331139356,0.0853846595,-0.171961844,0.0235794857,0.229488492,-0.1855428219,-0.3066665828,-0.2265532911,-0.2201996446,0.1819408536,-0.2934229374,-0.2255307734,-0.2099859565,-0.0752437338,-0.238996774,-0.0619035102,0.2569400966,0.1041302904,0.0636051819,-0.180517599,-0.112944454,-0.1477453262,0.2459634691,0.036579866,-0.0313800313,-0.0238822885,0.0492491797,-0.1442333758,0.0468462668,0.0617376976,0.0054562958,0.3023152053,-0.24537687,0.2731411457,0.1647228599,-0.3531705737,0.0257603191,0.2292658687,0.2352350354,0.1637060642,0.0439292341,-0.2445184886,-0.0798801333,0.1636724621,-0.3426403105,-0.1025773957,0.1320109069,-0.0503705293,-0.2299634367,0.1226769462,-0.0914564058,-0.0446527973,-0.1742426157,-0.1153788269,-0.3685084581,-0.023195779,-0.0259908866,-0.6067454815,-0.2589999139,-0.3911126852,0.0911107287,0.0767776966,0.3848111331,0.0314139649,-0.1286427826,0.03499135,0.0054492084,0.6665763259,-0.3588528037,0.1318899989,0.3888178766,0.1826010346,-0.5201593637,0.0464813337,-0.5408808589,-0.0677352473,0.5847951174,0.3107598424,-0.3087505102,-0.1876550168,0.2801081538,0.1125766486,-0.1401080936,-0.2347939163,-0.2696113884,-0.1781988889,-0.2698454261,0.0171936359,0.0557700023,0.4278587699,0.1232239604,-0.1406504214,0.0721854195,-0.1652846485,0.0866666138,0.1255809069,0.3151289225,0.0645750314,-0.0838090703,0.0077919164,0.4125732183,-0.5944031477,0.2864172161,-0.4851076901,0.1927944571,-0.2273615152,-0.328479737,0.4715566933,0.1258498877,0.1030341536,-0.1134480163,-0.4936938584,0.3033245504,0.0638441965,-0.0831680596,0.2360485345,0.0592235699,-0.3293114305,-0.1521269977,0.5013444424,-0.0071325549,-0.1690050662,0.2432100326,0.2352398634,-0.2630656958,0.4415557086,-0.2041729689,0.9265600443,0.0056159873,0.0258411262,0.2870517373,-0.2789625227,0.8090059161,0.1631895453,0.2161719203,-0.6210642457,-0.2518280447,0.0850686133,0.016058594,-0.2423630804,0.2976560295,-0.2355505526,0.4589768946,-0.0121192401,-0.1135623306,0.0279942062,0.2951656878,0.0156730972,-0.1405717432,-0.4649794996,0.1599099934,-0.0662333518,0.0484137945,-0.1477064788,-0.1353217065,-0.1125904545,-0.2703467011,0.0508750156,0.2572171986,-0.1887545288,-0.105067499,0.3840775192,-0.0969939381,-0.4764280915,0.1886082888,0.1645728797,0.1683147699,-0.0254909135,-0.0002758027,0.2108842283,-0.0234198254,0.2289783657,0.247936964,0.3653315306,-0.1223271564,-0.0973775461,-0.1338676363,0.064932853,-0.1845349669,-0.1692492962,0.127954796,-0.0452740155,-0.1366738677,-0.358142823,0.0514153466,-0.0300162919,-0.2975687087,0.2069406807,0.1305554956,-0.3883197904,0.1677464992,0.1961372644,-0.2346573919,-0.0455334298,0.2453290522,0.0336952768,0.3841931224,0.2047000676,0.0815010443,-0.0885000303,-0.3397399783,-0.1529927105,0.2341195047,-0.2854889333,0.1031944752,0.0246006195,-0.0432794392,-0.0187416896,0.0296234526,0.0010695281,0.1826490164,0.040926747,-0.5421478152,-0.1911755949,0.206156522,0.0697732717,0.264770031,0.0933379978,0.095425494,0.1538645178,0.3309948742,-0.4655721784,-0.0665434748,-0.2060849816,0.0350940004,0.3617003262,-0.0381532013,0.0318448804,0.1196528003,0.3590333164,0.1681550294,-0.0275896545,-0.3633230925,-0.1218012795,0.021022018,0.011237056,0.13607499,-0.0002368721,-0.0772914141,-0.0476324931,-0.0603776127,-0.1380553991,-0.1138837337,0.0573951937,0.4368974864,0.0120547824,0.1806413829,0.1900898963,0.073707737,0.0010534968,-0.1602472067,0.0041385815,0.0584418774,-0.1170266643,-0.1028447077,-0.084579587,-0.0688165799,-0.34674716,-0.0264487267,0.0425144657,-0.2716346085,-0.0720903724,0.0922813714,0.5431329012,0.0102823777,-0.3550232351,-0.0576087236,0.0179644432,0.366271317,-0.4257030785,-0.0731651261,0.3309276402,0.1604926288,0.1730724424,0.3325448632,-0.0447354093,-0.1781766862,-0.2087679654,0.0305946898,0.2009146214,-0.4336821437,-0.2525113225,0.2735182047,-0.1162681952,-0.1262279451,0.2873553932,0.1490314603,0.065661259,0.3069005609,0.3288626075,0.4082328975,0.1511743218,-0.0830094963,-0.0291998275,0.1579968631,-0.0893348232,0.5161421895,-0.031485036,0.1521834284,0.1059786007,0.4009474516,-0.1144165918,-0.1971788704,0.0786328465,0.2710168958,-0.0787690282,-0.1097806245,0.0644101277,-0.0432889052,-0.3472925425,-0.0621940307,0.0450896621,-0.2083047926,0.3097210526,-0.0583522022,0.022623511,-0.4173946977,-0.0902321488,-0.027181698,0.3587398231,-0.4057599306,0.0858086795,-0.2127826512,0.0595117696,0.0405168198,0.2847782075,0.2470644265,0.1909408569,-0.1159373224,-0.0872841701,-0.2469024509,-0.2750295103,-0.1577946395,0.1843676716,0.390973419,0.095044896,0.4424323738,0.193641305,-0.2382703125,0.1344816834,0.3262963593,0.1697658896,-0.0001857622,0.5534906387,-0.0781988204,-0.3453863859,-0.342066586,-0.1411952823,-0.5782538652,-0.0299621951,-0.1062523946,-0.0438843779,0.1868725717,-0.3955799639,0.1576487869,0.0748442337,0.3263625503,0.0356060751,0.2900480628,-0.1921820492,-0.1839094311,-0.6267316341,0.0342988707,-0.0493500568,-0.3736487329,0.1371129304,0.0432723425,-0.0165217854,0.0078129051,-0.0029565161,0.4577889442,-0.1141474843,-0.0648862943,-0.3475700021,0.0400977097,-0.1236049458,0.0119616399,0.1670319438,-0.2817007899,0.0828907341,-0.3192810416,0.230290845,0.0500635728,0.0984978974,-0.0559355356,0.2182603925,0.1619683355,0.0825658217,0.285207063,-0.2228188366,0.0076093934,-0.0092683733,-0.1123697758,-0.258662641,0.2557783127,0.2113101482,0.3842935264,0.1529252529,0.0440454222,-0.0555644222,0.0552817136,0.0396078154,0.2648043334,0.028891515,0.2277389765,0.1584652662,-0.0786509812,-0.0090829553,0.2013971508,0.0145823248,0.0666336939,-0.0542220213,-0.5628551245,0.5456069708,-0.2962230146,-0.361948967,0.1794127375,0.3243953586,0.1885161847,-0.1862468421,-0.5841639042,-0.0850754008,0.309058547,-0.0215421524,-0.3569181263,0.2974792123,-0.0412640646,0.0561956316,0.0890012905,0.0794592649,0.383600235,-0.1798284799,-0.0454019494,-0.3063307106]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/181","title":"Cannot upload my own dataset","comments":"It's my misunderstanding. I cannot just upload a csv. I need to write a dataset loading script too.","body":"I look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```","comment_length":18,"text":"Cannot upload my own dataset\nI look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```\nIt's my misunderstanding. I cannot just upload a csv. I need to write a dataset loading script too.","embeddings":[0.1318956316,-0.1267824173,0.0566898957,-0.0285366997,0.1169987619,-0.1541467309,0.3969922364,-0.0621060431,-0.2148291469,0.2518130839,-0.0867060944,0.1588994563,-0.1717998087,0.2942122519,0.3824804425,0.1044842899,0.0907262638,0.313213408,0.039624013,-0.1279833913,-0.2627111375,0.0586452149,-0.0162037704,-0.0731670484,-0.0752847642,-0.1867671013,-0.1350204051,0.2678759992,-0.2258439064,-0.3100970387,0.4256777167,0.2096183598,0.3465848863,0.5284516215,-0.0001206829,-0.0640514717,0.1734646112,-0.2484510839,-0.5683873296,-0.3589856923,0.0835861117,-0.134131372,0.0810334161,-0.2865523398,0.2272181511,0.0529175773,0.0508734211,0.1323656142,0.5366080403,0.4095071852,0.1244776621,0.0669096932,-0.21695292,-0.1015119627,0.0081962785,0.1687242389,-0.0312566347,0.2934638262,0.319208473,-0.4081138968,0.2917584777,-0.2426822186,-0.0390427783,-0.0072850897,0.5370677114,0.0545539185,-0.0537988208,-0.4257257581,0.0046937442,0.1571441889,0.2818421721,0.0336492211,-0.328982234,-0.1129509732,0.1115847304,-0.5599861145,0.209152475,0.3053469658,-0.4256556034,-0.038526576,-0.2067479193,0.3366687596,-0.3919653594,0.4759362936,0.115687333,0.3350119591,-0.0372300595,0.1301802993,0.286183387,-0.0934038758,-0.0096256062,0.0732585043,0.0251915678,0.2944706678,-0.1664416045,-0.1814347655,-0.030566372,-0.2799897492,-0.1150280088,0.0909378901,0.4687089622,-0.1892622262,-0.2184372246,0.1630108356,0.1149361059,0.3294886053,-0.0234385896,0.052271124,0.1015268862,-0.0250006132,0.3100626469,-0.0896840543,-0.2538237274,-0.1026260257,-0.1475954801,-0.0238781832,0.2796851993,0.1287663728,0.0125933634,-0.1362918764,0.0550912954,0.0633083433,-0.1616130918,0.3230094314,-0.0883520097,-0.2684028149,0.0020205732,0.2783103585,0.1604540348,-0.2968745828,0.1016653851,0.3348389566,-0.1052022278,-0.1559351981,0.3859619498,0.0261231866,0.1852370948,-0.059755411,0.2265562862,-0.1464707106,0.0220260415,-0.0348720923,-0.1694041044,0.0149150034,0.2372968048,0.3928617537,-0.039099019,-0.5896613002,-0.1137735993,0.0959934145,-0.4078327417,-0.3410403728,-0.4582659602,0.0567067675,0.0479315221,-0.1880596131,-0.3508000672,-0.1340391636,0.100757651,-0.2789129019,-0.0907559171,0.0103238439,-0.0874788389,-0.1320454329,0.0613688715,0.0411118232,-0.5356760025,0.2628814578,-0.0881018788,0.5520697236,0.3589294255,0.2969783545,-0.3132716715,0.3023732305,-0.0438669398,0.1726223975,0.5356270671,-0.4726380408,-0.2246148288,0.0079028793,-0.0368109755,-0.1761097908,0.1135248244,0.1971012652,0.1221957281,-0.0631406084,0.0917988345,0.3859574795,0.0503729843,0.0187241938,-0.1974925548,-0.1673998982,0.2501933873,-0.0269129984,-0.1687581539,0.118255876,0.2636357844,0.0289663654,0.0302107874,-0.3133385479,0.119590722,0.2181630135,0.3369296789,-0.1820133775,-0.1540909559,0.2823067605,-0.4028119147,0.1659925282,0.1726571023,0.3474863768,-0.0552689508,-0.0441463403,-0.4338443279,-0.1111389622,-0.020369539,-0.0253519472,0.0457756147,0.1155959293,0.0229346696,-0.2324661314,-0.2718806565,0.106889233,-0.260445863,0.0493842512,-0.2114430219,-0.0395634696,-0.3254660666,-0.0634859204,0.2559999228,0.1197216883,0.2038004696,-0.012564647,-0.1673868001,0.1078442335,-0.4246124029,0.2669951618,-0.0308554564,0.1198563874,0.0077462881,-0.0065248855,-0.2597014606,-0.0786653385,0.2222268283,-0.2051438391,-0.384041369,0.1664907783,-0.0096187619,0.0653093308,0.1281350255,0.0084469281,0.2979522347,-0.1877758503,0.0634716079,0.1433500201,0.3599375188,0.1193041354,0.3806166947,-0.0215531122,-0.1204044446,-0.1549840271,0.1685704142,0.0108447364,0.1322262585,0.0583917722,0.0619405694,-0.0786255747,0.3246955276,0.5969873667,0.3846938312,0.0355436765,-0.0649995431,-0.0080587473,-0.1009131745,-0.2345053107,-0.0254966617,-0.0430704616,0.0423278436,0.1619130969,0.3972329795,0.229597196,-0.0900117978,0.0274449047,0.1321057826,0.4263569117,-0.2129421979,0.2630126178,-0.10497576,-0.4120586812,0.281242758,-0.2462048382,-0.067263253,-0.154713124,-0.1818923205,0.1328585297,0.1023472026,-0.1452899575,0.1561025977,0.3768164814,-0.1118422374,-0.2587055266,-0.2300248295,-0.0944366455,-0.2048037052,0.0228916407,0.282281965,-0.0796214938,0.3960422277,0.2977897227,0.0591257364,0.0036398415,-0.3079999387,-0.0184835792,-0.0448784009,0.1012091786,0.0983481482,0.3818153739,0.0118331155,-0.5280569196,0.1464271843,0.0774341896,-0.2765727043,0.2484791279,-0.0213040691,-0.3578555286,-0.4833171666,-0.0259705801,-0.2373687327,-0.3723592162,0.2339429259,0.3423468173,0.2010428011,0.6263756156,0.0330815539,0.1055231318,0.3673251569,0.15832524,0.0377818011,-0.1425475925,0.4890125096,-0.2197071463,-0.4154717326,0.1447934657,-0.1579416543,0.5447530746,-0.0712947398,-0.3519460559,-0.1823217869,0.0811840668,0.1464411169,-0.0377867222,-0.0392003171,0.3170751929,-0.2450888455,0.0634948164,-0.0740778372,-0.3456253111,0.1111996695,0.6317626238,0.2669855654,0.0019004043,0.4668571055,0.0014619923,0.7395948768,0.0747235268,-0.292644769,0.4109621644,-0.1004676968,0.2557592094,-0.1761987358,-0.3271772563,0.0276228432,0.1809422523,-0.1293832064,0.1223582774,-0.0567714684,-0.0739911348,-0.4756852686,0.3063322902,-0.2941300869,0.0496866852,0.0100904191,0.1070074141,-0.0096911667,-0.0785009339,-0.2286139876,-0.3366196752,-0.4889740944,-0.1889460534,0.272603929,-0.2198997587,0.1720749885,-0.6565859914,-0.13269867,-0.6450010538,0.3082123995,-0.1679037213,0.3023480773,-0.1833533198,0.1802234501,0.1702102125,-0.0694299042,0.1479043961,-0.38897416,-0.201797545,-0.1040052176,0.318189472,-0.4269340336,-0.0653618574,-0.0728006884,0.082064949,0.4818255901,0.203553915,-0.0144850351,0.0970021486,-0.1921155304,-0.1825740337,-0.2068865448,-0.2178378552,-0.0810333714,-0.6945078969,-0.3678599894,-0.5016222,0.122737579,0.1390094757,0.1468932033,-0.030242376,0.2109566927,0.0878909156,0.4853440821,0.004931198,-0.0599366575,0.6008954048,-0.1651565284,-0.4717906415,0.4963052273,-0.1177104041,0.3781428039,0.0914768428,-0.1569361687,-0.1381997317,-0.3253128231,0.4744020998,0.2339234352,0.1031671911,0.2817531228,-0.0628667921,0.0260088351,-0.1506823003,0.4732232988,0.3427963257,-0.1364436746,-0.3455303013,-0.1474513859,0.4850682914,-0.2086669654,-0.0615280271,0.2201681435,-0.2048091143,-0.2317992002,0.2176339477,0.1648198217,1.2438877821,-0.0595956966,0.4256589115,0.0745675117,-0.0327597968,0.7743020058,-0.166334644,0.068158634,-0.2318925112,0.1349936873,-0.1414010078,-0.0612532757,-0.1139107794,0.1102680787,-0.2297602147,0.1084133759,-0.1135409251,-0.0428984538,0.0006202807,0.451251626,0.1017268375,-0.3414739072,-0.1738090664,0.0613442473,-0.2538861632,0.3133541048,-0.1798343807,-0.2180389464,-0.3014192581,-0.2326198518,-0.18177858,0.2445215434,-0.2698900998,0.0779517293,0.0035693035,-0.322989583,0.1731056571,0.4630560279,0.510530591,-0.0635785609,-0.0687162876,-0.0096662417,-0.0855765045,-0.0423675925,-0.0929513872,0.0601842515,0.1363020837,-0.2955884933,0.0736771077,0.3211552203,-0.0851583108,-0.3732730448,-0.21020706,0.1948630959,0.0594121739,-0.134041518,-0.3536510468,0.0659634098,-0.212487638,-0.0157401673,0.0045342492,0.0331519842,0.0830105096,-0.0901566744,0.0896118656,-0.1979517043,0.0944740847,0.0631254613,-0.0416487679,-0.0489926301,0.7007597089,-0.1211075038,0.0802455172,-0.0431203656,-0.2052936852,-0.3344253898,-0.1463949382,-0.0778771713,0.3191208541,-0.0304467604,0.0391583219,0.6615247726,0.3981347382,0.1872176528,0.0680559799,-0.1111062989,-0.3231564164,-0.3133419752,-0.034411747,0.2028845996,-0.0251049455,0.2131589502,0.2098329216,0.2966146469,-0.2147560418,-0.1815819591,-0.2030589283,0.1776479334,0.7618901134,-0.0131239332,-0.1111019403,-0.0103951786,-0.0331597254,0.0095930006,-0.1480641961,-0.103352204,-0.3793171346,0.1847985089,0.5306537747,-0.1196051612,-0.0148341628,-0.2393809557,-0.1119954512,-0.1213679239,-0.3297748566,0.0416503847,-0.0146457329,0.2055361718,-0.3102743924,0.3925669491,-0.1536854953,0.1730086356,0.0203283466,0.2511644065,0.0605651066,-0.0521633625,0.1532326937,0.0744908005,-0.3929888606,0.0065550376,-0.1775214225,-0.0990986601,0.0257294215,-0.4181020856,0.157932058,-0.2215919346,0.0222716834,0.0460855924,-0.5020128489,0.2189573348,0.4416899383,0.0885728374,0.0407045931,0.1663511842,0.0706042498,0.213899672,0.0241184905,0.275057137,-0.0007004683,-0.3031487465,-0.0489965864,-0.227034241,-0.1142665669,-0.2721327245,0.0433908887,-0.0608787835,-0.4242881238,0.0046088663,0.1704251915,0.1590811163,-0.1035111248,0.1558342278,-0.1988390684,0.0952536911,0.0563986525,-0.1158856153,0.1838596612,-0.2343326956,0.1283363849,-0.0451820865,-0.007980708,0.3380725682,0.1622446477,0.6163566113,-0.3498658836,-0.0113322129,-0.2669867575,0.0580319837,0.0238670036,0.1797101796,-0.660415709,-0.0053761932,0.0684240982,-0.0865073577,0.03860192,-0.3267984986,-0.0503800474,0.2629701197,-0.1634489298,0.1078680009,-0.1304238141,0.1925937831,0.1316483915,-0.1126357913,0.3684443831,0.0031906615,-0.0359711871,-0.0683030114,0.0586663149,0.5559293032,0.617588222,0.168113783,0.2705543339,0.2301716059,-0.0069069783,0.0774815455,0.1667327434,0.4616708755,0.3019829094,0.1201505885,0.0308131445,-0.1213352829,0.1484298706,0.0533193536,0.0956460834,-0.3158128858,0.2280452549,-0.0291061513,-0.3997559547,-0.1564083695,0.4017167091,-0.4579074383,-0.0140898237,0.190097183,0.1823272407,0.3674243689,-0.2338710874,0.0223054178,-0.1849538833,0.3975650668,0.4656152725,-0.0361273028,-0.1464556605,-0.219589904,-0.4871545732,0.177601248,-0.2229430079,-0.2007381618,-0.3758898377,0.1740441024,0.0622250251,-0.0186625812,0.0783121213,0.0861965939,-0.2741782069,0.0232395213,-0.3622923791,-0.020600779,-0.1225058585,-0.0025120997,-0.0768298209,-0.1665668488,0.0395683944,0.1100984663,0.0128767453,0.0241478793,0.2424088269,0.1370490789,0.3673104346,0.4786677659,-0.0718720704,0.7685264349,0.0670453832,0.0205732882,0.0492995679,-0.0422592536,-0.1771253943,-0.0018188178,0.1784064621,0.3781342804,-0.0588379614,0.349863708,-0.1977619529,0.0625041798,-0.0225883611,-0.5348159671,0.0552991778,-0.0063379309,-0.0871976614,0.230238542,0.4238304198,0.0884501636,0.0706020147,-0.1437548101,-0.4335863888,-0.2854062915,0.41403687,-0.5086393952,-0.3003112078,0.2236109227,0.2395336479,-0.0293427408,0.0527283587,-0.2230433822,-0.0850507692,0.135616079,-0.2508603036,-0.0518899672,0.3241963387,-0.5186468959,0.0162721314,-0.0552875437,0.3049903214,0.202418685,-0.3796506822,-0.0708045289,-0.2150799185]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/181","title":"Cannot upload my own dataset","comments":"I now try with the sample `datasets\/csv` folder. \r\n\r\n nlp-cli upload csv\r\n\r\nThe error is still the same\r\n\r\n```\r\n2020-05-21 17:20:56.394659: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/csv\/csv.py to S3 under filename csv\/csv.py and namespace korakot\r\nAbout to upload file \/content\/csv\/dummy\/0.0.0\/dummy_data.zip to S3 under filename csv\/dummy\/0.0.0\/dummy_data.zip and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```\r\n","body":"I look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```","comment_length":116,"text":"Cannot upload my own dataset\nI look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```\nI now try with the sample `datasets\/csv` folder. \r\n\r\n nlp-cli upload csv\r\n\r\nThe error is still the same\r\n\r\n```\r\n2020-05-21 17:20:56.394659: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/csv\/csv.py to S3 under filename csv\/csv.py and namespace korakot\r\nAbout to upload file \/content\/csv\/dummy\/0.0.0\/dummy_data.zip to S3 under filename csv\/dummy\/0.0.0\/dummy_data.zip and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```\r\n","embeddings":[0.0526046231,-0.0281094722,0.0553304181,-0.0020762177,0.145578444,-0.1847578585,0.3376964033,-0.0477224775,-0.192668274,0.1797223389,-0.1303296536,0.2436908185,-0.21156165,0.2791502476,0.3693570495,0.0698641911,0.0132651599,0.2872202396,0.0126219867,-0.1083282828,-0.2543199956,0.091935806,-0.019309666,-0.0809588656,-0.1234256402,-0.1938010007,-0.08119528,0.2602367401,-0.223822698,-0.3616123497,0.4911182821,0.234791398,0.3647105396,0.579341054,-0.0001194561,-0.0481620245,0.1620747,-0.221251756,-0.6242922544,-0.4079536498,0.0205320753,-0.1510945857,0.070411697,-0.2569880784,0.1892614663,0.023185974,0.0217074882,0.0470516086,0.489642024,0.4066984653,0.137758106,0.1278687418,-0.2263692915,-0.1397725344,0.0078123156,0.173207283,-0.0587553121,0.2595106363,0.334825933,-0.4004638493,0.2734425962,-0.1960979998,0.0127515486,-0.0379062071,0.4887399375,-0.0217540264,-0.0607262999,-0.3897155821,-0.0111776432,0.1542044729,0.3593457043,-0.0174938589,-0.3905940056,-0.12496797,0.1351312995,-0.4883084893,0.2485231161,0.2740954161,-0.4021524489,-0.0272089615,-0.2619934678,0.3479141295,-0.4030778706,0.4445774555,0.1077126414,0.3708469868,-0.0370596536,0.0913995504,0.2884841561,-0.1540009379,0.0222615097,0.0381423905,0.0280135628,0.2681338489,-0.2156573981,-0.1543159336,0.0116408691,-0.2373854071,-0.0875118226,0.1148514226,0.4278518558,-0.1259322017,-0.2469718456,0.1405272335,0.1018434167,0.3375079334,0.0230610128,0.1245041862,0.1315496266,0.0036603063,0.2448296547,-0.0774665773,-0.294364959,-0.0886220858,-0.1120098606,-0.0521204583,0.2688553929,0.0746984109,-0.0075792344,-0.1084796488,0.078623794,0.0423113853,-0.0832311437,0.3526001275,-0.0594029352,-0.2365911007,-0.0201667901,0.2576301992,0.1414344907,-0.2609570622,0.0658907294,0.2955011129,-0.1243116632,-0.1489530802,0.4141170382,0.0235561319,0.1915086955,-0.0685793608,0.2129649371,-0.114051573,0.0568696,0.0295375474,-0.1648236364,0.0172457229,0.2235611975,0.3928875923,-0.0523391664,-0.5404902697,-0.1403445899,0.0234505367,-0.3477195799,-0.357789427,-0.4238156378,0.0750945732,0.020578349,-0.2402535379,-0.3374996483,-0.0972320661,0.0914815366,-0.2820447981,-0.0950792953,-0.045189213,-0.1562713236,-0.1163597256,0.1120778993,0.062964119,-0.4992431104,0.2119519413,-0.0793736652,0.5196451545,0.2760166526,0.3038301468,-0.3427072167,0.2330141068,-0.0663762987,0.1847803295,0.5762096643,-0.4968986511,-0.3117572069,0.0467118211,-0.0305642225,-0.1179906055,0.0722416341,0.1715793163,0.1136374399,-0.0842440352,0.1021822467,0.4385216832,0.067252405,0.0401820317,-0.2208324224,-0.1542976499,0.2543805838,0.0369843766,-0.1506792009,0.083122924,0.260474056,0.0628705621,0.0104562873,-0.3288799822,0.1415969878,0.2101691961,0.3580575585,-0.1686786115,-0.1278675348,0.2173071802,-0.4310545027,0.2184393257,0.1891910285,0.3485833108,-0.1336560696,-0.0364643224,-0.4364340007,-0.1022443995,-0.1047702953,-0.0485605337,0.0417283028,0.1295845807,0.0738745704,-0.1800095141,-0.222815603,0.1598291248,-0.2588529289,0.0683703944,-0.210019201,0.0358617045,-0.2738796175,-0.0956867859,0.2633078992,0.1670373529,0.2156515867,-0.0159932766,-0.2017098516,0.0932927281,-0.4534426332,0.2857999206,0.0175593048,0.1001565158,0.0362415425,-0.0552057512,-0.2138194591,-0.1098800674,0.1715619713,-0.1524938941,-0.3677259088,0.1745695621,-0.0044881394,0.0360841341,0.166636318,0.0101487059,0.2959846854,-0.1779489815,0.0921189636,0.0983786359,0.417483747,0.093171522,0.403357625,-0.0110187316,-0.1426615566,-0.0857242867,0.1682548672,-0.0006490924,0.1505849063,0.0749115199,0.0242283382,-0.0512584634,0.3066678345,0.6624006629,0.3770546615,0.0858841166,-0.026365513,-0.0519595072,-0.1250855476,-0.2106573433,-0.0400317982,-0.0779025778,0.0588910915,0.1829154342,0.4417646825,0.2363879383,-0.1361708194,-0.0484123416,0.1532731056,0.4534134269,-0.1956255138,0.2583765388,-0.1609421372,-0.4299409091,0.2424896061,-0.2555158734,-0.0626863539,-0.1816957295,-0.2170829773,0.146137014,0.1071710587,-0.0841627643,0.1007013097,0.3315699697,-0.1059278175,-0.2866389453,-0.2544071376,-0.1137493402,-0.1918030977,0.0277642626,0.3250083029,-0.0884016827,0.4124760032,0.3248804808,0.0886461809,-0.0461392142,-0.3268382251,-0.0045968406,-0.0374101885,0.1208912134,0.1213874146,0.356371671,0.0086652311,-0.5445611477,0.1948808581,0.0784640685,-0.2436255813,0.251660347,-0.0325384811,-0.3485889137,-0.4676508904,-0.0208269879,-0.2737125158,-0.36877051,0.239671275,0.2940942347,0.2099166214,0.5754036307,0.0177124944,0.1252833754,0.3206279576,0.1239150017,0.0683422983,-0.1860464215,0.5255692005,-0.2455919087,-0.4252485633,0.1025658846,-0.1634910107,0.5601887107,-0.0097183557,-0.3501398861,-0.1784346998,0.0594795719,0.2183053195,-0.0234066676,-0.0674496144,0.2789646685,-0.2049719989,0.0612629019,-0.0819354728,-0.3369690478,0.1344738603,0.6249414682,0.2558960021,-0.0422820225,0.4419958591,0.0380829684,0.702817142,0.1071994007,-0.3385474384,0.4494320452,-0.0385142528,0.1877281666,-0.1323684603,-0.3693924844,0.0650609732,0.2254458368,-0.061309617,0.1397267282,-0.0222405326,-0.1021208689,-0.4349323511,0.2301696837,-0.2566095889,0.0486133657,-0.00560317,0.1051201746,0.0555104949,-0.0678321421,-0.2128119916,-0.3788439929,-0.4829711914,-0.1628735662,0.274995029,-0.2391632497,0.1747967452,-0.6903293133,-0.0932890475,-0.6602335572,0.300175041,-0.1387225091,0.3314152658,-0.2023537159,0.2067674547,0.1133597121,-0.0766510293,0.1280318499,-0.4650164843,-0.16798006,-0.0770433098,0.2747678459,-0.4723593593,-0.0419122875,-0.0276053585,0.0749130622,0.4365334809,0.2350970507,0.0189995226,0.0795564577,-0.1766324788,-0.2030602098,-0.1537004709,-0.2233584523,-0.0368927829,-0.6711586714,-0.3523872197,-0.4475812614,0.0672138184,0.1456528902,0.1405826807,-0.0376426019,0.1860164851,0.1109216437,0.5076415539,0.0428426117,-0.0462599546,0.6295244098,-0.0930514336,-0.4571355581,0.4725039005,-0.0099429451,0.389480412,0.0355944559,-0.194271937,-0.1696300507,-0.2850331366,0.4578328133,0.1723040938,0.0854799747,0.2172770202,-0.1069508418,0.0281256791,-0.1549408287,0.4699554443,0.3389653862,-0.0914734527,-0.4181351364,-0.1839520931,0.449573487,-0.1790728867,-0.0337394476,0.2428125292,-0.1924960315,-0.276345998,0.2006180882,0.1193980351,1.0911499262,-0.0767855495,0.3788450956,0.1221328303,-0.0487476774,0.7742523551,-0.1455495805,0.0457032733,-0.233587727,0.1004334539,-0.1133074388,-0.0588087589,-0.0676055551,0.1411717087,-0.2373992503,0.0841967762,-0.0400056466,-0.1112078428,0.0072581493,0.552326858,0.0727096274,-0.3941997886,-0.220260486,0.0838499814,-0.3112987876,0.3093791306,-0.1699573845,-0.2196554095,-0.2857396901,-0.1824826449,-0.207254529,0.2760237753,-0.2514284551,0.0993921608,0.0488731414,-0.3554010987,0.3192008138,0.4633531868,0.524256289,-0.0765047669,-0.0971658006,0.0498913489,-0.1302103251,-0.0141759282,-0.0387769528,0.0305916592,0.1902915686,-0.2695977986,0.0427540168,0.2530287504,-0.0257400349,-0.3122657835,-0.1774225831,0.1460345089,0.0467341729,-0.1632915735,-0.3530443311,0.0866535529,-0.1731006652,-0.0444597863,0.0278086308,0.0287349839,0.0737266764,-0.0680368617,0.0925293416,-0.2007798105,0.0808693543,0.1064461842,-0.1156514138,-0.0588401072,0.741160512,-0.0600612797,0.0792916864,-0.0856437087,-0.1985215098,-0.3181911707,-0.1407571435,-0.0897720978,0.2795092463,-0.0309429467,0.0831823945,0.6403141618,0.3384789526,0.1092205644,0.1618171632,-0.1482797414,-0.3228501081,-0.3060598969,-0.0438086875,0.1931378394,-0.1073027179,0.1915976405,0.2019474357,0.2981015444,-0.2282575071,-0.1637889296,-0.2005669475,0.1670419276,0.7540919781,-0.0116151581,-0.0665543601,-0.0367735103,-0.0189000815,0.002394943,-0.1610791385,-0.1169768274,-0.3814634383,0.1755768061,0.4809396863,-0.1895360351,-0.0424814411,-0.2404210716,-0.0995077714,-0.1417815685,-0.2967741787,0.0675631016,-0.0337111689,0.24088943,-0.2643099725,0.3360068798,-0.0966443717,0.2147415429,0.0065671541,0.1791659445,0.0874761567,-0.0227311812,0.195901826,0.0774845779,-0.3794807792,-0.0381360874,-0.20470348,-0.0682202503,0.0289715659,-0.4525889158,0.1307521611,-0.1122686714,0.0439764485,0.0357846245,-0.4939005375,0.1632531583,0.3917883933,0.1032580584,0.0149257984,0.118500866,0.1698049009,0.2038431317,0.0200777352,0.2445479929,0.0348778181,-0.2913975716,-0.0517354868,-0.2043577433,-0.0584846139,-0.2675398588,0.1212550327,-0.0928469226,-0.3782732785,-0.0432638302,0.153394416,0.1886675209,-0.1023825482,0.1934892535,-0.2057672143,0.1222846434,0.0327518173,-0.0728969797,0.2181709558,-0.2738108933,0.1113118231,-0.0672394633,-0.0237731002,0.32178092,0.1499329358,0.6510175467,-0.3514948189,-0.0054610702,-0.2674220204,0.1102579236,0.0091688205,0.1439594477,-0.6295008659,-0.0255037062,0.0543598905,-0.0833841786,0.0133432308,-0.3383154273,-0.0363659412,0.2841472328,-0.1411570013,0.0500104576,-0.1062960774,0.210164398,0.1268292814,-0.1563307792,0.4182304442,0.0171362571,-0.1046922132,-0.048103679,0.0715438277,0.5066614151,0.5816252232,0.1312174648,0.2830177248,0.181376785,0.0337430984,0.0606108978,0.1921532154,0.4136766195,0.2930335402,0.1462849975,0.02923158,-0.1140508279,0.0791373178,0.04990004,0.0570766628,-0.3188941777,0.2812895775,-0.0311866179,-0.3361921608,-0.1678582281,0.3929540217,-0.4424993396,0.0111441622,0.2605808079,0.2248203903,0.3537732363,-0.2567598224,0.0243905317,-0.1696015,0.4160138667,0.4884312451,-0.0781686604,-0.1432206929,-0.2934460938,-0.5263960958,0.1610699594,-0.216284737,-0.2315175235,-0.3568784297,0.1811114997,0.0572608374,-0.0962691456,0.0581554547,0.1449394971,-0.2326552123,-0.0044060578,-0.4434360862,0.0147196045,-0.1537976414,-0.0142655699,-0.03849338,-0.1682831347,0.0295242611,0.0920236483,0.0187680144,-0.0066740024,0.2200709283,0.1332621127,0.3443261683,0.4730619788,-0.0944239125,0.7783322334,0.092848666,0.0046176463,-0.0069649196,-0.0658429712,-0.1645212024,0.0251689069,0.136098966,0.4669876993,-0.0853173211,0.373172164,-0.2298848033,0.0358114019,-0.0314073078,-0.5221913457,0.0767316148,0.009969431,-0.0682161897,0.2494254112,0.4241001904,0.1379924715,0.0984187499,-0.1065407172,-0.4279006422,-0.3111038804,0.4142977893,-0.5499933958,-0.3333510756,0.1943292916,0.2238215804,-0.0524814017,0.0226925798,-0.2831411362,-0.0702407509,0.1561145186,-0.2359622866,-0.0192985907,0.3056487143,-0.4318902791,0.0704646185,-0.0529117808,0.3128914237,0.1947142333,-0.3801876009,-0.04024636,-0.1632496864]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/181","title":"Cannot upload my own dataset","comments":"We haven't tested the dataset upload feature yet cc @julien-c \r\nThis is on our short\/mid-term roadmap though","body":"I look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```","comment_length":17,"text":"Cannot upload my own dataset\nI look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```\nWe haven't tested the dataset upload feature yet cc @julien-c \r\nThis is on our short\/mid-term roadmap though","embeddings":[0.010165412,-0.0146975406,0.0639977977,-0.0472517423,0.1368083656,-0.1440622509,0.3703480661,-0.0352642201,-0.1715517193,0.1538790166,-0.1031072065,0.2589793801,-0.230888322,0.3168729246,0.3755461276,0.0948919877,0.019371815,0.2717739344,-0.0201163273,-0.0985039175,-0.2267878354,0.0701328591,-0.0022228684,-0.0957495868,-0.142616719,-0.2116264999,-0.0903275609,0.2581228912,-0.234555155,-0.341229558,0.4868592024,0.2359908074,0.3435915709,0.5804950595,-0.0001179577,-0.0519963317,0.1718314439,-0.2157051712,-0.6393771172,-0.3554878533,-0.0061307894,-0.134217307,0.0606778674,-0.2517402768,0.1928754896,0.0274626482,0.0349178202,0.0308800731,0.4822887778,0.4076221883,0.1489153057,0.1457220614,-0.1781223565,-0.1500244886,-0.0028327829,0.18243514,-0.0589718297,0.2539565563,0.3586966693,-0.4081481099,0.2644420862,-0.1995871067,0.0367807001,-0.0442407019,0.487485081,-0.0092500607,-0.0749194771,-0.4059659541,-0.0048401151,0.1771373749,0.3769998252,-0.0600269549,-0.3945055604,-0.145692274,0.1550535113,-0.4790179133,0.2537232935,0.2788683176,-0.424706161,-0.0225299057,-0.2711575925,0.3137456477,-0.4093555212,0.4221662879,0.1056479365,0.3526955247,-0.0337093771,0.0666819736,0.243746832,-0.1536932439,0.0117682517,0.0654737875,0.0214675888,0.2782631218,-0.2166639417,-0.1950291097,0.0027943654,-0.2405849695,-0.0581812523,0.129156217,0.4047325552,-0.1296833754,-0.2442011088,0.1506330818,0.133687526,0.3241246641,0.0014827683,0.1223634258,0.1223344952,-0.0058462271,0.2769157588,-0.051742468,-0.2742703557,-0.0508025326,-0.1130108014,-0.0688472167,0.2920281887,0.0677258074,-0.0055412739,-0.1293468624,0.0973584875,0.0165869407,-0.1025609896,0.3439234495,-0.0819455087,-0.2613025904,-0.0538049601,0.2326337695,0.1370870918,-0.2900106013,0.0570580028,0.2758283615,-0.1154935136,-0.1544326395,0.4356733561,0.0124119911,0.1956504881,-0.0544522628,0.2021948993,-0.0660793781,0.0722418875,0.0119863097,-0.1628797501,0.0086813662,0.2195927054,0.384788096,-0.0684984699,-0.5333330035,-0.1435816735,0.0261609294,-0.3532370031,-0.3473283052,-0.4594140351,0.0884744301,0.0037644345,-0.2817487121,-0.3244448006,-0.0715043247,0.0896539763,-0.2517111599,-0.0952162668,-0.0463004708,-0.1421903223,-0.122484535,0.1295616031,0.0880389139,-0.5273433328,0.184871614,-0.0860141665,0.4892792404,0.2712750137,0.2785508037,-0.3086805344,0.2438846678,-0.0801130012,0.1897517443,0.5997764468,-0.4807251692,-0.3346849978,0.0427862369,-0.0373171307,-0.0900270417,0.0372784212,0.1934347749,0.1080194935,-0.0963605344,0.0952949747,0.4479237199,0.0456785262,0.0166918822,-0.2161557674,-0.1669479311,0.2624107003,0.0407923721,-0.1555023193,0.0785308406,0.2734291255,0.0899971277,0.0241547208,-0.3324550092,0.1537160873,0.1996024102,0.3573826253,-0.1757684499,-0.1370925009,0.206030041,-0.4696049094,0.2170207202,0.1805566847,0.3490841985,-0.1210269406,-0.0183037631,-0.4437145293,-0.1005717665,-0.1240304336,-0.0586681627,0.0490877256,0.1163989827,0.0549807139,-0.1700619161,-0.2535095513,0.18731682,-0.2769084275,0.0612526424,-0.2197418362,0.0450597033,-0.2696869969,-0.1241740063,0.2587273717,0.1824533641,0.1657662839,0.0060952953,-0.1836488992,0.1066876352,-0.478485167,0.262398541,0.0336905457,0.1265535951,0.0412620455,-0.0676175654,-0.1832831055,-0.1436622143,0.1708056331,-0.1465517282,-0.4013840854,0.1432179064,-0.0279377457,0.0045213141,0.1643311828,0.0004223413,0.279594779,-0.179331392,0.1159435213,0.1068178341,0.3852045238,0.066510573,0.4020833373,-0.0413434729,-0.1261366606,-0.0323795192,0.2109304219,0.0192543678,0.1892586797,0.0736541077,0.0137931267,-0.0801438838,0.2915304005,0.6511793733,0.380454123,0.1110787094,-0.0320215598,-0.0562743917,-0.1082973033,-0.1881228238,-0.063018389,-0.1195570678,0.0715672448,0.1802178025,0.4236962199,0.2545208633,-0.1388218552,-0.0360268503,0.1298332661,0.4580771625,-0.1908658743,0.2338176519,-0.1396615803,-0.438040942,0.2469321489,-0.2298305929,-0.0904453471,-0.1691734046,-0.2033484131,0.1343929023,0.1450604498,-0.0874417275,0.1189224645,0.3378386497,-0.084738709,-0.2890054882,-0.2606640458,-0.0930047929,-0.2001828402,0.0532764979,0.3222643435,-0.1070774347,0.4432595372,0.3364593089,0.0863811001,-0.0549170114,-0.3232798576,0.0085717281,-0.0607416742,0.127227053,0.1032358482,0.343431294,-0.0030982832,-0.5138462782,0.1799697429,0.0642064512,-0.2702479959,0.2215505093,-0.0224287286,-0.3159378171,-0.4438218772,-0.0553473644,-0.2909525335,-0.3391727805,0.2574870884,0.3296154439,0.1998353153,0.5830568075,0.0271325316,0.1387715936,0.3214481771,0.1015031114,0.0490918346,-0.2018515617,0.5142780542,-0.2508527339,-0.4668107033,0.1272697002,-0.1354223341,0.5388168693,-0.0261844639,-0.3549962938,-0.2038773447,0.0651994944,0.2458015382,-0.0366456285,-0.0519297607,0.2837885618,-0.1914527267,0.0443885289,-0.0722914264,-0.3318912089,0.1452786326,0.6041272879,0.2712104023,-0.0585435331,0.4339739084,0.0434251577,0.7070642114,0.1327071488,-0.3311067522,0.4485009015,-0.0690921396,0.2065573931,-0.1401202977,-0.3747258186,0.1332441568,0.235465765,-0.0336518213,0.1127984077,-0.0309750363,-0.0822748169,-0.4156804979,0.2208882272,-0.2339075506,0.0483934321,-0.016791543,0.1189959347,0.0460602827,-0.0811108947,-0.2233671099,-0.36375615,-0.5096309185,-0.1245183349,0.245887652,-0.2464717627,0.2104229182,-0.6700918674,-0.055322621,-0.6859092712,0.2873030305,-0.1499044895,0.3086008728,-0.2161931694,0.1993662864,0.094112806,-0.0823159292,0.1505372077,-0.4754492044,-0.1588874161,-0.0853981897,0.2848472595,-0.4748277366,-0.0264239758,-0.0513210446,0.072651282,0.4124585986,0.1920997798,0.0395092256,0.0759668946,-0.1413953006,-0.2161689848,-0.1769444942,-0.2280471921,-0.0457699411,-0.6439383626,-0.3358868062,-0.4636676311,0.0563708134,0.1376702935,0.1639883518,-0.0686658174,0.1555040777,0.0947162509,0.5056194067,0.0479242764,-0.0662424415,0.638486445,-0.0932825431,-0.4260934293,0.4850158989,0.0034875595,0.3892193437,0.0280552804,-0.1498471946,-0.1559250206,-0.2998328805,0.4713130891,0.1373466998,0.0870597363,0.184803322,-0.0992770046,0.0281231198,-0.1918720007,0.5012036562,0.3349218369,-0.0805315524,-0.4222896099,-0.1881601512,0.4366788864,-0.1674681753,-0.0022928726,0.2426342219,-0.1802781969,-0.2683978975,0.1748557985,0.1420557499,1.0450088978,-0.0981931239,0.3577184379,0.1145982742,-0.0308236554,0.7806394696,-0.1800111979,0.045405522,-0.2553684711,0.0990400836,-0.1199556142,-0.0433890931,-0.0456890613,0.1368022859,-0.2528727055,0.085742645,-0.0294855516,-0.1470440775,0.0057264739,0.5664038658,0.0617874824,-0.3942605555,-0.2046339661,0.1085034385,-0.3170492947,0.2979918122,-0.1628501564,-0.207619071,-0.2826575041,-0.1551133841,-0.2209420204,0.3035310209,-0.2398694754,0.0794946998,0.0589173697,-0.3198938966,0.3326177597,0.4678548574,0.5250076652,-0.0823699012,-0.1008318141,0.0375755578,-0.1670936495,-0.0270536151,0.0090900119,0.0502924062,0.2291195691,-0.290697962,0.030212447,0.2814525366,-0.0009342713,-0.3020415008,-0.1756666899,0.1342540979,0.0712272525,-0.1674864888,-0.3050799072,0.1047319919,-0.133216843,-0.0392335393,0.0401462317,0.0476170965,0.0688897073,-0.0286139119,0.0777553022,-0.1839819551,0.0571828634,0.1178763658,-0.1195770353,-0.0726924911,0.7520653605,-0.0539941601,0.0932504386,-0.1099424139,-0.1821333915,-0.2736782432,-0.1381536722,-0.0994902924,0.2768892646,-0.0209478829,0.0904218033,0.6383234859,0.32130301,0.0976346955,0.1383455545,-0.1498945504,-0.3523440063,-0.3358486593,-0.0205375757,0.1858695447,-0.120178774,0.1458787769,0.2181804627,0.3104346097,-0.2514614761,-0.1950040907,-0.1998571903,0.1583553553,0.7369630337,-0.0176640227,-0.0129007399,-0.0611461364,0.0069881915,-0.0301093515,-0.173837319,-0.1364713758,-0.3693220317,0.1645641327,0.4616593719,-0.191413179,-0.0562297814,-0.2185690701,-0.1284788102,-0.1337906867,-0.2693399489,0.0604025833,-0.0575630292,0.2415007651,-0.2288512886,0.3202768862,-0.0882498547,0.1737593561,0.033312656,0.1657403857,0.0758051202,-0.0196456369,0.1906692535,0.0748085901,-0.3352306783,-0.0262000058,-0.1774330735,-0.0726978257,0.0177799221,-0.4788339138,0.1080649495,-0.1383731365,0.0509617925,0.0222610589,-0.4982845783,0.1378389895,0.3878335655,0.1195944846,-0.0020744093,0.1106700599,0.2088810354,0.2025484741,0.0134531064,0.2371901721,0.0218779296,-0.2497720271,-0.074469164,-0.1981972754,-0.0792121217,-0.2815967798,0.0992150232,-0.112468265,-0.3867094517,-0.0492038652,0.1662582755,0.1972377896,-0.0932935327,0.2138920426,-0.2051724792,0.1138751805,0.0320776924,-0.0619553998,0.2281326354,-0.2797224224,0.1353702396,-0.0985191688,0.005144659,0.3340873718,0.1458950788,0.6321770549,-0.316231519,-0.0161116756,-0.2486732155,0.0746075809,0.0175966788,0.1678756624,-0.6061634421,-0.0291333906,0.052960068,-0.0852579698,0.0107766585,-0.3678248823,-0.0482248999,0.2759406269,-0.157885164,0.0588335618,-0.0975589976,0.1953690648,0.1377755105,-0.1273939461,0.4026268125,-0.0086532393,-0.0919725671,-0.0548680648,0.1125578731,0.4699504375,0.5798239112,0.1491091996,0.3191974461,0.2029526681,0.0339744277,0.0652849302,0.1831146628,0.3685241342,0.2624925077,0.1308895499,0.0407573916,-0.1079788357,0.0592564046,0.0276295412,0.0705804899,-0.3081683815,0.2848696709,-0.0199584421,-0.3238851726,-0.1779014319,0.3976590335,-0.4280272126,0.0080058742,0.3059367537,0.2120259851,0.325581193,-0.2719441056,0.0336144455,-0.1892011315,0.4501140118,0.4820170701,-0.0971010476,-0.1016004384,-0.2517859638,-0.5138528347,0.1598959118,-0.1947072297,-0.2622525692,-0.3417943418,0.189845264,0.0361181274,-0.0889950693,0.1070143878,0.1390070766,-0.2102020085,-0.0200358927,-0.4624018669,0.0136038279,-0.1590491384,0.0098757949,-0.0202057511,-0.1756241471,0.0439807624,0.0640803799,0.01580064,-0.0002465911,0.1899518371,0.1183922291,0.3266207576,0.4968595505,-0.1008909941,0.7589022517,0.1091519967,-0.0081170388,-0.0122549543,-0.0607085526,-0.2116784155,0.0700843483,0.1279249191,0.4384543598,-0.0824115723,0.3603610992,-0.22094284,0.0750983357,-0.033126574,-0.5146051049,0.084510833,0.0228470489,-0.044427868,0.2764645219,0.4149300754,0.0924856067,0.0941309705,-0.0950982347,-0.4262842834,-0.274613589,0.3977466524,-0.5510283709,-0.2920255661,0.1532051265,0.2142148465,-0.0326685123,-0.009405748,-0.3166780174,-0.0671029761,0.1787447184,-0.2281676382,-0.0117133697,0.2962084115,-0.4524932802,0.0907235071,-0.0468803905,0.3302342296,0.1773254871,-0.371445328,-0.0162002109,-0.1414589882]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/181","title":"Cannot upload my own dataset","comments":"Even if I fix the `TypeError: __init__() got an unexpected keyword argument 'cdn'` error, it looks like it still uploads to `https:\/\/s3.amazonaws.com\/models.huggingface.co\/bert\/<namespace>\/<dataset_name>` instead of `https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/<namespace>\/<dataset_name>`","body":"I look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```","comment_length":25,"text":"Cannot upload my own dataset\nI look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```\nEven if I fix the `TypeError: __init__() got an unexpected keyword argument 'cdn'` error, it looks like it still uploads to `https:\/\/s3.amazonaws.com\/models.huggingface.co\/bert\/<namespace>\/<dataset_name>` instead of `https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/<namespace>\/<dataset_name>`","embeddings":[0.0766489804,-0.0518902279,0.0566434562,-0.0133813079,0.1621728241,-0.1869212389,0.3156289756,-0.0278216116,-0.1946031451,0.1959180981,-0.1271059215,0.2358573377,-0.2207303047,0.2926344573,0.3533962965,0.0726656169,0.005575384,0.2951037288,0.0090058008,-0.0938581005,-0.2198612839,0.0948086008,0.0193501543,-0.079871662,-0.1348169595,-0.1881403178,-0.0791992322,0.275642097,-0.1972334087,-0.3347654939,0.4486150742,0.2227426618,0.3325239122,0.600295186,-0.0001181157,-0.0586367175,0.1657020003,-0.2280745208,-0.6185942292,-0.3737724423,0.073014237,-0.1427434534,0.0919025093,-0.2383846194,0.1893360168,0.0374169759,0.0164695997,0.0653847903,0.5281576514,0.3778299391,0.1473527253,0.1031371132,-0.1705768555,-0.0949941054,0.0273437556,0.189317435,-0.0560426563,0.2944778502,0.3386437297,-0.3992356062,0.2927792072,-0.1987794042,-0.0034406974,-0.0603887588,0.482424438,-0.0279464349,0.0078290747,-0.4104376435,-0.029902583,0.1773936749,0.3018230498,-0.0301785674,-0.3591761887,-0.1138800085,0.1687852591,-0.4841401875,0.2560546398,0.2725909948,-0.4102144837,-0.0385259055,-0.2570946813,0.3279775381,-0.4180808961,0.4426954389,0.1196050569,0.3393745422,-0.0226807911,0.0766781867,0.2550292909,-0.1277886331,0.0078817355,0.0909430981,0.0217479058,0.2719116211,-0.1963246763,-0.1846594959,-0.0206889845,-0.28975752,-0.1094860509,0.0998324826,0.4507453442,-0.1309796423,-0.247373417,0.1494717002,0.0931351632,0.3291782141,-0.0070715589,0.0955357477,0.128275767,0.0212816391,0.2753561437,-0.0661067888,-0.2871695161,-0.0721225441,-0.1261585504,-0.0473260097,0.2747190893,0.0848323107,0.0117681734,-0.0948850736,0.0725315735,0.0604458526,-0.0801716447,0.341994524,-0.1117420197,-0.2417011708,0.0059717465,0.2359330803,0.1392108202,-0.2688980103,0.0752417669,0.3023251593,-0.111968182,-0.1445471346,0.4047563672,0.0083994977,0.1989607066,-0.0468069166,0.1874880344,-0.1364440322,0.0624775402,0.0066449237,-0.1763243079,0.0334348194,0.2295496315,0.3857797682,-0.0658191144,-0.5420824289,-0.1311396807,0.0309668574,-0.376873821,-0.3573203981,-0.4452536106,0.0883500725,0.0414214395,-0.2367779016,-0.3071362376,-0.1186672673,0.097911194,-0.293420583,-0.0951493606,-0.0466190241,-0.1209683344,-0.1328951567,0.1152042896,0.0470283739,-0.4755062461,0.2196235359,-0.0837467909,0.5124325752,0.3262296319,0.2987667918,-0.3122524619,0.2711575329,-0.0631727949,0.1732649654,0.5793398023,-0.4712457955,-0.2747651339,0.0279496312,-0.0340622887,-0.1028424278,0.0546566024,0.1710966527,0.0808372647,-0.0993412733,0.074894309,0.4404076636,0.0722699836,0.0462460928,-0.2394128293,-0.1674557775,0.2583916485,0.0339711159,-0.1678980887,0.0763512626,0.2412539423,0.089694351,0.0238092691,-0.3316558003,0.1337260306,0.2027384192,0.3554896116,-0.1569679976,-0.14370884,0.2355046421,-0.4078379571,0.1885665506,0.1236447915,0.3522777855,-0.1016198173,-0.0101924054,-0.4185565114,-0.0969462097,-0.0985112861,-0.0237516873,0.0604023002,0.1085125729,0.0505038947,-0.1806459129,-0.2265082896,0.1404366791,-0.3063790798,0.0572904982,-0.2142498791,0.0430720113,-0.2826433182,-0.1120526642,0.2531868219,0.1754763722,0.1947953105,-0.0150505332,-0.1971835494,0.0930893794,-0.4591653347,0.3227218091,-0.0023458232,0.1325975955,0.0305035859,-0.0227372944,-0.2367982566,-0.0917936787,0.1783348322,-0.1693652123,-0.3360210657,0.1382521838,-0.0299928598,0.0167360548,0.1587326378,0.0029888896,0.2892495096,-0.1894600242,0.1112693548,0.1314771324,0.401583761,0.097141467,0.3815457821,-0.0451287515,-0.1332823187,-0.0848595053,0.1830834597,0.0231890697,0.1644845456,0.0743936598,0.0209634881,-0.0921830609,0.3019861877,0.5998592377,0.3605666161,0.0762010813,-0.0564554706,-0.0607898794,-0.1149053946,-0.2258322686,-0.0362129882,-0.0858919993,0.0646131411,0.2199161798,0.439574331,0.2659415901,-0.1319428086,-0.0113981068,0.1058641523,0.4606073499,-0.1928886026,0.2437663823,-0.1272559911,-0.4152647257,0.2468311936,-0.2743420303,-0.0827637166,-0.1754637212,-0.1994408518,0.1508498639,0.0996925235,-0.1142217442,0.1735125631,0.3297783434,-0.1030074134,-0.2672968209,-0.2869341969,-0.0814498514,-0.2095011026,0.0438969024,0.3391340077,-0.1061070561,0.4255249202,0.3308122754,0.057941433,-0.0315249488,-0.3174512982,-0.0129783414,-0.0383924693,0.0903385654,0.0889073163,0.3747071624,0.0114755733,-0.5478414893,0.1954109967,0.0543338619,-0.2502206862,0.2478963584,-0.031556882,-0.3154171407,-0.4758595526,0.0024259677,-0.2676188648,-0.3631879389,0.2492437363,0.330157876,0.2313883603,0.5911985636,0.0197045896,0.1316882968,0.319717288,0.1506914645,0.0455147736,-0.1869827509,0.496663183,-0.2360724211,-0.4265482128,0.1076653004,-0.1438559294,0.5702787042,-0.0444080271,-0.3675763905,-0.1872586608,0.0848747119,0.1766058952,-0.0725205168,-0.0716762617,0.29713732,-0.192166999,0.0524090715,-0.0585342273,-0.2975660861,0.1311017275,0.6311603785,0.2633401752,-0.0588213019,0.4399241507,0.0179807954,0.6985471249,0.1050329879,-0.3126606047,0.4494108856,-0.038285967,0.1811314523,-0.1402674615,-0.3474922478,0.0887276456,0.2385052443,-0.0427024513,0.1445246786,-0.0436844006,-0.0786030367,-0.4418254793,0.2470037043,-0.2635871768,0.0589443296,0.0095812818,0.1064015776,0.0375937857,-0.1067169979,-0.2408442199,-0.3441419005,-0.516520083,-0.1217708141,0.263173908,-0.2403482646,0.1861877441,-0.6647042632,-0.12222974,-0.6593780518,0.3021117151,-0.153539449,0.2957101166,-0.206398651,0.1863965094,0.127656132,-0.0775054023,0.1092500016,-0.4198971093,-0.187713474,-0.0957765281,0.2693743408,-0.4779632092,-0.055133678,-0.0379424766,0.0814743415,0.4557517171,0.2085586339,0.012650786,0.0950661004,-0.1674311757,-0.2288460732,-0.1712035239,-0.2424796671,-0.047815945,-0.638192296,-0.3416767716,-0.4831197262,0.048536025,0.1684757024,0.1516406983,-0.0745965317,0.1877341419,0.0991791114,0.4985150993,0.0212043468,-0.0574468896,0.6246475577,-0.1178974286,-0.445033282,0.4890882373,-0.0353235714,0.3584004045,0.0672954246,-0.1445760876,-0.1746760905,-0.2812872529,0.4476060569,0.180800885,0.10797932,0.2188978493,-0.1151423529,0.048115056,-0.1877975762,0.4967008233,0.3341125548,-0.0950548872,-0.3954765201,-0.1563894153,0.4564099014,-0.1978865266,-0.0036152159,0.2535579503,-0.2283557057,-0.2812907696,0.2037982047,0.1643551737,1.1384706497,-0.0975452214,0.3732190132,0.1278036088,-0.0374404974,0.7614159584,-0.1807738543,0.0525587089,-0.2565995753,0.1456496418,-0.1063772142,-0.0524123684,-0.083264567,0.0872610137,-0.2380087972,0.0753851905,-0.0617038757,-0.1353008002,-0.0051054847,0.5711603761,0.100400418,-0.3970249593,-0.2120527625,0.0920322239,-0.3342669606,0.3082541823,-0.1684511006,-0.2339408845,-0.3076736629,-0.1439931691,-0.1809539795,0.2873732448,-0.2535720468,0.0750209093,0.0626567528,-0.3308961689,0.2630811334,0.4393514097,0.5037197471,-0.0771643072,-0.0923200473,0.0256033055,-0.1352894604,-0.0499084182,-0.0268991403,0.074930951,0.191163823,-0.3115491569,0.0795424655,0.2915777564,-0.0250289906,-0.3022929728,-0.1862049848,0.1503424048,0.0951563194,-0.1453766078,-0.3303045034,0.0915492624,-0.183080554,-0.0318600312,0.0402968638,0.0597610585,0.0436034128,-0.0649102926,0.0955716744,-0.1812672466,0.0707396716,0.093663238,-0.0932908431,-0.0547153205,0.7555968761,-0.0715325549,0.0950313359,-0.0853843689,-0.2100308388,-0.3262539804,-0.1590141803,-0.099872224,0.27881217,-0.0517821051,0.0801995993,0.6546943188,0.3361329734,0.1276161224,0.1409848034,-0.1391991526,-0.3442875147,-0.3336320817,-0.0259987656,0.1786498576,-0.1002606973,0.2080222517,0.1933750063,0.3369976282,-0.2439341396,-0.1885765791,-0.1968167871,0.1518177241,0.7084043026,-0.0343682058,-0.0673392937,-0.0254820026,-0.002678158,-0.025168594,-0.1707731336,-0.1274736524,-0.3797087669,0.1688812524,0.4832161963,-0.1887609959,-0.066865012,-0.2162546813,-0.1076687425,-0.1371475458,-0.3264832497,0.0357094444,-0.045685567,0.2057019323,-0.2262983918,0.3478626609,-0.0900916383,0.2073935866,0.0170755312,0.1803188473,0.0580898374,-0.0408446863,0.1712089628,0.0622973517,-0.3547430933,-0.0378981382,-0.1847419441,-0.1004066914,0.0257856902,-0.4382697344,0.1099519432,-0.1590729356,0.0434115119,0.0324844643,-0.50390172,0.1512496173,0.3939334154,0.1096911654,0.0425386131,0.1271428317,0.1619914621,0.2084404528,-0.0173917767,0.2256092429,0.0390386134,-0.2658639252,-0.0520798042,-0.1688262373,-0.1227191985,-0.2723004818,0.0729629323,-0.1405907124,-0.4185331762,-0.0591943935,0.1482284218,0.1680250019,-0.1055326462,0.1531299055,-0.2056105733,0.089794673,0.0189697761,-0.0849580094,0.2312179804,-0.2662844956,0.0731587335,-0.0885978341,-0.032750681,0.3354672492,0.1424110532,0.6379612684,-0.3282474875,-0.0250228085,-0.2442358583,0.1070632562,0.0051226136,0.1822274774,-0.6148509979,-0.0193567257,0.0426514335,-0.1046621799,0.0303557571,-0.3352691233,-0.0621046424,0.3056537509,-0.1784630865,0.0963635147,-0.142400682,0.2184893936,0.1633142084,-0.1621600389,0.4087911248,0.0342442803,-0.1055067927,-0.087574549,0.0800634176,0.4988902211,0.5946880579,0.1123402938,0.3067478538,0.2138982415,0.0375816785,0.0789859742,0.2048170865,0.4359585941,0.2809099555,0.114513658,0.0407944806,-0.1263172925,0.075223729,0.0311446674,0.1062242985,-0.2927734554,0.3019167483,-0.0107805952,-0.365827769,-0.174708724,0.3855642378,-0.4677880704,0.0231263172,0.2493375689,0.219366923,0.3385199308,-0.2580967844,0.0306783356,-0.1771263331,0.432107985,0.4894350767,-0.0882350057,-0.1246488318,-0.2862339318,-0.4793259203,0.1886095405,-0.1928080916,-0.210184142,-0.3527612388,0.1888302416,0.0455835722,-0.102539815,0.0731486082,0.0991485715,-0.2226377577,-0.0292674378,-0.4544503689,0.0493438393,-0.1757567376,-0.0339437835,-0.0286232922,-0.1621441543,0.0706113279,0.0598372966,0.0344206616,0.0202425625,0.2201696187,0.1197861433,0.3783572316,0.4785477221,-0.0986453146,0.7240328193,0.0831504315,-0.0094493954,0.0268242545,-0.0551792681,-0.1759870648,0.0395119414,0.1294778138,0.4713498652,-0.0773479789,0.3596563339,-0.2402812392,0.0741641894,-0.0500345677,-0.5376127958,0.0891820565,-0.0042181509,-0.0442380048,0.266225189,0.4257824123,0.1162227318,0.1147406176,-0.1251274496,-0.4667599201,-0.2830771804,0.418811053,-0.5340777636,-0.3015986681,0.1996172369,0.226903379,-0.0685337111,0.0504344665,-0.3018623888,-0.0672926307,0.1631371677,-0.2279830426,-0.0155877136,0.3134565949,-0.4714373946,0.0794140995,-0.0380180851,0.2798065841,0.2096854597,-0.3522427678,-0.0415447615,-0.1549486816]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/181","title":"Cannot upload my own dataset","comments":"@lhoestq The endpoints in https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/hf_api.py should be (depending on the type of file):\r\n```\r\nPOST \/api\/datasets\/presign\r\nGET \/api\/datasets\/listObjs\r\nDELETE \/api\/datasets\/deleteObj\r\nPOST \/api\/metrics\/presign \r\nGET \/api\/metrics\/listObjs\r\nDELETE \/api\/metrics\/deleteObj\r\n```\r\n\r\nIn addition to this, @thomwolf cleaned up the objects with dataclasses but you should revert this and re-align to the hf_api that's in this branch of transformers: https:\/\/github.com\/huggingface\/transformers\/pull\/4632 (so that potential new JSON attributes in the API output don't break existing versions of any library)","body":"I look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```","comment_length":72,"text":"Cannot upload my own dataset\nI look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```\n@lhoestq The endpoints in https:\/\/github.com\/huggingface\/nlp\/blob\/master\/src\/nlp\/hf_api.py should be (depending on the type of file):\r\n```\r\nPOST \/api\/datasets\/presign\r\nGET \/api\/datasets\/listObjs\r\nDELETE \/api\/datasets\/deleteObj\r\nPOST \/api\/metrics\/presign \r\nGET \/api\/metrics\/listObjs\r\nDELETE \/api\/metrics\/deleteObj\r\n```\r\n\r\nIn addition to this, @thomwolf cleaned up the objects with dataclasses but you should revert this and re-align to the hf_api that's in this branch of transformers: https:\/\/github.com\/huggingface\/transformers\/pull\/4632 (so that potential new JSON attributes in the API output don't break existing versions of any library)","embeddings":[0.0944484174,-0.030214861,0.0476927236,-0.0215730593,0.1586551219,-0.1987861693,0.309314847,-0.0286144279,-0.2004603595,0.1795501411,-0.1415876746,0.2648700476,-0.2190395445,0.2876112759,0.3653645813,0.0843751803,-0.0124915065,0.2811934948,-0.0141321747,-0.0984324738,-0.211942777,0.1044818237,0.0320216455,-0.049357675,-0.1195409074,-0.1753448844,-0.0963914841,0.2744733393,-0.2059195936,-0.3228287399,0.4480949938,0.2225773484,0.3027473688,0.5770031214,-0.0001172611,-0.0428950414,0.1430104822,-0.2234495729,-0.6293331981,-0.3826371729,0.0799077004,-0.1367630661,0.0705319941,-0.2529910207,0.1908079833,0.0417050384,0.000706739,0.0639598146,0.5112629533,0.3838021159,0.1552750617,0.0833508521,-0.1944427937,-0.0838215053,0.0314243957,0.1894314736,-0.0444487631,0.2973888218,0.364697814,-0.3999806941,0.2879976034,-0.1958285272,-0.0049449289,-0.0952834189,0.4774852991,-0.0244300906,0.0022098885,-0.4087409079,-0.004733745,0.1479023844,0.2761023939,-0.044268284,-0.3474467397,-0.1056420058,0.1462824494,-0.4893601835,0.2515532076,0.3014510572,-0.4247399569,-0.0290612895,-0.2817880511,0.3245869875,-0.4137291014,0.4369983971,0.1153259203,0.354593873,-0.0213738903,0.074619472,0.230759427,-0.1369190216,-0.0222479887,0.0737877786,0.0122102657,0.2807020247,-0.1889118105,-0.1876102984,-0.0186335649,-0.2808482349,-0.0978708342,0.1118517295,0.4561248124,-0.1378393024,-0.2398003489,0.1559489369,0.0908848122,0.3234421909,-0.0065494422,0.0977970436,0.1205292344,0.0003337557,0.2836908996,-0.0605229773,-0.2795391977,-0.0779379979,-0.1403634995,-0.0355208144,0.2655667067,0.0827493966,-0.0020347519,-0.088930212,0.065670073,0.0615277812,-0.083274506,0.3429426253,-0.080831036,-0.2568761408,-0.0040633054,0.252545774,0.147270903,-0.2681153715,0.0705913752,0.3221955895,-0.1211353242,-0.149605453,0.4027736783,0.0167119913,0.2068513781,-0.0547541231,0.1655029655,-0.1426860988,0.0625669211,0.0156807788,-0.1528277248,0.0331363566,0.237373203,0.3591507375,-0.0791967884,-0.5343106985,-0.125402838,0.049368076,-0.3811899722,-0.3526668251,-0.4539186656,0.0959174708,0.0441683568,-0.2248820662,-0.3203067183,-0.101484552,0.0998919755,-0.2611241937,-0.071775645,-0.0267491229,-0.1189917624,-0.1263092607,0.1208137274,0.033902891,-0.4636068344,0.2092765123,-0.0461566262,0.523032248,0.2966082692,0.2885639369,-0.3173555136,0.2671290934,-0.0691568628,0.1890891343,0.5893782973,-0.4831946194,-0.2531136274,0.0503122136,-0.0370711274,-0.1145570129,0.0292967111,0.1607625484,0.0855262578,-0.0990604013,0.0783930346,0.4290344119,0.0783739761,0.045694191,-0.2462299019,-0.1808586866,0.2245772332,0.0309159346,-0.1713451445,0.0790283978,0.2619319856,0.0561614595,0.0280440822,-0.3400062919,0.1454513073,0.1951549798,0.3312384486,-0.1613267809,-0.1434575915,0.2471031547,-0.4086436927,0.1843587905,0.1157301217,0.3616276681,-0.1448758245,0.0029632335,-0.4288184047,-0.092421107,-0.0853365138,-0.0376782976,0.0695079416,0.0906381607,0.0685435906,-0.1673062444,-0.2222517282,0.1170056313,-0.3020484149,0.0560998693,-0.1878899485,0.061975915,-0.28403005,-0.1122310981,0.2280951589,0.1870625764,0.2036272883,-0.0150089432,-0.1789750606,0.090549551,-0.4720056057,0.3271060586,0.0459110849,0.1694826037,0.0231936872,-0.0260973182,-0.2118454725,-0.1186255291,0.1891427487,-0.1595870256,-0.3298256695,0.1566566974,-0.022237977,0.0081614153,0.1694997698,0.0123370783,0.2830250859,-0.1977061927,0.1146022975,0.1360907406,0.4101681411,0.0970576778,0.3762931824,-0.0375031307,-0.1397168487,-0.0750534981,0.1551811099,0.0004623414,0.171068266,0.072201483,0.0252788514,-0.094703041,0.2884230018,0.5951282382,0.3534690738,0.0794587955,-0.0524825007,-0.044750642,-0.1115141809,-0.2275600135,-0.0162841342,-0.0774828494,0.0485457778,0.2117167711,0.4386914372,0.2605527639,-0.1326118112,-0.0168820564,0.1147994399,0.454028517,-0.1906829625,0.2663320601,-0.0992186666,-0.4326516092,0.2592717707,-0.2881084979,-0.0862180293,-0.1827846467,-0.1833827049,0.1253715754,0.0859248787,-0.1114077941,0.152381748,0.3510534465,-0.0952787101,-0.2716924846,-0.2880837321,-0.1030787975,-0.2051630914,0.048939541,0.330488205,-0.1106445193,0.4104598165,0.3433226645,0.0662654787,-0.0462405868,-0.3267802,-0.0032144643,-0.0442584604,0.0937911645,0.0913999304,0.3395482898,0.0049578729,-0.5441505313,0.1870235205,0.0585749187,-0.2473661751,0.2597736716,-0.0378395431,-0.3145050108,-0.4980634153,0.0100823594,-0.2758832276,-0.3460385501,0.2394315749,0.3302131295,0.236339286,0.5939539671,0.0181891359,0.131915465,0.3163865507,0.1342408955,0.0335440598,-0.2057852,0.4849305451,-0.236068815,-0.4347625673,0.1069922596,-0.1374837905,0.5780151486,-0.0456713438,-0.3929759264,-0.1988608539,0.0978337601,0.206396237,-0.0661416799,-0.0829823688,0.2729762197,-0.1973797083,0.0393159352,-0.062912859,-0.298491627,0.1541345567,0.6249304414,0.2360728532,-0.0714840367,0.4156574607,0.0507278517,0.699492991,0.0851802602,-0.3150883317,0.4526679218,-0.0317254253,0.1771961451,-0.1375230402,-0.3402383029,0.0954925716,0.2374717593,-0.0405364446,0.1565459967,-0.0283113699,-0.083195053,-0.4500622153,0.2251806557,-0.259337455,0.0497520231,0.0146007203,0.0814522952,0.0388735048,-0.0907918289,-0.2421220094,-0.3565268815,-0.5098557472,-0.1295820475,0.2505662739,-0.2341902852,0.1831369996,-0.6524074674,-0.103706941,-0.6693302989,0.2962004542,-0.1468888521,0.2754461169,-0.216462329,0.186621964,0.133478716,-0.0864000395,0.0892623737,-0.4160827398,-0.1734875143,-0.0992085561,0.2953804135,-0.4609894753,-0.0464688167,-0.0333322287,0.0832540914,0.4845872223,0.2163555622,0.0204087533,0.0979561955,-0.1728145927,-0.2555900514,-0.1645137966,-0.2275293916,-0.0332689621,-0.6557898521,-0.3437057734,-0.4793667197,0.0711590797,0.1532265693,0.151920408,-0.0705181509,0.2251535356,0.1137521788,0.4937580228,0.0177867748,-0.0832151994,0.6168334484,-0.1239348873,-0.4665552974,0.4901572168,-0.0620346144,0.3874391913,0.0747206137,-0.1377121359,-0.1564882696,-0.3009517789,0.4412570298,0.1783333123,0.1096612513,0.2135729939,-0.1321162283,0.0521494895,-0.1916648149,0.4733583033,0.338417381,-0.0971643776,-0.3771511316,-0.1397310197,0.4467119575,-0.215798527,-0.0079214675,0.2395016551,-0.2120080143,-0.2822959721,0.2108962685,0.1703666151,1.1364248991,-0.077402845,0.3717598915,0.132048741,-0.0545556135,0.7561656237,-0.189636603,0.0247636996,-0.2485022247,0.1745954603,-0.10031946,-0.0448153913,-0.0889813229,0.1152660772,-0.2281902432,0.057403028,-0.0467419252,-0.163637951,0.0073889745,0.5417450666,0.0994138047,-0.3780417442,-0.2032103688,0.1022660434,-0.3348423839,0.300632298,-0.1681476533,-0.234698981,-0.3243205249,-0.1621618122,-0.1864953637,0.3050611913,-0.237493977,0.0810238048,0.0531149097,-0.3184699118,0.2553343773,0.4358448982,0.5033400655,-0.1063321084,-0.0790950954,0.0308602154,-0.137772128,-0.029955158,0.0002161172,0.0722954124,0.1759189963,-0.3124389648,0.0652772188,0.2966577411,-0.0143026495,-0.3096224368,-0.1740131676,0.1471674442,0.0762593523,-0.14346008,-0.3124761581,0.0992090851,-0.1777139753,-0.0247426759,0.0412108861,0.0661772639,0.0359409004,-0.0725611746,0.0739263743,-0.1873590499,0.0651199147,0.0738555714,-0.1152715087,-0.0616941042,0.7477407455,-0.0671861991,0.0966047719,-0.0992188156,-0.2355632484,-0.3396821618,-0.1887059957,-0.1045092046,0.2684942186,-0.0451550037,0.0695585832,0.6650152206,0.3328910768,0.1416175365,0.1635205895,-0.1516109109,-0.3325670958,-0.340667665,-0.0334491134,0.1692481786,-0.0827003419,0.1978584081,0.2017530501,0.3437685668,-0.2551349401,-0.1751260608,-0.2131165564,0.1651638597,0.7187824845,-0.0541661382,-0.0610992089,-0.0191584751,0.0032399113,-0.0441674478,-0.1669820249,-0.1334289461,-0.3722266555,0.1622686237,0.4754085839,-0.2071845233,-0.0741108805,-0.2015526891,-0.1035787836,-0.1250377446,-0.3151989877,0.0409207828,-0.0583339781,0.2108713239,-0.2468243092,0.3526260555,-0.088084802,0.1950054765,0.0192929022,0.1832036376,0.0725326985,-0.0244797692,0.164117977,0.0505684167,-0.3545883894,-0.0314844102,-0.2035165578,-0.0913090855,0.0471166186,-0.4444263577,0.11381758,-0.1683659256,0.0478400998,0.0216457322,-0.5189088583,0.1451275051,0.3848022521,0.120114468,0.0483219996,0.1230035946,0.1684435457,0.2147819698,-0.0101585686,0.2320228219,0.0621912405,-0.2759912014,-0.0597312748,-0.1776445806,-0.1046361476,-0.2630263865,0.0702874213,-0.1269117296,-0.4082081318,-0.051094301,0.1315317899,0.1713171303,-0.1107316837,0.1462880671,-0.2145098299,0.0946293026,0.0280823838,-0.0888011679,0.2282719463,-0.2685782015,0.0781367794,-0.1073058844,-0.0419518165,0.3354445696,0.1483921111,0.6253684759,-0.3005793393,-0.0276781321,-0.2420210987,0.1444241852,-0.0007229528,0.1765722483,-0.6177632213,-0.0288763829,0.0507036746,-0.1046496034,0.0283250902,-0.3496130407,-0.0608217493,0.3098294139,-0.1774179339,0.1142836064,-0.1528502107,0.2142976224,0.148403734,-0.1513316631,0.4299252331,0.0596262366,-0.1039166003,-0.0821743533,0.0770898834,0.5147434473,0.5829105377,0.1159393415,0.3167451024,0.2150345147,0.0282079987,0.0716503635,0.2167905718,0.4343607128,0.2830868363,0.126561448,0.0477544814,-0.1217485219,0.0731860697,0.0519172363,0.1230235919,-0.2741186023,0.3007909954,0.001268163,-0.3716159761,-0.1921243072,0.3825390935,-0.4631416202,0.0365770943,0.2187487036,0.2108299136,0.3433951437,-0.2772810161,0.0388861001,-0.1634581834,0.4362384677,0.4967529774,-0.0677361339,-0.1250281036,-0.2891867161,-0.4506284297,0.1813104004,-0.198644951,-0.2299533933,-0.3568552732,0.1961074322,0.0484901294,-0.1008574814,0.0538350455,0.0999100357,-0.2170632184,-0.0169730466,-0.4612067938,0.0610630773,-0.158781141,-0.0148151172,-0.0287290849,-0.1493174881,0.0639240518,0.0733674765,0.0429255329,0.0296155084,0.2109252214,0.1070856377,0.3892157674,0.4718276858,-0.101585038,0.7334179878,0.0889010727,-0.0027499315,0.0124151427,-0.0510118417,-0.1712319553,0.0297250114,0.1415660977,0.4371804893,-0.0835016146,0.3541291654,-0.2191764861,0.0532273352,-0.0519568063,-0.5288754106,0.1035564765,-0.0058149975,-0.0341604315,0.2721144557,0.4090954959,0.1110354438,0.1126571596,-0.1088172644,-0.4588081539,-0.2800885141,0.3989613056,-0.5402231216,-0.3015978634,0.1958185136,0.2340399623,-0.0600510873,0.0478358455,-0.3058386743,-0.0596739389,0.1677778363,-0.2149410397,-0.0090568988,0.3197653294,-0.4719270468,0.0773251504,-0.032319963,0.2738975585,0.2047537565,-0.3582354486,-0.0377810188,-0.1672146171]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/181","title":"Cannot upload my own dataset","comments":"New commands are\r\n```\r\nnlp-cli upload_dataset <path\/to\/dataset>\r\nnlp-cli upload_metric <path\/to\/metric>\r\nnlp-cli s3_datasets {rm, ls}\r\nnlp-cli s3_metrics {rm, ls}\r\n```\r\nClosing this issue.","body":"I look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```","comment_length":22,"text":"Cannot upload my own dataset\nI look into `nlp-cli` and `user.py` to learn how to upload my own data.\r\n\r\nIt is supposed to work like this\r\n- Register to get username, password at huggingface.co\r\n- `nlp-cli login` and type username, passworld\r\n- I have a single file to upload at `.\/ttc\/ttc_freq_extra.csv`\r\n- `nlp-cli upload ttc\/ttc_freq_extra.csv`\r\n\r\nBut I got this error.\r\n\r\n```\r\n2020-05-21 16:33:52.722464: I tensorflow\/stream_executor\/platform\/default\/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nAbout to upload file \/content\/ttc\/ttc_freq_extra.csv to S3 under filename ttc\/ttc_freq_extra.csv and namespace korakot\r\nProceed? [Y\/n] y\r\nUploading... This might take a while if files are large\r\nTraceback (most recent call last):\r\n File \"\/usr\/local\/bin\/nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/commands\/user.py\", line 234, in run\r\n token=token, filename=filename, filepath=filepath, organization=self.args.organization\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 141, in presign_and_upload\r\n urls = self.presign(token, filename=filename, organization=organization)\r\n File \"\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/hf_api.py\", line 132, in presign\r\n return PresignedUrl(**d)\r\nTypeError: __init__() got an unexpected keyword argument 'cdn'\r\n```\nNew commands are\r\n```\r\nnlp-cli upload_dataset <path\/to\/dataset>\r\nnlp-cli upload_metric <path\/to\/metric>\r\nnlp-cli s3_datasets {rm, ls}\r\nnlp-cli s3_metrics {rm, ls}\r\n```\r\nClosing this issue.","embeddings":[0.1084671691,-0.044078771,0.0504012294,-0.0276059136,0.1381330788,-0.1811706275,0.3266917765,-0.0091856392,-0.1950231344,0.180015564,-0.1299543679,0.2376451492,-0.230676949,0.291965574,0.3680000901,0.0748568103,0.0062184273,0.2890027463,0.0118492749,-0.1012402996,-0.2111017108,0.1081708297,0.035998825,-0.0665371567,-0.1229436174,-0.1740513891,-0.1002068445,0.2709647417,-0.2011140138,-0.3351971507,0.4402616322,0.2242947519,0.3215373158,0.5771903992,-0.0001176812,-0.0532847866,0.13914451,-0.2255368829,-0.6261975169,-0.3673120439,0.0849663466,-0.1735269576,0.0826737806,-0.2572462857,0.1832292378,0.0250916239,0.0085984897,0.0770226046,0.5234438777,0.3687664866,0.1549511999,0.0958202556,-0.1830172092,-0.0987186655,0.0291224457,0.1872793585,-0.0424793661,0.2965939641,0.368725121,-0.3961860538,0.3029372394,-0.1861630976,-0.0089923665,-0.0887328684,0.4844445884,-0.0220308937,0.0068499739,-0.4027802944,-0.0080832988,0.1580432355,0.272163868,-0.0382779129,-0.3466467857,-0.1168711558,0.1563580781,-0.4871679246,0.2421206385,0.2832774222,-0.4218559861,-0.0465164334,-0.2582421303,0.3114126921,-0.4107922316,0.4234362543,0.1253170967,0.3503036499,-0.0261408836,0.0610153116,0.259301126,-0.1453029364,-0.0212428421,0.0925706029,0.0176385213,0.2819576263,-0.1919323057,-0.1905701011,-0.0215728953,-0.2899370193,-0.1039742082,0.1091003865,0.4545297623,-0.1308467239,-0.2231484801,0.1516526043,0.0871840715,0.3312409818,0.0086245565,0.0796040446,0.1109220982,0.0127320886,0.2880072892,-0.06930051,-0.2757668793,-0.0784996375,-0.1371893734,-0.0488898866,0.2654902041,0.0865564197,0.0027184128,-0.0861177742,0.0716395155,0.0399115048,-0.0725512505,0.3346659243,-0.0997375473,-0.2767594457,0.0082464684,0.2441070974,0.1290178746,-0.2690321803,0.0709089339,0.3014498949,-0.1232390478,-0.1411055028,0.4131863117,0.0026039535,0.1909866333,-0.0497684591,0.1741697937,-0.1157201827,0.0725953132,0.0268774275,-0.1545820236,0.0261169504,0.2355850488,0.375002414,-0.0722648129,-0.5426658988,-0.1255455166,0.0468735807,-0.3762072921,-0.3481229544,-0.4494222701,0.0936678872,0.0353323221,-0.2343602479,-0.290365845,-0.0844530612,0.0992024839,-0.2737713158,-0.0880437568,-0.0178201012,-0.1080890968,-0.1402437836,0.114935793,0.0539799929,-0.4861147404,0.2089095861,-0.0626698956,0.5137918591,0.3013716042,0.2810108066,-0.3043970466,0.2599909008,-0.0623657331,0.1773784161,0.5754054189,-0.474339664,-0.2507374883,0.0342788249,-0.0365614444,-0.1030127406,0.0546859354,0.1682045311,0.0883628651,-0.1026604846,0.074139908,0.4095371664,0.0759576038,0.0300125126,-0.2245140672,-0.1747812778,0.2542684674,0.0123875318,-0.1504040956,0.0748391226,0.2523053885,0.0847450495,0.0307703875,-0.3352099657,0.1500530988,0.2014988661,0.3359648585,-0.1352574378,-0.1372293383,0.2380945235,-0.4048440456,0.1965625882,0.1271633655,0.3596282303,-0.1108400673,0.0062363427,-0.4200919569,-0.1138808578,-0.0957373306,-0.0317725502,0.0666666776,0.0963794664,0.0702915415,-0.1855908334,-0.2302042246,0.1288204789,-0.3153890073,0.0610579215,-0.1971267909,0.0591705479,-0.2863599956,-0.1116008535,0.2181829214,0.1753217429,0.2073204368,-0.0180181973,-0.1725531071,0.0894984901,-0.4653667212,0.3087368309,0.0124266222,0.14155294,0.0335689224,-0.0215841755,-0.1996200234,-0.117882289,0.1772723198,-0.1705975533,-0.3431076109,0.1499734372,-0.0331483968,0.0194792952,0.1668228805,-0.0024819076,0.2705669999,-0.1924484968,0.1081033424,0.1190124899,0.3931277692,0.1005677804,0.3735919595,-0.0623499416,-0.126172334,-0.080317907,0.1737629175,0.0145931114,0.1585074812,0.0684892833,0.0332597308,-0.0781026036,0.2860080004,0.6209813356,0.3515605628,0.0983222872,-0.0484041534,-0.0506114289,-0.1180213094,-0.2226050347,-0.032334052,-0.0865320414,0.0450485833,0.2136892676,0.4292725623,0.2628835738,-0.1386042386,0.002830114,0.1121918634,0.441070199,-0.1806763709,0.2472006083,-0.1107613221,-0.4309422076,0.2381785214,-0.2824086249,-0.0893063396,-0.1644198149,-0.1953134388,0.1184250563,0.1135315597,-0.1269594133,0.1662990153,0.3391794562,-0.1104057431,-0.2544942796,-0.2794032395,-0.0882852823,-0.2085284889,0.0529598072,0.328294456,-0.11787384,0.4269639254,0.3558065593,0.0669828057,-0.0482534915,-0.3466056585,-0.0127200242,-0.0637017488,0.0786134899,0.0849280879,0.3438898325,-0.0008332819,-0.5517418385,0.1905883104,0.0630328357,-0.252502948,0.2460184395,-0.0219418537,-0.3063010275,-0.4941290319,-0.0074582994,-0.2783525586,-0.3438791931,0.2504079938,0.3512364328,0.2351835072,0.6135793328,0.0208297428,0.1394834816,0.3170115054,0.1529634148,0.0331702419,-0.1874231398,0.4657980502,-0.2530597448,-0.4329993427,0.1104678884,-0.1329057962,0.5748018622,-0.0416008197,-0.3659122884,-0.1997584552,0.0914410055,0.1881645173,-0.0772584751,-0.0661343262,0.2856583595,-0.1821654141,0.0421079136,-0.0591190793,-0.3117053509,0.1501697004,0.6340502501,0.2549774647,-0.0765854865,0.421613127,0.0424232557,0.7029533982,0.0894530341,-0.3097903132,0.4549658895,-0.0282193925,0.1947965622,-0.1329689324,-0.3579265773,0.0959679559,0.2346650064,-0.0307864323,0.137752369,-0.0272191986,-0.0677645653,-0.4466430247,0.2475075126,-0.2521286905,0.0607918128,-0.0016714791,0.10914772,0.0424583182,-0.1023130342,-0.2365525067,-0.3656544387,-0.5041989684,-0.129275158,0.2498649955,-0.2540321648,0.1969147325,-0.6588913202,-0.1237396374,-0.657538712,0.2880611122,-0.1481109113,0.259454906,-0.2137505859,0.1771367192,0.1310416162,-0.0893459842,0.117172353,-0.4182179272,-0.1867839694,-0.0956521258,0.2866274714,-0.4773975909,-0.0534710735,-0.0437615253,0.0851091668,0.4654028714,0.2201539278,0.0312621593,0.0952756032,-0.1582916826,-0.2440674454,-0.1730844378,-0.2345627546,-0.0407793522,-0.6392328739,-0.3375430405,-0.4886773527,0.0494773686,0.1570402235,0.1608155668,-0.0753104985,0.2124441862,0.1187975928,0.4977765977,0.0224337,-0.0697201788,0.6355295777,-0.1225436628,-0.4498655498,0.4896850884,-0.053558588,0.3719132245,0.0506037511,-0.1260614544,-0.1803298593,-0.2989445925,0.4641976357,0.1874573529,0.1153126881,0.2137034088,-0.1160859466,0.0494511984,-0.1887466908,0.4735016525,0.3472860456,-0.0927902907,-0.3748316765,-0.1661893725,0.4463634193,-0.188774094,-0.0016303288,0.2298584282,-0.2005378306,-0.2884903252,0.1962748468,0.1790330112,1.1350758076,-0.0961631313,0.3786780238,0.1360208243,-0.0524253398,0.7858073115,-0.1857078671,0.0341377594,-0.2420800179,0.1655093282,-0.1026663184,-0.0376381315,-0.0871461928,0.1050227806,-0.2391220629,0.0672172233,-0.0572762564,-0.1655515134,-0.0060338778,0.5472076535,0.0957107991,-0.3948434293,-0.199129194,0.1065407768,-0.3313684464,0.2973275483,-0.1884972006,-0.234790951,-0.2989434302,-0.1551221758,-0.1712496579,0.3010398149,-0.2388863713,0.0890778527,0.0547804125,-0.3258431554,0.2452927381,0.4202984273,0.505764842,-0.0947711393,-0.0716306344,0.0159043241,-0.1294851452,-0.0580585077,-0.016993206,0.0785609335,0.2045133561,-0.3050248027,0.0645935535,0.283226192,-0.0042025563,-0.2958987355,-0.1927382946,0.1700759977,0.1059851274,-0.1450627297,-0.3228810132,0.1075231433,-0.198575601,-0.0239279438,0.0431403145,0.0521441475,0.0230946746,-0.062853232,0.0758086815,-0.1846114546,0.0675204247,0.0913106576,-0.0864118338,-0.0853638053,0.7531554103,-0.0654557422,0.1021022126,-0.0943018049,-0.235720396,-0.3444405198,-0.1637298763,-0.1043168828,0.2887042463,-0.0454124101,0.1091612354,0.6310174465,0.3479663134,0.1186373159,0.1422429681,-0.130046308,-0.3389728963,-0.3504675031,-0.0227931347,0.1764518023,-0.0935845971,0.1920359284,0.206674099,0.3386301398,-0.2541728914,-0.1868463457,-0.1824915707,0.1629973054,0.7249684334,-0.0466189347,-0.0673949718,-0.0203771275,0.0019190285,-0.040756721,-0.1685425788,-0.1342892051,-0.3807841539,0.1612873077,0.4796664119,-0.2025006413,-0.0754027069,-0.2050625384,-0.0948038474,-0.1198447421,-0.3056564331,0.0439240448,-0.0480843335,0.2215616703,-0.232160002,0.3371523619,-0.0559340306,0.1899590343,0.0306708366,0.1726497412,0.0724193007,-0.0340202637,0.1680604368,0.0574886017,-0.3541554809,-0.0327512063,-0.191226989,-0.1038464829,0.0290145352,-0.4462993443,0.1061887741,-0.1760344058,0.0498005338,0.0224347301,-0.4888657331,0.1380916834,0.3895100057,0.1181805134,0.036484424,0.1172882542,0.1646392196,0.2187669873,-0.0181711446,0.2286812812,0.0425651446,-0.2698048949,-0.0645261779,-0.1723545492,-0.1230842248,-0.2726886868,0.0867757201,-0.1169707999,-0.3999556601,-0.0692690834,0.1547372639,0.1715611517,-0.0974047184,0.1438153982,-0.2051552385,0.1071855947,0.0338847078,-0.0860831141,0.2209541053,-0.2704435587,0.0801371187,-0.0875020772,-0.0271068905,0.3488442004,0.1409808248,0.6430830359,-0.3007900715,-0.038441591,-0.2372835726,0.1241429597,-0.0008916795,0.1930063963,-0.6118702888,-0.0161166321,0.0401755199,-0.1086053327,0.0398652777,-0.3541440666,-0.0675953031,0.3094727993,-0.1717279255,0.1138679758,-0.1698194295,0.2061772197,0.1469643265,-0.1484514177,0.3970651627,0.0542402454,-0.1030556858,-0.0744956359,0.0873932019,0.5002292395,0.60055691,0.116389811,0.3231106699,0.2016103417,0.0381331183,0.0725643411,0.2018144578,0.428368926,0.2745627463,0.1099810973,0.0479595251,-0.1261921376,0.0636164322,0.0479529388,0.0990201831,-0.2963976562,0.2942281961,0.0014726814,-0.37344715,-0.1934802979,0.3822582662,-0.4880563021,0.0280241873,0.2486962825,0.2155369371,0.3356565535,-0.2708441019,0.036630284,-0.1670725048,0.4374946058,0.4826229811,-0.0939953923,-0.1257395744,-0.2834657133,-0.4656836092,0.1705352217,-0.1989509165,-0.2232457548,-0.3570991457,0.2004116029,0.042870257,-0.1253056526,0.0718620196,0.0931004137,-0.2287151963,-0.0589637421,-0.4724918008,0.056482017,-0.1354804784,-0.0212811325,-0.0252808798,-0.1377789676,0.0507606156,0.0555005074,0.0349746868,0.0259417892,0.2124393731,0.1044918522,0.3972411454,0.488858819,-0.1169947833,0.7418341041,0.0934427604,-0.0103765568,0.0119380737,-0.0374658853,-0.1898798198,0.0539034717,0.1236977428,0.450558275,-0.0926324725,0.3479193449,-0.2200059444,0.0700672418,-0.0405461416,-0.5381629467,0.1111885831,-0.0031196105,-0.0312197655,0.2671132684,0.4321083426,0.124111712,0.1108286232,-0.105014421,-0.4677978754,-0.2764585018,0.406214118,-0.5333340168,-0.2964298725,0.1877313405,0.2369347662,-0.0829136074,0.0361353084,-0.2994179726,-0.0701772794,0.1632543802,-0.2294121087,-0.0128206266,0.3217462003,-0.4839883447,0.0631383583,-0.0361769013,0.2412737012,0.2241035849,-0.3566044569,-0.0330376141,-0.1516035646]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/179","title":"[Feature request] separate split name and split instructions","comments":"If your dataset is a collection of sub-datasets, you should probably consider having one config per sub-dataset. For example for Glue, we have sst2, mnli etc.\r\nIf you want to have multiple train sets (for example one per stage). The easiest solution would be to name them `nlp.Split(\"train_stage1\")`, `nlp.Split(\"train_stage2\")`, etc. or something like that.","body":"Currently, the name of an nlp.NamedSplit is parsed in arrow_reader.py and used as the instruction.\r\n\r\nThis makes it impossible to have several training sets, which can occur when:\r\n- A dataset corresponds to a collection of sub-datasets\r\n- A dataset was built in stages, adding new examples at each stage\r\n\r\nWould it be possible to have two separate fields in the Split class, a name \/instruction and a unique ID that is used as the key in the builder's split_dict ?","comment_length":54,"text":"[Feature request] separate split name and split instructions\nCurrently, the name of an nlp.NamedSplit is parsed in arrow_reader.py and used as the instruction.\r\n\r\nThis makes it impossible to have several training sets, which can occur when:\r\n- A dataset corresponds to a collection of sub-datasets\r\n- A dataset was built in stages, adding new examples at each stage\r\n\r\nWould it be possible to have two separate fields in the Split class, a name \/instruction and a unique ID that is used as the key in the builder's split_dict ?\nIf your dataset is a collection of sub-datasets, you should probably consider having one config per sub-dataset. For example for Glue, we have sst2, mnli etc.\r\nIf you want to have multiple train sets (for example one per stage). The easiest solution would be to name them `nlp.Split(\"train_stage1\")`, `nlp.Split(\"train_stage2\")`, etc. or something like that.","embeddings":[0.0433895998,-0.1981111467,0.0301857479,0.2106904835,0.1068341956,0.0847189277,0.5492762327,0.1950958967,-0.0144093074,0.1525928825,0.0023248887,0.4517517984,-0.2708511949,0.325571388,0.0190773923,-0.4061379731,-0.0645274445,0.061306797,0.2330210954,0.1022859216,0.0976755843,0.2273223996,-0.0382929817,0.2814404666,-0.2380087078,-0.2281862199,-0.2058773041,-0.1653984934,0.0188028179,-0.4043416381,0.1754710525,0.1728920341,-0.2417059243,0.0342779346,-0.0001190069,-0.1262154877,0.2135144621,-0.0090983743,-0.1567251086,-0.1734883636,-0.5583313107,-0.4940268993,0.4417530298,-0.5981582403,-0.004088195,-0.0666586384,-0.0093628475,0.2533458471,0.5818642378,-0.161096707,0.0953406096,-0.3728163242,-0.0064579169,0.121883586,0.3031226695,0.1706276685,-0.3206125498,-0.3328774869,-0.192162022,0.016418919,-0.0143435178,0.2186747342,0.0672667697,-0.1475302875,0.1250576824,0.2088906467,-0.1813259125,-0.1672047824,-0.0625163093,0.3708770573,-0.165844366,-0.1862094104,-0.2181791216,-0.7446480393,0.1928082108,-0.3991033435,0.0094767595,0.2820796669,-0.0232580472,0.0569057204,0.2280940711,-0.3026098609,-0.1733477861,0.0793732703,0.1580905467,0.6143723726,0.0829360858,0.2064342499,0.1980846226,0.0196455605,0.1997448802,0.1302580982,-0.041976776,0.1836853325,-0.2608934939,-0.3455759585,-0.3928423524,-0.0823004618,-0.2153024822,0.147732839,0.2048063278,-0.0372608788,0.0862403661,0.1348689944,0.2460693419,0.0795055479,0.4300121665,0.6358777285,0.112690866,-0.0192368124,-0.3608025312,-0.0242510345,-0.0676975027,0.0619240478,-0.1763537377,0.1009590253,-0.0996292084,0.2363564968,-0.058111269,-0.2936745286,-0.1154489666,-0.2922120988,0.2777602673,-0.0074071391,0.1323409379,0.0555965491,-0.3250802457,0.1376652718,-0.1903274506,-0.4007326663,-0.1060519591,0.123240687,-0.3039298356,0.2852865458,-0.0250473674,0.0895747021,0.1397495568,-0.023922611,-0.1711868048,-0.2618388236,0.1999876499,-0.0307732355,0.2510850132,0.0799315944,0.0168919601,0.0575356856,-0.0981862321,-0.1105860695,-0.3132466376,-0.0043426598,-0.1522113979,-0.3459362984,0.1437167972,0.059594404,0.1488416046,0.0675621256,0.2856252789,0.6215521097,0.347048521,-0.1152626202,0.1364320964,0.0587193295,-0.2213844806,-0.2414389998,-0.0208209138,-0.055000741,0.0404546596,-0.4513776004,-0.2863806486,0.1082007512,-0.1051907986,0.0911794305,-0.2293736637,0.1108224988,-0.0266857482,0.600189507,0.7154173255,-0.2713153958,0.0781716704,0.3304649293,0.0726041943,-0.3457192183,0.4756768346,0.0852043554,0.1321290582,-0.0576222539,-0.3968372643,0.1820663363,-0.380117774,-0.1711618751,0.203973785,-0.532147944,0.1683947593,-0.0909518749,-0.0555265844,-0.3709448576,0.0200215261,0.2963109016,0.4655571282,0.0344679244,0.2371775806,-0.1741747409,-0.077729702,0.2893902361,-0.1731615067,-0.3693116009,-0.2701151967,-0.0106132654,-0.1732822061,0.0567807704,0.0131774973,-0.4073964953,0.043092899,-0.2719112039,-0.1646472961,-0.2850073576,0.046643842,-0.2125976086,-0.0586675778,-0.3464665711,-0.3343035579,0.1712365896,-0.0637922063,0.2837935388,-0.3155242205,-0.1585672498,0.1928499043,0.0083349887,-0.2399851978,0.5741861463,0.2506141663,-0.0253385399,0.0648977458,0.3517275453,0.3894828558,-0.072423026,0.0233753882,-0.1261819899,0.2539397478,-0.0673269182,-0.0839314014,-0.1060166135,-0.2638687789,-0.1437775195,-0.1017713919,0.5019603372,0.1218737587,0.5909392834,0.290153265,0.0582061559,-0.0644508824,-0.2585398853,-0.1919443756,-0.4637328982,-0.2635037005,-0.3758354783,0.0825798362,-0.0438322984,-0.2890101969,0.2458765358,0.447573185,0.1451078653,0.0160814598,-0.2310429364,0.1197398156,0.1733066887,-0.2040424645,0.7559654713,0.1979498118,0.1695155501,-0.0935158879,0.0572643541,-0.0423333645,-0.1895749867,0.2760686576,0.282030195,0.0795545951,-0.0271394197,-0.1698196679,-0.1431918293,0.1030377299,-0.033338055,0.2293055356,-0.3285975754,-0.086150147,-0.0772327036,-0.1250096709,-0.3374963701,-0.6012638807,-0.0317822844,-0.2347899973,-0.241727218,0.1579475403,-0.0885935575,-0.2108806521,0.141749531,-0.1397173703,0.133240208,-0.3738991022,0.0399717614,0.1473181248,-0.3808600307,-0.0698838085,0.0259122401,0.0559975207,0.0349369906,0.2627146542,0.3256543875,-0.2808838487,0.1503280997,-0.3359712958,0.1768517792,-0.3270225525,-0.2894216478,0.2219965011,-0.3101728857,-0.0042195669,-0.468949914,0.3010498881,-0.1061424017,-0.1001413465,-0.3176913559,0.3218808174,0.2116595805,-0.1271355599,-0.319062382,-0.4289037585,-0.1418146342,0.3455915451,-0.0754193962,0.2965156734,0.0980440751,-0.4208855033,-0.0293276794,-0.0928120092,-0.1169013157,0.0434541591,0.1445828378,-0.0896989703,-0.4102134407,0.1566269845,0.0523340218,-0.1368256658,0.3339677155,-0.1761155427,-0.0114617981,0.1888323873,-0.0135303875,-0.0349345207,0.0357261188,0.1150890663,0.1559380293,0.2736525834,0.0665156841,-0.0738581121,-0.2324019819,0.434866786,0.1524572223,0.1515314728,-0.090065524,0.0241572894,-0.0799343213,0.6646896601,0.268293798,0.1996210963,-0.1777932346,-0.0047393981,-0.3417829573,0.0782982782,-0.2114556432,0.1659888923,-0.0030222298,-0.2993084788,0.3410924971,0.0300070345,-0.0614930429,-0.2507540286,-0.0688423216,-0.0362502448,-0.5644194484,0.0686152726,-0.1840482652,0.2156812251,-0.1999175549,0.0524547063,0.003842477,0.0624568537,-0.1463524252,-0.1538309157,-0.1487344205,-0.0714732707,-0.1142189056,-0.6203209758,-0.0538157262,-0.0422428325,0.2887146771,-0.1428101808,-0.021646969,-0.2228597552,0.2955176234,0.3017681837,0.2860278487,-0.0901259482,-0.1666277796,0.0933360085,0.0421495289,0.1922543794,-0.1983072758,0.0910089836,0.029229816,0.0748301074,0.0393599905,-0.5721217394,-0.2948225439,0.5887443423,-0.2423543781,-0.3340773284,0.0383950397,0.0384378098,0.1323857605,0.0734295174,-0.0404679142,0.1036643982,0.1270347089,0.1372128278,0.0833885819,-0.0328542553,-0.2343676835,0.0897765532,0.1343216747,0.0210472029,-0.0537205003,0.0877911523,0.2963995636,0.2657298446,-0.3124952614,0.4403767884,-0.4000655413,-0.0989199877,0.0095383339,-0.2170740068,0.5385269523,0.3315568566,0.1370120198,0.4471999109,0.0058187731,0.0808025002,-0.2052093893,-0.0684699491,0.2784641087,-0.1718565226,-0.4090302885,-0.7786411643,0.3944488466,0.0955311507,-0.1922816634,0.2213877589,0.1355137378,-0.4084827602,0.5755674243,0.0162111335,0.8848457932,-0.0789018646,0.2104570717,-0.0524022505,-0.2898178101,0.6745561957,-0.095918946,0.1330348998,-0.3664889336,0.294572413,0.057935711,0.1039666906,0.2314224243,0.4899133146,0.0067731775,0.2321578264,0.0844296291,0.2629140019,-0.3335983157,0.3395188153,0.1432533413,-0.1159840599,-0.0867068022,-0.048195269,0.0014830627,-0.2470320314,0.2088813037,-0.1991021186,0.0166246444,0.1703215688,-0.2553974986,-0.0069497125,-0.1202722564,0.211919412,-0.1201169491,-0.4156662524,0.3914086819,0.1776712537,0.2339615822,-0.1650224775,0.0192258451,0.4578537941,-0.0423441008,0.093565993,-0.171716392,-0.0125137512,0.3011683822,0.0534467623,0.0274007823,0.0310278311,0.018673433,-0.088381134,-0.2242789268,0.2394068837,0.4494660199,-0.5152654648,0.2077822387,0.0315666646,0.0019560773,-0.1789326668,0.0212221164,-0.2193907946,-0.2493930757,0.040107619,-0.1799755841,-0.0684349313,0.0049613854,-0.3192849755,0.1781567484,0.007905581,0.3882559836,-0.1814900637,0.160438925,-0.2399626821,0.1463244557,0.1513789594,-0.2061514407,0.2905279696,0.1530695409,-0.2651844919,0.1352849007,-0.0193828046,0.0580952913,0.6687638164,-0.180017069,0.0318539403,-0.220824793,-0.040196687,-0.1035734862,0.097322762,0.0068293237,0.4960907996,-0.1209783033,0.4065931439,-0.1664100289,-0.0965329707,0.2013273984,0.2539378405,-0.1084280238,0.017031135,0.1527231634,0.1292616427,0.0203810632,-0.0549706742,0.1774233878,-0.0882724524,-0.3262878656,0.1167680696,0.094508931,0.2027482688,-0.3695308864,-0.1411668658,0.1034345329,-0.1556040794,0.2073809057,0.0628075823,0.260486722,0.5012199879,0.2707221806,0.0106502911,-0.3788053393,0.216989398,-0.1248052493,0.3370325863,0.2525511682,0.3109061122,-0.2106348425,0.072893858,0.1331864148,-0.0557745993,0.1861089766,-0.2795474827,-0.0730740726,0.0360655226,-0.0971464515,-0.1405694485,0.2155104578,0.4737992585,0.1372170448,-0.034968555,0.2236468643,0.1104778424,-0.0722930878,0.2131297439,0.2800585926,-0.1797782928,0.0994811356,0.5531477332,0.1132161915,0.2622578144,-0.4264287353,-0.0630765706,-0.1069678143,-0.0306344386,0.5049614906,0.8284041286,0.2145945579,-0.0691812485,0.463724792,-0.182843864,0.2680802345,0.4525595903,-0.1814104319,0.3613754809,0.3433869779,-0.1725420356,0.5454282165,0.1076205522,-0.3695583642,0.0466442071,-0.0581809022,-0.2487653196,-0.1222338155,0.7737441659,0.0438777469,-0.3321120441,0.1711920649,0.3155758977,0.1680949628,-0.2501696646,-0.1361529082,0.1134822965,-0.3541958928,0.0006026044,-0.0543007143,-0.2277763486,0.3543234468,0.0822668746,0.3009105027,-0.0990269184,-0.0676979795,0.2107921392,0.4752950966,-0.3517262638,-0.1351253539,-0.0174337123,-0.2915224731,0.1551805288,-0.0719442889,0.0081623821,-0.1346800327,-0.2549780011,-0.2885667086,0.2460688055,-0.105483152,0.0232846346,-0.0697783157,0.1218494549,0.2911970019,0.2118598819,-0.0230831243,-0.0354093388,-0.2072150111,0.2606799304,-0.0189934094,-0.0599126555,0.5498864651,0.239816308,-0.1878667623,-0.0259961095,0.2288018763,-0.2921216786,-0.0178034883,0.3589164913,-0.0660984218,0.0889411047,-0.1679062545,-0.0227218699,-0.0211715996,0.485548228,-0.2224391699,-0.2036731392,-0.2754064798,-0.3702971339,-0.0224149283,-0.1403138936,-0.1252135485,-0.2741916478,-0.1313826442,0.1689521074,-0.1570608765,-0.1679461598,0.1619284749,0.0404398404,0.4329413772,0.3385254443,-0.2316653132,0.0057672281,-0.0215880945,0.1457760781,0.173848927,0.1357694268,-0.0744211674,0.3656345606,-0.2971501052,0.1566156,-0.0228856504,0.5849359632,0.1345315427,0.0016773812,-0.0207358468,0.2275291085,0.0298627596,0.0371166356,0.0406959765,0.0126986522,-0.5554838181,0.4114271998,0.0680289865,0.1626550704,-0.2748198509,0.4120602012,-0.1414839029,0.0037239632,-0.2320048958,0.1067834944,-0.0139237847,0.1969076842,-0.0230115913,0.296452105,0.3759607971,0.2861545086,-0.1174599826,-0.0658682138,-0.134728685,-0.3783155978,0.1620430648,-0.0366276652,-0.0999183878,0.0897413492,0.3389489651,0.1993412971,0.11615064,-0.7652572393,-0.2747168243,0.1887986511,-0.2403244525,-0.0242941268,0.1428132057,0.2829594612,-0.1662367135,0.0147062335,-0.1382516623,0.2797054648,-0.0631294399,0.0305362325,-0.1435409933]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/179","title":"[Feature request] separate split name and split instructions","comments":"Thanks for the tip! I ended up setting up three different versions of the dataset with their own configs.\r\n\r\nfor the named splits, I was trying with `nlp.Split(\"train-stage1\")`, which fails. Changing to `nlp.Split(\"train_stage1\")` works :) I looked for examples of what works in the code comments, it may be worth adding some examples of valid\/invalid names in there?","body":"Currently, the name of an nlp.NamedSplit is parsed in arrow_reader.py and used as the instruction.\r\n\r\nThis makes it impossible to have several training sets, which can occur when:\r\n- A dataset corresponds to a collection of sub-datasets\r\n- A dataset was built in stages, adding new examples at each stage\r\n\r\nWould it be possible to have two separate fields in the Split class, a name \/instruction and a unique ID that is used as the key in the builder's split_dict ?","comment_length":58,"text":"[Feature request] separate split name and split instructions\nCurrently, the name of an nlp.NamedSplit is parsed in arrow_reader.py and used as the instruction.\r\n\r\nThis makes it impossible to have several training sets, which can occur when:\r\n- A dataset corresponds to a collection of sub-datasets\r\n- A dataset was built in stages, adding new examples at each stage\r\n\r\nWould it be possible to have two separate fields in the Split class, a name \/instruction and a unique ID that is used as the key in the builder's split_dict ?\nThanks for the tip! I ended up setting up three different versions of the dataset with their own configs.\r\n\r\nfor the named splits, I was trying with `nlp.Split(\"train-stage1\")`, which fails. Changing to `nlp.Split(\"train_stage1\")` works :) I looked for examples of what works in the code comments, it may be worth adding some examples of valid\/invalid names in there?","embeddings":[0.1045852527,-0.0982461423,0.0275993329,0.2224785239,0.1107441559,0.1121136993,0.5969958305,0.2325306684,-0.0358218625,0.1119877174,0.0187910739,0.4395620525,-0.2157790214,0.250805527,-0.0480731539,-0.3708141148,-0.0314742029,0.0995834023,0.3009313643,0.1106136888,0.0753546134,0.2413922995,-0.0459942184,0.2722317874,-0.2111954093,-0.1831194758,-0.2387676984,-0.1546239704,0.0284153596,-0.4011641443,0.1909267604,0.1710229218,-0.2977707386,0.1150344014,-0.0001220013,-0.1493762881,0.269780159,-0.0088353129,-0.1772001237,-0.231570214,-0.5108584166,-0.4750356078,0.441298455,-0.6071299314,0.0064698858,-0.1662882417,0.0098626055,0.2345714271,0.5989475846,-0.1420638859,0.0794020444,-0.3933258355,0.0042510843,0.0590030104,0.2985037565,0.2553159893,-0.3581681252,-0.3896652758,-0.1325118542,0.0193850994,0.0135845514,0.1901755184,0.0480410121,-0.1508929282,0.0564224124,0.1882013828,-0.2405911535,-0.1051746458,-0.031475611,0.3205355704,-0.1470056772,-0.133174777,-0.2248992175,-0.7789208293,0.12296433,-0.4444979727,0.0231793597,0.308399111,-0.0855548605,0.0007133393,0.159852773,-0.2577973306,-0.1586889476,0.0799732357,0.187296465,0.5755344629,0.0426866375,0.2130333036,0.130654946,0.0514660664,0.2459288538,0.0974831581,0.0370741785,0.1481966823,-0.2402214557,-0.3462637365,-0.3629279137,-0.0612269901,-0.2478050739,0.0727624744,0.1947965175,-0.0930430442,0.0379200615,0.109423928,0.3325090706,0.0307519361,0.3663776517,0.5928646326,0.0970462933,0.0739011765,-0.3302590847,-0.0182994399,-0.0558232032,0.1093028933,-0.2353015542,0.0842660591,-0.100039497,0.1635543555,-0.0458830334,-0.259771049,-0.0991402268,-0.2428592294,0.2244431674,0.0378052592,0.1180723682,0.0390675887,-0.3129239082,0.2033401132,-0.1228090003,-0.4388536513,-0.08893843,0.0996953845,-0.3002477884,0.2296502441,0.1069265753,0.1424677372,0.1824060529,0.0084284814,-0.1476583183,-0.2453731745,0.1563497186,0.0066439323,0.2317557186,0.042394314,0.0324652158,0.0562312268,-0.0377421491,-0.0723695308,-0.3021252155,0.0177272614,-0.07855957,-0.3099419177,0.1863626838,0.0404328778,0.1440342516,0.061936494,0.3620258272,0.6019592881,0.4055770338,-0.1647848189,0.10734272,0.037772011,-0.2596309185,-0.1934563071,-0.0436960123,-0.0139565021,-0.0107302275,-0.448692143,-0.3815647066,0.1615701914,-0.1280699223,0.072770834,-0.1476602703,0.0584482104,-0.0028749532,0.6331416965,0.5671984553,-0.2285178602,0.0892338976,0.3364673257,0.1159049943,-0.3352688253,0.4838373959,0.0501568131,0.107458055,-0.1066686958,-0.4239080846,0.1804618239,-0.5068801045,-0.2254240811,0.1737490892,-0.5565368533,0.1767485738,-0.0706296936,-0.0389987677,-0.3304858506,-0.0028007983,0.2975155115,0.4849632978,0.0340141691,0.2780067325,-0.1932996809,-0.067157656,0.2642988563,-0.1216097176,-0.3572362065,-0.3117902577,-0.0428921767,-0.1212896779,0.0310803726,-0.0378559791,-0.3895498216,-0.0125864493,-0.2790810168,-0.1604308188,-0.2774960101,0.031024849,-0.2263091356,-0.0619041249,-0.3348519504,-0.3435290158,0.1144687012,-0.0875819921,0.2657094598,-0.3278540969,-0.1681105047,0.2228049189,-0.0255954638,-0.18210648,0.6047663689,0.2861999869,-0.0080928607,0.0620746203,0.3274420798,0.3911905885,-0.0306486897,0.0057343468,-0.1291932613,0.1832372993,-0.123016797,-0.0524583384,-0.0439426266,-0.2799320817,-0.2084253132,-0.2021620423,0.5459465384,0.0275820028,0.6292759776,0.2911064029,0.049969472,-0.0128314495,-0.335465461,-0.1722602248,-0.5019676685,-0.211684823,-0.4157887995,0.1152676716,-0.1337288767,-0.2776992917,0.2355115712,0.430552274,0.177992776,0.0336464979,-0.234156087,0.1341372579,0.1435284764,-0.1983996779,0.6893722415,0.1891099066,0.1522488743,-0.0407744497,0.0803330094,0.0330183059,-0.1813274026,0.285543412,0.2572939694,0.0831954777,-0.001740361,-0.1671254486,-0.1373580396,0.0530168675,0.0606803,0.2253384292,-0.3368180692,-0.0538131148,-0.074822925,-0.1295629889,-0.3124649823,-0.5548498631,-0.0334240794,-0.1863946468,-0.2262301296,0.2090401202,0.0143301617,-0.2173095047,0.1576506197,-0.1592102647,0.1664498001,-0.4105944335,0.0815151557,0.0501448661,-0.4121741652,-0.0365711115,0.0229783617,0.0629541054,0.0134071205,0.2887684703,0.3602593839,-0.2948746383,0.1842560172,-0.3726987541,0.1845199615,-0.3997818232,-0.2675514817,0.2295847237,-0.303517282,0.0146265998,-0.4285719693,0.3069125414,-0.1674591601,-0.1307964623,-0.3343504071,0.3282942176,0.1910609454,-0.1777388155,-0.3134026527,-0.4024408162,-0.1457711458,0.3905413747,-0.0697919652,0.275059253,0.0724082813,-0.4712016284,-0.0289214309,-0.0028673685,-0.1211846471,0.0688783303,0.2207591087,-0.0744477063,-0.4078603387,0.1298802346,0.0845913589,-0.1246915534,0.3680259883,-0.247337088,-0.0626762062,0.2727819085,0.0512867309,-0.061664667,0.1613779515,0.0468612462,0.1606241167,0.2921684682,0.0672668666,-0.1720883101,-0.241377458,0.3609777391,0.1215413287,0.1340132207,-0.0718494728,0.0910204872,-0.0506043695,0.6642805338,0.2965461314,0.2599332035,-0.1523855478,0.0683135614,-0.3416393399,0.0813866407,-0.2282858938,0.1414750516,-0.05142067,-0.3744937181,0.3664929271,0.0320485607,-0.1365581751,-0.2629485726,-0.0122708697,-0.08253254,-0.5328510404,0.0085465321,-0.1830297112,0.2694475353,-0.1307053268,0.0647173226,0.0322014354,-0.0341987573,-0.1484875977,-0.1926971525,-0.0961194858,-0.1442140341,-0.13510786,-0.5458646417,-0.0523284487,-0.0137120802,0.2454072833,-0.1310093999,-0.0434124656,-0.1996350735,0.2979269028,0.2764442265,0.2861575186,-0.0740296021,-0.1080965847,0.1467863023,0.0468646623,0.1941204667,-0.1502496153,0.075472191,0.025733415,0.0400082134,0.021056816,-0.5988478065,-0.2771080136,0.5790033937,-0.2179683894,-0.3166865706,0.0283836182,0.0546553992,0.1528549641,0.0856237188,-0.0284987781,0.126384452,0.1029291824,0.062597245,0.1315194219,-0.0376137607,-0.1843018383,0.1243039593,0.1284985542,0.0001874979,-0.0889901295,0.1555132568,0.2396557927,0.2841664553,-0.3276495337,0.4977139533,-0.3636688292,-0.100325197,0.0216880515,-0.2143646479,0.528365314,0.4158993959,0.142138347,0.3851136565,0.0473217331,0.0473606698,-0.1936742812,-0.1166506931,0.3290860355,-0.2601207793,-0.4622159004,-0.7009153366,0.4129931927,0.0718453452,-0.2251184285,0.2465713322,0.157099545,-0.3763374686,0.5711110234,-0.0197212305,0.8757333159,0.0014644306,0.1962808669,0.0199150406,-0.2845607102,0.6315014362,-0.0150956372,0.1031225547,-0.3777364194,0.3019941151,0.0610180013,0.0848897845,0.2285560668,0.5355205536,0.0136506241,0.1836872697,0.0637411103,0.2786479592,-0.2928549051,0.3380337954,0.2074407935,-0.0835545063,-0.1413369626,-0.0422163233,0.026887605,-0.2585878074,0.1981573999,-0.1889909953,-0.0228493884,0.1410942674,-0.2769814134,0.0659616664,-0.0040316191,0.1196418554,-0.1257599741,-0.44830814,0.432087779,0.1650522202,0.3249994218,-0.1801125854,-0.0380750634,0.4714801311,-0.0581888519,0.0914947391,-0.1675043106,0.0250935182,0.2214222401,0.0295112282,0.1062862128,0.0263106246,-0.0252752639,-0.0637804866,-0.2644971907,0.2150605917,0.4529538751,-0.5559501052,0.1746736169,-0.0129838968,0.0524932183,-0.2350749969,0.017459536,-0.2393327504,-0.2196990848,0.0163796227,-0.1804942191,-0.0482522883,-0.0266843103,-0.295588553,0.2380402386,-0.0153078651,0.4281603992,-0.1167142987,0.1651086062,-0.2321613133,0.076061897,0.1726447493,-0.2626329362,0.2844870687,0.1135672629,-0.2745529413,0.150613904,-0.0156046515,0.0522632822,0.6398307681,-0.1539235413,0.0154650565,-0.2086797804,0.0041732374,-0.0666328892,0.0633293837,-0.0451880172,0.4669850171,-0.1915590167,0.4138614237,-0.1375510097,-0.0185930189,0.2807841599,0.2623932064,-0.1522808224,0.0096322559,0.1951939613,0.075289689,0.0053583211,-0.0065126806,0.2042772621,-0.0604877993,-0.3154309392,0.1255770922,0.1368715912,0.1940066814,-0.3965695798,-0.1686493605,0.0718194842,-0.1536899358,0.1787269562,0.0668720901,0.2828257084,0.4819812477,0.3057694435,0.1115998402,-0.3644501567,0.1841042638,-0.1227383018,0.3808992803,0.2149303257,0.2762943208,-0.1651396453,0.0946763605,0.1048572361,-0.0490785502,0.1903125793,-0.3088617921,0.0137656527,0.0328457952,-0.0307464842,-0.1496913731,0.2588714361,0.567273736,0.126431793,-0.0025872372,0.2297625989,0.0861496627,-0.0910127535,0.2563649416,0.2815908492,-0.1594063789,0.0842419565,0.5541467071,0.1247868165,0.2655094862,-0.4000850916,-0.0335832685,-0.0566166006,-0.0443628542,0.5013854504,0.8824549317,0.2060433775,-0.019955853,0.4297123551,-0.1907733083,0.2422768772,0.4915966392,-0.1730972379,0.3304674327,0.3147028089,-0.1226274222,0.5907780528,0.1948813945,-0.4043379426,-0.0030669447,-0.0167705882,-0.2341190577,-0.1521472186,0.761743784,-0.0115821445,-0.3254048228,0.0941189826,0.2618059516,0.1801455617,-0.2026370764,-0.1622175872,0.0550562292,-0.3555076718,0.008617959,-0.0889240056,-0.2062522322,0.3511845469,0.0373201668,0.2648570538,-0.1526114941,-0.032471858,0.1779340357,0.502948761,-0.3182084262,-0.1020573974,-0.0024602599,-0.2936457694,0.1971894205,-0.092733793,-0.005650823,-0.1405810267,-0.3290474713,-0.3568285704,0.202842325,-0.1465294361,-0.0000996979,0.0366201065,0.1252532452,0.303735584,0.1848145723,-0.0483888835,-0.0262146536,-0.1517801881,0.2818922997,-0.0748927742,-0.0471144617,0.6482107639,0.1358209848,-0.1838163286,0.0230755992,0.1705509126,-0.2316404134,-0.0665177554,0.3410791755,-0.0482521579,0.117223911,-0.1129511595,-0.0336049758,-0.0348389633,0.4051429629,-0.2068143785,-0.2867167592,-0.261464417,-0.4266864061,-0.0807617828,-0.0739794672,-0.1493709087,-0.2486891598,-0.0993375704,0.1381161064,-0.0872694775,-0.1610429585,0.1189860851,-0.0762759894,0.3919470012,0.3623470962,-0.22487095,0.006812613,-0.0320697352,0.1607149243,0.1272042096,0.1112387627,-0.1088411063,0.3447792828,-0.3298472762,0.1718093008,-0.0711511001,0.5625978708,0.2020675987,-0.0593351908,0.0123897055,0.2377561629,0.0298560318,-0.063521862,0.0353422388,-0.0028580634,-0.5393224359,0.5002480745,0.0172593575,0.1763503999,-0.2386563718,0.4447182119,-0.146188125,-0.0446424596,-0.2594881058,0.1652462631,-0.0692133904,0.2242832482,-0.060414765,0.3944737613,0.4205290079,0.2640945911,-0.1234448254,-0.0252444558,-0.211769104,-0.3554520905,0.2351640761,-0.0745393187,-0.1371604204,0.1033388153,0.3187500536,0.1716340333,0.016238939,-0.7228699327,-0.269102782,0.1650721282,-0.2762211263,0.0203279126,0.0386570096,0.2319524735,-0.2397077531,0.0130427713,-0.0458081104,0.2986752987,-0.079456456,0.0646827444,-0.1154509261]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/172","title":"Clone not working on Windows environment","comments":"Should be fixed on master now :)","body":"Cloning in a windows environment is not working because of use of special character '?' in folder name ..\r\nPlease consider changing the folder name ....\r\nReference to folder -\r\nnlp\/datasets\/cnn_dailymail\/dummy\/3.0.0\/3.0.0\/dummy_data-zip-extracted\/dummy_data\/uc?export=download&id=0BwmD_VLjROrfM1BxdkxVaTY2bWs\/dailymail\/stories\/\r\n\r\nerror log:\r\nfatal: cannot create directory at 'datasets\/cnn_dailymail\/dummy\/3.0.0\/3.0.0\/dummy_data-zip-extracted\/dummy_data\/uc?export=download&id=0BwmD_VLjROrfM1BxdkxVaTY2bWs': Invalid argument\r\n\r\n","comment_length":7,"text":"Clone not working on Windows environment\nCloning in a windows environment is not working because of use of special character '?' in folder name ..\r\nPlease consider changing the folder name ....\r\nReference to folder -\r\nnlp\/datasets\/cnn_dailymail\/dummy\/3.0.0\/3.0.0\/dummy_data-zip-extracted\/dummy_data\/uc?export=download&id=0BwmD_VLjROrfM1BxdkxVaTY2bWs\/dailymail\/stories\/\r\n\r\nerror log:\r\nfatal: cannot create directory at 'datasets\/cnn_dailymail\/dummy\/3.0.0\/3.0.0\/dummy_data-zip-extracted\/dummy_data\/uc?export=download&id=0BwmD_VLjROrfM1BxdkxVaTY2bWs': Invalid argument\r\n\r\n\nShould be fixed on master now :)","embeddings":[-0.072222136,0.2448333353,0.023837233,0.0023602808,0.0249441881,0.0223568529,0.2989340425,0.2546470761,-0.1417268962,0.281490773,-0.0207462665,0.1428505033,-0.1080002785,0.0421381444,-0.0966366902,0.0482945628,0.0968503579,0.1433014274,-0.4318033159,-0.2644028068,-0.2747170329,0.2689005733,-0.0428499132,0.0453163758,-0.2072145641,0.2817907333,-0.2760969996,0.5728909969,-0.3486270905,0.235895142,-0.0706248954,-0.2487677038,0.2875396907,0.3314112425,-0.0001004764,0.1258598268,0.398378551,-0.0148943365,-0.1658406109,-0.1123942658,0.1845993996,0.1027483419,-0.2500542998,-0.3065808117,0.0319056921,0.0913026556,0.1163843423,0.4030434489,0.3522788882,0.3673146069,0.2976339757,0.1192450672,0.2610049844,-0.1825359017,-0.1404657662,0.2976892292,-0.0690009519,0.4594821036,-0.0376148708,0.017333312,0.3608736694,0.0096988184,0.0706227273,-0.0373014696,-0.254863441,-0.0095023979,0.3364208043,-0.1627759188,0.1408889741,-0.1342355311,0.2080250382,-0.0774089843,0.0524383709,0.5660685897,0.0754018202,-0.2615524232,0.1736626923,0.2411269546,-0.0961631835,0.1585930735,-0.2811504006,-0.0354615785,-0.3392667472,0.1116540805,0.1139140129,0.0279653464,-0.0728786513,-0.057787247,0.230243668,-0.047773283,0.216093868,-0.1623906493,-0.1273185462,0.1297035068,0.3606819808,-0.1460002363,-0.173347488,0.459835887,-0.1004980803,0.2395251095,0.0246754512,-0.2578923404,0.478312701,0.2191285193,-0.1387733519,0.2738290727,-0.1358165443,0.1978676468,0.1633841693,0.0309821144,-0.232168898,0.0354691967,-0.3530723751,-0.1297312081,0.1547057331,0.2091253698,0.2927966416,-0.2596560717,0.0705518201,-0.0804402828,-0.3337473273,-0.015225336,-0.1159836873,0.0960030556,0.1412661374,-0.1261460483,0.195800364,0.2296052426,-0.0671603307,0.043263983,-0.1429541558,0.2744924426,-0.3325434327,-0.2576068342,0.2653585672,-0.034149576,0.2292973548,-0.0239505693,-0.0270926375,0.0601533987,0.0934190527,0.1622604579,-0.1387079656,0.149292931,0.0394675918,0.0589042418,0.0506281704,-0.0766592994,0.172245428,0.0575826317,-0.1937683374,0.0199564993,-0.0844110548,0.3256950676,-0.0097129168,-0.1034265161,-0.1141900644,-0.2750374079,-0.0666876063,0.2283204794,-0.0954767168,0.1911850721,0.2163905799,-0.15256688,0.0384696834,0.049322702,-0.3134994507,0.1960304379,-0.1534813046,-0.1394629329,0.2925137281,0.3932116628,0.0005033063,0.1036287844,-0.1604764909,0.1192724407,-0.0427648425,-0.0014537257,-0.0906808898,0.0645554364,-0.2555452287,-0.3114949167,0.0565860607,-0.1767115891,0.209916532,-0.1165923327,-0.3363350034,0.2085995376,0.0553079322,0.0908081084,-0.3295405209,-0.1527791768,-0.0015719,-0.1187982485,-0.0757644996,0.2273662239,0.0129158469,-0.0196354631,-0.0727255419,-0.3213067949,-0.0393163897,0.1791120023,0.4310134351,-0.2057016492,0.0320165232,0.1405246109,0.0869954824,0.0064839213,-0.2331197113,0.4354610443,0.2754903138,0.1357610822,-0.2326268852,-0.0930869728,-0.0889473855,-0.1740344763,0.2615979016,0.1147457361,0.3382464945,-0.1222446412,0.1227741465,-0.0997201055,-0.1520085335,0.0606649667,-0.0528823845,0.3292134106,-0.5861177444,-0.2164282948,-0.0389043912,-0.0010102254,0.3021003604,-0.247842595,-0.0984012485,-0.0342432894,-0.3530802131,0.1976795048,0.1191911474,-0.2203859836,-0.1124910265,-0.34930709,0.3360770047,0.7471868396,0.0356415324,0.0213295948,-0.0720892474,-0.1365536153,0.2094161361,-0.1203447729,0.0347071849,-0.0330293439,0.0348995477,-0.3069928885,0.3250964582,-0.0512312353,0.3154071271,0.3885475099,0.0986228064,0.1378457695,0.0933081135,-0.0948079079,-0.0605315939,-0.0905895904,-0.0493021533,0.0714294985,-0.0261474401,-0.0636419728,-0.1859407723,0.2375640869,0.551318109,0.2097958028,0.2601013482,0.2422053069,-0.0246897936,-0.3146793246,0.5383503437,-0.4369964004,-0.0534145646,0.0410596728,-0.0869878083,-0.1486067623,-0.4210665226,0.2169430107,0.2712532878,0.2080214322,-0.1840510815,-0.1737326384,-0.1488974392,-0.054371424,-0.3322834074,-0.0948463231,-0.1173314378,-0.0869291872,0.1905817538,0.3967081308,-0.1015569121,-0.0464454144,0.2194819301,-0.1326781809,-0.1391257495,0.1374510825,-0.1876312196,0.1575722247,-0.2401415855,0.193195045,0.1112679243,-0.0203291345,0.2155266851,-0.0284942538,0.1375362575,-0.4268456697,-0.2420138717,-0.0687608495,0.0753046647,0.3825212419,0.0784514546,0.2060335428,0.2129829824,-0.3221800923,0.2472830117,0.0076336893,-0.3229382336,0.0563428253,-0.0147618242,-0.4120503068,-0.335398525,-0.0728751495,-0.3213318586,-0.1157653183,0.0112150088,0.0800300166,0.102025874,0.2178753763,-0.0490041599,-0.1886505038,0.2975283861,-0.0231101606,-0.3065622747,-0.4778422713,0.3166108727,-0.4249462187,-0.4392322004,-0.0566545911,0.1612659097,0.0422225669,-0.204241097,-0.1580410451,-0.2004250288,-0.0662673712,-0.1000047252,0.2305028141,0.2938475311,0.2642723024,-0.1342629343,-0.2331430316,-0.0545398146,0.0415431112,0.1485749781,0.0509871729,0.3026665747,0.1267609596,0.2182439268,-0.1913610101,0.3639514148,0.3213573098,0.2043498605,0.497913301,-0.0319769159,0.2838976383,0.0573247708,-0.3223094642,0.2388723791,-0.0967905298,-0.3097503483,0.3070206344,-0.0863972455,-0.2951928675,0.0660897642,-0.4048424661,-0.2270951867,-0.1458414644,-0.0704447031,0.2694574594,0.0571236573,-0.1120978147,-0.0658272654,0.0334771462,-0.03294475,0.1168264151,0.2789962292,-0.3421440721,0.1257906109,-0.5290076137,-0.4662916958,-0.4519834518,0.2021813542,-0.124905549,0.2663232386,-0.1641506106,-0.0429331474,0.013287208,-0.1681153029,0.5639018416,-0.2676967084,0.0450704657,0.2706233859,0.2872689962,0.0685524493,-0.1188501939,0.0777505636,-0.04120408,0.0391004793,0.0805139914,-0.04715975,0.0132007049,0.2983353734,0.1667170078,-0.2445832193,-0.0724719018,-0.0699666739,-0.3746201992,-0.2109761685,-0.1546019912,-0.1544813365,0.1461153328,-0.0463195369,0.2360853404,0.110289067,-0.1293008775,0.2526634037,0.0506031774,0.230439201,-0.0661839321,0.1272159368,-0.2387236953,0.2643558383,-0.0395134799,0.3928150833,-0.4844264984,0.1530029774,-0.1072685048,-0.2340084165,0.2857706249,0.2561480999,-0.1032700166,-0.338690877,-0.2820525765,-0.0491128564,-0.1600961089,-0.3124428093,0.0835359767,0.011379106,-0.0500588417,0.4226410985,0.1014351249,0.0519079044,0.0830269754,0.1142880991,-0.0063738241,-0.2640431523,-0.076966241,-0.1186283529,0.725125134,0.2714959085,0.2328570187,-0.010862248,-0.2590369284,-0.139045164,0.2291622162,-0.2503413558,-0.255176872,-0.2901045978,0.0693372712,-0.1040380821,0.0433895215,0.4309067428,-0.3841441572,0.1984419227,-0.1517775506,0.1327618062,0.2259584665,0.3442968726,-0.1810701489,0.0501961261,0.119255513,0.3374144435,0.1658715457,0.0703838691,-0.0725776255,-0.1188657209,0.0166421011,-0.1780635417,-0.1480855793,0.3903591633,-0.539506495,0.1396297067,-0.0458505899,-0.0108185345,-0.2320262194,-0.2847661972,0.4421077073,0.1606908441,-0.1732674688,-0.2006811649,0.2146002054,0.0211995766,-0.0568995923,-0.0590677485,0.2522714436,-0.1874774694,-0.146083504,-0.1275588274,0.0001821789,-0.0117863957,-0.2684288323,-0.2456561476,-0.0222869646,-0.0731253251,-0.3782897592,0.1309483796,-0.4865331352,-0.1150884628,0.2418204695,-0.0546628907,-0.3059399128,0.3050224483,-0.3368273973,-0.1428214461,-0.0399043374,0.3562360108,-0.0328445695,0.049416624,0.3531682789,0.0039427122,0.0354984365,-0.3449535072,0.2765756249,-0.0010422505,-0.1273308396,0.1510382295,0.0667617396,-0.1991696358,0.2096076906,0.3200406134,0.2535898089,0.171135366,-0.246814236,-0.27180475,-0.1921629012,-0.0417220592,0.091878809,0.2433555275,0.0444000177,0.3901061416,0.387011975,-0.2401217818,-0.4279720783,0.353241086,-0.0122047365,0.0915510803,0.1788835824,-0.3220150173,0.4172350168,-0.0467356108,0.2385679632,0.0164537691,-0.0917795077,-0.2980648279,-0.2070278227,0.0271326043,0.0193465631,-0.1528651267,0.4074489474,-0.2754353285,-0.0518073812,0.2478256375,-0.0961036906,-0.1376867145,-0.1707592756,0.2113671005,0.1478029042,-0.1243506819,-0.1174094453,-0.0427656062,-0.0935956389,0.3921006918,-0.4428366423,0.0922852084,-0.0389267951,-0.1825051606,-0.3932494223,-0.0622072034,-0.266626358,0.2478535026,0.2050534636,-0.0009702187,-0.2929422557,0.0083407983,0.3782891631,0.1257910877,0.0788734257,0.1873054653,0.4633595645,0.2992500067,-0.1994764954,-0.0673308671,-0.1439285278,-0.2540596128,0.0136024635,0.0490105227,-0.11787007,-0.3107616305,0.0001999888,-0.1415438801,0.3910624385,-0.0937714204,-0.1092248261,0.3724509776,-0.1257507503,0.1235124618,-0.0954608098,-0.2249797583,-0.1935902387,-0.0403877497,0.124329567,0.1266496181,0.1606333703,0.0360285118,0.332236141,0.0680777431,0.0299707334,-0.0996248722,0.3529308736,0.4193640947,-0.0420958214,0.4078550637,-0.3505252302,-0.2482173145,-0.2436001897,0.3625846803,0.0019077022,-0.0002637132,-0.4164118767,0.0368287601,-0.1900624931,0.0137319323,-0.0074484004,-0.0589605123,0.1648282111,0.326721251,-0.0182219315,-0.1797696352,-0.4267650843,0.1990403533,-0.1709714681,-0.3845579624,0.3275022209,0.0040303897,-0.1393790543,-0.0433249697,0.4704885781,0.4540001154,0.2992688119,-0.0126469731,0.0884597227,-0.1609196067,-0.133567065,-0.0493253432,0.2206128836,0.1732659787,-0.1592150331,0.0012932969,0.221387282,-0.1387868375,0.0289685745,0.180723533,-0.2821901143,0.284558326,0.2462350726,0.1375713497,0.0077698785,-0.3382768929,-0.2160674185,-0.565577209,0.1590144336,0.1435901523,-0.2783200443,0.0334108546,-0.1697915196,0.1671974361,0.1027230918,0.4168748856,0.2916252613,-0.0801814198,-0.4434960186,-0.2354782969,-0.254656136,-0.0892739147,-0.127441749,-0.1121546552,-0.2036813647,0.059575513,0.0885942876,-0.0216813069,0.0741848499,0.4026126564,-0.3024293184,0.2509666085,-0.262262404,0.1121180877,0.3672163188,0.0336301625,0.0848906562,-0.4120741487,0.2501886487,-0.0469695106,0.2139582932,0.0080642579,0.0601195917,0.0044974219,0.3935307264,-0.0779612884,-0.037178196,0.1817575544,-0.0839915499,-0.1857419014,-0.02172089,0.2324300557,-0.3299225867,-0.0219214913,0.4733914733,0.1162245125,-0.0949816182,-0.0793488771,-0.2898656726,-0.1030459553,0.0857874379,0.003220214,-0.2350972891,0.2881885767,0.0219509546,-0.088861756,0.1066681966,0.3720402122,-0.0909863859,0.1734986007,-0.3821181953,-0.2914468944,0.3656570315,-0.2981884778,-0.2493096739,-0.0522251688,0.1046447754,-0.468151778,0.0697628111,-0.3307999372,0.1101167724,0.0325100832,-0.1893388182,0.0065616369,0.417762965,0.2122422308,-0.0090517029,0.0786132365,0.219319582,0.3824792802,-0.2901660204,0.1917784214,-0.0367028527]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/172","title":"Clone not working on Windows environment","comments":"Thanks @lhoestq \ud83d\udc4d Now I can uninstall WSL and get back to work with windows.\ud83d\ude42","body":"Cloning in a windows environment is not working because of use of special character '?' in folder name ..\r\nPlease consider changing the folder name ....\r\nReference to folder -\r\nnlp\/datasets\/cnn_dailymail\/dummy\/3.0.0\/3.0.0\/dummy_data-zip-extracted\/dummy_data\/uc?export=download&id=0BwmD_VLjROrfM1BxdkxVaTY2bWs\/dailymail\/stories\/\r\n\r\nerror log:\r\nfatal: cannot create directory at 'datasets\/cnn_dailymail\/dummy\/3.0.0\/3.0.0\/dummy_data-zip-extracted\/dummy_data\/uc?export=download&id=0BwmD_VLjROrfM1BxdkxVaTY2bWs': Invalid argument\r\n\r\n","comment_length":15,"text":"Clone not working on Windows environment\nCloning in a windows environment is not working because of use of special character '?' in folder name ..\r\nPlease consider changing the folder name ....\r\nReference to folder -\r\nnlp\/datasets\/cnn_dailymail\/dummy\/3.0.0\/3.0.0\/dummy_data-zip-extracted\/dummy_data\/uc?export=download&id=0BwmD_VLjROrfM1BxdkxVaTY2bWs\/dailymail\/stories\/\r\n\r\nerror log:\r\nfatal: cannot create directory at 'datasets\/cnn_dailymail\/dummy\/3.0.0\/3.0.0\/dummy_data-zip-extracted\/dummy_data\/uc?export=download&id=0BwmD_VLjROrfM1BxdkxVaTY2bWs': Invalid argument\r\n\r\n\nThanks @lhoestq \ud83d\udc4d Now I can uninstall WSL and get back to work with windows.\ud83d\ude42","embeddings":[-0.138573125,0.0753224865,0.0432779938,-0.1176929697,0.2541937232,0.0252196249,0.2679175138,0.0847872943,0.3106636107,0.2151383311,0.0513728857,0.1750458777,-0.046355512,-0.0310346577,-0.0004376394,-0.0913323611,0.1437229663,0.0368737392,-0.157195121,-0.1312250197,-0.3091589808,0.0827655196,-0.1326295584,0.1460978687,0.0630805492,0.2135993987,-0.1923877001,0.4833094478,-0.2760517001,0.3006512821,0.1543378979,-0.3294268847,0.3273667097,0.1675844193,-0.0001031563,0.0690575689,0.3950113356,0.0083316192,-0.1028622612,-0.0074998499,0.1889351755,0.0553952344,-0.1991931498,-0.2031341046,0.1091696471,-0.0328162275,0.02187871,0.0962846726,0.4857789576,0.1962793022,0.2990598679,0.0518182032,0.3104912043,-0.0439594239,-0.027456129,0.4404866993,-0.173062861,0.373539865,0.2122645974,-0.1842420101,0.4183603823,0.1763753444,-0.0817390308,-0.0155053958,-0.2984766662,0.0446572527,0.0110998359,-0.0356381685,0.3890959322,-0.1020762846,0.237925306,-0.0215259846,0.1173015162,0.5437575579,0.1917372048,-0.188138023,0.1376507729,0.3476985097,-0.1208565086,0.1123520732,-0.1828539521,0.0639399067,-0.236844644,0.0026266705,-0.082869865,0.2680434287,-0.0517313965,0.0858067125,0.0731447637,-0.0074151186,0.2454094291,-0.0692484155,-0.2371811718,0.0520878099,0.3467778862,-0.1147866324,-0.1339465678,0.4888749123,-0.0232168008,0.1469351947,-0.0056854868,-0.1769247055,0.7225176096,0.079152219,-0.2115031779,0.4582766593,-0.2050025016,0.0769912377,0.1606249362,0.0404651724,-0.0885350853,0.0315060541,-0.2759741247,-0.1074961275,-0.066377148,0.1274151504,0.4179158509,-0.1042178795,-0.2353761643,-0.0724691227,-0.4965019226,-0.0096851755,-0.0889826789,0.2138242126,0.2212096751,0.0811585486,0.358427614,0.1072374284,-0.0226588529,0.1280086786,-0.0982530713,0.2453273535,-0.3467870653,-0.2594353855,0.2202048004,0.0218573976,0.1724885404,-0.0608316027,-0.1606435031,-0.0833962038,0.0607411489,0.2444689721,0.0078681074,0.0516349599,-0.0130821932,0.213446483,0.0937716588,-0.1900781989,0.0499609858,0.1315303743,-0.2052370757,0.0224688612,-0.1987413019,0.3022593558,-0.0231187381,-0.1072839871,-0.1288536489,-0.1972728819,-0.2097360641,0.5116514564,-0.1017253548,0.1558985263,0.1837887913,-0.0599549264,0.0429647341,0.160417363,-0.4028478861,0.1004882455,0.0967739597,0.0173545219,0.3772863746,0.3956894279,-0.066045709,0.1144016758,-0.0917334706,0.1366997063,0.1665385067,-0.0035983115,-0.1226901487,0.2038213015,-0.029619446,-0.1513480842,-0.0814859718,-0.1658518463,0.1067233086,0.070783481,-0.5377454162,0.2283424586,-0.0284293015,0.127339229,-0.3539429605,-0.3663363159,0.1281239837,-0.1663532853,-0.1027627736,0.0642471313,-0.0178903472,0.1416255534,0.0824364647,-0.2920839489,0.0351901688,0.0770203695,0.5139864683,-0.0549625568,0.1092457175,0.0046111215,-0.0629648417,0.0293066502,-0.032569401,0.2236389369,0.353995204,0.0798154175,-0.2595305443,0.023622334,-0.3107394576,-0.2830114067,0.2093310505,0.0824982077,0.2776260674,-0.2002555877,0.0743567646,-0.144382298,-0.3849539459,0.0631452873,-0.0664263144,0.2778089345,-0.6239449978,-0.1490679979,-0.2481540442,0.0238125641,0.1918224096,-0.0393068418,-0.0104343984,-0.1473257095,-0.4532571733,0.2423190176,0.0599578768,-0.2652384341,-0.1807869673,-0.5600167513,0.4569556713,0.6329239607,0.0921544209,-0.0033153198,-0.0767904669,-0.3254493475,0.3893317282,0.0301288106,-0.1231343597,0.0026799098,0.1022843942,-0.3693817258,0.2838276327,0.0592696257,0.3101095855,0.25448668,0.0336791724,-0.11739178,0.0737938657,-0.1460081339,-0.0049873465,-0.0511737429,0.0173758697,0.1301012784,-0.1792789847,-0.0658790097,-0.2332678139,0.2563056946,0.5190519094,0.1071394458,0.1860898286,0.1466433257,-0.113399744,-0.1873608977,0.5150583982,-0.22070539,-0.1726084054,0.0146081792,-0.0277135018,-0.0073046484,-0.4541345835,0.2388757318,0.1164531931,0.1234108582,-0.2458884716,-0.0041689659,-0.1013178825,-0.0989056826,-0.4457656741,-0.1620376706,0.019912025,-0.0953403115,0.1811239272,0.2544880211,-0.1376120001,0.0064111673,0.1947747171,-0.0749539956,-0.1997277439,0.248385042,-0.3092172146,-0.0027783758,-0.2196460366,0.1999790221,0.1419139057,-0.0363224149,0.3171083629,0.0374389738,0.0096639721,-0.6015461683,-0.323459506,-0.1355333477,0.0969577357,0.2650502026,0.1557774991,0.3194543719,0.1791657507,-0.335242033,0.2689843774,-0.0870768204,-0.1989187002,-0.0186926052,-0.0936733335,-0.2805403173,-0.3046529293,-0.1009982079,-0.3088158369,-0.1386726946,-0.0191424992,0.0718756467,-0.1592180431,0.0934967548,-0.0445632041,-0.2178962678,0.3227195144,-0.0125877485,-0.280407697,-0.2826620936,0.2275710404,-0.1196778268,-0.3790894449,0.0995166004,0.0553945675,-0.1492934674,-0.0890928954,-0.1426102221,-0.1828903407,-0.0977795124,-0.1053223759,0.1505232006,0.285766542,0.356402576,-0.1008033976,-0.1462134868,0.0264963135,0.0740749389,-0.045961041,-0.0136725353,0.4960773587,0.1189406961,0.1874797493,-0.2066273987,0.3111021519,0.3284835219,-0.0054594819,0.3849785626,0.076137118,0.1575180739,-0.0219362564,-0.275564909,0.2082762122,-0.1231391802,-0.0601361208,0.2984976768,0.1581945568,-0.173581332,0.0106586069,-0.1761052907,-0.1205290407,-0.0809909031,-0.0004312826,0.2925980389,0.0550485253,-0.0633985028,0.135133937,0.1222665012,-0.2105230838,0.2414478362,0.2745229602,-0.3730316758,0.0716911107,-0.3720548451,-0.3680813313,-0.536734283,0.228797704,-0.0892420411,0.1121103391,-0.0390670672,-0.2109885514,0.0503769182,-0.2737465799,0.3756831586,-0.1886711866,-0.047415439,0.2199393958,0.3986917436,0.0984791145,-0.1820390522,-0.0095616374,-0.3653271794,-0.0135474447,-0.0019389269,0.0447695665,0.0406387821,0.3384926617,0.3540420234,-0.3199862838,0.0118495924,-0.0725955591,-0.6396974325,-0.1330639571,-0.1848202348,-0.2118290663,0.018529268,-0.11885418,0.1521568149,-0.0241771955,-0.1165029854,0.1740528792,0.0055299932,0.1458612531,-0.0098207472,0.0449001081,-0.3741801381,0.7470428944,-0.0722432733,0.3618466854,-0.4434897304,0.1923591793,-0.1547881812,-0.1841194928,0.0560270958,0.1991999149,-0.183482036,-0.1484571099,-0.3923021853,0.0040170513,-0.5023021698,-0.3014469147,0.3122556508,0.0337166972,0.1784558594,0.5529619455,0.1499446332,-0.1118241921,-0.00094866,0.2827709317,0.18203035,-0.2480801046,-0.0984601378,0.0087709464,0.816706717,0.2510704398,0.058530774,0.1925565153,-0.3850495219,-0.0748106763,0.3348942697,-0.1361796409,-0.1669280678,-0.0834251419,0.0057119578,-0.0021724685,0.3211739361,0.2632608414,-0.2532701194,0.0722721294,-0.0547521599,0.3172453344,0.1574831456,0.1761447787,-0.1950591803,0.0928905904,0.0575552993,0.3123197556,-0.0228854641,0.2604388595,-0.0466499887,0.1577220857,0.1625836492,-0.4241305888,-0.3638002276,0.3909408748,-0.3333904445,0.0977929384,-0.0578624941,-0.2078972906,0.0762812495,-0.0810162425,0.5154328346,0.0731946602,-0.2263771147,0.0101651559,0.2000533342,0.1786486655,-0.0011338608,-0.0493704602,-0.0324649103,-0.1518778354,-0.1423225105,-0.1839989573,0.0207400564,0.0135681313,-0.4329721928,-0.1437359154,0.2765734494,-0.2024637759,-0.0397892855,0.0596053824,-0.2623001933,-0.2453421652,0.2204214036,-0.1586541384,-0.1671632528,0.4264075756,-0.4856819808,-0.1034528464,-0.0158170629,0.2208563834,-0.0908640251,-0.0688254759,0.3984332383,0.0872395113,-0.1678449363,-0.3704929948,0.2475438118,-0.0711465329,-0.1413691938,0.0072612646,0.2366068363,-0.2320280522,0.1005599201,0.3537056744,0.1876672506,0.1616673768,-0.1852502674,-0.2074624449,-0.0920225009,-0.0299089029,0.2349598259,0.3093287349,-0.1087051556,0.2906647325,0.3151295483,-0.1296312809,-0.3982983232,0.3040527403,-0.0158896372,0.0932113603,0.0534178354,-0.2377379686,0.4126282334,0.0458041243,0.2643339336,-0.0483549312,0.0617044158,-0.3289490342,-0.1707849801,0.0009535875,-0.1211513653,-0.1902440786,0.4958486259,-0.2976509929,0.0552861318,0.1906770021,-0.0124131916,0.0514259189,-0.13405931,0.1578755826,0.0681019202,0.0051059979,-0.0852242932,0.1064840928,-0.2583217025,0.6361700892,-0.3880790472,0.0247927103,0.2630324364,-0.0822810382,-0.4307810962,-0.0354134738,-0.1167553216,0.0948849022,0.1292642057,0.0648873374,-0.372153759,-0.0321842618,0.3071150482,0.1832013875,-0.1076284498,0.0731765777,0.4422773123,0.3079825044,-0.3165112138,0.0144415582,-0.1830841452,-0.3784549236,0.0607668087,-0.0683916137,-0.2303716242,-0.315823853,-0.0076384367,-0.2654208243,0.5080600381,-0.0405946597,-0.1633206457,0.3552668393,-0.2061000317,0.0550406389,-0.2795135379,-0.1755055636,-0.1286426187,-0.1876980811,-0.0359142348,0.197094366,-0.0568998381,0.3380410373,0.2133136839,0.0257699303,0.0026384736,-0.2461902946,0.0884447247,0.2186960727,0.0101402244,0.3654832542,-0.1213158742,-0.2451049834,-0.1636710465,0.3953779936,-0.0422700234,-0.13263686,-0.6377023458,0.1536540091,-0.0323935784,0.0411088243,-0.0243507344,-0.0895042941,0.0583391264,0.1364949793,-0.1026168168,-0.1083800644,-0.2655323744,-0.038994465,-0.0920030922,-0.3378475606,0.3289667666,0.2038041949,-0.0770935714,-0.0853537545,0.5034739971,0.3212512136,0.2087230086,-0.0095558064,0.2288976312,-0.0522151999,-0.1634650677,-0.081567049,0.2148257792,0.0825831145,-0.134690702,0.0815827176,0.1905126423,-0.1339314431,-0.1457169354,-0.0258021355,-0.0887648985,0.4599476159,0.2711461484,-0.0172412228,0.0317939781,-0.3238120973,-0.0211720895,-0.4186550081,0.2498486042,0.0230138786,-0.1193417907,0.1089133248,-0.2100325525,0.1473279148,0.0470000617,0.339964509,0.3005589843,-0.0573183224,-0.2957144082,-0.3858345747,-0.0746192634,0.0263589732,-0.2688777149,-0.2594425082,-0.413060993,0.0006401211,0.0617982559,-0.1395887136,-0.0281557404,0.4201024771,0.0073420354,0.3265376687,-0.1560195833,-0.0016211377,0.5315374136,0.1206562668,0.0477272049,-0.5655257106,0.2924608886,0.1440214664,0.0991826653,-0.1325057149,0.2148786336,-0.0736404955,0.3327274621,-0.1732082814,0.0544786118,0.1080718637,-0.0166857913,-0.1784480363,-0.0622335598,0.0602586791,-0.1126424968,-0.0280322433,0.4763785303,-0.0830427483,-0.0124268448,-0.0958502665,-0.2465522885,-0.2529411018,0.0301033519,0.1053107083,-0.429688096,0.1773757786,-0.0209224224,-0.0287912488,0.2134535462,0.2278049439,-0.1907121092,-0.0055095307,-0.3952746689,-0.2730879188,0.3617615402,-0.1816455126,-0.273068279,-0.1165433377,0.1142183095,-0.4685232341,0.1999057084,-0.2263014317,0.0858215466,0.0170557294,-0.2092676908,0.0691856742,0.3578792512,0.2712190151,-0.0923449695,0.0782063007,-0.1164605618,0.4681142271,-0.0542898215,0.1730793118,-0.0808440149]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/168","title":"Loading 'wikitext' dataset fails","comments":"Hi, make sure you have a recent version of pyarrow.\r\n\r\nAre you using it in Google Colab? In this case, this error is probably the same as #128","body":"Loading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'","comment_length":28,"text":"Loading 'wikitext' dataset fails\nLoading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\nHi, make sure you have a recent version of pyarrow.\r\n\r\nAre you using it in Google Colab? In this case, this error is probably the same as #128","embeddings":[-0.1193379238,0.0886276364,-0.0089776339,0.2101976722,0.3692684174,0.0260604266,0.4487809837,0.4131019711,0.4330818355,0.0125665311,0.1238447428,0.3954298794,-0.1505493075,0.0219624937,0.1209058017,-0.4289889932,-0.0337393396,0.3636732399,-0.0414666571,0.062973544,-0.1876838803,0.108516708,-0.2651918828,0.2550059557,-0.3057226241,-0.0570259951,0.2159208357,0.0820631012,-0.0789184123,-0.5590060353,0.2770831883,-0.2170494348,0.0644511506,0.1147467494,-0.0001099558,0.205306232,0.3326975107,0.0302131604,-0.5799459815,-0.1895776838,-0.4773317575,-0.1189394891,0.2974321842,-0.2497071624,-0.028513357,0.0538973026,0.0359112546,-0.0744121596,0.2967469394,0.5315061212,0.2378886342,0.5236920714,0.1545488685,0.0344940759,-0.0785685405,-0.1705711782,-0.0821156427,0.2239678055,0.0738770068,-0.3915446997,-0.0133740734,0.1269243062,0.1511732191,0.2637012899,0.5083716512,0.0143935969,0.240522787,-0.1506265402,0.0149465678,0.0764008239,0.3353828788,-0.1649338454,-0.2716008127,-0.2243597806,0.2476661652,-0.3347570896,0.2365306169,-0.0366766118,-0.1622047126,-0.0081965309,-0.2692284882,-0.0281475354,-0.2418207526,0.3526881039,-0.0582148433,0.4996492267,0.0435834154,0.1427477896,-0.0580658838,-0.075457558,0.1403762847,-0.0870907307,-0.1293002814,0.2578134239,-0.383156091,-0.0364675075,-0.0349461213,-0.1169757396,-0.0316241011,0.2635210156,0.219990015,-0.2248735428,0.3238382339,0.3154148161,0.2229899764,0.4204691052,0.0727646127,0.063592121,-0.0828047171,0.1360252202,-0.2337277681,-0.1554848254,-0.1555357128,-0.1765549034,0.0928974897,-0.1546022743,0.2260205299,-0.071116358,-0.4984739125,0.1147207841,-0.1434045881,0.046290502,0.2051148862,0.1794852465,-0.1276334524,0.2322352529,0.1441244632,0.2825672626,-0.2519766986,-0.0431360416,-0.1662892997,0.2492134273,-0.2343834639,-0.1822493821,0.2427540869,0.1934002936,0.3994777203,-0.1011680439,0.111264661,-0.2017549574,0.2273753434,-0.0455191471,-0.1288790852,0.2566530704,-0.2160614133,0.1548979729,0.2015910596,-0.2824236155,-0.0848238841,-0.10174492,-0.2726355791,-0.3080923855,-0.1262511909,0.2814154327,0.147381857,-0.2429646254,-0.1522630006,-0.0611871071,0.1474927962,-0.3687241077,-0.1181055456,-0.4612588286,-0.2206385881,-0.0230053104,0.3976424038,0.2049094886,0.1313870102,-0.1089720726,-0.1154211313,0.0778549612,-0.0317966752,0.1227136478,-0.4383572936,0.5421624184,-0.0610556342,0.118862763,0.6413468719,-0.3427866101,-0.4578541815,0.123568885,-0.0142798293,0.2093286514,-0.2253688425,-0.0167210288,-0.0132269189,-0.184551537,0.2123140246,0.6192135215,-0.0150432223,0.0824793652,-0.1268968582,-0.0155486753,0.5002205968,0.1098568216,0.0925252214,-0.0079286359,-0.0417051278,0.4966493547,-0.0013129428,-0.1259710789,-0.0148605099,-0.013630108,-0.0646863058,-0.104604587,-0.2178259641,-0.0738868713,-0.2776648998,0.3043035865,-0.117547363,-0.0585312918,-0.0471505858,0.1061868444,-0.3491126597,0.1684893072,-0.4232687652,-0.3773184419,0.1720929891,0.1551331282,-0.0108234957,0.2384415865,-0.2047417462,0.2194369435,-0.274028182,0.0555009618,-0.2189284861,0.1612891257,-0.0558832213,-0.0424223617,-0.0094791884,0.3087970316,0.0383313261,-0.0179966092,-0.3062886596,0.2860310078,-0.1254843473,0.0803563818,-0.0322646983,-0.1433701664,0.0688060746,-0.0715434849,-0.1751839221,0.2081920803,0.3319509625,-0.1801262945,-0.043142952,0.1128761917,0.1272121966,0.0814221054,0.0988641679,0.1639728397,0.1439921707,-0.0658260733,0.1012795344,0.0774310902,0.2337106764,-0.0949872434,0.2047105134,-0.1564514041,-0.167145744,-0.1188064739,0.3131195307,0.1700583696,0.167417407,0.0814866126,-0.275252223,-0.0948746875,0.0785581768,0.1460482925,0.4009645283,-0.0518327467,-0.1083415523,-0.0525707416,-0.4165976346,0.0711256638,0.1707080156,0.0366290882,0.2091285437,0.3018561602,0.0413160138,0.0873196498,-0.2952665389,-0.3570170999,0.2824332714,0.4405301809,-0.2284827828,0.2078800648,-0.0305316318,-0.0025111341,-0.0678482279,-0.3336339295,-0.188629657,-0.3429944813,-0.1213362366,0.4337745905,0.024119148,0.1810040623,-0.3977624178,-0.0220502764,0.1947504133,-0.1147857755,-0.0313790925,-0.3157985508,-0.269328475,0.0459421203,0.3818612099,0.0910761356,0.3510979414,-0.0667702854,-0.0925358012,-0.007033736,-0.0584152527,-0.0755677298,-0.1837216765,0.2440479547,0.009521856,0.2855836153,-0.1770296395,-0.5220327377,0.5017470121,-0.2557020187,-0.0577849969,0.3151076138,0.2711132169,-0.249190703,-0.0794775337,-0.3746882677,-0.1264498532,-0.280605495,-0.0594684482,0.0748294741,0.0620202571,0.3974690437,0.0708564147,0.2160111964,-0.0025732063,0.1134374887,-0.0482623316,-0.0900634304,0.1859846711,-0.310032934,-0.3173182309,-0.1213430315,0.0099359797,0.3347615004,0.1107365489,-0.3489591479,-0.2141319066,-0.1049579829,0.0853126571,-0.2213259488,-0.0400672704,0.1828597784,0.0723028034,-0.0657224879,-0.0866443887,0.035989143,0.0712399781,0.1523462385,0.1349620819,-0.0063045779,0.0892271996,-0.0848842263,0.6008986831,0.218693167,-0.2143773437,0.3002861738,-0.1032764912,0.005145404,-0.2086836398,-0.4398039281,0.2641642988,-0.11523287,0.1829283237,0.1723142117,-0.0794930086,-0.2247841209,-0.192558974,0.0688526183,-0.178447336,-0.2727790177,0.1461688429,-0.2173948586,0.2787112296,-0.0688640028,0.2225363404,-0.2052234709,-0.167981565,0.0698513761,0.166933611,-0.090545617,0.0796746612,-0.2440017164,-0.2761095464,-0.4799583256,-0.0278399177,0.2410321236,0.6437504888,-0.0863839611,0.0371890776,0.0807222873,0.0817412362,0.1373739988,-0.2301357985,0.1822634041,0.2891299725,0.0318248197,-0.3197474778,-0.0643306822,-0.1401119977,0.3580168784,0.0766504705,-0.0544316322,-0.037231572,-0.1676542461,0.1398317814,0.2599402964,-0.1872950047,-0.0441127233,-0.0917341635,-0.0898099318,-0.3077637553,-0.0061545791,-0.0099474099,0.3356595933,0.1229696721,0.2657254636,-0.5462990999,0.1166565344,0.0901664123,0.1745875478,0.0959533602,0.0579445958,0.187246576,-0.1666348726,0.1832098216,-0.145905599,0.5633814335,0.1473762542,-0.3307984769,-0.3195302188,0.0216632597,0.1674342602,0.1839805692,-0.0185859147,-0.1411214918,-0.2769463956,-0.061463993,-0.0087094773,0.3045702279,0.5610407591,0.0026275387,-0.3271814287,-0.2585409284,0.482839793,-0.2850035131,0.1173297018,0.321557641,0.053378012,-0.3131920397,0.0841390863,0.0778320804,0.6767753959,0.1695389003,0.311147064,0.3826704919,0.1839297265,0.4142697752,-0.043089699,0.2847148776,-0.3927352428,-0.1334252506,-0.0806859359,-0.065462403,0.1070814878,-0.0088732634,-0.0254380628,0.1506213844,0.1568897963,-0.004141801,0.057148274,0.4492524862,-0.0387835056,-0.1968196929,-0.3470262885,0.156016469,-0.2179998308,0.297508955,-0.0907334238,-0.0664841011,0.0552040972,-0.0673523247,-0.2555586398,0.2574389279,-0.299474448,0.459457159,0.0578220226,-0.4756816328,0.198415935,0.2070970535,0.1756633669,0.0056988141,-0.2995947301,0.2387785167,-0.2900909483,-0.2218656987,-0.0672868341,0.1119474173,0.4200738966,-0.1725511253,-0.146377489,0.1479405463,-0.1206544265,-0.2415021658,0.1039589867,0.0368667245,0.3524883091,-0.392575711,-0.5152193308,-0.00242018,-0.1820756793,-0.1749101579,0.1744999588,0.023450667,-0.189030841,0.1518952847,0.1498275995,-0.1989797652,0.1164065152,0.2480414361,-0.2495406866,0.1462838054,0.76375705,0.0339458175,0.0128051164,-0.3138304949,-0.0542888604,0.1250851899,-0.4819999039,-0.1171273738,0.1255033612,0.1691254526,-0.1442471892,0.4983651936,0.0630208924,-0.0188453011,0.2471855134,-0.542414248,-0.1062184349,-0.0430215374,0.0177533776,0.2377753258,-0.0026245497,0.114889577,0.1355090588,0.1033881679,-0.272637248,0.0142540857,-0.1576250941,0.137997821,0.2178348005,-0.1565312743,0.2232821733,-0.0769478604,0.1563859284,-0.0958470255,-0.1230344325,-0.2482931763,-0.2548377514,0.1001116261,0.2455675602,-0.1645215303,-0.0306790825,-0.102333799,0.0236772168,-0.2311056703,-0.0084627206,-0.1177286133,-0.2933430374,0.3894323707,0.0438978039,0.2783688009,-0.1116899401,0.1263558418,-0.3007574677,0.1491208375,-0.0184958763,-0.075229764,0.2758128941,0.0936535597,-0.4658874869,0.0054211435,-0.4026061594,0.0371483862,0.3698208928,-0.4551106691,0.1599583179,-0.1316279471,0.0954658315,0.4656842053,-0.2714086771,-0.0036663271,0.3496651351,0.2378421873,-0.2974172235,-0.1892595887,0.2399912179,0.0613459609,0.0324590541,0.0272394195,0.0466615967,-0.3037491441,-0.3212345541,0.0892789215,-0.0293396804,-0.2656017244,-0.0114079248,0.3393265009,-0.0023644918,-0.0595147535,-0.050050322,0.2110216618,0.0353926681,0.382693857,-0.2926004231,0.2069978416,0.0592639633,0.0439102203,-0.1111432463,-0.2192169279,0.0503001995,-0.1578414291,0.0338347405,0.2952830493,0.0998148471,0.1530806422,-0.2855040729,-0.1245369539,-0.2080229074,0.2279238254,-0.1790424585,-0.0489066392,-0.2429951131,0.0210169256,-0.0258763712,-0.0946998745,-0.1416703463,-0.038179189,0.3095561862,0.2532219291,-0.17226924,-0.4555348456,0.1082671434,0.3432940543,0.180047825,-0.2458842993,0.336853832,0.1938146502,-0.0687782988,-0.0141235748,0.397110045,0.4357684255,0.338393867,-0.3642207682,0.0811674222,-0.2033814788,-0.0568215884,-0.0232602563,0.2695125937,0.1060727611,0.1173768565,0.4624233842,0.1214860529,-0.1247266233,0.0436382517,-0.0880563408,-0.2605327666,-0.262947619,0.5191478729,0.1857949793,-0.1143881306,-0.3206512332,0.1899377853,-0.2749869525,-0.0809145123,0.0628518164,0.2405894995,0.1131533161,-0.2138276696,0.0705087706,0.0003153924,0.7061742544,0.4877223969,-0.0177194085,-0.3658049405,-0.1165233403,-0.5989871025,0.2178366035,-0.3639689684,-0.2690189183,0.2833368778,-0.0723317787,-0.2880821824,0.2485276014,0.1276198775,0.1581964493,-0.0572769679,0.09834712,-0.4907647073,-0.1486053318,-0.0306509864,-0.0076748356,-0.0831808746,-0.257052511,0.1453228593,-0.4349942505,0.1161028594,0.01198583,0.0693481863,-0.0002400207,0.0966407359,0.3544250131,0.1653244644,0.5012484193,-0.0355607159,0.0036567017,-0.1911553591,-0.1566222012,-0.2289921194,0.1318234056,0.0876291767,0.5133115053,-0.0261075124,-0.0685456768,-0.2640267909,0.4031677544,-0.0810070783,0.0883163437,-0.1350804716,0.0773147568,-0.0703499094,0.2275845408,0.2643646002,0.4245406091,-0.1783216298,0.2022095919,-0.2722912431,-0.4602457583,0.3289350867,-0.6614814401,-0.517790556,0.0751233175,0.1161924526,-0.1423345357,0.0581936017,-0.6006346941,0.0705664605,0.2426124364,-0.1114050075,-0.317119956,0.2179448009,0.0165422,0.0800011158,0.0435740389,0.44835639,-0.148473084,-0.1597679257,-0.0049338331,-0.1632519215]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/168","title":"Loading 'wikitext' dataset fails","comments":"Thanks!\r\n\r\nYes I'm using Google Colab, it seems like a duplicate then.","body":"Loading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'","comment_length":12,"text":"Loading 'wikitext' dataset fails\nLoading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\nThanks!\r\n\r\nYes I'm using Google Colab, it seems like a duplicate then.","embeddings":[-0.1193379238,0.0886276364,-0.0089776339,0.2101976722,0.3692684174,0.0260604266,0.4487809837,0.4131019711,0.4330818355,0.0125665311,0.1238447428,0.3954298794,-0.1505493075,0.0219624937,0.1209058017,-0.4289889932,-0.0337393396,0.3636732399,-0.0414666571,0.062973544,-0.1876838803,0.108516708,-0.2651918828,0.2550059557,-0.3057226241,-0.0570259951,0.2159208357,0.0820631012,-0.0789184123,-0.5590060353,0.2770831883,-0.2170494348,0.0644511506,0.1147467494,-0.0001099558,0.205306232,0.3326975107,0.0302131604,-0.5799459815,-0.1895776838,-0.4773317575,-0.1189394891,0.2974321842,-0.2497071624,-0.028513357,0.0538973026,0.0359112546,-0.0744121596,0.2967469394,0.5315061212,0.2378886342,0.5236920714,0.1545488685,0.0344940759,-0.0785685405,-0.1705711782,-0.0821156427,0.2239678055,0.0738770068,-0.3915446997,-0.0133740734,0.1269243062,0.1511732191,0.2637012899,0.5083716512,0.0143935969,0.240522787,-0.1506265402,0.0149465678,0.0764008239,0.3353828788,-0.1649338454,-0.2716008127,-0.2243597806,0.2476661652,-0.3347570896,0.2365306169,-0.0366766118,-0.1622047126,-0.0081965309,-0.2692284882,-0.0281475354,-0.2418207526,0.3526881039,-0.0582148433,0.4996492267,0.0435834154,0.1427477896,-0.0580658838,-0.075457558,0.1403762847,-0.0870907307,-0.1293002814,0.2578134239,-0.383156091,-0.0364675075,-0.0349461213,-0.1169757396,-0.0316241011,0.2635210156,0.219990015,-0.2248735428,0.3238382339,0.3154148161,0.2229899764,0.4204691052,0.0727646127,0.063592121,-0.0828047171,0.1360252202,-0.2337277681,-0.1554848254,-0.1555357128,-0.1765549034,0.0928974897,-0.1546022743,0.2260205299,-0.071116358,-0.4984739125,0.1147207841,-0.1434045881,0.046290502,0.2051148862,0.1794852465,-0.1276334524,0.2322352529,0.1441244632,0.2825672626,-0.2519766986,-0.0431360416,-0.1662892997,0.2492134273,-0.2343834639,-0.1822493821,0.2427540869,0.1934002936,0.3994777203,-0.1011680439,0.111264661,-0.2017549574,0.2273753434,-0.0455191471,-0.1288790852,0.2566530704,-0.2160614133,0.1548979729,0.2015910596,-0.2824236155,-0.0848238841,-0.10174492,-0.2726355791,-0.3080923855,-0.1262511909,0.2814154327,0.147381857,-0.2429646254,-0.1522630006,-0.0611871071,0.1474927962,-0.3687241077,-0.1181055456,-0.4612588286,-0.2206385881,-0.0230053104,0.3976424038,0.2049094886,0.1313870102,-0.1089720726,-0.1154211313,0.0778549612,-0.0317966752,0.1227136478,-0.4383572936,0.5421624184,-0.0610556342,0.118862763,0.6413468719,-0.3427866101,-0.4578541815,0.123568885,-0.0142798293,0.2093286514,-0.2253688425,-0.0167210288,-0.0132269189,-0.184551537,0.2123140246,0.6192135215,-0.0150432223,0.0824793652,-0.1268968582,-0.0155486753,0.5002205968,0.1098568216,0.0925252214,-0.0079286359,-0.0417051278,0.4966493547,-0.0013129428,-0.1259710789,-0.0148605099,-0.013630108,-0.0646863058,-0.104604587,-0.2178259641,-0.0738868713,-0.2776648998,0.3043035865,-0.117547363,-0.0585312918,-0.0471505858,0.1061868444,-0.3491126597,0.1684893072,-0.4232687652,-0.3773184419,0.1720929891,0.1551331282,-0.0108234957,0.2384415865,-0.2047417462,0.2194369435,-0.274028182,0.0555009618,-0.2189284861,0.1612891257,-0.0558832213,-0.0424223617,-0.0094791884,0.3087970316,0.0383313261,-0.0179966092,-0.3062886596,0.2860310078,-0.1254843473,0.0803563818,-0.0322646983,-0.1433701664,0.0688060746,-0.0715434849,-0.1751839221,0.2081920803,0.3319509625,-0.1801262945,-0.043142952,0.1128761917,0.1272121966,0.0814221054,0.0988641679,0.1639728397,0.1439921707,-0.0658260733,0.1012795344,0.0774310902,0.2337106764,-0.0949872434,0.2047105134,-0.1564514041,-0.167145744,-0.1188064739,0.3131195307,0.1700583696,0.167417407,0.0814866126,-0.275252223,-0.0948746875,0.0785581768,0.1460482925,0.4009645283,-0.0518327467,-0.1083415523,-0.0525707416,-0.4165976346,0.0711256638,0.1707080156,0.0366290882,0.2091285437,0.3018561602,0.0413160138,0.0873196498,-0.2952665389,-0.3570170999,0.2824332714,0.4405301809,-0.2284827828,0.2078800648,-0.0305316318,-0.0025111341,-0.0678482279,-0.3336339295,-0.188629657,-0.3429944813,-0.1213362366,0.4337745905,0.024119148,0.1810040623,-0.3977624178,-0.0220502764,0.1947504133,-0.1147857755,-0.0313790925,-0.3157985508,-0.269328475,0.0459421203,0.3818612099,0.0910761356,0.3510979414,-0.0667702854,-0.0925358012,-0.007033736,-0.0584152527,-0.0755677298,-0.1837216765,0.2440479547,0.009521856,0.2855836153,-0.1770296395,-0.5220327377,0.5017470121,-0.2557020187,-0.0577849969,0.3151076138,0.2711132169,-0.249190703,-0.0794775337,-0.3746882677,-0.1264498532,-0.280605495,-0.0594684482,0.0748294741,0.0620202571,0.3974690437,0.0708564147,0.2160111964,-0.0025732063,0.1134374887,-0.0482623316,-0.0900634304,0.1859846711,-0.310032934,-0.3173182309,-0.1213430315,0.0099359797,0.3347615004,0.1107365489,-0.3489591479,-0.2141319066,-0.1049579829,0.0853126571,-0.2213259488,-0.0400672704,0.1828597784,0.0723028034,-0.0657224879,-0.0866443887,0.035989143,0.0712399781,0.1523462385,0.1349620819,-0.0063045779,0.0892271996,-0.0848842263,0.6008986831,0.218693167,-0.2143773437,0.3002861738,-0.1032764912,0.005145404,-0.2086836398,-0.4398039281,0.2641642988,-0.11523287,0.1829283237,0.1723142117,-0.0794930086,-0.2247841209,-0.192558974,0.0688526183,-0.178447336,-0.2727790177,0.1461688429,-0.2173948586,0.2787112296,-0.0688640028,0.2225363404,-0.2052234709,-0.167981565,0.0698513761,0.166933611,-0.090545617,0.0796746612,-0.2440017164,-0.2761095464,-0.4799583256,-0.0278399177,0.2410321236,0.6437504888,-0.0863839611,0.0371890776,0.0807222873,0.0817412362,0.1373739988,-0.2301357985,0.1822634041,0.2891299725,0.0318248197,-0.3197474778,-0.0643306822,-0.1401119977,0.3580168784,0.0766504705,-0.0544316322,-0.037231572,-0.1676542461,0.1398317814,0.2599402964,-0.1872950047,-0.0441127233,-0.0917341635,-0.0898099318,-0.3077637553,-0.0061545791,-0.0099474099,0.3356595933,0.1229696721,0.2657254636,-0.5462990999,0.1166565344,0.0901664123,0.1745875478,0.0959533602,0.0579445958,0.187246576,-0.1666348726,0.1832098216,-0.145905599,0.5633814335,0.1473762542,-0.3307984769,-0.3195302188,0.0216632597,0.1674342602,0.1839805692,-0.0185859147,-0.1411214918,-0.2769463956,-0.061463993,-0.0087094773,0.3045702279,0.5610407591,0.0026275387,-0.3271814287,-0.2585409284,0.482839793,-0.2850035131,0.1173297018,0.321557641,0.053378012,-0.3131920397,0.0841390863,0.0778320804,0.6767753959,0.1695389003,0.311147064,0.3826704919,0.1839297265,0.4142697752,-0.043089699,0.2847148776,-0.3927352428,-0.1334252506,-0.0806859359,-0.065462403,0.1070814878,-0.0088732634,-0.0254380628,0.1506213844,0.1568897963,-0.004141801,0.057148274,0.4492524862,-0.0387835056,-0.1968196929,-0.3470262885,0.156016469,-0.2179998308,0.297508955,-0.0907334238,-0.0664841011,0.0552040972,-0.0673523247,-0.2555586398,0.2574389279,-0.299474448,0.459457159,0.0578220226,-0.4756816328,0.198415935,0.2070970535,0.1756633669,0.0056988141,-0.2995947301,0.2387785167,-0.2900909483,-0.2218656987,-0.0672868341,0.1119474173,0.4200738966,-0.1725511253,-0.146377489,0.1479405463,-0.1206544265,-0.2415021658,0.1039589867,0.0368667245,0.3524883091,-0.392575711,-0.5152193308,-0.00242018,-0.1820756793,-0.1749101579,0.1744999588,0.023450667,-0.189030841,0.1518952847,0.1498275995,-0.1989797652,0.1164065152,0.2480414361,-0.2495406866,0.1462838054,0.76375705,0.0339458175,0.0128051164,-0.3138304949,-0.0542888604,0.1250851899,-0.4819999039,-0.1171273738,0.1255033612,0.1691254526,-0.1442471892,0.4983651936,0.0630208924,-0.0188453011,0.2471855134,-0.542414248,-0.1062184349,-0.0430215374,0.0177533776,0.2377753258,-0.0026245497,0.114889577,0.1355090588,0.1033881679,-0.272637248,0.0142540857,-0.1576250941,0.137997821,0.2178348005,-0.1565312743,0.2232821733,-0.0769478604,0.1563859284,-0.0958470255,-0.1230344325,-0.2482931763,-0.2548377514,0.1001116261,0.2455675602,-0.1645215303,-0.0306790825,-0.102333799,0.0236772168,-0.2311056703,-0.0084627206,-0.1177286133,-0.2933430374,0.3894323707,0.0438978039,0.2783688009,-0.1116899401,0.1263558418,-0.3007574677,0.1491208375,-0.0184958763,-0.075229764,0.2758128941,0.0936535597,-0.4658874869,0.0054211435,-0.4026061594,0.0371483862,0.3698208928,-0.4551106691,0.1599583179,-0.1316279471,0.0954658315,0.4656842053,-0.2714086771,-0.0036663271,0.3496651351,0.2378421873,-0.2974172235,-0.1892595887,0.2399912179,0.0613459609,0.0324590541,0.0272394195,0.0466615967,-0.3037491441,-0.3212345541,0.0892789215,-0.0293396804,-0.2656017244,-0.0114079248,0.3393265009,-0.0023644918,-0.0595147535,-0.050050322,0.2110216618,0.0353926681,0.382693857,-0.2926004231,0.2069978416,0.0592639633,0.0439102203,-0.1111432463,-0.2192169279,0.0503001995,-0.1578414291,0.0338347405,0.2952830493,0.0998148471,0.1530806422,-0.2855040729,-0.1245369539,-0.2080229074,0.2279238254,-0.1790424585,-0.0489066392,-0.2429951131,0.0210169256,-0.0258763712,-0.0946998745,-0.1416703463,-0.038179189,0.3095561862,0.2532219291,-0.17226924,-0.4555348456,0.1082671434,0.3432940543,0.180047825,-0.2458842993,0.336853832,0.1938146502,-0.0687782988,-0.0141235748,0.397110045,0.4357684255,0.338393867,-0.3642207682,0.0811674222,-0.2033814788,-0.0568215884,-0.0232602563,0.2695125937,0.1060727611,0.1173768565,0.4624233842,0.1214860529,-0.1247266233,0.0436382517,-0.0880563408,-0.2605327666,-0.262947619,0.5191478729,0.1857949793,-0.1143881306,-0.3206512332,0.1899377853,-0.2749869525,-0.0809145123,0.0628518164,0.2405894995,0.1131533161,-0.2138276696,0.0705087706,0.0003153924,0.7061742544,0.4877223969,-0.0177194085,-0.3658049405,-0.1165233403,-0.5989871025,0.2178366035,-0.3639689684,-0.2690189183,0.2833368778,-0.0723317787,-0.2880821824,0.2485276014,0.1276198775,0.1581964493,-0.0572769679,0.09834712,-0.4907647073,-0.1486053318,-0.0306509864,-0.0076748356,-0.0831808746,-0.257052511,0.1453228593,-0.4349942505,0.1161028594,0.01198583,0.0693481863,-0.0002400207,0.0966407359,0.3544250131,0.1653244644,0.5012484193,-0.0355607159,0.0036567017,-0.1911553591,-0.1566222012,-0.2289921194,0.1318234056,0.0876291767,0.5133115053,-0.0261075124,-0.0685456768,-0.2640267909,0.4031677544,-0.0810070783,0.0883163437,-0.1350804716,0.0773147568,-0.0703499094,0.2275845408,0.2643646002,0.4245406091,-0.1783216298,0.2022095919,-0.2722912431,-0.4602457583,0.3289350867,-0.6614814401,-0.517790556,0.0751233175,0.1161924526,-0.1423345357,0.0581936017,-0.6006346941,0.0705664605,0.2426124364,-0.1114050075,-0.317119956,0.2179448009,0.0165422,0.0800011158,0.0435740389,0.44835639,-0.148473084,-0.1597679257,-0.0049338331,-0.1632519215]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/168","title":"Loading 'wikitext' dataset fails","comments":"Closing as it is a duplicate","body":"Loading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'","comment_length":6,"text":"Loading 'wikitext' dataset fails\nLoading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\nClosing as it is a duplicate","embeddings":[-0.1193379238,0.0886276364,-0.0089776339,0.2101976722,0.3692684174,0.0260604266,0.4487809837,0.4131019711,0.4330818355,0.0125665311,0.1238447428,0.3954298794,-0.1505493075,0.0219624937,0.1209058017,-0.4289889932,-0.0337393396,0.3636732399,-0.0414666571,0.062973544,-0.1876838803,0.108516708,-0.2651918828,0.2550059557,-0.3057226241,-0.0570259951,0.2159208357,0.0820631012,-0.0789184123,-0.5590060353,0.2770831883,-0.2170494348,0.0644511506,0.1147467494,-0.0001099558,0.205306232,0.3326975107,0.0302131604,-0.5799459815,-0.1895776838,-0.4773317575,-0.1189394891,0.2974321842,-0.2497071624,-0.028513357,0.0538973026,0.0359112546,-0.0744121596,0.2967469394,0.5315061212,0.2378886342,0.5236920714,0.1545488685,0.0344940759,-0.0785685405,-0.1705711782,-0.0821156427,0.2239678055,0.0738770068,-0.3915446997,-0.0133740734,0.1269243062,0.1511732191,0.2637012899,0.5083716512,0.0143935969,0.240522787,-0.1506265402,0.0149465678,0.0764008239,0.3353828788,-0.1649338454,-0.2716008127,-0.2243597806,0.2476661652,-0.3347570896,0.2365306169,-0.0366766118,-0.1622047126,-0.0081965309,-0.2692284882,-0.0281475354,-0.2418207526,0.3526881039,-0.0582148433,0.4996492267,0.0435834154,0.1427477896,-0.0580658838,-0.075457558,0.1403762847,-0.0870907307,-0.1293002814,0.2578134239,-0.383156091,-0.0364675075,-0.0349461213,-0.1169757396,-0.0316241011,0.2635210156,0.219990015,-0.2248735428,0.3238382339,0.3154148161,0.2229899764,0.4204691052,0.0727646127,0.063592121,-0.0828047171,0.1360252202,-0.2337277681,-0.1554848254,-0.1555357128,-0.1765549034,0.0928974897,-0.1546022743,0.2260205299,-0.071116358,-0.4984739125,0.1147207841,-0.1434045881,0.046290502,0.2051148862,0.1794852465,-0.1276334524,0.2322352529,0.1441244632,0.2825672626,-0.2519766986,-0.0431360416,-0.1662892997,0.2492134273,-0.2343834639,-0.1822493821,0.2427540869,0.1934002936,0.3994777203,-0.1011680439,0.111264661,-0.2017549574,0.2273753434,-0.0455191471,-0.1288790852,0.2566530704,-0.2160614133,0.1548979729,0.2015910596,-0.2824236155,-0.0848238841,-0.10174492,-0.2726355791,-0.3080923855,-0.1262511909,0.2814154327,0.147381857,-0.2429646254,-0.1522630006,-0.0611871071,0.1474927962,-0.3687241077,-0.1181055456,-0.4612588286,-0.2206385881,-0.0230053104,0.3976424038,0.2049094886,0.1313870102,-0.1089720726,-0.1154211313,0.0778549612,-0.0317966752,0.1227136478,-0.4383572936,0.5421624184,-0.0610556342,0.118862763,0.6413468719,-0.3427866101,-0.4578541815,0.123568885,-0.0142798293,0.2093286514,-0.2253688425,-0.0167210288,-0.0132269189,-0.184551537,0.2123140246,0.6192135215,-0.0150432223,0.0824793652,-0.1268968582,-0.0155486753,0.5002205968,0.1098568216,0.0925252214,-0.0079286359,-0.0417051278,0.4966493547,-0.0013129428,-0.1259710789,-0.0148605099,-0.013630108,-0.0646863058,-0.104604587,-0.2178259641,-0.0738868713,-0.2776648998,0.3043035865,-0.117547363,-0.0585312918,-0.0471505858,0.1061868444,-0.3491126597,0.1684893072,-0.4232687652,-0.3773184419,0.1720929891,0.1551331282,-0.0108234957,0.2384415865,-0.2047417462,0.2194369435,-0.274028182,0.0555009618,-0.2189284861,0.1612891257,-0.0558832213,-0.0424223617,-0.0094791884,0.3087970316,0.0383313261,-0.0179966092,-0.3062886596,0.2860310078,-0.1254843473,0.0803563818,-0.0322646983,-0.1433701664,0.0688060746,-0.0715434849,-0.1751839221,0.2081920803,0.3319509625,-0.1801262945,-0.043142952,0.1128761917,0.1272121966,0.0814221054,0.0988641679,0.1639728397,0.1439921707,-0.0658260733,0.1012795344,0.0774310902,0.2337106764,-0.0949872434,0.2047105134,-0.1564514041,-0.167145744,-0.1188064739,0.3131195307,0.1700583696,0.167417407,0.0814866126,-0.275252223,-0.0948746875,0.0785581768,0.1460482925,0.4009645283,-0.0518327467,-0.1083415523,-0.0525707416,-0.4165976346,0.0711256638,0.1707080156,0.0366290882,0.2091285437,0.3018561602,0.0413160138,0.0873196498,-0.2952665389,-0.3570170999,0.2824332714,0.4405301809,-0.2284827828,0.2078800648,-0.0305316318,-0.0025111341,-0.0678482279,-0.3336339295,-0.188629657,-0.3429944813,-0.1213362366,0.4337745905,0.024119148,0.1810040623,-0.3977624178,-0.0220502764,0.1947504133,-0.1147857755,-0.0313790925,-0.3157985508,-0.269328475,0.0459421203,0.3818612099,0.0910761356,0.3510979414,-0.0667702854,-0.0925358012,-0.007033736,-0.0584152527,-0.0755677298,-0.1837216765,0.2440479547,0.009521856,0.2855836153,-0.1770296395,-0.5220327377,0.5017470121,-0.2557020187,-0.0577849969,0.3151076138,0.2711132169,-0.249190703,-0.0794775337,-0.3746882677,-0.1264498532,-0.280605495,-0.0594684482,0.0748294741,0.0620202571,0.3974690437,0.0708564147,0.2160111964,-0.0025732063,0.1134374887,-0.0482623316,-0.0900634304,0.1859846711,-0.310032934,-0.3173182309,-0.1213430315,0.0099359797,0.3347615004,0.1107365489,-0.3489591479,-0.2141319066,-0.1049579829,0.0853126571,-0.2213259488,-0.0400672704,0.1828597784,0.0723028034,-0.0657224879,-0.0866443887,0.035989143,0.0712399781,0.1523462385,0.1349620819,-0.0063045779,0.0892271996,-0.0848842263,0.6008986831,0.218693167,-0.2143773437,0.3002861738,-0.1032764912,0.005145404,-0.2086836398,-0.4398039281,0.2641642988,-0.11523287,0.1829283237,0.1723142117,-0.0794930086,-0.2247841209,-0.192558974,0.0688526183,-0.178447336,-0.2727790177,0.1461688429,-0.2173948586,0.2787112296,-0.0688640028,0.2225363404,-0.2052234709,-0.167981565,0.0698513761,0.166933611,-0.090545617,0.0796746612,-0.2440017164,-0.2761095464,-0.4799583256,-0.0278399177,0.2410321236,0.6437504888,-0.0863839611,0.0371890776,0.0807222873,0.0817412362,0.1373739988,-0.2301357985,0.1822634041,0.2891299725,0.0318248197,-0.3197474778,-0.0643306822,-0.1401119977,0.3580168784,0.0766504705,-0.0544316322,-0.037231572,-0.1676542461,0.1398317814,0.2599402964,-0.1872950047,-0.0441127233,-0.0917341635,-0.0898099318,-0.3077637553,-0.0061545791,-0.0099474099,0.3356595933,0.1229696721,0.2657254636,-0.5462990999,0.1166565344,0.0901664123,0.1745875478,0.0959533602,0.0579445958,0.187246576,-0.1666348726,0.1832098216,-0.145905599,0.5633814335,0.1473762542,-0.3307984769,-0.3195302188,0.0216632597,0.1674342602,0.1839805692,-0.0185859147,-0.1411214918,-0.2769463956,-0.061463993,-0.0087094773,0.3045702279,0.5610407591,0.0026275387,-0.3271814287,-0.2585409284,0.482839793,-0.2850035131,0.1173297018,0.321557641,0.053378012,-0.3131920397,0.0841390863,0.0778320804,0.6767753959,0.1695389003,0.311147064,0.3826704919,0.1839297265,0.4142697752,-0.043089699,0.2847148776,-0.3927352428,-0.1334252506,-0.0806859359,-0.065462403,0.1070814878,-0.0088732634,-0.0254380628,0.1506213844,0.1568897963,-0.004141801,0.057148274,0.4492524862,-0.0387835056,-0.1968196929,-0.3470262885,0.156016469,-0.2179998308,0.297508955,-0.0907334238,-0.0664841011,0.0552040972,-0.0673523247,-0.2555586398,0.2574389279,-0.299474448,0.459457159,0.0578220226,-0.4756816328,0.198415935,0.2070970535,0.1756633669,0.0056988141,-0.2995947301,0.2387785167,-0.2900909483,-0.2218656987,-0.0672868341,0.1119474173,0.4200738966,-0.1725511253,-0.146377489,0.1479405463,-0.1206544265,-0.2415021658,0.1039589867,0.0368667245,0.3524883091,-0.392575711,-0.5152193308,-0.00242018,-0.1820756793,-0.1749101579,0.1744999588,0.023450667,-0.189030841,0.1518952847,0.1498275995,-0.1989797652,0.1164065152,0.2480414361,-0.2495406866,0.1462838054,0.76375705,0.0339458175,0.0128051164,-0.3138304949,-0.0542888604,0.1250851899,-0.4819999039,-0.1171273738,0.1255033612,0.1691254526,-0.1442471892,0.4983651936,0.0630208924,-0.0188453011,0.2471855134,-0.542414248,-0.1062184349,-0.0430215374,0.0177533776,0.2377753258,-0.0026245497,0.114889577,0.1355090588,0.1033881679,-0.272637248,0.0142540857,-0.1576250941,0.137997821,0.2178348005,-0.1565312743,0.2232821733,-0.0769478604,0.1563859284,-0.0958470255,-0.1230344325,-0.2482931763,-0.2548377514,0.1001116261,0.2455675602,-0.1645215303,-0.0306790825,-0.102333799,0.0236772168,-0.2311056703,-0.0084627206,-0.1177286133,-0.2933430374,0.3894323707,0.0438978039,0.2783688009,-0.1116899401,0.1263558418,-0.3007574677,0.1491208375,-0.0184958763,-0.075229764,0.2758128941,0.0936535597,-0.4658874869,0.0054211435,-0.4026061594,0.0371483862,0.3698208928,-0.4551106691,0.1599583179,-0.1316279471,0.0954658315,0.4656842053,-0.2714086771,-0.0036663271,0.3496651351,0.2378421873,-0.2974172235,-0.1892595887,0.2399912179,0.0613459609,0.0324590541,0.0272394195,0.0466615967,-0.3037491441,-0.3212345541,0.0892789215,-0.0293396804,-0.2656017244,-0.0114079248,0.3393265009,-0.0023644918,-0.0595147535,-0.050050322,0.2110216618,0.0353926681,0.382693857,-0.2926004231,0.2069978416,0.0592639633,0.0439102203,-0.1111432463,-0.2192169279,0.0503001995,-0.1578414291,0.0338347405,0.2952830493,0.0998148471,0.1530806422,-0.2855040729,-0.1245369539,-0.2080229074,0.2279238254,-0.1790424585,-0.0489066392,-0.2429951131,0.0210169256,-0.0258763712,-0.0946998745,-0.1416703463,-0.038179189,0.3095561862,0.2532219291,-0.17226924,-0.4555348456,0.1082671434,0.3432940543,0.180047825,-0.2458842993,0.336853832,0.1938146502,-0.0687782988,-0.0141235748,0.397110045,0.4357684255,0.338393867,-0.3642207682,0.0811674222,-0.2033814788,-0.0568215884,-0.0232602563,0.2695125937,0.1060727611,0.1173768565,0.4624233842,0.1214860529,-0.1247266233,0.0436382517,-0.0880563408,-0.2605327666,-0.262947619,0.5191478729,0.1857949793,-0.1143881306,-0.3206512332,0.1899377853,-0.2749869525,-0.0809145123,0.0628518164,0.2405894995,0.1131533161,-0.2138276696,0.0705087706,0.0003153924,0.7061742544,0.4877223969,-0.0177194085,-0.3658049405,-0.1165233403,-0.5989871025,0.2178366035,-0.3639689684,-0.2690189183,0.2833368778,-0.0723317787,-0.2880821824,0.2485276014,0.1276198775,0.1581964493,-0.0572769679,0.09834712,-0.4907647073,-0.1486053318,-0.0306509864,-0.0076748356,-0.0831808746,-0.257052511,0.1453228593,-0.4349942505,0.1161028594,0.01198583,0.0693481863,-0.0002400207,0.0966407359,0.3544250131,0.1653244644,0.5012484193,-0.0355607159,0.0036567017,-0.1911553591,-0.1566222012,-0.2289921194,0.1318234056,0.0876291767,0.5133115053,-0.0261075124,-0.0685456768,-0.2640267909,0.4031677544,-0.0810070783,0.0883163437,-0.1350804716,0.0773147568,-0.0703499094,0.2275845408,0.2643646002,0.4245406091,-0.1783216298,0.2022095919,-0.2722912431,-0.4602457583,0.3289350867,-0.6614814401,-0.517790556,0.0751233175,0.1161924526,-0.1423345357,0.0581936017,-0.6006346941,0.0705664605,0.2426124364,-0.1114050075,-0.317119956,0.2179448009,0.0165422,0.0800011158,0.0435740389,0.44835639,-0.148473084,-0.1597679257,-0.0049338331,-0.1632519215]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/168","title":"Loading 'wikitext' dataset fails","comments":"Hi,\r\nThe squad bug seems to be fixed, but the loading of the 'wikitext' still suffers from this problem (on Colab with pyarrow=0.17.1).","body":"Loading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'","comment_length":23,"text":"Loading 'wikitext' dataset fails\nLoading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\nHi,\r\nThe squad bug seems to be fixed, but the loading of the 'wikitext' still suffers from this problem (on Colab with pyarrow=0.17.1).","embeddings":[-0.1193379238,0.0886276364,-0.0089776339,0.2101976722,0.3692684174,0.0260604266,0.4487809837,0.4131019711,0.4330818355,0.0125665311,0.1238447428,0.3954298794,-0.1505493075,0.0219624937,0.1209058017,-0.4289889932,-0.0337393396,0.3636732399,-0.0414666571,0.062973544,-0.1876838803,0.108516708,-0.2651918828,0.2550059557,-0.3057226241,-0.0570259951,0.2159208357,0.0820631012,-0.0789184123,-0.5590060353,0.2770831883,-0.2170494348,0.0644511506,0.1147467494,-0.0001099558,0.205306232,0.3326975107,0.0302131604,-0.5799459815,-0.1895776838,-0.4773317575,-0.1189394891,0.2974321842,-0.2497071624,-0.028513357,0.0538973026,0.0359112546,-0.0744121596,0.2967469394,0.5315061212,0.2378886342,0.5236920714,0.1545488685,0.0344940759,-0.0785685405,-0.1705711782,-0.0821156427,0.2239678055,0.0738770068,-0.3915446997,-0.0133740734,0.1269243062,0.1511732191,0.2637012899,0.5083716512,0.0143935969,0.240522787,-0.1506265402,0.0149465678,0.0764008239,0.3353828788,-0.1649338454,-0.2716008127,-0.2243597806,0.2476661652,-0.3347570896,0.2365306169,-0.0366766118,-0.1622047126,-0.0081965309,-0.2692284882,-0.0281475354,-0.2418207526,0.3526881039,-0.0582148433,0.4996492267,0.0435834154,0.1427477896,-0.0580658838,-0.075457558,0.1403762847,-0.0870907307,-0.1293002814,0.2578134239,-0.383156091,-0.0364675075,-0.0349461213,-0.1169757396,-0.0316241011,0.2635210156,0.219990015,-0.2248735428,0.3238382339,0.3154148161,0.2229899764,0.4204691052,0.0727646127,0.063592121,-0.0828047171,0.1360252202,-0.2337277681,-0.1554848254,-0.1555357128,-0.1765549034,0.0928974897,-0.1546022743,0.2260205299,-0.071116358,-0.4984739125,0.1147207841,-0.1434045881,0.046290502,0.2051148862,0.1794852465,-0.1276334524,0.2322352529,0.1441244632,0.2825672626,-0.2519766986,-0.0431360416,-0.1662892997,0.2492134273,-0.2343834639,-0.1822493821,0.2427540869,0.1934002936,0.3994777203,-0.1011680439,0.111264661,-0.2017549574,0.2273753434,-0.0455191471,-0.1288790852,0.2566530704,-0.2160614133,0.1548979729,0.2015910596,-0.2824236155,-0.0848238841,-0.10174492,-0.2726355791,-0.3080923855,-0.1262511909,0.2814154327,0.147381857,-0.2429646254,-0.1522630006,-0.0611871071,0.1474927962,-0.3687241077,-0.1181055456,-0.4612588286,-0.2206385881,-0.0230053104,0.3976424038,0.2049094886,0.1313870102,-0.1089720726,-0.1154211313,0.0778549612,-0.0317966752,0.1227136478,-0.4383572936,0.5421624184,-0.0610556342,0.118862763,0.6413468719,-0.3427866101,-0.4578541815,0.123568885,-0.0142798293,0.2093286514,-0.2253688425,-0.0167210288,-0.0132269189,-0.184551537,0.2123140246,0.6192135215,-0.0150432223,0.0824793652,-0.1268968582,-0.0155486753,0.5002205968,0.1098568216,0.0925252214,-0.0079286359,-0.0417051278,0.4966493547,-0.0013129428,-0.1259710789,-0.0148605099,-0.013630108,-0.0646863058,-0.104604587,-0.2178259641,-0.0738868713,-0.2776648998,0.3043035865,-0.117547363,-0.0585312918,-0.0471505858,0.1061868444,-0.3491126597,0.1684893072,-0.4232687652,-0.3773184419,0.1720929891,0.1551331282,-0.0108234957,0.2384415865,-0.2047417462,0.2194369435,-0.274028182,0.0555009618,-0.2189284861,0.1612891257,-0.0558832213,-0.0424223617,-0.0094791884,0.3087970316,0.0383313261,-0.0179966092,-0.3062886596,0.2860310078,-0.1254843473,0.0803563818,-0.0322646983,-0.1433701664,0.0688060746,-0.0715434849,-0.1751839221,0.2081920803,0.3319509625,-0.1801262945,-0.043142952,0.1128761917,0.1272121966,0.0814221054,0.0988641679,0.1639728397,0.1439921707,-0.0658260733,0.1012795344,0.0774310902,0.2337106764,-0.0949872434,0.2047105134,-0.1564514041,-0.167145744,-0.1188064739,0.3131195307,0.1700583696,0.167417407,0.0814866126,-0.275252223,-0.0948746875,0.0785581768,0.1460482925,0.4009645283,-0.0518327467,-0.1083415523,-0.0525707416,-0.4165976346,0.0711256638,0.1707080156,0.0366290882,0.2091285437,0.3018561602,0.0413160138,0.0873196498,-0.2952665389,-0.3570170999,0.2824332714,0.4405301809,-0.2284827828,0.2078800648,-0.0305316318,-0.0025111341,-0.0678482279,-0.3336339295,-0.188629657,-0.3429944813,-0.1213362366,0.4337745905,0.024119148,0.1810040623,-0.3977624178,-0.0220502764,0.1947504133,-0.1147857755,-0.0313790925,-0.3157985508,-0.269328475,0.0459421203,0.3818612099,0.0910761356,0.3510979414,-0.0667702854,-0.0925358012,-0.007033736,-0.0584152527,-0.0755677298,-0.1837216765,0.2440479547,0.009521856,0.2855836153,-0.1770296395,-0.5220327377,0.5017470121,-0.2557020187,-0.0577849969,0.3151076138,0.2711132169,-0.249190703,-0.0794775337,-0.3746882677,-0.1264498532,-0.280605495,-0.0594684482,0.0748294741,0.0620202571,0.3974690437,0.0708564147,0.2160111964,-0.0025732063,0.1134374887,-0.0482623316,-0.0900634304,0.1859846711,-0.310032934,-0.3173182309,-0.1213430315,0.0099359797,0.3347615004,0.1107365489,-0.3489591479,-0.2141319066,-0.1049579829,0.0853126571,-0.2213259488,-0.0400672704,0.1828597784,0.0723028034,-0.0657224879,-0.0866443887,0.035989143,0.0712399781,0.1523462385,0.1349620819,-0.0063045779,0.0892271996,-0.0848842263,0.6008986831,0.218693167,-0.2143773437,0.3002861738,-0.1032764912,0.005145404,-0.2086836398,-0.4398039281,0.2641642988,-0.11523287,0.1829283237,0.1723142117,-0.0794930086,-0.2247841209,-0.192558974,0.0688526183,-0.178447336,-0.2727790177,0.1461688429,-0.2173948586,0.2787112296,-0.0688640028,0.2225363404,-0.2052234709,-0.167981565,0.0698513761,0.166933611,-0.090545617,0.0796746612,-0.2440017164,-0.2761095464,-0.4799583256,-0.0278399177,0.2410321236,0.6437504888,-0.0863839611,0.0371890776,0.0807222873,0.0817412362,0.1373739988,-0.2301357985,0.1822634041,0.2891299725,0.0318248197,-0.3197474778,-0.0643306822,-0.1401119977,0.3580168784,0.0766504705,-0.0544316322,-0.037231572,-0.1676542461,0.1398317814,0.2599402964,-0.1872950047,-0.0441127233,-0.0917341635,-0.0898099318,-0.3077637553,-0.0061545791,-0.0099474099,0.3356595933,0.1229696721,0.2657254636,-0.5462990999,0.1166565344,0.0901664123,0.1745875478,0.0959533602,0.0579445958,0.187246576,-0.1666348726,0.1832098216,-0.145905599,0.5633814335,0.1473762542,-0.3307984769,-0.3195302188,0.0216632597,0.1674342602,0.1839805692,-0.0185859147,-0.1411214918,-0.2769463956,-0.061463993,-0.0087094773,0.3045702279,0.5610407591,0.0026275387,-0.3271814287,-0.2585409284,0.482839793,-0.2850035131,0.1173297018,0.321557641,0.053378012,-0.3131920397,0.0841390863,0.0778320804,0.6767753959,0.1695389003,0.311147064,0.3826704919,0.1839297265,0.4142697752,-0.043089699,0.2847148776,-0.3927352428,-0.1334252506,-0.0806859359,-0.065462403,0.1070814878,-0.0088732634,-0.0254380628,0.1506213844,0.1568897963,-0.004141801,0.057148274,0.4492524862,-0.0387835056,-0.1968196929,-0.3470262885,0.156016469,-0.2179998308,0.297508955,-0.0907334238,-0.0664841011,0.0552040972,-0.0673523247,-0.2555586398,0.2574389279,-0.299474448,0.459457159,0.0578220226,-0.4756816328,0.198415935,0.2070970535,0.1756633669,0.0056988141,-0.2995947301,0.2387785167,-0.2900909483,-0.2218656987,-0.0672868341,0.1119474173,0.4200738966,-0.1725511253,-0.146377489,0.1479405463,-0.1206544265,-0.2415021658,0.1039589867,0.0368667245,0.3524883091,-0.392575711,-0.5152193308,-0.00242018,-0.1820756793,-0.1749101579,0.1744999588,0.023450667,-0.189030841,0.1518952847,0.1498275995,-0.1989797652,0.1164065152,0.2480414361,-0.2495406866,0.1462838054,0.76375705,0.0339458175,0.0128051164,-0.3138304949,-0.0542888604,0.1250851899,-0.4819999039,-0.1171273738,0.1255033612,0.1691254526,-0.1442471892,0.4983651936,0.0630208924,-0.0188453011,0.2471855134,-0.542414248,-0.1062184349,-0.0430215374,0.0177533776,0.2377753258,-0.0026245497,0.114889577,0.1355090588,0.1033881679,-0.272637248,0.0142540857,-0.1576250941,0.137997821,0.2178348005,-0.1565312743,0.2232821733,-0.0769478604,0.1563859284,-0.0958470255,-0.1230344325,-0.2482931763,-0.2548377514,0.1001116261,0.2455675602,-0.1645215303,-0.0306790825,-0.102333799,0.0236772168,-0.2311056703,-0.0084627206,-0.1177286133,-0.2933430374,0.3894323707,0.0438978039,0.2783688009,-0.1116899401,0.1263558418,-0.3007574677,0.1491208375,-0.0184958763,-0.075229764,0.2758128941,0.0936535597,-0.4658874869,0.0054211435,-0.4026061594,0.0371483862,0.3698208928,-0.4551106691,0.1599583179,-0.1316279471,0.0954658315,0.4656842053,-0.2714086771,-0.0036663271,0.3496651351,0.2378421873,-0.2974172235,-0.1892595887,0.2399912179,0.0613459609,0.0324590541,0.0272394195,0.0466615967,-0.3037491441,-0.3212345541,0.0892789215,-0.0293396804,-0.2656017244,-0.0114079248,0.3393265009,-0.0023644918,-0.0595147535,-0.050050322,0.2110216618,0.0353926681,0.382693857,-0.2926004231,0.2069978416,0.0592639633,0.0439102203,-0.1111432463,-0.2192169279,0.0503001995,-0.1578414291,0.0338347405,0.2952830493,0.0998148471,0.1530806422,-0.2855040729,-0.1245369539,-0.2080229074,0.2279238254,-0.1790424585,-0.0489066392,-0.2429951131,0.0210169256,-0.0258763712,-0.0946998745,-0.1416703463,-0.038179189,0.3095561862,0.2532219291,-0.17226924,-0.4555348456,0.1082671434,0.3432940543,0.180047825,-0.2458842993,0.336853832,0.1938146502,-0.0687782988,-0.0141235748,0.397110045,0.4357684255,0.338393867,-0.3642207682,0.0811674222,-0.2033814788,-0.0568215884,-0.0232602563,0.2695125937,0.1060727611,0.1173768565,0.4624233842,0.1214860529,-0.1247266233,0.0436382517,-0.0880563408,-0.2605327666,-0.262947619,0.5191478729,0.1857949793,-0.1143881306,-0.3206512332,0.1899377853,-0.2749869525,-0.0809145123,0.0628518164,0.2405894995,0.1131533161,-0.2138276696,0.0705087706,0.0003153924,0.7061742544,0.4877223969,-0.0177194085,-0.3658049405,-0.1165233403,-0.5989871025,0.2178366035,-0.3639689684,-0.2690189183,0.2833368778,-0.0723317787,-0.2880821824,0.2485276014,0.1276198775,0.1581964493,-0.0572769679,0.09834712,-0.4907647073,-0.1486053318,-0.0306509864,-0.0076748356,-0.0831808746,-0.257052511,0.1453228593,-0.4349942505,0.1161028594,0.01198583,0.0693481863,-0.0002400207,0.0966407359,0.3544250131,0.1653244644,0.5012484193,-0.0355607159,0.0036567017,-0.1911553591,-0.1566222012,-0.2289921194,0.1318234056,0.0876291767,0.5133115053,-0.0261075124,-0.0685456768,-0.2640267909,0.4031677544,-0.0810070783,0.0883163437,-0.1350804716,0.0773147568,-0.0703499094,0.2275845408,0.2643646002,0.4245406091,-0.1783216298,0.2022095919,-0.2722912431,-0.4602457583,0.3289350867,-0.6614814401,-0.517790556,0.0751233175,0.1161924526,-0.1423345357,0.0581936017,-0.6006346941,0.0705664605,0.2426124364,-0.1114050075,-0.317119956,0.2179448009,0.0165422,0.0800011158,0.0435740389,0.44835639,-0.148473084,-0.1597679257,-0.0049338331,-0.1632519215]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/168","title":"Loading 'wikitext' dataset fails","comments":"When you install `nlp` for the first time on a Colab runtime, it updates the `pyarrow` library that was already on colab. This update shows this message on colab:\r\n```\r\nWARNING: The following packages were previously imported in this runtime:\r\n [pyarrow]\r\nYou must restart the runtime in order to use newly installed versions.\r\n```\r\nYou just have to restart the runtime and it should be fine.","body":"Loading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'","comment_length":66,"text":"Loading 'wikitext' dataset fails\nLoading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\nWhen you install `nlp` for the first time on a Colab runtime, it updates the `pyarrow` library that was already on colab. This update shows this message on colab:\r\n```\r\nWARNING: The following packages were previously imported in this runtime:\r\n [pyarrow]\r\nYou must restart the runtime in order to use newly installed versions.\r\n```\r\nYou just have to restart the runtime and it should be fine.","embeddings":[-0.1193379238,0.0886276364,-0.0089776339,0.2101976722,0.3692684174,0.0260604266,0.4487809837,0.4131019711,0.4330818355,0.0125665311,0.1238447428,0.3954298794,-0.1505493075,0.0219624937,0.1209058017,-0.4289889932,-0.0337393396,0.3636732399,-0.0414666571,0.062973544,-0.1876838803,0.108516708,-0.2651918828,0.2550059557,-0.3057226241,-0.0570259951,0.2159208357,0.0820631012,-0.0789184123,-0.5590060353,0.2770831883,-0.2170494348,0.0644511506,0.1147467494,-0.0001099558,0.205306232,0.3326975107,0.0302131604,-0.5799459815,-0.1895776838,-0.4773317575,-0.1189394891,0.2974321842,-0.2497071624,-0.028513357,0.0538973026,0.0359112546,-0.0744121596,0.2967469394,0.5315061212,0.2378886342,0.5236920714,0.1545488685,0.0344940759,-0.0785685405,-0.1705711782,-0.0821156427,0.2239678055,0.0738770068,-0.3915446997,-0.0133740734,0.1269243062,0.1511732191,0.2637012899,0.5083716512,0.0143935969,0.240522787,-0.1506265402,0.0149465678,0.0764008239,0.3353828788,-0.1649338454,-0.2716008127,-0.2243597806,0.2476661652,-0.3347570896,0.2365306169,-0.0366766118,-0.1622047126,-0.0081965309,-0.2692284882,-0.0281475354,-0.2418207526,0.3526881039,-0.0582148433,0.4996492267,0.0435834154,0.1427477896,-0.0580658838,-0.075457558,0.1403762847,-0.0870907307,-0.1293002814,0.2578134239,-0.383156091,-0.0364675075,-0.0349461213,-0.1169757396,-0.0316241011,0.2635210156,0.219990015,-0.2248735428,0.3238382339,0.3154148161,0.2229899764,0.4204691052,0.0727646127,0.063592121,-0.0828047171,0.1360252202,-0.2337277681,-0.1554848254,-0.1555357128,-0.1765549034,0.0928974897,-0.1546022743,0.2260205299,-0.071116358,-0.4984739125,0.1147207841,-0.1434045881,0.046290502,0.2051148862,0.1794852465,-0.1276334524,0.2322352529,0.1441244632,0.2825672626,-0.2519766986,-0.0431360416,-0.1662892997,0.2492134273,-0.2343834639,-0.1822493821,0.2427540869,0.1934002936,0.3994777203,-0.1011680439,0.111264661,-0.2017549574,0.2273753434,-0.0455191471,-0.1288790852,0.2566530704,-0.2160614133,0.1548979729,0.2015910596,-0.2824236155,-0.0848238841,-0.10174492,-0.2726355791,-0.3080923855,-0.1262511909,0.2814154327,0.147381857,-0.2429646254,-0.1522630006,-0.0611871071,0.1474927962,-0.3687241077,-0.1181055456,-0.4612588286,-0.2206385881,-0.0230053104,0.3976424038,0.2049094886,0.1313870102,-0.1089720726,-0.1154211313,0.0778549612,-0.0317966752,0.1227136478,-0.4383572936,0.5421624184,-0.0610556342,0.118862763,0.6413468719,-0.3427866101,-0.4578541815,0.123568885,-0.0142798293,0.2093286514,-0.2253688425,-0.0167210288,-0.0132269189,-0.184551537,0.2123140246,0.6192135215,-0.0150432223,0.0824793652,-0.1268968582,-0.0155486753,0.5002205968,0.1098568216,0.0925252214,-0.0079286359,-0.0417051278,0.4966493547,-0.0013129428,-0.1259710789,-0.0148605099,-0.013630108,-0.0646863058,-0.104604587,-0.2178259641,-0.0738868713,-0.2776648998,0.3043035865,-0.117547363,-0.0585312918,-0.0471505858,0.1061868444,-0.3491126597,0.1684893072,-0.4232687652,-0.3773184419,0.1720929891,0.1551331282,-0.0108234957,0.2384415865,-0.2047417462,0.2194369435,-0.274028182,0.0555009618,-0.2189284861,0.1612891257,-0.0558832213,-0.0424223617,-0.0094791884,0.3087970316,0.0383313261,-0.0179966092,-0.3062886596,0.2860310078,-0.1254843473,0.0803563818,-0.0322646983,-0.1433701664,0.0688060746,-0.0715434849,-0.1751839221,0.2081920803,0.3319509625,-0.1801262945,-0.043142952,0.1128761917,0.1272121966,0.0814221054,0.0988641679,0.1639728397,0.1439921707,-0.0658260733,0.1012795344,0.0774310902,0.2337106764,-0.0949872434,0.2047105134,-0.1564514041,-0.167145744,-0.1188064739,0.3131195307,0.1700583696,0.167417407,0.0814866126,-0.275252223,-0.0948746875,0.0785581768,0.1460482925,0.4009645283,-0.0518327467,-0.1083415523,-0.0525707416,-0.4165976346,0.0711256638,0.1707080156,0.0366290882,0.2091285437,0.3018561602,0.0413160138,0.0873196498,-0.2952665389,-0.3570170999,0.2824332714,0.4405301809,-0.2284827828,0.2078800648,-0.0305316318,-0.0025111341,-0.0678482279,-0.3336339295,-0.188629657,-0.3429944813,-0.1213362366,0.4337745905,0.024119148,0.1810040623,-0.3977624178,-0.0220502764,0.1947504133,-0.1147857755,-0.0313790925,-0.3157985508,-0.269328475,0.0459421203,0.3818612099,0.0910761356,0.3510979414,-0.0667702854,-0.0925358012,-0.007033736,-0.0584152527,-0.0755677298,-0.1837216765,0.2440479547,0.009521856,0.2855836153,-0.1770296395,-0.5220327377,0.5017470121,-0.2557020187,-0.0577849969,0.3151076138,0.2711132169,-0.249190703,-0.0794775337,-0.3746882677,-0.1264498532,-0.280605495,-0.0594684482,0.0748294741,0.0620202571,0.3974690437,0.0708564147,0.2160111964,-0.0025732063,0.1134374887,-0.0482623316,-0.0900634304,0.1859846711,-0.310032934,-0.3173182309,-0.1213430315,0.0099359797,0.3347615004,0.1107365489,-0.3489591479,-0.2141319066,-0.1049579829,0.0853126571,-0.2213259488,-0.0400672704,0.1828597784,0.0723028034,-0.0657224879,-0.0866443887,0.035989143,0.0712399781,0.1523462385,0.1349620819,-0.0063045779,0.0892271996,-0.0848842263,0.6008986831,0.218693167,-0.2143773437,0.3002861738,-0.1032764912,0.005145404,-0.2086836398,-0.4398039281,0.2641642988,-0.11523287,0.1829283237,0.1723142117,-0.0794930086,-0.2247841209,-0.192558974,0.0688526183,-0.178447336,-0.2727790177,0.1461688429,-0.2173948586,0.2787112296,-0.0688640028,0.2225363404,-0.2052234709,-0.167981565,0.0698513761,0.166933611,-0.090545617,0.0796746612,-0.2440017164,-0.2761095464,-0.4799583256,-0.0278399177,0.2410321236,0.6437504888,-0.0863839611,0.0371890776,0.0807222873,0.0817412362,0.1373739988,-0.2301357985,0.1822634041,0.2891299725,0.0318248197,-0.3197474778,-0.0643306822,-0.1401119977,0.3580168784,0.0766504705,-0.0544316322,-0.037231572,-0.1676542461,0.1398317814,0.2599402964,-0.1872950047,-0.0441127233,-0.0917341635,-0.0898099318,-0.3077637553,-0.0061545791,-0.0099474099,0.3356595933,0.1229696721,0.2657254636,-0.5462990999,0.1166565344,0.0901664123,0.1745875478,0.0959533602,0.0579445958,0.187246576,-0.1666348726,0.1832098216,-0.145905599,0.5633814335,0.1473762542,-0.3307984769,-0.3195302188,0.0216632597,0.1674342602,0.1839805692,-0.0185859147,-0.1411214918,-0.2769463956,-0.061463993,-0.0087094773,0.3045702279,0.5610407591,0.0026275387,-0.3271814287,-0.2585409284,0.482839793,-0.2850035131,0.1173297018,0.321557641,0.053378012,-0.3131920397,0.0841390863,0.0778320804,0.6767753959,0.1695389003,0.311147064,0.3826704919,0.1839297265,0.4142697752,-0.043089699,0.2847148776,-0.3927352428,-0.1334252506,-0.0806859359,-0.065462403,0.1070814878,-0.0088732634,-0.0254380628,0.1506213844,0.1568897963,-0.004141801,0.057148274,0.4492524862,-0.0387835056,-0.1968196929,-0.3470262885,0.156016469,-0.2179998308,0.297508955,-0.0907334238,-0.0664841011,0.0552040972,-0.0673523247,-0.2555586398,0.2574389279,-0.299474448,0.459457159,0.0578220226,-0.4756816328,0.198415935,0.2070970535,0.1756633669,0.0056988141,-0.2995947301,0.2387785167,-0.2900909483,-0.2218656987,-0.0672868341,0.1119474173,0.4200738966,-0.1725511253,-0.146377489,0.1479405463,-0.1206544265,-0.2415021658,0.1039589867,0.0368667245,0.3524883091,-0.392575711,-0.5152193308,-0.00242018,-0.1820756793,-0.1749101579,0.1744999588,0.023450667,-0.189030841,0.1518952847,0.1498275995,-0.1989797652,0.1164065152,0.2480414361,-0.2495406866,0.1462838054,0.76375705,0.0339458175,0.0128051164,-0.3138304949,-0.0542888604,0.1250851899,-0.4819999039,-0.1171273738,0.1255033612,0.1691254526,-0.1442471892,0.4983651936,0.0630208924,-0.0188453011,0.2471855134,-0.542414248,-0.1062184349,-0.0430215374,0.0177533776,0.2377753258,-0.0026245497,0.114889577,0.1355090588,0.1033881679,-0.272637248,0.0142540857,-0.1576250941,0.137997821,0.2178348005,-0.1565312743,0.2232821733,-0.0769478604,0.1563859284,-0.0958470255,-0.1230344325,-0.2482931763,-0.2548377514,0.1001116261,0.2455675602,-0.1645215303,-0.0306790825,-0.102333799,0.0236772168,-0.2311056703,-0.0084627206,-0.1177286133,-0.2933430374,0.3894323707,0.0438978039,0.2783688009,-0.1116899401,0.1263558418,-0.3007574677,0.1491208375,-0.0184958763,-0.075229764,0.2758128941,0.0936535597,-0.4658874869,0.0054211435,-0.4026061594,0.0371483862,0.3698208928,-0.4551106691,0.1599583179,-0.1316279471,0.0954658315,0.4656842053,-0.2714086771,-0.0036663271,0.3496651351,0.2378421873,-0.2974172235,-0.1892595887,0.2399912179,0.0613459609,0.0324590541,0.0272394195,0.0466615967,-0.3037491441,-0.3212345541,0.0892789215,-0.0293396804,-0.2656017244,-0.0114079248,0.3393265009,-0.0023644918,-0.0595147535,-0.050050322,0.2110216618,0.0353926681,0.382693857,-0.2926004231,0.2069978416,0.0592639633,0.0439102203,-0.1111432463,-0.2192169279,0.0503001995,-0.1578414291,0.0338347405,0.2952830493,0.0998148471,0.1530806422,-0.2855040729,-0.1245369539,-0.2080229074,0.2279238254,-0.1790424585,-0.0489066392,-0.2429951131,0.0210169256,-0.0258763712,-0.0946998745,-0.1416703463,-0.038179189,0.3095561862,0.2532219291,-0.17226924,-0.4555348456,0.1082671434,0.3432940543,0.180047825,-0.2458842993,0.336853832,0.1938146502,-0.0687782988,-0.0141235748,0.397110045,0.4357684255,0.338393867,-0.3642207682,0.0811674222,-0.2033814788,-0.0568215884,-0.0232602563,0.2695125937,0.1060727611,0.1173768565,0.4624233842,0.1214860529,-0.1247266233,0.0436382517,-0.0880563408,-0.2605327666,-0.262947619,0.5191478729,0.1857949793,-0.1143881306,-0.3206512332,0.1899377853,-0.2749869525,-0.0809145123,0.0628518164,0.2405894995,0.1131533161,-0.2138276696,0.0705087706,0.0003153924,0.7061742544,0.4877223969,-0.0177194085,-0.3658049405,-0.1165233403,-0.5989871025,0.2178366035,-0.3639689684,-0.2690189183,0.2833368778,-0.0723317787,-0.2880821824,0.2485276014,0.1276198775,0.1581964493,-0.0572769679,0.09834712,-0.4907647073,-0.1486053318,-0.0306509864,-0.0076748356,-0.0831808746,-0.257052511,0.1453228593,-0.4349942505,0.1161028594,0.01198583,0.0693481863,-0.0002400207,0.0966407359,0.3544250131,0.1653244644,0.5012484193,-0.0355607159,0.0036567017,-0.1911553591,-0.1566222012,-0.2289921194,0.1318234056,0.0876291767,0.5133115053,-0.0261075124,-0.0685456768,-0.2640267909,0.4031677544,-0.0810070783,0.0883163437,-0.1350804716,0.0773147568,-0.0703499094,0.2275845408,0.2643646002,0.4245406091,-0.1783216298,0.2022095919,-0.2722912431,-0.4602457583,0.3289350867,-0.6614814401,-0.517790556,0.0751233175,0.1161924526,-0.1423345357,0.0581936017,-0.6006346941,0.0705664605,0.2426124364,-0.1114050075,-0.317119956,0.2179448009,0.0165422,0.0800011158,0.0435740389,0.44835639,-0.148473084,-0.1597679257,-0.0049338331,-0.1632519215]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/168","title":"Loading 'wikitext' dataset fails","comments":"That was it, thanks!","body":"Loading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'","comment_length":4,"text":"Loading 'wikitext' dataset fails\nLoading the 'wikitext' dataset fails with Attribute error:\r\n\r\nCode to reproduce (From example notebook):\r\n\r\nimport nlp\r\nwikitext_dataset = nlp.load_dataset('wikitext')\r\n\r\n\r\nError:\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-17-d5d9df94b13c> in <module>()\r\n 11 \r\n 12 # Load a dataset and print the first examples in the training set\r\n---> 13 wikitext_dataset = nlp.load_dataset('wikitext')\r\n 14 print(wikitext_dataset['train'][0])\r\n\r\n6 frames\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 518 download_mode=download_mode,\r\n 519 ignore_verifications=ignore_verifications,\r\n--> 520 save_infos=save_infos,\r\n 521 )\r\n 522 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 363 verify_infos = not save_infos and not ignore_verifications\r\n 364 self._download_and_prepare(\r\n--> 365 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 366 )\r\n 367 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 416 try:\r\n 417 # Prepare split will record examples associated to the split\r\n--> 418 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 419 except OSError:\r\n 420 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 594 example = self.info.features.encode_example(record)\r\n 595 writer.write(example)\r\n--> 596 num_examples, num_bytes = writer.finalize()\r\n 597 \r\n 598 assert num_examples == num_examples, f\"Expected to write {split_info.num_examples} but wrote {num_examples}\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in finalize(self, close_stream)\r\n 173 def finalize(self, close_stream=True):\r\n 174 if self.pa_writer is not None:\r\n--> 175 self.write_on_file()\r\n 176 self.pa_writer.close()\r\n 177 if close_stream:\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in write_on_file(self)\r\n 124 else:\r\n 125 # All good\r\n--> 126 self._write_array_on_file(pa_array)\r\n 127 self.current_rows = []\r\n 128 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/arrow_writer.py in _write_array_on_file(self, pa_array)\r\n 93 def _write_array_on_file(self, pa_array):\r\n 94 \"\"\"Write a PyArrow Array\"\"\"\r\n---> 95 pa_batch = pa.RecordBatch.from_struct_array(pa_array)\r\n 96 self._num_bytes += pa_array.nbytes\r\n 97 self.pa_writer.write_batch(pa_batch)\r\n\r\nAttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\nThat was it, thanks!","embeddings":[-0.1193379238,0.0886276364,-0.0089776339,0.2101976722,0.3692684174,0.0260604266,0.4487809837,0.4131019711,0.4330818355,0.0125665311,0.1238447428,0.3954298794,-0.1505493075,0.0219624937,0.1209058017,-0.4289889932,-0.0337393396,0.3636732399,-0.0414666571,0.062973544,-0.1876838803,0.108516708,-0.2651918828,0.2550059557,-0.3057226241,-0.0570259951,0.2159208357,0.0820631012,-0.0789184123,-0.5590060353,0.2770831883,-0.2170494348,0.0644511506,0.1147467494,-0.0001099558,0.205306232,0.3326975107,0.0302131604,-0.5799459815,-0.1895776838,-0.4773317575,-0.1189394891,0.2974321842,-0.2497071624,-0.028513357,0.0538973026,0.0359112546,-0.0744121596,0.2967469394,0.5315061212,0.2378886342,0.5236920714,0.1545488685,0.0344940759,-0.0785685405,-0.1705711782,-0.0821156427,0.2239678055,0.0738770068,-0.3915446997,-0.0133740734,0.1269243062,0.1511732191,0.2637012899,0.5083716512,0.0143935969,0.240522787,-0.1506265402,0.0149465678,0.0764008239,0.3353828788,-0.1649338454,-0.2716008127,-0.2243597806,0.2476661652,-0.3347570896,0.2365306169,-0.0366766118,-0.1622047126,-0.0081965309,-0.2692284882,-0.0281475354,-0.2418207526,0.3526881039,-0.0582148433,0.4996492267,0.0435834154,0.1427477896,-0.0580658838,-0.075457558,0.1403762847,-0.0870907307,-0.1293002814,0.2578134239,-0.383156091,-0.0364675075,-0.0349461213,-0.1169757396,-0.0316241011,0.2635210156,0.219990015,-0.2248735428,0.3238382339,0.3154148161,0.2229899764,0.4204691052,0.0727646127,0.063592121,-0.0828047171,0.1360252202,-0.2337277681,-0.1554848254,-0.1555357128,-0.1765549034,0.0928974897,-0.1546022743,0.2260205299,-0.071116358,-0.4984739125,0.1147207841,-0.1434045881,0.046290502,0.2051148862,0.1794852465,-0.1276334524,0.2322352529,0.1441244632,0.2825672626,-0.2519766986,-0.0431360416,-0.1662892997,0.2492134273,-0.2343834639,-0.1822493821,0.2427540869,0.1934002936,0.3994777203,-0.1011680439,0.111264661,-0.2017549574,0.2273753434,-0.0455191471,-0.1288790852,0.2566530704,-0.2160614133,0.1548979729,0.2015910596,-0.2824236155,-0.0848238841,-0.10174492,-0.2726355791,-0.3080923855,-0.1262511909,0.2814154327,0.147381857,-0.2429646254,-0.1522630006,-0.0611871071,0.1474927962,-0.3687241077,-0.1181055456,-0.4612588286,-0.2206385881,-0.0230053104,0.3976424038,0.2049094886,0.1313870102,-0.1089720726,-0.1154211313,0.0778549612,-0.0317966752,0.1227136478,-0.4383572936,0.5421624184,-0.0610556342,0.118862763,0.6413468719,-0.3427866101,-0.4578541815,0.123568885,-0.0142798293,0.2093286514,-0.2253688425,-0.0167210288,-0.0132269189,-0.184551537,0.2123140246,0.6192135215,-0.0150432223,0.0824793652,-0.1268968582,-0.0155486753,0.5002205968,0.1098568216,0.0925252214,-0.0079286359,-0.0417051278,0.4966493547,-0.0013129428,-0.1259710789,-0.0148605099,-0.013630108,-0.0646863058,-0.104604587,-0.2178259641,-0.0738868713,-0.2776648998,0.3043035865,-0.117547363,-0.0585312918,-0.0471505858,0.1061868444,-0.3491126597,0.1684893072,-0.4232687652,-0.3773184419,0.1720929891,0.1551331282,-0.0108234957,0.2384415865,-0.2047417462,0.2194369435,-0.274028182,0.0555009618,-0.2189284861,0.1612891257,-0.0558832213,-0.0424223617,-0.0094791884,0.3087970316,0.0383313261,-0.0179966092,-0.3062886596,0.2860310078,-0.1254843473,0.0803563818,-0.0322646983,-0.1433701664,0.0688060746,-0.0715434849,-0.1751839221,0.2081920803,0.3319509625,-0.1801262945,-0.043142952,0.1128761917,0.1272121966,0.0814221054,0.0988641679,0.1639728397,0.1439921707,-0.0658260733,0.1012795344,0.0774310902,0.2337106764,-0.0949872434,0.2047105134,-0.1564514041,-0.167145744,-0.1188064739,0.3131195307,0.1700583696,0.167417407,0.0814866126,-0.275252223,-0.0948746875,0.0785581768,0.1460482925,0.4009645283,-0.0518327467,-0.1083415523,-0.0525707416,-0.4165976346,0.0711256638,0.1707080156,0.0366290882,0.2091285437,0.3018561602,0.0413160138,0.0873196498,-0.2952665389,-0.3570170999,0.2824332714,0.4405301809,-0.2284827828,0.2078800648,-0.0305316318,-0.0025111341,-0.0678482279,-0.3336339295,-0.188629657,-0.3429944813,-0.1213362366,0.4337745905,0.024119148,0.1810040623,-0.3977624178,-0.0220502764,0.1947504133,-0.1147857755,-0.0313790925,-0.3157985508,-0.269328475,0.0459421203,0.3818612099,0.0910761356,0.3510979414,-0.0667702854,-0.0925358012,-0.007033736,-0.0584152527,-0.0755677298,-0.1837216765,0.2440479547,0.009521856,0.2855836153,-0.1770296395,-0.5220327377,0.5017470121,-0.2557020187,-0.0577849969,0.3151076138,0.2711132169,-0.249190703,-0.0794775337,-0.3746882677,-0.1264498532,-0.280605495,-0.0594684482,0.0748294741,0.0620202571,0.3974690437,0.0708564147,0.2160111964,-0.0025732063,0.1134374887,-0.0482623316,-0.0900634304,0.1859846711,-0.310032934,-0.3173182309,-0.1213430315,0.0099359797,0.3347615004,0.1107365489,-0.3489591479,-0.2141319066,-0.1049579829,0.0853126571,-0.2213259488,-0.0400672704,0.1828597784,0.0723028034,-0.0657224879,-0.0866443887,0.035989143,0.0712399781,0.1523462385,0.1349620819,-0.0063045779,0.0892271996,-0.0848842263,0.6008986831,0.218693167,-0.2143773437,0.3002861738,-0.1032764912,0.005145404,-0.2086836398,-0.4398039281,0.2641642988,-0.11523287,0.1829283237,0.1723142117,-0.0794930086,-0.2247841209,-0.192558974,0.0688526183,-0.178447336,-0.2727790177,0.1461688429,-0.2173948586,0.2787112296,-0.0688640028,0.2225363404,-0.2052234709,-0.167981565,0.0698513761,0.166933611,-0.090545617,0.0796746612,-0.2440017164,-0.2761095464,-0.4799583256,-0.0278399177,0.2410321236,0.6437504888,-0.0863839611,0.0371890776,0.0807222873,0.0817412362,0.1373739988,-0.2301357985,0.1822634041,0.2891299725,0.0318248197,-0.3197474778,-0.0643306822,-0.1401119977,0.3580168784,0.0766504705,-0.0544316322,-0.037231572,-0.1676542461,0.1398317814,0.2599402964,-0.1872950047,-0.0441127233,-0.0917341635,-0.0898099318,-0.3077637553,-0.0061545791,-0.0099474099,0.3356595933,0.1229696721,0.2657254636,-0.5462990999,0.1166565344,0.0901664123,0.1745875478,0.0959533602,0.0579445958,0.187246576,-0.1666348726,0.1832098216,-0.145905599,0.5633814335,0.1473762542,-0.3307984769,-0.3195302188,0.0216632597,0.1674342602,0.1839805692,-0.0185859147,-0.1411214918,-0.2769463956,-0.061463993,-0.0087094773,0.3045702279,0.5610407591,0.0026275387,-0.3271814287,-0.2585409284,0.482839793,-0.2850035131,0.1173297018,0.321557641,0.053378012,-0.3131920397,0.0841390863,0.0778320804,0.6767753959,0.1695389003,0.311147064,0.3826704919,0.1839297265,0.4142697752,-0.043089699,0.2847148776,-0.3927352428,-0.1334252506,-0.0806859359,-0.065462403,0.1070814878,-0.0088732634,-0.0254380628,0.1506213844,0.1568897963,-0.004141801,0.057148274,0.4492524862,-0.0387835056,-0.1968196929,-0.3470262885,0.156016469,-0.2179998308,0.297508955,-0.0907334238,-0.0664841011,0.0552040972,-0.0673523247,-0.2555586398,0.2574389279,-0.299474448,0.459457159,0.0578220226,-0.4756816328,0.198415935,0.2070970535,0.1756633669,0.0056988141,-0.2995947301,0.2387785167,-0.2900909483,-0.2218656987,-0.0672868341,0.1119474173,0.4200738966,-0.1725511253,-0.146377489,0.1479405463,-0.1206544265,-0.2415021658,0.1039589867,0.0368667245,0.3524883091,-0.392575711,-0.5152193308,-0.00242018,-0.1820756793,-0.1749101579,0.1744999588,0.023450667,-0.189030841,0.1518952847,0.1498275995,-0.1989797652,0.1164065152,0.2480414361,-0.2495406866,0.1462838054,0.76375705,0.0339458175,0.0128051164,-0.3138304949,-0.0542888604,0.1250851899,-0.4819999039,-0.1171273738,0.1255033612,0.1691254526,-0.1442471892,0.4983651936,0.0630208924,-0.0188453011,0.2471855134,-0.542414248,-0.1062184349,-0.0430215374,0.0177533776,0.2377753258,-0.0026245497,0.114889577,0.1355090588,0.1033881679,-0.272637248,0.0142540857,-0.1576250941,0.137997821,0.2178348005,-0.1565312743,0.2232821733,-0.0769478604,0.1563859284,-0.0958470255,-0.1230344325,-0.2482931763,-0.2548377514,0.1001116261,0.2455675602,-0.1645215303,-0.0306790825,-0.102333799,0.0236772168,-0.2311056703,-0.0084627206,-0.1177286133,-0.2933430374,0.3894323707,0.0438978039,0.2783688009,-0.1116899401,0.1263558418,-0.3007574677,0.1491208375,-0.0184958763,-0.075229764,0.2758128941,0.0936535597,-0.4658874869,0.0054211435,-0.4026061594,0.0371483862,0.3698208928,-0.4551106691,0.1599583179,-0.1316279471,0.0954658315,0.4656842053,-0.2714086771,-0.0036663271,0.3496651351,0.2378421873,-0.2974172235,-0.1892595887,0.2399912179,0.0613459609,0.0324590541,0.0272394195,0.0466615967,-0.3037491441,-0.3212345541,0.0892789215,-0.0293396804,-0.2656017244,-0.0114079248,0.3393265009,-0.0023644918,-0.0595147535,-0.050050322,0.2110216618,0.0353926681,0.382693857,-0.2926004231,0.2069978416,0.0592639633,0.0439102203,-0.1111432463,-0.2192169279,0.0503001995,-0.1578414291,0.0338347405,0.2952830493,0.0998148471,0.1530806422,-0.2855040729,-0.1245369539,-0.2080229074,0.2279238254,-0.1790424585,-0.0489066392,-0.2429951131,0.0210169256,-0.0258763712,-0.0946998745,-0.1416703463,-0.038179189,0.3095561862,0.2532219291,-0.17226924,-0.4555348456,0.1082671434,0.3432940543,0.180047825,-0.2458842993,0.336853832,0.1938146502,-0.0687782988,-0.0141235748,0.397110045,0.4357684255,0.338393867,-0.3642207682,0.0811674222,-0.2033814788,-0.0568215884,-0.0232602563,0.2695125937,0.1060727611,0.1173768565,0.4624233842,0.1214860529,-0.1247266233,0.0436382517,-0.0880563408,-0.2605327666,-0.262947619,0.5191478729,0.1857949793,-0.1143881306,-0.3206512332,0.1899377853,-0.2749869525,-0.0809145123,0.0628518164,0.2405894995,0.1131533161,-0.2138276696,0.0705087706,0.0003153924,0.7061742544,0.4877223969,-0.0177194085,-0.3658049405,-0.1165233403,-0.5989871025,0.2178366035,-0.3639689684,-0.2690189183,0.2833368778,-0.0723317787,-0.2880821824,0.2485276014,0.1276198775,0.1581964493,-0.0572769679,0.09834712,-0.4907647073,-0.1486053318,-0.0306509864,-0.0076748356,-0.0831808746,-0.257052511,0.1453228593,-0.4349942505,0.1161028594,0.01198583,0.0693481863,-0.0002400207,0.0966407359,0.3544250131,0.1653244644,0.5012484193,-0.0355607159,0.0036567017,-0.1911553591,-0.1566222012,-0.2289921194,0.1318234056,0.0876291767,0.5133115053,-0.0261075124,-0.0685456768,-0.2640267909,0.4031677544,-0.0810070783,0.0883163437,-0.1350804716,0.0773147568,-0.0703499094,0.2275845408,0.2643646002,0.4245406091,-0.1783216298,0.2022095919,-0.2722912431,-0.4602457583,0.3289350867,-0.6614814401,-0.517790556,0.0751233175,0.1161924526,-0.1423345357,0.0581936017,-0.6006346941,0.0705664605,0.2426124364,-0.1114050075,-0.317119956,0.2179448009,0.0165422,0.0800011158,0.0435740389,0.44835639,-0.148473084,-0.1597679257,-0.0049338331,-0.1632519215]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/166","title":"Add a method to shuffle a dataset","comments":"+1 for the naming convention\r\n\r\nAbout the `shuffle` method, from my understanding it should be done in `Dataloader` (better separation between dataset processing - usage)","body":"Could maybe be a `dataset.shuffle(generator=None, seed=None)` signature method.\r\n\r\nAlso, we could maybe have a clear indication of which method modify in-place and which methods return\/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-place. What do you think?","comment_length":25,"text":"Add a method to shuffle a dataset\nCould maybe be a `dataset.shuffle(generator=None, seed=None)` signature method.\r\n\r\nAlso, we could maybe have a clear indication of which method modify in-place and which methods return\/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-place. What do you think?\n+1 for the naming convention\r\n\r\nAbout the `shuffle` method, from my understanding it should be done in `Dataloader` (better separation between dataset processing - usage)","embeddings":[0.1534199119,-0.0373094566,-0.0626969859,-0.129071027,0.1771616787,0.0837633684,0.2535591722,0.2304865569,-0.2128239423,0.3302725852,-0.0707573071,0.736138761,-0.22251454,-0.201811865,0.0237806197,-0.0824748725,0.2442180812,-0.2117860466,-0.1650031358,0.0270733535,-0.114130117,-0.3897317946,0.0232441165,-0.1421938986,-0.1926327497,-0.0498330221,-0.1330985576,0.0474074855,-0.2461620569,-0.254447788,-0.4673665464,0.7014350891,-0.1527414173,0.4499371052,-0.0001127545,-0.2929293811,0.3552512527,0.0323526859,-0.3119530082,0.0747513101,-0.1722263396,-0.0720390752,0.066071108,-0.2017751336,0.2140467763,0.0577745289,0.056992393,-0.1298427731,0.1646680534,-0.0894344673,0.1747247726,-0.0997950658,-0.3013956249,-0.0162508078,0.5358915925,0.5136160254,-0.1551083922,0.1145031527,0.5382512808,0.3613963723,-0.1880674213,0.1561811715,-0.0936744362,-0.0767192766,0.4886149168,-0.2939814329,-0.5896620154,0.0441594161,0.1208570153,0.1790135801,0.600099206,-0.2012623996,-0.4169183969,-0.2404291332,0.2678801417,-0.1333735138,0.0538631491,-0.1913710237,-0.0122782988,-0.1094488278,-0.1305658072,-0.1574682295,-0.0056679989,-0.058629632,0.4977308512,0.5297002792,0.1231477931,-0.0194789357,0.1149401367,0.0041558417,0.5591573119,-0.1023659408,0.0144958962,0.2615667284,-0.1102365255,-0.1002175808,-0.0076730279,0.2563235462,0.2494795173,0.1132015958,0.3054137528,0.2701373994,-0.0686765388,0.0746165365,-0.0694335476,-0.0760908872,0.0326859951,-0.0448363349,0.2888597846,-0.315322727,0.1799629182,-0.1102357283,-0.1513262391,-0.0501809679,0.0384859368,0.1283468157,-0.3014697731,-0.1019995958,0.1980670094,-0.5291099548,0.063182883,-0.4272428155,0.078263551,0.1191479117,0.2903625369,0.1153954268,-0.1512455344,0.0494494848,0.2924679816,-0.126424253,-0.0667637065,0.0325260013,-0.4743708968,0.2319803834,0.1612490118,-0.2171575576,0.1098426655,0.0895114169,0.1500685811,0.2294976562,0.1636538059,0.2171195,0.3288630843,-0.1775386631,-0.4487595856,-0.0693501309,-0.0962270945,0.2949227989,-0.296510905,0.2620727122,-0.2736850977,-0.2621245086,-0.063977167,0.126588583,-0.025156701,-0.1012657806,-0.2362124771,0.2004535794,0.1188166961,-0.3033381701,0.3808519542,0.1356821954,0.1871736497,-0.1035546958,-0.0309804305,0.2598655522,-0.0952065289,-0.0204186831,-0.3389229476,-0.1029439494,0.1810968071,0.1123062298,-0.1615383327,0.0239715166,-0.1227995083,-0.1310105622,0.4819880426,0.2170279473,-0.1716065258,-0.0348062478,-0.1534935683,-0.1682638377,0.5057479739,0.3218181133,-0.0796090141,-0.2573798597,0.2834156454,-0.0926008001,-0.4131713212,0.0371191837,-0.2332087606,-0.134677574,0.3806222379,0.2380657196,-0.0918979049,0.0011277061,0.0451045111,-0.4092569351,0.1858002543,-0.1324031949,-0.1216470823,-0.252774626,-0.0293269381,0.3391713798,-0.0838654488,-0.0944200605,-0.0348933488,-0.1526423842,0.2866241336,-0.1084313169,0.0284079481,-0.2912257612,0.0380689725,-0.3927570283,-0.0045623919,-0.1235525161,0.0338835716,0.2487103641,-0.2404133081,-0.2340536416,-0.2861448824,-0.1929812282,-0.1690480709,-0.1284981519,-0.2281817049,-0.0566281676,-0.045760747,-0.0460523702,-0.2762995958,0.2086228579,-0.0301994905,-0.0818403065,-0.1712910533,0.261354208,0.0889698192,-0.1329397112,0.4365851283,0.3457510471,-0.0626278222,0.0924833342,0.1802336276,-0.0563746244,-0.0167035572,-0.0987445116,-0.2994470298,0.2603794634,-0.2665770054,0.0909133032,-0.2621858716,-0.3188471794,-0.0470446572,0.3720082641,-0.2649490833,0.1697956473,0.1168687865,0.2092733234,0.3513684571,0.1626689136,-0.3234312832,-0.1395172626,0.2649697959,-0.1599839479,-0.0220340025,0.3419928253,0.2032807618,-0.2494020611,0.1452873349,0.0685134083,0.3731107116,0.3408052027,0.0128899449,0.0719234049,0.2797623873,-0.1501169205,-0.0482702069,0.1665181518,-0.2826219499,-0.0111773713,0.1748601347,0.1202738807,-0.1523288488,-0.3830471933,0.076049,0.0322069414,-0.0361786485,-0.0776990876,-0.018525213,-0.0737834424,-0.2008932233,-0.1971309185,-0.3390847743,-0.2624280453,0.2674481273,0.0584906973,-0.2952111661,0.3880605102,0.1832676232,0.5487818718,-0.2152484506,-0.1079600081,-0.3561215401,-0.4411757588,0.1263572127,0.0409559198,0.2258597761,-0.1359730512,0.7025209069,-0.0510038771,-0.0799882561,-0.3822794259,-0.469830066,-0.0898152888,-0.0074408106,-0.2571099102,0.3828089237,-0.327059716,0.0884958357,-0.4728185236,-0.0689703077,-0.2552244365,-0.0102612013,0.0073529021,-0.0102490671,0.0092774797,-0.1842011958,-0.2975763679,-0.1312457919,-0.404979825,0.0740640685,-0.6318072677,-0.0500301234,0.0120368581,-0.2868341208,-0.132414937,-0.2205682993,0.1613693535,-0.1366978735,-0.7596729398,0.192911461,-0.2022341937,0.0575747043,-0.1422053277,-0.0697121099,0.0288929269,0.6446994543,0.2754819393,-0.1669305563,0.1485349536,0.2016544193,-0.0016149917,0.1561451703,0.3470445573,0.2847751081,-0.1229221076,0.0623417944,-0.2354115993,-0.1129372194,-0.1072700471,0.3154536486,0.1071916148,0.3211006522,0.016647812,0.6867966056,-0.6602865458,-0.0905801803,0.0193492956,0.2002179176,0.1372473538,-0.0679681301,-0.0909691602,0.5424605608,0.2456187457,-0.3071582317,0.1921598613,-0.0483933054,-0.1601231396,0.0377632938,0.1311082691,-0.1474983394,-0.142530784,0.3427986503,-0.3273083866,0.0862519071,-0.0400641672,0.173698917,-0.2951994836,0.0042597195,0.1468442082,0.2026420683,0.3288092315,-0.2704314888,-0.436099261,-0.0211798381,-0.1265943497,0.2958840132,0.2474629134,0.1288380325,-0.0136047583,-0.1818044186,0.122365132,0.2576231062,0.4825322628,-0.2469895333,-0.2833612263,-0.0055692671,-0.0168965477,-0.1124948934,0.0152133359,-0.0755571276,0.1418469101,0.0728932545,0.0938347504,-0.2272615582,-0.1768296063,0.1136337072,0.1961219311,0.1610937268,-0.3392724991,-0.082963489,-0.1653590798,-0.1599015594,0.1925914288,-0.0237951633,-0.3008366227,-0.0556079186,0.0588310175,-0.0334818028,0.3097732067,0.0266839061,0.1292271018,0.0320875123,0.1218265519,0.0449799709,0.0122386329,0.3770184517,0.4276652634,-0.2056345195,-0.4869756401,-0.170944348,0.0145150069,-0.5441596508,0.244069621,0.2382855713,0.2514321208,0.2967531383,-0.1404077411,0.1760651171,-0.2185067534,0.0735383555,0.0524369851,0.007403553,-0.5293944478,-0.9117104411,0.5416220427,0.37718606,-0.3224137127,0.5554153323,0.0005213641,-0.268488884,0.4150928557,0.2951985896,1.2055472136,-0.0814688802,0.1874591112,-0.0292165149,-0.376734674,0.2766079307,-0.180907473,0.1494001448,-0.1293389648,-0.3936464787,0.0539814308,-0.0889720023,0.0747401193,0.3279801905,-0.1248976961,0.3453121185,-0.3932631612,0.1730785072,0.1445833445,0.1214244366,-0.2439465374,-0.2914193273,-0.0822787583,0.0728245154,-0.0915910006,-0.2102664709,-0.0334864035,-0.0050693857,-0.0670093372,0.1408868134,-0.1443740129,-0.3057850897,0.1522462368,-0.0800425187,-0.2814941108,-0.17594935,0.4433329105,0.108589083,0.2535839379,0.0798494741,0.1667892337,0.2920091152,0.0875647515,0.4055556655,-0.0955692604,-0.0892812237,0.1074694842,0.082121104,-0.0487220064,0.0300491657,0.1108833477,-0.4286027253,-0.5298848152,-0.0848549306,0.5617539883,-0.1269346476,0.0926104039,0.0187527854,-0.1059865355,-0.0174996872,0.1370433122,0.2787863612,-0.2586558759,0.3623966575,-0.2182174474,0.0199905969,-0.1200647205,0.1049025431,-0.2883319557,-0.1277452856,0.2179229558,-0.2194974422,-0.1973453313,-0.1657759398,0.117488645,0.2331488281,-0.0043129758,0.0334511772,-0.1374044418,-0.2781903446,-0.1548792273,-0.1567089856,0.2195171118,0.3490116,-0.0353539214,0.1103747338,-0.4601895809,0.0767346695,0.1705332547,0.339868784,-0.1675649583,-0.0466486551,-0.1151122227,-0.0192351751,-0.3286210001,-0.1494592577,0.2203099132,0.2718959749,0.009397666,0.2210141569,-0.201407671,-0.2405496091,0.0655259863,0.061105147,-0.0186436642,-0.1480703652,-0.1365438849,0.1117107868,-0.0700048432,0.229429245,0.2037572265,-0.0818171501,-0.0028204131,-0.2503929734,0.0264820289,0.2947494686,0.0050876583,0.4968261421,0.6184055209,0.0818525106,-0.2173988819,0.2066225857,-0.0255023837,0.0314403586,0.1206334904,0.147725895,-0.0265630521,-0.0203822646,0.1190709919,0.321632266,0.1847999096,0.0678361505,-0.0905277133,0.2885020971,0.1930856705,0.4536784887,0.0995225012,0.4589992464,0.0012193552,-0.3575810492,0.4314953983,0.206557706,-0.0190283768,-0.0736285895,-0.2535411119,-0.0057583065,0.0805494487,-0.0053965994,-0.2296503931,0.015040027,-0.3449503779,-0.077599518,0.1451237798,-0.4788012803,0.3961937129,0.3324785233,-0.1217051372,-0.0279557817,0.3708786368,0.009444139,-0.2695912421,0.1324599832,0.4628453553,0.9553408027,0.3267761469,0.1606335044,0.13742733,0.1492273211,0.1605956256,0.6927061081,0.3411798775,0.0890379846,0.3850889206,0.2273181826,0.1477177292,-0.3402733207,0.2771571279,0.0776449144,0.2560738623,0.0305005293,-0.0663550645,-0.1942120492,-0.2519174218,-0.0834770873,0.0126168253,0.0515346602,0.5799618363,0.2349966764,0.259142518,-0.3528327048,0.1266424656,0.2718269229,0.4411756396,-0.1501090825,-0.4055159092,0.142024219,-0.1715123951,0.7451763153,-0.0197703708,-0.0012754285,0.2458312958,0.0297372229,0.1183305308,-0.3746955991,-0.0371007659,-0.0156203741,0.1232681945,0.1435822845,-0.1799301654,-0.0120897563,0.1765798777,-0.1434436142,0.0751900151,-0.0751836672,-0.046274323,-0.1835750788,0.3442087471,0.4996169806,-0.3088755608,0.1062352955,-0.0849186182,-0.1956761926,-0.1527646184,0.0405588895,0.1610234082,-0.2053282261,0.0913472623,0.0590674616,0.2930758297,0.2837913334,0.2766535282,-0.0072681,-0.1498103291,-0.3939853907,-0.3894713819,-0.1301595569,0.476010859,0.0036120068,-0.0354772098,-0.1831990331,0.0368414856,-0.0670794994,0.1456745714,-0.1224078983,0.0533313192,0.1979422718,-0.3570578098,-0.1212216914,-0.1313951761,-0.0225724969,0.183655411,-0.0610039197,-0.306797713,0.4100827277,0.0441316329,-0.309710741,0.1448233724,0.0698837414,0.1139738709,0.2117415965,-0.0195594691,0.3880785406,-0.0024604348,-0.1821307987,0.0263851825,-0.1421173811,-0.5724025369,0.1211909205,0.0075439629,-0.0112707885,-0.0383663736,-0.2087760568,-0.4421339333,0.5485701561,-0.3553211689,0.0855077878,-0.7277727127,0.1580511779,0.2472626567,0.0818874761,-0.0485803783,0.1143995747,0.0510065928,0.2707607746,-0.2295465171,0.1943167895,0.5219248533,0.1695054919,-0.2427388579,-0.3214066625,0.2131143212,-0.089612402,0.0295326803,-0.3195725381,-0.0837741867,0.2322507799,0.0034360823,0.0798486397,0.3839514256,0.128450036,-0.0466571972,-0.2227720618,-0.0014488882,-0.1201055646,-0.0303441267,-0.1148541272,-0.2884330153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/166","title":"Add a method to shuffle a dataset","comments":"+1 for shuffle in `Dataloader`. \r\nSome `Dataloader` just store idxs of dataset and just shuffle those idxs, which might(?) be faster than do shuffle in dataset, especially when doing shuffle every epoch.\r\n\r\nAlso +1 for the naming convention.","body":"Could maybe be a `dataset.shuffle(generator=None, seed=None)` signature method.\r\n\r\nAlso, we could maybe have a clear indication of which method modify in-place and which methods return\/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-place. What do you think?","comment_length":38,"text":"Add a method to shuffle a dataset\nCould maybe be a `dataset.shuffle(generator=None, seed=None)` signature method.\r\n\r\nAlso, we could maybe have a clear indication of which method modify in-place and which methods return\/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-place. What do you think?\n+1 for shuffle in `Dataloader`. \r\nSome `Dataloader` just store idxs of dataset and just shuffle those idxs, which might(?) be faster than do shuffle in dataset, especially when doing shuffle every epoch.\r\n\r\nAlso +1 for the naming convention.","embeddings":[0.0704690665,0.0160601158,-0.081076622,-0.1330937445,0.1867506802,-0.0187649392,0.3612542748,0.2494378537,-0.1316797584,0.398124367,-0.1151702181,0.7459426522,-0.3297457397,-0.2149706483,-0.0445238054,-0.0798154026,0.2653796673,-0.1031481922,-0.2052428573,0.0562242642,-0.1247075349,-0.3481492698,0.0236733481,-0.1405295581,-0.0896880254,0.012186124,-0.1130371764,0.0714675263,-0.2733206451,-0.2186136097,-0.4450689554,0.7157347202,-0.1049824506,0.5668814778,-0.0001091151,-0.2588721216,0.3414728045,0.0329268649,-0.3560408652,0.1413132101,-0.133158654,-0.0783198178,0.0386806056,-0.2729430497,0.2220702767,-0.0343981311,0.1158991307,-0.1127861366,0.1109896749,-0.1473737508,0.1819331348,-0.0993652567,-0.3846822679,0.0020908492,0.578519702,0.4610337317,-0.2091200054,0.1824525148,0.574575305,0.3853920698,-0.3338279128,0.1915495694,-0.1309427023,-0.1019477099,0.5779697299,-0.3001551926,-0.5396052003,0.0647645891,-0.0393920168,0.2036895007,0.6476197243,-0.1902356595,-0.4480341077,-0.1510591656,0.1816706657,-0.0496767871,0.0307899434,-0.2421191484,-0.0195767432,-0.0853262246,-0.0373785309,-0.1129700616,-0.0025622498,-0.1363610774,0.421797514,0.5597810745,0.1549821496,-0.0286143627,0.2360943109,-0.0222396906,0.5372052789,-0.0696752667,0.0733335465,0.2372976393,-0.217320621,-0.1808136553,-0.0155333262,0.2117105424,0.1879729182,0.1057438105,0.3026645482,0.2721392512,0.013139585,0.1095649898,-0.1144884303,-0.038552694,0.0334646702,-0.1094348207,0.3762949109,-0.2285953909,0.1526614726,-0.0816256106,-0.2623736262,-0.0178392008,0.1021517888,0.1414059252,-0.4133082628,0.0568721928,0.1791498959,-0.4913693964,0.1562060118,-0.4261989594,0.1031714454,0.18962273,0.1870161742,0.1320405155,-0.1550371498,-0.0127899544,0.2115512788,-0.184740752,-0.1022361442,0.0491869822,-0.507145524,0.2194849998,0.1013816968,-0.2031907737,0.1159764007,0.0387395322,0.136518091,0.3439450264,0.1533025652,0.2006931305,0.1924792677,-0.2556225657,-0.4391724169,-0.1295641661,-0.1443853378,0.3152359426,-0.2803953588,0.2423091829,-0.29306826,-0.2366222292,-0.1138081625,0.1717359573,-0.0018636793,-0.0619996898,-0.2293261737,0.246396184,0.0740043819,-0.3056387901,0.4155263007,0.0785193592,0.1712349504,-0.1233353764,-0.0725498423,0.2618198693,-0.1112703532,-0.0184363145,-0.230252713,-0.1622869819,0.205648914,0.1142255291,-0.1920033842,0.0078952899,-0.1419530213,-0.1474459022,0.4907092154,0.1812635809,-0.1588147283,-0.0123546282,-0.1144363135,-0.1025750786,0.6056385636,0.3467120826,0.0564940237,-0.3341640234,0.2879296243,-0.0612017065,-0.3102137446,0.0981529653,-0.2247581184,-0.1184130087,0.3814633787,0.2611493468,-0.1228186637,-0.0091771241,0.0056606396,-0.3373622,0.1947128624,-0.1716935784,-0.0731263235,-0.2363464832,-0.0819968805,0.3089695275,-0.0150543284,-0.0649891198,-0.0631167591,-0.1157574281,0.29175964,0.0176495314,0.1858094186,-0.2812825739,0.0290214121,-0.2989604473,-0.0154391713,-0.174696058,0.0616281666,0.2616629303,-0.2221287042,-0.2475360483,-0.2646149695,-0.1795042455,-0.2702847719,-0.1623175293,-0.2577840388,-0.1003571302,-0.0421096273,0.0273020007,-0.2138474584,0.2204702348,-0.1302502453,-0.0731815547,-0.1548641026,0.248049289,0.1106282026,0.0012288903,0.3039054275,0.3766371608,-0.0230145231,0.0787241757,0.3076277077,-0.0839393139,0.0128669385,-0.1136038601,-0.2761874795,0.2542312741,-0.2328188866,0.0878826529,-0.2713023722,-0.3496966064,-0.0082614282,0.359682858,-0.3368625641,0.2372022718,0.1287111342,0.2788371742,0.407399863,0.2508471906,-0.3290274739,-0.1328150779,0.3740357757,-0.1232951954,-0.079807274,0.3697065413,0.1284963787,-0.2949621975,0.1104044765,0.0889155045,0.3699173927,0.3470844626,0.0725460127,-0.0338136256,0.3311807811,-0.1331802458,-0.0100540407,0.096919775,-0.1369654089,-0.0208794028,0.1016628444,0.0581178144,-0.1998404711,-0.5052341223,0.033565145,0.0917880833,0.0397617519,-0.0772798061,0.0556890666,-0.0526593328,-0.1414629519,-0.1758359224,-0.2801253498,-0.2714925408,0.2531290948,0.1074848771,-0.3382922411,0.3336867094,0.1831430197,0.5316727161,-0.1584078223,-0.0584187023,-0.3735506237,-0.4471942186,0.1538641453,0.0767324269,0.2285131365,-0.1827137917,0.6215420961,-0.0355267115,-0.0638798401,-0.3513756096,-0.3830470443,-0.062467549,-0.0042403601,-0.2293925732,0.3913440108,-0.2218208611,0.1168422997,-0.4907086492,-0.0055111027,-0.3185169697,-0.1076779142,0.0271484479,0.0252681263,0.0133220227,-0.1668285429,-0.2421270907,-0.17180264,-0.4843709469,0.0358714052,-0.5416013598,-0.0025590253,0.0130536994,-0.2118822634,-0.1035678983,-0.1109935269,0.1574204564,-0.1544938833,-0.8162309527,0.2280825973,-0.1825128943,0.0700531453,-0.2289283276,-0.115426816,0.0835939795,0.6077492833,0.2217947245,-0.0639686808,0.1345303655,0.1378970593,0.0393652804,0.1684257388,0.2622101903,0.3026329577,-0.1817001849,0.0207795333,-0.2253602445,-0.0899050459,-0.1223536357,0.2259428054,0.1704076976,0.3607722819,-0.015084832,0.7392673492,-0.677593708,-0.0537548438,-0.0282049254,0.1690758616,0.1355460584,-0.031218864,-0.1497747749,0.5597659349,0.2193570882,-0.305631578,0.2308703214,-0.0871220976,-0.2488691658,-0.0113426326,0.0699885264,-0.0415941179,-0.0578062534,0.3630122244,-0.3087019026,0.0132483747,-0.0610507913,0.1225184873,-0.3269179761,0.0095565123,0.1524282992,0.2298136353,0.2655058503,-0.243544206,-0.4098938406,0.0442300551,-0.2146961242,0.1624693424,0.2266132534,0.0795100853,0.0974687189,-0.2776519656,0.06662599,0.2089674175,0.5011021495,-0.1220227405,-0.2938186526,-0.0181357153,-0.0350466967,-0.0899970159,-0.0249856599,-0.1166574135,0.125579074,0.0206605382,0.0976838693,-0.2659020126,-0.1665679216,0.0744646341,0.1325633675,0.0766402557,-0.3359169662,-0.1441160589,-0.0007817466,-0.1548809111,0.1778858304,-0.0962789133,-0.28561306,-0.0614372119,-0.0219436381,0.0364086069,0.2939795554,0.0735350326,0.2143480927,0.0576428771,0.0708463043,0.0482284315,-0.0650582463,0.2817289233,0.4365673959,-0.1428242773,-0.4651130736,-0.1063261703,0.1357996464,-0.4808528125,0.2497146279,0.1165674552,0.2149628401,0.1901762635,-0.1473195702,0.2178759128,-0.2683398724,-0.0105708428,0.0611147806,0.0282117017,-0.4887733757,-0.9323247671,0.5213588476,0.3647642732,-0.2582478821,0.574324131,0.0042627086,-0.2041105926,0.4123572707,0.3018024862,1.1519231796,-0.0961458981,0.2185096294,0.1011870131,-0.2109863758,0.2955074906,-0.2298139632,0.1557398885,-0.1591601372,-0.3745518923,0.0661106631,-0.1327633858,0.0498974174,0.2634173036,-0.2560466826,0.2536982298,-0.3084175289,0.2344588488,0.0204369165,0.1914422065,-0.2874631286,-0.3286153674,-0.0508241914,0.099836342,-0.0581801906,-0.30008775,-0.0473760888,-0.0546851791,-0.0937070772,0.2028754652,-0.1015469208,-0.2596932948,0.1305816919,-0.0596916713,-0.2657776773,-0.0993007943,0.3667414188,0.034363769,0.125494495,-0.0148783028,0.1513283998,0.2482614219,0.0390186422,0.466606766,-0.0450372621,-0.131803453,0.2440776229,0.0633664653,0.0054161726,0.0451870002,0.1168513149,-0.4874970913,-0.4285205007,-0.0431745201,0.599740088,-0.1540204287,0.1164660081,0.0491575375,-0.1950256079,-0.0481614582,0.1584091485,0.2610592842,-0.2325856835,0.2867864072,-0.1204797849,0.0426859967,-0.1251034886,0.0896789655,-0.1612811536,-0.1136116013,0.1927433163,-0.2361187637,-0.2805587947,-0.1918399483,0.1516358852,0.2131286114,-0.0291945506,0.1103460789,-0.2356516868,-0.3742775023,-0.0794693083,-0.1926430613,0.2327066064,0.3861809969,-0.070741564,0.1231981143,-0.4934320748,0.0880587548,0.0968146622,0.3548372686,-0.2407083958,-0.0927828476,-0.1096079871,-0.0184561573,-0.3788084686,-0.2516284883,0.1554828286,0.1993113309,0.0081897229,0.1161620244,-0.2287403196,-0.10670086,0.0968572423,0.0808686092,-0.0796146542,-0.1717631966,-0.1784939766,0.0964445025,-0.0995346829,0.2165154517,0.2164936066,-0.0610805377,-0.0931972265,-0.2493307739,0.0512282625,0.2615832388,-0.05960593,0.4868053198,0.6710011363,0.0972442627,-0.2060213387,0.1600845009,0.0736619756,0.0598272234,0.0875620097,0.1277291477,-0.0364175849,-0.0929659531,0.1047328562,0.4397211671,0.1550138891,0.1372467428,-0.0029528865,0.2808757722,0.1976499856,0.358643502,0.1862039715,0.4341942966,0.0411203131,-0.3968984187,0.4125515223,0.2342455685,-0.0356522202,-0.1306801736,-0.2073367238,-0.1162463576,0.1384901255,0.0103824027,-0.1483621895,0.0067654243,-0.2170861661,-0.131002292,0.0985452011,-0.4322230816,0.3699084818,0.2464427799,-0.1015389934,-0.0739281103,0.4425268769,0.0176673196,-0.1838265508,0.1610752046,0.388546586,0.887912631,0.387340188,0.1493379176,0.0943550244,0.1737465858,0.1442482471,0.7236198187,0.2329770774,0.1124048308,0.306879729,0.2533042133,0.102000244,-0.2223381698,0.2781133652,0.1487628073,0.132282123,0.0113918837,-0.0357112028,-0.2258719206,-0.3367291391,-0.0903291777,-0.0132283308,0.0849900842,0.524677515,0.199939996,0.2989271581,-0.2885638177,0.1400090009,0.2076197863,0.4345971644,-0.2177172899,-0.4291311204,0.1927115619,-0.1165731549,0.6738458872,0.0535891764,-0.0311514381,0.2896184027,0.0544737689,0.0748949945,-0.3837047815,-0.0371138044,-0.0679518655,0.1082939059,0.1000972688,-0.2352246791,0.0417208858,0.1770892143,-0.199933067,0.0167148579,0.0351784304,-0.1471107155,-0.1254286021,0.2332198471,0.5211764574,-0.2374134511,0.1050830707,-0.0206958391,-0.3141877055,-0.1590210497,-0.0296092015,0.0701116994,-0.1647299975,0.0476915836,0.0857360214,0.2440963984,0.3486326039,0.2794227302,0.0095753986,-0.1195922866,-0.4181688428,-0.4080618918,-0.1301293075,0.4473884106,0.0212766826,-0.023291016,-0.0365239196,0.0388338342,-0.0390183702,0.1655305028,-0.1370772123,0.0162820593,0.1175595,-0.369233191,-0.1099373102,-0.0945160389,-0.0524839647,0.187233597,-0.1290984154,-0.2749146521,0.3199726343,0.1062960774,-0.2490778714,0.0901098773,0.1567774862,0.1838329583,0.2313624024,-0.105749473,0.2222574502,-0.0139714135,-0.1876373291,0.0664204955,-0.0860761255,-0.4891245663,0.0098314798,0.0942463353,0.0846359357,-0.0203239042,-0.2257528603,-0.4579548836,0.5146232247,-0.3050029874,0.0837943479,-0.7182448506,0.2188802212,0.2385821491,0.051791992,-0.1557138115,0.1274532974,0.1018624157,0.2214878052,-0.3064768314,0.165153563,0.4485177696,0.1076698303,-0.2351182997,-0.286156714,0.1609076411,-0.0434874073,0.1759275496,-0.3066344857,0.0024164012,0.2160612643,0.0114686452,0.107178241,0.4145712554,0.0568224788,0.0054259296,-0.2095331997,0.0201552454,-0.1445862502,-0.0305705536,-0.1140463501,-0.2860078514]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/166","title":"Add a method to shuffle a dataset","comments":"As you might already know the issue of dataset shuffling came up in the nlp code [walkthrough](https:\/\/youtu.be\/G3pOvrKkFuk?t=3204) by Yannic Kilcher\r\n","body":"Could maybe be a `dataset.shuffle(generator=None, seed=None)` signature method.\r\n\r\nAlso, we could maybe have a clear indication of which method modify in-place and which methods return\/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-place. What do you think?","comment_length":20,"text":"Add a method to shuffle a dataset\nCould maybe be a `dataset.shuffle(generator=None, seed=None)` signature method.\r\n\r\nAlso, we could maybe have a clear indication of which method modify in-place and which methods return\/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-place. What do you think?\nAs you might already know the issue of dataset shuffling came up in the nlp code [walkthrough](https:\/\/youtu.be\/G3pOvrKkFuk?t=3204) by Yannic Kilcher\r\n","embeddings":[0.1168218851,0.0479950309,-0.0080725169,-0.2174368352,0.128882423,0.0286900382,0.1363591105,0.3606428206,-0.264738977,0.2979810238,0.0116364621,0.8511587381,-0.3536046445,-0.2320982218,0.2013928443,-0.0704999045,0.1584128886,-0.1035659686,-0.2414909154,0.0523635074,-0.1391384453,-0.3828111291,-0.0782272369,-0.2000364959,-0.1313398331,-0.0940043628,-0.0838561207,0.132300958,-0.1840190589,-0.2935966551,-0.4079650939,0.6355049014,-0.1021856815,0.1419230551,-0.0001081365,-0.2722443342,0.3262639046,0.1310779005,-0.4321397245,0.0446859859,-0.0818533301,-0.0996276066,0.0086366646,-0.1526380479,0.0655293167,0.1338957995,0.1654344797,-0.1464117467,0.2423582822,0.0503101274,0.2091447711,-0.0329092294,-0.1901006848,0.1584738642,0.5000153184,0.4068689048,-0.1012272462,0.0286281444,0.4395242333,0.2293315679,-0.265799284,0.2744675577,-0.0664383993,-0.1814917326,0.3078677058,-0.2879181802,-0.5689676404,0.0742852613,0.0025417558,0.1787862182,0.5037118793,-0.2853350341,-0.3903695643,-0.3526299596,0.330252111,-0.1884994954,0.0199903995,-0.0964175165,-0.0568163805,-0.0657435358,-0.2452473491,-0.1598949879,0.0368060954,-0.0217535887,0.5161419511,0.6055985093,0.2005672455,-0.0033858907,0.2449980676,0.0833965391,0.5854985714,-0.0093415622,0.0330339447,0.299644649,-0.1472086906,-0.2083730847,0.0550040305,0.2948370278,0.2364352643,0.1446129829,0.3270729184,0.2850995064,-0.0413031392,0.0373536497,0.0055521391,0.0381203219,-0.0531010404,0.0549554788,0.2810688913,-0.3015078008,0.231248647,0.0039709494,-0.0490369946,-0.0805971175,-0.0109761739,0.1324440092,-0.1945451051,-0.068498686,0.0361292921,-0.5815837979,-0.0496325232,-0.3245076537,0.1488962919,0.1077939048,0.277036041,0.1407219321,-0.1079227254,-0.048173897,0.1835096627,-0.2393871844,-0.1050535291,0.0258388333,-0.4939265847,0.170787856,0.1737404168,-0.0752229467,0.1390149891,0.1129442975,0.0670285374,0.2374949902,0.0788469091,0.2658469379,0.4057005644,-0.2036512047,-0.4380708635,-0.1334305704,-0.1084037796,0.2805595398,-0.3744524121,0.2528228462,-0.2730659544,-0.2303433716,-0.0922330618,0.1565239877,-0.0969693288,-0.2698227167,-0.1759152114,0.3530564308,0.0626457781,-0.4234461486,0.3546225131,0.0778842717,0.0622802265,-0.0040477095,0.0131624537,0.3176636398,0.0469545834,0.0204990134,-0.2403812259,0.0018366999,0.2378979623,0.2391389161,-0.1998987645,-0.0213025268,-0.1616416276,-0.0612693727,0.5229774714,0.1896463186,-0.2080697417,0.0131860506,-0.209504053,-0.0972790122,0.4091675282,0.3883785605,-0.1224163175,-0.3362308741,0.2075741887,0.1173946261,-0.410466373,0.1484215409,-0.3570763469,-0.0799920484,0.4687559903,0.3144295812,-0.2434637696,0.0008649587,0.0736546218,-0.1594797373,0.2940757275,-0.1337452233,-0.1370373666,-0.2356622666,-0.0692658871,0.1285503656,-0.1717895567,-0.1168628484,-0.1040293276,-0.2373951077,0.223628521,-0.0517794266,0.1164419204,-0.2859423459,0.0874580741,-0.3385697901,0.0740101263,-0.1682889313,0.0914287716,0.115264304,-0.1039150581,-0.1696283668,-0.2285808921,-0.0651362538,-0.1469430178,-0.18812038,-0.3230232,0.029987203,-0.1107577682,-0.0512310565,-0.0759257749,0.4667927623,-0.0837616101,-0.0623995699,-0.0960640013,0.2955295742,-0.0001282156,-0.2677128315,0.3739856482,0.2960723937,0.0336664915,0.0351459384,0.0965519771,0.0336218141,-0.0555981323,-0.0028627121,-0.3500953317,0.3163101971,-0.168394953,0.0289800316,-0.182195738,-0.2597624958,0.0674805865,0.22245875,-0.3342285156,0.1662608534,0.2199173868,0.2118288577,0.2813099325,0.2211573124,-0.4753617942,0.0671244785,0.4927187264,-0.0540447421,0.0184653047,0.3442239463,0.2549565434,-0.3329372406,0.1225965917,0.12657094,0.2766240239,0.4469533265,0.052660808,0.1430922449,0.095129028,-0.1651538163,-0.0443397202,-0.0692203566,-0.1712171882,-0.0281354692,0.2366150171,0.1459546834,-0.2876828909,-0.3273473978,-0.0469350293,0.1891978085,-0.1102173775,-0.0588323995,-0.1330818832,-0.306199789,-0.3187922537,-0.091669634,-0.3395288885,-0.3069963753,0.3917934895,0.0103865359,-0.3207394183,0.4042351246,0.1464307606,0.5348382592,-0.2747852206,-0.0112941647,-0.2897607386,-0.4303383827,-0.0809720308,0.0738233626,0.2886367142,-0.0477829576,0.7124966979,0.0789341927,-0.1327527761,-0.3044414222,-0.5318074822,-0.1408619136,0.031390328,-0.2137417048,0.3421264291,-0.3924229741,0.0901098996,-0.3882688582,-0.1876741946,-0.3428919315,-0.0802419633,0.1229611039,0.000893511,0.0224003438,-0.2999293804,-0.3448785543,-0.0723853931,-0.4497671723,0.0770968124,-0.6393800974,0.0720585957,0.0566139631,-0.2364704013,-0.1664166898,-0.372789681,0.0077646188,-0.1279707998,-0.5969657302,0.1661372185,-0.1103719324,-0.0612819679,-0.2556629777,-0.056040328,0.0138127496,0.6228568554,0.1948454231,-0.1397157311,0.1068795025,0.0764808729,-0.0811647028,0.1502829045,0.434381336,0.1620499194,-0.1234577894,0.1187695414,-0.182643339,0.0421155691,-0.0153984576,0.3286912739,0.034825176,0.2597435415,0.1325586289,0.5310879946,-0.5287257433,-0.0803312957,0.0568569489,0.1760204136,0.2583037019,-0.0953007266,0.0022558759,0.5387940407,0.2906549573,-0.2538530231,0.2548822463,-0.0269761998,-0.309207648,0.1209314317,0.113509208,-0.0975208133,-0.158199504,0.4676401019,-0.3359434903,0.0285512544,0.0540268198,0.122513026,-0.2675397098,-0.1677859277,0.0072456296,0.044103384,0.3468675017,-0.1331419349,-0.4249144495,-0.0456011519,-0.2952165306,0.1951901466,0.2627948821,0.2744700909,-0.0153546641,-0.1793951243,0.1379477233,0.1270139217,0.3991081715,-0.1852071136,-0.1568524092,-0.0833653361,0.1259493381,-0.1888726503,0.0536638573,-0.1135924235,0.1066893116,0.1624179184,0.0592937917,-0.263667196,-0.1953767985,0.1375112683,0.0862831995,0.163312152,-0.3234398365,-0.0784196705,-0.196593374,-0.1358692348,0.2577888668,-0.006181696,-0.2898062468,-0.0326427855,0.0835871771,-0.0508601405,0.2291354984,-0.0087268213,0.1734147966,0.0596777648,0.2108023018,-0.0244449601,-0.0085772062,0.3802026808,0.2015165538,-0.1710920483,-0.3553249836,0.0014369861,-0.0802206621,-0.36159724,0.2507227957,0.21086891,0.2100782841,0.2614562809,-0.1482924819,0.150756672,-0.2337452173,-0.0301890038,-0.0679102316,0.0075330227,-0.5339679122,-0.8507205248,0.5450390577,0.3243733048,-0.3054794669,0.4437339604,-0.0245010555,-0.3562779725,0.3782588542,0.3299740255,1.0034141541,-0.0732178017,0.2827602923,0.1653261781,-0.3735775054,0.4298719764,-0.1853653044,0.2949048281,-0.1998839974,-0.2698773742,0.1041074619,-0.0036482469,-0.0280221403,0.2834885418,-0.2357668728,0.3124027848,-0.168769151,0.0294253919,0.1295360774,0.2141415775,-0.2698855996,-0.2272395641,-0.2418516874,0.1364964843,-0.2273492217,-0.2124517858,0.0214568414,-0.1538651288,0.0076257312,0.0642189384,-0.1773075014,-0.1158239543,0.1199986413,-0.1337777376,-0.1947444528,-0.0684148148,0.5051291585,0.0154245934,0.2500440478,0.0793078095,0.1413497776,0.2305992097,0.0713906437,0.3418623507,-0.0460258573,-0.0362507179,0.1730342209,0.0257415045,0.0010229491,-0.059545856,0.1303381324,-0.6264162064,-0.4741675556,-0.0089476295,0.3204630613,-0.0256152079,0.1331469864,0.236908257,-0.0770763159,-0.0841195807,0.1552939415,0.2397764176,-0.247765854,0.3586952388,-0.1131172329,-0.0368933119,-0.0794554055,0.158766821,-0.2302314639,0.0152254645,0.3005438149,-0.2329336256,-0.1303836703,-0.2351307869,-0.019167589,0.2381070852,-0.0647713393,0.1519560069,-0.1365499794,-0.2242705524,-0.066215694,-0.1361082196,0.0521620102,0.2807179689,-0.0546429753,0.0137363272,-0.39536497,0.0208005272,0.0989729315,0.3021579981,-0.1149687394,0.1467446238,-0.1508698463,0.0054862527,-0.3782982826,-0.2402099669,0.114668414,0.3197236359,0.0731664449,0.1575392485,-0.210562259,-0.218443498,0.0611620545,0.0433409102,-0.0498469882,-0.2375105172,-0.1392947882,0.0682379007,-0.0031458421,0.2037642002,0.1567554623,-0.1276063472,-0.0396701284,-0.2093827724,-0.1145478487,0.1765206307,-0.0192615408,0.5665615201,0.5050375462,0.1023535281,0.0097768484,0.1858239323,-0.1405273825,-0.1123260632,0.0195766427,0.2767688632,-0.0136954617,0.0510714985,0.0243228078,0.2786236107,0.1418535709,0.0893693045,-0.0814656466,0.1931727529,0.1095469892,0.4665201306,0.0618303791,0.358240962,0.0823546052,-0.4855668545,0.4337492883,0.218169719,-0.1612998098,-0.0022095924,-0.1734406054,0.0734911859,0.1654214114,-0.0031532843,-0.1050325334,-0.0687328056,-0.2876114249,-0.0735954493,0.149368301,-0.5486269593,0.2823895216,0.2594433725,-0.1250715256,-0.1608252078,0.2170795351,-0.0982802659,-0.1600296944,0.1616698056,0.3601127267,1.0041590929,0.3350098431,0.2119106054,0.0243654475,0.1491099149,0.0933804959,0.6433224678,0.3787457049,0.1495160013,0.4294312,0.231285125,0.1897774786,-0.3021349907,0.2600722909,0.1661063433,0.1658439487,0.0658090487,-0.1154673249,-0.1193490773,-0.2659814656,-0.0077395556,-0.0538183488,0.1052822843,0.5309208035,0.186180234,0.1603646576,-0.4112891555,0.203455627,0.2584479153,0.470056802,-0.1868491471,-0.2418614626,0.0250696614,-0.2161088884,0.7542926669,0.0419594646,0.0245069489,0.2809267044,0.0492959842,0.0327876695,-0.4148062468,0.0094170477,-0.1502786428,0.1793757528,0.2596071362,-0.1214563325,0.021811489,0.1387145668,-0.1115119681,0.2013954967,-0.0467036217,-0.0709940791,-0.0858621523,0.3180975616,0.5659244657,-0.2598890066,0.0407957174,0.0292307623,-0.136031881,-0.0651412979,0.0066843005,0.1458758861,-0.1700687706,-0.0568631627,0.0926966146,0.352460742,0.4031998217,0.2684459984,-0.1074260399,-0.1138460487,-0.3688149154,-0.5009536147,-0.1890977621,0.3571243286,-0.051589638,-0.0281946808,-0.0222725067,-0.0188507978,-0.0188997742,-0.0447733589,0.0354644619,0.0138454894,0.070772849,-0.3659392297,-0.0138913551,-0.2021807134,-0.073253125,0.1431048065,-0.1260943115,-0.2787936032,0.3147032261,0.0446381196,-0.22550264,0.2509476542,-0.0806992874,0.146020934,0.0339386463,0.1352876872,0.18925412,0.0494482182,-0.113702029,-0.1537812352,-0.1095562801,-0.6963852644,0.0051313648,0.0585892014,0.0548669584,-0.0691279545,-0.1053768694,-0.411696136,0.5068254471,-0.2618786991,0.081107676,-0.582264781,0.174057588,0.1921132654,0.1904583424,-0.1475059837,0.2194094658,0.0977318361,0.18745552,-0.2305743098,0.1007898077,0.4633552432,0.0216554832,-0.1791426837,-0.2164991498,0.2422753274,0.0125419432,-0.1401490271,-0.3644134998,-0.1711257994,0.1759828627,-0.0416666791,0.2560968399,0.4652126729,0.0460455343,-0.0463175401,-0.1834834814,0.0035771886,-0.0613768436,-0.0453562103,0.0050765024,-0.4476947486]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/166","title":"Add a method to shuffle a dataset","comments":"We added the `.shuffle` method :)\r\n\r\nClosing this one.","body":"Could maybe be a `dataset.shuffle(generator=None, seed=None)` signature method.\r\n\r\nAlso, we could maybe have a clear indication of which method modify in-place and which methods return\/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-place. What do you think?","comment_length":9,"text":"Add a method to shuffle a dataset\nCould maybe be a `dataset.shuffle(generator=None, seed=None)` signature method.\r\n\r\nAlso, we could maybe have a clear indication of which method modify in-place and which methods return\/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-place. What do you think?\nWe added the `.shuffle` method :)\r\n\r\nClosing this one.","embeddings":[0.0458720177,-0.1113289595,-0.0959766954,-0.1793048233,0.2690799236,-0.0499059111,0.2569342554,0.3070525825,-0.2576756477,0.3601591885,-0.0516567677,0.7354241014,-0.2381035089,-0.2504498959,-0.072288245,-0.0192962978,0.1153434291,-0.1139178053,-0.1771237403,-0.0670236126,-0.1373585165,-0.2984468639,0.0276970826,-0.1496905833,-0.123050563,-0.0893146843,-0.0748811588,0.1313046962,-0.2035017163,-0.2265138924,-0.5108004212,0.6731533408,-0.2169247866,0.3351395726,-0.0001030913,-0.2821241617,0.3256727457,0.0078621283,-0.3106943667,0.2446469069,-0.1359430552,-0.0129044456,-0.0393867232,-0.1773285717,0.0327871703,0.004251658,0.0577146113,-0.1753917485,0.2576735616,-0.0251169987,0.2756093144,-0.075944826,-0.233118549,0.0283510331,0.603715837,0.3594251275,-0.1877135336,0.0461156256,0.4734654427,0.3093492985,-0.2634647191,0.306045711,-0.1440658271,-0.0703306347,0.4136994183,-0.2444642782,-0.4632970095,0.0632318705,0.0932609886,0.1565860957,0.4835579395,-0.1425240934,-0.3151530623,-0.1922137737,0.2000628263,-0.1778693199,-0.0293872021,-0.1784840077,-0.078160435,-0.0787242353,-0.1996803582,-0.1762767285,0.0138564566,-0.0886381939,0.4327900112,0.5116638541,0.1304726601,-0.0603986196,0.1898352057,0.0427527577,0.4852651954,0.0315047167,0.0097037954,0.1006228402,-0.0293277744,-0.1452980489,0.0917987004,0.2134477645,0.2611967325,0.1682271659,0.2666303813,0.3548190594,0.061765071,0.1140441149,-0.0193893202,0.021260608,0.0348588862,-0.0446445383,0.3571183085,-0.2392663062,0.0672773197,0.0059342245,-0.0659623891,-0.0958458707,0.0809947699,0.2291942835,-0.1833470762,-0.0774214491,0.0853422731,-0.4057099223,0.0186323188,-0.3623667955,0.0698446631,0.2454483956,0.2203303725,0.1025942415,-0.1053867489,-0.0386477262,0.2153200805,-0.1496504992,-0.1524930745,0.0329053067,-0.4822015166,0.1831743717,0.0852787271,-0.1832133532,0.1518912315,0.1000820026,0.0952294841,0.1769210696,0.1776031256,0.2300525904,0.4096446633,-0.1615964025,-0.3584439754,-0.1020788848,-0.0415987186,0.256916672,-0.2985394299,0.2519503236,-0.1761252731,-0.2260220796,-0.1386569142,0.2259449512,0.0193331838,-0.0946680978,-0.0350209139,0.2623944879,0.1074043363,-0.302228421,0.2929519415,0.1465701461,0.0766411349,-0.0758407637,0.0747128725,0.2707820535,-0.0625017062,0.1036330014,-0.209531948,-0.0757320076,0.2195853293,0.1947757304,-0.2059274912,0.0018352746,-0.1609993428,-0.0976684317,0.3933225274,0.273329258,-0.1918969452,-0.0319990776,-0.2216685712,-0.0706300363,0.4052089453,0.2933708727,-0.0902817547,-0.2705168724,0.313280642,-0.1059389636,-0.3940593898,0.1157250553,-0.4220017195,-0.1273064315,0.3307253122,0.193473801,-0.0720081925,0.0136185708,0.015444125,-0.3322635591,0.2917464972,-0.1880687773,-0.1288532764,-0.1772664785,0.112498492,0.2444533706,-0.1128949299,-0.125304386,0.0034814668,-0.1652956754,0.2376590371,0.0023675889,0.0588112958,-0.3837297261,-0.0313200429,-0.3312363327,-0.0106937876,-0.1794888675,0.1726694405,0.2711069882,-0.2710857093,-0.2714470625,-0.1912030876,-0.1888070256,-0.145286411,-0.2375813723,-0.2908133566,-0.1112392545,-0.1780436933,-0.001494133,-0.1518722177,0.247931391,0.0221056547,-0.1080174074,-0.1297329962,0.3100733161,0.0200686101,-0.168069005,0.3385531604,0.2982184291,-0.1091284677,0.0325255096,0.1076689959,0.0296664387,-0.0601701736,-0.0386882238,-0.3171703219,0.3175130486,-0.1023693681,0.0087013608,-0.2588205636,-0.2463442981,0.0494660959,0.3029254973,-0.3277424872,0.0913990512,0.0678253323,0.1817773581,0.2358746678,0.2082927376,-0.4082450271,0.0393293165,0.4428701699,-0.0699161068,0.0639063269,0.3319459856,0.1785602421,-0.2798582017,0.0846599042,0.0703862011,0.2293777019,0.4606849253,0.0729646906,0.0589890182,0.1669765264,-0.0820399001,0.0299253911,0.1010418534,-0.1694451869,0.0362370051,0.1699336469,0.059341874,-0.3282831609,-0.3209476769,-0.1130403578,0.1082758754,-0.0832009017,-0.0705521777,0.0088607343,0.0027428996,-0.1294484288,-0.1682452261,-0.200764969,-0.2878943384,0.4283651412,0.0980331153,-0.4116387069,0.4074201584,0.2210422009,0.5010541081,-0.1184359416,-0.0439946428,-0.344979465,-0.457796216,0.0623888336,0.144869715,0.2936738133,-0.0844372213,0.6192545891,-0.0490583591,-0.0403005518,-0.291495353,-0.4490326643,-0.0490827374,-0.0484979078,-0.1428057551,0.3809469342,-0.1830634326,0.0440053865,-0.4622418284,-0.0147941466,-0.3303591013,-0.0509017296,0.0261138007,0.0398266129,-0.0306528546,-0.2689436972,-0.4190496802,0.0758364573,-0.4857862592,0.0235373061,-0.5937173963,0.0080266213,-0.0035376952,-0.1406791657,-0.1888632029,-0.2334652245,0.1556701362,-0.2573850751,-0.6907612681,0.2910352349,-0.196481213,-0.0264577325,-0.1644118577,-0.0766239092,0.0306041911,0.5531592369,0.2581907213,-0.1571707129,0.0837730467,0.111499697,-0.120839037,0.0822467208,0.3833498955,0.329185307,-0.2680970132,0.0748319924,-0.2547262013,-0.0792996064,-0.113390021,0.2468069643,-0.1025898308,0.3798575401,0.0335110649,0.711714685,-0.6103314757,-0.1152564436,0.0468555279,0.0940176621,0.2336754948,-0.0823929459,-0.051362019,0.4309771955,0.2893259823,-0.2759527564,0.2385007292,-0.0594678447,-0.2445783317,0.0448974147,0.1597354263,-0.1669537723,-0.2382713556,0.3688036203,-0.2495758832,0.0105332267,0.0403970331,0.2668389678,-0.2186738104,0.0334969051,0.1496240199,0.0664243475,0.3061052859,-0.2413841039,-0.4083221555,0.0270522106,-0.113190271,0.212299183,0.2656019032,0.1117632315,-0.0462309495,-0.2907865942,0.0858518258,0.0410959758,0.4864780307,-0.2216674685,-0.2473633885,-0.0131935058,0.1039134488,-0.0765640885,0.0132250879,-0.1325265318,0.1276209801,0.0787512287,0.0638727993,-0.2494671345,-0.1128378361,0.0950650871,0.1593630612,0.1022979394,-0.2298390567,-0.2004694343,-0.2245715857,-0.2338870466,0.2995217741,0.0509421676,-0.2688263059,-0.1099589765,-0.0809146315,0.0252491534,0.2271619737,0.0272643063,0.2057647258,0.1570833027,0.1094299182,-0.0214469135,-0.0453927368,0.3104555905,0.3021821976,-0.0980833471,-0.3113131523,0.0746936724,0.0038678655,-0.5483671427,0.1199709624,0.2100631893,0.1964723468,0.1894735992,-0.1512186825,0.2807449698,-0.1430957913,0.0134094637,0.0173788331,-0.0155003853,-0.3659232259,-0.7962211967,0.470888406,0.2037581801,-0.3456250429,0.5366685987,-0.0224450212,-0.186042577,0.3972584009,0.2950812876,1.1401792765,-0.1050652564,0.1631899774,0.0408772118,-0.2971175909,0.1839446872,-0.2163823545,0.193524912,-0.1543485075,-0.3335132003,0.1145294085,-0.0144875329,-0.0331197903,0.2083741874,-0.2277849168,0.2560117543,-0.3314750791,0.1329427212,0.02948175,0.122382246,-0.2428648919,-0.2060503513,-0.1010179892,0.1695029289,-0.0830130652,-0.1980883628,-0.0494202226,-0.1433557272,-0.0803649649,0.0635742247,-0.125691697,-0.1733994633,0.1166292503,-0.1203736737,-0.2982142866,-0.1052981541,0.2738404274,0.0302976407,0.2439472228,0.1953489482,0.1067066118,0.2617944777,0.1603035182,0.4249142408,-0.1274801493,-0.0844669417,0.0961899161,0.0145056974,-0.0625255406,0.008857172,0.0690865815,-0.5279789567,-0.4174118936,-0.0481780171,0.4577102959,-0.0827749521,0.2169125825,0.1231351346,-0.1400517225,-0.1202400774,0.2246676236,0.2581078708,-0.2062673867,0.3084848821,-0.0699914396,0.0333280154,-0.1123146489,0.1934687793,-0.242219016,-0.0544280075,0.2033987939,-0.1974670142,-0.2406316847,-0.3222791553,0.0771834478,0.2134777904,-0.007521362,0.0817430392,-0.1959887892,-0.3431714475,-0.0306074247,-0.0338758565,0.1607040465,0.3321942687,-0.0565897524,0.1826602668,-0.5228874683,0.0437905006,0.1080434173,0.3851551414,-0.1446839273,0.0019476218,-0.099652119,0.0108968941,-0.4797002077,-0.290413022,0.0699096173,0.1782238334,-0.0175409447,0.0977740213,-0.1665294021,-0.2092351615,0.2050605118,0.1128353849,-0.1186860055,-0.2794760466,-0.058349397,0.0648306981,-0.113827236,0.3230375051,0.1987977326,0.0364881828,-0.0909192786,-0.2411839217,0.0523448586,0.2807867825,0.0412652194,0.4985972643,0.5115523934,0.0605928339,-0.1233784482,-0.0307783876,-0.0558955707,-0.054991696,0.0627696961,0.049337551,-0.0110596679,-0.1158013493,0.235854432,0.3357754946,0.1575369835,0.1658149362,-0.0860576481,0.2753851712,0.0273284409,0.4192214012,0.1496044397,0.363563925,-0.0514015146,-0.4441314042,0.3230185211,0.3578450382,-0.1724584848,-0.0882251635,-0.3139410615,-0.0758354068,0.2143248022,0.0370793566,-0.1355959326,-0.1052345186,-0.1679296494,0.0371065028,0.100303866,-0.6267274618,0.345285207,0.2603649497,-0.155790776,-0.150020808,0.3576497436,-0.0697030127,-0.1025049761,0.2092477381,0.4060266614,0.9255943894,0.2791836858,0.2084614336,0.0321313925,0.1530416161,0.198273927,0.619907558,0.3144522607,0.1797328293,0.4180613756,0.1720528901,0.1356167197,-0.3263702989,0.1702260226,0.094355315,0.1205515563,0.0034349947,-0.0265618805,-0.1899600029,-0.3044843674,-0.087004967,0.0391031764,0.1868640929,0.4560989439,0.1400663257,0.1640470028,-0.2780999541,0.1627019048,0.140650481,0.4945070744,-0.1901016533,-0.3359673321,0.1979904771,-0.281247586,0.6002868414,0.0131584229,0.0247540735,0.3014058173,0.0278385151,0.1445844918,-0.3272437155,-0.0836513042,-0.1674958766,0.0331129283,0.1577154249,-0.15763098,0.043819245,0.2565824687,-0.2684271634,0.1948699802,-0.0821619928,-0.008993878,-0.0448601395,0.3007904589,0.5004149675,-0.3531553447,0.0630802661,-0.1016678959,-0.2546260953,-0.1425479949,-0.037968006,0.1400996596,-0.1183824688,-0.0356023461,0.140039593,0.2927168906,0.3128968775,0.1572071463,-0.1014034897,-0.0957251191,-0.3603276014,-0.5620921254,-0.0379945412,0.3479022384,0.044967141,-0.0025432736,-0.0492276289,0.1054176167,-0.0368320644,0.0732586384,-0.104358606,0.0425966121,0.1126810461,-0.3693031967,-0.0228185784,-0.0822516084,-0.1051378399,0.2595389485,-0.1527068168,-0.2160933763,0.2289402187,0.2336125374,-0.2091847211,0.1175979972,0.0276017077,0.1568848193,0.2222479135,0.0208031405,0.1658019871,-0.1142530441,-0.0862495378,-0.0707385093,-0.0983880311,-0.5242286325,0.0274180155,0.1608378142,0.1392528862,0.0077225422,-0.2074720263,-0.4588046968,0.4828262031,-0.1580882519,0.1497069895,-0.5681374669,0.2545708716,0.2800040245,0.130523935,-0.1745940298,0.0943967327,0.0392215699,0.185761407,-0.3306389451,-0.0220087226,0.5680433512,0.1930760741,-0.2984186709,-0.223488465,0.3325754404,0.0683132783,-0.0310588516,-0.256616801,-0.0925081074,0.1692553014,0.0382529609,0.0141012836,0.3710389733,0.0480106957,-0.0711416304,-0.1527274251,0.0235366896,-0.0390283279,0.0262994803,-0.1156835556,-0.4015509188]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/164","title":"Add Spanish POR and NER Datasets","comments":"Hello @mrm8488, are these datasets official datasets published in an NLP\/CL\/ML venue?","body":"Hi guys,\r\nIn order to cover multilingual support a little step could be adding standard Datasets used for Spanish NER and POS tasks.\r\nI can provide it in raw and preprocessed formats.","comment_length":12,"text":"Add Spanish POR and NER Datasets\nHi guys,\r\nIn order to cover multilingual support a little step could be adding standard Datasets used for Spanish NER and POS tasks.\r\nI can provide it in raw and preprocessed formats.\nHello @mrm8488, are these datasets official datasets published in an NLP\/CL\/ML venue?","embeddings":[-0.143006593,-0.0090511432,-0.2522800267,0.0832095817,0.0416293368,-0.2631311417,-0.0207332764,0.0178046189,-0.0471933149,0.0691099688,-0.2256437987,0.2731508911,0.0523054563,0.2312858552,0.1396114528,-0.2148642987,0.0306870826,0.1047005653,-0.0000387967,-0.0829317048,-0.0544710681,0.2502807677,-0.0041145538,-0.0214006323,0.1294011921,-0.0855903625,-0.3412953913,-0.208059743,-0.3143412173,-0.3122626841,0.0042263241,0.1638319641,0.2610605955,0.0381873287,-0.0000991146,-0.2244775593,0.0253702458,-0.1137114614,-0.1451014578,-0.2402695566,-0.1693311334,-0.4515506029,-0.2404995412,-0.0445651151,-0.3123645782,-0.1802634448,0.1711746752,-0.1662742943,0.1782662123,0.4173791111,0.2644209564,0.0874056369,-0.0986619815,-0.2346584052,0.1405193061,0.3721913993,-0.14551799,0.1298943907,0.6116508245,-0.2004995495,-0.1290588081,0.2287990451,0.2066670358,-0.2706539631,-0.2570709586,0.0136854658,0.2845196128,-0.1680201441,0.1508240253,0.1731949896,0.0417536683,-0.0878835544,-0.1548787653,-0.101187095,0.0769763142,-0.2525152862,-0.1783319116,0.1712638438,-0.0455089062,0.231522873,0.3131119013,-0.3609355092,-0.0989178047,0.1965614408,-0.0038683484,0.0884288251,0.0410908349,-0.3410513103,0.066008985,-0.0585195944,0.3313038349,0.0296394099,0.0085169161,0.0905824006,0.0087466501,-0.0238047745,0.2353140712,0.2270486355,0.1045116335,-0.3004828095,0.252443403,0.1131100282,-0.0122764381,0.1072466522,0.1091026738,-0.2366190255,0.1779432297,0.148265332,0.036857713,0.0378463529,0.0468869247,0.0531071275,-0.0777294487,0.074743852,-0.2474761456,0.1774173081,-0.0275466945,-0.3670156598,0.1452553868,0.1427853405,0.0410597175,-0.1859596819,-0.0795085877,0.0769523755,0.0487770513,0.4059741199,-0.2991338372,0.1527245045,-0.1696414948,-0.4657156467,-0.100618206,0.1169296429,-0.4266336858,0.0832870677,0.3467898071,-0.0424930602,0.0842467993,-0.059175089,-0.2422454208,-0.0403479375,0.1903895587,-0.3815797269,0.1385026574,-0.2592106462,0.162951842,-0.1275840402,-0.3558652997,-0.0568048805,-0.1388413161,0.1386435181,0.0167093799,0.026344711,-0.3163227737,0.282365948,-0.0713531226,-0.2137777656,0.1418922991,0.6137128472,0.0468066372,-0.3149355054,-0.0659628063,-0.0513324253,-0.3176127076,0.0364080407,-0.0271708891,0.1505833417,-0.3354068696,-0.0871383771,0.0148240561,-0.1112109274,0.4288019836,0.298933357,-0.0076196375,0.1086558253,-0.1924104989,0.2521306872,0.2096559703,-0.1990686953,0.1448341906,0.0441987924,0.0345337391,-0.1596515775,0.1586140245,-0.1308264583,0.1507871598,0.0609863326,-0.0381002165,0.5285031199,-0.3202125132,-0.0564006791,-0.2278119922,0.1684605181,0.0756980255,0.4538359642,0.3122705221,-0.2184244692,0.1589494944,0.1764149368,0.0418590195,-0.2708284855,0.1077465862,0.0921725109,0.2431602031,-0.1303712279,-0.0048149996,0.0519852415,-0.0888029933,-0.2818301022,-0.3815155029,0.6595309377,0.027558133,-0.0691711903,-0.1041956544,-0.2359968722,-0.0004314713,-0.245372057,0.3663835227,0.1813619733,-0.0511360019,0.1958911717,0.1255218685,0.1733154804,-0.1701458097,-0.1079499498,-0.1024827659,0.3225261867,-0.2212775201,-0.1153662428,-0.120844841,0.440110445,0.0433207117,-0.0142631363,0.1320908666,-0.0906328037,0.145729363,0.3292986453,0.3170951903,0.0582115948,0.3087044358,-0.1627863497,0.0689331964,-0.0432092734,-0.0442487411,-0.0902183354,-0.1139038652,0.3719649911,0.1756590307,-0.0926986784,0.1530174613,-0.0777414069,0.2879241407,-0.0855304524,-0.167522639,-0.1456148177,0.4388900101,0.1150299385,-0.0723340213,0.287607491,-0.5719982982,-0.0274918973,0.3068737388,-0.0320215374,0.0602484271,0.0370295569,0.0363343768,-0.1086919382,0.2047660798,0.202158317,0.1700147092,0.3160887957,0.2807499468,0.041689679,0.1061834171,-0.2038034797,0.1172531918,-0.1276620626,0.0452269055,0.318459034,0.1265399903,-0.2389170527,-0.2926153839,-0.1135743484,-0.0906027183,0.1344220191,0.0535526276,-0.337710768,0.0161958095,-0.4229514003,-0.3033912778,0.14870134,-0.0665687993,-0.1271905154,0.3854347467,-0.2660675347,-0.0267704334,0.1679171026,-0.206177488,0.4689273834,-0.6129950285,0.1174631044,-0.0304129981,-0.3470191956,-0.1204907,0.2939471304,0.2659576833,-0.0583531,0.2323165983,-0.3808419704,0.1118677258,-0.040818166,-0.4948528111,0.2193280756,0.0650022253,0.0504040942,-0.1423784345,-0.2462464571,-0.0032392885,0.1141014099,-0.0096079558,-0.1017249674,-0.2147962004,-0.0218260754,-0.4776009917,0.0586833917,-0.2287277132,-0.6140753627,-0.515227735,-0.2934901416,0.0544960052,-0.1253526658,-0.0351013988,0.1183950454,-0.0393530875,-0.030009713,-0.1777374893,0.0546948723,-0.0884526595,-0.0652174205,0.1338513792,-0.322326988,-0.3049965799,-0.3332341611,-0.1268603206,0.4985563159,-0.0012752914,-0.2010248154,-0.0094697447,0.1774142236,0.1733534038,0.0597141199,-0.0136672538,0.2748804688,0.3685227633,-0.2405869365,-0.0435953885,-0.0810153335,-0.0317222625,0.2425564229,0.1355540454,-0.1717797518,-0.2542401254,0.1797935218,0.2427062839,-0.2982928753,-0.1300873309,0.454518944,0.1503161341,0.2184018344,-0.0298409238,-0.1972079873,0.0987100974,0.1361500472,0.1330411732,0.3518277109,0.0685360357,-0.6104001403,-0.3022808135,-0.1277559251,-0.3303301632,-0.0357583389,0.4287549853,0.0105967931,0.5536217093,0.0615678392,-0.23572357,-0.1912198812,-0.2424422204,-0.1048907191,-0.2979130745,0.0589422472,-0.1008454487,-0.2829050422,0.16203475,-0.1701632291,0.061140351,0.0146997366,0.1294210553,-0.3688777387,0.098966971,-0.0184760112,-0.0107067442,0.1287188381,-0.3316444755,-0.3810409009,-0.0254909266,-0.109295845,-0.065263778,0.1591274291,-0.487382859,-0.1219205931,0.2091849148,0.3610112071,-0.123977825,-0.1723548025,0.0418494008,0.0522403158,-0.0065425676,-0.2761469781,0.3122378588,-0.1895149946,-0.0306946672,0.1849955618,0.1741686761,-0.0704402775,-0.0069143418,-0.0489536561,0.1624384373,0.4874535203,0.3499164879,0.2081669718,-0.128394261,0.1858950704,0.0136729321,-0.1225124523,0.1358325779,0.3224493265,0.2043156475,-0.2167859972,-0.0698226988,0.1799222231,-0.5017117262,0.5630519986,0.2066629678,0.4713095129,0.2249850184,-0.0082393773,-0.2903646529,-0.0387165733,0.0210945774,0.2172367126,0.2239994109,-0.2145444155,-0.1925316155,-0.393582046,0.3166583478,-0.1720026135,-0.3212892413,0.0355002396,-0.2520988286,0.1025164202,0.222788468,1.0366873741,0.1154183447,0.102483578,-0.1327446848,-0.4121939242,0.1060788333,-0.3653350472,-0.1612376124,-0.0995022804,0.0620779395,0.0573371165,0.1271192133,0.0476022512,0.2004494667,-0.3734266162,0.3662820458,0.3384810388,-0.1908457428,0.09569446,0.3433718383,0.1259907186,-0.4679020643,-0.1131012365,0.1855319142,-0.2043593824,-0.1486060023,-0.0475723483,-0.1853172928,0.1587606817,0.1019115001,-0.1210927665,0.0961605683,-0.044771865,-0.199732691,0.0399180725,-0.0319937244,0.19302091,0.2675754726,0.2294471413,0.033350341,-0.0446208455,0.0400579832,0.1188934296,-0.0537821725,0.1806925684,-0.1348441988,0.179108277,-0.3347353339,-0.2110340297,-0.0653284118,0.1150636971,0.4119949639,-0.4447442591,-0.0767236203,-0.0684260204,0.0260339454,-0.0076941024,0.217745766,-0.1099289581,-0.3135855198,0.3005534112,0.1244731545,0.1145556569,-0.004991198,0.2611054778,-0.0402969643,-0.3660694957,0.0054865102,0.2639731169,-0.278906852,0.1675572991,0.1446138322,-0.1880043596,-0.3435233533,0.0292545278,0.1816164553,-0.4714886546,-0.3672687113,-0.0112134041,-0.2632039189,-0.113177903,-0.0013523641,0.0903703347,-0.0478350595,0.0832123756,-0.0334391706,-0.1888912022,0.1157691479,0.0362545624,0.0369236618,0.126868844,-0.223065868,0.0739578381,0.1579575688,-0.4613892734,0.0439845584,-0.2215579003,0.2725712955,0.1843876988,-0.0447042175,-0.1710215062,-0.024264425,0.2180684209,-0.0277313609,-0.0674924403,-0.3183730543,0.0254878979,0.0535202846,0.1064432114,0.2966007292,0.1798415333,0.1017875075,0.0541050509,-0.0577132329,-0.2903987169,-0.0613105558,-0.108088918,0.536374867,-0.1020564884,0.2667411566,0.2713515759,0.4112454653,0.1008100137,0.0591166951,0.0996016413,0.2310642302,0.1646936387,-0.1674426198,0.2192317694,0.110320285,0.3566216826,0.3138433397,-0.0803269148,0.1561998725,0.0961584002,0.0708414093,0.0675859377,-0.0817739516,-0.0260302182,0.1372433454,0.2849724293,0.3766402006,-0.2422289103,0.2925322056,0.4111239612,-0.0826985985,0.2797631025,0.3189455569,0.4042599499,0.2117823213,0.4376193583,-0.0297512375,-0.2826297581,-0.1582306325,-0.1299103945,-0.0220429525,0.0450875573,0.0483143888,0.0752239227,-0.0152289504,-0.1651907712,-0.0595459044,-0.08264485,0.427547127,0.3114112914,0.3471360803,0.3330835402,-0.0267302953,0.1845517606,0.1576726288,0.2611514032,0.1258048564,0.1330276728,0.5818778276,0.3300306797,-0.2803665698,-0.0158362351,0.0952820554,-0.1988240629,-0.1609638184,-0.345205009,-0.2405385971,-0.4255211353,-0.1161499545,0.0684121847,0.1044278741,0.1367599219,-0.0325922519,0.057713896,-0.1902481765,-0.0128305685,0.0090047065,-0.0564853251,-0.1864143461,0.4270148575,-0.1956822723,-0.0917207003,-0.1032248288,0.4831185341,0.3510151207,0.0620273799,-0.0308663603,0.0075436826,0.1229149923,-0.0190819874,-0.0618560873,0.3442749381,0.4006114304,0.3853791356,0.2640568018,0.2232484221,-0.2405479252,0.1797632724,0.1821371466,-0.2630798519,0.2401544005,-0.0111576831,0.3677970767,-0.1435966343,-0.2261284739,-0.1060877144,-0.4646854699,-0.1056542844,0.2733241022,-0.0401350632,0.0265045576,-0.1865986437,0.1671245098,0.1845276505,0.2202326804,0.3818471134,0.4262036681,-0.3077747822,-0.2591087222,-0.307837069,0.1897766441,-0.0741146281,0.1963239759,-0.3796331882,-0.1348987371,0.5004114509,0.2391741276,-0.2896980643,-0.1964026541,0.1697278172,0.361874491,-0.6076230407,0.105828315,0.2571285665,0.149852246,0.0532950312,0.1447021961,-0.1033925042,-0.1104070619,0.2391902208,-0.1947022825,-0.0093088103,-0.0606561191,0.4083637893,0.1375070065,0.0444738083,0.393605262,-0.2914887965,0.0620943718,0.0186926033,-0.1759134829,0.0100105656,0.2889062762,0.3557102084,-0.0736406669,-0.192803517,-0.3682310879,0.0713530034,0.0040082149,-0.1581542194,0.1706050634,-0.0817189962,0.0973616838,0.2086880505,-0.2495108396,-0.2872353196,-0.0130619342,0.060735479,-0.3488255143,-0.1551560611,-0.2538079917,0.0984101668,0.0904398561,0.014035861,-0.1843362302,-0.1579546034,0.168171972,0.1059111953,-0.4187383652,0.163714394,0.1904732287,0.0474443473,0.053321939,0.189474836,-0.0012616488,0.0536160469,-0.2121588141,-0.1412542015,-0.0229697246,-0.0637846068,-0.1599192768,-0.1812165231]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/164","title":"Add Spanish POR and NER Datasets","comments":"What about this one: https:\/\/github.com\/ccasimiro88\/TranslateAlignRetrieve?","body":"Hi guys,\r\nIn order to cover multilingual support a little step could be adding standard Datasets used for Spanish NER and POS tasks.\r\nI can provide it in raw and preprocessed formats.","comment_length":5,"text":"Add Spanish POR and NER Datasets\nHi guys,\r\nIn order to cover multilingual support a little step could be adding standard Datasets used for Spanish NER and POS tasks.\r\nI can provide it in raw and preprocessed formats.\nWhat about this one: https:\/\/github.com\/ccasimiro88\/TranslateAlignRetrieve?","embeddings":[-0.191217944,0.0006991289,-0.2643911541,-0.0178482644,0.1087873951,-0.2409905046,-0.0053022187,0.1182187572,-0.082876578,0.2340667993,-0.1802721769,0.1794563979,0.1459223181,0.2148519605,0.0853017345,-0.1813978106,0.0933858082,0.1108024269,0.0317240842,-0.0090871155,-0.1936511844,0.28742522,-0.0046590213,-0.0358820967,0.1478987187,-0.0105939014,-0.2947510779,-0.1110509709,-0.2369925082,-0.4156005979,-0.0936100855,0.1999781728,0.1560342163,0.022758523,-0.000096074,-0.079531081,-0.0174356103,-0.0286419392,-0.1398783028,-0.209782362,-0.1433549374,-0.4797947109,-0.2400019467,-0.0071786297,-0.2654830813,-0.1426571757,0.1724170893,-0.2112307996,0.261957258,0.3571873605,0.3139199018,0.0622405596,-0.0719801635,-0.1974152923,0.180973947,0.1571241319,-0.2207831442,0.1020912081,0.6522237659,-0.151423946,-0.145752728,0.3406205475,0.2352037281,-0.2694533765,-0.2711609006,-0.0143711921,0.3385477364,-0.127292186,0.1533605903,0.1494920254,-0.001076161,-0.040044494,-0.2017580569,-0.0929330364,0.0281650424,-0.2526643872,-0.1445648968,-0.0555564575,-0.0526205115,0.1953231692,0.2645739317,-0.4479946196,-0.1413099766,0.1212937236,-0.111315012,0.1267622709,0.0246785786,-0.3284650743,0.1279512197,-0.0176952202,0.1407421529,0.1714019626,0.0257145874,0.058196269,-0.1128908023,-0.0305798054,0.151747182,0.0074488786,0.2233203799,-0.2478235066,0.0948031396,0.2623134553,0.0729472861,0.1081124619,0.0528857112,-0.0444431119,0.0986879542,0.0504992679,0.100139603,-0.0063880319,0.0927999988,0.0312839188,-0.1034927368,-0.0929971412,-0.2633579671,0.2186910659,0.0231515504,-0.3192613423,0.1208675876,0.0599859692,0.0035948358,-0.0680763125,-0.1119800881,0.1072708741,0.0645975322,0.4673243165,-0.2355404645,0.1833821237,-0.1531332433,-0.5209925771,-0.1398406029,0.031862434,-0.4086742699,0.1721363664,0.1910591125,-0.0166827254,0.0829013139,-0.0710250661,-0.0656883419,-0.1739233732,0.2193692774,-0.270797044,0.1148942038,-0.3110767305,0.182417959,-0.1716168523,-0.2857591808,-0.1005441919,-0.1046765521,0.1424079239,-0.0244728774,0.0929296389,-0.3216660023,0.3085467815,0.0312575735,-0.2026760131,0.2229219228,0.652695477,-0.0231272634,-0.2766310573,-0.1241633669,0.0214018039,-0.3381906748,0.0379339196,0.0323750935,0.2505966127,-0.3059368134,-0.1237235665,0.0367226526,-0.0937586799,0.3354689777,0.2338565439,0.0592349768,0.0558500588,-0.1879372299,0.3529515564,0.2870531082,-0.2460926175,0.1174822599,0.0842940211,-0.1256451011,-0.1420713365,0.13545008,-0.0930180103,0.1932400912,0.0594749264,-0.0825201944,0.5718033314,-0.3723745942,0.0881439447,-0.2334652692,0.1546979547,0.0261800289,0.4478884637,0.3149843216,-0.135499388,0.1680226922,0.0868883803,0.0340016782,-0.2924573123,0.0945204198,0.0402635075,0.1975890249,-0.0693419725,0.0673200935,0.0076068202,-0.1411930025,-0.2229871601,-0.3773127496,0.5127241015,-0.0017178026,-0.0766158253,-0.127502799,-0.2426762879,-0.1060349643,-0.1437954903,0.4347969294,0.2632839382,-0.1500488073,0.1415145248,0.1752725542,0.1418909729,-0.0569556467,-0.0008685784,-0.1520911306,0.2122703344,-0.223951146,-0.0433822684,-0.1100988761,0.3125441372,0.0477915034,-0.0831329077,0.0973743126,0.0818456933,0.1645014137,0.3448532224,0.2613890767,0.103098765,0.2898136973,-0.1823209971,0.0778102502,-0.0822813362,-0.0977216661,-0.1300687939,0.0485114418,0.3489822745,0.1833806485,0.0182960927,0.1048964113,-0.0026169796,0.3516511023,-0.0817278624,-0.1567896008,-0.1958138198,0.3224622309,0.0590821467,-0.0945297107,0.0909565762,-0.3604397476,0.0097860219,0.3633840382,0.0331506059,0.0424017869,0.0781536326,0.0131209614,-0.0897200704,0.0725650489,0.1782468408,0.0366278067,0.3174859881,0.3299797177,0.1656823307,0.1734041274,-0.133462891,0.1158191562,-0.1751636863,0.0856040716,0.2831489444,-0.0040210504,-0.224911347,-0.2062587887,-0.1928686649,-0.1051993296,0.1140534878,0.0591056012,-0.2844158709,-0.0642350763,-0.3342674375,-0.2575910389,0.1171286553,-0.1262388229,-0.1708521247,0.4256019592,-0.2854487598,-0.0789101422,0.2865056396,-0.0510235876,0.4205230772,-0.5481014252,0.0357153676,-0.0252305362,-0.3674344122,-0.0989745036,0.2924828827,0.231793493,-0.1396892369,0.224928692,-0.4759376645,0.1070359945,-0.0357980393,-0.5442329645,0.2452809364,0.0565462224,-0.0345993303,-0.0553039275,-0.1955514699,-0.0000148164,0.2138851434,-0.0084255869,-0.1326287836,-0.3071199954,0.0111019462,-0.324796617,0.0318045728,-0.2350537777,-0.5595866442,-0.3729475737,-0.3813715577,0.104381755,-0.2052758187,-0.0495990813,0.13202326,0.0458339378,0.0161921754,-0.180601269,0.109592922,-0.1230973452,-0.007346137,0.0784930661,-0.3188099563,-0.2209194303,-0.2186207622,-0.0245840307,0.5036639571,0.1513725668,-0.0876968876,-0.0307691917,0.1748484522,0.2045455426,-0.0207661223,0.0128496028,0.4074030817,0.3471804261,-0.244155556,-0.0293131415,-0.1515554339,-0.085643284,0.2144882232,0.0591734387,-0.1634095311,-0.2074870765,0.1704758406,0.3465434909,-0.2696989179,-0.0545462295,0.3982076347,0.0778535381,0.1995124072,-0.015676938,-0.1426476389,0.1385662407,0.1291248798,0.1061420739,0.4041088521,0.0410137959,-0.668022275,-0.3720175922,-0.1881132871,-0.3960177898,-0.1014097854,0.3843832016,-0.0470806658,0.5312978625,0.0287629496,-0.2056684345,-0.0862125158,-0.1587765366,0.1005310118,-0.197074458,0.0241526868,-0.1060945392,-0.1055715904,-0.0062159812,-0.0464084856,0.1483425498,0.1094233021,0.0349792428,-0.3408914506,-0.0137764905,-0.0991125405,-0.1699038595,0.2351678908,-0.3000918627,-0.4332758188,-0.1272641569,-0.0833101422,0.0615360849,0.2040418684,-0.366271615,-0.2041462511,0.1068406254,0.3253265917,-0.166885227,-0.1379862428,-0.0255549382,0.172178641,-0.072690554,-0.1324134916,0.108339116,-0.2491745353,-0.0918181762,0.1922817081,0.1206696406,-0.1902936101,0.1074913442,-0.0969145,0.1603640467,0.359957397,0.1620516628,0.3301300108,-0.0565952361,0.0222890843,-0.0265973546,-0.0437338278,0.0517429672,0.3902280331,0.2014980912,-0.1410317123,-0.1966952682,0.0914515406,-0.3580254018,0.5285892487,0.2002400607,0.4443377554,0.139135018,-0.0581047349,-0.1757815033,-0.0782901049,0.0804357529,0.1939051896,0.2636688948,-0.1120557785,-0.1918566525,-0.4417550266,0.1900993139,-0.160727337,-0.2431439608,0.0924611837,-0.2775330544,0.0905489698,0.1324884892,1.1033917665,0.1686163992,0.018589627,-0.0934926271,-0.5399758816,0.1435104609,-0.4202431142,-0.1255483776,-0.15711537,0.0681318492,0.0445600376,0.2209005207,0.0693136528,0.0844487175,-0.5164357424,0.3140262663,0.3282980323,-0.241850704,0.0566561185,0.2590618432,0.1147610098,-0.4763607085,0.0016142578,0.2510300279,-0.2686349452,-0.2185137123,-0.0796429291,-0.1424877644,0.207404539,0.0661906078,-0.1124873459,0.182862103,-0.0717850178,-0.2107854038,0.0139968675,0.0827370882,0.2111281157,0.1642392427,0.3289702237,0.0193527639,0.0186165757,0.0590539649,0.1112194955,0.0582116842,0.1684823185,-0.2402994335,0.1886992455,-0.3803064227,-0.2146591842,-0.1015245914,0.1218297854,0.3520784676,-0.4099380672,-0.1131093875,-0.0667637885,0.1329622865,0.2181644887,0.2847052813,-0.0955124795,-0.3663686216,0.3259407282,0.0932205617,-0.0603231639,0.0637901053,0.1769742221,-0.0876177549,-0.2700235248,-0.0346205831,0.1882523745,-0.3601574302,0.1101618037,0.0521462262,-0.1896125227,-0.3910233676,0.0135331219,0.1332290471,-0.4133756757,-0.339163363,-0.0275348499,-0.2856507003,-0.1279578209,-0.0176684223,0.0322248004,-0.028844377,0.0975900516,-0.023822099,-0.1382746845,0.009138668,0.0149075696,0.1289699078,0.0924401507,-0.1759359241,0.0868759751,0.1472420543,-0.5050423741,-0.0934308767,-0.208038196,0.2702588737,0.031151535,-0.0900030956,-0.0796960443,-0.0335204192,0.2714110315,0.1041566655,-0.0668587163,-0.3483319879,-0.0102474662,0.0419476964,0.074471958,0.3501231074,0.2198439538,0.1968223155,0.0335757248,-0.0500432961,-0.1493551284,-0.0337403677,-0.0764895007,0.5140311718,-0.0324915983,0.2889122069,0.2089730948,0.2483797818,0.1328306645,0.1527619362,0.0902501568,0.2403993309,0.0690078586,-0.185591042,0.290931046,0.2171251923,0.4299529493,0.2555933595,-0.0402094387,0.2087807208,-0.0229622424,0.0555133559,0.0682897717,0.0299924351,-0.075015001,0.0408847742,0.3493697643,0.4360310137,-0.2219664603,0.2306095213,0.3321712911,-0.2582198679,0.3648055494,0.361635983,0.2640653253,0.1903021336,0.5040661693,-0.0173443444,-0.2511411309,-0.1277496517,-0.0487188548,-0.136809513,-0.0390874259,0.0303182434,0.0138478316,0.0027646914,-0.0963010415,0.0064488975,-0.2305142283,0.388012141,0.1642688811,0.3625433743,0.2226402462,0.0337020084,0.2058231682,0.1601379216,0.0520542264,0.1237766966,0.2139550596,0.4618104696,0.3227737546,-0.2995185852,0.0143626695,0.1419681311,-0.1660998911,-0.092378892,-0.3871394992,-0.1956970096,-0.5284217596,-0.1823373288,0.2277712971,0.091721192,0.2123763859,0.0475419164,0.1291822791,-0.0304145273,-0.0936864242,-0.0650502741,0.0728956088,-0.2042721957,0.3341454864,-0.0044179042,-0.1718678325,-0.040746633,0.4001725912,0.2823663354,0.0635338202,-0.1347172111,0.0582388788,0.1197995692,-0.0395859703,-0.09943524,0.2167884558,0.2455011159,0.3230012655,0.2351204604,0.2431241721,-0.2647364736,0.2468871921,0.1143135056,-0.2101012915,0.2531087399,0.007765356,0.3759463727,-0.0521442667,-0.2855297923,-0.0993689001,-0.5393313169,-0.2674993277,0.2050567269,-0.104503952,-0.0758670196,-0.0846419632,0.1758958399,0.3754512966,0.2512613833,0.3171058595,0.3367232084,-0.2770576775,-0.2797332704,-0.2190154344,0.2346130162,0.0534838811,0.1894771159,-0.3420219421,-0.0951866135,0.4739043415,0.3106839955,-0.1529535204,-0.2677822113,0.2969344854,0.2557627261,-0.5436837077,0.1440631896,0.2747417688,0.0481311753,0.0963225067,0.0668110475,-0.1106623337,-0.0525193773,0.2547315359,-0.1858159155,-0.0379972719,-0.0723374262,0.2572823763,0.1963914782,0.0145907272,0.2616878152,-0.3619594276,-0.0262216907,0.0736025572,-0.1364529878,-0.1064676121,0.1908137202,0.360242933,-0.1058873236,-0.1380241811,-0.3813676536,-0.0224525332,-0.0011139561,-0.088507086,0.187682122,-0.0971851349,0.196041286,0.3025095463,-0.2229755074,-0.2661528885,0.0313742533,0.0797657371,-0.2915389836,-0.2729875743,-0.3375311494,0.0733995438,0.2806568742,-0.1151208505,-0.0906459987,-0.090946041,0.0162581354,0.1169834137,-0.4237063229,0.2224249691,0.1825972944,0.2083545178,-0.0797395036,0.2478806973,-0.0210097916,-0.0256697088,-0.1409687549,-0.1247223243,0.1086561829,-0.0209282972,-0.1656817943,-0.1993289441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"Sounds good, @mariamabarham do you want to give a look?\r\nI think we should have two configurations so we can allow either version of the dataset to be loaded with the `1.0` version being the default maybe.\r\n\r\nCc some authors of the great cos-e: @nazneenrajani @bmccann","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":46,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\nSounds good, @mariamabarham do you want to give a look?\r\nI think we should have two configurations so we can allow either version of the dataset to be loaded with the `1.0` version being the default maybe.\r\n\r\nCc some authors of the great cos-e: @nazneenrajani @bmccann","embeddings":[-0.2522474825,-0.266276896,-0.1281618774,-0.3826157153,-0.4156885147,-0.100878194,0.019268686,0.2192727029,-0.0871498287,0.3809595406,0.0570257194,0.3763352334,-0.0155466124,0.3852823675,-0.0764763206,0.1405269355,0.0847294554,0.3658879399,-0.0866107494,-0.1383298188,-0.2019471526,0.1104134172,-0.0694255233,-0.0544472262,0.1405384243,0.170844987,-0.1972422153,-0.0724422336,-0.2242698073,-0.207714051,0.3550541103,0.4555105865,0.2251389772,0.0698326007,-0.0001058422,-0.2381846458,0.4025995731,-0.0729653761,-0.3339167237,-0.2384467125,-0.4179014564,-0.3771288991,0.0656943172,-0.0959776789,0.0635926723,0.0929082483,0.1650227159,-0.1966180652,0.0641362593,0.0119994245,0.3026269972,0.1741242707,-0.0875847489,-0.345148623,0.3349481523,0.4491071999,-0.3537368178,-0.17945382,0.3167724311,0.0130630191,-0.0673429519,0.4110816121,0.2277281731,-0.1817953587,-0.0252207909,-0.0642819852,0.181264773,0.0074197254,0.1417348683,0.1168777496,0.5061966777,0.129801333,-0.4155392349,0.2899827361,0.0042788964,-0.4935306907,0.2665919363,-0.1062386185,-0.0296140201,0.0642383844,0.1872216314,-0.338834852,-0.4957718849,0.206514433,-0.0779574066,0.0905718207,-0.1579037011,-0.153671056,0.1244990081,0.1076784879,-0.0434820652,0.0413679257,-0.176366359,0.2043868452,-0.2431399822,-0.2590625882,0.2833954096,0.0010792859,0.4588757753,0.0828110874,0.1540057808,0.0436965041,-0.0718054548,0.1041589752,0.1537681371,0.2015690207,0.4607427716,-0.0598572791,0.3992586732,0.2871789932,0.4768276215,0.0420061909,0.1551856697,0.0342297927,-0.0602732599,-0.0319348462,0.0725531429,-0.581581831,-0.0960409343,0.050024353,-0.0163888112,0.0166014414,-0.0381413028,0.0935134143,0.2691530883,0.22098203,0.2916286886,-0.0643905103,-0.035680864,-0.4872016907,-0.295339793,-0.0749426708,-0.2445647866,0.2993686795,0.3277596235,-0.4222892523,0.1382057667,-0.3130993247,0.2833765745,0.0493353382,-0.1800794899,0.3390871286,-0.2251762599,0.1247514635,0.0571318828,0.0853094459,-0.2785827518,0.5125001073,-0.1971521825,0.1246577799,-0.3382834792,-0.4485521913,-0.386895746,0.2870850861,-0.0740516558,-0.1555139869,-0.1518206298,0.3666457534,-0.2398147434,-0.164404273,-0.080504857,0.3010807335,-0.2718228102,-0.1208867654,0.0271402616,0.1432529241,-0.1281467229,-0.0400960408,-0.5082139969,-0.0936469659,-0.1425676942,-0.0299358666,-0.1208521128,-0.2735491395,-0.1699308753,-0.1735196859,0.399415046,-0.3332186639,-0.1463903934,0.232865721,-0.0286984947,-0.2497422695,0.1483421624,-0.1565070152,0.0467194282,-0.2703953385,-0.2281974107,0.0490062051,-0.1481479406,-0.104041703,-0.1473010927,-0.4412724972,-0.1827401817,0.260365665,0.2474036664,0.0334629491,0.2279241681,-0.3996139765,0.0900127292,-0.1699413955,0.175619036,-0.0197643433,0.7601087093,0.0327926986,0.1584606916,0.0639225468,-0.3966698945,0.0720174536,0.0905553699,0.1692190468,0.2283328474,-0.3766866326,-0.035115961,-0.0492086746,0.0924886987,-0.0949597433,0.2041813433,0.1204757318,-0.1321026087,-0.0152661344,-0.0282226633,0.2592668533,-0.240224421,-0.1505691111,-0.0757941753,0.1669904739,-0.1864804476,0.0391455293,-0.1825116575,0.1286381781,-0.0777062625,-0.0659887269,-0.1271273345,0.3903431594,-0.0975591019,0.2309856415,0.0551918037,0.3234147727,0.145233199,-0.1754473448,0.0759682357,-0.1623838544,-0.101344794,0.1879144907,-0.225928098,0.3841077685,0.0692224056,-0.1512343884,0.0097928289,-0.0087744044,0.0624555238,-0.0967298448,-0.3642013669,-0.11718826,0.187561065,0.1114010885,-0.4437871277,-0.2857011259,-0.3088617921,-0.0880208611,0.03273784,-0.0647144169,0.1102838963,0.2029710114,-0.0012311663,-0.2374154627,0.3081510365,-0.105986394,0.2734643519,0.385437876,-0.1669505537,-0.0213520527,-0.2048149109,0.1140058339,0.1383963078,0.281591028,0.0497884266,0.1022423878,0.3348846138,0.0942727476,-0.350233376,-0.3089445531,-0.0454191677,0.0576975048,-0.1550267488,-0.1957225949,-0.0336968303,0.0711773634,0.0014667144,0.147846356,0.1637208313,-0.191449821,0.4073889554,0.269248426,-0.2228077948,0.3243134916,0.0367607139,0.8047565818,-0.1273847669,0.185386315,-0.244890511,0.0535737164,-0.313298583,0.2546364963,-0.2267655283,-0.2278539538,0.4608531892,-0.1008097753,0.2458382845,-0.1181472465,-0.710154891,0.0993442237,-0.1033389196,-0.0934279189,0.0303849988,-0.2146870792,-0.07387501,0.019448733,0.059018217,-0.4485968351,-0.4825384319,-0.1099956259,-0.2480619997,0.1305108517,-0.3649932146,-0.4868312776,0.1457365006,-0.2606458068,-0.2076711208,-0.0529673696,0.1615409702,0.269811511,0.12356361,-0.05811961,0.0688649938,0.051312957,-0.3860318065,0.0458214581,0.3424375653,-0.3188912272,-0.4158439636,0.1217730492,0.0117609445,0.1593984067,0.1757793128,-0.3779386878,-0.3949556351,0.2752785683,0.1295326948,-0.1465403736,0.0690116808,0.534485817,0.069837749,-0.2125398964,-0.0777078271,-0.4181550741,0.1807310432,0.4679790437,0.1432108581,-0.2948800623,0.2655481696,0.1931300014,0.3202042878,0.164707303,-0.0755226389,0.1214160025,0.1337215602,0.218137756,-0.2059158087,0.010714449,0.0165832303,0.0526604392,0.1732346416,0.5349010229,0.1963907331,-0.4072147012,-0.3729563951,-0.2282298952,0.0977964178,0.0007365281,-0.1938165724,0.164902091,0.4404810965,-0.0690809339,-0.1524769813,-0.3290405869,-0.1078980491,0.0979669616,0.3418047726,0.1577134728,-0.0869842991,0.0985982642,0.1773262471,-0.2839725614,0.35610497,0.1472280174,-0.0044862209,0.0054541496,-0.1018800586,0.1518637985,0.0198382847,0.1859119833,-0.2136527151,-0.1691336483,-0.1400272548,-0.0866287872,0.0897074491,-0.0618215613,-0.111711435,-0.2805307209,-0.0905053839,0.0519463494,0.0272615496,-0.1453279704,0.6831689477,-0.1708133817,0.0108841285,-0.2852762043,0.0505160242,-0.2335623354,-0.0009022983,-0.0179394558,-0.1901804358,0.0005650664,-0.1017308161,-0.1505747885,0.239319548,0.0138585055,-0.0067094178,0.1419388354,-0.1401240379,0.0114140399,0.1147930473,0.0812957138,0.0478428304,0.3729476631,0.126290679,0.2275985926,-0.2665452659,0.176054284,-0.1937773973,0.0653737187,0.2491942793,0.0301405769,-0.0644270554,-0.2971716523,-0.2293914855,-0.1234592125,-0.2008915544,0.2619427741,0.0415447727,-0.3726584017,0.0283851177,0.324510783,0.0217527971,-0.2729001939,0.2598052621,0.2153334022,-0.2479898185,0.3315581083,0.0006591656,0.73477,0.2477429956,-0.0634477362,0.2377441525,-0.0478468202,0.4994876385,-0.1533699334,0.0137079852,0.0788927078,-0.132751897,-0.0938505977,0.2984379232,0.3770504296,-0.0920601338,-0.3057995439,0.1782729179,0.3676154017,0.2488740236,0.0815007612,0.1855843216,0.3359360695,-0.3099154234,0.1831716746,0.1883808374,0.007580678,0.169815585,0.0008189662,-0.0363445692,0.0140580693,-0.3261225522,-0.2256563306,-0.1407503188,0.3442864418,-0.0092998249,-0.0357621424,-0.0722307637,0.4353804588,-0.3726445735,0.2079820335,0.1910073608,-0.2896559536,0.1257748902,0.1657652259,0.0537250526,-0.0695897192,-0.0944856703,0.0770210028,-0.1679818481,-0.0914042816,0.1893950254,0.2116372287,-0.1577854455,-0.0802488774,-0.1400251389,-0.0321276039,0.0489335768,0.3648075759,0.5734391809,0.1982718855,-0.2803201973,0.2468814999,0.2291864306,-0.111687392,-0.0778991953,0.1520566344,-0.1000769734,-0.2093644887,0.4151041508,0.4110589623,-0.4059256315,-0.0472274423,-0.0828969032,-0.0895663053,-0.2743075788,-0.046447929,0.1201977059,-0.5605068803,-0.1997343004,0.0694171935,-0.2003938109,0.1811296642,0.7872422338,0.1409817934,0.1752446443,-0.0520094112,-0.018956881,-0.1822192073,0.0649248883,-0.4313209951,0.3021856248,-0.0268186629,-0.1548499465,-0.1137981787,0.0398096032,-0.4568734169,-0.192203939,0.2623077035,-0.1323944926,0.372192502,0.3105391562,0.2645424604,0.0035347384,0.2119169682,-0.1414116472,-0.3338809013,-0.2028174698,0.0646516681,0.0921234563,0.0667718425,0.1733727455,-0.038243819,-0.1149307862,-0.2642995715,0.0325512551,0.2622590661,-0.0297731422,-0.1581732184,0.1851063371,-0.244840011,-0.0719846189,0.1540466994,0.3823598027,0.189388901,0.0826760381,-0.0652610734,0.4259437025,-0.0072943703,-0.112665087,0.0466037504,0.3781134188,-0.2774096727,0.3002434373,0.0186753497,-0.0245852005,-0.0857216045,-0.0378537402,0.1339822412,-0.4113796651,-0.0707939714,0.0289072189,0.0336687006,0.3719576597,-0.2801135182,0.0525011942,-0.0581756309,-0.0551724024,0.2176862061,0.1967808008,0.3475120068,0.0154164843,0.7643714547,0.0325853378,0.0124498364,-0.0536110103,0.452549696,0.1095074117,0.0451967679,0.0928003043,0.0846388564,0.0675306246,0.1761347204,-0.0687372163,0.1197274029,-0.082432583,0.4220128953,0.5082917213,0.0091781029,-0.3382549882,0.0782883018,0.2829191387,0.089671433,0.0331840105,0.3009580374,-0.1659035981,0.0112869833,0.0791627765,-0.7445872426,-0.1521625221,0.0200079884,-0.155373469,-0.1584934145,-0.1347789466,-0.406162858,-0.0660934076,0.1643917859,0.0075460924,0.1547903568,-0.1995059103,0.0156907178,-0.1065372825,-0.0079827383,0.1111055091,0.2240459472,0.0266270656,0.3685717881,-0.4948239028,-0.2772006989,0.3008654416,0.3290163279,-0.0164614562,-0.0708062053,0.3832434416,0.1276649684,0.0833326802,-0.1014528275,0.0409205928,0.0849525928,0.0976473764,0.1466632783,0.2090859562,0.2040060759,-0.1369895041,0.2419367582,-0.0748776346,-0.139432475,-0.2820705771,0.1175025329,0.1690342724,0.0901318789,0.0302818026,0.1664307863,0.0411572829,-0.3709060848,-0.0729447827,-0.0576062538,-0.1295697838,-0.1690272987,0.1355428994,-0.169342652,-0.0485295914,0.2299575806,0.0072026076,-0.0136007695,-0.4024969935,0.0531442314,0.3580956757,0.1434370428,-0.0768830255,0.0670179725,0.2265027761,0.2597039342,-0.1298808903,0.1573919952,-0.2579096556,0.3699952662,0.0945108905,-0.3097195327,0.1618372798,0.4594863951,-0.0283689909,-0.1124437973,-0.1568193138,-0.1069958284,0.0478334799,0.0735950768,-0.2418142408,0.070189476,0.135439679,-0.0018165458,0.5354017615,0.0257030744,0.1743659377,-0.213264212,-0.0477884784,-0.4498699903,-0.2057503611,-0.0833876878,0.0280175004,0.0858354121,0.0876757503,-0.0047045005,0.0536871925,-0.0317796394,0.279851228,-0.0141204037,0.0002256404,-0.2494334877,0.4703898132,-0.3713132739,-0.270843327,-0.0491854288,-0.1553102136,0.1130798385,-0.1719789952,-0.1046572104,-0.176669091,0.287073791,-0.0711372718,0.0517831109,-0.1854549199,0.3238449097,0.3122844994,-0.0000629606,-0.3327043951,0.0687358379,0.1956884563,0.1794513911,-0.1364153773,0.2679242492,-0.0747957081,0.0234999582,-0.1861780733,-0.0111830328,-0.1901151091,-0.0674749911,0.0328602828,-0.3748263419]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"cos_e v1.0 is related to CQA v1.0 but only CQA v1.11 dataset is available on their website. Indeed their is lots of ids in cos_e v1, which are not in CQA v1.11 or the other way around.\r\n@sarahwie, @thomwolf, @nazneenrajani, @bmccann do you know where I can find CQA v1.0\r\n","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":50,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\ncos_e v1.0 is related to CQA v1.0 but only CQA v1.11 dataset is available on their website. Indeed their is lots of ids in cos_e v1, which are not in CQA v1.11 or the other way around.\r\n@sarahwie, @thomwolf, @nazneenrajani, @bmccann do you know where I can find CQA v1.0\r\n","embeddings":[-0.0232327729,-0.1867629588,-0.1784631461,-0.3313111365,-0.5435762405,0.0657164901,-0.1783175617,0.1259405315,-0.3440611362,0.4689726233,0.2354179025,0.1240532696,0.0749001652,0.4212986231,0.0156426728,0.0894112587,0.117030628,0.4064915478,0.1564118713,-0.0762307197,-0.0469662733,0.3286637664,-0.2688592672,-0.012009602,0.1689498425,0.112477757,-0.3255741596,-0.2680034637,-0.1470949948,-0.1698807329,0.4301481247,0.2873969078,0.354203701,-0.0082968352,-0.0001180449,-0.2494498491,0.2678097486,-0.0638853982,-0.2248022854,-0.2884759903,-0.3987270594,-0.3756753504,-0.1216315106,-0.0159799121,0.1313795745,0.0580618493,0.2008041292,0.0156754162,-0.1259213537,0.1071852669,0.1894945502,0.1807599962,0.141135782,-0.3795875609,0.1705751866,0.3191810846,-0.2734742761,-0.2285316288,0.2699380815,0.0258557796,0.3077663779,0.4816246331,0.2874330282,-0.3409462869,-0.1720889509,-0.013270352,0.1167342588,-0.1250804812,0.2783508301,-0.0452661663,0.5306007266,0.1798347831,-0.3857192695,0.4131044745,0.0453327149,-0.4731240571,0.2673746645,-0.0138167813,0.0682043806,0.118957445,0.237447843,-0.305306524,-0.4384269118,0.1660591513,-0.1632631868,0.0052230353,-0.1122972071,-0.1868192852,0.0656315535,0.045105312,-0.0487550758,0.1840634644,-0.2509777844,0.2603937089,-0.2296653688,-0.3830533922,0.2817987502,0.0218084604,0.3297159374,-0.0346217491,0.1586210579,-0.2194430083,-0.2475143075,0.0007930144,0.1206260175,0.3228637278,0.3589960635,0.0529479943,0.1459193528,0.235377714,0.4406010211,-0.0404021442,0.0712785497,0.0129208686,-0.1671769172,0.0568953156,0.060212113,-0.6049147248,-0.1965153962,-0.0045784405,-0.2337331623,0.0628118888,-0.1438020915,0.0319252163,0.418045342,0.2180632949,0.2200229913,-0.1935508996,-0.0098073846,-0.6159117222,-0.2765554786,-0.0146776028,-0.0672637373,0.3063664138,0.2101648301,-0.3618155718,0.2205503136,-0.3328654766,0.224240616,0.0553091727,-0.1585887969,0.4364677966,-0.2266087979,0.1858067364,-0.0229867026,0.0607716367,-0.2817752361,0.4902441502,-0.2170691788,0.1842231899,-0.3824463189,-0.4308210611,-0.501421392,0.1403851956,0.0507101119,-0.1981070489,0.0367361046,0.2465415746,-0.2765118778,-0.1712324917,0.0227238573,0.2964550257,-0.1747252792,-0.1326307803,-0.0055325804,0.1790474206,-0.2631231844,-0.0621967837,-0.392976135,0.0770601183,-0.1144097224,0.0586968064,-0.0611227006,-0.3164398372,-0.1432292461,-0.2597507536,0.3386314511,-0.6128681302,-0.5238029957,0.1491823345,-0.0313184857,-0.2901610732,0.2189871818,0.0052803205,0.0072041349,-0.1621967703,-0.2095164955,-0.0362694375,-0.2873411179,-0.3331362903,0.0219733696,-0.4311449528,-0.0796960965,0.1074301749,0.0924729109,0.0935381427,0.2323921025,-0.6215128303,0.1444454491,-0.0195611008,0.0103544975,0.0693640336,0.6898933649,0.0778632462,0.2611390352,0.0516130961,-0.3183967769,0.0562004223,-0.2152469456,0.0695454702,0.2679488957,-0.3772905767,-0.1154734194,0.0034697491,0.1692007184,-0.2824362814,0.0478885844,0.0488528237,0.1080219224,0.0012774151,0.0135006458,0.3367107213,-0.1517360806,-0.1799661964,-0.103039369,0.1771976799,-0.1950794309,0.0085162651,0.0162855051,0.2220596969,0.1066612378,0.019013105,-0.0894381478,0.1999128014,-0.1644220203,0.2517156601,0.2599522173,0.4301055074,0.2277344763,-0.2260206789,0.0658660084,-0.0969092548,-0.1978674233,0.3507429659,-0.0107880915,0.2520744205,0.1498487294,-0.0752098784,-0.0074323537,0.2028425783,0.0315937921,0.0416572914,-0.2068519294,-0.1189873368,0.2147415876,0.1000060216,-0.3482510448,-0.3811196685,-0.2063105851,0.1608173549,0.1386514008,-0.1160973459,-0.0007920525,0.1139478981,-0.0206832141,-0.0809205472,0.3460274637,-0.1036329493,0.0956860334,0.2869534791,-0.2408586293,-0.1860807091,-0.009621487,0.1602338552,0.1520680189,0.2799004018,-0.2412097156,-0.1741050929,0.3792101443,0.1788065135,-0.1269617975,-0.2324094176,-0.0148720704,0.0560344458,0.0285747927,-0.0962511376,-0.1388391256,0.005258176,-0.0699222237,-0.0594110079,0.1831770986,-0.1977694035,0.277212888,0.1787653118,0.0725568607,0.2747685313,-0.0002004471,0.7763874531,-0.2226073742,0.303317517,-0.0758502036,0.1394991577,-0.4241236448,0.1356927454,-0.2917244136,-0.1730993241,0.4106993079,-0.098780632,0.3334311545,-0.0885260105,-0.8045449853,-0.0008278571,-0.1077457815,-0.1097392961,0.1192290485,-0.3133375645,-0.0917769819,0.0245353598,-0.0257826895,-0.2734339237,-0.4853473604,-0.1646532863,-0.3983456492,0.0878533795,-0.1717593968,-0.3822391033,-0.0438728295,-0.0767247751,-0.1242800876,-0.1243780553,0.2652041018,0.1253896803,-0.0085397772,-0.0731472969,0.1770051867,-0.0466657057,-0.1972223371,0.2090691924,0.3373953998,-0.3322137594,-0.3296049237,0.3415119648,-0.0217028633,0.2813383043,0.0114787035,-0.4279914498,-0.1520687789,0.066889599,0.1309575886,0.0516323112,-0.0365510024,0.7915483713,-0.116004616,-0.0761181116,-0.0275548529,-0.5734854341,0.1022515222,0.5704916716,0.2422737032,-0.1116487533,0.3174265921,0.1585417092,0.5127474666,0.3592986465,0.0104253078,0.022733897,0.1787563711,0.0413279384,-0.1243185475,-0.2332143337,0.0308823455,0.0835504085,0.0810874254,0.3579852879,0.1712749749,-0.2575516999,-0.23424007,-0.4313467741,-0.0486306958,0.1353089809,-0.3467603326,0.0899708644,0.666692853,-0.0572716184,-0.1794426739,-0.4841734171,-0.0520420149,0.1998935342,0.4482497871,0.3122658134,-0.0023317491,0.159017995,-0.0632469729,-0.2990367413,0.3822370768,0.096446164,0.0762840435,-0.0236213636,0.05705259,0.1606549323,0.0039881067,0.2074627876,-0.1188340709,-0.2066369057,-0.0895554721,-0.0419886783,0.3214385211,-0.0323582925,0.0398222357,-0.2484740913,-0.1059365273,-0.1875948608,0.1950106621,-0.0317002088,0.6617469192,-0.2799368203,-0.1786487252,-0.1140607595,-0.0565543026,-0.0745498911,0.0456811525,-0.0588558465,-0.2610063851,-0.0176252518,-0.0392804556,-0.1723128557,0.2192323655,-0.2052269578,0.007135998,0.0485954471,-0.0689193159,-0.0186888855,0.0803602487,0.1086186916,-0.0285142343,0.2341809869,0.0030018315,0.3989965916,-0.3178144395,0.1732622534,-0.0441847555,0.2235638648,0.15306741,-0.0316438191,0.0737854019,-0.1780215204,-0.3110803664,-0.2239465117,-0.4165322483,0.3540849686,0.0946247801,-0.2084657848,0.0869978517,0.2448112667,0.091131404,-0.2714422345,0.3374062777,0.3942421675,-0.0137908729,0.178623721,-0.0409817547,0.7828717828,0.3878051639,-0.1163377315,0.0890960395,-0.1653100103,0.303385675,0.027321076,0.1015211418,0.1656834185,-0.1250873655,-0.1338010728,0.1904687732,0.3938689232,-0.1923112571,-0.2761599422,0.3090688884,0.2963235974,0.379137516,0.1851730049,0.0466067046,0.3087801635,-0.0121754948,0.2288380861,0.0856976807,0.1700183749,0.4446303844,-0.0150174806,-0.1496939063,-0.0736476034,-0.244804427,-0.3190774322,-0.2423902899,0.1053238213,-0.125343889,0.0637126341,-0.1350139976,0.4795577228,-0.1867608577,0.4332155883,0.0197667554,-0.3183547258,0.114286311,0.2496321946,-0.2459230721,0.0507218167,-0.0320617817,-0.2735675573,-0.1974431574,-0.1281524897,0.4675572813,0.2643013299,-0.1820860356,-0.2903381586,-0.1792751402,-0.0097718434,0.167353034,0.3781094849,0.4931951761,0.2174014747,-0.2325193286,0.1388516277,0.1454216689,-0.1246194467,-0.3335357308,-0.1084514186,0.1260614246,-0.1592849642,0.2072100192,0.4298124611,-0.322442919,0.0774456859,-0.1238440424,0.0354774296,-0.1322073638,-0.2717387676,0.0306166373,-0.4342004359,-0.1563019007,0.0317441002,-0.1710365564,0.1918587536,0.9666594863,0.0614945553,0.200777933,-0.187818855,-0.015118341,-0.20959647,0.1748205721,-0.5250988603,0.2180927396,-0.175972566,-0.0797948018,-0.1631566584,0.2668922544,-0.2849743664,-0.2289028019,0.1799532622,-0.1577756703,0.3327469826,0.2898228168,0.1722331494,0.1120604351,0.0433363393,-0.1313568205,-0.1408833563,-0.0471347831,0.1000950038,0.1224140152,0.0846709907,0.3393621743,-0.0374947563,-0.1875878423,-0.1911075264,0.199716121,0.249857083,-0.0930991024,0.0867886469,0.247604087,-0.2960228324,-0.1283229738,-0.0852326527,0.5458861589,0.2416352034,0.137349993,0.1575449258,0.3756002188,-0.0907988399,-0.0389034599,-0.0743579865,0.4465692937,-0.3893509209,0.2567912638,0.0368533209,0.092438899,-0.1318696886,0.0382294282,-0.2625072598,-0.5318465829,-0.0129731167,0.1900568008,-0.0572399423,0.2067933083,-0.3643628657,0.0220328569,-0.2610281408,-0.0993008763,-0.0296507869,0.2112357616,0.3775580525,0.0707441419,0.8590605259,0.1181167811,-0.1102918014,-0.0721647814,0.4866429865,0.2277108729,0.39008075,0.1389106512,0.0902569517,0.0017617153,0.1071945876,-0.1658710092,-0.0767112449,-0.1397224516,0.44157058,0.4549824893,0.3275453746,-0.2002153993,-0.2080340087,0.2436772287,0.1848271191,0.0143338824,0.2330923826,-0.081150651,0.1278197318,0.0943102166,-0.7621110082,-0.0989195779,0.0741957426,-0.1612007171,-0.1580147892,-0.1228053048,-0.2658144832,-0.1131069139,0.3679631352,0.0979596376,0.2267372459,-0.27596277,-0.1494999826,-0.0591042563,-0.2040783316,0.2876170576,0.2594886124,0.0613487028,0.4843765795,-0.5644532442,-0.3007172048,0.4297738969,0.2108376324,0.0191381201,-0.2560528517,0.2916177511,0.0555405617,-0.0217688959,-0.2667055428,0.2390976548,0.1827180088,-0.1402673274,0.1108371615,-0.0146557838,0.062215358,0.0434249826,-0.0588011518,0.0120082824,-0.0806274787,-0.475709796,0.1342348754,-0.2050315738,0.1652192175,-0.1219318509,0.4493810236,0.0837065279,-0.380233705,-0.2866356075,-0.0444584079,-0.2227216661,0.0143529475,0.0550222881,-0.4033516347,-0.1603308171,0.3848736584,0.1999774277,-0.1210138574,-0.6013125181,0.0616343617,0.2573510408,0.1335511655,-0.1310296357,0.1150699332,0.2602623999,0.1477566659,-0.2571093738,0.2359615415,-0.0140665593,0.2910215855,0.098843962,-0.1629007012,0.0829617977,0.3361525536,0.0039278325,-0.2415363044,-0.0137025341,-0.1254630685,0.2838470936,-0.125092715,-0.1482129991,-0.029125331,0.0956339613,0.1216610298,0.4608382285,0.0902078003,0.0214881059,-0.0657209754,-0.1475344598,-0.4012160599,-0.060062509,0.0328884833,-0.132617712,0.0142063173,0.0518731698,-0.1359093934,0.3132941425,-0.0092498017,0.436945051,0.1265654415,-0.024377048,-0.1482625455,0.4578133821,-0.4662644267,-0.1687988192,0.2816958129,-0.1104642153,0.0842431188,-0.1132378727,-0.1050944924,-0.0178550445,0.1795220077,-0.1586622596,0.2223753184,-0.200109601,0.4053361118,0.2907797396,-0.0312130433,-0.3527669609,-0.0743962377,0.1831317693,0.2774866521,-0.0178936105,0.3830201924,0.1775814146,0.059883818,-0.1240592301,0.2663044631,-0.1834813654,-0.0043845247,0.0480381325,-0.2990570962]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"@mariamabarham I'm also not sure where to find CQA 1.0. Perhaps it's not possible to include this version of the dataset. I'll close the issue if that's the case.","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":29,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\n@mariamabarham I'm also not sure where to find CQA 1.0. Perhaps it's not possible to include this version of the dataset. I'll close the issue if that's the case.","embeddings":[-0.2492516041,-0.1036452204,-0.1391218007,-0.421084702,-0.5307980776,-0.0687487349,-0.0787495673,0.079065986,-0.2197635472,0.439055711,0.1513827443,0.2936987579,0.0653416812,0.437533468,0.0067794044,0.2519571781,0.154427886,0.398141861,-0.0359028429,-0.0811108202,-0.0989973918,0.3010998368,-0.2269811034,-0.0594953373,0.0692324415,-0.0203682762,-0.2728616595,-0.1656872779,-0.1539622247,-0.1539584398,0.4113928378,0.362680614,0.2077950686,0.0043117353,-0.0001105137,-0.2109789252,0.2824904621,-0.1277124733,-0.2300390154,-0.1238049865,-0.3332126737,-0.3251547813,0.0086399382,0.0028271973,0.0867994577,0.1362752318,0.1289967,-0.1048249602,0.0015906467,0.1707078815,0.2644631267,0.2665335834,-0.0055657737,-0.3921883702,0.1672420353,0.2768290043,-0.3164582551,-0.1491539478,0.3068576455,-0.0171318669,0.0829685107,0.5250521898,0.2620768547,-0.1775963157,-0.129355222,-0.0538349077,0.1235546172,-0.1147581339,0.3260295987,0.0554937795,0.5153294802,0.1486225873,-0.4204435647,0.3833307028,-0.0019987151,-0.3895690739,0.2576098442,-0.0585736595,0.0061522052,0.0596882999,0.1455919445,-0.1512774974,-0.3991324902,0.1424817145,-0.1709094048,0.0312471241,-0.1322285086,-0.2196556777,0.0802680999,0.0961432829,-0.1033073887,0.209843874,-0.2582565248,0.1884111464,-0.2771467566,-0.2824535668,0.3214116096,0.0659707859,0.4131599069,-0.0623520128,0.1519739777,-0.0995100141,-0.1111688018,0.1464018226,0.051799383,0.250402391,0.2354280204,-0.0533718951,0.323148936,0.2254090458,0.4800464213,-0.039997153,0.1612970233,0.0416817218,-0.218578577,0.012427018,-0.0265025832,-0.6751200557,-0.2794663608,0.0029017271,-0.0757998899,0.05903209,-0.0788476914,0.1403900385,0.3872992694,0.1108708456,0.2383925766,-0.0379945524,-0.068619892,-0.50391078,-0.3114446104,0.0499008447,-0.0885656923,0.211431399,0.2986612022,-0.2297255397,0.3060235381,-0.3308390081,0.367343843,0.1144686192,-0.2047566473,0.3666538894,-0.192683965,0.2855257392,0.0603468567,0.0428677164,-0.2203834355,0.3571561873,-0.2228281796,0.0957194567,-0.2677157521,-0.4235020578,-0.4708480835,0.1999157965,0.0558829419,-0.2293753326,0.0492196679,0.3272183836,-0.2058468312,-0.1771707088,-0.025696652,0.3373240829,-0.107574068,-0.1229423359,0.1113110632,0.1878852695,-0.2905888259,-0.0097981887,-0.3686021566,0.0443041809,-0.0441482775,0.043566715,-0.1005467772,-0.2231190205,-0.1166252717,-0.2349683642,0.3060906827,-0.3721945584,-0.4113180339,0.1309095919,0.0586126447,-0.3415106833,0.1281526834,-0.1533479095,0.0193928108,-0.1947710663,-0.1779786199,0.1028158292,-0.180414021,-0.165331468,-0.0236465596,-0.4142696857,-0.1249917001,0.2005962878,0.1209188923,0.0646155849,0.2701731622,-0.6116059422,0.0557609871,-0.1391748935,0.1110891625,0.083754085,0.746485889,-0.0461580902,0.1895792931,0.1458414942,-0.3134403825,0.023550909,-0.2664182782,0.1402665079,0.1166450679,-0.3408264816,-0.0927930176,0.0023918157,0.0644075945,-0.3398041427,0.1386117637,0.0946590379,0.0508369282,0.0979671031,0.1258643866,0.3094986379,-0.1931175292,-0.2252657264,0.0018339381,0.1426391304,-0.2699936628,-0.1047979221,0.1443324983,0.2866582274,0.0250290446,-0.0401916504,-0.1668748856,0.2657507658,-0.1632743925,0.2737016678,0.0959017277,0.3830801547,0.1436947584,-0.14276281,0.0734449551,-0.1170784757,-0.1200567111,0.3190225363,-0.1159522161,0.3486655951,0.0862263963,-0.081145443,-0.0352938585,0.1888718605,0.1800477952,-0.037101604,-0.2181863189,-0.1006281078,0.1947361678,0.0921784863,-0.3050856292,-0.3667620122,-0.2442114949,0.0528481901,0.1815315783,-0.0088263582,0.0946418792,0.1603841931,-0.0385175683,-0.1638860404,0.4497121572,-0.0922288075,0.1084881797,0.3049634695,-0.2005963624,-0.1527107358,-0.1487498432,0.0477216914,0.1299113333,0.2641783953,-0.0851148292,-0.0490247756,0.339348942,0.1400030702,-0.353304714,-0.2279834747,-0.0583972521,0.0282817204,0.0008938876,-0.0952360854,-0.1337913722,0.0065198941,0.0302349683,0.0011713211,0.1366798133,-0.2366505861,0.3674173355,0.1476506442,-0.1177413762,0.3822770119,-0.0072095348,0.7501001358,-0.1117475033,0.3082492948,-0.1857604831,0.1219291463,-0.4235035479,0.2297057062,-0.2892770469,-0.1285903156,0.4955849051,-0.0979459882,0.4712076187,-0.1098539978,-0.7647408843,0.0673186854,-0.122909002,-0.0159366075,0.0147052314,-0.2569583356,-0.1591472924,-0.0193548314,-0.049182754,-0.3315587044,-0.4355299771,-0.051724676,-0.4801782668,-0.005570455,-0.351970762,-0.4595099688,-0.0211464334,-0.216278255,-0.1986924261,-0.2179501355,0.1730873585,0.2513789237,0.0254415795,0.0503351875,-0.0154360346,0.0072431732,-0.2543784976,0.1297785193,0.4986969829,-0.3865446746,-0.5030515194,0.2911550701,-0.0433312356,0.3502509296,-0.0062929997,-0.3598849773,-0.2641656995,0.000894843,0.1479327232,-0.0094680898,0.0382687859,0.762144208,-0.1158139035,-0.1910690814,-0.1162673011,-0.5392227173,0.1955431402,0.4982728362,0.2428084165,-0.2789062262,0.3810684681,0.2099152207,0.3193804026,0.351167798,-0.0565193929,0.1216356382,0.163393572,0.1477165073,-0.1963391006,-0.0166930668,0.1428921968,0.0754414797,0.1486087888,0.5334333777,0.1272828877,-0.3797419369,-0.2531403303,-0.2787438929,-0.0565796942,0.0524577238,-0.2920554578,0.015738802,0.5925149918,-0.0275712907,-0.1643763036,-0.2947055697,-0.0242578648,0.182671532,0.4032699764,0.2256295383,-0.0115686646,0.0843058154,-0.0759992227,-0.360498935,0.3819508553,0.056751553,0.1014717743,-0.128196314,0.0615387894,0.1697919816,0.0103952875,0.2331535965,-0.1957223862,-0.2707685232,-0.097538963,-0.0332345739,0.2088963538,-0.0756589323,-0.0604064725,-0.1760562211,-0.0584552586,-0.1786636561,0.0345810726,-0.0947062299,0.5656785965,-0.2811228633,-0.045757357,-0.0969693139,-0.0175605491,-0.2522436976,-0.0532965921,0.0190001577,-0.1461752802,-0.0296965074,-0.0632439628,-0.2030777931,0.3560205698,-0.1344767958,0.0051939827,0.2022386044,0.0130261537,0.053488031,0.0671311393,0.0382125638,0.0319467597,0.2077141702,0.0228809286,0.36382249,-0.1425088495,0.2619236708,-0.2598847151,0.2579028904,0.1795643717,-0.053154543,0.0472466573,-0.2090418339,-0.2319688946,-0.1119751111,-0.2654044926,0.2837344408,0.0360335708,-0.1409928203,0.1201900542,0.2442866862,0.0264968816,-0.2426864505,0.2338392437,0.341463536,-0.1066096202,0.3187501132,0.0144107556,0.7144321799,0.2847658098,-0.1513294727,0.1777028441,-0.09784697,0.2846783698,-0.0852066278,0.0308430307,0.1436238587,-0.0294141974,-0.1080328077,0.2240846455,0.4733096957,-0.1807129979,-0.4074777067,0.2393462956,0.3737254441,0.2526540458,0.1791731566,0.0258240346,0.2792443633,-0.1941186041,0.2228651047,0.1612623334,0.1186592281,0.2478431314,-0.0079302695,-0.1217643246,-0.1070735827,-0.3347503543,-0.2222693861,-0.1555014253,0.1758873016,-0.1371008009,0.1295870245,-0.1396807283,0.5002401471,-0.2163032144,0.3512282372,0.0769427493,-0.3939012289,0.1984991282,0.1237464473,-0.0937176794,-0.0484697483,0.011539422,-0.0930699632,-0.3150056303,-0.0404584855,0.2745904624,0.1350487471,-0.2294902802,-0.1504877061,-0.1553414762,-0.0517476201,0.1955786496,0.3932063878,0.496765554,0.2055808008,-0.2776392102,0.2118620574,0.1932183504,-0.0996155962,-0.0902924836,0.0016793218,0.0523810759,-0.2097777277,0.2024031579,0.3447616994,-0.2052671015,0.112269558,-0.0696332008,-0.0756431669,-0.2333086878,-0.3203868866,-0.0139438938,-0.5064803958,-0.2149947286,0.0462604426,-0.264634937,0.1625892371,0.9665586948,0.0193355326,0.0856622383,-0.15413028,-0.1277806014,-0.2720561028,0.0914573073,-0.6348624825,0.3025991321,-0.0970020816,-0.0106851896,-0.0588764921,0.2275673449,-0.4005646408,-0.2310591191,0.0858088061,-0.0754894614,0.1727146506,0.1850544959,0.3002167642,0.2000572383,0.108021386,-0.0364285596,-0.2978537381,-0.1409254074,0.0670611411,0.0902276784,0.085174419,0.2651061118,0.0155737298,-0.0961283818,-0.3636771739,0.1402250081,0.3158980608,-0.2554154992,-0.0834079161,0.2138915509,-0.2956832051,-0.1531853974,0.0740929917,0.2681745589,0.2032589763,0.0001840272,0.0226533581,0.4017669559,-0.0740264356,-0.1279597282,0.0669446215,0.3990156054,-0.3243765831,0.2326636016,0.0578705221,0.0217764545,-0.0746772215,-0.0274186879,-0.1153786108,-0.5985839367,-0.052579619,0.0898478925,0.0272688195,0.2741135061,-0.4083034694,0.10977038,-0.2336895615,-0.0654066801,0.1733333766,0.1268072426,0.5387257934,-0.0435409807,0.9180057645,0.1078987792,-0.0798937157,-0.0731531903,0.3522465825,0.0984481201,0.1931409687,0.202614665,0.1697167158,0.0173920747,0.0831242278,-0.0640716404,0.0103155775,-0.0951850712,0.3211268187,0.4013402164,0.1526316255,-0.3137331605,0.0371578522,0.1614001393,0.1964496225,0.0164234433,0.21400702,-0.0955813229,0.1256730705,0.1140469462,-0.8520111442,-0.1479996294,-0.0002812394,-0.1519284397,-0.1547406167,-0.1215368509,-0.2612949014,-0.1040913835,0.2915905416,0.1082489863,0.2901679277,-0.2956883907,-0.0885552764,-0.0715282485,-0.0806171969,0.1421355009,0.1428746432,0.0894800872,0.5019297004,-0.4905025065,-0.2560096085,0.3050370514,0.3041681349,0.1610272825,-0.2079575211,0.2815841138,0.1821937561,0.0196436178,-0.2664214373,0.167782262,0.2630542219,-0.0011915984,0.1215178892,0.1015298292,0.1110988334,-0.0689552799,0.0190045778,-0.1104255244,-0.0046383045,-0.2969081402,0.143189922,-0.0726609454,0.1251638085,-0.0830333307,0.2422427833,0.0606009401,-0.3382857144,-0.2355432361,-0.1456390172,-0.2185368091,-0.1448451579,0.1124263927,-0.2741731405,0.020281015,0.2404279113,0.2540443838,-0.0268278699,-0.5908102989,0.0029733039,0.274187088,0.1333941519,0.033586964,0.0834171772,0.2575049102,0.2082338631,-0.261369735,0.2973892987,-0.1402301788,0.3001264632,-0.0088128038,-0.3560077548,0.0668691546,0.2684217095,-0.0083763432,-0.1806321144,-0.0437190011,-0.1160851792,0.2443627566,0.0062126056,-0.2181475759,0.0143151358,0.0867173597,0.0467003919,0.4745104909,0.2091073543,0.1296901852,-0.1403009742,-0.1385899186,-0.4177076519,-0.0969316289,0.0402000919,-0.1644821614,0.0742763281,0.22511217,-0.0425951146,0.1396816075,-0.0036879184,0.4698243141,0.0791626945,-0.0848117843,-0.121596925,0.5496423244,-0.4163464904,-0.1756496578,0.123440586,-0.1344538927,0.1641334593,-0.184733361,-0.1444327086,-0.1051093265,0.2785086334,-0.2274986506,0.0910434574,-0.0843521729,0.3133118749,0.1195718721,-0.0108948322,-0.3383713663,0.0417312197,0.1130971685,0.231424436,-0.0858876854,0.4441743493,0.1086281165,0.0114214979,-0.1466144174,0.3638553917,-0.1538969576,-0.0409897678,0.0437493846,-0.3198182285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"I do have a copy of the dataset. I can upload it to our repo.","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":15,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\nI do have a copy of the dataset. I can upload it to our repo.","embeddings":[-0.1301398873,-0.3180444539,-0.1511415243,-0.3585762084,-0.5274255276,-0.0939866155,-0.095670484,0.0967687294,-0.1954925507,0.4636943042,0.0969164148,0.3103718758,0.0110549517,0.2462162822,-0.0645753592,0.1845471412,0.0454773903,0.3228810132,-0.0203287378,-0.1253295243,-0.1092714295,0.1784150451,-0.1254279613,-0.1435169876,0.157889232,0.1021316797,-0.2689062655,-0.2657912076,-0.2371212542,-0.10629756,0.2977129221,0.3701300621,0.1365868151,0.1024103016,-0.0001141522,-0.2017341703,0.3805154264,-0.0472319797,-0.2405022383,-0.1861137003,-0.3527407646,-0.4007867575,0.018492531,-0.1332497299,0.1369123608,0.0520547926,0.1307166219,-0.0537263267,0.0368565731,0.126551494,0.2153394073,0.2612070441,0.0466376431,-0.3750877082,0.3713698089,0.5215311646,-0.3999204338,-0.2230221927,0.25242576,0.0256787315,0.0530219786,0.4859327376,0.2914904058,-0.209337905,-0.1669053137,-0.1291134059,0.1156887487,-0.0294842441,0.24386473,0.0771595538,0.4434210956,0.2278308272,-0.4095323086,0.3944903314,0.0466179252,-0.420558542,0.2389674783,-0.1286545992,0.0233564842,0.0818191767,0.1710073799,-0.3995892406,-0.457126826,0.1633176208,-0.0319894552,0.1493175924,-0.1369164288,-0.1987553686,0.0882630944,0.0855623558,-0.1022077426,0.205327794,-0.1453804225,0.142112717,-0.1941294074,-0.2368096858,0.3193953335,-0.1187759489,0.4837409854,0.0551715046,0.1206807345,-0.1289093047,-0.1747512072,0.0475262031,0.1188868955,0.2328905165,0.3765419424,0.1113040745,0.2856770456,0.3706319034,0.5251098275,0.0471715257,0.1679468304,0.1290954202,-0.0703309029,0.0059320456,0.1129587442,-0.6104717255,-0.0715411529,0.0267226286,-0.1651549637,0.0067793126,-0.0938457921,0.0775082931,0.3279447556,0.1588679999,0.2542691231,-0.0754406825,0.0348116569,-0.492032975,-0.2880235314,0.0442299694,-0.1380849928,0.2874877751,0.2840173542,-0.3480616212,0.2365134805,-0.3418881297,0.2721745372,-0.0279921405,-0.2165766954,0.4410939217,-0.1569769382,0.2365999669,0.1015854403,0.1018674821,-0.2390626967,0.5781809688,-0.1842716187,0.1235166639,-0.3338504732,-0.4775677919,-0.4868704081,0.2007485777,-0.0901124328,-0.2106387615,-0.0502286069,0.2731007338,-0.1976573616,-0.0586690381,0.048704274,0.3209480941,-0.2871660888,-0.1133866534,0.0613894947,0.1384380758,-0.0575092696,-0.0124506019,-0.5742512345,0.0052600536,-0.0729009211,0.0190999024,-0.0711815506,-0.2797948122,-0.1768175215,-0.1025363579,0.3666108549,-0.4066238999,-0.3093332052,0.1051588356,-0.0828099474,-0.2622595131,0.2527459562,-0.1832671762,0.009008484,-0.2272030413,-0.1686164886,-0.0905781686,-0.142659843,-0.1856201589,-0.0608413741,-0.4203020334,-0.2023369521,0.2431054413,0.2142650932,0.0216510389,0.2697407901,-0.3960029483,0.0612482354,-0.1463420838,0.1326318234,0.0516977869,0.8249756694,0.0901127905,0.1683174521,0.0149085028,-0.230446279,0.0307847075,-0.004948142,0.1637962162,0.2023214549,-0.3884166181,-0.108568117,0.0186736174,0.1368117183,-0.2402224094,0.1545701921,0.2006679773,-0.1630881876,0.0166011509,0.0535856821,0.1790063679,-0.3140646815,-0.1629086733,-0.0846173838,0.1044438705,-0.2316454649,-0.0612418763,-0.0360046886,0.1628044546,-0.0336622857,-0.0257635731,-0.1328804195,0.2433867753,-0.079946436,0.2389126122,0.1943679899,0.4010286927,0.1040250361,-0.232250303,-0.0188024305,-0.1810192168,-0.2169619948,0.2639425695,-0.1946939677,0.3217251301,0.1168457717,-0.136943385,-0.059640076,0.1133780405,0.076882571,-0.0098923985,-0.3611055315,-0.0982194915,0.2110361606,0.0818047523,-0.4671806693,-0.3771185875,-0.2357837707,-0.0583927073,0.2133108824,-0.0220660213,0.1872704178,0.1643862575,0.0050910958,-0.2744437456,0.319427222,-0.1886038333,0.0777855292,0.373283416,-0.2062876225,-0.073380433,-0.1037016809,0.1833920181,0.0525205769,0.2987331152,-0.0065756799,0.0300998259,0.3607836068,0.1661983728,-0.3111253083,-0.2824747562,-0.0655037239,0.074151814,-0.1006521508,-0.2895946801,-0.0692490265,0.1158679575,-0.0269074682,0.1567314416,0.1119908094,-0.1380441189,0.3437453508,0.2363645285,-0.0801676139,0.3129836917,0.0532259718,0.7220324278,-0.116899386,0.2209923416,-0.1348646432,0.1317929476,-0.2741573751,0.2003955543,-0.3114868402,-0.1603976041,0.5469685197,-0.0862385482,0.3748144209,-0.1230098754,-0.7809898853,0.0182593185,-0.1341382712,-0.2427570671,0.1018486619,-0.2668537498,-0.0622558743,0.0510766879,-0.0508656502,-0.3411754072,-0.5344321132,-0.0868705437,-0.3874756992,0.137857005,-0.3504896462,-0.435923934,0.1890019029,-0.1960467696,-0.1383478642,-0.1172400191,0.1319788396,0.2151584178,0.0350902826,0.094738692,0.0459458977,0.0550254211,-0.2530817688,0.2055948526,0.3684577346,-0.2830173373,-0.3265470862,0.2879656553,0.0631728098,0.0560587756,0.1893425137,-0.4470998645,-0.4195119441,0.1434142739,0.1084284261,-0.1350639462,0.0135688521,0.7221037149,-0.0184357353,-0.1925778389,-0.1101450175,-0.53260988,0.1829788685,0.5376421809,0.188729763,-0.2814339399,0.2907653451,0.2473984063,0.3881089091,0.2700050175,-0.0351172537,0.0885069668,0.0548753105,0.1409626901,-0.1460552663,-0.0937455297,0.0280847922,0.0921357796,0.1769600213,0.462898761,0.1312185526,-0.3038619757,-0.2249210626,-0.3283392191,0.066381909,0.0750307143,-0.2265835404,0.1201867685,0.5775747895,-0.1256805658,-0.1779618561,-0.3741830587,-0.0288375411,0.1463678479,0.3406570554,0.1614884287,-0.0733088851,0.1061225235,0.1024344265,-0.3345720768,0.3589105606,0.1454829872,0.0309207905,-0.1277604252,0.0327313319,0.1614751667,0.0084108496,0.1002992094,-0.282572329,-0.0826107189,-0.1312268674,-0.1317445338,0.2073327601,-0.0904586837,-0.136823222,-0.235954985,-0.0495570526,0.0843731314,0.0452182479,-0.0389382876,0.7889192104,-0.2510404885,-0.0095922314,-0.1640725732,0.0475807264,-0.299313128,0.1067073792,0.0112735601,-0.1068122238,-0.0297446605,0.0377518535,-0.1933379024,0.3182402253,-0.0514369681,-0.086913608,-0.0083162822,-0.1852336377,0.0985035524,0.0364756249,0.0846017003,-0.0214071423,0.3214014173,0.0437427871,0.2807782292,-0.3044239581,0.1388286203,-0.2575796545,0.0701567605,0.1772870272,0.0339201838,-0.0396525413,-0.2024119943,-0.233607322,-0.0368360206,-0.2508929968,0.406306982,-0.00908734,-0.2250766158,0.0722713023,0.1944386065,0.0866824687,-0.2134993374,0.3258743584,0.2731483281,-0.1404895484,0.2669580877,-0.0046340288,0.7183266282,0.3461143374,-0.2011156827,0.2221728861,-0.1429887414,0.4602639377,-0.0865991041,-0.0558302216,0.2339565605,-0.1750221848,-0.0732434094,0.2484050989,0.3967687786,-0.2224189937,-0.3416428566,0.1455119699,0.3324221671,0.3283753991,0.1891310811,0.1741768271,0.4224458933,-0.1886374354,0.2463038266,0.1276967973,0.0522835217,0.2442716211,0.0974935144,-0.0891623497,-0.1090648472,-0.3499188423,-0.4321232438,-0.1855127811,0.3060638607,-0.0470060445,-0.0166960731,-0.0338804014,0.4815781116,-0.339944303,0.2531513274,0.1209667996,-0.3116705418,0.1763544977,0.1864216626,-0.1396140158,-0.0212109443,0.0043305219,-0.0761719942,-0.2574709654,-0.0868249312,0.211904034,0.1781587899,-0.2674174607,-0.1282445341,-0.1877415925,-0.1005506292,0.1406441033,0.4012115896,0.6115493178,0.1359340101,-0.3031636477,0.197253108,0.1780744344,-0.1567554474,-0.2139320076,0.0621183552,-0.0219098292,-0.1457689553,0.3291982412,0.4112918079,-0.3497694433,0.0214742646,-0.1107354537,-0.0446328148,-0.2222426087,-0.1674980968,0.0854737461,-0.6017765999,-0.2554288507,-0.0341730453,-0.2033930272,0.1958699375,0.7892417312,0.0958763137,0.183666721,-0.1987364739,0.0008463892,-0.1632602811,0.0411737971,-0.483723104,0.2972729206,-0.1245325655,-0.0901405811,-0.0784069449,0.163760528,-0.3665677905,-0.1950000077,0.1776394993,-0.1568974704,0.3508590162,0.244912982,0.2970455587,0.1198674813,0.1339178085,-0.0894430727,-0.2211396247,-0.1161564216,0.0442933626,0.1068718135,0.0547208227,0.3294250965,-0.0133787496,-0.1398479939,-0.2972803712,0.1549500376,0.1967650503,-0.0851160437,-0.038449727,0.1937953532,-0.2134046853,-0.0467604212,0.0949220732,0.3463662565,0.1956269145,0.0363109,-0.0518069416,0.4867759943,0.0589833558,-0.0996605232,-0.0848693252,0.3457535803,-0.3071815968,0.3234304786,-0.0567583852,0.0773623362,-0.1113125011,-0.0675248727,0.0764101073,-0.5348697305,-0.0658813938,0.1311749518,-0.0504486039,0.2584062815,-0.3335230947,0.0924864262,0.029950168,-0.1052725986,0.1402956694,0.1756880134,0.3292538226,0.2189726084,0.9014947414,0.0691317171,-0.0746648014,-0.100260675,0.431441009,0.1574966311,0.1056706682,0.1086158603,0.1242473871,-0.0149076572,0.2548089921,-0.1625202,0.165304333,-0.0493630879,0.4765473604,0.4369476438,0.2106315941,-0.2653195262,-0.0030056052,0.2411319017,0.197799623,0.1195811853,0.2271252573,-0.1336301416,0.0978337899,0.1558500975,-0.8371521831,-0.0796554759,0.0518329293,-0.1005760729,-0.1086158454,-0.0858419761,-0.4607277811,-0.0625791699,0.3197573721,0.1980779469,0.1276978254,-0.306539923,-0.1254784018,-0.0792138055,-0.0986674502,0.1094344705,0.224629581,-0.035257142,0.3084667027,-0.5054548979,-0.3232398629,0.2835775614,0.3739507198,0.0568750352,-0.2342677116,0.469181776,0.1160808206,0.0785395429,-0.1513750404,0.0903927907,0.1426035166,0.00019239,0.1041196138,0.1485255212,0.1165400445,-0.1035360992,0.2095784694,-0.1587474942,-0.1455764472,-0.3073990345,0.162215054,-0.0058911229,0.0732277781,-0.0806858242,0.2510504425,0.1243724525,-0.3320806324,-0.2047232389,0.0086557604,-0.2165244222,-0.1781672835,0.0944767371,-0.3145744205,-0.1393760741,0.3247355819,0.0284531787,0.0982789025,-0.5078366399,0.0375952348,0.3588588238,0.1305332631,-0.1337831616,0.1626102626,0.2412441373,0.1967028379,-0.1768679619,0.2850234807,-0.3694350421,0.3337841928,0.1166632995,-0.2275416255,0.1349611729,0.5156612396,-0.0545513555,-0.0957691073,-0.0271320008,-0.1117659137,0.19722718,0.019734282,-0.2587211728,-0.0780601948,0.1497837901,0.0306151267,0.4778962731,0.1017924249,0.1346043199,-0.1709789187,-0.1296743304,-0.4870501161,-0.1663849056,-0.0266479664,-0.0536070019,0.0532120056,0.2029839009,-0.052554097,0.0919133797,-0.0367506631,0.3837091923,-0.0350507312,-0.0741489679,-0.1484145671,0.4973812103,-0.4166463017,-0.2811608016,0.1262328625,-0.2493222207,0.1309021562,-0.2178701907,-0.1375610083,-0.1085695773,0.2321492583,-0.0757678896,0.0885247141,-0.172472775,0.3168586791,0.2556537092,-0.0150290299,-0.3104695082,-0.0098979333,0.2076014876,0.3018361032,-0.1535345763,0.4014832675,0.0109265698,0.0803888887,-0.0943641514,0.0632768273,-0.2406885177,-0.010490072,0.0499435626,-0.3058010936]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"Great @nazneenrajani. let me know once done.\r\nThanks","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":8,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\nGreat @nazneenrajani. let me know once done.\r\nThanks","embeddings":[-0.0984604284,-0.4266276062,-0.1598725915,-0.4335260093,-0.534044385,-0.2286310941,-0.0564631484,0.066483885,-0.1666012257,0.437336117,0.1297692209,0.406981498,0.0080552027,0.2869954705,-0.0414330028,0.1756015569,0.1165005565,0.2966299951,-0.0338153131,-0.1281320304,-0.2061457336,0.1972557753,-0.1409226656,-0.1230612248,0.1850394905,0.1169243455,-0.2242874205,-0.2301250547,-0.1777700931,-0.0845569223,0.1908569932,0.3428768814,0.0720593333,0.0885169208,-0.0001109753,-0.1993190795,0.4054781199,-0.0222613309,-0.2152526826,-0.1041497514,-0.3607130051,-0.3759299517,0.0395881049,-0.0977439359,0.1538605541,0.1472418755,0.1366618872,-0.0208774172,0.0175955966,0.061701186,0.2497207671,0.271813184,-0.0017262027,-0.3295991123,0.3175432682,0.4677054584,-0.3979580104,-0.178125456,0.3202885687,-0.0604353994,0.0024289533,0.5021643043,0.3001122773,-0.2166237533,-0.0631554797,-0.1113526523,0.1735108793,-0.0089481436,0.1851784438,-0.0156360902,0.3485785723,0.2119536549,-0.3804816008,0.3945019245,0.0111551434,-0.4396688938,0.223045662,-0.1543727368,0.0077737505,0.0137101291,0.2133329213,-0.384305805,-0.4598852992,0.1724333763,0.0081853606,0.151943326,-0.1224118546,-0.2328099459,0.1095026955,0.1040858775,-0.1632295549,0.1091082022,-0.1683257222,0.1990684271,-0.2451204211,-0.2474337518,0.3356184363,-0.0913635567,0.4005572498,0.0598473661,0.1289174855,-0.0366756879,-0.0895482749,0.1248070896,0.0209670402,0.2541719377,0.4179026186,0.0708148032,0.3600858748,0.3803395033,0.5166504979,0.0606947504,0.1579957306,0.1183166876,-0.0865857825,-0.0284505002,0.015290455,-0.5321254134,-0.0338326693,0.0831718668,-0.1513517201,0.0397769399,-0.0266042352,0.0614872016,0.3296309114,0.156623885,0.3452825844,-0.1240245476,-0.0081185773,-0.39647156,-0.2681236267,0.0686143562,-0.2103271931,0.308416456,0.2520850897,-0.2993870378,0.1570482403,-0.3120833039,0.3172906339,-0.048049394,-0.2255782485,0.3977645636,-0.161890313,0.1251000613,0.0237823725,0.0327066369,-0.262206614,0.6085239053,-0.1196332872,0.1114372686,-0.2860949934,-0.4498589933,-0.4717455208,0.2482513934,0.0143280076,-0.2026568502,-0.0471846759,0.2849008143,-0.2319584191,-0.0281196199,0.0760913342,0.3362406492,-0.2413535267,-0.0600832403,0.0320217945,0.0300394744,0.0584044121,-0.0581125952,-0.5410783291,-0.0256238729,-0.07588052,0.0512987822,-0.041771464,-0.2927590907,-0.1421889812,-0.1134520397,0.3554275036,-0.4287962019,-0.1894410104,0.1237104237,-0.0813907012,-0.2682003677,0.2166692466,-0.1851745248,-0.0507999361,-0.1693673134,-0.1564828902,0.0041229874,-0.0807625949,-0.0830439106,-0.1764528602,-0.3811270893,-0.2834825218,0.2789177895,0.2977291942,-0.0171506796,0.2642216086,-0.4367787838,0.0561293848,-0.1603308916,0.1372807473,0.0049327314,0.8683258295,0.1084533855,0.1635189056,0.0296751745,-0.1622567326,-0.0327982157,-0.0847034529,0.2062029243,0.2036293596,-0.3830429018,0.0196370743,0.052990444,0.1406262368,-0.203823328,0.1946005821,0.2063080668,-0.1738993675,0.0095504615,0.0841032937,0.1504216939,-0.2806483209,-0.1106739268,0.0166645944,0.09734267,-0.2716980875,-0.0705886036,-0.1080852076,0.1881189793,-0.0423441306,-0.0413440801,-0.1226869449,0.2323552966,-0.1439270377,0.2294135243,0.1535075903,0.3440109491,0.1432592869,-0.2222134322,-0.0467869304,-0.1453260183,-0.1514824331,0.262336731,-0.1263309717,0.3240723908,0.1412366331,-0.1349403113,-0.0013148338,0.0236961637,-0.0412930958,-0.0624412782,-0.3139080703,-0.070960775,0.2513736486,0.0849703401,-0.5345712304,-0.3922075927,-0.2302638292,-0.0778032616,0.1345421225,-0.0121850418,0.1820549071,0.1959723681,0.0329131745,-0.253354311,0.2659390867,-0.2539847791,0.0892140269,0.3087373972,-0.2507514954,-0.1015602052,-0.0924313366,0.1653366536,0.1007976532,0.3016465604,-0.0337727517,0.1047964767,0.3633017242,0.1118326038,-0.2911680937,-0.3247349858,-0.1040002257,-0.0242059361,-0.0698118061,-0.2901803851,-0.0062027532,0.0005253625,-0.0075095864,0.1253960431,0.1389173567,-0.1222207546,0.3855186403,0.2697809339,-0.1761918813,0.3530724347,0.0833029523,0.7253095508,-0.1425898671,0.220638752,-0.1638322473,0.1403510869,-0.2423991412,0.2245521992,-0.2775268853,-0.2140369266,0.5213732123,-0.0615886487,0.3416377306,-0.1116407663,-0.7797102928,0.0201704968,-0.1062431484,-0.2635073662,0.0211244635,-0.3150078952,-0.0356463082,0.0203677192,0.03406699,-0.3807293773,-0.499160409,-0.0941608772,-0.2860929668,0.1237568185,-0.4136055112,-0.3581493795,0.2455969751,-0.2284523696,-0.1169609204,-0.1637234837,0.1291491389,0.2520574331,-0.0150209414,0.0939551294,0.0366896056,0.0867816582,-0.2155776173,0.1983382255,0.3639414608,-0.2892199159,-0.3299863636,0.1710053682,0.0118791498,0.0820347741,0.1449565589,-0.3222258985,-0.4897066653,0.2258739173,0.1472156942,-0.1275795251,0.03430783,0.6753582358,0.016888069,-0.2478757352,-0.1361996084,-0.5315277576,0.2198706418,0.4798232019,0.1649738699,-0.2808046043,0.2201123983,0.2319739908,0.3428451717,0.1397179812,-0.0305475853,0.1152571887,0.1448338479,0.083935,-0.096269004,-0.0520715564,0.0321182422,0.1251002997,0.1701325029,0.5376249552,0.1098371148,-0.3530765176,-0.2589733601,-0.3753066063,0.0678908378,0.0987410769,-0.2270084918,0.1266188174,0.5603125095,-0.1954197586,-0.2261537611,-0.3260574639,0.0071706683,0.1899200976,0.3711421192,0.1662632078,-0.1318006814,0.124631159,0.0434735566,-0.2243656963,0.3287175298,0.1971143484,0.0112357326,-0.0879436806,-0.0860859305,0.1572379023,0.0000515519,0.0999178663,-0.2032807469,-0.1437650323,-0.1789914221,-0.1508964151,0.3045862913,-0.0907022431,-0.106299825,-0.2998533845,-0.0752291232,0.0927243531,0.0894179419,-0.0787037313,0.7179254293,-0.3208882213,0.0395925939,-0.2033752799,0.0566517785,-0.2558920681,0.0890552402,0.0637344867,-0.1101151556,0.0005520643,0.0640937015,-0.2262568623,0.356130451,-0.0610587485,-0.0902587771,0.0251982715,-0.219325617,-0.0145309186,0.0819769353,-0.022700347,-0.0335387662,0.3673194945,-0.1033195704,0.2554474473,-0.3018244207,0.0799029022,-0.2435807437,0.0928935558,0.1914339364,0.0616438426,0.0309815891,-0.2723288536,-0.1475184411,-0.0223553609,-0.3409327865,0.358671993,0.0289600249,-0.1850708276,0.1636091918,0.1534837186,0.0276302993,-0.2354666293,0.2956339121,0.2670017481,-0.2248173058,0.2836906314,0.0229911935,0.7322015166,0.3987801075,-0.1828980744,0.2236297131,-0.1654896736,0.363150537,-0.1205328926,0.0633383021,0.1760601103,-0.1720389873,-0.0676886737,0.2536482811,0.3925435245,-0.2450269461,-0.3485922217,0.1331379712,0.3342761695,0.2986865938,0.1326000392,0.1446879357,0.4117774665,-0.2320210934,0.2552525699,0.1556757838,0.0852180868,0.2246253043,-0.0106860688,-0.111427024,-0.0598288067,-0.3336454034,-0.3251220584,-0.1421130598,0.311229378,-0.0355610922,0.0218386557,0.0361021981,0.4501746893,-0.412443608,0.2289155573,0.1217476204,-0.2186650187,0.1592608094,0.2261567712,-0.0451408438,-0.1047586352,-0.0040880553,-0.0600921586,-0.2735109031,-0.0738251284,0.2116038352,0.1498630196,-0.2211978137,-0.1613753587,-0.1227523014,-0.0685958862,0.2137014866,0.4420020282,0.6429758668,0.0834419206,-0.2512280345,0.2307418883,0.1395422369,-0.1870850027,-0.1745970845,0.0319069996,-0.0239825975,-0.1659363508,0.2969283462,0.349940151,-0.3703781962,-0.0547349826,-0.1399014592,-0.1325106025,-0.2542572021,-0.2317085266,0.0166598726,-0.4957679808,-0.2309152186,0.0407668836,-0.2827278376,0.1944627464,0.7969889045,0.0437466279,0.2325035632,-0.1782571524,0.0278444681,-0.1201661155,0.0386152789,-0.5678523183,0.2495937794,-0.1351405531,-0.04861008,-0.0298950765,0.1356121004,-0.4141657948,-0.2027241439,0.1763293594,-0.1208021492,0.3438341916,0.1936060637,0.3152340949,0.1233968288,0.1822042614,-0.0451956131,-0.1822737902,-0.1644218862,0.0702410564,0.0978215262,0.090169698,0.3106128871,0.0236952733,-0.0238800291,-0.2478199601,0.1247383058,0.223154828,-0.1486893445,-0.0878370926,0.2176296562,-0.2237930894,-0.0280873664,0.1509175003,0.3690072,0.1980968863,0.061790131,-0.1251654476,0.4082355499,0.0326042771,-0.1418799013,-0.0479715466,0.3731529415,-0.3143652976,0.322190702,-0.0452548377,0.0763766468,-0.1678780317,-0.1054929495,0.0254887268,-0.5389515162,-0.0538084023,0.0653003901,-0.026799256,0.3157739639,-0.2585015595,0.1061976776,-0.0686965212,-0.2379803956,0.1828433573,0.1886925995,0.2967681289,0.1296900511,0.8816441894,0.0821610615,-0.1134992242,-0.0548322164,0.4301608801,0.0915418118,0.1125326529,0.1149110496,0.1156862974,-0.0032756771,0.2270473689,-0.1983436793,0.1425747871,-0.1261375546,0.4419122338,0.496453464,0.1903985441,-0.1597566754,-0.0099324528,0.270963788,0.1797107905,0.114636384,0.2088964432,-0.1558624059,0.0825867057,0.1108892933,-0.7427728772,-0.035708569,0.0152432136,-0.1038098857,-0.121755667,-0.0667323619,-0.4875343442,-0.0499455445,0.3673542738,0.2100887001,0.1189517826,-0.2260130495,-0.0539558679,-0.0264348369,-0.1299854368,0.1452940851,0.3049142361,-0.0739770383,0.267319411,-0.5132334232,-0.3003462553,0.2104585469,0.3346031308,0.0673639774,-0.181038931,0.4370150268,0.1177267656,0.0768127814,-0.1082873791,0.0886619687,0.1234456003,0.079241544,0.1540174037,0.0835218132,0.1396485418,-0.1514026374,0.2231218368,-0.1510170549,-0.2057478279,-0.229825303,0.1430386603,0.0941929594,0.0669031218,-0.1024062932,0.2410683483,0.1254101694,-0.3311318159,-0.2310711294,-0.0202584937,-0.2224680334,-0.1574594378,0.117741935,-0.232385233,-0.0819546059,0.2388717234,-0.0071216356,0.0568799153,-0.5143882632,0.1615179926,0.3483859599,0.2143351287,-0.0891227424,0.0726843923,0.2210086286,0.1801141948,-0.2455265522,0.2711420953,-0.3863960207,0.3698768914,0.1491878182,-0.2800353467,0.230658561,0.4645965099,-0.1118566096,-0.020349687,-0.0800010189,-0.1048074365,0.1383961737,0.0584727228,-0.2333701551,-0.0247031506,0.1252730787,0.0494321845,0.4898329079,0.0703211725,0.1366277188,-0.2135000825,-0.1882422119,-0.3239864409,-0.1468973905,-0.0286840368,-0.0358011574,0.0906978846,0.1643945873,0.0190772302,0.0827687234,-0.0605895445,0.4286050498,-0.047221303,-0.1281690449,-0.1053880453,0.4482275546,-0.3674612641,-0.2982043624,0.0524722561,-0.2243446261,0.1500609368,-0.2271207571,-0.2082016915,-0.1553517431,0.2081321329,-0.0957112387,0.074215956,-0.1542081833,0.3486942649,0.2159940749,-0.007937693,-0.3450245261,-0.0463073812,0.1748753339,0.3117300868,-0.1486849338,0.3757424355,0.0094702262,0.0522767827,-0.1377326101,-0.0923318565,-0.2208012789,0.0155854048,0.1036510319,-0.3115336597]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"@mariamabarham @sarahwie I added them to the cos-e repo https:\/\/github.com\/salesforce\/cos-e\/tree\/master\/data\/v1.0","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":10,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\n@mariamabarham @sarahwie I added them to the cos-e repo https:\/\/github.com\/salesforce\/cos-e\/tree\/master\/data\/v1.0","embeddings":[-0.1105099842,-0.2899199128,-0.1491891295,-0.3627903759,-0.4523007572,-0.2256634235,-0.0282360576,0.1408634484,-0.2538501322,0.4498185515,0.0751570761,0.38797158,0.0425402336,0.2660450935,-0.1850065738,0.1662538946,0.0821270868,0.2847425938,-0.0684490278,-0.1758020222,-0.1656155139,0.2111146897,-0.0834199861,-0.115658991,0.033161398,0.0681043193,-0.2680976987,-0.0516228527,-0.2817241549,-0.1220236048,0.2226493806,0.2825840712,-0.0390282236,0.1958428323,-0.0001047351,-0.1277198344,0.3686774969,-0.0441772491,-0.2485123128,-0.106799759,-0.2639452517,-0.3507713675,0.1011684164,-0.1152972728,0.098225683,0.0323078968,0.0871762857,-0.0522180274,0.027208399,0.1533835679,0.3173853457,0.2568055093,0.0589353628,-0.3289894462,0.3780804574,0.4017496109,-0.3586454093,-0.1548779756,0.2950598896,0.0049424847,-0.1088919118,0.443107307,0.2574525476,-0.1022046357,-0.1898992658,-0.083000645,0.1722013801,-0.0694470778,0.214844659,-0.0230911672,0.3844826221,0.1637047678,-0.358619839,0.3150598705,-0.0461464338,-0.3562252522,0.240582779,-0.1433638334,0.0267632231,-0.0377082527,0.2623124123,-0.2052147985,-0.4243731499,0.1624577343,-0.0486270301,0.1569897532,-0.1900076568,-0.2381605655,0.1165211126,0.0138016203,-0.0860471725,0.0987017527,-0.1471157372,0.1349691004,-0.2075955868,-0.1906109005,0.4211822748,-0.0787604824,0.3761399984,0.0931291506,0.1730197966,0.0004846188,-0.1060821712,0.1129802465,0.0336052962,0.2757493854,0.2848389745,0.0913650692,0.4189934433,0.3602530956,0.3961263895,0.0149178552,0.2029134184,0.0254098978,-0.0974419713,-0.004295243,0.1424585879,-0.6087972522,-0.1649029255,0.0642765984,-0.1096345186,0.1026610062,-0.0329363532,0.1136755273,0.2644775212,0.1995950192,0.2871039808,-0.0581997782,-0.0806010589,-0.2997106016,-0.3630349934,0.0672503188,-0.1996574551,0.2172202021,0.3029484451,-0.2711045444,0.2499898523,-0.2992222905,0.3183801174,-0.0596313439,-0.2030145675,0.3204038441,-0.1078100801,0.187747255,0.0853634998,0.0095685413,-0.1606739461,0.5273944736,-0.119097285,0.0238477495,-0.3180026412,-0.466019839,-0.396156162,0.3189098835,-0.0268830564,-0.1677015573,0.04300249,0.2596033514,-0.1348244399,-0.0574474074,0.0315677822,0.3269695342,-0.1613130569,-0.1671221554,0.1028940156,0.0845443532,0.1094989777,0.0143589936,-0.5378427505,0.0228069928,0.0059078885,0.1188835576,-0.0311900433,-0.1919428408,-0.1140955016,-0.071478039,0.2888525128,-0.3547953665,-0.2358079553,0.1270564497,-0.1077576354,-0.2689869702,0.1766440272,-0.2336010784,-0.0483663082,-0.186305508,-0.1793400496,-0.0582139008,-0.0263043866,-0.0365621746,-0.1860681772,-0.3688216805,-0.2411829233,0.1830926239,0.2149324715,0.0186881069,0.3132036626,-0.4172572792,0.0149887288,-0.2359952033,0.1279733777,0.0905384198,0.7598071098,0.0722570717,0.1728650182,0.0803303793,-0.1946442723,-0.0208071135,-0.0954768285,0.2619996667,0.0830688179,-0.4140618742,-0.0523125157,0.0455783978,0.0803837925,-0.1944339275,0.2920073569,0.2019822598,-0.0968950912,0.0520556159,0.1286410689,0.1699697822,-0.2986956537,-0.1192315817,-0.0396893993,0.1938036233,-0.2914066613,-0.1423063576,-0.0048026857,0.2254898697,-0.0289290212,-0.0231968686,-0.1766593903,0.3264501989,-0.1096506715,0.268840611,0.0806187317,0.3169659972,0.0955481678,-0.2157376707,0.0198598523,-0.0164306704,-0.0398904085,0.1833677888,-0.2314926833,0.3272468746,0.0899580196,-0.0770549104,0.0072816424,0.0344210379,0.0584678128,-0.0862323493,-0.2815707028,-0.0743353665,0.2570330501,0.0388418213,-0.4544445276,-0.3035358489,-0.2429041862,-0.0309668891,0.1322577298,0.0047111316,0.1663066447,0.092572093,0.0631185025,-0.2668700516,0.2874673903,-0.1866489053,0.1605634242,0.3469513953,-0.0273273401,-0.058162231,-0.1434065402,0.0420711711,0.0520873256,0.2671890557,0.0442623496,0.1776610911,0.396291256,0.0810671225,-0.4205490649,-0.2584124506,-0.0821333975,0.0400332585,-0.1307285428,-0.2619461715,-0.1203723177,0.0058675702,0.0538100861,0.0459432974,0.1297233105,-0.2296670079,0.4137788713,0.3302611411,-0.2560935915,0.2699486315,0.1063216925,0.562898159,-0.0812864825,0.2825179994,-0.1876783669,0.0953949168,-0.2783951759,0.2822469175,-0.310554266,-0.078857325,0.4669961929,-0.1425058097,0.3224419951,-0.1679263115,-0.7417755723,0.0288674682,-0.1358955055,-0.22500211,0.1121225581,-0.266083777,-0.0249672756,-0.0316549614,0.0133699933,-0.4153401852,-0.4691419303,-0.0657257289,-0.3117731512,0.100803338,-0.4696816802,-0.5729985833,0.2031171769,-0.2866070271,-0.0627894998,-0.0561583154,0.1614972651,0.2848571837,0.0746788532,0.1335701942,0.1317504346,0.1045799479,-0.2942233682,0.1401828676,0.4072445929,-0.3546135128,-0.4529688358,0.1857894659,0.056809403,0.144585669,0.0930700377,-0.3211679161,-0.457375139,0.1295356452,0.074867256,-0.0928788334,0.0901726633,0.649158597,-0.0063629551,-0.2661207616,-0.176330775,-0.3897010088,0.2315919548,0.4149267673,0.1685947329,-0.3144434988,0.2944246233,0.2283575386,0.2388053834,0.1602139175,-0.027494859,0.1815392971,0.0646454319,0.1722006351,-0.1445020437,0.0121825291,0.086100176,0.0490002334,0.1106905639,0.532754004,0.1815988123,-0.3559790552,-0.2464389503,-0.3523410857,0.0903465524,0.0089325998,-0.2327065021,0.1209648252,0.4588823617,-0.1164131761,-0.1524167806,-0.3212913275,-0.0191888288,0.1105570197,0.3004916012,0.1500369906,-0.1182974577,0.0352351815,0.0759960935,-0.3472131193,0.364580214,0.1756302118,-0.0543532334,-0.1394251138,-0.0757893324,0.2111914754,-0.0824480355,0.0674093813,-0.1855696589,-0.0563967265,-0.0683053732,-0.1292234212,0.1987124681,-0.0873127133,-0.2064336538,-0.1594371796,-0.008049259,0.0733562037,-0.0597424582,-0.0961034372,0.6179479957,-0.2649606466,0.0562479794,-0.1010571271,0.0047879624,-0.3558262289,-0.0046196589,0.0055762483,-0.063158296,0.0401855335,-0.0880034789,-0.2393122762,0.3484443426,-0.0182610974,-0.106532976,0.1072998717,-0.0978515372,-0.0622486919,0.0209841784,0.0261177849,0.1001017839,0.3730151951,0.1235639006,0.2890373766,-0.2046006173,0.1695156544,-0.2598277628,0.1132205725,0.1992325932,0.0279960074,-0.0129475184,-0.2774799466,-0.1307318062,0.024508357,-0.1425509453,0.3381821513,0.0447965488,-0.1438383609,0.1664600074,0.243338719,0.1385165304,-0.2295667678,0.2499170899,0.2386157662,-0.2088542581,0.3042526543,0.0301872864,0.6975969076,0.4159961641,-0.164186582,0.2802120447,-0.1853696257,0.3803130984,-0.0974107236,0.0441281423,0.155428499,-0.1878449023,-0.0506310277,0.3071325719,0.3443891704,-0.1850291491,-0.37178123,0.1791502982,0.2749327421,0.2863356173,0.0988564044,0.082405813,0.2994071841,-0.2097744048,0.2169862837,0.2429351509,0.0862018019,0.1907436699,0.0193680227,-0.0777261704,-0.1432931721,-0.3954125643,-0.1694811881,-0.0407643057,0.2547591329,0.0407763384,-0.0059950757,-0.0438704565,0.2918638587,-0.3711519539,0.2335197926,0.0788049474,-0.2628515065,0.1451218873,0.2281831652,-0.0208768398,-0.0839189962,0.0377429314,0.0450423807,-0.2623903751,-0.1283988506,0.1140303761,0.0966320261,-0.2215798646,-0.107605122,-0.1461038142,-0.0823222622,0.085809581,0.5179097056,0.5523250699,-0.0033419796,-0.2931996584,0.2802454531,0.1532690972,-0.1631631702,-0.0682566315,0.0698630586,-0.0753262714,-0.1806662828,0.3219983876,0.3726559877,-0.1846349686,0.0146117341,-0.0857422724,-0.1169222444,-0.2965170443,-0.2537272274,0.0571980253,-0.5642811656,-0.1426914483,-0.0200153887,-0.2880688012,0.2746486962,0.8212579489,0.0281812139,0.241767168,-0.2316553742,-0.0152693149,-0.2238126397,0.0895221606,-0.5219019055,0.314652741,-0.1487479955,0.0060522589,-0.0260257516,0.0523863398,-0.4775229096,-0.1338966191,0.0946689472,-0.0715300068,0.3194737732,0.097200878,0.3627783358,0.0971913412,0.2349120975,-0.0309780259,-0.1978642493,-0.2355967909,0.0415611528,0.0737291202,0.0188546069,0.2720523775,-0.0434822477,-0.0826031491,-0.352789253,0.0926365703,0.2803263068,-0.1191127971,-0.0864521563,0.1151786149,-0.2199023664,-0.0051848413,0.1899199784,0.2465656996,0.1942081004,-0.0913007706,-0.2004462332,0.3648982346,-0.0607910529,-0.1338200867,0.0118822856,0.271428138,-0.2372850031,0.2237846553,0.0541642494,0.0720824823,-0.1781941801,-0.1326787025,0.1481766999,-0.5120142102,0.0335779823,0.0610363148,-0.0491789505,0.3607150018,-0.2879689038,0.0703356713,-0.1026512757,-0.0621998459,0.2429291904,0.1457651258,0.3370699286,-0.0073085879,0.7488498688,0.1381836534,-0.0328688771,-0.0233806781,0.319083482,0.13052167,0.0222669784,0.1073380038,0.1697164029,-0.1119200289,0.2321375906,-0.1291299909,0.2049805373,0.0217676274,0.403354913,0.3894528747,0.1528523415,-0.1114502251,-0.0240901429,0.1942880154,0.2178132981,0.050361976,0.1989265829,-0.1413270384,-0.0307427999,0.0786099285,-0.7540009618,-0.013771602,-0.0790055469,-0.159894377,-0.1432564706,-0.1217869967,-0.4405770898,-0.1393083781,0.2518351674,0.2156273872,0.2049375623,-0.3067447245,-0.081661813,-0.1354237795,-0.1505788416,0.1114154384,0.2878088355,-0.015170278,0.3465285599,-0.3757772446,-0.2742931545,0.1826489121,0.3991624713,0.1760850549,-0.1293388009,0.3729693294,0.1191688478,0.1002201363,-0.1855455339,0.0839382857,0.1724279821,0.0112558175,0.0648662001,0.178646788,0.2047922164,-0.2081614584,0.2376348674,-0.1246998161,-0.1080382466,-0.2103197277,0.1783776879,0.050262589,0.0043722419,-0.0316194706,0.1201637164,0.0131646572,-0.3780833781,-0.2519471645,-0.07692004,-0.1520031542,-0.2323385477,0.1504140496,-0.2054394782,0.0078160837,0.2236978114,-0.022499688,0.1478531957,-0.5128591061,-0.0187211893,0.3707717657,0.1359564513,-0.1825648844,0.0508883558,0.1994834691,0.1350027919,-0.2513659596,0.2384128124,-0.3279833198,0.2555198073,0.0518879555,-0.3774158657,0.2089868337,0.3233862221,-0.1396609545,-0.0443156622,-0.1004155055,-0.1508947611,0.1285553277,0.1424012035,-0.2117396295,-0.0286195781,0.0846318901,0.0874105096,0.4348007143,0.1372354329,0.1435124725,-0.258476913,-0.1243948936,-0.4219454229,-0.18238765,-0.0295909159,-0.0046750568,0.0922084972,0.2397895753,0.03642435,0.081090495,0.027061278,0.3418522775,0.0226038229,-0.0338368677,-0.1158605441,0.5644187927,-0.36571455,-0.240909189,0.0250463169,-0.1482197344,0.1300824732,-0.2530510426,-0.1342455894,-0.3193638623,0.3017372191,-0.1424193233,0.0449940041,-0.1466437429,0.3297568858,0.266621381,-0.0356907323,-0.331687212,-0.0233679675,0.1581693888,0.2840459943,-0.1770391762,0.2929292917,-0.0418623313,-0.0669914857,-0.1183208302,0.0213894714,-0.1611606777,-0.0141603183,0.171817407,-0.3577920198]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"You can now do\r\n```python\r\nfrom nlp import load_dataset\r\ncos_e = load_dataset(\"cos_e\", \"v1.0\")\r\n```\r\nThanks @mariamabarham !","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":17,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\nYou can now do\r\n```python\r\nfrom nlp import load_dataset\r\ncos_e = load_dataset(\"cos_e\", \"v1.0\")\r\n```\r\nThanks @mariamabarham !","embeddings":[-0.2524742186,-0.2647848129,-0.1142224446,-0.4864544868,-0.3794747293,-0.208027944,0.080287829,0.1892701834,0.117797479,0.3005792499,-0.1557758898,0.5226972699,-0.1476927847,0.4125771224,0.1462089717,0.1017268375,0.0327461325,0.3626562059,-0.1460937411,-0.2017550319,-0.2499855608,0.2243396193,-0.170479089,-0.0546534024,0.2192804813,0.0769513175,-0.29594329,0.079357408,-0.0616971627,-0.1953394562,0.3291566968,0.3292326927,0.1830692142,0.0646769628,-0.0001042149,-0.2351597548,0.2574845254,-0.128415674,-0.2869003117,-0.1815093756,-0.1894630939,-0.4737758636,0.2212621272,-0.2176605612,0.1091976464,0.1245936006,0.2031544,-0.1243735552,0.1270561814,0.2281515151,0.3232874572,0.4082860947,-0.0342665724,-0.1849743575,0.2077931166,0.2706400156,-0.2086310685,-0.0129490839,0.159164995,-0.2754572034,-0.1968272328,0.4057946503,0.1401126981,-0.0958480313,-0.1321593523,-0.0657319725,0.160822317,0.0215867739,0.0102871815,0.035435807,0.2350435704,-0.0497151688,-0.3959866166,0.1401738226,0.0289128553,-0.5668572783,0.0933171883,-0.0871446431,-0.1245133057,-0.1430565119,0.2074773461,-0.1623462439,-0.4726059437,0.2572447062,0.0048299152,0.2947573066,-0.0636480153,-0.2017008066,0.2266446352,0.0425014384,-0.0799821243,0.1515029669,-0.0328291841,0.3442159891,-0.532577157,-0.280259788,0.3405756354,0.0903657824,0.2678717971,0.0426026657,0.2928168476,0.141662091,-0.0038665642,0.1931778044,0.0302175097,0.2401251048,0.2926645279,-0.0412991047,0.4281789064,0.265352577,0.3137462735,0.0563644879,0.0704402551,-0.1089823171,-0.1450002193,-0.0684927106,0.0081538325,-0.4685888886,-0.2638157904,-0.053575594,-0.0558466762,0.1455134749,-0.0139182229,0.2732973695,0.1701271832,0.2448029518,0.4300428033,0.0092152497,-0.2248602808,-0.3951990008,-0.2916809022,0.0973813161,-0.3190465868,0.0840040892,0.3731903732,-0.2006007433,0.2742324769,-0.4334007502,0.2543345988,0.0711415559,-0.1991913319,0.2366589457,-0.1144714132,0.0614363663,0.2251132578,-0.0429788195,-0.238359943,0.2062325627,-0.208387658,-0.0808214471,-0.3151924908,-0.3829509318,-0.4068824351,0.2756214142,-0.0551683418,-0.3148973286,-0.1343500316,0.3234367371,-0.1739309579,-0.1865482032,-0.0361932665,0.1392180175,-0.2541751266,-0.1285971105,0.0910563022,0.1125260368,0.0546063036,-0.1368647069,-0.5214048624,0.0013883548,-0.0100025153,0.0492770411,0.021899486,-0.1245969906,-0.1308999807,-0.1309102178,0.5791584253,-0.3678940833,-0.0639109239,0.1252523661,0.0537508205,-0.288112998,0.0940648243,-0.171425581,-0.1428023726,-0.1066986248,-0.1012187973,0.3418894708,-0.0704961494,0.0995737687,-0.1165204719,-0.3035582006,-0.012261508,0.3013634682,0.2912403047,-0.0810912326,0.233009398,-0.1553655267,0.2071705461,-0.2782928348,0.0341411792,0.0545125641,0.5859817266,0.0462901928,0.0990734473,-0.0517431013,-0.3171615601,0.0166788325,-0.1142725125,0.2951973379,0.1128331125,-0.2925976515,-0.0322293863,0.0540716574,0.1180363968,-0.1114925593,0.2614135742,0.1112518013,0.0232048314,0.0792037472,0.074082613,0.3704974949,-0.1423626393,-0.1394184828,-0.2519687116,0.1947321892,-0.2114014775,-0.0909537897,0.0903237984,0.3076283634,-0.0141470851,-0.0164915174,-0.0900452584,0.3420787156,-0.197724387,0.3473968506,0.0661397278,0.1199117377,0.031651888,-0.1870265901,0.1046212167,0.0204780828,0.105019249,0.1983437836,-0.1874218136,0.4752044082,0.0901664048,-0.0594587885,0.093907468,0.0329540819,0.1509225816,-0.23204647,-0.3025615215,0.0127726402,0.387075305,0.1158940941,-0.3680178821,-0.3732737005,-0.3721951842,-0.0663403645,0.156610027,0.0056959055,0.217738837,0.1688701808,-0.0229636282,-0.244408235,0.2332585305,-0.0959547982,0.312954396,0.333904177,-0.088454403,0.0366148315,-0.3495225012,0.0381885469,0.1847692132,0.2104589939,-0.0037193808,0.181754753,0.4397560954,-0.033547584,-0.4896511137,-0.4122570157,-0.2240456194,0.048277352,-0.2493998855,-0.0127507159,-0.061692778,-0.14514485,-0.0550506599,0.0944197103,0.0958611444,-0.3651608229,0.2836418748,0.3310318589,-0.2519614398,0.1996840537,-0.0112702306,0.6269270778,-0.0407500565,0.1283237338,-0.2383580953,-0.0503174812,-0.2411791235,0.2618356347,-0.241127938,0.0310927499,0.4830776155,-0.1881505549,0.2633421421,0.0512588657,-0.6950424314,0.0192280635,-0.1217725649,-0.0672427341,0.1074874252,-0.3113571405,-0.1011445522,-0.0694991052,0.0436006822,-0.6364917755,-0.3448207676,-0.0135144899,-0.2336722165,0.166044116,-0.4094178975,-0.4348336458,-0.1046345457,-0.3718426228,-0.1279685348,0.0990161672,0.1822295785,0.3901272416,-0.0058501712,0.1749867499,0.0610397719,0.0588261858,-0.3134266734,0.179340139,0.4206720293,-0.4030599892,-0.5246385932,0.0087475618,-0.0924663991,0.3225848973,0.0916909128,-0.1163844243,-0.3931524158,0.171018675,0.0336882435,-0.1324877441,0.270876646,0.4044108391,-0.0321612321,-0.210416168,-0.1765942127,-0.2458321452,0.1508519351,0.4518745542,0.1313406825,-0.2489530593,0.2793438137,0.0862229019,0.1133267358,0.1045460254,-0.1675508916,0.2823275328,-0.0276979227,0.2028851956,-0.00161795,-0.000430525,0.0732810274,0.0729724988,0.2253548503,0.5441488028,0.1401338428,-0.4167070389,-0.3592557013,-0.3044947684,-0.0382961594,-0.0881944522,-0.0575658157,0.2055100352,0.3595193624,-0.0280685741,-0.1992827356,-0.4128101468,-0.2024634033,-0.012766419,0.2373701632,0.0040865862,-0.0650852844,0.0963450149,0.0386572257,-0.4979262352,0.3819626868,0.1151311994,0.0731688514,0.0380203873,-0.0845394284,0.0638466775,-0.1043224111,0.1870593727,-0.1114227697,0.0054573738,-0.0564449206,-0.1029585376,-0.1770634502,0.0564194955,-0.3784660399,-0.1969332397,0.0758086443,-0.009620335,-0.0009147953,-0.1784561574,0.6600254178,-0.0800711811,-0.057160344,-0.1971149296,-0.0125376321,-0.4126778841,-0.0750907883,-0.0036725111,-0.1563310325,0.2179157734,-0.1202236786,-0.0820919499,0.1733660251,-0.0176468622,0.1186634898,0.2828460038,-0.0711320713,0.1216849163,-0.0152173536,-0.0258024856,0.0783257559,0.1132290587,0.1190390661,0.1873854548,-0.0626782551,0.0500757806,-0.0578701124,0.2915523052,0.1874192059,0.0929489955,0.0229431409,-0.3587073684,-0.0909989625,-0.0888163969,-0.0064678146,0.285156548,0.2302761823,-0.299284637,0.0183602031,0.4084853232,0.1127875745,-0.2255141884,0.1382583827,0.0354362801,-0.4326172471,0.5216682553,-0.0476075597,0.68146348,0.2664853334,0.0666795745,0.4089911878,0.0198040027,0.6191624403,-0.1282190531,0.0904502049,0.0086136619,-0.0749348849,-0.0283187665,0.2580912411,0.2802236676,-0.1453398019,-0.3026856184,0.3580352664,0.4741484225,-0.0282293465,0.0660802871,0.1906586587,0.2675009966,-0.3572769761,-0.1759170294,0.1953857094,-0.0574008711,0.1983058602,-0.0095139528,-0.0763243288,-0.0335256644,-0.3446197212,-0.0948351994,-0.0046190377,0.4087622762,0.0627283081,-0.0131351547,0.0001819658,0.3545754254,-0.2510665059,0.2068149447,0.1270359606,-0.2332005203,0.134161666,0.0304149128,0.0334758349,-0.0575642139,0.06580697,0.1772830039,-0.2168302238,-0.0924753696,0.1330929697,0.2801997364,-0.2276493907,0.0751197785,-0.0599246398,-0.0967974737,0.0991335809,0.2709746957,0.5199604034,0.0423270799,-0.2932431102,0.2494106591,0.2173015326,-0.247353673,0.1259591877,0.1490783542,-0.1499185562,-0.1261404753,0.4295510054,0.3287899494,-0.1759683192,0.1623607725,0.0284766518,-0.0659816861,-0.2654573321,-0.2450836152,-0.0956838578,-0.5750576854,-0.0652817339,0.1114283204,-0.1890367419,0.2021917999,0.7269644737,-0.039208848,0.2074633688,-0.1095714644,-0.2568229437,-0.2760748565,0.0506777503,-0.4521549344,0.2244724482,0.1467543393,0.0574889593,-0.057872992,0.08658503,-0.4668600857,-0.2237445265,0.0735446736,-0.0111879315,0.3666561544,0.0840821266,0.3317528069,0.070025079,0.2591494918,0.0170924775,-0.1813287735,-0.2714544237,-0.1252371222,0.0788597316,0.021880284,0.1711352617,0.0980693176,-0.0567502081,-0.273042053,-0.05970148,0.1882501394,-0.0853996575,-0.1864282787,0.3759683073,-0.2097960263,-0.1313131005,0.3740197718,0.2997764349,0.2019176185,0.0131763183,-0.1052276269,0.4919641018,-0.0627124608,-0.1072225645,0.0412870161,0.1994612068,-0.2113650888,0.1068022698,0.0834738985,0.0005001425,-0.2596773505,-0.2030632645,0.2125793546,-0.2958903313,0.0158599038,0.0698932633,0.0752128065,0.332896173,-0.4336086512,0.1551664025,-0.0699419901,-0.1500570625,0.3807333112,0.1659959257,0.2593375742,-0.1449136734,0.6386486888,-0.0111857625,-0.0193987377,0.0635231435,0.2700661123,0.0073323469,0.0831864923,0.0074915076,0.0672960877,0.0025866027,0.2614638507,-0.0244696736,0.043506965,-0.0263967104,0.1955511868,0.3980351985,-0.004165641,-0.2185803652,-0.0406354256,0.2692481279,0.1335300654,0.2203333974,0.1255196929,0.0139963087,-0.0264746305,0.0733945221,-0.6518132091,0.1022446081,0.0463339686,-0.2674894631,-0.1924035698,-0.0601472855,-0.5283681154,-0.0347709917,0.1602236331,-0.0957058147,0.1165722534,-0.0516646095,0.0750947148,-0.1783219725,-0.0644858629,0.1521918327,0.2097459137,-0.0724275559,0.2600594461,-0.3143092096,-0.1335666925,0.1131693721,0.3317927718,0.2536153495,-0.0798392743,0.2513802648,0.1893157512,0.0781903192,-0.063410677,0.0155419558,0.127496466,0.1751867682,0.1567927748,0.3456556201,0.1996743083,-0.2353816032,0.1090392023,0.018087171,-0.1851761341,-0.1878706813,0.2947611213,0.2506922781,0.1279386878,-0.1549279839,0.0566547923,-0.2226935625,-0.3151550293,-0.0565547682,-0.041626621,-0.0657914653,-0.2692038119,0.1421039999,-0.1014509797,0.2428993881,0.2020774037,0.3052968085,0.0752405673,-0.4313354492,-0.1671844423,0.2120477259,0.1855373532,-0.2335086912,-0.0426983982,0.1669409573,0.1464369744,-0.2412617654,0.1020103842,-0.1333176643,0.4249044359,-0.0015242463,-0.3902807236,0.159053728,0.1435409486,-0.0917147994,-0.0373319946,-0.2592243254,-0.3049336374,-0.1383423656,0.1372608393,-0.2919712365,0.0967427716,0.0066001336,0.1337219179,0.4191933274,0.0297125448,0.2779367268,-0.3339419961,-0.0068876762,-0.3844679594,-0.2501214445,-0.0622773357,0.0465739667,0.0962267295,0.1779742092,0.0200274773,0.186679095,0.0494944341,0.3927054703,0.0339296758,-0.2314746976,-0.0423311107,0.5437826514,-0.3518792689,-0.2071861178,-0.0782560557,-0.0048726927,0.039167285,-0.1339019388,-0.0551983826,-0.3065366447,0.3257248998,-0.3351883292,-0.0586281531,-0.071957387,0.347671777,0.1555516124,-0.0065174177,-0.3395654857,0.1631650776,0.1424868852,0.1623525918,-0.0340319201,0.398617655,-0.1481129676,-0.0120549574,-0.2176659405,-0.1681087315,-0.0643277019,-0.0848139673,0.131564185,-0.4343095124]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"Thanks!","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":1,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\nThanks!","embeddings":[-0.0335155129,-0.454898566,-0.1642210633,-0.4509503841,-0.5463636518,-0.2034063488,-0.1133813262,0.0461733714,-0.1995286793,0.4774852693,0.1932146102,0.3546816409,-0.016773615,0.2678667903,-0.0421214253,0.1539986432,0.1074425355,0.3011286855,-0.0259964839,-0.1337000132,-0.1678483188,0.2148818523,-0.1721268296,-0.1121395677,0.2067897171,0.1625837088,-0.2434387356,-0.3054930568,-0.1708604395,-0.0694880933,0.227175504,0.32649526,0.1099894717,0.0612294264,-0.0001173351,-0.2014687061,0.4456411898,0.0072261519,-0.1943401843,-0.1444909275,-0.3536531329,-0.4165662825,0.0426067188,-0.080226846,0.1977087557,0.1085069031,0.1735034585,-0.0057337349,-0.0222746544,0.0035838091,0.1903860271,0.2628206015,0.0180327594,-0.3309567571,0.327888757,0.5360342264,-0.4217381477,-0.1991113573,0.2862053216,-0.0675288439,0.0673314929,0.5051828027,0.3092705905,-0.2749612629,-0.0678363368,-0.1127932817,0.1979076713,0.0125789791,0.1636929065,-0.0068760025,0.3842322528,0.2601342201,-0.4226760864,0.3870058656,0.0293890834,-0.4492206872,0.2439102679,-0.1750587523,0.0295040365,0.0537144579,0.215265438,-0.4494636953,-0.4585649967,0.1869870871,-0.0008022494,0.1814904809,-0.0908074379,-0.209627524,0.0877698213,0.1102291942,-0.1747121662,0.121331498,-0.1757067144,0.2095564902,-0.1768392473,-0.2473857701,0.354469955,-0.1330892593,0.4062245488,0.0833318755,0.1383985579,-0.1437189132,-0.1412730068,0.0686470866,0.0880829096,0.2546902895,0.4674052,0.1247007027,0.3099298775,0.3725231886,0.5642582774,0.0692703649,0.153383404,0.1332696229,-0.0661924928,-0.0122173401,0.0785319582,-0.5476745367,0.0294280872,0.0581745282,-0.2073272616,-0.0095566111,-0.0423524454,-0.0013001775,0.3498433828,0.2072217613,0.3097729683,-0.1089371741,0.0284394044,-0.4191375077,-0.2518110275,0.0543908738,-0.1640347838,0.3380824924,0.2325703651,-0.3264151514,0.1324386448,-0.3107689321,0.3043555617,-0.0864846781,-0.2191865146,0.4543616176,-0.2005060464,0.1464598179,0.005387967,0.0701741353,-0.2688889503,0.66449821,-0.1416361034,0.1794375032,-0.2817288637,-0.4724205732,-0.4875608087,0.1915043592,-0.0135899875,-0.1871726662,-0.0518165044,0.2701691985,-0.203778103,-0.0238658991,0.0883181691,0.3138701916,-0.2938065827,-0.064651303,0.026175024,0.0121672768,0.0562251955,-0.0925775096,-0.5747742653,-0.0293175522,-0.0977717564,0.040335767,-0.0645620078,-0.2836513817,-0.1520722359,-0.1426868886,0.3531985879,-0.4514945447,-0.1993733644,0.1356552392,-0.1068740562,-0.2238928825,0.2575457692,-0.1578735411,-0.0572688319,-0.1649316549,-0.1503144205,-0.0715463683,-0.1120077595,-0.1760062724,-0.1292197853,-0.3942312002,-0.2309372127,0.2524298429,0.286626339,0.0009804267,0.2725093067,-0.4167281091,0.0662223771,-0.1254604757,0.1249754578,-0.0324447714,0.847279191,0.11871811,0.1682491153,0.051382523,-0.1455026418,0.0007967189,-0.0691466331,0.1753814369,0.2571578026,-0.3944887817,0.0370473862,0.0686966851,0.1952143162,-0.2120803893,0.1272846013,0.1968524009,-0.2059603035,0.0191932004,0.044541955,0.1349998116,-0.2843279541,-0.1042958349,0.022931084,0.100843668,-0.2325702608,-0.0582270101,-0.1489853114,0.1760208458,-0.0396253653,-0.0137259299,-0.1193952635,0.2054267377,-0.1190036759,0.2240747958,0.1721908748,0.3666245341,0.1409755349,-0.231292218,-0.0370084755,-0.1510464549,-0.2411237061,0.233299166,-0.1098321229,0.2960204482,0.1413518935,-0.1278763264,-0.0376372412,0.0495419949,-0.1021801978,0.0090900138,-0.3876464963,-0.0734171048,0.2515737414,0.0522753969,-0.5475145578,-0.4050464034,-0.2148596048,-0.1171039417,0.1403880566,-0.0437726155,0.1619343907,0.191712603,0.0637826622,-0.2577779591,0.2697466016,-0.2818771601,0.0427770689,0.3146449924,-0.2698205411,-0.0972133577,-0.0946353748,0.2285823375,0.0883710608,0.3109023571,-0.0580267422,0.0434798785,0.3607960343,0.1547819227,-0.2615215182,-0.2922334075,-0.1104988679,-0.0302600246,-0.0509891994,-0.3147245049,0.0233323369,0.0526040643,-0.0341056511,0.1398358792,0.1119583771,-0.0987113938,0.3535954952,0.2769578695,-0.1153570116,0.3558600545,0.0947201028,0.7465165257,-0.1531164944,0.2047653943,-0.0717782825,0.1892822385,-0.2513434291,0.1714068651,-0.2912474871,-0.2330379635,0.5002884865,-0.0475544818,0.3155162632,-0.0672079027,-0.8151515722,0.0039756084,-0.1317571998,-0.2981568575,0.0554655679,-0.316452533,-0.0179772135,0.0386409611,0.0199691169,-0.2945590615,-0.499740988,-0.1107104272,-0.2907910645,0.1502086222,-0.3772340119,-0.3173969686,0.2681907713,-0.1453136802,-0.1130178347,-0.1935667247,0.1251445562,0.2337103784,0.0026216542,0.0456763282,0.0397915468,0.0701327696,-0.1509322673,0.2640598118,0.3454025388,-0.2304075956,-0.2532916665,0.2111442983,0.0488136746,0.0093706353,0.1868614405,-0.4016985297,-0.4862574041,0.2285033166,0.1116337329,-0.1191316992,0.0032077669,0.7244369388,0.0137277879,-0.1800157279,-0.0832932591,-0.5568796992,0.2197959572,0.5421218276,0.1716326475,-0.2548394501,0.2390735894,0.2521260679,0.3514921665,0.2042569965,0.0168433171,0.0988065749,0.1415642202,0.0815884545,-0.1022141725,-0.1010157913,0.0182872042,0.1217142493,0.1206022054,0.4657464921,0.1153677851,-0.3298612237,-0.2474089563,-0.378349781,0.0995951146,0.1374293268,-0.2270612419,0.0881247893,0.6110296249,-0.1955145448,-0.2185589671,-0.3416725695,0.038672708,0.1873231828,0.3904876113,0.2172959298,-0.0788985267,0.1394876838,0.0548638217,-0.2743108571,0.2959477305,0.1899339706,-0.0094980104,-0.075405255,-0.0838751346,0.1460564435,-0.0083655771,0.0666340962,-0.1863406003,-0.1030577272,-0.205748722,-0.1520197392,0.3340769708,-0.1267046481,-0.0401890725,-0.2871422172,-0.1001511738,0.0825098604,0.1218746603,-0.0243685301,0.7594544291,-0.3247665465,0.0157974493,-0.1977483481,0.0144843739,-0.2085856795,0.1708633751,0.0600653216,-0.1410720199,-0.0175240319,0.0746595189,-0.2026903182,0.349006623,-0.1012941077,-0.0917105302,-0.026638614,-0.2314457893,-0.0216629058,0.0801123381,0.0068768347,-0.0617347956,0.3190976679,-0.075159356,0.2766931355,-0.3543196321,0.0533007048,-0.2451183349,0.0951386094,0.1917923391,0.0438237004,-0.0042694146,-0.2416734248,-0.1953306049,-0.0423428752,-0.3735588491,0.4086188972,0.0141824502,-0.213357538,0.1352727413,0.1404684335,0.0528428331,-0.2311856151,0.3453972638,0.3026511967,-0.1813116819,0.2694408,0.0037925986,0.7083385587,0.4133132994,-0.1876365691,0.22230497,-0.1856509447,0.3848887086,-0.0569227561,0.0316813663,0.2398763299,-0.1827246249,-0.0820878819,0.2281623781,0.3937471807,-0.2449363768,-0.3181838989,0.1441597641,0.3544702828,0.3368707299,0.1501451135,0.1637333184,0.4371649027,-0.1696397811,0.2969662547,0.084885776,0.0913375542,0.2700667977,0.0051021404,-0.1038464904,-0.0410936475,-0.3795398176,-0.4101157486,-0.1875353903,0.307962805,-0.0593064763,0.0018754317,0.0271858592,0.4652564824,-0.3908314109,0.2804296315,0.1206795797,-0.2110161185,0.1324018687,0.2591264844,-0.0673518032,-0.0917314589,-0.0164624564,-0.1262826324,-0.2515151501,-0.077438809,0.2256623656,0.1479275078,-0.2459356636,-0.2006443888,-0.1607678682,-0.0937055722,0.2307496369,0.4208414555,0.6775869131,0.1369774491,-0.2288987935,0.1790207028,0.1569697559,-0.146931231,-0.2471116781,-0.0049787168,0.001577749,-0.1396837085,0.2970603406,0.3662147224,-0.4236115813,-0.0744130611,-0.1866766959,-0.0803444907,-0.1827374697,-0.2223896384,0.0420035273,-0.5072385669,-0.2543535233,0.004553366,-0.2541323304,0.1867462993,0.814142704,0.0682360381,0.2618999183,-0.171870932,0.0829634219,-0.0917294547,0.0885645598,-0.547601223,0.2633841932,-0.151135534,-0.0485751554,-0.0556695834,0.1355786324,-0.329649508,-0.2360911816,0.2791733444,-0.1781122833,0.360735774,0.2608091235,0.2823237181,0.0981949493,0.1287194788,-0.0984594598,-0.166429624,-0.0856162384,0.0706988499,0.121013701,0.1028853729,0.3624244332,0.0315296389,-0.0774711967,-0.1851493418,0.1262309104,0.2014972568,-0.095324561,-0.0688648522,0.206802696,-0.2036649138,0.014854881,0.0951138958,0.4036404192,0.1902116686,0.0687889978,-0.062499769,0.4326686859,0.0888098702,-0.099083513,-0.1188613176,0.3598309159,-0.3768962324,0.3495980203,-0.0986997932,0.0946758538,-0.2165361345,-0.138494119,0.0316311382,-0.5281537175,-0.0321099274,0.0952986553,-0.0798631907,0.2523521781,-0.2650229037,0.0941003338,-0.0418109335,-0.1963367611,0.094403781,0.2049114704,0.2989526093,0.222501561,0.8804389834,0.0760115907,-0.1486577988,-0.0678777024,0.4888120592,0.1444507986,0.1421856284,0.0888006389,0.0459638126,-0.0170996431,0.2221054137,-0.2006523013,0.1492018998,-0.1591666043,0.4887678325,0.5134711266,0.2004890889,-0.179971382,-0.0581304953,0.2917058468,0.2063032985,0.1150191352,0.2427873909,-0.1956726164,0.104258053,0.105284892,-0.7772347331,-0.0585467033,0.0518833436,-0.1259681135,-0.0933523774,-0.0555330552,-0.5079789758,-0.0459091365,0.3941427469,0.2388593405,0.0854701549,-0.255715549,-0.0631274953,-0.0250692964,-0.1645293534,0.1471521705,0.318933636,-0.0774362609,0.2639330328,-0.5299558043,-0.3374872208,0.2492263913,0.2988563776,0.0416949056,-0.2024859637,0.4742391109,0.073124513,0.0406224802,-0.079665117,0.1087611988,0.076831378,0.0394030102,0.1540914476,0.0704783425,0.0855614543,-0.0750140697,0.2462936193,-0.1708794981,-0.2132565677,-0.2949430346,0.1570913196,0.039786756,0.1136367768,-0.0696730465,0.3271564245,0.173495084,-0.3411516249,-0.2707258165,0.0609895065,-0.2641049027,-0.138711825,0.0763167366,-0.2678321004,-0.204508245,0.3224084973,-0.0084613934,0.0448751412,-0.5479387641,0.1626827866,0.3549610674,0.1732343435,-0.0991472974,0.1168755814,0.2368862927,0.1842701435,-0.235496223,0.2643695772,-0.3658336699,0.3503040373,0.1303751022,-0.204452619,0.202426374,0.5179699063,-0.0970741138,-0.0738625377,-0.0217720661,-0.1242711544,0.1878062785,-0.0108897509,-0.2437902838,-0.0375194959,0.1283105463,0.0585870706,0.4941731691,0.0713350475,0.1180097908,-0.1694412529,-0.2142635137,-0.3806630969,-0.1393252611,-0.0360232592,-0.0473715588,0.0222143754,0.2044505775,-0.0394944698,0.1307848692,-0.0806980431,0.4303512871,-0.0516600758,-0.115675576,-0.1337549537,0.4427872002,-0.3892194629,-0.3260742128,0.0690066069,-0.2389815748,0.1384256333,-0.1861528009,-0.185459137,-0.1023749113,0.1747238636,-0.0876376629,0.1467013657,-0.1892235577,0.3292217851,0.2396405935,-0.0173970759,-0.3523094654,-0.0941024125,0.1928811669,0.342341274,-0.1571198851,0.3694726825,0.0411321335,0.0776690841,-0.123100847,-0.0624792837,-0.2716119587,0.0105895447,0.0931288376,-0.3121074438]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"@mariamabarham Just wanted to note that default behavior `cos_e = load_dataset(\"cos_e\")` now loads `v1.0`. Not sure if this is intentional (but the flag specification does work as intended). ","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":28,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\n@mariamabarham Just wanted to note that default behavior `cos_e = load_dataset(\"cos_e\")` now loads `v1.0`. Not sure if this is intentional (but the flag specification does work as intended). ","embeddings":[-0.2263446897,-0.2859195471,-0.1184170321,-0.4919193089,-0.4530381262,-0.2508601248,0.2163070142,0.2214708775,-0.0356374495,0.3164639473,0.0701321959,0.4571445286,0.0324561484,0.3533188105,-0.1624477506,0.4214379489,0.1171861738,0.2206687033,-0.0301127639,-0.2556170523,-0.2271460146,0.1309096217,-0.168599084,-0.1252735853,0.1629865915,0.2887752652,-0.2544233203,-0.0957462415,-0.041461423,-0.1833728552,0.4028756022,0.3478627801,0.1702445745,0.064628467,-0.0001100439,-0.1363886595,0.4567955732,-0.063955389,-0.2753710747,-0.1536118835,-0.355692625,-0.3650524616,0.0707520247,-0.0715934336,0.2172768563,0.162727356,0.1494596601,-0.0311017707,0.0402089953,0.166272521,0.2691470683,0.2860792875,-0.0316775627,-0.3328509927,0.2125618309,0.4736269712,-0.3600447178,-0.0748891607,0.0436958186,-0.0970935151,-0.1418292075,0.3820013404,0.2254563123,-0.1720419973,-0.2601476908,-0.0306741036,0.2146512419,0.1305552572,0.2605664432,0.0569030903,0.3529518545,0.138649404,-0.2649240196,0.2500213087,-0.0588390715,-0.4292012155,0.2250390798,-0.1623161435,0.021105051,0.00474513,0.3541200459,-0.1866617054,-0.5334498882,0.1561824828,-0.2166710198,0.1714299172,-0.2124951929,-0.1408740431,0.0942203104,0.1076117083,-0.0929835513,0.11195191,-0.2070815712,0.1804278344,-0.3474440873,-0.1026492268,0.3085197508,-0.063929081,0.4267698824,0.1589215398,0.1950719804,0.1905683875,-0.0156986732,0.1507025361,0.0115639623,0.3686622083,0.5664452314,0.0640151426,0.403051585,0.1425480098,0.4047579169,0.0165483244,0.0781297907,0.021182036,-0.0022484772,-0.1615668237,0.110374704,-0.5184797049,-0.1801729351,-0.0319606699,-0.1015330628,0.0927184895,0.0045885984,0.2791186571,0.288818419,0.1953786314,0.3214773834,-0.0221195426,0.0167544503,-0.2821889222,-0.2884854078,-0.081266515,-0.2485474199,0.2028684467,0.1990646422,-0.3987296522,0.2090887129,-0.3644556999,0.4094014168,-0.064132601,-0.2289320827,0.3046186566,-0.035715349,0.263448,0.0388117805,-0.0173004922,-0.0994563922,0.3986671865,-0.1659814268,0.1135759428,-0.217561841,-0.488876164,-0.2699548006,0.2650915682,-0.1804881543,-0.252689451,-0.2429434955,0.4173628092,-0.2470451295,-0.0076080086,0.0136847692,0.1410547048,-0.3214108944,-0.2032902241,-0.0824026614,0.1973445565,-0.043600142,-0.0886062086,-0.6579604149,-0.0992988124,0.0154747954,-0.1841049194,-0.0183929373,-0.308886081,-0.242937088,-0.0975850374,0.409198463,-0.5409089923,-0.0888950005,0.3007032871,-0.0528970063,-0.210860312,0.1997171491,-0.3001001775,-0.1599310935,-0.2848050296,-0.3112842739,0.0520441234,-0.0344547667,-0.0510548279,-0.0711145699,-0.3464411497,-0.3489783108,0.2585716248,0.3637625575,0.0329327658,0.3567892909,-0.4188337922,0.1504984498,-0.055605378,0.1332037747,-0.0255278852,0.6704112291,0.1037110835,0.1955670863,0.138107121,-0.3721389472,0.0969761461,0.182708323,0.1522473842,0.2096587718,-0.3624671996,-0.1504474878,0.1442919225,-0.0062052649,-0.1901966929,0.1890366822,0.1674598008,-0.0667391941,-0.0394476093,-0.0403212085,0.4120458663,-0.3333109021,-0.0893501341,0.1691672206,0.0670861527,-0.059089683,-0.0072165024,-0.0522315614,0.0911005214,-0.0032769458,0.0087143965,-0.1518038064,0.3978513181,-0.0146999247,0.1373352259,0.0756934658,0.2027658373,0.129919216,-0.1129686013,0.0156410038,0.03118724,-0.0605425537,0.1088652089,-0.2137699425,0.3678352237,-0.044001434,0.0786850303,-0.0030529334,-0.0895441696,0.0268371515,-0.0728679299,-0.5226480961,-0.2180042267,0.2421562076,0.106957756,-0.4233147204,-0.2397636473,-0.2204621434,-0.0805059969,0.1762452871,-0.1001153961,0.1500443816,0.3162189424,0.1081685126,-0.1911519617,0.2767570317,-0.1665317118,0.4063424468,0.3151409924,-0.1061262265,-0.0480204746,-0.3054266274,-0.0532287769,0.136480242,0.2142147422,0.1236786842,0.1257357746,0.4169747233,-0.1495238841,-0.5176540613,-0.4135940373,-0.0627236813,-0.1665929705,-0.1966625452,-0.1665610671,-0.0639953241,0.0942216516,-0.1074077711,0.0386362262,0.0833421871,-0.2362005562,0.2636577785,0.2246634364,-0.0613555647,0.2357615829,0.0254008397,0.6541308761,-0.0225258637,0.1364822388,-0.1587279141,0.0665988848,-0.4203677773,0.1941088885,-0.5060146451,-0.226005882,0.4339205921,-0.0075914143,0.3097059131,-0.0600994974,-0.6933544278,0.0736582577,-0.1926906854,-0.1378563941,0.0665964484,-0.2883520424,-0.1115364283,0.0407223403,0.0874927044,-0.3944211006,-0.3563586771,-0.0815198347,-0.2302690446,0.2790993452,-0.34588781,-0.4851728678,0.2348673791,-0.3389965296,-0.1109105125,-0.0839217007,0.1287858933,0.5446904302,0.1381399632,0.0874253511,0.1251380295,-0.014643562,-0.4106577933,0.2389359474,0.3401024342,-0.3349628448,-0.4000643492,0.1089507937,0.093590498,0.1544493735,0.2097405195,-0.1866329163,-0.5588251352,0.180184111,0.1557171494,-0.2045299858,0.1585502774,0.3840799332,-0.0192675143,-0.1956894249,-0.186851114,-0.3143155575,0.2257439196,0.3081616461,0.1298700124,-0.1642553657,0.2678792775,0.1611649394,0.1489787251,0.0820311978,0.0191196091,0.050539054,0.1108465493,0.3510953784,-0.0256932583,0.0552800484,0.1333315372,0.081922248,0.2069119513,0.4673638642,0.1389230341,-0.2309901714,-0.2874268293,-0.2523161471,0.1630087644,0.0316604376,-0.1666241139,0.1882855892,0.4335594475,-0.12905927,-0.0965869576,-0.3269830644,0.0328602158,0.0610745139,0.4213992655,0.0968269929,-0.1743542552,-0.0047745951,0.1403530538,-0.3281843066,0.359834969,0.1082926765,0.0513450354,-0.0035432954,-0.2092659622,0.2115894109,-0.029143583,0.4310415983,-0.1808991283,0.0137020852,-0.1111804396,-0.0778980777,-0.0930261016,-0.1693755984,-0.1178676635,-0.306042701,-0.0353091247,-0.0355612747,-0.0528145544,-0.0292600524,0.5762625933,-0.2663776577,0.0744910687,-0.2874293029,-0.2081997246,-0.4295221865,-0.100883618,0.0045454372,-0.0596228354,-0.0953707322,-0.2208274752,-0.1356152594,0.2697879076,-0.227451086,-0.1419631243,0.1268463284,-0.2542560995,-0.0007777258,-0.0330542438,0.1000150442,-0.1161509082,0.3154968917,0.016632162,0.2869828343,-0.3732557893,0.10811764,-0.1803288013,0.217107147,0.2719832957,0.0013414951,0.1454332024,-0.4411624372,-0.2380361855,-0.0139470939,-0.0844509304,0.2103102803,-0.0049422248,-0.2048134208,-0.0462406501,0.3997000158,0.0542028621,-0.2666934431,0.2724325657,0.1473266482,-0.2883755863,0.4483861923,0.0112594785,0.6613929272,0.2898104787,-0.2178915739,0.2965306044,-0.2164826393,0.5671082139,-0.1858257502,0.1396551728,0.0983451381,-0.1416409761,0.0241187159,0.2629132867,0.3644725084,-0.0419859998,-0.3291960061,0.2483590394,0.216752246,0.3008893728,0.2127334177,0.1163180247,0.0853663832,-0.3070466518,0.0013401267,0.1875015497,0.084092766,0.2702900469,-0.0411760025,0.0987709016,0.0285097472,-0.3716987967,-0.2279875726,0.0316319354,0.4609621763,0.1189183667,0.0127081107,0.0654435679,0.2390542328,-0.3236112297,0.2711402476,0.1031316072,-0.0883093402,0.2994325757,0.3193251193,0.0104971947,-0.1767601967,-0.1510275155,0.0243218224,-0.1588129997,-0.2019996196,0.2272274047,0.1741968542,-0.2386624813,-0.2017285675,0.0339090377,-0.0544710085,0.1423086822,0.533423245,0.4345429242,0.1062091887,-0.1701266021,0.2046646029,0.1269295216,-0.1147509217,0.0031191912,0.1529110819,-0.1906908602,-0.1832517982,0.5322161317,0.2791705132,-0.132044524,-0.0629270375,-0.1157118604,0.0162951294,-0.2150592506,-0.1890401691,0.1409250647,-0.3937105238,-0.127748549,0.1681396067,-0.2827530801,0.2493843287,0.8095263243,0.0720107853,0.2018945366,-0.2282631844,-0.1281666607,-0.025408417,-0.0795553178,-0.4771957695,0.1812091023,-0.0205051433,-0.0507149361,-0.087898761,-0.0128123062,-0.3956343532,-0.2420557439,0.1952884942,-0.0969548076,0.4021477699,0.2122156918,0.4604626,0.1858866215,0.2123562545,0.0151058715,-0.2088433802,-0.1787459105,-0.0136566488,0.1072642431,0.0870388076,0.2987443507,-0.1990333647,-0.1829626709,-0.2923521698,0.0222863201,0.335093081,0.1099198908,-0.1967455745,0.088270843,-0.4510148168,0.0102492506,0.3422644734,0.1872564703,0.1763869077,0.0917051062,-0.066082567,0.3762292266,-0.1261286438,-0.0265957899,0.1228471696,0.3449653089,-0.259908855,0.1292830557,0.0085767331,0.0821661502,-0.1988819838,-0.1735463142,0.0164010487,-0.3018582761,0.0068968311,-0.0566382632,-0.1672634631,0.3034840524,-0.2394814789,0.0696647763,-0.276270628,-0.1659895331,0.207940504,0.2772287428,0.2561792433,-0.0934531763,0.7590200305,0.106886737,0.1414371729,-0.0747290328,0.2207971811,0.1693820655,0.0426334031,0.1686858386,0.2758137286,-0.0878944546,0.3604963422,-0.0775493979,0.3008495867,0.0368034989,0.2739905119,0.5524520874,0.055059772,-0.3132965565,0.1077868268,0.4623661637,0.0511642061,0.0584337339,0.1863642931,-0.2256689817,-0.0209183302,0.1821378767,-0.763312459,-0.176771313,0.0875550136,-0.2771156132,-0.1974260658,-0.1056888402,-0.4732800722,-0.0067754006,0.2994881272,0.0507814847,-0.0161205605,-0.2081557512,0.0081938365,-0.0603774302,-0.2778089046,0.0737763494,0.2569638789,0.0064146561,0.2032146305,-0.4178889394,-0.1361067444,0.377004534,0.1393404007,0.1372516155,-0.0884928554,0.5053774118,0.1196648106,0.102590695,-0.0916008353,-0.0092941187,0.1886350662,0.123281315,0.1045760363,0.1803709865,0.1920618415,-0.186566487,0.1564943194,-0.035235133,-0.2429848313,-0.2443137169,0.1843425781,0.1063569039,0.1503627896,-0.0876581147,0.2204680592,0.0323285945,-0.3462284803,-0.0621121116,0.0205743536,-0.1276403815,-0.1460464448,0.1198637933,-0.0651046857,-0.1248826161,0.169946894,-0.1149884239,0.0625870898,-0.3730799854,-0.0043127052,0.365942508,0.187730819,0.0706614181,0.0547842979,0.1421366483,0.1427471787,-0.2538005114,0.122228153,-0.4326272011,0.5907484293,-0.087734513,-0.182802245,0.1407769173,0.4150139689,-0.112276569,-0.1277844906,-0.2960697711,-0.2116582543,0.0673025995,0.1088856086,-0.3329249322,-0.1272960454,0.1839385182,0.0378894098,0.4441289306,0.0480902381,0.213060841,-0.1285390705,-0.0711585358,-0.5014448166,-0.1346664578,-0.0239164252,0.0925919786,-0.1338444948,-0.0298392922,0.0431614891,0.119513154,0.0220345519,0.3425879478,0.0461307168,-0.0907569528,-0.095056802,0.4244766235,-0.5147942901,-0.3080168664,-0.1060278639,-0.1027720198,0.0668235049,-0.1294811517,-0.0408959799,-0.2425192446,0.3495898545,-0.2260924727,-0.0850382373,-0.0861115232,0.434343636,0.2038701773,0.0867725089,-0.2204902172,0.0985116661,0.2275275886,0.1977185607,-0.2760639191,0.150846377,-0.0448230989,0.0177721623,-0.2207440883,0.0235827975,-0.1709706783,0.0150426785,0.1533436626,-0.3238045871]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/163","title":"[Feature request] Add cos-e v1.0","comments":"> @mariamabarham Just wanted to note that default behavior `cos_e = load_dataset(\"cos_e\")` now loads `v1.0`. Not sure if this is intentional (but the flag specification does work as intended).\r\n\r\nIn the new version of `nlp`, if you try `cos_e = load_dataset(\"cos_e\")` it throws this error:\r\n```\r\nValueError: Config name is missing.\r\nPlease pick one among the available configs: ['v1.0', 'v1.11']\r\nExample of usage:\r\n\t`load_dataset('cos_e', 'v1.0')`\r\n```\r\nFor datasets with at least two configurations, we now force the user to pick one (no default)","body":"I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).","comment_length":83,"text":"[Feature request] Add cos-e v1.0\nI noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https:\/\/www.aclweb.org\/anthology\/P19-1487\/), and v1.11 has noted [annotation](https:\/\/github.com\/salesforce\/cos-e\/issues\/2) [issues](https:\/\/arxiv.org\/pdf\/2004.14546.pdf).\n> @mariamabarham Just wanted to note that default behavior `cos_e = load_dataset(\"cos_e\")` now loads `v1.0`. Not sure if this is intentional (but the flag specification does work as intended).\r\n\r\nIn the new version of `nlp`, if you try `cos_e = load_dataset(\"cos_e\")` it throws this error:\r\n```\r\nValueError: Config name is missing.\r\nPlease pick one among the available configs: ['v1.0', 'v1.11']\r\nExample of usage:\r\n\t`load_dataset('cos_e', 'v1.0')`\r\n```\r\nFor datasets with at least two configurations, we now force the user to pick one (no default)","embeddings":[-0.1603744179,-0.1905596256,-0.0409298912,-0.5251531005,-0.5301490426,-0.2091349959,0.0791700259,0.3027105331,-0.084867388,0.2993151248,0.0913367122,0.5883413553,-0.127389133,0.3007558584,0.1448273063,0.1922949404,0.0493811816,0.3573561311,0.0160913859,-0.1463334262,-0.2847516239,0.2668954134,-0.1590034366,0.022066094,0.1762136072,0.1823166907,-0.321796298,-0.02632479,-0.0610385649,-0.265681982,0.433855027,0.296884656,0.2419633865,-0.1544307172,-0.0001130699,-0.2487540096,0.5088639259,-0.0622031353,-0.4297463298,-0.3075879216,-0.3339643478,-0.5184953809,0.1690254658,-0.0987784192,0.1858433932,0.1506469548,0.3687634766,-0.0325421356,-0.0012714587,0.1081603914,0.2273946702,0.2823914587,-0.0840936154,-0.2982402146,0.1130262315,0.3750306368,-0.3086328804,-0.1634405702,-0.0422804318,-0.2713531256,-0.0524476096,0.3673297465,0.2016812265,-0.2252693772,-0.1958558112,-0.0478476547,0.1452473402,0.0442856513,0.0785973147,0.094222784,0.3499735892,0.0466120131,-0.2632308006,0.0423653424,0.0768405125,-0.5963779688,0.2815056145,-0.0695537999,-0.1712331176,-0.0754080042,0.2659721375,-0.2741574049,-0.4252688289,0.3804299533,-0.0608034842,0.2607444823,-0.0665872693,-0.0678771287,0.2350341678,0.0808953643,-0.1046843082,0.0276291668,-0.1441478878,0.2545187473,-0.4161166251,-0.1421226412,0.4012795389,0.0414198898,0.2769635022,0.1498364657,0.2790861428,0.0128509672,-0.0514417291,0.1853981018,0.0855170265,0.3146654069,0.6022409201,0.0287071243,0.2610608339,0.0919452533,0.5128621459,0.1002445891,0.0753753558,0.0205947272,-0.1638908088,-0.1278833449,0.1113580614,-0.38129282,-0.122656323,-0.0945472941,-0.0938417986,0.0630017295,0.0248092152,0.1908301115,0.2591687739,0.2565684021,0.3442850113,-0.0727608204,-0.1324234158,-0.3846977055,-0.2491017729,-0.0533332527,-0.3931257427,0.2702561021,0.3587175608,-0.23135297,0.2805841267,-0.4026278555,0.1932352036,0.0167675111,-0.2188313454,0.3100357056,-0.1656469554,0.0798900947,0.0468821861,0.0137233566,-0.3002426326,0.2614299953,-0.2551450133,0.095273219,-0.2514296174,-0.5217930675,-0.2898740172,0.2131017447,-0.2102160007,-0.3397215307,-0.1648019254,0.4888764322,-0.2243465483,-0.2618386745,-0.0197754744,0.0740577132,-0.3711227477,-0.1326876432,-0.0844178274,0.1536420733,0.0322751924,-0.2395110726,-0.6589326262,0.0237019323,0.0504046939,-0.1060368195,-0.1056155264,-0.2657812536,-0.1496505737,-0.059896972,0.7229759693,-0.4373762608,-0.1154996455,0.2689371109,-0.0676072016,-0.2052620798,0.1048451215,-0.1371217072,-0.2300135046,-0.1585985273,-0.1731006354,0.2563079596,-0.0513107441,-0.1048711464,-0.0849446282,-0.3723749518,-0.0840224028,0.2475238144,0.2598854303,0.0102373781,0.2654013932,-0.1716258377,0.2031891048,-0.0694108307,0.148471877,0.0382137522,0.4409762621,0.0429674685,0.0998889282,0.1025070101,-0.5050923824,0.0785629451,-0.0480642281,0.2747194171,0.3230306506,-0.2149771154,-0.0498168804,-0.0513810441,0.0627841502,-0.2429581732,0.1295619309,0.291669488,-0.0453818031,-0.027404394,-0.0881026238,0.4771677554,-0.2146998197,-0.1701603234,-0.225007683,0.0794707313,-0.1317952126,-0.038429413,0.0011605707,0.3066433072,-0.0057504578,0.031460952,-0.085896194,0.3410401046,-0.1707220674,0.2327118963,0.0020118935,0.0517823882,0.1553287953,-0.1710367948,0.0714992061,-0.0387482047,-0.0591474883,0.1741671711,-0.235874936,0.408785224,0.0555513836,-0.0049423906,0.0156648513,-0.025185002,0.0245814025,-0.2130450159,-0.5066873431,-0.1608449966,0.3907119632,0.0903762951,-0.41927737,-0.2182733417,-0.4219491482,-0.0539113581,0.2659189105,0.0109217949,0.0686053559,0.3461208642,0.1989822388,-0.2583040595,0.287675947,-0.0337567888,0.5213515162,0.3559094965,-0.1437416971,0.0553794093,-0.497071892,-0.0343345255,0.1937440932,0.2227172703,0.1649831682,0.0345131382,0.4799375534,0.0247187279,-0.4151760042,-0.4975746572,0.0300820656,-0.066778481,-0.170993045,-0.0767860264,-0.2356609255,-0.2485540956,-0.1343775839,0.1527370065,-0.0066434303,-0.2822704613,0.2530551255,0.2228182256,-0.0968491957,0.2704641521,-0.0881978497,0.7165158987,-0.1732389033,0.1325025558,-0.2111668289,-0.0197338276,-0.4386223555,0.1435934305,-0.3673205376,-0.1281872392,0.4306366146,0.0337064676,0.2416912466,0.1017245874,-0.6566215158,0.0737065673,-0.1674202383,-0.0849752128,0.0753148273,-0.3813074827,-0.1796533912,-0.0971656218,0.0917324945,-0.3468784392,-0.3567785919,-0.0608406849,-0.1778292805,0.1936436445,-0.386149019,-0.5043295026,-0.0028922216,-0.3854027689,-0.0603320077,-0.0328723416,0.2645630836,0.7287469506,-0.0292721316,0.0609108508,-0.0188146252,0.0539790057,-0.3222586513,0.4111197591,0.2417619824,-0.3663308024,-0.3256415725,-0.0940812826,-0.0090113189,0.275954932,0.137724191,-0.200719595,-0.3760524988,0.1728867441,0.0956325606,-0.1895506084,0.1695750803,0.4947325289,0.0356851928,-0.1434897184,-0.1199136004,-0.3978121579,0.2254878879,0.5513763428,0.3134689331,-0.0997380987,0.2379786521,0.1289057434,0.0546539351,0.1860589534,-0.031547334,0.2133136988,0.0751072168,0.3027555943,-0.0681923628,-0.0488389358,0.186204493,0.1191441193,0.3289926648,0.4975135624,0.1018099263,-0.4926541448,-0.3609602153,-0.2420579046,0.1246070266,0.0034182796,-0.0689061135,0.0607320331,0.4087852538,0.026039958,-0.2615892887,-0.4135234952,-0.1763294786,-0.1090498343,0.4381890297,0.0329929218,0.0128651941,-0.0826628059,0.0264654383,-0.4243906736,0.2321767807,0.0650167912,0.2848562002,-0.0542014651,-0.1326745898,0.1120835245,0.0853098333,0.2600694895,-0.2370330393,-0.1213978454,-0.0110551957,0.1241703928,-0.1962542683,-0.0632331893,-0.1607283503,-0.1735666096,0.0729999766,0.0165275447,0.0124157723,-0.184082076,0.5894098282,-0.1919881701,0.0277282801,-0.33776775,-0.1544549614,-0.2920678556,-0.0954473317,-0.0887523666,-0.1869193614,0.142253533,-0.2019129694,-0.0053323368,0.2286872268,-0.094863221,-0.0054370011,0.2288451642,-0.2483285218,0.1345135123,-0.0950883627,0.017488502,0.0978950188,0.0397709273,-0.0459276624,0.2397381663,-0.0904295072,0.183639273,0.0417141169,0.3428072333,0.2911259532,0.0761276633,0.1502442807,-0.3855303228,-0.2505571842,-0.0424469709,-0.080795832,0.2425474226,-0.0665718466,-0.3306765854,-0.0909568816,0.4912471175,0.066090785,-0.3306332827,0.3240830004,0.0861422867,-0.3378441036,0.4092399776,-0.1289198548,0.7491976619,0.3950465322,0.037740577,0.3253991008,-0.0398076698,0.8740342259,-0.0319583379,0.2768107057,0.0714416802,0.05260868,0.0337325186,0.2498622537,0.3760496378,-0.0150479833,-0.3114896119,0.3749997318,0.5791009665,0.0683800578,0.19329153,0.3513818681,0.1492741108,-0.4552726746,-0.1680758446,0.1207706854,-0.1080627739,0.3852906227,-0.01413175,-0.0065785898,0.0169402435,-0.3202203214,-0.210574612,0.0402919427,0.3929307163,0.0589028634,0.0736830831,0.0282569807,0.4237447083,-0.2301636487,0.182666868,0.160945043,-0.2040281445,0.1644722968,0.1395416707,-0.0185785703,-0.0731378421,-0.1439753026,0.0839561149,-0.1762894094,-0.2379287928,0.10621465,0.1955348849,-0.2047399133,-0.1186834946,-0.0075541562,-0.0384330601,0.2452460229,0.2921880186,0.5004431009,0.1078107506,-0.2376678139,0.1723292619,0.0980486646,-0.1080287248,0.0239805859,0.2002630681,-0.1512065232,-0.1128026396,0.5301057696,0.2133043259,-0.1893975735,0.1030639559,-0.1451846659,0.0732821673,-0.195994705,-0.1548787206,-0.0329858176,-0.5144613981,-0.1993074119,0.1846110523,-0.2377413213,0.3092066646,0.7213858366,0.0165548548,0.2494661361,-0.1088580564,-0.1884734929,-0.0534621663,0.0277340394,-0.5827199817,0.1609862447,0.1386865526,0.0597309731,-0.0396648198,0.0786677375,-0.364228487,-0.2929227948,0.1222373694,0.0475867279,0.528796196,0.1722330749,0.2284763604,0.0847081318,0.1357515305,-0.1003014892,-0.188078776,-0.1580127627,-0.1364351958,0.1190504357,0.1717754006,0.1710317731,-0.1796264797,-0.1982627511,-0.2071645409,-0.0241188258,0.1777467877,-0.070918411,-0.1936513186,0.3347195983,-0.4681262672,-0.0451326631,0.3891866207,0.3572050929,0.1698103547,0.0491305776,-0.0288290046,0.5353575349,-0.0368523784,-0.0511418656,-0.0861509517,0.243158564,-0.3270846009,0.2475724071,-0.0905899704,-0.0074878465,-0.2218110114,-0.2281967998,-0.030339662,-0.3078689873,0.0357371122,-0.044831153,0.063953042,0.2379177362,-0.314887464,0.0552234389,-0.0359492376,-0.0291319452,0.1841197759,0.2187729329,0.3217138946,-0.1154591441,0.6415475011,-0.0441875681,0.0149112055,-0.1816295981,0.2836932838,0.1905100644,0.1689391136,0.0888021886,0.1863639206,0.0084648598,0.3330289721,-0.0624589212,0.2202375829,-0.0485847332,0.3439473212,0.5547142029,0.0711765066,-0.3583300114,-0.0543022193,0.3893846869,0.1457907706,0.1239126027,0.1341465414,0.0070626298,-0.0710020438,0.1813876033,-0.6420034766,0.0533803329,0.0319610685,-0.3519986272,-0.3825597763,-0.0926098824,-0.510448575,-0.0621520579,0.1417339444,-0.0074940184,0.0329422466,-0.0786883458,0.0239071026,-0.2059449553,-0.0545887575,0.2270655036,0.1682695746,-0.0234084707,0.3689232469,-0.5310975909,-0.2112864405,0.1906959265,0.2477077544,0.2092950046,-0.0390018225,0.4058574438,0.1120250449,0.1455616951,-0.0707634687,-0.0598200969,0.161033228,0.3544895649,0.0419980362,0.2613224983,0.0980872735,-0.0997924507,0.1400725096,0.0434310734,-0.2990834415,-0.2118794322,0.3249712884,0.1753543317,0.1804134697,-0.0232999586,0.3757541478,-0.130172357,-0.2383160144,0.0811240897,0.1026385128,-0.1089486107,-0.2417626679,0.099444136,-0.0711653233,0.0070667439,0.1841976643,-0.0305572804,0.0211345758,-0.4140671194,-0.1286515594,0.2505950928,0.1137351692,-0.1267634034,0.0977417454,0.1597737223,0.0466861203,-0.2836061418,-0.1293572932,-0.0501629934,0.5626105666,-0.0194489006,-0.2112591118,0.1510004252,0.3584873378,-0.026390994,-0.1281693876,-0.1886705011,-0.2875412405,0.0438987501,-0.0081591923,-0.3241662085,0.0115262112,0.0436978713,0.2331978232,0.3404268324,0.1010014787,0.277570039,-0.1553439647,-0.0850162283,-0.5489779711,-0.1848103553,-0.0563795231,0.0664725825,-0.0956385285,0.05234465,-0.0347967558,0.3438424766,-0.0101212952,0.3170662224,-0.0208143368,-0.0981979743,-0.0326947607,0.4011127055,-0.4748795331,-0.1552657038,-0.0215120167,-0.0174943432,0.0346452929,-0.060585618,-0.0785984322,-0.2389976531,0.3525123596,-0.1608174294,-0.0447668619,-0.0203035772,0.424336195,0.2557458282,0.0769362748,-0.2594652176,-0.0163807999,0.2102701068,0.1324423403,-0.1359895766,0.2562758029,-0.1520243436,-0.025490256,-0.2026946396,-0.1097297147,-0.1159679368,-0.1583410501,0.1154208556,-0.3498463333]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"usually you can replace `download` in your dataset script with `download_and_prepare()` - could you share the code for your dataset here? :-) ","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":22,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nusually you can replace `download` in your dataset script with `download_and_prepare()` - could you share the code for your dataset here? :-) ","embeddings":[-0.298441112,0.2715887129,-0.0254015345,-0.1736125499,0.1113381907,-0.0324908607,0.3496927917,0.30469504,0.096391201,0.1327766478,0.150235936,0.1813911349,-0.2963347435,0.0544609018,0.1696768552,0.0301233977,-0.0018450774,0.2378813475,-0.1085268855,0.2644943595,-0.1570069939,0.0259409193,-0.083469741,0.0347202979,-0.0079233758,0.1446914077,-0.1539782584,0.0155087179,-0.3736417294,-0.9120110869,0.4226641059,0.0865378603,0.185269773,0.40165025,-0.0001230559,-0.0780507028,0.5848661065,-0.0812642276,-0.661423862,-0.2475264817,0.1799682081,-0.4401650727,0.3309346437,-0.030891018,0.0564301126,-0.1428359747,0.2256732881,0.0363934413,0.0625527129,0.3939165771,0.1415757239,0.2784300447,0.1039661169,-0.2289483398,-0.1139817685,0.0661353096,-0.1395557374,0.5098290443,0.3375481367,0.0808631182,-0.0261453837,-0.1172660813,0.0758388788,0.4826373458,0.1798252314,-0.0201055538,0.4507332742,-0.0264175273,-0.1909984201,0.0009862897,0.8557766676,-0.5572375059,-0.285890758,0.020644838,0.0972792208,-0.1694420427,0.2176303118,-0.2057976872,-0.3246448934,0.0170045309,-0.194869265,-0.3370635509,-0.0902516693,0.3106402159,-0.2195286453,0.1833121777,0.0680324584,0.212271452,0.0559982434,-0.09097334,0.2383900583,-0.0226798709,-0.0061221295,0.2574823499,-0.0842557028,-0.2459642887,-0.3158248663,-0.2029247284,0.1599836349,0.0582976267,-0.4827130139,-0.1835442036,0.0428904146,0.2146052569,0.1119600013,0.3177333474,0.3439334631,0.2845926583,0.1950680166,0.1744562089,0.2424224317,0.1537244022,-0.0711153746,-0.3969819844,0.106654942,0.0880114958,0.2400938421,-0.2806532681,-0.1707747579,-0.2619642913,-0.3381373882,0.0097823311,-0.077322863,0.2922813296,-0.0901487619,-0.233863011,0.049078092,0.4467692375,-0.1982865781,-0.1562927663,-0.0272893012,0.1053005457,-0.2146361023,0.0499136969,0.3252117336,0.014613213,0.404274106,-0.2409366816,-0.1593504548,-0.0467430316,0.1086579189,0.0592325144,-0.2635221481,0.3010063171,-0.072180815,0.0834098533,0.0486629382,-0.276293546,-0.2149082869,0.2309889048,0.0679151267,-0.4953610003,-0.0233423337,0.1133399755,-0.509395659,-0.0598592609,-0.0691320673,-0.1643345505,0.1056805477,-0.3929363191,-0.0308141876,-0.4348366261,-0.3722633421,-0.1116913408,-0.0335124917,0.296977818,-0.3369508982,-0.2700842023,-0.4456784427,-0.3823715448,0.2131670266,-0.2579701841,-0.0726997331,0.2081034333,-0.4191890657,-0.029213706,0.5602772832,-0.5683712363,-0.3613014221,0.4470447898,-0.4189874828,-0.2590119839,0.1865042597,-0.0008599867,0.066505529,-0.3295421898,-0.143052578,0.0646112785,0.0423734151,-0.1362273991,-0.1102428958,-0.2676572502,0.1991444677,0.2632280886,-0.0967826173,0.1312554926,0.1476364881,0.1027576998,0.3266959488,0.1425901502,0.1718100756,-0.0883590654,0.2301718295,-0.2256702781,-0.1406852901,-0.1093285978,-0.60998106,0.2383183986,-0.0662352294,0.2077953219,-0.141708836,0.0233580619,-0.6258832812,0.0158698317,-0.0591708645,-0.2286385894,0.0068517132,0.1909549236,0.1120264381,0.0182145387,0.0371074192,0.0626556799,-0.2197260559,0.1759645194,0.0102650728,-0.0012479722,-0.103381142,0.0382828675,0.2361369282,0.2201990187,0.086502634,-0.310854733,-0.0424862429,0.2862296104,0.2748507261,0.2039648443,0.1929637194,-0.0746738091,0.1389698535,0.1538354158,0.174069047,0.2167987525,0.0869945884,-0.2102238238,0.0805541277,0.1106917784,0.1924929768,-0.0216934215,-0.1749161184,0.0260542631,0.4049425721,0.1291377246,-0.0262372196,-0.3822834492,0.1004609317,0.2953742445,0.6366958022,-0.0471049361,0.3663630188,-0.1084176898,-0.1969179064,-0.3486185372,0.0082016429,0.1602071077,-0.1580492407,-0.0922559649,-0.0057400051,0.2688911855,0.8805611134,-0.0264900457,0.1716458499,0.1453922242,-0.1114395261,-0.2071067542,0.1883522272,0.0770070702,0.0482561029,0.1420410424,-0.014179443,-0.3831110597,-0.0887720957,-0.0530825257,0.4589431584,0.4843926132,-0.3677526414,-0.1962990165,-0.2473428398,-0.3671188354,-0.4130909145,-0.3259518147,-0.1526172459,-0.3070310354,-0.1207636446,0.2766251862,0.1386705935,0.2270921618,-0.2515863478,0.1858075559,-0.2386373878,-0.3114603162,0.0959394649,-0.1713583767,-0.3402530551,0.030635016,0.2242275923,0.0313207619,0.4151515365,-0.1468674242,-0.2182322741,-0.2512980402,-0.2445472479,0.1256610155,0.0889754444,0.5211691856,0.4599178433,-0.0276097395,0.4571679235,0.1171948463,0.0346041061,-0.2661328316,-0.4669435024,-0.1576691717,-0.0198037047,0.0831541792,-0.0382985845,-0.4752772748,-0.5595113635,-0.077983655,0.0170762204,0.1075389534,0.0264048297,0.044131618,-0.0996685699,0.2051924467,0.0257165655,-0.1166401729,-0.1441698968,0.0730155408,0.0200615227,-0.1757684648,-0.1796307266,0.1131969169,-0.0628765449,0.0819620714,-0.1548333019,-0.4178172052,-0.3486689031,0.074657999,0.1838562489,0.2805372477,-0.2723371983,0.2895947993,0.2498318702,0.1481546909,-0.1244997606,-0.392529279,0.1562229693,0.0194375068,0.2505045235,0.1468767226,0.4825636148,-0.1921384335,0.7506562471,0.1180375516,-0.1631099284,0.0205957256,-0.2556129098,0.0414721295,-0.0539504029,-0.1241387352,0.3178112805,-0.1397545934,-0.2466551512,0.1957453936,-0.2231131941,-0.0524620004,-0.0955221653,0.2213319987,-0.0654381514,-0.025436474,-0.0534771457,-0.2093604207,0.0576123409,0.0378894471,-0.0709614009,-0.1944077313,-0.214019984,0.0048630615,0.3135365844,0.0049294685,0.1602716893,-0.3005148768,0.0185747035,-0.4462014139,0.3039755523,0.27127406,0.5984053016,0.082950145,0.0767505318,0.1285491735,0.1027716696,0.2140497863,-0.028821256,0.1224334612,0.2516891658,-0.1339614391,-0.0891562477,-0.2406843603,-0.065881975,0.2125184238,-0.029731784,0.0958938748,-0.3500879705,-0.0892627239,0.2652928829,0.1409552246,-0.0859002993,-0.3551106453,-0.1233074218,-0.0726566762,-0.1503216326,-0.2127329409,-0.1758047193,0.1004404277,-0.1292781234,0.2820553482,0.0149539718,-0.0901893526,-0.0866159722,0.1156768799,0.140010193,-0.1844521463,0.3376259506,0.177307874,0.0468804277,0.4023389816,0.3133348823,-0.1016156822,-0.2746430039,0.0268891063,-0.0069021033,0.3866170049,0.1277942657,0.1136167198,0.0418583527,-0.2697535455,-0.1489514112,-0.2984609604,0.1984382719,0.3825050592,-0.0943648219,-0.051381804,-0.3670403957,0.6012390256,0.2057897598,-0.1570284218,0.4772731364,-0.251943022,-0.2803322673,-0.0551795214,0.0721907988,0.8358832002,0.1205168441,0.1654244512,0.3775365055,-0.2624785304,0.3626869023,0.0903796107,0.1614859849,-0.5544980764,-0.0848002434,0.0247876067,-0.2305979729,0.2437988073,-0.1052323282,-0.2362073064,0.3708560765,0.0479405336,0.2831426859,0.2545764744,0.3559742272,-0.1117859259,-0.0445249155,-0.1092494577,0.1016581953,-0.1658704877,0.4050753713,-0.2347889245,0.0895391926,-0.0864689425,-0.0123711238,-0.2658644319,-0.1177591085,-0.1435891539,0.0613323227,0.074063696,-0.1460015923,0.1079467461,0.1995042413,0.1334598958,-0.162147671,-0.1651239544,0.1512049288,-0.1544630527,0.103398934,0.3309924901,0.0573625676,0.239774093,-0.3374667466,-0.1727189869,0.1334928423,-0.238129735,-0.1332807094,-0.095524244,-0.186577782,0.1056491062,-0.3508002162,-0.3200396001,-0.1402075738,0.0729287788,-0.1441500634,0.0400046743,0.272886306,-0.3790879846,-0.006065995,-0.1270800084,-0.3006229699,0.0165792163,0.473544091,-0.0619314499,-0.1693366766,0.7641561031,-0.1777873486,0.068674162,-0.1367510855,0.0628245026,0.0116071478,-0.1386356354,0.025742637,-0.0488918163,-0.1807256937,0.1018419191,0.5352259874,0.3329122066,0.0048212628,0.0025991534,-0.4617116153,-0.4422576427,-0.0901934057,-0.154354915,0.1846049279,-0.076524578,0.1344404966,0.2172488421,0.1043150127,-0.2009827197,-0.1766957045,0.0075936401,0.2428712398,-0.1265760958,0.0439214222,-0.0267230105,0.0873127356,0.0314299613,-0.0270371456,-0.3348520398,-0.072183609,-0.0649949685,0.2020932287,0.0967857689,-0.0286305249,0.0722042546,-0.1510536522,-0.0254527908,0.1781967431,-0.0268533044,0.0856483504,-0.0972304195,0.1517250687,0.3238803446,-0.0716767684,-0.0500325561,0.2173234522,0.0562485456,0.162016511,-0.0630055815,0.2864282727,-0.1369599849,0.0546018332,0.2025117874,0.1091905832,0.0082698623,0.2538399398,0.1179974079,-0.1682082862,0.141395539,-0.0499028228,0.3557972312,0.7263293862,-0.2195244879,-0.1248159185,0.4329024255,0.0563653782,-0.2547045946,-0.043995589,0.309862107,0.1377622634,0.0462357812,0.2116364539,0.0015198017,-0.038648434,0.1667878777,0.1361153722,0.1286683381,-0.2262919992,-0.0590387285,0.4483224452,-0.220318228,0.1624891609,0.390001893,0.0647899881,0.2133036703,0.2186936736,0.0370326042,0.6142647862,0.0564462729,0.0178854503,0.0840262175,-0.0497642569,-0.0006026002,0.2766371071,-0.0300385933,-0.0860835388,0.3275442719,0.5549071431,-0.350287497,-0.1590734571,-0.2640811503,-0.0866027623,0.0319981426,-0.1442377865,-0.285094142,0.0815575644,-0.2161394954,-0.1304976344,-0.1188771203,-0.3093909919,0.1627067924,0.0170424953,-0.398336798,-0.2444033772,-0.2909784913,0.1936600208,-0.1094031036,-0.2699058354,0.2208736092,-0.0130408434,0.032856442,0.2911531329,0.324870944,0.4699465036,0.1625403166,-0.318148464,-0.2325124741,-0.1214049459,0.0972115174,0.0350800194,0.3806981742,-0.0498869494,-0.1070307493,0.4631488025,0.0726584941,0.0371339954,0.2186426818,0.432112664,-0.4861977696,-0.1715066582,0.7934110165,-0.0755286664,-0.1411666721,-0.1295779049,0.2552585304,-0.2604231834,0.1343458146,0.0995590985,-0.0734473094,0.1301887184,0.0254731644,0.0394775085,-0.1064054593,0.3546760976,0.3110141754,-0.0438164733,-0.3694728017,-0.2513988316,-0.7986485958,0.2274625599,-0.1135691106,-0.323698312,-0.0648540929,-0.2603073418,0.1893703192,0.0481114648,-0.1414918154,0.1951838583,0.0548699573,0.1351767182,-0.1899445057,-0.3179088831,-0.2887228429,-0.0239487588,-0.1038828567,-0.386927247,0.1137444377,-0.3304968178,-0.0795861483,0.1184127927,-0.0097816065,0.4779797196,-0.0057918886,0.3654905558,-0.1124819741,0.7752766609,0.1879506856,0.0632824153,-0.1649430543,0.1221766546,-0.1206476018,0.1348381341,-0.0703359619,0.104410015,-0.1189628094,0.5212569237,-0.4437208772,0.4605461657,-0.0281022657,0.074082166,-0.3021473289,0.0456480756,-0.2554614246,0.1218354926,0.3444595039,0.141145438,-0.0916104317,0.3550699055,-0.2149428278,-0.1903913617,0.3962554634,-0.1034895405,-0.0939862877,0.0215162858,0.2646273375,-0.0517231002,-0.1400916278,-0.4393869638,0.1963266879,0.2907605767,-0.0937120169,0.1183897778,0.2602344155,0.2837514579,0.1340215057,0.0183274616,0.9827128649,-0.0597431548,0.1760434508,-0.1053840145,-0.2278604656]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"I have an initial version here: https:\/\/github.com\/EntilZha\/nlp\/tree\/master\/datasets\/qanta Thats pretty close to what I'll do as a PR, but still want to do some more sanity checks\/tests (just got tests passing).\r\n\r\nI figured out how to get all tests passing by adding a download command and some finagling with the data zip https:\/\/github.com\/EntilZha\/nlp\/blob\/master\/tests\/utils.py#L127\r\n\r\n","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":52,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nI have an initial version here: https:\/\/github.com\/EntilZha\/nlp\/tree\/master\/datasets\/qanta Thats pretty close to what I'll do as a PR, but still want to do some more sanity checks\/tests (just got tests passing).\r\n\r\nI figured out how to get all tests passing by adding a download command and some finagling with the data zip https:\/\/github.com\/EntilZha\/nlp\/blob\/master\/tests\/utils.py#L127\r\n\r\n","embeddings":[-0.1962168664,0.4000370204,-0.0314685591,-0.1644130796,0.0442840867,-0.133292973,0.2794850469,0.376173526,0.0094886748,0.0829261169,0.2064323872,0.1580682844,-0.3005688488,0.0680410042,0.1369243413,0.0012009284,-0.0564629063,0.1698959321,-0.097953327,0.2125060409,-0.1272202581,0.0748247653,0.0204572827,0.1080618128,-0.0279993936,0.0887562558,-0.2158944607,0.0343908481,-0.3809469342,-0.78580755,0.3818258643,0.0705721527,0.1851999313,0.3587058783,-0.0001226896,-0.0862562507,0.5880787969,-0.0560768247,-0.7038213611,-0.2316204906,0.2454068959,-0.4804557264,0.3018535376,-0.0215079132,0.1885171235,-0.131638214,0.2442083061,0.1043512151,0.1617511511,0.3224481046,0.1187366098,0.2776325345,0.0514680147,-0.1797240227,-0.1009447202,0.0860042721,-0.0892682076,0.5028808713,0.3648626804,0.0311197303,-0.1123634428,-0.1311188638,0.0922635347,0.4931352437,0.1250221133,-0.0222941581,0.5011738539,0.0191125106,-0.2128935456,0.0410488024,0.7716670036,-0.5064089894,-0.3754853308,-0.0786100179,0.0564953834,-0.1375145763,0.2607557178,-0.1600877792,-0.3946450949,0.0198199078,-0.1663755476,-0.280372113,-0.067122072,0.2935138643,-0.2830935717,0.2682918012,0.0777481496,0.1755786836,0.0895025879,-0.1008501872,0.2573432922,-0.0618371181,0.0244054813,0.2348817885,-0.0302161761,-0.2238747627,-0.2610195279,-0.2221421152,0.1922143996,0.087184079,-0.462625742,-0.1906755567,-0.0312074535,0.1495987922,0.14161475,0.4171294868,0.3384237289,0.238900423,0.2432880104,0.1627921462,0.2434393466,0.2634249628,-0.0598132797,-0.3704581857,-0.0324971564,0.0922335014,0.148252517,-0.3287043869,-0.1533401161,-0.2270715833,-0.3853175938,-0.0463141352,-0.0549251363,0.2376481295,-0.1088894606,-0.1465936154,0.0490113646,0.4120434821,-0.2063009888,-0.1511349827,-0.0294886027,0.0445264392,-0.1915404499,-0.0126141282,0.3086444139,0.0557580702,0.3780356348,-0.3671215475,-0.1176601425,-0.0343115553,0.1131467894,0.1852356195,-0.2689662874,0.2899587452,-0.1555463374,0.0520323552,0.0785048008,-0.2927011251,-0.2210675478,0.16356498,0.1203946844,-0.456910491,0.0378944427,0.0930617675,-0.6060150266,-0.064664647,0.0127269747,-0.1157290488,0.0913880691,-0.3326220512,0.026576288,-0.3913691044,-0.2052655965,-0.0672703683,-0.0997263268,0.2391047776,-0.266743809,-0.3372322023,-0.4481431246,-0.3259230256,0.1292015463,-0.2248050272,-0.0771240368,0.2149810642,-0.3794620037,-0.0085159829,0.549783051,-0.5216640234,-0.3586078286,0.4454957545,-0.4292176366,-0.2467299998,0.2341967672,-0.0489816442,-0.0027357105,-0.3621473014,-0.0894957483,0.0822743401,0.0496208407,-0.1781890541,-0.1884840727,-0.3439170718,0.1136098728,0.2200979888,-0.0398818441,0.088763535,0.0967642218,0.0540974848,0.3513938189,0.1143978909,0.1550448537,-0.1440186054,0.1771669984,-0.2378764302,-0.171195209,-0.0517720506,-0.5992661119,0.2523061931,-0.1184304655,0.2130238414,-0.046452675,-0.01118204,-0.5071431398,0.0033061588,0.0198404863,-0.2721042633,0.0020702777,0.2044220567,0.1296175867,-0.0479584932,-0.0106063243,0.0322238915,-0.2107295841,0.192533195,-0.0098091802,-0.0202862602,-0.0461895429,0.0624085069,0.2495948821,0.2555881739,0.0547183454,-0.3409820795,0.0079531223,0.2820033729,0.2083241642,0.1894648671,0.1891921461,-0.0701106638,0.1703492254,0.2639250159,0.1645806283,0.2629550397,-0.0151816439,-0.2123049349,-0.0861642584,0.244685933,0.168041423,0.0441920497,-0.1283625215,0.0657141581,0.3238113225,0.1138124764,-0.1348553449,-0.2280585468,0.1960445642,0.2852371633,0.6019870639,-0.0504562259,0.3259406686,-0.0479220226,-0.2781182826,-0.3706560135,0.0179368667,0.113899745,-0.0962130725,-0.0562753715,0.0941435024,0.3609081209,0.8490371704,-0.039258536,0.2709162831,0.0866793543,-0.1844285578,-0.2316581011,0.1509413719,0.1144368127,-0.0264656991,0.1556097716,0.0009208352,-0.4011897147,-0.089022167,-0.000131221,0.4331214428,0.39993155,-0.3471736908,-0.1098056212,-0.3021007478,-0.4445498586,-0.3985895216,-0.3630511463,-0.2080827355,-0.3813646138,-0.0614553764,0.221244067,-0.0000809908,0.2355280966,-0.1949504614,0.2832934558,-0.3094020486,-0.3343959749,0.1164741814,-0.2011949271,-0.4004581273,0.038076669,0.2813215256,0.0704870895,0.4426928461,-0.1078199968,-0.2973609269,-0.2132588625,-0.3188506663,0.1328554153,0.0609576516,0.5699258447,0.4761799574,-0.014884566,0.4204102755,0.0788300261,0.0420439132,-0.2335481942,-0.4335296154,-0.2341788411,0.0918297097,0.096914418,-0.1396645457,-0.397292465,-0.5266936421,-0.0855906904,0.1703432649,0.1270349622,0.0732077435,0.0939834714,-0.1114279777,0.2511834502,0.0868324786,-0.1874495,-0.1168527007,0.0710896701,-0.0090840636,-0.1429180652,-0.1871635616,0.066673696,-0.1160890758,0.0868185386,-0.1572062522,-0.4886566997,-0.4423266053,0.1542490423,0.2309943885,0.3925921023,-0.2067830712,0.2817326486,0.2018010765,0.1231473163,-0.1095167696,-0.3401258588,0.1602251381,0.0937714949,0.311019212,0.1370342374,0.3423713446,-0.117230013,0.7626640201,0.1187916175,-0.0941151008,-0.0333751999,-0.1344078928,0.0153888641,-0.0178408455,-0.0405953638,0.3350592554,-0.2001538575,-0.3104267418,0.2072171867,-0.1168439463,-0.0370542593,-0.1353754848,0.2891805768,-0.0174129494,-0.067305617,0.0038466188,-0.2191674858,0.008076339,0.09175165,-0.0282482766,-0.1992084831,-0.158573404,0.0310441516,0.3056081533,0.0092986375,0.1733637154,-0.2502235472,0.0066730492,-0.380459249,0.3209520578,0.3083269894,0.479110539,0.0899409577,0.040502023,0.1281770617,0.023768926,0.2255100757,0.0795255825,0.1641315222,0.2530296743,-0.1310341358,-0.0451933332,-0.2479988933,-0.0610883795,0.0637976378,0.0043010772,0.0955445841,-0.4097131491,-0.0987334698,0.1882557571,0.0891149938,-0.0782415271,-0.235160768,-0.0749968588,-0.0863107145,-0.1113193184,-0.2114881426,-0.2218316346,0.0170182362,-0.2150207311,0.4063696861,-0.0038531555,-0.0881171972,-0.0601546206,0.0969459936,0.0898773596,-0.2820567191,0.2924055457,0.2651253343,0.0748448074,0.3037792742,0.3153021038,-0.090810962,-0.2188899964,0.0253689457,0.0552031994,0.421171993,0.3506498933,0.1043400913,-0.05036759,-0.2399419993,-0.2073724568,-0.2343254685,0.1969753355,0.2857733071,-0.0309117921,0.0785086527,-0.3184976876,0.7194924355,0.2203308493,-0.2501448691,0.4313560724,-0.2180647254,-0.2628088892,-0.0657833815,0.1275826544,0.9033368826,0.1479494125,0.2377247214,0.4199579954,-0.269123733,0.4643270969,0.0728289336,0.1551105231,-0.4947299659,-0.000164842,0.0196805149,-0.1823747754,0.20951958,-0.0957571343,-0.2247008681,0.3749459088,0.146523118,0.2810683548,0.3701077104,0.3088001609,-0.2517709732,-0.0336605757,-0.1858089268,0.0902788639,-0.1498014629,0.4019285142,-0.2816435397,-0.0003146477,-0.1442373395,-0.044247698,-0.2529449463,-0.1122367978,-0.1734399796,0.062142238,0.1558529437,-0.1320602596,-0.0661598593,0.1833309978,0.0826906189,-0.2143256515,-0.1497820169,0.1254417002,-0.1433178186,0.1684724092,0.3818271458,0.0453288518,0.3266818821,-0.3169654012,-0.0581666119,0.067108579,-0.2803243399,-0.1247270703,-0.1329447627,-0.1724696457,0.1416563988,-0.2999446094,-0.1425621808,-0.1025121063,-0.0098974369,-0.1159026474,0.0223953407,0.2810628116,-0.3552967012,-0.0610115267,-0.1189609841,-0.2299093753,0.0096130008,0.4747706354,-0.0672381148,-0.1376446187,0.6765255928,-0.210376367,0.0658826604,-0.1299179345,0.0620127432,-0.1010729373,-0.1118434221,0.0767836049,-0.1696745902,-0.3008672893,0.0437146164,0.5850269198,0.4418942034,0.0124066016,-0.0704073608,-0.4628920555,-0.4116662741,-0.0003565765,-0.1414971799,0.206992209,-0.2121649981,0.220739156,0.1632817239,0.0460311994,-0.1872499585,-0.2017060667,0.0494651757,0.20594953,-0.1384761482,0.0434278436,-0.0155620379,0.144931972,-0.0339252241,0.0634906814,-0.361533463,-0.0769442469,-0.1011360362,0.2054294497,0.1118431985,-0.038949497,0.1556028426,-0.1173465401,0.0397215597,0.2143973559,-0.06621629,-0.0483590253,-0.0926889181,0.1126715988,0.3458280861,0.0979500711,-0.0069502778,0.3014909029,0.1000383347,0.1763518304,-0.116132766,0.2669962943,-0.2339322418,0.0885548741,0.2024891824,0.1292089671,-0.0194327626,0.2600130439,0.1396011412,-0.1141727567,0.0804191083,-0.162528798,0.3586398363,0.8415077925,-0.1782603711,-0.1310917437,0.4703676403,0.0327278115,-0.193390727,0.0271165464,0.2276788503,0.2025407404,0.0396529026,0.2661379278,0.010690527,-0.1374742985,0.1961124092,0.1407437474,0.0947431326,-0.1786990017,-0.0726684406,0.5383141637,-0.2553034723,0.1620742381,0.3502657712,0.0004753117,0.2156861126,0.0850607976,0.0254514813,0.5338089466,-0.0121567799,0.058235649,0.1135306433,0.0191257987,-0.0442432649,0.2593924999,0.0514435731,-0.0117760701,0.3216550648,0.6301919222,-0.4306803346,-0.1747279167,-0.2778612077,-0.0494090877,0.0822818875,-0.1499838978,-0.3964487612,0.0435975343,-0.2213106155,-0.0733620599,-0.1479509026,-0.318523705,0.2299109846,0.0032966724,-0.3555911481,-0.3461935818,-0.3967627585,0.140846774,-0.0358082242,-0.2752868831,0.2508650124,0.0294072777,0.0185320955,0.2829647362,0.2839174867,0.4441568255,0.0862858072,-0.4387360513,-0.2540848553,-0.1913308799,0.1358326077,-0.0776217729,0.3995730281,-0.0598165542,-0.1680218875,0.4992049932,0.0588679723,0.0435323976,0.2664315403,0.3724866211,-0.5206157565,-0.1689815074,0.8208150864,-0.0679488629,-0.1667515486,-0.027104279,0.1776790619,-0.2319854498,0.082595095,-0.0357043147,-0.047500506,0.1877605617,0.1194052771,0.0410767011,-0.1780636758,0.3992719054,0.3388817906,-0.0715001374,-0.3425282836,-0.2139325887,-0.8131753206,0.0885048583,-0.1005767658,-0.373714596,-0.0635504723,-0.2374126762,0.2535687387,0.1047838777,-0.2574350536,0.1771554649,-0.0015825713,0.1259434223,-0.2627958953,-0.3659875691,-0.2140853405,-0.0257645454,-0.0297405347,-0.4025479257,0.0784246027,-0.3133718371,-0.0992647335,0.2237176895,-0.0393956155,0.4940953255,0.0730826408,0.2537780106,-0.012197407,0.686855793,0.2427031398,-0.0313030891,-0.2473426312,0.1787670255,-0.1333087087,0.12535806,-0.1070144847,0.1350857019,-0.1702579111,0.6143173575,-0.4141833186,0.4506014287,0.074951686,0.0781901032,-0.2869772017,0.0479707234,-0.3413873613,0.1438796371,0.3280658424,0.3049733639,-0.1195661575,0.3443971872,-0.2657007277,-0.1519378871,0.4117482007,-0.2489467412,-0.0583861545,0.0572358184,0.2190218866,0.0009550565,-0.1586714834,-0.44350034,0.1210438833,0.2671409845,-0.1325504929,0.0848161206,0.1458554268,0.2846496701,0.0264419019,-0.0357408114,0.9360089898,-0.0723667219,0.1544691324,-0.0185700357,-0.2431814373]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"I'm quite positive that you can just replace the `dl_manager.download()` statements here: https:\/\/github.com\/EntilZha\/nlp\/blob\/4d46443b65f1f756921db8275594e6af008a1de7\/datasets\/qanta\/qanta.py#L194 with `dl_manager.download_and_extract()` even though you don't extract anything. I would prefer to avoid adding more functions to the MockDataLoadManager and keep it as simple as possible (It's already to complex now IMO). \r\n\r\nCould you check if you can replace the `download()` function? ","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":55,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nI'm quite positive that you can just replace the `dl_manager.download()` statements here: https:\/\/github.com\/EntilZha\/nlp\/blob\/4d46443b65f1f756921db8275594e6af008a1de7\/datasets\/qanta\/qanta.py#L194 with `dl_manager.download_and_extract()` even though you don't extract anything. I would prefer to avoid adding more functions to the MockDataLoadManager and keep it as simple as possible (It's already to complex now IMO). \r\n\r\nCould you check if you can replace the `download()` function? ","embeddings":[-0.1008703038,0.502432704,-0.0397975706,-0.0897842571,0.0968036205,-0.122407198,0.2872873247,0.3135527372,-0.0543848462,0.1157420203,0.046182245,0.1649335027,-0.2744551301,-0.0059635905,0.1659567803,-0.0370165035,-0.0780216157,0.2551133931,-0.0911455303,0.2149973661,-0.2306919545,0.1244610325,-0.0709374547,0.0745792389,0.136530906,0.1996980011,-0.1658058912,0.0543188453,-0.3160357475,-0.8764674664,0.3479210436,0.1809952855,0.2423820496,0.238788411,-0.0001248056,-0.1538030356,0.5954886079,-0.1265063286,-0.6724078059,-0.1728314757,0.2416220158,-0.4436886013,0.2643420398,-0.0716193542,0.2790195942,-0.1683751345,0.1821474433,0.1102833897,0.0740535781,0.3863360286,0.1194621101,0.2407578826,0.0858564675,-0.1201937124,-0.0584492311,0.0240839645,-0.0008228987,0.6254271865,0.40664801,0.0214351378,0.021082202,-0.0100838691,-0.0617214106,0.4501312375,0.2495122999,-0.0478202999,0.4997144639,-0.0700687692,-0.1631965488,-0.0136147235,0.6971593499,-0.5866323709,-0.3620384037,-0.0547511168,0.1543341428,-0.2042603791,0.1349426657,-0.2809634805,-0.4394977689,0.0580523983,-0.1019300595,-0.5488113761,0.0178486109,0.2542369962,-0.3011981547,0.2867898345,0.113982752,0.1454486102,0.2424277663,-0.0527879111,0.2136763483,-0.0221389402,-0.0824553668,0.1698232442,0.0019520939,-0.3044983447,-0.3149802387,-0.2155281156,0.0969157442,0.1349225342,-0.3737144768,-0.1438041925,-0.0563400127,0.1341730803,0.0918924063,0.4165343642,0.3381826282,0.220636338,0.1082044318,0.2282307893,0.2354674041,0.1206872538,0.0386737771,-0.361107856,0.0198340695,0.19144997,0.134788096,-0.1944695115,-0.0890687183,-0.2402393818,-0.4562417865,-0.0577771999,-0.1779349297,0.2095633745,-0.0875371099,-0.0898265392,0.1073288992,0.4088947475,-0.1238566339,-0.2714741826,0.0032230511,0.1967519969,-0.0980255678,0.1111704409,0.2187103182,0.0249425564,0.3383955359,-0.3740005493,-0.0343965143,-0.0379927233,0.1008777991,-0.01392633,-0.1752636433,0.297956109,-0.1332934499,0.0354545861,0.0203104541,-0.3518683016,-0.2085856944,0.2470930368,0.2287328094,-0.457187593,-0.0947381333,0.0599102452,-0.500505805,-0.0223911442,-0.0530996174,-0.0767954215,-0.0751673281,-0.5091549754,0.1566996872,-0.3630222678,-0.289362967,-0.0502463989,-0.0594202727,0.3765130639,-0.2985684574,-0.2998124659,-0.5405735373,-0.3079764843,0.1497205496,-0.2449992448,-0.164001286,0.2662369907,-0.3465722799,0.043164555,0.667219758,-0.6826146841,-0.3817332983,0.3879492581,-0.2968750894,-0.2016317993,0.1372314394,-0.0817428753,0.0714176744,-0.3140597939,-0.2028535903,-0.0059267748,0.091598101,-0.2129977643,-0.1380927116,-0.4739530683,0.1392556578,0.2211888134,0.000785049,0.2456368357,0.0335898325,0.1470759213,0.3634047806,0.1003377736,0.1316768378,-0.1588163972,0.2499465644,-0.188180238,-0.1837327778,-0.0881156474,-0.5891324878,0.2165798545,-0.1606362164,0.2297133803,-0.0706375539,-0.0018133485,-0.4160644412,-0.0398944467,0.0938376635,-0.2353090942,-0.025887005,0.1969702244,0.1955227703,0.0044661476,-0.0157303475,0.0643759146,-0.1620636135,0.2334327549,0.0740081891,-0.0516507477,-0.0220283866,0.1428957134,0.2040299326,0.2436191887,0.0309016332,-0.2502624691,-0.0366772115,0.3134894073,0.14573057,0.2744227052,0.1060874239,-0.040288914,0.1848415285,0.2803111076,0.1604785472,0.2589093149,0.1125444993,-0.1240083873,0.0815058127,0.1140125692,0.3063770533,0.0369060151,-0.1014102846,0.0488382876,0.2526002526,0.0661955997,-0.1218185648,-0.3277988434,-0.0246300884,0.2636627555,0.6258199811,-0.0983051136,0.3067036271,-0.1081027836,-0.2765682042,-0.3729353547,0.027126193,0.1354558617,-0.2431386113,-0.0593481697,0.0059099658,0.4128134847,0.8891434669,-0.0876502767,0.2127116024,0.2576935589,0.0994640961,-0.1476352811,0.3102177083,0.2731757164,0.036704842,0.1560715735,0.0172955096,-0.2442479581,-0.032426957,-0.0920960009,0.3309157193,0.3591976166,-0.3003346622,-0.1063411608,-0.3063665032,-0.5244471431,-0.369181931,-0.3438985646,-0.2026222795,-0.3367348611,-0.1670247614,0.2561099529,-0.0168017894,0.2510785758,-0.1769596338,0.23437576,-0.3581812084,-0.2618915737,0.0140888477,-0.0698123351,-0.2865777612,0.0055031381,0.2328591645,0.1618770957,0.2496272922,-0.2545626462,-0.2123430222,-0.3464898169,-0.1836167276,0.1987740993,0.2477274835,0.4137883484,0.5279138684,-0.0702143088,0.3567334414,0.0594498403,-0.0660828054,-0.1920313388,-0.3716994226,-0.3034237325,-0.0005273036,0.2784799933,-0.0836251453,-0.3137311041,-0.6029018164,-0.1412212849,0.0342323072,0.0753482878,0.1252205074,0.0305907503,-0.2410265803,0.2264879048,-0.0675989687,-0.2334192097,-0.20004794,0.1599311978,-0.0091451891,-0.1631018668,-0.0454618149,0.0740075111,-0.1086775586,0.0565040857,-0.2311774343,-0.439026922,-0.4993732572,0.2032101899,-0.0114732869,0.1493686587,-0.1808473319,0.2266940176,0.2446268052,0.1441554874,0.0077745435,-0.4050812125,0.1296161264,0.1733418554,0.259601444,0.2025204897,0.4707221985,-0.1895778328,0.6809964776,0.0801900625,-0.0654067695,-0.0069155768,-0.1938064098,0.1298341453,0.091098316,-0.0736017227,0.2188321948,-0.190089196,-0.2723953426,0.260730058,-0.3641332686,-0.0907397792,-0.1529102176,0.2148099244,-0.1826370656,-0.1637751907,0.077838324,-0.2045095861,0.0020857533,0.1484215856,0.0000834311,-0.2190662026,-0.0584676079,0.1079579443,0.2922641635,0.0909208879,0.1173679158,-0.2829025984,-0.0073552672,-0.4521165788,0.357547313,0.3472754955,0.4865115583,0.2158142477,0.0392039232,0.0082475133,0.0407738611,0.2137794644,-0.0299481209,0.013005143,0.2556817532,-0.1364886314,0.0614763647,-0.2255053818,-0.0926549658,0.0719013214,0.0817237124,-0.070730336,-0.3118750453,-0.1512361765,0.3328527212,0.1557936221,-0.1117549837,-0.2423667908,0.008510354,-0.1531419456,-0.1970023662,-0.1900701225,-0.245558694,0.138157025,-0.1238618493,0.4470300078,0.0135273598,-0.0988405049,0.0870625898,0.0247686282,0.1871044934,-0.1544833183,0.1312623471,0.1785686314,0.0028186338,0.3041517138,0.2868061364,-0.024945613,-0.2621819675,0.0249094851,0.1132410467,0.3784664869,0.3422511518,0.1974110007,0.0638271943,-0.2843548059,-0.1820515543,-0.1446656585,0.1674748361,0.3507558405,-0.1033928171,0.0512712412,-0.2381137162,0.6501333117,0.1520873159,-0.217585668,0.2924677134,-0.1957999617,-0.2512966692,-0.12040589,0.1437388211,0.9772185683,0.1787049472,0.1921579391,0.3799656034,-0.270180136,0.5282026529,0.1216726899,0.114315249,-0.4974674284,-0.0368093438,0.0127716009,-0.2195836157,0.2260390371,-0.1551375389,-0.2371827513,0.3542981744,-0.0105408495,0.2823988795,0.3365881741,0.3203299046,-0.2227639556,-0.0262488239,-0.1613601446,0.0425570756,-0.0497173704,0.4177896678,-0.2992789745,0.0098755704,-0.0863077492,-0.0455262996,-0.3256154656,-0.2006005049,-0.2927237749,0.0171926655,0.0827980638,-0.1399679929,-0.0106162168,0.177964285,0.0898156464,-0.2160074413,-0.099816151,0.0924351439,-0.1403939575,0.2634117603,0.3869937658,0.0964832827,0.2686981559,-0.3461075127,-0.0996442512,0.3046475649,-0.2349199653,-0.2469947338,-0.0418010913,-0.2066766471,0.1384481788,-0.1916412562,-0.3267768025,-0.025652919,0.0002871492,-0.1344084144,-0.0031316129,0.1647348553,-0.3638896346,-0.0080352966,-0.2035221756,-0.2534616292,0.0494406484,0.4804393351,-0.0290465169,-0.2206923962,0.665512979,-0.2630746365,0.1176692098,-0.1220518202,-0.0074471156,-0.0930061489,-0.1092986688,0.0760425776,-0.1761204749,-0.1763357222,-0.0106830439,0.5820413232,0.3484705985,0.0852458999,-0.1574584991,-0.4427753985,-0.4029706717,0.0175911728,-0.0671456382,0.137936607,-0.1804800183,0.1810571551,0.1291481555,0.0830077603,-0.1761809289,-0.1088118926,0.0277180225,0.101576902,-0.1334145218,0.1775039285,0.0053310166,0.1419469863,-0.0623018146,0.1782146394,-0.2920298278,-0.052992478,-0.1140303165,0.2268325239,0.0238888226,0.0608929805,0.1892109811,-0.0479664057,0.0154750412,0.2455648631,-0.1041474491,-0.0335548967,-0.0547389202,0.2197332233,0.3419250846,0.042103827,-0.136351794,0.2392885983,0.2469777465,0.2195747197,-0.1397819966,0.3792563081,-0.117185995,0.0650610402,0.2873711288,0.045283135,-0.079984732,0.2675745487,0.0938475803,0.0307124015,0.1152932867,-0.0701815635,0.2299199998,0.8423094153,-0.2426541299,-0.116752483,0.4015055597,0.0097817425,-0.2332358062,-0.0803140104,0.2138213217,0.0644630119,0.0096819159,0.2943811119,0.049018532,-0.0173016116,0.3186424971,0.0939385295,0.0876453668,-0.1457333267,-0.0733506009,0.5390332937,-0.2602725029,0.212284714,0.3923049271,0.139943406,0.3845630586,0.1599566191,-0.0668777972,0.5641658306,-0.0652574375,0.0076857517,0.2044231594,-0.0826479867,-0.089759849,0.3919528723,0.0287756976,0.0224167928,0.3010084629,0.3995398581,-0.3488951027,-0.167156443,-0.2033413798,0.0306608472,0.1244575903,-0.1107562184,-0.2374506593,0.0058272076,-0.2603170276,-0.0581110939,-0.1811089665,-0.3672569096,0.2537269294,-0.0513663068,-0.3961371481,-0.1772397012,-0.589852035,0.0297317766,-0.1069158465,-0.269297272,0.0516767986,0.0317018926,0.0756488368,0.2259105146,0.4408133328,0.3618397117,0.1011862904,-0.3247079551,-0.3104074299,-0.2057196498,0.1689282358,-0.0008590583,0.2647931278,0.0646267012,-0.1833536923,0.4019779861,0.034111727,0.051967375,0.181642279,0.4109761119,-0.4785670042,-0.271351248,0.9588337541,0.0185881164,-0.0753741264,-0.1051436961,0.2092437148,-0.245941177,0.153588891,-0.0540698208,-0.025079608,0.1652690768,0.2022853792,0.011583522,-0.0484386608,0.2834643126,0.2929687798,-0.0652442202,-0.2998435199,-0.32373631,-0.8680157065,0.1610240638,0.0929507688,-0.2784693241,-0.0194217283,-0.2899569869,0.1758711487,0.1876767278,-0.2499030977,0.2563643157,-0.0072891396,-0.035454493,-0.2857216895,-0.2891827226,-0.2412427068,-0.0238645561,-0.0229866449,-0.3951801658,0.1954310685,-0.4377284646,-0.1360102147,0.1604931056,-0.0511066616,0.4187010825,0.0776288733,0.4448586702,-0.0341765992,0.50524652,0.153489694,-0.0501542352,-0.0478207953,0.2508709133,-0.1411208659,0.1136065871,-0.0732937083,0.094286494,-0.175388813,0.6784105301,-0.4178821445,0.4013880491,0.011650905,0.1236110851,-0.4917762876,-0.002498212,-0.173773855,0.146238029,0.379739821,0.2535791397,-0.1776374876,0.3676301241,-0.422621578,0.0261848643,0.5046986341,-0.1647151858,-0.0771925226,-0.0422559828,0.3168366849,-0.0462424122,-0.2065518647,-0.5365794301,0.065256767,0.2432874888,-0.0635266379,0.1409955323,0.2062073201,0.1786971688,0.0790715143,-0.0381838605,0.8222998381,-0.0023479762,0.1193780079,-0.1942015737,-0.2611887753]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"I might be doing something wrong, but swapping those two gives this error:\r\n```\r\n> with open(path) as f:\r\nE IsADirectoryError: [Errno 21] Is a directory: 'datasets\/qanta\/dummy\/mode=first,char_skip=25\/2018.4.18\/dummy_data-zip-extracted\/dummy_data'\r\n\r\nsrc\/nlp\/datasets\/qanta\/3d965403133687b819905ead4b69af7bcee365865279b2f797c79f809b4490c3\/qanta.py:280: IsADirectoryError\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n```\r\n\r\nSo it seems like the directory name is getting passed. Is this not functioning as expected, or is there some caching happening maybe? I deleted the dummy files and re-ran the import script with no changes. I'm digging a bit in with a debugger, but no clear reason yet","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":88,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nI might be doing something wrong, but swapping those two gives this error:\r\n```\r\n> with open(path) as f:\r\nE IsADirectoryError: [Errno 21] Is a directory: 'datasets\/qanta\/dummy\/mode=first,char_skip=25\/2018.4.18\/dummy_data-zip-extracted\/dummy_data'\r\n\r\nsrc\/nlp\/datasets\/qanta\/3d965403133687b819905ead4b69af7bcee365865279b2f797c79f809b4490c3\/qanta.py:280: IsADirectoryError\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n```\r\n\r\nSo it seems like the directory name is getting passed. Is this not functioning as expected, or is there some caching happening maybe? I deleted the dummy files and re-ran the import script with no changes. I'm digging a bit in with a debugger, but no clear reason yet","embeddings":[0.0540467873,0.074223496,0.0096600102,0.1716468334,0.0184579045,-0.1827674061,0.4075564146,0.1807362586,0.0317732841,0.2498853356,0.0380608626,0.068305321,-0.0652733743,-0.2527975738,0.0166334137,0.1595787108,0.1347910911,0.0446650498,-0.2199352086,-0.0055111041,-0.3515838087,0.2071037889,-0.187350601,0.1459729224,-0.0956168994,0.270640105,-0.2202364653,0.2536099851,-0.1994191855,-0.5675512552,0.1467316002,0.0295442697,0.1657156944,0.5746880174,-0.0001326538,-0.0771422461,0.5401037335,-0.1227790937,-0.6687539816,-0.2304881364,0.2879289985,-0.1529602557,0.2021133006,-0.0812509581,0.270616293,-0.1741900593,0.204366371,-0.0653551742,0.1248996928,0.3529885709,0.0195018034,0.1126961336,-0.0109825647,0.0728081167,0.0934385285,0.2516900599,-0.0975718126,0.4664695859,0.2114457041,0.0007738983,-0.045916073,-0.0158069301,0.0191086475,0.4609082341,0.3064597249,0.1869644523,0.4650607109,0.1037350446,-0.0125133693,-0.0378652699,0.6569068432,-0.4838958979,-0.4344105422,-0.0811828449,-0.12728405,-0.2094848007,0.3831119239,-0.1031870842,-0.3200292885,0.0950442329,0.0144911129,-0.26167503,-0.0353864059,0.2144244164,-0.3736029565,0.3995922506,0.0347100683,0.193312183,0.0728287771,-0.0495952033,0.3031981289,-0.342491895,0.0469243564,0.1650336087,0.0996335447,-0.0047323732,-0.1664235592,-0.0624356754,-0.1453322917,0.0344626084,-0.4331528246,-0.1392920613,-0.1853588372,0.2131967545,0.1850105375,0.4487712383,0.2529617548,0.49305287,0.1789587885,0.1321790516,0.1347615123,0.1885264963,-0.007200378,-0.4462205172,-0.0033650703,0.1070362031,0.2182107419,-0.2600002587,-0.2631745636,-0.2860175371,-0.4661550224,-0.0361489393,-0.0577055514,0.2854334116,-0.2194109708,-0.1112207621,0.0685082227,0.3858869076,-0.1995007694,0.0582178086,0.0005531253,0.0390481763,-0.2579016089,0.0363133103,0.0975848287,0.1127365455,0.2136280388,-0.327388227,0.1367980242,-0.1455704868,0.0729196742,-0.0004475712,0.1181760207,0.3578861058,-0.1497733146,0.190330863,0.1012113616,-0.4527332783,-0.2562183142,0.2993798554,-0.1043073907,-0.4040018916,0.0991009325,-0.0051061912,-0.426192373,0.0608746558,0.0788302869,-0.3964565992,0.1935182959,-0.5260314941,0.0671956837,-0.2752553523,-0.2804777026,-0.1202278137,-0.0989401713,0.4806630909,-0.3474727273,-0.2877002656,-0.3806169927,-0.0432820767,0.1235175207,-0.0717245266,-0.1209199354,0.0439575091,-0.5625425577,0.0187574141,0.4358156621,-0.5859128833,-0.3239840269,0.3623735905,-0.502648592,-0.1966268867,0.4976197779,-0.1181100309,-0.0193111189,-0.4362893999,0.0002922537,0.0490591675,0.0698962659,-0.041126553,-0.1771132052,-0.2051967978,-0.0250027739,0.0876833275,-0.1762067229,0.1093476638,0.0455018543,-0.105764769,0.4786712527,0.0737257749,0.0964748412,-0.0445868708,0.2020455003,0.0504911467,-0.0411830917,0.2556861043,-0.569576025,0.1495259553,-0.1553898454,0.1262127012,-0.2437577695,-0.0856458768,-0.3687703311,0.064841494,-0.1458710283,-0.156398654,-0.1060830504,0.5268377662,-0.0165023785,-0.0627749637,0.0430927873,0.0877717286,-0.1442557126,0.2422386706,-0.1104285866,0.0053613787,-0.0516392998,0.0858540013,-0.0734896287,0.0308981612,0.2242912948,-0.4081809521,0.033209011,0.3498995602,0.2747839093,0.4627654254,-0.0225549079,-0.13486889,0.1559964418,0.2479833215,0.3091241717,0.344550848,0.08198791,-0.2899188101,-0.0191937685,0.2700664103,0.1689306349,0.3189726174,-0.0317878425,-0.0494062528,0.0412943549,0.133926481,-0.0626230836,-0.2984604239,0.1893189549,0.1693602651,0.5996592045,0.0712876692,0.3279828131,0.0256755203,-0.1727201939,-0.2790479958,-0.0857859626,-0.0000382862,-0.1887648851,-0.0837015808,0.0913325027,0.405679822,0.9681350589,-0.0937558413,0.3328851759,0.0330499038,-0.1080778763,-0.2747818828,0.2888103724,0.2276344448,-0.0749421865,0.1733702123,-0.0313987061,-0.3184027076,-0.0401051641,-0.0503057837,0.5000056624,0.1426006258,-0.5107773542,0.2243059576,-0.2545761764,-0.3097943664,-0.6035177708,-0.3119944334,-0.0605385303,-0.5021187663,-0.095515579,0.3961226344,-0.0304831285,0.2094062269,-0.0747649297,0.084592104,-0.443292588,-0.4474219978,-0.0140934829,0.0082636159,-0.4845774174,-0.1461125612,0.3326219916,-0.0875561163,0.2277921587,-0.2338760495,-0.20161964,-0.2357240617,-0.2972360849,0.1263270974,0.1579475105,0.4473168254,0.2323800027,0.0002066567,0.4607324302,0.0031133909,0.1273820102,-0.2060612142,-0.3891748786,-0.1575384289,0.127281487,0.1342288107,-0.1593408734,-0.391477704,-0.4792369008,-0.0361469463,-0.095289968,0.0591819324,0.1256530136,0.2484227568,-0.1625206172,0.14527677,0.0638242066,-0.0957263485,-0.0330404565,0.0895208791,0.0384316444,-0.2388996333,-0.0015528752,0.0442732498,0.0077284467,0.0566343069,-0.0904529095,-0.4622611701,-0.4703610241,0.1932881773,0.0555998646,0.1804823577,-0.1026051864,0.4552404583,0.2974390686,0.1375872344,-0.0904007852,-0.3450211287,0.0245696828,0.080518268,0.3078872859,0.1523377001,0.3881365955,-0.1925664693,0.8443017602,0.1262982041,0.0352677852,0.0838550553,-0.1537719369,0.1894443929,-0.0821860284,-0.0315487906,0.1504146904,-0.0836085528,-0.5522133708,0.3674583733,-0.0571217388,-0.2313218266,-0.0664763227,0.227354452,-0.1440651268,-0.1588514894,0.0742010176,-0.1906691194,-0.0124345142,0.1656012833,0.0025182383,-0.2638314962,-0.2158576101,0.0124256136,0.3022841215,0.0273345821,0.0839223415,-0.3305439949,-0.1284147054,-0.1475516111,0.3084338605,0.3300766051,0.410194248,-0.0159425363,-0.1134252399,0.2813284695,0.0076485169,0.3506708443,0.0673457757,0.2124343663,0.195330739,-0.0024008385,-0.1179331318,-0.1861052811,-0.0773607567,0.1471854001,0.1093931645,0.1820794195,-0.2384732217,-0.2643302977,0.2099795491,0.196324572,-0.1552478224,-0.2132079154,-0.2094291151,-0.1290473193,-0.1241961047,-0.3455802202,-0.1391733736,-0.0629343912,-0.0944626927,0.3878278434,0.0180963613,-0.1835798025,0.0552175306,0.0047634016,0.3092903495,-0.1802646071,0.153189078,0.3662014008,0.1074476168,0.4060262442,0.3897089064,-0.1806146204,-0.1275579929,0.0257556569,0.0957939625,0.3917335272,0.6246490479,0.046178434,-0.1896109134,-0.2906071842,-0.2193540782,-0.2398288548,0.0960141644,0.4089126289,0.0091063883,0.1007163748,-0.285456717,0.5554720759,0.208112061,-0.3564368784,0.5225517154,-0.0035550273,-0.2918550968,0.1154966503,-0.0511861555,1.1163867712,0.2942337692,0.4239664972,0.3608600795,-0.20886603,0.0946512669,0.1167899743,0.1752223819,-0.4606740475,-0.0215319972,-0.0190337971,-0.3330588043,0.2362798452,-0.0354699418,-0.0998866409,0.4064423144,-0.0400789194,0.3567273915,0.1936495155,0.1573718339,-0.2964200974,0.0282647684,0.0261446033,-0.0406553186,0.2072830051,0.4589496255,-0.1629308909,0.0317801163,-0.2632970214,-0.0773327425,-0.0307313856,-0.0730969682,-0.1340214908,0.0746179894,0.0222444478,-0.2519100904,-0.2052024752,0.3793680668,0.1928956509,-0.2292574197,0.0226006396,0.1731875092,-0.1718836427,0.3654879034,0.3521059752,-0.0413430743,0.304631412,-0.1348404884,-0.1324376911,0.1016066894,-0.2835325897,-0.1992300898,-0.2421751618,-0.0925786421,0.225916639,-0.3246323168,-0.1483439952,-0.1495778114,-0.0864616334,-0.1007796079,-0.0521918572,0.1412243545,-0.2876558602,0.0100730034,-0.2608498931,-0.4288968742,-0.0010192889,0.5187016726,-0.1206838861,-0.0558586083,0.8927490115,-0.2553901076,0.070040524,0.0378354825,0.160714224,-0.2221290022,-0.2651758194,0.0671496913,-0.1813219786,-0.2439156473,0.0216876529,0.5689189434,0.3260739744,0.0817907229,-0.3220901489,-0.4693024755,-0.2484159619,0.0068611214,-0.1358154863,0.1807656735,-0.2004820555,0.0283805616,0.182130903,-0.1174349263,-0.1076878458,-0.0199809838,0.1194526404,0.2296413183,-0.1515396088,0.0516107455,0.0627184659,0.0399596728,-0.0837176368,0.1038524806,-0.2190838158,0.0567347407,-0.0864666179,0.264562428,0.0287008043,0.0045508579,0.0061855926,-0.3258795142,0.1301833242,0.0681178644,-0.0231584329,0.0288918614,-0.0509508103,0.0426695421,0.3008279204,0.061457824,0.0294049513,0.3885272741,0.0578088611,0.2583294809,-0.1853506267,0.2191310674,-0.2305950075,0.0207664892,0.0404156595,0.1000647545,-0.0187642165,0.2092515379,0.1339735091,-0.1802221984,0.1182851866,-0.1002387851,0.3639490008,0.9179779887,-0.1468938738,-0.2300969958,0.4129927754,-0.0907869786,-0.0570111573,-0.1269749701,0.0788692236,0.1358591765,-0.0584469289,0.3512894809,-0.0528068207,-0.1174375862,0.3474311233,0.2113447636,0.2010032088,-0.2374160737,-0.1016400009,0.4046128988,-0.3055979908,0.4028894901,0.4748873711,0.0623087436,0.2003504038,0.325150013,0.0861221999,0.5218539238,-0.145374164,-0.0289070141,0.1209087968,-0.1082878634,-0.0298490319,0.2552218437,-0.005975917,0.044361677,0.3560171723,0.5558943748,-0.4765009582,-0.1052879617,-0.2333438694,-0.1378047317,0.0167639405,-0.2931414843,-0.2188497186,-0.0358028077,-0.1372567117,-0.0325974971,-0.1987469494,-0.199868381,0.2898299396,0.0561583489,-0.2658578753,-0.2286198884,-0.3638243377,0.0198975503,0.0399810299,-0.3475835323,0.2545584738,-0.0002539147,0.1358793974,0.2975616753,0.1314556748,0.4875786901,0.19343777,-0.2696227133,-0.2617158294,-0.2879703939,-0.0018290355,-0.077898927,0.3554931581,-0.0298691578,-0.0914938524,0.3583312631,-0.0157092474,-0.0056668585,0.2206005305,0.4076079428,-0.2764085233,-0.1498058587,0.8099423647,-0.2403461784,-0.210106045,-0.0757306516,0.0336879827,-0.4592143297,0.1271804422,0.011043923,-0.1447659731,0.1875212491,0.1113790199,-0.0391637161,-0.0905533805,0.342149049,0.4022055566,-0.1118356511,-0.3300437927,-0.1995304823,-0.8534030318,0.2650276721,-0.0500908159,-0.1436564773,-0.2096838355,-0.0795226991,0.2140345126,-0.0171053782,-0.110200502,0.1108071133,0.0584933311,0.1832573861,-0.3073981106,-0.3775624037,-0.1809949279,0.0396495759,-0.0084772902,-0.5915566683,0.1994027644,-0.417517215,-0.1523204446,0.161968857,-0.0717150271,0.3689966202,0.2090159506,0.3645198643,-0.1086046845,0.6161589622,0.1779766977,-0.1468058676,-0.2628196776,0.0578591675,-0.0972580612,0.0183628853,-0.0892748833,0.1736892015,-0.3242898881,0.4118089378,-0.3782767355,0.220134005,0.0850958452,-0.0034846333,-0.2222467661,0.1844734699,-0.3588902652,0.1709943861,0.324440062,0.327229321,0.0407465696,0.3295919299,-0.3035615683,-0.2424766272,0.6405374408,-0.0990562588,-0.1643478274,0.136787042,0.2945452929,-0.0111584049,-0.2771060765,-0.5036939979,0.1726923585,0.2413635254,0.0693380758,0.0241263155,0.0519338213,0.2529217601,0.0171920918,-0.0656626821,0.918079555,0.0310803764,0.0706129819,0.0990286246,-0.2034779936]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"From what I can tell here: https:\/\/github.com\/huggingface\/nlp\/blob\/master\/tests\/utils.py#L115\r\n\r\n1. `data_url` is the correct http link\r\n2. `path_to_dummy_data` is a directory, which is causing the issue\r\n\r\nThat path comes from `download_dummy_data`, which I think assumes that the data comes from the zip file, but isn't aware of individual files. So it seems like it data manager needs to be aware if the url its getting is for a file or a zip\/directory, and pass this information along. This might happen in `download_dummy_data`, but probably better to happen in `download_and_extract`? Maybe a simple check to see if `os.path.basename` returns the dummy data zip filename, if not then join paths with the basename of the url?","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":112,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nFrom what I can tell here: https:\/\/github.com\/huggingface\/nlp\/blob\/master\/tests\/utils.py#L115\r\n\r\n1. `data_url` is the correct http link\r\n2. `path_to_dummy_data` is a directory, which is causing the issue\r\n\r\nThat path comes from `download_dummy_data`, which I think assumes that the data comes from the zip file, but isn't aware of individual files. So it seems like it data manager needs to be aware if the url its getting is for a file or a zip\/directory, and pass this information along. This might happen in `download_dummy_data`, but probably better to happen in `download_and_extract`? Maybe a simple check to see if `os.path.basename` returns the dummy data zip filename, if not then join paths with the basename of the url?","embeddings":[-0.0789025053,0.2855144143,0.0002484272,0.052754432,0.0506349877,-0.260754168,0.2143824548,0.2759754658,-0.054033909,0.0801737085,0.207228303,0.223573193,-0.2214686573,-0.0345815979,0.1244285852,0.0544766523,-0.1065842211,0.0847872198,-0.2250418067,0.155033648,-0.2493224442,0.1753319949,0.0034214412,0.1672303677,0.0154142119,0.1461061686,-0.2237707675,0.2018211782,-0.4194307923,-0.7930223346,0.3121376932,0.0302134454,0.2352778167,0.3840055466,-0.0001236156,-0.0918525681,0.589799881,-0.1559888422,-0.7334056497,-0.4215240777,0.2433728427,-0.5072749257,0.2773010731,-0.1551025063,0.185713321,-0.0634940639,0.2981808186,0.2096097767,0.1883738339,0.4003967941,0.0724132881,0.3003302813,0.0519982874,-0.1126803607,0.0085209534,0.1745276898,-0.0521616563,0.587381959,0.2025229931,0.0743161216,-0.1562705785,-0.0790689439,0.1053197011,0.5026990175,0.1522777379,0.0629803464,0.3990126252,0.0465288758,-0.2675873637,0.0866298825,0.6675570607,-0.5247456431,-0.3587066829,-0.0709844455,-0.0790122375,-0.0574314632,0.3807904422,-0.1414193809,-0.412658751,0.09209878,-0.1064498127,-0.1712896377,-0.0319949836,0.3967043757,-0.2983646691,0.4051716924,-0.0184657462,0.1622784734,0.1562570482,0.059117496,0.2566516697,-0.1750605255,0.1734549403,0.1806305349,0.1060590521,-0.2094244361,-0.2895837724,-0.2742270827,0.1452919543,0.1114526391,-0.3082630634,-0.1924095154,-0.0626133084,0.2252860814,0.2307136059,0.324116379,0.1796365976,0.269238323,0.2589327395,0.3053899407,0.2700368166,0.1619523466,0.0072083711,-0.3309308589,-0.121018663,0.1566589326,0.204818055,-0.212076515,-0.1972972006,-0.2485962361,-0.4048675597,-0.1265592724,-0.0687749311,0.311609596,-0.1900963336,-0.1030601785,0.092429176,0.3832482696,-0.1280384213,-0.0707097128,0.0286854729,0.04314382,-0.2482326329,0.0286369994,0.2963480651,0.2202127427,0.4854156077,-0.2904739082,-0.1611180604,-0.0584542193,0.1296471208,0.0434607528,-0.0472125523,0.2191486508,-0.0758894384,0.1827891171,0.0915671811,-0.3037326634,-0.299829036,0.1725249887,-0.0722899139,-0.4023593962,0.1072350442,0.0654562116,-0.5473223329,0.0217342135,0.0958953723,-0.1695723236,-0.0575396009,-0.3571426868,0.04796464,-0.2368825972,-0.2540795207,-0.1595515907,-0.0131576732,0.3331227899,-0.1687354594,-0.3116174638,-0.4004590511,-0.2978602946,0.2116440386,-0.1955409944,-0.1178338006,0.2728709877,-0.4967008233,0.0652605966,0.6163495183,-0.4722355604,-0.2997187078,0.5806676149,-0.4525257647,-0.0682009533,0.2638137341,-0.0369549468,-0.2212249935,-0.2977394462,-0.084492892,0.1349065602,0.0260605756,-0.1500234604,-0.1978701055,-0.3536165655,0.1459461004,0.2388371378,-0.0879810899,0.092689544,0.0530852154,-0.070395194,0.4528979957,0.106686607,0.0637685955,-0.0611375459,0.1207884699,-0.1968197525,-0.1071662679,0.0261884611,-0.5845915079,0.1251316816,-0.3148132861,0.1976306736,-0.1885021478,-0.0742158294,-0.5055440068,-0.018279003,0.0228491519,-0.2564983964,0.0097338939,0.3550091982,0.0908067226,-0.0491765253,0.0322510339,0.0155851385,-0.2026489973,0.2942970097,-0.0109322155,0.0641305745,-0.0877616107,0.1153794304,0.2468134761,0.25044626,0.1680199355,-0.4039673805,0.0306897704,0.3176622689,0.1910518259,0.1839980334,0.201267451,-0.0668875054,0.1683950275,0.1911004186,0.0835079551,0.4694286585,0.0475594029,-0.1295868456,0.0458018295,0.2079724371,0.1549796015,-0.0021298323,-0.0742816553,0.0792869031,0.3448398411,0.0221743993,-0.1013446897,-0.254516542,0.1040521711,0.3349339366,0.6334264874,-0.0067082136,0.2735137343,-0.1097401306,-0.1404369324,-0.3109175265,0.0469735488,0.1404445171,-0.1851261407,-0.1090435013,0.000903753,0.3755458593,0.8084435463,-0.008677111,0.3378500342,0.0986356437,-0.1634431928,-0.23178491,0.246971488,0.1726222634,-0.091688782,0.1258708537,-0.1599158794,-0.3018528819,-0.1450392008,0.0199347269,0.4233738482,0.3181325495,-0.4547547996,-0.0568398125,-0.2974067032,-0.5187103748,-0.5051652193,-0.3165683746,-0.2091601044,-0.3562681377,-0.056082841,0.2834823728,-0.1309288889,0.1668241769,-0.1150368974,0.2502073944,-0.3294556737,-0.3059851527,0.0541405305,-0.0925322548,-0.4306950271,0.0013973453,0.4053305984,0.0908558294,0.2507201433,-0.2317565978,-0.2034363896,-0.262078166,-0.2770365477,0.1347783506,0.1804633737,0.5565587878,0.4462135732,-0.0005814062,0.4154549241,0.0234471709,0.021318879,-0.1771230251,-0.3594786823,-0.1665748209,0.1146169975,0.1163782179,-0.1748237312,-0.435265094,-0.5285280347,-0.0585300401,0.0292403344,0.0863750428,0.1540934891,0.09188582,-0.1715522707,0.1888977438,0.042821262,-0.1449873298,-0.1551212221,-0.0749296099,0.0263330638,-0.1407961845,-0.1739190817,0.0086749513,-0.2095603198,0.0579381362,-0.1622326523,-0.4062244296,-0.5565926433,0.1409670264,0.1812706739,0.2534226179,-0.150748536,0.3773111403,0.182440564,0.0918047354,-0.1326707155,-0.3363351524,0.2459822148,0.1070011258,0.4016033113,0.2060058713,0.2957031429,-0.0263223983,0.7256447673,0.149187237,-0.0722948611,0.0998502597,-0.1860619634,0.1038331017,0.1562849879,0.0059008389,0.1845206171,-0.1581729054,-0.2670913637,0.318619132,-0.1976814717,-0.0466639511,-0.1197159663,0.2846116424,-0.2596027255,-0.1273642629,0.1273560971,-0.0586019456,-0.0435282215,0.0980659053,-0.0489263907,-0.1612470448,-0.0613584928,0.1211147308,0.3515174389,0.0332875401,0.1457083374,-0.3123940229,-0.0485616364,-0.3304588497,0.2873315513,0.3822647929,0.4291678369,0.0589802563,0.0130436877,0.0952789038,0.0012857775,0.2869000137,0.0190521572,0.1585067511,0.1310186386,-0.1548553109,0.0247319937,-0.1552329212,-0.0956421122,0.0984688178,0.0075187609,0.1507926136,-0.4683267176,-0.1387944072,0.1901915222,0.1147199869,-0.0580222867,-0.2220772505,-0.0159309171,-0.121208258,-0.0473067313,-0.2221316844,-0.1463329196,-0.0098731993,-0.1014291421,0.2902652323,0.0774345249,-0.1481548101,-0.014837685,0.0885698199,0.2045015246,-0.3280986845,0.11342334,0.2718537152,0.1991793513,0.2238492817,0.4006248415,-0.1000059247,-0.2102216184,0.0339924544,0.0597412959,0.3214836419,0.40066275,0.1043481901,-0.0390596539,-0.2513598502,-0.1566225886,-0.3356704414,0.2235788554,0.4120943844,-0.1030944213,0.1445936859,-0.2291204035,0.6646540165,0.1315937787,-0.238384977,0.5089071393,-0.2838887274,-0.2806597054,-0.0773645267,0.0467766859,1.0281779766,0.2466938645,0.2519368827,0.4927608669,-0.2603980303,0.292863071,0.0694463551,0.0985822156,-0.4321812391,0.0190815199,-0.0640499592,-0.2019298822,0.2541381717,0.0323692188,-0.2334097326,0.3385886252,0.1327560693,0.3894303739,0.2429597378,0.4002089202,-0.2431952655,-0.0215322971,-0.2314871401,0.0444953218,-0.0291538797,0.3578965068,-0.2018131912,-0.0817456171,-0.2016923726,-0.132237047,-0.0011776419,-0.0818459913,-0.3922837377,0.0288296789,0.0667122155,-0.1053429767,-0.2333946228,0.2026789635,0.1094043851,-0.2765069306,-0.131254971,0.1421453804,-0.3156776428,0.2580564022,0.3230242729,-0.0061197849,0.2491148561,-0.3315944374,-0.0262771808,0.0729192495,-0.3236356378,-0.1923995912,0.0774704143,-0.2175783515,0.0755516961,-0.3103999496,-0.2757736444,-0.1316889673,-0.0882785022,-0.1178445965,-0.0042067799,0.2224297822,-0.3529714048,-0.0780835301,-0.0619572327,-0.2608759701,-0.0054623163,0.4155068099,-0.2048797607,-0.0484172329,0.6307442188,-0.4179269373,0.0933704078,-0.07985349,0.1335898638,-0.1575944424,-0.1162409782,0.0598994382,-0.0821805894,-0.3145850301,-0.0502756275,0.5699495077,0.4229193032,0.1599030793,-0.1802933514,-0.4978502095,-0.302197814,0.0407070331,-0.191613242,0.1929024011,-0.0863239095,0.21488747,0.2381756604,0.0008385406,-0.1775393188,-0.1106198281,-0.0448225141,0.1870532185,-0.1764591783,0.0281169005,0.0877773538,0.1109380201,-0.0514413752,0.0712796301,-0.2663800418,-0.0333771035,-0.1041757911,0.2260269523,0.1543931216,-0.0742285848,0.1739836335,-0.1489111781,0.028481856,0.1557397544,-0.1324018687,-0.0511802137,-0.0304492712,0.1164702177,0.2340570837,0.0443301871,-0.2025707811,0.282455653,-0.0093792528,0.1550862491,-0.0946004093,0.2090190649,-0.3223199844,0.0316015035,0.171443373,0.0984688103,-0.0670006275,0.2769372463,0.1365858912,0.0017015988,0.0024578536,-0.0474307016,0.3074745834,0.7473389506,-0.2027828395,-0.1288805157,0.4639034271,-0.0035370653,-0.2118543237,0.061207965,0.1913787276,0.0043711797,0.0706608295,0.2556616962,0.0453903973,0.079801172,0.253644824,0.0567863509,0.1590507478,-0.0625450611,-0.0556710288,0.516302526,-0.2475844473,0.3452169895,0.344825387,-0.0267718304,0.3172128201,0.0437358692,-0.0222516749,0.5483084321,-0.0056327889,0.007812988,0.1248813346,0.0073669599,-0.1026034653,0.2061543763,0.0983363241,0.0502793305,0.2542821467,0.6443097591,-0.5301635861,-0.1942593455,-0.2695648372,-0.098445408,0.0875171423,-0.1178481355,-0.4535405636,-0.0087994607,-0.3257640004,-0.1410659254,-0.137820825,-0.3443897367,0.2559643388,0.0038795404,-0.3412180543,-0.2516852617,-0.4334020913,0.1374484301,-0.0445603579,-0.3266586959,0.2011279911,0.0334377028,0.0312454775,0.2748129964,0.299069494,0.4600579441,0.0771294385,-0.286952585,-0.3478589952,-0.1940520704,0.0755387917,-0.0769362822,0.296520561,0.0063506509,-0.1687133163,0.5234034061,0.0489704572,0.0155934971,0.2165281177,0.4239832461,-0.4900086522,-0.1489179879,0.7677970529,-0.0795136392,-0.0862946585,-0.0531122461,-0.0124602718,-0.2715530097,0.0650703162,-0.1014520526,-0.1340539902,0.2226326019,0.1248547509,0.0261102933,-0.0593128651,0.2937095463,0.3687492013,-0.1291076392,-0.4022060037,-0.2841154635,-0.9730965495,0.0777431354,0.01921927,-0.3939653337,-0.0716051981,-0.2766572833,0.2171499431,0.1105116978,-0.2178167999,0.2072971314,0.2407176644,0.0459835157,-0.2624928653,-0.3137991726,-0.242788747,0.0363880917,-0.0418955721,-0.4255926311,0.1177315414,-0.2274348587,-0.1400522441,0.289314568,-0.0407455117,0.3404312432,0.0762931705,0.2889729738,-0.0720176473,0.6170152426,0.1039070636,0.0108980788,-0.1175371185,0.2103998661,-0.1169168875,0.0672923476,-0.0797016621,0.2090822309,-0.1803915352,0.6635019779,-0.5068436861,0.4374557734,0.0402261838,0.0813539028,-0.2613565028,0.1021741554,-0.3644091189,0.1315890402,0.3581798375,0.2829104364,-0.1217759997,0.3061517179,-0.137870878,-0.1978137493,0.3982644677,-0.1261218637,-0.0415316448,0.0497804359,0.2727915943,-0.0928295329,-0.0611093417,-0.4452347159,0.0542220511,0.2737287879,0.0292110145,0.1129289195,0.0959343538,0.1567934453,0.1195365116,-0.062240202,0.8743882775,0.0446121134,0.0175980385,0.0011319175,-0.2442411035]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"I think the dataset script works correctly. Just the dummy data structure seems to be wrong. I will soon add more commands that should make the create of the dummy data easier.\r\n\r\nI'd recommend that you won't concentrate too much on the dummy data.\r\nIf you manage to load the dataset correctly via:\r\n\r\n```python \r\n# use local path to qanta\r\nnlp.load_dataset(\".\/datasets\/qanta\")\r\n```\r\n\r\nthen feel free to open a PR and we will look into the dummy data problem together :-) \r\n\r\nAlso please make sure that the Version is in the format 1.0.0 (three numbers separated by two points) - not a date. ","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":102,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nI think the dataset script works correctly. Just the dummy data structure seems to be wrong. I will soon add more commands that should make the create of the dummy data easier.\r\n\r\nI'd recommend that you won't concentrate too much on the dummy data.\r\nIf you manage to load the dataset correctly via:\r\n\r\n```python \r\n# use local path to qanta\r\nnlp.load_dataset(\".\/datasets\/qanta\")\r\n```\r\n\r\nthen feel free to open a PR and we will look into the dummy data problem together :-) \r\n\r\nAlso please make sure that the Version is in the format 1.0.0 (three numbers separated by two points) - not a date. ","embeddings":[-0.1612339467,0.2551890612,0.0169224199,-0.0688923895,-0.018732788,-0.0799422488,0.403591305,0.3387187719,0.0754055381,0.0486796238,0.1928060502,0.1107183471,-0.2993968129,0.1028159931,0.1005925834,-0.1166138947,0.0225548334,0.2291203886,-0.1793874353,0.2209168226,-0.2534815669,0.1003351882,-0.1108291149,0.1783715039,-0.01125177,0.158291623,-0.2870258987,0.05930347,-0.3393084407,-0.7205921412,0.3562000096,0.0720799118,0.2591177523,0.4941640794,-0.0001224414,-0.0743295774,0.5197001696,-0.1447191983,-0.6755980849,-0.2325552404,0.2401564121,-0.3556254506,0.3359572589,0.0014614864,0.0567581132,-0.2257977575,0.2479344457,0.1120189801,0.2297286987,0.3637806773,0.1369117051,0.2928271592,0.0548726954,-0.1232561469,-0.1417387426,-0.011339915,-0.1185133755,0.5441453457,0.34197703,0.0331580415,0.007737142,-0.0545326471,0.0872461647,0.3912862539,0.1923445761,0.0001654503,0.5235515237,-0.0445559323,-0.1164836586,0.0907762498,0.7365930676,-0.5787578225,-0.3857163489,-0.0000173481,0.0516150072,-0.1114275083,0.1721991748,-0.1005596966,-0.2993264794,0.059376657,-0.1098723412,-0.2239533365,-0.1002402231,0.2735372484,-0.2138689458,0.2330448031,0.0602377243,0.2284818441,0.0010024061,-0.1071489081,0.1952328086,-0.1300069392,-0.0561490357,0.2684306204,-0.0915718302,-0.1815170199,-0.3443870246,-0.1629769355,0.1395751983,-0.1011953279,-0.3182562292,-0.1162263155,0.0414253436,0.1830254197,0.1564452052,0.3665716052,0.2711389065,0.2031321079,0.1212837324,0.1413872093,0.2718673944,0.2031423599,-0.0669590905,-0.3533773124,-0.0125099681,0.0931031555,0.2051630765,-0.2799247801,-0.161819607,-0.2108635902,-0.3621450365,0.0115342895,-0.1201581508,0.3297589719,-0.1922219098,-0.1893845946,0.079003334,0.5592823625,-0.2009380907,-0.2160238475,-0.0144221354,0.0249405243,-0.2150385082,-0.0653627217,0.2374851704,0.0095679089,0.4028269351,-0.255792588,-0.2903880179,-0.0060508512,0.2314983308,0.0201615579,-0.1914067864,0.2558790743,-0.0500898287,-0.0247730706,0.1001884639,-0.3326967955,-0.2278572917,0.221253261,0.1008477136,-0.4667082727,-0.2118012309,0.0906530023,-0.4503858685,0.0216746088,0.0170063488,-0.0464019701,0.0985519364,-0.3617475927,-0.0825721323,-0.4229156375,-0.2915905714,-0.1143477708,-0.0574087463,0.3280137777,-0.4315405488,-0.2729130685,-0.4097924531,-0.2997344732,0.1033358723,-0.2614889145,-0.0754129365,0.2706587613,-0.3903686106,-0.107451193,0.5073695183,-0.5006065965,-0.2522059381,0.4450431466,-0.3928951621,-0.1923517138,0.1796453297,0.0358868502,0.0022722199,-0.2983292341,-0.1209911555,0.0788602307,0.0852051452,-0.1336452961,-0.1600969434,-0.3697938621,0.1391663253,0.2701596618,-0.0896331817,0.1255898625,0.0214272775,-0.0110596046,0.3364551961,0.0801436007,0.1863798201,-0.0893134251,0.1101852059,-0.1883333921,-0.0934507847,-0.0114891641,-0.6761333346,0.210529834,-0.2046797872,0.3002503216,-0.0144708166,0.0340374336,-0.6020737886,0.0189521722,-0.0877689868,-0.2432880998,0.0422781669,0.1985485405,0.1348025203,0.0554199256,0.0055906614,0.1368088275,-0.1502519995,0.254986763,-0.0785130784,-0.0289891474,-0.1147481203,0.0913884491,0.30210495,0.2819165289,0.0884689987,-0.3099198937,0.0524374358,0.2828422785,0.2346712202,0.1435471922,0.2016164809,-0.1117907837,0.0856142491,0.2256001234,0.2614989877,0.3159290552,0.1025579721,-0.2118208855,-0.0016611454,0.2494020164,0.245536685,-0.0044797841,-0.2242422402,0.0585732125,0.2959597111,0.1056821421,-0.1060553566,-0.3111675084,0.0934354514,0.3330492675,0.593619585,-0.0991462022,0.1983140111,-0.0352447107,-0.2650436759,-0.3206388056,0.1122196689,0.1166355014,-0.2833050191,-0.1018628478,0.0325119309,0.2716315091,0.801594615,-0.0550957881,0.059599679,0.1137067303,-0.1535864174,-0.194641009,0.1942010671,0.1026120856,-0.0795992538,0.1113403663,-0.0458141193,-0.3477388322,-0.0373355448,-0.0098788086,0.4176739156,0.4067234099,-0.407340765,-0.0569835417,-0.1535446942,-0.3752649724,-0.3857826591,-0.3960626125,-0.2094003856,-0.3791491091,-0.0610079877,0.3446573913,0.0182159357,0.2503041923,-0.0940140411,0.2637229264,-0.2055865973,-0.3096750379,0.1600998193,-0.1387048662,-0.4587657154,0.0301654451,0.242906943,0.1665126234,0.3658002317,-0.2689116597,-0.2122471929,-0.2011979371,-0.3196367919,0.1471571624,-0.0106315445,0.6500738263,0.4507491887,0.0585350841,0.2979269922,0.1112190932,0.0039671566,-0.1363391131,-0.3817054629,-0.0809629411,-0.0211391002,0.0634535402,-0.1135388464,-0.4393889904,-0.5979081392,-0.0280654766,0.0741738155,0.1146289185,0.121243827,0.1300856769,-0.0153652998,0.1810198575,0.0319088437,-0.196292311,-0.1938026994,0.0531015843,0.0823288187,-0.1754688174,-0.2817417681,0.1657281816,-0.2176589519,0.0891386271,-0.2676085532,-0.4637163877,-0.4648181796,0.2448098361,0.163631469,0.3045325875,-0.2045297623,0.2921668589,0.2518639863,0.1732723713,-0.1214886978,-0.3563105166,0.203666985,0.0156517271,0.3223896027,0.1992477626,0.3970181644,-0.2766500711,0.766987741,0.1820705831,-0.0812233463,-0.0390784182,-0.2118773013,0.0991342962,-0.0870854184,-0.0576467067,0.2388328761,-0.1382336915,-0.2020785511,0.3131056726,-0.1702451259,-0.0092434473,-0.1277004629,0.2386337221,-0.1114007533,-0.0631121472,0.0595799163,-0.166543901,0.037804883,0.0986809954,-0.0149557078,-0.261767894,-0.1944641918,-0.0525464118,0.3353534341,0.0021535689,0.1182075217,-0.3188006282,0.0112374481,-0.4265463054,0.2786066234,0.2302430719,0.5973353386,0.1270480454,0.0818643048,0.1029395461,0.1030560136,0.2114652693,-0.0074452725,0.0996928588,0.2119988203,-0.0997580811,-0.2154737562,-0.2416671962,-0.0882357657,0.1525653899,-0.0412069224,0.1243562773,-0.4224524796,-0.062765196,0.2315380275,0.196136564,-0.1854176074,-0.3014689088,-0.1950827241,-0.0132913189,-0.1063403785,-0.2276051193,-0.2369610667,0.0754648373,-0.1983977705,0.4346013665,0.0265482236,-0.0950341821,-0.0094672209,0.2204019874,0.218563959,-0.22556521,0.3471061289,0.2126541734,0.0442229062,0.1937830597,0.2733416259,0.0466998704,-0.269918263,0.0717186704,-0.0559654869,0.3882953823,0.2309399396,0.0786185861,-0.0000542772,-0.2286577523,-0.1555216908,-0.2668571174,0.1135368347,0.3650881946,-0.0120891407,-0.0575675443,-0.323779285,0.5772801638,0.1279290318,-0.2223548591,0.36209777,-0.2282128632,-0.234174639,0.1770774573,0.0581521243,0.8921769261,0.1632307172,0.2606756389,0.4901679754,-0.222150892,0.3511627614,0.2147481591,0.1131199077,-0.652251184,-0.0027107783,0.039679084,-0.1463422328,0.2141655535,-0.1434666812,-0.3366465271,0.3871507943,0.0043895761,0.2776869237,0.2860884368,0.3106579781,-0.1787121445,-0.0270796698,-0.1167604476,0.082982488,-0.0814638957,0.372913152,-0.23302643,-0.0751855373,-0.2692853212,-0.0624626167,-0.1918040961,-0.1346476823,-0.1693446785,-0.0520100556,0.1199265867,-0.110771127,-0.0419672877,0.2312603593,0.2426555008,-0.1447335333,-0.2273568213,0.0562331006,-0.2652635574,0.0754645616,0.3290458024,0.061140541,0.2503130436,-0.2712118328,-0.1735932827,0.1091701463,-0.2485246509,-0.0686342865,-0.1494593322,-0.2002931386,0.1652695835,-0.300275147,-0.345749557,-0.1171413288,-0.0490969196,-0.0921911448,0.0285915788,0.2284182012,-0.3406799138,-0.005800467,-0.1140653789,-0.2673190832,-0.0284326077,0.4541558325,-0.0294729583,-0.0137649635,0.8067979813,-0.0969641432,0.1509829462,-0.1248980835,0.0640550777,-0.0021405877,-0.2087340355,0.0074982024,-0.0727609098,-0.3054572344,0.0633394793,0.6002621651,0.3502784073,-0.0064058686,-0.0158133972,-0.5503163338,-0.4705408514,-0.0174367521,-0.133722797,0.1088274345,-0.0900086984,0.1982869059,0.2599764764,0.0302022845,-0.2092874497,-0.138788417,-0.0238549169,0.1949311793,-0.1560316086,0.0156250093,0.0409127884,0.1092598364,-0.0156606119,0.0169974547,-0.3918665349,-0.0713815168,-0.100626424,0.2022781968,0.1108470634,-0.114285782,0.063302651,-0.1228529289,-0.1114388108,0.1994763911,-0.0942070782,0.0171757322,-0.1065144017,0.1853143871,0.3287239671,0.0513389185,-0.090477623,0.2512061,0.1026214138,0.1897009909,-0.1157941893,0.2186831832,-0.2577763498,0.044404909,0.1239484549,0.1364471316,-0.0643078759,0.3193918467,0.1831649244,-0.1417762339,0.1457554996,-0.1233255044,0.341140002,0.7627295256,-0.259604156,-0.1603022814,0.435028106,0.0420883633,-0.2385641336,0.0053894501,0.2081551254,0.1026714295,0.0534216389,0.2619720995,0.054714784,-0.081309773,0.1798189431,0.1486181021,0.1108655632,-0.1900277734,-0.0886750519,0.5046733022,-0.2913792729,0.2332158387,0.3728398085,0.0424090065,0.1642456055,0.1657638103,0.0522631817,0.4733754992,0.0552747138,0.0517030284,0.1152092963,-0.0260249674,-0.004513477,0.399151206,-0.0094315223,0.008123151,0.3110034466,0.59464252,-0.4087997675,-0.1550912261,-0.3245675266,-0.1048493758,0.0352178626,-0.1329291761,-0.3802722991,0.0711787194,-0.2537857592,-0.1552360207,-0.1535773873,-0.3261702955,0.1191238016,0.0025213526,-0.3347837627,-0.2304136902,-0.4620945156,0.2961378396,-0.1024323106,-0.2105663717,0.2456554919,0.0673215687,0.1390589923,0.3560238481,0.3868747354,0.4162228405,0.1502303779,-0.4203753173,-0.1778589189,-0.1542007327,0.1086269319,0.0784463435,0.348234266,-0.0188990738,-0.188468188,0.5514342189,0.0635804236,0.0415649973,0.2617346048,0.4324657321,-0.4136838615,-0.1912736744,0.7280190587,-0.1477663964,-0.1276976764,-0.1156830862,0.174031958,-0.339132905,0.1881848425,0.0550409295,-0.0119350972,0.201720506,0.0042705783,0.0377386734,-0.1704934984,0.429818511,0.3963538706,-0.0303225517,-0.4015493393,-0.2416230887,-0.8262498379,0.1943665892,-0.0957244039,-0.317833364,-0.0378299914,-0.2311189324,0.1911775619,0.1463651657,-0.2504749596,0.2390635014,-0.0157609005,0.1646005064,-0.3513507843,-0.3869226277,-0.2707547843,0.1131700277,-0.0845247731,-0.4501944482,0.1749817282,-0.2466263175,-0.0976357684,0.1637708396,-0.0195154995,0.4524601698,0.1002552137,0.3120700121,-0.0456242487,0.5762505531,0.1779057831,-0.0407508276,-0.2322151661,0.1957597584,-0.14506042,0.0977167562,0.0491046831,0.1591257304,-0.1298924387,0.5070652962,-0.3636959791,0.4363593161,0.0866974592,0.1201671362,-0.4500047565,0.1737584472,-0.2223962098,0.1384892166,0.2480433285,0.2696770728,-0.1054699868,0.4141568542,-0.1896025091,-0.0589252375,0.4289875925,-0.206317842,-0.1036186963,0.0680237114,0.196911037,-0.1153511927,-0.1730975211,-0.4120732248,0.2724251151,0.2450025827,-0.0622012168,0.1468649805,0.1653281152,0.2617563307,0.0928101614,-0.0122945048,0.8619530797,-0.0430600345,0.1236886606,-0.0738010705,-0.2325073183]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"The script loading seems to work fine so I'll work on getting a PR open after a few sanity checks on the data.\r\n\r\nOn version, we currently have it versioned with YYYY.MM.DD scheme so it would be nice to not change that, but will it cause issues?","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":47,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nThe script loading seems to work fine so I'll work on getting a PR open after a few sanity checks on the data.\r\n\r\nOn version, we currently have it versioned with YYYY.MM.DD scheme so it would be nice to not change that, but will it cause issues?","embeddings":[-0.1767385751,0.2111754417,-0.0094802435,-0.1317930073,0.073443301,-0.1427535415,0.4176616967,0.2573601604,0.0499721952,0.1190561801,0.2233332545,0.069634974,-0.288746506,0.0840551481,0.0976942331,-0.0076815896,0.0639337674,0.1413368583,-0.1044421196,0.2076696455,-0.2553172708,-0.0057773921,-0.0539509095,0.1049623638,-0.0217238683,0.1230244637,-0.2095181495,0.1034018323,-0.436101228,-0.8527143598,0.3802421987,0.1069500297,0.2888356149,0.3949593008,-0.0001284344,-0.0574783869,0.5988144875,-0.1096532643,-0.6252249479,-0.1496018916,0.2785562873,-0.3912249804,0.3590856194,-0.0352212638,0.1500796974,-0.1874664128,0.2187322229,0.0418383554,0.1443508565,0.288277179,0.0764790773,0.2813302279,0.0681955814,-0.2176912427,-0.1155698746,0.0595381968,-0.0826354846,0.5680220127,0.4420571923,0.1645011157,-0.0479074344,-0.1347484142,0.0992561951,0.4558956921,0.1890973747,-0.0451924801,0.6421710253,0.017217312,-0.1417569667,-0.0005506419,0.9479646683,-0.5009502769,-0.3974162638,0.0502950549,-0.0116671054,-0.121365875,0.2301284969,-0.1927763224,-0.2900263965,0.0472511202,-0.1721949726,-0.3062464595,-0.1238460988,0.3003010154,-0.3025889993,0.3052990437,0.1033435538,0.2493603081,0.0188441053,-0.1182857454,0.2843585312,-0.0363516286,-0.0146416137,0.1863449365,0.0785246491,-0.2858282626,-0.3459248543,-0.2041811943,0.1131222174,-0.0412339792,-0.4233128428,-0.1887902319,-0.0349557586,0.176213488,0.2227994055,0.2793657184,0.3226895034,0.1964391321,0.2440313399,0.0767488107,0.3281917572,0.2670004368,-0.0798517466,-0.4030307531,0.0162077155,0.0597614497,0.2114082873,-0.3521575928,-0.0871662945,-0.2609820664,-0.3343007267,-0.0478125885,-0.1376745254,0.2543797791,-0.1169835031,-0.0950978771,0.0304058231,0.4238702059,-0.1529856771,-0.1591299325,-0.0107982727,0.0142800193,-0.1677898616,-0.0182297546,0.2101244032,0.0889372975,0.3157947659,-0.2774038315,-0.2017735839,0.0035693622,-0.015546673,0.0973913372,-0.1956209242,0.2981408238,-0.2475427687,-0.0211074036,0.0438069291,-0.2993895113,-0.2070538849,0.2779168189,0.0965090245,-0.4538351297,-0.0178777836,0.0486825742,-0.5866084099,-0.0063493485,-0.0185742844,-0.1163564175,0.1194530055,-0.3823113441,0.0136968838,-0.4451647103,-0.3795362413,-0.1074555367,-0.1562435031,0.2226886451,-0.3308186829,-0.2761642933,-0.3116811514,-0.3430005908,0.1597062796,-0.2424592674,-0.0967334658,0.0930971727,-0.3562758863,-0.2131888866,0.5165299773,-0.6141238809,-0.3168637156,0.479444325,-0.3933302164,-0.3192726672,0.2487607449,-0.0207883809,0.0985716134,-0.4370949566,-0.1215800643,-0.015840156,-0.0188480616,-0.1695326418,-0.2131860703,-0.343208313,0.1577182412,0.2769614756,-0.0671554729,0.1459979415,0.1615187675,0.135269165,0.3999652565,0.1997321099,0.1501396149,-0.1701677144,0.24233374,-0.2819080353,-0.034803208,-0.0649490952,-0.6823078394,0.2038476914,-0.0831387043,0.1682871282,0.0175625514,-0.0103570186,-0.509771347,-0.0043713702,0.0005460257,-0.1865508854,-0.0790807828,0.1180780455,0.1709702611,-0.060119167,-0.0290490147,0.0088177472,-0.2176761776,0.1949049234,0.0657253489,0.0218320377,-0.0467246063,0.0551206432,0.2055900395,0.2732928991,0.0578795038,-0.3460522294,-0.0076025217,0.2738026381,0.3753958642,0.2467864156,0.1457493305,0.0340562537,0.2007615715,0.2431774288,0.2701938748,0.2410458177,0.0138973389,-0.1885058433,0.0430782996,0.1935187876,0.165984109,-0.0397758484,-0.1623610109,-0.0281060059,0.273868978,0.1649010032,-0.1608117968,-0.3247186542,0.0163578633,0.321111232,0.6261168718,0.0213598311,0.2209395766,-0.1034023091,-0.2408190966,-0.368386507,-0.035853982,0.1324319243,-0.1466745883,-0.1312391907,0.0345725529,0.2909605801,0.8139572144,-0.0582411364,0.1461983919,0.0932113752,-0.086801596,-0.2109648883,0.1355714053,0.1335216016,0.0800888911,0.1352105886,0.0244504642,-0.3316611648,-0.0123217488,0.0506467633,0.4254483879,0.3760308027,-0.3850040436,-0.1450611502,-0.1898256391,-0.4163719714,-0.4648425579,-0.260025382,-0.2401769608,-0.3224987388,-0.1134426072,0.3259061873,0.0177899059,0.2050327212,-0.1692858189,0.2984496057,-0.2949590385,-0.2706933022,0.1276618838,-0.2065273076,-0.3765336573,-0.0059781531,0.2203705907,-0.0590680018,0.4229106009,-0.155665502,-0.3003937006,-0.1599166989,-0.3217740059,0.143232435,0.1649394631,0.5009924769,0.4258591533,-0.0190750472,0.4650248885,0.1807303876,0.0294762701,-0.1994131953,-0.4520103931,-0.2094367743,-0.0139306746,0.0920100883,-0.0975565389,-0.3627032936,-0.6268529296,-0.0494347513,0.0924643278,0.0464919582,0.086845383,0.07008598,-0.1743046492,0.1262971163,0.064890869,-0.1463151574,-0.139249593,0.1288548112,-0.0637990981,-0.1144398302,-0.2529369295,0.087097913,-0.0972301587,-0.0221045632,-0.0692762062,-0.4163517654,-0.4980790913,0.1860776693,0.1202047467,0.3934659064,-0.3054569364,0.351828903,0.2290000319,0.1961491853,-0.072484605,-0.3492940366,0.218824923,0.117685616,0.2108539343,0.3140124083,0.396840781,-0.1700614244,0.694856286,0.1315309554,-0.1743964106,-0.0818143189,-0.1846346408,0.0265189037,-0.071657151,0.0252696108,0.2892469764,-0.1218499541,-0.3842003942,0.2331876308,-0.1759000868,-0.0446932092,-0.1819266826,0.3031354547,0.0455004945,-0.0235685483,-0.0176596977,-0.1874298006,0.0621086098,0.0636762083,-0.0382253192,-0.169171378,-0.1928943247,0.0061616171,0.3494343758,0.0028033648,0.201798588,-0.2370390147,0.0679618791,-0.3249550164,0.3246842921,0.3511456549,0.6161112189,0.1235387251,0.0064037275,0.1294781119,0.0452519655,0.1685182154,0.0158663038,0.2025444657,0.19419007,-0.1533940881,-0.1033152565,-0.2794733644,-0.055624526,0.107362859,-0.0144901657,0.2064722776,-0.373070091,-0.1632523537,0.2161459476,0.1357463151,-0.1004923731,-0.3330310881,-0.0852660686,-0.0171893686,-0.0329428948,-0.287905246,-0.2201474011,-0.0035187057,-0.1704875082,0.3609251976,-0.0098711252,-0.1548066586,-0.0450737625,0.1220652387,0.1488846838,-0.2825213969,0.3064956069,0.2818534076,0.109121047,0.4292038381,0.2410032749,-0.0791716352,-0.2065414935,0.0624480732,-0.0080296807,0.3869711757,0.3669582307,0.1010065153,-0.0404612459,-0.1914743334,-0.1860705316,-0.3599002659,0.104932785,0.3145605922,-0.0783961639,-0.0809592828,-0.3749858439,0.6471113563,0.1889345348,-0.2820150256,0.4349669218,-0.2932124138,-0.2415936589,0.016749328,0.1147200093,0.9402288795,0.1011528671,0.1793756932,0.407402128,-0.256912291,0.2777275145,0.2090794742,0.0930837691,-0.5795202255,0.0369095914,0.006616855,-0.1812874377,0.2285750955,-0.1084260941,-0.1412522048,0.3231203556,0.0675940216,0.344671309,0.2996440828,0.3267160356,-0.1354597062,0.0387337022,-0.1763438731,0.0148163596,-0.1632350683,0.3315987885,-0.2451390177,0.0207077116,-0.2024508268,-0.0634012967,-0.2631101012,-0.2040712088,-0.145793885,0.029359702,0.0580561087,-0.1723593473,-0.0035784678,0.2685664892,0.1853124052,-0.2835444212,-0.1179847494,0.0657290071,-0.1478354782,0.124173075,0.3541330695,0.0912996978,0.3243589699,-0.3461557925,-0.041703824,0.0784574673,-0.3147274554,-0.1103460863,-0.1682372987,-0.1697039455,0.1779033691,-0.3520860076,-0.2943040431,-0.1413059533,0.0285021663,-0.0528791882,-0.0212047677,0.268715173,-0.2445107549,-0.0112315835,-0.1383842081,-0.2438734472,0.0144760571,0.4472572803,0.0349842608,-0.1098534912,0.6268347502,-0.2000739574,0.0783017427,-0.0809643418,0.0423324928,-0.0304042976,-0.1384970844,0.0794522688,-0.0652863234,-0.309253484,0.0309992395,0.6317122579,0.4538192451,-0.0222006775,-0.0184409525,-0.4137720466,-0.4684455991,0.0602294505,-0.1337691993,0.1552896202,-0.1445984244,0.1882132292,0.106248416,-0.0461607464,-0.148394689,-0.1558260173,0.1758184135,0.2297770977,-0.1876090467,0.0470787585,-0.0200724117,0.0423388928,-0.0639493987,-0.004503449,-0.3734774888,-0.0004291504,-0.063051872,0.2435456812,0.1059922799,-0.1003908366,0.1039548814,-0.1913582832,-0.0130948108,0.2529708445,-0.0426039174,0.0964034796,-0.0803560242,0.1069929972,0.344679445,0.1548197567,-0.008971517,0.3556565046,0.0866757259,0.3026209176,-0.0783770382,0.3181912601,-0.2581302524,0.0790608302,0.1899051517,0.1577679068,-0.0170426518,0.2019144446,0.1141490489,-0.1714568436,0.1450867951,-0.0436180197,0.4042550027,0.8169980049,-0.2331590652,-0.2094206512,0.3954935074,-0.0064650001,-0.2016218603,0.0467281453,0.2742975056,0.1663990319,-0.0245610662,0.3101134896,0.0983721912,-0.0165255927,0.1389919221,0.2003036588,0.1752066463,-0.2545608878,-0.0143157793,0.4361817837,-0.1911935061,0.2348254323,0.3925985992,0.1005083919,0.194804132,0.09433534,0.0811318159,0.5245569944,0.1252108663,0.103516221,0.1761603951,-0.0139163779,0.0260474272,0.4011744261,-0.0035628832,-0.0460843332,0.4040578902,0.6079488397,-0.3851246834,-0.2014325261,-0.2968819439,-0.1906679571,0.0513568036,-0.1312872171,-0.2509272397,-0.0213656723,-0.1806859523,-0.0604484156,-0.2423968017,-0.2871671021,0.2314747274,-0.0044749221,-0.2735227048,-0.2810953557,-0.3688079119,0.1331094503,-0.029771274,-0.2417410165,0.289027065,-0.0109147131,0.0470148884,0.3937853873,0.2517336309,0.4028508663,0.1396003664,-0.3574990332,-0.2865712643,-0.206185773,0.0576338321,0.0451568365,0.3485940099,-0.0030628,-0.1303939372,0.4741200507,0.0290321,0.1062214002,0.2274840325,0.3550856113,-0.4659959674,-0.1002222747,0.8348633051,-0.083296515,-0.2012382895,0.0472789109,0.1799412966,-0.2012376487,0.0493576452,-0.0665802658,0.0086866915,0.1635265201,0.1432312429,-0.0007181851,-0.0985775441,0.2886527479,0.3661210835,0.0254539419,-0.284355104,-0.2156102359,-0.7466059923,0.1942778081,-0.0566345602,-0.3099111021,-0.0215904377,-0.2513841391,0.2322980911,0.0619805604,-0.1773173958,0.1805739552,0.0964598656,0.1078421772,-0.1980058402,-0.2759743631,-0.2155529112,0.0227827728,-0.1038480029,-0.4249195755,0.1188997105,-0.3571048975,-0.1592427641,0.0888787955,0.0176619142,0.4647809267,0.0055893078,0.2521135211,-0.0838739276,0.7441824079,0.2783362269,-0.0673998818,-0.1677696705,0.1265472025,-0.1436953694,0.1740306616,-0.1216218621,0.0844411701,-0.1829200387,0.5333905816,-0.4123538733,0.4977735877,-0.1279706806,0.0690186024,-0.3403318822,0.059882354,-0.263789326,0.0313907415,0.3083661795,0.1954005957,-0.0378026851,0.3024753332,-0.2264569104,-0.1237143502,0.4578659534,-0.1712019295,-0.0323081091,0.0850580186,0.2551979721,0.0569543019,-0.1403321326,-0.4825669229,0.1895668507,0.2474295944,-0.0960592479,0.1211329475,0.1639314592,0.2560333014,0.0140033476,-0.0236373562,1.0229690075,-0.1114954725,0.0829375386,-0.1492390633,-0.2332779467]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"> The script loading seems to work fine so I'll work on getting a PR open after a few sanity checks on the data.\r\n> \r\n> On version, we currently have it versioned with YYYY.MM.DD scheme so it would be nice to not change that, but will it cause issues?\r\n\r\nIt would cause issues for sure for the tests....not sure if it would also cause issues otherwise.\r\n\r\nI would prefer to keep the same version style as we have for other models. You could for example simply add version 1.0.0 and add a comment with the date you currently use for the versioning.\r\n\r\n What is your opinion regarding the version here @lhoestq @mariamabarham @thomwolf ? ","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":115,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\n> The script loading seems to work fine so I'll work on getting a PR open after a few sanity checks on the data.\r\n> \r\n> On version, we currently have it versioned with YYYY.MM.DD scheme so it would be nice to not change that, but will it cause issues?\r\n\r\nIt would cause issues for sure for the tests....not sure if it would also cause issues otherwise.\r\n\r\nI would prefer to keep the same version style as we have for other models. You could for example simply add version 1.0.0 and add a comment with the date you currently use for the versioning.\r\n\r\n What is your opinion regarding the version here @lhoestq @mariamabarham @thomwolf ? ","embeddings":[-0.0784065276,0.2176980078,-0.007995056,-0.2187167704,-0.0167497806,-0.150138393,0.4548567832,0.3005523384,0.0172307827,0.0683669299,0.2465083301,0.0794769451,-0.3443652093,0.0902385265,0.0619925447,0.0040574609,0.1139812395,0.08718732,-0.0638602227,0.1532087475,-0.2385701835,-0.0466873348,0.0585397147,0.1851983666,0.0103353141,0.1138955727,-0.2475723326,0.0872460753,-0.4137390256,-0.8389000893,0.3315024674,0.1653555185,0.3825393319,0.4212989211,-0.000126961,-0.0945958197,0.6248885393,-0.0205539875,-0.6024017334,-0.1692791432,0.3652236462,-0.3112033904,0.4110463262,0.0247803349,0.1619848311,-0.202710256,0.1706410348,0.1620692313,0.1313155591,0.1334400028,0.0678538755,0.221517399,0.0524370447,-0.1449633092,-0.199095726,0.0519704409,-0.065805234,0.6167839766,0.5382683277,0.1772143394,-0.0214810427,-0.0021334172,0.1414722949,0.3280438781,0.2071106285,-0.0733768344,0.7348244786,0.0613605045,-0.1309859008,0.008156308,0.9527406096,-0.4113693833,-0.5182249546,0.0556686558,0.0488390252,-0.1310151368,0.1828904897,-0.1695913076,-0.197383821,0.1067790389,-0.181289494,-0.3202574551,-0.2214025706,0.281229347,-0.2812559009,0.2253086269,0.0833764374,0.2423145622,-0.1092570871,-0.1724252552,0.2602316141,-0.0996678397,-0.0187064912,0.1975751519,0.1821485162,-0.3591315746,-0.3890568614,-0.110869959,0.1458980888,-0.109128885,-0.3968161643,-0.2765681744,-0.0512405969,0.0908384994,0.2513261139,0.224924013,0.4186557233,0.1173015162,0.2379641384,0.028204415,0.3856794238,0.2755427063,-0.0382748879,-0.3110348284,0.0303875543,0.0722006261,0.1840481311,-0.4060976803,0.0470503718,-0.2323033363,-0.3417426944,-0.1388191432,-0.1190886945,0.1863187551,-0.1159295738,0.0003319214,-0.007227229,0.3783921897,-0.1264192164,-0.1867788434,-0.0292388052,-0.0574388653,-0.1304681748,-0.0203393735,0.1908874214,0.0325555168,0.2897745073,-0.2891440988,-0.1841714382,0.0204521529,-0.0876472443,0.124909386,-0.2652384937,0.2683077157,-0.4012267292,-0.0153727951,-0.0719234869,-0.3612963557,-0.1501754522,0.288268894,0.0795195475,-0.434409827,-0.0159452744,0.0527834632,-0.4855265021,-0.0122479582,-0.0292706341,-0.0582246967,0.1487240195,-0.3259877264,0.0435404964,-0.4276154339,-0.3264013231,-0.1250606626,-0.2064811736,0.2373211086,-0.377964586,-0.246578306,-0.2162880152,-0.3262242377,0.051763501,-0.2061721832,-0.0900932252,0.0343936421,-0.1435723901,-0.2408736795,0.5255958438,-0.6463606358,-0.2875416875,0.5019090772,-0.2739059329,-0.275549978,0.2525528073,-0.0246022884,0.1583237648,-0.5408318639,-0.1535225958,-0.1046664491,-0.0388403796,-0.1779832244,-0.3315536678,-0.4214683473,0.2005831003,0.3345440626,-0.1466783583,0.1094430238,0.0924078524,0.0771398544,0.503700912,0.1834124029,0.051097393,-0.2177707404,0.3656322658,-0.2329492867,-0.0781683475,-0.0145330513,-0.7206129432,0.1898140907,-0.1022906527,0.09577097,0.1370865256,0.0051964805,-0.435996592,-0.0237317048,-0.0138875563,-0.2627038956,-0.0815754607,-0.061651174,0.0520687103,-0.0294860303,-0.0105231004,0.022948971,-0.202406019,0.2221765965,0.085792236,-0.0024018316,-0.0921844542,0.047818061,0.1557611972,0.2939287722,0.0750231817,-0.3025203049,-0.0203419719,0.320471853,0.3806015551,0.2452131063,0.1120913252,0.1005762294,0.2997726798,0.3446676731,0.3244914114,0.1484442353,-0.0438332222,-0.0966134816,-0.0020459061,0.1990581751,0.1993008703,-0.1508339643,-0.2492024451,-0.0845803246,0.1860621423,0.2040824741,-0.229447633,-0.2984946966,-0.0223975703,0.3367255926,0.5842097998,0.0103343045,0.1890866011,-0.0359901078,-0.2254911661,-0.3813836277,-0.1191439852,0.1591709703,-0.1531739682,-0.1768587083,0.0373908207,0.2746086419,0.6558873653,-0.0508092195,0.0616934188,0.079688549,-0.0912752002,-0.1582047343,0.1329773217,0.195186466,0.0797488391,0.0744832456,0.0270143524,-0.2232914865,0.0546599478,0.074182868,0.2964885831,0.3012321889,-0.3033635616,-0.1087811887,-0.1436949223,-0.3550186455,-0.4484522045,-0.2716024816,-0.361220181,-0.2839112282,-0.0444087945,0.3055320382,-0.0080493754,0.1616134197,-0.0884679705,0.3513897955,-0.3062573075,-0.0802176818,0.130917117,-0.1687153429,-0.402138859,-0.0075454758,0.1693283617,-0.1414070725,0.3990307748,-0.0735815465,-0.3104703426,-0.073119387,-0.4179688096,0.0762952268,0.136833787,0.568072319,0.4437888563,-0.0158590879,0.4559708834,0.2141160369,0.0265395232,-0.1878950596,-0.3747596741,-0.2456979305,0.0024519786,0.0622133464,-0.0628887266,-0.3411558867,-0.6747949123,0.0159527473,0.1417423785,0.0415407754,0.0917668045,0.0913704336,-0.2075153291,0.0098946961,-0.0334914811,-0.1260697246,-0.0970021933,0.1426061243,-0.1185902357,-0.1148613021,-0.212409392,0.0941465646,0.0500380322,-0.0497131981,-0.1521892548,-0.3812296987,-0.5387281179,0.2752443552,0.0959388912,0.4432166219,-0.3750834763,0.3168643713,0.2574598491,0.2106015086,-0.0305581577,-0.3124099076,0.2227205187,0.2024612278,0.2387276441,0.3416226506,0.4278680086,-0.171560064,0.6987609267,0.2607888579,-0.2120236456,-0.2311721742,-0.0054807258,-0.1285502464,-0.0936508253,0.0392338559,0.2602946758,-0.1145472825,-0.3659272492,0.283010602,-0.0718556121,0.000910308,-0.1715389639,0.2326366156,0.0482660942,-0.0258478206,-0.0169917885,-0.1041960493,0.0901826099,0.0361693613,-0.0619631521,-0.1635523289,-0.2494882196,0.0705663338,0.451167345,-0.0071659945,0.2389381379,-0.1418856531,0.1438901573,-0.3231490552,0.2919078171,0.3646897674,0.5911159515,0.2054839879,-0.0649881214,0.1646294892,0.0394922979,0.1643059254,0.0681182221,0.2253869325,0.0840039626,-0.2383330166,-0.080511719,-0.3215266466,0.0013930256,0.0430134125,0.0670979172,0.2078156173,-0.3719227016,-0.225464493,0.1916741133,0.1168142781,-0.1499802917,-0.3239277899,-0.0455497131,0.1384336948,0.0778777003,-0.3222942054,-0.3112131059,-0.0988619179,-0.2305609882,0.355545193,-0.0032921985,-0.1536962837,-0.0603078716,0.1760216951,0.1231407449,-0.3502573669,0.323148638,0.3036520779,0.1975912899,0.4199138284,0.1198775098,-0.1202519014,-0.1092016101,-0.0389911532,-0.0917914137,0.3305354714,0.4466269314,0.0294082426,-0.0755067021,-0.1519654691,-0.2144212127,-0.369579643,-0.0305054374,0.2852317095,-0.0497960113,-0.0676759407,-0.3444427252,0.6544519067,0.1329582185,-0.353995204,0.3894945681,-0.2603273988,-0.2049239129,0.0144690285,0.1230352148,0.9271790981,0.1357528567,0.2261446863,0.3561149836,-0.2628798485,0.191061303,0.207308352,-0.0211730618,-0.5951221585,0.0795613304,-0.010455111,-0.1193761155,0.2597505748,-0.068187125,-0.0672565848,0.3064362109,0.1151743382,0.3037268221,0.2747290432,0.3166321516,-0.0745716244,0.1246765926,-0.1457638741,-0.0027337288,-0.1987129152,0.2215567678,-0.2173818201,-0.0591210909,-0.1345620304,-0.1353386045,-0.1971942186,-0.2788552642,-0.141729638,0.0025871682,0.0993950889,-0.2422435284,0.098872222,0.2078379393,0.2124312073,-0.2769320607,-0.1058308184,-0.006890195,-0.0941553563,0.1918313503,0.2326048017,0.1145533174,0.4135675728,-0.3656519949,-0.0298331175,-0.0021501007,-0.3015243113,-0.0788200796,-0.2447600216,-0.2292693555,0.2069189399,-0.3499625623,-0.2433496118,-0.0717149153,0.0609528013,-0.0496355183,-0.0165895112,0.2987670004,-0.1676464826,0.002461198,-0.0652825981,-0.1483661979,0.0035220273,0.4236818254,0.0353416502,-0.1286203414,0.5450736284,-0.1776935756,-0.0122192102,-0.0652453825,0.042219758,-0.0569341667,-0.2640587389,0.0248498525,-0.0826938972,-0.2541989982,0.0266472138,0.6445194483,0.5140348673,-0.0881612226,0.0085108485,-0.3115551174,-0.4337587357,0.1188798174,-0.1496616006,0.1279402226,-0.2016358823,0.1935626417,0.0896285027,-0.0807940438,-0.164430663,-0.1689248979,0.2872902155,0.1661549211,-0.207466796,0.0557495691,-0.0566674098,0.0003578013,-0.0773194209,-0.0914462358,-0.4320655465,-0.009923242,-0.0681832209,0.2238519639,0.0970201194,-0.0993718207,0.0750118792,-0.1820569187,0.0329624228,0.333746314,-0.0292700659,0.0577458441,0.0408788361,0.2147264332,0.3124193847,0.2434008867,-0.0267592873,0.4136707485,0.0528822802,0.2778565288,-0.0312035643,0.3052886128,-0.2380787432,0.0376361199,0.0707495809,0.1923795342,-0.1881693006,0.3230146468,0.0442904495,-0.1993772537,0.1765965521,0.0082335752,0.3879286051,0.7847018242,-0.1964469105,-0.3009437025,0.4536357522,0.0278083384,-0.1245294362,0.0831708461,0.2623588443,0.192595154,-0.0484889224,0.3369885981,0.142313689,0.0378630012,0.1024456844,0.2327052057,0.1155837104,-0.3071411848,-0.0141030094,0.4850961864,-0.1911139786,0.2585493922,0.4261291623,0.2196157277,0.0425951369,0.0173609033,0.0899879485,0.4031569362,0.2000469714,0.0417233929,0.2126084417,0.0728520229,-0.0520071983,0.3586238027,0.0100920359,-0.0732287839,0.4328912199,0.6263275146,-0.2457696199,-0.2735121846,-0.2652320862,-0.2573295236,0.1176027581,-0.1507389396,-0.3253526688,-0.0313863792,-0.1723254621,-0.0664226934,-0.3928255737,-0.2323384732,0.2464115471,0.0609279163,-0.2492012382,-0.352198869,-0.4377792478,0.0422928035,0.0669203922,-0.1486774981,0.2484370768,0.0039673005,0.0939529538,0.4052653313,0.2469263375,0.2404518723,0.0969857872,-0.3979862034,-0.3323000073,-0.2425734848,0.076023519,0.1692405492,0.2883461714,0.0202966575,-0.1690831482,0.4392830729,0.0049516847,0.1724260002,0.3671842813,0.3042013645,-0.4158707559,-0.1131071001,0.8669896126,-0.0103636039,-0.2329439521,0.1480515748,0.1992991716,-0.2076277882,0.1263024211,-0.0411046259,0.0579413511,0.1284018308,0.1566936225,0.0126448208,-0.0319884308,0.1883247346,0.3991008103,-0.0195389893,-0.236924395,-0.1239281148,-0.5654705763,0.0769627914,0.0513411053,-0.2140652537,-0.0187728386,-0.2449738681,0.2046446204,0.0657124892,-0.1154638007,0.0221598428,-0.0508057587,0.1486472338,-0.1791858524,-0.1962109953,-0.1313890517,0.1199856028,-0.1198876351,-0.304007858,0.0513783321,-0.3177831471,-0.1912519634,0.0181918908,0.047366187,0.4646494389,0.0385754667,0.1306618452,-0.0679712445,0.5171782374,0.3847463131,-0.1775899678,-0.1369657964,0.1095674336,-0.1280329227,0.1817882657,-0.1369916797,0.1513926387,-0.122855179,0.4253528416,-0.3331277966,0.418358475,-0.2275635749,0.0001634949,-0.4170820713,0.1047742665,-0.1155537814,-0.1177250594,0.2918638587,0.2757244706,-0.0717188194,0.2615314126,-0.2811012864,-0.074364908,0.4302862287,-0.2299857438,0.0475661568,0.0962654427,0.2981061637,0.1159942299,-0.0538569354,-0.5310581923,0.1282913387,0.2365075201,-0.1469854116,0.0226728115,0.2222896665,0.3116584718,-0.0690558776,-0.0391331799,1.0794568062,-0.173514232,0.1126835793,-0.1564429253,-0.1975001395]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"Maybe use the YYYY.MM.DD as the config name ? That's what we are doing for wikipedia","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":16,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nMaybe use the YYYY.MM.DD as the config name ? That's what we are doing for wikipedia","embeddings":[-0.14118056,0.2805646956,-0.0102292299,-0.1734388322,0.0996211544,-0.056226477,0.3951387703,0.3567785025,0.1325159967,0.1633097976,0.2544322908,0.116898261,-0.263643682,-0.0322963297,0.088267535,-0.0225812048,0.0243194513,0.1539965272,-0.0692435578,0.2538096011,-0.1577265114,0.0276444573,-0.076210849,0.0034683063,-0.0320668034,0.1495646238,-0.1243555024,-0.02923223,-0.3976539373,-0.8835687637,0.3851564825,0.0803036019,0.1571981609,0.4587191641,-0.0001239302,-0.0921837762,0.6603642106,-0.0954594538,-0.712616384,-0.1102857441,0.2140960246,-0.4431758225,0.3148986101,-0.1227040365,0.0747777,-0.1257148683,0.2423224896,0.0029505764,0.0423441492,0.2779585421,0.1137095019,0.2098547369,0.1447016746,-0.2050856799,-0.0632630885,0.0726819262,-0.1255190372,0.5561836958,0.2988155484,0.0897956714,-0.0114447912,-0.0887281001,0.085425891,0.4745651782,0.2024761587,-0.0256978609,0.5517067313,0.010823681,-0.1635490507,-0.0021501598,0.9606093168,-0.5900264382,-0.2856666446,0.055982098,0.0618151985,-0.0441574343,0.2293256074,-0.1933582425,-0.3710887432,0.0055937292,-0.1138648912,-0.3460586667,-0.1390753537,0.386338383,-0.2816330492,0.3920342922,0.1204347461,0.2018029094,0.0296665747,-0.1227106974,0.1766296327,-0.0373546295,0.0529412478,0.2122679502,0.0531617999,-0.2913708091,-0.3612178564,-0.2329002172,0.1176927611,-0.0716995001,-0.5114038587,-0.2140529752,0.0601356998,0.2136065364,0.1194375455,0.2886677682,0.3164390028,0.2632665634,0.2690865695,0.1591008455,0.2307039499,0.2268164903,-0.0012361812,-0.3905970752,0.007877118,0.0435170084,0.2372446805,-0.2792246342,-0.0018139748,-0.2233662307,-0.3548255265,-0.0376896523,-0.1549624354,0.2712012231,-0.2122434676,-0.2190591544,0.0354916044,0.3669433594,-0.1903371811,-0.0829747617,-0.0300624408,0.1026915759,-0.2094676048,0.0413397625,0.2736006379,0.1198736206,0.3815645874,-0.221165508,-0.1752733141,-0.0626638979,0.0488902032,0.1650394201,-0.2584552169,0.3138112128,-0.1754798591,0.0914820582,0.0757019594,-0.398679167,-0.2606751025,0.1905807257,0.0879774913,-0.3984540999,-0.0753719285,0.0986795574,-0.5322725773,-0.0472827367,0.0367800593,-0.1087662131,0.1694989651,-0.4394990802,0.0399811491,-0.3296199441,-0.3275543153,-0.1206928715,-0.0447751544,0.2245151997,-0.3344630599,-0.3448539674,-0.3996011317,-0.2620831728,0.1910858899,-0.2722562551,-0.0516153537,0.1798928976,-0.344676137,-0.0936113968,0.5451645255,-0.5609610677,-0.3666338623,0.4103248119,-0.3971425295,-0.3093581796,0.2160005718,-0.0019485553,0.0721393675,-0.295432657,-0.1350308359,0.0266989823,0.054789938,-0.1295778751,-0.1770810634,-0.3129167855,0.1571338028,0.2404756695,-0.0577673204,0.1276278049,0.1767682582,0.2212318033,0.3395685852,0.121658124,0.1946088821,-0.0848003924,0.2039909363,-0.1755814403,-0.0735082105,-0.0987343937,-0.6286855936,0.2173955292,-0.0404789001,0.2884503901,-0.0710285231,-0.0512765013,-0.6030596495,-0.028186148,0.0071453885,-0.2543840706,0.0009686919,0.2109040022,0.1257520169,0.0718231425,0.0874273703,-0.0562531427,-0.2094862014,0.1899579614,-0.0399362333,-0.0160773695,-0.1167456508,0.1136765182,0.1223235205,0.2010829896,0.0323860869,-0.2665897608,-0.0113995541,0.1885953248,0.3017372787,0.2669649422,0.2012650073,-0.0286410134,0.1848205328,0.143307209,0.2356161028,0.3190945387,0.0731490329,-0.2795488834,0.0007631116,0.138755396,0.2109912932,0.0036716217,-0.2251781821,-0.0080106547,0.3341830671,0.1293625087,-0.129444167,-0.3318291605,0.0935377553,0.3465369642,0.5930159092,-0.0306564644,0.3187490702,-0.1770489216,-0.1961135566,-0.3319160044,0.0105721857,0.1472771913,-0.1484428495,-0.1327450722,0.0047822339,0.2607828677,0.7473016381,-0.017134035,0.2102444768,0.0828128532,-0.1943518817,-0.1811460108,0.1961622536,0.1460686028,0.0584960617,0.1173257157,0.0083537688,-0.3501943946,-0.0421719179,-0.0106375404,0.461648345,0.4212019444,-0.3261991143,-0.1254191548,-0.20866476,-0.4111089408,-0.4504458606,-0.4019859433,-0.3150023818,-0.3542329967,-0.1066545695,0.2702236176,0.0867956728,0.2829833031,-0.2247167528,0.3037927151,-0.256911844,-0.3586894274,0.0704028532,-0.2922025025,-0.3351594508,0.0113521395,0.2335169166,0.0032293529,0.4592936039,-0.1246683151,-0.3249701262,-0.2268050313,-0.3663176894,0.1080345362,0.0643457845,0.5244490504,0.4165665507,0.0403012782,0.4771682024,0.0490936302,0.0349766128,-0.1871902347,-0.4240949452,-0.2361034602,-0.0126340063,0.1769871861,-0.092719838,-0.3785631955,-0.5373612642,-0.1031319723,0.0880439878,0.0672019348,0.0066941795,0.1528783739,-0.1076127812,0.1514076442,0.1071259081,-0.0557063259,-0.1339660436,0.1255902946,0.0207218658,-0.1688687354,-0.1097411364,0.1061746031,-0.1115940139,0.0373692065,-0.1364319921,-0.4577694535,-0.4049997628,0.1082664281,0.1651178598,0.4095954299,-0.2527129948,0.294341445,0.2326197177,0.1478602141,-0.0971041173,-0.3855850399,0.1825864464,0.0358474478,0.2548633814,0.2382111251,0.4276255369,-0.1880529225,0.8638376594,0.1257771254,-0.0933354497,-0.0604630746,-0.220559746,0.0037176982,-0.0592096448,-0.0444653966,0.3229119778,-0.1166205853,-0.3214450181,0.2759305239,-0.2046300471,-0.1540329158,-0.1347916275,0.2869546413,-0.0061089247,-0.0675118268,-0.0446652137,-0.1488482952,0.0453065373,0.0540636368,-0.0495025031,-0.2289303243,-0.2520631254,0.0469618738,0.3166661561,-0.1577635854,0.1885858774,-0.2466304898,0.0316434838,-0.4474693239,0.294185698,0.333185941,0.5381495357,0.102789335,-0.0153011773,0.1133804172,0.1049145013,0.2085276246,-0.0177592486,0.1344524175,0.2412903458,-0.0960642919,-0.124616392,-0.2597484291,-0.0087198196,0.1897026598,-0.0042952993,0.134266898,-0.4184399545,-0.0704549849,0.2634961605,0.1959713846,-0.0976237059,-0.3244104683,-0.0867622122,-0.0282972548,-0.0801302567,-0.2394549996,-0.1685003489,0.101182349,-0.0758366883,0.362282604,0.0368489064,-0.0670245886,-0.0823661238,0.1554445177,0.1467124522,-0.2071237415,0.264832139,0.2287517488,0.0274729338,0.3402643502,0.2413718402,-0.1182683557,-0.2586965263,-0.0475783423,0.0072170664,0.3554090858,0.2310941964,0.1883851588,-0.0193964634,-0.1601124853,-0.1480425447,-0.349137336,0.1983934492,0.3260372579,-0.0945455655,-0.0815766975,-0.3744671643,0.6699296236,0.2106671482,-0.2216611505,0.5757380724,-0.2158820778,-0.3613097966,-0.0209628139,0.0972931534,0.9390996695,0.1230465174,0.2218455374,0.3271255493,-0.2036059499,0.4068246782,0.1139772385,0.1631513536,-0.5522786379,0.0355900973,0.0416921675,-0.180774942,0.1868134439,-0.1286399215,-0.2766340375,0.3458251655,0.0909400731,0.3162331581,0.28049317,0.3736039102,-0.0628255829,-0.0473872833,-0.1677778065,0.0675245151,-0.1731278002,0.3932382464,-0.2501639724,0.0991219506,-0.1918776184,-0.0384524092,-0.3120805323,-0.1937751174,-0.1224613935,0.1201615632,0.0544199795,-0.1686798632,-0.0250516869,0.2270249277,0.1252934039,-0.1694914103,-0.1770975143,0.1419795752,-0.1813607663,0.0641489998,0.3031472564,0.0513926744,0.302105993,-0.3474114835,-0.0908854902,0.1219640896,-0.2841303349,-0.2052353472,-0.1263772398,-0.1708372831,0.0842736363,-0.3267584145,-0.2828609049,-0.1330208927,0.046896778,-0.1448683888,0.0261616129,0.2255216241,-0.3177241385,0.0012350624,-0.0340485647,-0.3448729217,0.0183053538,0.4467866421,-0.0853858441,-0.0789619982,0.7237137556,-0.1690612137,0.0381769463,-0.1186438054,0.0230471399,-0.1872940511,-0.1536633372,0.0496012606,-0.0515940562,-0.2729286253,0.0106120696,0.5785095692,0.4408791065,0.045046214,-0.0207257774,-0.4588910341,-0.4163409472,0.0077948477,-0.1810985655,0.2008613348,-0.0820304081,0.1893222779,0.1570998281,0.0159545057,-0.1874679476,-0.1633620113,0.0402951799,0.2436018884,-0.1337099373,0.0225273203,-0.0700732395,0.0874569342,0.001811015,-0.0883047134,-0.2857427895,-0.0525176972,-0.1365345567,0.2203616947,0.0687349066,-0.0029467314,0.0809546039,-0.1938171238,-0.0055879429,0.1685533673,-0.0551455468,0.1077399552,-0.0969745964,0.1997775882,0.3340001702,0.0480222926,-0.0194162242,0.2950715125,0.119937472,0.2520236671,-0.0479513444,0.3378598988,-0.1911024898,0.0625755712,0.1788640916,0.2028529793,-0.086178124,0.1728704125,0.0845031589,-0.1091340706,0.1038738489,0.0455032364,0.3643860817,0.7686905861,-0.2697173357,-0.1473634839,0.433665961,0.0431382433,-0.2070350647,-0.0615150221,0.3559984565,0.0967504755,0.0595653653,0.242922917,-0.0034863208,-0.0197570827,0.1555714309,0.1688243002,0.0833289549,-0.2238809019,-0.0141823161,0.3811250329,-0.2043302953,0.161302641,0.3652935326,0.1226838976,0.2558839917,0.1170235947,0.1278586686,0.4928233922,0.066528514,0.0349112563,0.0712296292,0.0762958899,0.0614202395,0.269656986,-0.0918271616,-0.0211747643,0.3710107505,0.5512713194,-0.3462454975,-0.1316279918,-0.2795209587,-0.1425871402,0.0464556217,-0.180913046,-0.2402541935,0.013754948,-0.1374774426,-0.1102682427,-0.1549270302,-0.3587170839,0.2438014001,0.008351272,-0.3695170879,-0.3090163171,-0.2120955586,0.1773448288,-0.0065538902,-0.3145365715,0.1816492528,-0.0601580814,0.0093021439,0.2860340476,0.2964823842,0.4562503994,0.0833472237,-0.4143469334,-0.2541948855,-0.1805916578,0.0638355464,-0.0275508538,0.3468858004,0.0370091014,-0.1990600824,0.4957928061,0.0606259257,0.035857223,0.1340503097,0.4426796436,-0.5077698827,-0.1259602457,0.6570252776,-0.0669105649,-0.1711490303,-0.0144005073,0.2682855427,-0.313123107,0.0917255282,-0.0007791146,-0.0426418819,0.1423483938,0.0749628544,0.0250100791,-0.1708717793,0.4039265811,0.319394201,0.0195973273,-0.2624467909,-0.2591548562,-0.7384793758,0.1910721064,-0.1162960976,-0.3426277339,-0.0722841993,-0.1999579817,0.2421927601,0.0468117632,-0.143645227,0.101925008,0.0297586638,0.1439784467,-0.1850314885,-0.3121943176,-0.2006991804,-0.0220741108,-0.0688942671,-0.4064727128,0.1168579757,-0.3669164777,-0.0984173641,0.0744385496,0.0026448462,0.5323455334,0.0727048665,0.2942675352,-0.1021233872,0.8140289783,0.1740232557,-0.0489028804,-0.1733034849,0.1147557572,-0.1063158289,0.1628691107,-0.1273643523,0.1119908169,-0.2088825405,0.4868567884,-0.3890764415,0.6018158197,-0.0178611204,0.0819393247,-0.2825646698,0.0114689134,-0.2549231648,0.089187853,0.2956136763,0.1952596009,-0.076654017,0.3548053801,-0.1530628651,-0.2579870522,0.4175685346,-0.1606322974,-0.1191652119,0.031318523,0.2214600742,-0.0125493053,-0.090629153,-0.4313506484,0.1975592822,0.3064984679,-0.0652703494,0.0352906212,0.1964746118,0.2254900932,0.0979530513,-0.0058735618,0.9912316799,-0.1093053967,0.1168520153,-0.1647678018,-0.2565422058]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"> Maybe use the YYYY.MM.DD as the config name ? That's what we are doing for wikipedia\r\n\r\nI'm not sure if this will work because the name should be unique and it seems that he has multiple config name in his data with the same version.\r\nAs @patrickvonplaten suggested, I think you can add a comment about the version in the data description.","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":63,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\n> Maybe use the YYYY.MM.DD as the config name ? That's what we are doing for wikipedia\r\n\r\nI'm not sure if this will work because the name should be unique and it seems that he has multiple config name in his data with the same version.\r\nAs @patrickvonplaten suggested, I think you can add a comment about the version in the data description.","embeddings":[-0.0867200196,0.273745954,0.0112488763,-0.1407608241,0.06271144,-0.0481668338,0.3938752115,0.3804911971,0.089334771,0.2072925568,0.2584843934,0.1472849846,-0.2399412692,-0.015442811,0.0475122295,-0.0067649507,-0.0585154556,0.212169677,-0.0290324856,0.2738312185,-0.084899053,-0.0111241778,0.0752315372,0.0901787728,-0.0447208472,0.1121766046,-0.1227182001,-0.0627685487,-0.2637858689,-0.9248585701,0.3450054228,0.1243726015,0.1424257159,0.505398035,-0.000122348,-0.0531865545,0.6260417104,-0.1073716879,-0.6992914081,-0.1502024084,0.2308250368,-0.4619852304,0.2858994007,-0.1848255545,0.0345431305,-0.1320323795,0.2310450673,0.0503384918,0.0913198218,0.1461351812,0.1093361527,0.1425985098,0.1153796017,-0.1623488367,-0.1014972478,0.0278073363,-0.0967587978,0.5679475069,0.2544053793,0.1326054186,0.0163471717,-0.0107254805,0.0691266283,0.4541423321,0.130886063,-0.0550915748,0.5264000893,0.0546021387,-0.1274577975,0.1345887929,0.9771099687,-0.5941920877,-0.3401079178,0.0099459672,0.1404200494,-0.0033941194,0.2467773855,-0.1763136387,-0.3652087152,0.008559336,-0.0973320231,-0.2902490199,-0.1943590641,0.3475721776,-0.2534486353,0.424346745,0.1328440309,0.2170556635,-0.050475046,-0.2341308892,0.1444251835,-0.0475862585,0.0646230131,0.1204755977,0.0650520921,-0.3192999363,-0.4705287814,-0.328173697,0.112379916,-0.1774080396,-0.4003559053,-0.2411047369,0.0657076314,0.2026884854,0.1215652898,0.1992538273,0.3851887584,0.2224493027,0.2539424002,0.1447160691,0.1746027619,0.2078223377,0.0239006337,-0.4101084173,0.0025961222,0.030733617,0.2849099338,-0.2973404825,0.0172445662,-0.1251343787,-0.3631385267,-0.0263323933,-0.2165974975,0.3584257662,-0.2512477338,-0.2243993431,0.0818880796,0.3014504015,-0.1021782681,-0.0719931349,-0.0232946165,0.0245751254,-0.1246247068,0.1102535948,0.2417030931,0.0867036879,0.4142282903,-0.2295178771,-0.1562173814,-0.1250013113,-0.0037395095,0.2119347602,-0.236028567,0.3118693531,-0.1720710844,0.0903826058,0.0259262845,-0.4385855794,-0.3004496396,0.1951527148,0.0872920007,-0.3440077007,-0.1151545867,0.0968005285,-0.5595495701,-0.0811964795,0.1679898053,-0.1205598414,0.1339519322,-0.4694450796,0.036544919,-0.2592833936,-0.3031272888,-0.1586189866,-0.0653035566,0.2700704336,-0.3147053719,-0.3589433432,-0.3312329948,-0.207387194,0.0995778441,-0.3180844188,-0.0369836502,0.1321742237,-0.235939458,-0.074274011,0.5720569491,-0.5332984924,-0.3166052401,0.4164279103,-0.3450374603,-0.2490305007,0.2598500848,0.0241569262,0.1698991209,-0.2648550868,-0.1230879799,-0.0503988676,0.1220727861,-0.1038053483,-0.1910211146,-0.4137368798,0.0998167545,0.2717950046,-0.1085234359,0.1087996587,0.1280963123,0.190027833,0.4033350646,0.0872235,0.1264391541,-0.0844396129,0.1908574551,-0.0995471925,-0.155177623,-0.037927635,-0.6501656175,0.2827853262,-0.0225358084,0.2207427323,0.0293448586,-0.111420773,-0.594740808,-0.0711897612,-0.0147449775,-0.2298111916,0.0382472761,0.2451983094,0.040380951,0.0287419427,0.132302165,-0.0691986158,-0.1727848798,0.1583565623,-0.0575009398,-0.0359481126,-0.1088355333,0.1005447209,0.1413194686,0.1920615435,0.0882061794,-0.2097434551,-0.0082583241,0.1989328116,0.2664547861,0.2994746566,0.1711785495,-0.0397524312,0.1177480444,0.1686531454,0.2221972495,0.2807427347,0.0006243974,-0.2647777498,0.0124517307,0.1801750064,0.2809550166,-0.0212842952,-0.3174166083,-0.0282782242,0.319344759,0.0637189522,-0.1611047089,-0.2894610763,0.0456849746,0.3787927032,0.5581903458,-0.0528634191,0.1500062793,-0.2476996928,-0.2477092743,-0.3083853126,0.0146637214,0.0906416029,-0.1740710735,-0.0253096484,0.0588444695,0.2664115429,0.6372907758,0.0469010733,0.1413996965,0.0812915266,-0.2351587266,-0.1235398203,0.1703160852,0.1829021573,0.1264069676,0.1413461119,-0.0091574667,-0.2856316268,0.0274155401,-0.0330830887,0.4702055454,0.3248136342,-0.3375098705,-0.1011722386,-0.2301952243,-0.3934481442,-0.5377241969,-0.4357345104,-0.4151235223,-0.3019935489,-0.0380007587,0.2989151478,0.0214011818,0.3244083226,-0.2172331661,0.290810436,-0.2720581591,-0.4005092382,0.0599124655,-0.2860786915,-0.316165477,0.0083329827,0.2427163571,-0.0401453376,0.4774542153,-0.0952487439,-0.399439007,-0.2494901866,-0.3282950222,0.0896219239,0.0833929926,0.5332785845,0.4108942747,0.1148156673,0.5304839015,-0.0375305414,0.0177491456,-0.0744116679,-0.3561416566,-0.2321403772,-0.0141687626,0.2195192724,-0.0558517948,-0.3374412954,-0.5183429122,-0.1113724113,0.0803157538,0.0580617674,0.0783332139,0.2125877589,-0.0920212641,0.1044832841,0.0639614314,0.0126589676,-0.1317116618,0.1249209195,0.0439360775,-0.1699934155,-0.033919435,0.1276836991,-0.0689528733,0.048999235,-0.1415831447,-0.4356604815,-0.4000315368,0.1655731648,0.1715940833,0.3695299029,-0.2987922132,0.253908664,0.2139564604,0.1299117208,-0.1091062874,-0.4278604686,0.2279863507,0.1012911052,0.3282501698,0.2778740823,0.3687109649,-0.1983781159,0.8586021662,0.2069288045,-0.2106228769,-0.140954569,-0.2455222905,-0.0657978356,0.0359346904,0.0135255847,0.2188209742,-0.0273560937,-0.3129955828,0.308288157,-0.2308590263,-0.1324726045,-0.1215602458,0.2083230317,-0.0257658549,-0.1131550148,0.0600839481,-0.1421558261,0.0608767308,0.0080190385,-0.0666141436,-0.24221237,-0.2538880408,0.0287346989,0.338599205,-0.187069878,0.1998989433,-0.1525164396,0.0184741355,-0.4077702761,0.3384023309,0.3403618038,0.5474078655,0.120121479,-0.0353846215,0.1302523911,0.1420368254,0.2074140608,-0.093703486,0.1443245262,0.2877646983,-0.075814575,-0.1367583573,-0.3179686069,0.0366009325,0.1823500693,0.0333267562,0.1228108853,-0.4709644318,-0.0984624997,0.2863663733,0.1944670379,-0.163132146,-0.3334002495,-0.1024145409,0.0487912968,-0.078747496,-0.2163063586,-0.2403726131,0.0442989878,-0.0388559289,0.3086199462,0.0088628298,-0.0353341028,-0.0612189472,0.2663486004,0.2031404525,-0.2001658827,0.2893233299,0.2616367638,0.0841667056,0.3153355122,0.2017226964,-0.0826863274,-0.2403506488,-0.0825027376,0.0174107999,0.305663377,0.3083480895,0.240355432,-0.0085575823,-0.1876197159,-0.1448667198,-0.2892587185,0.1690632105,0.2935922444,-0.1301059425,-0.1003356948,-0.3282798827,0.7530415058,0.1722810715,-0.2961041629,0.5668509603,-0.2393666655,-0.4109598994,-0.0135607421,0.1026329622,0.983656466,0.0768217742,0.2185542136,0.223509416,-0.1969558448,0.3902522624,0.1606567353,0.0613651015,-0.5651336908,0.0880424008,0.029159693,-0.0989540517,0.2444810569,-0.0641176105,-0.3517038524,0.3407964408,0.0821813196,0.2972544432,0.2518068552,0.3703328967,-0.0462054089,-0.0392567925,-0.2405529767,0.0510329902,-0.2243731618,0.3719371855,-0.2197385132,0.0341649279,-0.1888172477,-0.0778594241,-0.3279076815,-0.2395801097,-0.1501383036,0.069476828,0.0191968121,-0.1875432432,-0.0277949627,0.2390358895,0.0630716607,-0.1204936728,-0.190208599,0.1141537279,-0.2027817369,0.0889564678,0.2375631183,-0.017683804,0.3286878169,-0.3449379206,-0.1181571558,0.1530845463,-0.1872831434,-0.2327241898,-0.0833960921,-0.1945824176,0.0458210893,-0.3402771652,-0.2401780933,-0.0763223916,0.0278451648,-0.1041361094,0.0382427908,0.1613780856,-0.3257567585,0.0171801113,0.05420563,-0.3329342008,0.0469637513,0.3687141538,-0.0863739923,-0.0008776627,0.6735665798,-0.2475475669,0.0873534679,-0.1080074608,0.0003370644,-0.3327262402,-0.1332299262,-0.0411363691,-0.0439339541,-0.2875957191,-0.0657156706,0.559666276,0.5117819309,0.1010083482,0.0624178946,-0.3830115795,-0.3422269225,0.0754092261,-0.1296175718,0.2656584978,-0.0739362612,0.2888073921,0.0922791287,0.0448268913,-0.2038615793,-0.142798081,0.0174056236,0.2378127277,-0.1625674963,0.1230777279,-0.1156911999,0.0284341387,-0.0143928677,-0.1034309044,-0.2604857087,-0.0468020886,-0.1658411175,0.2251224518,0.0308049582,0.0834361389,0.0127724931,-0.1899289638,0.0993688181,0.1725460589,0.0049480349,0.1131477281,-0.067547664,0.272829324,0.3285062611,0.1184525266,-0.1246718094,0.2750124931,0.1070988476,0.2108289301,-0.006237552,0.3608352542,-0.2627816796,0.0366214104,0.1833157241,0.1857272387,-0.2079302669,0.1354490668,0.0163808838,-0.0431311578,0.0332614109,0.1499871612,0.3344955742,0.7781231999,-0.3100931048,-0.1703517586,0.4529953003,0.0666804016,-0.1927295476,-0.0510592312,0.331204772,0.040324863,0.0768392682,0.260233283,0.0116452342,-0.064655453,0.199753046,0.1704651117,-0.0107983472,-0.2592610717,0.0898494273,0.3426214755,-0.1821499467,0.1657043397,0.3916592002,0.1482624412,0.2784762681,0.0073830113,0.135996744,0.4200585485,0.0414295606,-0.079955481,0.0923589468,0.119661808,0.1031177938,0.2714883089,-0.1196326092,0.0387167297,0.4046548605,0.5331205726,-0.2811604738,-0.1410221308,-0.3413613439,-0.1561206579,0.0632823333,-0.2039290369,-0.3100763559,0.0280646849,-0.124559544,-0.124959968,-0.171752125,-0.3785264492,0.269066453,0.0695129186,-0.3724465072,-0.3337205052,-0.2165624499,0.2483433336,0.0970796794,-0.355741024,0.1587710828,-0.0236980896,-0.0322236679,0.2971594036,0.3373472691,0.4456104636,-0.0167821292,-0.3285935521,-0.2467666715,-0.1693964303,0.0949044079,0.0320788212,0.2031139731,0.0107955914,-0.300865829,0.5508679748,0.0714309141,0.0513558686,0.1238610223,0.3914077878,-0.437710613,-0.1757072806,0.5832920671,-0.0174543113,-0.1820178181,0.0052442565,0.2448502779,-0.3289756477,0.1818322688,0.0218245555,0.0414459184,0.1453244239,0.0867565721,0.021520054,-0.1136463135,0.4138198793,0.2832201719,-0.067356661,-0.2639584243,-0.2231560946,-0.7593222857,0.1801719666,-0.1377020329,-0.3707537055,0.0003886807,-0.212786451,0.2463701367,0.0333916992,-0.1169734523,-0.0176872704,-0.0045459913,0.1856840551,-0.2392715365,-0.3589062691,-0.1332222819,0.0147405118,0.0135994731,-0.3658695221,0.1153110042,-0.2808386981,-0.1155010313,0.0610263944,0.0120415259,0.5453346968,0.0536651872,0.2477727085,-0.1688180417,0.7727634907,0.1596137881,-0.0645501837,-0.1531021446,0.0954500735,-0.1310413629,0.1424975693,-0.0821687505,0.1293337196,-0.1667124778,0.4938798547,-0.3666076064,0.540687263,-0.0992247015,0.0808291808,-0.221819073,-0.0370958373,-0.1514172107,0.1198634058,0.2276909351,0.2716071904,-0.1073482186,0.3570457697,-0.2341275513,-0.353611201,0.481494993,-0.100370191,-0.1550838798,0.0216877535,0.3206322789,-0.0950705558,0.0744412988,-0.4297247529,0.2376840711,0.3241073787,-0.0406027175,0.0087048104,0.2114334255,0.2268542349,0.0831518993,0.0017669218,1.0600345135,-0.1092360318,0.0234473143,-0.2541800737,-0.3114488721]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"Actually maybe our versioning format (inherited from tfds) is too strong for what we use it for?\r\nWe could allow any string maybe?\r\n\r\nI see it more and more like an identifier for the user that we will back with a serious hashing\/versioning system.- so we could let the user quite free on it.","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":54,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nActually maybe our versioning format (inherited from tfds) is too strong for what we use it for?\r\nWe could allow any string maybe?\r\n\r\nI see it more and more like an identifier for the user that we will back with a serious hashing\/versioning system.- so we could let the user quite free on it.","embeddings":[-0.1527510583,0.2105686069,0.0053558247,-0.1521225572,0.1695314944,-0.1229730174,0.3444485664,0.3772257268,0.0139800562,0.2296853364,0.1677791476,0.0152248899,-0.3342413306,0.0313071348,0.0228184462,-0.1002439708,-0.0594148934,0.1970065981,-0.07303565,0.3432954848,-0.060112685,-0.0731619373,0.0019713789,0.1934303194,0.0108176172,0.1458041072,-0.1153508872,-0.0187536702,-0.388517797,-0.8583580256,0.1343808174,0.0918508172,0.2320994288,0.4368851781,-0.0001261235,-0.0739656165,0.5866602063,-0.0240417346,-0.76744169,-0.1698298305,0.3446432054,-0.3756413162,0.2535519004,-0.0657789782,0.1137201115,-0.1838426143,0.1813541353,0.0753962398,0.1361462176,0.2455598861,0.0991945639,0.3106926978,0.028130766,-0.1111011356,-0.0595696643,0.12354514,-0.1438006163,0.5142219663,0.4895508885,0.2406701893,-0.078602612,-0.0708241239,0.062615484,0.4359807074,0.1371402293,-0.0326458737,0.5384781361,0.0442370549,-0.1900421828,0.1604462713,0.8945170641,-0.5086080432,-0.4655534029,-0.0676140115,0.0179396346,-0.1036736593,0.1076579317,-0.1938883811,-0.3774500787,0.208910659,-0.1669587791,-0.3321849406,-0.099039562,0.2889301181,-0.1762079895,0.2312367558,0.1270467192,0.2515535355,-0.0295838341,-0.0673802122,0.1549281627,-0.0520944595,0.0152693242,0.1340529621,0.0880432501,-0.3319088221,-0.4149749279,-0.2981059253,0.1995219588,0.0256361235,-0.5338360667,-0.1899030656,-0.0685549602,0.1762414128,0.0877644643,0.2744112909,0.3668691516,0.2205429822,0.3118174374,-0.0107747652,0.2858647406,0.2193337232,-0.0055318722,-0.2797216773,-0.0023883595,0.0753944889,0.1061919481,-0.2118771672,-0.0213846564,-0.2597497106,-0.4228295386,-0.0277204104,-0.0929705203,0.2329256982,-0.1784337461,-0.1466088146,-0.0199730396,0.3341645598,-0.10098546,-0.1949630827,-0.0001091545,-0.0203206725,-0.2484293729,-0.0608610325,0.2246825844,-0.1317874342,0.3598705232,-0.3554433286,-0.1441572607,-0.0381403603,-0.0225203186,0.0308944341,-0.1706526428,0.2882429957,-0.2357190102,0.0170724653,0.0558260903,-0.300290674,-0.2188138962,0.2182501704,-0.00322932,-0.436083585,-0.1364366561,0.0600697175,-0.5637034774,0.0217235312,0.0472865775,-0.1695634872,0.1888354123,-0.2967651188,0.1417250484,-0.2503827214,-0.360743463,-0.1515683681,-0.178382352,0.2280643433,-0.3107924759,-0.2898622751,-0.2297293544,-0.3846806288,0.0779804513,-0.1305169165,-0.0723214597,0.2086261064,-0.3361051977,-0.0359190889,0.6512774229,-0.5617482066,-0.2883246839,0.6501528621,-0.3996012509,-0.2595348358,0.2737458944,-0.0726999491,0.1499108374,-0.4444636405,-0.1325359792,0.0042216242,-0.0445069894,-0.1131415665,-0.0871563405,-0.4246346653,0.2098808438,0.2037368417,-0.0647907332,0.2234140486,0.1218448356,0.1067118198,0.447171241,0.0352130979,0.1692753583,-0.2102168351,0.2047894895,-0.1228681058,-0.1074659973,-0.1356255412,-0.5853859186,0.2420368791,-0.0473356619,0.1534970552,-0.0063745119,-0.0640541762,-0.4417180419,0.0285028666,0.1057558134,-0.2158669084,-0.0553909466,0.1458471715,0.0250658635,-0.0346959829,-0.0263200961,-0.0075410814,-0.2913226485,0.1625150293,0.2089127749,-0.1484843045,-0.0833220556,0.0995793715,0.1504308134,0.2522899508,-0.069793798,-0.3053160906,0.0206990167,0.3233892918,0.2586054206,0.1494383961,0.0798541307,0.0505342968,0.1862144172,0.2668224275,0.1463747621,0.2067105025,0.0248774067,-0.2785701752,0.023999054,0.1552948952,0.1547406465,-0.0246730689,-0.2123955339,-0.0428837165,0.2966695726,0.0858922228,-0.1595072448,-0.2886157036,0.1007187441,0.2132441103,0.6271631718,0.0130698625,0.234406665,-0.1607029736,-0.1911903769,-0.4197960496,0.0476297401,0.2504348159,-0.0863969773,-0.061566107,-0.0387018584,0.245597735,0.8113176823,0.0071627758,0.2236732692,0.059365768,-0.0622192919,-0.2031270564,0.2153962553,0.1874169111,-0.0199555196,0.1495851129,-0.0095342221,-0.2130720466,0.0111513054,-0.0299495626,0.3745067418,0.3891143501,-0.4598509073,-0.0712630153,-0.2434232235,-0.4637842774,-0.3513340056,-0.3214049935,-0.2830208838,-0.2886622846,-0.0760621578,0.3205254376,0.042951379,0.2218059152,-0.2701076567,0.2553033531,-0.2692339122,-0.2900582552,0.039597556,-0.1383657604,-0.3783037663,-0.0101342872,0.2034051865,-0.1298724264,0.4923704267,-0.0401080884,-0.2758656442,-0.2192289233,-0.3464267254,0.1785963625,-0.0080206348,0.5385330915,0.5283279419,0.011746807,0.4921999574,0.045098193,-0.1030585617,-0.282466799,-0.4460033178,-0.1996397674,0.0454242565,0.1237611547,-0.225199461,-0.2420060486,-0.5144339204,-0.1097066924,0.144529447,-0.034998782,0.1203451455,0.2705279887,-0.1861665696,0.1380648166,-0.0068364283,-0.0619552657,-0.1589343548,0.0721360371,-0.0934135169,-0.1512038261,-0.1503474265,0.0513874143,0.0091444002,-0.0261237267,-0.035869427,-0.3401893377,-0.4658867419,0.1552981436,0.1391379982,0.3482474387,-0.228883341,0.1996987313,0.2338796109,0.1612381637,-0.0610769652,-0.3452841043,0.1479627788,0.0475275852,0.244982183,0.2692882717,0.419529587,-0.0195856169,0.7755751014,0.1433252096,-0.0958903879,-0.0783183575,-0.2671538889,0.0062824115,-0.0542478226,-0.0055671064,0.3108572662,-0.1007471681,-0.2828229666,0.3185684979,-0.189475134,0.0514922217,-0.0904453546,0.2656157613,-0.068093136,-0.0441911742,0.0500732288,-0.2819502354,0.0393402129,-0.0126572419,-0.0375200808,-0.1332028955,-0.2186045051,0.0995264724,0.3623550534,-0.0186795276,0.290342927,-0.1158390492,0.1568785608,-0.3423962593,0.2570711374,0.4411435425,0.6185429692,0.0831977725,0.091347672,0.1832656115,0.1200193688,0.2273238003,-0.0364771076,0.2337812036,0.0632607862,-0.2360342294,-0.0178660657,-0.2264618725,0.0005947087,0.096052289,-0.0757207423,0.1736176461,-0.4774000943,-0.2320781052,0.217721507,0.1053232551,-0.0766578987,-0.3893097937,-0.0650529936,0.0382843912,-0.0528653599,-0.3129262924,-0.1361307204,-0.0052228402,-0.0890213475,0.4332496226,-0.1022617891,-0.1455698311,-0.0051328964,0.1043287963,0.0447094217,-0.1961445808,0.206121102,0.1978488266,0.174650982,0.2170516551,0.1883539855,-0.0840262175,-0.1561632752,-0.0374351442,-0.0573025383,0.3986184895,0.3663240969,0.1006257311,0.0427717082,-0.2101763338,-0.1530300081,-0.3422592282,0.2205568552,0.2823307514,-0.1639903784,-0.0510298498,-0.2840905488,0.6471725702,0.143563509,-0.2456749976,0.4512060583,-0.1504861265,-0.3036877811,-0.025458103,0.12860246,0.9857160449,0.043363329,0.2286954075,0.3851466775,-0.3389442563,0.4281964898,0.1474850476,0.0266634673,-0.5198190808,-0.0551795512,-0.0594308749,-0.1654727906,0.2741520107,-0.1585927755,-0.2002357841,0.3456085324,0.0879504681,0.373614043,0.4137506485,0.3992369473,-0.1054462269,-0.0578830987,-0.1608775258,0.0220652111,-0.1849373877,0.2843340337,-0.2157171965,-0.0206527654,-0.1866103262,-0.028540181,-0.2467509657,-0.2373770177,-0.1255302131,0.0179843996,0.1404604614,-0.1092204005,0.0576988496,0.1687144786,0.1828474253,-0.0949667394,-0.1353602111,0.0253714379,-0.2317045778,0.218809098,0.304461062,0.0282770302,0.3999041319,-0.2815256715,-0.0707609653,0.1486912966,-0.2799600959,-0.2452793866,-0.1716294736,-0.2162201405,0.1609037519,-0.3544024825,-0.2652533054,-0.1846260726,-0.0204339474,-0.0361430943,-0.0385485403,0.2329090387,-0.3300033808,-0.0132328169,-0.0374223404,-0.3219469488,0.0307154506,0.4821372628,-0.0463778749,-0.1701239645,0.6307936907,-0.2415558249,0.0289761797,-0.0897456855,0.0322232544,-0.1857169122,-0.186218068,0.0385230407,-0.0700711384,-0.3210484982,-0.034937378,0.6218970418,0.5002502799,0.0262285322,-0.0709335208,-0.4239472151,-0.4768605232,-0.0198271219,-0.2159661651,0.14384152,-0.2070553154,0.3047217727,0.0950570405,0.0363291018,-0.1691189408,-0.2332280725,0.1270900369,0.2611347735,-0.2023116499,0.0358815566,-0.0104004256,0.0488338023,-0.0775552988,0.0545224883,-0.34435305,-0.0227407049,-0.2212177217,0.2403481305,0.1290156692,0.0158836357,0.1009758562,-0.129224658,-0.0302050952,0.1971331239,-0.0388315432,0.1435365081,-0.0886921808,0.2655785978,0.3191796839,0.1678487211,-0.0894197896,0.2904235125,0.0813600197,0.26330176,-0.0373626016,0.302998513,-0.2278717309,0.0369661525,0.1037138775,0.2465103567,-0.0189531446,0.2433108091,0.1185162961,-0.0831339732,0.0995228812,0.0027662588,0.3833668232,0.7761166692,-0.1842091978,-0.3148204684,0.4969838262,0.0178063698,-0.1070319563,0.0064506615,0.2476189882,0.1680014133,0.0253084674,0.3604952097,0.0300943144,-0.0359100327,0.0922939479,0.1062340364,0.0886682495,-0.2641745806,-0.2379089296,0.521091938,-0.2259115875,0.2412495762,0.3535765409,0.1358070523,0.2959990501,0.0376607962,0.0756625533,0.5019494891,-0.01001353,0.0431271903,0.1404643059,0.1329786032,0.0145008508,0.2870317698,0.0750098079,0.0137452614,0.3375614285,0.590536654,-0.202177301,-0.1644719392,-0.1727119088,-0.1674891114,0.0769596547,-0.2280699909,-0.2735538483,0.1288489848,-0.2124824226,-0.1030298695,-0.2302060276,-0.2926721275,0.2116537541,0.0527508408,-0.3386245668,-0.2714465857,-0.4014511406,0.1491457671,0.0506777056,-0.3459292054,0.1293285191,-0.0709914714,0.1156202555,0.3956112862,0.2540735006,0.3397339284,0.0673303902,-0.3097273707,-0.3134120107,-0.324321121,0.150709942,0.0087401113,0.3977159858,0.0177150257,-0.1132778525,0.6163253784,0.004213539,0.1003301814,0.2640595734,0.3659212291,-0.5369965434,-0.1176469028,0.785967648,0.0799899474,-0.1935423762,0.0794038549,0.228684321,-0.191302672,0.1738030463,0.0464404859,-0.1126520261,0.1843629032,0.1114716977,0.0151839051,-0.0522953756,0.3484433591,0.4192720652,-0.0645557716,-0.2744944394,-0.0155785959,-0.6457430124,0.1813395023,-0.0532816313,-0.2521620095,0.0043504289,-0.2423717231,0.17777659,0.1159846485,-0.2468877435,0.1393566281,-0.0888076723,0.0866631791,-0.1472109109,-0.3679318428,-0.1788335145,0.0805126131,-0.0754692256,-0.4362683296,0.2002203465,-0.420611918,-0.1224083677,0.0762811378,0.0151648251,0.5335341096,-0.0193109345,0.2926970422,-0.0549521968,0.5333089828,0.1951011866,0.0060291658,-0.1851986349,0.1820676029,-0.1992658079,0.1308487207,-0.0489592366,0.0695416033,-0.1477246135,0.5281546116,-0.421425283,0.5098819733,-0.0507756993,0.1615394503,-0.3290122747,0.1024018377,-0.240339011,0.086355485,0.2888398767,0.2186271846,-0.1191042587,0.3717104495,-0.2540441751,-0.2021583617,0.4779550135,-0.1859750897,-0.1001986414,0.1542516202,0.303208679,0.1987037659,-0.1865209788,-0.5294916034,0.1576288044,0.326366365,-0.1043697298,-0.045027554,0.2218457162,0.2079702765,0.0608655289,0.0075389463,0.9585243464,-0.1330401152,0.0889604837,-0.0688256547,-0.2480966747]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/161","title":"Discussion on version identifier & MockDataLoaderManager for test data","comments":"I'm good with either putting it in description, adding it to the config, or loosening version formatting. I mostly don't have a full conceptual grasp of what each identifier ends up meaning in the datasets code so hard to evaluate the best approach.\r\n\r\nFor background, the multiple formats is a consequence of:\r\n\r\n1. Each example is one multi-sentence trivia question\r\n2. For training, its better to treat each sentence as an example\r\n3. For evaluation, should test on: (1) first sentence, (2) full question, and (3) partial questions (does the model get the question right having seen the first half)\r\n\r\nWe use the date format for version since: (1) we expect some degree of updates since new questions come in every year and (2) the timestamp itself matches the Wikipedia dump that it is dependent on (so similar to the Wikipedia dataset).\r\n\r\nperhaps this is better discussed in https:\/\/github.com\/huggingface\/nlp\/pull\/169 or update title?","body":"Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.","comment_length":152,"text":"Discussion on version identifier & MockDataLoaderManager for test data\nHi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp\/utils\/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests\/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers the error. If I can get something to work, I can include it in my data PR once I'm done.\nI'm good with either putting it in description, adding it to the config, or loosening version formatting. I mostly don't have a full conceptual grasp of what each identifier ends up meaning in the datasets code so hard to evaluate the best approach.\r\n\r\nFor background, the multiple formats is a consequence of:\r\n\r\n1. Each example is one multi-sentence trivia question\r\n2. For training, its better to treat each sentence as an example\r\n3. For evaluation, should test on: (1) first sentence, (2) full question, and (3) partial questions (does the model get the question right having seen the first half)\r\n\r\nWe use the date format for version since: (1) we expect some degree of updates since new questions come in every year and (2) the timestamp itself matches the Wikipedia dump that it is dependent on (so similar to the Wikipedia dataset).\r\n\r\nperhaps this is better discussed in https:\/\/github.com\/huggingface\/nlp\/pull\/169 or update title?","embeddings":[0.0576032586,0.2424968183,-0.0005986306,-0.0168732125,-0.0262222327,-0.1091884449,0.370344609,0.2744979262,0.001739301,-0.0733632967,0.1906664371,0.1067379192,-0.2590659261,0.0527149476,0.0636911318,-0.1715938449,0.0611246936,0.2052573115,-0.0903218836,0.1993689686,-0.1066960767,-0.013935524,-0.0217242874,0.2025265396,-0.0918588638,-0.0448577106,-0.2902240157,0.0635661334,-0.3564002514,-0.951844871,0.2559961677,0.1613283157,0.2148887366,0.3902352154,-0.000122996,-0.1342219561,0.5136099458,-0.0304478332,-0.7506933212,-0.1976183653,0.3274134099,-0.4366217852,0.3908320367,-0.0833521858,0.2027073652,-0.2667598128,0.3141605854,0.2079798728,0.1440406889,0.2145274431,0.0708214268,0.2529842854,0.0297172219,-0.0984105095,-0.0401779562,0.1192993447,-0.0344780795,0.5099808574,0.3937709033,0.1262135357,-0.1632188261,0.1172782481,0.1400750875,0.3983493447,0.2259790897,-0.0179536529,0.5039620399,-0.0475987755,-0.2023796588,0.054717876,0.8494493961,-0.5009226203,-0.4326483309,-0.1416949332,0.0322893485,-0.1647890955,0.215471074,-0.1092026532,-0.155073002,0.0571248904,-0.2444538623,-0.1113326922,-0.1574691385,0.3368920684,-0.2251197249,0.2915144563,-0.010365339,0.1961043775,-0.0307248626,-0.1835584342,0.0579076596,-0.201194644,0.076302357,0.1746196151,0.0161247719,-0.2886019647,-0.2943039536,-0.1872648001,0.1245486215,0.0148579227,-0.4481178224,-0.1796251088,-0.037364874,0.2117079794,0.2502318025,0.3405126631,0.3932392597,0.1903862357,0.1532818228,0.080182299,0.1514482349,0.1508202106,-0.0418833457,-0.2256963253,-0.0119262999,0.0255419165,0.1661311239,-0.2824716866,-0.0516791269,-0.125729844,-0.5409556031,-0.0757431984,-0.1315049231,0.232893914,-0.1798986048,-0.045072075,0.0321231186,0.4197336733,-0.1784877181,-0.1591355801,-0.0156887863,0.074614495,-0.2581368089,-0.0456648469,0.1743211299,-0.0906316936,0.4508666694,-0.2195088267,-0.3382495344,-0.1793952286,0.0695329607,0.0131425904,-0.0920684114,0.2377603501,-0.195138067,0.0676418543,0.0803676695,-0.4109777808,-0.278220892,0.0108870352,0.0562765449,-0.3433366418,-0.0272840112,0.0628061891,-0.4833724201,-0.010563083,0.1242057756,0.0902784467,0.1491734535,-0.3166505992,0.1007444933,-0.4133774042,-0.2347713262,-0.09195555,-0.1889970303,0.3042599559,-0.2840290666,-0.2755050063,-0.1664578617,-0.3127403855,0.0908932686,-0.1886774749,-0.0761043727,0.2745265365,-0.1902465969,0.0301888958,0.5996155143,-0.5389534235,-0.2087565511,0.5325268507,-0.2751671076,-0.0529346727,0.2893841267,-0.0739789233,0.11017441,-0.3594630957,-0.1081959158,0.0583431013,0.0039684311,-0.1924611479,-0.2157147676,-0.2680107057,0.1802842617,0.2851307094,-0.1920588017,0.0187028758,0.0654814392,0.1850584894,0.3273562193,0.0951834172,0.1283269376,-0.1886839867,-0.0551352724,-0.078462109,-0.1236384809,-0.0161704775,-0.7373569012,0.1525451243,-0.1034967005,0.2063196748,0.0440025479,0.0035778468,-0.4655615985,-0.0496902689,-0.0746113583,-0.3006815314,0.030343568,0.2375243455,0.0760563537,-0.0039685084,0.0089023514,-0.0123311263,-0.3098595738,0.2481218874,-0.0493221693,-0.1834705025,-0.0836690515,0.1499595046,0.2053993195,0.2515467107,0.0339780934,-0.2759887874,0.104669176,0.325684011,0.2371479571,0.2414845824,0.1715386063,0.0690022781,0.2337163538,0.3315992653,0.3070375621,0.1655813605,0.0867639706,-0.2693616748,-0.091931887,0.1549016237,0.2427976876,0.0045868633,-0.1618411988,-0.0673988536,0.3385397494,0.1212914437,-0.2309662998,-0.2655969262,0.0126359593,0.2951552868,0.6402708888,0.0631799996,0.0867368728,-0.034492854,-0.1113729998,-0.4682069421,-0.0232922081,0.0999129266,-0.2714600861,-0.1655065864,-0.0195121579,0.2567766905,0.7403603792,0.0143573759,0.2837748528,0.0430517532,-0.2606042922,-0.2131257802,0.1901426464,0.2923331857,-0.0724745616,0.1426078528,0.0594439395,-0.221412003,-0.011801376,-0.1442519426,0.3677423298,0.3290628791,-0.4422619939,0.0115682436,-0.2243393809,-0.4463371336,-0.5092366934,-0.3774478734,-0.260625273,-0.3482601643,-0.054240752,0.2230655998,-0.0988534838,0.2085058987,-0.044455193,0.3514878154,-0.3339144886,-0.0618906543,0.1812186688,-0.2874465883,-0.4304261804,-0.0212775953,0.3774447143,0.0103003662,0.3699012995,-0.1629614234,-0.3067913949,-0.1718641669,-0.3577466607,0.1757736653,-0.0513365977,0.4902473688,0.4374170303,0.0568049252,0.4220124185,-0.0797566921,-0.005828253,-0.0616052635,-0.3539553881,-0.1985042393,0.0080538932,0.0115295136,-0.222992599,-0.4676481187,-0.5795462132,-0.0671649203,0.2384971231,0.0237400886,0.1362574548,0.3584893942,-0.2352586985,0.2309435606,-0.1168214455,0.0617523566,-0.1307684928,0.0484241471,0.0334694274,-0.1410069615,-0.1465092748,0.0830686688,-0.0692913532,-0.0562386177,-0.125167191,-0.4092945457,-0.4603893161,0.1925333738,-0.0198786259,0.3854576647,-0.119753927,0.1941302568,0.2285906225,0.1526881754,-0.0922198668,-0.2539879084,0.3085112274,0.0481679998,0.2682614923,0.2667805552,0.309961766,-0.0824953392,0.8548858762,0.2343316674,-0.2229711413,-0.1782390922,-0.0280719642,-0.0743275732,0.0089906864,0.0390544869,0.2459827513,-0.193739146,-0.2576619685,0.4539285004,-0.1147165969,-0.0794108808,-0.0128904246,0.2891291082,-0.1092850417,-0.0057352274,0.1946333349,-0.1576353163,0.0234707966,0.010373733,-0.0337058231,-0.158334136,-0.0684642196,-0.0331967063,0.3419372737,-0.1114081889,0.2197247148,-0.2726880014,-0.0640707836,-0.2533636987,0.2327743173,0.397885114,0.4647169113,0.1509021223,-0.0316694863,0.2120310217,0.2042482048,0.0879232287,-0.0319000632,0.1063778996,0.1811224371,-0.2204211056,-0.0303714629,-0.3027999401,-0.2254156619,0.0427549593,0.0199663751,0.1791830808,-0.4684458375,-0.2401077598,0.321111083,0.0621437728,-0.0500153005,-0.2363810092,0.0235095154,0.0056043281,-0.0498346388,-0.1732706875,-0.1989137232,-0.0138894431,-0.1176551953,0.4240598679,-0.0383676067,0.0453142002,-0.0219754316,0.2474600226,0.0965982005,-0.149227038,0.1423807144,0.2692522407,0.1429665536,0.2716625631,0.1835299283,-0.1302198619,-0.2218181491,0.0519532263,-0.0687578768,0.3754239976,0.4415144622,0.2074637115,0.0312544368,-0.2436330765,-0.1729833782,-0.3376980722,0.277703613,0.3622957766,-0.1226688623,-0.0188634358,-0.3984581828,0.7639327645,0.1151419804,-0.3186496198,0.3197761774,-0.1712697148,-0.2754097283,0.0667418242,0.0449261591,0.9632908702,0.1875797659,0.2697296739,0.3356108069,-0.3050235212,0.5088133812,0.077299729,0.0832989067,-0.5266803503,-0.0049511627,0.0118988277,-0.0852637663,0.2526693046,-0.091750659,-0.2902073562,0.3714599907,0.1541250646,0.3369058371,0.2666241229,0.3904858828,-0.1026519164,-0.0818484798,-0.3148290217,0.032741759,-0.1812620759,0.2942303121,-0.2555636466,-0.0545356236,-0.194590196,-0.0489115939,-0.1663556546,-0.2321008742,-0.1798075587,0.0440441109,0.0973836705,-0.244547978,0.0070837503,0.2463037223,0.134790346,-0.3061777949,-0.0829956084,-0.0007387234,-0.2038984448,0.1973680407,0.2699485123,-0.0520637333,0.4594920278,-0.2614246309,0.0046782661,0.0620057583,-0.2999454141,-0.011490711,-0.0774812549,-0.1792584211,0.1897273064,-0.2973540127,-0.2757717669,0.0346750095,0.0157294106,-0.0169355441,-0.013132113,0.1991764605,-0.2707007825,0.0163088311,-0.0908088312,-0.2381539047,-0.0127136232,0.3433263004,-0.0879753977,-0.0714719296,0.7205213308,-0.1423330456,-0.0130480044,-0.0877578631,0.078926459,-0.0772123188,-0.2102147341,-0.0246679764,-0.0939804763,-0.3097340465,-0.116279088,0.666424334,0.5130177736,0.0873771086,-0.0127653768,-0.3916791677,-0.3361733258,0.1102444753,-0.2133310288,0.2046304941,-0.1537820399,0.2301100343,0.2059984058,0.0917343274,-0.198963508,-0.2128663212,-0.0524187237,0.321977973,-0.208108753,0.008620494,-0.0228313915,0.0747560933,-0.07202214,-0.0303917695,-0.3601396084,-0.0361162908,-0.1606751531,0.2308598608,0.1678742766,-0.1009952351,0.051195465,-0.0566672683,0.0830277354,0.1583657861,-0.0659609139,-0.0692204237,-0.0469820686,0.315620482,0.221127525,0.2653516531,-0.0612607375,0.1533363611,0.0779319182,0.1418083906,0.0833170861,0.1833310425,-0.3230113983,0.0137753617,0.0764905363,0.2949419618,-0.1126141399,0.2398710251,0.1162517816,-0.0320994705,0.1410957724,0.086941205,0.2844581604,0.8438323736,-0.1847463548,-0.288144052,0.410600692,0.0176033266,-0.0421866216,0.0829827115,0.2660801411,0.1146043241,0.0222730618,0.2854967415,0.0558502674,-0.0035334702,0.0395242162,0.168595627,0.1924369633,-0.2128523737,-0.0687089562,0.4142083526,-0.244790107,0.2616270781,0.3799393773,0.0584362149,0.2171715647,-0.0937651396,0.0666021183,0.4720223546,0.1034332067,0.0724764243,0.1404718012,0.1095653176,-0.0547242835,0.4614464045,0.0272937566,0.0198597983,0.2861617506,0.6132851839,-0.3343770504,-0.3217316568,-0.1734317243,-0.0768055394,0.1072123274,-0.1811950952,-0.5276597142,0.0149069289,-0.2530696988,-0.0724482164,-0.2701844275,-0.35571751,0.3199391067,0.0411813967,-0.3105317652,-0.416092366,-0.3604919612,0.0854643658,0.1281212717,-0.3214929104,0.2644841373,0.2310384363,0.0690568313,0.4039296508,0.3324464262,0.4023066461,0.0916779116,-0.3861660659,-0.3036434054,-0.253918469,0.1070750058,0.1250625551,0.2783580124,-0.0427696779,-0.214996919,0.5813984275,0.0155170904,0.0790656954,0.2922722399,0.4092685878,-0.5363095403,-0.2418405265,0.6709066033,-0.0029570032,-0.2478880435,0.1084637493,0.1560738534,-0.360806793,0.0843800232,0.0010097483,0.071505107,0.1984622329,0.093607679,0.0260886624,-0.0099243,0.4939213991,0.4914433956,-0.1078462005,-0.3457841277,-0.1026550606,-0.7086454034,0.0232039858,-0.0249228291,-0.3069429398,-0.0286518726,-0.1313230395,0.1805092543,0.0959316418,-0.2593300939,0.0591554977,-0.0345859528,0.0785580575,-0.3168296218,-0.2308543026,-0.1634583026,0.0117537268,-0.1121792644,-0.247342363,0.1147055179,-0.3909784555,-0.1484862268,0.2285928875,0.0486631319,0.4706296325,0.0439793691,0.1680421978,-0.0149448048,0.5717009902,0.271389991,-0.1270384043,-0.2280654609,0.2154112458,-0.1466310024,0.17723988,-0.1287577748,0.3634102046,-0.1993662864,0.5129454136,-0.205855906,0.4702887237,-0.0446837172,0.0570265912,-0.3319069445,0.0854544565,-0.2292844355,0.0452085137,0.3392937183,0.3558689654,-0.0692561418,0.2981765866,-0.2602950335,-0.2253068984,0.3200793266,-0.3188410401,-0.1184044704,0.1429835558,0.2893081903,0.140463382,0.0245359689,-0.6277331114,0.0738109499,0.2311153114,-0.1705221385,-0.0932829529,0.1762322485,0.2381398529,-0.0625921115,-0.0359239504,0.9138431549,-0.0750885457,-0.0677718595,-0.1077091768,-0.3335293233]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/160","title":"caching in map causes same result to be returned for train, validation and test","comments":"Hi @dpressel, \r\n\r\nthanks for posting your issue! Can you maybe add a complete code snippet that we can copy paste to reproduce the error? For example, I'm not sure where the variable `train_set` comes from in your code and it seems like you are loading multiple datasets at once? ","body":"hello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n","comment_length":49,"text":"caching in map causes same result to be returned for train, validation and test\nhello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n\nHi @dpressel, \r\n\r\nthanks for posting your issue! Can you maybe add a complete code snippet that we can copy paste to reproduce the error? For example, I'm not sure where the variable `train_set` comes from in your code and it seems like you are loading multiple datasets at once? ","embeddings":[-0.1263227314,-0.2207089961,-0.100632146,0.2780107558,0.2454756796,-0.2052229047,0.1372107863,0.5128592849,0.1882870793,-0.0623357445,0.1278197765,0.3891343772,0.0037762846,-0.1606175303,-0.0107436497,0.0242125317,0.0960802659,0.2350524068,-0.1455620974,-0.1648233086,0.081894055,0.2129840404,-0.0424102098,0.1165268347,-0.3903492093,0.0412896834,0.2123562694,-0.2308897674,0.0684574693,-0.2570649683,0.1678505987,-0.0932429209,0.0258333106,0.1663175523,-0.0001097382,0.0067983456,-0.0153878154,-0.1385697424,-0.0072745453,0.0401948094,0.1181339324,-0.2117293179,-0.1091826707,-0.3121318519,-0.3320115805,-0.0376609191,0.0178094879,-0.3556600213,0.3128580153,0.1474524736,0.2607103586,0.2249089628,-0.2538127303,0.2111362219,0.1451275945,-0.0293276701,-0.174197495,-0.042305138,-0.043978855,-0.3344650865,-0.3767647743,0.5254573822,-0.0765818357,0.1701412946,0.2146666646,0.2250556946,0.1726054549,0.03593757,0.1429362446,-0.1139624342,0.0110948198,-0.2269175202,-0.0271165781,-0.3733721972,-0.4186404347,-0.0785259828,0.137625128,0.1810450852,-0.1363935322,-0.06371627,-0.4319375753,0.2392898649,0.0652496368,-0.0127885444,0.1195269376,0.2951607704,0.0438062623,0.1952228397,0.1294467449,-0.069732815,-0.0225592721,-0.2145169973,-0.0589715466,0.2264011949,-0.4359855056,0.0720824599,0.4315197766,-0.1404137909,-0.0532427132,0.0306374077,0.3919309974,0.0792137086,0.1168181524,-0.0084918672,-0.104160659,0.5898677111,0.0443217419,0.171184063,-0.0149702216,-0.1982782632,-0.3560169637,0.0049879928,0.288017571,-0.286601156,0.4284112751,0.4023602307,-0.0422602966,-0.1215295419,-0.0301021244,0.0940235928,-0.3308537602,0.055669792,0.0587296784,0.3139583766,-0.2254502922,0.149265632,-0.1737674624,-0.1033549458,-0.3849695921,0.1273514628,-0.3378459215,-0.0333239511,-0.4962185323,0.0015165606,0.285384804,-0.0028597817,0.2853520513,-0.0384451598,-0.1372202933,-0.3019890487,0.1031737626,-0.3985118568,0.5702316761,-0.10722965,-0.197017163,0.2410212457,0.1303783357,0.0852641463,-0.2976881862,-0.1983774155,-0.1698076725,-0.2090685666,0.2598606944,0.2422455698,-0.0794276446,0.1413960606,0.0924820974,0.0776791126,0.3432580829,-0.232855022,0.1028404012,-0.2014892399,-0.2819799483,-0.2043585628,0.1625683904,0.2623100579,0.0667765215,-0.1373723596,0.4017969966,0.2371831983,0.1472906321,0.4138066471,-0.1965338886,0.3094943762,-0.4334916174,0.348619163,0.5064768195,-0.3267318308,-0.4963446558,0.0994484499,-0.1563160419,0.1002884582,0.1024493128,0.151373148,0.0290875416,0.0144196171,0.1285368353,0.0187024176,-0.0005147557,0.2084205002,-0.316927731,-0.0219340678,0.5500642657,-0.2524672449,0.0284720566,-0.131236881,-0.0550430529,0.0219331458,0.0067682369,-0.1121900007,0.1218594089,0.1146048158,-0.0568789616,-0.1581844538,-0.106438823,-0.0246925242,-0.2882628739,0.2383016348,-0.1498185694,-0.0999678373,0.0617249236,-0.0970491692,-0.1620027572,-0.129275158,-0.1809165776,-0.2730642855,0.2013016045,0.2305243611,0.3202398419,-0.0729073882,0.2252094597,0.3680819571,0.0483636484,-0.2160722017,-0.4300346375,0.0856799036,-0.1777346134,-0.2188281417,-0.0748463124,0.2230703831,0.2090260535,-0.0479263812,-0.1424278915,0.1482760459,-0.2220858783,0.4189184308,-0.1370279044,0.2694525421,-0.0696570575,0.0293060318,0.016667014,0.2540052533,0.0588200353,-0.0898012966,-0.0727058947,0.4742337465,0.1814022362,0.1701161861,-0.0480709597,-0.2162680477,0.0519842356,-0.1783178449,-0.1209156215,-0.1188160628,0.1364075691,-0.0805720538,0.148888275,0.152328819,-0.1069871038,0.3120084703,0.7600834966,0.1969186664,-0.0110169193,0.0157407038,-0.0572130904,-0.2177918702,0.3338562846,-0.058682885,0.31038782,0.1047316417,0.1271647215,-0.0462091751,-0.2196502686,-0.154435575,0.1246479601,-0.2143685222,-0.1919764429,0.3080338836,0.2820868194,-0.1583673358,-0.3537832797,0.0940154344,-0.0127645368,0.0261916555,-0.1580014974,0.2172367424,-0.166529417,-0.278084904,-0.2864152193,0.136156708,-0.1222038716,-0.0921654701,-0.0412631668,0.4254295826,-0.036654897,0.1951516867,-0.3149722517,0.0895192102,0.2266037613,-0.2663277388,-0.0716857836,-0.1902881712,-0.3949196339,0.042231977,-0.0952302068,-0.0929960757,0.3783293962,-0.0473160893,-0.067959927,-0.2881456017,-0.1725400686,0.1071778685,-0.0926118121,-0.1553680301,0.1672483534,-0.0313384198,-0.1011587083,-0.2987454534,0.2937971354,-0.3053128421,-0.299143523,-0.0519867688,0.0989064649,0.0301090628,-0.2803472877,-0.428624332,0.2735781968,-0.3172441423,0.0754678994,-0.1489290893,0.1862263232,0.3309345841,0.0089533916,-0.0145607088,-0.0640404522,0.1643940359,-0.4674099684,-0.1409540623,0.3225782216,0.0187610649,-0.2163045555,-0.3245913982,-0.2757247686,0.3477241099,0.1923889369,-0.5803956985,-0.2144230455,-0.048900146,0.2480322272,-0.0512586385,-0.0464955159,0.5745424032,0.0197255258,-0.179463625,-0.1672781259,-0.3318405151,0.359731406,0.4192632437,0.2494875044,-0.0111588463,0.1778150499,0.2009202093,0.6707463264,0.2019197792,-0.2590580583,0.331730634,0.1072319448,0.0749976784,-0.2034970075,-0.3130226135,0.1857604682,-0.1816030145,0.0197321847,0.2719842196,0.0062508965,-0.2801841497,0.0177421588,0.2188196182,-0.2253448963,-0.3280583322,0.2597456276,-0.3424280584,0.2409481704,0.0363704525,0.3158006966,-0.582013905,-0.2427242249,0.1946279705,-0.3359463811,0.2798449695,0.0444980972,-0.5836006999,0.1214706451,-0.5380226374,0.1819416732,0.3146281838,0.2724077702,-0.092025429,-0.2465260625,0.043347083,-0.1471314728,0.5711411834,0.0409730114,0.4085795283,0.1699264944,-0.1302466393,-0.1994734555,-0.2247125357,0.0172346737,0.2801373005,0.3264995515,0.4008480906,-0.3452449739,-0.3632731438,-0.2657663524,0.0073641897,-0.0461810417,0.007107567,-0.1258338541,0.1893848181,0.0164867323,0.1743457466,-0.0149804754,0.118302688,0.0160397142,-0.1365461946,-0.0391333476,-0.2194266319,0.1282413155,0.2264375687,0.4735942185,0.1503249407,0.0324611329,0.0332145244,0.2151578218,-0.567109406,0.2659578621,-0.0888494328,-0.0041319211,0.0746441931,0.1444383562,-0.030301474,0.337897867,-0.042524904,0.1457136869,-0.1286294758,0.0089567872,-0.5539684892,0.4854543805,0.3874592483,0.1324663013,-0.0448895507,-0.246857211,0.1734628826,0.0741201192,-0.2765311003,0.1322598755,-0.4306930304,-0.4319946468,0.1857973337,0.2094861418,0.9232475162,0.332021594,0.3719581068,0.1007037163,0.1035806388,0.2642515898,-0.2341118157,0.4092507362,-0.3693645,-0.0463177823,0.0245385505,-0.1667332053,-0.2273836434,-0.0104256496,-0.0109939184,0.3025950491,0.3240507543,0.5999364257,-0.1121197641,0.1666993201,0.2156913131,-0.1581103951,-0.2492317557,0.2010731846,-0.2923038006,0.315240562,-0.1124083251,-0.0741084665,-0.0165099371,-0.3513420522,0.0671400949,0.0609418713,-0.3365171552,0.30484128,0.1334620416,-0.2726664841,-0.2969880998,0.2550771236,0.1119868383,0.0825273544,-0.0486002751,-0.0351850241,0.4006448984,0.3504586816,-0.0068906583,-0.2015887648,0.1864127815,0.0638751984,-0.0128940148,0.0302428547,-0.0564054921,-0.3015830219,-0.2525120974,0.1482045352,0.0033722101,-0.4129621685,-0.055318471,0.0614447072,-0.0667884797,-0.0941368639,0.150880307,-0.0260680113,-0.1079400256,0.315154314,0.1173381507,-0.2430422902,0.0523580648,0.4748320878,-0.0403974019,0.0318177901,0.4154281914,-0.15281865,-0.103806369,-0.2553910613,-0.2841726243,-0.1020596772,-0.4488408566,0.1122078001,-0.219405964,-0.4019113779,0.0442324318,0.2808186412,0.0147068594,0.3184957206,-0.0116181625,-0.1249581054,-0.3769723475,-0.1037905142,-0.1065821052,0.2671778798,0.2795186639,0.284348309,-0.2958665192,0.2492261082,-0.3831097186,-0.1071424857,-0.187170282,0.43791309,-0.0391381346,-0.3655827045,-0.0091469325,0.0018549742,0.1299711019,0.1775448769,-0.63865906,-0.2926191092,-0.0909917802,0.0818608031,-0.0075877849,-0.1598218381,-0.2392821759,-0.0172140636,0.1823851168,-0.4274035096,0.0921567082,0.1512164474,0.047031749,0.2371866852,-0.1610301137,0.1214557961,0.2881679535,-0.1865271479,-0.1841488928,-0.0762282088,0.0804811493,0.2570166588,-0.0694677755,0.0204793625,-0.2104029655,0.030472707,0.0892741159,0.1427949369,0.2369287759,-0.0247526355,-0.2156609595,0.0005509924,0.1827996075,0.114120096,-0.0785531625,-0.3289754689,0.2802228332,0.2783651054,-0.1175179482,-0.0102367681,0.167707786,-0.0993882194,0.2528515756,0.2071537971,0.1075878143,0.0526108742,0.2696116865,0.1264938265,0.4050408304,-0.348328352,0.0894215927,0.0644161254,0.1016490906,0.037777856,0.3812146783,-0.1532443315,0.0899222344,0.1483831406,0.1883367747,0.2025395483,0.2628896534,0.2135414332,-0.187863946,-0.0341212004,-0.2410863489,0.1436366886,-0.142726168,0.4156598449,0.1505924016,0.1928557456,-0.1148526296,-0.4619386494,-0.2462679744,0.2353635579,-0.3245418668,-0.3092090786,-0.0486779399,-0.0801849738,0.0559291169,0.1945610195,-0.0792901143,0.2038590163,0.3428387344,-0.0086603099,-0.0137500595,-0.0843536034,-0.1706176847,0.1139787063,0.2748646736,-0.0817389637,0.4469365478,0.0108006606,-0.0515694395,-0.2344710529,0.4114173651,0.2861286402,0.1949478239,-0.1688625664,0.0639532804,0.200646624,0.0306134559,-0.3546894789,0.3286518455,0.1511448473,-0.2930275798,0.368005842,0.1237954348,-0.2321557552,0.2764032483,-0.1041960269,-0.2014673799,-0.0498889387,0.3224370182,0.0997074991,-0.039932128,-0.1949630082,0.1111204624,-0.1445716172,-0.2853394747,0.1593456715,-0.0238983259,0.4676609337,-0.2756756544,0.1172051802,-0.0618249103,0.6220863461,-0.0292275213,0.1243288815,-0.3501377404,0.1026433855,-0.3331668377,0.1984097809,-0.0838177949,0.1537422538,-0.3975837231,0.2042651922,-0.1154986024,-0.0234658923,-0.1180424616,-0.0078867814,0.147406593,0.178931728,-0.3869177699,0.1554814875,-0.0862594694,-0.0062670191,0.2917736471,-0.3766753376,0.2219908834,0.042479679,0.2069104314,0.0109865274,-0.021723574,-0.0548064671,0.3365739584,0.6232414246,0.3470292389,0.1425508261,-0.1010277867,0.0815534294,-0.347438246,0.0141040748,-0.0608881786,-0.2096496969,0.0975683033,0.5258308053,-0.2256272584,0.1319463998,-0.1780457795,0.2366945595,0.3822763264,-0.1527291089,0.0617997609,0.0482259318,-0.2598406374,-0.0051989588,-0.125866577,0.4537958205,0.0071826167,0.3238305449,-0.3592907786,-0.5417625308,0.4329102337,-0.4511258006,-0.3025252223,0.1232788339,0.238785848,0.3495687544,-0.0600003339,-0.5436675549,-0.1724909097,0.2691140175,-0.0306607224,-0.1570452452,0.2076845467,-0.1486439705,-0.0410752967,-0.1167542785,0.1764688343,0.101526916,-0.2914130688,0.2436990142,-0.3701035082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/160","title":"caching in map causes same result to be returned for train, validation and test","comments":"Hi, the full example was listed in the PR above, but here is the exact link:\r\n\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/blob\/3c1aa3ca062cb23f303ca98ac40b6652b37ee971\/api-examples\/layers-classify-hf-datasets.py\r\n\r\nThe problem is coming from\r\n```\r\n if cache_file_name is None:\r\n # we create a unique hash from the function, current dataset file and the mapping args\r\n cache_kwargs = {\r\n \"with_indices\": with_indices,\r\n \"batched\": batched,\r\n \"batch_size\": batch_size,\r\n \"remove_columns\": remove_columns,\r\n \"keep_in_memory\": keep_in_memory,\r\n \"load_from_cache_file\": load_from_cache_file,\r\n \"cache_file_name\": cache_file_name,\r\n \"writer_batch_size\": writer_batch_size,\r\n \"arrow_schema\": arrow_schema,\r\n \"disable_nullable\": disable_nullable,\r\n }\r\n cache_file_name = self._get_cache_file_path(function, cache_kwargs)\r\n```\r\nThe cached value is always the same, but I was able to change that by just renaming the function each time which seems to fix the issue.","body":"hello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n","comment_length":99,"text":"caching in map causes same result to be returned for train, validation and test\nhello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n\nHi, the full example was listed in the PR above, but here is the exact link:\r\n\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/blob\/3c1aa3ca062cb23f303ca98ac40b6652b37ee971\/api-examples\/layers-classify-hf-datasets.py\r\n\r\nThe problem is coming from\r\n```\r\n if cache_file_name is None:\r\n # we create a unique hash from the function, current dataset file and the mapping args\r\n cache_kwargs = {\r\n \"with_indices\": with_indices,\r\n \"batched\": batched,\r\n \"batch_size\": batch_size,\r\n \"remove_columns\": remove_columns,\r\n \"keep_in_memory\": keep_in_memory,\r\n \"load_from_cache_file\": load_from_cache_file,\r\n \"cache_file_name\": cache_file_name,\r\n \"writer_batch_size\": writer_batch_size,\r\n \"arrow_schema\": arrow_schema,\r\n \"disable_nullable\": disable_nullable,\r\n }\r\n cache_file_name = self._get_cache_file_path(function, cache_kwargs)\r\n```\r\nThe cached value is always the same, but I was able to change that by just renaming the function each time which seems to fix the issue.","embeddings":[-0.1263227314,-0.2207089961,-0.100632146,0.2780107558,0.2454756796,-0.2052229047,0.1372107863,0.5128592849,0.1882870793,-0.0623357445,0.1278197765,0.3891343772,0.0037762846,-0.1606175303,-0.0107436497,0.0242125317,0.0960802659,0.2350524068,-0.1455620974,-0.1648233086,0.081894055,0.2129840404,-0.0424102098,0.1165268347,-0.3903492093,0.0412896834,0.2123562694,-0.2308897674,0.0684574693,-0.2570649683,0.1678505987,-0.0932429209,0.0258333106,0.1663175523,-0.0001097382,0.0067983456,-0.0153878154,-0.1385697424,-0.0072745453,0.0401948094,0.1181339324,-0.2117293179,-0.1091826707,-0.3121318519,-0.3320115805,-0.0376609191,0.0178094879,-0.3556600213,0.3128580153,0.1474524736,0.2607103586,0.2249089628,-0.2538127303,0.2111362219,0.1451275945,-0.0293276701,-0.174197495,-0.042305138,-0.043978855,-0.3344650865,-0.3767647743,0.5254573822,-0.0765818357,0.1701412946,0.2146666646,0.2250556946,0.1726054549,0.03593757,0.1429362446,-0.1139624342,0.0110948198,-0.2269175202,-0.0271165781,-0.3733721972,-0.4186404347,-0.0785259828,0.137625128,0.1810450852,-0.1363935322,-0.06371627,-0.4319375753,0.2392898649,0.0652496368,-0.0127885444,0.1195269376,0.2951607704,0.0438062623,0.1952228397,0.1294467449,-0.069732815,-0.0225592721,-0.2145169973,-0.0589715466,0.2264011949,-0.4359855056,0.0720824599,0.4315197766,-0.1404137909,-0.0532427132,0.0306374077,0.3919309974,0.0792137086,0.1168181524,-0.0084918672,-0.104160659,0.5898677111,0.0443217419,0.171184063,-0.0149702216,-0.1982782632,-0.3560169637,0.0049879928,0.288017571,-0.286601156,0.4284112751,0.4023602307,-0.0422602966,-0.1215295419,-0.0301021244,0.0940235928,-0.3308537602,0.055669792,0.0587296784,0.3139583766,-0.2254502922,0.149265632,-0.1737674624,-0.1033549458,-0.3849695921,0.1273514628,-0.3378459215,-0.0333239511,-0.4962185323,0.0015165606,0.285384804,-0.0028597817,0.2853520513,-0.0384451598,-0.1372202933,-0.3019890487,0.1031737626,-0.3985118568,0.5702316761,-0.10722965,-0.197017163,0.2410212457,0.1303783357,0.0852641463,-0.2976881862,-0.1983774155,-0.1698076725,-0.2090685666,0.2598606944,0.2422455698,-0.0794276446,0.1413960606,0.0924820974,0.0776791126,0.3432580829,-0.232855022,0.1028404012,-0.2014892399,-0.2819799483,-0.2043585628,0.1625683904,0.2623100579,0.0667765215,-0.1373723596,0.4017969966,0.2371831983,0.1472906321,0.4138066471,-0.1965338886,0.3094943762,-0.4334916174,0.348619163,0.5064768195,-0.3267318308,-0.4963446558,0.0994484499,-0.1563160419,0.1002884582,0.1024493128,0.151373148,0.0290875416,0.0144196171,0.1285368353,0.0187024176,-0.0005147557,0.2084205002,-0.316927731,-0.0219340678,0.5500642657,-0.2524672449,0.0284720566,-0.131236881,-0.0550430529,0.0219331458,0.0067682369,-0.1121900007,0.1218594089,0.1146048158,-0.0568789616,-0.1581844538,-0.106438823,-0.0246925242,-0.2882628739,0.2383016348,-0.1498185694,-0.0999678373,0.0617249236,-0.0970491692,-0.1620027572,-0.129275158,-0.1809165776,-0.2730642855,0.2013016045,0.2305243611,0.3202398419,-0.0729073882,0.2252094597,0.3680819571,0.0483636484,-0.2160722017,-0.4300346375,0.0856799036,-0.1777346134,-0.2188281417,-0.0748463124,0.2230703831,0.2090260535,-0.0479263812,-0.1424278915,0.1482760459,-0.2220858783,0.4189184308,-0.1370279044,0.2694525421,-0.0696570575,0.0293060318,0.016667014,0.2540052533,0.0588200353,-0.0898012966,-0.0727058947,0.4742337465,0.1814022362,0.1701161861,-0.0480709597,-0.2162680477,0.0519842356,-0.1783178449,-0.1209156215,-0.1188160628,0.1364075691,-0.0805720538,0.148888275,0.152328819,-0.1069871038,0.3120084703,0.7600834966,0.1969186664,-0.0110169193,0.0157407038,-0.0572130904,-0.2177918702,0.3338562846,-0.058682885,0.31038782,0.1047316417,0.1271647215,-0.0462091751,-0.2196502686,-0.154435575,0.1246479601,-0.2143685222,-0.1919764429,0.3080338836,0.2820868194,-0.1583673358,-0.3537832797,0.0940154344,-0.0127645368,0.0261916555,-0.1580014974,0.2172367424,-0.166529417,-0.278084904,-0.2864152193,0.136156708,-0.1222038716,-0.0921654701,-0.0412631668,0.4254295826,-0.036654897,0.1951516867,-0.3149722517,0.0895192102,0.2266037613,-0.2663277388,-0.0716857836,-0.1902881712,-0.3949196339,0.042231977,-0.0952302068,-0.0929960757,0.3783293962,-0.0473160893,-0.067959927,-0.2881456017,-0.1725400686,0.1071778685,-0.0926118121,-0.1553680301,0.1672483534,-0.0313384198,-0.1011587083,-0.2987454534,0.2937971354,-0.3053128421,-0.299143523,-0.0519867688,0.0989064649,0.0301090628,-0.2803472877,-0.428624332,0.2735781968,-0.3172441423,0.0754678994,-0.1489290893,0.1862263232,0.3309345841,0.0089533916,-0.0145607088,-0.0640404522,0.1643940359,-0.4674099684,-0.1409540623,0.3225782216,0.0187610649,-0.2163045555,-0.3245913982,-0.2757247686,0.3477241099,0.1923889369,-0.5803956985,-0.2144230455,-0.048900146,0.2480322272,-0.0512586385,-0.0464955159,0.5745424032,0.0197255258,-0.179463625,-0.1672781259,-0.3318405151,0.359731406,0.4192632437,0.2494875044,-0.0111588463,0.1778150499,0.2009202093,0.6707463264,0.2019197792,-0.2590580583,0.331730634,0.1072319448,0.0749976784,-0.2034970075,-0.3130226135,0.1857604682,-0.1816030145,0.0197321847,0.2719842196,0.0062508965,-0.2801841497,0.0177421588,0.2188196182,-0.2253448963,-0.3280583322,0.2597456276,-0.3424280584,0.2409481704,0.0363704525,0.3158006966,-0.582013905,-0.2427242249,0.1946279705,-0.3359463811,0.2798449695,0.0444980972,-0.5836006999,0.1214706451,-0.5380226374,0.1819416732,0.3146281838,0.2724077702,-0.092025429,-0.2465260625,0.043347083,-0.1471314728,0.5711411834,0.0409730114,0.4085795283,0.1699264944,-0.1302466393,-0.1994734555,-0.2247125357,0.0172346737,0.2801373005,0.3264995515,0.4008480906,-0.3452449739,-0.3632731438,-0.2657663524,0.0073641897,-0.0461810417,0.007107567,-0.1258338541,0.1893848181,0.0164867323,0.1743457466,-0.0149804754,0.118302688,0.0160397142,-0.1365461946,-0.0391333476,-0.2194266319,0.1282413155,0.2264375687,0.4735942185,0.1503249407,0.0324611329,0.0332145244,0.2151578218,-0.567109406,0.2659578621,-0.0888494328,-0.0041319211,0.0746441931,0.1444383562,-0.030301474,0.337897867,-0.042524904,0.1457136869,-0.1286294758,0.0089567872,-0.5539684892,0.4854543805,0.3874592483,0.1324663013,-0.0448895507,-0.246857211,0.1734628826,0.0741201192,-0.2765311003,0.1322598755,-0.4306930304,-0.4319946468,0.1857973337,0.2094861418,0.9232475162,0.332021594,0.3719581068,0.1007037163,0.1035806388,0.2642515898,-0.2341118157,0.4092507362,-0.3693645,-0.0463177823,0.0245385505,-0.1667332053,-0.2273836434,-0.0104256496,-0.0109939184,0.3025950491,0.3240507543,0.5999364257,-0.1121197641,0.1666993201,0.2156913131,-0.1581103951,-0.2492317557,0.2010731846,-0.2923038006,0.315240562,-0.1124083251,-0.0741084665,-0.0165099371,-0.3513420522,0.0671400949,0.0609418713,-0.3365171552,0.30484128,0.1334620416,-0.2726664841,-0.2969880998,0.2550771236,0.1119868383,0.0825273544,-0.0486002751,-0.0351850241,0.4006448984,0.3504586816,-0.0068906583,-0.2015887648,0.1864127815,0.0638751984,-0.0128940148,0.0302428547,-0.0564054921,-0.3015830219,-0.2525120974,0.1482045352,0.0033722101,-0.4129621685,-0.055318471,0.0614447072,-0.0667884797,-0.0941368639,0.150880307,-0.0260680113,-0.1079400256,0.315154314,0.1173381507,-0.2430422902,0.0523580648,0.4748320878,-0.0403974019,0.0318177901,0.4154281914,-0.15281865,-0.103806369,-0.2553910613,-0.2841726243,-0.1020596772,-0.4488408566,0.1122078001,-0.219405964,-0.4019113779,0.0442324318,0.2808186412,0.0147068594,0.3184957206,-0.0116181625,-0.1249581054,-0.3769723475,-0.1037905142,-0.1065821052,0.2671778798,0.2795186639,0.284348309,-0.2958665192,0.2492261082,-0.3831097186,-0.1071424857,-0.187170282,0.43791309,-0.0391381346,-0.3655827045,-0.0091469325,0.0018549742,0.1299711019,0.1775448769,-0.63865906,-0.2926191092,-0.0909917802,0.0818608031,-0.0075877849,-0.1598218381,-0.2392821759,-0.0172140636,0.1823851168,-0.4274035096,0.0921567082,0.1512164474,0.047031749,0.2371866852,-0.1610301137,0.1214557961,0.2881679535,-0.1865271479,-0.1841488928,-0.0762282088,0.0804811493,0.2570166588,-0.0694677755,0.0204793625,-0.2104029655,0.030472707,0.0892741159,0.1427949369,0.2369287759,-0.0247526355,-0.2156609595,0.0005509924,0.1827996075,0.114120096,-0.0785531625,-0.3289754689,0.2802228332,0.2783651054,-0.1175179482,-0.0102367681,0.167707786,-0.0993882194,0.2528515756,0.2071537971,0.1075878143,0.0526108742,0.2696116865,0.1264938265,0.4050408304,-0.348328352,0.0894215927,0.0644161254,0.1016490906,0.037777856,0.3812146783,-0.1532443315,0.0899222344,0.1483831406,0.1883367747,0.2025395483,0.2628896534,0.2135414332,-0.187863946,-0.0341212004,-0.2410863489,0.1436366886,-0.142726168,0.4156598449,0.1505924016,0.1928557456,-0.1148526296,-0.4619386494,-0.2462679744,0.2353635579,-0.3245418668,-0.3092090786,-0.0486779399,-0.0801849738,0.0559291169,0.1945610195,-0.0792901143,0.2038590163,0.3428387344,-0.0086603099,-0.0137500595,-0.0843536034,-0.1706176847,0.1139787063,0.2748646736,-0.0817389637,0.4469365478,0.0108006606,-0.0515694395,-0.2344710529,0.4114173651,0.2861286402,0.1949478239,-0.1688625664,0.0639532804,0.200646624,0.0306134559,-0.3546894789,0.3286518455,0.1511448473,-0.2930275798,0.368005842,0.1237954348,-0.2321557552,0.2764032483,-0.1041960269,-0.2014673799,-0.0498889387,0.3224370182,0.0997074991,-0.039932128,-0.1949630082,0.1111204624,-0.1445716172,-0.2853394747,0.1593456715,-0.0238983259,0.4676609337,-0.2756756544,0.1172051802,-0.0618249103,0.6220863461,-0.0292275213,0.1243288815,-0.3501377404,0.1026433855,-0.3331668377,0.1984097809,-0.0838177949,0.1537422538,-0.3975837231,0.2042651922,-0.1154986024,-0.0234658923,-0.1180424616,-0.0078867814,0.147406593,0.178931728,-0.3869177699,0.1554814875,-0.0862594694,-0.0062670191,0.2917736471,-0.3766753376,0.2219908834,0.042479679,0.2069104314,0.0109865274,-0.021723574,-0.0548064671,0.3365739584,0.6232414246,0.3470292389,0.1425508261,-0.1010277867,0.0815534294,-0.347438246,0.0141040748,-0.0608881786,-0.2096496969,0.0975683033,0.5258308053,-0.2256272584,0.1319463998,-0.1780457795,0.2366945595,0.3822763264,-0.1527291089,0.0617997609,0.0482259318,-0.2598406374,-0.0051989588,-0.125866577,0.4537958205,0.0071826167,0.3238305449,-0.3592907786,-0.5417625308,0.4329102337,-0.4511258006,-0.3025252223,0.1232788339,0.238785848,0.3495687544,-0.0600003339,-0.5436675549,-0.1724909097,0.2691140175,-0.0306607224,-0.1570452452,0.2076845467,-0.1486439705,-0.0410752967,-0.1167542785,0.1764688343,0.101526916,-0.2914130688,0.2436990142,-0.3701035082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/160","title":"caching in map causes same result to be returned for train, validation and test","comments":"Ok, I think @lhoestq has already found a solution :-) Maybe you can chime in @lhoestq ","body":"hello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n","comment_length":16,"text":"caching in map causes same result to be returned for train, validation and test\nhello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n\nOk, I think @lhoestq has already found a solution :-) Maybe you can chime in @lhoestq ","embeddings":[-0.1263227314,-0.2207089961,-0.100632146,0.2780107558,0.2454756796,-0.2052229047,0.1372107863,0.5128592849,0.1882870793,-0.0623357445,0.1278197765,0.3891343772,0.0037762846,-0.1606175303,-0.0107436497,0.0242125317,0.0960802659,0.2350524068,-0.1455620974,-0.1648233086,0.081894055,0.2129840404,-0.0424102098,0.1165268347,-0.3903492093,0.0412896834,0.2123562694,-0.2308897674,0.0684574693,-0.2570649683,0.1678505987,-0.0932429209,0.0258333106,0.1663175523,-0.0001097382,0.0067983456,-0.0153878154,-0.1385697424,-0.0072745453,0.0401948094,0.1181339324,-0.2117293179,-0.1091826707,-0.3121318519,-0.3320115805,-0.0376609191,0.0178094879,-0.3556600213,0.3128580153,0.1474524736,0.2607103586,0.2249089628,-0.2538127303,0.2111362219,0.1451275945,-0.0293276701,-0.174197495,-0.042305138,-0.043978855,-0.3344650865,-0.3767647743,0.5254573822,-0.0765818357,0.1701412946,0.2146666646,0.2250556946,0.1726054549,0.03593757,0.1429362446,-0.1139624342,0.0110948198,-0.2269175202,-0.0271165781,-0.3733721972,-0.4186404347,-0.0785259828,0.137625128,0.1810450852,-0.1363935322,-0.06371627,-0.4319375753,0.2392898649,0.0652496368,-0.0127885444,0.1195269376,0.2951607704,0.0438062623,0.1952228397,0.1294467449,-0.069732815,-0.0225592721,-0.2145169973,-0.0589715466,0.2264011949,-0.4359855056,0.0720824599,0.4315197766,-0.1404137909,-0.0532427132,0.0306374077,0.3919309974,0.0792137086,0.1168181524,-0.0084918672,-0.104160659,0.5898677111,0.0443217419,0.171184063,-0.0149702216,-0.1982782632,-0.3560169637,0.0049879928,0.288017571,-0.286601156,0.4284112751,0.4023602307,-0.0422602966,-0.1215295419,-0.0301021244,0.0940235928,-0.3308537602,0.055669792,0.0587296784,0.3139583766,-0.2254502922,0.149265632,-0.1737674624,-0.1033549458,-0.3849695921,0.1273514628,-0.3378459215,-0.0333239511,-0.4962185323,0.0015165606,0.285384804,-0.0028597817,0.2853520513,-0.0384451598,-0.1372202933,-0.3019890487,0.1031737626,-0.3985118568,0.5702316761,-0.10722965,-0.197017163,0.2410212457,0.1303783357,0.0852641463,-0.2976881862,-0.1983774155,-0.1698076725,-0.2090685666,0.2598606944,0.2422455698,-0.0794276446,0.1413960606,0.0924820974,0.0776791126,0.3432580829,-0.232855022,0.1028404012,-0.2014892399,-0.2819799483,-0.2043585628,0.1625683904,0.2623100579,0.0667765215,-0.1373723596,0.4017969966,0.2371831983,0.1472906321,0.4138066471,-0.1965338886,0.3094943762,-0.4334916174,0.348619163,0.5064768195,-0.3267318308,-0.4963446558,0.0994484499,-0.1563160419,0.1002884582,0.1024493128,0.151373148,0.0290875416,0.0144196171,0.1285368353,0.0187024176,-0.0005147557,0.2084205002,-0.316927731,-0.0219340678,0.5500642657,-0.2524672449,0.0284720566,-0.131236881,-0.0550430529,0.0219331458,0.0067682369,-0.1121900007,0.1218594089,0.1146048158,-0.0568789616,-0.1581844538,-0.106438823,-0.0246925242,-0.2882628739,0.2383016348,-0.1498185694,-0.0999678373,0.0617249236,-0.0970491692,-0.1620027572,-0.129275158,-0.1809165776,-0.2730642855,0.2013016045,0.2305243611,0.3202398419,-0.0729073882,0.2252094597,0.3680819571,0.0483636484,-0.2160722017,-0.4300346375,0.0856799036,-0.1777346134,-0.2188281417,-0.0748463124,0.2230703831,0.2090260535,-0.0479263812,-0.1424278915,0.1482760459,-0.2220858783,0.4189184308,-0.1370279044,0.2694525421,-0.0696570575,0.0293060318,0.016667014,0.2540052533,0.0588200353,-0.0898012966,-0.0727058947,0.4742337465,0.1814022362,0.1701161861,-0.0480709597,-0.2162680477,0.0519842356,-0.1783178449,-0.1209156215,-0.1188160628,0.1364075691,-0.0805720538,0.148888275,0.152328819,-0.1069871038,0.3120084703,0.7600834966,0.1969186664,-0.0110169193,0.0157407038,-0.0572130904,-0.2177918702,0.3338562846,-0.058682885,0.31038782,0.1047316417,0.1271647215,-0.0462091751,-0.2196502686,-0.154435575,0.1246479601,-0.2143685222,-0.1919764429,0.3080338836,0.2820868194,-0.1583673358,-0.3537832797,0.0940154344,-0.0127645368,0.0261916555,-0.1580014974,0.2172367424,-0.166529417,-0.278084904,-0.2864152193,0.136156708,-0.1222038716,-0.0921654701,-0.0412631668,0.4254295826,-0.036654897,0.1951516867,-0.3149722517,0.0895192102,0.2266037613,-0.2663277388,-0.0716857836,-0.1902881712,-0.3949196339,0.042231977,-0.0952302068,-0.0929960757,0.3783293962,-0.0473160893,-0.067959927,-0.2881456017,-0.1725400686,0.1071778685,-0.0926118121,-0.1553680301,0.1672483534,-0.0313384198,-0.1011587083,-0.2987454534,0.2937971354,-0.3053128421,-0.299143523,-0.0519867688,0.0989064649,0.0301090628,-0.2803472877,-0.428624332,0.2735781968,-0.3172441423,0.0754678994,-0.1489290893,0.1862263232,0.3309345841,0.0089533916,-0.0145607088,-0.0640404522,0.1643940359,-0.4674099684,-0.1409540623,0.3225782216,0.0187610649,-0.2163045555,-0.3245913982,-0.2757247686,0.3477241099,0.1923889369,-0.5803956985,-0.2144230455,-0.048900146,0.2480322272,-0.0512586385,-0.0464955159,0.5745424032,0.0197255258,-0.179463625,-0.1672781259,-0.3318405151,0.359731406,0.4192632437,0.2494875044,-0.0111588463,0.1778150499,0.2009202093,0.6707463264,0.2019197792,-0.2590580583,0.331730634,0.1072319448,0.0749976784,-0.2034970075,-0.3130226135,0.1857604682,-0.1816030145,0.0197321847,0.2719842196,0.0062508965,-0.2801841497,0.0177421588,0.2188196182,-0.2253448963,-0.3280583322,0.2597456276,-0.3424280584,0.2409481704,0.0363704525,0.3158006966,-0.582013905,-0.2427242249,0.1946279705,-0.3359463811,0.2798449695,0.0444980972,-0.5836006999,0.1214706451,-0.5380226374,0.1819416732,0.3146281838,0.2724077702,-0.092025429,-0.2465260625,0.043347083,-0.1471314728,0.5711411834,0.0409730114,0.4085795283,0.1699264944,-0.1302466393,-0.1994734555,-0.2247125357,0.0172346737,0.2801373005,0.3264995515,0.4008480906,-0.3452449739,-0.3632731438,-0.2657663524,0.0073641897,-0.0461810417,0.007107567,-0.1258338541,0.1893848181,0.0164867323,0.1743457466,-0.0149804754,0.118302688,0.0160397142,-0.1365461946,-0.0391333476,-0.2194266319,0.1282413155,0.2264375687,0.4735942185,0.1503249407,0.0324611329,0.0332145244,0.2151578218,-0.567109406,0.2659578621,-0.0888494328,-0.0041319211,0.0746441931,0.1444383562,-0.030301474,0.337897867,-0.042524904,0.1457136869,-0.1286294758,0.0089567872,-0.5539684892,0.4854543805,0.3874592483,0.1324663013,-0.0448895507,-0.246857211,0.1734628826,0.0741201192,-0.2765311003,0.1322598755,-0.4306930304,-0.4319946468,0.1857973337,0.2094861418,0.9232475162,0.332021594,0.3719581068,0.1007037163,0.1035806388,0.2642515898,-0.2341118157,0.4092507362,-0.3693645,-0.0463177823,0.0245385505,-0.1667332053,-0.2273836434,-0.0104256496,-0.0109939184,0.3025950491,0.3240507543,0.5999364257,-0.1121197641,0.1666993201,0.2156913131,-0.1581103951,-0.2492317557,0.2010731846,-0.2923038006,0.315240562,-0.1124083251,-0.0741084665,-0.0165099371,-0.3513420522,0.0671400949,0.0609418713,-0.3365171552,0.30484128,0.1334620416,-0.2726664841,-0.2969880998,0.2550771236,0.1119868383,0.0825273544,-0.0486002751,-0.0351850241,0.4006448984,0.3504586816,-0.0068906583,-0.2015887648,0.1864127815,0.0638751984,-0.0128940148,0.0302428547,-0.0564054921,-0.3015830219,-0.2525120974,0.1482045352,0.0033722101,-0.4129621685,-0.055318471,0.0614447072,-0.0667884797,-0.0941368639,0.150880307,-0.0260680113,-0.1079400256,0.315154314,0.1173381507,-0.2430422902,0.0523580648,0.4748320878,-0.0403974019,0.0318177901,0.4154281914,-0.15281865,-0.103806369,-0.2553910613,-0.2841726243,-0.1020596772,-0.4488408566,0.1122078001,-0.219405964,-0.4019113779,0.0442324318,0.2808186412,0.0147068594,0.3184957206,-0.0116181625,-0.1249581054,-0.3769723475,-0.1037905142,-0.1065821052,0.2671778798,0.2795186639,0.284348309,-0.2958665192,0.2492261082,-0.3831097186,-0.1071424857,-0.187170282,0.43791309,-0.0391381346,-0.3655827045,-0.0091469325,0.0018549742,0.1299711019,0.1775448769,-0.63865906,-0.2926191092,-0.0909917802,0.0818608031,-0.0075877849,-0.1598218381,-0.2392821759,-0.0172140636,0.1823851168,-0.4274035096,0.0921567082,0.1512164474,0.047031749,0.2371866852,-0.1610301137,0.1214557961,0.2881679535,-0.1865271479,-0.1841488928,-0.0762282088,0.0804811493,0.2570166588,-0.0694677755,0.0204793625,-0.2104029655,0.030472707,0.0892741159,0.1427949369,0.2369287759,-0.0247526355,-0.2156609595,0.0005509924,0.1827996075,0.114120096,-0.0785531625,-0.3289754689,0.2802228332,0.2783651054,-0.1175179482,-0.0102367681,0.167707786,-0.0993882194,0.2528515756,0.2071537971,0.1075878143,0.0526108742,0.2696116865,0.1264938265,0.4050408304,-0.348328352,0.0894215927,0.0644161254,0.1016490906,0.037777856,0.3812146783,-0.1532443315,0.0899222344,0.1483831406,0.1883367747,0.2025395483,0.2628896534,0.2135414332,-0.187863946,-0.0341212004,-0.2410863489,0.1436366886,-0.142726168,0.4156598449,0.1505924016,0.1928557456,-0.1148526296,-0.4619386494,-0.2462679744,0.2353635579,-0.3245418668,-0.3092090786,-0.0486779399,-0.0801849738,0.0559291169,0.1945610195,-0.0792901143,0.2038590163,0.3428387344,-0.0086603099,-0.0137500595,-0.0843536034,-0.1706176847,0.1139787063,0.2748646736,-0.0817389637,0.4469365478,0.0108006606,-0.0515694395,-0.2344710529,0.4114173651,0.2861286402,0.1949478239,-0.1688625664,0.0639532804,0.200646624,0.0306134559,-0.3546894789,0.3286518455,0.1511448473,-0.2930275798,0.368005842,0.1237954348,-0.2321557552,0.2764032483,-0.1041960269,-0.2014673799,-0.0498889387,0.3224370182,0.0997074991,-0.039932128,-0.1949630082,0.1111204624,-0.1445716172,-0.2853394747,0.1593456715,-0.0238983259,0.4676609337,-0.2756756544,0.1172051802,-0.0618249103,0.6220863461,-0.0292275213,0.1243288815,-0.3501377404,0.1026433855,-0.3331668377,0.1984097809,-0.0838177949,0.1537422538,-0.3975837231,0.2042651922,-0.1154986024,-0.0234658923,-0.1180424616,-0.0078867814,0.147406593,0.178931728,-0.3869177699,0.1554814875,-0.0862594694,-0.0062670191,0.2917736471,-0.3766753376,0.2219908834,0.042479679,0.2069104314,0.0109865274,-0.021723574,-0.0548064671,0.3365739584,0.6232414246,0.3470292389,0.1425508261,-0.1010277867,0.0815534294,-0.347438246,0.0141040748,-0.0608881786,-0.2096496969,0.0975683033,0.5258308053,-0.2256272584,0.1319463998,-0.1780457795,0.2366945595,0.3822763264,-0.1527291089,0.0617997609,0.0482259318,-0.2598406374,-0.0051989588,-0.125866577,0.4537958205,0.0071826167,0.3238305449,-0.3592907786,-0.5417625308,0.4329102337,-0.4511258006,-0.3025252223,0.1232788339,0.238785848,0.3495687544,-0.0600003339,-0.5436675549,-0.1724909097,0.2691140175,-0.0306607224,-0.1570452452,0.2076845467,-0.1486439705,-0.0410752967,-0.1167542785,0.1764688343,0.101526916,-0.2914130688,0.2436990142,-0.3701035082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/160","title":"caching in map causes same result to be returned for train, validation and test","comments":"This fixed my issue (I think)\r\n\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/commit\/48aa8ecde4b307bd3e7dde5fe71e43a1d4769ee1","body":"hello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n","comment_length":7,"text":"caching in map causes same result to be returned for train, validation and test\nhello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n\nThis fixed my issue (I think)\r\n\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/commit\/48aa8ecde4b307bd3e7dde5fe71e43a1d4769ee1","embeddings":[-0.1263227314,-0.2207089961,-0.100632146,0.2780107558,0.2454756796,-0.2052229047,0.1372107863,0.5128592849,0.1882870793,-0.0623357445,0.1278197765,0.3891343772,0.0037762846,-0.1606175303,-0.0107436497,0.0242125317,0.0960802659,0.2350524068,-0.1455620974,-0.1648233086,0.081894055,0.2129840404,-0.0424102098,0.1165268347,-0.3903492093,0.0412896834,0.2123562694,-0.2308897674,0.0684574693,-0.2570649683,0.1678505987,-0.0932429209,0.0258333106,0.1663175523,-0.0001097382,0.0067983456,-0.0153878154,-0.1385697424,-0.0072745453,0.0401948094,0.1181339324,-0.2117293179,-0.1091826707,-0.3121318519,-0.3320115805,-0.0376609191,0.0178094879,-0.3556600213,0.3128580153,0.1474524736,0.2607103586,0.2249089628,-0.2538127303,0.2111362219,0.1451275945,-0.0293276701,-0.174197495,-0.042305138,-0.043978855,-0.3344650865,-0.3767647743,0.5254573822,-0.0765818357,0.1701412946,0.2146666646,0.2250556946,0.1726054549,0.03593757,0.1429362446,-0.1139624342,0.0110948198,-0.2269175202,-0.0271165781,-0.3733721972,-0.4186404347,-0.0785259828,0.137625128,0.1810450852,-0.1363935322,-0.06371627,-0.4319375753,0.2392898649,0.0652496368,-0.0127885444,0.1195269376,0.2951607704,0.0438062623,0.1952228397,0.1294467449,-0.069732815,-0.0225592721,-0.2145169973,-0.0589715466,0.2264011949,-0.4359855056,0.0720824599,0.4315197766,-0.1404137909,-0.0532427132,0.0306374077,0.3919309974,0.0792137086,0.1168181524,-0.0084918672,-0.104160659,0.5898677111,0.0443217419,0.171184063,-0.0149702216,-0.1982782632,-0.3560169637,0.0049879928,0.288017571,-0.286601156,0.4284112751,0.4023602307,-0.0422602966,-0.1215295419,-0.0301021244,0.0940235928,-0.3308537602,0.055669792,0.0587296784,0.3139583766,-0.2254502922,0.149265632,-0.1737674624,-0.1033549458,-0.3849695921,0.1273514628,-0.3378459215,-0.0333239511,-0.4962185323,0.0015165606,0.285384804,-0.0028597817,0.2853520513,-0.0384451598,-0.1372202933,-0.3019890487,0.1031737626,-0.3985118568,0.5702316761,-0.10722965,-0.197017163,0.2410212457,0.1303783357,0.0852641463,-0.2976881862,-0.1983774155,-0.1698076725,-0.2090685666,0.2598606944,0.2422455698,-0.0794276446,0.1413960606,0.0924820974,0.0776791126,0.3432580829,-0.232855022,0.1028404012,-0.2014892399,-0.2819799483,-0.2043585628,0.1625683904,0.2623100579,0.0667765215,-0.1373723596,0.4017969966,0.2371831983,0.1472906321,0.4138066471,-0.1965338886,0.3094943762,-0.4334916174,0.348619163,0.5064768195,-0.3267318308,-0.4963446558,0.0994484499,-0.1563160419,0.1002884582,0.1024493128,0.151373148,0.0290875416,0.0144196171,0.1285368353,0.0187024176,-0.0005147557,0.2084205002,-0.316927731,-0.0219340678,0.5500642657,-0.2524672449,0.0284720566,-0.131236881,-0.0550430529,0.0219331458,0.0067682369,-0.1121900007,0.1218594089,0.1146048158,-0.0568789616,-0.1581844538,-0.106438823,-0.0246925242,-0.2882628739,0.2383016348,-0.1498185694,-0.0999678373,0.0617249236,-0.0970491692,-0.1620027572,-0.129275158,-0.1809165776,-0.2730642855,0.2013016045,0.2305243611,0.3202398419,-0.0729073882,0.2252094597,0.3680819571,0.0483636484,-0.2160722017,-0.4300346375,0.0856799036,-0.1777346134,-0.2188281417,-0.0748463124,0.2230703831,0.2090260535,-0.0479263812,-0.1424278915,0.1482760459,-0.2220858783,0.4189184308,-0.1370279044,0.2694525421,-0.0696570575,0.0293060318,0.016667014,0.2540052533,0.0588200353,-0.0898012966,-0.0727058947,0.4742337465,0.1814022362,0.1701161861,-0.0480709597,-0.2162680477,0.0519842356,-0.1783178449,-0.1209156215,-0.1188160628,0.1364075691,-0.0805720538,0.148888275,0.152328819,-0.1069871038,0.3120084703,0.7600834966,0.1969186664,-0.0110169193,0.0157407038,-0.0572130904,-0.2177918702,0.3338562846,-0.058682885,0.31038782,0.1047316417,0.1271647215,-0.0462091751,-0.2196502686,-0.154435575,0.1246479601,-0.2143685222,-0.1919764429,0.3080338836,0.2820868194,-0.1583673358,-0.3537832797,0.0940154344,-0.0127645368,0.0261916555,-0.1580014974,0.2172367424,-0.166529417,-0.278084904,-0.2864152193,0.136156708,-0.1222038716,-0.0921654701,-0.0412631668,0.4254295826,-0.036654897,0.1951516867,-0.3149722517,0.0895192102,0.2266037613,-0.2663277388,-0.0716857836,-0.1902881712,-0.3949196339,0.042231977,-0.0952302068,-0.0929960757,0.3783293962,-0.0473160893,-0.067959927,-0.2881456017,-0.1725400686,0.1071778685,-0.0926118121,-0.1553680301,0.1672483534,-0.0313384198,-0.1011587083,-0.2987454534,0.2937971354,-0.3053128421,-0.299143523,-0.0519867688,0.0989064649,0.0301090628,-0.2803472877,-0.428624332,0.2735781968,-0.3172441423,0.0754678994,-0.1489290893,0.1862263232,0.3309345841,0.0089533916,-0.0145607088,-0.0640404522,0.1643940359,-0.4674099684,-0.1409540623,0.3225782216,0.0187610649,-0.2163045555,-0.3245913982,-0.2757247686,0.3477241099,0.1923889369,-0.5803956985,-0.2144230455,-0.048900146,0.2480322272,-0.0512586385,-0.0464955159,0.5745424032,0.0197255258,-0.179463625,-0.1672781259,-0.3318405151,0.359731406,0.4192632437,0.2494875044,-0.0111588463,0.1778150499,0.2009202093,0.6707463264,0.2019197792,-0.2590580583,0.331730634,0.1072319448,0.0749976784,-0.2034970075,-0.3130226135,0.1857604682,-0.1816030145,0.0197321847,0.2719842196,0.0062508965,-0.2801841497,0.0177421588,0.2188196182,-0.2253448963,-0.3280583322,0.2597456276,-0.3424280584,0.2409481704,0.0363704525,0.3158006966,-0.582013905,-0.2427242249,0.1946279705,-0.3359463811,0.2798449695,0.0444980972,-0.5836006999,0.1214706451,-0.5380226374,0.1819416732,0.3146281838,0.2724077702,-0.092025429,-0.2465260625,0.043347083,-0.1471314728,0.5711411834,0.0409730114,0.4085795283,0.1699264944,-0.1302466393,-0.1994734555,-0.2247125357,0.0172346737,0.2801373005,0.3264995515,0.4008480906,-0.3452449739,-0.3632731438,-0.2657663524,0.0073641897,-0.0461810417,0.007107567,-0.1258338541,0.1893848181,0.0164867323,0.1743457466,-0.0149804754,0.118302688,0.0160397142,-0.1365461946,-0.0391333476,-0.2194266319,0.1282413155,0.2264375687,0.4735942185,0.1503249407,0.0324611329,0.0332145244,0.2151578218,-0.567109406,0.2659578621,-0.0888494328,-0.0041319211,0.0746441931,0.1444383562,-0.030301474,0.337897867,-0.042524904,0.1457136869,-0.1286294758,0.0089567872,-0.5539684892,0.4854543805,0.3874592483,0.1324663013,-0.0448895507,-0.246857211,0.1734628826,0.0741201192,-0.2765311003,0.1322598755,-0.4306930304,-0.4319946468,0.1857973337,0.2094861418,0.9232475162,0.332021594,0.3719581068,0.1007037163,0.1035806388,0.2642515898,-0.2341118157,0.4092507362,-0.3693645,-0.0463177823,0.0245385505,-0.1667332053,-0.2273836434,-0.0104256496,-0.0109939184,0.3025950491,0.3240507543,0.5999364257,-0.1121197641,0.1666993201,0.2156913131,-0.1581103951,-0.2492317557,0.2010731846,-0.2923038006,0.315240562,-0.1124083251,-0.0741084665,-0.0165099371,-0.3513420522,0.0671400949,0.0609418713,-0.3365171552,0.30484128,0.1334620416,-0.2726664841,-0.2969880998,0.2550771236,0.1119868383,0.0825273544,-0.0486002751,-0.0351850241,0.4006448984,0.3504586816,-0.0068906583,-0.2015887648,0.1864127815,0.0638751984,-0.0128940148,0.0302428547,-0.0564054921,-0.3015830219,-0.2525120974,0.1482045352,0.0033722101,-0.4129621685,-0.055318471,0.0614447072,-0.0667884797,-0.0941368639,0.150880307,-0.0260680113,-0.1079400256,0.315154314,0.1173381507,-0.2430422902,0.0523580648,0.4748320878,-0.0403974019,0.0318177901,0.4154281914,-0.15281865,-0.103806369,-0.2553910613,-0.2841726243,-0.1020596772,-0.4488408566,0.1122078001,-0.219405964,-0.4019113779,0.0442324318,0.2808186412,0.0147068594,0.3184957206,-0.0116181625,-0.1249581054,-0.3769723475,-0.1037905142,-0.1065821052,0.2671778798,0.2795186639,0.284348309,-0.2958665192,0.2492261082,-0.3831097186,-0.1071424857,-0.187170282,0.43791309,-0.0391381346,-0.3655827045,-0.0091469325,0.0018549742,0.1299711019,0.1775448769,-0.63865906,-0.2926191092,-0.0909917802,0.0818608031,-0.0075877849,-0.1598218381,-0.2392821759,-0.0172140636,0.1823851168,-0.4274035096,0.0921567082,0.1512164474,0.047031749,0.2371866852,-0.1610301137,0.1214557961,0.2881679535,-0.1865271479,-0.1841488928,-0.0762282088,0.0804811493,0.2570166588,-0.0694677755,0.0204793625,-0.2104029655,0.030472707,0.0892741159,0.1427949369,0.2369287759,-0.0247526355,-0.2156609595,0.0005509924,0.1827996075,0.114120096,-0.0785531625,-0.3289754689,0.2802228332,0.2783651054,-0.1175179482,-0.0102367681,0.167707786,-0.0993882194,0.2528515756,0.2071537971,0.1075878143,0.0526108742,0.2696116865,0.1264938265,0.4050408304,-0.348328352,0.0894215927,0.0644161254,0.1016490906,0.037777856,0.3812146783,-0.1532443315,0.0899222344,0.1483831406,0.1883367747,0.2025395483,0.2628896534,0.2135414332,-0.187863946,-0.0341212004,-0.2410863489,0.1436366886,-0.142726168,0.4156598449,0.1505924016,0.1928557456,-0.1148526296,-0.4619386494,-0.2462679744,0.2353635579,-0.3245418668,-0.3092090786,-0.0486779399,-0.0801849738,0.0559291169,0.1945610195,-0.0792901143,0.2038590163,0.3428387344,-0.0086603099,-0.0137500595,-0.0843536034,-0.1706176847,0.1139787063,0.2748646736,-0.0817389637,0.4469365478,0.0108006606,-0.0515694395,-0.2344710529,0.4114173651,0.2861286402,0.1949478239,-0.1688625664,0.0639532804,0.200646624,0.0306134559,-0.3546894789,0.3286518455,0.1511448473,-0.2930275798,0.368005842,0.1237954348,-0.2321557552,0.2764032483,-0.1041960269,-0.2014673799,-0.0498889387,0.3224370182,0.0997074991,-0.039932128,-0.1949630082,0.1111204624,-0.1445716172,-0.2853394747,0.1593456715,-0.0238983259,0.4676609337,-0.2756756544,0.1172051802,-0.0618249103,0.6220863461,-0.0292275213,0.1243288815,-0.3501377404,0.1026433855,-0.3331668377,0.1984097809,-0.0838177949,0.1537422538,-0.3975837231,0.2042651922,-0.1154986024,-0.0234658923,-0.1180424616,-0.0078867814,0.147406593,0.178931728,-0.3869177699,0.1554814875,-0.0862594694,-0.0062670191,0.2917736471,-0.3766753376,0.2219908834,0.042479679,0.2069104314,0.0109865274,-0.021723574,-0.0548064671,0.3365739584,0.6232414246,0.3470292389,0.1425508261,-0.1010277867,0.0815534294,-0.347438246,0.0141040748,-0.0608881786,-0.2096496969,0.0975683033,0.5258308053,-0.2256272584,0.1319463998,-0.1780457795,0.2366945595,0.3822763264,-0.1527291089,0.0617997609,0.0482259318,-0.2598406374,-0.0051989588,-0.125866577,0.4537958205,0.0071826167,0.3238305449,-0.3592907786,-0.5417625308,0.4329102337,-0.4511258006,-0.3025252223,0.1232788339,0.238785848,0.3495687544,-0.0600003339,-0.5436675549,-0.1724909097,0.2691140175,-0.0306607224,-0.1570452452,0.2076845467,-0.1486439705,-0.0410752967,-0.1167542785,0.1764688343,0.101526916,-0.2914130688,0.2436990142,-0.3701035082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/160","title":"caching in map causes same result to be returned for train, validation and test","comments":"> Ok, I think @lhoestq has already found a solution :-) Maybe you can chime in @lhoestq\r\n\r\nOh, awesome! I see the PR, Ill check it out","body":"hello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n","comment_length":27,"text":"caching in map causes same result to be returned for train, validation and test\nhello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n\n> Ok, I think @lhoestq has already found a solution :-) Maybe you can chime in @lhoestq\r\n\r\nOh, awesome! I see the PR, Ill check it out","embeddings":[-0.1263227314,-0.2207089961,-0.100632146,0.2780107558,0.2454756796,-0.2052229047,0.1372107863,0.5128592849,0.1882870793,-0.0623357445,0.1278197765,0.3891343772,0.0037762846,-0.1606175303,-0.0107436497,0.0242125317,0.0960802659,0.2350524068,-0.1455620974,-0.1648233086,0.081894055,0.2129840404,-0.0424102098,0.1165268347,-0.3903492093,0.0412896834,0.2123562694,-0.2308897674,0.0684574693,-0.2570649683,0.1678505987,-0.0932429209,0.0258333106,0.1663175523,-0.0001097382,0.0067983456,-0.0153878154,-0.1385697424,-0.0072745453,0.0401948094,0.1181339324,-0.2117293179,-0.1091826707,-0.3121318519,-0.3320115805,-0.0376609191,0.0178094879,-0.3556600213,0.3128580153,0.1474524736,0.2607103586,0.2249089628,-0.2538127303,0.2111362219,0.1451275945,-0.0293276701,-0.174197495,-0.042305138,-0.043978855,-0.3344650865,-0.3767647743,0.5254573822,-0.0765818357,0.1701412946,0.2146666646,0.2250556946,0.1726054549,0.03593757,0.1429362446,-0.1139624342,0.0110948198,-0.2269175202,-0.0271165781,-0.3733721972,-0.4186404347,-0.0785259828,0.137625128,0.1810450852,-0.1363935322,-0.06371627,-0.4319375753,0.2392898649,0.0652496368,-0.0127885444,0.1195269376,0.2951607704,0.0438062623,0.1952228397,0.1294467449,-0.069732815,-0.0225592721,-0.2145169973,-0.0589715466,0.2264011949,-0.4359855056,0.0720824599,0.4315197766,-0.1404137909,-0.0532427132,0.0306374077,0.3919309974,0.0792137086,0.1168181524,-0.0084918672,-0.104160659,0.5898677111,0.0443217419,0.171184063,-0.0149702216,-0.1982782632,-0.3560169637,0.0049879928,0.288017571,-0.286601156,0.4284112751,0.4023602307,-0.0422602966,-0.1215295419,-0.0301021244,0.0940235928,-0.3308537602,0.055669792,0.0587296784,0.3139583766,-0.2254502922,0.149265632,-0.1737674624,-0.1033549458,-0.3849695921,0.1273514628,-0.3378459215,-0.0333239511,-0.4962185323,0.0015165606,0.285384804,-0.0028597817,0.2853520513,-0.0384451598,-0.1372202933,-0.3019890487,0.1031737626,-0.3985118568,0.5702316761,-0.10722965,-0.197017163,0.2410212457,0.1303783357,0.0852641463,-0.2976881862,-0.1983774155,-0.1698076725,-0.2090685666,0.2598606944,0.2422455698,-0.0794276446,0.1413960606,0.0924820974,0.0776791126,0.3432580829,-0.232855022,0.1028404012,-0.2014892399,-0.2819799483,-0.2043585628,0.1625683904,0.2623100579,0.0667765215,-0.1373723596,0.4017969966,0.2371831983,0.1472906321,0.4138066471,-0.1965338886,0.3094943762,-0.4334916174,0.348619163,0.5064768195,-0.3267318308,-0.4963446558,0.0994484499,-0.1563160419,0.1002884582,0.1024493128,0.151373148,0.0290875416,0.0144196171,0.1285368353,0.0187024176,-0.0005147557,0.2084205002,-0.316927731,-0.0219340678,0.5500642657,-0.2524672449,0.0284720566,-0.131236881,-0.0550430529,0.0219331458,0.0067682369,-0.1121900007,0.1218594089,0.1146048158,-0.0568789616,-0.1581844538,-0.106438823,-0.0246925242,-0.2882628739,0.2383016348,-0.1498185694,-0.0999678373,0.0617249236,-0.0970491692,-0.1620027572,-0.129275158,-0.1809165776,-0.2730642855,0.2013016045,0.2305243611,0.3202398419,-0.0729073882,0.2252094597,0.3680819571,0.0483636484,-0.2160722017,-0.4300346375,0.0856799036,-0.1777346134,-0.2188281417,-0.0748463124,0.2230703831,0.2090260535,-0.0479263812,-0.1424278915,0.1482760459,-0.2220858783,0.4189184308,-0.1370279044,0.2694525421,-0.0696570575,0.0293060318,0.016667014,0.2540052533,0.0588200353,-0.0898012966,-0.0727058947,0.4742337465,0.1814022362,0.1701161861,-0.0480709597,-0.2162680477,0.0519842356,-0.1783178449,-0.1209156215,-0.1188160628,0.1364075691,-0.0805720538,0.148888275,0.152328819,-0.1069871038,0.3120084703,0.7600834966,0.1969186664,-0.0110169193,0.0157407038,-0.0572130904,-0.2177918702,0.3338562846,-0.058682885,0.31038782,0.1047316417,0.1271647215,-0.0462091751,-0.2196502686,-0.154435575,0.1246479601,-0.2143685222,-0.1919764429,0.3080338836,0.2820868194,-0.1583673358,-0.3537832797,0.0940154344,-0.0127645368,0.0261916555,-0.1580014974,0.2172367424,-0.166529417,-0.278084904,-0.2864152193,0.136156708,-0.1222038716,-0.0921654701,-0.0412631668,0.4254295826,-0.036654897,0.1951516867,-0.3149722517,0.0895192102,0.2266037613,-0.2663277388,-0.0716857836,-0.1902881712,-0.3949196339,0.042231977,-0.0952302068,-0.0929960757,0.3783293962,-0.0473160893,-0.067959927,-0.2881456017,-0.1725400686,0.1071778685,-0.0926118121,-0.1553680301,0.1672483534,-0.0313384198,-0.1011587083,-0.2987454534,0.2937971354,-0.3053128421,-0.299143523,-0.0519867688,0.0989064649,0.0301090628,-0.2803472877,-0.428624332,0.2735781968,-0.3172441423,0.0754678994,-0.1489290893,0.1862263232,0.3309345841,0.0089533916,-0.0145607088,-0.0640404522,0.1643940359,-0.4674099684,-0.1409540623,0.3225782216,0.0187610649,-0.2163045555,-0.3245913982,-0.2757247686,0.3477241099,0.1923889369,-0.5803956985,-0.2144230455,-0.048900146,0.2480322272,-0.0512586385,-0.0464955159,0.5745424032,0.0197255258,-0.179463625,-0.1672781259,-0.3318405151,0.359731406,0.4192632437,0.2494875044,-0.0111588463,0.1778150499,0.2009202093,0.6707463264,0.2019197792,-0.2590580583,0.331730634,0.1072319448,0.0749976784,-0.2034970075,-0.3130226135,0.1857604682,-0.1816030145,0.0197321847,0.2719842196,0.0062508965,-0.2801841497,0.0177421588,0.2188196182,-0.2253448963,-0.3280583322,0.2597456276,-0.3424280584,0.2409481704,0.0363704525,0.3158006966,-0.582013905,-0.2427242249,0.1946279705,-0.3359463811,0.2798449695,0.0444980972,-0.5836006999,0.1214706451,-0.5380226374,0.1819416732,0.3146281838,0.2724077702,-0.092025429,-0.2465260625,0.043347083,-0.1471314728,0.5711411834,0.0409730114,0.4085795283,0.1699264944,-0.1302466393,-0.1994734555,-0.2247125357,0.0172346737,0.2801373005,0.3264995515,0.4008480906,-0.3452449739,-0.3632731438,-0.2657663524,0.0073641897,-0.0461810417,0.007107567,-0.1258338541,0.1893848181,0.0164867323,0.1743457466,-0.0149804754,0.118302688,0.0160397142,-0.1365461946,-0.0391333476,-0.2194266319,0.1282413155,0.2264375687,0.4735942185,0.1503249407,0.0324611329,0.0332145244,0.2151578218,-0.567109406,0.2659578621,-0.0888494328,-0.0041319211,0.0746441931,0.1444383562,-0.030301474,0.337897867,-0.042524904,0.1457136869,-0.1286294758,0.0089567872,-0.5539684892,0.4854543805,0.3874592483,0.1324663013,-0.0448895507,-0.246857211,0.1734628826,0.0741201192,-0.2765311003,0.1322598755,-0.4306930304,-0.4319946468,0.1857973337,0.2094861418,0.9232475162,0.332021594,0.3719581068,0.1007037163,0.1035806388,0.2642515898,-0.2341118157,0.4092507362,-0.3693645,-0.0463177823,0.0245385505,-0.1667332053,-0.2273836434,-0.0104256496,-0.0109939184,0.3025950491,0.3240507543,0.5999364257,-0.1121197641,0.1666993201,0.2156913131,-0.1581103951,-0.2492317557,0.2010731846,-0.2923038006,0.315240562,-0.1124083251,-0.0741084665,-0.0165099371,-0.3513420522,0.0671400949,0.0609418713,-0.3365171552,0.30484128,0.1334620416,-0.2726664841,-0.2969880998,0.2550771236,0.1119868383,0.0825273544,-0.0486002751,-0.0351850241,0.4006448984,0.3504586816,-0.0068906583,-0.2015887648,0.1864127815,0.0638751984,-0.0128940148,0.0302428547,-0.0564054921,-0.3015830219,-0.2525120974,0.1482045352,0.0033722101,-0.4129621685,-0.055318471,0.0614447072,-0.0667884797,-0.0941368639,0.150880307,-0.0260680113,-0.1079400256,0.315154314,0.1173381507,-0.2430422902,0.0523580648,0.4748320878,-0.0403974019,0.0318177901,0.4154281914,-0.15281865,-0.103806369,-0.2553910613,-0.2841726243,-0.1020596772,-0.4488408566,0.1122078001,-0.219405964,-0.4019113779,0.0442324318,0.2808186412,0.0147068594,0.3184957206,-0.0116181625,-0.1249581054,-0.3769723475,-0.1037905142,-0.1065821052,0.2671778798,0.2795186639,0.284348309,-0.2958665192,0.2492261082,-0.3831097186,-0.1071424857,-0.187170282,0.43791309,-0.0391381346,-0.3655827045,-0.0091469325,0.0018549742,0.1299711019,0.1775448769,-0.63865906,-0.2926191092,-0.0909917802,0.0818608031,-0.0075877849,-0.1598218381,-0.2392821759,-0.0172140636,0.1823851168,-0.4274035096,0.0921567082,0.1512164474,0.047031749,0.2371866852,-0.1610301137,0.1214557961,0.2881679535,-0.1865271479,-0.1841488928,-0.0762282088,0.0804811493,0.2570166588,-0.0694677755,0.0204793625,-0.2104029655,0.030472707,0.0892741159,0.1427949369,0.2369287759,-0.0247526355,-0.2156609595,0.0005509924,0.1827996075,0.114120096,-0.0785531625,-0.3289754689,0.2802228332,0.2783651054,-0.1175179482,-0.0102367681,0.167707786,-0.0993882194,0.2528515756,0.2071537971,0.1075878143,0.0526108742,0.2696116865,0.1264938265,0.4050408304,-0.348328352,0.0894215927,0.0644161254,0.1016490906,0.037777856,0.3812146783,-0.1532443315,0.0899222344,0.1483831406,0.1883367747,0.2025395483,0.2628896534,0.2135414332,-0.187863946,-0.0341212004,-0.2410863489,0.1436366886,-0.142726168,0.4156598449,0.1505924016,0.1928557456,-0.1148526296,-0.4619386494,-0.2462679744,0.2353635579,-0.3245418668,-0.3092090786,-0.0486779399,-0.0801849738,0.0559291169,0.1945610195,-0.0792901143,0.2038590163,0.3428387344,-0.0086603099,-0.0137500595,-0.0843536034,-0.1706176847,0.1139787063,0.2748646736,-0.0817389637,0.4469365478,0.0108006606,-0.0515694395,-0.2344710529,0.4114173651,0.2861286402,0.1949478239,-0.1688625664,0.0639532804,0.200646624,0.0306134559,-0.3546894789,0.3286518455,0.1511448473,-0.2930275798,0.368005842,0.1237954348,-0.2321557552,0.2764032483,-0.1041960269,-0.2014673799,-0.0498889387,0.3224370182,0.0997074991,-0.039932128,-0.1949630082,0.1111204624,-0.1445716172,-0.2853394747,0.1593456715,-0.0238983259,0.4676609337,-0.2756756544,0.1172051802,-0.0618249103,0.6220863461,-0.0292275213,0.1243288815,-0.3501377404,0.1026433855,-0.3331668377,0.1984097809,-0.0838177949,0.1537422538,-0.3975837231,0.2042651922,-0.1154986024,-0.0234658923,-0.1180424616,-0.0078867814,0.147406593,0.178931728,-0.3869177699,0.1554814875,-0.0862594694,-0.0062670191,0.2917736471,-0.3766753376,0.2219908834,0.042479679,0.2069104314,0.0109865274,-0.021723574,-0.0548064671,0.3365739584,0.6232414246,0.3470292389,0.1425508261,-0.1010277867,0.0815534294,-0.347438246,0.0141040748,-0.0608881786,-0.2096496969,0.0975683033,0.5258308053,-0.2256272584,0.1319463998,-0.1780457795,0.2366945595,0.3822763264,-0.1527291089,0.0617997609,0.0482259318,-0.2598406374,-0.0051989588,-0.125866577,0.4537958205,0.0071826167,0.3238305449,-0.3592907786,-0.5417625308,0.4329102337,-0.4511258006,-0.3025252223,0.1232788339,0.238785848,0.3495687544,-0.0600003339,-0.5436675549,-0.1724909097,0.2691140175,-0.0306607224,-0.1570452452,0.2076845467,-0.1486439705,-0.0410752967,-0.1167542785,0.1764688343,0.101526916,-0.2914130688,0.2436990142,-0.3701035082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/160","title":"caching in map causes same result to be returned for train, validation and test","comments":"The PR should prevent the cache from losing track of the of the dataset type (based on the location of its data). Not sure about your second problem though (cache off).","body":"hello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n","comment_length":31,"text":"caching in map causes same result to be returned for train, validation and test\nhello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n\nThe PR should prevent the cache from losing track of the of the dataset type (based on the location of its data). Not sure about your second problem though (cache off).","embeddings":[-0.1263227314,-0.2207089961,-0.100632146,0.2780107558,0.2454756796,-0.2052229047,0.1372107863,0.5128592849,0.1882870793,-0.0623357445,0.1278197765,0.3891343772,0.0037762846,-0.1606175303,-0.0107436497,0.0242125317,0.0960802659,0.2350524068,-0.1455620974,-0.1648233086,0.081894055,0.2129840404,-0.0424102098,0.1165268347,-0.3903492093,0.0412896834,0.2123562694,-0.2308897674,0.0684574693,-0.2570649683,0.1678505987,-0.0932429209,0.0258333106,0.1663175523,-0.0001097382,0.0067983456,-0.0153878154,-0.1385697424,-0.0072745453,0.0401948094,0.1181339324,-0.2117293179,-0.1091826707,-0.3121318519,-0.3320115805,-0.0376609191,0.0178094879,-0.3556600213,0.3128580153,0.1474524736,0.2607103586,0.2249089628,-0.2538127303,0.2111362219,0.1451275945,-0.0293276701,-0.174197495,-0.042305138,-0.043978855,-0.3344650865,-0.3767647743,0.5254573822,-0.0765818357,0.1701412946,0.2146666646,0.2250556946,0.1726054549,0.03593757,0.1429362446,-0.1139624342,0.0110948198,-0.2269175202,-0.0271165781,-0.3733721972,-0.4186404347,-0.0785259828,0.137625128,0.1810450852,-0.1363935322,-0.06371627,-0.4319375753,0.2392898649,0.0652496368,-0.0127885444,0.1195269376,0.2951607704,0.0438062623,0.1952228397,0.1294467449,-0.069732815,-0.0225592721,-0.2145169973,-0.0589715466,0.2264011949,-0.4359855056,0.0720824599,0.4315197766,-0.1404137909,-0.0532427132,0.0306374077,0.3919309974,0.0792137086,0.1168181524,-0.0084918672,-0.104160659,0.5898677111,0.0443217419,0.171184063,-0.0149702216,-0.1982782632,-0.3560169637,0.0049879928,0.288017571,-0.286601156,0.4284112751,0.4023602307,-0.0422602966,-0.1215295419,-0.0301021244,0.0940235928,-0.3308537602,0.055669792,0.0587296784,0.3139583766,-0.2254502922,0.149265632,-0.1737674624,-0.1033549458,-0.3849695921,0.1273514628,-0.3378459215,-0.0333239511,-0.4962185323,0.0015165606,0.285384804,-0.0028597817,0.2853520513,-0.0384451598,-0.1372202933,-0.3019890487,0.1031737626,-0.3985118568,0.5702316761,-0.10722965,-0.197017163,0.2410212457,0.1303783357,0.0852641463,-0.2976881862,-0.1983774155,-0.1698076725,-0.2090685666,0.2598606944,0.2422455698,-0.0794276446,0.1413960606,0.0924820974,0.0776791126,0.3432580829,-0.232855022,0.1028404012,-0.2014892399,-0.2819799483,-0.2043585628,0.1625683904,0.2623100579,0.0667765215,-0.1373723596,0.4017969966,0.2371831983,0.1472906321,0.4138066471,-0.1965338886,0.3094943762,-0.4334916174,0.348619163,0.5064768195,-0.3267318308,-0.4963446558,0.0994484499,-0.1563160419,0.1002884582,0.1024493128,0.151373148,0.0290875416,0.0144196171,0.1285368353,0.0187024176,-0.0005147557,0.2084205002,-0.316927731,-0.0219340678,0.5500642657,-0.2524672449,0.0284720566,-0.131236881,-0.0550430529,0.0219331458,0.0067682369,-0.1121900007,0.1218594089,0.1146048158,-0.0568789616,-0.1581844538,-0.106438823,-0.0246925242,-0.2882628739,0.2383016348,-0.1498185694,-0.0999678373,0.0617249236,-0.0970491692,-0.1620027572,-0.129275158,-0.1809165776,-0.2730642855,0.2013016045,0.2305243611,0.3202398419,-0.0729073882,0.2252094597,0.3680819571,0.0483636484,-0.2160722017,-0.4300346375,0.0856799036,-0.1777346134,-0.2188281417,-0.0748463124,0.2230703831,0.2090260535,-0.0479263812,-0.1424278915,0.1482760459,-0.2220858783,0.4189184308,-0.1370279044,0.2694525421,-0.0696570575,0.0293060318,0.016667014,0.2540052533,0.0588200353,-0.0898012966,-0.0727058947,0.4742337465,0.1814022362,0.1701161861,-0.0480709597,-0.2162680477,0.0519842356,-0.1783178449,-0.1209156215,-0.1188160628,0.1364075691,-0.0805720538,0.148888275,0.152328819,-0.1069871038,0.3120084703,0.7600834966,0.1969186664,-0.0110169193,0.0157407038,-0.0572130904,-0.2177918702,0.3338562846,-0.058682885,0.31038782,0.1047316417,0.1271647215,-0.0462091751,-0.2196502686,-0.154435575,0.1246479601,-0.2143685222,-0.1919764429,0.3080338836,0.2820868194,-0.1583673358,-0.3537832797,0.0940154344,-0.0127645368,0.0261916555,-0.1580014974,0.2172367424,-0.166529417,-0.278084904,-0.2864152193,0.136156708,-0.1222038716,-0.0921654701,-0.0412631668,0.4254295826,-0.036654897,0.1951516867,-0.3149722517,0.0895192102,0.2266037613,-0.2663277388,-0.0716857836,-0.1902881712,-0.3949196339,0.042231977,-0.0952302068,-0.0929960757,0.3783293962,-0.0473160893,-0.067959927,-0.2881456017,-0.1725400686,0.1071778685,-0.0926118121,-0.1553680301,0.1672483534,-0.0313384198,-0.1011587083,-0.2987454534,0.2937971354,-0.3053128421,-0.299143523,-0.0519867688,0.0989064649,0.0301090628,-0.2803472877,-0.428624332,0.2735781968,-0.3172441423,0.0754678994,-0.1489290893,0.1862263232,0.3309345841,0.0089533916,-0.0145607088,-0.0640404522,0.1643940359,-0.4674099684,-0.1409540623,0.3225782216,0.0187610649,-0.2163045555,-0.3245913982,-0.2757247686,0.3477241099,0.1923889369,-0.5803956985,-0.2144230455,-0.048900146,0.2480322272,-0.0512586385,-0.0464955159,0.5745424032,0.0197255258,-0.179463625,-0.1672781259,-0.3318405151,0.359731406,0.4192632437,0.2494875044,-0.0111588463,0.1778150499,0.2009202093,0.6707463264,0.2019197792,-0.2590580583,0.331730634,0.1072319448,0.0749976784,-0.2034970075,-0.3130226135,0.1857604682,-0.1816030145,0.0197321847,0.2719842196,0.0062508965,-0.2801841497,0.0177421588,0.2188196182,-0.2253448963,-0.3280583322,0.2597456276,-0.3424280584,0.2409481704,0.0363704525,0.3158006966,-0.582013905,-0.2427242249,0.1946279705,-0.3359463811,0.2798449695,0.0444980972,-0.5836006999,0.1214706451,-0.5380226374,0.1819416732,0.3146281838,0.2724077702,-0.092025429,-0.2465260625,0.043347083,-0.1471314728,0.5711411834,0.0409730114,0.4085795283,0.1699264944,-0.1302466393,-0.1994734555,-0.2247125357,0.0172346737,0.2801373005,0.3264995515,0.4008480906,-0.3452449739,-0.3632731438,-0.2657663524,0.0073641897,-0.0461810417,0.007107567,-0.1258338541,0.1893848181,0.0164867323,0.1743457466,-0.0149804754,0.118302688,0.0160397142,-0.1365461946,-0.0391333476,-0.2194266319,0.1282413155,0.2264375687,0.4735942185,0.1503249407,0.0324611329,0.0332145244,0.2151578218,-0.567109406,0.2659578621,-0.0888494328,-0.0041319211,0.0746441931,0.1444383562,-0.030301474,0.337897867,-0.042524904,0.1457136869,-0.1286294758,0.0089567872,-0.5539684892,0.4854543805,0.3874592483,0.1324663013,-0.0448895507,-0.246857211,0.1734628826,0.0741201192,-0.2765311003,0.1322598755,-0.4306930304,-0.4319946468,0.1857973337,0.2094861418,0.9232475162,0.332021594,0.3719581068,0.1007037163,0.1035806388,0.2642515898,-0.2341118157,0.4092507362,-0.3693645,-0.0463177823,0.0245385505,-0.1667332053,-0.2273836434,-0.0104256496,-0.0109939184,0.3025950491,0.3240507543,0.5999364257,-0.1121197641,0.1666993201,0.2156913131,-0.1581103951,-0.2492317557,0.2010731846,-0.2923038006,0.315240562,-0.1124083251,-0.0741084665,-0.0165099371,-0.3513420522,0.0671400949,0.0609418713,-0.3365171552,0.30484128,0.1334620416,-0.2726664841,-0.2969880998,0.2550771236,0.1119868383,0.0825273544,-0.0486002751,-0.0351850241,0.4006448984,0.3504586816,-0.0068906583,-0.2015887648,0.1864127815,0.0638751984,-0.0128940148,0.0302428547,-0.0564054921,-0.3015830219,-0.2525120974,0.1482045352,0.0033722101,-0.4129621685,-0.055318471,0.0614447072,-0.0667884797,-0.0941368639,0.150880307,-0.0260680113,-0.1079400256,0.315154314,0.1173381507,-0.2430422902,0.0523580648,0.4748320878,-0.0403974019,0.0318177901,0.4154281914,-0.15281865,-0.103806369,-0.2553910613,-0.2841726243,-0.1020596772,-0.4488408566,0.1122078001,-0.219405964,-0.4019113779,0.0442324318,0.2808186412,0.0147068594,0.3184957206,-0.0116181625,-0.1249581054,-0.3769723475,-0.1037905142,-0.1065821052,0.2671778798,0.2795186639,0.284348309,-0.2958665192,0.2492261082,-0.3831097186,-0.1071424857,-0.187170282,0.43791309,-0.0391381346,-0.3655827045,-0.0091469325,0.0018549742,0.1299711019,0.1775448769,-0.63865906,-0.2926191092,-0.0909917802,0.0818608031,-0.0075877849,-0.1598218381,-0.2392821759,-0.0172140636,0.1823851168,-0.4274035096,0.0921567082,0.1512164474,0.047031749,0.2371866852,-0.1610301137,0.1214557961,0.2881679535,-0.1865271479,-0.1841488928,-0.0762282088,0.0804811493,0.2570166588,-0.0694677755,0.0204793625,-0.2104029655,0.030472707,0.0892741159,0.1427949369,0.2369287759,-0.0247526355,-0.2156609595,0.0005509924,0.1827996075,0.114120096,-0.0785531625,-0.3289754689,0.2802228332,0.2783651054,-0.1175179482,-0.0102367681,0.167707786,-0.0993882194,0.2528515756,0.2071537971,0.1075878143,0.0526108742,0.2696116865,0.1264938265,0.4050408304,-0.348328352,0.0894215927,0.0644161254,0.1016490906,0.037777856,0.3812146783,-0.1532443315,0.0899222344,0.1483831406,0.1883367747,0.2025395483,0.2628896534,0.2135414332,-0.187863946,-0.0341212004,-0.2410863489,0.1436366886,-0.142726168,0.4156598449,0.1505924016,0.1928557456,-0.1148526296,-0.4619386494,-0.2462679744,0.2353635579,-0.3245418668,-0.3092090786,-0.0486779399,-0.0801849738,0.0559291169,0.1945610195,-0.0792901143,0.2038590163,0.3428387344,-0.0086603099,-0.0137500595,-0.0843536034,-0.1706176847,0.1139787063,0.2748646736,-0.0817389637,0.4469365478,0.0108006606,-0.0515694395,-0.2344710529,0.4114173651,0.2861286402,0.1949478239,-0.1688625664,0.0639532804,0.200646624,0.0306134559,-0.3546894789,0.3286518455,0.1511448473,-0.2930275798,0.368005842,0.1237954348,-0.2321557552,0.2764032483,-0.1041960269,-0.2014673799,-0.0498889387,0.3224370182,0.0997074991,-0.039932128,-0.1949630082,0.1111204624,-0.1445716172,-0.2853394747,0.1593456715,-0.0238983259,0.4676609337,-0.2756756544,0.1172051802,-0.0618249103,0.6220863461,-0.0292275213,0.1243288815,-0.3501377404,0.1026433855,-0.3331668377,0.1984097809,-0.0838177949,0.1537422538,-0.3975837231,0.2042651922,-0.1154986024,-0.0234658923,-0.1180424616,-0.0078867814,0.147406593,0.178931728,-0.3869177699,0.1554814875,-0.0862594694,-0.0062670191,0.2917736471,-0.3766753376,0.2219908834,0.042479679,0.2069104314,0.0109865274,-0.021723574,-0.0548064671,0.3365739584,0.6232414246,0.3470292389,0.1425508261,-0.1010277867,0.0815534294,-0.347438246,0.0141040748,-0.0608881786,-0.2096496969,0.0975683033,0.5258308053,-0.2256272584,0.1319463998,-0.1780457795,0.2366945595,0.3822763264,-0.1527291089,0.0617997609,0.0482259318,-0.2598406374,-0.0051989588,-0.125866577,0.4537958205,0.0071826167,0.3238305449,-0.3592907786,-0.5417625308,0.4329102337,-0.4511258006,-0.3025252223,0.1232788339,0.238785848,0.3495687544,-0.0600003339,-0.5436675549,-0.1724909097,0.2691140175,-0.0306607224,-0.1570452452,0.2076845467,-0.1486439705,-0.0410752967,-0.1167542785,0.1764688343,0.101526916,-0.2914130688,0.2436990142,-0.3701035082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/160","title":"caching in map causes same result to be returned for train, validation and test","comments":"Yes, with caching on, it seems to work without the function renaming hack, I mentioned this also in the PR. Thanks!","body":"hello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n","comment_length":21,"text":"caching in map causes same result to be returned for train, validation and test\nhello,\r\n\r\nI am working on a program that uses the `nlp` library with the `SST2` dataset.\r\n\r\nThe rough outline of the program is:\r\n\r\n```\r\nimport nlp as nlp_datasets\r\n...\r\nparser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+')\r\n...\r\ndataset = nlp_datasets.load_dataset(*args.dataset)\r\n...\r\n# Create feature vocabs\r\nvocabs = create_vocabs(dataset.values(), vectorizers)\r\n...\r\n# Create a function to vectorize based on vectorizers and vocabs:\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n# factory method to create a `convert_to_features` function based on vocabs\r\nconvert_to_features = create_featurizer(vectorizers, vocabs)\r\ntrain_set = train_set.map(convert_to_features, batched=True)\r\ntrain_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntrain_loader = torch.utils.data.DataLoader(train_set, batch_size=args.batchsz)\r\n\r\nvalid_set = valid_set.map(convert_to_features, batched=True)\r\nvalid_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\nvalid_loader = torch.utils.data.DataLoader(valid_set, batch_size=args.batchsz)\r\n\r\ntest_set = test_set.map(convert_to_features, batched=True)\r\ntest_set.set_format(type='torch', columns=list(vectorizers.keys()) + ['y', 'lengths'])\r\ntest_loader = torch.utils.data.DataLoader(test_set, batch_size=args.batchsz)\r\n\r\nprint('TS', train_set.num_rows)\r\nprint('VS', valid_set.num_rows)\r\nprint('ES', test_set.num_rows)\r\n\r\n```\r\nIm not sure if Im using it incorrectly, but the results are not what I expect. Namely, the `.map()` seems to grab the datset from the cache and then loses track of what the specific dataset is, instead using my training data for all datasets:\r\n\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\nTS 67349\r\nVS 67349\r\nES 67349\r\n```\r\n\r\nThe behavior changes if I turn off the caching but then the results fail:\r\n\r\n```\r\ntrain_set = train_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\nvalid_set = valid_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n...\r\ntest_set = test_set.map(convert_to_features, batched=True, load_from_cache_file=False)\r\n```\r\n\r\nNow I get the right set of features back...\r\n```\r\nTS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 68\/68 [00:00<00:00, 92.78it\/s]\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1\/1 [00:00<00:00, 75.47it\/s]\r\n 0%| | 0\/2 [00:00<?, ?it\/s]TS 67349\r\nVS 872\r\nES 1821\r\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2\/2 [00:00<00:00, 77.19it\/s]\r\n```\r\nbut I think its losing track of the original training set:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"\/home\/dpressel\/dev\/work\/baseline\/api-examples\/layers-classify-hf-datasets.py\", line 148, in <module>\r\n for x in train_loader:\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 345, in __next__\r\n data = self._next_data()\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/dataloader.py\", line 385, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/torch\/utils\/data\/_utils\/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 338, in __getitem__\r\n output_all_columns=self._output_all_columns,\r\n File \"\/home\/dpressel\/anaconda3\/lib\/python3.7\/site-packages\/nlp\/arrow_dataset.py\", line 294, in _getitem\r\n outputs = self._unnest(self._data.slice(key, 1).to_pydict())\r\n File \"pyarrow\/table.pxi\", line 1211, in pyarrow.lib.Table.slice\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 3: In chunk 0: Invalid: Length spanned by list offsets (15859698) larger than values array (length 100000)\r\n\r\nProcess finished with exit code 1\r\n```\r\n\r\nThe full-example program (minus the print stmts) is here:\r\nhttps:\/\/github.com\/dpressel\/mead-baseline\/pull\/620\/files\r\n\r\n\nYes, with caching on, it seems to work without the function renaming hack, I mentioned this also in the PR. Thanks!","embeddings":[-0.1263227314,-0.2207089961,-0.100632146,0.2780107558,0.2454756796,-0.2052229047,0.1372107863,0.5128592849,0.1882870793,-0.0623357445,0.1278197765,0.3891343772,0.0037762846,-0.1606175303,-0.0107436497,0.0242125317,0.0960802659,0.2350524068,-0.1455620974,-0.1648233086,0.081894055,0.2129840404,-0.0424102098,0.1165268347,-0.3903492093,0.0412896834,0.2123562694,-0.2308897674,0.0684574693,-0.2570649683,0.1678505987,-0.0932429209,0.0258333106,0.1663175523,-0.0001097382,0.0067983456,-0.0153878154,-0.1385697424,-0.0072745453,0.0401948094,0.1181339324,-0.2117293179,-0.1091826707,-0.3121318519,-0.3320115805,-0.0376609191,0.0178094879,-0.3556600213,0.3128580153,0.1474524736,0.2607103586,0.2249089628,-0.2538127303,0.2111362219,0.1451275945,-0.0293276701,-0.174197495,-0.042305138,-0.043978855,-0.3344650865,-0.3767647743,0.5254573822,-0.0765818357,0.1701412946,0.2146666646,0.2250556946,0.1726054549,0.03593757,0.1429362446,-0.1139624342,0.0110948198,-0.2269175202,-0.0271165781,-0.3733721972,-0.4186404347,-0.0785259828,0.137625128,0.1810450852,-0.1363935322,-0.06371627,-0.4319375753,0.2392898649,0.0652496368,-0.0127885444,0.1195269376,0.2951607704,0.0438062623,0.1952228397,0.1294467449,-0.069732815,-0.0225592721,-0.2145169973,-0.0589715466,0.2264011949,-0.4359855056,0.0720824599,0.4315197766,-0.1404137909,-0.0532427132,0.0306374077,0.3919309974,0.0792137086,0.1168181524,-0.0084918672,-0.104160659,0.5898677111,0.0443217419,0.171184063,-0.0149702216,-0.1982782632,-0.3560169637,0.0049879928,0.288017571,-0.286601156,0.4284112751,0.4023602307,-0.0422602966,-0.1215295419,-0.0301021244,0.0940235928,-0.3308537602,0.055669792,0.0587296784,0.3139583766,-0.2254502922,0.149265632,-0.1737674624,-0.1033549458,-0.3849695921,0.1273514628,-0.3378459215,-0.0333239511,-0.4962185323,0.0015165606,0.285384804,-0.0028597817,0.2853520513,-0.0384451598,-0.1372202933,-0.3019890487,0.1031737626,-0.3985118568,0.5702316761,-0.10722965,-0.197017163,0.2410212457,0.1303783357,0.0852641463,-0.2976881862,-0.1983774155,-0.1698076725,-0.2090685666,0.2598606944,0.2422455698,-0.0794276446,0.1413960606,0.0924820974,0.0776791126,0.3432580829,-0.232855022,0.1028404012,-0.2014892399,-0.2819799483,-0.2043585628,0.1625683904,0.2623100579,0.0667765215,-0.1373723596,0.4017969966,0.2371831983,0.1472906321,0.4138066471,-0.1965338886,0.3094943762,-0.4334916174,0.348619163,0.5064768195,-0.3267318308,-0.4963446558,0.0994484499,-0.1563160419,0.1002884582,0.1024493128,0.151373148,0.0290875416,0.0144196171,0.1285368353,0.0187024176,-0.0005147557,0.2084205002,-0.316927731,-0.0219340678,0.5500642657,-0.2524672449,0.0284720566,-0.131236881,-0.0550430529,0.0219331458,0.0067682369,-0.1121900007,0.1218594089,0.1146048158,-0.0568789616,-0.1581844538,-0.106438823,-0.0246925242,-0.2882628739,0.2383016348,-0.1498185694,-0.0999678373,0.0617249236,-0.0970491692,-0.1620027572,-0.129275158,-0.1809165776,-0.2730642855,0.2013016045,0.2305243611,0.3202398419,-0.0729073882,0.2252094597,0.3680819571,0.0483636484,-0.2160722017,-0.4300346375,0.0856799036,-0.1777346134,-0.2188281417,-0.0748463124,0.2230703831,0.2090260535,-0.0479263812,-0.1424278915,0.1482760459,-0.2220858783,0.4189184308,-0.1370279044,0.2694525421,-0.0696570575,0.0293060318,0.016667014,0.2540052533,0.0588200353,-0.0898012966,-0.0727058947,0.4742337465,0.1814022362,0.1701161861,-0.0480709597,-0.2162680477,0.0519842356,-0.1783178449,-0.1209156215,-0.1188160628,0.1364075691,-0.0805720538,0.148888275,0.152328819,-0.1069871038,0.3120084703,0.7600834966,0.1969186664,-0.0110169193,0.0157407038,-0.0572130904,-0.2177918702,0.3338562846,-0.058682885,0.31038782,0.1047316417,0.1271647215,-0.0462091751,-0.2196502686,-0.154435575,0.1246479601,-0.2143685222,-0.1919764429,0.3080338836,0.2820868194,-0.1583673358,-0.3537832797,0.0940154344,-0.0127645368,0.0261916555,-0.1580014974,0.2172367424,-0.166529417,-0.278084904,-0.2864152193,0.136156708,-0.1222038716,-0.0921654701,-0.0412631668,0.4254295826,-0.036654897,0.1951516867,-0.3149722517,0.0895192102,0.2266037613,-0.2663277388,-0.0716857836,-0.1902881712,-0.3949196339,0.042231977,-0.0952302068,-0.0929960757,0.3783293962,-0.0473160893,-0.067959927,-0.2881456017,-0.1725400686,0.1071778685,-0.0926118121,-0.1553680301,0.1672483534,-0.0313384198,-0.1011587083,-0.2987454534,0.2937971354,-0.3053128421,-0.299143523,-0.0519867688,0.0989064649,0.0301090628,-0.2803472877,-0.428624332,0.2735781968,-0.3172441423,0.0754678994,-0.1489290893,0.1862263232,0.3309345841,0.0089533916,-0.0145607088,-0.0640404522,0.1643940359,-0.4674099684,-0.1409540623,0.3225782216,0.0187610649,-0.2163045555,-0.3245913982,-0.2757247686,0.3477241099,0.1923889369,-0.5803956985,-0.2144230455,-0.048900146,0.2480322272,-0.0512586385,-0.0464955159,0.5745424032,0.0197255258,-0.179463625,-0.1672781259,-0.3318405151,0.359731406,0.4192632437,0.2494875044,-0.0111588463,0.1778150499,0.2009202093,0.6707463264,0.2019197792,-0.2590580583,0.331730634,0.1072319448,0.0749976784,-0.2034970075,-0.3130226135,0.1857604682,-0.1816030145,0.0197321847,0.2719842196,0.0062508965,-0.2801841497,0.0177421588,0.2188196182,-0.2253448963,-0.3280583322,0.2597456276,-0.3424280584,0.2409481704,0.0363704525,0.3158006966,-0.582013905,-0.2427242249,0.1946279705,-0.3359463811,0.2798449695,0.0444980972,-0.5836006999,0.1214706451,-0.5380226374,0.1819416732,0.3146281838,0.2724077702,-0.092025429,-0.2465260625,0.043347083,-0.1471314728,0.5711411834,0.0409730114,0.4085795283,0.1699264944,-0.1302466393,-0.1994734555,-0.2247125357,0.0172346737,0.2801373005,0.3264995515,0.4008480906,-0.3452449739,-0.3632731438,-0.2657663524,0.0073641897,-0.0461810417,0.007107567,-0.1258338541,0.1893848181,0.0164867323,0.1743457466,-0.0149804754,0.118302688,0.0160397142,-0.1365461946,-0.0391333476,-0.2194266319,0.1282413155,0.2264375687,0.4735942185,0.1503249407,0.0324611329,0.0332145244,0.2151578218,-0.567109406,0.2659578621,-0.0888494328,-0.0041319211,0.0746441931,0.1444383562,-0.030301474,0.337897867,-0.042524904,0.1457136869,-0.1286294758,0.0089567872,-0.5539684892,0.4854543805,0.3874592483,0.1324663013,-0.0448895507,-0.246857211,0.1734628826,0.0741201192,-0.2765311003,0.1322598755,-0.4306930304,-0.4319946468,0.1857973337,0.2094861418,0.9232475162,0.332021594,0.3719581068,0.1007037163,0.1035806388,0.2642515898,-0.2341118157,0.4092507362,-0.3693645,-0.0463177823,0.0245385505,-0.1667332053,-0.2273836434,-0.0104256496,-0.0109939184,0.3025950491,0.3240507543,0.5999364257,-0.1121197641,0.1666993201,0.2156913131,-0.1581103951,-0.2492317557,0.2010731846,-0.2923038006,0.315240562,-0.1124083251,-0.0741084665,-0.0165099371,-0.3513420522,0.0671400949,0.0609418713,-0.3365171552,0.30484128,0.1334620416,-0.2726664841,-0.2969880998,0.2550771236,0.1119868383,0.0825273544,-0.0486002751,-0.0351850241,0.4006448984,0.3504586816,-0.0068906583,-0.2015887648,0.1864127815,0.0638751984,-0.0128940148,0.0302428547,-0.0564054921,-0.3015830219,-0.2525120974,0.1482045352,0.0033722101,-0.4129621685,-0.055318471,0.0614447072,-0.0667884797,-0.0941368639,0.150880307,-0.0260680113,-0.1079400256,0.315154314,0.1173381507,-0.2430422902,0.0523580648,0.4748320878,-0.0403974019,0.0318177901,0.4154281914,-0.15281865,-0.103806369,-0.2553910613,-0.2841726243,-0.1020596772,-0.4488408566,0.1122078001,-0.219405964,-0.4019113779,0.0442324318,0.2808186412,0.0147068594,0.3184957206,-0.0116181625,-0.1249581054,-0.3769723475,-0.1037905142,-0.1065821052,0.2671778798,0.2795186639,0.284348309,-0.2958665192,0.2492261082,-0.3831097186,-0.1071424857,-0.187170282,0.43791309,-0.0391381346,-0.3655827045,-0.0091469325,0.0018549742,0.1299711019,0.1775448769,-0.63865906,-0.2926191092,-0.0909917802,0.0818608031,-0.0075877849,-0.1598218381,-0.2392821759,-0.0172140636,0.1823851168,-0.4274035096,0.0921567082,0.1512164474,0.047031749,0.2371866852,-0.1610301137,0.1214557961,0.2881679535,-0.1865271479,-0.1841488928,-0.0762282088,0.0804811493,0.2570166588,-0.0694677755,0.0204793625,-0.2104029655,0.030472707,0.0892741159,0.1427949369,0.2369287759,-0.0247526355,-0.2156609595,0.0005509924,0.1827996075,0.114120096,-0.0785531625,-0.3289754689,0.2802228332,0.2783651054,-0.1175179482,-0.0102367681,0.167707786,-0.0993882194,0.2528515756,0.2071537971,0.1075878143,0.0526108742,0.2696116865,0.1264938265,0.4050408304,-0.348328352,0.0894215927,0.0644161254,0.1016490906,0.037777856,0.3812146783,-0.1532443315,0.0899222344,0.1483831406,0.1883367747,0.2025395483,0.2628896534,0.2135414332,-0.187863946,-0.0341212004,-0.2410863489,0.1436366886,-0.142726168,0.4156598449,0.1505924016,0.1928557456,-0.1148526296,-0.4619386494,-0.2462679744,0.2353635579,-0.3245418668,-0.3092090786,-0.0486779399,-0.0801849738,0.0559291169,0.1945610195,-0.0792901143,0.2038590163,0.3428387344,-0.0086603099,-0.0137500595,-0.0843536034,-0.1706176847,0.1139787063,0.2748646736,-0.0817389637,0.4469365478,0.0108006606,-0.0515694395,-0.2344710529,0.4114173651,0.2861286402,0.1949478239,-0.1688625664,0.0639532804,0.200646624,0.0306134559,-0.3546894789,0.3286518455,0.1511448473,-0.2930275798,0.368005842,0.1237954348,-0.2321557552,0.2764032483,-0.1041960269,-0.2014673799,-0.0498889387,0.3224370182,0.0997074991,-0.039932128,-0.1949630082,0.1111204624,-0.1445716172,-0.2853394747,0.1593456715,-0.0238983259,0.4676609337,-0.2756756544,0.1172051802,-0.0618249103,0.6220863461,-0.0292275213,0.1243288815,-0.3501377404,0.1026433855,-0.3331668377,0.1984097809,-0.0838177949,0.1537422538,-0.3975837231,0.2042651922,-0.1154986024,-0.0234658923,-0.1180424616,-0.0078867814,0.147406593,0.178931728,-0.3869177699,0.1554814875,-0.0862594694,-0.0062670191,0.2917736471,-0.3766753376,0.2219908834,0.042479679,0.2069104314,0.0109865274,-0.021723574,-0.0548064671,0.3365739584,0.6232414246,0.3470292389,0.1425508261,-0.1010277867,0.0815534294,-0.347438246,0.0141040748,-0.0608881786,-0.2096496969,0.0975683033,0.5258308053,-0.2256272584,0.1319463998,-0.1780457795,0.2366945595,0.3822763264,-0.1527291089,0.0617997609,0.0482259318,-0.2598406374,-0.0051989588,-0.125866577,0.4537958205,0.0071826167,0.3238305449,-0.3592907786,-0.5417625308,0.4329102337,-0.4511258006,-0.3025252223,0.1232788339,0.238785848,0.3495687544,-0.0600003339,-0.5436675549,-0.1724909097,0.2691140175,-0.0306607224,-0.1570452452,0.2076845467,-0.1486439705,-0.0410752967,-0.1167542785,0.1764688343,0.101526916,-0.2914130688,0.2436990142,-0.3701035082]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/159","title":"How can we add more datasets to nlp library?","comments":"Found it. https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets","body":"","comment_length":3,"text":"How can we add more datasets to nlp library?\n\nFound it. https:\/\/github.com\/huggingface\/nlp\/tree\/master\/datasets","embeddings":[-0.1335406452,0.1999934465,-0.2406620532,0.2252554893,0.0810523108,0.1613626033,0.0168796554,0.1892843097,0.0033402608,-0.0771228969,-0.1592342854,0.4524565935,-0.1887387484,0.3291336,0.5956285,-0.4399953187,0.0195516236,0.1699261665,-0.245976612,-0.1265086085,-0.14989838,0.1877745986,0.0256802831,-0.2843306661,-0.2348057628,-0.1059194431,-0.2711085081,-0.02096368,-0.1145843938,-0.5019678473,-0.0869348943,0.0950045362,0.0909845382,-0.0099270474,-0.0000999333,-0.2091178298,0.2795499563,0.0025929061,-0.384901464,-0.2607463598,-0.082572706,-0.410484165,0.109910965,-0.1736689657,-0.1049469858,-0.2483086288,0.2519674599,0.1185774356,0.1293655634,0.3523300588,0.3330751061,0.2948721647,0.1775335222,-0.0024465271,0.2442770004,0.1760163009,-0.135796234,0.0743897855,0.2295242846,-0.0218744185,-0.0678738207,0.0465088151,0.1537063271,-0.1446955502,0.065355964,0.0216108859,-0.1422040015,-0.1187957823,-0.086151436,0.1341631413,0.4753353596,-0.1693312079,-0.28078565,-0.3158725202,0.2523893118,-0.2316648066,-0.0637258813,0.3033251762,0.0782851353,-0.0479708202,-0.1140153259,-0.4021839499,-0.1647203863,0.3876830339,0.4785958827,0.4122327864,-0.0745439827,-0.0536767058,0.319023937,-0.029807888,0.0332277231,0.0717366189,-0.2978582084,0.0695188791,-0.1501054317,-0.2853544056,0.1159969568,0.2642504573,0.3516438901,0.1664254963,0.0188953001,0.0472366214,-0.3197778165,-0.0247459728,0.2957212627,-0.1152371913,0.1393503547,0.1067295149,0.0382820256,-0.0911423713,-0.0070458325,0.0463166758,0.0926590487,0.0894760266,-0.1321589351,-0.1104139537,-0.0392894,-0.0495345555,0.1895581484,-0.1637314409,-0.0123388125,-0.2696214318,0.0926019028,0.13343063,0.0558044091,0.0346179791,-0.2261692584,0.1762873977,-0.0182700641,-0.4319476485,-0.2136096656,0.3066769838,-0.1892397255,0.0293266866,0.2973871529,-0.2043558955,0.5927449465,-0.0159081761,-0.197298184,0.1002195477,0.1506810784,0.0340301879,0.1300349236,-0.1674223244,0.0074217492,-0.0455638282,-0.1730204374,0.007653615,-0.2342523783,-0.140082553,-0.2257750034,-0.1409555525,-0.4460218251,0.2544766366,-0.0758722126,-0.4450891912,-0.3160660863,0.5898552537,-0.0551020764,-0.0943828598,0.1678987592,-0.0063964338,-0.353079468,-0.1583533734,-0.1260597706,0.1047586799,0.0641503558,-0.2278089672,-0.0604392514,-0.0400538072,-0.0592638403,0.2985629141,-0.3287353814,0.0206859764,-0.2316683233,0.1122634634,0.5914571881,-0.1240472943,-0.2430337667,0.0467122123,0.070261687,-0.1881186664,0.0892356485,0.2912685275,0.0456054136,-0.0139385657,-0.1230360717,0.3655094206,-0.1366529614,-0.098027803,-0.0229041968,-0.0774451569,0.3363763094,0.1914257854,-0.0334399603,-0.0109504955,0.2080673575,0.1253336519,0.219971776,-0.2193825096,0.0428953134,0.2837908268,0.1296700984,0.1320646405,0.0165707842,-0.4904898107,-0.1906900555,-0.2443612218,-0.215105325,0.299709022,0.2886498868,-0.1620037258,-0.0477799214,-0.0604577139,0.0634474009,-0.1799525023,0.2647878528,0.2254356146,0.190323934,-0.00530113,-0.1913066804,0.1897924691,-0.2025238276,0.1406408697,-0.3546248674,0.2136448622,0.0128577556,-0.0891744345,0.301240027,0.4054248035,-0.0446298718,-0.248119697,0.1203448474,-0.0048918976,-0.2380119264,0.1503117234,0.3699913621,0.1436352432,0.3687567413,-0.2028590441,0.0892643929,0.0350976475,0.0029545894,0.0158351585,-0.0425813384,0.0029916968,0.2171903998,-0.126162976,0.1338384151,-0.0325368643,0.3441290557,-0.0057652481,0.0104464786,-0.079882659,0.2211545706,0.1072561145,0.2460805178,0.0693575144,-0.3039842248,0.277310431,0.4510338306,-0.0542265251,0.0689785182,0.1516648382,-0.1481006742,-0.2788386941,0.2786778808,0.1161556244,0.2962078452,0.4504546225,0.1247814,-0.1067658514,-0.1921533495,-0.1891237497,0.0977319628,-0.0263967067,-0.0331555121,0.0259766374,0.3112738729,-0.0820097402,-0.5328738093,-0.2268768251,-0.0273422934,0.17600362,0.1752070189,-0.0616667792,-0.3212575912,-0.7216424942,-0.0757442415,-0.1026117653,-0.3114608526,-0.1535623521,0.1423117965,0.0920075104,0.012428374,0.1342668831,0.0664306134,0.3270722926,-0.252335459,0.3285568058,-0.0212801415,-0.0578744337,-0.1878436655,0.2380752563,0.2645125687,0.0091326982,0.5530657768,-0.0274392087,0.0576486476,-0.4187124968,-0.4277716279,0.063429445,-0.092858389,0.3142046928,0.1008081585,-0.2064902782,0.0080453027,-0.189840287,0.0007667152,-0.2729120553,-0.0299755372,-0.2399253398,-0.2484616041,0.0014745818,-0.0779883116,-0.5138895512,-0.3617379665,-0.3681296408,0.4638181329,0.3387599587,0.1943744272,0.1074677557,-0.2269403934,0.1523561329,-0.277146548,0.2314647734,-0.1693885475,-0.2630432546,0.1228863001,-0.3903217912,-0.3714213669,-0.1077890918,-0.2754690945,0.3491194844,-0.2340399474,-0.3201162815,-0.2453724295,-0.1784490496,0.0305680688,-0.0529426299,0.2542542219,0.1459284723,-0.065092504,-0.1169044897,-0.0002327055,-0.2666864097,0.1338651776,0.3148212135,0.3010255396,-0.1235065237,0.1685944647,-0.1228209361,0.2416014522,0.1806844175,-0.0800805762,0.5001093149,-0.1274173409,0.2298307866,-0.1230072007,-0.3266549706,0.2871787846,0.1371640712,0.2804775536,0.3723748326,0.2090105265,-0.1762290746,0.0771544054,-0.4530126452,-0.2185431421,-0.0662719682,0.0972096771,0.0278982408,0.0776231512,-0.0381579548,-0.4353512228,-0.4400433302,-0.2639373839,-0.0994647071,0.0898439959,0.0233531315,0.1354276091,-0.4755792916,-0.1326186955,-0.4534306228,0.1232541278,0.2035360336,0.1173385307,-0.1345309317,0.1569252014,-0.0587870032,0.1689026803,0.3308016658,-0.2247410864,-0.3150060773,0.0004523364,-0.1441341639,-0.0877763778,0.2658806741,-0.1846583337,0.2601927817,0.2988402545,0.1838731021,-0.2033327222,-0.0062979162,0.4520226419,0.1113886833,0.0478207953,-0.3812561631,0.1852353364,0.0235544182,-0.1269459128,-0.0851396471,-0.1835716367,0.2259104997,0.1859068722,-0.2954384983,0.0740383193,-0.3120179474,0.1112242118,-0.0653377101,-0.1121871024,0.3958625793,-0.0041482495,-0.4416771829,0.2079727948,-0.0531757735,0.494299233,-0.103599377,-0.0794822276,-0.0804468095,-0.061346367,0.5910966992,-0.1763149947,0.2147365212,0.5073437691,-0.1693212837,-0.0542012118,-0.0899144858,0.2447328717,0.0526672937,0.2833581567,-0.1724648476,-0.3722310066,0.3977561295,0.052932471,0.1171398088,0.2029625773,-0.0509205833,-0.2823122442,0.0311011653,-0.0548632108,0.5648096204,-0.051976271,0.2402831763,-0.0033581983,-0.2404185534,0.8079634309,-0.1601192504,0.1449321955,-0.1521190852,0.0144081768,0.1299853921,0.0817263871,-0.0439451188,0.1221086681,-0.0180419404,0.2865327895,0.4060718119,-0.1061451808,-0.1722283512,0.6839924455,-0.1587841064,-0.3064936101,-0.5154694915,0.2594402432,-0.0833602324,-0.1040883884,-0.0893619657,-0.2415319383,0.1011879146,-0.0491485894,-0.2093528211,-0.0508578643,0.0632777959,-0.0003044298,0.3115297556,-0.2914239764,0.1588661522,0.1056731343,0.2688990533,0.0791351944,-0.149276495,-0.0986581743,-0.5794999003,-0.1806600392,-0.045007322,0.2516980469,0.2487960905,-0.1453986466,-0.1473960876,0.0242087692,0.2118916661,-0.3574889302,-0.1288211197,-0.0893364996,-0.0330672786,-0.1616015583,-0.0592811406,0.2779622376,-0.201993227,-0.2204404622,0.2499269098,0.3106345832,-0.1495818198,-0.1366345435,0.319273144,0.0140753314,-0.3008784652,0.1790655553,0.2081694752,0.0955212116,0.3621181846,0.1707606018,0.0461801477,-0.3094235957,-0.1551953703,0.1100846156,-0.1713668108,0.156402275,0.3066341579,-0.0415061824,-0.046380233,0.0741301253,-0.0691770092,0.1868039519,-0.0259992536,-0.2497932166,-0.5798814297,0.2423989922,-0.3692207932,0.2138345242,0.1777445972,0.1073682085,0.1290202886,0.359695673,-0.4105057418,-0.1084878147,-0.2615824938,0.0865874887,0.3704781234,-0.141598016,-0.2241394967,0.2723584473,0.2161535919,-0.130759269,-0.1625023186,-0.2899391353,-0.1773122549,0.020196775,0.1920600832,0.0582296476,0.0196274761,-0.0970007852,-0.0400527306,-0.1090442613,-0.0722374767,-0.2886530757,-0.0877607539,0.4601534605,0.0511234738,-0.3587554097,0.1381819248,0.2278088927,-0.0953808799,-0.065047279,0.1049649268,0.1383260787,0.1468421817,0.0682649761,-0.1363201141,0.0525396131,0.0267356839,-0.0613617152,0.1481487751,0.1850190312,-0.0511845946,0.29288885,0.0760764852,-0.0953093544,-0.017536059,0.3519930542,0.2867044508,0.2604888678,-0.3781736195,-0.1735209972,0.2324880511,0.0812512338,0.2040199637,0.1865293086,0.2776302397,-0.0854997784,0.4448855221,-0.0653338507,-0.0694187954,-0.162834689,0.1486913264,-0.042389445,-0.0171948075,-0.1453141719,0.3029360771,0.0961693376,0.1659192145,0.125116691,-0.0669469014,0.3911999762,0.0129607823,0.052794043,0.4942438006,0.1641958803,-0.0703725964,0.1747995764,0.2477718294,0.1639556885,0.0619069748,0.6308472753,0.1184157133,-0.1244700849,-0.0901787058,0.4095795155,-0.201705128,-0.1735340655,-0.0195567794,-0.0060505951,-0.0434364416,0.0930594057,-0.1290021837,-0.2444870919,0.1135217026,-0.1031239703,-0.1108415648,-0.4786436856,0.249922201,0.2681171298,-0.1131946892,-0.056335371,0.3791209161,-0.1896380335,-0.1202309802,0.0543105565,0.3533880711,0.0409336835,-0.0135541074,0.2910169959,0.0431528762,-0.054229781,0.0152226537,-0.1003461033,0.2382562608,0.2102110237,0.0859878808,0.1227213591,0.1685882509,-0.1836933047,0.1728665382,0.0523833781,-0.2447250634,-0.4073019922,0.1545042396,0.2836321294,-0.2948231101,-0.3668750823,0.0825244561,-0.3997980356,-0.1746642441,0.3004310131,-0.1143867895,0.0535423271,-0.3095144033,0.1591244787,-0.2277983725,0.4056253135,0.2534192801,0.1969311833,0.0589150302,-0.3001429439,-0.2808842957,-0.071220018,0.1046813354,-0.1971353143,-0.1770918667,0.0270696953,0.0219800081,0.0487918556,-0.021075394,0.1069547608,0.1803552508,-0.1950109601,-0.3059131503,-0.0508167818,0.0075041121,0.1850045472,0.0853630826,0.2851344943,0.0427959301,-0.0452753268,0.0452368483,-0.2873758376,0.1741995662,-0.2233403474,0.1665140986,-0.0308966842,0.1551309824,0.3072985411,-0.2111860663,0.1113241166,-0.318849206,-0.3840657175,-0.2382074296,0.1619792581,0.1285139322,0.2065514475,-0.2644241154,0.2039341182,0.1446891129,0.1718867719,-0.1405296475,0.1691948771,-0.0279970076,0.3262259066,-0.1360148937,-0.0489914641,0.1092481017,0.198446691,0.1030922458,-0.2329329103,0.0792168304,-0.1998874396,0.0871883109,-0.2807664573,-0.0923199952,-0.13941136,0.1674966216,0.1780989021,-0.2341704816,-0.5522790551,-0.1351096332,0.3741740584,0.0456082039,0.0194926523,0.6326849461,-0.0136145903,0.0989050195,-0.0845122635,0.273673445,0.1917676926,-0.0527961291,0.1451330632,-0.2625232935]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"You can just run: \r\n`val = nlp.load_dataset('squad')` \r\n\r\nif you want to have just the validation script you can also do:\r\n\r\n`val = nlp.load_dataset('squad', split=\"validation\")`","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":24,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nYou can just run: \r\n`val = nlp.load_dataset('squad')` \r\n\r\nif you want to have just the validation script you can also do:\r\n\r\n`val = nlp.load_dataset('squad', split=\"validation\")`","embeddings":[-0.0479782261,0.0604858324,-0.0539134853,0.2444889098,0.248679772,0.1877578199,0.2922648191,0.3124673367,0.1899903119,-0.0237427168,0.0557831042,0.4075591564,-0.1347435117,-0.0079085575,0.2922923267,-0.2682780921,-0.1929560602,0.3292351067,0.1848603189,-0.06305556,-0.0684112608,0.1494150758,-0.2461704016,0.1525854617,-0.023350941,-0.1903261989,-0.0376415066,0.2420651466,-0.0751293227,-0.5705810785,0.2770506442,-0.1046772152,0.2867957354,-0.0100499829,-0.000113336,-0.0096305776,0.343242377,-0.2085264027,-0.285394311,-0.2410947084,-0.2574113607,-0.2226096839,0.4168426096,-0.2224058658,-0.1222060472,-0.0949645787,0.2801200449,0.0533588342,0.4846663475,0.4920531213,0.1911881864,0.2663962841,-0.2778134346,-0.2081823498,-0.2290327847,-0.0079413252,0.1164551675,-0.0115763806,0.0253511015,-0.1536256969,0.1678515226,0.0487463437,-0.1323892474,0.0573955439,0.2352476269,-0.1945355237,-0.1781451106,-0.2897628248,-0.0892733857,0.2384080291,0.3575641215,-0.0509690531,-0.1911022961,-0.0961414874,0.2534791529,-0.2780938447,0.1107778773,0.2565834224,-0.1418790221,-0.0119595416,-0.2243394703,-0.1495238245,-0.1682105362,0.400627166,0.2719267607,0.3005495965,0.0308479033,0.2198183835,0.4388645291,0.0139065729,0.0014640251,-0.1625842303,-0.0448136143,0.2608413398,-0.5341176987,-0.2073041797,-0.0824583173,0.186087817,0.0907499194,0.0411117375,0.2947389185,-0.0586644523,-0.0574181005,0.1990403235,0.508874774,0.1017294005,0.5623295903,0.2690935731,-0.1490220129,-0.1593780369,0.1682632715,0.0441463403,-0.1811117679,-0.1665756404,0.1207617223,0.1402468085,0.1163438037,-0.2006215453,-0.2761594355,-0.0517085679,-0.1991660744,-0.1768129617,0.160023123,0.5370280147,-0.0615797006,0.3261684477,0.0482984595,0.1132508442,-0.0572101362,-0.4197391868,-0.1476211399,0.3679234087,-0.072527118,-0.0302502383,0.4270748198,-0.0798367113,0.4679146409,-0.0434464402,-0.1424437314,-0.0100079579,0.2092293948,-0.1657940745,-0.0835500583,0.0574743077,0.379348278,-0.0478072017,0.2634988427,-0.2280272841,-0.1109257713,0.1095157266,-0.2067266852,-0.1016824841,0.0654660612,0.1888945103,-0.022871878,-0.131171301,-0.2601767182,0.1217983738,0.1881858557,-0.3063747585,-0.3416291177,-0.2337040454,-0.1677280217,-0.1121920049,-0.0668830872,0.1756419539,-0.2444734275,-0.0179958232,-0.0542966984,-0.0141140576,-0.1190045178,0.4451571107,-0.5217581987,0.4278976917,-0.1365601718,0.3605508506,0.9311334491,-0.3305851221,-0.0189148113,0.2314217836,0.0573704392,-0.2136706859,-0.0486658849,0.2341494858,0.1505651623,-0.0110798329,-0.0230927672,0.4446450472,-0.1218503714,0.069405362,-0.0394676141,0.0232371893,0.4373553097,0.159803167,-0.0309362561,0.1563685089,-0.034846168,0.495734483,0.3241556585,-0.2226721942,-0.1649337858,0.3829777539,-0.1546869427,-0.2373206764,-0.1041943505,-0.1764299124,-0.6095246673,-0.1106410772,-0.1527284682,-0.1248762459,0.2874699533,-0.0954664946,-0.3576019108,-0.0612154044,-0.0567842312,0.1104555428,0.0263876989,0.2085289955,0.1452484578,-0.2851122916,-0.3506065905,0.2338497043,-0.0766568854,-0.0593928844,-0.5049280524,0.1625627279,-0.1459736973,-0.1190996319,0.3205271661,-0.0516694896,0.2130184472,-0.0669354722,0.0374815054,0.3208777606,0.1133230999,-0.1674942374,-0.0458593778,-0.2503063977,0.1178534776,-0.1653389484,-0.082371451,0.2613770068,0.2618916631,-0.1377428621,-0.2714335918,0.3402355313,-0.191697821,0.2066088021,-0.0078753596,0.1686403155,0.3393786252,-0.092761755,-0.2258027941,-0.0144079085,0.2171371728,0.1687519699,0.1625562608,0.0007494665,-0.4013223052,-0.240463987,0.4795569777,0.1693531424,0.1784728318,-0.0569449067,0.2091763914,-0.025562143,0.1343107671,-0.1709836125,0.5497861505,0.1769136786,-0.2952726483,-0.0186025538,-0.2508745193,-0.1387842894,0.1455360502,-0.1979945451,0.2664521635,-0.001689852,-0.0935718343,-0.1395549476,-0.3698905706,-0.1032401547,0.2697143555,0.281714946,-0.3868989944,0.1220140383,-0.26945436,-0.3693347871,-0.1848937869,-0.0826629475,-0.2725851834,-0.2749228477,-0.0287234131,0.1694056988,0.187444672,0.219473809,-0.0533871651,0.1847233474,0.0460549854,0.122603789,-0.0121899676,-0.0678897053,-0.2740035057,0.0332430974,0.1798577458,0.2468972653,0.2460195273,-0.1143594384,-0.256623894,-0.0241387524,-0.0743716657,0.1052277461,-0.0097435946,0.1189465523,0.2973913252,0.3271968365,-0.0906634629,-0.1738584936,0.182153821,0.095028244,-0.0800516754,0.1346051395,0.0533426777,-0.1240585148,-0.2009021491,-0.5390253067,-0.1912579685,-0.2482042313,0.0243600085,0.2482373118,0.312301755,0.3128586411,0.0227755103,0.2032660842,-0.2936572134,0.0312259886,-0.2376646549,0.1938053221,0.1072007045,-0.0977869555,-0.458922714,0.03418741,0.0552466288,0.1852573752,-0.1203773245,-0.1107425466,-0.5271216631,0.0227927882,-0.1498503983,-0.3265810609,0.1050831974,0.3287192583,-0.2310402989,0.0733636245,-0.0461928286,-0.2906912267,0.2586800158,0.0462521799,0.365660876,-0.1040059254,0.4038248956,0.0919928402,0.2871785164,0.2608461082,-0.2382416129,0.2712729275,-0.2320260108,0.2584132552,0.2038302124,-0.4950603843,-0.0847604647,0.0858800337,0.0767835528,0.145523116,-0.3406988382,-0.2208052129,-0.1559219211,0.0492631346,-0.2039716095,-0.0615674779,0.1771132052,-0.2447161674,-0.0318931863,0.1497231424,0.1860647798,-0.337456435,-0.0963255987,-0.2990653813,0.0830657035,0.1593766361,-0.0426067039,-0.625782609,-0.2466262728,-0.2012993395,0.2173017114,-0.0509004593,0.5578956008,-0.0921858698,0.1652602255,0.103403464,0.063557826,0.5721996427,-0.1936546862,0.0759709626,0.055616565,0.1774015278,-0.4670735896,0.07038147,-0.3027267754,0.4461676776,0.3199688792,0.2301204503,-0.3077149391,-0.2546513379,0.23427625,0.0646215454,-0.393843919,-0.1513882577,-0.3432256281,-0.2765830159,-0.3246466219,0.0321079157,-0.099881418,0.3159375489,0.2772572041,0.1718533784,-0.1050194651,-0.1600511819,0.2639561296,0.2484743893,0.1817944944,0.2546616197,-0.2329726219,-0.342979759,0.2199472785,-0.3948250711,0.356292516,-0.0791478455,-0.2387234122,-0.0230564401,-0.1043956205,0.6284286976,0.2375016958,-0.0911512449,0.0006651203,-0.0127430558,-0.0968484432,0.1345953643,0.1691795886,0.2501634657,-0.1953588426,-0.2360312492,-0.4948010743,0.1556424201,0.0602439344,-0.0246247686,0.1110296547,-0.4291245639,-0.2869417071,0.1914908588,0.0332706906,0.7203471065,-0.1108689308,0.2077295184,0.3283588886,0.220575422,0.5223879814,0.1833228916,0.1967042685,-0.5505108237,-0.0884159207,0.0163084902,-0.1553128064,0.0538390502,0.5587201118,-0.3569832444,0.4578364193,0.050852634,0.0682478547,0.1064595208,0.7116006613,0.1096309572,-0.2842377722,-0.7899659872,0.1267939359,-0.4717500508,0.0126648434,-0.1502150446,-0.2712709606,0.241021201,-0.4336148202,-0.2351602614,0.4454449713,-0.0808401629,0.1855414212,0.1141599342,-0.0236039553,0.147852093,0.6605551839,0.0363125913,0.246555835,-0.0157780256,0.0055703488,-0.2947255075,-0.0225113537,-0.3566211164,0.1375783086,0.2956491411,-0.1165754423,-0.1978688836,0.2053549141,-0.0247939918,-0.3660856783,0.0459995903,0.3323644698,-0.1188983545,-0.2764553726,-0.6099253893,0.0294744279,-0.080860436,0.0434286259,0.0761771649,0.1332083195,-0.371081382,0.1783420444,0.3886562586,-0.096660167,0.2243727148,0.3577516973,0.1713744253,-0.0878246948,0.6338263154,0.2177982479,0.2036120147,-0.3888981938,-0.0880616158,-0.0150341922,-0.197756514,0.1592378765,0.087997824,0.0146867493,0.0019775692,0.2046621442,0.0043500373,0.301401943,0.0867831334,-0.4593501687,-0.0889242515,0.064652659,0.0695057213,0.0562211275,0.279078722,0.059227623,0.0778173357,0.2417650372,-0.2793127894,-0.0918926448,-0.3776773214,0.343188554,0.03254788,-0.1979169697,-0.0710441545,-0.0240744743,0.0881160796,0.0156934857,-0.1319460571,-0.2038966864,-0.050701391,0.1307488382,0.1073057279,0.1659913361,-0.2348480523,-0.2871066928,-0.1090496182,-0.0624414794,-0.2199464589,-0.1065064147,0.0051832069,0.4725281,0.3254975677,0.0438868627,-0.0175343249,-0.0497449487,-0.4380334914,0.0234620776,0.0018056475,0.3404190838,0.3198777735,-0.0159825087,-0.513222456,-0.043813508,0.1945156157,0.1373007149,0.0317850746,-0.3519869745,0.1275759339,-0.0616409108,0.1934378594,0.4508307278,-0.2184756845,0.3574877083,-0.0891413689,0.1199365035,-0.2625109851,0.0365061574,0.4811249077,0.0406514667,-0.152913779,0.3276671171,-0.015340751,0.0515311584,0.1491159797,-0.1526934505,0.3365018964,-0.1881406307,0.0778990239,0.1299314648,-0.1292014718,-0.215659067,0.1182215288,-0.0606214106,-0.0235823486,0.0413591228,-0.3875613213,0.2492305785,0.0137856891,0.0008749777,0.2242959142,-0.2114547938,0.0166216251,0.3268390596,0.0624799058,0.0522455573,-0.0715785399,0.2790295482,-0.210857898,-0.2023442835,-0.4202898741,0.3439577818,-0.0849120021,-0.0385797471,-0.6270728707,0.0836673602,-0.163369,0.031855002,0.0022622251,-0.0733572468,0.2121777236,0.077726163,0.0106974244,-0.3616233766,0.2238462716,0.3764090836,-0.1437568218,-0.280601114,0.3382047713,-0.1337412596,0.1470300257,0.1625492275,0.2682202458,0.2369654179,0.3678376079,0.202509433,0.0287724212,-0.1917094439,-0.1172227859,-0.0705474988,0.1597934365,0.2279228121,0.4306640327,0.2431881875,0.1567758322,-0.0250945818,0.1761787534,0.024499217,-0.1788792163,-0.2493199408,0.5915536284,0.1455297917,-0.0688037872,-0.4301485717,0.2644997239,-0.2930421531,0.0178336836,0.2483346015,0.3755817711,0.2748734951,-0.2838099003,0.0425066166,-0.1007415056,0.5464732647,0.0283303149,0.0814291239,-0.2912140191,0.0581281818,-0.6404224038,-0.0053227572,-0.1164413914,-0.1911016405,0.1937851906,0.2920219004,-0.0143481446,0.3954774141,-0.2576817572,0.2110134959,-0.0510238372,0.011996381,-0.2657921016,0.024393877,0.2058585137,0.3686724305,0.0010752419,-0.2293819487,0.0077199587,-0.0486579351,0.0039965338,-0.1935209334,-0.1978333294,0.0505485013,0.3167338371,-0.1383808702,0.1197504476,0.5713409781,-0.0127289081,0.0639797747,-0.0495303161,-0.1536050141,-0.306173116,0.2583884001,0.111023739,0.3103417754,-0.2092684358,0.3361272216,-0.2611889243,0.267896831,-0.287009865,-0.137658447,-0.1113615781,-0.0054199952,-0.0969751999,0.2928805351,-0.0532116033,0.0230846386,0.0039807898,0.1733213067,-0.2453134209,-0.1909745336,0.3813109398,-0.3571410477,-0.338865757,0.0115267495,0.2736949027,-0.20245336,-0.1865314245,-0.5078135133,-0.2016619891,0.3373521566,-0.2002734393,-0.1685628444,0.2662326992,-0.118324481,0.0067152418,-0.0444777831,-0.2023692578,0.1116032377,-0.3773908019,-0.1624005437,-0.3672365844]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"If you want to load a local dataset, make sure you include a `.\/` before the folder name. ","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":18,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nIf you want to load a local dataset, make sure you include a `.\/` before the folder name. ","embeddings":[-0.1817345172,0.1649202257,-0.037762545,0.3960107863,0.2116742432,0.0808849931,0.2252638936,0.2089404166,0.2736941576,-0.0764815882,0.0226199981,0.5386043787,-0.1813205481,-0.0346676223,0.266469568,-0.3476087451,-0.1070255116,0.3149689138,-0.065006569,-0.2064713985,-0.2161854804,0.2870637476,-0.2059969455,0.1824874878,0.032629434,-0.1372919679,-0.1574772447,0.3898998797,-0.0368854105,-0.4234907329,0.3749963343,-0.1197535917,0.3778780699,0.09478122,-0.0001058318,-0.0234758928,0.4701956809,-0.1295060962,-0.3101872802,-0.3924332559,-0.2680528164,-0.3620916307,0.4882286489,-0.2194731832,-0.0158795342,-0.2597392797,0.2412916571,0.0264009163,0.2533610463,0.3210203648,0.2670232356,0.1399294883,-0.2075907886,-0.2245950252,0.0458516367,0.1774975359,0.1134933233,0.1632965505,0.0754975826,-0.1619021595,0.1334751695,-0.0079500368,-0.1729715914,0.2086657733,0.280320853,-0.0951739773,-0.1461393088,-0.0608950183,-0.1281757504,0.2202229351,0.59426862,0.0227832869,-0.0991348475,-0.0659364983,0.1251932681,-0.1856793761,0.1517226398,0.282074064,-0.2423073351,-0.0171662848,-0.2248252332,-0.2694967091,-0.1921048313,0.4482887685,0.2587141395,0.2339924872,-0.1192826629,0.3122357726,0.3364270329,-0.0419210531,0.1714497656,-0.1993420869,-0.0360708274,0.2063961625,-0.3502089083,0.1492244601,0.0149744209,0.3842139244,0.05865217,0.0374896303,0.2414616644,-0.0637469664,-0.2141748369,0.1980293542,0.4698019028,0.0535757579,0.3585375249,0.2473382503,-0.1437019259,-0.2810351849,0.0694392845,0.0354901962,-0.2649764717,-0.0487420484,-0.1562064737,-0.1527792662,0.1816767603,-0.2239640802,-0.122301437,-0.2737606168,-0.1520317495,-0.1668013781,0.1815780699,0.4278885722,0.0906486213,0.298003763,0.0953643769,0.1498009562,-0.1026716456,-0.2856143117,-0.2298306674,0.1412283182,-0.1225494593,-0.1385333836,0.3747474551,-0.0187855698,0.4567199349,-0.0382334664,-0.2033585906,0.1767637879,0.1393191069,-0.0794531778,-0.2095606625,0.1328118742,0.311127305,0.0777217746,0.3022845984,-0.1965751052,-0.2245519012,0.1427998692,-0.3064493239,-0.1693765372,-0.146987319,0.2282441705,0.0955979079,-0.1747365445,-0.3551029563,0.1555950344,0.0108269304,-0.298584342,-0.3533901572,-0.1492172182,-0.1736250669,-0.1328094453,-0.0644598827,0.2461013347,-0.2640273571,-0.0133301113,-0.2493095249,-0.0340757743,0.1630187035,0.5130516291,-0.6319713593,0.327796489,-0.2078633755,0.391122818,0.755734086,-0.0893503949,-0.0843905434,0.4410306513,0.0372952782,-0.3427175581,-0.0661499798,0.3210615814,-0.073901467,0.0138341179,0.0989211723,0.5014772415,-0.0818543658,0.0567304194,-0.1080970913,-0.1147874147,0.3714604378,0.2119090855,-0.2093863338,0.2779374421,0.1506364942,0.4080718756,0.2564155757,-0.2179135233,-0.1621754169,0.3962439299,-0.1703134179,-0.1292039603,-0.1280872673,0.0495205075,-0.5174013376,-0.0258927997,-0.1493400037,-0.0429201834,0.2196694314,0.0843060687,-0.2865209877,0.0157589074,-0.0331116728,0.1085644737,0.0955378339,0.416079998,0.0223479699,-0.1381156743,-0.2010146528,0.1912926733,-0.1028567702,-0.188134402,-0.3548856974,0.2209466398,-0.1161208898,-0.1161725074,0.3425400257,-0.1195624396,0.2132452428,-0.0790648386,-0.0497983433,0.2547598481,-0.1155869365,0.0983075649,0.1258719265,-0.240930602,0.0365218557,-0.2851820886,0.1046782658,0.2166626602,0.3297916353,-0.1805115491,-0.3861504197,0.1478818059,-0.0720380694,0.2228576839,0.002266607,0.0943218246,0.4035722315,-0.0046275281,0.0257201679,0.1011727303,0.3559525013,0.3568498492,0.255599767,0.0850090161,-0.4724255502,-0.1945292503,0.2176447213,0.1026500836,0.1617790163,0.099341698,0.0584509447,-0.207224384,0.2317270637,-0.0166771412,0.6256718636,0.2888614833,-0.062880002,-0.0638780519,-0.2169774175,-0.2650143504,0.1154557988,-0.1677089036,0.2183244228,-0.0975164622,0.0157232191,-0.1876262873,-0.4450456798,-0.1722655594,0.1685965806,0.3621346951,-0.2655333281,0.1152258292,-0.3535382748,-0.5387765765,-0.0472406931,-0.2088954449,-0.2409414798,-0.2782775462,-0.1543750018,0.2228322178,0.0525749214,0.1316055655,-0.1063755602,0.2991155982,0.0982860923,0.0415978134,-0.1295877248,-0.0537600778,-0.4215118289,0.0692125857,0.3457301557,0.3664290309,0.2584702075,-0.1480545402,-0.0782568902,-0.1410404593,0.1053793803,0.1019720212,0.0545456074,0.2188705057,0.2128574997,0.2562524974,-0.1707851738,-0.2850758433,0.2987315059,-0.037657328,-0.0720784739,0.1106504947,-0.0113610905,-0.1662681848,-0.2586746514,-0.350299418,-0.2985755503,-0.3545612693,0.1117272228,0.4160958827,0.2045212537,0.2739645243,0.0068997019,0.3108428121,-0.1213606074,-0.0468598045,-0.1962010264,-0.1599239409,0.1169947758,-0.1250785738,-0.4658184946,-0.0345596671,0.1089994088,0.1442873031,-0.0446926579,-0.3045373261,-0.2324936241,-0.0408554673,-0.0980673134,-0.1314141303,0.1571474522,0.2371114045,-0.199622333,-0.0385133922,-0.0430273414,-0.2611531913,0.3550449312,0.0848537534,0.2365913391,0.0896642506,0.2459274977,-0.0686401725,0.2741695642,0.1861712635,-0.1141704693,0.4656280577,-0.1571580768,0.2868110538,-0.0404679328,-0.3110336959,0.0076055527,0.153792575,-0.1186019704,0.3943250179,-0.0438434258,-0.474465996,-0.159142524,-0.0732011944,-0.1936897784,-0.0308248252,0.150909692,-0.1315547973,0.0101213707,0.0121903848,-0.1107870787,-0.2863282561,-0.2158373743,-0.1978515834,0.0440807864,0.0579860285,0.101277858,-0.5978105068,-0.0726605281,-0.3372056186,0.2020888478,-0.0547832586,0.6055858135,-0.2325655222,0.0905442387,0.086597234,0.0851052403,0.434148252,-0.0885391086,0.0687682778,0.2186235487,0.1043595597,-0.4736560583,0.1123694107,-0.2625893652,0.426608026,0.0638911724,0.3056513667,-0.1797006428,-0.2165188491,0.032158494,0.1714894325,-0.3322618008,-0.2128717154,-0.1436229795,-0.3173755705,-0.2811239362,-0.2209930122,-0.2068213969,0.2611551583,0.274191469,0.1687981188,-0.1617898345,-0.0724516213,0.1129973754,0.0159394555,0.220764488,0.1923493445,-0.2700276971,-0.1895432621,0.2842895687,-0.269683063,0.493673116,-0.1684418619,-0.0299714953,0.0951571837,-0.1637626439,0.4011535645,0.1444170177,-0.1325805783,-0.1154080033,-0.0261235442,-0.0657980964,-0.0323980339,0.143111214,0.216551438,-0.2004897445,-0.0818391815,-0.4670593441,0.4266169667,0.0104191201,-0.1125413254,0.2631275058,-0.3820769489,-0.2714160085,0.3012168407,-0.0442962274,0.6325564384,-0.2386367619,0.327773422,0.3764719665,0.2650222182,0.5416570306,0.3012344539,0.0501421206,-0.3201082945,-0.0265939403,-0.0102580804,-0.1556917131,-0.0250745527,0.4631872177,-0.3125990033,0.2900803983,0.0298200995,-0.1297743022,0.178837195,0.5517043471,0.0873577371,-0.2644074261,-0.8121371865,0.2312103957,-0.3571052849,0.0336257257,-0.1185195521,-0.2851407528,0.1299961805,-0.2430936098,-0.1001197696,0.4421347976,0.0696010739,0.2493881136,0.0401049666,-0.113990441,0.001416078,0.509995997,0.0257469825,0.0212784614,-0.2656074762,-0.0200615264,-0.4163828194,0.0331865437,-0.2824682295,0.1166371331,0.1996988356,-0.1085787863,-0.0332871191,-0.0084261699,-0.0831880569,-0.4496490955,0.0269915573,0.1286884248,0.0059284279,-0.0635255724,-0.6231586337,0.1030215621,-0.0645145178,0.087068826,0.1293968558,0.2588725686,-0.1879108846,-0.1054455042,0.3657416403,-0.1877488047,0.1452238858,0.350510776,0.0596682951,-0.0442221276,0.5409495234,0.3071980178,0.1004933044,-0.2943285406,0.0865003914,-0.0235573817,-0.1124792024,0.1435960531,0.0707216337,0.1952658296,0.0336386263,0.3984982669,0.1098665446,0.1383949369,0.1088799983,-0.5213492513,-0.1246241182,0.1196967736,-0.1379736364,0.1303384304,0.2263134271,-0.0683588684,0.3383349776,-0.0009442222,-0.3291742802,0.0199278165,-0.4328753352,0.189015016,0.1621651351,-0.0478649363,0.0083954362,0.1013941988,0.1302054524,-0.192321822,-0.1621753871,-0.2228136212,0.0071930089,0.1053413749,0.1618339419,-0.001895958,-0.1585660875,-0.4806967378,-0.0356518626,0.0098626465,-0.2220204324,-0.0680057108,-0.1960415542,0.3081574142,0.2666198015,0.0075459206,-0.1272087246,0.1897006929,-0.4071490765,-0.0079394579,-0.1021506116,0.217810452,0.3239803612,-0.1289044917,-0.3088864982,0.0502029322,0.0872261077,0.1594093442,0.1079364195,-0.2871794701,0.2242118567,0.0353735238,0.2761371732,0.2336180508,-0.190080151,0.3175784051,0.129858315,0.1888832599,-0.2455729395,-0.1276438981,0.3067604005,0.0863678828,-0.1376015395,0.1601119488,0.045871567,-0.2316598147,-0.0420913957,-0.1830097735,0.3428035378,-0.2322958708,-0.0806592107,0.2617074549,-0.2149302065,-0.0969157442,0.038874466,0.1646554619,0.1013289094,0.4209339321,-0.2638904154,0.1601245701,0.0836495236,0.1174877584,0.1718549579,-0.4299756289,0.1223289669,0.3331015706,0.2134533077,0.0374951474,-0.1245476082,0.2487550229,-0.2603924274,-0.0767705366,-0.2401843965,0.3554292619,-0.1318928301,-0.1976371557,-0.5061555505,-0.0679941326,-0.106144011,0.0043146112,-0.1662182212,-0.1071427688,0.2077886015,0.0833817348,0.070195578,-0.5049276948,0.1436966509,0.3242912591,-0.3905444741,-0.2331097722,0.3381960094,-0.1477967203,-0.0204021577,0.1488845199,0.2539529502,0.2729275823,0.5522407889,0.1210468113,0.0156944804,-0.0684001073,-0.1189374849,-0.2865125537,0.1086864322,0.1187445372,0.4551871121,0.2992774844,0.2070508003,-0.0704742074,0.3092947602,-0.0950277522,-0.0797183663,-0.2047464848,0.5447618961,-0.0118542351,-0.0640676171,-0.2537020743,0.1976335049,-0.3508046567,0.016287405,0.3670661747,0.1425317973,0.2990792096,-0.1212267503,0.0745313615,-0.1345239133,0.6167289615,0.1394236684,0.3182078898,-0.2117817849,-0.2057549357,-0.6120727658,0.0046664784,-0.1309651434,-0.1546331942,0.2182592601,0.2286336571,0.1870756447,0.2939270735,-0.2870681286,0.2898222804,-0.1562624127,-0.0523818098,-0.245122388,-0.098032102,0.0994081497,0.2930144072,-0.1242988482,-0.3223796487,0.0919067785,-0.1275496632,0.0963745341,-0.2215860784,-0.0952981487,-0.0226894263,0.2877829373,-0.0963434204,0.189383924,0.4010134637,0.0056790477,0.0589185953,-0.1345385611,-0.3568746746,-0.2299482971,0.05842641,-0.0629678592,0.3110414743,-0.4552665055,0.4634615481,-0.3302415311,0.016149085,-0.1946723908,0.0959753841,-0.1624245942,0.1784181297,-0.2633728683,0.0764708146,0.0178104043,0.0366632119,0.1744442284,0.0213586949,-0.1419451684,-0.0555831604,0.4751200378,-0.4024738669,-0.333753407,0.1060202867,0.0798722655,-0.1831517667,-0.1411480606,-0.2714113891,-0.1178144291,0.4060054719,-0.1564534009,-0.1739277542,0.1983254552,-0.0602462366,-0.0233197883,-0.116582945,-0.009512376,0.1224937215,-0.3754437566,-0.127486378,-0.4122233689]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"This happens by just doing run all cells on colab ... I assumed the colab example is broken. ","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":18,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nThis happens by just doing run all cells on colab ... I assumed the colab example is broken. ","embeddings":[0.0188431982,0.0450551845,0.005720363,0.3553819954,0.0485084765,0.0158459563,0.3805162013,0.0017092479,0.0825143605,-0.0696413219,0.0609751008,0.6260883808,-0.1543067843,0.0177123547,0.2570040226,-0.1367096156,-0.0055067092,0.4703560174,-0.0491582714,-0.1032542363,-0.2574828565,0.298918128,-0.4601308405,0.0343436487,-0.1575132459,-0.1436825246,-0.2623033822,0.0875611082,-0.0323915444,-0.269321382,0.4908155203,-0.0376952179,0.159505859,0.2392023057,-0.0001199438,-0.1626406014,0.381934613,-0.186781764,-0.5980202556,-0.3188870847,-0.4396148622,-0.3556466997,0.5800086856,-0.2417218685,0.0015593286,0.1016805172,0.2442041337,0.2477970272,0.2513045967,0.2270814329,0.1859471053,0.0525158793,-0.2828712165,-0.2271952927,0.1020056307,0.0441174731,0.0202824473,0.2397984862,0.2175246328,-0.2253180593,0.2749579251,0.0992982015,-0.2582588494,0.1961469203,0.0591719523,-0.109739691,-0.190532133,-0.0806436539,-0.0734864399,0.2947736979,0.4633131027,0.1014977396,-0.0970329121,-0.1807356477,0.3349782526,-0.3535569608,0.250633657,0.1607245505,-0.3357418776,-0.1015135497,-0.2035338283,-0.1373180896,-0.0174727831,0.4194001853,0.1564692557,0.2983800769,-0.2599446476,0.2820228636,0.3466849923,0.0848815516,0.4672145844,-0.093956925,0.0315395258,0.1105768159,-0.3193621635,0.0937875733,-0.076409094,0.505810082,-0.0126253339,-0.2257357985,0.3272550106,-0.1687375307,-0.2140426338,0.2629674375,0.3502497971,-0.0067188279,0.3153295815,0.0902891457,-0.1434864402,-0.1687917709,-0.0142858811,0.0906808451,-0.3385536671,0.0852106139,-0.1945907325,-0.1255175024,0.1719267815,-0.2188440561,-0.2274805903,-0.2521557212,-0.3828362823,-0.0296623483,0.205755651,0.4171341062,-0.0021472008,0.1563944817,0.1263044477,0.0278350506,-0.168648079,-0.3810290396,-0.1644434929,0.1161909774,-0.216347456,-0.0061730864,0.3435818553,0.1051118821,0.4116132855,0.0553373024,-0.0924310535,0.1041881815,-0.0682782307,-0.330776304,-0.1765857339,0.1721393317,0.3596381247,0.0152125657,0.3292273283,-0.3052338958,-0.0320951939,0.3024981022,-0.2944398522,-0.0814929903,0.0301410239,0.0952656344,-0.0199779645,-0.0424025916,-0.5787671208,0.3203098178,0.3033498228,-0.5106570721,-0.3087828755,-0.3915750384,-0.2988180518,-0.125619635,-0.0953124538,0.2600243986,-0.1438380927,0.0360734016,-0.0859902725,0.2935400903,0.2463108003,0.3319583237,-0.5736114383,0.3138142526,-0.2048135698,0.2774712741,0.6275325418,0.1795435995,-0.1897026598,0.1105682924,0.1969002634,-0.141344741,-0.0844787359,0.2261341959,0.0577903315,-0.1141991615,0.2488618195,0.2980818748,-0.1208086163,0.0419587828,-0.1281765401,-0.0940658003,0.3527612388,0.0641502291,-0.1838858277,0.3009245098,-0.0646800622,0.3978241682,0.3637025356,-0.1571041346,-0.249550283,0.2259039879,-0.0780370981,-0.2327934653,-0.1085262969,0.1071589887,-0.4472696483,-0.1631046981,0.1275351197,0.0777038932,0.2421364337,0.004942107,-0.2343296111,0.046577353,0.1284436882,0.1167520732,-0.0086372411,0.3010903895,-0.1533358693,-0.176939562,-0.1945761591,0.3904251754,-0.1771713942,-0.1402009428,-0.2490786612,0.253318876,0.0073272223,-0.2020684928,0.2641592026,-0.1066093072,0.243747279,-0.0436390564,-0.1106656715,0.0648488924,0.0664474815,-0.1450033188,-0.1039982289,-0.2828156054,0.0237217899,-0.0298546068,-0.0533010475,0.0986916944,0.3930721283,-0.2130803764,-0.190332368,0.2627229095,-0.0860733241,0.2650961876,-0.0253574699,0.1188815162,0.2733763158,0.1228569672,-0.0279482231,0.1626714617,0.30422616,0.4033070505,0.0843398198,0.1841973513,-0.4527641237,0.0301091485,0.3008410335,0.1992160529,0.2322489023,0.0064755501,0.0678080991,-0.181238696,0.4324338734,-0.0643158108,0.6145316362,0.1886730939,-0.209035486,-0.1017479748,-0.1130301058,-0.2819030881,0.1876810342,-0.1347732395,0.4792538583,-0.3165477216,0.1207220033,-0.1672209501,-0.3869832754,-0.2485431433,0.2913652956,0.4233625829,-0.3012811542,0.2501380742,-0.3041700721,-0.5747203827,-0.0290657133,-0.2319538295,-0.0292114057,-0.2642417252,-0.1924958229,0.4679270387,0.2398442924,0.1629972905,-0.0738499388,0.4010507464,0.0614264533,0.2328597307,-0.0830552876,-0.0852954164,-0.4240858555,0.0136715453,0.1367811561,0.2953564227,0.0146755073,-0.2348014712,-0.244991526,0.1951456219,-0.0592318028,0.2362014651,-0.1503009498,0.1712605357,0.2429655641,0.1682321727,-0.562274456,-0.1670964211,0.0664915964,0.0026001146,-0.0596728399,0.0490776971,-0.0145267639,-0.2116800249,-0.1921958923,-0.0983457044,-0.2190615237,-0.1602042913,0.0941900536,0.2214883566,0.2686265707,0.3534071445,0.0601304024,0.3149447143,0.099223353,-0.0172488596,-0.2775678635,0.0167652499,0.0780190602,-0.0512637049,-0.2939233482,-0.0442048237,-0.00979516,0.1574009955,-0.0309384875,-0.3212945759,-0.4990661144,0.0285517722,-0.2699047029,-0.1555912346,0.1095616743,0.1630758941,-0.1443767846,0.111897856,0.0000050772,-0.3624994159,0.2137898505,0.0498596504,0.2663090825,0.0962129757,0.2913758755,-0.0136638312,0.1354088187,0.2348683625,-0.0575198159,0.3559513688,-0.1440768987,0.2338394672,0.0250850432,-0.4332297444,0.1090022624,0.0465004519,-0.1990036815,0.3114969432,-0.147105217,-0.4056760669,-0.1029729992,0.0083678681,-0.2438027263,-0.1495487839,0.2076834738,-0.2232062221,-0.0823238939,0.1431369931,0.0253842678,-0.3486604989,-0.1726819873,-0.5382248759,0.1132899299,0.1111399159,-0.0439457223,-0.1952192783,-0.2355070263,-0.3916324377,0.2378993481,-0.0027547735,0.644654572,-0.01960776,0.1463010907,0.225941807,0.1995600313,0.520560801,0.0790817738,-0.1048813239,0.2530302107,0.0220156834,-0.3757421374,-0.0236886814,-0.3089157045,0.7288348079,0.1766836345,0.0662020072,-0.1405593157,-0.0932530537,0.1985871345,0.1059836,-0.2695719004,-0.1981988251,-0.2073459774,-0.4354094863,-0.166730091,-0.0864372998,-0.09890683,0.1802148223,0.0933300033,0.1534482986,-0.2424213439,-0.1563525051,0.2513750792,0.278608501,0.0825914815,0.3105608523,-0.2169618756,-0.4663379788,0.0374566056,-0.1172831729,0.4401257634,-0.1858834475,-0.0011638994,0.1874312609,-0.2095351964,0.4377322793,0.3021081686,-0.1805868149,0.0607620701,-0.2463490069,-0.1503035873,0.3578484356,0.0961571783,0.1690749675,-0.0551968515,-0.1837340295,-0.5243776441,0.3472335935,0.0418853238,-0.1612322927,0.2484366894,-0.3539991677,-0.2773648798,0.2996807098,-0.110945642,0.8321897984,-0.1266225427,0.3009485602,0.5401297212,0.4465813041,0.6798884273,0.3885042965,0.28030774,-0.3319865167,-0.3079812527,-0.0232887045,-0.2238513082,-0.0678480193,0.3628305793,-0.411280334,0.3376837373,-0.0278395787,0.0099772355,0.3609881401,0.1533302367,0.0450644828,-0.2034812719,-1.0773434639,0.1218764707,-0.3595482111,0.0161050875,-0.2016160935,-0.140235275,-0.0046623233,-0.1980497539,-0.1052008197,0.4397706091,-0.2203986645,0.0834875256,0.0974925607,-0.1798941195,0.1056626439,0.5122752786,0.09337271,0.1080747619,-0.0910507292,-0.083584033,-0.0587084889,0.0929103419,-0.2837283015,-0.0805540159,0.2159754187,-0.0194965061,-0.0091663823,0.0911281854,0.007368383,-0.4445006847,0.253213644,0.3105249107,-0.0502014868,-0.1410366893,-0.5808526874,0.2743006349,0.0435475037,-0.0704939663,0.0331012607,0.184922561,-0.2720305026,0.2190475911,0.332603842,-0.1810185015,0.1619433165,0.2498017997,0.2368979603,0.1104009748,0.6567296982,0.2099250257,0.0971609056,-0.1672793776,-0.0176593363,-0.1620028019,-0.2184258997,0.2865261436,-0.0895909891,0.2046187371,0.0978025571,0.6650971174,0.1055327281,0.2791699469,0.1492865682,-0.4848052263,-0.0581244528,0.1349754333,-0.0207399335,-0.0040335562,0.3663780391,0.1007788926,0.1261033267,0.0105664032,-0.2117127031,0.0826556757,-0.3914166093,0.1768348664,0.1331156194,-0.0426456779,-0.0066920207,0.0670283213,0.017817596,0.0762797743,0.0228195954,-0.1450711638,-0.0434146188,0.1743065268,0.1842657477,-0.0238612741,-0.3253518641,-0.3507259488,-0.1480284631,-0.0674501508,-0.2227751315,-0.1148283333,-0.1403856874,0.066670157,0.3694604039,0.1092573181,-0.0782339945,0.2560463846,-0.4652413428,-0.0711704567,0.0496548451,0.2274327576,0.4017964602,0.0891726613,0.0078914193,0.0463515259,-0.1639361233,0.0542703122,0.1405020803,-0.4412919283,0.3304993808,-0.2029120177,0.0232587699,0.3467922807,-0.1097396687,0.1332849413,0.1245111153,0.0473389812,-0.2183277458,0.0846269727,0.4756410122,0.2179626673,-0.1550711095,0.5182563066,0.0236436706,-0.1052301526,-0.0653344095,-0.1741686314,0.1846060306,0.0539895184,-0.072395198,0.0390972756,-0.2783755064,0.0799001157,0.0419401079,0.0248784255,0.0975355804,0.1697599292,-0.29428038,0.1484504789,0.0453889742,0.0297055505,0.2482106537,-0.4228446782,0.2555705309,0.3607267439,0.2210371941,0.0565914698,-0.1093819216,0.2068733275,-0.483093977,-0.1412041187,-0.2477655411,0.3302391171,-0.2287642211,-0.094954595,-0.4242097735,0.079579927,-0.1871645898,-0.0233737696,-0.3070195019,-0.0418385901,0.2699076533,0.0926278979,0.0871969834,-0.4432079196,-0.056602031,0.4643257558,-0.163533628,-0.1599615216,0.3797776699,-0.3749161363,0.0549514033,-0.0238972604,0.3070366383,0.3153679669,0.5219579339,0.2288387567,-0.2398249954,0.0558712259,-0.0621773228,-0.2804094851,0.0601927377,0.3870798051,0.5215955973,0.1456529647,0.0719813704,-0.0021659688,0.5232652426,0.0035667347,-0.2093378007,-0.2321674228,0.2833399773,-0.0056838575,-0.0855437368,-0.2916556299,0.2349633425,-0.2003579736,0.2329581827,0.2146114111,0.0314249136,0.2805596888,0.1035914868,-0.0037103964,-0.0278489739,0.4672179818,0.1311962456,0.3271560669,-0.3506749868,-0.2680202425,-0.4513780177,-0.0028916376,-0.2462493777,-0.1295735836,0.0787519738,0.0188913122,0.1488024443,0.2157799006,-0.2876194715,0.2032906562,0.005948063,-0.037146017,-0.2495369017,-0.3417991996,0.1039941683,0.0421294756,-0.1263803095,-0.1977806538,0.0610503666,-0.1242572218,0.0076415986,-0.3355777562,-0.1542006284,-0.0714069903,0.1461085379,-0.1325714588,0.227431491,0.428499192,0.0202704445,-0.0244105309,0.0168936942,-0.3717769384,-0.2659429014,0.1922811121,0.0226681586,0.2352156341,-0.2476634681,0.5727081895,-0.204247579,0.0751401484,-0.0598158464,0.2185975462,-0.1207904816,0.2023949325,-0.3862576187,0.2626579702,-0.0179839395,0.0288401954,0.0260131545,0.1373887807,-0.2358583063,0.0463779308,0.5678297877,-0.4367074668,-0.2931664884,-0.0399847478,0.1249179617,-0.1795124263,-0.2163804322,-0.4107367098,-0.2804537714,0.3210200369,-0.2351149023,-0.2631281912,0.4131294489,-0.1097609475,-0.0727500916,-0.1793256849,0.2224081755,0.0321225934,-0.4685643911,-0.2566356063,-0.3495836854]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"Oh I see you might have a wrong version of pyarrow install on the colab -> could you try the following. Add these lines to the beginning of your notebook, restart the runtime and run it again:\r\n```\r\n!pip uninstall -y -qq pyarrow\r\n!pip uninstall -y -qq nlp\r\n!pip install -qq git+https:\/\/github.com\/huggingface\/nlp.git\r\n```","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":53,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nOh I see you might have a wrong version of pyarrow install on the colab -> could you try the following. Add these lines to the beginning of your notebook, restart the runtime and run it again:\r\n```\r\n!pip uninstall -y -qq pyarrow\r\n!pip uninstall -y -qq nlp\r\n!pip install -qq git+https:\/\/github.com\/huggingface\/nlp.git\r\n```","embeddings":[-0.1487929672,0.177358225,-0.0415252633,0.3428395987,0.1430013031,-0.1136175469,0.1551361531,0.1810095608,-0.1386966854,-0.0235533211,-0.1086137295,0.824855566,-0.1088479236,0.0025473528,0.2896052301,-0.273955971,-0.0902934968,0.4029020965,-0.076595448,-0.0639468357,-0.1341270208,0.3876859844,-0.2954268754,0.1872269958,-0.1233507991,-0.1301395297,-0.1366636753,0.1866870224,-0.2076076269,-0.4219763577,0.3289388418,-0.0730949938,0.2341246754,0.2311662287,-0.0001097652,-0.0474842824,0.2344948649,-0.0567841865,-0.2778341174,-0.3594297469,-0.09514191,-0.3835330904,0.6282963753,-0.3426682949,-0.061945729,-0.013691307,0.2064764798,0.5288489461,0.3657016456,0.3335526288,0.2460253388,0.2046816349,0.012928457,-0.1019961089,0.3749512136,-0.0657254681,-0.0736380145,0.4479817748,0.2363180369,-0.2795596421,0.1032444909,0.0722953826,-0.2124039084,0.1664639562,0.1779628992,0.0769344494,-0.0671167746,-0.123390466,-0.1497291774,0.115436919,0.3498918712,-0.199906677,0.0064630695,-0.0781896412,0.1789008677,-0.3727684319,0.1460716277,0.3057048917,-0.3531241417,0.0832639858,-0.0329890773,-0.181336239,-0.1379807144,0.3700060248,0.1592617184,0.3465178907,-0.0843496472,0.2189971656,0.3824252486,0.0736254156,0.4075033367,0.008381254,-0.036072854,0.0784764364,-0.1490985155,0.1150645837,0.0018030468,0.4292596579,-0.0158935711,-0.1278882474,0.2102989256,-0.0398970097,0.1295638382,0.2672181726,0.0784155056,0.1771277487,0.1082350761,0.0198865347,0.0342881978,-0.0640786588,0.0264675692,0.0537834838,-0.1929517835,-0.1257613897,-0.0986878201,-0.1596484482,0.2492925823,-0.1578544676,-0.3151191175,-0.3915263712,-0.3593124747,-0.0904713646,0.0977066159,0.4054092169,-0.1776591688,0.0079453774,0.1842356175,0.2125342637,-0.3329145312,-0.2086358815,-0.141910702,0.1688598245,-0.3038463891,-0.0869803727,0.3669921458,-0.086929746,0.4350621104,0.0915842801,-0.090644829,0.0990218073,-0.0513410829,-0.1203421578,-0.2850756645,0.1791954339,0.2982243299,-0.0311093982,0.1415540427,-0.3812169433,-0.1345567703,0.3453596234,-0.2035247684,-0.2004272789,-0.1206995025,0.1847997904,-0.022888355,-0.1863526404,-0.3829165697,0.1329852343,0.1375317723,-0.2288273126,-0.2935593426,-0.2705018818,0.0196134262,-0.2460684329,-0.0110319657,-0.009930037,-0.2936679721,-0.0903062448,-0.1332146525,0.04531762,0.2246524394,0.3951141834,-0.529274106,0.4179309309,-0.1316703111,0.3782483339,0.6474629045,-0.0438345782,-0.3523705602,0.077712886,0.0100973314,-0.131932348,-0.1612619311,0.2264738828,0.0355200097,0.0866481587,0.2476299405,0.2930698097,0.0733571351,0.1554504782,-0.340482682,-0.2222441286,0.2725386918,0.198863402,0.0550471209,0.0438532643,0.0352111273,0.1302620322,0.3609609306,-0.2110332102,-0.2131267786,0.1836090833,0.1716726571,-0.0342067294,-0.1786791086,0.0559241809,-0.3501668572,-0.0080574574,-0.1969958842,0.2563146055,-0.0577348173,-0.0100110248,-0.2123849392,0.1170311123,0.0096824169,0.0178452879,0.1186448783,0.2356435657,0.0754365325,-0.0449417457,-0.0608420111,0.2400655001,-0.0117313126,0.0717653781,-0.1706349701,0.3285128176,-0.2863952219,-0.4844674468,0.2462784499,0.0502860248,0.1879673451,-0.0319809169,0.0527980439,0.1622606963,-0.1116875559,-0.202608183,-0.1910264045,-0.1893150061,0.0705778226,-0.318967849,0.0696857125,0.0835235938,0.2004163563,0.0006734337,0.0999838635,0.2382402271,0.0537171625,0.0103423893,-0.0235097781,0.0162903983,0.1499005109,-0.1002154723,0.1077844501,0.3029182255,0.3645248115,0.319503665,0.1586574167,0.0025900907,-0.4805756211,-0.024724666,0.3976528347,0.1972628981,0.1791994125,0.0537569784,-0.0763690695,-0.045942273,0.3068538308,-0.3248250484,0.4616414607,0.3096235693,-0.2889766693,0.0340869911,-0.2853967547,-0.2893974185,0.137168631,-0.093845509,0.308326751,-0.0694536492,0.2494655699,-0.1373800337,-0.468562901,-0.1146634743,0.0716922432,0.3634001315,-0.2764385343,0.1655787677,-0.278968811,-0.5766425729,-0.0026118536,-0.3778031766,-0.2139499485,-0.1601440459,-0.0080164643,0.3329255581,0.251232326,0.2919051349,0.2842282951,0.1532880813,-0.0797626078,-0.0617443994,-0.1177826375,-0.2035913169,-0.4503104985,0.0792020112,0.2295377702,0.2759786546,0.1467198431,-0.2758328021,-0.1061238423,-0.0484323017,-0.2091184407,0.0643989444,-0.1999296844,0.4625612795,0.1053710431,0.1716955453,-0.4071997106,-0.3779241443,0.3300964832,-0.1246658117,-0.1140806824,0.0141204325,-0.066068992,-0.1999975592,-0.1609916091,-0.2568092942,-0.3521341681,-0.3320152164,0.2843401432,0.3300653398,0.183504656,0.2762922049,0.1325851828,0.3061435819,-0.0540185533,0.0262596291,-0.1248756647,0.0732353032,0.0929615498,-0.1124174595,-0.4067388177,0.0862567574,-0.0165202506,0.1904916912,-0.0490399972,-0.4750238955,-0.4325562119,-0.0386491977,0.3313347399,-0.1380579025,0.1436620653,0.3142635226,-0.077102609,0.0089175552,0.0811164603,-0.2799062431,0.1458631903,0.1194532737,0.2612983584,-0.0505969338,0.0772078484,-0.0448262095,0.5389634371,-0.0806145668,-0.0731455907,0.4156589806,-0.1554380357,0.1861876696,-0.0001052632,-0.3473629653,0.0560716242,-0.0079891598,-0.0489121489,0.4476047158,-0.1462386996,-0.4494740069,-0.1367118955,0.0507570691,-0.3940086663,-0.2110321671,0.2257069796,0.240978539,0.0329285525,0.1777135432,-0.0429693572,-0.1713814288,-0.2843979597,-0.2230813801,0.0512322485,-0.1399466097,-0.0382215604,-0.2054777294,-0.3146076202,-0.5384998322,0.2506504357,0.0899122804,0.1832946092,-0.0573804826,0.0498424694,0.3154447675,0.0160135515,0.4428637624,0.1561343521,-0.1624365747,0.2648038566,0.1484660953,-0.4350921512,-0.1480485797,-0.3881222904,0.4172604382,0.3217190802,0.113253504,-0.3579915464,-0.0952270254,0.1670903265,0.1364838183,-0.1777441353,-0.098592639,-0.3406661153,-0.4646257162,-0.2337496728,-0.0615029037,-0.0564704873,0.3894998133,0.0827373266,0.2343580425,-0.1019006446,-0.0572529361,0.0408180505,0.2335130721,0.2398320287,0.0712827519,-0.1877723634,-0.4298273027,0.0924191549,0.0448195785,0.4507398009,-0.1198705211,-0.1727010012,-0.0238322522,-0.1074680239,0.3415601254,0.1617317498,-0.1103234291,0.0401134342,-0.2229598016,0.0378062017,0.2822964489,0.1910190433,0.273284018,-0.0520990416,-0.1497420818,-0.1134603322,0.3765210807,0.0413663462,-0.0979813412,0.0757907927,-0.2930542231,-0.3040894866,0.336338222,-0.0172186401,0.8864101768,-0.2272506207,0.3322289586,0.6621405482,0.1691019386,0.5987753868,0.1311426014,0.3515995145,-0.5247064233,-0.1329174638,0.0987381712,-0.1258141994,-0.0833816603,0.2061784714,-0.4300662577,0.2210561186,0.2416702807,-0.1076160371,0.2421873659,0.2347363234,0.1185161844,-0.0621187091,-0.8111822009,0.1840382516,-0.4015726149,0.1016213745,-0.3028538525,-0.1754026413,-0.0820066482,-0.2573376596,-0.1481730342,0.2711634636,-0.4599965513,0.0919734538,0.2421209365,-0.3207073212,-0.0675393566,0.3489339054,-0.0428769402,0.0800161138,-0.0843871385,0.0468044765,-0.1234174296,-0.2193293869,-0.1237828881,0.1357638687,0.1368616968,-0.1435557455,-0.1878620833,0.130559817,-0.0406766236,-0.4170719385,0.1590569466,0.1850437969,-0.0326084234,-0.138705954,-0.3320928812,0.155729562,0.0752578154,-0.1019872501,0.1313642114,0.086199984,-0.1734533012,0.2305032611,0.2186102271,-0.2491340041,0.0388650224,0.3013150692,0.0388518609,0.0676339865,0.4714342058,0.344555378,-0.0536276847,-0.2484150231,-0.1738594025,-0.3065652847,-0.3511999249,0.1581276506,0.1769619137,0.1844577044,0.1050079986,0.7097862363,0.2515309155,0.1808908582,0.0911257267,-0.3801748455,-0.0572451912,0.2111606151,0.0917122737,-0.0420955457,0.4282212853,0.1901574433,0.3156459332,-0.0352550857,-0.3332714736,0.1473167688,-0.4213024974,0.2532201707,0.0160000809,-0.2545118034,0.0547904596,0.1210601628,0.1523280591,-0.0012898976,-0.0823921934,-0.2743921578,-0.0119351558,0.0950860456,0.0822044015,-0.0388382301,-0.1826976091,-0.2417631745,0.1309063286,-0.0190201029,0.0004280359,-0.0873029009,-0.1583036482,0.1622682065,0.3484469056,0.0771066546,-0.0654413924,0.1697943211,-0.4177083075,0.1224730238,0.1762703806,0.1888025254,0.2052430212,0.024384452,0.0199927296,-0.0593147315,0.0609834269,0.1278047413,0.0298902187,-0.4590283632,0.0442098975,-0.2240297645,0.1944714934,0.2536652982,-0.1431824714,-0.0038238815,0.3047591746,0.2181076258,-0.3287240863,-0.0076390072,0.3376449943,0.0793381929,-0.079542987,0.2885996103,0.0738833174,-0.1289464682,0.0766499713,-0.0376940258,0.2600875795,0.0634455159,-0.1863857657,0.0894152299,-0.1593292505,-0.0205677915,0.1780695617,0.1595655978,0.1040639803,0.0581663996,-0.4287500381,0.0446683876,-0.1730945259,0.1529623866,0.0876862556,-0.2966554463,-0.0928230211,0.2903237343,0.2871262431,-0.1108151525,-0.0921285897,0.4882983267,-0.0525403582,-0.1273430884,-0.1616660506,0.4248256385,-0.3100778759,-0.0690412149,-0.2973684967,0.0200318117,-0.1517922133,0.0838347003,-0.2869226635,-0.18728365,0.2494694293,0.2284654379,-0.0803936869,-0.5375557542,0.0824732259,0.3234293461,-0.2376470864,-0.0669940338,0.3535562754,-0.2561561465,-0.0691310018,-0.2235434651,0.4734697044,0.3910302222,0.4207758904,-0.0952729136,-0.1849465817,-0.0888456777,0.0087611228,-0.2569767535,0.0604863726,0.4327039421,0.6803268194,0.1491388679,0.1645188183,-0.1627497822,0.3199814558,-0.057844948,-0.2080640644,-0.092957072,0.2306423187,0.0797842145,-0.2521179914,-0.1542225927,0.0813414231,-0.3745661974,0.3997107744,0.1591039598,0.0810647532,0.1519899964,-0.1392211616,0.0746280551,-0.1444069296,0.5761428475,0.3628770709,0.5610783696,-0.3690725565,-0.3454785943,-0.507691741,0.1809564829,-0.3644898832,-0.1576549262,0.0980733335,0.1633273065,-0.0162168555,0.3809639215,-0.2118591815,0.2227020562,-0.0548863076,-0.0331606083,-0.4557057321,0.057696566,0.2077846229,-0.1075277925,-0.0050814212,-0.1848664731,0.0224077515,-0.3520956039,0.0984448865,-0.3358082175,-0.1187064573,0.0855861902,0.2473828197,0.237248227,0.1274773329,0.3981240988,0.0424264781,0.0091262367,0.0060326774,-0.4042062759,-0.2243943363,0.137031123,0.0498385914,0.2697416246,-0.189566642,0.3870378733,-0.2441124618,0.2650884688,0.0390811525,0.0683938786,0.1275362223,0.0294493176,-0.1323510408,0.0084868716,-0.111344412,0.1461021453,0.1385046691,0.0935584754,-0.4371018708,-0.1693151742,0.4644240141,-0.4737013876,-0.2482483685,0.0474209897,0.0651879013,-0.1243952066,-0.2313201427,-0.3892985582,-0.1924515367,0.3040780425,-0.1035837606,-0.2666764855,0.3316687346,-0.1010993049,-0.0110998983,-0.0621120818,0.1679121852,0.1046942547,-0.3704235554,-0.3236937523,-0.4475359023]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"> Oh I see you might have a wrong version of pyarrow install on the colab -> could you try the following. Add these lines to the beginning of your notebook, restart the runtime and run it again:\r\n> \r\n> ```\r\n> !pip uninstall -y -qq pyarrow\r\n> !pip uninstall -y -qq nlp\r\n> !pip install -qq git+https:\/\/github.com\/huggingface\/nlp.git\r\n> ```\r\n\r\nTried, having the same error.","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":65,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\n> Oh I see you might have a wrong version of pyarrow install on the colab -> could you try the following. Add these lines to the beginning of your notebook, restart the runtime and run it again:\r\n> \r\n> ```\r\n> !pip uninstall -y -qq pyarrow\r\n> !pip uninstall -y -qq nlp\r\n> !pip install -qq git+https:\/\/github.com\/huggingface\/nlp.git\r\n> ```\r\n\r\nTried, having the same error.","embeddings":[-0.1167454496,0.1842945516,-0.0243378729,0.4034665227,0.1603583395,-0.0801019296,0.2126228809,0.1767457724,-0.1010505483,-0.0292401705,-0.0643654838,0.8222458959,-0.1264523864,-0.0137211913,0.3064845204,-0.3175931871,-0.0604344048,0.4053914547,-0.0765027627,-0.073327899,-0.1381295174,0.4036580622,-0.3136930764,0.175076738,-0.0893364996,-0.1419267952,-0.1608567387,0.1332477778,-0.1906817406,-0.4354628921,0.3544144034,-0.0906923413,0.2426740378,0.2432136685,-0.0001137994,-0.0411222465,0.241860643,-0.0652759075,-0.3070584536,-0.3983516097,-0.1334444284,-0.3729758859,0.6542971134,-0.3005972505,-0.0274334233,-0.0415602624,0.2386431545,0.5376312733,0.372884959,0.3187704384,0.2171253115,0.1901440918,-0.0212035589,-0.1347409785,0.3434083462,-0.0393818244,-0.0503408164,0.4133885205,0.2552083433,-0.264321357,0.1400381029,0.0521855764,-0.2430432737,0.1733251661,0.210436061,0.0216798708,-0.0455762856,-0.1478258073,-0.1497154981,0.1675877124,0.3823014796,-0.1461519599,-0.0075807711,-0.0868382379,0.1697665751,-0.3536817133,0.1502903253,0.2991668582,-0.3662483096,0.0429872498,-0.038737949,-0.2004193366,-0.1347132027,0.4069931209,0.1866998523,0.3191030025,-0.08083345,0.2694012523,0.4425932467,0.0977581143,0.3867422938,0.0186899193,-0.0305929445,0.0945214853,-0.2340413332,0.1411139369,-0.0283439737,0.4344560504,-0.0264845882,-0.1392008662,0.1912462711,-0.0719764531,0.0989132524,0.2541507781,0.1694355756,0.1823100895,0.1392900199,0.0455916077,-0.0352401771,-0.0994457081,0.0498521775,0.075848937,-0.2112872303,-0.1360410899,-0.0762299001,-0.1687743366,0.2506899238,-0.1541397721,-0.3055714071,-0.3601968884,-0.3899203837,-0.1199926063,0.1078698933,0.3979915977,-0.114564009,0.0822231323,0.1959000975,0.2271794677,-0.3056085706,-0.2729312181,-0.1145991012,0.1753754765,-0.288784802,-0.0823075995,0.3766977489,-0.1048520356,0.3892615438,0.1520421654,-0.1241002977,0.0676041543,-0.0555735901,-0.1365541667,-0.3021424711,0.1692290157,0.2672598362,-0.0021346167,0.2127933055,-0.3595523238,-0.150388822,0.3658851981,-0.2205357254,-0.2247951478,-0.1220563129,0.1412147284,0.028386917,-0.1768839955,-0.444557786,0.1517373621,0.1788907051,-0.2466423362,-0.2768021226,-0.3186874688,0.0077811056,-0.2150366306,-0.0710537508,0.008767209,-0.2900655568,-0.08582706,-0.1462416798,0.0773742869,0.2142516971,0.4388743043,-0.5672301054,0.4444164336,-0.1419268101,0.3385083973,0.678235054,-0.0183102805,-0.32352826,0.0883158296,0.0305543877,-0.1364798397,-0.1348195374,0.2449487597,0.0429064482,0.0743577257,0.2120756805,0.2548865378,0.037367519,0.1282247007,-0.3341502249,-0.2208392173,0.3260882199,0.1719200313,0.0449679792,0.0858539268,0.0007421771,0.1515171081,0.3673003316,-0.194008559,-0.2241534442,0.2160496712,0.1658680588,-0.0558227673,-0.1567902863,0.0522089601,-0.3774954975,-0.0272505376,-0.1940360069,0.2364625037,-0.0366313718,-0.0170107838,-0.2251413763,0.1077531055,0.0393063426,0.0416889675,0.0596334226,0.2118602544,0.0463873968,-0.0546706319,-0.0957266763,0.2498774827,-0.0153152961,0.045964919,-0.2043724954,0.2887839079,-0.2584067881,-0.4483318329,0.2522760332,-0.024650488,0.231608212,-0.0450183041,0.0484644398,0.1449431181,-0.109044157,-0.1970139444,-0.2191610485,-0.1847279966,0.0860859305,-0.3083516061,0.0279370472,0.1044317931,0.2074577808,-0.0422420092,0.0940594897,0.2297780812,0.0535009876,0.0939685702,-0.0705597028,0.0453821644,0.1401529163,-0.056574136,0.1223993227,0.3100036383,0.3866485655,0.3548178375,0.1566017568,0.0273696184,-0.4818149507,-0.0249916818,0.3454561234,0.2102758139,0.1618045419,0.0827300772,-0.0298936423,-0.0505774803,0.3143811822,-0.3307668865,0.4674462378,0.2590411007,-0.3121702969,0.0196408443,-0.2590818405,-0.276958257,0.1626333147,-0.1040933803,0.306971252,-0.1221771836,0.2225687504,-0.1370518208,-0.4423549771,-0.1141143963,0.1217770055,0.4049737155,-0.3246611655,0.1975838691,-0.2822492123,-0.5722625256,-0.0103235701,-0.3882720172,-0.2582880259,-0.1644416749,-0.0355504192,0.3743381202,0.258903563,0.2822013795,0.2758613229,0.220667094,-0.043692071,-0.0294897668,-0.1224362999,-0.1835069507,-0.4691478014,0.0328801945,0.232235983,0.2515956163,0.1045099199,-0.2957598269,-0.1997369677,-0.0619781204,-0.1777971983,0.0704857782,-0.2100956291,0.4583040476,0.1514371932,0.2439417243,-0.4031871855,-0.3655143976,0.3061071634,-0.0678867996,-0.0983480066,0.056174513,-0.101801835,-0.1774573326,-0.1754218638,-0.250941813,-0.299520731,-0.2887907624,0.3235304654,0.2983064353,0.234187752,0.2694034576,0.1197186485,0.2639577687,-0.0644705892,0.0239801705,-0.1363691986,0.0608765669,0.0808318183,-0.062325146,-0.3854481876,0.0866534561,0.0299976803,0.1967360675,-0.0461696647,-0.489589572,-0.4529497921,-0.0650658906,0.3251758516,-0.158857137,0.1583993137,0.3022809923,-0.0926661566,0.0465169959,0.1072315574,-0.2898403406,0.2133358717,0.1094611287,0.2854808271,-0.0139041962,0.1235152185,-0.0094498489,0.548443675,-0.001754429,-0.0684523657,0.3987729549,-0.1526180655,0.1563472599,0.0027434924,-0.3681469262,0.0373674221,-0.0233739745,-0.0695350617,0.4168310463,-0.1659400463,-0.4626934528,-0.1225786954,0.0619474314,-0.3594362736,-0.2139088511,0.2408260852,0.1935857832,0.0531270988,0.1630569994,-0.0553099774,-0.2096183747,-0.2863791883,-0.2207653224,0.0878088847,-0.0869478509,-0.0201714057,-0.2250835747,-0.2970820367,-0.5088524818,0.2509250343,0.0528013371,0.2817326784,-0.0670774579,0.0950974226,0.3330951333,0.0427666269,0.5087607503,0.1918318719,-0.1700252146,0.2541190088,0.1583372504,-0.4275511205,-0.1468093246,-0.3640681505,0.439719528,0.3146083653,0.074054569,-0.3733355701,-0.1124892533,0.1465473324,0.15660353,-0.2032291144,-0.1083292514,-0.3091737926,-0.4813622236,-0.2311211675,-0.0402821638,-0.0518497713,0.3798216283,0.1105995923,0.2730440497,-0.0871806219,-0.0902626365,0.0736916214,0.1693742126,0.2320990413,0.0732723475,-0.2000032365,-0.4365999103,0.0514616184,0.0156591311,0.4611103833,-0.1330996156,-0.1863541156,-0.0111387735,-0.1330201775,0.3412382901,0.2154463083,-0.1333307624,0.0738788173,-0.2124446183,0.0228207223,0.2874016464,0.135630697,0.2971384525,-0.0679823458,-0.1731099933,-0.1615945101,0.3831522167,0.0418414213,-0.1218295395,0.1178747788,-0.3031163514,-0.3134897649,0.346806258,-0.0441373251,0.9466443062,-0.2616964877,0.3528294563,0.6437761188,0.2162544131,0.6181933284,0.2029841095,0.3590887785,-0.5230041146,-0.1619655192,0.0692746937,-0.1588440984,-0.0696186125,0.2365116924,-0.4519900978,0.2443508357,0.1636417806,-0.1168134063,0.2646887898,0.2421243042,0.1128875688,-0.0383287109,-0.8219125271,0.1458720416,-0.4373737872,0.0863008648,-0.3026201725,-0.1839605719,-0.0540654138,-0.2869914472,-0.2158607394,0.2583190799,-0.4710239768,0.106359072,0.2658198774,-0.3001655936,-0.0291272122,0.3820653856,-0.0198591147,0.0335832126,-0.0936835781,0.0491719283,-0.1269505322,-0.1726061851,-0.1665309221,0.1063394025,0.1398371011,-0.1086089313,-0.1786043197,0.1221827418,-0.0622655116,-0.4694956541,0.132144168,0.1977358162,-0.0869303197,-0.1385254115,-0.3822513819,0.1494188011,0.0735637397,-0.0707661211,0.0941329375,0.1329763234,-0.1765105873,0.2160206288,0.2014975846,-0.2504373789,0.065322265,0.3018831611,0.0679555088,0.0366781764,0.5193562508,0.3891953528,-0.0338010341,-0.2297708988,-0.1690052748,-0.3114826381,-0.3626984954,0.1972350776,0.1226458326,0.2215544879,0.0838548839,0.7151266336,0.2434430122,0.227590397,0.126963377,-0.3996832371,-0.0358035639,0.191385448,0.0748594701,-0.0275887735,0.3770537972,0.1519477665,0.3207682371,-0.0033023332,-0.2881686985,0.150667116,-0.4082630575,0.2566252053,0.0366633758,-0.224699989,0.0363082737,0.1087992042,0.1077068001,-0.0298582241,-0.0883532539,-0.2236032039,-0.0288090035,0.129560858,0.0890350044,-0.0266496092,-0.1756920516,-0.277728796,0.1461302191,-0.0094038015,-0.0264653489,-0.0564798862,-0.1902318448,0.1657527387,0.3668989539,0.058313027,-0.0908922553,0.1691874564,-0.4560787082,0.1232586354,0.1953537166,0.2099875212,0.222993806,0.0554313622,-0.039780315,-0.0212050658,0.031973023,0.1031654328,0.042560596,-0.4590923786,0.0736305639,-0.2225982845,0.1734437495,0.2976996303,-0.1824147999,0.0458635017,0.2626213133,0.16768457,-0.2930977941,0.0001989702,0.3582189977,0.1203975528,-0.1432800889,0.3454214633,0.0932518616,-0.1498586088,0.0990256965,-0.07402464,0.2875246704,0.0109214364,-0.1412149072,0.086641863,-0.1567338258,-0.0386293977,0.1619806737,0.1447479427,0.0718041211,0.054146722,-0.4902257919,0.0099507505,-0.1230560839,0.137706548,0.1187772006,-0.2979780436,-0.0554410703,0.3504873514,0.2865201533,-0.117576465,-0.0961047262,0.4848793745,-0.0940283239,-0.1632601768,-0.1643852293,0.4357664883,-0.2799874246,-0.0767805725,-0.3393844366,0.0047870479,-0.1908966601,0.070670791,-0.2964791059,-0.2076821476,0.2984222174,0.2486559004,-0.0768513978,-0.614030838,0.0274349004,0.3349641263,-0.204739809,-0.0625639781,0.358453691,-0.258689642,-0.1024391204,-0.169890672,0.4860388637,0.4006546736,0.4735705853,-0.0663425177,-0.1409594417,-0.0994687229,0.0307329632,-0.2324665934,0.0167827252,0.406073153,0.6930806637,0.1262867749,0.1186055467,-0.1056386009,0.3538590968,-0.0857059658,-0.1582997292,-0.1626994014,0.2995930314,0.0455165356,-0.2240646631,-0.1628620028,0.1120547131,-0.3453021049,0.3758247197,0.204516679,0.115626581,0.1848961413,-0.1063636318,0.046108868,-0.1243736148,0.5372791886,0.373082757,0.5633397698,-0.3966864944,-0.3218538463,-0.4960481524,0.1362046152,-0.3848460615,-0.1097486466,0.1707843989,0.1453402638,-0.0112818312,0.3810256124,-0.1682498604,0.1817333251,-0.0657246932,0.0121305259,-0.4602122307,0.0471042469,0.2258985639,-0.0463340692,-0.0137390085,-0.1818442494,0.0513876751,-0.3250573575,0.0591216646,-0.3371163309,-0.1214639172,0.1017951369,0.2731238008,0.1851113588,0.1538303047,0.3686857224,0.1065936312,-0.0249294508,-0.0054215672,-0.3920840323,-0.2548120022,0.1399204135,0.0209414903,0.2942177057,-0.2489256412,0.3734029531,-0.2700332999,0.26078251,0.0249035824,0.0597705767,0.0525586158,0.0587041378,-0.1462624967,0.0600807033,-0.0949366689,0.1293771714,0.1723308861,0.0974653065,-0.4798222482,-0.1424223483,0.4880381525,-0.4941724539,-0.2818080187,0.0410798788,0.0696433783,-0.1752572954,-0.2010941058,-0.436858058,-0.2603852451,0.3175222576,-0.1229455173,-0.2913716137,0.3025758266,-0.1120312735,-0.0274077374,-0.0758701041,0.1726703495,0.121152997,-0.4006265104,-0.3341223896,-0.4647310078]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"Can you post a link here of your colab? I'll make a copy of it and see what's wrong","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":19,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nCan you post a link here of your colab? I'll make a copy of it and see what's wrong","embeddings":[-0.0944177732,0.0577484891,-0.0563767403,0.4275850356,0.163964808,0.0742794946,0.2727659643,0.117012091,0.1710002273,-0.0799954087,0.0186920296,0.6328831911,-0.2422109395,0.0163328592,0.2107996494,-0.2766317427,-0.0907839686,0.3511370122,-0.0092701586,-0.0968043208,-0.1953075379,0.3334060013,-0.3628482819,0.0484167077,-0.0346082114,-0.2740375698,-0.1557290107,0.1878253669,-0.1242923886,-0.3305407763,0.4099054337,-0.0354296677,0.3256011903,0.1078403667,-0.0001117283,-0.1089981943,0.3479312658,-0.1518660933,-0.4430922568,-0.3908513784,-0.4066255093,-0.3311358094,0.5561751127,-0.2386394292,-0.0143266022,-0.0093313446,0.2354069352,0.2007259279,0.3323273659,0.2858671546,0.2596550584,0.0442861244,-0.2355460078,-0.2654969394,0.050662417,0.0879464373,0.0988134146,0.163086772,0.0925930962,-0.0973792523,0.1707882583,0.0481715389,-0.2419967353,0.1752585024,0.1400122643,-0.1347237825,-0.1181351542,-0.1624150425,-0.0806269869,0.1478787512,0.5082919598,0.1030740067,-0.0987901837,-0.1000477523,0.1896934956,-0.286546886,0.1539595574,0.1754411161,-0.2739141583,-0.1033988893,-0.1660094112,-0.2197421938,-0.1112829596,0.4456929862,0.2409761995,0.3562979996,-0.2257326692,0.298609972,0.3900255263,-0.0510845967,0.2951830328,-0.1574724019,-0.0734995455,0.131653443,-0.4317269921,0.074965544,-0.0500257723,0.4330421984,-0.0119189359,-0.0132096596,0.1902484596,-0.061675325,-0.2391259819,0.1778040081,0.4587350488,0.0673510954,0.3295468688,0.2016974092,-0.201498732,-0.1453443468,-0.0136575485,0.0482084043,-0.2767910063,-0.0094084013,-0.0984970555,-0.0740562975,0.1964335889,-0.2669284046,-0.2038516402,-0.3185021281,-0.2895675004,-0.1457579732,0.2442164272,0.4916263223,0.020174643,0.3007639349,0.1570000947,0.0578117892,-0.0931734443,-0.3615260124,-0.2038238198,0.1619384289,-0.1468799859,-0.0615600124,0.2930540442,0.0340116471,0.448358655,0.0203078482,-0.1338737011,0.0599753261,0.06086272,-0.2019215226,-0.2287667394,0.0642865524,0.3478239477,0.0982966796,0.3062968552,-0.2995078564,-0.1201545522,0.1795665324,-0.3466897309,-0.07242091,-0.0808773637,0.1977057308,0.0796237215,-0.141462639,-0.4857795537,0.2512125671,0.2131908238,-0.3777722716,-0.3341540992,-0.2774744034,-0.2586355507,-0.1477241069,-0.1644667387,0.192593798,-0.1633934379,0.0110202869,-0.1721020788,0.1691295505,0.2013225406,0.4515893161,-0.6045714617,0.3232270479,-0.1837305427,0.3216724992,0.6954843998,0.0425289907,-0.1176911518,0.2242796719,0.1377721727,-0.1988940984,-0.0700929463,0.3250871301,0.0055549396,-0.0742184818,0.194830209,0.2920792699,-0.070395194,0.0888627917,-0.1292112172,-0.085827291,0.3985470533,0.1036651358,-0.1579093635,0.2847554684,0.0225734469,0.387608707,0.318256706,-0.2133907676,-0.204695195,0.3692450523,-0.0771973133,-0.2113891542,-0.1649078429,0.0796319917,-0.5164647102,-0.148133561,0.0552004538,0.0568614788,0.2273904979,0.0358464494,-0.2773908675,0.0528135635,0.0634371564,0.1129316464,0.060874518,0.3140868247,-0.0475826748,-0.1477338076,-0.1583509892,0.3639943004,-0.1363474876,-0.1746114641,-0.3826625645,0.2217883319,-0.0929973945,-0.1971967965,0.3563100994,-0.1209462881,0.3216608763,0.0056677344,-0.076805912,0.1255416572,-0.0518943481,-0.0745484009,0.0108739641,-0.2348690331,0.0723031461,-0.2150939852,-0.0121060992,0.1338879019,0.3557282984,-0.1919794381,-0.1114845201,0.2194107622,-0.0630843714,0.2133897245,-0.0319266282,0.1152900681,0.3205166459,0.0824629217,0.0267626606,0.1700761914,0.3108051121,0.338758409,0.1589118838,0.096321106,-0.4620636702,-0.0288305581,0.3394735157,0.2506683767,0.1097115129,0.067457974,0.0738063157,-0.2183005214,0.3347753584,-0.0456084646,0.5907067657,0.2570639849,-0.1428965777,-0.0963693261,-0.1601371169,-0.2419663072,0.1867603809,-0.1417369246,0.3764169216,-0.2677262723,0.0374711044,-0.0967654884,-0.3154970706,-0.2467064857,0.2001125365,0.4145114422,-0.3021208644,0.1845778227,-0.3184514642,-0.6122184992,-0.0246766657,-0.2226867229,-0.2144084424,-0.2517209351,-0.2436459512,0.3359796405,0.1686677635,0.1485218108,0.0210639033,0.300331533,0.1470391154,0.2221864313,-0.0986351073,-0.1473153979,-0.3738417029,0.0583635867,0.2481480837,0.2452021241,0.1374836713,-0.1306252778,-0.1697440445,0.0066691185,0.0038388085,0.1205846593,-0.0324912146,0.216229409,0.2202266753,0.2286649942,-0.3838885128,-0.2931216359,0.1811462492,0.0501118302,-0.0635097176,-0.0123661328,-0.0663048998,-0.1467030644,-0.221685499,-0.2276797295,-0.2492044866,-0.2721238732,0.1024553478,0.3819594681,0.2514652908,0.4363299012,0.0330109,0.2322150022,0.018253345,0.0557875149,-0.3442589939,-0.1378060877,0.061183773,-0.0679440498,-0.3699589372,0.0090392902,0.0155556751,0.243000254,-0.0309312586,-0.3123928308,-0.4251820147,0.005747512,-0.1629866362,-0.1913186908,0.1102052927,0.1411829144,-0.162750721,0.049729567,-0.0122396694,-0.4197314084,0.32576859,0.0682948679,0.3534807265,0.0213272069,0.2272073478,-0.0540841185,0.2415468991,0.2091078907,-0.1829463542,0.3785313964,-0.102518931,0.1686152518,0.0769261047,-0.4180788398,-0.0118579948,0.0942209139,-0.1602270603,0.3974246979,-0.1701873392,-0.4542646706,-0.1027727947,0.0393196717,-0.2190036774,-0.1078045368,0.1393344253,-0.1023523435,-0.0790980086,0.08506383,-0.052753102,-0.3985595107,-0.2537576258,-0.3477157056,0.0529253818,0.0223826636,-0.0295982212,-0.3345445395,-0.1595572978,-0.3149395883,0.2395560443,-0.0176621545,0.5357069969,-0.1317476779,0.2084483057,0.1599405706,0.160514757,0.5391771793,-0.0042330432,-0.0770473629,0.2137627453,0.0934879258,-0.3116676509,0.1002289206,-0.2798253596,0.608453691,0.1788612902,0.1645065546,-0.197702229,-0.1621483266,0.0838124901,0.0849958509,-0.289794296,-0.1497540474,-0.167356208,-0.3817217052,-0.2253378481,-0.1936150342,-0.1779948026,0.248066619,0.2114923894,0.1661565006,-0.1954188347,-0.1426351964,0.262514919,0.1269871444,0.1998727471,0.3181171119,-0.2371673286,-0.3918939829,0.1985253692,-0.2465757281,0.4274762273,-0.1285825372,-0.055504173,0.0647449568,-0.1245541647,0.4006834328,0.1586538851,-0.1465181112,0.0738389939,-0.1670156717,-0.0785278082,0.2051887661,0.1892433912,0.2028227746,-0.0654627308,-0.301053673,-0.4792555869,0.3388665617,0.0600720793,-0.1279197633,0.263863802,-0.4850989878,-0.3184443116,0.2754377127,-0.0522007085,0.7567285299,-0.2063623667,0.2419603318,0.4539169967,0.3747141361,0.7180554867,0.3555465341,0.1714846045,-0.3743323386,-0.134727791,0.0468176343,-0.1648560911,-0.0524149984,0.4183011949,-0.4593026638,0.2971616387,0.0908390433,-0.0813943148,0.2911205292,0.3949204683,0.1572984308,-0.2269384712,-0.9205312133,0.184787944,-0.5212965012,-0.0144491913,-0.2380586267,-0.2703250945,0.1267392635,-0.215494886,-0.1769770533,0.3852037489,-0.1663262546,0.2285098284,0.1150970459,-0.1887537539,0.136617884,0.4601589143,0.0072502908,0.055386804,-0.1231565922,0.0063700136,-0.2630661428,-0.0035396079,-0.3486834168,0.0370588824,0.2620079517,-0.0467847735,0.006012396,0.0992759243,-0.0075838417,-0.4604245126,0.0909323618,0.2387206107,-0.1011239812,-0.1637809277,-0.6422834396,0.224445194,0.0163368955,0.0392378457,0.0909119695,0.2503003776,-0.2858886123,0.0664465651,0.3971079588,-0.1849339157,0.1832328737,0.2834944129,0.1624584943,-0.0123606501,0.6115808487,0.2992283106,0.1012273878,-0.2387717515,-0.0709977672,-0.2970315516,-0.1364714503,0.2583658695,0.0133151738,0.2169068605,0.0536291711,0.4977575839,0.0889904574,0.3045551479,0.187696591,-0.4823527634,-0.1277923435,0.0479105413,-0.1715131104,0.0695128813,0.2219394445,0.0558224879,0.1981600225,0.0360918753,-0.2870244384,0.0237156395,-0.4934332371,0.1951367408,0.1133662611,-0.1215149835,-0.0395379551,0.0931236669,0.0978412777,-0.0100641837,-0.100552842,-0.2164046019,-0.0038482568,0.1184237674,0.1497483552,-0.0376645923,-0.229934901,-0.3803801239,0.0246472098,-0.033925157,-0.2382895947,-0.0828281567,-0.0945709869,0.2162041068,0.2533518076,0.0786233619,-0.1160022765,0.2192775905,-0.422082901,0.0385239832,0.0625497922,0.2341906726,0.3411723673,-0.0144390073,-0.1992273331,0.0177835934,-0.050154008,0.057997629,0.1186567545,-0.3623912036,0.2170702964,-0.0274253692,0.0745919272,0.2473858148,-0.1721195579,0.2376957983,0.0521599799,0.1425526291,-0.2117356956,0.1076828018,0.4506113529,0.1334870011,-0.155875966,0.3369011879,-0.0323223397,-0.1187548339,0.0966094211,-0.1801873147,0.2794786394,-0.0734927431,0.0404457822,0.0206928793,-0.221008718,-0.0513839982,0.0358043015,0.1041699648,0.1150034219,0.2257714123,-0.3926174045,0.1944237351,0.07317397,0.0463243239,0.2405050099,-0.4207023978,0.1766010523,0.4530297518,0.1612999141,0.0059533995,-0.1181485653,0.2535956502,-0.3474000692,-0.1329341829,-0.182741493,0.4126595855,-0.2337231934,-0.0821047425,-0.4265695214,0.0290308129,-0.2017479092,0.0100179752,-0.2226506919,-0.1094484329,0.259529829,0.1130048186,0.0602780171,-0.5401207805,0.0179282352,0.3882117569,-0.2209554911,-0.1750015765,0.3850456178,-0.2396029383,-0.0174902938,0.0739506632,0.2733232975,0.3615348041,0.5325784683,0.1489320397,-0.0635512024,0.0309473928,-0.1205691695,-0.2335693389,0.0351258218,0.2925216258,0.533411026,0.2173664421,0.158681646,-0.0499518402,0.3641262352,0.0384728387,-0.198728919,-0.2504662573,0.4240936339,-0.0341579467,-0.0619152412,-0.2673442066,0.1769925654,-0.2738175094,0.0713171884,0.2920418978,0.0804787278,0.3284246027,0.0028235484,0.0526658334,-0.0924561098,0.507956624,0.1694988906,0.4341136813,-0.3372322619,-0.2290872633,-0.5142739415,0.0001975119,-0.0819985643,-0.1618801206,0.0642590672,0.1983007193,0.126223743,0.2476507723,-0.3073264658,0.2391534746,0.024386663,-0.1103061736,-0.3272208273,-0.1412585229,0.1078135595,0.1055522114,-0.0454790667,-0.1948754787,-0.0092759971,-0.0601524673,0.0494782701,-0.3424968421,-0.1001690775,-0.041887559,0.2790625989,-0.1178992838,0.209799096,0.455666244,0.0206730701,0.0670856908,-0.0001961423,-0.4117968976,-0.1785275042,0.1516951621,-0.0067165061,0.3829131424,-0.307526499,0.570931673,-0.271702081,0.1020732597,-0.1034890041,0.0722227395,-0.1684909314,0.1881438196,-0.2853586078,0.1580480635,-0.0204589069,0.0486682616,0.1286192089,0.119767651,-0.2522920966,-0.0476760305,0.5260148048,-0.4421868324,-0.3493867517,0.0554379784,0.0884970427,-0.2051024437,-0.2034863979,-0.4745284915,-0.2382493615,0.4172509611,-0.221027419,-0.1876802593,0.3302192986,-0.0670964196,-0.0023840307,-0.1136216745,0.1298249364,0.1202918887,-0.3896759748,-0.1308841556,-0.4098169208]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"This should be fixed in the current version of the notebook. You can try it again","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":16,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nThis should be fixed in the current version of the notebook. You can try it again","embeddings":[-0.14710401,0.0616018027,-0.0582842454,0.4253875017,0.1894991249,0.0785849765,0.2161241472,0.2134936303,0.1519843489,-0.1275146753,-0.0001826019,0.6510865688,-0.1451359391,-0.0014385447,0.3353012502,-0.3381648958,-0.0270487517,0.2980506122,-0.1423709244,-0.139957279,-0.251378715,0.4030191302,-0.403865695,0.1327827126,0.0217111856,-0.2106446177,-0.1193493381,0.1600602567,-0.0570843592,-0.4420138299,0.2749314308,-0.1162693277,0.3361519277,0.0596038662,-0.0001093193,-0.1074158251,0.4051618874,-0.1475655437,-0.324451685,-0.3263498247,-0.2186392695,-0.3992671371,0.4858598411,-0.236934334,-0.0837220028,-0.2342163175,0.2003939152,0.205556497,0.2355740815,0.3608186841,0.2482601255,0.2883740366,-0.0050674905,-0.2518624961,0.113341637,-0.0598288365,0.0556300804,0.0564066246,0.1227304563,-0.2810391188,0.1201603934,0.1123895049,-0.1244824529,0.140409559,0.32755813,-0.1153151914,-0.1606588662,-0.0343969502,-0.2053136081,0.139947474,0.6114565134,0.0074364212,-0.096150212,-0.0970607251,0.1883187741,-0.1917824,0.1913514286,0.1349308044,-0.2412607372,-0.0447604172,-0.0612316094,-0.2859053314,-0.1500891894,0.5021399856,0.2336249799,0.2242006361,-0.1589904577,0.2421950251,0.3899206519,-0.0083063142,0.2520094216,-0.0961503759,0.0210110079,0.1517125517,-0.4725143313,0.0512609258,-0.0563090779,0.2896998525,-0.0615376681,0.0276436303,0.3553225398,-0.0759071782,0.0143087953,0.1539303809,0.4142656624,0.0991450697,0.3659394681,0.142717585,-0.1154273152,-0.0909942761,0.1692388952,-0.0011916872,-0.1991012245,-0.1065868363,0.0003799207,-0.1910157353,0.2669310272,-0.1325427741,-0.2063051611,-0.1803048551,-0.2383127362,-0.2072892189,0.1692766249,0.386274755,-0.0172584821,0.2506605983,0.3198311925,0.1109934077,-0.1958430409,-0.4604044259,-0.2068077028,0.1390091777,-0.1498809159,-0.1078221053,0.4200163186,0.1649085879,0.4011068642,0.0820918605,-0.2094210684,0.0880855918,0.2217310965,-0.0882465318,-0.2672819197,0.0673334524,0.2882325351,0.0381790437,0.2759985328,-0.1466917247,-0.0678562969,0.2202020735,-0.1441067457,0.0143563533,-0.0321486965,0.1994994134,0.2235759497,-0.1856115609,-0.4228378534,0.261685729,0.184596464,-0.2667084038,-0.3265155852,-0.2504821122,-0.1455685645,-0.0913558602,-0.1373983771,0.0388476476,-0.134717375,-0.0570279807,-0.1768084913,-0.0377319232,-0.0007948495,0.4105513692,-0.5428647995,0.3664244711,-0.1221485138,0.3923411667,0.6489169598,-0.0662516207,-0.2524014711,0.2570919991,0.1391685456,-0.2042081952,-0.0069453134,0.4016405046,0.0115900282,0.0561003871,0.1736215502,0.4234896302,0.0155259436,-0.0658941641,-0.2105989456,0.0261427388,0.3903925419,0.2192244977,-0.0415549055,0.1964025646,-0.0098716207,0.4987840354,0.3632690609,-0.2133999169,-0.1948372573,0.411316663,-0.1184312031,-0.0656650141,-0.037386097,0.0369416513,-0.5206589103,-0.1199092865,-0.196219787,-0.0006293911,0.2955867946,0.0231932737,-0.1861868799,-0.0230503753,0.0493758731,0.0864595547,0.0804345906,0.2497896999,-0.0671436712,-0.0959643349,-0.1293256581,0.3399986029,-0.2296832353,-0.083242774,-0.4938007295,0.1520338804,-0.0635071248,-0.127120167,0.3829320073,-0.0805837885,0.2247433364,-0.0184061583,-0.0018849538,0.2358590513,0.0078952163,-0.1829628348,-0.1754606217,-0.2189997435,0.0580682904,-0.2521507442,0.003222249,0.3137159944,0.3588713408,-0.1463735104,-0.2388788909,0.2632762194,-0.058813192,0.077931717,0.036876671,0.1455331147,0.3402607143,-0.0000087834,-0.0452250615,0.1210005656,0.363027066,0.2222457528,0.2289000452,-0.0109761916,-0.5415526628,-0.166392535,0.236135602,0.1183262765,0.2350784242,0.0695296824,0.0505542159,-0.134726584,0.3139561117,-0.0789088756,0.564648807,0.258366704,-0.1720236689,-0.0106734736,-0.2668840289,-0.2177450806,0.2246934474,-0.1679311544,0.2259047627,-0.1199601144,-0.0455366187,-0.2276471853,-0.3853067458,-0.0195681211,0.2285125256,0.3585436642,-0.3488200903,0.0707585886,-0.2900004983,-0.5700064301,-0.0472625345,-0.0852352977,-0.2404441535,-0.2873762548,-0.1835914999,0.372112602,0.2075570822,0.1819979995,0.1255182475,0.3339088559,0.0761348605,0.3449693024,-0.032909546,-0.0340282954,-0.3831047416,0.0628434196,0.2666669488,0.2797673643,0.1618136168,-0.1823485941,-0.1599879563,-0.1668965816,-0.0793588087,0.0458260551,-0.019588232,0.2268973887,0.2402881831,0.1081629395,-0.2062222213,-0.1785273105,0.2137962878,-0.0022068836,-0.1256932914,0.1754659116,0.024077747,-0.2419172376,-0.1336048245,-0.3721008599,-0.2026903331,-0.2859755754,0.2898208201,0.2603527606,0.1501604766,0.2237322927,-0.0443140492,0.2782638967,-0.0599956773,0.059785068,-0.2615355551,0.1652655751,0.131751895,-0.1247980446,-0.474758625,0.0284088776,0.1332520843,0.2597393692,-0.0391957946,-0.3962197602,-0.2797995806,-0.0628721341,-0.0415767431,-0.2037835866,0.1857272685,0.2249608934,-0.2239194959,-0.0108788814,-0.0394017845,-0.2407147884,0.3065955043,-0.0597395748,0.3338377476,0.036213845,0.2909496725,0.0233216863,0.1673305929,0.2028820068,-0.0382280461,0.3601816893,-0.0221900027,0.2296243757,0.023822844,-0.3766422272,-0.0938368216,0.0425865166,0.0202813298,0.3478186131,-0.1894874573,-0.5239443183,-0.1018915102,-0.0794728249,-0.1185208261,-0.0848715678,0.2039989084,-0.0853741542,0.1128859222,0.1295377016,0.0575596765,-0.2600976825,-0.2331086546,-0.3029342294,0.0225759726,0.0735481158,-0.0427397192,-0.5263240337,-0.0761232972,-0.4482276738,0.2564470172,-0.130502671,0.718678236,-0.1613697559,0.0535440296,0.1632243246,0.1382237077,0.4650244415,-0.0487650447,-0.1252088696,0.295535028,0.0455329902,-0.4988198578,0.0340653844,-0.3732386827,0.4948996603,0.2369887382,0.3017042577,-0.2380533516,-0.2031725198,0.1209122762,0.0351136364,-0.3111352324,-0.1628223509,-0.1695384234,-0.3093952537,-0.2191194594,-0.1260666996,-0.0663816184,0.2515184283,0.2691265345,0.2146985978,-0.2542358637,-0.1581294537,0.1210542545,0.0447969474,0.2813987434,0.1676735282,-0.1435081959,-0.3304818571,0.2267173976,-0.1561353654,0.3116641343,-0.2288347036,-0.2642146945,0.0904803127,-0.3270346522,0.4051325023,0.2067093104,-0.1592459977,0.0610428639,-0.0167705119,-0.0720344707,0.0523717403,0.0226381905,0.3762821555,-0.0188071933,-0.1746628284,-0.4443615675,0.2924353778,0.012178936,-0.1129357964,0.290335536,-0.502586782,-0.3571293652,0.311984539,-0.0031953913,0.6695187688,-0.1953851283,0.3539968729,0.3911889195,0.152891174,0.5130825639,0.2707496881,0.2401058078,-0.4175071716,-0.2042742223,0.0612088218,-0.132992059,0.0462803058,0.4067360461,-0.4205967784,0.2720881104,0.0809929818,-0.1781464964,0.2040997595,0.4951287508,0.0710523501,-0.1758463085,-0.8010065556,0.1794166565,-0.4259902537,-0.0819003657,-0.1838507056,-0.2049735039,0.1103641391,-0.2486983389,-0.1459380388,0.2737151086,-0.143836394,0.3215153813,0.1304227859,-0.1658082604,0.1141297594,0.5232134461,0.0828576535,0.0857463777,-0.1633158922,0.013356043,-0.2903116345,-0.0479972661,-0.2451193184,0.0846784636,0.2167606205,-0.1243449375,-0.1024546772,0.0242515448,-0.0125924181,-0.5824511051,-0.0329106636,0.1788655668,0.0091795763,-0.178531602,-0.5684044361,0.1453222781,-0.0660546049,-0.0733007267,0.145985648,0.2204097658,-0.2523961365,0.063000001,0.2761664987,-0.143791616,0.1990672052,0.4851745069,0.1575772464,-0.1270971596,0.5813419223,0.4267323315,0.0470233746,-0.2995758057,0.0804791227,-0.0023925039,-0.2117257118,0.2594318986,0.0485198684,0.2076459378,-0.0603215322,0.5625927448,0.1421716213,0.1815800816,0.1138183251,-0.4061147273,-0.0834067985,0.1661806405,0.0083853602,0.0418401025,0.2487579584,-0.1257813722,0.240810886,-0.0025826602,-0.3201105297,0.0575891659,-0.3998419642,0.1390158534,0.1902564466,-0.2089867443,-0.0938785151,0.1192636639,0.1347805411,-0.0924397334,-0.098681584,-0.2281137407,-0.0275938753,0.0995097086,0.0621742457,0.030855177,-0.1956920177,-0.4680973291,-0.0687435865,-0.0431046523,-0.2906481922,-0.0174293369,-0.1362618357,0.3661518395,0.2362370491,0.0131259505,-0.0676069409,0.1398829967,-0.4628416896,-0.0739330798,-0.0567501634,0.2288001776,0.3603369892,0.0388829522,-0.3254885972,0.0196876433,-0.0969237015,0.073729232,0.1605698168,-0.3796150684,0.1734416336,0.0566922799,0.1947249323,0.2639690638,-0.2307887524,0.1366111785,0.0543897189,0.1980296671,-0.3040991127,-0.1324525177,0.348549664,0.1209521666,-0.1900663972,0.24346219,0.0969154164,-0.2263965607,0.010111222,-0.1134702414,0.1359642595,-0.2534936666,-0.0058573973,0.1649450809,-0.0938960835,-0.0936056897,0.1276295334,0.202522561,0.1171358228,0.2931626737,-0.2662751377,0.0618479773,0.0373290963,0.1988638788,0.0534297414,-0.3770519495,0.0391864888,0.3416778743,0.1264641136,0.0550488457,-0.1870340854,0.2549545467,-0.218196556,-0.1266997606,-0.2538814843,0.2500620186,-0.290099442,-0.1299736649,-0.4836012721,0.0175619591,-0.1633219868,-0.0217761416,-0.2347826511,-0.0343216769,0.2000267357,0.0977106094,0.0068534832,-0.5503773689,0.0679188818,0.3271187842,-0.2268446535,-0.206884414,0.2573148906,-0.220730707,0.1030768901,0.0693538859,0.3882035613,0.2815446556,0.5223617554,0.131636858,0.0157513879,-0.0084802983,-0.0625543594,-0.2425931543,0.1332502514,0.3340332508,0.5506145954,0.202540338,0.1368731111,-0.0447126329,0.3973864615,-0.0148728611,-0.1966416091,-0.235560745,0.4564155042,-0.041589722,-0.1657175124,-0.2481363118,0.2467006147,-0.3824774921,0.0845846161,0.3379653096,0.2099988163,0.2183213234,-0.1004958078,0.068880327,-0.2012833953,0.5259037018,0.1325386018,0.2430855781,-0.3363080919,-0.3151684403,-0.592972219,-0.030399682,-0.1867649555,-0.0836672857,0.2598384321,0.2768538296,0.1286007017,0.2379028946,-0.1078584492,0.0375849567,-0.1580642164,0.006950133,-0.3478974104,-0.067231372,0.3516533673,0.1494029313,-0.0589876994,-0.1482323706,0.0019771345,-0.0525819995,0.0660509989,-0.345415324,-0.0664856881,-0.0121003781,0.3986493051,-0.0508856215,0.1690225303,0.341196835,0.0692432076,-0.0755392238,-0.0546249785,-0.341304183,-0.2765317857,0.2060128152,-0.0200255141,0.3978819549,-0.3307775259,0.3640452623,-0.2454458475,0.279085964,-0.1147223487,-0.0889033452,-0.2284966707,0.2759385109,-0.2228842229,0.1294908822,0.0211059134,0.1572134942,0.1121838838,0.1212484315,-0.2977690101,-0.1270810068,0.4186914861,-0.4630413651,-0.2886116803,0.0755283684,0.1234209985,-0.1864893734,-0.1444676369,-0.3811195195,-0.3344034255,0.3654739857,-0.1666970998,-0.230718419,0.3508667946,-0.0595547222,0.0238494985,-0.0895109698,0.1427756846,0.1065756306,-0.4077624679,-0.1727150381,-0.5183374286]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"Also see: https:\/\/github.com\/huggingface\/nlp\/issues\/222","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":3,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nAlso see: https:\/\/github.com\/huggingface\/nlp\/issues\/222","embeddings":[-0.0126522174,-0.0246375147,-0.0722459927,0.380612731,0.1857978851,0.0949039161,0.3264930248,0.1880550683,0.2901073992,-0.1611640751,0.0399563126,0.4761996269,-0.1600545347,0.0138206724,0.3445964456,-0.4098393023,-0.0900351703,0.3455869555,-0.035433311,-0.0953039899,-0.1943152249,0.3255010247,-0.285798192,0.1441764534,0.0105972253,-0.2351358831,-0.1263173372,0.2266370058,0.0013775341,-0.4403240681,0.3849108517,-0.0674823225,0.2382890731,0.0213521253,-0.0001093491,-0.0507475063,0.3791428804,-0.1183919758,-0.375620842,-0.4032427967,-0.236971572,-0.4013931155,0.5054411888,-0.1960688829,-0.0281280465,-0.1962764412,0.2491616905,0.2580249906,0.3401263952,0.2542785406,0.2662699819,0.2117005587,-0.1316090822,-0.2593427896,-0.0140938303,0.07800401,0.1292455643,0.1018620506,0.0107793501,-0.1475755423,0.1373133361,0.0533914343,-0.1645040661,0.1440097988,0.1624133289,-0.1610451043,-0.1429128945,-0.0647350624,-0.1858589202,0.2578802705,0.4186974764,0.1232275963,-0.1750700176,-0.1929650903,0.158933565,-0.2773229182,0.1696619391,0.1651919335,-0.2426309586,-0.1099988744,-0.2409773022,-0.2766531408,-0.1167300195,0.459469825,0.2871930599,0.1930487156,-0.195667848,0.295809418,0.4224937558,-0.0574891716,-0.0726357549,-0.0789577886,-0.0242570974,0.1515191048,-0.4674142003,0.0636653975,-0.0116165737,0.2845612764,0.0485873893,0.1006273627,0.1501687914,-0.0687961429,-0.2441064417,0.1335614026,0.5120695233,0.1002753973,0.3289682269,0.0782482401,-0.1891340464,-0.1439639926,0.1319978982,0.0522485599,-0.2219604701,0.0198332854,-0.0871245265,-0.1494176686,0.1811903864,-0.1786748171,-0.1734318733,-0.2075805366,-0.2416976392,-0.2196392715,0.159626171,0.4279192686,0.0475107804,0.4032963514,0.2339086235,0.1131292954,-0.1348406672,-0.3891661167,-0.2057365775,0.1546826661,-0.1109441817,-0.0969506055,0.2789571583,-0.0521753207,0.5093408823,0.0190578215,-0.1758990288,0.0508189797,0.1158559993,-0.1145036966,-0.2091635019,0.0119395787,0.2974591851,0.1027802601,0.3155650198,-0.1145581976,-0.1201774403,0.0447872393,-0.2693708837,-0.0460889414,-0.0433883406,0.2115037441,0.1591280103,-0.1398023814,-0.4289461672,0.2984135151,0.1433789581,-0.3182417452,-0.277764529,-0.2426440567,-0.2064383328,-0.0971897915,-0.1277104765,0.1654132456,-0.0056359093,-0.1037401333,-0.0323035941,0.0176092573,0.0400378183,0.5338773727,-0.6055860519,0.3361652792,-0.1810061634,0.3795588017,0.7665454745,-0.0013225747,-0.0457546636,0.2584697902,0.057528045,-0.1569900811,-0.0431547277,0.3349553347,-0.0641293153,-0.0492418483,0.0991904661,0.3698593378,-0.0186707899,0.01062629,-0.0801815987,-0.0767880082,0.4376591444,0.1232091933,-0.2100781947,0.2248265594,-0.038888704,0.4488624334,0.2915731072,-0.1921007633,-0.2042241544,0.3885711133,-0.262725383,-0.074484624,-0.1075956523,0.0122502055,-0.5454056263,-0.1647608876,-0.0681501552,-0.0346816331,0.3996149004,0.0549670421,-0.2651139498,-0.0162282865,0.0568085052,0.1098521277,0.0689062476,0.312631309,0.0616249144,-0.0968547314,-0.2283500731,0.2779490948,-0.0979777947,-0.1044645384,-0.5283600688,0.1460050344,-0.0407336652,-0.0909034535,0.3838220537,0.0011240343,0.2719336748,-0.0470277555,0.0215525553,0.1502118558,-0.1892229468,-0.0071768668,-0.0311240852,-0.2471677214,0.095604755,-0.2150157392,-0.0546603985,0.235809207,0.3517908752,-0.1765045524,-0.1499076337,0.1531682462,0.0281830933,0.204100132,0.0520638004,0.1361578405,0.3612595201,0.0414571092,-0.0042563966,0.1743855774,0.4097516537,0.2529943883,0.2093637437,0.0233195573,-0.519566834,-0.1230003908,0.3658306599,0.1778496653,0.1830345839,0.0768179521,0.0809859708,-0.2122323215,0.2149214447,-0.2610073388,0.5002450943,0.2802160978,-0.1507855207,-0.1087419614,-0.2155293673,-0.2179338783,0.1632426679,-0.1237036586,0.2140616626,-0.2089090496,0.0262012854,-0.1710212082,-0.4249663949,-0.1535183489,0.1668901742,0.3078089952,-0.3366973996,0.1197474524,-0.2964127064,-0.5232054591,-0.0227638464,-0.2288310379,-0.3212434351,-0.269092977,-0.1904140115,0.2975833416,0.1025721878,0.1450109929,-0.0419217795,0.2860243917,0.1275295317,0.291767329,-0.1653503776,-0.0466390438,-0.3879055977,0.0597755983,0.1839917153,0.2234416306,0.247221157,-0.109595418,-0.2232345194,-0.0677759126,-0.0070579126,0.1042051911,-0.0910245776,0.1760556251,0.2969223559,0.2178561687,-0.2038486153,-0.1563934237,0.2528845072,-0.0306917764,-0.1142783016,0.1006298214,0.0217268858,-0.0922763199,-0.2423951924,-0.3469438553,-0.1714244336,-0.3131513298,0.2998683453,0.3234166801,0.2378020138,0.3593803346,-0.1040752083,0.2916773856,-0.1835149676,0.1297673285,-0.3554510474,-0.0204549786,0.0195219517,-0.0570576452,-0.4337068796,-0.0498819388,0.0708000287,0.1333042234,-0.0358149074,-0.2555756569,-0.4691180587,0.0012346962,-0.146984905,-0.1998371482,0.2012149096,0.2256180495,-0.2095003128,-0.0182220992,0.0020097031,-0.3617865443,0.4500448108,-0.0155357756,0.2453411222,0.0879180282,0.3249625862,-0.0284127686,0.2412416786,0.3112297654,-0.1280827522,0.4474761486,-0.1249618679,0.2412916422,0.0296919532,-0.3647510409,-0.0669807643,0.1254737079,0.1033921018,0.3512487411,-0.1423090994,-0.32720837,-0.0651881024,-0.0532175787,-0.1792300642,-0.0412328169,0.1034047157,-0.154436022,-0.0434268825,0.0380228013,-0.0172110628,-0.3584653437,-0.2389853448,-0.2526334226,0.0505923927,0.1230382472,-0.0008173272,-0.6070681214,-0.1093800068,-0.3653319776,0.2412635684,-0.0310822316,0.6489038467,-0.166791603,0.1110176221,0.1639056057,0.1393776536,0.6179256439,0.0323629379,-0.053441871,0.1514113247,0.079909116,-0.3985539079,0.1620472074,-0.2630522549,0.4878306985,0.2307167649,0.2546696961,-0.3237624168,-0.2486155629,0.142271921,0.0800535604,-0.2982649803,-0.2112228423,-0.1879907399,-0.3443235159,-0.262903899,-0.1448234022,-0.147154361,0.2366800159,0.2181999087,0.1835787445,-0.1930060983,-0.1544593871,0.2197824121,0.0037197687,0.1648908257,0.2278585732,-0.2259671837,-0.2211022675,0.2367426902,-0.3826127052,0.4511163831,-0.1492830515,-0.1224235073,0.0394655652,-0.0658047944,0.4321969151,0.1841388494,-0.0964459106,0.1321901232,-0.0777033418,-0.0333301276,-0.0086995428,0.0851036608,0.3092630804,-0.0875017866,-0.1672192812,-0.4607818723,0.361766398,-0.0100485533,-0.0528206266,0.2460148931,-0.381492734,-0.2988764644,0.20333606,-0.1815159321,0.6627342105,-0.1840265989,0.3714972734,0.4747862816,0.2930856943,0.7021633387,0.3320554197,0.1593614519,-0.379191041,-0.0049397335,0.0083533861,-0.185641855,-0.0204307325,0.423791945,-0.3176655173,0.3630723059,0.0611482076,-0.0908428878,0.1726378351,0.4783190191,0.0921053365,-0.2583820224,-0.9693938494,0.1839290857,-0.4200972319,0.0242757387,-0.1602538973,-0.2959502637,0.1801200509,-0.2945791781,-0.1072254404,0.4189432859,-0.0359007157,0.2328572124,0.272159189,-0.1355077773,0.1183291525,0.4173803926,0.0271792542,-0.0244457163,-0.1198212653,-0.0744181052,-0.3890822828,0.025106499,-0.2970232666,0.1155794412,0.289339155,-0.0707561821,-0.0280129183,0.1441044807,-0.049599845,-0.4226925373,-0.0780199692,0.1969454885,-0.1368101835,-0.2013379186,-0.6303848624,0.1822657734,0.0300831385,0.0541640446,0.1012781784,0.3042830527,-0.2918907702,-0.1060979217,0.390553683,-0.1097506285,0.1775039881,0.3524293303,0.1413332224,-0.0734750554,0.6722478867,0.2883760035,0.0770841315,-0.246379301,-0.1417151242,-0.0618339851,-0.2301283479,0.2618446052,0.1181230545,0.1906996965,-0.0595346987,0.3992159069,-0.0132006062,0.2187735885,0.2311671227,-0.4305692017,-0.0812872797,0.1453864872,-0.1666389257,0.0251773018,0.3127034009,-0.0402338095,0.2152516693,0.0977265909,-0.2982608676,-0.0185421612,-0.4786123931,0.21448946,0.2064589858,-0.2119238079,-0.0061726486,0.1489423066,0.1278576553,-0.0397920944,-0.1937703788,-0.2197831869,-0.0477354787,0.107740365,0.1254994422,-0.0601320043,-0.1213938892,-0.3484645784,-0.0172705725,-0.0198874082,-0.2118807137,-0.1081696227,-0.1218593493,0.3042336702,0.2101454437,0.023514675,-0.1255548596,0.1708803773,-0.4810462892,-0.0493083782,-0.0295034945,0.201386109,0.2814672887,0.0142548038,-0.3927134275,0.1037121564,0.1241472065,0.0353120752,0.1416443735,-0.3268879652,0.1662661731,-0.0164247695,0.189438656,0.2617892325,-0.262981087,0.3128752708,0.047004804,0.1648625433,-0.27547732,0.054499723,0.4443087578,0.0940574408,-0.1617704332,0.2317110747,0.0611235686,-0.1450628787,0.0549395792,-0.1911232322,0.2021757662,-0.2439127117,-0.0614327081,0.0284879468,-0.1582842469,-0.1281987578,0.0983517617,0.1494558007,0.1470436752,0.2050511539,-0.2928059101,0.1842297167,0.0433472134,0.1475574374,0.1674596816,-0.3911655843,0.0785673931,0.4035182297,0.1925048083,0.0712211728,-0.1961481273,0.3424704373,-0.3128987551,-0.1442000121,-0.2076455355,0.4404121637,-0.1975830495,-0.1835874021,-0.4669170082,0.0181865785,-0.1652553678,0.005655726,-0.2020735294,-0.1497284919,0.2632114291,0.0925386772,0.0790988281,-0.6583909988,-0.0401922129,0.3390706182,-0.2180782855,-0.1587317884,0.3838413358,-0.1239895001,0.0099214073,0.1958113015,0.3436056077,0.3114562035,0.4332994521,0.1998897195,-0.0493265763,-0.0394054465,-0.0537397973,-0.2290721238,0.1361859739,0.2383690923,0.4903230071,0.1675968319,0.1495812386,-0.0681288242,0.3037060797,-0.0309554227,-0.1561143696,-0.362424612,0.574354589,-0.1099871323,-0.1003412157,-0.3649092913,0.1372545362,-0.3015061021,0.0538001545,0.3656354845,0.2124220133,0.2850552499,-0.1062818244,0.075400658,-0.1193940341,0.4610932469,0.1619059294,0.2841157317,-0.1999046803,-0.0961971208,-0.5014604926,-0.0589028187,-0.057718955,-0.1119532734,0.1666118801,0.2070364654,0.1208840087,0.2173322588,-0.2549437881,0.1701681316,0.0129317706,-0.1251276731,-0.2487983108,-0.0791163966,0.1293335706,0.1990769655,-0.016303869,-0.1837031096,0.0187758226,0.0115782358,0.0296742208,-0.326379478,0.051893618,-0.0733188465,0.2707438469,-0.1850114614,0.2028604299,0.3654871881,0.0595532842,-0.0386117324,-0.069501847,-0.3588421941,-0.2364203483,0.1369562745,-0.1141235232,0.3749925196,-0.3501645029,0.4390982091,-0.2570733428,0.1105172187,-0.1760481745,0.0356906615,-0.2041308731,0.2011554986,-0.2668835223,0.1289761811,0.0775112435,0.1233419776,0.1472762674,0.0850060731,-0.2315612584,-0.0645575598,0.4690610468,-0.4914423227,-0.3233278394,0.087965779,0.1403710842,-0.1642584801,-0.2124602497,-0.5195890069,-0.2230149657,0.4404174685,-0.1605864763,-0.1194936261,0.3180796802,-0.1487314701,-0.0361532643,-0.0845101774,0.0672527403,0.1508588046,-0.3751016557,0.0157349724,-0.4035296142]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"I am getting this error when running this command\r\n```\r\nval = nlp.load_dataset('squad', split=\"validation\")\r\n```\r\n\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/root\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/dataset_info.json'\r\n\r\nCan anybody help?","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":27,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nI am getting this error when running this command\r\n```\r\nval = nlp.load_dataset('squad', split=\"validation\")\r\n```\r\n\r\nFileNotFoundError: [Errno 2] No such file or directory: '\/root\/.cache\/huggingface\/datasets\/squad\/plain_text\/1.0.0\/dataset_info.json'\r\n\r\nCan anybody help?","embeddings":[0.0945204496,-0.0742244646,-0.0324147865,0.4672505856,0.2305199057,0.2202830017,0.1371435225,0.2303495407,0.1524401158,-0.1321380436,0.0483731888,0.4689956903,-0.1310244799,-0.1885149926,0.2510724068,-0.2700096667,-0.1781660765,0.3005647063,0.2233232409,-0.0836042911,-0.0806101561,0.3751226962,-0.2587866485,0.1794959158,-0.0696576014,-0.1662869304,-0.1388200074,0.3408648074,-0.0691159666,-0.5248632431,0.3120110035,-0.2391539067,0.3402944803,0.1358475387,-0.0001175047,-0.0599622279,0.3421852589,-0.2002910823,-0.4428662658,-0.4260238707,-0.2834005356,-0.4043212235,0.4492724538,-0.2842895389,0.0768049881,-0.1121732667,0.3357010484,0.1443691105,0.5330747962,0.2761425376,0.1718373299,0.0707812831,-0.1612328887,-0.1983360797,-0.0910207927,0.1816204339,0.1172778234,-0.0071002003,0.0522170886,-0.0216119569,0.2575652599,0.1014679894,-0.2076477408,0.1450597048,0.2254332453,-0.0419415273,-0.1588707566,-0.2111566216,-0.0347859897,0.2176091373,0.459060967,-0.0522607416,-0.3170992434,-0.1796333343,0.2186481208,-0.1685426831,0.1927463114,0.3058081567,-0.2965562642,-0.0683341622,-0.1642871052,-0.3450191021,-0.1650012136,0.3975941539,0.2750882208,0.153581202,-0.2077714801,0.2252542078,0.3455044031,-0.0361194052,-0.1980067343,-0.1579736769,-0.0737132952,0.2626844943,-0.3977152109,-0.0694361851,-0.0915700495,0.1090952829,0.1139032021,0.0260958076,0.2392785549,-0.1550077498,-0.2616507709,0.2248243392,0.4656402171,0.1235243306,0.5095347166,0.1183806881,-0.1002926752,-0.1797017157,0.2383677065,-0.0380186066,-0.3207312226,-0.0792199075,-0.0185729247,-0.1307921857,0.2193070948,-0.3075301647,-0.2229974419,-0.1044669822,-0.277685076,-0.1764388531,0.1635672897,0.5018176436,0.0014942345,0.2334463447,0.1880491823,0.1749494523,0.0047678435,-0.347222209,-0.1687656045,0.2098423243,-0.0978954658,0.0369908288,0.3579716086,-0.0562426485,0.5068039894,0.0184472632,-0.2095424831,0.0261955187,0.1318286508,-0.118098937,-0.2231777161,0.0519675463,0.3733535707,0.1931856722,0.3314566612,-0.2338434309,-0.1246823817,0.1174580306,-0.4283104241,-0.103522025,-0.0473563373,0.1175819486,-0.0290024411,-0.1308630854,-0.4799637198,0.1306148916,0.1781004667,-0.4104081988,-0.2551843226,-0.1001109034,-0.1320746094,-0.159317404,-0.0487702191,0.3155113161,-0.1800805777,-0.2033411115,-0.0260751415,0.0535214245,-0.0411030874,0.5622394085,-0.5059040189,0.2713315785,-0.2215330005,0.3962130845,0.8447894454,-0.2791198194,-0.0745531693,0.3374209404,-0.0547371395,-0.1533975005,0.0275876578,0.3057096899,-0.0191108789,-0.0021677173,0.1813527495,0.3129216731,-0.0446244329,0.073014535,-0.094625555,-0.150456652,0.3942299187,0.1227044836,-0.2146468014,0.199151516,-0.0499236733,0.4528456032,0.2658727467,-0.1584924608,-0.1263805032,0.5303464532,-0.0729102418,-0.0173793975,-0.0843213871,0.0034207061,-0.6630390286,-0.0856323913,-0.079698585,-0.1029730737,0.0860074535,0.0132862804,-0.1351171136,-0.0978125557,-0.037882328,0.0486694463,-0.0134728076,0.2081670761,0.0850289389,-0.1480762511,-0.2450399697,0.2566251457,-0.2056079656,0.0003631883,-0.5732815266,0.2165286392,-0.0633266196,-0.1461006701,0.2288919985,-0.0382383727,0.2623355687,-0.0182976779,0.0105227828,0.1740931422,-0.0960973129,-0.0670991987,0.144589439,-0.1476926357,0.0615387186,-0.2101720721,-0.0106029259,0.2204058766,0.3082589805,-0.170662865,-0.3535010517,0.2086140811,-0.0569116659,0.3180491328,-0.0306315385,0.0880767256,0.2986536622,0.0288564973,-0.0738666952,0.1052178144,0.3209812343,0.2160905302,0.1885679662,0.0780155137,-0.3490473926,-0.198389262,0.3542863131,0.176728189,0.1097875759,0.1547648162,0.0201387741,-0.1638192981,0.1139688417,-0.1261243671,0.610334456,0.2790957689,-0.2174886912,-0.1357534379,-0.1227157116,-0.1258223504,0.1132868454,-0.1588297933,0.2757975757,0.004947497,0.0412526987,-0.0545984097,-0.317029804,-0.2727826238,0.1860586852,0.4085494578,-0.3574673235,0.0801150724,-0.2669263482,-0.4085353315,-0.0564550236,-0.1330900043,-0.3703535795,-0.2320026606,-0.1946538538,0.2208274752,0.3217781484,0.0679833293,-0.0536397845,0.3023664951,0.1571761072,0.010820169,-0.1574364752,-0.0606794804,-0.3003083169,-0.0399848297,0.1811008006,0.1816272438,0.1177246571,-0.1682049036,-0.189560473,-0.1653555036,-0.0380487554,0.0624029227,-0.0236980654,0.1431701332,0.3579214513,0.4521507919,-0.1662943363,-0.1952404678,0.2785131037,0.1151324362,-0.1622625589,0.0508082695,0.0326858833,-0.0698714033,-0.215606451,-0.3680701554,-0.1935552806,-0.2798214853,0.2910511792,0.309706986,0.2572421134,0.3628795743,0.0027240445,0.3270461857,-0.2055666298,0.2300758809,-0.3146944046,-0.146342501,0.1815907806,-0.0957210213,-0.3216098845,0.1054399759,0.1184149384,0.1621737182,-0.1154639795,-0.3580621779,-0.4874969721,0.0048743812,-0.0502634533,-0.3202095628,0.1484276205,0.3125919998,-0.274921149,0.0704339072,-0.0273816306,-0.480682075,0.3043287396,0.1231428012,0.4395596981,0.0070705772,0.465110451,0.0839045346,0.4180484712,0.2728458941,-0.1289700866,0.3143639565,-0.1395247728,0.2892531753,0.0977075323,-0.3613308072,-0.005515845,0.2341443896,0.0285627563,0.2181877196,0.0020773406,-0.1427938938,-0.0738524124,0.0572479255,-0.2445681989,-0.0016677544,0.1255098134,-0.0108063472,-0.1184798703,0.0733584464,0.0380210318,-0.4303795099,-0.2144970298,-0.3226793408,0.3040024936,0.1819106936,0.0907660201,-0.595271945,-0.2950332463,-0.2407021224,0.2234677672,-0.0160447434,0.5692745447,-0.1884070337,0.1972352713,0.1648418158,0.1474805474,0.6530005932,-0.0889932811,-0.0057106218,0.0284773596,0.1281398237,-0.4141221941,0.1168201193,-0.1514577717,0.6537325382,0.307310909,0.4086618423,-0.3931210041,-0.3012160063,0.0673375279,0.0989908352,-0.3241265416,-0.1676250249,-0.2476393282,-0.269257158,-0.2987031937,-0.1954523176,-0.1579847932,0.2615727186,0.3339886069,0.2174839526,-0.0739692599,-0.0597332492,0.2119123191,0.0225699879,0.2418562025,0.1694159955,-0.2763483822,-0.0864098221,0.271286428,-0.303406626,0.4241046309,-0.0572852753,-0.183334589,0.0778551325,-0.077429682,0.4648774564,0.2289249599,-0.0216726381,0.0373545066,-0.0037187748,-0.0163167957,0.0841489732,-0.0525827147,0.4038655758,-0.1450893879,-0.206680432,-0.4859923124,0.3761895001,0.1591276377,-0.0214494467,0.3054713309,-0.2700211108,-0.267129153,0.2274037153,-0.1958310157,0.8024116755,-0.0994682088,0.408102423,0.4356691837,0.2962632775,0.8019766212,0.1236784905,0.1407842934,-0.4002182484,-0.0436968692,-0.0313347057,-0.2122370005,0.0230466109,0.4677306116,-0.304666698,0.3824668229,0.0094721336,0.0154947778,0.1847152412,0.5063360333,0.0510775633,-0.3034004271,-0.9086315632,0.0838200673,-0.460206598,0.1457359642,-0.1305831671,-0.2854569852,0.0649591759,-0.2466886789,-0.1898802221,0.3769304454,-0.1219206527,0.1978493631,0.2616352141,-0.1367861778,0.1148827374,0.6384857297,0.0748051554,0.0469581597,-0.1343822181,-0.016911976,-0.3627629578,0.0178055707,-0.256657958,0.0931969136,0.2086672634,-0.0349645615,-0.0508150905,0.2315024883,-0.0677069575,-0.4613187015,-0.0208379943,0.1927846819,-0.2391491532,-0.1979088634,-0.7866969109,0.0266452879,-0.0650695562,0.0931719989,0.0293474328,0.3400616348,-0.400275439,-0.074763827,0.3850586712,-0.1993654072,0.2121276408,0.4944164753,-0.0325931981,-0.1524095684,0.6422560811,0.2634295821,0.0940716714,-0.2485849112,0.0126413889,-0.1036128327,-0.1308942139,0.1215966567,0.1789048761,0.157901153,-0.0204341393,0.147237137,0.0738912076,0.2489591092,0.1935808808,-0.5574479699,-0.1018656567,0.2492240667,-0.0811532289,0.0217872374,0.2138766944,-0.0820286348,0.1733286083,0.0159258917,-0.2284913212,0.1480927914,-0.3470257521,0.2211472243,0.1807365268,-0.1060174331,0.0744995549,0.0298022181,0.0348679423,-0.066113025,-0.1237300858,-0.1322527677,-0.0326611064,0.1824914068,0.1036966369,0.0389149413,-0.186268732,-0.3422485292,-0.0194664896,-0.0565465242,-0.1517760903,0.0580182523,-0.008630448,0.3131299913,0.1051790938,0.0541062169,-0.1583132595,0.1678390503,-0.3720066845,0.0594945699,0.0716723651,0.2414957136,0.1420675516,0.0604033843,-0.5040307045,0.036519248,-0.0073998244,0.0685614944,0.1255654097,-0.4524102807,0.1612499952,0.050101351,0.1488492936,0.3014197052,-0.298420012,0.3430105746,0.0157609228,0.0933736637,-0.1277236044,0.0259063374,0.5305549502,0.0562080406,-0.2049935162,0.3098514974,-0.1405711323,-0.1258148402,0.1287858337,-0.1223197952,0.4390665591,-0.3158116341,0.0028662262,0.0072843325,-0.1066027656,-0.1523946375,0.0978690907,0.1274053901,0.1369557828,0.1636426896,-0.3774887323,0.0929714069,0.0766735002,0.0822270215,0.1945706904,-0.3949976861,-0.161449194,0.4666148722,0.183962822,0.02096406,-0.2303614169,0.458430022,-0.3903162479,-0.1258833706,-0.2265261859,0.3225133717,-0.1883447617,-0.093386665,-0.5020495057,-0.0742425621,-0.0915476605,-0.1035261378,-0.1040730178,-0.2166135311,0.0948215425,0.1172204912,-0.0321893245,-0.5349878669,0.0956763476,0.4124471843,-0.1561525464,-0.2419692725,0.3715268373,-0.1406010389,-0.0175679326,0.2602107823,0.2587348819,0.3944669366,0.4811031818,0.1643831581,0.0550997853,-0.1380093694,-0.1544767618,-0.1036757082,0.1291246116,0.1398021728,0.2022048533,0.0663730428,0.1531262994,0.0224308036,0.2971623838,0.0227452498,-0.0627914965,-0.2193392515,0.4865588546,-0.1630722433,-0.1093456894,-0.3018390238,0.1257065237,-0.325258702,-0.0098167537,0.2018197179,0.3219429553,0.2425882816,-0.0543666705,0.0230505597,-0.1263326406,0.3122853935,0.2782086134,0.096190162,-0.271876663,0.0222724117,-0.6221786737,-0.0661117733,-0.0654824972,-0.2500227094,0.1446917206,0.3395778239,0.0061544371,0.2411964983,-0.1994547397,0.194353804,-0.0239312574,-0.0095548378,-0.3606649339,-0.1835725456,0.1819235086,0.3209441006,0.023058122,-0.2056955993,-0.0616287142,0.0127901603,-0.0675777495,-0.2296615541,0.018732395,-0.1072269231,0.3282791674,-0.1081203744,0.0464643911,0.4340124428,0.0673873499,-0.0733740628,-0.0713890716,-0.3200552762,-0.2393113077,0.2040486634,-0.0679474771,0.4824776351,-0.4668584764,0.3478718102,-0.3590073586,0.147755906,-0.2117471844,-0.0926740915,-0.2156452388,0.2352151871,-0.2525527775,0.3473977149,-0.0045701545,0.085600175,0.0493667424,0.1745089591,-0.2338332683,0.0256593172,0.5776774287,-0.4646982551,-0.2673557103,0.1539919227,0.2269408256,-0.0230029132,-0.2422227114,-0.3909176588,-0.1950158775,0.4414580166,-0.1283150017,-0.2740136981,0.1108235866,-0.1983668506,-0.0869564563,-0.0698463693,0.0026228814,0.1056982577,-0.3569948077,-0.0054749544,-0.4018365145]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"It seems like your download was corrupted :-\/ Can you run the following command: \r\n\r\n```\r\nrm -r \/root\/.cache\/huggingface\/datasets\r\n```\r\n\r\nto delete the cache completely and rerun the download? ","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":28,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nIt seems like your download was corrupted :-\/ Can you run the following command: \r\n\r\n```\r\nrm -r \/root\/.cache\/huggingface\/datasets\r\n```\r\n\r\nto delete the cache completely and rerun the download? ","embeddings":[-0.1208358333,-0.03516794,-0.0817960054,0.5541924834,0.302541554,0.1315348595,0.0039048679,0.1393863559,0.2804363668,-0.1001443416,-0.1951356232,0.3699634373,-0.1204647198,0.058258269,0.3326439559,-0.4019226432,-0.1763337702,0.2947899103,-0.1826591343,-0.0056841513,-0.0904862583,0.433855027,-0.1845223308,0.1346078217,0.0080535682,-0.187505886,-0.1384477317,0.4052764773,-0.2076117843,-0.49975124,0.3896077275,-0.0273205675,0.2414442897,0.1131715178,-0.0001096911,-0.0584160723,0.4368598461,-0.0966122523,-0.3577699065,-0.2647203505,-0.192922473,-0.350591898,0.3675930798,-0.2010151744,-0.0211331937,-0.0549998209,0.1864340752,0.1668653637,0.4770371318,0.2255970538,0.2994112968,0.3363180757,0.0674541742,-0.2147987038,0.0665507019,0.0725695267,0.0687307045,0.1771294177,0.0832876563,-0.0365010425,0.2348442078,0.1266146451,-0.0852469131,0.108896099,0.2413454205,-0.1036856472,-0.144518733,-0.0758871436,-0.083059907,0.183268562,0.4341345727,0.0184461698,-0.1963406056,-0.181243062,0.158436954,-0.3754579723,0.2296156585,0.1866718382,-0.1867710054,0.0540544204,-0.2525212467,-0.2835464478,-0.0359016508,0.333701849,0.1003297344,0.1187488586,-0.1964879036,0.1338862032,0.3819933534,-0.0455963202,-0.2346821278,-0.047804378,-0.1841405034,0.1175477728,-0.3242128193,0.127813831,-0.0064956178,0.3758376539,0.1633100808,0.1040660217,0.1301358342,0.0321311951,-0.2173241973,0.0430441052,0.4068357646,0.1638967842,0.2254475653,0.0350734815,-0.0719189197,0.0226660836,0.2335491329,0.0246645808,-0.1768237203,-0.0131368367,-0.1947178096,-0.1669892073,0.1918569654,-0.3178799748,-0.2259164155,-0.1674319208,-0.1062630415,-0.1299868673,0.1750203818,0.4389184713,-0.1331017017,0.2702865899,0.1546382159,0.2108220011,-0.0719088763,-0.3060039878,-0.2306702286,-0.022825975,-0.2534001172,-0.0309476517,0.3860832155,-0.1861059219,0.5473421216,0.0135182906,-0.1677367985,0.0134690646,-0.0357146971,0.0495146737,-0.3188311756,0.1386560649,0.2437673658,0.1978790015,0.2121309936,-0.1395173967,-0.0113407131,0.0430387855,-0.2953130007,-0.1464065462,-0.0801675171,0.2135620564,-0.1552684158,-0.1995833814,-0.4402094185,0.0507312864,0.1212089658,-0.2245112509,-0.2888998687,-0.1243139431,-0.1727029383,-0.1074631363,0.0130304741,0.3030004799,-0.1179322079,-0.2333218455,-0.0909878165,-0.1433808953,0.0395836607,0.5990224481,-0.4720256031,0.20328179,-0.2796126902,0.3399639726,0.5836390257,-0.0662525818,-0.3358995914,0.227344349,-0.0008168123,-0.01343079,-0.0993626267,0.1922669858,0.0965110585,0.0118185421,0.1164100394,0.3620453775,0.0287526902,0.1049309522,-0.1732213944,-0.1945458204,0.3117160201,0.0709671602,-0.093405135,0.3505002856,-0.0114897657,0.3573697209,0.4544574618,-0.087701261,-0.1222721264,0.5004968643,-0.1453568935,0.0492671989,-0.1668965369,-0.0882610455,-0.5441409945,0.0593109317,-0.1901802868,0.0315848924,0.0135777965,0.0000582695,-0.2903513312,-0.0430091172,-0.0103999386,-0.0132164368,0.1037643775,0.3401206732,0.0623906478,-0.1085375249,-0.1861585528,0.3082202673,-0.0766590238,0.0399836786,-0.5488080382,0.2430409193,-0.146076411,-0.184444204,0.3218106329,-0.0020530643,0.2478424311,-0.1411488652,-0.0050197276,0.2052924186,-0.2015739083,0.0224538557,0.0889196545,-0.2364064008,0.1708909124,-0.303313911,-0.0202919561,0.0921814814,0.2264540344,-0.1342033297,-0.1724210531,0.1043648496,-0.0441480801,0.1301497817,0.0137395272,0.1632827073,0.4534036517,-0.0649586022,-0.0343797356,0.0723703653,0.4089065492,0.1714710444,0.1922938973,0.0571836159,-0.4705581665,-0.0130068632,0.3720381856,0.0831544399,0.0520668589,0.1183371395,-0.0280825831,-0.1841575354,0.1566457301,-0.0922177061,0.4385109544,0.2807067037,0.021200642,0.0197020657,-0.1619890928,-0.2165618688,0.0882186219,-0.1398135424,0.1459080428,0.0268743746,0.0015689586,-0.1073058695,-0.5094411969,-0.0787741169,0.1528925449,0.4604257345,-0.344897747,0.0262953509,-0.4339517653,-0.5732690692,0.0247580074,-0.2075873166,-0.3741030991,-0.3285754025,-0.2151392549,0.3307510316,0.1884204149,0.2206446826,-0.0567403026,0.2805197239,0.0792720169,0.1388975084,-0.2551280558,-0.0306231957,-0.3055675328,0.0591915175,0.2956994176,0.1850128323,0.2659867406,-0.2459721565,-0.1127665192,-0.2350987643,0.0023600659,0.0553314984,-0.0132859424,0.2023050934,0.2612726092,0.3959534168,-0.1890043318,-0.1414331198,0.3529102206,-0.0753984004,-0.1788623333,0.0558834635,0.0728590265,-0.1878098398,-0.3019116521,-0.2071998566,-0.1801958829,-0.3855598569,0.2631172836,0.3475292027,0.15318349,0.4041582644,-0.1057731956,0.2807362676,-0.1813134402,0.1571633518,-0.3872302771,-0.178345561,0.1482327878,-0.1565333307,-0.482779175,0.0307455119,0.1828864962,0.2376771271,0.0537138619,-0.4516231418,-0.2450682372,-0.0631421953,-0.0250612721,-0.155939579,0.1497789621,0.2201609015,-0.2043899149,0.0002099902,-0.0150725851,-0.4173906147,0.2231901586,-0.0412197709,0.3961802423,-0.0561774448,0.3297177553,0.0454111546,0.3947474957,0.2492125481,-0.0296319164,0.4718222022,-0.0854204744,0.4036557376,0.0172223747,-0.3875497878,-0.0199890807,0.0398939177,0.1378875375,0.3058994114,-0.0331907868,-0.2092963308,-0.1294094771,0.1402889788,-0.2421462983,-0.0437906459,0.0807234719,-0.037313655,-0.0279166084,0.1074752435,0.0142168337,-0.2243624181,-0.2369502336,-0.0941373035,0.1967789829,0.0384115502,0.1247633696,-0.4978688657,-0.0171323139,-0.4774362445,0.2559385002,-0.0917933285,0.5045907497,-0.1847157031,0.1850707531,0.196822077,-0.0239254683,0.7373422384,0.0003915665,0.0439621136,0.1132054925,0.1021771878,-0.3814080358,0.0360465907,-0.2373385727,0.3178756237,0.2889144719,0.3312782943,-0.4328977466,-0.2238558531,0.0705868453,0.1767569482,-0.3416845798,-0.1191545799,-0.2778836191,-0.4018354416,-0.2972097099,-0.1775015146,-0.1243700087,0.2542223334,0.1983023137,0.1885098219,-0.2070012987,-0.0196100436,0.0393012054,0.0126312207,0.3488481045,0.2138466537,-0.1760130227,-0.02989191,0.3170618713,-0.3836450577,0.6565624475,-0.2302845269,-0.0801241994,-0.0019898429,-0.0588679016,0.1196519136,0.283590138,0.0258420464,0.0661927536,0.1482639462,0.0225745086,0.0024656362,0.1120658591,0.4678326547,-0.086416617,-0.236718148,-0.4672760665,0.3942334354,0.062698096,-0.0873977914,0.2672941983,-0.0925363898,-0.2375618964,0.0946522951,-0.2230311781,0.7779065371,-0.2666202188,0.2676980197,0.465626061,0.1332101673,0.7914779782,0.3020038605,0.055827301,-0.3450185955,-0.0534346364,0.0034793576,-0.0830922425,-0.0145146539,0.2470213473,-0.2202867866,0.3266405165,0.0761027187,-0.0174217615,0.1612612009,0.4570517838,0.0335229561,-0.2320164591,-0.825848937,0.2221596092,-0.4772251844,0.2146332115,-0.2210186571,-0.2377548665,0.1717681587,-0.2165990025,-0.2919359505,0.330537796,-0.0632522553,0.1522871554,0.1296484619,-0.1401220262,-0.0292421952,0.3727943897,0.1012073979,0.0817443877,-0.2231071442,0.0493483692,-0.4282253981,-0.1093116552,-0.1711350977,0.1482419521,0.1803590208,-0.1669814885,-0.0889637917,0.063490063,-0.0151438238,-0.4146514237,0.0380000249,0.0872356072,-0.2390982956,-0.0853257999,-0.5879573226,0.1076952741,0.1023370922,-0.1085282564,0.1018533111,0.3348296583,-0.2579452097,-0.0709144846,0.3010928035,-0.1090114415,0.0297327265,0.5251571536,-0.0630115047,-0.2667009532,0.4769779444,0.3307652771,-0.0190023631,-0.221119225,-0.2141519338,-0.2117604166,-0.2999354005,0.1025577039,0.019615056,0.128610298,-0.0449616946,0.3471592367,0.196378395,0.0761312321,0.1904612482,-0.5562463999,-0.2069181651,0.1641508341,-0.0187878553,0.1562838405,0.2512000203,-0.0482756644,0.2054535896,0.0045395819,-0.3312322795,0.1287466884,-0.3835963607,0.1300290525,0.231796056,-0.1469204426,0.0800262764,-0.1164825484,0.1197986379,0.0081298798,-0.1924992055,-0.2152577341,-0.1304066628,0.0840822384,-0.0064272401,-0.1149520278,-0.0763693228,-0.2749094367,0.1007032841,-0.0419499017,-0.0774721205,0.0395103768,0.020404648,0.1975651979,0.1334065646,-0.0004852299,-0.1075720787,0.1885932386,-0.2641467154,-0.0612763204,0.0023727312,0.2145817727,0.1448704153,-0.070588164,-0.2833717465,0.0135017578,0.0492412001,0.1117857844,0.2707844079,-0.4229501486,0.1147240475,0.1553975642,0.3200045228,0.3243114054,-0.1496700495,0.2079665214,0.1099298373,0.2093261033,-0.2810993493,-0.0381194837,0.3883262575,0.1568452269,-0.1864844412,0.1853397489,0.0217627939,-0.1800826788,0.0756686553,-0.060421329,0.4124303162,-0.1338834167,0.0184364989,0.0799999461,-0.0961672142,-0.1573920697,0.0937553495,0.1930660605,0.0726382956,0.3452230692,-0.3517149091,0.1164216772,-0.1140093133,0.1874355674,0.061584428,-0.4589631557,-0.0249234103,0.41391325,0.095237501,0.1886266619,-0.1540615559,0.481189996,-0.3659304082,-0.3362007141,-0.2584637403,0.4400587976,-0.2293966264,-0.1707068831,-0.4461744726,-0.0980627239,-0.1007109731,0.0837694481,-0.1396288276,-0.1747603416,0.2323826551,0.0708606243,-0.0887764767,-0.7703475356,-0.1341928393,0.2694258094,-0.1755435616,-0.1846483499,0.3699839413,0.0142251719,-0.0626866519,0.1248962283,0.3632325828,0.3218103647,0.2891969383,0.2349758595,0.0119543839,-0.0369728357,-0.0304125845,-0.2882830501,0.2068495452,0.2686154246,0.2998308837,0.1915788352,0.2024134845,-0.1214675456,0.2347262353,-0.097241275,0.0367967747,-0.1842853427,0.521306932,-0.221200794,-0.0775651783,-0.2360417992,0.0832680166,-0.4021274745,0.1651912928,0.2254088521,0.1096487045,0.2508280277,-0.1404033005,0.0821206495,-0.0586435683,0.5475358367,0.4613279998,0.3466024101,-0.1919843704,-0.1506362557,-0.8086357713,0.0455873385,-0.0774512962,-0.136884585,0.1033005044,0.1629898548,0.0158608407,0.250456512,-0.123708047,0.2035199255,0.0803961679,-0.1093621552,-0.3904360831,-0.0249690264,0.1333875954,0.1654972434,-0.0090693161,-0.2018320709,0.1789035499,-0.0910183117,0.0404824913,-0.1891566515,0.0775694177,-0.193812862,0.10293594,-0.0825136751,0.2240863889,0.4087136984,0.0647912994,-0.1705442965,-0.2324148417,-0.4169975519,-0.159077093,0.097555615,-0.0611262694,0.4269742966,-0.2808963656,0.3999004364,-0.3593806922,0.2727791071,-0.0009129657,0.1102970392,-0.3232365847,0.1900504678,-0.2084136903,0.0575464293,0.1520988643,0.2662848234,0.1349390596,0.1428112388,-0.2719806135,-0.0622669272,0.6023275256,-0.3892844915,-0.2362944782,0.1108532026,0.1529450715,0.0849868804,-0.3550404906,-0.5108585358,-0.0798035786,0.2991117239,-0.2087141126,-0.2410825193,0.2545332611,-0.2331948876,-0.0458343588,-0.0497135371,0.2441420853,0.2074665427,-0.4690235853,0.1218286753,-0.2944378853]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/157","title":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"","comments":"I tried the notebook again today and it worked without barfing. \ud83d\udc4c ","body":"I'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a","comment_length":12,"text":"nlp.load_dataset() gives \"TypeError: list_() takes exactly one argument (2 given)\"\nI'm trying to load datasets from nlp but there seems to have error saying \r\n\"TypeError: list_() takes exactly one argument (2 given)\"\r\n\r\ngist can be found here\r\nhttps:\/\/gist.github.com\/saahiluppal\/c4b878f330b10b9ab9762bc0776c0a6a\nI tried the notebook again today and it worked without barfing. \ud83d\udc4c ","embeddings":[-0.0978619978,-0.0089837192,-0.0532620177,0.4294988811,0.1991301775,0.0352936201,0.2032472938,0.1589664966,0.1813582778,-0.1757220477,-0.0188250244,0.5022028685,-0.1602034718,0.1090852171,0.39484936,-0.2542797923,-0.1012669727,0.3150129318,-0.0448288284,-0.1017404795,-0.2462023497,0.2946237326,-0.3840457201,0.1119248345,0.1356037557,-0.0901679248,-0.1356754601,0.0763962045,-0.1005969048,-0.4047166705,0.3889181614,-0.2122971714,0.3395319879,0.1157031506,-0.0001146432,-0.0390841104,0.3371038735,-0.1087208837,-0.2853006721,-0.2994198203,-0.1884863526,-0.3863472641,0.4669420421,-0.2349059582,-0.128158614,-0.2595922947,0.2763047814,0.1965456605,0.3216397762,0.3153101504,0.195456773,0.2817656398,-0.0390088744,-0.183592096,0.1520634443,-0.1054124311,-0.0238059442,0.1257887781,0.2429684401,-0.25004673,0.1999377459,0.1079043522,-0.2201446146,0.1891566813,0.2595740557,-0.0613289066,-0.1449576914,0.0273404103,-0.1197630093,0.1792055815,0.5293779969,0.0147023862,-0.1124156266,-0.0741047263,0.1544126868,-0.1590465903,0.1588237733,0.1769175678,-0.2853136659,0.0613020845,-0.2173841,-0.266271174,-0.1087110564,0.459132731,0.238284111,0.3390488327,-0.1690659821,0.2597351074,0.3135816455,0.0514371246,0.2095700353,-0.0877829045,-0.051280681,0.1309477985,-0.3934925199,0.1425008178,0.0431219339,0.3004518747,0.0013978166,0.1071383879,0.4295507669,-0.0649253801,0.0068699857,0.0498249158,0.3018165529,0.123710677,0.3777247965,0.0664694011,-0.1263407022,-0.0863259435,0.1085587293,-0.0262266397,-0.1506413519,-0.117975831,-0.0337450542,-0.1927677095,0.2790391743,-0.1832551658,-0.1459616721,-0.0696067438,-0.2710930109,-0.0945197344,0.204409644,0.4471167922,-0.0019583092,0.1441871971,0.3796589375,0.092295602,-0.0600716844,-0.4339272976,-0.1541426927,0.0048640412,-0.2387216836,-0.0052834158,0.3184240162,0.3021641076,0.3497276902,0.0543026738,-0.2918973565,0.1491129994,0.1097396687,-0.1224058717,-0.1366995573,0.1042790785,0.3602097034,0.077242516,0.3055233359,-0.1340763271,-0.0467659086,0.1821885705,0.066529423,-0.0329272449,0.0181008317,0.1410869509,0.2045599371,-0.1297108233,-0.4601496756,0.1573881805,0.2682464421,-0.1812391132,-0.2610780597,-0.2719019353,-0.1646433324,-0.1509912908,-0.1370006502,0.0690039098,-0.1631943733,-0.0459189303,0.0016565803,0.0496716686,0.1104422733,0.4201090336,-0.5359603167,0.3988125622,-0.0942384675,0.4001959264,0.6274641752,-0.0559797697,-0.2904140055,0.3295037746,0.1168447584,-0.0881893635,0.1138694957,0.3696369529,-0.0377262719,0.1590976268,0.2487129122,0.430698216,-0.0396516807,-0.072639592,-0.2433909625,0.0012067874,0.266325891,0.1786847115,0.0468168929,0.2002728134,0.0129981348,0.4121426344,0.4917290211,-0.2527033985,-0.1457127631,0.4598867893,-0.0050027231,-0.0350397639,-0.0823846534,0.0729731843,-0.4390936792,-0.1072219685,-0.1135967523,0.0217016973,0.2566843927,0.001586751,-0.1125757694,-0.0219517928,0.0378663428,0.1104240716,0.0026378667,0.2085205615,0.0198194124,-0.1444324106,-0.1343076527,0.4303342998,-0.341661334,-0.0209880695,-0.5724426508,0.1712132096,0.0071348874,-0.2511111498,0.413018167,-0.1617671996,0.2454603612,-0.0032754478,0.0362396091,0.101827167,-0.1006248966,-0.2796386182,-0.1620208174,-0.2998875976,-0.003588523,-0.2099698335,-0.0566281453,0.2792499661,0.3338078856,-0.1727267653,-0.1193921939,0.1653707325,-0.0019619958,0.1450850517,0.0682739243,0.1529564261,0.3096627593,-0.0586588271,-0.0226122681,0.1538106948,0.5671073198,0.177664414,0.1956718117,-0.0445046015,-0.5615799427,-0.1312890798,0.2210336179,0.218929857,0.222558111,0.0333628915,0.0582254305,-0.1018479913,0.2786838412,-0.0915045291,0.591077745,0.2655672729,-0.2054132521,-0.0290865041,-0.2516041994,-0.249490872,0.1562234163,-0.2788498402,0.1994558722,-0.1157755703,-0.0280742291,-0.2013272792,-0.3627894223,-0.1022069082,0.2165239304,0.3375803828,-0.3848187625,0.0063304566,-0.2545509934,-0.5454970002,-0.0032484301,-0.0270620026,-0.1632596105,-0.2869267762,-0.2043155432,0.4848935008,0.2474765778,0.0918483362,0.1453402489,0.3846697211,0.0186529271,0.2804993689,-0.1147565097,-0.0394028872,-0.4135298431,0.0078767966,0.2781000137,0.228095904,0.1280113459,-0.2747892141,-0.1785196513,-0.2819597721,0.002456882,-0.0129542891,-0.0785581172,0.3120702803,0.295352608,0.2263077646,-0.3449520171,-0.0854619592,0.2050204724,0.0375055932,-0.1109408885,0.1764075011,0.0336489864,-0.3005732298,-0.1557135135,-0.1912007034,-0.1766583025,-0.2805616558,0.3866593242,0.2394661903,0.115062356,0.129916057,-0.1109364182,0.2773230374,-0.0492957607,0.0846207142,-0.2481299043,0.0943707824,0.1134093851,-0.0867193118,-0.4741086066,-0.0662310869,0.1969617009,0.2349889725,0.0057486934,-0.503236413,-0.3343475759,-0.1053868979,-0.1276756376,-0.1661843657,0.139932707,0.1196789369,-0.2126663029,0.0081580067,-0.0293085687,-0.2314333022,0.2975244224,-0.0972963944,0.4022951722,0.1110139117,0.2579305768,0.1180591583,0.1812429279,0.2061791271,-0.0844239667,0.4403686523,-0.0699269399,0.1545097232,0.0226734951,-0.4391704798,-0.2091757357,-0.0211713705,0.0185799841,0.2509815097,-0.1741808206,-0.2976959646,-0.0929051489,-0.0459469855,-0.134152621,-0.1165451109,0.2195416689,-0.1186275408,0.0290138051,0.1462859511,0.0260526128,-0.2790657878,-0.2745437026,-0.3538789153,0.0806777999,0.0449392386,-0.0249784309,-0.5919908881,-0.1338556856,-0.4912366867,0.3891040981,-0.2014274299,0.7287322283,-0.2141005397,0.0865165293,0.1594589353,0.1173432469,0.6006242037,-0.1070461273,-0.0436090752,0.2674657404,0.0612583607,-0.5946859717,0.0424160734,-0.3509787917,0.5119432211,0.2798431218,0.2605009079,-0.2502328455,-0.0876714811,0.1498061121,0.0665134639,-0.3474057019,-0.1096751541,-0.3124244213,-0.4429957271,-0.1401431859,-0.1461155564,-0.0970234498,0.244403854,0.2890482545,0.1901312917,-0.3126401901,-0.1623568237,0.1908832341,0.0253109522,0.1838146001,0.1223661974,-0.0965070203,-0.5125684738,0.2716950476,-0.1252891421,0.3193442822,-0.1984947622,-0.2647243738,0.108805351,-0.3016397655,0.3653151691,0.2799583375,-0.1561599374,0.1649865508,-0.0526542589,0.0166685488,0.238557905,0.0118312463,0.2850512266,0.0214608274,-0.1987568438,-0.3836485744,0.3372885287,0.0963810757,-0.1202161312,0.3328419626,-0.4054137766,-0.4113137722,0.3338047564,-0.0322231203,0.8277196288,-0.3161780536,0.362308234,0.4090558589,0.1549281478,0.465034157,0.2053243518,0.2387046814,-0.3766338229,-0.1457997561,0.0340666734,-0.2132339627,0.1208868623,0.3613424897,-0.3645220399,0.3114232719,0.0061561652,-0.1468157768,0.2733067572,0.4449298084,0.0149683291,-0.0932615399,-0.8966153264,0.1126933619,-0.366856575,-0.0067616832,-0.1915483624,-0.2040933669,0.0536539629,-0.2616089582,-0.2631019652,0.2550935745,-0.175802201,0.3015985191,0.2011833787,-0.2187446952,-0.0420658179,0.5920344591,0.2008507848,0.0958814472,-0.1344189048,-0.0183134023,-0.1942329407,-0.0579062514,-0.2895255089,0.0170125049,0.2692034245,-0.1104205549,-0.1191323623,0.148215726,0.045631431,-0.6622515917,-0.0002029672,0.2162501216,-0.0657070875,-0.2417714,-0.5623164177,0.2371325791,-0.0559547432,-0.0459660031,0.0719102323,0.1340413839,-0.3150815964,0.1285991073,0.2757146657,-0.2715463936,0.1436865479,0.4671690464,0.1631237119,0.0298801996,0.5125285387,0.4460131526,0.103065744,-0.2575864196,0.0139159309,0.0881260857,-0.2019198239,0.1976523101,0.0063134367,0.2714785933,-0.026469633,0.4838833809,0.2002754062,0.2109103054,0.1447580904,-0.4210511744,-0.0123798167,0.2433562428,0.0651828125,-0.0172807202,0.3174306452,0.0431148224,0.2636613846,-0.0220599938,-0.2574436069,0.0649960414,-0.5414392948,0.1924872696,0.2275379598,-0.10687197,-0.086033538,0.0582296103,0.0954226777,0.0190168507,-0.0599927939,-0.1900402755,-0.000413234,0.1352550238,0.0496475771,0.0961034149,-0.3026443124,-0.4682095051,0.0097818505,-0.0058011641,-0.2426864207,0.0854097679,-0.1084098071,0.1779192388,0.2018107921,0.0185737871,-0.2067110837,0.1641792804,-0.620678246,-0.0417487696,-0.0101152183,0.2454765141,0.4074579179,0.090598233,-0.286361903,-0.0452394709,-0.083501786,0.027735563,0.1683492512,-0.3948402405,0.168196559,-0.0858857259,0.1370601654,0.308799684,-0.3008833826,0.1227114573,0.0048965886,0.1178456023,-0.2751494646,-0.1487390995,0.2606804073,0.1469627172,-0.2551198602,0.3159499168,0.0754181594,-0.2191940546,0.0265093893,-0.1230942458,0.1597686559,0.0132663641,-0.1031458601,0.0598339066,-0.1134908646,-0.0755908117,0.099434495,0.2103749365,0.0462175719,0.2308793515,-0.3761764169,0.1713639051,-0.1463924795,0.2666099966,-0.0251709018,-0.4329718947,-0.0536174662,0.3739406168,0.0026755345,0.1121626273,-0.2378611118,0.2476841062,-0.2264543027,-0.2675724626,-0.2559577525,0.3068457246,-0.227009356,-0.1147747189,-0.5423227549,0.0279445052,-0.1563853472,0.070781894,-0.1486895084,-0.0256317388,0.2050488889,0.0688232556,0.0193660595,-0.5419484377,0.0295504816,0.3440537751,-0.2000709325,-0.2498634607,0.2643384635,-0.3195311129,0.1199509948,0.0576709285,0.3666535616,0.2471931726,0.5309934616,0.1126041263,0.0581745021,0.0265466738,-0.0684190914,-0.2935711741,0.1557208449,0.3904161453,0.5884839296,0.1910630316,0.1177950576,0.0121752284,0.4502870142,-0.0912760943,-0.0802513435,-0.2324325144,0.4795342386,-0.0951204896,-0.1855834872,-0.2508372664,0.2504933178,-0.4046950042,0.1532838792,0.4174050689,0.15550448,0.3054456711,-0.0938141271,0.0222526975,-0.118665576,0.5114337206,0.2192262411,0.3019443154,-0.3648154736,-0.3583559692,-0.5298042297,-0.0835923627,-0.2474442869,-0.1439376175,0.1287345886,0.2622145116,0.1348589808,0.2003440112,-0.1104463264,0.1027742326,-0.3175977767,0.0581828691,-0.3000189066,-0.1037764922,0.2996056676,0.2267735451,-0.1005984396,-0.2516804338,0.1046309993,-0.1253588796,0.0364931785,-0.3264747858,-0.0773985535,0.1183808967,0.4085708857,0.049801264,0.1397907585,0.3639867604,0.1001831517,-0.1035083011,-0.0524041466,-0.3355380595,-0.1574976295,0.2597566545,-0.0452014916,0.2497619987,-0.3794930577,0.5019409657,-0.2306190282,0.2582268119,-0.0217779092,0.0427951925,-0.1630870551,0.1985079497,-0.2387802899,0.1250852644,0.0484081767,0.1869886816,0.1173457429,0.0935017169,-0.3586775959,-0.1005586311,0.4519633651,-0.4143419564,-0.2553896904,0.1384246498,0.1509450823,-0.2147861272,-0.1069563925,-0.5959550738,-0.4257127047,0.3575249016,-0.1929703951,-0.2994723916,0.364148736,0.0571422614,-0.0644742772,-0.1432805955,-0.0904047787,0.089421533,-0.4399746656,-0.1626235694,-0.4628075957]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/156","title":"SyntaxError with WMT datasets","comments":"Jeez - don't know what happened there :D Should be fixed now! \r\n\r\nThanks a lot for reporting this @tomhosking !","body":"The following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook","comment_length":20,"text":"SyntaxError with WMT datasets\nThe following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook\nJeez - don't know what happened there :D Should be fixed now! \r\n\r\nThanks a lot for reporting this @tomhosking !","embeddings":[-0.32973665,0.0324432179,-0.0193531662,0.0127459355,0.2164737433,-0.0149221374,0.3609470129,0.5264438987,0.1301548183,-0.1121621057,0.0932864323,0.4350022972,-0.3767533898,0.1414317936,0.1508708149,-0.1754975915,-0.0268027131,0.1413965672,-0.3586361706,0.151956737,-0.2946045697,0.1583642811,-0.420961529,0.2173392326,-0.2542883456,-0.0902357399,-0.021146521,0.144950211,-0.1394597441,-0.462849319,0.140802905,-0.0065226816,0.2389953732,0.3692950606,-0.0001028357,-0.0074911187,0.1617367566,-0.0390187725,-0.3923019171,-0.3421501517,-0.4517919123,-0.1481353492,0.143838644,-0.2942019701,0.0407955088,-0.1019728407,0.012090249,-0.2335641831,0.5227859616,0.470978111,0.2622492909,0.4046530128,-0.0788426101,-0.1995568573,0.0400845148,0.1868067235,-0.0251018368,0.2721501887,-0.0462756231,-0.3078270257,0.0443689711,0.2844180167,-0.0971550569,0.2046322972,0.2288134843,0.0462167524,0.1737399995,0.0146226026,-0.1336928159,0.111689046,0.3766534328,-0.3340502977,-0.1922972947,-0.0254198257,-0.0237156339,-0.3796640635,0.1737448573,0.2510075867,-0.2430952042,0.2042771727,-0.2803643048,0.0721920952,-0.0495335385,0.280293107,-0.0419157632,0.7289190888,-0.1515552253,0.0830518156,-0.0326680392,-0.1897672564,0.1182416454,-0.0233938098,-0.0012605379,0.1437297612,-0.0696197599,-0.0688927695,0.2810450196,-0.0273937248,-0.2340159565,0.0267737973,0.1816953868,-0.1459489912,0.1988346726,0.1966277063,0.283844322,-0.064487271,0.0399041548,-0.0328306854,0.2843933702,0.2534207404,-0.027510548,0.0475533232,-0.1733619273,-0.4689917266,-0.0253391545,0.0136443032,0.5950089097,0.1803969294,-0.2701009512,-0.0289219059,-0.2787000239,-0.302798748,-0.0676975995,0.2093273848,0.121819444,-0.1600497216,0.3483577371,0.0752085149,-0.2440336943,-0.1647191793,-0.2930479646,0.340895921,-0.3188335896,-0.1684278697,0.2188079953,0.2253614962,0.0922131017,0.0471641496,-0.0809147358,-0.1988425553,0.1389292777,-0.0077622063,0.1081167087,0.1051954702,-0.1785956919,0.093339853,0.2775603235,-0.1004941463,-0.0483997054,0.2007462978,-0.0364864543,-0.0897136256,-0.1614140123,0.288959831,-0.1801721752,-0.0983834192,0.0543797128,0.0768853873,0.2552024722,-0.0877907947,0.0312968753,-0.3928907812,-0.2555431128,-0.164903596,0.1820681095,0.4742359221,-0.4417522848,-0.1943070441,-0.1998421103,-0.1128745079,0.196072191,0.1654726416,-0.1848081648,0.6032976508,-0.1668522954,0.1821197867,0.4813289344,-0.2698017657,0.1708316952,0.0655022413,-0.2529610395,0.076276429,-0.0907965153,-0.045418676,-0.0019625076,-0.0203296375,0.3685547709,0.2570970356,-0.0868278965,-0.0704206899,0.025802223,0.028126996,0.593431592,0.0462836511,-0.0008243334,-0.2142312676,0.0578708202,0.4939417243,0.2460762858,-0.057706587,-0.0333488211,-0.0137065565,0.0170148499,-0.1148447394,-0.3352916241,-0.2411742806,-0.22320728,-0.1093647555,0.0478304662,0.3630459607,-0.1524047107,0.2927405536,-0.2438348085,0.0848255977,-0.2252929807,-0.2423848957,0.2631049156,0.0465616249,-0.2522629201,-0.1086916551,-0.0686345994,-0.0800719112,0.0088367239,0.0938628986,-0.1322340667,0.1556389332,-0.2579882741,-0.3206021488,0.1305004209,0.256996721,0.1654368639,-0.0436956733,-0.2140459865,0.2780266702,0.1649134308,-0.0692972168,-0.2854145169,-0.0294151921,-0.0743030012,-0.160280183,0.1651479155,0.179478243,0.1970850676,-0.1973392218,-0.2075010985,0.1806471944,-0.023483254,0.0752795711,0.1489133984,0.1009815335,0.1457805783,0.0210331753,0.0285009481,-0.2971132398,0.1780157238,-0.0241094306,-0.1505931169,-0.0727762729,-0.1188315302,-0.1056578383,0.5293703079,0.2276985347,0.2241090834,-0.2177880704,-0.262104243,-0.081199497,0.0782550648,0.070517458,0.4799798429,0.1782226712,-0.0041789506,0.1772151589,-0.0156665258,-0.2165574729,0.216871202,0.1818329394,0.0331380144,-0.0351118222,0.0385485813,0.085117124,-0.1735186428,-0.2818623483,-0.0842829421,0.1642438173,-0.3306981623,0.0497132167,-0.4721483588,-0.3931072354,-0.4514593482,-0.3841335177,-0.1739959866,-0.4288794994,-0.1867683232,-0.0351117514,-0.2173171639,0.3002482057,-0.1404795647,-0.0482548997,0.1101874709,0.0663792789,0.098848708,-0.2670576572,-0.1530420929,0.1411233395,0.3205446899,0.1450353414,0.3701429069,-0.0794283301,0.0111533152,0.0473328829,-0.3408435583,0.1160992384,-0.1009821147,0.2626492679,0.4324879348,0.3414855003,0.0838382915,-0.2842497528,0.4270731211,-0.0637356862,0.0510102436,0.049875319,0.1940965652,-0.14796938,-0.3029948175,-0.4622400105,-0.3657283187,-0.4703692198,-0.12164215,0.1440427452,0.131994009,0.4903213084,-0.0182412937,0.3564141393,-0.0871031806,-0.0005423531,0.07355012,0.0614891574,0.2821263671,-0.312462002,-0.1920034438,0.0803076327,0.03580717,0.2169488519,-0.1713449955,-0.2317764312,0.1126878113,-0.105670698,0.1334697753,0.0481273234,0.101513952,0.1759227514,0.2410977781,-0.2015783489,-0.3006727397,-0.3467811942,0.0994992778,0.1069888249,0.125855729,-0.1017532796,0.1422888935,0.1000687107,0.3848873377,0.0874497145,-0.1828735769,0.3111949563,0.0411565527,0.3453080058,-0.0618227795,-0.3201507926,0.2449907064,-0.0587742515,0.1667978317,0.2006136924,-0.1382380724,0.0803111568,-0.2032093555,-0.0921038389,-0.0643962622,-0.418659091,0.111884892,0.0074180295,0.1880878508,0.2058223784,0.0365455262,-0.194925338,0.0036354265,-0.0097371815,0.2998480797,0.0105650686,0.1356204152,-0.1349021494,-0.0225577783,-0.3069858551,0.1838999689,0.2323766798,0.4915707707,-0.0271859914,-0.1784804314,-0.2188410014,0.0941773877,0.1146481782,-0.13565588,0.1829944998,0.4571828842,0.0663695559,-0.2134140432,-0.1721869856,-0.5033162832,0.1326623857,0.4361288249,0.0677459612,-0.0008687158,-0.3626334071,0.4444147646,0.1816581339,-0.1891018599,-0.2773975432,-0.0000782053,-0.0716028959,-0.336502254,0.0441519283,0.0950581655,0.1696865559,0.2260651141,0.0341112204,0.1553541869,0.0793138146,0.1081611663,0.1017457172,-0.0415175855,0.1275674403,0.0413583294,-0.101653263,0.1490106136,0.2546506524,0.3937421143,-0.1238749921,-0.1940542758,-0.164419353,0.0299259908,0.3551030457,0.0970911607,-0.0819731206,0.1334188133,0.0647364482,0.1381150633,-0.2426775098,0.2743785083,0.2958749831,0.170900777,-0.3747496903,-0.3495922685,-0.0297986846,-0.0531896651,0.1575489193,0.1404528022,-0.0618960597,-0.3537502587,-0.0751761198,-0.0094296345,0.5648851395,0.1630987078,0.289714843,0.4404497743,-0.0870268047,0.473436296,0.1777279079,0.0051305699,-0.246722281,0.1247948483,0.0227170885,-0.1888537854,0.2077721357,0.1377238184,-0.2474355549,0.2713086009,-0.0384190902,0.2527678311,0.065176256,0.090902254,0.3361058533,-0.1986477971,-0.4095683694,0.2058805674,-0.4133999646,0.0716988146,-0.2496188134,-0.3231573999,0.159161225,-0.2507033348,-0.2010598481,0.1776943803,-0.2587882876,0.2284641117,0.0260749403,-0.0663734078,0.4716348946,-0.1141988561,-0.1069330126,0.3006836772,-0.0471390635,-0.0889480561,-0.1888510734,0.0094235633,-0.0697389618,0.247573629,0.5000484586,-0.1218189448,-0.0368878432,-0.1750988066,-0.0812497661,0.1250171214,0.0917121768,0.2032328099,0.1998405457,-0.2620665729,-0.2598944306,-0.0320798419,-0.1772939861,-0.2796158493,0.2072875649,0.0089059155,-0.0291699897,0.2409033477,0.0525551066,-0.1873457432,-0.0364574045,0.1764351875,-0.176746279,-0.066592209,0.5022160411,0.4060555696,-0.2406093776,-0.3803833723,0.1318946034,-0.0628116056,-0.2885809243,0.3547709584,0.1688399315,0.3007945418,0.0165438652,0.3438568413,0.0157201346,-0.0099177929,0.1101292968,-0.3858225644,-0.182272926,0.027388623,-0.342487514,0.0165286101,0.173354581,0.6295419335,0.1122876555,0.0537228286,-0.4009521902,-0.1003775001,-0.0929586515,0.2274139524,0.4283628762,-0.1619137526,-0.0227419026,0.0861172155,0.1760029942,0.0346549228,-0.2526120842,-0.314812988,-0.2940256298,0.0661309883,0.1720108539,-0.4291438162,0.3799619973,0.0724910498,-0.0009614042,-0.3068479002,-0.1781092137,0.0893365964,-0.1307528168,0.3698425293,0.1245556027,0.1172236055,0.175313279,0.0280542914,0.1229266375,0.1096993908,-0.0064216722,0.1405479908,0.0346085429,-0.0513197146,-0.196865961,0.0078801606,0.0852092505,0.2760110795,0.1000634953,-0.4387730658,-0.1408093125,-0.1907574236,0.331359297,0.4964479804,-0.2855861187,-0.0179454908,0.1170451567,0.2279493809,-0.431779623,0.0991013646,0.4069510996,0.0616520308,0.0558136776,0.1605772227,-0.196325466,-0.0263518207,-0.2005510032,-0.015522846,-0.0812871903,-0.2450969666,0.1379801035,0.2535305619,-0.1621550173,0.0723396018,0.2234296203,0.2248426974,0.2094131559,0.4411480725,-0.0880139023,0.0335321389,0.3215833306,-0.1977277696,0.0530192964,-0.4163442254,-0.2213221192,0.004601588,0.1555845737,0.0431443006,-0.0308006573,0.0894242376,0.0554312021,-0.0492895246,0.0911091119,0.1842016131,-0.1754979193,-0.0029799072,0.07021337,0.1035524011,-0.1834444851,-0.2007070631,-0.0523935743,0.0233831629,-0.0611661263,0.0867941976,0.1566937417,-0.5509833097,-0.0985001028,0.028403068,-0.3185489476,-0.4562836587,0.0733105093,0.4109583497,0.0954850167,-0.1203892529,-0.0074952352,0.6174410582,0.4333248734,-0.0764670596,-0.1840572208,-0.2809346318,-0.0467595868,-0.0322065093,0.0866955072,0.1216947138,0.2625560462,0.5520280004,0.1766430885,-0.1991811842,0.1269986182,0.2516283095,-0.2234730422,-0.1569376588,0.2527378798,-0.0195222571,-0.1268068105,-0.2825405598,0.1944848895,-0.2954165936,0.1875001341,0.4098545313,0.2683029473,0.1570728123,-0.1060613543,0.1378096938,0.1873126626,0.5516356826,0.1906979531,0.3860901296,-0.3571788371,-0.0554284528,-0.5711801052,-0.0245900769,-0.1053961366,-0.0316519812,-0.1443446428,-0.0839933902,0.0324577168,0.2593309879,0.1967318356,0.1654993743,0.1051031351,-0.0875484273,-0.2856193185,-0.1786556393,0.0377548598,-0.1408987641,0.1282964647,-0.2223509252,-0.0397654586,-0.4887033701,0.1265330911,-0.3787437677,-0.2208055407,0.0734933987,0.4357133806,0.2232686132,0.2265576571,0.5930403471,-0.0317760035,-0.0286308639,-0.3275698721,-0.2059123963,-0.1545397788,0.3455994725,0.0363045111,0.4889930487,-0.1071662307,-0.0158568323,-0.3778460324,0.1949944496,-0.3336418271,0.0198232438,-0.1027710661,0.1275236309,-0.1381614953,0.1817010939,0.0611189194,0.0021520669,-0.1441734731,0.2395447195,-0.3893561363,-0.4405485988,0.5200942159,-0.4261544347,-0.6753060222,-0.1937554777,0.0424238034,-0.2450479567,-0.1419898719,-0.4619947672,0.062090721,0.2820001543,-0.0690544322,-0.2712805569,0.3742442727,-0.1549424082,0.0212926641,0.0864809677,0.2581366599,0.0770811439,-0.2606575787,-0.2184218317,-0.2339115441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/156","title":"SyntaxError with WMT datasets","comments":"Hi @patrickvonplaten!\r\n\r\nI'm now getting the below error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-3206959998b9> in <module>\r\n 1 import nlp\r\n 2 \r\n----> 3 dataset = nlp.load_dataset('wmt14')\r\n 4 print(dataset['train'][0])\r\n\r\n~\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 507 # Instantiate the dataset builder\r\n 508 builder_instance = builder_cls(\r\n--> 509 cache_dir=cache_dir, name=name, version=version, data_dir=data_dir, data_files=data_files, **config_kwargs,\r\n 510 )\r\n 511 \r\n\r\nTypeError: Can't instantiate abstract class Wmt with abstract methods _subsets\r\n```\r\n\r\n","body":"The following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook","comment_length":76,"text":"SyntaxError with WMT datasets\nThe following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook\nHi @patrickvonplaten!\r\n\r\nI'm now getting the below error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-28-3206959998b9> in <module>\r\n 1 import nlp\r\n 2 \r\n----> 3 dataset = nlp.load_dataset('wmt14')\r\n 4 print(dataset['train'][0])\r\n\r\n~\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 507 # Instantiate the dataset builder\r\n 508 builder_instance = builder_cls(\r\n--> 509 cache_dir=cache_dir, name=name, version=version, data_dir=data_dir, data_files=data_files, **config_kwargs,\r\n 510 )\r\n 511 \r\n\r\nTypeError: Can't instantiate abstract class Wmt with abstract methods _subsets\r\n```\r\n\r\n","embeddings":[-0.32973665,0.0324432179,-0.0193531662,0.0127459355,0.2164737433,-0.0149221374,0.3609470129,0.5264438987,0.1301548183,-0.1121621057,0.0932864323,0.4350022972,-0.3767533898,0.1414317936,0.1508708149,-0.1754975915,-0.0268027131,0.1413965672,-0.3586361706,0.151956737,-0.2946045697,0.1583642811,-0.420961529,0.2173392326,-0.2542883456,-0.0902357399,-0.021146521,0.144950211,-0.1394597441,-0.462849319,0.140802905,-0.0065226816,0.2389953732,0.3692950606,-0.0001028357,-0.0074911187,0.1617367566,-0.0390187725,-0.3923019171,-0.3421501517,-0.4517919123,-0.1481353492,0.143838644,-0.2942019701,0.0407955088,-0.1019728407,0.012090249,-0.2335641831,0.5227859616,0.470978111,0.2622492909,0.4046530128,-0.0788426101,-0.1995568573,0.0400845148,0.1868067235,-0.0251018368,0.2721501887,-0.0462756231,-0.3078270257,0.0443689711,0.2844180167,-0.0971550569,0.2046322972,0.2288134843,0.0462167524,0.1737399995,0.0146226026,-0.1336928159,0.111689046,0.3766534328,-0.3340502977,-0.1922972947,-0.0254198257,-0.0237156339,-0.3796640635,0.1737448573,0.2510075867,-0.2430952042,0.2042771727,-0.2803643048,0.0721920952,-0.0495335385,0.280293107,-0.0419157632,0.7289190888,-0.1515552253,0.0830518156,-0.0326680392,-0.1897672564,0.1182416454,-0.0233938098,-0.0012605379,0.1437297612,-0.0696197599,-0.0688927695,0.2810450196,-0.0273937248,-0.2340159565,0.0267737973,0.1816953868,-0.1459489912,0.1988346726,0.1966277063,0.283844322,-0.064487271,0.0399041548,-0.0328306854,0.2843933702,0.2534207404,-0.027510548,0.0475533232,-0.1733619273,-0.4689917266,-0.0253391545,0.0136443032,0.5950089097,0.1803969294,-0.2701009512,-0.0289219059,-0.2787000239,-0.302798748,-0.0676975995,0.2093273848,0.121819444,-0.1600497216,0.3483577371,0.0752085149,-0.2440336943,-0.1647191793,-0.2930479646,0.340895921,-0.3188335896,-0.1684278697,0.2188079953,0.2253614962,0.0922131017,0.0471641496,-0.0809147358,-0.1988425553,0.1389292777,-0.0077622063,0.1081167087,0.1051954702,-0.1785956919,0.093339853,0.2775603235,-0.1004941463,-0.0483997054,0.2007462978,-0.0364864543,-0.0897136256,-0.1614140123,0.288959831,-0.1801721752,-0.0983834192,0.0543797128,0.0768853873,0.2552024722,-0.0877907947,0.0312968753,-0.3928907812,-0.2555431128,-0.164903596,0.1820681095,0.4742359221,-0.4417522848,-0.1943070441,-0.1998421103,-0.1128745079,0.196072191,0.1654726416,-0.1848081648,0.6032976508,-0.1668522954,0.1821197867,0.4813289344,-0.2698017657,0.1708316952,0.0655022413,-0.2529610395,0.076276429,-0.0907965153,-0.045418676,-0.0019625076,-0.0203296375,0.3685547709,0.2570970356,-0.0868278965,-0.0704206899,0.025802223,0.028126996,0.593431592,0.0462836511,-0.0008243334,-0.2142312676,0.0578708202,0.4939417243,0.2460762858,-0.057706587,-0.0333488211,-0.0137065565,0.0170148499,-0.1148447394,-0.3352916241,-0.2411742806,-0.22320728,-0.1093647555,0.0478304662,0.3630459607,-0.1524047107,0.2927405536,-0.2438348085,0.0848255977,-0.2252929807,-0.2423848957,0.2631049156,0.0465616249,-0.2522629201,-0.1086916551,-0.0686345994,-0.0800719112,0.0088367239,0.0938628986,-0.1322340667,0.1556389332,-0.2579882741,-0.3206021488,0.1305004209,0.256996721,0.1654368639,-0.0436956733,-0.2140459865,0.2780266702,0.1649134308,-0.0692972168,-0.2854145169,-0.0294151921,-0.0743030012,-0.160280183,0.1651479155,0.179478243,0.1970850676,-0.1973392218,-0.2075010985,0.1806471944,-0.023483254,0.0752795711,0.1489133984,0.1009815335,0.1457805783,0.0210331753,0.0285009481,-0.2971132398,0.1780157238,-0.0241094306,-0.1505931169,-0.0727762729,-0.1188315302,-0.1056578383,0.5293703079,0.2276985347,0.2241090834,-0.2177880704,-0.262104243,-0.081199497,0.0782550648,0.070517458,0.4799798429,0.1782226712,-0.0041789506,0.1772151589,-0.0156665258,-0.2165574729,0.216871202,0.1818329394,0.0331380144,-0.0351118222,0.0385485813,0.085117124,-0.1735186428,-0.2818623483,-0.0842829421,0.1642438173,-0.3306981623,0.0497132167,-0.4721483588,-0.3931072354,-0.4514593482,-0.3841335177,-0.1739959866,-0.4288794994,-0.1867683232,-0.0351117514,-0.2173171639,0.3002482057,-0.1404795647,-0.0482548997,0.1101874709,0.0663792789,0.098848708,-0.2670576572,-0.1530420929,0.1411233395,0.3205446899,0.1450353414,0.3701429069,-0.0794283301,0.0111533152,0.0473328829,-0.3408435583,0.1160992384,-0.1009821147,0.2626492679,0.4324879348,0.3414855003,0.0838382915,-0.2842497528,0.4270731211,-0.0637356862,0.0510102436,0.049875319,0.1940965652,-0.14796938,-0.3029948175,-0.4622400105,-0.3657283187,-0.4703692198,-0.12164215,0.1440427452,0.131994009,0.4903213084,-0.0182412937,0.3564141393,-0.0871031806,-0.0005423531,0.07355012,0.0614891574,0.2821263671,-0.312462002,-0.1920034438,0.0803076327,0.03580717,0.2169488519,-0.1713449955,-0.2317764312,0.1126878113,-0.105670698,0.1334697753,0.0481273234,0.101513952,0.1759227514,0.2410977781,-0.2015783489,-0.3006727397,-0.3467811942,0.0994992778,0.1069888249,0.125855729,-0.1017532796,0.1422888935,0.1000687107,0.3848873377,0.0874497145,-0.1828735769,0.3111949563,0.0411565527,0.3453080058,-0.0618227795,-0.3201507926,0.2449907064,-0.0587742515,0.1667978317,0.2006136924,-0.1382380724,0.0803111568,-0.2032093555,-0.0921038389,-0.0643962622,-0.418659091,0.111884892,0.0074180295,0.1880878508,0.2058223784,0.0365455262,-0.194925338,0.0036354265,-0.0097371815,0.2998480797,0.0105650686,0.1356204152,-0.1349021494,-0.0225577783,-0.3069858551,0.1838999689,0.2323766798,0.4915707707,-0.0271859914,-0.1784804314,-0.2188410014,0.0941773877,0.1146481782,-0.13565588,0.1829944998,0.4571828842,0.0663695559,-0.2134140432,-0.1721869856,-0.5033162832,0.1326623857,0.4361288249,0.0677459612,-0.0008687158,-0.3626334071,0.4444147646,0.1816581339,-0.1891018599,-0.2773975432,-0.0000782053,-0.0716028959,-0.336502254,0.0441519283,0.0950581655,0.1696865559,0.2260651141,0.0341112204,0.1553541869,0.0793138146,0.1081611663,0.1017457172,-0.0415175855,0.1275674403,0.0413583294,-0.101653263,0.1490106136,0.2546506524,0.3937421143,-0.1238749921,-0.1940542758,-0.164419353,0.0299259908,0.3551030457,0.0970911607,-0.0819731206,0.1334188133,0.0647364482,0.1381150633,-0.2426775098,0.2743785083,0.2958749831,0.170900777,-0.3747496903,-0.3495922685,-0.0297986846,-0.0531896651,0.1575489193,0.1404528022,-0.0618960597,-0.3537502587,-0.0751761198,-0.0094296345,0.5648851395,0.1630987078,0.289714843,0.4404497743,-0.0870268047,0.473436296,0.1777279079,0.0051305699,-0.246722281,0.1247948483,0.0227170885,-0.1888537854,0.2077721357,0.1377238184,-0.2474355549,0.2713086009,-0.0384190902,0.2527678311,0.065176256,0.090902254,0.3361058533,-0.1986477971,-0.4095683694,0.2058805674,-0.4133999646,0.0716988146,-0.2496188134,-0.3231573999,0.159161225,-0.2507033348,-0.2010598481,0.1776943803,-0.2587882876,0.2284641117,0.0260749403,-0.0663734078,0.4716348946,-0.1141988561,-0.1069330126,0.3006836772,-0.0471390635,-0.0889480561,-0.1888510734,0.0094235633,-0.0697389618,0.247573629,0.5000484586,-0.1218189448,-0.0368878432,-0.1750988066,-0.0812497661,0.1250171214,0.0917121768,0.2032328099,0.1998405457,-0.2620665729,-0.2598944306,-0.0320798419,-0.1772939861,-0.2796158493,0.2072875649,0.0089059155,-0.0291699897,0.2409033477,0.0525551066,-0.1873457432,-0.0364574045,0.1764351875,-0.176746279,-0.066592209,0.5022160411,0.4060555696,-0.2406093776,-0.3803833723,0.1318946034,-0.0628116056,-0.2885809243,0.3547709584,0.1688399315,0.3007945418,0.0165438652,0.3438568413,0.0157201346,-0.0099177929,0.1101292968,-0.3858225644,-0.182272926,0.027388623,-0.342487514,0.0165286101,0.173354581,0.6295419335,0.1122876555,0.0537228286,-0.4009521902,-0.1003775001,-0.0929586515,0.2274139524,0.4283628762,-0.1619137526,-0.0227419026,0.0861172155,0.1760029942,0.0346549228,-0.2526120842,-0.314812988,-0.2940256298,0.0661309883,0.1720108539,-0.4291438162,0.3799619973,0.0724910498,-0.0009614042,-0.3068479002,-0.1781092137,0.0893365964,-0.1307528168,0.3698425293,0.1245556027,0.1172236055,0.175313279,0.0280542914,0.1229266375,0.1096993908,-0.0064216722,0.1405479908,0.0346085429,-0.0513197146,-0.196865961,0.0078801606,0.0852092505,0.2760110795,0.1000634953,-0.4387730658,-0.1408093125,-0.1907574236,0.331359297,0.4964479804,-0.2855861187,-0.0179454908,0.1170451567,0.2279493809,-0.431779623,0.0991013646,0.4069510996,0.0616520308,0.0558136776,0.1605772227,-0.196325466,-0.0263518207,-0.2005510032,-0.015522846,-0.0812871903,-0.2450969666,0.1379801035,0.2535305619,-0.1621550173,0.0723396018,0.2234296203,0.2248426974,0.2094131559,0.4411480725,-0.0880139023,0.0335321389,0.3215833306,-0.1977277696,0.0530192964,-0.4163442254,-0.2213221192,0.004601588,0.1555845737,0.0431443006,-0.0308006573,0.0894242376,0.0554312021,-0.0492895246,0.0911091119,0.1842016131,-0.1754979193,-0.0029799072,0.07021337,0.1035524011,-0.1834444851,-0.2007070631,-0.0523935743,0.0233831629,-0.0611661263,0.0867941976,0.1566937417,-0.5509833097,-0.0985001028,0.028403068,-0.3185489476,-0.4562836587,0.0733105093,0.4109583497,0.0954850167,-0.1203892529,-0.0074952352,0.6174410582,0.4333248734,-0.0764670596,-0.1840572208,-0.2809346318,-0.0467595868,-0.0322065093,0.0866955072,0.1216947138,0.2625560462,0.5520280004,0.1766430885,-0.1991811842,0.1269986182,0.2516283095,-0.2234730422,-0.1569376588,0.2527378798,-0.0195222571,-0.1268068105,-0.2825405598,0.1944848895,-0.2954165936,0.1875001341,0.4098545313,0.2683029473,0.1570728123,-0.1060613543,0.1378096938,0.1873126626,0.5516356826,0.1906979531,0.3860901296,-0.3571788371,-0.0554284528,-0.5711801052,-0.0245900769,-0.1053961366,-0.0316519812,-0.1443446428,-0.0839933902,0.0324577168,0.2593309879,0.1967318356,0.1654993743,0.1051031351,-0.0875484273,-0.2856193185,-0.1786556393,0.0377548598,-0.1408987641,0.1282964647,-0.2223509252,-0.0397654586,-0.4887033701,0.1265330911,-0.3787437677,-0.2208055407,0.0734933987,0.4357133806,0.2232686132,0.2265576571,0.5930403471,-0.0317760035,-0.0286308639,-0.3275698721,-0.2059123963,-0.1545397788,0.3455994725,0.0363045111,0.4889930487,-0.1071662307,-0.0158568323,-0.3778460324,0.1949944496,-0.3336418271,0.0198232438,-0.1027710661,0.1275236309,-0.1381614953,0.1817010939,0.0611189194,0.0021520669,-0.1441734731,0.2395447195,-0.3893561363,-0.4405485988,0.5200942159,-0.4261544347,-0.6753060222,-0.1937554777,0.0424238034,-0.2450479567,-0.1419898719,-0.4619947672,0.062090721,0.2820001543,-0.0690544322,-0.2712805569,0.3742442727,-0.1549424082,0.0212926641,0.0864809677,0.2581366599,0.0770811439,-0.2606575787,-0.2184218317,-0.2339115441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/156","title":"SyntaxError with WMT datasets","comments":"To correct this error I think you need the master branch of `nlp`. Can you try to install `nlp` with. `WMT` was not included at the beta release of the library. \r\n\r\nCan you try:\r\n`pip install git+https:\/\/github.com\/huggingface\/nlp.git`\r\n\r\nand check again? ","body":"The following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook","comment_length":40,"text":"SyntaxError with WMT datasets\nThe following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook\nTo correct this error I think you need the master branch of `nlp`. Can you try to install `nlp` with. `WMT` was not included at the beta release of the library. \r\n\r\nCan you try:\r\n`pip install git+https:\/\/github.com\/huggingface\/nlp.git`\r\n\r\nand check again? ","embeddings":[-0.32973665,0.0324432179,-0.0193531662,0.0127459355,0.2164737433,-0.0149221374,0.3609470129,0.5264438987,0.1301548183,-0.1121621057,0.0932864323,0.4350022972,-0.3767533898,0.1414317936,0.1508708149,-0.1754975915,-0.0268027131,0.1413965672,-0.3586361706,0.151956737,-0.2946045697,0.1583642811,-0.420961529,0.2173392326,-0.2542883456,-0.0902357399,-0.021146521,0.144950211,-0.1394597441,-0.462849319,0.140802905,-0.0065226816,0.2389953732,0.3692950606,-0.0001028357,-0.0074911187,0.1617367566,-0.0390187725,-0.3923019171,-0.3421501517,-0.4517919123,-0.1481353492,0.143838644,-0.2942019701,0.0407955088,-0.1019728407,0.012090249,-0.2335641831,0.5227859616,0.470978111,0.2622492909,0.4046530128,-0.0788426101,-0.1995568573,0.0400845148,0.1868067235,-0.0251018368,0.2721501887,-0.0462756231,-0.3078270257,0.0443689711,0.2844180167,-0.0971550569,0.2046322972,0.2288134843,0.0462167524,0.1737399995,0.0146226026,-0.1336928159,0.111689046,0.3766534328,-0.3340502977,-0.1922972947,-0.0254198257,-0.0237156339,-0.3796640635,0.1737448573,0.2510075867,-0.2430952042,0.2042771727,-0.2803643048,0.0721920952,-0.0495335385,0.280293107,-0.0419157632,0.7289190888,-0.1515552253,0.0830518156,-0.0326680392,-0.1897672564,0.1182416454,-0.0233938098,-0.0012605379,0.1437297612,-0.0696197599,-0.0688927695,0.2810450196,-0.0273937248,-0.2340159565,0.0267737973,0.1816953868,-0.1459489912,0.1988346726,0.1966277063,0.283844322,-0.064487271,0.0399041548,-0.0328306854,0.2843933702,0.2534207404,-0.027510548,0.0475533232,-0.1733619273,-0.4689917266,-0.0253391545,0.0136443032,0.5950089097,0.1803969294,-0.2701009512,-0.0289219059,-0.2787000239,-0.302798748,-0.0676975995,0.2093273848,0.121819444,-0.1600497216,0.3483577371,0.0752085149,-0.2440336943,-0.1647191793,-0.2930479646,0.340895921,-0.3188335896,-0.1684278697,0.2188079953,0.2253614962,0.0922131017,0.0471641496,-0.0809147358,-0.1988425553,0.1389292777,-0.0077622063,0.1081167087,0.1051954702,-0.1785956919,0.093339853,0.2775603235,-0.1004941463,-0.0483997054,0.2007462978,-0.0364864543,-0.0897136256,-0.1614140123,0.288959831,-0.1801721752,-0.0983834192,0.0543797128,0.0768853873,0.2552024722,-0.0877907947,0.0312968753,-0.3928907812,-0.2555431128,-0.164903596,0.1820681095,0.4742359221,-0.4417522848,-0.1943070441,-0.1998421103,-0.1128745079,0.196072191,0.1654726416,-0.1848081648,0.6032976508,-0.1668522954,0.1821197867,0.4813289344,-0.2698017657,0.1708316952,0.0655022413,-0.2529610395,0.076276429,-0.0907965153,-0.045418676,-0.0019625076,-0.0203296375,0.3685547709,0.2570970356,-0.0868278965,-0.0704206899,0.025802223,0.028126996,0.593431592,0.0462836511,-0.0008243334,-0.2142312676,0.0578708202,0.4939417243,0.2460762858,-0.057706587,-0.0333488211,-0.0137065565,0.0170148499,-0.1148447394,-0.3352916241,-0.2411742806,-0.22320728,-0.1093647555,0.0478304662,0.3630459607,-0.1524047107,0.2927405536,-0.2438348085,0.0848255977,-0.2252929807,-0.2423848957,0.2631049156,0.0465616249,-0.2522629201,-0.1086916551,-0.0686345994,-0.0800719112,0.0088367239,0.0938628986,-0.1322340667,0.1556389332,-0.2579882741,-0.3206021488,0.1305004209,0.256996721,0.1654368639,-0.0436956733,-0.2140459865,0.2780266702,0.1649134308,-0.0692972168,-0.2854145169,-0.0294151921,-0.0743030012,-0.160280183,0.1651479155,0.179478243,0.1970850676,-0.1973392218,-0.2075010985,0.1806471944,-0.023483254,0.0752795711,0.1489133984,0.1009815335,0.1457805783,0.0210331753,0.0285009481,-0.2971132398,0.1780157238,-0.0241094306,-0.1505931169,-0.0727762729,-0.1188315302,-0.1056578383,0.5293703079,0.2276985347,0.2241090834,-0.2177880704,-0.262104243,-0.081199497,0.0782550648,0.070517458,0.4799798429,0.1782226712,-0.0041789506,0.1772151589,-0.0156665258,-0.2165574729,0.216871202,0.1818329394,0.0331380144,-0.0351118222,0.0385485813,0.085117124,-0.1735186428,-0.2818623483,-0.0842829421,0.1642438173,-0.3306981623,0.0497132167,-0.4721483588,-0.3931072354,-0.4514593482,-0.3841335177,-0.1739959866,-0.4288794994,-0.1867683232,-0.0351117514,-0.2173171639,0.3002482057,-0.1404795647,-0.0482548997,0.1101874709,0.0663792789,0.098848708,-0.2670576572,-0.1530420929,0.1411233395,0.3205446899,0.1450353414,0.3701429069,-0.0794283301,0.0111533152,0.0473328829,-0.3408435583,0.1160992384,-0.1009821147,0.2626492679,0.4324879348,0.3414855003,0.0838382915,-0.2842497528,0.4270731211,-0.0637356862,0.0510102436,0.049875319,0.1940965652,-0.14796938,-0.3029948175,-0.4622400105,-0.3657283187,-0.4703692198,-0.12164215,0.1440427452,0.131994009,0.4903213084,-0.0182412937,0.3564141393,-0.0871031806,-0.0005423531,0.07355012,0.0614891574,0.2821263671,-0.312462002,-0.1920034438,0.0803076327,0.03580717,0.2169488519,-0.1713449955,-0.2317764312,0.1126878113,-0.105670698,0.1334697753,0.0481273234,0.101513952,0.1759227514,0.2410977781,-0.2015783489,-0.3006727397,-0.3467811942,0.0994992778,0.1069888249,0.125855729,-0.1017532796,0.1422888935,0.1000687107,0.3848873377,0.0874497145,-0.1828735769,0.3111949563,0.0411565527,0.3453080058,-0.0618227795,-0.3201507926,0.2449907064,-0.0587742515,0.1667978317,0.2006136924,-0.1382380724,0.0803111568,-0.2032093555,-0.0921038389,-0.0643962622,-0.418659091,0.111884892,0.0074180295,0.1880878508,0.2058223784,0.0365455262,-0.194925338,0.0036354265,-0.0097371815,0.2998480797,0.0105650686,0.1356204152,-0.1349021494,-0.0225577783,-0.3069858551,0.1838999689,0.2323766798,0.4915707707,-0.0271859914,-0.1784804314,-0.2188410014,0.0941773877,0.1146481782,-0.13565588,0.1829944998,0.4571828842,0.0663695559,-0.2134140432,-0.1721869856,-0.5033162832,0.1326623857,0.4361288249,0.0677459612,-0.0008687158,-0.3626334071,0.4444147646,0.1816581339,-0.1891018599,-0.2773975432,-0.0000782053,-0.0716028959,-0.336502254,0.0441519283,0.0950581655,0.1696865559,0.2260651141,0.0341112204,0.1553541869,0.0793138146,0.1081611663,0.1017457172,-0.0415175855,0.1275674403,0.0413583294,-0.101653263,0.1490106136,0.2546506524,0.3937421143,-0.1238749921,-0.1940542758,-0.164419353,0.0299259908,0.3551030457,0.0970911607,-0.0819731206,0.1334188133,0.0647364482,0.1381150633,-0.2426775098,0.2743785083,0.2958749831,0.170900777,-0.3747496903,-0.3495922685,-0.0297986846,-0.0531896651,0.1575489193,0.1404528022,-0.0618960597,-0.3537502587,-0.0751761198,-0.0094296345,0.5648851395,0.1630987078,0.289714843,0.4404497743,-0.0870268047,0.473436296,0.1777279079,0.0051305699,-0.246722281,0.1247948483,0.0227170885,-0.1888537854,0.2077721357,0.1377238184,-0.2474355549,0.2713086009,-0.0384190902,0.2527678311,0.065176256,0.090902254,0.3361058533,-0.1986477971,-0.4095683694,0.2058805674,-0.4133999646,0.0716988146,-0.2496188134,-0.3231573999,0.159161225,-0.2507033348,-0.2010598481,0.1776943803,-0.2587882876,0.2284641117,0.0260749403,-0.0663734078,0.4716348946,-0.1141988561,-0.1069330126,0.3006836772,-0.0471390635,-0.0889480561,-0.1888510734,0.0094235633,-0.0697389618,0.247573629,0.5000484586,-0.1218189448,-0.0368878432,-0.1750988066,-0.0812497661,0.1250171214,0.0917121768,0.2032328099,0.1998405457,-0.2620665729,-0.2598944306,-0.0320798419,-0.1772939861,-0.2796158493,0.2072875649,0.0089059155,-0.0291699897,0.2409033477,0.0525551066,-0.1873457432,-0.0364574045,0.1764351875,-0.176746279,-0.066592209,0.5022160411,0.4060555696,-0.2406093776,-0.3803833723,0.1318946034,-0.0628116056,-0.2885809243,0.3547709584,0.1688399315,0.3007945418,0.0165438652,0.3438568413,0.0157201346,-0.0099177929,0.1101292968,-0.3858225644,-0.182272926,0.027388623,-0.342487514,0.0165286101,0.173354581,0.6295419335,0.1122876555,0.0537228286,-0.4009521902,-0.1003775001,-0.0929586515,0.2274139524,0.4283628762,-0.1619137526,-0.0227419026,0.0861172155,0.1760029942,0.0346549228,-0.2526120842,-0.314812988,-0.2940256298,0.0661309883,0.1720108539,-0.4291438162,0.3799619973,0.0724910498,-0.0009614042,-0.3068479002,-0.1781092137,0.0893365964,-0.1307528168,0.3698425293,0.1245556027,0.1172236055,0.175313279,0.0280542914,0.1229266375,0.1096993908,-0.0064216722,0.1405479908,0.0346085429,-0.0513197146,-0.196865961,0.0078801606,0.0852092505,0.2760110795,0.1000634953,-0.4387730658,-0.1408093125,-0.1907574236,0.331359297,0.4964479804,-0.2855861187,-0.0179454908,0.1170451567,0.2279493809,-0.431779623,0.0991013646,0.4069510996,0.0616520308,0.0558136776,0.1605772227,-0.196325466,-0.0263518207,-0.2005510032,-0.015522846,-0.0812871903,-0.2450969666,0.1379801035,0.2535305619,-0.1621550173,0.0723396018,0.2234296203,0.2248426974,0.2094131559,0.4411480725,-0.0880139023,0.0335321389,0.3215833306,-0.1977277696,0.0530192964,-0.4163442254,-0.2213221192,0.004601588,0.1555845737,0.0431443006,-0.0308006573,0.0894242376,0.0554312021,-0.0492895246,0.0911091119,0.1842016131,-0.1754979193,-0.0029799072,0.07021337,0.1035524011,-0.1834444851,-0.2007070631,-0.0523935743,0.0233831629,-0.0611661263,0.0867941976,0.1566937417,-0.5509833097,-0.0985001028,0.028403068,-0.3185489476,-0.4562836587,0.0733105093,0.4109583497,0.0954850167,-0.1203892529,-0.0074952352,0.6174410582,0.4333248734,-0.0764670596,-0.1840572208,-0.2809346318,-0.0467595868,-0.0322065093,0.0866955072,0.1216947138,0.2625560462,0.5520280004,0.1766430885,-0.1991811842,0.1269986182,0.2516283095,-0.2234730422,-0.1569376588,0.2527378798,-0.0195222571,-0.1268068105,-0.2825405598,0.1944848895,-0.2954165936,0.1875001341,0.4098545313,0.2683029473,0.1570728123,-0.1060613543,0.1378096938,0.1873126626,0.5516356826,0.1906979531,0.3860901296,-0.3571788371,-0.0554284528,-0.5711801052,-0.0245900769,-0.1053961366,-0.0316519812,-0.1443446428,-0.0839933902,0.0324577168,0.2593309879,0.1967318356,0.1654993743,0.1051031351,-0.0875484273,-0.2856193185,-0.1786556393,0.0377548598,-0.1408987641,0.1282964647,-0.2223509252,-0.0397654586,-0.4887033701,0.1265330911,-0.3787437677,-0.2208055407,0.0734933987,0.4357133806,0.2232686132,0.2265576571,0.5930403471,-0.0317760035,-0.0286308639,-0.3275698721,-0.2059123963,-0.1545397788,0.3455994725,0.0363045111,0.4889930487,-0.1071662307,-0.0158568323,-0.3778460324,0.1949944496,-0.3336418271,0.0198232438,-0.1027710661,0.1275236309,-0.1381614953,0.1817010939,0.0611189194,0.0021520669,-0.1441734731,0.2395447195,-0.3893561363,-0.4405485988,0.5200942159,-0.4261544347,-0.6753060222,-0.1937554777,0.0424238034,-0.2450479567,-0.1419898719,-0.4619947672,0.062090721,0.2820001543,-0.0690544322,-0.2712805569,0.3742442727,-0.1549424082,0.0212926641,0.0864809677,0.2581366599,0.0770811439,-0.2606575787,-0.2184218317,-0.2339115441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/156","title":"SyntaxError with WMT datasets","comments":"That works, thanks :)\r\n\r\nThe WMT datasets are listed in by `list_datasets()` in the beta release on pypi - it would be good to only show datasets that are actually supported by that version?","body":"The following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook","comment_length":34,"text":"SyntaxError with WMT datasets\nThe following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook\nThat works, thanks :)\r\n\r\nThe WMT datasets are listed in by `list_datasets()` in the beta release on pypi - it would be good to only show datasets that are actually supported by that version?","embeddings":[-0.32973665,0.0324432179,-0.0193531662,0.0127459355,0.2164737433,-0.0149221374,0.3609470129,0.5264438987,0.1301548183,-0.1121621057,0.0932864323,0.4350022972,-0.3767533898,0.1414317936,0.1508708149,-0.1754975915,-0.0268027131,0.1413965672,-0.3586361706,0.151956737,-0.2946045697,0.1583642811,-0.420961529,0.2173392326,-0.2542883456,-0.0902357399,-0.021146521,0.144950211,-0.1394597441,-0.462849319,0.140802905,-0.0065226816,0.2389953732,0.3692950606,-0.0001028357,-0.0074911187,0.1617367566,-0.0390187725,-0.3923019171,-0.3421501517,-0.4517919123,-0.1481353492,0.143838644,-0.2942019701,0.0407955088,-0.1019728407,0.012090249,-0.2335641831,0.5227859616,0.470978111,0.2622492909,0.4046530128,-0.0788426101,-0.1995568573,0.0400845148,0.1868067235,-0.0251018368,0.2721501887,-0.0462756231,-0.3078270257,0.0443689711,0.2844180167,-0.0971550569,0.2046322972,0.2288134843,0.0462167524,0.1737399995,0.0146226026,-0.1336928159,0.111689046,0.3766534328,-0.3340502977,-0.1922972947,-0.0254198257,-0.0237156339,-0.3796640635,0.1737448573,0.2510075867,-0.2430952042,0.2042771727,-0.2803643048,0.0721920952,-0.0495335385,0.280293107,-0.0419157632,0.7289190888,-0.1515552253,0.0830518156,-0.0326680392,-0.1897672564,0.1182416454,-0.0233938098,-0.0012605379,0.1437297612,-0.0696197599,-0.0688927695,0.2810450196,-0.0273937248,-0.2340159565,0.0267737973,0.1816953868,-0.1459489912,0.1988346726,0.1966277063,0.283844322,-0.064487271,0.0399041548,-0.0328306854,0.2843933702,0.2534207404,-0.027510548,0.0475533232,-0.1733619273,-0.4689917266,-0.0253391545,0.0136443032,0.5950089097,0.1803969294,-0.2701009512,-0.0289219059,-0.2787000239,-0.302798748,-0.0676975995,0.2093273848,0.121819444,-0.1600497216,0.3483577371,0.0752085149,-0.2440336943,-0.1647191793,-0.2930479646,0.340895921,-0.3188335896,-0.1684278697,0.2188079953,0.2253614962,0.0922131017,0.0471641496,-0.0809147358,-0.1988425553,0.1389292777,-0.0077622063,0.1081167087,0.1051954702,-0.1785956919,0.093339853,0.2775603235,-0.1004941463,-0.0483997054,0.2007462978,-0.0364864543,-0.0897136256,-0.1614140123,0.288959831,-0.1801721752,-0.0983834192,0.0543797128,0.0768853873,0.2552024722,-0.0877907947,0.0312968753,-0.3928907812,-0.2555431128,-0.164903596,0.1820681095,0.4742359221,-0.4417522848,-0.1943070441,-0.1998421103,-0.1128745079,0.196072191,0.1654726416,-0.1848081648,0.6032976508,-0.1668522954,0.1821197867,0.4813289344,-0.2698017657,0.1708316952,0.0655022413,-0.2529610395,0.076276429,-0.0907965153,-0.045418676,-0.0019625076,-0.0203296375,0.3685547709,0.2570970356,-0.0868278965,-0.0704206899,0.025802223,0.028126996,0.593431592,0.0462836511,-0.0008243334,-0.2142312676,0.0578708202,0.4939417243,0.2460762858,-0.057706587,-0.0333488211,-0.0137065565,0.0170148499,-0.1148447394,-0.3352916241,-0.2411742806,-0.22320728,-0.1093647555,0.0478304662,0.3630459607,-0.1524047107,0.2927405536,-0.2438348085,0.0848255977,-0.2252929807,-0.2423848957,0.2631049156,0.0465616249,-0.2522629201,-0.1086916551,-0.0686345994,-0.0800719112,0.0088367239,0.0938628986,-0.1322340667,0.1556389332,-0.2579882741,-0.3206021488,0.1305004209,0.256996721,0.1654368639,-0.0436956733,-0.2140459865,0.2780266702,0.1649134308,-0.0692972168,-0.2854145169,-0.0294151921,-0.0743030012,-0.160280183,0.1651479155,0.179478243,0.1970850676,-0.1973392218,-0.2075010985,0.1806471944,-0.023483254,0.0752795711,0.1489133984,0.1009815335,0.1457805783,0.0210331753,0.0285009481,-0.2971132398,0.1780157238,-0.0241094306,-0.1505931169,-0.0727762729,-0.1188315302,-0.1056578383,0.5293703079,0.2276985347,0.2241090834,-0.2177880704,-0.262104243,-0.081199497,0.0782550648,0.070517458,0.4799798429,0.1782226712,-0.0041789506,0.1772151589,-0.0156665258,-0.2165574729,0.216871202,0.1818329394,0.0331380144,-0.0351118222,0.0385485813,0.085117124,-0.1735186428,-0.2818623483,-0.0842829421,0.1642438173,-0.3306981623,0.0497132167,-0.4721483588,-0.3931072354,-0.4514593482,-0.3841335177,-0.1739959866,-0.4288794994,-0.1867683232,-0.0351117514,-0.2173171639,0.3002482057,-0.1404795647,-0.0482548997,0.1101874709,0.0663792789,0.098848708,-0.2670576572,-0.1530420929,0.1411233395,0.3205446899,0.1450353414,0.3701429069,-0.0794283301,0.0111533152,0.0473328829,-0.3408435583,0.1160992384,-0.1009821147,0.2626492679,0.4324879348,0.3414855003,0.0838382915,-0.2842497528,0.4270731211,-0.0637356862,0.0510102436,0.049875319,0.1940965652,-0.14796938,-0.3029948175,-0.4622400105,-0.3657283187,-0.4703692198,-0.12164215,0.1440427452,0.131994009,0.4903213084,-0.0182412937,0.3564141393,-0.0871031806,-0.0005423531,0.07355012,0.0614891574,0.2821263671,-0.312462002,-0.1920034438,0.0803076327,0.03580717,0.2169488519,-0.1713449955,-0.2317764312,0.1126878113,-0.105670698,0.1334697753,0.0481273234,0.101513952,0.1759227514,0.2410977781,-0.2015783489,-0.3006727397,-0.3467811942,0.0994992778,0.1069888249,0.125855729,-0.1017532796,0.1422888935,0.1000687107,0.3848873377,0.0874497145,-0.1828735769,0.3111949563,0.0411565527,0.3453080058,-0.0618227795,-0.3201507926,0.2449907064,-0.0587742515,0.1667978317,0.2006136924,-0.1382380724,0.0803111568,-0.2032093555,-0.0921038389,-0.0643962622,-0.418659091,0.111884892,0.0074180295,0.1880878508,0.2058223784,0.0365455262,-0.194925338,0.0036354265,-0.0097371815,0.2998480797,0.0105650686,0.1356204152,-0.1349021494,-0.0225577783,-0.3069858551,0.1838999689,0.2323766798,0.4915707707,-0.0271859914,-0.1784804314,-0.2188410014,0.0941773877,0.1146481782,-0.13565588,0.1829944998,0.4571828842,0.0663695559,-0.2134140432,-0.1721869856,-0.5033162832,0.1326623857,0.4361288249,0.0677459612,-0.0008687158,-0.3626334071,0.4444147646,0.1816581339,-0.1891018599,-0.2773975432,-0.0000782053,-0.0716028959,-0.336502254,0.0441519283,0.0950581655,0.1696865559,0.2260651141,0.0341112204,0.1553541869,0.0793138146,0.1081611663,0.1017457172,-0.0415175855,0.1275674403,0.0413583294,-0.101653263,0.1490106136,0.2546506524,0.3937421143,-0.1238749921,-0.1940542758,-0.164419353,0.0299259908,0.3551030457,0.0970911607,-0.0819731206,0.1334188133,0.0647364482,0.1381150633,-0.2426775098,0.2743785083,0.2958749831,0.170900777,-0.3747496903,-0.3495922685,-0.0297986846,-0.0531896651,0.1575489193,0.1404528022,-0.0618960597,-0.3537502587,-0.0751761198,-0.0094296345,0.5648851395,0.1630987078,0.289714843,0.4404497743,-0.0870268047,0.473436296,0.1777279079,0.0051305699,-0.246722281,0.1247948483,0.0227170885,-0.1888537854,0.2077721357,0.1377238184,-0.2474355549,0.2713086009,-0.0384190902,0.2527678311,0.065176256,0.090902254,0.3361058533,-0.1986477971,-0.4095683694,0.2058805674,-0.4133999646,0.0716988146,-0.2496188134,-0.3231573999,0.159161225,-0.2507033348,-0.2010598481,0.1776943803,-0.2587882876,0.2284641117,0.0260749403,-0.0663734078,0.4716348946,-0.1141988561,-0.1069330126,0.3006836772,-0.0471390635,-0.0889480561,-0.1888510734,0.0094235633,-0.0697389618,0.247573629,0.5000484586,-0.1218189448,-0.0368878432,-0.1750988066,-0.0812497661,0.1250171214,0.0917121768,0.2032328099,0.1998405457,-0.2620665729,-0.2598944306,-0.0320798419,-0.1772939861,-0.2796158493,0.2072875649,0.0089059155,-0.0291699897,0.2409033477,0.0525551066,-0.1873457432,-0.0364574045,0.1764351875,-0.176746279,-0.066592209,0.5022160411,0.4060555696,-0.2406093776,-0.3803833723,0.1318946034,-0.0628116056,-0.2885809243,0.3547709584,0.1688399315,0.3007945418,0.0165438652,0.3438568413,0.0157201346,-0.0099177929,0.1101292968,-0.3858225644,-0.182272926,0.027388623,-0.342487514,0.0165286101,0.173354581,0.6295419335,0.1122876555,0.0537228286,-0.4009521902,-0.1003775001,-0.0929586515,0.2274139524,0.4283628762,-0.1619137526,-0.0227419026,0.0861172155,0.1760029942,0.0346549228,-0.2526120842,-0.314812988,-0.2940256298,0.0661309883,0.1720108539,-0.4291438162,0.3799619973,0.0724910498,-0.0009614042,-0.3068479002,-0.1781092137,0.0893365964,-0.1307528168,0.3698425293,0.1245556027,0.1172236055,0.175313279,0.0280542914,0.1229266375,0.1096993908,-0.0064216722,0.1405479908,0.0346085429,-0.0513197146,-0.196865961,0.0078801606,0.0852092505,0.2760110795,0.1000634953,-0.4387730658,-0.1408093125,-0.1907574236,0.331359297,0.4964479804,-0.2855861187,-0.0179454908,0.1170451567,0.2279493809,-0.431779623,0.0991013646,0.4069510996,0.0616520308,0.0558136776,0.1605772227,-0.196325466,-0.0263518207,-0.2005510032,-0.015522846,-0.0812871903,-0.2450969666,0.1379801035,0.2535305619,-0.1621550173,0.0723396018,0.2234296203,0.2248426974,0.2094131559,0.4411480725,-0.0880139023,0.0335321389,0.3215833306,-0.1977277696,0.0530192964,-0.4163442254,-0.2213221192,0.004601588,0.1555845737,0.0431443006,-0.0308006573,0.0894242376,0.0554312021,-0.0492895246,0.0911091119,0.1842016131,-0.1754979193,-0.0029799072,0.07021337,0.1035524011,-0.1834444851,-0.2007070631,-0.0523935743,0.0233831629,-0.0611661263,0.0867941976,0.1566937417,-0.5509833097,-0.0985001028,0.028403068,-0.3185489476,-0.4562836587,0.0733105093,0.4109583497,0.0954850167,-0.1203892529,-0.0074952352,0.6174410582,0.4333248734,-0.0764670596,-0.1840572208,-0.2809346318,-0.0467595868,-0.0322065093,0.0866955072,0.1216947138,0.2625560462,0.5520280004,0.1766430885,-0.1991811842,0.1269986182,0.2516283095,-0.2234730422,-0.1569376588,0.2527378798,-0.0195222571,-0.1268068105,-0.2825405598,0.1944848895,-0.2954165936,0.1875001341,0.4098545313,0.2683029473,0.1570728123,-0.1060613543,0.1378096938,0.1873126626,0.5516356826,0.1906979531,0.3860901296,-0.3571788371,-0.0554284528,-0.5711801052,-0.0245900769,-0.1053961366,-0.0316519812,-0.1443446428,-0.0839933902,0.0324577168,0.2593309879,0.1967318356,0.1654993743,0.1051031351,-0.0875484273,-0.2856193185,-0.1786556393,0.0377548598,-0.1408987641,0.1282964647,-0.2223509252,-0.0397654586,-0.4887033701,0.1265330911,-0.3787437677,-0.2208055407,0.0734933987,0.4357133806,0.2232686132,0.2265576571,0.5930403471,-0.0317760035,-0.0286308639,-0.3275698721,-0.2059123963,-0.1545397788,0.3455994725,0.0363045111,0.4889930487,-0.1071662307,-0.0158568323,-0.3778460324,0.1949944496,-0.3336418271,0.0198232438,-0.1027710661,0.1275236309,-0.1381614953,0.1817010939,0.0611189194,0.0021520669,-0.1441734731,0.2395447195,-0.3893561363,-0.4405485988,0.5200942159,-0.4261544347,-0.6753060222,-0.1937554777,0.0424238034,-0.2450479567,-0.1419898719,-0.4619947672,0.062090721,0.2820001543,-0.0690544322,-0.2712805569,0.3742442727,-0.1549424082,0.0212926641,0.0864809677,0.2581366599,0.0770811439,-0.2606575787,-0.2184218317,-0.2339115441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/156","title":"SyntaxError with WMT datasets","comments":"Usually, the idea is that a dataset can be added without releasing a new version. The problem in the case of `WMT` was that some \"core\" code of the library had to be changed as well. \r\n\r\n@thomwolf @lhoestq @julien-c - How should we go about this. If we add a dataset that also requires \"core\" code changes, how do we handle the versioning? The moment a dataset is on AWS it will actually be listed with `list_datasets()` in all earlier versions...\r\n\r\nIs there a way to somehow insert the `pip version` to the HfApi() and get only the datasets that were available for this version (at the date of the release of the version) @julien-c ? ","body":"The following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook","comment_length":116,"text":"SyntaxError with WMT datasets\nThe following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook\nUsually, the idea is that a dataset can be added without releasing a new version. The problem in the case of `WMT` was that some \"core\" code of the library had to be changed as well. \r\n\r\n@thomwolf @lhoestq @julien-c - How should we go about this. If we add a dataset that also requires \"core\" code changes, how do we handle the versioning? The moment a dataset is on AWS it will actually be listed with `list_datasets()` in all earlier versions...\r\n\r\nIs there a way to somehow insert the `pip version` to the HfApi() and get only the datasets that were available for this version (at the date of the release of the version) @julien-c ? ","embeddings":[-0.32973665,0.0324432179,-0.0193531662,0.0127459355,0.2164737433,-0.0149221374,0.3609470129,0.5264438987,0.1301548183,-0.1121621057,0.0932864323,0.4350022972,-0.3767533898,0.1414317936,0.1508708149,-0.1754975915,-0.0268027131,0.1413965672,-0.3586361706,0.151956737,-0.2946045697,0.1583642811,-0.420961529,0.2173392326,-0.2542883456,-0.0902357399,-0.021146521,0.144950211,-0.1394597441,-0.462849319,0.140802905,-0.0065226816,0.2389953732,0.3692950606,-0.0001028357,-0.0074911187,0.1617367566,-0.0390187725,-0.3923019171,-0.3421501517,-0.4517919123,-0.1481353492,0.143838644,-0.2942019701,0.0407955088,-0.1019728407,0.012090249,-0.2335641831,0.5227859616,0.470978111,0.2622492909,0.4046530128,-0.0788426101,-0.1995568573,0.0400845148,0.1868067235,-0.0251018368,0.2721501887,-0.0462756231,-0.3078270257,0.0443689711,0.2844180167,-0.0971550569,0.2046322972,0.2288134843,0.0462167524,0.1737399995,0.0146226026,-0.1336928159,0.111689046,0.3766534328,-0.3340502977,-0.1922972947,-0.0254198257,-0.0237156339,-0.3796640635,0.1737448573,0.2510075867,-0.2430952042,0.2042771727,-0.2803643048,0.0721920952,-0.0495335385,0.280293107,-0.0419157632,0.7289190888,-0.1515552253,0.0830518156,-0.0326680392,-0.1897672564,0.1182416454,-0.0233938098,-0.0012605379,0.1437297612,-0.0696197599,-0.0688927695,0.2810450196,-0.0273937248,-0.2340159565,0.0267737973,0.1816953868,-0.1459489912,0.1988346726,0.1966277063,0.283844322,-0.064487271,0.0399041548,-0.0328306854,0.2843933702,0.2534207404,-0.027510548,0.0475533232,-0.1733619273,-0.4689917266,-0.0253391545,0.0136443032,0.5950089097,0.1803969294,-0.2701009512,-0.0289219059,-0.2787000239,-0.302798748,-0.0676975995,0.2093273848,0.121819444,-0.1600497216,0.3483577371,0.0752085149,-0.2440336943,-0.1647191793,-0.2930479646,0.340895921,-0.3188335896,-0.1684278697,0.2188079953,0.2253614962,0.0922131017,0.0471641496,-0.0809147358,-0.1988425553,0.1389292777,-0.0077622063,0.1081167087,0.1051954702,-0.1785956919,0.093339853,0.2775603235,-0.1004941463,-0.0483997054,0.2007462978,-0.0364864543,-0.0897136256,-0.1614140123,0.288959831,-0.1801721752,-0.0983834192,0.0543797128,0.0768853873,0.2552024722,-0.0877907947,0.0312968753,-0.3928907812,-0.2555431128,-0.164903596,0.1820681095,0.4742359221,-0.4417522848,-0.1943070441,-0.1998421103,-0.1128745079,0.196072191,0.1654726416,-0.1848081648,0.6032976508,-0.1668522954,0.1821197867,0.4813289344,-0.2698017657,0.1708316952,0.0655022413,-0.2529610395,0.076276429,-0.0907965153,-0.045418676,-0.0019625076,-0.0203296375,0.3685547709,0.2570970356,-0.0868278965,-0.0704206899,0.025802223,0.028126996,0.593431592,0.0462836511,-0.0008243334,-0.2142312676,0.0578708202,0.4939417243,0.2460762858,-0.057706587,-0.0333488211,-0.0137065565,0.0170148499,-0.1148447394,-0.3352916241,-0.2411742806,-0.22320728,-0.1093647555,0.0478304662,0.3630459607,-0.1524047107,0.2927405536,-0.2438348085,0.0848255977,-0.2252929807,-0.2423848957,0.2631049156,0.0465616249,-0.2522629201,-0.1086916551,-0.0686345994,-0.0800719112,0.0088367239,0.0938628986,-0.1322340667,0.1556389332,-0.2579882741,-0.3206021488,0.1305004209,0.256996721,0.1654368639,-0.0436956733,-0.2140459865,0.2780266702,0.1649134308,-0.0692972168,-0.2854145169,-0.0294151921,-0.0743030012,-0.160280183,0.1651479155,0.179478243,0.1970850676,-0.1973392218,-0.2075010985,0.1806471944,-0.023483254,0.0752795711,0.1489133984,0.1009815335,0.1457805783,0.0210331753,0.0285009481,-0.2971132398,0.1780157238,-0.0241094306,-0.1505931169,-0.0727762729,-0.1188315302,-0.1056578383,0.5293703079,0.2276985347,0.2241090834,-0.2177880704,-0.262104243,-0.081199497,0.0782550648,0.070517458,0.4799798429,0.1782226712,-0.0041789506,0.1772151589,-0.0156665258,-0.2165574729,0.216871202,0.1818329394,0.0331380144,-0.0351118222,0.0385485813,0.085117124,-0.1735186428,-0.2818623483,-0.0842829421,0.1642438173,-0.3306981623,0.0497132167,-0.4721483588,-0.3931072354,-0.4514593482,-0.3841335177,-0.1739959866,-0.4288794994,-0.1867683232,-0.0351117514,-0.2173171639,0.3002482057,-0.1404795647,-0.0482548997,0.1101874709,0.0663792789,0.098848708,-0.2670576572,-0.1530420929,0.1411233395,0.3205446899,0.1450353414,0.3701429069,-0.0794283301,0.0111533152,0.0473328829,-0.3408435583,0.1160992384,-0.1009821147,0.2626492679,0.4324879348,0.3414855003,0.0838382915,-0.2842497528,0.4270731211,-0.0637356862,0.0510102436,0.049875319,0.1940965652,-0.14796938,-0.3029948175,-0.4622400105,-0.3657283187,-0.4703692198,-0.12164215,0.1440427452,0.131994009,0.4903213084,-0.0182412937,0.3564141393,-0.0871031806,-0.0005423531,0.07355012,0.0614891574,0.2821263671,-0.312462002,-0.1920034438,0.0803076327,0.03580717,0.2169488519,-0.1713449955,-0.2317764312,0.1126878113,-0.105670698,0.1334697753,0.0481273234,0.101513952,0.1759227514,0.2410977781,-0.2015783489,-0.3006727397,-0.3467811942,0.0994992778,0.1069888249,0.125855729,-0.1017532796,0.1422888935,0.1000687107,0.3848873377,0.0874497145,-0.1828735769,0.3111949563,0.0411565527,0.3453080058,-0.0618227795,-0.3201507926,0.2449907064,-0.0587742515,0.1667978317,0.2006136924,-0.1382380724,0.0803111568,-0.2032093555,-0.0921038389,-0.0643962622,-0.418659091,0.111884892,0.0074180295,0.1880878508,0.2058223784,0.0365455262,-0.194925338,0.0036354265,-0.0097371815,0.2998480797,0.0105650686,0.1356204152,-0.1349021494,-0.0225577783,-0.3069858551,0.1838999689,0.2323766798,0.4915707707,-0.0271859914,-0.1784804314,-0.2188410014,0.0941773877,0.1146481782,-0.13565588,0.1829944998,0.4571828842,0.0663695559,-0.2134140432,-0.1721869856,-0.5033162832,0.1326623857,0.4361288249,0.0677459612,-0.0008687158,-0.3626334071,0.4444147646,0.1816581339,-0.1891018599,-0.2773975432,-0.0000782053,-0.0716028959,-0.336502254,0.0441519283,0.0950581655,0.1696865559,0.2260651141,0.0341112204,0.1553541869,0.0793138146,0.1081611663,0.1017457172,-0.0415175855,0.1275674403,0.0413583294,-0.101653263,0.1490106136,0.2546506524,0.3937421143,-0.1238749921,-0.1940542758,-0.164419353,0.0299259908,0.3551030457,0.0970911607,-0.0819731206,0.1334188133,0.0647364482,0.1381150633,-0.2426775098,0.2743785083,0.2958749831,0.170900777,-0.3747496903,-0.3495922685,-0.0297986846,-0.0531896651,0.1575489193,0.1404528022,-0.0618960597,-0.3537502587,-0.0751761198,-0.0094296345,0.5648851395,0.1630987078,0.289714843,0.4404497743,-0.0870268047,0.473436296,0.1777279079,0.0051305699,-0.246722281,0.1247948483,0.0227170885,-0.1888537854,0.2077721357,0.1377238184,-0.2474355549,0.2713086009,-0.0384190902,0.2527678311,0.065176256,0.090902254,0.3361058533,-0.1986477971,-0.4095683694,0.2058805674,-0.4133999646,0.0716988146,-0.2496188134,-0.3231573999,0.159161225,-0.2507033348,-0.2010598481,0.1776943803,-0.2587882876,0.2284641117,0.0260749403,-0.0663734078,0.4716348946,-0.1141988561,-0.1069330126,0.3006836772,-0.0471390635,-0.0889480561,-0.1888510734,0.0094235633,-0.0697389618,0.247573629,0.5000484586,-0.1218189448,-0.0368878432,-0.1750988066,-0.0812497661,0.1250171214,0.0917121768,0.2032328099,0.1998405457,-0.2620665729,-0.2598944306,-0.0320798419,-0.1772939861,-0.2796158493,0.2072875649,0.0089059155,-0.0291699897,0.2409033477,0.0525551066,-0.1873457432,-0.0364574045,0.1764351875,-0.176746279,-0.066592209,0.5022160411,0.4060555696,-0.2406093776,-0.3803833723,0.1318946034,-0.0628116056,-0.2885809243,0.3547709584,0.1688399315,0.3007945418,0.0165438652,0.3438568413,0.0157201346,-0.0099177929,0.1101292968,-0.3858225644,-0.182272926,0.027388623,-0.342487514,0.0165286101,0.173354581,0.6295419335,0.1122876555,0.0537228286,-0.4009521902,-0.1003775001,-0.0929586515,0.2274139524,0.4283628762,-0.1619137526,-0.0227419026,0.0861172155,0.1760029942,0.0346549228,-0.2526120842,-0.314812988,-0.2940256298,0.0661309883,0.1720108539,-0.4291438162,0.3799619973,0.0724910498,-0.0009614042,-0.3068479002,-0.1781092137,0.0893365964,-0.1307528168,0.3698425293,0.1245556027,0.1172236055,0.175313279,0.0280542914,0.1229266375,0.1096993908,-0.0064216722,0.1405479908,0.0346085429,-0.0513197146,-0.196865961,0.0078801606,0.0852092505,0.2760110795,0.1000634953,-0.4387730658,-0.1408093125,-0.1907574236,0.331359297,0.4964479804,-0.2855861187,-0.0179454908,0.1170451567,0.2279493809,-0.431779623,0.0991013646,0.4069510996,0.0616520308,0.0558136776,0.1605772227,-0.196325466,-0.0263518207,-0.2005510032,-0.015522846,-0.0812871903,-0.2450969666,0.1379801035,0.2535305619,-0.1621550173,0.0723396018,0.2234296203,0.2248426974,0.2094131559,0.4411480725,-0.0880139023,0.0335321389,0.3215833306,-0.1977277696,0.0530192964,-0.4163442254,-0.2213221192,0.004601588,0.1555845737,0.0431443006,-0.0308006573,0.0894242376,0.0554312021,-0.0492895246,0.0911091119,0.1842016131,-0.1754979193,-0.0029799072,0.07021337,0.1035524011,-0.1834444851,-0.2007070631,-0.0523935743,0.0233831629,-0.0611661263,0.0867941976,0.1566937417,-0.5509833097,-0.0985001028,0.028403068,-0.3185489476,-0.4562836587,0.0733105093,0.4109583497,0.0954850167,-0.1203892529,-0.0074952352,0.6174410582,0.4333248734,-0.0764670596,-0.1840572208,-0.2809346318,-0.0467595868,-0.0322065093,0.0866955072,0.1216947138,0.2625560462,0.5520280004,0.1766430885,-0.1991811842,0.1269986182,0.2516283095,-0.2234730422,-0.1569376588,0.2527378798,-0.0195222571,-0.1268068105,-0.2825405598,0.1944848895,-0.2954165936,0.1875001341,0.4098545313,0.2683029473,0.1570728123,-0.1060613543,0.1378096938,0.1873126626,0.5516356826,0.1906979531,0.3860901296,-0.3571788371,-0.0554284528,-0.5711801052,-0.0245900769,-0.1053961366,-0.0316519812,-0.1443446428,-0.0839933902,0.0324577168,0.2593309879,0.1967318356,0.1654993743,0.1051031351,-0.0875484273,-0.2856193185,-0.1786556393,0.0377548598,-0.1408987641,0.1282964647,-0.2223509252,-0.0397654586,-0.4887033701,0.1265330911,-0.3787437677,-0.2208055407,0.0734933987,0.4357133806,0.2232686132,0.2265576571,0.5930403471,-0.0317760035,-0.0286308639,-0.3275698721,-0.2059123963,-0.1545397788,0.3455994725,0.0363045111,0.4889930487,-0.1071662307,-0.0158568323,-0.3778460324,0.1949944496,-0.3336418271,0.0198232438,-0.1027710661,0.1275236309,-0.1381614953,0.1817010939,0.0611189194,0.0021520669,-0.1441734731,0.2395447195,-0.3893561363,-0.4405485988,0.5200942159,-0.4261544347,-0.6753060222,-0.1937554777,0.0424238034,-0.2450479567,-0.1419898719,-0.4619947672,0.062090721,0.2820001543,-0.0690544322,-0.2712805569,0.3742442727,-0.1549424082,0.0212926641,0.0864809677,0.2581366599,0.0770811439,-0.2606575787,-0.2184218317,-0.2339115441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/156","title":"SyntaxError with WMT datasets","comments":"We plan to have something like a `requirements.txt` per dataset to prevent user from loading dataset with old version of `nlp` or any other libraries. Right now the solution is just to keep `nlp` up to date when you want to load a dataset that leverages the latests features of `nlp`.\r\n\r\nFor datasets that are on AWS but that use features that are not released yet we should be able to filter those from the `list_dataset` as soon as we have the `requirements.txt` feature on (filter datasets that need a future version of `nlp`).\r\n\r\nShall we rename this issue to be more explicit about the problem ?\r\nSomething like `Specify the minimum version of the nlp library required for each dataset` ?","body":"The following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook","comment_length":122,"text":"SyntaxError with WMT datasets\nThe following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook\nWe plan to have something like a `requirements.txt` per dataset to prevent user from loading dataset with old version of `nlp` or any other libraries. Right now the solution is just to keep `nlp` up to date when you want to load a dataset that leverages the latests features of `nlp`.\r\n\r\nFor datasets that are on AWS but that use features that are not released yet we should be able to filter those from the `list_dataset` as soon as we have the `requirements.txt` feature on (filter datasets that need a future version of `nlp`).\r\n\r\nShall we rename this issue to be more explicit about the problem ?\r\nSomething like `Specify the minimum version of the nlp library required for each dataset` ?","embeddings":[-0.32973665,0.0324432179,-0.0193531662,0.0127459355,0.2164737433,-0.0149221374,0.3609470129,0.5264438987,0.1301548183,-0.1121621057,0.0932864323,0.4350022972,-0.3767533898,0.1414317936,0.1508708149,-0.1754975915,-0.0268027131,0.1413965672,-0.3586361706,0.151956737,-0.2946045697,0.1583642811,-0.420961529,0.2173392326,-0.2542883456,-0.0902357399,-0.021146521,0.144950211,-0.1394597441,-0.462849319,0.140802905,-0.0065226816,0.2389953732,0.3692950606,-0.0001028357,-0.0074911187,0.1617367566,-0.0390187725,-0.3923019171,-0.3421501517,-0.4517919123,-0.1481353492,0.143838644,-0.2942019701,0.0407955088,-0.1019728407,0.012090249,-0.2335641831,0.5227859616,0.470978111,0.2622492909,0.4046530128,-0.0788426101,-0.1995568573,0.0400845148,0.1868067235,-0.0251018368,0.2721501887,-0.0462756231,-0.3078270257,0.0443689711,0.2844180167,-0.0971550569,0.2046322972,0.2288134843,0.0462167524,0.1737399995,0.0146226026,-0.1336928159,0.111689046,0.3766534328,-0.3340502977,-0.1922972947,-0.0254198257,-0.0237156339,-0.3796640635,0.1737448573,0.2510075867,-0.2430952042,0.2042771727,-0.2803643048,0.0721920952,-0.0495335385,0.280293107,-0.0419157632,0.7289190888,-0.1515552253,0.0830518156,-0.0326680392,-0.1897672564,0.1182416454,-0.0233938098,-0.0012605379,0.1437297612,-0.0696197599,-0.0688927695,0.2810450196,-0.0273937248,-0.2340159565,0.0267737973,0.1816953868,-0.1459489912,0.1988346726,0.1966277063,0.283844322,-0.064487271,0.0399041548,-0.0328306854,0.2843933702,0.2534207404,-0.027510548,0.0475533232,-0.1733619273,-0.4689917266,-0.0253391545,0.0136443032,0.5950089097,0.1803969294,-0.2701009512,-0.0289219059,-0.2787000239,-0.302798748,-0.0676975995,0.2093273848,0.121819444,-0.1600497216,0.3483577371,0.0752085149,-0.2440336943,-0.1647191793,-0.2930479646,0.340895921,-0.3188335896,-0.1684278697,0.2188079953,0.2253614962,0.0922131017,0.0471641496,-0.0809147358,-0.1988425553,0.1389292777,-0.0077622063,0.1081167087,0.1051954702,-0.1785956919,0.093339853,0.2775603235,-0.1004941463,-0.0483997054,0.2007462978,-0.0364864543,-0.0897136256,-0.1614140123,0.288959831,-0.1801721752,-0.0983834192,0.0543797128,0.0768853873,0.2552024722,-0.0877907947,0.0312968753,-0.3928907812,-0.2555431128,-0.164903596,0.1820681095,0.4742359221,-0.4417522848,-0.1943070441,-0.1998421103,-0.1128745079,0.196072191,0.1654726416,-0.1848081648,0.6032976508,-0.1668522954,0.1821197867,0.4813289344,-0.2698017657,0.1708316952,0.0655022413,-0.2529610395,0.076276429,-0.0907965153,-0.045418676,-0.0019625076,-0.0203296375,0.3685547709,0.2570970356,-0.0868278965,-0.0704206899,0.025802223,0.028126996,0.593431592,0.0462836511,-0.0008243334,-0.2142312676,0.0578708202,0.4939417243,0.2460762858,-0.057706587,-0.0333488211,-0.0137065565,0.0170148499,-0.1148447394,-0.3352916241,-0.2411742806,-0.22320728,-0.1093647555,0.0478304662,0.3630459607,-0.1524047107,0.2927405536,-0.2438348085,0.0848255977,-0.2252929807,-0.2423848957,0.2631049156,0.0465616249,-0.2522629201,-0.1086916551,-0.0686345994,-0.0800719112,0.0088367239,0.0938628986,-0.1322340667,0.1556389332,-0.2579882741,-0.3206021488,0.1305004209,0.256996721,0.1654368639,-0.0436956733,-0.2140459865,0.2780266702,0.1649134308,-0.0692972168,-0.2854145169,-0.0294151921,-0.0743030012,-0.160280183,0.1651479155,0.179478243,0.1970850676,-0.1973392218,-0.2075010985,0.1806471944,-0.023483254,0.0752795711,0.1489133984,0.1009815335,0.1457805783,0.0210331753,0.0285009481,-0.2971132398,0.1780157238,-0.0241094306,-0.1505931169,-0.0727762729,-0.1188315302,-0.1056578383,0.5293703079,0.2276985347,0.2241090834,-0.2177880704,-0.262104243,-0.081199497,0.0782550648,0.070517458,0.4799798429,0.1782226712,-0.0041789506,0.1772151589,-0.0156665258,-0.2165574729,0.216871202,0.1818329394,0.0331380144,-0.0351118222,0.0385485813,0.085117124,-0.1735186428,-0.2818623483,-0.0842829421,0.1642438173,-0.3306981623,0.0497132167,-0.4721483588,-0.3931072354,-0.4514593482,-0.3841335177,-0.1739959866,-0.4288794994,-0.1867683232,-0.0351117514,-0.2173171639,0.3002482057,-0.1404795647,-0.0482548997,0.1101874709,0.0663792789,0.098848708,-0.2670576572,-0.1530420929,0.1411233395,0.3205446899,0.1450353414,0.3701429069,-0.0794283301,0.0111533152,0.0473328829,-0.3408435583,0.1160992384,-0.1009821147,0.2626492679,0.4324879348,0.3414855003,0.0838382915,-0.2842497528,0.4270731211,-0.0637356862,0.0510102436,0.049875319,0.1940965652,-0.14796938,-0.3029948175,-0.4622400105,-0.3657283187,-0.4703692198,-0.12164215,0.1440427452,0.131994009,0.4903213084,-0.0182412937,0.3564141393,-0.0871031806,-0.0005423531,0.07355012,0.0614891574,0.2821263671,-0.312462002,-0.1920034438,0.0803076327,0.03580717,0.2169488519,-0.1713449955,-0.2317764312,0.1126878113,-0.105670698,0.1334697753,0.0481273234,0.101513952,0.1759227514,0.2410977781,-0.2015783489,-0.3006727397,-0.3467811942,0.0994992778,0.1069888249,0.125855729,-0.1017532796,0.1422888935,0.1000687107,0.3848873377,0.0874497145,-0.1828735769,0.3111949563,0.0411565527,0.3453080058,-0.0618227795,-0.3201507926,0.2449907064,-0.0587742515,0.1667978317,0.2006136924,-0.1382380724,0.0803111568,-0.2032093555,-0.0921038389,-0.0643962622,-0.418659091,0.111884892,0.0074180295,0.1880878508,0.2058223784,0.0365455262,-0.194925338,0.0036354265,-0.0097371815,0.2998480797,0.0105650686,0.1356204152,-0.1349021494,-0.0225577783,-0.3069858551,0.1838999689,0.2323766798,0.4915707707,-0.0271859914,-0.1784804314,-0.2188410014,0.0941773877,0.1146481782,-0.13565588,0.1829944998,0.4571828842,0.0663695559,-0.2134140432,-0.1721869856,-0.5033162832,0.1326623857,0.4361288249,0.0677459612,-0.0008687158,-0.3626334071,0.4444147646,0.1816581339,-0.1891018599,-0.2773975432,-0.0000782053,-0.0716028959,-0.336502254,0.0441519283,0.0950581655,0.1696865559,0.2260651141,0.0341112204,0.1553541869,0.0793138146,0.1081611663,0.1017457172,-0.0415175855,0.1275674403,0.0413583294,-0.101653263,0.1490106136,0.2546506524,0.3937421143,-0.1238749921,-0.1940542758,-0.164419353,0.0299259908,0.3551030457,0.0970911607,-0.0819731206,0.1334188133,0.0647364482,0.1381150633,-0.2426775098,0.2743785083,0.2958749831,0.170900777,-0.3747496903,-0.3495922685,-0.0297986846,-0.0531896651,0.1575489193,0.1404528022,-0.0618960597,-0.3537502587,-0.0751761198,-0.0094296345,0.5648851395,0.1630987078,0.289714843,0.4404497743,-0.0870268047,0.473436296,0.1777279079,0.0051305699,-0.246722281,0.1247948483,0.0227170885,-0.1888537854,0.2077721357,0.1377238184,-0.2474355549,0.2713086009,-0.0384190902,0.2527678311,0.065176256,0.090902254,0.3361058533,-0.1986477971,-0.4095683694,0.2058805674,-0.4133999646,0.0716988146,-0.2496188134,-0.3231573999,0.159161225,-0.2507033348,-0.2010598481,0.1776943803,-0.2587882876,0.2284641117,0.0260749403,-0.0663734078,0.4716348946,-0.1141988561,-0.1069330126,0.3006836772,-0.0471390635,-0.0889480561,-0.1888510734,0.0094235633,-0.0697389618,0.247573629,0.5000484586,-0.1218189448,-0.0368878432,-0.1750988066,-0.0812497661,0.1250171214,0.0917121768,0.2032328099,0.1998405457,-0.2620665729,-0.2598944306,-0.0320798419,-0.1772939861,-0.2796158493,0.2072875649,0.0089059155,-0.0291699897,0.2409033477,0.0525551066,-0.1873457432,-0.0364574045,0.1764351875,-0.176746279,-0.066592209,0.5022160411,0.4060555696,-0.2406093776,-0.3803833723,0.1318946034,-0.0628116056,-0.2885809243,0.3547709584,0.1688399315,0.3007945418,0.0165438652,0.3438568413,0.0157201346,-0.0099177929,0.1101292968,-0.3858225644,-0.182272926,0.027388623,-0.342487514,0.0165286101,0.173354581,0.6295419335,0.1122876555,0.0537228286,-0.4009521902,-0.1003775001,-0.0929586515,0.2274139524,0.4283628762,-0.1619137526,-0.0227419026,0.0861172155,0.1760029942,0.0346549228,-0.2526120842,-0.314812988,-0.2940256298,0.0661309883,0.1720108539,-0.4291438162,0.3799619973,0.0724910498,-0.0009614042,-0.3068479002,-0.1781092137,0.0893365964,-0.1307528168,0.3698425293,0.1245556027,0.1172236055,0.175313279,0.0280542914,0.1229266375,0.1096993908,-0.0064216722,0.1405479908,0.0346085429,-0.0513197146,-0.196865961,0.0078801606,0.0852092505,0.2760110795,0.1000634953,-0.4387730658,-0.1408093125,-0.1907574236,0.331359297,0.4964479804,-0.2855861187,-0.0179454908,0.1170451567,0.2279493809,-0.431779623,0.0991013646,0.4069510996,0.0616520308,0.0558136776,0.1605772227,-0.196325466,-0.0263518207,-0.2005510032,-0.015522846,-0.0812871903,-0.2450969666,0.1379801035,0.2535305619,-0.1621550173,0.0723396018,0.2234296203,0.2248426974,0.2094131559,0.4411480725,-0.0880139023,0.0335321389,0.3215833306,-0.1977277696,0.0530192964,-0.4163442254,-0.2213221192,0.004601588,0.1555845737,0.0431443006,-0.0308006573,0.0894242376,0.0554312021,-0.0492895246,0.0911091119,0.1842016131,-0.1754979193,-0.0029799072,0.07021337,0.1035524011,-0.1834444851,-0.2007070631,-0.0523935743,0.0233831629,-0.0611661263,0.0867941976,0.1566937417,-0.5509833097,-0.0985001028,0.028403068,-0.3185489476,-0.4562836587,0.0733105093,0.4109583497,0.0954850167,-0.1203892529,-0.0074952352,0.6174410582,0.4333248734,-0.0764670596,-0.1840572208,-0.2809346318,-0.0467595868,-0.0322065093,0.0866955072,0.1216947138,0.2625560462,0.5520280004,0.1766430885,-0.1991811842,0.1269986182,0.2516283095,-0.2234730422,-0.1569376588,0.2527378798,-0.0195222571,-0.1268068105,-0.2825405598,0.1944848895,-0.2954165936,0.1875001341,0.4098545313,0.2683029473,0.1570728123,-0.1060613543,0.1378096938,0.1873126626,0.5516356826,0.1906979531,0.3860901296,-0.3571788371,-0.0554284528,-0.5711801052,-0.0245900769,-0.1053961366,-0.0316519812,-0.1443446428,-0.0839933902,0.0324577168,0.2593309879,0.1967318356,0.1654993743,0.1051031351,-0.0875484273,-0.2856193185,-0.1786556393,0.0377548598,-0.1408987641,0.1282964647,-0.2223509252,-0.0397654586,-0.4887033701,0.1265330911,-0.3787437677,-0.2208055407,0.0734933987,0.4357133806,0.2232686132,0.2265576571,0.5930403471,-0.0317760035,-0.0286308639,-0.3275698721,-0.2059123963,-0.1545397788,0.3455994725,0.0363045111,0.4889930487,-0.1071662307,-0.0158568323,-0.3778460324,0.1949944496,-0.3336418271,0.0198232438,-0.1027710661,0.1275236309,-0.1381614953,0.1817010939,0.0611189194,0.0021520669,-0.1441734731,0.2395447195,-0.3893561363,-0.4405485988,0.5200942159,-0.4261544347,-0.6753060222,-0.1937554777,0.0424238034,-0.2450479567,-0.1419898719,-0.4619947672,0.062090721,0.2820001543,-0.0690544322,-0.2712805569,0.3742442727,-0.1549424082,0.0212926641,0.0864809677,0.2581366599,0.0770811439,-0.2606575787,-0.2184218317,-0.2339115441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/156","title":"SyntaxError with WMT datasets","comments":"Closing this one.\r\nFeel free to re-open if you have other questions :)","body":"The following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook","comment_length":13,"text":"SyntaxError with WMT datasets\nThe following snippet produces a syntax error:\r\n\r\n```\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('wmt14')\r\nprint(dataset['train'][0])\r\n```\r\n\r\n```\r\nTraceback (most recent call last):\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/IPython\/core\/interactiveshell.py\", line 3326, in run_code\r\n exec(code_obj, self.user_global_ns, self.user_ns)\r\n\r\n File \"<ipython-input-8-3206959998b9>\", line 3, in <module>\r\n dataset = nlp.load_dataset('wmt14')\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 505, in load_dataset\r\n builder_cls = import_main_class(module_path, dataset=True)\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/load.py\", line 56, in import_main_class\r\n module = importlib.import_module(module_path)\r\n\r\n File \"\/usr\/lib\/python3.6\/importlib\/__init__.py\", line 126, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n\r\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\r\n\r\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\r\n\r\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\r\n\r\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\r\n\r\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt14.py\", line 21, in <module>\r\n from .wmt_utils import Wmt, WmtConfig\r\n\r\n File \"\/home\/tom\/.local\/lib\/python3.6\/site-packages\/nlp\/datasets\/wmt14\/c258d646f4f5870b0245f783b7aa0af85c7117e06aacf1e0340bd81935094de2\/wmt_utils.py\", line 659\r\n <<<<<<< HEAD\r\n ^\r\nSyntaxError: invalid syntax\r\n```\r\n\r\nPython version:\r\n`3.6.9 (default, Apr 18 2020, 01:56:04) [GCC 8.4.0]`\r\nRunning on Ubuntu 18.04, via a Jupyter notebook\nClosing this one.\r\nFeel free to re-open if you have other questions :)","embeddings":[-0.32973665,0.0324432179,-0.0193531662,0.0127459355,0.2164737433,-0.0149221374,0.3609470129,0.5264438987,0.1301548183,-0.1121621057,0.0932864323,0.4350022972,-0.3767533898,0.1414317936,0.1508708149,-0.1754975915,-0.0268027131,0.1413965672,-0.3586361706,0.151956737,-0.2946045697,0.1583642811,-0.420961529,0.2173392326,-0.2542883456,-0.0902357399,-0.021146521,0.144950211,-0.1394597441,-0.462849319,0.140802905,-0.0065226816,0.2389953732,0.3692950606,-0.0001028357,-0.0074911187,0.1617367566,-0.0390187725,-0.3923019171,-0.3421501517,-0.4517919123,-0.1481353492,0.143838644,-0.2942019701,0.0407955088,-0.1019728407,0.012090249,-0.2335641831,0.5227859616,0.470978111,0.2622492909,0.4046530128,-0.0788426101,-0.1995568573,0.0400845148,0.1868067235,-0.0251018368,0.2721501887,-0.0462756231,-0.3078270257,0.0443689711,0.2844180167,-0.0971550569,0.2046322972,0.2288134843,0.0462167524,0.1737399995,0.0146226026,-0.1336928159,0.111689046,0.3766534328,-0.3340502977,-0.1922972947,-0.0254198257,-0.0237156339,-0.3796640635,0.1737448573,0.2510075867,-0.2430952042,0.2042771727,-0.2803643048,0.0721920952,-0.0495335385,0.280293107,-0.0419157632,0.7289190888,-0.1515552253,0.0830518156,-0.0326680392,-0.1897672564,0.1182416454,-0.0233938098,-0.0012605379,0.1437297612,-0.0696197599,-0.0688927695,0.2810450196,-0.0273937248,-0.2340159565,0.0267737973,0.1816953868,-0.1459489912,0.1988346726,0.1966277063,0.283844322,-0.064487271,0.0399041548,-0.0328306854,0.2843933702,0.2534207404,-0.027510548,0.0475533232,-0.1733619273,-0.4689917266,-0.0253391545,0.0136443032,0.5950089097,0.1803969294,-0.2701009512,-0.0289219059,-0.2787000239,-0.302798748,-0.0676975995,0.2093273848,0.121819444,-0.1600497216,0.3483577371,0.0752085149,-0.2440336943,-0.1647191793,-0.2930479646,0.340895921,-0.3188335896,-0.1684278697,0.2188079953,0.2253614962,0.0922131017,0.0471641496,-0.0809147358,-0.1988425553,0.1389292777,-0.0077622063,0.1081167087,0.1051954702,-0.1785956919,0.093339853,0.2775603235,-0.1004941463,-0.0483997054,0.2007462978,-0.0364864543,-0.0897136256,-0.1614140123,0.288959831,-0.1801721752,-0.0983834192,0.0543797128,0.0768853873,0.2552024722,-0.0877907947,0.0312968753,-0.3928907812,-0.2555431128,-0.164903596,0.1820681095,0.4742359221,-0.4417522848,-0.1943070441,-0.1998421103,-0.1128745079,0.196072191,0.1654726416,-0.1848081648,0.6032976508,-0.1668522954,0.1821197867,0.4813289344,-0.2698017657,0.1708316952,0.0655022413,-0.2529610395,0.076276429,-0.0907965153,-0.045418676,-0.0019625076,-0.0203296375,0.3685547709,0.2570970356,-0.0868278965,-0.0704206899,0.025802223,0.028126996,0.593431592,0.0462836511,-0.0008243334,-0.2142312676,0.0578708202,0.4939417243,0.2460762858,-0.057706587,-0.0333488211,-0.0137065565,0.0170148499,-0.1148447394,-0.3352916241,-0.2411742806,-0.22320728,-0.1093647555,0.0478304662,0.3630459607,-0.1524047107,0.2927405536,-0.2438348085,0.0848255977,-0.2252929807,-0.2423848957,0.2631049156,0.0465616249,-0.2522629201,-0.1086916551,-0.0686345994,-0.0800719112,0.0088367239,0.0938628986,-0.1322340667,0.1556389332,-0.2579882741,-0.3206021488,0.1305004209,0.256996721,0.1654368639,-0.0436956733,-0.2140459865,0.2780266702,0.1649134308,-0.0692972168,-0.2854145169,-0.0294151921,-0.0743030012,-0.160280183,0.1651479155,0.179478243,0.1970850676,-0.1973392218,-0.2075010985,0.1806471944,-0.023483254,0.0752795711,0.1489133984,0.1009815335,0.1457805783,0.0210331753,0.0285009481,-0.2971132398,0.1780157238,-0.0241094306,-0.1505931169,-0.0727762729,-0.1188315302,-0.1056578383,0.5293703079,0.2276985347,0.2241090834,-0.2177880704,-0.262104243,-0.081199497,0.0782550648,0.070517458,0.4799798429,0.1782226712,-0.0041789506,0.1772151589,-0.0156665258,-0.2165574729,0.216871202,0.1818329394,0.0331380144,-0.0351118222,0.0385485813,0.085117124,-0.1735186428,-0.2818623483,-0.0842829421,0.1642438173,-0.3306981623,0.0497132167,-0.4721483588,-0.3931072354,-0.4514593482,-0.3841335177,-0.1739959866,-0.4288794994,-0.1867683232,-0.0351117514,-0.2173171639,0.3002482057,-0.1404795647,-0.0482548997,0.1101874709,0.0663792789,0.098848708,-0.2670576572,-0.1530420929,0.1411233395,0.3205446899,0.1450353414,0.3701429069,-0.0794283301,0.0111533152,0.0473328829,-0.3408435583,0.1160992384,-0.1009821147,0.2626492679,0.4324879348,0.3414855003,0.0838382915,-0.2842497528,0.4270731211,-0.0637356862,0.0510102436,0.049875319,0.1940965652,-0.14796938,-0.3029948175,-0.4622400105,-0.3657283187,-0.4703692198,-0.12164215,0.1440427452,0.131994009,0.4903213084,-0.0182412937,0.3564141393,-0.0871031806,-0.0005423531,0.07355012,0.0614891574,0.2821263671,-0.312462002,-0.1920034438,0.0803076327,0.03580717,0.2169488519,-0.1713449955,-0.2317764312,0.1126878113,-0.105670698,0.1334697753,0.0481273234,0.101513952,0.1759227514,0.2410977781,-0.2015783489,-0.3006727397,-0.3467811942,0.0994992778,0.1069888249,0.125855729,-0.1017532796,0.1422888935,0.1000687107,0.3848873377,0.0874497145,-0.1828735769,0.3111949563,0.0411565527,0.3453080058,-0.0618227795,-0.3201507926,0.2449907064,-0.0587742515,0.1667978317,0.2006136924,-0.1382380724,0.0803111568,-0.2032093555,-0.0921038389,-0.0643962622,-0.418659091,0.111884892,0.0074180295,0.1880878508,0.2058223784,0.0365455262,-0.194925338,0.0036354265,-0.0097371815,0.2998480797,0.0105650686,0.1356204152,-0.1349021494,-0.0225577783,-0.3069858551,0.1838999689,0.2323766798,0.4915707707,-0.0271859914,-0.1784804314,-0.2188410014,0.0941773877,0.1146481782,-0.13565588,0.1829944998,0.4571828842,0.0663695559,-0.2134140432,-0.1721869856,-0.5033162832,0.1326623857,0.4361288249,0.0677459612,-0.0008687158,-0.3626334071,0.4444147646,0.1816581339,-0.1891018599,-0.2773975432,-0.0000782053,-0.0716028959,-0.336502254,0.0441519283,0.0950581655,0.1696865559,0.2260651141,0.0341112204,0.1553541869,0.0793138146,0.1081611663,0.1017457172,-0.0415175855,0.1275674403,0.0413583294,-0.101653263,0.1490106136,0.2546506524,0.3937421143,-0.1238749921,-0.1940542758,-0.164419353,0.0299259908,0.3551030457,0.0970911607,-0.0819731206,0.1334188133,0.0647364482,0.1381150633,-0.2426775098,0.2743785083,0.2958749831,0.170900777,-0.3747496903,-0.3495922685,-0.0297986846,-0.0531896651,0.1575489193,0.1404528022,-0.0618960597,-0.3537502587,-0.0751761198,-0.0094296345,0.5648851395,0.1630987078,0.289714843,0.4404497743,-0.0870268047,0.473436296,0.1777279079,0.0051305699,-0.246722281,0.1247948483,0.0227170885,-0.1888537854,0.2077721357,0.1377238184,-0.2474355549,0.2713086009,-0.0384190902,0.2527678311,0.065176256,0.090902254,0.3361058533,-0.1986477971,-0.4095683694,0.2058805674,-0.4133999646,0.0716988146,-0.2496188134,-0.3231573999,0.159161225,-0.2507033348,-0.2010598481,0.1776943803,-0.2587882876,0.2284641117,0.0260749403,-0.0663734078,0.4716348946,-0.1141988561,-0.1069330126,0.3006836772,-0.0471390635,-0.0889480561,-0.1888510734,0.0094235633,-0.0697389618,0.247573629,0.5000484586,-0.1218189448,-0.0368878432,-0.1750988066,-0.0812497661,0.1250171214,0.0917121768,0.2032328099,0.1998405457,-0.2620665729,-0.2598944306,-0.0320798419,-0.1772939861,-0.2796158493,0.2072875649,0.0089059155,-0.0291699897,0.2409033477,0.0525551066,-0.1873457432,-0.0364574045,0.1764351875,-0.176746279,-0.066592209,0.5022160411,0.4060555696,-0.2406093776,-0.3803833723,0.1318946034,-0.0628116056,-0.2885809243,0.3547709584,0.1688399315,0.3007945418,0.0165438652,0.3438568413,0.0157201346,-0.0099177929,0.1101292968,-0.3858225644,-0.182272926,0.027388623,-0.342487514,0.0165286101,0.173354581,0.6295419335,0.1122876555,0.0537228286,-0.4009521902,-0.1003775001,-0.0929586515,0.2274139524,0.4283628762,-0.1619137526,-0.0227419026,0.0861172155,0.1760029942,0.0346549228,-0.2526120842,-0.314812988,-0.2940256298,0.0661309883,0.1720108539,-0.4291438162,0.3799619973,0.0724910498,-0.0009614042,-0.3068479002,-0.1781092137,0.0893365964,-0.1307528168,0.3698425293,0.1245556027,0.1172236055,0.175313279,0.0280542914,0.1229266375,0.1096993908,-0.0064216722,0.1405479908,0.0346085429,-0.0513197146,-0.196865961,0.0078801606,0.0852092505,0.2760110795,0.1000634953,-0.4387730658,-0.1408093125,-0.1907574236,0.331359297,0.4964479804,-0.2855861187,-0.0179454908,0.1170451567,0.2279493809,-0.431779623,0.0991013646,0.4069510996,0.0616520308,0.0558136776,0.1605772227,-0.196325466,-0.0263518207,-0.2005510032,-0.015522846,-0.0812871903,-0.2450969666,0.1379801035,0.2535305619,-0.1621550173,0.0723396018,0.2234296203,0.2248426974,0.2094131559,0.4411480725,-0.0880139023,0.0335321389,0.3215833306,-0.1977277696,0.0530192964,-0.4163442254,-0.2213221192,0.004601588,0.1555845737,0.0431443006,-0.0308006573,0.0894242376,0.0554312021,-0.0492895246,0.0911091119,0.1842016131,-0.1754979193,-0.0029799072,0.07021337,0.1035524011,-0.1834444851,-0.2007070631,-0.0523935743,0.0233831629,-0.0611661263,0.0867941976,0.1566937417,-0.5509833097,-0.0985001028,0.028403068,-0.3185489476,-0.4562836587,0.0733105093,0.4109583497,0.0954850167,-0.1203892529,-0.0074952352,0.6174410582,0.4333248734,-0.0764670596,-0.1840572208,-0.2809346318,-0.0467595868,-0.0322065093,0.0866955072,0.1216947138,0.2625560462,0.5520280004,0.1766430885,-0.1991811842,0.1269986182,0.2516283095,-0.2234730422,-0.1569376588,0.2527378798,-0.0195222571,-0.1268068105,-0.2825405598,0.1944848895,-0.2954165936,0.1875001341,0.4098545313,0.2683029473,0.1570728123,-0.1060613543,0.1378096938,0.1873126626,0.5516356826,0.1906979531,0.3860901296,-0.3571788371,-0.0554284528,-0.5711801052,-0.0245900769,-0.1053961366,-0.0316519812,-0.1443446428,-0.0839933902,0.0324577168,0.2593309879,0.1967318356,0.1654993743,0.1051031351,-0.0875484273,-0.2856193185,-0.1786556393,0.0377548598,-0.1408987641,0.1282964647,-0.2223509252,-0.0397654586,-0.4887033701,0.1265330911,-0.3787437677,-0.2208055407,0.0734933987,0.4357133806,0.2232686132,0.2265576571,0.5930403471,-0.0317760035,-0.0286308639,-0.3275698721,-0.2059123963,-0.1545397788,0.3455994725,0.0363045111,0.4889930487,-0.1071662307,-0.0158568323,-0.3778460324,0.1949944496,-0.3336418271,0.0198232438,-0.1027710661,0.1275236309,-0.1381614953,0.1817010939,0.0611189194,0.0021520669,-0.1441734731,0.2395447195,-0.3893561363,-0.4405485988,0.5200942159,-0.4261544347,-0.6753060222,-0.1937554777,0.0424238034,-0.2450479567,-0.1419898719,-0.4619947672,0.062090721,0.2820001543,-0.0690544322,-0.2712805569,0.3742442727,-0.1549424082,0.0212926641,0.0864809677,0.2581366599,0.0770811439,-0.2606575787,-0.2184218317,-0.2339115441]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/153","title":"Meta-datasets (GLUE\/XTREME\/...) \u2013 Special care to attributions and citations","comments":"As @yoavgo suggested, there should be the possibility to call a function like nlp.bib that outputs all bibtex ref from the datasets and models actually used and eventually nlp.bib.forreadme that would output the same info + versions numbers so they can be included in a readme.md file.","body":"Meta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation\/specific homepage\/etc are very visible and accessible and not only the generic citation of the meta-dataset itself.\r\n\r\nLet's take GLUE as an example:\r\n\r\nThe configuration has the citation for each dataset included (e.g. [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/glue\/glue.py#L154-L161)) but it should be copied inside the dataset info so that, when people access `dataset.info.citation` they get both the citation for GLUE and the citation for the specific datasets inside GLUE that they have loaded.","comment_length":47,"text":"Meta-datasets (GLUE\/XTREME\/...) \u2013 Special care to attributions and citations\nMeta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation\/specific homepage\/etc are very visible and accessible and not only the generic citation of the meta-dataset itself.\r\n\r\nLet's take GLUE as an example:\r\n\r\nThe configuration has the citation for each dataset included (e.g. [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/glue\/glue.py#L154-L161)) but it should be copied inside the dataset info so that, when people access `dataset.info.citation` they get both the citation for GLUE and the citation for the specific datasets inside GLUE that they have loaded.\nAs @yoavgo suggested, there should be the possibility to call a function like nlp.bib that outputs all bibtex ref from the datasets and models actually used and eventually nlp.bib.forreadme that would output the same info + versions numbers so they can be included in a readme.md file.","embeddings":[0.0246247817,0.118831113,-0.0451854467,0.0450591594,0.0441776179,-0.0376197174,0.1662002355,0.4080994725,-0.0163466334,-0.0382265486,-0.1563763916,0.4495483637,0.0651095584,0.1650387347,0.2494661063,0.1432768852,-0.0831086859,0.126027897,-0.0986576602,-0.1266354769,-0.0226358771,-0.000577053,0.2753751874,0.1213619411,-0.0726753324,0.0683133081,-0.0837138742,-0.1722621918,0.0561943837,-0.3044899702,0.0694034994,0.2606022656,-0.0170912296,-0.2680614293,-0.0000978849,0.0377667025,0.3440962732,0.0101948828,-0.3477296531,-0.429287374,-0.1936511248,-0.103100203,0.1493570805,-0.3307287097,0.003330522,-0.0333790928,0.2674756348,-0.1879940331,0.2923165858,-0.0892808661,0.2798752487,0.1032133102,-0.0715401173,0.163459152,0.3679850101,0.3241039813,0.0542786568,0.1083626524,0.4975113273,-0.0500445403,-0.1755764931,0.6305480599,0.0459234454,0.0036622845,0.2669553161,-0.0709657818,0.5887317657,-0.0492290631,-0.096299313,0.5263558626,0.2227196842,-0.6074836254,-0.4398880899,-0.3692813814,0.0451757535,0.2873602211,-0.1065417528,0.2674505413,-0.0276875682,0.1063177064,-0.2240421623,-0.0661293715,0.1180195734,-0.0483124331,0.4212715924,0.3543231487,-0.1338099837,0.119181104,-0.0516938493,-0.2293422073,-0.0088687297,-0.1549333185,0.0400073752,-0.0808936879,0.123238802,-0.3047404587,0.2001968026,0.0923784673,0.3088464141,0.1367182136,0.122009322,0.0266151652,-0.146077022,0.2947081029,0.5069969296,-0.1077478752,0.0833440572,-0.1293120384,0.412833482,-0.0750439763,-0.0995054394,0.1161198542,0.0377844684,0.2770670056,-0.4017126858,-0.0321898535,0.1351315826,-0.356089294,-0.1103997305,-0.209833771,-0.004396481,-0.1123190522,-0.1871229559,0.0738015249,-0.0747267827,0.0355087519,-0.2478306144,-0.0633666143,-0.0975732803,-0.2529425323,-0.1559454054,-0.0071639037,-0.4421660006,0.0629012138,0.1075693518,0.0717808902,0.43800354,-0.1486310512,-0.2177499384,-0.0932364985,0.0806349963,-0.1154959723,0.141805172,0.0463996716,-0.0120881284,0.1354492456,-0.0623426326,-0.1585858762,-0.5453760028,-0.281229496,-0.2335349023,-0.1142608896,-0.1217738688,0.2140998244,-0.2378810495,-0.0336710252,-0.1656148732,0.3436548412,0.0028632823,-0.2042075396,-0.1521623731,0.1259879768,-0.1789656878,-0.2321767509,0.167627275,0.4192044139,0.0847230032,-0.3906548917,0.0857027918,-0.1678960025,-0.224933967,0.0824193656,-0.2435826659,0.2264718413,0.0546728633,-0.0510314219,0.4179317355,-0.4451332986,-0.2007088959,0.1056299061,0.0069582798,-0.0058151111,0.1629912853,0.2626295388,0.2648093998,-0.4369670153,0.115082562,0.2264191359,-0.0909755379,0.1131956801,-0.0979316756,-0.3599979579,-0.1971422583,0.1153209433,-0.1953724027,-0.1405662,0.3687634468,0.2460271269,0.2487063557,-0.0708749965,0.0915198848,0.1125273183,0.0391069204,-0.1303308904,-0.0861907005,-0.1503329128,-0.1870399266,0.0249141753,-0.0964708105,-0.064290747,0.269320637,-0.4242702723,0.1844859421,-0.0245297439,0.0834137797,-0.3695567548,0.3480744064,0.1537958831,0.1634940058,0.1520516574,-0.0422113165,0.065561384,-0.3310511112,-0.0113126291,-0.5354913473,0.0942881033,-0.1187246069,0.0748171806,0.3878872097,0.4740018249,-0.0167377945,0.1244141161,0.0938351825,0.1163709238,-0.0099770743,0.5856169462,0.1358384788,0.544441402,0.3542374372,-0.2291153371,0.1989946216,-0.2761800289,-0.0334047265,-0.2197769135,-0.2923256159,0.0742766932,0.0806699544,0.1615872085,0.2401180714,-0.0031547656,0.0206981767,-0.127638787,-0.3540003598,-0.2646840811,-0.1925270706,0.2989507318,0.1064232588,0.2376115471,-0.0457903147,0.3892675638,0.588792026,0.2582814097,0.0722084194,-0.1035115495,-0.3918624818,-0.3607375324,0.1827531457,0.3668745756,0.0106802741,0.4147942066,0.2208257169,0.0703670457,0.0306203049,-0.0070171594,-0.0146290259,0.0639665723,0.0039685089,-0.1023051664,0.1829533577,0.2220139652,-0.0197271071,0.050205797,-0.0858023539,-0.2350770533,-0.3454783857,0.0444184206,-0.1313546598,-0.326444,-0.3867818415,-0.1903593838,-0.1276940554,-0.2918636501,0.0501473173,0.2012407184,-0.2010379136,0.2224684209,0.1106221527,0.2077444345,-0.3440878391,0.2071490139,0.0854750797,-0.3341499865,-0.0840324461,0.1393871307,0.2725419402,0.2580163479,0.3359660804,-0.06465175,0.0261683837,-0.4569425285,-0.515994966,0.1827281415,-0.2701183558,0.2415601015,0.156835109,-0.4601936936,0.122871004,-0.1160470769,-0.050750494,-0.1861074865,-0.0536630154,-0.0073613338,-0.0501456186,-0.1278600544,-0.1668647081,-0.2610401511,-0.2157847136,-0.3370561302,0.1664503664,0.0380573906,0.2831123173,0.0774715915,-0.2718665004,-0.0369202457,-0.4590601921,0.4280962944,-0.1247672215,-0.3134246171,0.1134670302,-0.2323602587,-0.28865394,-0.1188957989,0.0080536623,0.0566300452,-0.043539118,-0.406419903,-0.3878153265,0.1752029359,-0.4479329586,0.1624153554,0.3990525603,0.2329633683,0.0885439441,-0.1322378069,-0.0513090417,-0.1691470444,0.1176300421,0.0073705055,0.3056493402,-0.0071056825,-0.2780187428,0.260004878,0.4271394908,0.3567825258,-0.219072178,-0.0146385347,0.1176844388,0.1751934439,0.2322545946,-0.0309029538,0.2735729218,-0.2118646652,0.2610421479,0.4443438053,0.1279819459,-0.1374716014,-0.1472721696,-0.0137350857,-0.1451564133,-0.3161193728,-0.0702141449,-0.1799120307,0.0761957839,0.239365831,0.1343641281,-0.0113189053,-0.0088720424,0.0753617212,0.0345781781,0.0030817979,0.2330604643,-0.1782840937,-0.080824092,-0.2139741629,-0.0003337846,-0.0437253267,-0.095579356,0.0804316401,-0.2571632266,0.044679109,0.0432965979,0.0802348182,-0.4035222828,0.0505080223,0.0212400984,-0.0759911835,-0.211048767,0.2083843797,-0.1404273212,0.0276485905,0.3837427497,0.3798472285,-0.3838362992,-0.0999628901,0.2218699604,-0.2906836569,-0.085378252,0.0127613042,0.0433048606,0.0845265687,-0.1861722171,-0.334657371,-0.2545770705,-0.2915425003,0.0688392669,0.0817854777,-0.0503190309,0.1893037111,0.2378100604,0.5395802259,0.036606539,-0.0214021225,0.1680479497,0.2743470371,0.2082947642,0.0686147213,0.474973619,-0.3105337024,0.1759103835,0.004865841,0.0305425264,0.3277992308,0.1406104118,0.3335471749,-0.2353598326,-0.0730708539,-0.3103371561,-0.3443017304,0.0666368306,-0.1151014939,-0.0017656671,-0.3705739975,-0.3775325716,0.4986186624,0.2680627406,-0.2094174623,0.2011909932,0.4137504995,-0.2342975736,0.2603648603,0.272367239,0.9191636443,0.1498700082,0.1800039262,-0.2219550312,-0.1232447773,0.7734662294,-0.2378213257,-0.1032318845,-0.4030280709,-0.050889615,-0.0875690728,-0.0483462177,-0.0609739348,-0.0097106639,-0.156202212,0.0388742387,0.461021781,0.0530400425,-0.0207449831,0.2822639644,-0.0285837352,-0.4216649532,-0.318543911,0.1946771741,-0.2890468538,0.054595843,0.0236220434,-0.3277182579,0.1445377171,0.0117120286,-0.1225617751,0.0325646698,-0.2189267129,-0.0182802528,0.235180527,0.0059492667,0.128790766,-0.0862660706,0.4078099728,0.1767650992,-0.1866821796,-0.1501105428,0.0778151155,0.0461609177,-0.0545196682,-0.1122355387,0.4139033258,0.0515861996,-0.0872314274,-0.160015747,0.2364194095,-0.1824629903,0.0922735259,-0.2390926331,-0.0663060471,0.0143176336,0.1541275233,0.4742465615,-0.0743652359,-0.1986968815,0.1723300815,0.0487580486,-0.1633841842,0.3867480755,0.1621730328,-0.0545145422,-0.0019206168,0.1609745473,0.0643419996,0.2746698558,-0.1538702101,0.0138627533,-0.0589305647,-0.236745894,-0.1681011915,-0.2533482909,-0.0477730222,-0.0620882139,-0.1614845693,-0.2547166944,-0.1479824036,0.4201218784,0.2420392334,0.0814246014,0.012998471,-0.3477077484,-0.1274092942,0.524877131,-0.2895520329,0.2338656187,0.2778621614,-0.0251269527,-0.0746016875,0.352840811,-0.4569594264,-0.2785537839,-0.1057296395,0.2241243422,0.2803422809,-0.0470543653,0.0641498044,-0.1529431641,0.0628762618,0.2037899196,-0.2687698305,-0.2959519029,-0.1975250691,0.0467847511,0.0747405738,0.2488690317,-0.0278540216,0.1466547251,-0.0155415516,-0.1831372976,0.2784824967,0.0426162854,0.0743094161,0.2869681716,-0.2478322387,0.0583494306,0.0973381475,0.1801584512,0.0274759959,0.0700874478,-0.0790478513,0.1992201358,-0.2649215162,-0.070161745,-0.0950428322,0.2005637437,0.1385750473,0.1087582633,0.2033931911,0.2841627002,-0.1086218804,0.230853498,0.3003301919,0.2504829466,-0.0701154619,0.1288276762,0.0497634001,0.2858296633,-0.1443909705,0.0265506171,0.1703904718,-0.0648998916,0.1966520697,0.3415986598,0.6348131299,0.1437174976,-0.0312792808,0.0165652428,0.4313782454,-0.3090530038,0.1085073426,-0.1150923297,0.0771559551,-0.0341038108,0.2147912681,0.3228097856,0.3315179348,-0.1194634065,0.1553512365,0.4127912521,0.3552703857,-0.0143117215,0.4019970894,0.191492945,0.6244620681,0.1322709173,0.3288288116,0.1824136972,-0.0930975303,0.0609375536,0.2356578708,-0.1685187072,-0.1678877324,0.174408257,0.1029071063,-0.2838237584,-0.0977063328,-0.0319373496,0.0447320491,-0.1604592055,-0.1514484882,-0.3046040535,0.2555143237,-0.0021699106,-0.0198319778,-0.5262685418,0.1322362572,0.0754948109,0.3279767036,-0.2590014935,0.1548555791,-0.1746222377,0.0382031128,-0.031889122,0.6182312965,0.2011750787,0.0509652458,-0.0920288861,-0.0575843677,-0.1386909038,0.0254335515,-0.2176241428,-0.0062837545,0.0841073543,-0.2035938501,0.4979931116,0.1799909621,-0.0580427572,0.0022594375,-0.0878293738,-0.1749539822,-0.0972588137,0.179677397,0.1109033599,-0.0554575697,-0.0902675092,0.1233485118,-0.3803673685,-0.1134387404,0.5566621423,-0.053652104,0.200256452,-0.327170074,0.1841356605,0.1794322282,0.2267818302,0.4506832361,-0.1333667636,0.091468066,0.0038500936,-0.5929581523,-0.095376268,0.0523439907,-0.1913161874,0.0228088293,0.2161401659,0.0670496896,0.3399184048,-0.209751606,-0.0620342717,-0.2910601497,-0.0899993181,-0.1750152409,-0.2186882794,-0.5035049319,-0.2248722315,0.1228632629,-0.1133475602,-0.0700516477,-0.0803341642,0.1012629643,-0.09100499,0.0183238219,0.2306822389,0.2187523395,0.3465239704,0.0849297494,-0.0043655522,0.0331295095,-0.1806721985,-0.3137551248,-0.0466434211,-0.1333763748,-0.0495013669,-0.1501333714,0.1640447974,-0.3214978874,-0.0536162928,-0.1226253659,-0.1029359028,0.209572956,-0.0407544374,-0.331140399,-0.0751542822,-0.1512630433,0.060662169,0.0814384446,0.3346470892,-0.1246187836,0.0323785841,-0.0844900161,-0.3636503518,0.0457067974,-0.4250534475,-0.0730658099,-0.1039496884,0.4401104152,-0.169357866,-0.0339414515,-0.8108168244,0.0374916531,0.3362652063,-0.2119695395,-0.0372975543,0.1978939176,-0.105373688,-0.1026731655,-0.1749888211,-0.2042030394,-0.0627431422,-0.21993348,-0.032796979,-0.3536889255]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/153","title":"Meta-datasets (GLUE\/XTREME\/...) \u2013 Special care to attributions and citations","comments":"Actually, double checking with @mariamabarham, we already have this feature I think.\r\n\r\nIt's like this currently:\r\n```python\r\n>>> from nlp import load_dataset\r\n>>> \r\n>>> dataset = load_dataset('glue', 'cola', split='train')\r\n>>> print(dataset.info.citation)\r\n@article{warstadt2018neural,\r\n title={Neural Network Acceptability Judgments},\r\n author={Warstadt, Alex and Singh, Amanpreet and Bowman, Samuel R},\r\n journal={arXiv preprint arXiv:1805.12471},\r\n year={2018}\r\n}\r\n@inproceedings{wang2019glue,\r\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\r\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\r\n note={In the Proceedings of ICLR.},\r\n year={2019}\r\n}\r\n\r\nNote that each GLUE dataset has its own citation. Please see the source to see\r\nthe correct citation for each contained dataset.\r\n```\r\n\r\nWhat do you think @dseddah?","body":"Meta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation\/specific homepage\/etc are very visible and accessible and not only the generic citation of the meta-dataset itself.\r\n\r\nLet's take GLUE as an example:\r\n\r\nThe configuration has the citation for each dataset included (e.g. [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/glue\/glue.py#L154-L161)) but it should be copied inside the dataset info so that, when people access `dataset.info.citation` they get both the citation for GLUE and the citation for the specific datasets inside GLUE that they have loaded.","comment_length":115,"text":"Meta-datasets (GLUE\/XTREME\/...) \u2013 Special care to attributions and citations\nMeta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation\/specific homepage\/etc are very visible and accessible and not only the generic citation of the meta-dataset itself.\r\n\r\nLet's take GLUE as an example:\r\n\r\nThe configuration has the citation for each dataset included (e.g. [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/glue\/glue.py#L154-L161)) but it should be copied inside the dataset info so that, when people access `dataset.info.citation` they get both the citation for GLUE and the citation for the specific datasets inside GLUE that they have loaded.\nActually, double checking with @mariamabarham, we already have this feature I think.\r\n\r\nIt's like this currently:\r\n```python\r\n>>> from nlp import load_dataset\r\n>>> \r\n>>> dataset = load_dataset('glue', 'cola', split='train')\r\n>>> print(dataset.info.citation)\r\n@article{warstadt2018neural,\r\n title={Neural Network Acceptability Judgments},\r\n author={Warstadt, Alex and Singh, Amanpreet and Bowman, Samuel R},\r\n journal={arXiv preprint arXiv:1805.12471},\r\n year={2018}\r\n}\r\n@inproceedings{wang2019glue,\r\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\r\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\r\n note={In the Proceedings of ICLR.},\r\n year={2019}\r\n}\r\n\r\nNote that each GLUE dataset has its own citation. Please see the source to see\r\nthe correct citation for each contained dataset.\r\n```\r\n\r\nWhat do you think @dseddah?","embeddings":[0.1459961236,0.1260700375,-0.0172191672,0.089572154,0.0626265556,-0.0980593339,0.1901406646,0.3274239004,-0.1365870982,0.0089798607,-0.145271495,0.5205239654,0.1002962738,0.0326508507,0.37821877,0.1075446531,-0.1473995,-0.0629663318,0.0565607883,-0.0082267206,0.0059154364,-0.1218897626,0.2556420267,-0.0212328024,-0.0183687899,0.1203053519,-0.097508274,-0.2274171412,0.1760930717,-0.3408681154,0.1228426918,0.3797631562,-0.0701436698,-0.2853853703,-0.0001086736,0.1405759752,0.4462239444,0.0706145167,-0.18459405,-0.4417585135,-0.3292541802,-0.2576588988,0.1386142522,-0.4017037153,-0.0440296866,0.118548505,0.1867649555,-0.2359999418,0.2564884722,-0.1184171513,0.1935309619,0.2313036621,-0.2722188234,0.2109432369,0.4602375031,0.3790766299,0.0113808513,0.0496142618,0.3216421902,-0.0529675297,-0.1204490811,0.562353909,-0.1215746328,0.0837367252,0.2450886071,-0.0117531838,0.320433706,-0.0062420312,-0.1649091095,0.5180585384,0.3185868263,-0.4948662221,-0.4512437284,-0.5133804083,0.1050352007,0.4594183266,-0.0621292926,0.3865552545,0.0475593098,0.1687434316,-0.1067017838,-0.0047656014,0.1005403474,0.0102560399,0.3739118278,0.1554258317,-0.1244882941,0.1860373616,-0.019477956,-0.0803717822,0.0637969524,-0.2097226381,-0.0261022337,-0.0222852975,0.1380210072,-0.1774203926,0.1316171736,0.0892143995,0.2456714958,0.3150606751,0.1443891376,-0.0571823828,-0.1413040012,0.3117470741,0.5554292202,-0.1336813122,-0.0731785223,0.0855107456,0.4026112854,-0.087894313,-0.0893845409,0.1603705436,0.0179317407,0.2412939668,-0.4022683501,-0.0895631835,0.2245129496,-0.2601105571,-0.2941981852,-0.1213130504,-0.0415150188,-0.2908771336,-0.2496638298,0.0559084155,0.0048680846,-0.0641784221,-0.290843308,0.1059360504,0.0068274052,-0.2634836733,-0.2163226604,0.1354009956,-0.3099243045,0.1489433646,0.1042680815,0.1488614231,0.37087515,-0.1605480015,-0.10679508,-0.1834675074,0.0414253213,-0.1364021748,0.1206368729,0.1558703333,-0.0061822073,0.309397012,0.0233883373,-0.1726247817,-0.4984369576,0.0189275499,-0.1504622847,-0.0613014959,0.1211765558,0.1209563762,-0.4014960527,-0.1249210164,-0.3572137952,0.3441362083,0.0831248239,-0.1271319836,-0.0298275966,0.0958377048,-0.1046818718,-0.2167442143,0.20299685,0.6037052274,0.094476603,-0.3677794635,0.0785753503,0.0083179902,-0.1182465777,0.1067932174,-0.2318627685,0.4463392496,0.04635581,0.0373966135,0.2166668475,-0.574362278,-0.27188766,0.1112951934,-0.1332582831,0.0907881409,0.2957184613,0.2834581137,0.2771713436,-0.4240314662,0.0544349253,0.2461315542,-0.0672732145,0.1181529537,-0.0978874788,-0.3014930487,-0.0465724804,0.0463174433,-0.2792060375,0.0065148612,0.2964585125,0.2114416212,0.303152591,0.1095388085,0.05622622,0.0741520971,0.0705616847,-0.1719370037,-0.0168912783,-0.0143582625,-0.1421349943,-0.0081006167,0.1352147758,0.0016021794,-0.0393484756,-0.4302903414,0.143793717,-0.091854699,-0.0512564518,-0.3166414797,0.2020543069,0.1330290139,0.2059656829,0.059566047,0.0042645358,0.1732726395,-0.3910719454,0.0728156939,-0.4693699479,0.1274827272,-0.0700266212,0.0972751603,0.2877371609,0.3687362373,0.2272826284,-0.0500230901,-0.0158360507,0.1769267172,0.1354346126,0.6815310717,-0.0163425338,0.4602656662,0.2904715538,-0.5873613954,0.1933742315,-0.2625128627,-0.0854331106,-0.2483108044,-0.3497190475,0.1101983935,-0.0919145346,0.1990024894,0.1812399477,0.0126496563,0.0414172411,-0.1380126029,-0.3803361058,-0.4260339439,-0.1419214308,0.4059378207,0.0526502654,0.1769868135,-0.0444141477,0.3668403327,0.3053131104,0.2437003702,0.0134226372,-0.1738654673,-0.3735981286,-0.2639075518,0.185659498,0.5812264681,0.0588332973,0.3410807848,0.1458032876,0.0462883525,0.0564634614,-0.107605733,0.0538325384,0.0469712839,-0.1568640321,-0.1075319126,0.1514116377,0.2761743963,0.048563648,0.129694894,-0.0437429436,-0.2355324924,-0.5040350556,-0.0668784827,-0.1607832313,-0.3451222777,-0.3923563659,-0.2192247808,0.026634993,-0.335919112,-0.0565615892,0.1736796796,-0.3128882051,0.1817235798,-0.0108183147,0.2305707783,-0.4563739598,-0.0576599091,0.1130955145,-0.2517158687,0.2092413455,0.061892271,0.2074334919,0.2063422203,0.2276794463,0.0069877878,-0.0038602711,-0.4430010915,-0.4584119618,0.1721507162,-0.2430887073,0.223734051,-0.068184562,-0.381262958,0.1325632781,-0.1528228372,0.0077382294,-0.164008975,-0.0976217315,0.0144634107,0.0926716849,-0.0532015525,-0.1190430447,-0.1074228212,-0.3545091152,-0.2291745543,0.1439630091,0.0865277648,0.151639998,0.1420312226,-0.3330469429,-0.0992255136,-0.2470338494,0.4689602554,-0.1464550942,-0.4327390492,0.2442490458,-0.3043130934,-0.2188072503,-0.382504493,-0.0708175898,0.0686381385,0.0416203998,-0.3691650331,-0.3203602433,0.0656622276,-0.3541071415,0.0406556539,0.4220488071,0.1390880942,-0.0038265618,-0.0347152166,-0.043318402,-0.1257904619,0.0180331022,0.0695181489,0.0802453682,-0.0301386248,-0.4453698993,0.2384728044,0.6093105674,0.357383132,-0.0821101516,0.0699076876,0.0932861716,0.2277640104,0.2781896293,-0.1315300465,0.2565012574,-0.1018916667,0.191035673,0.3069961667,0.025632238,-0.0925609693,0.0231936779,0.1315832734,-0.2536820173,-0.2625791728,-0.1427263469,-0.2425165921,0.0999456123,0.2951813936,0.0251958352,-0.0353681892,-0.1885510832,0.0486023463,0.0317493118,-0.0509754606,0.1091302186,-0.1552980095,-0.1770073026,-0.2219056487,-0.0770820156,-0.0770994201,0.0108810598,0.0537928902,-0.2894570529,0.0265580174,0.0722320452,0.2669827342,-0.4505029023,0.2121725678,-0.032419499,0.10508506,-0.3226864636,0.0640611872,-0.1325684488,0.1368396878,0.336933434,0.2544022202,-0.2235413939,0.0444946364,0.2816637754,-0.2996258438,-0.0928030014,-0.1158425137,0.0406182297,0.0177642647,-0.0124280443,-0.3173237741,0.0742201209,-0.1216428056,0.0145625388,0.0549925864,-0.1626982391,0.2023650408,0.2333432734,0.3720487952,0.0684486479,0.0558963157,0.3628766835,0.4473117888,0.0581358485,0.0180231053,0.7409008145,-0.2952211201,0.0563455075,-0.1472657025,-0.1130052432,0.357837081,0.1461554766,0.2450062633,-0.183985427,-0.1946238428,-0.3163897693,-0.5575904846,0.0014644146,-0.1903021336,-0.0404741988,-0.3032287657,-0.4608917236,0.5262104869,0.274582088,-0.1825014353,0.2443663627,0.4735277593,-0.2609789073,0.2932690084,0.1494868398,1.0307896137,0.0836325884,0.0950735882,-0.2820116282,-0.0443105064,0.8447420597,-0.2465769947,-0.1482634544,-0.4357126355,-0.154820621,-0.0630074218,-0.1531615108,0.0450283438,0.1211143956,-0.0931899399,-0.0749623403,0.3189808726,0.1369545013,-0.0409425721,0.3400014937,0.0210309625,-0.3488111496,-0.1081471667,0.1052887514,-0.2712959349,0.1158172563,0.1484314203,-0.1422663629,0.0167673733,0.0847001523,-0.2729050219,0.1285783798,-0.3497876525,0.0474079847,0.2040601671,-0.099740468,0.2513721287,-0.0815373957,0.362647891,0.0766853392,-0.077524282,-0.1095659956,0.1261809468,-0.0558620989,-0.0091482205,-0.1079880372,0.3444417417,0.2466157377,-0.1017064601,-0.1256991327,0.2732780576,-0.3068853319,0.2283775508,-0.3318193555,-0.0723537728,-0.0460721143,0.0602134913,0.4835633636,0.056153737,-0.1791360825,0.0989872143,0.1726373583,-0.1966995895,0.39186427,0.0239054747,-0.0214624442,0.1064641178,0.2095681131,0.1908181459,0.3242694438,-0.1016838923,-0.078185603,-0.0653061047,-0.1796934754,-0.0285092909,-0.4060038626,0.0183630344,-0.0783497617,-0.2472977787,-0.0412862115,-0.2350756079,0.2162176818,0.1541727632,0.1411766112,-0.13114658,-0.3797957003,-0.128626436,0.3909105957,-0.3289385736,0.2858897746,0.2250097245,0.0502598397,-0.09526366,0.5021759868,-0.3409575522,-0.3235459328,0.0221151281,0.1368127912,0.4950767159,0.1555399299,0.0880417526,-0.2358874977,-0.0016130841,0.1939478964,-0.1645644307,-0.2098518908,-0.2161638588,0.0900487378,0.1563793272,0.2249210328,-0.1129648983,0.1534243673,0.1410470605,-0.1431699544,0.2750482261,0.0476489775,0.1550292969,0.2671999633,-0.1416065693,0.0867233202,-0.0700107589,0.2183841765,0.034005817,0.0793861151,0.0670531243,0.242281124,-0.2143163234,0.0508971177,-0.1181933507,0.067170687,0.1286502928,0.1199416444,0.0316485949,0.1299973428,-0.1956668198,0.218742907,0.2273792624,0.2240438163,-0.1611781418,0.0137801282,0.0545649901,0.1908657253,-0.0332820415,0.0127062462,0.1787575781,-0.0020338323,0.2374046892,0.3935462236,0.6818829179,0.2885891497,-0.2439573556,-0.0527394712,0.6587566137,-0.2186493427,0.2375546992,-0.2637001574,0.0659804493,-0.0107328268,0.2295999378,0.272857517,0.3210789263,-0.2714985013,0.220680967,0.4711928964,0.4068385661,-0.1425220817,0.5252481699,0.1469684243,0.6082965136,0.1113488227,0.2239979953,0.1076094508,0.0341427475,0.1358426064,0.1710914224,-0.0182897635,-0.0725185573,0.0644753128,0.143590346,-0.3939923048,-0.1767218411,-0.0251856484,0.0049262443,-0.1533123255,-0.0583387539,-0.4621622264,0.4027003944,0.00104166,0.0057613649,-0.4525317252,0.2401925921,0.0155174956,0.3872750401,-0.2879736722,0.0384007767,-0.362652123,-0.0686880797,-0.0212081559,0.6027603149,0.1913992167,-0.0533110201,0.0139935687,-0.2052620202,-0.1629516482,-0.0150786182,-0.1240524799,0.1879218966,-0.0923248678,-0.4238505363,0.4680993855,0.1247907877,-0.022123171,-0.0545001328,0.401155591,-0.1949981898,-0.048451025,0.2918994725,0.0863604322,-0.1618094444,0.057146281,-0.0852502212,-0.3430182934,-0.1911881119,0.658212781,-0.2030204087,0.1391563714,-0.2320263833,0.1199664623,0.226610288,0.2377991527,0.5384885073,-0.0507965647,-0.1292383075,0.0027813399,-0.6148341894,0.0713938847,-0.006926117,-0.4261507392,0.1241677925,0.353978008,0.101686582,0.4172505438,-0.1787737161,0.0932393968,-0.3028385341,-0.1100546569,-0.2607184649,-0.1933708489,-0.4956416488,-0.1775759459,0.0974054486,-0.0375238433,-0.1832970232,-0.024274312,-0.0276758447,-0.1284870505,-0.0492253415,0.1767484844,0.0764692202,0.5610907078,-0.1291882843,0.167000398,0.0468154214,-0.0033250903,-0.2533151805,-0.0164968017,-0.2715819478,0.108231172,-0.1755800545,0.1581413597,-0.3534299433,0.0128837721,-0.163362518,-0.2471809983,0.1130566373,-0.1065551788,-0.232426703,-0.0351042673,-0.172358498,0.1253805757,0.0198150165,0.2141256332,-0.1107533202,0.2137901038,-0.034669932,-0.4619117379,0.0293232221,-0.5250598788,0.0538856722,0.0534540787,0.4995156527,-0.3100973666,0.1177526042,-0.8463413715,0.0515844785,0.4032821953,-0.277438581,0.0514607653,0.1651724577,0.0618404485,-0.0724133328,-0.1524946094,-0.458316952,-0.0309897605,-0.133245185,-0.2018910795,-0.2705309391]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/153","title":"Meta-datasets (GLUE\/XTREME\/...) \u2013 Special care to attributions and citations","comments":"Looks good but why would there be a difference between the ref in the source and the one to be printed? ","body":"Meta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation\/specific homepage\/etc are very visible and accessible and not only the generic citation of the meta-dataset itself.\r\n\r\nLet's take GLUE as an example:\r\n\r\nThe configuration has the citation for each dataset included (e.g. [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/glue\/glue.py#L154-L161)) but it should be copied inside the dataset info so that, when people access `dataset.info.citation` they get both the citation for GLUE and the citation for the specific datasets inside GLUE that they have loaded.","comment_length":21,"text":"Meta-datasets (GLUE\/XTREME\/...) \u2013 Special care to attributions and citations\nMeta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation\/specific homepage\/etc are very visible and accessible and not only the generic citation of the meta-dataset itself.\r\n\r\nLet's take GLUE as an example:\r\n\r\nThe configuration has the citation for each dataset included (e.g. [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/glue\/glue.py#L154-L161)) but it should be copied inside the dataset info so that, when people access `dataset.info.citation` they get both the citation for GLUE and the citation for the specific datasets inside GLUE that they have loaded.\nLooks good but why would there be a difference between the ref in the source and the one to be printed? ","embeddings":[0.3710706532,-0.189554438,0.0191334877,0.1910193712,0.2161493897,-0.2085573524,0.1987189054,0.1203194708,-0.4126298726,0.1728314608,-0.069900319,0.2978672087,0.5788611174,-0.0941340774,0.0515486635,0.1253734082,0.0821742415,-0.0912789628,0.0445878953,-0.2442144752,0.0848147348,-0.0631177574,0.0506394282,-0.1104451492,-0.2710995376,0.2647356689,0.0209034123,-0.206914708,0.0229565091,-0.2019968927,0.11226587,0.231772393,-0.131111443,-0.1376200765,-0.0001118262,0.1149284765,0.5276104212,0.0706696734,-0.2119960338,-0.3199276626,-0.5297551155,0.1398907006,0.0771626979,-0.2583320737,0.0313991494,0.1823125184,0.1165360734,-0.352959007,-0.0093296245,-0.0394008905,0.1819293499,0.0825059935,-0.2895170748,0.1119412854,0.5242987871,0.4687557817,-0.0029224367,-0.0349701755,0.305057168,0.1795527488,-0.0236520581,0.4301710427,-0.0821437016,0.1121853068,0.3664802611,-0.0346075259,0.1556817293,0.1016899049,0.2127133757,0.5792899132,0.5639744401,-0.3503526747,-0.4455969632,-0.0643575191,-0.2279416621,0.7072249055,0.0023293032,0.3356460631,0.2440561801,0.1996308267,-0.2165645659,0.0758417472,0.1721116006,-0.0429942757,0.1275491565,0.063676253,-0.2236409485,0.2789629102,-0.114293687,0.1594837159,0.0819841176,-0.420992285,-0.0768155605,-0.0632445887,0.1375056356,-0.0246544182,0.1798446774,0.24040775,0.1731196642,0.3289603293,0.10105519,-0.0538053177,-0.1195121408,0.213840425,0.4652202427,-0.058295507,-0.0239802003,0.1139115691,0.3770868778,-0.0983374491,-0.1944268197,0.0007585639,0.195794493,0.191299364,-0.1929052472,-0.1453934163,0.1623052359,-0.463285774,-0.2766082585,-0.0817648098,-0.1609376371,-0.1545790881,-0.2178323716,-0.0247255452,-0.1249387711,-0.145170778,-0.403859973,0.0624164529,-0.0965505093,-0.1777428389,-0.2701953948,0.0317755081,-0.4211231172,-0.0052372869,-0.1362654716,-0.0847613662,0.459787637,0.0404549167,-0.1569712758,-0.2084572762,-0.1347543299,-0.3011267781,0.1222964078,0.2674717009,-0.1152221411,0.5308942795,0.0520881526,-0.0636469051,-0.2705881,0.0262673702,-0.2270457894,-0.0266717244,0.0976305902,0.1063080281,-0.3084685504,0.0589838736,-0.2892626226,-0.0296794437,-0.0860867947,-0.1939717382,-0.03830152,0.0436109938,-0.0529656373,-0.3850648999,0.28538239,0.4911711812,-0.0857126787,-0.1355216652,-0.0564725474,-0.1587516814,0.1264362782,0.1260387152,-0.1900234967,0.3097705245,-0.0281818714,-0.096990332,0.0837962702,-0.6937301755,-0.3254210055,0.1508454531,-0.1544984579,0.1510373801,0.2302343547,0.3041201234,0.2933088243,-0.2983376682,0.0998613089,-0.0219012257,-0.038767688,-0.0237236302,-0.1900504827,0.1869901568,-0.2530880868,-0.253964901,-0.3992248774,0.0562268831,0.5640982985,0.0704080537,0.1283388287,0.1379701346,0.1192498207,0.1941003799,0.1917908341,-0.0824429095,0.0546214208,0.187761724,-0.1661573201,0.0924347341,0.1991955936,-0.0035535593,0.024902042,-0.4276168346,0.0427061208,0.0044620815,0.0468668975,-0.4431465864,0.1442899406,0.1915252656,0.2306707352,0.1354448199,0.1846791208,0.1833567172,-0.365303874,0.0683992058,-0.2830339968,0.1996475905,-0.0797477588,0.1337218434,0.1323318928,0.1372403353,0.2207589298,0.0120502645,-0.0985877812,0.3556069136,0.3995122612,0.5275145173,0.1119190454,0.5339223742,0.1610894203,-0.4089438021,0.2597426176,-0.3462619185,0.0743694752,-0.3508377671,-0.3630340695,-0.0470067561,-0.0801867545,0.135768339,0.2215691954,-0.0757960975,-0.0397508293,-0.0386020765,-0.4364562333,-0.3766801953,-0.2181799412,0.4416787922,0.1016835794,0.4957173169,-0.1061612293,0.3637826145,0.4405107796,0.1013623327,-0.1691761315,-0.1962750405,-0.5381575227,-0.3298218548,0.1286329925,0.5671320558,0.2782896757,0.2022490352,0.1697903275,-0.0425620712,0.0787486658,-0.1126358956,-0.0997967198,0.0701482669,-0.1495620608,-0.285508126,0.0689722747,0.2679756582,0.1870436817,0.2309084386,-0.0301250722,-0.2843521535,-0.3232039511,0.04307466,-0.3240039349,-0.0407652333,-0.5766679049,-0.3918191493,-0.0104558105,-0.4315826297,-0.1078247204,0.1349769384,-0.1740663052,0.0823604986,0.1595657617,0.0688121915,-0.2842487097,-0.0293627661,0.1395149082,-0.1835513711,-0.1281135529,-0.0044218297,0.0175670516,0.1599948257,0.0829468295,-0.1667425781,0.068767786,-0.4904097915,-0.4789209664,0.1207591221,-0.2202389538,0.3203408122,-0.0384610891,-0.2237913311,0.1036441401,0.1945070475,-0.0011661422,-0.0906019956,-0.1277207136,-0.0154326176,0.0522626042,-0.1932349801,0.0065811984,-0.1787864119,0.0111029781,-0.0925588161,-0.1211540774,-0.0922701806,0.0162157174,-0.0017527689,-0.1606836617,-0.2513112724,-0.2981009781,0.5867748857,-0.1880421191,-0.53858459,0.1440544128,-0.1071985289,-0.3634325862,-0.1922984421,0.096953012,0.1107210442,0.0934633613,-0.4405764937,-0.2818561494,-0.075861305,-0.3825753629,0.0945296958,0.5671178102,0.1745609343,-0.0541002937,-0.0420439877,-0.3272089362,-0.2793985903,-0.0696472153,-0.0173725821,0.1907824576,0.0211140905,-0.4315021932,0.1551573426,0.7282778621,0.1914071292,-0.2086044848,0.0034594361,0.1567714959,0.3637274802,0.0413013101,-0.0831241533,0.2958693206,-0.0617816038,0.1687472165,0.3897436559,0.1141489223,0.038711518,0.060301315,0.0741800517,-0.139454782,-0.266595453,-0.202994898,-0.1411861032,0.0127278119,0.4299002886,0.1416197419,-0.0056187999,0.0206397045,0.0869596526,0.0698442906,0.0233843252,0.2531621456,-0.1308822185,-0.2867669165,-0.0971667841,-0.0664034486,-0.0387144946,-0.049675703,0.1097015068,-0.2118722647,0.1932049394,0.3003405929,0.4350329936,-0.417619139,0.0963036418,-0.1038191095,0.0313386917,-0.3685772121,-0.0163237527,-0.0223166198,0.1766922772,0.3939031661,0.1989233643,-0.1897853315,-0.1051552594,0.2461254597,-0.221157074,0.0724719092,0.029019108,-0.1387590915,0.0929427743,-0.0739365444,-0.6030511856,0.0187375359,-0.1044100747,0.0593464002,0.0096813273,-0.1021588296,0.3224580288,0.039136596,0.3551358283,0.0348695628,-0.0689317584,0.2235226482,0.4400111437,0.290381819,0.2235567272,0.7098864913,-0.2559742928,-0.0053986264,-0.0037876987,-0.0892186537,0.1618127674,0.009479885,0.335227102,-0.269246459,-0.1782659739,-0.2961285412,-0.374494493,0.0646152571,0.008303171,0.2035894245,-0.4331896603,-0.4817038476,0.3110720515,0.2815311253,-0.2063479424,0.3583265841,0.6577655077,-0.0655946434,0.0816554353,0.0890252143,0.9955168962,0.2413372695,0.0766954347,-0.269965589,-0.2412139922,0.4257271588,-0.3257038593,-0.1483791918,-0.3584233522,-0.2681060433,-0.1424036771,-0.2714770734,0.0717015639,0.0205782652,-0.0349523127,-0.0553455688,0.2691440582,0.3271937668,0.1875399947,0.1650419235,0.3169312775,-0.1339176446,0.2843642533,0.069775112,-0.0130591681,0.2349355817,0.0731813014,-0.0077356631,-0.2012022138,0.0773546696,-0.2886352837,-0.0520210378,-0.2754355669,-0.0249541905,0.3710817695,0.0561725087,0.1464805901,0.0981513187,0.4332255721,0.2281130701,0.0255290866,0.135901913,0.3105177283,0.0054032556,-0.0137964124,-0.1647181809,0.3166963756,0.1933377385,-0.3221420646,-0.1632477194,-0.0083784489,-0.1810971051,0.1805637479,-0.2429842502,-0.1155785248,-0.1082163602,0.1011651382,0.2719067335,0.1162843257,-0.1731322408,0.0870802253,0.0594926141,-0.0209991112,0.4342000484,0.0265938807,-0.3213006556,-0.0477682166,0.3901553452,0.1667378247,0.4073674083,-0.2308252007,-0.1439904273,-0.1466341913,-0.1201416925,0.0269581843,-0.4322423637,0.0031581086,-0.046739269,-0.308405906,-0.1173509285,-0.2169158906,0.3540162444,0.1399853677,0.0820275843,-0.1853622347,-0.3884167373,-0.0328477025,0.194494918,-0.3568926752,0.1773681045,0.3129887283,-0.1322024763,-0.0110282153,0.3492723405,-0.3185352981,-0.2000140101,0.0586608872,0.3657361269,0.3873246908,0.1077220216,0.2268543094,-0.0975844264,0.0175084099,-0.0145992562,-0.1545727849,-0.1769678146,-0.0821051002,0.08822155,-0.0011965152,0.150762409,-0.1493171901,0.0870346427,0.1106299385,-0.1158173084,0.5244326591,0.2708853185,0.1177692786,0.1045741737,-0.3514069021,0.2301556617,0.0059177098,0.1880840361,0.0719177052,0.2233790308,0.0242969524,-0.0159117635,-0.0473979786,-0.1082611158,-0.0254385583,0.0831866488,0.0861197934,0.0690012574,0.0336107127,0.0939615071,0.1466305405,0.1949994862,0.1073074266,-0.0318157002,-0.1631299555,-0.1450655162,0.0851987526,0.1608677208,0.0558377877,-0.1083912626,0.2466586828,0.0391708761,0.2540230453,0.2905915678,0.6154165864,0.2174603045,-0.1451695114,-0.1075324044,0.5513700247,-0.1269809008,0.1778581738,-0.1611260325,0.0784477815,0.1300777346,0.4158474505,0.3707647622,0.056975279,-0.0141050247,0.3265420794,0.3966307342,0.3768800795,-0.1455526352,0.4099495113,0.0208457615,0.7864277363,0.1345523149,0.0224277619,-0.078428261,-0.0664540157,-0.2014079839,0.2849704921,0.0579815097,-0.1270243078,-0.0439570844,0.0324937813,-0.4116569757,-0.2509120107,-0.0901643559,0.2941636145,-0.1861133426,-0.0275498107,-0.2019683868,0.3964211047,-0.103546612,-0.1722547412,-0.3388819098,0.3151679635,-0.1727359742,0.3633085787,-0.0682506412,-0.0829107687,-0.4237008989,0.0068067913,0.1390761882,0.5101337433,0.1666077375,0.1185481995,0.2297815531,-0.3136428893,0.0034597046,-0.1398490071,-0.0611804463,0.2551799119,-0.1935110837,-0.3489827216,0.4559244514,0.1366758794,-0.0269264132,-0.0449476242,0.145423606,0.0375789553,-0.1526239365,0.1556871831,-0.2055647522,0.0173788015,-0.0842236206,0.1623413861,-0.2820925713,-0.1711754203,0.5271524191,-0.4188021719,0.0813421607,-0.3827397525,0.1020132899,0.1935077757,0.2114935964,0.6204615235,-0.1037727818,-0.1469076127,-0.1095471978,-0.4079175293,0.1644622087,-0.1582414806,-0.3184521794,0.0875775069,0.3191933632,0.0396871381,0.1900385022,0.1147427708,-0.0094537437,-0.3201802373,0.0919591486,-0.0765101984,-0.197186619,-0.608258903,-0.375665158,-0.0783490017,-0.0393030047,-0.0120838135,-0.069850266,0.0790322721,-0.1023752317,0.1293747425,0.2980475426,-0.0536593124,0.5025126934,0.0655843392,0.1570371389,0.1724780351,-0.0988541469,-0.2360389531,0.1621499807,-0.0947708338,0.092025876,-0.3200398684,0.0492448956,-0.1787626147,-0.0936449096,-0.156188041,-0.0930555761,0.2712773085,0.0855677947,-0.3052106202,-0.0915409327,-0.3662164807,-0.0475925319,0.1131279394,0.187800929,-0.0185386967,0.1009799168,-0.0904691666,-0.3305736482,-0.0814046413,-0.4412478209,0.0493259579,-0.0729332939,0.3531402946,-0.396494478,0.0092270244,-0.8006329536,0.093430236,0.3816471696,-0.2947748005,-0.1552791893,0.0444764681,-0.0449647717,0.1206379831,0.0081219655,-0.3054713011,-0.072244063,0.0304554421,0.007134594,-0.1575198174]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/153","title":"Meta-datasets (GLUE\/XTREME\/...) \u2013 Special care to attributions and citations","comments":"Yes, I think we should remove this warning @mariamabarham.\r\n\r\nIt's probably a relic of tfds which didn't have the same way to access citations. ","body":"Meta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation\/specific homepage\/etc are very visible and accessible and not only the generic citation of the meta-dataset itself.\r\n\r\nLet's take GLUE as an example:\r\n\r\nThe configuration has the citation for each dataset included (e.g. [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/glue\/glue.py#L154-L161)) but it should be copied inside the dataset info so that, when people access `dataset.info.citation` they get both the citation for GLUE and the citation for the specific datasets inside GLUE that they have loaded.","comment_length":24,"text":"Meta-datasets (GLUE\/XTREME\/...) \u2013 Special care to attributions and citations\nMeta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation\/specific homepage\/etc are very visible and accessible and not only the generic citation of the meta-dataset itself.\r\n\r\nLet's take GLUE as an example:\r\n\r\nThe configuration has the citation for each dataset included (e.g. [here](https:\/\/github.com\/huggingface\/nlp\/blob\/master\/datasets\/glue\/glue.py#L154-L161)) but it should be copied inside the dataset info so that, when people access `dataset.info.citation` they get both the citation for GLUE and the citation for the specific datasets inside GLUE that they have loaded.\nYes, I think we should remove this warning @mariamabarham.\r\n\r\nIt's probably a relic of tfds which didn't have the same way to access citations. ","embeddings":[0.0700624064,-0.1445659846,-0.0095198443,0.0153763946,0.3223144114,0.1147492006,0.2942697406,0.2142912596,-0.0758150816,0.2329561561,-0.199527964,0.0947120413,0.2110273242,-0.1058298051,-0.0246724095,0.1326556653,-0.094471477,-0.0668338239,0.006023536,-0.0338115767,-0.0002153188,-0.1118184403,0.1891224533,0.4480978847,-0.1449099481,0.0250754803,0.0706492141,-0.0829554275,0.0318240412,-0.6247040629,0.1060621962,0.2154619247,-0.1298779696,-0.1473544389,-0.0000993931,0.2487725765,0.5883553624,0.0202402454,-0.043448884,-0.1799246371,-0.3961481452,-0.1768084615,0.1669310182,-0.4316168427,-0.0729646087,-0.0108446572,0.1668516099,-0.2109289765,0.0664901882,-0.1339367628,0.2768972218,0.345723927,-0.0828671306,-0.002588365,0.4308311939,0.2690129876,-0.1015345007,0.0697410926,0.2587527633,0.1499801278,-0.1898740679,0.5585029721,-0.2547703087,0.2304152101,0.3343616128,-0.0307750218,0.4076521695,-0.1766723543,-0.0784500688,0.551199913,0.4213293195,-0.2138153911,-0.2104596198,-0.3318564892,-0.0781147629,0.4283214509,0.116151616,0.264708966,0.0232580155,0.2159827799,-0.070054926,-0.1256978512,0.0333455764,-0.1293752789,0.3187437356,0.1101646498,-0.2037925571,0.2475543916,0.2031625956,-0.0234887023,-0.0203431789,0.0218580663,-0.2034279555,-0.1890849173,0.1768645346,-0.2052761167,0.2073391229,-0.0320457481,0.3742009699,0.1205046102,0.0731947422,-0.1955579668,0.0892424658,0.2061319351,0.5328937769,-0.2574990988,-0.0131412949,-0.019457059,0.4737762511,0.0001122737,0.1296988279,0.0355440378,0.2302269936,0.0140746674,-0.3832718134,0.1017331108,0.3223795593,-0.2117628753,-0.3561384976,-0.1078935638,0.0404996201,-0.2033008486,-0.2048760355,0.0807845742,0.1384629458,-0.2026667148,-0.3101668656,0.0859198272,0.0247680023,-0.3047363758,-0.3057552278,0.0872142762,-0.4357516468,0.0824755132,-0.1928705871,-0.0116642546,0.2965302765,-0.1448454857,-0.0302492436,0.1089663804,-0.140047878,-0.2112825662,0.2341417968,0.4489609897,0.0946052149,0.3177887797,0.1769146174,-0.0336496532,-0.3718382716,0.1521982253,-0.3379709125,-0.490080595,-0.0837172195,0.2100351751,-0.349639535,0.0844409093,-0.3771421015,0.1509403735,0.1434283555,0.0649405122,0.0921902284,0.1776473969,-0.2130921036,-0.3454888463,-0.0422288962,0.5463482738,-0.0827619508,-0.0866190866,-0.019095622,-0.4698160887,-0.1299040914,0.260160327,-0.4755264521,0.329053849,-0.2679233551,0.1474170387,0.0395524837,-0.4772801697,-0.1736409366,0.3119107187,-0.1044778526,-0.0479145609,0.1176473051,0.1089152396,0.1659002155,-0.5870681405,-0.0304698683,-0.1759415418,0.1104888618,0.1192117855,-0.1627025455,-0.3829751015,0.0566528663,-0.2763588428,-0.2641004622,-0.0128724966,0.4228345156,0.120945245,0.2197252661,0.0305886101,0.0377682596,0.2376689315,0.1482820213,-0.1065012589,-0.0714381114,0.1029490158,-0.2082951963,0.0652793571,0.145004645,0.0664696246,0.1364903152,-0.4238216281,0.0991497263,-0.0161927138,-0.1137727126,-0.2195951492,0.2538335323,-0.0830094963,0.0947018638,0.1340948641,-0.0194897838,0.1364644319,-0.6126280427,0.0462036319,-0.2305437773,-0.1165874153,-0.2976146638,0.2311675698,0.0554820672,0.0804915801,0.1117280722,0.0237842239,-0.0524694696,0.3400121629,0.051771,0.3949137032,-0.0406735316,0.6138724089,0.223179847,-0.5713280439,0.1378154755,-0.215835169,-0.1533059627,-0.1040812507,-0.1582480073,-0.135267809,-0.0885922164,0.060109023,0.0451331735,0.0000661988,0.0415534154,-0.006741839,-0.250873059,-0.4003402889,-0.2829267681,0.500271976,-0.072398968,0.2863230705,-0.108521983,0.2606117129,0.3621717393,0.0548170544,-0.0393292792,0.0296875592,-0.0757627413,-0.319732219,0.1559106708,0.496068716,0.2209616452,0.3079429567,0.145032391,0.1003675088,0.0301022269,-0.2011739463,0.097958982,0.1186491847,-0.0690442175,-0.1829247326,0.0365337618,0.3410843015,-0.0014196211,0.1646095514,0.0206258371,-0.0387642011,-0.6539142132,-0.1158526167,-0.1930624992,-0.3580485284,-0.0513720177,-0.0068286932,-0.2454954535,-0.4027350545,0.164192602,0.0823664516,-0.5194941759,0.2355529219,0.0435500294,0.2066192627,-0.1651600301,0.1021998078,-0.0209218953,-0.1009680852,-0.0554131791,0.0996649414,-0.0378540866,0.0698290169,0.2612133622,-0.1642963141,-0.0497960672,-0.5962061882,-0.382376492,0.2521505356,-0.3279669881,0.2222896665,0.1489731222,-0.1134576797,0.0981157571,-0.1255473793,-0.0929596573,-0.1906380951,-0.1266705096,-0.0908182263,-0.0062617552,-0.035066884,-0.1715420634,-0.1666036695,-0.2416954339,-0.3181687295,-0.1549535394,-0.1906793118,0.2219138294,0.3152846992,-0.1027839258,-0.082713455,-0.2239649892,0.3041283488,-0.4019365013,-0.5048869848,0.1658492684,-0.2215511352,-0.3774491847,-0.042088896,0.0598330535,0.1222768798,-0.1280859709,-0.3601261377,-0.4373358488,-0.0798779204,-0.3631129861,0.0020136079,0.2800777853,0.2119290233,-0.0499435849,-0.1743852198,-0.0333191976,-0.0635367334,0.1218525544,-0.1271622181,0.2680666447,-0.0562351793,-0.3644718528,0.2132196575,0.5952408314,0.175683111,0.0378459729,0.1164054051,-0.0386803783,0.6246584654,0.0095963432,-0.1521571279,0.2418382466,0.0098849526,0.0968593433,0.2258625627,-0.1410880089,0.0289778188,0.0751975402,0.1054384559,0.0982205346,-0.373752594,-0.1792002171,-0.0816590339,0.0180538408,0.250085026,0.0515308417,0.2197839767,-0.2101409733,0.194371298,0.3538516462,0.1270569265,0.1666173786,-0.1615899205,0.0003354756,-0.208024025,-0.1089843735,-0.0246526599,-0.0357160531,-0.1357513517,-0.3393703699,0.1776392013,0.2876363397,0.203537181,-0.4313399792,0.1474874318,0.0751998648,0.125007838,-0.2347596735,-0.1022746414,-0.10592594,0.0317357667,0.3103686273,0.2519684434,-0.2847986519,-0.2675506473,0.1913034618,-0.2784380019,0.0206548311,-0.0241041034,-0.2883765697,-0.0973786041,-0.3232779801,-0.1715882123,0.2962934971,-0.1423625946,-0.1365234852,0.082355015,0.0970979035,0.1753094494,0.1114523858,0.260116756,0.0136224134,0.0649727061,0.0734224617,0.3374759853,0.1691444963,0.1002798378,0.6567738056,-0.1104599237,0.0931354761,0.1282741576,-0.0122837927,0.1127998382,0.2854925096,0.132607162,-0.1569599956,-0.239324376,-0.107843101,-0.3550827205,0.0783154815,-0.0448400714,-0.1880510151,-0.1419253051,-0.1826450229,0.3122469485,0.1790773869,-0.1519476175,0.1807038635,0.7234703302,-0.1564106494,0.1299906671,0.3690554798,1.0367549658,0.0786096826,0.1230297834,0.0089302622,0.0527470782,0.7456656098,-0.1144990325,-0.1441476047,-0.357393086,-0.2628279328,-0.0622423179,-0.1339672804,0.095784612,0.0564467721,0.0164953005,-0.0189593304,0.207292214,0.1737813354,-0.1430290788,0.2870823443,0.0599219985,-0.1960026473,0.2088269293,0.184065029,-0.1905962676,0.2169747353,0.1216054708,-0.0859945193,0.2198132128,-0.1362178773,-0.2206011117,-0.0281082895,-0.2595712543,0.3074919283,0.150179863,0.0294347126,0.3491312563,-0.0557836555,0.4425501823,0.1760608554,-0.140961498,-0.0623013005,0.151614964,-0.1039143279,-0.1533171386,-0.1062075123,0.370236814,0.1290439069,-0.2317489982,-0.1106465459,0.0887440518,-0.3262949288,0.0484152734,-0.3035492003,-0.2184044719,-0.2886973023,0.0248752609,0.2225869894,0.041234117,0.0508071221,0.1517779082,0.0382645726,-0.1730937958,0.2421165109,0.0426589213,-0.3017169535,0.1082696691,0.2322665453,0.1175742373,0.1416342109,-0.1748856157,-0.1543360949,-0.1903945953,-0.2253008187,-0.0359266251,-0.1865978837,-0.1278928518,-0.0533694588,0.0019444082,-0.1060341224,-0.3190714121,0.2784805596,0.1176414788,0.2540142238,0.029478861,-0.1734063774,-0.2294342816,0.4788809419,-0.5582040548,0.2304531485,0.161242947,0.015269882,0.083411105,0.2891397774,-0.4294477999,-0.1126134843,-0.0422561541,-0.0008414167,0.4520377815,0.0395564921,0.1620170325,-0.1332273483,0.0221148022,0.0787496045,-0.3577288389,-0.2111248523,-0.1421762109,0.1017502025,0.1264156103,0.1620751172,0.0904510915,-0.0684406608,-0.0206578001,0.1031246185,0.4096180499,0.2586770058,0.1639599502,0.2574076653,-0.2574314475,0.0512582883,0.007144107,0.1007031426,-0.0765478909,0.2108139992,-0.0785492659,-0.0009626267,-0.2708190978,-0.1260650009,-0.0733054131,0.2384915501,0.1439597607,0.2079132795,-0.0476420671,0.2139858305,-0.0866488516,0.1197579205,0.231268242,-0.1606967747,-0.2761586308,-0.3324230909,0.0647148415,0.2520394325,-0.1905215085,-0.0617568232,0.0368080437,0.1374163926,0.0982210636,0.3613865972,0.7539802194,0.3685592711,0.1130199879,0.0363460444,0.6952106357,-0.3063836992,0.1092279777,-0.1038421616,-0.2854996324,-0.0144362357,0.3212291598,0.1427951157,0.3815550208,-0.0452330559,0.1222537905,0.3876748681,0.3558711708,-0.0638079941,0.3027633131,-0.1111487076,0.6664072275,0.1474333107,0.0913889259,-0.2211162746,0.0329661965,-0.0991560444,0.0657408535,0.0010138151,-0.1221516207,0.0528853238,-0.0396427438,-0.16328381,0.0566708967,0.0481022038,0.0374542698,-0.1255088896,0.0347319096,-0.2353889346,0.3433109224,-0.2406151146,-0.1162596345,-0.2253903151,0.2782294452,0.0665475652,0.3283776045,-0.0249047633,-0.0077510038,-0.2384950519,-0.0517232828,0.2002740055,0.7102639079,0.2245142758,0.1078248024,0.2929628491,-0.2177311927,-0.1748894602,-0.0290930886,0.031633202,0.2168760896,-0.0662583187,-0.2111458778,0.5035734177,0.1081087142,-0.0754254311,-0.0261127446,0.2137738913,0.0172242634,-0.0072238375,0.1925745159,0.0698869675,-0.0549637713,0.0361170173,0.0007773835,-0.264354974,-0.1431745738,0.7343968153,-0.2106710374,0.1840776205,-0.3899570107,0.1711172611,0.2917602956,0.4459947348,0.5310922265,-0.2351712286,-0.2693157792,0.1810176224,-0.8228289485,0.3952476084,0.0089494567,-0.0218157973,0.2428773493,0.2482739836,0.0618534908,0.3399040997,0.0026374878,-0.0453979038,-0.3527440727,-0.2476007938,-0.0640814826,-0.0895553306,-0.4320930541,-0.1581158787,-0.0024585817,-0.063932322,0.0824957788,0.0692600161,0.0386226699,-0.0147044603,0.147907123,0.2448121309,0.0497076362,0.4475648105,0.1053285375,0.175612852,0.0448156595,-0.0231368709,-0.1482611448,-0.060125228,-0.2957175672,0.1472607702,0.0451177694,-0.0438434742,-0.1834724247,-0.1543112099,-0.3483727276,-0.1378935277,0.1143561751,0.1803752184,-0.3972219229,0.0922557712,-0.0055910121,0.1035329029,0.0397807248,-0.0528207086,0.0071038874,0.2563713789,-0.1417813897,-0.5460961461,0.3538501859,-0.2556179464,0.0452880599,-0.0195357949,0.4716043174,-0.111215651,0.00824349,-0.9045564532,0.0236185286,0.3609058857,-0.1116673574,-0.2511820197,0.1106984019,-0.1293355823,-0.1062828228,-0.0434472226,-0.2299886644,-0.0897579119,-0.285236001,-0.0782153904,-0.3147262335]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/149","title":"[Feature request] Add Ubuntu Dialogue Corpus dataset","comments":"@AlphaMycelium the Ubuntu Dialogue Corpus [version 2]( https:\/\/github.com\/rkadlec\/ubuntu-ranking-dataset-creator) is added. Note that it requires a manual download by following the download instructions in the [repos]( https:\/\/github.com\/rkadlec\/ubuntu-ranking-dataset-creator).\r\nMaybe we can close this issue for now?","body":"https:\/\/github.com\/rkadlec\/ubuntu-ranking-dataset-creator or http:\/\/dataset.cs.mcgill.ca\/ubuntu-corpus-1.0\/","comment_length":34,"text":"[Feature request] Add Ubuntu Dialogue Corpus dataset\nhttps:\/\/github.com\/rkadlec\/ubuntu-ranking-dataset-creator or http:\/\/dataset.cs.mcgill.ca\/ubuntu-corpus-1.0\/\n@AlphaMycelium the Ubuntu Dialogue Corpus [version 2]( https:\/\/github.com\/rkadlec\/ubuntu-ranking-dataset-creator) is added. Note that it requires a manual download by following the download instructions in the [repos]( https:\/\/github.com\/rkadlec\/ubuntu-ranking-dataset-creator).\r\nMaybe we can close this issue for now?","embeddings":[-0.1457340121,0.2197993249,-0.065370962,-0.0233403649,0.1595499516,0.2317842543,0.2344495654,0.1404034793,-0.093551524,0.2077977061,-0.1967509389,0.2113592923,0.0908540338,-0.0198894702,-0.0083212974,-0.283608079,-0.0770926997,-0.0565560795,-0.0296263378,-0.1516617835,-0.5537121296,0.192510739,-0.3070515692,-0.0567773059,0.1344804913,-0.1004216671,0.0116874911,0.0735562816,-0.1892565042,-0.4569538534,-0.0369172283,0.2100033462,-0.3256537616,0.1866213679,-0.0000998173,-0.2117406577,0.305957973,0.0414143987,-0.3893639445,-0.3456751704,-0.1688473374,0.0708190203,-0.1088221669,0.0231406745,-0.3398527801,-0.028168736,0.0886452049,-0.2531458735,-0.1794046909,0.1825007796,0.3645730019,0.0986470953,0.0132904667,-0.5131924748,0.0464070328,-0.0297200866,-0.3338403106,0.1761569828,0.3506742418,0.234850511,-0.2275696397,0.3493562043,-0.0850725919,-0.0861073732,-0.1641721427,-0.1617786288,0.2012357861,-0.2212982625,0.2724966407,0.0855337828,0.8510781527,-0.2054710537,-0.1884159744,0.0553838126,0.2368637025,-0.1846892387,0.0259006098,0.0517957173,-0.1399551183,-0.0659550726,-0.1907348484,-0.2145889848,-0.1629605591,0.1234750226,0.0775852725,0.2758606076,-0.0982586443,-0.1178449169,-0.0648626909,-0.0285703354,0.1747392416,0.0444263257,0.0044974941,-0.0994881392,-0.4505813122,-0.1421777755,0.293445915,-0.013916715,0.1805759519,0.0792424455,-0.068962127,0.1509007066,0.2425786257,-0.1074475646,0.1982827485,0.1836605072,0.2357859761,0.1007084846,0.2788398862,0.1521478742,0.0445494503,0.0953177065,-0.1455168277,-0.1416531205,0.0203809589,0.2877107859,0.0293086581,-0.1949759722,-0.2099930197,0.0331568345,-0.010143891,0.1828418821,0.095018059,0.3426792026,-0.0702010691,0.0008310307,0.0004483556,0.1347321272,-0.1633306742,-0.237931639,-0.1635294855,0.0435231775,-0.3320834339,0.0789924711,0.3352038264,0.1977042109,0.0810485184,-0.0731780455,0.6526392102,-0.0325590707,0.0887615308,-0.0946499631,-0.1276352108,-0.0256183911,0.0849453881,0.1249081716,0.1755470335,0.1631166935,-0.1329624206,0.1215442047,-0.097945407,-0.2292851061,-0.1485717297,0.354133904,0.2843053341,-0.345941186,-0.2182160169,0.1816513538,0.0304691866,-0.1242463291,0.0360024199,0.1793025285,-0.5638617277,-0.1198016629,-0.1937254369,0.1882115752,-0.144085601,-0.0578094311,-0.3757381141,-0.0396629199,-0.0179012269,0.0567414649,-0.0736379698,0.1494026929,-0.145664528,0.2531201541,-0.0436911695,-0.4407576323,-0.4889676571,0.1472562701,-0.179641068,-0.2009970099,0.0854589269,0.0793593824,0.4515788853,-0.2775592804,0.1006058455,0.225910306,0.2275553942,0.2256549746,-0.0608956106,-0.1173453256,-0.2582629621,0.1127173305,0.1965082437,0.1305604875,0.0260700583,0.0918694586,0.1809975207,-0.1875294745,0.1367051005,0.1900651455,0.4828184545,-0.0919792801,0.3573668003,-0.1160516739,-0.0181473736,-0.0106611922,-0.2372460961,0.3643128574,0.0819884092,-0.3734312356,-0.3915190399,0.0386187695,-0.0672150254,-0.1480859667,0.2796754837,0.1446676999,-0.2751811445,-0.0765579566,-0.1154840514,0.0268470794,-0.0738224313,0.0028135788,0.0991961807,0.0292198714,-0.0367991515,0.0031727643,0.3169911504,0.1818198711,0.0539361499,0.2323714942,-0.1097493693,0.3216853738,-0.053882625,0.1996225119,0.0129265524,-0.0571823306,0.0831375122,-0.2413344234,0.1716485471,0.314267993,-0.1589084417,0.0413433015,-0.1335089952,0.0982968137,0.243168816,0.01510973,0.1955194473,0.1304554492,0.195156917,-0.2241234481,0.1276551783,-0.319267869,0.0511935018,0.2708053291,0.1071858108,-0.0309547093,-0.2921259701,0.2595498264,0.3622592986,-0.0280064922,0.3814353645,0.2174120396,-0.1518546194,-0.1470986158,0.1664779484,-0.2648390532,0.0843152478,0.3542546928,0.2902134359,0.0116119152,0.0655246228,-0.3486016095,0.0931589082,-0.0934586301,-0.1122279614,0.4130340815,0.2647446692,-0.0552241988,-0.5771736503,0.1376950294,-0.0525519997,0.0329414345,-0.2305874079,-0.2867489457,-0.1552271843,-0.4819123745,0.2490713596,0.0291622411,0.1454894543,-0.109712936,0.4188368618,0.0285632256,-0.0700844377,0.471575737,0.2065003365,0.4535330236,-0.0588373616,0.0771489516,0.2722302973,-0.2450772971,-0.084478192,0.3063847125,-0.2414114773,-0.026740944,0.5109686255,0.054128021,0.2273268104,-0.3986753225,-0.2502072752,0.2745244503,0.2686484456,0.2051804364,0.145625487,-0.1614298075,-0.2069333941,-0.1343294382,-0.1148727834,-0.0758794174,-0.3143543303,0.0239254441,-0.0332681946,-0.1479963511,-0.1009343937,-0.4821455181,0.2255835384,-0.5429046154,-0.0721463412,-0.0649805292,0.1399098039,0.0545005463,-0.1080094203,0.0203583539,0.2380866855,0.2772634327,-0.2891526222,-0.0476002768,0.3393503726,-0.4837678969,-0.3549007177,-0.2117679715,-0.1903628558,0.1838318259,0.0908524692,-0.1896077245,-0.0120774899,0.0065510469,0.0783627927,0.3021087945,-0.0885489807,0.3156242073,0.0954385102,-0.3790741563,-0.2183194607,-0.0315257944,0.043476779,0.0937239602,0.2295888662,-0.1382554621,0.1959206462,0.1474936754,0.2408270687,0.2726469338,0.3439339995,0.1083914489,0.1125005335,0.1874090582,-0.1723178774,-0.2078281194,0.2516317964,0.0688262731,-0.0784474984,0.5799384117,0.3527544737,-0.3229682148,-0.0441805422,-0.2059842944,-0.094933778,-0.2647589147,-0.1428055614,-0.0631783754,0.4356000423,-0.0135947661,0.041785989,-0.1648344845,-0.2783921659,0.3895579278,-0.051098194,0.0315871164,-0.3228036165,0.2417618483,0.5417657495,-0.3342849314,0.2472952902,0.1395982355,-0.2831557691,-0.2363410294,-0.031259384,0.0474420711,0.1733107865,0.4277935922,-0.384716928,0.1248625815,0.1616610736,0.0252456497,0.2655962408,0.3384222388,-0.1732677668,0.0109521979,0.0534408167,0.3378056884,-0.2153678387,0.3740749955,0.2895160913,-0.1248382777,-0.0601081587,-0.26088202,-0.0309962388,-0.0748172551,-0.0637135655,0.0653591082,-0.1835860908,-0.0489762612,0.1201496199,-0.0919861794,-0.0969497338,-0.1271463037,0.1077818722,-0.0239680726,0.0495240912,0.3088514209,0.2255610228,-0.0301389433,-0.2712610066,0.2891284227,0.2875470519,0.0030461396,-0.3202451169,-0.0187487602,-0.2834385037,-0.0177802201,0.2611631751,-0.0801034793,0.1032154337,-0.2141304463,0.038524773,0.0321226381,-0.2263955921,0.1077724025,-0.0683878213,-0.2188723236,-0.4211242795,0.3298643827,-0.1022923142,-0.2822283804,0.1920567602,0.0519663244,-0.2273859233,0.105306007,0.2425385118,0.9062684774,-0.1012231708,-0.145975098,0.2176080197,0.0047934144,0.3197511733,0.0895394385,0.3081050813,-0.1948556304,-0.062122751,0.0099997753,0.0492278673,-0.0072252597,0.2924642861,-0.3950828314,0.066003561,-0.0555661395,-0.0773386955,0.059215121,0.0383181237,0.0392659083,-0.3930009007,-0.0438488238,0.3037889004,0.1409310997,0.013818698,-0.0424112082,0.0637060106,0.1813243777,-0.0245167967,-0.3403656483,0.1083220094,-0.0531971492,0.2331746966,0.1822421402,-0.2906413078,0.3123782277,0.2073211223,0.3979808986,0.3287528157,-0.2152677625,0.1981971264,0.4791021049,0.0006112132,0.1144865006,0.0868551582,0.3496516645,-0.2161878198,-0.2021831274,0.025055116,0.2770274282,-0.2289780676,-0.3953163922,-0.1918429434,-0.033247225,-0.0389100127,0.2239557505,0.082939066,0.0776084661,-0.3248807192,0.2740041316,-0.0897005275,0.0833290219,-0.0385384299,0.1145118847,-0.1745885015,-0.1383467168,0.3227700591,0.2980443537,-0.2873389423,0.0797953233,0.3821919858,-0.3586542606,-0.4391874969,-0.3443102241,-0.1838161349,-0.0743341148,0.2853102982,0.0049283504,-0.0564049706,0.0785309523,0.1789056361,0.012647667,0.125992626,-0.1419865638,0.0705614686,0.0192044768,0.0071516116,-0.1402675509,-0.176906839,-0.208302334,-0.1948184073,0.2179618925,-0.0506890602,-0.4734625518,-0.1639012396,-0.3634221256,-0.0077872365,-0.0267494824,0.0167206358,-0.0337219462,0.0006393655,0.2583466768,-0.2795390487,-0.2122946233,-0.3477820158,-0.3026463389,0.0763725862,-0.0170776863,0.1839533448,0.1910753697,-0.064907223,-0.1648267657,-0.1892855167,0.2007458955,0.3675888479,-0.037102595,-0.0127657559,-0.1142714545,0.0583863258,0.2455140054,0.096690312,0.0115866745,-0.0399800241,-0.0249286275,-0.1150882021,0.0173326433,0.0672490224,0.0794697404,0.1012893245,0.2966344357,-0.2234760523,0.2667152286,-0.0709162131,0.1485743821,0.2088928968,0.3306438625,-0.1091108993,-0.0084630484,-0.1643293798,-0.0769793242,0.352930814,-0.4000604749,-0.1541776806,-0.3860528767,-0.2121091932,0.1627434939,0.1340810061,-0.0262010228,-0.0305801332,0.3415963948,0.1099040881,0.2402597666,-0.3467566371,0.0539669022,0.1017192081,0.1444538981,0.0434430875,0.2272928655,0.023086274,0.1974098831,-0.0081568453,0.3512206078,0.22099711,0.08935716,0.2808228135,0.1898697168,0.0683039427,0.3816756606,0.3104638457,-0.1492519826,-0.0836074054,0.1683249772,0.075111039,-0.0174036734,0.1823724955,-0.2535954714,0.0285787303,-0.1589749455,-0.1432743073,-0.3131225109,-0.2097185999,-0.2728105187,-0.02302761,-0.1024280787,-0.023780413,-0.1523753405,-0.2327272594,0.0678313747,-0.4618993104,0.2440065742,-0.1306282133,0.2266205102,-0.2389585674,-0.0932062641,-0.1087381765,0.1765353382,0.1268898994,0.5840149522,0.0268169828,0.207146734,-0.0601523817,-0.1499152035,-0.0992480069,-0.0444229953,-0.1103360802,0.1674551517,0.1072185338,0.3521735072,0.1139855012,0.2665700018,-0.2999924421,0.4324698448,0.1590471864,-0.0051832017,-0.1671378314,0.3814975917,-0.1099770963,-0.1448549628,0.0426694378,-0.1786764115,-0.3601490855,-0.1388969868,0.5247937441,-0.1214096248,-0.0610766485,-0.3290764689,0.2009472996,0.0469033271,0.258603543,0.2924341857,-0.0568693541,0.0858515501,-0.1661030352,-0.4019676149,0.4331298172,0.1938847154,0.0503123067,0.0095250374,0.012734253,0.0714429989,0.4110077024,-0.3716356158,-0.3558700383,-0.4108234346,-0.0205852706,-0.4024207294,0.0023752037,0.2220056355,0.0762670338,-0.0202505924,-0.1922042519,-0.1259808838,-0.0466455966,0.2020660192,-0.0103786495,-0.1062001064,0.2348825037,-0.0697827712,0.0883780867,0.036886435,0.3651401103,-0.1855439693,0.0432211384,0.0604685172,-0.3253747523,-0.1535695344,0.1066719517,0.106730327,0.1684603244,0.0724487305,-0.1535844952,0.1489978284,0.082560204,0.1491424739,-0.0495881811,-0.2522642314,0.2731970847,0.0124631291,-0.0587328859,-0.1050488651,0.0041218172,-0.0167050697,-0.0452459194,-0.0455753021,-0.6543854475,0.1914206743,-0.4609658122,-0.1696773022,-0.1917192042,0.4397036731,-0.1815086156,-0.4255829453,-0.2430221587,0.1648899615,0.2478934675,0.1533915848,-0.1385324895,0.2155104429,-0.2603447139,0.1555553675,-0.1125279963,0.374379456,0.148469016,-0.1316868961,-0.2542992532,-0.1657210737]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/148","title":"_download_and_prepare() got an unexpected keyword argument 'verify_infos'","comments":"Same error for dataset 'wiki40b'","body":"# Reproduce\r\nIn Colab,\r\n```\r\n%pip install -q nlp\r\n%pip install -q apache_beam mwparserfromhell\r\n\r\ndataset = nlp.load_dataset('wikipedia')\r\n```\r\nget\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wikipedia\/20200501.aa\/1.0.0...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-6-52471d2a0088> in <module>()\r\n----> 1 dataset = nlp.load_dataset('wikipedia')\r\n\r\n1 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 515 download_mode=download_mode,\r\n 516 ignore_verifications=ignore_verifications,\r\n--> 517 save_infos=save_infos,\r\n 518 )\r\n 519 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 361 verify_infos = not save_infos and not ignore_verifications\r\n 362 self._download_and_prepare(\r\n--> 363 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 364 )\r\n 365 # Sync info\r\n\r\nTypeError: _download_and_prepare() got an unexpected keyword argument 'verify_infos'\r\n```","comment_length":5,"text":"_download_and_prepare() got an unexpected keyword argument 'verify_infos'\n# Reproduce\r\nIn Colab,\r\n```\r\n%pip install -q nlp\r\n%pip install -q apache_beam mwparserfromhell\r\n\r\ndataset = nlp.load_dataset('wikipedia')\r\n```\r\nget\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wikipedia\/20200501.aa\/1.0.0...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-6-52471d2a0088> in <module>()\r\n----> 1 dataset = nlp.load_dataset('wikipedia')\r\n\r\n1 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 515 download_mode=download_mode,\r\n 516 ignore_verifications=ignore_verifications,\r\n--> 517 save_infos=save_infos,\r\n 518 )\r\n 519 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 361 verify_infos = not save_infos and not ignore_verifications\r\n 362 self._download_and_prepare(\r\n--> 363 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 364 )\r\n 365 # Sync info\r\n\r\nTypeError: _download_and_prepare() got an unexpected keyword argument 'verify_infos'\r\n```\nSame error for dataset 'wiki40b'","embeddings":[-0.1235572174,-0.0397848673,-0.0055092555,0.1611870676,0.0169578884,-0.1350489408,0.0657530054,0.4551704824,0.1232152507,0.021626275,-0.03391321,0.3797163069,-0.0624290854,-0.1007759199,0.1021122187,-0.2043866515,-0.0299466662,0.3831298947,-0.2216906548,0.0933285505,-0.2810808718,0.3655325472,-0.2893398404,0.0260600615,-0.0329860486,-0.16864115,-0.0508995838,-0.0743941888,-0.4602272213,-0.547026217,0.0951261222,0.0306446236,0.1206287369,0.0053344509,-0.0001163738,-0.0381661132,0.4436964095,-0.2771382034,-0.7427003384,-0.1261380613,-0.3340919018,-0.2321169674,0.2440309674,-0.4157205522,0.1843524277,-0.0059965216,0.075056389,0.3486465514,0.4171301425,0.4644598961,0.200536415,0.2971923649,0.1138166189,0.0195616204,0.27884534,0.0641594753,-0.0619661398,0.2426051497,-0.1783657372,-0.0709384382,-0.0052235937,0.4772388339,-0.2131444365,0.0855332315,0.1330017447,-0.1308964044,0.2469562739,-0.3498190939,0.0208939873,-0.1071821302,0.1359665692,-0.1527432352,-0.0982338935,-0.1605093777,-0.0864290595,-0.04295801,0.3078669608,0.1401110888,-0.3331155479,-0.1578291953,-0.1270762235,-0.0841358155,0.0353925563,0.1835552305,0.3300491571,0.5899922848,-0.0497779772,0.1509763151,0.4151127934,-0.0357216299,0.0207407828,-0.4068606496,-0.1482583433,0.4026424289,0.049302239,-0.055824995,0.0863497257,0.2894313633,0.1310058236,-0.0609081164,-0.1524266154,-0.2973771691,0.2357500345,0.2540418506,-0.115595974,0.3927487731,0.2608386874,-0.1045859084,-0.0721451491,0.4366670251,-0.0106147798,0.0045231227,0.0814360529,-0.2072258741,-0.1146633253,0.1720440686,0.3206189871,-0.075715512,-0.2830899954,-0.0527558997,-0.3759811223,-0.2118949443,0.0546104163,0.2799731195,-0.0596625991,0.1641170382,0.2523317337,0.0639396012,-0.1830996424,-0.355871588,-0.0060372939,0.5307106972,-0.3692677617,0.1088422984,0.2773804665,0.0297607947,0.3715512753,0.0710571632,0.1084663719,-0.1922290027,0.0081847114,-0.1136943549,0.0170879643,0.191891551,0.0751349553,0.2280997485,0.1702846438,-0.2731007636,-0.0559133478,0.0734336749,-0.1701537222,-0.3210542202,0.0298237447,0.1287616193,-0.1424731463,0.1064565182,-0.2300979048,-0.2174750268,0.2913390994,-0.1481356472,-0.0797236487,-0.2410677671,-0.1138117686,-0.3118925691,0.2701602876,0.3090493083,0.0610449389,-0.0764651075,-0.126256749,0.0815221891,0.1397793442,0.1771391183,-0.05673787,0.5519248247,-0.2873392701,-0.0218319762,0.5885543227,-0.5047643185,-0.1221137494,0.4325270951,0.2941926718,0.1136151776,0.0804054961,-0.0886105746,0.0824073032,-0.0841640383,0.136400938,0.2611859143,-0.0708677918,0.1725219935,-0.4314493239,-0.2597858608,-0.0033856754,-0.1935753971,0.1093644202,0.0988344029,-0.0665592104,0.4609386921,0.2081838399,-0.0398749746,-0.0668114498,0.0072144158,0.0813858882,-0.2442791611,-0.1745290607,-0.0502461195,-0.092803888,-0.0233252645,-0.2406600863,0.3740319312,-0.2776536644,-0.0398809612,-0.5766931176,0.0680673793,-0.0913610086,-0.2638744414,0.1735809147,0.2431659549,0.1676423103,0.2170452923,0.1381660104,0.2460132837,-0.0369323343,0.1189965606,-0.1959366351,0.0400885455,-0.530447185,-0.1659938693,0.0018557716,0.0367873162,0.4112120867,-0.0404937305,-0.2196119428,0.3641524911,-0.0062647909,-0.0896337926,-0.0294939,-0.1216243207,0.0175363477,0.0025386226,0.1115113422,-0.1192253157,0.2949276268,0.0110115567,0.133936435,0.2002941072,0.0966166705,0.1122702584,0.0558015853,0.019035155,0.1759367734,0.1094659939,-0.0207830779,-0.162983492,0.3710613549,0.287098974,-0.2384943664,-0.1333288252,0.2498748302,-0.2145897001,0.2947457731,0.1404752731,0.1344620138,0.0455050506,0.0054756431,-0.0899126083,0.0880696028,0.0852347538,0.5152243376,-0.0017243662,-0.1640182137,0.1468628347,-0.1637807786,-0.046249412,0.4427818656,0.1440772861,0.1238524616,0.0260960497,0.077842772,0.0450725779,-0.1442724168,-0.0836225078,0.0802928209,0.2956393361,-0.2856363952,0.2834445834,-0.1164082512,-0.6694436073,-0.2616386116,0.0013871792,-0.1963309348,-0.5623303652,-0.0344885066,0.2329371274,0.1686943322,0.3313954175,0.0539554842,0.0384349711,0.0201330297,-0.3091154099,-0.1174745113,-0.1735436469,-0.2210492641,-0.0196194369,0.2301711142,-0.091287151,0.4829075336,-0.2198187858,-0.0029718918,-0.0347283073,-0.4303334653,0.2199233323,-0.2484197468,0.2980496883,0.2876155078,0.2704826295,0.065573521,-0.3325261176,0.2070403397,-0.084687382,-0.1312547028,-0.182498768,-0.0159427449,-0.1466005296,-0.325193584,-0.2690192759,-0.1084341407,-0.2765624821,-0.1300962865,0.2832934856,0.1702609658,0.291947484,0.0225821566,0.1300492138,-0.0574910827,0.1404466331,-0.15466775,-0.1020002663,0.2008778602,-0.2406257391,-0.1484370083,0.0161580555,-0.2373637557,0.31983459,-0.0741384178,-0.4126252532,-0.4417607784,-0.1262500435,0.1026174724,-0.1170484275,0.030252317,0.1435332596,0.0594573803,0.0833234563,-0.1012964621,-0.358184278,-0.0216267277,0.0594813824,0.5545659661,-0.0118366294,0.2215062529,0.1181129441,0.8230003715,-0.0310095605,-0.0929989219,0.362737447,0.0134160174,0.1705204546,-0.2054594457,-0.3809252679,0.0194545221,-0.0674738362,0.0495414622,0.1275547296,-0.1995813996,-0.0910216793,-0.3627660871,0.0624519773,-0.4654777646,-0.3720410764,0.317986846,-0.0665973797,0.176747486,0.1028034166,0.200889945,-0.1752512157,-0.1498139799,-0.089904882,0.3439922333,-0.0096800849,-0.016205065,0.2878166139,-0.2316004783,-0.4206951857,0.3904573023,0.3236595094,0.2822977602,-0.0737732053,0.1631087661,0.0476977341,-0.0359379314,0.2337878644,-0.2132974565,0.037365295,0.1142063215,0.1863577366,-0.3354548216,-0.1879960597,-0.327027142,-0.0530778728,0.4943103194,0.1524488181,-0.4010412693,0.1333307624,0.3525338769,0.063237533,-0.154259786,-0.0000922625,-0.2981521785,-0.485463649,-0.4669547379,0.0269721281,0.1650410742,0.3092245162,-0.0053489166,0.5527027845,0.2564508617,0.2369553894,0.1115218475,0.1491895467,0.0742204115,0.1745903045,-0.1303040087,-0.4405701756,0.2134807855,-0.2293315828,0.173782438,-0.0487777181,0.0191855691,-0.3581656218,0.0430152267,0.0899861082,0.2455186397,-0.107349284,0.0280592442,-0.1718208194,-0.0809289664,0.0484767482,0.4078331888,0.3670092225,-0.161844328,-0.397736609,-0.3018179238,0.471419394,0.0427819677,-0.1380018294,0.1858424395,0.1023061275,-0.2666895688,0.365398556,0.09796758,1.0379098654,0.1603966504,0.0617936626,0.5151295066,-0.2044004947,0.7474738955,-0.2016752362,0.1778167486,-0.3029268384,0.0606557913,0.0282601472,-0.0470006838,0.1230213493,-0.2754657567,-0.6714712977,0.3089575768,0.1118707806,0.0441568308,0.0274046622,0.2391885519,-0.0861304104,-0.1027354375,-0.2655648291,0.1117288917,-0.3630501032,0.4866702557,-0.3031411469,-0.1454939842,0.0735564157,-0.5422187448,-0.4056447744,0.1347865164,-0.2200269252,0.3026019633,-0.1894472241,-0.0419460796,0.0287013575,0.0511229597,0.0494193584,0.2312056273,-0.0734436437,0.2083618641,-0.1264243275,-0.1591736972,-0.157670632,0.2229796499,0.1178622022,-0.1599719226,-0.1607328206,0.0287085604,-0.132377848,-0.0551206805,-0.0210096762,-0.0180984288,0.1744486988,-0.2757012844,-0.3031266928,-0.0291795898,-0.3162955046,-0.1168425232,0.0777494982,-0.140821442,-0.2035886943,0.2582942545,-0.004404407,-0.3116005659,0.0391671993,0.2989181578,-0.0587782227,-0.1035333499,0.4948186278,-0.0174766071,-0.2341255397,-0.2265408337,0.0059656831,-0.8320417404,-0.4196950197,0.0923220366,0.1404572725,0.0354364514,-0.1578514278,0.4367974997,0.230864495,0.2776599824,0.2272067666,-0.4679569304,-0.0105704796,0.101254575,-0.280731082,-0.0497216471,0.1797121316,0.3023636937,0.3777132928,0.2600378692,-0.2399078608,-0.1359714866,-0.2728255689,0.4363948703,0.4620951414,-0.0453722961,0.2471175194,0.1813136637,0.0653796345,0.2439451218,-0.1171858385,-0.1759735644,-0.1848390251,0.1749594957,0.1496821493,-0.0600156933,-0.0156633221,0.0482636467,0.1012092903,-0.1342943758,-0.2212226242,-0.1341987103,-0.1018091291,0.403714329,-0.0811604187,0.350672394,0.0903647467,0.2299607396,-0.2027548105,0.2579711378,0.2215570658,0.1295814812,0.0725676417,-0.1162279621,-0.3555429578,-0.1598296165,0.030582333,0.1808655709,0.0709812716,-0.0929418877,-0.1310455203,-0.3081482351,0.0658408925,0.2741805017,-0.1258384883,-0.1190997139,0.1458713859,0.0900861919,-0.1429093331,0.2187800854,0.3888690174,0.2636514604,0.0101662474,0.3116286099,-0.1266066134,0.0549860857,0.064990662,-0.0614425093,0.1470943987,-0.1211681664,-0.0370414108,0.2725234628,0.0567122214,0.1953904182,0.0161142703,-0.1064417809,0.1345743388,0.2616095841,-0.3864254057,0.259992063,0.0923803076,-0.0890370905,0.13858971,-0.1082067639,0.030560758,0.2163309604,0.1461498439,0.222229749,0.1549704522,0.2731763721,0.0785089582,-0.1611618251,-0.3665848374,0.364913702,-0.1428804994,-0.0714778602,-0.2635197639,-0.094346635,-0.2115145773,0.054637596,0.1483044773,-0.1082378924,0.3571384549,0.3224037886,-0.1065469906,-0.274723053,0.0086299479,0.2282196879,0.049541872,-0.2872061729,0.2931395173,0.1396940947,0.1063329875,-0.0078268209,0.2379901558,0.725264132,0.5168668032,-0.3115155697,-0.1636020094,-0.176910907,0.0911142901,-0.0807719231,0.0747022033,0.2983618677,0.4166287184,0.4595347941,0.1323583126,-0.1695204079,0.0517079942,0.1032559797,-0.321909517,-0.0616908297,0.5042173266,-0.0247077644,-0.0180153437,-0.2502755225,0.346177727,-0.4587026536,0.2066499889,-0.0599659234,0.2616643906,0.2048780918,-0.0402288064,0.0371435806,-0.0611800887,0.6178979874,0.3663236797,0.3425807953,-0.6118857265,-0.1754017174,-0.4910422266,-0.0601615161,-0.1484051943,-0.3703594506,-0.2915735245,0.0971980393,-0.0788731202,0.3884087205,-0.0742788315,-0.0682195649,0.1689771712,0.0846798047,-0.3359518051,-0.1068405658,0.031062372,-0.3000177443,-0.0120873703,-0.2438317984,0.0211333092,-0.2394023836,0.0584723428,-0.1020724401,-0.3599013388,0.347179085,0.3313420713,0.378216207,0.2259520888,0.8266328573,-0.1996865422,-0.1380767971,0.0664436743,-0.0932086334,0.0112206126,0.0206320137,0.0810936093,0.2432469726,-0.0941913053,0.3419338167,-0.5246486664,0.7450121641,-0.0435727648,-0.0832773969,-0.1926093102,-0.3051315844,-0.1723100841,0.2594204247,-0.1198933721,0.0273916256,-0.1963588744,0.3427009583,-0.4602262974,-0.415148735,0.5375388265,-0.3625352383,-0.388100028,0.0640936047,0.2020009309,-0.3227238953,-0.1909432262,-0.7663069367,-0.0921797231,0.2804146111,-0.1019756794,-0.2139068246,0.1753968,-0.0223213769,0.0938506499,-0.0044266516,-0.0552640632,-0.0458026119,-0.254857868,-0.1172515601,-0.2409612238]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/148","title":"_download_and_prepare() got an unexpected keyword argument 'verify_infos'","comments":"Should be fixed on master :)","body":"# Reproduce\r\nIn Colab,\r\n```\r\n%pip install -q nlp\r\n%pip install -q apache_beam mwparserfromhell\r\n\r\ndataset = nlp.load_dataset('wikipedia')\r\n```\r\nget\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wikipedia\/20200501.aa\/1.0.0...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-6-52471d2a0088> in <module>()\r\n----> 1 dataset = nlp.load_dataset('wikipedia')\r\n\r\n1 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 515 download_mode=download_mode,\r\n 516 ignore_verifications=ignore_verifications,\r\n--> 517 save_infos=save_infos,\r\n 518 )\r\n 519 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 361 verify_infos = not save_infos and not ignore_verifications\r\n 362 self._download_and_prepare(\r\n--> 363 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 364 )\r\n 365 # Sync info\r\n\r\nTypeError: _download_and_prepare() got an unexpected keyword argument 'verify_infos'\r\n```","comment_length":6,"text":"_download_and_prepare() got an unexpected keyword argument 'verify_infos'\n# Reproduce\r\nIn Colab,\r\n```\r\n%pip install -q nlp\r\n%pip install -q apache_beam mwparserfromhell\r\n\r\ndataset = nlp.load_dataset('wikipedia')\r\n```\r\nget\r\n```\r\nDownloading and preparing dataset wikipedia\/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to \/root\/.cache\/huggingface\/datasets\/wikipedia\/20200501.aa\/1.0.0...\r\n\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-6-52471d2a0088> in <module>()\r\n----> 1 dataset = nlp.load_dataset('wikipedia')\r\n\r\n1 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 515 download_mode=download_mode,\r\n 516 ignore_verifications=ignore_verifications,\r\n--> 517 save_infos=save_infos,\r\n 518 )\r\n 519 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 361 verify_infos = not save_infos and not ignore_verifications\r\n 362 self._download_and_prepare(\r\n--> 363 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 364 )\r\n 365 # Sync info\r\n\r\nTypeError: _download_and_prepare() got an unexpected keyword argument 'verify_infos'\r\n```\nShould be fixed on master :)","embeddings":[-0.1235572174,-0.0397848673,-0.0055092555,0.1611870676,0.0169578884,-0.1350489408,0.0657530054,0.4551704824,0.1232152507,0.021626275,-0.03391321,0.3797163069,-0.0624290854,-0.1007759199,0.1021122187,-0.2043866515,-0.0299466662,0.3831298947,-0.2216906548,0.0933285505,-0.2810808718,0.3655325472,-0.2893398404,0.0260600615,-0.0329860486,-0.16864115,-0.0508995838,-0.0743941888,-0.4602272213,-0.547026217,0.0951261222,0.0306446236,0.1206287369,0.0053344509,-0.0001163738,-0.0381661132,0.4436964095,-0.2771382034,-0.7427003384,-0.1261380613,-0.3340919018,-0.2321169674,0.2440309674,-0.4157205522,0.1843524277,-0.0059965216,0.075056389,0.3486465514,0.4171301425,0.4644598961,0.200536415,0.2971923649,0.1138166189,0.0195616204,0.27884534,0.0641594753,-0.0619661398,0.2426051497,-0.1783657372,-0.0709384382,-0.0052235937,0.4772388339,-0.2131444365,0.0855332315,0.1330017447,-0.1308964044,0.2469562739,-0.3498190939,0.0208939873,-0.1071821302,0.1359665692,-0.1527432352,-0.0982338935,-0.1605093777,-0.0864290595,-0.04295801,0.3078669608,0.1401110888,-0.3331155479,-0.1578291953,-0.1270762235,-0.0841358155,0.0353925563,0.1835552305,0.3300491571,0.5899922848,-0.0497779772,0.1509763151,0.4151127934,-0.0357216299,0.0207407828,-0.4068606496,-0.1482583433,0.4026424289,0.049302239,-0.055824995,0.0863497257,0.2894313633,0.1310058236,-0.0609081164,-0.1524266154,-0.2973771691,0.2357500345,0.2540418506,-0.115595974,0.3927487731,0.2608386874,-0.1045859084,-0.0721451491,0.4366670251,-0.0106147798,0.0045231227,0.0814360529,-0.2072258741,-0.1146633253,0.1720440686,0.3206189871,-0.075715512,-0.2830899954,-0.0527558997,-0.3759811223,-0.2118949443,0.0546104163,0.2799731195,-0.0596625991,0.1641170382,0.2523317337,0.0639396012,-0.1830996424,-0.355871588,-0.0060372939,0.5307106972,-0.3692677617,0.1088422984,0.2773804665,0.0297607947,0.3715512753,0.0710571632,0.1084663719,-0.1922290027,0.0081847114,-0.1136943549,0.0170879643,0.191891551,0.0751349553,0.2280997485,0.1702846438,-0.2731007636,-0.0559133478,0.0734336749,-0.1701537222,-0.3210542202,0.0298237447,0.1287616193,-0.1424731463,0.1064565182,-0.2300979048,-0.2174750268,0.2913390994,-0.1481356472,-0.0797236487,-0.2410677671,-0.1138117686,-0.3118925691,0.2701602876,0.3090493083,0.0610449389,-0.0764651075,-0.126256749,0.0815221891,0.1397793442,0.1771391183,-0.05673787,0.5519248247,-0.2873392701,-0.0218319762,0.5885543227,-0.5047643185,-0.1221137494,0.4325270951,0.2941926718,0.1136151776,0.0804054961,-0.0886105746,0.0824073032,-0.0841640383,0.136400938,0.2611859143,-0.0708677918,0.1725219935,-0.4314493239,-0.2597858608,-0.0033856754,-0.1935753971,0.1093644202,0.0988344029,-0.0665592104,0.4609386921,0.2081838399,-0.0398749746,-0.0668114498,0.0072144158,0.0813858882,-0.2442791611,-0.1745290607,-0.0502461195,-0.092803888,-0.0233252645,-0.2406600863,0.3740319312,-0.2776536644,-0.0398809612,-0.5766931176,0.0680673793,-0.0913610086,-0.2638744414,0.1735809147,0.2431659549,0.1676423103,0.2170452923,0.1381660104,0.2460132837,-0.0369323343,0.1189965606,-0.1959366351,0.0400885455,-0.530447185,-0.1659938693,0.0018557716,0.0367873162,0.4112120867,-0.0404937305,-0.2196119428,0.3641524911,-0.0062647909,-0.0896337926,-0.0294939,-0.1216243207,0.0175363477,0.0025386226,0.1115113422,-0.1192253157,0.2949276268,0.0110115567,0.133936435,0.2002941072,0.0966166705,0.1122702584,0.0558015853,0.019035155,0.1759367734,0.1094659939,-0.0207830779,-0.162983492,0.3710613549,0.287098974,-0.2384943664,-0.1333288252,0.2498748302,-0.2145897001,0.2947457731,0.1404752731,0.1344620138,0.0455050506,0.0054756431,-0.0899126083,0.0880696028,0.0852347538,0.5152243376,-0.0017243662,-0.1640182137,0.1468628347,-0.1637807786,-0.046249412,0.4427818656,0.1440772861,0.1238524616,0.0260960497,0.077842772,0.0450725779,-0.1442724168,-0.0836225078,0.0802928209,0.2956393361,-0.2856363952,0.2834445834,-0.1164082512,-0.6694436073,-0.2616386116,0.0013871792,-0.1963309348,-0.5623303652,-0.0344885066,0.2329371274,0.1686943322,0.3313954175,0.0539554842,0.0384349711,0.0201330297,-0.3091154099,-0.1174745113,-0.1735436469,-0.2210492641,-0.0196194369,0.2301711142,-0.091287151,0.4829075336,-0.2198187858,-0.0029718918,-0.0347283073,-0.4303334653,0.2199233323,-0.2484197468,0.2980496883,0.2876155078,0.2704826295,0.065573521,-0.3325261176,0.2070403397,-0.084687382,-0.1312547028,-0.182498768,-0.0159427449,-0.1466005296,-0.325193584,-0.2690192759,-0.1084341407,-0.2765624821,-0.1300962865,0.2832934856,0.1702609658,0.291947484,0.0225821566,0.1300492138,-0.0574910827,0.1404466331,-0.15466775,-0.1020002663,0.2008778602,-0.2406257391,-0.1484370083,0.0161580555,-0.2373637557,0.31983459,-0.0741384178,-0.4126252532,-0.4417607784,-0.1262500435,0.1026174724,-0.1170484275,0.030252317,0.1435332596,0.0594573803,0.0833234563,-0.1012964621,-0.358184278,-0.0216267277,0.0594813824,0.5545659661,-0.0118366294,0.2215062529,0.1181129441,0.8230003715,-0.0310095605,-0.0929989219,0.362737447,0.0134160174,0.1705204546,-0.2054594457,-0.3809252679,0.0194545221,-0.0674738362,0.0495414622,0.1275547296,-0.1995813996,-0.0910216793,-0.3627660871,0.0624519773,-0.4654777646,-0.3720410764,0.317986846,-0.0665973797,0.176747486,0.1028034166,0.200889945,-0.1752512157,-0.1498139799,-0.089904882,0.3439922333,-0.0096800849,-0.016205065,0.2878166139,-0.2316004783,-0.4206951857,0.3904573023,0.3236595094,0.2822977602,-0.0737732053,0.1631087661,0.0476977341,-0.0359379314,0.2337878644,-0.2132974565,0.037365295,0.1142063215,0.1863577366,-0.3354548216,-0.1879960597,-0.327027142,-0.0530778728,0.4943103194,0.1524488181,-0.4010412693,0.1333307624,0.3525338769,0.063237533,-0.154259786,-0.0000922625,-0.2981521785,-0.485463649,-0.4669547379,0.0269721281,0.1650410742,0.3092245162,-0.0053489166,0.5527027845,0.2564508617,0.2369553894,0.1115218475,0.1491895467,0.0742204115,0.1745903045,-0.1303040087,-0.4405701756,0.2134807855,-0.2293315828,0.173782438,-0.0487777181,0.0191855691,-0.3581656218,0.0430152267,0.0899861082,0.2455186397,-0.107349284,0.0280592442,-0.1718208194,-0.0809289664,0.0484767482,0.4078331888,0.3670092225,-0.161844328,-0.397736609,-0.3018179238,0.471419394,0.0427819677,-0.1380018294,0.1858424395,0.1023061275,-0.2666895688,0.365398556,0.09796758,1.0379098654,0.1603966504,0.0617936626,0.5151295066,-0.2044004947,0.7474738955,-0.2016752362,0.1778167486,-0.3029268384,0.0606557913,0.0282601472,-0.0470006838,0.1230213493,-0.2754657567,-0.6714712977,0.3089575768,0.1118707806,0.0441568308,0.0274046622,0.2391885519,-0.0861304104,-0.1027354375,-0.2655648291,0.1117288917,-0.3630501032,0.4866702557,-0.3031411469,-0.1454939842,0.0735564157,-0.5422187448,-0.4056447744,0.1347865164,-0.2200269252,0.3026019633,-0.1894472241,-0.0419460796,0.0287013575,0.0511229597,0.0494193584,0.2312056273,-0.0734436437,0.2083618641,-0.1264243275,-0.1591736972,-0.157670632,0.2229796499,0.1178622022,-0.1599719226,-0.1607328206,0.0287085604,-0.132377848,-0.0551206805,-0.0210096762,-0.0180984288,0.1744486988,-0.2757012844,-0.3031266928,-0.0291795898,-0.3162955046,-0.1168425232,0.0777494982,-0.140821442,-0.2035886943,0.2582942545,-0.004404407,-0.3116005659,0.0391671993,0.2989181578,-0.0587782227,-0.1035333499,0.4948186278,-0.0174766071,-0.2341255397,-0.2265408337,0.0059656831,-0.8320417404,-0.4196950197,0.0923220366,0.1404572725,0.0354364514,-0.1578514278,0.4367974997,0.230864495,0.2776599824,0.2272067666,-0.4679569304,-0.0105704796,0.101254575,-0.280731082,-0.0497216471,0.1797121316,0.3023636937,0.3777132928,0.2600378692,-0.2399078608,-0.1359714866,-0.2728255689,0.4363948703,0.4620951414,-0.0453722961,0.2471175194,0.1813136637,0.0653796345,0.2439451218,-0.1171858385,-0.1759735644,-0.1848390251,0.1749594957,0.1496821493,-0.0600156933,-0.0156633221,0.0482636467,0.1012092903,-0.1342943758,-0.2212226242,-0.1341987103,-0.1018091291,0.403714329,-0.0811604187,0.350672394,0.0903647467,0.2299607396,-0.2027548105,0.2579711378,0.2215570658,0.1295814812,0.0725676417,-0.1162279621,-0.3555429578,-0.1598296165,0.030582333,0.1808655709,0.0709812716,-0.0929418877,-0.1310455203,-0.3081482351,0.0658408925,0.2741805017,-0.1258384883,-0.1190997139,0.1458713859,0.0900861919,-0.1429093331,0.2187800854,0.3888690174,0.2636514604,0.0101662474,0.3116286099,-0.1266066134,0.0549860857,0.064990662,-0.0614425093,0.1470943987,-0.1211681664,-0.0370414108,0.2725234628,0.0567122214,0.1953904182,0.0161142703,-0.1064417809,0.1345743388,0.2616095841,-0.3864254057,0.259992063,0.0923803076,-0.0890370905,0.13858971,-0.1082067639,0.030560758,0.2163309604,0.1461498439,0.222229749,0.1549704522,0.2731763721,0.0785089582,-0.1611618251,-0.3665848374,0.364913702,-0.1428804994,-0.0714778602,-0.2635197639,-0.094346635,-0.2115145773,0.054637596,0.1483044773,-0.1082378924,0.3571384549,0.3224037886,-0.1065469906,-0.274723053,0.0086299479,0.2282196879,0.049541872,-0.2872061729,0.2931395173,0.1396940947,0.1063329875,-0.0078268209,0.2379901558,0.725264132,0.5168668032,-0.3115155697,-0.1636020094,-0.176910907,0.0911142901,-0.0807719231,0.0747022033,0.2983618677,0.4166287184,0.4595347941,0.1323583126,-0.1695204079,0.0517079942,0.1032559797,-0.321909517,-0.0616908297,0.5042173266,-0.0247077644,-0.0180153437,-0.2502755225,0.346177727,-0.4587026536,0.2066499889,-0.0599659234,0.2616643906,0.2048780918,-0.0402288064,0.0371435806,-0.0611800887,0.6178979874,0.3663236797,0.3425807953,-0.6118857265,-0.1754017174,-0.4910422266,-0.0601615161,-0.1484051943,-0.3703594506,-0.2915735245,0.0971980393,-0.0788731202,0.3884087205,-0.0742788315,-0.0682195649,0.1689771712,0.0846798047,-0.3359518051,-0.1068405658,0.031062372,-0.3000177443,-0.0120873703,-0.2438317984,0.0211333092,-0.2394023836,0.0584723428,-0.1020724401,-0.3599013388,0.347179085,0.3313420713,0.378216207,0.2259520888,0.8266328573,-0.1996865422,-0.1380767971,0.0664436743,-0.0932086334,0.0112206126,0.0206320137,0.0810936093,0.2432469726,-0.0941913053,0.3419338167,-0.5246486664,0.7450121641,-0.0435727648,-0.0832773969,-0.1926093102,-0.3051315844,-0.1723100841,0.2594204247,-0.1198933721,0.0273916256,-0.1963588744,0.3427009583,-0.4602262974,-0.415148735,0.5375388265,-0.3625352383,-0.388100028,0.0640936047,0.2020009309,-0.3227238953,-0.1909432262,-0.7663069367,-0.0921797231,0.2804146111,-0.1019756794,-0.2139068246,0.1753968,-0.0223213769,0.0938506499,-0.0044266516,-0.0552640632,-0.0458026119,-0.254857868,-0.1172515601,-0.2409612238]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/147","title":"Error with sklearn train_test_split","comments":"Indeed. Probably we will want to have a similar method directly in the library","body":"It would be nice if we could use sklearn `train_test_split` to quickly generate subsets from the dataset objects returned by `nlp.load_dataset`. At the moment the code:\r\n\r\n```python\r\ndata = nlp.load_dataset('imdb', cache_dir=data_cache)\r\nf_half, s_half = train_test_split(data['train'], test_size=0.5, random_state=seed)\r\n```\r\nthrows:\r\n```\r\nValueError: Can only get row(s) (int or slice) or columns (string).\r\n```\r\nIt's not a big deal, since there are other ways to split the data, but it would be a cool thing to have.","comment_length":14,"text":"Error with sklearn train_test_split\nIt would be nice if we could use sklearn `train_test_split` to quickly generate subsets from the dataset objects returned by `nlp.load_dataset`. At the moment the code:\r\n\r\n```python\r\ndata = nlp.load_dataset('imdb', cache_dir=data_cache)\r\nf_half, s_half = train_test_split(data['train'], test_size=0.5, random_state=seed)\r\n```\r\nthrows:\r\n```\r\nValueError: Can only get row(s) (int or slice) or columns (string).\r\n```\r\nIt's not a big deal, since there are other ways to split the data, but it would be a cool thing to have.\nIndeed. Probably we will want to have a similar method directly in the library","embeddings":[-0.2563788295,-0.0644306913,-0.035498403,0.2747470438,-0.0786235258,0.2247386575,0.3027848601,0.4740771949,0.2557899058,0.0850956365,-0.1294074953,0.5314365625,-0.4141556323,0.3943921328,0.3385440409,-0.3769304454,-0.0129884733,0.117341511,-0.1270778626,0.0081687514,-0.2582624257,0.139677465,-0.299203068,0.0605836324,-0.4176435769,-0.2679511607,-0.0432194397,0.1660378426,-0.1671729535,-0.1191947088,0.2936882675,-0.0456483141,0.1982234269,-0.0163362902,-0.0001202757,-0.1970368475,0.3166990578,-0.1053743064,-0.459097892,-0.0985875875,0.0543059744,-0.2379517108,0.1257642508,-0.4602694511,-0.1291868538,0.1880947202,0.0822216645,-0.4847090244,0.1998903304,0.2955559194,0.0700132847,0.2954943478,-0.3317224085,-0.0208559334,0.5024752021,0.1594042182,-0.2306277007,0.1308294833,0.061148487,-0.2062914819,-0.0707155168,0.0318351127,-0.1665589511,-0.0234412476,0.3209356368,0.3114843369,-0.3555578887,-0.3718710542,-0.226439774,0.4149742126,0.2031561583,-0.1924046278,-0.4140833318,-0.4694017172,0.2938275635,-0.4254930317,-0.3060105443,0.5813971162,-0.1824505776,0.0774712563,-0.5900189877,-0.3342083097,-0.2069658339,0.1471570432,0.3202574551,0.1352224499,0.2428734154,0.1803170443,0.5178642869,0.0859259292,0.2303887755,0.0697344914,-0.175899744,0.0956755579,-0.3105896711,-0.3996487856,-0.3374843001,-0.0816816911,0.1933583915,0.5096212029,0.097629346,0.0014179993,0.1385325491,0.2062973082,0.3332417011,0.0119127063,0.2409065664,0.3225541413,0.1830050796,-0.0564937964,-0.1412444711,0.1620720476,-0.148314178,-0.0109004853,0.0292095337,-0.0598597638,-0.1965491474,0.150877133,-0.1434649825,-0.1017727181,-0.3230958283,-0.0822922662,0.1210210919,0.1586916596,-0.0305864289,0.0163765587,0.143876493,0.1725385785,-0.1841085553,-0.4094184935,-0.1361972094,0.4354753792,-0.2738498747,-0.0422226973,0.33452335,-0.1520641297,0.2028423697,-0.3122412264,0.1585047245,-0.1634952128,0.1819376498,-0.0709482282,-0.0509082489,-0.0674863532,-0.0159958228,-0.1999692172,-0.075440757,0.0435768627,-0.311545521,0.1118528694,-0.1866648644,-0.4073170125,-0.1227839813,0.0286443457,-0.0923167616,-0.0714214146,-0.3388644755,0.324049592,0.047076378,0.172972247,-0.0122022098,-0.1483084708,-0.2726034522,-0.1007549316,0.361774385,0.1880431175,-0.0538327098,-0.1248503998,-0.2262964547,0.0400578827,0.193218261,0.3626339734,-0.1419844329,0.4681950212,-0.1472182274,0.3395015597,0.4338316619,-0.0933114439,0.0036732885,0.3458417654,-0.0161364228,-0.343672514,0.1544907242,0.3161483407,0.1733408272,0.0234221015,0.105421707,0.9238656163,-0.3765714169,-0.0543399975,0.043083325,-0.4824441075,0.5269793272,0.3176592588,0.3459016085,0.2239728868,0.1350872219,0.6375965476,0.2573592365,-0.1745417267,0.0479181148,-0.3107813299,-0.0653136969,-0.1867287159,-0.1998771876,-0.4482831061,-0.3805777133,0.1778324991,0.0176689252,-0.0052181007,0.1922239065,-0.2317246795,-0.2663811743,0.0454467461,0.3348067999,0.0173546653,-0.0590923578,-0.2479762584,0.4909092486,-0.0883893669,-0.319606483,-0.09619537,-0.2584815919,0.1328234524,-0.5131517649,0.402400583,0.1533718854,-0.0879908204,0.0381007493,0.4042316675,-0.0779418424,-0.3041173816,0.2671818137,0.4619360566,0.1711166948,-0.0776355788,0.1420949548,-0.2199203372,0.12536107,-0.0361787938,0.0647408292,-0.1168909594,-0.0569130778,-0.0468566827,-0.1526348442,0.5931372643,-0.461797148,0.3079969883,0.0863834471,0.2797944844,-0.258867979,-0.1027004644,-0.100128226,-0.2280480564,0.030880874,-0.2744944096,-0.0290004387,-0.1312364042,-0.3971855342,-0.0178775415,0.5642334819,-0.2550504208,0.0369858816,0.0801369026,0.101041168,-0.0781374723,0.0795487836,0.2340876311,0.2341329008,0.1205509752,0.0103305047,-0.0444708578,-0.1899305135,-0.1115485281,0.1899092346,0.1785846055,0.2659332454,0.0184565578,-0.0536246821,-0.2218679339,-0.1834138036,-0.1153930202,0.1425026357,0.1939889789,-0.0523621179,0.0535825379,-0.1669272631,-0.2567556798,-0.2590998113,-0.0336953625,0.1862380207,-0.176160723,0.0310846511,0.1443035156,-0.0679659471,0.0368778817,-0.3373915553,0.3077219129,-0.0952457562,0.0487343296,0.0281828828,-0.308634311,-0.0889451206,0.0532422327,0.2118645161,0.2271987647,0.3930493593,-0.1673459858,-0.049573496,-0.0643755421,0.1103550419,0.03636536,-0.109863326,0.3478211164,0.2598256469,-0.0590962209,-0.3074292839,-0.5575748086,0.0834034979,-0.1699415892,-0.1205193922,-0.0174682215,0.2569476366,0.1458992213,-0.1217034236,-0.5037764311,-0.3771938682,-0.1364512742,0.2819857001,0.1353143901,0.1397358924,0.1747769117,0.0675605908,0.1872658283,0.0702715591,-0.255097717,0.0235643815,-0.1180053502,0.1783989072,-0.1786459982,-0.147701323,-0.2250197977,-0.1983319074,-0.1229961216,0.3168367445,-0.3121153414,0.0467343107,-0.0172009543,0.0705589503,-0.089644514,-0.1677795202,-0.0084794508,-0.139190793,0.1154391095,0.0718043894,0.0189211685,0.2505007982,0.1257946193,0.0957762673,-0.0736251771,0.3596738279,-0.0190995764,0.6095506549,0.1091391072,0.0170682725,-0.1329470277,-0.106902048,0.0774692595,-0.012748694,-0.1763843596,0.2936709821,-0.1902980506,-0.298291415,0.1354304701,-0.0569631346,-0.4140163064,-0.1714672297,0.2115373611,-0.2598895133,-0.3451435864,0.1164083406,-0.3923815787,0.0280913822,-0.093161121,0.2099121809,-0.0289114993,-0.1996933371,-0.1053896248,-0.2863877714,0.5048568845,0.1104924977,-0.5086128116,-0.4263687134,-0.2703323662,0.0091581196,0.2041931152,0.4436284006,0.0901085809,-0.0617330223,0.1061705053,0.1901395768,0.5987061858,-0.1073387712,0.020644933,0.0413502604,-0.1180150211,-0.2646685243,0.0089356974,-0.2493711114,0.2459192872,0.2119954973,0.0828321874,-0.268522948,-0.0406379215,0.6046745777,0.3252046406,-0.0287797656,0.1603163034,0.2589775026,0.0314845443,-0.2066659927,0.2154177874,0.0481158197,0.1618179679,-0.1380858868,-0.0165908486,-0.1854343861,-0.1347726882,0.2413573116,0.170693472,0.2111901045,0.1743105054,0.2809842229,-0.077496618,0.0503891371,-0.0587404631,0.2672086954,-0.2009343803,-0.0589822978,-0.015685318,0.0731432512,0.5022845268,0.3091515303,-0.0625841469,0.0256795716,-0.1087269709,0.023732394,-0.12132027,0.2012005448,0.1082483679,-0.3782393634,-0.3602783978,-0.6372742653,0.4691144228,0.0339207314,0.1179716587,-0.0212406181,-0.3231775463,-0.3897924721,0.5576679111,0.316888541,0.8628214598,-0.1642515808,0.1902048588,-0.0143389264,0.2521319389,1.0715600252,-0.2784531116,0.284389168,-0.5403776169,-0.0387838297,0.0270936433,-0.0678363368,-0.1983575672,0.3843798041,-0.0974614993,0.2163694799,0.0530878268,0.0698267296,-0.3057222366,0.5422883034,-0.0760633722,-0.382863462,-0.1593525708,0.026511468,-0.1598715633,-0.165008232,0.0785272941,0.0064500128,-0.3177283406,0.1431534141,-0.3328509629,0.1053497419,0.115870975,0.4089378417,0.0585269183,-0.494721055,-0.0305618662,0.150126338,0.0452112369,0.3260215521,-0.0279975627,0.0441324972,-0.1474583894,-0.2814707756,0.1420121491,0.0265624449,0.407836169,0.1032635048,-0.1660955995,-0.0966184363,0.0156090707,-0.2325754613,0.0888252705,0.1562630981,0.168102473,-0.1035697386,-0.5118636489,0.0788946822,-0.1375501454,-0.2402336895,0.0197804235,0.1528166234,-0.1752555519,0.0896385685,0.1239166409,-0.1246537045,-0.0088110399,0.0602963977,0.1995832622,-0.0178375263,0.3614655733,0.1788297743,0.0272287764,-0.2554223835,0.2523908317,0.1545137912,-0.4393794835,0.3917892277,-0.0034396367,0.1752132177,0.1799971908,0.199115783,0.1146655157,0.0907526538,-0.0992985964,-0.464813143,-0.0136209093,0.2882833779,0.0208369177,0.2554058135,0.0637198761,0.2751032114,-0.2857967615,0.1870660633,-0.1359284818,-0.0006194993,-0.2840985358,0.2282785773,0.2304070294,0.0850352049,-0.2304810733,-0.0407900363,-0.0486887507,0.2448448241,-0.023803344,-0.1006009579,-0.2697364986,0.163958773,0.0910053104,-0.0319474488,0.037788514,-0.0892272592,-0.3234681487,-0.3198547065,-0.0629401729,-0.0093971826,0.1502421349,0.3009359837,0.425865382,0.0426073261,-0.0751170516,0.1522837132,-0.2053519636,0.2097007483,-0.0803709105,0.369115591,0.5045804381,0.0509102233,-0.139090687,-0.1251948774,-0.0454152003,0.0229504928,0.2342877984,-0.0373448953,-0.0829716623,0.0989815816,0.2699003518,0.6565375924,0.16306144,-0.0595031045,0.3408803046,0.0735268369,-0.1461358964,-0.1445009261,0.3284503818,-0.1542434543,-0.0851493403,0.459587276,-0.0316406973,0.3820690215,-0.3228369057,-0.1634370089,0.0500013083,-0.042537082,0.4517760873,0.5726828575,0.0101627288,-0.1869971305,-0.1408829838,-0.0037417291,0.4036642313,0.3601395786,0.0431505181,0.4196452498,0.0598751567,0.2523577809,0.218060717,-0.1314421594,0.0102914572,0.1348485351,0.026819773,0.2913077176,-0.1844392866,0.7214781642,-0.3545428813,-0.1130718067,-0.2718466818,0.3055517673,0.060774114,-0.1600049436,-0.0903750882,-0.2057932615,-0.2210249752,0.2242019325,-0.039345149,-0.323903203,0.0114937853,0.1725100577,0.1049667522,-0.1130418032,0.154133901,0.155181244,0.1115009263,-0.4149500728,-0.2298942357,-0.2810633183,0.0339167751,0.258724153,-0.0231290776,0.1134907529,0.2737301588,-0.2614902556,-0.0689047351,0.0131000448,0.0383136645,-0.0697700381,0.2757574022,0.3598179221,-0.0031430058,0.4360384643,0.0199894551,-0.0205488447,0.0896214247,0.2222483307,-0.225742951,-0.4241244793,0.3700315654,0.1916135997,-0.1305879205,-0.2467597872,0.0233099032,-0.0844789147,-0.2156492621,-0.027564982,-0.0733689219,0.055828847,0.0095563252,0.0037111084,-0.0745841563,0.6697283387,0.0892073885,0.1837118715,-0.2656835616,-0.0826346874,-0.2583622336,-0.2813948989,-0.0876498669,-0.173741132,0.2166977525,0.3500255346,-0.0326275714,0.3149937987,-0.4215781093,0.5697966218,0.1595301926,0.3315419853,-0.1674699038,-0.1807719916,-0.3066803217,-0.037350066,-0.066092357,-0.3598473072,0.0206665099,-0.0312069003,-0.3014675975,-0.0024159802,-0.2434080541,0.2996055782,0.3074098527,0.1039586738,0.0568419024,0.4448255897,0.0150189986,0.0702329129,-0.4384061992,-0.4519982338,-0.3424550593,0.2438722402,0.104706645,0.4713110924,-0.0331988633,0.2415590286,-0.2706903219,0.3779142201,0.0455007702,-0.0480831489,-0.1101142168,-0.0352189355,-0.4694276154,0.0731915906,0.3426528871,0.2636346817,-0.1662855297,0.1308384985,0.0251958165,-0.2840974033,0.2643692791,-0.37921381,-0.3485509753,-0.0959481075,0.096818015,0.3055576384,-0.2053628862,-0.4704428911,-0.2265264839,0.1700679511,-0.1535383612,0.2280954421,0.218410641,0.1744409651,0.0046481714,-0.0906179771,0.2904072702,0.0273292344,-0.3758798838,0.0322177671,-0.3449985981]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/147","title":"Error with sklearn train_test_split","comments":"Related: #166 ","body":"It would be nice if we could use sklearn `train_test_split` to quickly generate subsets from the dataset objects returned by `nlp.load_dataset`. At the moment the code:\r\n\r\n```python\r\ndata = nlp.load_dataset('imdb', cache_dir=data_cache)\r\nf_half, s_half = train_test_split(data['train'], test_size=0.5, random_state=seed)\r\n```\r\nthrows:\r\n```\r\nValueError: Can only get row(s) (int or slice) or columns (string).\r\n```\r\nIt's not a big deal, since there are other ways to split the data, but it would be a cool thing to have.","comment_length":2,"text":"Error with sklearn train_test_split\nIt would be nice if we could use sklearn `train_test_split` to quickly generate subsets from the dataset objects returned by `nlp.load_dataset`. At the moment the code:\r\n\r\n```python\r\ndata = nlp.load_dataset('imdb', cache_dir=data_cache)\r\nf_half, s_half = train_test_split(data['train'], test_size=0.5, random_state=seed)\r\n```\r\nthrows:\r\n```\r\nValueError: Can only get row(s) (int or slice) or columns (string).\r\n```\r\nIt's not a big deal, since there are other ways to split the data, but it would be a cool thing to have.\nRelated: #166 ","embeddings":[-0.2569483519,-0.1140865311,-0.0456743501,0.2932060063,-0.0182216056,0.2564948201,0.3482436836,0.4857951105,0.2606162727,0.0623584315,-0.1505012363,0.4861782193,-0.3912051618,0.4599109292,0.2672834992,-0.3708164394,-0.0134636881,0.0748469606,-0.1207511052,0.013030258,-0.2829493284,0.1548824012,-0.315553844,0.0942894295,-0.4775365889,-0.2517726719,-0.0465350747,0.1642509401,-0.1807188839,-0.1188640743,0.2750198841,-0.1205005795,0.2114437521,0.0100398846,-0.000119863,-0.1646049023,0.3370469809,-0.1011632904,-0.4652327299,-0.1025070548,0.0439553745,-0.2354095876,0.1043180302,-0.4516631067,-0.1418360174,0.1884279251,0.0721964389,-0.4296096563,0.2348443866,0.3951334059,0.0723653734,0.2794694006,-0.3508390188,-0.0557404235,0.4652903378,0.1655650288,-0.2305738032,0.1407256573,-0.0044395784,-0.2081433535,-0.0609252043,0.0380884223,-0.1692789048,0.0332914181,0.2814157605,0.3301760852,-0.2894614637,-0.3925126493,-0.153069526,0.4183163941,0.1841216832,-0.1357042044,-0.4171216786,-0.451628685,0.2825879753,-0.4374984801,-0.2754886448,0.5758084059,-0.2000607252,0.0803909674,-0.6369239688,-0.2779822946,-0.1870245785,0.1448346376,0.2201359272,0.1323650181,0.1985774785,0.1728741527,0.478415817,0.0546521917,0.2277911603,0.0634007454,-0.1501333416,0.0968041196,-0.3364036083,-0.3428457081,-0.3502821028,-0.1050110385,0.1591578722,0.4503516257,0.0915023834,-0.0523260757,0.1639137715,0.1950986683,0.3311975896,0.0052181315,0.2349870354,0.3810406923,0.1094160005,-0.07453987,-0.1634575129,0.1327302307,-0.1530514061,-0.0274471249,0.0265386365,-0.0624873713,-0.1417855173,0.1597489119,-0.1975298375,-0.0667410567,-0.3275434673,-0.0591846853,0.1011251137,0.1712584794,-0.0065921489,-0.0186716523,0.1105090454,0.2627281249,-0.1398673654,-0.4214609265,-0.1646440774,0.4236009717,-0.2524010539,-0.0562339947,0.3488939703,-0.1114456058,0.2414795607,-0.2876233459,0.1411073059,-0.2246326655,0.1424154937,-0.114384003,-0.0630212575,-0.0368102975,0.0246089436,-0.1575921774,-0.0321826898,0.0039899936,-0.2516914606,0.1170345917,-0.2050728351,-0.3792132139,-0.0634475425,0.0360029154,-0.0746891573,-0.0379351489,-0.3269699216,0.260887444,0.061293222,0.196274817,-0.0274356809,-0.1904699951,-0.2497649193,-0.0892976299,0.3830611706,0.1365506649,-0.0567108206,-0.1068019345,-0.1871412843,0.0706358328,0.2040556669,0.398139298,-0.0995157361,0.4651035368,-0.1536919475,0.367555052,0.405600816,-0.0892759264,0.0188801531,0.376881808,-0.039000567,-0.3654694557,0.1524698883,0.2740006447,0.1873836666,0.0465060771,0.1391949356,0.8914152384,-0.3921906352,-0.0926930979,0.0256874003,-0.47062549,0.5408143401,0.3322565258,0.3680728674,0.2402869612,0.062538527,0.5893423557,0.2266324759,-0.1284473687,0.0442358889,-0.2786621153,-0.1033967957,-0.1848076284,-0.2237640172,-0.4369088709,-0.3681192994,0.1881679893,0.0357415676,-0.0135162733,0.2145391703,-0.2034063637,-0.3391771615,0.0444157831,0.3161648512,0.0075936257,-0.0475113317,-0.2430584878,0.5013682842,-0.1027901992,-0.29519701,-0.0886213183,-0.2627440393,0.1236070544,-0.445694834,0.3641107082,0.0853056237,-0.1343327612,0.0796685368,0.3754869699,-0.0288696419,-0.2562999129,0.2504473627,0.4605531096,0.1744277179,-0.116595313,0.1269983351,-0.2496749461,0.0645517707,-0.0342737287,0.0458104089,-0.0955944955,-0.0484262519,-0.0699278861,-0.1606822312,0.5612823963,-0.4839864671,0.3451966345,0.0558419079,0.3143748045,-0.2548056245,-0.1261195093,-0.1040838584,-0.2312958837,0.0422628485,-0.3178004622,-0.0551734976,-0.1599614471,-0.405405134,-0.0303131305,0.5520142317,-0.2579165101,0.0254625157,0.0627810434,0.0688973367,-0.0516178384,0.0547661558,0.194868207,0.2260340005,0.1281138361,0.0283427015,-0.049291499,-0.2222720534,-0.1232296005,0.191549167,0.1956789345,0.2513378859,0.0384018458,-0.0450071022,-0.2341302186,-0.1417838782,-0.1079509407,0.1661491543,0.214027822,-0.1248808131,0.0350012071,-0.1736256033,-0.2547493279,-0.220108524,-0.0654914901,0.2168937773,-0.2087290287,0.015602245,0.1496811509,-0.0763095841,0.0514439791,-0.3736026287,0.2741066813,-0.0541514382,0.1232143119,0.0597609207,-0.3143505156,-0.087536186,0.0651828796,0.2163099945,0.266587019,0.3715928495,-0.2142543495,-0.0740729049,-0.0023426823,0.1604521573,0.0458440855,-0.1502849311,0.3611646295,0.2298995256,0.0157603063,-0.3316036165,-0.5483182669,0.1450603008,-0.150122568,-0.119999215,0.0480633937,0.303842783,0.1254639626,-0.0837157741,-0.4755842686,-0.3440048993,-0.1397309601,0.2625153661,0.1334497333,0.1164336428,0.187238574,0.0969770774,0.2155495435,0.0834667236,-0.2630026639,-0.0209955256,-0.0971684158,0.1900808811,-0.162645191,-0.1504536569,-0.2044730037,-0.1551111192,-0.1428550482,0.2752623856,-0.2988185883,0.0601560473,-0.0304624829,0.1010631993,-0.0936393365,-0.2068511695,0.0016110578,-0.1570055336,0.0987529755,0.1018804014,0.0225174762,0.2507362366,0.1272616982,0.0668571368,-0.0783914328,0.3970980346,0.0186268613,0.6054373384,0.1635080725,0.0452331081,-0.1405336112,-0.1312538385,0.0647960231,-0.0588942692,-0.2104960084,0.2843697071,-0.2199316472,-0.2959861457,0.1312840432,-0.0922289789,-0.3537943959,-0.1559274048,0.2327768356,-0.2885988653,-0.3449057639,0.1514621377,-0.4150838256,0.0396984257,-0.0800149292,0.2476316988,-0.0253144819,-0.1858376265,-0.117815271,-0.2977883816,0.4540146291,0.0806962848,-0.4848507047,-0.4575248361,-0.276512146,0.0087181609,0.1644867957,0.4792730808,0.0386130624,-0.0612570569,0.1048864573,0.2199711353,0.6277074218,-0.1710844785,0.0607880615,0.0891879275,-0.082927689,-0.2905177772,-0.0332521126,-0.3189702928,0.237621963,0.2770929933,0.1167735606,-0.2370818555,-0.0313001387,0.6310045719,0.3025521636,-0.0167750921,0.1539767385,0.2354175001,0.0457645357,-0.2051997781,0.2402983159,0.1202969179,0.1800380498,-0.1707599461,0.0173539501,-0.2341251522,-0.1430230439,0.2392040193,0.1449515373,0.2219616175,0.186364606,0.3058387935,-0.0829100609,0.0212799497,-0.0926416293,0.2891876996,-0.1871744692,-0.074647069,0.004902469,0.0317690559,0.4808318615,0.3356862068,-0.0602839291,-0.0752472952,-0.0906944871,0.0654864088,-0.1421135217,0.2122607827,0.1479070634,-0.3756569326,-0.3040614128,-0.5986819267,0.4623316526,0.0368541144,0.1120779216,-0.0749410242,-0.2756934166,-0.382117182,0.6074006557,0.2819454074,0.8676666617,-0.1112470776,0.1446232051,0.0336582772,0.2972201109,0.9760351181,-0.2311370075,0.2776483893,-0.4913999438,-0.0835624859,0.0179475471,-0.0891759321,-0.228184104,0.4061365128,-0.0952719003,0.1984728873,-0.0209157951,0.1062931344,-0.2804551125,0.5027315617,-0.063131474,-0.3235112727,-0.1350439191,0.0441684872,-0.1490650177,-0.1507719606,0.1143179759,0.0312258229,-0.3838447928,0.1171634793,-0.374946028,0.1045828685,0.1374202073,0.4235677421,0.0233054515,-0.4984021187,-0.0692886338,0.1551826596,0.0793038756,0.2915505171,-0.029660603,0.0205415767,-0.1107919589,-0.3066287339,0.1232242063,0.0125903878,0.4121911824,0.0893353969,-0.1576717496,-0.0531839915,0.0128456475,-0.1434264779,0.0922475681,0.1584906429,0.0939853862,-0.2053002566,-0.5513880849,0.0308802314,-0.101369895,-0.2447228879,0.0366106294,0.1257126331,-0.1672928184,0.0812736675,0.149694562,-0.1015563086,0.0096036997,0.069084689,0.1989646703,0.0257685594,0.4087576568,0.169669196,0.0149909277,-0.2603653669,0.2899105251,0.1772492528,-0.4158745706,0.3919133544,-0.0153332464,0.1874467582,0.1923105866,0.2008404136,0.140096724,0.0699369013,-0.0945183784,-0.4688631892,0.0129682114,0.2745258212,0.0409168825,0.2310877144,0.0311103128,0.3206758201,-0.2679524422,0.239792943,-0.1488738805,-0.0006701621,-0.3188642263,0.239845112,0.2180347294,0.0314730145,-0.2219565511,-0.0797616988,-0.0356955938,0.2125149816,-0.0042834533,-0.1130599752,-0.2567227781,0.1588771194,0.1360127777,-0.0374593362,0.0217735525,-0.0772800744,-0.3235512674,-0.2925919294,-0.1174314246,0.0064431359,0.1835477799,0.3285855949,0.4019186199,0.1216199547,-0.045230329,0.1237986088,-0.2156037986,0.214045763,-0.0645233989,0.356154561,0.5456652045,0.0848277286,-0.1876543611,-0.1474977881,-0.0566251054,-0.0010571873,0.2182969898,-0.0950098112,-0.0432679057,0.0673433617,0.2713105381,0.6639734507,0.1024219841,-0.0629001781,0.3388394117,0.0908331648,-0.1755443513,-0.0995423347,0.3765287995,-0.1287986785,-0.0712413192,0.429168582,-0.0855185688,0.3902583122,-0.3430961668,-0.1161453798,0.0440844484,-0.0747175962,0.4296702445,0.5549436212,0.0382409841,-0.1699738204,-0.1131744087,0.0045002829,0.3752913773,0.3634791672,0.0169985425,0.342877239,-0.0046630655,0.2291453779,0.21812132,-0.121668458,0.0039306907,0.1215187162,-0.0687024593,0.2627701461,-0.2282209992,0.6733939052,-0.4179520607,-0.0944250375,-0.3160182238,0.2916048169,0.0200028401,-0.1502272785,-0.1161034852,-0.1986629814,-0.2487082332,0.1865767539,-0.0251565762,-0.3126351237,0.0533652902,0.1627152115,0.0811766088,-0.1097970977,0.1012448594,0.1731932014,0.0533245206,-0.4251617491,-0.2190986276,-0.2906339169,0.0523394905,0.2768039703,-0.023917608,0.2154978663,0.2652592361,-0.295358479,-0.0683991984,0.0338048004,-0.033724986,-0.0673800632,0.2811967731,0.3494327068,-0.0055433693,0.4732094407,0.0471644662,-0.0386035666,0.0756110772,0.2257124633,-0.1984585226,-0.430564791,0.4200257361,0.1698119342,-0.1430337131,-0.2151940912,-0.0371317379,-0.130097419,-0.1500243098,-0.0708069727,-0.0634077415,0.0785081014,0.037266288,0.0070697968,-0.0588932373,0.6655322909,0.073628068,0.1777789891,-0.3108721972,-0.0419580415,-0.2678310573,-0.3081647158,-0.1510178894,-0.2046530396,0.2512850463,0.3652299047,-0.0313160457,0.3217016459,-0.3977333903,0.607329905,0.1069925949,0.3801503479,-0.167644456,-0.1916350126,-0.2867304087,0.0025048188,-0.0445203781,-0.347338438,0.0558306873,-0.0132053699,-0.282705307,0.0444982648,-0.2396046966,0.3159409165,0.2768389881,0.061905954,0.0673540756,0.4858365655,0.0360371023,0.0912104249,-0.4766132236,-0.4490068853,-0.3434131145,0.2813602388,0.1486535072,0.4828758836,-0.0696072802,0.1963066608,-0.2598830462,0.333879739,0.0002245899,-0.0466096587,-0.0987671539,-0.0088584414,-0.4488429725,0.1247706041,0.3381944597,0.1940866411,-0.1749995947,0.1647375971,0.0100950664,-0.3686333597,0.2831700444,-0.3759089708,-0.3778351843,-0.1025133505,0.1043999195,0.3132382035,-0.1834149063,-0.4612798095,-0.2482479662,0.1673630923,-0.1410721391,0.2221338004,0.2094789892,0.2086268514,0.0194430426,-0.075892739,0.3094868362,0.0191698261,-0.4151189923,0.0251551364,-0.3244137764]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/143","title":"ArrowTypeError in squad metrics","comments":"There was an issue in the format, thanks.\r\nNow you can do\r\n```python3\r\nsquad_dset = nlp.load_dataset(\"squad\")\r\nsquad_metric = nlp.load_metric(\"\/Users\/quentinlhoest\/Desktop\/hf\/nlp-bis\/metrics\/squad\")\r\npredictions = [\r\n {\"id\": v[\"id\"], \"prediction_text\": v[\"answers\"][\"text\"][0]} # take first possible answer\r\n for v in squad_dset[\"validation\"]\r\n]\r\nsquad_metric.compute(predictions, squad_dset[\"validation\"])\r\n```\r\n\r\nand the expected format is \r\n```\r\nArgs:\r\n predictions: List of question-answers dictionaries with the following key-values:\r\n - 'id': id of the question-answer pair as given in the references (see below)\r\n - 'prediction_text': the text of the answer\r\n references: List of question-answers dictionaries with the following key-values:\r\n - 'id': id of the question-answer pair (see above),\r\n - 'answers': a Dict {'text': list of possible texts for the answer, as a list of strings}\r\n```","body":"`squad_metric.compute` is giving following error\r\n```\r\nArrowTypeError: Could not convert [{'text': 'Denver Broncos'}, {'text': 'Denver Broncos'}, {'text': 'Denver Broncos'}] with type list: was not a dict, tuple, or recognized null value for conversion to struct type\r\n```\r\n\r\nThis is how my predictions and references look like\r\n```\r\npredictions[0]\r\n# {'id': '56be4db0acb8001400a502ec', 'prediction_text': 'Denver Broncos'}\r\n```\r\n\r\n```\r\nreferences[0]\r\n# {'answers': [{'text': 'Denver Broncos'},\r\n {'text': 'Denver Broncos'},\r\n {'text': 'Denver Broncos'}],\r\n 'id': '56be4db0acb8001400a502ec'}\r\n```\r\n\r\nThese are structured as per the `squad_metric.compute` help string.","comment_length":112,"text":"ArrowTypeError in squad metrics\n`squad_metric.compute` is giving following error\r\n```\r\nArrowTypeError: Could not convert [{'text': 'Denver Broncos'}, {'text': 'Denver Broncos'}, {'text': 'Denver Broncos'}] with type list: was not a dict, tuple, or recognized null value for conversion to struct type\r\n```\r\n\r\nThis is how my predictions and references look like\r\n```\r\npredictions[0]\r\n# {'id': '56be4db0acb8001400a502ec', 'prediction_text': 'Denver Broncos'}\r\n```\r\n\r\n```\r\nreferences[0]\r\n# {'answers': [{'text': 'Denver Broncos'},\r\n {'text': 'Denver Broncos'},\r\n {'text': 'Denver Broncos'}],\r\n 'id': '56be4db0acb8001400a502ec'}\r\n```\r\n\r\nThese are structured as per the `squad_metric.compute` help string.\nThere was an issue in the format, thanks.\r\nNow you can do\r\n```python3\r\nsquad_dset = nlp.load_dataset(\"squad\")\r\nsquad_metric = nlp.load_metric(\"\/Users\/quentinlhoest\/Desktop\/hf\/nlp-bis\/metrics\/squad\")\r\npredictions = [\r\n {\"id\": v[\"id\"], \"prediction_text\": v[\"answers\"][\"text\"][0]} # take first possible answer\r\n for v in squad_dset[\"validation\"]\r\n]\r\nsquad_metric.compute(predictions, squad_dset[\"validation\"])\r\n```\r\n\r\nand the expected format is \r\n```\r\nArgs:\r\n predictions: List of question-answers dictionaries with the following key-values:\r\n - 'id': id of the question-answer pair as given in the references (see below)\r\n - 'prediction_text': the text of the answer\r\n references: List of question-answers dictionaries with the following key-values:\r\n - 'id': id of the question-answer pair (see above),\r\n - 'answers': a Dict {'text': list of possible texts for the answer, as a list of strings}\r\n```","embeddings":[0.0508040339,0.0322123691,-0.0459481664,0.3747369051,0.3514669836,-0.1831914186,0.1283157915,0.2945096195,-0.0273575988,-0.016604308,0.0504663512,0.7067572474,-0.2432426214,-0.1496466845,-0.2859814763,-0.1268263459,-0.0782126337,0.2369157672,0.2221025676,0.1331424415,-0.106973663,0.3732542396,-0.2818196714,0.0420603156,-0.1871123612,-0.092268154,0.1176642776,-0.01809345,-0.4135136008,-0.6164665818,0.3017502725,-0.2972587049,0.0052366056,0.3161635101,-0.0001209838,0.0202212855,0.3255316317,-0.1212441251,-0.0995187759,-0.362477839,-0.1135787591,-0.2849887013,0.5184355378,-0.2206499279,-0.0529840961,0.1079014987,-0.2984890044,-0.3533740044,0.3960222304,0.4213611186,0.1145917401,0.5141720176,0.17944251,0.015836034,0.2684787214,0.0566137582,-0.2425853908,0.303730458,-0.0190495215,0.0541987233,0.2327842861,0.1645570397,-0.1349526644,0.0857612267,0.0732456818,0.1215908527,0.1143218726,-0.2737191319,-0.3592942655,-0.100736253,0.2150742114,-0.2580164075,-0.2781116366,0.1453799456,0.1172601357,-0.2204061449,-0.1395785362,-0.1348273009,-0.1219677031,-0.140086785,-0.0538091734,0.059328109,-0.2679055631,0.1362895519,0.0089108674,0.2972382307,-0.1583654433,0.0389316566,0.0348842293,-0.0563536137,0.0081363628,0.0158318728,-0.0161510687,0.0318318345,-0.3313934207,-0.3471526504,-0.2688043416,-0.4717374742,0.0875613466,-0.2626972497,0.4471974671,0.1234180257,0.3111550808,0.1956386268,0.1954374909,0.2209554464,0.3190047443,0.0902371332,-0.0424138866,0.0770395622,0.0266692955,-0.0146498308,-0.1296214163,-0.2506048381,-0.0045461426,0.1766952276,0.2899586856,-0.0501544364,-0.4034970403,-0.0299778003,-0.1991444379,0.1590966582,0.1704541743,0.0738310292,-0.0638078675,0.5698382854,0.1854057759,0.3461947441,-0.0511372909,0.0961199626,-0.019367015,0.0788623318,-0.009686959,-0.1923852414,-0.0270005018,0.2162169069,0.1956911385,0.0224326346,0.1642465293,-0.3621160388,-0.0669123754,-0.2479941249,0.128231436,0.0124068372,-0.0250415243,0.0036076822,0.463848561,-0.4866758883,0.0531590544,0.1031382233,-0.255577296,0.0349704623,-0.2557273805,0.079171896,-0.1247742623,-0.045316387,-0.2481400073,0.1151024848,0.2089597434,0.1211374551,-0.1219426841,-0.2758208513,-0.1292758137,-0.3359872997,0.1065729186,0.0315504782,-0.4582893848,0.0349679813,0.1058815569,0.1642006934,0.0140298782,0.3637439013,0.0033610323,0.2000980228,0.1050294712,0.0284748748,0.8762322664,-0.8044187427,-0.0820784718,-0.0090432046,0.001266808,-0.2809941769,-0.0873116031,-0.022015214,0.4402936697,0.0533394963,0.1420265436,0.0503537953,0.1453561038,-0.0085408613,-0.2638967633,-0.1193146557,0.1466754526,-0.1425349861,-0.1730227321,0.1039286926,-0.079743579,0.2684730589,0.1105622649,-0.0664785951,0.0611145757,0.1640529037,0.474242419,-0.2778576314,-0.0457406491,-0.3377385736,-0.3439454734,-0.0655030683,-0.5480741262,0.0769053474,-0.2180027813,0.0058844876,-0.4639632702,0.0206557568,-0.0563175716,0.3359313905,0.1233770177,-0.1026299596,0.0918412507,-0.09350577,-0.255533576,-0.6211904287,-0.2620673776,0.0830386728,0.1390779316,0.3291445971,-0.146387592,-0.3838292062,0.1651629359,0.4464730322,0.2130034715,0.0080681993,-0.0658437535,0.1213375106,0.3540560305,-0.1507598013,0.0106922667,-0.1044697389,0.1438919008,-0.269831419,-0.0107435836,0.2080315799,0.2057911307,-0.2543959916,-0.0049187574,0.5869116783,0.107732676,0.3149624765,-0.2861791849,0.3985180557,0.120303452,0.17695117,-0.0821278244,-0.2799807191,-0.3304685056,0.1522513032,-0.0254360698,-0.1620459706,-0.063786827,-0.1675985307,0.4572305083,0.3228217065,0.0994222239,0.1478133649,0.0465434566,0.0683184788,-0.2657335401,-0.1572823972,0.323928386,0.0054938663,-0.0097191008,-0.0885386541,0.0001503421,0.0225493163,0.2572911382,0.1510605365,0.5275332332,0.1311570555,0.3096666932,-0.1022444069,0.069310084,-0.0438624881,0.0674388111,0.1808807105,-0.3641236126,0.0111573087,-0.0118701011,0.0072927079,-0.238510862,-0.4707314968,-0.0267353449,-0.4714015424,-0.1158625707,0.0100719975,0.1432020515,0.1250333488,0.4857570827,0.0345742926,0.2919036746,-0.111506559,0.0830251873,-0.4020986557,-0.0498573147,-0.0003868992,0.1272578686,0.032095138,-0.1930103898,-0.0936159268,0.1062563732,-0.2555807829,-0.2851400673,0.0597852431,-0.3108468056,0.2366630435,0.3271429539,0.2808583081,-0.3724357188,-0.1666135937,0.3785741925,-0.0288569517,-0.1685851067,0.2441929281,-0.1426271647,-0.0243622791,-0.0988962501,-0.1292325556,-0.3072052598,-0.3354568779,0.0683564991,0.3457452357,0.3730772138,0.0531388633,0.2724443376,0.4034454525,0.0807920545,-0.000171115,-0.0012312595,-0.0164435767,0.3754096627,-0.2786477208,-0.3549692929,0.1033586785,-0.0014872467,0.3966333568,-0.1320361644,-0.1763366461,-0.4355768263,0.0963414907,-0.1231454015,-0.0684964806,0.0432276279,0.1222132444,0.1184096038,-0.1052376404,-0.1198866367,-0.0428601243,0.2057044506,0.2839810848,0.2661149502,0.0998692885,0.0670479015,0.1417685002,0.6796893477,0.5670326352,-0.3287319541,0.1067522913,0.0736637637,-0.0548754036,0.1150363311,-0.1393596083,-0.1889291406,-0.2117861211,0.1986008435,-0.0188459586,0.0248565525,-0.2308124602,-0.014350527,0.3296389878,-0.2408663929,-0.1322581172,0.1366748661,-0.4730261266,0.166498661,-0.1369110942,-0.0744425729,-0.1887187511,-0.0989009663,-0.0806802735,-0.021703463,-0.0713026896,-0.0749707296,-0.2790544331,-0.1767881066,-0.1064951718,0.3435658216,0.0650887266,0.4897403121,0.0626077279,0.099967584,-0.177779451,0.0076987888,0.3392592669,-0.0674403161,-0.2268365324,0.2448605448,0.2739031017,-0.6055728793,0.0356464982,-0.4606864154,0.358271867,0.3156456947,0.3402145207,-0.4001620412,-0.1986040771,-0.1345352978,0.0810137764,-0.1265990138,0.1817157716,-0.3149038851,-0.3180416524,-0.0891573429,0.1461329609,0.3113482893,0.3501497209,0.1483452022,-0.2247531265,-0.1554919332,-0.1953377873,-0.0279312544,-0.0508835092,0.3561355174,-0.0212169793,-0.1980430931,-0.0924143791,-0.0144238546,-0.0264717508,0.2210911363,0.2409333885,-0.0601793341,0.1075479016,-0.0482941493,0.5898953676,0.2381630987,0.0565027967,0.2755434215,-0.2692461908,0.0346872993,0.1243365705,-0.3535950184,0.363142103,-0.0017891934,-0.0033939842,-0.1160988137,0.3981674016,0.0824366733,0.1899142712,0.1251915693,0.2695427239,-0.099305585,0.5887526274,0.5044378042,1.0051785707,-0.0487237722,0.015972957,0.2873916328,-0.2012463361,0.7326346636,0.3245358467,0.0919515416,-0.4316605628,0.3172285855,-0.0308437347,-0.2598410845,0.18461667,0.1067727357,-0.4360992908,0.1860759705,-0.4134887159,0.1592618376,-0.0080980444,0.1717671156,0.0267341882,-0.231411472,-0.3901663423,0.0879284814,-0.3127359748,-0.2019169182,-0.0495227873,-0.2381231189,-0.381518662,-0.0429041982,-0.214808926,0.0469144285,-0.2930558622,0.142516166,0.4490362406,-0.4361013472,0.455305934,0.7702955604,0.127358079,0.0408997498,-0.1438682377,-0.0859153494,0.0762470439,-0.0254365224,0.2017617822,0.0004917998,-0.0491873585,-0.1677304953,-0.1297774017,0.1660690308,-0.043672353,-0.3717532158,0.3717103302,0.2580517828,-0.0624023266,-0.274951607,-0.3037238419,-0.109939754,-0.4069612324,-0.1118243635,0.0860736594,0.1400077641,-0.3264703155,0.4341596961,0.3683526814,-0.0564818271,0.4477924407,0.0986703634,0.0883863345,0.1416399777,0.4436476231,0.1274261922,-0.181484431,-0.2552116811,-0.0645526722,-0.4761027992,-0.0928403214,0.328677386,-0.0406082235,-0.213104412,-0.010864106,0.5669576526,0.2606894672,0.5758245587,-0.2606144845,-0.468760848,-0.2717659473,0.1897423416,-0.0198517628,0.018873971,-0.3593865633,0.0818080679,-0.0819114521,-0.0139126172,-0.2698000371,0.2134248316,-0.5696810484,0.3495379686,-0.6191996336,-0.4032836258,0.1359457374,-0.0089497408,0.0068187355,0.2225113958,-0.0832767263,-0.1549550444,-0.0731140822,0.1527770907,0.1555103064,-0.0514137894,-0.0118819149,0.1529569775,-0.1409272999,-0.1005185544,-0.4417811632,0.0856767669,-0.0675195232,0.3428399563,-0.1434997171,0.192402631,0.2198769599,-0.0856729746,-0.0103805698,0.1345742643,-0.0955531895,0.3077327609,0.170032531,-0.0654290542,-0.0525703505,0.2172368616,0.0633547232,0.0598932877,0.2724036872,-0.2143556327,-0.1789783835,0.0925299749,0.3538969457,0.2754508853,-0.1857608259,0.1347743124,-0.2238708287,0.0362664685,-0.1793578118,-0.0393188111,-0.0215975326,0.148763299,-0.0412245616,0.3793972433,0.3718069792,-0.0231432375,-0.0803523809,-0.0416756012,0.1542371213,-0.1227213368,0.2475917786,-0.395832181,0.0751777738,-0.0249076467,0.2735928595,-0.1237475723,0.0956041887,-0.2288337201,-0.0142173003,0.2819485068,0.0854726359,-0.0859705508,0.1893067658,0.3144551516,0.1553159505,0.1471986622,0.2227032036,0.0477154404,0.0851128623,0.5225045085,-0.2288807929,-0.0636966303,-0.0335921571,-0.0285828784,-0.0177003201,0.0306018703,-0.4115174413,-0.099260509,-0.0890588462,-0.4046374559,-0.0264153983,-0.1401497424,0.2578201294,0.1654344946,0.1977567375,-0.2232021838,0.274603188,0.4362157285,0.2820158601,-0.2543868124,0.2374145985,0.1238757446,0.2761558592,-0.0829126909,0.5080304146,0.32897228,0.3917831779,-0.4950323105,-0.0732566193,-0.1334456354,-0.3794184923,0.2263944447,0.0825960413,0.1340554208,-0.34693259,0.2593830228,0.1478417963,-0.0690117031,-0.1664567441,-0.0432168506,0.0238654707,0.1994794011,0.5430538058,-0.2604905367,-0.1268405318,-0.0791174099,-0.1818795651,-0.4427400529,-0.1996071637,0.1990776062,-0.0479121767,0.130148828,0.0324307345,0.0542692132,0.0615516603,0.3101579249,0.1081362292,0.2539687753,-0.3027979732,-0.1292401999,-0.126330018,0.0504745245,0.1605040282,0.0505028889,0.2596039474,0.4793994725,-0.1471509188,0.5145978928,0.0665208921,0.0356536545,-0.1728395373,-0.1022697017,-0.6833411455,0.0515285097,0.1717000157,-0.0296040457,0.1330193281,-0.1457442045,-0.0005717617,0.0577815659,0.0241931882,-0.3936920464,-0.1017920151,0.0639713034,0.2878846526,-0.0906261504,0.0134073263,0.4009183645,0.1135181263,-0.2874775529,0.1107888222,-0.2655998766,-0.0452516414,0.3098522127,-0.0559046008,0.5316380262,-0.0672464445,0.2168063968,-0.1924794912,0.0966423899,-0.0552821234,-0.2583383918,-0.4051496089,0.232138887,-0.194350183,0.0025616775,-0.2144562155,0.2064037323,-0.0294166803,0.437497288,-0.3823658228,-0.2853169441,0.4036141634,-0.1072982773,-0.3715754747,0.3903608024,0.1150050536,0.1893109977,0.1503645331,-0.6484135985,-0.2527800798,0.2446094006,-0.092293337,-0.1966515779,-0.0019911255,-0.1386492252,-0.0159693863,-0.0140757784,0.1917027086,-0.1523194909,0.1952579618,-0.320800215,-0.3016756475]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"I would suggest `nlds`. NLP is a very general, broad and ambiguous term, the library is not about NLP (as in processing) per se, it is about accessing Natural Language related datasets. So the name should reflect its purpose.\r\n","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":39,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nI would suggest `nlds`. NLP is a very general, broad and ambiguous term, the library is not about NLP (as in processing) per se, it is about accessing Natural Language related datasets. So the name should reflect its purpose.\r\n","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"Chiming in to second everything @honnibal said, and to add that I think the current name is going to impact the discoverability of this library. People who are looking for \"NLP Datasets\" through a search engine are going to see a library called `nlp` and think it's too broad. People who are looking to do NLP in python are going to search \"Python NLP\" and end up here, confused that this is a collection of datasets.\r\n\r\nThe names of the other huggingface libraries work because they're the only game in town: there are not very many robust, distinct libraries for `tokenizers` or `transformers` in python, for example. But there are several options for NLP in python, and adding this as a possible search result for \"python nlp\" when datasets are likely not what someone is searching for adds noise and frustrates potential users.","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":143,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nChiming in to second everything @honnibal said, and to add that I think the current name is going to impact the discoverability of this library. People who are looking for \"NLP Datasets\" through a search engine are going to see a library called `nlp` and think it's too broad. People who are looking to do NLP in python are going to search \"Python NLP\" and end up here, confused that this is a collection of datasets.\r\n\r\nThe names of the other huggingface libraries work because they're the only game in town: there are not very many robust, distinct libraries for `tokenizers` or `transformers` in python, for example. But there are several options for NLP in python, and adding this as a possible search result for \"python nlp\" when datasets are likely not what someone is searching for adds noise and frustrates potential users.","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"I'm also not sure whether the naming of `nlp` is the problem itself, as long as it comes with the appropriate identifier, so maybe something like `huggingface_nlp`? This is analogous to what @honnibal and spacy are doing for `spacy-transformers`. Of course, this is a \"step back\" from the recent changes\/renaming of transformers, but may be some middle ground between a complete rebranding, and keeping it identifiable.","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":66,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nI'm also not sure whether the naming of `nlp` is the problem itself, as long as it comes with the appropriate identifier, so maybe something like `huggingface_nlp`? This is analogous to what @honnibal and spacy are doing for `spacy-transformers`. Of course, this is a \"step back\" from the recent changes\/renaming of transformers, but may be some middle ground between a complete rebranding, and keeping it identifiable.","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"Interesting, thanks for sharing your thoughts.\r\n\r\nAs we\u2019ll move toward a first non-beta release, we will pool the community of contributors\/users of the library for their opinions on a good final name (like when we renamed the beautifully (?) named `pytorch-pretrained-bert`)\r\n\r\nIn the meantime, using `from nlp import load_dataset, load_metric` should work \ud83d\ude09","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":53,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nInteresting, thanks for sharing your thoughts.\r\n\r\nAs we\u2019ll move toward a first non-beta release, we will pool the community of contributors\/users of the library for their opinions on a good final name (like when we renamed the beautifully (?) named `pytorch-pretrained-bert`)\r\n\r\nIn the meantime, using `from nlp import load_dataset, load_metric` should work \ud83d\ude09","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"I feel like we are conflating two distinct subjects here:\r\n\r\n1. @honnibal's point is that using `nlp` as a package name might break existing code and bring developer usability issues in the future\r\n2. @pmbaumgartner's point is that the `nlp` package name is too broad and shouldn't be used by a package that exposes only datasets and metrics\r\n\r\n(let me know if I mischaracterize your point)\r\n\r\nI'll chime in to say that the first point is a bit silly IMO. As Python developers due to the limitations of the import system we already have to share:\r\n- a single flat namespace for packages\r\n- which also conflicts with local modules i.e. local files\r\n\r\nIf we add the constraint that this flat namespace also be shared with variable names this gets untractable pretty fast :)\r\n\r\nI also think all Python software developers\/ML engineers\/scientists are capable of at least a subset of:\r\n- importing only the methods that they need like @thomwolf suggested\r\n- aliasing their import\r\n- renaming a local variable","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":170,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nI feel like we are conflating two distinct subjects here:\r\n\r\n1. @honnibal's point is that using `nlp` as a package name might break existing code and bring developer usability issues in the future\r\n2. @pmbaumgartner's point is that the `nlp` package name is too broad and shouldn't be used by a package that exposes only datasets and metrics\r\n\r\n(let me know if I mischaracterize your point)\r\n\r\nI'll chime in to say that the first point is a bit silly IMO. As Python developers due to the limitations of the import system we already have to share:\r\n- a single flat namespace for packages\r\n- which also conflicts with local modules i.e. local files\r\n\r\nIf we add the constraint that this flat namespace also be shared with variable names this gets untractable pretty fast :)\r\n\r\nI also think all Python software developers\/ML engineers\/scientists are capable of at least a subset of:\r\n- importing only the methods that they need like @thomwolf suggested\r\n- aliasing their import\r\n- renaming a local variable","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"By the way, `nlp` will very likely not be only about datasets, and not even just about datasets and metrics.\r\n\r\nI see it as a laboratory for testing several long-term ideas about how we could do NLP in terms of research as well as open-source and community sharing, most of these ideas being too experimental\/big to fit in `transformers`.\r\n\r\nSome of the directions we would like to explore are about sharing, traceability and more experimental models, as well as seeing a model as the community-based process of creating a composite entity from data, optimization, and code.\r\n\r\nWe'll see how these ideas end up being implemented and we'll better know how we should define the library when we start to dive into these topics. I'll try to get the `nlp` team to draft a roadmap on these topics at some point.","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":140,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nBy the way, `nlp` will very likely not be only about datasets, and not even just about datasets and metrics.\r\n\r\nI see it as a laboratory for testing several long-term ideas about how we could do NLP in terms of research as well as open-source and community sharing, most of these ideas being too experimental\/big to fit in `transformers`.\r\n\r\nSome of the directions we would like to explore are about sharing, traceability and more experimental models, as well as seeing a model as the community-based process of creating a composite entity from data, optimization, and code.\r\n\r\nWe'll see how these ideas end up being implemented and we'll better know how we should define the library when we start to dive into these topics. I'll try to get the `nlp` team to draft a roadmap on these topics at some point.","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"> If we add the constraint that this flat namespace also be shared with variable names this gets untractable pretty fast :)\r\n\r\nI'm sort of confused by your point here. The namespace *is* shared by variable names. You should not use local variables that are named the same as modules, because then you cannot use the module within the scope of your function.\r\n\r\nFor instance,\r\n\r\n```python\r\n\r\nimport nlp\r\nimport transformers\r\n\r\nnlp = transformers.pipeline(\"sentiment-analysis\")\r\n```\r\n\r\nThis is a bug: you've just overwritten the module, so now you can't use it. Or instead:\r\n\r\n```python\r\n\r\nimport transformers\r\n\r\nnlp = transformers.pipeline(\"sentiment-analysis\")\r\n# (Later, e.g. in a notebook)\r\nimport nlp\r\n```\r\n\r\nThis is also a bug: you've overwritten your variable with an import.\r\n\r\nIf you have a module named `nlp`, you should avoid using `nlp` as a variable, or you'll have bugs in some contexts and inconsistencies in other contexts. You'll have situations where you need to import differently in one module vs another, or name variables differently in one context vs another, which is bad.\r\n\r\n> importing only the methods that they need like @thomwolf suggested\r\n\r\nOkay but the same logic applies to naming the module *literally anything else*. There's absolutely no point in having a module name that's 3 letters if you always plan to do `import from`! It would be entirely better to name it `nlp_datasets` if you don't want people to do `import nlp`.\r\n\r\nAnd finally:\r\n\r\n> By the way, nlp will very likely not be only about datasets, and not even just about datasets and metrics.\r\n\r\nSo...it isn't a datasets library? https:\/\/twitter.com\/Thom_Wolf\/status\/1261282491622731781\r\n\r\nI'm confused \ud83d\ude15 ","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":265,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\n> If we add the constraint that this flat namespace also be shared with variable names this gets untractable pretty fast :)\r\n\r\nI'm sort of confused by your point here. The namespace *is* shared by variable names. You should not use local variables that are named the same as modules, because then you cannot use the module within the scope of your function.\r\n\r\nFor instance,\r\n\r\n```python\r\n\r\nimport nlp\r\nimport transformers\r\n\r\nnlp = transformers.pipeline(\"sentiment-analysis\")\r\n```\r\n\r\nThis is a bug: you've just overwritten the module, so now you can't use it. Or instead:\r\n\r\n```python\r\n\r\nimport transformers\r\n\r\nnlp = transformers.pipeline(\"sentiment-analysis\")\r\n# (Later, e.g. in a notebook)\r\nimport nlp\r\n```\r\n\r\nThis is also a bug: you've overwritten your variable with an import.\r\n\r\nIf you have a module named `nlp`, you should avoid using `nlp` as a variable, or you'll have bugs in some contexts and inconsistencies in other contexts. You'll have situations where you need to import differently in one module vs another, or name variables differently in one context vs another, which is bad.\r\n\r\n> importing only the methods that they need like @thomwolf suggested\r\n\r\nOkay but the same logic applies to naming the module *literally anything else*. There's absolutely no point in having a module name that's 3 letters if you always plan to do `import from`! It would be entirely better to name it `nlp_datasets` if you don't want people to do `import nlp`.\r\n\r\nAnd finally:\r\n\r\n> By the way, nlp will very likely not be only about datasets, and not even just about datasets and metrics.\r\n\r\nSo...it isn't a datasets library? https:\/\/twitter.com\/Thom_Wolf\/status\/1261282491622731781\r\n\r\nI'm confused \ud83d\ude15 ","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"Dropping by as I noticed that the library has been renamed `datasets` so I wonder if the conversation above is settled (`nlp` not used anymore) :) ","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":26,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nDropping by as I noticed that the library has been renamed `datasets` so I wonder if the conversation above is settled (`nlp` not used anymore) :) ","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"I guess indeed","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":3,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nI guess indeed","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"I'd argue that `datasets` is worse than `nlp`. Datasets should be a user specific decision and not encapsulate all of python (`pip install datasets`). If this package contained every dataset in the world (NLP \/ vision \/ etc) then it would make sense =\/","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":44,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nI'd argue that `datasets` is worse than `nlp`. Datasets should be a user specific decision and not encapsulate all of python (`pip install datasets`). If this package contained every dataset in the world (NLP \/ vision \/ etc) then it would make sense =\/","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"I can't speak for the HF team @jramapuram, but as member of the community it looks to me that HF wanted to avoid the past path of changing names as scope broadened over time:\r\n\r\nRemember\r\nhttps:\/\/github.com\/huggingface\/pytorch-openai-transformer-lm\r\nhttps:\/\/github.com\/huggingface\/pytorch-pretrained-BERT\r\nhttps:\/\/github.com\/huggingface\/pytorch-transformers\r\nand now\r\nhttps:\/\/github.com\/huggingface\/transformers\r\n\r\n;) \r\n\r\nJokes aside, seems that the library is growing in a multi-modal direction (https:\/\/github.com\/huggingface\/datasets\/pull\/363) so the current name is not that implausible. Possibly HF ambition is really to grow its community and bring here a large chunk of datasets of the world (including tabular \/ vision \/ audio?).","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":89,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nI can't speak for the HF team @jramapuram, but as member of the community it looks to me that HF wanted to avoid the past path of changing names as scope broadened over time:\r\n\r\nRemember\r\nhttps:\/\/github.com\/huggingface\/pytorch-openai-transformer-lm\r\nhttps:\/\/github.com\/huggingface\/pytorch-pretrained-BERT\r\nhttps:\/\/github.com\/huggingface\/pytorch-transformers\r\nand now\r\nhttps:\/\/github.com\/huggingface\/transformers\r\n\r\n;) \r\n\r\nJokes aside, seems that the library is growing in a multi-modal direction (https:\/\/github.com\/huggingface\/datasets\/pull\/363) so the current name is not that implausible. Possibly HF ambition is really to grow its community and bring here a large chunk of datasets of the world (including tabular \/ vision \/ audio?).","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/138","title":"Consider renaming to nld","comments":"Yea I see your point. However, wouldn't scoping solve the entire problem? \r\n\r\n```python\r\nimport huggingface.datasets as D\r\nimport huggingface.transformers as T\r\n```\r\n\r\nCalling something `datasets` is akin to saying I'm going to name my package `python` --> `import python` ","body":"Hey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.","comment_length":39,"text":"Consider renaming to nld\nHey :)\r\n\r\nJust making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing.\r\n\r\nThe issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This means the package makes `nlp` a bad variable name everywhere in the codebase. I've always used `nlp` as the canonical variable name of spaCy's `Language` objects, and this is a convention that a lot of other code has followed (Stanza, flair, etc). And actually, your `transformers` library uses `nlp` as the name for its `Pipeline` instance in your readme.\r\n\r\nIf you stick with the `nlp` name for this package, if anyone uses it then they should rewrite all of that code. If `nlp` is a bad choice of variable anywhere, it's a bad choice of variable everywhere --- because you shouldn't have to notice whether some other function uses a module when you're naming variables within a function. You want to have one convention that you can stick to everywhere.\r\n\r\nIf people use your `nlp` package and continue to use the `nlp` variable name, they'll find themselves with confusing bugs. There will be many many bits of code cut-and-paste from tutorials that give confusing results when combined with the data loading from the `nlp` library. The problem will be especially bad for shadowed modules (people might reasonably have a module named `nlp.py` within their codebase) and notebooks, as people might run notebook cells for data loading out-of-order.\r\n\r\nI don't think it's an exaggeration to say that if your library becomes popular, we'll all be answering issues around this about once a week for the next few years. That seems pretty unideal, so I do hope you'll reconsider.\r\n\r\nI suggest `nld` as a better name. It more accurately represents what the package actually does. It's pretty unideal to have a package named `nlp` that doesn't do any processing, and contains data about natural language generation or other non-NLP tasks. The name is equally short, and is sort of a visual pun on `nlp`, since a d is a rotated p.\nYea I see your point. However, wouldn't scoping solve the entire problem? \r\n\r\n```python\r\nimport huggingface.datasets as D\r\nimport huggingface.transformers as T\r\n```\r\n\r\nCalling something `datasets` is akin to saying I'm going to name my package `python` --> `import python` ","embeddings":[0.354604274,0.0250588469,-0.0843049809,-0.3304423392,0.1876242757,-0.2360132337,0.281853199,0.2058823705,-0.1652067304,0.0582918525,0.2198298723,0.294048816,-0.3030344248,0.1246296465,0.3023298979,-0.2793754935,0.2086221278,0.0491361916,0.1632865071,-0.1286140531,-0.1061515063,0.1799212396,0.1566794068,0.4202884436,0.0016413978,0.0405684561,0.2510459423,-0.1025928706,0.0196671039,-0.3666577637,0.0917976722,0.3173364997,-0.0723683611,0.0078898594,-0.000103016,-0.1667028368,0.5983600616,-0.0208175331,-0.1951327026,-0.2738524973,0.349458456,-0.4347401261,0.0873838514,-0.5374493003,0.1049486846,0.2493642569,0.1611560434,-0.2769201994,0.1723022461,-0.08421278,0.2791434228,-0.0322697014,-0.2336917669,0.2664329112,-0.0823114812,0.1612648368,-0.1856753528,-0.1456345767,0.1019584611,-0.0416525863,0.0535839312,0.2429705113,-0.1371386796,0.0111304624,0.4416790307,0.1220082119,0.3055425584,0.0429867581,-0.0890983418,0.1462106407,0.016002601,-0.3754999638,-0.227576077,-0.1618045717,0.0141653046,-0.4211737514,0.0508646816,-0.0911910012,-0.0744974241,-0.0652683526,0.3642609417,-0.1167289987,0.0555218421,0.1869729459,0.0563985147,0.6406509876,0.1287473142,0.1855204701,0.0186784174,-0.2868248522,0.1609270722,0.1716974378,0.2048975378,0.0659400746,0.1589836776,-0.1076443791,0.0360815078,-0.0260845404,-0.1019777954,-0.0282877255,-0.0157899559,-0.0529250763,0.1841489673,-0.1026077643,-0.0229196306,-0.0395875387,0.2494286448,0.06806878,0.3341001272,0.0483051986,0.1616839319,0.0051702829,-0.017028138,0.1136647984,0.0018841785,0.0539529808,0.1125142872,0.0556235351,0.0231455937,-0.0699977055,0.139364928,-0.0318878219,0.012281891,0.2282057405,0.0296439026,0.4047164619,0.1179114431,-0.0512215644,-0.0545108691,0.0529048257,-0.1768298894,-0.0293129664,-0.5875118375,0.1920241714,-0.0093172109,0.5461207032,0.3421474099,-0.2364342958,-0.1096441969,0.0759178326,-0.0507352985,0.1224967614,-0.0358524546,-0.0496277064,-0.4342936575,0.1021536589,-0.2238204181,-0.3176436424,-0.2762625515,0.1289541721,-0.2020074129,0.0938013867,-0.1377503276,0.3023263812,-0.2851534784,-0.0645210966,0.4819487929,0.1596666574,0.13196069,-0.2455682158,0.1019591764,-0.0245489255,0.1727491468,-0.168171823,-0.2812473476,0.0770907551,0.1417668909,-0.4519435167,-0.5630795956,0.0712455139,0.3810745776,0.0145662418,-0.1681337059,-0.0279533472,0.1246093661,0.0463554375,0.5702139735,-0.3770810962,-0.1004941016,0.110865213,-0.2571691275,-0.1666597575,0.0733325183,0.1724896133,-0.0463332757,-0.1781515777,0.2236127108,0.4268013537,0.168226257,0.1578955799,-0.2671319842,-0.1424053609,0.367061913,0.0201186519,0.0713867247,0.1450184584,-0.2018309981,0.7421375513,0.1609519124,0.0481394902,0.0549914017,0.1300477087,-0.0001047266,0.2264642864,0.052703809,-0.04857428,-0.2530002296,-0.1717445701,-0.3851439953,0.2766966522,-0.140701443,-0.3179916739,-0.0011529682,-0.0611749887,0.0840609074,-0.2455073893,0.3142194748,0.1664161682,-0.0728963241,-0.1845241189,-0.3374127746,0.2081885487,0.1162087247,0.0268573985,-0.1347388923,0.0125924638,-0.0472036041,-0.1227353066,-0.0485320874,0.4621073008,0.0617956594,0.0230287537,-0.018729629,0.1212337241,-0.0920708776,-0.1717240363,-0.1780800521,0.352711916,0.1953412741,0.0335730724,0.1522716731,0.217854023,0.0052392827,0.0609577782,0.1654213071,0.0910024494,-0.2310306877,0.172860235,0.1610245407,-0.1268762499,0.2499854863,-0.3104695678,-0.3951842785,-0.1594946086,0.2015538812,-0.1744659841,0.3688205183,0.2622098029,-0.0718516931,-0.0054821074,0.7487162948,-0.102349095,-0.0018591094,0.0346931629,0.0061895815,0.0344388932,-0.0982624963,0.4255813956,0.3692655265,0.2573755383,-0.2147238851,0.0731224418,-0.1584894955,-0.2395242006,0.1479514241,0.2729809284,-0.02106102,-0.0159713048,0.0722971559,-0.0208247695,-0.1442234069,-0.3758417368,0.1088091359,0.0435103104,0.1615836322,0.196914956,-0.2253728807,-0.3355076909,-0.3560675979,-0.1077126339,-0.5679337978,-0.212283805,0.0389435887,-0.319509536,-0.3220556676,0.1848625839,0.0278035179,0.0538053773,-0.3259667754,-0.4077161551,-0.2541850805,-0.2610703707,-0.2709332705,0.101353474,-0.0553875603,0.121204786,0.3438507915,0.075265944,0.0450470969,0.079796873,-0.4759514928,0.0199883431,-0.0811674446,0.2733127177,0.2104487866,-0.0825033486,0.1517145187,-0.4660096467,0.1668881476,-0.2525931001,0.1693013757,0.0579317883,0.152862072,-0.048633296,-0.3818205595,-0.2003627717,-0.4499168396,-0.431799382,-0.1190716177,-0.0393395089,0.0792104751,0.4727215767,-0.575882256,-0.0751322135,0.1555481106,0.2960141897,-0.1504534483,0.3245946169,-0.0218335297,-0.1896303445,0.0030149538,-0.3701182008,-0.4601126313,0.5443202853,-0.0602085814,0.0033655989,0.0939403698,0.062001247,-0.1237863004,0.2940612435,-0.1174602509,0.3699588478,0.3588147461,-0.1884727329,-0.0632570162,-0.1866520196,0.1633792967,-0.0045336215,0.3885166049,0.1221950576,-0.1400718987,-0.0451746397,0.021036813,-0.2248969823,-0.0801913887,0.2310665101,0.3564607799,0.0327313393,-0.0163829681,-0.1170815378,0.0942937359,0.1725271791,-0.1195733771,0.2910572588,-0.084376283,-0.2946411967,-0.0038447217,0.216943264,-0.1007227972,-0.250341922,0.1929998994,0.0891832486,0.2827481031,0.1753280312,-0.1619341373,-0.1459892541,-0.3150652945,0.1284447312,0.225634858,-0.2269299328,0.0859745294,-0.5130316615,-0.3030593395,-0.085491918,-0.0233060848,-0.0840170681,0.2252946943,-0.0976991206,-0.0682784021,0.0381608866,-0.0622195601,0.1355437487,0.0053455988,-0.1475559324,0.3124498725,0.2203025371,-0.0112975044,-0.2311531156,-0.1663774699,0.0577343851,0.2586653233,0.2246444225,-0.5831884146,-0.3143754601,0.1958585382,0.0615963861,-0.1887239814,-0.2447848618,-0.4139582813,-0.2432189435,-0.2267659754,-0.2075638622,-0.1328572482,0.3199290633,0.0897640884,0.0595341958,0.1999856383,-0.2511616051,0.2202599198,0.1723578721,0.1237168163,0.0274906456,-0.1680532396,-0.0923233703,0.2246270031,-0.5235324502,0.0450229906,-0.0399270877,-0.0523353666,-0.1777237654,0.053336218,0.3100294471,0.3402785361,0.0900301486,0.2383213341,-0.1819750071,0.064058736,-0.0300185215,0.2332521677,0.1558601111,0.0364174955,-0.2073982507,-0.4303261638,0.2744522691,0.2052860409,-0.259930253,-0.0256312788,0.020775266,-0.380427599,0.5902324319,0.1180906296,1.0089893341,-0.0025154315,0.2391269356,0.0166731719,0.0414852723,0.7351976633,-0.2273076028,0.0742217749,-0.2871640325,0.433824122,0.2762079835,0.0291307736,0.1893188506,0.1501968354,-0.2432343513,0.4527542293,0.2205487341,-0.3383801281,0.0835853443,0.4536221325,0.165615797,-0.3907467425,-0.1250927299,0.1206355318,-0.4738864005,0.2193241715,-0.159949705,-0.2089595199,-0.113875024,-0.1106687263,0.0419254713,-0.0078130104,-0.0255674571,0.1110911146,-0.1030508131,0.1334174424,-0.1465845257,-0.2319064736,0.3092885911,0.445115298,-0.041815605,0.0672231317,-0.252936542,0.0395187624,0.1476794928,0.2119449675,0.3646717072,-0.1735542417,-0.1169806719,-0.1212965623,0.0693862513,-0.081658043,-0.3511219025,-0.1439371854,0.3010071814,-0.1273344457,-0.0553538501,0.269705236,-0.2059503943,-0.0515002906,0.2535747588,0.1286807805,-0.2365042269,0.1522155851,0.2154997736,-0.137998566,-0.0277184751,-0.2845080197,0.2966329157,-0.2628991008,-0.0141333407,-0.2800219059,-0.1900845021,-0.1783280671,-0.214416489,-0.4408949316,-0.2746851742,0.2680206299,0.3854089677,-0.6574815512,-0.1242854074,-0.0738777742,0.1040904894,0.299398005,-0.3066925704,-0.0836870968,-0.2827759683,0.445016712,-0.1112826467,-0.0271209963,0.050892815,0.6008216739,-0.0077933767,0.0307250191,-0.4114221931,-0.2275705636,-0.2836341858,0.350359112,0.0242581032,-0.4060619771,-0.2008109242,0.1526447088,0.195652619,0.0458796993,-0.1764339209,-0.2835904658,-0.3188830316,0.1050617769,0.0745702311,0.134276852,0.0811204165,-0.1392993331,0.2841313183,-0.0353498235,-0.1470662206,-0.0785914734,0.1584579647,0.6412721276,0.1998638064,0.2127105147,0.2033691555,0.1251465231,-0.0351963043,0.2976452112,0.0217048787,0.2803504467,0.0442090183,-0.0622191727,-0.2232208103,-0.1420169771,-0.237342447,0.3103158772,0.0100148749,-0.1164275259,-0.1670699567,0.0692104995,-0.0368862636,0.358656019,-0.1238265708,-0.2624826431,0.2713677287,0.2969161272,-0.3307812512,-0.0596150421,-0.0284350161,0.0032089942,0.1284840107,0.2637521923,0.0675547868,-0.083330974,0.2330919206,-0.0484737791,0.4031938016,-0.3050514758,0.0193954241,0.1950927377,0.0151604628,0.1632584333,0.305824995,0.0028242378,0.0275518503,0.3956401646,0.1797771901,0.1889588237,0.3612711728,0.121343635,0.1256159544,0.5455904603,-0.229820326,0.1712244898,0.4129389524,0.361943692,0.1206745133,-0.3275887668,0.0675624683,-0.2218260616,0.0876736641,0.1275337487,-0.1785115004,-0.2183451205,0.2007020563,0.349712044,-0.4002241492,-0.1088244915,-0.0759370029,-0.0582725666,0.0093693361,0.0185799748,0.079120867,-0.1426203996,0.095606342,-0.0189035628,0.161790356,-0.1936123818,0.0031982968,0.0838395506,0.1933111548,-0.0581986196,0.0555206537,-0.0990936682,0.0866527781,-0.2270681113,-0.0916607678,-0.2489612699,0.0196249783,-0.2773872018,0.2346076816,0.4263944924,0.5700036287,0.1071951762,0.1036822498,-0.2625008225,-0.1466628313,0.2125212848,-0.3936911225,-0.0943321735,0.8848905563,0.2048438638,-0.1527258903,-0.2799520791,0.1773991287,-0.7367970943,0.0776504874,0.2374967635,0.1197139844,-0.0409560129,-0.4269070625,0.1364692748,0.1251665205,0.2762799561,-0.0520887934,0.4124681354,-0.0601217262,-0.0616205819,-0.4459729493,0.0975326449,-0.0362376943,0.1628293991,0.0209497977,-0.1065493226,-0.0380697697,0.0391867161,-0.293356806,0.3067795038,-0.3621272743,-0.342715621,-0.2458036542,0.0772687942,0.014666358,0.0181849785,0.1471748501,-0.2030318826,-0.0983984843,-0.1064845994,0.0504437648,-0.2307023406,-0.1302485317,0.482693702,0.4234507978,0.1074437574,0.267604351,0.1654308736,-0.1382891238,-0.2191936374,-0.2272736281,-0.151767835,-0.2155966461,0.127609238,0.1082273871,0.1141461655,0.1075206473,0.0941387489,-0.2939921618,0.0820912346,-0.1551926881,0.0133389626,-0.2642064095,0.0743817091,0.1357637495,0.1215907633,-0.4070495069,0.4321894646,0.006867812,0.0536034741,-0.1758133322,-0.3169422746,0.5040283799,-0.0870549455,-0.3847106397,0.3119255602,0.2451522946,0.1422896087,0.0038849178,-0.3304694593,-0.1347045004,0.0382817574,0.0473495126,-0.2229124755,0.1495041102,-0.0536104068,-0.1715639681,-0.0246996321,0.0344616696,0.1170691848,-0.3060974479,0.2864738405,-0.232171908]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"I second this request. The bottom line is that **scores produced with different reference tokenizations are not comparable**. To discourage (even inadvertent) cheating, the user should never touch the reference. The `v13a` tokenization standard is not ideal, but at least it has been consistently used at matrix.statmt.org, facilitating comparisons.\r\n\r\nSacrebleu exposes [all its data sources](https:\/\/github.com\/mjpost\/sacrebleu\/blob\/master\/sacrebleu\/dataset.py) and additionally provides [an API](https:\/\/github.com\/mjpost\/sacrebleu\/blob\/master\/sacrebleu\/__init__.py) to accessing the references, which seem to fit within the spirit of your codebase.","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":74,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nI second this request. The bottom line is that **scores produced with different reference tokenizations are not comparable**. To discourage (even inadvertent) cheating, the user should never touch the reference. The `v13a` tokenization standard is not ideal, but at least it has been consistently used at matrix.statmt.org, facilitating comparisons.\r\n\r\nSacrebleu exposes [all its data sources](https:\/\/github.com\/mjpost\/sacrebleu\/blob\/master\/sacrebleu\/dataset.py) and additionally provides [an API](https:\/\/github.com\/mjpost\/sacrebleu\/blob\/master\/sacrebleu\/__init__.py) to accessing the references, which seem to fit within the spirit of your codebase.","embeddings":[-0.2068963349,-0.2239939272,-0.0241725501,-0.1960566938,0.1427302063,-0.2892453372,0.204468295,0.2425775528,-0.5539250374,0.4257799089,-0.2530024648,0.3626277447,-0.013167588,-0.078750357,-0.1742032766,0.1995098889,0.1531567276,0.1220696718,0.5717875361,-0.0796073452,-0.098993592,-0.1696630269,0.0176084861,0.2470569164,-0.3446577191,0.3597022593,0.2996474802,-0.2794865966,-0.1673764884,-0.2797946036,-0.1365122795,0.2227185071,-0.279530853,0.2289049923,-0.0001093052,-0.3120647073,-0.0601100661,-0.0350928679,-0.265420258,-0.4642713666,0.241164431,0.0276452024,-0.055413615,-0.1108257845,-0.0504698008,-0.2427429855,0.04569114,0.069631584,0.3667047322,-0.1348719299,0.1769615263,0.429012835,-0.4011058509,0.154327929,0.1258075386,0.2085044235,-0.0326999277,0.267272681,-0.0411467366,0.3646010756,0.0584651232,0.0565019138,-0.1211043224,-0.1393247098,0.1359600276,-0.0589449778,-0.2566533685,0.0117939906,0.2010882646,0.457672745,-0.0228146631,0.0737749264,-0.2328153402,-0.1242059618,0.0908815861,-0.3574267328,0.0922226533,-0.0738506392,0.1251289546,-0.175254941,-0.0502342358,0.3113988638,0.1208132952,0.063946411,0.1400031447,0.0747234672,-0.0462033637,0.0561071634,-0.0482494794,-0.0372716971,-0.1836458296,0.102579698,0.0456452817,0.0721746534,-0.2898781598,-0.1666918546,-0.1875944585,-0.5826480985,0.1071780995,0.061543446,0.506480217,0.2740062177,-0.1688214391,0.1206374317,-0.1836244464,0.0766869038,-0.266623348,-0.0895998031,0.3690304756,0.0124740377,0.2950677574,0.2305250466,0.2302126139,-0.1735688746,-0.1869598627,0.2104580849,-0.3297913969,-0.2119521946,-0.6489787102,0.0244919956,-0.2098716348,-0.0039327047,-0.1035745293,0.155200839,-0.2394098043,-0.1720015258,-0.0923475325,0.0038391147,-0.1732512563,-0.1808503717,-0.0416512713,-0.2452546954,-0.1640869975,-0.0652837753,-0.0555017143,0.0853482932,0.2829969823,0.0482707955,0.5071291924,-0.1670744121,0.0395633914,-0.1643660367,0.1284259409,-0.0809514001,-0.3301097155,-0.1837484092,-0.1064810678,-0.1753861755,-0.0757575631,0.1259743869,-0.265085876,-0.1970326006,0.219871372,0.1541852504,0.108176291,0.13014552,-0.4073212147,0.0986018553,-0.1138292104,-0.0174881369,0.2405324578,-0.1037751287,-0.1610575616,-0.2548727691,0.2549302876,-0.2436173111,0.1269932836,-0.184952125,0.5565254092,0.2552277148,0.1769501865,0.3732342124,-0.2868903279,0.2685480714,0.0757662728,0.175688982,0.0247070957,-0.0729766861,-0.2471740544,-0.1694583148,-0.1363464594,-0.2782011032,0.0569716506,-0.0022902002,0.4476939738,-0.1601079553,0.0129348934,-0.0264290869,0.1478295028,-0.0873768926,-0.3542244434,-0.5578355789,0.0816820934,-0.1322261095,0.0789171681,-0.0438447334,0.2843369544,0.5939742327,0.3166588843,-0.3041429222,-0.1074525639,-0.282163918,0.5762838125,-0.0202748664,0.071305953,0.1576715857,0.4550969601,0.1679539531,-0.0573650338,0.1576075256,0.1244589686,-0.4581418335,-0.0465322882,0.016293034,0.2220989764,0.0366160981,0.0769572705,-0.1002092585,-0.2054637372,0.1872119606,-0.4104688466,-0.2364171296,-0.5289868116,0.1547191739,0.1777647734,-0.0728960037,0.032057967,0.0623840354,0.136107713,0.3638375103,0.0351115428,-0.0589805134,-0.2578704953,0.3002849519,0.0624148734,0.0660421774,-0.3768453598,0.6890140772,0.3303513229,0.2215900868,-0.2227139771,0.1267667115,-0.1263362169,0.131438598,0.3574001491,0.5529800057,0.0214897841,-0.0761900619,0.0505527705,-0.2885354161,-0.2127612978,0.1523654759,-0.0368538648,-0.2658447027,0.0508229509,0.2942456603,-0.255817771,0.113773331,-0.183015734,-0.0119641284,0.1617596,0.0664129779,0.0226847734,0.0442922451,0.3693778217,-0.1273409277,0.0505143665,0.1427875906,0.0556401312,0.0483270921,0.0977223888,0.1243292987,-0.0088804727,-0.3289555907,-0.142603606,-0.0059857522,0.0289363433,0.1889076829,-0.0027061421,0.3297759295,-0.0355942398,0.016512664,-0.044131849,-0.1556869149,-0.0953174159,0.1654359102,-0.2257624418,0.2198485583,-0.2951077819,0.0351456702,0.0170469228,-0.1059285402,0.2216109484,0.2815543413,-0.1472128183,0.3347761631,0.0994495004,0.6170318723,-0.2452239096,0.0372117497,0.1176237017,0.1728842258,-0.323358357,0.0711140335,-0.1702132821,-0.3984040022,0.2559493482,-0.1248623133,-0.2755087018,-0.1597879082,-0.5918516517,0.0459023975,0.0396379679,0.0232888125,0.141442582,-0.5394570827,-0.4638618231,-0.2547611594,-0.1199979037,-0.2510250509,-0.1145920381,-0.0606336184,-0.0050708628,-0.1690470427,0.0162940826,-0.2947101593,0.1643231362,0.11852213,0.1338732839,-0.2773308754,0.1693721563,-0.2163564712,0.0118126739,0.1174274459,0.1643926054,0.3501387239,-0.212271437,0.1290461868,0.1318792701,-0.1976172775,-0.3267780542,-0.0503234938,0.0647123381,0.4233525395,0.1470820457,-0.2939503193,-0.2734202445,0.3738844395,-0.1731691062,-0.1254466772,-0.2122832835,-0.2605549991,0.0583349876,-0.1946227103,0.1922216117,0.1158105209,0.2132240683,-0.0956216007,0.4491807222,0.1187028587,-0.0395866819,0.6627027988,0.6913072467,0.1017229483,-0.1586233526,-0.0426356196,0.1583472788,-0.0875410885,-0.0045996848,0.0941019729,0.1971177161,-0.1294559687,-0.2560731173,0.3780324161,-0.1585333496,0.1232013479,0.1082131788,-0.1611289233,0.2057590187,-0.1643342227,-0.0010713548,-0.0017701952,0.095025897,0.0110170702,0.0680541769,0.1290991604,-0.4127894938,0.3153252304,0.157787919,-0.2213689834,-0.0144194122,-0.2104133219,0.0052283504,-0.3326506615,0.288944751,-0.0296189301,0.6624233723,0.0117718112,0.0834729746,0.1709177047,-0.0997462347,0.0504183136,0.1579196155,0.0903204456,-0.2175955772,-0.0257565677,-0.1992521286,-0.0704754218,-0.0727046654,-0.099099122,0.3267843127,0.2176624686,-0.3120601475,-0.1621986479,0.01944042,0.0851323828,-0.1260425299,0.1237159669,-0.4543351531,-0.0649406537,-0.0557297394,-0.0247003864,-0.1872989088,0.045034606,-0.2511964738,-0.05046257,-0.4296769202,-0.0678881183,0.2717933357,0.0868355855,-0.0631870106,0.1846376061,0.031705156,0.0364549123,0.3250904679,0.2090857923,-0.1376556307,-0.2794860601,-0.3214729726,-0.1144843623,-0.2028517127,0.0953914672,0.4196400046,-0.1740463525,0.4768792987,-0.0157344434,0.0112018771,-0.2375890166,0.3723275661,-0.0375473909,0.325984925,0.1532134265,-0.1466741711,0.0493439026,0.1555635929,-0.3171954155,0.0094123278,0.2422333807,-0.129157573,0.6084893942,0.72089082,0.8764437437,0.1259093136,0.1141559109,0.2070011497,-0.1199536249,0.2565965652,-0.6042691469,0.2072224319,-0.3428195417,-0.4102155566,0.0232140534,-0.1400771886,-0.1746553332,0.136953786,0.1619715244,-0.0327282101,-0.090376012,0.301217407,0.0788518488,0.2614023685,0.1506234705,-0.0877257064,0.2132781446,0.1011496708,-0.15149194,-0.0306136571,-0.082629092,0.352784574,-0.2943233252,0.1250689477,0.1902915388,-0.0342689902,-0.3502195477,0.1137002781,0.1317372322,0.1885910928,0.2218080759,0.2108040899,0.8279297352,0.433014214,0.0082829501,-0.1075669453,0.0648680702,0.1331890374,0.0712042227,-0.3538218439,0.3103142679,-0.0606912225,-0.1048767269,0.2527482808,0.5873185992,-0.1746493131,0.2403130382,-0.252007395,0.048712261,-0.1159991995,0.1788545102,0.2137910724,0.217363745,-0.0209003314,0.1036422849,-0.2565672696,-0.21369192,-0.0037656033,0.2833054364,-0.4414062202,0.0381095186,0.2712467313,0.1722855866,-0.3404607773,0.1972760409,-0.5864215493,-0.0897703543,0.0221272744,-0.0384003147,-0.0596899316,-0.0833159089,-0.0765719563,-0.4314348996,-0.1352991611,-0.2460096776,0.5607563853,0.6052772999,-0.0774891451,0.0336492844,0.135410741,-0.1630395055,0.2133660764,0.1951995939,0.2581731975,0.0349092335,0.0533160977,-0.3790338337,0.1731288135,-0.2918667495,0.0644071326,-0.1014000699,-0.003360217,-0.2342807949,-0.0441532694,0.0775413066,0.1023194939,-0.0481504984,0.0743544549,-0.4080381095,-0.0953465104,-0.3948292732,0.1093238667,0.2661676705,-0.1497953683,-0.3956966698,-0.1315514445,0.0635767505,-0.1323179007,0.1926935613,0.1348089427,0.1516891569,0.1362864226,-0.1986281276,0.3296236098,0.0634748638,-0.1438185722,0.1717550308,0.2343725264,0.3583288491,0.1249591634,0.1032916084,-0.0910669416,-0.0482546799,0.1773389727,0.2758865356,-0.1333847046,-0.1384397894,-0.1345929652,-0.1364497244,-0.3582069576,0.0131529672,0.2146361023,-0.1268167943,-0.7645155787,0.4478642941,0.148692295,-0.0601478033,0.0531819463,-0.1367063671,0.2776691616,-0.0705252662,0.3877398372,0.9065315723,-0.0126677956,-0.0706975758,-0.0528647676,-0.1292687953,-0.4736337364,0.1823346466,-0.3895036876,0.0389170796,0.348603487,0.2279839814,-0.0590303205,0.1209801883,-0.1988398135,-0.0722639039,0.6388521791,-0.0506130829,0.0819868296,0.100170061,0.2426890731,0.4570908844,0.1935916692,-0.1912609637,0.05711044,0.2640196681,0.4476599693,0.1914542466,-0.3882978857,-0.1500757933,-0.0603224076,-0.208774209,-0.2626775801,-0.1044183969,-0.2759130299,-0.2582128644,0.1911939979,-0.1134684682,0.1549495012,-0.4421712458,0.0690820068,-0.0469769835,0.1085620895,-0.0824249387,0.2035619169,0.4655303657,-0.103663519,-0.256113112,-0.2180809826,0.2237272114,0.1714991778,0.2189146429,-0.1410621554,0.1432940662,-0.142743513,-0.4396961331,-0.1719785035,0.4242268503,0.189185366,0.2286869884,0.2140737623,-0.3947884142,0.3290734887,0.0312089864,-0.1289631277,-0.2156376988,0.2847797871,0.1506447196,-0.2831283808,0.1975285262,0.1725680977,0.1861095726,0.1726138592,0.1169576719,-0.2801846266,0.3426784575,0.2673277557,-0.1990315318,0.269644767,-0.0757648051,0.0998394564,0.232134074,0.1680521667,-0.0124239177,-0.012970834,-0.1786152422,0.1300774068,-0.2532233894,0.4477965534,0.437833637,0.308998704,-0.0724688023,0.0309735481,-0.1884554327,0.1401493996,-0.195552513,-0.0951177105,-0.2162395567,-0.1527872086,-0.1156388819,-0.0487922728,-0.3362261951,-0.0385470018,-0.2555797398,-0.0223907195,-0.0252561923,-0.3975221217,0.0080473479,-0.3003574312,-0.1571476609,0.1437079757,0.2694339156,0.4956658781,0.2392160892,-0.1024476364,0.2482331842,-0.2371189147,0.3361227512,0.2244512588,-0.2408749014,0.1394227147,-0.0071252505,-0.0012073439,0.0741440132,0.2098388374,-0.2156430483,0.4213053882,0.452026993,-0.6470118165,-0.2663901448,0.0020055878,0.2271243036,-0.1412020475,-0.0937133133,0.4196790755,0.1627278924,0.425247252,-0.1409706622,0.12925978,0.2413696945,-0.4877031446,0.2607187033,-0.205283761,0.3699526191,0.434586674,0.0155136725,-0.6833723187,-0.2294603735,0.1836051047,-0.0996363461,-0.2657682896,0.2009209394,-0.1180143878,-0.1178948507,-0.2908832431,0.2428510636,-0.2789718509,-0.312399894,-0.017279679,-0.2478001267]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"Didn't we have a slide and discussion at WMT admitting that, for production-quality models, BLEU doesn't correlate with human eval anyway?\r\n","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":21,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nDidn't we have a slide and discussion at WMT admitting that, for production-quality models, BLEU doesn't correlate with human eval anyway?\r\n","embeddings":[-0.2491149306,-0.3357783258,-0.0347295776,-0.3980183005,0.1861638427,-0.2238558233,0.3552849293,0.1395354867,-0.3892639577,0.4466587007,-0.0967207775,0.2908814251,-0.1526997685,0.0279197823,-0.1170512065,0.204105854,0.1373981833,-0.009141366,0.6749675274,-0.1052772626,-0.1155280322,-0.2147949338,0.0487757102,0.3603987098,-0.2367162108,0.1981560141,0.3498717248,-0.2481336147,-0.1169564575,-0.2331206053,-0.2685071528,0.1880480051,-0.1370758861,0.0329764225,-0.0001076734,-0.3473198414,0.0011731707,0.0119105773,-0.3443155587,-0.4313924015,0.1730440706,-0.0576675162,-0.0739226341,-0.0539854728,0.1012308076,-0.0790880695,0.1463137269,0.0757335424,0.3050436378,-0.0967882499,0.1885267943,0.329490006,-0.5472923517,0.2496242374,-0.003958114,0.2622883618,-0.0547813512,0.1251739413,0.0734651908,0.287226826,0.0471568629,-0.0566911995,0.0409773998,-0.2722392082,0.0702089295,-0.0097833555,-0.1277758926,-0.1290709823,0.1420107037,0.5171830654,-0.0779175162,0.1853195429,-0.2592955828,-0.16466254,0.1634196043,-0.490164578,0.1457343847,0.0001003969,0.0894526467,-0.108816199,-0.0343419313,0.1923368722,0.1737521142,0.0310319327,0.122849606,0.1572470963,-0.0906207114,-0.011721557,-0.1866638958,-0.0387266576,-0.2872310579,0.1434111148,0.0214997251,-0.0525690764,-0.2434096038,-0.0831254348,-0.1516965479,-0.4444977343,0.0388033465,0.0838735029,0.4004847109,0.2629838288,-0.1080263034,0.1032519564,-0.1735090315,0.163530618,-0.1840124726,-0.2250757217,0.3822994232,0.0213678628,0.3974565864,0.1239967048,0.1788043231,-0.1456840634,-0.1180793121,0.1900229305,-0.3556284904,-0.1888513863,-0.7770127654,-0.0080201868,-0.2706565559,-0.0599770397,0.003641607,0.1757609993,-0.2525804043,-0.0737681836,-0.1152373552,0.0448658764,-0.0611337647,-0.247363314,-0.0125320265,-0.3729315698,-0.208776176,-0.1156876832,-0.1807078868,0.0068790852,0.2841663361,0.1106683388,0.4803144634,-0.1324008554,-0.0027704767,-0.101350978,-0.0555226691,-0.0896622017,-0.3398228288,-0.1305007935,-0.174497366,-0.2144024819,0.057472147,0.1234654337,-0.2911875844,-0.323586911,0.1811128706,0.1271613091,0.122085087,0.1307637393,-0.361689955,0.2228198647,0.0336335003,0.2459506691,0.2967767715,-0.2096000612,-0.0243234225,-0.2422040999,0.1906651407,-0.387170136,0.0432689153,-0.2791077793,0.5192180872,0.2964473069,0.3169187307,0.4025864303,-0.2807274461,0.3707182109,0.0726590082,0.2284861505,0.114445664,0.0976476148,-0.0794755891,-0.0209455863,-0.1425353587,-0.1864728034,0.0257972851,-0.0516279377,0.3966948986,-0.1670402288,0.0254637133,0.0985415876,0.0494866148,-0.1148233861,-0.354819715,-0.5579895377,0.3521686494,-0.1113125458,0.1058301032,-0.0760886297,0.216993615,0.5703735948,0.329774797,-0.3007261157,-0.1977592707,-0.4047829211,0.6597766876,0.0785528421,0.1278491765,0.316111207,0.4889910221,0.0977325514,-0.1304489076,0.1759990007,0.1449162364,-0.3879233599,0.0350165144,0.0397131629,0.2069048285,0.0064133396,0.0390690416,0.0081260586,-0.4279516935,0.1908796579,-0.5212842226,-0.1482566744,-0.5427168608,0.1451872289,0.1458370984,-0.1478000283,-0.000831613,0.1121125296,-0.0124797821,0.2398341894,-0.0534303263,-0.0295754876,-0.2044357061,0.2522409558,-0.0833936855,0.0158802737,-0.4189137518,0.5335503221,0.3664586544,0.1977936178,-0.351101011,0.2762394547,-0.1181578338,0.1622233093,0.3843787611,0.4806502461,0.1684254259,-0.0413100459,-0.0516036674,-0.2080800384,-0.1055176258,0.0972293988,0.0367422998,-0.1439665556,0.1675713211,0.1346848011,-0.1431772709,-0.0929022431,-0.2031818181,-0.0054223579,0.2441968322,0.0609391332,0.1806289405,0.0542811044,0.5331862569,-0.2062664479,0.1067077667,-0.077581048,-0.0115666008,0.0520081781,-0.0053719799,0.1069076881,-0.144276917,-0.3583911061,-0.0926376283,0.0081982678,-0.1586504132,0.1531457156,-0.0389769785,0.343019098,-0.1362837255,-0.1561332941,-0.0443779379,-0.1271745116,-0.1358420104,0.1161138266,-0.2197365165,0.2930183113,-0.0165710635,0.1127933189,-0.0896846876,-0.1608657539,0.2289372683,0.0212383978,-0.1587416977,0.3358259499,0.1481372416,0.7075200677,-0.2578309178,0.1896396279,0.0150311645,0.2776121795,-0.384010911,0.0889866129,-0.1946355999,-0.3346905708,0.2537218928,-0.1041299254,-0.2548139989,-0.0129498215,-0.6315752864,0.0518371612,0.0044754497,0.2070017755,0.1606281102,-0.5352048278,-0.4766362607,-0.1466354281,-0.1487207711,-0.1698218733,-0.0632988364,-0.0788986683,0.0721126348,-0.1897255927,-0.0883985981,-0.2601359189,0.1102479771,0.0771814957,0.1701355129,-0.3146591783,0.0592580363,-0.1035391241,-0.0372062549,0.1083503515,0.1465893388,0.2767812312,-0.1805095524,0.1025464013,0.0277055427,-0.1088031679,-0.2397813648,-0.1181817427,0.0440286621,0.2457215637,0.2073970139,-0.313262403,-0.096177876,0.2511529624,-0.1047439501,-0.1060047001,-0.1914811879,-0.278131485,0.0661624372,-0.1178401709,0.1875624657,0.0107328687,0.2696737647,0.1827258468,0.6394481063,0.1039488018,0.29801476,0.6273650527,0.57050699,-0.0594344176,-0.0164150055,0.0593583062,0.1587062925,0.0083983736,0.0667040125,0.0605471209,0.3135679066,-0.1482876986,-0.3254031837,0.4511334598,-0.1269151866,0.2275267541,0.1028302014,-0.1834723353,0.3342120945,-0.1366466284,0.1213156804,0.0884219259,0.0364703573,-0.0104183033,0.1635176539,0.2236900926,-0.2778138518,0.3969779313,0.2739136815,-0.0166277867,-0.0167518016,-0.121634163,0.0847386718,-0.3916296363,0.2628649473,-0.0066007129,0.6821075678,-0.0934061706,0.0792734697,0.044729609,-0.0611068904,-0.15126203,-0.0286299735,0.1222140864,-0.2704930604,-0.1011148393,-0.0246298052,-0.1035890654,-0.1537911445,-0.1808048636,0.3842013478,0.3214759231,-0.2883849442,-0.1265276521,-0.120101355,0.0310469884,-0.1715571731,0.0255027432,-0.382930994,-0.1697028577,-0.1091871262,0.0862319618,-0.1638846844,-0.0306878835,-0.2805267274,-0.0166414604,-0.4479308426,-0.0697767735,0.3496539593,0.0713341311,-0.2968046069,0.0560244881,-0.0998280644,-0.1421016157,0.3210002482,0.0437869616,-0.1180242375,-0.2759182155,-0.4726768434,-0.0391826406,-0.3159280121,0.1876202673,0.4698003829,-0.1936874241,0.4596888125,0.1652486324,0.1701928675,-0.1093983576,0.3558776975,0.0933343098,0.3376340568,0.0883948132,-0.1126320213,-0.1156195104,-0.0135779474,-0.2385885268,-0.0815633237,0.0567915067,-0.189568013,0.5808910728,0.7374231219,0.7756371498,0.178337872,0.0565868765,0.2356171459,-0.0151220048,0.1067235544,-0.6140139103,0.1484862417,-0.2301591337,-0.2497630864,0.0240746997,-0.043017596,-0.0399257466,0.1476687193,0.08560092,0.0164027363,-0.0707679614,0.3859116435,0.2548276186,0.271723479,0.0050371788,-0.0826184005,0.1035080031,0.0934236646,-0.1128809899,-0.107017301,-0.1548112929,0.1870467514,-0.1098317504,0.1068734229,0.2101290226,0.0572547391,-0.4577381313,0.1091430187,0.054270409,0.2002634853,0.3329734504,0.1272501796,0.8865144849,0.4321020544,0.0522489697,-0.1075202823,0.0868957639,0.1058642641,0.1058270335,-0.212264955,0.3860618472,-0.1290135831,0.0208047982,0.2148111463,0.4777830839,-0.1414243579,0.0681176856,-0.198374629,0.121390447,-0.1612010747,0.2055467367,0.2079623491,0.2873873115,-0.0521354638,0.0807134956,-0.0847521722,-0.1285315156,0.0266626086,0.2749696374,-0.4813677669,-0.1052286774,-0.0779462457,0.0974870324,-0.3360464871,0.1934803277,-0.608861208,-0.1146191955,-0.0163587965,-0.026393922,0.0452100523,-0.1461046487,-0.017829828,-0.3392192423,-0.2676384449,-0.2369656861,0.4567983449,0.4852994382,-0.1897087693,-0.1219143644,0.0583025031,-0.1139344573,0.1198730543,0.1304615736,0.0766293779,-0.0425431319,0.0220175907,-0.2366043478,0.0770407617,-0.2955486774,-0.1003503278,-0.0611502342,0.0482699908,-0.3870336711,-0.0554673187,0.1917675436,0.2615491152,-0.0251001157,0.0242408626,-0.3521004617,-0.0983275548,-0.4465918839,0.0942737013,0.309029907,-0.0684354082,-0.2979888022,-0.1364800483,-0.0203428268,-0.1440384984,0.0725734085,0.1827053726,0.1109868065,0.0628782436,-0.0982583314,0.3248747885,0.0486986749,-0.1236827821,0.1181345731,0.2344575673,0.2698459923,0.0933326483,0.0864959434,-0.0431058072,-0.15436095,0.1854565442,0.3218781352,0.0153115736,-0.131757006,-0.3005586565,-0.0854595974,-0.4511830211,-0.1068735644,0.2111316025,-0.1489962488,-0.6765805483,0.5133197308,0.1551615447,-0.0166077763,0.1863090545,-0.10034547,0.2901722789,-0.1238019913,0.4071457088,0.8698999286,-0.0672917515,0.0254613888,-0.1722702235,-0.2336156368,-0.4984537661,0.0652022734,-0.2628164887,-0.0970572755,0.4935136139,0.2357557267,-0.0586889349,0.1337695718,-0.207017988,-0.0297145508,0.5148485899,0.0178485811,0.1681435406,0.1661583334,0.1768041849,0.3574434817,0.069594115,-0.2493189126,0.0858768076,0.2473218441,0.4950659275,0.1484310329,-0.3257381916,-0.1173863187,-0.121986337,-0.2884106636,-0.1631172001,0.0093743075,-0.0569685288,-0.3132694066,0.1945052147,-0.1368418485,0.1540198922,-0.7071640491,0.0881195962,-0.0194029659,0.0836522281,-0.2503421009,0.1371775568,0.4700477421,-0.0295292921,-0.1994045526,-0.3230015337,0.1716652811,0.1711809635,0.224564597,-0.0988538936,0.3145775795,-0.1097397134,-0.3795018792,-0.0909525231,0.4449700713,0.1257590353,0.2684619129,0.3025090992,-0.1292747706,0.3866565526,0.0088189347,-0.1395197362,-0.1878393739,0.1800678819,0.0385491252,-0.2090242505,0.1225460246,0.2051604539,0.2142786235,0.0545157306,0.0446707457,-0.3483164012,0.3001547158,0.3341041803,-0.1491232067,0.2284933478,-0.1657892466,0.1003110707,0.2017822415,0.078084439,0.0800423399,0.0509261824,-0.0703282878,0.3531655073,-0.2789777517,0.4088290036,0.5149395466,0.3129907548,0.0254858788,-0.0116536804,-0.113834694,0.1585877985,-0.1413673013,-0.0180660803,-0.2098267078,-0.2153179795,-0.0428574048,0.0292316955,-0.0955213383,0.1794953048,-0.2284444273,0.0105049787,0.0080249328,-0.3193148673,-0.0018461946,-0.3061870337,-0.1885704398,0.0242178496,0.1980624348,0.5790305734,0.2496821731,-0.1685442328,0.1697681397,-0.2593251765,0.3227337599,0.1401127875,-0.2081028968,0.1342635304,0.0774539113,0.0544936322,0.1737288088,0.1709427088,-0.3139975369,0.4748363495,0.4173718691,-0.6377521753,-0.2072801441,0.028964268,0.2442902774,-0.180915609,-0.0765071064,0.3799448609,0.1829595119,0.2499195635,-0.1036512181,0.0794251114,0.2229080796,-0.4927923977,0.1913879365,-0.3053251207,0.2832438052,0.6400994658,0.0169496778,-0.6514305472,-0.4217723906,0.0981689617,-0.1173684299,-0.2218728513,0.128202036,-0.0157384779,-0.0883966163,-0.2711744905,0.1746778488,-0.2746869326,-0.3154573441,0.0952087417,-0.2452498674]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"Yes, there are slides like that at WMT every year :) BLEU correlates with human judgment only at coarse levels, and it seems to be getting worse when people try to use it to do model selection among high-performing neural systems.\r\n\r\nHowever, the point isn't whether BLEU is a good metric, but whether your BLEU score can be compared to other BLEU scores. They only can be compared if you use the same reference tokenization (similar to how you [can't compare LM perplexities across different segmentations](https:\/\/sjmielke.com\/comparing-perplexities.htm)). sacrebleu was an attempt to get everyone to use WMT's reference tokenization (meaning, your system has to first remove its own tokenization) so that you could just compare across papers. This also prevents scores from being gamed.","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":123,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nYes, there are slides like that at WMT every year :) BLEU correlates with human judgment only at coarse levels, and it seems to be getting worse when people try to use it to do model selection among high-performing neural systems.\r\n\r\nHowever, the point isn't whether BLEU is a good metric, but whether your BLEU score can be compared to other BLEU scores. They only can be compared if you use the same reference tokenization (similar to how you [can't compare LM perplexities across different segmentations](https:\/\/sjmielke.com\/comparing-perplexities.htm)). sacrebleu was an attempt to get everyone to use WMT's reference tokenization (meaning, your system has to first remove its own tokenization) so that you could just compare across papers. This also prevents scores from being gamed.","embeddings":[-0.3135464489,-0.2768908739,-0.0518379956,-0.2917878032,0.0934216827,-0.2474938184,0.28668347,0.2192002237,-0.4277971685,0.3128701746,-0.2075859904,0.1104491651,-0.1642134935,-0.1205747128,-0.1374371052,0.1132958829,0.1067669168,-0.0139779262,0.6649669409,-0.0207699798,-0.1770703495,-0.1872543395,0.0037461594,0.3304761648,-0.3724300563,0.2632550597,0.2985367477,-0.1913367212,-0.1498094052,-0.3490214944,-0.2689947486,0.2102922946,-0.2063372135,0.1252582371,-0.0001022047,-0.3871833086,0.0541020557,-0.0368929841,-0.4439689517,-0.3665273786,0.1544646025,-0.1669862866,-0.1151911169,-0.1171803251,-0.0134163201,-0.1435834169,0.040178448,0.1061420739,0.3251447082,-0.1058374122,0.2449501008,0.3034593165,-0.4683806598,0.1498442888,0.0699931011,0.2710111737,-0.0925994515,0.1770909429,-0.0050246725,0.3587901592,0.0353532992,-0.0747972876,0.0237663742,-0.0775674731,0.0670090392,-0.0880065784,-0.272153765,0.004597412,0.1218435839,0.5070339441,0.0015664555,0.1673952341,-0.3386724591,-0.0811631382,0.1015543714,-0.4909129441,0.0544757843,-0.0133016165,0.152076304,-0.0793089047,0.0021795162,0.281907171,0.2130946815,0.0366321206,0.0277076922,0.1824213862,-0.1342664212,0.0382129811,-0.1514034867,-0.0254612658,-0.447825551,0.113600567,0.0265502501,-0.0206227507,-0.3019562066,-0.1138956845,-0.1809839755,-0.4280613661,0.0257194713,0.0539783239,0.3791999221,0.2382393628,-0.0665691346,0.08494322,-0.1064646766,0.1340885609,-0.2781978846,-0.0645568818,0.3806477487,0.0334777944,0.3261171877,0.1682181954,0.2315180302,-0.11943423,-0.0887901485,0.170956403,-0.3551331162,-0.1987959892,-0.8085628748,-0.0013848902,-0.1617463082,-0.1357214153,0.0229595546,0.2048790455,-0.1913792789,-0.2177047133,-0.1282643378,-0.0339852013,-0.0903071016,-0.1296689361,-0.08550217,-0.3203369677,-0.2990040779,0.0311049167,-0.1111095175,0.2265412658,0.2177430242,-0.066437155,0.3926664591,-0.0662084669,0.0135622285,-0.1394260973,0.0663151592,-0.104272157,-0.4170202017,-0.1082601026,-0.1174196303,-0.1446509361,-0.0133492369,-0.0166644994,-0.21624358,-0.2710017562,0.1235542968,0.1748419851,0.0567967854,0.2153903395,-0.3138505816,0.2010301501,0.0067009279,0.0871113166,0.3875291049,-0.1698922664,-0.1422985494,-0.2447662652,0.1321950108,-0.2628275156,0.0004181841,-0.3629073203,0.4552856386,0.2354488224,0.2896308601,0.4723249972,-0.3027026653,0.2717089951,0.1163449958,0.2523094714,0.1024568826,0.0904901996,-0.1697993428,-0.0509173051,-0.0876499563,-0.2525357008,-0.0231531877,0.060714297,0.4108531773,-0.1739975959,0.0423741415,-0.0023212475,0.1505367011,-0.1329603195,-0.3309361041,-0.5472711921,0.2306468189,-0.0918670446,0.0819598064,-0.1068120971,0.2979665399,0.6883991957,0.3179251552,-0.2438719273,-0.0879760012,-0.3974711597,0.5804697871,-0.0651317611,0.0949912295,0.107622236,0.4525400996,0.1938538998,-0.0534838475,0.2195281535,0.2961948216,-0.3364754021,0.0217113681,0.0140826507,0.2248198092,-0.0177137926,0.1083814278,0.0481035113,-0.3625598252,0.1451941431,-0.4818155468,-0.2320880741,-0.4947704375,0.1124652028,0.2530227005,-0.1261920184,0.0631493703,0.0302841384,0.0864325762,0.3047133386,-0.0851937979,0.0517084897,-0.2287273556,0.1790742576,0.0435498394,0.0120681729,-0.428494662,0.5606502295,0.2933938503,0.1943801045,-0.2939884663,0.2260260731,-0.0991017297,0.1268149614,0.3718375564,0.452625066,0.1289929003,-0.1137036383,0.033375226,-0.255538106,-0.139011085,0.0403778031,-0.0013857797,-0.1774316728,0.1480552405,0.1533436924,-0.1985276043,-0.0082235774,-0.0935473815,-0.1130372509,0.1701189727,-0.0313551277,0.2223936617,0.0351333134,0.4144179821,-0.1636601239,0.1096955761,0.076553382,0.0223658849,0.0529046655,0.1112138629,0.0436987206,-0.0731369182,-0.2892229855,-0.1087961197,0.0376401506,-0.1532531232,0.1441552788,-0.0483636372,0.3560223579,-0.0613275059,-0.1281310618,-0.0646713674,-0.1471245587,-0.0709936842,0.1334352046,-0.1606141031,0.3112232089,-0.0550840683,0.1209589019,0.0132962177,-0.1642047912,0.173934564,0.0952125937,-0.0954497904,0.3084385097,-0.0024335445,0.7073996067,-0.2083141059,0.0458778627,0.0995651111,0.2048485279,-0.2296585888,0.1501955688,-0.142242372,-0.1898702532,0.2541303337,-0.0357539095,-0.2220229357,-0.0826374888,-0.5311999321,0.1318363249,-0.0413252227,0.1282416582,0.205958873,-0.502579093,-0.4749952257,-0.1504636407,-0.0743283704,-0.2248393297,-0.1113214493,-0.0876680017,-0.0031401471,-0.2432128638,-0.0484583378,-0.2018825412,0.0908645764,-0.0008779332,0.1538136005,-0.2386196554,0.0360030346,-0.1215343475,-0.198466897,0.1900894791,0.2075915039,0.1916429549,-0.2571556568,0.1882054508,0.0815550983,-0.1382200718,-0.2136996388,-0.0471130647,0.0370419435,0.3381594419,0.1280739754,-0.3612490296,-0.0142398439,0.3136072159,-0.1650910676,0.0089935008,-0.2673547566,-0.2607000768,0.0362970568,-0.2162789851,0.1994867474,0.003855034,0.1457368135,0.1224850863,0.6416712999,-0.0021231908,-0.026488401,0.5659835339,0.7179704905,0.0041275383,-0.1526336372,0.0798186883,0.2234243453,-0.0822401345,0.0420573466,0.0421513841,0.4057151377,-0.1733709723,-0.3698149323,0.4804112315,-0.1209396347,0.0635356009,0.0604742505,-0.1702916026,0.2942318022,-0.0911962911,0.0525649525,0.2059789747,0.0299623664,-0.0783183798,0.0455713645,0.2988677025,-0.3732349575,0.382648617,0.245017454,-0.1451509595,-0.0480050258,-0.1520748883,0.1740009487,-0.3281794488,0.2320007384,-0.0149284899,0.5163071752,-0.0459952205,0.0073827044,-0.0164377131,-0.0909691304,-0.2292496562,0.0067145866,0.1202262118,-0.1380390376,-0.0550658964,0.0235484391,-0.067479603,-0.1871848255,-0.1754924655,0.3741059899,0.2670319974,-0.2924336791,-0.0644136295,-0.0704974979,0.0761059076,-0.1893185079,0.0398724452,-0.4901160896,-0.0322923884,-0.0888096914,0.0647249967,-0.2311668247,-0.0291660987,-0.1579938233,-0.0025014449,-0.4603503942,-0.0593859777,0.2877703905,0.134414643,-0.3237524033,0.1645163298,0.0165825896,-0.0550508462,0.3421105444,0.224314332,-0.0592534654,-0.2839721441,-0.3086964488,-0.0284662396,-0.0766311213,0.0317634456,0.4103709161,-0.1020365432,0.4372866154,0.1617634445,0.1972983032,-0.2993662953,0.3632666767,0.052226916,0.2916179597,-0.0139050269,-0.0761201084,-0.0174428206,0.0481399298,-0.2702102363,-0.0599429794,0.3076438308,-0.1373657137,0.5109136701,0.7239353061,0.7825475335,0.1221410483,0.1194015741,0.1865378916,0.0223677699,0.1898472458,-0.390563041,0.1180919483,-0.2003886402,-0.1962484866,0.0201406125,-0.0396292359,-0.043919351,0.0889305621,0.0276265591,0.0048433491,-0.0461555086,0.3588597775,0.1058226824,0.2374973893,0.1095868349,-0.0569515601,0.1410305053,0.1804325581,-0.2120290548,-0.0806303844,-0.1090534478,0.2739864588,-0.1722732633,0.1797644794,0.1267143935,-0.0359193273,-0.4334559143,0.0524674617,0.0259466488,0.1520060748,0.3283515871,0.0159815941,0.7727355957,0.3558005393,-0.0424426384,0.0008841029,-0.0866600871,0.1440268755,0.0325928144,-0.3202248812,0.4579420984,-0.0804563388,-0.002943076,0.2582782209,0.5422329903,-0.09910243,0.2184824497,-0.1154132113,0.1114745066,-0.2073843777,0.2063444108,0.1947823465,0.2042612433,-0.0815079734,0.1459706277,-0.1154578999,-0.1760990471,0.0336234197,0.3827241957,-0.411639452,-0.1374210566,-0.0139681585,0.0277262721,-0.4150917232,0.093141824,-0.5964990258,-0.1652957499,-0.05322925,-0.0029554607,0.0427981243,-0.1159626171,-0.0431225561,-0.3712256551,-0.1380927116,-0.176166594,0.5500761867,0.5743538737,-0.0680378452,-0.0840881318,0.1526635289,-0.1294334531,0.1499759853,0.0810379609,0.2317016721,-0.1623604745,-0.0262844916,-0.2641332746,0.2796523571,-0.3633126915,-0.0482716225,-0.1661304533,0.0123861181,-0.3028207719,-0.0695281029,0.129830569,0.1473763883,0.0080074454,0.0723862872,-0.3600257039,-0.1478528976,-0.4487479925,0.0770938993,0.3181324303,-0.1365174949,-0.368360728,-0.0406293496,0.0914732963,-0.1284944564,0.123615101,0.1681539565,0.1460053325,0.0943040773,-0.2224002779,0.1836116016,0.0644032359,-0.0988318101,0.1749599576,0.2933544815,0.3220136464,0.161504969,0.1782365143,-0.0752100796,-0.1370176077,0.1689587831,0.3891145587,0.0350415595,-0.095660612,-0.2563347518,-0.1240466908,-0.4172921777,0.0690471679,0.222341314,-0.1401326507,-0.6945355535,0.4876029491,0.2131176293,-0.0483808517,0.1338071823,-0.1213914827,0.2667773962,-0.0532372594,0.3924629986,0.7362117171,-0.0372774899,-0.1512629688,-0.1831576973,-0.2062224299,-0.4393009841,0.074401699,-0.3138430417,-0.0038275898,0.4468527436,0.3168330193,0.0150227221,0.2523865402,-0.0869463533,-0.0798948333,0.7330470085,0.0549992062,0.2388372421,0.107382454,0.3212557435,0.4389120042,0.1100015193,-0.3505596519,-0.0311522949,0.131404981,0.2895136774,0.122955218,-0.3349921405,-0.1196227148,-0.074651204,-0.2652316093,-0.165761888,-0.0878552049,-0.1341841668,-0.2626985013,0.1637266129,-0.1358130723,0.1129834577,-0.494094938,0.021720998,0.0600214601,0.1324975491,-0.2049178034,0.1868079752,0.4067946374,-0.0961914361,-0.3172369301,-0.1704368889,0.215271309,0.1738672107,0.2623382509,-0.0425804034,0.1893675327,-0.1666003317,-0.3925199807,-0.1494614482,0.3882748485,0.0997849479,0.2310325503,0.1746230721,-0.301823169,0.4194815159,0.065508388,-0.1832059473,-0.1239532679,0.1407373399,-0.0470190942,-0.3344610929,0.1462144554,0.0389607064,0.2136737555,0.0602511056,0.1322596371,-0.2523955107,0.3149112165,0.3648401201,-0.281165719,0.279499501,-0.0469734482,0.1297311485,0.1206570864,0.2035244256,-0.0275657009,0.0962560773,-0.1082019508,0.3452294767,-0.2989458442,0.3674924672,0.4220760167,0.2841754258,-0.164042905,0.0246410966,-0.1880404055,0.1937737614,-0.1932515055,-0.0404883884,-0.1663774401,-0.2015745789,-0.0442335159,-0.0023901544,-0.1404891163,0.104660362,-0.2722037435,-0.0407936238,-0.0068744742,-0.43567577,-0.0047315164,-0.2395194471,-0.263361454,0.1253314912,0.2563939691,0.5456875563,0.2020896971,-0.0291165505,0.1459059864,-0.2464653999,0.3427484035,0.1208008602,-0.080365546,0.0564773679,0.0297700129,0.1523703337,0.2380371243,0.2812529206,-0.246946916,0.4508235157,0.3743038774,-0.6172069311,-0.2060279697,-0.0594086461,0.2945282757,-0.1835798174,-0.0771442726,0.3536556661,0.2103448659,0.3012430072,-0.0424631089,0.0642004088,0.2312901318,-0.4930180013,0.2911452353,-0.257902503,0.2946453094,0.5700720549,-0.0015902165,-0.668594718,-0.3495563567,0.1682346463,-0.053129945,-0.3244286776,0.1273411661,0.0007559858,-0.1607611626,-0.2262654603,0.1748576462,-0.351477176,-0.3605736792,0.0551531948,-0.1777174771]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"I do not consider as a sufficient solution switching this library's default metric from BLEU to the wrapper around SacreBLEU. \r\n\r\nAs currently implemented, the wrapper allows end users to toggle SacreBLEU options, but doesn't pass along the SacreBLEU signature. As @mjpost showed in [Post18](https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf), it's simply not credible to assume that people will stick to the defaults, therefore, the signature is necessary to be explicit about what options were used. \r\n\r\nIn addition to the `v13a` or `intl` options for the SacreBLEU `tokenize` argument, which was pointed out earlier, papers frequently differ on whether they lowercase text before scoring (`lowercase`) and the smoothing method used (`smooth_method`). BLEU scores can differ substantially (over 1 BLEU) just by changing these options. \r\n\r\nLosing the SacreBLEU signature is a regression in reproducibility and clarity.\r\n\r\n(Perhaps this should belong in a separate issue?)","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":137,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nI do not consider as a sufficient solution switching this library's default metric from BLEU to the wrapper around SacreBLEU. \r\n\r\nAs currently implemented, the wrapper allows end users to toggle SacreBLEU options, but doesn't pass along the SacreBLEU signature. As @mjpost showed in [Post18](https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf), it's simply not credible to assume that people will stick to the defaults, therefore, the signature is necessary to be explicit about what options were used. \r\n\r\nIn addition to the `v13a` or `intl` options for the SacreBLEU `tokenize` argument, which was pointed out earlier, papers frequently differ on whether they lowercase text before scoring (`lowercase`) and the smoothing method used (`smooth_method`). BLEU scores can differ substantially (over 1 BLEU) just by changing these options. \r\n\r\nLosing the SacreBLEU signature is a regression in reproducibility and clarity.\r\n\r\n(Perhaps this should belong in a separate issue?)","embeddings":[-0.27546224,-0.2523212731,0.015029124,-0.3404414356,0.2117930949,-0.3596958518,0.1853222847,0.3614321649,-0.3344036043,0.3967298269,0.0437221453,0.3229061961,-0.1908613592,0.0206783451,-0.186046496,0.0257372204,0.0594054833,-0.0886750668,0.6912773848,-0.0719551891,-0.1700499654,-0.1265845448,0.008774207,0.1469752491,-0.1182374805,0.3319087327,0.3068475127,-0.2342270315,-0.3007888198,-0.3260027468,-0.0953568295,0.2849736512,-0.2850325406,0.1056501493,-0.0001081686,-0.3039608002,0.0533862971,-0.1401547939,-0.1782739311,-0.4625362158,0.1213937998,-0.2437465489,-0.1138690561,-0.1544534713,-0.0162539575,-0.1707026213,0.0801403299,0.0198843144,0.346075356,0.0222148467,0.1744837016,0.2951719165,-0.4948261678,0.1890507042,-0.0312422365,0.2374393344,-0.0844698101,0.2880865335,-0.0362244211,0.313665539,-0.1438014805,-0.0414782017,-0.0263566021,-0.0947214589,0.1497689188,-0.0572514161,-0.2816728055,0.0068677953,0.0345946103,0.493599534,-0.0878522396,-0.0055465871,-0.2163961083,-0.1053453833,0.0727963522,-0.4867216349,0.1393391937,-0.0989826173,0.13512218,-0.2426536679,0.0057271724,0.1746472567,0.087215662,0.1095076799,-0.0680324584,0.0190903898,-0.1622909009,0.0086817127,-0.0108169317,0.0278702639,-0.48339656,0.2477413714,0.1340903193,-0.0004387455,-0.1850897521,-0.0631029382,-0.0277451854,-0.5770332217,-0.0194555204,0.1404000521,0.3119617403,0.3449689448,-0.2340863049,0.1241531,-0.3108345866,0.1665844172,-0.1009587497,-0.2342416197,0.4526522756,0.1669716686,0.4540477097,0.0995844603,0.2723415792,-0.2017757446,-0.0753199384,0.2687521577,-0.2019764185,-0.1338178813,-0.6423498988,0.0248230789,0.0201711301,-0.1173330247,0.1609819233,0.1125367731,-0.3188554943,0.0448210575,-0.0249752216,0.0525739416,-0.1805677861,-0.0498966724,-0.0631348863,-0.1982997954,-0.2298302203,-0.0254416279,-0.0078387856,0.0413170122,0.2758931816,0.1014923453,0.5700212717,-0.0659460798,0.0850098804,-0.0034184742,0.031156214,-0.1083388701,-0.4364647567,-0.1272356659,-0.0389608517,-0.1705790758,-0.0058359266,-0.0333688557,-0.3677676618,-0.3043262661,0.1216996014,0.1472866088,-0.0167448223,0.1506099403,-0.4524779618,0.1997803152,-0.1081032753,0.0593562685,0.2417585701,-0.1245697886,-0.1582509875,-0.2411295325,0.2471058667,-0.3424717486,0.1747466922,-0.3750631511,0.4064902067,0.1518936008,0.2413390875,0.2540333569,-0.2383617759,0.1914658993,0.1538760364,0.2080127448,0.116146557,-0.1404009461,-0.2718495131,-0.194881767,-0.0188464299,-0.3495473862,0.0538361631,0.0247635916,0.4324914217,-0.1881874502,-0.0882972255,0.0986308679,0.0389308892,-0.1022564918,-0.36290434,-0.6038562655,0.0815442055,-0.0762142763,0.2625020742,0.0183706656,0.1317540705,0.6068987846,0.1456057131,-0.1953678131,-0.0812748298,-0.2756708264,0.6100926995,-0.188805595,0.2301256061,0.1286764443,0.3836823106,0.124755688,-0.1045504957,0.1484696269,0.0867827535,-0.3845891953,-0.1050526574,-0.0811828747,0.3070400655,0.1821350157,0.0624115504,-0.0500540435,-0.4763344228,0.1471545249,-0.5149652362,-0.2080215812,-0.4747824073,0.2136285603,0.1432303637,-0.1589784026,0.1644805968,0.1681167036,0.1220598295,0.3748025596,0.025054723,-0.0488163941,-0.2073414028,0.2976349294,0.0177839305,0.1419627964,-0.4085462689,0.5593419671,0.3624275625,0.1577276438,-0.3639058173,0.1839713305,-0.2214429826,0.1356986761,0.1031718031,0.5868310928,0.1048102304,-0.0521509312,0.0152480276,-0.164219752,-0.2311976403,-0.0265012309,-0.1726716906,-0.2763185501,0.0324626751,-0.0138006862,-0.1483166069,-0.0102157332,-0.1258916259,-0.0296058804,0.2117739022,0.1503810585,0.1223942265,0.2029440552,0.56578058,-0.1657807082,0.0235720724,0.1294998527,0.0724493042,-0.0312146079,0.0524427965,0.1006217599,-0.083846271,-0.2964494228,-0.1408361644,-0.0505328849,-0.2003326863,0.1889407933,-0.0410737284,0.2605827749,-0.1415332556,0.017123526,0.0461693928,-0.0348761193,-0.1211702675,0.1572966874,-0.2206382006,0.2652708888,-0.0518266819,-0.0629634634,-0.0514928289,-0.2756368816,0.2165699899,0.1610555351,-0.2347649932,0.3221915364,0.0509920418,0.7734154463,-0.2841901481,-0.0480795801,0.0800386891,0.1919495016,-0.1521677524,0.0710826218,-0.1971363723,-0.288234055,0.2453038841,-0.1230286285,-0.169133395,-0.0259956568,-0.6473637819,0.0570211597,0.0846215561,0.1302390099,0.2541700602,-0.4206206203,-0.3183861673,-0.1246137917,-0.0045476472,-0.1607361734,-0.0812279433,-0.0185571797,-0.0673048794,-0.099458158,-0.1805820614,-0.3190949261,0.2195562869,-0.0247370955,0.074115105,-0.2699486017,0.1559746116,-0.2022841722,-0.1049693674,0.1877568364,0.3002032638,0.2783682048,-0.1742997766,0.0282519534,0.0802432597,-0.1298372895,-0.2182052284,-0.0646490678,0.02784664,0.5163828135,0.1747506261,-0.4123682082,-0.3798092604,0.2071429789,-0.042192813,-0.0351000726,-0.2209966481,-0.1542415321,0.0020488976,-0.1901153475,0.0419289246,0.0104154442,0.2413175404,-0.0365926735,0.5588755012,0.0531289354,0.0504403189,0.6644158959,0.759304285,-0.0096892109,-0.1742100716,0.0397704542,0.2477197796,0.013521269,0.108068645,0.0331627019,0.3516508639,-0.0321335457,-0.137116611,0.2237671763,-0.0456753969,0.1619409174,0.0685172454,-0.0137168625,0.2787080407,-0.2318697274,0.1013233736,0.0676323846,0.0986978933,-0.071184732,0.1271313578,0.0789783001,-0.2700660229,0.4005420208,0.3368866146,-0.0587390512,0.0102421828,-0.1101641208,0.1993377954,-0.3824777603,0.2657038271,-0.0751997083,0.5752350688,-0.0746018365,-0.024021538,-0.0173620097,-0.1876342744,0.0007758942,0.0691756606,-0.0253186785,-0.0942797959,0.0667790174,-0.1650070846,-0.0758966431,-0.0452968106,0.0164157581,0.3930713236,0.3482965827,-0.4058703184,-0.1958315074,-0.0629343688,0.057091441,-0.0748103335,0.0695133805,-0.4463495612,-0.2412393987,-0.0490483344,0.1686138213,-0.1845019609,0.0791845992,-0.1489431113,0.1644654274,-0.4173012972,-0.2147604823,0.2511439621,-0.0419026911,-0.1035649627,0.0442009158,-0.0292806383,-0.1310567558,0.260324955,0.1678565592,-0.1522385776,-0.2501449287,-0.3347626925,0.0205855537,-0.1456274539,0.0475460552,0.4614115655,-0.2092407644,0.5445093513,0.1642320603,0.316198349,-0.3125008047,0.5190663934,0.0436014347,0.1755461097,0.034788996,-0.1021186933,0.1291785985,0.1523217559,-0.2497835159,0.0631949231,0.1665419787,-0.0888466537,0.4626339078,0.740138948,0.8428682685,0.2177067399,-0.0320535265,0.2129051536,-0.0446677394,0.1443548203,-0.527851522,0.1338583678,-0.1948666722,-0.323053658,-0.0298948716,-0.0854226202,-0.1247150525,-0.0828242972,0.0162194818,-0.0688038468,-0.1390706003,0.2516281307,0.1149614826,0.2827929258,-0.1189624071,-0.1095579416,0.2673264146,0.1080279276,-0.2531181276,-0.1006174088,-0.0463677421,0.1394370794,0.0164078139,0.0881399363,0.1082650498,-0.0877597928,-0.3123394847,-0.1063389033,0.1527085006,0.2015444785,0.3342866302,0.0425663553,0.9651348591,0.385335505,-0.1243092492,0.045668412,-0.07019867,0.2441375405,0.1586636752,-0.3231444359,0.4519591033,-0.059423618,0.0911648571,0.1279832423,0.5075562,-0.0959443823,0.2298789173,-0.2519479096,0.1899130493,-0.2021712959,0.221261397,0.184471339,0.2525806427,-0.0824064314,0.1104957461,-0.0369913392,-0.0545950457,0.1019165739,0.1210890785,-0.4966746271,-0.039740555,0.1212442294,0.1668065041,-0.5853478312,0.0432191491,-0.5460817814,-0.0875242427,-0.014060461,-0.1513988823,0.0507144891,-0.2165408134,0.0371552221,-0.3308941722,-0.148248896,-0.1380577385,0.414191097,0.3518089354,-0.1259289682,-0.0658845678,0.0494555682,-0.1918944716,0.2395288348,-0.0224025175,0.2596870065,-0.104719013,-0.061847426,-0.2455209941,0.122344397,-0.3115983307,0.1537911147,-0.0017352327,0.0077153486,-0.2866838276,0.0747498125,0.1957452744,0.0206990074,-0.0793006271,0.0688399225,-0.3790499866,-0.0977245793,-0.5446706414,0.1280363202,0.2180264741,-0.1567624956,-0.2039008141,-0.0855124071,-0.0936808437,-0.1068139002,0.1108098328,0.2017198354,0.0471960939,0.0561057255,-0.3299783766,0.3696166277,0.2226047069,-0.2691073418,0.3405194581,0.2849809825,0.2048351169,0.1909551322,-0.0529008582,-0.0580233969,-0.0697703063,0.1431015581,0.3519707322,-0.0585142486,-0.1372234076,-0.1452504247,-0.1037769392,-0.3953814805,0.0443785116,0.1021114811,-0.1861859411,-0.5499740243,0.3633194864,0.1897132546,0.0583079681,0.1050949246,-0.2831101716,0.3252624869,0.0049520084,0.3199234903,0.8110058308,0.0550708883,0.0376190655,-0.1587773412,-0.1927165836,-0.4535197914,0.2451011688,-0.3284197152,-0.1791834086,0.3774274886,0.3361741602,-0.0284708869,0.1215713918,-0.1943570673,0.1049060524,0.7144160867,0.1136490926,0.1442257315,-0.0147125404,0.1472213268,0.3415810466,0.1295295954,-0.217777133,0.1098513231,0.2844247818,0.5538887382,0.0397790186,-0.4556814134,-0.0375084281,-0.0527021438,-0.2113131732,-0.239655599,-0.1923447996,-0.1665890366,-0.3287785649,0.2214094847,-0.1422390789,0.1728059351,-0.512806654,0.0530703478,0.049440559,0.0916843936,-0.288529247,0.1962189376,0.5878984332,-0.0366564356,-0.1349345297,-0.2193233222,0.0615804084,0.1573639214,0.3112172782,-0.0482836142,0.2059929669,-0.0467672534,-0.4113273025,0.011513887,0.374545455,0.1042153314,0.2700223625,0.3401529193,-0.276666671,0.2666606307,0.0687853321,-0.1562050134,-0.0245249346,0.0578132905,0.058238823,-0.2215091884,0.3447814882,0.3188422918,0.290083915,0.0622412972,-0.0034845138,-0.3254074454,0.147002995,0.3225288391,-0.105074726,0.3041751683,-0.2578832209,0.1071747839,0.3149333596,0.0865031406,-0.0894462988,-0.0131632015,-0.1644266546,0.1942091137,-0.2547185421,0.3230781853,0.5965038538,0.4035923779,0.0537025705,-0.042528253,-0.0979509205,0.1583048701,-0.0984668881,0.0471700057,-0.0377769135,-0.3022095859,-0.1488029063,0.0870131776,-0.1137644425,0.0482138097,-0.2075438499,-0.042838756,0.028820999,-0.4005045295,0.033614248,-0.2890281975,-0.1741446853,0.0606934689,0.2600748837,0.4975219071,0.2622143924,-0.1672336161,0.1346783191,-0.2747911215,0.2201880366,0.2679390013,-0.3194968402,0.003196744,-0.0054445323,0.2004410028,0.2150100172,0.1578175128,-0.3420923054,0.3807724118,0.3974065781,-0.6234899759,-0.2568434477,0.0357195251,0.1461793184,-0.1603844315,-0.0514152013,0.4731211066,0.1850257367,0.4181157649,-0.139153257,-0.0352263413,0.2448301613,-0.3790570796,0.2942575216,-0.1554787159,0.3669971526,0.5302295685,0.0581172965,-0.7508431673,-0.3180319667,0.0478923805,-0.0993853509,-0.3459317982,0.1738738865,-0.095916152,-0.043594487,-0.3346836865,0.3441054523,-0.0979178548,-0.3366897106,0.0911635607,-0.2113001049]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"Thanks for sharing your thoughts. This is a very important discussion.\r\n\r\nAlso one of the first items on our mid-term roadmap (we will try to clean it and share it soon) is to introduce mechanisms to get high-quality traceability and reproducibility for all the processes related to the library.\r\n\r\nSo having the signature for `sacrebleu` is really important!\r\n\r\nRegarding BLEU, I guess we can just remove it from the canonical metrics included in the repo itself (it won't prevent people to add it as \"user-metrics\" but at least we won't be promoting it).\r\n\r\nOn a more general note (definitely too large for the scope of this issue) we are wondering, with @srush in particular, how we could handle the selection of metrics\/datasets with the most community-based and bottom-up approach possible. If you have opinions on this, please share!","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":138,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nThanks for sharing your thoughts. This is a very important discussion.\r\n\r\nAlso one of the first items on our mid-term roadmap (we will try to clean it and share it soon) is to introduce mechanisms to get high-quality traceability and reproducibility for all the processes related to the library.\r\n\r\nSo having the signature for `sacrebleu` is really important!\r\n\r\nRegarding BLEU, I guess we can just remove it from the canonical metrics included in the repo itself (it won't prevent people to add it as \"user-metrics\" but at least we won't be promoting it).\r\n\r\nOn a more general note (definitely too large for the scope of this issue) we are wondering, with @srush in particular, how we could handle the selection of metrics\/datasets with the most community-based and bottom-up approach possible. If you have opinions on this, please share!","embeddings":[-0.0818871707,-0.1245273948,-0.0124915289,-0.2027842849,0.1472497433,-0.3798782527,0.2327238023,0.216726169,-0.3655436039,0.433247298,-0.0330452919,0.3025978804,-0.1015592813,-0.023105178,-0.062252719,0.0280110352,0.0301390756,-0.0420776941,0.5825217366,-0.0730280578,-0.2260937393,-0.1260984689,0.1246969253,0.3257729113,-0.1314591616,0.1447937042,0.3365076482,-0.2166612744,-0.3565535247,-0.2769578993,-0.1404503733,0.3231076896,-0.2540121675,0.2577394545,-0.0001069882,-0.4084021449,0.0452209786,-0.0130942352,-0.191257149,-0.4243074059,0.159482792,-0.0473010316,-0.0152656948,-0.1188468859,0.2306433469,-0.0742480829,-0.0054342658,0.0063780774,0.3002331853,-0.0364047252,0.1854625195,0.4392635226,-0.5508165359,0.1971772164,0.034031257,0.3261634111,-0.1520132571,0.26788342,0.2389596552,0.3190007508,-0.0405545309,0.0115692103,-0.0356705301,-0.1282536834,0.1492723525,-0.08414983,-0.1127899885,-0.0924187675,0.0481288433,0.5816909671,-0.0507145151,0.0061406051,-0.2000901103,-0.1411981583,0.1366505027,-0.6313545704,0.167587176,-0.0666436255,0.078595683,-0.1725691706,-0.098167941,0.1102956459,0.0271075442,0.1739035696,0.1366734952,0.14517878,-0.1113369986,0.0330863521,-0.0647934452,0.0700325444,-0.3146150112,0.1770358831,0.0844018683,-0.0167076942,-0.1335647106,-0.1489966363,-0.0528645217,-0.497538656,0.1695331484,0.1334266663,0.2738729715,0.211594373,-0.1975943297,0.1667269617,-0.0983183905,0.1832697988,-0.1140599623,-0.2469255626,0.5300056934,0.1867011935,0.5439881086,0.2406797111,0.1755176336,-0.1962686479,-0.1066871956,0.2655930519,-0.2863332927,-0.257281512,-0.6572712064,0.0352051295,-0.0842982605,-0.1380368769,0.1935329288,0.0698646754,-0.2418748289,0.039669916,-0.1248391047,0.1039018631,-0.1976496279,0.0453195758,-0.0443241633,-0.2019354701,-0.2791752815,-0.0301847011,0.0450530238,-0.0970879272,0.3239656389,0.0443930887,0.522208035,-0.1612733454,-0.0141982082,-0.066585727,-0.1375818253,0.0075469054,-0.4400563836,-0.0836923495,-0.1814333349,-0.1485814452,-0.154213056,0.0515527762,-0.3326386511,-0.3480359018,0.0988398045,0.0916899592,-0.0631538555,0.1043326855,-0.5079072714,0.1652381718,-0.0534167103,0.3217261434,0.3081628084,-0.0437036827,-0.1204959154,-0.2699185908,0.1820601821,-0.2673655748,0.1021958962,-0.3500091434,0.3089857399,0.1232982427,0.2470039427,0.4162701368,-0.2819542289,0.270813942,0.1471827775,0.1384106874,0.1858864427,-0.1708893031,-0.1119875386,-0.1249221414,-0.100658372,-0.2894634008,0.0726195723,-0.0181345604,0.3934051096,-0.322113663,-0.0021629999,0.0995085537,-0.0746621415,-0.093705982,-0.2949598134,-0.6365936995,0.1811713576,-0.0640156567,0.2618018687,-0.0739233643,0.1493518651,0.5474885106,0.16225411,-0.2383558452,-0.0671942011,-0.2915121615,0.6991117001,-0.0566018224,0.1714491099,0.1690240353,0.3516166508,0.0984459445,-0.1594759822,0.0942376703,0.1024674624,-0.3705070019,-0.1538808346,0.0854519531,0.3287083507,0.1383348405,0.0405930281,-0.152090326,-0.4923216403,0.1525607258,-0.5571346879,-0.20369789,-0.5812091231,0.2410837561,0.2647257447,-0.1177807823,0.0340831541,0.1121772379,0.1171182171,0.3213680089,0.0805639327,-0.0825177282,-0.1740756184,0.2082176507,0.0185832959,0.1545181572,-0.2574837506,0.6626389623,0.5278886557,0.292298317,-0.29538095,0.0023913991,-0.201923579,0.1292420179,0.2877389789,0.532376647,-0.0018848474,-0.035863556,-0.0381867066,-0.1980602145,-0.2115884125,0.0169116706,-0.0812572464,-0.1598342508,0.1188878566,0.0781567767,-0.2059928775,0.0583183654,-0.1707789749,-0.0270369127,0.3027314842,0.1237444431,0.0557776839,0.1786347181,0.5540725589,-0.1804165989,0.0339521207,0.0605224185,0.068394348,-0.0285854191,0.0326615758,0.0732317492,-0.0773920566,-0.3613891304,-0.1430312544,0.1695746034,-0.1167992651,0.1390595585,-0.0318722986,0.3440871239,-0.1423461884,0.0464832485,-0.027767878,-0.0538764186,-0.0934589282,0.1594306082,-0.1906055808,0.2088365704,-0.0212829746,0.0130718919,-0.1947479248,-0.1988852471,0.3032418489,0.1695010662,-0.2259410024,0.3890611529,0.2144977599,0.759288609,-0.3860781491,0.1463853121,-0.0089942282,0.1416170001,-0.2407531589,0.0906016231,-0.0337107368,-0.2865643799,0.4852043688,-0.0415745005,-0.1511919498,-0.1072937697,-0.7395250201,0.1265028715,-0.1127201095,0.2133208215,0.0958522856,-0.4332280755,-0.4236867428,-0.1633209884,-0.1294266284,-0.211745739,-0.006442185,-0.1051587239,-0.0112356981,-0.0615267754,-0.1734590679,-0.3504240215,0.0531871468,-0.1200858653,0.1620345116,-0.233478412,0.1778959334,-0.164712131,-0.1721254885,0.1713183969,0.2623412013,0.2695657611,-0.2081600428,0.0154860895,0.0330490433,-0.1602005363,-0.2397180945,0.0083256131,0.1209514812,0.3449020088,0.1680004001,-0.3970922232,-0.5138249993,0.4018003643,-0.1077529415,0.0158451013,-0.2317853421,-0.1203455031,0.0657151267,-0.1335249841,0.0531875901,-0.1063811705,0.2502326071,0.171159789,0.5836225748,0.122861445,0.1070338339,0.5068244338,0.8114039898,0.0053955093,-0.0740964636,0.0504551902,0.4288633466,0.0942585245,0.0926601067,0.1649817228,0.3664279282,-0.0541517064,-0.1900217086,0.3386750817,0.0103479335,0.1769568771,0.1430216134,0.008204259,0.1964699775,-0.248614043,0.1646997631,0.0811779052,0.0679921061,-0.0400025509,0.0257301889,0.093983911,-0.248870939,0.3483304381,0.3045796752,0.0546898767,0.0405344628,-0.1039697602,0.1161505282,-0.4229324162,0.2114220858,-0.0604516268,0.5206593275,-0.184816584,0.11328464,0.1800219119,-0.0277350601,-0.1479236484,-0.0034005365,0.0157108624,-0.1394667178,-0.1521192938,-0.0733044073,0.0512397327,-0.0800605118,-0.0391259789,0.269965589,0.3430534601,-0.4303210676,-0.3011679947,-0.0452266559,0.0184950512,-0.1099382043,-0.0034069763,-0.2263357937,-0.3198582828,-0.1545668542,0.0949595869,-0.1078187749,-0.0230285991,-0.1868997067,0.0459805578,-0.3090237081,-0.0789217427,0.3344667852,-0.0043194094,-0.2447860092,0.1004079953,-0.0108831096,-0.0265458263,0.2270487696,0.2180063426,-0.1570498347,-0.2422677279,-0.3172673881,0.0060723042,-0.2187599391,0.1110729054,0.5873727202,-0.147653088,0.5054621696,0.1277804971,0.2469361275,-0.2689008415,0.4203834534,0.1096115187,0.2378640324,-0.0426125713,-0.0586549304,0.0642313063,0.1824519932,-0.3406453133,0.1243078187,0.049310647,-0.1399400979,0.4881197214,0.6820531487,0.9486029148,0.0485911928,-0.0662658066,0.2085430324,-0.1202166528,0.2622591853,-0.6480964422,0.0284756273,-0.2240751237,-0.1116249934,-0.0257987678,-0.0222448278,-0.056581594,-0.132077232,0.1193469539,0.0120922439,-0.1102351621,0.2542645037,0.1482194066,0.254537493,0.0193994977,-0.2727670074,0.0965916589,0.1027325988,-0.2582747042,-0.0273498707,-0.0858013183,0.1085177734,-0.0950675011,0.0794278309,0.1427856535,-0.174799785,-0.3708379567,-0.0117371436,0.1937059015,0.0838192329,0.3218397498,0.0133659244,1.0088208914,0.2802313864,-0.0377495959,0.1045229211,-0.101484172,0.1387185752,0.0913878828,-0.180818513,0.3549826443,-0.1826035976,0.0734123439,0.0921842381,0.4235817492,-0.1495436877,0.1895076931,-0.2699933052,0.2245288491,-0.0113026304,0.349182725,0.1361858994,0.2687270939,-0.008623899,0.1014510021,0.0327432342,-0.1309930533,0.0639822558,0.0913363844,-0.4561839402,-0.1453894377,-0.1443601549,0.1587893367,-0.564089179,0.0211050119,-0.4806112051,-0.0929630175,-0.0470437817,-0.209301576,-0.0937464908,-0.2119074166,-0.0343812928,-0.2623495162,-0.3972578943,-0.1727908701,0.478001982,0.446559459,-0.1847447753,-0.0732816756,0.0263588745,-0.2604839504,0.1498878896,0.0247075055,0.168755874,-0.1278529316,-0.076144442,-0.2950933874,0.1716803908,-0.3109256029,0.0143059185,-0.1636409611,0.0499809831,-0.1476820856,0.0125603611,0.2107366025,0.1065220088,-0.0869776607,-0.015522345,-0.5122631788,-0.124829717,-0.5443797708,0.1313416809,0.218108803,-0.1643410921,-0.158419311,-0.0501495525,-0.0653146505,-0.0582474098,0.1735433638,0.1902839243,0.0992499813,0.0434955545,-0.1439116299,0.352876395,0.2970356047,-0.067028828,0.2380244285,0.4041190445,0.2627183795,0.1373497844,0.1009254754,-0.0760799795,-0.0886748731,0.0872029215,0.3086190522,-0.0052407989,-0.1128635779,-0.1133657023,-0.0444421582,-0.316631645,-0.0139574828,0.1433970034,-0.2234318554,-0.6269469261,0.4784930646,0.1365952343,0.0922521427,0.1207950711,-0.1289558113,0.3484460413,-0.0992507711,0.3846881092,0.9073587656,0.0232122932,0.0948701575,-0.1444477141,-0.066853784,-0.4905072451,0.2459781915,-0.3952052891,-0.1683618277,0.4017065465,0.2897826433,0.0517583489,0.1405251175,-0.3349552453,0.0645502657,0.5856606364,0.0408751629,0.1640686691,0.1820465028,0.2152882218,0.2936815321,0.1227067858,-0.1782294065,-0.0386111215,0.2500814497,0.6316232085,0.2004634291,-0.4587742686,-0.0536152683,-0.0187727474,-0.2213171422,-0.1653882712,-0.2391542494,-0.0933974534,-0.2279879898,0.1647833437,-0.1332666427,0.1003108472,-0.4623878598,0.1629349142,-0.025216097,-0.0677808821,-0.2709767222,0.0659302771,0.4809463322,0.0542778932,-0.1677126437,-0.3265404999,0.1090090349,0.1438055336,0.2817612886,-0.1133834422,0.1096930578,-0.0980447531,-0.3302260041,-0.0046820473,0.4351281524,0.1858336478,0.2620585263,0.2869015336,-0.1878167838,0.2493180186,0.0133465026,-0.1446851045,-0.1115254313,0.0267667025,-0.0584781282,-0.1519646049,0.3079360723,0.1857694983,0.2157376558,0.0239844322,-0.0553950258,-0.2486840189,0.2157903761,0.3891343474,-0.093095243,0.2307754308,-0.2237196565,0.1051014662,0.1636086851,0.168473497,-0.0344291925,-0.0173891801,-0.1156633124,0.23293598,-0.3763502836,0.3879484534,0.5590148568,0.3002312183,0.1045146883,-0.1096533015,-0.1282139719,0.2044407427,-0.1847605705,-0.013387844,-0.0819420964,-0.3096718192,-0.2206022143,0.0356080793,-0.0696326792,-0.018621156,-0.1757169664,0.0097310403,0.0133415945,-0.3776822686,-0.0208002888,-0.3679774106,-0.20080024,0.1272331178,0.3974537551,0.5252466798,0.3474843204,-0.1461521387,0.2490387857,-0.3213901222,0.2069022059,0.0504212156,-0.1658450514,0.002802032,-0.0882347673,0.1128154472,0.0582917631,0.2106526494,-0.4103106856,0.4649715126,0.2110950202,-0.6535023451,-0.1779310107,0.0222193226,0.2327961922,-0.0758989006,-0.0391966738,0.4191893041,0.2298194915,0.3273516297,-0.1562206596,0.043168705,0.2353197336,-0.4685609937,0.1130018532,-0.1070158929,0.271256268,0.5209808946,-0.0823811591,-0.7038664818,-0.3048591614,0.0180623624,-0.1834253222,-0.3575614989,0.185520649,-0.237533778,-0.2151159793,-0.1924316138,0.0931557193,-0.162741065,-0.335516274,0.1316692531,-0.1930133998]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"Yeah, I would love to have discussions about ways this project can have an community-based, transparent process to arrive at strong default metrics. @kpu \/ @mjpost do you have any suggestions of how that might work or pointers to places where this is done right? Perhaps this question can be template for what is likely to be repeated for other datasets.","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":61,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nYeah, I would love to have discussions about ways this project can have an community-based, transparent process to arrive at strong default metrics. @kpu \/ @mjpost do you have any suggestions of how that might work or pointers to places where this is done right? Perhaps this question can be template for what is likely to be repeated for other datasets.","embeddings":[-0.1317217052,-0.1521494538,0.0231624022,-0.2831539512,0.183517158,-0.4405026138,0.1937044263,0.0987422317,-0.3281913102,0.5617699027,0.0751185417,0.3197999895,-0.081899412,0.0583840534,-0.1340554357,0.2278729379,-0.0086308643,0.0542876907,0.5130555034,-0.0493563786,-0.0386517942,-0.2196662724,0.0258152429,0.32539469,-0.2940972149,0.1930957735,0.1951013207,-0.2362080663,-0.371363759,-0.3428968191,-0.0088307867,0.3811078072,-0.3085629046,0.2572025359,-0.0001145226,-0.6301751137,-0.0216781627,-0.0355617106,-0.095885314,-0.4296736419,0.1394093484,-0.1895246357,0.1505590975,-0.1307041198,0.0372194983,-0.2564647198,0.0646442622,-0.0690381005,0.4480718076,-0.1285511404,0.0754716024,0.2396064252,-0.6704871058,0.201708883,-0.2710964084,0.4848407507,-0.0352256298,0.172689423,0.1593183279,0.0996103436,-0.0320559107,-0.1317351609,0.0314794369,-0.2312244922,0.159571439,-0.0798824728,-0.1958988309,-0.2130770832,0.0196533427,0.6208432317,-0.0205363967,0.0141185746,-0.0521378741,-0.2381584197,0.0941509455,-0.3407889307,0.0405829214,-0.0104211178,-0.0858964175,-0.1650648266,-0.0927200541,0.0869041011,0.0539114401,0.0782720372,0.1201411933,0.1560006887,-0.071630843,-0.0470873192,-0.2539201975,0.0088730408,-0.239177987,0.0736634284,0.2476436943,0.1963906139,-0.1704584509,-0.2651331425,0.175860554,-0.3175501227,0.1600955129,0.1721196473,0.2469453514,0.3688746691,-0.3298505843,0.1381085664,-0.222165823,0.020336641,-0.0540364385,-0.2640292645,0.5498660803,0.0829504654,0.320584923,0.2071528137,0.0846228004,-0.0641749725,-0.0983875915,0.3398284316,-0.2787473202,-0.2240519077,-0.5955854654,0.1028965637,0.0405031256,-0.0313931964,0.2300077081,0.0795434266,-0.2954149544,0.0870085135,-0.1917365193,0.2282932997,-0.1924938262,-0.1655410081,-0.0242982265,-0.3788319528,-0.2206811756,0.0180576015,0.1756093204,-0.0845315456,0.3501525223,0.1961226761,0.5397789478,-0.1819541007,0.125162214,0.0132761793,-0.0413131379,-0.1623577774,-0.3348377347,-0.0199481305,0.0880881995,-0.1897372156,-0.0577017069,0.006054183,-0.1901556104,-0.4022370875,0.104691267,0.0129621625,0.0042207995,0.1865674853,-0.4194647968,0.2759493887,0.0240425915,0.2129327804,0.1609382182,-0.0490422025,-0.0290443432,-0.2775923312,0.3328245878,-0.1258491427,0.0587262623,-0.3050249219,0.4064477086,0.1433938295,0.0249315128,0.3754587173,-0.2546948493,0.3886653185,0.1451201886,0.0814343616,0.0819286555,0.0403046981,-0.1115597934,-0.0964670032,-0.1659022272,-0.2990907729,0.1548764855,-0.0652782097,0.4166331887,-0.2565686405,0.0135699064,0.2856188715,-0.0146813439,-0.0663690194,-0.2931022048,-0.5398417115,0.1623521,-0.1193594337,0.2307261676,-0.0030943905,0.2129999548,0.4048301876,0.1429204494,-0.329359889,-0.0748436823,-0.2866929471,0.807256043,-0.0472326502,0.2115935087,0.1206554994,0.4799895585,0.0730224401,-0.0094607957,0.146462515,0.0227705929,-0.4362933934,-0.2402880788,-0.0036761994,0.2924123406,0.3435100615,-0.0270517971,0.0217133202,-0.5986208916,0.2004327774,-0.5100762844,-0.0765308142,-0.6761402488,0.2280262113,0.2782546282,0.0366126001,0.1347272098,0.0282939859,0.1308628619,0.2619227767,-0.0363815874,0.037637081,-0.1882143319,0.3101923466,0.0034291248,0.3451182842,-0.2935290337,0.6801664233,0.4701668024,0.036768917,-0.2388762236,0.1942770481,-0.2076038122,0.152276963,0.0659182221,0.6317341924,-0.0831610411,-0.0194512699,-0.0466136448,-0.1456830353,-0.2612492442,0.1453503221,-0.1501732916,-0.1219528615,0.1595245153,-0.0031712626,-0.0224011224,0.092191197,-0.108019717,0.0781614706,0.0624217875,0.2085682303,0.1686545759,0.061268758,0.5531918406,-0.2921215594,0.1145000905,-0.03073694,-0.014236033,-0.0168158356,-0.0519746616,0.0130428346,-0.2535009384,-0.1804722548,-0.156113416,0.0416792333,-0.1151286811,0.22166574,0.0259586163,0.3211395144,-0.1789851338,-0.0223924872,-0.0044539077,-0.0689853132,-0.1949260533,0.0165647361,-0.1251769811,0.2771679461,0.0805467963,0.0781259313,0.0140539464,-0.223878324,0.2112262547,0.0069617061,-0.1106193289,0.3951956332,0.158703059,0.9091917872,-0.1935369223,0.0631669834,0.002672917,0.0989362821,-0.0681502298,-0.0395331569,-0.0933667645,-0.1807143092,0.2783333957,-0.0459964648,-0.0017116976,-0.0095604891,-0.8639891148,0.0664968863,-0.0342518725,0.3804614544,0.1018430889,-0.2858391106,-0.3674679995,-0.1909762025,-0.096568875,-0.2331636548,-0.0682037324,-0.1002083346,-0.0169144161,-0.1037079915,-0.0993808061,-0.2951435745,0.1288128942,-0.0396173932,0.0516296364,-0.1908406615,0.2184209377,-0.3723987341,-0.0398934819,0.1063477024,0.1864712983,0.4068046808,-0.0509282574,-0.0777145028,0.1177450866,-0.3018659055,-0.1925506294,-0.2463868111,0.008397405,0.1748577058,0.1515838802,-0.3545621336,-0.3071526587,0.3153419495,0.0824796259,-0.0236576833,-0.1282675266,-0.0953922644,0.1880735755,-0.0482724272,-0.0030243811,-0.0281915851,0.0706559047,0.1070397198,0.5870996118,0.1151951626,0.1844270378,0.6375483274,0.8891425133,-0.0007610413,-0.1633932143,0.0511823706,0.3028786182,0.1601830423,0.102950938,0.0815984756,0.2571055889,0.1289468408,-0.1197657809,0.4167870581,0.1457354128,0.2156973481,0.0906536132,-0.0239538271,0.197006613,-0.0854272023,-0.1111895069,-0.2246178687,0.1880072951,0.0435619913,0.1278184652,-0.1375352144,-0.321179986,0.2428238392,0.258426249,0.0896539167,0.0685442239,-0.2249547243,0.2045938075,-0.3556163013,0.0844413564,-0.1366856098,0.451364398,-0.1491567492,0.1695667952,0.1197335273,-0.0852375776,-0.1516986191,-0.0945376679,-0.0530011505,-0.2609036863,-0.0596010908,-0.1646456718,0.1009160206,0.054479979,-0.0627342835,0.3288066685,0.4255048931,-0.503269136,-0.2165864557,-0.1952255815,-0.0866937861,-0.0723471642,0.1607702523,-0.1748057157,-0.2509565055,-0.050684683,0.1388920248,-0.2789742649,-0.0109682027,-0.2774152458,0.2627337277,-0.3664969206,0.064585939,0.3063522875,-0.0061462717,-0.2401485443,-0.0119028045,-0.0181686841,-0.003760817,0.234649241,0.0261543076,-0.2223065197,-0.1262037605,-0.2266125679,0.0377428122,-0.2525560856,0.2906058133,0.5935212374,-0.1272306144,0.3823327124,0.1300915331,0.2303987592,-0.3150281906,0.5367191434,0.1276200414,0.0791494548,-0.0040222192,-0.0593618117,0.1446081847,0.0980578139,-0.4596040845,0.2224507481,0.1280931979,-0.1608809084,0.7775840163,0.6960657239,0.9820846319,0.0039522955,-0.0721495599,0.1592865288,-0.0825304985,0.199428454,-0.5290594101,-0.1130350754,-0.1436690241,-0.2249892652,-0.0384616219,-0.1100104079,-0.0650706142,0.0460722297,0.1236453429,-0.0724264011,-0.0467435718,0.2753323913,0.2509953082,0.3288512826,-0.0310064089,-0.213625446,0.1472299993,0.0207603276,-0.1194540411,-0.0915803015,-0.1198377162,0.1407984644,-0.1180053949,0.0931623653,0.1807111353,-0.1499829888,-0.1702421457,-0.0228035506,0.3271114826,0.014522451,0.394567132,0.3165232539,1.026964426,0.5743095875,-0.2811959684,0.0714817345,0.0430507921,0.1191386208,0.2622564137,-0.2490708679,0.2586089969,-0.0113669261,0.0802525133,-0.0578174554,0.4538450539,-0.1262694001,0.0497017093,-0.4980685711,0.3181197047,-0.1468976289,0.2496243119,0.0851846561,0.182674095,-0.1274155825,0.0465359911,0.0235979669,-0.1247427464,0.2199732512,0.1748704165,-0.3204934001,-0.1589328796,0.0713864714,0.0423103906,-0.5956210494,0.0675251633,-0.4984449744,-0.0839834884,-0.0324962363,0.0393840596,0.001255934,-0.0194753353,-0.1504121572,-0.373549968,-0.3433409333,-0.0752502829,0.5095216632,0.5031796098,-0.0972423926,-0.0008873265,-0.0953390896,-0.2102210969,0.2857161164,0.0276878327,0.1540140063,-0.1543096155,-0.11516238,-0.2759826183,0.2546269,-0.2463447899,0.0769503936,-0.0853328854,0.0626340434,-0.225712046,0.0481401831,-0.1233188063,0.1022133753,-0.0997200534,0.0495045669,-0.2831122279,-0.0458190627,-0.4601477385,0.1936975271,0.2755983472,-0.1209088862,-0.1152131632,-0.1827187985,-0.2387259901,-0.1846671849,0.0051862583,0.1594102234,-0.0069738836,-0.1387788057,-0.108800441,0.3613179326,0.1449435502,-0.1934643537,0.2268932909,0.1107136309,0.25266096,0.1844537109,-0.024416374,-0.0828995258,-0.0088782273,0.3117852807,0.4678493738,-0.1286132634,0.0326309875,-0.2066221088,-0.1152115166,-0.4381731153,0.1446431428,0.0749288052,-0.2063016742,-0.4469510615,0.4616006315,0.0484655388,0.0467185676,0.1793043017,-0.2502280176,0.3303585351,-0.0393883884,0.4395038784,1.0022841692,-0.0217681061,0.0498468615,-0.1830810457,0.0363966636,-0.4998032451,0.1945626885,-0.3181962669,-0.1974983513,0.328212738,0.4570223689,-0.00530395,0.0269596092,-0.2245129794,0.1818577945,0.5301968455,0.0648684874,0.0644666702,0.0855077058,0.1574232727,0.4331811666,-0.0030112832,-0.1559503078,-0.1252391785,0.117864117,0.6670604348,0.0246087965,-0.293050617,-0.1094190925,0.0293158591,-0.1487194151,-0.3060190082,-0.4843788445,-0.2215511352,-0.0690992922,0.0571747348,-0.1304025948,-0.0349585712,-0.5705887675,0.134568125,0.0988407806,-0.1397086829,-0.1329700649,0.1201607585,0.5725661516,0.008557138,-0.0652304888,-0.4107311666,0.1811183095,0.1349814981,0.423820287,-0.1089939624,0.2727339566,-0.3250204027,-0.3037956357,-0.0341927223,0.3719822168,0.260786593,0.2767417133,0.3163920641,-0.2537290454,0.2587060034,-0.0347202979,-0.1072571725,-0.2261417955,0.0305062067,-0.0201508943,-0.2125938684,0.2808629572,0.1410847306,0.1751459986,0.2730838656,-0.0957812145,-0.2364907414,0.2019379288,0.2923271358,0.0150731159,0.3490414917,-0.2561854422,0.0650185496,0.0890456587,0.0466663502,0.0944809839,0.1193936244,-0.1784594506,-0.0508390777,-0.4504283667,0.3851357996,0.5483930707,0.113631919,0.0287724119,-0.0870806947,-0.0440967418,0.2510688007,-0.3328098059,0.0829497203,-0.1790488064,-0.1630083919,-0.1391744912,0.0519334376,-0.0260232035,0.1079453006,-0.4039847255,0.0211441182,0.0460030511,-0.3065639436,-0.0477998406,-0.1658798605,-0.0909435898,0.1227347404,0.3015193045,0.4139880836,0.281003356,-0.1094645485,0.2570030093,-0.2844336927,0.0847519413,0.0271946006,-0.1665120721,0.0108845383,-0.0228679795,0.2588674128,0.0451126471,0.0209871475,-0.2498360276,0.2707715929,0.3818492591,-0.6561614871,-0.1667830348,-0.219191581,0.0337466858,-0.1286293119,0.005715745,0.361328721,0.086560078,0.5356072783,-0.0908006057,0.0742844641,0.101832144,-0.4056444168,0.0896817073,-0.4716910124,0.4330797195,0.4989500046,-0.1134616807,-0.3102812767,-0.2785300314,0.0997307077,-0.2346298695,-0.2983546555,0.0771568194,-0.1645468026,-0.1145866215,-0.3277128637,0.072635226,-0.2579048276,-0.1299065948,-0.0640186518,-0.1937729418]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"I think @bittlingmayer is referring to Figure 6 in http:\/\/statmt.org\/wmt19\/pdf\/53\/WMT02.pdf . When you look at Appendix A there are some cases where metrics fall apart at the high end and some where they correlate well. en-zh is arguably production-quality. \r\n\r\nThis could evolve into a metrics Bazaar where the value add is really the packaging and consistency: it installs\/compiles the metrics for me, gives a reproducible name to use in publication (involve the authors; you don't want a different sacrebleu hash system), a version number, and evaluation of the metrics like http:\/\/ufallab.ms.mff.cuni.cz\/~bojar\/wmt19-metrics-task-package.tgz but run when code changes rather than once a year. ","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":101,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nI think @bittlingmayer is referring to Figure 6 in http:\/\/statmt.org\/wmt19\/pdf\/53\/WMT02.pdf . When you look at Appendix A there are some cases where metrics fall apart at the high end and some where they correlate well. en-zh is arguably production-quality. \r\n\r\nThis could evolve into a metrics Bazaar where the value add is really the packaging and consistency: it installs\/compiles the metrics for me, gives a reproducible name to use in publication (involve the authors; you don't want a different sacrebleu hash system), a version number, and evaluation of the metrics like http:\/\/ufallab.ms.mff.cuni.cz\/~bojar\/wmt19-metrics-task-package.tgz but run when code changes rather than once a year. ","embeddings":[-0.2435240746,-0.3005855381,-0.0543989949,-0.2982117534,0.0659073368,-0.2633567154,0.3223109543,0.1647462398,-0.3544036448,0.3798716962,-0.0478951,0.277641356,0.0203537755,0.0332901925,-0.0991292,0.1471895427,0.0948516726,0.0232937969,0.4395563006,-0.2320107222,-0.1570619494,-0.2324187309,0.0439792275,0.3475279212,-0.0711236671,0.3041613996,0.2054768205,-0.1155458763,-0.2697204649,-0.2693089843,-0.1757273972,0.2499205321,-0.2212212086,-0.0528266951,-0.0001057406,-0.4228647649,0.0307961851,-0.1002201959,-0.3133630455,-0.4734850526,0.0775697753,-0.3150503635,-0.230293557,-0.2247039825,0.3282446265,-0.1976822764,0.074719131,0.0410479419,0.2474143505,-0.0185286161,0.1679522544,0.4588983953,-0.5752641559,0.0494604595,0.0216069166,0.1226708516,-0.0413586721,0.1909916997,0.1126721501,0.3767433763,-0.1389332861,-0.0326629393,0.1147210747,-0.2191132307,0.1291884184,-0.0577258244,0.1619864702,-0.0667688549,0.0551364981,0.5585500002,-0.0398028456,0.1188730001,-0.3280954063,-0.038841553,0.0573453493,-0.4812346995,-0.0124926111,0.0695373937,0.0995743573,-0.1530035585,-0.1505928636,0.1242723092,-0.0253851172,0.161043033,0.1169870496,0.2148509175,-0.0379261188,0.0777445808,-0.1787698269,0.1447507441,-0.1647797078,-0.0201989692,-0.0308875758,-0.1030920148,-0.1020114943,-0.1833823472,-0.0556736961,-0.4369937778,0.1886058748,0.0985735953,0.2170275897,0.3052583337,-0.0982056484,0.1443403959,-0.1386347115,0.188333258,0.0615375303,-0.3308641315,0.3270785809,0.0987469554,0.4898584783,0.1428254098,0.1948443651,-0.1527297348,-0.05732207,0.1905162334,-0.5074321032,-0.3131028712,-0.7329565883,0.1527253985,-0.0937250257,-0.1700952947,0.1190794334,0.1927861124,-0.2124164104,-0.0652727559,-0.2983232141,0.1727706492,-0.1432597935,-0.0799588859,-0.0384171531,-0.2377151847,-0.2817411721,-0.0238065831,-0.1102033556,0.0284593925,0.3472851813,0.1121465787,0.5056610703,-0.0675382242,0.0374996625,-0.0223302878,0.0134182582,-0.1105607897,-0.3938734233,-0.1357773095,-0.0827027485,-0.2144176066,-0.0691758096,-0.0673356727,-0.4156737626,-0.3330410123,0.0599923916,0.1190014035,0.0205701366,0.175370425,-0.1397430897,0.1198915243,0.057757549,0.1586414427,0.3195588291,-0.2251321673,-0.121294044,-0.1884955168,0.1353850365,-0.3119530678,0.0680036768,-0.2918475866,0.4261809886,0.2250981331,0.2090041488,0.3553487957,-0.3932777047,0.3532279134,0.1524048448,0.0343509801,0.2255767584,-0.3598628938,-0.2300913334,-0.1324645132,-0.0051909862,-0.3282734752,0.0394367389,0.0217808187,0.4474590719,-0.3550918102,0.154987514,-0.0488220751,0.0130619975,-0.2862188518,-0.2684639692,-0.4319381416,0.183356449,-0.0564962253,0.2309053242,-0.1190957651,0.2098727822,0.588730514,0.1590449661,-0.3271045983,-0.0990328118,-0.2509108484,0.5768555999,-0.1068919823,0.1817075759,0.1966934204,0.416318655,0.1697402894,-0.3384238183,0.1830931753,0.1225446984,-0.334172368,0.0349693969,0.0486892313,0.260468781,-0.0303768795,0.0571796373,-0.0929929689,-0.5172767639,0.1460862905,-0.4348348379,-0.1211315393,-0.520383954,0.055131115,0.2200426906,-0.0343097076,-0.1001343355,0.2275537252,-0.0255614165,0.3954468966,-0.0854102448,0.0379673205,-0.2463258952,0.1997847706,0.0000620177,-0.0702047646,-0.1267820597,0.5026574135,0.412874788,0.1263271719,-0.1499578357,0.1077883393,-0.1631533802,0.1223770455,0.1445949376,0.5002867579,0.0605948344,-0.1426591277,0.0556550436,-0.2584556937,-0.1794898957,0.0480559543,-0.168738097,-0.2170414031,0.2031293958,-0.0375201888,-0.2641672492,0.0120406179,-0.0362933204,-0.004235981,0.529820025,0.0890455693,0.0981049761,0.0477167778,0.4727903605,-0.2430368364,0.1521220058,0.1712397933,0.151741609,0.0604051799,0.0962107703,0.2173137516,-0.2510479987,-0.2809517384,-0.1332310438,0.1576150507,-0.090426825,0.2553810775,-0.0998248681,0.4683686793,-0.061147321,-0.1624720097,-0.0547122993,-0.1097156778,-0.0894951522,0.189157933,-0.1510652751,0.4585855305,-0.0595226102,-0.0563360937,-0.1826070994,-0.2930006385,0.2888199687,-0.0517311879,-0.0868313909,0.3431066871,0.08635851,0.6125349998,-0.2508165538,0.1708940268,0.0203667507,0.0727959424,-0.2392204702,0.0346356444,-0.2137923986,-0.1237191781,0.3305430114,-0.0502122454,-0.1832588315,-0.1312793493,-0.7176876068,0.0469175093,-0.0709975362,0.2612710595,0.1533103883,-0.4949451387,-0.5975624323,-0.1073579192,0.0488347746,-0.089778848,-0.0834562406,-0.0011627523,-0.0557501502,-0.3762321472,-0.0986449718,-0.336466074,0.0467078649,-0.1315041035,-0.052976653,-0.2506543994,0.0623446479,0.0420722365,-0.1769198477,0.1221345663,0.2492327094,0.2456513941,-0.2242902815,0.0694021136,0.113811709,-0.2636437714,-0.3411216438,-0.1655087024,0.0977164507,0.3935262859,0.2329139411,-0.3390344381,-0.3138285577,0.3934340477,-0.1104428917,-0.0339713879,-0.1589855552,-0.0925576314,0.1727854013,-0.1794543117,-0.02423691,0.0970856994,0.3053620458,0.2019963413,0.5900418162,-0.0375420824,0.1662781537,0.6131906509,0.640231967,0.1696110666,-0.0285704639,0.1647343338,0.3031334281,0.2151878923,0.0773576498,0.0306413323,0.3047842681,-0.0848558471,-0.0923094824,0.4445631802,-0.0694647953,0.2872692049,0.0495462567,-0.133614406,0.4021225572,-0.1581473053,0.0557464361,0.006533613,0.062239483,-0.0458670966,0.2046813071,0.1008251235,-0.2437957823,0.4465916455,0.1306043863,-0.2133849561,0.042942442,-0.2132988423,0.1193669811,-0.3069757223,0.2397652864,0.0184674468,0.618598938,-0.1211643592,0.1879020333,0.0202082582,-0.1702572852,-0.2018499076,-0.1748800427,0.0748123229,-0.2568154931,-0.0122299157,-0.3013005853,-0.041637145,-0.1608273238,-0.0306203775,0.24577263,0.3923413754,-0.3552401066,-0.1640493423,-0.0328795537,-0.0073479032,-0.1909746379,-0.043513298,-0.3017178178,-0.2088459283,-0.1884808838,-0.0239130389,-0.313796699,-0.0148679828,-0.1688668579,-0.0454512723,-0.2747916281,0.1033465862,0.3190315068,0.0205892026,-0.4130380452,-0.0418808572,-0.0249932855,-0.2193040103,0.2226120681,0.2926400006,-0.0402610824,-0.2669600546,-0.3590785861,-0.0327232331,-0.2441360354,0.1974163353,0.4649688005,-0.0027600266,0.1967924386,0.2424983829,0.2750263512,-0.1388515383,0.3099518418,0.1948876381,0.298732549,-0.2369791418,0.0276008062,0.0976976082,0.0072554015,-0.2170911878,-0.0346039198,0.1834515184,-0.1264139861,0.5017802715,0.712295711,0.9287544489,0.2515047491,0.0910652056,0.1519112736,0.0069029187,0.1555326879,-0.6266521811,0.0212753434,-0.2069049627,-0.3031750321,-0.0490292832,-0.0868972167,0.051091373,0.1322292536,0.017037129,0.0770195723,0.1021883115,0.2828607559,0.1404205412,0.2446164042,-0.0685038716,-0.1947111636,0.2222368121,0.1003433093,-0.0990998223,-0.0594044849,-0.1588498205,0.0864407122,-0.2373163104,0.1734232008,0.2514417768,-0.0805179179,-0.1611944884,0.1702409089,0.0391096696,0.0772239193,0.313806802,-0.012533877,0.9258645773,0.3882945776,-0.0377799422,0.0006564173,-0.156676203,0.1428188831,0.2550014555,-0.2980361879,0.4793711901,-0.2001491338,-0.0457986221,0.2099068165,0.4611559808,0.0084433155,0.177986905,-0.1972344965,0.1296453625,-0.0717757791,0.2466275543,0.2059281468,0.2094620466,-0.0797127038,0.0913936198,-0.0414937958,-0.1508346051,0.1583678722,0.1381304562,-0.4143410325,-0.1061489657,-0.0900033489,0.0441560075,-0.2025603205,-0.1179731786,-0.5925447345,-0.1234831214,-0.0317794532,-0.0961137861,0.0238800645,-0.2189496905,-0.1911744475,-0.3086482286,-0.1834496856,-0.2132280916,0.4570011199,0.5905830264,-0.2766248584,-0.0699192658,0.0085556554,-0.1150544733,0.185362637,0.1301269233,0.2473983765,-0.0786296278,-0.1963204592,-0.2025486976,0.2570351958,-0.3319363892,-0.0646318123,-0.12603046,0.0538067222,-0.271432519,-0.0365145542,0.1022507101,0.0821414068,-0.0650741458,-0.1030980498,-0.586448431,-0.1108660325,-0.4371490479,0.1243138462,0.356556952,-0.0725397617,-0.1796937287,-0.0323166288,-0.0666691959,-0.1997256428,0.0696001127,0.1717759818,0.0666337386,0.2180798799,-0.0696023777,0.4152755141,0.1327683628,-0.0613161214,0.3083026409,0.4883522391,0.2198681682,0.2059158534,0.0483122133,-0.0671828613,-0.335375607,0.2804253399,0.4356293976,-0.0782048926,0.0479052663,-0.166779846,0.051418595,-0.3711835146,0.1109727919,0.2058674693,-0.1694428623,-0.6405297518,0.4711530209,0.1569986045,-0.0250190552,0.0025432603,-0.0524795949,0.2136241794,-0.0841054395,0.3174225092,0.9295864701,-0.2271815389,0.1400176883,-0.1663311869,0.0165334195,-0.5663512349,0.1403526664,-0.3035806119,-0.0016455076,0.4644860029,0.281345427,-0.0494692698,0.2601789236,-0.0808808953,0.1376149654,0.6137875319,0.2619395256,0.2576929331,0.0514376462,0.3690963089,0.3331646323,0.1753263474,-0.1328465044,-0.0645269677,0.3387180567,0.3999513686,0.1680492759,-0.3725417256,-0.2459826022,0.087704435,-0.250395298,-0.1470989287,0.0127507905,-0.0871345252,-0.135726437,0.2197871357,0.0109248403,0.0635628179,-0.5164676309,0.2015509754,0.0450630337,0.0306923594,-0.1617545187,0.1428954452,0.3391336799,-0.0526367724,-0.2491468489,-0.1385196149,0.0254379511,-0.0300068203,0.1781278402,0.0066906777,0.0554639697,-0.1432782561,-0.2987905443,-0.0175468922,0.384249568,0.1524682045,0.3373552561,0.2429819107,-0.277423948,0.3964841366,0.0460667647,-0.1380787045,-0.2208641618,0.1454420984,-0.2153338492,-0.1817860156,0.2156696916,0.3408409953,0.208739385,0.1939417273,-0.004321089,-0.3412438035,0.1143863499,0.4114850461,-0.1555652916,0.2408095151,-0.2376056612,0.1159142554,0.1119886786,0.0707859471,0.0259179864,0.1004330218,-0.0963151008,0.3075036407,-0.3312196732,0.3280593753,0.4525681436,0.2827616334,0.0231592152,-0.0858401135,-0.1080540642,0.1935911626,-0.0431341976,0.0167215709,-0.0409584232,-0.2080360949,-0.130908668,0.2298148274,0.0387792289,0.0189993419,-0.2677189708,-0.0503974631,-0.018615691,-0.4426839352,-0.0158457998,-0.2415637225,-0.0676975548,0.1904706955,0.1778808385,0.6185246706,0.1885393709,-0.0592788905,0.2392137349,-0.251529932,0.3539308608,0.1815538257,-0.0367932916,0.0922107249,-0.0303095467,-0.0256262496,0.1017006189,0.0696415901,-0.2569894493,0.4339093864,0.3335089087,-0.6859284043,-0.2197222859,0.1127903685,0.1804609001,-0.1644693464,-0.0388458632,0.2896651626,0.2651185095,0.3415920436,-0.0562999658,-0.013357047,0.2013347149,-0.3302735388,0.185841769,-0.2076367289,0.3184798062,0.4439646006,0.0746592879,-0.5812065601,-0.4063215554,0.1298279613,-0.0837375373,-0.4195715785,0.1802417487,-0.0641491413,-0.164132461,-0.2556312084,0.1596225351,-0.4075241387,-0.5336329937,0.1822560579,-0.2612043917]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"While a Bazaar setup works for models \/ datasets, I am not sure it is ideal for metrics ? Ideal from my perspective would be to have tasks with metrics moderated by experts who document, cite, and codify known pitchfalls (as above^) and make it non-trivial for beginners to mess it up. ","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":52,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nWhile a Bazaar setup works for models \/ datasets, I am not sure it is ideal for metrics ? Ideal from my perspective would be to have tasks with metrics moderated by experts who document, cite, and codify known pitchfalls (as above^) and make it non-trivial for beginners to mess it up. ","embeddings":[-0.2716112435,-0.0951522961,0.0032678251,-0.3119646907,0.0734552667,-0.3363656998,0.4298071563,0.1751324683,-0.1235072389,0.4883823991,-0.2467203438,0.3768928349,0.0635004863,-0.0168901887,-0.1135296822,0.0980852991,0.0930720791,-0.055050876,0.3823439181,-0.0338674523,0.0045135221,-0.1527675241,0.0314827897,0.3090196252,-0.2237772942,0.0283721518,0.1682085246,-0.2984610498,-0.4094160795,-0.3030361235,-0.2503064275,0.3803891838,-0.1480116546,0.2917005718,-0.0001151704,-0.4013819993,-0.0981762186,-0.059069518,-0.3827992976,-0.4775865078,0.366587162,-0.3013215661,0.0795258358,-0.1160105169,-0.0342504419,0.0047137495,0.1514754444,0.1454437673,0.3491646945,-0.2688573003,0.1075736284,0.0655706525,-0.6408105493,0.2294312865,0.0010638142,0.0314926282,-0.1193256304,0.386238575,0.1450565159,0.3580458462,-0.1749091446,0.0084542409,-0.0281790588,-0.2371222973,0.0730595812,-0.0983801559,-0.1292035878,-0.1403798014,-0.1035128832,0.4384154081,0.0049730209,-0.0973153636,-0.3281689584,-0.2922053933,0.1905197054,-0.5373380184,-0.0360234864,0.023728719,0.0971384794,-0.0445607081,-0.0681914166,0.1502565444,0.1634793282,0.1564424038,0.2562790811,-0.1061417088,0.0252091084,-0.0557945408,-0.1979496032,-0.02891366,-0.3988059461,0.0176837333,-0.1329602599,0.1177844629,-0.4859588444,-0.2479260117,-0.0482075587,-0.1725309044,0.1114022583,0.2292976081,0.489972204,0.4232983291,0.0896841511,0.1236561462,-0.0064928848,0.1109312624,-0.0501702875,-0.5000490546,0.4304535389,0.0321430378,0.5204985738,0.1228310391,0.0374303497,-0.2252338976,-0.1111551151,0.2451990098,-0.2509506941,-0.1593144387,-0.5886031985,-0.0043277587,0.0603515133,-0.0698112622,0.1557022333,0.2324747294,-0.3494310081,-0.2075756043,-0.1804874539,0.273121804,-0.1225871667,-0.0826057792,0.0487322956,-0.242915079,-0.2589273453,0.0374127179,0.1405649632,-0.1522822827,0.2905605137,0.1604045779,0.4069100916,-0.116666846,0.0408542007,-0.0850826427,-0.1896258146,-0.078815937,-0.2262635678,-0.1687243879,-0.2212388366,0.0400544405,-0.0979273915,0.0506458916,-0.5633580089,-0.3021613657,0.3407365084,0.0659127384,-0.1206859499,0.1320510507,-0.2986834049,0.2521048486,0.0110187316,0.0953460634,0.2481513172,0.003308872,0.1383832097,-0.1485918164,0.0209884644,-0.3284705877,0.1403999925,-0.2713214457,0.3406400681,0.2785710692,0.2445607334,0.2890323997,-0.218217656,0.5324132442,0.1623586118,0.2161554247,0.5340870619,-0.2281572819,-0.2581608891,-0.1710297614,-0.0924945176,-0.2751013339,0.1718738526,-0.0064961622,0.5837255716,-0.3159643114,0.2218964547,-0.0814052448,0.0416901223,0.0293860584,-0.2764899135,-0.6142013669,0.3131126463,0.0284156688,0.181056127,-0.1728828102,0.237477079,0.6267751455,0.2742649913,-0.2361765057,-0.1974798292,-0.3107468188,0.5519845486,-0.1135756001,0.2790036201,-0.031590011,0.6670126319,0.0853865296,-0.2899171412,0.3918997645,0.2974933684,-0.4375243783,-0.0417100638,-0.1009925902,0.2452884465,0.2353353053,0.0026308983,-0.1728491038,-0.4428907633,-0.0358213075,-0.5390484929,-0.2909192741,-0.4662596881,0.2581130266,0.2478227168,-0.2186290026,0.0027993179,0.1067201793,0.0158346407,0.4310149252,0.0987120122,-0.0730316415,-0.0549239628,0.2245063335,0.1114878431,-0.0461935811,-0.4230187535,0.5532001257,0.4567530751,0.3584671319,-0.3324280977,0.1174089536,-0.1097384989,0.1080507413,0.3026809394,0.5004863143,0.0299102534,-0.075213857,-0.1159226522,-0.3373015225,-0.143930763,0.1615448147,-0.2121624947,0.025925681,0.1985714734,-0.1820227802,-0.0841758251,0.0348400287,0.0897824466,0.1336916685,0.2881728411,0.1314880997,0.3784699738,0.0170595255,0.4983482361,-0.0974344462,0.0174654704,-0.0808536783,0.064042829,0.1013893709,0.0783709958,0.1416007131,-0.0842412412,-0.3140433431,-0.0360682569,0.0506229699,-0.2652188838,0.1510672569,0.0040256651,0.2384016514,-0.1925639063,-0.1132325456,0.054701861,-0.0137849664,-0.1061579064,0.0103835119,-0.1112547815,0.3257693946,0.0154993515,-0.0412037857,-0.1149904802,-0.0449267402,0.4147253335,0.156196624,-0.1513459533,0.46862638,0.1331629753,0.889862597,-0.3588173687,0.1824448556,0.0361486748,0.074647449,-0.142360419,0.0306678433,-0.1382612735,-0.217179969,0.2067509145,-0.115179956,-0.1540721506,-0.090349175,-0.5959184766,0.1574498564,0.0161785372,0.2387338579,0.1615057141,-0.4224116802,-0.3875550926,-0.0576563142,-0.1962904781,-0.0018593288,-0.077414602,-0.0070118071,-0.0564371496,-0.2640835047,-0.1981953382,-0.1449561715,0.0665342286,-0.2154529393,-0.0363835581,-0.3496148288,0.1044955328,-0.0434419923,-0.1285260767,0.0353559144,0.172819525,0.3967497945,-0.1508047879,-0.1824913025,0.1661066413,-0.2788862586,-0.3064621091,-0.0038894336,0.0966389254,0.2173039317,0.3234605193,-0.0545761324,-0.2681362629,0.4015207589,-0.2236777693,0.239669323,-0.1192862242,-0.1183609366,0.1183238253,-0.0608622953,0.0557308123,-0.116234988,0.2423720807,0.2551181614,0.6405506134,-0.0614358671,-0.0171129853,0.5148363113,0.7666274905,-0.0480543934,0.0574715622,0.0293509457,0.3302820027,0.2424585968,0.0963620096,0.0738739967,0.2383838743,0.048527658,-0.1731742322,0.5399326086,0.0569874458,0.1875561625,-0.0069722421,-0.2128625959,0.1256726533,-0.2127611488,0.1273830831,-0.2768737972,0.0255960859,0.103937082,0.361195296,0.0558472797,-0.3032553196,0.3420774341,0.2542433143,0.0076066842,0.0097127771,-0.3011253774,-0.071728386,-0.3700399697,0.0960406438,-0.1140948758,0.6226420403,-0.20561634,0.1095124111,0.0963866785,-0.0412041843,-0.1886873245,-0.0084529575,0.0644264072,-0.3706533909,-0.1643958539,-0.1212552413,-0.0005474538,-0.076663062,0.0823048279,0.1608067006,0.3634998798,-0.4524725378,-0.0515175015,-0.1945276111,-0.0967377424,-0.2220489085,0.1207466498,-0.3650090694,-0.2577638924,-0.126439631,0.0808782801,-0.2026577443,0.023087319,-0.2422652096,0.1475729942,-0.3873409033,0.0930539817,0.3559022546,0.1420830637,-0.2137625962,0.0310045499,0.0007338201,-0.0288174991,0.3135266006,0.0844775438,-0.1006045938,0.044471968,-0.4483461678,0.0712972283,-0.2901828587,0.168091476,0.5892304778,-0.0501261055,0.4563713968,0.0410432145,0.1389462948,-0.1175750867,0.22737737,0.0125183593,0.3300785124,-0.1189336702,-0.0671241209,-0.1357755512,-0.0378434099,-0.30568856,0.0756613016,-0.0748805851,-0.1807442456,0.6868856549,0.5173199773,1.0146079063,0.0787370056,-0.021251468,0.3734481335,-0.009397815,0.2425405234,-0.7861776352,0.2814872563,-0.212210089,-0.0778138563,0.0191638824,-0.0871705413,-0.0081453631,0.1584015489,0.1307339072,0.1510109752,-0.0363263749,0.2764057815,0.0565410107,0.2006917447,-0.1347208321,-0.2353448272,0.0579283275,0.0223406199,-0.245577544,0.0298246276,-0.2363340855,0.1488296688,-0.1936109513,0.0741281733,0.155123055,-0.1749054641,0.0226380434,-0.0027896948,0.0246134568,0.0327717811,0.1727050096,0.0954932421,0.8078366518,0.3346098959,-0.0223898888,0.048615735,-0.2773946822,0.0121139437,0.183823958,-0.0739840344,0.2357072532,-0.1665159464,-0.0722345188,0.1235775128,0.4882237613,-0.1231636405,0.2154799104,-0.0793339163,0.2397370338,-0.0945619792,0.1983039677,0.1508797854,0.2667447329,-0.0931283459,0.0235778578,0.0221510883,-0.1578657627,0.12619403,0.218570739,-0.34509781,-0.1069146469,0.0952891931,0.1272649616,-0.4926974773,0.1087140068,-0.628934145,-0.1132393926,-0.0409731045,-0.1492630839,0.0094431108,-0.0940173194,-0.1626641154,-0.2838168144,-0.3629394472,-0.1701837331,0.4401834011,0.5001074672,-0.2889373899,-0.2283189148,-0.0966437086,-0.0193130095,0.3344841301,0.268068403,0.1076015979,0.0018625521,-0.1334115863,-0.1324552,0.0562288426,-0.2366832942,-0.0054784124,-0.1397372484,0.1391025186,-0.3680700958,0.0175799131,-0.1285096109,0.1051099449,-0.1527921557,-0.1187517717,-0.47953251,-0.0772052482,-0.4588771462,0.15162085,0.2532347441,0.0194800962,-0.23688896,-0.1563991755,-0.0646227449,-0.1802512109,-0.0223042481,0.2361235917,0.1033843756,0.0372079536,0.0258725062,0.1993011832,0.0319711789,-0.0736460015,0.3168280423,0.1739187986,0.2954163253,0.1351005137,-0.0611759946,-0.0937800109,0.0185369812,0.2630788982,0.5530714393,-0.0146352639,-0.0177889187,-0.2480931133,0.173448354,-0.2884624302,-0.0184250344,0.1141863167,-0.0781220719,-0.5648810863,0.4308204353,0.0625794381,-0.0822058916,0.0752090812,-0.0946582705,0.242233187,-0.0402170569,0.4068408906,0.7529960871,-0.0795914829,0.0255434513,-0.1882151514,0.1550857872,-0.2804107964,0.2354139984,-0.299374193,-0.2011638284,0.4214918613,0.4120867252,-0.0972430035,0.1152781472,-0.0703907236,0.0667415634,0.7104312778,0.1327818632,0.0329430625,-0.0580306277,0.1763578951,0.3949290514,0.1140035987,-0.0641239509,-0.1775739044,0.3646510839,0.4139314294,0.1880567968,-0.4847736955,-0.015321577,-0.0688169524,-0.3042999208,-0.0641747862,-0.1743083447,-0.1911013275,0.1235446408,0.02496529,-0.0924861133,0.1001883224,-0.3238366544,-0.0213374924,0.166619435,-0.0541971028,-0.4020608664,0.3309745193,0.443274647,-0.0125337765,-0.2154942751,-0.3891790807,0.0132996636,0.1771953851,0.4048067927,-0.0600416213,0.1703035384,-0.1342032701,-0.2803755403,-0.1264846325,0.4139946401,0.0436047018,0.2523820996,0.3450066447,-0.1774354279,0.3462061286,-0.0125081912,-0.0972227901,-0.1465275735,0.1158097088,-0.2892940044,-0.1744138151,0.2822023034,0.1961959749,0.1369196773,0.2372735441,-0.0769483298,-0.3263736963,0.1698492914,0.4115197361,-0.1071407273,0.2010266185,-0.331749171,0.0844250768,0.3444564044,0.2096613795,0.0686434954,0.227065146,-0.0830578879,0.3441199064,-0.3188240528,0.2829729617,0.1815317273,0.4891186357,-0.0119347647,-0.1267027557,0.0683197379,0.4705054164,-0.0840444043,0.0952895507,-0.1528708488,-0.1357558668,-0.2434614152,0.0611323416,-0.1101982668,0.2194024622,-0.2932601869,-0.1439932883,0.0557411574,-0.2541647851,-0.1065774262,-0.1062932462,-0.0946832374,0.0392295569,-0.1367759109,0.5326122642,0.1308070868,-0.0100217741,0.3000430763,-0.1281919777,0.1992919445,0.0591074452,0.0110045895,0.1053414717,-0.0249094944,0.034636993,0.0779797807,0.0038931286,-0.1783265471,0.3678911328,0.2766422629,-0.5722068548,-0.0486135967,0.1121278405,0.0945875496,-0.0071757091,-0.1297579855,0.3596749306,0.100467138,0.2457060218,-0.1034886017,0.1777023375,0.2419718355,-0.5511252284,0.0983385742,-0.3128454983,0.5175263286,0.5478052497,-0.0443065614,-0.5008807182,-0.3835437894,-0.0597351231,-0.2204342782,-0.2770809233,0.1814487576,-0.265383929,-0.226154536,-0.2388602942,0.2256522179,-0.3230361342,-0.2756550908,-0.0900107697,-0.3829523623]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"@srush @thomwolf \r\n\r\nModelFront could provide (automated, \"QE-based\") evaluation for all the pretrained translation models you host. Not bottom-up and not valid for claiming SoTA, but independent, practical for builders and not top-down.\r\n\r\nFor that I would also suggest some diverse benchmarks (so split it out into datasets with only user-generated data, or only constants, or only UI strings, or only READMEs) which tease out known trade-offs. Even hypothetical magic eval is limited if we always reduce it to a single number.\r\n\r\nRealistically people want to know how a model compares to an API like Google Translate, Microsoft Translator, DeepL or Yandex (especially for a language pair like EN:RU, or for the many languages that only Yandex supports), and that could be done too.\r\n","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":123,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \n@srush @thomwolf \r\n\r\nModelFront could provide (automated, \"QE-based\") evaluation for all the pretrained translation models you host. Not bottom-up and not valid for claiming SoTA, but independent, practical for builders and not top-down.\r\n\r\nFor that I would also suggest some diverse benchmarks (so split it out into datasets with only user-generated data, or only constants, or only UI strings, or only READMEs) which tease out known trade-offs. Even hypothetical magic eval is limited if we always reduce it to a single number.\r\n\r\nRealistically people want to know how a model compares to an API like Google Translate, Microsoft Translator, DeepL or Yandex (especially for a language pair like EN:RU, or for the many languages that only Yandex supports), and that could be done too.\r\n","embeddings":[-0.3010783792,-0.3047813177,-0.0611800104,-0.3138869703,0.1273147911,-0.3273704052,0.3816562593,0.1755339503,-0.4096497595,0.3841941953,-0.1201138794,0.1421944797,-0.0175862517,0.0955633819,-0.0426307619,0.1114330664,0.2059900016,0.0186355095,0.5584244132,-0.1555034071,-0.2218505144,-0.1242716014,0.0931644812,0.4027066529,-0.1611104757,0.1850787848,0.2451403141,-0.2799305618,-0.2717313766,-0.2062995434,-0.3766919374,0.3417181969,-0.101731725,0.1029788107,-0.0001036642,-0.2463439405,0.0133144446,0.0624925233,-0.4220637083,-0.3517745435,0.2042213678,-0.0018871224,-0.1989042163,-0.1293878406,0.1683181077,-0.0654822737,0.0924509913,-0.0769087896,0.4604594111,-0.0036925988,0.1910533607,0.2854010165,-0.4029682279,0.1959451288,-0.0800039023,0.0667359978,-0.043820899,0.0918971449,0.0659917668,0.3166938126,-0.0174476504,0.1742993444,0.1360616386,-0.3998275101,-0.0907147378,-0.0763450116,-0.0710826963,-0.2661143839,0.0737977847,0.632391572,-0.069395408,0.0650852993,-0.517341733,-0.2082910389,0.0953331292,-0.3892348111,0.0115180286,-0.0865085796,0.1589214951,0.0082930885,0.0806455389,0.246793732,0.1504050493,0.0348283611,-0.0706624538,0.1887613535,-0.0210120417,-0.0120099187,-0.0475762039,-0.0762665495,-0.3279489875,0.0747205839,-0.109516494,0.0245236661,-0.3229111135,-0.1526347846,0.1233857051,-0.4519011974,0.0032077837,-0.0374149494,0.3347592354,0.441319555,-0.3616331816,0.0896280631,-0.1255673319,0.3435740769,-0.0596448183,-0.0491025187,0.453595072,-0.0634102151,0.4054988921,0.1880567521,0.0947800949,-0.0840573087,-0.3517459631,0.2669509053,-0.3329097629,-0.1914095581,-0.5132962465,0.0346818268,-0.1944807172,-0.0897806883,0.0164489429,0.1560226083,-0.2040168196,0.1376789808,-0.1604218632,0.0174841601,-0.0186054017,-0.3845176101,-0.0334108844,-0.2106874287,-0.2203496844,-0.0961611345,-0.1874143779,0.0846683681,0.2580227256,-0.0988033116,0.3942308426,-0.2225654274,0.0133917965,-0.2726571858,0.0337810293,-0.2382860035,-0.2118438035,-0.181231007,-0.2084265798,-0.1865707487,-0.0181796998,-0.0470120348,-0.1525343359,-0.289927572,0.2821365595,0.1840290427,0.0047596879,0.1546301246,-0.4161112905,0.5570933819,-0.1702479124,0.3113904893,0.3434077501,-0.1283844262,-0.1843543053,-0.1070308164,0.1684682071,0.0197692737,-0.1154248491,-0.1988828331,0.4583144188,0.1314865053,0.2196232229,0.3778914511,-0.3014671206,0.3738218248,0.1338273287,0.2919662297,0.1472800821,0.2005857378,0.0152869308,-0.0110682882,-0.0741768405,-0.2649732828,-0.013581221,-0.1358347982,0.4207935929,-0.2460363656,-0.0291044172,0.2523927689,0.0322386362,0.0370513424,-0.4075156748,-0.4062932134,0.2187842578,0.020536704,0.1168178767,-0.1498483419,0.2145441175,0.6380792856,0.1009495333,-0.4478616416,-0.1879836768,-0.3772436678,0.4720911384,-0.1045633033,0.1327014267,0.2647719383,0.3468116522,0.2244712263,-0.0971404687,0.272069633,0.2956106961,-0.2033835799,0.1709327698,0.1735499948,0.1529824287,0.1029701754,0.1602182686,0.0174833219,-0.2046836019,0.2456154823,-0.4802745879,-0.2275651842,-0.5535259247,0.0107631469,0.2309435457,-0.1112869754,0.0813828409,0.0827927217,-0.0050035091,0.357495755,-0.1463560611,-0.0156822782,-0.1896278113,0.2357197404,-0.0501512736,0.173104018,-0.2341891229,0.4826675355,0.3344714344,0.2200555205,-0.2113866657,0.1971500516,-0.0720547363,0.1580653489,0.304595083,0.6122882962,0.1674378067,0.066242516,0.1328540593,-0.4178977609,0.0257421639,0.1547708064,-0.1491043419,-0.2237517536,0.2564687729,0.2144478261,-0.1147010252,0.0463883281,-0.1337213963,0.0569404811,0.557829082,-0.1428235918,0.2295326442,0.1989189535,0.5189016461,-0.3050459921,-0.0725493133,0.0159229934,-0.137386322,0.0877700076,0.2015790641,0.1478748769,-0.0273405556,-0.2188295573,0.1252611876,0.0157827772,-0.1527891159,0.1578863561,-0.0572472923,0.2114070952,-0.2554963827,-0.1779022813,-0.0976835787,-0.0542230755,-0.0601527095,0.1648679823,-0.1169072539,0.1615856886,-0.1253595799,-0.0872274861,-0.1981623918,-0.2564946115,0.3206938505,-0.2945397198,-0.1807955503,0.2102748603,0.106284596,0.5638193488,-0.3794335425,0.0047400771,0.0382160768,0.0092616407,-0.3987799883,0.0889996737,0.004158393,-0.3560526371,0.3400898874,-0.1060509533,-0.3122848868,-0.0148054091,-0.6370803714,0.1654572934,-0.0358800516,0.0560886189,0.0953678638,-0.5258800387,-0.34286201,-0.0420559756,-0.080934085,-0.0219958797,-0.1218088865,0.1219160482,0.0355795287,-0.2740758061,-0.1570341438,-0.3323710561,0.1546514183,-0.0127200242,0.2127460539,-0.4659501016,0.1562655121,-0.1724230349,0.0156060383,0.1617344916,0.0289321616,0.2153354287,-0.1713782996,0.1383953094,0.0528838113,0.0204353072,-0.1943559945,-0.0703226849,0.0171600785,0.244169876,0.1987923086,-0.2404878139,-0.153113544,0.2164021134,-0.0376479924,0.0399203189,-0.1564818323,-0.0306669958,0.0451151244,-0.1633165479,0.1383824944,0.0089752218,0.2126473188,0.152288273,0.4314678907,0.0553292222,0.4029931426,0.5166801214,0.418168366,-0.1030318886,-0.1397066414,0.141138643,0.2512727976,-0.1293347776,-0.0605842769,0.138382256,0.3226554096,-0.1787824333,-0.213416189,0.5223706365,-0.1267057061,-0.0794635415,-0.0309164301,-0.2024281621,0.0861635059,-0.1707430482,0.2229557782,-0.0606429763,0.1329788417,-0.1167042404,-0.0526942126,0.2400886863,-0.1161271185,0.413440764,0.1377022713,-0.078609094,-0.0248051155,-0.1858156919,-0.0177487377,-0.1989762336,0.2066971809,-0.0497013442,0.6125745773,-0.0224640146,0.0416000597,-0.0182192568,-0.3017787933,-0.1325733215,-0.0910971612,0.0358232036,-0.1997811794,-0.194438979,-0.0023742544,-0.0775745586,-0.2604102194,-0.2908519208,0.3734757304,0.345025003,-0.3941194117,-0.1345342547,-0.173740387,0.1843829453,-0.1679183096,0.2255705297,-0.4885704517,-0.181758672,0.0039799451,0.1672822386,-0.266772747,-0.0977736264,-0.0358691476,0.077849783,-0.4336225986,-0.0363891572,0.2320016176,0.302026093,-0.2426457703,0.1992149353,0.0103907222,-0.133296147,0.2695195377,-0.0817988068,-0.0510806069,-0.2775348127,-0.4561480284,-0.1318356097,-0.2702732384,0.096823886,0.4356077313,-0.1423236132,0.4133484066,0.0749862865,0.1741437614,-0.0798217431,0.3434535563,0.0658073351,0.4377069175,0.0558745526,-0.1456004679,-0.026328193,-0.0390202478,-0.2419624776,-0.0920764357,0.1118782386,-0.2722546756,0.5859879851,0.6461005211,0.8278374076,0.0803755149,-0.047875192,0.2468944788,-0.0552865304,0.1243924648,-0.5560688972,0.034564469,-0.1392500997,-0.0212092102,0.0536285415,0.1082494333,0.002543072,0.1189902797,-0.1170467138,0.0566629432,0.1732517928,0.1567191929,0.2821531892,0.2916399837,-0.0329690352,-0.2444657236,-0.0386993363,0.1101935282,-0.1400664896,-0.2128028125,-0.106635727,0.1880790293,0.0331032872,0.0293956883,0.078700915,0.0888462141,-0.5987715721,-0.0650123507,-0.0856707767,0.2696971297,0.3580605388,-0.0188839119,0.8316283822,0.2832247615,0.119116962,-0.1778423488,-0.0494898073,0.1510476023,0.1721254587,-0.3283566535,0.3913442791,-0.2600150108,-0.0559406579,0.1997480243,0.4209336042,0.1908631325,-0.2129945755,-0.1195623204,0.1156532094,0.0422233082,0.2432251722,0.4383271635,0.1440800875,-0.0786429569,0.1281515211,-0.117427811,-0.233821556,0.0916192159,0.1162770689,-0.5345737934,-0.0330856293,-0.1756121069,0.0224823132,-0.2264664173,0.1342695951,-0.4359618127,-0.1769833714,-0.1010553837,-0.160472542,-0.0274049863,-0.320879817,-0.0187319033,-0.361594975,-0.391741693,-0.2454905808,0.3045246601,0.3522576094,-0.1608040184,-0.0098429695,0.115289852,0.0346931927,0.1334688514,0.272924751,0.0233790483,-0.1093133762,-0.0115124583,-0.1055809557,0.1600701064,-0.3418622315,-0.0706229061,-0.0519283935,0.2057733387,-0.4521644115,-0.0775003657,0.2171656042,0.1973765641,-0.0114864297,0.0005601675,-0.3841929734,-0.1844679564,-0.455478251,0.0981119871,0.2321031392,0.0263405927,-0.125706017,-0.0285964441,0.0790148526,-0.184972018,0.0533491746,0.032103043,-0.0278661624,0.2553341985,-0.0528879315,0.4102685452,-0.0099444613,-0.187106654,0.1511584669,0.3071472943,0.3186525404,0.2298206687,0.1377829015,-0.1820972115,-0.1341100335,0.2317277938,0.1455397904,0.0820587501,-0.1794007123,-0.1591881812,0.0763330385,-0.3153572977,-0.0199869629,0.2084306926,-0.1109768599,-0.6128420234,0.5199989676,0.1922833323,-0.0951463282,0.1334492117,-0.0495554022,0.0025591722,0.0175715871,0.4665986598,0.6622140408,0.0724929571,0.1799515635,-0.1895277798,-0.2662142515,-0.407558471,0.0489830002,-0.0582186952,-0.0481622741,0.4330044091,0.1064982116,-0.017890349,0.1348910928,-0.1890152842,-0.1548712254,0.6038553119,-0.0263801422,0.2872196734,0.0575812086,0.1438649297,0.4932638109,0.1847047061,-0.211464867,0.1327314973,0.1473629177,0.3855796754,0.3346921206,-0.2888309658,-0.1308003962,0.0546635874,-0.1389887184,-0.3133107722,-0.1855576485,-0.0357175246,-0.5018911958,0.2657041252,-0.0431043729,0.1155032367,-0.6338984966,0.2443313152,0.0073595904,0.0681277961,-0.213164106,-0.0517546758,0.5043628812,-0.1438052654,-0.0407439433,-0.0937774852,0.0617833063,0.0430583507,0.2118077725,-0.2009401023,0.1442930251,-0.2338577211,-0.0210406538,-0.0939821377,0.4442207515,0.0067855772,0.1781806499,0.237249136,0.0456485972,0.3661987185,0.0389941968,-0.2013029754,-0.1067333668,0.0866034627,-0.0056920792,-0.2740407884,0.2464192808,0.2823229432,0.3083327711,-0.1360838115,0.1157031953,-0.445237875,0.3003881276,0.3731364906,-0.108938612,0.0536351949,-0.1768936664,0.135211736,0.3026176393,0.2551554739,0.1132306829,0.1672952771,-0.0844840631,0.2983357906,-0.151146695,0.1852400005,0.4153453112,0.3259380162,-0.0650975704,0.0395726375,0.087272726,0.3574017584,-0.1685951203,0.0177781414,-0.3066978157,-0.1291559637,-0.043979533,0.0610176735,-0.1576666981,0.2799113393,-0.2105532438,-0.1177376136,-0.1154096127,-0.2146639377,0.0347236581,-0.1586821079,-0.2142564505,0.1090103164,0.1373441815,0.4826810062,0.3121455014,-0.2878768742,0.1421314925,-0.168791011,0.2134721428,0.0941193476,-0.2045440972,0.1016779616,0.1986633092,0.0027230305,0.1577430367,0.0377338082,-0.2912996113,0.311696738,0.428334713,-0.5875482559,-0.1743982583,0.0046637626,0.3589839041,-0.2686443627,-0.1424582303,0.6110182405,0.2124613076,0.1437303424,-0.2312262803,0.0382120684,0.1179813072,-0.5521712899,0.2420517057,-0.2142174691,0.2684102654,0.5262132883,0.0395510644,-0.7220555544,-0.2957613468,0.1374918073,-0.0636387467,-0.3452251554,0.2037790418,0.1262586266,-0.0357222334,-0.2306516767,0.1873379946,-0.2571241558,-0.3657433987,0.1439216137,-0.2025856823]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"Very important discussion.\r\nI am trying to understand the effects of tokenization.\r\nI wanted to ask which is a good practice.\r\nSacrebleu should be used on top of the tokenized output, or detokenized(raw) text?","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":34,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nVery important discussion.\r\nI am trying to understand the effects of tokenization.\r\nI wanted to ask which is a good practice.\r\nSacrebleu should be used on top of the tokenized output, or detokenized(raw) text?","embeddings":[-0.0534889176,-0.3386299312,-0.0087702982,-0.3742045164,-0.0085886149,-0.2285602838,0.3205294013,0.0761310235,-0.4184632897,0.2326123863,0.0009287069,0.2679917216,-0.0949827507,0.0452182107,-0.1583247781,0.1368268579,0.2894061804,-0.1308312267,0.7297717929,0.0212527793,-0.0430663414,-0.1947475672,0.001422337,0.2860234678,-0.3367886245,0.1425233632,0.2514053583,-0.2285656035,-0.0521239303,-0.292340517,-0.2806226015,0.2520862222,-0.1373460144,0.0093058804,-0.0001160272,-0.4126324654,-0.1296647489,-0.0198328681,-0.1429169029,-0.4186006486,0.1919959784,-0.1563176513,-0.0356621593,0.0456132814,0.0545156784,-0.2076709569,0.0848316252,0.1671738178,0.2774563134,-0.3143969476,0.0799156353,0.225090608,-0.6144962907,0.2736075222,-0.054839097,0.2317456603,0.0535063818,-0.0118859429,-0.0252933614,0.4090677202,0.0175442323,-0.1041721478,-0.0559770539,-0.3861667812,0.1353247911,0.06579642,-0.335260272,-0.0958414823,0.2196234167,0.2872524858,0.0224596988,0.1869004369,-0.1534449458,-0.2841993272,0.0761333331,-0.4379484355,0.1803535074,-0.0302007608,0.2803532183,-0.1332246661,-0.0537090786,0.0750859082,0.1532894224,0.1280846298,-0.102976732,0.0057184761,-0.1361026466,-0.0367109291,-0.2315266579,-0.0838124007,-0.4070293009,0.0609028302,-0.0091075264,0.0110439574,-0.0429762974,-0.0523035862,0.0025271492,-0.7345796824,-0.0853615925,0.1489671767,0.2097123712,0.3566478491,-0.1289371103,0.159062326,-0.056948971,0.1275474876,-0.0987708047,0.0199274924,0.3785921037,0.0142448433,0.2808940113,-0.0214221701,0.0007674813,0.2486759424,-0.1548784375,0.1420980096,-0.408283025,-0.3000940681,-0.6668434739,0.1036392599,-0.2481211424,0.0378221795,-0.0483978987,0.1572759002,-0.1483570337,0.0623549558,-0.1810033768,0.1936271638,0.1048920229,-0.2167397738,0.0502771772,-0.3493444622,-0.1972103715,-0.0617683642,-0.1969302297,0.2116462141,0.183921814,0.0738020614,0.4163661897,-0.2031268328,0.0313051939,-0.2946106195,0.1265487522,0.0019550612,-0.2798082829,0.0322897248,-0.2459756583,-0.1797044128,0.0388398208,0.4029069543,-0.1808570772,-0.327364713,0.2659843862,0.0760198459,0.0202188585,0.1035282612,-0.432877332,0.2859940231,0.1217677519,0.1663894504,0.2306155562,-0.0684603304,-0.0736714825,-0.3263158798,0.2303990275,-0.3518730998,-0.1408069134,-0.2563331425,0.5405233502,0.1719583273,0.3885608315,0.4680672586,-0.2848960757,0.5760139227,0.0932207182,0.4495755732,0.01896693,0.048596397,-0.2619140446,-0.0025355495,-0.1162557006,-0.0547501259,-0.0172038674,-0.0843821466,0.5590431094,-0.0869396254,0.1992362142,0.0768202171,0.1631886959,-0.2353538871,-0.3542741239,-0.397136867,0.2236181647,-0.2758860886,0.1850797087,0.096283339,-0.0378320292,0.5277570486,0.2873274982,-0.0656038076,-0.0629128218,-0.291706562,0.6088622212,0.0836235508,0.0710543394,0.1710272431,0.4758094251,0.0462910868,-0.2629953325,0.0352913477,-0.1314114183,-0.4512753487,0.0885267481,-0.1049226671,0.0758570731,-0.0942520425,-0.0322296023,0.0450401567,-0.3564468324,0.164209798,-0.5558924675,-0.2223709524,-0.4970424771,0.2961452901,0.2391628623,-0.0020704775,0.1577320695,-0.0906916857,-0.0740430057,0.1381121576,0.011025819,-0.1357451379,-0.1228665933,0.3264986873,-0.046636343,0.12500453,-0.5223798156,0.400347203,0.3916243315,0.2758909762,-0.206725806,0.2369775027,-0.1858052313,0.0750563741,0.404109329,0.3920301795,0.0217135958,-0.1045066342,-0.1196122468,-0.3385766745,-0.159617424,0.1461785734,0.0319236256,-0.3729406595,-0.1031898111,0.2717977464,-0.1802096814,-0.0249539744,-0.1335631162,0.1292451918,0.1978729218,-0.0696284845,0.012273916,0.0273110252,0.3356278837,-0.1597333252,0.1968339086,-0.0091647413,0.0043587391,-0.1094205976,0.1001755893,-0.0053064274,0.0634325743,-0.247174561,-0.0943734273,0.110315159,-0.1918456852,0.2059087604,-0.1342611015,0.3183043897,0.035325326,0.0935448855,-0.1182367653,-0.1221438348,-0.0410002731,0.0142045282,-0.1001667976,0.0179272126,-0.0039315796,0.0271524508,-0.0183164552,-0.1413636655,0.1531798095,0.019389566,-0.2313043922,0.1562374979,0.1821723878,0.5524813533,-0.3020157516,0.1339546889,0.1936060786,0.3675565422,-0.3302900493,0.0363846235,-0.2999408841,-0.2240291983,0.1970828623,-0.0278861169,-0.3574439883,0.0504803061,-0.4842433333,0.0669598281,0.1395796984,0.0275166202,0.1065052301,-0.3922307193,-0.5467971563,-0.1757953316,-0.0042574196,-0.0964547545,0.0331297778,-0.1423620731,0.1674597263,-0.0377144068,0.0826196,-0.1102223024,0.1233957708,0.1389751136,0.1535558552,-0.3033349812,0.0906351656,-0.2730766833,-0.1930274069,0.0952649489,0.1624789536,0.1974751949,-0.1028229818,-0.0607130863,-0.0131074544,-0.0805634484,-0.2119524032,-0.1738860309,0.0929426104,0.1628528684,0.3696706891,-0.1624283195,0.0751223266,0.2577899992,0.0042396588,-0.1037947088,-0.2740035057,-0.203591615,0.0812192634,-0.0519025065,0.1165928394,0.0817931369,0.2820691168,0.076432541,0.5138465762,0.2153431922,0.0580351874,0.7999769449,0.7114078999,0.0817101449,0.0359829925,-0.1352850497,0.2996606529,-0.1601791233,0.1094593331,0.0748464242,0.2245965749,-0.1328710765,-0.3674926758,0.4549716413,-0.0962663442,0.2582733035,0.2384827584,-0.208620742,0.0582688674,-0.1274103224,0.1739630252,-0.0556664132,0.1843208969,-0.0986244008,0.1526645273,0.2859437168,-0.1766542792,0.3779893219,0.2085440755,0.0653846636,-0.0473963283,-0.3087049723,0.1234575212,-0.5473151207,0.0934267566,-0.0639090538,0.5431884527,0.0296074059,-0.0022998857,0.1439680606,-0.1351860017,-0.0970320255,0.0714728311,0.041777499,-0.2586293519,-0.1854123473,0.0914046392,-0.138263464,-0.2102320194,0.0290809181,0.5118290782,0.2903952003,-0.3197952807,-0.0996137336,-0.2138155401,0.1121413633,-0.0750301778,0.0956911668,-0.3345774412,0.0132236816,-0.141260758,0.0880283266,-0.1450709552,-0.0727718025,-0.1694633812,0.0055153794,-0.2798148394,-0.2080267817,0.1647835523,-0.0230039172,-0.1223689765,-0.0823845714,0.0094243204,-0.0722908229,0.272473067,0.311545223,-0.227441296,-0.1168709695,-0.4276043773,-0.1647058725,-0.1958160698,0.1057078019,0.3983944952,-0.1843330264,0.5690668225,0.004811876,0.1230712235,-0.1870932728,0.3980107009,0.0956771523,0.2545207441,0.0360730514,-0.099510707,-0.130692631,0.0619059727,-0.3289783597,0.0217716582,0.132412672,-0.2924468517,0.6204172969,0.6178931594,0.8301172853,0.0846483335,0.1251003295,0.1030754969,0.0327746458,0.2072517127,-0.6080394387,0.2170162648,-0.193660304,-0.1063490883,0.0239130147,-0.0080045583,0.0482156128,0.2031485289,0.0128195565,-0.1044121385,-0.1504555345,0.4825916588,0.1195061877,0.1735836864,0.1527306437,-0.1248426661,0.1393548399,-0.0281603895,-0.1646536142,-0.0294523928,-0.3362514973,0.3733489215,-0.094514735,0.2675045431,0.2772245109,0.0196562503,-0.3073723614,0.0228392258,-0.0109279975,-0.0246544182,0.1751667559,0.2104125023,0.7899873853,0.4659531713,0.2458577305,-0.0486486815,0.1249255612,-0.0467612892,0.0160929784,-0.2824114859,0.3794047832,-0.1792993397,-0.1056238785,0.1692982018,0.4437360466,0.0467780419,0.0936776847,-0.2310969681,0.2133798748,-0.3391326666,0.0827495158,0.4338933825,0.2404444069,-0.0835730284,0.0179650672,0.0970852748,-0.2547543347,-0.0175267775,0.1035506725,-0.2994256616,-0.1116393432,0.0624579117,0.1543577015,-0.5081811547,0.153384164,-0.5861129165,-0.1312573701,0.056773901,0.1227254793,-0.0542148799,0.0861294419,-0.0087828413,-0.3815504909,-0.1786795259,-0.265306741,0.4815694988,0.4449788332,-0.2063005269,-0.1911075562,0.2366999984,-0.3283568621,0.1744791865,0.219710961,0.0498582572,0.07396321,0.0947087929,-0.1963036507,0.1503926218,-0.1986085773,0.0172987804,-0.0096511766,0.0963604227,-0.4073476493,-0.1041634679,0.1298940331,0.1363520324,-0.0441093929,0.1370586157,-0.1685771495,-0.0380905494,-0.2837465405,0.1469806731,0.4286892116,-0.3974762857,-0.3142196536,0.0536310337,0.0646048784,-0.0767826214,0.3690259159,0.1853734851,0.1484677494,0.1393382996,-0.1879214048,0.3939815462,-0.1106282547,-0.1560994536,0.1274504066,0.2338510603,0.3141628206,0.164848119,0.1651278734,-0.0950680301,-0.184711203,0.1635130495,0.3346990645,0.0109329307,0.0062641115,-0.4087081552,-0.0509607829,-0.4910421669,0.0622508153,0.2071519345,-0.0036128154,-0.7337759733,0.4949357808,0.1029315963,0.1583255529,0.2687261403,-0.1319207847,0.0446651541,-0.1563638896,0.3830019832,0.7513081431,0.110281311,-0.1851554662,-0.0803580284,0.0365786888,-0.4534380138,0.2319751382,-0.2504219711,-0.0887408108,0.4726869464,-0.0184531957,0.0739924535,0.1111711934,-0.1799923033,0.2386817336,0.6313253045,0.1956053674,0.1330754906,0.0964031741,0.056211628,0.2380404621,0.2916322351,-0.3061421812,0.0503402986,0.2470412999,0.3151422441,0.115483366,-0.4584792554,0.2262699753,0.1190007329,-0.351793915,-0.2572354674,-0.1431496143,-0.128045246,-0.1637375057,0.208720997,-0.006529747,0.0363614745,-0.7331231833,0.0608991571,-0.0053502815,0.1665577739,-0.0927113593,-0.0820778087,0.5893756747,-0.0017918412,-0.0726919472,-0.2931450307,0.2624198794,0.2441455722,0.1507084221,-0.0842477903,0.1998927146,-0.1149915606,-0.2833669484,-0.1429377049,0.4707008898,0.0886059999,0.3637767136,0.2629124522,-0.0730766281,0.2400520146,-0.0417576395,-0.1373883784,-0.2229875922,0.3520814478,-0.1452921033,-0.2077229321,0.1517062038,0.0927271843,-0.0673967153,0.1797779202,-0.228827104,-0.1746732891,0.113188304,0.2865765095,-0.1737413704,0.2427581251,-0.0289975964,0.0552601553,0.2686699033,-0.1070958972,0.1887899488,0.0644707903,-0.0453248993,0.1264551878,-0.0885060281,0.4814583659,0.4274739325,0.1564151049,-0.0851601586,-0.0036629841,-0.1007591933,0.132827729,-0.1662563086,0.0295942463,-0.2819787264,-0.2470763773,0.0164004844,0.180824697,-0.3414705396,0.2594128549,-0.3715002537,0.0746630058,-0.2038798034,-0.0241516717,-0.0495332181,-0.0924307704,-0.213645786,-0.1159372628,-0.0197798777,0.5663449764,0.262696296,-0.1448654234,0.15551503,-0.2822676599,0.5056078434,0.2490209192,-0.3081420362,0.1997314394,0.0465038829,0.1736227721,0.1470647454,0.3265783191,-0.2225539684,0.3439588845,0.4440348744,-0.5293257236,-0.3123128414,0.050039582,0.2352320403,-0.3383472562,0.0453185402,0.5623493195,0.1424375027,0.3567664623,-0.3271564543,0.1723992378,0.1337708831,-0.4792151749,0.2275817096,-0.2920597196,0.1575166881,0.573379755,0.0559313335,-0.743095696,-0.4670759439,0.1347171068,-0.1734416485,-0.3048739433,-0.0742864907,-0.0171921737,-0.3328159451,-0.2436640412,0.3680664599,-0.3271870911,-0.3053837121,0.1229454726,-0.1342660636]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/137","title":"Tokenized BLEU considered harmful - Discussion on community-based process","comments":"Use sacrebleu on detokenized output and raw unmodified references. ","body":"https:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. ","comment_length":9,"text":"Tokenized BLEU considered harmful - Discussion on community-based process\nhttps:\/\/github.com\/huggingface\/nlp\/blob\/7d1526dfeeb29248d832f1073192dbf03ad642da\/metrics\/bleu\/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokenizer is usually not the same as the one used by `mteval-v13a.pl`, the closest thing we have to a standard. Moreover, tokenizers are like window managers: they can be endlessly customized and nobody has quite the same options. \r\n\r\nAs @mjpost reported in https:\/\/www.aclweb.org\/anthology\/W18-6319.pdf BLEU configurations can vary by 1.8. Yet people are incorrectly putting non-comparable BLEU scores in the same table, such as Table 1 in https:\/\/arxiv.org\/abs\/2004.04902 . \r\n\r\nThere are a few use cases for tokenized BLEU like Thai. For Chinese, people seem to use character BLEU for better or worse.\r\n\r\nThe default easy option should be the one that's correct more often. And that is sacrebleu. Please don't make it easy for people to run what is usually the wrong option; it definitely shouldn't be `bleu`. \r\n\r\nAlso, I know this is inherited from TensorFlow and, paging @lmthang, they should discourage it too. \nUse sacrebleu on detokenized output and raw unmodified references. ","embeddings":[-0.1997387111,-0.3002173603,-0.0499641113,-0.3133287132,0.1553713381,-0.3138303161,0.2694100738,0.2305928171,-0.4391962588,0.4396157861,-0.1379041374,0.3319519162,-0.0998260975,0.01742048,-0.1475791186,0.1475464851,0.0647070408,0.0487196445,0.567234695,-0.0939040184,-0.1126418337,-0.2011696696,0.0313608348,0.2928083241,-0.2823537588,0.2498696446,0.3497069776,-0.2821489573,-0.1465739459,-0.2620095015,-0.2448139191,0.2275571674,-0.3441912532,0.1618193537,-0.0001048519,-0.3154124916,0.0023697056,-0.090604566,-0.3431861103,-0.4482063949,0.1331160069,-0.0905154645,-0.1050386205,-0.1324004382,0.0085874721,-0.199540779,0.1430263966,0.1090291888,0.3640665114,-0.0477227829,0.2165119648,0.3130126595,-0.5499479771,0.1918250471,0.0779710785,0.2889163494,-0.0494903475,0.1266257465,-0.0436118059,0.3868537843,0.039927043,0.0083381068,-0.0358274169,-0.2473973632,0.0469313897,-0.0220237207,-0.2477305532,-0.0883097872,0.2017854005,0.4947493076,-0.0512267239,0.1007076427,-0.2836279273,-0.2053323388,0.1010039598,-0.4695952535,0.1014689282,-0.0305460021,0.1215590015,-0.1150144488,0.0378520116,0.2168111354,0.115751572,0.0625238866,0.1224924475,0.0666300356,-0.140709281,0.0083784843,-0.1362400949,-0.0313326269,-0.3769300878,0.1300305575,-0.0052407705,0.0190333743,-0.1861538142,-0.0531506427,-0.1410668194,-0.5963930488,-0.0384503044,0.0276304651,0.3901568949,0.3056848347,-0.1996835768,0.1507122517,-0.2101283222,0.1008831784,-0.1685279459,-0.1698961556,0.4606572688,-0.002647541,0.3432299495,0.1185021251,0.2235511094,-0.0690561384,-0.1332838684,0.158619836,-0.3325148821,-0.2211337835,-0.7714271545,-0.0030731247,-0.1160404459,-0.0507523641,-0.03665727,0.2052527666,-0.3007713556,-0.0547076985,-0.0898911953,0.0525918752,-0.1307134479,-0.1758969426,-0.0611737296,-0.2430723459,-0.1962323636,-0.0817603767,-0.0746510625,0.1375359148,0.2489928305,-0.0027013747,0.5282544494,-0.1558327973,0.0226536877,-0.1340212524,0.0884089619,-0.0351488143,-0.3108409643,-0.086257197,-0.2226266265,-0.1916077435,0.0248575266,0.1450810581,-0.2518231571,-0.3038840294,0.1406217664,0.1831118315,0.0678695962,0.1400945336,-0.3797347844,0.1979618669,0.0636861548,0.1426651031,0.3059016168,-0.0725588799,-0.110048607,-0.2830224931,0.2356856912,-0.345235765,0.0187359396,-0.2711615264,0.4340251386,0.1593878269,0.2927749157,0.4076354504,-0.2251587361,0.3979407251,0.1065583974,0.2155897319,0.1441556364,-0.0226894896,-0.2381714582,-0.1252594888,-0.088570796,-0.2884486914,0.1215204671,-0.0289188903,0.4915941358,-0.1788553447,0.1038349643,0.0693510473,0.1120644063,-0.1198194101,-0.3791587055,-0.5272731781,0.1664494276,-0.0808078721,0.1480536461,-0.0401071794,0.2209029198,0.5987330079,0.2670991123,-0.3072123528,-0.089499712,-0.3125240207,0.6350923777,0.0437138081,0.1319063157,0.196200192,0.4403845072,0.1091533378,-0.1691111922,0.201659888,0.0840526149,-0.4515349865,-0.00327099,-0.0249532741,0.2254187167,0.0596736111,0.1152669117,-0.0544820502,-0.2757781744,0.2581782937,-0.4869832695,-0.2275692374,-0.5113255978,0.1554826647,0.1724786907,-0.1157297567,0.0940934047,-0.0005609007,0.0758048743,0.3380241692,-0.0105099585,-0.0063509019,-0.2178938091,0.2931793928,-0.0425518006,0.1196117476,-0.470571667,0.5725618005,0.3792096972,0.2799773216,-0.2934612036,0.178817004,-0.1218893826,0.1069269329,0.3522537649,0.5215818286,0.0691461861,-0.127759859,-0.0154760061,-0.258638978,-0.0965040773,0.100876376,-0.0299845506,-0.2897162735,0.0715295225,0.1579093486,-0.2074283659,-0.0380440019,-0.2037254572,-0.0137810325,0.2337262332,-0.0013619554,0.2114053965,0.0633663833,0.4815567732,-0.0882221311,0.0991752148,0.0603413433,0.0333880931,0.0563645884,0.0568458959,0.1072356254,-0.0726427212,-0.3915997744,-0.1073406488,-0.0179766491,-0.1033725962,0.1817964315,-0.0411617719,0.3003271222,-0.1462497264,0.0509894863,-0.022095399,-0.0839328691,-0.077956453,0.0962213799,-0.2418697327,0.0889592096,-0.0711634085,0.0687688962,0.0626971349,-0.1555515975,0.2290656269,0.0890121534,-0.2408920228,0.3154833913,0.16932033,0.6235182285,-0.1924426854,0.1023438051,0.0122120949,0.2039497793,-0.3464924693,0.1256743371,-0.2027878016,-0.3157309592,0.3201164901,-0.0844383389,-0.3003954291,-0.0844879299,-0.5661140084,0.0868472531,-0.0109746363,0.0030979845,0.2019969076,-0.4598691463,-0.3947058022,-0.1702326983,-0.037829496,-0.2211409211,-0.0408382155,-0.0694172457,-0.0093190754,-0.1503886729,-0.1534661204,-0.2378029972,0.1424928457,0.0268623605,0.1958590001,-0.2992166579,0.1475102603,-0.1910695136,-0.0993905962,0.1804561615,0.2059526891,0.237872079,-0.2010253966,0.1040793732,0.1685343087,-0.1851334721,-0.314906925,-0.0773000345,0.0693260133,0.3185837269,0.1270611584,-0.311131686,-0.1937095523,0.250523001,-0.0906230733,0.0107547063,-0.1511992514,-0.3017153442,0.0450218543,-0.1932752579,0.1248431504,0.0501810685,0.2979814708,0.1067670658,0.5134654045,0.1084602848,0.1049153283,0.699057579,0.5992892981,-0.0062749484,-0.068843171,0.0163900945,0.1837301403,-0.0461585037,0.081147112,0.1096112281,0.3403282762,-0.1393206269,-0.2915149033,0.371619314,-0.1608059257,0.1374006271,0.057022851,-0.0964822322,0.220478788,-0.17940332,0.1034840569,-0.041802343,0.0168717355,-0.004534483,0.1234496608,0.2256627679,-0.3214838207,0.395681262,0.1075667664,-0.086980097,-0.059189871,-0.1624429971,0.0494617857,-0.4588431716,0.2817772925,-0.031508144,0.5841936469,-0.0768669769,0.053021837,0.0591964275,-0.0184582826,-0.0505947992,-0.0070365123,0.0562494136,-0.2323496342,-0.1039325371,-0.1062863022,-0.0859469399,-0.2023563534,-0.0539919324,0.3796958923,0.2733723819,-0.3758244812,-0.1288552433,-0.0566113703,0.0046331217,-0.1853283793,0.0592083894,-0.4394654334,-0.1894272268,-0.117909275,0.013548661,-0.1083062738,0.0192950591,-0.2895965278,-0.015979778,-0.3817168772,-0.0156509876,0.2842929661,0.0477846898,-0.1772667319,0.0996186733,-0.0474547148,-0.0962059721,0.2600926459,0.17234914,-0.1212577,-0.3096446097,-0.3585148454,-0.0317386091,-0.3003956676,0.1136874855,0.4413614869,-0.142151922,0.5035508275,0.0786492825,0.1751225591,-0.2126282901,0.3714292347,0.0508981645,0.2740515471,0.05037608,-0.2063299417,-0.031754896,0.0488614552,-0.2408408076,-0.0067179557,0.1679429114,-0.2084169388,0.5476562381,0.7207805514,0.8223052025,0.0547992997,0.0788691714,0.2238840312,-0.0113985576,0.2791097462,-0.600566566,0.1417916566,-0.2631811798,-0.2848194838,0.0661709905,-0.0286987294,-0.0788056105,0.1271051764,-0.0043361299,0.0046942681,-0.027200222,0.308162868,0.1723925471,0.2854908109,0.0600847639,-0.186670512,0.0794385523,0.1205385104,-0.1349593997,-0.090185523,-0.1738809496,0.2068017721,-0.0843843147,0.1483253241,0.1755369455,0.0044345302,-0.401016444,0.0856088176,0.1092658639,0.1210803539,0.1736362129,0.1489587128,0.8056714535,0.4180629849,0.0905952603,-0.0741612092,-0.011288492,0.135130465,0.0738020912,-0.2522870004,0.3949091733,-0.143046543,-0.067001611,0.2998654246,0.5205954313,-0.0691307634,0.2198588103,-0.2103224397,0.0319954716,-0.1313221306,0.127882719,0.2007831335,0.1988162249,-0.0263432432,0.121653311,-0.080147922,-0.2430415004,0.0204920936,0.2827753425,-0.4851855636,-0.0674837828,0.023402676,0.1827867329,-0.3943020999,0.1625677347,-0.5352604389,-0.1035157517,-0.0365495421,-0.0386258662,-0.0643804297,-0.0816050991,0.0136980545,-0.4021200836,-0.2054730356,-0.2605602145,0.4690966308,0.4411887527,-0.1279838085,-0.0790620074,0.1002039909,-0.1438676864,0.1406670958,0.1795743704,0.1486173868,-0.0226407405,-0.0117011154,-0.2220982313,0.1500996798,-0.3390025496,-0.0162419081,-0.1009790674,0.0903964192,-0.2805962563,-0.0178253055,0.2061178535,0.1763446927,0.0023618035,0.118594937,-0.3814789057,-0.149154529,-0.3916102946,0.1014747694,0.2942251563,-0.1352667958,-0.3693255782,-0.1002373472,-0.0471406616,-0.10325104,0.2333122641,0.2027706057,0.050433971,0.2135924399,-0.1804823875,0.3417735696,0.1416638643,-0.1594304442,0.177157402,0.2600477636,0.3130069673,0.1491719782,0.1107983664,-0.0709850267,-0.0616344698,0.0850974992,0.2939630747,-0.0744316056,-0.0997399017,-0.3174083829,-0.0595466718,-0.4170998037,0.0142265921,0.2062364072,-0.1929650903,-0.6870406866,0.4475169182,0.2049552947,0.0001216259,0.1766634732,-0.0904152617,0.3070899546,-0.0331113674,0.4514761865,0.8138970137,-0.0482531153,-0.0496378355,-0.1031860784,-0.2761394382,-0.5143803358,0.1060544252,-0.2583383024,-0.0299169496,0.4381925762,0.1627823561,-0.0185552705,0.1302128285,-0.1190778911,0.1161666214,0.6584087014,0.0302349962,0.1087572724,0.0805546343,0.1907402277,0.333291769,0.1548113823,-0.3312582076,0.0717322975,0.2596799135,0.4034895003,0.1734175384,-0.3353610635,-0.0407035165,-0.056765642,-0.3370977044,-0.1863140166,-0.0467263907,-0.1045980081,-0.2463611215,0.1499691606,-0.0737761259,0.1721674502,-0.5628749132,0.1561705917,0.0715015009,0.1612634063,-0.1235200688,0.1271834075,0.5605511069,-0.0712649524,-0.1943125427,-0.24822402,0.1721220165,0.1783405691,0.2080319971,-0.0746180117,0.2382847965,-0.2188554406,-0.2753576636,-0.1047399417,0.399197042,0.0590835214,0.3146723211,0.3570926189,-0.1201423556,0.3659511209,0.0720595047,-0.2032527477,-0.204381749,0.2653701901,0.0181104112,-0.2471011728,0.1614717543,0.1667134464,0.1375555396,0.0695152134,0.0335347615,-0.3702684343,0.224678725,0.3129011691,-0.2363210469,0.2909861505,-0.1313014627,0.1243387237,0.2436964065,0.1759516001,0.0514290929,0.0872622505,-0.0298413895,0.218033731,-0.2389480472,0.4522778988,0.5037154555,0.3143425882,-0.0904747024,0.0580693148,-0.1360955834,0.1894173473,-0.1332973689,-0.0802348703,-0.2378631383,-0.1733326912,-0.092855908,0.0180563536,-0.2844088972,0.0846212208,-0.2603448033,-0.0195386577,0.0113665182,-0.3492625952,0.0673021376,-0.2630202472,-0.1569335908,-0.0015126549,0.281632483,0.559525013,0.2815879285,-0.1432114989,0.1546209306,-0.2911225259,0.3851341903,0.2123356313,-0.2266463339,0.1069469675,0.0500986539,0.0816547349,0.1689533293,0.2403203845,-0.2804475427,0.4810648859,0.4415680468,-0.6723828912,-0.222390756,0.0408890061,0.2743962109,-0.2000888884,-0.0533166789,0.4335959554,0.2095841318,0.3187227845,-0.1613584608,0.0848994553,0.2281397879,-0.5267804861,0.2124322057,-0.2560077608,0.3015178144,0.5199456811,0.0057655191,-0.6995115876,-0.3256812692,0.107206814,-0.1711836159,-0.3172636628,0.063783817,-0.0798808485,-0.1207376644,-0.2735656202,0.252845645,-0.2784125209,-0.3449241221,0.0816528276,-0.2897794545]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/133","title":"[Question] Using\/adding a local dataset","comments":"Hi @zphang,\r\n\r\nSo you can just give the local path to a dataset script file and it should work.\r\n\r\nHere is an example:\r\n- you can download one of the scripts in the `datasets` folder of the present repo (or clone the repo)\r\n- then you can load it with `load_dataset('PATH\/TO\/YOUR\/LOCAL\/SCRIPT.py')`\r\n\r\nDoes it make sense?","body":"Users may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.","comment_length":55,"text":"[Question] Using\/adding a local dataset\nUsers may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.\nHi @zphang,\r\n\r\nSo you can just give the local path to a dataset script file and it should work.\r\n\r\nHere is an example:\r\n- you can download one of the scripts in the `datasets` folder of the present repo (or clone the repo)\r\n- then you can load it with `load_dataset('PATH\/TO\/YOUR\/LOCAL\/SCRIPT.py')`\r\n\r\nDoes it make sense?","embeddings":[-0.3090123832,0.3447009623,-0.0485156663,-0.1978272647,-0.0229685567,0.0207580794,0.1884503365,0.1586086899,0.1726605445,0.2399257869,0.0062487917,0.3857524097,0.0464574918,0.2451036572,0.3998591304,0.0255310666,-0.0244459994,0.070049651,-0.1732779294,-0.1373276114,-0.3640020192,-0.2514444888,0.3551029265,0.2580877542,0.1759759635,0.0368698873,-0.0416280627,0.4159365594,-0.1249530241,-0.4263148904,0.3296694756,0.2230904102,0.2731530368,0.5622919798,-0.0001124956,0.0487476103,0.0971881226,0.091562666,-0.3861498237,-0.5210891366,-0.4382898808,-0.1007484943,0.3260928094,-0.3913037777,0.0804672316,-0.275677979,0.2364041507,-0.6570729017,0.2124788761,0.334617734,0.1648003459,0.1231686771,-0.2467631102,-0.2063307464,0.1589167416,0.4795048237,-0.1128517985,0.2791862786,0.4473838508,-0.1877632588,0.2337213159,-0.1688640118,-0.0824528709,0.1349943578,0.6063190103,0.0902402252,-0.0422737189,-0.1100494266,0.2116334736,0.2509287894,0.7525584698,-0.4354602695,-0.4185048044,-0.1471859962,-0.1568042338,-0.0201355182,-0.0830314532,0.1686895937,-0.1593190134,0.3493252397,-0.2758412063,-0.379021883,-0.3844536245,0.4024395049,0.0384801887,0.010832279,0.2473139465,0.1539154202,0.0587831475,0.0556046069,0.292114526,-0.2781497836,0.4651041627,0.2353744507,0.2413295805,0.1399903148,0.1555410177,0.030088963,0.2262865901,0.0460031852,0.0414500348,-0.0955523476,-0.3153201342,0.1158604622,0.1871736497,-0.1279511005,-0.0290783346,0.286554873,0.1502451301,-0.1007895246,0.2262394279,-0.1162564084,-0.2261371017,-0.1064446196,-0.1391726881,-0.0561173148,0.3408097029,-0.3458411694,0.167849198,-0.205803141,0.0207194872,-0.1034033746,-0.0321571082,0.1876645237,0.0221797042,-0.0610919669,-0.1767026633,0.2906900942,-0.017483335,0.0387220569,-0.0492266379,0.0404996872,-0.1510084122,0.2130775005,-0.0267855357,0.0682646632,0.3417430222,-0.2578414977,-0.0675443783,0.1835035682,0.3229322433,0.054926496,-0.0518588163,0.1016049609,0.0698422268,0.2327794433,0.0810619667,-0.2629656792,-0.3304671943,0.2878436744,0.0846007392,-0.512612462,-0.2862553,0.117992498,-0.1039315686,-0.3487535417,-0.0508175157,-0.1008580253,-0.0907144323,-0.1254523396,-0.3346060812,0.0529804528,-0.2649670243,-0.2240497321,0.2608115077,0.4978345037,-0.6259242296,-0.0063599949,-0.2309514731,-0.1024579108,0.0164717808,0.0993955135,-0.5138141513,0.4445572495,-0.0029953418,0.0540964566,0.4258202016,-0.55223912,-0.2102416903,0.3178724945,-0.3017854691,-0.1860096008,0.140557155,0.2060299963,0.048614312,-0.0618532263,0.0390185602,0.4804842472,-0.1015206277,0.1506304294,-0.1494632214,-0.4203814268,-0.083973147,0.1870279312,0.1132033616,0.2471871376,0.3971060514,-0.1074997783,0.1809024066,-0.3337996006,0.0671204701,-0.1638286114,0.566373229,0.0118478071,-0.2202934921,0.1150465161,-0.467292726,0.3003493845,-0.1240665317,-0.0614873096,-0.1178948656,-0.3920164704,-0.5102698207,0.0492816567,-0.1401500553,0.1149683595,0.0089603104,0.2614189088,0.1601195335,-0.0975656658,-0.4219164848,-0.1015219837,-0.1182910055,0.1643473655,-0.041011896,0.2645519674,0.0092755565,0.2183258086,-0.1034366637,0.0791499466,0.0184264518,-0.10625723,0.0431232154,0.4312104881,-0.1817670465,0.3442499042,0.3935149014,0.4217561483,0.1872560978,-0.0497749075,0.303897202,-0.0257912911,0.2131172866,0.1154727861,-0.4095541239,0.0820389912,-0.1498382837,0.3559699655,0.077401787,0.0688540936,0.1919163615,-0.1492986381,-0.2090765685,-0.0831065625,-0.1376328021,0.1600797623,0.306304723,0.0021112622,-0.2828633487,0.2838191986,-0.2052478492,-0.1391828805,0.1246972978,-0.0414204374,-0.1701616496,0.1188541725,0.1547646075,0.6942638159,0.2962358594,0.1909158379,0.2645978034,0.2782218754,-0.1092056707,-0.156890735,-0.1746372283,0.009422401,-0.1520839334,-0.0372975692,-0.1935871094,-0.1481100768,0.1039798111,0.0082543362,-0.0298509616,0.2094391435,-0.2639261186,-0.0796205029,-0.4523699284,-0.3008153439,-0.4785174429,-0.3388303816,0.1356676668,-0.294691354,0.0774368048,0.0399066582,-0.2128749043,0.0000551179,0.0094636651,0.3449990451,-0.4868342876,-0.4742683172,0.1611764878,-0.2804466784,-0.0841309726,0.0790378526,0.5324950218,0.2608543038,0.40120399,-0.3653788865,0.0251296554,-0.4508334994,0.1964248717,0.0116490582,0.2860533595,0.3462802172,-0.0934295431,0.1495042741,0.19244048,-0.1640766114,0.1141793802,0.0246386826,0.2139932662,-0.4466781914,-0.0829798952,-0.0619533695,-0.0256237648,-0.3897651434,-0.4546720088,-0.2127760351,0.2069655806,0.3351814449,0.2323654145,-0.1442607939,0.264713347,0.3041373193,0.0645773634,-0.0183743499,0.1228377521,-0.4090555906,0.2555212677,-0.3162050247,-0.287995249,0.1857584417,-0.057696227,0.090583995,0.1595321596,-0.3731519878,-0.3830951154,-0.0730348602,0.4951599836,0.2606619,0.2412657142,0.3888583779,-0.0440204777,-0.0119954329,0.0503383987,-0.3888966739,0.2352429628,0.0624073818,0.0916620567,0.263176024,0.2989981174,-0.0743815675,0.8578977585,-0.1455602944,-0.1107754335,0.1480316818,-0.3166589141,0.2758001089,-0.1176741496,-0.0122438986,-0.1448895484,-0.0289503355,-0.3250271082,0.1221099794,0.203612417,-0.089010641,-0.1741441935,-0.085378252,-0.4541608691,-0.2337510437,0.0254911017,-0.0534089953,0.1687844843,0.2771443725,-0.1507515758,-0.3245590925,0.1861573011,-0.003603278,0.2770047188,0.1198659241,0.1058981717,-0.5349532962,0.162843734,-0.5217799544,0.2875427604,-0.044600755,-0.0548369586,-0.0213947706,-0.1630278379,0.1599017829,-0.1261882186,0.0439266823,-0.1828114539,-0.1097598597,-0.0068712505,-0.1080960631,-0.3330803514,0.1410881281,0.283320576,0.264331907,-0.245342806,0.2438479662,-0.0562427901,0.044547677,0.2595143318,0.311139971,-0.1650022268,-0.0057193129,0.2337888181,-0.4701072276,-0.2506866455,-0.1330276281,-0.3484069109,-0.044062037,0.3391861916,0.1561364233,-0.0916157514,0.1820578724,-0.0224278867,0.0625032634,0.1596190035,0.1343283355,0.1258118898,0.2432863563,0.1354044378,0.3282774687,0.5877701044,-0.3588279486,-0.2533798814,-0.1741580665,-0.1767531633,0.2243775278,0.3292057514,0.12520051,-0.133790791,0.3102718592,-0.2872840464,-0.1254223883,0.2608675659,-0.1050272658,-0.1434653401,0.0244967602,-0.8044292331,0.28952685,0.1553311199,-0.243293494,0.2668421566,0.0078174416,-0.1743748337,0.2095232755,0.162942186,0.7579632401,-0.4437439442,0.1145592928,0.1241101772,-0.4018707275,0.0330545306,-0.2425469011,-0.3431080878,-0.306245029,0.0938952863,-0.1047331691,-0.1054000184,0.1476497054,0.2390692383,0.1314321756,0.2230501622,-0.0568470359,0.3353710473,0.0288255699,0.3337692916,0.1361625195,-0.2415827364,0.0377577357,0.1356452405,0.0677470043,0.0438893139,-0.0665081292,0.0850727037,-0.2782536745,-0.0189432334,0.0477679931,-0.1156926006,-0.1494702548,-0.1841482669,-0.2691521049,-0.2467095554,-0.0767551064,0.4303747714,0.4167241454,0.0722389668,-0.3530828655,-0.1185159236,-0.1004868001,0.0056711286,-0.0354961269,-0.0554677024,-0.0404178314,-0.1505733877,-0.1983326077,0.2179647684,-0.1036985666,-0.3711073101,0.0277004167,-0.1833144873,0.5654058456,-0.0788399652,-0.3157847226,0.3617791235,0.2291212231,-0.059753038,0.0932733417,0.2492666543,0.0130075915,0.0274539348,-0.1529621631,-0.165779382,-0.1468960643,0.009503237,0.0281961877,-0.1100358516,0.3025347292,-0.12020123,0.1202395335,-0.1613371521,0.2237156779,0.2726371586,-0.1207309961,-0.348100245,-0.0674351379,0.0865759775,-0.2907437086,0.1579110026,0.3332560956,0.0087620029,-0.1873578727,-0.1596955955,-0.3220474422,0.0357323736,0.1746580601,0.3953210711,0.1721738428,-0.1684340537,0.0949205682,-0.1443105489,-0.277505666,0.0315797329,0.0918971673,-0.008668609,0.4190226197,0.3323517144,0.1582546681,-0.256983012,0.0122306198,-0.3626359105,-0.2923771739,-0.1488960087,-0.0217148401,0.1732617915,0.0727509707,-0.1938297153,-0.0358959548,-0.3776186109,0.0853224173,-0.0042174147,-0.014367979,0.3106650114,-0.1886311173,0.0378824472,0.3566713035,-0.0390641615,-0.1609387845,0.3279899955,0.1096002236,0.1133506894,-0.0317191035,0.3117019236,-0.232244283,-0.2384360284,0.2176868767,0.0534631722,0.145015344,0.359978497,0.0809576362,-0.178669259,0.252915889,0.0985219702,0.3109072149,0.1625376642,-0.1273652762,0.237199232,0.5944063067,0.1551053375,-0.0100596547,0.0452671424,0.4078793526,-0.0945123956,0.1689719409,0.3568082154,0.3670763075,0.0080307983,-0.0850092396,-0.1580796391,0.2763581574,-0.0155335562,0.1728954613,0.4302865565,-0.1704941392,0.0022108338,-0.2089142799,0.5364076495,0.1826850921,0.1011999622,-0.3112928867,0.2255074531,0.2389800251,0.044595547,-0.0218509957,-0.3141853511,0.4493926466,-0.1371200234,0.069722496,-0.2145035267,0.1119652987,0.3023663163,-0.0259370729,0.0698579922,-0.0393655933,0.225037992,0.3174158633,-0.3979130983,-0.5131573677,-0.2063063532,0.09434513,-0.1360853761,-0.1871704012,-0.3797954619,-0.0526379123,-0.3703563213,0.1468514651,-0.0262237471,0.3386091888,-0.0082071982,-0.3097346127,0.0072235921,-0.0132661555,-0.1417602897,0.1345342845,0.0774541423,0.2011660486,0.03180122,0.2910661399,0.06561777,0.1547477245,-0.0896606222,0.1127415895,-0.0500756763,-0.3120886683,-0.2572543621,0.0974961519,0.5104216337,0.1776400506,-0.1427344084,0.2959182858,-0.2829830945,0.2022327185,-0.1543823928,0.0820171237,0.2602854073,-0.2395662963,0.1110521108,0.1815539598,-0.1238755137,-0.1938827336,0.5840380788,-0.1556454301,0.1547323763,-0.1188833043,0.0732828304,0.0356058143,0.3977754116,0.2855275571,0.345795244,-0.2175602466,-0.1462509334,-0.6215121746,0.0547005162,-0.326185286,-0.3606292605,-0.089747183,0.0179488454,0.2037896067,0.4190505445,0.0250480492,0.0577828027,-0.3072550893,0.2750271261,0.1199231818,-0.3218171,-0.1491023898,0.1752468199,0.0802077204,-0.2508783638,-0.0228299834,-0.0424541757,0.0099298032,0.0608876236,-0.0181423649,0.2218649536,-0.2996989489,0.318922013,-0.1086551473,0.1874166727,0.1608805358,0.1467810124,-0.1272587031,0.0153376572,-0.4648637176,0.0709749982,-0.3032184243,0.0352576822,-0.4548551738,0.0857747346,-0.2342859656,-0.1904199421,0.0644101277,0.0428802259,0.1759908646,-0.0478285998,-0.0736400336,-0.1847977042,0.2476388812,0.0585627444,0.0441952497,0.2459204346,-0.0293015689,0.0700095966,0.3953747749,-0.1033558697,0.0685854703,0.1040190905,0.0215658527,-0.1921858788,-0.0868418068,-0.1821876317,0.1563757211,0.3387641311,-0.0737505108,0.1943217218,-0.0017367044,-0.1784482598,0.0873631909,-0.1296693981,0.1940760612,-0.2205123007,-0.0282898638,-0.4687397182,-0.4691467285]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/133","title":"[Question] Using\/adding a local dataset","comments":"Could you give a more concrete example, please? \r\n\r\nI looked up wikitext dataset script from the repo. Should I just overwrite the `data_file` on line 98 to point to the local dataset directory? Would it work for different configurations of wikitext (wikitext2, wikitext103 etc.)?\r\n\r\nOr maybe we can use DownloadManager to specify local dataset location? In that case, where do we use DownloadManager instance?\r\n\r\nThanks","body":"Users may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.","comment_length":65,"text":"[Question] Using\/adding a local dataset\nUsers may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.\nCould you give a more concrete example, please? \r\n\r\nI looked up wikitext dataset script from the repo. Should I just overwrite the `data_file` on line 98 to point to the local dataset directory? Would it work for different configurations of wikitext (wikitext2, wikitext103 etc.)?\r\n\r\nOr maybe we can use DownloadManager to specify local dataset location? In that case, where do we use DownloadManager instance?\r\n\r\nThanks","embeddings":[-0.201106742,0.1899249405,-0.003256344,-0.224240467,-0.1611983478,0.0761197284,0.1614465714,0.0639025941,0.3210599422,0.2902062237,0.1616271585,0.3396971226,0.0044046212,0.17841582,0.3171544075,0.0068329861,0.0992788449,0.0957340449,-0.3213845789,-0.0826752335,-0.4165589511,-0.2904864848,0.1600189358,0.251401633,-0.023760464,0.0018262565,-0.033562541,0.2927275002,-0.1446347088,-0.429789573,0.248342976,0.0722853988,0.2206123918,0.4754655957,-0.0001182133,-0.0005177689,0.1042939499,0.1279481351,-0.3687489629,-0.3679015636,-0.5386099815,-0.1528670937,0.3830012679,-0.3592471182,0.1713761836,-0.2725833654,0.3547175229,-0.5175001025,0.1106041372,0.3297718763,0.0845135003,0.01644684,-0.3243149519,-0.0814011768,-0.0194652434,0.3938502073,-0.1172908619,0.092780441,0.5590108037,-0.151700139,0.1013459563,-0.1845551729,0.0188766383,0.0987392515,0.73269701,0.1822927743,-0.133181721,-0.1888921559,0.265627265,0.2752160728,0.978749454,-0.4884048104,-0.3078885674,-0.1967763007,-0.1440268159,0.17021586,0.1675825864,0.0440687016,-0.2567725182,0.3383143842,-0.2870934606,-0.3765918612,-0.3952593207,0.3760219514,-0.0366000086,0.1612858921,0.2452769428,0.1619727463,-0.0502662919,0.100068979,0.376927793,-0.4633914232,0.533257544,0.2898092866,0.2012470514,0.0762998387,0.0185708217,0.2050853223,0.1075656563,0.0456787981,0.093530573,-0.115116179,-0.2197816968,0.1503829807,0.1955340952,-0.1072269902,0.0580995679,0.0583123267,-0.0129448092,-0.1241884157,0.2272546291,-0.2052033693,-0.0919414535,-0.1886608005,-0.0670248047,-0.1223851517,0.1509943902,-0.1734285653,0.1557186991,-0.1623838097,0.1865758598,-0.044490926,0.0042570811,0.0590122454,0.0116429953,-0.0666119158,-0.2304314077,0.2163502872,-0.10072992,0.1422902346,-0.0023032848,0.10835886,-0.1681965142,0.2011038661,-0.014153772,0.1593428403,0.3369290233,-0.107068643,-0.2198595256,0.2733101249,0.3086557388,0.0411727354,-0.058824297,-0.0002561307,0.00132361,0.4736920595,-0.0012951762,-0.21119003,-0.3106511235,0.2809778452,-0.0066266754,-0.4388183653,-0.2618158758,0.0476111062,-0.0542419292,-0.3454589546,0.0852806643,-0.0601730831,-0.1355484575,-0.2552620471,-0.247264728,0.0612340458,-0.3180578053,-0.2794026136,0.1472722292,0.4335565269,-0.6197720766,0.1046252996,-0.3968454897,-0.0975053906,-0.0415219069,0.0253652614,-0.5023852587,0.644315064,-0.0937030166,0.1674820185,0.5209028125,-0.5649005175,-0.359269768,0.3398555219,-0.4539186954,-0.1321177036,0.0977254957,0.2600652575,0.0936468244,-0.109752059,-0.0989375859,0.4697247744,-0.112903282,0.2188141644,-0.1347548366,-0.4183905423,-0.0843808427,-0.0157877374,0.0025854514,0.2211381197,0.3674027622,-0.0033046179,0.1627334505,-0.2260329276,0.2156279534,-0.1381243616,0.3959729075,-0.0033242169,-0.4116642773,0.05300951,-0.5292719007,0.2987862527,-0.0821761191,-0.2978417575,-0.061711669,-0.2491993606,-0.4075194895,-0.0099819209,-0.1845238209,0.1506024748,0.0044379039,0.2695951462,0.2963570356,0.0519347936,-0.4162099957,-0.133428067,-0.0967299938,0.1747345775,0.0542285517,0.115951024,0.1357028633,0.2156604379,-0.2228408903,-0.1024402231,-0.0811592117,-0.0526875108,-0.0750327036,0.4124063253,-0.1210081875,0.3154419959,0.3610307872,0.3907224834,0.0826732293,-0.0380303599,0.3107642233,-0.0806855708,0.2622424364,-0.0040192534,-0.4382065535,-0.08983206,-0.2542896569,0.3789340854,-0.0098817246,0.0602177344,0.0066504758,-0.1252065301,0.0367149115,-0.1321915537,-0.2798055708,0.1747564971,0.368437469,0.0706078857,-0.038574338,0.2817542851,-0.0782720298,-0.0912842751,0.1970616132,0.0334147401,-0.1778915823,0.0533279665,0.1861423999,0.6401782632,0.3523538113,0.0190552641,0.3730781078,0.2265361547,-0.0659944788,-0.1597479135,-0.0807786435,0.0429389887,-0.2280526906,0.0812424123,-0.3045521677,-0.1317316592,0.2525743842,0.1863312721,0.0300050098,0.1881730855,-0.1252739131,-0.1555030942,-0.5117800236,-0.1336259395,-0.4786206782,-0.1955774873,0.1407467425,-0.3628897369,0.0657002181,0.161475718,-0.1322312951,-0.1513798088,-0.0597913526,0.3881176412,-0.4753079116,-0.3518950045,0.1394779831,-0.1854702234,-0.2374277711,0.0315733068,0.6115478873,0.4463932514,0.33825019,-0.2882735729,0.1297725886,-0.4036678076,0.2660573423,0.0261176154,0.2925158739,0.3645951152,-0.075043954,0.2648330927,-0.04500857,-0.2584792972,0.0405011736,0.0190757979,0.2235125601,-0.4360733628,0.0972263515,0.0809806734,0.1280534416,-0.277589649,-0.4195917547,-0.0940385833,0.1597174108,0.0941893235,0.1976951361,-0.3276100457,0.2694450319,0.2887520194,-0.0299179982,0.0139780194,0.177430138,-0.5027784109,0.143203795,-0.3212492168,-0.3227063417,0.1215478629,-0.0575171374,0.0851605013,0.1391380578,-0.3284092247,-0.3233733773,-0.1079703346,0.4302555025,0.2456706315,0.2824125886,0.4292845726,-0.0518874153,0.0977899283,0.01058493,-0.1871977895,0.2799861133,0.0446822867,0.122363016,0.4100353122,0.1412941962,-0.021201428,0.8346470594,-0.0451971404,-0.0149274031,0.172093451,-0.1510705501,0.1046918929,-0.1567704529,0.0099297296,0.028863512,0.164471224,-0.1726556718,0.2619543374,0.1630268246,-0.0238371864,-0.1079985201,-0.0019458395,-0.6516395211,-0.281501621,0.1576974541,0.0779751986,0.1179106534,0.1506599784,-0.1593615115,-0.3685023487,0.0843603536,0.0596061274,0.3265198171,0.0653825924,0.2057809979,-0.5275418758,-0.0217870884,-0.4887795448,0.1494187564,0.1151234657,0.0102096265,-0.0593723021,-0.1659352183,0.2630721629,0.0466136597,0.0171362106,-0.3798713088,-0.2030287981,-0.1739846915,-0.0245621875,-0.3165881038,0.1540765315,0.1968533844,0.3650214672,-0.3394563794,-0.0869276449,0.0256624669,-0.0603919178,0.0684168711,0.330776453,-0.1495114118,0.0022934345,0.3297540545,-0.3381733,-0.0906088501,-0.1742623448,-0.3632776439,-0.0439927839,0.2933518291,0.1321927607,-0.1658266783,0.2575239241,-0.0628170297,-0.0434783138,0.061550729,0.1208431199,0.0758472085,0.1748593897,0.258961767,0.3091345727,0.4725776911,-0.2592527866,-0.2494719177,-0.1277246922,-0.1911685318,0.2383447289,0.3395708799,0.086214669,-0.0512472093,0.2160194516,-0.2398550957,-0.1491377205,0.4135980606,0.0208387021,-0.121659942,-0.052811645,-0.5506029129,0.412348181,0.1758911908,-0.267665714,0.4514496922,0.1657428294,-0.2359360605,0.0756049156,0.2379730046,0.6936482191,-0.3097140193,0.0802344903,0.02729325,-0.330005914,0.1457707137,-0.2029700875,-0.2584664524,-0.2801539898,0.0056253076,-0.1864778847,-0.0664820373,0.20956783,0.3264792562,0.1327617764,0.1352020502,-0.1065572873,0.4445719421,-0.0459575541,0.3641988635,-0.0763155073,-0.1981815696,0.2768371999,0.0686809197,0.1540248096,0.0718904659,-0.1086257249,0.1333896816,-0.350099951,0.0650733635,0.0456935056,-0.2884739637,-0.2221194357,-0.1750264019,-0.3311592638,-0.2256614119,0.0866429731,0.5593302846,0.3667827547,-0.0627468526,-0.3632653058,-0.1484799832,-0.2045289427,0.0321372263,-0.0109100034,-0.1073164195,-0.0927572995,-0.1285810173,-0.2543644607,0.2171223015,-0.0968962312,-0.1919464022,0.1000003442,-0.1579968184,0.5252968073,-0.1653541923,-0.5256086588,0.1742151976,0.190550521,0.0368741304,0.0375824273,0.0699380487,0.1399573833,0.1634882987,-0.1984801143,-0.0739290565,-0.1062669456,0.0556706786,-0.0819461346,-0.2178378701,0.3355741203,-0.2186214924,0.1150698811,-0.1149398163,0.3450754881,0.2544180155,-0.0306781176,-0.4708593488,-0.0176131502,0.1544541717,-0.2405828983,0.2427038401,0.3589853346,0.0093313269,-0.2977697551,-0.0323534086,-0.2179199159,0.1263128817,0.2344694883,0.3669865429,-0.0639409423,-0.182125181,0.2110611349,-0.1015860736,-0.1906231046,0.0003964903,0.2516076267,0.0104195243,0.3795425892,0.4396514893,0.0391277224,-0.1012218297,-0.0470709503,-0.2827145755,-0.2453538626,-0.0945762396,0.0638762489,0.204818815,0.181913659,-0.27643314,0.0072122142,-0.3616904616,0.1659701765,-0.0349282101,-0.0330117121,0.1484006345,-0.2603638172,0.1269669831,0.4068417549,-0.056566488,-0.2221179456,0.3254331946,0.0639169514,0.1191478446,-0.106442377,0.240022555,-0.0784545466,-0.2176354825,0.2401969582,0.120996736,0.15088588,0.2979306579,-0.0415278338,-0.1601878554,0.2301127613,0.1253978163,0.1095735654,0.1534783542,-0.0760620758,0.0315713398,0.6513589621,0.0897535831,0.2572369277,-0.1281398684,0.4234576821,-0.0802422836,0.1287795007,0.2507908046,0.3465698361,-0.1704728007,-0.2700290978,-0.2467248887,0.1901026964,0.0330273323,0.0979629308,0.5943115354,-0.1580969095,0.0035581889,-0.2296649069,0.5587201715,0.1998936087,0.1838400513,-0.2704486549,0.2301089764,0.0386982746,0.1439985633,-0.0744861364,-0.4671472609,0.5340021253,-0.1560768783,-0.0725030228,-0.1407627761,0.1164827421,0.1071745977,-0.0234851968,-0.0369571559,-0.1162528023,0.1878910661,0.2257583439,-0.3190586865,-0.5075492263,-0.2590656877,0.2057570219,-0.0734732226,-0.1691740155,-0.4071730971,0.0426339358,-0.3005787432,0.1249659285,0.0526028164,0.2633955479,0.0351422988,-0.1760602146,0.0864574537,-0.0337918252,-0.1701626778,0.1610926986,0.030221751,0.1943311989,-0.1201313138,0.3845150173,-0.0404019728,0.0399163887,-0.1322108805,0.1113831848,-0.0112475911,-0.23578839,-0.3786452413,-0.0522416495,0.4906333983,0.1209496483,-0.0634204969,0.0866601244,-0.225922063,0.0904838815,-0.1498255283,-0.1556241065,0.2959861457,-0.1465162784,0.2571232319,0.2059735805,-0.212365061,-0.2468205243,0.4340804219,-0.1987023503,0.1876056939,-0.090591602,0.0302089751,0.0852276981,0.4398606718,0.2210974544,0.2189484686,-0.2097180188,-0.3358183503,-0.5563833117,-0.0022010186,-0.2462023348,-0.3601617217,-0.0721914992,-0.0268875044,-0.0603751987,0.3827115297,0.0915344879,0.3271607161,-0.3082077503,0.5737892985,0.1382458508,-0.1990304142,-0.0954757258,0.1498658806,-0.0002131665,-0.2718093097,0.0871363655,-0.0924968272,-0.0473982766,0.0550046973,0.1706447899,0.1005482674,-0.4378389418,0.3770071566,-0.0032647334,0.1779544502,0.1080736816,0.1435927302,-0.0703194216,0.2697401643,-0.5151308775,0.1531899273,-0.3629555106,0.0959678367,-0.4167467356,0.1517499834,-0.1196899042,-0.2285772413,0.1843283325,0.2477381825,0.1524385661,-0.2565478683,-0.0983575508,-0.0676255226,0.5636066198,0.1586187184,0.0167138577,0.2635459006,-0.0313635096,0.0390514359,0.2246009856,-0.0727228895,0.0968283638,0.0811699778,0.1358646154,-0.2718605101,-0.0182824545,-0.131401971,0.0559088401,0.3561341763,-0.1358777583,0.3260549605,0.0356968082,-0.1461341232,0.0269031599,-0.0898805633,0.1970147043,-0.2921511233,0.0877321288,-0.4093596339,-0.4107897282]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/133","title":"[Question] Using\/adding a local dataset","comments":"Hi @MaveriQ , although what I am doing is to commit a new dataset, but I think looking at imdb script might help.\r\nYou may want to use `dl_manager.download_custom`, give it a url(arbitrary string), a custom_download(arbitrary function) and return a path, and finally use _get sample to fetch a sample.","body":"Users may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.","comment_length":50,"text":"[Question] Using\/adding a local dataset\nUsers may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.\nHi @MaveriQ , although what I am doing is to commit a new dataset, but I think looking at imdb script might help.\r\nYou may want to use `dl_manager.download_custom`, give it a url(arbitrary string), a custom_download(arbitrary function) and return a path, and finally use _get sample to fetch a sample.","embeddings":[-0.3848367631,0.426192522,-0.0582928807,-0.062286865,-0.1617802233,-0.0290353093,0.1196396947,0.0627896935,0.225906685,0.2672970891,-0.1065552235,0.3037168384,-0.1359498352,0.2354285568,0.3147620559,0.0563904569,-0.1727443933,0.0608874746,-0.2876501977,-0.1493894309,-0.4367071986,-0.1395202726,0.3666732311,0.0160693154,0.0315403603,-0.0743042752,-0.0310812704,0.4312176406,-0.1881704032,-0.4960770309,0.1156177148,0.0801915526,0.3331568539,0.4255272746,-0.0001110298,-0.1615571827,0.1255319417,0.0057520876,-0.3756665885,-0.2831614316,-0.4848709106,-0.0562204309,0.323540777,-0.3018951416,0.167841509,-0.1485852748,0.1979242265,-0.6253156066,0.1218306199,0.2712398767,0.2022996396,0.0922849402,-0.43065539,-0.3477675915,0.1746295989,0.4100430906,0.0017399631,0.3648223281,0.5526695848,-0.1044886857,0.041821409,-0.1705145836,-0.1301363111,0.090017952,0.5631837845,0.1887885332,-0.0688499361,-0.2714115381,0.2703507245,0.3865032494,0.8187440038,-0.3395529687,-0.2601167262,-0.1455864012,-0.0715409219,-0.0685554668,-0.0204075594,0.1074622124,-0.2264563143,0.269659549,-0.1549880505,-0.2469827384,-0.304423511,0.2749157846,-0.0413943343,-0.0181669779,0.2146755755,0.0636865571,0.1867269427,0.0977482572,0.3312122822,-0.2048025727,0.3516189456,0.1316544563,0.2721080482,0.0842457563,0.0519595481,0.0258223396,0.1119041592,0.1751177609,0.1289793253,-0.1803898811,-0.2821822762,0.0377658159,0.0307112504,-0.0086092046,-0.0257195737,0.1011313871,0.2328979969,-0.0870419145,0.161633715,-0.0817736015,-0.1669723094,-0.0070390333,-0.0731824115,-0.0299173594,0.0432662331,-0.1567643732,0.0560780875,-0.083348088,0.1362375617,-0.0285525918,-0.0489856414,0.3104062974,0.0150913829,-0.2721305192,-0.0798953325,0.2013561279,0.1322698295,-0.0393841267,-0.0412744544,0.0199241992,-0.1046452373,0.2363117486,0.147088021,0.1627573073,0.2879635394,-0.2300513238,-0.1614208817,0.2307295501,0.4086382985,-0.0029009632,0.1960332394,0.064433001,0.2319631577,0.3098809719,-0.0775888711,-0.3319489062,-0.2738490701,0.1951435953,0.3463362753,-0.477000922,-0.4734295905,0.1855375469,-0.1103097498,-0.4235783517,0.0389089249,0.0183148254,-0.2762017548,-0.089792572,-0.2889486551,0.2255990356,-0.1902697235,-0.2724776864,0.1481330842,0.6925089955,-0.7137827873,0.1149784029,-0.3483240008,-0.0896897987,0.0683882684,0.1120432988,-0.3913522661,0.3761354089,-0.1457839012,0.2146044075,0.4088515937,-0.6111477017,-0.1721384227,0.3494308889,-0.4038502872,-0.1963203251,0.2511584461,0.3222974539,0.4519783556,0.066643618,-0.0624450967,0.3906957507,-0.0220193658,0.1316788346,-0.1510139406,-0.305136174,-0.0586398244,0.0699212626,0.2266173661,0.2623304129,0.4574160576,0.0484189354,0.1695820093,-0.1636099368,0.1698448062,-0.144837454,0.4279435277,0.0851072595,-0.3243453503,0.0992945433,-0.4985300601,0.2049115449,-0.1063704491,0.0182877183,0.0512507632,-0.3985586762,-0.4555648267,0.0084442375,-0.0241559129,0.0394430272,0.0740873367,0.4098875523,0.111603789,-0.0808666795,-0.3878562748,-0.1711488217,-0.0823819265,0.0974678844,-0.0757940486,0.134936586,-0.0323764905,0.0282301623,-0.1085282415,-0.0606007166,0.0247230791,-0.1131200939,0.031650383,0.3478105366,-0.1357044429,0.2376403511,0.4116074741,0.5253426433,0.2657344043,0.0615926273,0.4397469163,-0.1330983788,0.2462782711,0.037404716,-0.3912294507,0.0947273076,-0.1505209953,0.2330704778,0.1097806469,0.1505803466,0.2660111189,-0.2243696302,-0.037805669,-0.3007338047,-0.1880720407,0.1886543036,0.2170049101,0.1190810204,-0.3386602998,0.30360502,-0.2050985545,-0.1305982172,0.2531043291,-0.0729730576,-0.1957225502,0.1222395748,0.265190959,0.719281733,0.257367909,0.2213434875,0.4219477773,0.1621901691,0.0460409932,-0.2819206119,-0.0837263688,-0.137734741,-0.2034482658,0.2263743281,-0.1388097107,-0.1602101028,0.1037717015,0.0657160878,0.0637394339,0.1325288117,-0.2115382999,-0.2972091734,-0.5328249931,-0.4521932006,-0.471578002,-0.2514738739,0.1549419165,-0.2584532201,0.0688430518,0.1870961487,-0.2059519291,-0.0985751599,0.0135416742,0.3428137302,-0.5756483078,-0.3648793399,0.1012740433,-0.1248009875,-0.0904797688,0.1322442591,0.5557697415,0.2368973345,0.4106332958,-0.2774346471,0.1927794218,-0.5172348619,0.2275174409,0.1171501726,0.2604715526,0.30322963,0.0053092577,0.0604221113,0.230099991,-0.3323422372,-0.1318967938,0.0288817026,0.1702225208,-0.4290680885,-0.2287544757,-0.0765331984,0.0283667129,-0.3620597422,-0.386800766,-0.1854599714,0.108247824,0.3310581446,0.2041646689,-0.2248843163,0.1045309231,0.2954856455,0.1064800695,-0.0391984172,0.0378971286,-0.295904547,0.2268674225,-0.5191518664,-0.3212379217,0.2769160867,-0.1316582114,0.1797231436,-0.084203355,-0.3934152424,-0.4176706076,0.0114080608,0.4556054175,0.1660173833,0.1292218566,0.3921355605,-0.0643505752,-0.0798713714,0.0666899383,-0.3398392797,0.3800733089,0.1752509624,0.1266064048,0.1159877554,0.2497957796,-0.0579179563,0.6440163255,-0.1830210686,-0.0704624951,0.2175628245,-0.3027667999,0.3039292097,-0.1950046569,0.0160667337,-0.1849907488,0.0593994446,-0.3080708385,0.1053065062,0.1063763276,-0.2208043188,-0.1077624634,0.0710956156,-0.5568802357,-0.218927443,0.1655976623,-0.0701704174,-0.0457322039,0.1608373523,-0.1061761975,-0.2573771775,0.0402342826,0.017962005,0.2764771581,0.114589572,0.0426733419,-0.4099330902,0.2472340167,-0.4953665137,0.3042174578,0.0162102003,-0.0775231645,-0.1177611277,-0.2123426199,0.2518219948,-0.0044415556,0.1961312592,-0.3412710726,-0.120858334,-0.0693047047,-0.1481302679,-0.2724314034,0.1889242679,0.2091514617,0.0962155983,-0.0688104928,-0.0275923144,0.077221185,0.120781824,0.2681333125,0.1969593912,-0.1796223819,0.0488439277,0.2749687433,-0.4902130067,-0.3052376509,-0.1285990924,-0.3804083169,-0.1004881188,0.3143920302,-0.0193944145,-0.0093049109,0.1335303485,-0.0119812861,0.0977726802,0.0215975456,0.2703413665,0.1306134462,0.1557264775,0.2251377851,0.2261177301,0.4576103389,-0.5124971867,0.0661022589,0.0373427793,-0.2493051291,0.2425736636,0.5589843988,0.1534711719,-0.1143529192,0.1337856501,-0.3730230629,-0.1092103869,0.2899789512,-0.1280820817,-0.2938512564,0.0737991035,-0.5813837647,0.2530484796,0.0172076169,-0.3314665258,0.2307874262,0.0136135668,-0.177219972,0.2753067017,0.2678606212,0.8365145326,-0.4570670724,0.103018336,0.0011445108,-0.3318505585,0.046555955,-0.1733440161,-0.3231115341,-0.4147609174,0.0445171148,-0.0893562362,0.0026177175,-0.0699599832,0.2252667397,0.1070017666,0.1200768128,-0.0926552117,0.3599123657,-0.1197258607,0.5791432261,-0.0485459268,-0.2690149546,0.2073564529,0.1682010293,0.2366123497,0.0454038158,-0.13148579,0.0847711489,-0.3010174632,0.1168324053,-0.1397565454,-0.2063902915,-0.2243941724,-0.2509864867,-0.3670594692,-0.2147752047,-0.289758116,0.4370842278,0.1813788116,0.0487442687,-0.2492340654,-0.1513504982,-0.0564189143,0.0952757969,0.2041546553,-0.074277781,0.0794266239,-0.1869462281,-0.1399560124,0.3438461125,0.0996903777,-0.3000402749,0.1464681923,-0.1366907954,0.3839766085,-0.1102889851,-0.3345853388,0.2412901223,0.1035766304,-0.0345637016,0.1176462695,0.080399774,0.0547618344,-0.0088795945,-0.0843754113,-0.1235312894,-0.2258912623,0.0984864905,-0.0630132779,-0.0329018421,0.0670758784,-0.0384742022,0.1117518172,-0.2123567164,0.2502036691,0.1913926899,-0.1529754698,-0.2262974232,-0.0672005862,-0.0583535768,-0.150317356,0.3297863901,0.5043991804,0.0446604043,-0.2884420455,-0.0395113006,-0.305175215,0.1208179072,0.1991970241,0.3645035028,-0.1128467917,-0.1020559818,0.0982349813,-0.0357392654,-0.3419883251,0.032480564,0.0278009251,0.0813556015,0.3493483067,0.33746171,0.1446841359,-0.1031700075,0.0259710588,-0.2057397068,-0.2100810856,-0.1573517323,0.0012234058,0.1340416521,-0.0000421047,-0.2480038404,-0.1373692304,-0.3587089479,0.0457345843,-0.0177547634,0.1086430252,0.1591466814,-0.2690216005,0.1355026364,0.2857588232,-0.0394928306,-0.0789088011,0.4189853072,0.1688017398,0.242767483,-0.1434381902,0.3528724611,-0.2446471453,-0.2967463434,0.3589386046,0.0398059636,0.0226757843,0.3075048029,0.0224844795,0.0471258946,0.1039671674,0.2862005234,0.2158544064,0.113455534,-0.0325708836,0.1587417275,0.4928452671,0.2049648315,0.08445777,0.048349075,0.2964135706,-0.0804208294,0.1584670544,0.4488151073,0.5492035747,-0.0297647361,-0.1080661491,-0.2132347226,0.2940467,0.0685781091,0.0706083775,0.3314423263,-0.2698527575,0.0526083224,-0.0960480347,0.4973633885,0.2098038197,0.1848232448,-0.2710791826,0.3261646628,0.074494116,0.0849038661,0.1035947427,-0.4527049065,0.5264686942,-0.1542585492,0.1644180268,0.0150644034,0.0815414116,0.138286531,-0.1083258539,0.0551921278,-0.1803573221,0.2047864348,0.2272732109,-0.2318090945,-0.4666737318,-0.3928039372,0.0798891485,0.0555024147,-0.1574798971,-0.4056291878,-0.1510347128,-0.1940854192,0.057830818,0.060583625,0.2785733938,-0.1412671208,-0.2922287881,-0.1030336171,-0.0668308958,-0.0191861466,0.2330534607,0.1710578501,0.1807079762,0.0117062675,0.2945524752,-0.0442339405,0.1430578232,-0.1285067052,0.1688940674,-0.1636389196,-0.1438398212,-0.1164140403,0.1027717665,0.503702879,0.1779298782,-0.1836162508,0.1393096149,-0.1325043291,0.060336791,-0.006865656,-0.0011450228,0.3038974404,-0.2055754513,-0.0140965478,0.1673094332,-0.1579139382,-0.207157135,0.4206857383,-0.4214175045,0.2027395517,-0.1293613911,0.0933696181,0.0240692366,0.4000601172,0.1421524584,0.3937480152,-0.2291720659,-0.1997952759,-0.6256083846,0.0719341934,-0.0309096444,-0.2188179344,-0.1814703196,0.2008784264,0.0258024372,0.3286497891,-0.0908542499,0.3316740692,-0.1790325344,0.4084464014,0.0946504921,-0.2586821616,-0.1537325829,0.06199646,-0.0393062271,-0.1248531863,-0.0020094612,-0.2757777572,0.0361882262,0.0092468672,-0.2016094774,0.2079782188,0.0503706411,0.5563039184,-0.1002469733,0.3293508887,-0.0786158815,0.1105795205,-0.1071064174,0.0856865719,-0.5030032396,-0.0771812722,-0.1009547114,-0.0053004622,-0.2584873438,0.3241891563,-0.2446036637,0.0442472845,0.062897265,0.204973653,-0.1208750904,-0.2290636003,-0.1516967863,-0.1979351044,0.3643102348,0.1954901963,0.0609970465,0.2987620234,-0.038192343,0.05876204,0.3519357741,0.0225912314,0.2312604934,-0.0169116352,0.0948237628,-0.062231537,-0.0154617131,-0.2078003883,0.0430318639,0.2420939952,-0.145723924,0.2573820353,0.0028638418,-0.2082130611,0.0160817113,-0.1767822802,0.0718663633,-0.1037820429,-0.0878303871,-0.5066398382,-0.4559910595]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/133","title":"[Question] Using\/adding a local dataset","comments":"The download manager supports local directories. You can specify a local directory instead of a url and it should work.","body":"Users may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.","comment_length":20,"text":"[Question] Using\/adding a local dataset\nUsers may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.\nThe download manager supports local directories. You can specify a local directory instead of a url and it should work.","embeddings":[-0.3328148127,0.3316176832,-0.0937174559,-0.1150168031,-0.0208442621,-0.0157956686,0.1370421052,-0.0018574436,0.2311633378,0.2999218702,-0.2155137956,0.3249565363,-0.0286778808,0.2468104064,0.256311506,-0.019887669,-0.0942598656,-0.0035388255,-0.3042692542,-0.0223879162,-0.3524807692,-0.2041261196,0.2866287529,0.0946041495,0.1109767631,0.0382468104,-0.1240868047,0.3648514152,-0.2470917106,-0.5490741134,0.1527569741,0.1418217719,0.2310725302,0.437787205,-0.0001046002,0.0189364236,0.0589842759,0.0831919685,-0.3596253693,-0.3652914166,-0.4895199537,-0.1823952645,0.2321391404,-0.3313549757,0.1816500872,-0.2091437876,0.2616906166,-0.4603243172,0.1957477331,0.2347235978,0.2444289774,0.1967382729,-0.3012832999,-0.277736485,0.1513493806,0.3050220907,-0.0953642949,0.3225678205,0.480772227,0.0270891711,0.0487667248,-0.1207921654,0.0089251734,0.2245352864,0.5624610186,0.1625576466,-0.0116053196,-0.1680624634,0.1289849132,0.2235559523,0.8339118958,-0.2505113184,-0.2758901715,-0.1794393808,-0.1428814977,0.0745863169,0.1766538024,0.1667137146,-0.2424696088,0.3480572999,-0.2073036879,-0.3318106234,-0.4209996462,0.3002486229,-0.0823545232,0.0901305154,0.148043111,0.0346384458,0.1784930974,0.1505463272,0.3084104061,-0.2626871765,0.4867874682,0.1493736058,0.3469060063,0.0186692104,0.0091671869,0.0597288497,0.0581502803,0.0049298303,0.0845637694,-0.2019646764,-0.2501284778,0.1012817323,0.194765076,-0.1616328359,-0.1639881432,0.0453511216,0.1654997468,-0.1145339757,0.2640037537,-0.1346506476,-0.04747748,-0.1054486111,-0.2436897457,-0.04822696,0.1499397308,-0.2155676484,0.1587688476,-0.1933264285,0.143144086,-0.0827113613,-0.1080665141,0.1716115475,-0.0643270239,-0.1997919977,-0.1747719944,0.1752836108,0.0788822398,-0.003174942,-0.0325122513,0.0397609323,-0.1757469475,0.1771128178,0.092284672,0.1191456839,0.2410404533,-0.2445326596,-0.1945561469,0.2696327865,0.3167909682,0.0556184538,0.0812009349,0.0831421614,0.149827823,0.2602837384,-0.1244604066,-0.2472457439,-0.2761300802,0.1556815803,0.1210807711,-0.4608033597,-0.346748054,0.2225116342,-0.0693982542,-0.4279256165,0.1279610246,-0.0919031501,-0.3035586178,-0.1095278934,-0.3762881756,0.2411279082,-0.2823703885,-0.2036432922,0.061444737,0.4221624732,-0.5632275939,0.0680270419,-0.3140809238,-0.1536068022,0.081208393,0.0673325583,-0.4035091996,0.4359643161,-0.0880141854,0.1984979063,0.4498338401,-0.3739206195,-0.248434633,0.2898443043,-0.4758916795,-0.275029093,0.1604036838,0.2817880511,0.2360337824,-0.1272835732,-0.0010094335,0.4449621439,-0.0301013403,0.1336784959,-0.1492882669,-0.3828783035,-0.1846954525,0.1658445001,0.0935244188,0.2740213573,0.4902617931,-0.0867759064,0.1048768535,-0.2554169893,0.1711494178,-0.1441415399,0.3760519028,0.0958574936,-0.3978231251,0.10250251,-0.4131017625,0.2748021185,-0.2554860711,0.0173003785,-0.0387647673,-0.3529527187,-0.4677009583,-0.0808891878,-0.0483333506,0.1280551106,0.1403770745,0.3553267419,0.1730312705,-0.042150069,-0.2173710763,-0.1355109811,0.0085204896,0.0731845647,0.0879561082,0.1203848943,-0.1296135187,0.1671655625,-0.0807571933,-0.1476216316,-0.0683528632,-0.0872266889,0.0538804159,0.3684260845,-0.3002171814,0.3217330277,0.4200192988,0.3561283648,0.2749055624,0.0052998383,0.3201113939,0.0622730963,0.2744009793,0.0814503059,-0.445967108,0.0579807535,-0.0812322944,0.123615317,0.1355774552,0.1881991774,0.3140018284,-0.1872361451,0.0499427095,-0.1889361441,-0.1364404112,0.2132397145,0.2067412287,0.1099324375,-0.2441462129,0.3518172503,-0.2140483111,-0.0292856954,0.151651457,0.0076449458,-0.2025773823,0.1463209689,0.2122214735,0.7142144442,0.3711208999,0.2744206786,0.5198994279,0.310362488,-0.0486618765,-0.311712265,-0.0655756891,-0.0484361984,-0.3203289509,0.1223462224,-0.1544592977,-0.1924382299,0.002362818,0.1162005588,0.0746062472,0.1901597679,-0.2044833899,-0.2033991814,-0.4943545461,-0.4443024695,-0.3851741254,-0.176476419,0.1432518214,-0.3896022439,0.2325002253,0.2980713248,-0.3325253725,-0.0755388886,-0.0236256495,0.4244071245,-0.5079961419,-0.2385831326,0.0261594914,-0.1457481384,-0.1004976183,0.1987191439,0.6238669753,0.1396591812,0.4617751539,-0.3050397635,0.1750437319,-0.5360557437,0.2101890594,0.086243324,0.2917476594,0.267960012,-0.0695425346,0.142178908,0.1346641034,-0.1786262542,0.0194231831,-0.0481260866,0.0891558528,-0.4775236845,-0.0579296574,-0.0515679717,-0.0693865716,-0.4508682191,-0.4622930586,-0.2281161696,0.1067738831,0.2294006944,0.2104959488,-0.1482412517,0.1822174788,0.2864893973,-0.0247466024,-0.0642585307,-0.0540307835,-0.4249995053,0.3460866809,-0.4094592035,-0.3963734806,0.2634205818,-0.0048495037,0.0607354268,0.127043575,-0.3452188373,-0.2589051723,-0.0152641898,0.4125747979,0.2224730551,0.3128193021,0.3615954518,-0.0700532943,-0.1416071951,0.0418695509,-0.1188239977,0.2748989165,0.1181513742,0.1624618918,0.1214215234,0.0242045149,0.041386649,0.7666494846,-0.2841893435,0.0296369959,0.3063009083,-0.2404668629,0.2930563688,-0.1120333746,0.0548743047,-0.0382472873,-0.0130241495,-0.2256661206,0.2389036417,0.2462767214,-0.2232904136,-0.2001853287,-0.003566846,-0.4502517283,-0.2875757515,0.0985797346,0.2173651755,0.1011390314,0.2530359626,-0.2309326082,-0.2907760143,0.0605341643,0.086262092,0.2632459402,0.0246931762,0.0811339319,-0.4827651381,0.221823141,-0.5248218775,0.2419671714,0.1376827508,-0.1598442793,-0.0748456419,-0.19123061,0.2349520624,-0.0130826803,0.1374215782,-0.3182955384,-0.1204765961,-0.0853927433,-0.0559155755,-0.1883340329,0.1953862011,0.1754147112,0.1091788784,-0.1051429138,0.024798153,0.0241363011,0.060125351,0.1284392476,0.1796256751,-0.1894996762,-0.1080818549,0.2534787059,-0.4953245223,-0.243418023,-0.1111792549,-0.3200092614,-0.1051257551,0.3820067346,0.0263364762,-0.0429683104,0.2296731472,-0.0256568156,0.0640229136,0.014057193,0.2018829882,0.0301009081,0.2087985724,0.3099249601,0.2337269634,0.5688285828,-0.2552954257,-0.132339403,0.004919888,-0.365829885,0.1158285961,0.4357488155,0.1503397375,-0.1368775815,0.2269542366,-0.2468608022,-0.1247578785,0.3887690902,-0.1464826018,-0.1978249997,0.114044182,-0.545856595,0.2291252315,0.1126792803,-0.2533116937,0.3556565344,0.0142334048,-0.1803978682,-0.034232717,0.2393798083,0.8441696167,-0.3914889991,0.1342200786,0.1728963256,-0.4487317204,0.0640468299,-0.1675946862,-0.1599902213,-0.199022159,0.047585085,-0.1961748302,-0.0012954429,0.1033269465,0.2738250196,0.1472558081,0.1584479809,0.1287865043,0.3942500055,0.040109653,0.4644725621,-0.0648256391,-0.3008779287,0.1104108915,0.1808762103,0.1821755469,0.0285436641,-0.1747424901,0.023429133,-0.143666327,0.0528764613,0.1441975236,-0.1073759571,-0.1736645997,-0.2426132709,-0.3312175572,-0.0961977169,-0.1404302716,0.3021621108,0.231749773,-0.1153027937,-0.3340063989,-0.1055551618,-0.2329513729,0.1065126657,0.0841599554,-0.1033118367,-0.0349668972,-0.2349572182,-0.0964194015,0.2735937834,0.0137448935,-0.3880422711,0.1033767015,-0.0830756202,0.3187538087,-0.0601771027,-0.2184089869,0.1928029805,0.0720682368,-0.0842296258,0.1714804173,0.0040411479,0.06381841,0.0323038734,-0.0927064419,-0.128319636,-0.2378640026,0.1213048697,-0.0376541093,-0.1120582819,0.1412198991,-0.1316526085,0.0228834134,-0.229651913,0.2076699138,-0.0647614822,-0.0941110328,-0.2112787068,-0.0986709893,0.0131404661,-0.3099612892,0.292590946,0.4640975296,0.0650695339,-0.2433964014,-0.0453474037,-0.3601169288,0.0367979184,0.1390079707,0.2664690614,0.0454233587,-0.0703886449,0.2597786486,-0.2342162281,-0.3999510705,-0.0085374769,0.1805464923,-0.0074770218,0.2454086542,0.2613684237,0.1355552077,-0.1929587275,0.0586796142,-0.2924510837,-0.3188830912,-0.1917698532,0.0480895378,0.1074440479,0.0741114095,-0.3049527109,-0.0367656201,-0.4195577502,0.0747928247,0.1151763722,-0.0306945778,0.216125235,-0.2699196637,-0.0341940112,0.2148441523,-0.0833459124,-0.1578927785,0.3615546823,0.1708088964,0.2164482474,-0.0841170028,0.3144187927,-0.2583162189,-0.2349190414,0.279196173,0.0948935524,0.1129544526,0.275683701,0.134462148,-0.1123438329,0.0850342661,0.2172557861,0.2484403104,0.1143139452,-0.0378024466,0.1662724018,0.5602965355,0.2351333052,-0.0242154431,0.0052662264,0.2990368605,-0.0792749599,0.2069281638,0.3484202325,0.4614018798,-0.0520196445,-0.1128157526,-0.120528385,0.2037849575,0.1180578694,0.0149883423,0.4109712839,-0.2127862573,0.0593318567,-0.077200368,0.4835878611,0.1506009996,0.1024382487,-0.2601339519,0.3067755103,0.1027875915,0.0971295238,-0.1219797358,-0.3945372701,0.3556652665,-0.1409206986,0.0882926881,-0.070576705,0.1609783918,0.0493551791,-0.0906499997,-0.0427053012,-0.149849236,0.1929709762,0.2787309289,-0.2605365217,-0.5395200849,-0.3284620047,0.1070910469,0.1022425443,-0.0866559148,-0.3576251268,0.0754332319,-0.3292023242,0.1798388213,-0.0610151291,0.2364042252,-0.0184434727,-0.2858761847,-0.1096456274,-0.0603121705,-0.0596978813,0.0980016217,0.0946279764,0.2085938156,0.0144422036,0.2151188254,0.0339634642,-0.0314734913,-0.1862135828,0.0961282253,-0.2353504896,-0.1415176541,-0.0941005722,0.0512998626,0.5619301796,0.2237608433,-0.2359908074,0.3325533867,-0.2451448143,0.0877619684,-0.0076323226,-0.0057697417,0.3148099184,-0.1601537019,0.1945454031,0.0263813082,-0.2031202316,-0.1971875727,0.4323668778,-0.3405088186,0.2886347175,-0.1056365445,0.1221227869,0.137906149,0.4599789083,0.2430542856,0.3819717765,-0.1633378714,-0.2031960785,-0.7250159383,0.0014586527,-0.2276556641,-0.2019017637,-0.2462693006,0.0608271286,0.1831451803,0.5207514167,0.0384065583,0.2574203908,-0.1446758807,0.2845542431,0.0956900418,-0.2066713721,0.0241013467,0.210052833,-0.041422762,-0.2633382678,0.0799541473,-0.1722462773,0.0905362293,0.0933067352,0.0456164144,0.0172933955,-0.1484568268,0.4618344009,-0.1388920993,0.2116686106,-0.0301938262,0.0760782436,-0.123342216,0.1470332444,-0.4226232767,-0.1421387494,-0.1032996923,0.0295119286,-0.4267243147,0.3114053011,-0.2295437157,-0.1130969226,0.0896334872,0.3376246989,0.0192064717,-0.0996322259,-0.1746704578,-0.0605944693,0.2805763781,0.1080289632,0.0237143263,0.1330379844,-0.0562866181,-0.0110771246,0.3543693721,0.075547263,0.0758679658,0.0505732447,0.0024057857,-0.0291858129,0.0282703619,-0.1315639615,0.0883868709,0.2798042297,-0.1602360606,0.2538670003,0.0465172008,-0.1900811046,0.0160288513,-0.1646509767,0.1697913706,-0.1910301,-0.0753482431,-0.4665113986,-0.3920325339]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/133","title":"[Question] Using\/adding a local dataset","comments":"Closing this one.\r\nFeel free to re-open if you have other questions :)","body":"Users may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.","comment_length":13,"text":"[Question] Using\/adding a local dataset\nUsers may want to either create\/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets.\r\n\r\nIt appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this.\r\n\r\nA notebook\/example script demonstrating this would be very helpful.\nClosing this one.\r\nFeel free to re-open if you have other questions :)","embeddings":[-0.2340447307,0.0623236187,-0.1817659289,-0.1030367538,-0.1069895327,-0.1180460379,0.1800584644,-0.0092992103,0.2396566421,0.2702416778,-0.0248189699,0.3204395473,0.0419566818,0.3004589081,0.1500639766,-0.0574494079,-0.0028222739,0.1545833051,-0.2663291395,-0.1534758508,-0.4335796535,-0.2221324146,0.1845320165,0.0389680192,0.0839577541,-0.0319967382,-0.0995219573,0.4171653092,-0.1163977832,-0.4459762871,0.0869862437,0.130461514,0.0982927158,0.3471679986,-0.0000991708,-0.0978306159,-0.001086283,0.1619111449,-0.402551651,-0.176244989,-0.4742349088,-0.0627873465,0.2713688016,-0.3262061179,0.0122529222,-0.2950710058,0.1670463979,-0.6466331482,0.3660861254,0.2184128463,0.305755496,0.0819686279,-0.3102839887,-0.2864428163,0.0685531348,0.387534678,-0.0607775599,0.2625134587,0.4655345976,-0.1872126162,0.034356907,-0.0636488348,-0.0341813788,0.1888313144,0.5157752037,-0.0517737009,-0.0766924545,-0.1875923872,0.2455046326,0.2083102614,0.8219038844,-0.2154753208,-0.1962447762,-0.1062156856,-0.1752970666,0.0034485862,0.1389121562,0.0854434446,-0.1615106612,0.2759390175,-0.1765829921,-0.2980995774,-0.5053555369,0.2485807836,-0.0834107921,0.0303758979,0.0569580011,0.023557689,-0.0576950908,0.0686924085,0.28944242,-0.203019619,0.5600421429,0.1357426196,0.3157347143,-0.0120027168,0.0951921269,0.1650746763,0.1491907686,0.0883774906,0.1141545996,-0.1789590418,-0.2119991928,0.1444872022,0.1161604598,-0.1660731286,-0.0472291149,-0.0099805873,0.0785669386,-0.2127217501,0.0247576423,-0.0736438036,-0.1201349348,-0.1817115396,-0.186872974,0.1189934388,0.2740235329,-0.2628394067,0.0643348396,-0.0334318057,0.0907256529,-0.0742138028,-0.0382633507,0.2778432369,-0.089288868,-0.1569483131,-0.1282004416,0.150724411,-0.0631038249,-0.0671646297,-0.0618271269,0.1390417665,-0.2256961167,0.1613721848,0.028345162,0.2288921624,0.2170259058,-0.1051082835,-0.1911388636,0.2308250368,0.4124793112,-0.1451580226,0.2955024242,0.0443007015,0.0159696732,0.1239180341,-0.0410541072,-0.201072529,-0.2775067687,0.2061688006,0.0859148428,-0.3888732493,-0.3660963476,0.293123126,-0.0032295205,-0.4035815299,0.2647457719,0.0188626535,-0.2525904477,-0.1765717268,-0.3261222541,0.1680154651,-0.2663998604,-0.3591791093,0.0872679949,0.4117487073,-0.6436815858,0.1500520855,-0.3070729375,-0.1742684841,0.1417776495,0.1566924751,-0.3989138901,0.4139659405,-0.0504376888,0.1867071688,0.3808718026,-0.2965492308,-0.0626646653,0.217777133,-0.4783066511,-0.145410493,0.1472225785,0.1285156608,0.0499540716,-0.0794444084,0.1636555195,0.3535853624,-0.0161045045,0.0446406491,-0.2439544052,-0.3484984934,-0.190740034,0.1716372222,0.1046807021,0.1716074497,0.4113115668,-0.1049613655,0.1795637012,-0.3056634068,0.0104720946,-0.1012714282,0.5436905026,0.0576179922,-0.3521859944,0.0826254189,-0.3381786942,0.1245172173,-0.2403568178,-0.0366176702,0.0402812511,-0.3118146658,-0.5337172151,-0.0445402451,-0.0089946399,0.1224489212,0.2293052673,0.385519892,-0.0136336805,-0.1228340641,-0.2671014369,-0.0958808362,-0.0974722877,-0.0154830022,0.0083421404,0.1124456674,-0.1948833466,0.1666066647,-0.0225395709,0.0180010758,0.0485482961,-0.1234131977,0.0760154128,0.323096931,-0.2260910273,0.3416344821,0.4593682587,0.4408616424,0.180811733,-0.0346235372,0.2955825329,-0.1238637567,0.2348626852,0.1294243485,-0.3870770335,0.1037916839,-0.1160750315,0.0634738281,0.1204606816,0.1385178417,0.168935284,-0.3431692421,-0.0486665741,-0.177181676,0.011533197,0.164793402,0.1527663171,0.0910719484,-0.3921922445,0.3364053667,-0.1538802832,-0.0690778419,0.1373099089,-0.0650197268,-0.1643146425,0.1196354032,0.2948622704,0.4775601327,0.4031919241,0.343985796,0.4855925143,0.2693591714,-0.0859143808,-0.2834724188,-0.0168015789,-0.0003299483,-0.2478617132,0.1201950386,-0.1164511368,-0.0874390155,-0.1367408931,0.0182495881,-0.076775305,0.0889969468,-0.2155269533,-0.1668376029,-0.3519216776,-0.3126979768,-0.3334333599,-0.2116796672,0.2430629134,-0.3562237322,0.2966968715,0.1667497009,-0.4134536386,0.0263662674,-0.0069775456,0.5561195612,-0.497464627,-0.1548998654,0.1004190072,-0.2089561373,-0.0147156781,0.2798289359,0.4877380133,0.2950192392,0.5456094742,-0.2726047337,0.1664591283,-0.4077503681,0.1020412594,0.0549767092,0.2733381987,0.2737232149,-0.0685653463,0.2141100317,0.02062333,-0.310228616,0.0416363254,-0.0180058889,0.2001317888,-0.4790883362,-0.0170662366,-0.0782339871,-0.0166142732,-0.548636198,-0.2890625596,-0.2840363085,0.2108244449,0.2542214692,0.2451482266,-0.0246518143,0.254675895,0.3121144176,-0.0166108217,0.1208282188,-0.1680164486,-0.3653850853,0.1442697644,-0.3721332848,-0.3896836638,0.0484033376,-0.0525028817,0.1493432224,0.086292997,-0.3456423283,-0.3098781407,-0.010481583,0.388081044,0.1448433697,0.1383310854,0.4290950894,-0.043018695,-0.2826294899,-0.0481538326,-0.1192321926,0.1749422103,0.0377307683,0.1061021313,0.1541514993,0.0231395587,-0.0700898319,0.723616004,-0.1816315353,-0.0046591018,0.2998164296,-0.1400341541,0.2192247659,-0.0948968455,-0.0819178969,-0.0564594194,-0.0305207204,-0.23833552,0.2920174599,0.1995367706,-0.176544711,-0.1502692848,-0.0425757654,-0.6182237267,-0.3028629124,0.1731316149,0.0526345074,0.1535172909,0.1927720159,-0.2593116164,-0.3553339541,0.2690370083,-0.0452153906,0.2123460919,0.1046256125,-0.1179795563,-0.4842033386,0.2253107131,-0.3686075211,0.2601789832,-0.0698490068,-0.1286122948,-0.2732779682,-0.20356372,0.1894687265,-0.0189031698,0.1449096948,-0.2828314304,-0.1964805424,-0.0067266859,0.0100262752,-0.2284708023,0.2947481275,0.0482077077,0.1364623606,-0.0414153561,0.1467256695,0.1025760621,-0.0177126024,0.2952991724,0.2017811537,-0.0610762388,-0.0401869006,0.2910271585,-0.4672080576,-0.1972751617,-0.0252491608,-0.2999446392,-0.1395520121,0.3061011136,0.0136907697,0.0147501538,0.1678325832,0.1021149382,0.0549999736,-0.0327622928,0.1585744321,-0.011421782,0.1443578154,0.3292292953,0.2730075419,0.4998633862,-0.2941319942,-0.0329894274,0.0244506095,-0.2542631924,0.2938095033,0.5023182034,0.135807693,-0.1024566889,0.1792096794,-0.078979671,-0.1927848011,0.3268970847,-0.1273101717,-0.1792757362,0.0695736483,-0.486895591,0.2543378174,0.1652390957,-0.4311656952,0.4019594789,0.043104697,-0.2291384339,0.2390274256,0.2632145584,1.0068784952,-0.262650609,0.04497559,0.1469345987,-0.4212185144,-0.0308272392,-0.1589654535,-0.2502621412,-0.2789667547,-0.0068193013,-0.1130568534,-0.1287905574,0.1592760682,0.2173370123,0.0240318142,0.0026503571,-0.0129043004,0.2823114395,-0.0824458003,0.4282697439,0.0445356667,-0.1839808971,0.1394833177,0.2685530186,0.2380207181,-0.0571176335,-0.1718681753,0.1008396223,-0.2135462612,0.00058345,0.1387145221,-0.1073796749,-0.1952395439,-0.1596758515,-0.3355304599,-0.1731835157,-0.0303351805,0.3206864893,0.395281285,-0.0575888976,-0.3516646922,-0.0630966127,-0.0589498021,0.075987868,-0.0608622208,-0.1320289224,0.0852269754,-0.2531914115,-0.0954105482,0.2613358796,-0.0865725577,-0.246145606,-0.1592905521,0.0107836714,0.4053976834,-0.2496642023,-0.1967829317,0.2224326581,0.0350344628,-0.092913501,0.2478069514,0.0474300273,0.1734297723,0.110276401,0.044584699,-0.1405653208,-0.2639960349,0.1130712107,0.0934561491,-0.0472445898,0.2261512727,-0.1566031724,-0.0561277233,-0.4371493459,0.1719939858,0.1172479466,-0.0871128887,-0.2099872977,-0.0714192539,-0.0255184732,-0.1817005128,0.2636330426,0.4381730855,0.1590638161,-0.2424417734,-0.0368713997,-0.3174369037,0.1280519068,0.0556709766,0.3867987394,0.0977492332,-0.0678585395,0.1697053313,-0.0368913598,-0.5023216009,0.0323475264,0.0096138986,0.0710608512,0.2756697237,0.0604960248,0.0787763298,-0.2122939974,0.1883202046,-0.4003522396,-0.3579204082,-0.3027029037,0.1022326723,0.111400418,0.0361161418,-0.2791388333,-0.0597667582,-0.2974494994,0.154539004,0.054519102,-0.0464790948,0.1833060235,-0.2172808349,-0.0022286503,0.228149116,0.0082570333,-0.0811683536,0.2518073916,0.1825716197,0.0354309082,-0.160504654,0.1362122148,-0.3092097342,-0.3426457644,0.309307009,0.1838472486,0.2559071183,0.2386463284,0.0125938449,0.0111577092,0.0418265909,0.1350480914,0.2305101752,0.0637259781,-0.1040398553,0.1870907694,0.42646119,0.3415212333,-0.0220582355,0.0572044402,0.324224472,-0.1104958355,0.2144369632,0.22143659,0.4543704391,-0.0733810812,-0.0461507775,-0.1076486632,0.1140897721,-0.0000622018,0.04361099,0.3415201604,-0.23516047,-0.0469012596,-0.0819074959,0.4879409969,0.1134516522,0.0695640296,-0.238362968,0.2829244733,0.141702041,0.1095002145,-0.0188087244,-0.3168478012,0.554138124,-0.0696045086,-0.0238267146,-0.0586008877,0.1159224361,0.1440183669,-0.1167613938,-0.0492446907,-0.0456961878,0.2790544629,0.1182091758,-0.2542775273,-0.5589908957,-0.3193452656,0.0153020089,-0.163917616,-0.1532765478,-0.203448832,0.0843838379,-0.3356080055,0.0735878572,0.0675596818,0.2749829292,0.014930401,-0.1848711222,-0.0827663317,0.0016960189,0.0263610054,0.1817945093,0.0063589569,0.260366708,0.1720883548,0.2375580817,-0.0701462775,0.0784862265,-0.1099044979,0.0286813453,-0.132122308,-0.1483792067,-0.0729823634,0.1221996322,0.6002767086,0.3067511618,-0.2952425182,0.3751896918,-0.1512859315,0.0812696964,-0.1485119909,-0.1634014994,0.2722460032,-0.3174121976,0.1185677126,0.0256197155,-0.2066636384,-0.2176396251,0.3958863914,-0.2242447436,0.2425207049,-0.0550014824,0.1558466554,0.1021876186,0.4254706502,0.2391021699,0.2879161537,-0.2511937022,-0.3088442385,-0.5597196817,0.0273455326,-0.1773285121,-0.1529171318,-0.0898719057,-0.0010790532,0.24556005,0.4013886452,-0.082918331,0.1533690542,-0.1948561966,0.2117271572,-0.074538894,-0.02354756,0.2103841007,0.1126540378,0.0167007651,-0.1788679212,0.151513204,-0.0829238668,0.1897274256,0.0376730263,-0.0782870278,0.0873937905,-0.0743302405,0.3973536789,-0.0717712417,0.3200063407,-0.1342616975,0.0566522963,-0.0636165887,0.0719977096,-0.3411487639,0.0872846693,0.012277252,0.1168903038,-0.2723591924,0.0874424726,0.0177268293,-0.0829086006,-0.0146624045,0.259616375,-0.0019159735,0.0081234965,0.0308917835,0.0094227381,0.1343782544,0.1298496872,0.0005772363,0.1758455783,-0.1303355843,-0.0672528148,0.4800922871,0.0500495657,0.0108384471,0.0705886558,0.1969242394,-0.059997838,0.0538472049,-0.1275982708,-0.0240501594,0.3294790983,-0.104561016,0.2125699073,0.1246753484,-0.170885399,0.0193986949,-0.15161708,0.1392984092,-0.0974646807,-0.1235640049,-0.4502566159,-0.443315953]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/132","title":"[Feature Request] Add the OpenWebText dataset","comments":"We're experimenting with hosting the OpenWebText corpus on Zenodo for easier downloading. https:\/\/zenodo.org\/record\/3834942#.Xs1w8i-z2J8","body":"The OpenWebText dataset is an open clone of OpenAI's WebText dataset. It can be used to train ELECTRA as is specified in the [README](https:\/\/www.github.com\/google-research\/electra).\r\n\r\nMore information and the download link are available [here](https:\/\/skylion007.github.io\/OpenWebTextCorpus\/).","comment_length":13,"text":"[Feature Request] Add the OpenWebText dataset\nThe OpenWebText dataset is an open clone of OpenAI's WebText dataset. It can be used to train ELECTRA as is specified in the [README](https:\/\/www.github.com\/google-research\/electra).\r\n\r\nMore information and the download link are available [here](https:\/\/skylion007.github.io\/OpenWebTextCorpus\/).\nWe're experimenting with hosting the OpenWebText corpus on Zenodo for easier downloading. https:\/\/zenodo.org\/record\/3834942#.Xs1w8i-z2J8","embeddings":[-0.3416550756,0.0541993938,-0.1440849304,0.0401994474,0.1345212013,0.238294214,0.2991812229,0.2828027904,-0.1627614945,0.1327498853,-0.1687744707,0.1038049459,-0.0543265902,0.0419227108,0.1092259139,-0.2998088002,-0.1331175715,0.0652201474,0.0598098002,-0.2368174046,-0.1483214498,-0.1343982816,-0.2954437435,0.003287452,0.1121514514,0.069874078,-0.0302277487,0.0101079913,-0.0484651141,-0.2079993486,0.0732650384,-0.0241804738,0.187254101,0.212294504,-0.0000995318,-0.0908312947,-0.0499620549,-0.052070763,-0.2286730558,-0.1240527928,0.161016956,0.4190413654,-0.1013150364,-0.1493400782,-0.2204504609,-0.2088854164,0.3798259199,-0.1344406903,0.2971353829,0.3739475012,0.2415615171,0.4758035839,0.1870892197,0.0018353906,0.276990205,0.1021200046,-0.3548754156,-0.1740919799,0.3438061178,0.4092445672,-0.1062408462,0.2103044689,-0.0155081972,0.0210423712,-0.0163112544,0.0517000258,0.0689275935,-0.0672649965,-0.0548704378,0.4577101469,0.8695401549,-0.2064647675,-0.2438505888,0.2027133405,0.0450923964,0.077061832,0.0762129948,0.5061758757,-0.0081848614,0.4195893109,0.0165992156,-0.1815844029,-0.350672394,0.3086508811,0.1385599673,0.4003516138,0.029391259,-0.1291525513,0.0129015353,0.1995172352,0.3634063005,-0.0433269739,0.142138496,0.1371688545,0.1107874364,-0.298530519,-0.0903960615,-0.3708994091,0.331145227,0.2275838852,0.4582754374,0.1895662397,-0.303894341,0.1532961726,0.1554412246,-0.1198849231,0.0781551972,0.0762245804,0.2302327454,-0.1369104981,0.0664063096,0.0273375418,-0.2349761426,0.197774604,-0.1001581773,-0.0679594129,0.1375867277,-0.1571395993,-0.3402013183,-0.1625039577,-0.1236519814,-0.1919787824,-0.1331262887,0.0220856667,-0.2259699404,-0.0462540016,0.3709908724,0.1853947192,-0.1153065413,-0.5262021422,-0.0375421606,-0.0504118316,-0.0914681032,0.0703931898,0.2178780288,-0.0439450294,0.4347185493,-0.3388530016,-0.089258261,0.1874636263,0.3554647863,0.0764435381,0.1862937957,-0.1129118875,-0.0902589485,-0.2343527228,-0.2227841169,-0.0569833666,-0.2070690542,0.0934539437,-0.1236706078,-0.2657000721,-0.4540025592,0.2137362659,0.1499472111,-0.5109845996,0.0776997134,0.4250895083,0.0915563479,-0.1204427555,0.0168181378,0.2556272447,-0.1877433807,0.1907373816,0.1961418837,0.2280317396,0.0522253141,-0.1356109828,-0.0120484354,-0.0487793982,-0.2190424055,0.1398778558,-0.262046963,0.2281204164,0.1388502717,-0.0253516715,0.3520481884,-0.2142710537,-0.1915847212,-0.2567312419,0.1274421513,-0.3861259222,0.0223216787,0.018625889,0.2496936172,-0.1026858166,-0.1427347511,0.5792079568,-0.1119054258,0.1061472297,-0.1747275144,-0.2929340899,-0.0828349516,0.3579361141,-0.0588361509,-0.1591908634,0.0122420127,-0.0811278522,0.2287109494,-0.2200174034,0.1251217723,-0.1684464812,-0.1370516419,0.0657486171,0.0612899251,-0.1769488901,0.0469119884,-0.2467340082,0.1233018115,0.4557655156,0.3941364884,-0.3246976733,-0.0837066174,-0.1129464954,-0.1131170541,-0.2447959185,0.1833631247,0.0181464273,0.003304169,0.3506526947,-0.1210548505,-0.4486702979,-0.2372497618,0.1080167741,-0.1517303884,0.0622771606,-0.1660573184,0.1393907517,-0.0437201336,0.1566061974,-0.5391842127,-0.0107400725,-0.0151865836,0.2697483301,-0.4115535021,-0.0993571207,0.2405686378,0.4206374884,0.1517095864,-0.5572317839,0.2132436484,0.3098583817,0.1009841412,-0.1322560906,-0.2079693973,-0.0029611653,0.7288300395,-0.0632025748,0.0657449216,0.1638941914,-0.0169372447,-0.1122326404,-0.1638704985,0.0835229531,-0.2345880121,0.1059795693,0.074868314,-0.1551554501,-0.3463962972,-0.1415009499,0.346149534,0.0371975042,0.3818967938,0.1624904573,-0.5147320628,-0.1246497184,0.000826115,-0.2088274509,-0.0621309578,0.3045709431,0.1649667472,0.0683360621,0.0014974754,-0.0812461302,0.2212968469,0.2711848021,0.0863139853,-0.022593664,-0.0819015205,0.0685784146,-0.5081288815,-0.2816069126,-0.1346300244,-0.0094331615,-0.0521076359,-0.0454091839,0.1190504655,-0.040719606,-0.16778256,0.4195336998,0.1723707318,0.0056042746,0.3686810136,0.0377899967,-0.2071402371,0.2020102292,-0.057842575,0.5851584077,0.0425676554,0.2310312986,-0.0333863534,-0.134466514,-0.3236749172,0.2408467084,0.3488429785,-0.271972388,0.4540918767,-0.0963846818,0.173424989,-0.1643756628,-0.3966038525,0.1093534306,0.184969455,-0.1872778237,-0.0430469625,0.0064339917,-0.0410708226,-0.4515667558,-0.0681815073,-0.3204523623,-0.1990663111,-0.0817044228,0.0583978519,-0.1387014985,-0.0096551469,-0.8732765317,-0.0678644925,-0.2977582216,0.1441752911,-0.0199334603,-0.0106103923,0.1842948496,0.0666734502,0.0896256045,-0.1792040616,0.2841283381,-0.0811011493,-0.3858577907,0.4303667247,-0.2689103186,-0.3699898422,-0.2373476028,-0.2709103525,0.1706854999,0.0938930511,-0.3043581545,0.4677475393,0.114090018,-0.0313689001,0.0739698857,0.1081587896,0.3753954172,0.2612234056,-0.2660072148,0.0638164058,0.141771093,0.1371694356,0.0439077392,0.2617016137,-0.1572106481,-0.0834728032,0.0440246239,0.3705181181,-0.3164688647,-0.3068282008,0.2530996203,0.0491551831,-0.1117405593,-0.0923208296,0.0380803682,0.1283822954,-0.0443946123,0.1926495135,0.4442175031,0.3622460365,-0.4790538549,-0.2996875346,-0.1055109724,-0.232256487,-0.2077607363,0.0726971105,0.0188834332,0.2400633842,0.1455320716,0.0304731447,-0.1064880788,-0.2369484752,0.3415915966,-0.0550782308,-0.0510219,-0.1766169071,-0.2012384087,0.5690963268,-0.2576702833,0.1889905781,0.0350292251,-0.3098205328,0.1246698052,-0.1708921939,0.1245368943,0.1621868461,0.1331946105,-0.3949792385,-0.3374094367,-0.0910300463,0.3537002504,-0.0003866527,0.1490053982,-0.1752251089,-0.1137025282,0.0770967901,-0.3061680496,-0.1872982681,0.0844798386,0.6793684363,-0.0368533321,-0.1997461915,-0.0758721903,0.1459561586,-0.1037639156,-0.1317135692,0.0343384072,-0.096700035,-0.0157232564,0.1630744487,-0.0394295491,-0.215842098,0.3165538907,0.1303464919,0.0013063904,-0.1289670169,-0.0312915295,0.1151872799,-0.0646569058,-0.0073298998,0.3007616997,0.3237712979,0.0063380403,-0.2607872784,-0.1418108493,-0.4909071922,0.1463313252,0.0271840375,0.268764466,-0.1243201792,0.1977666318,-0.2094762325,-0.160842821,0.1815361083,0.238269344,0.139015317,-0.1660521179,-0.6486974359,0.293862015,-0.1727990657,0.0160119068,0.3939470053,-0.0225223135,-0.1447549015,0.3007645607,0.2392561734,0.9410755038,-0.1018494442,0.2914465368,0.1715653539,0.2178132385,-0.1211438254,-0.5121845007,0.0925285146,-0.045734942,-0.1621296853,-0.1449636519,0.0506944694,0.0361125655,0.2518134117,-0.2437445074,-0.0899406597,0.4621210992,-0.1314624101,0.1780579835,0.4241439998,-0.1022310033,-0.3146031201,0.0430879332,0.1476499438,0.2182667404,0.1382867843,0.1188396662,-0.260044992,0.0750306696,-0.1138517633,-0.2644019425,0.2084536999,-0.016160978,0.2436692268,-0.128902331,-0.0697784275,0.4377861917,0.2316160649,0.3585156798,0.0347482339,-0.281490922,0.1411949247,-0.080698356,-0.128795132,0.0833392292,0.0755191967,0.4916356802,-0.1459369957,-0.2151434422,0.3867561519,0.2107381374,-0.3158980012,-0.0797280595,-0.2938491702,0.49553442,-0.3037103713,0.2986175716,0.4633345902,-0.4224887192,-0.2079270333,0.2227209955,0.0680372491,0.0153544322,0.0444674045,0.3719411194,-0.0560906865,-0.1081303284,0.44019261,0.249855876,-0.0266859736,0.3439136446,0.1052713469,-0.0885656327,-0.3111062348,0.2990408838,0.5705190301,-0.628118217,-0.0101140672,-0.3074256778,-0.0161342882,-0.1165586859,0.487259835,0.1161918789,0.0269253328,-0.1268949062,0.2628557682,-0.215741232,0.039657943,0.1034291908,0.1985804439,-0.1197097003,0.0769043714,0.2621165216,-0.2367507219,-0.4181477129,0.0085617006,0.0082195234,0.0544789359,0.0789802521,0.0202755239,0.3161352277,0.0779509321,0.1458290964,-0.2162072212,-0.2228221446,-0.2964443564,-0.2569650114,0.0970709622,0.1131598651,-0.1727364212,0.0564326644,-0.0021583848,-0.2491900474,-0.0870446712,-0.0424979366,-0.192028895,-0.0237108711,0.2554457784,-0.0038500989,0.1805750281,-0.0784945861,-0.2276600003,-0.2619599998,0.3273430765,0.2250787914,0.0051597413,0.0705077052,-0.1824246645,-0.4175609946,0.1973008513,0.1508399844,-0.104810968,0.4084991217,0.0717317164,-0.1288691312,0.2340230793,0.1002642661,0.0500566475,-0.2019118518,-0.051341448,0.1480071992,0.3270938993,-0.3531073034,0.0617230013,-0.0781919137,-0.1491479129,0.1500087082,-0.0571009517,-0.11045257,0.0071199038,-0.103078045,-0.0710928887,0.0197486933,-0.1262753606,0.2625103295,0.4590882957,-0.0214599315,-0.1309042871,-0.0905266628,0.003974922,0.3096196055,-0.1580948085,-0.0030193964,0.4676138759,0.0987666845,0.3048251271,0.034587644,0.0281525329,0.2708967328,-0.1558602899,0.3019318581,0.3603947759,0.2809625864,0.2318606675,0.2192941308,0.0426739193,-0.3681986034,0.1026704833,0.059414845,-0.024440879,-0.0664692596,-0.182642594,-0.0233226046,0.0621953867,-0.0007720849,0.0172709581,0.1268072426,-0.1462226212,-0.1598113179,-0.1522328556,0.2438760847,-0.0997404456,0.1921796203,-0.2200870961,-0.022860935,0.2204122692,-0.215414077,0.0018844883,0.6561310291,-0.1262687296,0.034621302,-0.0564585812,0.0889652967,-0.2498203516,-0.1030735895,-0.1180217639,-0.0029303734,0.3171373308,0.0845222399,0.1932401359,0.1077018827,-0.1012548357,0.2016032189,-0.0577942133,0.0585693792,-0.1437858045,-0.3098757863,0.3371222913,-0.1900720149,-0.2066694796,0.0402961969,-0.2285959125,0.1614197791,-0.157110393,-0.1405712813,-0.0859856009,-0.1503361613,0.1288059503,0.0258252844,0.3386015892,0.302911669,0.1203877181,0.1545193493,-0.1798345745,-0.2010110021,0.1093736291,0.0706846565,-0.0825974867,-0.1306046098,0.1812009811,0.2614890039,0.3060030341,-0.0530240312,-0.3603710234,-0.1272686273,0.1842636913,-0.2383072674,0.071606867,0.233710736,0.3969296217,-0.0298623536,-0.0127485888,-0.0548710153,-0.1551704705,0.1224559397,-0.3688395321,0.0123423403,-0.0152056124,-0.0262555089,0.3564382792,-0.3443374932,0.1521074921,-0.2628708482,-0.0847820416,-0.1120615974,-0.1616512835,-0.1716986448,0.0398839861,0.410033226,0.0443215966,-0.1875505894,-0.2317012995,0.1421611756,0.2952626348,-0.0701477677,0.1697651148,-0.4030431211,-0.0734670535,0.0692079216,-0.0965164676,0.071432054,-0.0496656485,-0.2464208752,-0.4774279296,-0.0598765463,-0.3348174095,0.21588552,0.0383685865,0.2812624872,-0.4329897463,0.0350762419,0.215300411,0.1842904538,-0.7420243621,0.119530566,0.3623094261,0.0608482733,-0.2816313207,0.2801265717,0.3309626877,0.0552879907,-0.236408323,0.3154873848,-0.2009149194,-0.2137447447,-0.2145553529,-0.398398757]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/132","title":"[Feature Request] Add the OpenWebText dataset","comments":"Closing since it's been added in #660 ","body":"The OpenWebText dataset is an open clone of OpenAI's WebText dataset. It can be used to train ELECTRA as is specified in the [README](https:\/\/www.github.com\/google-research\/electra).\r\n\r\nMore information and the download link are available [here](https:\/\/skylion007.github.io\/OpenWebTextCorpus\/).","comment_length":7,"text":"[Feature Request] Add the OpenWebText dataset\nThe OpenWebText dataset is an open clone of OpenAI's WebText dataset. It can be used to train ELECTRA as is specified in the [README](https:\/\/www.github.com\/google-research\/electra).\r\n\r\nMore information and the download link are available [here](https:\/\/skylion007.github.io\/OpenWebTextCorpus\/).\nClosing since it's been added in #660 ","embeddings":[-0.3877815902,-0.0403194763,-0.1449296921,0.0515605286,0.1259210408,0.1362337172,0.368734777,0.2851602137,-0.1509412378,0.1764705032,-0.1019843221,0.1499531269,-0.0325923674,0.1335982084,0.1235400736,-0.2554698884,-0.0946017355,0.0875556991,0.055192668,-0.1899047196,-0.1755563766,-0.1045066714,-0.3305323422,0.0187032484,0.0080581727,0.0806048661,-0.0092836022,-0.0037962631,-0.0368271731,-0.1264729947,0.070007287,-0.0175121464,0.1833328903,0.1889968812,-0.0001014954,-0.0077340938,-0.0399677828,-0.0653062686,-0.1965959072,-0.0421554521,0.2006065696,0.4618940353,-0.0289071407,-0.1317331046,-0.2191943675,-0.1348990053,0.3582674265,-0.1650926471,0.3187970519,0.3941795826,0.2256833166,0.5161272287,0.1578972638,0.0575811267,0.2182250917,0.0788553432,-0.3408220708,-0.2025042325,0.2783665359,0.3472372591,-0.0947293118,0.3117616773,0.0521567538,0.0620321296,0.0190339517,0.1367871016,0.0609671809,-0.0571123809,0.0049640182,0.4471552372,0.8217999339,-0.2054247558,-0.223657608,0.1806342751,0.037247017,0.0077597094,0.1588153541,0.4549292028,-0.0239185933,0.4458140433,-0.0305058416,-0.1711679399,-0.3621671796,0.2912054062,0.0987679735,0.3703212142,0.0135396235,-0.1225320622,0.0377464183,0.1606004387,0.3667850494,-0.0358585902,0.1552544385,0.1161049008,0.03355667,-0.2954063714,-0.1284913123,-0.2616719902,0.2888487875,0.2260974348,0.4608129561,0.1467177421,-0.3142684996,0.2314681858,0.1001526192,-0.0693192482,0.0577681586,0.0906542167,0.274332732,-0.1397467554,0.0504975207,0.0567351431,-0.1986501664,0.1601760089,-0.1007301062,-0.1224919632,0.1775112152,-0.0689507946,-0.4112971425,-0.1297802776,-0.1674078852,-0.2306410223,-0.1071455032,0.0267978907,-0.2057501376,-0.0465597771,0.4278745651,0.2000532448,-0.1046921909,-0.5032296181,-0.0400838554,-0.0367808752,-0.0990439728,0.0577349812,0.1819639057,0.0604644902,0.4321676791,-0.2937154174,-0.1417885274,0.1505284309,0.4040749371,0.0945248231,0.1992779076,-0.1252820045,-0.1097523496,-0.2455857098,-0.2314138114,-0.1065759361,-0.1558239162,0.1048486754,-0.1339914054,-0.2115803808,-0.4920811653,0.2178239673,0.2056051046,-0.532615602,0.1008690223,0.4794957936,0.0845950767,-0.1169925109,0.0638774931,0.2188691646,-0.140970692,0.2324639112,0.2385520041,0.1919833124,-0.0291985087,-0.1511632502,-0.0227124114,-0.1026485786,-0.1858363599,0.217670545,-0.2734113336,0.2300826609,0.1716304272,-0.0135773858,0.3588407338,-0.1211051196,-0.2257569432,-0.268063426,0.0945944935,-0.4357400537,0.0287707653,0.0736039877,0.1296274066,-0.035390757,-0.1294262111,0.5638279915,-0.1008990929,0.1003933921,-0.2106841207,-0.3266806901,0.0754398853,0.3824287355,-0.0676751137,-0.1972745508,-0.0522886477,-0.0918024927,0.1890305132,-0.2591564357,0.1227320805,-0.2049455047,-0.116087079,0.0839333013,0.0592856184,-0.1854113787,0.0204060096,-0.2792411149,0.0521056391,0.4073783755,0.3977132142,-0.3542044759,-0.0568257943,-0.1018222719,-0.1733376086,-0.2822097838,0.1736875772,0.0938324332,-0.1177155524,0.3278291225,-0.1867308617,-0.4087429345,-0.2542396486,0.1016770154,-0.1550387293,0.1054275855,-0.2264609486,0.1502800733,-0.0574032255,0.1698767096,-0.5192562342,-0.0083447034,-0.0045254845,0.2610065043,-0.4526666403,-0.128607288,0.312805593,0.3532956839,0.1462346613,-0.5679572821,0.2290876359,0.2972147763,0.1013734788,-0.1089597121,-0.2331583202,0.0240023248,0.7703670263,-0.0651438236,0.0504427589,0.1000975668,-0.0774429739,-0.1379517168,-0.1183978468,0.0553533696,-0.2191727012,0.1190148368,0.1418917179,-0.1043063104,-0.4318858981,-0.1164130419,0.4207346439,0.0720381737,0.4207285643,0.2099877596,-0.459987253,-0.1841591299,-0.016949743,-0.2307055295,-0.0569763221,0.2738055289,0.0968585834,0.0946595147,-0.0529983863,-0.05611239,0.2073777616,0.2640318274,0.0325306542,-0.07980147,-0.0841491446,0.0265679751,-0.5563346744,-0.3130489886,-0.1172910184,-0.0014113039,-0.0793995038,-0.0191740803,0.0866606086,-0.0096890861,-0.22001867,0.4411745369,0.0897994637,0.0465565734,0.3832089901,-0.0245131943,-0.1839576364,0.2452980578,0.013222469,0.6161791086,0.0609317012,0.2628774643,-0.011903516,-0.1409990191,-0.3268576562,0.2364426851,0.3396543264,-0.2394582182,0.4305123985,-0.1129269153,0.1567393243,-0.1345756352,-0.4524295032,0.1055437252,0.1215145886,-0.1569504887,-0.0753604844,0.0213137753,-0.0973471552,-0.4712342322,-0.07033813,-0.2627441883,-0.2086256891,-0.1264191717,0.057490766,-0.1270720512,-0.0156925116,-0.8919383883,-0.119393833,-0.2571570277,0.2425616533,-0.0770832449,-0.0169104971,0.1199329272,0.0790969655,0.125092566,-0.2250371128,0.2384039462,-0.0932058468,-0.3829116523,0.3821946681,-0.2327339649,-0.3294761777,-0.2716021836,-0.2869305909,0.2067214996,0.0370258316,-0.2675123513,0.4028686881,0.0803063214,-0.0270764865,0.0551768877,0.1366366744,0.3926542699,0.2493353933,-0.2337570488,0.0543927439,0.0523979254,0.1728836447,-0.0292384904,0.2577363253,-0.143760398,-0.0013702903,-0.0215620305,0.3786074817,-0.2967534363,-0.3733518124,0.2822592854,0.0228955429,-0.0625976399,-0.119459942,0.0657942593,0.1830470711,-0.0055367248,0.1873258948,0.4459917843,0.2516974807,-0.4682088196,-0.2623145282,-0.0922858715,-0.2785878479,-0.2120842636,0.0890496522,0.0681721866,0.2034151256,0.0899923518,0.0264682323,-0.1268654317,-0.2039203644,0.3061027825,-0.0052839611,-0.0961911455,-0.206951499,-0.2939528227,0.4862935841,-0.2510069907,0.2009486705,0.0014517533,-0.1989254206,0.1124418899,-0.1637655497,0.1092712134,0.160771057,0.1369478106,-0.3870561123,-0.3548296988,-0.0974529088,0.3523896337,-0.0022407738,0.1641947329,-0.1728425622,-0.1059028804,0.1167501286,-0.3565562069,-0.1675637215,0.0597808063,0.6983379722,-0.0190197732,-0.1891335994,-0.0182206333,0.0898893028,-0.0758744627,-0.1104825065,0.0677561313,-0.0925467536,-0.0310386047,0.2083208263,-0.045513507,-0.2445542067,0.1958695799,0.1422866434,0.0439169109,-0.1163060889,-0.059187986,0.0957981497,-0.0563205332,0.0372640789,0.2535972595,0.2921512425,-0.004963615,-0.2836885452,-0.1432183087,-0.4728403687,0.2290361226,0.01153176,0.2251929194,-0.0428362563,0.2254658788,-0.1349596679,-0.1721255183,0.2196086794,0.2595805228,0.1713427305,-0.1405500472,-0.6014659405,0.323359251,-0.2178632617,-0.013818372,0.4246948063,-0.0457008965,-0.1391953528,0.2683946788,0.2296640575,0.8854385614,-0.0310222171,0.2890104353,0.1486107856,0.1543272734,-0.1530069858,-0.4866296649,0.0872864053,-0.0766761154,-0.113341257,-0.12366198,0.0425640233,0.0437268279,0.2063527107,-0.2192773968,-0.0745179802,0.4723415971,-0.1277459413,0.1536539048,0.4327242076,-0.1129104421,-0.2712637782,0.0577818751,0.129152149,0.2123001516,0.0983379856,0.1385897994,-0.2463244051,0.0438914448,-0.1938430369,-0.1844409257,0.1972681731,-0.1429845244,0.2465395033,-0.1377528459,-0.0748151913,0.4715871811,0.2500533164,0.4118134677,0.0127269812,-0.3041696548,0.1961467117,-0.162379548,-0.0625988096,0.0272323508,0.0283336658,0.5142261982,-0.1434932053,-0.2318817079,0.3554238975,0.22911717,-0.2941799164,-0.118901439,-0.2643765211,0.5663958192,-0.3097813427,0.3556226492,0.4194905162,-0.4346344173,-0.2398388386,0.2116590589,0.0750978515,0.0218264908,0.0524153523,0.3804723918,-0.0963802412,-0.1078628823,0.4786551595,0.2486327589,0.0296514872,0.4014142454,0.065577656,-0.0917664021,-0.3381180465,0.3314412832,0.6408321857,-0.6385036707,-0.0075447895,-0.274938345,-0.0702193975,-0.1215099469,0.5387367606,0.0850663781,0.0422591008,-0.1287598014,0.2572239935,-0.2716786861,0.042488955,0.0579413138,0.2608305514,-0.1173960343,0.1092964262,0.3252531588,-0.2559546828,-0.4107108712,0.0857389271,0.0026566877,0.1048321947,0.0597433113,-0.058290612,0.3389481008,0.1056333631,0.1603587717,-0.2157565504,-0.2359897494,-0.293695271,-0.2654471695,0.101140216,0.1096083224,-0.1536694467,0.0531124622,0.0547113419,-0.2621219158,-0.0967189297,0.0236885063,-0.2852711082,0.0040986724,0.2632400393,0.0082294894,0.0811253861,-0.0597714745,-0.3462846279,-0.254303515,0.303311497,0.1687547266,-0.0099953394,0.1004838198,-0.1927586496,-0.3776525557,0.1912083626,0.1621132195,-0.0790222287,0.3093197048,0.0105808629,-0.186629042,0.2124942392,0.0983381197,0.0894021988,-0.2179104835,-0.0410887934,0.1976447254,0.320189476,-0.3655350208,0.0717479363,0.010260527,-0.2152016014,0.1788294911,-0.1030413359,-0.0697328821,0.0642135963,-0.1231485903,-0.0722273141,-0.039207086,-0.1782136112,0.2840287387,0.4556744695,-0.0554209277,-0.0973291695,0.003134524,0.0078027309,0.3593313992,-0.0754332915,-0.0177058354,0.4935851395,0.0286958013,0.3113474548,0.091881305,0.0811404213,0.2171713561,-0.2187197059,0.2161054015,0.3700172305,0.2816602886,0.2890434265,0.1537790298,0.0588008948,-0.3837465048,0.0668499321,-0.0355866663,-0.0379459225,-0.0393544175,-0.1761793345,-0.0592756197,0.0712583587,-0.0270878132,0.0653369278,0.1770770103,-0.1137616932,-0.2127595544,-0.1434528679,0.2059846669,-0.0467800684,0.1903217286,-0.232079342,0.0013866078,0.215817973,-0.2469071001,0.0360720865,0.6709459424,-0.1151626781,0.0482663773,-0.0996639952,0.0629285723,-0.2260181904,-0.1138397157,-0.1179306209,-0.045549985,0.3118437827,0.0968402848,0.2352796048,0.0931884646,-0.1079428568,0.1775942743,-0.0625093952,0.0597955994,-0.1966835856,-0.3303006291,0.3023185432,-0.1630831361,-0.3181438744,0.0154390875,-0.2685100436,0.1846439838,-0.1995553672,-0.1066725552,-0.1283148676,-0.1667050123,0.1145398095,0.0064220023,0.3865298927,0.2747035921,0.0462843627,0.126290217,-0.2219376713,-0.2263638824,0.0898158848,0.1036984622,-0.0901469067,-0.1076349244,0.2040016651,0.2055097073,0.2586574554,0.0556356609,-0.2891088128,-0.1027566567,0.10220249,-0.283083111,0.1065571904,0.2305533886,0.3443319499,0.0179205127,-0.0352390483,-0.0142999543,-0.1472082883,0.1079862565,-0.3851658106,0.073937647,-0.0816582739,-0.1184856743,0.4436280727,-0.2781649232,0.1816434711,-0.2921372354,-0.0505095087,-0.0693266168,-0.1719315797,-0.2208983451,0.0625932217,0.4457108378,0.1359845698,-0.1765760928,-0.2248434424,0.0847439542,0.334923774,-0.0630496666,0.1498791426,-0.3228122294,-0.0599846281,0.0994840786,-0.0884238631,0.1236699522,-0.0454683118,-0.2438144684,-0.4222666323,-0.0523005091,-0.3576660752,0.2622029185,0.0196288675,0.2595988214,-0.4588454962,0.0475041382,0.2969136536,0.2203830481,-0.7713189125,0.1214495674,0.3653688431,0.0592060387,-0.2787189782,0.3049948812,0.3781586587,0.0607180186,-0.2137670964,0.3226258457,-0.1427860707,-0.1476722658,-0.1455879658,-0.4568995237]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/131","title":"[Feature request] Add Toronto BookCorpus dataset","comments":"As far as I understand, `wikitext` is refer to `WikiText-103` and `WikiText-2` that created by researchers in Salesforce, and mostly used in traditional language modeling.\r\n\r\nYou might want to say `wikipedia`, a dump from wikimedia foundation.\r\n\r\nAlso I would like to have Toronto BookCorpus too ! Though it involves copyright problem...","body":"I know the copyright\/distribution of this one is complex, but it would be great to have! That, combined with the existing `wikitext`, would provide a complete dataset for pretraining models like BERT.","comment_length":51,"text":"[Feature request] Add Toronto BookCorpus dataset\nI know the copyright\/distribution of this one is complex, but it would be great to have! That, combined with the existing `wikitext`, would provide a complete dataset for pretraining models like BERT.\nAs far as I understand, `wikitext` is refer to `WikiText-103` and `WikiText-2` that created by researchers in Salesforce, and mostly used in traditional language modeling.\r\n\r\nYou might want to say `wikipedia`, a dump from wikimedia foundation.\r\n\r\nAlso I would like to have Toronto BookCorpus too ! Though it involves copyright problem...","embeddings":[-0.1203030646,-0.0427291431,-0.0033754625,-0.0371080488,-0.0014099005,0.2495970875,0.2591406405,0.0787601024,-0.1759214103,0.1513458639,-0.1105382219,0.3519927561,-0.2013259381,0.3885519207,0.2654998004,-0.3297358155,0.1974350363,0.3152870834,0.1295732111,-0.3444743752,-0.2801202536,0.1701482236,-0.1321317255,-0.0258767437,-0.2297755778,-0.089881368,-0.2280260175,-0.170895949,-0.1450006217,-0.2987955809,0.1079742983,0.231218949,0.2329899967,0.3308149576,-0.0001178023,-0.129299745,-0.0035570574,-0.1993113309,-0.1886769086,0.0027899798,-0.1899906099,-0.0180445425,0.0683518425,-0.1603112966,-0.2583996058,0.0763194486,0.1850883365,-0.1007540375,0.0858062133,0.343433857,0.10101863,0.0015432481,0.0614229962,-0.0526098087,0.1530392468,0.1496795565,-0.3662744164,-0.0291989185,0.6788930893,0.0175581891,0.0982917249,0.3453907967,0.2276426107,-0.4172407985,0.4147026837,0.0883816779,0.1217895448,-0.510502398,0.0668936521,0.6966727376,0.7019848228,-0.2438404262,-0.3179569244,0.0559540316,0.2425318509,0.1563386321,-0.0448534824,0.2160191983,0.2315979898,0.2801142037,-0.2702526748,-0.5717220902,-0.2372366488,0.2509011626,-0.0221447758,0.5618988276,0.0817074627,-0.1417115331,-0.0033399714,-0.0242395122,-0.0552432351,-0.0694842413,-0.0548124015,0.1566242576,-0.0169132166,-0.6129307747,0.0411336236,0.2980645597,-0.092630215,-0.3859571815,-0.0651953295,-0.0231579095,-0.4463918507,-0.0368636101,0.1946208179,-0.156463176,0.2730415761,0.2248607427,0.0094607463,-0.1019401476,-0.1289753616,-0.0116830179,-0.0784892812,0.4305069447,-0.2459729165,-0.1600877792,-0.0489459783,-0.2302552909,0.1001332998,0.2012497336,0.1455676854,0.0592588224,0.1095893085,0.1507774144,0.0652142018,-0.0314450525,-0.2356081754,0.2275933921,-0.0804409906,-0.1276546568,-0.003691592,0.206608668,-0.125838846,0.1776729971,0.2078405768,0.0144338217,0.1936472654,-0.0902615339,-0.2388184667,0.3665217459,0.1816204786,-0.1185825467,0.16314964,-0.1314094514,-0.1708646566,0.1456835568,-0.1973750591,0.1564054042,-0.5758010745,0.1816931665,-0.4668942392,-0.3915768862,-0.5993227959,0.141055882,0.1419877559,-0.5276520252,0.2362723798,0.8123851418,0.064204134,-0.5009546876,0.3067885339,-0.0126405172,-0.4330202341,-0.0228663236,-0.0258893203,0.2668816745,-0.6962561607,-0.1327310801,-0.3971264362,-0.0115828551,0.0968688577,-0.0610348433,-0.3117049336,0.261877656,0.1716336161,0.2281699479,0.0753216147,0.2323235422,-0.1851024777,-0.1036177576,0.0874260664,-0.1750548035,-0.1340025961,0.2259523273,-0.1113839075,-0.0359231196,-0.0478876196,0.3111105561,-0.1302880347,-0.0960222334,0.0157367457,-0.4841164947,0.318810761,0.3002837896,0.272297591,-0.1253351569,0.0118789161,0.2225655764,0.0961411893,-0.110868372,0.1566894799,0.2868021131,0.3329593241,0.1994375437,0.0469315462,-0.3570227325,-0.1672731191,0.0165462494,-0.2720005214,0.1726179272,0.2039352655,-0.3268772662,-0.0627511144,-0.1004491746,-0.035023246,-0.1459440291,0.0689660385,0.0011789732,-0.0445479229,0.350902617,-0.1993529499,-0.4671763182,-0.2539493442,-0.2026211023,-0.0942984521,0.131910786,-0.1313410103,0.0148439491,-0.1028534845,0.3714749813,-0.1952738315,0.0429734103,0.0932413712,0.0261615906,-0.0709653571,0.2398361117,0.5780489445,0.2656174302,0.3966757953,-0.340257287,-0.1551364362,0.2055173814,0.0773994699,-0.1563781351,-0.3925898373,0.2254051119,0.5341241956,-0.0376358107,-0.0605878048,0.0535709634,-0.1059387699,-0.0207489822,0.1491303891,-0.3045207858,-0.0067129848,0.0497156717,0.0128582735,-0.0142342448,0.0356460363,0.2842810154,0.4608014524,0.0758397952,0.2276088148,0.2138712853,-0.093118906,-0.1594765633,0.1642937064,-0.6020973921,0.0047437302,0.1380546838,0.0668518692,-0.1709865779,0.2515239716,0.036465317,0.257309854,0.1864288747,-0.4395307004,-0.1851505339,0.0494308583,0.1068220288,-0.2343508005,-0.1974150687,0.0767159909,0.100461185,0.1266270876,0.0548214391,0.0118365353,0.2088557631,0.1573198736,-0.3849427998,-0.185869813,-0.0615806282,0.440664053,-0.1395955533,-0.1214400232,0.0538734049,-0.3997397125,0.7127889991,-0.0584130622,-0.1999062151,-0.0424081981,-0.4109727442,-0.2314403802,0.199538663,0.0768301263,-0.1943102032,0.3325208127,-0.1586371958,-0.0028328951,-0.1757369339,-0.6405138969,0.0662712306,-0.2856981456,0.106942758,0.1301275045,-0.0954240635,-0.4952315092,-0.6794479489,-0.0619290061,-0.1864671409,-0.1427813619,-0.192903474,-0.0497932509,-0.1747597009,-0.0376340449,-0.2620522678,-0.2260607034,-0.0736269429,0.0616546385,0.1563372761,-0.1010700092,-0.0249270294,-0.1283663958,0.1685890704,-0.1529408842,0.1515389979,-0.0851017162,-0.0959619284,0.3563258946,-0.1869836301,-0.0367288254,-0.0954204723,-0.3324646354,-0.0849369615,-0.119814001,-0.5708934665,0.0190490633,-0.3292813599,0.2812100351,0.1669231504,0.2473456115,0.2538973391,0.2934454978,-0.0000012368,0.0077695264,-0.2143929154,0.0833687261,0.4451698661,0.2463767081,-0.2431333661,-0.4042190015,-0.4259373546,0.9085752368,0.1068758518,-0.0638644919,0.3335602582,0.1069840416,0.1849312335,-0.3890581429,-0.3985509574,0.3437674642,0.0076479972,0.1845640242,0.6781775951,0.1935609579,-0.4262855947,-0.1212160587,-0.0828766301,-0.3206272125,-0.2982353568,0.1666903347,-0.0389205776,0.3487040997,-0.1588196605,-0.0854739323,-0.2506280243,-0.4194843769,0.2019395083,0.2093776315,-0.0223288648,0.0369570553,-0.1125409827,-0.0446531028,-0.3651693463,-0.0709725767,0.2231381088,-0.0861496031,-0.3378555179,0.0023914452,0.016174471,0.3255724609,0.1124800295,-0.3400231302,-0.4251874089,0.090305008,-0.0814472064,0.3436067998,0.0666731596,-0.3359043598,0.0442589484,0.0270873886,0.3047460914,-0.247565046,0.3161991835,0.1803212911,0.0627989694,-0.2615940571,-0.0206120219,0.0859699771,0.152007252,0.0093697226,0.1521784812,-0.1388440281,0.252455622,0.1712992787,0.0154283429,0.0630820766,0.1046419889,0.4216129184,0.2655829787,0.1079315543,0.3327842653,0.3009247184,-0.2129462659,-0.2090461254,0.2195741385,0.3560739756,-0.2384302169,-0.2108798027,0.0228045173,-0.417827785,0.3979946673,0.2974001169,0.2010069191,0.2210659385,0.3151341081,-0.2512997389,-0.0383732319,0.3850078583,0.1623445451,-0.129419148,-0.6602057219,-0.3214538395,0.3941152096,-0.0947410986,-0.0545481667,0.297124356,0.7003638148,-0.2230717987,0.3394843936,0.0942504555,1.2495231628,0.1217972562,0.2638053894,-0.1027573943,0.4040849805,0.4660696089,-0.2418933213,0.2373077124,-0.0360522829,-0.1895733774,-0.1081976444,0.0944936275,-0.2266768366,0.1481458396,-0.4586884081,-0.0527448393,0.0114851166,0.1850906014,-0.0073274374,0.1905531585,0.1831702143,-0.2310636938,0.1273107827,0.0228653494,0.2265528291,0.189373374,-0.0929233432,-0.3157699704,0.0324303433,0.127848655,-0.4980014861,-0.2815515697,0.0249724258,0.1483407468,-0.0530507751,-0.3387379944,0.5112301111,0.0929107592,0.4893570244,0.3941795528,-0.6226494312,0.1022606492,-0.0717197657,-0.1073729917,-0.1507455707,-0.0580609366,0.1940758377,-0.0611594059,-0.0825773254,0.3468394876,-0.0433920026,-0.1836774796,-0.6170405746,-0.090930514,0.3270126879,-0.2303509414,-0.1068833321,0.3243516684,-0.0353589915,-0.2372574508,0.0805064067,-0.0059942203,0.0026254442,0.0350049436,0.1640794575,-0.1183576956,-0.3099952936,0.0841729715,0.3252533376,-0.0996682495,0.4742273986,0.0861170813,-0.1129944474,-0.2319325805,-0.0448301211,0.4195480049,-0.2819791734,0.019798873,0.2061307281,-0.2327864021,-0.0839593336,0.4906166494,0.286111176,-0.0229897369,0.0821240395,0.0601215623,-0.2261487991,-0.0554719456,-0.2626163363,0.4064219892,0.0816564411,0.1091464832,0.1652400643,0.3366719186,-0.268812865,0.0010864609,-0.164252311,0.1797138751,-0.0653520823,-0.0068507642,-0.3436772227,0.033064194,-0.07670331,-0.3228807747,-0.1143826172,-0.0593982264,-0.3355407417,0.1810092479,0.1219650656,0.0069084899,-0.03232003,0.0321049318,-0.1256710738,-0.0758570135,0.0010258225,-0.0158414282,0.1976877004,0.4515360594,0.308994323,-0.1009596288,-0.3155783117,0.1131053716,0.1481800079,0.2062325478,0.3296627998,0.1669562459,0.1293245554,-0.1134046763,0.2945017517,0.4461606443,0.0404467136,0.0471049659,0.0449854992,0.3489354253,0.2054125071,0.2897990942,0.0164017417,0.2954607904,-0.137049824,0.0868140534,0.4369346797,0.1470106542,-0.4036336541,-0.3313179016,0.3603761196,-0.222819075,0.1565473229,0.1027681306,0.1354373395,0.0339056812,0.4411041439,0.0601487905,-0.4326860607,0.228925243,0.2286005467,0.5060856938,-0.0302599147,-0.0213914979,0.2700184584,0.2049050182,0.0924483463,0.2611296475,-0.0638318434,0.3037702143,-0.0329920314,0.1306117624,0.4755954146,-0.0753741935,0.2632484138,-0.4242350757,0.1863837242,0.2121647149,0.2218834758,0.2924069762,-0.2385537475,-0.1996545941,-0.4172307253,0.058275748,-0.2770192623,-0.10289184,0.2746732831,-0.3297052383,0.0710519478,0.0745954961,-0.0042716926,-0.0868161917,0.171675086,0.0161286257,-0.079866603,-0.2260412127,0.2717641592,0.0839497969,0.2781046033,-0.179423362,0.0297811981,-0.2411809564,-0.1518139094,-0.0983170196,0.2840474844,-0.1001974493,0.1047799364,-0.066455774,0.1172985435,-0.0512697026,-0.0188821703,0.0610758848,0.3471687734,0.3465960026,0.1170335859,0.3457550704,-0.0296342559,-0.0525401719,0.1200158149,0.116375789,0.1492111534,-0.189067632,-0.1836669594,0.2674719989,-0.1307184547,-0.2795535922,0.2237781882,-0.1392241269,0.2970333695,0.2243644893,0.042345915,-0.0313698202,-0.1629267037,0.0301333778,0.1524742544,0.4141487181,0.5447956324,-0.1703626066,-0.3880797923,-0.0804081932,-0.1971433759,0.0741244182,-0.0081351772,0.2461234778,0.0413562655,0.0459554307,0.0988351554,0.3108892441,-0.1848080605,0.075326778,-0.5170527697,0.1700903028,-0.3715028167,0.0418563001,-0.1737041622,0.0655958354,-0.0548095405,0.1317092627,0.3381567299,-0.1705050766,-0.0524503626,-0.2627622485,-0.4465302229,0.171893239,0.1726240367,0.3508761227,0.0367633365,0.1563948095,-0.3625408113,0.0843450129,-0.1391148716,-0.1857397258,-0.062195994,-0.0658125728,0.3058333993,0.1891340315,-0.3044047356,-0.1787871271,0.0287971254,0.4081341028,0.3072485626,0.2441803068,-0.2350321859,-0.0048551154,-0.0840900391,-0.1166811213,0.2761300802,0.3668229878,-0.033745341,-0.2301010638,0.0434890464,-0.3992784023,0.2303056121,-0.0316715129,-0.2247760296,-0.3862753212,-0.1471910924,0.2313435078,0.1427815109,-0.6531757116,0.0309710559,0.2807719409,-0.1503988653,-0.1293620318,0.3325277567,0.0073560243,0.0614835769,-0.2657532096,0.3298508227,-0.0010019829,-0.0641314611,-0.1747636646,-0.2928825021]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/131","title":"[Feature request] Add Toronto BookCorpus dataset","comments":"Hi, @lhoestq, just a reminder that this is solved by #248 .\ud83d\ude09 ","body":"I know the copyright\/distribution of this one is complex, but it would be great to have! That, combined with the existing `wikitext`, would provide a complete dataset for pretraining models like BERT.","comment_length":12,"text":"[Feature request] Add Toronto BookCorpus dataset\nI know the copyright\/distribution of this one is complex, but it would be great to have! That, combined with the existing `wikitext`, would provide a complete dataset for pretraining models like BERT.\nHi, @lhoestq, just a reminder that this is solved by #248 .\ud83d\ude09 ","embeddings":[-0.3530713916,0.0818351209,-0.0512021929,-0.0562965684,0.1030769795,0.1598965079,0.2854468226,0.1542768627,-0.2026538253,0.1954537332,-0.1945031881,0.4634687603,-0.2993954718,0.5999765992,0.2449026257,-0.2644297779,0.1414255351,0.3474517465,0.0667473748,-0.3228992522,-0.2888145149,0.1182890683,-0.0818758756,-0.1069841683,-0.3090595603,-0.1505122036,-0.2716774046,-0.0325037837,-0.1199504063,-0.2613653243,0.0278117564,0.2644578218,0.2675823569,0.3438077271,-0.0001135843,-0.0585605949,-0.1081648394,-0.2323906571,-0.1039653793,-0.0057008304,-0.1783859879,0.0983427241,0.1327925175,-0.1806357056,-0.3417238295,0.2791766822,0.1748458892,-0.0383970253,0.1419355571,0.3271052539,0.1491801292,0.2304730564,0.0555296279,-0.1316100359,0.2170151174,0.085074231,-0.4428583682,-0.0069279061,0.6354027987,-0.0787376612,-0.030719012,0.4032549858,0.278303653,-0.3690599203,0.3999812603,0.1626937389,0.1467266381,-0.5039070845,-0.0844702423,0.6871304512,0.5590509772,-0.1902266443,-0.3911696672,0.0173519906,0.2627430558,-0.0020228452,-0.0646487325,0.227801919,0.188866362,0.2231326401,-0.3767464757,-0.4018161595,-0.2587823272,0.1201045364,-0.0888696909,0.3482936621,0.0181965549,-0.204556644,0.0676023886,0.0931415483,-0.0513329804,-0.0168664306,-0.0237673428,0.0630133823,-0.1711230725,-0.660140872,0.1104794964,0.340424329,-0.0933348611,-0.2067766786,0.1112683192,-0.0527363978,-0.4054372013,-0.0392729826,0.2862157822,-0.2154929489,0.0773248151,0.266277045,0.0068214084,0.0015110233,-0.1948895007,-0.0262225848,-0.0596174896,0.4250321686,-0.2118391395,-0.2515172064,-0.0822372213,-0.2189814001,0.0427666381,0.1609366834,0.2324040085,0.0904442146,0.2137736678,0.2297777385,-0.0084169526,-0.133411184,-0.2920756042,0.3014232814,0.0131757287,0.018794585,-0.1116219833,0.020169368,-0.1280305535,0.0972481892,0.2295675725,0.0213323738,0.1727430969,-0.1225285232,-0.2133776546,0.3953699768,0.2022215873,-0.1189098433,0.1313255727,-0.0065382179,-0.1242757067,-0.0445319191,-0.1347397268,0.2393669039,-0.5524098873,0.0609942339,-0.3389555812,-0.4029352665,-0.4738232493,0.1847163737,0.1238063499,-0.4891537726,0.1798185855,0.6850425601,0.0096650934,-0.3333271146,0.1865439266,-0.0279926639,-0.3171575367,-0.0731904581,-0.0160568375,0.2111773342,-0.5390386581,-0.1289507598,-0.3195516169,-0.0129324608,0.1050254777,-0.0675212592,-0.2512214184,0.2571516335,0.1737056524,0.1071732715,0.2058099359,0.2996785045,-0.1902100742,-0.095382385,-0.0035975764,-0.214791894,-0.1459126174,0.285248965,-0.1660559922,0.0618912578,0.0182154607,0.2498273104,-0.2710356116,-0.0299949534,0.0236378461,-0.4691954851,0.284445852,0.3631178439,0.3294969499,-0.0258909948,-0.0532556139,0.1511998773,0.1603419483,-0.0624436513,0.0097879097,0.2339785099,0.3884833455,0.0950002298,0.0210911855,-0.3001109064,-0.0884518325,0.0025261219,-0.3351885974,0.0821895599,0.2839167714,-0.2929804921,-0.0610351376,-0.041577518,-0.0073852013,-0.1258832365,0.1319384426,-0.0251212846,-0.0895885825,0.2389881313,-0.2965579927,-0.3924486637,-0.237929076,-0.1192500666,-0.1590929925,0.1711838841,-0.1475754827,-0.0197852962,-0.048665829,0.3318711519,-0.2050869763,0.0680591837,0.0781360865,0.0499756634,-0.1077140123,0.1685992926,0.5811066031,0.2977339029,0.3056699038,-0.2758526504,-0.2440401465,0.172952503,0.0270780381,-0.1330551505,-0.5253860354,0.2636268437,0.4727225602,-0.0084802769,-0.0266865492,0.1053815633,-0.2098847181,-0.0864084214,0.1691185683,-0.1916768849,-0.0027123431,-0.0736141801,0.1142366603,0.044668749,-0.1398601979,0.2297053486,0.260619998,0.1344564259,0.1404762715,0.1407901943,-0.0942115933,-0.1160090119,0.2029231042,-0.5030984879,-0.0809803233,0.172503829,0.0874712244,-0.2393335849,0.1906817853,-0.059278395,0.1258677393,0.2230143547,-0.3664851487,-0.1972886771,0.1338398755,0.1744339764,-0.325378567,-0.2206885219,0.0457479768,0.1039743647,0.1199013367,0.1119407415,0.135046795,0.2634909153,0.250179857,-0.2869062722,-0.0801319331,-0.0138486056,0.4831455052,0.013174192,-0.1418718249,0.0235730838,-0.416972816,0.7396331429,-0.02195579,-0.196649611,0.0022886691,-0.2560915947,-0.1490013003,0.3308675885,0.1087094098,-0.2662095129,0.3676841557,-0.1753557175,0.034951143,-0.0942805633,-0.6390859485,0.0839085951,-0.2832455933,0.0863266289,0.2546829879,-0.099369511,-0.5634621978,-0.6116763949,-0.0091522429,-0.1429014504,-0.0804653093,-0.2683009803,-0.0629727766,-0.1535622329,-0.1109692752,-0.288218528,-0.1556944549,-0.1096823737,0.2387394905,0.2579617798,-0.1307810992,0.0299821533,-0.0955391601,0.1272521615,-0.1016827226,0.1058778241,0.0087048113,-0.1887646616,0.4039832354,-0.2219723165,-0.0735071078,-0.1701087058,-0.4423953593,-0.0159738977,-0.1636707336,-0.6092448831,-0.0072703967,-0.3321243823,0.1712649763,0.1855120957,0.1704378724,0.2090994418,0.2574058473,-0.0828415006,-0.0213904567,-0.1560780704,0.1461199373,0.4504311979,0.3514805734,-0.3593667448,-0.4878981411,-0.4969362319,0.7895026803,0.0241604093,-0.1183571294,0.3893427551,-0.0344803333,0.2317912579,-0.3303797543,-0.3117989898,0.2264675051,0.0291125923,0.0391533002,0.5663431883,0.2310748845,-0.3861228824,-0.1239110753,0.0151527729,-0.237560004,-0.3146321476,0.1676384211,-0.0290563367,0.2578862607,-0.1056678668,-0.0169243161,-0.2118214369,-0.3835376501,0.1079934537,0.1333487183,0.1000312045,-0.0985748768,-0.0911260247,0.0325333774,-0.4511336684,-0.1237985492,0.1290263087,-0.2660599649,-0.3289949894,-0.0824463591,-0.0775993913,0.2312130183,0.238392055,-0.1988155693,-0.4160836935,0.1095189676,-0.2606290877,0.3492560387,0.0796920136,-0.3689533472,0.0275904927,0.0107672475,0.4101486206,-0.1222265884,0.2937948108,0.2571627498,-0.0709461272,-0.2180236727,0.0089340247,0.0917362869,0.2005823702,-0.0131655792,0.2538703382,-0.2337388247,0.1975591332,0.0501574501,-0.1267797053,-0.0000917551,0.1219678074,0.4350592196,0.3168491423,0.262956053,0.2515999079,0.3808172345,-0.233199209,-0.1600285619,0.3447858989,0.4562059343,-0.2658222616,-0.1082517058,0.1253189445,-0.3459869325,0.3353765607,0.3743786812,0.1277441531,0.1225835681,0.2972899973,-0.1302846372,-0.0558442064,0.4407849908,0.1283885092,-0.1778684705,-0.5448563099,-0.2383829206,0.443215847,-0.1106247082,-0.0504176691,0.187182948,0.590031445,-0.066264689,0.3532591462,0.1125704125,1.1742764711,0.0579103082,0.2691525519,0.0418681987,0.3343931437,0.3043076992,-0.2695567608,0.2024072409,-0.1682733893,-0.2162494063,-0.1130810454,0.0808848664,-0.2641293108,0.2026871145,-0.3303598464,-0.0548721477,-0.0174442958,0.2853525579,-0.0743438005,0.0908841118,0.1218804643,-0.2132061273,0.0564059354,0.0829800218,0.254488945,0.1922994554,-0.0461959951,-0.4058105648,-0.0578150228,0.1218593866,-0.3544528782,-0.2294527739,-0.0064511537,0.0311129782,0.0727765337,-0.420992732,0.2956435382,0.2246629745,0.4153340161,0.386112988,-0.627442956,0.1474931091,-0.0022730397,-0.044508379,-0.1345216185,0.0469555072,0.123977907,-0.0455266647,0.0006766248,0.211545825,-0.0984642655,-0.2238708586,-0.3990501761,-0.0807346478,0.3533549309,-0.3402844667,0.0245301444,0.2219348848,-0.0580640472,-0.2396477908,0.1268497705,0.0246005394,0.0929384381,0.1228447855,0.2556537986,-0.1122233644,-0.2827653587,0.0371740833,0.3307575583,-0.0496480688,0.3765894175,0.0235415027,-0.0868113786,-0.3078312874,-0.0542515367,0.6219502687,-0.2399242818,0.0953597277,0.2075728178,-0.2176828384,0.0534915254,0.5695549846,0.2437940091,0.001492755,0.0744213909,0.134211272,-0.3346592784,0.0030410744,-0.2236312926,0.5563611984,0.0801088214,0.1106335372,0.1237306446,0.2785244584,-0.3402215242,-0.0506156124,-0.2924838066,0.1870934814,-0.1047564745,0.0030146141,-0.2522844672,-0.0578013845,-0.0017460083,-0.2160876095,-0.2029037476,-0.133037135,-0.2524747849,0.1397436708,0.1029133424,-0.0356582329,-0.0830610096,0.1615696549,-0.2261074185,-0.0484898947,0.0036189989,-0.0997241884,0.2949822247,0.2190819979,0.3462601304,-0.2225853652,-0.2889631093,0.056392286,0.1393466145,0.1009168997,0.1710449755,0.0963358283,0.0243503992,-0.2036394328,0.4045351446,0.37783131,0.1149256602,0.0316662192,0.0574861355,0.292737633,0.1111876443,0.23688218,0.0922051147,0.2531791031,-0.1055907011,0.1134812459,0.316167742,0.1974385232,-0.4240779877,-0.2772727907,0.2071357965,-0.1941078603,0.1854434162,0.1880311817,0.1805674136,0.055643227,0.3464222848,0.135581091,-0.4149521291,0.2608425617,0.2213757187,0.5095142722,-0.160080269,0.0083132777,0.2322040349,0.1866162568,0.0523970351,0.2950533628,-0.1322898567,0.3157801926,-0.0267529283,0.0576014854,0.4815668762,-0.1385781169,0.1304540038,-0.4692881703,0.217563808,0.2372852713,0.2014225274,0.377638042,-0.319866538,-0.2914625108,-0.3563239574,-0.0193266049,-0.2669887543,0.0034911481,0.2811425328,-0.3437141776,0.0792974979,0.0778418481,-0.0040180096,-0.0191985983,0.2073154151,-0.0819394737,0.0021550029,-0.2645290792,0.2155743688,0.1637546271,0.2516279519,-0.1718136817,0.1083473861,-0.220675081,-0.1467503756,0.0463265069,0.2716297507,-0.0404167548,-0.0113482848,0.0727947503,0.0648011565,0.0329129174,-0.0582956299,0.0761484653,0.1884684861,0.3378543258,0.101609908,0.3864337206,0.015535647,-0.1025246307,0.2758005857,-0.0661797002,0.3051455617,-0.1319627017,-0.1218654513,0.3402505219,-0.117672801,-0.2950793803,0.0557326674,-0.0447345562,0.2531768084,0.1830644906,0.0657372922,-0.0177029595,-0.1696589738,0.065183036,0.1298998743,0.4278731942,0.4784700572,-0.1147143468,-0.4671115577,-0.0516655929,-0.1992997527,0.1428870857,-0.0061023468,0.2872783542,0.0556401722,0.0380976796,0.1537520289,0.3480788171,-0.0662874952,-0.0870205238,-0.4227252603,0.1524617374,-0.3983856142,0.0812174156,-0.2558302283,0.0397118852,0.0126817161,0.1134594679,0.3701181114,-0.2032474428,0.0140354987,-0.2011418194,-0.4450189173,0.0956254378,0.1167504266,0.3339002728,-0.0918285176,0.1622655541,-0.3778140843,0.0562093481,-0.0789315924,-0.2871648967,-0.2027499676,-0.2450581789,0.2750613987,0.3006405234,-0.2316131592,-0.081600748,0.0098084183,0.3548703492,0.4189389944,0.3003314137,-0.1389712989,-0.0530018881,-0.0226600245,-0.0337900147,0.3255939186,0.3307943642,0.0249421559,-0.2284136862,0.038185548,-0.4998102784,0.3786394596,0.0978988037,-0.184520036,-0.4066109359,-0.2397980839,0.3181039095,0.2063694298,-0.8003483415,0.0128229642,0.181505695,-0.2265817225,-0.0773545355,0.3136056066,0.0670312345,0.0541073978,-0.2679041922,0.3143021166,-0.0931516141,-0.045451805,-0.2162485272,-0.2869776487]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/130","title":"Loading GLUE dataset loads CoLA by default","comments":"As a follow-up to this: It looks like the actual GLUE task name is supplied as the `name` argument. Is there a way to check what `name`s\/sub-datasets are available under a grouping like GLUE? That information doesn't seem to be readily available in info from `nlp.list_datasets()`.\r\n\r\nEdit: I found the info under `Glue.BUILDER_CONFIGS`","body":"If I run:\r\n\r\n```python\r\ndataset = nlp.load_dataset('glue')\r\n```\r\nThe resultant dataset seems to be CoLA be default, without throwing any error. This is in contrast to calling:\r\n\r\n```python\r\nmetric = nlp.load_metric(\"glue\")\r\n```\r\nwhich throws an error telling the user that they need to specify a task in GLUE. Should the same apply for loading datasets?","comment_length":53,"text":"Loading GLUE dataset loads CoLA by default\nIf I run:\r\n\r\n```python\r\ndataset = nlp.load_dataset('glue')\r\n```\r\nThe resultant dataset seems to be CoLA be default, without throwing any error. This is in contrast to calling:\r\n\r\n```python\r\nmetric = nlp.load_metric(\"glue\")\r\n```\r\nwhich throws an error telling the user that they need to specify a task in GLUE. Should the same apply for loading datasets?\nAs a follow-up to this: It looks like the actual GLUE task name is supplied as the `name` argument. Is there a way to check what `name`s\/sub-datasets are available under a grouping like GLUE? That information doesn't seem to be readily available in info from `nlp.list_datasets()`.\r\n\r\nEdit: I found the info under `Glue.BUILDER_CONFIGS`","embeddings":[-0.1725947261,-0.2484223098,0.0658995882,0.3159227967,-0.0194631834,0.0501955673,0.3762942553,-0.1403564811,0.7129033804,0.1203574985,-0.281144768,0.5193799138,0.0818037093,0.0636687577,0.2311221063,0.0292404853,0.0766592845,0.1667412966,-0.1115129814,-0.1865060031,-0.4865836501,0.3221149743,-0.3004097939,0.2563777268,-0.1656361967,-0.0245170966,-0.1407751441,0.1467231661,-0.108386524,-0.2313846946,0.4261036217,0.4168219864,0.1703007817,-0.1210095137,-0.0001230205,-0.0875958875,0.6260707974,-0.1760393083,-0.1699001342,-0.2148626596,-0.6149526238,-0.3502326012,0.4615489542,-0.2409356982,0.121970579,0.3319019675,0.3726021349,-0.1849743277,0.0137700653,0.0481360778,0.0500271097,0.5045632124,-0.3903687894,-0.0787917599,-0.0217167828,-0.1140737087,-0.0577430613,0.6744437814,-0.0432515889,-0.3659249842,0.0440582149,-0.0565739125,-0.0486897714,0.3303422928,0.3373464048,0.0048554521,0.0214751195,-0.4663365483,-0.1145747453,0.5373286605,0.207733199,-0.1874137819,-0.1875194311,-0.306093961,0.1226772144,-0.0138648329,0.2320161909,0.311306864,0.0170250237,0.0447694771,0.1531568319,0.0357270986,0.1865292341,0.3109622598,-0.1849832684,0.5227513313,-0.1189951822,0.2692030966,0.2278757244,-0.0366255753,-0.1399613768,-0.3156047761,-0.2983661294,0.1015648022,-0.3652482033,-0.0635139495,0.0324664265,0.5708153248,0.1710638106,0.007658991,0.2041781992,0.0784858391,0.065491356,0.3932666183,0.3801899552,0.124688603,0.4200635254,-0.0918590575,0.0543452762,-0.0628205836,-0.102940686,0.1195995882,-0.3057979941,0.0590026192,-0.1013354659,-0.2553478181,0.1419204175,-0.0149535704,-0.4296116829,-0.2354436517,-0.093179971,-0.1769378632,0.1591098756,0.2330816537,0.0021019669,0.1385073811,-0.3767081499,0.3707320392,-0.1587339044,-0.0748586357,-0.138585791,-0.0942317322,-0.3390645385,0.1342394501,0.2060617507,-0.0339633897,0.2835069895,-0.1634259075,0.0305538792,0.138215974,0.1435979456,0.0048819045,0.1341758668,0.2735054791,0.307185322,0.0825897828,0.0443132818,-0.4303753078,-0.3361852467,0.0561994314,-0.2327112406,-0.2623212337,0.0821429417,0.0417890511,-0.6383655667,-0.0251228437,-0.6208206415,0.3129177392,-0.0949081033,0.1121767014,-0.0261938348,-0.1705806404,-0.2382139862,-0.2332176268,-0.0261172801,0.3035363257,-0.0929569378,-0.3242765367,-0.2786442935,-0.2820478678,-0.098320514,0.0907391831,-0.1283429861,-0.018978525,-0.1342946589,0.06959676,0.4640235603,-0.1827951819,-0.3434996605,0.1787410825,0.0483872779,0.0153116807,-0.0854013339,0.1239225566,0.0589953214,-0.0509528853,0.4281704724,0.4943158329,0.0499853827,-0.0940643623,0.0150875002,-0.0785889849,0.1891945302,0.0466191955,-0.0622825809,0.2793145478,0.0416076593,-0.0561059751,0.0146746729,0.341131717,-0.4115764499,-0.0953838304,-0.0471466668,-0.2374942303,0.1808024943,0.0512195081,-0.4635836184,0.2720180452,0.0998440608,0.0805387944,0.2446931005,-0.1569918841,-0.1589941382,0.1996584535,-0.0505542904,0.3557608724,-0.0723996907,0.312301755,0.1252313852,-0.2424221486,-0.2602823079,0.6475881934,-0.2580867112,0.1739526242,-0.1627800763,0.1830734015,0.2841739953,0.042868495,0.0041607409,0.2706499398,0.0965240225,-0.0510825478,0.0416191444,0.3302467465,0.2279451787,-0.0203640107,0.1437880844,-0.0672411025,-0.0037752101,-0.0256039631,0.1983072311,0.0699307695,0.0402127951,0.067323871,-0.0200360883,0.3690297604,0.0603734516,0.2613092959,-0.0212721918,0.0521854237,0.1354755014,0.1276657432,-0.5614256859,-0.2807866931,0.0287649464,0.082913734,0.5041057467,0.3679535687,-0.3384752274,-0.1453409046,0.4069750607,-0.060414087,0.1469535232,-0.1286323071,-0.3089123964,-0.0977878347,0.2786588669,0.4591606855,0.5872238278,0.1190103218,0.0607877187,-0.2157073617,-0.191721797,-0.2105320394,0.0080239661,0.2727658451,0.010636013,-0.2120886445,0.3194173872,-0.1029439196,-0.4509445131,0.3798750043,-0.0225264635,-0.1781170666,-0.2922337353,0.0603942722,-0.0216050968,-0.6061503291,0.0375328586,-0.24335289,-0.2842607796,-0.4053376913,-0.1251335293,-0.1692967415,0.0879081711,0.1985606402,0.2002056241,0.4580504596,-0.4821966588,-0.1907385141,0.0049896152,-0.0888826028,-0.1672079861,-0.0923187062,-0.0698061809,0.2826128006,0.1935006976,-0.2446638495,0.0474125408,0.1971144676,-0.1925168931,0.329113096,-0.1074580848,0.5722259879,0.1572949886,-0.0895838812,0.0413080975,-0.1016609296,0.1898089051,-0.1826377362,0.033904884,-0.2854732871,0.0542666055,0.0052303951,-0.1092132106,0.0582989752,-0.5250714421,-0.2473227084,0.0293148477,0.1289645731,-0.0038830021,0.1676345617,-0.4640073478,-0.0216197148,0.1260849386,-0.0622136071,-0.2979718745,-0.1379568577,0.2206344604,-0.1228186637,-0.1601770967,-0.0570614971,-0.3852192461,0.1511649191,0.0822596699,-0.0723710805,-0.5356999636,-0.153814733,-0.0764898434,0.3229403794,-0.0198661089,-0.0556149371,-0.1602959633,0.2289456278,-0.1405753344,-0.3786470294,0.0012552876,0.3115509152,0.1157942042,0.1630406976,0.0895351171,-0.1561243087,0.4338520765,-0.1202090532,-0.3725539446,0.2616612613,-0.1179798171,0.5720015168,0.1639204621,-0.1120459288,0.0150117064,0.2438841164,0.1999145895,0.4289881885,0.1312764734,-0.5775893331,-0.1473947167,0.3481875658,-0.217397958,-0.0927844793,-0.1649266928,-0.1839098036,0.0381601043,0.2785153985,0.1254333258,-0.240496397,-0.2598409057,-0.2579221129,0.5730073452,0.1214921325,-0.0010677513,-0.0006986006,-0.188199833,-0.4797840416,0.3396484852,-0.1379379928,0.2610199153,0.0711839572,-0.1808648705,0.054503005,0.028228024,0.4136380553,-0.4197868705,-0.1091813445,0.2105709612,0.2621342838,-0.5395376682,-0.2534891367,-0.1327023804,0.161636278,0.2831813395,0.3954243958,-0.1778280884,-0.3551356792,0.1704839468,0.0274270549,-0.041103784,-0.3452648818,-0.4882954359,-0.0566726252,0.0115762101,-0.153635323,-0.3646239936,0.0242681224,0.1641968936,0.1713278145,-0.2304840088,-0.0268829986,0.2792695165,0.3903970122,0.3077055216,0.3962610662,0.0455122665,0.0234219078,0.3396551311,-0.2034998685,0.2640024424,-0.2735615373,0.0892031863,0.0983520076,0.2199610621,0.4898403287,0.1392580867,-0.1975558102,-0.2086831331,-0.3556486368,-0.193010062,-0.2211498022,0.5052555203,0.2868886292,0.1124188527,-0.3608316481,-0.4029764831,0.6374166608,0.3167223632,-0.3363929689,0.1907069683,-0.1471171528,-0.214784801,0.211468935,-0.133271575,0.6008566022,-0.2714727521,-0.0229018908,0.2934980094,-0.0912845582,0.5923182368,0.1955212653,0.0557716861,-0.168160826,-0.3279605508,0.0124272415,-0.1805083603,-0.0889320597,0.4985411465,-0.0640219003,0.3416544497,0.1716861725,0.0928528607,0.129314661,0.4685433209,-0.1334171295,-0.4109864831,-0.321007818,-0.0561121553,-0.1951428056,0.3181063831,-0.1777684987,0.0801919103,0.2733339071,0.0147976791,-0.2270472348,-0.1246654987,-0.1243035048,-0.2731578648,0.2614934146,-0.1384985298,-0.0948751941,0.0144327404,0.0576809086,0.2713041306,-0.0127286185,0.3360319436,0.0560382456,0.0094601102,0.074154675,0.0778452232,0.2644193769,0.262159884,0.1019129157,-0.3572232425,0.078108415,0.0508938208,-0.1172807142,0.2615503073,-0.0818520635,-0.0782536715,-0.3958770335,0.0864315256,0.1346022636,0.1327538192,0.0166734215,0.0140067711,0.0466091037,0.2174340487,-0.0068892422,-0.4923045635,-0.1174881756,0.1458957791,0.0056935074,-0.1417082697,-0.1675385833,-0.2822417021,0.0204147045,-0.0521882698,-0.0765814483,-0.1556699574,0.0530719236,0.1401907504,0.198268488,-0.1143056154,-0.0026671886,0.0962615609,0.2301116586,-0.0851900429,-0.0016917229,-0.4102162719,-0.0616993681,0.3642499447,-0.2124478221,0.1922762692,0.1400474459,-0.2225620151,0.2793596983,0.0068772477,-0.1233970672,-0.0267786607,-0.2066753656,0.0513302907,0.4680585265,0.2214337587,-0.0004541745,-0.1304057539,-0.0374140069,-0.1361543983,-0.207468614,-0.094070375,0.0528177768,0.2285226583,-0.0060461997,0.0739196092,0.1231034994,-0.286813885,-0.0283010341,-0.381631434,0.2295446247,-0.1392904818,-0.2746681273,-0.1037012637,0.236836046,0.2031371891,-0.013137849,0.3210723102,0.0591191538,0.2595496178,0.1804192364,0.0758329704,-0.0321286134,0.1945056021,0.0709731802,0.1289192736,0.1149740517,-0.0108192312,-0.1177629977,-0.1611794531,-0.0711769313,-0.2137155384,0.2505786121,0.1549585909,-0.0079257218,-0.0201234538,-0.0746395737,-0.0352764092,-0.155285731,0.0498857796,0.4855635166,0.0985367373,0.2936947644,0.4415092468,0.4212818146,0.2450571805,0.1583522409,-0.1554617584,0.4559428394,0.1695654392,0.2591851354,-0.020420257,-0.0706861541,0.3534201682,0.4438771307,0.0906872377,0.0828980282,0.1025343463,0.4269812405,0.359785974,0.4493689835,0.0207054224,0.4012000263,-0.3365100324,-0.0249276888,0.2151338905,0.0700939894,0.1877699345,0.1934136599,0.3586158752,-0.4535290897,-0.1445495337,-0.2608300745,-0.1262645125,-0.0567817762,-0.3152939081,-0.0631950498,-0.0281352848,-0.2097603381,0.1032829881,-0.0767308027,-0.3128421903,0.2273436785,-0.1764271408,0.2926792204,-0.0310119223,-0.0866282657,0.2989379466,-0.0908518285,-0.1975506991,0.2206462771,-0.6988401413,0.1629049182,0.186883837,0.2296291888,0.2515773773,0.1685351133,-0.0951669738,-0.2500780821,0.1212634891,0.0343983397,-0.3038606644,0.1564115882,0.2277084142,0.0262300652,0.3221802711,0.066687122,-0.0435386859,-0.0528642721,-0.0116808387,-0.1098030433,-0.335627526,0.3870970905,-0.1120915934,0.0079677012,-0.0141070895,0.159694016,-0.3824214637,0.2323531806,0.1740886122,0.0809750259,0.086835973,-0.1557409018,0.017270213,-0.0826461539,0.6005551815,0.2271195352,0.5752949715,-0.2252593189,-0.3279740512,-0.7001205683,0.0368187316,-0.213243112,0.0002161046,0.231488362,-0.0133432727,0.1313794851,0.309060216,-0.2131347656,0.6716585755,0.0967309624,-0.3353390992,-0.2283659428,-0.2301148176,-0.2177292109,-0.2783159912,-0.1690437347,-0.3825414181,-0.286383599,-0.3098651767,-0.1817477345,0.0221986026,-0.1952802986,0.2178573161,0.3884981871,0.2519211769,0.059437748,0.5966924429,-0.0084712552,0.364321053,-0.0634283051,-0.4590592086,-0.190564692,0.1081524417,-0.3577147126,-0.0781206042,-0.1493205279,0.3721561134,0.0095064174,-0.1486745179,0.1301596761,0.2861028314,0.1310577393,-0.1102912277,-0.4144431055,-0.1108416244,0.0648282766,0.3997950554,0.1975503117,-0.1892720014,0.274508357,0.0350961722,0.1692058146,-0.4382280409,-0.3190026581,-0.2125051767,0.2093898356,-0.0648908988,-0.0385554731,-0.3037867546,-0.1511371434,0.4055892825,-0.2444157451,-0.1811446697,0.0704870075,0.0627478436,-0.0557158031,-0.2137461752,-0.2511053085,0.1618840843,-0.3231654465,-0.1326719671,-0.2384339273]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/130","title":"Loading GLUE dataset loads CoLA by default","comments":"Yes so the first config is loaded by default when no `name` is supplied but for GLUE this should probably throw an error indeed.\r\n\r\nWe can probably just add an `__init__` at the top of the `class Glue(nlp.GeneratorBasedBuilder)` in the `glue.py` script which does this check:\r\n```\r\nclass Glue(nlp.GeneratorBasedBuilder):\r\n def __init__(self, *args, **kwargs):\r\n assert 'name' in kwargs and kwargs[name] is not None, \"Glue has to be called with a configuration name\"\r\n super(Glue, self).__init__(*args, **kwargs)\r\n```","body":"If I run:\r\n\r\n```python\r\ndataset = nlp.load_dataset('glue')\r\n```\r\nThe resultant dataset seems to be CoLA be default, without throwing any error. This is in contrast to calling:\r\n\r\n```python\r\nmetric = nlp.load_metric(\"glue\")\r\n```\r\nwhich throws an error telling the user that they need to specify a task in GLUE. Should the same apply for loading datasets?","comment_length":75,"text":"Loading GLUE dataset loads CoLA by default\nIf I run:\r\n\r\n```python\r\ndataset = nlp.load_dataset('glue')\r\n```\r\nThe resultant dataset seems to be CoLA be default, without throwing any error. This is in contrast to calling:\r\n\r\n```python\r\nmetric = nlp.load_metric(\"glue\")\r\n```\r\nwhich throws an error telling the user that they need to specify a task in GLUE. Should the same apply for loading datasets?\nYes so the first config is loaded by default when no `name` is supplied but for GLUE this should probably throw an error indeed.\r\n\r\nWe can probably just add an `__init__` at the top of the `class Glue(nlp.GeneratorBasedBuilder)` in the `glue.py` script which does this check:\r\n```\r\nclass Glue(nlp.GeneratorBasedBuilder):\r\n def __init__(self, *args, **kwargs):\r\n assert 'name' in kwargs and kwargs[name] is not None, \"Glue has to be called with a configuration name\"\r\n super(Glue, self).__init__(*args, **kwargs)\r\n```","embeddings":[-0.1694328934,-0.1131431013,0.1923100948,0.1447666287,0.1055869684,-0.1608262062,0.3246065676,-0.2416682988,0.4928927422,0.2975996435,0.0066172071,0.5925444365,0.0851709545,0.0806321874,0.1479725689,0.0871441364,-0.067354098,0.3852500021,-0.23223418,-0.1193545908,-0.4771755338,0.2173205316,-0.2149813771,0.2020166218,-0.2388361394,0.0284569897,-0.0420778804,0.2567212284,-0.1525195092,-0.3225887418,0.5009709597,0.4004223049,0.048006732,-0.0071245264,-0.0001277717,-0.0651666522,0.6264618039,-0.2052075714,0.0044319453,-0.1768510491,-0.579139173,-0.3032614589,0.4434992373,-0.3734254241,0.0086701922,0.5683819056,0.2329163998,-0.1811600924,0.1121183187,-0.0361293517,0.0350640565,0.2749432027,-0.3470981419,-0.1134784892,-0.1467447728,-0.0202015154,-0.0451229811,0.8158906102,-0.2364584059,-0.4017541707,0.0876230225,0.0711732581,-0.1125771478,0.2932894826,0.5203597546,0.053309124,0.2042470425,-0.3093051016,-0.0650203153,0.5827364922,0.1253231466,-0.3110702932,-0.0930168554,-0.2010276914,0.2905486822,-0.3094304502,0.3871704638,0.1643277705,-0.1318155974,0.1253418028,0.0307994727,-0.1639344394,0.1551112086,0.2616843879,-0.1995371282,0.3903371096,-0.0643599406,0.2511925101,0.156022504,0.0305497944,0.0102956006,-0.4243236482,-0.2983722985,0.0067564799,-0.10550116,-0.0589881539,-0.0094818156,0.477258116,0.0725090951,-0.1048418954,0.1459956169,-0.0078614382,0.1422875226,0.3124381602,0.1497373879,0.1795002073,0.4674269557,0.0466182716,0.1715822667,0.081407927,-0.1534212679,0.0676890388,0.0424037576,-0.0310727488,0.0605365895,-0.1745251268,0.2021006346,-0.1210566685,-0.0995601937,-0.1149303168,0.0140529908,-0.1394963861,0.0938220173,0.3139578104,-0.1833715737,0.2045091093,-0.2903826833,0.425496012,-0.2119328678,0.0371902063,-0.1563826501,-0.0690169781,-0.2381932139,0.3441038132,0.2139392048,-0.0218732934,0.2847322822,-0.1093067005,0.0480423309,0.1019340605,0.2422238886,-0.0931617916,0.1004795283,0.3126707375,0.1332187355,0.0953530222,0.0670015663,-0.5786064267,-0.2952351868,0.1635297686,-0.3343775272,-0.3553062677,0.2480690777,0.0628692061,-0.708633244,0.0345576443,-0.4189407229,0.2075431347,0.0170405768,-0.1871376634,-0.1002299637,-0.1505235881,-0.2170741707,-0.3105587661,0.1083692759,0.4598320127,-0.1007290632,-0.3130807877,-0.1742125303,-0.1292668432,0.0828600004,-0.1217455417,-0.2707915306,0.1485442072,-0.1997079998,-0.0768881589,0.6406319141,-0.3604440093,-0.3108129203,0.0916327387,-0.0039357585,0.1509331316,0.0407514088,0.1100211963,-0.0455563627,-0.0431681164,0.4982181489,0.3668849468,0.1026644483,0.0229516774,-0.0810709968,-0.2251185328,0.0689027458,-0.0152990492,0.0260320511,0.3120307028,0.0078258645,0.0009660467,0.0331266336,0.1660285294,-0.4011260271,-0.021597825,0.0519483946,-0.2292003185,0.0509481281,0.2154651582,-0.3800513148,0.2970751226,0.0964184925,0.1881465614,0.1825900525,-0.1498969048,-0.0632588044,0.0635967255,-0.1701562703,0.2489450872,-0.0731581971,0.2950997651,-0.0338743962,-0.1346231401,-0.2492021471,0.540307343,-0.3059586585,0.2340622991,-0.2367593944,0.1333027035,0.2101884633,0.0561619699,-0.2033628672,0.1293233335,0.0567595251,-0.122350499,-0.0068923528,0.4079770744,0.2588035166,-0.0673183724,-0.0300810356,-0.056121897,0.0861489326,-0.0556969047,0.0633730441,0.0601268969,0.1221924052,-0.1220948771,-0.1145645678,0.4290385544,0.1284840107,0.254658401,-0.0721655264,-0.0801663324,0.0726670325,-0.0405235142,-0.6427885294,-0.3399472833,-0.0774796456,0.1088022515,0.3928149045,0.3404849172,-0.1719955206,-0.3024609983,0.4272810519,0.1237078607,0.1707774401,-0.1698995084,-0.1243429556,-0.0518664233,0.233638823,0.4314949512,0.7753482461,0.1059535295,-0.119349733,-0.0878806338,-0.252744019,-0.1804577857,0.1540833861,0.1202417016,0.2086475492,-0.1038482413,0.225382641,-0.0485094711,-0.3323350549,0.2384047061,0.0789695233,-0.237333253,-0.3911013305,0.0130334143,0.037892051,-0.4379425645,-0.0596230589,-0.1697868258,-0.2356129885,-0.2868682146,-0.2006106824,-0.1280094832,0.0884272307,0.2005221397,0.1486909837,0.3373689651,-0.4282748699,-0.5522496104,0.060178142,-0.1075852513,-0.2194833755,-0.1756982356,-0.1533198953,0.0817978829,0.0827822983,-0.2737563252,-0.0729155093,0.2294443995,-0.3642032444,0.3472115099,0.0735491365,0.5990024805,0.2174128741,-0.1311398894,0.1475277245,-0.099021472,0.3466440439,-0.0230739973,0.1609727442,-0.1113976613,0.0236246232,-0.0147377988,-0.2133246213,0.083220005,-0.3554655313,-0.2513581216,-0.1286612749,0.2116760761,-0.0156708062,0.3157366216,-0.1783219427,-0.1097887307,0.0914314911,0.0718853697,-0.1312550902,-0.1693262905,0.2335375547,-0.0578243844,-0.1815220714,-0.1683603376,-0.4493404031,0.2137179971,0.0865022764,-0.2035409808,-0.6979241967,-0.1644452661,0.1394333541,0.1271078438,0.1680632532,-0.002659562,0.0592408292,0.1707666516,-0.1051993743,-0.4053518772,0.1375135779,0.2413716316,0.2316518575,0.2321867496,0.1489290744,-0.2921331227,0.2373870462,-0.1976976842,-0.3449137509,0.3137609065,-0.1818259358,0.7343224287,0.0804427266,-0.0825398117,-0.060086254,0.3341647685,0.0770700946,0.3019992411,0.0270407423,-0.3756336272,-0.3720228076,0.1613937616,-0.1581159234,-0.239488557,-0.1362718046,-0.2108805478,-0.0219349302,0.365162313,0.2551844418,-0.2354743034,-0.1455219388,-0.2013722509,0.516302526,-0.012953219,0.0295677632,-0.0201277267,-0.2775000334,-0.5343539119,0.2455603778,-0.146355316,0.1572123021,0.1345406473,-0.1922781318,-0.0072503155,-0.0329219066,0.4484995604,-0.4122419953,-0.0996281654,0.2113329023,0.2511311173,-0.3981359601,-0.2474796921,-0.1553961337,0.174159795,0.3201672435,0.4943858981,-0.2636165917,-0.2303315103,0.0899358764,0.1175076663,-0.1248111278,-0.4126553237,-0.6122552156,-0.2123478949,0.027997477,-0.2161383629,-0.4284980893,0.1542112082,0.113321498,0.1765402257,-0.3216019869,0.0860445052,0.3135072887,0.5341850519,0.2590313256,0.2268689424,0.0850269124,-0.1945073307,0.1815155894,-0.1779831946,0.2631509006,-0.1842771769,0.0911116004,-0.0229095444,0.1386466026,0.4430520236,0.2633058131,-0.1904398799,-0.2521757483,-0.3842644095,0.0120926322,-0.0886511803,0.4244211018,0.2282304764,0.0691799223,-0.4083652794,-0.4107196927,0.7090920806,0.2342594713,-0.3464303017,0.2325852513,0.0968148187,-0.2642559111,0.2385286242,-0.0687904879,0.821739614,-0.1572079659,0.0378649607,0.4619811773,-0.084375307,0.8485870957,0.2168823183,0.0733861402,-0.2406611592,-0.2776116431,0.0754704773,-0.2156395763,0.0073834793,0.3476289213,-0.2095759809,0.4331807494,0.1528299451,-0.0585146248,0.1437228322,0.3877589107,-0.1267999709,-0.3651953638,-0.3336216211,-0.0929782614,-0.1426859051,0.3854745328,-0.2091261297,0.0100323046,0.2175322324,-0.2441766411,0.064075917,0.0053117,-0.3357228339,-0.146776557,0.1981844157,0.042095378,-0.1830551624,0.095242545,0.0516483448,0.5375210643,-0.0338200815,0.2843132317,0.1906497478,0.0319073312,0.097601302,0.1305136234,0.1728519499,0.1522521824,0.0914795548,-0.2071563751,-0.2150570899,0.0594864786,0.0880543813,0.1970224082,0.0150062609,-0.0814502537,-0.393012017,0.1681260467,0.1702818274,0.1798896641,-0.0060126376,-0.0813502967,0.0563100688,0.3447432816,0.0487872958,-0.6621549129,-0.0542145297,0.1478797048,0.0002175454,-0.1335227042,-0.0022439761,-0.6259785891,0.0941376686,-0.0127480337,-0.0586571395,-0.3170337677,-0.0323306285,0.0398210622,0.4780427516,-0.1849782765,0.0811521709,0.1349265128,0.2699980438,-0.0187520869,0.1475009024,-0.3171009123,-0.0256846063,0.2583192289,-0.1051417887,0.1135639474,0.2958765328,-0.0457141548,0.0959492028,0.0761814043,-0.116967015,-0.0398203731,-0.0438477322,0.0436805077,0.3629110456,0.230517149,-0.0599704459,-0.0625936165,0.0021006148,-0.2558599412,-0.2553121746,-0.0671655163,-0.0285246521,0.2817902267,-0.0392715186,0.0817243606,0.0318439901,-0.320789963,-0.0137261013,-0.3447394669,0.2159488797,-0.0390916727,-0.2512979209,-0.1333875209,0.234627381,0.2099357396,-0.1666316986,0.4368416369,0.0915486217,0.1507894546,-0.0342852958,0.1278306395,-0.1758308113,0.1013853326,0.1230714172,0.053345602,-0.0466476493,0.0449270569,-0.098375231,-0.2339438349,0.1022058353,-0.1674125493,0.1383025348,0.2927308679,-0.0598358437,-0.0348771848,-0.1222479641,-0.0147356661,-0.1832727492,0.0411267802,0.2083768547,0.0906107724,0.282905221,0.5284818411,0.2885523438,0.1527935266,0.0795444623,-0.2467986494,0.4222283959,0.2285944819,0.2029920816,0.0729553252,-0.128305912,0.2980453074,0.4683024585,0.1064262465,0.058200106,0.0640890896,0.3717340827,0.280800283,0.3173823357,-0.0448246114,0.2279088199,-0.5367950797,0.1424016654,0.1578690261,-0.0403629243,0.0215254817,0.1811546534,0.2548321784,-0.5046090484,-0.1440710723,-0.2625884414,-0.17491436,0.0065932428,-0.2983582616,-0.0540819429,0.0035191451,-0.1413202435,-0.0033879736,-0.1103818789,-0.1116216704,0.2767394781,0.0200334527,0.1978140324,-0.0182344429,0.0003213403,0.2182855308,0.0053632739,-0.2908353806,0.2773526609,-0.4761333466,0.1482859254,0.0949409828,0.3791989982,0.3747552335,0.338196218,0.1080377772,-0.2391609997,0.1539077759,0.0631620586,-0.2858421803,0.0304326043,0.4256488979,-0.0481047519,0.3314395845,0.081542775,-0.0818149596,-0.050274618,0.02068251,-0.0032848353,-0.2690691352,0.3264235258,0.0490596443,0.1140367389,-0.0136946701,0.2425748259,-0.4371364713,0.2823233604,0.2107557058,0.0228382479,0.036306899,-0.199520275,0.009381175,-0.0806474686,0.4938765168,0.2020367533,0.3574380577,-0.3710487783,-0.2412195802,-0.7887631059,0.0977469161,-0.2246285677,-0.1129087359,0.1279349923,-0.0284431577,0.1264050007,0.2226118594,-0.4586646855,0.6459430456,0.0432609282,-0.3621830344,-0.2518526614,-0.2155232131,-0.1494945288,-0.3778801262,-0.1263384372,-0.5542500615,-0.1973572671,-0.3510094881,-0.0958448648,0.1185839251,-0.032318227,0.2601847649,0.4510059953,0.3935413659,0.1403742433,0.6321727633,0.0271028988,0.1805553436,0.0666143745,-0.3282628655,-0.2012848705,0.0727162957,-0.1741974652,-0.0222712234,-0.0957804471,0.3400003612,-0.0634004772,0.0831931233,0.0744192973,0.1784865409,0.1372774839,-0.1144179702,-0.3300217986,0.0174730383,0.0189866796,0.3895493746,0.098460868,0.1057167128,0.0687498748,-0.115683116,0.4444267154,-0.3427819014,-0.1141097546,-0.2470716089,0.3522530198,-0.008845306,-0.0612932295,-0.3550993502,-0.0582106374,0.3809674084,-0.254418999,-0.3067000508,0.1960640997,-0.2356848419,-0.0591070466,-0.2299783677,-0.1528535187,-0.0320820622,-0.3341134787,-0.3963763714,-0.1641500294]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/130","title":"Loading GLUE dataset loads CoLA by default","comments":"An error is raised if the sub-dataset is not specified :)\r\n```\r\nValueError: Config name is missing.\r\nPlease pick one among the available configs: ['cola', 'sst2', 'mrpc', 'qqp', 'stsb', 'mnli', 'mnli_mismatched', 'mnli_matched', 'qnli', 'rte', 'wnli', 'ax']\r\nExample of usage:\r\n\t`load_dataset('glue', 'cola')`\r\n```","body":"If I run:\r\n\r\n```python\r\ndataset = nlp.load_dataset('glue')\r\n```\r\nThe resultant dataset seems to be CoLA be default, without throwing any error. This is in contrast to calling:\r\n\r\n```python\r\nmetric = nlp.load_metric(\"glue\")\r\n```\r\nwhich throws an error telling the user that they need to specify a task in GLUE. Should the same apply for loading datasets?","comment_length":42,"text":"Loading GLUE dataset loads CoLA by default\nIf I run:\r\n\r\n```python\r\ndataset = nlp.load_dataset('glue')\r\n```\r\nThe resultant dataset seems to be CoLA be default, without throwing any error. This is in contrast to calling:\r\n\r\n```python\r\nmetric = nlp.load_metric(\"glue\")\r\n```\r\nwhich throws an error telling the user that they need to specify a task in GLUE. Should the same apply for loading datasets?\nAn error is raised if the sub-dataset is not specified :)\r\n```\r\nValueError: Config name is missing.\r\nPlease pick one among the available configs: ['cola', 'sst2', 'mrpc', 'qqp', 'stsb', 'mnli', 'mnli_mismatched', 'mnli_matched', 'qnli', 'rte', 'wnli', 'ax']\r\nExample of usage:\r\n\t`load_dataset('glue', 'cola')`\r\n```","embeddings":[-0.1234988421,-0.2316408008,0.1299612522,0.1788021624,-0.0260566305,0.0349783376,0.4479486942,-0.2386648804,0.6000370383,0.1970048845,-0.1358669847,0.5252872109,0.1504832804,0.1636874825,0.136348784,-0.0717558712,0.0006625268,0.1845783591,-0.2312728316,-0.1488693208,-0.5079345703,0.2756892443,-0.3465511799,0.1943188459,-0.1970527023,0.0579489991,-0.087690033,0.1988683343,-0.1275700033,-0.2630568445,0.5758833289,0.1909217834,0.231526792,-0.0689435974,-0.0001251231,-0.0415128693,0.5903518796,-0.2156631202,-0.0600046255,-0.2415104061,-0.6910247207,-0.3064957559,0.5161479115,-0.2989683151,0.1165501028,0.413461864,0.2769863009,-0.2617215514,0.0181751121,0.033144556,0.061910592,0.4095025063,-0.3501492739,-0.1138348579,-0.1835411638,-0.2181191593,-0.0447623059,0.8315686584,-0.141043663,-0.3725712895,0.0942742974,-0.0238960367,-0.0733997524,0.2983854711,0.4873611927,0.0610513911,0.1099266782,-0.3827915192,-0.039928399,0.4658715427,0.2723920047,-0.3111106157,-0.1319310367,-0.2014098018,0.0698197782,-0.1846679151,0.2438081205,0.2750248313,-0.0093281763,0.1062035561,0.0593750142,-0.0998046547,0.1585218608,0.2778344154,-0.2515942454,0.4114090502,-0.0927316546,0.2883330286,0.0994251817,0.0096532451,-0.0466165245,-0.4762985706,-0.2465341389,-0.038551949,-0.3014807403,0.042502448,-0.0595233031,0.3228663504,0.0277488157,0.0063540041,0.2096154392,0.0352419354,0.1282928586,0.3927739859,0.304125011,0.2012167871,0.4864655733,0.0798439384,0.0103814853,-0.1019407436,-0.0146400519,0.0106120808,-0.0358130187,0.0316042155,0.0298986714,-0.1671725065,0.0716416985,-0.0642414317,-0.3528066576,-0.1836000979,-0.0994028077,-0.1517701298,0.0608579703,0.210056603,-0.1752351224,0.2339484543,-0.3219704628,0.5184216499,-0.2057177424,0.0290766004,-0.1946338117,-0.0567978807,-0.2896016836,0.2027562559,0.1402870864,-0.0241539683,0.3323242664,-0.225214988,0.0736501217,0.1254807413,0.2058135867,-0.0744053721,0.0132715581,0.2799834013,0.3132457733,0.074169077,0.0476088338,-0.4513415694,-0.2670765817,0.2466621697,-0.3530212343,-0.3260582685,0.1507556587,0.0712697655,-0.6081245542,-0.125782758,-0.5980647206,0.2011917681,-0.0330206305,-0.0957994461,-0.125877291,-0.1952590793,-0.2787999511,-0.3485414088,0.0488444194,0.3585067987,-0.1044200361,-0.2893980145,-0.1159591973,-0.1736197919,0.1300078481,0.0689161271,-0.2402690798,0.1021810845,-0.1363590956,0.0296844188,0.5617924929,-0.4231231213,-0.3428970277,0.1687544882,0.0245062672,0.0866550356,-0.0903093442,0.1507560164,0.0957414657,-0.0008755716,0.5150431991,0.3908320069,0.1038895845,0.0325811356,-0.0275330525,-0.1966059059,0.1574236304,0.1146840379,0.048747357,0.3507826924,0.0906138048,-0.0561980754,0.0775560662,0.2598617375,-0.3014622033,-0.027242735,0.0217312295,-0.171361953,0.0665832087,0.1311919689,-0.3722841442,0.2801187634,0.1047654822,0.1512985826,0.3379476964,-0.0076634968,-0.1042713001,0.1097762957,-0.1633646339,0.195718959,-0.0947120413,0.220807448,-0.003428662,-0.1852053851,-0.2477206588,0.7136369348,-0.2071398348,0.2480587512,-0.3053863943,0.2926473618,0.1744356602,0.0416671261,-0.1472857893,0.2093914151,0.0162382126,-0.0904520825,-0.0169345699,0.4237202406,0.2346641123,-0.0982944146,0.0180235934,-0.1781243682,0.1551904529,-0.1540311277,0.1796039641,0.0569764003,0.1660908461,-0.0245173275,-0.1312085688,0.2916741967,0.1013282016,0.2943254709,-0.0451603346,-0.0105011053,-0.0684616864,0.118326731,-0.5679361224,-0.321479857,-0.0514614061,0.1028148755,0.3370561004,0.3288535774,-0.1660836339,-0.2440533787,0.369302243,-0.002039029,0.1345632374,-0.0618418343,-0.2000744194,-0.1146795005,0.2307555825,0.4542492032,0.780992806,0.1426455528,-0.0287820864,-0.1476744711,-0.1922801286,-0.2099156529,0.1177241728,0.180589959,0.0754990652,-0.1641794145,0.2235136032,-0.0315686688,-0.2807987332,0.33404392,0.0296085179,-0.1554500163,-0.3178513348,0.0475432277,-0.1117244437,-0.4101986587,0.0206767414,-0.2049304247,-0.1609475464,-0.3836710453,-0.2601151168,-0.1813657731,0.1334932595,0.1927686334,0.1012802944,0.2962879837,-0.4001510441,-0.4196333289,0.1201057658,-0.0600765683,-0.2728140354,-0.1278371066,-0.2320308536,0.2656525075,0.111680612,-0.3139145672,0.0864860266,0.1176205128,-0.278377533,0.3208665252,0.0015642951,0.5977897644,0.2011018991,-0.0882596076,0.128596127,-0.1251037866,0.3491458595,0.0017467653,0.1210567877,-0.1546037048,0.0655654967,-0.02046551,-0.03798474,-0.0115168812,-0.4604470432,-0.2555038035,-0.0506604277,0.1130422205,-0.0491565764,0.2361868918,-0.1544955224,-0.1646802276,0.0937767923,0.016965868,-0.1407371163,-0.1789999008,0.1827668697,-0.1058648154,-0.2507736683,-0.1905013323,-0.2628196776,0.2189360112,0.1154186428,-0.1900068671,-0.6095163822,-0.2136401385,0.0528340675,0.1841755211,0.1485554278,-0.0460244454,-0.1051724032,0.2209769934,-0.0925368965,-0.5195083618,0.1842223555,0.3203352988,0.1941149682,0.2446210384,0.2085207999,-0.2775198221,0.2897791862,-0.0849903449,-0.3120298684,0.3616764545,-0.1904333234,0.7031662464,0.0129345711,-0.139397949,0.0729867518,0.2007292956,0.1995618939,0.3539246023,0.0539097041,-0.4473100305,-0.331104368,0.100653775,-0.1744276434,-0.1516626477,-0.1699431092,-0.1528659165,0.0169214886,0.336558789,0.2076553702,-0.2597447932,-0.1404970437,-0.2507315874,0.5781279206,0.0702034831,0.0374327973,-0.0382413603,-0.2341484576,-0.5961073637,0.270255357,-0.1358282119,0.3426012695,0.1182705909,-0.2289820313,-0.0021415395,0.0184903499,0.3585397005,-0.4094825089,-0.0682685301,0.0962973237,0.2511437535,-0.5397173166,-0.1926564574,-0.1136872545,0.2541304529,0.2612880468,0.4388514459,-0.1797167212,-0.2849872112,0.0872998312,0.0643102676,-0.0630421191,-0.469121933,-0.6102338433,-0.1010913402,-0.064572826,-0.3660079241,-0.346842736,0.1801565886,0.2016358972,0.1500584334,-0.2731436491,-0.0162039101,0.2527001202,0.3925921619,0.2719098926,0.1817427427,0.184546113,-0.1595913023,0.2297361195,-0.0564077199,0.2807782292,-0.354215771,-0.0151209068,0.1904881895,0.1704812646,0.4698705077,0.1108675078,-0.16137366,-0.2684822083,-0.2951902747,-0.0972744152,-0.0785639659,0.3524290621,0.2164832801,0.1594810039,-0.3811665773,-0.4419957101,0.6043346524,0.2233275026,-0.2405327111,0.187607348,-0.0667403638,-0.2117961943,0.2143482566,-0.0669703856,0.5985327363,-0.1284909844,0.0916655436,0.369563818,-0.1243647188,0.6237491369,0.2063853741,0.0759058371,-0.2269101292,-0.3838103116,0.0506905466,-0.1940048039,0.0028884739,0.296800077,-0.0549072027,0.4263043702,0.1700823307,0.008350742,0.1928048283,0.3605599701,-0.0489434451,-0.3714997172,-0.2259421051,-0.052800823,-0.0827042386,0.3465657532,-0.2169736177,0.0431168787,0.1682090163,-0.1101052165,-0.0678092614,-0.0234714281,-0.2826137841,-0.1039946079,0.2303590029,-0.0313927382,-0.079272747,0.1958858371,0.0613814704,0.3434627354,0.0009841764,0.2596989274,0.0874426141,-0.0194611326,0.0819290057,0.1264114827,0.1152312309,0.1161387935,-0.0388714857,-0.2807229459,-0.1271611303,-0.0238107517,-0.025350526,0.3009958863,-0.1114900708,-0.0929967239,-0.4669117332,0.1423485279,0.1914154291,0.1634110957,0.0115316408,-0.042653162,0.1352909952,0.344422996,0.0088965567,-0.6847134233,-0.0675039217,0.1736076325,0.0209343378,-0.1076227129,-0.0012278006,-0.4774571061,0.0368442871,-0.0020239425,-0.0172328539,-0.2136154324,0.1042880863,0.0255871844,0.2290693074,-0.0791649744,0.0066934018,0.1148319244,0.2489208579,-0.150047034,0.1294598281,-0.4029648602,-0.0795957297,0.3129899204,-0.2300515622,0.0742010549,0.1916969419,-0.0439088419,0.2274306864,-0.0690702721,-0.123583883,0.0653165951,-0.0522512309,0.0089519704,0.5287432075,0.1569641829,-0.0210093148,-0.0435691513,-0.0112620117,-0.1743117273,-0.1910108179,-0.0618028268,0.0024230122,0.2551753223,0.0011204638,-0.0146853523,0.0184637159,-0.310103178,-0.174723804,-0.3459266424,0.2098681182,0.0033632405,-0.2649254799,-0.1162271127,0.1137511134,0.1385004222,-0.1187486649,0.4296660423,0.1253481507,0.1957650483,0.0459073521,0.0943782702,-0.0456265621,0.1868713945,0.0879961476,0.0450324565,0.0660444647,0.0409797877,-0.1049304381,-0.285474509,0.1505993456,-0.1932396442,0.2039231509,0.2462192923,-0.0559262484,-0.0440675765,-0.1102957949,-0.0094901528,-0.1574669927,0.0320285708,0.3289686441,0.1058955938,0.2635232508,0.4555968046,0.3814673126,0.2769702971,0.1282123029,-0.2218068093,0.4917271733,0.2261070758,0.2155487835,0.098766692,-0.1463597566,0.3802175522,0.4052405357,0.1994757503,0.0747835711,0.2466503233,0.4022906125,0.3567189872,0.4274443686,-0.0589060858,0.1925158203,-0.5156714916,0.0274077598,0.1548264474,-0.0806977898,0.0456089526,0.1194177046,0.2797237635,-0.5002767444,-0.1349911392,-0.2369687259,-0.1610701382,-0.0834138244,-0.3425095379,-0.0503795259,-0.0827671289,-0.2414561808,0.0232107136,-0.1896066368,-0.2008666396,0.2027769238,-0.1356223226,0.2215157151,-0.0626322553,-0.1490359902,0.2326163352,-0.1483922154,-0.1509531587,0.34864977,-0.5462238193,0.0526319966,0.1990113556,0.3418708146,0.3917393386,0.3730882108,0.0622666553,-0.2511269152,0.1511781961,0.0322664008,-0.2045397311,0.1231361777,0.4545445144,-0.0257386062,0.3601792455,0.083403714,-0.0871942788,0.0170110054,0.1199719012,-0.0551810041,-0.395190537,0.2881284058,-0.087728709,0.106680885,0.0497434549,0.2013334334,-0.5093894005,0.28519243,0.1834153384,0.0092073623,0.0748609826,-0.1200309098,0.011583114,-0.1305601746,0.6386980414,0.2398784459,0.4619357586,-0.2891969681,-0.1900742352,-0.7429805994,0.0833733901,-0.1862056702,-0.020553397,0.1557284147,-0.1198331267,0.1317612678,0.3032123148,-0.2857263386,0.8249191046,0.0994757786,-0.3034417033,-0.2338256985,-0.2445177883,-0.0684641525,-0.2900053263,-0.1076962575,-0.5415430069,-0.2475466579,-0.2017037421,-0.1008103192,-0.0254446864,-0.1396081448,0.1976482868,0.2801486254,0.444722712,0.1520667374,0.7404744029,0.0077932877,0.1641880721,0.0688155293,-0.4169103503,-0.1511935145,0.0532518476,-0.2265490592,-0.0972032547,-0.1434023678,0.3322530091,-0.0548000857,0.0224726703,0.1328810751,0.1327273548,0.0146934167,-0.1066536903,-0.2414740026,-0.1502703577,0.045996882,0.4398674965,0.1370453238,0.0483141877,0.1230973601,0.0523004346,0.3566411734,-0.3256351352,-0.1902870089,-0.2003345042,0.1451602131,0.0799769759,0.0199826043,-0.3159400821,-0.0165485144,0.4208723903,-0.2796034813,-0.2696284354,0.1668289304,-0.1810707599,-0.111652106,-0.1702020317,-0.1608195901,0.0220315754,-0.4292684197,-0.2138603628,-0.1999558806]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/129","title":"[Feature request] Add Google Natural Question dataset","comments":"Indeed, I think this one is almost ready cc @lhoestq ","body":"Would be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.","comment_length":10,"text":"[Feature request] Add Google Natural Question dataset\nWould be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.\nIndeed, I think this one is almost ready cc @lhoestq ","embeddings":[-0.2195455432,0.1499305815,-0.3216252029,-0.3452427983,0.0086130407,0.082326673,0.4086747766,0.1682614535,0.0568478145,0.2701212466,-0.1504386514,0.355076015,-0.300471276,0.1427343786,0.4933852255,0.0691987798,0.1555067599,0.0366067812,0.2344898582,-0.0378882363,-0.1965658814,0.1177461073,-0.1599753648,-0.1527878344,0.1038858667,-0.0813527033,-0.0448364243,0.1714063436,0.0340784192,-0.4504149258,-0.2057361007,0.193761006,-0.1478844136,0.1838628203,-0.0000995509,-0.2497134805,0.1432181001,0.0928870738,-0.4245201945,-0.0801598951,0.0140724462,-0.0226005334,-0.0507291555,0.0911515206,-0.2717536688,-0.0771518201,0.3021962941,-0.6053178906,0.0254131332,0.4482088089,0.2553070486,-0.0971344635,-0.150298059,-0.288988322,0.2953046858,0.3042961359,-0.3419626653,0.106181398,0.3485744596,-0.0624598898,0.0961427093,0.0222570579,0.016316345,-0.1584353447,0.0516601652,0.0856701061,-0.1142828614,-0.6237665415,-0.0973016471,0.1450857371,0.4909499884,-0.0973819271,-0.0529192165,-0.1794435829,0.0810113996,0.155585885,-0.1723972857,0.3338635564,-0.3287171423,0.0612178817,-0.0380047634,-0.51929003,-0.0663720816,-0.0696083009,0.1670112163,0.1441807151,0.0784827545,-0.0553014316,-0.0678143948,-0.1465065032,0.1961405128,0.0713063926,0.0140561331,0.1937569976,-0.3554562926,-0.5174686313,0.2079886794,-0.0767196417,0.3289734125,0.3620480001,-0.0185011216,0.0141245229,0.0909904987,-0.0980228558,0.3101476431,-0.4181229472,0.3670832813,-0.0051976377,0.216540128,-0.163209483,0.0014236268,0.182143271,0.0128298225,-0.0320817418,0.0507653654,0.163714394,0.3604471385,-0.0127929235,-0.0107702967,0.0132319089,0.0293630604,-0.0459672734,-0.0661774501,0.0883109495,0.0496142469,-0.2501470447,-0.2311675251,0.238502264,0.1993708611,-0.4556008875,-0.0688435808,-0.046014268,0.076422736,0.320815444,0.4330477715,-0.1929329038,0.1556885242,-0.0060110101,0.0534321219,0.0776478574,0.208035931,-0.0534464978,-0.0058471924,-0.1918716133,0.0385602079,-0.364636898,-0.0810332298,-0.1539277434,-0.228346929,0.225647673,-0.1277515143,-0.1985879093,-0.3816975355,0.2776164711,0.0123097468,-0.3487060964,0.1700978726,0.6133918762,-0.2183579803,-0.3541502655,-0.0159220304,0.3320671618,-0.2150813937,-0.1593517214,-0.0686533973,0.2601793706,-0.1652263105,0.0725153461,-0.0935814977,0.0571076535,-0.1860306561,-0.083621949,0.0667555779,0.1905135959,-0.0947520062,0.0565320142,0.3139103055,-0.1441750228,-0.0803966373,-0.0832770541,-0.4169888794,-0.2054464519,-0.2015326619,0.3464259207,0.0423047505,-0.1096167043,0.1950757056,0.5934324265,-0.2120502144,0.0249251556,-0.057229016,-0.214148581,0.1492818892,0.466230005,-0.0652576685,-0.0241303034,0.427924484,-0.316821903,-0.1987075359,-0.1232914478,-0.2518391013,0.0234934166,0.5425813198,-0.240645647,0.00866549,-0.3614624441,-0.6854640245,-0.0536889695,-0.4700409174,0.3513811529,0.2445983142,-0.2542477548,-0.2324532568,-0.1272621751,0.0132997064,-0.060193751,0.2890169919,0.093389079,-0.2169727236,0.2843669355,-0.2039668858,-0.0719564036,-0.0213607568,0.1156332567,0.0581409484,0.2255847901,-0.0420941003,-0.0688757002,0.0875929296,0.1432512254,0.1119365469,0.2200958431,0.1679893881,0.1487051994,-0.2161658108,0.2402569056,0.6387906075,0.0000026424,0.0782241002,-0.3701251149,-0.1345954686,0.225515455,-0.2021925598,-0.0157794245,-0.5144201517,0.0788861215,0.3136683702,-0.0661565587,-0.2258936912,0.1332137734,-0.0215557031,-0.1148962006,0.1433193535,0.0687787011,-0.2120974362,0.2623585165,0.1202278286,-0.258690387,-0.2521428764,0.3613994122,0.3068109155,-0.146061033,0.3606644869,0.1461704373,-0.0283351336,-0.2668964267,0.3600512147,-0.19408454,0.233622089,0.2400392741,0.179833293,0.2107689381,-0.0197899826,-0.0338124558,-0.0973325372,-0.0780251771,0.0100142341,0.0335140303,0.4484058917,0.1243027449,-0.2833541036,-0.0780121386,-0.0459358208,0.0369258225,-0.0523263551,-0.1716937721,0.0867344886,0.130508408,0.2661084831,0.355933696,-0.1286354214,-0.1764157563,0.3921942413,-0.0646345988,-0.2070232183,0.0793863609,0.1960138679,0.8522104621,-0.0276734233,-0.3615861833,-0.00778232,-0.0578414984,0.1093152463,0.3126023114,0.1878884733,0.0076489854,0.6228275299,-0.1154340655,0.1923084259,-0.5637335777,-0.3716578782,0.1653897315,0.1471683681,0.6805805564,-0.0408331789,0.027354259,-0.1241847575,0.0276194643,-0.1102003753,-0.1720717996,-0.1562938541,-0.4351576567,0.0469945669,0.1626403183,-0.1839473397,-0.5961667299,-0.1830637455,-0.3234260976,0.2282421142,0.104643181,0.1045549661,0.1330212504,-0.1033460274,-0.0427478403,0.0430314168,0.1007362232,-0.2335173637,0.0403277762,0.0650234893,-0.3776592612,0.0426739082,-0.1393127888,-0.049170807,-0.2209099531,0.0068064239,-0.2698206306,0.0476273447,0.0568588115,0.4031853676,0.0673889071,-0.1771364361,0.3521080017,0.0837517902,-0.2038389295,0.0553630665,-0.1745213866,0.091508083,0.4010717571,-0.0022932948,0.2771012485,0.0003530251,0.0696001798,0.6412567496,-0.090214245,0.1398363113,0.0286024436,-0.2327840477,0.2266936004,-0.0395933054,-0.0410326049,0.2653284371,0.1177358255,-0.0731708035,0.3487163186,0.344632864,-0.5520005226,-0.0875098705,-0.0392642319,-0.1995066553,-0.0173002724,-0.0461302809,0.036166966,0.1053824201,-0.0693519861,-0.295322746,-0.2081553191,-0.3496506214,0.0233487654,-0.0004418762,0.2460575402,-0.1142508984,-0.7296465039,0.3268490732,-0.1920619756,0.2020509392,0.3000995219,-0.1893317401,-0.2315276265,-0.160863027,-0.1375146955,0.0222192574,-0.0221524518,-0.5022993088,-0.254904598,-0.0745976195,0.3034452796,0.1777596623,0.1673330516,0.0116278762,0.066146411,0.1158282384,0.3427494168,-0.1517309397,-0.0239003599,0.4543728828,0.0643576458,0.0524737202,-0.162406534,0.1468925625,-0.1401303858,-0.0416170917,0.1657700688,0.1425540298,0.0109942937,-0.0016255032,-0.2934008837,0.0505453236,0.0867887884,0.206217438,-0.0379454195,-0.0283478368,0.2086362094,-0.0130199734,-0.0014980227,-0.1814134121,-0.1667376906,0.3786190152,-0.0798447281,-0.0412881151,0.1678691208,-0.2593721449,0.383823365,0.2300424725,0.2830732167,0.0616598316,-0.135128051,0.0626456141,-0.0425701663,0.1051696166,0.0652175024,-0.2067465633,-0.1279575378,-0.2844725549,0.4317781031,0.0016295934,-0.2735476792,0.2773709893,0.0062215114,0.0370880403,-0.1428923905,0.1719768047,0.6904017329,0.0018893197,0.0117170457,0.1127387658,-0.28648296,0.3095478117,-0.2371689379,0.0903158337,-0.3876378238,-0.1118871719,0.1067608446,-0.1060188785,-0.0359562412,0.2487955987,-0.2143115252,0.2115284503,0.1136755198,0.4631509483,0.3249306679,0.5195552111,0.1012013778,-0.0264239274,0.281889081,0.3316406608,-0.0071582692,0.0409275554,-0.1324460953,-0.0608283505,0.0921340212,0.0501005426,-0.2557135522,0.1433587223,0.0589564852,-0.1693807095,0.1304886788,-0.367185384,0.1296345443,0.3258569837,0.3113828003,0.3057224154,-0.1727153659,-0.0213274993,0.0384955816,0.2396238446,0.0865685418,0.0020282338,0.2759577632,-0.2259387523,-0.2618417442,0.2119627148,0.0606510788,-0.4469602704,-0.7831354141,0.1457665563,-0.1211323738,-0.161846891,0.117284365,0.0761859864,0.0979178473,-0.2171312869,0.2314415425,-0.032140363,0.2518272698,-0.2203581184,0.3276579976,0.0261983685,-0.2047995329,-0.2207152247,0.2016609013,-0.3431309164,0.0139536355,-0.0001723166,-0.0773770958,-0.5084995627,0.0727636144,-0.0498543456,-0.1023441851,0.115000993,-0.2363931686,-0.3069461286,0.0232241936,-0.1523042172,-0.1256197244,0.0485619344,0.0874010697,0.0346143469,-0.0184837207,-0.0066573056,-0.197661981,0.2097932547,0.1737775952,-0.2796476185,0.0994764343,0.0182349496,-0.4733228683,0.0329716094,-0.2294195592,0.0221200231,-0.0958972424,-0.0910790637,-0.1366851628,-0.146274671,0.1474505514,-0.4785277247,-0.0633867458,-0.3253521323,0.0999032184,0.0373624898,-0.0380679071,0.2740706801,-0.1078775004,-0.1169020161,-0.026409803,0.0424542576,0.1014011353,-0.1560594141,0.0626463741,0.0654221922,0.1205515042,-0.0407277495,0.1319249272,0.2629074454,-0.015074674,0.0018349169,-0.0898675621,0.1405021548,-0.0032107211,-0.0523937158,-0.0617310293,0.3776588142,0.4259517491,0.1561464071,-0.0076075573,-0.13975811,-0.233479768,0.0996391848,-0.1802663505,-0.2448442131,-0.3768388331,0.3652623594,0.1851318628,0.3409548998,-0.1657432616,-0.1396816522,0.2121840715,-0.0860356465,0.2010423094,0.2082812637,0.160802722,-0.0486686341,0.5175109506,0.1215657145,0.2700885832,-0.0338108055,0.157461822,-0.2158748209,-0.0155717256,-0.1141371056,0.208940804,-0.0308563281,-0.0973149315,0.0391784012,0.0416607261,0.073468864,0.1533000022,0.0135783842,0.4659326077,0.2916106582,0.2850903571,0.2747941911,0.0440253578,0.0346821435,0.3668931127,0.6151239872,-0.2032653838,-0.2772495449,-0.1381533593,0.1938444674,0.0116229709,0.1506089121,-0.1385049075,-0.1506204754,0.0814506412,0.0065889144,-0.0527981967,-0.4704925418,-0.2042126358,-0.1064030975,0.1052564681,0.2935042977,0.3478497267,0.2246894985,0.0768481046,-0.0309607256,-0.0152955651,-0.226500988,0.0615942031,0.1443723738,0.291403085,-0.1389779449,0.1323116869,-0.4571181238,-0.0492315292,-0.1335261017,-0.1361500174,0.0151949693,-0.1487470567,-0.0232613739,-0.2254363,0.2779797912,0.2979105115,-0.2051640004,0.3004294634,0.210325703,0.0147823878,-0.2990427613,-0.0571686812,-0.0084872628,-0.1583294421,0.1287121624,-0.0589757524,-0.0690144375,-0.2339840978,0.1484526694,-0.1381042451,0.0390410386,-0.0523348078,0.173126772,0.0441943593,-0.0328479856,0.1069332287,0.219512701,0.1665666848,-0.1593213528,-0.3635842502,0.0813682452,0.1478851438,0.0806486979,0.1323091388,-0.0152659975,0.0436465293,0.3653179109,-0.1468109637,-0.2758375406,-0.2343283743,-0.0709410384,-0.4580126703,0.0763198957,-0.2786455154,0.2380641997,0.092924282,-0.1179995313,0.0970254764,-0.1727838218,0.1100055426,0.0030354098,-0.0276145954,0.0010831915,-0.216470167,0.2851892114,-0.1282620728,0.3956661224,-0.1626134962,-0.1967356503,0.0075123576,-0.2292032093,-0.1228288859,-0.1429303885,0.0739279613,0.2861866653,-0.0382130593,-0.18700023,0.3473990858,0.2301069349,0.2833473086,0.3080895245,-0.2257643789,0.1057095602,0.1948071718,0.1060007885,-0.0684532225,-0.0634011403,-0.1667783558,-0.1145764068,0.0053930916,-0.0458499715,0.3459529877,0.0767225251,-0.1860336512,-0.1804625839,0.0482382216,0.2191384882,0.0058687469,-0.032793805,-0.2193779796,0.2676590979,0.0487740971,0.1519828439,0.3244445026,-0.0804462209,0.0505191162,-0.1965596527,-0.2168501765,0.1094829738,-0.0175508726,-0.2583505511,-0.3284038901]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/129","title":"[Feature request] Add Google Natural Question dataset","comments":"I'm doing the latest adjustments to make the processing of the dataset run on Dataflow","body":"Would be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.","comment_length":15,"text":"[Feature request] Add Google Natural Question dataset\nWould be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.\nI'm doing the latest adjustments to make the processing of the dataset run on Dataflow","embeddings":[-0.2179996967,0.1545338035,-0.2591143548,-0.3583598733,-0.0440472923,0.0838627815,0.3318873346,0.0975727439,0.0507082716,0.1603799462,-0.1157890558,0.3596675694,-0.3388666213,0.2146274149,0.4577982724,-0.0239018425,0.1525196284,0.0205585789,0.2067489624,-0.1588350236,-0.2168020457,0.0686258078,-0.2159248441,-0.281272918,0.240082249,-0.1803402454,-0.0219869204,0.1209494099,-0.036586985,-0.3671535552,-0.1910734326,0.2446611971,-0.0824514925,0.2459131479,-0.0001080934,-0.2357440889,0.2440791428,0.0615384951,-0.3804223835,-0.1635967493,0.0016236238,-0.0495487489,-0.0880728811,0.1411439776,-0.1709051728,-0.1036372483,0.3383852541,-0.6215036511,0.0559730083,0.5883984566,0.1537480801,-0.0524607897,-0.1026437357,-0.2149047703,0.2332810014,0.3189528286,-0.3424163163,0.1812490076,0.4033689499,-0.0424945951,0.1822645515,0.0149612986,0.0099131316,-0.1554568261,0.1342455745,0.0279845111,-0.2061950862,-0.6845499277,-0.0862463713,0.2037118673,0.5091105103,-0.0882809535,-0.160022825,-0.1470947713,0.1495689601,0.0405366272,-0.1540489644,0.3040738702,-0.3751213253,-0.0042525111,-0.1231859401,-0.6675485373,-0.1362742782,-0.0005686472,0.1819574982,0.1631731987,0.0971654356,0.0468151867,-0.1563278139,-0.0351237282,0.3182655275,0.0226872396,0.058690723,0.1606395692,-0.5331799388,-0.3885344565,0.1903072298,-0.0767762437,0.2892724276,0.372569859,-0.0351594798,-0.0299092587,0.0394708812,-0.1323409528,0.1896201819,-0.3920537531,0.3275400698,0.04526373,0.1606197208,-0.2197638005,0.0696920529,0.0826804191,0.0034183527,-0.0181928612,-0.1106972471,0.0449822061,0.4899743497,0.0307568312,0.0101270685,0.0249524601,0.1298995763,-0.1274974793,-0.0474503711,0.0378808863,0.0811309665,-0.2980341613,-0.1868679374,0.359562099,0.2587091327,-0.4306229949,-0.0884071216,-0.086670652,0.1312408149,0.2118064612,0.3744327426,-0.242779538,0.1817566901,0.0544743985,-0.0396250524,0.2892996073,0.2402534187,-0.0690466464,-0.0688710883,-0.1048531756,0.0347611606,-0.2990918159,-0.0419084914,0.0453434065,-0.179396376,0.2066881955,-0.1867924482,-0.2676823139,-0.4045217335,0.1823180318,-0.1011939049,-0.4326204062,0.1003653184,0.5696130991,-0.2715473473,-0.3116195202,-0.0423158184,0.2220230848,-0.2834436893,-0.1569850445,-0.0473171957,0.2369686216,-0.3297266662,0.1770819724,-0.1558014452,0.139089033,-0.0162997842,-0.051361464,0.1151800826,0.2450859994,-0.0810756758,0.0059431638,0.4584914744,-0.1925172806,-0.1262059808,0.0137925474,-0.4337471724,-0.247739017,-0.2781277299,0.3208724856,0.0098593608,0.0263517555,0.2498120964,0.6514527798,-0.2508278191,-0.1044783443,-0.0296709146,-0.2870114744,0.2383397222,0.4557878971,-0.0139175504,0.0840974748,0.3782624006,-0.4788369238,-0.280058831,-0.1185711846,-0.2246852815,0.0147780217,0.6001603603,-0.0999112129,-0.0110722259,-0.36136204,-0.6427802444,0.0744694322,-0.4407260418,0.4363308251,0.2642551064,-0.1636290103,-0.2012892812,-0.0907096937,-0.033664193,-0.036255423,0.1844715029,0.1235721335,-0.2098376751,0.2629602849,-0.3348825872,-0.0694257766,-0.1860673875,0.1632377654,-0.0161107704,0.3075799644,0.0179838445,-0.0714886561,0.0329025537,0.0870356038,0.1064238921,0.1748946011,0.1053918973,0.15018031,-0.377468437,0.3212545514,0.6691417694,-0.0655392781,0.0157480538,-0.2589434683,-0.0599770099,0.2474687099,-0.1885609776,-0.0294217877,-0.5087790489,0.0234515816,0.2398290634,-0.0517516024,-0.176521942,0.1813620031,-0.0003734985,-0.0226567052,0.1805463284,0.1952475905,-0.1604543626,0.1958818585,0.2460069656,-0.3380490541,-0.3136392534,0.28199175,0.4085603952,-0.0943176299,0.3153499663,0.134929195,-0.056008473,-0.2883775234,0.3818978667,-0.0849742293,0.3509969115,0.2542995214,0.1780038476,0.2163211554,-0.0229379237,-0.0532630458,-0.0779515728,-0.0307672881,0.0335207991,0.0682789311,0.5030938387,0.1904067993,-0.3470931351,-0.0917151719,0.064971596,0.0983365551,-0.028838845,-0.1813691258,0.0814923495,0.0590499043,0.255990386,0.4688203931,-0.1806893796,-0.1826143712,0.3295137882,-0.0610258207,-0.1620487273,0.1303142011,0.1461297423,0.8277571797,0.0707886443,-0.5137224197,-0.0842678547,0.1070977598,0.0156783517,0.2106571645,0.206618309,0.0237875413,0.6294592619,0.0164429098,0.1977297366,-0.6266110539,-0.3953923583,0.1118383557,0.0588971004,0.7985547781,-0.0070765689,0.1157534495,-0.0317719691,0.0261901449,-0.0860108733,-0.2356061488,-0.168840915,-0.4003618062,-0.0413089022,0.163316384,-0.1245256066,-0.6447999477,-0.2203363627,-0.3174123168,0.3182625473,0.1032394096,0.0519657768,0.2328984141,-0.1603973061,-0.0198078435,0.1189898401,0.0814796388,-0.1791548431,0.0113397827,0.1251571178,-0.2787204981,0.0064424262,-0.0865782276,-0.0221536867,-0.1509017795,0.106171146,-0.3694597185,0.0612712614,-0.0427647308,0.4555973113,0.0943543166,-0.1156546026,0.4253980815,-0.0499399789,-0.0687067509,0.0140993735,-0.1753053963,0.1014819071,0.4894390404,-0.0025048184,0.3873247206,0.1992509961,0.0074785058,0.677937746,-0.156603083,0.1536458582,0.0921955332,-0.3264673948,0.3406627178,-0.0661818981,-0.2011291534,0.3574688137,0.1512657404,-0.0800447762,0.2300456911,0.2714007199,-0.5368145108,-0.052909825,-0.0592689551,-0.1194605827,0.0890736729,-0.1037509888,0.0625391155,0.118818827,0.0511323959,-0.28390944,-0.2373143435,-0.3555634618,-0.122348845,0.0093541481,0.2872795463,-0.0781408697,-0.7651305795,0.359400481,-0.3864434361,0.2735097706,0.2004330158,-0.0341601409,-0.3089960814,0.0104332296,-0.0728665516,0.0233815927,0.1088576019,-0.7435159683,-0.3173729777,-0.0470336303,0.379866302,0.2106955647,0.1210449561,-0.0211904626,0.230088681,0.2283167839,0.4277050197,-0.0412525125,-0.0614730008,0.4541471899,0.094013527,0.1008674204,-0.2045516223,0.0823221505,-0.1993183047,-0.0525046922,0.1381705403,0.1436239183,0.0080957683,-0.0564959757,-0.3190083504,-0.0096270721,0.1351199746,0.2639384568,-0.0880054981,-0.1218074188,0.2053897083,-0.04350302,-0.0045568356,-0.2350113392,-0.1898552626,0.3567549288,0.0154573154,-0.2365453839,0.1835369766,-0.1939477175,0.4100670815,0.3064618111,0.2238501012,0.0753097087,-0.0660401508,0.1068545058,-0.0242488403,0.1710259169,0.1663762182,-0.153086409,-0.1800569296,-0.3547867835,0.3473017812,0.0010766205,-0.2488620132,0.4459112883,0.0999012738,-0.0010695551,-0.1737370789,0.2049194574,0.7647126317,-0.0318128318,-0.0372272395,0.2310293168,-0.2539766431,0.4919417202,-0.2764192224,0.1125659943,-0.2762091458,-0.2386709303,0.0321511589,-0.1241957769,0.0398683622,0.2155921161,-0.1197409481,0.2994552553,0.2297387123,0.5279407501,0.420941323,0.6744312048,0.1046165526,0.0392391458,0.3217152357,0.232869342,0.0065888958,0.0733880699,-0.1549262851,-0.1483414471,0.1325162053,-0.0436155088,-0.323418498,0.1186518446,-0.0948146656,-0.1743389815,0.029662706,-0.3824208975,0.2388327569,0.4224671125,0.3599102199,0.3088006377,-0.2793653607,0.0617832951,0.0686683357,0.1519653648,0.0829939768,-0.0520386249,0.203589201,-0.2492265552,-0.2949021459,0.1584331691,0.0195121672,-0.5485479832,-0.7936475277,0.1507455558,-0.1129564345,-0.1328016222,-0.0267982557,0.0848262459,0.1079034731,-0.1453471929,0.1471416056,-0.0510998666,0.2705442011,-0.2581840456,0.2887170613,0.0237514284,-0.2000342757,-0.3065992594,0.1542833,-0.4896828234,0.1349181235,0.0097950455,-0.0338865556,-0.3987733424,-0.0079304418,-0.0043746796,-0.1455568671,0.0442974754,-0.228374362,-0.2362195253,0.0047270269,-0.1531527787,-0.2126428336,-0.0249182228,0.0877569988,-0.0873413682,0.0419255681,0.0112740472,-0.2709859908,0.1651536077,0.1931200922,-0.3436056077,0.1802379787,-0.0400638916,-0.3642570972,0.0264419075,-0.1578775495,-0.0312927775,-0.0793689042,-0.0743839219,-0.0955175534,-0.1609707475,0.0522025935,-0.516296804,-0.0794346929,-0.21946536,0.2099575996,0.0983298048,-0.0142110204,0.146991998,-0.1046954617,-0.246967867,-0.1630811095,0.0923242569,0.0790068284,-0.2398420721,-0.0284243841,0.0531516038,0.1522846818,0.0873106122,0.1047477573,0.2738344073,0.0261496995,-0.0690003783,-0.146511212,0.2446545959,0.1240958497,-0.0305383895,-0.1977203637,0.2834594846,0.3672688603,0.238955304,0.0419480689,-0.2664948106,-0.1760826558,0.0367882363,-0.099167034,-0.1723373532,-0.3390679657,0.3538172543,0.2241566777,0.2358088493,-0.2959581614,-0.1317160875,0.3385924697,-0.012547127,0.1669308841,0.1488070339,0.1741617322,0.0861782953,0.4112332165,0.1512111127,0.2925000191,-0.0992680788,0.2323456854,-0.1876226217,-0.0580708832,-0.1015321612,0.2011732459,0.0388407744,-0.2796900868,0.2108698487,0.0624837428,0.0765061826,0.2616155148,-0.0553315319,0.4540587068,0.1193418279,0.2365199775,0.2464748621,0.0924037322,0.1790685803,0.3107661009,0.4752652347,-0.1943520457,-0.2687090635,-0.1582583487,0.1198563278,-0.0305504147,0.1915524006,-0.1853521019,-0.1351974756,0.1379520744,0.0566713586,-0.002815231,-0.4687034786,-0.3783359528,-0.1206032187,0.0626652986,0.2460693419,0.4210193753,0.2411831319,0.043712806,-0.0312653929,0.0838015601,-0.2139580846,0.0893021449,0.1483651847,0.2698636949,-0.0834606439,0.2029880732,-0.4250545502,0.0824699178,-0.0901706219,-0.1194617227,0.0214345884,0.0033260456,0.0527272411,-0.0991140306,0.2240280956,0.2158416063,-0.1296737641,0.3890728056,0.1564440578,0.0467368737,-0.432469219,0.0201105271,-0.0380898118,-0.1201017722,0.0847479627,-0.0681646913,-0.0100461133,-0.1643251032,0.2975455821,-0.0769889131,0.0077375607,-0.0248868149,0.1326921731,-0.0933398977,-0.1064783037,0.1598197967,0.365242511,0.137995854,-0.1553000808,-0.4534423053,0.0180766415,0.2144034654,0.0501360074,0.1882308275,0.0366938561,0.1256079823,0.248039946,-0.1576543748,-0.3140295744,-0.2912433743,-0.0613763928,-0.401552856,0.0880080983,-0.276737988,0.3210288286,0.1032246202,-0.1532758623,0.1161855236,-0.1099915281,-0.0095861852,-0.0143044405,-0.0145053994,-0.0055839163,-0.1852730364,0.2004684955,-0.0607065596,0.4985478818,-0.1191785112,-0.1407915652,-0.0996492878,-0.1834765077,-0.2575906515,-0.0775332004,-0.0232282449,0.3223912418,-0.0498257689,-0.2262111604,0.2259539515,0.2891069055,0.2565287352,0.2309908569,-0.2826024294,0.0998661071,0.0369680971,0.2190325707,-0.0565354116,-0.1476031393,-0.0768844783,-0.1402772814,-0.0242876206,0.010814785,0.4033837914,-0.0752706602,-0.2677646577,-0.1799449325,0.0240620896,0.3526474833,0.0503216311,-0.1502027363,-0.2318199724,0.2983199656,0.0692803636,0.183968097,0.3835159242,-0.0705052316,0.0961417034,-0.1214161441,-0.2192945033,0.0007252392,-0.1108163521,-0.2532394528,-0.2919079065]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/129","title":"[Feature request] Add Google Natural Question dataset","comments":"Is there an update to this? It will be very beneficial for the QA community!","body":"Would be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.","comment_length":15,"text":"[Feature request] Add Google Natural Question dataset\nWould be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.\nIs there an update to this? It will be very beneficial for the QA community!","embeddings":[-0.1010675654,0.0765452012,-0.2365692705,-0.3144036829,-0.0334924869,-0.0076650172,0.2399665266,0.176968649,-0.1138928309,0.1455584317,-0.0857489184,0.2047868073,-0.1504556984,0.1042284444,0.4593848884,0.0886675492,0.2996316254,-0.068120636,0.2519078553,-0.1587471068,-0.2599771619,0.2212228179,-0.387586534,-0.2592709661,0.1899092495,-0.054177694,-0.1278436929,-0.0491153933,0.1351170689,-0.4108201265,-0.1453486234,0.2697668076,-0.1087826937,0.2635703087,-0.0001219884,-0.4112681746,0.2723810375,0.0837880969,-0.2601504624,0.0351123363,-0.1126703471,-0.1927349269,-0.0816436708,0.2262728959,-0.0328740738,-0.0935305059,0.3543232977,-0.5727344155,-0.1994510591,0.3950356841,0.0527461804,-0.1526897848,0.0650295913,-0.1869562119,0.259344846,0.3684511483,-0.2592239976,0.1567201614,0.5362657309,-0.0450151786,0.3531877995,-0.0232208669,-0.0001043734,-0.255238533,0.2147888392,0.1287193447,-0.1948243678,-0.5777590871,0.0234453157,0.1271579713,0.6865088344,0.0089062667,-0.2538946271,0.010260581,0.163433522,0.0885295123,-0.2201989293,0.2475360185,-0.3156529069,-0.0901514068,-0.0217820667,-0.8032174706,-0.1340380311,0.0160401445,0.2142431438,0.196440205,0.1881554574,0.0180549566,-0.1540137082,-0.0830362439,0.2173684984,0.144208312,-0.0094360253,0.2730645239,-0.5609337091,-0.5661659837,0.1351905763,-0.1371896565,0.2753344774,0.3951441348,0.058884345,0.0617923662,-0.0386122614,-0.2027409673,0.2873837054,-0.2782306373,0.4319748282,0.0388146415,0.2748508155,-0.1480566114,0.0005507009,0.1165650412,0.0623143613,-0.0904476568,-0.0047270693,0.0777924061,0.5833980441,-0.0655112937,0.0262869559,0.0375685729,-0.0521974415,-0.2313048244,-0.3605746627,-0.2029981613,0.15489842,-0.1666064709,-0.2400586158,0.1613769829,0.2498717755,-0.5339114666,-0.0208050497,-0.2312472165,0.1131042838,0.3698703945,0.1373325139,-0.2452622801,0.0163330287,-0.0755158141,0.0041882689,0.2474119216,-0.0077148527,-0.088051945,-0.1854547262,-0.0689413697,-0.2075314522,-0.3263189495,-0.1206215024,0.0509124249,-0.2388976365,0.3113809228,-0.1544907838,-0.2129653841,-0.4981547594,0.0496327542,0.1611327678,-0.4291840196,0.0909919143,0.6563743353,-0.2078250349,-0.4771808684,0.0107139265,0.2077094316,-0.30487445,-0.218168065,-0.0896556005,0.1684746593,-0.4043701589,0.0999798477,-0.1423337609,0.161483705,-0.2107609063,0.1126692817,0.0202970374,0.0953067914,-0.048475042,-0.1073002443,0.3588015437,-0.2646079659,-0.2147095501,-0.049946744,-0.2568301857,-0.3506476879,-0.3096123338,0.4229902029,-0.110302031,-0.0689715147,0.1523991972,0.5050196052,-0.1757497638,-0.1815144271,-0.0510696396,-0.3274396658,0.3543447256,0.3331774175,-0.0206866562,0.1855437756,0.47288993,-0.3912593722,-0.0438151285,0.1368673295,-0.2004066557,0.0549577363,0.7342354059,-0.083897166,0.0736362264,-0.2963742316,-0.672206223,0.0509146005,-0.5351147056,0.4320774078,0.266994983,-0.180907011,-0.0721710399,0.0019979409,0.1379591078,-0.2491020262,0.0125830993,0.0479104258,-0.2240985036,0.3084010482,-0.3176258206,-0.0122633576,-0.0032180022,0.0531838126,0.0006056227,0.3049460351,0.0666741654,-0.0303492844,-0.0553496815,0.2117913067,0.1433418989,0.1926550567,0.0898464322,0.1413283497,-0.3195588887,0.2454628795,0.7631553411,-0.0666939691,0.1328092217,-0.3938724399,-0.0363429114,0.2106523067,-0.1960317343,-0.0354942158,-0.3586273789,-0.1216646135,0.1773193777,-0.0246565994,-0.3588504791,0.0700533316,-0.0974456519,0.0825232416,0.1346327662,0.3190042078,-0.3979439139,0.2564481199,0.2365657836,-0.3167081475,-0.2337542176,0.3793131411,0.3472525179,-0.0808475316,0.0851756036,0.2252023071,0.0106021697,-0.2767315507,0.3379987776,-0.1156407148,0.4185311794,0.1121404469,0.013811714,0.1742069572,0.1718930155,0.0012921088,0.0073250672,0.0227481183,-0.0783609822,-0.0673502088,0.5299584866,0.2391272336,-0.1240775511,0.0544634387,0.0314134285,0.0735067055,0.0391697921,-0.1800275296,0.0255431365,0.0338046625,0.1801949441,0.2936970294,-0.3189242184,-0.272395581,0.1620416939,-0.1218972579,-0.0224347729,0.2279008329,0.2456881702,0.8468668461,0.0396012887,-0.4998513162,0.0935306102,0.2211406231,0.1367274523,0.088802062,0.0800238773,-0.0806985945,0.3972251415,-0.0224493407,0.3097508252,-0.7418852448,-0.4688921273,-0.0235529114,0.0321022607,0.7047140598,-0.0092732562,-0.0286057591,-0.0523764007,0.0885889307,-0.1691515446,-0.3356925845,-0.2078730762,-0.4979295433,0.0407531857,0.1285530329,-0.038403552,-0.5499298573,-0.3428254724,-0.2109512687,0.217997551,-0.0078427121,0.0879272744,0.3036643267,-0.1053264067,-0.1120751128,-0.0006804232,-0.1389132291,-0.1123749465,0.0112472894,0.1134038568,-0.0985508934,0.0861797705,0.0697232336,0.1470580697,-0.0609030426,0.0186019465,-0.4702960253,0.1178056672,0.0311567858,0.5400927663,0.0600467734,-0.2139683515,0.5682368279,-0.0940685272,-0.0377454646,0.1168287918,-0.3363195062,0.0710854977,0.551961422,0.1305512637,0.3927941918,0.1162196621,-0.0395122319,0.7269040346,-0.0608605258,0.1465231627,0.0472224504,-0.1969857067,0.4011974931,0.0370641835,-0.231480673,0.2062162161,0.2081607878,-0.0526549891,0.1756215841,0.2865813375,-0.4918379784,-0.1091481,-0.1102547571,-0.015647145,0.0878659412,-0.1488128603,0.045328714,0.2233940363,0.1012084633,-0.2681474984,-0.2615084946,-0.3988901377,-0.1148633361,0.0535774268,0.3183894157,-0.0703593343,-0.7360619903,0.2618208528,-0.2096351087,0.4035932422,0.2142528445,0.1140531823,-0.247589469,0.0345165282,0.0434571058,0.1482781023,0.0000763075,-0.5058178902,-0.3511660993,0.1138740182,0.3284647465,0.3696104884,0.0137452018,0.2555041909,0.0800401494,0.1323554218,0.348123461,0.055831857,-0.1636172533,0.4706936181,0.0919450969,-0.0003536604,-0.0361462273,0.045119375,0.0044657765,0.1937972009,0.0771085247,0.0034981458,-0.0636682734,-0.0759742856,-0.2435655296,0.0828849748,-0.0252533983,0.192877546,-0.257181704,-0.004938052,0.2136623114,-0.0649792552,0.0782649443,-0.3720301688,-0.1322643608,0.2277768552,-0.0561487712,-0.1660398841,0.1050819457,-0.2521620393,0.5323531032,0.3740146458,0.1936174929,0.2492868751,-0.1138674989,0.0491899773,0.1181877181,-0.1212639064,0.295214206,-0.1891121715,-0.1116654873,-0.3006031811,0.3440159261,-0.0939860046,-0.3023914993,0.3865417838,0.1408640593,0.073380813,-0.3040342629,0.1384289116,0.8610535264,0.2466318756,0.1106521785,-0.0271276627,-0.3660755455,0.2948797643,-0.1601068974,0.2104753405,-0.3451253772,-0.1427899599,0.005568211,-0.0996798426,0.0845232755,0.062920548,-0.1844354272,0.3384518921,0.1117875576,0.5297511816,0.3909307122,0.3912557065,0.1349047869,0.2777563334,0.5042269826,0.1204475313,-0.041464828,0.2043401152,-0.138385877,-0.1088856161,0.1456083059,0.0011259292,-0.4926309586,0.088307783,0.1094215289,-0.2061664313,0.2410414815,-0.4640955031,0.2982094884,0.4526820183,0.3846502304,0.1924867332,-0.2723700106,0.1175958738,0.0731045604,0.3343651891,0.0479323454,-0.0401373692,0.1369674057,-0.1125797406,-0.4219459593,0.2949152291,0.0147395385,-0.6358892918,-1.0165194273,0.2444266677,-0.1498328,-0.1326871216,0.1795390099,0.1717730165,0.2637260258,-0.0975048989,0.0477506332,0.1164508536,0.3188813627,-0.214798525,0.2866464257,0.004915969,-0.0391651243,-0.1206632182,0.2632009983,-0.5288813114,0.207372725,-0.068442747,0.0705850422,-0.2660738826,0.1366051137,-0.2846512794,-0.2107201964,0.0983410999,-0.3333328068,-0.0915434435,0.0497766435,-0.0269080717,-0.3213665187,0.0466650799,-0.0064007388,-0.0800089762,0.0508830026,0.112494424,-0.4528143108,0.1340732276,0.0161151513,-0.4492944479,0.1828424782,0.081458956,-0.1940292269,0.219525069,0.0021794292,-0.0768840834,-0.1843096167,0.0074381265,-0.1483889967,-0.24266693,-0.1231933758,-0.5521556735,-0.0266625788,-0.0577046871,0.1885733306,0.1945190132,-0.0436751992,0.2304674536,-0.1601957977,-0.2774597108,-0.033650104,0.190666005,-0.0059324619,-0.2338650376,0.0705270842,0.1639136374,0.3055597842,0.0679854453,0.1288475096,0.4594943821,0.0921664461,0.1367528886,0.007945505,0.1478838623,0.2071207017,0.0092025958,-0.2072319239,0.4453361928,0.240385592,0.2153585255,0.0808959529,-0.2756208777,-0.1466834098,-0.0373485796,-0.1638063192,-0.1801985055,-0.2928020656,0.2480837107,0.1187405586,0.0763487518,-0.1861695945,-0.1179122552,0.3463779092,0.1130429581,-0.1407885402,0.0564460158,0.2777611017,0.0275907256,0.7248439789,0.2056929618,0.2650901079,-0.0912509412,0.326394856,-0.097755909,0.0378387533,-0.120978348,0.2949841321,-0.0723655894,-0.3227367103,0.2190043032,0.0791238993,-0.0132637406,0.2806507349,-0.071889326,0.5289364457,0.2781876624,0.2434257716,0.3552532196,0.1055455431,-0.060404148,0.4066284299,0.5845900178,-0.1426929832,-0.2211654037,-0.1213170066,0.1747686416,0.0297350381,0.1372907609,-0.0133262277,-0.1637679338,0.1272211224,-0.0108121047,0.025235435,-0.382505089,-0.2431117594,-0.1528201103,-0.058535181,0.2119433582,0.4083942473,0.1436356604,0.1509049386,0.0349616297,-0.0681974143,-0.230273366,0.0862208083,0.3404197693,0.4625248015,-0.2864993513,0.1518698037,-0.4861615598,-0.0235120468,-0.1425990462,-0.0433137827,0.2357735485,0.0593087003,0.0015687319,-0.1434129477,0.0546201877,0.0887996927,0.0147143165,0.5427064896,0.222540915,0.1482056081,-0.533878088,0.1628819257,-0.2310628444,-0.2537183762,0.0769824088,0.0893983543,0.1070019081,-0.1366950572,0.2034406364,-0.0969915166,-0.1342535615,-0.0322636887,0.0509794578,-0.0662144348,-0.064380303,0.1991900951,0.1647539586,0.2006705254,-0.252098918,-0.425514251,0.0671986043,0.4483526349,0.2352419943,0.0712155253,0.0313500576,0.0281469468,0.2623105943,0.0560842156,-0.3688796461,-0.3364644945,-0.0484514087,-0.4906277955,0.0678211525,-0.3346514106,0.294066757,-0.0483367331,-0.0301254727,0.1512941718,-0.2186060697,-0.1427232027,-0.0034486577,-0.145679459,0.0773520768,-0.03115822,0.2697182596,-0.1271001399,0.3465590179,0.0673134923,-0.2953623235,-0.19350712,-0.3479121029,-0.1002618521,-0.0240911413,-0.0751576796,0.2864362597,-0.1406211257,-0.194703117,0.2435571253,0.4499562681,0.0641725361,0.2257223576,-0.3899097443,0.0346183367,0.1383999586,-0.0323022678,0.0258708,-0.2026813328,-0.0974158123,-0.0611961558,-0.2312194556,0.2005805969,0.3475272357,-0.0399407297,-0.17849271,-0.2595869899,0.0066010007,0.2199456841,-0.1413224787,-0.1959666312,-0.3471494019,0.2996377349,0.2433196157,0.09838368,0.4283132851,-0.0721405596,-0.0055664526,-0.0860208273,-0.1282110959,-0.0213420093,0.013511681,-0.3342831433,-0.1812717319]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/129","title":"[Feature request] Add Google Natural Question dataset","comments":"Still work in progress :)\r\nThe idea is to have the dataset already processed somewhere so that the user only have to download the processed files. I'm also doing it for wikipedia.","body":"Would be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.","comment_length":32,"text":"[Feature request] Add Google Natural Question dataset\nWould be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.\nStill work in progress :)\r\nThe idea is to have the dataset already processed somewhere so that the user only have to download the processed files. I'm also doing it for wikipedia.","embeddings":[-0.1264829934,0.118322663,-0.2279913723,-0.291672945,-0.0624212101,0.2136778831,0.2895945609,0.1539695859,-0.0013700712,0.1868419945,-0.0968561098,0.3380507231,-0.2951551378,0.1478503346,0.5057772398,0.0203669779,0.1696388274,0.0553913526,0.2871233225,-0.0953791589,-0.2698328197,0.0765784606,-0.1687618196,-0.2782278955,0.1056088507,-0.1428774446,-0.0722409263,0.1143308133,0.027486071,-0.4552854896,-0.2237554342,0.2545962334,-0.1129967123,0.2231184244,-0.0001096427,-0.2789277732,0.2226358354,0.112318866,-0.3725150228,-0.1047181562,-0.0413002409,-0.1243512332,-0.0853371993,0.0625224411,-0.0817401633,-0.1325051785,0.3818160295,-0.6124858856,-0.0268362463,0.3899215162,0.122078225,-0.199852109,-0.190206483,-0.1897235811,0.3511377573,0.4608781636,-0.2872081995,0.1540661454,0.3880565763,0.0529771447,0.2539665103,-0.0058899093,-0.0131472936,-0.2195429206,0.2016466409,0.1184510663,-0.2630254626,-0.7430915833,-0.0995844826,0.2322649062,0.6047319174,-0.0574460588,-0.1416893601,-0.1579467952,0.1478342265,0.1208298281,-0.1835998148,0.3654572368,-0.3039933741,0.0501606464,-0.0627103373,-0.6823121905,-0.1605512798,-0.0132761523,0.332806319,0.0657699406,0.1374448985,-0.0156134404,-0.0974986404,0.0450189225,0.2130221277,-0.0901884958,0.0328702033,0.2595645189,-0.3627632856,-0.5124523044,0.0899482518,-0.1171922311,0.3360565603,0.3056876659,0.0137720397,0.0356598198,-0.0117772799,-0.1386509538,0.2288579047,-0.4270021617,0.3070000708,-0.0288840905,0.2250990868,-0.2379026413,0.0268371664,0.0528682731,0.1061939001,-0.0000614115,-0.0319908895,0.0561245605,0.5262048841,-0.0002247452,0.1605979651,0.0187635701,0.1277631223,-0.1333844811,-0.1766401231,-0.0510196127,0.0360507779,-0.2246823609,-0.1994835138,0.3086479902,0.2672205269,-0.4665071368,0.0621694513,-0.0965680555,0.1323516518,0.4155916274,0.4051395357,-0.2413119674,0.169897154,0.0360411592,-0.0391495638,0.1876740903,0.3271625042,-0.1256587803,-0.0916793495,-0.1870537549,-0.000780109,-0.2132493407,-0.1154736876,0.0056673461,-0.2466759682,0.2631090879,-0.1965469569,-0.2714584172,-0.4861047268,0.1716238558,-0.0501413532,-0.3752221167,0.0850140899,0.583350122,-0.2131745964,-0.4421684444,-0.0384575091,0.3340290785,-0.3543671966,-0.1925550848,-0.1020908952,0.3394127488,-0.2921144962,0.1228870004,-0.2250705659,0.17884399,-0.1179101095,-0.0456080697,0.0647220686,0.3149306774,-0.088625446,0.0736361966,0.4298567176,-0.0873505324,-0.1254737377,0.0056469161,-0.3732474148,-0.1461568326,-0.2327250093,0.4704762399,0.0164358076,-0.0243042242,0.1361232847,0.620953083,-0.1872262359,-0.0575591065,-0.0665313378,-0.2913662791,0.2745901048,0.4272626638,-0.0473464541,0.1423153281,0.4425697625,-0.4155982435,-0.1925052851,-0.0963218063,-0.1358813792,0.0439686589,0.5920690894,-0.0666090399,0.048606135,-0.3970466554,-0.6555466056,0.0584423877,-0.5253989697,0.4157306254,0.1717705131,-0.3447129428,-0.190633297,-0.1787100285,-0.0554477647,-0.0577215739,0.1324526221,0.0833631903,-0.210131675,0.3074828088,-0.2489636242,-0.0875366628,-0.0911918357,0.1868052036,-0.0371299572,0.2840295434,0.0193647444,0.0078887753,-0.0581408702,0.101112023,0.0558958165,0.1696424037,0.1618481278,0.147439912,-0.2663387358,0.3004567623,0.7892981768,-0.1025377437,0.0999988914,-0.433324337,-0.0779074207,0.2329799235,-0.0746343285,-0.0570602901,-0.5146198869,-0.0429202132,0.1678945869,-0.1031115949,-0.2638575435,0.1069280803,-0.0922851861,-0.0043124165,0.198727712,0.1549771726,-0.2380850166,0.3278709054,0.200280875,-0.3042111993,-0.197980538,0.2825926244,0.3525802493,-0.0804718658,0.2441768944,0.1669814289,-0.0494250208,-0.2179890275,0.3993863761,-0.2034090459,0.3370552361,0.1952108741,0.1115920618,0.2344420403,0.035474468,0.0314401388,-0.0775914565,0.039846383,0.0043040374,-0.0123272175,0.3985608816,0.1814754307,-0.1965930611,-0.0778731927,0.0317704566,0.0802367106,-0.0097108316,-0.1559304148,0.0805395395,0.019641934,0.1897779554,0.3993789554,-0.2257104963,-0.2347687036,0.3144590259,-0.1198318526,-0.1439290643,0.0316768438,0.0501617342,0.8729278445,-0.0273239724,-0.5442200303,-0.0153269405,0.0397060923,0.0890801772,0.1639343798,0.2189647704,0.0372849554,0.5934820771,-0.1173360422,0.2164494097,-0.617814064,-0.3918794096,0.1022077575,0.0896863565,0.710786283,-0.1210797578,0.1686468422,-0.0542224012,-0.0059996038,-0.1753214896,-0.1285113394,-0.1234858558,-0.3988273144,0.1177428067,0.1473958343,-0.1068145633,-0.5827357173,-0.2150991261,-0.2387956828,0.275292784,0.0514024273,0.0896681324,0.1455763876,-0.0885479748,-0.0370756611,0.0952314511,0.1224287525,-0.1380489469,0.0188020095,0.1814914942,-0.3380017877,0.020610299,-0.0145539213,-0.0074496451,-0.2079558372,0.1188184544,-0.4295735955,0.1003184393,-0.0406882092,0.5455446839,-0.0113011291,-0.1634618342,0.4637937546,0.0664937198,-0.0375897437,0.0745828524,-0.2076437324,0.0464941859,0.4402966797,0.0109204203,0.3696685135,0.1138623878,0.0918771029,0.7774419785,-0.0929285362,0.165503487,0.1381973475,-0.2433730811,0.3671025038,-0.0329846963,-0.2301270515,0.2571269274,0.0395342149,-0.045794744,0.2923860252,0.3140208125,-0.5405169725,-0.1084464192,-0.0975295454,-0.2241243869,0.0787612796,-0.0909119174,0.0495339632,0.1687605977,-0.0146236168,-0.3095278442,-0.2697105408,-0.4782190025,0.0006556556,0.1415156722,0.3210869133,-0.0214643907,-0.7990790606,0.2610267401,-0.2326617837,0.2963690758,0.3586941659,-0.1683469266,-0.2834871113,-0.0251544919,-0.0751447529,0.105555512,0.0950825959,-0.6820047498,-0.3342575133,-0.0939260647,0.3469977975,0.2276742458,0.1512900889,0.153706044,0.2113018632,0.1694778651,0.3991979361,-0.1164277047,0.0147713209,0.4677456617,0.2445376962,0.0231523495,-0.2225554734,0.1410969198,-0.0984069332,-0.0531165451,0.0483176857,0.0898004249,-0.038260147,0.0054527274,-0.3349076509,0.0222195257,0.1141530424,0.2640430331,-0.0694898367,-0.1289881319,0.2777032554,-0.0161061529,-0.0457972996,-0.2402308285,-0.2297029793,0.3972441852,-0.0402562022,-0.1828676164,0.075965412,-0.3203580678,0.4652192891,0.2720444202,0.2872103453,0.1312916279,0.0119014829,-0.0484231338,0.0305479765,0.1681139469,0.1870979518,-0.2057365626,-0.26636675,-0.3575922251,0.396810472,0.0434698313,-0.239214614,0.3576796353,0.0763992444,0.0213085897,-0.2220416516,0.1333397776,0.8947169185,-0.0358984992,0.0528780892,-0.044707153,-0.2402792275,0.513348937,-0.2384704798,0.1108819917,-0.2768147588,-0.2028582692,0.0088015804,-0.120022282,0.0389480926,0.2103425115,-0.1931476295,0.2788774371,0.2045099288,0.5296322107,0.3737516701,0.729147613,0.1295298487,-0.0700048655,0.4096052945,0.1770173758,-0.0639947355,0.112131387,-0.1489973068,-0.1916503459,0.0413915254,0.0596120693,-0.3205528557,0.1084137484,0.0594030023,-0.1732536107,-0.0464015007,-0.412935555,0.2577240169,0.3421130478,0.3003041744,0.3664965928,-0.2876014411,-0.0446529351,-0.0216342118,0.1876133531,0.0895207301,-0.028689744,0.235544309,-0.2369134277,-0.2917117476,0.3072361648,0.0121785002,-0.5410516262,-0.7716203332,0.1476327926,-0.1288671792,-0.200255096,0.0093943998,0.1515381485,0.0769796222,-0.201329574,0.1030309051,-0.0029128834,0.2968446612,-0.1928766817,0.2775415778,0.0477330796,-0.2167565972,-0.2681976557,0.1629424989,-0.4700159431,0.1171531379,-0.0344330594,-0.0447847545,-0.3277244866,0.0771810338,-0.33488518,-0.1297084987,-0.0249946043,-0.2824254036,-0.1224708557,-0.1122542322,-0.2110599577,-0.1402383,-0.0135644367,0.1458732486,-0.0593236424,-0.0550620258,-0.0300885625,-0.2311188579,0.2259329706,0.1825275272,-0.3293658197,0.1830652654,0.0149295023,-0.3160926104,0.0486816801,-0.1452112794,-0.0063239583,-0.0528488085,0.0427395627,-0.2127816528,-0.2501588464,-0.0078132525,-0.4964265525,-0.0221406408,-0.1543390602,0.1246522591,0.1322606951,0.0031429064,0.1324242502,-0.1171665266,-0.2033440918,-0.0441999212,0.1001467183,-0.0077485377,-0.2282249928,-0.036107596,0.1218543947,0.1777943671,-0.0034432192,0.0384112932,0.4273817837,-0.0317000002,0.0171849038,-0.0129487114,0.2087147832,0.0780423433,-0.0089410096,-0.1340428144,0.2957901061,0.4136580825,0.1380592585,0.1150139943,-0.2549239397,-0.1582354456,0.0589896142,-0.1757811904,-0.1221647561,-0.3631984591,0.3854271173,0.2200687677,0.1688584834,-0.1993879825,-0.1094223037,0.4071269035,-0.043178536,0.0932114944,0.224827975,0.180392161,0.060044799,0.5306475163,0.1425167024,0.2679504752,-0.0027292904,0.2815707624,-0.2442645282,-0.005794893,-0.0698049664,0.1937907189,0.093527399,-0.2080988884,0.0658863634,0.013777866,-0.0094902487,0.1928471625,-0.0599220768,0.4543343782,0.2252905071,0.284716785,0.2496798933,0.1088786349,0.16216892,0.4742131233,0.5453689694,-0.1347162873,-0.2987945676,-0.1159657538,0.2142247856,0.0841614604,0.2528367937,-0.1838054955,-0.17157574,0.1785861105,0.0249804631,0.0176922642,-0.5543310046,-0.2713360488,-0.0649453178,-0.0385675132,0.3036853075,0.5384804606,0.2075599134,0.0961764455,-0.0577886403,-0.1321073771,-0.3161457777,0.0671726614,0.1705609262,0.2660114765,-0.2431790382,0.2275885493,-0.4661037326,-0.0193287153,-0.1350448281,-0.0441155918,0.1105103642,-0.026977405,0.0380791016,-0.1401662976,0.272539705,0.1590077728,-0.0692598373,0.2817730904,0.1698110402,0.0800339282,-0.4773024917,-0.0394711457,-0.0188462064,-0.2353262156,0.1025566459,-0.0457152389,-0.0405280516,-0.1752206832,0.2825731337,-0.0891309455,-0.0471125133,-0.0324287191,0.1031923145,-0.025537841,-0.1508350968,0.2183919549,0.2567389309,0.0751330331,-0.1158794165,-0.4217970967,0.0160843097,0.1543062478,0.05462097,0.1882624328,-0.0203694813,0.0373376831,0.430306673,-0.1522462964,-0.23574467,-0.3084613681,-0.0191211961,-0.4197785556,-0.0007807341,-0.3621883392,0.2439667135,-0.0121728163,-0.1025440544,0.1342561394,-0.139633894,-0.0383384861,-0.0095764324,-0.0967921764,-0.0821815878,-0.1409668028,0.3854480684,-0.0765448511,0.4136479497,-0.144494772,-0.1302660406,-0.0255282391,-0.1579268277,-0.1202981919,-0.060478311,0.039265383,0.2337161303,-0.1245405897,-0.1382558942,0.2238861173,0.4119002819,0.2467154711,0.2627681196,-0.2406850755,-0.0000555273,0.0660140216,0.1587697417,-0.0062568118,-0.1299206614,-0.1807848215,-0.1478940994,-0.0897652656,0.0770090297,0.3406490386,0.0047929096,-0.2193232179,-0.1407265365,0.0838686079,0.2768307328,0.0240295045,-0.0091434624,-0.3159788847,0.3554436266,0.044292599,0.2519076765,0.367336601,-0.1636931896,0.0251702964,-0.1388675719,-0.206733048,-0.0635093451,-0.0509295166,-0.3128146231,-0.3202603459]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/129","title":"[Feature request] Add Google Natural Question dataset","comments":"Super appreciate your hard work !!\r\nI'll cross my fingers and hope easily loadable wikipedia dataset will come soon. ","body":"Would be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.","comment_length":19,"text":"[Feature request] Add Google Natural Question dataset\nWould be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.\nSuper appreciate your hard work !!\r\nI'll cross my fingers and hope easily loadable wikipedia dataset will come soon. ","embeddings":[-0.2017221898,0.1256560832,-0.243652463,-0.1397192776,-0.0242803767,0.1866627932,0.3122389019,0.2764431834,0.2076867372,0.1970580816,-0.1712979823,0.4047228992,-0.2165549845,0.0521147177,0.4223890603,-0.073818326,0.0963116959,0.0344126187,0.1412131041,-0.0688865259,-0.2413297445,0.1189085916,-0.2048509419,-0.3114299476,0.1137431785,-0.0265876912,0.0056279604,0.1296857446,-0.0191482026,-0.46474877,-0.1019493714,0.2669221461,-0.0335932523,0.0589358844,-0.0001016455,-0.3032364547,0.4032799602,0.0779839605,-0.5168819427,-0.0655971393,-0.0521263182,-0.185677886,0.0872990787,-0.0902722701,-0.0940551013,-0.0729103014,0.3084506392,-0.4208047092,-0.0311204437,0.1934424341,0.2163048536,-0.2274320722,-0.054045517,-0.2763386965,0.4652504623,0.2870284319,-0.252453059,0.1178008839,0.2060125023,-0.1705203652,0.0489440747,0.1386658549,-0.1220482662,-0.1587690711,0.2546662092,0.0132162627,-0.2298582345,-0.6946879029,0.0819802582,0.3509275019,0.6737273932,-0.0681870505,0.0174818672,-0.0835055038,0.0691689998,0.316935569,-0.0121149784,0.2640858293,-0.280595392,-0.0260666758,0.0806760117,-0.5248840451,-0.1353522688,0.0794723779,0.1235963479,0.2329439372,0.0704725161,-0.0491512567,0.0704620034,-0.1361278296,0.121880129,-0.0182035733,0.0354017429,0.289712131,-0.2823018432,-0.3730688691,0.1729209274,-0.0806743652,0.4334471226,0.1434795558,-0.182099402,0.0895427465,0.069713749,-0.1381901503,0.3977148831,-0.4390159249,0.2682743967,0.0185330082,0.3492509425,-0.1019775867,-0.0591323152,0.1783821285,0.0601319596,-0.0503683947,-0.0196788795,0.0031638169,0.4461967051,-0.1417485774,0.066461727,0.0695009306,0.0996486098,-0.2425523549,-0.2930368483,0.1231226251,-0.0758641809,-0.1061744615,-0.1502657533,0.1875163466,0.0976303965,-0.3231859803,0.0109538548,-0.0869039446,0.0609172173,0.3313105702,0.3955783248,-0.2209326923,0.1768432707,-0.0045005567,0.0399462804,0.1234167367,0.1304859966,-0.0674804077,-0.0571334474,-0.0708311573,0.1281122863,-0.1609143764,-0.0249504615,-0.2242379934,-0.2743986547,0.2609122396,-0.1720436364,-0.2196110785,-0.4405394495,0.2128851414,-0.0390662141,-0.2638796568,0.037721362,0.6213995814,-0.2066448182,-0.3917209506,-0.094340086,0.3771166205,-0.274625957,-0.19942002,0.0378392227,0.3590111434,-0.2019740492,0.0762342066,0.0023042518,0.2200215161,-0.1345555037,-0.0042752698,-0.0283836406,0.1991046965,-0.0155628538,0.0061412142,0.3151330352,-0.1164556816,-0.1977724731,0.0492177941,-0.1604313701,-0.1088204905,-0.2988303006,0.365462482,0.2050008625,0.0367247686,0.1548392773,0.6484239101,-0.0959142074,0.0700780302,-0.1999340355,-0.3467964232,0.2524615228,0.4127126634,0.1526065767,0.0555251464,0.4181648791,0.0242204256,-0.0879720002,-0.0933704674,-0.1246289089,0.1427963227,0.4049690068,-0.0752707794,0.1650160104,-0.2869382799,-0.7554092407,0.0070732986,-0.3759455681,0.4506393671,0.1198987663,-0.3328285217,-0.2033619285,-0.1555897892,-0.0963467956,-0.0173717532,0.2725466192,-0.0232194792,-0.2063346207,0.3823437393,-0.0413553193,-0.2588642538,-0.0988851562,0.0557651408,-0.0398643427,0.3208885491,-0.0731618032,0.0140761416,0.0993167683,0.1940563768,0.1240271628,0.1539308727,0.0964250118,0.0013577669,-0.1384818256,0.2336906344,0.6280089021,-0.1288906485,0.1499207616,-0.4823338687,-0.0783839598,0.249270305,-0.1126050279,-0.1443093419,-0.5968993902,0.0453750193,0.1952260137,-0.0277763419,-0.2514410317,0.1576115042,0.0838201791,-0.0013669209,0.1719455719,0.0266576335,-0.275231719,0.3787720501,0.0927775428,-0.2781858444,-0.2065993696,0.3358213305,0.3354873955,-0.0571714304,0.2678935528,0.2196593732,-0.1162183285,-0.327996105,0.3910807669,-0.327238977,0.0438269526,0.2073225379,0.1478286833,0.1060358733,-0.0045650909,-0.0442673899,-0.0006884474,0.0218425579,0.1484501511,0.0149370134,0.3489830494,0.1786848307,-0.3392166793,-0.0444479957,-0.0846968368,0.1069925427,-0.0341456383,-0.1682205349,0.1051002145,-0.0143145034,0.3418886662,0.3167032897,-0.2188689411,-0.3700164557,0.2243956029,-0.0030721356,-0.0534685664,0.021066064,0.1321425587,0.76685673,0.081600219,-0.2681781948,0.0104911365,-0.3074510992,0.0731090978,0.2322458774,0.3219352961,-0.0507351495,0.4783683121,-0.0127267512,0.1537485868,-0.5923330188,-0.4827515781,0.1192220896,0.1201204881,0.5215231776,-0.0767149255,0.2746658623,-0.1981059313,-0.0962232202,-0.1624536216,-0.1311424971,-0.0499704815,-0.2253646106,-0.081558533,0.1146077812,-0.0186095368,-0.5942140818,-0.1366326213,-0.2467777282,0.2635306716,0.1249779984,0.0920455754,0.2029576004,-0.0379214957,-0.0611100495,0.1131454334,0.1624169797,-0.2250624597,0.0724944845,0.2097474635,-0.3217253089,-0.1396625191,-0.0905296504,-0.0590848178,-0.1587044597,0.0375077203,-0.502446413,0.1750737429,0.0547256917,0.5136936307,0.0465577058,-0.0670346469,0.2833913565,0.1509447247,-0.1449300051,0.1244263202,-0.0931254029,-0.0735590309,0.3112142086,-0.0499760732,0.340729326,0.0626762062,0.1072977185,0.8626490831,0.0262238849,0.2461064011,0.0816391706,-0.1486225724,0.2170805633,-0.041842103,-0.1956444085,0.2293414474,-0.0688142702,0.0712457523,0.3811711073,0.3485940695,-0.7127806544,-0.132243529,0.0294274185,-0.149114579,-0.0603660606,0.0778016225,0.054173734,0.0699850172,0.0360005386,-0.1101418585,-0.2480125427,-0.5357368588,0.1295739114,0.0482935049,0.1851278692,-0.0859129429,-0.5935693383,0.1470861733,-0.3421041667,0.1983744055,0.2628055811,-0.2051061541,-0.2329393774,-0.091507487,-0.0430252999,0.1226267293,-0.047290951,-0.5968506932,-0.1789013594,0.0909118652,0.3216105103,0.0200854819,0.1138465628,-0.0045610457,0.1625011414,0.2883402407,0.2708817124,-0.2645644546,-0.0858377218,0.4172804654,0.1865865737,-0.0441888906,-0.2176682055,0.1468117833,-0.073518768,-0.1281828135,0.0271384828,0.0516687222,0.0792045966,-0.0753632337,-0.2486678511,-0.0119471876,0.2135819644,0.0522740707,0.029155992,0.0081770048,0.4452054501,-0.0595232695,0.059198536,-0.175677225,-0.2734086812,0.3783250749,-0.1988500208,-0.1228888929,0.0863266438,-0.3101259768,0.1966827661,0.280584693,0.2280436754,0.0192760769,-0.0044815065,-0.0680993721,-0.0929078236,0.0332280435,0.0595763288,-0.2315208018,-0.1461089104,-0.4073536396,0.5744869113,-0.0246708542,-0.2062188387,0.4350853562,0.1027261242,-0.0777830482,-0.0194744263,0.2194956094,0.8299000859,-0.1335147321,-0.0075181597,0.063693732,-0.0539618842,0.4576734304,-0.2505891621,0.0735540092,-0.4168210924,-0.0084919827,0.045606479,0.0198429022,-0.1410610974,0.1322646588,-0.3169585466,0.2361897081,0.162144348,0.4432865679,0.2552448809,0.6676918268,0.2424916327,-0.1550122499,0.1867098808,0.2357006371,-0.0828566328,0.1624566764,-0.1243340671,-0.125570327,0.1676034182,0.0658551008,-0.3938299716,0.1224125996,0.1045020074,0.1199975237,-0.0506299734,-0.4421891272,0.1672293246,0.3943251669,0.1877704114,0.2768916488,-0.3447802067,-0.1187880114,-0.0482082218,0.0517186411,0.0695094243,0.0050473572,0.3263136148,-0.2268957794,-0.2697200477,0.2798440158,-0.0280678868,-0.3937918544,-0.7479723692,0.1891156733,-0.1511424482,-0.2068119943,0.014640389,0.0870558545,-0.0576682799,-0.2293117195,0.1886188686,-0.0282722376,0.3348191679,-0.0941612422,0.5585499406,0.0249891207,-0.1245562956,-0.0176535267,0.2369760424,-0.3065421581,0.0392976478,0.1258615553,-0.2271726876,-0.3767748177,0.0916154534,-0.3147225082,-0.1326134205,-0.0033261278,-0.27339077,-0.057555642,-0.0747801512,-0.0820466429,-0.1083166972,-0.0323698893,0.1667633951,-0.0741983801,0.0460748598,0.1443843842,-0.303845793,0.1990856528,0.2698339224,-0.4061629176,0.1455001235,0.1015824303,-0.3921451271,0.0970488489,-0.1440386772,0.0267851707,-0.0550875142,-0.0830566138,-0.2625126243,-0.1238798648,0.0842673704,-0.5643660426,-0.024732668,-0.2464054525,0.1013985649,0.0770656317,-0.0061122389,0.1267730743,-0.2499560863,-0.2763819993,-0.0238529667,-0.0266975183,0.1340636611,-0.2562166154,-0.0630080551,0.1146491915,0.0420220755,0.1150593832,0.0218585059,0.3890399635,-0.0226261448,0.0821869969,0.0305915419,0.1802603155,0.0513875186,-0.0341734365,-0.18733868,0.4169662893,0.3195123076,0.0212557428,0.1506751627,-0.2277188599,-0.1770617217,0.1769993007,-0.2267283946,-0.0846477374,-0.3461987972,0.2979512513,0.0658283457,0.2433163971,-0.1541794389,-0.1555650979,0.2816893756,-0.0220122747,0.1283847392,0.193256706,0.2165102214,-0.2247624993,0.461284101,0.1914697438,0.1502158642,-0.0829280913,0.1770392656,-0.3337228,0.0258027054,-0.1233357713,0.0699183866,0.1436952949,-0.1244452,0.0552729405,0.1204426885,-0.029124178,0.1607766449,0.0560952276,0.3877538741,0.1583577096,0.3579882383,0.2694165707,-0.0543300249,-0.000401969,0.3794995248,0.4973192215,-0.0085080191,-0.2578886747,-0.1455724239,0.1235439405,-0.0840489864,0.1343507916,-0.1973325014,-0.2540834546,0.2803653777,0.0516729578,-0.0954664052,-0.5176731944,-0.1210098118,0.0048650233,0.0199109167,0.0933810323,0.5804944038,0.2324518859,0.1397990435,-0.0496818982,-0.0689874142,-0.2643327415,0.1004764214,0.0505643375,0.2765887976,-0.1153285801,0.2246717513,-0.5835821629,-0.0967264548,-0.0239767861,-0.158259511,0.073010847,-0.1432618648,0.0423458144,-0.2617450655,0.3171426654,0.2427728623,-0.155713439,0.1199158654,0.3222721517,0.0269380175,-0.1829206496,-0.3265081048,0.0956234485,-0.1970596462,0.1611655056,-0.0911294371,-0.1361400187,-0.1043128893,0.2041631937,0.0825366601,0.0682390556,-0.1296533942,0.1413918883,0.1740397513,0.068771936,0.1836816967,0.196671769,0.0824553594,-0.1812617332,-0.4225731194,0.0702468306,0.0128514664,0.0565258972,0.3268133402,0.1963738054,0.0592327304,0.4628643394,-0.2050184011,-0.3977901042,-0.0790096894,-0.0306949206,-0.530298233,-0.0237284172,-0.0909201354,0.1974270344,-0.1180870011,-0.0832891837,0.1738403738,-0.1959283054,0.0899704769,-0.2011751682,0.0082131661,-0.0598205887,0.0493348949,0.258613497,-0.163077876,0.4206116796,-0.1707490534,-0.2628891468,-0.0634048879,-0.2587566078,0.0949070007,-0.0358782634,0.0971381441,0.2370991111,0.0201058462,-0.2771774828,0.2679741681,0.3956780434,0.1753614098,0.3635539114,-0.3275954723,0.0337469913,0.1932550222,0.1581031829,-0.1832087189,-0.0189898685,-0.2257447392,-0.1279949844,-0.0483816192,-0.1600064188,0.1892322302,-0.0581814311,-0.2753027976,-0.1643755883,0.0284312256,0.1633923799,0.135329172,-0.0095459791,-0.1871438622,0.2839349806,-0.0327428468,0.0291293897,0.3169107139,0.0377589762,0.0842675045,-0.3207291067,-0.3377174139,0.0936820507,-0.063525252,-0.4468116164,-0.3785190582]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/129","title":"[Feature request] Add Google Natural Question dataset","comments":"Quick update on NQ: due to some limitations I met using apache beam + parquet I was not able to use the dataset in a nested parquet structure in python to convert it to our Apache Arrow format yet.\r\nHowever we had planned to change this conversion step anyways so we'll make just sure that it enables to process and convert the NQ dataset to arrow.","body":"Would be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.","comment_length":66,"text":"[Feature request] Add Google Natural Question dataset\nWould be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.\nQuick update on NQ: due to some limitations I met using apache beam + parquet I was not able to use the dataset in a nested parquet structure in python to convert it to our Apache Arrow format yet.\r\nHowever we had planned to change this conversion step anyways so we'll make just sure that it enables to process and convert the NQ dataset to arrow.","embeddings":[-0.2953702807,0.2493468076,-0.2129656225,-0.2288988531,-0.2177684903,-0.007762617,0.1600123793,0.2642006576,0.0725733489,0.2330668569,-0.1219623908,0.4617756605,-0.3594296277,0.3047490716,0.6277964711,-0.1398961693,0.109149985,0.1171826348,0.1310335547,0.0052731531,-0.317882508,0.2159188092,-0.2996281385,-0.1805972159,0.3880342245,-0.2200264335,-0.0794761404,0.0181822274,-0.0265263487,-0.5745174289,-0.3112954199,0.0473093987,0.0476757661,0.040043842,-0.0001121784,-0.2186170071,0.174267292,0.06726899,-0.4605724514,0.0060868203,-0.0046946011,-0.2376577556,0.1306650788,0.1122267917,-0.2582050264,-0.4748965502,0.3113420904,-0.5965595841,0.151903674,0.5919366479,0.1091722175,0.0871775076,-0.1037180349,-0.0694532841,0.3513969481,0.054879088,-0.3796557486,0.0306329634,0.287867099,0.0260598753,0.0931188986,0.0747287571,0.1412568241,-0.2341369092,0.0540361963,0.1074217334,-0.2296620011,-0.5888482332,-0.1962655783,0.2285999805,0.5203668475,-0.1335051209,-0.1554478258,-0.1241949126,0.1033762544,0.0856694356,-0.2679816484,0.4884107411,-0.3231521845,0.1151645854,0.0538653061,-0.5424550772,-0.2480218709,-0.0013973647,0.2195855975,0.2661553621,0.1762827635,0.0917402655,-0.1932388097,-0.1051129922,0.3093812168,0.1191854104,0.0989951715,0.1564710885,-0.5393491387,-0.4150444269,-0.0024330709,-0.1480288506,0.2973623276,0.3861842752,0.1967416853,0.0959154814,0.1803384125,-0.084174104,0.2228099853,-0.4225268066,0.3411365449,0.0574572906,0.0475361161,-0.2187972814,0.0957734585,0.0870952979,0.098573409,-0.1406960487,-0.1046526283,0.0877152234,0.402862668,0.1496584564,0.0862818137,0.0376020521,-0.0851916671,-0.0380429663,-0.1958857477,0.1254626513,0.085096769,-0.1955902576,-0.1212322637,0.4383507669,0.1599331796,-0.5843251944,0.0355459787,0.0209727865,0.1136966199,0.2827502489,0.3090487719,0.0690916926,0.0580869094,-0.0036550178,0.2363467515,0.2372781187,0.29205814,0.1680350602,0.0874804184,-0.1461002082,-0.0065881619,-0.3226618171,-0.0357294939,0.0072311522,-0.2910197377,0.2138700783,-0.2787839472,-0.2301320434,-0.4146444499,0.0961111709,0.0838122815,-0.3740875125,0.0682291985,0.4454984963,-0.325990051,-0.3123779595,-0.0269231871,0.1304176599,-0.2292339057,-0.236409381,-0.0510749854,0.2156182379,-0.3670797348,0.0269612279,-0.1518088579,0.2874581516,-0.1973644048,0.087819919,0.0525866225,0.1360738277,-0.1900932491,0.02233533,0.6011225581,-0.2211753428,-0.0514523238,0.0367660411,-0.2813921571,-0.1744874716,-0.3171973825,0.3433481157,0.1696347147,0.0938416496,0.0310205203,0.6536281705,-0.0924195722,-0.0391355455,-0.039010603,-0.227211833,0.199558422,0.4691130519,-0.1735947877,0.0021728354,0.365701139,-0.5542101264,-0.0867786556,-0.1103414893,-0.054489959,-0.0242103953,0.4067526758,-0.2000357509,0.1616885513,-0.251570195,-0.7552241683,0.0615149029,-0.7966159582,0.4462405145,-0.0178984851,-0.1940711886,-0.25786376,0.0020970935,-0.1257833838,-0.1862559766,0.1563038826,-0.0292648375,-0.2549140751,0.223178193,-0.2163216472,0.0227223746,0.1051910967,0.2016458809,-0.1075953543,0.2680731416,-0.0149880806,-0.039805036,0.1759323627,0.1753741652,0.1201355383,0.1826214492,0.2836145163,0.1856957078,-0.282862246,0.2047657818,0.5085203648,-0.122134909,0.0383016057,-0.5743350983,-0.0077638873,0.3941744268,-0.2608610094,-0.0196069311,-0.3717058301,0.33925578,0.2843263447,0.1068695635,-0.3091539741,0.2991065681,-0.0987705365,0.0346748345,-0.0478970148,0.1723697633,-0.2749818265,0.3048970997,0.1159953177,-0.379665345,-0.2862068117,0.2668562829,0.3423602879,-0.0679126307,0.3512673676,0.0810949281,-0.084396556,-0.1377175897,0.4772306383,-0.1296388656,0.1275278181,0.2070845962,0.1430482268,0.2882402539,0.0135387015,-0.0293025617,-0.0542867482,-0.062663354,0.0930217803,-0.0140057923,0.5086836219,0.166303426,-0.2894741595,-0.0866594762,-0.1338922232,-0.0666268617,-0.1769340485,0.0182868522,-0.0329752751,0.1535910517,0.0673317537,0.4362247288,-0.2621743679,-0.3076348901,0.4359920323,-0.2957397401,-0.1318915635,0.1456815302,0.1299084425,0.7573097944,0.1741632521,-0.5157591701,-0.1659678221,-0.0387541354,0.0553572103,0.18978028,0.0895486251,0.0912854597,0.5234966278,0.0052641835,0.3065762222,-0.5940279961,-0.3140815794,0.1414429992,0.069832623,0.6522313356,0.0418383628,0.1898486465,-0.2482081503,0.036749661,-0.1394643635,-0.1734526753,-0.2007226944,-0.1394867748,-0.036491517,0.3687531352,0.000965357,-0.7123110294,-0.3366532326,-0.3728920221,0.2275636792,0.1590357572,0.1532221735,0.0394760035,-0.186305359,-0.0669428557,-0.021684425,0.0722760931,0.036198739,0.2218100727,0.3625462353,-0.4247932732,-0.0341427363,-0.0311215539,-0.2124654055,0.0582344793,0.2095229626,-0.3300026655,0.2758338451,0.0560238436,0.4175790548,0.038538821,-0.1555835307,0.4633901417,0.0285181478,-0.0280735157,0.054876592,0.0071400232,-0.0003175375,0.4249238968,0.0339552462,0.4035647511,0.171717152,0.0728459805,0.8291463256,-0.101885356,0.2454564571,0.1981819868,-0.3811302185,0.192614466,-0.094975315,0.0792257935,0.1595588177,0.0482387394,0.0363795869,0.2618745565,0.2772433162,-0.26420331,-0.0219521169,-0.0085883616,-0.2737037838,-0.0691514611,0.0772752091,-0.1758162826,0.208435744,-0.057199046,-0.3028361201,-0.0077579985,-0.3353803754,0.0962579772,-0.0371425599,0.2221387476,-0.0566493906,-0.5201749206,0.2741071582,-0.2561059892,0.3044739962,0.1520831734,0.0187370442,-0.0972888917,-0.1100977659,0.0510124676,0.0260034334,0.2799509466,-0.5850881338,-0.2880302668,-0.0004771212,0.2919858992,-0.1805926263,0.1485688984,-0.1216200516,-0.1709078252,-0.0084621329,0.2437309027,-0.2748570442,0.0256144181,0.2547246516,0.1436057091,-0.0830470175,-0.0870349184,0.0610063337,0.00381171,-0.0679193065,0.0861342922,0.1751767844,0.0368140228,0.0221585408,-0.3207511604,-0.1898104548,0.0385239385,0.0654068664,-0.0426951721,0.1252105385,0.0773346201,-0.119737938,-0.0457897224,-0.1115374938,-0.2641516626,0.3890094459,0.0105832098,-0.1641115248,0.0933910012,-0.0709336475,0.5043094754,0.2509220541,0.2337360233,0.0052509913,-0.1095805913,0.0317121036,0.0412797928,-0.0287786033,0.0354567692,-0.3161974251,-0.1466639042,-0.2415972054,0.6507223845,0.0372220501,-0.2402258962,0.2970208228,0.4208500683,0.0302587636,0.1934558153,0.2243903726,0.8491632938,-0.1841994822,0.2300492823,0.3015240729,-0.5155704618,0.4732300341,-0.2944478691,0.0363535881,-0.5128238797,0.0509699546,0.0217953324,-0.2157734334,0.0110963546,0.3124728799,-0.3961852491,0.3208960295,0.3178912103,0.2970656455,0.2469931841,0.5372269154,0.1069167629,0.0190015882,-0.033440195,0.1644966006,-0.1671494544,-0.0511111766,-0.1643350422,-0.2703466713,0.0554557815,-0.0509196296,-0.0999383926,0.1822422147,-0.0203763228,-0.1326396763,0.1385697573,-0.6162393093,0.3407854438,0.3981691003,0.1293951571,0.2212397456,-0.3203514814,0.0915318877,0.0743872747,0.0871158317,0.1228626966,0.0445478745,0.147400111,-0.1661927551,-0.1829893887,0.3560411036,-0.0840281099,-0.5819593072,-0.5803861022,0.2874083221,0.0599017031,-0.051909592,0.0874738619,-0.0114848018,0.0312291961,-0.1485454887,0.101768516,-0.1339637339,0.2199296653,-0.2898611128,0.5137260556,0.0661385804,-0.0919781402,-0.2159300596,0.1337400377,-0.2603694201,0.1668200791,0.2180063576,-0.0852472931,-0.3821596503,0.0981840566,0.1488300413,-0.2141097486,0.0560086519,-0.3263069093,-0.1686580032,0.0470834449,-0.1273081452,-0.1296937019,-0.1492377669,0.0034557462,0.0149493152,0.0116738454,0.1407546252,-0.3681993484,0.2026133388,0.3773682714,-0.4822437763,0.0854319632,-0.0428501144,-0.3180077672,0.0114279129,-0.0445798114,0.126340732,-0.0620591789,0.0007136732,-0.1197288558,0.0260773227,-0.0179867502,-0.4380299449,-0.106077157,-0.1333417147,0.1605287641,0.1040503159,0.0511585884,0.2821306288,-0.1498065442,-0.1852013916,-0.1185877919,-0.0300844684,0.0556870326,-0.3347616792,-0.0019837236,0.0977141634,0.3494845033,-0.0445081629,-0.1010216624,0.3270165324,-0.0091348113,-0.0247055497,-0.0248528197,0.1827057302,0.0151356449,-0.1065310314,-0.1402987987,0.3998171687,0.4279946685,0.0667518824,0.1130566821,-0.2174636126,-0.3045727611,-0.2593292296,0.0946189612,-0.18020989,-0.2480643541,0.1867720038,0.1791692674,0.1481642574,-0.2058574706,0.0145484982,0.1814024746,-0.0175495651,0.1872916371,0.2526591718,0.1082973257,-0.1395985186,0.2712236047,0.0994569212,0.0944782495,-0.1087876856,0.1274227649,-0.1205447018,0.0775215104,-0.2013966888,0.2472883761,-0.0433115438,-0.1436530948,0.1596127003,-0.137327686,0.2263962179,0.4787703454,-0.0233069286,0.3734986782,0.4101337492,0.0915490687,0.2185010165,0.2861803174,0.0649857223,0.4564537704,0.5201380253,0.3051720858,-0.3560695946,-0.1103175879,0.218999669,0.0466102362,0.1319053173,-0.1664593816,-0.0457989164,-0.0366330706,-0.0358387493,-0.1343392432,-0.5893675685,-0.1926735342,0.0665653795,0.2404344678,0.1041243598,0.5153942108,0.150959447,-0.0223057922,-0.0879775062,0.0629826114,-0.2974987626,0.1008124724,0.0913930088,0.1373390406,-0.0905936509,0.0450599641,-0.5651184916,0.0931614116,-0.0444074869,-0.1106403172,0.0804785788,-0.2457737923,0.0204879269,-0.1154539138,0.4060715735,0.1286485195,-0.1022803485,0.1689956337,0.2748243213,0.0741764829,-0.4765068293,0.0160645582,-0.1441310495,-0.1782052964,0.019906966,-0.0834218562,-0.1489900649,-0.1782407016,0.0712680817,-0.1267171502,-0.0334695205,-0.0527068377,0.098326318,0.0045979391,0.1004149541,0.2653541267,0.3687852919,0.0852377787,-0.2007394135,-0.4986420274,0.069699198,0.080859147,0.0986251533,0.1873314977,0.0979688913,0.1143833101,0.4632999897,-0.2479228079,0.0577324443,-0.3735258579,0.148249805,-0.4224083424,-0.0023336348,-0.1219204813,0.3359934986,0.0025945252,0.0205018464,0.1152354106,-0.0752210543,-0.023245547,0.0893219784,-0.2059785873,0.0207429081,-0.3161965907,0.4156787694,-0.0646224022,0.2781089842,-0.1155630574,-0.2146965116,-0.0343846455,-0.2159999907,-0.0555058233,0.0031665242,-0.0039882967,0.3707588315,-0.0244906247,-0.1919500828,0.2959436178,0.1552345306,0.1495912671,0.2460536063,-0.1522095948,0.2334620208,0.1154718697,0.2583086193,-0.2823133171,-0.1274650097,-0.0627356395,-0.0903532133,0.0303452499,-0.1563047618,0.3553386629,-0.0998722166,-0.2949932516,0.1143989265,0.0260068271,0.0625382736,-0.0148762632,-0.0783558413,-0.267486006,0.2445240021,-0.0303307567,0.2791244686,0.2451996654,0.0041028326,-0.0061915792,-0.2805902064,-0.1679349989,0.161710158,0.0688722134,-0.4143075347,-0.4462714791]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/129","title":"[Feature request] Add Google Natural Question dataset","comments":"NQ was added in #427 \ud83c\udf89","body":"Would be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.","comment_length":6,"text":"[Feature request] Add Google Natural Question dataset\nWould be great to have https:\/\/github.com\/google-research-datasets\/natural-questions as an alternative to SQuAD.\nNQ was added in #427 \ud83c\udf89","embeddings":[-0.2527337074,0.1290001571,-0.2742703259,-0.315703094,-0.0950332806,0.0368855484,0.3137984872,0.2305101454,0.0487668216,0.2945367098,-0.1171070486,0.3264758885,-0.2139460593,0.2714367509,0.5496010184,0.0337451957,0.0646804795,0.1489827335,0.2004495859,-0.036181841,-0.3141700327,0.1927435398,-0.2773861289,-0.1913581789,0.0563797355,-0.0995788649,-0.0115417577,0.068137534,-0.0337115787,-0.4519464672,-0.2582683563,0.2499224842,-0.0840723068,0.0860409215,-0.0001051355,-0.2827883065,0.1821442842,0.0513489395,-0.3456768394,-0.0347926281,-0.0174633246,-0.1295782775,-0.0183914416,0.0926863849,-0.2038366944,-0.0579449497,0.3608071506,-0.4548892975,-0.0826795846,0.5543115139,0.1982922703,-0.1357459575,-0.1412546635,-0.2538290024,0.2878347039,0.1678301245,-0.3066085875,0.076115936,0.3764795661,-0.0195343662,0.135050118,0.0169384181,0.0497187339,-0.1323202103,-0.0036079141,0.1171334237,-0.1289617866,-0.6843480468,-0.0792006105,0.2337775528,0.5078117847,-0.0194177665,-0.0291734226,-0.050647229,0.0237228852,0.0176606067,-0.1277789026,0.2541032732,-0.3335121274,-0.0488616265,-0.1667219251,-0.5302762985,0.0148112522,-0.0604420714,0.1267403811,0.1878856868,0.0769249722,-0.013652036,-0.076482296,-0.1557200998,0.0885438323,0.1056940779,-0.0105381068,0.0909256712,-0.4890934825,-0.4635598361,0.1700668335,-0.1397471279,0.3135802746,0.4097675383,-0.0336529203,0.0267333239,0.0929720029,-0.153264299,0.1566860825,-0.3028660119,0.4077055752,-0.054452695,0.2403903306,-0.1820364147,-0.0092216749,0.1100260392,0.0998199359,-0.1522863656,0.0112673771,0.131529972,0.4641113579,0.0166596919,0.033714205,-0.0051198248,0.0186543148,-0.0925260633,-0.1707203388,0.0841769725,0.0905071944,-0.383588016,-0.251455754,0.257920593,0.1832225919,-0.5213155746,-0.0850636661,-0.1121249497,0.1808356494,0.3054837883,0.353967905,-0.1076666266,0.1765501052,-0.007587634,0.0701817572,0.1814748198,0.2366342098,-0.0705352724,-0.0884149969,-0.1363772452,-0.0022061954,-0.2924681306,-0.0716486797,-0.0351047255,-0.2457170635,0.2450486273,-0.2556976974,-0.2154146731,-0.3790736198,0.2390798777,0.1283082664,-0.3648141623,0.1726303697,0.5687582493,-0.2461989671,-0.38498348,0.0731581599,0.2215532511,-0.242608726,-0.1652911901,-0.0710679516,0.2356467247,-0.3245968223,0.0232084766,-0.0809504017,0.0488465503,-0.128904283,0.0473971032,0.0808794796,0.1751401871,-0.130793184,0.0889322981,0.4615530968,-0.1307281554,-0.217887044,0.0049140095,-0.4037440419,-0.2407877147,-0.2903078198,0.4152729511,0.0318386927,0.0135575552,0.2827268839,0.5643343329,-0.1923328191,-0.0683899671,-0.0674415305,-0.3347279727,0.2583280206,0.4795278013,-0.0359135903,0.0341909938,0.3129233718,-0.5101425648,-0.247682482,-0.0871613845,-0.2336502671,0.0138206314,0.6307244301,-0.2623747885,0.0501167551,-0.2923575044,-0.7749915719,0.0240755491,-0.6488557458,0.4140391648,0.1793031991,-0.2475732118,-0.2037326992,-0.1768354625,-0.072310403,-0.1129630879,0.2492409199,0.090860039,-0.1940571815,0.3377588987,-0.1741749793,-0.0700844824,-0.0103842746,0.1269384623,-0.0142457495,0.1670178622,-0.0745927244,-0.0604112446,0.0944024846,0.1097969338,0.1730532944,0.2467653751,0.141350016,0.1317427307,-0.3262675107,0.2720399797,0.6080576777,-0.0690242127,0.0421452373,-0.3710555732,-0.0941862091,0.2859421074,-0.2532652318,0.000012043,-0.4063949585,0.0812307671,0.2368597686,-0.1317242384,-0.3069317043,0.149899289,-0.001648225,-0.0739688203,0.1767439544,0.0966764167,-0.1344906986,0.2761140764,0.1251755804,-0.2720324397,-0.2294206917,0.3365075886,0.4397426844,-0.0419717245,0.3540340066,0.1028771177,-0.0056468211,-0.3185594678,0.3663855791,-0.2820371389,0.2201289237,0.1901420951,0.1132575572,0.2351866215,0.0434673876,-0.1316639483,-0.1065895036,-0.0969796628,-0.0728295967,0.0138162207,0.4610729516,0.2030226737,-0.2871963382,-0.090868108,0.0363162085,0.00463332,-0.0433751531,-0.1785627604,-0.0443225689,0.134379819,0.3379104733,0.3540254235,-0.2060436606,-0.2249843329,0.3601371348,-0.1840630323,-0.1268028319,0.1610554755,0.23939161,0.7757292986,0.105584152,-0.4842696786,-0.0012578459,0.0235989112,0.154698953,0.2592198849,0.0224268176,0.0609948449,0.5688983798,-0.135776341,0.2915812433,-0.5896384716,-0.3888340294,0.1180806533,0.0731614381,0.7766543031,-0.0229488462,0.0574789084,-0.187044099,0.0810115933,-0.1593973488,-0.1545093656,-0.186725989,-0.3893817365,-0.0862162709,0.2318009883,-0.1634772569,-0.6823755503,-0.2215225995,-0.2928783,0.2829406261,0.0934055001,0.1294092536,0.1912019998,-0.245415628,0.0381811969,-0.0064684036,0.1050786674,-0.1935687363,0.059051156,0.1449239403,-0.3302711248,0.046135094,-0.0664934739,-0.037880864,-0.1172069982,0.032511998,-0.3810953498,0.058361616,-0.0117219724,0.6068057418,0.1660741568,-0.2003594041,0.4389730692,-0.0674028918,-0.1336413771,0.1393980831,-0.2596430182,0.0320508033,0.4961146414,0.0707495958,0.244841516,0.1379818916,-0.0141608212,0.7149358988,-0.053389743,0.2290126532,0.1140324026,-0.2784878016,0.2584958076,-0.153524667,-0.0387995876,0.3525215983,0.1350791603,0.014235273,0.2969287336,0.245875448,-0.4059833288,-0.0382097065,-0.0897934064,-0.1576552093,-0.0091014504,-0.0642541125,0.0345188007,0.1232458353,-0.021304952,-0.2291148156,-0.1180077493,-0.3479931951,-0.0982072651,0.0428679883,0.265231818,-0.1375091374,-0.8128235936,0.2275952399,-0.2039649636,0.2579613328,0.2609704733,-0.0928385332,-0.2700346708,-0.0286880638,-0.1032815129,0.0204335507,0.1612982303,-0.5830146074,-0.3195915222,-0.0381675139,0.2653430998,0.2786916494,0.1645591408,-0.0095897,0.0522523224,0.1778275073,0.3042349219,-0.119622156,-0.0252228379,0.4190097749,0.0646650791,0.0249595214,-0.1178389862,0.1149757653,-0.110998705,-0.0871423408,0.1515057236,0.0331177413,-0.0287523363,-0.1021471024,-0.3206798136,0.0019042036,0.1373140663,0.1875373572,-0.1037986279,-0.01302958,0.209539935,-0.0018618543,-0.0720406994,-0.2560637891,-0.2145930231,0.3589407504,-0.0616973191,-0.0580192618,0.1452374756,-0.1564146727,0.424186945,0.2587097883,0.239418909,0.1126338542,-0.0993095115,0.1253152639,0.0748432875,-0.0110378629,0.1852273643,-0.236016199,-0.1910108626,-0.1986269057,0.416857034,0.0221910439,-0.1971594095,0.2382443994,0.0902553648,0.0632430911,-0.2113408446,0.1874680072,0.6505560875,0.0538558885,0.0028812962,0.1546304822,-0.2217882276,0.4366540015,-0.1246380284,0.1061577573,-0.3624223173,-0.1800239384,0.0781555027,-0.0810363442,0.0260120481,0.1771468669,-0.3852204382,0.3052292168,0.1915695667,0.4466146231,0.2601419985,0.4316146374,0.1296646148,0.0560605116,0.2731465697,0.2569665313,-0.0664188266,0.1007289216,-0.1220439672,-0.1258832216,0.1515092701,0.0051607182,-0.1332656145,0.1164218038,0.0600696541,-0.1232271269,0.1770816594,-0.4235325754,0.1528260261,0.3555067182,0.3635520041,0.3504352868,-0.3028739989,0.0186076462,0.1460704803,0.1749701649,0.1557552516,0.0287564006,0.1323413998,-0.22983329,-0.2768349648,0.2929086089,0.0242806021,-0.4365566671,-0.7986387014,0.1129990816,-0.1979412884,-0.125644654,0.0068525434,0.0068251071,0.1123844758,-0.205389291,0.1827926189,-0.0333824195,0.325053066,-0.1652936935,0.4134865403,-0.0169382263,-0.1863029003,-0.2001411915,0.2485128641,-0.3346118033,0.0619173087,0.0632575601,-0.0622582287,-0.457968235,0.0942916796,-0.0391872004,-0.1936356723,0.1115753576,-0.2082905769,-0.1658214629,0.0546750836,0.0224101972,-0.2638406157,-0.0313324556,0.1341946572,0.0676268786,-0.0408276394,0.0776804313,-0.3442235887,0.1960053146,0.2953461409,-0.415920794,0.1143322363,0.1569093913,-0.4049677253,0.1075739935,-0.2897248566,0.0507724658,-0.1104954705,-0.1138477102,-0.0929012075,-0.0246881265,0.0403508469,-0.3996304274,-0.1170810238,-0.2464606613,0.1634318084,0.0493811741,-0.0032289203,0.2299975902,-0.0943709388,-0.1159803644,-0.173068732,0.1214524806,0.1072428077,-0.2527323067,0.1120522693,0.12668477,0.107425265,-0.0558374524,0.0536071509,0.3191287816,0.0256850123,-0.0519302711,-0.1439642608,0.142652303,-0.1402638704,-0.0331931524,-0.1843569279,0.3490825295,0.4414985478,0.1380598843,-0.0029099998,-0.2463584095,-0.174610883,-0.0654706582,-0.195113197,-0.2284046263,-0.3950070441,0.3261699975,0.1780872345,0.2563956678,-0.1801643223,-0.0787723139,0.2414267659,-0.0257768631,0.1900009811,0.1667548865,0.1547299922,-0.0524136946,0.5459644794,0.1356307119,0.1944392622,-0.1256396472,0.1363459677,-0.1736024469,0.0475440286,-0.1234202683,0.3765498698,0.0420401096,-0.2073365748,0.0657506809,-0.0311138853,0.0097791301,0.2182739973,0.0663951263,0.5363291502,0.2772472203,0.1966473311,0.3244958818,-0.0029351877,0.0041841939,0.3522984087,0.5268390179,-0.1565608531,-0.2269152403,-0.1771160215,0.2306699902,-0.0880824625,0.1161644608,-0.1536933035,-0.0917560831,0.0181249324,0.0223014615,-0.1196839511,-0.3771083653,-0.3395436108,-0.1010718122,0.0427574739,0.1885415018,0.4380831718,0.1258512586,-0.0147209344,-0.03131238,0.0306759682,-0.3445282578,0.1279976815,0.1890842319,0.3103131056,-0.1803544462,0.154103443,-0.4117557406,-0.0110979816,-0.0652974248,-0.154461354,0.075049296,-0.0848789364,0.1346468031,-0.076929301,0.3095929027,0.2322029173,-0.1700032204,0.3350429237,0.2265887558,0.1125898436,-0.3665296435,0.0489985533,-0.087690413,-0.13482894,0.1017734706,-0.1017195731,-0.0457222499,-0.1532407701,0.1228250712,-0.0673015565,0.0070868866,-0.0765011832,0.1543448269,-0.0650704801,-0.0126316706,0.1490267515,0.2957654893,0.1402834654,-0.1110872552,-0.4470304251,0.1024677753,0.2343010753,0.1918700486,0.1994575411,0.0819578469,0.0082944101,0.3595209122,-0.2569714189,-0.2254187316,-0.2453509122,-0.0826914832,-0.4532651901,0.0348186642,-0.306324631,0.2460764498,0.052349057,-0.100864239,0.2390150279,-0.1797226965,0.0369538888,-0.0081918435,-0.0842550844,0.0087671373,-0.1030402258,0.3571089506,-0.061859455,0.3873288035,-0.1535038501,-0.1953894645,-0.00413953,-0.2365693897,-0.1016502976,-0.1752275527,0.0727273822,0.3559392393,0.0657651648,-0.1092958897,0.3376969695,0.3359389007,0.2009320706,0.2751390636,-0.1573511958,0.0447387397,0.2051317841,0.0897679329,-0.0643929467,-0.1168532223,-0.0869660676,-0.1206421331,-0.006965402,-0.0899256095,0.307363987,0.1455935389,-0.1540305912,-0.2132871002,0.0868562981,0.2311483324,-0.0763985589,0.0085517196,-0.2524156868,0.3156168163,0.0405935384,0.1884806901,0.3922862113,-0.0378692783,0.0665828064,-0.1959484369,-0.1987729967,0.1618304849,-0.0211329795,-0.2869441807,-0.3070507348]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/128","title":"Some error inside nlp.load_dataset()","comments":"Google colab has an old version of Apache Arrow built-in.\r\nBe sure you execute the \"pip install\" cell and restart the notebook environment if the colab asks for it.","body":"First of all, nice work!\r\n\r\nI am going through [this overview notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb)\r\n\r\nIn simple step `dataset = nlp.load_dataset('squad', split='validation[:10%]')`\r\n\r\nI get an error, which is connected with some inner code, I think:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-8-d848d3a99b8c> in <module>()\r\n 1 # Downloading and loading a dataset\r\n 2 \r\n----> 3 dataset = nlp.load_dataset('squad', split='validation[:10%]')\r\n\r\n8 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 515 download_mode=download_mode,\r\n 516 ignore_verifications=ignore_verifications,\r\n--> 517 save_infos=save_infos,\r\n 518 )\r\n 519 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 361 verify_infos = not save_infos and not ignore_verifications\r\n 362 self._download_and_prepare(\r\n--> 363 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 364 )\r\n 365 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 414 try:\r\n 415 # Prepare split will record examples associated to the split\r\n--> 416 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 417 except OSError:\r\n 418 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 585 fname = \"{}-{}.arrow\".format(self.name, split_generator.name)\r\n 586 fpath = os.path.join(self._cache_dir, fname)\r\n--> 587 examples_type = self.info.features.type\r\n 588 writer = ArrowWriter(data_type=examples_type, path=fpath, writer_batch_size=self._writer_batch_size)\r\n 589 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in type(self)\r\n 460 @property\r\n 461 def type(self):\r\n--> 462 return get_nested_type(self)\r\n 463 \r\n 464 @classmethod\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in get_nested_type(schema)\r\n 370 # Nested structures: we allow dict, list\/tuples, sequences\r\n 371 if isinstance(schema, dict):\r\n--> 372 return pa.struct({key: get_nested_type(value) for key, value in schema.items()})\r\n 373 elif isinstance(schema, (list, tuple)):\r\n 374 assert len(schema) == 1, \"We defining list feature, you should just provide one example of the inner type\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in <dictcomp>(.0)\r\n 370 # Nested structures: we allow dict, list\/tuples, sequences\r\n 371 if isinstance(schema, dict):\r\n--> 372 return pa.struct({key: get_nested_type(value) for key, value in schema.items()})\r\n 373 elif isinstance(schema, (list, tuple)):\r\n 374 assert len(schema) == 1, \"We defining list feature, you should just provide one example of the inner type\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in get_nested_type(schema)\r\n 379 # We allow to reverse list of dict => dict of list for compatiblity with tfds\r\n 380 if isinstance(inner_type, pa.StructType):\r\n--> 381 return pa.struct(dict((f.name, pa.list_(f.type, schema.length)) for f in inner_type))\r\n 382 return pa.list_(inner_type, schema.length)\r\n 383 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in <genexpr>(.0)\r\n 379 # We allow to reverse list of dict => dict of list for compatiblity with tfds\r\n 380 if isinstance(inner_type, pa.StructType):\r\n--> 381 return pa.struct(dict((f.name, pa.list_(f.type, schema.length)) for f in inner_type))\r\n 382 return pa.list_(inner_type, schema.length)\r\n 383 \r\n\r\nTypeError: list_() takes exactly one argument (2 given)\r\n```","comment_length":29,"text":"Some error inside nlp.load_dataset()\nFirst of all, nice work!\r\n\r\nI am going through [this overview notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb)\r\n\r\nIn simple step `dataset = nlp.load_dataset('squad', split='validation[:10%]')`\r\n\r\nI get an error, which is connected with some inner code, I think:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-8-d848d3a99b8c> in <module>()\r\n 1 # Downloading and loading a dataset\r\n 2 \r\n----> 3 dataset = nlp.load_dataset('squad', split='validation[:10%]')\r\n\r\n8 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 515 download_mode=download_mode,\r\n 516 ignore_verifications=ignore_verifications,\r\n--> 517 save_infos=save_infos,\r\n 518 )\r\n 519 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 361 verify_infos = not save_infos and not ignore_verifications\r\n 362 self._download_and_prepare(\r\n--> 363 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 364 )\r\n 365 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 414 try:\r\n 415 # Prepare split will record examples associated to the split\r\n--> 416 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 417 except OSError:\r\n 418 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 585 fname = \"{}-{}.arrow\".format(self.name, split_generator.name)\r\n 586 fpath = os.path.join(self._cache_dir, fname)\r\n--> 587 examples_type = self.info.features.type\r\n 588 writer = ArrowWriter(data_type=examples_type, path=fpath, writer_batch_size=self._writer_batch_size)\r\n 589 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in type(self)\r\n 460 @property\r\n 461 def type(self):\r\n--> 462 return get_nested_type(self)\r\n 463 \r\n 464 @classmethod\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in get_nested_type(schema)\r\n 370 # Nested structures: we allow dict, list\/tuples, sequences\r\n 371 if isinstance(schema, dict):\r\n--> 372 return pa.struct({key: get_nested_type(value) for key, value in schema.items()})\r\n 373 elif isinstance(schema, (list, tuple)):\r\n 374 assert len(schema) == 1, \"We defining list feature, you should just provide one example of the inner type\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in <dictcomp>(.0)\r\n 370 # Nested structures: we allow dict, list\/tuples, sequences\r\n 371 if isinstance(schema, dict):\r\n--> 372 return pa.struct({key: get_nested_type(value) for key, value in schema.items()})\r\n 373 elif isinstance(schema, (list, tuple)):\r\n 374 assert len(schema) == 1, \"We defining list feature, you should just provide one example of the inner type\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in get_nested_type(schema)\r\n 379 # We allow to reverse list of dict => dict of list for compatiblity with tfds\r\n 380 if isinstance(inner_type, pa.StructType):\r\n--> 381 return pa.struct(dict((f.name, pa.list_(f.type, schema.length)) for f in inner_type))\r\n 382 return pa.list_(inner_type, schema.length)\r\n 383 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in <genexpr>(.0)\r\n 379 # We allow to reverse list of dict => dict of list for compatiblity with tfds\r\n 380 if isinstance(inner_type, pa.StructType):\r\n--> 381 return pa.struct(dict((f.name, pa.list_(f.type, schema.length)) for f in inner_type))\r\n 382 return pa.list_(inner_type, schema.length)\r\n 383 \r\n\r\nTypeError: list_() takes exactly one argument (2 given)\r\n```\nGoogle colab has an old version of Apache Arrow built-in.\r\nBe sure you execute the \"pip install\" cell and restart the notebook environment if the colab asks for it.","embeddings":[-0.0670081377,0.0724993795,-0.0823872089,0.1897586733,0.267567426,0.0750563815,0.2924349904,0.3888992071,0.0351886898,-0.1877863556,-0.1114239022,0.2747960091,-0.0162892751,0.0643214062,0.2407227159,-0.2864512801,-0.1546422988,0.2549157739,-0.0124721359,0.1232353002,-0.2655300796,0.1166809052,-0.3398716748,0.3442742527,-0.0802765116,-0.1929238886,-0.0296067595,0.3408279419,-0.2490498424,-0.5879958868,0.3756798208,-0.1130219474,0.2241377831,0.1312045604,-0.0001051211,0.0366863534,0.4223832488,-0.0525458157,-0.4962961674,-0.2144066691,-0.2681881189,-0.1776150465,0.2765499055,-0.5071807504,0.3534993827,0.0584660582,0.1453215331,0.172617957,0.4719077647,0.3926600218,0.3172274828,0.4065113962,0.1332868487,-0.0378221758,-0.0242808741,-0.0132257333,0.0424994789,0.3613389134,-0.0597006828,-0.2916482091,-0.1058744639,0.1836791486,-0.0739921257,0.0822866559,0.2149083763,-0.0782490224,0.1761753112,-0.2062117159,-0.2290777713,0.0764528736,0.310865581,-0.3317778707,-0.1878501326,-0.2332880795,0.1658896357,-0.286799103,0.2173205912,0.2049746066,-0.2205486894,-0.0430262759,-0.1023475379,0.0367132388,-0.1697550267,0.2968499064,0.1530664116,0.1422135085,-0.0430708043,0.0544458181,0.1611634493,0.0255567972,0.1349268258,-0.1027902737,-0.0119791068,0.3151762187,-0.3757234514,-0.2236605138,0.0898646191,0.1502705365,0.0355190895,0.0423011705,0.1994195133,-0.1857152432,0.0624949113,0.3097718954,0.3490634561,0.2397695631,0.3099018633,0.1482294202,0.111098066,0.2224611193,0.0778009146,0.040816877,-0.0988591313,-0.2621631026,0.239755556,0.0685242489,0.3718980551,-0.0113207232,-0.4528195858,0.0513457023,-0.2397605479,-0.0740910098,0.1411995888,0.2162542045,0.0511003211,0.0173275508,0.2893837094,0.0570257194,-0.2770854235,-0.3184550405,-0.1165348962,0.2498125285,-0.3269898295,0.0069971788,0.1883507669,0.0352810137,0.308734715,0.0512832366,-0.0036526511,-0.1259551197,0.2600570321,-0.1636987627,-0.2521975636,0.1093505993,-0.0698628202,-0.1896069795,0.1693443358,-0.2162506133,-0.0904491022,-0.0141093191,-0.1204682961,-0.3169983029,-0.2073783726,0.2719335258,-0.1719684303,0.0835194215,-0.173584044,-0.1093321443,0.2492954582,-0.3055495918,0.0167168751,-0.3455445468,-0.1564008296,-0.1015432626,0.0918391347,0.3969186246,0.0699042231,-0.1808037758,-0.2149003297,-0.2028656602,-0.0220352262,0.2457960546,-0.2600751817,0.622546494,-0.1959222555,0.2986060679,0.6382418275,-0.3533470035,-0.2437379956,0.2599744499,-0.0882795528,0.0478558242,-0.2249871194,0.0980274081,0.1081014648,-0.0667474642,0.3803537786,0.510879159,-0.1190070361,0.0474854112,-0.1602506638,0.0501506105,0.290140897,0.1346265078,-0.0792289376,0.1688669324,0.0797369778,0.6542676091,0.0420449451,-0.0276956894,-0.1052355617,-0.0001355516,-0.0991272107,-0.1206663325,-0.1381555945,-0.2238782793,-0.5490896106,-0.0626133382,-0.4242883027,0.0873748809,-0.013877809,-0.0138693107,-0.4102626741,0.0395544954,-0.2429813296,-0.1580200195,0.233346194,0.1386093199,-0.1312867552,-0.1246051863,-0.0883770362,0.3406808376,-0.3272353113,0.0233292747,-0.5331027508,0.2011083066,-0.2686674297,-0.1432891786,0.1279355735,0.2801431417,0.0276369806,-0.0509063974,-0.1709376276,0.4143425524,-0.0930758566,0.0388934501,0.0303059649,-0.1917450279,-0.0529045127,-0.0611756183,0.0195933953,0.1188744232,0.2759116292,-0.0820460692,0.1410550177,0.2947699726,-0.0729208514,0.07114885,0.0039901254,0.1084019095,0.1370729059,-0.1486365795,-0.0938241631,-0.0202759448,0.3909056187,0.2834856808,0.2081480324,-0.0288621206,-0.2282043695,-0.1797693074,0.4322507083,-0.0689732358,0.1922805607,0.2028542161,0.1720713526,-0.1804119796,-0.0145860836,-0.0284012984,0.541462481,0.0389190987,-0.3035569489,0.2416815013,-0.0952259153,-0.0702836663,0.2528953254,0.06303069,0.1794157326,0.2073055506,0.1111282334,0.1453695595,-0.3022251725,-0.2434382737,0.1295956373,0.3865086734,-0.3687125742,0.3285272717,-0.0910153762,-0.5053382516,-0.1362622678,-0.0972985923,-0.1995762736,-0.3681539893,-0.2554415464,0.197640568,0.0036706775,0.0604116842,-0.0738964379,0.1609850377,0.079625167,0.0558333285,-0.0202164091,-0.2739411294,-0.1665540338,0.0571618229,0.4820735753,0.0963714942,0.3322121799,-0.2350907773,-0.2184936702,0.0778685883,-0.3148480058,0.0317500941,-0.0699617192,0.2388782352,0.3108826578,0.21828942,-0.096302405,-0.2679318786,0.3181727231,-0.0013899575,-0.3025505543,0.1445501298,0.0671869069,0.0353902094,-0.1147907376,-0.4487812817,-0.4314314425,-0.4160797298,0.168674171,-0.156861797,0.2440702766,0.4337669313,-0.2115263343,0.3001113534,-0.1510654837,-0.0303636268,-0.1450505853,-0.0210048556,0.139748618,-0.2263175845,-0.3450107276,-0.2016336471,-0.0866467878,0.2221891433,-0.2155761123,-0.4385277331,-0.0517624728,-0.0057823048,0.1955138892,-0.2977797985,-0.1599109322,0.2887176275,-0.0905646086,-0.1065410152,-0.016771609,-0.1417757273,-0.0614282452,0.0821514055,0.3034346104,-0.1351955384,0.4471290112,-0.1231755018,0.2869756818,0.0732987374,-0.1082974225,0.3293793499,0.1539563686,0.2039957941,-0.2005158216,-0.4926657081,0.2181332558,-0.1572080106,0.0515012555,0.028721964,-0.3280373514,-0.3047070503,-0.2021842599,-0.1398532987,-0.3305560946,-0.2503878474,0.2730174959,-0.0773454905,0.171475485,0.1585309803,-0.0780505911,-0.2217248529,-0.0377909839,-0.2130164653,0.319760114,0.1046187356,-0.058427982,-0.297986567,-0.2966352403,-0.1739130169,0.1791519821,0.2524517179,0.4256486595,-0.1788269728,0.1140436083,0.0183622744,0.067863062,0.1849156171,0.1055376679,0.1085020453,0.3210605383,0.0974963158,-0.3292140365,-0.0738473982,-0.325737983,0.5837900639,0.5781292915,0.0980626494,-0.2752339542,-0.0338024981,0.483458966,0.0486783423,-0.2065960318,-0.3602990508,-0.2959495187,-0.1465177685,-0.3288712502,-0.024636358,-0.025520673,0.2704153061,0.0710084662,0.2534430325,0.1216343418,0.09270484,0.0010231338,0.0216788035,0.0657800436,0.1994562894,-0.0630532056,-0.2098067403,0.1590036899,-0.0702687502,0.3327355683,0.0725932494,-0.205061093,0.1818727404,0.0516673587,0.2031390965,0.1788525134,0.0591571778,0.0491884798,-0.08486709,0.0960116386,-0.0209740791,0.1371421516,0.3824803829,0.0255470555,-0.2554465532,-0.3978533745,0.4386381805,-0.3213837743,0.0976416245,0.3634882569,-0.2074615657,-0.2057627738,0.1143071502,0.0049699652,0.7501723766,0.2101755142,0.2123607546,0.5577793121,0.0075994348,0.4680124521,0.0951069742,0.2008805424,-0.3666546941,0.1826377809,0.0357554182,-0.1490111351,0.0260926094,-0.0038814822,-0.2142940909,0.1134230122,-0.045701921,-0.0226703752,0.1832748204,0.4041080773,-0.075043492,-0.1945253164,-0.6791546345,0.2014897615,-0.4122564793,0.1889532804,-0.2822135985,-0.2469041348,0.0997457504,-0.1837922633,-0.2537021935,0.2055188566,-0.0916898698,0.3260078132,0.1200041324,-0.1432432979,0.2602951229,0.2704222202,0.0525325984,-0.0416510701,-0.028946707,-0.0233650003,-0.1941991895,-0.1629676372,0.0426979363,0.121391423,0.3357355595,-0.1928718835,-0.3291578889,0.1110835746,-0.0343966708,-0.3594432175,0.1536705941,0.1568349153,0.1917549074,-0.3811148405,-0.3047457635,-0.2029924691,-0.143666774,-0.1514397711,0.2000862807,0.0843164995,-0.2529889643,0.0350682214,-0.009234854,-0.2911499739,0.180242613,0.2770356536,0.0140180718,0.0106047653,0.7288349867,0.0609149821,-0.1173972413,-0.3024675548,0.3176200688,-0.0881229565,-0.3163417578,0.1033362895,-0.0070597157,0.1721570641,-0.2600637674,0.2397180796,-0.0039600749,0.202087909,0.2013679445,-0.703392446,-0.322152853,0.1701672822,-0.1725695729,-0.0260471031,0.2704144716,0.269151032,0.1669470072,0.1838987321,-0.3495567441,-0.1555121094,-0.3398022652,0.2904283404,0.2434929907,-0.3183088601,-0.0067416327,-0.0603752173,0.1436004639,0.1854965389,-0.1724689752,-0.3484415412,-0.0325862393,0.1216217205,0.0408617109,-0.0962840319,0.0424279943,-0.0707946792,-0.1155121252,-0.0079113021,-0.1634440571,-0.1701862514,0.0744745508,0.3102091253,0.0170277562,0.2306427956,0.1136413664,0.1596212238,-0.1087767109,0.1454434991,-0.0863041356,0.2423229069,-0.0232036151,0.0559719913,-0.3918284476,-0.1013244018,-0.0728679448,0.21742028,0.1941603124,-0.4284442961,0.0244407319,-0.2036377639,0.0187180359,0.4487906396,-0.3464199603,-0.0054861447,-0.053349223,0.1968188733,-0.2553157806,0.2166371346,0.3410317898,0.0716980919,-0.1447334588,-0.0323971212,-0.1158764958,0.045765236,-0.0180209335,0.0056249658,0.1477858424,-0.0067933681,0.1899334788,0.214968577,-0.0566866286,0.052880343,0.1113122925,0.1312787682,0.133976832,0.2176308334,-0.1340116113,0.1117198914,-0.1489599496,0.1516866684,0.1375527233,-0.2686516643,-0.2173861116,0.3156462908,0.2006194592,0.130423665,0.0018125098,0.3766193688,-0.1620232165,-0.1020948291,-0.0075148148,0.1957176328,-0.1902849376,0.0911125615,-0.5262409449,0.0250650346,-0.1286998391,-0.0458877571,-0.0549714193,0.0364595614,0.2260546088,0.2589622438,-0.2998779118,-0.3869462907,0.1207441464,0.3827261627,-0.1209206283,-0.3748805225,0.2868905365,0.2392374426,0.1293732226,0.0250822231,0.6040561199,0.5194359422,0.4611826837,-0.2293996066,0.0646776184,-0.1322274059,-0.0579322875,0.072303243,0.0647797212,0.3152109385,0.1912102103,0.2648988068,0.1848494858,-0.1656361222,0.1554615796,0.3279311955,-0.2607153952,-0.1104304567,0.4586729407,-0.004156006,-0.0905640796,-0.3590586782,0.2188563496,-0.3501581252,0.2154807001,0.2738781273,0.3120318949,0.0712650567,-0.1063841432,0.108919248,-0.1599286944,0.703853786,0.3672445714,0.1457680762,-0.5256857276,-0.2018235773,-0.7177901268,0.1828651577,-0.0225400962,-0.137879312,0.288698554,0.0494952835,-0.1539878547,0.2997170389,-0.0527194403,0.0517334007,-0.0982843265,-0.053851448,-0.3332030773,-0.1750536859,0.1419279128,0.0070070261,-0.1369071603,-0.3033785224,0.0991408378,-0.2753921747,0.0693376288,-0.1080029532,-0.190194115,0.0246990621,0.440284133,0.0714419782,0.1907030344,0.5019017458,-0.0452676229,-0.0860020146,-0.0191513691,-0.2989412248,-0.092126064,0.2473773211,0.1353030801,0.3523185849,-0.1346607953,0.1831392497,-0.4423944056,0.495952785,-0.277638495,-0.3152654469,-0.2573377788,-0.107219696,-0.0311579686,0.2137938589,0.0032327722,0.1233839318,-0.0819337443,0.2954353988,-0.4231437743,-0.3038302362,0.5512321591,-0.4347351193,-0.2962920964,0.167625621,0.2419217825,-0.1240828708,-0.1673004329,-0.5922073722,0.0424591042,0.3712547123,-0.1085435748,-0.2428871244,0.2297098637,-0.1221716404,0.1525463164,0.0340049341,0.3174511194,0.0109068071,-0.2272163779,0.0310515221,-0.3468204141]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/128","title":"Some error inside nlp.load_dataset()","comments":"Thanks for reply, worked fine!\r\n","body":"First of all, nice work!\r\n\r\nI am going through [this overview notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb)\r\n\r\nIn simple step `dataset = nlp.load_dataset('squad', split='validation[:10%]')`\r\n\r\nI get an error, which is connected with some inner code, I think:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-8-d848d3a99b8c> in <module>()\r\n 1 # Downloading and loading a dataset\r\n 2 \r\n----> 3 dataset = nlp.load_dataset('squad', split='validation[:10%]')\r\n\r\n8 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 515 download_mode=download_mode,\r\n 516 ignore_verifications=ignore_verifications,\r\n--> 517 save_infos=save_infos,\r\n 518 )\r\n 519 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 361 verify_infos = not save_infos and not ignore_verifications\r\n 362 self._download_and_prepare(\r\n--> 363 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 364 )\r\n 365 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 414 try:\r\n 415 # Prepare split will record examples associated to the split\r\n--> 416 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 417 except OSError:\r\n 418 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 585 fname = \"{}-{}.arrow\".format(self.name, split_generator.name)\r\n 586 fpath = os.path.join(self._cache_dir, fname)\r\n--> 587 examples_type = self.info.features.type\r\n 588 writer = ArrowWriter(data_type=examples_type, path=fpath, writer_batch_size=self._writer_batch_size)\r\n 589 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in type(self)\r\n 460 @property\r\n 461 def type(self):\r\n--> 462 return get_nested_type(self)\r\n 463 \r\n 464 @classmethod\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in get_nested_type(schema)\r\n 370 # Nested structures: we allow dict, list\/tuples, sequences\r\n 371 if isinstance(schema, dict):\r\n--> 372 return pa.struct({key: get_nested_type(value) for key, value in schema.items()})\r\n 373 elif isinstance(schema, (list, tuple)):\r\n 374 assert len(schema) == 1, \"We defining list feature, you should just provide one example of the inner type\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in <dictcomp>(.0)\r\n 370 # Nested structures: we allow dict, list\/tuples, sequences\r\n 371 if isinstance(schema, dict):\r\n--> 372 return pa.struct({key: get_nested_type(value) for key, value in schema.items()})\r\n 373 elif isinstance(schema, (list, tuple)):\r\n 374 assert len(schema) == 1, \"We defining list feature, you should just provide one example of the inner type\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in get_nested_type(schema)\r\n 379 # We allow to reverse list of dict => dict of list for compatiblity with tfds\r\n 380 if isinstance(inner_type, pa.StructType):\r\n--> 381 return pa.struct(dict((f.name, pa.list_(f.type, schema.length)) for f in inner_type))\r\n 382 return pa.list_(inner_type, schema.length)\r\n 383 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in <genexpr>(.0)\r\n 379 # We allow to reverse list of dict => dict of list for compatiblity with tfds\r\n 380 if isinstance(inner_type, pa.StructType):\r\n--> 381 return pa.struct(dict((f.name, pa.list_(f.type, schema.length)) for f in inner_type))\r\n 382 return pa.list_(inner_type, schema.length)\r\n 383 \r\n\r\nTypeError: list_() takes exactly one argument (2 given)\r\n```","comment_length":5,"text":"Some error inside nlp.load_dataset()\nFirst of all, nice work!\r\n\r\nI am going through [this overview notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb)\r\n\r\nIn simple step `dataset = nlp.load_dataset('squad', split='validation[:10%]')`\r\n\r\nI get an error, which is connected with some inner code, I think:\r\n```\r\n---------------------------------------------------------------------------\r\n\r\nTypeError Traceback (most recent call last)\r\n\r\n<ipython-input-8-d848d3a99b8c> in <module>()\r\n 1 # Downloading and loading a dataset\r\n 2 \r\n----> 3 dataset = nlp.load_dataset('squad', split='validation[:10%]')\r\n\r\n8 frames\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs)\r\n 515 download_mode=download_mode,\r\n 516 ignore_verifications=ignore_verifications,\r\n--> 517 save_infos=save_infos,\r\n 518 )\r\n 519 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, dl_manager, **download_and_prepare_kwargs)\r\n 361 verify_infos = not save_infos and not ignore_verifications\r\n 362 self._download_and_prepare(\r\n--> 363 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs\r\n 364 )\r\n 365 # Sync info\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 414 try:\r\n 415 # Prepare split will record examples associated to the split\r\n--> 416 self._prepare_split(split_generator, **prepare_split_kwargs)\r\n 417 except OSError:\r\n 418 raise OSError(\"Cannot find data file. \" + (self.MANUAL_DOWNLOAD_INSTRUCTIONS or \"\"))\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/builder.py in _prepare_split(self, split_generator)\r\n 585 fname = \"{}-{}.arrow\".format(self.name, split_generator.name)\r\n 586 fpath = os.path.join(self._cache_dir, fname)\r\n--> 587 examples_type = self.info.features.type\r\n 588 writer = ArrowWriter(data_type=examples_type, path=fpath, writer_batch_size=self._writer_batch_size)\r\n 589 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in type(self)\r\n 460 @property\r\n 461 def type(self):\r\n--> 462 return get_nested_type(self)\r\n 463 \r\n 464 @classmethod\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in get_nested_type(schema)\r\n 370 # Nested structures: we allow dict, list\/tuples, sequences\r\n 371 if isinstance(schema, dict):\r\n--> 372 return pa.struct({key: get_nested_type(value) for key, value in schema.items()})\r\n 373 elif isinstance(schema, (list, tuple)):\r\n 374 assert len(schema) == 1, \"We defining list feature, you should just provide one example of the inner type\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in <dictcomp>(.0)\r\n 370 # Nested structures: we allow dict, list\/tuples, sequences\r\n 371 if isinstance(schema, dict):\r\n--> 372 return pa.struct({key: get_nested_type(value) for key, value in schema.items()})\r\n 373 elif isinstance(schema, (list, tuple)):\r\n 374 assert len(schema) == 1, \"We defining list feature, you should just provide one example of the inner type\"\r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in get_nested_type(schema)\r\n 379 # We allow to reverse list of dict => dict of list for compatiblity with tfds\r\n 380 if isinstance(inner_type, pa.StructType):\r\n--> 381 return pa.struct(dict((f.name, pa.list_(f.type, schema.length)) for f in inner_type))\r\n 382 return pa.list_(inner_type, schema.length)\r\n 383 \r\n\r\n\/usr\/local\/lib\/python3.6\/dist-packages\/nlp\/features.py in <genexpr>(.0)\r\n 379 # We allow to reverse list of dict => dict of list for compatiblity with tfds\r\n 380 if isinstance(inner_type, pa.StructType):\r\n--> 381 return pa.struct(dict((f.name, pa.list_(f.type, schema.length)) for f in inner_type))\r\n 382 return pa.list_(inner_type, schema.length)\r\n 383 \r\n\r\nTypeError: list_() takes exactly one argument (2 given)\r\n```\nThanks for reply, worked fine!\r\n","embeddings":[-0.0670081377,0.0724993795,-0.0823872089,0.1897586733,0.267567426,0.0750563815,0.2924349904,0.3888992071,0.0351886898,-0.1877863556,-0.1114239022,0.2747960091,-0.0162892751,0.0643214062,0.2407227159,-0.2864512801,-0.1546422988,0.2549157739,-0.0124721359,0.1232353002,-0.2655300796,0.1166809052,-0.3398716748,0.3442742527,-0.0802765116,-0.1929238886,-0.0296067595,0.3408279419,-0.2490498424,-0.5879958868,0.3756798208,-0.1130219474,0.2241377831,0.1312045604,-0.0001051211,0.0366863534,0.4223832488,-0.0525458157,-0.4962961674,-0.2144066691,-0.2681881189,-0.1776150465,0.2765499055,-0.5071807504,0.3534993827,0.0584660582,0.1453215331,0.172617957,0.4719077647,0.3926600218,0.3172274828,0.4065113962,0.1332868487,-0.0378221758,-0.0242808741,-0.0132257333,0.0424994789,0.3613389134,-0.0597006828,-0.2916482091,-0.1058744639,0.1836791486,-0.0739921257,0.0822866559,0.2149083763,-0.0782490224,0.1761753112,-0.2062117159,-0.2290777713,0.0764528736,0.310865581,-0.3317778707,-0.1878501326,-0.2332880795,0.1658896357,-0.286799103,0.2173205912,0.2049746066,-0.2205486894,-0.0430262759,-0.1023475379,0.0367132388,-0.1697550267,0.2968499064,0.1530664116,0.1422135085,-0.0430708043,0.0544458181,0.1611634493,0.0255567972,0.1349268258,-0.1027902737,-0.0119791068,0.3151762187,-0.3757234514,-0.2236605138,0.0898646191,0.1502705365,0.0355190895,0.0423011705,0.1994195133,-0.1857152432,0.0624949113,0.3097718954,0.3490634561,0.2397695631,0.3099018633,0.1482294202,0.111098066,0.2224611193,0.0778009146,0.040816877,-0.0988591313,-0.2621631026,0.239755556,0.0685242489,0.3718980551,-0.0113207232,-0.4528195858,0.0513457023,-0.2397605479,-0.0740910098,0.1411995888,0.2162542045,0.0511003211,0.0173275508,0.2893837094,0.0570257194,-0.2770854235,-0.3184550405,-0.1165348962,0.2498125285,-0.3269898295,0.0069971788,0.1883507669,0.0352810137,0.308734715,0.0512832366,-0.0036526511,-0.1259551197,0.2600570321,-0.1636987627,-0.2521975636,0.1093505993,-0.0698628202,-0.1896069795,0.1693443358,-0.2162506133,-0.0904491022,-0.0141093191,-0.1204682961,-0.3169983029,-0.2073783726,0.2719335258,-0.1719684303,0.0835194215,-0.173584044,-0.1093321443,0.2492954582,-0.3055495918,0.0167168751,-0.3455445468,-0.1564008296,-0.1015432626,0.0918391347,0.3969186246,0.0699042231,-0.1808037758,-0.2149003297,-0.2028656602,-0.0220352262,0.2457960546,-0.2600751817,0.622546494,-0.1959222555,0.2986060679,0.6382418275,-0.3533470035,-0.2437379956,0.2599744499,-0.0882795528,0.0478558242,-0.2249871194,0.0980274081,0.1081014648,-0.0667474642,0.3803537786,0.510879159,-0.1190070361,0.0474854112,-0.1602506638,0.0501506105,0.290140897,0.1346265078,-0.0792289376,0.1688669324,0.0797369778,0.6542676091,0.0420449451,-0.0276956894,-0.1052355617,-0.0001355516,-0.0991272107,-0.1206663325,-0.1381555945,-0.2238782793,-0.5490896106,-0.0626133382,-0.4242883027,0.0873748809,-0.013877809,-0.0138693107,-0.4102626741,0.0395544954,-0.2429813296,-0.1580200195,0.233346194,0.1386093199,-0.1312867552,-0.1246051863,-0.0883770362,0.3406808376,-0.3272353113,0.0233292747,-0.5331027508,0.2011083066,-0.2686674297,-0.1432891786,0.1279355735,0.2801431417,0.0276369806,-0.0509063974,-0.1709376276,0.4143425524,-0.0930758566,0.0388934501,0.0303059649,-0.1917450279,-0.0529045127,-0.0611756183,0.0195933953,0.1188744232,0.2759116292,-0.0820460692,0.1410550177,0.2947699726,-0.0729208514,0.07114885,0.0039901254,0.1084019095,0.1370729059,-0.1486365795,-0.0938241631,-0.0202759448,0.3909056187,0.2834856808,0.2081480324,-0.0288621206,-0.2282043695,-0.1797693074,0.4322507083,-0.0689732358,0.1922805607,0.2028542161,0.1720713526,-0.1804119796,-0.0145860836,-0.0284012984,0.541462481,0.0389190987,-0.3035569489,0.2416815013,-0.0952259153,-0.0702836663,0.2528953254,0.06303069,0.1794157326,0.2073055506,0.1111282334,0.1453695595,-0.3022251725,-0.2434382737,0.1295956373,0.3865086734,-0.3687125742,0.3285272717,-0.0910153762,-0.5053382516,-0.1362622678,-0.0972985923,-0.1995762736,-0.3681539893,-0.2554415464,0.197640568,0.0036706775,0.0604116842,-0.0738964379,0.1609850377,0.079625167,0.0558333285,-0.0202164091,-0.2739411294,-0.1665540338,0.0571618229,0.4820735753,0.0963714942,0.3322121799,-0.2350907773,-0.2184936702,0.0778685883,-0.3148480058,0.0317500941,-0.0699617192,0.2388782352,0.3108826578,0.21828942,-0.096302405,-0.2679318786,0.3181727231,-0.0013899575,-0.3025505543,0.1445501298,0.0671869069,0.0353902094,-0.1147907376,-0.4487812817,-0.4314314425,-0.4160797298,0.168674171,-0.156861797,0.2440702766,0.4337669313,-0.2115263343,0.3001113534,-0.1510654837,-0.0303636268,-0.1450505853,-0.0210048556,0.139748618,-0.2263175845,-0.3450107276,-0.2016336471,-0.0866467878,0.2221891433,-0.2155761123,-0.4385277331,-0.0517624728,-0.0057823048,0.1955138892,-0.2977797985,-0.1599109322,0.2887176275,-0.0905646086,-0.1065410152,-0.016771609,-0.1417757273,-0.0614282452,0.0821514055,0.3034346104,-0.1351955384,0.4471290112,-0.1231755018,0.2869756818,0.0732987374,-0.1082974225,0.3293793499,0.1539563686,0.2039957941,-0.2005158216,-0.4926657081,0.2181332558,-0.1572080106,0.0515012555,0.028721964,-0.3280373514,-0.3047070503,-0.2021842599,-0.1398532987,-0.3305560946,-0.2503878474,0.2730174959,-0.0773454905,0.171475485,0.1585309803,-0.0780505911,-0.2217248529,-0.0377909839,-0.2130164653,0.319760114,0.1046187356,-0.058427982,-0.297986567,-0.2966352403,-0.1739130169,0.1791519821,0.2524517179,0.4256486595,-0.1788269728,0.1140436083,0.0183622744,0.067863062,0.1849156171,0.1055376679,0.1085020453,0.3210605383,0.0974963158,-0.3292140365,-0.0738473982,-0.325737983,0.5837900639,0.5781292915,0.0980626494,-0.2752339542,-0.0338024981,0.483458966,0.0486783423,-0.2065960318,-0.3602990508,-0.2959495187,-0.1465177685,-0.3288712502,-0.024636358,-0.025520673,0.2704153061,0.0710084662,0.2534430325,0.1216343418,0.09270484,0.0010231338,0.0216788035,0.0657800436,0.1994562894,-0.0630532056,-0.2098067403,0.1590036899,-0.0702687502,0.3327355683,0.0725932494,-0.205061093,0.1818727404,0.0516673587,0.2031390965,0.1788525134,0.0591571778,0.0491884798,-0.08486709,0.0960116386,-0.0209740791,0.1371421516,0.3824803829,0.0255470555,-0.2554465532,-0.3978533745,0.4386381805,-0.3213837743,0.0976416245,0.3634882569,-0.2074615657,-0.2057627738,0.1143071502,0.0049699652,0.7501723766,0.2101755142,0.2123607546,0.5577793121,0.0075994348,0.4680124521,0.0951069742,0.2008805424,-0.3666546941,0.1826377809,0.0357554182,-0.1490111351,0.0260926094,-0.0038814822,-0.2142940909,0.1134230122,-0.045701921,-0.0226703752,0.1832748204,0.4041080773,-0.075043492,-0.1945253164,-0.6791546345,0.2014897615,-0.4122564793,0.1889532804,-0.2822135985,-0.2469041348,0.0997457504,-0.1837922633,-0.2537021935,0.2055188566,-0.0916898698,0.3260078132,0.1200041324,-0.1432432979,0.2602951229,0.2704222202,0.0525325984,-0.0416510701,-0.028946707,-0.0233650003,-0.1941991895,-0.1629676372,0.0426979363,0.121391423,0.3357355595,-0.1928718835,-0.3291578889,0.1110835746,-0.0343966708,-0.3594432175,0.1536705941,0.1568349153,0.1917549074,-0.3811148405,-0.3047457635,-0.2029924691,-0.143666774,-0.1514397711,0.2000862807,0.0843164995,-0.2529889643,0.0350682214,-0.009234854,-0.2911499739,0.180242613,0.2770356536,0.0140180718,0.0106047653,0.7288349867,0.0609149821,-0.1173972413,-0.3024675548,0.3176200688,-0.0881229565,-0.3163417578,0.1033362895,-0.0070597157,0.1721570641,-0.2600637674,0.2397180796,-0.0039600749,0.202087909,0.2013679445,-0.703392446,-0.322152853,0.1701672822,-0.1725695729,-0.0260471031,0.2704144716,0.269151032,0.1669470072,0.1838987321,-0.3495567441,-0.1555121094,-0.3398022652,0.2904283404,0.2434929907,-0.3183088601,-0.0067416327,-0.0603752173,0.1436004639,0.1854965389,-0.1724689752,-0.3484415412,-0.0325862393,0.1216217205,0.0408617109,-0.0962840319,0.0424279943,-0.0707946792,-0.1155121252,-0.0079113021,-0.1634440571,-0.1701862514,0.0744745508,0.3102091253,0.0170277562,0.2306427956,0.1136413664,0.1596212238,-0.1087767109,0.1454434991,-0.0863041356,0.2423229069,-0.0232036151,0.0559719913,-0.3918284476,-0.1013244018,-0.0728679448,0.21742028,0.1941603124,-0.4284442961,0.0244407319,-0.2036377639,0.0187180359,0.4487906396,-0.3464199603,-0.0054861447,-0.053349223,0.1968188733,-0.2553157806,0.2166371346,0.3410317898,0.0716980919,-0.1447334588,-0.0323971212,-0.1158764958,0.045765236,-0.0180209335,0.0056249658,0.1477858424,-0.0067933681,0.1899334788,0.214968577,-0.0566866286,0.052880343,0.1113122925,0.1312787682,0.133976832,0.2176308334,-0.1340116113,0.1117198914,-0.1489599496,0.1516866684,0.1375527233,-0.2686516643,-0.2173861116,0.3156462908,0.2006194592,0.130423665,0.0018125098,0.3766193688,-0.1620232165,-0.1020948291,-0.0075148148,0.1957176328,-0.1902849376,0.0911125615,-0.5262409449,0.0250650346,-0.1286998391,-0.0458877571,-0.0549714193,0.0364595614,0.2260546088,0.2589622438,-0.2998779118,-0.3869462907,0.1207441464,0.3827261627,-0.1209206283,-0.3748805225,0.2868905365,0.2392374426,0.1293732226,0.0250822231,0.6040561199,0.5194359422,0.4611826837,-0.2293996066,0.0646776184,-0.1322274059,-0.0579322875,0.072303243,0.0647797212,0.3152109385,0.1912102103,0.2648988068,0.1848494858,-0.1656361222,0.1554615796,0.3279311955,-0.2607153952,-0.1104304567,0.4586729407,-0.004156006,-0.0905640796,-0.3590586782,0.2188563496,-0.3501581252,0.2154807001,0.2738781273,0.3120318949,0.0712650567,-0.1063841432,0.108919248,-0.1599286944,0.703853786,0.3672445714,0.1457680762,-0.5256857276,-0.2018235773,-0.7177901268,0.1828651577,-0.0225400962,-0.137879312,0.288698554,0.0494952835,-0.1539878547,0.2997170389,-0.0527194403,0.0517334007,-0.0982843265,-0.053851448,-0.3332030773,-0.1750536859,0.1419279128,0.0070070261,-0.1369071603,-0.3033785224,0.0991408378,-0.2753921747,0.0693376288,-0.1080029532,-0.190194115,0.0246990621,0.440284133,0.0714419782,0.1907030344,0.5019017458,-0.0452676229,-0.0860020146,-0.0191513691,-0.2989412248,-0.092126064,0.2473773211,0.1353030801,0.3523185849,-0.1346607953,0.1831392497,-0.4423944056,0.495952785,-0.277638495,-0.3152654469,-0.2573377788,-0.107219696,-0.0311579686,0.2137938589,0.0032327722,0.1233839318,-0.0819337443,0.2954353988,-0.4231437743,-0.3038302362,0.5512321591,-0.4347351193,-0.2962920964,0.167625621,0.2419217825,-0.1240828708,-0.1673004329,-0.5922073722,0.0424591042,0.3712547123,-0.1085435748,-0.2428871244,0.2297098637,-0.1221716404,0.1525463164,0.0340049341,0.3174511194,0.0109068071,-0.2272163779,0.0310515221,-0.3468204141]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/120","title":"\ud83d\udc1b `map` not working","comments":"I didn't assign the output \ud83e\udd26\u200d\u2642\ufe0f\r\n\r\n```python\r\ndataset.map(test)\r\n```\r\n\r\nshould be :\r\n\r\n```python\r\ndataset = dataset.map(test)\r\n```","body":"I'm trying to run a basic example (mapping function to add a prefix). \r\n[Here is the colab notebook I'm using.](https:\/\/colab.research.google.com\/drive\/1YH4JCAy0R1MMSc-k_Vlik_s1LEzP_t1h?usp=sharing)\r\n\r\n```python\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('squad', split='validation[:10%]')\r\n\r\ndef test(sample):\r\n sample['title'] = \"test prefix @@@ \" + sample[\"title\"]\r\n return sample\r\n\r\nprint(dataset[0]['title'])\r\ndataset.map(test)\r\nprint(dataset[0]['title'])\r\n```\r\nOutput :\r\n> Super_Bowl_50\r\nSuper_Bowl_50\r\n\r\nExpected output :\r\n> Super_Bowl_50\r\ntest prefix @@@ Super_Bowl_50","comment_length":17,"text":"\ud83d\udc1b `map` not working\nI'm trying to run a basic example (mapping function to add a prefix). \r\n[Here is the colab notebook I'm using.](https:\/\/colab.research.google.com\/drive\/1YH4JCAy0R1MMSc-k_Vlik_s1LEzP_t1h?usp=sharing)\r\n\r\n```python\r\nimport nlp\r\n\r\ndataset = nlp.load_dataset('squad', split='validation[:10%]')\r\n\r\ndef test(sample):\r\n sample['title'] = \"test prefix @@@ \" + sample[\"title\"]\r\n return sample\r\n\r\nprint(dataset[0]['title'])\r\ndataset.map(test)\r\nprint(dataset[0]['title'])\r\n```\r\nOutput :\r\n> Super_Bowl_50\r\nSuper_Bowl_50\r\n\r\nExpected output :\r\n> Super_Bowl_50\r\ntest prefix @@@ Super_Bowl_50\nI didn't assign the output \ud83e\udd26\u200d\u2642\ufe0f\r\n\r\n```python\r\ndataset.map(test)\r\n```\r\n\r\nshould be :\r\n\r\n```python\r\ndataset = dataset.map(test)\r\n```","embeddings":[-0.1308908612,-0.288844794,-0.0395002961,-0.0892400667,0.1618371606,0.0317668915,0.0189053081,0.2528122365,0.372751385,0.1373792738,0.2324204594,0.6317356229,-0.0919273198,0.1015725359,0.1246013343,0.0128668025,0.1166422889,0.2223274559,-0.0682751015,-0.0151178883,-0.455271095,0.1827889383,-0.3452918828,0.3195168376,-0.1788795739,-0.0197323374,0.0069232848,-0.1686056405,-0.150720045,-0.1539473236,0.0808377117,-0.1376018077,-0.1134997159,0.3354356885,-0.0001208801,-0.1301519424,0.3381017148,-0.0655615479,-0.2692093551,-0.7517463565,-0.3737787306,0.1315031052,0.1735256761,-0.1138142198,-0.1845378578,0.2965347171,0.4201627076,-0.2645874619,0.1648977846,0.3474882543,0.1447360814,-0.1399454474,0.0086326376,-0.0474374108,0.0732381865,0.2379055172,-0.1838501692,0.2487777621,0.3222712278,-0.3623713851,0.392286092,0.118521519,-0.0787296072,-0.1327400059,0.1144522801,0.2852788866,0.2475356609,-0.2126544565,0.1879269779,-0.0303253382,-0.0482015125,-0.2822025716,0.033467032,0.05807136,-0.1231693774,-0.5438141823,0.0971134081,-0.0498420335,-0.0030328662,-0.062092647,-0.204874143,-0.0853242427,0.2212250531,0.6745916009,0.1022634059,0.9548653364,-0.116338104,0.2279542983,0.3377545476,-0.099991098,0.0041731172,-0.0862776488,0.067313768,0.376660198,0.069268167,0.1327460706,0.1648412198,-0.0948004797,-0.2642460465,-0.6372027397,0.1999678314,0.010157818,-0.1684218198,0.3996458948,0.020142179,0.2257401496,0.4909593761,0.3822185695,0.1175036356,0.0059167417,-0.2318648845,-0.1034115329,-0.0083997566,0.0281142052,-0.3582536876,0.1919237375,0.1363413334,0.1164997518,0.0773293898,0.0594358034,-0.4156877398,-0.0138432207,0.0812271386,0.32006827,-0.0052135736,-0.3745110929,-0.0273591131,0.1046239659,0.070869796,0.1158993021,-0.1659529507,0.1215727329,-0.6696619987,-0.1169036627,-0.0992712975,0.145935297,0.3452649713,-0.1467160434,0.0364236012,-0.0586835556,0.0628519133,-0.2476706505,0.4761578143,0.0555966832,-0.1343472749,0.4023709893,0.3105251193,-0.5544050932,-0.0962489694,0.4196369052,-0.1514539719,-0.0798978359,-0.1413527727,0.0859790742,-0.1877979487,0.0859404057,-0.3463787138,0.3524733782,0.3490440249,-0.3453374207,0.119552128,-0.1349669546,-0.6410399079,-0.195941776,0.1157927066,0.0416710228,-0.3937189579,-0.0446945652,-0.2229492366,0.4102565944,0.0790214166,0.1991873682,-0.1474320889,0.4485391974,-0.1345473379,0.4181539118,0.3669397533,-0.3884114027,-0.3279597461,-0.0438183732,-0.2272126973,-0.1566497236,-0.3628614247,0.0593842678,0.0771250501,0.0182608292,0.3440069854,0.0222455766,0.0512007624,0.0110291466,-0.0668046698,-0.1032929271,0.2660877705,-0.094947353,-0.0731077194,0.1685606688,-0.0119153848,-0.0485044308,0.2499534935,-0.1668195575,0.0707987323,-0.1735836267,0.2821959853,0.0722110048,-0.174994126,-0.0447287112,-0.1231511161,0.0357340761,-0.1855429709,-0.0179974753,-0.0927846208,-0.0023257572,-0.157992214,0.0448443182,-0.164195314,-0.0482518822,0.0877189264,0.0776939988,0.0191209968,0.0309899244,-0.0614586249,0.109588325,0.1743949205,0.1306148171,-0.5375675559,-0.1109388694,0.0721308738,-0.312258482,-0.3377268016,0.262784332,0.2720017433,-0.0913677216,-0.1593883634,0.1826328486,0.0240379274,-0.1603485346,-0.0225018281,-0.2081052512,0.1836200356,-0.1446668357,-0.0301504731,0.5182896256,0.2055364549,-0.2129921168,-0.0350321531,0.1274992675,0.1038670465,0.2728381455,-0.0886465237,0.0565799288,-0.0319980644,-0.0433970131,-0.0903621167,-0.1823167056,-0.0143293645,0.0340844467,-0.0532754324,0.0036533964,0.0778258592,-0.1119713336,0.4006410837,0.0231645592,0.3708415031,0.0209428277,-0.1153282151,0.1659086645,0.1863711178,-0.0547924601,0.5865674615,0.0131229572,-0.0254018661,0.1955808252,0.0845860764,-0.1995898783,0.1008771956,-0.1404303163,0.066596806,-0.2407288998,0.128489241,-0.1666507721,-0.0924958214,0.056909997,0.0684152171,0.2624983788,-0.2354493439,-0.0236500502,-0.1048641726,-0.2389103025,0.1042293683,-0.4910825789,-0.1033510417,-0.5496193171,-0.061219532,0.2564050555,0.0575671531,0.2079385668,0.2172444165,-0.3307384849,-0.0395485573,-0.4727929831,0.0556713194,-0.3682712615,-0.2575454116,0.0285215862,-0.2591771185,-0.1270316094,-0.0272990596,0.1257539988,-0.076078482,0.0362054817,-0.591994524,0.1960734278,-0.1373319924,0.0644340292,0.1637121886,-0.229882285,-0.5939878225,-0.4168896675,0.3337340355,-0.0796898231,-0.0332238451,-0.1684315205,0.0010464024,-0.3191904426,-0.310695529,-0.052027043,-0.1261584014,-0.3103807867,0.4487834573,-0.1254446954,0.2938332558,0.342856735,-0.1564956754,0.3314344585,-0.1706441194,-0.3483972251,-0.2987544835,-0.0064586718,0.3493944705,-0.3635314107,-0.148914963,0.0703558922,-0.1867864132,0.18225573,-0.0701196045,-0.1043677777,-0.2003175616,-0.0184744895,-0.0312952399,0.3904739618,0.2745307386,0.0836583078,-0.0805731341,-0.0049564466,-0.1555367857,-0.5929970741,0.278050065,0.6513228416,0.5089289546,0.1921819746,0.2489903569,-0.0075511946,0.5325362682,0.161803931,-0.2125650644,0.2115820795,-0.3211400807,-0.0320385434,-0.1424285471,-0.4234916568,0.0245851129,-0.0067756288,-0.1400478482,0.2757472098,0.0166609921,-0.2540222108,-0.0493890457,0.2344820797,-0.4783283472,-0.21114254,0.0979933888,-0.1960964352,0.0742845163,0.1765285432,-0.002119283,-0.2126140296,-0.2381561697,-0.0159716811,-0.2040700316,0.093843922,-0.0728586465,-0.0798611045,-0.5401269794,-0.4879357815,0.4711795747,0.0462231115,0.2800812721,-0.4085733891,-0.0222517047,-0.1019747257,0.3389689326,0.7204770446,-0.4146663249,-0.1517051607,-0.1285315305,0.0266340896,-0.3732301593,0.0504202843,-0.1387645453,0.7048385739,0.2851270735,0.4269518554,-0.0718691424,0.1912510693,0.3845406473,-0.3727276325,0.0515594035,0.0104761086,-0.4230521619,-0.340342015,-0.136509791,0.0031712106,0.190131411,-0.0638536811,0.4496664703,-0.1266025007,0.1970550269,-0.0817779601,0.363324672,0.2228880525,0.2051602751,0.4692052901,-0.219152689,0.1418530792,-0.1080152914,-0.4818500578,0.1222387925,-0.0705552176,-0.1371371895,0.0297262166,-0.1465840489,0.7096356153,0.3122912645,-0.1099983454,0.061601229,-0.1118347868,0.0894456431,0.0629385039,0.1518034637,0.3356262147,0.2160641551,-0.2308477759,-0.3069339395,0.3739404082,0.2010587454,-0.2294181734,0.4848844409,-0.2888379693,-0.3693415821,0.4401736259,-0.0552504994,0.8812676668,-0.3288820088,-0.0781108737,0.0262057409,-0.0904381275,0.3551356196,0.3642558455,0.2386020571,-0.2705540359,-0.2682383955,-0.0676918179,-0.0946341529,0.111542955,0.2740836442,-0.5248658061,0.3611488938,0.0136299655,0.3058551252,0.1772102714,-0.1273939908,0.3114953637,-0.1406894475,-0.3770224154,0.0788080916,-0.0949940756,0.2723190188,-0.1273154914,-0.0706704929,-0.4240265787,-0.2328914404,-0.0435348377,0.2279338837,-0.7580811977,-0.0378287137,0.2483018339,-0.0059414692,0.133612752,0.4042817056,0.2174759209,0.3187231123,-0.0584830008,-0.024190031,0.2432533801,0.1190760285,-0.2066161186,0.0010836093,-0.2842412889,0.0681241453,-0.222366944,0.1566010416,0.1069237664,-0.1556615084,-0.1290683597,0.2709940374,0.1125932634,-0.1063869745,-0.4957402349,0.115818806,-0.2005815655,0.0929071307,0.0137986848,-0.0961244479,-0.2572940588,0.1917704493,0.0472915657,-0.4564117193,0.1453085691,0.3444761932,-0.0730798393,0.097623311,0.3154871762,0.0668098181,0.0293462519,-0.0775230229,-0.2012252212,0.0359979756,0.2118708491,0.1406421363,0.0459912159,-0.224490568,0.0806289092,0.3399581909,-0.0683124363,0.0892965719,-0.223807916,-0.2267852575,-0.2857347727,0.2419134378,0.1176860407,0.0042890166,0.3015356362,0.1707770526,0.435931623,0.112964049,-0.2406932861,0.0985802487,0.0716508776,0.384924829,0.2092826962,0.0064116158,0.2786640525,-0.0815784633,0.0873332545,0.1576545686,-0.3437969387,-0.2149541527,-0.1669623107,0.1524948776,0.3685129881,0.1399248689,-0.3078056872,-0.1329809278,0.014440353,-0.3739818633,0.0259651318,0.1786461771,-0.1250319779,0.3014205992,-0.0177775249,0.0482214764,0.1343896836,-0.0410700962,-0.1205487326,0.1740269065,0.2187933624,0.1700946093,0.006505074,-0.0358884819,0.1499822736,-0.1678553224,-0.0587953068,0.0520397872,0.2513245344,-0.2683354914,0.0198334288,-0.1027357504,0.2284679413,-0.150015071,-0.086946547,-0.115996033,-0.0329861492,0.0067000752,-0.2695997059,0.1006250158,0.5648444891,0.20892784,0.3598713875,0.3155955672,0.1374083012,0.0718387067,-0.2548124194,-0.1165946722,0.1607795954,-0.1605867296,0.1015514284,0.5504282713,-0.1623869091,0.3152948618,0.190196529,-0.0635565147,-0.0931730196,0.3896214366,0.3885452449,0.4713999331,0.270885855,0.0840103999,0.3230815232,-0.0734597892,0.127055794,0.2542310655,-0.0447488315,0.4074877799,0.2904060781,0.3753799796,-0.2979837358,0.0869580433,0.249066934,-0.0140675725,-0.1390672177,-0.1929689944,-0.1051259413,0.186953336,-0.3700275421,-0.1924062073,-0.0126225706,-0.1224955916,0.6040125489,0.0263606273,0.1525217593,0.1616507918,-0.1923268884,0.1634815037,0.1731297374,-0.1715735048,0.252620399,-0.4696709216,0.1061891243,-0.210182637,0.3927227557,0.1896379292,0.4295219481,-0.4504497349,0.4053703249,-0.1087130606,-0.2128284723,0.0145011218,0.2407447696,0.6244417429,0.1723867059,0.0321058817,0.1274470389,-0.0888697729,0.2410169989,0.2507748902,0.1872519702,0.3839383721,0.20809111,0.0156714171,-0.3604923487,-0.2400274277,-0.0676507875,-0.4239911437,-0.001406859,0.4004197121,-0.2318472862,0.2212595344,-0.146082893,0.0396972112,0.2456219196,0.2564711869,0.2404112816,0.2309992611,-0.45482862,-0.203508988,-0.4605851769,0.0590339638,-0.1616377831,0.1461365521,-0.0401878059,-0.1157577485,0.1455264986,0.2692285478,0.007954658,-0.3788704872,-0.151646778,0.1226706207,0.2159106433,-0.1012304798,-0.1711682528,0.134629339,0.1686708182,-0.2212593257,-0.0315539762,-0.2509437501,-0.0465140268,-0.3802003562,-0.1483202577,-0.1365135461,0.0572493449,0.5121200085,0.2238942981,0.2454637736,-0.2290041447,0.1886906922,0.1829675734,0.032173302,-0.037736997,0.2693751752,0.1605891734,0.2671964169,-0.0952196494,0.01210225,-0.1513287276,-0.0638699085,0.2292641103,-0.2815912068,-0.3133399487,0.0606127828,-0.1826198846,-0.0654367208,-0.0771833062,0.4257065952,-0.2217243165,0.2012743205,0.0630278513,-0.2539878488,0.5734155774,-0.4769709408,-0.472315371,0.0201122481,0.1540259421,-0.0972416401,-0.3128697872,-0.4151430726,-0.2179453224,0.1861940175,-0.1381082982,-0.1734426171,0.0823902562,-0.2065389454,0.0544588529,-0.1094942614,0.3343848884,0.1170885861,-0.1189925,0.0300853085,-0.3970053196]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/119","title":"\ud83d\udc1b Colab : type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'","comments":"It's strange, after installing `nlp` on Colab, the `pyarrow` version seems fine from `pip` but not from python :\r\n\r\n```python\r\nimport pyarrow\r\n\r\n!pip show pyarrow\r\nprint(\"version = {}\".format(pyarrow.__version__))\r\n```\r\n\r\n> Name: pyarrow\r\nVersion: 0.17.0\r\nSummary: Python library for Apache Arrow\r\nHome-page: https:\/\/arrow.apache.org\/\r\nAuthor: None\r\nAuthor-email: None\r\nLicense: Apache License, Version 2.0\r\nLocation: \/usr\/local\/lib\/python3.6\/dist-packages\r\nRequires: numpy\r\nRequired-by: nlp, feather-format\r\n> \r\n> version = 0.14.1","body":"I'm trying to load CNN\/DM dataset on Colab.\r\n\r\n[Colab notebook](https:\/\/colab.research.google.com\/drive\/11Mf7iNhIyt6GpgA1dBEtg3cyMHmMhtZS?usp=sharing)\r\n\r\nBut I meet this error :\r\n\r\n> AttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\r\n","comment_length":63,"text":"\ud83d\udc1b Colab : type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\nI'm trying to load CNN\/DM dataset on Colab.\r\n\r\n[Colab notebook](https:\/\/colab.research.google.com\/drive\/11Mf7iNhIyt6GpgA1dBEtg3cyMHmMhtZS?usp=sharing)\r\n\r\nBut I meet this error :\r\n\r\n> AttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\r\n\nIt's strange, after installing `nlp` on Colab, the `pyarrow` version seems fine from `pip` but not from python :\r\n\r\n```python\r\nimport pyarrow\r\n\r\n!pip show pyarrow\r\nprint(\"version = {}\".format(pyarrow.__version__))\r\n```\r\n\r\n> Name: pyarrow\r\nVersion: 0.17.0\r\nSummary: Python library for Apache Arrow\r\nHome-page: https:\/\/arrow.apache.org\/\r\nAuthor: None\r\nAuthor-email: None\r\nLicense: Apache License, Version 2.0\r\nLocation: \/usr\/local\/lib\/python3.6\/dist-packages\r\nRequires: numpy\r\nRequired-by: nlp, feather-format\r\n> \r\n> version = 0.14.1","embeddings":[-0.2273421288,-0.023100283,-0.0340336785,0.1723956466,0.0568722859,-0.0996569172,0.2892366946,0.1857061237,-0.0378169827,0.0181524269,0.0521476455,0.5537141562,-0.2503296733,0.0401885919,0.0903182849,-0.0751916692,0.0834185407,0.258730948,0.1037930027,-0.1277628094,-0.1502255648,0.1527535319,-0.3422175944,0.1068625599,-0.1202139482,-0.0682862923,0.1318603307,-0.0488690846,-0.0279800203,-0.4014903009,0.4233409464,-0.2384961843,0.1887197942,0.0650445744,-0.0001270101,0.0349720083,0.3435460329,-0.1084538177,-0.355566293,-0.3720389307,-0.054751683,-0.3115216196,0.6195546389,-0.1986419708,0.088460274,-0.2497655153,0.050522767,0.1458174437,0.1289491057,0.2576676607,0.0944548026,0.2429156899,0.2680660784,0.0369408764,0.3030516207,0.1586951762,-0.3170281649,0.8054688573,0.1521752775,-0.0587130897,0.382936269,0.1836001426,-0.292855233,0.3461164534,0.2409609109,0.0738360882,0.1236388534,0.0077660764,-0.024109859,0.0774950683,0.2109683603,-0.1711868197,-0.1291289479,0.1297546625,0.2963559628,-0.0789032876,-0.0245266836,0.3779229522,-0.1516014338,0.2225867808,-0.0472658351,-0.0163547583,-0.1646301001,0.2858141661,-0.0288101323,0.4169096351,-0.1104350016,0.2412788719,0.0684108511,0.1585405916,0.6358198524,0.1410775185,0.0020679836,0.0277524069,0.2584057152,0.2047149241,-0.1335516125,0.0248839911,-0.1483647078,-0.0341440775,0.4241458774,-0.0339717269,0.1124347597,0.3809981048,-0.1169970557,0.2114268839,-0.1434601694,-0.0326790698,0.1138976887,0.1078684628,-0.2203160524,-0.0274381042,-0.1393719316,0.0647004098,0.1614550799,-0.2065222114,0.431833595,-0.1551369578,-0.2683421671,-0.2181916088,-0.6080505252,-0.0528738946,-0.0578332469,0.3415149748,-0.2353737801,0.1496099532,-0.0124659156,0.1326818317,-0.2515012324,-0.0333122015,-0.0183452722,0.1814994514,-0.2161078304,-0.1013979167,0.1927607954,0.143014133,0.2412786335,-0.2968501151,-0.015250206,0.0636128634,-0.0285659172,-0.2196339965,0.3401510417,0.4856801331,0.2516511381,0.06944561,0.1663389951,-0.3547950089,-0.270447284,0.4337829351,-0.4250109196,-0.3489814997,-0.3356326222,0.01754486,-0.2543338239,-0.073577866,-0.6186476946,-0.1014759764,-0.0699049011,-0.3670096397,-0.078219384,-0.2841028869,0.0642401576,-0.3631058335,0.0506600216,-0.071565561,-0.4645743072,-0.1888604909,-0.0797881633,0.1347500682,0.3948203921,0.1182053462,-0.1718877703,0.0022878549,0.1080621406,-0.0528102703,0.5477809906,-0.1122525558,-0.4989821017,0.0723102242,0.1771580875,-0.1638492942,0.1235975251,0.1389488131,-0.0887546241,0.078792803,0.1198111773,0.2991141379,0.2241226733,-0.031194346,-0.3039538562,-0.2280001044,0.1243885756,0.1765549928,0.0341269746,0.0752119794,0.1721445918,-0.1110548824,0.0998981372,-0.1193781942,-0.1763902605,-0.0833880082,0.4347991049,-0.2393099964,-0.1224682555,-0.0388059244,-0.1566177905,0.0591565333,-0.0501139611,0.032014966,-0.3557682931,0.0814519748,-0.1497117132,0.2907872796,-0.083903797,-0.1440961957,-0.0464455634,0.1238604635,-0.0979486555,0.1587922126,-0.1432317346,0.0955498591,0.0379814878,0.0007210464,0.040151082,0.4449893236,-0.2541424632,-0.3926926851,-0.1927075535,0.226912573,0.0311912671,0.0506366491,-0.0731994063,0.0104083186,-0.1712086052,-0.236125201,-0.2261316031,-0.0824178606,0.1172009781,-0.4324964881,0.2024841905,0.0100899572,0.240073055,0.0297513362,0.4912329614,0.3507786691,0.0764851272,0.2080870569,0.1864064187,0.1246459708,0.0424841605,0.2603189647,0.2518983185,-0.0253212322,0.0110155623,0.115363799,0.3049418926,-0.0108530624,-0.3738047183,0.0419913903,0.3567649424,0.052017875,0.3117726445,-0.0289514475,-0.4617649019,-0.0467014909,0.3583218455,-0.0207948927,0.4363690913,0.1499262005,-0.0362185761,-0.1294074506,-0.1859571636,-0.1250617057,0.2619031072,0.091383338,0.430811733,-0.1021552235,0.2690778375,0.0374647975,-0.2176920027,-0.4670942426,-0.0732046142,0.2502478659,-0.181112811,0.3723154068,-0.2413871139,-0.3120823205,-0.2413733453,-0.6629678011,-0.0929108709,-0.2774499357,-0.1045342311,0.351057142,0.3144078553,0.1537207067,0.0195632689,-0.1171601936,0.0267101247,-0.2134560943,0.0538363643,-0.2400348037,-0.5079825521,-0.0529078059,0.2017836422,0.2395821214,-0.0820834786,-0.0431232676,0.0429058,0.0961140171,-0.2651324272,0.1731713414,0.0140371248,0.3970383406,0.0427309796,0.0957200006,-0.564878583,-0.4018689692,0.2459059954,-0.2879649699,-0.1525113881,0.1136663482,-0.1439387649,-0.2021039128,-0.1280095577,-0.0126290526,-0.5180642009,-0.2005015761,0.2381270528,0.2255993783,0.2146229595,0.4167945385,0.367084384,0.2150843143,0.3779624403,0.2771554887,0.0209961869,0.0206607785,0.3824847341,-0.1830014288,-0.4004779458,0.081339635,-0.2128242254,0.198611632,0.1585609168,-0.3589545488,-0.4112590551,-0.0249733888,-0.0041604419,-0.0937071666,0.0064305365,0.0887785032,0.1368145347,0.126965493,-0.0449938364,-0.1077996641,-0.2498589158,-0.0095360894,0.035297215,-0.0117583796,-0.0443688855,-0.151064977,0.7354260683,-0.3515334427,-0.1901317388,0.4313213825,-0.1782525629,0.2434347123,-0.1233844459,-0.428332746,-0.0222853068,-0.0430831946,-0.0439529717,0.3376128972,-0.0383584797,-0.5813614726,-0.2494640946,0.0765072405,-0.3812740743,-0.1667638719,0.1639233083,0.2693271339,0.0802882463,0.0635868013,0.0131593188,-0.1439971775,-0.2827734649,-0.1369134784,0.0401388891,-0.2952933908,-0.0084800199,0.2029032558,-0.7626090646,-0.2404322922,0.4146600068,0.2852545679,0.2476827651,0.0957996249,-0.2617818713,0.2249414325,0.0091588125,0.1963609308,0.1493815929,0.0703866854,0.1387833953,0.1577401608,-0.3971831501,-0.3087859154,-0.1306043267,0.1879579425,0.0510905944,-0.1843996793,0.0000043631,0.0832094848,0.3705067337,0.1232452169,0.0100015271,0.2355479002,-0.4408878684,-0.4533112943,-0.1411542296,-0.3123024106,-0.1471474618,0.2033472657,-0.0009640277,0.1586270183,-0.4254463613,-0.1607260257,0.0539994612,0.3005611598,0.2176273912,0.0466334857,-0.2314663827,-0.387650609,0.2860001922,0.350459516,0.4171788692,0.1805090904,-0.1292809099,0.0856091082,0.1106171086,0.4416895509,0.224926129,-0.0695738643,-0.1265290082,-0.0970632806,-0.0114017539,0.4370592237,0.0542652011,0.0889404416,-0.1472579092,-0.2033494264,-0.123573333,0.6838407516,0.2170898914,-0.0134178204,-0.1140771285,-0.0493055247,-0.1263804883,0.5223632455,0.090720512,1.1277776957,-0.0168337375,0.2786756754,0.5416336656,0.0329480954,0.1964418292,0.0991426036,0.1452222615,-0.3288172185,-0.3356903493,-0.196082145,-0.1755356491,0.173887223,-0.0160523709,-0.1755716056,0.2271187603,0.1601848751,0.1184690297,0.2924094796,-0.1929136217,-0.031146761,0.0040136236,-0.7306958437,-0.0800056532,-0.2024460137,-0.020557737,-0.1546234339,-0.2079721838,-0.2679944634,-0.1427392215,-0.1489298195,0.1014357954,-0.4181284308,0.2435303032,0.3503923416,-0.4392693937,-0.1902593225,0.0290632024,-0.1315020025,-0.1416925788,-0.0977706537,0.0657374635,0.2824130654,-0.157059297,-0.0103112916,0.0394270942,0.5105876923,-0.1625644714,-0.091076754,0.1391649842,-0.0809079185,-0.2074200511,0.2585479319,0.3215390444,0.3235269189,-0.079331249,-0.3273769319,-0.0956643969,0.0364747345,0.0193114188,-0.0172787905,0.0058736191,-0.1353987157,0.2033499032,0.41230914,-0.2906852067,0.0073793088,0.2270115912,0.1058614925,0.4746120572,0.3311219811,-0.0327561349,-0.158861205,-0.0273490585,-0.1565034539,-0.1622785777,-0.6085425615,0.1317338198,0.1362909526,-0.0720681846,0.1540685743,0.6013589501,0.2857756019,0.3519016802,-0.2581928968,-0.5689340234,-0.0419965014,0.4747011065,0.0831502825,-0.0886518806,0.1389125586,0.332262218,0.2429087907,-0.2398879826,-0.1197650135,0.0642960817,-0.0183652714,0.1527070552,-0.4946833849,-0.0499270186,0.1966678202,0.0385345705,-0.0119276997,-0.0322385691,-0.1452204734,-0.095877476,0.0005260142,0.2258659303,0.1175636575,-0.1088495702,-0.2788051069,-0.1316891164,-0.023935128,-0.1247390062,-0.1053410769,-0.0863497779,-0.1400244981,0.0613604672,0.3017079234,0.2998593748,-0.24013035,0.3917749524,-0.2197278887,0.4072280824,0.166893363,0.0009881462,0.0527887642,0.2019532323,0.0988019928,-0.1682947576,-0.4009670317,-0.0271095634,0.421603471,-0.4594939053,0.0482727475,-0.339794457,0.2195504159,0.0650723502,0.0874450803,-0.2689504325,0.2230324149,-0.028731538,-0.4750271142,0.1333149374,0.361492008,0.208487764,-0.0552632436,0.516254127,0.0567856953,-0.0576882139,-0.2037632167,0.1861198097,-0.0415574946,0.1579738706,0.012569651,0.1212744638,0.0556830019,0.3084196746,0.0515967607,0.1649468094,0.3111001551,0.1194946393,-0.2167848796,0.2646975219,0.1837601811,-0.1751306206,0.2735774219,-0.1674000025,0.0444005504,0.2159041613,0.3815360963,0.1580124199,0.2519905865,0.609924078,-0.1743126214,-0.0356752016,-0.0002921621,0.2645956278,-0.0474145971,0.0049241642,0.0395850837,0.1466959715,-0.1381296664,0.026298834,-0.1146885604,-0.1200741231,0.3901174664,0.4043715,-0.2081314772,-0.2560096681,0.0645320788,0.362274766,0.0507559516,-0.0023758998,0.1668530107,-0.0628212318,-0.0786540061,-0.0965215042,0.3266403377,0.4390550852,0.2800118029,-0.3530528843,-0.2642388642,0.0318567716,-0.2656352222,-0.108605966,-0.0279251281,0.5955819488,0.4304202497,0.0805637315,0.0338481329,0.0643038899,0.5642895699,0.082594201,-0.0700400546,-0.0192197654,0.3611959219,-0.1765223593,-0.3979310989,-0.2160397023,0.0124515127,-0.3533458114,0.1628708541,-0.0022757822,-0.0497106388,-0.0197402481,0.003781792,-0.062652126,0.0078473929,0.4651158452,0.5102041364,0.5972729325,-0.0888904408,-0.0550756976,-0.3583928943,0.2906204164,-0.1737446338,-0.1128890887,-0.0008863649,0.0702733472,0.020050507,0.3194665611,-0.0278414693,0.2376878113,0.0380444154,0.0470623299,-0.4800111353,-0.3692433834,-0.0701784566,-0.1957493871,-0.1922422349,-0.2766743004,-0.0515329614,-0.6590490937,-0.1110530868,-0.18732813,-0.4949440062,0.082383804,0.5267144442,0.4822739065,-0.0843054652,0.6200261712,0.0461450741,0.0247717984,0.1293837279,-0.3383233249,0.0459184386,0.4102053344,0.1469039768,0.330563128,-0.0211198796,0.4421643615,-0.0969297737,0.4875105023,-0.0563006327,0.0907277986,-0.1164334863,0.0269789286,-0.3615356684,-0.0369251333,-0.2916876972,-0.0247042086,-0.0811568722,0.0831420347,-0.3877400756,-0.368583858,0.5129152536,-0.5418502688,-0.2892979085,0.0782183856,0.1129932925,0.0209979787,-0.0002982196,-0.5997846723,-0.1234206706,0.2359479964,-0.0325088724,-0.1359734237,0.1711479723,-0.0848128051,0.0840384662,-0.2127688676,0.6645045877,-0.2815516293,-0.2294520587,-0.4506763518,-0.2799414992]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/119","title":"\ud83d\udc1b Colab : type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'","comments":"Ok I just had to restart the runtime after installing `nlp`. After restarting, the version of `pyarrow` is fine.","body":"I'm trying to load CNN\/DM dataset on Colab.\r\n\r\n[Colab notebook](https:\/\/colab.research.google.com\/drive\/11Mf7iNhIyt6GpgA1dBEtg3cyMHmMhtZS?usp=sharing)\r\n\r\nBut I meet this error :\r\n\r\n> AttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\r\n","comment_length":19,"text":"\ud83d\udc1b Colab : type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\nI'm trying to load CNN\/DM dataset on Colab.\r\n\r\n[Colab notebook](https:\/\/colab.research.google.com\/drive\/11Mf7iNhIyt6GpgA1dBEtg3cyMHmMhtZS?usp=sharing)\r\n\r\nBut I meet this error :\r\n\r\n> AttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'\r\n\nOk I just had to restart the runtime after installing `nlp`. After restarting, the version of `pyarrow` is fine.","embeddings":[-0.2835429907,0.0447813608,0.0172028225,0.2262002379,0.1537768245,-0.0492425673,0.2736166418,0.1078590304,-0.0629185885,0.0287807398,0.089954637,0.4489680529,-0.3376855552,0.2220071256,0.2667759061,-0.1032789871,0.0401548296,0.3651624024,0.1162208095,-0.0348985642,-0.1869616807,0.1577655524,-0.4699403346,0.093881458,-0.2458846867,-0.1728018373,0.0579281226,0.0365408249,-0.0127620483,-0.4258046448,0.4096870124,-0.2087133527,0.2067955136,0.049235519,-0.0001234279,-0.0009604987,0.3468475044,-0.1835941374,-0.380650878,-0.2610020041,-0.0870168731,-0.0719363689,0.5392192006,-0.1955817789,0.0109692076,-0.2137584835,0.0638213009,0.1740896404,0.0559797697,0.2633334696,0.1195242032,0.139874205,0.2437542379,0.0178375617,0.1536610126,0.1872360259,-0.2336520851,0.7389292717,0.162230134,-0.1054732352,0.36913082,0.0969696939,-0.2743653059,0.2058781385,0.3785057664,0.06716723,0.0230651107,-0.0656105727,-0.0372657552,0.1121522561,0.3022116423,-0.2212343365,-0.002140336,0.1088491306,0.4271304309,-0.2622922361,0.0225428641,0.3496264815,-0.1930697113,0.1046563014,-0.1546211243,-0.1691000164,-0.1028717309,0.2732112408,0.0031538589,0.2862986326,-0.0820575356,0.1322266757,0.2594425678,0.1244037822,0.5737629533,0.1832592785,-0.0675802603,0.0166732483,0.1255553961,0.1166650206,-0.1151927635,0.0048528053,-0.1534453928,-0.2217015773,0.3211809099,-0.1065697223,0.1329921782,0.3500325382,-0.0483986437,0.3419393897,-0.0776829645,-0.0088762799,0.0688883886,0.0841941833,-0.3237350583,0.0261882674,-0.2106641978,0.0583889484,0.155764237,-0.1791136414,0.3839304447,-0.1505262107,-0.1758469343,-0.2368793041,-0.5335425138,-0.0388216041,0.1062474921,0.4182965457,-0.1107630804,0.0536781214,0.0200206507,0.0310178827,-0.1992997229,-0.1668626815,-0.1107777283,0.312034905,-0.349057585,-0.0040702019,0.2279451638,0.1600241065,0.1629523784,-0.2321518511,0.0262949038,0.0044045448,-0.0706854612,-0.2387650311,0.230105266,0.5033140779,0.165932864,-0.0782710835,0.1211165339,-0.3699175417,-0.191652149,0.5548939109,-0.3915857971,-0.494168669,-0.3462758958,0.0253200568,-0.3040960729,-0.1140843108,-0.7277987003,0.0179966483,-0.0531854145,-0.3441113234,-0.15384835,-0.3110789955,0.0408108942,-0.2952149808,0.135038808,0.081266351,-0.4829117954,-0.1567835361,-0.0012833568,0.0355808325,0.4002494812,0.0959426835,-0.2945821583,0.1581172794,0.0410972238,-0.0093365349,0.4867193401,-0.0099859219,-0.38403368,-0.0328746997,0.1367969364,-0.1818748564,-0.0236539785,0.131323576,0.0161779411,0.1076828092,0.0527130589,0.3254430592,0.1134636849,0.0453519225,-0.3021544516,-0.1602824181,0.2848570347,0.272996366,0.1056291163,0.2323065102,0.0395339988,0.0595537499,0.1080310643,-0.1597004682,-0.205755353,-0.1156278551,0.4424152374,-0.259305656,-0.0949249566,-0.1033822745,-0.200037688,0.0993315503,0.0219705366,0.1486597806,-0.2642741799,0.098718375,-0.0018053843,0.2425141782,-0.0544146299,-0.0932020172,-0.0472624265,0.1068520397,-0.1139084846,0.0855095237,-0.2942034304,0.1732182652,-0.0855095536,-0.059121605,0.1172953025,0.4961355627,-0.221056357,-0.4288746417,-0.1441909969,0.2088763863,0.0196503736,-0.0017652157,-0.1202741787,0.0807258189,-0.1485061496,-0.2879565656,-0.2266209424,-0.0763427243,0.0713797435,-0.3917328715,0.1217708066,-0.2625131309,0.1892559826,0.0111224782,0.6011775136,0.2312677354,0.0353085808,0.0962717757,0.1675921679,0.0979729146,0.1538674533,0.0913398787,0.4108107686,0.0572190322,0.0587004013,0.0357715562,0.1198372245,0.0239027534,-0.5300848484,0.0398728959,0.4848031402,0.0172754433,0.2846191525,0.0506549031,-0.3280873299,-0.0911349952,0.2563597858,0.0343006924,0.4824927151,0.1402957737,-0.2031363547,-0.0206070263,-0.165212214,-0.1110758409,0.260269165,0.0133426962,0.5372642875,-0.1529492736,0.1179368421,0.0928597376,-0.2979283035,-0.5027729273,0.0765952989,0.4357075393,-0.1862333268,0.249039039,-0.252643317,-0.4363840818,-0.1391378045,-0.4521755278,-0.0466637239,-0.174242124,-0.1480227411,0.2751824856,0.3541748226,0.1759913266,0.0556834601,0.1341069043,-0.036357563,-0.0861927941,0.1157094687,-0.2787898779,-0.4251944125,-0.0438833944,0.3219062686,0.2654452622,-0.1445865333,-0.0429668278,0.0609350614,0.2248010486,-0.2154298723,0.1809117198,0.0597249344,0.3478369415,0.0290864334,0.1726853997,-0.6037281156,-0.4763202965,0.1973819137,-0.2548491061,-0.1514302343,0.2011635602,-0.1185197905,-0.2979263067,-0.1392681599,-0.0399456732,-0.5196630955,-0.2402305007,0.1384040415,0.0515491664,0.2625725269,0.4227849841,0.2467351109,0.2516053319,0.2856173217,0.2782490253,0.0416850634,-0.0272825621,0.206046313,-0.2266697288,-0.3076445162,0.1467666179,-0.1775440127,0.1536619216,0.0756456777,-0.5399981737,-0.3905145228,-0.0119577246,-0.1464466751,-0.1086419597,-0.1064012274,0.1131096035,0.1331327409,0.1270314604,0.0461503081,-0.2463925928,-0.1610321999,0.1056986079,0.0864791498,-0.0364234038,-0.0304755699,-0.209296912,0.5693787336,-0.4026015699,-0.2776995301,0.4051408768,-0.1440754384,0.2460667193,-0.089598462,-0.5039294362,0.0884152278,0.0083330488,-0.0451650918,0.3507307768,-0.1640148908,-0.647126019,-0.2834515572,-0.020089224,-0.3335896134,-0.1855003983,0.2301662415,0.2210153937,0.1495528072,0.01602423,-0.1176654026,-0.1176837161,-0.240801543,-0.2220288515,0.0213847067,-0.1458092332,-0.0059299683,0.0796282291,-0.8123748302,-0.1659664065,0.2598355711,0.2342234105,0.3524184823,-0.0374385454,-0.0922249183,0.1741700321,0.019621795,0.3843098283,0.1001141965,0.0802244544,0.1886648983,0.0988742709,-0.1602258235,-0.2760578692,-0.0472392775,0.2767828703,0.0847614184,-0.0996068195,-0.0005061375,0.1258312017,0.4356416762,0.1801761389,-0.0695371255,0.0788680911,-0.4101521671,-0.4919433296,-0.1469111741,-0.1934156716,-0.1702412963,0.3370666504,0.0920142084,0.1263026744,-0.3354412615,-0.2364562601,0.0757105127,0.3256473541,0.2640870512,0.2453938723,-0.1104420424,-0.4127435386,0.1787739545,0.3454219103,0.4031330347,0.0899149999,-0.0048779706,0.1280058026,0.0996113122,0.5552995801,0.2658326626,-0.1084294543,-0.0626028329,-0.0699036419,0.0412332602,0.4241522849,0.086490266,0.0685795769,-0.0943540335,-0.2509576976,-0.1393634081,0.4788408875,0.0622922555,0.0435181372,-0.1303570271,0.0010191656,-0.1650122553,0.5379965901,0.1355877221,1.2302488089,-0.0798116997,0.113510564,0.427278161,0.0972904563,0.344057858,0.1468312293,0.2983436584,-0.3297371268,-0.5403360128,-0.1466910392,-0.2585095465,0.1084064245,-0.0777798146,-0.2165906727,0.0931953415,0.1336765587,0.0605889149,0.2407919317,-0.0724502131,0.1816533357,-0.086043261,-0.6862668395,-0.0634382814,-0.3170798123,-0.0801262408,-0.14438501,-0.2058357447,-0.0329403356,0.0021592076,-0.2719330788,0.1684134603,-0.3516063988,0.1474134177,0.3816626668,-0.2477458119,-0.0673009306,0.0983501673,-0.0222047735,-0.0337441862,-0.0974004194,0.0890250206,0.2403408438,-0.1929229796,-0.0155281918,0.0943442062,0.5773585439,-0.1882115752,-0.1423975825,0.1596543044,-0.0297807939,-0.2199529111,0.2198917568,0.1586842686,0.1554428339,-0.0714244768,-0.3738238811,0.0186094567,0.0511658825,-0.0071008625,0.0193584654,0.0567212366,-0.2727185786,0.2915140092,0.3602831066,-0.2769806087,0.011714112,0.1982056797,0.2166066766,0.5088031888,0.3107889891,0.1555577666,-0.1226476952,-0.0622748807,-0.2113235593,-0.1434185207,-0.5613009334,0.2085903734,0.2144378275,-0.0192412734,0.1904455423,0.6016796231,0.2383433431,0.4067668021,-0.1961831599,-0.5690779686,-0.0830445811,0.3884062767,0.0704281107,-0.0219318364,0.2153734863,0.469776839,0.2040328681,-0.041186206,-0.1725406647,0.0276707001,-0.2810370922,0.1568967998,-0.3452838659,-0.0871520564,0.1488282681,-0.0540606715,-0.0080369972,0.0230866298,-0.0838181153,-0.1111237183,0.0463638678,0.2178961784,0.1191333681,-0.020109361,-0.2711077929,-0.2300772518,-0.0756807551,-0.0553609543,-0.1189887971,-0.171586588,-0.2497189939,0.0661625192,0.2646791041,0.1477046609,-0.2316978127,0.3580917716,-0.1722728759,0.3114063442,0.1581256092,0.0033557692,0.0392439403,0.1404027343,0.0489908122,-0.1449489892,-0.4863519669,-0.0621226095,0.4573004246,-0.4543430209,0.0370992385,-0.2756279707,0.1962378472,-0.0416425318,0.050406903,-0.1343413889,0.332824558,-0.0076247058,-0.4974961579,0.0592225976,0.3732177615,0.2192761004,-0.1509863436,0.5697948337,0.0828205869,-0.0880134404,-0.2487197518,0.1719935685,0.1382824779,0.1886371225,0.0161365457,0.057177823,-0.1314997673,0.1863767356,0.082692109,0.2691390812,0.2436386049,0.2580545843,-0.2570932806,0.1916866452,0.0885266513,-0.1264073402,0.2733305693,-0.268920511,0.1977313459,0.3990250528,0.3675202727,0.0204290673,0.0263473894,0.5964528322,-0.184897393,-0.1633167565,0.0365588851,0.4003457427,-0.1985926479,0.2459598333,0.013610567,0.15648444,-0.145643428,0.0391396917,-0.0670530349,-0.1625826657,0.3168179095,0.3569125533,-0.2791519165,-0.1688157171,0.0278107375,0.3662383854,-0.003270173,-0.0418250449,0.2396450639,-0.1418609619,-0.1152706742,-0.0356536806,0.3584585786,0.3520517945,0.4302458465,-0.2706185877,-0.1951931417,0.1002783403,-0.1823779792,-0.0459233299,-0.0653905272,0.7172917724,0.5224655867,0.0196235664,0.0389848761,0.0533650666,0.5616181493,0.1299623102,0.0132284882,0.0963453874,0.2652920485,0.0220605303,-0.3668479323,-0.2870575488,0.0700067133,-0.3549190164,0.1026443169,0.0585613512,-0.1038258895,0.0053131306,0.0027310462,-0.035607893,0.0250225198,0.5348510146,0.4581922293,0.5921170115,-0.2136776,-0.1385885179,-0.4208382666,0.2827702165,-0.1084440649,-0.1752826124,0.1029847115,0.022684738,-0.0546757802,0.2285357863,-0.119230397,0.3515221775,-0.0360822156,-0.0729368851,-0.5151742697,-0.3620357215,-0.1695677787,-0.1291205436,-0.2445668131,-0.1161457822,0.0950784609,-0.7269615531,-0.0848274752,-0.1624622494,-0.4739104807,-0.0573199317,0.5800807476,0.3454466462,-0.1207529157,0.716689527,0.0790596157,0.1381374151,0.2532726228,-0.3978876472,-0.0601039231,0.2693285346,0.2259960622,0.3219339252,-0.0466300547,0.5462441444,-0.2087995857,0.3317684829,-0.0491798706,0.2054100931,-0.2002651095,0.0489458889,-0.2735572755,-0.0352357179,-0.0208943654,0.0293851662,-0.1030030325,-0.0169485584,-0.3844216764,-0.2835633159,0.6003670692,-0.4355747104,-0.3089817762,0.0686277375,0.0962733924,-0.0252474621,-0.045725815,-0.5969157815,-0.1464795917,0.1604412496,0.0224280078,-0.1952164322,0.2663180828,-0.1609741002,0.1292503476,-0.1807330996,0.4809121192,-0.2284750789,-0.2534156144,-0.4889084697,-0.2165667415]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/118","title":"\u2753 How to apply a map to all subsets ?","comments":"That's the way!","body":"I'm working with CNN\/DM dataset, where I have 3 subsets : `train`, `test`, `validation`.\r\n\r\nShould I apply my map function on the subsets one by one ?\r\n\r\n```python\r\nimport nlp\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nfor corpus in ['train', 'test', 'validation']:\r\n cnn_dm[corpus] = cnn_dm[corpus].map(my_func)\r\n```\r\n\r\nOr is there a better way to do this ?","comment_length":3,"text":"\u2753 How to apply a map to all subsets ?\nI'm working with CNN\/DM dataset, where I have 3 subsets : `train`, `test`, `validation`.\r\n\r\nShould I apply my map function on the subsets one by one ?\r\n\r\n```python\r\nimport nlp\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nfor corpus in ['train', 'test', 'validation']:\r\n cnn_dm[corpus] = cnn_dm[corpus].map(my_func)\r\n```\r\n\r\nOr is there a better way to do this ?\nThat's the way!","embeddings":[0.0147184702,-0.4408499002,-0.2191771567,-0.1899598688,-0.2989725471,0.1243559569,0.054131005,0.3102736175,0.1604536325,0.0873898864,0.0930391923,0.4565735161,-0.2604030073,-0.0573265664,0.2850784957,-0.100307703,0.1645949483,0.2718676627,-0.2464935631,-0.1829447448,-0.3428681493,-0.2058437914,-0.2778635621,-0.1447101086,-0.1044876426,-0.3943483531,-0.1481689364,-0.0933613554,0.2229522169,-0.2135799676,0.190821439,0.0199113879,0.0802730396,0.1382663846,-0.000112549,-0.0940667093,0.0706392154,-0.0981380865,0.0494530387,-0.2748911381,-0.311816901,-0.3726471364,0.121913664,-0.4168619514,0.0560577065,0.026560409,0.2041710764,-0.266651541,0.2426314056,-0.0798195153,0.1225233674,0.1188694164,-0.2016891241,0.0720776469,-0.133309111,0.1394457221,0.241972506,0.293800354,0.3626162708,-0.4156958163,0.2328511029,0.2570594847,-0.3154198229,0.27773422,0.4873460531,-0.1516582966,0.1046083942,-0.5774194002,-0.3162068427,0.1234372035,0.3150307238,-0.1677197963,-0.073306039,-0.0436789766,-0.2062227875,-0.3253844678,-0.2046141326,0.1137282178,0.0947153345,-0.2701574266,-0.716911912,-0.0448183455,0.0297885295,0.633916676,-0.3696345687,0.4150871933,-0.173082456,0.2482126355,0.2735759616,-0.0863138661,-0.1554481983,-0.0337360129,0.0060470253,0.4146464765,-0.0750099123,-0.3132121861,0.1297567487,0.0443959907,-0.0343082733,0.0285978522,-0.2126865387,0.0572950915,-0.1877037287,0.3147087991,0.4100811481,0.0552190505,0.385877043,0.6236424446,0.2068817168,-0.3980059624,-0.3207308054,0.3920006752,0.1008677259,-0.0965765566,-0.0322490409,0.1750874668,0.1630067676,-0.1050071344,0.1777083725,-0.3961575031,-0.4213402271,-0.3179362416,-0.1166030914,0.1954880804,0.1510435343,-0.0724301487,-0.2274181098,0.3431016505,-0.0656266734,0.3664130867,-0.1784325093,0.2733654082,-0.1361103207,0.1191522181,-0.0020067152,-0.1793031245,0.184699133,-0.3458022177,-0.251807183,0.032944642,0.0749178231,-0.0649888813,0.0640867501,0.0777259693,0.1255937666,0.222927317,0.276268661,0.2730693221,-0.2841854692,0.39382568,-0.1296219826,-0.2305816114,0.2861991525,0.142140016,-0.2728927135,-0.2016637623,-0.2415083349,0.5372298956,0.140845418,-0.0927921608,0.1432565749,0.2090305984,-0.2544732094,-0.1576039642,0.0293197017,0.3020511568,-0.167654708,0.007802851,0.0516034067,0.0572213642,0.1359664351,0.2370682359,-0.2826177478,0.3070636988,0.1281833202,0.0701621026,0.7291414142,0.0487254411,-0.2612171471,0.0896240622,-0.1039213613,-0.1404579282,-0.3533899486,0.1460213214,0.1700755954,-0.1654422134,0.1211969182,0.4095906317,-0.0198548436,0.2734885812,-0.0199103281,0.1302362084,0.4042031169,0.1956016272,0.154847905,0.4226060212,0.387391746,-0.2005483657,0.2112674117,0.0177641828,-0.122617282,0.1027633026,0.0369337723,0.1049816608,-0.1853458732,-0.3026823997,-0.2737396061,-0.0160694607,0.1242413893,-0.2392200828,0.2451300174,0.0663407966,-0.0298823714,-0.2867467999,-0.0083352551,0.0714734867,0.0394173674,0.1102622151,0.2704003453,0.0698171183,-0.4579935074,-0.3474185765,0.1517035961,0.0758124366,-0.2192331851,0.0689463839,0.2675577998,-0.0189865399,-0.2657428682,0.0552398711,0.2072344869,-0.3247493804,-0.0193946585,0.082230635,-0.1089788303,-0.124755092,-0.1685072482,0.0943605527,0.2574067414,-0.3492699564,-0.0012715879,0.0770416781,0.0374644287,-0.2408648431,-0.2252986878,0.0303203184,-0.0083339028,0.3794999421,-0.2728161216,0.1088346317,-0.2037817836,0.1066530868,-0.2001684606,-0.1808640808,-0.3133023381,0.0936442316,0.1538196355,0.2447134852,0.1583609134,-0.0856839195,0.1778128296,-0.4150200486,0.294040978,0.1356187314,-0.0065027662,0.2894739211,0.2508842945,-0.0469171815,0.135107547,0.0908341408,0.1811917275,0.0876293704,0.1503839046,0.0102306344,-0.0917462036,0.168288067,0.2751570344,-0.3935909271,0.1173689961,0.0705024377,0.0121657941,0.0119724208,0.0986556709,0.3968876898,-0.0744666457,-0.0319188163,-0.2503604293,-0.2357503325,0.188651666,-0.2500239313,-0.2018368095,-0.0827581361,-0.2698657811,-0.3156387508,-0.2845276296,0.2515695393,0.149199307,0.0654856339,-0.0795319602,0.0000093888,0.2545885444,-0.3420854211,0.0842508301,0.1118056849,0.0858539343,0.0837690011,0.2699506283,-0.0221259966,-0.1147832349,-0.0456781872,-0.220791772,-0.0719360188,-0.2751031518,-0.2407620251,0.0123918178,0.3630637228,-0.4689585865,-0.4008159339,-0.0334208384,0.2728086412,-0.0141492775,-0.0257588997,-0.1722587347,-0.0232361574,0.1104000211,-0.1823539883,-0.3897988796,-0.4695669115,0.4340716302,-0.0022272381,0.4755172431,0.3498813212,-0.2227888703,-0.2277327925,-0.3337567449,0.0533057451,0.1961311996,0.0593283772,0.3909672499,-0.0712433681,-0.0153328609,-0.0589640215,-0.3241891563,-0.1117422953,0.3016292751,-0.319794476,-0.2963643074,-0.2197163105,-0.3705636263,0.2016861141,0.5963969827,0.0404520743,-0.0684807971,0.0274935272,0.0765486658,-0.1149018556,0.1333310008,0.5127104521,-0.0457310788,0.2259085476,0.2016773522,0.1363451034,0.130717352,0.5562351942,-0.1837156713,0.4135272503,0.1172994375,-0.0695249885,0.2995041013,-0.5126637816,0.1960013062,0.037316259,0.031881161,0.2471691221,-0.1981822699,-0.6117690802,-0.110662967,-0.3204655349,-0.5573168993,-0.0057524298,0.1147149801,-0.2643935382,0.305106312,0.0083394004,-0.2482305616,-0.162642315,0.0056717978,0.0757664368,0.0938232318,0.3397384286,0.1481895745,-0.3755730987,-0.2797023952,-0.0307394955,0.3100639582,-0.027074961,0.094660908,0.0054682465,0.0150120631,-0.1242058352,0.2167576104,0.3206866086,-0.3616738021,0.023249045,-0.0372747742,-0.2636862695,-0.338596791,0.0843621194,-0.015294265,0.2661997378,-0.0007810814,0.451525867,-0.0028929198,0.0498188324,0.2977621555,0.0040654391,-0.0262203887,-0.3989346921,-0.0858621746,-0.3038944602,-0.1806055158,0.1384629011,-0.125402987,-0.0534690171,0.6754009724,-0.4109427929,-0.0384374335,-0.1643060595,0.0398279056,0.2246585786,0.0514435843,0.7657280564,-0.1078651622,0.0919225439,-0.2431751192,-0.4160998762,0.3302589059,-0.3608972728,-0.1958621889,-0.1200618595,0.3429864645,0.6795859337,0.1841572821,0.299412787,-0.0359069593,-0.1163837388,0.0630756021,-0.1164652258,0.1447984576,-0.1813229322,0.115350157,-0.0391874909,-0.6432768106,0.2892981172,0.1427851617,-0.1518750787,0.340279609,-0.1736778617,0.0077112941,0.3372370899,0.2406185269,0.7881451249,-0.1372397393,-0.1632315069,-0.0205552913,0.7625697851,0.6641044021,-0.2361039221,-0.0323252417,-0.1785178185,-0.2925643325,-0.0352246165,-0.213368997,0.2048580796,0.3708994389,-0.1785287112,0.5122699738,0.3150480092,0.1817046106,0.038264256,-0.3554986417,0.1721025705,-0.1785200238,-0.5172115564,0.0954223946,-0.3391722739,0.3816054761,-0.0494047664,-0.0234678164,0.0703422576,-0.2255607098,0.0153989214,-0.0242839679,-0.660474956,0.0222635083,-0.1948675066,0.0638944656,-0.0007190148,0.5309045911,0.1167225465,0.3031152487,0.2959515452,-0.1647960991,0.3881299198,0.1856658608,-0.2093065381,-0.0450366773,0.0566418618,-0.0369068421,-0.1922156662,0.1433984786,-0.1357508898,0.1986469775,-0.2690739334,0.3723040521,-0.2057614625,0.1727426797,-0.3646772504,0.3819588125,-0.0672499537,0.1517970264,0.0795325711,0.0804218799,-0.2926427424,0.5523737669,-0.2703634202,-0.2291410863,0.2495518476,0.0520577282,0.0709508732,-0.1107843071,0.0105687901,0.0422277041,-0.0927993208,-0.0826717541,-0.0159973465,0.1524496526,0.3950795531,0.0012186496,-0.1351241022,0.0030161564,-0.0863255411,0.3975607753,-0.0003212961,0.1630861014,-0.1143250167,-0.1965066493,-0.2969972491,0.3902535141,0.1569690406,0.1204823926,0.1778272092,-0.061930526,0.0480783693,0.6994156837,-0.2451494187,-0.134376049,0.0100318408,0.4785548449,-0.0982972756,-0.0434188657,-0.1925055832,-0.0898675844,0.0768195465,0.3194635212,-0.2773693204,-0.2354900837,0.1009974852,0.1174335927,0.1800746024,-0.2703796625,0.1865201294,-0.0907022581,0.097317107,-0.2262709886,0.0403580889,-0.2605613768,0.0476573072,0.213626653,0.2292589247,-0.4193480015,0.1375194788,0.0446359217,-0.3279216886,-0.1831148267,0.1258450001,0.0362509266,0.3402777612,-0.0572264269,-0.109859772,-0.280793041,-0.0153201958,0.1317791641,0.1925884783,-0.0280733611,0.1194089726,0.1281414181,0.2289205641,0.0805702433,-0.1616594791,0.2995042205,0.1766488999,0.1875767708,-0.2355950028,0.1448364109,0.6082422733,0.1890717,0.2752243578,0.3260062337,-0.2391822934,0.0774006397,-0.3245652914,-0.0767550394,0.1247605011,-0.2143122703,0.1947287917,0.4218851626,-0.02379876,-0.3411378264,-0.1802949756,0.3591167629,0.2005905509,-0.1298374087,0.2101326138,0.2938283086,-0.0397457071,-0.1286824942,0.2178430855,-0.5101249814,0.5423719883,0.2294928282,0.1952519417,0.1537064314,0.2123911828,-0.1266253442,-0.0500667617,-0.35012272,0.3728628457,0.3403010368,0.0668193176,0.1080910191,-0.2700201869,0.0761764422,0.0666221306,-0.1224667728,0.022128351,-0.7313489914,0.0843331292,0.4144398868,0.1705940068,-0.2066475004,-0.1368201524,0.1261467189,0.3320560157,-0.0143156517,0.2223874629,-0.0677830502,-0.1387091428,-0.200136438,0.1937905997,-0.0657305196,0.2071232498,-0.1204869375,0.4959211946,-0.0995185226,-0.1675240546,-0.0192283411,-0.0857645497,-0.0332747586,0.3065463603,0.0285603348,0.1185394451,-0.0567149147,0.3490054011,-0.3261077106,-0.2722033858,-0.112437062,0.0818281546,0.1908946186,-0.1562822163,-0.0066830711,0.0659257248,-0.4544784725,-0.2709435225,0.2789897323,-0.0782099664,0.1175413057,-0.0914003327,0.022312142,-0.0264507551,0.2839033902,0.0979386121,0.5311249495,-0.295505017,-0.1327778697,-0.3451567292,0.0091733485,-0.1468972862,-0.101229839,0.2663759887,-0.1248848811,0.4124905467,0.2681246996,-0.354583621,0.3331113756,-0.0226475298,0.2262219489,-0.0922166035,-0.0225973632,-0.4070357084,0.0666814819,-0.0067180512,0.0243441164,0.1069131941,-0.2150768489,0.0053401403,-0.0653874576,0.1105947047,0.0862649605,0.1714591831,0.2196976691,0.0542849675,0.5418863297,-0.1141247526,0.2397624701,0.0830029026,-0.0316899121,-0.3055813313,0.1571092308,-0.1852190197,0.259133786,-0.2265454382,-0.1178789437,-0.1746517718,-0.1370829344,-0.0195567794,0.0343185328,0.0991390422,-0.1442154348,0.1081581488,-0.1184978187,0.1772360802,0.231183067,0.074653402,0.1332533509,-0.1550638974,-0.14821814,0.3490909338,-0.4168328345,-0.263535887,0.0534020886,0.08451581,0.2143753022,-0.0040828502,-0.2363670915,-0.2570492625,0.3445166945,-0.250657171,-0.1805705428,0.244221434,-0.0873721987,-0.2828139961,-0.1181635559,0.060755197,-0.1219865084,-0.232496202,-0.231673792,-0.3682834208]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/117","title":"\u2753 How to remove specific rows of a dataset ?","comments":"Hi, you can't do that at the moment.","body":"I saw on the [example notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb#scrollTo=efFhDWhlvSVC) how to remove a specific column :\r\n\r\n```python\r\ndataset.drop('id')\r\n```\r\n\r\nBut I didn't find how to remove a specific row. \r\n\r\n**For example, how can I remove all sample with `id` < 10 ?**","comment_length":8,"text":"\u2753 How to remove specific rows of a dataset ?\nI saw on the [example notebook](https:\/\/colab.research.google.com\/github\/huggingface\/nlp\/blob\/master\/notebooks\/Overview.ipynb#scrollTo=efFhDWhlvSVC) how to remove a specific column :\r\n\r\n```python\r\ndataset.drop('id')\r\n```\r\n\r\nBut I didn't find how to remove a specific row. \r\n\r\n**For example, how can I remove all sample with `id` < 10 ?**\nHi, you can't do that at the moment.","embeddings":[0.3195144534,-0.1150931641,-0.1154173687,0.0429543667,-0.148636952,0.1822795272,0.4059865177,0.2484357357,-0.0649004579,0.1665265709,-0.0613705814,0.3026041985,-0.0005447132,0.2589341104,0.1649005711,-0.1265789419,0.107849136,0.3622194827,0.1438154578,0.0357861109,-0.12901932,0.0133167645,-0.4908776283,0.1374447197,-0.1689482331,-0.3568274081,0.0657751784,-0.1326453835,0.0282774735,-0.0145286778,0.4869680703,0.2867501974,0.012244788,-0.0229781214,-0.0001171699,-0.284116298,-0.0174166635,0.034599185,-0.0954544023,-0.0155803217,-0.1928839386,0.0369167067,0.019420851,-0.236968711,0.0230175462,-0.3532491624,-0.2305090874,0.0727846697,-0.0413926989,0.6193379164,0.1458798498,0.1371319294,-0.1837350577,0.2167124897,0.2721369267,0.1181053966,0.0212151241,0.1122839004,0.3035219014,-0.1664502323,0.3842768371,0.0209946968,-0.1652896553,0.2441660762,0.1945979744,0.0298904572,-0.3561968505,-0.5581288338,0.3898473084,0.3672844172,0.526933372,-0.1987100095,0.0276576709,-0.0429956876,0.1142777726,-0.2458896488,-0.0986908451,0.3125723004,-0.3729000986,0.2107402235,-0.3712187707,-0.1312124133,-0.1721686125,0.2429539859,-0.1244104058,0.1477076113,0.0317163132,0.0177166145,0.3931277394,-0.007985469,0.2494326085,0.2570914626,0.1877815872,0.5053792596,-0.4821478724,-0.3959023356,-0.2118965983,0.210126996,0.3495928347,0.0268982481,0.1377289891,-0.1541674435,0.1830329448,0.239538312,0.4213148057,0.090785563,0.0488179885,0.2753781378,0.1400183439,0.0425068699,0.0540948883,0.082783632,0.1670957506,0.0397296026,0.1483879536,-0.0059447456,0.4720786214,-0.3539786339,-0.716535449,0.3472976983,-0.663333714,-0.0075972825,-0.2256766111,-0.0256129298,0.3905863762,-0.2702774704,-0.1877991259,0.2828870118,-0.491548866,-0.1478595436,-0.0807580501,0.417632997,0.0192718804,-0.4172714651,0.3264673352,-0.253839612,0.3160030842,0.0969843045,-0.1196758151,-0.2436455935,0.1815968007,-0.1873383224,0.1577031612,0.4009289742,-0.2604040802,-0.2491678596,0.0618541017,-0.0206850152,-0.2729716599,0.3853448629,0.0266993046,-0.0578257516,-0.1221964732,0.0757750422,0.0062102843,0.1634425521,-0.1454113275,0.1817570031,-0.2496588528,0.2580812275,0.1171284243,-0.2936651409,0.0222596712,-0.1024721861,0.1278496981,-0.1541228294,-0.5935997367,-0.0321826003,-0.0056834263,-0.2117248923,-0.0274786428,0.1812636703,-0.1195086241,0.3713586628,-0.0280134585,0.0748820975,0.1643677503,0.0073022074,-0.310577333,0.1052219197,-0.0620642677,0.0033335851,-0.4492588937,0.1022063419,0.4106502235,0.0245166086,-0.0165190399,0.199386254,-0.0557825342,-0.2123739123,0.1492182016,-0.2712383568,0.2267190963,0.1249991655,0.0296498295,0.0309542865,0.3292108476,-0.022484282,0.265307188,0.211957708,-0.0534084477,0.2985645831,0.2766297162,-0.2218134403,0.0406000167,-0.379009068,-0.4411446452,0.0750983804,0.0106861517,-0.2429122776,-0.0379546694,-0.2811900079,-0.1543458849,-0.1675964594,0.1501708925,-0.0770617351,0.0285589378,-0.5415883064,0.2341533154,-0.1981397569,-0.6277296543,0.2048902363,-0.1460404694,0.2200158834,-0.424307853,-0.1609514505,0.117100969,-0.1311726719,0.3385491371,0.1777295321,0.1417569667,-0.0655929446,-0.0024840338,0.1695273817,0.3137876987,-0.4677219689,0.0184425078,0.1246956885,-0.0172228832,-0.1497854292,0.2254935354,0.2237529904,-0.0310708061,0.1247879714,-0.307302177,0.5156716704,-0.4067123234,0.1849460304,-0.2728995085,-0.0739486367,0.2019453049,-0.2781563699,-0.1101066619,-0.0572779439,0.2786914408,0.2526318133,0.1506437361,0.0238297582,-0.4241728485,0.0088173496,-0.1845770478,-0.1310366988,0.0457037799,0.1417726725,-0.1361454427,-0.0373879671,-0.034859363,0.3764725327,0.3724847138,0.1472775042,0.0819054618,0.1434179544,0.2359379381,-0.1057497263,-0.0289989971,0.1087472066,-0.1482945383,-0.2959666252,0.1322414726,0.1576522142,-0.3020285368,-0.0134719545,-0.1942407638,0.5720694661,-0.2220956981,0.2641701698,-0.3117522299,-0.4328802824,0.1755643636,-0.1225748807,-0.2159014493,-0.0650420114,0.0270747021,0.004371786,0.0242254529,0.2891863883,-0.1688486487,0.4449323118,-0.2257707119,0.4436966777,-0.0312025826,-0.0104155336,0.1170916557,0.176558435,-0.1655818671,0.763374269,0.433752656,-0.1559776217,-0.3736501634,-0.1115748286,0.0844970867,-0.0350156389,-0.2024555504,0.3999218047,0.3613839746,0.5423498154,-0.4491961896,0.2632742226,-0.3641259372,-0.1741336882,-0.044154197,-0.1127873361,-0.018710563,0.3037776053,0.019826198,-0.0775432363,-0.0653871149,-0.1613538116,0.1182699427,-0.0239558145,0.2999584675,-0.1405731738,-0.1769367307,0.0591099598,0.012956595,-0.1737507433,0.0367431045,-0.6337899566,0.0970459282,0.2331214994,-0.347766608,0.2170951366,0.2604687512,-0.19588691,0.0516694821,-0.1124493703,-0.2835792601,-0.1960168332,0.1196214184,-0.1860349178,0.0449274443,0.1933713406,-0.2039949745,0.016647663,-0.1289180666,-0.0591687709,-0.1474360824,0.19821769,0.0519673079,0.3370753527,0.213517487,0.0724880025,0.3554419279,0.3928786814,-0.0831176341,0.1044803783,0.2467607558,0.1051203609,0.1517246366,-0.3325433731,0.038802173,0.2459724993,-0.0024637417,0.3660835326,-0.0474806838,-0.1110640913,0.1292196214,-0.2193748355,-0.1516161263,-0.1602499932,0.2399027348,-0.2172161788,0.1685813069,-0.0926323235,-0.2270818949,0.0620174147,-0.1549541056,-0.2321469933,0.0070868763,0.3379510343,0.1019186154,-0.3722668886,-0.4741045237,-0.5162418485,0.2805936038,-0.1781900078,0.17373541,0.0921820849,0.3119992316,0.3345897198,0.116243571,0.3842256367,0.1312207282,0.1969231963,0.0067712069,0.0221276395,-0.1147764772,0.205423519,-0.3171752989,0.253752619,0.1974568218,0.0852990225,-0.2143047899,-0.5378309488,0.5150711536,0.3362970352,-0.2268784493,-0.3404119015,-0.0915860385,-0.4938678145,-0.2321663499,0.3251407146,0.2498505116,-0.1315113157,0.1287425309,0.0312443841,0.1091367975,-0.2323795855,0.1734576374,-0.0596805066,-0.002110516,0.057715036,-0.1387811601,0.2739112079,-0.0895967856,0.3219866157,0.2958584726,-0.0084488215,-0.3716020286,-0.3503487408,-0.2643249631,0.5421429873,-0.0271670893,0.0813360885,0.4981842339,-0.0909335166,0.1206037104,0.2798022032,-0.0579344071,0.0177281946,-0.2233750075,-0.0714828745,-0.4139633179,0.18378672,-0.118119821,-0.1093953922,-0.058896184,0.2627877593,-0.273755759,0.1736771315,0.1810863465,0.8084107041,-0.1068662554,-0.241911605,-0.4169403911,0.0649784133,0.4524412751,0.1022863388,0.0189097449,-0.2148054093,-0.2473249137,-0.2409019768,0.0049320017,0.3633077443,0.2643376589,0.2434782833,0.3407578468,-0.4103443623,-0.2800236046,0.2318773717,-0.1295079142,0.2403042912,0.0682208091,-0.4648582935,0.0463437811,-0.4436355531,0.1261204034,0.0224465784,0.4067936838,0.1164493859,-0.2702439427,0.0250151549,-0.249681592,-0.2047379017,0.0406082384,-0.0122528775,-0.666903019,-0.1141446382,0.2339348048,0.0957636088,-0.2037639767,0.2530967593,0.0703448281,0.272652179,-0.0805615187,0.1217546612,0.1717057377,0.079768464,-0.079038173,-0.4070646167,0.1824309379,0.0905253291,-0.232837081,0.1116219163,0.0808613822,-0.3483428955,-0.15165025,-0.253605783,-0.3784405887,0.2578878403,-0.1360731125,0.0988551974,0.4056170881,-0.0938894525,0.3096752167,0.013764523,-0.1814522743,-0.1672369093,-0.0622890405,0.3044808209,0.0618307739,0.3587664366,0.1878426671,-0.2491004616,-0.2708142102,0.2529279292,0.0438622534,0.0838874206,-0.0093003642,-0.0337954387,0.4598587453,0.105659157,0.0543784574,-0.2745966911,-0.2314849049,-0.0291188788,-0.4467261732,-0.3174610436,0.2357432395,0.2806824744,0.2126339525,-0.0908714086,-0.0459955893,-0.2005638629,0.3882760704,-0.1892061532,0.2704887092,-0.1037368774,0.3504751623,0.3002181947,0.4645816982,-0.2163719684,-0.2744876444,-0.0604768321,-0.3467896581,0.2020756304,-0.2215917706,-0.3779397011,0.1566741616,-0.1985049695,-0.2817323208,-0.1910455525,0.1041088104,-0.0283291955,0.0079919053,0.2434876114,0.1352207214,0.4007274806,0.6270326972,0.5967047811,-0.2021615803,-0.2360228449,-0.0456733443,-0.0701104179,-0.0625355169,0.2089289129,0.0540908501,0.2158486098,0.1069788039,-0.2055553049,-0.0608560704,0.1302799731,-0.2146727294,-0.043414101,-0.0745585337,0.2164342999,-0.0368532091,0.1009351388,-0.0942161903,-0.1893699318,0.1826564968,0.1411887258,0.115688026,-0.326369971,0.0146624334,0.551097095,-0.017560713,-0.1486718655,0.131392017,-0.2150288075,0.1394655406,-0.3185628653,-0.0252200551,0.4702924192,-0.0914627984,-0.1209942475,0.3464027643,0.159976989,-0.0936273932,0.3752292097,0.3284036815,0.0633213073,0.3597099483,0.0024134752,0.2649602294,0.1723473817,0.0970616341,-0.1423364729,-0.3119592071,-0.051384531,0.4301647246,-0.1044527739,-0.1611288339,0.2774640024,0.1627943218,0.1305462271,-0.2144176364,0.1853726804,0.250703156,0.11806117,0.0212173313,-0.244486779,-0.0798726156,0.3067832291,0.0126046184,-0.1452722847,-0.2019272745,0.3505740166,0.1172055602,-0.1492522508,-0.3351097703,0.1645326763,-0.2707330883,-0.028621316,0.0361121781,-0.296772331,-0.3477041423,0.2757103741,0.1249609292,0.0838681236,-0.0801272914,-0.2138256282,0.0082750097,0.0389328003,-0.3822173774,-0.0173004903,-0.0008150119,0.0203352775,0.0894079283,0.186372444,-0.1908777207,0.0735877752,-0.0631586835,0.1849762052,0.3243218958,0.1733853668,-0.6030876637,0.4084729552,-0.089494586,-0.3953003883,0.0134837693,0.0602511428,0.106521368,0.3144837022,0.0732493773,0.2356896102,0.0807632282,0.1479012966,0.0649542585,-0.2855641246,0.1393537223,0.2469562888,0.0352709182,-0.2926810086,-0.0713005811,-0.5504330397,0.0522704422,-0.4476322234,-0.1090249941,0.5795437098,-0.1650570631,0.0913846195,0.2037355006,0.1251773387,0.063348271,-0.0756596625,0.2547255754,-0.1718556732,-0.1367690265,-0.3951804042,0.3768397272,-0.232987687,-0.1285158396,0.0522026829,-0.3823714852,-0.0255681966,-0.1453130841,0.0966025516,0.4148057699,0.0043511875,-0.1165730059,0.074149996,0.3338783383,0.0679166764,0.0185257476,-0.0078606876,0.2295695394,-0.2426483184,0.2582051158,-0.06899672,0.1123594791,0.0347421095,-0.0828165337,0.179306373,0.2987140715,-0.3361296356,0.1300754249,-0.1687180251,-0.1085173339,0.190850541,0.0330564119,0.2978493273,0.1636612266,-0.1788155437,0.0902692452,-0.4294871092,0.0506075695,0.1570049971,-0.3811496794,-0.4620200098,-0.1487098336,0.3585850894,-0.2993304729,-0.4676374793,-0.3070907891,0.12633802,0.1233866811,-0.3441030085,-0.2083365619,0.3825998604,0.017445296,0.062715359,0.2475798875,0.2119242102,0.3137841225,0.3052134216,-0.150949955,-0.2519218624]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/116","title":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323","comments":"Can you share your data files or a minimally reproducible example?","body":"I'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)","comment_length":11,"text":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\nI'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)\nCan you share your data files or a minimally reproducible example?","embeddings":[-0.0484108999,0.1494438648,-0.0144368587,0.2324064672,0.2381852567,-0.2273712456,0.0248729773,0.3209259808,-0.3282232583,0.1879350692,-0.0436761267,0.5396180153,-0.028553484,-0.5326482654,0.0905098915,-0.1926954836,0.0212951638,0.2724654377,0.3803286254,0.0611108318,-0.1704670042,0.370685637,-0.1781456918,0.2907916903,-0.2035011947,-0.02081066,-0.1948845237,0.0629046708,-0.3188581169,-0.4638895094,-0.1054828689,-0.1024661958,0.0979508758,0.2689797282,-0.0001244346,-0.1751092374,0.1868264824,-0.0762416348,-0.0275540017,-0.3462415338,0.1128415912,-0.4930665195,0.4031451643,-0.4177653491,0.170830071,-0.6244263053,-0.2995930612,0.1952298582,0.4168790281,0.3945330977,0.0441393778,0.0936739668,0.1324967444,-0.2007661611,0.3989597857,-0.2748504877,0.066698499,0.89983356,0.4558342397,-0.1448000073,-0.0401322581,0.2228201926,0.0730428994,0.0149453497,0.4327254295,-0.0533877984,0.2792081237,0.2867449522,-0.0949268043,0.3374998569,-0.0839335546,-0.5187918544,-0.2566540539,-0.1076821983,0.185643062,-0.8163092732,-0.1286412477,0.2110202461,-0.0102362223,-0.0445661247,-0.1762859374,0.0043327361,-0.202665478,0.296056509,0.0443157963,-0.0123850191,-0.0464996807,0.0817476511,0.1358697712,-0.137044251,-0.3099374473,0.320639044,-0.3387844265,0.0737477094,-0.3465700746,0.1309144795,-0.0482992455,0.0948025733,0.2325206995,0.0529242381,0.4461768866,-0.1346886605,0.087285392,0.1181821451,-0.1563505679,0.6945234537,0.2288198471,0.1688499004,-0.2089546472,0.0280153565,-0.0058058403,-0.2579502463,-0.2268987298,-0.478284359,0.0641953051,0.197565183,-0.1124412417,-0.1427302808,-0.3596061766,-0.0701090172,-0.3943303525,-0.0711067244,0.1218513921,0.3590536118,-0.10140495,0.1525862217,0.0726632923,0.4597468078,-0.0258450303,-0.2102440894,-0.0580233335,0.2112616152,-0.5268654227,0.2261209041,0.1351095885,0.4433135092,0.21714665,-0.0752042532,0.3243650794,-0.3267936707,0.1445037276,-0.2008171231,-0.1258250028,-0.2789407074,-0.0631981045,-0.1082888767,0.0142105436,-0.5029014349,-0.1532485634,0.0990718454,-0.3687569797,-0.1662993282,-0.1026472896,0.0120840343,-0.1833140552,0.0745351911,0.1055172533,0.2640973032,0.2056305707,-0.4256345332,0.18212834,-0.100942336,-0.0436721109,-0.203020826,0.2189773619,-0.0758219659,-0.3536063135,-0.0894801766,0.4526171684,0.4528063536,0.1727126688,0.5272013545,0.0137373563,0.3521324396,0.0813144743,-0.0354538858,0.8424562812,-0.6846866608,-0.2231179178,0.1149474159,-0.0709354579,-0.4116330445,0.2602656484,-0.2383692265,0.051168099,-0.0143754147,0.4653161764,-0.0849721134,0.0472368076,-0.2508241534,-0.2991683781,0.0040747006,0.0588710941,0.0084642153,-0.2811980247,-0.0897741318,-0.0488331206,-0.2200994492,0.3138251305,-0.1383665502,0.0722981393,0.4024355114,-0.2080475241,0.1982601285,0.2201012075,-0.0539953969,-0.0007820445,-0.2002329826,-0.6982911825,-0.2092123181,-0.3664566875,-0.322666347,-0.301951617,0.0502164252,0.3635741174,0.0460539348,-0.0738419965,-0.0264036693,0.0759646818,0.106605649,-0.02407947,-0.3763007522,-0.2422619313,0.0683166161,-0.0283436328,-0.1287142783,-0.0405510403,-0.3031941354,0.2618334889,0.5115119219,0.2998487949,0.2185798436,0.0127100432,0.3357186913,-0.0530818366,0.0385160632,0.0031763704,0.166727677,0.217645973,-0.3526036143,0.1132125407,-0.1348578483,0.1628439128,-0.1702944785,-0.2253958434,0.2233021259,-0.2498648614,0.2700778544,-0.0867912471,-0.1561401337,0.0460926257,-0.0317321122,-0.3636042774,-0.1251443326,0.3140366673,0.1993118674,0.1983759552,0.014862584,-0.0893616229,-0.2343324125,0.5126395226,0.2382301986,0.0453368314,0.0966138914,-0.201578334,0.0357357375,-0.0162527058,-0.4718225896,0.2397544384,0.3644063473,-0.0316462107,-0.0221429951,-0.0429177172,-0.3460890949,-0.0413118601,0.1400948316,-0.0198436752,0.4519988,0.4667825401,-0.0790605098,-0.0504423156,-0.3444618285,0.1056927517,0.018145632,-0.2869449556,0.0854994953,-0.0928440914,-0.08512263,-0.1231979355,-0.5676968694,0.1299020797,-0.0849809274,-0.0126286801,-0.1245169491,0.0956618786,0.2610791326,0.3190778494,0.1672564,0.1541433483,-0.0122026289,-0.0860986412,-0.3945364356,-0.0906882733,-0.090967983,-0.1061119139,0.0678256005,0.1470550597,-0.0549308024,0.0349823758,0.0622430481,-0.3803509474,0.0295127071,0.0643559173,0.0784752294,0.228306964,-0.1341003478,-0.3366881311,0.0902048871,0.5146914721,0.1201803684,-0.0716831684,0.2031467259,-0.2457646728,0.0037806747,-0.023147881,-0.0450857617,0.0027600732,-0.5757048726,0.5463731289,0.1712348312,0.0098638982,0.1131588593,0.1204748452,0.1200160831,0.0486254953,0.1461982727,0.2107058614,0.2465798557,0.358859092,-0.2526665628,-0.3568253815,-0.1079798043,-0.082408011,-0.013710835,-0.0473912358,-0.003807538,-0.3536386788,0.0570606552,0.0840445757,-0.1737281978,-0.0157784019,0.3159812093,-0.1218546927,-0.0439049937,-0.2451833636,-0.0274841283,0.2959914207,-0.1531527638,0.2863889635,-0.0060745403,0.2986389697,0.1093694419,0.2220389545,0.2839267254,-0.0812405199,0.1299683601,-0.2488677502,0.0734260902,0.0621435903,0.0287926625,-0.1874070317,0.0320397615,0.0243994705,0.2567745149,0.0383370779,0.1051502079,0.0316436552,0.082252942,-0.1534430832,0.0680896863,0.1561764926,-0.4296875,0.1365559101,0.0142846908,-0.2073156387,-0.1874194294,-0.1745846421,0.059278477,-0.20021227,-0.1064156592,0.0999170914,-0.2455765903,-0.2447274774,-0.530651629,0.4471029341,0.197530061,0.0662917346,-0.0049310951,0.0438427553,0.0788660273,-0.1666196287,0.3349824846,0.118058145,-0.0518371798,0.3908744156,0.1489331722,-0.7703570127,0.0682836473,-0.2295815498,0.0862705261,0.4799775779,0.2262803465,-0.3940051794,0.1249990612,0.1921350956,0.0500689819,-0.0873920247,-0.1040006131,-0.2994387746,-0.1654075384,-0.0009769896,0.0253288932,0.1024899408,0.530312717,0.3234949112,-0.0581373498,0.073858358,-0.1855965704,0.08614178,-0.0218464937,0.3967503011,-0.2346101999,-0.2176067084,-0.08635851,0.180011332,0.03742579,0.1994260997,-0.077761434,-0.3718473017,0.0304684937,-0.1235133111,0.484002471,0.2642645538,-0.0983748734,0.1202832162,-0.4368756711,0.2414758205,0.0975490883,0.0749985427,0.3098908663,-0.0936266333,-0.1041781604,-0.0282442588,0.4248251021,0.299140662,-0.1348480731,0.4283899963,-0.1488250196,-0.1779367477,0.3555758595,0.0096270265,1.2682539225,0.1330196708,0.1622652709,0.3792473376,-0.1121312007,0.5393418074,-0.5706241727,0.060709957,-0.328659296,0.0411214307,0.1694737673,-0.2111629546,-0.1630334407,0.2189146578,-0.1712940931,0.1046974733,-0.0355872586,-0.0068458761,0.0451460667,0.213304922,0.0584504046,-0.0736657903,-0.1618646681,-0.1069007367,0.0053723026,-0.3707917333,-0.2145620137,-0.2374691665,-0.4095864594,-0.1064157113,-0.0987536237,-0.0213390458,-0.1782251894,0.1174107641,0.2197524607,-0.2645194232,-0.2058391124,0.479606837,0.1684680283,0.020650709,-0.1389461756,0.4515068531,-0.0753368139,-0.1914983243,0.0527978726,0.1406685114,-0.1533929855,-0.1301668435,-0.0539266951,0.2589164078,0.053553097,-0.1797603369,-0.1086256057,0.1197632998,0.0353318974,-0.0817672089,-0.0935092717,-0.0750810057,-0.0371411033,-0.114557445,0.0121744955,0.1973262727,-0.2239903212,0.0889098197,0.1238299385,-0.4186739922,0.1349578351,0.2161296308,-0.0281959437,-0.0298421886,0.2064988464,0.1256942004,0.0638106242,-0.0820103437,-0.1771955937,0.0062144287,-0.6206239462,0.15438205,0.2448551506,0.2416170537,0.0788989067,0.5654673576,0.4303180277,0.2418850213,0.1621096879,-0.5253363848,-0.138990283,0.393784225,0.0411351584,-0.0012862665,0.0131383855,0.1209195182,0.0944202915,0.4704739451,-0.1819507778,-0.0967711508,-0.1706084311,0.2223894298,0.0085048629,-0.0934698209,-0.0153280394,0.2147357017,-0.0700510144,0.3550094962,-0.0328202285,-0.1032798588,-0.1331746727,0.1831610054,0.1164912358,0.1827006042,-0.0876715034,-0.0247142911,0.1785601377,-0.2734658718,-0.1133090034,0.039793916,0.0571353585,0.1176882088,-0.0504314564,0.2396285534,-0.0993257463,0.0896985754,0.1377465874,0.1381872147,-0.1886792481,0.0439348482,0.0833187774,0.0111554433,-0.1698138416,-0.0239825509,0.116372034,-0.2154205292,0.1004935279,-0.0341722928,0.1504902244,-0.3534037471,0.3587049544,0.2903285027,-0.2548568249,0.1399508268,-0.1936226338,0.0444997698,-0.2725124955,0.1284924001,-0.1254521906,0.1464235485,-0.1526830047,0.0971887112,0.247060746,-0.2337091565,-0.0371146835,-0.0647044182,0.1030392125,-0.0930313617,-0.3742157519,-0.0057695187,0.0689240173,-0.0456708334,0.4323097765,0.4596277773,0.3494678736,0.3223532736,0.0015933571,0.1053435653,0.289255619,0.1755398363,-0.2148514092,0.0220537316,-0.3992324173,0.2694867253,0.2940640152,-0.3090040982,-0.2928163707,0.505859673,0.4063033164,-0.1386972666,0.2061246037,0.3077871501,0.15110147,-0.2271926701,-0.6417418718,-0.0092034042,-0.2397990376,-0.288621515,0.0163221769,0.3405055106,0.156011939,0.1659024507,0.1301397085,-0.411116153,0.035026174,0.2408800721,0.2254923433,-0.3800949156,0.4668016136,0.1068027914,0.1648666114,0.1657349169,0.2970757484,0.5358623266,0.3125315905,-0.0251224246,-0.2034834176,-0.5653469563,-0.0419349633,0.0613391809,0.4115430117,0.6082406044,-0.0325935334,0.2404124141,-0.0346725993,-0.1047688946,0.5051207542,-0.0761076659,-0.0347303823,-0.0782301649,0.2921956778,-0.0342162475,-0.3350933194,-0.2405047715,-0.1278740615,-0.4312065244,0.0331073776,-0.1096360385,0.0538936332,0.1234204099,-0.2271036953,0.0029862849,0.0392661281,0.0182844531,0.4407159388,0.4052073956,-0.2754095495,-0.4549683034,-0.3413536847,0.0139644276,-0.2679379582,-0.282499522,0.1010745466,0.2545700669,-0.0266171489,0.1249921471,-0.0678727776,0.7530362606,-0.2742034197,0.2584146559,-0.4447611272,0.2618383467,-0.0108801685,0.0784202591,0.0263246782,-0.0474146642,-0.0208975039,-0.3052112758,-0.1192710176,-0.1387607753,-0.1895947307,0.2239844352,0.6866971254,0.5715528727,-0.0459099412,0.6867462397,0.1533572078,0.2292416245,0.0432044342,-0.1098712757,0.0579686798,0.2281459272,0.0379402041,0.0418352969,-0.3195984364,-0.0232584272,0.1560480297,0.4577078521,-0.3642401099,-0.5386852622,-0.0340450443,-0.1406246722,0.0022141691,0.1572956443,-0.1394826472,0.1909557134,0.0375712141,0.2512924075,-0.2371732593,-0.3950963616,0.510828197,-0.4227675498,-0.2113946974,0.4054193199,0.4281797111,0.5891724229,-0.0405182317,-0.2571867108,-0.0649866015,0.4788776338,0.4345226586,-0.221860677,0.1284220517,-0.106599547,0.0334945433,-0.189041391,-0.0248714462,-0.1043417007,-0.041730918,-0.4219146669,-0.4326018989]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/116","title":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323","comments":"Sure, [here is a Colab notebook](https:\/\/colab.research.google.com\/drive\/1uiS89fnHMG7HV_cYxp3r-_LqJQvNNKs9?usp=sharing) reproducing the error.\r\n\r\n> ArrowInvalid: Column 1 named references expected length 36 but got length 56","body":"I'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)","comment_length":22,"text":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\nI'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)\nSure, [here is a Colab notebook](https:\/\/colab.research.google.com\/drive\/1uiS89fnHMG7HV_cYxp3r-_LqJQvNNKs9?usp=sharing) reproducing the error.\r\n\r\n> ArrowInvalid: Column 1 named references expected length 36 but got length 56","embeddings":[-0.0459093675,0.1151693016,-0.0150177451,0.2721492946,0.2404057682,-0.2635479867,0.0431782566,0.2833093703,-0.4245963097,0.2897650599,-0.1165682301,0.6135146618,0.002955948,-0.4883929193,0.0284862295,-0.1368684173,-0.0081863794,0.2913643122,0.4083463848,0.1040178239,-0.2367764115,0.4247642457,-0.2803144157,0.2404751033,-0.2772250473,-0.0431662947,-0.140914157,0.0241196789,-0.3879240453,-0.4071284831,-0.0969800726,-0.129422754,0.1402386576,0.2845337987,-0.0001243008,-0.1726977825,0.2185771465,-0.0763504952,-0.0784560442,-0.355471164,0.0908430293,-0.4335229099,0.4626730978,-0.4217267931,0.1493797451,-0.444547832,-0.2770910561,0.2346983254,0.3155119419,0.4180500209,0.0594261922,0.0443487726,0.1355344653,-0.1406978667,0.3923070431,-0.3124579489,-0.0148894368,0.9393885136,0.4045062363,-0.123344034,0.0088452995,0.28397879,0.063688755,0.0859920904,0.3879432082,-0.0302904956,0.2522357702,0.2284366637,-0.0218096506,0.2644379735,-0.1019031778,-0.4539357126,-0.1760019958,-0.0415138938,0.21072267,-0.8138436675,-0.1296514571,0.2143556327,-0.0087865172,-0.0206669904,-0.1671189815,0.0948061123,-0.1624212414,0.2428711802,0.0433369242,0.1015653685,-0.1131134704,0.1020043194,0.1957405508,-0.0519757085,-0.1000677571,0.2911117673,-0.3222543597,0.0251701009,-0.2827189565,0.1632133722,-0.0535844453,0.227842316,0.151659444,-0.0624881648,0.4369321764,-0.1569254845,0.0929231942,0.1149883196,-0.2760958076,0.694979012,0.2233356237,0.184092924,-0.1416840255,0.1369490623,-0.0960068703,-0.2293881476,-0.1860226691,-0.480227083,0.0426827371,0.2551446557,-0.0416479856,-0.1482476741,-0.3896528184,-0.1070984676,-0.4597559571,-0.0251150914,0.1216923967,0.3067443669,-0.1807587892,0.1014436632,0.0864019021,0.3240639567,-0.1184371188,-0.1620375365,-0.0594235398,0.2187091708,-0.5450322032,0.24067837,0.1006991565,0.4013178945,0.2102111727,-0.1323679686,0.2888313532,-0.3561969399,0.0973299295,-0.2760156691,-0.1638509184,-0.2382118106,-0.0878214017,-0.1219403297,0.0132112503,-0.5898448825,-0.1037444472,0.1584692597,-0.3831334114,-0.1743612289,-0.1052634716,0.0049848054,-0.1634496152,0.12345732,0.027296681,0.221239686,0.2564776242,-0.4194496274,0.1893670857,-0.1130669788,-0.1574188918,-0.2067606896,0.1711912751,-0.0747364238,-0.3709995151,-0.0177110434,0.4077173769,0.5180056691,0.2838170826,0.5089685917,0.0264923573,0.3261210024,0.0735975057,-0.0801501572,0.6995396614,-0.6300969124,-0.2905769944,0.0334739424,-0.0851893649,-0.3690969646,0.2691858709,-0.2439285219,0.0441662557,-0.013352273,0.5227802992,-0.1308207512,0.0557931922,-0.2241363525,-0.3720876575,0.036407847,0.0222385153,-0.0950129405,-0.2133183926,-0.0826494023,-0.0538708083,-0.1896583587,0.3228634894,-0.1274841428,0.0176559612,0.3466716409,-0.0873065442,0.1222033277,0.2171980292,0.0220382046,-0.0281907972,-0.2014203072,-0.6301739216,-0.1126510277,-0.431871742,-0.3525094092,-0.3410968482,0.1011865586,0.3651586175,-0.001393555,-0.0714071915,0.0488335751,0.042067498,0.1768305451,0.0685995519,-0.2902733386,-0.2386568338,0.0704261884,-0.0020123774,-0.1128121763,-0.1153134629,-0.322853148,0.2446822822,0.5144621134,0.2985491455,0.2018042952,-0.0622772612,0.2856212556,0.0043485234,-0.0654836297,-0.0669207796,0.1077349484,0.2339478135,-0.3387617469,0.0726139173,-0.1530004889,0.2296396941,-0.1233316883,-0.0984923616,0.2231531292,-0.2790377736,0.2572297752,-0.1153825074,-0.2152660191,0.0184745844,0.0237298124,-0.27996099,-0.1046834961,0.2680798769,0.2856266797,0.1711651832,0.0939781219,-0.0549585819,-0.1399534792,0.5200073123,0.2920409739,-0.0245912131,0.0778038725,-0.1622531563,0.0071549374,0.1081201062,-0.4270858765,0.3236562908,0.32295838,-0.0640644655,-0.0057762745,-0.0237933304,-0.3719567657,-0.0604034699,0.1462898552,0.0762989148,0.3581416905,0.5070212483,-0.0947278589,-0.0556469858,-0.3841948509,0.1101764441,0.093338713,-0.2602670491,0.1308114827,-0.0515703894,-0.1690995693,-0.097034134,-0.6012462974,0.1804671288,-0.0802451968,-0.0359985828,0.0063303011,0.1556383371,0.2581205666,0.3968037665,0.1313301921,0.157966122,0.0120585654,-0.0660515577,-0.3850284517,-0.1293904781,-0.0779900327,-0.0796602517,0.03261251,0.048235666,-0.1015558913,0.0065349024,0.1033419743,-0.4134378731,0.0513205007,0.0072536194,0.1206091046,0.2042333186,-0.1892071664,-0.477563709,0.0355365649,0.4449720085,0.1436422914,-0.0621073171,0.1123714671,-0.3018159568,-0.0244500022,-0.0187459383,0.0003442197,0.0026015707,-0.5757029653,0.4874514937,0.1324093342,0.0313988514,0.2158802301,0.1193303168,0.139752686,0.1555221975,0.1511530876,0.1885483563,0.2938373685,0.3376951516,-0.2272808999,-0.3196132183,-0.0752933547,-0.1763201058,0.0593584999,-0.1004610062,-0.0242937524,-0.4497425556,0.0767455101,0.0626864806,-0.1489302665,-0.0185114648,0.293641597,-0.1113280877,-0.0409678742,-0.2013934404,-0.0951743871,0.2215320766,-0.1303657293,0.3529987037,-0.0387268104,0.2882715166,0.0644610152,0.2265393436,0.2385138273,-0.0950826108,0.1198192686,-0.2171099037,0.0581075288,-0.0009196081,-0.0051776296,-0.1468934864,0.0323146954,-0.0584757626,0.3397779167,0.0088871401,0.0543765947,0.0446124859,0.0703489259,-0.1768777221,0.0167973395,0.1315166056,-0.3431315124,0.1045738012,0.0665161833,-0.2047301829,-0.1528976411,-0.1957311183,-0.0199007019,-0.1526229978,-0.1908104867,0.0952243954,-0.082950294,-0.3003459275,-0.5376380682,0.4880761802,0.2718830705,0.03943168,0.0597293749,0.1059129015,0.1667040437,-0.1199396104,0.2913585901,0.1649096608,-0.139945671,0.3811170757,0.0911114216,-0.7312483191,0.0372770168,-0.1806705892,0.105029732,0.4936748147,0.1741348356,-0.3769854009,0.2013276666,0.2614145577,-0.0201314762,-0.0615326688,-0.1179474145,-0.3705139756,-0.2486398965,0.0301801413,0.0106780687,0.141023919,0.5643630028,0.3776751757,-0.0929028392,0.0685540661,-0.2486028522,0.0542128794,0.0503444411,0.3621605635,-0.0843937472,-0.2571892738,-0.211995393,0.1519052684,0.0750098303,0.1960412562,-0.145062238,-0.3731332421,0.0012289984,-0.1362247467,0.4701750278,0.2824201882,-0.1490755379,0.111876674,-0.4701266885,0.2755333781,0.2338600308,0.1581372023,0.2137172371,-0.0551842228,-0.120167017,-0.0093368366,0.3632326424,0.2750260532,-0.1148686484,0.3932282925,-0.1611337215,-0.1602805406,0.3675287664,0.0127028096,1.2714098692,0.0975431502,0.0793970451,0.3910428882,-0.1035361961,0.5623705983,-0.4354495406,0.1505535096,-0.3372240663,-0.0208264012,0.1981648207,-0.1487899572,-0.2604249418,0.054230772,-0.1902705282,0.1042467952,0.0511453189,-0.0547909625,0.062358465,0.1138234064,0.0617331713,-0.0636809394,-0.1832271069,-0.099617064,-0.0773857087,-0.2662287056,-0.2663745582,-0.187515974,-0.4289110005,-0.1219716221,-0.0295866318,-0.0172172002,-0.3545282185,0.1140032038,0.2464299947,-0.3073646128,-0.2578003407,0.4245165586,0.1749410629,0.0302233752,-0.1034406275,0.3559255004,-0.0102874348,-0.2247642428,0.0141628757,0.1017747372,-0.0789612681,-0.1271600574,-0.0435363464,0.1797403246,0.0908436328,-0.2109033465,0.0210864749,0.1918752044,0.1124206111,-0.0391528644,-0.0934672877,-0.1108878329,-0.0429537669,-0.1440030634,0.0203764066,0.1023555472,-0.2453034669,0.1471945941,0.1183391213,-0.4757426679,0.1392678022,0.1966353506,0.0275300462,0.0539885573,0.1919196844,0.114958398,0.0123342052,-0.041663859,-0.2283478677,-0.1194720641,-0.6108919978,0.2342528105,0.2643496692,0.268345058,0.1053332463,0.7349308133,0.4408649504,0.3101646006,0.1042270139,-0.4634284079,-0.1832597405,0.3580051363,-0.0033854714,0.0080309268,0.0990699306,0.2240065783,0.1526346356,0.4532804489,-0.1785029918,-0.1258416772,-0.2407491058,0.264547348,-0.005641181,-0.1428903192,-0.0223507006,0.2614609003,-0.0678479895,0.4470047951,-0.0577817708,-0.1261119843,-0.1382829845,0.1772471368,0.1550782919,0.1593081653,-0.1568727642,-0.0480731241,0.1710494906,-0.2919183969,-0.1099194661,0.0438834056,0.0842028335,0.0323538929,-0.0337266475,0.2420573682,-0.0670753419,0.1501717716,0.065464668,0.1689889431,-0.1195601597,0.0390871875,0.1156015769,0.0121837407,-0.0692270324,-0.1027964354,0.0157409944,-0.1640983373,0.0536844693,-0.0620916188,0.1002121046,-0.3933312297,0.2680687606,0.222346276,-0.1340806335,0.0152251748,-0.161917612,0.0142182959,-0.2476196289,0.1726635844,-0.1349610686,0.1790925264,-0.0944838151,0.1975019276,0.2050222009,-0.144759953,-0.0439680591,-0.0366867483,0.0615684763,0.0083008809,-0.4156885445,-0.099110432,0.0441412926,0.0513112061,0.5212087631,0.3527199328,0.3001064062,0.2329666466,-0.0012876226,0.1306740493,0.2468909472,0.1943112165,-0.1645900011,0.035485439,-0.3330209553,0.2299549133,0.291893512,-0.3531016707,-0.2284728438,0.4662558436,0.3584325612,-0.0544250757,0.2475519478,0.3130892515,0.067882061,-0.2025711685,-0.5934307575,0.0560357459,-0.2338709533,-0.2285305858,0.0355643816,0.3799592555,0.2097530365,0.167124182,0.0456029549,-0.322860837,-0.0597860925,0.3357740045,0.2310250998,-0.3742822409,0.4791252911,0.0166130159,0.1928634644,0.0226326007,0.3341029882,0.5191258192,0.3489403427,-0.01741964,-0.3148105443,-0.5039317608,0.010652977,0.0554369763,0.3430719972,0.6552648544,0.0480343886,0.2076675296,-0.0326169915,-0.1121252701,0.5894531012,-0.0735190734,-0.1071110666,0.0037505825,0.2051315308,-0.0056832265,-0.3008825481,-0.3026952147,-0.0356636494,-0.3989062011,0.1145052016,-0.1866627187,-0.1142666712,0.1240372062,-0.2243680358,-0.0001946484,0.0147111397,0.0164383072,0.4261868,0.4437749088,-0.3615905941,-0.4749903977,-0.2958472371,0.0605123341,-0.2756667733,-0.2944253981,0.0337536,0.2232059985,-0.0990083143,0.111352697,-0.1051389426,0.7662674189,-0.2335243821,0.2276912928,-0.4450758398,0.2036018521,-0.029354129,-0.1003778577,0.026040053,-0.0760598555,-0.0351051651,-0.3225160539,-0.1146166995,-0.1549395025,-0.249151364,0.1952333599,0.6734793782,0.6036807895,-0.0284575652,0.6332760453,0.1161303446,0.2928337157,0.1177119538,-0.1052152738,0.0930417404,0.2406363785,0.15859887,0.1079438627,-0.2546256781,0.0573633276,0.1524136961,0.5619996786,-0.2601129413,-0.4716579914,0.0167368371,-0.1678411812,-0.0719887093,0.1382744312,-0.1904106736,0.1949079931,0.0541488789,0.2557867169,-0.2786217332,-0.4741185009,0.5731935501,-0.4175247848,-0.2266870737,0.4212970436,0.4433149397,0.5793071389,-0.1704610288,-0.2849518061,-0.1100476086,0.4246287346,0.3773316145,-0.2665534019,0.1867272854,-0.0989618152,0.1017554477,-0.1822587401,-0.0169705208,-0.1173922643,-0.0213973522,-0.4595490098,-0.4233966172]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/116","title":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323","comments":"This is because `add` takes as input a batch of elements and you provided only one. I think we should have `add` for one prediction\/reference and `add_batch` for a batch of predictions\/references. This would make it more coherent with the way we use Arrow.\r\n\r\nLet me do this change","body":"I'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)","comment_length":49,"text":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\nI'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)\nThis is because `add` takes as input a batch of elements and you provided only one. I think we should have `add` for one prediction\/reference and `add_batch` for a batch of predictions\/references. This would make it more coherent with the way we use Arrow.\r\n\r\nLet me do this change","embeddings":[-0.0529837795,0.1234703138,-0.01561666,0.2224005014,0.2445526421,-0.2058066279,0.0348038748,0.2737767696,-0.2766898572,0.2284607589,-0.0470915511,0.5116131902,-0.0322659165,-0.5062631965,0.1793055087,-0.2050362378,0.0496256724,0.2373667061,0.3766706586,0.0053776717,-0.2605624497,0.3032194376,-0.1880754679,0.1640558392,-0.2273846716,-0.0277239792,-0.2081149369,-0.0000733137,-0.3736788034,-0.5136927962,-0.186548993,-0.1602295339,0.0929599479,0.2773670852,-0.0001235569,-0.1549440473,0.149187848,-0.0811496451,-0.0515882559,-0.2576403916,0.0698857605,-0.4877816439,0.2812328041,-0.3927665353,0.2623868883,-0.5535238981,-0.2693206966,0.120566301,0.3221309185,0.3828631043,0.0593123138,0.1814615875,0.1997821927,-0.1504540741,0.4799138904,-0.2872894108,0.027054552,0.8451249599,0.4745265245,-0.1043577269,0.0101642683,0.2115784734,0.0698365495,-0.008395087,0.4609832466,-0.0085368883,0.3220505416,0.2491591722,-0.1645759642,0.2763655186,-0.068574734,-0.45033741,-0.2426938415,-0.1303236336,0.2197299004,-0.8170561194,-0.2259736806,0.1794520617,-0.0250247847,-0.1299295276,-0.1462098211,-0.0006306816,-0.1770025939,0.3596486449,0.1693440825,-0.0147993406,-0.0118942177,0.1285102665,0.1847855151,-0.0393699519,-0.2734961808,0.2802458107,-0.3638677597,0.1397887319,-0.3178015053,0.1936397105,0.0414396375,0.0906878859,0.2250478268,-0.0171071608,0.4222173393,-0.13881585,0.0712493807,0.1222760528,-0.2013107091,0.6905140281,0.2165013701,0.1341722608,-0.3007668257,0.0094200037,0.0774845406,-0.2331835628,-0.1528362483,-0.4813553989,0.0929708183,0.2240232825,-0.0433338583,-0.0264970958,-0.3014723659,-0.0460714623,-0.3360155821,-0.1674297005,0.0867106766,0.3681385517,-0.0388477556,0.0624952391,0.1104905382,0.4761027694,-0.0439197794,-0.1354417205,-0.0921947658,0.2260674685,-0.5083327293,0.2136933655,0.0997615308,0.3832262158,0.1763621569,-0.073057346,0.2946143448,-0.3448928595,0.087014392,-0.1961714923,-0.0883425027,-0.2993743122,-0.0402428135,-0.0425539501,0.030005252,-0.4934621453,-0.0675721541,0.1587438285,-0.3589678109,-0.1973158568,-0.1149041355,0.0197511129,-0.0926990733,0.083531037,0.0243349411,0.2035319954,0.3013353646,-0.396302402,0.1792049855,-0.1033511907,-0.0777109861,-0.2220205814,0.1725850999,0.0104020853,-0.3534255326,-0.0505670086,0.4460203946,0.4284258485,0.219517529,0.5122495294,0.0170855112,0.4416695237,0.0511136688,-0.0548066832,0.8497806191,-0.6675328612,-0.2030715495,0.1031758562,-0.1517938375,-0.3849717081,0.244235605,-0.1403275728,0.0424472056,0.0017697939,0.5252922773,-0.1263539046,0.0111441053,-0.2413899004,-0.2935152948,0.1216426417,0.0770450234,-0.0832394958,-0.2776447237,-0.0424484909,-0.0174259041,-0.2651268244,0.315461129,-0.185546279,0.094031401,0.4362598062,-0.1067258939,0.1817872077,0.1525934935,-0.0213390924,-0.0370824598,-0.187252149,-0.7418581843,-0.1519334614,-0.3425426781,-0.3136158288,-0.3647011817,0.1320697665,0.3085981309,0.0722291172,-0.065481782,0.024303494,0.0326466747,0.0266355537,0.0122129973,-0.3676487803,-0.2076220512,0.0716270953,-0.1128770486,-0.1675752699,-0.1139455065,-0.3280251026,0.2027525008,0.5103162527,0.3298405707,0.1804865003,0.0027423869,0.3516777754,-0.1016335711,-0.0717662424,-0.1064506024,0.0596692003,0.1854149103,-0.2278652936,0.0522937067,-0.2333274633,0.2378190607,-0.2452935576,-0.1454570889,0.1176463068,-0.247784391,0.2879173756,-0.1668803096,-0.1577744037,0.0029124611,0.020116711,-0.2688775361,-0.1750835478,0.2672446668,0.2518250942,0.2535242438,-0.0183255989,0.023780413,-0.2061971873,0.4903864563,0.2489996552,0.0117215551,0.1085847393,-0.167279169,0.0327808559,-0.033713907,-0.458175391,0.2858690619,0.3511742353,-0.0713107288,-0.0535919406,-0.0323589295,-0.3229382336,-0.1007634327,0.1633617282,-0.0217885561,0.4516548514,0.4819218814,-0.0829909071,-0.0522987023,-0.2824621797,0.156576857,0.0922544301,-0.3211486638,0.1029351428,0.0322973467,-0.0499642231,-0.1294693947,-0.6580086946,0.1291842908,-0.1266931444,0.0177926142,-0.0792982504,0.1163699999,0.2354958206,0.3414893448,0.1112762541,0.1976393759,-0.0952695683,-0.0614368916,-0.3655001521,-0.0988830477,-0.0739974678,-0.0825133026,0.0339725949,0.1985394806,0.0282311291,0.0514330678,0.0575689711,-0.3675534427,-0.0267351288,-0.0360474586,0.103749834,0.1836600006,-0.1784429103,-0.3185208738,-0.0348143652,0.4909922183,0.1498887837,-0.0801961944,0.1976710111,-0.2899659276,-0.1214714423,-0.0277298242,-0.1028282419,0.0555745959,-0.6066274047,0.4826376438,0.2180045247,0.0196515191,0.1390763074,0.161531806,0.1337260157,0.1015784591,0.1502329856,0.188487187,0.2855364382,0.3695700467,-0.2453286499,-0.3526135981,-0.0996478871,-0.212758705,0.0100670941,-0.0160778034,0.0084737213,-0.3637005389,0.0287342574,0.1017535329,-0.1553876251,0.0992477313,0.3726509511,-0.1118815467,-0.0563923642,-0.2225678712,-0.0112460554,0.2974607646,-0.0985563323,0.338360697,-0.1027389243,0.3412115276,0.0712851882,0.3087331653,0.2600158751,-0.0898883343,0.2145285755,-0.2634678185,0.1250749379,0.0214319285,-0.0775774866,-0.1835475564,-0.0070341597,0.0651078746,0.1468191743,-0.0149853649,0.0971959308,0.0405044183,0.1509201378,-0.1699974686,0.0514253341,0.1844325513,-0.4027772248,0.1682156622,0.0393504053,-0.2191617936,-0.2926399112,-0.1878985167,-0.0037849485,-0.2331756055,-0.0678836927,0.0419052802,-0.3083762527,-0.2258057594,-0.5613568425,0.5143626332,0.2656938136,0.1224104241,0.0351366363,0.0294201188,0.1295226514,-0.1196372882,0.4070136547,0.0622383244,-0.1854687035,0.3979620337,0.253379494,-0.7920082211,0.0881141946,-0.242290467,0.1198909879,0.532846868,0.1347442269,-0.4058455825,0.2194062322,0.246075511,0.0621094517,-0.094024308,-0.133276239,-0.3536236286,-0.1603842229,-0.0559742525,0.0075835558,0.1301679313,0.5748971105,0.4040496647,-0.0979499966,0.1008284241,-0.1471957415,0.0985399112,-0.0460633002,0.3488911688,-0.0430460349,-0.2175797075,-0.2036944032,0.1121959388,0.0683082342,0.1961613744,-0.1751299798,-0.3364208937,-0.012685813,-0.1817173362,0.4985184968,0.2646092772,-0.0381798595,0.1476729214,-0.4133817554,0.2689514458,0.145388931,0.1187986359,0.3516078293,-0.114142634,-0.1327005923,-0.053757336,0.3737565279,0.2154425085,-0.103802748,0.4574333429,-0.1623682827,-0.139893353,0.3795520663,-0.0466680825,1.2245256901,0.1065460816,0.1257059276,0.4437693357,-0.073207356,0.5044654012,-0.4942473471,0.1356800497,-0.3452019989,0.052658435,0.1872795373,-0.2526928186,-0.1908869594,0.1537221968,-0.1415637732,0.1200093478,-0.0190732609,0.0004702446,0.0279732756,0.2141709775,0.0780109391,-0.1477119774,-0.06492614,-0.0891720653,0.0468232445,-0.2853758931,-0.2525690794,-0.1806266159,-0.3987855613,-0.1978798211,-0.0816202983,-0.0249775015,-0.1934519559,0.0791226476,0.2814867198,-0.2885399461,-0.2183078378,0.4451541901,0.1607865542,0.0353191905,-0.0677491054,0.3672395349,0.0054890737,-0.1944261789,0.0791476294,0.0584230237,-0.1377999485,-0.0912399441,-0.0723859146,0.3397003412,0.0489698239,-0.2909993529,-0.1121116877,0.2122796476,0.0954592079,-0.169405058,-0.1494031549,-0.0012263339,-0.0458358936,-0.159769848,0.0027230121,0.1189912334,-0.2679605782,0.1396853328,0.0382477604,-0.4058614075,0.1576608568,0.1739564091,0.0133692073,-0.0927389488,0.1960027069,0.142060712,0.040985588,-0.0514809974,-0.2509515882,0.0064283679,-0.6246702671,0.1300439388,0.3372043669,0.2366605848,0.0571026206,0.5329243541,0.3399436474,0.3049383759,0.1352116019,-0.4270899892,-0.2202938497,0.2987324297,0.1002271771,0.0221988037,0.0831799582,0.1857971996,0.1129416525,0.6304265857,-0.187393859,-0.1925971955,-0.1350554377,0.3047333658,0.0813753456,-0.117572479,0.0184587501,0.2227160335,-0.0670713931,0.3135066032,-0.0954195783,-0.1269567758,-0.1054324433,0.1703345925,0.1472406387,0.1388580203,-0.0746194795,-0.0503378287,0.2078300714,-0.2769307196,-0.0654866621,0.0310652014,-0.0053381654,0.0142431408,0.0108615588,0.2386567146,-0.0158688556,0.121025756,0.016570121,0.1443902701,-0.1419911385,0.0147411795,0.1334257573,-0.0300457571,-0.2427770942,-0.0422282927,0.1408337057,-0.158628121,0.072535947,-0.1146461889,0.0446701683,-0.3494180441,0.2841543853,0.2033738494,-0.203712225,0.1615557075,-0.2249836922,0.0506676286,-0.2578105927,0.0797745064,-0.1561123133,0.1253433377,-0.0308502764,0.1467737705,0.2249477655,-0.2415276021,-0.0554378629,-0.0418150499,0.0138770398,-0.0724344552,-0.4040735662,-0.0099067837,0.0720861554,-0.0764176697,0.4535537362,0.3923403025,0.287193805,0.229751274,-0.0046435352,0.1363933831,0.287887603,0.1203811914,-0.2763402164,0.0090057179,-0.3835601509,0.2529487014,0.331979394,-0.246913448,-0.1659370512,0.4449226558,0.4305506945,-0.0974193662,0.2416949123,0.3374014795,0.1081164777,-0.1779177785,-0.6260396838,0.0332708508,-0.2198946774,-0.2954455912,0.0575806834,0.2543746829,0.1096747294,0.1720699966,0.0758072659,-0.3366914093,0.0180775952,0.2719969451,0.1824454814,-0.4048291743,0.4678040743,0.101282239,0.179356277,0.1541534066,0.2735800445,0.550137639,0.3750266731,0.0532439314,-0.1164380088,-0.5201444626,-0.032899335,0.0351073109,0.3914085925,0.5328321457,0.0087496284,0.2254718393,-0.0226206202,-0.1150094643,0.4778467715,-0.092689693,-0.0166104697,-0.050837189,0.3062936068,-0.0524580777,-0.3750989437,-0.3568495512,-0.0904021785,-0.4538401365,0.068874523,-0.0712363794,-0.0935439691,0.1073963866,-0.2740226984,0.0096903937,0.0120879067,0.0106808245,0.4688245058,0.4138969481,-0.3861910701,-0.4042206407,-0.3711745441,0.0177390482,-0.3070272505,-0.2642475963,-0.0213969089,0.2950696945,-0.0288525838,0.110348992,-0.0030994129,0.7121225595,-0.315707922,0.2841114402,-0.4020727575,0.2923105359,-0.0185886454,0.0423711389,0.0393847898,-0.0799309835,-0.0461327061,-0.2612860799,-0.0898224264,-0.1520295739,-0.2171119452,0.1971558481,0.714482069,0.5996641517,-0.0183611866,0.6578068137,0.1130572185,0.2553167343,0.0129335793,-0.138968423,0.038907934,0.2253329158,0.1198836192,0.1031522602,-0.3069996834,-0.0306272302,0.1291933954,0.5416575074,-0.2737932801,-0.4963235557,-0.0378850475,-0.0884252191,-0.1001896337,0.2050063312,-0.0978763103,0.1428104937,0.1161091849,0.2808979452,-0.2200459242,-0.4489737451,0.5190483332,-0.3741150796,-0.2154400647,0.3690963387,0.4454098642,0.5610118508,-0.0898234472,-0.2587349415,-0.1160245463,0.4629229307,0.4017496407,-0.2320597768,0.1813247949,-0.1719188839,0.0393691994,-0.1715845764,-0.0857508555,-0.1080271229,-0.0240790304,-0.396851629,-0.433147788]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/116","title":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323","comments":"Thanks for noticing though. I was mainly used to do `.compute` directly ^^","body":"I'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)","comment_length":13,"text":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\nI'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)\nThanks for noticing though. I was mainly used to do `.compute` directly ^^","embeddings":[-0.0906596184,0.1422121823,-0.0103905555,0.2536019087,0.2622912526,-0.2561780512,0.0166131314,0.3302355111,-0.3349124491,0.2990702391,-0.044984445,0.5559690595,0.0599088855,-0.5039998293,0.1049175635,-0.1441465169,-0.01154791,0.2800841928,0.390884012,0.0705950409,-0.2649939954,0.360801667,-0.1966688186,0.2647157609,-0.1890148371,-0.0929352045,-0.1321041882,0.0510196649,-0.3382510543,-0.434096992,-0.1080586314,-0.1712291092,0.0882133469,0.2875922918,-0.0001243917,-0.1350644529,0.2345233262,-0.085697569,-0.0255331509,-0.2422120571,0.1949889362,-0.4378356338,0.4129045904,-0.4462184906,0.1987752169,-0.5338203907,-0.3209138513,0.1098099947,0.3061314523,0.3397934735,0.0488059409,0.1721343249,0.1662994325,-0.16676642,0.390850395,-0.3572039902,0.0793499053,0.8499205112,0.3666140735,-0.1302367747,-0.0155977588,0.1937702149,0.0767655149,0.0389419608,0.4126162529,-0.0458381772,0.3380200863,0.2968806028,-0.0899528265,0.3095541894,-0.1001740769,-0.4494247437,-0.234443903,0.0043439493,0.1509010941,-0.8078969717,-0.1113685668,0.1858337373,-0.0433407612,-0.123537451,-0.1658288985,0.0399771854,-0.1853912622,0.2673507631,0.0205618106,0.0012171593,-0.0955046117,0.1550161391,0.2211948931,-0.0541516542,-0.2171892077,0.3026699126,-0.2954831421,0.0934066102,-0.3498264849,0.1850242466,-0.0012912217,0.0857894197,0.1933615357,-0.0013980606,0.4760616422,-0.0619981848,0.1196958497,0.1364840716,-0.2325964272,0.7211273313,0.1757261008,0.1624204367,-0.1809781045,0.0843411908,-0.0217731968,-0.2392669767,-0.1902516484,-0.5169257522,0.0920119509,0.2457526624,-0.1187002659,-0.1383070201,-0.3434146345,-0.0839941204,-0.4101729393,-0.0778795034,0.0958394036,0.3080675602,-0.0714189932,0.1173966751,0.1073470712,0.368372947,-0.0992584303,-0.2277431488,-0.0907088742,0.2464898825,-0.4702476561,0.2611674964,0.1059781089,0.3061449826,0.215099141,-0.1141523868,0.3205482066,-0.4079195261,0.1416653842,-0.2163691074,-0.1525625885,-0.2626957297,-0.109656319,-0.1516564041,0.0624704175,-0.5091994405,-0.1182116196,0.1432695985,-0.329785645,-0.2191476822,-0.0391146541,0.0119432285,-0.1649461538,0.123931244,0.0659466311,0.239244014,0.285899967,-0.4008102119,0.1382102221,-0.0924829841,-0.0451243594,-0.2087444961,0.2582022548,-0.1161596999,-0.3996031582,0.0171792507,0.3765339553,0.4172064662,0.2147257775,0.5687553287,-0.0193901192,0.3946840465,0.061961744,-0.0765030533,0.7636829615,-0.6570129991,-0.2176803201,0.0625430867,-0.102401495,-0.4136026204,0.2137619406,-0.2277290374,-0.0006546963,0.0154079599,0.4701844454,-0.1692610681,0.0622791871,-0.2835194767,-0.3280038238,0.0500122085,0.1121104136,-0.0465356484,-0.2383959889,-0.0389828756,-0.0592503399,-0.277638495,0.2425759733,-0.1595663577,0.0512853749,0.3760402203,-0.2096524239,0.1330359578,0.2566767037,-0.0637749657,-0.0259688087,-0.1519137472,-0.7368127108,-0.1392633915,-0.3950161934,-0.291702956,-0.3110801578,0.1171491444,0.3529608548,0.0612578467,-0.0756436512,-0.0120892767,0.149772048,0.0977028981,0.0078505296,-0.3995055854,-0.2914209068,0.0447440334,-0.049994491,-0.1572068483,-0.0637762174,-0.3110693395,0.2403077483,0.4711670876,0.318900764,0.2200011164,-0.0659950599,0.3223726451,-0.0657327101,0.0264109317,-0.020041205,0.0921887159,0.155339554,-0.384108603,0.1057649031,-0.1224541888,0.2476519197,-0.1561136991,-0.1459343135,0.1726836115,-0.2644754648,0.3023426831,-0.0906106904,-0.1807369441,-0.0004914169,0.0604642145,-0.2892223299,-0.0864667967,0.2793085277,0.3279035985,0.2238644511,-0.0091120144,0.0011809525,-0.2657687366,0.5572475791,0.2123651654,-0.0214049835,0.1013840809,-0.1441063583,0.0190348756,-0.0290780105,-0.5026097298,0.2240033299,0.3246925771,-0.0075848321,-0.0022693945,-0.0452853255,-0.3263436258,-0.0493302494,0.1756205708,0.0270397682,0.4072049558,0.4612495601,-0.137380898,-0.0754566193,-0.3109734356,0.1637173444,0.0812421441,-0.2918257713,0.1251361221,-0.0318665393,-0.0200701933,-0.0800893009,-0.5301100016,0.0896300599,-0.1132786348,-0.0310390536,-0.1002783775,0.0969469249,0.2710415423,0.3880288303,0.1355627328,0.1951618493,-0.0587256625,-0.0606985912,-0.416552484,-0.1513792872,-0.1104125902,-0.1171100885,0.0353574492,0.1153844595,-0.0560144223,0.0227686651,0.0102811987,-0.4021193385,-0.0375810973,0.0291017257,0.111422196,0.2343678623,-0.1756646186,-0.3528187871,0.1194427982,0.5194316506,0.1146398708,-0.0884103179,0.1702069491,-0.2477121651,-0.0187275205,-0.0207083821,-0.1064321548,0.0524124093,-0.5989528298,0.5768093467,0.1302159131,0.0342234969,0.079991132,0.0870131478,0.119130969,0.1310640424,0.1049044058,0.1405299157,0.2476617694,0.3633092642,-0.1933511347,-0.3571261764,-0.1192596629,-0.1390074939,0.0610608086,0.0228652339,0.0074941013,-0.2747006416,0.0511338226,0.107269071,-0.2150589973,0.027754318,0.3566780984,-0.1309958994,-0.0372505747,-0.2433843911,-0.0612253062,0.280256182,-0.1444690824,0.2004085481,0.0158728641,0.3328732848,0.0848826766,0.2663414478,0.2492568642,-0.0711355656,0.1290451586,-0.2435156554,0.0370000266,0.0372804888,0.024524074,-0.1966536492,0.0001138412,-0.005698937,0.2327747941,-0.0094152922,0.0608531274,0.0437735133,0.0631193295,-0.1445583105,0.0587171726,0.1424943656,-0.4261609614,0.2212381214,0.0913032591,-0.1779242903,-0.2431979179,-0.1389169991,0.0120888883,-0.254693538,-0.1019973606,0.0611860938,-0.2649870813,-0.2613135278,-0.4981923997,0.5372238159,0.2452147901,0.092982538,0.0451361798,0.0893483087,0.1172697917,-0.1792796105,0.3844201565,0.0831295028,-0.0849517509,0.4586258531,0.1032738015,-0.8232374191,0.0791621134,-0.2199029177,0.0949417576,0.4622676969,0.2210554779,-0.4160294235,0.1585108638,0.1887862384,0.0656961724,-0.0545334965,-0.1737246215,-0.3569804132,-0.1934367716,0.0460676253,0.0420493037,0.1082852259,0.5454239249,0.329297781,-0.102440685,0.0702801645,-0.1125568375,0.0348427184,-0.0325998776,0.3685538173,-0.2099765241,-0.2346964031,-0.0976840556,0.1183421761,0.0110854553,0.2063160539,-0.0858700201,-0.3682450056,0.0331177637,-0.0762999579,0.4623513222,0.2323903143,-0.1487327218,0.1035095081,-0.3720928431,0.3006772399,0.1049685851,0.037560083,0.2588431835,-0.0492205545,-0.0604110472,-0.0453449637,0.4652400911,0.3283904195,-0.0795536488,0.438459605,-0.1424498409,-0.1445973217,0.2653193772,0.0336370096,1.2064902782,0.0601321273,0.0575764515,0.4308720529,-0.0686127543,0.6177080274,-0.57915169,0.1375730634,-0.352912575,0.0332840532,0.1709459722,-0.2172376066,-0.1819012761,0.1677140445,-0.1448425651,0.1737529933,-0.0408771746,-0.054301016,0.1389935017,0.2117535919,0.0314154588,-0.1338758916,-0.0990733802,-0.1033709943,-0.0280522909,-0.2647143304,-0.2504669726,-0.152474016,-0.3848483562,-0.1502784044,-0.0574610159,-0.0152388467,-0.2412904501,0.1570369154,0.1960908026,-0.203182444,-0.196855098,0.4225140214,0.1934716403,0.0161526222,-0.1127168685,0.3685764372,-0.0390804522,-0.2096883208,0.1137988716,0.1892215312,-0.1506533325,-0.1509993374,-0.1287656575,0.2206357867,-0.0355890915,-0.2032233179,-0.0656743124,0.1429569274,0.0689901561,-0.0633752719,-0.0461910293,-0.1312114149,-0.0722107738,-0.140075773,0.0083656423,0.1477631778,-0.2989741266,0.1216964498,0.0986803249,-0.4573228657,0.2114047259,0.1938071996,0.0244867653,0.012205122,0.2126618326,0.1479967684,0.0449252129,-0.0767881721,-0.2602609694,-0.0570234433,-0.6208195686,0.2204971611,0.2242001891,0.2779609859,0.0258955918,0.642154336,0.4051526785,0.2733242214,0.0865967348,-0.4662806094,-0.1956146657,0.3406147063,0.0362966359,-0.0262909736,0.0808339342,0.195468381,0.0654312745,0.4910428226,-0.168165803,-0.0746752843,-0.2382239252,0.2624144256,0.0084920712,-0.1134613231,0.0269227438,0.2875312567,-0.0678241849,0.4470340014,-0.1078181118,-0.1091954634,-0.0936487243,0.1929898709,0.0629462674,0.1591702104,-0.0323777534,-0.021006899,0.1315713227,-0.2826581597,-0.1193824559,0.017780168,0.0354059711,0.0648795292,-0.0585970171,0.2642292976,-0.1283651888,0.1201598719,0.0986300334,0.1446028501,-0.2336951196,0.086223267,0.1001207978,-0.0039781053,-0.2087833434,-0.0680337325,0.0618511289,-0.1868331432,0.1133186668,-0.0600158684,0.0772932172,-0.439858526,0.329559058,0.3231623471,-0.1999489963,0.0856817961,-0.2296781987,0.0258382093,-0.246127978,0.1423899084,-0.1253623217,0.1261954457,-0.0915312544,0.1415339112,0.2410854846,-0.19500193,-0.0508812033,-0.092345804,0.0907861739,-0.1378851086,-0.3848363161,-0.0635353178,0.0628953129,0.0172776133,0.4942233562,0.4071734548,0.3115758896,0.3141024709,0.0146875093,0.0631050169,0.2535632253,0.2005936652,-0.2814982235,0.0281980541,-0.3666320741,0.2916855812,0.300406605,-0.3402061164,-0.2947749496,0.4836648405,0.3845360279,-0.0546481386,0.2044994086,0.3319914937,0.1290189177,-0.2017076164,-0.6036969423,0.0043455362,-0.1726204902,-0.2581816912,0.076270327,0.371658206,0.1795210689,0.1708245277,0.0917857289,-0.3700613379,0.0064180326,0.300133884,0.2314876616,-0.3948168755,0.4765639007,0.1578555554,0.15734303,0.0872859731,0.3048834503,0.5455594063,0.3309475183,-0.0569597371,-0.1884077638,-0.5316607356,-0.0099213338,0.0377742164,0.3701798916,0.5824454427,0.0364216976,0.289940536,-0.0473558009,-0.1020723656,0.403101176,-0.0890262499,0.0613815337,-0.1002253518,0.291169107,-0.0685193464,-0.2958585024,-0.2867617905,-0.0906367451,-0.3997000456,0.0877996758,-0.151632756,-0.0144685805,0.0968434513,-0.2008751929,0.0082096281,-0.0274387598,0.0564026386,0.4386729002,0.4052660763,-0.3591628671,-0.4522491693,-0.3574154973,0.0627477542,-0.2676723897,-0.3265411258,0.1124564782,0.3182986081,-0.0456976146,0.0977995247,-0.0271958001,0.6616536379,-0.2611368001,0.2136194706,-0.4128443003,0.1964006722,-0.0901668146,-0.048540052,0.0189624764,-0.0928407013,-0.0556784533,-0.2644849718,-0.1096026823,-0.1816672981,-0.1713114977,0.2581601143,0.7088486552,0.6046629548,-0.0040197233,0.665968895,0.1805022806,0.2365671247,0.0756088123,-0.0859422907,0.0749578327,0.2927341461,0.0435742959,0.069861494,-0.3089294136,0.0026523496,0.0698300526,0.5351088643,-0.3081310391,-0.4730901122,-0.0283882115,-0.1662610918,-0.0043858928,0.140949443,-0.0807819515,0.2009618282,0.087770462,0.2574394941,-0.2930470705,-0.3751836121,0.5320130587,-0.3674864471,-0.3083664179,0.4433623254,0.457426846,0.6294953227,-0.1077038348,-0.2725806236,-0.0816680863,0.5271197557,0.3998463154,-0.2235261947,0.140517652,-0.1722929031,0.0659892634,-0.1518850029,-0.0940876901,-0.1480116844,-0.0265990943,-0.4083969295,-0.4235137701]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/116","title":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323","comments":"Thanks @lhoestq it works :)","body":"I'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)","comment_length":5,"text":"\ud83d\udc1b Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\nI'm trying to use rouge metric.\r\n\r\nI have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. \r\nI tried :\r\n\r\n```python\r\nimport nlp\r\n\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"test.pred.tokenized\") as p, open(\"test.gold.tokenized\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\n```\r\n\r\nBut I meet following error :\r\n\r\n> pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n\r\n---\r\n\r\nFull stack-trace :\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 3, in <module>\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/metric.py\", line 224, in add\r\n self.writer.write_batch(batch)\r\n File \"\/home\/me\/.venv\/transformers\/lib\/python3.6\/site-packages\/nlp\/arrow_writer.py\", line 148, in write_batch\r\n pa_table: pa.Table = pa.Table.from_pydict(batch_examples, schema=self._schema)\r\n File \"pyarrow\/table.pxi\", line 1550, in pyarrow.lib.Table.from_pydict\r\n File \"pyarrow\/table.pxi\", line 1503, in pyarrow.lib.Table.from_arrays\r\n File \"pyarrow\/public-api.pxi\", line 390, in pyarrow.lib.pyarrow_wrap_table\r\n File \"pyarrow\/error.pxi\", line 85, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323\r\n```\r\n\r\n(`nlp` installed from source)\nThanks @lhoestq it works :)","embeddings":[-0.0292924568,0.0778478906,-0.0154838525,0.254104346,0.2674146593,-0.2241524607,0.0466230549,0.3121848702,-0.3159381449,0.2265444249,-0.0514323264,0.5397802591,-0.0169005431,-0.5313331485,0.0781625137,-0.2136966437,0.0339472704,0.2652906477,0.4080455303,0.066221267,-0.2505090237,0.3747792244,-0.2374881506,0.282266438,-0.193803221,-0.0091253566,-0.143543452,0.019681992,-0.3356235623,-0.4658494592,-0.1323114485,-0.1641162932,0.120737806,0.2650203109,-0.0001259799,-0.1457512379,0.2013971657,-0.0667272061,0.0065367306,-0.3372959495,0.1047564298,-0.4477454126,0.3895814121,-0.4135539532,0.1784151942,-0.5855218768,-0.3212720752,0.1467482597,0.3320436478,0.3892081976,0.0403764546,0.0915543362,0.1753196865,-0.1588526964,0.3869362772,-0.3302285075,0.0618621074,0.874548018,0.4089440405,-0.0961387753,0.0001163339,0.2396064252,0.073540315,0.0413305536,0.4391814172,-0.0547365174,0.3036572039,0.2972774506,-0.0955765098,0.296558708,-0.0635294467,-0.464019686,-0.2184080929,-0.0609223992,0.1765401661,-0.810436666,-0.1475406587,0.2100195289,0.0051871021,-0.0433178991,-0.1868558377,0.0190736987,-0.2072926313,0.2889181972,0.0675477535,-0.0102802301,-0.0482140109,0.1478859782,0.1862677634,-0.0782043934,-0.2577134073,0.3290608823,-0.3406787217,0.0651191995,-0.358174473,0.1514505595,-0.0089338925,0.1321983039,0.2322725952,0.0161160622,0.388710171,-0.1369689405,0.109251745,0.1006389111,-0.1886223406,0.6885242462,0.2326201051,0.1947225183,-0.2331375033,0.0410981253,-0.0243895724,-0.251332432,-0.1929354966,-0.4929925799,0.1228334457,0.2154284567,-0.0800875574,-0.1368253231,-0.3080562949,-0.0586186312,-0.4048902392,-0.0937563702,0.0851602554,0.3502746522,-0.0714199021,0.1442349553,0.0870931223,0.4140856862,-0.0451203175,-0.1978018582,-0.0568912849,0.2037269175,-0.5134835839,0.2221945077,0.0903806016,0.3945741951,0.1903140992,-0.1031709909,0.2862511277,-0.3878107071,0.1184027642,-0.2132984251,-0.1498909146,-0.2816998661,-0.0958734602,-0.1074193493,0.0304209758,-0.4725669026,-0.1027102917,0.1594968289,-0.3870345354,-0.2038678229,-0.1080172583,-0.0026938561,-0.1242676154,0.0882943273,0.090204969,0.2241629064,0.246260494,-0.4227000773,0.1833693087,-0.0688054338,-0.0553320535,-0.1820823699,0.1700290889,-0.0864185393,-0.3446798325,-0.0561261997,0.4421240687,0.4455679953,0.2046446502,0.5284485817,0.0127697978,0.3639628887,0.061596144,-0.0877823085,0.823939383,-0.6951028705,-0.2112309486,0.1092879847,-0.0970165506,-0.4027778804,0.2707204819,-0.2427999824,0.0385673493,0.0253333226,0.4515143633,-0.156599775,0.0538574047,-0.2434365451,-0.3281935453,0.054219678,0.080800429,-0.052230183,-0.2551908195,-0.0521686301,-0.0397430956,-0.2123774588,0.330311656,-0.1104789674,0.0644507408,0.4013207853,-0.1780180484,0.1991139203,0.2389692515,-0.0399001837,-0.0045625078,-0.2118433267,-0.7099912763,-0.1861295849,-0.366826117,-0.3042334318,-0.3329870999,0.0932805538,0.3661900163,0.0444841534,-0.0964851752,-0.0095256409,0.0673323274,0.1280117929,-0.0011316882,-0.3982480168,-0.2359134257,0.078887783,-0.0154603794,-0.1542717367,-0.0686057359,-0.2690803409,0.2205531299,0.5052906871,0.3314108253,0.2279728502,-0.0060774717,0.3458492756,-0.059216287,-0.0163020436,-0.0709033459,0.1059245244,0.1899036169,-0.387444526,0.1122422218,-0.181903258,0.1865585297,-0.1549761891,-0.1510400027,0.1727300584,-0.2395796627,0.2994927466,-0.1384748071,-0.1861088276,-0.0021722668,0.0026852814,-0.3084315062,-0.1465221345,0.2940438688,0.2257282138,0.1935143769,0.0439133979,-0.0546735451,-0.2491929829,0.5238227248,0.2514137328,-0.0155062601,0.1191446781,-0.1646430343,0.0408029296,-0.0251149312,-0.484457314,0.2842749357,0.3542759418,-0.076383844,-0.031838432,-0.018908469,-0.3372303545,-0.0451315269,0.1521672457,0.0075176014,0.4300640821,0.4664379358,-0.1224052459,-0.0428383164,-0.3373470008,0.1287602335,0.0606873371,-0.3019217849,0.0879833475,-0.046230752,-0.0734029561,-0.1329786628,-0.5823301673,0.1401795447,-0.0826310813,-0.0326405726,-0.0881361067,0.1429036409,0.267441988,0.3176137507,0.1122043431,0.1750696152,-0.0146674328,-0.072006382,-0.3813501894,-0.0806277767,-0.1032643095,-0.1339068264,0.0002133483,0.1247572973,-0.0342989042,0.0001795834,0.0376785062,-0.375338763,-0.0020804396,0.0137010906,0.0492019542,0.2253049463,-0.1395380795,-0.3353738487,0.1058215946,0.5181372762,0.1406052262,-0.0563031845,0.2109601051,-0.2734842896,0.0028405942,0.0029368913,-0.0717532933,0.0511500724,-0.5843036771,0.5186074972,0.139600426,0.0041125328,0.1258280873,0.1110185906,0.0830518752,0.07102938,0.1489839256,0.1952774823,0.2602818608,0.3470641971,-0.2158378959,-0.3271070719,-0.1099951491,-0.1038206294,0.0156000499,-0.0824076086,0.0034084283,-0.3410354257,0.0692094862,0.0582535602,-0.1910032481,-0.0005929196,0.312852025,-0.1239128709,-0.0313827805,-0.2368501425,-0.0379435197,0.2942646444,-0.1623771638,0.3167879283,0.0134380208,0.3597766161,0.0707467869,0.1811981648,0.2877449691,-0.0580083877,0.1430675983,-0.2589712143,0.0561520755,0.0282560494,0.0074512968,-0.193608135,0.0413655415,0.0275329128,0.2393076569,0.0103808334,0.0836343616,0.0428053103,0.0724281296,-0.1461106986,0.0844372138,0.1279774755,-0.438637495,0.1483950615,0.0452418476,-0.2213304341,-0.189771235,-0.1733294129,0.0279369671,-0.2145543993,-0.1060647294,0.100534685,-0.243789345,-0.2426598072,-0.4910633564,0.4856000841,0.216401428,0.0950118601,0.0390346423,0.0775394738,0.1392115057,-0.1179732755,0.3810396791,0.135258615,-0.0930913612,0.3560099006,0.1601720005,-0.7960019708,0.0607758276,-0.1688161641,0.0655291155,0.4415141642,0.1850249469,-0.395292908,0.1715342999,0.2288758755,0.0062116883,-0.07090763,-0.172578156,-0.3400179744,-0.1425941885,0.0025489356,-0.0104702525,0.1103178933,0.5665931106,0.3636984825,-0.0950993001,0.0908160061,-0.2104704678,0.0665022656,-0.0473516174,0.4089695811,-0.2019826472,-0.2384409606,-0.1159174293,0.1772384942,0.0382960588,0.197682023,-0.1087793559,-0.3640206754,0.0066850656,-0.1322062463,0.5071527958,0.2320847064,-0.1299999207,0.1483830214,-0.4241616428,0.2957583964,0.130834356,0.0255306959,0.310670495,-0.0725185052,-0.133388862,-0.0244497899,0.3947096169,0.2625080645,-0.101518102,0.4479624033,-0.1138631552,-0.1540249437,0.3414285183,-0.0317720063,1.2724379301,0.1150318086,0.1047347486,0.3783414066,-0.1227607578,0.538675189,-0.51753968,0.0949997827,-0.3517797887,0.0531704836,0.172499761,-0.1941200048,-0.175351873,0.1817398965,-0.1517702341,0.1172143593,-0.0282105487,-0.0323212855,0.0336226374,0.1612071842,0.0790720358,-0.0662612915,-0.1250317097,-0.1172011867,0.0001704977,-0.3228432834,-0.237760365,-0.1821803153,-0.387840718,-0.1715877652,-0.0811145753,-0.0481274202,-0.2105592936,0.12802881,0.2723637521,-0.2395343482,-0.2129468769,0.4673573375,0.1461892128,0.0103779072,-0.1248033345,0.4157861173,-0.0275834203,-0.2139935195,0.0644100159,0.1593516618,-0.1727981269,-0.1291632503,-0.1242754534,0.2703377903,0.0365476571,-0.228237927,-0.0938684195,0.1594550014,0.0893634856,-0.0990184769,-0.1270142943,-0.1317063868,-0.0328008682,-0.1198416874,0.0021237549,0.1472036242,-0.2354814559,0.1064143702,0.1036282405,-0.4642044008,0.1555896103,0.2080752403,0.0001373551,-0.0131971994,0.2046350688,0.1273053885,0.0496315621,-0.0627417788,-0.1838468164,-0.0330398008,-0.638315022,0.1665601432,0.2977756858,0.2802018523,0.0891406089,0.5774689913,0.3999930918,0.2989989221,0.1424770504,-0.4933888912,-0.1508901417,0.3546253443,-0.0130647123,0.0091130603,0.0337902308,0.1643492132,0.1166773215,0.4888832569,-0.1605895758,-0.1223532706,-0.1859466136,0.2449905872,0.0425045341,-0.0951822102,-0.0044355225,0.2293655276,-0.0717439279,0.3723808229,-0.0499518588,-0.0996090323,-0.1328769028,0.1912985593,0.1339643151,0.1818137318,-0.1078379825,-0.0483125336,0.1725717187,-0.2489381582,-0.1258244514,0.0641177148,0.0702760369,0.1012232229,-0.0351100564,0.1855639815,-0.1103793681,0.10519097,0.0878161564,0.1529423594,-0.183643043,0.020333942,0.1114198193,0.013248587,-0.224165827,-0.0480295494,0.0936000794,-0.1897127181,0.1136535928,-0.041636128,0.1315306574,-0.3501258194,0.3317594528,0.2407617271,-0.2207341641,0.0941103175,-0.2276819795,0.0259870365,-0.2533003688,0.1261125654,-0.1612351388,0.1547038406,-0.1588453203,0.118904531,0.2067670077,-0.218967095,-0.0581536219,-0.0520948879,0.0953921825,-0.0777764469,-0.4098263085,-0.0312836319,0.0717411265,-0.037581414,0.4741306901,0.4270884097,0.3271844983,0.3053548932,0.0011712393,0.0817466453,0.2590588629,0.185540691,-0.2418730557,0.0476387404,-0.3908937871,0.2711747587,0.2728778422,-0.3510964513,-0.2784826458,0.4859693348,0.4540313184,-0.1083970293,0.2535942197,0.3208224773,0.1278632432,-0.2296557724,-0.6354447603,0.0280656852,-0.2173542976,-0.2790626884,0.0513021797,0.3433256149,0.1757888347,0.1834556758,0.0946277902,-0.4060703516,-0.0147506287,0.2744940221,0.2232565284,-0.4054993093,0.4787207544,0.1238620803,0.1607536077,0.1900290251,0.2858456969,0.5313690901,0.3307036757,0.0231536217,-0.2041836828,-0.5608857274,-0.0215803068,0.099269338,0.3792313337,0.6029282212,-0.0041969442,0.2102840394,-0.0472308956,-0.0933258235,0.5234521031,-0.0790946633,0.0118207242,-0.0626694635,0.3058934212,-0.0446391925,-0.3172926009,-0.3033355474,-0.0784777999,-0.4084243476,0.0529432632,-0.1144759953,-0.0235252101,0.1011941135,-0.2464455217,-0.0115388324,0.0213379916,-0.0056052813,0.418076843,0.401396215,-0.324363023,-0.4508088231,-0.3072156012,0.0317121223,-0.2721164823,-0.2910737097,0.1040478498,0.2535765767,-0.0561595969,0.0935481116,-0.0324513689,0.7720280886,-0.303940028,0.2827268541,-0.4087822139,0.2374813855,-0.0314168818,0.0584597848,0.0080185579,-0.0437475853,0.001164223,-0.3094387949,-0.1332157999,-0.148350656,-0.1608166546,0.2190837711,0.7097871304,0.5899258852,-0.0393648259,0.6841077209,0.168714121,0.2832725644,0.0700654015,-0.114995867,0.0623367466,0.2244765759,0.0871995911,0.0604098365,-0.3101277351,-0.0351035967,0.1546854377,0.5102920532,-0.3530781567,-0.5100502372,-0.0357801728,-0.161401853,-0.0047473637,0.1477394998,-0.1386429369,0.1588378549,0.10222736,0.2611150444,-0.2499983311,-0.4334172904,0.5402110219,-0.3780837059,-0.2396193743,0.4585520923,0.4540185332,0.5672485232,-0.1053858176,-0.3216618299,-0.090591155,0.4940842986,0.4393001497,-0.2174186707,0.1314985901,-0.1156747341,0.071582593,-0.1601027697,-0.0470308177,-0.1173348054,-0.0029799147,-0.4168791771,-0.3838069141]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/115","title":"AttributeError: 'dict' object has no attribute 'info'","comments":"I could access the info by first accessing the different splits :\r\n\r\n```python\r\nimport nlp\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nprint(cnn_dm['train'].info)\r\n```\r\n\r\nInformation seems to be duplicated between the subsets :\r\n\r\n```python\r\nprint(cnn_dm[\"train\"].info == cnn_dm[\"test\"].info == cnn_dm[\"validation\"].info)\r\n# True\r\n```\r\n\r\nIs it expected ?","body":"I'm trying to access the information of CNN\/DM dataset :\r\n\r\n```python\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nprint(cnn_dm.info)\r\n```\r\n\r\nreturns :\r\n\r\n> AttributeError: 'dict' object has no attribute 'info'","comment_length":42,"text":"AttributeError: 'dict' object has no attribute 'info'\nI'm trying to access the information of CNN\/DM dataset :\r\n\r\n```python\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nprint(cnn_dm.info)\r\n```\r\n\r\nreturns :\r\n\r\n> AttributeError: 'dict' object has no attribute 'info'\nI could access the info by first accessing the different splits :\r\n\r\n```python\r\nimport nlp\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nprint(cnn_dm['train'].info)\r\n```\r\n\r\nInformation seems to be duplicated between the subsets :\r\n\r\n```python\r\nprint(cnn_dm[\"train\"].info == cnn_dm[\"test\"].info == cnn_dm[\"validation\"].info)\r\n# True\r\n```\r\n\r\nIs it expected ?","embeddings":[0.1249497235,-0.4288868904,-0.0755954757,0.6906054616,0.0929051787,0.0969590843,0.269435823,0.0933502242,-0.0262442082,0.3825279176,-0.1993025392,0.2700487673,-0.0093511529,0.1719216257,-0.0433850512,-0.210740611,-0.1371984631,0.1066760346,0.0732271969,-0.2672348619,-0.2168749422,0.0030046206,-0.2907536328,0.0620870106,-0.243092075,-0.0518735573,-0.1551669389,0.2124351561,0.1969709694,-0.3359895647,0.068953611,-0.2129002362,-0.0541770272,0.1591783911,-0.0001246873,0.2159624696,0.3259490728,0.0212656278,-0.4290561676,-0.2216002643,-0.3981970549,-0.3013712764,0.3539854288,-0.4983978271,-0.0128480112,-0.2380619645,0.1509514451,-0.2423049957,0.4023640454,0.0351480581,0.0804313198,0.0912048742,-0.0358201005,0.1423137188,0.0212015919,0.2588219941,0.2370947301,0.1636524647,-0.2527903616,0.0627270564,0.1914429963,0.425129056,0.0440268442,0.3171995282,-0.1918286681,0.5741863251,0.1710591316,-0.2567908466,0.0959282145,0.2467696667,0.1706724763,-0.1224946007,-0.4816110134,-0.3179220855,0.1392760724,-0.2829391956,-0.0797850713,0.2207458317,-0.0475859083,0.1868588924,-0.2426729202,-0.0948003083,-0.2555106282,0.34411937,-0.3710780144,0.2993047535,-0.1208858192,0.2345955819,-0.1352393031,-0.0211658757,0.059795592,-0.0942324921,-0.0770025775,-0.0593116246,0.0755457431,-0.1603479981,-0.01686172,-0.3014565706,-0.3027259111,0.2690218389,0.0478206091,-0.1880151033,-0.0445899889,0.0476634391,0.4869711697,-0.000109636,-0.0382666774,0.6492273211,0.1455626488,-0.0843479037,-0.3494222462,-0.1001153514,-0.0615426525,0.3459921181,0.0822181925,0.0042108246,0.6258033514,-0.0217756461,-0.3278491497,-0.0295965802,-0.3862670362,-0.3342377543,-0.0254537836,0.3758757412,-0.2405271232,0.2571175992,-0.0810012892,0.1662137806,-0.1648440957,0.0016576174,-0.2084467858,0.1559554487,-0.2232463211,-0.0478444584,0.0609903559,0.1417697221,0.3283350468,-0.1531536877,0.137735337,-0.3344838917,0.0545238182,-0.1950105727,0.1769057661,0.272721231,0.1509708464,0.2214309722,0.3293368816,0.0471366122,-0.1581557691,0.0576178357,-0.2800487876,-0.393958509,0.0948265791,0.125705868,0.1640482694,0.1380668581,0.2096283287,0.2284054905,0.0393176824,-0.217241779,-0.086212799,0.1584498584,-0.2489687502,-0.3260108531,0.0923232064,0.3571796417,-0.3308548927,-0.1999840885,-0.1797043085,0.0452080742,0.2457810342,0.2800107002,-0.2833347917,0.0341763683,-0.0315481387,0.0204127356,0.5325303674,-0.5967894197,-0.4960958958,0.2557237446,-0.0120022874,0.0755669326,0.3176009953,0.2206098139,0.243834272,0.174723804,0.2448198795,0.3377783,0.0100636194,-0.2427791357,0.1311953366,-0.018845398,0.5743771195,0.024627924,-0.1827917993,0.1722548008,0.179322198,0.2157419771,0.3071216941,0.1184071898,-0.1560167372,-0.0024089026,-0.0152380494,0.0578808635,-0.0360682309,-0.2658931017,0.0435572006,0.1698504686,-0.0126617411,-0.1012088358,0.0056865183,-0.1509738266,-0.2884414792,0.0347993784,-0.2785722017,-0.491494,0.0957370102,0.2937062681,0.1070410609,0.0497993976,-0.3838412464,0.1931142658,-0.1739235818,0.0594511293,-0.2641113698,0.2616459131,-0.071247302,0.1353574991,-0.2566051781,0.561059773,0.2033924013,0.1142710522,0.0233576857,0.2034628838,0.2300905138,-0.0946370885,-0.1387583762,-0.1457545161,0.257663995,-0.4002567232,0.0830498561,0.1750844419,0.2735494971,-0.1525911093,0.0078616152,0.3907418847,0.3698183298,0.4349882603,0.2422243506,0.2169765234,0.0280832835,0.0046050744,0.142694205,-0.4672394693,-0.1400223374,-0.2043556273,0.1088456437,-0.1078538373,-0.3479730487,0.286169678,0.4567412138,0.0138227446,0.0307955127,-0.2911535203,-0.3108551204,-0.0372543037,-0.078284055,0.2733258009,0.4840683341,0.2343032658,0.0746500492,0.0855285451,-0.0124238692,0.0048267907,0.1198433116,0.1941518635,0.2420474291,0.1419252902,0.1857102364,0.0966679379,0.1877738088,-0.272166729,0.2104979455,0.3406482637,-0.3003823757,0.1629361212,-0.4099977314,-0.4880042076,0.0144571178,-0.3423974514,0.0018421698,-0.5070488453,-0.0561672412,-0.0347428843,0.1162312776,-0.0149056157,-0.0840855092,-0.1331908703,0.0325239599,0.2032119483,0.1661236584,-0.3341940045,-0.2673010528,-0.0231218804,-0.0567047298,0.2377941459,0.0999237671,0.0278999768,-0.189251259,0.0388244055,-0.084839575,0.2069585621,0.0250350237,0.2025088966,0.1381623596,0.2134179175,0.0887719393,-0.1765597314,0.1649943441,0.2929334342,-0.3445080519,0.2239868641,0.1372479945,-0.0419559702,0.0637377724,-0.3967239261,-0.6464790702,-0.2083216608,0.1116249487,-0.0307405274,0.3225349486,0.3150753677,-0.401143074,0.1219071299,0.0677896291,0.3995665014,-0.3273166418,0.0046430989,0.0294392891,-0.2837055326,-0.1603331566,-0.0964622349,-0.2722267807,0.1293747127,-0.0154739451,-0.4079187512,-0.2342712432,-0.3203280866,-0.2996292412,-0.0103180036,0.046045877,0.2568389177,-0.0417731553,-0.075642392,0.0562416837,-0.0911545753,0.0311829057,0.1106230691,-0.0718172714,-0.1088000983,0.0765248314,-0.0211603157,0.2974599302,0.3613356352,0.2205090523,-0.092190899,-0.3419997394,0.1614146084,0.1079079434,-0.6722040176,-0.1568453014,0.032436084,-0.1027038842,0.2190039307,-0.0459264517,-0.195047453,-0.2466960847,-0.2480788231,-0.6623730063,-0.2552473545,-0.2224983871,-0.1908550113,0.0593469143,0.1209391579,0.2826324105,-0.1851935983,-0.0073138508,-0.4037668109,0.1687467992,-0.1218697429,-0.1089019477,-0.2181362212,-0.6910570264,-0.0773044154,0.1956616193,0.0323370025,0.4622499049,-0.0999986529,-0.0388961099,0.3081653118,0.0503740087,0.4035009146,-0.3674364984,0.0445706025,0.1530432105,-0.4430539012,-0.3725542128,-0.0338469483,-0.3618171215,0.1378152221,0.1949038804,0.1364648491,-0.2649859488,-0.0264044851,0.5953834057,-0.0804161653,-0.3547804058,-0.0740955696,-0.1484934092,-0.078068234,-0.205273971,-0.3540726304,-0.1436256617,0.3512363732,0.2910938859,0.2157930136,-0.4060229957,-0.194933027,0.5014382601,0.2625548542,0.0735823661,0.0825917721,-0.0869189426,0.2178036869,0.3571122885,-0.2109902501,0.6885301471,-0.0523779541,-0.3731905818,-0.0004462712,0.0574517325,0.5497580171,0.4288550019,-0.1412372589,0.1208152473,-0.098271884,-0.1491336226,-0.2115520835,-0.2318600118,0.3229676187,0.0726093873,-0.3635558486,-0.4884039164,0.2722448409,-0.0849142075,-0.0906401873,0.1458915025,0.0629895404,-0.2087931931,0.3545900285,-0.3809763789,0.9903618693,0.329277426,0.4495568573,0.3513564467,0.1354907006,0.735176146,0.1040998921,0.258238554,-0.2864293754,-0.1503847986,-0.2274817526,-0.1266020238,0.1126266196,0.0128647871,-0.1182800159,0.2837458253,-0.3521834612,0.0155713195,-0.1753759831,0.2185531557,0.1026427299,0.1495238841,-0.3357542157,-0.0622499883,-0.0671954378,0.3927616477,0.1136748344,-0.1045120135,-0.3019219637,-0.0686690286,-0.3666780889,0.061556071,0.0243283845,0.4778870642,-0.029201068,-0.4192871153,-0.2649336159,0.2284439802,0.1736293435,0.1698459834,-0.0976153165,-0.0065581724,-0.1813513488,-0.028641209,0.0009764929,0.0205423385,0.3865458667,0.0427641943,-0.4486111701,0.2900660634,-0.0063763079,0.0870849043,-0.1067259014,0.4001533985,0.0151130231,-0.3704199195,-0.2464222014,0.1145144105,-0.2462421358,-0.2208717167,-0.0222668797,-0.1909097135,-0.4741311073,0.0980500206,0.0997473747,-0.3421039283,0.0581395552,0.1928592473,0.5048798323,0.2213872969,0.4437157214,-0.1308045387,-0.1161542162,-0.130745545,0.1097265109,0.0628418848,-0.0933685005,0.105019927,0.0379055813,0.2710072398,-0.0583926253,-0.0869445056,0.0529418923,0.3689321876,-0.3398185372,-0.7617678046,-0.0953113586,0.4501473308,-0.0682357103,0.2793732285,-0.0826314911,0.4166557491,-0.0972739309,0.449621439,-0.1855598837,0.0874326229,-0.2266394645,0.1080848649,0.1004183739,-0.1510281712,0.3465953171,0.2598976493,0.1185414568,-0.0120902611,0.1103910804,-0.0956456363,-0.2598609328,0.2280461192,-0.0838566422,-0.123056218,0.0620282702,-0.2112430632,0.0025263119,-0.0480793752,-0.0685603619,0.2314646542,0.2323455811,0.4411837459,0.1262153089,0.2127239108,-0.1653215885,0.4683231413,-0.2051687986,0.4228176773,0.0863542929,0.1467078328,-0.0884291083,-0.0318222754,-0.3322874904,-0.2322847843,-0.2675160468,-0.2130786628,0.1666942835,-0.1475897729,0.2354432642,-0.1891659647,0.414274931,0.2301889509,0.0721537545,0.1936382502,0.1391346753,0.0371111669,-0.4144486487,0.1488356739,0.454993248,0.1194748431,0.0236286353,0.4790083468,-0.2553528845,0.3260786831,-0.2668020129,-0.1055664644,-0.0370652489,-0.1044924557,0.227946192,0.3263755441,-0.048150856,-0.1252164096,-0.0117482059,0.3561750948,0.1935323775,0.599709928,-0.038131427,0.2293096483,0.0385161117,-0.0413721576,0.2349705398,-0.2034702748,0.2185642272,0.0990774855,-0.1482100487,0.0345255546,-0.1047384143,0.2622841895,-0.1495844722,0.2826263607,0.0317238308,0.3951663673,-0.0697281957,-0.169736594,0.0825276375,0.0043179058,0.1090886295,-0.2155872732,0.1921231747,-0.3596949279,0.1351113468,0.0958110988,-0.006021318,-0.5095930099,0.2014565617,0.3303108215,0.1504404992,-0.2705943584,0.0683202595,0.2008242756,0.1757639199,0.434227854,0.353349179,0.2161447704,0.0241985302,-0.3509183526,0.0147401532,-0.0097053749,-0.4079248309,0.0778094605,0.109809503,0.1038501486,0.1833731234,0.3941486776,0.0116242608,0.0506091677,0.006906935,0.2788461447,-0.1356189102,-0.3819633424,0.7057721019,-0.1346540302,-0.0302854329,-0.0852226019,0.0394843109,-0.7841269374,-0.4474335313,0.2592221498,0.2150035948,0.2384361178,0.1179916561,-0.0187265836,-0.1833498329,0.2197689414,0.3294548392,0.5710617304,-0.2078905404,0.0956083685,-0.3913311064,0.1525500566,-0.245697245,-0.4644716382,0.0544782691,0.364710331,-0.2238105237,0.1497281045,0.0324502848,0.2397390604,0.0054366691,0.2687495947,-0.3348238468,-0.490411371,-0.1722480357,-0.0827267617,-0.1528235674,-0.1034834906,0.0875313729,-0.1272981763,-0.133444041,0.0049410253,0.0592920259,0.0846502855,0.4907219708,-0.2772696912,0.3279724419,0.7240934968,0.0451794267,0.0452400632,-0.0002587912,-0.2852587402,-0.0813498423,0.4049207568,0.3777348995,0.4518817663,-0.2942051888,-0.1741612405,-0.0510410443,0.0535423942,-0.0227197632,-0.1496456265,-0.0280041862,0.2312571257,0.0170167163,0.2565079629,0.0652557313,0.1513559222,-0.0064846766,0.1476945728,0.0025767989,-0.4071927369,0.5725112557,-0.4473517537,-0.3716686964,-0.0119449934,0.1999332607,-0.1460211426,0.1406576037,-0.5610473752,-0.2154170126,0.6048803926,-0.161528945,-0.0200687945,0.0389875136,0.0219917726,-0.174516812,-0.0873249248,0.123953484,0.013977658,-0.1805021465,-0.0339610428,-0.1548305154]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/115","title":"AttributeError: 'dict' object has no attribute 'info'","comments":"Good point @Colanim ! What happens under the hood when running:\r\n\r\n```python\r\nimport nlp\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\n```\r\n\r\nis that for every split in `cnn_dailymail`, a different dataset object (which all holds the same info) is created. This has the advantages that the datasets are easily separable in a training setup. \r\nAlso note that you can load e.g. only the `train` split of the dataset via:\r\n\r\n```python\r\ncnn_dm_train = nlp.load_dataset('cnn_dailymail', split=\"train\")\r\nprint(cnn_dm_train.info)\r\n```\r\n\r\nI think we should make the `info` object slightly different when creating the dataset for each split - at the moment it contains for example the variable `splits` which should maybe be renamed to `split` and contain only one `SplitInfo` object ...\r\n","body":"I'm trying to access the information of CNN\/DM dataset :\r\n\r\n```python\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nprint(cnn_dm.info)\r\n```\r\n\r\nreturns :\r\n\r\n> AttributeError: 'dict' object has no attribute 'info'","comment_length":115,"text":"AttributeError: 'dict' object has no attribute 'info'\nI'm trying to access the information of CNN\/DM dataset :\r\n\r\n```python\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nprint(cnn_dm.info)\r\n```\r\n\r\nreturns :\r\n\r\n> AttributeError: 'dict' object has no attribute 'info'\nGood point @Colanim ! What happens under the hood when running:\r\n\r\n```python\r\nimport nlp\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\n```\r\n\r\nis that for every split in `cnn_dailymail`, a different dataset object (which all holds the same info) is created. This has the advantages that the datasets are easily separable in a training setup. \r\nAlso note that you can load e.g. only the `train` split of the dataset via:\r\n\r\n```python\r\ncnn_dm_train = nlp.load_dataset('cnn_dailymail', split=\"train\")\r\nprint(cnn_dm_train.info)\r\n```\r\n\r\nI think we should make the `info` object slightly different when creating the dataset for each split - at the moment it contains for example the variable `splits` which should maybe be renamed to `split` and contain only one `SplitInfo` object ...\r\n","embeddings":[0.0860448629,-0.2287642211,-0.0386711396,0.5566271544,0.2393110842,0.0820070282,0.2878679335,0.2890189886,-0.009584601,0.1997731477,-0.026269732,0.3800343275,-0.2866004109,0.3110379577,-0.1329599321,-0.312898308,-0.0617810637,0.0819354728,0.0685543194,-0.1877292842,-0.2176921517,-0.0186453275,-0.2146317512,0.0927675217,-0.3512641191,-0.1529839337,-0.1275163889,0.1742342263,0.1626358628,-0.3590449393,0.2186698765,-0.0824093074,0.0676054358,0.1456326097,-0.0001200967,0.1501245797,0.3937638998,-0.0132749826,-0.4775072932,-0.1808106154,-0.3652683496,-0.4784733653,0.4104360938,-0.4280000329,0.0141701633,-0.1955828071,0.061412517,-0.2636317611,0.2931585908,0.3056741059,0.1423611492,-0.0766032413,0.0717008263,0.0712740049,-0.0116626862,0.2337168306,0.1357189268,0.0999801829,-0.1666555256,-0.1774171889,0.2367302626,0.3461387157,-0.057550244,0.2939850092,0.1243030056,0.5018519759,0.2756029963,-0.3079268336,-0.1120869219,0.2656952441,0.3015528619,-0.1667890847,-0.4109053612,-0.402497828,0.3017542958,-0.3622320592,-0.0417770557,0.205669567,-0.1510541141,0.1317649484,-0.1509331018,-0.162933141,-0.3153492808,0.4602555335,-0.1868041456,0.1968062371,-0.0741694197,0.2910074592,-0.1482135206,0.0615462661,0.1554133147,-0.004855847,-0.0361904167,-0.064291656,0.1266234815,-0.1908593476,0.0237800386,-0.3372946084,-0.230488047,0.1992776692,0.0300014876,-0.2213547528,-0.0144096715,0.1017793939,0.4677263796,0.0468038544,0.0606660657,0.543048203,0.0864747465,-0.0984051377,-0.3730103374,-0.09964858,-0.1022699848,0.2678650618,0.0410548337,0.0573246442,0.6547595263,0.1710321009,-0.2175796479,0.0475258753,-0.291619122,-0.1841300875,0.0196716096,0.4202069044,-0.3758158088,0.3166821003,-0.0023546722,0.2598299384,-0.1711176187,-0.1351026297,-0.1672390699,0.1883716136,-0.1644888967,-0.0454338044,0.0658375174,0.2751301527,0.1876869053,-0.1203981861,0.1502534896,-0.2108028531,0.0118785836,-0.1996326745,0.1468948424,0.2428472042,-0.0314818099,0.1773031205,0.3328693211,-0.1857574582,-0.2288423777,0.0898406729,-0.3860589862,-0.4279601276,-0.0504821837,0.1436499208,0.0643262267,0.001024521,0.1183593422,0.4320581853,-0.0101195741,-0.2594801784,-0.1091433689,0.0236568023,-0.2464594245,-0.329369843,0.1277022958,0.3755084276,-0.3210703731,-0.1966145337,-0.2238684297,0.0265125725,0.2153689116,0.0777888522,-0.294102788,0.3019200861,0.157410264,-0.1787673086,0.5169852972,-0.5471842885,-0.408711195,0.2692839503,0.072537303,0.1466144621,0.1654193103,0.1860263199,0.1700222194,0.1644580811,0.2275981903,0.5364189148,-0.1210606396,-0.1559235752,0.1025620252,-0.0822920352,0.4554801881,0.1035026759,-0.1737706363,0.3286602795,0.1283843666,0.3014736474,0.2027495503,0.0885989442,0.0090019833,-0.0317896008,0.0372528955,0.1113415211,-0.142565012,-0.129997924,-0.2095660865,0.1428060383,0.1349204779,0.15136531,-0.1725528538,-0.1128297523,-0.14405711,-0.034804374,-0.2341735959,-0.5324178338,0.1261230409,0.0717895478,-0.0386830643,0.0219994839,-0.4613384604,0.3799110055,-0.1166355982,0.074172087,-0.267467767,0.1598748565,0.024300117,0.0912194103,-0.2332624048,0.4407404959,0.0516265556,0.0769978166,-0.0225503724,0.2596832514,0.1651898474,-0.2301405966,-0.0918694288,0.0093692252,0.1660282463,-0.3179225326,0.1193693578,0.1209877506,0.2671662867,-0.1914663017,0.0119792027,0.3789295554,0.3295072913,0.4914184213,0.0843806192,0.2783165574,0.0613161698,-0.0651727021,0.1059273928,-0.3576053381,-0.1298234612,-0.2990133762,0.1558816433,-0.1758946329,-0.3190110028,0.0923191234,0.3692837954,0.1103923321,0.0295224693,-0.039108187,-0.2058977187,0.0333850421,-0.2275181413,0.3525831401,0.5222452879,0.2023171037,-0.1882100403,0.1882207245,-0.0132415779,0.0064639673,0.1345752031,0.2327063233,0.3685837388,0.0253146086,0.089261502,0.1285708994,0.2245946378,-0.5728698969,0.18870157,0.5563440323,-0.3044978976,0.1860845089,-0.1966259629,-0.426761061,-0.0486705042,-0.3421123028,-0.0448546223,-0.3559026122,-0.103885524,-0.1297439635,0.0746125579,-0.0035421327,-0.131395027,-0.0159502942,0.0439389907,-0.0229033977,0.1642870754,-0.443105191,-0.2144471407,-0.0498669632,0.045073919,0.1782763898,0.1592863947,-0.0440746322,-0.0607202239,0.1528818458,-0.048673138,0.0986357853,-0.0524646044,0.3286854029,0.2460146993,0.2065265924,0.0840366185,-0.2817554176,0.2767193615,0.0393044353,-0.1586167961,0.2133670747,0.136747092,0.1096086502,0.105488278,-0.5136440396,-0.62358284,-0.274356246,0.0848078653,-0.0427183509,0.3290720582,0.4674614072,-0.0300893597,0.079797782,0.1688726693,0.4912845492,-0.1182611957,-0.1812795103,0.0237383693,-0.3036913574,-0.1763626784,-0.1564724147,-0.2721151114,0.0525374189,-0.0739744976,-0.4470912814,-0.2999522984,-0.2285856009,-0.2480921298,-0.088008076,0.060194239,0.1543685943,0.0106969262,-0.0379680768,0.0280986261,-0.2343032211,0.0801368356,0.1232224703,-0.0749570727,0.01107259,0.1864682436,-0.1427345574,0.1945560277,0.1889092326,0.0911906809,-0.0227014516,-0.2615151703,0.1633200943,0.0860735402,-0.6142836809,-0.0744868666,0.0496811159,-0.1890935451,0.1765649319,-0.0135553777,-0.2396260053,-0.3505911827,-0.082280077,-0.43670398,-0.301920712,0.0377715118,-0.349950105,-0.003130391,-0.0234195143,0.208810702,-0.2183731049,-0.0478227101,-0.2737944722,0.1834969968,-0.265694797,-0.0467953496,-0.3311159015,-0.6209935546,-0.187449351,0.2123217881,-0.0389576852,0.5240558982,-0.0960634053,-0.0959115177,0.2434383035,0.1877692342,0.3413788676,-0.3294027448,-0.2207898796,0.1993754357,-0.3081077635,-0.2661581933,-0.0125897462,-0.2624736726,0.2308905274,0.0840221941,0.1518732458,-0.2635540366,-0.2725986242,0.4996643662,-0.047410991,-0.3775448501,-0.1506830454,-0.1242490262,-0.1735735089,-0.2629612386,-0.3738561273,-0.1396512091,0.3977634609,0.3071323633,0.2809025049,-0.4559299946,-0.0161795542,0.5261110067,0.1319102943,-0.0191649981,0.1588785797,-0.0147514725,0.2452945709,0.34801054,-0.1547147334,0.6451739669,-0.2067790329,-0.4464270771,-0.2170442492,0.096778743,0.5147438049,0.3108335137,0.0173721388,0.0406193957,-0.1535697281,-0.0493383743,-0.2197557688,0.0758718699,0.2726365626,-0.0844428912,-0.5733374357,-0.7434114814,0.3695496321,-0.031038275,0.0652467757,0.1506867856,0.0506816134,-0.220097661,0.5178812742,-0.2569654584,0.9998452663,0.1222015619,0.4613041282,0.3156251907,0.0823382288,0.7598273158,-0.0633251593,0.2366248369,-0.2869423926,-0.1132774875,-0.0995151922,-0.0225576945,0.1267461479,0.0571400337,-0.0209355671,0.2738402188,-0.3052709103,-0.2133144438,-0.1185837984,0.4240011871,0.1744492352,-0.0879977942,-0.377571404,-0.0073220115,-0.1439826339,0.2474928051,0.0678809956,-0.2410566807,-0.2992667258,-0.0153948199,-0.2301784009,-0.0690459758,-0.0014922715,0.3913457394,0.1123484671,-0.4293933511,-0.1548964977,0.2263776362,0.1400167346,-0.0001003022,-0.1346225739,0.0734717622,-0.2021020055,-0.0619430877,0.0565335713,-0.0294130091,0.545021832,0.1291365772,-0.3338742554,0.2432878017,-0.0680089369,-0.0829130486,-0.0214943159,0.1990619153,0.1631254703,-0.3908890188,-0.276775986,0.0643763989,-0.1077012941,-0.1513024718,0.0519833937,0.0644566044,-0.444658041,0.2857235074,0.1659431159,-0.3559823036,0.0334116854,-0.0293755271,0.3441929519,0.1920981109,0.4179258645,-0.0483346991,0.0034172733,-0.1632783115,0.2689672709,0.2116617858,-0.0965425521,0.1407293081,0.2174390852,0.2164722681,-0.0511373132,-0.0398687795,-0.064002648,0.3094502985,-0.1029320881,-0.6809609532,-0.2260452062,0.3558220565,-0.0888810754,0.2200123072,-0.087990962,0.3695601225,-0.0372142904,0.404543072,-0.2320103943,-0.086496681,-0.13750045,0.0756261498,0.2385258228,-0.0305666011,0.362940222,0.1003727913,0.0797182396,-0.2587348521,0.0479663499,-0.1424860805,-0.3373849988,0.2134509087,0.0242026653,-0.0830624029,0.0481113158,-0.2655965686,-0.0406380072,-0.172041595,-0.1661501825,0.2982859313,-0.0029610328,0.6121310592,0.0350435637,0.3176383674,-0.1252473444,0.3119830191,-0.0618791468,0.3593980968,0.0486883298,0.1033123359,-0.0997074768,0.0335169286,-0.1895166785,-0.1363866031,-0.4034407437,-0.1352075487,0.2273533791,-0.1296814531,0.3118281662,-0.0196355414,0.2376512438,0.414401412,-0.1229048446,0.2113859653,0.3546824455,0.0849061087,-0.2380658686,0.2278732508,0.3830392659,0.1268699169,-0.0226749424,0.4299106896,-0.3232186139,0.251219511,-0.6819221377,0.0088795153,-0.065978907,-0.0778630897,0.1492553651,0.3625010848,-0.0338524021,-0.1559127271,0.1123943776,0.3487477899,0.1415754855,0.7695274949,-0.0935687795,0.178314954,0.1059763432,-0.1595848501,0.1738319844,-0.100675717,0.2519041002,0.1397706121,-0.0410985127,-0.0888485983,-0.1140481159,0.3358059227,-0.1596241146,0.111944139,0.09883295,0.2086175382,-0.00731715,-0.1386510134,0.0531157665,0.0646685287,0.0841273889,-0.2945115566,0.0844567418,-0.4773041904,0.2394621521,0.1333764046,-0.1010264531,-0.4834129512,0.2167518884,0.3828240931,0.1768933237,-0.2703917325,-0.1381626725,0.0654561818,0.035403695,0.3502778411,0.2074105144,0.3174919784,0.3150425255,-0.378813684,0.0372988507,0.0596941747,-0.3285831511,-0.0161813889,0.1548212022,0.1722812653,0.2621048391,0.2520404458,0.0226565786,0.0306113176,0.0622330867,0.2664444447,-0.100099124,-0.2283045352,0.7168126106,0.0259244349,-0.0125075988,-0.0889935791,0.066341728,-0.6105392575,-0.472433418,0.4058845043,0.2474890351,0.2331459075,0.0243962388,-0.0183270965,-0.1358532161,0.3031717241,0.2887056172,0.4536974728,-0.2951976657,0.1704634577,-0.4494920075,0.0891910419,-0.1738184988,-0.4273295403,0.0366424248,0.2685402334,-0.161010921,0.1174895316,-0.2216634005,0.0648126975,0.0111656319,0.2018957287,-0.3736800551,-0.4560360909,-0.3366246521,0.166943565,-0.0198303033,-0.1136479378,-0.0316366926,-0.1077481136,-0.0775697008,-0.1166220605,0.0480959043,0.1713340282,0.4316164255,-0.1280899942,0.2200974226,0.6994802952,-0.0704041943,0.1590989828,-0.125527516,-0.2694271803,-0.165231958,0.365824461,0.2203820944,0.403408438,-0.2829338908,0.0553558208,-0.1666609943,0.0688510165,-0.0296469331,-0.1650469899,-0.1649488509,0.3107210398,0.1461276263,0.2835080028,-0.0227956809,0.2195109576,-0.0578629375,0.1536682993,-0.1119299084,-0.40609622,0.5831545591,-0.3409291506,-0.4388330281,0.0932766274,0.2884216011,-0.087667793,0.1906590313,-0.5406758189,-0.141812712,0.4634838104,-0.1775232404,-0.2315238565,0.127287671,0.1137921065,-0.1065553129,-0.0174286589,0.2092203647,-0.1040872037,-0.1821131706,-0.0223513041,-0.1531525701]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/114","title":"Couldn't reach CNN\/DM dataset","comments":"Installing from source (instead of Pypi package) solved the problem.","body":"I can't get CNN \/ DailyMail dataset.\r\n\r\n```python\r\nimport nlp\r\n\r\nassert \"cnn_dailymail\" in [dataset.id for dataset in nlp.list_datasets()]\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\n```\r\n\r\n[Colab notebook](https:\/\/colab.research.google.com\/drive\/1zQ3bYAVzm1h0mw0yWPqKAg_4EUlSx5Ex?usp=sharing)\r\n\r\ngives following error :\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/cnn_dailymail\/cnn_dailymail.py\r\n```","comment_length":10,"text":"Couldn't reach CNN\/DM dataset\nI can't get CNN \/ DailyMail dataset.\r\n\r\n```python\r\nimport nlp\r\n\r\nassert \"cnn_dailymail\" in [dataset.id for dataset in nlp.list_datasets()]\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\n```\r\n\r\n[Colab notebook](https:\/\/colab.research.google.com\/drive\/1zQ3bYAVzm1h0mw0yWPqKAg_4EUlSx5Ex?usp=sharing)\r\n\r\ngives following error :\r\n\r\n```\r\nConnectionError: Couldn't reach https:\/\/s3.amazonaws.com\/datasets.huggingface.co\/nlp\/cnn_dailymail\/cnn_dailymail.py\r\n```\nInstalling from source (instead of Pypi package) solved the problem.","embeddings":[-0.1325635612,-0.1728316844,-0.1248831898,0.2880879045,-0.0853366256,0.1052738354,0.1285750419,0.088076137,-0.2978582978,0.0596829355,0.0546678603,0.3479437828,-0.0498145446,0.292714268,-0.0873555467,-0.0530289933,-0.0281037446,0.1780724227,-0.0441002548,-0.0091223195,-0.2079378217,0.2669224441,-0.4211028814,-0.0771987066,-0.2897993326,-0.2828618288,-0.202975899,0.1027728915,-0.2907704711,-0.2986947298,0.5752847791,-0.1030247286,0.1514659822,0.2612523735,-0.0001162649,-0.0256794877,0.4707442224,0.0104750236,-0.4517286718,-0.462490797,-0.1799417585,-0.3112878799,0.2668949962,-0.1383743435,-0.1209165454,-0.0421658978,0.096762538,0.0134510715,0.08351852,0.4198019207,0.1649901569,0.2614762187,0.2515462339,0.0076733814,-0.418274641,-0.1246746629,0.0812214091,0.5046625137,-0.0197427664,-0.2011644244,0.5380072594,0.0774497092,-0.2067647576,0.233229056,0.0647727549,0.1424549818,-0.0698840693,-0.4046576917,0.1229405627,0.1860254854,0.2632254064,0.0207848679,-0.2496980727,0.1561032385,0.2608366907,-0.254660964,0.1154409274,0.2897347212,-0.2844154239,0.1010440141,-0.312133491,-0.1097168177,-0.3314070106,0.5778590441,-0.3505963683,0.561886847,-0.157541886,0.1198096797,-0.001442411,-0.0607826337,0.5817742944,0.0670132935,0.221601814,0.0222872216,-0.0639433265,-0.1390542835,-0.0486659855,0.0062479246,-0.2156125307,-0.0621648617,-0.0854288712,-0.2308879048,-0.0174072664,-0.0122635402,-0.0389455259,0.1391890943,-0.0445690416,0.3787418604,0.3034226,0.3267440796,-0.0350960605,0.000849766,-0.2349606603,0.148383677,-0.1538766026,0.0780672878,0.4210164249,-0.2496808618,-0.3272556067,-0.1246191412,-0.5902455449,-0.0415871367,-0.0766843259,0.3471421599,-0.2759477198,-0.1510892063,-0.1269339919,0.1947115362,0.0722402334,0.1027369499,-0.1625682414,0.2143969983,-0.3106198013,0.0132001555,0.3655536771,-0.2886527777,0.4667890072,-0.3166426122,0.1405730695,-0.0117087476,0.015848998,-0.0840539187,0.0520681627,0.1326437294,0.0388751552,0.1506881714,0.0524970666,0.2183288187,-0.092939131,-0.0050630118,-0.2030860484,-0.506724596,0.1224947199,0.1542264372,-0.2542626262,-0.365760535,-0.0457823649,-0.040944919,-0.0599384606,-0.1148598343,-0.2098013908,0.0997427404,0.0228772685,-0.1202754304,0.1241058484,0.181479916,-0.3485296965,0.0323791355,-0.3189458847,0.1148401499,0.2245602459,0.3488176763,-0.2739505172,-0.0811710954,-0.0901051387,-0.0934087262,0.1898580492,-0.193391338,-0.5875319242,0.2199116498,-0.0862824172,-0.1562704444,0.0849595517,0.0095996726,-0.1190767735,0.0927276537,0.1836238354,0.5643208623,-0.1288073808,-0.0819425881,-0.1086177453,-0.3027399182,0.3777284324,0.1380462497,0.4464080036,0.0282017458,0.1606990248,0.103356719,0.2263633311,0.2492647916,-0.2601758838,0.1426169276,0.2498369813,-0.3017876148,0.0386034101,-0.1208028719,-0.0913432017,0.0687004402,0.1558278501,0.4970203638,-0.3082802296,0.0533939078,-0.3634628057,0.0029898037,0.0478360802,-0.1825904548,0.1315939575,0.209769398,0.1940524578,0.0906137228,-0.2397648096,0.3270562589,-0.0782081857,0.1575426608,-0.269038707,0.4490187466,-0.1895166934,-0.0600907989,0.2260601372,0.243904382,0.1746883839,-0.1591024101,0.0094464701,-0.1551231146,-0.3085186779,0.1087886915,0.1964936852,0.018409105,0.0911932066,-0.4585068524,0.2721798122,0.3323068619,0.0052503198,0.0965819806,0.3478583395,0.1764314771,0.0668430626,0.2715499699,0.1259653121,0.1260421127,0.2236893028,-0.0087250285,0.0658099279,-0.0991242453,0.1928074658,-0.1029638276,0.0901126713,-0.0987841189,-0.2173469812,0.1026085615,0.2663487494,-0.108567521,0.0243994873,0.009276717,-0.3923324645,-0.1980831176,0.0500488915,0.3951503038,0.5056527853,0.1812977195,0.1702767611,0.1510153115,-0.1535665542,-0.202889502,0.3293245435,0.2363599539,0.127284959,-0.1012239158,0.1367552131,0.1771022081,-0.0246134195,-0.3869362175,0.3189814985,0.389642626,-0.0599981174,0.1498281062,-0.4569960833,-0.7941927314,0.0936678275,-0.517932713,-0.1013013795,-0.3695623875,-0.2360026091,0.0965080857,0.4111672342,-0.0255898759,0.0730052963,0.2779136598,0.0486686677,-0.1464455128,-0.0374835208,-0.2067879289,-0.2355880141,0.1359621882,0.132337153,0.1807712913,0.279124409,-0.0822471753,-0.1051652059,-0.0013818628,-0.4353199899,0.164740622,-0.1001120508,0.4872505963,0.2220571488,0.3688397706,-0.1099040508,-0.1360710114,0.011276653,0.0095075462,-0.1547336727,-0.0523435436,-0.2469918579,-0.1470591575,0.1963800639,-0.4646048844,-0.6944027543,-0.1339432597,0.0800441504,-0.1830586642,0.0721943155,0.3371898532,-0.190896675,0.2026571929,-0.0030705547,0.1966808438,-0.1297685206,-0.2496330887,0.3328925669,-0.2424345016,-0.3275431991,0.0774986818,-0.128086701,0.296826601,-0.278146714,-0.5325272083,-0.4542021751,-0.1868614256,-0.1426704228,0.205377385,-0.0846498758,-0.0755332783,-0.2146513611,-0.0516522937,0.0597497709,-0.1765699238,-0.1856133938,0.1436050385,0.3393605649,0.1087776497,0.3504311442,0.0201077387,0.4086458087,0.1517719924,-0.2701259255,0.3417194784,0.0543476157,0.1267336756,0.1717814356,-0.7655741572,-0.0065748272,0.0912296474,-0.0668333694,0.1470218599,-0.1787853241,-0.5367721915,-0.3859931529,-0.3819732666,-0.3233430088,-0.1329277754,-0.0593478382,0.1476626843,0.2565658391,0.377110064,-0.0510040745,-0.0973518863,-0.4532804787,-0.1782214344,0.1795861721,-0.2396686822,0.1090718806,0.3474123478,-0.5611612797,-0.5499814153,0.3448520899,0.0073899459,0.5087954998,-0.102531068,0.3972463906,0.0546196327,0.0298866984,0.2965787351,-0.2233580202,0.0728127137,0.0396578573,-0.1232895926,-0.2315326333,-0.0653571337,-0.368086189,0.3236838579,0.2515344918,0.5398221612,-0.1138928533,0.0881233513,0.62126261,0.0039635752,0.0103232218,-0.2245854437,-0.1475253999,-0.3410786092,-0.1033845991,-0.1514238119,-0.2004840672,0.443209976,0.0021470075,-0.2474341094,-0.0421351343,-0.1324265748,0.2131520063,0.2957430184,-0.0066730897,0.2070727795,0.1785353571,-0.0301412996,0.1145196259,0.1856297851,0.5156526566,-0.0015376169,-0.1153807938,0.0000420053,0.1581770182,0.35194695,0.360719502,-0.2770816088,0.0924345553,0.0574347489,0.0534855872,0.1918984205,0.2186706364,0.3287962079,0.1763449758,-0.4129094183,-0.1328891516,0.1336050928,-0.0672575682,0.0509281754,-0.0549677759,-0.1390229017,-0.1006922796,0.1614903063,-0.0174133107,1.037468791,0.2919613123,0.002318691,0.2239404023,0.2178325206,0.7991517782,0.099978216,0.2419227809,-0.1753991991,-0.3233377635,-0.0625794753,-0.2219343483,0.0347262621,0.0150288967,-0.0406508595,0.3622539639,0.3035144508,-0.3166083395,0.1926882863,0.3814951181,0.0245449767,0.2351955175,-0.609574914,0.0585664771,-0.3075296581,0.2712094486,-0.1717493087,-0.1482972056,-0.1477733701,-0.2711003423,-0.2732656002,0.0795527846,-0.2700736523,0.1955267042,-0.0381975211,-0.0692579299,-0.0345884487,-0.0462486744,0.2845804691,0.2179158926,-0.2708780766,-0.1162046641,-0.1818929911,-0.577752471,0.2065443397,0.0872507021,0.2250651866,-0.1442297399,-0.2639597654,0.0407794863,0.0180438757,-0.0709014758,0.0984596387,0.0465386771,-0.1192148924,0.1024810821,-0.2075965106,-0.0240462329,-0.1034606546,-0.0584720075,0.0812420845,0.0178407524,-0.1611663848,0.3155433834,0.2823424935,-0.2927349508,-0.0979283527,0.3588816226,0.0731356442,-0.0729212835,0.3552950025,0.2626289427,-0.1983146071,-0.1047220305,0.1433139741,-0.1502799392,-0.2271710336,0.0364004858,0.0810453221,0.251837492,0.1154314876,0.3974749148,0.36241436,0.2371757179,-0.1498031467,-0.8817759752,0.045655828,0.2567014992,-0.0799689591,0.1709131002,0.1449917257,0.1070346758,0.1447328031,0.0060784719,-0.2633076906,0.1596488655,-0.3147291243,0.0778047219,0.0707805604,-0.2056909204,0.5693100691,0.0569759384,0.0516615249,-0.0314292386,0.0767887011,-0.2028081566,-0.0225495193,0.174501285,0.1545321196,-0.1612207741,0.1194525734,-0.39075014,-0.2148760408,0.1611218303,-0.0728096813,-0.0194771904,0.2153815776,-0.1133029461,0.2339196056,-0.0924223512,-0.1778906584,0.2602176666,0.0710813627,0.114909865,0.0716567263,0.230172202,0.2254589349,-0.0114111742,0.0960250273,-0.3350308239,-0.2162459195,0.0695268512,0.4948171675,-0.3508018553,0.0346787125,-0.0042346092,0.2899496853,0.1791649312,-0.0193964727,0.1102544293,0.4123606086,0.1099774167,-0.4181596637,-0.0011047429,0.1513393819,0.2846257091,-0.1197600961,0.2549881339,-0.1618804187,0.2145800442,-0.0887738243,-0.0918790922,0.3628063202,0.1284062713,0.0122099379,0.1048171222,-0.1531488001,0.0331835002,0.2334901989,0.0868138149,0.0434163101,0.2407686859,-0.0562379658,0.1444862187,0.1230523959,0.099700056,0.3605891466,-0.0678949207,0.213650614,-0.0886248648,0.0817642957,0.0263017528,-0.0117902625,0.5023523569,-0.3560862839,0.1989364922,-0.3046838939,0.3021398485,-0.1053864136,0.0908286124,0.0352430306,-0.0559555441,-0.0468923822,-0.0881964043,0.0591056161,-0.3133745193,0.0382143445,0.3155565262,-0.2332573682,-0.3098788261,-0.2016659081,0.3513458669,-0.0598239414,-0.237096563,0.2391025722,-0.1353395879,-0.0041188467,-0.1782781184,0.4842951,0.3357543349,0.3539469838,-0.1005510017,-0.0737016425,0.2034792155,-0.2730831206,0.0777913257,0.1291016638,0.6509997249,0.3340233564,0.199404031,0.0753578395,-0.09192577,0.3270461261,0.114399381,-0.0040690727,-0.1479695737,0.3844363987,-0.0046534324,-0.0458432175,-0.3573884964,0.0319541208,-0.8125851154,0.1583398134,0.2939425409,0.0305210929,0.0037894475,-0.2046182901,0.0379175693,-0.1855387837,0.5830961466,0.4075213671,0.6273577213,-0.4677210748,-0.1413228363,-0.4765673578,0.1515961438,0.1181604043,-0.0517737605,-0.0453413837,-0.1152540892,0.0591024794,0.3323251605,-0.0433765799,0.310321033,-0.1162115186,0.1951557696,-0.3050020933,-0.2512748241,-0.0047763768,-0.1232846901,-0.052886039,-0.0019079692,0.0796685442,-0.5957305431,-0.1269090027,-0.1195782349,-0.2830270231,-0.0021205919,0.302339524,0.1112060323,0.1750337481,0.8783550858,0.0528383739,-0.0245374665,-0.2862677574,-0.2784084082,-0.1354616731,0.3752395213,0.3330828846,0.2964968979,-0.0681653842,-0.0808845013,-0.2100301981,-0.0145186931,-0.2229393423,-0.1670486182,-0.0098537849,-0.068918027,-0.1792086512,0.1277781278,-0.0229540803,0.0108392118,-0.0194859747,-0.0958750024,-0.1091805845,-0.399528116,0.4943462908,-0.480047375,-0.1276861876,-0.0721073002,0.0939348936,-0.0479499102,-0.094088234,-0.3853567839,0.0581914783,0.3180679381,-0.0586052686,-0.0958255902,0.1164541543,0.1829300076,0.2491690814,-0.1439557523,0.2249824852,-0.0333175771,-0.2978137732,-0.30293414,-0.0462341383]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/38","title":"[Checksums] Error for some datasets","comments":"@lhoestq - could you take a look? It's not very urgent though!","body":"The checksums command works very nicely for `squad`. But for `crime_and_punish` and `xnli`, \r\nthe same bug happens:\r\n\r\nWhen running: \r\n```\r\npython nlp-cli nlp-cli test xnli --save_checksums\r\n```\r\n\r\nleads to:\r\n\r\n```\r\n File \"nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/home\/patrick\/python_bin\/nlp\/commands\/test.py\", line 61, in run\r\n ignore_checksums=self._ignore_checksums,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 383, in download_and_prepare\r\n self._download_and_prepare(dl_manager=dl_manager, download_config=download_config)\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 627, in _download_and_prepare\r\n dl_manager=dl_manager, max_examples_per_split=download_config.max_examples_per_split,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 431, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/patrick\/python_bin\/nlp\/datasets\/xnli\/8bf4185a2da1ef2a523186dd660d9adcf0946189e7fa5942ea31c63c07b68a7f\/xnli.py\", line 95, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 246, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 186, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 166, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum(path)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/checksums_utils.py\", line 81, in get_size_checksum\r\n with open(path, \"rb\") as f:\r\nTypeError: expected str, bytes or os.PathLike object, not tuple\r\n```\r\n","comment_length":12,"text":"[Checksums] Error for some datasets\nThe checksums command works very nicely for `squad`. But for `crime_and_punish` and `xnli`, \r\nthe same bug happens:\r\n\r\nWhen running: \r\n```\r\npython nlp-cli nlp-cli test xnli --save_checksums\r\n```\r\n\r\nleads to:\r\n\r\n```\r\n File \"nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/home\/patrick\/python_bin\/nlp\/commands\/test.py\", line 61, in run\r\n ignore_checksums=self._ignore_checksums,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 383, in download_and_prepare\r\n self._download_and_prepare(dl_manager=dl_manager, download_config=download_config)\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 627, in _download_and_prepare\r\n dl_manager=dl_manager, max_examples_per_split=download_config.max_examples_per_split,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 431, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/patrick\/python_bin\/nlp\/datasets\/xnli\/8bf4185a2da1ef2a523186dd660d9adcf0946189e7fa5942ea31c63c07b68a7f\/xnli.py\", line 95, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 246, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 186, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 166, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum(path)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/checksums_utils.py\", line 81, in get_size_checksum\r\n with open(path, \"rb\") as f:\r\nTypeError: expected str, bytes or os.PathLike object, not tuple\r\n```\r\n\n@lhoestq - could you take a look? It's not very urgent though!","embeddings":[-0.0627012923,0.4912028015,0.0221911166,0.0352558196,0.1421761513,-0.1060484871,0.2264277935,0.5860058665,0.0751720741,-0.0241698995,-0.1206225306,0.3043448329,0.1317296326,-0.1162971184,-0.0193382669,0.267116785,0.1346074939,0.1131177247,0.1047997475,-0.1083303615,-0.1602604091,0.0813890547,-0.0319696777,-0.053366188,-0.1634702533,0.0185005739,0.0922335386,0.1823868304,-0.2103388906,-0.3178810775,0.1828085035,0.2279720455,0.1180133522,0.1371126771,-0.0001219498,-0.0516471677,0.3258031607,-0.0858075172,-0.3152633905,0.0974388048,-0.2606828809,-0.3950639367,-0.0185791161,-0.4255894721,0.37205109,0.3203902245,0.0684392527,-0.0244946051,0.2928688824,0.2679957747,0.1374267936,0.5577475429,-0.1816007644,0.0434500761,0.1389746815,0.0707661435,-0.1192683354,0.2082669735,0.2688108087,-0.0789423734,0.1626288891,0.148846224,-0.3349234462,-0.3724066317,-0.0468121246,-0.1594828665,0.4003409147,-0.0175617374,-0.1278445572,0.2270002663,-0.3293230832,-0.2747108936,-0.2557941675,-0.1852754354,0.1262196153,-0.4058308303,0.3126735985,0.3615929782,-0.1322477013,-0.3856508434,-0.2356782705,0.3625898659,0.1786392778,0.1741923839,0.2103234828,0.3407019973,0.260371536,0.1711850762,0.2275183797,-0.0792823508,-0.217236042,0.0797711089,-0.2267632335,0.1732077897,-0.7504251003,0.0182806868,0.1907794774,0.1760831624,0.198428303,0.4816954732,0.195856899,-0.10183236,0.066376932,0.2116571963,0.0479924604,0.1656826884,0.5243514776,0.3478689194,0.4528412521,0.1817325801,0.0125049716,0.111005716,0.1427177042,-0.0208090693,0.0522172265,0.2191821039,-0.1132478863,-0.2716455162,-0.4418676496,0.1631049812,0.1084636375,-0.1463720649,0.470731169,0.2857308388,0.068439275,-0.0357059985,-0.1219876334,0.4323750138,-0.099094063,-0.100993596,-0.1348179877,-0.1447826028,-0.2988794148,-0.0133951716,0.3063559234,0.0413430072,0.2411156595,-0.1284284443,0.1660990715,-0.0986437798,-0.0269741975,-0.1339724809,-0.219033137,0.1573349684,0.0276565365,-0.1535784751,0.0760046393,-0.1846380979,-0.1296412945,0.1298684925,-0.1619165987,0.1030448377,0.0288048442,0.1467627883,-0.3563662469,-0.0005756225,0.2564969957,-0.3233272135,0.3759012222,-0.3198791146,0.1601622552,-0.1939737648,-0.0694036558,-0.2678243518,0.0827553272,0.1281574517,-0.0813919008,0.0126430159,-0.1524866223,-0.2828461528,0.2618383467,0.3076504171,-0.0190558992,0.1066120788,-0.2534818649,0.3117203414,0.6236253381,-0.5391952395,-0.3489595354,-0.103822574,0.2453111261,-0.2230147272,0.0535941944,0.191665262,0.2072511315,-0.3625619411,-0.4169750512,0.2400106043,0.0361243151,0.2027406842,-0.2288201153,-0.0826072171,0.2481996417,0.0433604345,-0.0329150036,-0.0492722131,0.1717432737,0.1826472133,0.2804702818,-0.151297465,0.1541961581,-0.0797671005,0.3585596979,-0.0942279845,0.0306893755,-0.1185662597,-0.2086607218,0.1897599101,-0.6152535081,0.1698651761,-0.0682334527,-0.1599045098,-0.2258606553,-0.076962322,-0.1803464442,0.099957794,0.1058685854,0.2672636807,0.2386163771,0.0891887695,0.006184604,0.1085713282,-0.6231620908,0.0500973426,-0.1666647792,0.3818546236,-0.109388046,-0.4351024926,-0.0777804404,0.3842042089,-0.0393123589,0.211449638,-0.2845964134,0.4408490956,0.0782611892,0.0158712585,0.0451843925,0.1725321412,0.0855831727,-0.216851294,0.0500729978,0.6702576876,0.2047982663,-0.0931499153,-0.1812061667,0.1928213537,-0.0809176266,0.0449230187,-0.0232682955,0.0919856131,0.2223759443,-0.2604334354,-0.2387338877,-0.2169543505,0.1933977008,0.2466154695,0.3714800179,-0.0275688227,0.0609269477,-0.0080922544,0.2202794701,-0.0130476607,0.1423615962,0.0530194193,0.4494906366,0.0485340543,-0.1523383409,0.4089898765,0.440344274,-0.047346212,-0.0382508226,-0.0022969323,-0.1980335712,-0.1626869887,0.0150017915,-0.1399281025,0.1665215641,0.2899983823,0.2119664103,0.0607976951,-0.1362030357,-0.0183633827,-0.0110555505,0.2668966353,-0.2982129753,-0.064503096,-0.3084881604,-0.467023313,-0.1548491269,0.0802347362,-0.4626788795,-0.4291225076,0.038176991,-0.091297403,-0.2140588909,0.1140495762,-0.3381961882,0.1988053173,0.0887017027,0.1420766413,0.3029599488,-0.2489279211,-0.1107928976,-0.0527097583,0.4460258484,0.216684252,0.4916332066,-0.1002859771,-0.1132466942,-0.1128514335,-0.3865994513,0.0523086786,-0.1023961753,0.1059584245,0.3008179963,-0.0112966755,-0.3379214704,-0.3600876331,0.3471969962,-0.1742161661,-0.1782204956,0.3479773402,0.0316541046,-0.0400286242,-0.2230281532,0.1262023896,-0.1741138548,-0.3688487113,-0.0119620208,0.0021293128,0.2077801377,0.1534917057,-0.2540101111,0.2515192032,-0.2722303271,0.1960293502,-0.0344826356,-0.2537587285,0.218549028,-0.103799589,-0.5492884517,-0.2848454118,0.0491536111,0.0365566164,-0.0146186268,-0.3372529447,-0.0577521324,-0.2105458826,0.1740269214,0.3083756566,0.0900353342,0.1295183897,-0.2349826396,0.054937344,-0.3663519621,-0.0224972423,0.2733776867,0.0242557637,0.6932616234,-0.104063876,-0.0190503821,0.0218197126,-0.1689798832,0.5697202682,0.0551941022,0.4835469127,0.2954323292,0.2614176571,0.1312999725,-0.1737048477,0.3005037308,0.2075375319,-0.0891058892,0.0702408105,0.0197217502,-0.1144013628,-0.1642323732,0.063145794,-0.0991122052,-0.3288484514,0.0574325472,-0.2308800966,0.2036144584,-0.0569812879,0.0050396742,-0.0520032607,-0.2992756963,-0.1895920932,0.4029568136,0.0352074802,0.0825007558,-0.8941776156,-0.3734428883,-0.0580198318,-0.0230483562,0.1503069699,0.4448044598,-0.052956935,-0.0915122777,-0.0444783419,-0.1056422666,0.1772993207,-0.0168644059,0.2455539256,0.1647070795,0.461846292,-0.4230664074,-0.3309165835,-0.1746821254,-0.2900449634,0.772881031,0.1094585955,-0.3473772407,0.0844293758,0.018153118,0.262878716,-0.1098704189,-0.4754602313,-0.343226999,-0.0321217552,-0.5686990619,0.0901678875,-0.1974218637,0.2255198807,-0.2028981,-0.007569442,0.1711058617,-0.1430691928,0.2128360271,0.1074751765,0.3902181387,0.1014437973,0.1900574118,-0.6925087571,-0.0128672719,-0.3469009399,0.4314152598,-0.3060939908,-0.0719964132,0.0285656154,-0.2489468753,0.2488690466,0.2073641419,0.0063013439,-0.0379808694,-0.1942691356,-0.078680329,0.2778099179,-0.1101286784,0.0465850681,-0.1467799395,0.0360385217,0.1356880516,0.0634812564,0.0873834863,0.1289671808,0.0959914476,-0.1394473463,0.0520357676,0.1729335189,0.2594872415,0.8522511721,0.187081933,-0.0413960367,0.322491914,0.0436897911,0.5114994645,0.2631316483,-0.045972731,-0.3159870207,-0.2445199192,0.2126174122,-0.1158060953,0.0311515462,0.0026393128,-0.136906743,0.0526393913,-0.3026596904,0.0340925716,0.0386014171,0.1042212695,-0.2920203805,-0.1956309825,-0.2934521139,0.1118570045,-0.0340618156,-0.0716128424,-0.1246580109,-0.0158204231,-0.2190474421,-0.1162916198,-0.241841346,0.3268517256,0.375132829,0.2795345187,0.3088382781,-0.1053697243,-0.172169596,0.4308393598,0.3833262622,0.3241157532,-0.0469214544,-0.0500255041,0.2153294533,0.1384762526,0.1145162135,0.0529842786,0.4081784189,-0.2819083333,-0.3036103845,-0.2963477671,0.0005155209,-0.3420434594,-0.0163207203,0.0408508219,-0.3057286143,-0.0060490808,-0.1209774986,-0.1276033074,-0.1601526439,-0.1505137384,0.0940362066,0.3216533065,-0.1465225667,0.4087108672,-0.1629143953,-0.0995132029,0.0657532439,0.2217114717,0.1253989488,-0.151916489,0.3232159317,0.1185117215,-0.0750734359,-0.2994795442,-0.2114419341,-0.2859397531,-0.3340995014,0.0424082093,-0.3305135965,-0.1347117424,-0.0309969354,0.3189557493,0.0510884859,0.0008771661,0.1723090112,-0.8057903647,-0.3234634697,-0.0176723246,0.1075909957,0.1114090905,-0.1055119708,0.1768844426,0.0630181581,0.0597595684,-0.2304637283,-0.1936600506,-0.4387562573,0.2453925908,0.3086216152,-0.023415165,-0.1464534551,-0.0446066856,0.0212506596,0.1878371835,-0.2749356925,-0.1673557758,-0.2905701101,0.1769852042,0.1942325085,-0.2884044945,0.2095006257,-0.1954605579,-0.0210173819,-0.128417179,-0.227964282,-0.287465483,-0.1371353865,0.2967792153,-0.0696971491,0.3016501963,0.3740829527,0.1151087135,-0.1468939185,-0.1791381091,0.0058008372,0.3783899546,-0.032310456,-0.0504013747,-0.39962852,-0.0413638614,-0.1118234098,0.3709046245,0.2230171263,-0.1674646884,-0.0073870653,-0.0221952535,0.1894589663,0.5158637166,-0.2272296697,-0.0111601707,0.2188180834,0.1058459729,-0.3579701781,0.0742272884,0.1458030194,0.3003220856,-0.1130555421,0.1453100145,0.3143996596,-0.5344158411,0.0374046676,-0.1251326799,0.2200834602,-0.0348914042,0.1869916618,0.6134514213,0.0214274246,-0.0261587482,-0.0164434612,-0.5574680567,-0.1759773493,-0.2727543414,-0.1565748006,0.5726572871,-0.0191970393,0.3378191888,0.295250237,-0.1438747197,0.3550578058,0.0667523593,-0.08251784,-0.0492106751,0.030975379,0.6170881391,-0.1240669042,-0.0032552867,-0.2620034516,0.4167793095,0.0118920216,0.0029244698,-0.7036758661,-0.122521095,0.0530188456,-0.1440849453,0.0717613623,0.2183354199,-0.0618503615,0.2090263516,-0.1070683524,-0.3118742108,-0.047426831,0.1139507219,-0.0056263413,-0.2752245069,0.2127086222,0.2968198359,0.096461758,0.0482514389,0.2512210906,0.517332077,0.6780068278,0.4037903249,-0.0311537273,0.0277381334,0.0893184468,-0.191722706,0.4383191168,0.3782945573,0.1211130098,0.370534271,0.022735348,-0.0717986077,-0.37267223,0.0469345525,-0.1442011744,0.3953694701,0.6808131933,0.2013915479,0.0490980893,-0.2055376917,0.0190036744,-0.4090503752,0.0908682421,0.1557385921,0.1374776661,0.3923416436,-0.2375065684,0.0626278892,-0.0439575091,0.4432863891,0.0786579773,-0.222279042,-0.2107690722,-0.3914448917,-0.5945703387,0.3091796637,-0.0900205746,-0.1103112027,0.0388355702,0.1946269125,-0.3591995835,0.0433389023,0.2884758413,0.0080677904,0.0107477577,-0.1215259731,-0.305639416,0.0861879215,-0.1210249439,-0.155523181,0.0220309086,-0.2280657887,0.1425570995,0.0081492495,0.0296611357,0.0584044643,-0.3096063435,-0.1291687489,0.2887797356,0.2433728129,0.3375445604,0.7790846229,-0.0342981219,-0.2687160671,-0.2744385898,0.1235166788,-0.0708819255,-0.0984067768,-0.1125384122,0.181746304,-0.1458387375,0.3906442821,-0.2668605447,0.4579822123,0.0540485047,-0.4571493864,-0.2478503436,-0.06851767,-0.0613558963,0.1012249887,0.0592227057,-0.0084168809,-0.0344311148,0.3457834125,-0.2023301125,-0.282543391,0.3413538337,0.0277229026,-0.0280720685,-0.3101087511,0.2275165617,-0.3426457644,0.1476496309,-0.8451424837,-0.1373260319,0.1091175228,-0.1255626231,-0.0498035774,0.1818447858,-0.1117396951,0.0617349483,0.0560518466,-0.4437292814,0.100038752,-0.2363803089,0.2210969925,-0.0943480134]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/38","title":"[Checksums] Error for some datasets","comments":"Fixed with 06882b4\r\n\r\nNow your command works :)\r\nNote that you can also do\r\n```\r\nnlp-cli test datasets\/nlp\/xnli --save_checksums\r\n```\r\nSo that it will save the checksums directly in the right directory.","body":"The checksums command works very nicely for `squad`. But for `crime_and_punish` and `xnli`, \r\nthe same bug happens:\r\n\r\nWhen running: \r\n```\r\npython nlp-cli nlp-cli test xnli --save_checksums\r\n```\r\n\r\nleads to:\r\n\r\n```\r\n File \"nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/home\/patrick\/python_bin\/nlp\/commands\/test.py\", line 61, in run\r\n ignore_checksums=self._ignore_checksums,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 383, in download_and_prepare\r\n self._download_and_prepare(dl_manager=dl_manager, download_config=download_config)\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 627, in _download_and_prepare\r\n dl_manager=dl_manager, max_examples_per_split=download_config.max_examples_per_split,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 431, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/patrick\/python_bin\/nlp\/datasets\/xnli\/8bf4185a2da1ef2a523186dd660d9adcf0946189e7fa5942ea31c63c07b68a7f\/xnli.py\", line 95, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 246, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 186, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 166, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum(path)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/checksums_utils.py\", line 81, in get_size_checksum\r\n with open(path, \"rb\") as f:\r\nTypeError: expected str, bytes or os.PathLike object, not tuple\r\n```\r\n","comment_length":32,"text":"[Checksums] Error for some datasets\nThe checksums command works very nicely for `squad`. But for `crime_and_punish` and `xnli`, \r\nthe same bug happens:\r\n\r\nWhen running: \r\n```\r\npython nlp-cli nlp-cli test xnli --save_checksums\r\n```\r\n\r\nleads to:\r\n\r\n```\r\n File \"nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/home\/patrick\/python_bin\/nlp\/commands\/test.py\", line 61, in run\r\n ignore_checksums=self._ignore_checksums,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 383, in download_and_prepare\r\n self._download_and_prepare(dl_manager=dl_manager, download_config=download_config)\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 627, in _download_and_prepare\r\n dl_manager=dl_manager, max_examples_per_split=download_config.max_examples_per_split,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 431, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/patrick\/python_bin\/nlp\/datasets\/xnli\/8bf4185a2da1ef2a523186dd660d9adcf0946189e7fa5942ea31c63c07b68a7f\/xnli.py\", line 95, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 246, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 186, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 166, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum(path)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/checksums_utils.py\", line 81, in get_size_checksum\r\n with open(path, \"rb\") as f:\r\nTypeError: expected str, bytes or os.PathLike object, not tuple\r\n```\r\n\nFixed with 06882b4\r\n\r\nNow your command works :)\r\nNote that you can also do\r\n```\r\nnlp-cli test datasets\/nlp\/xnli --save_checksums\r\n```\r\nSo that it will save the checksums directly in the right directory.","embeddings":[-0.0627012923,0.4912028015,0.0221911166,0.0352558196,0.1421761513,-0.1060484871,0.2264277935,0.5860058665,0.0751720741,-0.0241698995,-0.1206225306,0.3043448329,0.1317296326,-0.1162971184,-0.0193382669,0.267116785,0.1346074939,0.1131177247,0.1047997475,-0.1083303615,-0.1602604091,0.0813890547,-0.0319696777,-0.053366188,-0.1634702533,0.0185005739,0.0922335386,0.1823868304,-0.2103388906,-0.3178810775,0.1828085035,0.2279720455,0.1180133522,0.1371126771,-0.0001219498,-0.0516471677,0.3258031607,-0.0858075172,-0.3152633905,0.0974388048,-0.2606828809,-0.3950639367,-0.0185791161,-0.4255894721,0.37205109,0.3203902245,0.0684392527,-0.0244946051,0.2928688824,0.2679957747,0.1374267936,0.5577475429,-0.1816007644,0.0434500761,0.1389746815,0.0707661435,-0.1192683354,0.2082669735,0.2688108087,-0.0789423734,0.1626288891,0.148846224,-0.3349234462,-0.3724066317,-0.0468121246,-0.1594828665,0.4003409147,-0.0175617374,-0.1278445572,0.2270002663,-0.3293230832,-0.2747108936,-0.2557941675,-0.1852754354,0.1262196153,-0.4058308303,0.3126735985,0.3615929782,-0.1322477013,-0.3856508434,-0.2356782705,0.3625898659,0.1786392778,0.1741923839,0.2103234828,0.3407019973,0.260371536,0.1711850762,0.2275183797,-0.0792823508,-0.217236042,0.0797711089,-0.2267632335,0.1732077897,-0.7504251003,0.0182806868,0.1907794774,0.1760831624,0.198428303,0.4816954732,0.195856899,-0.10183236,0.066376932,0.2116571963,0.0479924604,0.1656826884,0.5243514776,0.3478689194,0.4528412521,0.1817325801,0.0125049716,0.111005716,0.1427177042,-0.0208090693,0.0522172265,0.2191821039,-0.1132478863,-0.2716455162,-0.4418676496,0.1631049812,0.1084636375,-0.1463720649,0.470731169,0.2857308388,0.068439275,-0.0357059985,-0.1219876334,0.4323750138,-0.099094063,-0.100993596,-0.1348179877,-0.1447826028,-0.2988794148,-0.0133951716,0.3063559234,0.0413430072,0.2411156595,-0.1284284443,0.1660990715,-0.0986437798,-0.0269741975,-0.1339724809,-0.219033137,0.1573349684,0.0276565365,-0.1535784751,0.0760046393,-0.1846380979,-0.1296412945,0.1298684925,-0.1619165987,0.1030448377,0.0288048442,0.1467627883,-0.3563662469,-0.0005756225,0.2564969957,-0.3233272135,0.3759012222,-0.3198791146,0.1601622552,-0.1939737648,-0.0694036558,-0.2678243518,0.0827553272,0.1281574517,-0.0813919008,0.0126430159,-0.1524866223,-0.2828461528,0.2618383467,0.3076504171,-0.0190558992,0.1066120788,-0.2534818649,0.3117203414,0.6236253381,-0.5391952395,-0.3489595354,-0.103822574,0.2453111261,-0.2230147272,0.0535941944,0.191665262,0.2072511315,-0.3625619411,-0.4169750512,0.2400106043,0.0361243151,0.2027406842,-0.2288201153,-0.0826072171,0.2481996417,0.0433604345,-0.0329150036,-0.0492722131,0.1717432737,0.1826472133,0.2804702818,-0.151297465,0.1541961581,-0.0797671005,0.3585596979,-0.0942279845,0.0306893755,-0.1185662597,-0.2086607218,0.1897599101,-0.6152535081,0.1698651761,-0.0682334527,-0.1599045098,-0.2258606553,-0.076962322,-0.1803464442,0.099957794,0.1058685854,0.2672636807,0.2386163771,0.0891887695,0.006184604,0.1085713282,-0.6231620908,0.0500973426,-0.1666647792,0.3818546236,-0.109388046,-0.4351024926,-0.0777804404,0.3842042089,-0.0393123589,0.211449638,-0.2845964134,0.4408490956,0.0782611892,0.0158712585,0.0451843925,0.1725321412,0.0855831727,-0.216851294,0.0500729978,0.6702576876,0.2047982663,-0.0931499153,-0.1812061667,0.1928213537,-0.0809176266,0.0449230187,-0.0232682955,0.0919856131,0.2223759443,-0.2604334354,-0.2387338877,-0.2169543505,0.1933977008,0.2466154695,0.3714800179,-0.0275688227,0.0609269477,-0.0080922544,0.2202794701,-0.0130476607,0.1423615962,0.0530194193,0.4494906366,0.0485340543,-0.1523383409,0.4089898765,0.440344274,-0.047346212,-0.0382508226,-0.0022969323,-0.1980335712,-0.1626869887,0.0150017915,-0.1399281025,0.1665215641,0.2899983823,0.2119664103,0.0607976951,-0.1362030357,-0.0183633827,-0.0110555505,0.2668966353,-0.2982129753,-0.064503096,-0.3084881604,-0.467023313,-0.1548491269,0.0802347362,-0.4626788795,-0.4291225076,0.038176991,-0.091297403,-0.2140588909,0.1140495762,-0.3381961882,0.1988053173,0.0887017027,0.1420766413,0.3029599488,-0.2489279211,-0.1107928976,-0.0527097583,0.4460258484,0.216684252,0.4916332066,-0.1002859771,-0.1132466942,-0.1128514335,-0.3865994513,0.0523086786,-0.1023961753,0.1059584245,0.3008179963,-0.0112966755,-0.3379214704,-0.3600876331,0.3471969962,-0.1742161661,-0.1782204956,0.3479773402,0.0316541046,-0.0400286242,-0.2230281532,0.1262023896,-0.1741138548,-0.3688487113,-0.0119620208,0.0021293128,0.2077801377,0.1534917057,-0.2540101111,0.2515192032,-0.2722303271,0.1960293502,-0.0344826356,-0.2537587285,0.218549028,-0.103799589,-0.5492884517,-0.2848454118,0.0491536111,0.0365566164,-0.0146186268,-0.3372529447,-0.0577521324,-0.2105458826,0.1740269214,0.3083756566,0.0900353342,0.1295183897,-0.2349826396,0.054937344,-0.3663519621,-0.0224972423,0.2733776867,0.0242557637,0.6932616234,-0.104063876,-0.0190503821,0.0218197126,-0.1689798832,0.5697202682,0.0551941022,0.4835469127,0.2954323292,0.2614176571,0.1312999725,-0.1737048477,0.3005037308,0.2075375319,-0.0891058892,0.0702408105,0.0197217502,-0.1144013628,-0.1642323732,0.063145794,-0.0991122052,-0.3288484514,0.0574325472,-0.2308800966,0.2036144584,-0.0569812879,0.0050396742,-0.0520032607,-0.2992756963,-0.1895920932,0.4029568136,0.0352074802,0.0825007558,-0.8941776156,-0.3734428883,-0.0580198318,-0.0230483562,0.1503069699,0.4448044598,-0.052956935,-0.0915122777,-0.0444783419,-0.1056422666,0.1772993207,-0.0168644059,0.2455539256,0.1647070795,0.461846292,-0.4230664074,-0.3309165835,-0.1746821254,-0.2900449634,0.772881031,0.1094585955,-0.3473772407,0.0844293758,0.018153118,0.262878716,-0.1098704189,-0.4754602313,-0.343226999,-0.0321217552,-0.5686990619,0.0901678875,-0.1974218637,0.2255198807,-0.2028981,-0.007569442,0.1711058617,-0.1430691928,0.2128360271,0.1074751765,0.3902181387,0.1014437973,0.1900574118,-0.6925087571,-0.0128672719,-0.3469009399,0.4314152598,-0.3060939908,-0.0719964132,0.0285656154,-0.2489468753,0.2488690466,0.2073641419,0.0063013439,-0.0379808694,-0.1942691356,-0.078680329,0.2778099179,-0.1101286784,0.0465850681,-0.1467799395,0.0360385217,0.1356880516,0.0634812564,0.0873834863,0.1289671808,0.0959914476,-0.1394473463,0.0520357676,0.1729335189,0.2594872415,0.8522511721,0.187081933,-0.0413960367,0.322491914,0.0436897911,0.5114994645,0.2631316483,-0.045972731,-0.3159870207,-0.2445199192,0.2126174122,-0.1158060953,0.0311515462,0.0026393128,-0.136906743,0.0526393913,-0.3026596904,0.0340925716,0.0386014171,0.1042212695,-0.2920203805,-0.1956309825,-0.2934521139,0.1118570045,-0.0340618156,-0.0716128424,-0.1246580109,-0.0158204231,-0.2190474421,-0.1162916198,-0.241841346,0.3268517256,0.375132829,0.2795345187,0.3088382781,-0.1053697243,-0.172169596,0.4308393598,0.3833262622,0.3241157532,-0.0469214544,-0.0500255041,0.2153294533,0.1384762526,0.1145162135,0.0529842786,0.4081784189,-0.2819083333,-0.3036103845,-0.2963477671,0.0005155209,-0.3420434594,-0.0163207203,0.0408508219,-0.3057286143,-0.0060490808,-0.1209774986,-0.1276033074,-0.1601526439,-0.1505137384,0.0940362066,0.3216533065,-0.1465225667,0.4087108672,-0.1629143953,-0.0995132029,0.0657532439,0.2217114717,0.1253989488,-0.151916489,0.3232159317,0.1185117215,-0.0750734359,-0.2994795442,-0.2114419341,-0.2859397531,-0.3340995014,0.0424082093,-0.3305135965,-0.1347117424,-0.0309969354,0.3189557493,0.0510884859,0.0008771661,0.1723090112,-0.8057903647,-0.3234634697,-0.0176723246,0.1075909957,0.1114090905,-0.1055119708,0.1768844426,0.0630181581,0.0597595684,-0.2304637283,-0.1936600506,-0.4387562573,0.2453925908,0.3086216152,-0.023415165,-0.1464534551,-0.0446066856,0.0212506596,0.1878371835,-0.2749356925,-0.1673557758,-0.2905701101,0.1769852042,0.1942325085,-0.2884044945,0.2095006257,-0.1954605579,-0.0210173819,-0.128417179,-0.227964282,-0.287465483,-0.1371353865,0.2967792153,-0.0696971491,0.3016501963,0.3740829527,0.1151087135,-0.1468939185,-0.1791381091,0.0058008372,0.3783899546,-0.032310456,-0.0504013747,-0.39962852,-0.0413638614,-0.1118234098,0.3709046245,0.2230171263,-0.1674646884,-0.0073870653,-0.0221952535,0.1894589663,0.5158637166,-0.2272296697,-0.0111601707,0.2188180834,0.1058459729,-0.3579701781,0.0742272884,0.1458030194,0.3003220856,-0.1130555421,0.1453100145,0.3143996596,-0.5344158411,0.0374046676,-0.1251326799,0.2200834602,-0.0348914042,0.1869916618,0.6134514213,0.0214274246,-0.0261587482,-0.0164434612,-0.5574680567,-0.1759773493,-0.2727543414,-0.1565748006,0.5726572871,-0.0191970393,0.3378191888,0.295250237,-0.1438747197,0.3550578058,0.0667523593,-0.08251784,-0.0492106751,0.030975379,0.6170881391,-0.1240669042,-0.0032552867,-0.2620034516,0.4167793095,0.0118920216,0.0029244698,-0.7036758661,-0.122521095,0.0530188456,-0.1440849453,0.0717613623,0.2183354199,-0.0618503615,0.2090263516,-0.1070683524,-0.3118742108,-0.047426831,0.1139507219,-0.0056263413,-0.2752245069,0.2127086222,0.2968198359,0.096461758,0.0482514389,0.2512210906,0.517332077,0.6780068278,0.4037903249,-0.0311537273,0.0277381334,0.0893184468,-0.191722706,0.4383191168,0.3782945573,0.1211130098,0.370534271,0.022735348,-0.0717986077,-0.37267223,0.0469345525,-0.1442011744,0.3953694701,0.6808131933,0.2013915479,0.0490980893,-0.2055376917,0.0190036744,-0.4090503752,0.0908682421,0.1557385921,0.1374776661,0.3923416436,-0.2375065684,0.0626278892,-0.0439575091,0.4432863891,0.0786579773,-0.222279042,-0.2107690722,-0.3914448917,-0.5945703387,0.3091796637,-0.0900205746,-0.1103112027,0.0388355702,0.1946269125,-0.3591995835,0.0433389023,0.2884758413,0.0080677904,0.0107477577,-0.1215259731,-0.305639416,0.0861879215,-0.1210249439,-0.155523181,0.0220309086,-0.2280657887,0.1425570995,0.0081492495,0.0296611357,0.0584044643,-0.3096063435,-0.1291687489,0.2887797356,0.2433728129,0.3375445604,0.7790846229,-0.0342981219,-0.2687160671,-0.2744385898,0.1235166788,-0.0708819255,-0.0984067768,-0.1125384122,0.181746304,-0.1458387375,0.3906442821,-0.2668605447,0.4579822123,0.0540485047,-0.4571493864,-0.2478503436,-0.06851767,-0.0613558963,0.1012249887,0.0592227057,-0.0084168809,-0.0344311148,0.3457834125,-0.2023301125,-0.282543391,0.3413538337,0.0277229026,-0.0280720685,-0.3101087511,0.2275165617,-0.3426457644,0.1476496309,-0.8451424837,-0.1373260319,0.1091175228,-0.1255626231,-0.0498035774,0.1818447858,-0.1117396951,0.0617349483,0.0560518466,-0.4437292814,0.100038752,-0.2363803089,0.2210969925,-0.0943480134]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/38","title":"[Checksums] Error for some datasets","comments":"Awesome!","body":"The checksums command works very nicely for `squad`. But for `crime_and_punish` and `xnli`, \r\nthe same bug happens:\r\n\r\nWhen running: \r\n```\r\npython nlp-cli nlp-cli test xnli --save_checksums\r\n```\r\n\r\nleads to:\r\n\r\n```\r\n File \"nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/home\/patrick\/python_bin\/nlp\/commands\/test.py\", line 61, in run\r\n ignore_checksums=self._ignore_checksums,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 383, in download_and_prepare\r\n self._download_and_prepare(dl_manager=dl_manager, download_config=download_config)\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 627, in _download_and_prepare\r\n dl_manager=dl_manager, max_examples_per_split=download_config.max_examples_per_split,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 431, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/patrick\/python_bin\/nlp\/datasets\/xnli\/8bf4185a2da1ef2a523186dd660d9adcf0946189e7fa5942ea31c63c07b68a7f\/xnli.py\", line 95, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 246, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 186, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 166, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum(path)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/checksums_utils.py\", line 81, in get_size_checksum\r\n with open(path, \"rb\") as f:\r\nTypeError: expected str, bytes or os.PathLike object, not tuple\r\n```\r\n","comment_length":1,"text":"[Checksums] Error for some datasets\nThe checksums command works very nicely for `squad`. But for `crime_and_punish` and `xnli`, \r\nthe same bug happens:\r\n\r\nWhen running: \r\n```\r\npython nlp-cli nlp-cli test xnli --save_checksums\r\n```\r\n\r\nleads to:\r\n\r\n```\r\n File \"nlp-cli\", line 33, in <module>\r\n service.run()\r\n File \"\/home\/patrick\/python_bin\/nlp\/commands\/test.py\", line 61, in run\r\n ignore_checksums=self._ignore_checksums,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 383, in download_and_prepare\r\n self._download_and_prepare(dl_manager=dl_manager, download_config=download_config)\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 627, in _download_and_prepare\r\n dl_manager=dl_manager, max_examples_per_split=download_config.max_examples_per_split,\r\n File \"\/home\/patrick\/python_bin\/nlp\/builder.py\", line 431, in _download_and_prepare\r\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\r\n File \"\/home\/patrick\/python_bin\/nlp\/datasets\/xnli\/8bf4185a2da1ef2a523186dd660d9adcf0946189e7fa5942ea31c63c07b68a7f\/xnli.py\", line 95, in _split_generators\r\n dl_dir = dl_manager.download_and_extract(_DATA_URL)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 246, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 186, in download\r\n self._record_sizes_checksums(url_or_urls, downloaded_path_or_paths)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/download_manager.py\", line 166, in _record_sizes_checksums\r\n self._recorded_sizes_checksums[url] = get_size_checksum(path)\r\n File \"\/home\/patrick\/python_bin\/nlp\/utils\/checksums_utils.py\", line 81, in get_size_checksum\r\n with open(path, \"rb\") as f:\r\nTypeError: expected str, bytes or os.PathLike object, not tuple\r\n```\r\n\nAwesome!","embeddings":[-0.0627012923,0.4912028015,0.0221911166,0.0352558196,0.1421761513,-0.1060484871,0.2264277935,0.5860058665,0.0751720741,-0.0241698995,-0.1206225306,0.3043448329,0.1317296326,-0.1162971184,-0.0193382669,0.267116785,0.1346074939,0.1131177247,0.1047997475,-0.1083303615,-0.1602604091,0.0813890547,-0.0319696777,-0.053366188,-0.1634702533,0.0185005739,0.0922335386,0.1823868304,-0.2103388906,-0.3178810775,0.1828085035,0.2279720455,0.1180133522,0.1371126771,-0.0001219498,-0.0516471677,0.3258031607,-0.0858075172,-0.3152633905,0.0974388048,-0.2606828809,-0.3950639367,-0.0185791161,-0.4255894721,0.37205109,0.3203902245,0.0684392527,-0.0244946051,0.2928688824,0.2679957747,0.1374267936,0.5577475429,-0.1816007644,0.0434500761,0.1389746815,0.0707661435,-0.1192683354,0.2082669735,0.2688108087,-0.0789423734,0.1626288891,0.148846224,-0.3349234462,-0.3724066317,-0.0468121246,-0.1594828665,0.4003409147,-0.0175617374,-0.1278445572,0.2270002663,-0.3293230832,-0.2747108936,-0.2557941675,-0.1852754354,0.1262196153,-0.4058308303,0.3126735985,0.3615929782,-0.1322477013,-0.3856508434,-0.2356782705,0.3625898659,0.1786392778,0.1741923839,0.2103234828,0.3407019973,0.260371536,0.1711850762,0.2275183797,-0.0792823508,-0.217236042,0.0797711089,-0.2267632335,0.1732077897,-0.7504251003,0.0182806868,0.1907794774,0.1760831624,0.198428303,0.4816954732,0.195856899,-0.10183236,0.066376932,0.2116571963,0.0479924604,0.1656826884,0.5243514776,0.3478689194,0.4528412521,0.1817325801,0.0125049716,0.111005716,0.1427177042,-0.0208090693,0.0522172265,0.2191821039,-0.1132478863,-0.2716455162,-0.4418676496,0.1631049812,0.1084636375,-0.1463720649,0.470731169,0.2857308388,0.068439275,-0.0357059985,-0.1219876334,0.4323750138,-0.099094063,-0.100993596,-0.1348179877,-0.1447826028,-0.2988794148,-0.0133951716,0.3063559234,0.0413430072,0.2411156595,-0.1284284443,0.1660990715,-0.0986437798,-0.0269741975,-0.1339724809,-0.219033137,0.1573349684,0.0276565365,-0.1535784751,0.0760046393,-0.1846380979,-0.1296412945,0.1298684925,-0.1619165987,0.1030448377,0.0288048442,0.1467627883,-0.3563662469,-0.0005756225,0.2564969957,-0.3233272135,0.3759012222,-0.3198791146,0.1601622552,-0.1939737648,-0.0694036558,-0.2678243518,0.0827553272,0.1281574517,-0.0813919008,0.0126430159,-0.1524866223,-0.2828461528,0.2618383467,0.3076504171,-0.0190558992,0.1066120788,-0.2534818649,0.3117203414,0.6236253381,-0.5391952395,-0.3489595354,-0.103822574,0.2453111261,-0.2230147272,0.0535941944,0.191665262,0.2072511315,-0.3625619411,-0.4169750512,0.2400106043,0.0361243151,0.2027406842,-0.2288201153,-0.0826072171,0.2481996417,0.0433604345,-0.0329150036,-0.0492722131,0.1717432737,0.1826472133,0.2804702818,-0.151297465,0.1541961581,-0.0797671005,0.3585596979,-0.0942279845,0.0306893755,-0.1185662597,-0.2086607218,0.1897599101,-0.6152535081,0.1698651761,-0.0682334527,-0.1599045098,-0.2258606553,-0.076962322,-0.1803464442,0.099957794,0.1058685854,0.2672636807,0.2386163771,0.0891887695,0.006184604,0.1085713282,-0.6231620908,0.0500973426,-0.1666647792,0.3818546236,-0.109388046,-0.4351024926,-0.0777804404,0.3842042089,-0.0393123589,0.211449638,-0.2845964134,0.4408490956,0.0782611892,0.0158712585,0.0451843925,0.1725321412,0.0855831727,-0.216851294,0.0500729978,0.6702576876,0.2047982663,-0.0931499153,-0.1812061667,0.1928213537,-0.0809176266,0.0449230187,-0.0232682955,0.0919856131,0.2223759443,-0.2604334354,-0.2387338877,-0.2169543505,0.1933977008,0.2466154695,0.3714800179,-0.0275688227,0.0609269477,-0.0080922544,0.2202794701,-0.0130476607,0.1423615962,0.0530194193,0.4494906366,0.0485340543,-0.1523383409,0.4089898765,0.440344274,-0.047346212,-0.0382508226,-0.0022969323,-0.1980335712,-0.1626869887,0.0150017915,-0.1399281025,0.1665215641,0.2899983823,0.2119664103,0.0607976951,-0.1362030357,-0.0183633827,-0.0110555505,0.2668966353,-0.2982129753,-0.064503096,-0.3084881604,-0.467023313,-0.1548491269,0.0802347362,-0.4626788795,-0.4291225076,0.038176991,-0.091297403,-0.2140588909,0.1140495762,-0.3381961882,0.1988053173,0.0887017027,0.1420766413,0.3029599488,-0.2489279211,-0.1107928976,-0.0527097583,0.4460258484,0.216684252,0.4916332066,-0.1002859771,-0.1132466942,-0.1128514335,-0.3865994513,0.0523086786,-0.1023961753,0.1059584245,0.3008179963,-0.0112966755,-0.3379214704,-0.3600876331,0.3471969962,-0.1742161661,-0.1782204956,0.3479773402,0.0316541046,-0.0400286242,-0.2230281532,0.1262023896,-0.1741138548,-0.3688487113,-0.0119620208,0.0021293128,0.2077801377,0.1534917057,-0.2540101111,0.2515192032,-0.2722303271,0.1960293502,-0.0344826356,-0.2537587285,0.218549028,-0.103799589,-0.5492884517,-0.2848454118,0.0491536111,0.0365566164,-0.0146186268,-0.3372529447,-0.0577521324,-0.2105458826,0.1740269214,0.3083756566,0.0900353342,0.1295183897,-0.2349826396,0.054937344,-0.3663519621,-0.0224972423,0.2733776867,0.0242557637,0.6932616234,-0.104063876,-0.0190503821,0.0218197126,-0.1689798832,0.5697202682,0.0551941022,0.4835469127,0.2954323292,0.2614176571,0.1312999725,-0.1737048477,0.3005037308,0.2075375319,-0.0891058892,0.0702408105,0.0197217502,-0.1144013628,-0.1642323732,0.063145794,-0.0991122052,-0.3288484514,0.0574325472,-0.2308800966,0.2036144584,-0.0569812879,0.0050396742,-0.0520032607,-0.2992756963,-0.1895920932,0.4029568136,0.0352074802,0.0825007558,-0.8941776156,-0.3734428883,-0.0580198318,-0.0230483562,0.1503069699,0.4448044598,-0.052956935,-0.0915122777,-0.0444783419,-0.1056422666,0.1772993207,-0.0168644059,0.2455539256,0.1647070795,0.461846292,-0.4230664074,-0.3309165835,-0.1746821254,-0.2900449634,0.772881031,0.1094585955,-0.3473772407,0.0844293758,0.018153118,0.262878716,-0.1098704189,-0.4754602313,-0.343226999,-0.0321217552,-0.5686990619,0.0901678875,-0.1974218637,0.2255198807,-0.2028981,-0.007569442,0.1711058617,-0.1430691928,0.2128360271,0.1074751765,0.3902181387,0.1014437973,0.1900574118,-0.6925087571,-0.0128672719,-0.3469009399,0.4314152598,-0.3060939908,-0.0719964132,0.0285656154,-0.2489468753,0.2488690466,0.2073641419,0.0063013439,-0.0379808694,-0.1942691356,-0.078680329,0.2778099179,-0.1101286784,0.0465850681,-0.1467799395,0.0360385217,0.1356880516,0.0634812564,0.0873834863,0.1289671808,0.0959914476,-0.1394473463,0.0520357676,0.1729335189,0.2594872415,0.8522511721,0.187081933,-0.0413960367,0.322491914,0.0436897911,0.5114994645,0.2631316483,-0.045972731,-0.3159870207,-0.2445199192,0.2126174122,-0.1158060953,0.0311515462,0.0026393128,-0.136906743,0.0526393913,-0.3026596904,0.0340925716,0.0386014171,0.1042212695,-0.2920203805,-0.1956309825,-0.2934521139,0.1118570045,-0.0340618156,-0.0716128424,-0.1246580109,-0.0158204231,-0.2190474421,-0.1162916198,-0.241841346,0.3268517256,0.375132829,0.2795345187,0.3088382781,-0.1053697243,-0.172169596,0.4308393598,0.3833262622,0.3241157532,-0.0469214544,-0.0500255041,0.2153294533,0.1384762526,0.1145162135,0.0529842786,0.4081784189,-0.2819083333,-0.3036103845,-0.2963477671,0.0005155209,-0.3420434594,-0.0163207203,0.0408508219,-0.3057286143,-0.0060490808,-0.1209774986,-0.1276033074,-0.1601526439,-0.1505137384,0.0940362066,0.3216533065,-0.1465225667,0.4087108672,-0.1629143953,-0.0995132029,0.0657532439,0.2217114717,0.1253989488,-0.151916489,0.3232159317,0.1185117215,-0.0750734359,-0.2994795442,-0.2114419341,-0.2859397531,-0.3340995014,0.0424082093,-0.3305135965,-0.1347117424,-0.0309969354,0.3189557493,0.0510884859,0.0008771661,0.1723090112,-0.8057903647,-0.3234634697,-0.0176723246,0.1075909957,0.1114090905,-0.1055119708,0.1768844426,0.0630181581,0.0597595684,-0.2304637283,-0.1936600506,-0.4387562573,0.2453925908,0.3086216152,-0.023415165,-0.1464534551,-0.0446066856,0.0212506596,0.1878371835,-0.2749356925,-0.1673557758,-0.2905701101,0.1769852042,0.1942325085,-0.2884044945,0.2095006257,-0.1954605579,-0.0210173819,-0.128417179,-0.227964282,-0.287465483,-0.1371353865,0.2967792153,-0.0696971491,0.3016501963,0.3740829527,0.1151087135,-0.1468939185,-0.1791381091,0.0058008372,0.3783899546,-0.032310456,-0.0504013747,-0.39962852,-0.0413638614,-0.1118234098,0.3709046245,0.2230171263,-0.1674646884,-0.0073870653,-0.0221952535,0.1894589663,0.5158637166,-0.2272296697,-0.0111601707,0.2188180834,0.1058459729,-0.3579701781,0.0742272884,0.1458030194,0.3003220856,-0.1130555421,0.1453100145,0.3143996596,-0.5344158411,0.0374046676,-0.1251326799,0.2200834602,-0.0348914042,0.1869916618,0.6134514213,0.0214274246,-0.0261587482,-0.0164434612,-0.5574680567,-0.1759773493,-0.2727543414,-0.1565748006,0.5726572871,-0.0191970393,0.3378191888,0.295250237,-0.1438747197,0.3550578058,0.0667523593,-0.08251784,-0.0492106751,0.030975379,0.6170881391,-0.1240669042,-0.0032552867,-0.2620034516,0.4167793095,0.0118920216,0.0029244698,-0.7036758661,-0.122521095,0.0530188456,-0.1440849453,0.0717613623,0.2183354199,-0.0618503615,0.2090263516,-0.1070683524,-0.3118742108,-0.047426831,0.1139507219,-0.0056263413,-0.2752245069,0.2127086222,0.2968198359,0.096461758,0.0482514389,0.2512210906,0.517332077,0.6780068278,0.4037903249,-0.0311537273,0.0277381334,0.0893184468,-0.191722706,0.4383191168,0.3782945573,0.1211130098,0.370534271,0.022735348,-0.0717986077,-0.37267223,0.0469345525,-0.1442011744,0.3953694701,0.6808131933,0.2013915479,0.0490980893,-0.2055376917,0.0190036744,-0.4090503752,0.0908682421,0.1557385921,0.1374776661,0.3923416436,-0.2375065684,0.0626278892,-0.0439575091,0.4432863891,0.0786579773,-0.222279042,-0.2107690722,-0.3914448917,-0.5945703387,0.3091796637,-0.0900205746,-0.1103112027,0.0388355702,0.1946269125,-0.3591995835,0.0433389023,0.2884758413,0.0080677904,0.0107477577,-0.1215259731,-0.305639416,0.0861879215,-0.1210249439,-0.155523181,0.0220309086,-0.2280657887,0.1425570995,0.0081492495,0.0296611357,0.0584044643,-0.3096063435,-0.1291687489,0.2887797356,0.2433728129,0.3375445604,0.7790846229,-0.0342981219,-0.2687160671,-0.2744385898,0.1235166788,-0.0708819255,-0.0984067768,-0.1125384122,0.181746304,-0.1458387375,0.3906442821,-0.2668605447,0.4579822123,0.0540485047,-0.4571493864,-0.2478503436,-0.06851767,-0.0613558963,0.1012249887,0.0592227057,-0.0084168809,-0.0344311148,0.3457834125,-0.2023301125,-0.282543391,0.3413538337,0.0277229026,-0.0280720685,-0.3101087511,0.2275165617,-0.3426457644,0.1476496309,-0.8451424837,-0.1373260319,0.1091175228,-0.1255626231,-0.0498035774,0.1818447858,-0.1117396951,0.0617349483,0.0560518466,-0.4437292814,0.100038752,-0.2363803089,0.2210969925,-0.0943480134]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/6","title":"Error when citation is not given in the DatasetInfo","comments":"Yes looks good to me.\r\nNote that we may refactor quite strongly the `info.py` to make it a lot simpler (it's very complicated for basically a dictionary of info I think)","body":"The following error is raised when the `citation` parameter is missing when we instantiate a `DatasetInfo`:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/info.py\", line 338, in __repr__\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\nAttributeError: 'NoneType' object has no attribute 'strip'\r\n```\r\n\r\nI propose to do the following change in the `info.py` file. The method:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nBecomes:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n ## the strip is done only is the citation is given\r\n citation_pprint = self.citation\r\n\r\n if self.citation:\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nAnd now it is ok. @thomwolf are you ok with this fix?","comment_length":31,"text":"Error when citation is not given in the DatasetInfo\nThe following error is raised when the `citation` parameter is missing when we instantiate a `DatasetInfo`:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/info.py\", line 338, in __repr__\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\nAttributeError: 'NoneType' object has no attribute 'strip'\r\n```\r\n\r\nI propose to do the following change in the `info.py` file. The method:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nBecomes:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n ## the strip is done only is the citation is given\r\n citation_pprint = self.citation\r\n\r\n if self.citation:\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nAnd now it is ok. @thomwolf are you ok with this fix?\nYes looks good to me.\r\nNote that we may refactor quite strongly the `info.py` to make it a lot simpler (it's very complicated for basically a dictionary of info I think)","embeddings":[0.2855120897,0.0732322857,0.1477971971,0.1601161659,0.2827774882,0.2352983654,0.2543569803,0.4805189073,-0.4466493428,0.16624026,0.403686285,0.3917041719,0.0535713695,-0.203330785,-0.0758434907,-0.3508134186,-0.1062086448,0.2190288305,0.3720492125,-0.0522521213,-0.3091630936,0.1654009521,-0.3137940764,-0.0903745443,-0.2572643757,-0.1726606786,-0.2614687681,0.2065153718,-0.3102674484,-0.5589748025,0.1857405305,0.063550368,0.0325638615,0.0146645615,-0.0001185848,0.0302900337,0.3959799111,-0.1021313891,-0.2354656905,-0.2480834574,-0.5433374643,-0.2237135321,-0.029977249,-0.3961987495,-0.1640129983,-0.0125304721,-0.1367189437,0.1232259646,0.1175461933,0.2436022162,0.1624839008,0.0612453744,0.0864233896,-0.3654048443,0.0225184187,0.4103301764,-0.0947481841,-0.2211684287,-0.0262204297,0.0773107931,0.0846517608,0.2534255385,-0.196715951,-0.1238166988,0.4814615846,-0.0951773897,-0.0258708559,-0.0744219944,-0.0097684907,0.1676390171,0.5024431944,-0.2813969254,-0.5132006407,-0.0163041782,0.2781536281,-0.0212765168,-0.0074698143,0.0304121226,-0.0812080204,0.0895343795,0.2815128863,-0.2978960872,-0.0270233434,-0.1198625565,0.0443734415,0.2116338611,0.0791531727,0.1897022575,0.006169613,-0.1191703305,-0.1341097206,-0.0683485717,-0.2433660775,0.1042299494,-0.0947714001,-0.2623224258,0.3013670146,-0.1240373775,0.3263944983,0.2747790217,-0.0126671297,-0.3645196557,0.1062225774,0.1398010999,0.2596132457,0.0484170578,-0.0095627923,0.2912044525,0.3808363378,-0.2724874616,-0.0530267395,-0.0224923305,0.1467389613,-0.0760321692,0.1871124804,0.3163543642,0.6639338136,-0.2459217012,-0.2278142571,0.0181359351,-0.4110317528,0.0834341347,-0.0681851432,0.1200495139,0.0531694703,0.1811803877,0.0788510516,0.1461924911,0.0442111008,-0.3349076211,-0.3145993948,0.2601854205,-0.141174674,-0.1581093818,0.0291049238,0.3195924461,0.3016165793,0.0878599286,-0.2863543034,-0.1176711768,0.0519980602,-0.3555636108,-0.15103136,0.2481955886,-0.1378612816,0.0533519797,0.3309005797,-0.3365867436,-0.207349807,0.5098431706,-0.2286172807,-0.3707093,-0.1757764667,0.1305076778,-0.2939586639,-0.1857750118,-0.0267878566,0.1469056308,0.0339358598,-0.2264343053,-0.1586720645,-0.1596159935,-0.1011698768,-0.1808520108,0.210424751,0.7028290629,-0.2065771967,-0.2376949489,-0.1792689413,0.0066007464,0.3217360973,0.0637401119,-0.2051242739,0.4242260456,-0.1891367584,0.309779495,0.4551816881,-0.5058348179,-0.264592886,0.0924514309,-0.090154022,-0.1495330036,0.3402810097,0.0750064477,0.3930941522,-0.3331330121,0.1752074659,0.1514261812,-0.1647031307,-0.0128044328,-0.2054263502,-0.025216924,0.2687395513,-0.0305352192,0.1777623147,0.4301315248,0.026665451,0.23543787,0.0071865981,-0.0845381394,0.09592361,0.6712453961,0.4971642196,-0.0815782472,0.0407697707,0.0891511142,0.1425515115,-0.0065857228,-0.0049517527,0.5135893822,-0.4911940694,-0.3580025136,-0.2441371828,-0.0477071702,0.0384993032,-0.128395915,0.0801430494,0.1488384604,0.2277501523,0.153040424,-0.1139558703,0.0341832377,-0.3096472323,0.1729861945,-0.2149274796,0.0709076151,-0.0962565392,-0.1759302914,0.2825001478,0.4030955136,0.3565137982,0.1395501345,-0.2577136159,0.2078028768,0.3365240097,-0.0840731561,0.3318354785,-0.0264528077,0.2214430571,-0.5601446629,-0.106444627,0.1312949508,0.2259838134,-0.0394202173,-0.1677945852,0.1209984645,-0.0971575826,0.1369831562,0.0983100757,0.2548840344,0.2972361147,-0.1475794017,-0.2540400028,-0.1345717907,-0.4073702097,0.3577827811,0.0442956835,0.1807568073,-0.3323710263,-0.06579981,0.2814854383,0.0483954698,-0.083436124,0.0813504905,-0.2417711169,-0.0494092368,0.0345413573,0.504362762,0.558821857,0.2867323756,-0.1355875134,0.0329405963,0.0807511434,0.0325014032,0.1965010613,0.172426492,0.0140458588,0.1181470826,0.0886255354,0.264751792,-0.0481837392,-0.3537189364,0.2952138186,0.4577527642,-0.2807684243,-0.150439322,-0.3010823727,-0.3772722483,-0.2303933054,-0.1566966474,0.1558211148,-0.2895430624,-0.1114748344,0.0278519038,-0.226184845,0.2793827951,-0.2176556587,0.0751801282,0.1010244191,-0.2260205597,-0.0099929068,-0.1933918744,0.0566482544,-0.0408326052,-0.0482504629,0.2714774609,0.2608515024,0.0312365536,0.0936894119,-0.4852143526,-0.499157995,0.185065195,-0.2016006261,0.2247467786,0.3787457943,0.254168421,-0.056391865,-0.1754277796,0.3002544045,0.1890562177,-0.0711904615,0.2082998604,-0.2221645266,-0.1809340864,-0.1976263523,-0.4478684068,-0.1524800062,-0.3519616425,-0.324232161,0.0325689949,-0.1595597714,0.5114241242,-0.0376598351,0.107879214,0.4586960673,0.3906828761,0.0442582294,0.0350523777,0.3263026476,-0.1154825166,-0.3304879963,0.1488804519,-0.2420458794,0.050730776,0.1023525372,-0.1610190421,-0.4054076076,0.085415177,0.2321101427,-0.0968379304,0.2345744967,0.3909744322,0.2886578441,-0.1030012518,-0.1418347955,-0.2641219795,0.0721986815,0.0509948917,0.2560450733,0.161925137,0.1512395144,-0.0068999906,0.3923106492,0.1909050792,0.0996040404,0.221127972,-0.0202333014,0.1393013299,0.0169179533,-0.0862640515,-0.037109077,0.2619738877,-0.1297571361,0.0234721079,-0.0368065387,-0.1350055486,-0.0652676299,-0.2120331526,-0.2937174439,-0.2581718266,-0.1676616669,-0.2041821033,0.5298532844,0.1539096385,0.0566982999,-0.2150213122,0.1116406694,0.0398819856,0.3978654742,0.4186894894,-0.0655552149,-0.1002212837,-0.1168369129,-0.4247890711,0.0079289991,0.0285914894,0.289878875,-0.0292112362,-0.118423447,0.1176300943,0.1090158671,0.4440200627,-0.3201776147,-0.094239153,0.2171335667,-0.0520307347,-0.0749424621,0.0966636464,-0.1407677382,0.3265134692,0.1625849009,-0.0012651245,-0.3431600332,-0.1991199255,0.5652010441,0.0161457453,-0.0239477362,0.0118602039,-0.0996702164,-0.5567368269,-0.2893752158,0.007464183,-0.0616957918,0.3286765516,0.4054490328,0.0357954055,-0.3641443551,-0.0787933394,0.0652355626,0.1031478569,0.4205913842,-0.0471299216,0.1016341671,-0.2359844148,-0.1014064476,0.1652707905,0.2810600698,-0.0035395133,-0.4454729557,-0.1814178377,-0.0911664367,0.2103846818,0.0349834226,-0.1234335527,0.2475430071,0.0253230613,-0.0295028333,-0.1880521923,-0.3054905832,0.177183643,-0.217938289,-0.3578876555,-0.4102500975,0.0567891188,0.2962569296,0.003355328,0.1699745357,0.6390137076,-0.3207844198,0.3601374328,0.0156328343,1.0747797489,0.2924572229,-0.2725102603,0.2915841639,0.0433302224,0.5590177774,0.235897705,0.2341372669,-0.4298227131,-0.3786416948,-0.0747423917,-0.0519499518,0.2784281969,-0.0809192508,-0.5020384192,0.1756699979,-0.1523383707,0.0022322566,0.1220385805,0.0872596353,0.0060309148,-0.2312285155,-0.2017602473,0.0670160204,-0.1527360678,0.3128901124,-0.085725978,-0.243533656,-0.1070018113,-0.3578900099,-0.2665623128,0.000804026,-0.018101396,0.2025057226,0.1674943566,-0.2818686068,-0.0193199199,0.2606441677,0.3911963999,0.1438509226,-0.1460814923,0.1348209083,-0.3412264287,0.0674619228,0.0010561617,0.3297884166,0.6635107398,0.0570278242,-0.2791726291,0.0722952187,-0.3045301437,-0.2996293604,0.2755072117,-0.3606644571,-0.2726466358,-0.1459128559,-0.3561815023,0.1865038574,-0.1307550222,-0.3097569942,0.1204930171,0.1125206426,-0.2872116864,0.1730938405,0.0209985767,-0.1482337713,0.088712588,0.1320442706,0.4319786429,0.5363356471,0.352291584,-0.0066535491,0.0746256039,-0.1703940928,0.0755943954,-0.2000356019,-0.2088153064,0.4038462937,0.0661254823,0.328760922,-0.0338510461,0.3243108392,-0.147584945,0.211573422,-0.245580107,-0.3820871413,-0.1751209497,0.2641634643,-0.1494313031,-0.0163528882,-0.0356309339,0.239595592,0.0800135434,-0.148766011,-0.2949716449,-0.00972299,-0.4209564328,0.2777429521,0.3176947534,0.2704794109,0.0551025309,-0.0695449412,0.0922539011,0.0423996262,0.0950587615,-0.0792630613,-0.3869243264,0.1389072239,0.1080889925,0.2912574112,-0.0833027735,-0.1454300731,-0.1910989136,-0.0237107761,0.0475815721,0.2399301529,-0.0320061967,0.3826493323,-0.3793178201,0.1382137537,-0.2307718247,0.0374536328,0.114278093,0.1407286227,0.0325585678,0.1824617386,0.0813172907,-0.0186538789,-0.5630632043,0.1317278445,-0.1979646534,-0.1241242513,0.3146758676,0.009368144,0.1651304662,0.2314257771,-0.014451636,-0.0592810959,-0.1193889305,0.2787908614,0.0533665046,0.14934735,-0.4110959172,-0.0266342256,0.1290837824,-0.1473093182,0.1387932301,0.2995389104,0.0316387713,0.0877545178,-0.0951115862,-0.1150851697,0.7280268073,-0.3307806253,0.1762421429,0.4804491401,0.0208486728,0.1105091348,-0.0597954094,0.1357881129,0.1895601004,0.1926511824,0.0215189178,0.3219555616,0.0450908169,-0.1167576984,0.0563818365,-0.2132242769,0.7372803092,0.4240525365,0.280253619,-0.2031773329,0.0000221593,0.3681063056,-0.1852521598,-0.0167535301,-0.0322262719,0.1175446063,-0.1886272877,0.1395931393,-0.3298102915,-0.0430390052,-0.203642711,-0.2156832516,0.119598411,0.0806257576,0.0927326381,-0.1072928235,-0.0609588064,-0.425447315,0.3382223547,0.2237199843,0.111854054,-0.3376896977,-0.1360161602,-0.1737665087,-0.1550251991,0.1986856312,0.6460391879,0.380517453,0.2580200434,0.294634372,-0.3353052735,-0.110161528,-0.2738885283,0.149237141,0.1907613426,0.033480823,0.2015863955,0.0069947504,0.058482565,-0.0647767931,0.1245979741,0.1289161146,-0.1045876369,-0.1857670695,0.6569797397,0.0692303404,-0.3637074232,0.1040506735,0.1762638688,-0.3564894795,0.0153637268,0.2813159227,0.238305375,-0.0361465551,0.0326536335,0.0407718979,-0.1304047555,-0.0026093845,0.5073683262,-0.439085722,-0.2159501761,-0.083373405,-0.5888130069,-0.0206927042,0.1485571265,-0.3750435114,0.1395829469,0.0455849618,-0.0831533447,0.1431595236,-0.1116364747,-0.2610347271,-0.1053233817,-0.1459404528,-0.1304328293,-0.3109001517,-0.3554877639,-0.2258844674,-0.202214241,-0.1785319895,-0.0908836499,0.0453533381,0.0243956372,-0.1597588509,-0.1778942794,0.1169612333,-0.2261746973,0.4319359362,0.3429265618,0.4568071663,-0.0935688764,-0.1591013968,0.093947731,-0.1951591969,-0.2731350362,0.1776688695,0.0527599342,0.5685744286,0.0545593575,-0.1818611026,-0.2560166717,0.1016356423,-0.0009115308,-0.0363247953,-0.6818224788,0.3935556412,-0.2956717908,0.1442215443,0.1971506625,-0.0676475614,-0.0124600008,0.2009128332,-0.2974279523,-0.306786865,0.4448246062,-0.1353919655,-0.1814392656,-0.182621181,0.46392259,-0.3398220539,-0.065836668,-0.7032311559,0.0131412279,0.2306145877,0.0301840845,-0.332703799,0.2300281972,0.2124577016,0.0812509134,0.1095385626,-0.2603391409,-0.1833191216,-0.0552592017,0.0039653582,0.028439153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/6","title":"Error when citation is not given in the DatasetInfo","comments":"No, problem ^^ It might just be a temporary fix :)","body":"The following error is raised when the `citation` parameter is missing when we instantiate a `DatasetInfo`:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/info.py\", line 338, in __repr__\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\nAttributeError: 'NoneType' object has no attribute 'strip'\r\n```\r\n\r\nI propose to do the following change in the `info.py` file. The method:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nBecomes:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n ## the strip is done only is the citation is given\r\n citation_pprint = self.citation\r\n\r\n if self.citation:\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nAnd now it is ok. @thomwolf are you ok with this fix?","comment_length":11,"text":"Error when citation is not given in the DatasetInfo\nThe following error is raised when the `citation` parameter is missing when we instantiate a `DatasetInfo`:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/info.py\", line 338, in __repr__\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\nAttributeError: 'NoneType' object has no attribute 'strip'\r\n```\r\n\r\nI propose to do the following change in the `info.py` file. The method:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nBecomes:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n ## the strip is done only is the citation is given\r\n citation_pprint = self.citation\r\n\r\n if self.citation:\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nAnd now it is ok. @thomwolf are you ok with this fix?\nNo, problem ^^ It might just be a temporary fix :)","embeddings":[0.2855120897,0.0732322857,0.1477971971,0.1601161659,0.2827774882,0.2352983654,0.2543569803,0.4805189073,-0.4466493428,0.16624026,0.403686285,0.3917041719,0.0535713695,-0.203330785,-0.0758434907,-0.3508134186,-0.1062086448,0.2190288305,0.3720492125,-0.0522521213,-0.3091630936,0.1654009521,-0.3137940764,-0.0903745443,-0.2572643757,-0.1726606786,-0.2614687681,0.2065153718,-0.3102674484,-0.5589748025,0.1857405305,0.063550368,0.0325638615,0.0146645615,-0.0001185848,0.0302900337,0.3959799111,-0.1021313891,-0.2354656905,-0.2480834574,-0.5433374643,-0.2237135321,-0.029977249,-0.3961987495,-0.1640129983,-0.0125304721,-0.1367189437,0.1232259646,0.1175461933,0.2436022162,0.1624839008,0.0612453744,0.0864233896,-0.3654048443,0.0225184187,0.4103301764,-0.0947481841,-0.2211684287,-0.0262204297,0.0773107931,0.0846517608,0.2534255385,-0.196715951,-0.1238166988,0.4814615846,-0.0951773897,-0.0258708559,-0.0744219944,-0.0097684907,0.1676390171,0.5024431944,-0.2813969254,-0.5132006407,-0.0163041782,0.2781536281,-0.0212765168,-0.0074698143,0.0304121226,-0.0812080204,0.0895343795,0.2815128863,-0.2978960872,-0.0270233434,-0.1198625565,0.0443734415,0.2116338611,0.0791531727,0.1897022575,0.006169613,-0.1191703305,-0.1341097206,-0.0683485717,-0.2433660775,0.1042299494,-0.0947714001,-0.2623224258,0.3013670146,-0.1240373775,0.3263944983,0.2747790217,-0.0126671297,-0.3645196557,0.1062225774,0.1398010999,0.2596132457,0.0484170578,-0.0095627923,0.2912044525,0.3808363378,-0.2724874616,-0.0530267395,-0.0224923305,0.1467389613,-0.0760321692,0.1871124804,0.3163543642,0.6639338136,-0.2459217012,-0.2278142571,0.0181359351,-0.4110317528,0.0834341347,-0.0681851432,0.1200495139,0.0531694703,0.1811803877,0.0788510516,0.1461924911,0.0442111008,-0.3349076211,-0.3145993948,0.2601854205,-0.141174674,-0.1581093818,0.0291049238,0.3195924461,0.3016165793,0.0878599286,-0.2863543034,-0.1176711768,0.0519980602,-0.3555636108,-0.15103136,0.2481955886,-0.1378612816,0.0533519797,0.3309005797,-0.3365867436,-0.207349807,0.5098431706,-0.2286172807,-0.3707093,-0.1757764667,0.1305076778,-0.2939586639,-0.1857750118,-0.0267878566,0.1469056308,0.0339358598,-0.2264343053,-0.1586720645,-0.1596159935,-0.1011698768,-0.1808520108,0.210424751,0.7028290629,-0.2065771967,-0.2376949489,-0.1792689413,0.0066007464,0.3217360973,0.0637401119,-0.2051242739,0.4242260456,-0.1891367584,0.309779495,0.4551816881,-0.5058348179,-0.264592886,0.0924514309,-0.090154022,-0.1495330036,0.3402810097,0.0750064477,0.3930941522,-0.3331330121,0.1752074659,0.1514261812,-0.1647031307,-0.0128044328,-0.2054263502,-0.025216924,0.2687395513,-0.0305352192,0.1777623147,0.4301315248,0.026665451,0.23543787,0.0071865981,-0.0845381394,0.09592361,0.6712453961,0.4971642196,-0.0815782472,0.0407697707,0.0891511142,0.1425515115,-0.0065857228,-0.0049517527,0.5135893822,-0.4911940694,-0.3580025136,-0.2441371828,-0.0477071702,0.0384993032,-0.128395915,0.0801430494,0.1488384604,0.2277501523,0.153040424,-0.1139558703,0.0341832377,-0.3096472323,0.1729861945,-0.2149274796,0.0709076151,-0.0962565392,-0.1759302914,0.2825001478,0.4030955136,0.3565137982,0.1395501345,-0.2577136159,0.2078028768,0.3365240097,-0.0840731561,0.3318354785,-0.0264528077,0.2214430571,-0.5601446629,-0.106444627,0.1312949508,0.2259838134,-0.0394202173,-0.1677945852,0.1209984645,-0.0971575826,0.1369831562,0.0983100757,0.2548840344,0.2972361147,-0.1475794017,-0.2540400028,-0.1345717907,-0.4073702097,0.3577827811,0.0442956835,0.1807568073,-0.3323710263,-0.06579981,0.2814854383,0.0483954698,-0.083436124,0.0813504905,-0.2417711169,-0.0494092368,0.0345413573,0.504362762,0.558821857,0.2867323756,-0.1355875134,0.0329405963,0.0807511434,0.0325014032,0.1965010613,0.172426492,0.0140458588,0.1181470826,0.0886255354,0.264751792,-0.0481837392,-0.3537189364,0.2952138186,0.4577527642,-0.2807684243,-0.150439322,-0.3010823727,-0.3772722483,-0.2303933054,-0.1566966474,0.1558211148,-0.2895430624,-0.1114748344,0.0278519038,-0.226184845,0.2793827951,-0.2176556587,0.0751801282,0.1010244191,-0.2260205597,-0.0099929068,-0.1933918744,0.0566482544,-0.0408326052,-0.0482504629,0.2714774609,0.2608515024,0.0312365536,0.0936894119,-0.4852143526,-0.499157995,0.185065195,-0.2016006261,0.2247467786,0.3787457943,0.254168421,-0.056391865,-0.1754277796,0.3002544045,0.1890562177,-0.0711904615,0.2082998604,-0.2221645266,-0.1809340864,-0.1976263523,-0.4478684068,-0.1524800062,-0.3519616425,-0.324232161,0.0325689949,-0.1595597714,0.5114241242,-0.0376598351,0.107879214,0.4586960673,0.3906828761,0.0442582294,0.0350523777,0.3263026476,-0.1154825166,-0.3304879963,0.1488804519,-0.2420458794,0.050730776,0.1023525372,-0.1610190421,-0.4054076076,0.085415177,0.2321101427,-0.0968379304,0.2345744967,0.3909744322,0.2886578441,-0.1030012518,-0.1418347955,-0.2641219795,0.0721986815,0.0509948917,0.2560450733,0.161925137,0.1512395144,-0.0068999906,0.3923106492,0.1909050792,0.0996040404,0.221127972,-0.0202333014,0.1393013299,0.0169179533,-0.0862640515,-0.037109077,0.2619738877,-0.1297571361,0.0234721079,-0.0368065387,-0.1350055486,-0.0652676299,-0.2120331526,-0.2937174439,-0.2581718266,-0.1676616669,-0.2041821033,0.5298532844,0.1539096385,0.0566982999,-0.2150213122,0.1116406694,0.0398819856,0.3978654742,0.4186894894,-0.0655552149,-0.1002212837,-0.1168369129,-0.4247890711,0.0079289991,0.0285914894,0.289878875,-0.0292112362,-0.118423447,0.1176300943,0.1090158671,0.4440200627,-0.3201776147,-0.094239153,0.2171335667,-0.0520307347,-0.0749424621,0.0966636464,-0.1407677382,0.3265134692,0.1625849009,-0.0012651245,-0.3431600332,-0.1991199255,0.5652010441,0.0161457453,-0.0239477362,0.0118602039,-0.0996702164,-0.5567368269,-0.2893752158,0.007464183,-0.0616957918,0.3286765516,0.4054490328,0.0357954055,-0.3641443551,-0.0787933394,0.0652355626,0.1031478569,0.4205913842,-0.0471299216,0.1016341671,-0.2359844148,-0.1014064476,0.1652707905,0.2810600698,-0.0035395133,-0.4454729557,-0.1814178377,-0.0911664367,0.2103846818,0.0349834226,-0.1234335527,0.2475430071,0.0253230613,-0.0295028333,-0.1880521923,-0.3054905832,0.177183643,-0.217938289,-0.3578876555,-0.4102500975,0.0567891188,0.2962569296,0.003355328,0.1699745357,0.6390137076,-0.3207844198,0.3601374328,0.0156328343,1.0747797489,0.2924572229,-0.2725102603,0.2915841639,0.0433302224,0.5590177774,0.235897705,0.2341372669,-0.4298227131,-0.3786416948,-0.0747423917,-0.0519499518,0.2784281969,-0.0809192508,-0.5020384192,0.1756699979,-0.1523383707,0.0022322566,0.1220385805,0.0872596353,0.0060309148,-0.2312285155,-0.2017602473,0.0670160204,-0.1527360678,0.3128901124,-0.085725978,-0.243533656,-0.1070018113,-0.3578900099,-0.2665623128,0.000804026,-0.018101396,0.2025057226,0.1674943566,-0.2818686068,-0.0193199199,0.2606441677,0.3911963999,0.1438509226,-0.1460814923,0.1348209083,-0.3412264287,0.0674619228,0.0010561617,0.3297884166,0.6635107398,0.0570278242,-0.2791726291,0.0722952187,-0.3045301437,-0.2996293604,0.2755072117,-0.3606644571,-0.2726466358,-0.1459128559,-0.3561815023,0.1865038574,-0.1307550222,-0.3097569942,0.1204930171,0.1125206426,-0.2872116864,0.1730938405,0.0209985767,-0.1482337713,0.088712588,0.1320442706,0.4319786429,0.5363356471,0.352291584,-0.0066535491,0.0746256039,-0.1703940928,0.0755943954,-0.2000356019,-0.2088153064,0.4038462937,0.0661254823,0.328760922,-0.0338510461,0.3243108392,-0.147584945,0.211573422,-0.245580107,-0.3820871413,-0.1751209497,0.2641634643,-0.1494313031,-0.0163528882,-0.0356309339,0.239595592,0.0800135434,-0.148766011,-0.2949716449,-0.00972299,-0.4209564328,0.2777429521,0.3176947534,0.2704794109,0.0551025309,-0.0695449412,0.0922539011,0.0423996262,0.0950587615,-0.0792630613,-0.3869243264,0.1389072239,0.1080889925,0.2912574112,-0.0833027735,-0.1454300731,-0.1910989136,-0.0237107761,0.0475815721,0.2399301529,-0.0320061967,0.3826493323,-0.3793178201,0.1382137537,-0.2307718247,0.0374536328,0.114278093,0.1407286227,0.0325585678,0.1824617386,0.0813172907,-0.0186538789,-0.5630632043,0.1317278445,-0.1979646534,-0.1241242513,0.3146758676,0.009368144,0.1651304662,0.2314257771,-0.014451636,-0.0592810959,-0.1193889305,0.2787908614,0.0533665046,0.14934735,-0.4110959172,-0.0266342256,0.1290837824,-0.1473093182,0.1387932301,0.2995389104,0.0316387713,0.0877545178,-0.0951115862,-0.1150851697,0.7280268073,-0.3307806253,0.1762421429,0.4804491401,0.0208486728,0.1105091348,-0.0597954094,0.1357881129,0.1895601004,0.1926511824,0.0215189178,0.3219555616,0.0450908169,-0.1167576984,0.0563818365,-0.2132242769,0.7372803092,0.4240525365,0.280253619,-0.2031773329,0.0000221593,0.3681063056,-0.1852521598,-0.0167535301,-0.0322262719,0.1175446063,-0.1886272877,0.1395931393,-0.3298102915,-0.0430390052,-0.203642711,-0.2156832516,0.119598411,0.0806257576,0.0927326381,-0.1072928235,-0.0609588064,-0.425447315,0.3382223547,0.2237199843,0.111854054,-0.3376896977,-0.1360161602,-0.1737665087,-0.1550251991,0.1986856312,0.6460391879,0.380517453,0.2580200434,0.294634372,-0.3353052735,-0.110161528,-0.2738885283,0.149237141,0.1907613426,0.033480823,0.2015863955,0.0069947504,0.058482565,-0.0647767931,0.1245979741,0.1289161146,-0.1045876369,-0.1857670695,0.6569797397,0.0692303404,-0.3637074232,0.1040506735,0.1762638688,-0.3564894795,0.0153637268,0.2813159227,0.238305375,-0.0361465551,0.0326536335,0.0407718979,-0.1304047555,-0.0026093845,0.5073683262,-0.439085722,-0.2159501761,-0.083373405,-0.5888130069,-0.0206927042,0.1485571265,-0.3750435114,0.1395829469,0.0455849618,-0.0831533447,0.1431595236,-0.1116364747,-0.2610347271,-0.1053233817,-0.1459404528,-0.1304328293,-0.3109001517,-0.3554877639,-0.2258844674,-0.202214241,-0.1785319895,-0.0908836499,0.0453533381,0.0243956372,-0.1597588509,-0.1778942794,0.1169612333,-0.2261746973,0.4319359362,0.3429265618,0.4568071663,-0.0935688764,-0.1591013968,0.093947731,-0.1951591969,-0.2731350362,0.1776688695,0.0527599342,0.5685744286,0.0545593575,-0.1818611026,-0.2560166717,0.1016356423,-0.0009115308,-0.0363247953,-0.6818224788,0.3935556412,-0.2956717908,0.1442215443,0.1971506625,-0.0676475614,-0.0124600008,0.2009128332,-0.2974279523,-0.306786865,0.4448246062,-0.1353919655,-0.1814392656,-0.182621181,0.46392259,-0.3398220539,-0.065836668,-0.7032311559,0.0131412279,0.2306145877,0.0301840845,-0.332703799,0.2300281972,0.2124577016,0.0812509134,0.1095385626,-0.2603391409,-0.1833191216,-0.0552592017,0.0039653582,0.028439153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/6","title":"Error when citation is not given in the DatasetInfo","comments":"Fixed.","body":"The following error is raised when the `citation` parameter is missing when we instantiate a `DatasetInfo`:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/info.py\", line 338, in __repr__\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\nAttributeError: 'NoneType' object has no attribute 'strip'\r\n```\r\n\r\nI propose to do the following change in the `info.py` file. The method:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nBecomes:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n ## the strip is done only is the citation is given\r\n citation_pprint = self.citation\r\n\r\n if self.citation:\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nAnd now it is ok. @thomwolf are you ok with this fix?","comment_length":1,"text":"Error when citation is not given in the DatasetInfo\nThe following error is raised when the `citation` parameter is missing when we instantiate a `DatasetInfo`:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/info.py\", line 338, in __repr__\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\nAttributeError: 'NoneType' object has no attribute 'strip'\r\n```\r\n\r\nI propose to do the following change in the `info.py` file. The method:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nBecomes:\r\n```python\r\ndef __repr__(self):\r\n splits_pprint = _indent(\"\\n\".join([\"{\"] + [\r\n \" '{}': {},\".format(k, split.num_examples)\r\n for k, split in sorted(self.splits.items())\r\n ] + [\"}\"]))\r\n features_pprint = _indent(repr(self.features))\r\n ## the strip is done only is the citation is given\r\n citation_pprint = self.citation\r\n\r\n if self.citation:\r\n citation_pprint = _indent('\"\"\"{}\"\"\"'.format(self.citation.strip()))\r\n return INFO_STR.format(\r\n name=self.name,\r\n version=self.version,\r\n description=self.description,\r\n total_num_examples=self.splits.total_num_examples,\r\n features=features_pprint,\r\n splits=splits_pprint,\r\n citation=citation_pprint,\r\n homepage=self.homepage,\r\n supervised_keys=self.supervised_keys,\r\n # Proto add a \\n that we strip.\r\n license=str(self.license).strip())\r\n```\r\nAnd now it is ok. @thomwolf are you ok with this fix?\nFixed.","embeddings":[0.2855120897,0.0732322857,0.1477971971,0.1601161659,0.2827774882,0.2352983654,0.2543569803,0.4805189073,-0.4466493428,0.16624026,0.403686285,0.3917041719,0.0535713695,-0.203330785,-0.0758434907,-0.3508134186,-0.1062086448,0.2190288305,0.3720492125,-0.0522521213,-0.3091630936,0.1654009521,-0.3137940764,-0.0903745443,-0.2572643757,-0.1726606786,-0.2614687681,0.2065153718,-0.3102674484,-0.5589748025,0.1857405305,0.063550368,0.0325638615,0.0146645615,-0.0001185848,0.0302900337,0.3959799111,-0.1021313891,-0.2354656905,-0.2480834574,-0.5433374643,-0.2237135321,-0.029977249,-0.3961987495,-0.1640129983,-0.0125304721,-0.1367189437,0.1232259646,0.1175461933,0.2436022162,0.1624839008,0.0612453744,0.0864233896,-0.3654048443,0.0225184187,0.4103301764,-0.0947481841,-0.2211684287,-0.0262204297,0.0773107931,0.0846517608,0.2534255385,-0.196715951,-0.1238166988,0.4814615846,-0.0951773897,-0.0258708559,-0.0744219944,-0.0097684907,0.1676390171,0.5024431944,-0.2813969254,-0.5132006407,-0.0163041782,0.2781536281,-0.0212765168,-0.0074698143,0.0304121226,-0.0812080204,0.0895343795,0.2815128863,-0.2978960872,-0.0270233434,-0.1198625565,0.0443734415,0.2116338611,0.0791531727,0.1897022575,0.006169613,-0.1191703305,-0.1341097206,-0.0683485717,-0.2433660775,0.1042299494,-0.0947714001,-0.2623224258,0.3013670146,-0.1240373775,0.3263944983,0.2747790217,-0.0126671297,-0.3645196557,0.1062225774,0.1398010999,0.2596132457,0.0484170578,-0.0095627923,0.2912044525,0.3808363378,-0.2724874616,-0.0530267395,-0.0224923305,0.1467389613,-0.0760321692,0.1871124804,0.3163543642,0.6639338136,-0.2459217012,-0.2278142571,0.0181359351,-0.4110317528,0.0834341347,-0.0681851432,0.1200495139,0.0531694703,0.1811803877,0.0788510516,0.1461924911,0.0442111008,-0.3349076211,-0.3145993948,0.2601854205,-0.141174674,-0.1581093818,0.0291049238,0.3195924461,0.3016165793,0.0878599286,-0.2863543034,-0.1176711768,0.0519980602,-0.3555636108,-0.15103136,0.2481955886,-0.1378612816,0.0533519797,0.3309005797,-0.3365867436,-0.207349807,0.5098431706,-0.2286172807,-0.3707093,-0.1757764667,0.1305076778,-0.2939586639,-0.1857750118,-0.0267878566,0.1469056308,0.0339358598,-0.2264343053,-0.1586720645,-0.1596159935,-0.1011698768,-0.1808520108,0.210424751,0.7028290629,-0.2065771967,-0.2376949489,-0.1792689413,0.0066007464,0.3217360973,0.0637401119,-0.2051242739,0.4242260456,-0.1891367584,0.309779495,0.4551816881,-0.5058348179,-0.264592886,0.0924514309,-0.090154022,-0.1495330036,0.3402810097,0.0750064477,0.3930941522,-0.3331330121,0.1752074659,0.1514261812,-0.1647031307,-0.0128044328,-0.2054263502,-0.025216924,0.2687395513,-0.0305352192,0.1777623147,0.4301315248,0.026665451,0.23543787,0.0071865981,-0.0845381394,0.09592361,0.6712453961,0.4971642196,-0.0815782472,0.0407697707,0.0891511142,0.1425515115,-0.0065857228,-0.0049517527,0.5135893822,-0.4911940694,-0.3580025136,-0.2441371828,-0.0477071702,0.0384993032,-0.128395915,0.0801430494,0.1488384604,0.2277501523,0.153040424,-0.1139558703,0.0341832377,-0.3096472323,0.1729861945,-0.2149274796,0.0709076151,-0.0962565392,-0.1759302914,0.2825001478,0.4030955136,0.3565137982,0.1395501345,-0.2577136159,0.2078028768,0.3365240097,-0.0840731561,0.3318354785,-0.0264528077,0.2214430571,-0.5601446629,-0.106444627,0.1312949508,0.2259838134,-0.0394202173,-0.1677945852,0.1209984645,-0.0971575826,0.1369831562,0.0983100757,0.2548840344,0.2972361147,-0.1475794017,-0.2540400028,-0.1345717907,-0.4073702097,0.3577827811,0.0442956835,0.1807568073,-0.3323710263,-0.06579981,0.2814854383,0.0483954698,-0.083436124,0.0813504905,-0.2417711169,-0.0494092368,0.0345413573,0.504362762,0.558821857,0.2867323756,-0.1355875134,0.0329405963,0.0807511434,0.0325014032,0.1965010613,0.172426492,0.0140458588,0.1181470826,0.0886255354,0.264751792,-0.0481837392,-0.3537189364,0.2952138186,0.4577527642,-0.2807684243,-0.150439322,-0.3010823727,-0.3772722483,-0.2303933054,-0.1566966474,0.1558211148,-0.2895430624,-0.1114748344,0.0278519038,-0.226184845,0.2793827951,-0.2176556587,0.0751801282,0.1010244191,-0.2260205597,-0.0099929068,-0.1933918744,0.0566482544,-0.0408326052,-0.0482504629,0.2714774609,0.2608515024,0.0312365536,0.0936894119,-0.4852143526,-0.499157995,0.185065195,-0.2016006261,0.2247467786,0.3787457943,0.254168421,-0.056391865,-0.1754277796,0.3002544045,0.1890562177,-0.0711904615,0.2082998604,-0.2221645266,-0.1809340864,-0.1976263523,-0.4478684068,-0.1524800062,-0.3519616425,-0.324232161,0.0325689949,-0.1595597714,0.5114241242,-0.0376598351,0.107879214,0.4586960673,0.3906828761,0.0442582294,0.0350523777,0.3263026476,-0.1154825166,-0.3304879963,0.1488804519,-0.2420458794,0.050730776,0.1023525372,-0.1610190421,-0.4054076076,0.085415177,0.2321101427,-0.0968379304,0.2345744967,0.3909744322,0.2886578441,-0.1030012518,-0.1418347955,-0.2641219795,0.0721986815,0.0509948917,0.2560450733,0.161925137,0.1512395144,-0.0068999906,0.3923106492,0.1909050792,0.0996040404,0.221127972,-0.0202333014,0.1393013299,0.0169179533,-0.0862640515,-0.037109077,0.2619738877,-0.1297571361,0.0234721079,-0.0368065387,-0.1350055486,-0.0652676299,-0.2120331526,-0.2937174439,-0.2581718266,-0.1676616669,-0.2041821033,0.5298532844,0.1539096385,0.0566982999,-0.2150213122,0.1116406694,0.0398819856,0.3978654742,0.4186894894,-0.0655552149,-0.1002212837,-0.1168369129,-0.4247890711,0.0079289991,0.0285914894,0.289878875,-0.0292112362,-0.118423447,0.1176300943,0.1090158671,0.4440200627,-0.3201776147,-0.094239153,0.2171335667,-0.0520307347,-0.0749424621,0.0966636464,-0.1407677382,0.3265134692,0.1625849009,-0.0012651245,-0.3431600332,-0.1991199255,0.5652010441,0.0161457453,-0.0239477362,0.0118602039,-0.0996702164,-0.5567368269,-0.2893752158,0.007464183,-0.0616957918,0.3286765516,0.4054490328,0.0357954055,-0.3641443551,-0.0787933394,0.0652355626,0.1031478569,0.4205913842,-0.0471299216,0.1016341671,-0.2359844148,-0.1014064476,0.1652707905,0.2810600698,-0.0035395133,-0.4454729557,-0.1814178377,-0.0911664367,0.2103846818,0.0349834226,-0.1234335527,0.2475430071,0.0253230613,-0.0295028333,-0.1880521923,-0.3054905832,0.177183643,-0.217938289,-0.3578876555,-0.4102500975,0.0567891188,0.2962569296,0.003355328,0.1699745357,0.6390137076,-0.3207844198,0.3601374328,0.0156328343,1.0747797489,0.2924572229,-0.2725102603,0.2915841639,0.0433302224,0.5590177774,0.235897705,0.2341372669,-0.4298227131,-0.3786416948,-0.0747423917,-0.0519499518,0.2784281969,-0.0809192508,-0.5020384192,0.1756699979,-0.1523383707,0.0022322566,0.1220385805,0.0872596353,0.0060309148,-0.2312285155,-0.2017602473,0.0670160204,-0.1527360678,0.3128901124,-0.085725978,-0.243533656,-0.1070018113,-0.3578900099,-0.2665623128,0.000804026,-0.018101396,0.2025057226,0.1674943566,-0.2818686068,-0.0193199199,0.2606441677,0.3911963999,0.1438509226,-0.1460814923,0.1348209083,-0.3412264287,0.0674619228,0.0010561617,0.3297884166,0.6635107398,0.0570278242,-0.2791726291,0.0722952187,-0.3045301437,-0.2996293604,0.2755072117,-0.3606644571,-0.2726466358,-0.1459128559,-0.3561815023,0.1865038574,-0.1307550222,-0.3097569942,0.1204930171,0.1125206426,-0.2872116864,0.1730938405,0.0209985767,-0.1482337713,0.088712588,0.1320442706,0.4319786429,0.5363356471,0.352291584,-0.0066535491,0.0746256039,-0.1703940928,0.0755943954,-0.2000356019,-0.2088153064,0.4038462937,0.0661254823,0.328760922,-0.0338510461,0.3243108392,-0.147584945,0.211573422,-0.245580107,-0.3820871413,-0.1751209497,0.2641634643,-0.1494313031,-0.0163528882,-0.0356309339,0.239595592,0.0800135434,-0.148766011,-0.2949716449,-0.00972299,-0.4209564328,0.2777429521,0.3176947534,0.2704794109,0.0551025309,-0.0695449412,0.0922539011,0.0423996262,0.0950587615,-0.0792630613,-0.3869243264,0.1389072239,0.1080889925,0.2912574112,-0.0833027735,-0.1454300731,-0.1910989136,-0.0237107761,0.0475815721,0.2399301529,-0.0320061967,0.3826493323,-0.3793178201,0.1382137537,-0.2307718247,0.0374536328,0.114278093,0.1407286227,0.0325585678,0.1824617386,0.0813172907,-0.0186538789,-0.5630632043,0.1317278445,-0.1979646534,-0.1241242513,0.3146758676,0.009368144,0.1651304662,0.2314257771,-0.014451636,-0.0592810959,-0.1193889305,0.2787908614,0.0533665046,0.14934735,-0.4110959172,-0.0266342256,0.1290837824,-0.1473093182,0.1387932301,0.2995389104,0.0316387713,0.0877545178,-0.0951115862,-0.1150851697,0.7280268073,-0.3307806253,0.1762421429,0.4804491401,0.0208486728,0.1105091348,-0.0597954094,0.1357881129,0.1895601004,0.1926511824,0.0215189178,0.3219555616,0.0450908169,-0.1167576984,0.0563818365,-0.2132242769,0.7372803092,0.4240525365,0.280253619,-0.2031773329,0.0000221593,0.3681063056,-0.1852521598,-0.0167535301,-0.0322262719,0.1175446063,-0.1886272877,0.1395931393,-0.3298102915,-0.0430390052,-0.203642711,-0.2156832516,0.119598411,0.0806257576,0.0927326381,-0.1072928235,-0.0609588064,-0.425447315,0.3382223547,0.2237199843,0.111854054,-0.3376896977,-0.1360161602,-0.1737665087,-0.1550251991,0.1986856312,0.6460391879,0.380517453,0.2580200434,0.294634372,-0.3353052735,-0.110161528,-0.2738885283,0.149237141,0.1907613426,0.033480823,0.2015863955,0.0069947504,0.058482565,-0.0647767931,0.1245979741,0.1289161146,-0.1045876369,-0.1857670695,0.6569797397,0.0692303404,-0.3637074232,0.1040506735,0.1762638688,-0.3564894795,0.0153637268,0.2813159227,0.238305375,-0.0361465551,0.0326536335,0.0407718979,-0.1304047555,-0.0026093845,0.5073683262,-0.439085722,-0.2159501761,-0.083373405,-0.5888130069,-0.0206927042,0.1485571265,-0.3750435114,0.1395829469,0.0455849618,-0.0831533447,0.1431595236,-0.1116364747,-0.2610347271,-0.1053233817,-0.1459404528,-0.1304328293,-0.3109001517,-0.3554877639,-0.2258844674,-0.202214241,-0.1785319895,-0.0908836499,0.0453533381,0.0243956372,-0.1597588509,-0.1778942794,0.1169612333,-0.2261746973,0.4319359362,0.3429265618,0.4568071663,-0.0935688764,-0.1591013968,0.093947731,-0.1951591969,-0.2731350362,0.1776688695,0.0527599342,0.5685744286,0.0545593575,-0.1818611026,-0.2560166717,0.1016356423,-0.0009115308,-0.0363247953,-0.6818224788,0.3935556412,-0.2956717908,0.1442215443,0.1971506625,-0.0676475614,-0.0124600008,0.2009128332,-0.2974279523,-0.306786865,0.4448246062,-0.1353919655,-0.1814392656,-0.182621181,0.46392259,-0.3398220539,-0.065836668,-0.7032311559,0.0131412279,0.2306145877,0.0301840845,-0.332703799,0.2300281972,0.2124577016,0.0812509134,0.1095385626,-0.2603391409,-0.1833191216,-0.0552592017,0.0039653582,0.028439153]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/5","title":"ValueError when a split is empty","comments":"To fix this I propose to modify only the file `arrow_reader.py` with few updates. First update, the following method:\r\n```python\r\ndef _make_file_instructions_from_absolutes(\r\n name,\r\n name2len,\r\n absolute_instructions,\r\n):\r\n \"\"\"Returns the files instructions from the absolute instructions list.\"\"\"\r\n # For each split, return the files instruction (skip\/take)\r\n file_instructions = []\r\n num_examples = 0\r\n for abs_instr in absolute_instructions:\r\n length = name2len[abs_instr.splitname]\r\n if not length:\r\n raise ValueError(\r\n 'Split empty. This might means that dataset hasn\\'t been generated '\r\n 'yet and info not restored from GCS, or that legacy dataset is used.')\r\n filename = filename_for_dataset_split(\r\n dataset_name=name,\r\n split=abs_instr.splitname,\r\n filetype_suffix='arrow')\r\n from_ = 0 if abs_instr.from_ is None else abs_instr.from_\r\n to = length if abs_instr.to is None else abs_instr.to\r\n num_examples += to - from_\r\n single_file_instructions = [{\"filename\": filename, \"skip\": from_, \"take\": to - from_}]\r\n file_instructions.extend(single_file_instructions)\r\n return FileInstructions(\r\n num_examples=num_examples,\r\n file_instructions=file_instructions,\r\n )\r\n```\r\nBecomes:\r\n```python\r\ndef _make_file_instructions_from_absolutes(\r\n name,\r\n name2len,\r\n absolute_instructions,\r\n):\r\n \"\"\"Returns the files instructions from the absolute instructions list.\"\"\"\r\n # For each split, return the files instruction (skip\/take)\r\n file_instructions = []\r\n num_examples = 0\r\n for abs_instr in absolute_instructions:\r\n length = name2len[abs_instr.splitname]\r\n ## Delete the if not length and the raise\r\n filename = filename_for_dataset_split(\r\n dataset_name=name,\r\n split=abs_instr.splitname,\r\n filetype_suffix='arrow')\r\n from_ = 0 if abs_instr.from_ is None else abs_instr.from_\r\n to = length if abs_instr.to is None else abs_instr.to\r\n num_examples += to - from_\r\n single_file_instructions = [{\"filename\": filename, \"skip\": from_, \"take\": to - from_}]\r\n file_instructions.extend(single_file_instructions)\r\n return FileInstructions(\r\n num_examples=num_examples,\r\n file_instructions=file_instructions,\r\n )\r\n```\r\n\r\nSecond update the following method:\r\n```python\r\ndef _read_files(files, info):\r\n \"\"\"Returns Dataset for given file instructions.\r\n\r\n Args:\r\n files: List[dict(filename, skip, take)], the files information.\r\n The filenames contain the absolute path, not relative.\r\n skip\/take indicates which example read in the file: `ds.slice(skip, take)`\r\n \"\"\"\r\n pa_batches = []\r\n for f_dict in files:\r\n pa_table: pa.Table = _get_dataset_from_filename(f_dict)\r\n pa_batches.extend(pa_table.to_batches())\r\n pa_table = pa.Table.from_batches(pa_batches)\r\n ds = Dataset(arrow_table=pa_table, data_files=files, info=info)\r\n return ds\r\n```\r\nBecomes:\r\n```python\r\ndef _read_files(files, info):\r\n \"\"\"Returns Dataset for given file instructions.\r\n\r\n Args:\r\n files: List[dict(filename, skip, take)], the files information.\r\n The filenames contain the absolute path, not relative.\r\n skip\/take indicates which example read in the file: `ds.slice(skip, take)`\r\n \"\"\"\r\n pa_batches = []\r\n for f_dict in files:\r\n pa_table: pa.Table = _get_dataset_from_filename(f_dict)\r\n pa_batches.extend(pa_table.to_batches())\r\n ## we modify the table only if there are some batches\r\n if pa_batches:\r\n pa_table = pa.Table.from_batches(pa_batches)\r\n ds = Dataset(arrow_table=pa_table, data_files=files, info=info)\r\n return ds\r\n```\r\n\r\nWith these two updates it works now. @thomwolf are you ok with this changes?","body":"When a split is empty either TEST, VALIDATION or TRAIN I get the following error:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 295, in load\r\n ds = dbuilder.as_dataset(**as_dataset_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 587, in as_dataset\r\n datasets = utils.map_nested(build_single_dataset, split, map_tuple=True)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in map_nested\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in <dictcomp>\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 601, in _build_single_dataset\r\n split=split,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 625, in _as_dataset\r\n split_infos=self.info.splits.values(),\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 200, in read\r\n return py_utils.map_nested(_read_instruction_to_ds, instructions)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 191, in _read_instruction_to_ds\r\n file_instructions = make_file_instructions(name, split_infos, instruction)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 104, in make_file_instructions\r\n absolute_instructions=absolute_instructions,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 122, in _make_file_instructions_from_absolutes\r\n 'Split empty. This might means that dataset hasn\\'t been generated '\r\nValueError: Split empty. This might means that dataset hasn't been generated yet and info not restored from GCS, or that legacy dataset is used.\r\n``` \r\n\r\nHow to reproduce:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(self, **config):\r\n self.train = config.pop(\"train\", None)\r\n self.validation = config.pop(\"validation\", None)\r\n super(Bbc, self).__init__(**config)\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, description=\"bla\", features=nlp.features.FeaturesDict({\"id\": nlp.int32, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": self.train}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": self.validation}),\r\n nlp.SplitGenerator(name=nlp.Split.TEST, gen_kwargs={\"filepath\": None})]\r\n\r\n def _generate_examples(self, filepath):\r\n if not filepath:\r\n return None, {}\r\n\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"id\": idx, \"text\": line[1], \"label\": line[0]}\r\n```\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\"})\r\n```","comment_length":379,"text":"ValueError when a split is empty\nWhen a split is empty either TEST, VALIDATION or TRAIN I get the following error:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 295, in load\r\n ds = dbuilder.as_dataset(**as_dataset_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 587, in as_dataset\r\n datasets = utils.map_nested(build_single_dataset, split, map_tuple=True)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in map_nested\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in <dictcomp>\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 601, in _build_single_dataset\r\n split=split,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 625, in _as_dataset\r\n split_infos=self.info.splits.values(),\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 200, in read\r\n return py_utils.map_nested(_read_instruction_to_ds, instructions)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 191, in _read_instruction_to_ds\r\n file_instructions = make_file_instructions(name, split_infos, instruction)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 104, in make_file_instructions\r\n absolute_instructions=absolute_instructions,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 122, in _make_file_instructions_from_absolutes\r\n 'Split empty. This might means that dataset hasn\\'t been generated '\r\nValueError: Split empty. This might means that dataset hasn't been generated yet and info not restored from GCS, or that legacy dataset is used.\r\n``` \r\n\r\nHow to reproduce:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(self, **config):\r\n self.train = config.pop(\"train\", None)\r\n self.validation = config.pop(\"validation\", None)\r\n super(Bbc, self).__init__(**config)\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, description=\"bla\", features=nlp.features.FeaturesDict({\"id\": nlp.int32, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": self.train}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": self.validation}),\r\n nlp.SplitGenerator(name=nlp.Split.TEST, gen_kwargs={\"filepath\": None})]\r\n\r\n def _generate_examples(self, filepath):\r\n if not filepath:\r\n return None, {}\r\n\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"id\": idx, \"text\": line[1], \"label\": line[0]}\r\n```\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\"})\r\n```\nTo fix this I propose to modify only the file `arrow_reader.py` with few updates. First update, the following method:\r\n```python\r\ndef _make_file_instructions_from_absolutes(\r\n name,\r\n name2len,\r\n absolute_instructions,\r\n):\r\n \"\"\"Returns the files instructions from the absolute instructions list.\"\"\"\r\n # For each split, return the files instruction (skip\/take)\r\n file_instructions = []\r\n num_examples = 0\r\n for abs_instr in absolute_instructions:\r\n length = name2len[abs_instr.splitname]\r\n if not length:\r\n raise ValueError(\r\n 'Split empty. This might means that dataset hasn\\'t been generated '\r\n 'yet and info not restored from GCS, or that legacy dataset is used.')\r\n filename = filename_for_dataset_split(\r\n dataset_name=name,\r\n split=abs_instr.splitname,\r\n filetype_suffix='arrow')\r\n from_ = 0 if abs_instr.from_ is None else abs_instr.from_\r\n to = length if abs_instr.to is None else abs_instr.to\r\n num_examples += to - from_\r\n single_file_instructions = [{\"filename\": filename, \"skip\": from_, \"take\": to - from_}]\r\n file_instructions.extend(single_file_instructions)\r\n return FileInstructions(\r\n num_examples=num_examples,\r\n file_instructions=file_instructions,\r\n )\r\n```\r\nBecomes:\r\n```python\r\ndef _make_file_instructions_from_absolutes(\r\n name,\r\n name2len,\r\n absolute_instructions,\r\n):\r\n \"\"\"Returns the files instructions from the absolute instructions list.\"\"\"\r\n # For each split, return the files instruction (skip\/take)\r\n file_instructions = []\r\n num_examples = 0\r\n for abs_instr in absolute_instructions:\r\n length = name2len[abs_instr.splitname]\r\n ## Delete the if not length and the raise\r\n filename = filename_for_dataset_split(\r\n dataset_name=name,\r\n split=abs_instr.splitname,\r\n filetype_suffix='arrow')\r\n from_ = 0 if abs_instr.from_ is None else abs_instr.from_\r\n to = length if abs_instr.to is None else abs_instr.to\r\n num_examples += to - from_\r\n single_file_instructions = [{\"filename\": filename, \"skip\": from_, \"take\": to - from_}]\r\n file_instructions.extend(single_file_instructions)\r\n return FileInstructions(\r\n num_examples=num_examples,\r\n file_instructions=file_instructions,\r\n )\r\n```\r\n\r\nSecond update the following method:\r\n```python\r\ndef _read_files(files, info):\r\n \"\"\"Returns Dataset for given file instructions.\r\n\r\n Args:\r\n files: List[dict(filename, skip, take)], the files information.\r\n The filenames contain the absolute path, not relative.\r\n skip\/take indicates which example read in the file: `ds.slice(skip, take)`\r\n \"\"\"\r\n pa_batches = []\r\n for f_dict in files:\r\n pa_table: pa.Table = _get_dataset_from_filename(f_dict)\r\n pa_batches.extend(pa_table.to_batches())\r\n pa_table = pa.Table.from_batches(pa_batches)\r\n ds = Dataset(arrow_table=pa_table, data_files=files, info=info)\r\n return ds\r\n```\r\nBecomes:\r\n```python\r\ndef _read_files(files, info):\r\n \"\"\"Returns Dataset for given file instructions.\r\n\r\n Args:\r\n files: List[dict(filename, skip, take)], the files information.\r\n The filenames contain the absolute path, not relative.\r\n skip\/take indicates which example read in the file: `ds.slice(skip, take)`\r\n \"\"\"\r\n pa_batches = []\r\n for f_dict in files:\r\n pa_table: pa.Table = _get_dataset_from_filename(f_dict)\r\n pa_batches.extend(pa_table.to_batches())\r\n ## we modify the table only if there are some batches\r\n if pa_batches:\r\n pa_table = pa.Table.from_batches(pa_batches)\r\n ds = Dataset(arrow_table=pa_table, data_files=files, info=info)\r\n return ds\r\n```\r\n\r\nWith these two updates it works now. @thomwolf are you ok with this changes?","embeddings":[-0.2972342372,-0.0131540569,-0.103067942,0.2931129634,0.0207595415,-0.0219010543,0.6005375385,0.4749840498,-0.0488540977,0.4204655886,0.0286984555,0.2306589633,-0.3276998699,0.1830051839,-0.4715830386,-0.2210291326,-0.0620970726,0.4033246934,0.0316321403,0.0022019364,-0.2077705711,0.1781739295,-0.3489648104,0.2001751065,-0.4412805736,-0.3235142529,-0.0962038264,0.0336118527,-0.1655961424,-0.5333554745,0.2795565426,-0.315582782,-0.2108621746,0.4383441806,-0.0001089997,0.2471223772,0.3216657937,0.0212966055,-0.2349680364,-0.1730071604,-0.2148883939,0.0468612611,-0.0701180547,-0.1868547648,-0.1008067131,-0.2197800577,-0.3655616343,0.0298773181,0.2869431078,0.2550378144,0.1966028512,0.1572323292,0.0249177329,-0.1763806641,0.0856723338,-0.105694443,0.0970320776,0.0854190886,-0.1803481877,-0.2269463241,-0.1018456817,0.2128955573,-0.0124384062,0.0009043886,-0.2380745709,0.1348021328,0.2368194163,-0.3553718626,-0.1338815093,0.0872519761,-0.0105639342,-0.0024150251,-0.1980995387,-0.3114365339,0.0216911975,-0.3575952351,0.0597179197,0.2921065986,-0.1286965311,0.124095358,0.228433311,0.1736139357,-0.2168529481,-0.0249617305,-0.3034926653,0.3541932404,-0.031331487,0.0348216146,0.2476863563,-0.0247122962,0.0857289061,-0.0367395617,-0.218579337,0.115537703,-0.2246417701,-0.3038747907,-0.1532901824,-0.3125609457,-0.0919000655,-0.0317804776,0.027630873,-0.1767440885,0.1064907983,0.4088112414,0.304340452,-0.0534912385,0.3710521758,0.4308705032,0.0564235486,0.1503341496,0.0145205986,0.02915808,0.0300026834,-0.276969254,0.0414252691,0.1946728826,0.4200007617,0.0776329115,-0.3998927772,-0.0317487083,-0.2662963867,-0.0253004767,0.2187399864,0.377743125,0.0436375402,0.1135604605,-0.059801314,0.1989057064,-0.0594728105,-0.4426795244,-0.3030682206,0.3334758282,-0.1042798758,-0.3177535534,0.0193641763,-0.0207902119,0.1957095265,0.2191386372,-0.1739814132,-0.0592373982,0.1233527735,-0.4605067074,0.1437606066,0.3618825972,-0.0033455135,-0.2083796561,0.2578766048,-0.2094171792,0.0035285249,0.2536199987,-0.0682576448,-0.3559485078,-0.0321290009,0.337768048,0.1614764482,0.1738892049,-0.116377309,0.1625697762,0.3533759117,-0.2079333812,0.024277322,-0.085838072,-0.2200120091,-0.2932094634,0.1470615864,0.3619220257,-0.3413058817,-0.133083865,-0.1345353574,-0.1824868917,0.0983030871,0.0809265226,-0.1797881573,0.3938573897,-0.2386934012,0.5446959734,0.4584326446,-0.2361291498,-0.0809233189,0.3233044147,-0.0367279463,-0.2104673386,-0.0367180035,-0.1798265278,0.1973543465,0.0377613567,0.1241152808,0.0272634588,-0.2770866752,-0.1476682574,-0.2296878099,0.0351514332,0.3037281632,0.0866832137,0.2002630979,-0.2132453322,-0.0600378513,0.0495811887,-0.0122905448,-0.0569790304,0.1059752181,0.2511630356,0.2062113136,-0.2719920576,-0.1631948948,-0.1595900804,-0.2165904641,0.1562371105,-0.4682896137,0.0489555225,-0.1579335034,-0.110664919,-0.1586215347,0.1668361574,-0.3561002612,-0.3218687773,0.244806394,-0.0483381748,0.0161141399,0.1176600084,-0.3002466261,-0.0519229732,-0.1977204233,0.2331566364,-0.0668309703,0.4000051022,-0.2685908675,-0.2137828469,-0.1045412198,0.1410851032,0.1222287789,-0.0281012356,-0.0013925924,0.5620461702,0.3317559361,-0.1843186319,-0.1320330203,-0.4199966788,0.218624562,0.0801085904,-0.0030808279,0.043158669,0.1956190467,-0.1354965121,0.0702928603,0.4215479791,-0.164619267,0.099870339,0.0960523933,0.110236004,0.0731214285,-0.1226166189,0.1938954443,-0.1256381422,-0.1501625627,-0.6316099167,0.1066081002,0.0297481883,-0.1727101952,-0.1840963364,0.3182244897,-0.0157901887,0.1076047346,-0.0694290102,0.08997076,-0.0430511348,0.0124480771,0.4540242851,0.3595222831,0.1093105897,0.0691784099,-0.026388742,0.1020538881,0.0043796748,0.23595047,-0.0829031691,0.212587297,0.4033504426,-0.2370361388,-0.1057201624,-0.0774152577,-0.0378504284,0.2675491273,0.3059709072,-0.0998275802,-0.164961651,-0.1096301898,-0.0298333131,0.0824984685,-0.0792456567,-0.205833748,-0.3919481933,0.0027045219,0.0255116206,-0.1684821695,0.2499877661,-0.0448413081,-0.0304837208,0.3654894531,-0.0320747755,-0.087266393,-0.0385822132,-0.3025594652,0.1351388246,-0.0923736244,0.1497679502,0.1493162215,0.2277183682,-0.1718711555,0.1362490058,-0.1538341492,0.1338604391,-0.2160104364,-0.0281831641,0.456261605,0.1762324274,0.1133828387,-0.2925429642,0.3600921035,-0.1658401489,-0.2999567688,0.1651653945,0.3752211034,0.1846021414,-0.3102334738,-0.2788729072,-0.3881157041,-0.4727134705,-0.0969036743,-0.1738376915,0.1903258264,0.1340660453,0.2356193662,0.2885960937,0.1857780218,-0.271815896,0.0637562424,0.1844049394,0.0797061399,-0.1597715169,-0.1220592856,0.1154370606,-0.3134942651,0.4794272482,-0.2793750167,-0.3259856701,-0.0821823329,0.0239235349,0.2465979904,-0.1440027952,-0.1793990135,0.2091835737,0.2842224538,-0.1756088287,0.0167780239,-0.0508549362,0.2462119609,-0.1252524406,0.0307019632,-0.0990921035,0.3232720196,0.2634009719,0.4122442603,0.2139680237,-0.1056070551,0.0975423306,-0.0699536577,0.182574898,-0.0814448372,-0.4792007208,-0.2696222663,0.1803120822,-0.0901969522,0.0706486776,-0.2812636197,-0.1435808986,-0.2160851806,-0.4142453671,-0.1067027375,-0.1011577547,0.0656697154,-0.3637958467,0.2715679407,-0.3162961006,0.0702568293,0.1731541008,0.2696449161,-0.0772327557,-0.0503921136,0.1772670448,-0.2216293514,-0.1413160861,0.2567235231,-0.1036249548,0.3498116732,0.1539374739,0.0673135296,-0.0289432034,-0.0545444526,-0.1466884017,-0.0163864382,0.5797001123,-0.2601287663,0.0176654123,0.164336428,-0.1536818147,0.0434219129,-0.0261435527,-0.2320132256,0.1479611099,0.2719551325,0.3046005666,-0.3562911153,-0.249252364,0.5426259637,-0.1976947486,-0.0373904891,-0.0235336404,-0.1609675139,-0.1525467336,-0.1793020517,0.2017104179,0.1008620188,0.3100736141,0.1250273138,-0.1130562723,-0.1938491166,-0.2577957213,0.1798548996,0.0540498272,0.3140077293,-0.3528229296,0.1488126069,-0.1904231906,-0.0026105491,0.4281252623,0.4972411692,-0.1910155565,-0.1812524199,0.0633946732,-0.056969665,0.4667569101,0.2683847249,-0.1765778512,0.2876428068,-0.1283185035,0.2308446467,0.0327771083,0.0581566989,0.2633811533,-0.070523724,-0.3526120484,-0.268350333,0.0413162746,-0.2034849226,0.2236248553,-0.0684950277,0.0261809099,-0.1003900915,0.4952727556,-0.0874894187,0.5495759845,0.30720523,0.0596139021,0.5087492466,0.0492096394,0.0741647631,-0.0918361694,0.1703942418,-0.3895156085,-0.1453877538,-0.0813769251,-0.0178872477,0.2603551745,0.299153775,-0.3176331818,0.1703550816,-0.0689568222,0.3271101117,-0.0951449051,0.1279287487,-0.1128715277,0.1373051107,-0.1373913437,0.1528431475,-0.1778098643,-0.0433916524,-0.0117512671,-0.2617189884,0.0929747671,-0.3384669125,0.0450530387,0.1761831492,-0.2778231502,0.3936631083,0.08988107,-0.0701066107,-0.0150365978,0.3869258463,-0.2067174017,0.2277069837,0.2768901289,0.0989972353,-0.0945708081,-0.0682174414,-0.0461658984,0.1985012144,0.5305140615,0.0187036991,-0.237316221,0.2738576531,-0.3552470803,-0.2206637859,0.0997032076,0.2115415931,-0.1121062562,-0.54317981,-0.287638396,-0.1500448436,-0.1843277216,-0.1001694575,0.1664669663,-0.0357040353,-0.5352680087,-0.0308117177,-0.0213693138,-0.38414976,0.0178021565,0.156004265,0.11668998,0.2432755828,0.6686916947,-0.1233044639,0.0458799936,-0.405903101,0.2833765447,0.5040166378,-0.4621737897,0.5352696776,-0.0003103273,-0.0040415013,0.1493380666,0.0950817913,0.2635626197,0.4651308954,-0.0686901808,-0.30382967,-0.2341786176,0.119249694,-0.0470512509,0.3636304736,0.0785157979,0.3002308607,-0.087351121,0.1233468205,-0.3635303378,0.2422537655,-0.5364352465,0.2647243142,-0.1799982339,-0.2914522588,0.2737426162,0.0470660515,0.2638888359,0.1261848062,-0.0724792406,-0.2448962033,-0.0077675153,0.0940875709,-0.0158843081,0.1719727814,-0.0965202227,-0.0099471901,-0.282669574,0.016326366,-0.0709557459,-0.1472244859,-0.0752624795,0.3406225741,0.2054663002,0.3667328656,-0.0849789083,0.0152907269,-0.2456668317,-0.0473569818,-0.0918312147,0.1580786407,-0.1518214196,-0.100118272,0.2340717614,-0.0137567082,-0.2714761794,-0.3891197145,0.3647781014,-0.3388309479,-0.031513866,-0.2740297318,0.3758507073,0.1427515298,-0.1823082268,0.1834774166,0.0731686726,0.2262777686,-0.3055286109,0.2211140245,0.0991859511,0.0586696416,0.2623972893,0.3369342685,0.0075353426,0.2085572928,-0.1443812102,0.2394004017,0.1947534084,-0.3221794069,0.2501594424,0.7153099775,-0.3566303551,0.0208050869,0.3262009621,0.1710805893,0.1138685793,0.3657097518,-0.1454454958,0.3020301163,-0.0067227227,-0.2384236455,0.0666034892,-0.3710921407,-0.2538041472,0.4186913669,-0.0792862922,-0.1696757972,0.0349570885,0.181840241,-0.2336432338,-0.3206714988,0.163492009,0.19812271,-0.383675009,0.2351748943,-0.0976288095,-0.1467548609,-0.4992577434,-0.2183255106,0.3450300097,-0.0672331378,0.0651980788,0.1184729785,-0.0711433589,-0.0004677467,0.0417842418,0.3717813492,0.0315411761,-0.4196262658,0.2352451384,0.5135622025,0.1162745357,0.1104368046,0.2298839539,0.6503498554,0.3178335726,-0.395114392,0.0024227239,0.0518526211,-0.285060674,0.1031493247,0.0671083629,0.0683465004,0.4216350913,0.4693383873,0.1729317605,-0.1969882399,0.124450475,0.1749673933,0.2270068228,-0.1749568731,0.5054811239,0.1467603296,-0.075184986,-0.3003476858,-0.0588061251,-0.5807249546,-0.2967453301,0.2346037477,0.0357339755,0.2273804247,0.0481621139,0.105007261,0.0746003985,0.2745174766,0.3521292806,-0.0363088734,-0.4053151011,-0.0144129973,-0.3202983141,-0.1394312531,-0.2130356282,0.0831121951,-0.1005976796,0.1630856097,0.0160826985,0.011790283,0.0935773328,0.0428467579,0.0924524516,0.0422077477,-0.3283329606,0.0516575798,0.0035041969,-0.0702390447,0.0970961228,-0.3198404908,-0.0519565977,0.0440273546,0.0714518875,0.2712695003,-0.1841316521,0.1653142571,0.0590149164,-0.0674508587,0.1311156154,0.275832057,0.0277970303,0.0532533862,0.0716786757,-0.0929611176,-0.2856594324,0.2750591636,0.4175396562,0.4672616422,0.1319558471,-0.1317616999,-0.2883381844,0.1405082196,0.012215591,0.0503441505,-0.1526644081,0.2168417871,0.1254860312,0.0466474332,0.1734493077,-0.0986990035,-0.2047881037,0.0616952553,-0.4222971797,-0.4047438502,0.3502146602,-0.1958712339,-0.2600913048,-0.0018100692,0.0211870205,0.4787614644,0.1570665985,-0.6342811584,-0.2474755943,0.3716071844,-0.2042856514,-0.3200725019,-0.0337961279,0.1368300617,-0.0045459424,0.0947797,-0.1115361154,0.0558272153,-0.2615488172,0.2261002809,-0.2754898667]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/5","title":"ValueError when a split is empty","comments":"Yes sounds good to me!\r\nDo you want to make a PR? or I can do it as well","body":"When a split is empty either TEST, VALIDATION or TRAIN I get the following error:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 295, in load\r\n ds = dbuilder.as_dataset(**as_dataset_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 587, in as_dataset\r\n datasets = utils.map_nested(build_single_dataset, split, map_tuple=True)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in map_nested\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in <dictcomp>\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 601, in _build_single_dataset\r\n split=split,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 625, in _as_dataset\r\n split_infos=self.info.splits.values(),\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 200, in read\r\n return py_utils.map_nested(_read_instruction_to_ds, instructions)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 191, in _read_instruction_to_ds\r\n file_instructions = make_file_instructions(name, split_infos, instruction)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 104, in make_file_instructions\r\n absolute_instructions=absolute_instructions,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 122, in _make_file_instructions_from_absolutes\r\n 'Split empty. This might means that dataset hasn\\'t been generated '\r\nValueError: Split empty. This might means that dataset hasn't been generated yet and info not restored from GCS, or that legacy dataset is used.\r\n``` \r\n\r\nHow to reproduce:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(self, **config):\r\n self.train = config.pop(\"train\", None)\r\n self.validation = config.pop(\"validation\", None)\r\n super(Bbc, self).__init__(**config)\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, description=\"bla\", features=nlp.features.FeaturesDict({\"id\": nlp.int32, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": self.train}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": self.validation}),\r\n nlp.SplitGenerator(name=nlp.Split.TEST, gen_kwargs={\"filepath\": None})]\r\n\r\n def _generate_examples(self, filepath):\r\n if not filepath:\r\n return None, {}\r\n\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"id\": idx, \"text\": line[1], \"label\": line[0]}\r\n```\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\"})\r\n```","comment_length":19,"text":"ValueError when a split is empty\nWhen a split is empty either TEST, VALIDATION or TRAIN I get the following error:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 295, in load\r\n ds = dbuilder.as_dataset(**as_dataset_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 587, in as_dataset\r\n datasets = utils.map_nested(build_single_dataset, split, map_tuple=True)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in map_nested\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in <dictcomp>\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 601, in _build_single_dataset\r\n split=split,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 625, in _as_dataset\r\n split_infos=self.info.splits.values(),\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 200, in read\r\n return py_utils.map_nested(_read_instruction_to_ds, instructions)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 191, in _read_instruction_to_ds\r\n file_instructions = make_file_instructions(name, split_infos, instruction)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 104, in make_file_instructions\r\n absolute_instructions=absolute_instructions,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 122, in _make_file_instructions_from_absolutes\r\n 'Split empty. This might means that dataset hasn\\'t been generated '\r\nValueError: Split empty. This might means that dataset hasn't been generated yet and info not restored from GCS, or that legacy dataset is used.\r\n``` \r\n\r\nHow to reproduce:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(self, **config):\r\n self.train = config.pop(\"train\", None)\r\n self.validation = config.pop(\"validation\", None)\r\n super(Bbc, self).__init__(**config)\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, description=\"bla\", features=nlp.features.FeaturesDict({\"id\": nlp.int32, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": self.train}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": self.validation}),\r\n nlp.SplitGenerator(name=nlp.Split.TEST, gen_kwargs={\"filepath\": None})]\r\n\r\n def _generate_examples(self, filepath):\r\n if not filepath:\r\n return None, {}\r\n\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"id\": idx, \"text\": line[1], \"label\": line[0]}\r\n```\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\"})\r\n```\nYes sounds good to me!\r\nDo you want to make a PR? or I can do it as well","embeddings":[-0.2972342372,-0.0131540569,-0.103067942,0.2931129634,0.0207595415,-0.0219010543,0.6005375385,0.4749840498,-0.0488540977,0.4204655886,0.0286984555,0.2306589633,-0.3276998699,0.1830051839,-0.4715830386,-0.2210291326,-0.0620970726,0.4033246934,0.0316321403,0.0022019364,-0.2077705711,0.1781739295,-0.3489648104,0.2001751065,-0.4412805736,-0.3235142529,-0.0962038264,0.0336118527,-0.1655961424,-0.5333554745,0.2795565426,-0.315582782,-0.2108621746,0.4383441806,-0.0001089997,0.2471223772,0.3216657937,0.0212966055,-0.2349680364,-0.1730071604,-0.2148883939,0.0468612611,-0.0701180547,-0.1868547648,-0.1008067131,-0.2197800577,-0.3655616343,0.0298773181,0.2869431078,0.2550378144,0.1966028512,0.1572323292,0.0249177329,-0.1763806641,0.0856723338,-0.105694443,0.0970320776,0.0854190886,-0.1803481877,-0.2269463241,-0.1018456817,0.2128955573,-0.0124384062,0.0009043886,-0.2380745709,0.1348021328,0.2368194163,-0.3553718626,-0.1338815093,0.0872519761,-0.0105639342,-0.0024150251,-0.1980995387,-0.3114365339,0.0216911975,-0.3575952351,0.0597179197,0.2921065986,-0.1286965311,0.124095358,0.228433311,0.1736139357,-0.2168529481,-0.0249617305,-0.3034926653,0.3541932404,-0.031331487,0.0348216146,0.2476863563,-0.0247122962,0.0857289061,-0.0367395617,-0.218579337,0.115537703,-0.2246417701,-0.3038747907,-0.1532901824,-0.3125609457,-0.0919000655,-0.0317804776,0.027630873,-0.1767440885,0.1064907983,0.4088112414,0.304340452,-0.0534912385,0.3710521758,0.4308705032,0.0564235486,0.1503341496,0.0145205986,0.02915808,0.0300026834,-0.276969254,0.0414252691,0.1946728826,0.4200007617,0.0776329115,-0.3998927772,-0.0317487083,-0.2662963867,-0.0253004767,0.2187399864,0.377743125,0.0436375402,0.1135604605,-0.059801314,0.1989057064,-0.0594728105,-0.4426795244,-0.3030682206,0.3334758282,-0.1042798758,-0.3177535534,0.0193641763,-0.0207902119,0.1957095265,0.2191386372,-0.1739814132,-0.0592373982,0.1233527735,-0.4605067074,0.1437606066,0.3618825972,-0.0033455135,-0.2083796561,0.2578766048,-0.2094171792,0.0035285249,0.2536199987,-0.0682576448,-0.3559485078,-0.0321290009,0.337768048,0.1614764482,0.1738892049,-0.116377309,0.1625697762,0.3533759117,-0.2079333812,0.024277322,-0.085838072,-0.2200120091,-0.2932094634,0.1470615864,0.3619220257,-0.3413058817,-0.133083865,-0.1345353574,-0.1824868917,0.0983030871,0.0809265226,-0.1797881573,0.3938573897,-0.2386934012,0.5446959734,0.4584326446,-0.2361291498,-0.0809233189,0.3233044147,-0.0367279463,-0.2104673386,-0.0367180035,-0.1798265278,0.1973543465,0.0377613567,0.1241152808,0.0272634588,-0.2770866752,-0.1476682574,-0.2296878099,0.0351514332,0.3037281632,0.0866832137,0.2002630979,-0.2132453322,-0.0600378513,0.0495811887,-0.0122905448,-0.0569790304,0.1059752181,0.2511630356,0.2062113136,-0.2719920576,-0.1631948948,-0.1595900804,-0.2165904641,0.1562371105,-0.4682896137,0.0489555225,-0.1579335034,-0.110664919,-0.1586215347,0.1668361574,-0.3561002612,-0.3218687773,0.244806394,-0.0483381748,0.0161141399,0.1176600084,-0.3002466261,-0.0519229732,-0.1977204233,0.2331566364,-0.0668309703,0.4000051022,-0.2685908675,-0.2137828469,-0.1045412198,0.1410851032,0.1222287789,-0.0281012356,-0.0013925924,0.5620461702,0.3317559361,-0.1843186319,-0.1320330203,-0.4199966788,0.218624562,0.0801085904,-0.0030808279,0.043158669,0.1956190467,-0.1354965121,0.0702928603,0.4215479791,-0.164619267,0.099870339,0.0960523933,0.110236004,0.0731214285,-0.1226166189,0.1938954443,-0.1256381422,-0.1501625627,-0.6316099167,0.1066081002,0.0297481883,-0.1727101952,-0.1840963364,0.3182244897,-0.0157901887,0.1076047346,-0.0694290102,0.08997076,-0.0430511348,0.0124480771,0.4540242851,0.3595222831,0.1093105897,0.0691784099,-0.026388742,0.1020538881,0.0043796748,0.23595047,-0.0829031691,0.212587297,0.4033504426,-0.2370361388,-0.1057201624,-0.0774152577,-0.0378504284,0.2675491273,0.3059709072,-0.0998275802,-0.164961651,-0.1096301898,-0.0298333131,0.0824984685,-0.0792456567,-0.205833748,-0.3919481933,0.0027045219,0.0255116206,-0.1684821695,0.2499877661,-0.0448413081,-0.0304837208,0.3654894531,-0.0320747755,-0.087266393,-0.0385822132,-0.3025594652,0.1351388246,-0.0923736244,0.1497679502,0.1493162215,0.2277183682,-0.1718711555,0.1362490058,-0.1538341492,0.1338604391,-0.2160104364,-0.0281831641,0.456261605,0.1762324274,0.1133828387,-0.2925429642,0.3600921035,-0.1658401489,-0.2999567688,0.1651653945,0.3752211034,0.1846021414,-0.3102334738,-0.2788729072,-0.3881157041,-0.4727134705,-0.0969036743,-0.1738376915,0.1903258264,0.1340660453,0.2356193662,0.2885960937,0.1857780218,-0.271815896,0.0637562424,0.1844049394,0.0797061399,-0.1597715169,-0.1220592856,0.1154370606,-0.3134942651,0.4794272482,-0.2793750167,-0.3259856701,-0.0821823329,0.0239235349,0.2465979904,-0.1440027952,-0.1793990135,0.2091835737,0.2842224538,-0.1756088287,0.0167780239,-0.0508549362,0.2462119609,-0.1252524406,0.0307019632,-0.0990921035,0.3232720196,0.2634009719,0.4122442603,0.2139680237,-0.1056070551,0.0975423306,-0.0699536577,0.182574898,-0.0814448372,-0.4792007208,-0.2696222663,0.1803120822,-0.0901969522,0.0706486776,-0.2812636197,-0.1435808986,-0.2160851806,-0.4142453671,-0.1067027375,-0.1011577547,0.0656697154,-0.3637958467,0.2715679407,-0.3162961006,0.0702568293,0.1731541008,0.2696449161,-0.0772327557,-0.0503921136,0.1772670448,-0.2216293514,-0.1413160861,0.2567235231,-0.1036249548,0.3498116732,0.1539374739,0.0673135296,-0.0289432034,-0.0545444526,-0.1466884017,-0.0163864382,0.5797001123,-0.2601287663,0.0176654123,0.164336428,-0.1536818147,0.0434219129,-0.0261435527,-0.2320132256,0.1479611099,0.2719551325,0.3046005666,-0.3562911153,-0.249252364,0.5426259637,-0.1976947486,-0.0373904891,-0.0235336404,-0.1609675139,-0.1525467336,-0.1793020517,0.2017104179,0.1008620188,0.3100736141,0.1250273138,-0.1130562723,-0.1938491166,-0.2577957213,0.1798548996,0.0540498272,0.3140077293,-0.3528229296,0.1488126069,-0.1904231906,-0.0026105491,0.4281252623,0.4972411692,-0.1910155565,-0.1812524199,0.0633946732,-0.056969665,0.4667569101,0.2683847249,-0.1765778512,0.2876428068,-0.1283185035,0.2308446467,0.0327771083,0.0581566989,0.2633811533,-0.070523724,-0.3526120484,-0.268350333,0.0413162746,-0.2034849226,0.2236248553,-0.0684950277,0.0261809099,-0.1003900915,0.4952727556,-0.0874894187,0.5495759845,0.30720523,0.0596139021,0.5087492466,0.0492096394,0.0741647631,-0.0918361694,0.1703942418,-0.3895156085,-0.1453877538,-0.0813769251,-0.0178872477,0.2603551745,0.299153775,-0.3176331818,0.1703550816,-0.0689568222,0.3271101117,-0.0951449051,0.1279287487,-0.1128715277,0.1373051107,-0.1373913437,0.1528431475,-0.1778098643,-0.0433916524,-0.0117512671,-0.2617189884,0.0929747671,-0.3384669125,0.0450530387,0.1761831492,-0.2778231502,0.3936631083,0.08988107,-0.0701066107,-0.0150365978,0.3869258463,-0.2067174017,0.2277069837,0.2768901289,0.0989972353,-0.0945708081,-0.0682174414,-0.0461658984,0.1985012144,0.5305140615,0.0187036991,-0.237316221,0.2738576531,-0.3552470803,-0.2206637859,0.0997032076,0.2115415931,-0.1121062562,-0.54317981,-0.287638396,-0.1500448436,-0.1843277216,-0.1001694575,0.1664669663,-0.0357040353,-0.5352680087,-0.0308117177,-0.0213693138,-0.38414976,0.0178021565,0.156004265,0.11668998,0.2432755828,0.6686916947,-0.1233044639,0.0458799936,-0.405903101,0.2833765447,0.5040166378,-0.4621737897,0.5352696776,-0.0003103273,-0.0040415013,0.1493380666,0.0950817913,0.2635626197,0.4651308954,-0.0686901808,-0.30382967,-0.2341786176,0.119249694,-0.0470512509,0.3636304736,0.0785157979,0.3002308607,-0.087351121,0.1233468205,-0.3635303378,0.2422537655,-0.5364352465,0.2647243142,-0.1799982339,-0.2914522588,0.2737426162,0.0470660515,0.2638888359,0.1261848062,-0.0724792406,-0.2448962033,-0.0077675153,0.0940875709,-0.0158843081,0.1719727814,-0.0965202227,-0.0099471901,-0.282669574,0.016326366,-0.0709557459,-0.1472244859,-0.0752624795,0.3406225741,0.2054663002,0.3667328656,-0.0849789083,0.0152907269,-0.2456668317,-0.0473569818,-0.0918312147,0.1580786407,-0.1518214196,-0.100118272,0.2340717614,-0.0137567082,-0.2714761794,-0.3891197145,0.3647781014,-0.3388309479,-0.031513866,-0.2740297318,0.3758507073,0.1427515298,-0.1823082268,0.1834774166,0.0731686726,0.2262777686,-0.3055286109,0.2211140245,0.0991859511,0.0586696416,0.2623972893,0.3369342685,0.0075353426,0.2085572928,-0.1443812102,0.2394004017,0.1947534084,-0.3221794069,0.2501594424,0.7153099775,-0.3566303551,0.0208050869,0.3262009621,0.1710805893,0.1138685793,0.3657097518,-0.1454454958,0.3020301163,-0.0067227227,-0.2384236455,0.0666034892,-0.3710921407,-0.2538041472,0.4186913669,-0.0792862922,-0.1696757972,0.0349570885,0.181840241,-0.2336432338,-0.3206714988,0.163492009,0.19812271,-0.383675009,0.2351748943,-0.0976288095,-0.1467548609,-0.4992577434,-0.2183255106,0.3450300097,-0.0672331378,0.0651980788,0.1184729785,-0.0711433589,-0.0004677467,0.0417842418,0.3717813492,0.0315411761,-0.4196262658,0.2352451384,0.5135622025,0.1162745357,0.1104368046,0.2298839539,0.6503498554,0.3178335726,-0.395114392,0.0024227239,0.0518526211,-0.285060674,0.1031493247,0.0671083629,0.0683465004,0.4216350913,0.4693383873,0.1729317605,-0.1969882399,0.124450475,0.1749673933,0.2270068228,-0.1749568731,0.5054811239,0.1467603296,-0.075184986,-0.3003476858,-0.0588061251,-0.5807249546,-0.2967453301,0.2346037477,0.0357339755,0.2273804247,0.0481621139,0.105007261,0.0746003985,0.2745174766,0.3521292806,-0.0363088734,-0.4053151011,-0.0144129973,-0.3202983141,-0.1394312531,-0.2130356282,0.0831121951,-0.1005976796,0.1630856097,0.0160826985,0.011790283,0.0935773328,0.0428467579,0.0924524516,0.0422077477,-0.3283329606,0.0516575798,0.0035041969,-0.0702390447,0.0970961228,-0.3198404908,-0.0519565977,0.0440273546,0.0714518875,0.2712695003,-0.1841316521,0.1653142571,0.0590149164,-0.0674508587,0.1311156154,0.275832057,0.0277970303,0.0532533862,0.0716786757,-0.0929611176,-0.2856594324,0.2750591636,0.4175396562,0.4672616422,0.1319558471,-0.1317616999,-0.2883381844,0.1405082196,0.012215591,0.0503441505,-0.1526644081,0.2168417871,0.1254860312,0.0466474332,0.1734493077,-0.0986990035,-0.2047881037,0.0616952553,-0.4222971797,-0.4047438502,0.3502146602,-0.1958712339,-0.2600913048,-0.0018100692,0.0211870205,0.4787614644,0.1570665985,-0.6342811584,-0.2474755943,0.3716071844,-0.2042856514,-0.3200725019,-0.0337961279,0.1368300617,-0.0045459424,0.0947797,-0.1115361154,0.0558272153,-0.2615488172,0.2261002809,-0.2754898667]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/5","title":"ValueError when a split is empty","comments":"Fixed.","body":"When a split is empty either TEST, VALIDATION or TRAIN I get the following error:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 295, in load\r\n ds = dbuilder.as_dataset(**as_dataset_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 587, in as_dataset\r\n datasets = utils.map_nested(build_single_dataset, split, map_tuple=True)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in map_nested\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in <dictcomp>\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 601, in _build_single_dataset\r\n split=split,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 625, in _as_dataset\r\n split_infos=self.info.splits.values(),\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 200, in read\r\n return py_utils.map_nested(_read_instruction_to_ds, instructions)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 191, in _read_instruction_to_ds\r\n file_instructions = make_file_instructions(name, split_infos, instruction)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 104, in make_file_instructions\r\n absolute_instructions=absolute_instructions,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 122, in _make_file_instructions_from_absolutes\r\n 'Split empty. This might means that dataset hasn\\'t been generated '\r\nValueError: Split empty. This might means that dataset hasn't been generated yet and info not restored from GCS, or that legacy dataset is used.\r\n``` \r\n\r\nHow to reproduce:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(self, **config):\r\n self.train = config.pop(\"train\", None)\r\n self.validation = config.pop(\"validation\", None)\r\n super(Bbc, self).__init__(**config)\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, description=\"bla\", features=nlp.features.FeaturesDict({\"id\": nlp.int32, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": self.train}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": self.validation}),\r\n nlp.SplitGenerator(name=nlp.Split.TEST, gen_kwargs={\"filepath\": None})]\r\n\r\n def _generate_examples(self, filepath):\r\n if not filepath:\r\n return None, {}\r\n\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"id\": idx, \"text\": line[1], \"label\": line[0]}\r\n```\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\"})\r\n```","comment_length":1,"text":"ValueError when a split is empty\nWhen a split is empty either TEST, VALIDATION or TRAIN I get the following error:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 295, in load\r\n ds = dbuilder.as_dataset(**as_dataset_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 587, in as_dataset\r\n datasets = utils.map_nested(build_single_dataset, split, map_tuple=True)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in map_nested\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 158, in <dictcomp>\r\n for k, v in data_struct.items()\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 601, in _build_single_dataset\r\n split=split,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/builder.py\", line 625, in _as_dataset\r\n split_infos=self.info.splits.values(),\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 200, in read\r\n return py_utils.map_nested(_read_instruction_to_ds, instructions)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/utils\/py_utils.py\", line 172, in map_nested\r\n return function(data_struct)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 191, in _read_instruction_to_ds\r\n file_instructions = make_file_instructions(name, split_infos, instruction)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 104, in make_file_instructions\r\n absolute_instructions=absolute_instructions,\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/arrow_reader.py\", line 122, in _make_file_instructions_from_absolutes\r\n 'Split empty. This might means that dataset hasn\\'t been generated '\r\nValueError: Split empty. This might means that dataset hasn't been generated yet and info not restored from GCS, or that legacy dataset is used.\r\n``` \r\n\r\nHow to reproduce:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(self, **config):\r\n self.train = config.pop(\"train\", None)\r\n self.validation = config.pop(\"validation\", None)\r\n super(Bbc, self).__init__(**config)\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, description=\"bla\", features=nlp.features.FeaturesDict({\"id\": nlp.int32, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": self.train}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": self.validation}),\r\n nlp.SplitGenerator(name=nlp.Split.TEST, gen_kwargs={\"filepath\": None})]\r\n\r\n def _generate_examples(self, filepath):\r\n if not filepath:\r\n return None, {}\r\n\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"id\": idx, \"text\": line[1], \"label\": line[0]}\r\n```\r\n\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\"})\r\n```\nFixed.","embeddings":[-0.2972342372,-0.0131540569,-0.103067942,0.2931129634,0.0207595415,-0.0219010543,0.6005375385,0.4749840498,-0.0488540977,0.4204655886,0.0286984555,0.2306589633,-0.3276998699,0.1830051839,-0.4715830386,-0.2210291326,-0.0620970726,0.4033246934,0.0316321403,0.0022019364,-0.2077705711,0.1781739295,-0.3489648104,0.2001751065,-0.4412805736,-0.3235142529,-0.0962038264,0.0336118527,-0.1655961424,-0.5333554745,0.2795565426,-0.315582782,-0.2108621746,0.4383441806,-0.0001089997,0.2471223772,0.3216657937,0.0212966055,-0.2349680364,-0.1730071604,-0.2148883939,0.0468612611,-0.0701180547,-0.1868547648,-0.1008067131,-0.2197800577,-0.3655616343,0.0298773181,0.2869431078,0.2550378144,0.1966028512,0.1572323292,0.0249177329,-0.1763806641,0.0856723338,-0.105694443,0.0970320776,0.0854190886,-0.1803481877,-0.2269463241,-0.1018456817,0.2128955573,-0.0124384062,0.0009043886,-0.2380745709,0.1348021328,0.2368194163,-0.3553718626,-0.1338815093,0.0872519761,-0.0105639342,-0.0024150251,-0.1980995387,-0.3114365339,0.0216911975,-0.3575952351,0.0597179197,0.2921065986,-0.1286965311,0.124095358,0.228433311,0.1736139357,-0.2168529481,-0.0249617305,-0.3034926653,0.3541932404,-0.031331487,0.0348216146,0.2476863563,-0.0247122962,0.0857289061,-0.0367395617,-0.218579337,0.115537703,-0.2246417701,-0.3038747907,-0.1532901824,-0.3125609457,-0.0919000655,-0.0317804776,0.027630873,-0.1767440885,0.1064907983,0.4088112414,0.304340452,-0.0534912385,0.3710521758,0.4308705032,0.0564235486,0.1503341496,0.0145205986,0.02915808,0.0300026834,-0.276969254,0.0414252691,0.1946728826,0.4200007617,0.0776329115,-0.3998927772,-0.0317487083,-0.2662963867,-0.0253004767,0.2187399864,0.377743125,0.0436375402,0.1135604605,-0.059801314,0.1989057064,-0.0594728105,-0.4426795244,-0.3030682206,0.3334758282,-0.1042798758,-0.3177535534,0.0193641763,-0.0207902119,0.1957095265,0.2191386372,-0.1739814132,-0.0592373982,0.1233527735,-0.4605067074,0.1437606066,0.3618825972,-0.0033455135,-0.2083796561,0.2578766048,-0.2094171792,0.0035285249,0.2536199987,-0.0682576448,-0.3559485078,-0.0321290009,0.337768048,0.1614764482,0.1738892049,-0.116377309,0.1625697762,0.3533759117,-0.2079333812,0.024277322,-0.085838072,-0.2200120091,-0.2932094634,0.1470615864,0.3619220257,-0.3413058817,-0.133083865,-0.1345353574,-0.1824868917,0.0983030871,0.0809265226,-0.1797881573,0.3938573897,-0.2386934012,0.5446959734,0.4584326446,-0.2361291498,-0.0809233189,0.3233044147,-0.0367279463,-0.2104673386,-0.0367180035,-0.1798265278,0.1973543465,0.0377613567,0.1241152808,0.0272634588,-0.2770866752,-0.1476682574,-0.2296878099,0.0351514332,0.3037281632,0.0866832137,0.2002630979,-0.2132453322,-0.0600378513,0.0495811887,-0.0122905448,-0.0569790304,0.1059752181,0.2511630356,0.2062113136,-0.2719920576,-0.1631948948,-0.1595900804,-0.2165904641,0.1562371105,-0.4682896137,0.0489555225,-0.1579335034,-0.110664919,-0.1586215347,0.1668361574,-0.3561002612,-0.3218687773,0.244806394,-0.0483381748,0.0161141399,0.1176600084,-0.3002466261,-0.0519229732,-0.1977204233,0.2331566364,-0.0668309703,0.4000051022,-0.2685908675,-0.2137828469,-0.1045412198,0.1410851032,0.1222287789,-0.0281012356,-0.0013925924,0.5620461702,0.3317559361,-0.1843186319,-0.1320330203,-0.4199966788,0.218624562,0.0801085904,-0.0030808279,0.043158669,0.1956190467,-0.1354965121,0.0702928603,0.4215479791,-0.164619267,0.099870339,0.0960523933,0.110236004,0.0731214285,-0.1226166189,0.1938954443,-0.1256381422,-0.1501625627,-0.6316099167,0.1066081002,0.0297481883,-0.1727101952,-0.1840963364,0.3182244897,-0.0157901887,0.1076047346,-0.0694290102,0.08997076,-0.0430511348,0.0124480771,0.4540242851,0.3595222831,0.1093105897,0.0691784099,-0.026388742,0.1020538881,0.0043796748,0.23595047,-0.0829031691,0.212587297,0.4033504426,-0.2370361388,-0.1057201624,-0.0774152577,-0.0378504284,0.2675491273,0.3059709072,-0.0998275802,-0.164961651,-0.1096301898,-0.0298333131,0.0824984685,-0.0792456567,-0.205833748,-0.3919481933,0.0027045219,0.0255116206,-0.1684821695,0.2499877661,-0.0448413081,-0.0304837208,0.3654894531,-0.0320747755,-0.087266393,-0.0385822132,-0.3025594652,0.1351388246,-0.0923736244,0.1497679502,0.1493162215,0.2277183682,-0.1718711555,0.1362490058,-0.1538341492,0.1338604391,-0.2160104364,-0.0281831641,0.456261605,0.1762324274,0.1133828387,-0.2925429642,0.3600921035,-0.1658401489,-0.2999567688,0.1651653945,0.3752211034,0.1846021414,-0.3102334738,-0.2788729072,-0.3881157041,-0.4727134705,-0.0969036743,-0.1738376915,0.1903258264,0.1340660453,0.2356193662,0.2885960937,0.1857780218,-0.271815896,0.0637562424,0.1844049394,0.0797061399,-0.1597715169,-0.1220592856,0.1154370606,-0.3134942651,0.4794272482,-0.2793750167,-0.3259856701,-0.0821823329,0.0239235349,0.2465979904,-0.1440027952,-0.1793990135,0.2091835737,0.2842224538,-0.1756088287,0.0167780239,-0.0508549362,0.2462119609,-0.1252524406,0.0307019632,-0.0990921035,0.3232720196,0.2634009719,0.4122442603,0.2139680237,-0.1056070551,0.0975423306,-0.0699536577,0.182574898,-0.0814448372,-0.4792007208,-0.2696222663,0.1803120822,-0.0901969522,0.0706486776,-0.2812636197,-0.1435808986,-0.2160851806,-0.4142453671,-0.1067027375,-0.1011577547,0.0656697154,-0.3637958467,0.2715679407,-0.3162961006,0.0702568293,0.1731541008,0.2696449161,-0.0772327557,-0.0503921136,0.1772670448,-0.2216293514,-0.1413160861,0.2567235231,-0.1036249548,0.3498116732,0.1539374739,0.0673135296,-0.0289432034,-0.0545444526,-0.1466884017,-0.0163864382,0.5797001123,-0.2601287663,0.0176654123,0.164336428,-0.1536818147,0.0434219129,-0.0261435527,-0.2320132256,0.1479611099,0.2719551325,0.3046005666,-0.3562911153,-0.249252364,0.5426259637,-0.1976947486,-0.0373904891,-0.0235336404,-0.1609675139,-0.1525467336,-0.1793020517,0.2017104179,0.1008620188,0.3100736141,0.1250273138,-0.1130562723,-0.1938491166,-0.2577957213,0.1798548996,0.0540498272,0.3140077293,-0.3528229296,0.1488126069,-0.1904231906,-0.0026105491,0.4281252623,0.4972411692,-0.1910155565,-0.1812524199,0.0633946732,-0.056969665,0.4667569101,0.2683847249,-0.1765778512,0.2876428068,-0.1283185035,0.2308446467,0.0327771083,0.0581566989,0.2633811533,-0.070523724,-0.3526120484,-0.268350333,0.0413162746,-0.2034849226,0.2236248553,-0.0684950277,0.0261809099,-0.1003900915,0.4952727556,-0.0874894187,0.5495759845,0.30720523,0.0596139021,0.5087492466,0.0492096394,0.0741647631,-0.0918361694,0.1703942418,-0.3895156085,-0.1453877538,-0.0813769251,-0.0178872477,0.2603551745,0.299153775,-0.3176331818,0.1703550816,-0.0689568222,0.3271101117,-0.0951449051,0.1279287487,-0.1128715277,0.1373051107,-0.1373913437,0.1528431475,-0.1778098643,-0.0433916524,-0.0117512671,-0.2617189884,0.0929747671,-0.3384669125,0.0450530387,0.1761831492,-0.2778231502,0.3936631083,0.08988107,-0.0701066107,-0.0150365978,0.3869258463,-0.2067174017,0.2277069837,0.2768901289,0.0989972353,-0.0945708081,-0.0682174414,-0.0461658984,0.1985012144,0.5305140615,0.0187036991,-0.237316221,0.2738576531,-0.3552470803,-0.2206637859,0.0997032076,0.2115415931,-0.1121062562,-0.54317981,-0.287638396,-0.1500448436,-0.1843277216,-0.1001694575,0.1664669663,-0.0357040353,-0.5352680087,-0.0308117177,-0.0213693138,-0.38414976,0.0178021565,0.156004265,0.11668998,0.2432755828,0.6686916947,-0.1233044639,0.0458799936,-0.405903101,0.2833765447,0.5040166378,-0.4621737897,0.5352696776,-0.0003103273,-0.0040415013,0.1493380666,0.0950817913,0.2635626197,0.4651308954,-0.0686901808,-0.30382967,-0.2341786176,0.119249694,-0.0470512509,0.3636304736,0.0785157979,0.3002308607,-0.087351121,0.1233468205,-0.3635303378,0.2422537655,-0.5364352465,0.2647243142,-0.1799982339,-0.2914522588,0.2737426162,0.0470660515,0.2638888359,0.1261848062,-0.0724792406,-0.2448962033,-0.0077675153,0.0940875709,-0.0158843081,0.1719727814,-0.0965202227,-0.0099471901,-0.282669574,0.016326366,-0.0709557459,-0.1472244859,-0.0752624795,0.3406225741,0.2054663002,0.3667328656,-0.0849789083,0.0152907269,-0.2456668317,-0.0473569818,-0.0918312147,0.1580786407,-0.1518214196,-0.100118272,0.2340717614,-0.0137567082,-0.2714761794,-0.3891197145,0.3647781014,-0.3388309479,-0.031513866,-0.2740297318,0.3758507073,0.1427515298,-0.1823082268,0.1834774166,0.0731686726,0.2262777686,-0.3055286109,0.2211140245,0.0991859511,0.0586696416,0.2623972893,0.3369342685,0.0075353426,0.2085572928,-0.1443812102,0.2394004017,0.1947534084,-0.3221794069,0.2501594424,0.7153099775,-0.3566303551,0.0208050869,0.3262009621,0.1710805893,0.1138685793,0.3657097518,-0.1454454958,0.3020301163,-0.0067227227,-0.2384236455,0.0666034892,-0.3710921407,-0.2538041472,0.4186913669,-0.0792862922,-0.1696757972,0.0349570885,0.181840241,-0.2336432338,-0.3206714988,0.163492009,0.19812271,-0.383675009,0.2351748943,-0.0976288095,-0.1467548609,-0.4992577434,-0.2183255106,0.3450300097,-0.0672331378,0.0651980788,0.1184729785,-0.0711433589,-0.0004677467,0.0417842418,0.3717813492,0.0315411761,-0.4196262658,0.2352451384,0.5135622025,0.1162745357,0.1104368046,0.2298839539,0.6503498554,0.3178335726,-0.395114392,0.0024227239,0.0518526211,-0.285060674,0.1031493247,0.0671083629,0.0683465004,0.4216350913,0.4693383873,0.1729317605,-0.1969882399,0.124450475,0.1749673933,0.2270068228,-0.1749568731,0.5054811239,0.1467603296,-0.075184986,-0.3003476858,-0.0588061251,-0.5807249546,-0.2967453301,0.2346037477,0.0357339755,0.2273804247,0.0481621139,0.105007261,0.0746003985,0.2745174766,0.3521292806,-0.0363088734,-0.4053151011,-0.0144129973,-0.3202983141,-0.1394312531,-0.2130356282,0.0831121951,-0.1005976796,0.1630856097,0.0160826985,0.011790283,0.0935773328,0.0428467579,0.0924524516,0.0422077477,-0.3283329606,0.0516575798,0.0035041969,-0.0702390447,0.0970961228,-0.3198404908,-0.0519565977,0.0440273546,0.0714518875,0.2712695003,-0.1841316521,0.1653142571,0.0590149164,-0.0674508587,0.1311156154,0.275832057,0.0277970303,0.0532533862,0.0716786757,-0.0929611176,-0.2856594324,0.2750591636,0.4175396562,0.4672616422,0.1319558471,-0.1317616999,-0.2883381844,0.1405082196,0.012215591,0.0503441505,-0.1526644081,0.2168417871,0.1254860312,0.0466474332,0.1734493077,-0.0986990035,-0.2047881037,0.0616952553,-0.4222971797,-0.4047438502,0.3502146602,-0.1958712339,-0.2600913048,-0.0018100692,0.0211870205,0.4787614644,0.1570665985,-0.6342811584,-0.2474755943,0.3716071844,-0.2042856514,-0.3200725019,-0.0337961279,0.1368300617,-0.0045459424,0.0947797,-0.1115361154,0.0558272153,-0.2615488172,0.2261002809,-0.2754898667]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/4","title":"[Feature] Keep the list of labels of a dataset as metadata","comments":"Yes! I see mostly two options for this:\r\n- a `Feature` approach like currently (but we might deprecate features)\r\n- wrapping in a smart way the Dictionary arrays of Arrow: https:\/\/arrow.apache.org\/docs\/python\/data.html?highlight=dictionary%20encode#dictionary-arrays","body":"It would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.","comment_length":31,"text":"[Feature] Keep the list of labels of a dataset as metadata\nIt would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.\nYes! I see mostly two options for this:\r\n- a `Feature` approach like currently (but we might deprecate features)\r\n- wrapping in a smart way the Dictionary arrays of Arrow: https:\/\/arrow.apache.org\/docs\/python\/data.html?highlight=dictionary%20encode#dictionary-arrays","embeddings":[0.1531038135,-0.0619831346,-0.1409771144,0.0409961455,0.1746009439,0.1912366003,0.2029764354,0.2303460389,-0.0681677684,0.0453065932,-0.0701301843,0.5808182359,-0.3312079906,0.1982669532,0.0061332081,-0.1212565675,-0.0128118554,-0.0272393189,0.0372485667,-0.0533793308,-0.1979878545,-0.3048540056,0.1820646673,-0.0400751047,-0.2287419736,-0.0672599226,0.1197463796,-0.3560012281,0.0509823933,-0.683021307,-0.1670466661,-0.0050464547,-0.0410597622,-0.1845041811,-0.0001063943,-0.0223298315,0.2254130542,-0.0063495575,-0.3393560648,0.0036049434,-0.3735336661,-0.252666384,0.2441437989,-0.2920621037,0.0635793507,-0.4220448136,-0.2341690958,-0.3475583792,0.0337780938,0.0731797069,0.2945555151,-0.0099700466,-0.0633411705,0.198744297,0.4615858495,0.3243832886,-0.4491933882,-0.0578366034,0.2165960819,0.1747603565,0.1609930098,0.6276299953,-0.1296104938,-0.2748526335,0.4868668318,0.0433791801,0.053918384,-0.296800077,-0.0868220106,0.0617774017,1.0295807123,-0.3937677443,-0.3564585447,-0.1362314075,0.0429887623,-0.248009935,-0.0468659699,0.1938485056,0.0802288949,0.3612401783,0.1168773398,-0.3947039247,-0.1558552384,0.0775375143,0.0123679871,0.5747609735,0.0079296753,0.006995487,-0.0464662872,-0.1192316413,0.1567903757,-0.1131044254,-0.078289628,0.0232752524,0.0664356947,-0.2106243074,-0.1160638183,-0.2866198421,0.2796501815,0.498293668,0.0537300706,-0.1367437243,-0.3827081621,0.1011157855,0.2913881838,-0.0538900159,0.5860031247,0.1966705918,0.2218898088,-0.3975260258,0.0835797042,-0.0928669795,-0.0077863103,0.3977659643,0.5791301727,-0.036672432,0.4260240793,-0.2016580999,-0.0679505244,0.152439177,-0.2635843158,-0.1534396708,0.1068716496,0.2219767272,0.1317170113,0.2757698596,-0.124416247,0.0658217221,0.545681417,-0.106338717,-0.0713844672,-0.0793322474,0.0519186817,0.1652719229,-0.0826038718,0.0689334869,0.4504217803,-0.1267027855,0.1890272796,-0.064078778,0.2548243701,-0.059005443,0.2395191789,0.2200737596,-0.3890978396,0.2038203627,-0.0463676862,0.1786973476,-0.3816194534,0.3463989794,-0.0534163564,-0.2060552239,-0.1616810411,0.2132520229,0.1692087203,-0.1648845226,-0.1126016676,0.0359026939,0.0138854189,0.2280050516,0.3043180704,0.1635724306,-0.0215761587,-0.398511678,-0.2363695353,0.0136657599,-0.3425946534,-0.0080059143,0.2239617854,0.1316719353,-0.2075072825,0.2535156012,-0.2940642834,0.1013360992,-0.0034681449,0.0235082712,0.3565423787,-0.4369632602,-0.2732064724,-0.0279662944,0.1864790022,-0.1035629436,0.0481274948,0.4568344951,0.1580266356,-0.2445974052,-0.0531273261,0.1354698092,-0.1033434644,0.1129797995,0.0414424762,-0.0668416023,0.1230785027,-0.0180224981,-0.3156913221,-0.102321066,-0.1786926091,-0.1581577808,0.2403324544,-0.0035172047,0.1564743668,0.4551902115,0.4405421913,0.1935774088,-0.0041880249,-0.2268556058,-0.3842477202,-0.0581124127,0.2511888444,-0.138198331,-0.3032629788,-0.6113615632,-0.1767916977,0.2631886303,-0.0754908696,0.0055043357,0.2138353884,-0.1993266791,-0.2779326737,0.0624175295,-0.2574976385,0.0297128856,0.0468112342,0.1070219353,0.0149146356,0.1145128533,0.0927567706,0.0290389154,-0.1173444837,0.2132788002,0.0408873037,0.0829877034,-0.0345223844,0.145313099,0.3163555562,-0.3232550919,0.1213525236,0.4607732594,0.3032994568,-0.5314986706,0.0622610189,0.1226724684,0.0218913965,-0.0518396385,-0.2709649801,-0.0260816962,0.1027444303,0.3627811372,-0.0343949236,0.2423748374,0.1144367009,0.1110746562,-0.3544942141,-0.2036374658,-0.3353032768,-0.0867559314,0.135872066,0.1299476922,-0.3200688064,0.484347105,0.6123461127,0.0640560612,0.2658021748,0.0981492102,-0.352642417,0.1297808439,0.2959808409,-0.1598252803,-0.0144370655,0.3183603287,-0.0008385434,-0.0637735426,0.0819299892,0.1354631335,0.3058955371,0.2960719168,0.3277533054,-0.1924916208,0.0786267519,-0.0426282175,0.0444987938,-0.3453595936,-0.158987388,-0.2152090371,0.0066773626,0.0380661525,-0.2455751896,-0.112329334,-0.0882889852,-0.5619063377,-0.4075858891,-0.3165303767,0.0344989039,0.1153464913,-0.3554001153,0.3305261731,-0.3464923799,0.0453386232,0.0523465686,-0.2609804571,-0.1679685116,-0.3869166672,0.1574691236,0.0538238436,0.0889512673,-0.0401540585,0.5392515063,0.3545880616,0.159402445,-0.6103399396,-0.2528915405,0.1790262461,0.0170618836,-0.2253209203,0.1091634557,0.1069799885,-0.1393186003,-0.0697339848,-0.0014452366,0.0785157382,-0.0372331813,-0.0896277502,0.2730149329,-0.0314972065,-0.0204621516,-0.3882804215,-0.4076157212,-0.4055927694,0.3017730117,0.1276456714,0.1697412729,0.1785371453,-0.2694990337,0.0260159299,-0.1096279845,0.0952151269,-0.1074527949,0.1441744864,0.2464603335,-0.1360030472,-0.0080136368,0.0874869153,-0.1389940232,0.17115435,0.0423163511,-0.0191368796,-0.1977375746,-0.0855146572,0.3730426431,-0.1065643579,0.1364077777,0.2442816198,0.3777547777,-0.1017339379,-0.0099786222,-0.1178264394,0.2335264981,-0.0354419388,0.089684315,0.0194216222,0.0804448053,-0.1720803976,0.4238517582,0.0187355634,0.1440928727,-0.0549170449,0.0537979268,0.0363465585,-0.0803152546,0.0127990982,-0.2057098895,-0.0890100896,-0.0760111883,0.0934896171,0.1513704956,-0.1037450135,0.1576051265,0.1566733867,-0.1015150324,-0.1684867591,0.0172242075,-0.1306100637,0.1751759201,-0.0098245917,0.0268270727,-0.037248902,0.0644017234,0.244250685,0.0227099694,0.0414308794,-0.1529840529,-0.4404383302,-0.0904365554,-0.4571638107,-0.0186682735,0.1574329436,0.067332767,0.1258171946,-0.1683335304,-0.0783452988,0.2440267652,0.2816129923,-0.1813260168,-0.4476846159,-0.0188780259,0.1774274558,-0.0620651506,0.0060735582,-0.1370530576,0.0500413068,0.0479589477,0.0873184577,-0.3140090108,-0.1462479085,0.1048215255,0.0610036328,-0.1581060588,-0.0232596826,-0.1054871082,-0.3103759289,-0.072210215,-0.1279830486,0.3736522496,-0.22800605,0.115080297,-0.0935125574,-0.1260614097,-0.2332110554,0.0052719163,-0.0157447234,0.4306237698,0.025954809,0.1231583357,0.0240179673,0.4069707394,0.2633194327,0.5096574426,-0.0239819288,-0.5610494614,-0.3903732896,-0.3155612051,0.4798541069,-0.075795874,0.0816516876,0.2385451496,-0.5130556822,-0.1403168887,-0.2198767215,-0.1562792659,-0.1549161077,-0.0766614303,-0.300317198,-0.4329050481,0.8527727723,0.1414794773,-0.1323336363,0.5135406256,0.0858340859,-0.2635056973,0.3207180798,0.3549795747,0.9080423713,0.1340634078,0.2197116762,0.2137010396,-0.1758752912,0.5592682362,-0.2411009073,-0.0584962964,-0.4896416664,-0.1990523934,-0.1038560644,-0.0818371773,-0.0474385992,0.1922317296,-0.3542010188,0.1373840421,-0.2455499172,-0.1133809835,-0.0277624037,0.0846347213,0.0244097319,-0.3767651618,-0.0888493136,0.1496566236,-0.2634575069,-0.0647317991,-0.1604021937,-0.185431987,-0.1244848594,0.0089534037,0.0690332279,0.0361811668,-0.0694402531,0.3164689839,-0.0331775099,0.0394300632,0.1400249898,-0.2949136794,0.3375794291,-0.0080010062,-0.0537775718,0.0220292974,0.1087084785,0.1741507351,-0.0164033361,-0.0383144394,0.3151322305,0.084761031,-0.0962397233,0.152751267,-0.082610935,-0.3614305258,-0.205161199,-0.1234936789,0.2789392471,-0.1750292629,0.0053626667,0.2863708138,-0.4388440251,-0.1329395473,0.1619930267,0.2034303993,-0.1601743996,0.2107394189,0.0597958155,0.0342131816,-0.01033027,-0.2540884018,0.3012002409,0.0377209336,0.0974835083,-0.1277719289,-0.1516406387,-0.2891377807,0.3465970159,0.0953676403,-0.182810843,0.1077958867,0.0206010416,-0.0663892254,-0.2406278402,0.1129579321,0.0351880416,0.0675411969,-0.3194276392,0.0897001103,-0.4712828994,0.401877135,0.0008574196,0.2631204426,-0.269004494,-0.0221454147,-0.1851147711,-0.0290774684,-0.3835372031,-0.3776127994,0.0324241519,0.0163850244,0.2647833824,0.1624726206,-0.0006888488,-0.0970997587,0.1627370417,0.0842982158,0.0273614358,-0.2790388167,-0.1625831872,0.0456121601,0.122341603,-0.0057719145,-0.2549293935,-0.1305669099,0.0186504181,-0.1509663761,-0.1034641713,0.0977511927,0.1634016335,0.3027569354,-0.0969430655,-0.0073341141,-0.1510326117,0.2586010396,0.0882661194,0.0381171815,0.0888106227,0.2020591795,-0.1580035537,-0.0724754557,-0.1776504666,0.115978457,-0.2789205909,0.0626973584,0.1140846908,-0.0370479859,0.0618271306,0.3791281879,0.1837725639,0.1133081838,-0.1546559185,-0.0656302497,0.2212628722,0.4003598988,-0.2861902118,-0.4286292791,0.0216550268,0.053979665,0.2183307111,0.321728915,0.1728089899,0.373780787,0.0311294682,-0.0377844237,-0.074126929,-0.2500039935,0.0995302796,0.3859386444,0.2482805252,-0.0500040986,0.4153280258,0.3016577363,0.4897958636,0.3244734704,0.1961953938,0.5447490215,0.5794219375,-0.1124871671,0.2589476407,0.2721155584,-0.1472965181,0.0009824876,0.0080813561,0.1034234166,0.2910740376,-0.0611256957,0.1336740255,-0.5219960809,0.0359727181,0.1585479081,-0.0773001462,0.0589157008,0.0387311913,0.0738617182,0.0119208386,-0.3031667173,0.055681549,-0.3524497151,0.2914430499,-0.0364606269,0.3144467175,-0.2338736355,0.0263108928,0.2244738042,0.4883762896,-0.0834985077,-0.2560598254,-0.1256843209,-0.1655036509,-0.1606784463,0.0785372704,-0.1051097736,0.0427173078,0.1470707655,0.1667228192,-0.3128276169,-0.1395053566,0.0200569332,-0.0416333303,0.0261395797,-0.255317986,0.2161889076,0.1669336557,-0.0802818537,0.2079637647,-0.0241285805,-0.1080368012,-0.0740808547,0.0546040386,0.3698731065,-0.6536043882,0.0966494009,0.0538565367,-0.1759699732,-0.2538166046,0.6161900163,0.0870841742,0.096521847,-0.1378525496,0.1350959092,-0.0521588773,0.2209243625,0.2347933501,0.0537722409,0.0318052284,0.0512791462,-0.2802792788,-0.0921993032,-0.1210640669,-0.5091294646,0.3963055909,0.059604343,0.1690883487,0.4241621792,0.1171318889,-0.0809314102,0.1277048737,-0.0703379735,-0.334805131,0.0001409011,0.1554784328,0.349424988,0.0199729651,-0.1777938008,0.0416195616,0.1490759104,0.1344936788,-0.4543763995,0.0134989778,0.1714762151,0.3090107739,0.4693944156,-0.1916486919,0.0279902723,0.0782829523,-0.1085646525,0.1821728647,-0.3972954452,-0.0272018593,0.0350831226,0.0457653403,0.3254002929,-0.1416345686,0.0597179979,-0.275221169,0.1590811312,-0.1686425507,-0.0186356865,-0.2717111409,0.1797593832,-0.0870108828,0.1033272818,-0.0918986797,0.3254535198,-0.0733680949,-0.0042084926,-0.047981184,-0.2741729915,0.4397296906,0.2359516323,-0.3103941679,0.1537979394,0.2523867488,0.247675404,-0.0316913724,-0.4815124571,-0.1608425826,0.3907756507,-0.0851691291,-0.2077005059,-0.0706798509,0.1268517971,0.038436491,-0.1619018465,0.3741303086,-0.2902053893,0.0696642771,-0.1280392706,-0.2311172932]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/4","title":"[Feature] Keep the list of labels of a dataset as metadata","comments":"I would have a preference for the second bullet point.","body":"It would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.","comment_length":10,"text":"[Feature] Keep the list of labels of a dataset as metadata\nIt would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.\nI would have a preference for the second bullet point.","embeddings":[0.0293631349,-0.1155054048,-0.1700816751,-0.0855561122,0.1147266328,0.2889238596,0.3354568183,0.1456108689,-0.0391280055,0.1645747274,-0.1343861371,0.508287251,-0.2241065055,0.2445494235,-0.1276418567,0.012875095,-0.0614322536,0.0968658775,0.1445268095,-0.0327437297,-0.2494076192,-0.2335203141,0.0732480437,-0.0196467582,-0.2349418551,-0.0671437159,0.2197063416,-0.1342632025,-0.2024480551,-0.5270950794,-0.1980384141,0.1842437685,-0.1053666398,-0.2167119831,-0.0000976233,-0.0932864994,0.1665011793,0.0139487758,-0.1421146691,0.0862590298,-0.5326735973,-0.057015378,0.1112352535,-0.2279454172,0.0710529536,-0.0664561391,-0.1489659697,-0.1670911312,-0.2293688655,0.0890429094,0.3654920757,-0.1167983636,-0.1181882024,-0.0421036035,0.3731845319,0.2562757432,-0.4062798321,0.0188663378,0.1401552558,0.207828775,0.1568725407,0.6802257299,-0.0925508663,-0.1941654682,0.5777368546,-0.0301727727,0.057900507,-0.3308675885,0.0740896612,-0.0511584245,0.9914309978,-0.28361395,-0.1064194739,-0.0309693497,-0.0332931951,-0.351216346,0.0469574519,0.0286442935,0.1562867761,0.2269727588,0.0759441927,-0.5117239356,-0.1343560815,-0.0837039053,-0.0352762416,0.486716181,-0.0675417632,-0.0890751928,0.0977420285,-0.1662443876,0.0714858323,-0.0084275454,-0.0666198581,-0.1097609252,0.0712882876,-0.220387131,-0.0087112421,-0.126612246,0.2281254679,0.3837567568,-0.1104578301,-0.1356578469,-0.3003953993,0.1470444202,0.173389256,0.0551691093,0.5556361675,0.0576207638,0.193176657,-0.2940514982,0.0169160739,-0.1918577552,-0.0322023481,0.284262985,0.3659068048,0.0000796334,0.4380069673,-0.3349049985,-0.303882122,0.0127337864,-0.0542830564,-0.2575193048,0.1386717409,0.2540123165,0.2054247707,0.0807438344,-0.2998623252,-0.03120175,0.3855664432,-0.2087136209,-0.1481612325,-0.0409316085,-0.1580993235,0.0860511661,-0.106703423,-0.0550019965,0.4361275136,-0.0484784544,0.0419687964,-0.0646527112,0.1543109268,-0.1556528807,0.1624344289,0.1833079606,-0.3666541874,0.2202202231,-0.0449558944,0.0563018136,-0.3380034864,0.3447945118,-0.2276990563,-0.3190176189,-0.3217469454,0.303706944,0.1989821494,-0.2674617767,-0.030594429,0.1397759467,-0.0089408001,0.0978407934,0.2904456854,0.1919048727,-0.0517072827,-0.2905842066,-0.3571213484,-0.0024291002,-0.3665712774,0.122284174,0.1344277263,-0.109695904,-0.0086258492,0.1961066872,-0.3544827104,0.2320493609,-0.0589431264,0.2861635089,0.2865117192,-0.2762065828,-0.1640850753,-0.0291478857,-0.0129812872,-0.1851045489,0.023031516,0.3416969776,0.119349353,-0.3289370835,0.0576023608,-0.2153084427,-0.1176959425,0.1182277352,-0.0456952825,-0.0986547023,0.2802162766,-0.1168975011,-0.2748385072,0.0945385173,-0.1193174347,0.0977855399,0.1927859932,-0.0143712088,-0.0319283903,0.3965570629,0.5150435567,0.1218417585,-0.1631626338,-0.1682334542,-0.3645655215,-0.0774930418,0.1846098006,-0.0274414271,0.0269178264,-0.5255357027,-0.194751054,0.1179134101,-0.0723055825,-0.1164974943,0.3166493475,-0.0144461226,-0.3392838836,-0.0226134956,-0.3307933211,-0.0871046185,-0.1194623411,-0.0160945673,0.0026151382,0.0107084215,-0.0490803979,0.1372010857,-0.0674502328,0.0732338205,0.0991355553,0.0904839635,-0.0761967003,0.3091894984,0.310023278,-0.1852348894,0.0981465206,0.5589535832,0.2525705397,-0.4247962832,-0.1068705022,-0.0662422478,-0.0453117937,0.1039235219,-0.3207565248,-0.1803239286,0.0944962129,0.1153637096,-0.0113998959,0.0250142124,0.0377020277,0.0711887255,-0.2363298833,-0.2525538802,-0.2609181106,-0.0628478527,-0.0824598521,0.2729126215,-0.4908310473,0.6171912551,0.746331811,0.1077093631,0.1486454159,0.1882977486,-0.1215975285,-0.0594343208,0.381865412,-0.1565885395,0.0737794191,0.327349633,0.017138347,-0.0510080121,0.0610906556,0.0396685749,0.2763455212,0.2166715562,0.2809692323,-0.1169894785,0.0609551184,0.0305745658,-0.0809142888,-0.2426536381,-0.0739465654,-0.0445464961,-0.0087260026,-0.1341098994,-0.3153580129,-0.2026081979,-0.0154910451,-0.3056362569,-0.3594334126,-0.2527287006,0.1839900017,0.1616832465,-0.2925862074,0.4213606119,-0.2673611939,0.2812131941,-0.1215308309,0.0517118536,-0.0357982814,-0.0911067873,0.2202523947,0.1849782616,0.0002467555,0.1033456326,0.4883606136,0.1968716383,0.2178294063,-0.5993577838,-0.3195835054,0.2563534975,0.0524729192,-0.1525758952,0.1146904528,0.1339157075,-0.1988527477,-0.1582288444,-0.0660651401,0.0767699555,-0.0531890132,-0.2592608631,0.1213608086,-0.0179240424,-0.0522591025,-0.4172685146,-0.2143857628,-0.4609072506,0.1729961336,0.0000452881,0.2046743631,0.231676206,-0.2372834533,0.0349810012,-0.1767355204,0.1649785489,-0.2417177707,-0.0997990891,0.0161346961,-0.1851119697,-0.0064692497,0.1020333096,0.1180599928,0.1175624505,-0.1119825095,-0.0087343799,-0.3104288876,-0.0067088981,0.268055439,0.0131676123,0.0047988817,0.3132352233,0.3182920516,-0.2525475919,0.0175135862,-0.2876842618,0.3120698929,-0.0231176242,0.153870225,-0.0328892544,-0.0563321672,-0.1562765688,0.310026288,0.030610295,0.1131650954,-0.1069273353,0.0519238412,0.1778342426,-0.1403193176,-0.1016229242,0.0497970209,-0.0174714774,-0.1231159419,0.3137035966,0.0050025797,-0.1039185375,0.1992286742,0.0594747961,-0.0157395098,-0.1324871033,0.030968383,0.1217619181,0.2026556581,0.0386287607,-0.0658455715,-0.0772148222,0.0620655417,0.222931236,0.1858320981,0.163164407,-0.144255802,-0.5091451406,-0.1402577013,-0.3970173597,-0.0916294083,0.1157711819,-0.0766397342,-0.0364926346,-0.2160254419,-0.1845762581,0.2108242512,0.2426296622,-0.2459412515,-0.3399982452,-0.0626247898,0.221957013,0.2840518951,0.0444555469,-0.0168387983,0.0920192823,0.1536584795,0.2163560241,-0.2012830526,-0.1775057167,0.2528308034,-0.0860018507,-0.1124704257,-0.0409399606,-0.0759316832,-0.2729162872,-0.1317006052,-0.1199992672,0.2626651525,-0.2285382599,0.0541083924,0.054576043,0.039651107,-0.2076414526,0.0843709409,0.1129943207,0.340995729,0.1705305278,-0.0278130323,-0.0117341187,0.3839263022,0.2778154314,0.4740895331,-0.0903165042,-0.392192781,-0.1299423873,-0.4084880352,0.3755406737,-0.0058596414,0.0410697497,0.2361614853,-0.42234236,-0.0661218166,-0.2248270959,-0.0246575605,-0.0506050698,-0.120130986,-0.1650893241,-0.2483627945,0.60002774,-0.0515270531,-0.1207789257,0.3429099321,0.095471248,-0.2221287936,0.1095691249,0.3870481551,0.9329563379,0.2158791572,-0.0533486418,0.1089625731,-0.0051520928,0.4670397043,-0.1995007694,-0.0341705941,-0.3818032742,-0.4078201056,-0.0519884937,0.0484660715,0.0020992742,0.0461547375,-0.3114880919,-0.016582774,-0.1116350517,0.0978526324,0.0251094438,0.058750011,0.2792433202,-0.3199702203,-0.0237636101,0.2399484515,-0.2019415051,-0.0836633965,-0.1539947242,-0.1056864336,0.0804490149,0.0970474333,0.1149585769,-0.041414313,0.0112455795,0.2076626122,0.0108664874,0.1626600772,0.07083188,-0.296146363,0.3953908086,0.0384122692,-0.0977376401,-0.0362519734,0.2942314744,0.083241865,-0.1687631756,-0.0921263471,0.2872538567,-0.0876586661,-0.0803339407,-0.0153924804,-0.0758094937,-0.2622149587,-0.2399798483,-0.2419921607,0.1996817887,-0.2033775002,-0.0497335605,0.3355275393,-0.3405804038,-0.2305642068,0.2280198485,0.1104580462,-0.0398879573,0.0952288881,0.0067399703,0.0969547257,-0.1452371478,-0.1831871569,0.2275594622,0.019642869,-0.0406491347,-0.2151032835,-0.1773633212,-0.4102487862,0.2690861821,0.0514975786,-0.1699638069,0.1337250322,-0.0383066311,-0.1462388933,-0.1750444919,0.2692436874,0.0916305557,0.0916952491,-0.1592703313,0.1456498653,-0.3970549703,0.3116949797,-0.1535526961,0.2798321545,-0.3818862438,0.12715514,-0.0474860929,-0.105287917,-0.5086778402,-0.3076809645,-0.206594795,0.0280528162,0.3604106009,0.063332893,-0.0792186707,-0.0634498224,0.2065096647,0.1012878194,0.0045686942,-0.3690008223,-0.1541891545,0.0525771417,0.0457245596,0.0966603085,-0.1158901677,0.0506719127,-0.1548327804,0.0159268137,0.0102286106,0.003445125,0.1346754581,0.3203670979,-0.2030471116,-0.0291082785,-0.1092294157,0.0500102602,0.2414834499,0.1209510118,-0.0136446534,0.0847910792,-0.2779938877,-0.0250456557,-0.1911264062,0.2183375657,-0.224212274,0.1879443675,0.0030418162,-0.0316051953,0.0795314014,0.3583785892,0.1132599115,-0.0495669954,-0.197298035,-0.1813026071,0.1939198226,0.4950839579,-0.3167037666,-0.2517769039,-0.0151647441,-0.1191818938,0.2578955293,0.3807979226,0.3594509661,0.3687084615,0.258530885,0.095388703,0.0783028826,-0.3733274341,0.1544091702,0.3238869011,0.0223993491,0.1175339967,0.5194658041,0.4067981243,0.4247867763,0.2606678307,0.3275885582,0.5395296812,0.4494192004,-0.0209759101,0.2939913273,0.1614499241,-0.0022673674,0.2340125442,-0.0278205965,0.0132077709,0.2519075274,-0.1801075339,-0.0412426628,-0.4753208458,0.030594945,0.1864268333,-0.1187018082,0.0530144237,-0.0072867954,-0.0969926417,-0.042261295,-0.2770887017,0.0448363088,-0.0530334003,0.2590140104,-0.2794308662,0.187245965,-0.1004588082,-0.0149079552,0.0382001437,0.4100560546,0.0462177657,-0.0535165556,-0.1873351932,-0.1051606238,0.0661433563,0.3526640832,-0.0973153561,0.1855368167,0.3282094896,-0.0887401104,-0.1850625426,-0.2142762691,0.0358252823,-0.0511681587,0.0901203454,-0.136347115,0.3121339083,0.2222328484,-0.192411676,0.4421311319,-0.0703689381,-0.0490778461,0.0068507977,-0.0585423261,0.5139121413,-0.5202402472,0.0520928279,0.0025407157,-0.1378689557,-0.2414245158,0.5240742564,0.0741697699,0.120551303,-0.1086953655,0.1904869825,0.0237877239,0.2624394,0.132757023,-0.0778857321,-0.0353010707,-0.0480220541,-0.3655478358,0.0909809098,0.0613607131,-0.37662673,0.4014663696,0.1003997922,0.0847956389,0.3992516398,0.0748316944,-0.2993969619,0.0958550349,-0.2252707779,-0.2548469305,0.1245557368,0.1526490152,0.1999219656,0.0274677612,-0.1359715611,0.1311460733,0.0834694579,0.2109908462,-0.3626901209,0.0062981909,0.0788330585,0.3330643773,0.4325463176,-0.0502489693,-0.0313294977,-0.1235998347,-0.1263602674,0.1279098094,-0.2984011471,-0.2024737895,-0.1165564433,0.2075112462,0.3578585982,-0.0166751184,0.123992987,-0.3874228299,0.0362362005,0.0548281036,0.1785241812,-0.4267524183,0.3297547698,0.0264446922,-0.0294060912,0.0806065723,0.222508505,0.065313153,-0.0271705482,-0.2159770876,-0.2258985788,0.4730822742,0.2977980077,-0.2629348636,0.0684802458,0.352374047,0.2424940169,-0.0033522048,-0.5809319615,-0.1460276395,0.4197126031,-0.0886651874,-0.2584050894,0.0746999308,0.0087102419,-0.0585333519,-0.0553145148,0.26551193,-0.1166743338,-0.0439825654,-0.1050012261,-0.120362781]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/4","title":"[Feature] Keep the list of labels of a dataset as metadata","comments":"This should be accessible now as a feature in dataset.info.features (and even have the mapping methods).","body":"It would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.","comment_length":16,"text":"[Feature] Keep the list of labels of a dataset as metadata\nIt would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.\nThis should be accessible now as a feature in dataset.info.features (and even have the mapping methods).","embeddings":[-0.0496049151,-0.0984715894,-0.1723788232,-0.0273312796,0.1680133641,0.1654706001,0.3242646754,0.2305074632,-0.0478426032,0.2778837681,-0.2302711606,0.492888689,-0.1844331771,0.2717520297,-0.0722042844,0.0264097042,-0.1009540185,0.0530562438,0.0414123982,-0.0347817875,-0.3080717027,-0.1635333449,0.1349450946,0.0643810108,-0.237500295,-0.0460765511,0.1412658542,-0.195451051,-0.0466829576,-0.6009489894,-0.1726687551,0.2099639624,-0.075079076,-0.1280405819,-0.0000924084,-0.1215024292,0.1399082392,0.0204979815,-0.2751190066,0.0876628309,-0.4400596917,-0.0634489655,0.1277536601,-0.3124091029,-0.030301543,-0.1620437503,0.021702636,-0.2083199173,-0.1103460044,0.0456910953,0.4085198641,0.0035763024,0.000951154,-0.0858163834,0.4976731837,0.1833201498,-0.4111306965,-0.0564074665,0.2255617231,0.1022936627,0.0467004068,0.5869672298,-0.0641263872,-0.1524451226,0.3551450074,0.0096309436,0.1121637821,-0.3324599266,0.0456431322,0.102847904,0.9027412534,-0.3083907068,-0.2470870912,-0.1317891926,-0.0125023052,-0.1988036931,0.0801150128,0.0526473895,0.0992975682,0.1977138817,0.0551455878,-0.3937343657,-0.1392278224,-0.0967758074,-0.0178305861,0.4943385422,-0.1232156828,-0.0832449645,0.103380844,-0.2010066211,0.1059985161,-0.0191349257,-0.0495446175,-0.0015812089,0.0675469264,-0.2441413254,0.0989901572,-0.1532108039,0.2423734367,0.3187515438,-0.0798826888,-0.0288356263,-0.3631315231,0.1258137077,0.209936142,-0.0072107795,0.4834262133,0.1031986848,0.2926647365,-0.2207577825,-0.1145367622,-0.1516707838,0.0008645065,0.2730922401,0.3092392385,-0.0071154884,0.3624818921,-0.3352712393,-0.2529607117,-0.0697743148,0.0482400991,-0.1187166423,0.1934691668,0.3123272359,0.0354918763,0.047175996,-0.2396335304,-0.0242575929,0.2669627964,-0.1388695389,-0.2123287916,-0.1292961538,-0.0829126015,0.0255032964,0.0194381252,0.0227207169,0.3629740775,-0.1533164829,-0.0136335632,-0.0287381373,0.1597516388,-0.0463826209,0.2886559665,0.1537839919,-0.2408829629,0.1265989691,-0.0432015248,0.0505391397,-0.3477985859,0.0889652073,-0.2663639188,-0.2532556951,-0.2703177035,0.3592046201,0.1903045326,-0.1899776608,-0.0029621404,0.1976010948,-0.0891306251,0.0178295821,0.0847133398,0.1344020367,-0.1058998704,-0.2697375417,-0.1391242892,0.0331959985,-0.3472645581,-0.0254422724,0.1159432456,-0.1032407358,-0.0487671383,0.1077130139,-0.3057391047,0.0840149745,-0.1304382533,0.1342795938,0.2191059738,-0.1828376204,-0.2081879526,-0.0410217308,-0.0110607604,-0.22094585,-0.0191750973,0.3735283017,0.1456188858,-0.2916695476,-0.0252589881,-0.0990169123,-0.072570242,0.1471474916,-0.0109407846,-0.0757462531,0.1559707969,-0.0825422555,-0.219349876,0.1205210015,0.0938543454,0.108762823,0.0994633436,-0.069374688,0.1036927253,0.4537687898,0.4328600168,0.1737988144,-0.0954231545,-0.2658981085,-0.3110403419,-0.0317870229,0.062756367,-0.0624678358,0.0312345717,-0.5854985118,-0.1437659562,0.1872646958,-0.014401502,0.0042471611,0.3816923201,0.1099592149,-0.3010726571,-0.0189692788,-0.2340240628,-0.0605973862,-0.0038256675,-0.0431618318,-0.0519668497,-0.0022551881,0.0235307403,0.0451432839,0.0022891283,0.135599494,0.1382024139,0.0514880531,-0.0487503,0.2879352868,0.2015812844,-0.1524143219,0.1315054297,0.4913338125,0.267747283,-0.3744827211,-0.1273801625,0.0072628055,-0.0027249097,0.026938403,-0.3749998212,0.0071977749,0.1368162632,0.1164636984,0.1178022772,0.1408797503,0.216995582,0.0723883733,-0.1767579913,-0.3400178552,-0.2923746109,-0.1043406054,-0.0824241117,0.0987219289,-0.477979064,0.5590272546,0.7271511555,0.1874971688,0.2642273903,0.0688684508,-0.2115924507,0.021986587,0.3479870558,-0.0654762313,0.0490362495,0.4268992841,0.2220215499,-0.0561690293,-0.0363544859,-0.0129512409,0.2276465297,0.1589504331,0.1350108385,-0.1153644174,0.1157374904,-0.0805163011,-0.1274996698,-0.2985969782,-0.0738099664,-0.0514241382,-0.0466167927,-0.1763063222,-0.3657177389,-0.1905111521,0.029067006,-0.2515901327,-0.3413636386,-0.3003245592,0.2787344158,0.2025237679,-0.3914099634,0.4054634869,-0.2809786499,0.180066824,-0.012809217,-0.0476994738,-0.0472429618,-0.2777006626,0.0947754681,0.2540524006,-0.0370873697,0.079837583,0.45062235,0.1521618217,0.2697623372,-0.6478309035,-0.3487976193,0.2086029351,0.0181230512,-0.0781253949,0.1263718158,0.1510206908,-0.144692868,-0.0811225548,-0.0505003259,-0.0534176379,-0.1017615944,-0.1608781815,0.1139837503,-0.089772895,-0.1069203094,-0.4496890008,-0.1781235337,-0.5046770573,0.2615948915,0.0235694367,0.2280349433,0.179664582,-0.2393412888,0.1410049647,-0.1516529322,0.1815710217,-0.2869572937,-0.1689717621,0.1275098324,-0.268332094,-0.1008078903,0.1138846576,0.0540677682,0.0919373706,-0.0693853572,-0.057525456,-0.197721526,-0.1064798087,0.3291487992,0.1518095285,0.1072690412,0.3619349301,0.317442894,-0.2667703927,-0.0292521175,-0.226419583,0.277375102,0.0564540923,0.1877759546,-0.0833340362,-0.0693659559,-0.1662255824,0.2725535929,-0.0823972896,0.0848290026,0.0662150308,-0.0174633972,0.2688052654,-0.1162871644,-0.136541307,0.0221418776,-0.06840121,0.0213083643,0.2833125591,0.1046236157,-0.1605488658,0.0719438642,0.0609424114,-0.0409467444,-0.2158065885,-0.0451207981,0.0695250258,0.1812651306,0.0980483368,-0.0176504571,-0.1575074792,-0.0406275913,0.1534331441,0.0189441405,0.0940211788,-0.1688886434,-0.6407893896,-0.0228613261,-0.2375487685,-0.0048509273,0.079965435,-0.1354018748,-0.0997044668,-0.2216232866,-0.1857279241,0.1717873365,0.2978361845,-0.2914413512,-0.1940685809,0.0473770685,0.174042508,0.1496276706,0.1061806604,-0.1042316183,0.0721667409,0.1377886832,0.3421629369,-0.2507076561,-0.130092904,0.2145001739,0.0027836969,-0.1060275808,0.0286506303,-0.0539324172,-0.2279928178,-0.1222900748,0.0036638402,0.1679654419,-0.126489982,0.0231005102,-0.0387707762,-0.1081668362,-0.1523584723,0.0796496868,0.2087122947,0.2823014855,0.1310623437,0.0923201293,0.0415388085,0.3975786269,0.2388134748,0.4212321937,-0.0804597959,-0.3947167695,-0.0684074238,-0.2574702501,0.3553962708,-0.0132448478,0.0715203211,0.1550484598,-0.3175726831,-0.0690394267,-0.235242039,0.0467910767,-0.0560228899,-0.0785978958,-0.2056134641,-0.262540102,0.5939211845,-0.0425712243,-0.1731755883,0.3066683412,0.0504121073,-0.2644334435,0.1184930503,0.3276302814,0.8606021404,0.1222230345,-0.0537577681,0.0809399188,-0.064808853,0.5247746706,-0.1475501657,0.0755364001,-0.4238331914,-0.3841570318,-0.0377392396,0.0851701573,-0.0223440565,0.1731523722,-0.2174615711,0.0564646833,-0.0340877101,0.1059449315,-0.081872344,0.129647851,0.1797146052,-0.4037360251,-0.0829447582,0.3114581406,-0.1685132533,-0.0216136985,-0.0532036424,-0.1298722327,0.1390577108,0.0984611362,0.0490276292,0.1109285653,-0.0927612633,0.147949636,0.015854001,0.1491802037,-0.0347529538,-0.1864575893,0.4739907384,0.0897607058,-0.1023961827,0.0387800261,0.1525944173,0.1193607226,-0.0934809893,0.0345099494,0.3152235448,-0.0283829048,-0.1219002008,0.0688555688,0.0050387145,-0.2486807406,-0.2861433327,-0.1798698008,0.1733433455,-0.1654538065,0.1035719588,0.1675272733,-0.2941785753,-0.3173500001,0.2873852551,0.0916209072,-0.0947530046,0.1600699574,0.1074649915,0.1371638924,-0.1262545139,-0.0400859527,0.0961308703,0.0823537111,0.0036407269,-0.1572650373,-0.1986840218,-0.4302117527,0.2172503918,0.056150239,-0.1555856168,0.1959596723,-0.0319295749,-0.2095275223,-0.0483820327,0.1660515219,-0.0033530309,0.0846330076,-0.1702111065,0.1072602421,-0.4351981878,0.2706967592,-0.182945773,0.345410645,-0.2357186377,0.0502329879,-0.0859367624,-0.0480709225,-0.558170855,-0.2756898403,-0.2299216092,0.0350648165,0.2522653937,0.0905505717,0.0081944717,-0.0666727051,0.2577536404,0.0664045289,0.0030390371,-0.4120518863,-0.219226405,-0.0005245918,-0.0515116341,0.1661136746,-0.219269082,-0.0418004096,-0.1061743423,-0.1460494995,-0.0169929452,0.0699336901,0.103895627,0.3313807249,-0.202977404,-0.0209295582,0.0031163334,0.1315289736,0.2356397808,0.0610738508,0.0241907164,0.0824208111,-0.3954671919,-0.0931766629,-0.1182155311,0.2054079473,-0.2415837646,0.0571347252,0.1239662915,-0.0197928194,0.0389057994,0.4792376161,0.1772829145,-0.16278404,-0.0241418369,-0.1127079874,0.2879496813,0.4854298234,-0.3828673661,-0.3012910485,0.1474419385,-0.0106521081,0.3686859012,0.335085541,0.2474940121,0.2425523251,0.2311017662,0.0421506651,0.0638672039,-0.3651693165,0.0823835209,0.3219587207,0.1054731384,0.0369813442,0.423851043,0.3885628581,0.4059867263,0.3636251986,0.2105579227,0.4172862768,0.4277817309,-0.0787116438,0.3518086374,0.1038472503,-0.064244695,0.1441380531,-0.0626457334,0.1139072552,0.1288525611,-0.0965228304,-0.1344093382,-0.3736048639,-0.0066508846,0.2293714881,-0.1778656989,-0.0393972322,-0.0244953036,-0.1111564115,0.0197917111,-0.2580574453,0.0502885506,-0.1362705529,0.2617529035,-0.3673520982,0.3187538385,-0.2376466841,0.0193837602,0.0612944439,0.3824278116,0.0097051738,0.0181866232,-0.1317526698,-0.1554831564,0.0369391292,0.2386711836,-0.0669498965,0.0630168617,0.2173569202,0.0632435456,-0.2458054423,-0.2655652761,-0.1004000232,-0.0394907594,0.0020971969,-0.1866284311,0.3683693409,0.2563747466,-0.2082833946,0.3538438678,-0.057259459,0.0278933626,0.0193740632,-0.0197794214,0.4352951944,-0.4933036268,0.0994581208,-0.048345264,-0.1053133681,-0.3252818882,0.5150409937,0.0269297827,0.1437045485,-0.1438029259,0.2109610289,0.078725636,0.240708515,0.1314755231,-0.1165205538,0.0942088217,-0.0172769856,-0.4910584688,0.0508875512,-0.0163881648,-0.3304605782,0.2815282643,0.1093787178,0.1281301379,0.2880233228,0.0509758964,-0.3023867905,0.1641387194,-0.1748100817,-0.2904703617,-0.019254053,0.1376743019,0.1960679144,0.0584706813,-0.1869938523,0.1094982624,-0.0065719229,0.2122161686,-0.282808125,0.078359887,0.0352970399,0.2787852585,0.335129261,-0.0861840695,0.0458692685,-0.183117941,-0.0132579831,0.0696531311,-0.3647841811,-0.0705898404,-0.1177367195,0.1854351163,0.3088005185,-0.0070980019,0.095221132,-0.2999936044,0.0367259718,0.0878407136,0.1955060065,-0.2710282505,0.3101607263,0.0018104687,-0.0283746105,0.0632261708,0.217079699,0.0452684946,-0.0553325377,-0.0132549312,-0.3287933469,0.4158826172,0.1384561807,-0.3457223177,0.0865024403,0.2781175673,0.2457077801,-0.0114862565,-0.4668861628,-0.084787257,0.4394419491,-0.1099125594,-0.2286850065,0.023750443,-0.0242877286,0.0577272251,-0.110079959,0.2805046141,0.0308379121,0.0172548201,0.0041121664,-0.1801844537]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/4","title":"[Feature] Keep the list of labels of a dataset as metadata","comments":"Perfect! Well done!!","body":"It would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.","comment_length":3,"text":"[Feature] Keep the list of labels of a dataset as metadata\nIt would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.\nPerfect! Well done!!","embeddings":[-0.0049427026,-0.1597746611,-0.1961742938,-0.0051457658,0.1053665429,0.1452949941,0.3001495004,0.1979843974,-0.0403422602,0.163125515,-0.123466298,0.5530841351,-0.2473976761,0.2839092016,-0.1811384857,0.0297433455,-0.0983592495,0.0971284658,0.1255857944,-0.0606868081,-0.2341280878,-0.1253540963,0.0262677409,0.022950897,-0.2908366621,-0.0303058587,0.1890719235,-0.1540237665,-0.2558837831,-0.466140449,-0.1371995062,0.1662941873,-0.1736577302,-0.0808503106,-0.000096654,-0.1044963673,0.2082877755,-0.0077037709,-0.1397406906,0.0939582661,-0.4590600729,0.024942603,0.0838402882,-0.1802059263,0.0040844604,-0.135450989,-0.1818469316,-0.2181414515,-0.1461574137,0.0798633322,0.4114482105,-0.1234153062,-0.0750244632,-0.0490121655,0.4761606455,0.2670746446,-0.3902377784,-0.0076924106,0.1799071729,0.1295792311,0.1330809891,0.7216393352,-0.0679600686,-0.2717407644,0.4838833511,-0.0405855738,0.0715406984,-0.3105361462,0.0801098123,-0.0427716039,0.9933025837,-0.2302710712,-0.0657023042,-0.0136260791,-0.1222143248,-0.2838305235,-0.0100116124,0.0113500906,0.0907816365,0.1683920473,-0.010208617,-0.4722045064,-0.1826412529,-0.0819756463,-0.0370308869,0.539804101,-0.1269577742,-0.0969839469,0.0903788507,-0.1523478031,0.0848117173,-0.0039303391,-0.1348201931,-0.0243925843,0.0262977947,-0.2217567861,0.094714202,-0.1503588706,0.2545779049,0.4379733205,-0.0661695376,-0.055802986,-0.2937495708,0.1409989148,0.2021223605,0.1088422462,0.5464553833,0.0503693782,0.2371758372,-0.2717501521,-0.0248766728,-0.1401839852,-0.020864496,0.2568811476,0.3013596535,-0.0157584455,0.4407480061,-0.4111038148,-0.3162680864,0.0507045537,-0.0091274437,-0.2446366847,0.1288081408,0.3065906465,0.1361767948,0.0517671593,-0.221197769,-0.0347798318,0.3173003793,-0.0973022357,-0.1966799945,-0.0273517389,-0.1651950777,-0.0117119625,-0.0791479275,0.0363716409,0.3695669472,-0.0914533064,0.0341322906,-0.1172682717,0.1985382587,-0.1308670789,0.2064289302,0.1639556885,-0.3200687468,0.1290684491,-0.0055714617,0.0249058101,-0.300971657,0.2896812856,-0.2329612076,-0.312849462,-0.3901703954,0.349393487,0.198458001,-0.233376056,-0.0243524872,0.1399419755,-0.0194629543,0.1363996863,0.2584696412,0.2040382028,-0.0139367655,-0.2660694122,-0.2436650097,-0.0072382283,-0.3851862252,0.0588313974,0.0905199051,-0.124469772,-0.0177358631,0.1691356897,-0.3195768893,0.174481675,-0.088150911,0.2750786841,0.3242750466,-0.2458982021,-0.1534033418,-0.0159979332,-0.0662918314,-0.2005014271,-0.0291288309,0.3394254148,0.0556685925,-0.2397167087,0.0214047451,-0.112587966,-0.0973720178,0.0741112158,-0.0916808695,-0.0971060768,0.1873191446,-0.0391849279,-0.2273071706,-0.0217701644,0.0115468213,0.0929929093,0.2789686024,-0.0334584527,-0.0021385797,0.3999299705,0.5643230081,0.1322357357,-0.0383848995,-0.2299525738,-0.3466582894,-0.0785549134,0.1130993441,-0.0475082994,-0.0032641371,-0.4887993932,-0.2110845894,0.1514522582,-0.1491174698,-0.1210131645,0.364659965,0.041199185,-0.3596985042,0.0051087853,-0.3411982656,-0.0658990666,-0.0650719032,-0.0630109832,-0.0259881131,0.0611186177,-0.0434747711,0.0864067823,-0.0779403299,0.0719288215,0.149945274,0.0306347404,-0.0674778819,0.3047033548,0.2547572553,-0.1484712958,0.1018039063,0.3903384805,0.2561263442,-0.4918302,-0.1127123237,-0.035636209,0.0152697396,0.0281379558,-0.3446542323,-0.1239383817,0.0540817343,0.1074937135,-0.047347147,0.0398416258,0.0499319099,-0.020581089,-0.2060991228,-0.2440981716,-0.1389468461,-0.0907227769,-0.1245270893,0.1630226225,-0.4003733993,0.5827518702,0.7063413858,0.2011756748,0.1275979131,0.2103029042,-0.1632165313,-0.0495944098,0.3204748333,-0.1258469969,0.0997074693,0.3344145715,0.0849291384,-0.0672286376,0.0908531919,0.0481097586,0.2853122056,0.1720513105,0.2410943806,0.014808841,0.1211275086,-0.0328971483,-0.1398342103,-0.2799704969,-0.0799756721,-0.0361163914,-0.0010580268,-0.0964278057,-0.3101136684,-0.1959535927,0.0446282066,-0.4007124305,-0.2970173657,-0.2305703908,0.2444017977,0.1736670583,-0.3192749918,0.4304821789,-0.2385501415,0.2827530205,-0.0603540018,0.0797536224,-0.0294416137,-0.1595160961,0.1817659289,0.2231632918,0.045290608,0.0669123605,0.4846773446,0.237337321,0.2196707278,-0.6060312986,-0.3348538578,0.2684273124,0.020025922,-0.1582012624,0.1165852249,0.0861003846,-0.1867509037,-0.1469798088,-0.0452635065,0.0462637283,-0.1263544112,-0.2138727754,0.1248260811,-0.064655453,-0.1035172418,-0.4099293947,-0.1686361432,-0.4259869456,0.2497263551,0.0464531071,0.2003126442,0.2268297523,-0.2203791142,0.1272613853,-0.1939286143,0.0517396964,-0.287722379,-0.0910830349,0.0405238643,-0.2301250547,-0.0648522153,0.111265555,0.1677405387,0.0724832937,-0.1408929825,-0.0349353068,-0.2663846612,-0.0706500486,0.2242217213,0.0024530778,0.0054389699,0.3446237147,0.3036802411,-0.2777374387,-0.0607246757,-0.2841645181,0.2422917932,-0.0299229082,0.2096351832,-0.1207577735,-0.0400821008,-0.109895274,0.320042789,0.0070331027,0.0703070089,0.01150278,-0.0066237873,0.1706097424,-0.1140713617,-0.1841535568,0.0541748516,-0.0748829246,-0.0850195438,0.2711383402,0.0410193428,-0.1928955913,0.0666516572,0.0689898357,-0.0022312431,-0.1028285325,-0.0122937569,0.1049735546,0.1079922393,0.0895640925,-0.0576395467,-0.0629502684,0.0079138875,0.2293405682,0.0701426566,0.1508974433,-0.1791799068,-0.524992764,-0.0684750751,-0.2856722772,-0.0325444899,0.0520859808,-0.1590142995,-0.0684174225,-0.2288868427,-0.1264598966,0.1298195571,0.3494583368,-0.2347156852,-0.2571791112,-0.012817516,0.232905522,0.2429898828,0.0761484057,-0.0551106967,0.0749796107,0.1325607449,0.3135743439,-0.1608978659,-0.1160039529,0.2259185016,-0.1250160933,-0.1344153136,-0.0702254847,-0.1055418253,-0.2745369077,-0.1255641431,0.0024299074,0.2265334725,-0.1224007756,0.0611995235,-0.0511159189,-0.0025181214,-0.1682798117,0.0442197323,0.1049373075,0.3581658602,0.105211176,0.0160705764,-0.0323009975,0.3750706017,0.1783087403,0.4684070051,-0.0120358905,-0.3239444494,-0.0612508394,-0.4099839926,0.3722437918,0.0312143452,0.064187035,0.200343892,-0.366871506,0.061670322,-0.1949052215,-0.000481349,-0.0203757565,-0.0959347561,-0.2208674401,-0.1555047035,0.6444688439,-0.0643648356,-0.1539199203,0.4251870513,0.172319904,-0.2317101955,0.1257274151,0.3783744276,1.0254029036,0.2526032925,-0.0774812996,0.0697240457,-0.0668391585,0.4803986251,-0.2407032251,0.0332683474,-0.4284316003,-0.4246185422,-0.0527292378,0.0253535695,-0.0378947333,0.0172545705,-0.3774061501,-0.0916657746,-0.1707300693,0.0902982503,-0.0353731513,0.0386748724,0.2670757771,-0.2774054408,0.0258587673,0.3010373712,-0.1238402352,0.0102730058,-0.2058480978,-0.1243798286,0.0720546842,0.024799658,0.0217345599,0.0479623228,0.0249737594,0.1979530901,-0.0156969875,0.1758662909,-0.0117191747,-0.3055869937,0.4540694356,0.0255076028,-0.1475705802,0.06699045,0.3249039352,0.1222148314,-0.1661458611,-0.1267701685,0.2951304018,-0.1233937219,-0.0959848538,0.0339021496,0.0010259306,-0.2686879635,-0.299058944,-0.1045692861,0.2195101529,-0.1449511647,-0.0279293787,0.2664353848,-0.320849061,-0.2290141582,0.274923265,0.148934409,-0.0985475108,0.1550949961,0.094369024,0.0402730405,-0.1097739339,-0.1133397445,0.1755098552,-0.0302165709,-0.0161929298,-0.1587824076,-0.2266312242,-0.4319523871,0.2302051038,0.0989285335,-0.2621452212,0.1082201526,-0.0827669129,-0.1436189711,-0.0604949445,0.2695351243,0.0831219926,0.1283098459,-0.1689441949,0.0792668536,-0.3354309797,0.2626658976,-0.2204947472,0.3034475744,-0.4239976704,0.0567635708,0.0460266285,-0.0451340824,-0.5389589667,-0.2348728329,-0.2432571948,0.0560347177,0.309577316,0.0605258271,-0.1109156758,-0.1264839619,0.242204532,0.1585630178,-0.0044378741,-0.4058422148,-0.2080591619,0.0328799412,0.1130121648,0.0935746208,-0.156002596,0.0541641898,-0.1026494429,0.0415994115,0.006674991,0.0279233027,0.0995508581,0.2364812046,-0.1470196098,-0.0956625342,-0.0791528374,0.0743737146,0.2953699231,0.0330844335,0.007991191,0.0560919009,-0.2636613548,-0.0585752241,-0.0890542418,0.2165733874,-0.1826757789,0.1899277568,0.0313737392,-0.0896043554,0.0651656538,0.2927963436,0.1904449016,-0.0020338784,-0.1686446071,-0.1886713654,0.2070233077,0.5019389987,-0.3451790214,-0.3007515073,0.0093368357,-0.0778953657,0.2203213573,0.271197468,0.2348156124,0.3036780357,0.2559859455,0.1725147963,0.0807644725,-0.368280828,0.1030128747,0.2738997042,-0.0803083852,0.0651743189,0.5343902111,0.3798068166,0.4374248683,0.3024464548,0.2823150456,0.4787766337,0.4010951817,-0.0477376059,0.2929828763,0.1605397761,-0.0267705023,0.2009379715,-0.0826346949,0.0450647473,0.2521844506,-0.1208511516,0.0117055159,-0.4629037082,0.0229909848,0.2912282646,-0.1527652293,0.0585314557,-0.0326383375,-0.1913733631,-0.0913723707,-0.2284371257,0.0387979783,-0.0148613006,0.259565413,-0.2320512831,0.1268305033,-0.1032001972,-0.0220839605,0.0477427617,0.3864571154,-0.0215503629,-0.0089242551,-0.1032252982,-0.1161539182,0.0129617099,0.2914168239,-0.0827725977,0.2136318833,0.1919876486,0.0364913344,-0.2225012928,-0.2518489361,-0.0466788374,0.0038426225,0.0928239673,-0.2481469512,0.3177725971,0.2747697532,-0.2251709849,0.4771801829,0.002687132,0.0240936149,-0.0283880848,-0.0489997678,0.4762868881,-0.462477088,0.0432100967,-0.0272330996,-0.1681993306,-0.2463163435,0.4897706211,0.0655042231,0.1544370204,-0.1387079209,0.1925322264,0.0435743332,0.2640231848,0.1028416157,-0.0690691173,0.0140907913,-0.1058536544,-0.437739104,0.0786690935,0.0307836216,-0.3364422917,0.3698449731,0.082052961,0.0985895023,0.3276998401,0.1078378931,-0.2431470603,0.1392769068,-0.1717433035,-0.3487941623,0.1292327493,0.201515764,0.1912297904,0.0499856472,-0.1741539091,0.1585155278,0.0084595466,0.2409703732,-0.2721626759,-0.0277731456,0.0101953419,0.3349409401,0.4295875132,-0.0655629411,0.0365433767,-0.1404824704,-0.0762398988,0.1161229983,-0.2840474844,-0.1850147098,-0.125250116,0.2663552463,0.3263175189,-0.0400602221,0.1249161139,-0.3470651805,0.0613279305,0.0742959306,0.1809833944,-0.3886091113,0.2718046606,0.021545995,-0.0189409498,0.0172088072,0.2121260017,0.0199410971,0.0029867936,-0.2296372056,-0.3007526398,0.5508956909,0.2880313098,-0.342564851,0.0752125606,0.3618096113,0.2975891232,-0.0679052025,-0.5035438538,-0.1691170037,0.41502285,-0.1046972722,-0.216114074,0.0883089378,0.0241023693,-0.0422621332,-0.1166921481,0.2706602514,0.010125705,0.0211258437,-0.0424769893,-0.1561575681]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/4","title":"[Feature] Keep the list of labels of a dataset as metadata","comments":"Hi,\r\nI hope we could get a better documentation.\r\nIt took me more than 1 hour to found this way to get the label information.","body":"It would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.","comment_length":25,"text":"[Feature] Keep the list of labels of a dataset as metadata\nIt would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.\nHi,\r\nI hope we could get a better documentation.\r\nIt took me more than 1 hour to found this way to get the label information.","embeddings":[-0.0012369895,-0.1648714691,-0.1444353759,0.0895124599,0.1662273556,0.1849589199,0.3181912005,0.0862610564,-0.0790702775,0.3192556798,-0.1647137552,0.5471563935,-0.200616017,0.2491293252,-0.1022710279,0.005948856,-0.157364741,0.0784152895,0.1580577493,-0.1261553466,-0.1806293577,-0.1693413109,0.0251516663,0.0729987621,-0.357034713,-0.0805934444,0.2116435468,-0.2055696845,-0.175666973,-0.5510213375,-0.1353230625,0.0937980413,-0.0786830112,-0.2112112939,-0.0000956425,-0.0900051296,0.2437764853,-0.080670476,-0.1377635747,0.0274874195,-0.450014025,-0.0306218527,0.1355545819,-0.2670372725,-0.0036933725,-0.0796642378,-0.0997258052,-0.2003891319,-0.2043870687,0.045795612,0.3885522187,-0.0868356749,0.04931321,-0.0398686752,0.5041478276,0.253072679,-0.4126293361,-0.0206843466,0.2439529151,0.1936398596,0.1692546904,0.6854000688,0.0045043714,-0.2017082423,0.3443518579,-0.0197564326,0.0436132103,-0.4302619994,0.1069989279,0.0780438334,1.148717761,-0.2577243149,-0.1661703885,0.03488601,0.013556974,-0.2583199143,-0.0203066878,0.1106656119,0.0717397258,0.1847582757,0.0073038996,-0.3757540286,-0.2281251401,-0.1638335288,0.0684202462,0.5441606641,-0.1537513584,-0.0204880033,0.1099063233,-0.1912332475,0.0578916371,0.071008563,-0.120502539,-0.0050507295,-0.0082447464,-0.1972566098,-0.0211164281,-0.1875525862,0.2372861058,0.4003069401,-0.1271066666,-0.095906429,-0.2914321125,0.152131781,0.1925049275,0.0913550705,0.6101221442,0.0721338987,0.1942231208,-0.3319151103,-0.0696623251,-0.1991147995,-0.0890012234,0.3325089812,0.3371145129,-0.0426591523,0.3909405768,-0.4748960435,-0.2659723461,-0.042298682,-0.0798114538,-0.1603301316,0.19779706,0.3338942826,0.0164155792,0.0565250404,-0.3122140169,-0.1739887893,0.3510822654,-0.2051125467,-0.1845497191,-0.0596301965,-0.1094477922,0.0480876528,-0.073043339,-0.0498026721,0.4779724777,-0.1632076651,0.0568900257,-0.0626576021,0.1679087281,-0.179626286,0.2364334315,0.2492837161,-0.1773935407,0.1520492584,-0.1059272513,0.0674741194,-0.3431613147,0.1716541052,-0.3339016736,-0.3239254355,-0.328152746,0.3036516309,0.2867658138,-0.1956075579,-0.081105046,0.1323769242,-0.078157641,0.1557225883,0.1960145384,0.1929073781,-0.0881436765,-0.357396245,-0.2089812756,-0.0679304376,-0.3548083007,0.0634033903,0.1835010946,-0.0971572623,0.0263426118,0.1660446525,-0.3402350545,0.1804993749,-0.0747926459,0.152998805,0.2817381322,-0.304576695,-0.1977692395,-0.0991869122,0.0565673783,-0.3040274084,0.0313151367,0.357199192,0.1371988356,-0.2525860965,-0.037988916,-0.119259201,-0.1827456206,0.1229213178,0.0769780204,-0.0729242191,0.1534150094,-0.0192936305,-0.3171247244,-0.0344871543,0.0391572975,0.12291646,0.202541247,-0.0204833131,0.0531006716,0.595230639,0.5330174565,0.1924235523,-0.0073460038,-0.1274936199,-0.2925902605,-0.0048785121,0.1233347654,0.0435006395,-0.0476257652,-0.5623975992,-0.1672822535,0.2020263374,-0.0502532795,-0.1356932819,0.3412768245,0.0692978352,-0.1592821032,0.0307356305,-0.234834671,-0.0835729614,-0.1016063467,-0.1031384096,-0.0791600943,-0.0126527939,0.0224931482,0.0734020025,0.0195239559,0.0770109221,0.1528437883,0.0858817697,-0.0826209188,0.2503280938,0.3642770052,-0.191136539,0.0896234885,0.4628484845,0.202989012,-0.5049231052,-0.1265562475,-0.10458868,-0.0068428423,0.074111253,-0.3349622786,-0.1465952992,0.1267175674,0.1205837429,-0.0052573672,0.1091275066,0.1501394361,0.0851366669,-0.1811238974,-0.2047689259,-0.2610884905,-0.122795172,-0.054593306,0.1272089481,-0.4433560073,0.6448712945,0.81930691,0.1280813813,0.1116261333,0.0814267173,-0.156567961,-0.0312818065,0.3874503076,-0.1714462042,-0.0298576802,0.3746444583,0.0640372112,-0.120387584,-0.0335312001,0.0954781398,0.2456456423,0.1229071245,0.2739792168,-0.1923180372,0.0251686499,-0.1045745313,-0.053110417,-0.3405828774,-0.037565954,-0.024300104,0.0644434541,-0.1721327007,-0.3638795912,-0.2382689118,-0.0409300476,-0.3866483569,-0.3366088271,-0.2427175641,0.1281423122,0.1403514147,-0.2408833057,0.4542850554,-0.2238083333,0.2266671658,0.0122665986,0.114090085,-0.0426628329,-0.2075577229,0.0608816408,0.2083187401,-0.0507894829,0.0117369825,0.4336638153,0.1324025244,0.2497446537,-0.5540035963,-0.3498836756,0.2959485352,-0.0070695118,-0.1748352796,0.0527373217,0.1413435936,-0.1761558354,-0.0824228302,-0.1549874097,-0.0203714985,-0.1302684546,-0.233631894,0.107998386,-0.1518330574,-0.123505421,-0.4730251729,-0.3118480444,-0.334597379,0.2471403182,0.0232421719,0.2466925234,0.252849251,-0.3076382279,0.2085217386,-0.1585282683,0.1713126302,-0.2391826808,-0.1494201422,0.0749913082,-0.2606092393,-0.1155941561,0.2041754723,0.1375212818,0.084371984,-0.2092921883,-0.0700515434,-0.2334060371,-0.0555215925,0.2725538909,0.0570352077,0.1138924956,0.3041983247,0.2385294139,-0.1843321323,-0.0857886299,-0.2510438859,0.356795907,-0.0861879438,0.2458522916,-0.0341413543,-0.1150797084,-0.1341923177,0.3058535755,-0.0768330172,0.0881058425,-0.0944919884,-0.0230640844,0.2096022367,-0.061435461,-0.0265810397,0.0251490269,-0.0652383789,-0.039539177,0.2249218822,0.060064856,-0.1194644943,0.1041637734,0.1115606129,0.0047303387,-0.1024364159,-0.0938312337,0.1650921255,0.1451711506,0.0665776432,-0.0159239154,-0.0180987343,-0.0130033055,0.2016478777,0.09713424,0.0922558084,-0.1459082663,-0.594145298,-0.0588250495,-0.2968173623,-0.1020794958,0.1661741585,-0.1209907606,-0.1129442006,-0.0747667328,-0.0998607203,0.1974823475,0.2224697173,-0.2555692792,-0.1733750701,0.0044721235,0.1745814234,0.27035743,0.0569524989,-0.1367483288,0.0740187913,0.1512488425,0.2651050687,-0.221526742,-0.2185596675,0.2028308213,-0.1147377044,-0.1887829602,0.0338839777,-0.0532906465,-0.3151333928,-0.0555494577,-0.0646812692,0.2114087492,-0.0791066289,-0.015470502,0.0376386456,-0.0672188774,-0.12499208,0.0175752211,0.1285692155,0.3588565886,0.1678734273,0.0796643198,0.0354860462,0.4213563502,0.3082743883,0.4567632973,-0.0362382568,-0.3652471304,-0.0549165644,-0.3048213422,0.3599268794,-0.050885912,0.0300819073,0.2228411734,-0.403685987,-0.0858376399,-0.1862980425,0.01415426,-0.0314853415,-0.0606366359,-0.1823998988,-0.2255865484,0.6065032482,-0.0310318358,-0.1482856423,0.3782891035,0.1398421228,-0.2171355933,0.0844929516,0.4369863868,0.9195872545,0.2937507033,-0.1067114323,0.0844583139,-0.1077143401,0.4760145843,-0.2062119246,-0.0425350554,-0.4645941257,-0.4050450325,-0.0478519686,0.0685212165,-0.1141045392,0.0732415393,-0.2745035291,0.0024094363,-0.0741713643,0.0284497123,-0.0318055861,0.1433521062,0.2166115642,-0.3721882701,0.072267428,0.2561986148,-0.1443861276,-0.0086841211,-0.0669339523,-0.1190872341,0.1041477248,0.0687935501,-0.0215216614,0.0928882062,-0.0396408923,0.2475081086,0.004642501,0.2571249902,-0.0125184422,-0.2462338954,0.4121073484,0.1033278853,-0.1521440297,0.0583823659,0.2776153088,0.0871054754,-0.1127850562,-0.0382049195,0.3457883298,-0.0559502803,-0.1426967233,0.0610188954,-0.0773511454,-0.2373097837,-0.2455615699,-0.2331226766,0.1190004647,-0.053512346,0.0234284122,0.2704649866,-0.3187799454,-0.1941487491,0.2453618497,0.1659117341,-0.1222819388,0.1812703311,0.089996092,0.0735302046,-0.1401903778,-0.117897369,0.2483602911,0.0603412278,-0.0290227663,-0.1574220657,-0.1389187425,-0.3266373575,0.3035323918,0.0681060478,-0.2784837782,0.0669027939,-0.019135667,-0.0039054197,-0.0589890033,0.2535939217,0.0287114885,0.0881969929,-0.2199205458,0.0105091687,-0.4449212551,0.3048158288,-0.2428059727,0.2791714072,-0.2988988459,-0.0400789566,-0.028939575,-0.0025078582,-0.5056153536,-0.3050804734,-0.2678802907,0.0716246292,0.3104784191,0.063939184,0.0683621168,-0.0452102087,0.187783733,0.1508858502,0.0330270492,-0.3446173072,-0.2024808526,0.0183732025,0.0463424958,0.1228948832,-0.3053214252,-0.1195162684,-0.1498734057,0.1034253985,-0.0127863716,0.0593541227,0.1396947205,0.2568347156,-0.1656391025,0.0006458853,-0.1029189676,0.1477177143,0.2645990252,0.0748073757,0.0030838002,0.0812361762,-0.3653029501,-0.0929747969,-0.2094047219,0.1605017334,-0.1986073554,0.0196455065,0.1048589945,-0.0261649005,0.1322061867,0.311432898,0.1559106261,-0.1090626568,-0.1002719477,-0.1722406,0.3041315973,0.4529170394,-0.2942444086,-0.302731514,-0.0992136821,-0.0347641222,0.2156742066,0.3279439211,0.2934311628,0.4048697352,0.3351889253,0.0780645981,0.0476002507,-0.2987958789,-0.0002297973,0.265429765,0.0078931432,0.0940517932,0.5532249212,0.4313891232,0.4667008519,0.1995991915,0.2376894802,0.4809706807,0.4447413981,-0.0968546644,0.3799720407,0.1630188674,-0.0923212394,0.1164420173,-0.0926215202,0.0603369288,0.087514475,-0.1604095399,-0.0309883058,-0.4357770383,-0.0316673815,0.2741468847,-0.1772161722,0.044856783,-0.0410772189,-0.0920217708,-0.0446954779,-0.1975095868,0.1018282697,-0.07103239,0.1375837475,-0.362408489,0.1395973116,-0.0862417594,-0.0924478918,0.1920213848,0.4139567018,0.0493411198,0.020673234,-0.2146125138,-0.0767424107,-0.0649268627,0.293551892,-0.0297299977,0.1483553797,0.2943935394,-0.037196476,-0.115806289,-0.24979496,0.0679294989,0.0402478725,0.0486520715,-0.2836843133,0.3421565294,0.2083597332,-0.164416045,0.3860663474,-0.0611929819,0.085527651,0.0496991649,-0.1185661778,0.4763045311,-0.4626595676,-0.0060808435,0.0001812959,-0.1738995314,-0.216846168,0.4887299538,0.127983734,0.1148267388,-0.2230089158,0.186453864,-0.0178236086,0.3101934493,0.1780806333,-0.0462342054,0.0877066851,-0.0644420609,-0.3680131137,0.0221317671,0.0745854303,-0.4415938556,0.3512912393,0.0816530883,0.1450376213,0.3208815157,0.0781862661,-0.2648010254,0.1972446442,-0.1202231646,-0.3618985415,0.0410246328,0.1647651643,0.198595643,0.0445908792,-0.1590082496,0.1538600326,0.0185912456,0.1810941398,-0.3610509932,0.0255804434,0.0831687301,0.3645092547,0.3108390868,-0.1341510266,0.0127661815,-0.1271163821,-0.0687103942,0.0942865908,-0.308377713,-0.0817371756,-0.1672578156,0.2020456791,0.2749411166,0.0515005328,0.1004326269,-0.3757931292,0.0677645653,0.0364500619,0.1660328954,-0.281642884,0.2044068426,-0.0710698441,-0.0331357904,0.1783468723,0.1479926705,0.1054490954,-0.0864951238,-0.1680939645,-0.2889731228,0.5036252141,0.2576980591,-0.2148969918,0.0834675878,0.2721806765,0.3186980188,-0.0604453832,-0.5303121805,-0.069999814,0.4421144724,-0.1397153884,-0.232890591,-0.0145276748,0.0453348942,0.0968215019,-0.1318294704,0.2383090854,-0.0106113786,0.012284941,-0.0843884572,-0.1290216446]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/4","title":"[Feature] Keep the list of labels of a dataset as metadata","comments":"Yes we are working on the doc right now, should be in the next release quite soon.","body":"It would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.","comment_length":17,"text":"[Feature] Keep the list of labels of a dataset as metadata\nIt would be useful to keep the list of the labels of a dataset as metadata. Either directly in the `DatasetInfo` or in the Arrow metadata.\nYes we are working on the doc right now, should be in the next release quite soon.","embeddings":[0.0434394106,-0.0362067781,-0.1563860625,-0.0524546467,0.122303389,0.2208477557,0.2759690583,0.1788536161,0.0087437974,0.2267382592,-0.1574053764,0.5626071692,-0.2707872689,0.2366517782,-0.0232619904,0.0165955648,-0.0474637486,0.0209634155,0.1306467503,-0.0336884186,-0.2415300161,-0.1698799729,0.1512940228,-0.053203024,-0.2328320593,-0.0394080877,0.1785150617,-0.2916420698,-0.1576616466,-0.5575762391,-0.1708637029,0.2196195722,-0.0708721131,-0.1691180617,-0.0000974572,-0.127391696,0.1770223379,-0.023043802,-0.1470951587,-0.0159274116,-0.4960756302,-0.0295409262,0.1479242295,-0.1194488406,0.0284162294,-0.0955274627,-0.0997408554,-0.3222854733,-0.183641836,0.0122567974,0.3661126196,-0.0640486255,-0.0614348017,0.0069510075,0.4805946052,0.2897692919,-0.4747666121,0.0097427862,0.2417715788,0.1432863623,0.1520195752,0.6420621872,-0.0295930393,-0.2510326505,0.3693681657,0.0115175229,0.0201355666,-0.3688250482,0.0242838394,0.0691424683,1.0715801716,-0.3872752488,-0.1921420395,-0.0374775976,0.0414030552,-0.3004676402,0.0241514053,0.0755021051,0.09690658,0.214543134,0.0616907664,-0.5372310281,-0.1946929991,-0.0914590582,0.0487755053,0.5409656763,-0.0751591995,-0.1670041233,0.1555884331,-0.1719274372,0.1570845395,-0.0032776885,-0.1599087715,-0.0129470052,-0.0063076848,-0.2228289694,-0.0958382264,-0.1572147161,0.2525980175,0.4223838151,-0.0732687041,-0.0907001048,-0.3351193964,0.1235924363,0.295575887,0.0198883619,0.6246948838,0.0946136788,0.2880927324,-0.4285324514,0.0352327786,-0.1624549031,-0.0993715748,0.3854803741,0.2950223088,-0.0776431784,0.3249313235,-0.4315893948,-0.1809603125,-0.04522264,0.0039805737,-0.1701399386,0.1890342236,0.2550528646,0.1284542531,0.1242187545,-0.2564404905,-0.0217115339,0.4035780132,-0.1800673902,-0.114364773,-0.1187709272,-0.0754052177,0.123991631,0.0000380957,-0.0456723347,0.3899737895,-0.1539159268,0.0750819892,-0.0887620822,0.1447477788,-0.0744216144,0.2771876752,0.1985365301,-0.3157619536,0.1249477193,-0.1547252536,0.0169929005,-0.3274433613,0.2830433547,-0.2704171836,-0.2312248945,-0.4020668566,0.2658469081,0.2844426036,-0.3097572625,-0.1067962945,0.1678143293,0.0092426715,0.091192238,0.2200801373,0.1911843866,-0.0632189438,-0.2448177338,-0.3248766959,-0.020181708,-0.3555791974,0.0212505255,0.1612933874,-0.0053835595,-0.1127403751,0.1281253695,-0.2790532112,0.1148020327,-0.0166740548,0.198962599,0.215004921,-0.3138137162,-0.1832923293,-0.140606612,0.089103654,-0.2974984944,-0.0027125869,0.3696067631,0.1509943008,-0.3287335336,-0.0952910036,-0.1117072329,-0.1460364312,0.1848975122,0.0566564873,-0.0997147709,0.157763496,-0.0430928208,-0.2826045752,-0.0194783807,0.0270490441,0.1429832429,0.1487599015,-0.0547486246,0.0540669486,0.5306127071,0.5022466183,0.1811080128,-0.0339322388,-0.193663612,-0.3577806354,-0.1424719691,0.1433421373,-0.1000240296,-0.023233626,-0.5768047571,-0.1213204712,0.142217055,-0.0736330152,-0.0558607951,0.2889319062,0.0115037095,-0.3270288408,0.1184311211,-0.335976392,-0.1061220169,-0.0946167037,-0.0481744744,-0.0669817477,0.0370242558,0.0716620609,0.0317729451,0.0576182678,0.1245610192,0.0394380726,0.042498406,-0.0618643463,0.2732851505,0.3557746708,-0.1737657636,0.140822202,0.5224788785,0.3935517073,-0.4445533454,-0.0970880389,-0.0675663874,-0.092899926,0.005917463,-0.4204341471,-0.1387980729,0.0643701181,0.1765261739,-0.0108582927,0.0879018828,0.1306762546,0.0643801913,-0.2210184932,-0.2471513599,-0.3081650138,-0.1115167513,0.0108659621,0.0825834423,-0.4147004187,0.6170598865,0.7159015536,0.1884166747,0.1927088052,0.1729831696,-0.1423829943,0.0155096315,0.3298496306,-0.132500574,0.0031863532,0.3548626304,0.1074496582,-0.1206543893,0.0597314164,-0.0086461874,0.2671056092,0.141038999,0.2787774205,-0.0590398572,0.0770412982,-0.0987058207,-0.1839791536,-0.223000899,-0.0659468621,-0.0925993472,0.0474029146,-0.1751674265,-0.4039583504,-0.2162207812,-0.0144815538,-0.4419514537,-0.4217895269,-0.2748585343,0.2110532075,0.2215564251,-0.3491579592,0.4842574596,-0.2305960208,0.3170522153,-0.0948730186,0.0737082437,-0.1031873599,-0.3096196651,0.1940344125,0.168467626,0.0646788701,-0.0462103449,0.4901993275,0.2274284065,0.2941660285,-0.6255378127,-0.3937917352,0.2680423558,0.1080576032,-0.2010456026,0.0211653765,0.0601061098,-0.1602723598,-0.1049769372,-0.1912260801,-0.0714137927,-0.1114473119,-0.2610792518,0.1679469198,-0.0486777611,-0.0943757668,-0.412730366,-0.3196935356,-0.3933529258,0.3229835927,0.047798086,0.1956526041,0.2367887348,-0.2738618255,0.1109074652,-0.2667829096,0.0560957827,-0.2118016779,-0.1464274973,0.040188659,-0.3039529026,-0.0693357587,0.1233008057,0.1781906635,0.1063380837,-0.1853940785,-0.0211831313,-0.1820605099,0.0329698324,0.2561916411,0.107886672,0.0982023552,0.2715085149,0.30254969,-0.1601308286,-0.0782744735,-0.2235371023,0.3034810722,-0.0258755721,0.2311404049,-0.0255304892,-0.1827392578,-0.1518349499,0.248751238,-0.0658458993,0.121062085,-0.0589036047,0.0601909719,0.2668934464,-0.1118384823,-0.0111772427,0.0304732341,-0.0972397625,-0.0587645955,0.2651949823,0.13476412,-0.1844831854,0.2004963905,0.1009731069,0.0333162323,-0.0763027519,-0.112285845,0.0593859665,0.0727922246,0.1100728884,-0.0937862098,-0.028882321,-0.0867157802,0.2341371179,0.0882664993,0.1332021654,-0.1511579752,-0.4961045384,-0.0867211521,-0.3261429667,-0.1355999112,0.1741979867,-0.0836375058,-0.0742302388,-0.0959644914,-0.1763468087,0.2213132083,0.2468971759,-0.2746393681,-0.3400365412,-0.0794511884,0.2384649515,0.3246187568,0.1152687967,-0.0302219596,0.0080616139,0.1005545333,0.2541942596,-0.1792628318,-0.1938325614,0.2142207921,-0.117186591,-0.1312997639,-0.0480093285,-0.032134451,-0.2642597258,-0.0507320687,0.0047004195,0.2307537347,-0.1946417391,0.0197138153,-0.0166253708,0.0100512514,-0.2149720788,0.008316013,0.0952983052,0.3919167519,0.1922385544,0.0882560089,0.0017747023,0.3346697688,0.4027009308,0.4051616788,-0.1491683573,-0.3887189031,-0.0762826428,-0.3767279685,0.4026464522,-0.1080439761,0.1047671959,0.2102488726,-0.395393461,-0.1026771739,-0.2025729269,-0.0433469005,-0.1292918921,-0.1171878651,-0.1737458706,-0.2503151,0.7172234058,0.0203566905,-0.1120926887,0.513060987,0.0829407498,-0.2472289205,0.0864144787,0.4482965767,0.8322601318,0.2280032039,-0.0528047644,0.0237781238,-0.1513478756,0.4931334555,-0.2395317256,0.0114548625,-0.4578917027,-0.3646808565,-0.0469420403,0.1008851901,-0.0444332398,0.0961261392,-0.2933894396,-0.0768104345,-0.0855771005,0.1010476798,-0.0362258479,0.14803271,0.2260167152,-0.4164177775,0.087406978,0.2416806519,-0.1608041823,-0.1103582457,-0.11802385,-0.1503984928,0.1359771341,0.1578750163,0.0194362607,0.0825114176,0.0534790568,0.2472418696,-0.0511249937,0.2083416581,0.0534807704,-0.2338529825,0.4566830397,0.0253251828,-0.1820273846,0.1007103473,0.2212894857,0.1215675846,-0.1458111107,-0.0431773923,0.2938301563,-0.0340799913,-0.0782395601,-0.0266606137,-0.0056681526,-0.3166435361,-0.2957299054,-0.2071878016,0.1741409451,-0.0244430713,0.175775215,0.3065809309,-0.3132876158,-0.2669222355,0.2318620831,0.1378940344,-0.0916738659,0.162243709,0.0242805127,0.0810886398,-0.182470426,-0.1041540653,0.2632157505,-0.0078621181,-0.0150172226,-0.10657987,-0.2068449259,-0.3972658813,0.3107145131,0.0443361476,-0.1850654483,0.1632600129,-0.056431137,-0.1413294673,-0.0493741632,0.2088932842,0.0147228707,0.0290503968,-0.2001441568,0.100595206,-0.3834915757,0.2265595794,-0.2507708669,0.3285464346,-0.3352530897,-0.0198477972,-0.0424778424,-0.0354992859,-0.4664963484,-0.3121863306,-0.1677401066,0.0158843603,0.3144163489,0.1175070629,0.0002261839,-0.1307317466,0.1292648315,0.1647517234,0.0711842775,-0.3367094696,-0.2378472984,0.020653123,0.0642883256,0.1974440068,-0.2520066202,-0.0365505256,-0.1360204518,-0.0062989434,-0.0214355104,0.0665380433,0.1169795468,0.2628073692,-0.0627685711,-0.0530670881,0.0545101799,0.1585968882,0.3177578747,0.0519035347,0.0091197155,0.1115512624,-0.332687974,-0.004548938,-0.0989475697,0.1941012889,-0.1683584303,0.0285615027,0.091192089,-0.1114879325,0.06439922,0.3876011074,0.1390179694,-0.01243761,-0.0604139641,-0.1483981758,0.2694877386,0.4179908931,-0.224919796,-0.3678020537,0.0189449955,-0.0062456368,0.1890119314,0.2914510071,0.3563767076,0.3418659568,0.2424625158,0.1027166173,0.0587759018,-0.3888030052,0.1311996877,0.2608471215,0.0531240143,0.0675649121,0.477567941,0.3761873245,0.4159620404,0.249059394,0.2351549715,0.4531279206,0.4466481805,-0.1421216279,0.3544519246,0.2385438979,-0.0762028396,0.1016236469,-0.0155798113,0.0840309858,0.2076202482,-0.0984833539,0.0064950949,-0.5115835667,-0.0214262139,0.2802046537,-0.0889278129,-0.0108711608,-0.0475592315,-0.0755883232,0.069201231,-0.2171291262,0.0890483111,-0.1171615645,0.2474506944,-0.2408570796,0.2792401016,-0.1811423153,-0.0920725986,0.1192620173,0.3826780915,0.0472017415,-0.0246769171,-0.2306434959,-0.166396156,-0.0187213831,0.278301537,-0.1242557913,0.0902309939,0.3227208555,0.0015389592,-0.2542091608,-0.2029975206,0.0883399546,0.0786236897,0.0293894876,-0.2283680737,0.3250943124,0.2124478221,-0.1215018779,0.4036473334,-0.0172289833,0.036403399,0.0660319105,-0.1407794803,0.5390024781,-0.4593968093,0.0848291293,-0.0046444666,-0.1158535108,-0.2132770717,0.5708605647,0.0442866161,0.0722281635,-0.133733198,0.1802088916,0.0906191543,0.3319391906,0.1743141711,-0.1310208142,0.0620356351,-0.0890992954,-0.3275170624,0.0419921763,0.0372598395,-0.3228781521,0.3510718048,0.0857380182,0.1607897729,0.3787332475,0.0880778357,-0.2828834951,0.160938859,-0.2028871179,-0.3434160352,0.0972482711,0.1517733932,0.2282914966,0.0468576625,-0.1009306312,0.1427977532,0.0792722255,0.1364851296,-0.3838135302,0.1300282627,0.013951336,0.3490138948,0.3182300627,-0.1503994018,-0.0340052024,-0.0932925418,-0.0818849877,0.0732897967,-0.428881973,-0.1255489439,-0.1211812496,0.1574782729,0.2775984406,-0.0780612305,0.0220227856,-0.351292789,0.058491651,-0.0090997564,0.2142266184,-0.3489816487,0.2006735802,0.0022644664,0.0270617083,0.1181811541,0.1586289555,0.0363150388,0.0254806858,-0.0948070884,-0.2343761027,0.3852598071,0.2222097665,-0.2268879116,0.1126205176,0.2603065968,0.2898999155,-0.0853694156,-0.5906236768,-0.1029810011,0.4209432304,-0.088877596,-0.1769334823,0.0716728866,0.0451232754,0.096405223,-0.1804311424,0.2323370427,0.0496095568,0.0416403897,-0.0830894783,-0.1776939183]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/3","title":"[Feature] More dataset outputs","comments":"Yes!\r\n- pandas will be a one-liner in `arrow_dataset`: https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.Table.html#pyarrow.Table.to_pandas\r\n- for Spark I have no idea. let's investigate that at some point","body":"Add the following dataset outputs:\r\n\r\n- Spark\r\n- Pandas","comment_length":23,"text":"[Feature] More dataset outputs\nAdd the following dataset outputs:\r\n\r\n- Spark\r\n- Pandas\nYes!\r\n- pandas will be a one-liner in `arrow_dataset`: https:\/\/arrow.apache.org\/docs\/python\/generated\/pyarrow.Table.html#pyarrow.Table.to_pandas\r\n- for Spark I have no idea. let's investigate that at some point","embeddings":[-0.4236600101,-0.061805103,-0.2500792146,0.0073686498,0.302544415,0.1241440624,0.2948102653,0.3709729016,-0.146498695,0.0166035481,-0.2502678633,0.7553219199,-0.0170536079,0.3141035438,0.2778237462,-0.1931229085,0.0942965299,0.1060331464,-0.2115752995,-0.0662083998,-0.2025430948,0.0692163706,0.1552253366,0.1883412302,-0.3077115119,0.0542252064,-0.2953985035,-0.3888697922,-0.0519564971,-0.4054460824,-0.0291128717,-0.0066005536,0.0361478627,0.0782574043,-0.000105874,-0.0859250426,0.2928694189,0.0189206824,-0.0985230654,-0.1385447383,0.0015435003,-0.2957569063,0.4588668346,-0.2712087929,0.0664705336,-0.4761643708,0.1820202321,0.1816089451,0.3071880639,0.1816480309,0.3140155971,0.2772028148,0.0876397192,-0.0478978492,0.4405783117,0.1783980727,-0.1809587032,0.2180856317,0.035754513,-0.1531905979,-0.2881404161,0.2587610483,-0.0024940812,0.0705226511,0.1081708595,0.2521232665,-0.0431856662,-0.2616991699,-0.01423347,0.2416135669,0.5697915554,-0.4277185798,-0.2202006131,-0.101631999,-0.0081955735,-0.1537102312,-0.0653768033,0.1404832602,-0.0338151343,0.159204334,0.2372844219,-0.148344636,-0.1802538931,0.0324250311,0.0428766832,0.567420423,0.0350545309,0.0761028975,0.2279883325,0.2201832831,0.2787986398,-0.0737675801,-0.0235871468,-0.0989851728,-0.1824269742,-0.0830923989,0.0241274517,-0.1216841117,0.0777469501,0.1878507435,0.2261743397,0.2018346637,0.0698906407,0.1147333905,0.1123695225,-0.2256430835,-0.056435246,0.1779745817,0.3263490796,-0.2604716718,0.1060847715,-0.1591226906,-0.0053732484,-0.219436273,0.2726423144,-0.0774719715,0.0982927978,0.1101387367,-0.1355004162,-0.0550318398,-0.0814183578,-0.1464684159,-0.155445084,0.2199733704,-0.0913158059,0.0208556242,-0.1420201659,0.2134454995,0.1208521575,-0.0601051189,-0.1055602953,0.0647919104,-0.0890365019,-0.0799133927,0.0018244983,-0.0327111855,0.2158912271,0.0448193289,0.03563587,0.4225840271,0.1090576053,-0.0523547567,0.1695752144,0.1312266439,0.0576992407,0.0786736831,-0.0038996153,0.0589622185,-0.2611331642,0.0690426007,-0.0885624662,-0.0840344429,-0.4620334208,0.2259626389,-0.0248054769,-0.200250417,-0.3925516605,0.4292556345,-0.1183581427,-0.0206985641,0.2050465941,-0.0127050057,0.1893754601,-0.3556747735,0.0144588277,0.1952510327,-0.5921708345,0.0633017942,0.0456995703,-0.1552539915,-0.0064759669,0.0487558916,-0.2209919989,0.0197519101,-0.1579330564,0.29813537,0.4278999567,-0.0897898898,-0.1441705823,0.0430087447,0.0285683107,-0.1067304537,0.2111125886,0.2186039537,0.1423521638,0.271194905,0.0414086394,0.0939130038,-0.1085531414,-0.1136269495,-0.0003044448,-0.139150694,0.1358144879,-0.0325311013,0.3364757001,-0.1443516761,0.1346239001,-0.4724796116,0.263956368,-0.00486856,0.0789878666,-0.0003532473,0.4330609739,0.2121476084,0.1850803643,-0.3109677136,-0.1032641381,-0.0511830486,0.1801474541,0.0830069631,-0.3254921436,-0.3626220822,-0.0088911122,0.2873338163,-0.0360436514,0.034847606,0.2639262974,-0.1477839798,0.1388918161,-0.021156868,-0.072000958,-0.0205897335,-0.0253083408,0.1436189711,-0.0922502726,0.4329188466,-0.118888095,-0.2623587549,0.0483898148,-0.2049252838,-0.13349244,0.1074999869,0.0801312253,0.3656724393,-0.0655185953,0.1203395724,0.2987856269,0.31216079,0.1776942015,-0.3076853156,0.0522706099,0.0754826665,-0.0292481426,0.1495366991,-0.1281578094,0.346421659,-0.0796327218,0.0508350097,-0.0992205292,0.2070183605,0.2490616143,0.2576343715,-0.1731717288,0.0004901166,-0.2870499194,0.1379067749,0.1510300487,-0.0820539072,-0.5496090651,0.2932040989,0.3700991571,0.0327968001,0.3454669714,-0.0250067133,-0.401443541,-0.1105823889,0.5292955637,0.0525189526,0.0905274898,0.4710197449,-0.0126012042,-0.2390518636,-0.4071599841,-0.137552157,0.1281996667,0.2362999171,0.2711461782,0.0875641704,0.1468279809,-0.1913910359,-0.4449611604,-0.1730959713,-0.1262284517,0.089934364,-0.2430396527,0.1468863636,-0.1808814257,-0.2581594288,0.0371693932,-0.1066882983,0.0025671637,-0.3480190039,0.2545681596,0.1055138856,-0.2233173102,0.2132592946,-0.0406424962,-0.1390573233,0.0891872868,-0.0437626615,-0.4960609674,-0.2942949831,-0.0096845673,0.2562079728,0.020722039,0.0339410342,0.2570437789,-0.036506407,0.2048762292,-0.4996928573,-0.3178368807,0.0938514993,-0.2857542634,0.2574905157,0.1663200557,-0.0129058249,-0.2927206457,-0.2263267338,0.2496741563,-0.2605257928,-0.0466354154,-0.0209565405,-0.1518792659,0.1266611665,0.0478392355,-0.1570951194,-0.0969642028,-0.3295823336,0.2770689726,0.2057902813,0.0709744468,-0.3107313216,0.4204165339,0.1273530871,-0.0021180054,-0.1289887279,-0.1200925931,0.2234212309,0.5940716267,-0.2815119922,-0.2537537813,-0.0943998173,-0.2281398326,0.1199764907,-0.0334272273,-0.1439121813,-0.2743971646,-0.0952565372,0.5558851957,0.0151029443,0.2108995467,0.2947681248,0.2981844842,-0.2353915125,0.014389283,0.0349209607,-0.0512359925,0.1449640095,-0.026564721,-0.1152961478,0.3901530206,-0.1239051446,0.6111701727,-0.4408535957,-0.1811609864,0.3107129931,-0.3898284733,0.3139994442,-0.0590240844,-0.0743202418,-0.4171909988,-0.3363457918,0.1835570633,0.028889332,-0.0978000686,-0.0701976717,0.0535304509,-0.0173596796,-0.3781405687,-0.149671942,0.3706858158,-0.3826171458,-0.0449076891,-0.0318744555,0.0616622791,-0.2474861443,-0.0242293961,0.129541561,-0.1503232569,0.1282164603,-0.1866502613,-0.159518674,-0.1391901523,-0.5074170232,0.2244947255,0.08834631,-0.3453115523,0.150164485,-0.2426865548,0.0441008843,-0.1037171185,0.7440726757,-0.1512461454,-0.4287917316,0.0065509151,-0.1091931239,-0.4215226173,-0.0463161767,-0.465247035,-0.0164195057,-0.1166235656,0.2729661167,-0.361923337,-0.0509220958,0.2919286191,0.0509537607,-0.1701606512,0.0673455298,-0.2369609177,0.0373370126,-0.2488288283,0.0919299126,0.1083174348,-0.084851481,-0.1848137081,-0.3347016275,-0.1315171123,-0.2330950648,-0.0510703549,0.2955830097,0.2444735616,-0.1521165818,-0.0414713211,-0.0711243451,0.0722673312,0.2425003797,0.2282215357,-0.0261796378,-0.2925851047,0.1497313976,-0.1393755227,0.3812773526,-0.0219439473,-0.0449749865,0.2856168747,-0.2105940878,-0.0285500195,-0.0717061311,-0.0063276864,0.2856150568,0.233872816,-0.0286578927,-0.4190359414,0.5137844086,0.0793262422,-0.0663892478,0.4169153273,0.1874301285,-0.1192047819,0.4170823991,0.2910560668,0.8528812528,-0.3674771786,0.1441677213,0.2792282403,-0.2981651723,0.3815590739,-0.0733482987,0.1263207644,-0.3640948832,-0.0649174377,-0.0619075969,0.0030821585,0.0039406451,0.1856360137,-0.0547005273,0.0131366309,0.0989299789,0.2386181206,-0.0860847235,0.192790091,-0.0799527764,-0.1208278686,-0.2532425523,0.1723967642,0.0492238402,-0.2195205092,-0.158089608,-0.2827788889,-0.20585379,-0.1249750406,-0.0509357266,-0.1033991352,-0.1315752864,0.3394502699,-0.0046163145,-0.5491881967,-0.2443902045,0.2781844437,-0.1453282833,-0.0514825024,-0.0131185856,0.006196911,0.1524531394,-0.0414575301,-0.037162587,-0.1098245531,0.0374339409,-0.032830473,-0.3800393045,0.3853589296,0.2410804927,-0.5668436885,0.1609778106,-0.0169084724,0.2184248716,-0.2202437669,0.1394702792,-0.0754644424,-0.4161069691,-0.2774191499,0.1953768879,0.126256004,0.0463337861,0.2286078334,0.2153229266,-0.0742994472,-0.2823758423,0.0974317566,0.033988338,0.4001073837,0.0114465915,0.1470696777,-0.3426958919,-0.241380468,-0.0467287973,0.3023665249,-0.6356937289,0.1141931787,0.0490903407,0.0380219258,-0.09951628,0.0368563458,0.1409469694,0.1654648781,-0.3828516006,0.0914212912,-0.3278032839,0.5099631548,0.1044449434,0.215803653,0.2300736308,-0.2173577398,0.0876064971,0.0678537861,-0.4121639729,-0.2144198269,0.067929633,0.1426175833,-0.0907981396,0.0899497941,0.0608389117,0.0438593328,0.2920072973,0.1050189584,0.0492106117,-0.3632779717,0.0336147808,0.032568749,-0.0775837675,0.0871944577,-0.1540614665,0.0951250046,-0.2058449835,-0.2835697532,0.0083017275,0.0809051916,-0.1132796407,0.095698446,0.5496812463,-0.1312663853,-0.0233380925,0.0118745826,-0.1684227884,0.1172573045,-0.0229233988,0.2631512582,0.075496383,-0.0470000468,0.1029868722,0.1477090865,0.0002244463,-0.1104510501,0.2196855843,0.1343410462,-0.2223708183,-0.1797736585,0.3855793476,0.1455974728,0.1459499002,-0.0591587722,0.2067237943,0.3235148489,-0.4859023392,-0.1007405594,-0.215610221,-0.1383678615,0.331548512,0.2487018704,0.3619480729,0.1890194565,0.1776093245,0.0943158343,-0.0886645466,0.0484385751,0.0201540999,0.6003825665,0.1240898371,0.1301896274,0.6020786762,0.0362828784,0.1585396677,0.1191584393,-0.0292069912,0.2612558901,0.2183488905,-0.1731966883,0.3991560936,0.0464419872,-0.1838891804,-0.3725125492,0.0198999122,0.0921855271,0.417124182,0.3947752714,0.2483377606,-0.0407497361,-0.2025268227,0.1162446886,-0.2078961283,-0.0644785017,-0.0614173934,-0.1641442329,-0.0436220467,0.1553644985,-0.0065030945,-0.2308659106,0.4870080948,-0.119844079,0.144540742,-0.0890762657,0.5264891982,0.0031989007,0.1854021996,-0.0827387124,-0.2578554153,-0.0145939142,-0.1432345212,-0.2359838933,0.2723373175,0.256008029,-0.1224136353,-0.0381277874,0.1001228243,-0.0988265127,-0.1525486559,-0.1079179943,-0.1098882109,0.3089714944,0.0051677665,0.0478055775,0.3477054834,-0.2107414007,0.1764133871,0.0003710983,0.0005046175,-0.0676341429,0.1346570104,0.2658907473,-0.4134854674,-0.0095928926,-0.0151889492,-0.2957991362,-0.0114339795,0.3258183599,-0.1867858469,-0.1942660511,-0.2431714833,0.1492814273,0.0415519029,0.3240580857,0.2847059965,0.3985266089,-0.135204643,0.0422858931,-0.3762682378,0.1351507008,-0.1144200712,-0.0440750793,-0.1052158624,0.2420796305,0.2215883434,0.2693567574,0.2978172302,-0.1322226673,-0.1728221774,0.3995055556,-0.2361762673,0.092445761,-0.1232868806,0.2140400261,0.010721121,-0.0861671939,0.1957567483,-0.1794981211,0.2275644392,-0.3031543791,-0.1252721995,0.2259432077,0.0071529741,0.7032317519,-0.1147768348,0.2997299135,-0.2697787881,0.190140903,-0.1035612151,-0.2823874056,-0.1613539159,0.3027569652,0.0590027273,0.1038334668,0.0345875956,-0.1951782405,-0.0299111344,0.2045649588,0.0181104969,0.1058937684,0.1295786053,-0.000353654,-0.0422811806,-0.2819921076,-0.1773630977,0.0588661656,0.1037272364,-0.2369870543,-0.1459033787,-0.3007537723,0.0751755014,-0.3455227315,-0.2172939181,-0.3983820379,-0.0591066666,0.6210557818,0.147074461,-0.2711721957,-0.1042654067,0.3030769527,0.129969269,-0.0202270467,0.1612374932,0.1881453991,0.2072905749,-0.2551644146,0.069487974,-0.1023021489,-0.0435779691,-0.4128997028,-0.580579102]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/3","title":"[Feature] More dataset outputs","comments":"For Spark it looks to be pretty straightforward as well https:\/\/spark.apache.org\/docs\/latest\/sql-pyspark-pandas-with-arrow.html but looks to be having a dependency to Spark is necessary, then nevermind we can skip it","body":"Add the following dataset outputs:\r\n\r\n- Spark\r\n- Pandas","comment_length":28,"text":"[Feature] More dataset outputs\nAdd the following dataset outputs:\r\n\r\n- Spark\r\n- Pandas\nFor Spark it looks to be pretty straightforward as well https:\/\/spark.apache.org\/docs\/latest\/sql-pyspark-pandas-with-arrow.html but looks to be having a dependency to Spark is necessary, then nevermind we can skip it","embeddings":[-0.5538601875,-0.1469933093,-0.2703717053,0.0605323277,0.2588167489,0.1869446039,0.2700168788,0.4753178656,0.0849914774,0.1245463192,-0.0857295692,0.6752998233,-0.1100908369,0.3995358944,0.02501739,-0.2358332276,-0.10844253,-0.0002741395,-0.2984122932,0.0295078866,-0.2555410862,0.1111409962,0.2057302296,0.106230326,-0.3432416022,-0.0539068729,-0.2266675979,-0.2679175436,0.0396039635,-0.3622856736,0.0098006399,-0.0368889011,0.0572327338,0.2427306771,-0.0000990341,-0.1879158467,0.2158855051,0.0625416562,-0.0935503468,-0.1431126744,-0.0893051028,-0.2987647355,0.499340713,-0.32741943,-0.001760466,-0.2864935398,0.0751593783,-0.04923236,0.2396622151,0.0574651584,0.3537534475,0.2032936215,-0.0505121388,-0.0931669399,0.011724608,-0.0443656407,-0.070485726,0.0733082592,0.0989308506,-0.0770636648,-0.3784050047,0.3305183053,-0.0419841483,-0.0166584868,-0.0480699576,0.0941856578,0.1756498963,-0.2204088122,0.0856401324,0.1887379289,0.4700126052,-0.3086578846,-0.273935169,-0.2869238257,-0.1321463734,-0.1571960449,-0.165258348,0.2294098437,0.061391741,0.2063027918,0.1382098645,-0.0838116929,-0.259355247,-0.0061442731,-0.1574677974,0.4023459852,0.1091234013,-0.0698471963,0.2539541125,0.065071553,0.090452224,0.07416673,-0.0713789165,-0.0784911439,-0.2035108805,-0.2129990608,0.055978369,-0.2913092673,0.0196784977,0.1325111538,0.1347487867,0.2045744061,0.021805767,0.1224140972,0.1788022816,-0.2655582726,0.072490029,0.2539013624,0.2871673405,-0.2409293056,0.1024187282,-0.1289289296,-0.0519943759,-0.2499616891,-0.0389688984,0.0761710703,-0.009905586,0.0484883711,-0.242231369,-0.0391204469,0.0224282481,-0.0806409717,-0.120958887,0.3000917733,-0.0717795864,0.0651730448,-0.0388911031,0.1913029999,0.0923086181,-0.0686771646,-0.1064642146,0.1821031868,-0.018811876,-0.1073064059,-0.0210037772,0.0676506981,0.2616631091,0.0312227253,-0.1493034661,0.3017959297,0.1890586466,-0.023165673,-0.0656429827,0.089925915,0.1306271106,-0.039786581,0.1047995016,0.0877235308,-0.2037502974,-0.0406651348,0.0616203696,-0.2846140563,-0.5234966278,0.3075758815,0.0326080211,-0.2044956237,-0.238145262,0.5409141779,-0.1498934776,0.0261262842,0.0736213773,0.1660899371,0.1113569811,-0.2901499569,0.0518780835,0.3494935036,-0.6666471362,-0.0364144258,-0.0275494326,0.053966336,-0.0140019711,-0.0442776121,-0.1260465533,0.0602844059,-0.2055065185,0.0935079157,0.3954156935,-0.1698444188,0.0932532847,0.2512268722,-0.1106658131,-0.1646519452,0.2606687844,0.2550144196,0.2499963492,0.2460156977,-0.1078328639,0.2227351069,-0.1699499637,-0.064878419,0.048115544,-0.1972186863,0.1155436337,0.05259661,0.387093395,-0.0340709463,0.221007064,-0.2865006328,0.3103354275,-0.0058452999,0.0196932796,0.1366246194,0.3954476714,0.22259444,-0.058859475,-0.4964853227,0.0338161886,0.0449425727,0.1493114084,0.0796359181,-0.2523971498,-0.3368175328,-0.0894295946,0.1325584203,0.0984070748,0.0260067414,0.303365767,-0.2732362449,-0.0066490863,-0.0025284952,-0.042477794,0.1733339429,0.1019584239,0.0758345351,-0.1531866789,0.2987890244,-0.1364869624,-0.1301478446,0.1363058239,-0.1347593963,0.0113679273,0.0759536847,0.1194988862,0.3512020111,-0.0334719419,0.1418432146,0.4699138701,0.3609049916,0.193582803,-0.0714738369,0.0232465882,0.0834303871,-0.1012845635,0.0524637662,-0.1680175513,0.5050488114,0.0285401754,0.0658488646,-0.0646449924,0.2344291806,0.3114738762,0.1539533585,-0.1260727197,-0.1162480488,-0.2289201617,0.0588110834,-0.1082051024,-0.1587497592,-0.5301884413,0.3296276927,0.1728732139,0.0653062761,0.2982239723,-0.0369402841,-0.2525236309,0.0748354495,0.5008333921,0.2263532579,-0.0477875732,0.5283264518,-0.0352791362,-0.1022377089,-0.2568717301,-0.0895506516,0.2588435709,0.1919327676,0.0886877179,0.024516983,0.0556643084,-0.1531887949,-0.3598842919,-0.2306941748,-0.1705384254,-0.0150525235,-0.1025401726,-0.0200323071,-0.2984920144,-0.391721487,0.2167875916,-0.1473570466,0.0638981611,-0.3715459704,0.1317882091,0.2747735083,-0.2373939455,0.3069716692,-0.2229562253,0.1047448516,0.2739242911,-0.0615652278,-0.1887937337,-0.1890191883,-0.0125589827,0.2616713047,-0.0673963949,0.1524406374,0.4395678937,0.0669552833,0.2104679495,-0.5713222027,-0.2410607636,0.0058486396,-0.1094828844,0.1648474038,0.2017240971,0.0509246401,-0.2241855264,-0.2447012961,0.1638811082,-0.1372713894,0.0477252379,-0.0188017413,-0.0957382917,0.0998383239,0.0703514144,-0.1633325666,-0.1662694663,-0.5500902534,0.0828612894,0.186529398,0.1722971052,-0.377078414,0.2456801385,-0.0064157583,-0.0042385207,-0.1589741558,-0.1405054033,0.1762330979,0.5741929412,-0.2814626396,-0.2942431271,0.0368870944,-0.2076482326,0.0857882202,-0.1217321232,-0.2171171457,-0.2361728847,-0.1228895411,0.4963285327,0.0257592537,0.1940105408,0.3538024426,0.3066567481,-0.1938488036,0.0947411582,-0.1446152478,0.0669935867,0.1638549119,0.1552448869,-0.1388577968,0.5101762414,-0.0036213815,0.3623269498,-0.1596821994,-0.1353380978,0.2480469346,-0.4136278033,0.1829379201,0.0529923774,-0.1954563707,-0.4070477188,-0.2088878453,0.079246603,0.1046092138,-0.0215933714,0.1064467058,-0.042693276,0.1092934012,-0.4261368811,-0.1591197401,0.1760043949,-0.5998126864,-0.0185944643,-0.1473730952,0.1710504442,-0.4020419717,0.1407187879,0.1962788403,-0.0827495307,0.2097803205,-0.1657092869,-0.0217232611,0.1618666798,-0.3959148824,0.2811082304,-0.0157675594,-0.2961775362,0.0769507661,-0.1391290873,-0.0845751315,0.0637777671,0.797721982,-0.0942019075,-0.3272272944,-0.1421581209,-0.4589719176,-0.4927669168,0.0491793118,-0.4003291726,-0.1040405184,0.0499262474,0.2751608789,-0.3529361486,0.081784904,0.2856051922,-0.0333222896,-0.3227697313,-0.1220464185,-0.0563027412,0.0124278367,-0.1439657807,0.0038630879,0.1078337952,0.0371832885,-0.0827960148,-0.4590067267,-0.1218592897,-0.2888567448,0.1653406173,0.2567783296,0.277081877,-0.1532198042,0.1080461815,-0.073895812,0.1903036237,0.1817007512,0.1720701754,-0.0119922785,-0.1438145339,0.1125291288,-0.2174853832,0.41462183,0.0601004884,-0.019789407,0.4138478041,-0.1720751971,-0.0474327877,-0.1397332698,-0.0289316606,0.219427228,0.3861899078,0.0908285826,-0.4624679685,0.3850292265,0.0525398478,-0.0734120458,0.2635189295,-0.0065077529,-0.1166533008,0.4208479822,0.1781477481,0.6603938341,-0.2464488894,0.0212263968,0.2166665345,-0.1932483912,0.5082377195,0.0458221994,0.1675451547,-0.4494053721,-0.088238962,-0.0032246073,-0.0208325889,-0.0487342104,0.1761500686,-0.1165792122,0.1440375298,0.038928248,0.0252577346,-0.0963331014,0.1879375428,-0.0851209313,-0.1762478948,-0.3370991051,0.2434593141,-0.036345277,-0.2323079407,-0.2233270109,-0.1924232095,-0.0084550409,-0.0551238209,-0.1408630759,-0.18514283,-0.0445708819,0.3750240207,-0.0438330099,-0.3344646096,-0.1694511175,0.2673597038,-0.1261745393,0.1092490703,0.00398074,-0.0005696745,-0.0059605092,-0.0760789663,-0.0963567048,0.0754360184,0.090562962,-0.0184686743,-0.3136177957,0.3293158412,0.2190896869,-0.5377386808,0.1065106988,-0.0581550039,0.0350746922,-0.2206358165,-0.0327825472,-0.1255033612,-0.2831252515,-0.3058258593,0.233323887,0.2005994022,-0.0552861877,0.2352708429,0.3613854945,0.0522262305,-0.2309640944,0.119760029,-0.0991700664,0.1457154006,-0.0590493307,0.2901027501,-0.3653233051,-0.2521643341,-0.0394930802,0.1479850113,-0.2624279559,0.1793011576,-0.0067111175,-0.1661863029,-0.0007604156,-0.0954088047,0.3354943395,0.2000451386,-0.2635275126,-0.0343284532,-0.4292801917,0.4129947722,-0.187420547,0.2908354104,0.0477416553,-0.1660306156,-0.1871522516,0.0321107507,-0.4850105345,-0.2301731557,-0.0357287489,0.0593255907,0.0069926633,-0.0477631763,0.0056666569,0.0631515905,0.3455852866,-0.0027798316,-0.003431014,-0.3484045565,-0.0082270186,0.0137176346,-0.0804243386,0.1694184393,-0.0967044979,-0.0827886388,-0.3235602677,-0.1857091933,-0.0599632226,0.2771925926,-0.1297049671,0.1470701993,0.4295483828,-0.3150501847,-0.0670119673,0.05285028,-0.0652609468,0.2217627466,0.0057284567,0.2473478913,0.0480536483,-0.0362059027,0.2022471875,0.1298313737,0.096727103,-0.0383809842,0.2857320905,0.081359975,-0.251101166,0.0581488423,0.5268175602,0.092079863,0.0782798901,0.0456989333,0.2018512189,0.4354124367,-0.5172768235,-0.0245171096,-0.1613010168,-0.2021863014,0.4359527826,0.2774684131,0.2886758745,0.023715863,-0.0110470848,0.1545763463,0.088868998,-0.0540204011,-0.0214465149,0.4221894443,0.1871911138,-0.0488412231,0.5675747395,0.2676724195,0.0662400275,0.1882762015,0.0367020555,0.1498597711,0.3013271987,-0.176353544,0.2299538851,0.014013377,-0.018525809,-0.2787474692,-0.1560726613,-0.1053986549,0.3894237876,0.170496434,0.033648748,-0.1212964952,-0.1965693682,-0.0020069296,-0.163424015,0.0229379572,-0.0623772666,-0.1905083954,0.0014819438,0.0027373147,0.1688669026,-0.2232734263,0.3259580731,-0.1593616456,0.2217888236,-0.1897013336,0.3834525645,-0.0696084723,0.2390525788,-0.1094118208,-0.199721843,-0.0254955702,-0.0693785325,-0.155263707,0.3439956009,0.1947231591,-0.1392541826,-0.0998403579,0.0485636219,-0.2059310824,-0.1380547285,-0.1576255113,-0.1479854882,0.1985811293,-0.0086966902,0.211063683,0.3714939356,-0.2656693459,0.2914207578,0.0902619213,0.1243373305,-0.0337071344,0.220386222,0.2795262635,-0.3818406463,0.0499219522,-0.1054142937,-0.2984302342,-0.212967068,0.4633649886,-0.2778484523,-0.0615501739,-0.0243083108,0.2047193497,0.0068233414,0.0913381577,0.2940889001,0.2840595543,-0.0201185457,0.0968312249,-0.3403924108,0.1523322314,0.0418291613,-0.1389815807,-0.1681732088,0.167258814,0.2638446689,0.1836084872,0.2422299236,-0.0132133821,-0.0441721343,0.3362065852,-0.3135394156,-0.0912041962,-0.063454777,0.2471351326,-0.0174588151,-0.024579389,0.1253470331,-0.1420980841,0.2855964005,-0.160123542,0.0010203043,0.0976110175,-0.088498421,0.3796849549,0.0489379764,0.3516635597,-0.3393688202,0.2192280293,-0.0766993463,-0.2807760835,-0.1411431283,0.1555172205,0.1545360833,0.1638946831,0.1129808649,-0.2505370378,-0.0831802264,0.1967108548,0.0114984736,-0.1834111512,-0.0000354241,0.2110721916,-0.04290317,-0.1575335562,-0.158992961,0.1991970241,0.0546980202,-0.0331352502,0.0876633078,-0.2895196378,0.071242325,-0.2618819773,-0.3346387446,-0.2955847383,0.0571167357,0.4148213267,0.2260582447,-0.0717805699,-0.1028660461,0.228287369,-0.0454057716,-0.0539877564,0.2059537172,0.0364880301,0.0366009362,-0.1425175667,0.0098954123,0.0355803519,-0.018932689,-0.3302782476,-0.4358122349]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/3","title":"[Feature] More dataset outputs","comments":"Now Pandas is available.","body":"Add the following dataset outputs:\r\n\r\n- Spark\r\n- Pandas","comment_length":4,"text":"[Feature] More dataset outputs\nAdd the following dataset outputs:\r\n\r\n- Spark\r\n- Pandas\nNow Pandas is available.","embeddings":[-0.5605735183,-0.2810478806,-0.3054264486,-0.0202812366,0.423039794,0.1238518879,0.4131855071,0.4143275917,0.0950963944,0.1730397195,-0.2762993574,0.46549353,-0.0291814078,0.6266698837,0.0863993093,-0.2519063354,0.0025802348,-0.0078618126,-0.1247697026,-0.0834056139,-0.1543336064,0.1792408079,0.0557166338,0.0634401813,-0.382357657,-0.0815719292,-0.2427468747,-0.233208254,-0.0887345225,-0.2141350508,0.0032287149,-0.0216396302,0.101666458,0.2231214046,-0.0000948048,-0.1426877826,0.3449600041,0.0059489026,-0.0713188201,-0.1804521084,-0.0158907361,-0.2163418233,0.3632938862,-0.250832051,-0.1199385598,-0.1625438035,0.2464133352,0.0878912881,0.2174548805,0.1959855109,0.3818102777,0.118739225,0.1606528163,-0.2769883871,0.0605093986,0.1036021337,-0.1592815816,0.0728720725,0.154288426,-0.1278614402,-0.223054871,0.1584609747,-0.0065630595,-0.0704571977,-0.1470735967,0.0771920457,-0.0313883089,-0.3854606748,0.071273841,0.2064695805,0.7287855148,-0.0774312541,-0.3670248389,-0.1298147738,-0.0934322253,-0.110809505,0.0437804125,0.1213161573,0.0367335044,0.2012685835,0.1808751971,-0.184784025,-0.112901248,0.1061843634,-0.0001378999,0.5311102867,-0.0609724782,-0.075129278,0.2726037204,0.0401786901,0.0839943588,0.1263711005,-0.0013597836,-0.0541342534,-0.2541925311,-0.2182553411,0.0532635786,-0.2972331345,0.0074528134,0.1017227322,-0.0190287791,0.2302965075,-0.1384139955,0.1463084668,0.142423749,-0.1111324951,0.0726535395,0.3017404675,0.4835239053,-0.0342323333,0.1072895527,-0.1857201904,-0.1346861571,-0.1148521006,-0.1808778346,0.200067699,-0.0687409788,0.0387212038,-0.1013096273,-0.1149253845,0.0723308027,-0.1356645226,-0.0991640016,0.1802845746,-0.0904305428,-0.0756697282,-0.1146319881,0.2313925922,-0.0181881227,-0.0489992313,-0.1948415786,0.3051385581,-0.1226211563,-0.3080699742,0.125004828,-0.0404645652,0.2734992206,-0.0065062963,-0.0041559455,0.3253841102,-0.0071257986,-0.0153654898,-0.1694819778,0.1860501319,0.0630691648,0.1012741029,0.1129188612,-0.0020006935,-0.1815344542,-0.0317662768,-0.0191411413,-0.1807615906,-0.4876521826,0.294763267,-0.0130519858,-0.1253746152,-0.2298713475,0.4765610099,-0.1761758924,0.0283098705,-0.0035199907,0.0755738094,0.120968245,-0.2208720744,0.0753712133,0.2852086127,-0.5380459428,0.0721722692,-0.0507098027,-0.1001561433,0.1358697563,0.0324235521,-0.0798579231,0.0007258386,-0.2192491293,0.0981721506,0.25197047,-0.1772865206,0.0138912592,0.1885907352,-0.0249510873,-0.1528000832,0.1500916034,0.1654195487,0.181151256,0.2337187529,0.1552174836,0.0856589079,0.0076142047,-0.1174200624,0.0075881453,-0.22055462,0.0700652748,0.0417803861,0.5401762724,0.1441537887,0.2044215947,-0.1661584079,0.2014704496,0.1359596103,-0.0697166324,0.1901840419,0.4344332814,0.4043529332,0.0226896442,-0.4787076116,-0.0081957998,0.0664163008,0.3725768626,0.3009994626,-0.0487072654,-0.2460282147,-0.0217069481,0.0543324463,0.1968643963,-0.0269891601,0.3363642395,-0.1078399867,0.0876373872,0.1878613234,-0.1111258641,0.112817578,-0.0125463419,0.0462078303,-0.1610002667,0.2060691267,-0.208450526,-0.1796584576,0.2370758206,-0.2700490057,-0.024391951,-0.0107206395,-0.0448796228,0.3032186031,-0.0085769128,0.082892023,0.3056030273,0.4450861216,0.230081439,-0.0435081609,-0.1764153391,0.0084616803,-0.0617111064,0.1593791693,-0.2148915678,0.2638334036,-0.0093872212,-0.1657611579,-0.0810587257,0.2210919857,0.5023257136,0.1264224052,-0.1067883,-0.0853123665,-0.1146957055,-0.0014639128,-0.1901257187,-0.1857460737,-0.5784315467,0.0764239952,0.285320133,0.0544704273,0.2389473468,0.0042073573,-0.2317944765,0.1113900989,0.4582757056,0.135410741,0.0016279647,0.4541945755,0.1045402735,-0.2721461952,-0.2331648916,-0.1740688533,0.1451184303,0.2583128214,0.0707582384,0.0955379829,0.1215280369,-0.2188212425,-0.5103621483,-0.015106854,-0.1527068168,0.0529386327,-0.1023004204,-0.0776967704,-0.3180305064,-0.3197216392,0.3011444509,0.0200242437,0.0487386473,-0.3283027411,0.1829854399,0.2900992036,-0.1077369079,0.3175334632,0.0108905146,-0.0419596359,0.1230173782,0.0965358689,-0.3462607265,-0.1322319806,-0.0165052116,0.3435741961,-0.0356403552,0.0216553863,0.3010875881,0.0059211696,0.1793174446,-0.599198997,-0.4405664802,0.0796404257,-0.1310638338,0.3994837105,0.210849002,0.0851671025,-0.2739203274,-0.1750560105,0.1714029163,-0.2912486196,0.0171745699,-0.1276642978,-0.0685581192,0.0376335345,-0.1749272794,-0.177394852,-0.1200257912,-0.5138010383,0.2114157975,0.1841824055,0.0303549916,-0.2707962394,0.2177503705,-0.0213775635,0.1028519794,-0.0757539719,-0.2451559007,0.0105328597,0.5832479,-0.3347376585,-0.356210649,0.0132578937,-0.1989607811,0.0199488625,-0.187410593,-0.1755174398,-0.328805387,-0.1931298822,0.2403131872,0.0095026232,0.0934632942,0.3450869322,0.1958967298,-0.2722289562,0.0006106629,-0.1494863033,-0.0556139685,0.2936645448,0.0974730402,-0.0534014255,0.5065205097,0.0023067973,0.3367478549,-0.2898570299,-0.0682153627,0.2670586407,-0.3250327408,0.2417049259,-0.1058027223,-0.2424339354,-0.2890582681,-0.2316591591,0.1595398933,0.0189887341,-0.0560348071,-0.0865928158,-0.1490261704,0.0295225345,-0.3370073438,-0.1086558998,0.1543356478,-0.3889641166,0.0858907774,-0.0008442936,0.1756362915,-0.2427693158,0.0069116866,0.1443759054,0.0352479704,0.1494486779,-0.1105104461,-0.1672111601,-0.0885647163,-0.4571197331,0.0987799764,0.0396570861,-0.1687824279,-0.1550947875,-0.0513181202,0.0755318999,0.0283715464,0.8782949448,-0.2219055593,-0.3527308702,0.072358042,-0.404725194,-0.0755736232,-0.0197900683,-0.4483616352,-0.0470821708,0.1619489342,0.4874007404,-0.3362988532,0.0453025773,0.2467626929,-0.0747840926,-0.2219170034,-0.0865525007,-0.0657860562,0.0498539992,-0.3677957356,-0.021906646,0.176725477,-0.0736308694,-0.1566028893,-0.270894289,-0.106637612,-0.190312773,0.2195025533,0.1826120764,0.1212244853,0.0516403355,0.0886588395,0.0236739498,0.0043713758,-0.0170570295,0.1285455525,-0.0184324738,-0.126098752,0.2170646191,-0.2059428394,0.3460121453,-0.0253653582,-0.0723286048,0.1494430751,-0.2140778154,0.0317167342,-0.2130894661,0.1812683195,0.4028437436,0.2887655199,-0.0823870301,-0.5523777008,0.1910067201,-0.0744587779,-0.0575884283,0.2728565931,0.0935505182,-0.0352263376,0.1793181747,0.0605924986,0.6855868101,-0.3739259541,-0.0600604005,0.0884885639,-0.0440304093,0.5817507505,-0.0480769016,0.2983770669,-0.2996833622,-0.2692142427,-0.0357454233,0.0829030424,0.0665609911,0.1150633544,-0.1283308268,0.0994718894,0.1507948488,0.1987924576,-0.1000038534,0.2081843466,-0.0234450344,-0.1670836508,-0.1373332441,0.3173078597,-0.040886756,0.0146359736,-0.0702573732,-0.167535305,-0.0029417854,-0.0755562484,-0.2504606545,-0.1162345484,0.0852141157,0.5641095042,-0.0178463254,-0.2880498469,-0.1537417471,0.2848911583,0.0746188611,0.2659211159,-0.0496371351,0.0518127829,0.0133740194,-0.1886545122,-0.1566685438,0.1627478749,0.1032561734,-0.0677636936,-0.3047872186,0.3190539479,0.2171248198,-0.3423257172,-0.0963698924,-0.0687277019,0.0339904465,-0.2214225382,0.0525771342,-0.0461896323,-0.2415038943,-0.3203812242,0.2794230878,0.1294573992,0.1194539219,0.0933848545,0.2140615433,0.1116963327,-0.2393906415,0.2794671953,-0.0085600298,0.1449804604,0.0178069118,0.2477379143,-0.3582419455,-0.2684030533,-0.1692912132,-0.1615322381,-0.2244459093,0.1871069521,-0.0278936699,-0.19314529,-0.0125143612,-0.0747047365,0.1687876284,-0.0627877265,-0.2604813576,0.0257015284,-0.3593687415,0.3351815641,-0.196124211,0.3223096132,0.1477462202,-0.1462188214,0.0138628157,0.0016586004,-0.4939669371,-0.2257616669,-0.1515235752,0.0194490552,0.2490849495,0.0184926726,-0.0030868447,-0.0040895822,0.3408665955,0.0728568062,0.0415344313,-0.3801183105,-0.0260917582,-0.0056847306,-0.0325670689,0.1204404458,-0.0521378964,0.0173814818,-0.3111100495,-0.1295497864,-0.0785105079,0.2200359404,-0.1028982624,0.3636810482,0.2160535157,-0.1050383598,0.2245359868,0.079274781,-0.0435377508,0.2095227093,0.0002123346,0.0851582214,0.2700780928,-0.0095188692,-0.0032267796,0.0291238837,-0.0664864779,0.1133029982,0.2777047753,0.1600766927,-0.0637320504,0.0638066307,0.2948674262,0.0085744886,0.1293479949,0.0983286127,0.3077301383,0.3923206925,-0.4944099486,-0.1569782346,-0.0777174607,-0.1964470148,0.3806734085,0.1432294846,0.3788931072,0.0784265772,0.179358393,0.1655058116,0.1176585481,-0.093368955,0.11517483,0.4874716401,0.0072200643,-0.0305301081,0.4043377638,0.1300925165,0.0470393896,0.4098467231,0.2214134783,-0.0268930811,0.0102639403,-0.0515296087,0.2898014188,-0.1186611876,-0.0318470299,-0.3437842131,-0.2092956752,0.0752588883,0.3516072631,0.1690769792,-0.1826175749,0.1014818549,-0.3154641986,0.0248937402,-0.3009927869,-0.0070432257,0.0793082565,-0.1746337414,-0.0107114697,0.1721325815,0.1174693182,0.0315906741,0.353726238,-0.1804667711,0.1429551542,-0.0154998684,0.2432420254,-0.0905116796,0.1912090331,-0.1220990643,-0.2279142141,-0.0827303082,-0.09575627,-0.1972584277,0.2812512517,0.2357123792,-0.19360587,0.0182609055,-0.0738815516,-0.1812521666,-0.1428563297,-0.0803702772,0.098079972,0.2053376585,0.1492051631,0.097399272,0.3660574555,-0.2904982269,0.2368472666,-0.0093842922,0.0460706465,-0.1104869843,0.0732707232,0.1850090474,-0.2846384943,-0.0303193759,-0.0847255066,-0.3373473287,-0.1452928036,0.202979967,-0.2199544609,-0.0211428832,-0.1853465885,0.2060076594,-0.0245829672,0.2111359537,0.2360705733,0.073357515,0.1107978374,0.0348058008,-0.3061364293,0.0605645329,-0.0012051674,0.0400061719,-0.2184312791,0.0434642024,0.2461922616,0.0692259222,0.1149502173,-0.237454161,-0.1204208955,0.2032379508,-0.22926566,-0.0817837492,-0.1315024793,0.1224967837,0.0025970449,0.0093974192,0.1823441535,-0.1394926906,0.2917674184,-0.204755187,-0.0329437256,0.0095944023,-0.0905954838,0.3981857896,0.0937932283,0.3259840012,-0.4571531117,0.1272998303,-0.0965878293,-0.3735011518,-0.1124190837,0.1340454221,0.0838972107,0.2763010561,-0.0603058785,-0.1031813994,-0.0794479325,0.2862478495,0.019548988,-0.0081442939,-0.1282301396,0.0700694025,-0.1110526994,-0.2587448061,0.0453246646,0.2137951702,0.0182177164,-0.315553546,0.0733469427,-0.2938712835,0.0089822123,-0.2037923187,-0.3982305527,-0.4965318143,-0.2157392502,0.5891416669,0.3988456428,-0.1114158258,-0.0471769273,0.1822317243,0.0891050473,0.0741813481,0.3256694078,0.197248444,-0.0462139025,-0.1195436046,-0.129973188,0.1732568741,0.0160978995,-0.155228287,-0.2832098305]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2","title":"Issue to read a local dataset","comments":"My first bug report \u2764\ufe0f\r\nLooking into this right now!","body":"Hello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)","comment_length":10,"text":"Issue to read a local dataset\nHello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)\nMy first bug report \u2764\ufe0f\r\nLooking into this right now!","embeddings":[-0.2719613612,0.1254656762,-0.0766852647,0.256811142,0.3318964839,0.0312691405,0.2642410994,0.3003315628,0.3266820014,0.2098683715,0.1191096306,0.2447595894,-0.2942002714,0.1220210195,0.1568034738,-0.1723625809,-0.1275069416,0.3326410651,0.1501982808,-0.3661575317,-0.27604267,0.1413950026,-0.1191338152,0.1533625871,-0.2043232173,0.0607061796,-0.036645785,0.4012748301,-0.2069403231,-0.2449603528,0.1894469708,0.0119974464,0.2054079175,0.122409679,-0.0001004647,0.1904105544,0.2995985448,0.0436597355,-0.1575898379,-0.364929229,-0.0587465651,-0.3323357999,0.2688184083,-0.3986444175,-0.3326914608,-0.1401537061,0.0905135646,-0.4378168285,0.4685541093,0.3634622395,0.3068140745,0.1048855186,-0.0699259415,0.1463657767,0.195211947,0.152147904,-0.0873053744,0.2656306922,0.031915091,-0.2012286931,-0.1960862428,0.1223725975,-0.2222957164,0.0449407585,0.2718259394,0.1370849758,0.182052508,-0.1032298878,-0.032771349,-0.0018344095,0.3769168258,-0.2492007166,-0.2248517573,-0.13559407,0.0572931319,-0.2591866851,0.0394319892,0.342464596,-0.349896282,0.3455857933,-0.2791967094,0.1082817465,-0.3008818924,0.2382965088,0.1193012074,0.3104308248,-0.076890111,0.1048834771,0.0907402039,0.2240543067,0.1786538064,-0.1133791059,0.2185333967,0.2561581731,-0.0302851815,0.0967433602,0.3372043669,0.0830674171,-0.0235993508,-0.0186839495,0.1047985032,0.0249034651,0.0130941896,0.3297557831,0.1096323282,0.1400410384,0.1735097319,0.087680243,0.0717137232,-0.0622796975,-0.2097682357,0.0256421678,-0.1461680084,-0.283438772,0.0297457296,0.2404010743,0.4468615353,-0.1885111183,-0.1420501471,0.2193927914,-0.2050481737,-0.0884669423,0.1696829945,0.4059306085,-0.0357793234,0.2412724048,0.1384448409,0.0997360051,-0.3508428931,-0.1973964125,-0.3274970055,-0.0029654084,-0.2937141359,-0.0887341276,0.3137897849,0.0290612634,0.3492821157,-0.0983311981,-0.2490224689,-0.006291463,0.1585656404,-0.2560243607,0.1404405981,0.138631627,0.0651841164,0.034821,0.2272037119,-0.4230361283,-0.1299738437,0.2165416181,0.0924713016,-0.4230499268,-0.1779252738,0.3226445317,-0.0838615969,0.1769313514,0.180574581,0.1790369749,0.1407811642,-0.036249388,-0.2524119318,-0.0215898771,-0.459605664,-0.2430503964,0.2107181102,0.3516464829,-0.4640030563,0.0853260383,-0.164036572,-0.1117871031,0.3510994613,0.2487842441,-0.3300900161,0.4454449415,-0.1017358303,0.1409201324,0.6580390334,-0.3004144132,-0.3807192743,0.6322368979,-0.3527038991,-0.1465917528,0.0498830862,0.0073254914,-0.0018453643,0.2033673376,0.2166674733,0.3950749636,-0.1738853306,0.1002998799,-0.4054657817,-0.152086556,0.1437031329,0.1029602066,-0.0783379376,0.1322496831,0.1525419801,-0.05174575,0.497027427,-0.1389278173,0.1745269895,0.3185678422,0.2489011884,-0.0089441948,-0.1572516412,-0.0782212839,-0.2163548321,0.04727301,-0.1540113091,-0.0111839129,-0.1268217266,-0.1966494024,-0.3752773702,0.1024699733,-0.4161543548,-0.0244457163,0.2532124221,0.2463398874,0.0947190672,0.0092416257,-0.0782992616,-0.1027992368,-0.1485022753,-0.2201791555,-0.3223364353,0.2954417169,-0.3488661051,-0.1378353983,0.1312768757,0.3829959631,0.1777834296,-0.1002675742,-0.052053079,0.2365419269,-0.0244584288,0.0325130597,0.0145486901,-0.1853820533,-0.0120613975,-0.279640317,0.1465234011,0.0235882532,0.348398447,0.0077597764,-0.3682603538,0.377537787,-0.0371985137,0.0586498044,-0.0751952454,-0.0948600769,0.3058986366,-0.2762518227,-0.0493951924,-0.0236396324,0.2183714211,0.0369678438,0.1693601906,0.04129586,-0.3924960494,0.047164917,0.3549182117,-0.072537221,-0.1731103212,0.0018169068,0.0174868628,-0.0824427828,0.164626956,0.2774141133,0.3839050233,0.2610786259,0.0662332624,0.1700644791,-0.2135237604,-0.1831107736,-0.054769367,-0.0941865519,-0.027594151,0.3564697206,0.1723606884,-0.1707368493,-0.3508794904,0.0280967355,0.155304268,0.3651978076,-0.1970971674,0.2237509936,-0.2249526829,-0.2432209104,-0.3664650917,-0.0820040554,0.0938911885,-0.3452377021,-0.1890164614,0.1978519261,-0.107149139,-0.1096806899,-0.1519416273,-0.0034873104,0.0685041398,-0.1828078628,0.0319254473,-0.29726246,-0.4607281685,0.1170991212,0.3691140115,0.3623128831,0.4312880337,0.0150138075,-0.1073622257,0.0393667519,-0.0697979704,0.076720871,-0.100895606,0.4542248547,-0.0569396578,0.0416780375,-0.2283503711,-0.4602437317,0.5433725715,-0.0930568278,-0.1630839258,0.003641258,0.1127770245,-0.2605376542,-0.2423969209,-0.5966646671,-0.376578927,-0.3775201142,0.2393841147,0.2145242095,0.2878387868,0.3024253845,0.0545422919,0.227584064,0.2486261129,0.0220368877,-0.1577897519,-0.2745801508,0.1092553139,-0.3437816501,-0.3863990009,-0.0418258309,-0.2934792042,0.2532537878,-0.177119717,-0.5053718686,0.106159687,-0.2085334063,0.175987035,0.1183275729,-0.0756837502,0.3462190628,0.0716120824,-0.0953156427,-0.1507954448,-0.1724995971,0.17420201,0.0590066873,0.0601394363,0.060319189,0.2799823284,-0.1291005462,0.4796222448,0.0056985766,0.0647447631,0.2293583602,-0.1257217079,0.0351171568,-0.0269690529,-0.5060185194,0.0983495489,0.1094464958,-0.324074626,0.239359051,-0.0309577566,-0.332950443,-0.2473927289,0.0161873512,-0.5706443191,-0.3236039579,0.0986715853,-0.0077600973,0.1682422012,0.0484663807,-0.0216584858,-0.1972861886,-0.0092988918,-0.1825920939,0.2409341633,0.0875305682,0.0187797509,-0.3264673948,-0.0633295923,-0.5391075611,0.1027067751,0.0346259587,0.3136318624,-0.1360362172,-0.1588585377,0.2374024689,-0.1259416044,0.355168581,-0.1997641325,0.1949021965,0.2806993425,0.0566505939,-0.2779286504,0.0598577335,-0.0828077719,0.194408074,0.0551685654,0.1930948049,-0.0358043946,-0.2404814959,0.3322658539,0.086165607,-0.3055275679,-0.4114955664,-0.1551193148,-0.2620900869,-0.1998126209,-0.0636808574,-0.2033237219,0.3930025399,-0.0087458072,-0.0178830903,0.0776314586,0.0366086811,0.1659790725,0.0659319609,0.144485414,0.1646545529,0.1758186221,0.0877508596,0.4288354814,-0.1880310923,0.6513575315,-0.0744920522,-0.1693410277,0.0129821599,-0.1879096776,0.2662377656,0.2502310276,-0.1011858135,-0.2100187987,0.1666356325,-0.111804314,-0.1770980358,0.394836545,0.107521981,-0.1158342734,-0.0676748827,-0.5222005844,0.3812000751,-0.114734821,-0.0565005615,-0.0064784149,-0.1909905374,-0.1881240904,0.5036737323,0.1323155314,0.864397049,-0.1980929971,0.0349103101,0.4668910205,0.007784619,0.1646999121,-0.1840792,0.2341465205,-0.5701379776,0.019398775,0.0076868804,-0.0287059639,0.1016457453,0.2519910634,-0.0923879817,0.0385436043,0.1525755972,-0.1554612368,-0.1020387262,0.283041954,0.0544179939,-0.1383448094,-0.3233302832,0.2222985327,-0.0313629843,0.1917642057,0.0116054919,-0.2589574754,-0.1316404045,-0.1493984312,0.0529658459,0.1751309335,-0.1544432491,-0.0379619598,-0.0486805253,-0.3220744729,0.0414558649,0.3110913336,0.1537470967,0.2356782109,-0.2747209668,-0.1066584215,-0.0653964505,-0.1012914628,-0.1185328737,0.0051975329,0.1416031569,0.0174175873,-0.4252744019,-0.0209827181,-0.0953727588,-0.1059336886,0.128332004,0.1576503962,0.2031957209,-0.3334443569,-0.406455636,0.0477226153,-0.1087376252,-0.0600517206,0.2278128713,-0.021711776,-0.1207553223,-0.107445024,0.0386740565,-0.2580270767,0.0760376602,0.3284548521,0.261800617,0.1757760346,0.3383722305,0.1642873436,-0.0343141928,-0.3920804858,-0.0356164388,0.0704319179,-0.2614100575,-0.1708102077,0.1431283057,0.0959377959,0.2589941323,0.3358263373,0.0868119672,0.0314114615,-0.0264190324,-0.4197938442,-0.4787349105,-0.156475544,-0.0960422233,0.2361431718,0.3992731869,0.350523293,0.0761040971,0.1712295115,-0.4348987043,-0.1582422704,-0.3672046959,0.0972431526,0.10911531,-0.1289702207,0.0431214049,-0.0544742681,0.231937632,-0.0266497321,-0.1715811938,-0.3512310386,0.0337690935,0.0571922138,0.1569814384,-0.1603884697,-0.126356557,-0.1384025216,-0.037456084,-0.0434973948,-0.0566055775,0.1242901683,-0.0668676496,0.4345211685,-0.0098284902,0.1701683551,-0.0747041106,0.0009316503,-0.2399915755,0.1899886429,-0.0329511464,0.0423244759,0.1744868606,-0.0874913782,-0.150878042,-0.0646028668,-0.0523427799,0.3434982598,0.4014250636,-0.3132493496,-0.0029764506,0.1609406322,0.3720909357,0.1368874609,-0.0667902604,0.1870818138,0.3331491053,0.3209904134,-0.2170785218,-0.0078470316,0.1768234521,0.0243657287,0.1571125537,0.2079402208,0.1602826715,-0.1185155734,-0.0839414075,0.0089579439,0.219958663,0.0269431118,-0.172602579,0.5578839779,-0.106967777,0.0147087779,0.0153596047,0.2626159787,0.1503474712,0.3625153303,0.0332685411,-0.0691114068,-0.0515283495,0.3558601439,-0.2451538295,-0.492842555,0.2119537294,0.1049965546,-0.138162747,0.0064495583,-0.1462257951,0.2908702791,0.0504665002,-0.045047041,-0.0586939566,0.4341705739,-0.2295295745,-0.1355336159,-0.441393733,-0.0720626786,-0.2986844182,-0.0694366246,0.0210932959,0.1438496411,-0.0314354375,0.0074021979,-0.0485220365,-0.1798207313,0.1309196651,0.2044498771,-0.136657387,-0.3524982631,0.3145772517,0.1482188404,0.1692373157,-0.1363149136,0.1530724317,0.2172753513,0.4162555933,-0.19093889,0.0166432783,-0.1435080916,-0.1691758782,-0.3510366082,0.2251788527,0.0193293262,0.3033943474,0.5757574439,0.2563101649,-0.1848143041,0.2783590257,0.1571856737,0.0825569481,0.0694978014,0.2987569869,0.157189101,-0.1272757649,-0.2587654889,-0.1263581514,-0.4825097919,-0.0734842867,0.337828815,0.0948294923,0.5023238063,-0.2306268215,0.1396726817,-0.1246094182,0.4588422775,0.269364357,0.1685478389,-0.4026035964,-0.2152403295,-0.5258612633,0.0776555389,-0.0867535397,-0.2213983834,-0.0802540183,0.2547663152,0.0406547263,0.1811081916,-0.5375036001,0.1512551457,-0.1450660825,0.2304359972,-0.0383698754,-0.1621061862,-0.1003011614,-0.0384932198,-0.0739434585,-0.4356856048,0.1343759298,-0.106498912,0.2026229054,-0.2383623719,0.0076444671,-0.0351861045,0.0809182376,0.03337726,0.2150260955,0.3726360202,-0.24440898,0.2414368838,-0.2125802338,0.0813362673,-0.2655959129,0.1747843325,0.1388967186,0.5628277659,-0.2568778694,0.2340295315,-0.3165405989,0.0263359509,0.0514674932,0.0794786662,-0.0276819076,0.1196670309,-0.3101905584,0.1019593179,0.1867545545,0.158185035,0.0220817178,0.1333450824,-0.2682808638,-0.2907093763,0.4097493589,-0.2822920084,-0.2055113912,0.1762141585,0.1532312185,-0.0029166152,-0.0402859598,-0.4433356225,0.0254867394,0.3729545176,-0.1966754198,0.0635944456,0.2106608301,0.0638403893,0.0155626731,-0.0240852218,-0.1181434765,0.1259868592,-0.0181906056,-0.1313410401,-0.2844596803]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2","title":"Issue to read a local dataset","comments":"Ok, there are some news, most good than bad :laughing: \r\n\r\nThe dataset script now became:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(self, **config):\r\n self.train = config.pop(\"train\", None)\r\n self.validation = config.pop(\"validation\", None)\r\n super(Bbc, self).__init__(**config)\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, description=\"bla\", features=nlp.features.FeaturesDict({\"id\": nlp.int32, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": self.train}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": self.validation})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"id\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nAnd the dataset folder becomes:\r\n```\r\n.\r\n\u251c\u2500\u2500 bbc\r\n\u2502 \u251c\u2500\u2500 bbc.py\r\n\u2502 \u2514\u2500\u2500 data\r\n\u2502 \u251c\u2500\u2500 test.csv\r\n\u2502 \u2514\u2500\u2500 train.csv\r\n```\r\nI can load the dataset by using the keywords arguments like this:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\"})\r\n```\r\n\r\nThat was the good part ^^ Because it took me some time to understand that the script itself is put in cache in `datasets\/src\/nlp\/datasets\/some-hash\/bbc.py` which is very difficult to discover without checking the source code. It means that doesn't matter the changes you do to your original script it is taken into account. I think instead of doing a hash on the name (I suppose it is the name), a hash on the content of the script itself should be a better solution.\r\n\r\nThen by diving a bit in the code I found the `force_reload` parameter [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/nlp\/load.py#L50) but the call of this `load_dataset` method is done with the `builder_kwargs` as seen [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/nlp\/load.py#L166) which is ok until the call to the builder is done as the builder do not have this `force_reload` parameter. To show as example, the previous load becomes:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\", \"force_reload\": True})\r\n```\r\nRaises\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 283, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 170, in builder\r\n builder_instance = builder_cls(**builder_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/datasets\/84d638d2a8ca919d1021a554e741766f50679dc6553d5a0612b6094311babd39\/bbc.py\", line 12, in __init__\r\n super(Bbc, self).__init__(**config)\r\nTypeError: __init__() got an unexpected keyword argument 'force_reload'\r\n```\r\nSo yes the cache is refreshed with the new script but then raises this error.","body":"Hello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)","comment_length":353,"text":"Issue to read a local dataset\nHello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)\nOk, there are some news, most good than bad :laughing: \r\n\r\nThe dataset script now became:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(self, **config):\r\n self.train = config.pop(\"train\", None)\r\n self.validation = config.pop(\"validation\", None)\r\n super(Bbc, self).__init__(**config)\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, description=\"bla\", features=nlp.features.FeaturesDict({\"id\": nlp.int32, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": self.train}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": self.validation})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"id\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nAnd the dataset folder becomes:\r\n```\r\n.\r\n\u251c\u2500\u2500 bbc\r\n\u2502 \u251c\u2500\u2500 bbc.py\r\n\u2502 \u2514\u2500\u2500 data\r\n\u2502 \u251c\u2500\u2500 test.csv\r\n\u2502 \u2514\u2500\u2500 train.csv\r\n```\r\nI can load the dataset by using the keywords arguments like this:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\"})\r\n```\r\n\r\nThat was the good part ^^ Because it took me some time to understand that the script itself is put in cache in `datasets\/src\/nlp\/datasets\/some-hash\/bbc.py` which is very difficult to discover without checking the source code. It means that doesn't matter the changes you do to your original script it is taken into account. I think instead of doing a hash on the name (I suppose it is the name), a hash on the content of the script itself should be a better solution.\r\n\r\nThen by diving a bit in the code I found the `force_reload` parameter [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/nlp\/load.py#L50) but the call of this `load_dataset` method is done with the `builder_kwargs` as seen [here](https:\/\/github.com\/huggingface\/datasets\/blob\/master\/src\/nlp\/load.py#L166) which is ok until the call to the builder is done as the builder do not have this `force_reload` parameter. To show as example, the previous load becomes:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc\", builder_kwargs={\"train\": \"bbc\/data\/train.csv\", \"validation\": \"bbc\/data\/test.csv\", \"force_reload\": True})\r\n```\r\nRaises\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 283, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/load.py\", line 170, in builder\r\n builder_instance = builder_cls(**builder_kwargs)\r\n File \"\/home\/jplu\/dev\/jplu\/datasets\/src\/nlp\/datasets\/84d638d2a8ca919d1021a554e741766f50679dc6553d5a0612b6094311babd39\/bbc.py\", line 12, in __init__\r\n super(Bbc, self).__init__(**config)\r\nTypeError: __init__() got an unexpected keyword argument 'force_reload'\r\n```\r\nSo yes the cache is refreshed with the new script but then raises this error.","embeddings":[-0.2719613612,0.1254656762,-0.0766852647,0.256811142,0.3318964839,0.0312691405,0.2642410994,0.3003315628,0.3266820014,0.2098683715,0.1191096306,0.2447595894,-0.2942002714,0.1220210195,0.1568034738,-0.1723625809,-0.1275069416,0.3326410651,0.1501982808,-0.3661575317,-0.27604267,0.1413950026,-0.1191338152,0.1533625871,-0.2043232173,0.0607061796,-0.036645785,0.4012748301,-0.2069403231,-0.2449603528,0.1894469708,0.0119974464,0.2054079175,0.122409679,-0.0001004647,0.1904105544,0.2995985448,0.0436597355,-0.1575898379,-0.364929229,-0.0587465651,-0.3323357999,0.2688184083,-0.3986444175,-0.3326914608,-0.1401537061,0.0905135646,-0.4378168285,0.4685541093,0.3634622395,0.3068140745,0.1048855186,-0.0699259415,0.1463657767,0.195211947,0.152147904,-0.0873053744,0.2656306922,0.031915091,-0.2012286931,-0.1960862428,0.1223725975,-0.2222957164,0.0449407585,0.2718259394,0.1370849758,0.182052508,-0.1032298878,-0.032771349,-0.0018344095,0.3769168258,-0.2492007166,-0.2248517573,-0.13559407,0.0572931319,-0.2591866851,0.0394319892,0.342464596,-0.349896282,0.3455857933,-0.2791967094,0.1082817465,-0.3008818924,0.2382965088,0.1193012074,0.3104308248,-0.076890111,0.1048834771,0.0907402039,0.2240543067,0.1786538064,-0.1133791059,0.2185333967,0.2561581731,-0.0302851815,0.0967433602,0.3372043669,0.0830674171,-0.0235993508,-0.0186839495,0.1047985032,0.0249034651,0.0130941896,0.3297557831,0.1096323282,0.1400410384,0.1735097319,0.087680243,0.0717137232,-0.0622796975,-0.2097682357,0.0256421678,-0.1461680084,-0.283438772,0.0297457296,0.2404010743,0.4468615353,-0.1885111183,-0.1420501471,0.2193927914,-0.2050481737,-0.0884669423,0.1696829945,0.4059306085,-0.0357793234,0.2412724048,0.1384448409,0.0997360051,-0.3508428931,-0.1973964125,-0.3274970055,-0.0029654084,-0.2937141359,-0.0887341276,0.3137897849,0.0290612634,0.3492821157,-0.0983311981,-0.2490224689,-0.006291463,0.1585656404,-0.2560243607,0.1404405981,0.138631627,0.0651841164,0.034821,0.2272037119,-0.4230361283,-0.1299738437,0.2165416181,0.0924713016,-0.4230499268,-0.1779252738,0.3226445317,-0.0838615969,0.1769313514,0.180574581,0.1790369749,0.1407811642,-0.036249388,-0.2524119318,-0.0215898771,-0.459605664,-0.2430503964,0.2107181102,0.3516464829,-0.4640030563,0.0853260383,-0.164036572,-0.1117871031,0.3510994613,0.2487842441,-0.3300900161,0.4454449415,-0.1017358303,0.1409201324,0.6580390334,-0.3004144132,-0.3807192743,0.6322368979,-0.3527038991,-0.1465917528,0.0498830862,0.0073254914,-0.0018453643,0.2033673376,0.2166674733,0.3950749636,-0.1738853306,0.1002998799,-0.4054657817,-0.152086556,0.1437031329,0.1029602066,-0.0783379376,0.1322496831,0.1525419801,-0.05174575,0.497027427,-0.1389278173,0.1745269895,0.3185678422,0.2489011884,-0.0089441948,-0.1572516412,-0.0782212839,-0.2163548321,0.04727301,-0.1540113091,-0.0111839129,-0.1268217266,-0.1966494024,-0.3752773702,0.1024699733,-0.4161543548,-0.0244457163,0.2532124221,0.2463398874,0.0947190672,0.0092416257,-0.0782992616,-0.1027992368,-0.1485022753,-0.2201791555,-0.3223364353,0.2954417169,-0.3488661051,-0.1378353983,0.1312768757,0.3829959631,0.1777834296,-0.1002675742,-0.052053079,0.2365419269,-0.0244584288,0.0325130597,0.0145486901,-0.1853820533,-0.0120613975,-0.279640317,0.1465234011,0.0235882532,0.348398447,0.0077597764,-0.3682603538,0.377537787,-0.0371985137,0.0586498044,-0.0751952454,-0.0948600769,0.3058986366,-0.2762518227,-0.0493951924,-0.0236396324,0.2183714211,0.0369678438,0.1693601906,0.04129586,-0.3924960494,0.047164917,0.3549182117,-0.072537221,-0.1731103212,0.0018169068,0.0174868628,-0.0824427828,0.164626956,0.2774141133,0.3839050233,0.2610786259,0.0662332624,0.1700644791,-0.2135237604,-0.1831107736,-0.054769367,-0.0941865519,-0.027594151,0.3564697206,0.1723606884,-0.1707368493,-0.3508794904,0.0280967355,0.155304268,0.3651978076,-0.1970971674,0.2237509936,-0.2249526829,-0.2432209104,-0.3664650917,-0.0820040554,0.0938911885,-0.3452377021,-0.1890164614,0.1978519261,-0.107149139,-0.1096806899,-0.1519416273,-0.0034873104,0.0685041398,-0.1828078628,0.0319254473,-0.29726246,-0.4607281685,0.1170991212,0.3691140115,0.3623128831,0.4312880337,0.0150138075,-0.1073622257,0.0393667519,-0.0697979704,0.076720871,-0.100895606,0.4542248547,-0.0569396578,0.0416780375,-0.2283503711,-0.4602437317,0.5433725715,-0.0930568278,-0.1630839258,0.003641258,0.1127770245,-0.2605376542,-0.2423969209,-0.5966646671,-0.376578927,-0.3775201142,0.2393841147,0.2145242095,0.2878387868,0.3024253845,0.0545422919,0.227584064,0.2486261129,0.0220368877,-0.1577897519,-0.2745801508,0.1092553139,-0.3437816501,-0.3863990009,-0.0418258309,-0.2934792042,0.2532537878,-0.177119717,-0.5053718686,0.106159687,-0.2085334063,0.175987035,0.1183275729,-0.0756837502,0.3462190628,0.0716120824,-0.0953156427,-0.1507954448,-0.1724995971,0.17420201,0.0590066873,0.0601394363,0.060319189,0.2799823284,-0.1291005462,0.4796222448,0.0056985766,0.0647447631,0.2293583602,-0.1257217079,0.0351171568,-0.0269690529,-0.5060185194,0.0983495489,0.1094464958,-0.324074626,0.239359051,-0.0309577566,-0.332950443,-0.2473927289,0.0161873512,-0.5706443191,-0.3236039579,0.0986715853,-0.0077600973,0.1682422012,0.0484663807,-0.0216584858,-0.1972861886,-0.0092988918,-0.1825920939,0.2409341633,0.0875305682,0.0187797509,-0.3264673948,-0.0633295923,-0.5391075611,0.1027067751,0.0346259587,0.3136318624,-0.1360362172,-0.1588585377,0.2374024689,-0.1259416044,0.355168581,-0.1997641325,0.1949021965,0.2806993425,0.0566505939,-0.2779286504,0.0598577335,-0.0828077719,0.194408074,0.0551685654,0.1930948049,-0.0358043946,-0.2404814959,0.3322658539,0.086165607,-0.3055275679,-0.4114955664,-0.1551193148,-0.2620900869,-0.1998126209,-0.0636808574,-0.2033237219,0.3930025399,-0.0087458072,-0.0178830903,0.0776314586,0.0366086811,0.1659790725,0.0659319609,0.144485414,0.1646545529,0.1758186221,0.0877508596,0.4288354814,-0.1880310923,0.6513575315,-0.0744920522,-0.1693410277,0.0129821599,-0.1879096776,0.2662377656,0.2502310276,-0.1011858135,-0.2100187987,0.1666356325,-0.111804314,-0.1770980358,0.394836545,0.107521981,-0.1158342734,-0.0676748827,-0.5222005844,0.3812000751,-0.114734821,-0.0565005615,-0.0064784149,-0.1909905374,-0.1881240904,0.5036737323,0.1323155314,0.864397049,-0.1980929971,0.0349103101,0.4668910205,0.007784619,0.1646999121,-0.1840792,0.2341465205,-0.5701379776,0.019398775,0.0076868804,-0.0287059639,0.1016457453,0.2519910634,-0.0923879817,0.0385436043,0.1525755972,-0.1554612368,-0.1020387262,0.283041954,0.0544179939,-0.1383448094,-0.3233302832,0.2222985327,-0.0313629843,0.1917642057,0.0116054919,-0.2589574754,-0.1316404045,-0.1493984312,0.0529658459,0.1751309335,-0.1544432491,-0.0379619598,-0.0486805253,-0.3220744729,0.0414558649,0.3110913336,0.1537470967,0.2356782109,-0.2747209668,-0.1066584215,-0.0653964505,-0.1012914628,-0.1185328737,0.0051975329,0.1416031569,0.0174175873,-0.4252744019,-0.0209827181,-0.0953727588,-0.1059336886,0.128332004,0.1576503962,0.2031957209,-0.3334443569,-0.406455636,0.0477226153,-0.1087376252,-0.0600517206,0.2278128713,-0.021711776,-0.1207553223,-0.107445024,0.0386740565,-0.2580270767,0.0760376602,0.3284548521,0.261800617,0.1757760346,0.3383722305,0.1642873436,-0.0343141928,-0.3920804858,-0.0356164388,0.0704319179,-0.2614100575,-0.1708102077,0.1431283057,0.0959377959,0.2589941323,0.3358263373,0.0868119672,0.0314114615,-0.0264190324,-0.4197938442,-0.4787349105,-0.156475544,-0.0960422233,0.2361431718,0.3992731869,0.350523293,0.0761040971,0.1712295115,-0.4348987043,-0.1582422704,-0.3672046959,0.0972431526,0.10911531,-0.1289702207,0.0431214049,-0.0544742681,0.231937632,-0.0266497321,-0.1715811938,-0.3512310386,0.0337690935,0.0571922138,0.1569814384,-0.1603884697,-0.126356557,-0.1384025216,-0.037456084,-0.0434973948,-0.0566055775,0.1242901683,-0.0668676496,0.4345211685,-0.0098284902,0.1701683551,-0.0747041106,0.0009316503,-0.2399915755,0.1899886429,-0.0329511464,0.0423244759,0.1744868606,-0.0874913782,-0.150878042,-0.0646028668,-0.0523427799,0.3434982598,0.4014250636,-0.3132493496,-0.0029764506,0.1609406322,0.3720909357,0.1368874609,-0.0667902604,0.1870818138,0.3331491053,0.3209904134,-0.2170785218,-0.0078470316,0.1768234521,0.0243657287,0.1571125537,0.2079402208,0.1602826715,-0.1185155734,-0.0839414075,0.0089579439,0.219958663,0.0269431118,-0.172602579,0.5578839779,-0.106967777,0.0147087779,0.0153596047,0.2626159787,0.1503474712,0.3625153303,0.0332685411,-0.0691114068,-0.0515283495,0.3558601439,-0.2451538295,-0.492842555,0.2119537294,0.1049965546,-0.138162747,0.0064495583,-0.1462257951,0.2908702791,0.0504665002,-0.045047041,-0.0586939566,0.4341705739,-0.2295295745,-0.1355336159,-0.441393733,-0.0720626786,-0.2986844182,-0.0694366246,0.0210932959,0.1438496411,-0.0314354375,0.0074021979,-0.0485220365,-0.1798207313,0.1309196651,0.2044498771,-0.136657387,-0.3524982631,0.3145772517,0.1482188404,0.1692373157,-0.1363149136,0.1530724317,0.2172753513,0.4162555933,-0.19093889,0.0166432783,-0.1435080916,-0.1691758782,-0.3510366082,0.2251788527,0.0193293262,0.3033943474,0.5757574439,0.2563101649,-0.1848143041,0.2783590257,0.1571856737,0.0825569481,0.0694978014,0.2987569869,0.157189101,-0.1272757649,-0.2587654889,-0.1263581514,-0.4825097919,-0.0734842867,0.337828815,0.0948294923,0.5023238063,-0.2306268215,0.1396726817,-0.1246094182,0.4588422775,0.269364357,0.1685478389,-0.4026035964,-0.2152403295,-0.5258612633,0.0776555389,-0.0867535397,-0.2213983834,-0.0802540183,0.2547663152,0.0406547263,0.1811081916,-0.5375036001,0.1512551457,-0.1450660825,0.2304359972,-0.0383698754,-0.1621061862,-0.1003011614,-0.0384932198,-0.0739434585,-0.4356856048,0.1343759298,-0.106498912,0.2026229054,-0.2383623719,0.0076444671,-0.0351861045,0.0809182376,0.03337726,0.2150260955,0.3726360202,-0.24440898,0.2414368838,-0.2125802338,0.0813362673,-0.2655959129,0.1747843325,0.1388967186,0.5628277659,-0.2568778694,0.2340295315,-0.3165405989,0.0263359509,0.0514674932,0.0794786662,-0.0276819076,0.1196670309,-0.3101905584,0.1019593179,0.1867545545,0.158185035,0.0220817178,0.1333450824,-0.2682808638,-0.2907093763,0.4097493589,-0.2822920084,-0.2055113912,0.1762141585,0.1532312185,-0.0029166152,-0.0402859598,-0.4433356225,0.0254867394,0.3729545176,-0.1966754198,0.0635944456,0.2106608301,0.0638403893,0.0155626731,-0.0240852218,-0.1181434765,0.1259868592,-0.0181906056,-0.1313410401,-0.2844596803]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2","title":"Issue to read a local dataset","comments":"Ok great, so as discussed today, let's:\r\n- have a main dataset directory inside the lib with sub-directories hashed by the content of the file\r\n- keep a cache for downloading the scripts from S3 for now\r\n- later: add methods to list and clean the local versions of the datasets (and the distant versions on S3 as well)\r\n\r\nSide question: do you often use `builder_kwargs` for other things than supplying file paths? I was thinking about having a more easy to read and remember `data_files` argument maybe.","body":"Hello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)","comment_length":88,"text":"Issue to read a local dataset\nHello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)\nOk great, so as discussed today, let's:\r\n- have a main dataset directory inside the lib with sub-directories hashed by the content of the file\r\n- keep a cache for downloading the scripts from S3 for now\r\n- later: add methods to list and clean the local versions of the datasets (and the distant versions on S3 as well)\r\n\r\nSide question: do you often use `builder_kwargs` for other things than supplying file paths? I was thinking about having a more easy to read and remember `data_files` argument maybe.","embeddings":[-0.2719613612,0.1254656762,-0.0766852647,0.256811142,0.3318964839,0.0312691405,0.2642410994,0.3003315628,0.3266820014,0.2098683715,0.1191096306,0.2447595894,-0.2942002714,0.1220210195,0.1568034738,-0.1723625809,-0.1275069416,0.3326410651,0.1501982808,-0.3661575317,-0.27604267,0.1413950026,-0.1191338152,0.1533625871,-0.2043232173,0.0607061796,-0.036645785,0.4012748301,-0.2069403231,-0.2449603528,0.1894469708,0.0119974464,0.2054079175,0.122409679,-0.0001004647,0.1904105544,0.2995985448,0.0436597355,-0.1575898379,-0.364929229,-0.0587465651,-0.3323357999,0.2688184083,-0.3986444175,-0.3326914608,-0.1401537061,0.0905135646,-0.4378168285,0.4685541093,0.3634622395,0.3068140745,0.1048855186,-0.0699259415,0.1463657767,0.195211947,0.152147904,-0.0873053744,0.2656306922,0.031915091,-0.2012286931,-0.1960862428,0.1223725975,-0.2222957164,0.0449407585,0.2718259394,0.1370849758,0.182052508,-0.1032298878,-0.032771349,-0.0018344095,0.3769168258,-0.2492007166,-0.2248517573,-0.13559407,0.0572931319,-0.2591866851,0.0394319892,0.342464596,-0.349896282,0.3455857933,-0.2791967094,0.1082817465,-0.3008818924,0.2382965088,0.1193012074,0.3104308248,-0.076890111,0.1048834771,0.0907402039,0.2240543067,0.1786538064,-0.1133791059,0.2185333967,0.2561581731,-0.0302851815,0.0967433602,0.3372043669,0.0830674171,-0.0235993508,-0.0186839495,0.1047985032,0.0249034651,0.0130941896,0.3297557831,0.1096323282,0.1400410384,0.1735097319,0.087680243,0.0717137232,-0.0622796975,-0.2097682357,0.0256421678,-0.1461680084,-0.283438772,0.0297457296,0.2404010743,0.4468615353,-0.1885111183,-0.1420501471,0.2193927914,-0.2050481737,-0.0884669423,0.1696829945,0.4059306085,-0.0357793234,0.2412724048,0.1384448409,0.0997360051,-0.3508428931,-0.1973964125,-0.3274970055,-0.0029654084,-0.2937141359,-0.0887341276,0.3137897849,0.0290612634,0.3492821157,-0.0983311981,-0.2490224689,-0.006291463,0.1585656404,-0.2560243607,0.1404405981,0.138631627,0.0651841164,0.034821,0.2272037119,-0.4230361283,-0.1299738437,0.2165416181,0.0924713016,-0.4230499268,-0.1779252738,0.3226445317,-0.0838615969,0.1769313514,0.180574581,0.1790369749,0.1407811642,-0.036249388,-0.2524119318,-0.0215898771,-0.459605664,-0.2430503964,0.2107181102,0.3516464829,-0.4640030563,0.0853260383,-0.164036572,-0.1117871031,0.3510994613,0.2487842441,-0.3300900161,0.4454449415,-0.1017358303,0.1409201324,0.6580390334,-0.3004144132,-0.3807192743,0.6322368979,-0.3527038991,-0.1465917528,0.0498830862,0.0073254914,-0.0018453643,0.2033673376,0.2166674733,0.3950749636,-0.1738853306,0.1002998799,-0.4054657817,-0.152086556,0.1437031329,0.1029602066,-0.0783379376,0.1322496831,0.1525419801,-0.05174575,0.497027427,-0.1389278173,0.1745269895,0.3185678422,0.2489011884,-0.0089441948,-0.1572516412,-0.0782212839,-0.2163548321,0.04727301,-0.1540113091,-0.0111839129,-0.1268217266,-0.1966494024,-0.3752773702,0.1024699733,-0.4161543548,-0.0244457163,0.2532124221,0.2463398874,0.0947190672,0.0092416257,-0.0782992616,-0.1027992368,-0.1485022753,-0.2201791555,-0.3223364353,0.2954417169,-0.3488661051,-0.1378353983,0.1312768757,0.3829959631,0.1777834296,-0.1002675742,-0.052053079,0.2365419269,-0.0244584288,0.0325130597,0.0145486901,-0.1853820533,-0.0120613975,-0.279640317,0.1465234011,0.0235882532,0.348398447,0.0077597764,-0.3682603538,0.377537787,-0.0371985137,0.0586498044,-0.0751952454,-0.0948600769,0.3058986366,-0.2762518227,-0.0493951924,-0.0236396324,0.2183714211,0.0369678438,0.1693601906,0.04129586,-0.3924960494,0.047164917,0.3549182117,-0.072537221,-0.1731103212,0.0018169068,0.0174868628,-0.0824427828,0.164626956,0.2774141133,0.3839050233,0.2610786259,0.0662332624,0.1700644791,-0.2135237604,-0.1831107736,-0.054769367,-0.0941865519,-0.027594151,0.3564697206,0.1723606884,-0.1707368493,-0.3508794904,0.0280967355,0.155304268,0.3651978076,-0.1970971674,0.2237509936,-0.2249526829,-0.2432209104,-0.3664650917,-0.0820040554,0.0938911885,-0.3452377021,-0.1890164614,0.1978519261,-0.107149139,-0.1096806899,-0.1519416273,-0.0034873104,0.0685041398,-0.1828078628,0.0319254473,-0.29726246,-0.4607281685,0.1170991212,0.3691140115,0.3623128831,0.4312880337,0.0150138075,-0.1073622257,0.0393667519,-0.0697979704,0.076720871,-0.100895606,0.4542248547,-0.0569396578,0.0416780375,-0.2283503711,-0.4602437317,0.5433725715,-0.0930568278,-0.1630839258,0.003641258,0.1127770245,-0.2605376542,-0.2423969209,-0.5966646671,-0.376578927,-0.3775201142,0.2393841147,0.2145242095,0.2878387868,0.3024253845,0.0545422919,0.227584064,0.2486261129,0.0220368877,-0.1577897519,-0.2745801508,0.1092553139,-0.3437816501,-0.3863990009,-0.0418258309,-0.2934792042,0.2532537878,-0.177119717,-0.5053718686,0.106159687,-0.2085334063,0.175987035,0.1183275729,-0.0756837502,0.3462190628,0.0716120824,-0.0953156427,-0.1507954448,-0.1724995971,0.17420201,0.0590066873,0.0601394363,0.060319189,0.2799823284,-0.1291005462,0.4796222448,0.0056985766,0.0647447631,0.2293583602,-0.1257217079,0.0351171568,-0.0269690529,-0.5060185194,0.0983495489,0.1094464958,-0.324074626,0.239359051,-0.0309577566,-0.332950443,-0.2473927289,0.0161873512,-0.5706443191,-0.3236039579,0.0986715853,-0.0077600973,0.1682422012,0.0484663807,-0.0216584858,-0.1972861886,-0.0092988918,-0.1825920939,0.2409341633,0.0875305682,0.0187797509,-0.3264673948,-0.0633295923,-0.5391075611,0.1027067751,0.0346259587,0.3136318624,-0.1360362172,-0.1588585377,0.2374024689,-0.1259416044,0.355168581,-0.1997641325,0.1949021965,0.2806993425,0.0566505939,-0.2779286504,0.0598577335,-0.0828077719,0.194408074,0.0551685654,0.1930948049,-0.0358043946,-0.2404814959,0.3322658539,0.086165607,-0.3055275679,-0.4114955664,-0.1551193148,-0.2620900869,-0.1998126209,-0.0636808574,-0.2033237219,0.3930025399,-0.0087458072,-0.0178830903,0.0776314586,0.0366086811,0.1659790725,0.0659319609,0.144485414,0.1646545529,0.1758186221,0.0877508596,0.4288354814,-0.1880310923,0.6513575315,-0.0744920522,-0.1693410277,0.0129821599,-0.1879096776,0.2662377656,0.2502310276,-0.1011858135,-0.2100187987,0.1666356325,-0.111804314,-0.1770980358,0.394836545,0.107521981,-0.1158342734,-0.0676748827,-0.5222005844,0.3812000751,-0.114734821,-0.0565005615,-0.0064784149,-0.1909905374,-0.1881240904,0.5036737323,0.1323155314,0.864397049,-0.1980929971,0.0349103101,0.4668910205,0.007784619,0.1646999121,-0.1840792,0.2341465205,-0.5701379776,0.019398775,0.0076868804,-0.0287059639,0.1016457453,0.2519910634,-0.0923879817,0.0385436043,0.1525755972,-0.1554612368,-0.1020387262,0.283041954,0.0544179939,-0.1383448094,-0.3233302832,0.2222985327,-0.0313629843,0.1917642057,0.0116054919,-0.2589574754,-0.1316404045,-0.1493984312,0.0529658459,0.1751309335,-0.1544432491,-0.0379619598,-0.0486805253,-0.3220744729,0.0414558649,0.3110913336,0.1537470967,0.2356782109,-0.2747209668,-0.1066584215,-0.0653964505,-0.1012914628,-0.1185328737,0.0051975329,0.1416031569,0.0174175873,-0.4252744019,-0.0209827181,-0.0953727588,-0.1059336886,0.128332004,0.1576503962,0.2031957209,-0.3334443569,-0.406455636,0.0477226153,-0.1087376252,-0.0600517206,0.2278128713,-0.021711776,-0.1207553223,-0.107445024,0.0386740565,-0.2580270767,0.0760376602,0.3284548521,0.261800617,0.1757760346,0.3383722305,0.1642873436,-0.0343141928,-0.3920804858,-0.0356164388,0.0704319179,-0.2614100575,-0.1708102077,0.1431283057,0.0959377959,0.2589941323,0.3358263373,0.0868119672,0.0314114615,-0.0264190324,-0.4197938442,-0.4787349105,-0.156475544,-0.0960422233,0.2361431718,0.3992731869,0.350523293,0.0761040971,0.1712295115,-0.4348987043,-0.1582422704,-0.3672046959,0.0972431526,0.10911531,-0.1289702207,0.0431214049,-0.0544742681,0.231937632,-0.0266497321,-0.1715811938,-0.3512310386,0.0337690935,0.0571922138,0.1569814384,-0.1603884697,-0.126356557,-0.1384025216,-0.037456084,-0.0434973948,-0.0566055775,0.1242901683,-0.0668676496,0.4345211685,-0.0098284902,0.1701683551,-0.0747041106,0.0009316503,-0.2399915755,0.1899886429,-0.0329511464,0.0423244759,0.1744868606,-0.0874913782,-0.150878042,-0.0646028668,-0.0523427799,0.3434982598,0.4014250636,-0.3132493496,-0.0029764506,0.1609406322,0.3720909357,0.1368874609,-0.0667902604,0.1870818138,0.3331491053,0.3209904134,-0.2170785218,-0.0078470316,0.1768234521,0.0243657287,0.1571125537,0.2079402208,0.1602826715,-0.1185155734,-0.0839414075,0.0089579439,0.219958663,0.0269431118,-0.172602579,0.5578839779,-0.106967777,0.0147087779,0.0153596047,0.2626159787,0.1503474712,0.3625153303,0.0332685411,-0.0691114068,-0.0515283495,0.3558601439,-0.2451538295,-0.492842555,0.2119537294,0.1049965546,-0.138162747,0.0064495583,-0.1462257951,0.2908702791,0.0504665002,-0.045047041,-0.0586939566,0.4341705739,-0.2295295745,-0.1355336159,-0.441393733,-0.0720626786,-0.2986844182,-0.0694366246,0.0210932959,0.1438496411,-0.0314354375,0.0074021979,-0.0485220365,-0.1798207313,0.1309196651,0.2044498771,-0.136657387,-0.3524982631,0.3145772517,0.1482188404,0.1692373157,-0.1363149136,0.1530724317,0.2172753513,0.4162555933,-0.19093889,0.0166432783,-0.1435080916,-0.1691758782,-0.3510366082,0.2251788527,0.0193293262,0.3033943474,0.5757574439,0.2563101649,-0.1848143041,0.2783590257,0.1571856737,0.0825569481,0.0694978014,0.2987569869,0.157189101,-0.1272757649,-0.2587654889,-0.1263581514,-0.4825097919,-0.0734842867,0.337828815,0.0948294923,0.5023238063,-0.2306268215,0.1396726817,-0.1246094182,0.4588422775,0.269364357,0.1685478389,-0.4026035964,-0.2152403295,-0.5258612633,0.0776555389,-0.0867535397,-0.2213983834,-0.0802540183,0.2547663152,0.0406547263,0.1811081916,-0.5375036001,0.1512551457,-0.1450660825,0.2304359972,-0.0383698754,-0.1621061862,-0.1003011614,-0.0384932198,-0.0739434585,-0.4356856048,0.1343759298,-0.106498912,0.2026229054,-0.2383623719,0.0076444671,-0.0351861045,0.0809182376,0.03337726,0.2150260955,0.3726360202,-0.24440898,0.2414368838,-0.2125802338,0.0813362673,-0.2655959129,0.1747843325,0.1388967186,0.5628277659,-0.2568778694,0.2340295315,-0.3165405989,0.0263359509,0.0514674932,0.0794786662,-0.0276819076,0.1196670309,-0.3101905584,0.1019593179,0.1867545545,0.158185035,0.0220817178,0.1333450824,-0.2682808638,-0.2907093763,0.4097493589,-0.2822920084,-0.2055113912,0.1762141585,0.1532312185,-0.0029166152,-0.0402859598,-0.4433356225,0.0254867394,0.3729545176,-0.1966754198,0.0635944456,0.2106608301,0.0638403893,0.0155626731,-0.0240852218,-0.1181434765,0.1259868592,-0.0181906056,-0.1313410401,-0.2844596803]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2","title":"Issue to read a local dataset","comments":"Good plan!\r\n\r\nYes I do use `builder_kwargs` for other things such as:\r\n- dataset name\r\n- properties to know how to properly read a CSV file: do I have to skip the first line in a CSV, which delimiter is used, and the columns ids to use.\r\n- properties to know how to properly read a JSON file: which properties in a JSON object to read","body":"Hello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)","comment_length":66,"text":"Issue to read a local dataset\nHello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)\nGood plan!\r\n\r\nYes I do use `builder_kwargs` for other things such as:\r\n- dataset name\r\n- properties to know how to properly read a CSV file: do I have to skip the first line in a CSV, which delimiter is used, and the columns ids to use.\r\n- properties to know how to properly read a JSON file: which properties in a JSON object to read","embeddings":[-0.2719613612,0.1254656762,-0.0766852647,0.256811142,0.3318964839,0.0312691405,0.2642410994,0.3003315628,0.3266820014,0.2098683715,0.1191096306,0.2447595894,-0.2942002714,0.1220210195,0.1568034738,-0.1723625809,-0.1275069416,0.3326410651,0.1501982808,-0.3661575317,-0.27604267,0.1413950026,-0.1191338152,0.1533625871,-0.2043232173,0.0607061796,-0.036645785,0.4012748301,-0.2069403231,-0.2449603528,0.1894469708,0.0119974464,0.2054079175,0.122409679,-0.0001004647,0.1904105544,0.2995985448,0.0436597355,-0.1575898379,-0.364929229,-0.0587465651,-0.3323357999,0.2688184083,-0.3986444175,-0.3326914608,-0.1401537061,0.0905135646,-0.4378168285,0.4685541093,0.3634622395,0.3068140745,0.1048855186,-0.0699259415,0.1463657767,0.195211947,0.152147904,-0.0873053744,0.2656306922,0.031915091,-0.2012286931,-0.1960862428,0.1223725975,-0.2222957164,0.0449407585,0.2718259394,0.1370849758,0.182052508,-0.1032298878,-0.032771349,-0.0018344095,0.3769168258,-0.2492007166,-0.2248517573,-0.13559407,0.0572931319,-0.2591866851,0.0394319892,0.342464596,-0.349896282,0.3455857933,-0.2791967094,0.1082817465,-0.3008818924,0.2382965088,0.1193012074,0.3104308248,-0.076890111,0.1048834771,0.0907402039,0.2240543067,0.1786538064,-0.1133791059,0.2185333967,0.2561581731,-0.0302851815,0.0967433602,0.3372043669,0.0830674171,-0.0235993508,-0.0186839495,0.1047985032,0.0249034651,0.0130941896,0.3297557831,0.1096323282,0.1400410384,0.1735097319,0.087680243,0.0717137232,-0.0622796975,-0.2097682357,0.0256421678,-0.1461680084,-0.283438772,0.0297457296,0.2404010743,0.4468615353,-0.1885111183,-0.1420501471,0.2193927914,-0.2050481737,-0.0884669423,0.1696829945,0.4059306085,-0.0357793234,0.2412724048,0.1384448409,0.0997360051,-0.3508428931,-0.1973964125,-0.3274970055,-0.0029654084,-0.2937141359,-0.0887341276,0.3137897849,0.0290612634,0.3492821157,-0.0983311981,-0.2490224689,-0.006291463,0.1585656404,-0.2560243607,0.1404405981,0.138631627,0.0651841164,0.034821,0.2272037119,-0.4230361283,-0.1299738437,0.2165416181,0.0924713016,-0.4230499268,-0.1779252738,0.3226445317,-0.0838615969,0.1769313514,0.180574581,0.1790369749,0.1407811642,-0.036249388,-0.2524119318,-0.0215898771,-0.459605664,-0.2430503964,0.2107181102,0.3516464829,-0.4640030563,0.0853260383,-0.164036572,-0.1117871031,0.3510994613,0.2487842441,-0.3300900161,0.4454449415,-0.1017358303,0.1409201324,0.6580390334,-0.3004144132,-0.3807192743,0.6322368979,-0.3527038991,-0.1465917528,0.0498830862,0.0073254914,-0.0018453643,0.2033673376,0.2166674733,0.3950749636,-0.1738853306,0.1002998799,-0.4054657817,-0.152086556,0.1437031329,0.1029602066,-0.0783379376,0.1322496831,0.1525419801,-0.05174575,0.497027427,-0.1389278173,0.1745269895,0.3185678422,0.2489011884,-0.0089441948,-0.1572516412,-0.0782212839,-0.2163548321,0.04727301,-0.1540113091,-0.0111839129,-0.1268217266,-0.1966494024,-0.3752773702,0.1024699733,-0.4161543548,-0.0244457163,0.2532124221,0.2463398874,0.0947190672,0.0092416257,-0.0782992616,-0.1027992368,-0.1485022753,-0.2201791555,-0.3223364353,0.2954417169,-0.3488661051,-0.1378353983,0.1312768757,0.3829959631,0.1777834296,-0.1002675742,-0.052053079,0.2365419269,-0.0244584288,0.0325130597,0.0145486901,-0.1853820533,-0.0120613975,-0.279640317,0.1465234011,0.0235882532,0.348398447,0.0077597764,-0.3682603538,0.377537787,-0.0371985137,0.0586498044,-0.0751952454,-0.0948600769,0.3058986366,-0.2762518227,-0.0493951924,-0.0236396324,0.2183714211,0.0369678438,0.1693601906,0.04129586,-0.3924960494,0.047164917,0.3549182117,-0.072537221,-0.1731103212,0.0018169068,0.0174868628,-0.0824427828,0.164626956,0.2774141133,0.3839050233,0.2610786259,0.0662332624,0.1700644791,-0.2135237604,-0.1831107736,-0.054769367,-0.0941865519,-0.027594151,0.3564697206,0.1723606884,-0.1707368493,-0.3508794904,0.0280967355,0.155304268,0.3651978076,-0.1970971674,0.2237509936,-0.2249526829,-0.2432209104,-0.3664650917,-0.0820040554,0.0938911885,-0.3452377021,-0.1890164614,0.1978519261,-0.107149139,-0.1096806899,-0.1519416273,-0.0034873104,0.0685041398,-0.1828078628,0.0319254473,-0.29726246,-0.4607281685,0.1170991212,0.3691140115,0.3623128831,0.4312880337,0.0150138075,-0.1073622257,0.0393667519,-0.0697979704,0.076720871,-0.100895606,0.4542248547,-0.0569396578,0.0416780375,-0.2283503711,-0.4602437317,0.5433725715,-0.0930568278,-0.1630839258,0.003641258,0.1127770245,-0.2605376542,-0.2423969209,-0.5966646671,-0.376578927,-0.3775201142,0.2393841147,0.2145242095,0.2878387868,0.3024253845,0.0545422919,0.227584064,0.2486261129,0.0220368877,-0.1577897519,-0.2745801508,0.1092553139,-0.3437816501,-0.3863990009,-0.0418258309,-0.2934792042,0.2532537878,-0.177119717,-0.5053718686,0.106159687,-0.2085334063,0.175987035,0.1183275729,-0.0756837502,0.3462190628,0.0716120824,-0.0953156427,-0.1507954448,-0.1724995971,0.17420201,0.0590066873,0.0601394363,0.060319189,0.2799823284,-0.1291005462,0.4796222448,0.0056985766,0.0647447631,0.2293583602,-0.1257217079,0.0351171568,-0.0269690529,-0.5060185194,0.0983495489,0.1094464958,-0.324074626,0.239359051,-0.0309577566,-0.332950443,-0.2473927289,0.0161873512,-0.5706443191,-0.3236039579,0.0986715853,-0.0077600973,0.1682422012,0.0484663807,-0.0216584858,-0.1972861886,-0.0092988918,-0.1825920939,0.2409341633,0.0875305682,0.0187797509,-0.3264673948,-0.0633295923,-0.5391075611,0.1027067751,0.0346259587,0.3136318624,-0.1360362172,-0.1588585377,0.2374024689,-0.1259416044,0.355168581,-0.1997641325,0.1949021965,0.2806993425,0.0566505939,-0.2779286504,0.0598577335,-0.0828077719,0.194408074,0.0551685654,0.1930948049,-0.0358043946,-0.2404814959,0.3322658539,0.086165607,-0.3055275679,-0.4114955664,-0.1551193148,-0.2620900869,-0.1998126209,-0.0636808574,-0.2033237219,0.3930025399,-0.0087458072,-0.0178830903,0.0776314586,0.0366086811,0.1659790725,0.0659319609,0.144485414,0.1646545529,0.1758186221,0.0877508596,0.4288354814,-0.1880310923,0.6513575315,-0.0744920522,-0.1693410277,0.0129821599,-0.1879096776,0.2662377656,0.2502310276,-0.1011858135,-0.2100187987,0.1666356325,-0.111804314,-0.1770980358,0.394836545,0.107521981,-0.1158342734,-0.0676748827,-0.5222005844,0.3812000751,-0.114734821,-0.0565005615,-0.0064784149,-0.1909905374,-0.1881240904,0.5036737323,0.1323155314,0.864397049,-0.1980929971,0.0349103101,0.4668910205,0.007784619,0.1646999121,-0.1840792,0.2341465205,-0.5701379776,0.019398775,0.0076868804,-0.0287059639,0.1016457453,0.2519910634,-0.0923879817,0.0385436043,0.1525755972,-0.1554612368,-0.1020387262,0.283041954,0.0544179939,-0.1383448094,-0.3233302832,0.2222985327,-0.0313629843,0.1917642057,0.0116054919,-0.2589574754,-0.1316404045,-0.1493984312,0.0529658459,0.1751309335,-0.1544432491,-0.0379619598,-0.0486805253,-0.3220744729,0.0414558649,0.3110913336,0.1537470967,0.2356782109,-0.2747209668,-0.1066584215,-0.0653964505,-0.1012914628,-0.1185328737,0.0051975329,0.1416031569,0.0174175873,-0.4252744019,-0.0209827181,-0.0953727588,-0.1059336886,0.128332004,0.1576503962,0.2031957209,-0.3334443569,-0.406455636,0.0477226153,-0.1087376252,-0.0600517206,0.2278128713,-0.021711776,-0.1207553223,-0.107445024,0.0386740565,-0.2580270767,0.0760376602,0.3284548521,0.261800617,0.1757760346,0.3383722305,0.1642873436,-0.0343141928,-0.3920804858,-0.0356164388,0.0704319179,-0.2614100575,-0.1708102077,0.1431283057,0.0959377959,0.2589941323,0.3358263373,0.0868119672,0.0314114615,-0.0264190324,-0.4197938442,-0.4787349105,-0.156475544,-0.0960422233,0.2361431718,0.3992731869,0.350523293,0.0761040971,0.1712295115,-0.4348987043,-0.1582422704,-0.3672046959,0.0972431526,0.10911531,-0.1289702207,0.0431214049,-0.0544742681,0.231937632,-0.0266497321,-0.1715811938,-0.3512310386,0.0337690935,0.0571922138,0.1569814384,-0.1603884697,-0.126356557,-0.1384025216,-0.037456084,-0.0434973948,-0.0566055775,0.1242901683,-0.0668676496,0.4345211685,-0.0098284902,0.1701683551,-0.0747041106,0.0009316503,-0.2399915755,0.1899886429,-0.0329511464,0.0423244759,0.1744868606,-0.0874913782,-0.150878042,-0.0646028668,-0.0523427799,0.3434982598,0.4014250636,-0.3132493496,-0.0029764506,0.1609406322,0.3720909357,0.1368874609,-0.0667902604,0.1870818138,0.3331491053,0.3209904134,-0.2170785218,-0.0078470316,0.1768234521,0.0243657287,0.1571125537,0.2079402208,0.1602826715,-0.1185155734,-0.0839414075,0.0089579439,0.219958663,0.0269431118,-0.172602579,0.5578839779,-0.106967777,0.0147087779,0.0153596047,0.2626159787,0.1503474712,0.3625153303,0.0332685411,-0.0691114068,-0.0515283495,0.3558601439,-0.2451538295,-0.492842555,0.2119537294,0.1049965546,-0.138162747,0.0064495583,-0.1462257951,0.2908702791,0.0504665002,-0.045047041,-0.0586939566,0.4341705739,-0.2295295745,-0.1355336159,-0.441393733,-0.0720626786,-0.2986844182,-0.0694366246,0.0210932959,0.1438496411,-0.0314354375,0.0074021979,-0.0485220365,-0.1798207313,0.1309196651,0.2044498771,-0.136657387,-0.3524982631,0.3145772517,0.1482188404,0.1692373157,-0.1363149136,0.1530724317,0.2172753513,0.4162555933,-0.19093889,0.0166432783,-0.1435080916,-0.1691758782,-0.3510366082,0.2251788527,0.0193293262,0.3033943474,0.5757574439,0.2563101649,-0.1848143041,0.2783590257,0.1571856737,0.0825569481,0.0694978014,0.2987569869,0.157189101,-0.1272757649,-0.2587654889,-0.1263581514,-0.4825097919,-0.0734842867,0.337828815,0.0948294923,0.5023238063,-0.2306268215,0.1396726817,-0.1246094182,0.4588422775,0.269364357,0.1685478389,-0.4026035964,-0.2152403295,-0.5258612633,0.0776555389,-0.0867535397,-0.2213983834,-0.0802540183,0.2547663152,0.0406547263,0.1811081916,-0.5375036001,0.1512551457,-0.1450660825,0.2304359972,-0.0383698754,-0.1621061862,-0.1003011614,-0.0384932198,-0.0739434585,-0.4356856048,0.1343759298,-0.106498912,0.2026229054,-0.2383623719,0.0076444671,-0.0351861045,0.0809182376,0.03337726,0.2150260955,0.3726360202,-0.24440898,0.2414368838,-0.2125802338,0.0813362673,-0.2655959129,0.1747843325,0.1388967186,0.5628277659,-0.2568778694,0.2340295315,-0.3165405989,0.0263359509,0.0514674932,0.0794786662,-0.0276819076,0.1196670309,-0.3101905584,0.1019593179,0.1867545545,0.158185035,0.0220817178,0.1333450824,-0.2682808638,-0.2907093763,0.4097493589,-0.2822920084,-0.2055113912,0.1762141585,0.1532312185,-0.0029166152,-0.0402859598,-0.4433356225,0.0254867394,0.3729545176,-0.1966754198,0.0635944456,0.2106608301,0.0638403893,0.0155626731,-0.0240852218,-0.1181434765,0.1259868592,-0.0181906056,-0.1313410401,-0.2844596803]} {"html_url":"https:\/\/github.com\/huggingface\/datasets\/issues\/2","title":"Issue to read a local dataset","comments":"Done!","body":"Hello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)","comment_length":1,"text":"Issue to read a local dataset\nHello,\r\n\r\nAs proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:\r\n```python\r\nimport os\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass BbcConfig(nlp.BuilderConfig):\r\n def __init__(self, **kwargs):\r\n super(BbcConfig, self).__init__(**kwargs)\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n _DIR = \".\/data\"\r\n _DEV_FILE = \"test.csv\"\r\n _TRAINING_FILE = \"train.csv\"\r\n\r\n BUILDER_CONFIGS = [BbcConfig(name=\"bbc\", version=nlp.Version(\"1.0.0\"))]\r\n\r\n def _info(self):\r\n return nlp.DatasetInfo(builder=self, features=nlp.features.FeaturesDict({\"id\": nlp.string, \"text\": nlp.string, \"label\": nlp.string}))\r\n\r\n def _split_generators(self, dl_manager):\r\n files = {\"train\": os.path.join(self._DIR, self._TRAINING_FILE), \"dev\": os.path.join(self._DIR, self._DEV_FILE)}\r\n\r\n return [nlp.SplitGenerator(name=nlp.Split.TRAIN, gen_kwargs={\"filepath\": files[\"train\"]}),\r\n nlp.SplitGenerator(name=nlp.Split.VALIDATION, gen_kwargs={\"filepath\": files[\"dev\"]})]\r\n\r\n def _generate_examples(self, filepath):\r\n with open(filepath) as f:\r\n reader = csv.reader(f, delimiter=',', quotechar=\"\\\"\")\r\n lines = list(reader)[1:]\r\n\r\n for idx, line in enumerate(lines):\r\n yield idx, {\"idx\": idx, \"text\": line[1], \"label\": line[0]}\r\n\r\n```\r\n\r\nThe dataset is attached to this issue as well:\r\n[data.zip](https:\/\/github.com\/huggingface\/datasets\/files\/4476928\/data.zip)\r\n\r\nNow the steps to reproduce what I would like to do:\r\n1. unzip data locally (I know the nlp lib can detect and extract archives but I want to reduce and facilitate the reproduction as much as possible)\r\n2. create the `bbc.py` script as above at the same location than the unziped `data` folder.\r\n\r\nNow I try to load the dataset in three different ways and none works, the first one with the name of the dataset like I would do with TFDS:\r\n```python\r\nimport nlp\r\nfrom bbc import Bbc\r\ndataset = nlp.load(\"bbc\")\r\n```\r\n\r\nI get:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 88, in load_dataset\r\n local_files_only=local_files_only,\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/utils\/file_utils.py\", line 214, in cached_path\r\n if not is_zipfile(output_path) and not tarfile.is_tarfile(output_path):\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/zipfile.py\", line 203, in is_zipfile\r\n with open(filename, \"rb\") as fp:\r\nTypeError: expected str, bytes or os.PathLike object, not NoneType\r\n```\r\n\r\nBut @thomwolf told me that no need to import the script, just put the path of it, then I tried three different way to do:\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\".\/bbc.py\")\r\n```\r\nAnd\r\n```python\r\nimport nlp\r\ndataset = nlp.load(\"\/absolute\/path\/to\/bbc.py\")\r\n```\r\n\r\nThese three ways gives me:\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 280, in load\r\n dbuilder: DatasetBuilder = builder(path, name, data_dir=data_dir, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 166, in builder\r\n builder_cls = load_dataset(path, name=name, **builder_kwargs)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/site-packages\/nlp\/load.py\", line 124, in load_dataset\r\n dataset_module = importlib.import_module(module_path)\r\n File \"\/opt\/anaconda3\/envs\/transformers\/lib\/python3.7\/importlib\/__init__.py\", line 127, in import_module\r\n return _bootstrap._gcd_import(name[level:], package, level)\r\n File \"<frozen importlib._bootstrap>\", line 1006, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 983, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 965, in _find_and_load_unlocked\r\nModuleNotFoundError: No module named 'nlp.datasets.2fd72627d92c328b3e9c4a3bf7ec932c48083caca09230cebe4c618da6e93688.bbc'\r\n```\r\nAny idea of what I'm missing? or I might have spot a bug :)\nDone!","embeddings":[-0.2719613612,0.1254656762,-0.0766852647,0.256811142,0.3318964839,0.0312691405,0.2642410994,0.3003315628,0.3266820014,0.2098683715,0.1191096306,0.2447595894,-0.2942002714,0.1220210195,0.1568034738,-0.1723625809,-0.1275069416,0.3326410651,0.1501982808,-0.3661575317,-0.27604267,0.1413950026,-0.1191338152,0.1533625871,-0.2043232173,0.0607061796,-0.036645785,0.4012748301,-0.2069403231,-0.2449603528,0.1894469708,0.0119974464,0.2054079175,0.122409679,-0.0001004647,0.1904105544,0.2995985448,0.0436597355,-0.1575898379,-0.364929229,-0.0587465651,-0.3323357999,0.2688184083,-0.3986444175,-0.3326914608,-0.1401537061,0.0905135646,-0.4378168285,0.4685541093,0.3634622395,0.3068140745,0.1048855186,-0.0699259415,0.1463657767,0.195211947,0.152147904,-0.0873053744,0.2656306922,0.031915091,-0.2012286931,-0.1960862428,0.1223725975,-0.2222957164,0.0449407585,0.2718259394,0.1370849758,0.182052508,-0.1032298878,-0.032771349,-0.0018344095,0.3769168258,-0.2492007166,-0.2248517573,-0.13559407,0.0572931319,-0.2591866851,0.0394319892,0.342464596,-0.349896282,0.3455857933,-0.2791967094,0.1082817465,-0.3008818924,0.2382965088,0.1193012074,0.3104308248,-0.076890111,0.1048834771,0.0907402039,0.2240543067,0.1786538064,-0.1133791059,0.2185333967,0.2561581731,-0.0302851815,0.0967433602,0.3372043669,0.0830674171,-0.0235993508,-0.0186839495,0.1047985032,0.0249034651,0.0130941896,0.3297557831,0.1096323282,0.1400410384,0.1735097319,0.087680243,0.0717137232,-0.0622796975,-0.2097682357,0.0256421678,-0.1461680084,-0.283438772,0.0297457296,0.2404010743,0.4468615353,-0.1885111183,-0.1420501471,0.2193927914,-0.2050481737,-0.0884669423,0.1696829945,0.4059306085,-0.0357793234,0.2412724048,0.1384448409,0.0997360051,-0.3508428931,-0.1973964125,-0.3274970055,-0.0029654084,-0.2937141359,-0.0887341276,0.3137897849,0.0290612634,0.3492821157,-0.0983311981,-0.2490224689,-0.006291463,0.1585656404,-0.2560243607,0.1404405981,0.138631627,0.0651841164,0.034821,0.2272037119,-0.4230361283,-0.1299738437,0.2165416181,0.0924713016,-0.4230499268,-0.1779252738,0.3226445317,-0.0838615969,0.1769313514,0.180574581,0.1790369749,0.1407811642,-0.036249388,-0.2524119318,-0.0215898771,-0.459605664,-0.2430503964,0.2107181102,0.3516464829,-0.4640030563,0.0853260383,-0.164036572,-0.1117871031,0.3510994613,0.2487842441,-0.3300900161,0.4454449415,-0.1017358303,0.1409201324,0.6580390334,-0.3004144132,-0.3807192743,0.6322368979,-0.3527038991,-0.1465917528,0.0498830862,0.0073254914,-0.0018453643,0.2033673376,0.2166674733,0.3950749636,-0.1738853306,0.1002998799,-0.4054657817,-0.152086556,0.1437031329,0.1029602066,-0.0783379376,0.1322496831,0.1525419801,-0.05174575,0.497027427,-0.1389278173,0.1745269895,0.3185678422,0.2489011884,-0.0089441948,-0.1572516412,-0.0782212839,-0.2163548321,0.04727301,-0.1540113091,-0.0111839129,-0.1268217266,-0.1966494024,-0.3752773702,0.1024699733,-0.4161543548,-0.0244457163,0.2532124221,0.2463398874,0.0947190672,0.0092416257,-0.0782992616,-0.1027992368,-0.1485022753,-0.2201791555,-0.3223364353,0.2954417169,-0.3488661051,-0.1378353983,0.1312768757,0.3829959631,0.1777834296,-0.1002675742,-0.052053079,0.2365419269,-0.0244584288,0.0325130597,0.0145486901,-0.1853820533,-0.0120613975,-0.279640317,0.1465234011,0.0235882532,0.348398447,0.0077597764,-0.3682603538,0.377537787,-0.0371985137,0.0586498044,-0.0751952454,-0.0948600769,0.3058986366,-0.2762518227,-0.0493951924,-0.0236396324,0.2183714211,0.0369678438,0.1693601906,0.04129586,-0.3924960494,0.047164917,0.3549182117,-0.072537221,-0.1731103212,0.0018169068,0.0174868628,-0.0824427828,0.164626956,0.2774141133,0.3839050233,0.2610786259,0.0662332624,0.1700644791,-0.2135237604,-0.1831107736,-0.054769367,-0.0941865519,-0.027594151,0.3564697206,0.1723606884,-0.1707368493,-0.3508794904,0.0280967355,0.155304268,0.3651978076,-0.1970971674,0.2237509936,-0.2249526829,-0.2432209104,-0.3664650917,-0.0820040554,0.0938911885,-0.3452377021,-0.1890164614,0.1978519261,-0.107149139,-0.1096806899,-0.1519416273,-0.0034873104,0.0685041398,-0.1828078628,0.0319254473,-0.29726246,-0.4607281685,0.1170991212,0.3691140115,0.3623128831,0.4312880337,0.0150138075,-0.1073622257,0.0393667519,-0.0697979704,0.076720871,-0.100895606,0.4542248547,-0.0569396578,0.0416780375,-0.2283503711,-0.4602437317,0.5433725715,-0.0930568278,-0.1630839258,0.003641258,0.1127770245,-0.2605376542,-0.2423969209,-0.5966646671,-0.376578927,-0.3775201142,0.2393841147,0.2145242095,0.2878387868,0.3024253845,0.0545422919,0.227584064,0.2486261129,0.0220368877,-0.1577897519,-0.2745801508,0.1092553139,-0.3437816501,-0.3863990009,-0.0418258309,-0.2934792042,0.2532537878,-0.177119717,-0.5053718686,0.106159687,-0.2085334063,0.175987035,0.1183275729,-0.0756837502,0.3462190628,0.0716120824,-0.0953156427,-0.1507954448,-0.1724995971,0.17420201,0.0590066873,0.0601394363,0.060319189,0.2799823284,-0.1291005462,0.4796222448,0.0056985766,0.0647447631,0.2293583602,-0.1257217079,0.0351171568,-0.0269690529,-0.5060185194,0.0983495489,0.1094464958,-0.324074626,0.239359051,-0.0309577566,-0.332950443,-0.2473927289,0.0161873512,-0.5706443191,-0.3236039579,0.0986715853,-0.0077600973,0.1682422012,0.0484663807,-0.0216584858,-0.1972861886,-0.0092988918,-0.1825920939,0.2409341633,0.0875305682,0.0187797509,-0.3264673948,-0.0633295923,-0.5391075611,0.1027067751,0.0346259587,0.3136318624,-0.1360362172,-0.1588585377,0.2374024689,-0.1259416044,0.355168581,-0.1997641325,0.1949021965,0.2806993425,0.0566505939,-0.2779286504,0.0598577335,-0.0828077719,0.194408074,0.0551685654,0.1930948049,-0.0358043946,-0.2404814959,0.3322658539,0.086165607,-0.3055275679,-0.4114955664,-0.1551193148,-0.2620900869,-0.1998126209,-0.0636808574,-0.2033237219,0.3930025399,-0.0087458072,-0.0178830903,0.0776314586,0.0366086811,0.1659790725,0.0659319609,0.144485414,0.1646545529,0.1758186221,0.0877508596,0.4288354814,-0.1880310923,0.6513575315,-0.0744920522,-0.1693410277,0.0129821599,-0.1879096776,0.2662377656,0.2502310276,-0.1011858135,-0.2100187987,0.1666356325,-0.111804314,-0.1770980358,0.394836545,0.107521981,-0.1158342734,-0.0676748827,-0.5222005844,0.3812000751,-0.114734821,-0.0565005615,-0.0064784149,-0.1909905374,-0.1881240904,0.5036737323,0.1323155314,0.864397049,-0.1980929971,0.0349103101,0.4668910205,0.007784619,0.1646999121,-0.1840792,0.2341465205,-0.5701379776,0.019398775,0.0076868804,-0.0287059639,0.1016457453,0.2519910634,-0.0923879817,0.0385436043,0.1525755972,-0.1554612368,-0.1020387262,0.283041954,0.0544179939,-0.1383448094,-0.3233302832,0.2222985327,-0.0313629843,0.1917642057,0.0116054919,-0.2589574754,-0.1316404045,-0.1493984312,0.0529658459,0.1751309335,-0.1544432491,-0.0379619598,-0.0486805253,-0.3220744729,0.0414558649,0.3110913336,0.1537470967,0.2356782109,-0.2747209668,-0.1066584215,-0.0653964505,-0.1012914628,-0.1185328737,0.0051975329,0.1416031569,0.0174175873,-0.4252744019,-0.0209827181,-0.0953727588,-0.1059336886,0.128332004,0.1576503962,0.2031957209,-0.3334443569,-0.406455636,0.0477226153,-0.1087376252,-0.0600517206,0.2278128713,-0.021711776,-0.1207553223,-0.107445024,0.0386740565,-0.2580270767,0.0760376602,0.3284548521,0.261800617,0.1757760346,0.3383722305,0.1642873436,-0.0343141928,-0.3920804858,-0.0356164388,0.0704319179,-0.2614100575,-0.1708102077,0.1431283057,0.0959377959,0.2589941323,0.3358263373,0.0868119672,0.0314114615,-0.0264190324,-0.4197938442,-0.4787349105,-0.156475544,-0.0960422233,0.2361431718,0.3992731869,0.350523293,0.0761040971,0.1712295115,-0.4348987043,-0.1582422704,-0.3672046959,0.0972431526,0.10911531,-0.1289702207,0.0431214049,-0.0544742681,0.231937632,-0.0266497321,-0.1715811938,-0.3512310386,0.0337690935,0.0571922138,0.1569814384,-0.1603884697,-0.126356557,-0.1384025216,-0.037456084,-0.0434973948,-0.0566055775,0.1242901683,-0.0668676496,0.4345211685,-0.0098284902,0.1701683551,-0.0747041106,0.0009316503,-0.2399915755,0.1899886429,-0.0329511464,0.0423244759,0.1744868606,-0.0874913782,-0.150878042,-0.0646028668,-0.0523427799,0.3434982598,0.4014250636,-0.3132493496,-0.0029764506,0.1609406322,0.3720909357,0.1368874609,-0.0667902604,0.1870818138,0.3331491053,0.3209904134,-0.2170785218,-0.0078470316,0.1768234521,0.0243657287,0.1571125537,0.2079402208,0.1602826715,-0.1185155734,-0.0839414075,0.0089579439,0.219958663,0.0269431118,-0.172602579,0.5578839779,-0.106967777,0.0147087779,0.0153596047,0.2626159787,0.1503474712,0.3625153303,0.0332685411,-0.0691114068,-0.0515283495,0.3558601439,-0.2451538295,-0.492842555,0.2119537294,0.1049965546,-0.138162747,0.0064495583,-0.1462257951,0.2908702791,0.0504665002,-0.045047041,-0.0586939566,0.4341705739,-0.2295295745,-0.1355336159,-0.441393733,-0.0720626786,-0.2986844182,-0.0694366246,0.0210932959,0.1438496411,-0.0314354375,0.0074021979,-0.0485220365,-0.1798207313,0.1309196651,0.2044498771,-0.136657387,-0.3524982631,0.3145772517,0.1482188404,0.1692373157,-0.1363149136,0.1530724317,0.2172753513,0.4162555933,-0.19093889,0.0166432783,-0.1435080916,-0.1691758782,-0.3510366082,0.2251788527,0.0193293262,0.3033943474,0.5757574439,0.2563101649,-0.1848143041,0.2783590257,0.1571856737,0.0825569481,0.0694978014,0.2987569869,0.157189101,-0.1272757649,-0.2587654889,-0.1263581514,-0.4825097919,-0.0734842867,0.337828815,0.0948294923,0.5023238063,-0.2306268215,0.1396726817,-0.1246094182,0.4588422775,0.269364357,0.1685478389,-0.4026035964,-0.2152403295,-0.5258612633,0.0776555389,-0.0867535397,-0.2213983834,-0.0802540183,0.2547663152,0.0406547263,0.1811081916,-0.5375036001,0.1512551457,-0.1450660825,0.2304359972,-0.0383698754,-0.1621061862,-0.1003011614,-0.0384932198,-0.0739434585,-0.4356856048,0.1343759298,-0.106498912,0.2026229054,-0.2383623719,0.0076444671,-0.0351861045,0.0809182376,0.03337726,0.2150260955,0.3726360202,-0.24440898,0.2414368838,-0.2125802338,0.0813362673,-0.2655959129,0.1747843325,0.1388967186,0.5628277659,-0.2568778694,0.2340295315,-0.3165405989,0.0263359509,0.0514674932,0.0794786662,-0.0276819076,0.1196670309,-0.3101905584,0.1019593179,0.1867545545,0.158185035,0.0220817178,0.1333450824,-0.2682808638,-0.2907093763,0.4097493589,-0.2822920084,-0.2055113912,0.1762141585,0.1532312185,-0.0029166152,-0.0402859598,-0.4433356225,0.0254867394,0.3729545176,-0.1966754198,0.0635944456,0.2106608301,0.0638403893,0.0155626731,-0.0240852218,-0.1181434765,0.1259868592,-0.0181906056,-0.1313410401,-0.2844596803]}